From 8ebc90992de88bb86331f670950e869b94a4befa Mon Sep 17 00:00:00 2001 From: Liu Yiqun Date: Tue, 20 Jan 2026 10:31:38 +0800 Subject: [PATCH 01/10] Enable to record tolerance for models. --- .../subgraph_decompose_and_evaluation_step.py | 64 ++++++++++++++++++- graph_net_bench/analysis_util.py | 24 +++++++ 2 files changed, 87 insertions(+), 1 deletion(-) diff --git a/graph_net/subgraph_decompose_and_evaluation_step.py b/graph_net/subgraph_decompose_and_evaluation_step.py index a6223d8b6..80b3fe881 100755 --- a/graph_net/subgraph_decompose_and_evaluation_step.py +++ b/graph_net/subgraph_decompose_and_evaluation_step.py @@ -9,7 +9,7 @@ import glob from dataclasses import dataclass, field, asdict from typing import List, Dict -from graph_net_bench.analysis_util import get_incorrect_models +from graph_net_bench.analysis_util import get_incorrect_models, get_min_passed_tolerance from graph_net.graph_net_root import get_graphnet_root from graph_net_bench import path_utils @@ -297,6 +297,64 @@ def update_running_state_with_incorrect_models( ) +class ToleranceRecord: + model_name2subgraph_tolerance_record = {} + filename = "tolerance_record.json" + + @classmethod + def load(cls, pass_id, output_dir): + if pass_id >= 0: + work_dir = get_decompose_workspace_path(output_dir, pass_id) + filepath = os.path.join(work_dir, cls.filename) + with open(filepath, "r") as f: + data = json.load(f) + cls.model_name2subgraph_tolerance_record = data + + @classmethod + def save(cls, pass_id, output_dir): + work_dir = get_decompose_workspace_path(output_dir, pass_id) + filepath = os.path.join(work_dir, cls.filename) + print(f"Save tolerance record to: {filepath}.") + with open(filepath, "w") as f: + json.dump(cls.model_name2subgraph_tolerance_record, f, indent=4) + + @classmethod + def update(cls, pass_id, output_dir, decompose_config, log_path): + cls.load(pass_id - 1, output_dir) + + subgraph_path2tolerance = get_min_passed_tolerance(log_path) + running_state = decompose_config.get_running_state(pass_id) + for subgraph_path, tolerance in subgraph_path2tolerance.items(): + model_name, subgraph_idx = extract_model_name_and_subgraph_idx( + subgraph_path + ) + if model_name not in running_state.model_name2record: + continue + + split_positions = running_state.model_name2record[ + model_name + ].get_split_positions(decompose_config.decompose_method) + assert len(split_positions) >= 2 + subgraph_split_point = int(split_positions[1]) + if model_name not in cls.model_name2subgraph_tolerance_record: + cls.model_name2subgraph_tolerance_record[model_name] = {} + cls.model_name2subgraph_tolerance_record[model_name][ + subgraph_split_point + ] = tolerance + + cls.model_name2subgraph_tolerance_record = dict( + sorted(cls.model_name2subgraph_tolerance_record.items()) + ) + for ( + model_name, + subgraph_tolerance_record, + ) in cls.model_name2subgraph_tolerance_record.items(): + cls.model_name2subgraph_tolerance_record[model_name] = dict( + sorted(subgraph_tolerance_record.items(), key=lambda x: int(x[0])) + ) + cls.save(pass_id, output_dir) + + def get_rectfied_model_path(model_path): graphnet_root = get_graphnet_root() return os.path.join(graphnet_root, model_path.split("GraphNet/")[-1]) @@ -792,6 +850,10 @@ def main(args): print_incorrect_models( decompose_config, current_pass_id, log_prompt="[Analysis]" ) + + ToleranceRecord.update( + current_pass_id, base_output_dir, decompose_config, log_path + ) print_summary_and_suggestion(decompose_config, current_pass_id) # --- Step 5: Save States --- diff --git a/graph_net_bench/analysis_util.py b/graph_net_bench/analysis_util.py index d8df6d2dd..8e1c7b74f 100644 --- a/graph_net_bench/analysis_util.py +++ b/graph_net_bench/analysis_util.py @@ -488,3 +488,27 @@ def get_incorrect_models( failed_models.add(sample.get("model_path")) if not is_correct else None return failed_models + + +def get_min_passed_tolerance(log_file_path: str, type: str = "ESt") -> int: + model_path2tolerance = {} + samples = parse_logs_to_data(log_file_path) + + for sample in samples: + model_path = sample.get("model_path") + for tolerance in range(-10, 5, 1): + is_correct, fail_type = check_sample_correctness(sample, tolerance) + if is_correct: + model_path2tolerance[model_path] = tolerance + break + + if type == "ESt": + for sample in samples: + model_path = sample.get("model_path") + if ( + model_path not in model_path2tolerance + or model_path2tolerance[model_path] > 1 + ): + model_path2tolerance[model_path] = 1 + + return model_path2tolerance From 7f3c8d901f239f85c7ad371bbd4f32d6bb3d4a7a Mon Sep 17 00:00:00 2001 From: Liu Yiqun Date: Tue, 20 Jan 2026 10:32:27 +0800 Subject: [PATCH 02/10] Update auto_fault_locator.py. --- graph_net/auto_fault_locator.py | 83 ++++++++++++++++++++++----------- 1 file changed, 56 insertions(+), 27 deletions(-) diff --git a/graph_net/auto_fault_locator.py b/graph_net/auto_fault_locator.py index cf541ccb7..979fb80a0 100644 --- a/graph_net/auto_fault_locator.py +++ b/graph_net/auto_fault_locator.py @@ -23,9 +23,9 @@ def __init__(self, args): self.machine = args.machine self.port = args.port - def get_one_step_cmd(self, config_str): - config_b64 = convert_json_to_b64_string(config_str) - return [ + def execute_one_step_cmd(self, test_config): + test_config_b64_str = convert_json_to_b64_string(test_config) + cmd = [ sys.executable, "-m", "graph_net.subgraph_decompose_and_evaluation_step", @@ -36,7 +36,7 @@ def get_one_step_cmd(self, config_str): "--framework", self.framework, "--test-config", - config_b64, + test_config_b64_str, "--decompose-method", self.decompose_method, "--tolerance", @@ -45,49 +45,78 @@ def get_one_step_cmd(self, config_str): self.max_subgraph_size, ] - def run_remote_test_reference(self): + print(f"[AutoFaultLocator] Executing: {' '.join(cmd)}", flush=True) + result = subprocess.run(cmd, check=True, text=True) + return result + + def run_test_reference_device(self, is_remote): print( - "\n>>> [Step 1] Run Remote Reference Device (Decomposition And Evaluation)\n" + "\n>>> [AutoFaultLocator 2/1] Run Test Reference Device (Decomposition And Evaluation)\n", + flush=True, ) - test_remote_reference_device_config_str = { - "test_module_name": "test_remote_reference_device", - "test_remote_reference_device_arguments": { + test_module_name = ( + "test_remote_reference_device" if is_remote else "test_reference_device" + ) + test_reference_device_config = { + "test_module_name": test_module_name, + f"{test_module_name}_arguments": { "model-path": None, "reference-dir": None, "compiler": "nope", "device": self.reference_device, - "op-lib": "default", "warmup": 5, "trials": 20, "seed": 123, - "machine": self.machine, - "port": self.port, }, } - - cmd = self.get_one_step_cmd(test_remote_reference_device_config_str) - print(f"Executing: {' '.join(cmd)}") - result = subprocess.run(cmd, check=True, text=True) + if args.framework == "torch": + test_reference_device_config[f"{test_module_name}_arguments"].update( + {"op-lib": "default"} + ) + if is_remote: + test_reference_device_config[f"{test_module_name}_arguments"].update( + { + "machine": self.machine, + "port": self.port, + } + ) + + result = self.execute_one_step_cmd(test_reference_device_config) assert ( result.returncode == 0 ), f"Run Remote Reference Device failed with return code {result.returncode}" - def run_local_test_target(self): - print("\n>>> [Step 2] Run Local Target Device (Evaluation And Analysis)\n") + def run_test_target_device(self, is_remote): + print( + "\n>>> [AutoFaultLocator 2/2] Run Test Target Device (Evaluation And Analysis)\n", + flush=True, + ) - test_target_device_config_str = { - "test_module_name": "test_target_device", - "test_target_device_arguments": { + test_module_name = ( + "test_remote_target_device" if is_remote else "test_target_device" + ) + test_target_device_config = { + "test_module_name": test_module_name, + f"{test_module_name}_arguments": { "model-path": None, "reference-dir": None, + "compiler": "nope", "device": self.target_device, + "warmup": 5, + "trials": 20, + "seed": 123, }, } - - cmd = self.get_one_step_cmd(test_target_device_config_str) - print(f"Executing: {' '.join(cmd)}") - result = subprocess.run(cmd, check=True, text=True) + if is_remote: + test_target_device_config[f"{test_module_name}_arguments"].update( + { + "machine": self.machine, + "port": self.port, + } + ) + + result = self.execute_one_step_cmd(test_target_device_config) assert ( result.returncode == 0 ), f"Run Local Target Device failed with return code {result.returncode}" @@ -114,8 +143,8 @@ def analyze_and_decide_next(self): def main(args): locator = AutoFaultLocator(args) while True: - locator.run_remote_test_reference() - locator.run_local_test_target() + locator.run_test_reference_device(is_remote=False) + locator.run_test_target_device(is_remote=True) should_continue = locator.analyze_and_decide_next() if not should_continue: break From 5ec032d51023de38f55ef04a7924a96876df9e41 Mon Sep 17 00:00:00 2001 From: Liu Yiqun Date: Tue, 20 Jan 2026 15:21:39 +0800 Subject: [PATCH 03/10] Fix prophetnet-large-uncased. --- graph_net/test/paddle_nlp_model_getter.py | 40 +- graph_net/test/paddle_nlp_test.py | 6 + .../prophetnet-large-uncased/input_meta.py | 34 +- .../prophetnet-large-uncased/model.py | 1432 ++++--- .../prophetnet-large-uncased/weight_meta.py | 3710 ++++++++++++----- 5 files changed, 3470 insertions(+), 1752 deletions(-) diff --git a/graph_net/test/paddle_nlp_model_getter.py b/graph_net/test/paddle_nlp_model_getter.py index eb24dabb1..43e39286c 100644 --- a/graph_net/test/paddle_nlp_model_getter.py +++ b/graph_net/test/paddle_nlp_model_getter.py @@ -1,3 +1,6 @@ +# Reference implementation: https://github.com/PaddlePaddle/PaddleTest/tree/develop/framework/e2e/PaddleLT_new/layerNLPcase/transformers + + def get_auto_model_and_inputs(model_name, text, dtype): from paddlenlp.transformers import AutoConfig, AutoModelForCausalLM, AutoTokenizer @@ -139,18 +142,18 @@ def get_xlnet_model_and_inputs(model_name, text, dtype): tokenizer = XLNetTokenizer.from_pretrained(model_name) - enc = tokenizer( + inputs = tokenizer( text, return_tensors="pd", padding=True, truncation=True, # max_length=512, ) - if "attention_mask" not in enc: - input_ids = enc["input_ids"] + if "attention_mask" not in inputs: + input_ids = inputs["input_ids"] pad_id = tokenizer.pad_token_id - enc["attention_mask"] = (input_ids != pad_id).astype("int64") - return model, enc + inputs["attention_mask"] = (input_ids != pad_id).astype("int64") + return model, inputs def get_fnet_model_and_inputs(model_name, text, dtype): @@ -163,3 +166,30 @@ def get_fnet_model_and_inputs(model_name, text, dtype): tokenizer = FNetTokenizer.from_pretrained(model_name) inputs = tokenizer(text, return_tensors="pd") return model, inputs + + +def get_prophetnet_model_and_inputs(model_name, text, dtype): + import paddle + from paddlenlp.transformers import ProphetNetModel, ProphetNetConfig + from paddlenlp.transformers import ProphetNetTokenizer + + config = ProphetNetConfig.from_pretrained(model_name) + model = ProphetNetModel(config) + + tokenizer = ProphetNetTokenizer.from_pretrained(model_name) + inputs = tokenizer(text, return_tensors="pd") + inputs.pop("token_type_ids", None) + + if "attention_mask" not in inputs: + input_ids = inputs["input_ids"] + pad_id = tokenizer.pad_token_id + inputs["attention_mask"] = (input_ids != pad_id).astype("int64") + + if "decoder_input_ids" not in inputs: + batch_size = inputs["input_ids"].shape[0] + decoder_input_ids = paddle.full( + [batch_size, 1], tokenizer.bos_token_id, dtype="int64" + ) + inputs["decoder_input_ids"] = decoder_input_ids + + return model, inputs diff --git a/graph_net/test/paddle_nlp_test.py b/graph_net/test/paddle_nlp_test.py index 9495c2acc..87d020dc8 100644 --- a/graph_net/test/paddle_nlp_test.py +++ b/graph_net/test/paddle_nlp_test.py @@ -275,6 +275,12 @@ def extract_fnet_models(text_en, text_cn): process_model(model_name, nlp_model_getter.get_fnet_model_and_inputs, text_en) +def extract_prophetnet_models(text_en, text_cn): + # prophetnet-large-uncased models + model_name = "prophetnet-large-uncased" + process_model(model_name, nlp_model_getter.get_prophetnet_model_and_inputs, text_en) + + def main(): current_dir = os.path.dirname(os.path.abspath(__file__)) dump_dir = os.path.join(current_dir, "dump") diff --git a/paddle_samples/PaddleNLP/prophetnet-large-uncased/input_meta.py b/paddle_samples/PaddleNLP/prophetnet-large-uncased/input_meta.py index 544223cfa..e009ae95f 100644 --- a/paddle_samples/PaddleNLP/prophetnet-large-uncased/input_meta.py +++ b/paddle_samples/PaddleNLP/prophetnet-large-uncased/input_meta.py @@ -1,19 +1,43 @@ class Program_weight_tensor_data_0: name = "data_0" - shape = [1, 10] + original_name = "input_ids" + shape = [1, 20] dtype = "int64" - data = [100, 100, 100, 100, 100, 100, 100, 100, 999, 102] + data = [ + 7592, + 1010, + 2026, + 2171, + 2003, + 3960, + 1012, + 1045, + 2572, + 4083, + 2055, + 2312, + 2653, + 4275, + 1998, + 2037, + 4294, + 2015, + 1012, + 102, + ] class Program_weight_tensor_data_1: name = "data_1" - shape = [1, 10] + original_name = "attention_mask" + shape = [1, 20] dtype = "int64" - data = [1, 1, 1, 1, 1, 1, 1, 1, 1, 1] + data = [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1] class Program_weight_tensor_data_2: name = "data_2" + original_name = "decoder_input_ids" shape = [1, 1] dtype = "int64" - data = [0] + data = [102] diff --git a/paddle_samples/PaddleNLP/prophetnet-large-uncased/model.py b/paddle_samples/PaddleNLP/prophetnet-large-uncased/model.py index b883dc171..e3f59cd30 100644 --- a/paddle_samples/PaddleNLP/prophetnet-large-uncased/model.py +++ b/paddle_samples/PaddleNLP/prophetnet-large-uncased/model.py @@ -564,18 +564,18 @@ def forward( cast_0 = paddle._C_ops.cast(not_equal_0, paddle.float32) del not_equal_0 - # pd_op.embedding: (1x10x1024xf32) <- (1x10xi64, 30522x1024xf32) + # pd_op.embedding: (1x20x1024xf32) <- (1x20xi64, 30522x1024xf32) embedding_0 = paddle._C_ops.embedding(data_0, parameter_539, 0, False) del data_0 # pd_op.full_int_array: (1xi64) <- () full_int_array_0 = [1] - # pd_op.unsqueeze: (1x1x10xi64) <- (1x10xi64, 1xi64) + # pd_op.unsqueeze: (1x1x20xi64) <- (1x20xi64, 1xi64) unsqueeze_0 = paddle._C_ops.unsqueeze(data_1, full_int_array_0) del data_1 - # pd_op.cast: (1x1x10xf32) <- (1x1x10xi64) + # pd_op.cast: (1x1x20xf32) <- (1x1x20xi64) cast_1 = paddle._C_ops.cast(unsqueeze_0, paddle.float32) # pd_op.full: (1xf32) <- () @@ -583,14 +583,14 @@ def forward( [1], float("-1"), paddle.float32, paddle.core.CPUPlace() ) - # pd_op.scale: (1x1x10xf32) <- (1x1x10xf32, 1xf32) + # pd_op.scale: (1x1x20xf32) <- (1x1x20xf32, 1xf32) scale_0 = paddle._C_ops.scale(cast_1, full_1, float("1"), True) del cast_1 # pd_op.full_int_array: (3xi64) <- () full_int_array_1 = [16, 1, 1] - # pd_op.tile: (16x1x10xf32) <- (1x1x10xf32, 3xi64) + # pd_op.tile: (16x1x20xf32) <- (1x1x20xf32, 3xi64) tile_0 = paddle._C_ops.tile(scale_0, full_int_array_1) del scale_0 @@ -599,17 +599,17 @@ def forward( [1], float("-10000"), paddle.float32, paddle.core.CPUPlace() ) - # pd_op.scale: (16x1x10xf32) <- (16x1x10xf32, 1xf32) + # pd_op.scale: (16x1x20xf32) <- (16x1x20xf32, 1xf32) scale_1 = paddle._C_ops.scale(tile_0, full_2, float("0"), True) del tile_0 - # pd_op.cast: (16x1x10xf32) <- (16x1x10xf32) + # pd_op.cast: (16x1x20xf32) <- (16x1x20xf32) cast_2 = paddle._C_ops.cast(scale_1, paddle.float32) del scale_1 - # pd_op.full: (1x10xi64) <- () + # pd_op.full: (1x20xi64) <- () full_3 = paddle._C_ops.full( - [1, 10], + [1, 20], float("1"), paddle.int64, paddle.framework._current_expected_place(), @@ -620,18 +620,18 @@ def forward( [1], float("1"), paddle.int32, paddle.core.CPUPlace() ) - # pd_op.cumsum: (1x10xi64) <- (1x10xi64, 1xi32) + # pd_op.cumsum: (1x20xi64) <- (1x20xi64, 1xi32) cumsum_0 = paddle._C_ops.cumsum(full_3, full_4, False, False, False) - # pd_op.cast: (1x10xi64) <- (1x10xi64) + # pd_op.cast: (1x20xi64) <- (1x20xi64) cast_3 = paddle._C_ops.cast(cumsum_0, paddle.int64) del cumsum_0 - # pd_op.multiply: (1x10xi64) <- (1x10xi64, 1x10xi64) + # pd_op.multiply: (1x20xi64) <- (1x20xi64, 1x20xi64) multiply_0 = paddle._C_ops.multiply(cast_3, full_3) del cast_3, full_3 - # pd_op.cast: (1x10xi64) <- (1x10xi64) + # pd_op.cast: (1x20xi64) <- (1x20xi64) cast_4 = paddle._C_ops.cast(multiply_0, paddle.int64) del multiply_0 @@ -640,7 +640,7 @@ def forward( [1], float("1"), paddle.float32, paddle.core.CPUPlace() ) - # pd_op.scale: (1x10xi64) <- (1x10xi64, 1xf32) + # pd_op.scale: (1x20xi64) <- (1x20xi64, 1xf32) scale_2 = paddle._C_ops.scale(cast_4, full_5, float("0"), True) del cast_4 @@ -654,19 +654,19 @@ def forward( [1], float("511"), paddle.float32, paddle.core.CPUPlace() ) - # pd_op.clip: (1x10xi64) <- (1x10xi64, 1xf32, 1xf32) + # pd_op.clip: (1x20xi64) <- (1x20xi64, 1xf32, 1xf32) clip_0 = paddle._C_ops.clip(scale_2, full_6, full_7) del scale_2 - # pd_op.embedding: (1x10x1024xf32) <- (1x10xi64, 512x1024xf32) + # pd_op.embedding: (1x20x1024xf32) <- (1x20xi64, 512x1024xf32) embedding_1 = paddle._C_ops.embedding(clip_0, parameter_538, 0, False) del clip_0, parameter_538 - # pd_op.add: (1x10x1024xf32) <- (1x10x1024xf32, 1x10x1024xf32) + # pd_op.add: (1x20x1024xf32) <- (1x20x1024xf32, 1x20x1024xf32) add_0 = paddle._C_ops.add(embedding_0, embedding_1) del embedding_0, embedding_1 - # pd_op.layer_norm: (1x10x1024xf32, 1x10xf32, 1x10xf32) <- (1x10x1024xf32, 1024xf32, 1024xf32) + # pd_op.layer_norm: (1x20x1024xf32, 1x20xf32, 1x20xf32) <- (1x20x1024xf32, 1024xf32, 1024xf32) layer_norm_0, layer_norm_1, layer_norm_2 = (lambda x, f: f(x))( paddle._C_ops.layer_norm( add_0, parameter_537, parameter_536, float("1e-05"), 2 @@ -680,7 +680,7 @@ def forward( [1], float("0.1"), paddle.float32, paddle.core.CPUPlace() ) - # pd_op.dropout: (1x10x1024xf32, 1x10x1024xui8) <- (1x10x1024xf32, None, 1xf32) + # pd_op.dropout: (1x20x1024xf32, 1x20x1024xui8) <- (1x20x1024xf32, None, 1xf32) dropout_0, dropout_1 = (lambda x, f: f(x))( paddle._C_ops.dropout( layer_norm_0, None, full_8, True, "upscale_in_train", 0, False @@ -689,11 +689,11 @@ def forward( ) del layer_norm_0 - # pd_op.matmul: (1x10x1024xf32) <- (1x10x1024xf32, 1024x1024xf32) + # pd_op.matmul: (1x20x1024xf32) <- (1x20x1024xf32, 1024x1024xf32) matmul_0 = paddle._C_ops.matmul(dropout_0, parameter_531, False, False) del parameter_531 - # pd_op.add: (1x10x1024xf32) <- (1x10x1024xf32, 1024xf32) + # pd_op.add: (1x20x1024xf32) <- (1x20x1024xf32, 1024xf32) add_1 = paddle._C_ops.add(matmul_0, parameter_530) del matmul_0, parameter_530 @@ -702,88 +702,88 @@ def forward( [1], float("0.125"), paddle.float32, paddle.core.CPUPlace() ) - # pd_op.scale: (1x10x1024xf32) <- (1x10x1024xf32, 1xf32) + # pd_op.scale: (1x20x1024xf32) <- (1x20x1024xf32, 1xf32) scale_3 = paddle._C_ops.scale(add_1, full_9, float("0"), True) del add_1 - # pd_op.matmul: (1x10x1024xf32) <- (1x10x1024xf32, 1024x1024xf32) + # pd_op.matmul: (1x20x1024xf32) <- (1x20x1024xf32, 1024x1024xf32) matmul_1 = paddle._C_ops.matmul(dropout_0, parameter_535, False, False) del parameter_535 - # pd_op.add: (1x10x1024xf32) <- (1x10x1024xf32, 1024xf32) + # pd_op.add: (1x20x1024xf32) <- (1x20x1024xf32, 1024xf32) add_2 = paddle._C_ops.add(matmul_1, parameter_534) del matmul_1, parameter_534 # pd_op.full_int_array: (4xi64) <- () full_int_array_2 = [1, -1, 16, 64] - # pd_op.reshape: (1x10x16x64xf32) <- (1x10x1024xf32, 4xi64) + # pd_op.reshape: (1x20x16x64xf32) <- (1x20x1024xf32, 4xi64) reshape_24 = paddle._C_ops.reshape(add_2, full_int_array_2) del add_2 - # pd_op.transpose: (1x16x10x64xf32) <- (1x10x16x64xf32) + # pd_op.transpose: (1x16x20x64xf32) <- (1x20x16x64xf32) transpose_0 = paddle._C_ops.transpose(reshape_24, [0, 2, 1, 3]) del reshape_24 - # pd_op.matmul: (1x10x1024xf32) <- (1x10x1024xf32, 1024x1024xf32) + # pd_op.matmul: (1x20x1024xf32) <- (1x20x1024xf32, 1024x1024xf32) matmul_2 = paddle._C_ops.matmul(dropout_0, parameter_533, False, False) del parameter_533 - # pd_op.add: (1x10x1024xf32) <- (1x10x1024xf32, 1024xf32) + # pd_op.add: (1x20x1024xf32) <- (1x20x1024xf32, 1024xf32) add_3 = paddle._C_ops.add(matmul_2, parameter_532) del matmul_2, parameter_532 - # pd_op.reshape: (1x10x16x64xf32) <- (1x10x1024xf32, 4xi64) + # pd_op.reshape: (1x20x16x64xf32) <- (1x20x1024xf32, 4xi64) reshape_25 = paddle._C_ops.reshape(add_3, full_int_array_2) del add_3 - # pd_op.transpose: (1x16x10x64xf32) <- (1x10x16x64xf32) + # pd_op.transpose: (1x16x20x64xf32) <- (1x20x16x64xf32) transpose_1 = paddle._C_ops.transpose(reshape_25, [0, 2, 1, 3]) del reshape_25 # pd_op.full_int_array: (4xi64) <- () - full_int_array_3 = [1, 10, 16, 64] + full_int_array_3 = [1, 20, 16, 64] - # pd_op.reshape: (1x10x16x64xf32) <- (1x10x1024xf32, 4xi64) + # pd_op.reshape: (1x20x16x64xf32) <- (1x20x1024xf32, 4xi64) reshape_26 = paddle._C_ops.reshape(scale_3, full_int_array_3) del scale_3 - # pd_op.transpose: (1x16x10x64xf32) <- (1x10x16x64xf32) + # pd_op.transpose: (1x16x20x64xf32) <- (1x20x16x64xf32) transpose_2 = paddle._C_ops.transpose(reshape_26, [0, 2, 1, 3]) del reshape_26 # pd_op.full_int_array: (3xi64) <- () full_int_array_4 = [16, -1, 64] - # pd_op.reshape: (16x10x64xf32) <- (1x16x10x64xf32, 3xi64) + # pd_op.reshape: (16x20x64xf32) <- (1x16x20x64xf32, 3xi64) reshape_27 = paddle._C_ops.reshape(transpose_2, full_int_array_4) del transpose_2 - # pd_op.reshape: (16x10x64xf32) <- (1x16x10x64xf32, 3xi64) + # pd_op.reshape: (16x20x64xf32) <- (1x16x20x64xf32, 3xi64) reshape_28 = paddle._C_ops.reshape(transpose_0, full_int_array_4) del transpose_0 - # pd_op.reshape: (16x10x64xf32) <- (1x16x10x64xf32, 3xi64) + # pd_op.reshape: (16x20x64xf32) <- (1x16x20x64xf32, 3xi64) reshape_29 = paddle._C_ops.reshape(transpose_1, full_int_array_4) del transpose_1 - # pd_op.transpose: (16x64x10xf32) <- (16x10x64xf32) + # pd_op.transpose: (16x64x20xf32) <- (16x20x64xf32) transpose_3 = paddle._C_ops.transpose(reshape_28, [0, 2, 1]) del reshape_28 - # pd_op.bmm: (16x10x10xf32) <- (16x10x64xf32, 16x64x10xf32) + # pd_op.bmm: (16x20x20xf32) <- (16x20x64xf32, 16x64x20xf32) bmm_0 = paddle._C_ops.bmm(reshape_27, transpose_3) del reshape_27, transpose_3 - # pd_op.add: (16x10x10xf32) <- (16x10x10xf32, 16x1x10xf32) + # pd_op.add: (16x20x20xf32) <- (16x20x20xf32, 16x1x20xf32) add_4 = paddle._C_ops.add(bmm_0, cast_2) del bmm_0 - # pd_op.softmax: (16x10x10xf32) <- (16x10x10xf32) + # pd_op.softmax: (16x20x20xf32) <- (16x20x20xf32) softmax_0 = paddle._C_ops.softmax(add_4, -1) del add_4 - # pd_op.dropout: (16x10x10xf32, 16x10x10xui8) <- (16x10x10xf32, None, 1xf32) + # pd_op.dropout: (16x20x20xf32, 16x20x20xui8) <- (16x20x20xf32, None, 1xf32) dropout_2, dropout_3 = (lambda x, f: f(x))( paddle._C_ops.dropout( softmax_0, None, full_8, True, "upscale_in_train", 0, False @@ -792,37 +792,37 @@ def forward( ) del softmax_0 - # pd_op.bmm: (16x10x64xf32) <- (16x10x10xf32, 16x10x64xf32) + # pd_op.bmm: (16x20x64xf32) <- (16x20x20xf32, 16x20x64xf32) bmm_1 = paddle._C_ops.bmm(dropout_2, reshape_29) del dropout_2, reshape_29 # pd_op.full_int_array: (4xi64) <- () - full_int_array_5 = [1, 16, 10, 64] + full_int_array_5 = [1, 16, 20, 64] - # pd_op.reshape: (1x16x10x64xf32) <- (16x10x64xf32, 4xi64) + # pd_op.reshape: (1x16x20x64xf32) <- (16x20x64xf32, 4xi64) reshape_30 = paddle._C_ops.reshape(bmm_1, full_int_array_5) del bmm_1 - # pd_op.transpose: (1x10x16x64xf32) <- (1x16x10x64xf32) + # pd_op.transpose: (1x20x16x64xf32) <- (1x16x20x64xf32) transpose_4 = paddle._C_ops.transpose(reshape_30, [0, 2, 1, 3]) del reshape_30 # pd_op.full_int_array: (3xi64) <- () - full_int_array_6 = [1, 10, 1024] + full_int_array_6 = [1, 20, 1024] - # pd_op.reshape: (1x10x1024xf32) <- (1x10x16x64xf32, 3xi64) + # pd_op.reshape: (1x20x1024xf32) <- (1x20x16x64xf32, 3xi64) reshape_31 = paddle._C_ops.reshape(transpose_4, full_int_array_6) del transpose_4 - # pd_op.matmul: (1x10x1024xf32) <- (1x10x1024xf32, 1024x1024xf32) + # pd_op.matmul: (1x20x1024xf32) <- (1x20x1024xf32, 1024x1024xf32) matmul_3 = paddle._C_ops.matmul(reshape_31, parameter_529, False, False) del parameter_529, reshape_31 - # pd_op.add: (1x10x1024xf32) <- (1x10x1024xf32, 1024xf32) + # pd_op.add: (1x20x1024xf32) <- (1x20x1024xf32, 1024xf32) add_5 = paddle._C_ops.add(matmul_3, parameter_528) del matmul_3, parameter_528 - # pd_op.dropout: (1x10x1024xf32, 1x10x1024xui8) <- (1x10x1024xf32, None, 1xf32) + # pd_op.dropout: (1x20x1024xf32, 1x20x1024xui8) <- (1x20x1024xf32, None, 1xf32) dropout_4, dropout_5 = (lambda x, f: f(x))( paddle._C_ops.dropout( add_5, None, full_8, True, "upscale_in_train", 0, False @@ -831,11 +831,11 @@ def forward( ) del add_5 - # pd_op.add: (1x10x1024xf32) <- (1x10x1024xf32, 1x10x1024xf32) + # pd_op.add: (1x20x1024xf32) <- (1x20x1024xf32, 1x20x1024xf32) add_6 = paddle._C_ops.add(dropout_4, dropout_0) del dropout_0, dropout_4 - # pd_op.layer_norm: (1x10x1024xf32, 1x10xf32, 1x10xf32) <- (1x10x1024xf32, 1024xf32, 1024xf32) + # pd_op.layer_norm: (1x20x1024xf32, 1x20xf32, 1x20xf32) <- (1x20x1024xf32, 1024xf32, 1024xf32) layer_norm_3, layer_norm_4, layer_norm_5 = (lambda x, f: f(x))( paddle._C_ops.layer_norm( add_6, parameter_527, parameter_526, float("1e-05"), 2 @@ -844,19 +844,19 @@ def forward( ) del add_6, parameter_526, parameter_527 - # pd_op.matmul: (1x10x4096xf32) <- (1x10x1024xf32, 1024x4096xf32) + # pd_op.matmul: (1x20x4096xf32) <- (1x20x1024xf32, 1024x4096xf32) matmul_4 = paddle._C_ops.matmul(layer_norm_3, parameter_525, False, False) del parameter_525 - # pd_op.add: (1x10x4096xf32) <- (1x10x4096xf32, 4096xf32) + # pd_op.add: (1x20x4096xf32) <- (1x20x4096xf32, 4096xf32) add_7 = paddle._C_ops.add(matmul_4, parameter_524) del matmul_4, parameter_524 - # pd_op.gelu: (1x10x4096xf32) <- (1x10x4096xf32) + # pd_op.gelu: (1x20x4096xf32) <- (1x20x4096xf32) gelu_0 = paddle._C_ops.gelu(add_7, False) del add_7 - # pd_op.dropout: (1x10x4096xf32, 1x10x4096xui8) <- (1x10x4096xf32, None, 1xf32) + # pd_op.dropout: (1x20x4096xf32, 1x20x4096xui8) <- (1x20x4096xf32, None, 1xf32) dropout_6, dropout_7 = (lambda x, f: f(x))( paddle._C_ops.dropout( gelu_0, None, full_8, True, "upscale_in_train", 0, False @@ -865,15 +865,15 @@ def forward( ) del gelu_0 - # pd_op.matmul: (1x10x1024xf32) <- (1x10x4096xf32, 4096x1024xf32) + # pd_op.matmul: (1x20x1024xf32) <- (1x20x4096xf32, 4096x1024xf32) matmul_5 = paddle._C_ops.matmul(dropout_6, parameter_523, False, False) del dropout_6, parameter_523 - # pd_op.add: (1x10x1024xf32) <- (1x10x1024xf32, 1024xf32) + # pd_op.add: (1x20x1024xf32) <- (1x20x1024xf32, 1024xf32) add_8 = paddle._C_ops.add(matmul_5, parameter_522) del matmul_5, parameter_522 - # pd_op.dropout: (1x10x1024xf32, 1x10x1024xui8) <- (1x10x1024xf32, None, 1xf32) + # pd_op.dropout: (1x20x1024xf32, 1x20x1024xui8) <- (1x20x1024xf32, None, 1xf32) dropout_8, dropout_9 = (lambda x, f: f(x))( paddle._C_ops.dropout( add_8, None, full_8, True, "upscale_in_train", 0, False @@ -882,11 +882,11 @@ def forward( ) del add_8 - # pd_op.add: (1x10x1024xf32) <- (1x10x1024xf32, 1x10x1024xf32) + # pd_op.add: (1x20x1024xf32) <- (1x20x1024xf32, 1x20x1024xf32) add_9 = paddle._C_ops.add(dropout_8, layer_norm_3) del dropout_8, layer_norm_3 - # pd_op.layer_norm: (1x10x1024xf32, 1x10xf32, 1x10xf32) <- (1x10x1024xf32, 1024xf32, 1024xf32) + # pd_op.layer_norm: (1x20x1024xf32, 1x20xf32, 1x20xf32) <- (1x20x1024xf32, 1024xf32, 1024xf32) layer_norm_6, layer_norm_7, layer_norm_8 = (lambda x, f: f(x))( paddle._C_ops.layer_norm( add_9, parameter_521, parameter_520, float("1e-05"), 2 @@ -895,87 +895,87 @@ def forward( ) del add_9, parameter_520, parameter_521 - # pd_op.matmul: (1x10x1024xf32) <- (1x10x1024xf32, 1024x1024xf32) + # pd_op.matmul: (1x20x1024xf32) <- (1x20x1024xf32, 1024x1024xf32) matmul_6 = paddle._C_ops.matmul(layer_norm_6, parameter_515, False, False) del parameter_515 - # pd_op.add: (1x10x1024xf32) <- (1x10x1024xf32, 1024xf32) + # pd_op.add: (1x20x1024xf32) <- (1x20x1024xf32, 1024xf32) add_10 = paddle._C_ops.add(matmul_6, parameter_514) del matmul_6, parameter_514 - # pd_op.scale: (1x10x1024xf32) <- (1x10x1024xf32, 1xf32) + # pd_op.scale: (1x20x1024xf32) <- (1x20x1024xf32, 1xf32) scale_4 = paddle._C_ops.scale(add_10, full_9, float("0"), True) del add_10 - # pd_op.matmul: (1x10x1024xf32) <- (1x10x1024xf32, 1024x1024xf32) + # pd_op.matmul: (1x20x1024xf32) <- (1x20x1024xf32, 1024x1024xf32) matmul_7 = paddle._C_ops.matmul(layer_norm_6, parameter_519, False, False) del parameter_519 - # pd_op.add: (1x10x1024xf32) <- (1x10x1024xf32, 1024xf32) + # pd_op.add: (1x20x1024xf32) <- (1x20x1024xf32, 1024xf32) add_11 = paddle._C_ops.add(matmul_7, parameter_518) del matmul_7, parameter_518 - # pd_op.reshape: (1x10x16x64xf32) <- (1x10x1024xf32, 4xi64) + # pd_op.reshape: (1x20x16x64xf32) <- (1x20x1024xf32, 4xi64) reshape_32 = paddle._C_ops.reshape(add_11, full_int_array_2) del add_11 - # pd_op.transpose: (1x16x10x64xf32) <- (1x10x16x64xf32) + # pd_op.transpose: (1x16x20x64xf32) <- (1x20x16x64xf32) transpose_5 = paddle._C_ops.transpose(reshape_32, [0, 2, 1, 3]) del reshape_32 - # pd_op.matmul: (1x10x1024xf32) <- (1x10x1024xf32, 1024x1024xf32) + # pd_op.matmul: (1x20x1024xf32) <- (1x20x1024xf32, 1024x1024xf32) matmul_8 = paddle._C_ops.matmul(layer_norm_6, parameter_517, False, False) del parameter_517 - # pd_op.add: (1x10x1024xf32) <- (1x10x1024xf32, 1024xf32) + # pd_op.add: (1x20x1024xf32) <- (1x20x1024xf32, 1024xf32) add_12 = paddle._C_ops.add(matmul_8, parameter_516) del matmul_8, parameter_516 - # pd_op.reshape: (1x10x16x64xf32) <- (1x10x1024xf32, 4xi64) + # pd_op.reshape: (1x20x16x64xf32) <- (1x20x1024xf32, 4xi64) reshape_33 = paddle._C_ops.reshape(add_12, full_int_array_2) del add_12 - # pd_op.transpose: (1x16x10x64xf32) <- (1x10x16x64xf32) + # pd_op.transpose: (1x16x20x64xf32) <- (1x20x16x64xf32) transpose_6 = paddle._C_ops.transpose(reshape_33, [0, 2, 1, 3]) del reshape_33 - # pd_op.reshape: (1x10x16x64xf32) <- (1x10x1024xf32, 4xi64) + # pd_op.reshape: (1x20x16x64xf32) <- (1x20x1024xf32, 4xi64) reshape_34 = paddle._C_ops.reshape(scale_4, full_int_array_3) del scale_4 - # pd_op.transpose: (1x16x10x64xf32) <- (1x10x16x64xf32) + # pd_op.transpose: (1x16x20x64xf32) <- (1x20x16x64xf32) transpose_7 = paddle._C_ops.transpose(reshape_34, [0, 2, 1, 3]) del reshape_34 - # pd_op.reshape: (16x10x64xf32) <- (1x16x10x64xf32, 3xi64) + # pd_op.reshape: (16x20x64xf32) <- (1x16x20x64xf32, 3xi64) reshape_35 = paddle._C_ops.reshape(transpose_7, full_int_array_4) del transpose_7 - # pd_op.reshape: (16x10x64xf32) <- (1x16x10x64xf32, 3xi64) + # pd_op.reshape: (16x20x64xf32) <- (1x16x20x64xf32, 3xi64) reshape_36 = paddle._C_ops.reshape(transpose_5, full_int_array_4) del transpose_5 - # pd_op.reshape: (16x10x64xf32) <- (1x16x10x64xf32, 3xi64) + # pd_op.reshape: (16x20x64xf32) <- (1x16x20x64xf32, 3xi64) reshape_37 = paddle._C_ops.reshape(transpose_6, full_int_array_4) del transpose_6 - # pd_op.transpose: (16x64x10xf32) <- (16x10x64xf32) + # pd_op.transpose: (16x64x20xf32) <- (16x20x64xf32) transpose_8 = paddle._C_ops.transpose(reshape_36, [0, 2, 1]) del reshape_36 - # pd_op.bmm: (16x10x10xf32) <- (16x10x64xf32, 16x64x10xf32) + # pd_op.bmm: (16x20x20xf32) <- (16x20x64xf32, 16x64x20xf32) bmm_2 = paddle._C_ops.bmm(reshape_35, transpose_8) del reshape_35, transpose_8 - # pd_op.add: (16x10x10xf32) <- (16x10x10xf32, 16x1x10xf32) + # pd_op.add: (16x20x20xf32) <- (16x20x20xf32, 16x1x20xf32) add_13 = paddle._C_ops.add(bmm_2, cast_2) del bmm_2 - # pd_op.softmax: (16x10x10xf32) <- (16x10x10xf32) + # pd_op.softmax: (16x20x20xf32) <- (16x20x20xf32) softmax_1 = paddle._C_ops.softmax(add_13, -1) del add_13 - # pd_op.dropout: (16x10x10xf32, 16x10x10xui8) <- (16x10x10xf32, None, 1xf32) + # pd_op.dropout: (16x20x20xf32, 16x20x20xui8) <- (16x20x20xf32, None, 1xf32) dropout_10, dropout_11 = (lambda x, f: f(x))( paddle._C_ops.dropout( softmax_1, None, full_8, True, "upscale_in_train", 0, False @@ -984,31 +984,31 @@ def forward( ) del softmax_1 - # pd_op.bmm: (16x10x64xf32) <- (16x10x10xf32, 16x10x64xf32) + # pd_op.bmm: (16x20x64xf32) <- (16x20x20xf32, 16x20x64xf32) bmm_3 = paddle._C_ops.bmm(dropout_10, reshape_37) del dropout_10, reshape_37 - # pd_op.reshape: (1x16x10x64xf32) <- (16x10x64xf32, 4xi64) + # pd_op.reshape: (1x16x20x64xf32) <- (16x20x64xf32, 4xi64) reshape_38 = paddle._C_ops.reshape(bmm_3, full_int_array_5) del bmm_3 - # pd_op.transpose: (1x10x16x64xf32) <- (1x16x10x64xf32) + # pd_op.transpose: (1x20x16x64xf32) <- (1x16x20x64xf32) transpose_9 = paddle._C_ops.transpose(reshape_38, [0, 2, 1, 3]) del reshape_38 - # pd_op.reshape: (1x10x1024xf32) <- (1x10x16x64xf32, 3xi64) + # pd_op.reshape: (1x20x1024xf32) <- (1x20x16x64xf32, 3xi64) reshape_39 = paddle._C_ops.reshape(transpose_9, full_int_array_6) del transpose_9 - # pd_op.matmul: (1x10x1024xf32) <- (1x10x1024xf32, 1024x1024xf32) + # pd_op.matmul: (1x20x1024xf32) <- (1x20x1024xf32, 1024x1024xf32) matmul_9 = paddle._C_ops.matmul(reshape_39, parameter_513, False, False) del parameter_513, reshape_39 - # pd_op.add: (1x10x1024xf32) <- (1x10x1024xf32, 1024xf32) + # pd_op.add: (1x20x1024xf32) <- (1x20x1024xf32, 1024xf32) add_14 = paddle._C_ops.add(matmul_9, parameter_512) del matmul_9, parameter_512 - # pd_op.dropout: (1x10x1024xf32, 1x10x1024xui8) <- (1x10x1024xf32, None, 1xf32) + # pd_op.dropout: (1x20x1024xf32, 1x20x1024xui8) <- (1x20x1024xf32, None, 1xf32) dropout_12, dropout_13 = (lambda x, f: f(x))( paddle._C_ops.dropout( add_14, None, full_8, True, "upscale_in_train", 0, False @@ -1017,11 +1017,11 @@ def forward( ) del add_14 - # pd_op.add: (1x10x1024xf32) <- (1x10x1024xf32, 1x10x1024xf32) + # pd_op.add: (1x20x1024xf32) <- (1x20x1024xf32, 1x20x1024xf32) add_15 = paddle._C_ops.add(dropout_12, layer_norm_6) del dropout_12, layer_norm_6 - # pd_op.layer_norm: (1x10x1024xf32, 1x10xf32, 1x10xf32) <- (1x10x1024xf32, 1024xf32, 1024xf32) + # pd_op.layer_norm: (1x20x1024xf32, 1x20xf32, 1x20xf32) <- (1x20x1024xf32, 1024xf32, 1024xf32) layer_norm_9, layer_norm_10, layer_norm_11 = (lambda x, f: f(x))( paddle._C_ops.layer_norm( add_15, parameter_511, parameter_510, float("1e-05"), 2 @@ -1030,19 +1030,19 @@ def forward( ) del add_15, parameter_510, parameter_511 - # pd_op.matmul: (1x10x4096xf32) <- (1x10x1024xf32, 1024x4096xf32) + # pd_op.matmul: (1x20x4096xf32) <- (1x20x1024xf32, 1024x4096xf32) matmul_10 = paddle._C_ops.matmul(layer_norm_9, parameter_509, False, False) del parameter_509 - # pd_op.add: (1x10x4096xf32) <- (1x10x4096xf32, 4096xf32) + # pd_op.add: (1x20x4096xf32) <- (1x20x4096xf32, 4096xf32) add_16 = paddle._C_ops.add(matmul_10, parameter_508) del matmul_10, parameter_508 - # pd_op.gelu: (1x10x4096xf32) <- (1x10x4096xf32) + # pd_op.gelu: (1x20x4096xf32) <- (1x20x4096xf32) gelu_1 = paddle._C_ops.gelu(add_16, False) del add_16 - # pd_op.dropout: (1x10x4096xf32, 1x10x4096xui8) <- (1x10x4096xf32, None, 1xf32) + # pd_op.dropout: (1x20x4096xf32, 1x20x4096xui8) <- (1x20x4096xf32, None, 1xf32) dropout_14, dropout_15 = (lambda x, f: f(x))( paddle._C_ops.dropout( gelu_1, None, full_8, True, "upscale_in_train", 0, False @@ -1051,15 +1051,15 @@ def forward( ) del gelu_1 - # pd_op.matmul: (1x10x1024xf32) <- (1x10x4096xf32, 4096x1024xf32) + # pd_op.matmul: (1x20x1024xf32) <- (1x20x4096xf32, 4096x1024xf32) matmul_11 = paddle._C_ops.matmul(dropout_14, parameter_507, False, False) del dropout_14, parameter_507 - # pd_op.add: (1x10x1024xf32) <- (1x10x1024xf32, 1024xf32) + # pd_op.add: (1x20x1024xf32) <- (1x20x1024xf32, 1024xf32) add_17 = paddle._C_ops.add(matmul_11, parameter_506) del matmul_11, parameter_506 - # pd_op.dropout: (1x10x1024xf32, 1x10x1024xui8) <- (1x10x1024xf32, None, 1xf32) + # pd_op.dropout: (1x20x1024xf32, 1x20x1024xui8) <- (1x20x1024xf32, None, 1xf32) dropout_16, dropout_17 = (lambda x, f: f(x))( paddle._C_ops.dropout( add_17, None, full_8, True, "upscale_in_train", 0, False @@ -1068,11 +1068,11 @@ def forward( ) del add_17 - # pd_op.add: (1x10x1024xf32) <- (1x10x1024xf32, 1x10x1024xf32) + # pd_op.add: (1x20x1024xf32) <- (1x20x1024xf32, 1x20x1024xf32) add_18 = paddle._C_ops.add(dropout_16, layer_norm_9) del dropout_16, layer_norm_9 - # pd_op.layer_norm: (1x10x1024xf32, 1x10xf32, 1x10xf32) <- (1x10x1024xf32, 1024xf32, 1024xf32) + # pd_op.layer_norm: (1x20x1024xf32, 1x20xf32, 1x20xf32) <- (1x20x1024xf32, 1024xf32, 1024xf32) layer_norm_12, layer_norm_13, layer_norm_14 = (lambda x, f: f(x))( paddle._C_ops.layer_norm( add_18, parameter_505, parameter_504, float("1e-05"), 2 @@ -1081,87 +1081,87 @@ def forward( ) del add_18, parameter_504, parameter_505 - # pd_op.matmul: (1x10x1024xf32) <- (1x10x1024xf32, 1024x1024xf32) + # pd_op.matmul: (1x20x1024xf32) <- (1x20x1024xf32, 1024x1024xf32) matmul_12 = paddle._C_ops.matmul(layer_norm_12, parameter_499, False, False) del parameter_499 - # pd_op.add: (1x10x1024xf32) <- (1x10x1024xf32, 1024xf32) + # pd_op.add: (1x20x1024xf32) <- (1x20x1024xf32, 1024xf32) add_19 = paddle._C_ops.add(matmul_12, parameter_498) del matmul_12, parameter_498 - # pd_op.scale: (1x10x1024xf32) <- (1x10x1024xf32, 1xf32) + # pd_op.scale: (1x20x1024xf32) <- (1x20x1024xf32, 1xf32) scale_5 = paddle._C_ops.scale(add_19, full_9, float("0"), True) del add_19 - # pd_op.matmul: (1x10x1024xf32) <- (1x10x1024xf32, 1024x1024xf32) + # pd_op.matmul: (1x20x1024xf32) <- (1x20x1024xf32, 1024x1024xf32) matmul_13 = paddle._C_ops.matmul(layer_norm_12, parameter_503, False, False) del parameter_503 - # pd_op.add: (1x10x1024xf32) <- (1x10x1024xf32, 1024xf32) + # pd_op.add: (1x20x1024xf32) <- (1x20x1024xf32, 1024xf32) add_20 = paddle._C_ops.add(matmul_13, parameter_502) del matmul_13, parameter_502 - # pd_op.reshape: (1x10x16x64xf32) <- (1x10x1024xf32, 4xi64) + # pd_op.reshape: (1x20x16x64xf32) <- (1x20x1024xf32, 4xi64) reshape_40 = paddle._C_ops.reshape(add_20, full_int_array_2) del add_20 - # pd_op.transpose: (1x16x10x64xf32) <- (1x10x16x64xf32) + # pd_op.transpose: (1x16x20x64xf32) <- (1x20x16x64xf32) transpose_10 = paddle._C_ops.transpose(reshape_40, [0, 2, 1, 3]) del reshape_40 - # pd_op.matmul: (1x10x1024xf32) <- (1x10x1024xf32, 1024x1024xf32) + # pd_op.matmul: (1x20x1024xf32) <- (1x20x1024xf32, 1024x1024xf32) matmul_14 = paddle._C_ops.matmul(layer_norm_12, parameter_501, False, False) del parameter_501 - # pd_op.add: (1x10x1024xf32) <- (1x10x1024xf32, 1024xf32) + # pd_op.add: (1x20x1024xf32) <- (1x20x1024xf32, 1024xf32) add_21 = paddle._C_ops.add(matmul_14, parameter_500) del matmul_14, parameter_500 - # pd_op.reshape: (1x10x16x64xf32) <- (1x10x1024xf32, 4xi64) + # pd_op.reshape: (1x20x16x64xf32) <- (1x20x1024xf32, 4xi64) reshape_41 = paddle._C_ops.reshape(add_21, full_int_array_2) del add_21 - # pd_op.transpose: (1x16x10x64xf32) <- (1x10x16x64xf32) + # pd_op.transpose: (1x16x20x64xf32) <- (1x20x16x64xf32) transpose_11 = paddle._C_ops.transpose(reshape_41, [0, 2, 1, 3]) del reshape_41 - # pd_op.reshape: (1x10x16x64xf32) <- (1x10x1024xf32, 4xi64) + # pd_op.reshape: (1x20x16x64xf32) <- (1x20x1024xf32, 4xi64) reshape_42 = paddle._C_ops.reshape(scale_5, full_int_array_3) del scale_5 - # pd_op.transpose: (1x16x10x64xf32) <- (1x10x16x64xf32) + # pd_op.transpose: (1x16x20x64xf32) <- (1x20x16x64xf32) transpose_12 = paddle._C_ops.transpose(reshape_42, [0, 2, 1, 3]) del reshape_42 - # pd_op.reshape: (16x10x64xf32) <- (1x16x10x64xf32, 3xi64) + # pd_op.reshape: (16x20x64xf32) <- (1x16x20x64xf32, 3xi64) reshape_43 = paddle._C_ops.reshape(transpose_12, full_int_array_4) del transpose_12 - # pd_op.reshape: (16x10x64xf32) <- (1x16x10x64xf32, 3xi64) + # pd_op.reshape: (16x20x64xf32) <- (1x16x20x64xf32, 3xi64) reshape_44 = paddle._C_ops.reshape(transpose_10, full_int_array_4) del transpose_10 - # pd_op.reshape: (16x10x64xf32) <- (1x16x10x64xf32, 3xi64) + # pd_op.reshape: (16x20x64xf32) <- (1x16x20x64xf32, 3xi64) reshape_45 = paddle._C_ops.reshape(transpose_11, full_int_array_4) del transpose_11 - # pd_op.transpose: (16x64x10xf32) <- (16x10x64xf32) + # pd_op.transpose: (16x64x20xf32) <- (16x20x64xf32) transpose_13 = paddle._C_ops.transpose(reshape_44, [0, 2, 1]) del reshape_44 - # pd_op.bmm: (16x10x10xf32) <- (16x10x64xf32, 16x64x10xf32) + # pd_op.bmm: (16x20x20xf32) <- (16x20x64xf32, 16x64x20xf32) bmm_4 = paddle._C_ops.bmm(reshape_43, transpose_13) del reshape_43, transpose_13 - # pd_op.add: (16x10x10xf32) <- (16x10x10xf32, 16x1x10xf32) + # pd_op.add: (16x20x20xf32) <- (16x20x20xf32, 16x1x20xf32) add_22 = paddle._C_ops.add(bmm_4, cast_2) del bmm_4 - # pd_op.softmax: (16x10x10xf32) <- (16x10x10xf32) + # pd_op.softmax: (16x20x20xf32) <- (16x20x20xf32) softmax_2 = paddle._C_ops.softmax(add_22, -1) del add_22 - # pd_op.dropout: (16x10x10xf32, 16x10x10xui8) <- (16x10x10xf32, None, 1xf32) + # pd_op.dropout: (16x20x20xf32, 16x20x20xui8) <- (16x20x20xf32, None, 1xf32) dropout_18, dropout_19 = (lambda x, f: f(x))( paddle._C_ops.dropout( softmax_2, None, full_8, True, "upscale_in_train", 0, False @@ -1170,31 +1170,31 @@ def forward( ) del softmax_2 - # pd_op.bmm: (16x10x64xf32) <- (16x10x10xf32, 16x10x64xf32) + # pd_op.bmm: (16x20x64xf32) <- (16x20x20xf32, 16x20x64xf32) bmm_5 = paddle._C_ops.bmm(dropout_18, reshape_45) del dropout_18, reshape_45 - # pd_op.reshape: (1x16x10x64xf32) <- (16x10x64xf32, 4xi64) + # pd_op.reshape: (1x16x20x64xf32) <- (16x20x64xf32, 4xi64) reshape_46 = paddle._C_ops.reshape(bmm_5, full_int_array_5) del bmm_5 - # pd_op.transpose: (1x10x16x64xf32) <- (1x16x10x64xf32) + # pd_op.transpose: (1x20x16x64xf32) <- (1x16x20x64xf32) transpose_14 = paddle._C_ops.transpose(reshape_46, [0, 2, 1, 3]) del reshape_46 - # pd_op.reshape: (1x10x1024xf32) <- (1x10x16x64xf32, 3xi64) + # pd_op.reshape: (1x20x1024xf32) <- (1x20x16x64xf32, 3xi64) reshape_47 = paddle._C_ops.reshape(transpose_14, full_int_array_6) del transpose_14 - # pd_op.matmul: (1x10x1024xf32) <- (1x10x1024xf32, 1024x1024xf32) + # pd_op.matmul: (1x20x1024xf32) <- (1x20x1024xf32, 1024x1024xf32) matmul_15 = paddle._C_ops.matmul(reshape_47, parameter_497, False, False) del parameter_497, reshape_47 - # pd_op.add: (1x10x1024xf32) <- (1x10x1024xf32, 1024xf32) + # pd_op.add: (1x20x1024xf32) <- (1x20x1024xf32, 1024xf32) add_23 = paddle._C_ops.add(matmul_15, parameter_496) del matmul_15, parameter_496 - # pd_op.dropout: (1x10x1024xf32, 1x10x1024xui8) <- (1x10x1024xf32, None, 1xf32) + # pd_op.dropout: (1x20x1024xf32, 1x20x1024xui8) <- (1x20x1024xf32, None, 1xf32) dropout_20, dropout_21 = (lambda x, f: f(x))( paddle._C_ops.dropout( add_23, None, full_8, True, "upscale_in_train", 0, False @@ -1203,11 +1203,11 @@ def forward( ) del add_23 - # pd_op.add: (1x10x1024xf32) <- (1x10x1024xf32, 1x10x1024xf32) + # pd_op.add: (1x20x1024xf32) <- (1x20x1024xf32, 1x20x1024xf32) add_24 = paddle._C_ops.add(dropout_20, layer_norm_12) del dropout_20, layer_norm_12 - # pd_op.layer_norm: (1x10x1024xf32, 1x10xf32, 1x10xf32) <- (1x10x1024xf32, 1024xf32, 1024xf32) + # pd_op.layer_norm: (1x20x1024xf32, 1x20xf32, 1x20xf32) <- (1x20x1024xf32, 1024xf32, 1024xf32) layer_norm_15, layer_norm_16, layer_norm_17 = (lambda x, f: f(x))( paddle._C_ops.layer_norm( add_24, parameter_495, parameter_494, float("1e-05"), 2 @@ -1216,19 +1216,19 @@ def forward( ) del add_24, parameter_494, parameter_495 - # pd_op.matmul: (1x10x4096xf32) <- (1x10x1024xf32, 1024x4096xf32) + # pd_op.matmul: (1x20x4096xf32) <- (1x20x1024xf32, 1024x4096xf32) matmul_16 = paddle._C_ops.matmul(layer_norm_15, parameter_493, False, False) del parameter_493 - # pd_op.add: (1x10x4096xf32) <- (1x10x4096xf32, 4096xf32) + # pd_op.add: (1x20x4096xf32) <- (1x20x4096xf32, 4096xf32) add_25 = paddle._C_ops.add(matmul_16, parameter_492) del matmul_16, parameter_492 - # pd_op.gelu: (1x10x4096xf32) <- (1x10x4096xf32) + # pd_op.gelu: (1x20x4096xf32) <- (1x20x4096xf32) gelu_2 = paddle._C_ops.gelu(add_25, False) del add_25 - # pd_op.dropout: (1x10x4096xf32, 1x10x4096xui8) <- (1x10x4096xf32, None, 1xf32) + # pd_op.dropout: (1x20x4096xf32, 1x20x4096xui8) <- (1x20x4096xf32, None, 1xf32) dropout_22, dropout_23 = (lambda x, f: f(x))( paddle._C_ops.dropout( gelu_2, None, full_8, True, "upscale_in_train", 0, False @@ -1237,15 +1237,15 @@ def forward( ) del gelu_2 - # pd_op.matmul: (1x10x1024xf32) <- (1x10x4096xf32, 4096x1024xf32) + # pd_op.matmul: (1x20x1024xf32) <- (1x20x4096xf32, 4096x1024xf32) matmul_17 = paddle._C_ops.matmul(dropout_22, parameter_491, False, False) del dropout_22, parameter_491 - # pd_op.add: (1x10x1024xf32) <- (1x10x1024xf32, 1024xf32) + # pd_op.add: (1x20x1024xf32) <- (1x20x1024xf32, 1024xf32) add_26 = paddle._C_ops.add(matmul_17, parameter_490) del matmul_17, parameter_490 - # pd_op.dropout: (1x10x1024xf32, 1x10x1024xui8) <- (1x10x1024xf32, None, 1xf32) + # pd_op.dropout: (1x20x1024xf32, 1x20x1024xui8) <- (1x20x1024xf32, None, 1xf32) dropout_24, dropout_25 = (lambda x, f: f(x))( paddle._C_ops.dropout( add_26, None, full_8, True, "upscale_in_train", 0, False @@ -1254,11 +1254,11 @@ def forward( ) del add_26 - # pd_op.add: (1x10x1024xf32) <- (1x10x1024xf32, 1x10x1024xf32) + # pd_op.add: (1x20x1024xf32) <- (1x20x1024xf32, 1x20x1024xf32) add_27 = paddle._C_ops.add(dropout_24, layer_norm_15) del dropout_24, layer_norm_15 - # pd_op.layer_norm: (1x10x1024xf32, 1x10xf32, 1x10xf32) <- (1x10x1024xf32, 1024xf32, 1024xf32) + # pd_op.layer_norm: (1x20x1024xf32, 1x20xf32, 1x20xf32) <- (1x20x1024xf32, 1024xf32, 1024xf32) layer_norm_18, layer_norm_19, layer_norm_20 = (lambda x, f: f(x))( paddle._C_ops.layer_norm( add_27, parameter_489, parameter_488, float("1e-05"), 2 @@ -1267,87 +1267,87 @@ def forward( ) del add_27, parameter_488, parameter_489 - # pd_op.matmul: (1x10x1024xf32) <- (1x10x1024xf32, 1024x1024xf32) + # pd_op.matmul: (1x20x1024xf32) <- (1x20x1024xf32, 1024x1024xf32) matmul_18 = paddle._C_ops.matmul(layer_norm_18, parameter_483, False, False) del parameter_483 - # pd_op.add: (1x10x1024xf32) <- (1x10x1024xf32, 1024xf32) + # pd_op.add: (1x20x1024xf32) <- (1x20x1024xf32, 1024xf32) add_28 = paddle._C_ops.add(matmul_18, parameter_482) del matmul_18, parameter_482 - # pd_op.scale: (1x10x1024xf32) <- (1x10x1024xf32, 1xf32) + # pd_op.scale: (1x20x1024xf32) <- (1x20x1024xf32, 1xf32) scale_6 = paddle._C_ops.scale(add_28, full_9, float("0"), True) del add_28 - # pd_op.matmul: (1x10x1024xf32) <- (1x10x1024xf32, 1024x1024xf32) + # pd_op.matmul: (1x20x1024xf32) <- (1x20x1024xf32, 1024x1024xf32) matmul_19 = paddle._C_ops.matmul(layer_norm_18, parameter_487, False, False) del parameter_487 - # pd_op.add: (1x10x1024xf32) <- (1x10x1024xf32, 1024xf32) + # pd_op.add: (1x20x1024xf32) <- (1x20x1024xf32, 1024xf32) add_29 = paddle._C_ops.add(matmul_19, parameter_486) del matmul_19, parameter_486 - # pd_op.reshape: (1x10x16x64xf32) <- (1x10x1024xf32, 4xi64) + # pd_op.reshape: (1x20x16x64xf32) <- (1x20x1024xf32, 4xi64) reshape_48 = paddle._C_ops.reshape(add_29, full_int_array_2) del add_29 - # pd_op.transpose: (1x16x10x64xf32) <- (1x10x16x64xf32) + # pd_op.transpose: (1x16x20x64xf32) <- (1x20x16x64xf32) transpose_15 = paddle._C_ops.transpose(reshape_48, [0, 2, 1, 3]) del reshape_48 - # pd_op.matmul: (1x10x1024xf32) <- (1x10x1024xf32, 1024x1024xf32) + # pd_op.matmul: (1x20x1024xf32) <- (1x20x1024xf32, 1024x1024xf32) matmul_20 = paddle._C_ops.matmul(layer_norm_18, parameter_485, False, False) del parameter_485 - # pd_op.add: (1x10x1024xf32) <- (1x10x1024xf32, 1024xf32) + # pd_op.add: (1x20x1024xf32) <- (1x20x1024xf32, 1024xf32) add_30 = paddle._C_ops.add(matmul_20, parameter_484) del matmul_20, parameter_484 - # pd_op.reshape: (1x10x16x64xf32) <- (1x10x1024xf32, 4xi64) + # pd_op.reshape: (1x20x16x64xf32) <- (1x20x1024xf32, 4xi64) reshape_49 = paddle._C_ops.reshape(add_30, full_int_array_2) del add_30 - # pd_op.transpose: (1x16x10x64xf32) <- (1x10x16x64xf32) + # pd_op.transpose: (1x16x20x64xf32) <- (1x20x16x64xf32) transpose_16 = paddle._C_ops.transpose(reshape_49, [0, 2, 1, 3]) del reshape_49 - # pd_op.reshape: (1x10x16x64xf32) <- (1x10x1024xf32, 4xi64) + # pd_op.reshape: (1x20x16x64xf32) <- (1x20x1024xf32, 4xi64) reshape_50 = paddle._C_ops.reshape(scale_6, full_int_array_3) del scale_6 - # pd_op.transpose: (1x16x10x64xf32) <- (1x10x16x64xf32) + # pd_op.transpose: (1x16x20x64xf32) <- (1x20x16x64xf32) transpose_17 = paddle._C_ops.transpose(reshape_50, [0, 2, 1, 3]) del reshape_50 - # pd_op.reshape: (16x10x64xf32) <- (1x16x10x64xf32, 3xi64) + # pd_op.reshape: (16x20x64xf32) <- (1x16x20x64xf32, 3xi64) reshape_51 = paddle._C_ops.reshape(transpose_17, full_int_array_4) del transpose_17 - # pd_op.reshape: (16x10x64xf32) <- (1x16x10x64xf32, 3xi64) + # pd_op.reshape: (16x20x64xf32) <- (1x16x20x64xf32, 3xi64) reshape_52 = paddle._C_ops.reshape(transpose_15, full_int_array_4) del transpose_15 - # pd_op.reshape: (16x10x64xf32) <- (1x16x10x64xf32, 3xi64) + # pd_op.reshape: (16x20x64xf32) <- (1x16x20x64xf32, 3xi64) reshape_53 = paddle._C_ops.reshape(transpose_16, full_int_array_4) del transpose_16 - # pd_op.transpose: (16x64x10xf32) <- (16x10x64xf32) + # pd_op.transpose: (16x64x20xf32) <- (16x20x64xf32) transpose_18 = paddle._C_ops.transpose(reshape_52, [0, 2, 1]) del reshape_52 - # pd_op.bmm: (16x10x10xf32) <- (16x10x64xf32, 16x64x10xf32) + # pd_op.bmm: (16x20x20xf32) <- (16x20x64xf32, 16x64x20xf32) bmm_6 = paddle._C_ops.bmm(reshape_51, transpose_18) del reshape_51, transpose_18 - # pd_op.add: (16x10x10xf32) <- (16x10x10xf32, 16x1x10xf32) + # pd_op.add: (16x20x20xf32) <- (16x20x20xf32, 16x1x20xf32) add_31 = paddle._C_ops.add(bmm_6, cast_2) del bmm_6 - # pd_op.softmax: (16x10x10xf32) <- (16x10x10xf32) + # pd_op.softmax: (16x20x20xf32) <- (16x20x20xf32) softmax_3 = paddle._C_ops.softmax(add_31, -1) del add_31 - # pd_op.dropout: (16x10x10xf32, 16x10x10xui8) <- (16x10x10xf32, None, 1xf32) + # pd_op.dropout: (16x20x20xf32, 16x20x20xui8) <- (16x20x20xf32, None, 1xf32) dropout_26, dropout_27 = (lambda x, f: f(x))( paddle._C_ops.dropout( softmax_3, None, full_8, True, "upscale_in_train", 0, False @@ -1356,31 +1356,31 @@ def forward( ) del softmax_3 - # pd_op.bmm: (16x10x64xf32) <- (16x10x10xf32, 16x10x64xf32) + # pd_op.bmm: (16x20x64xf32) <- (16x20x20xf32, 16x20x64xf32) bmm_7 = paddle._C_ops.bmm(dropout_26, reshape_53) del dropout_26, reshape_53 - # pd_op.reshape: (1x16x10x64xf32) <- (16x10x64xf32, 4xi64) + # pd_op.reshape: (1x16x20x64xf32) <- (16x20x64xf32, 4xi64) reshape_54 = paddle._C_ops.reshape(bmm_7, full_int_array_5) del bmm_7 - # pd_op.transpose: (1x10x16x64xf32) <- (1x16x10x64xf32) + # pd_op.transpose: (1x20x16x64xf32) <- (1x16x20x64xf32) transpose_19 = paddle._C_ops.transpose(reshape_54, [0, 2, 1, 3]) del reshape_54 - # pd_op.reshape: (1x10x1024xf32) <- (1x10x16x64xf32, 3xi64) + # pd_op.reshape: (1x20x1024xf32) <- (1x20x16x64xf32, 3xi64) reshape_55 = paddle._C_ops.reshape(transpose_19, full_int_array_6) del transpose_19 - # pd_op.matmul: (1x10x1024xf32) <- (1x10x1024xf32, 1024x1024xf32) + # pd_op.matmul: (1x20x1024xf32) <- (1x20x1024xf32, 1024x1024xf32) matmul_21 = paddle._C_ops.matmul(reshape_55, parameter_481, False, False) del parameter_481, reshape_55 - # pd_op.add: (1x10x1024xf32) <- (1x10x1024xf32, 1024xf32) + # pd_op.add: (1x20x1024xf32) <- (1x20x1024xf32, 1024xf32) add_32 = paddle._C_ops.add(matmul_21, parameter_480) del matmul_21, parameter_480 - # pd_op.dropout: (1x10x1024xf32, 1x10x1024xui8) <- (1x10x1024xf32, None, 1xf32) + # pd_op.dropout: (1x20x1024xf32, 1x20x1024xui8) <- (1x20x1024xf32, None, 1xf32) dropout_28, dropout_29 = (lambda x, f: f(x))( paddle._C_ops.dropout( add_32, None, full_8, True, "upscale_in_train", 0, False @@ -1389,11 +1389,11 @@ def forward( ) del add_32 - # pd_op.add: (1x10x1024xf32) <- (1x10x1024xf32, 1x10x1024xf32) + # pd_op.add: (1x20x1024xf32) <- (1x20x1024xf32, 1x20x1024xf32) add_33 = paddle._C_ops.add(dropout_28, layer_norm_18) del dropout_28, layer_norm_18 - # pd_op.layer_norm: (1x10x1024xf32, 1x10xf32, 1x10xf32) <- (1x10x1024xf32, 1024xf32, 1024xf32) + # pd_op.layer_norm: (1x20x1024xf32, 1x20xf32, 1x20xf32) <- (1x20x1024xf32, 1024xf32, 1024xf32) layer_norm_21, layer_norm_22, layer_norm_23 = (lambda x, f: f(x))( paddle._C_ops.layer_norm( add_33, parameter_479, parameter_478, float("1e-05"), 2 @@ -1402,19 +1402,19 @@ def forward( ) del add_33, parameter_478, parameter_479 - # pd_op.matmul: (1x10x4096xf32) <- (1x10x1024xf32, 1024x4096xf32) + # pd_op.matmul: (1x20x4096xf32) <- (1x20x1024xf32, 1024x4096xf32) matmul_22 = paddle._C_ops.matmul(layer_norm_21, parameter_477, False, False) del parameter_477 - # pd_op.add: (1x10x4096xf32) <- (1x10x4096xf32, 4096xf32) + # pd_op.add: (1x20x4096xf32) <- (1x20x4096xf32, 4096xf32) add_34 = paddle._C_ops.add(matmul_22, parameter_476) del matmul_22, parameter_476 - # pd_op.gelu: (1x10x4096xf32) <- (1x10x4096xf32) + # pd_op.gelu: (1x20x4096xf32) <- (1x20x4096xf32) gelu_3 = paddle._C_ops.gelu(add_34, False) del add_34 - # pd_op.dropout: (1x10x4096xf32, 1x10x4096xui8) <- (1x10x4096xf32, None, 1xf32) + # pd_op.dropout: (1x20x4096xf32, 1x20x4096xui8) <- (1x20x4096xf32, None, 1xf32) dropout_30, dropout_31 = (lambda x, f: f(x))( paddle._C_ops.dropout( gelu_3, None, full_8, True, "upscale_in_train", 0, False @@ -1423,15 +1423,15 @@ def forward( ) del gelu_3 - # pd_op.matmul: (1x10x1024xf32) <- (1x10x4096xf32, 4096x1024xf32) + # pd_op.matmul: (1x20x1024xf32) <- (1x20x4096xf32, 4096x1024xf32) matmul_23 = paddle._C_ops.matmul(dropout_30, parameter_475, False, False) del dropout_30, parameter_475 - # pd_op.add: (1x10x1024xf32) <- (1x10x1024xf32, 1024xf32) + # pd_op.add: (1x20x1024xf32) <- (1x20x1024xf32, 1024xf32) add_35 = paddle._C_ops.add(matmul_23, parameter_474) del matmul_23, parameter_474 - # pd_op.dropout: (1x10x1024xf32, 1x10x1024xui8) <- (1x10x1024xf32, None, 1xf32) + # pd_op.dropout: (1x20x1024xf32, 1x20x1024xui8) <- (1x20x1024xf32, None, 1xf32) dropout_32, dropout_33 = (lambda x, f: f(x))( paddle._C_ops.dropout( add_35, None, full_8, True, "upscale_in_train", 0, False @@ -1440,11 +1440,11 @@ def forward( ) del add_35 - # pd_op.add: (1x10x1024xf32) <- (1x10x1024xf32, 1x10x1024xf32) + # pd_op.add: (1x20x1024xf32) <- (1x20x1024xf32, 1x20x1024xf32) add_36 = paddle._C_ops.add(dropout_32, layer_norm_21) del dropout_32, layer_norm_21 - # pd_op.layer_norm: (1x10x1024xf32, 1x10xf32, 1x10xf32) <- (1x10x1024xf32, 1024xf32, 1024xf32) + # pd_op.layer_norm: (1x20x1024xf32, 1x20xf32, 1x20xf32) <- (1x20x1024xf32, 1024xf32, 1024xf32) layer_norm_24, layer_norm_25, layer_norm_26 = (lambda x, f: f(x))( paddle._C_ops.layer_norm( add_36, parameter_473, parameter_472, float("1e-05"), 2 @@ -1453,87 +1453,87 @@ def forward( ) del add_36, parameter_472, parameter_473 - # pd_op.matmul: (1x10x1024xf32) <- (1x10x1024xf32, 1024x1024xf32) + # pd_op.matmul: (1x20x1024xf32) <- (1x20x1024xf32, 1024x1024xf32) matmul_24 = paddle._C_ops.matmul(layer_norm_24, parameter_467, False, False) del parameter_467 - # pd_op.add: (1x10x1024xf32) <- (1x10x1024xf32, 1024xf32) + # pd_op.add: (1x20x1024xf32) <- (1x20x1024xf32, 1024xf32) add_37 = paddle._C_ops.add(matmul_24, parameter_466) del matmul_24, parameter_466 - # pd_op.scale: (1x10x1024xf32) <- (1x10x1024xf32, 1xf32) + # pd_op.scale: (1x20x1024xf32) <- (1x20x1024xf32, 1xf32) scale_7 = paddle._C_ops.scale(add_37, full_9, float("0"), True) del add_37 - # pd_op.matmul: (1x10x1024xf32) <- (1x10x1024xf32, 1024x1024xf32) + # pd_op.matmul: (1x20x1024xf32) <- (1x20x1024xf32, 1024x1024xf32) matmul_25 = paddle._C_ops.matmul(layer_norm_24, parameter_471, False, False) del parameter_471 - # pd_op.add: (1x10x1024xf32) <- (1x10x1024xf32, 1024xf32) + # pd_op.add: (1x20x1024xf32) <- (1x20x1024xf32, 1024xf32) add_38 = paddle._C_ops.add(matmul_25, parameter_470) del matmul_25, parameter_470 - # pd_op.reshape: (1x10x16x64xf32) <- (1x10x1024xf32, 4xi64) + # pd_op.reshape: (1x20x16x64xf32) <- (1x20x1024xf32, 4xi64) reshape_56 = paddle._C_ops.reshape(add_38, full_int_array_2) del add_38 - # pd_op.transpose: (1x16x10x64xf32) <- (1x10x16x64xf32) + # pd_op.transpose: (1x16x20x64xf32) <- (1x20x16x64xf32) transpose_20 = paddle._C_ops.transpose(reshape_56, [0, 2, 1, 3]) del reshape_56 - # pd_op.matmul: (1x10x1024xf32) <- (1x10x1024xf32, 1024x1024xf32) + # pd_op.matmul: (1x20x1024xf32) <- (1x20x1024xf32, 1024x1024xf32) matmul_26 = paddle._C_ops.matmul(layer_norm_24, parameter_469, False, False) del parameter_469 - # pd_op.add: (1x10x1024xf32) <- (1x10x1024xf32, 1024xf32) + # pd_op.add: (1x20x1024xf32) <- (1x20x1024xf32, 1024xf32) add_39 = paddle._C_ops.add(matmul_26, parameter_468) del matmul_26, parameter_468 - # pd_op.reshape: (1x10x16x64xf32) <- (1x10x1024xf32, 4xi64) + # pd_op.reshape: (1x20x16x64xf32) <- (1x20x1024xf32, 4xi64) reshape_57 = paddle._C_ops.reshape(add_39, full_int_array_2) del add_39 - # pd_op.transpose: (1x16x10x64xf32) <- (1x10x16x64xf32) + # pd_op.transpose: (1x16x20x64xf32) <- (1x20x16x64xf32) transpose_21 = paddle._C_ops.transpose(reshape_57, [0, 2, 1, 3]) del reshape_57 - # pd_op.reshape: (1x10x16x64xf32) <- (1x10x1024xf32, 4xi64) + # pd_op.reshape: (1x20x16x64xf32) <- (1x20x1024xf32, 4xi64) reshape_58 = paddle._C_ops.reshape(scale_7, full_int_array_3) del scale_7 - # pd_op.transpose: (1x16x10x64xf32) <- (1x10x16x64xf32) + # pd_op.transpose: (1x16x20x64xf32) <- (1x20x16x64xf32) transpose_22 = paddle._C_ops.transpose(reshape_58, [0, 2, 1, 3]) del reshape_58 - # pd_op.reshape: (16x10x64xf32) <- (1x16x10x64xf32, 3xi64) + # pd_op.reshape: (16x20x64xf32) <- (1x16x20x64xf32, 3xi64) reshape_59 = paddle._C_ops.reshape(transpose_22, full_int_array_4) del transpose_22 - # pd_op.reshape: (16x10x64xf32) <- (1x16x10x64xf32, 3xi64) + # pd_op.reshape: (16x20x64xf32) <- (1x16x20x64xf32, 3xi64) reshape_60 = paddle._C_ops.reshape(transpose_20, full_int_array_4) del transpose_20 - # pd_op.reshape: (16x10x64xf32) <- (1x16x10x64xf32, 3xi64) + # pd_op.reshape: (16x20x64xf32) <- (1x16x20x64xf32, 3xi64) reshape_61 = paddle._C_ops.reshape(transpose_21, full_int_array_4) del transpose_21 - # pd_op.transpose: (16x64x10xf32) <- (16x10x64xf32) + # pd_op.transpose: (16x64x20xf32) <- (16x20x64xf32) transpose_23 = paddle._C_ops.transpose(reshape_60, [0, 2, 1]) del reshape_60 - # pd_op.bmm: (16x10x10xf32) <- (16x10x64xf32, 16x64x10xf32) + # pd_op.bmm: (16x20x20xf32) <- (16x20x64xf32, 16x64x20xf32) bmm_8 = paddle._C_ops.bmm(reshape_59, transpose_23) del reshape_59, transpose_23 - # pd_op.add: (16x10x10xf32) <- (16x10x10xf32, 16x1x10xf32) + # pd_op.add: (16x20x20xf32) <- (16x20x20xf32, 16x1x20xf32) add_40 = paddle._C_ops.add(bmm_8, cast_2) del bmm_8 - # pd_op.softmax: (16x10x10xf32) <- (16x10x10xf32) + # pd_op.softmax: (16x20x20xf32) <- (16x20x20xf32) softmax_4 = paddle._C_ops.softmax(add_40, -1) del add_40 - # pd_op.dropout: (16x10x10xf32, 16x10x10xui8) <- (16x10x10xf32, None, 1xf32) + # pd_op.dropout: (16x20x20xf32, 16x20x20xui8) <- (16x20x20xf32, None, 1xf32) dropout_34, dropout_35 = (lambda x, f: f(x))( paddle._C_ops.dropout( softmax_4, None, full_8, True, "upscale_in_train", 0, False @@ -1542,31 +1542,31 @@ def forward( ) del softmax_4 - # pd_op.bmm: (16x10x64xf32) <- (16x10x10xf32, 16x10x64xf32) + # pd_op.bmm: (16x20x64xf32) <- (16x20x20xf32, 16x20x64xf32) bmm_9 = paddle._C_ops.bmm(dropout_34, reshape_61) del dropout_34, reshape_61 - # pd_op.reshape: (1x16x10x64xf32) <- (16x10x64xf32, 4xi64) + # pd_op.reshape: (1x16x20x64xf32) <- (16x20x64xf32, 4xi64) reshape_62 = paddle._C_ops.reshape(bmm_9, full_int_array_5) del bmm_9 - # pd_op.transpose: (1x10x16x64xf32) <- (1x16x10x64xf32) + # pd_op.transpose: (1x20x16x64xf32) <- (1x16x20x64xf32) transpose_24 = paddle._C_ops.transpose(reshape_62, [0, 2, 1, 3]) del reshape_62 - # pd_op.reshape: (1x10x1024xf32) <- (1x10x16x64xf32, 3xi64) + # pd_op.reshape: (1x20x1024xf32) <- (1x20x16x64xf32, 3xi64) reshape_63 = paddle._C_ops.reshape(transpose_24, full_int_array_6) del transpose_24 - # pd_op.matmul: (1x10x1024xf32) <- (1x10x1024xf32, 1024x1024xf32) + # pd_op.matmul: (1x20x1024xf32) <- (1x20x1024xf32, 1024x1024xf32) matmul_27 = paddle._C_ops.matmul(reshape_63, parameter_465, False, False) del parameter_465, reshape_63 - # pd_op.add: (1x10x1024xf32) <- (1x10x1024xf32, 1024xf32) + # pd_op.add: (1x20x1024xf32) <- (1x20x1024xf32, 1024xf32) add_41 = paddle._C_ops.add(matmul_27, parameter_464) del matmul_27, parameter_464 - # pd_op.dropout: (1x10x1024xf32, 1x10x1024xui8) <- (1x10x1024xf32, None, 1xf32) + # pd_op.dropout: (1x20x1024xf32, 1x20x1024xui8) <- (1x20x1024xf32, None, 1xf32) dropout_36, dropout_37 = (lambda x, f: f(x))( paddle._C_ops.dropout( add_41, None, full_8, True, "upscale_in_train", 0, False @@ -1575,11 +1575,11 @@ def forward( ) del add_41 - # pd_op.add: (1x10x1024xf32) <- (1x10x1024xf32, 1x10x1024xf32) + # pd_op.add: (1x20x1024xf32) <- (1x20x1024xf32, 1x20x1024xf32) add_42 = paddle._C_ops.add(dropout_36, layer_norm_24) del dropout_36, layer_norm_24 - # pd_op.layer_norm: (1x10x1024xf32, 1x10xf32, 1x10xf32) <- (1x10x1024xf32, 1024xf32, 1024xf32) + # pd_op.layer_norm: (1x20x1024xf32, 1x20xf32, 1x20xf32) <- (1x20x1024xf32, 1024xf32, 1024xf32) layer_norm_27, layer_norm_28, layer_norm_29 = (lambda x, f: f(x))( paddle._C_ops.layer_norm( add_42, parameter_463, parameter_462, float("1e-05"), 2 @@ -1588,19 +1588,19 @@ def forward( ) del add_42, parameter_462, parameter_463 - # pd_op.matmul: (1x10x4096xf32) <- (1x10x1024xf32, 1024x4096xf32) + # pd_op.matmul: (1x20x4096xf32) <- (1x20x1024xf32, 1024x4096xf32) matmul_28 = paddle._C_ops.matmul(layer_norm_27, parameter_461, False, False) del parameter_461 - # pd_op.add: (1x10x4096xf32) <- (1x10x4096xf32, 4096xf32) + # pd_op.add: (1x20x4096xf32) <- (1x20x4096xf32, 4096xf32) add_43 = paddle._C_ops.add(matmul_28, parameter_460) del matmul_28, parameter_460 - # pd_op.gelu: (1x10x4096xf32) <- (1x10x4096xf32) + # pd_op.gelu: (1x20x4096xf32) <- (1x20x4096xf32) gelu_4 = paddle._C_ops.gelu(add_43, False) del add_43 - # pd_op.dropout: (1x10x4096xf32, 1x10x4096xui8) <- (1x10x4096xf32, None, 1xf32) + # pd_op.dropout: (1x20x4096xf32, 1x20x4096xui8) <- (1x20x4096xf32, None, 1xf32) dropout_38, dropout_39 = (lambda x, f: f(x))( paddle._C_ops.dropout( gelu_4, None, full_8, True, "upscale_in_train", 0, False @@ -1609,15 +1609,15 @@ def forward( ) del gelu_4 - # pd_op.matmul: (1x10x1024xf32) <- (1x10x4096xf32, 4096x1024xf32) + # pd_op.matmul: (1x20x1024xf32) <- (1x20x4096xf32, 4096x1024xf32) matmul_29 = paddle._C_ops.matmul(dropout_38, parameter_459, False, False) del dropout_38, parameter_459 - # pd_op.add: (1x10x1024xf32) <- (1x10x1024xf32, 1024xf32) + # pd_op.add: (1x20x1024xf32) <- (1x20x1024xf32, 1024xf32) add_44 = paddle._C_ops.add(matmul_29, parameter_458) del matmul_29, parameter_458 - # pd_op.dropout: (1x10x1024xf32, 1x10x1024xui8) <- (1x10x1024xf32, None, 1xf32) + # pd_op.dropout: (1x20x1024xf32, 1x20x1024xui8) <- (1x20x1024xf32, None, 1xf32) dropout_40, dropout_41 = (lambda x, f: f(x))( paddle._C_ops.dropout( add_44, None, full_8, True, "upscale_in_train", 0, False @@ -1626,11 +1626,11 @@ def forward( ) del add_44 - # pd_op.add: (1x10x1024xf32) <- (1x10x1024xf32, 1x10x1024xf32) + # pd_op.add: (1x20x1024xf32) <- (1x20x1024xf32, 1x20x1024xf32) add_45 = paddle._C_ops.add(dropout_40, layer_norm_27) del dropout_40, layer_norm_27 - # pd_op.layer_norm: (1x10x1024xf32, 1x10xf32, 1x10xf32) <- (1x10x1024xf32, 1024xf32, 1024xf32) + # pd_op.layer_norm: (1x20x1024xf32, 1x20xf32, 1x20xf32) <- (1x20x1024xf32, 1024xf32, 1024xf32) layer_norm_30, layer_norm_31, layer_norm_32 = (lambda x, f: f(x))( paddle._C_ops.layer_norm( add_45, parameter_457, parameter_456, float("1e-05"), 2 @@ -1639,87 +1639,87 @@ def forward( ) del add_45, parameter_456, parameter_457 - # pd_op.matmul: (1x10x1024xf32) <- (1x10x1024xf32, 1024x1024xf32) + # pd_op.matmul: (1x20x1024xf32) <- (1x20x1024xf32, 1024x1024xf32) matmul_30 = paddle._C_ops.matmul(layer_norm_30, parameter_451, False, False) del parameter_451 - # pd_op.add: (1x10x1024xf32) <- (1x10x1024xf32, 1024xf32) + # pd_op.add: (1x20x1024xf32) <- (1x20x1024xf32, 1024xf32) add_46 = paddle._C_ops.add(matmul_30, parameter_450) del matmul_30, parameter_450 - # pd_op.scale: (1x10x1024xf32) <- (1x10x1024xf32, 1xf32) + # pd_op.scale: (1x20x1024xf32) <- (1x20x1024xf32, 1xf32) scale_8 = paddle._C_ops.scale(add_46, full_9, float("0"), True) del add_46 - # pd_op.matmul: (1x10x1024xf32) <- (1x10x1024xf32, 1024x1024xf32) + # pd_op.matmul: (1x20x1024xf32) <- (1x20x1024xf32, 1024x1024xf32) matmul_31 = paddle._C_ops.matmul(layer_norm_30, parameter_455, False, False) del parameter_455 - # pd_op.add: (1x10x1024xf32) <- (1x10x1024xf32, 1024xf32) + # pd_op.add: (1x20x1024xf32) <- (1x20x1024xf32, 1024xf32) add_47 = paddle._C_ops.add(matmul_31, parameter_454) del matmul_31, parameter_454 - # pd_op.reshape: (1x10x16x64xf32) <- (1x10x1024xf32, 4xi64) + # pd_op.reshape: (1x20x16x64xf32) <- (1x20x1024xf32, 4xi64) reshape_64 = paddle._C_ops.reshape(add_47, full_int_array_2) del add_47 - # pd_op.transpose: (1x16x10x64xf32) <- (1x10x16x64xf32) + # pd_op.transpose: (1x16x20x64xf32) <- (1x20x16x64xf32) transpose_25 = paddle._C_ops.transpose(reshape_64, [0, 2, 1, 3]) del reshape_64 - # pd_op.matmul: (1x10x1024xf32) <- (1x10x1024xf32, 1024x1024xf32) + # pd_op.matmul: (1x20x1024xf32) <- (1x20x1024xf32, 1024x1024xf32) matmul_32 = paddle._C_ops.matmul(layer_norm_30, parameter_453, False, False) del parameter_453 - # pd_op.add: (1x10x1024xf32) <- (1x10x1024xf32, 1024xf32) + # pd_op.add: (1x20x1024xf32) <- (1x20x1024xf32, 1024xf32) add_48 = paddle._C_ops.add(matmul_32, parameter_452) del matmul_32, parameter_452 - # pd_op.reshape: (1x10x16x64xf32) <- (1x10x1024xf32, 4xi64) + # pd_op.reshape: (1x20x16x64xf32) <- (1x20x1024xf32, 4xi64) reshape_65 = paddle._C_ops.reshape(add_48, full_int_array_2) del add_48 - # pd_op.transpose: (1x16x10x64xf32) <- (1x10x16x64xf32) + # pd_op.transpose: (1x16x20x64xf32) <- (1x20x16x64xf32) transpose_26 = paddle._C_ops.transpose(reshape_65, [0, 2, 1, 3]) del reshape_65 - # pd_op.reshape: (1x10x16x64xf32) <- (1x10x1024xf32, 4xi64) + # pd_op.reshape: (1x20x16x64xf32) <- (1x20x1024xf32, 4xi64) reshape_66 = paddle._C_ops.reshape(scale_8, full_int_array_3) del scale_8 - # pd_op.transpose: (1x16x10x64xf32) <- (1x10x16x64xf32) + # pd_op.transpose: (1x16x20x64xf32) <- (1x20x16x64xf32) transpose_27 = paddle._C_ops.transpose(reshape_66, [0, 2, 1, 3]) del reshape_66 - # pd_op.reshape: (16x10x64xf32) <- (1x16x10x64xf32, 3xi64) + # pd_op.reshape: (16x20x64xf32) <- (1x16x20x64xf32, 3xi64) reshape_67 = paddle._C_ops.reshape(transpose_27, full_int_array_4) del transpose_27 - # pd_op.reshape: (16x10x64xf32) <- (1x16x10x64xf32, 3xi64) + # pd_op.reshape: (16x20x64xf32) <- (1x16x20x64xf32, 3xi64) reshape_68 = paddle._C_ops.reshape(transpose_25, full_int_array_4) del transpose_25 - # pd_op.reshape: (16x10x64xf32) <- (1x16x10x64xf32, 3xi64) + # pd_op.reshape: (16x20x64xf32) <- (1x16x20x64xf32, 3xi64) reshape_69 = paddle._C_ops.reshape(transpose_26, full_int_array_4) del transpose_26 - # pd_op.transpose: (16x64x10xf32) <- (16x10x64xf32) + # pd_op.transpose: (16x64x20xf32) <- (16x20x64xf32) transpose_28 = paddle._C_ops.transpose(reshape_68, [0, 2, 1]) del reshape_68 - # pd_op.bmm: (16x10x10xf32) <- (16x10x64xf32, 16x64x10xf32) + # pd_op.bmm: (16x20x20xf32) <- (16x20x64xf32, 16x64x20xf32) bmm_10 = paddle._C_ops.bmm(reshape_67, transpose_28) del reshape_67, transpose_28 - # pd_op.add: (16x10x10xf32) <- (16x10x10xf32, 16x1x10xf32) + # pd_op.add: (16x20x20xf32) <- (16x20x20xf32, 16x1x20xf32) add_49 = paddle._C_ops.add(bmm_10, cast_2) del bmm_10 - # pd_op.softmax: (16x10x10xf32) <- (16x10x10xf32) + # pd_op.softmax: (16x20x20xf32) <- (16x20x20xf32) softmax_5 = paddle._C_ops.softmax(add_49, -1) del add_49 - # pd_op.dropout: (16x10x10xf32, 16x10x10xui8) <- (16x10x10xf32, None, 1xf32) + # pd_op.dropout: (16x20x20xf32, 16x20x20xui8) <- (16x20x20xf32, None, 1xf32) dropout_42, dropout_43 = (lambda x, f: f(x))( paddle._C_ops.dropout( softmax_5, None, full_8, True, "upscale_in_train", 0, False @@ -1728,31 +1728,31 @@ def forward( ) del softmax_5 - # pd_op.bmm: (16x10x64xf32) <- (16x10x10xf32, 16x10x64xf32) + # pd_op.bmm: (16x20x64xf32) <- (16x20x20xf32, 16x20x64xf32) bmm_11 = paddle._C_ops.bmm(dropout_42, reshape_69) del dropout_42, reshape_69 - # pd_op.reshape: (1x16x10x64xf32) <- (16x10x64xf32, 4xi64) + # pd_op.reshape: (1x16x20x64xf32) <- (16x20x64xf32, 4xi64) reshape_70 = paddle._C_ops.reshape(bmm_11, full_int_array_5) del bmm_11 - # pd_op.transpose: (1x10x16x64xf32) <- (1x16x10x64xf32) + # pd_op.transpose: (1x20x16x64xf32) <- (1x16x20x64xf32) transpose_29 = paddle._C_ops.transpose(reshape_70, [0, 2, 1, 3]) del reshape_70 - # pd_op.reshape: (1x10x1024xf32) <- (1x10x16x64xf32, 3xi64) + # pd_op.reshape: (1x20x1024xf32) <- (1x20x16x64xf32, 3xi64) reshape_71 = paddle._C_ops.reshape(transpose_29, full_int_array_6) del transpose_29 - # pd_op.matmul: (1x10x1024xf32) <- (1x10x1024xf32, 1024x1024xf32) + # pd_op.matmul: (1x20x1024xf32) <- (1x20x1024xf32, 1024x1024xf32) matmul_33 = paddle._C_ops.matmul(reshape_71, parameter_449, False, False) del parameter_449, reshape_71 - # pd_op.add: (1x10x1024xf32) <- (1x10x1024xf32, 1024xf32) + # pd_op.add: (1x20x1024xf32) <- (1x20x1024xf32, 1024xf32) add_50 = paddle._C_ops.add(matmul_33, parameter_448) del matmul_33, parameter_448 - # pd_op.dropout: (1x10x1024xf32, 1x10x1024xui8) <- (1x10x1024xf32, None, 1xf32) + # pd_op.dropout: (1x20x1024xf32, 1x20x1024xui8) <- (1x20x1024xf32, None, 1xf32) dropout_44, dropout_45 = (lambda x, f: f(x))( paddle._C_ops.dropout( add_50, None, full_8, True, "upscale_in_train", 0, False @@ -1761,11 +1761,11 @@ def forward( ) del add_50 - # pd_op.add: (1x10x1024xf32) <- (1x10x1024xf32, 1x10x1024xf32) + # pd_op.add: (1x20x1024xf32) <- (1x20x1024xf32, 1x20x1024xf32) add_51 = paddle._C_ops.add(dropout_44, layer_norm_30) del dropout_44, layer_norm_30 - # pd_op.layer_norm: (1x10x1024xf32, 1x10xf32, 1x10xf32) <- (1x10x1024xf32, 1024xf32, 1024xf32) + # pd_op.layer_norm: (1x20x1024xf32, 1x20xf32, 1x20xf32) <- (1x20x1024xf32, 1024xf32, 1024xf32) layer_norm_33, layer_norm_34, layer_norm_35 = (lambda x, f: f(x))( paddle._C_ops.layer_norm( add_51, parameter_447, parameter_446, float("1e-05"), 2 @@ -1774,19 +1774,19 @@ def forward( ) del add_51, parameter_446, parameter_447 - # pd_op.matmul: (1x10x4096xf32) <- (1x10x1024xf32, 1024x4096xf32) + # pd_op.matmul: (1x20x4096xf32) <- (1x20x1024xf32, 1024x4096xf32) matmul_34 = paddle._C_ops.matmul(layer_norm_33, parameter_445, False, False) del parameter_445 - # pd_op.add: (1x10x4096xf32) <- (1x10x4096xf32, 4096xf32) + # pd_op.add: (1x20x4096xf32) <- (1x20x4096xf32, 4096xf32) add_52 = paddle._C_ops.add(matmul_34, parameter_444) del matmul_34, parameter_444 - # pd_op.gelu: (1x10x4096xf32) <- (1x10x4096xf32) + # pd_op.gelu: (1x20x4096xf32) <- (1x20x4096xf32) gelu_5 = paddle._C_ops.gelu(add_52, False) del add_52 - # pd_op.dropout: (1x10x4096xf32, 1x10x4096xui8) <- (1x10x4096xf32, None, 1xf32) + # pd_op.dropout: (1x20x4096xf32, 1x20x4096xui8) <- (1x20x4096xf32, None, 1xf32) dropout_46, dropout_47 = (lambda x, f: f(x))( paddle._C_ops.dropout( gelu_5, None, full_8, True, "upscale_in_train", 0, False @@ -1795,15 +1795,15 @@ def forward( ) del gelu_5 - # pd_op.matmul: (1x10x1024xf32) <- (1x10x4096xf32, 4096x1024xf32) + # pd_op.matmul: (1x20x1024xf32) <- (1x20x4096xf32, 4096x1024xf32) matmul_35 = paddle._C_ops.matmul(dropout_46, parameter_443, False, False) del dropout_46, parameter_443 - # pd_op.add: (1x10x1024xf32) <- (1x10x1024xf32, 1024xf32) + # pd_op.add: (1x20x1024xf32) <- (1x20x1024xf32, 1024xf32) add_53 = paddle._C_ops.add(matmul_35, parameter_442) del matmul_35, parameter_442 - # pd_op.dropout: (1x10x1024xf32, 1x10x1024xui8) <- (1x10x1024xf32, None, 1xf32) + # pd_op.dropout: (1x20x1024xf32, 1x20x1024xui8) <- (1x20x1024xf32, None, 1xf32) dropout_48, dropout_49 = (lambda x, f: f(x))( paddle._C_ops.dropout( add_53, None, full_8, True, "upscale_in_train", 0, False @@ -1812,11 +1812,11 @@ def forward( ) del add_53 - # pd_op.add: (1x10x1024xf32) <- (1x10x1024xf32, 1x10x1024xf32) + # pd_op.add: (1x20x1024xf32) <- (1x20x1024xf32, 1x20x1024xf32) add_54 = paddle._C_ops.add(dropout_48, layer_norm_33) del dropout_48, layer_norm_33 - # pd_op.layer_norm: (1x10x1024xf32, 1x10xf32, 1x10xf32) <- (1x10x1024xf32, 1024xf32, 1024xf32) + # pd_op.layer_norm: (1x20x1024xf32, 1x20xf32, 1x20xf32) <- (1x20x1024xf32, 1024xf32, 1024xf32) layer_norm_36, layer_norm_37, layer_norm_38 = (lambda x, f: f(x))( paddle._C_ops.layer_norm( add_54, parameter_441, parameter_440, float("1e-05"), 2 @@ -1825,87 +1825,87 @@ def forward( ) del add_54, parameter_440, parameter_441 - # pd_op.matmul: (1x10x1024xf32) <- (1x10x1024xf32, 1024x1024xf32) + # pd_op.matmul: (1x20x1024xf32) <- (1x20x1024xf32, 1024x1024xf32) matmul_36 = paddle._C_ops.matmul(layer_norm_36, parameter_435, False, False) del parameter_435 - # pd_op.add: (1x10x1024xf32) <- (1x10x1024xf32, 1024xf32) + # pd_op.add: (1x20x1024xf32) <- (1x20x1024xf32, 1024xf32) add_55 = paddle._C_ops.add(matmul_36, parameter_434) del matmul_36, parameter_434 - # pd_op.scale: (1x10x1024xf32) <- (1x10x1024xf32, 1xf32) + # pd_op.scale: (1x20x1024xf32) <- (1x20x1024xf32, 1xf32) scale_9 = paddle._C_ops.scale(add_55, full_9, float("0"), True) del add_55 - # pd_op.matmul: (1x10x1024xf32) <- (1x10x1024xf32, 1024x1024xf32) + # pd_op.matmul: (1x20x1024xf32) <- (1x20x1024xf32, 1024x1024xf32) matmul_37 = paddle._C_ops.matmul(layer_norm_36, parameter_439, False, False) del parameter_439 - # pd_op.add: (1x10x1024xf32) <- (1x10x1024xf32, 1024xf32) + # pd_op.add: (1x20x1024xf32) <- (1x20x1024xf32, 1024xf32) add_56 = paddle._C_ops.add(matmul_37, parameter_438) del matmul_37, parameter_438 - # pd_op.reshape: (1x10x16x64xf32) <- (1x10x1024xf32, 4xi64) + # pd_op.reshape: (1x20x16x64xf32) <- (1x20x1024xf32, 4xi64) reshape_72 = paddle._C_ops.reshape(add_56, full_int_array_2) del add_56 - # pd_op.transpose: (1x16x10x64xf32) <- (1x10x16x64xf32) + # pd_op.transpose: (1x16x20x64xf32) <- (1x20x16x64xf32) transpose_30 = paddle._C_ops.transpose(reshape_72, [0, 2, 1, 3]) del reshape_72 - # pd_op.matmul: (1x10x1024xf32) <- (1x10x1024xf32, 1024x1024xf32) + # pd_op.matmul: (1x20x1024xf32) <- (1x20x1024xf32, 1024x1024xf32) matmul_38 = paddle._C_ops.matmul(layer_norm_36, parameter_437, False, False) del parameter_437 - # pd_op.add: (1x10x1024xf32) <- (1x10x1024xf32, 1024xf32) + # pd_op.add: (1x20x1024xf32) <- (1x20x1024xf32, 1024xf32) add_57 = paddle._C_ops.add(matmul_38, parameter_436) del matmul_38, parameter_436 - # pd_op.reshape: (1x10x16x64xf32) <- (1x10x1024xf32, 4xi64) + # pd_op.reshape: (1x20x16x64xf32) <- (1x20x1024xf32, 4xi64) reshape_73 = paddle._C_ops.reshape(add_57, full_int_array_2) del add_57 - # pd_op.transpose: (1x16x10x64xf32) <- (1x10x16x64xf32) + # pd_op.transpose: (1x16x20x64xf32) <- (1x20x16x64xf32) transpose_31 = paddle._C_ops.transpose(reshape_73, [0, 2, 1, 3]) del reshape_73 - # pd_op.reshape: (1x10x16x64xf32) <- (1x10x1024xf32, 4xi64) + # pd_op.reshape: (1x20x16x64xf32) <- (1x20x1024xf32, 4xi64) reshape_74 = paddle._C_ops.reshape(scale_9, full_int_array_3) del scale_9 - # pd_op.transpose: (1x16x10x64xf32) <- (1x10x16x64xf32) + # pd_op.transpose: (1x16x20x64xf32) <- (1x20x16x64xf32) transpose_32 = paddle._C_ops.transpose(reshape_74, [0, 2, 1, 3]) del reshape_74 - # pd_op.reshape: (16x10x64xf32) <- (1x16x10x64xf32, 3xi64) + # pd_op.reshape: (16x20x64xf32) <- (1x16x20x64xf32, 3xi64) reshape_75 = paddle._C_ops.reshape(transpose_32, full_int_array_4) del transpose_32 - # pd_op.reshape: (16x10x64xf32) <- (1x16x10x64xf32, 3xi64) + # pd_op.reshape: (16x20x64xf32) <- (1x16x20x64xf32, 3xi64) reshape_76 = paddle._C_ops.reshape(transpose_30, full_int_array_4) del transpose_30 - # pd_op.reshape: (16x10x64xf32) <- (1x16x10x64xf32, 3xi64) + # pd_op.reshape: (16x20x64xf32) <- (1x16x20x64xf32, 3xi64) reshape_77 = paddle._C_ops.reshape(transpose_31, full_int_array_4) del transpose_31 - # pd_op.transpose: (16x64x10xf32) <- (16x10x64xf32) + # pd_op.transpose: (16x64x20xf32) <- (16x20x64xf32) transpose_33 = paddle._C_ops.transpose(reshape_76, [0, 2, 1]) del reshape_76 - # pd_op.bmm: (16x10x10xf32) <- (16x10x64xf32, 16x64x10xf32) + # pd_op.bmm: (16x20x20xf32) <- (16x20x64xf32, 16x64x20xf32) bmm_12 = paddle._C_ops.bmm(reshape_75, transpose_33) del reshape_75, transpose_33 - # pd_op.add: (16x10x10xf32) <- (16x10x10xf32, 16x1x10xf32) + # pd_op.add: (16x20x20xf32) <- (16x20x20xf32, 16x1x20xf32) add_58 = paddle._C_ops.add(bmm_12, cast_2) del bmm_12 - # pd_op.softmax: (16x10x10xf32) <- (16x10x10xf32) + # pd_op.softmax: (16x20x20xf32) <- (16x20x20xf32) softmax_6 = paddle._C_ops.softmax(add_58, -1) del add_58 - # pd_op.dropout: (16x10x10xf32, 16x10x10xui8) <- (16x10x10xf32, None, 1xf32) + # pd_op.dropout: (16x20x20xf32, 16x20x20xui8) <- (16x20x20xf32, None, 1xf32) dropout_50, dropout_51 = (lambda x, f: f(x))( paddle._C_ops.dropout( softmax_6, None, full_8, True, "upscale_in_train", 0, False @@ -1914,31 +1914,31 @@ def forward( ) del softmax_6 - # pd_op.bmm: (16x10x64xf32) <- (16x10x10xf32, 16x10x64xf32) + # pd_op.bmm: (16x20x64xf32) <- (16x20x20xf32, 16x20x64xf32) bmm_13 = paddle._C_ops.bmm(dropout_50, reshape_77) del dropout_50, reshape_77 - # pd_op.reshape: (1x16x10x64xf32) <- (16x10x64xf32, 4xi64) + # pd_op.reshape: (1x16x20x64xf32) <- (16x20x64xf32, 4xi64) reshape_78 = paddle._C_ops.reshape(bmm_13, full_int_array_5) del bmm_13 - # pd_op.transpose: (1x10x16x64xf32) <- (1x16x10x64xf32) + # pd_op.transpose: (1x20x16x64xf32) <- (1x16x20x64xf32) transpose_34 = paddle._C_ops.transpose(reshape_78, [0, 2, 1, 3]) del reshape_78 - # pd_op.reshape: (1x10x1024xf32) <- (1x10x16x64xf32, 3xi64) + # pd_op.reshape: (1x20x1024xf32) <- (1x20x16x64xf32, 3xi64) reshape_79 = paddle._C_ops.reshape(transpose_34, full_int_array_6) del transpose_34 - # pd_op.matmul: (1x10x1024xf32) <- (1x10x1024xf32, 1024x1024xf32) + # pd_op.matmul: (1x20x1024xf32) <- (1x20x1024xf32, 1024x1024xf32) matmul_39 = paddle._C_ops.matmul(reshape_79, parameter_433, False, False) del parameter_433, reshape_79 - # pd_op.add: (1x10x1024xf32) <- (1x10x1024xf32, 1024xf32) + # pd_op.add: (1x20x1024xf32) <- (1x20x1024xf32, 1024xf32) add_59 = paddle._C_ops.add(matmul_39, parameter_432) del matmul_39, parameter_432 - # pd_op.dropout: (1x10x1024xf32, 1x10x1024xui8) <- (1x10x1024xf32, None, 1xf32) + # pd_op.dropout: (1x20x1024xf32, 1x20x1024xui8) <- (1x20x1024xf32, None, 1xf32) dropout_52, dropout_53 = (lambda x, f: f(x))( paddle._C_ops.dropout( add_59, None, full_8, True, "upscale_in_train", 0, False @@ -1947,11 +1947,11 @@ def forward( ) del add_59 - # pd_op.add: (1x10x1024xf32) <- (1x10x1024xf32, 1x10x1024xf32) + # pd_op.add: (1x20x1024xf32) <- (1x20x1024xf32, 1x20x1024xf32) add_60 = paddle._C_ops.add(dropout_52, layer_norm_36) del dropout_52, layer_norm_36 - # pd_op.layer_norm: (1x10x1024xf32, 1x10xf32, 1x10xf32) <- (1x10x1024xf32, 1024xf32, 1024xf32) + # pd_op.layer_norm: (1x20x1024xf32, 1x20xf32, 1x20xf32) <- (1x20x1024xf32, 1024xf32, 1024xf32) layer_norm_39, layer_norm_40, layer_norm_41 = (lambda x, f: f(x))( paddle._C_ops.layer_norm( add_60, parameter_431, parameter_430, float("1e-05"), 2 @@ -1960,19 +1960,19 @@ def forward( ) del add_60, parameter_430, parameter_431 - # pd_op.matmul: (1x10x4096xf32) <- (1x10x1024xf32, 1024x4096xf32) + # pd_op.matmul: (1x20x4096xf32) <- (1x20x1024xf32, 1024x4096xf32) matmul_40 = paddle._C_ops.matmul(layer_norm_39, parameter_429, False, False) del parameter_429 - # pd_op.add: (1x10x4096xf32) <- (1x10x4096xf32, 4096xf32) + # pd_op.add: (1x20x4096xf32) <- (1x20x4096xf32, 4096xf32) add_61 = paddle._C_ops.add(matmul_40, parameter_428) del matmul_40, parameter_428 - # pd_op.gelu: (1x10x4096xf32) <- (1x10x4096xf32) + # pd_op.gelu: (1x20x4096xf32) <- (1x20x4096xf32) gelu_6 = paddle._C_ops.gelu(add_61, False) del add_61 - # pd_op.dropout: (1x10x4096xf32, 1x10x4096xui8) <- (1x10x4096xf32, None, 1xf32) + # pd_op.dropout: (1x20x4096xf32, 1x20x4096xui8) <- (1x20x4096xf32, None, 1xf32) dropout_54, dropout_55 = (lambda x, f: f(x))( paddle._C_ops.dropout( gelu_6, None, full_8, True, "upscale_in_train", 0, False @@ -1981,15 +1981,15 @@ def forward( ) del gelu_6 - # pd_op.matmul: (1x10x1024xf32) <- (1x10x4096xf32, 4096x1024xf32) + # pd_op.matmul: (1x20x1024xf32) <- (1x20x4096xf32, 4096x1024xf32) matmul_41 = paddle._C_ops.matmul(dropout_54, parameter_427, False, False) del dropout_54, parameter_427 - # pd_op.add: (1x10x1024xf32) <- (1x10x1024xf32, 1024xf32) + # pd_op.add: (1x20x1024xf32) <- (1x20x1024xf32, 1024xf32) add_62 = paddle._C_ops.add(matmul_41, parameter_426) del matmul_41, parameter_426 - # pd_op.dropout: (1x10x1024xf32, 1x10x1024xui8) <- (1x10x1024xf32, None, 1xf32) + # pd_op.dropout: (1x20x1024xf32, 1x20x1024xui8) <- (1x20x1024xf32, None, 1xf32) dropout_56, dropout_57 = (lambda x, f: f(x))( paddle._C_ops.dropout( add_62, None, full_8, True, "upscale_in_train", 0, False @@ -1998,11 +1998,11 @@ def forward( ) del add_62 - # pd_op.add: (1x10x1024xf32) <- (1x10x1024xf32, 1x10x1024xf32) + # pd_op.add: (1x20x1024xf32) <- (1x20x1024xf32, 1x20x1024xf32) add_63 = paddle._C_ops.add(dropout_56, layer_norm_39) del dropout_56, layer_norm_39 - # pd_op.layer_norm: (1x10x1024xf32, 1x10xf32, 1x10xf32) <- (1x10x1024xf32, 1024xf32, 1024xf32) + # pd_op.layer_norm: (1x20x1024xf32, 1x20xf32, 1x20xf32) <- (1x20x1024xf32, 1024xf32, 1024xf32) layer_norm_42, layer_norm_43, layer_norm_44 = (lambda x, f: f(x))( paddle._C_ops.layer_norm( add_63, parameter_425, parameter_424, float("1e-05"), 2 @@ -2011,87 +2011,87 @@ def forward( ) del add_63, parameter_424, parameter_425 - # pd_op.matmul: (1x10x1024xf32) <- (1x10x1024xf32, 1024x1024xf32) + # pd_op.matmul: (1x20x1024xf32) <- (1x20x1024xf32, 1024x1024xf32) matmul_42 = paddle._C_ops.matmul(layer_norm_42, parameter_419, False, False) del parameter_419 - # pd_op.add: (1x10x1024xf32) <- (1x10x1024xf32, 1024xf32) + # pd_op.add: (1x20x1024xf32) <- (1x20x1024xf32, 1024xf32) add_64 = paddle._C_ops.add(matmul_42, parameter_418) del matmul_42, parameter_418 - # pd_op.scale: (1x10x1024xf32) <- (1x10x1024xf32, 1xf32) + # pd_op.scale: (1x20x1024xf32) <- (1x20x1024xf32, 1xf32) scale_10 = paddle._C_ops.scale(add_64, full_9, float("0"), True) del add_64 - # pd_op.matmul: (1x10x1024xf32) <- (1x10x1024xf32, 1024x1024xf32) + # pd_op.matmul: (1x20x1024xf32) <- (1x20x1024xf32, 1024x1024xf32) matmul_43 = paddle._C_ops.matmul(layer_norm_42, parameter_423, False, False) del parameter_423 - # pd_op.add: (1x10x1024xf32) <- (1x10x1024xf32, 1024xf32) + # pd_op.add: (1x20x1024xf32) <- (1x20x1024xf32, 1024xf32) add_65 = paddle._C_ops.add(matmul_43, parameter_422) del matmul_43, parameter_422 - # pd_op.reshape: (1x10x16x64xf32) <- (1x10x1024xf32, 4xi64) + # pd_op.reshape: (1x20x16x64xf32) <- (1x20x1024xf32, 4xi64) reshape_80 = paddle._C_ops.reshape(add_65, full_int_array_2) del add_65 - # pd_op.transpose: (1x16x10x64xf32) <- (1x10x16x64xf32) + # pd_op.transpose: (1x16x20x64xf32) <- (1x20x16x64xf32) transpose_35 = paddle._C_ops.transpose(reshape_80, [0, 2, 1, 3]) del reshape_80 - # pd_op.matmul: (1x10x1024xf32) <- (1x10x1024xf32, 1024x1024xf32) + # pd_op.matmul: (1x20x1024xf32) <- (1x20x1024xf32, 1024x1024xf32) matmul_44 = paddle._C_ops.matmul(layer_norm_42, parameter_421, False, False) del parameter_421 - # pd_op.add: (1x10x1024xf32) <- (1x10x1024xf32, 1024xf32) + # pd_op.add: (1x20x1024xf32) <- (1x20x1024xf32, 1024xf32) add_66 = paddle._C_ops.add(matmul_44, parameter_420) del matmul_44, parameter_420 - # pd_op.reshape: (1x10x16x64xf32) <- (1x10x1024xf32, 4xi64) + # pd_op.reshape: (1x20x16x64xf32) <- (1x20x1024xf32, 4xi64) reshape_81 = paddle._C_ops.reshape(add_66, full_int_array_2) del add_66 - # pd_op.transpose: (1x16x10x64xf32) <- (1x10x16x64xf32) + # pd_op.transpose: (1x16x20x64xf32) <- (1x20x16x64xf32) transpose_36 = paddle._C_ops.transpose(reshape_81, [0, 2, 1, 3]) del reshape_81 - # pd_op.reshape: (1x10x16x64xf32) <- (1x10x1024xf32, 4xi64) + # pd_op.reshape: (1x20x16x64xf32) <- (1x20x1024xf32, 4xi64) reshape_82 = paddle._C_ops.reshape(scale_10, full_int_array_3) del scale_10 - # pd_op.transpose: (1x16x10x64xf32) <- (1x10x16x64xf32) + # pd_op.transpose: (1x16x20x64xf32) <- (1x20x16x64xf32) transpose_37 = paddle._C_ops.transpose(reshape_82, [0, 2, 1, 3]) del reshape_82 - # pd_op.reshape: (16x10x64xf32) <- (1x16x10x64xf32, 3xi64) + # pd_op.reshape: (16x20x64xf32) <- (1x16x20x64xf32, 3xi64) reshape_83 = paddle._C_ops.reshape(transpose_37, full_int_array_4) del transpose_37 - # pd_op.reshape: (16x10x64xf32) <- (1x16x10x64xf32, 3xi64) + # pd_op.reshape: (16x20x64xf32) <- (1x16x20x64xf32, 3xi64) reshape_84 = paddle._C_ops.reshape(transpose_35, full_int_array_4) del transpose_35 - # pd_op.reshape: (16x10x64xf32) <- (1x16x10x64xf32, 3xi64) + # pd_op.reshape: (16x20x64xf32) <- (1x16x20x64xf32, 3xi64) reshape_85 = paddle._C_ops.reshape(transpose_36, full_int_array_4) del transpose_36 - # pd_op.transpose: (16x64x10xf32) <- (16x10x64xf32) + # pd_op.transpose: (16x64x20xf32) <- (16x20x64xf32) transpose_38 = paddle._C_ops.transpose(reshape_84, [0, 2, 1]) del reshape_84 - # pd_op.bmm: (16x10x10xf32) <- (16x10x64xf32, 16x64x10xf32) + # pd_op.bmm: (16x20x20xf32) <- (16x20x64xf32, 16x64x20xf32) bmm_14 = paddle._C_ops.bmm(reshape_83, transpose_38) del reshape_83, transpose_38 - # pd_op.add: (16x10x10xf32) <- (16x10x10xf32, 16x1x10xf32) + # pd_op.add: (16x20x20xf32) <- (16x20x20xf32, 16x1x20xf32) add_67 = paddle._C_ops.add(bmm_14, cast_2) del bmm_14 - # pd_op.softmax: (16x10x10xf32) <- (16x10x10xf32) + # pd_op.softmax: (16x20x20xf32) <- (16x20x20xf32) softmax_7 = paddle._C_ops.softmax(add_67, -1) del add_67 - # pd_op.dropout: (16x10x10xf32, 16x10x10xui8) <- (16x10x10xf32, None, 1xf32) + # pd_op.dropout: (16x20x20xf32, 16x20x20xui8) <- (16x20x20xf32, None, 1xf32) dropout_58, dropout_59 = (lambda x, f: f(x))( paddle._C_ops.dropout( softmax_7, None, full_8, True, "upscale_in_train", 0, False @@ -2100,31 +2100,31 @@ def forward( ) del softmax_7 - # pd_op.bmm: (16x10x64xf32) <- (16x10x10xf32, 16x10x64xf32) + # pd_op.bmm: (16x20x64xf32) <- (16x20x20xf32, 16x20x64xf32) bmm_15 = paddle._C_ops.bmm(dropout_58, reshape_85) del dropout_58, reshape_85 - # pd_op.reshape: (1x16x10x64xf32) <- (16x10x64xf32, 4xi64) + # pd_op.reshape: (1x16x20x64xf32) <- (16x20x64xf32, 4xi64) reshape_86 = paddle._C_ops.reshape(bmm_15, full_int_array_5) del bmm_15 - # pd_op.transpose: (1x10x16x64xf32) <- (1x16x10x64xf32) + # pd_op.transpose: (1x20x16x64xf32) <- (1x16x20x64xf32) transpose_39 = paddle._C_ops.transpose(reshape_86, [0, 2, 1, 3]) del reshape_86 - # pd_op.reshape: (1x10x1024xf32) <- (1x10x16x64xf32, 3xi64) + # pd_op.reshape: (1x20x1024xf32) <- (1x20x16x64xf32, 3xi64) reshape_87 = paddle._C_ops.reshape(transpose_39, full_int_array_6) del transpose_39 - # pd_op.matmul: (1x10x1024xf32) <- (1x10x1024xf32, 1024x1024xf32) + # pd_op.matmul: (1x20x1024xf32) <- (1x20x1024xf32, 1024x1024xf32) matmul_45 = paddle._C_ops.matmul(reshape_87, parameter_417, False, False) del parameter_417, reshape_87 - # pd_op.add: (1x10x1024xf32) <- (1x10x1024xf32, 1024xf32) + # pd_op.add: (1x20x1024xf32) <- (1x20x1024xf32, 1024xf32) add_68 = paddle._C_ops.add(matmul_45, parameter_416) del matmul_45, parameter_416 - # pd_op.dropout: (1x10x1024xf32, 1x10x1024xui8) <- (1x10x1024xf32, None, 1xf32) + # pd_op.dropout: (1x20x1024xf32, 1x20x1024xui8) <- (1x20x1024xf32, None, 1xf32) dropout_60, dropout_61 = (lambda x, f: f(x))( paddle._C_ops.dropout( add_68, None, full_8, True, "upscale_in_train", 0, False @@ -2133,11 +2133,11 @@ def forward( ) del add_68 - # pd_op.add: (1x10x1024xf32) <- (1x10x1024xf32, 1x10x1024xf32) + # pd_op.add: (1x20x1024xf32) <- (1x20x1024xf32, 1x20x1024xf32) add_69 = paddle._C_ops.add(dropout_60, layer_norm_42) del dropout_60, layer_norm_42 - # pd_op.layer_norm: (1x10x1024xf32, 1x10xf32, 1x10xf32) <- (1x10x1024xf32, 1024xf32, 1024xf32) + # pd_op.layer_norm: (1x20x1024xf32, 1x20xf32, 1x20xf32) <- (1x20x1024xf32, 1024xf32, 1024xf32) layer_norm_45, layer_norm_46, layer_norm_47 = (lambda x, f: f(x))( paddle._C_ops.layer_norm( add_69, parameter_415, parameter_414, float("1e-05"), 2 @@ -2146,19 +2146,19 @@ def forward( ) del add_69, parameter_414, parameter_415 - # pd_op.matmul: (1x10x4096xf32) <- (1x10x1024xf32, 1024x4096xf32) + # pd_op.matmul: (1x20x4096xf32) <- (1x20x1024xf32, 1024x4096xf32) matmul_46 = paddle._C_ops.matmul(layer_norm_45, parameter_413, False, False) del parameter_413 - # pd_op.add: (1x10x4096xf32) <- (1x10x4096xf32, 4096xf32) + # pd_op.add: (1x20x4096xf32) <- (1x20x4096xf32, 4096xf32) add_70 = paddle._C_ops.add(matmul_46, parameter_412) del matmul_46, parameter_412 - # pd_op.gelu: (1x10x4096xf32) <- (1x10x4096xf32) + # pd_op.gelu: (1x20x4096xf32) <- (1x20x4096xf32) gelu_7 = paddle._C_ops.gelu(add_70, False) del add_70 - # pd_op.dropout: (1x10x4096xf32, 1x10x4096xui8) <- (1x10x4096xf32, None, 1xf32) + # pd_op.dropout: (1x20x4096xf32, 1x20x4096xui8) <- (1x20x4096xf32, None, 1xf32) dropout_62, dropout_63 = (lambda x, f: f(x))( paddle._C_ops.dropout( gelu_7, None, full_8, True, "upscale_in_train", 0, False @@ -2167,15 +2167,15 @@ def forward( ) del gelu_7 - # pd_op.matmul: (1x10x1024xf32) <- (1x10x4096xf32, 4096x1024xf32) + # pd_op.matmul: (1x20x1024xf32) <- (1x20x4096xf32, 4096x1024xf32) matmul_47 = paddle._C_ops.matmul(dropout_62, parameter_411, False, False) del dropout_62, parameter_411 - # pd_op.add: (1x10x1024xf32) <- (1x10x1024xf32, 1024xf32) + # pd_op.add: (1x20x1024xf32) <- (1x20x1024xf32, 1024xf32) add_71 = paddle._C_ops.add(matmul_47, parameter_410) del matmul_47, parameter_410 - # pd_op.dropout: (1x10x1024xf32, 1x10x1024xui8) <- (1x10x1024xf32, None, 1xf32) + # pd_op.dropout: (1x20x1024xf32, 1x20x1024xui8) <- (1x20x1024xf32, None, 1xf32) dropout_64, dropout_65 = (lambda x, f: f(x))( paddle._C_ops.dropout( add_71, None, full_8, True, "upscale_in_train", 0, False @@ -2184,11 +2184,11 @@ def forward( ) del add_71 - # pd_op.add: (1x10x1024xf32) <- (1x10x1024xf32, 1x10x1024xf32) + # pd_op.add: (1x20x1024xf32) <- (1x20x1024xf32, 1x20x1024xf32) add_72 = paddle._C_ops.add(dropout_64, layer_norm_45) del dropout_64, layer_norm_45 - # pd_op.layer_norm: (1x10x1024xf32, 1x10xf32, 1x10xf32) <- (1x10x1024xf32, 1024xf32, 1024xf32) + # pd_op.layer_norm: (1x20x1024xf32, 1x20xf32, 1x20xf32) <- (1x20x1024xf32, 1024xf32, 1024xf32) layer_norm_48, layer_norm_49, layer_norm_50 = (lambda x, f: f(x))( paddle._C_ops.layer_norm( add_72, parameter_409, parameter_408, float("1e-05"), 2 @@ -2197,87 +2197,87 @@ def forward( ) del add_72, parameter_408, parameter_409 - # pd_op.matmul: (1x10x1024xf32) <- (1x10x1024xf32, 1024x1024xf32) + # pd_op.matmul: (1x20x1024xf32) <- (1x20x1024xf32, 1024x1024xf32) matmul_48 = paddle._C_ops.matmul(layer_norm_48, parameter_403, False, False) del parameter_403 - # pd_op.add: (1x10x1024xf32) <- (1x10x1024xf32, 1024xf32) + # pd_op.add: (1x20x1024xf32) <- (1x20x1024xf32, 1024xf32) add_73 = paddle._C_ops.add(matmul_48, parameter_402) del matmul_48, parameter_402 - # pd_op.scale: (1x10x1024xf32) <- (1x10x1024xf32, 1xf32) + # pd_op.scale: (1x20x1024xf32) <- (1x20x1024xf32, 1xf32) scale_11 = paddle._C_ops.scale(add_73, full_9, float("0"), True) del add_73 - # pd_op.matmul: (1x10x1024xf32) <- (1x10x1024xf32, 1024x1024xf32) + # pd_op.matmul: (1x20x1024xf32) <- (1x20x1024xf32, 1024x1024xf32) matmul_49 = paddle._C_ops.matmul(layer_norm_48, parameter_407, False, False) del parameter_407 - # pd_op.add: (1x10x1024xf32) <- (1x10x1024xf32, 1024xf32) + # pd_op.add: (1x20x1024xf32) <- (1x20x1024xf32, 1024xf32) add_74 = paddle._C_ops.add(matmul_49, parameter_406) del matmul_49, parameter_406 - # pd_op.reshape: (1x10x16x64xf32) <- (1x10x1024xf32, 4xi64) + # pd_op.reshape: (1x20x16x64xf32) <- (1x20x1024xf32, 4xi64) reshape_88 = paddle._C_ops.reshape(add_74, full_int_array_2) del add_74 - # pd_op.transpose: (1x16x10x64xf32) <- (1x10x16x64xf32) + # pd_op.transpose: (1x16x20x64xf32) <- (1x20x16x64xf32) transpose_40 = paddle._C_ops.transpose(reshape_88, [0, 2, 1, 3]) del reshape_88 - # pd_op.matmul: (1x10x1024xf32) <- (1x10x1024xf32, 1024x1024xf32) + # pd_op.matmul: (1x20x1024xf32) <- (1x20x1024xf32, 1024x1024xf32) matmul_50 = paddle._C_ops.matmul(layer_norm_48, parameter_405, False, False) del parameter_405 - # pd_op.add: (1x10x1024xf32) <- (1x10x1024xf32, 1024xf32) + # pd_op.add: (1x20x1024xf32) <- (1x20x1024xf32, 1024xf32) add_75 = paddle._C_ops.add(matmul_50, parameter_404) del matmul_50, parameter_404 - # pd_op.reshape: (1x10x16x64xf32) <- (1x10x1024xf32, 4xi64) + # pd_op.reshape: (1x20x16x64xf32) <- (1x20x1024xf32, 4xi64) reshape_89 = paddle._C_ops.reshape(add_75, full_int_array_2) del add_75 - # pd_op.transpose: (1x16x10x64xf32) <- (1x10x16x64xf32) + # pd_op.transpose: (1x16x20x64xf32) <- (1x20x16x64xf32) transpose_41 = paddle._C_ops.transpose(reshape_89, [0, 2, 1, 3]) del reshape_89 - # pd_op.reshape: (1x10x16x64xf32) <- (1x10x1024xf32, 4xi64) + # pd_op.reshape: (1x20x16x64xf32) <- (1x20x1024xf32, 4xi64) reshape_90 = paddle._C_ops.reshape(scale_11, full_int_array_3) del scale_11 - # pd_op.transpose: (1x16x10x64xf32) <- (1x10x16x64xf32) + # pd_op.transpose: (1x16x20x64xf32) <- (1x20x16x64xf32) transpose_42 = paddle._C_ops.transpose(reshape_90, [0, 2, 1, 3]) del reshape_90 - # pd_op.reshape: (16x10x64xf32) <- (1x16x10x64xf32, 3xi64) + # pd_op.reshape: (16x20x64xf32) <- (1x16x20x64xf32, 3xi64) reshape_91 = paddle._C_ops.reshape(transpose_42, full_int_array_4) del transpose_42 - # pd_op.reshape: (16x10x64xf32) <- (1x16x10x64xf32, 3xi64) + # pd_op.reshape: (16x20x64xf32) <- (1x16x20x64xf32, 3xi64) reshape_92 = paddle._C_ops.reshape(transpose_40, full_int_array_4) del transpose_40 - # pd_op.reshape: (16x10x64xf32) <- (1x16x10x64xf32, 3xi64) + # pd_op.reshape: (16x20x64xf32) <- (1x16x20x64xf32, 3xi64) reshape_93 = paddle._C_ops.reshape(transpose_41, full_int_array_4) del transpose_41 - # pd_op.transpose: (16x64x10xf32) <- (16x10x64xf32) + # pd_op.transpose: (16x64x20xf32) <- (16x20x64xf32) transpose_43 = paddle._C_ops.transpose(reshape_92, [0, 2, 1]) del reshape_92 - # pd_op.bmm: (16x10x10xf32) <- (16x10x64xf32, 16x64x10xf32) + # pd_op.bmm: (16x20x20xf32) <- (16x20x64xf32, 16x64x20xf32) bmm_16 = paddle._C_ops.bmm(reshape_91, transpose_43) del reshape_91, transpose_43 - # pd_op.add: (16x10x10xf32) <- (16x10x10xf32, 16x1x10xf32) + # pd_op.add: (16x20x20xf32) <- (16x20x20xf32, 16x1x20xf32) add_76 = paddle._C_ops.add(bmm_16, cast_2) del bmm_16 - # pd_op.softmax: (16x10x10xf32) <- (16x10x10xf32) + # pd_op.softmax: (16x20x20xf32) <- (16x20x20xf32) softmax_8 = paddle._C_ops.softmax(add_76, -1) del add_76 - # pd_op.dropout: (16x10x10xf32, 16x10x10xui8) <- (16x10x10xf32, None, 1xf32) + # pd_op.dropout: (16x20x20xf32, 16x20x20xui8) <- (16x20x20xf32, None, 1xf32) dropout_66, dropout_67 = (lambda x, f: f(x))( paddle._C_ops.dropout( softmax_8, None, full_8, True, "upscale_in_train", 0, False @@ -2286,31 +2286,31 @@ def forward( ) del softmax_8 - # pd_op.bmm: (16x10x64xf32) <- (16x10x10xf32, 16x10x64xf32) + # pd_op.bmm: (16x20x64xf32) <- (16x20x20xf32, 16x20x64xf32) bmm_17 = paddle._C_ops.bmm(dropout_66, reshape_93) del dropout_66, reshape_93 - # pd_op.reshape: (1x16x10x64xf32) <- (16x10x64xf32, 4xi64) + # pd_op.reshape: (1x16x20x64xf32) <- (16x20x64xf32, 4xi64) reshape_94 = paddle._C_ops.reshape(bmm_17, full_int_array_5) del bmm_17 - # pd_op.transpose: (1x10x16x64xf32) <- (1x16x10x64xf32) + # pd_op.transpose: (1x20x16x64xf32) <- (1x16x20x64xf32) transpose_44 = paddle._C_ops.transpose(reshape_94, [0, 2, 1, 3]) del reshape_94 - # pd_op.reshape: (1x10x1024xf32) <- (1x10x16x64xf32, 3xi64) + # pd_op.reshape: (1x20x1024xf32) <- (1x20x16x64xf32, 3xi64) reshape_95 = paddle._C_ops.reshape(transpose_44, full_int_array_6) del transpose_44 - # pd_op.matmul: (1x10x1024xf32) <- (1x10x1024xf32, 1024x1024xf32) + # pd_op.matmul: (1x20x1024xf32) <- (1x20x1024xf32, 1024x1024xf32) matmul_51 = paddle._C_ops.matmul(reshape_95, parameter_401, False, False) del parameter_401, reshape_95 - # pd_op.add: (1x10x1024xf32) <- (1x10x1024xf32, 1024xf32) + # pd_op.add: (1x20x1024xf32) <- (1x20x1024xf32, 1024xf32) add_77 = paddle._C_ops.add(matmul_51, parameter_400) del matmul_51, parameter_400 - # pd_op.dropout: (1x10x1024xf32, 1x10x1024xui8) <- (1x10x1024xf32, None, 1xf32) + # pd_op.dropout: (1x20x1024xf32, 1x20x1024xui8) <- (1x20x1024xf32, None, 1xf32) dropout_68, dropout_69 = (lambda x, f: f(x))( paddle._C_ops.dropout( add_77, None, full_8, True, "upscale_in_train", 0, False @@ -2319,11 +2319,11 @@ def forward( ) del add_77 - # pd_op.add: (1x10x1024xf32) <- (1x10x1024xf32, 1x10x1024xf32) + # pd_op.add: (1x20x1024xf32) <- (1x20x1024xf32, 1x20x1024xf32) add_78 = paddle._C_ops.add(dropout_68, layer_norm_48) del dropout_68, layer_norm_48 - # pd_op.layer_norm: (1x10x1024xf32, 1x10xf32, 1x10xf32) <- (1x10x1024xf32, 1024xf32, 1024xf32) + # pd_op.layer_norm: (1x20x1024xf32, 1x20xf32, 1x20xf32) <- (1x20x1024xf32, 1024xf32, 1024xf32) layer_norm_51, layer_norm_52, layer_norm_53 = (lambda x, f: f(x))( paddle._C_ops.layer_norm( add_78, parameter_399, parameter_398, float("1e-05"), 2 @@ -2332,19 +2332,19 @@ def forward( ) del add_78, parameter_398, parameter_399 - # pd_op.matmul: (1x10x4096xf32) <- (1x10x1024xf32, 1024x4096xf32) + # pd_op.matmul: (1x20x4096xf32) <- (1x20x1024xf32, 1024x4096xf32) matmul_52 = paddle._C_ops.matmul(layer_norm_51, parameter_397, False, False) del parameter_397 - # pd_op.add: (1x10x4096xf32) <- (1x10x4096xf32, 4096xf32) + # pd_op.add: (1x20x4096xf32) <- (1x20x4096xf32, 4096xf32) add_79 = paddle._C_ops.add(matmul_52, parameter_396) del matmul_52, parameter_396 - # pd_op.gelu: (1x10x4096xf32) <- (1x10x4096xf32) + # pd_op.gelu: (1x20x4096xf32) <- (1x20x4096xf32) gelu_8 = paddle._C_ops.gelu(add_79, False) del add_79 - # pd_op.dropout: (1x10x4096xf32, 1x10x4096xui8) <- (1x10x4096xf32, None, 1xf32) + # pd_op.dropout: (1x20x4096xf32, 1x20x4096xui8) <- (1x20x4096xf32, None, 1xf32) dropout_70, dropout_71 = (lambda x, f: f(x))( paddle._C_ops.dropout( gelu_8, None, full_8, True, "upscale_in_train", 0, False @@ -2353,15 +2353,15 @@ def forward( ) del gelu_8 - # pd_op.matmul: (1x10x1024xf32) <- (1x10x4096xf32, 4096x1024xf32) + # pd_op.matmul: (1x20x1024xf32) <- (1x20x4096xf32, 4096x1024xf32) matmul_53 = paddle._C_ops.matmul(dropout_70, parameter_395, False, False) del dropout_70, parameter_395 - # pd_op.add: (1x10x1024xf32) <- (1x10x1024xf32, 1024xf32) + # pd_op.add: (1x20x1024xf32) <- (1x20x1024xf32, 1024xf32) add_80 = paddle._C_ops.add(matmul_53, parameter_394) del matmul_53, parameter_394 - # pd_op.dropout: (1x10x1024xf32, 1x10x1024xui8) <- (1x10x1024xf32, None, 1xf32) + # pd_op.dropout: (1x20x1024xf32, 1x20x1024xui8) <- (1x20x1024xf32, None, 1xf32) dropout_72, dropout_73 = (lambda x, f: f(x))( paddle._C_ops.dropout( add_80, None, full_8, True, "upscale_in_train", 0, False @@ -2370,11 +2370,11 @@ def forward( ) del add_80 - # pd_op.add: (1x10x1024xf32) <- (1x10x1024xf32, 1x10x1024xf32) + # pd_op.add: (1x20x1024xf32) <- (1x20x1024xf32, 1x20x1024xf32) add_81 = paddle._C_ops.add(dropout_72, layer_norm_51) del dropout_72, layer_norm_51 - # pd_op.layer_norm: (1x10x1024xf32, 1x10xf32, 1x10xf32) <- (1x10x1024xf32, 1024xf32, 1024xf32) + # pd_op.layer_norm: (1x20x1024xf32, 1x20xf32, 1x20xf32) <- (1x20x1024xf32, 1024xf32, 1024xf32) layer_norm_54, layer_norm_55, layer_norm_56 = (lambda x, f: f(x))( paddle._C_ops.layer_norm( add_81, parameter_393, parameter_392, float("1e-05"), 2 @@ -2383,87 +2383,87 @@ def forward( ) del add_81, parameter_392, parameter_393 - # pd_op.matmul: (1x10x1024xf32) <- (1x10x1024xf32, 1024x1024xf32) + # pd_op.matmul: (1x20x1024xf32) <- (1x20x1024xf32, 1024x1024xf32) matmul_54 = paddle._C_ops.matmul(layer_norm_54, parameter_387, False, False) del parameter_387 - # pd_op.add: (1x10x1024xf32) <- (1x10x1024xf32, 1024xf32) + # pd_op.add: (1x20x1024xf32) <- (1x20x1024xf32, 1024xf32) add_82 = paddle._C_ops.add(matmul_54, parameter_386) del matmul_54, parameter_386 - # pd_op.scale: (1x10x1024xf32) <- (1x10x1024xf32, 1xf32) + # pd_op.scale: (1x20x1024xf32) <- (1x20x1024xf32, 1xf32) scale_12 = paddle._C_ops.scale(add_82, full_9, float("0"), True) del add_82 - # pd_op.matmul: (1x10x1024xf32) <- (1x10x1024xf32, 1024x1024xf32) + # pd_op.matmul: (1x20x1024xf32) <- (1x20x1024xf32, 1024x1024xf32) matmul_55 = paddle._C_ops.matmul(layer_norm_54, parameter_391, False, False) del parameter_391 - # pd_op.add: (1x10x1024xf32) <- (1x10x1024xf32, 1024xf32) + # pd_op.add: (1x20x1024xf32) <- (1x20x1024xf32, 1024xf32) add_83 = paddle._C_ops.add(matmul_55, parameter_390) del matmul_55, parameter_390 - # pd_op.reshape: (1x10x16x64xf32) <- (1x10x1024xf32, 4xi64) + # pd_op.reshape: (1x20x16x64xf32) <- (1x20x1024xf32, 4xi64) reshape_96 = paddle._C_ops.reshape(add_83, full_int_array_2) del add_83 - # pd_op.transpose: (1x16x10x64xf32) <- (1x10x16x64xf32) + # pd_op.transpose: (1x16x20x64xf32) <- (1x20x16x64xf32) transpose_45 = paddle._C_ops.transpose(reshape_96, [0, 2, 1, 3]) del reshape_96 - # pd_op.matmul: (1x10x1024xf32) <- (1x10x1024xf32, 1024x1024xf32) + # pd_op.matmul: (1x20x1024xf32) <- (1x20x1024xf32, 1024x1024xf32) matmul_56 = paddle._C_ops.matmul(layer_norm_54, parameter_389, False, False) del parameter_389 - # pd_op.add: (1x10x1024xf32) <- (1x10x1024xf32, 1024xf32) + # pd_op.add: (1x20x1024xf32) <- (1x20x1024xf32, 1024xf32) add_84 = paddle._C_ops.add(matmul_56, parameter_388) del matmul_56, parameter_388 - # pd_op.reshape: (1x10x16x64xf32) <- (1x10x1024xf32, 4xi64) + # pd_op.reshape: (1x20x16x64xf32) <- (1x20x1024xf32, 4xi64) reshape_97 = paddle._C_ops.reshape(add_84, full_int_array_2) del add_84 - # pd_op.transpose: (1x16x10x64xf32) <- (1x10x16x64xf32) + # pd_op.transpose: (1x16x20x64xf32) <- (1x20x16x64xf32) transpose_46 = paddle._C_ops.transpose(reshape_97, [0, 2, 1, 3]) del reshape_97 - # pd_op.reshape: (1x10x16x64xf32) <- (1x10x1024xf32, 4xi64) + # pd_op.reshape: (1x20x16x64xf32) <- (1x20x1024xf32, 4xi64) reshape_98 = paddle._C_ops.reshape(scale_12, full_int_array_3) del scale_12 - # pd_op.transpose: (1x16x10x64xf32) <- (1x10x16x64xf32) + # pd_op.transpose: (1x16x20x64xf32) <- (1x20x16x64xf32) transpose_47 = paddle._C_ops.transpose(reshape_98, [0, 2, 1, 3]) del reshape_98 - # pd_op.reshape: (16x10x64xf32) <- (1x16x10x64xf32, 3xi64) + # pd_op.reshape: (16x20x64xf32) <- (1x16x20x64xf32, 3xi64) reshape_99 = paddle._C_ops.reshape(transpose_47, full_int_array_4) del transpose_47 - # pd_op.reshape: (16x10x64xf32) <- (1x16x10x64xf32, 3xi64) + # pd_op.reshape: (16x20x64xf32) <- (1x16x20x64xf32, 3xi64) reshape_100 = paddle._C_ops.reshape(transpose_45, full_int_array_4) del transpose_45 - # pd_op.reshape: (16x10x64xf32) <- (1x16x10x64xf32, 3xi64) + # pd_op.reshape: (16x20x64xf32) <- (1x16x20x64xf32, 3xi64) reshape_101 = paddle._C_ops.reshape(transpose_46, full_int_array_4) del transpose_46 - # pd_op.transpose: (16x64x10xf32) <- (16x10x64xf32) + # pd_op.transpose: (16x64x20xf32) <- (16x20x64xf32) transpose_48 = paddle._C_ops.transpose(reshape_100, [0, 2, 1]) del reshape_100 - # pd_op.bmm: (16x10x10xf32) <- (16x10x64xf32, 16x64x10xf32) + # pd_op.bmm: (16x20x20xf32) <- (16x20x64xf32, 16x64x20xf32) bmm_18 = paddle._C_ops.bmm(reshape_99, transpose_48) del reshape_99, transpose_48 - # pd_op.add: (16x10x10xf32) <- (16x10x10xf32, 16x1x10xf32) + # pd_op.add: (16x20x20xf32) <- (16x20x20xf32, 16x1x20xf32) add_85 = paddle._C_ops.add(bmm_18, cast_2) del bmm_18 - # pd_op.softmax: (16x10x10xf32) <- (16x10x10xf32) + # pd_op.softmax: (16x20x20xf32) <- (16x20x20xf32) softmax_9 = paddle._C_ops.softmax(add_85, -1) del add_85 - # pd_op.dropout: (16x10x10xf32, 16x10x10xui8) <- (16x10x10xf32, None, 1xf32) + # pd_op.dropout: (16x20x20xf32, 16x20x20xui8) <- (16x20x20xf32, None, 1xf32) dropout_74, dropout_75 = (lambda x, f: f(x))( paddle._C_ops.dropout( softmax_9, None, full_8, True, "upscale_in_train", 0, False @@ -2472,31 +2472,31 @@ def forward( ) del softmax_9 - # pd_op.bmm: (16x10x64xf32) <- (16x10x10xf32, 16x10x64xf32) + # pd_op.bmm: (16x20x64xf32) <- (16x20x20xf32, 16x20x64xf32) bmm_19 = paddle._C_ops.bmm(dropout_74, reshape_101) del dropout_74, reshape_101 - # pd_op.reshape: (1x16x10x64xf32) <- (16x10x64xf32, 4xi64) + # pd_op.reshape: (1x16x20x64xf32) <- (16x20x64xf32, 4xi64) reshape_102 = paddle._C_ops.reshape(bmm_19, full_int_array_5) del bmm_19 - # pd_op.transpose: (1x10x16x64xf32) <- (1x16x10x64xf32) + # pd_op.transpose: (1x20x16x64xf32) <- (1x16x20x64xf32) transpose_49 = paddle._C_ops.transpose(reshape_102, [0, 2, 1, 3]) del reshape_102 - # pd_op.reshape: (1x10x1024xf32) <- (1x10x16x64xf32, 3xi64) + # pd_op.reshape: (1x20x1024xf32) <- (1x20x16x64xf32, 3xi64) reshape_103 = paddle._C_ops.reshape(transpose_49, full_int_array_6) del transpose_49 - # pd_op.matmul: (1x10x1024xf32) <- (1x10x1024xf32, 1024x1024xf32) + # pd_op.matmul: (1x20x1024xf32) <- (1x20x1024xf32, 1024x1024xf32) matmul_57 = paddle._C_ops.matmul(reshape_103, parameter_385, False, False) del parameter_385, reshape_103 - # pd_op.add: (1x10x1024xf32) <- (1x10x1024xf32, 1024xf32) + # pd_op.add: (1x20x1024xf32) <- (1x20x1024xf32, 1024xf32) add_86 = paddle._C_ops.add(matmul_57, parameter_384) del matmul_57, parameter_384 - # pd_op.dropout: (1x10x1024xf32, 1x10x1024xui8) <- (1x10x1024xf32, None, 1xf32) + # pd_op.dropout: (1x20x1024xf32, 1x20x1024xui8) <- (1x20x1024xf32, None, 1xf32) dropout_76, dropout_77 = (lambda x, f: f(x))( paddle._C_ops.dropout( add_86, None, full_8, True, "upscale_in_train", 0, False @@ -2505,11 +2505,11 @@ def forward( ) del add_86 - # pd_op.add: (1x10x1024xf32) <- (1x10x1024xf32, 1x10x1024xf32) + # pd_op.add: (1x20x1024xf32) <- (1x20x1024xf32, 1x20x1024xf32) add_87 = paddle._C_ops.add(dropout_76, layer_norm_54) del dropout_76, layer_norm_54 - # pd_op.layer_norm: (1x10x1024xf32, 1x10xf32, 1x10xf32) <- (1x10x1024xf32, 1024xf32, 1024xf32) + # pd_op.layer_norm: (1x20x1024xf32, 1x20xf32, 1x20xf32) <- (1x20x1024xf32, 1024xf32, 1024xf32) layer_norm_57, layer_norm_58, layer_norm_59 = (lambda x, f: f(x))( paddle._C_ops.layer_norm( add_87, parameter_383, parameter_382, float("1e-05"), 2 @@ -2518,19 +2518,19 @@ def forward( ) del add_87, parameter_382, parameter_383 - # pd_op.matmul: (1x10x4096xf32) <- (1x10x1024xf32, 1024x4096xf32) + # pd_op.matmul: (1x20x4096xf32) <- (1x20x1024xf32, 1024x4096xf32) matmul_58 = paddle._C_ops.matmul(layer_norm_57, parameter_381, False, False) del parameter_381 - # pd_op.add: (1x10x4096xf32) <- (1x10x4096xf32, 4096xf32) + # pd_op.add: (1x20x4096xf32) <- (1x20x4096xf32, 4096xf32) add_88 = paddle._C_ops.add(matmul_58, parameter_380) del matmul_58, parameter_380 - # pd_op.gelu: (1x10x4096xf32) <- (1x10x4096xf32) + # pd_op.gelu: (1x20x4096xf32) <- (1x20x4096xf32) gelu_9 = paddle._C_ops.gelu(add_88, False) del add_88 - # pd_op.dropout: (1x10x4096xf32, 1x10x4096xui8) <- (1x10x4096xf32, None, 1xf32) + # pd_op.dropout: (1x20x4096xf32, 1x20x4096xui8) <- (1x20x4096xf32, None, 1xf32) dropout_78, dropout_79 = (lambda x, f: f(x))( paddle._C_ops.dropout( gelu_9, None, full_8, True, "upscale_in_train", 0, False @@ -2539,15 +2539,15 @@ def forward( ) del gelu_9 - # pd_op.matmul: (1x10x1024xf32) <- (1x10x4096xf32, 4096x1024xf32) + # pd_op.matmul: (1x20x1024xf32) <- (1x20x4096xf32, 4096x1024xf32) matmul_59 = paddle._C_ops.matmul(dropout_78, parameter_379, False, False) del dropout_78, parameter_379 - # pd_op.add: (1x10x1024xf32) <- (1x10x1024xf32, 1024xf32) + # pd_op.add: (1x20x1024xf32) <- (1x20x1024xf32, 1024xf32) add_89 = paddle._C_ops.add(matmul_59, parameter_378) del matmul_59, parameter_378 - # pd_op.dropout: (1x10x1024xf32, 1x10x1024xui8) <- (1x10x1024xf32, None, 1xf32) + # pd_op.dropout: (1x20x1024xf32, 1x20x1024xui8) <- (1x20x1024xf32, None, 1xf32) dropout_80, dropout_81 = (lambda x, f: f(x))( paddle._C_ops.dropout( add_89, None, full_8, True, "upscale_in_train", 0, False @@ -2556,11 +2556,11 @@ def forward( ) del add_89 - # pd_op.add: (1x10x1024xf32) <- (1x10x1024xf32, 1x10x1024xf32) + # pd_op.add: (1x20x1024xf32) <- (1x20x1024xf32, 1x20x1024xf32) add_90 = paddle._C_ops.add(dropout_80, layer_norm_57) del dropout_80, layer_norm_57 - # pd_op.layer_norm: (1x10x1024xf32, 1x10xf32, 1x10xf32) <- (1x10x1024xf32, 1024xf32, 1024xf32) + # pd_op.layer_norm: (1x20x1024xf32, 1x20xf32, 1x20xf32) <- (1x20x1024xf32, 1024xf32, 1024xf32) layer_norm_60, layer_norm_61, layer_norm_62 = (lambda x, f: f(x))( paddle._C_ops.layer_norm( add_90, parameter_377, parameter_376, float("1e-05"), 2 @@ -2569,87 +2569,87 @@ def forward( ) del add_90, parameter_376, parameter_377 - # pd_op.matmul: (1x10x1024xf32) <- (1x10x1024xf32, 1024x1024xf32) + # pd_op.matmul: (1x20x1024xf32) <- (1x20x1024xf32, 1024x1024xf32) matmul_60 = paddle._C_ops.matmul(layer_norm_60, parameter_371, False, False) del parameter_371 - # pd_op.add: (1x10x1024xf32) <- (1x10x1024xf32, 1024xf32) + # pd_op.add: (1x20x1024xf32) <- (1x20x1024xf32, 1024xf32) add_91 = paddle._C_ops.add(matmul_60, parameter_370) del matmul_60, parameter_370 - # pd_op.scale: (1x10x1024xf32) <- (1x10x1024xf32, 1xf32) + # pd_op.scale: (1x20x1024xf32) <- (1x20x1024xf32, 1xf32) scale_13 = paddle._C_ops.scale(add_91, full_9, float("0"), True) del add_91 - # pd_op.matmul: (1x10x1024xf32) <- (1x10x1024xf32, 1024x1024xf32) + # pd_op.matmul: (1x20x1024xf32) <- (1x20x1024xf32, 1024x1024xf32) matmul_61 = paddle._C_ops.matmul(layer_norm_60, parameter_375, False, False) del parameter_375 - # pd_op.add: (1x10x1024xf32) <- (1x10x1024xf32, 1024xf32) + # pd_op.add: (1x20x1024xf32) <- (1x20x1024xf32, 1024xf32) add_92 = paddle._C_ops.add(matmul_61, parameter_374) del matmul_61, parameter_374 - # pd_op.reshape: (1x10x16x64xf32) <- (1x10x1024xf32, 4xi64) + # pd_op.reshape: (1x20x16x64xf32) <- (1x20x1024xf32, 4xi64) reshape_104 = paddle._C_ops.reshape(add_92, full_int_array_2) del add_92 - # pd_op.transpose: (1x16x10x64xf32) <- (1x10x16x64xf32) + # pd_op.transpose: (1x16x20x64xf32) <- (1x20x16x64xf32) transpose_50 = paddle._C_ops.transpose(reshape_104, [0, 2, 1, 3]) del reshape_104 - # pd_op.matmul: (1x10x1024xf32) <- (1x10x1024xf32, 1024x1024xf32) + # pd_op.matmul: (1x20x1024xf32) <- (1x20x1024xf32, 1024x1024xf32) matmul_62 = paddle._C_ops.matmul(layer_norm_60, parameter_373, False, False) del parameter_373 - # pd_op.add: (1x10x1024xf32) <- (1x10x1024xf32, 1024xf32) + # pd_op.add: (1x20x1024xf32) <- (1x20x1024xf32, 1024xf32) add_93 = paddle._C_ops.add(matmul_62, parameter_372) del matmul_62, parameter_372 - # pd_op.reshape: (1x10x16x64xf32) <- (1x10x1024xf32, 4xi64) + # pd_op.reshape: (1x20x16x64xf32) <- (1x20x1024xf32, 4xi64) reshape_105 = paddle._C_ops.reshape(add_93, full_int_array_2) del add_93 - # pd_op.transpose: (1x16x10x64xf32) <- (1x10x16x64xf32) + # pd_op.transpose: (1x16x20x64xf32) <- (1x20x16x64xf32) transpose_51 = paddle._C_ops.transpose(reshape_105, [0, 2, 1, 3]) del reshape_105 - # pd_op.reshape: (1x10x16x64xf32) <- (1x10x1024xf32, 4xi64) + # pd_op.reshape: (1x20x16x64xf32) <- (1x20x1024xf32, 4xi64) reshape_106 = paddle._C_ops.reshape(scale_13, full_int_array_3) del scale_13 - # pd_op.transpose: (1x16x10x64xf32) <- (1x10x16x64xf32) + # pd_op.transpose: (1x16x20x64xf32) <- (1x20x16x64xf32) transpose_52 = paddle._C_ops.transpose(reshape_106, [0, 2, 1, 3]) del reshape_106 - # pd_op.reshape: (16x10x64xf32) <- (1x16x10x64xf32, 3xi64) + # pd_op.reshape: (16x20x64xf32) <- (1x16x20x64xf32, 3xi64) reshape_107 = paddle._C_ops.reshape(transpose_52, full_int_array_4) del transpose_52 - # pd_op.reshape: (16x10x64xf32) <- (1x16x10x64xf32, 3xi64) + # pd_op.reshape: (16x20x64xf32) <- (1x16x20x64xf32, 3xi64) reshape_108 = paddle._C_ops.reshape(transpose_50, full_int_array_4) del transpose_50 - # pd_op.reshape: (16x10x64xf32) <- (1x16x10x64xf32, 3xi64) + # pd_op.reshape: (16x20x64xf32) <- (1x16x20x64xf32, 3xi64) reshape_109 = paddle._C_ops.reshape(transpose_51, full_int_array_4) del transpose_51 - # pd_op.transpose: (16x64x10xf32) <- (16x10x64xf32) + # pd_op.transpose: (16x64x20xf32) <- (16x20x64xf32) transpose_53 = paddle._C_ops.transpose(reshape_108, [0, 2, 1]) del reshape_108 - # pd_op.bmm: (16x10x10xf32) <- (16x10x64xf32, 16x64x10xf32) + # pd_op.bmm: (16x20x20xf32) <- (16x20x64xf32, 16x64x20xf32) bmm_20 = paddle._C_ops.bmm(reshape_107, transpose_53) del reshape_107, transpose_53 - # pd_op.add: (16x10x10xf32) <- (16x10x10xf32, 16x1x10xf32) + # pd_op.add: (16x20x20xf32) <- (16x20x20xf32, 16x1x20xf32) add_94 = paddle._C_ops.add(bmm_20, cast_2) del bmm_20 - # pd_op.softmax: (16x10x10xf32) <- (16x10x10xf32) + # pd_op.softmax: (16x20x20xf32) <- (16x20x20xf32) softmax_10 = paddle._C_ops.softmax(add_94, -1) del add_94 - # pd_op.dropout: (16x10x10xf32, 16x10x10xui8) <- (16x10x10xf32, None, 1xf32) + # pd_op.dropout: (16x20x20xf32, 16x20x20xui8) <- (16x20x20xf32, None, 1xf32) dropout_82, dropout_83 = (lambda x, f: f(x))( paddle._C_ops.dropout( softmax_10, None, full_8, True, "upscale_in_train", 0, False @@ -2658,31 +2658,31 @@ def forward( ) del softmax_10 - # pd_op.bmm: (16x10x64xf32) <- (16x10x10xf32, 16x10x64xf32) + # pd_op.bmm: (16x20x64xf32) <- (16x20x20xf32, 16x20x64xf32) bmm_21 = paddle._C_ops.bmm(dropout_82, reshape_109) del dropout_82, reshape_109 - # pd_op.reshape: (1x16x10x64xf32) <- (16x10x64xf32, 4xi64) + # pd_op.reshape: (1x16x20x64xf32) <- (16x20x64xf32, 4xi64) reshape_110 = paddle._C_ops.reshape(bmm_21, full_int_array_5) del bmm_21 - # pd_op.transpose: (1x10x16x64xf32) <- (1x16x10x64xf32) + # pd_op.transpose: (1x20x16x64xf32) <- (1x16x20x64xf32) transpose_54 = paddle._C_ops.transpose(reshape_110, [0, 2, 1, 3]) del reshape_110 - # pd_op.reshape: (1x10x1024xf32) <- (1x10x16x64xf32, 3xi64) + # pd_op.reshape: (1x20x1024xf32) <- (1x20x16x64xf32, 3xi64) reshape_111 = paddle._C_ops.reshape(transpose_54, full_int_array_6) del transpose_54 - # pd_op.matmul: (1x10x1024xf32) <- (1x10x1024xf32, 1024x1024xf32) + # pd_op.matmul: (1x20x1024xf32) <- (1x20x1024xf32, 1024x1024xf32) matmul_63 = paddle._C_ops.matmul(reshape_111, parameter_369, False, False) del parameter_369, reshape_111 - # pd_op.add: (1x10x1024xf32) <- (1x10x1024xf32, 1024xf32) + # pd_op.add: (1x20x1024xf32) <- (1x20x1024xf32, 1024xf32) add_95 = paddle._C_ops.add(matmul_63, parameter_368) del matmul_63, parameter_368 - # pd_op.dropout: (1x10x1024xf32, 1x10x1024xui8) <- (1x10x1024xf32, None, 1xf32) + # pd_op.dropout: (1x20x1024xf32, 1x20x1024xui8) <- (1x20x1024xf32, None, 1xf32) dropout_84, dropout_85 = (lambda x, f: f(x))( paddle._C_ops.dropout( add_95, None, full_8, True, "upscale_in_train", 0, False @@ -2691,11 +2691,11 @@ def forward( ) del add_95 - # pd_op.add: (1x10x1024xf32) <- (1x10x1024xf32, 1x10x1024xf32) + # pd_op.add: (1x20x1024xf32) <- (1x20x1024xf32, 1x20x1024xf32) add_96 = paddle._C_ops.add(dropout_84, layer_norm_60) del dropout_84, layer_norm_60 - # pd_op.layer_norm: (1x10x1024xf32, 1x10xf32, 1x10xf32) <- (1x10x1024xf32, 1024xf32, 1024xf32) + # pd_op.layer_norm: (1x20x1024xf32, 1x20xf32, 1x20xf32) <- (1x20x1024xf32, 1024xf32, 1024xf32) layer_norm_63, layer_norm_64, layer_norm_65 = (lambda x, f: f(x))( paddle._C_ops.layer_norm( add_96, parameter_367, parameter_366, float("1e-05"), 2 @@ -2704,19 +2704,19 @@ def forward( ) del add_96, parameter_366, parameter_367 - # pd_op.matmul: (1x10x4096xf32) <- (1x10x1024xf32, 1024x4096xf32) + # pd_op.matmul: (1x20x4096xf32) <- (1x20x1024xf32, 1024x4096xf32) matmul_64 = paddle._C_ops.matmul(layer_norm_63, parameter_365, False, False) del parameter_365 - # pd_op.add: (1x10x4096xf32) <- (1x10x4096xf32, 4096xf32) + # pd_op.add: (1x20x4096xf32) <- (1x20x4096xf32, 4096xf32) add_97 = paddle._C_ops.add(matmul_64, parameter_364) del matmul_64, parameter_364 - # pd_op.gelu: (1x10x4096xf32) <- (1x10x4096xf32) + # pd_op.gelu: (1x20x4096xf32) <- (1x20x4096xf32) gelu_10 = paddle._C_ops.gelu(add_97, False) del add_97 - # pd_op.dropout: (1x10x4096xf32, 1x10x4096xui8) <- (1x10x4096xf32, None, 1xf32) + # pd_op.dropout: (1x20x4096xf32, 1x20x4096xui8) <- (1x20x4096xf32, None, 1xf32) dropout_86, dropout_87 = (lambda x, f: f(x))( paddle._C_ops.dropout( gelu_10, None, full_8, True, "upscale_in_train", 0, False @@ -2725,15 +2725,15 @@ def forward( ) del gelu_10 - # pd_op.matmul: (1x10x1024xf32) <- (1x10x4096xf32, 4096x1024xf32) + # pd_op.matmul: (1x20x1024xf32) <- (1x20x4096xf32, 4096x1024xf32) matmul_65 = paddle._C_ops.matmul(dropout_86, parameter_363, False, False) del dropout_86, parameter_363 - # pd_op.add: (1x10x1024xf32) <- (1x10x1024xf32, 1024xf32) + # pd_op.add: (1x20x1024xf32) <- (1x20x1024xf32, 1024xf32) add_98 = paddle._C_ops.add(matmul_65, parameter_362) del matmul_65, parameter_362 - # pd_op.dropout: (1x10x1024xf32, 1x10x1024xui8) <- (1x10x1024xf32, None, 1xf32) + # pd_op.dropout: (1x20x1024xf32, 1x20x1024xui8) <- (1x20x1024xf32, None, 1xf32) dropout_88, dropout_89 = (lambda x, f: f(x))( paddle._C_ops.dropout( add_98, None, full_8, True, "upscale_in_train", 0, False @@ -2742,11 +2742,11 @@ def forward( ) del add_98 - # pd_op.add: (1x10x1024xf32) <- (1x10x1024xf32, 1x10x1024xf32) + # pd_op.add: (1x20x1024xf32) <- (1x20x1024xf32, 1x20x1024xf32) add_99 = paddle._C_ops.add(dropout_88, layer_norm_63) del dropout_88, layer_norm_63 - # pd_op.layer_norm: (1x10x1024xf32, 1x10xf32, 1x10xf32) <- (1x10x1024xf32, 1024xf32, 1024xf32) + # pd_op.layer_norm: (1x20x1024xf32, 1x20xf32, 1x20xf32) <- (1x20x1024xf32, 1024xf32, 1024xf32) layer_norm_66, layer_norm_67, layer_norm_68 = (lambda x, f: f(x))( paddle._C_ops.layer_norm( add_99, parameter_361, parameter_360, float("1e-05"), 2 @@ -2755,87 +2755,87 @@ def forward( ) del add_99, parameter_360, parameter_361 - # pd_op.matmul: (1x10x1024xf32) <- (1x10x1024xf32, 1024x1024xf32) + # pd_op.matmul: (1x20x1024xf32) <- (1x20x1024xf32, 1024x1024xf32) matmul_66 = paddle._C_ops.matmul(layer_norm_66, parameter_355, False, False) del parameter_355 - # pd_op.add: (1x10x1024xf32) <- (1x10x1024xf32, 1024xf32) + # pd_op.add: (1x20x1024xf32) <- (1x20x1024xf32, 1024xf32) add_100 = paddle._C_ops.add(matmul_66, parameter_354) del matmul_66, parameter_354 - # pd_op.scale: (1x10x1024xf32) <- (1x10x1024xf32, 1xf32) + # pd_op.scale: (1x20x1024xf32) <- (1x20x1024xf32, 1xf32) scale_14 = paddle._C_ops.scale(add_100, full_9, float("0"), True) del add_100 - # pd_op.matmul: (1x10x1024xf32) <- (1x10x1024xf32, 1024x1024xf32) + # pd_op.matmul: (1x20x1024xf32) <- (1x20x1024xf32, 1024x1024xf32) matmul_67 = paddle._C_ops.matmul(layer_norm_66, parameter_359, False, False) del parameter_359 - # pd_op.add: (1x10x1024xf32) <- (1x10x1024xf32, 1024xf32) + # pd_op.add: (1x20x1024xf32) <- (1x20x1024xf32, 1024xf32) add_101 = paddle._C_ops.add(matmul_67, parameter_358) del matmul_67, parameter_358 - # pd_op.reshape: (1x10x16x64xf32) <- (1x10x1024xf32, 4xi64) + # pd_op.reshape: (1x20x16x64xf32) <- (1x20x1024xf32, 4xi64) reshape_112 = paddle._C_ops.reshape(add_101, full_int_array_2) del add_101 - # pd_op.transpose: (1x16x10x64xf32) <- (1x10x16x64xf32) + # pd_op.transpose: (1x16x20x64xf32) <- (1x20x16x64xf32) transpose_55 = paddle._C_ops.transpose(reshape_112, [0, 2, 1, 3]) del reshape_112 - # pd_op.matmul: (1x10x1024xf32) <- (1x10x1024xf32, 1024x1024xf32) + # pd_op.matmul: (1x20x1024xf32) <- (1x20x1024xf32, 1024x1024xf32) matmul_68 = paddle._C_ops.matmul(layer_norm_66, parameter_357, False, False) del parameter_357 - # pd_op.add: (1x10x1024xf32) <- (1x10x1024xf32, 1024xf32) + # pd_op.add: (1x20x1024xf32) <- (1x20x1024xf32, 1024xf32) add_102 = paddle._C_ops.add(matmul_68, parameter_356) del matmul_68, parameter_356 - # pd_op.reshape: (1x10x16x64xf32) <- (1x10x1024xf32, 4xi64) + # pd_op.reshape: (1x20x16x64xf32) <- (1x20x1024xf32, 4xi64) reshape_113 = paddle._C_ops.reshape(add_102, full_int_array_2) del add_102 - # pd_op.transpose: (1x16x10x64xf32) <- (1x10x16x64xf32) + # pd_op.transpose: (1x16x20x64xf32) <- (1x20x16x64xf32) transpose_56 = paddle._C_ops.transpose(reshape_113, [0, 2, 1, 3]) del reshape_113 - # pd_op.reshape: (1x10x16x64xf32) <- (1x10x1024xf32, 4xi64) + # pd_op.reshape: (1x20x16x64xf32) <- (1x20x1024xf32, 4xi64) reshape_114 = paddle._C_ops.reshape(scale_14, full_int_array_3) del full_int_array_3, scale_14 - # pd_op.transpose: (1x16x10x64xf32) <- (1x10x16x64xf32) + # pd_op.transpose: (1x16x20x64xf32) <- (1x20x16x64xf32) transpose_57 = paddle._C_ops.transpose(reshape_114, [0, 2, 1, 3]) del reshape_114 - # pd_op.reshape: (16x10x64xf32) <- (1x16x10x64xf32, 3xi64) + # pd_op.reshape: (16x20x64xf32) <- (1x16x20x64xf32, 3xi64) reshape_115 = paddle._C_ops.reshape(transpose_57, full_int_array_4) del transpose_57 - # pd_op.reshape: (16x10x64xf32) <- (1x16x10x64xf32, 3xi64) + # pd_op.reshape: (16x20x64xf32) <- (1x16x20x64xf32, 3xi64) reshape_116 = paddle._C_ops.reshape(transpose_55, full_int_array_4) del transpose_55 - # pd_op.reshape: (16x10x64xf32) <- (1x16x10x64xf32, 3xi64) + # pd_op.reshape: (16x20x64xf32) <- (1x16x20x64xf32, 3xi64) reshape_117 = paddle._C_ops.reshape(transpose_56, full_int_array_4) del transpose_56 - # pd_op.transpose: (16x64x10xf32) <- (16x10x64xf32) + # pd_op.transpose: (16x64x20xf32) <- (16x20x64xf32) transpose_58 = paddle._C_ops.transpose(reshape_116, [0, 2, 1]) del reshape_116 - # pd_op.bmm: (16x10x10xf32) <- (16x10x64xf32, 16x64x10xf32) + # pd_op.bmm: (16x20x20xf32) <- (16x20x64xf32, 16x64x20xf32) bmm_22 = paddle._C_ops.bmm(reshape_115, transpose_58) del reshape_115, transpose_58 - # pd_op.add: (16x10x10xf32) <- (16x10x10xf32, 16x1x10xf32) + # pd_op.add: (16x20x20xf32) <- (16x20x20xf32, 16x1x20xf32) add_103 = paddle._C_ops.add(bmm_22, cast_2) del bmm_22, cast_2 - # pd_op.softmax: (16x10x10xf32) <- (16x10x10xf32) + # pd_op.softmax: (16x20x20xf32) <- (16x20x20xf32) softmax_11 = paddle._C_ops.softmax(add_103, -1) del add_103 - # pd_op.dropout: (16x10x10xf32, 16x10x10xui8) <- (16x10x10xf32, None, 1xf32) + # pd_op.dropout: (16x20x20xf32, 16x20x20xui8) <- (16x20x20xf32, None, 1xf32) dropout_90, dropout_91 = (lambda x, f: f(x))( paddle._C_ops.dropout( softmax_11, None, full_8, True, "upscale_in_train", 0, False @@ -2844,31 +2844,31 @@ def forward( ) del softmax_11 - # pd_op.bmm: (16x10x64xf32) <- (16x10x10xf32, 16x10x64xf32) + # pd_op.bmm: (16x20x64xf32) <- (16x20x20xf32, 16x20x64xf32) bmm_23 = paddle._C_ops.bmm(dropout_90, reshape_117) del dropout_90, reshape_117 - # pd_op.reshape: (1x16x10x64xf32) <- (16x10x64xf32, 4xi64) + # pd_op.reshape: (1x16x20x64xf32) <- (16x20x64xf32, 4xi64) reshape_118 = paddle._C_ops.reshape(bmm_23, full_int_array_5) del bmm_23, full_int_array_5 - # pd_op.transpose: (1x10x16x64xf32) <- (1x16x10x64xf32) + # pd_op.transpose: (1x20x16x64xf32) <- (1x16x20x64xf32) transpose_59 = paddle._C_ops.transpose(reshape_118, [0, 2, 1, 3]) del reshape_118 - # pd_op.reshape: (1x10x1024xf32) <- (1x10x16x64xf32, 3xi64) + # pd_op.reshape: (1x20x1024xf32) <- (1x20x16x64xf32, 3xi64) reshape_119 = paddle._C_ops.reshape(transpose_59, full_int_array_6) del full_int_array_6, transpose_59 - # pd_op.matmul: (1x10x1024xf32) <- (1x10x1024xf32, 1024x1024xf32) + # pd_op.matmul: (1x20x1024xf32) <- (1x20x1024xf32, 1024x1024xf32) matmul_69 = paddle._C_ops.matmul(reshape_119, parameter_353, False, False) del parameter_353, reshape_119 - # pd_op.add: (1x10x1024xf32) <- (1x10x1024xf32, 1024xf32) + # pd_op.add: (1x20x1024xf32) <- (1x20x1024xf32, 1024xf32) add_104 = paddle._C_ops.add(matmul_69, parameter_352) del matmul_69, parameter_352 - # pd_op.dropout: (1x10x1024xf32, 1x10x1024xui8) <- (1x10x1024xf32, None, 1xf32) + # pd_op.dropout: (1x20x1024xf32, 1x20x1024xui8) <- (1x20x1024xf32, None, 1xf32) dropout_92, dropout_93 = (lambda x, f: f(x))( paddle._C_ops.dropout( add_104, None, full_8, True, "upscale_in_train", 0, False @@ -2877,11 +2877,11 @@ def forward( ) del add_104 - # pd_op.add: (1x10x1024xf32) <- (1x10x1024xf32, 1x10x1024xf32) + # pd_op.add: (1x20x1024xf32) <- (1x20x1024xf32, 1x20x1024xf32) add_105 = paddle._C_ops.add(dropout_92, layer_norm_66) del dropout_92, layer_norm_66 - # pd_op.layer_norm: (1x10x1024xf32, 1x10xf32, 1x10xf32) <- (1x10x1024xf32, 1024xf32, 1024xf32) + # pd_op.layer_norm: (1x20x1024xf32, 1x20xf32, 1x20xf32) <- (1x20x1024xf32, 1024xf32, 1024xf32) layer_norm_69, layer_norm_70, layer_norm_71 = (lambda x, f: f(x))( paddle._C_ops.layer_norm( add_105, parameter_351, parameter_350, float("1e-05"), 2 @@ -2890,19 +2890,19 @@ def forward( ) del add_105, parameter_350, parameter_351 - # pd_op.matmul: (1x10x4096xf32) <- (1x10x1024xf32, 1024x4096xf32) + # pd_op.matmul: (1x20x4096xf32) <- (1x20x1024xf32, 1024x4096xf32) matmul_70 = paddle._C_ops.matmul(layer_norm_69, parameter_349, False, False) del parameter_349 - # pd_op.add: (1x10x4096xf32) <- (1x10x4096xf32, 4096xf32) + # pd_op.add: (1x20x4096xf32) <- (1x20x4096xf32, 4096xf32) add_106 = paddle._C_ops.add(matmul_70, parameter_348) del matmul_70, parameter_348 - # pd_op.gelu: (1x10x4096xf32) <- (1x10x4096xf32) + # pd_op.gelu: (1x20x4096xf32) <- (1x20x4096xf32) gelu_11 = paddle._C_ops.gelu(add_106, False) del add_106 - # pd_op.dropout: (1x10x4096xf32, 1x10x4096xui8) <- (1x10x4096xf32, None, 1xf32) + # pd_op.dropout: (1x20x4096xf32, 1x20x4096xui8) <- (1x20x4096xf32, None, 1xf32) dropout_94, dropout_95 = (lambda x, f: f(x))( paddle._C_ops.dropout( gelu_11, None, full_8, True, "upscale_in_train", 0, False @@ -2911,15 +2911,15 @@ def forward( ) del gelu_11 - # pd_op.matmul: (1x10x1024xf32) <- (1x10x4096xf32, 4096x1024xf32) + # pd_op.matmul: (1x20x1024xf32) <- (1x20x4096xf32, 4096x1024xf32) matmul_71 = paddle._C_ops.matmul(dropout_94, parameter_347, False, False) del dropout_94, parameter_347 - # pd_op.add: (1x10x1024xf32) <- (1x10x1024xf32, 1024xf32) + # pd_op.add: (1x20x1024xf32) <- (1x20x1024xf32, 1024xf32) add_107 = paddle._C_ops.add(matmul_71, parameter_346) del matmul_71, parameter_346 - # pd_op.dropout: (1x10x1024xf32, 1x10x1024xui8) <- (1x10x1024xf32, None, 1xf32) + # pd_op.dropout: (1x20x1024xf32, 1x20x1024xui8) <- (1x20x1024xf32, None, 1xf32) dropout_96, dropout_97 = (lambda x, f: f(x))( paddle._C_ops.dropout( add_107, None, full_8, True, "upscale_in_train", 0, False @@ -2928,11 +2928,11 @@ def forward( ) del add_107 - # pd_op.add: (1x10x1024xf32) <- (1x10x1024xf32, 1x10x1024xf32) + # pd_op.add: (1x20x1024xf32) <- (1x20x1024xf32, 1x20x1024xf32) add_108 = paddle._C_ops.add(dropout_96, layer_norm_69) del dropout_96, layer_norm_69 - # pd_op.layer_norm: (1x10x1024xf32, 1x10xf32, 1x10xf32) <- (1x10x1024xf32, 1024xf32, 1024xf32) + # pd_op.layer_norm: (1x20x1024xf32, 1x20xf32, 1x20xf32) <- (1x20x1024xf32, 1024xf32, 1024xf32) layer_norm_72, layer_norm_73, layer_norm_74 = (lambda x, f: f(x))( paddle._C_ops.layer_norm( add_108, parameter_345, parameter_344, float("1e-05"), 2 @@ -3194,23 +3194,23 @@ def forward( cast_8 = paddle._C_ops.cast(tile_4, paddle.float32) del tile_4 - # pd_op.tile: (16x1x10xi64) <- (1x1x10xi64, 3xi64) + # pd_op.tile: (16x1x20xi64) <- (1x1x20xi64, 3xi64) tile_5 = paddle._C_ops.tile(unsqueeze_0, full_int_array_1) del full_int_array_1, unsqueeze_0 - # pd_op.cast: (16x1x10xf32) <- (16x1x10xi64) + # pd_op.cast: (16x1x20xf32) <- (16x1x20xi64) cast_9 = paddle._C_ops.cast(tile_5, paddle.float32) del tile_5 - # pd_op.scale: (16x1x10xf32) <- (16x1x10xf32, 1xf32) + # pd_op.scale: (16x1x20xf32) <- (16x1x20xf32, 1xf32) scale_21 = paddle._C_ops.scale(cast_9, full_1, float("1"), True) del cast_9, full_1 - # pd_op.scale: (16x1x10xf32) <- (16x1x10xf32, 1xf32) + # pd_op.scale: (16x1x20xf32) <- (16x1x20xf32, 1xf32) scale_22 = paddle._C_ops.scale(scale_21, full_2, float("0"), True) del full_2, scale_21 - # pd_op.cast: (16x1x10xf32) <- (16x1x10xf32) + # pd_op.cast: (16x1x20xf32) <- (16x1x20xf32) cast_10 = paddle._C_ops.cast(scale_22, paddle.float32) del scale_22 @@ -3914,35 +3914,35 @@ def forward( scale_24 = paddle._C_ops.scale(add_126, full_9, float("0"), True) del add_126 - # pd_op.matmul: (1x10x1024xf32) <- (1x10x1024xf32, 1024x1024xf32) + # pd_op.matmul: (1x20x1024xf32) <- (1x20x1024xf32, 1024x1024xf32) matmul_80 = paddle._C_ops.matmul(layer_norm_72, parameter_329, False, False) del parameter_329 - # pd_op.add: (1x10x1024xf32) <- (1x10x1024xf32, 1024xf32) + # pd_op.add: (1x20x1024xf32) <- (1x20x1024xf32, 1024xf32) add_127 = paddle._C_ops.add(matmul_80, parameter_328) del matmul_80, parameter_328 - # pd_op.reshape: (1x10x16x64xf32) <- (1x10x1024xf32, 4xi64) + # pd_op.reshape: (1x20x16x64xf32) <- (1x20x1024xf32, 4xi64) reshape_145 = paddle._C_ops.reshape(add_127, full_int_array_2) del add_127 - # pd_op.transpose: (1x16x10x64xf32) <- (1x10x16x64xf32) + # pd_op.transpose: (1x16x20x64xf32) <- (1x20x16x64xf32) transpose_70 = paddle._C_ops.transpose(reshape_145, [0, 2, 1, 3]) del reshape_145 - # pd_op.matmul: (1x10x1024xf32) <- (1x10x1024xf32, 1024x1024xf32) + # pd_op.matmul: (1x20x1024xf32) <- (1x20x1024xf32, 1024x1024xf32) matmul_81 = paddle._C_ops.matmul(layer_norm_72, parameter_327, False, False) del parameter_327 - # pd_op.add: (1x10x1024xf32) <- (1x10x1024xf32, 1024xf32) + # pd_op.add: (1x20x1024xf32) <- (1x20x1024xf32, 1024xf32) add_128 = paddle._C_ops.add(matmul_81, parameter_326) del matmul_81, parameter_326 - # pd_op.reshape: (1x10x16x64xf32) <- (1x10x1024xf32, 4xi64) + # pd_op.reshape: (1x20x16x64xf32) <- (1x20x1024xf32, 4xi64) reshape_146 = paddle._C_ops.reshape(add_128, full_int_array_2) del add_128 - # pd_op.transpose: (1x16x10x64xf32) <- (1x10x16x64xf32) + # pd_op.transpose: (1x16x20x64xf32) <- (1x20x16x64xf32) transpose_71 = paddle._C_ops.transpose(reshape_146, [0, 2, 1, 3]) del reshape_146 @@ -3958,29 +3958,31 @@ def forward( reshape_148 = paddle._C_ops.reshape(transpose_72, full_int_array_4) del transpose_72 - # pd_op.reshape: (16x10x64xf32) <- (1x16x10x64xf32, 3xi64) + # pd_op.reshape: (16x20x64xf32) <- (1x16x20x64xf32, 3xi64) reshape_149 = paddle._C_ops.reshape(transpose_70, full_int_array_4) + del transpose_70 - # pd_op.reshape: (16x10x64xf32) <- (1x16x10x64xf32, 3xi64) + # pd_op.reshape: (16x20x64xf32) <- (1x16x20x64xf32, 3xi64) reshape_150 = paddle._C_ops.reshape(transpose_71, full_int_array_4) + del transpose_71 - # pd_op.transpose: (16x64x10xf32) <- (16x10x64xf32) + # pd_op.transpose: (16x64x20xf32) <- (16x20x64xf32) transpose_73 = paddle._C_ops.transpose(reshape_149, [0, 2, 1]) del reshape_149 - # pd_op.bmm: (16x3x10xf32) <- (16x3x64xf32, 16x64x10xf32) + # pd_op.bmm: (16x3x20xf32) <- (16x3x64xf32, 16x64x20xf32) bmm_26 = paddle._C_ops.bmm(reshape_148, transpose_73) del reshape_148, transpose_73 - # pd_op.add: (16x3x10xf32) <- (16x3x10xf32, 16x1x10xf32) + # pd_op.add: (16x3x20xf32) <- (16x3x20xf32, 16x1x20xf32) add_129 = paddle._C_ops.add(bmm_26, cast_10) del bmm_26 - # pd_op.softmax: (16x3x10xf32) <- (16x3x10xf32) + # pd_op.softmax: (16x3x20xf32) <- (16x3x20xf32) softmax_14 = paddle._C_ops.softmax(add_129, -1) del add_129 - # pd_op.dropout: (16x3x10xf32, 16x3x10xui8) <- (16x3x10xf32, None, 1xf32) + # pd_op.dropout: (16x3x20xf32, 16x3x20xui8) <- (16x3x20xf32, None, 1xf32) dropout_106, dropout_107 = (lambda x, f: f(x))( paddle._C_ops.dropout( softmax_14, None, full_8, True, "upscale_in_train", 0, False @@ -3989,7 +3991,7 @@ def forward( ) del softmax_14 - # pd_op.bmm: (16x3x64xf32) <- (16x3x10xf32, 16x10x64xf32) + # pd_op.bmm: (16x3x64xf32) <- (16x3x20xf32, 16x20x64xf32) bmm_27 = paddle._C_ops.bmm(dropout_106, reshape_150) del dropout_106, reshape_150 @@ -4605,35 +4607,35 @@ def forward( scale_26 = paddle._C_ops.scale(add_147, full_9, float("0"), True) del add_147 - # pd_op.matmul: (1x10x1024xf32) <- (1x10x1024xf32, 1024x1024xf32) + # pd_op.matmul: (1x20x1024xf32) <- (1x20x1024xf32, 1024x1024xf32) matmul_93 = paddle._C_ops.matmul(layer_norm_72, parameter_301, False, False) del parameter_301 - # pd_op.add: (1x10x1024xf32) <- (1x10x1024xf32, 1024xf32) + # pd_op.add: (1x20x1024xf32) <- (1x20x1024xf32, 1024xf32) add_148 = paddle._C_ops.add(matmul_93, parameter_300) del matmul_93, parameter_300 - # pd_op.reshape: (1x10x16x64xf32) <- (1x10x1024xf32, 4xi64) + # pd_op.reshape: (1x20x16x64xf32) <- (1x20x1024xf32, 4xi64) reshape_176 = paddle._C_ops.reshape(add_148, full_int_array_2) del add_148 - # pd_op.transpose: (1x16x10x64xf32) <- (1x10x16x64xf32) + # pd_op.transpose: (1x16x20x64xf32) <- (1x20x16x64xf32) transpose_85 = paddle._C_ops.transpose(reshape_176, [0, 2, 1, 3]) del reshape_176 - # pd_op.matmul: (1x10x1024xf32) <- (1x10x1024xf32, 1024x1024xf32) + # pd_op.matmul: (1x20x1024xf32) <- (1x20x1024xf32, 1024x1024xf32) matmul_94 = paddle._C_ops.matmul(layer_norm_72, parameter_299, False, False) del parameter_299 - # pd_op.add: (1x10x1024xf32) <- (1x10x1024xf32, 1024xf32) + # pd_op.add: (1x20x1024xf32) <- (1x20x1024xf32, 1024xf32) add_149 = paddle._C_ops.add(matmul_94, parameter_298) del matmul_94, parameter_298 - # pd_op.reshape: (1x10x16x64xf32) <- (1x10x1024xf32, 4xi64) + # pd_op.reshape: (1x20x16x64xf32) <- (1x20x1024xf32, 4xi64) reshape_177 = paddle._C_ops.reshape(add_149, full_int_array_2) del add_149 - # pd_op.transpose: (1x16x10x64xf32) <- (1x10x16x64xf32) + # pd_op.transpose: (1x16x20x64xf32) <- (1x20x16x64xf32) transpose_86 = paddle._C_ops.transpose(reshape_177, [0, 2, 1, 3]) del reshape_177 @@ -4649,29 +4651,31 @@ def forward( reshape_179 = paddle._C_ops.reshape(transpose_87, full_int_array_4) del transpose_87 - # pd_op.reshape: (16x10x64xf32) <- (1x16x10x64xf32, 3xi64) + # pd_op.reshape: (16x20x64xf32) <- (1x16x20x64xf32, 3xi64) reshape_180 = paddle._C_ops.reshape(transpose_85, full_int_array_4) + del transpose_85 - # pd_op.reshape: (16x10x64xf32) <- (1x16x10x64xf32, 3xi64) + # pd_op.reshape: (16x20x64xf32) <- (1x16x20x64xf32, 3xi64) reshape_181 = paddle._C_ops.reshape(transpose_86, full_int_array_4) + del transpose_86 - # pd_op.transpose: (16x64x10xf32) <- (16x10x64xf32) + # pd_op.transpose: (16x64x20xf32) <- (16x20x64xf32) transpose_88 = paddle._C_ops.transpose(reshape_180, [0, 2, 1]) del reshape_180 - # pd_op.bmm: (16x3x10xf32) <- (16x3x64xf32, 16x64x10xf32) + # pd_op.bmm: (16x3x20xf32) <- (16x3x64xf32, 16x64x20xf32) bmm_30 = paddle._C_ops.bmm(reshape_179, transpose_88) del reshape_179, transpose_88 - # pd_op.add: (16x3x10xf32) <- (16x3x10xf32, 16x1x10xf32) + # pd_op.add: (16x3x20xf32) <- (16x3x20xf32, 16x1x20xf32) add_150 = paddle._C_ops.add(bmm_30, cast_10) del bmm_30 - # pd_op.softmax: (16x3x10xf32) <- (16x3x10xf32) + # pd_op.softmax: (16x3x20xf32) <- (16x3x20xf32) softmax_17 = paddle._C_ops.softmax(add_150, -1) del add_150 - # pd_op.dropout: (16x3x10xf32, 16x3x10xui8) <- (16x3x10xf32, None, 1xf32) + # pd_op.dropout: (16x3x20xf32, 16x3x20xui8) <- (16x3x20xf32, None, 1xf32) dropout_120, dropout_121 = (lambda x, f: f(x))( paddle._C_ops.dropout( softmax_17, None, full_8, True, "upscale_in_train", 0, False @@ -4680,7 +4684,7 @@ def forward( ) del softmax_17 - # pd_op.bmm: (16x3x64xf32) <- (16x3x10xf32, 16x10x64xf32) + # pd_op.bmm: (16x3x64xf32) <- (16x3x20xf32, 16x20x64xf32) bmm_31 = paddle._C_ops.bmm(dropout_120, reshape_181) del dropout_120, reshape_181 @@ -5290,35 +5294,35 @@ def forward( scale_28 = paddle._C_ops.scale(add_168, full_9, float("0"), True) del add_168 - # pd_op.matmul: (1x10x1024xf32) <- (1x10x1024xf32, 1024x1024xf32) + # pd_op.matmul: (1x20x1024xf32) <- (1x20x1024xf32, 1024x1024xf32) matmul_106 = paddle._C_ops.matmul(layer_norm_72, parameter_273, False, False) del parameter_273 - # pd_op.add: (1x10x1024xf32) <- (1x10x1024xf32, 1024xf32) + # pd_op.add: (1x20x1024xf32) <- (1x20x1024xf32, 1024xf32) add_169 = paddle._C_ops.add(matmul_106, parameter_272) del matmul_106, parameter_272 - # pd_op.reshape: (1x10x16x64xf32) <- (1x10x1024xf32, 4xi64) + # pd_op.reshape: (1x20x16x64xf32) <- (1x20x1024xf32, 4xi64) reshape_207 = paddle._C_ops.reshape(add_169, full_int_array_2) del add_169 - # pd_op.transpose: (1x16x10x64xf32) <- (1x10x16x64xf32) + # pd_op.transpose: (1x16x20x64xf32) <- (1x20x16x64xf32) transpose_100 = paddle._C_ops.transpose(reshape_207, [0, 2, 1, 3]) del reshape_207 - # pd_op.matmul: (1x10x1024xf32) <- (1x10x1024xf32, 1024x1024xf32) + # pd_op.matmul: (1x20x1024xf32) <- (1x20x1024xf32, 1024x1024xf32) matmul_107 = paddle._C_ops.matmul(layer_norm_72, parameter_271, False, False) del parameter_271 - # pd_op.add: (1x10x1024xf32) <- (1x10x1024xf32, 1024xf32) + # pd_op.add: (1x20x1024xf32) <- (1x20x1024xf32, 1024xf32) add_170 = paddle._C_ops.add(matmul_107, parameter_270) del matmul_107, parameter_270 - # pd_op.reshape: (1x10x16x64xf32) <- (1x10x1024xf32, 4xi64) + # pd_op.reshape: (1x20x16x64xf32) <- (1x20x1024xf32, 4xi64) reshape_208 = paddle._C_ops.reshape(add_170, full_int_array_2) del add_170 - # pd_op.transpose: (1x16x10x64xf32) <- (1x10x16x64xf32) + # pd_op.transpose: (1x16x20x64xf32) <- (1x20x16x64xf32) transpose_101 = paddle._C_ops.transpose(reshape_208, [0, 2, 1, 3]) del reshape_208 @@ -5334,29 +5338,31 @@ def forward( reshape_210 = paddle._C_ops.reshape(transpose_102, full_int_array_4) del transpose_102 - # pd_op.reshape: (16x10x64xf32) <- (1x16x10x64xf32, 3xi64) + # pd_op.reshape: (16x20x64xf32) <- (1x16x20x64xf32, 3xi64) reshape_211 = paddle._C_ops.reshape(transpose_100, full_int_array_4) + del transpose_100 - # pd_op.reshape: (16x10x64xf32) <- (1x16x10x64xf32, 3xi64) + # pd_op.reshape: (16x20x64xf32) <- (1x16x20x64xf32, 3xi64) reshape_212 = paddle._C_ops.reshape(transpose_101, full_int_array_4) + del transpose_101 - # pd_op.transpose: (16x64x10xf32) <- (16x10x64xf32) + # pd_op.transpose: (16x64x20xf32) <- (16x20x64xf32) transpose_103 = paddle._C_ops.transpose(reshape_211, [0, 2, 1]) del reshape_211 - # pd_op.bmm: (16x3x10xf32) <- (16x3x64xf32, 16x64x10xf32) + # pd_op.bmm: (16x3x20xf32) <- (16x3x64xf32, 16x64x20xf32) bmm_34 = paddle._C_ops.bmm(reshape_210, transpose_103) del reshape_210, transpose_103 - # pd_op.add: (16x3x10xf32) <- (16x3x10xf32, 16x1x10xf32) + # pd_op.add: (16x3x20xf32) <- (16x3x20xf32, 16x1x20xf32) add_171 = paddle._C_ops.add(bmm_34, cast_10) del bmm_34 - # pd_op.softmax: (16x3x10xf32) <- (16x3x10xf32) + # pd_op.softmax: (16x3x20xf32) <- (16x3x20xf32) softmax_20 = paddle._C_ops.softmax(add_171, -1) del add_171 - # pd_op.dropout: (16x3x10xf32, 16x3x10xui8) <- (16x3x10xf32, None, 1xf32) + # pd_op.dropout: (16x3x20xf32, 16x3x20xui8) <- (16x3x20xf32, None, 1xf32) dropout_134, dropout_135 = (lambda x, f: f(x))( paddle._C_ops.dropout( softmax_20, None, full_8, True, "upscale_in_train", 0, False @@ -5365,7 +5371,7 @@ def forward( ) del softmax_20 - # pd_op.bmm: (16x3x64xf32) <- (16x3x10xf32, 16x10x64xf32) + # pd_op.bmm: (16x3x64xf32) <- (16x3x20xf32, 16x20x64xf32) bmm_35 = paddle._C_ops.bmm(dropout_134, reshape_212) del dropout_134, reshape_212 @@ -5975,35 +5981,35 @@ def forward( scale_30 = paddle._C_ops.scale(add_189, full_9, float("0"), True) del add_189 - # pd_op.matmul: (1x10x1024xf32) <- (1x10x1024xf32, 1024x1024xf32) + # pd_op.matmul: (1x20x1024xf32) <- (1x20x1024xf32, 1024x1024xf32) matmul_119 = paddle._C_ops.matmul(layer_norm_72, parameter_245, False, False) del parameter_245 - # pd_op.add: (1x10x1024xf32) <- (1x10x1024xf32, 1024xf32) + # pd_op.add: (1x20x1024xf32) <- (1x20x1024xf32, 1024xf32) add_190 = paddle._C_ops.add(matmul_119, parameter_244) del matmul_119, parameter_244 - # pd_op.reshape: (1x10x16x64xf32) <- (1x10x1024xf32, 4xi64) + # pd_op.reshape: (1x20x16x64xf32) <- (1x20x1024xf32, 4xi64) reshape_238 = paddle._C_ops.reshape(add_190, full_int_array_2) del add_190 - # pd_op.transpose: (1x16x10x64xf32) <- (1x10x16x64xf32) + # pd_op.transpose: (1x16x20x64xf32) <- (1x20x16x64xf32) transpose_115 = paddle._C_ops.transpose(reshape_238, [0, 2, 1, 3]) del reshape_238 - # pd_op.matmul: (1x10x1024xf32) <- (1x10x1024xf32, 1024x1024xf32) + # pd_op.matmul: (1x20x1024xf32) <- (1x20x1024xf32, 1024x1024xf32) matmul_120 = paddle._C_ops.matmul(layer_norm_72, parameter_243, False, False) del parameter_243 - # pd_op.add: (1x10x1024xf32) <- (1x10x1024xf32, 1024xf32) + # pd_op.add: (1x20x1024xf32) <- (1x20x1024xf32, 1024xf32) add_191 = paddle._C_ops.add(matmul_120, parameter_242) del matmul_120, parameter_242 - # pd_op.reshape: (1x10x16x64xf32) <- (1x10x1024xf32, 4xi64) + # pd_op.reshape: (1x20x16x64xf32) <- (1x20x1024xf32, 4xi64) reshape_239 = paddle._C_ops.reshape(add_191, full_int_array_2) del add_191 - # pd_op.transpose: (1x16x10x64xf32) <- (1x10x16x64xf32) + # pd_op.transpose: (1x16x20x64xf32) <- (1x20x16x64xf32) transpose_116 = paddle._C_ops.transpose(reshape_239, [0, 2, 1, 3]) del reshape_239 @@ -6019,29 +6025,31 @@ def forward( reshape_241 = paddle._C_ops.reshape(transpose_117, full_int_array_4) del transpose_117 - # pd_op.reshape: (16x10x64xf32) <- (1x16x10x64xf32, 3xi64) + # pd_op.reshape: (16x20x64xf32) <- (1x16x20x64xf32, 3xi64) reshape_242 = paddle._C_ops.reshape(transpose_115, full_int_array_4) + del transpose_115 - # pd_op.reshape: (16x10x64xf32) <- (1x16x10x64xf32, 3xi64) + # pd_op.reshape: (16x20x64xf32) <- (1x16x20x64xf32, 3xi64) reshape_243 = paddle._C_ops.reshape(transpose_116, full_int_array_4) + del transpose_116 - # pd_op.transpose: (16x64x10xf32) <- (16x10x64xf32) + # pd_op.transpose: (16x64x20xf32) <- (16x20x64xf32) transpose_118 = paddle._C_ops.transpose(reshape_242, [0, 2, 1]) del reshape_242 - # pd_op.bmm: (16x3x10xf32) <- (16x3x64xf32, 16x64x10xf32) + # pd_op.bmm: (16x3x20xf32) <- (16x3x64xf32, 16x64x20xf32) bmm_38 = paddle._C_ops.bmm(reshape_241, transpose_118) del reshape_241, transpose_118 - # pd_op.add: (16x3x10xf32) <- (16x3x10xf32, 16x1x10xf32) + # pd_op.add: (16x3x20xf32) <- (16x3x20xf32, 16x1x20xf32) add_192 = paddle._C_ops.add(bmm_38, cast_10) del bmm_38 - # pd_op.softmax: (16x3x10xf32) <- (16x3x10xf32) + # pd_op.softmax: (16x3x20xf32) <- (16x3x20xf32) softmax_23 = paddle._C_ops.softmax(add_192, -1) del add_192 - # pd_op.dropout: (16x3x10xf32, 16x3x10xui8) <- (16x3x10xf32, None, 1xf32) + # pd_op.dropout: (16x3x20xf32, 16x3x20xui8) <- (16x3x20xf32, None, 1xf32) dropout_148, dropout_149 = (lambda x, f: f(x))( paddle._C_ops.dropout( softmax_23, None, full_8, True, "upscale_in_train", 0, False @@ -6050,7 +6058,7 @@ def forward( ) del softmax_23 - # pd_op.bmm: (16x3x64xf32) <- (16x3x10xf32, 16x10x64xf32) + # pd_op.bmm: (16x3x64xf32) <- (16x3x20xf32, 16x20x64xf32) bmm_39 = paddle._C_ops.bmm(dropout_148, reshape_243) del dropout_148, reshape_243 @@ -6660,35 +6668,35 @@ def forward( scale_32 = paddle._C_ops.scale(add_210, full_9, float("0"), True) del add_210 - # pd_op.matmul: (1x10x1024xf32) <- (1x10x1024xf32, 1024x1024xf32) + # pd_op.matmul: (1x20x1024xf32) <- (1x20x1024xf32, 1024x1024xf32) matmul_132 = paddle._C_ops.matmul(layer_norm_72, parameter_217, False, False) del parameter_217 - # pd_op.add: (1x10x1024xf32) <- (1x10x1024xf32, 1024xf32) + # pd_op.add: (1x20x1024xf32) <- (1x20x1024xf32, 1024xf32) add_211 = paddle._C_ops.add(matmul_132, parameter_216) del matmul_132, parameter_216 - # pd_op.reshape: (1x10x16x64xf32) <- (1x10x1024xf32, 4xi64) + # pd_op.reshape: (1x20x16x64xf32) <- (1x20x1024xf32, 4xi64) reshape_269 = paddle._C_ops.reshape(add_211, full_int_array_2) del add_211 - # pd_op.transpose: (1x16x10x64xf32) <- (1x10x16x64xf32) + # pd_op.transpose: (1x16x20x64xf32) <- (1x20x16x64xf32) transpose_130 = paddle._C_ops.transpose(reshape_269, [0, 2, 1, 3]) del reshape_269 - # pd_op.matmul: (1x10x1024xf32) <- (1x10x1024xf32, 1024x1024xf32) + # pd_op.matmul: (1x20x1024xf32) <- (1x20x1024xf32, 1024x1024xf32) matmul_133 = paddle._C_ops.matmul(layer_norm_72, parameter_215, False, False) del parameter_215 - # pd_op.add: (1x10x1024xf32) <- (1x10x1024xf32, 1024xf32) + # pd_op.add: (1x20x1024xf32) <- (1x20x1024xf32, 1024xf32) add_212 = paddle._C_ops.add(matmul_133, parameter_214) del matmul_133, parameter_214 - # pd_op.reshape: (1x10x16x64xf32) <- (1x10x1024xf32, 4xi64) + # pd_op.reshape: (1x20x16x64xf32) <- (1x20x1024xf32, 4xi64) reshape_270 = paddle._C_ops.reshape(add_212, full_int_array_2) del add_212 - # pd_op.transpose: (1x16x10x64xf32) <- (1x10x16x64xf32) + # pd_op.transpose: (1x16x20x64xf32) <- (1x20x16x64xf32) transpose_131 = paddle._C_ops.transpose(reshape_270, [0, 2, 1, 3]) del reshape_270 @@ -6704,29 +6712,31 @@ def forward( reshape_272 = paddle._C_ops.reshape(transpose_132, full_int_array_4) del transpose_132 - # pd_op.reshape: (16x10x64xf32) <- (1x16x10x64xf32, 3xi64) + # pd_op.reshape: (16x20x64xf32) <- (1x16x20x64xf32, 3xi64) reshape_273 = paddle._C_ops.reshape(transpose_130, full_int_array_4) + del transpose_130 - # pd_op.reshape: (16x10x64xf32) <- (1x16x10x64xf32, 3xi64) + # pd_op.reshape: (16x20x64xf32) <- (1x16x20x64xf32, 3xi64) reshape_274 = paddle._C_ops.reshape(transpose_131, full_int_array_4) + del transpose_131 - # pd_op.transpose: (16x64x10xf32) <- (16x10x64xf32) + # pd_op.transpose: (16x64x20xf32) <- (16x20x64xf32) transpose_133 = paddle._C_ops.transpose(reshape_273, [0, 2, 1]) del reshape_273 - # pd_op.bmm: (16x3x10xf32) <- (16x3x64xf32, 16x64x10xf32) + # pd_op.bmm: (16x3x20xf32) <- (16x3x64xf32, 16x64x20xf32) bmm_42 = paddle._C_ops.bmm(reshape_272, transpose_133) del reshape_272, transpose_133 - # pd_op.add: (16x3x10xf32) <- (16x3x10xf32, 16x1x10xf32) + # pd_op.add: (16x3x20xf32) <- (16x3x20xf32, 16x1x20xf32) add_213 = paddle._C_ops.add(bmm_42, cast_10) del bmm_42 - # pd_op.softmax: (16x3x10xf32) <- (16x3x10xf32) + # pd_op.softmax: (16x3x20xf32) <- (16x3x20xf32) softmax_26 = paddle._C_ops.softmax(add_213, -1) del add_213 - # pd_op.dropout: (16x3x10xf32, 16x3x10xui8) <- (16x3x10xf32, None, 1xf32) + # pd_op.dropout: (16x3x20xf32, 16x3x20xui8) <- (16x3x20xf32, None, 1xf32) dropout_162, dropout_163 = (lambda x, f: f(x))( paddle._C_ops.dropout( softmax_26, None, full_8, True, "upscale_in_train", 0, False @@ -6735,7 +6745,7 @@ def forward( ) del softmax_26 - # pd_op.bmm: (16x3x64xf32) <- (16x3x10xf32, 16x10x64xf32) + # pd_op.bmm: (16x3x64xf32) <- (16x3x20xf32, 16x20x64xf32) bmm_43 = paddle._C_ops.bmm(dropout_162, reshape_274) del dropout_162, reshape_274 @@ -7345,35 +7355,35 @@ def forward( scale_34 = paddle._C_ops.scale(add_231, full_9, float("0"), True) del add_231 - # pd_op.matmul: (1x10x1024xf32) <- (1x10x1024xf32, 1024x1024xf32) + # pd_op.matmul: (1x20x1024xf32) <- (1x20x1024xf32, 1024x1024xf32) matmul_145 = paddle._C_ops.matmul(layer_norm_72, parameter_189, False, False) del parameter_189 - # pd_op.add: (1x10x1024xf32) <- (1x10x1024xf32, 1024xf32) + # pd_op.add: (1x20x1024xf32) <- (1x20x1024xf32, 1024xf32) add_232 = paddle._C_ops.add(matmul_145, parameter_188) del matmul_145, parameter_188 - # pd_op.reshape: (1x10x16x64xf32) <- (1x10x1024xf32, 4xi64) + # pd_op.reshape: (1x20x16x64xf32) <- (1x20x1024xf32, 4xi64) reshape_300 = paddle._C_ops.reshape(add_232, full_int_array_2) del add_232 - # pd_op.transpose: (1x16x10x64xf32) <- (1x10x16x64xf32) + # pd_op.transpose: (1x16x20x64xf32) <- (1x20x16x64xf32) transpose_145 = paddle._C_ops.transpose(reshape_300, [0, 2, 1, 3]) del reshape_300 - # pd_op.matmul: (1x10x1024xf32) <- (1x10x1024xf32, 1024x1024xf32) + # pd_op.matmul: (1x20x1024xf32) <- (1x20x1024xf32, 1024x1024xf32) matmul_146 = paddle._C_ops.matmul(layer_norm_72, parameter_187, False, False) del parameter_187 - # pd_op.add: (1x10x1024xf32) <- (1x10x1024xf32, 1024xf32) + # pd_op.add: (1x20x1024xf32) <- (1x20x1024xf32, 1024xf32) add_233 = paddle._C_ops.add(matmul_146, parameter_186) del matmul_146, parameter_186 - # pd_op.reshape: (1x10x16x64xf32) <- (1x10x1024xf32, 4xi64) + # pd_op.reshape: (1x20x16x64xf32) <- (1x20x1024xf32, 4xi64) reshape_301 = paddle._C_ops.reshape(add_233, full_int_array_2) del add_233 - # pd_op.transpose: (1x16x10x64xf32) <- (1x10x16x64xf32) + # pd_op.transpose: (1x16x20x64xf32) <- (1x20x16x64xf32) transpose_146 = paddle._C_ops.transpose(reshape_301, [0, 2, 1, 3]) del reshape_301 @@ -7389,29 +7399,31 @@ def forward( reshape_303 = paddle._C_ops.reshape(transpose_147, full_int_array_4) del transpose_147 - # pd_op.reshape: (16x10x64xf32) <- (1x16x10x64xf32, 3xi64) + # pd_op.reshape: (16x20x64xf32) <- (1x16x20x64xf32, 3xi64) reshape_304 = paddle._C_ops.reshape(transpose_145, full_int_array_4) + del transpose_145 - # pd_op.reshape: (16x10x64xf32) <- (1x16x10x64xf32, 3xi64) + # pd_op.reshape: (16x20x64xf32) <- (1x16x20x64xf32, 3xi64) reshape_305 = paddle._C_ops.reshape(transpose_146, full_int_array_4) + del transpose_146 - # pd_op.transpose: (16x64x10xf32) <- (16x10x64xf32) + # pd_op.transpose: (16x64x20xf32) <- (16x20x64xf32) transpose_148 = paddle._C_ops.transpose(reshape_304, [0, 2, 1]) del reshape_304 - # pd_op.bmm: (16x3x10xf32) <- (16x3x64xf32, 16x64x10xf32) + # pd_op.bmm: (16x3x20xf32) <- (16x3x64xf32, 16x64x20xf32) bmm_46 = paddle._C_ops.bmm(reshape_303, transpose_148) del reshape_303, transpose_148 - # pd_op.add: (16x3x10xf32) <- (16x3x10xf32, 16x1x10xf32) + # pd_op.add: (16x3x20xf32) <- (16x3x20xf32, 16x1x20xf32) add_234 = paddle._C_ops.add(bmm_46, cast_10) del bmm_46 - # pd_op.softmax: (16x3x10xf32) <- (16x3x10xf32) + # pd_op.softmax: (16x3x20xf32) <- (16x3x20xf32) softmax_29 = paddle._C_ops.softmax(add_234, -1) del add_234 - # pd_op.dropout: (16x3x10xf32, 16x3x10xui8) <- (16x3x10xf32, None, 1xf32) + # pd_op.dropout: (16x3x20xf32, 16x3x20xui8) <- (16x3x20xf32, None, 1xf32) dropout_176, dropout_177 = (lambda x, f: f(x))( paddle._C_ops.dropout( softmax_29, None, full_8, True, "upscale_in_train", 0, False @@ -7420,7 +7432,7 @@ def forward( ) del softmax_29 - # pd_op.bmm: (16x3x64xf32) <- (16x3x10xf32, 16x10x64xf32) + # pd_op.bmm: (16x3x64xf32) <- (16x3x20xf32, 16x20x64xf32) bmm_47 = paddle._C_ops.bmm(dropout_176, reshape_305) del dropout_176, reshape_305 @@ -8030,35 +8042,35 @@ def forward( scale_36 = paddle._C_ops.scale(add_252, full_9, float("0"), True) del add_252 - # pd_op.matmul: (1x10x1024xf32) <- (1x10x1024xf32, 1024x1024xf32) + # pd_op.matmul: (1x20x1024xf32) <- (1x20x1024xf32, 1024x1024xf32) matmul_158 = paddle._C_ops.matmul(layer_norm_72, parameter_161, False, False) del parameter_161 - # pd_op.add: (1x10x1024xf32) <- (1x10x1024xf32, 1024xf32) + # pd_op.add: (1x20x1024xf32) <- (1x20x1024xf32, 1024xf32) add_253 = paddle._C_ops.add(matmul_158, parameter_160) del matmul_158, parameter_160 - # pd_op.reshape: (1x10x16x64xf32) <- (1x10x1024xf32, 4xi64) + # pd_op.reshape: (1x20x16x64xf32) <- (1x20x1024xf32, 4xi64) reshape_331 = paddle._C_ops.reshape(add_253, full_int_array_2) del add_253 - # pd_op.transpose: (1x16x10x64xf32) <- (1x10x16x64xf32) + # pd_op.transpose: (1x16x20x64xf32) <- (1x20x16x64xf32) transpose_160 = paddle._C_ops.transpose(reshape_331, [0, 2, 1, 3]) del reshape_331 - # pd_op.matmul: (1x10x1024xf32) <- (1x10x1024xf32, 1024x1024xf32) + # pd_op.matmul: (1x20x1024xf32) <- (1x20x1024xf32, 1024x1024xf32) matmul_159 = paddle._C_ops.matmul(layer_norm_72, parameter_159, False, False) del parameter_159 - # pd_op.add: (1x10x1024xf32) <- (1x10x1024xf32, 1024xf32) + # pd_op.add: (1x20x1024xf32) <- (1x20x1024xf32, 1024xf32) add_254 = paddle._C_ops.add(matmul_159, parameter_158) del matmul_159, parameter_158 - # pd_op.reshape: (1x10x16x64xf32) <- (1x10x1024xf32, 4xi64) + # pd_op.reshape: (1x20x16x64xf32) <- (1x20x1024xf32, 4xi64) reshape_332 = paddle._C_ops.reshape(add_254, full_int_array_2) del add_254 - # pd_op.transpose: (1x16x10x64xf32) <- (1x10x16x64xf32) + # pd_op.transpose: (1x16x20x64xf32) <- (1x20x16x64xf32) transpose_161 = paddle._C_ops.transpose(reshape_332, [0, 2, 1, 3]) del reshape_332 @@ -8074,29 +8086,31 @@ def forward( reshape_334 = paddle._C_ops.reshape(transpose_162, full_int_array_4) del transpose_162 - # pd_op.reshape: (16x10x64xf32) <- (1x16x10x64xf32, 3xi64) + # pd_op.reshape: (16x20x64xf32) <- (1x16x20x64xf32, 3xi64) reshape_335 = paddle._C_ops.reshape(transpose_160, full_int_array_4) + del transpose_160 - # pd_op.reshape: (16x10x64xf32) <- (1x16x10x64xf32, 3xi64) + # pd_op.reshape: (16x20x64xf32) <- (1x16x20x64xf32, 3xi64) reshape_336 = paddle._C_ops.reshape(transpose_161, full_int_array_4) + del transpose_161 - # pd_op.transpose: (16x64x10xf32) <- (16x10x64xf32) + # pd_op.transpose: (16x64x20xf32) <- (16x20x64xf32) transpose_163 = paddle._C_ops.transpose(reshape_335, [0, 2, 1]) del reshape_335 - # pd_op.bmm: (16x3x10xf32) <- (16x3x64xf32, 16x64x10xf32) + # pd_op.bmm: (16x3x20xf32) <- (16x3x64xf32, 16x64x20xf32) bmm_50 = paddle._C_ops.bmm(reshape_334, transpose_163) del reshape_334, transpose_163 - # pd_op.add: (16x3x10xf32) <- (16x3x10xf32, 16x1x10xf32) + # pd_op.add: (16x3x20xf32) <- (16x3x20xf32, 16x1x20xf32) add_255 = paddle._C_ops.add(bmm_50, cast_10) del bmm_50 - # pd_op.softmax: (16x3x10xf32) <- (16x3x10xf32) + # pd_op.softmax: (16x3x20xf32) <- (16x3x20xf32) softmax_32 = paddle._C_ops.softmax(add_255, -1) del add_255 - # pd_op.dropout: (16x3x10xf32, 16x3x10xui8) <- (16x3x10xf32, None, 1xf32) + # pd_op.dropout: (16x3x20xf32, 16x3x20xui8) <- (16x3x20xf32, None, 1xf32) dropout_190, dropout_191 = (lambda x, f: f(x))( paddle._C_ops.dropout( softmax_32, None, full_8, True, "upscale_in_train", 0, False @@ -8105,7 +8119,7 @@ def forward( ) del softmax_32 - # pd_op.bmm: (16x3x64xf32) <- (16x3x10xf32, 16x10x64xf32) + # pd_op.bmm: (16x3x64xf32) <- (16x3x20xf32, 16x20x64xf32) bmm_51 = paddle._C_ops.bmm(dropout_190, reshape_336) del dropout_190, reshape_336 @@ -8715,35 +8729,35 @@ def forward( scale_38 = paddle._C_ops.scale(add_273, full_9, float("0"), True) del add_273 - # pd_op.matmul: (1x10x1024xf32) <- (1x10x1024xf32, 1024x1024xf32) + # pd_op.matmul: (1x20x1024xf32) <- (1x20x1024xf32, 1024x1024xf32) matmul_171 = paddle._C_ops.matmul(layer_norm_72, parameter_133, False, False) del parameter_133 - # pd_op.add: (1x10x1024xf32) <- (1x10x1024xf32, 1024xf32) + # pd_op.add: (1x20x1024xf32) <- (1x20x1024xf32, 1024xf32) add_274 = paddle._C_ops.add(matmul_171, parameter_132) del matmul_171, parameter_132 - # pd_op.reshape: (1x10x16x64xf32) <- (1x10x1024xf32, 4xi64) + # pd_op.reshape: (1x20x16x64xf32) <- (1x20x1024xf32, 4xi64) reshape_362 = paddle._C_ops.reshape(add_274, full_int_array_2) del add_274 - # pd_op.transpose: (1x16x10x64xf32) <- (1x10x16x64xf32) + # pd_op.transpose: (1x16x20x64xf32) <- (1x20x16x64xf32) transpose_175 = paddle._C_ops.transpose(reshape_362, [0, 2, 1, 3]) del reshape_362 - # pd_op.matmul: (1x10x1024xf32) <- (1x10x1024xf32, 1024x1024xf32) + # pd_op.matmul: (1x20x1024xf32) <- (1x20x1024xf32, 1024x1024xf32) matmul_172 = paddle._C_ops.matmul(layer_norm_72, parameter_131, False, False) del parameter_131 - # pd_op.add: (1x10x1024xf32) <- (1x10x1024xf32, 1024xf32) + # pd_op.add: (1x20x1024xf32) <- (1x20x1024xf32, 1024xf32) add_275 = paddle._C_ops.add(matmul_172, parameter_130) del matmul_172, parameter_130 - # pd_op.reshape: (1x10x16x64xf32) <- (1x10x1024xf32, 4xi64) + # pd_op.reshape: (1x20x16x64xf32) <- (1x20x1024xf32, 4xi64) reshape_363 = paddle._C_ops.reshape(add_275, full_int_array_2) del add_275 - # pd_op.transpose: (1x16x10x64xf32) <- (1x10x16x64xf32) + # pd_op.transpose: (1x16x20x64xf32) <- (1x20x16x64xf32) transpose_176 = paddle._C_ops.transpose(reshape_363, [0, 2, 1, 3]) del reshape_363 @@ -8759,29 +8773,31 @@ def forward( reshape_365 = paddle._C_ops.reshape(transpose_177, full_int_array_4) del transpose_177 - # pd_op.reshape: (16x10x64xf32) <- (1x16x10x64xf32, 3xi64) + # pd_op.reshape: (16x20x64xf32) <- (1x16x20x64xf32, 3xi64) reshape_366 = paddle._C_ops.reshape(transpose_175, full_int_array_4) + del transpose_175 - # pd_op.reshape: (16x10x64xf32) <- (1x16x10x64xf32, 3xi64) + # pd_op.reshape: (16x20x64xf32) <- (1x16x20x64xf32, 3xi64) reshape_367 = paddle._C_ops.reshape(transpose_176, full_int_array_4) + del transpose_176 - # pd_op.transpose: (16x64x10xf32) <- (16x10x64xf32) + # pd_op.transpose: (16x64x20xf32) <- (16x20x64xf32) transpose_178 = paddle._C_ops.transpose(reshape_366, [0, 2, 1]) del reshape_366 - # pd_op.bmm: (16x3x10xf32) <- (16x3x64xf32, 16x64x10xf32) + # pd_op.bmm: (16x3x20xf32) <- (16x3x64xf32, 16x64x20xf32) bmm_54 = paddle._C_ops.bmm(reshape_365, transpose_178) del reshape_365, transpose_178 - # pd_op.add: (16x3x10xf32) <- (16x3x10xf32, 16x1x10xf32) + # pd_op.add: (16x3x20xf32) <- (16x3x20xf32, 16x1x20xf32) add_276 = paddle._C_ops.add(bmm_54, cast_10) del bmm_54 - # pd_op.softmax: (16x3x10xf32) <- (16x3x10xf32) + # pd_op.softmax: (16x3x20xf32) <- (16x3x20xf32) softmax_35 = paddle._C_ops.softmax(add_276, -1) del add_276 - # pd_op.dropout: (16x3x10xf32, 16x3x10xui8) <- (16x3x10xf32, None, 1xf32) + # pd_op.dropout: (16x3x20xf32, 16x3x20xui8) <- (16x3x20xf32, None, 1xf32) dropout_204, dropout_205 = (lambda x, f: f(x))( paddle._C_ops.dropout( softmax_35, None, full_8, True, "upscale_in_train", 0, False @@ -8790,7 +8806,7 @@ def forward( ) del softmax_35 - # pd_op.bmm: (16x3x64xf32) <- (16x3x10xf32, 16x10x64xf32) + # pd_op.bmm: (16x3x64xf32) <- (16x3x20xf32, 16x20x64xf32) bmm_55 = paddle._C_ops.bmm(dropout_204, reshape_367) del dropout_204, reshape_367 @@ -9400,35 +9416,35 @@ def forward( scale_40 = paddle._C_ops.scale(add_294, full_9, float("0"), True) del add_294 - # pd_op.matmul: (1x10x1024xf32) <- (1x10x1024xf32, 1024x1024xf32) + # pd_op.matmul: (1x20x1024xf32) <- (1x20x1024xf32, 1024x1024xf32) matmul_184 = paddle._C_ops.matmul(layer_norm_72, parameter_105, False, False) del parameter_105 - # pd_op.add: (1x10x1024xf32) <- (1x10x1024xf32, 1024xf32) + # pd_op.add: (1x20x1024xf32) <- (1x20x1024xf32, 1024xf32) add_295 = paddle._C_ops.add(matmul_184, parameter_104) del matmul_184, parameter_104 - # pd_op.reshape: (1x10x16x64xf32) <- (1x10x1024xf32, 4xi64) + # pd_op.reshape: (1x20x16x64xf32) <- (1x20x1024xf32, 4xi64) reshape_393 = paddle._C_ops.reshape(add_295, full_int_array_2) del add_295 - # pd_op.transpose: (1x16x10x64xf32) <- (1x10x16x64xf32) + # pd_op.transpose: (1x16x20x64xf32) <- (1x20x16x64xf32) transpose_190 = paddle._C_ops.transpose(reshape_393, [0, 2, 1, 3]) del reshape_393 - # pd_op.matmul: (1x10x1024xf32) <- (1x10x1024xf32, 1024x1024xf32) + # pd_op.matmul: (1x20x1024xf32) <- (1x20x1024xf32, 1024x1024xf32) matmul_185 = paddle._C_ops.matmul(layer_norm_72, parameter_103, False, False) del parameter_103 - # pd_op.add: (1x10x1024xf32) <- (1x10x1024xf32, 1024xf32) + # pd_op.add: (1x20x1024xf32) <- (1x20x1024xf32, 1024xf32) add_296 = paddle._C_ops.add(matmul_185, parameter_102) del matmul_185, parameter_102 - # pd_op.reshape: (1x10x16x64xf32) <- (1x10x1024xf32, 4xi64) + # pd_op.reshape: (1x20x16x64xf32) <- (1x20x1024xf32, 4xi64) reshape_394 = paddle._C_ops.reshape(add_296, full_int_array_2) del add_296 - # pd_op.transpose: (1x16x10x64xf32) <- (1x10x16x64xf32) + # pd_op.transpose: (1x16x20x64xf32) <- (1x20x16x64xf32) transpose_191 = paddle._C_ops.transpose(reshape_394, [0, 2, 1, 3]) del reshape_394 @@ -9444,29 +9460,31 @@ def forward( reshape_396 = paddle._C_ops.reshape(transpose_192, full_int_array_4) del transpose_192 - # pd_op.reshape: (16x10x64xf32) <- (1x16x10x64xf32, 3xi64) + # pd_op.reshape: (16x20x64xf32) <- (1x16x20x64xf32, 3xi64) reshape_397 = paddle._C_ops.reshape(transpose_190, full_int_array_4) + del transpose_190 - # pd_op.reshape: (16x10x64xf32) <- (1x16x10x64xf32, 3xi64) + # pd_op.reshape: (16x20x64xf32) <- (1x16x20x64xf32, 3xi64) reshape_398 = paddle._C_ops.reshape(transpose_191, full_int_array_4) + del transpose_191 - # pd_op.transpose: (16x64x10xf32) <- (16x10x64xf32) + # pd_op.transpose: (16x64x20xf32) <- (16x20x64xf32) transpose_193 = paddle._C_ops.transpose(reshape_397, [0, 2, 1]) del reshape_397 - # pd_op.bmm: (16x3x10xf32) <- (16x3x64xf32, 16x64x10xf32) + # pd_op.bmm: (16x3x20xf32) <- (16x3x64xf32, 16x64x20xf32) bmm_58 = paddle._C_ops.bmm(reshape_396, transpose_193) del reshape_396, transpose_193 - # pd_op.add: (16x3x10xf32) <- (16x3x10xf32, 16x1x10xf32) + # pd_op.add: (16x3x20xf32) <- (16x3x20xf32, 16x1x20xf32) add_297 = paddle._C_ops.add(bmm_58, cast_10) del bmm_58 - # pd_op.softmax: (16x3x10xf32) <- (16x3x10xf32) + # pd_op.softmax: (16x3x20xf32) <- (16x3x20xf32) softmax_38 = paddle._C_ops.softmax(add_297, -1) del add_297 - # pd_op.dropout: (16x3x10xf32, 16x3x10xui8) <- (16x3x10xf32, None, 1xf32) + # pd_op.dropout: (16x3x20xf32, 16x3x20xui8) <- (16x3x20xf32, None, 1xf32) dropout_218, dropout_219 = (lambda x, f: f(x))( paddle._C_ops.dropout( softmax_38, None, full_8, True, "upscale_in_train", 0, False @@ -9475,7 +9493,7 @@ def forward( ) del softmax_38 - # pd_op.bmm: (16x3x64xf32) <- (16x3x10xf32, 16x10x64xf32) + # pd_op.bmm: (16x3x64xf32) <- (16x3x20xf32, 16x20x64xf32) bmm_59 = paddle._C_ops.bmm(dropout_218, reshape_398) del dropout_218, reshape_398 @@ -10085,35 +10103,35 @@ def forward( scale_42 = paddle._C_ops.scale(add_315, full_9, float("0"), True) del add_315 - # pd_op.matmul: (1x10x1024xf32) <- (1x10x1024xf32, 1024x1024xf32) + # pd_op.matmul: (1x20x1024xf32) <- (1x20x1024xf32, 1024x1024xf32) matmul_197 = paddle._C_ops.matmul(layer_norm_72, parameter_77, False, False) del parameter_77 - # pd_op.add: (1x10x1024xf32) <- (1x10x1024xf32, 1024xf32) + # pd_op.add: (1x20x1024xf32) <- (1x20x1024xf32, 1024xf32) add_316 = paddle._C_ops.add(matmul_197, parameter_76) del matmul_197, parameter_76 - # pd_op.reshape: (1x10x16x64xf32) <- (1x10x1024xf32, 4xi64) + # pd_op.reshape: (1x20x16x64xf32) <- (1x20x1024xf32, 4xi64) reshape_424 = paddle._C_ops.reshape(add_316, full_int_array_2) del add_316 - # pd_op.transpose: (1x16x10x64xf32) <- (1x10x16x64xf32) + # pd_op.transpose: (1x16x20x64xf32) <- (1x20x16x64xf32) transpose_205 = paddle._C_ops.transpose(reshape_424, [0, 2, 1, 3]) del reshape_424 - # pd_op.matmul: (1x10x1024xf32) <- (1x10x1024xf32, 1024x1024xf32) + # pd_op.matmul: (1x20x1024xf32) <- (1x20x1024xf32, 1024x1024xf32) matmul_198 = paddle._C_ops.matmul(layer_norm_72, parameter_75, False, False) del parameter_75 - # pd_op.add: (1x10x1024xf32) <- (1x10x1024xf32, 1024xf32) + # pd_op.add: (1x20x1024xf32) <- (1x20x1024xf32, 1024xf32) add_317 = paddle._C_ops.add(matmul_198, parameter_74) del matmul_198, parameter_74 - # pd_op.reshape: (1x10x16x64xf32) <- (1x10x1024xf32, 4xi64) + # pd_op.reshape: (1x20x16x64xf32) <- (1x20x1024xf32, 4xi64) reshape_425 = paddle._C_ops.reshape(add_317, full_int_array_2) del add_317 - # pd_op.transpose: (1x16x10x64xf32) <- (1x10x16x64xf32) + # pd_op.transpose: (1x16x20x64xf32) <- (1x20x16x64xf32) transpose_206 = paddle._C_ops.transpose(reshape_425, [0, 2, 1, 3]) del reshape_425 @@ -10129,29 +10147,31 @@ def forward( reshape_427 = paddle._C_ops.reshape(transpose_207, full_int_array_4) del transpose_207 - # pd_op.reshape: (16x10x64xf32) <- (1x16x10x64xf32, 3xi64) + # pd_op.reshape: (16x20x64xf32) <- (1x16x20x64xf32, 3xi64) reshape_428 = paddle._C_ops.reshape(transpose_205, full_int_array_4) + del transpose_205 - # pd_op.reshape: (16x10x64xf32) <- (1x16x10x64xf32, 3xi64) + # pd_op.reshape: (16x20x64xf32) <- (1x16x20x64xf32, 3xi64) reshape_429 = paddle._C_ops.reshape(transpose_206, full_int_array_4) + del transpose_206 - # pd_op.transpose: (16x64x10xf32) <- (16x10x64xf32) + # pd_op.transpose: (16x64x20xf32) <- (16x20x64xf32) transpose_208 = paddle._C_ops.transpose(reshape_428, [0, 2, 1]) del reshape_428 - # pd_op.bmm: (16x3x10xf32) <- (16x3x64xf32, 16x64x10xf32) + # pd_op.bmm: (16x3x20xf32) <- (16x3x64xf32, 16x64x20xf32) bmm_62 = paddle._C_ops.bmm(reshape_427, transpose_208) del reshape_427, transpose_208 - # pd_op.add: (16x3x10xf32) <- (16x3x10xf32, 16x1x10xf32) + # pd_op.add: (16x3x20xf32) <- (16x3x20xf32, 16x1x20xf32) add_318 = paddle._C_ops.add(bmm_62, cast_10) del bmm_62 - # pd_op.softmax: (16x3x10xf32) <- (16x3x10xf32) + # pd_op.softmax: (16x3x20xf32) <- (16x3x20xf32) softmax_41 = paddle._C_ops.softmax(add_318, -1) del add_318 - # pd_op.dropout: (16x3x10xf32, 16x3x10xui8) <- (16x3x10xf32, None, 1xf32) + # pd_op.dropout: (16x3x20xf32, 16x3x20xui8) <- (16x3x20xf32, None, 1xf32) dropout_232, dropout_233 = (lambda x, f: f(x))( paddle._C_ops.dropout( softmax_41, None, full_8, True, "upscale_in_train", 0, False @@ -10160,7 +10180,7 @@ def forward( ) del softmax_41 - # pd_op.bmm: (16x3x64xf32) <- (16x3x10xf32, 16x10x64xf32) + # pd_op.bmm: (16x3x64xf32) <- (16x3x20xf32, 16x20x64xf32) bmm_63 = paddle._C_ops.bmm(dropout_232, reshape_429) del dropout_232, reshape_429 @@ -10770,35 +10790,35 @@ def forward( scale_44 = paddle._C_ops.scale(add_336, full_9, float("0"), True) del add_336 - # pd_op.matmul: (1x10x1024xf32) <- (1x10x1024xf32, 1024x1024xf32) + # pd_op.matmul: (1x20x1024xf32) <- (1x20x1024xf32, 1024x1024xf32) matmul_210 = paddle._C_ops.matmul(layer_norm_72, parameter_49, False, False) del parameter_49 - # pd_op.add: (1x10x1024xf32) <- (1x10x1024xf32, 1024xf32) + # pd_op.add: (1x20x1024xf32) <- (1x20x1024xf32, 1024xf32) add_337 = paddle._C_ops.add(matmul_210, parameter_48) del matmul_210, parameter_48 - # pd_op.reshape: (1x10x16x64xf32) <- (1x10x1024xf32, 4xi64) + # pd_op.reshape: (1x20x16x64xf32) <- (1x20x1024xf32, 4xi64) reshape_455 = paddle._C_ops.reshape(add_337, full_int_array_2) del add_337 - # pd_op.transpose: (1x16x10x64xf32) <- (1x10x16x64xf32) + # pd_op.transpose: (1x16x20x64xf32) <- (1x20x16x64xf32) transpose_220 = paddle._C_ops.transpose(reshape_455, [0, 2, 1, 3]) del reshape_455 - # pd_op.matmul: (1x10x1024xf32) <- (1x10x1024xf32, 1024x1024xf32) + # pd_op.matmul: (1x20x1024xf32) <- (1x20x1024xf32, 1024x1024xf32) matmul_211 = paddle._C_ops.matmul(layer_norm_72, parameter_47, False, False) del parameter_47 - # pd_op.add: (1x10x1024xf32) <- (1x10x1024xf32, 1024xf32) + # pd_op.add: (1x20x1024xf32) <- (1x20x1024xf32, 1024xf32) add_338 = paddle._C_ops.add(matmul_211, parameter_46) del matmul_211, parameter_46 - # pd_op.reshape: (1x10x16x64xf32) <- (1x10x1024xf32, 4xi64) + # pd_op.reshape: (1x20x16x64xf32) <- (1x20x1024xf32, 4xi64) reshape_456 = paddle._C_ops.reshape(add_338, full_int_array_2) del add_338 - # pd_op.transpose: (1x16x10x64xf32) <- (1x10x16x64xf32) + # pd_op.transpose: (1x16x20x64xf32) <- (1x20x16x64xf32) transpose_221 = paddle._C_ops.transpose(reshape_456, [0, 2, 1, 3]) del reshape_456 @@ -10814,29 +10834,31 @@ def forward( reshape_458 = paddle._C_ops.reshape(transpose_222, full_int_array_4) del transpose_222 - # pd_op.reshape: (16x10x64xf32) <- (1x16x10x64xf32, 3xi64) + # pd_op.reshape: (16x20x64xf32) <- (1x16x20x64xf32, 3xi64) reshape_459 = paddle._C_ops.reshape(transpose_220, full_int_array_4) + del transpose_220 - # pd_op.reshape: (16x10x64xf32) <- (1x16x10x64xf32, 3xi64) + # pd_op.reshape: (16x20x64xf32) <- (1x16x20x64xf32, 3xi64) reshape_460 = paddle._C_ops.reshape(transpose_221, full_int_array_4) + del transpose_221 - # pd_op.transpose: (16x64x10xf32) <- (16x10x64xf32) + # pd_op.transpose: (16x64x20xf32) <- (16x20x64xf32) transpose_223 = paddle._C_ops.transpose(reshape_459, [0, 2, 1]) del reshape_459 - # pd_op.bmm: (16x3x10xf32) <- (16x3x64xf32, 16x64x10xf32) + # pd_op.bmm: (16x3x20xf32) <- (16x3x64xf32, 16x64x20xf32) bmm_66 = paddle._C_ops.bmm(reshape_458, transpose_223) del reshape_458, transpose_223 - # pd_op.add: (16x3x10xf32) <- (16x3x10xf32, 16x1x10xf32) + # pd_op.add: (16x3x20xf32) <- (16x3x20xf32, 16x1x20xf32) add_339 = paddle._C_ops.add(bmm_66, cast_10) del bmm_66 - # pd_op.softmax: (16x3x10xf32) <- (16x3x10xf32) + # pd_op.softmax: (16x3x20xf32) <- (16x3x20xf32) softmax_44 = paddle._C_ops.softmax(add_339, -1) del add_339 - # pd_op.dropout: (16x3x10xf32, 16x3x10xui8) <- (16x3x10xf32, None, 1xf32) + # pd_op.dropout: (16x3x20xf32, 16x3x20xui8) <- (16x3x20xf32, None, 1xf32) dropout_246, dropout_247 = (lambda x, f: f(x))( paddle._C_ops.dropout( softmax_44, None, full_8, True, "upscale_in_train", 0, False @@ -10845,7 +10867,7 @@ def forward( ) del softmax_44 - # pd_op.bmm: (16x3x64xf32) <- (16x3x10xf32, 16x10x64xf32) + # pd_op.bmm: (16x3x64xf32) <- (16x3x20xf32, 16x20x64xf32) bmm_67 = paddle._C_ops.bmm(dropout_246, reshape_460) del dropout_246, reshape_460 @@ -11458,35 +11480,35 @@ def forward( scale_46 = paddle._C_ops.scale(add_357, full_9, float("0"), True) del add_357, full_9 - # pd_op.matmul: (1x10x1024xf32) <- (1x10x1024xf32, 1024x1024xf32) + # pd_op.matmul: (1x20x1024xf32) <- (1x20x1024xf32, 1024x1024xf32) matmul_223 = paddle._C_ops.matmul(layer_norm_72, parameter_21, False, False) del parameter_21 - # pd_op.add: (1x10x1024xf32) <- (1x10x1024xf32, 1024xf32) + # pd_op.add: (1x20x1024xf32) <- (1x20x1024xf32, 1024xf32) add_358 = paddle._C_ops.add(matmul_223, parameter_20) del matmul_223, parameter_20 - # pd_op.reshape: (1x10x16x64xf32) <- (1x10x1024xf32, 4xi64) + # pd_op.reshape: (1x20x16x64xf32) <- (1x20x1024xf32, 4xi64) reshape_486 = paddle._C_ops.reshape(add_358, full_int_array_2) del add_358 - # pd_op.transpose: (1x16x10x64xf32) <- (1x10x16x64xf32) + # pd_op.transpose: (1x16x20x64xf32) <- (1x20x16x64xf32) transpose_235 = paddle._C_ops.transpose(reshape_486, [0, 2, 1, 3]) del reshape_486 - # pd_op.matmul: (1x10x1024xf32) <- (1x10x1024xf32, 1024x1024xf32) + # pd_op.matmul: (1x20x1024xf32) <- (1x20x1024xf32, 1024x1024xf32) matmul_224 = paddle._C_ops.matmul(layer_norm_72, parameter_19, False, False) - del parameter_19 + del layer_norm_72, parameter_19 - # pd_op.add: (1x10x1024xf32) <- (1x10x1024xf32, 1024xf32) + # pd_op.add: (1x20x1024xf32) <- (1x20x1024xf32, 1024xf32) add_359 = paddle._C_ops.add(matmul_224, parameter_18) del matmul_224, parameter_18 - # pd_op.reshape: (1x10x16x64xf32) <- (1x10x1024xf32, 4xi64) + # pd_op.reshape: (1x20x16x64xf32) <- (1x20x1024xf32, 4xi64) reshape_487 = paddle._C_ops.reshape(add_359, full_int_array_2) del add_359, full_int_array_2 - # pd_op.transpose: (1x16x10x64xf32) <- (1x10x16x64xf32) + # pd_op.transpose: (1x16x20x64xf32) <- (1x20x16x64xf32) transpose_236 = paddle._C_ops.transpose(reshape_487, [0, 2, 1, 3]) del reshape_487 @@ -11502,30 +11524,31 @@ def forward( reshape_489 = paddle._C_ops.reshape(transpose_237, full_int_array_4) del transpose_237 - # pd_op.reshape: (16x10x64xf32) <- (1x16x10x64xf32, 3xi64) + # pd_op.reshape: (16x20x64xf32) <- (1x16x20x64xf32, 3xi64) reshape_490 = paddle._C_ops.reshape(transpose_235, full_int_array_4) + del transpose_235 - # pd_op.reshape: (16x10x64xf32) <- (1x16x10x64xf32, 3xi64) + # pd_op.reshape: (16x20x64xf32) <- (1x16x20x64xf32, 3xi64) reshape_491 = paddle._C_ops.reshape(transpose_236, full_int_array_4) - del full_int_array_4 + del full_int_array_4, transpose_236 - # pd_op.transpose: (16x64x10xf32) <- (16x10x64xf32) + # pd_op.transpose: (16x64x20xf32) <- (16x20x64xf32) transpose_238 = paddle._C_ops.transpose(reshape_490, [0, 2, 1]) del reshape_490 - # pd_op.bmm: (16x3x10xf32) <- (16x3x64xf32, 16x64x10xf32) + # pd_op.bmm: (16x3x20xf32) <- (16x3x64xf32, 16x64x20xf32) bmm_70 = paddle._C_ops.bmm(reshape_489, transpose_238) del reshape_489, transpose_238 - # pd_op.add: (16x3x10xf32) <- (16x3x10xf32, 16x1x10xf32) + # pd_op.add: (16x3x20xf32) <- (16x3x20xf32, 16x1x20xf32) add_360 = paddle._C_ops.add(bmm_70, cast_10) del bmm_70, cast_10 - # pd_op.softmax: (16x3x10xf32) <- (16x3x10xf32) + # pd_op.softmax: (16x3x20xf32) <- (16x3x20xf32) softmax_47 = paddle._C_ops.softmax(add_360, -1) del add_360 - # pd_op.dropout: (16x3x10xf32, 16x3x10xui8) <- (16x3x10xf32, None, 1xf32) + # pd_op.dropout: (16x3x20xf32, 16x3x20xui8) <- (16x3x20xf32, None, 1xf32) dropout_260, dropout_261 = (lambda x, f: f(x))( paddle._C_ops.dropout( softmax_47, None, full_8, True, "upscale_in_train", 0, False @@ -11534,7 +11557,7 @@ def forward( ) del softmax_47 - # pd_op.bmm: (16x3x64xf32) <- (16x3x10xf32, 16x10x64xf32) + # pd_op.bmm: (16x3x64xf32) <- (16x3x20xf32, 16x20x64xf32) bmm_71 = paddle._C_ops.bmm(dropout_260, reshape_491) del dropout_260, reshape_491 @@ -11641,36 +11664,7 @@ def forward( slice_1 = paddle._C_ops.slice( layer_norm_183, [1], full_int_array_0, full_int_array_13, [1], [] ) - del ( - full_int_array_0, - full_int_array_13, - layer_norm_183, - layer_norm_72, - transpose_100, - transpose_101, - transpose_115, - transpose_116, - transpose_130, - transpose_131, - transpose_145, - transpose_146, - transpose_160, - transpose_161, - transpose_175, - transpose_176, - transpose_190, - transpose_191, - transpose_205, - transpose_206, - transpose_220, - transpose_221, - transpose_235, - transpose_236, - transpose_70, - transpose_71, - transpose_85, - transpose_86, - ) + del full_int_array_0, full_int_array_13, layer_norm_183 return ( slice_0, diff --git a/paddle_samples/PaddleNLP/prophetnet-large-uncased/weight_meta.py b/paddle_samples/PaddleNLP/prophetnet-large-uncased/weight_meta.py index 652f18551..35defed3a 100644 --- a/paddle_samples/PaddleNLP/prophetnet-large-uncased/weight_meta.py +++ b/paddle_samples/PaddleNLP/prophetnet-large-uncased/weight_meta.py @@ -1,25 +1,26 @@ class Program_weight_tensor_parameter_0: name = "parameter_0" + original_name = "eager_tmp_3" shape = [2, 512, 1024] dtype = "float32" - min_val = float("-inf") - mean = float("-inf") - std = float("nan") + min_val = float("-0.06") + max_val = float("0.06") data = None class Program_weight_tensor_parameter_1: name = "parameter_1" + original_name = "eager_tmp_2" shape = [512, 512] dtype = "float32" - min_val = float("-inf") - mean = float("-inf") - std = float("nan") + min_val = float("-0.06") + max_val = float("0.06") data = None class Program_weight_tensor_parameter_2: name = "parameter_2" + original_name = "eager_tmp_1" shape = [1, 512, 1024] dtype = "int32" min_val = 0 @@ -29,6 +30,7 @@ class Program_weight_tensor_parameter_2: class Program_weight_tensor_parameter_3: name = "parameter_3" + original_name = "eager_tmp_0" shape = [1, 512, 512] dtype = "int32" min_val = 0 @@ -38,4769 +40,6431 @@ class Program_weight_tensor_parameter_3: class Program_weight_tensor_parameter_4: name = "parameter_4" + original_name = "layer_norm_61.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.452148") + max_val = float("0.236206") + mean = float("0.000814992") + std = float("0.103339") data = None class Program_weight_tensor_parameter_5: name = "parameter_5" + original_name = "layer_norm_61.w_0" shape = [1024] dtype = "float32" - min_val = float("1.0") - max_val = float("1.0") - mean = float("1.0") + min_val = float("0.853027") + max_val = float("1.00586") + mean = float("0.970052") + std = float("0.00928717") data = None class Program_weight_tensor_parameter_6: name = "parameter_6" + original_name = "layer_norm_60.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.358154") + max_val = float("0.38916") + mean = float("-0.0103484") + std = float("0.148209") data = None class Program_weight_tensor_parameter_7: name = "parameter_7" + original_name = "layer_norm_60.w_0" shape = [1024] dtype = "float32" - min_val = float("1.0") - max_val = float("1.0") - mean = float("1.0") + min_val = float("0.817383") + max_val = float("1.44141") + mean = float("1.24825") + std = float("0.0237268") data = None class Program_weight_tensor_parameter_8: name = "parameter_8" + original_name = "linear_203.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.083374") + max_val = float("0.101013") + mean = float("-0.000367784") + std = float("0.0323448") data = None class Program_weight_tensor_parameter_9: name = "parameter_9" + original_name = "linear_203.w_0" shape = [4096, 1024] dtype = "float32" - min_val = float("-0.106072") - max_val = float("0.101304") - mean = float("2.18394e-06") - std = float("0.0200064") + min_val = float("-1.27246") + max_val = float("0.722656") + mean = float("-7.20212e-07") + std = float("0.0252089") data = None class Program_weight_tensor_parameter_10: name = "parameter_10" + original_name = "linear_202.b_0" shape = [4096] dtype = "float32" + min_val = float("-0.231567") + max_val = float("0.17041") + mean = float("-0.0420246") + std = float("0.0617493") data = None class Program_weight_tensor_parameter_11: name = "parameter_11" + original_name = "linear_202.w_0" shape = [1024, 4096] dtype = "float32" - min_val = float("-0.0996486") - max_val = float("0.107113") - mean = float("8.23004e-06") - std = float("0.0199933") + min_val = float("-0.158569") + max_val = float("0.177124") + mean = float("0.000207353") + std = float("0.0290842") data = None class Program_weight_tensor_parameter_12: name = "parameter_12" + original_name = "layer_norm_59.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.518555") + max_val = float("0.220459") + mean = float("-0.0238952") + std = float("0.0907258") data = None class Program_weight_tensor_parameter_13: name = "parameter_13" + original_name = "layer_norm_59.w_0" shape = [1024] dtype = "float32" - min_val = float("1.0") - max_val = float("1.0") - mean = float("1.0") + min_val = float("0.941895") + max_val = float("1.30371") + mean = float("0.961005") + std = float("0.0144994") data = None class Program_weight_tensor_parameter_14: name = "parameter_14" + original_name = "linear_201.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.160767") + max_val = float("0.17688") + mean = float("-0.000136284") + std = float("0.0538224") data = None class Program_weight_tensor_parameter_15: name = "parameter_15" + original_name = "linear_201.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.0979996") - max_val = float("0.0973597") - mean = float("-1.07633e-05") - std = float("0.0200143") + min_val = float("-0.354492") + max_val = float("0.34668") + mean = float("1.53171e-05") + std = float("0.0346223") data = None class Program_weight_tensor_parameter_16: name = "parameter_16" + original_name = "linear_200.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.21582") + max_val = float("0.205322") + mean = float("-0.00214459") + std = float("0.109487") data = None class Program_weight_tensor_parameter_17: name = "parameter_17" + original_name = "linear_200.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.0945349") - max_val = float("0.10205") - mean = float("-7.23719e-06") - std = float("0.0199873") + min_val = float("-0.144409") + max_val = float("0.147217") + mean = float("0.000119139") + std = float("0.0346979") data = None class Program_weight_tensor_parameter_18: name = "parameter_18" + original_name = "linear_199.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.127563") + max_val = float("0.138306") + mean = float("0.0014313") + std = float("0.0343896") data = None class Program_weight_tensor_parameter_19: name = "parameter_19" + original_name = "linear_199.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.0994254") - max_val = float("0.0952829") - mean = float("-3.52851e-05") - std = float("0.0199933") + min_val = float("-0.157837") + max_val = float("0.194458") + mean = float("-7.6559e-05") + std = float("0.0373485") data = None class Program_weight_tensor_parameter_20: name = "parameter_20" + original_name = "linear_198.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.259033") + max_val = float("0.274414") + mean = float("-0.00787324") + std = float("0.161596") data = None class Program_weight_tensor_parameter_21: name = "parameter_21" + original_name = "linear_198.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.094652") - max_val = float("0.104842") - mean = float("4.39518e-06") - std = float("0.0200327") + min_val = float("-0.165771") + max_val = float("0.159058") + mean = float("2.62323e-05") + std = float("0.0337999") data = None class Program_weight_tensor_parameter_22: name = "parameter_22" + original_name = "layer_norm_58.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.375488") + max_val = float("0.186279") + mean = float("-0.0433858") + std = float("0.0759863") data = None class Program_weight_tensor_parameter_23: name = "parameter_23" + original_name = "layer_norm_58.w_0" shape = [1024] dtype = "float32" - min_val = float("1.0") - max_val = float("1.0") - mean = float("1.0") + min_val = float("0.905273") + max_val = float("1.02051") + mean = float("0.963117") + std = float("0.00600066") data = None class Program_weight_tensor_parameter_24: name = "parameter_24" + original_name = "linear_197.b_0" shape = [512] dtype = "float32" + min_val = float("-0.193481") + max_val = float("0.212769") + mean = float("-0.0222641") + std = float("0.0540544") data = None class Program_weight_tensor_parameter_25: name = "parameter_25" + original_name = "linear_197.w_0" shape = [1024, 512] dtype = "float32" - min_val = float("-0.091934") - max_val = float("0.105912") - mean = float("5.73166e-05") - std = float("0.019992") + min_val = float("-0.14563") + max_val = float("0.110107") + mean = float("-3.26498e-05") + std = float("0.0179843") data = None class Program_weight_tensor_parameter_26: name = "parameter_26" + original_name = "linear_196.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.0911255") + max_val = float("0.0791626") + mean = float("6.19112e-05") + std = float("0.0321936") data = None class Program_weight_tensor_parameter_27: name = "parameter_27" + original_name = "linear_196.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.0970294") - max_val = float("0.107901") - mean = float("-9.68548e-06") - std = float("0.0200058") + min_val = float("-0.256592") + max_val = float("0.262207") + mean = float("2.89491e-06") + std = float("0.0256758") data = None class Program_weight_tensor_parameter_28: name = "parameter_28" + original_name = "linear_195.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.155273") + max_val = float("0.165894") + mean = float("0.00209596") + std = float("0.0597834") data = None class Program_weight_tensor_parameter_29: name = "parameter_29" + original_name = "linear_195.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.107093") - max_val = float("0.0924737") - mean = float("-9.76095e-06") - std = float("0.0199887") + min_val = float("-0.149414") + max_val = float("0.151855") + mean = float("-2.34763e-05") + std = float("0.0361948") data = None class Program_weight_tensor_parameter_30: name = "parameter_30" + original_name = "linear_194.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.0662842") + max_val = float("0.072998") + mean = float("0.000509212") + std = float("0.0210998") data = None class Program_weight_tensor_parameter_31: name = "parameter_31" + original_name = "linear_194.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.0989044") - max_val = float("0.0983128") - mean = float("-6.70061e-06") - std = float("0.0200172") + min_val = float("-0.137207") + max_val = float("0.143433") + mean = float("3.79538e-05") + std = float("0.0287667") data = None class Program_weight_tensor_parameter_32: name = "parameter_32" + original_name = "linear_193.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.12146") + max_val = float("0.10675") + mean = float("-0.000732174") + std = float("0.0245368") data = None class Program_weight_tensor_parameter_33: name = "parameter_33" + original_name = "linear_193.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.0980593") - max_val = float("0.0977565") - mean = float("2.52491e-06") - std = float("0.0200202") + min_val = float("-0.145264") + max_val = float("0.153931") + mean = float("6.56715e-06") + std = float("0.0360211") data = None class Program_weight_tensor_parameter_34: name = "parameter_34" + original_name = "layer_norm_57.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.414551") + max_val = float("0.280273") + mean = float("-0.0147455") + std = float("0.065716") data = None class Program_weight_tensor_parameter_35: name = "parameter_35" + original_name = "layer_norm_57.w_0" shape = [1024] dtype = "float32" - min_val = float("1.0") - max_val = float("1.0") - mean = float("1.0") + min_val = float("0.627441") + max_val = float("1.11133") + mean = float("0.962764") + std = float("0.0144757") data = None class Program_weight_tensor_parameter_36: name = "parameter_36" + original_name = "linear_192.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.145996") + max_val = float("0.143555") + mean = float("0.000352888") + std = float("0.0680989") data = None class Program_weight_tensor_parameter_37: name = "parameter_37" + original_name = "linear_192.w_0" shape = [4096, 1024] dtype = "float32" - min_val = float("-0.101016") - max_val = float("0.105924") - mean = float("3.44544e-06") - std = float("0.0200024") + min_val = float("-1.4668") + max_val = float("0.720703") + mean = float("-4.75082e-06") + std = float("0.0297387") data = None class Program_weight_tensor_parameter_38: name = "parameter_38" + original_name = "linear_191.b_0" shape = [4096] dtype = "float32" + min_val = float("-0.206421") + max_val = float("0.166992") + mean = float("-0.0285632") + std = float("0.0322798") data = None class Program_weight_tensor_parameter_39: name = "parameter_39" + original_name = "linear_191.w_0" shape = [1024, 4096] dtype = "float32" - min_val = float("-0.101964") - max_val = float("0.107844") - mean = float("-3.41908e-06") - std = float("0.0200059") + min_val = float("-0.481689") + max_val = float("0.411377") + mean = float("0.000168527") + std = float("0.0309878") data = None class Program_weight_tensor_parameter_40: name = "parameter_40" + original_name = "layer_norm_56.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.569336") + max_val = float("0.292725") + mean = float("-0.00974464") + std = float("0.069752") data = None class Program_weight_tensor_parameter_41: name = "parameter_41" + original_name = "layer_norm_56.w_0" shape = [1024] dtype = "float32" - min_val = float("1.0") - max_val = float("1.0") - mean = float("1.0") + min_val = float("0.951172") + max_val = float("1.19922") + mean = float("0.96072") + std = float("0.012215") data = None class Program_weight_tensor_parameter_42: name = "parameter_42" + original_name = "linear_190.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.181519") + max_val = float("0.11377") + mean = float("-0.000114971") + std = float("0.0430907") data = None class Program_weight_tensor_parameter_43: name = "parameter_43" + original_name = "linear_190.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.0984877") - max_val = float("0.104757") - mean = float("3.25959e-05") - std = float("0.0200231") + min_val = float("-0.372314") + max_val = float("0.382568") + mean = float("-2.75028e-06") + std = float("0.0346781") data = None class Program_weight_tensor_parameter_44: name = "parameter_44" + original_name = "linear_189.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.178833") + max_val = float("0.210693") + mean = float("-0.00213549") + std = float("0.0841748") data = None class Program_weight_tensor_parameter_45: name = "parameter_45" + original_name = "linear_189.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.0959707") - max_val = float("0.0994419") - mean = float("1.21262e-05") - std = float("0.0199948") + min_val = float("-0.169434") + max_val = float("0.161865") + mean = float("3.2143e-05") + std = float("0.035845") data = None class Program_weight_tensor_parameter_46: name = "parameter_46" + original_name = "linear_188.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.143555") + max_val = float("0.0921021") + mean = float("-0.000680979") + std = float("0.0297051") data = None class Program_weight_tensor_parameter_47: name = "parameter_47" + original_name = "linear_188.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.0994287") - max_val = float("0.0958598") - mean = float("2.86432e-05") - std = float("0.0199895") + min_val = float("-0.169434") + max_val = float("0.166138") + mean = float("-1.99112e-06") + std = float("0.0372977") data = None class Program_weight_tensor_parameter_48: name = "parameter_48" + original_name = "linear_187.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.287354") + max_val = float("0.311279") + mean = float("-0.00105768") + std = float("0.114683") data = None class Program_weight_tensor_parameter_49: name = "parameter_49" + original_name = "linear_187.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.0991599") - max_val = float("0.0980661") - mean = float("1.7976e-05") - std = float("0.0200211") + min_val = float("-0.200195") + max_val = float("0.22168") + mean = float("2.08729e-05") + std = float("0.0351082") data = None class Program_weight_tensor_parameter_50: name = "parameter_50" + original_name = "layer_norm_55.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.706055") + max_val = float("0.266846") + mean = float("-0.00908414") + std = float("0.0627384") data = None class Program_weight_tensor_parameter_51: name = "parameter_51" + original_name = "layer_norm_55.w_0" shape = [1024] dtype = "float32" - min_val = float("1.0") - max_val = float("1.0") - mean = float("1.0") + min_val = float("0.889648") + max_val = float("1.11328") + mean = float("0.969057") + std = float("0.00649421") data = None class Program_weight_tensor_parameter_52: name = "parameter_52" + original_name = "linear_186.b_0" shape = [512] dtype = "float32" + min_val = float("-0.180664") + max_val = float("0.151367") + mean = float("-0.00755014") + std = float("0.0427569") data = None class Program_weight_tensor_parameter_53: name = "parameter_53" + original_name = "linear_186.w_0" shape = [1024, 512] dtype = "float32" - min_val = float("-0.0994995") - max_val = float("0.0933616") - mean = float("-1.87231e-05") - std = float("0.0200211") + min_val = float("-0.124451") + max_val = float("0.14502") + mean = float("-0.000264523") + std = float("0.0194731") data = None class Program_weight_tensor_parameter_54: name = "parameter_54" + original_name = "linear_185.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.0805664") + max_val = float("0.0880127") + mean = float("5.25069e-05") + std = float("0.0372208") data = None class Program_weight_tensor_parameter_55: name = "parameter_55" + original_name = "linear_185.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.102435") - max_val = float("0.0976003") - mean = float("-2.92068e-05") - std = float("0.0199896") + min_val = float("-0.340576") + max_val = float("0.346924") + mean = float("4.05886e-06") + std = float("0.0256414") data = None class Program_weight_tensor_parameter_56: name = "parameter_56" + original_name = "linear_184.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.152954") + max_val = float("0.165894") + mean = float("0.000853512") + std = float("0.0610069") data = None class Program_weight_tensor_parameter_57: name = "parameter_57" + original_name = "linear_184.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.0953802") - max_val = float("0.0960691") - mean = float("-4.18383e-05") - std = float("0.0199888") + min_val = float("-0.149902") + max_val = float("0.14502") + mean = float("-3.08742e-06") + std = float("0.03508") data = None class Program_weight_tensor_parameter_58: name = "parameter_58" + original_name = "linear_183.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.067688") + max_val = float("0.0870972") + mean = float("8.50256e-05") + std = float("0.0258201") data = None class Program_weight_tensor_parameter_59: name = "parameter_59" + original_name = "linear_183.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.0964754") - max_val = float("0.0889254") - mean = float("-3.38984e-05") - std = float("0.0200151") + min_val = float("-0.145874") + max_val = float("0.1521") + mean = float("5.52254e-06") + std = float("0.0289329") data = None class Program_weight_tensor_parameter_60: name = "parameter_60" + original_name = "linear_182.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.155762") + max_val = float("0.137085") + mean = float("0.000497557") + std = float("0.025787") data = None class Program_weight_tensor_parameter_61: name = "parameter_61" + original_name = "linear_182.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.101191") - max_val = float("0.102416") - mean = float("8.40873e-06") - std = float("0.0199744") + min_val = float("-0.15332") + max_val = float("0.138794") + mean = float("-2.64437e-07") + std = float("0.0345397") data = None class Program_weight_tensor_parameter_62: name = "parameter_62" + original_name = "layer_norm_54.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.529785") + max_val = float("0.138184") + mean = float("-0.0242569") + std = float("0.0448547") data = None class Program_weight_tensor_parameter_63: name = "parameter_63" + original_name = "layer_norm_54.w_0" shape = [1024] dtype = "float32" - min_val = float("1.0") - max_val = float("1.0") - mean = float("1.0") + min_val = float("0.864258") + max_val = float("1.1377") + mean = float("0.966406") + std = float("0.0078016") data = None class Program_weight_tensor_parameter_64: name = "parameter_64" + original_name = "linear_181.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.187622") + max_val = float("0.144287") + mean = float("0.000594991") + std = float("0.0611641") data = None class Program_weight_tensor_parameter_65: name = "parameter_65" + original_name = "linear_181.w_0" shape = [4096, 1024] dtype = "float32" - min_val = float("-0.09745") - max_val = float("0.0971174") - mean = float("2.00812e-06") - std = float("0.0199952") + min_val = float("-1.25684") + max_val = float("0.603027") + mean = float("-6.09526e-06") + std = float("0.0322566") data = None class Program_weight_tensor_parameter_66: name = "parameter_66" + original_name = "linear_180.b_0" shape = [4096] dtype = "float32" + min_val = float("-0.132324") + max_val = float("0.162476") + mean = float("-0.0308557") + std = float("0.0253492") data = None class Program_weight_tensor_parameter_67: name = "parameter_67" + original_name = "linear_180.w_0" shape = [1024, 4096] dtype = "float32" - min_val = float("-0.0985694") - max_val = float("0.105974") - mean = float("-9.26347e-06") - std = float("0.0199926") + min_val = float("-0.388672") + max_val = float("0.597656") + mean = float("0.000202614") + std = float("0.0338258") data = None class Program_weight_tensor_parameter_68: name = "parameter_68" + original_name = "layer_norm_53.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.555664") + max_val = float("0.289795") + mean = float("-0.0080598") + std = float("0.0748146") data = None class Program_weight_tensor_parameter_69: name = "parameter_69" + original_name = "layer_norm_53.w_0" shape = [1024] dtype = "float32" - min_val = float("1.0") - max_val = float("1.0") - mean = float("1.0") + min_val = float("0.944336") + max_val = float("1.16309") + mean = float("0.955787") + std = float("0.00961034") data = None class Program_weight_tensor_parameter_70: name = "parameter_70" + original_name = "linear_179.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.194092") + max_val = float("0.139282") + mean = float("-0.000102304") + std = float("0.0507224") data = None class Program_weight_tensor_parameter_71: name = "parameter_71" + original_name = "linear_179.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.0996717") - max_val = float("0.098865") - mean = float("1.52941e-05") - std = float("0.0199961") + min_val = float("-0.378906") + max_val = float("0.431641") + mean = float("8.66301e-06") + std = float("0.0309595") data = None class Program_weight_tensor_parameter_72: name = "parameter_72" + original_name = "linear_178.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.17395") + max_val = float("0.177002") + mean = float("0.00147189") + std = float("0.0558104") data = None class Program_weight_tensor_parameter_73: name = "parameter_73" + original_name = "linear_178.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.100568") - max_val = float("0.0962315") - mean = float("-1.95311e-06") - std = float("0.019972") + min_val = float("-0.168823") + max_val = float("0.161743") + mean = float("-1.61008e-06") + std = float("0.0374168") data = None class Program_weight_tensor_parameter_74: name = "parameter_74" + original_name = "linear_177.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.191406") + max_val = float("0.159424") + mean = float("-0.00190845") + std = float("0.0390091") data = None class Program_weight_tensor_parameter_75: name = "parameter_75" + original_name = "linear_177.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.0946121") - max_val = float("0.101219") - mean = float("-3.34689e-05") - std = float("0.0200091") + min_val = float("-0.154907") + max_val = float("0.151367") + mean = float("9.02231e-05") + std = float("0.0338327") data = None class Program_weight_tensor_parameter_76: name = "parameter_76" + original_name = "linear_176.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.267822") + max_val = float("0.242798") + mean = float("0.00277793") + std = float("0.0635837") data = None class Program_weight_tensor_parameter_77: name = "parameter_77" + original_name = "linear_176.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.0978156") - max_val = float("0.098342") - mean = float("3.59747e-05") - std = float("0.0200065") + min_val = float("-0.18457") + max_val = float("0.19812") + mean = float("1.35952e-05") + std = float("0.0370587") data = None class Program_weight_tensor_parameter_78: name = "parameter_78" + original_name = "layer_norm_52.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.881836") + max_val = float("0.260498") + mean = float("-0.00350322") + std = float("0.0612452") data = None class Program_weight_tensor_parameter_79: name = "parameter_79" + original_name = "layer_norm_52.w_0" shape = [1024] dtype = "float32" - min_val = float("1.0") - max_val = float("1.0") - mean = float("1.0") + min_val = float("0.961426") + max_val = float("1.06934") + mean = float("0.97059") + std = float("0.0051963") data = None class Program_weight_tensor_parameter_80: name = "parameter_80" + original_name = "linear_175.b_0" shape = [512] dtype = "float32" + min_val = float("-0.159424") + max_val = float("0.155029") + mean = float("-0.00313954") + std = float("0.0403184") data = None class Program_weight_tensor_parameter_81: name = "parameter_81" + original_name = "linear_175.w_0" shape = [1024, 512] dtype = "float32" - min_val = float("-0.0909577") - max_val = float("0.105643") - mean = float("-3.24223e-05") - std = float("0.0199898") + min_val = float("-0.150757") + max_val = float("0.179565") + mean = float("-0.000184175") + std = float("0.0191668") data = None class Program_weight_tensor_parameter_82: name = "parameter_82" + original_name = "linear_174.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.118896") + max_val = float("0.0726929") + mean = float("3.34025e-05") + std = float("0.0306484") data = None class Program_weight_tensor_parameter_83: name = "parameter_83" + original_name = "linear_174.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.0950324") - max_val = float("0.0925353") - mean = float("-2.63852e-06") - std = float("0.0200208") + min_val = float("-0.358398") + max_val = float("0.391113") + mean = float("7.80795e-06") + std = float("0.0278148") data = None class Program_weight_tensor_parameter_84: name = "parameter_84" + original_name = "linear_173.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.198364") + max_val = float("0.204956") + mean = float("-0.00156342") + std = float("0.0618878") data = None class Program_weight_tensor_parameter_85: name = "parameter_85" + original_name = "linear_173.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.0963575") - max_val = float("0.0985771") - mean = float("-2.63966e-06") - std = float("0.0199886") + min_val = float("-0.153809") + max_val = float("0.147949") + mean = float("2.39708e-05") + std = float("0.0341002") data = None class Program_weight_tensor_parameter_86: name = "parameter_86" + original_name = "linear_172.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.0991211") + max_val = float("0.0668945") + mean = float("-0.000914387") + std = float("0.0195478") data = None class Program_weight_tensor_parameter_87: name = "parameter_87" + original_name = "linear_172.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.0945037") - max_val = float("0.0961924") - mean = float("1.87324e-05") - std = float("0.0199879") + min_val = float("-0.156372") + max_val = float("0.154297") + mean = float("0.000119402") + std = float("0.0321379") data = None class Program_weight_tensor_parameter_88: name = "parameter_88" + original_name = "linear_171.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.105835") + max_val = float("0.0956421") + mean = float("-0.000832019") + std = float("0.0124758") data = None class Program_weight_tensor_parameter_89: name = "parameter_89" + original_name = "linear_171.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.0953841") - max_val = float("0.0933501") - mean = float("-7.68459e-06") - std = float("0.0199911") + min_val = float("-0.143188") + max_val = float("0.152954") + mean = float("-2.57198e-05") + std = float("0.0338278") data = None class Program_weight_tensor_parameter_90: name = "parameter_90" + original_name = "layer_norm_51.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.460938") + max_val = float("0.146118") + mean = float("-0.0267166") + std = float("0.0370838") data = None class Program_weight_tensor_parameter_91: name = "parameter_91" + original_name = "layer_norm_51.w_0" shape = [1024] dtype = "float32" - min_val = float("1.0") - max_val = float("1.0") - mean = float("1.0") + min_val = float("0.81543") + max_val = float("1.06934") + mean = float("0.965765") + std = float("0.00759213") data = None class Program_weight_tensor_parameter_92: name = "parameter_92" + original_name = "linear_170.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.206665") + max_val = float("0.148682") + mean = float("0.000572778") + std = float("0.0661158") data = None class Program_weight_tensor_parameter_93: name = "parameter_93" + original_name = "linear_170.w_0" shape = [4096, 1024] dtype = "float32" - min_val = float("-0.0984076") - max_val = float("0.104354") - mean = float("-3.47192e-06") - std = float("0.0199897") + min_val = float("-1.35938") + max_val = float("0.608398") + mean = float("-7.68569e-06") + std = float("0.0317401") data = None class Program_weight_tensor_parameter_94: name = "parameter_94" + original_name = "linear_169.b_0" shape = [4096] dtype = "float32" + min_val = float("-0.131348") + max_val = float("0.174683") + mean = float("-0.0282512") + std = float("0.0242934") data = None class Program_weight_tensor_parameter_95: name = "parameter_95" + original_name = "linear_169.w_0" shape = [1024, 4096] dtype = "float32" - min_val = float("-0.107353") - max_val = float("0.102096") - mean = float("-1.11733e-05") - std = float("0.0200026") + min_val = float("-0.406006") + max_val = float("0.394531") + mean = float("0.000225634") + std = float("0.0339562") data = None class Program_weight_tensor_parameter_96: name = "parameter_96" + original_name = "layer_norm_50.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.601074") + max_val = float("0.244019") + mean = float("-0.00798247") + std = float("0.0805236") data = None class Program_weight_tensor_parameter_97: name = "parameter_97" + original_name = "layer_norm_50.w_0" shape = [1024] dtype = "float32" - min_val = float("1.0") - max_val = float("1.0") - mean = float("1.0") + min_val = float("0.949219") + max_val = float("1.15039") + mean = float("0.957792") + std = float("0.0077259") data = None class Program_weight_tensor_parameter_98: name = "parameter_98" + original_name = "linear_168.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.188354") + max_val = float("0.122681") + mean = float("-0.000121107") + std = float("0.0476223") data = None class Program_weight_tensor_parameter_99: name = "parameter_99" + original_name = "linear_168.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.0939327") - max_val = float("0.0954823") - mean = float("2.40252e-05") - std = float("0.0199775") + min_val = float("-0.456055") + max_val = float("0.455078") + mean = float("-1.28983e-05") + std = float("0.0300814") data = None class Program_weight_tensor_parameter_100: name = "parameter_100" + original_name = "linear_167.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.221924") + max_val = float("0.241089") + mean = float("0.00108492") + std = float("0.0686862") data = None class Program_weight_tensor_parameter_101: name = "parameter_101" + original_name = "linear_167.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.0986293") - max_val = float("0.0968711") - mean = float("2.27255e-05") - std = float("0.0199851") + min_val = float("-0.165039") + max_val = float("0.144897") + mean = float("-4.00397e-05") + std = float("0.0355893") data = None class Program_weight_tensor_parameter_102: name = "parameter_102" + original_name = "linear_166.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.147461") + max_val = float("0.137207") + mean = float("0.000434436") + std = float("0.0387579") data = None class Program_weight_tensor_parameter_103: name = "parameter_103" + original_name = "linear_166.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.102733") - max_val = float("0.0890922") - mean = float("2.51087e-06") - std = float("0.0199766") + min_val = float("-0.154053") + max_val = float("0.166016") + mean = float("-1.76975e-05") + std = float("0.0327625") data = None class Program_weight_tensor_parameter_104: name = "parameter_104" + original_name = "linear_165.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.276855") + max_val = float("0.26416") + mean = float("0.00428854") + std = float("0.0685763") data = None class Program_weight_tensor_parameter_105: name = "parameter_105" + original_name = "linear_165.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.0965806") - max_val = float("0.104679") - mean = float("-3.11027e-05") - std = float("0.0200046") + min_val = float("-0.239624") + max_val = float("0.264893") + mean = float("-1.36181e-05") + std = float("0.0351549") data = None class Program_weight_tensor_parameter_106: name = "parameter_106" + original_name = "layer_norm_49.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.760742") + max_val = float("0.185303") + mean = float("-0.0163815") + std = float("0.0523827") data = None class Program_weight_tensor_parameter_107: name = "parameter_107" + original_name = "layer_norm_49.w_0" shape = [1024] dtype = "float32" - min_val = float("1.0") - max_val = float("1.0") - mean = float("1.0") + min_val = float("0.931641") + max_val = float("1.08496") + mean = float("0.971287") + std = float("0.00509524") data = None class Program_weight_tensor_parameter_108: name = "parameter_108" + original_name = "linear_164.b_0" shape = [512] dtype = "float32" + min_val = float("-0.174927") + max_val = float("0.185303") + mean = float("-0.00891726") + std = float("0.0478504") data = None class Program_weight_tensor_parameter_109: name = "parameter_109" + original_name = "linear_164.w_0" shape = [1024, 512] dtype = "float32" - min_val = float("-0.0924062") - max_val = float("0.0954859") - mean = float("-9.49412e-06") - std = float("0.0199823") + min_val = float("-0.141479") + max_val = float("0.143188") + mean = float("-0.000134034") + std = float("0.0201543") data = None class Program_weight_tensor_parameter_110: name = "parameter_110" + original_name = "linear_163.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.131714") + max_val = float("0.0727539") + mean = float("2.05343e-05") + std = float("0.0291319") data = None class Program_weight_tensor_parameter_111: name = "parameter_111" + original_name = "linear_163.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.105761") - max_val = float("0.0939133") - mean = float("-3.26191e-05") - std = float("0.0200009") + min_val = float("-0.313721") + max_val = float("0.275146") + mean = float("-1.14887e-05") + std = float("0.0256521") data = None class Program_weight_tensor_parameter_112: name = "parameter_112" + original_name = "linear_162.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.171631") + max_val = float("0.152954") + mean = float("0.00198737") + std = float("0.0496255") data = None class Program_weight_tensor_parameter_113: name = "parameter_113" + original_name = "linear_162.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.0926243") - max_val = float("0.092981") - mean = float("3.16142e-05") - std = float("0.0199863") + min_val = float("-0.144287") + max_val = float("0.142212") + mean = float("-6.92853e-05") + std = float("0.0329723") data = None class Program_weight_tensor_parameter_114: name = "parameter_114" + original_name = "linear_161.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.0858765") + max_val = float("0.0872803") + mean = float("0.00116466") + std = float("0.0203333") data = None class Program_weight_tensor_parameter_115: name = "parameter_115" + original_name = "linear_161.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.0986968") - max_val = float("0.0952158") - mean = float("3.60616e-05") - std = float("0.0200047") + min_val = float("-0.130615") + max_val = float("0.126587") + mean = float("-3.38525e-05") + std = float("0.0289626") data = None class Program_weight_tensor_parameter_116: name = "parameter_116" + original_name = "linear_160.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.0911255") + max_val = float("0.0819092") + mean = float("-0.000309498") + std = float("0.00906396") data = None class Program_weight_tensor_parameter_117: name = "parameter_117" + original_name = "linear_160.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.0984313") - max_val = float("0.0955871") - mean = float("-1.58957e-05") - std = float("0.0200101") + min_val = float("-0.151367") + max_val = float("0.152222") + mean = float("-2.87351e-05") + std = float("0.0325833") data = None class Program_weight_tensor_parameter_118: name = "parameter_118" + original_name = "layer_norm_48.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.388672") + max_val = float("0.0941772") + mean = float("-0.0264367") + std = float("0.0461682") data = None class Program_weight_tensor_parameter_119: name = "parameter_119" + original_name = "layer_norm_48.w_0" shape = [1024] dtype = "float32" - min_val = float("1.0") - max_val = float("1.0") - mean = float("1.0") + min_val = float("0.830078") + max_val = float("1.01562") + mean = float("0.963964") + std = float("0.00555931") data = None class Program_weight_tensor_parameter_120: name = "parameter_120" + original_name = "linear_159.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.221924") + max_val = float("0.146729") + mean = float("0.000569073") + std = float("0.0631338") data = None class Program_weight_tensor_parameter_121: name = "parameter_121" + original_name = "linear_159.w_0" shape = [4096, 1024] dtype = "float32" - min_val = float("-0.0994665") - max_val = float("0.0968224") - mean = float("5.21698e-06") - std = float("0.0199962") + min_val = float("-1.24512") + max_val = float("0.429199") + mean = float("-7.60458e-06") + std = float("0.0321452") data = None class Program_weight_tensor_parameter_122: name = "parameter_122" + original_name = "linear_158.b_0" shape = [4096] dtype = "float32" + min_val = float("-0.111511") + max_val = float("0.127563") + mean = float("-0.0287542") + std = float("0.0224263") data = None class Program_weight_tensor_parameter_123: name = "parameter_123" + original_name = "linear_158.w_0" shape = [1024, 4096] dtype = "float32" - min_val = float("-0.105515") - max_val = float("0.100532") - mean = float("-4.34489e-06") - std = float("0.0200139") + min_val = float("-0.279297") + max_val = float("0.308594") + mean = float("0.000188713") + std = float("0.0340141") data = None class Program_weight_tensor_parameter_124: name = "parameter_124" + original_name = "layer_norm_47.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.543457") + max_val = float("0.248291") + mean = float("-0.00802594") + std = float("0.0804152") data = None class Program_weight_tensor_parameter_125: name = "parameter_125" + original_name = "layer_norm_47.w_0" shape = [1024] dtype = "float32" - min_val = float("1.0") - max_val = float("1.0") - mean = float("1.0") + min_val = float("0.946289") + max_val = float("1.19141") + mean = float("0.95713") + std = float("0.00936032") data = None class Program_weight_tensor_parameter_126: name = "parameter_126" + original_name = "linear_157.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.203369") + max_val = float("0.0757446") + mean = float("0.000161026") + std = float("0.0337974") data = None class Program_weight_tensor_parameter_127: name = "parameter_127" + original_name = "linear_157.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.0912865") - max_val = float("0.097693") - mean = float("-2.40648e-05") - std = float("0.0199969") + min_val = float("-0.375244") + max_val = float("0.369873") + mean = float("-5.38651e-06") + std = float("0.0332589") data = None class Program_weight_tensor_parameter_128: name = "parameter_128" + original_name = "linear_156.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.243652") + max_val = float("0.274658") + mean = float("-0.00101064") + std = float("0.060821") data = None class Program_weight_tensor_parameter_129: name = "parameter_129" + original_name = "linear_156.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.0987604") - max_val = float("0.10159") - mean = float("-5.13199e-06") - std = float("0.0199903") + min_val = float("-0.149292") + max_val = float("0.1604") + mean = float("-1.68702e-05") + std = float("0.0348152") data = None class Program_weight_tensor_parameter_130: name = "parameter_130" + original_name = "linear_155.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.124756") + max_val = float("0.103149") + mean = float("-0.000962713") + std = float("0.030538") data = None class Program_weight_tensor_parameter_131: name = "parameter_131" + original_name = "linear_155.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.0993724") - max_val = float("0.0908072") - mean = float("9.18208e-06") - std = float("0.0199948") + min_val = float("-0.150513") + max_val = float("0.154175") + mean = float("2.11838e-05") + std = float("0.036386") data = None class Program_weight_tensor_parameter_132: name = "parameter_132" + original_name = "linear_154.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.266846") + max_val = float("0.254639") + mean = float("-5.10952e-05") + std = float("0.0602811") data = None class Program_weight_tensor_parameter_133: name = "parameter_133" + original_name = "linear_154.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.0932493") - max_val = float("0.0920728") - mean = float("1.79921e-09") - std = float("0.0199774") + min_val = float("-0.20752") + max_val = float("0.208618") + mean = float("3.2997e-05") + std = float("0.034413") data = None class Program_weight_tensor_parameter_134: name = "parameter_134" + original_name = "layer_norm_46.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.757324") + max_val = float("0.196167") + mean = float("-0.00779874") + std = float("0.0556228") data = None class Program_weight_tensor_parameter_135: name = "parameter_135" + original_name = "layer_norm_46.w_0" shape = [1024] dtype = "float32" - min_val = float("1.0") - max_val = float("1.0") - mean = float("1.0") + min_val = float("0.925781") + max_val = float("1.06055") + mean = float("0.96836") + std = float("0.00463433") data = None class Program_weight_tensor_parameter_136: name = "parameter_136" + original_name = "linear_153.b_0" shape = [512] dtype = "float32" + min_val = float("-0.16626") + max_val = float("0.179688") + mean = float("-0.00614995") + std = float("0.0444669") data = None class Program_weight_tensor_parameter_137: name = "parameter_137" + original_name = "linear_153.w_0" shape = [1024, 512] dtype = "float32" - min_val = float("-0.0932386") - max_val = float("0.0868364") - mean = float("9.6794e-06") - std = float("0.0200106") + min_val = float("-0.144043") + max_val = float("0.146973") + mean = float("-3.84003e-05") + std = float("0.0194451") data = None class Program_weight_tensor_parameter_138: name = "parameter_138" + original_name = "linear_152.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.134521") + max_val = float("0.0673218") + mean = float("-0.000204232") + std = float("0.02609") data = None class Program_weight_tensor_parameter_139: name = "parameter_139" + original_name = "linear_152.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.105008") - max_val = float("0.105821") - mean = float("-3.18322e-06") - std = float("0.0199929") + min_val = float("-0.341797") + max_val = float("0.356445") + mean = float("-3.92821e-07") + std = float("0.0279225") data = None class Program_weight_tensor_parameter_140: name = "parameter_140" + original_name = "linear_151.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.184448") + max_val = float("0.17627") + mean = float("0.000799665") + std = float("0.0539382") data = None class Program_weight_tensor_parameter_141: name = "parameter_141" + original_name = "linear_151.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.10429") - max_val = float("0.0932657") - mean = float("-2.46593e-05") - std = float("0.0200134") + min_val = float("-0.145386") + max_val = float("0.138062") + mean = float("4.38639e-05") + std = float("0.0332583") data = None class Program_weight_tensor_parameter_142: name = "parameter_142" + original_name = "linear_150.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.074585") + max_val = float("0.0632935") + mean = float("0.000558878") + std = float("0.016439") data = None class Program_weight_tensor_parameter_143: name = "parameter_143" + original_name = "linear_150.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.0984979") - max_val = float("0.0986825") - mean = float("1.82065e-05") - std = float("0.0199669") + min_val = float("-0.144165") + max_val = float("0.137451") + mean = float("-9.06656e-05") + std = float("0.0313416") data = None class Program_weight_tensor_parameter_144: name = "parameter_144" + original_name = "linear_149.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.0557556") + max_val = float("0.0479126") + mean = float("-0.000258846") + std = float("0.00578725") data = None class Program_weight_tensor_parameter_145: name = "parameter_145" + original_name = "linear_149.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.0894438") - max_val = float("0.0984713") - mean = float("9.38225e-06") - std = float("0.020014") + min_val = float("-0.162842") + max_val = float("0.160767") + mean = float("-3.52168e-06") + std = float("0.0332435") data = None class Program_weight_tensor_parameter_146: name = "parameter_146" + original_name = "layer_norm_45.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.425537") + max_val = float("0.113892") + mean = float("-0.0255635") + std = float("0.0426301") data = None class Program_weight_tensor_parameter_147: name = "parameter_147" + original_name = "layer_norm_45.w_0" shape = [1024] dtype = "float32" - min_val = float("1.0") - max_val = float("1.0") - mean = float("1.0") + min_val = float("0.830078") + max_val = float("0.995117") + mean = float("0.967191") + std = float("0.00784191") data = None class Program_weight_tensor_parameter_148: name = "parameter_148" + original_name = "linear_148.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.214111") + max_val = float("0.150513") + mean = float("0.000269823") + std = float("0.0645") data = None class Program_weight_tensor_parameter_149: name = "parameter_149" + original_name = "linear_148.w_0" shape = [4096, 1024] dtype = "float32" - min_val = float("-0.101356") - max_val = float("0.0993031") - mean = float("-5.21466e-06") - std = float("0.0199991") + min_val = float("-1.17871") + max_val = float("0.383057") + mean = float("-1.25885e-05") + std = float("0.0328996") data = None class Program_weight_tensor_parameter_150: name = "parameter_150" + original_name = "linear_147.b_0" shape = [4096] dtype = "float32" + min_val = float("-0.139771") + max_val = float("0.134521") + mean = float("-0.0287656") + std = float("0.021017") data = None class Program_weight_tensor_parameter_151: name = "parameter_151" + original_name = "linear_147.w_0" shape = [1024, 4096] dtype = "float32" - min_val = float("-0.104468") - max_val = float("0.0981822") - mean = float("-1.13551e-05") - std = float("0.0199927") + min_val = float("-0.308838") + max_val = float("0.320068") + mean = float("0.000273633") + std = float("0.0345419") data = None class Program_weight_tensor_parameter_152: name = "parameter_152" + original_name = "layer_norm_44.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.516602") + max_val = float("0.217651") + mean = float("-0.00908048") + std = float("0.0833479") data = None class Program_weight_tensor_parameter_153: name = "parameter_153" + original_name = "layer_norm_44.w_0" shape = [1024] dtype = "float32" - min_val = float("1.0") - max_val = float("1.0") - mean = float("1.0") + min_val = float("0.946289") + max_val = float("1.10547") + mean = float("0.955975") + std = float("0.00745813") data = None class Program_weight_tensor_parameter_154: name = "parameter_154" + original_name = "linear_146.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.199707") + max_val = float("0.0733032") + mean = float("-6.55206e-05") + std = float("0.0319126") data = None class Program_weight_tensor_parameter_155: name = "parameter_155" + original_name = "linear_146.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.0930349") - max_val = float("0.0936954") - mean = float("-1.07343e-05") - std = float("0.0199803") + min_val = float("-0.366211") + max_val = float("0.337891") + mean = float("7.3169e-06") + std = float("0.035957") data = None class Program_weight_tensor_parameter_156: name = "parameter_156" + original_name = "linear_145.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.191895") + max_val = float("0.215454") + mean = float("0.00193751") + std = float("0.0660667") data = None class Program_weight_tensor_parameter_157: name = "parameter_157" + original_name = "linear_145.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.0973248") - max_val = float("0.0917487") - mean = float("-2.55272e-05") - std = float("0.0200155") + min_val = float("-0.182861") + max_val = float("0.146118") + mean = float("3.58973e-05") + std = float("0.0347333") data = None class Program_weight_tensor_parameter_158: name = "parameter_158" + original_name = "linear_144.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.117615") + max_val = float("0.10791") + mean = float("-8.55081e-05") + std = float("0.0239493") data = None class Program_weight_tensor_parameter_159: name = "parameter_159" + original_name = "linear_144.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.0942376") - max_val = float("0.0934242") - mean = float("5.75674e-05") - std = float("0.0199938") + min_val = float("-0.161133") + max_val = float("0.154541") + mean = float("2.28046e-07") + std = float("0.0389436") data = None class Program_weight_tensor_parameter_160: name = "parameter_160" + original_name = "linear_143.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.270996") + max_val = float("0.282959") + mean = float("-0.000525738") + std = float("0.0458231") data = None class Program_weight_tensor_parameter_161: name = "parameter_161" + original_name = "linear_143.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.0930816") - max_val = float("0.0968099") - mean = float("-3.57151e-06") - std = float("0.0199964") + min_val = float("-0.259277") + max_val = float("0.237549") + mean = float("4.58715e-06") + std = float("0.0342737") data = None class Program_weight_tensor_parameter_162: name = "parameter_162" + original_name = "layer_norm_43.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.643066") + max_val = float("0.172485") + mean = float("-0.00575765") + std = float("0.059832") data = None class Program_weight_tensor_parameter_163: name = "parameter_163" + original_name = "layer_norm_43.w_0" shape = [1024] dtype = "float32" - min_val = float("1.0") - max_val = float("1.0") - mean = float("1.0") + min_val = float("0.944824") + max_val = float("1.02246") + mean = float("0.964204") + std = float("0.00400266") data = None class Program_weight_tensor_parameter_164: name = "parameter_164" + original_name = "linear_142.b_0" shape = [512] dtype = "float32" + min_val = float("-0.134521") + max_val = float("0.184082") + mean = float("-0.00922243") + std = float("0.0428727") data = None class Program_weight_tensor_parameter_165: name = "parameter_165" + original_name = "linear_142.w_0" shape = [1024, 512] dtype = "float32" - min_val = float("-0.0988311") - max_val = float("0.0890383") - mean = float("-2.70537e-05") - std = float("0.0199875") + min_val = float("-0.179443") + max_val = float("0.14624") + mean = float("-5.69185e-05") + std = float("0.0204847") data = None class Program_weight_tensor_parameter_166: name = "parameter_166" + original_name = "linear_141.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.191895") + max_val = float("0.0673828") + mean = float("-7.08969e-05") + std = float("0.0237132") data = None class Program_weight_tensor_parameter_167: name = "parameter_167" + original_name = "linear_141.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.0953095") - max_val = float("0.0977238") - mean = float("-1.25859e-05") - std = float("0.02001") + min_val = float("-0.303223") + max_val = float("0.304688") + mean = float("-1.55993e-07") + std = float("0.0257737") data = None class Program_weight_tensor_parameter_168: name = "parameter_168" + original_name = "linear_140.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.153809") + max_val = float("0.14917") + mean = float("-0.000686316") + std = float("0.0459238") data = None class Program_weight_tensor_parameter_169: name = "parameter_169" + original_name = "linear_140.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.0951552") - max_val = float("0.102059") - mean = float("-1.44694e-05") - std = float("0.0200185") + min_val = float("-0.136597") + max_val = float("0.141357") + mean = float("3.14721e-05") + std = float("0.0335683") data = None class Program_weight_tensor_parameter_170: name = "parameter_170" + original_name = "linear_139.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.0888062") + max_val = float("0.0791016") + mean = float("0.00121661") + std = float("0.0267297") data = None class Program_weight_tensor_parameter_171: name = "parameter_171" + original_name = "linear_139.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.0964233") - max_val = float("0.0933983") - mean = float("1.36442e-05") - std = float("0.0199999") + min_val = float("-0.123291") + max_val = float("0.130005") + mean = float("-3.93562e-05") + std = float("0.0282541") data = None class Program_weight_tensor_parameter_172: name = "parameter_172" + original_name = "linear_138.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.0317688") + max_val = float("0.030777") + mean = float("-9.28889e-05") + std = float("0.00367174") data = None class Program_weight_tensor_parameter_173: name = "parameter_173" + original_name = "linear_138.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.0949449") - max_val = float("0.0975806") - mean = float("-2.29517e-05") - std = float("0.019993") + min_val = float("-0.138672") + max_val = float("0.137939") + mean = float("-1.70677e-05") + std = float("0.0333659") data = None class Program_weight_tensor_parameter_174: name = "parameter_174" + original_name = "layer_norm_42.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.441162") + max_val = float("0.123047") + mean = float("-0.020248") + std = float("0.053088") data = None class Program_weight_tensor_parameter_175: name = "parameter_175" + original_name = "layer_norm_42.w_0" shape = [1024] dtype = "float32" - min_val = float("1.0") - max_val = float("1.0") - mean = float("1.0") + min_val = float("0.838867") + max_val = float("1.03027") + mean = float("0.9661") + std = float("0.00643206") data = None class Program_weight_tensor_parameter_176: name = "parameter_176" + original_name = "linear_137.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.256348") + max_val = float("0.135986") + mean = float("3.48757e-06") + std = float("0.0593899") data = None class Program_weight_tensor_parameter_177: name = "parameter_177" + original_name = "linear_137.w_0" shape = [4096, 1024] dtype = "float32" - min_val = float("-0.102834") - max_val = float("0.0985034") - mean = float("-3.2641e-06") - std = float("0.0200017") + min_val = float("-1.04297") + max_val = float("0.43042") + mean = float("-1.03813e-05") + std = float("0.0339604") data = None class Program_weight_tensor_parameter_178: name = "parameter_178" + original_name = "linear_136.b_0" shape = [4096] dtype = "float32" + min_val = float("-0.120972") + max_val = float("0.144897") + mean = float("-0.0279362") + std = float("0.0181329") data = None class Program_weight_tensor_parameter_179: name = "parameter_179" + original_name = "linear_136.w_0" shape = [1024, 4096] dtype = "float32" - min_val = float("-0.0961081") - max_val = float("0.0969126") - mean = float("-1.71144e-06") - std = float("0.0199969") + min_val = float("-0.231689") + max_val = float("0.195557") + mean = float("0.000256277") + std = float("0.0350948") data = None class Program_weight_tensor_parameter_180: name = "parameter_180" + original_name = "layer_norm_41.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.523926") + max_val = float("0.265381") + mean = float("-0.00876716") + std = float("0.0833981") data = None class Program_weight_tensor_parameter_181: name = "parameter_181" + original_name = "layer_norm_41.w_0" shape = [1024] dtype = "float32" - min_val = float("1.0") - max_val = float("1.0") - mean = float("1.0") + min_val = float("0.939941") + max_val = float("1.2168") + mean = float("0.952867") + std = float("0.00965346") data = None class Program_weight_tensor_parameter_182: name = "parameter_182" + original_name = "linear_135.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.163086") + max_val = float("0.0844727") + mean = float("0.000504746") + std = float("0.0399244") data = None class Program_weight_tensor_parameter_183: name = "parameter_183" + original_name = "linear_135.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.107998") - max_val = float("0.0934133") - mean = float("2.58743e-06") - std = float("0.0199802") + min_val = float("-0.383789") + max_val = float("0.345703") + mean = float("5.07133e-06") + std = float("0.0366016") data = None class Program_weight_tensor_parameter_184: name = "parameter_184" + original_name = "linear_134.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.214722") + max_val = float("0.219971") + mean = float("0.00176371") + std = float("0.0591464") data = None class Program_weight_tensor_parameter_185: name = "parameter_185" + original_name = "linear_134.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.100444") - max_val = float("0.0946931") - mean = float("6.94006e-06") - std = float("0.0199927") + min_val = float("-0.144409") + max_val = float("0.163818") + mean = float("-1.30126e-05") + std = float("0.0344449") data = None class Program_weight_tensor_parameter_186: name = "parameter_186" + original_name = "linear_133.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.126221") + max_val = float("0.10968") + mean = float("0.000189028") + std = float("0.0274629") data = None class Program_weight_tensor_parameter_187: name = "parameter_187" + original_name = "linear_133.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.0984793") - max_val = float("0.0928984") - mean = float("-5.61687e-06") - std = float("0.0199963") + min_val = float("-0.17395") + max_val = float("0.161011") + mean = float("-3.1963e-06") + std = float("0.0396784") data = None class Program_weight_tensor_parameter_188: name = "parameter_188" + original_name = "linear_132.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.273682") + max_val = float("0.250732") + mean = float("0.00030725") + std = float("0.0327842") data = None class Program_weight_tensor_parameter_189: name = "parameter_189" + original_name = "linear_132.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.0945283") - max_val = float("0.0934422") - mean = float("4.37499e-06") - std = float("0.0200078") + min_val = float("-0.202637") + max_val = float("0.21228") + mean = float("-1.79887e-05") + std = float("0.0343667") data = None class Program_weight_tensor_parameter_190: name = "parameter_190" + original_name = "layer_norm_40.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.652832") + max_val = float("0.214111") + mean = float("-0.00531139") + std = float("0.0621133") data = None class Program_weight_tensor_parameter_191: name = "parameter_191" + original_name = "layer_norm_40.w_0" shape = [1024] dtype = "float32" - min_val = float("1.0") - max_val = float("1.0") - mean = float("1.0") + min_val = float("0.941406") + max_val = float("1.01953") + mean = float("0.960682") + std = float("0.00439287") data = None class Program_weight_tensor_parameter_192: name = "parameter_192" + original_name = "linear_131.b_0" shape = [512] dtype = "float32" + min_val = float("-0.241089") + max_val = float("0.208008") + mean = float("-0.0143757") + std = float("0.0480623") data = None class Program_weight_tensor_parameter_193: name = "parameter_193" + original_name = "linear_131.w_0" shape = [1024, 512] dtype = "float32" - min_val = float("-0.100323") - max_val = float("0.0899183") - mean = float("-9.67633e-06") - std = float("0.0199855") + min_val = float("-0.129883") + max_val = float("0.128662") + mean = float("1.55831e-05") + std = float("0.020567") data = None class Program_weight_tensor_parameter_194: name = "parameter_194" + original_name = "linear_130.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.177612") + max_val = float("0.0778198") + mean = float("9.69057e-05") + std = float("0.0319041") data = None class Program_weight_tensor_parameter_195: name = "parameter_195" + original_name = "linear_130.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.0931948") - max_val = float("0.103837") - mean = float("3.35368e-05") - std = float("0.0199896") + min_val = float("-0.308594") + max_val = float("0.350098") + mean = float("-5.04435e-06") + std = float("0.0256714") data = None class Program_weight_tensor_parameter_196: name = "parameter_196" + original_name = "linear_129.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.161743") + max_val = float("0.165649") + mean = float("-0.00190979") + std = float("0.0485538") data = None class Program_weight_tensor_parameter_197: name = "parameter_197" + original_name = "linear_129.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.0946333") - max_val = float("0.103733") - mean = float("4.11378e-05") - std = float("0.0199926") + min_val = float("-0.136475") + max_val = float("0.130493") + mean = float("1.74217e-05") + std = float("0.0337152") data = None class Program_weight_tensor_parameter_198: name = "parameter_198" + original_name = "linear_128.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.078125") + max_val = float("0.0743408") + mean = float("0.000979235") + std = float("0.0256346") data = None class Program_weight_tensor_parameter_199: name = "parameter_199" + original_name = "linear_128.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.0956644") - max_val = float("0.107132") - mean = float("2.45002e-05") - std = float("0.0199913") + min_val = float("-0.11792") + max_val = float("0.124451") + mean = float("-3.44983e-05") + std = float("0.0274477") data = None class Program_weight_tensor_parameter_200: name = "parameter_200" + original_name = "linear_127.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.0223541") + max_val = float("0.0136032") + mean = float("-0.000110644") + std = float("0.00292466") data = None class Program_weight_tensor_parameter_201: name = "parameter_201" + original_name = "linear_127.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.106028") - max_val = float("0.0984437") - mean = float("-2.52862e-05") - std = float("0.0200209") + min_val = float("-0.136841") + max_val = float("0.136353") + mean = float("2.19697e-05") + std = float("0.0335088") data = None class Program_weight_tensor_parameter_202: name = "parameter_202" + original_name = "layer_norm_39.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.600098") + max_val = float("0.118469") + mean = float("-0.0143629") + std = float("0.0607705") data = None class Program_weight_tensor_parameter_203: name = "parameter_203" + original_name = "layer_norm_39.w_0" shape = [1024] dtype = "float32" - min_val = float("1.0") - max_val = float("1.0") - mean = float("1.0") + min_val = float("0.907715") + max_val = float("1.06445") + mean = float("0.966736") + std = float("0.00530408") data = None class Program_weight_tensor_parameter_204: name = "parameter_204" + original_name = "linear_126.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.322021") + max_val = float("0.133179") + mean = float("0.000107616") + std = float("0.0526926") data = None class Program_weight_tensor_parameter_205: name = "parameter_205" + original_name = "linear_126.w_0" shape = [4096, 1024] dtype = "float32" - min_val = float("-0.106987") - max_val = float("0.103369") - mean = float("-1.00166e-05") - std = float("0.019987") + min_val = float("-1.27344") + max_val = float("0.459717") + mean = float("-1.21559e-05") + std = float("0.0343892") data = None class Program_weight_tensor_parameter_206: name = "parameter_206" + original_name = "linear_125.b_0" shape = [4096] dtype = "float32" + min_val = float("-0.120239") + max_val = float("0.111206") + mean = float("-0.0285688") + std = float("0.015663") data = None class Program_weight_tensor_parameter_207: name = "parameter_207" + original_name = "linear_125.w_0" shape = [1024, 4096] dtype = "float32" - min_val = float("-0.0979837") - max_val = float("0.110512") - mean = float("1.08441e-05") - std = float("0.0200008") + min_val = float("-0.204712") + max_val = float("0.325928") + mean = float("0.000271252") + std = float("0.0356911") data = None class Program_weight_tensor_parameter_208: name = "parameter_208" + original_name = "layer_norm_38.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.557129") + max_val = float("0.20874") + mean = float("-0.00828665") + std = float("0.0886667") data = None class Program_weight_tensor_parameter_209: name = "parameter_209" + original_name = "layer_norm_38.w_0" shape = [1024] dtype = "float32" - min_val = float("1.0") - max_val = float("1.0") - mean = float("1.0") + min_val = float("0.941895") + max_val = float("0.998047") + mean = float("0.951185") + std = float("0.00392054") data = None class Program_weight_tensor_parameter_210: name = "parameter_210" + original_name = "linear_124.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.219604") + max_val = float("0.0977173") + mean = float("0.000263113") + std = float("0.0382921") data = None class Program_weight_tensor_parameter_211: name = "parameter_211" + original_name = "linear_124.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.0995923") - max_val = float("0.0965366") - mean = float("6.8185e-06") - std = float("0.0199911") + min_val = float("-0.403809") + max_val = float("0.393799") + mean = float("-1.29511e-06") + std = float("0.033811") data = None class Program_weight_tensor_parameter_212: name = "parameter_212" + original_name = "linear_123.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.183105") + max_val = float("0.160522") + mean = float("0.000225446") + std = float("0.0600205") data = None class Program_weight_tensor_parameter_213: name = "parameter_213" + original_name = "linear_123.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.0992627") - max_val = float("0.10944") - mean = float("-2.49174e-06") - std = float("0.0199839") + min_val = float("-0.156372") + max_val = float("0.154541") + mean = float("8.98549e-06") + std = float("0.0355561") data = None class Program_weight_tensor_parameter_214: name = "parameter_214" + original_name = "linear_122.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.137085") + max_val = float("0.127319") + mean = float("-0.000637583") + std = float("0.032504") data = None class Program_weight_tensor_parameter_215: name = "parameter_215" + original_name = "linear_122.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.097401") - max_val = float("0.0967019") - mean = float("1.13902e-05") - std = float("0.0199877") + min_val = float("-0.167847") + max_val = float("0.154907") + mean = float("1.46823e-05") + std = float("0.0366147") data = None class Program_weight_tensor_parameter_216: name = "parameter_216" + original_name = "linear_121.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.240356") + max_val = float("0.192261") + mean = float("-0.000497012") + std = float("0.0380276") data = None class Program_weight_tensor_parameter_217: name = "parameter_217" + original_name = "linear_121.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.0981034") - max_val = float("0.10037") - mean = float("-4.5789e-06") - std = float("0.0199754") + min_val = float("-0.179565") + max_val = float("0.192627") + mean = float("-3.64932e-05") + std = float("0.0358548") data = None class Program_weight_tensor_parameter_218: name = "parameter_218" + original_name = "layer_norm_37.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.730957") + max_val = float("0.183472") + mean = float("-0.00297988") + std = float("0.0790677") data = None class Program_weight_tensor_parameter_219: name = "parameter_219" + original_name = "layer_norm_37.w_0" shape = [1024] dtype = "float32" - min_val = float("1.0") - max_val = float("1.0") - mean = float("1.0") + min_val = float("0.931152") + max_val = float("1.03516") + mean = float("0.962269") + std = float("0.00463529") data = None class Program_weight_tensor_parameter_220: name = "parameter_220" + original_name = "linear_120.b_0" shape = [512] dtype = "float32" + min_val = float("-0.157837") + max_val = float("0.215698") + mean = float("-0.0245563") + std = float("0.0510572") data = None class Program_weight_tensor_parameter_221: name = "parameter_221" + original_name = "linear_120.w_0" shape = [1024, 512] dtype = "float32" - min_val = float("-0.0884315") - max_val = float("0.0915591") - mean = float("-3.34334e-06") - std = float("0.0200283") + min_val = float("-0.15686") + max_val = float("0.1521") + mean = float("0.000179217") + std = float("0.0243807") data = None class Program_weight_tensor_parameter_222: name = "parameter_222" + original_name = "linear_119.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.223389") + max_val = float("0.0814819") + mean = float("-0.000133873") + std = float("0.0329361") data = None class Program_weight_tensor_parameter_223: name = "parameter_223" + original_name = "linear_119.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.0935903") - max_val = float("0.0936766") - mean = float("-1.17989e-05") - std = float("0.0200204") + min_val = float("-0.428711") + max_val = float("0.422852") + mean = float("9.95933e-07") + std = float("0.0212588") data = None class Program_weight_tensor_parameter_224: name = "parameter_224" + original_name = "linear_118.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.13147") + max_val = float("0.134155") + mean = float("-0.00125458") + std = float("0.0350922") data = None class Program_weight_tensor_parameter_225: name = "parameter_225" + original_name = "linear_118.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.0987654") - max_val = float("0.095979") - mean = float("-1.96123e-05") - std = float("0.0200064") + min_val = float("-0.128906") + max_val = float("0.139526") + mean = float("6.47289e-06") + std = float("0.0328073") data = None class Program_weight_tensor_parameter_226: name = "parameter_226" + original_name = "linear_117.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.0944824") + max_val = float("0.0935059") + mean = float("-4.75917e-05") + std = float("0.0382414") data = None class Program_weight_tensor_parameter_227: name = "parameter_227" + original_name = "linear_117.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.0972514") - max_val = float("0.0945943") - mean = float("1.71259e-06") - std = float("0.0199927") + min_val = float("-0.0997314") + max_val = float("0.098938") + mean = float("8.28139e-06") + std = float("0.0227143") data = None class Program_weight_tensor_parameter_228: name = "parameter_228" + original_name = "linear_116.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.0211639") + max_val = float("0.0116577") + mean = float("2.68113e-05") + std = float("0.00215626") data = None class Program_weight_tensor_parameter_229: name = "parameter_229" + original_name = "linear_116.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.094617") - max_val = float("0.101108") - mean = float("-2.5317e-05") - std = float("0.0199948") + min_val = float("-0.138306") + max_val = float("0.131226") + mean = float("-5.5017e-06") + std = float("0.032351") data = None class Program_weight_tensor_parameter_230: name = "parameter_230" + original_name = "layer_norm_36.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.476562") + max_val = float("0.13147") + mean = float("-0.0151846") + std = float("0.0656483") data = None class Program_weight_tensor_parameter_231: name = "parameter_231" + original_name = "layer_norm_36.w_0" shape = [1024] dtype = "float32" - min_val = float("1.0") - max_val = float("1.0") - mean = float("1.0") + min_val = float("0.874023") + max_val = float("0.985352") + mean = float("0.972496") + std = float("0.0057502") data = None class Program_weight_tensor_parameter_232: name = "parameter_232" + original_name = "linear_115.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.121216") + max_val = float("0.135498") + mean = float("6.16143e-05") + std = float("0.050325") data = None class Program_weight_tensor_parameter_233: name = "parameter_233" + original_name = "linear_115.w_0" shape = [4096, 1024] dtype = "float32" - min_val = float("-0.0994821") - max_val = float("0.0982011") - mean = float("2.35682e-05") - std = float("0.0199952") + min_val = float("-0.88916") + max_val = float("0.366211") + mean = float("-1.60717e-06") + std = float("0.0349685") data = None class Program_weight_tensor_parameter_234: name = "parameter_234" + original_name = "linear_114.b_0" shape = [4096] dtype = "float32" + min_val = float("-0.0734863") + max_val = float("0.0766602") + mean = float("-0.0294882") + std = float("0.0120966") data = None class Program_weight_tensor_parameter_235: name = "parameter_235" + original_name = "linear_114.w_0" shape = [1024, 4096] dtype = "float32" - min_val = float("-0.0999906") - max_val = float("0.0969719") - mean = float("-2.61974e-07") - std = float("0.0199946") + min_val = float("-0.161743") + max_val = float("0.159058") + mean = float("0.00021809") + std = float("0.0368582") data = None class Program_weight_tensor_parameter_236: name = "parameter_236" + original_name = "layer_norm_35.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.538086") + max_val = float("0.165283") + mean = float("-0.00640279") + std = float("0.0815196") data = None class Program_weight_tensor_parameter_237: name = "parameter_237" + original_name = "layer_norm_35.w_0" shape = [1024] dtype = "float32" - min_val = float("1.0") - max_val = float("1.0") - mean = float("1.0") + min_val = float("0.9375") + max_val = float("1.25977") + mean = float("0.94835") + std = float("0.0105667") data = None class Program_weight_tensor_parameter_238: name = "parameter_238" + original_name = "linear_113.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.199585") + max_val = float("0.0912476") + mean = float("1.07998e-05") + std = float("0.0367528") data = None class Program_weight_tensor_parameter_239: name = "parameter_239" + original_name = "linear_113.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.0950171") - max_val = float("0.0926724") - mean = float("-2.55019e-05") - std = float("0.0199999") + min_val = float("-0.370117") + max_val = float("0.364258") + mean = float("7.10388e-06") + std = float("0.0335488") data = None class Program_weight_tensor_parameter_240: name = "parameter_240" + original_name = "linear_112.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.192871") + max_val = float("0.194702") + mean = float("-0.000471501") + std = float("0.059078") data = None class Program_weight_tensor_parameter_241: name = "parameter_241" + original_name = "linear_112.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.0998045") - max_val = float("0.092358") - mean = float("2.34201e-05") - std = float("0.0199979") + min_val = float("-0.147705") + max_val = float("0.144775") + mean = float("-2.60025e-05") + std = float("0.0344203") data = None class Program_weight_tensor_parameter_242: name = "parameter_242" + original_name = "linear_111.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.132202") + max_val = float("0.128174") + mean = float("0.000177023") + std = float("0.0272657") data = None class Program_weight_tensor_parameter_243: name = "parameter_243" + original_name = "linear_111.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.0916965") - max_val = float("0.0948372") - mean = float("3.52695e-05") - std = float("0.0199848") + min_val = float("-0.143921") + max_val = float("0.162354") + mean = float("2.41112e-06") + std = float("0.0365185") data = None class Program_weight_tensor_parameter_244: name = "parameter_244" + original_name = "linear_110.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.237305") + max_val = float("0.234863") + mean = float("-7.11976e-05") + std = float("0.0264564") data = None class Program_weight_tensor_parameter_245: name = "parameter_245" + original_name = "linear_110.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.0969563") - max_val = float("0.0939267") - mean = float("-1.45276e-05") - std = float("0.0200178") + min_val = float("-0.189453") + max_val = float("0.177002") + mean = float("-2.83879e-05") + std = float("0.0345099") data = None class Program_weight_tensor_parameter_246: name = "parameter_246" + original_name = "layer_norm_34.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.590332") + max_val = float("0.16626") + mean = float("-0.00170094") + std = float("0.0679209") data = None class Program_weight_tensor_parameter_247: name = "parameter_247" + original_name = "layer_norm_34.w_0" shape = [1024] dtype = "float32" - min_val = float("1.0") - max_val = float("1.0") - mean = float("1.0") + min_val = float("0.923828") + max_val = float("0.999512") + mean = float("0.961038") + std = float("0.00421373") data = None class Program_weight_tensor_parameter_248: name = "parameter_248" + original_name = "linear_109.b_0" shape = [512] dtype = "float32" + min_val = float("-0.233032") + max_val = float("0.277832") + mean = float("-0.0249556") + std = float("0.0520129") data = None class Program_weight_tensor_parameter_249: name = "parameter_249" + original_name = "linear_109.w_0" shape = [1024, 512] dtype = "float32" - min_val = float("-0.0982625") - max_val = float("0.112291") - mean = float("3.77821e-05") - std = float("0.0199698") + min_val = float("-0.179199") + max_val = float("0.142578") + mean = float("8.92417e-05") + std = float("0.023044") data = None class Program_weight_tensor_parameter_250: name = "parameter_250" + original_name = "linear_108.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.1875") + max_val = float("0.0821533") + mean = float("-0.000341071") + std = float("0.0346441") data = None class Program_weight_tensor_parameter_251: name = "parameter_251" + original_name = "linear_108.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.0982452") - max_val = float("0.0938883") - mean = float("4.49206e-06") - std = float("0.0199994") + min_val = float("-0.391602") + max_val = float("0.353027") + mean = float("5.99148e-06") + std = float("0.0210915") data = None class Program_weight_tensor_parameter_252: name = "parameter_252" + original_name = "linear_107.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.139526") + max_val = float("0.14209") + mean = float("0.00235831") + std = float("0.0355021") data = None class Program_weight_tensor_parameter_253: name = "parameter_253" + original_name = "linear_107.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.0927645") - max_val = float("0.0912421") - mean = float("1.76549e-06") - std = float("0.0200088") + min_val = float("-0.137451") + max_val = float("0.134521") + mean = float("-8.4717e-07") + std = float("0.0327593") data = None class Program_weight_tensor_parameter_254: name = "parameter_254" + original_name = "linear_106.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.0725098") + max_val = float("0.0779419") + mean = float("0.000243429") + std = float("0.0284383") data = None class Program_weight_tensor_parameter_255: name = "parameter_255" + original_name = "linear_106.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.0952781") - max_val = float("0.096158") - mean = float("-7.59462e-06") - std = float("0.0200257") + min_val = float("-0.102844") + max_val = float("0.101868") + mean = float("-1.24259e-05") + std = float("0.0223102") data = None class Program_weight_tensor_parameter_256: name = "parameter_256" + original_name = "linear_105.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.00919342") + max_val = float("0.0101089") + mean = float("-9.55271e-05") + std = float("0.00162568") data = None class Program_weight_tensor_parameter_257: name = "parameter_257" + original_name = "linear_105.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.0986601") - max_val = float("0.0912809") - mean = float("-3.25412e-06") - std = float("0.0200009") + min_val = float("-0.139404") + max_val = float("0.145752") + mean = float("-5.39504e-06") + std = float("0.0321462") data = None class Program_weight_tensor_parameter_258: name = "parameter_258" + original_name = "layer_norm_33.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.463623") + max_val = float("0.129395") + mean = float("-0.00914398") + std = float("0.0645008") data = None class Program_weight_tensor_parameter_259: name = "parameter_259" + original_name = "layer_norm_33.w_0" shape = [1024] dtype = "float32" - min_val = float("1.0") - max_val = float("1.0") - mean = float("1.0") + min_val = float("0.773926") + max_val = float("0.987305") + mean = float("0.974666") + std = float("0.00893311") data = None class Program_weight_tensor_parameter_260: name = "parameter_260" + original_name = "linear_104.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.118591") + max_val = float("0.128052") + mean = float("0.00050242") + std = float("0.0450276") data = None class Program_weight_tensor_parameter_261: name = "parameter_261" + original_name = "linear_104.w_0" shape = [4096, 1024] dtype = "float32" - min_val = float("-0.10063") - max_val = float("0.109828") - mean = float("2.13232e-05") - std = float("0.0199861") + min_val = float("-0.725098") + max_val = float("0.45166") + mean = float("7.70257e-06") + std = float("0.0354617") data = None class Program_weight_tensor_parameter_262: name = "parameter_262" + original_name = "linear_103.b_0" shape = [4096] dtype = "float32" + min_val = float("-0.1026") + max_val = float("0.0713501") + mean = float("-0.0302563") + std = float("0.0110595") data = None class Program_weight_tensor_parameter_263: name = "parameter_263" + original_name = "linear_103.w_0" shape = [1024, 4096] dtype = "float32" - min_val = float("-0.100968") - max_val = float("0.0985829") - mean = float("1.28652e-05") - std = float("0.0199902") + min_val = float("-0.170532") + max_val = float("0.162598") + mean = float("0.000178452") + std = float("0.036939") data = None class Program_weight_tensor_parameter_264: name = "parameter_264" + original_name = "layer_norm_32.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.607422") + max_val = float("0.171997") + mean = float("-0.00593873") + std = float("0.0841203") data = None class Program_weight_tensor_parameter_265: name = "parameter_265" + original_name = "layer_norm_32.w_0" shape = [1024] dtype = "float32" - min_val = float("1.0") - max_val = float("1.0") - mean = float("1.0") + min_val = float("0.937012") + max_val = float("1.41895") + mean = float("0.949653") + std = float("0.0156994") data = None class Program_weight_tensor_parameter_266: name = "parameter_266" + original_name = "linear_102.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.166016") + max_val = float("0.129761") + mean = float("0.000521014") + std = float("0.0483909") data = None class Program_weight_tensor_parameter_267: name = "parameter_267" + original_name = "linear_102.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.0911595") - max_val = float("0.0941098") - mean = float("1.104e-05") - std = float("0.019992") + min_val = float("-0.32373") + max_val = float("0.329834") + mean = float("3.78237e-06") + std = float("0.0345458") data = None class Program_weight_tensor_parameter_268: name = "parameter_268" + original_name = "linear_101.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.162354") + max_val = float("0.213501") + mean = float("0.00118156") + std = float("0.0511273") data = None class Program_weight_tensor_parameter_269: name = "parameter_269" + original_name = "linear_101.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.0991791") - max_val = float("0.0945547") - mean = float("-2.58733e-05") - std = float("0.0199884") + min_val = float("-0.182251") + max_val = float("0.14624") + mean = float("-1.56188e-05") + std = float("0.03546") data = None class Program_weight_tensor_parameter_270: name = "parameter_270" + original_name = "linear_100.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.180176") + max_val = float("0.183105") + mean = float("-0.000548592") + std = float("0.0409745") data = None class Program_weight_tensor_parameter_271: name = "parameter_271" + original_name = "linear_100.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.0951836") - max_val = float("0.0975088") - mean = float("5.78947e-07") - std = float("0.020024") + min_val = float("-0.14917") + max_val = float("0.146729") + mean = float("3.85424e-05") + std = float("0.0373351") data = None class Program_weight_tensor_parameter_272: name = "parameter_272" + original_name = "linear_99.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.124451") + max_val = float("0.166626") + mean = float("0.000307189") + std = float("0.0191431") data = None class Program_weight_tensor_parameter_273: name = "parameter_273" + original_name = "linear_99.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.0950242") - max_val = float("0.0889688") - mean = float("-4.62415e-06") - std = float("0.019974") + min_val = float("-0.209351") + max_val = float("0.180298") + mean = float("-3.29862e-05") + std = float("0.0355982") data = None class Program_weight_tensor_parameter_274: name = "parameter_274" + original_name = "layer_norm_31.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.498779") + max_val = float("0.151245") + mean = float("-0.000517263") + std = float("0.0791918") data = None class Program_weight_tensor_parameter_275: name = "parameter_275" + original_name = "layer_norm_31.w_0" shape = [1024] dtype = "float32" - min_val = float("1.0") - max_val = float("1.0") - mean = float("1.0") + min_val = float("0.888184") + max_val = float("0.98584") + mean = float("0.958619") + std = float("0.00434506") data = None class Program_weight_tensor_parameter_276: name = "parameter_276" + original_name = "linear_98.b_0" shape = [512] dtype = "float32" + min_val = float("-0.245728") + max_val = float("0.289795") + mean = float("-0.0147373") + std = float("0.0493199") data = None class Program_weight_tensor_parameter_277: name = "parameter_277" + original_name = "linear_98.w_0" shape = [1024, 512] dtype = "float32" - min_val = float("-0.0910291") - max_val = float("0.0953185") - mean = float("-8.25923e-06") - std = float("0.0199867") + min_val = float("-0.206909") + max_val = float("0.198242") + mean = float("4.65898e-05") + std = float("0.0215197") data = None class Program_weight_tensor_parameter_278: name = "parameter_278" + original_name = "linear_97.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.185791") + max_val = float("0.127563") + mean = float("-0.000264936") + std = float("0.0497672") data = None class Program_weight_tensor_parameter_279: name = "parameter_279" + original_name = "linear_97.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.0940048") - max_val = float("0.100451") - mean = float("5.45249e-06") - std = float("0.0199828") + min_val = float("-0.255127") + max_val = float("0.266602") + mean = float("3.6921e-07") + std = float("0.0207555") data = None class Program_weight_tensor_parameter_280: name = "parameter_280" + original_name = "linear_96.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.186401") + max_val = float("0.137817") + mean = float("-0.000506923") + std = float("0.0335921") data = None class Program_weight_tensor_parameter_281: name = "parameter_281" + original_name = "linear_96.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.100147") - max_val = float("0.0928783") - mean = float("9.2778e-06") - std = float("0.0199992") + min_val = float("-0.150146") + max_val = float("0.147461") + mean = float("2.03841e-05") + std = float("0.0339246") data = None class Program_weight_tensor_parameter_282: name = "parameter_282" + original_name = "linear_95.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.074646") + max_val = float("0.0725098") + mean = float("-0.000315445") + std = float("0.0226842") data = None class Program_weight_tensor_parameter_283: name = "parameter_283" + original_name = "linear_95.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.11436") - max_val = float("0.0941297") - mean = float("-7.02163e-06") - std = float("0.0200214") + min_val = float("-0.098877") + max_val = float("0.0971069") + mean = float("-1.64952e-05") + std = float("0.0218104") data = None class Program_weight_tensor_parameter_284: name = "parameter_284" + original_name = "linear_94.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.00806427") + max_val = float("0.00700378") + mean = float("5.51337e-06") + std = float("0.00153575") data = None class Program_weight_tensor_parameter_285: name = "parameter_285" + original_name = "linear_94.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.101258") - max_val = float("0.0940722") - mean = float("4.6099e-05") - std = float("0.0200132") + min_val = float("-0.150391") + max_val = float("0.158569") + mean = float("4.99343e-06") + std = float("0.0333562") data = None class Program_weight_tensor_parameter_286: name = "parameter_286" + original_name = "layer_norm_30.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.42749") + max_val = float("0.133545") + mean = float("-0.00672979") + std = float("0.0544821") data = None class Program_weight_tensor_parameter_287: name = "parameter_287" + original_name = "layer_norm_30.w_0" shape = [1024] dtype = "float32" - min_val = float("1.0") - max_val = float("1.0") - mean = float("1.0") + min_val = float("0.697754") + max_val = float("0.987793") + mean = float("0.977366") + std = float("0.0135496") data = None class Program_weight_tensor_parameter_288: name = "parameter_288" + original_name = "linear_93.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.137573") + max_val = float("0.135132") + mean = float("0.000137495") + std = float("0.0509008") data = None class Program_weight_tensor_parameter_289: name = "parameter_289" + original_name = "linear_93.w_0" shape = [4096, 1024] dtype = "float32" - min_val = float("-0.105675") - max_val = float("0.106266") - mean = float("1.33711e-05") - std = float("0.0200036") + min_val = float("-0.685547") + max_val = float("0.591309") + mean = float("8.39059e-06") + std = float("0.0346649") data = None class Program_weight_tensor_parameter_290: name = "parameter_290" + original_name = "linear_92.b_0" shape = [4096] dtype = "float32" + min_val = float("-0.0772705") + max_val = float("0.0779419") + mean = float("-0.0328036") + std = float("0.0103623") data = None class Program_weight_tensor_parameter_291: name = "parameter_291" + original_name = "linear_92.w_0" shape = [1024, 4096] dtype = "float32" - min_val = float("-0.104936") - max_val = float("0.0980911") - mean = float("-1.7719e-06") - std = float("0.0199966") + min_val = float("-0.156494") + max_val = float("0.153687") + mean = float("0.000126656") + std = float("0.0361335") data = None class Program_weight_tensor_parameter_292: name = "parameter_292" + original_name = "layer_norm_29.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.65918") + max_val = float("0.349609") + mean = float("-0.00371468") + std = float("0.0944608") data = None class Program_weight_tensor_parameter_293: name = "parameter_293" + original_name = "layer_norm_29.w_0" shape = [1024] dtype = "float32" - min_val = float("1.0") - max_val = float("1.0") - mean = float("1.0") + min_val = float("0.942383") + max_val = float("1.44238") + mean = float("0.953568") + std = float("0.0174744") data = None class Program_weight_tensor_parameter_294: name = "parameter_294" + original_name = "linear_91.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.166992") + max_val = float("0.134644") + mean = float("-0.000406027") + std = float("0.0543582") data = None class Program_weight_tensor_parameter_295: name = "parameter_295" + original_name = "linear_91.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.101153") - max_val = float("0.0981075") - mean = float("-3.71591e-05") - std = float("0.020003") + min_val = float("-0.314697") + max_val = float("0.349609") + mean = float("-1.88661e-05") + std = float("0.0338309") data = None class Program_weight_tensor_parameter_296: name = "parameter_296" + original_name = "linear_90.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.155029") + max_val = float("0.147583") + mean = float("-0.00174626") + std = float("0.054196") data = None class Program_weight_tensor_parameter_297: name = "parameter_297" + original_name = "linear_90.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.0976708") - max_val = float("0.0952972") - mean = float("3.79038e-05") - std = float("0.0200112") + min_val = float("-0.163574") + max_val = float("0.169434") + mean = float("-1.65566e-05") + std = float("0.0368129") data = None class Program_weight_tensor_parameter_298: name = "parameter_298" + original_name = "linear_89.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.157471") + max_val = float("0.152466") + mean = float("-0.000854221") + std = float("0.0325367") data = None class Program_weight_tensor_parameter_299: name = "parameter_299" + original_name = "linear_89.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.0996401") - max_val = float("0.0958766") - mean = float("-7.93499e-06") - std = float("0.0200037") + min_val = float("-0.147583") + max_val = float("0.159668") + mean = float("2.62521e-05") + std = float("0.0362974") data = None class Program_weight_tensor_parameter_300: name = "parameter_300" + original_name = "linear_88.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.123596") + max_val = float("0.0946655") + mean = float("-0.00158462") + std = float("0.0178123") data = None class Program_weight_tensor_parameter_301: name = "parameter_301" + original_name = "linear_88.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.0984399") - max_val = float("0.0973662") - mean = float("2.52004e-05") - std = float("0.0199824") + min_val = float("-0.171875") + max_val = float("0.200684") + mean = float("1.39106e-05") + std = float("0.0361574") data = None class Program_weight_tensor_parameter_302: name = "parameter_302" + original_name = "layer_norm_28.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.446777") + max_val = float("0.223389") + mean = float("-0.000111504") + std = float("0.081971") data = None class Program_weight_tensor_parameter_303: name = "parameter_303" + original_name = "layer_norm_28.w_0" shape = [1024] dtype = "float32" - min_val = float("1.0") - max_val = float("1.0") - mean = float("1.0") + min_val = float("0.92041") + max_val = float("0.972656") + mean = float("0.961025") + std = float("0.0039625") data = None class Program_weight_tensor_parameter_304: name = "parameter_304" + original_name = "linear_87.b_0" shape = [512] dtype = "float32" + min_val = float("-0.153198") + max_val = float("0.263184") + mean = float("-0.0248493") + std = float("0.0569705") data = None class Program_weight_tensor_parameter_305: name = "parameter_305" + original_name = "linear_87.w_0" shape = [1024, 512] dtype = "float32" - min_val = float("-0.101357") - max_val = float("0.094174") - mean = float("4.46963e-05") - std = float("0.0200347") + min_val = float("-0.132446") + max_val = float("0.134277") + mean = float("2.79023e-06") + std = float("0.0204691") data = None class Program_weight_tensor_parameter_306: name = "parameter_306" + original_name = "linear_86.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.217529") + max_val = float("0.114685") + mean = float("-0.000144818") + std = float("0.0400873") data = None class Program_weight_tensor_parameter_307: name = "parameter_307" + original_name = "linear_86.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.0929681") - max_val = float("0.0891093") - mean = float("-1.62389e-05") - std = float("0.0199894") + min_val = float("-0.211426") + max_val = float("0.210327") + mean = float("1.19935e-07") + std = float("0.0207701") data = None class Program_weight_tensor_parameter_308: name = "parameter_308" + original_name = "linear_85.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.153809") + max_val = float("0.161133") + mean = float("-0.00063892") + std = float("0.04344") data = None class Program_weight_tensor_parameter_309: name = "parameter_309" + original_name = "linear_85.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.0928909") - max_val = float("0.0975276") - mean = float("-1.76411e-05") - std = float("0.0199951") + min_val = float("-0.155151") + max_val = float("0.151123") + mean = float("-8.7411e-06") + std = float("0.0355235") data = None class Program_weight_tensor_parameter_310: name = "parameter_310" + original_name = "linear_84.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.0715942") + max_val = float("0.0753174") + mean = float("7.68704e-05") + std = float("0.0220949") data = None class Program_weight_tensor_parameter_311: name = "parameter_311" + original_name = "linear_84.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.0939867") - max_val = float("0.0927453") - mean = float("1.03521e-05") - std = float("0.0199962") + min_val = float("-0.0960693") + max_val = float("0.0983276") + mean = float("-4.30156e-06") + std = float("0.021972") data = None class Program_weight_tensor_parameter_312: name = "parameter_312" + original_name = "linear_83.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.0143051") + max_val = float("0.0152588") + mean = float("-2.92414e-05") + std = float("0.00193818") data = None class Program_weight_tensor_parameter_313: name = "parameter_313" + original_name = "linear_83.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.0898961") - max_val = float("0.0931385") - mean = float("2.22165e-06") - std = float("0.0200037") + min_val = float("-0.145386") + max_val = float("0.149292") + mean = float("-2.06775e-07") + std = float("0.0349698") data = None class Program_weight_tensor_parameter_314: name = "parameter_314" + original_name = "layer_norm_27.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.33374") + max_val = float("0.144287") + mean = float("-0.00200115") + std = float("0.0577484") data = None class Program_weight_tensor_parameter_315: name = "parameter_315" + original_name = "layer_norm_27.w_0" shape = [1024] dtype = "float32" - min_val = float("1.0") - max_val = float("1.0") - mean = float("1.0") + min_val = float("0.674805") + max_val = float("0.990723") + mean = float("0.978185") + std = float("0.0151641") data = None class Program_weight_tensor_parameter_316: name = "parameter_316" + original_name = "linear_82.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.197388") + max_val = float("0.134644") + mean = float("0.000302168") + std = float("0.0550577") data = None class Program_weight_tensor_parameter_317: name = "parameter_317" + original_name = "linear_82.w_0" shape = [4096, 1024] dtype = "float32" - min_val = float("-0.0980051") - max_val = float("0.106621") - mean = float("2.69303e-06") - std = float("0.019993") + min_val = float("-0.623535") + max_val = float("0.497559") + mean = float("5.5003e-06") + std = float("0.0333114") data = None class Program_weight_tensor_parameter_318: name = "parameter_318" + original_name = "linear_81.b_0" shape = [4096] dtype = "float32" + min_val = float("-0.102295") + max_val = float("0.0664062") + mean = float("-0.0371166") + std = float("0.0113836") data = None class Program_weight_tensor_parameter_319: name = "parameter_319" + original_name = "linear_81.w_0" shape = [1024, 4096] dtype = "float32" - min_val = float("-0.0995924") - max_val = float("0.105663") - mean = float("8.26453e-06") - std = float("0.0199945") + min_val = float("-0.148315") + max_val = float("0.194092") + mean = float("7.16009e-05") + std = float("0.0351085") data = None class Program_weight_tensor_parameter_320: name = "parameter_320" + original_name = "layer_norm_26.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.729004") + max_val = float("0.450439") + mean = float("-0.00242681") + std = float("0.100021") data = None class Program_weight_tensor_parameter_321: name = "parameter_321" + original_name = "layer_norm_26.w_0" shape = [1024] dtype = "float32" - min_val = float("1.0") - max_val = float("1.0") - mean = float("1.0") + min_val = float("0.936523") + max_val = float("1.48926") + mean = float("0.956938") + std = float("0.0197076") data = None class Program_weight_tensor_parameter_322: name = "parameter_322" + original_name = "linear_80.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.272461") + max_val = float("0.13562") + mean = float("4.79873e-05") + std = float("0.0347092") data = None class Program_weight_tensor_parameter_323: name = "parameter_323" + original_name = "linear_80.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.100632") - max_val = float("0.0994276") - mean = float("-7.67635e-06") - std = float("0.0200208") + min_val = float("-0.356934") + max_val = float("0.33667") + mean = float("-8.0903e-06") + std = float("0.0330627") data = None class Program_weight_tensor_parameter_324: name = "parameter_324" + original_name = "linear_79.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.181885") + max_val = float("0.187134") + mean = float("-0.00143465") + std = float("0.0762566") data = None class Program_weight_tensor_parameter_325: name = "parameter_325" + original_name = "linear_79.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.100072") - max_val = float("0.115185") - mean = float("-6.00974e-06") - std = float("0.020031") + min_val = float("-0.152832") + max_val = float("0.147827") + mean = float("-9.56824e-06") + std = float("0.0374546") data = None class Program_weight_tensor_parameter_326: name = "parameter_326" + original_name = "linear_78.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.104675") + max_val = float("0.102722") + mean = float("-0.000618589") + std = float("0.0280027") data = None class Program_weight_tensor_parameter_327: name = "parameter_327" + original_name = "linear_78.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.102729") - max_val = float("0.11009") - mean = float("1.67291e-05") - std = float("0.0200101") + min_val = float("-0.163452") + max_val = float("0.142456") + mean = float("2.21057e-05") + std = float("0.0353334") data = None class Program_weight_tensor_parameter_328: name = "parameter_328" + original_name = "linear_77.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.111694") + max_val = float("0.114258") + mean = float("6.53763e-05") + std = float("0.00974636") data = None class Program_weight_tensor_parameter_329: name = "parameter_329" + original_name = "linear_77.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.0939058") - max_val = float("0.0964981") - mean = float("-1.1446e-05") - std = float("0.0199957") + min_val = float("-0.153809") + max_val = float("0.167358") + mean = float("3.71841e-06") + std = float("0.0352176") data = None class Program_weight_tensor_parameter_330: name = "parameter_330" + original_name = "layer_norm_25.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.651367") + max_val = float("0.416992") + mean = float("-0.00141882") + std = float("0.119634") data = None class Program_weight_tensor_parameter_331: name = "parameter_331" + original_name = "layer_norm_25.w_0" shape = [1024] dtype = "float32" - min_val = float("1.0") - max_val = float("1.0") - mean = float("1.0") + min_val = float("0.932617") + max_val = float("1.06836") + mean = float("0.952882") + std = float("0.00663826") data = None class Program_weight_tensor_parameter_332: name = "parameter_332" + original_name = "linear_76.b_0" shape = [512] dtype = "float32" + min_val = float("-0.324707") + max_val = float("0.493652") + mean = float("-0.0661616") + std = float("0.109566") data = None class Program_weight_tensor_parameter_333: name = "parameter_333" + original_name = "linear_76.w_0" shape = [1024, 512] dtype = "float32" - min_val = float("-0.101455") - max_val = float("0.0967113") - mean = float("1.91953e-05") - std = float("0.0199993") + min_val = float("-0.171509") + max_val = float("0.185913") + mean = float("-4.78964e-05") + std = float("0.018896") data = None class Program_weight_tensor_parameter_334: name = "parameter_334" + original_name = "linear_75.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.302002") + max_val = float("0.186646") + mean = float("4.17548e-05") + std = float("0.0502882") data = None class Program_weight_tensor_parameter_335: name = "parameter_335" + original_name = "linear_75.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.0943566") - max_val = float("0.100414") - mean = float("2.8966e-05") - std = float("0.0200044") + min_val = float("-0.211548") + max_val = float("0.162842") + mean = float("-4.47718e-06") + std = float("0.0199127") data = None class Program_weight_tensor_parameter_336: name = "parameter_336" + original_name = "linear_74.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.239014") + max_val = float("0.324219") + mean = float("0.00237037") + std = float("0.0873101") data = None class Program_weight_tensor_parameter_337: name = "parameter_337" + original_name = "linear_74.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.101385") - max_val = float("0.0995255") - mean = float("3.2353e-05") - std = float("0.0200098") + min_val = float("-0.127563") + max_val = float("0.136108") + mean = float("-7.06233e-06") + std = float("0.0290819") data = None class Program_weight_tensor_parameter_338: name = "parameter_338" + original_name = "linear_73.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.179565") + max_val = float("0.182373") + mean = float("0.000747683") + std = float("0.056607") data = None class Program_weight_tensor_parameter_339: name = "parameter_339" + original_name = "linear_73.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.0952493") - max_val = float("0.100119") - mean = float("-2.63876e-05") - std = float("0.0199764") + min_val = float("-0.108948") + max_val = float("0.106445") + mean = float("9.96184e-07") + std = float("0.0205952") data = None class Program_weight_tensor_parameter_340: name = "parameter_340" + original_name = "linear_72.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.0062561") + max_val = float("0.00666046") + mean = float("3.19497e-05") + std = float("0.000941104") data = None class Program_weight_tensor_parameter_341: name = "parameter_341" + original_name = "linear_72.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.10141") - max_val = float("0.100573") - mean = float("-1.76694e-05") - std = float("0.0199793") + min_val = float("-0.135498") + max_val = float("0.139282") + mean = float("-1.32945e-05") + std = float("0.028954") data = None class Program_weight_tensor_parameter_342: name = "parameter_342" + original_name = "embedding_1.w_0" shape = [2, 1024] dtype = "float32" - min_val = float("-0.0764307") - max_val = float("0.0763354") - mean = float("-0.0015711") - std = float("0.0444587") + min_val = float("-1.04395") + max_val = float("0.357666") + mean = float("0.00267661") + std = float("0.0483818") data = None class Program_weight_tensor_parameter_343: name = "parameter_343" + original_name = "prophet_net_positional_embeddings_1.w_0" shape = [512, 1024] dtype = "float32" - min_val = float("-0.0624998") - max_val = float("0.0624997") - mean = float("4.01361e-05") - std = float("0.0360647") + min_val = float("-0.171997") + max_val = float("0.121765") + mean = float("4.2471e-06") + std = float("0.0142559") data = None class Program_weight_tensor_parameter_344: name = "parameter_344" + original_name = "layer_norm_24.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.578613") + max_val = float("0.742676") + mean = float("-0.00978997") + std = float("0.0743837") data = None class Program_weight_tensor_parameter_345: name = "parameter_345" + original_name = "layer_norm_24.w_0" shape = [1024] dtype = "float32" - min_val = float("1.0") - max_val = float("1.0") - mean = float("1.0") + min_val = float("0.435303") + max_val = float("1.01172") + mean = float("0.976902") + std = float("0.0378715") data = None class Program_weight_tensor_parameter_346: name = "parameter_346" + original_name = "linear_71.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.139526") + max_val = float("0.139526") + mean = float("0.000325459") + std = float("0.028331") data = None class Program_weight_tensor_parameter_347: name = "parameter_347" + original_name = "linear_71.w_0" shape = [4096, 1024] dtype = "float32" - min_val = float("-0.101248") - max_val = float("0.105454") - mean = float("-6.2514e-06") - std = float("0.0199929") + min_val = float("-0.663574") + max_val = float("0.67627") + mean = float("1.78765e-05") + std = float("0.0292763") data = None class Program_weight_tensor_parameter_348: name = "parameter_348" + original_name = "linear_70.b_0" shape = [4096] dtype = "float32" + min_val = float("-0.155884") + max_val = float("0.119385") + mean = float("-0.0261288") + std = float("0.0368081") data = None class Program_weight_tensor_parameter_349: name = "parameter_349" + original_name = "linear_70.w_0" shape = [1024, 4096] dtype = "float32" - min_val = float("-0.102697") - max_val = float("0.106561") - mean = float("-1.11099e-05") - std = float("0.0200002") + min_val = float("-0.261719") + max_val = float("0.24231") + mean = float("-7.60152e-05") + std = float("0.0370367") data = None class Program_weight_tensor_parameter_350: name = "parameter_350" + original_name = "layer_norm_23.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.637207") + max_val = float("0.49707") + mean = float("0.000320229") + std = float("0.0660371") data = None class Program_weight_tensor_parameter_351: name = "parameter_351" + original_name = "layer_norm_23.w_0" shape = [1024] dtype = "float32" - min_val = float("1.0") - max_val = float("1.0") - mean = float("1.0") + min_val = float("0.928223") + max_val = float("1.03516") + mean = float("0.947912") + std = float("0.00794533") data = None class Program_weight_tensor_parameter_352: name = "parameter_352" + original_name = "linear_69.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.196655") + max_val = float("0.15918") + mean = float("-0.00013439") + std = float("0.0326094") data = None class Program_weight_tensor_parameter_353: name = "parameter_353" + original_name = "linear_69.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.0987179") - max_val = float("0.0987132") - mean = float("3.34531e-05") - std = float("0.0199919") + min_val = float("-0.290771") + max_val = float("0.30835") + mean = float("7.00671e-06") + std = float("0.0291129") data = None class Program_weight_tensor_parameter_354: name = "parameter_354" + original_name = "linear_68.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.155396") + max_val = float("0.173096") + mean = float("-0.00348107") + std = float("0.0592561") data = None class Program_weight_tensor_parameter_355: name = "parameter_355" + original_name = "linear_68.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.105429") - max_val = float("0.100721") - mean = float("3.39269e-05") - std = float("0.0200355") + min_val = float("-0.192871") + max_val = float("0.171387") + mean = float("4.90223e-05") + std = float("0.0359445") data = None class Program_weight_tensor_parameter_356: name = "parameter_356" + original_name = "linear_67.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.13623") + max_val = float("0.162842") + mean = float("8.54137e-05") + std = float("0.0272196") data = None class Program_weight_tensor_parameter_357: name = "parameter_357" + original_name = "linear_67.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.0983065") - max_val = float("0.103653") - mean = float("1.07552e-06") - std = float("0.020017") + min_val = float("-0.166626") + max_val = float("0.144409") + mean = float("-2.58679e-05") + std = float("0.0324919") data = None class Program_weight_tensor_parameter_358: name = "parameter_358" + original_name = "linear_66.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.268311") + max_val = float("0.264648") + mean = float("0.00191343") + std = float("0.102959") data = None class Program_weight_tensor_parameter_359: name = "parameter_359" + original_name = "linear_66.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.0919331") - max_val = float("0.0933776") - mean = float("1.38431e-05") - std = float("0.0200012") + min_val = float("-0.197266") + max_val = float("0.197998") + mean = float("-8.61982e-06") + std = float("0.0355623") data = None class Program_weight_tensor_parameter_360: name = "parameter_360" + original_name = "layer_norm_22.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.492432") + max_val = float("0.415039") + mean = float("-0.00239725") + std = float("0.0678067") data = None class Program_weight_tensor_parameter_361: name = "parameter_361" + original_name = "layer_norm_22.w_0" shape = [1024] dtype = "float32" - min_val = float("1.0") - max_val = float("1.0") - mean = float("1.0") + min_val = float("0.847168") + max_val = float("1.03711") + mean = float("0.960694") + std = float("0.00932765") data = None class Program_weight_tensor_parameter_362: name = "parameter_362" + original_name = "linear_65.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.226685") + max_val = float("0.1604") + mean = float("-0.000432078") + std = float("0.04833") data = None class Program_weight_tensor_parameter_363: name = "parameter_363" + original_name = "linear_65.w_0" shape = [4096, 1024] dtype = "float32" - min_val = float("-0.111633") - max_val = float("0.10628") - mean = float("-8.28941e-06") - std = float("0.0200021") + min_val = float("-0.624512") + max_val = float("0.682129") + mean = float("-4.43721e-06") + std = float("0.0306632") data = None class Program_weight_tensor_parameter_364: name = "parameter_364" + original_name = "linear_64.b_0" shape = [4096] dtype = "float32" + min_val = float("-0.142212") + max_val = float("0.157959") + mean = float("-0.0298145") + std = float("0.0352929") data = None class Program_weight_tensor_parameter_365: name = "parameter_365" + original_name = "linear_64.w_0" shape = [1024, 4096] dtype = "float32" - min_val = float("-0.106146") - max_val = float("0.108352") - mean = float("1.10053e-05") - std = float("0.020005") + min_val = float("-0.257812") + max_val = float("0.203003") + mean = float("7.41007e-07") + std = float("0.0374019") data = None class Program_weight_tensor_parameter_366: name = "parameter_366" + original_name = "layer_norm_21.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.449951") + max_val = float("0.356934") + mean = float("-0.00147732") + std = float("0.0808913") data = None class Program_weight_tensor_parameter_367: name = "parameter_367" + original_name = "layer_norm_21.w_0" shape = [1024] dtype = "float32" - min_val = float("1.0") - max_val = float("1.0") - mean = float("1.0") + min_val = float("0.928223") + max_val = float("1.0293") + mean = float("0.947881") + std = float("0.00847948") data = None class Program_weight_tensor_parameter_368: name = "parameter_368" + original_name = "linear_63.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.154053") + max_val = float("0.18335") + mean = float("0.000197993") + std = float("0.0289427") data = None class Program_weight_tensor_parameter_369: name = "parameter_369" + original_name = "linear_63.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.0962445") - max_val = float("0.100698") - mean = float("-5.98533e-06") - std = float("0.0200153") + min_val = float("-0.313965") + max_val = float("0.320801") + mean = float("-6.56826e-06") + std = float("0.0323046") data = None class Program_weight_tensor_parameter_370: name = "parameter_370" + original_name = "linear_62.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.200439") + max_val = float("0.229004") + mean = float("0.00319841") + std = float("0.0597114") data = None class Program_weight_tensor_parameter_371: name = "parameter_371" + original_name = "linear_62.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.10165") - max_val = float("0.0968901") - mean = float("2.19517e-06") - std = float("0.0199986") + min_val = float("-0.141479") + max_val = float("0.140381") + mean = float("-1.16073e-05") + std = float("0.0335006") data = None class Program_weight_tensor_parameter_372: name = "parameter_372" + original_name = "linear_61.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.0933838") + max_val = float("0.13147") + mean = float("0.00103397") + std = float("0.0285467") data = None class Program_weight_tensor_parameter_373: name = "parameter_373" + original_name = "linear_61.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.095301") - max_val = float("0.0951553") - mean = float("-3.61915e-05") - std = float("0.0200034") + min_val = float("-0.154419") + max_val = float("0.149292") + mean = float("-4.98463e-06") + std = float("0.0350574") data = None class Program_weight_tensor_parameter_374: name = "parameter_374" + original_name = "linear_60.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.258057") + max_val = float("0.237793") + mean = float("-0.000386028") + std = float("0.0676746") data = None class Program_weight_tensor_parameter_375: name = "parameter_375" + original_name = "linear_60.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.106467") - max_val = float("0.0928625") - mean = float("3.15848e-05") - std = float("0.0199846") + min_val = float("-0.250732") + max_val = float("0.217407") + mean = float("-4.34083e-06") + std = float("0.033369") data = None class Program_weight_tensor_parameter_376: name = "parameter_376" + original_name = "layer_norm_20.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.507324") + max_val = float("0.343262") + mean = float("-0.00322494") + std = float("0.0635818") data = None class Program_weight_tensor_parameter_377: name = "parameter_377" + original_name = "layer_norm_20.w_0" shape = [1024] dtype = "float32" - min_val = float("1.0") - max_val = float("1.0") - mean = float("1.0") + min_val = float("0.792969") + max_val = float("1.02637") + mean = float("0.965108") + std = float("0.0126219") data = None class Program_weight_tensor_parameter_378: name = "parameter_378" + original_name = "linear_59.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.205933") + max_val = float("0.1521") + mean = float("0.000216002") + std = float("0.0518186") data = None class Program_weight_tensor_parameter_379: name = "parameter_379" + original_name = "linear_59.w_0" shape = [4096, 1024] dtype = "float32" - min_val = float("-0.0976331") - max_val = float("0.0991648") - mean = float("7.93172e-06") - std = float("0.019996") + min_val = float("-0.785645") + max_val = float("0.742188") + mean = float("-1.0929e-05") + std = float("0.0316186") data = None class Program_weight_tensor_parameter_380: name = "parameter_380" + original_name = "linear_58.b_0" shape = [4096] dtype = "float32" + min_val = float("-0.147461") + max_val = float("0.158447") + mean = float("-0.0280018") + std = float("0.0332087") data = None class Program_weight_tensor_parameter_381: name = "parameter_381" + original_name = "linear_58.w_0" shape = [1024, 4096] dtype = "float32" - min_val = float("-0.0999039") - max_val = float("0.0986416") - mean = float("-4.63473e-06") - std = float("0.0200024") + min_val = float("-0.243286") + max_val = float("0.226929") + mean = float("6.68727e-06") + std = float("0.0365013") data = None class Program_weight_tensor_parameter_382: name = "parameter_382" + original_name = "layer_norm_19.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.458252") + max_val = float("0.360107") + mean = float("-0.0010696") + std = float("0.0897329") data = None class Program_weight_tensor_parameter_383: name = "parameter_383" + original_name = "layer_norm_19.w_0" shape = [1024] dtype = "float32" - min_val = float("1.0") - max_val = float("1.0") - mean = float("1.0") + min_val = float("0.916504") + max_val = float("1.03613") + mean = float("0.947683") + std = float("0.00943618") data = None class Program_weight_tensor_parameter_384: name = "parameter_384" + original_name = "linear_57.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.175781") + max_val = float("0.139893") + mean = float("6.26309e-05") + std = float("0.0433647") data = None class Program_weight_tensor_parameter_385: name = "parameter_385" + original_name = "linear_57.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.095238") - max_val = float("0.102729") - mean = float("1.36443e-05") - std = float("0.0199918") + min_val = float("-0.246216") + max_val = float("0.238281") + mean = float("7.48987e-06") + std = float("0.0316142") data = None class Program_weight_tensor_parameter_386: name = "parameter_386" + original_name = "linear_56.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.150513") + max_val = float("0.150513") + mean = float("0.00278748") + std = float("0.0502907") data = None class Program_weight_tensor_parameter_387: name = "parameter_387" + original_name = "linear_56.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.0975117") - max_val = float("0.0892062") - mean = float("8.21871e-06") - std = float("0.019997") + min_val = float("-0.152222") + max_val = float("0.157593") + mean = float("8.61042e-06") + std = float("0.0347166") data = None class Program_weight_tensor_parameter_388: name = "parameter_388" + original_name = "linear_55.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.135376") + max_val = float("0.131958") + mean = float("0.00105707") + std = float("0.0245789") data = None class Program_weight_tensor_parameter_389: name = "parameter_389" + original_name = "linear_55.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.103816") - max_val = float("0.0932822") - mean = float("-3.86739e-05") - std = float("0.020014") + min_val = float("-0.144653") + max_val = float("0.144531") + mean = float("-5.92249e-06") + std = float("0.0338656") data = None class Program_weight_tensor_parameter_390: name = "parameter_390" + original_name = "linear_54.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.256836") + max_val = float("0.248657") + mean = float("0.00221521") + std = float("0.0539023") data = None class Program_weight_tensor_parameter_391: name = "parameter_391" + original_name = "linear_54.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.101861") - max_val = float("0.108144") - mean = float("-1.35054e-05") - std = float("0.0200085") + min_val = float("-0.2323") + max_val = float("0.199341") + mean = float("-2.25472e-05") + std = float("0.0341958") data = None class Program_weight_tensor_parameter_392: name = "parameter_392" + original_name = "layer_norm_18.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.515137") + max_val = float("0.345459") + mean = float("-0.00730822") + std = float("0.0560214") data = None class Program_weight_tensor_parameter_393: name = "parameter_393" + original_name = "layer_norm_18.w_0" shape = [1024] dtype = "float32" - min_val = float("1.0") - max_val = float("1.0") - mean = float("1.0") + min_val = float("0.838867") + max_val = float("1.01367") + mean = float("0.966789") + std = float("0.0094262") data = None class Program_weight_tensor_parameter_394: name = "parameter_394" + original_name = "linear_53.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.248413") + max_val = float("0.165771") + mean = float("-4.61696e-05") + std = float("0.0522568") data = None class Program_weight_tensor_parameter_395: name = "parameter_395" + original_name = "linear_53.w_0" shape = [4096, 1024] dtype = "float32" - min_val = float("-0.102811") - max_val = float("0.098902") - mean = float("-4.73518e-06") - std = float("0.0199993") + min_val = float("-0.750977") + max_val = float("0.700684") + mean = float("-1.15972e-05") + std = float("0.0324169") data = None class Program_weight_tensor_parameter_396: name = "parameter_396" + original_name = "linear_52.b_0" shape = [4096] dtype = "float32" + min_val = float("-0.133423") + max_val = float("0.153442") + mean = float("-0.0303658") + std = float("0.0329256") data = None class Program_weight_tensor_parameter_397: name = "parameter_397" + original_name = "linear_52.w_0" shape = [1024, 4096] dtype = "float32" - min_val = float("-0.110147") - max_val = float("0.100048") - mean = float("2.74642e-06") - std = float("0.0199955") + min_val = float("-0.249756") + max_val = float("0.196533") + mean = float("-5.40115e-05") + std = float("0.0361284") data = None class Program_weight_tensor_parameter_398: name = "parameter_398" + original_name = "layer_norm_17.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.477051") + max_val = float("0.364014") + mean = float("-0.000599697") + std = float("0.0990258") data = None class Program_weight_tensor_parameter_399: name = "parameter_399" + original_name = "layer_norm_17.w_0" shape = [1024] dtype = "float32" - min_val = float("1.0") - max_val = float("1.0") - mean = float("1.0") + min_val = float("0.908691") + max_val = float("1.08496") + mean = float("0.948136") + std = float("0.00962775") data = None class Program_weight_tensor_parameter_400: name = "parameter_400" + original_name = "linear_51.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.197388") + max_val = float("0.178345") + mean = float("0.000330423") + std = float("0.0376253") data = None class Program_weight_tensor_parameter_401: name = "parameter_401" + original_name = "linear_51.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.103979") - max_val = float("0.0978683") - mean = float("8.37984e-06") - std = float("0.0200129") + min_val = float("-0.306885") + max_val = float("0.333008") + mean = float("1.01402e-06") + std = float("0.0323392") data = None class Program_weight_tensor_parameter_402: name = "parameter_402" + original_name = "linear_50.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.198486") + max_val = float("0.208252") + mean = float("-0.00273028") + std = float("0.0570606") data = None class Program_weight_tensor_parameter_403: name = "parameter_403" + original_name = "linear_50.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.0947675") - max_val = float("0.0976186") - mean = float("2.18974e-05") - std = float("0.0199866") + min_val = float("-0.145508") + max_val = float("0.139893") + mean = float("-1.88165e-06") + std = float("0.0334635") data = None class Program_weight_tensor_parameter_404: name = "parameter_404" + original_name = "linear_49.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.129272") + max_val = float("0.0943604") + mean = float("1.36403e-05") + std = float("0.0214269") data = None class Program_weight_tensor_parameter_405: name = "parameter_405" + original_name = "linear_49.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.0951906") - max_val = float("0.0923431") - mean = float("1.84878e-05") - std = float("0.0200122") + min_val = float("-0.159058") + max_val = float("0.153687") + mean = float("-3.64445e-05") + std = float("0.0348143") data = None class Program_weight_tensor_parameter_406: name = "parameter_406" + original_name = "linear_48.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.244385") + max_val = float("0.244751") + mean = float("-0.00203093") + std = float("0.03992") data = None class Program_weight_tensor_parameter_407: name = "parameter_407" + original_name = "linear_48.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.0905452") - max_val = float("0.101149") - mean = float("-1.44355e-05") - std = float("0.0200056") + min_val = float("-0.249023") + max_val = float("0.255615") + mean = float("-1.43333e-05") + std = float("0.0330586") data = None class Program_weight_tensor_parameter_408: name = "parameter_408" + original_name = "layer_norm_16.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.432373") + max_val = float("0.289062") + mean = float("-0.0111983") + std = float("0.0672973") data = None class Program_weight_tensor_parameter_409: name = "parameter_409" + original_name = "layer_norm_16.w_0" shape = [1024] dtype = "float32" - min_val = float("1.0") - max_val = float("1.0") - mean = float("1.0") + min_val = float("0.90918") + max_val = float("0.996582") + mean = float("0.967111") + std = float("0.00621475") data = None class Program_weight_tensor_parameter_410: name = "parameter_410" + original_name = "linear_47.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.203613") + max_val = float("0.144409") + mean = float("0.000351486") + std = float("0.0506849") data = None class Program_weight_tensor_parameter_411: name = "parameter_411" + original_name = "linear_47.w_0" shape = [4096, 1024] dtype = "float32" - min_val = float("-0.0991167") - max_val = float("0.0982659") - mean = float("-1.85633e-06") - std = float("0.0199953") + min_val = float("-0.712402") + max_val = float("0.62207") + mean = float("-1.24704e-05") + std = float("0.0334039") data = None class Program_weight_tensor_parameter_412: name = "parameter_412" + original_name = "linear_46.b_0" shape = [4096] dtype = "float32" + min_val = float("-0.141235") + max_val = float("0.138428") + mean = float("-0.031518") + std = float("0.0282451") data = None class Program_weight_tensor_parameter_413: name = "parameter_413" + original_name = "linear_46.w_0" shape = [1024, 4096] dtype = "float32" - min_val = float("-0.112306") - max_val = float("0.102346") - mean = float("8.72672e-06") - std = float("0.0199986") + min_val = float("-0.231445") + max_val = float("0.203125") + mean = float("-9.01988e-05") + std = float("0.03683") data = None class Program_weight_tensor_parameter_414: name = "parameter_414" + original_name = "layer_norm_15.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.536621") + max_val = float("0.376953") + mean = float("0.000161601") + std = float("0.100089") data = None class Program_weight_tensor_parameter_415: name = "parameter_415" + original_name = "layer_norm_15.w_0" shape = [1024] dtype = "float32" - min_val = float("1.0") - max_val = float("1.0") - mean = float("1.0") + min_val = float("0.916016") + max_val = float("1.14648") + mean = float("0.947086") + std = float("0.0113242") data = None class Program_weight_tensor_parameter_416: name = "parameter_416" + original_name = "linear_45.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.235596") + max_val = float("0.231079") + mean = float("0.000222951") + std = float("0.0342212") data = None class Program_weight_tensor_parameter_417: name = "parameter_417" + original_name = "linear_45.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.0956527") - max_val = float("0.100001") - mean = float("-8.76764e-07") - std = float("0.020012") + min_val = float("-0.285645") + max_val = float("0.298584") + mean = float("-3.17309e-06") + std = float("0.0314163") data = None class Program_weight_tensor_parameter_418: name = "parameter_418" + original_name = "linear_44.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.192505") + max_val = float("0.148682") + mean = float("-0.000102659") + std = float("0.0610946") data = None class Program_weight_tensor_parameter_419: name = "parameter_419" + original_name = "linear_44.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.0944597") - max_val = float("0.0978589") - mean = float("2.0862e-05") - std = float("0.0200228") + min_val = float("-0.140381") + max_val = float("0.144287") + mean = float("-2.79974e-06") + std = float("0.03273") data = None class Program_weight_tensor_parameter_420: name = "parameter_420" + original_name = "linear_43.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.132935") + max_val = float("0.069519") + mean = float("0.000201435") + std = float("0.0179718") data = None class Program_weight_tensor_parameter_421: name = "parameter_421" + original_name = "linear_43.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.0917877") - max_val = float("0.105241") - mean = float("1.62806e-05") - std = float("0.0200282") + min_val = float("-0.13855") + max_val = float("0.150635") + mean = float("-2.35134e-05") + std = float("0.0348853") data = None class Program_weight_tensor_parameter_422: name = "parameter_422" + original_name = "linear_42.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.23584") + max_val = float("0.150146") + mean = float("-0.00209914") + std = float("0.0316112") data = None class Program_weight_tensor_parameter_423: name = "parameter_423" + original_name = "linear_42.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.100093") - max_val = float("0.108774") - mean = float("4.95743e-05") - std = float("0.0200085") + min_val = float("-0.221924") + max_val = float("0.214722") + mean = float("3.28411e-05") + std = float("0.0320514") data = None class Program_weight_tensor_parameter_424: name = "parameter_424" + original_name = "layer_norm_14.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.487305") + max_val = float("0.303711") + mean = float("-0.0113358") + std = float("0.0726874") data = None class Program_weight_tensor_parameter_425: name = "parameter_425" + original_name = "layer_norm_14.w_0" shape = [1024] dtype = "float32" - min_val = float("1.0") + min_val = float("0.895508") max_val = float("1.0") - mean = float("1.0") + mean = float("0.962132") + std = float("0.00626821") data = None class Program_weight_tensor_parameter_426: name = "parameter_426" + original_name = "linear_41.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.206055") + max_val = float("0.124817") + mean = float("7.19376e-05") + std = float("0.0461399") data = None class Program_weight_tensor_parameter_427: name = "parameter_427" + original_name = "linear_41.w_0" shape = [4096, 1024] dtype = "float32" - min_val = float("-0.10039") - max_val = float("0.104795") - mean = float("9.7877e-06") - std = float("0.0200006") + min_val = float("-0.667969") + max_val = float("0.52832") + mean = float("2.80091e-06") + std = float("0.0338533") data = None class Program_weight_tensor_parameter_428: name = "parameter_428" + original_name = "linear_40.b_0" shape = [4096] dtype = "float32" + min_val = float("-0.153076") + max_val = float("0.129028") + mean = float("-0.0315053") + std = float("0.0257946") data = None class Program_weight_tensor_parameter_429: name = "parameter_429" + original_name = "linear_40.w_0" shape = [1024, 4096] dtype = "float32" - min_val = float("-0.0993053") - max_val = float("0.0993167") - mean = float("3.41737e-06") - std = float("0.0199956") + min_val = float("-0.26123") + max_val = float("0.207886") + mean = float("5.26586e-05") + std = float("0.0365606") data = None class Program_weight_tensor_parameter_430: name = "parameter_430" + original_name = "layer_norm_13.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.52832") + max_val = float("0.298584") + mean = float("-0.00191944") + std = float("0.106183") data = None class Program_weight_tensor_parameter_431: name = "parameter_431" + original_name = "layer_norm_13.w_0" shape = [1024] dtype = "float32" - min_val = float("1.0") - max_val = float("1.0") - mean = float("1.0") + min_val = float("0.920898") + max_val = float("1.12988") + mean = float("0.948469") + std = float("0.010686") data = None class Program_weight_tensor_parameter_432: name = "parameter_432" + original_name = "linear_39.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.230591") + max_val = float("0.151489") + mean = float("0.000343485") + std = float("0.0441325") data = None class Program_weight_tensor_parameter_433: name = "parameter_433" + original_name = "linear_39.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.101647") - max_val = float("0.0953956") - mean = float("-1.33769e-05") - std = float("0.0199901") + min_val = float("-0.293945") + max_val = float("0.283936") + mean = float("1.32653e-05") + std = float("0.030798") data = None class Program_weight_tensor_parameter_434: name = "parameter_434" + original_name = "linear_38.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.191895") + max_val = float("0.143433") + mean = float("0.00348649") + std = float("0.0567407") data = None class Program_weight_tensor_parameter_435: name = "parameter_435" + original_name = "linear_38.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.0927683") - max_val = float("0.100082") - mean = float("-1.29715e-05") - std = float("0.0199813") + min_val = float("-0.146118") + max_val = float("0.144165") + mean = float("-3.60339e-05") + std = float("0.0325118") data = None class Program_weight_tensor_parameter_436: name = "parameter_436" + original_name = "linear_37.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.065979") + max_val = float("0.0675049") + mean = float("-0.000866723") + std = float("0.0188837") data = None class Program_weight_tensor_parameter_437: name = "parameter_437" + original_name = "linear_37.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.093213") - max_val = float("0.106487") - mean = float("2.38763e-05") - std = float("0.0199641") + min_val = float("-0.139771") + max_val = float("0.154053") + mean = float("1.12351e-05") + std = float("0.0340043") data = None class Program_weight_tensor_parameter_438: name = "parameter_438" + original_name = "linear_36.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.184448") + max_val = float("0.165039") + mean = float("-0.00119359") + std = float("0.0350682") data = None class Program_weight_tensor_parameter_439: name = "parameter_439" + original_name = "linear_36.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.100074") - max_val = float("0.0930649") - mean = float("2.25092e-05") - std = float("0.0199914") + min_val = float("-0.208496") + max_val = float("0.211548") + mean = float("5.80237e-07") + std = float("0.03247") data = None class Program_weight_tensor_parameter_440: name = "parameter_440" + original_name = "layer_norm_12.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.516113") + max_val = float("0.292969") + mean = float("-0.00740127") + std = float("0.0761911") data = None class Program_weight_tensor_parameter_441: name = "parameter_441" + original_name = "layer_norm_12.w_0" shape = [1024] dtype = "float32" - min_val = float("1.0") - max_val = float("1.0") - mean = float("1.0") + min_val = float("0.86377") + max_val = float("0.993164") + mean = float("0.965313") + std = float("0.00643953") data = None class Program_weight_tensor_parameter_442: name = "parameter_442" + original_name = "linear_35.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.158325") + max_val = float("0.135132") + mean = float("0.000126876") + std = float("0.043139") data = None class Program_weight_tensor_parameter_443: name = "parameter_443" + original_name = "linear_35.w_0" shape = [4096, 1024] dtype = "float32" - min_val = float("-0.0957292") - max_val = float("0.099185") - mean = float("3.23641e-06") - std = float("0.0199912") + min_val = float("-0.807129") + max_val = float("0.501465") + mean = float("9.42509e-06") + std = float("0.0338706") data = None class Program_weight_tensor_parameter_444: name = "parameter_444" + original_name = "linear_34.b_0" shape = [4096] dtype = "float32" + min_val = float("-0.144653") + max_val = float("0.100342") + mean = float("-0.0321978") + std = float("0.0226705") data = None class Program_weight_tensor_parameter_445: name = "parameter_445" + original_name = "linear_34.w_0" shape = [1024, 4096] dtype = "float32" - min_val = float("-0.101875") - max_val = float("0.105822") - mean = float("-1.76224e-06") - std = float("0.0200038") + min_val = float("-0.231079") + max_val = float("0.184814") + mean = float("8.64958e-05") + std = float("0.036274") data = None class Program_weight_tensor_parameter_446: name = "parameter_446" + original_name = "layer_norm_11.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.54541") + max_val = float("0.328857") + mean = float("-0.00185309") + std = float("0.109378") data = None class Program_weight_tensor_parameter_447: name = "parameter_447" + original_name = "layer_norm_11.w_0" shape = [1024] dtype = "float32" - min_val = float("1.0") - max_val = float("1.0") - mean = float("1.0") + min_val = float("0.929199") + max_val = float("1.07031") + mean = float("0.950308") + std = float("0.00941139") data = None class Program_weight_tensor_parameter_448: name = "parameter_448" + original_name = "linear_33.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.216431") + max_val = float("0.123108") + mean = float("0.000372993") + std = float("0.0479976") data = None class Program_weight_tensor_parameter_449: name = "parameter_449" + original_name = "linear_33.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.0923893") - max_val = float("0.097908") - mean = float("-5.97937e-07") - std = float("0.0199918") + min_val = float("-0.265137") + max_val = float("0.230347") + mean = float("7.57999e-07") + std = float("0.028029") data = None class Program_weight_tensor_parameter_450: name = "parameter_450" + original_name = "linear_32.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.153564") + max_val = float("0.16394") + mean = float("-0.00154176") + std = float("0.0493504") data = None class Program_weight_tensor_parameter_451: name = "parameter_451" + original_name = "linear_32.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.0943266") - max_val = float("0.0993989") - mean = float("-1.24631e-05") - std = float("0.0200019") + min_val = float("-0.199829") + max_val = float("0.183716") + mean = float("1.91652e-05") + std = float("0.0341771") data = None class Program_weight_tensor_parameter_452: name = "parameter_452" + original_name = "linear_31.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.0783691") + max_val = float("0.121277") + mean = float("0.000587615") + std = float("0.0217441") data = None class Program_weight_tensor_parameter_453: name = "parameter_453" + original_name = "linear_31.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.105057") - max_val = float("0.105276") - mean = float("-8.65849e-06") - std = float("0.020002") + min_val = float("-0.138428") + max_val = float("0.134888") + mean = float("-8.8477e-06") + std = float("0.0307705") data = None class Program_weight_tensor_parameter_454: name = "parameter_454" + original_name = "linear_30.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.159058") + max_val = float("0.164185") + mean = float("0.000531139") + std = float("0.0295991") data = None class Program_weight_tensor_parameter_455: name = "parameter_455" + original_name = "linear_30.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.0945603") - max_val = float("0.105948") - mean = float("-4.1908e-05") - std = float("0.0200044") + min_val = float("-0.186279") + max_val = float("0.198853") + mean = float("1.17172e-06") + std = float("0.0339334") data = None class Program_weight_tensor_parameter_456: name = "parameter_456" + original_name = "layer_norm_10.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.443359") + max_val = float("0.189941") + mean = float("-0.000672378") + std = float("0.0785599") data = None class Program_weight_tensor_parameter_457: name = "parameter_457" + original_name = "layer_norm_10.w_0" shape = [1024] dtype = "float32" - min_val = float("1.0") - max_val = float("1.0") - mean = float("1.0") + min_val = float("0.859375") + max_val = float("1.00195") + mean = float("0.965514") + std = float("0.00685749") data = None class Program_weight_tensor_parameter_458: name = "parameter_458" + original_name = "linear_29.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.148193") + max_val = float("0.133667") + mean = float("1.79122e-05") + std = float("0.0392385") data = None class Program_weight_tensor_parameter_459: name = "parameter_459" + original_name = "linear_29.w_0" shape = [4096, 1024] dtype = "float32" - min_val = float("-0.0979283") - max_val = float("0.103085") - mean = float("-1.30321e-06") - std = float("0.0200036") + min_val = float("-0.695801") + max_val = float("0.449707") + mean = float("1.464e-05") + std = float("0.0343405") data = None class Program_weight_tensor_parameter_460: name = "parameter_460" + original_name = "linear_28.b_0" shape = [4096] dtype = "float32" + min_val = float("-0.140991") + max_val = float("0.12561") + mean = float("-0.0315909") + std = float("0.0236769") data = None class Program_weight_tensor_parameter_461: name = "parameter_461" + original_name = "linear_28.w_0" shape = [1024, 4096] dtype = "float32" - min_val = float("-0.106927") - max_val = float("0.0984713") - mean = float("-4.55845e-06") - std = float("0.0199977") + min_val = float("-0.227417") + max_val = float("0.178589") + mean = float("-3.66555e-05") + std = float("0.0366089") data = None class Program_weight_tensor_parameter_462: name = "parameter_462" + original_name = "layer_norm_9.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.443604") + max_val = float("0.364258") + mean = float("3.89351e-05") + std = float("0.112748") data = None class Program_weight_tensor_parameter_463: name = "parameter_463" + original_name = "layer_norm_9.w_0" shape = [1024] dtype = "float32" - min_val = float("1.0") - max_val = float("1.0") - mean = float("1.0") + min_val = float("0.935059") + max_val = float("1.06348") + mean = float("0.952088") + std = float("0.00814208") data = None class Program_weight_tensor_parameter_464: name = "parameter_464" + original_name = "linear_27.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.160156") + max_val = float("0.133667") + mean = float("0.000401551") + std = float("0.0438754") data = None class Program_weight_tensor_parameter_465: name = "parameter_465" + original_name = "linear_27.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.0998297") - max_val = float("0.0963412") - mean = float("5.75533e-05") - std = float("0.0199767") + min_val = float("-0.189575") + max_val = float("0.214355") + mean = float("5.21725e-06") + std = float("0.0274209") data = None class Program_weight_tensor_parameter_466: name = "parameter_466" + original_name = "linear_26.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.13855") + max_val = float("0.109802") + mean = float("-0.00102558") + std = float("0.0334611") data = None class Program_weight_tensor_parameter_467: name = "parameter_467" + original_name = "linear_26.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.0941039") - max_val = float("0.0932667") - mean = float("2.67678e-05") - std = float("0.0199896") + min_val = float("-0.195801") + max_val = float("0.190552") + mean = float("1.18238e-05") + std = float("0.037996") data = None class Program_weight_tensor_parameter_468: name = "parameter_468" + original_name = "linear_25.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.0916138") + max_val = float("0.105103") + mean = float("0.000390278") + std = float("0.0195086") data = None class Program_weight_tensor_parameter_469: name = "parameter_469" + original_name = "linear_25.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.0969902") - max_val = float("0.1042") - mean = float("-5.98067e-06") - std = float("0.0200062") + min_val = float("-0.143188") + max_val = float("0.143188") + mean = float("-1.49328e-05") + std = float("0.0289919") data = None class Program_weight_tensor_parameter_470: name = "parameter_470" + original_name = "linear_24.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.162354") + max_val = float("0.168091") + mean = float("-0.000479993") + std = float("0.0285619") data = None class Program_weight_tensor_parameter_471: name = "parameter_471" + original_name = "linear_24.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.100687") - max_val = float("0.0994866") - mean = float("-3.56567e-06") - std = float("0.0200013") + min_val = float("-0.188477") + max_val = float("0.186279") + mean = float("-1.3847e-05") + std = float("0.0379183") data = None class Program_weight_tensor_parameter_472: name = "parameter_472" + original_name = "layer_norm_8.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.224121") + max_val = float("0.30957") + mean = float("0.00255697") + std = float("0.0715734") data = None class Program_weight_tensor_parameter_473: name = "parameter_473" + original_name = "layer_norm_8.w_0" shape = [1024] dtype = "float32" - min_val = float("1.0") - max_val = float("1.0") - mean = float("1.0") + min_val = float("0.891113") + max_val = float("0.996582") + mean = float("0.969747") + std = float("0.00624836") data = None class Program_weight_tensor_parameter_474: name = "parameter_474" + original_name = "linear_23.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.116821") + max_val = float("0.108948") + mean = float("-0.000144747") + std = float("0.024838") data = None class Program_weight_tensor_parameter_475: name = "parameter_475" + original_name = "linear_23.w_0" shape = [4096, 1024] dtype = "float32" - min_val = float("-0.105598") - max_val = float("0.106107") - mean = float("-4.1694e-06") - std = float("0.0200048") + min_val = float("-0.584961") + max_val = float("0.464111") + mean = float("2.77091e-06") + std = float("0.0357016") data = None class Program_weight_tensor_parameter_476: name = "parameter_476" + original_name = "linear_22.b_0" shape = [4096] dtype = "float32" + min_val = float("-0.142456") + max_val = float("0.0545654") + mean = float("-0.0291114") + std = float("0.0195116") data = None class Program_weight_tensor_parameter_477: name = "parameter_477" + original_name = "linear_22.w_0" shape = [1024, 4096] dtype = "float32" - min_val = float("-0.0978924") - max_val = float("0.109131") - mean = float("-4.1312e-06") - std = float("0.0200027") + min_val = float("-0.205322") + max_val = float("0.215332") + mean = float("-3.75746e-05") + std = float("0.0366461") data = None class Program_weight_tensor_parameter_478: name = "parameter_478" + original_name = "layer_norm_7.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.346436") + max_val = float("0.36499") + mean = float("0.00184274") + std = float("0.106411") data = None class Program_weight_tensor_parameter_479: name = "parameter_479" + original_name = "layer_norm_7.w_0" shape = [1024] dtype = "float32" - min_val = float("1.0") - max_val = float("1.0") - mean = float("1.0") + min_val = float("0.929688") + max_val = float("1.06055") + mean = float("0.947949") + std = float("0.00811328") data = None class Program_weight_tensor_parameter_480: name = "parameter_480" + original_name = "linear_21.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.125366") + max_val = float("0.185303") + mean = float("0.0002676") + std = float("0.050709") data = None class Program_weight_tensor_parameter_481: name = "parameter_481" + original_name = "linear_21.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.101286") - max_val = float("0.101403") - mean = float("-2.16372e-05") - std = float("0.02") + min_val = float("-0.250977") + max_val = float("0.246338") + mean = float("1.21174e-06") + std = float("0.0255601") data = None class Program_weight_tensor_parameter_482: name = "parameter_482" + original_name = "linear_20.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.125366") + max_val = float("0.129761") + mean = float("-0.00142984") + std = float("0.028807") data = None class Program_weight_tensor_parameter_483: name = "parameter_483" + original_name = "linear_20.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.113911") - max_val = float("0.0975718") - mean = float("-1.28185e-06") - std = float("0.0200102") + min_val = float("-0.15332") + max_val = float("0.164307") + mean = float("-1.36974e-05") + std = float("0.0346356") data = None class Program_weight_tensor_parameter_484: name = "parameter_484" + original_name = "linear_19.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.0667114") + max_val = float("0.0656738") + mean = float("0.000935968") + std = float("0.0163501") data = None class Program_weight_tensor_parameter_485: name = "parameter_485" + original_name = "linear_19.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.0997715") - max_val = float("0.10386") - mean = float("-1.88894e-05") - std = float("0.020004") + min_val = float("-0.131836") + max_val = float("0.127686") + mean = float("-7.7012e-06") + std = float("0.0277387") data = None class Program_weight_tensor_parameter_486: name = "parameter_486" + original_name = "linear_18.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.0966797") + max_val = float("0.125") + mean = float("-0.000312715") + std = float("0.0229982") data = None class Program_weight_tensor_parameter_487: name = "parameter_487" + original_name = "linear_18.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.109369") - max_val = float("0.0947769") - mean = float("1.80859e-05") - std = float("0.0200146") + min_val = float("-0.170044") + max_val = float("0.153931") + mean = float("1.21947e-05") + std = float("0.034743") data = None class Program_weight_tensor_parameter_488: name = "parameter_488" + original_name = "layer_norm_6.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.29248") + max_val = float("0.381104") + mean = float("0.00333835") + std = float("0.0778582") data = None class Program_weight_tensor_parameter_489: name = "parameter_489" + original_name = "layer_norm_6.w_0" shape = [1024] dtype = "float32" - min_val = float("1.0") - max_val = float("1.0") - mean = float("1.0") + min_val = float("0.942383") + max_val = float("0.981445") + mean = float("0.963295") + std = float("0.00434195") data = None class Program_weight_tensor_parameter_490: name = "parameter_490" + original_name = "linear_17.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.101379") + max_val = float("0.152954") + mean = float("-0.000272748") + std = float("0.0255054") data = None class Program_weight_tensor_parameter_491: name = "parameter_491" + original_name = "linear_17.w_0" shape = [4096, 1024] dtype = "float32" - min_val = float("-0.119484") - max_val = float("0.101903") - mean = float("4.90297e-06") - std = float("0.0199966") + min_val = float("-0.313232") + max_val = float("0.415771") + mean = float("-3.64991e-06") + std = float("0.0353684") data = None class Program_weight_tensor_parameter_492: name = "parameter_492" + original_name = "linear_16.b_0" shape = [4096] dtype = "float32" + min_val = float("-0.147217") + max_val = float("0.0557556") + mean = float("-0.029135") + std = float("0.0155605") data = None class Program_weight_tensor_parameter_493: name = "parameter_493" + original_name = "linear_16.w_0" shape = [1024, 4096] dtype = "float32" - min_val = float("-0.0975582") - max_val = float("0.095477") - mean = float("-6.09357e-06") - std = float("0.0199984") + min_val = float("-0.185425") + max_val = float("0.216431") + mean = float("-2.47e-05") + std = float("0.0370089") data = None class Program_weight_tensor_parameter_494: name = "parameter_494" + original_name = "layer_norm_5.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.446045") + max_val = float("0.493164") + mean = float("0.00143073") + std = float("0.0999569") data = None class Program_weight_tensor_parameter_495: name = "parameter_495" + original_name = "layer_norm_5.w_0" shape = [1024] dtype = "float32" - min_val = float("1.0") - max_val = float("1.0") - mean = float("1.0") + min_val = float("0.93457") + max_val = float("1.0625") + mean = float("0.948988") + std = float("0.00773386") data = None class Program_weight_tensor_parameter_496: name = "parameter_496" + original_name = "linear_15.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.163086") + max_val = float("0.17981") + mean = float("5.56792e-05") + std = float("0.0546707") data = None class Program_weight_tensor_parameter_497: name = "parameter_497" + original_name = "linear_15.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.102974") - max_val = float("0.103186") - mean = float("-1.30775e-05") - std = float("0.0199967") + min_val = float("-0.247559") + max_val = float("0.255859") + mean = float("-4.50185e-06") + std = float("0.0237154") data = None class Program_weight_tensor_parameter_498: name = "parameter_498" + original_name = "linear_14.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.133789") + max_val = float("0.162231") + mean = float("0.000103115") + std = float("0.0359187") data = None class Program_weight_tensor_parameter_499: name = "parameter_499" + original_name = "linear_14.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.115446") - max_val = float("0.101411") - mean = float("8.97768e-06") - std = float("0.0200146") + min_val = float("-0.181519") + max_val = float("0.193359") + mean = float("6.25932e-05") + std = float("0.0373285") data = None class Program_weight_tensor_parameter_500: name = "parameter_500" + original_name = "linear_13.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.0597534") + max_val = float("0.0596924") + mean = float("-3.01423e-05") + std = float("0.0130638") data = None class Program_weight_tensor_parameter_501: name = "parameter_501" + original_name = "linear_13.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.0952101") - max_val = float("0.0975177") - mean = float("-1.7404e-06") - std = float("0.0200067") + min_val = float("-0.133667") + max_val = float("0.124634") + mean = float("1.79606e-06") + std = float("0.0257102") data = None class Program_weight_tensor_parameter_502: name = "parameter_502" + original_name = "linear_12.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.0639038") + max_val = float("0.0612793") + mean = float("7.53078e-05") + std = float("0.0144499") data = None class Program_weight_tensor_parameter_503: name = "parameter_503" + original_name = "linear_12.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.0950495") - max_val = float("0.101216") - mean = float("-9.00584e-06") - std = float("0.0199843") + min_val = float("-0.192871") + max_val = float("0.189209") + mean = float("1.0414e-05") + std = float("0.0375774") data = None class Program_weight_tensor_parameter_504: name = "parameter_504" + original_name = "layer_norm_4.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.388672") + max_val = float("0.455811") + mean = float("0.0026274") + std = float("0.0750707") data = None class Program_weight_tensor_parameter_505: name = "parameter_505" + original_name = "layer_norm_4.w_0" shape = [1024] dtype = "float32" - min_val = float("1.0") - max_val = float("1.0") - mean = float("1.0") + min_val = float("0.935059") + max_val = float("0.980469") + mean = float("0.967252") + std = float("0.00393033") data = None class Program_weight_tensor_parameter_506: name = "parameter_506" + original_name = "linear_11.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.100769") + max_val = float("0.144653") + mean = float("0.000408601") + std = float("0.03289") data = None class Program_weight_tensor_parameter_507: name = "parameter_507" + original_name = "linear_11.w_0" shape = [4096, 1024] dtype = "float32" - min_val = float("-0.100961") - max_val = float("0.102036") - mean = float("2.24352e-06") - std = float("0.0199959") + min_val = float("-0.446533") + max_val = float("0.594727") + mean = float("6.18537e-06") + std = float("0.0348272") data = None class Program_weight_tensor_parameter_508: name = "parameter_508" + original_name = "linear_10.b_0" shape = [4096] dtype = "float32" + min_val = float("-0.151611") + max_val = float("0.0605774") + mean = float("-0.0324846") + std = float("0.0201777") data = None class Program_weight_tensor_parameter_509: name = "parameter_509" + original_name = "linear_10.w_0" shape = [1024, 4096] dtype = "float32" - min_val = float("-0.107485") - max_val = float("0.105211") - mean = float("8.14967e-06") - std = float("0.0199918") + min_val = float("-0.164673") + max_val = float("0.193604") + mean = float("-2.90386e-06") + std = float("0.0356818") data = None class Program_weight_tensor_parameter_510: name = "parameter_510" + original_name = "layer_norm_3.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.53418") + max_val = float("0.563477") + mean = float("0.000232919") + std = float("0.122637") data = None class Program_weight_tensor_parameter_511: name = "parameter_511" + original_name = "layer_norm_3.w_0" shape = [1024] dtype = "float32" - min_val = float("1.0") - max_val = float("1.0") - mean = float("1.0") + min_val = float("0.902832") + max_val = float("1.12402") + mean = float("0.943692") + std = float("0.00928432") data = None class Program_weight_tensor_parameter_512: name = "parameter_512" + original_name = "linear_9.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.240601") + max_val = float("0.268799") + mean = float("0.00010442") + std = float("0.0482944") data = None class Program_weight_tensor_parameter_513: name = "parameter_513" + original_name = "linear_9.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.0953353") - max_val = float("0.0968531") - mean = float("2.98822e-05") - std = float("0.0199998") + min_val = float("-0.371094") + max_val = float("0.361084") + mean = float("3.12926e-06") + std = float("0.0247985") data = None class Program_weight_tensor_parameter_514: name = "parameter_514" + original_name = "linear_8.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.172241") + max_val = float("0.184448") + mean = float("0.000718303") + std = float("0.0743942") data = None class Program_weight_tensor_parameter_515: name = "parameter_515" + original_name = "linear_8.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.102585") - max_val = float("0.0961275") - mean = float("1.66904e-05") - std = float("0.0199954") + min_val = float("-0.197632") + max_val = float("0.166382") + mean = float("1.58143e-05") + std = float("0.0328662") data = None class Program_weight_tensor_parameter_516: name = "parameter_516" + original_name = "linear_7.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.071228") + max_val = float("0.06427") + mean = float("8.765e-05") + std = float("0.0151535") data = None class Program_weight_tensor_parameter_517: name = "parameter_517" + original_name = "linear_7.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.0994374") - max_val = float("0.0981079") - mean = float("3.86858e-06") - std = float("0.02001") + min_val = float("-0.123718") + max_val = float("0.119263") + mean = float("7.24318e-06") + std = float("0.0267841") data = None class Program_weight_tensor_parameter_518: name = "parameter_518" + original_name = "linear_6.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.0733032") + max_val = float("0.0714111") + mean = float("-0.000100577") + std = float("0.0191359") data = None class Program_weight_tensor_parameter_519: name = "parameter_519" + original_name = "linear_6.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.0917555") - max_val = float("0.100264") - mean = float("1.41448e-05") - std = float("0.0199943") + min_val = float("-0.145996") + max_val = float("0.142212") + mean = float("-1.23464e-05") + std = float("0.0320953") data = None class Program_weight_tensor_parameter_520: name = "parameter_520" + original_name = "layer_norm_2.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.559082") + max_val = float("0.65332") + mean = float("0.00581997") + std = float("0.0662343") data = None class Program_weight_tensor_parameter_521: name = "parameter_521" + original_name = "layer_norm_2.w_0" shape = [1024] dtype = "float32" - min_val = float("1.0") - max_val = float("1.0") - mean = float("1.0") + min_val = float("0.907715") + max_val = float("1.00781") + mean = float("0.958642") + std = float("0.0052672") data = None class Program_weight_tensor_parameter_522: name = "parameter_522" + original_name = "linear_5.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.188232") + max_val = float("0.235352") + mean = float("-0.00013931") + std = float("0.0389407") data = None class Program_weight_tensor_parameter_523: name = "parameter_523" + original_name = "linear_5.w_0" shape = [4096, 1024] dtype = "float32" - min_val = float("-0.100341") - max_val = float("0.10552") - mean = float("-1.13665e-05") - std = float("0.0200074") + min_val = float("-0.284424") + max_val = float("0.352539") + mean = float("-1.76049e-05") + std = float("0.0289235") data = None class Program_weight_tensor_parameter_524: name = "parameter_524" + original_name = "linear_4.b_0" shape = [4096] dtype = "float32" + min_val = float("-0.172363") + max_val = float("0.0493469") + mean = float("-0.0433741") + std = float("0.0164521") data = None class Program_weight_tensor_parameter_525: name = "parameter_525" + original_name = "linear_4.w_0" shape = [1024, 4096] dtype = "float32" - min_val = float("-0.100763") - max_val = float("0.100267") - mean = float("-1.73864e-06") - std = float("0.0200133") + min_val = float("-0.150757") + max_val = float("0.187256") + mean = float("7.21155e-07") + std = float("0.0305391") data = None class Program_weight_tensor_parameter_526: name = "parameter_526" + original_name = "layer_norm_1.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.561035") + max_val = float("0.585938") + mean = float("-0.00012587") + std = float("0.126379") data = None class Program_weight_tensor_parameter_527: name = "parameter_527" + original_name = "layer_norm_1.w_0" shape = [1024] dtype = "float32" - min_val = float("1.0") - max_val = float("1.0") - mean = float("1.0") + min_val = float("0.915527") + max_val = float("1.15039") + mean = float("0.961136") + std = float("0.00945497") data = None class Program_weight_tensor_parameter_528: name = "parameter_528" + original_name = "linear_3.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.234009") + max_val = float("0.248413") + mean = float("0.000159109") + std = float("0.0444183") data = None class Program_weight_tensor_parameter_529: name = "parameter_529" + original_name = "linear_3.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.101869") - max_val = float("0.0909238") - mean = float("2.55491e-05") - std = float("0.0200053") + min_val = float("-0.23999") + max_val = float("0.247803") + mean = float("1.75202e-06") + std = float("0.0280071") data = None class Program_weight_tensor_parameter_530: name = "parameter_530" + original_name = "linear_2.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.336914") + max_val = float("0.33667") + mean = float("0.00675366") + std = float("0.134018") data = None class Program_weight_tensor_parameter_531: name = "parameter_531" + original_name = "linear_2.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.0949849") - max_val = float("0.0928406") - mean = float("-1.91436e-05") - std = float("0.0199963") + min_val = float("-0.137329") + max_val = float("0.138672") + mean = float("3.63768e-06") + std = float("0.0295815") data = None class Program_weight_tensor_parameter_532: name = "parameter_532" + original_name = "linear_1.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.0773926") + max_val = float("0.0787964") + mean = float("0.000254078") + std = float("0.0165606") data = None class Program_weight_tensor_parameter_533: name = "parameter_533" + original_name = "linear_1.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.096276") - max_val = float("0.10672") - mean = float("5.91881e-06") - std = float("0.0199956") + min_val = float("-0.12793") + max_val = float("0.12561") + mean = float("6.87668e-06") + std = float("0.0290476") data = None class Program_weight_tensor_parameter_534: name = "parameter_534" + original_name = "linear_0.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.0793457") + max_val = float("0.0842285") + mean = float("-0.00033909") + std = float("0.0257897") data = None class Program_weight_tensor_parameter_535: name = "parameter_535" + original_name = "linear_0.w_0" shape = [1024, 1024] dtype = "float32" - min_val = float("-0.106394") - max_val = float("0.0981533") - mean = float("1.25676e-05") - std = float("0.0200069") + min_val = float("-0.139526") + max_val = float("0.134888") + mean = float("8.30226e-06") + std = float("0.0285206") data = None class Program_weight_tensor_parameter_536: name = "parameter_536" + original_name = "layer_norm_0.b_0" shape = [1024] dtype = "float32" + min_val = float("-0.338379") + max_val = float("0.212036") + mean = float("0.000953586") + std = float("0.105768") data = None class Program_weight_tensor_parameter_537: name = "parameter_537" + original_name = "layer_norm_0.w_0" shape = [1024] dtype = "float32" - min_val = float("1.0") - max_val = float("1.0") - mean = float("1.0") + min_val = float("0.831543") + max_val = float("0.986328") + mean = float("0.963245") + std = float("0.0099629") data = None class Program_weight_tensor_parameter_538: name = "parameter_538" + original_name = "prophet_net_positional_embeddings_0.w_0" shape = [512, 1024] dtype = "float32" - min_val = float("-0.0624994") - max_val = float("0.0624999") - mean = float("-8.09357e-05") - std = float("0.0360525") + min_val = float("-0.138184") + max_val = float("0.143188") + mean = float("-3.20669e-05") + std = float("0.0229547") data = None class Program_weight_tensor_parameter_539: name = "parameter_539" + original_name = "embedding_0.w_0" shape = [30522, 1024] dtype = "float32" - min_val = float("-0.0137912") - max_val = float("0.0137912") - mean = float("7.76548e-07") - std = float("0.00796211") + min_val = float("-0.380615") + max_val = float("0.479492") + mean = float("-0.000843613") + std = float("0.0634129") data = None From ec2f60d3e87166ae5aa63b1a1ae65b37634f0c29 Mon Sep 17 00:00:00 2001 From: Liu Yiqun Date: Tue, 20 Jan 2026 16:08:21 +0800 Subject: [PATCH 04/10] Add original_name for serveral PaddleNLP samples. --- .../PaddleNLP/bert-base-cased/input_meta.py | 2 + .../PaddleNLP/bert-base-cased/model.py | 1056 +++------ .../PaddleNLP/bert-base-cased/weight_meta.py | 199 ++ .../PaddleNLP/bert-large-cased/input_meta.py | 2 + .../PaddleNLP/bert-large-cased/model.py | 2076 +++++------------ .../PaddleNLP/bert-large-cased/weight_meta.py | 391 ++++ .../PaddleNLP/ernie-1.0/input_meta.py | 2 + paddle_samples/PaddleNLP/ernie-1.0/model.py | 1032 +++----- .../PaddleNLP/ernie-1.0/weight_meta.py | 199 ++ .../PaddleNLP/ernie-2.0-base-zh/input_meta.py | 2 + .../PaddleNLP/ernie-2.0-base-zh/model.py | 1032 +++----- .../ernie-2.0-base-zh/weight_meta.py | 199 ++ .../ernie-2.0-large-zh/input_meta.py | 2 + .../PaddleNLP/ernie-2.0-large-zh/model.py | 2028 +++++----------- .../ernie-2.0-large-zh/weight_meta.py | 391 ++++ .../PaddleNLP/ernie-3.0-base-zh/input_meta.py | 2 + .../PaddleNLP/ernie-3.0-base-zh/model.py | 1062 +++------ .../ernie-3.0-base-zh/weight_meta.py | 200 ++ .../ernie-3.0-medium-zh/input_meta.py | 2 + .../PaddleNLP/ernie-3.0-medium-zh/model.py | 552 ++--- .../ernie-3.0-medium-zh/weight_meta.py | 104 + .../ernie-3.0-micro-zh/input_meta.py | 2 + .../PaddleNLP/ernie-3.0-micro-zh/model.py | 382 +-- .../ernie-3.0-micro-zh/weight_meta.py | 72 + .../ernie-3.0-xbase-zh/input_meta.py | 2 + .../PaddleNLP/ernie-3.0-xbase-zh/model.py | 1742 ++++---------- .../ernie-3.0-xbase-zh/weight_meta.py | 328 +++ .../PaddleNLP/ernie-m-large/input_meta.py | 1 + .../PaddleNLP/ernie-m-large/model.py | 2073 +++++----------- .../PaddleNLP/ernie-m-large/weight_meta.py | 390 ++++ .../input_meta.py | 2 + .../model.py | 2076 +++++------------ .../weight_meta.py | 391 ++++ .../PaddleNLP/ernie-tiny/input_meta.py | 2 + paddle_samples/PaddleNLP/ernie-tiny/model.py | 285 +-- .../PaddleNLP/ernie-tiny/weight_meta.py | 55 + .../PaddleNLP/ppminilm-6l-768h/input_meta.py | 2 + .../PaddleNLP/ppminilm-6l-768h/model.py | 534 ++--- .../PaddleNLP/ppminilm-6l-768h/weight_meta.py | 103 + .../rocketqa-base-cross-encoder/input_meta.py | 2 + .../rocketqa-base-cross-encoder/model.py | 1062 +++------ .../weight_meta.py | 200 ++ .../input_meta.py | 2 + .../rocketqa-medium-cross-encoder/model.py | 552 ++--- .../weight_meta.py | 104 + .../rocketqa-mini-cross-encoder/input_meta.py | 2 + .../rocketqa-mini-cross-encoder/model.py | 552 ++--- .../weight_meta.py | 106 +- .../skep_ernie_1.0_large_ch/input_meta.py | 2 + .../skep_ernie_1.0_large_ch/model.py | 2028 +++++----------- .../skep_ernie_1.0_large_ch/weight_meta.py | 391 ++++ .../skep_ernie_2.0_large_en/input_meta.py | 2 + .../skep_ernie_2.0_large_en/model.py | 2076 +++++------------ .../skep_ernie_2.0_large_en/weight_meta.py | 391 ++++ .../PaddleNLP/utc-large/input_meta.py | 2 + paddle_samples/PaddleNLP/utc-large/model.py | 2034 +++++----------- .../PaddleNLP/utc-large/weight_meta.py | 392 ++++ .../PaddleNLP/utc-xbase/input_meta.py | 2 + paddle_samples/PaddleNLP/utc-xbase/model.py | 1742 ++++---------- .../PaddleNLP/utc-xbase/weight_meta.py | 328 +++ .../subgraph_1/input_meta.py | 1 + .../subgraph_1/weight_meta.py | 641 +++++ .../subgraph_1/input_meta.py | 1 + .../subgraph_1/weight_meta.py | 384 +++ 64 files changed, 13307 insertions(+), 18669 deletions(-) diff --git a/paddle_samples/PaddleNLP/bert-base-cased/input_meta.py b/paddle_samples/PaddleNLP/bert-base-cased/input_meta.py index 46bea809b..09fce9f05 100644 --- a/paddle_samples/PaddleNLP/bert-base-cased/input_meta.py +++ b/paddle_samples/PaddleNLP/bert-base-cased/input_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_data_0: name = "data_0" + original_name = "input_ids" shape = [1, 21] dtype = "int64" data = [ @@ -29,6 +30,7 @@ class Program_weight_tensor_data_0: class Program_weight_tensor_data_1: name = "data_1" + original_name = "token_type_ids" shape = [1, 21] dtype = "int64" data = [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0] diff --git a/paddle_samples/PaddleNLP/bert-base-cased/model.py b/paddle_samples/PaddleNLP/bert-base-cased/model.py index 2c1752a9d..abd9eac8b 100644 --- a/paddle_samples/PaddleNLP/bert-base-cased/model.py +++ b/paddle_samples/PaddleNLP/bert-base-cased/model.py @@ -268,7 +268,7 @@ def forward( # pd_op.embedding: (1x21x768xf32) <- (1x21xi64, 512x768xf32) embedding_1 = paddle._C_ops.embedding(subtract_0, parameter_197, -1, False) - del parameter_197 + del parameter_197, subtract_0 # pd_op.embedding: (1x21x768xf32) <- (1x21xi64, 2x768xf32) embedding_2 = paddle._C_ops.embedding(data_1, parameter_196, -1, False) @@ -276,9 +276,11 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 1x21x768xf32) add_0 = paddle._C_ops.add(embedding_0, embedding_1) + del embedding_0, embedding_1 # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 1x21x768xf32) add_1 = paddle._C_ops.add(add_0, embedding_2) + del add_0, embedding_2 # pd_op.layer_norm: (1x21x768xf32, 1x21xf32, 1x21xf32) <- (1x21x768xf32, 768xf32, 768xf32) layer_norm_0, layer_norm_1, layer_norm_2 = (lambda x, f: f(x))( @@ -287,125 +289,17 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_194, parameter_195 + del add_1, parameter_194, parameter_195 # pd_op.full: (1xf32) <- () full_4 = paddle._C_ops.full( [1], float("0.1"), paddle.float32, paddle.core.CPUPlace() ) - # pd_op.assign: (1xf32) <- (1xf32) - assign_0 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_1 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_2 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_3 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_4 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_5 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_6 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_7 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_8 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_9 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_10 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_11 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_12 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_13 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_14 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_15 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_16 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_17 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_18 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_19 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_20 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_21 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_22 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_23 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_24 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_25 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_26 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_27 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_28 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_29 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_30 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_31 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_32 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_33 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_34 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_35 = full_4 - # pd_op.dropout: (1x21x768xf32, 1x21x768xui8) <- (1x21x768xf32, None, 1xf32) dropout_0, dropout_1 = (lambda x, f: f(x))( paddle._C_ops.dropout( - layer_norm_0, None, full_4, False, "upscale_in_train", 0, False + layer_norm_0, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -417,13 +311,14 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_2 = paddle._C_ops.add(matmul_0, parameter_192) - del parameter_192 + del matmul_0, parameter_192 # pd_op.full_int_array: (4xi64) <- () full_int_array_1 = [0, 0, 12, 64] # pd_op.reshape: (1x21x12x64xf32) <- (1x21x768xf32, 4xi64) reshape_0 = paddle._C_ops.reshape(add_2, full_int_array_1) + del add_2 # pd_op.transpose: (1x12x21x64xf32) <- (1x21x12x64xf32) transpose_0 = paddle._C_ops.transpose(reshape_0, [0, 2, 1, 3]) @@ -435,7 +330,7 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_3 = paddle._C_ops.add(matmul_1, parameter_190) - del parameter_190 + del matmul_1, parameter_190 # pd_op.matmul: (1x21x768xf32) <- (1x21x768xf32, 768x768xf32) matmul_2 = paddle._C_ops.matmul(dropout_0, parameter_189, False, False) @@ -443,10 +338,11 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_4 = paddle._C_ops.add(matmul_2, parameter_188) - del parameter_188 + del matmul_2, parameter_188 # pd_op.reshape: (1x21x12x64xf32) <- (1x21x768xf32, 4xi64) reshape_1 = paddle._C_ops.reshape(add_3, full_int_array_1) + del add_3 # pd_op.transpose: (1x12x21x64xf32) <- (1x21x12x64xf32) transpose_1 = paddle._C_ops.transpose(reshape_1, [0, 2, 1, 3]) @@ -454,6 +350,7 @@ def forward( # pd_op.reshape: (1x21x12x64xf32) <- (1x21x768xf32, 4xi64) reshape_2 = paddle._C_ops.reshape(add_4, full_int_array_1) + del add_4 # pd_op.transpose: (1x12x21x64xf32) <- (1x21x12x64xf32) transpose_2 = paddle._C_ops.transpose(reshape_2, [0, 2, 1, 3]) @@ -464,48 +361,17 @@ def forward( [1], float("0.125"), paddle.float32, paddle.core.CPUPlace() ) - # pd_op.assign: (1xf32) <- (1xf32) - assign_36 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_37 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_38 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_39 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_40 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_41 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_42 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_43 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_44 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_45 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_46 = full_5 - # pd_op.scale: (1x12x21x64xf32) <- (1x12x21x64xf32, 1xf32) scale_1 = paddle._C_ops.scale(transpose_0, full_5, float("0"), True) del transpose_0 # pd_op.matmul: (1x12x21x21xf32) <- (1x12x21x64xf32, 1x12x21x64xf32) matmul_3 = paddle._C_ops.matmul(scale_1, transpose_1, False, True) + del scale_1, transpose_1 # pd_op.add: (1x12x21x21xf32) <- (1x12x21x21xf32, 1x1x1x21xf32) add_5 = paddle._C_ops.add(matmul_3, unsqueeze_0) + del matmul_3 # pd_op.softmax: (1x12x21x21xf32) <- (1x12x21x21xf32) softmax_0 = paddle._C_ops.softmax(add_5, -1) @@ -514,13 +380,15 @@ def forward( # pd_op.dropout: (1x12x21x21xf32, 1x12x21x21xui8) <- (1x12x21x21xf32, None, 1xf32) dropout_2, dropout_3 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_0, None, full_4, False, "upscale_in_train", 0, False + softmax_0, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_0 # pd_op.matmul: (1x12x21x64xf32) <- (1x12x21x21xf32, 1x12x21x64xf32) matmul_4 = paddle._C_ops.matmul(dropout_2, transpose_2, False, False) + del dropout_2, transpose_2 # pd_op.transpose: (1x21x12x64xf32) <- (1x12x21x64xf32) transpose_3 = paddle._C_ops.transpose(matmul_4, [0, 2, 1, 3]) @@ -531,19 +399,20 @@ def forward( # pd_op.reshape: (1x21x768xf32) <- (1x21x12x64xf32, 3xi64) reshape_3 = paddle._C_ops.reshape(transpose_3, full_int_array_2) + del transpose_3 # pd_op.matmul: (1x21x768xf32) <- (1x21x768xf32, 768x768xf32) matmul_5 = paddle._C_ops.matmul(reshape_3, parameter_187, False, False) - del parameter_187 + del parameter_187, reshape_3 # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_6 = paddle._C_ops.add(matmul_5, parameter_186) - del parameter_186 + del matmul_5, parameter_186 # pd_op.dropout: (1x21x768xf32, 1x21x768xui8) <- (1x21x768xf32, None, 1xf32) dropout_4, dropout_5 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_6, None, full_4, False, "upscale_in_train", 0, False + add_6, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -551,6 +420,7 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 1x21x768xf32) add_7 = paddle._C_ops.add(dropout_0, dropout_4) + del dropout_0, dropout_4 # pd_op.layer_norm: (1x21x768xf32, 1x21xf32, 1x21xf32) <- (1x21x768xf32, 768xf32, 768xf32) layer_norm_3, layer_norm_4, layer_norm_5 = (lambda x, f: f(x))( @@ -559,7 +429,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_180, parameter_181 + del add_7, parameter_180, parameter_181 # pd_op.matmul: (1x21x3072xf32) <- (1x21x768xf32, 768x3072xf32) matmul_6 = paddle._C_ops.matmul(layer_norm_3, parameter_185, False, False) @@ -567,23 +437,24 @@ def forward( # pd_op.add: (1x21x3072xf32) <- (1x21x3072xf32, 3072xf32) add_8 = paddle._C_ops.add(matmul_6, parameter_184) - del parameter_184 + del matmul_6, parameter_184 # pd_op.gelu: (1x21x3072xf32) <- (1x21x3072xf32) gelu_0 = paddle._C_ops.gelu(add_8, False) + del add_8 # pd_op.matmul: (1x21x768xf32) <- (1x21x3072xf32, 3072x768xf32) matmul_7 = paddle._C_ops.matmul(gelu_0, parameter_183, False, False) - del parameter_183 + del gelu_0, parameter_183 # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_9 = paddle._C_ops.add(matmul_7, parameter_182) - del parameter_182 + del matmul_7, parameter_182 # pd_op.dropout: (1x21x768xf32, 1x21x768xui8) <- (1x21x768xf32, None, 1xf32) dropout_6, dropout_7 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_9, None, full_4, False, "upscale_in_train", 0, False + add_9, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -591,6 +462,7 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 1x21x768xf32) add_10 = paddle._C_ops.add(layer_norm_3, dropout_6) + del dropout_6, layer_norm_3 # pd_op.layer_norm: (1x21x768xf32, 1x21xf32, 1x21xf32) <- (1x21x768xf32, 768xf32, 768xf32) layer_norm_6, layer_norm_7, layer_norm_8 = (lambda x, f: f(x))( @@ -599,7 +471,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_178, parameter_179 + del add_10, parameter_178, parameter_179 # pd_op.matmul: (1x21x768xf32) <- (1x21x768xf32, 768x768xf32) matmul_8 = paddle._C_ops.matmul(layer_norm_6, parameter_177, False, False) @@ -607,10 +479,11 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_11 = paddle._C_ops.add(matmul_8, parameter_176) - del parameter_176 + del matmul_8, parameter_176 # pd_op.reshape: (1x21x12x64xf32) <- (1x21x768xf32, 4xi64) reshape_4 = paddle._C_ops.reshape(add_11, full_int_array_1) + del add_11 # pd_op.transpose: (1x12x21x64xf32) <- (1x21x12x64xf32) transpose_4 = paddle._C_ops.transpose(reshape_4, [0, 2, 1, 3]) @@ -622,7 +495,7 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_12 = paddle._C_ops.add(matmul_9, parameter_174) - del parameter_174 + del matmul_9, parameter_174 # pd_op.matmul: (1x21x768xf32) <- (1x21x768xf32, 768x768xf32) matmul_10 = paddle._C_ops.matmul(layer_norm_6, parameter_173, False, False) @@ -630,10 +503,11 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_13 = paddle._C_ops.add(matmul_10, parameter_172) - del parameter_172 + del matmul_10, parameter_172 # pd_op.reshape: (1x21x12x64xf32) <- (1x21x768xf32, 4xi64) reshape_5 = paddle._C_ops.reshape(add_12, full_int_array_1) + del add_12 # pd_op.transpose: (1x12x21x64xf32) <- (1x21x12x64xf32) transpose_5 = paddle._C_ops.transpose(reshape_5, [0, 2, 1, 3]) @@ -641,6 +515,7 @@ def forward( # pd_op.reshape: (1x21x12x64xf32) <- (1x21x768xf32, 4xi64) reshape_6 = paddle._C_ops.reshape(add_13, full_int_array_1) + del add_13 # pd_op.transpose: (1x12x21x64xf32) <- (1x21x12x64xf32) transpose_6 = paddle._C_ops.transpose(reshape_6, [0, 2, 1, 3]) @@ -652,9 +527,11 @@ def forward( # pd_op.matmul: (1x12x21x21xf32) <- (1x12x21x64xf32, 1x12x21x64xf32) matmul_11 = paddle._C_ops.matmul(scale_2, transpose_5, False, True) + del scale_2, transpose_5 # pd_op.add: (1x12x21x21xf32) <- (1x12x21x21xf32, 1x1x1x21xf32) add_14 = paddle._C_ops.add(matmul_11, unsqueeze_0) + del matmul_11 # pd_op.softmax: (1x12x21x21xf32) <- (1x12x21x21xf32) softmax_1 = paddle._C_ops.softmax(add_14, -1) @@ -663,13 +540,15 @@ def forward( # pd_op.dropout: (1x12x21x21xf32, 1x12x21x21xui8) <- (1x12x21x21xf32, None, 1xf32) dropout_8, dropout_9 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_1, None, full_4, False, "upscale_in_train", 0, False + softmax_1, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_1 # pd_op.matmul: (1x12x21x64xf32) <- (1x12x21x21xf32, 1x12x21x64xf32) matmul_12 = paddle._C_ops.matmul(dropout_8, transpose_6, False, False) + del dropout_8, transpose_6 # pd_op.transpose: (1x21x12x64xf32) <- (1x12x21x64xf32) transpose_7 = paddle._C_ops.transpose(matmul_12, [0, 2, 1, 3]) @@ -677,19 +556,20 @@ def forward( # pd_op.reshape: (1x21x768xf32) <- (1x21x12x64xf32, 3xi64) reshape_7 = paddle._C_ops.reshape(transpose_7, full_int_array_2) + del transpose_7 # pd_op.matmul: (1x21x768xf32) <- (1x21x768xf32, 768x768xf32) matmul_13 = paddle._C_ops.matmul(reshape_7, parameter_171, False, False) - del parameter_171 + del parameter_171, reshape_7 # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_15 = paddle._C_ops.add(matmul_13, parameter_170) - del parameter_170 + del matmul_13, parameter_170 # pd_op.dropout: (1x21x768xf32, 1x21x768xui8) <- (1x21x768xf32, None, 1xf32) dropout_10, dropout_11 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_15, None, full_4, False, "upscale_in_train", 0, False + add_15, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -697,6 +577,7 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 1x21x768xf32) add_16 = paddle._C_ops.add(layer_norm_6, dropout_10) + del dropout_10, layer_norm_6 # pd_op.layer_norm: (1x21x768xf32, 1x21xf32, 1x21xf32) <- (1x21x768xf32, 768xf32, 768xf32) layer_norm_9, layer_norm_10, layer_norm_11 = (lambda x, f: f(x))( @@ -705,7 +586,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_164, parameter_165 + del add_16, parameter_164, parameter_165 # pd_op.matmul: (1x21x3072xf32) <- (1x21x768xf32, 768x3072xf32) matmul_14 = paddle._C_ops.matmul(layer_norm_9, parameter_169, False, False) @@ -713,23 +594,24 @@ def forward( # pd_op.add: (1x21x3072xf32) <- (1x21x3072xf32, 3072xf32) add_17 = paddle._C_ops.add(matmul_14, parameter_168) - del parameter_168 + del matmul_14, parameter_168 # pd_op.gelu: (1x21x3072xf32) <- (1x21x3072xf32) gelu_1 = paddle._C_ops.gelu(add_17, False) + del add_17 # pd_op.matmul: (1x21x768xf32) <- (1x21x3072xf32, 3072x768xf32) matmul_15 = paddle._C_ops.matmul(gelu_1, parameter_167, False, False) - del parameter_167 + del gelu_1, parameter_167 # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_18 = paddle._C_ops.add(matmul_15, parameter_166) - del parameter_166 + del matmul_15, parameter_166 # pd_op.dropout: (1x21x768xf32, 1x21x768xui8) <- (1x21x768xf32, None, 1xf32) dropout_12, dropout_13 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_18, None, full_4, False, "upscale_in_train", 0, False + add_18, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -737,6 +619,7 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 1x21x768xf32) add_19 = paddle._C_ops.add(layer_norm_9, dropout_12) + del dropout_12, layer_norm_9 # pd_op.layer_norm: (1x21x768xf32, 1x21xf32, 1x21xf32) <- (1x21x768xf32, 768xf32, 768xf32) layer_norm_12, layer_norm_13, layer_norm_14 = (lambda x, f: f(x))( @@ -745,7 +628,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_162, parameter_163 + del add_19, parameter_162, parameter_163 # pd_op.matmul: (1x21x768xf32) <- (1x21x768xf32, 768x768xf32) matmul_16 = paddle._C_ops.matmul(layer_norm_12, parameter_161, False, False) @@ -753,10 +636,11 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_20 = paddle._C_ops.add(matmul_16, parameter_160) - del parameter_160 + del matmul_16, parameter_160 # pd_op.reshape: (1x21x12x64xf32) <- (1x21x768xf32, 4xi64) reshape_8 = paddle._C_ops.reshape(add_20, full_int_array_1) + del add_20 # pd_op.transpose: (1x12x21x64xf32) <- (1x21x12x64xf32) transpose_8 = paddle._C_ops.transpose(reshape_8, [0, 2, 1, 3]) @@ -768,7 +652,7 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_21 = paddle._C_ops.add(matmul_17, parameter_158) - del parameter_158 + del matmul_17, parameter_158 # pd_op.matmul: (1x21x768xf32) <- (1x21x768xf32, 768x768xf32) matmul_18 = paddle._C_ops.matmul(layer_norm_12, parameter_157, False, False) @@ -776,10 +660,11 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_22 = paddle._C_ops.add(matmul_18, parameter_156) - del parameter_156 + del matmul_18, parameter_156 # pd_op.reshape: (1x21x12x64xf32) <- (1x21x768xf32, 4xi64) reshape_9 = paddle._C_ops.reshape(add_21, full_int_array_1) + del add_21 # pd_op.transpose: (1x12x21x64xf32) <- (1x21x12x64xf32) transpose_9 = paddle._C_ops.transpose(reshape_9, [0, 2, 1, 3]) @@ -787,6 +672,7 @@ def forward( # pd_op.reshape: (1x21x12x64xf32) <- (1x21x768xf32, 4xi64) reshape_10 = paddle._C_ops.reshape(add_22, full_int_array_1) + del add_22 # pd_op.transpose: (1x12x21x64xf32) <- (1x21x12x64xf32) transpose_10 = paddle._C_ops.transpose(reshape_10, [0, 2, 1, 3]) @@ -798,9 +684,11 @@ def forward( # pd_op.matmul: (1x12x21x21xf32) <- (1x12x21x64xf32, 1x12x21x64xf32) matmul_19 = paddle._C_ops.matmul(scale_3, transpose_9, False, True) + del scale_3, transpose_9 # pd_op.add: (1x12x21x21xf32) <- (1x12x21x21xf32, 1x1x1x21xf32) add_23 = paddle._C_ops.add(matmul_19, unsqueeze_0) + del matmul_19 # pd_op.softmax: (1x12x21x21xf32) <- (1x12x21x21xf32) softmax_2 = paddle._C_ops.softmax(add_23, -1) @@ -809,13 +697,15 @@ def forward( # pd_op.dropout: (1x12x21x21xf32, 1x12x21x21xui8) <- (1x12x21x21xf32, None, 1xf32) dropout_14, dropout_15 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_2, None, full_4, False, "upscale_in_train", 0, False + softmax_2, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_2 # pd_op.matmul: (1x12x21x64xf32) <- (1x12x21x21xf32, 1x12x21x64xf32) matmul_20 = paddle._C_ops.matmul(dropout_14, transpose_10, False, False) + del dropout_14, transpose_10 # pd_op.transpose: (1x21x12x64xf32) <- (1x12x21x64xf32) transpose_11 = paddle._C_ops.transpose(matmul_20, [0, 2, 1, 3]) @@ -823,19 +713,20 @@ def forward( # pd_op.reshape: (1x21x768xf32) <- (1x21x12x64xf32, 3xi64) reshape_11 = paddle._C_ops.reshape(transpose_11, full_int_array_2) + del transpose_11 # pd_op.matmul: (1x21x768xf32) <- (1x21x768xf32, 768x768xf32) matmul_21 = paddle._C_ops.matmul(reshape_11, parameter_155, False, False) - del parameter_155 + del parameter_155, reshape_11 # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_24 = paddle._C_ops.add(matmul_21, parameter_154) - del parameter_154 + del matmul_21, parameter_154 # pd_op.dropout: (1x21x768xf32, 1x21x768xui8) <- (1x21x768xf32, None, 1xf32) dropout_16, dropout_17 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_24, None, full_4, False, "upscale_in_train", 0, False + add_24, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -843,6 +734,7 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 1x21x768xf32) add_25 = paddle._C_ops.add(layer_norm_12, dropout_16) + del dropout_16, layer_norm_12 # pd_op.layer_norm: (1x21x768xf32, 1x21xf32, 1x21xf32) <- (1x21x768xf32, 768xf32, 768xf32) layer_norm_15, layer_norm_16, layer_norm_17 = (lambda x, f: f(x))( @@ -851,7 +743,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_148, parameter_149 + del add_25, parameter_148, parameter_149 # pd_op.matmul: (1x21x3072xf32) <- (1x21x768xf32, 768x3072xf32) matmul_22 = paddle._C_ops.matmul(layer_norm_15, parameter_153, False, False) @@ -859,23 +751,24 @@ def forward( # pd_op.add: (1x21x3072xf32) <- (1x21x3072xf32, 3072xf32) add_26 = paddle._C_ops.add(matmul_22, parameter_152) - del parameter_152 + del matmul_22, parameter_152 # pd_op.gelu: (1x21x3072xf32) <- (1x21x3072xf32) gelu_2 = paddle._C_ops.gelu(add_26, False) + del add_26 # pd_op.matmul: (1x21x768xf32) <- (1x21x3072xf32, 3072x768xf32) matmul_23 = paddle._C_ops.matmul(gelu_2, parameter_151, False, False) - del parameter_151 + del gelu_2, parameter_151 # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_27 = paddle._C_ops.add(matmul_23, parameter_150) - del parameter_150 + del matmul_23, parameter_150 # pd_op.dropout: (1x21x768xf32, 1x21x768xui8) <- (1x21x768xf32, None, 1xf32) dropout_18, dropout_19 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_27, None, full_4, False, "upscale_in_train", 0, False + add_27, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -883,6 +776,7 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 1x21x768xf32) add_28 = paddle._C_ops.add(layer_norm_15, dropout_18) + del dropout_18, layer_norm_15 # pd_op.layer_norm: (1x21x768xf32, 1x21xf32, 1x21xf32) <- (1x21x768xf32, 768xf32, 768xf32) layer_norm_18, layer_norm_19, layer_norm_20 = (lambda x, f: f(x))( @@ -891,7 +785,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_146, parameter_147 + del add_28, parameter_146, parameter_147 # pd_op.matmul: (1x21x768xf32) <- (1x21x768xf32, 768x768xf32) matmul_24 = paddle._C_ops.matmul(layer_norm_18, parameter_145, False, False) @@ -899,10 +793,11 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_29 = paddle._C_ops.add(matmul_24, parameter_144) - del parameter_144 + del matmul_24, parameter_144 # pd_op.reshape: (1x21x12x64xf32) <- (1x21x768xf32, 4xi64) reshape_12 = paddle._C_ops.reshape(add_29, full_int_array_1) + del add_29 # pd_op.transpose: (1x12x21x64xf32) <- (1x21x12x64xf32) transpose_12 = paddle._C_ops.transpose(reshape_12, [0, 2, 1, 3]) @@ -914,7 +809,7 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_30 = paddle._C_ops.add(matmul_25, parameter_142) - del parameter_142 + del matmul_25, parameter_142 # pd_op.matmul: (1x21x768xf32) <- (1x21x768xf32, 768x768xf32) matmul_26 = paddle._C_ops.matmul(layer_norm_18, parameter_141, False, False) @@ -922,10 +817,11 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_31 = paddle._C_ops.add(matmul_26, parameter_140) - del parameter_140 + del matmul_26, parameter_140 # pd_op.reshape: (1x21x12x64xf32) <- (1x21x768xf32, 4xi64) reshape_13 = paddle._C_ops.reshape(add_30, full_int_array_1) + del add_30 # pd_op.transpose: (1x12x21x64xf32) <- (1x21x12x64xf32) transpose_13 = paddle._C_ops.transpose(reshape_13, [0, 2, 1, 3]) @@ -933,6 +829,7 @@ def forward( # pd_op.reshape: (1x21x12x64xf32) <- (1x21x768xf32, 4xi64) reshape_14 = paddle._C_ops.reshape(add_31, full_int_array_1) + del add_31 # pd_op.transpose: (1x12x21x64xf32) <- (1x21x12x64xf32) transpose_14 = paddle._C_ops.transpose(reshape_14, [0, 2, 1, 3]) @@ -944,9 +841,11 @@ def forward( # pd_op.matmul: (1x12x21x21xf32) <- (1x12x21x64xf32, 1x12x21x64xf32) matmul_27 = paddle._C_ops.matmul(scale_4, transpose_13, False, True) + del scale_4, transpose_13 # pd_op.add: (1x12x21x21xf32) <- (1x12x21x21xf32, 1x1x1x21xf32) add_32 = paddle._C_ops.add(matmul_27, unsqueeze_0) + del matmul_27 # pd_op.softmax: (1x12x21x21xf32) <- (1x12x21x21xf32) softmax_3 = paddle._C_ops.softmax(add_32, -1) @@ -955,13 +854,15 @@ def forward( # pd_op.dropout: (1x12x21x21xf32, 1x12x21x21xui8) <- (1x12x21x21xf32, None, 1xf32) dropout_20, dropout_21 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_3, None, full_4, False, "upscale_in_train", 0, False + softmax_3, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_3 # pd_op.matmul: (1x12x21x64xf32) <- (1x12x21x21xf32, 1x12x21x64xf32) matmul_28 = paddle._C_ops.matmul(dropout_20, transpose_14, False, False) + del dropout_20, transpose_14 # pd_op.transpose: (1x21x12x64xf32) <- (1x12x21x64xf32) transpose_15 = paddle._C_ops.transpose(matmul_28, [0, 2, 1, 3]) @@ -969,19 +870,20 @@ def forward( # pd_op.reshape: (1x21x768xf32) <- (1x21x12x64xf32, 3xi64) reshape_15 = paddle._C_ops.reshape(transpose_15, full_int_array_2) + del transpose_15 # pd_op.matmul: (1x21x768xf32) <- (1x21x768xf32, 768x768xf32) matmul_29 = paddle._C_ops.matmul(reshape_15, parameter_139, False, False) - del parameter_139 + del parameter_139, reshape_15 # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_33 = paddle._C_ops.add(matmul_29, parameter_138) - del parameter_138 + del matmul_29, parameter_138 # pd_op.dropout: (1x21x768xf32, 1x21x768xui8) <- (1x21x768xf32, None, 1xf32) dropout_22, dropout_23 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_33, None, full_4, False, "upscale_in_train", 0, False + add_33, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -989,6 +891,7 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 1x21x768xf32) add_34 = paddle._C_ops.add(layer_norm_18, dropout_22) + del dropout_22, layer_norm_18 # pd_op.layer_norm: (1x21x768xf32, 1x21xf32, 1x21xf32) <- (1x21x768xf32, 768xf32, 768xf32) layer_norm_21, layer_norm_22, layer_norm_23 = (lambda x, f: f(x))( @@ -997,7 +900,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_132, parameter_133 + del add_34, parameter_132, parameter_133 # pd_op.matmul: (1x21x3072xf32) <- (1x21x768xf32, 768x3072xf32) matmul_30 = paddle._C_ops.matmul(layer_norm_21, parameter_137, False, False) @@ -1005,23 +908,24 @@ def forward( # pd_op.add: (1x21x3072xf32) <- (1x21x3072xf32, 3072xf32) add_35 = paddle._C_ops.add(matmul_30, parameter_136) - del parameter_136 + del matmul_30, parameter_136 # pd_op.gelu: (1x21x3072xf32) <- (1x21x3072xf32) gelu_3 = paddle._C_ops.gelu(add_35, False) + del add_35 # pd_op.matmul: (1x21x768xf32) <- (1x21x3072xf32, 3072x768xf32) matmul_31 = paddle._C_ops.matmul(gelu_3, parameter_135, False, False) - del parameter_135 + del gelu_3, parameter_135 # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_36 = paddle._C_ops.add(matmul_31, parameter_134) - del parameter_134 + del matmul_31, parameter_134 # pd_op.dropout: (1x21x768xf32, 1x21x768xui8) <- (1x21x768xf32, None, 1xf32) dropout_24, dropout_25 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_36, None, full_4, False, "upscale_in_train", 0, False + add_36, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1029,6 +933,7 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 1x21x768xf32) add_37 = paddle._C_ops.add(layer_norm_21, dropout_24) + del dropout_24, layer_norm_21 # pd_op.layer_norm: (1x21x768xf32, 1x21xf32, 1x21xf32) <- (1x21x768xf32, 768xf32, 768xf32) layer_norm_24, layer_norm_25, layer_norm_26 = (lambda x, f: f(x))( @@ -1037,7 +942,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_130, parameter_131 + del add_37, parameter_130, parameter_131 # pd_op.matmul: (1x21x768xf32) <- (1x21x768xf32, 768x768xf32) matmul_32 = paddle._C_ops.matmul(layer_norm_24, parameter_129, False, False) @@ -1045,10 +950,11 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_38 = paddle._C_ops.add(matmul_32, parameter_128) - del parameter_128 + del matmul_32, parameter_128 # pd_op.reshape: (1x21x12x64xf32) <- (1x21x768xf32, 4xi64) reshape_16 = paddle._C_ops.reshape(add_38, full_int_array_1) + del add_38 # pd_op.transpose: (1x12x21x64xf32) <- (1x21x12x64xf32) transpose_16 = paddle._C_ops.transpose(reshape_16, [0, 2, 1, 3]) @@ -1060,7 +966,7 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_39 = paddle._C_ops.add(matmul_33, parameter_126) - del parameter_126 + del matmul_33, parameter_126 # pd_op.matmul: (1x21x768xf32) <- (1x21x768xf32, 768x768xf32) matmul_34 = paddle._C_ops.matmul(layer_norm_24, parameter_125, False, False) @@ -1068,10 +974,11 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_40 = paddle._C_ops.add(matmul_34, parameter_124) - del parameter_124 + del matmul_34, parameter_124 # pd_op.reshape: (1x21x12x64xf32) <- (1x21x768xf32, 4xi64) reshape_17 = paddle._C_ops.reshape(add_39, full_int_array_1) + del add_39 # pd_op.transpose: (1x12x21x64xf32) <- (1x21x12x64xf32) transpose_17 = paddle._C_ops.transpose(reshape_17, [0, 2, 1, 3]) @@ -1079,6 +986,7 @@ def forward( # pd_op.reshape: (1x21x12x64xf32) <- (1x21x768xf32, 4xi64) reshape_18 = paddle._C_ops.reshape(add_40, full_int_array_1) + del add_40 # pd_op.transpose: (1x12x21x64xf32) <- (1x21x12x64xf32) transpose_18 = paddle._C_ops.transpose(reshape_18, [0, 2, 1, 3]) @@ -1090,9 +998,11 @@ def forward( # pd_op.matmul: (1x12x21x21xf32) <- (1x12x21x64xf32, 1x12x21x64xf32) matmul_35 = paddle._C_ops.matmul(scale_5, transpose_17, False, True) + del scale_5, transpose_17 # pd_op.add: (1x12x21x21xf32) <- (1x12x21x21xf32, 1x1x1x21xf32) add_41 = paddle._C_ops.add(matmul_35, unsqueeze_0) + del matmul_35 # pd_op.softmax: (1x12x21x21xf32) <- (1x12x21x21xf32) softmax_4 = paddle._C_ops.softmax(add_41, -1) @@ -1101,13 +1011,15 @@ def forward( # pd_op.dropout: (1x12x21x21xf32, 1x12x21x21xui8) <- (1x12x21x21xf32, None, 1xf32) dropout_26, dropout_27 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_4, None, full_4, False, "upscale_in_train", 0, False + softmax_4, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_4 # pd_op.matmul: (1x12x21x64xf32) <- (1x12x21x21xf32, 1x12x21x64xf32) matmul_36 = paddle._C_ops.matmul(dropout_26, transpose_18, False, False) + del dropout_26, transpose_18 # pd_op.transpose: (1x21x12x64xf32) <- (1x12x21x64xf32) transpose_19 = paddle._C_ops.transpose(matmul_36, [0, 2, 1, 3]) @@ -1115,19 +1027,20 @@ def forward( # pd_op.reshape: (1x21x768xf32) <- (1x21x12x64xf32, 3xi64) reshape_19 = paddle._C_ops.reshape(transpose_19, full_int_array_2) + del transpose_19 # pd_op.matmul: (1x21x768xf32) <- (1x21x768xf32, 768x768xf32) matmul_37 = paddle._C_ops.matmul(reshape_19, parameter_123, False, False) - del parameter_123 + del parameter_123, reshape_19 # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_42 = paddle._C_ops.add(matmul_37, parameter_122) - del parameter_122 + del matmul_37, parameter_122 # pd_op.dropout: (1x21x768xf32, 1x21x768xui8) <- (1x21x768xf32, None, 1xf32) dropout_28, dropout_29 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_42, None, full_4, False, "upscale_in_train", 0, False + add_42, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1135,6 +1048,7 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 1x21x768xf32) add_43 = paddle._C_ops.add(layer_norm_24, dropout_28) + del dropout_28, layer_norm_24 # pd_op.layer_norm: (1x21x768xf32, 1x21xf32, 1x21xf32) <- (1x21x768xf32, 768xf32, 768xf32) layer_norm_27, layer_norm_28, layer_norm_29 = (lambda x, f: f(x))( @@ -1143,7 +1057,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_116, parameter_117 + del add_43, parameter_116, parameter_117 # pd_op.matmul: (1x21x3072xf32) <- (1x21x768xf32, 768x3072xf32) matmul_38 = paddle._C_ops.matmul(layer_norm_27, parameter_121, False, False) @@ -1151,23 +1065,24 @@ def forward( # pd_op.add: (1x21x3072xf32) <- (1x21x3072xf32, 3072xf32) add_44 = paddle._C_ops.add(matmul_38, parameter_120) - del parameter_120 + del matmul_38, parameter_120 # pd_op.gelu: (1x21x3072xf32) <- (1x21x3072xf32) gelu_4 = paddle._C_ops.gelu(add_44, False) + del add_44 # pd_op.matmul: (1x21x768xf32) <- (1x21x3072xf32, 3072x768xf32) matmul_39 = paddle._C_ops.matmul(gelu_4, parameter_119, False, False) - del parameter_119 + del gelu_4, parameter_119 # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_45 = paddle._C_ops.add(matmul_39, parameter_118) - del parameter_118 + del matmul_39, parameter_118 # pd_op.dropout: (1x21x768xf32, 1x21x768xui8) <- (1x21x768xf32, None, 1xf32) dropout_30, dropout_31 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_45, None, full_4, False, "upscale_in_train", 0, False + add_45, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1175,6 +1090,7 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 1x21x768xf32) add_46 = paddle._C_ops.add(layer_norm_27, dropout_30) + del dropout_30, layer_norm_27 # pd_op.layer_norm: (1x21x768xf32, 1x21xf32, 1x21xf32) <- (1x21x768xf32, 768xf32, 768xf32) layer_norm_30, layer_norm_31, layer_norm_32 = (lambda x, f: f(x))( @@ -1183,7 +1099,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_114, parameter_115 + del add_46, parameter_114, parameter_115 # pd_op.matmul: (1x21x768xf32) <- (1x21x768xf32, 768x768xf32) matmul_40 = paddle._C_ops.matmul(layer_norm_30, parameter_113, False, False) @@ -1191,10 +1107,11 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_47 = paddle._C_ops.add(matmul_40, parameter_112) - del parameter_112 + del matmul_40, parameter_112 # pd_op.reshape: (1x21x12x64xf32) <- (1x21x768xf32, 4xi64) reshape_20 = paddle._C_ops.reshape(add_47, full_int_array_1) + del add_47 # pd_op.transpose: (1x12x21x64xf32) <- (1x21x12x64xf32) transpose_20 = paddle._C_ops.transpose(reshape_20, [0, 2, 1, 3]) @@ -1206,7 +1123,7 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_48 = paddle._C_ops.add(matmul_41, parameter_110) - del parameter_110 + del matmul_41, parameter_110 # pd_op.matmul: (1x21x768xf32) <- (1x21x768xf32, 768x768xf32) matmul_42 = paddle._C_ops.matmul(layer_norm_30, parameter_109, False, False) @@ -1214,10 +1131,11 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_49 = paddle._C_ops.add(matmul_42, parameter_108) - del parameter_108 + del matmul_42, parameter_108 # pd_op.reshape: (1x21x12x64xf32) <- (1x21x768xf32, 4xi64) reshape_21 = paddle._C_ops.reshape(add_48, full_int_array_1) + del add_48 # pd_op.transpose: (1x12x21x64xf32) <- (1x21x12x64xf32) transpose_21 = paddle._C_ops.transpose(reshape_21, [0, 2, 1, 3]) @@ -1225,6 +1143,7 @@ def forward( # pd_op.reshape: (1x21x12x64xf32) <- (1x21x768xf32, 4xi64) reshape_22 = paddle._C_ops.reshape(add_49, full_int_array_1) + del add_49 # pd_op.transpose: (1x12x21x64xf32) <- (1x21x12x64xf32) transpose_22 = paddle._C_ops.transpose(reshape_22, [0, 2, 1, 3]) @@ -1236,9 +1155,11 @@ def forward( # pd_op.matmul: (1x12x21x21xf32) <- (1x12x21x64xf32, 1x12x21x64xf32) matmul_43 = paddle._C_ops.matmul(scale_6, transpose_21, False, True) + del scale_6, transpose_21 # pd_op.add: (1x12x21x21xf32) <- (1x12x21x21xf32, 1x1x1x21xf32) add_50 = paddle._C_ops.add(matmul_43, unsqueeze_0) + del matmul_43 # pd_op.softmax: (1x12x21x21xf32) <- (1x12x21x21xf32) softmax_5 = paddle._C_ops.softmax(add_50, -1) @@ -1247,13 +1168,15 @@ def forward( # pd_op.dropout: (1x12x21x21xf32, 1x12x21x21xui8) <- (1x12x21x21xf32, None, 1xf32) dropout_32, dropout_33 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_5, None, full_4, False, "upscale_in_train", 0, False + softmax_5, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_5 # pd_op.matmul: (1x12x21x64xf32) <- (1x12x21x21xf32, 1x12x21x64xf32) matmul_44 = paddle._C_ops.matmul(dropout_32, transpose_22, False, False) + del dropout_32, transpose_22 # pd_op.transpose: (1x21x12x64xf32) <- (1x12x21x64xf32) transpose_23 = paddle._C_ops.transpose(matmul_44, [0, 2, 1, 3]) @@ -1261,19 +1184,20 @@ def forward( # pd_op.reshape: (1x21x768xf32) <- (1x21x12x64xf32, 3xi64) reshape_23 = paddle._C_ops.reshape(transpose_23, full_int_array_2) + del transpose_23 # pd_op.matmul: (1x21x768xf32) <- (1x21x768xf32, 768x768xf32) matmul_45 = paddle._C_ops.matmul(reshape_23, parameter_107, False, False) - del parameter_107 + del parameter_107, reshape_23 # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_51 = paddle._C_ops.add(matmul_45, parameter_106) - del parameter_106 + del matmul_45, parameter_106 # pd_op.dropout: (1x21x768xf32, 1x21x768xui8) <- (1x21x768xf32, None, 1xf32) dropout_34, dropout_35 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_51, None, full_4, False, "upscale_in_train", 0, False + add_51, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1281,6 +1205,7 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 1x21x768xf32) add_52 = paddle._C_ops.add(layer_norm_30, dropout_34) + del dropout_34, layer_norm_30 # pd_op.layer_norm: (1x21x768xf32, 1x21xf32, 1x21xf32) <- (1x21x768xf32, 768xf32, 768xf32) layer_norm_33, layer_norm_34, layer_norm_35 = (lambda x, f: f(x))( @@ -1289,7 +1214,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_100, parameter_101 + del add_52, parameter_100, parameter_101 # pd_op.matmul: (1x21x3072xf32) <- (1x21x768xf32, 768x3072xf32) matmul_46 = paddle._C_ops.matmul(layer_norm_33, parameter_105, False, False) @@ -1297,23 +1222,24 @@ def forward( # pd_op.add: (1x21x3072xf32) <- (1x21x3072xf32, 3072xf32) add_53 = paddle._C_ops.add(matmul_46, parameter_104) - del parameter_104 + del matmul_46, parameter_104 # pd_op.gelu: (1x21x3072xf32) <- (1x21x3072xf32) gelu_5 = paddle._C_ops.gelu(add_53, False) + del add_53 # pd_op.matmul: (1x21x768xf32) <- (1x21x3072xf32, 3072x768xf32) matmul_47 = paddle._C_ops.matmul(gelu_5, parameter_103, False, False) - del parameter_103 + del gelu_5, parameter_103 # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_54 = paddle._C_ops.add(matmul_47, parameter_102) - del parameter_102 + del matmul_47, parameter_102 # pd_op.dropout: (1x21x768xf32, 1x21x768xui8) <- (1x21x768xf32, None, 1xf32) dropout_36, dropout_37 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_54, None, full_4, False, "upscale_in_train", 0, False + add_54, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1321,6 +1247,7 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 1x21x768xf32) add_55 = paddle._C_ops.add(layer_norm_33, dropout_36) + del dropout_36, layer_norm_33 # pd_op.layer_norm: (1x21x768xf32, 1x21xf32, 1x21xf32) <- (1x21x768xf32, 768xf32, 768xf32) layer_norm_36, layer_norm_37, layer_norm_38 = (lambda x, f: f(x))( @@ -1329,7 +1256,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_98, parameter_99 + del add_55, parameter_98, parameter_99 # pd_op.matmul: (1x21x768xf32) <- (1x21x768xf32, 768x768xf32) matmul_48 = paddle._C_ops.matmul(layer_norm_36, parameter_97, False, False) @@ -1337,10 +1264,11 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_56 = paddle._C_ops.add(matmul_48, parameter_96) - del parameter_96 + del matmul_48, parameter_96 # pd_op.reshape: (1x21x12x64xf32) <- (1x21x768xf32, 4xi64) reshape_24 = paddle._C_ops.reshape(add_56, full_int_array_1) + del add_56 # pd_op.transpose: (1x12x21x64xf32) <- (1x21x12x64xf32) transpose_24 = paddle._C_ops.transpose(reshape_24, [0, 2, 1, 3]) @@ -1352,7 +1280,7 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_57 = paddle._C_ops.add(matmul_49, parameter_94) - del parameter_94 + del matmul_49, parameter_94 # pd_op.matmul: (1x21x768xf32) <- (1x21x768xf32, 768x768xf32) matmul_50 = paddle._C_ops.matmul(layer_norm_36, parameter_93, False, False) @@ -1360,10 +1288,11 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_58 = paddle._C_ops.add(matmul_50, parameter_92) - del parameter_92 + del matmul_50, parameter_92 # pd_op.reshape: (1x21x12x64xf32) <- (1x21x768xf32, 4xi64) reshape_25 = paddle._C_ops.reshape(add_57, full_int_array_1) + del add_57 # pd_op.transpose: (1x12x21x64xf32) <- (1x21x12x64xf32) transpose_25 = paddle._C_ops.transpose(reshape_25, [0, 2, 1, 3]) @@ -1371,6 +1300,7 @@ def forward( # pd_op.reshape: (1x21x12x64xf32) <- (1x21x768xf32, 4xi64) reshape_26 = paddle._C_ops.reshape(add_58, full_int_array_1) + del add_58 # pd_op.transpose: (1x12x21x64xf32) <- (1x21x12x64xf32) transpose_26 = paddle._C_ops.transpose(reshape_26, [0, 2, 1, 3]) @@ -1382,9 +1312,11 @@ def forward( # pd_op.matmul: (1x12x21x21xf32) <- (1x12x21x64xf32, 1x12x21x64xf32) matmul_51 = paddle._C_ops.matmul(scale_7, transpose_25, False, True) + del scale_7, transpose_25 # pd_op.add: (1x12x21x21xf32) <- (1x12x21x21xf32, 1x1x1x21xf32) add_59 = paddle._C_ops.add(matmul_51, unsqueeze_0) + del matmul_51 # pd_op.softmax: (1x12x21x21xf32) <- (1x12x21x21xf32) softmax_6 = paddle._C_ops.softmax(add_59, -1) @@ -1393,13 +1325,15 @@ def forward( # pd_op.dropout: (1x12x21x21xf32, 1x12x21x21xui8) <- (1x12x21x21xf32, None, 1xf32) dropout_38, dropout_39 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_6, None, full_4, False, "upscale_in_train", 0, False + softmax_6, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_6 # pd_op.matmul: (1x12x21x64xf32) <- (1x12x21x21xf32, 1x12x21x64xf32) matmul_52 = paddle._C_ops.matmul(dropout_38, transpose_26, False, False) + del dropout_38, transpose_26 # pd_op.transpose: (1x21x12x64xf32) <- (1x12x21x64xf32) transpose_27 = paddle._C_ops.transpose(matmul_52, [0, 2, 1, 3]) @@ -1407,19 +1341,20 @@ def forward( # pd_op.reshape: (1x21x768xf32) <- (1x21x12x64xf32, 3xi64) reshape_27 = paddle._C_ops.reshape(transpose_27, full_int_array_2) + del transpose_27 # pd_op.matmul: (1x21x768xf32) <- (1x21x768xf32, 768x768xf32) matmul_53 = paddle._C_ops.matmul(reshape_27, parameter_91, False, False) - del parameter_91 + del parameter_91, reshape_27 # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_60 = paddle._C_ops.add(matmul_53, parameter_90) - del parameter_90 + del matmul_53, parameter_90 # pd_op.dropout: (1x21x768xf32, 1x21x768xui8) <- (1x21x768xf32, None, 1xf32) dropout_40, dropout_41 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_60, None, full_4, False, "upscale_in_train", 0, False + add_60, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1427,6 +1362,7 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 1x21x768xf32) add_61 = paddle._C_ops.add(layer_norm_36, dropout_40) + del dropout_40, layer_norm_36 # pd_op.layer_norm: (1x21x768xf32, 1x21xf32, 1x21xf32) <- (1x21x768xf32, 768xf32, 768xf32) layer_norm_39, layer_norm_40, layer_norm_41 = (lambda x, f: f(x))( @@ -1435,7 +1371,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_84, parameter_85 + del add_61, parameter_84, parameter_85 # pd_op.matmul: (1x21x3072xf32) <- (1x21x768xf32, 768x3072xf32) matmul_54 = paddle._C_ops.matmul(layer_norm_39, parameter_89, False, False) @@ -1443,23 +1379,24 @@ def forward( # pd_op.add: (1x21x3072xf32) <- (1x21x3072xf32, 3072xf32) add_62 = paddle._C_ops.add(matmul_54, parameter_88) - del parameter_88 + del matmul_54, parameter_88 # pd_op.gelu: (1x21x3072xf32) <- (1x21x3072xf32) gelu_6 = paddle._C_ops.gelu(add_62, False) + del add_62 # pd_op.matmul: (1x21x768xf32) <- (1x21x3072xf32, 3072x768xf32) matmul_55 = paddle._C_ops.matmul(gelu_6, parameter_87, False, False) - del parameter_87 + del gelu_6, parameter_87 # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_63 = paddle._C_ops.add(matmul_55, parameter_86) - del parameter_86 + del matmul_55, parameter_86 # pd_op.dropout: (1x21x768xf32, 1x21x768xui8) <- (1x21x768xf32, None, 1xf32) dropout_42, dropout_43 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_63, None, full_4, False, "upscale_in_train", 0, False + add_63, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1467,6 +1404,7 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 1x21x768xf32) add_64 = paddle._C_ops.add(layer_norm_39, dropout_42) + del dropout_42, layer_norm_39 # pd_op.layer_norm: (1x21x768xf32, 1x21xf32, 1x21xf32) <- (1x21x768xf32, 768xf32, 768xf32) layer_norm_42, layer_norm_43, layer_norm_44 = (lambda x, f: f(x))( @@ -1475,7 +1413,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_82, parameter_83 + del add_64, parameter_82, parameter_83 # pd_op.matmul: (1x21x768xf32) <- (1x21x768xf32, 768x768xf32) matmul_56 = paddle._C_ops.matmul(layer_norm_42, parameter_81, False, False) @@ -1483,10 +1421,11 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_65 = paddle._C_ops.add(matmul_56, parameter_80) - del parameter_80 + del matmul_56, parameter_80 # pd_op.reshape: (1x21x12x64xf32) <- (1x21x768xf32, 4xi64) reshape_28 = paddle._C_ops.reshape(add_65, full_int_array_1) + del add_65 # pd_op.transpose: (1x12x21x64xf32) <- (1x21x12x64xf32) transpose_28 = paddle._C_ops.transpose(reshape_28, [0, 2, 1, 3]) @@ -1498,7 +1437,7 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_66 = paddle._C_ops.add(matmul_57, parameter_78) - del parameter_78 + del matmul_57, parameter_78 # pd_op.matmul: (1x21x768xf32) <- (1x21x768xf32, 768x768xf32) matmul_58 = paddle._C_ops.matmul(layer_norm_42, parameter_77, False, False) @@ -1506,10 +1445,11 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_67 = paddle._C_ops.add(matmul_58, parameter_76) - del parameter_76 + del matmul_58, parameter_76 # pd_op.reshape: (1x21x12x64xf32) <- (1x21x768xf32, 4xi64) reshape_29 = paddle._C_ops.reshape(add_66, full_int_array_1) + del add_66 # pd_op.transpose: (1x12x21x64xf32) <- (1x21x12x64xf32) transpose_29 = paddle._C_ops.transpose(reshape_29, [0, 2, 1, 3]) @@ -1517,6 +1457,7 @@ def forward( # pd_op.reshape: (1x21x12x64xf32) <- (1x21x768xf32, 4xi64) reshape_30 = paddle._C_ops.reshape(add_67, full_int_array_1) + del add_67 # pd_op.transpose: (1x12x21x64xf32) <- (1x21x12x64xf32) transpose_30 = paddle._C_ops.transpose(reshape_30, [0, 2, 1, 3]) @@ -1528,9 +1469,11 @@ def forward( # pd_op.matmul: (1x12x21x21xf32) <- (1x12x21x64xf32, 1x12x21x64xf32) matmul_59 = paddle._C_ops.matmul(scale_8, transpose_29, False, True) + del scale_8, transpose_29 # pd_op.add: (1x12x21x21xf32) <- (1x12x21x21xf32, 1x1x1x21xf32) add_68 = paddle._C_ops.add(matmul_59, unsqueeze_0) + del matmul_59 # pd_op.softmax: (1x12x21x21xf32) <- (1x12x21x21xf32) softmax_7 = paddle._C_ops.softmax(add_68, -1) @@ -1539,13 +1482,15 @@ def forward( # pd_op.dropout: (1x12x21x21xf32, 1x12x21x21xui8) <- (1x12x21x21xf32, None, 1xf32) dropout_44, dropout_45 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_7, None, full_4, False, "upscale_in_train", 0, False + softmax_7, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_7 # pd_op.matmul: (1x12x21x64xf32) <- (1x12x21x21xf32, 1x12x21x64xf32) matmul_60 = paddle._C_ops.matmul(dropout_44, transpose_30, False, False) + del dropout_44, transpose_30 # pd_op.transpose: (1x21x12x64xf32) <- (1x12x21x64xf32) transpose_31 = paddle._C_ops.transpose(matmul_60, [0, 2, 1, 3]) @@ -1553,19 +1498,20 @@ def forward( # pd_op.reshape: (1x21x768xf32) <- (1x21x12x64xf32, 3xi64) reshape_31 = paddle._C_ops.reshape(transpose_31, full_int_array_2) + del transpose_31 # pd_op.matmul: (1x21x768xf32) <- (1x21x768xf32, 768x768xf32) matmul_61 = paddle._C_ops.matmul(reshape_31, parameter_75, False, False) - del parameter_75 + del parameter_75, reshape_31 # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_69 = paddle._C_ops.add(matmul_61, parameter_74) - del parameter_74 + del matmul_61, parameter_74 # pd_op.dropout: (1x21x768xf32, 1x21x768xui8) <- (1x21x768xf32, None, 1xf32) dropout_46, dropout_47 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_69, None, full_4, False, "upscale_in_train", 0, False + add_69, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1573,6 +1519,7 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 1x21x768xf32) add_70 = paddle._C_ops.add(layer_norm_42, dropout_46) + del dropout_46, layer_norm_42 # pd_op.layer_norm: (1x21x768xf32, 1x21xf32, 1x21xf32) <- (1x21x768xf32, 768xf32, 768xf32) layer_norm_45, layer_norm_46, layer_norm_47 = (lambda x, f: f(x))( @@ -1581,7 +1528,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_68, parameter_69 + del add_70, parameter_68, parameter_69 # pd_op.matmul: (1x21x3072xf32) <- (1x21x768xf32, 768x3072xf32) matmul_62 = paddle._C_ops.matmul(layer_norm_45, parameter_73, False, False) @@ -1589,23 +1536,24 @@ def forward( # pd_op.add: (1x21x3072xf32) <- (1x21x3072xf32, 3072xf32) add_71 = paddle._C_ops.add(matmul_62, parameter_72) - del parameter_72 + del matmul_62, parameter_72 # pd_op.gelu: (1x21x3072xf32) <- (1x21x3072xf32) gelu_7 = paddle._C_ops.gelu(add_71, False) + del add_71 # pd_op.matmul: (1x21x768xf32) <- (1x21x3072xf32, 3072x768xf32) matmul_63 = paddle._C_ops.matmul(gelu_7, parameter_71, False, False) - del parameter_71 + del gelu_7, parameter_71 # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_72 = paddle._C_ops.add(matmul_63, parameter_70) - del parameter_70 + del matmul_63, parameter_70 # pd_op.dropout: (1x21x768xf32, 1x21x768xui8) <- (1x21x768xf32, None, 1xf32) dropout_48, dropout_49 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_72, None, full_4, False, "upscale_in_train", 0, False + add_72, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1613,6 +1561,7 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 1x21x768xf32) add_73 = paddle._C_ops.add(layer_norm_45, dropout_48) + del dropout_48, layer_norm_45 # pd_op.layer_norm: (1x21x768xf32, 1x21xf32, 1x21xf32) <- (1x21x768xf32, 768xf32, 768xf32) layer_norm_48, layer_norm_49, layer_norm_50 = (lambda x, f: f(x))( @@ -1621,7 +1570,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_66, parameter_67 + del add_73, parameter_66, parameter_67 # pd_op.matmul: (1x21x768xf32) <- (1x21x768xf32, 768x768xf32) matmul_64 = paddle._C_ops.matmul(layer_norm_48, parameter_65, False, False) @@ -1629,10 +1578,11 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_74 = paddle._C_ops.add(matmul_64, parameter_64) - del parameter_64 + del matmul_64, parameter_64 # pd_op.reshape: (1x21x12x64xf32) <- (1x21x768xf32, 4xi64) reshape_32 = paddle._C_ops.reshape(add_74, full_int_array_1) + del add_74 # pd_op.transpose: (1x12x21x64xf32) <- (1x21x12x64xf32) transpose_32 = paddle._C_ops.transpose(reshape_32, [0, 2, 1, 3]) @@ -1644,7 +1594,7 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_75 = paddle._C_ops.add(matmul_65, parameter_62) - del parameter_62 + del matmul_65, parameter_62 # pd_op.matmul: (1x21x768xf32) <- (1x21x768xf32, 768x768xf32) matmul_66 = paddle._C_ops.matmul(layer_norm_48, parameter_61, False, False) @@ -1652,10 +1602,11 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_76 = paddle._C_ops.add(matmul_66, parameter_60) - del parameter_60 + del matmul_66, parameter_60 # pd_op.reshape: (1x21x12x64xf32) <- (1x21x768xf32, 4xi64) reshape_33 = paddle._C_ops.reshape(add_75, full_int_array_1) + del add_75 # pd_op.transpose: (1x12x21x64xf32) <- (1x21x12x64xf32) transpose_33 = paddle._C_ops.transpose(reshape_33, [0, 2, 1, 3]) @@ -1663,6 +1614,7 @@ def forward( # pd_op.reshape: (1x21x12x64xf32) <- (1x21x768xf32, 4xi64) reshape_34 = paddle._C_ops.reshape(add_76, full_int_array_1) + del add_76 # pd_op.transpose: (1x12x21x64xf32) <- (1x21x12x64xf32) transpose_34 = paddle._C_ops.transpose(reshape_34, [0, 2, 1, 3]) @@ -1674,9 +1626,11 @@ def forward( # pd_op.matmul: (1x12x21x21xf32) <- (1x12x21x64xf32, 1x12x21x64xf32) matmul_67 = paddle._C_ops.matmul(scale_9, transpose_33, False, True) + del scale_9, transpose_33 # pd_op.add: (1x12x21x21xf32) <- (1x12x21x21xf32, 1x1x1x21xf32) add_77 = paddle._C_ops.add(matmul_67, unsqueeze_0) + del matmul_67 # pd_op.softmax: (1x12x21x21xf32) <- (1x12x21x21xf32) softmax_8 = paddle._C_ops.softmax(add_77, -1) @@ -1685,13 +1639,15 @@ def forward( # pd_op.dropout: (1x12x21x21xf32, 1x12x21x21xui8) <- (1x12x21x21xf32, None, 1xf32) dropout_50, dropout_51 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_8, None, full_4, False, "upscale_in_train", 0, False + softmax_8, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_8 # pd_op.matmul: (1x12x21x64xf32) <- (1x12x21x21xf32, 1x12x21x64xf32) matmul_68 = paddle._C_ops.matmul(dropout_50, transpose_34, False, False) + del dropout_50, transpose_34 # pd_op.transpose: (1x21x12x64xf32) <- (1x12x21x64xf32) transpose_35 = paddle._C_ops.transpose(matmul_68, [0, 2, 1, 3]) @@ -1699,19 +1655,20 @@ def forward( # pd_op.reshape: (1x21x768xf32) <- (1x21x12x64xf32, 3xi64) reshape_35 = paddle._C_ops.reshape(transpose_35, full_int_array_2) + del transpose_35 # pd_op.matmul: (1x21x768xf32) <- (1x21x768xf32, 768x768xf32) matmul_69 = paddle._C_ops.matmul(reshape_35, parameter_59, False, False) - del parameter_59 + del parameter_59, reshape_35 # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_78 = paddle._C_ops.add(matmul_69, parameter_58) - del parameter_58 + del matmul_69, parameter_58 # pd_op.dropout: (1x21x768xf32, 1x21x768xui8) <- (1x21x768xf32, None, 1xf32) dropout_52, dropout_53 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_78, None, full_4, False, "upscale_in_train", 0, False + add_78, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1719,6 +1676,7 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 1x21x768xf32) add_79 = paddle._C_ops.add(layer_norm_48, dropout_52) + del dropout_52, layer_norm_48 # pd_op.layer_norm: (1x21x768xf32, 1x21xf32, 1x21xf32) <- (1x21x768xf32, 768xf32, 768xf32) layer_norm_51, layer_norm_52, layer_norm_53 = (lambda x, f: f(x))( @@ -1727,7 +1685,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_52, parameter_53 + del add_79, parameter_52, parameter_53 # pd_op.matmul: (1x21x3072xf32) <- (1x21x768xf32, 768x3072xf32) matmul_70 = paddle._C_ops.matmul(layer_norm_51, parameter_57, False, False) @@ -1735,23 +1693,24 @@ def forward( # pd_op.add: (1x21x3072xf32) <- (1x21x3072xf32, 3072xf32) add_80 = paddle._C_ops.add(matmul_70, parameter_56) - del parameter_56 + del matmul_70, parameter_56 # pd_op.gelu: (1x21x3072xf32) <- (1x21x3072xf32) gelu_8 = paddle._C_ops.gelu(add_80, False) + del add_80 # pd_op.matmul: (1x21x768xf32) <- (1x21x3072xf32, 3072x768xf32) matmul_71 = paddle._C_ops.matmul(gelu_8, parameter_55, False, False) - del parameter_55 + del gelu_8, parameter_55 # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_81 = paddle._C_ops.add(matmul_71, parameter_54) - del parameter_54 + del matmul_71, parameter_54 # pd_op.dropout: (1x21x768xf32, 1x21x768xui8) <- (1x21x768xf32, None, 1xf32) dropout_54, dropout_55 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_81, None, full_4, False, "upscale_in_train", 0, False + add_81, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1759,6 +1718,7 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 1x21x768xf32) add_82 = paddle._C_ops.add(layer_norm_51, dropout_54) + del dropout_54, layer_norm_51 # pd_op.layer_norm: (1x21x768xf32, 1x21xf32, 1x21xf32) <- (1x21x768xf32, 768xf32, 768xf32) layer_norm_54, layer_norm_55, layer_norm_56 = (lambda x, f: f(x))( @@ -1767,7 +1727,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_50, parameter_51 + del add_82, parameter_50, parameter_51 # pd_op.matmul: (1x21x768xf32) <- (1x21x768xf32, 768x768xf32) matmul_72 = paddle._C_ops.matmul(layer_norm_54, parameter_49, False, False) @@ -1775,10 +1735,11 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_83 = paddle._C_ops.add(matmul_72, parameter_48) - del parameter_48 + del matmul_72, parameter_48 # pd_op.reshape: (1x21x12x64xf32) <- (1x21x768xf32, 4xi64) reshape_36 = paddle._C_ops.reshape(add_83, full_int_array_1) + del add_83 # pd_op.transpose: (1x12x21x64xf32) <- (1x21x12x64xf32) transpose_36 = paddle._C_ops.transpose(reshape_36, [0, 2, 1, 3]) @@ -1790,7 +1751,7 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_84 = paddle._C_ops.add(matmul_73, parameter_46) - del parameter_46 + del matmul_73, parameter_46 # pd_op.matmul: (1x21x768xf32) <- (1x21x768xf32, 768x768xf32) matmul_74 = paddle._C_ops.matmul(layer_norm_54, parameter_45, False, False) @@ -1798,10 +1759,11 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_85 = paddle._C_ops.add(matmul_74, parameter_44) - del parameter_44 + del matmul_74, parameter_44 # pd_op.reshape: (1x21x12x64xf32) <- (1x21x768xf32, 4xi64) reshape_37 = paddle._C_ops.reshape(add_84, full_int_array_1) + del add_84 # pd_op.transpose: (1x12x21x64xf32) <- (1x21x12x64xf32) transpose_37 = paddle._C_ops.transpose(reshape_37, [0, 2, 1, 3]) @@ -1809,6 +1771,7 @@ def forward( # pd_op.reshape: (1x21x12x64xf32) <- (1x21x768xf32, 4xi64) reshape_38 = paddle._C_ops.reshape(add_85, full_int_array_1) + del add_85 # pd_op.transpose: (1x12x21x64xf32) <- (1x21x12x64xf32) transpose_38 = paddle._C_ops.transpose(reshape_38, [0, 2, 1, 3]) @@ -1820,9 +1783,11 @@ def forward( # pd_op.matmul: (1x12x21x21xf32) <- (1x12x21x64xf32, 1x12x21x64xf32) matmul_75 = paddle._C_ops.matmul(scale_10, transpose_37, False, True) + del scale_10, transpose_37 # pd_op.add: (1x12x21x21xf32) <- (1x12x21x21xf32, 1x1x1x21xf32) add_86 = paddle._C_ops.add(matmul_75, unsqueeze_0) + del matmul_75 # pd_op.softmax: (1x12x21x21xf32) <- (1x12x21x21xf32) softmax_9 = paddle._C_ops.softmax(add_86, -1) @@ -1831,13 +1796,15 @@ def forward( # pd_op.dropout: (1x12x21x21xf32, 1x12x21x21xui8) <- (1x12x21x21xf32, None, 1xf32) dropout_56, dropout_57 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_9, None, full_4, False, "upscale_in_train", 0, False + softmax_9, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_9 # pd_op.matmul: (1x12x21x64xf32) <- (1x12x21x21xf32, 1x12x21x64xf32) matmul_76 = paddle._C_ops.matmul(dropout_56, transpose_38, False, False) + del dropout_56, transpose_38 # pd_op.transpose: (1x21x12x64xf32) <- (1x12x21x64xf32) transpose_39 = paddle._C_ops.transpose(matmul_76, [0, 2, 1, 3]) @@ -1845,19 +1812,20 @@ def forward( # pd_op.reshape: (1x21x768xf32) <- (1x21x12x64xf32, 3xi64) reshape_39 = paddle._C_ops.reshape(transpose_39, full_int_array_2) + del transpose_39 # pd_op.matmul: (1x21x768xf32) <- (1x21x768xf32, 768x768xf32) matmul_77 = paddle._C_ops.matmul(reshape_39, parameter_43, False, False) - del parameter_43 + del parameter_43, reshape_39 # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_87 = paddle._C_ops.add(matmul_77, parameter_42) - del parameter_42 + del matmul_77, parameter_42 # pd_op.dropout: (1x21x768xf32, 1x21x768xui8) <- (1x21x768xf32, None, 1xf32) dropout_58, dropout_59 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_87, None, full_4, False, "upscale_in_train", 0, False + add_87, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1865,6 +1833,7 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 1x21x768xf32) add_88 = paddle._C_ops.add(layer_norm_54, dropout_58) + del dropout_58, layer_norm_54 # pd_op.layer_norm: (1x21x768xf32, 1x21xf32, 1x21xf32) <- (1x21x768xf32, 768xf32, 768xf32) layer_norm_57, layer_norm_58, layer_norm_59 = (lambda x, f: f(x))( @@ -1873,7 +1842,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_36, parameter_37 + del add_88, parameter_36, parameter_37 # pd_op.matmul: (1x21x3072xf32) <- (1x21x768xf32, 768x3072xf32) matmul_78 = paddle._C_ops.matmul(layer_norm_57, parameter_41, False, False) @@ -1881,23 +1850,24 @@ def forward( # pd_op.add: (1x21x3072xf32) <- (1x21x3072xf32, 3072xf32) add_89 = paddle._C_ops.add(matmul_78, parameter_40) - del parameter_40 + del matmul_78, parameter_40 # pd_op.gelu: (1x21x3072xf32) <- (1x21x3072xf32) gelu_9 = paddle._C_ops.gelu(add_89, False) + del add_89 # pd_op.matmul: (1x21x768xf32) <- (1x21x3072xf32, 3072x768xf32) matmul_79 = paddle._C_ops.matmul(gelu_9, parameter_39, False, False) - del parameter_39 + del gelu_9, parameter_39 # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_90 = paddle._C_ops.add(matmul_79, parameter_38) - del parameter_38 + del matmul_79, parameter_38 # pd_op.dropout: (1x21x768xf32, 1x21x768xui8) <- (1x21x768xf32, None, 1xf32) dropout_60, dropout_61 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_90, None, full_4, False, "upscale_in_train", 0, False + add_90, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1905,6 +1875,7 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 1x21x768xf32) add_91 = paddle._C_ops.add(layer_norm_57, dropout_60) + del dropout_60, layer_norm_57 # pd_op.layer_norm: (1x21x768xf32, 1x21xf32, 1x21xf32) <- (1x21x768xf32, 768xf32, 768xf32) layer_norm_60, layer_norm_61, layer_norm_62 = (lambda x, f: f(x))( @@ -1913,7 +1884,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_34, parameter_35 + del add_91, parameter_34, parameter_35 # pd_op.matmul: (1x21x768xf32) <- (1x21x768xf32, 768x768xf32) matmul_80 = paddle._C_ops.matmul(layer_norm_60, parameter_33, False, False) @@ -1921,10 +1892,11 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_92 = paddle._C_ops.add(matmul_80, parameter_32) - del parameter_32 + del matmul_80, parameter_32 # pd_op.reshape: (1x21x12x64xf32) <- (1x21x768xf32, 4xi64) reshape_40 = paddle._C_ops.reshape(add_92, full_int_array_1) + del add_92 # pd_op.transpose: (1x12x21x64xf32) <- (1x21x12x64xf32) transpose_40 = paddle._C_ops.transpose(reshape_40, [0, 2, 1, 3]) @@ -1936,7 +1908,7 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_93 = paddle._C_ops.add(matmul_81, parameter_30) - del parameter_30 + del matmul_81, parameter_30 # pd_op.matmul: (1x21x768xf32) <- (1x21x768xf32, 768x768xf32) matmul_82 = paddle._C_ops.matmul(layer_norm_60, parameter_29, False, False) @@ -1944,10 +1916,11 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_94 = paddle._C_ops.add(matmul_82, parameter_28) - del parameter_28 + del matmul_82, parameter_28 # pd_op.reshape: (1x21x12x64xf32) <- (1x21x768xf32, 4xi64) reshape_41 = paddle._C_ops.reshape(add_93, full_int_array_1) + del add_93 # pd_op.transpose: (1x12x21x64xf32) <- (1x21x12x64xf32) transpose_41 = paddle._C_ops.transpose(reshape_41, [0, 2, 1, 3]) @@ -1955,6 +1928,7 @@ def forward( # pd_op.reshape: (1x21x12x64xf32) <- (1x21x768xf32, 4xi64) reshape_42 = paddle._C_ops.reshape(add_94, full_int_array_1) + del add_94 # pd_op.transpose: (1x12x21x64xf32) <- (1x21x12x64xf32) transpose_42 = paddle._C_ops.transpose(reshape_42, [0, 2, 1, 3]) @@ -1966,9 +1940,11 @@ def forward( # pd_op.matmul: (1x12x21x21xf32) <- (1x12x21x64xf32, 1x12x21x64xf32) matmul_83 = paddle._C_ops.matmul(scale_11, transpose_41, False, True) + del scale_11, transpose_41 # pd_op.add: (1x12x21x21xf32) <- (1x12x21x21xf32, 1x1x1x21xf32) add_95 = paddle._C_ops.add(matmul_83, unsqueeze_0) + del matmul_83 # pd_op.softmax: (1x12x21x21xf32) <- (1x12x21x21xf32) softmax_10 = paddle._C_ops.softmax(add_95, -1) @@ -1977,13 +1953,15 @@ def forward( # pd_op.dropout: (1x12x21x21xf32, 1x12x21x21xui8) <- (1x12x21x21xf32, None, 1xf32) dropout_62, dropout_63 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_10, None, full_4, False, "upscale_in_train", 0, False + softmax_10, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_10 # pd_op.matmul: (1x12x21x64xf32) <- (1x12x21x21xf32, 1x12x21x64xf32) matmul_84 = paddle._C_ops.matmul(dropout_62, transpose_42, False, False) + del dropout_62, transpose_42 # pd_op.transpose: (1x21x12x64xf32) <- (1x12x21x64xf32) transpose_43 = paddle._C_ops.transpose(matmul_84, [0, 2, 1, 3]) @@ -1991,19 +1969,20 @@ def forward( # pd_op.reshape: (1x21x768xf32) <- (1x21x12x64xf32, 3xi64) reshape_43 = paddle._C_ops.reshape(transpose_43, full_int_array_2) + del transpose_43 # pd_op.matmul: (1x21x768xf32) <- (1x21x768xf32, 768x768xf32) matmul_85 = paddle._C_ops.matmul(reshape_43, parameter_27, False, False) - del parameter_27 + del parameter_27, reshape_43 # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_96 = paddle._C_ops.add(matmul_85, parameter_26) - del parameter_26 + del matmul_85, parameter_26 # pd_op.dropout: (1x21x768xf32, 1x21x768xui8) <- (1x21x768xf32, None, 1xf32) dropout_64, dropout_65 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_96, None, full_4, False, "upscale_in_train", 0, False + add_96, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -2011,6 +1990,7 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 1x21x768xf32) add_97 = paddle._C_ops.add(layer_norm_60, dropout_64) + del dropout_64, layer_norm_60 # pd_op.layer_norm: (1x21x768xf32, 1x21xf32, 1x21xf32) <- (1x21x768xf32, 768xf32, 768xf32) layer_norm_63, layer_norm_64, layer_norm_65 = (lambda x, f: f(x))( @@ -2019,7 +1999,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_20, parameter_21 + del add_97, parameter_20, parameter_21 # pd_op.matmul: (1x21x3072xf32) <- (1x21x768xf32, 768x3072xf32) matmul_86 = paddle._C_ops.matmul(layer_norm_63, parameter_25, False, False) @@ -2027,23 +2007,24 @@ def forward( # pd_op.add: (1x21x3072xf32) <- (1x21x3072xf32, 3072xf32) add_98 = paddle._C_ops.add(matmul_86, parameter_24) - del parameter_24 + del matmul_86, parameter_24 # pd_op.gelu: (1x21x3072xf32) <- (1x21x3072xf32) gelu_10 = paddle._C_ops.gelu(add_98, False) + del add_98 # pd_op.matmul: (1x21x768xf32) <- (1x21x3072xf32, 3072x768xf32) matmul_87 = paddle._C_ops.matmul(gelu_10, parameter_23, False, False) - del parameter_23 + del gelu_10, parameter_23 # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_99 = paddle._C_ops.add(matmul_87, parameter_22) - del parameter_22 + del matmul_87, parameter_22 # pd_op.dropout: (1x21x768xf32, 1x21x768xui8) <- (1x21x768xf32, None, 1xf32) dropout_66, dropout_67 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_99, None, full_4, False, "upscale_in_train", 0, False + add_99, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -2051,6 +2032,7 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 1x21x768xf32) add_100 = paddle._C_ops.add(layer_norm_63, dropout_66) + del dropout_66, layer_norm_63 # pd_op.layer_norm: (1x21x768xf32, 1x21xf32, 1x21xf32) <- (1x21x768xf32, 768xf32, 768xf32) layer_norm_66, layer_norm_67, layer_norm_68 = (lambda x, f: f(x))( @@ -2059,7 +2041,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_18, parameter_19 + del add_100, parameter_18, parameter_19 # pd_op.matmul: (1x21x768xf32) <- (1x21x768xf32, 768x768xf32) matmul_88 = paddle._C_ops.matmul(layer_norm_66, parameter_17, False, False) @@ -2067,10 +2049,11 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_101 = paddle._C_ops.add(matmul_88, parameter_16) - del parameter_16 + del matmul_88, parameter_16 # pd_op.reshape: (1x21x12x64xf32) <- (1x21x768xf32, 4xi64) reshape_44 = paddle._C_ops.reshape(add_101, full_int_array_1) + del add_101 # pd_op.transpose: (1x12x21x64xf32) <- (1x21x12x64xf32) transpose_44 = paddle._C_ops.transpose(reshape_44, [0, 2, 1, 3]) @@ -2082,7 +2065,7 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_102 = paddle._C_ops.add(matmul_89, parameter_14) - del parameter_14 + del matmul_89, parameter_14 # pd_op.matmul: (1x21x768xf32) <- (1x21x768xf32, 768x768xf32) matmul_90 = paddle._C_ops.matmul(layer_norm_66, parameter_13, False, False) @@ -2090,10 +2073,11 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_103 = paddle._C_ops.add(matmul_90, parameter_12) - del parameter_12 + del matmul_90, parameter_12 # pd_op.reshape: (1x21x12x64xf32) <- (1x21x768xf32, 4xi64) reshape_45 = paddle._C_ops.reshape(add_102, full_int_array_1) + del add_102 # pd_op.transpose: (1x12x21x64xf32) <- (1x21x12x64xf32) transpose_45 = paddle._C_ops.transpose(reshape_45, [0, 2, 1, 3]) @@ -2101,7 +2085,7 @@ def forward( # pd_op.reshape: (1x21x12x64xf32) <- (1x21x768xf32, 4xi64) reshape_46 = paddle._C_ops.reshape(add_103, full_int_array_1) - del full_int_array_1 + del add_103, full_int_array_1 # pd_op.transpose: (1x12x21x64xf32) <- (1x21x12x64xf32) transpose_46 = paddle._C_ops.transpose(reshape_46, [0, 2, 1, 3]) @@ -2109,13 +2093,15 @@ def forward( # pd_op.scale: (1x12x21x64xf32) <- (1x12x21x64xf32, 1xf32) scale_12 = paddle._C_ops.scale(transpose_44, full_5, float("0"), True) - del transpose_44 + del full_5, transpose_44 # pd_op.matmul: (1x12x21x21xf32) <- (1x12x21x64xf32, 1x12x21x64xf32) matmul_91 = paddle._C_ops.matmul(scale_12, transpose_45, False, True) + del scale_12, transpose_45 # pd_op.add: (1x12x21x21xf32) <- (1x12x21x21xf32, 1x1x1x21xf32) add_104 = paddle._C_ops.add(matmul_91, unsqueeze_0) + del matmul_91, unsqueeze_0 # pd_op.softmax: (1x12x21x21xf32) <- (1x12x21x21xf32) softmax_11 = paddle._C_ops.softmax(add_104, -1) @@ -2124,13 +2110,15 @@ def forward( # pd_op.dropout: (1x12x21x21xf32, 1x12x21x21xui8) <- (1x12x21x21xf32, None, 1xf32) dropout_68, dropout_69 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_11, None, full_4, False, "upscale_in_train", 0, False + softmax_11, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_11 # pd_op.matmul: (1x12x21x64xf32) <- (1x12x21x21xf32, 1x12x21x64xf32) matmul_92 = paddle._C_ops.matmul(dropout_68, transpose_46, False, False) + del dropout_68, transpose_46 # pd_op.transpose: (1x21x12x64xf32) <- (1x12x21x64xf32) transpose_47 = paddle._C_ops.transpose(matmul_92, [0, 2, 1, 3]) @@ -2138,20 +2126,20 @@ def forward( # pd_op.reshape: (1x21x768xf32) <- (1x21x12x64xf32, 3xi64) reshape_47 = paddle._C_ops.reshape(transpose_47, full_int_array_2) - del full_int_array_2 + del full_int_array_2, transpose_47 # pd_op.matmul: (1x21x768xf32) <- (1x21x768xf32, 768x768xf32) matmul_93 = paddle._C_ops.matmul(reshape_47, parameter_11, False, False) - del parameter_11 + del parameter_11, reshape_47 # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_105 = paddle._C_ops.add(matmul_93, parameter_10) - del parameter_10 + del matmul_93, parameter_10 # pd_op.dropout: (1x21x768xf32, 1x21x768xui8) <- (1x21x768xf32, None, 1xf32) dropout_70, dropout_71 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_105, None, full_4, False, "upscale_in_train", 0, False + add_105, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -2159,6 +2147,7 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 1x21x768xf32) add_106 = paddle._C_ops.add(layer_norm_66, dropout_70) + del dropout_70, layer_norm_66 # pd_op.layer_norm: (1x21x768xf32, 1x21xf32, 1x21xf32) <- (1x21x768xf32, 768xf32, 768xf32) layer_norm_69, layer_norm_70, layer_norm_71 = (lambda x, f: f(x))( @@ -2167,7 +2156,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_4, parameter_5 + del add_106, parameter_4, parameter_5 # pd_op.matmul: (1x21x3072xf32) <- (1x21x768xf32, 768x3072xf32) matmul_94 = paddle._C_ops.matmul(layer_norm_69, parameter_9, False, False) @@ -2175,30 +2164,32 @@ def forward( # pd_op.add: (1x21x3072xf32) <- (1x21x3072xf32, 3072xf32) add_107 = paddle._C_ops.add(matmul_94, parameter_8) - del parameter_8 + del matmul_94, parameter_8 # pd_op.gelu: (1x21x3072xf32) <- (1x21x3072xf32) gelu_11 = paddle._C_ops.gelu(add_107, False) + del add_107 # pd_op.matmul: (1x21x768xf32) <- (1x21x3072xf32, 3072x768xf32) matmul_95 = paddle._C_ops.matmul(gelu_11, parameter_7, False, False) - del parameter_7 + del gelu_11, parameter_7 # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_108 = paddle._C_ops.add(matmul_95, parameter_6) - del parameter_6 + del matmul_95, parameter_6 # pd_op.dropout: (1x21x768xf32, 1x21x768xui8) <- (1x21x768xf32, None, 1xf32) dropout_72, dropout_73 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_108, None, full_4, False, "upscale_in_train", 0, False + add_108, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) - del add_108 + del add_108, full_4 # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 1x21x768xf32) add_109 = paddle._C_ops.add(layer_norm_69, dropout_72) + del dropout_72, layer_norm_69 # pd_op.layer_norm: (1x21x768xf32, 1x21xf32, 1x21xf32) <- (1x21x768xf32, 768xf32, 768xf32) layer_norm_72, layer_norm_73, layer_norm_74 = (lambda x, f: f(x))( @@ -2207,7 +2198,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_2, parameter_3 + del add_109, parameter_2, parameter_3 # pd_op.full_int_array: (1xi64) <- () full_int_array_3 = [0] @@ -2219,467 +2210,18 @@ def forward( slice_0 = paddle._C_ops.slice( layer_norm_72, [1], full_int_array_3, full_int_array_4, [1], [1] ) + del full_int_array_3, full_int_array_4, layer_norm_72 # pd_op.matmul: (1x768xf32) <- (1x768xf32, 768x768xf32) matmul_96 = paddle._C_ops.matmul(slice_0, parameter_1, False, False) - del parameter_1 + del parameter_1, slice_0 # pd_op.add: (1x768xf32) <- (1x768xf32, 768xf32) add_110 = paddle._C_ops.add(matmul_96, parameter_0) - del parameter_0 + del matmul_96, parameter_0 # pd_op.tanh: (1x768xf32) <- (1x768xf32) tanh_0 = paddle._C_ops.tanh(add_110) - del ( - add_0, - add_1, - add_10, - add_100, - add_101, - add_102, - add_103, - add_106, - add_107, - add_109, - add_11, - add_110, - add_12, - add_13, - add_16, - add_17, - add_19, - add_2, - add_20, - add_21, - add_22, - add_25, - add_26, - add_28, - add_29, - add_3, - add_30, - add_31, - add_34, - add_35, - add_37, - add_38, - add_39, - add_4, - add_40, - add_43, - add_44, - add_46, - add_47, - add_48, - add_49, - add_52, - add_53, - add_55, - add_56, - add_57, - add_58, - add_61, - add_62, - add_64, - add_65, - add_66, - add_67, - add_7, - add_70, - add_71, - add_73, - add_74, - add_75, - add_76, - add_79, - add_8, - add_80, - add_82, - add_83, - add_84, - add_85, - add_88, - add_89, - add_91, - add_92, - add_93, - add_94, - add_97, - add_98, - assign_0, - assign_1, - assign_10, - assign_11, - assign_12, - assign_13, - assign_14, - assign_15, - assign_16, - assign_17, - assign_18, - assign_19, - assign_2, - assign_20, - assign_21, - assign_22, - assign_23, - assign_24, - assign_25, - assign_26, - assign_27, - assign_28, - assign_29, - assign_3, - assign_30, - assign_31, - assign_32, - assign_33, - assign_34, - assign_35, - assign_36, - assign_37, - assign_38, - assign_39, - assign_4, - assign_40, - assign_41, - assign_42, - assign_43, - assign_44, - assign_45, - assign_46, - assign_5, - assign_6, - assign_7, - assign_8, - assign_9, - dropout_0, - dropout_1, - dropout_10, - dropout_11, - dropout_12, - dropout_13, - dropout_14, - dropout_15, - dropout_16, - dropout_17, - dropout_18, - dropout_19, - dropout_2, - dropout_20, - dropout_21, - dropout_22, - dropout_23, - dropout_24, - dropout_25, - dropout_26, - dropout_27, - dropout_28, - dropout_29, - dropout_3, - dropout_30, - dropout_31, - dropout_32, - dropout_33, - dropout_34, - dropout_35, - dropout_36, - dropout_37, - dropout_38, - dropout_39, - dropout_4, - dropout_40, - dropout_41, - dropout_42, - dropout_43, - dropout_44, - dropout_45, - dropout_46, - dropout_47, - dropout_48, - dropout_49, - dropout_5, - dropout_50, - dropout_51, - dropout_52, - dropout_53, - dropout_54, - dropout_55, - dropout_56, - dropout_57, - dropout_58, - dropout_59, - dropout_6, - dropout_60, - dropout_61, - dropout_62, - dropout_63, - dropout_64, - dropout_65, - dropout_66, - dropout_67, - dropout_68, - dropout_69, - dropout_7, - dropout_70, - dropout_71, - dropout_72, - dropout_73, - dropout_8, - dropout_9, - embedding_0, - embedding_1, - embedding_2, - full_4, - full_5, - full_int_array_3, - full_int_array_4, - gelu_0, - gelu_1, - gelu_10, - gelu_11, - gelu_2, - gelu_3, - gelu_4, - gelu_5, - gelu_6, - gelu_7, - gelu_8, - gelu_9, - layer_norm_1, - layer_norm_10, - layer_norm_11, - layer_norm_12, - layer_norm_13, - layer_norm_14, - layer_norm_15, - layer_norm_16, - layer_norm_17, - layer_norm_18, - layer_norm_19, - layer_norm_2, - layer_norm_20, - layer_norm_21, - layer_norm_22, - layer_norm_23, - layer_norm_24, - layer_norm_25, - layer_norm_26, - layer_norm_27, - layer_norm_28, - layer_norm_29, - layer_norm_3, - layer_norm_30, - layer_norm_31, - layer_norm_32, - layer_norm_33, - layer_norm_34, - layer_norm_35, - layer_norm_36, - layer_norm_37, - layer_norm_38, - layer_norm_39, - layer_norm_4, - layer_norm_40, - layer_norm_41, - layer_norm_42, - layer_norm_43, - layer_norm_44, - layer_norm_45, - layer_norm_46, - layer_norm_47, - layer_norm_48, - layer_norm_49, - layer_norm_5, - layer_norm_50, - layer_norm_51, - layer_norm_52, - layer_norm_53, - layer_norm_54, - layer_norm_55, - layer_norm_56, - layer_norm_57, - layer_norm_58, - layer_norm_59, - layer_norm_6, - layer_norm_60, - layer_norm_61, - layer_norm_62, - layer_norm_63, - layer_norm_64, - layer_norm_65, - layer_norm_66, - layer_norm_67, - layer_norm_68, - layer_norm_69, - layer_norm_7, - layer_norm_70, - layer_norm_71, - layer_norm_72, - layer_norm_73, - layer_norm_74, - layer_norm_8, - layer_norm_9, - matmul_0, - matmul_1, - matmul_10, - matmul_11, - matmul_13, - matmul_14, - matmul_15, - matmul_16, - matmul_17, - matmul_18, - matmul_19, - matmul_2, - matmul_21, - matmul_22, - matmul_23, - matmul_24, - matmul_25, - matmul_26, - matmul_27, - matmul_29, - matmul_3, - matmul_30, - matmul_31, - matmul_32, - matmul_33, - matmul_34, - matmul_35, - matmul_37, - matmul_38, - matmul_39, - matmul_40, - matmul_41, - matmul_42, - matmul_43, - matmul_45, - matmul_46, - matmul_47, - matmul_48, - matmul_49, - matmul_5, - matmul_50, - matmul_51, - matmul_53, - matmul_54, - matmul_55, - matmul_56, - matmul_57, - matmul_58, - matmul_59, - matmul_6, - matmul_61, - matmul_62, - matmul_63, - matmul_64, - matmul_65, - matmul_66, - matmul_67, - matmul_69, - matmul_7, - matmul_70, - matmul_71, - matmul_72, - matmul_73, - matmul_74, - matmul_75, - matmul_77, - matmul_78, - matmul_79, - matmul_8, - matmul_80, - matmul_81, - matmul_82, - matmul_83, - matmul_85, - matmul_86, - matmul_87, - matmul_88, - matmul_89, - matmul_9, - matmul_90, - matmul_91, - matmul_93, - matmul_94, - matmul_95, - matmul_96, - reshape_11, - reshape_15, - reshape_19, - reshape_23, - reshape_27, - reshape_3, - reshape_31, - reshape_35, - reshape_39, - reshape_43, - reshape_47, - reshape_7, - scale_1, - scale_10, - scale_11, - scale_12, - scale_2, - scale_3, - scale_4, - scale_5, - scale_6, - scale_7, - scale_8, - scale_9, - slice_0, - softmax_0, - softmax_1, - softmax_10, - softmax_11, - softmax_2, - softmax_3, - softmax_4, - softmax_5, - softmax_6, - softmax_7, - softmax_8, - softmax_9, - subtract_0, - transpose_1, - transpose_10, - transpose_11, - transpose_13, - transpose_14, - transpose_15, - transpose_17, - transpose_18, - transpose_19, - transpose_2, - transpose_21, - transpose_22, - transpose_23, - transpose_25, - transpose_26, - transpose_27, - transpose_29, - transpose_3, - transpose_30, - transpose_31, - transpose_33, - transpose_34, - transpose_35, - transpose_37, - transpose_38, - transpose_39, - transpose_41, - transpose_42, - transpose_43, - transpose_45, - transpose_46, - transpose_47, - transpose_5, - transpose_6, - transpose_7, - transpose_9, - unsqueeze_0, - ) + del add_110 return tanh_0 diff --git a/paddle_samples/PaddleNLP/bert-base-cased/weight_meta.py b/paddle_samples/PaddleNLP/bert-base-cased/weight_meta.py index ad85e1ff2..bc0904edd 100644 --- a/paddle_samples/PaddleNLP/bert-base-cased/weight_meta.py +++ b/paddle_samples/PaddleNLP/bert-base-cased/weight_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_parameter_0: name = "parameter_0" + original_name = "linear_72.b_0" shape = [768] dtype = "float32" min_val = float("-0.0799701") @@ -11,6 +12,7 @@ class Program_weight_tensor_parameter_0: class Program_weight_tensor_parameter_1: name = "parameter_1" + original_name = "linear_72.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.229597") @@ -22,6 +24,7 @@ class Program_weight_tensor_parameter_1: class Program_weight_tensor_parameter_2: name = "parameter_2" + original_name = "layer_norm_24.b_0" shape = [768] dtype = "float32" min_val = float("-0.175058") @@ -33,6 +36,7 @@ class Program_weight_tensor_parameter_2: class Program_weight_tensor_parameter_3: name = "parameter_3" + original_name = "layer_norm_24.w_0" shape = [768] dtype = "float32" min_val = float("0.375238") @@ -44,6 +48,7 @@ class Program_weight_tensor_parameter_3: class Program_weight_tensor_parameter_4: name = "parameter_4" + original_name = "layer_norm_23.b_0" shape = [768] dtype = "float32" min_val = float("-1.49129") @@ -55,6 +60,7 @@ class Program_weight_tensor_parameter_4: class Program_weight_tensor_parameter_5: name = "parameter_5" + original_name = "layer_norm_23.w_0" shape = [768] dtype = "float32" min_val = float("0.805317") @@ -66,6 +72,7 @@ class Program_weight_tensor_parameter_5: class Program_weight_tensor_parameter_6: name = "parameter_6" + original_name = "linear_71.b_0" shape = [768] dtype = "float32" min_val = float("-0.539593") @@ -77,6 +84,7 @@ class Program_weight_tensor_parameter_6: class Program_weight_tensor_parameter_7: name = "parameter_7" + original_name = "linear_71.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-0.598362") @@ -88,6 +96,7 @@ class Program_weight_tensor_parameter_7: class Program_weight_tensor_parameter_8: name = "parameter_8" + original_name = "linear_70.b_0" shape = [3072] dtype = "float32" min_val = float("-0.369677") @@ -99,6 +108,7 @@ class Program_weight_tensor_parameter_8: class Program_weight_tensor_parameter_9: name = "parameter_9" + original_name = "linear_70.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.31387") @@ -110,6 +120,7 @@ class Program_weight_tensor_parameter_9: class Program_weight_tensor_parameter_10: name = "parameter_10" + original_name = "linear_69.b_0" shape = [768] dtype = "float32" min_val = float("-0.119822") @@ -121,6 +132,7 @@ class Program_weight_tensor_parameter_10: class Program_weight_tensor_parameter_11: name = "parameter_11" + original_name = "linear_69.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.416684") @@ -132,6 +144,7 @@ class Program_weight_tensor_parameter_11: class Program_weight_tensor_parameter_12: name = "parameter_12" + original_name = "linear_68.b_0" shape = [768] dtype = "float32" min_val = float("-0.0562481") @@ -143,6 +156,7 @@ class Program_weight_tensor_parameter_12: class Program_weight_tensor_parameter_13: name = "parameter_13" + original_name = "linear_68.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.222997") @@ -154,6 +168,7 @@ class Program_weight_tensor_parameter_13: class Program_weight_tensor_parameter_14: name = "parameter_14" + original_name = "linear_67.b_0" shape = [768] dtype = "float32" min_val = float("-0.013506") @@ -165,6 +180,7 @@ class Program_weight_tensor_parameter_14: class Program_weight_tensor_parameter_15: name = "parameter_15" + original_name = "linear_67.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.378095") @@ -176,6 +192,7 @@ class Program_weight_tensor_parameter_15: class Program_weight_tensor_parameter_16: name = "parameter_16" + original_name = "linear_66.b_0" shape = [768] dtype = "float32" min_val = float("-0.542432") @@ -187,6 +204,7 @@ class Program_weight_tensor_parameter_16: class Program_weight_tensor_parameter_17: name = "parameter_17" + original_name = "linear_66.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.273163") @@ -198,6 +216,7 @@ class Program_weight_tensor_parameter_17: class Program_weight_tensor_parameter_18: name = "parameter_18" + original_name = "layer_norm_22.b_0" shape = [768] dtype = "float32" min_val = float("-0.813338") @@ -209,6 +228,7 @@ class Program_weight_tensor_parameter_18: class Program_weight_tensor_parameter_19: name = "parameter_19" + original_name = "layer_norm_22.w_0" shape = [768] dtype = "float32" min_val = float("0.12822") @@ -220,6 +240,7 @@ class Program_weight_tensor_parameter_19: class Program_weight_tensor_parameter_20: name = "parameter_20" + original_name = "layer_norm_21.b_0" shape = [768] dtype = "float32" min_val = float("-1.60877") @@ -231,6 +252,7 @@ class Program_weight_tensor_parameter_20: class Program_weight_tensor_parameter_21: name = "parameter_21" + original_name = "layer_norm_21.w_0" shape = [768] dtype = "float32" min_val = float("0.820765") @@ -242,6 +264,7 @@ class Program_weight_tensor_parameter_21: class Program_weight_tensor_parameter_22: name = "parameter_22" + original_name = "linear_65.b_0" shape = [768] dtype = "float32" min_val = float("-0.577303") @@ -253,6 +276,7 @@ class Program_weight_tensor_parameter_22: class Program_weight_tensor_parameter_23: name = "parameter_23" + original_name = "linear_65.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-3.37904") @@ -264,6 +288,7 @@ class Program_weight_tensor_parameter_23: class Program_weight_tensor_parameter_24: name = "parameter_24" + original_name = "linear_64.b_0" shape = [3072] dtype = "float32" min_val = float("-0.200296") @@ -275,6 +300,7 @@ class Program_weight_tensor_parameter_24: class Program_weight_tensor_parameter_25: name = "parameter_25" + original_name = "linear_64.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.355452") @@ -286,6 +312,7 @@ class Program_weight_tensor_parameter_25: class Program_weight_tensor_parameter_26: name = "parameter_26" + original_name = "linear_63.b_0" shape = [768] dtype = "float32" min_val = float("-0.148465") @@ -297,6 +324,7 @@ class Program_weight_tensor_parameter_26: class Program_weight_tensor_parameter_27: name = "parameter_27" + original_name = "linear_63.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.327008") @@ -308,6 +336,7 @@ class Program_weight_tensor_parameter_27: class Program_weight_tensor_parameter_28: name = "parameter_28" + original_name = "linear_62.b_0" shape = [768] dtype = "float32" min_val = float("-0.112973") @@ -319,6 +348,7 @@ class Program_weight_tensor_parameter_28: class Program_weight_tensor_parameter_29: name = "parameter_29" + original_name = "linear_62.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.179322") @@ -330,6 +360,7 @@ class Program_weight_tensor_parameter_29: class Program_weight_tensor_parameter_30: name = "parameter_30" + original_name = "linear_61.b_0" shape = [768] dtype = "float32" min_val = float("-0.0127674") @@ -341,6 +372,7 @@ class Program_weight_tensor_parameter_30: class Program_weight_tensor_parameter_31: name = "parameter_31" + original_name = "linear_61.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.224773") @@ -352,6 +384,7 @@ class Program_weight_tensor_parameter_31: class Program_weight_tensor_parameter_32: name = "parameter_32" + original_name = "linear_60.b_0" shape = [768] dtype = "float32" min_val = float("-0.570541") @@ -363,6 +396,7 @@ class Program_weight_tensor_parameter_32: class Program_weight_tensor_parameter_33: name = "parameter_33" + original_name = "linear_60.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.256972") @@ -374,6 +408,7 @@ class Program_weight_tensor_parameter_33: class Program_weight_tensor_parameter_34: name = "parameter_34" + original_name = "layer_norm_20.b_0" shape = [768] dtype = "float32" min_val = float("-0.532742") @@ -385,6 +420,7 @@ class Program_weight_tensor_parameter_34: class Program_weight_tensor_parameter_35: name = "parameter_35" + original_name = "layer_norm_20.w_0" shape = [768] dtype = "float32" min_val = float("0.362089") @@ -396,6 +432,7 @@ class Program_weight_tensor_parameter_35: class Program_weight_tensor_parameter_36: name = "parameter_36" + original_name = "layer_norm_19.b_0" shape = [768] dtype = "float32" min_val = float("-1.13946") @@ -407,6 +444,7 @@ class Program_weight_tensor_parameter_36: class Program_weight_tensor_parameter_37: name = "parameter_37" + original_name = "layer_norm_19.w_0" shape = [768] dtype = "float32" min_val = float("0.814329") @@ -418,6 +456,7 @@ class Program_weight_tensor_parameter_37: class Program_weight_tensor_parameter_38: name = "parameter_38" + original_name = "linear_59.b_0" shape = [768] dtype = "float32" min_val = float("-0.595962") @@ -429,6 +468,7 @@ class Program_weight_tensor_parameter_38: class Program_weight_tensor_parameter_39: name = "parameter_39" + original_name = "linear_59.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-3.96064") @@ -440,6 +480,7 @@ class Program_weight_tensor_parameter_39: class Program_weight_tensor_parameter_40: name = "parameter_40" + original_name = "linear_58.b_0" shape = [3072] dtype = "float32" min_val = float("-0.205959") @@ -451,6 +492,7 @@ class Program_weight_tensor_parameter_40: class Program_weight_tensor_parameter_41: name = "parameter_41" + original_name = "linear_58.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.818119") @@ -462,6 +504,7 @@ class Program_weight_tensor_parameter_41: class Program_weight_tensor_parameter_42: name = "parameter_42" + original_name = "linear_57.b_0" shape = [768] dtype = "float32" min_val = float("-0.137153") @@ -473,6 +516,7 @@ class Program_weight_tensor_parameter_42: class Program_weight_tensor_parameter_43: name = "parameter_43" + original_name = "linear_57.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.351773") @@ -484,6 +528,7 @@ class Program_weight_tensor_parameter_43: class Program_weight_tensor_parameter_44: name = "parameter_44" + original_name = "linear_56.b_0" shape = [768] dtype = "float32" min_val = float("-0.145221") @@ -495,6 +540,7 @@ class Program_weight_tensor_parameter_44: class Program_weight_tensor_parameter_45: name = "parameter_45" + original_name = "linear_56.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.172291") @@ -506,6 +552,7 @@ class Program_weight_tensor_parameter_45: class Program_weight_tensor_parameter_46: name = "parameter_46" + original_name = "linear_55.b_0" shape = [768] dtype = "float32" min_val = float("-0.0164098") @@ -517,6 +564,7 @@ class Program_weight_tensor_parameter_46: class Program_weight_tensor_parameter_47: name = "parameter_47" + original_name = "linear_55.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.218005") @@ -528,6 +576,7 @@ class Program_weight_tensor_parameter_47: class Program_weight_tensor_parameter_48: name = "parameter_48" + original_name = "linear_54.b_0" shape = [768] dtype = "float32" min_val = float("-0.345731") @@ -539,6 +588,7 @@ class Program_weight_tensor_parameter_48: class Program_weight_tensor_parameter_49: name = "parameter_49" + original_name = "linear_54.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.194926") @@ -550,6 +600,7 @@ class Program_weight_tensor_parameter_49: class Program_weight_tensor_parameter_50: name = "parameter_50" + original_name = "layer_norm_18.b_0" shape = [768] dtype = "float32" min_val = float("-0.484985") @@ -561,6 +612,7 @@ class Program_weight_tensor_parameter_50: class Program_weight_tensor_parameter_51: name = "parameter_51" + original_name = "layer_norm_18.w_0" shape = [768] dtype = "float32" min_val = float("0.307961") @@ -572,6 +624,7 @@ class Program_weight_tensor_parameter_51: class Program_weight_tensor_parameter_52: name = "parameter_52" + original_name = "layer_norm_17.b_0" shape = [768] dtype = "float32" min_val = float("-0.998358") @@ -583,6 +636,7 @@ class Program_weight_tensor_parameter_52: class Program_weight_tensor_parameter_53: name = "parameter_53" + original_name = "layer_norm_17.w_0" shape = [768] dtype = "float32" min_val = float("0.808405") @@ -594,6 +648,7 @@ class Program_weight_tensor_parameter_53: class Program_weight_tensor_parameter_54: name = "parameter_54" + original_name = "linear_53.b_0" shape = [768] dtype = "float32" min_val = float("-0.697645") @@ -605,6 +660,7 @@ class Program_weight_tensor_parameter_54: class Program_weight_tensor_parameter_55: name = "parameter_55" + original_name = "linear_53.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-4.59653") @@ -616,6 +672,7 @@ class Program_weight_tensor_parameter_55: class Program_weight_tensor_parameter_56: name = "parameter_56" + original_name = "linear_52.b_0" shape = [3072] dtype = "float32" min_val = float("-0.283332") @@ -627,6 +684,7 @@ class Program_weight_tensor_parameter_56: class Program_weight_tensor_parameter_57: name = "parameter_57" + original_name = "linear_52.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.546636") @@ -638,6 +696,7 @@ class Program_weight_tensor_parameter_57: class Program_weight_tensor_parameter_58: name = "parameter_58" + original_name = "linear_51.b_0" shape = [768] dtype = "float32" min_val = float("-0.14698") @@ -649,6 +708,7 @@ class Program_weight_tensor_parameter_58: class Program_weight_tensor_parameter_59: name = "parameter_59" + original_name = "linear_51.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.215355") @@ -660,6 +720,7 @@ class Program_weight_tensor_parameter_59: class Program_weight_tensor_parameter_60: name = "parameter_60" + original_name = "linear_50.b_0" shape = [768] dtype = "float32" min_val = float("-0.145905") @@ -671,6 +732,7 @@ class Program_weight_tensor_parameter_60: class Program_weight_tensor_parameter_61: name = "parameter_61" + original_name = "linear_50.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.18626") @@ -682,6 +744,7 @@ class Program_weight_tensor_parameter_61: class Program_weight_tensor_parameter_62: name = "parameter_62" + original_name = "linear_49.b_0" shape = [768] dtype = "float32" min_val = float("-0.014998") @@ -693,6 +756,7 @@ class Program_weight_tensor_parameter_62: class Program_weight_tensor_parameter_63: name = "parameter_63" + original_name = "linear_49.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.250592") @@ -704,6 +768,7 @@ class Program_weight_tensor_parameter_63: class Program_weight_tensor_parameter_64: name = "parameter_64" + original_name = "linear_48.b_0" shape = [768] dtype = "float32" min_val = float("-0.431732") @@ -715,6 +780,7 @@ class Program_weight_tensor_parameter_64: class Program_weight_tensor_parameter_65: name = "parameter_65" + original_name = "linear_48.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.217635") @@ -726,6 +792,7 @@ class Program_weight_tensor_parameter_65: class Program_weight_tensor_parameter_66: name = "parameter_66" + original_name = "layer_norm_16.b_0" shape = [768] dtype = "float32" min_val = float("-0.52343") @@ -737,6 +804,7 @@ class Program_weight_tensor_parameter_66: class Program_weight_tensor_parameter_67: name = "parameter_67" + original_name = "layer_norm_16.w_0" shape = [768] dtype = "float32" min_val = float("0.199035") @@ -748,6 +816,7 @@ class Program_weight_tensor_parameter_67: class Program_weight_tensor_parameter_68: name = "parameter_68" + original_name = "layer_norm_15.b_0" shape = [768] dtype = "float32" min_val = float("-1.05465") @@ -759,6 +828,7 @@ class Program_weight_tensor_parameter_68: class Program_weight_tensor_parameter_69: name = "parameter_69" + original_name = "layer_norm_15.w_0" shape = [768] dtype = "float32" min_val = float("0.793999") @@ -770,6 +840,7 @@ class Program_weight_tensor_parameter_69: class Program_weight_tensor_parameter_70: name = "parameter_70" + original_name = "linear_47.b_0" shape = [768] dtype = "float32" min_val = float("-0.791517") @@ -781,6 +852,7 @@ class Program_weight_tensor_parameter_70: class Program_weight_tensor_parameter_71: name = "parameter_71" + original_name = "linear_47.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-2.78969") @@ -792,6 +864,7 @@ class Program_weight_tensor_parameter_71: class Program_weight_tensor_parameter_72: name = "parameter_72" + original_name = "linear_46.b_0" shape = [3072] dtype = "float32" min_val = float("-0.31943") @@ -803,6 +876,7 @@ class Program_weight_tensor_parameter_72: class Program_weight_tensor_parameter_73: name = "parameter_73" + original_name = "linear_46.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.370931") @@ -814,6 +888,7 @@ class Program_weight_tensor_parameter_73: class Program_weight_tensor_parameter_74: name = "parameter_74" + original_name = "linear_45.b_0" shape = [768] dtype = "float32" min_val = float("-0.132828") @@ -825,6 +900,7 @@ class Program_weight_tensor_parameter_74: class Program_weight_tensor_parameter_75: name = "parameter_75" + original_name = "linear_45.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.262555") @@ -836,6 +912,7 @@ class Program_weight_tensor_parameter_75: class Program_weight_tensor_parameter_76: name = "parameter_76" + original_name = "linear_44.b_0" shape = [768] dtype = "float32" min_val = float("-0.188478") @@ -847,6 +924,7 @@ class Program_weight_tensor_parameter_76: class Program_weight_tensor_parameter_77: name = "parameter_77" + original_name = "linear_44.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.156855") @@ -858,6 +936,7 @@ class Program_weight_tensor_parameter_77: class Program_weight_tensor_parameter_78: name = "parameter_78" + original_name = "linear_43.b_0" shape = [768] dtype = "float32" min_val = float("-0.0204448") @@ -869,6 +948,7 @@ class Program_weight_tensor_parameter_78: class Program_weight_tensor_parameter_79: name = "parameter_79" + original_name = "linear_43.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.336674") @@ -880,6 +960,7 @@ class Program_weight_tensor_parameter_79: class Program_weight_tensor_parameter_80: name = "parameter_80" + original_name = "linear_42.b_0" shape = [768] dtype = "float32" min_val = float("-0.465063") @@ -891,6 +972,7 @@ class Program_weight_tensor_parameter_80: class Program_weight_tensor_parameter_81: name = "parameter_81" + original_name = "linear_42.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.237067") @@ -902,6 +984,7 @@ class Program_weight_tensor_parameter_81: class Program_weight_tensor_parameter_82: name = "parameter_82" + original_name = "layer_norm_14.b_0" shape = [768] dtype = "float32" min_val = float("-0.480276") @@ -913,6 +996,7 @@ class Program_weight_tensor_parameter_82: class Program_weight_tensor_parameter_83: name = "parameter_83" + original_name = "layer_norm_14.w_0" shape = [768] dtype = "float32" min_val = float("0.156593") @@ -924,6 +1008,7 @@ class Program_weight_tensor_parameter_83: class Program_weight_tensor_parameter_84: name = "parameter_84" + original_name = "layer_norm_13.b_0" shape = [768] dtype = "float32" min_val = float("-1.01171") @@ -935,6 +1020,7 @@ class Program_weight_tensor_parameter_84: class Program_weight_tensor_parameter_85: name = "parameter_85" + original_name = "layer_norm_13.w_0" shape = [768] dtype = "float32" min_val = float("0.743275") @@ -946,6 +1032,7 @@ class Program_weight_tensor_parameter_85: class Program_weight_tensor_parameter_86: name = "parameter_86" + original_name = "linear_41.b_0" shape = [768] dtype = "float32" min_val = float("-0.740611") @@ -957,6 +1044,7 @@ class Program_weight_tensor_parameter_86: class Program_weight_tensor_parameter_87: name = "parameter_87" + original_name = "linear_41.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-1.10993") @@ -968,6 +1056,7 @@ class Program_weight_tensor_parameter_87: class Program_weight_tensor_parameter_88: name = "parameter_88" + original_name = "linear_40.b_0" shape = [3072] dtype = "float32" min_val = float("-0.323713") @@ -979,6 +1068,7 @@ class Program_weight_tensor_parameter_88: class Program_weight_tensor_parameter_89: name = "parameter_89" + original_name = "linear_40.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.280618") @@ -990,6 +1080,7 @@ class Program_weight_tensor_parameter_89: class Program_weight_tensor_parameter_90: name = "parameter_90" + original_name = "linear_39.b_0" shape = [768] dtype = "float32" min_val = float("-0.136652") @@ -1001,6 +1092,7 @@ class Program_weight_tensor_parameter_90: class Program_weight_tensor_parameter_91: name = "parameter_91" + original_name = "linear_39.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.31155") @@ -1012,6 +1104,7 @@ class Program_weight_tensor_parameter_91: class Program_weight_tensor_parameter_92: name = "parameter_92" + original_name = "linear_38.b_0" shape = [768] dtype = "float32" min_val = float("-0.0913504") @@ -1023,6 +1116,7 @@ class Program_weight_tensor_parameter_92: class Program_weight_tensor_parameter_93: name = "parameter_93" + original_name = "linear_38.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.170674") @@ -1034,6 +1128,7 @@ class Program_weight_tensor_parameter_93: class Program_weight_tensor_parameter_94: name = "parameter_94" + original_name = "linear_37.b_0" shape = [768] dtype = "float32" min_val = float("-0.00918644") @@ -1045,6 +1140,7 @@ class Program_weight_tensor_parameter_94: class Program_weight_tensor_parameter_95: name = "parameter_95" + original_name = "linear_37.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.298481") @@ -1056,6 +1152,7 @@ class Program_weight_tensor_parameter_95: class Program_weight_tensor_parameter_96: name = "parameter_96" + original_name = "linear_36.b_0" shape = [768] dtype = "float32" min_val = float("-0.444081") @@ -1067,6 +1164,7 @@ class Program_weight_tensor_parameter_96: class Program_weight_tensor_parameter_97: name = "parameter_97" + original_name = "linear_36.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.259543") @@ -1078,6 +1176,7 @@ class Program_weight_tensor_parameter_97: class Program_weight_tensor_parameter_98: name = "parameter_98" + original_name = "layer_norm_12.b_0" shape = [768] dtype = "float32" min_val = float("-0.493288") @@ -1089,6 +1188,7 @@ class Program_weight_tensor_parameter_98: class Program_weight_tensor_parameter_99: name = "parameter_99" + original_name = "layer_norm_12.w_0" shape = [768] dtype = "float32" min_val = float("0.170401") @@ -1100,6 +1200,7 @@ class Program_weight_tensor_parameter_99: class Program_weight_tensor_parameter_100: name = "parameter_100" + original_name = "layer_norm_11.b_0" shape = [768] dtype = "float32" min_val = float("-1.05787") @@ -1111,6 +1212,7 @@ class Program_weight_tensor_parameter_100: class Program_weight_tensor_parameter_101: name = "parameter_101" + original_name = "layer_norm_11.w_0" shape = [768] dtype = "float32" min_val = float("0.75461") @@ -1122,6 +1224,7 @@ class Program_weight_tensor_parameter_101: class Program_weight_tensor_parameter_102: name = "parameter_102" + original_name = "linear_35.b_0" shape = [768] dtype = "float32" min_val = float("-0.621434") @@ -1133,6 +1236,7 @@ class Program_weight_tensor_parameter_102: class Program_weight_tensor_parameter_103: name = "parameter_103" + original_name = "linear_35.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-1.41945") @@ -1144,6 +1248,7 @@ class Program_weight_tensor_parameter_103: class Program_weight_tensor_parameter_104: name = "parameter_104" + original_name = "linear_34.b_0" shape = [3072] dtype = "float32" min_val = float("-0.359667") @@ -1155,6 +1260,7 @@ class Program_weight_tensor_parameter_104: class Program_weight_tensor_parameter_105: name = "parameter_105" + original_name = "linear_34.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.442576") @@ -1166,6 +1272,7 @@ class Program_weight_tensor_parameter_105: class Program_weight_tensor_parameter_106: name = "parameter_106" + original_name = "linear_33.b_0" shape = [768] dtype = "float32" min_val = float("-0.162778") @@ -1177,6 +1284,7 @@ class Program_weight_tensor_parameter_106: class Program_weight_tensor_parameter_107: name = "parameter_107" + original_name = "linear_33.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.197525") @@ -1188,6 +1296,7 @@ class Program_weight_tensor_parameter_107: class Program_weight_tensor_parameter_108: name = "parameter_108" + original_name = "linear_32.b_0" shape = [768] dtype = "float32" min_val = float("-0.0915653") @@ -1199,6 +1308,7 @@ class Program_weight_tensor_parameter_108: class Program_weight_tensor_parameter_109: name = "parameter_109" + original_name = "linear_32.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.167026") @@ -1210,6 +1320,7 @@ class Program_weight_tensor_parameter_109: class Program_weight_tensor_parameter_110: name = "parameter_110" + original_name = "linear_31.b_0" shape = [768] dtype = "float32" min_val = float("-0.0118429") @@ -1221,6 +1332,7 @@ class Program_weight_tensor_parameter_110: class Program_weight_tensor_parameter_111: name = "parameter_111" + original_name = "linear_31.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.31232") @@ -1232,6 +1344,7 @@ class Program_weight_tensor_parameter_111: class Program_weight_tensor_parameter_112: name = "parameter_112" + original_name = "linear_30.b_0" shape = [768] dtype = "float32" min_val = float("-0.34628") @@ -1243,6 +1356,7 @@ class Program_weight_tensor_parameter_112: class Program_weight_tensor_parameter_113: name = "parameter_113" + original_name = "linear_30.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.275421") @@ -1254,6 +1368,7 @@ class Program_weight_tensor_parameter_113: class Program_weight_tensor_parameter_114: name = "parameter_114" + original_name = "layer_norm_10.b_0" shape = [768] dtype = "float32" min_val = float("-0.534413") @@ -1265,6 +1380,7 @@ class Program_weight_tensor_parameter_114: class Program_weight_tensor_parameter_115: name = "parameter_115" + original_name = "layer_norm_10.w_0" shape = [768] dtype = "float32" min_val = float("0.332643") @@ -1276,6 +1392,7 @@ class Program_weight_tensor_parameter_115: class Program_weight_tensor_parameter_116: name = "parameter_116" + original_name = "layer_norm_9.b_0" shape = [768] dtype = "float32" min_val = float("-1.24542") @@ -1287,6 +1404,7 @@ class Program_weight_tensor_parameter_116: class Program_weight_tensor_parameter_117: name = "parameter_117" + original_name = "layer_norm_9.w_0" shape = [768] dtype = "float32" min_val = float("0.726918") @@ -1298,6 +1416,7 @@ class Program_weight_tensor_parameter_117: class Program_weight_tensor_parameter_118: name = "parameter_118" + original_name = "linear_29.b_0" shape = [768] dtype = "float32" min_val = float("-0.430119") @@ -1309,6 +1428,7 @@ class Program_weight_tensor_parameter_118: class Program_weight_tensor_parameter_119: name = "parameter_119" + original_name = "linear_29.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-1.32494") @@ -1320,6 +1440,7 @@ class Program_weight_tensor_parameter_119: class Program_weight_tensor_parameter_120: name = "parameter_120" + original_name = "linear_28.b_0" shape = [3072] dtype = "float32" min_val = float("-0.372176") @@ -1331,6 +1452,7 @@ class Program_weight_tensor_parameter_120: class Program_weight_tensor_parameter_121: name = "parameter_121" + original_name = "linear_28.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.419171") @@ -1342,6 +1464,7 @@ class Program_weight_tensor_parameter_121: class Program_weight_tensor_parameter_122: name = "parameter_122" + original_name = "linear_27.b_0" shape = [768] dtype = "float32" min_val = float("-0.28534") @@ -1353,6 +1476,7 @@ class Program_weight_tensor_parameter_122: class Program_weight_tensor_parameter_123: name = "parameter_123" + original_name = "linear_27.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.370497") @@ -1364,6 +1488,7 @@ class Program_weight_tensor_parameter_123: class Program_weight_tensor_parameter_124: name = "parameter_124" + original_name = "linear_26.b_0" shape = [768] dtype = "float32" min_val = float("-0.0847082") @@ -1375,6 +1500,7 @@ class Program_weight_tensor_parameter_124: class Program_weight_tensor_parameter_125: name = "parameter_125" + original_name = "linear_26.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.16733") @@ -1386,6 +1512,7 @@ class Program_weight_tensor_parameter_125: class Program_weight_tensor_parameter_126: name = "parameter_126" + original_name = "linear_25.b_0" shape = [768] dtype = "float32" min_val = float("-0.0158805") @@ -1397,6 +1524,7 @@ class Program_weight_tensor_parameter_126: class Program_weight_tensor_parameter_127: name = "parameter_127" + original_name = "linear_25.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.313149") @@ -1408,6 +1536,7 @@ class Program_weight_tensor_parameter_127: class Program_weight_tensor_parameter_128: name = "parameter_128" + original_name = "linear_24.b_0" shape = [768] dtype = "float32" min_val = float("-0.449185") @@ -1419,6 +1548,7 @@ class Program_weight_tensor_parameter_128: class Program_weight_tensor_parameter_129: name = "parameter_129" + original_name = "linear_24.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.253666") @@ -1430,6 +1560,7 @@ class Program_weight_tensor_parameter_129: class Program_weight_tensor_parameter_130: name = "parameter_130" + original_name = "layer_norm_8.b_0" shape = [768] dtype = "float32" min_val = float("-0.459685") @@ -1441,6 +1572,7 @@ class Program_weight_tensor_parameter_130: class Program_weight_tensor_parameter_131: name = "parameter_131" + original_name = "layer_norm_8.w_0" shape = [768] dtype = "float32" min_val = float("0.276944") @@ -1452,6 +1584,7 @@ class Program_weight_tensor_parameter_131: class Program_weight_tensor_parameter_132: name = "parameter_132" + original_name = "layer_norm_7.b_0" shape = [768] dtype = "float32" min_val = float("-1.28979") @@ -1463,6 +1596,7 @@ class Program_weight_tensor_parameter_132: class Program_weight_tensor_parameter_133: name = "parameter_133" + original_name = "layer_norm_7.w_0" shape = [768] dtype = "float32" min_val = float("0.764197") @@ -1474,6 +1608,7 @@ class Program_weight_tensor_parameter_133: class Program_weight_tensor_parameter_134: name = "parameter_134" + original_name = "linear_23.b_0" shape = [768] dtype = "float32" min_val = float("-0.501497") @@ -1485,6 +1620,7 @@ class Program_weight_tensor_parameter_134: class Program_weight_tensor_parameter_135: name = "parameter_135" + original_name = "linear_23.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-1.6714") @@ -1496,6 +1632,7 @@ class Program_weight_tensor_parameter_135: class Program_weight_tensor_parameter_136: name = "parameter_136" + original_name = "linear_22.b_0" shape = [3072] dtype = "float32" min_val = float("-0.329911") @@ -1507,6 +1644,7 @@ class Program_weight_tensor_parameter_136: class Program_weight_tensor_parameter_137: name = "parameter_137" + original_name = "linear_22.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.415751") @@ -1518,6 +1656,7 @@ class Program_weight_tensor_parameter_137: class Program_weight_tensor_parameter_138: name = "parameter_138" + original_name = "linear_21.b_0" shape = [768] dtype = "float32" min_val = float("-0.205201") @@ -1529,6 +1668,7 @@ class Program_weight_tensor_parameter_138: class Program_weight_tensor_parameter_139: name = "parameter_139" + original_name = "linear_21.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.170956") @@ -1540,6 +1680,7 @@ class Program_weight_tensor_parameter_139: class Program_weight_tensor_parameter_140: name = "parameter_140" + original_name = "linear_20.b_0" shape = [768] dtype = "float32" min_val = float("-0.161129") @@ -1551,6 +1692,7 @@ class Program_weight_tensor_parameter_140: class Program_weight_tensor_parameter_141: name = "parameter_141" + original_name = "linear_20.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.16405") @@ -1562,6 +1704,7 @@ class Program_weight_tensor_parameter_141: class Program_weight_tensor_parameter_142: name = "parameter_142" + original_name = "linear_19.b_0" shape = [768] dtype = "float32" min_val = float("-0.00904723") @@ -1573,6 +1716,7 @@ class Program_weight_tensor_parameter_142: class Program_weight_tensor_parameter_143: name = "parameter_143" + original_name = "linear_19.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.506249") @@ -1584,6 +1728,7 @@ class Program_weight_tensor_parameter_143: class Program_weight_tensor_parameter_144: name = "parameter_144" + original_name = "linear_18.b_0" shape = [768] dtype = "float32" min_val = float("-0.335773") @@ -1595,6 +1740,7 @@ class Program_weight_tensor_parameter_144: class Program_weight_tensor_parameter_145: name = "parameter_145" + original_name = "linear_18.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.433325") @@ -1606,6 +1752,7 @@ class Program_weight_tensor_parameter_145: class Program_weight_tensor_parameter_146: name = "parameter_146" + original_name = "layer_norm_6.b_0" shape = [768] dtype = "float32" min_val = float("-0.743646") @@ -1617,6 +1764,7 @@ class Program_weight_tensor_parameter_146: class Program_weight_tensor_parameter_147: name = "parameter_147" + original_name = "layer_norm_6.w_0" shape = [768] dtype = "float32" min_val = float("0.218041") @@ -1628,6 +1776,7 @@ class Program_weight_tensor_parameter_147: class Program_weight_tensor_parameter_148: name = "parameter_148" + original_name = "layer_norm_5.b_0" shape = [768] dtype = "float32" min_val = float("-1.64501") @@ -1639,6 +1788,7 @@ class Program_weight_tensor_parameter_148: class Program_weight_tensor_parameter_149: name = "parameter_149" + original_name = "layer_norm_5.w_0" shape = [768] dtype = "float32" min_val = float("0.701674") @@ -1650,6 +1800,7 @@ class Program_weight_tensor_parameter_149: class Program_weight_tensor_parameter_150: name = "parameter_150" + original_name = "linear_17.b_0" shape = [768] dtype = "float32" min_val = float("-0.491351") @@ -1661,6 +1812,7 @@ class Program_weight_tensor_parameter_150: class Program_weight_tensor_parameter_151: name = "parameter_151" + original_name = "linear_17.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-1.63238") @@ -1672,6 +1824,7 @@ class Program_weight_tensor_parameter_151: class Program_weight_tensor_parameter_152: name = "parameter_152" + original_name = "linear_16.b_0" shape = [3072] dtype = "float32" min_val = float("-0.250001") @@ -1683,6 +1836,7 @@ class Program_weight_tensor_parameter_152: class Program_weight_tensor_parameter_153: name = "parameter_153" + original_name = "linear_16.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.300051") @@ -1694,6 +1848,7 @@ class Program_weight_tensor_parameter_153: class Program_weight_tensor_parameter_154: name = "parameter_154" + original_name = "linear_15.b_0" shape = [768] dtype = "float32" min_val = float("-0.328843") @@ -1705,6 +1860,7 @@ class Program_weight_tensor_parameter_154: class Program_weight_tensor_parameter_155: name = "parameter_155" + original_name = "linear_15.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.195385") @@ -1716,6 +1872,7 @@ class Program_weight_tensor_parameter_155: class Program_weight_tensor_parameter_156: name = "parameter_156" + original_name = "linear_14.b_0" shape = [768] dtype = "float32" min_val = float("-0.244596") @@ -1727,6 +1884,7 @@ class Program_weight_tensor_parameter_156: class Program_weight_tensor_parameter_157: name = "parameter_157" + original_name = "linear_14.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.14529") @@ -1738,6 +1896,7 @@ class Program_weight_tensor_parameter_157: class Program_weight_tensor_parameter_158: name = "parameter_158" + original_name = "linear_13.b_0" shape = [768] dtype = "float32" min_val = float("-0.00912743") @@ -1749,6 +1908,7 @@ class Program_weight_tensor_parameter_158: class Program_weight_tensor_parameter_159: name = "parameter_159" + original_name = "linear_13.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.568666") @@ -1760,6 +1920,7 @@ class Program_weight_tensor_parameter_159: class Program_weight_tensor_parameter_160: name = "parameter_160" + original_name = "linear_12.b_0" shape = [768] dtype = "float32" min_val = float("-0.411989") @@ -1771,6 +1932,7 @@ class Program_weight_tensor_parameter_160: class Program_weight_tensor_parameter_161: name = "parameter_161" + original_name = "linear_12.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.465481") @@ -1782,6 +1944,7 @@ class Program_weight_tensor_parameter_161: class Program_weight_tensor_parameter_162: name = "parameter_162" + original_name = "layer_norm_4.b_0" shape = [768] dtype = "float32" min_val = float("-0.920298") @@ -1793,6 +1956,7 @@ class Program_weight_tensor_parameter_162: class Program_weight_tensor_parameter_163: name = "parameter_163" + original_name = "layer_norm_4.w_0" shape = [768] dtype = "float32" min_val = float("0.120482") @@ -1804,6 +1968,7 @@ class Program_weight_tensor_parameter_163: class Program_weight_tensor_parameter_164: name = "parameter_164" + original_name = "layer_norm_3.b_0" shape = [768] dtype = "float32" min_val = float("-2.14232") @@ -1815,6 +1980,7 @@ class Program_weight_tensor_parameter_164: class Program_weight_tensor_parameter_165: name = "parameter_165" + original_name = "layer_norm_3.w_0" shape = [768] dtype = "float32" min_val = float("0.741052") @@ -1826,6 +1992,7 @@ class Program_weight_tensor_parameter_165: class Program_weight_tensor_parameter_166: name = "parameter_166" + original_name = "linear_11.b_0" shape = [768] dtype = "float32" min_val = float("-0.415744") @@ -1837,6 +2004,7 @@ class Program_weight_tensor_parameter_166: class Program_weight_tensor_parameter_167: name = "parameter_167" + original_name = "linear_11.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-2.24656") @@ -1848,6 +2016,7 @@ class Program_weight_tensor_parameter_167: class Program_weight_tensor_parameter_168: name = "parameter_168" + original_name = "linear_10.b_0" shape = [3072] dtype = "float32" min_val = float("-0.229282") @@ -1859,6 +2028,7 @@ class Program_weight_tensor_parameter_168: class Program_weight_tensor_parameter_169: name = "parameter_169" + original_name = "linear_10.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.340072") @@ -1870,6 +2040,7 @@ class Program_weight_tensor_parameter_169: class Program_weight_tensor_parameter_170: name = "parameter_170" + original_name = "linear_9.b_0" shape = [768] dtype = "float32" min_val = float("-0.316372") @@ -1881,6 +2052,7 @@ class Program_weight_tensor_parameter_170: class Program_weight_tensor_parameter_171: name = "parameter_171" + original_name = "linear_9.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.365415") @@ -1892,6 +2064,7 @@ class Program_weight_tensor_parameter_171: class Program_weight_tensor_parameter_172: name = "parameter_172" + original_name = "linear_8.b_0" shape = [768] dtype = "float32" min_val = float("-0.192833") @@ -1903,6 +2076,7 @@ class Program_weight_tensor_parameter_172: class Program_weight_tensor_parameter_173: name = "parameter_173" + original_name = "linear_8.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.150234") @@ -1914,6 +2088,7 @@ class Program_weight_tensor_parameter_173: class Program_weight_tensor_parameter_174: name = "parameter_174" + original_name = "linear_7.b_0" shape = [768] dtype = "float32" min_val = float("-0.0127015") @@ -1925,6 +2100,7 @@ class Program_weight_tensor_parameter_174: class Program_weight_tensor_parameter_175: name = "parameter_175" + original_name = "linear_7.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.332539") @@ -1936,6 +2112,7 @@ class Program_weight_tensor_parameter_175: class Program_weight_tensor_parameter_176: name = "parameter_176" + original_name = "linear_6.b_0" shape = [768] dtype = "float32" min_val = float("-0.502453") @@ -1947,6 +2124,7 @@ class Program_weight_tensor_parameter_176: class Program_weight_tensor_parameter_177: name = "parameter_177" + original_name = "linear_6.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.297714") @@ -1958,6 +2136,7 @@ class Program_weight_tensor_parameter_177: class Program_weight_tensor_parameter_178: name = "parameter_178" + original_name = "layer_norm_2.b_0" shape = [768] dtype = "float32" min_val = float("-1.54433") @@ -1969,6 +2148,7 @@ class Program_weight_tensor_parameter_178: class Program_weight_tensor_parameter_179: name = "parameter_179" + original_name = "layer_norm_2.w_0" shape = [768] dtype = "float32" min_val = float("0.194415") @@ -1980,6 +2160,7 @@ class Program_weight_tensor_parameter_179: class Program_weight_tensor_parameter_180: name = "parameter_180" + original_name = "layer_norm_1.b_0" shape = [768] dtype = "float32" min_val = float("-3.63679") @@ -1991,6 +2172,7 @@ class Program_weight_tensor_parameter_180: class Program_weight_tensor_parameter_181: name = "parameter_181" + original_name = "layer_norm_1.w_0" shape = [768] dtype = "float32" min_val = float("0.563539") @@ -2002,6 +2184,7 @@ class Program_weight_tensor_parameter_181: class Program_weight_tensor_parameter_182: name = "parameter_182" + original_name = "linear_5.b_0" shape = [768] dtype = "float32" min_val = float("-0.314555") @@ -2013,6 +2196,7 @@ class Program_weight_tensor_parameter_182: class Program_weight_tensor_parameter_183: name = "parameter_183" + original_name = "linear_5.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-1.51271") @@ -2024,6 +2208,7 @@ class Program_weight_tensor_parameter_183: class Program_weight_tensor_parameter_184: name = "parameter_184" + original_name = "linear_4.b_0" shape = [3072] dtype = "float32" min_val = float("-0.349466") @@ -2035,6 +2220,7 @@ class Program_weight_tensor_parameter_184: class Program_weight_tensor_parameter_185: name = "parameter_185" + original_name = "linear_4.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.298772") @@ -2046,6 +2232,7 @@ class Program_weight_tensor_parameter_185: class Program_weight_tensor_parameter_186: name = "parameter_186" + original_name = "linear_3.b_0" shape = [768] dtype = "float32" min_val = float("-0.523742") @@ -2057,6 +2244,7 @@ class Program_weight_tensor_parameter_186: class Program_weight_tensor_parameter_187: name = "parameter_187" + original_name = "linear_3.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.522258") @@ -2068,6 +2256,7 @@ class Program_weight_tensor_parameter_187: class Program_weight_tensor_parameter_188: name = "parameter_188" + original_name = "linear_2.b_0" shape = [768] dtype = "float32" min_val = float("-0.2252") @@ -2079,6 +2268,7 @@ class Program_weight_tensor_parameter_188: class Program_weight_tensor_parameter_189: name = "parameter_189" + original_name = "linear_2.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.135266") @@ -2090,6 +2280,7 @@ class Program_weight_tensor_parameter_189: class Program_weight_tensor_parameter_190: name = "parameter_190" + original_name = "linear_1.b_0" shape = [768] dtype = "float32" min_val = float("-0.00716202") @@ -2101,6 +2292,7 @@ class Program_weight_tensor_parameter_190: class Program_weight_tensor_parameter_191: name = "parameter_191" + original_name = "linear_1.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.569508") @@ -2112,6 +2304,7 @@ class Program_weight_tensor_parameter_191: class Program_weight_tensor_parameter_192: name = "parameter_192" + original_name = "linear_0.b_0" shape = [768] dtype = "float32" min_val = float("-0.721302") @@ -2123,6 +2316,7 @@ class Program_weight_tensor_parameter_192: class Program_weight_tensor_parameter_193: name = "parameter_193" + original_name = "linear_0.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.290126") @@ -2134,6 +2328,7 @@ class Program_weight_tensor_parameter_193: class Program_weight_tensor_parameter_194: name = "parameter_194" + original_name = "layer_norm_0.b_0" shape = [768] dtype = "float32" min_val = float("-0.165663") @@ -2145,6 +2340,7 @@ class Program_weight_tensor_parameter_194: class Program_weight_tensor_parameter_195: name = "parameter_195" + original_name = "layer_norm_0.w_0" shape = [768] dtype = "float32" min_val = float("0.167558") @@ -2156,6 +2352,7 @@ class Program_weight_tensor_parameter_195: class Program_weight_tensor_parameter_196: name = "parameter_196" + original_name = "embedding_2.w_0" shape = [2, 768] dtype = "float32" min_val = float("-0.682424") @@ -2167,6 +2364,7 @@ class Program_weight_tensor_parameter_196: class Program_weight_tensor_parameter_197: name = "parameter_197" + original_name = "embedding_1.w_0" shape = [512, 768] dtype = "float32" min_val = float("-0.943522") @@ -2178,6 +2376,7 @@ class Program_weight_tensor_parameter_197: class Program_weight_tensor_parameter_198: name = "parameter_198" + original_name = "embedding_0.w_0" shape = [28996, 768] dtype = "float32" min_val = float("-0.972329") diff --git a/paddle_samples/PaddleNLP/bert-large-cased/input_meta.py b/paddle_samples/PaddleNLP/bert-large-cased/input_meta.py index 46bea809b..09fce9f05 100644 --- a/paddle_samples/PaddleNLP/bert-large-cased/input_meta.py +++ b/paddle_samples/PaddleNLP/bert-large-cased/input_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_data_0: name = "data_0" + original_name = "input_ids" shape = [1, 21] dtype = "int64" data = [ @@ -29,6 +30,7 @@ class Program_weight_tensor_data_0: class Program_weight_tensor_data_1: name = "data_1" + original_name = "token_type_ids" shape = [1, 21] dtype = "int64" data = [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0] diff --git a/paddle_samples/PaddleNLP/bert-large-cased/model.py b/paddle_samples/PaddleNLP/bert-large-cased/model.py index 14c81b6d4..7f815b1ad 100644 --- a/paddle_samples/PaddleNLP/bert-large-cased/model.py +++ b/paddle_samples/PaddleNLP/bert-large-cased/model.py @@ -460,7 +460,7 @@ def forward( # pd_op.embedding: (1x21x1024xf32) <- (1x21xi64, 512x1024xf32) embedding_1 = paddle._C_ops.embedding(subtract_0, parameter_389, -1, False) - del parameter_389 + del parameter_389, subtract_0 # pd_op.embedding: (1x21x1024xf32) <- (1x21xi64, 2x1024xf32) embedding_2 = paddle._C_ops.embedding(data_1, parameter_388, -1, False) @@ -468,9 +468,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_0 = paddle._C_ops.add(embedding_0, embedding_1) + del embedding_0, embedding_1 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_1 = paddle._C_ops.add(add_0, embedding_2) + del add_0, embedding_2 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_0, layer_norm_1, layer_norm_2 = (lambda x, f: f(x))( @@ -479,233 +481,17 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_386, parameter_387 + del add_1, parameter_386, parameter_387 # pd_op.full: (1xf32) <- () full_4 = paddle._C_ops.full( [1], float("0.1"), paddle.float32, paddle.core.CPUPlace() ) - # pd_op.assign: (1xf32) <- (1xf32) - assign_0 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_1 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_2 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_3 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_4 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_5 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_6 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_7 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_8 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_9 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_10 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_11 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_12 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_13 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_14 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_15 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_16 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_17 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_18 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_19 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_20 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_21 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_22 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_23 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_24 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_25 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_26 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_27 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_28 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_29 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_30 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_31 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_32 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_33 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_34 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_35 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_36 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_37 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_38 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_39 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_40 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_41 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_42 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_43 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_44 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_45 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_46 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_47 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_48 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_49 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_50 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_51 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_52 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_53 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_54 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_55 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_56 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_57 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_58 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_59 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_60 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_61 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_62 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_63 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_64 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_65 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_66 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_67 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_68 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_69 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_70 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_71 = full_4 - # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_0, dropout_1 = (lambda x, f: f(x))( paddle._C_ops.dropout( - layer_norm_0, None, full_4, False, "upscale_in_train", 0, False + layer_norm_0, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -717,13 +503,14 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_2 = paddle._C_ops.add(matmul_0, parameter_384) - del parameter_384 + del matmul_0, parameter_384 # pd_op.full_int_array: (4xi64) <- () full_int_array_1 = [0, 0, 16, 64] # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_0 = paddle._C_ops.reshape(add_2, full_int_array_1) + del add_2 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_0 = paddle._C_ops.transpose(reshape_0, [0, 2, 1, 3]) @@ -735,7 +522,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_3 = paddle._C_ops.add(matmul_1, parameter_382) - del parameter_382 + del matmul_1, parameter_382 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_2 = paddle._C_ops.matmul(dropout_0, parameter_381, False, False) @@ -743,10 +530,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_4 = paddle._C_ops.add(matmul_2, parameter_380) - del parameter_380 + del matmul_2, parameter_380 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_1 = paddle._C_ops.reshape(add_3, full_int_array_1) + del add_3 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_1 = paddle._C_ops.transpose(reshape_1, [0, 2, 1, 3]) @@ -754,6 +542,7 @@ def forward( # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_2 = paddle._C_ops.reshape(add_4, full_int_array_1) + del add_4 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_2 = paddle._C_ops.transpose(reshape_2, [0, 2, 1, 3]) @@ -764,84 +553,17 @@ def forward( [1], float("0.125"), paddle.float32, paddle.core.CPUPlace() ) - # pd_op.assign: (1xf32) <- (1xf32) - assign_72 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_73 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_74 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_75 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_76 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_77 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_78 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_79 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_80 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_81 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_82 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_83 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_84 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_85 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_86 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_87 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_88 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_89 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_90 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_91 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_92 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_93 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_94 = full_5 - # pd_op.scale: (1x16x21x64xf32) <- (1x16x21x64xf32, 1xf32) scale_1 = paddle._C_ops.scale(transpose_0, full_5, float("0"), True) del transpose_0 # pd_op.matmul: (1x16x21x21xf32) <- (1x16x21x64xf32, 1x16x21x64xf32) matmul_3 = paddle._C_ops.matmul(scale_1, transpose_1, False, True) + del scale_1, transpose_1 # pd_op.add: (1x16x21x21xf32) <- (1x16x21x21xf32, 1x1x1x21xf32) add_5 = paddle._C_ops.add(matmul_3, unsqueeze_0) + del matmul_3 # pd_op.softmax: (1x16x21x21xf32) <- (1x16x21x21xf32) softmax_0 = paddle._C_ops.softmax(add_5, -1) @@ -850,13 +572,15 @@ def forward( # pd_op.dropout: (1x16x21x21xf32, 1x16x21x21xui8) <- (1x16x21x21xf32, None, 1xf32) dropout_2, dropout_3 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_0, None, full_4, False, "upscale_in_train", 0, False + softmax_0, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_0 # pd_op.matmul: (1x16x21x64xf32) <- (1x16x21x21xf32, 1x16x21x64xf32) matmul_4 = paddle._C_ops.matmul(dropout_2, transpose_2, False, False) + del dropout_2, transpose_2 # pd_op.transpose: (1x21x16x64xf32) <- (1x16x21x64xf32) transpose_3 = paddle._C_ops.transpose(matmul_4, [0, 2, 1, 3]) @@ -867,19 +591,20 @@ def forward( # pd_op.reshape: (1x21x1024xf32) <- (1x21x16x64xf32, 3xi64) reshape_3 = paddle._C_ops.reshape(transpose_3, full_int_array_2) + del transpose_3 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_5 = paddle._C_ops.matmul(reshape_3, parameter_379, False, False) - del parameter_379 + del parameter_379, reshape_3 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_6 = paddle._C_ops.add(matmul_5, parameter_378) - del parameter_378 + del matmul_5, parameter_378 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_4, dropout_5 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_6, None, full_4, False, "upscale_in_train", 0, False + add_6, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -887,6 +612,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_7 = paddle._C_ops.add(dropout_0, dropout_4) + del dropout_0, dropout_4 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_3, layer_norm_4, layer_norm_5 = (lambda x, f: f(x))( @@ -895,7 +621,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_372, parameter_373 + del add_7, parameter_372, parameter_373 # pd_op.matmul: (1x21x4096xf32) <- (1x21x1024xf32, 1024x4096xf32) matmul_6 = paddle._C_ops.matmul(layer_norm_3, parameter_377, False, False) @@ -903,23 +629,24 @@ def forward( # pd_op.add: (1x21x4096xf32) <- (1x21x4096xf32, 4096xf32) add_8 = paddle._C_ops.add(matmul_6, parameter_376) - del parameter_376 + del matmul_6, parameter_376 # pd_op.gelu: (1x21x4096xf32) <- (1x21x4096xf32) gelu_0 = paddle._C_ops.gelu(add_8, False) + del add_8 # pd_op.matmul: (1x21x1024xf32) <- (1x21x4096xf32, 4096x1024xf32) matmul_7 = paddle._C_ops.matmul(gelu_0, parameter_375, False, False) - del parameter_375 + del gelu_0, parameter_375 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_9 = paddle._C_ops.add(matmul_7, parameter_374) - del parameter_374 + del matmul_7, parameter_374 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_6, dropout_7 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_9, None, full_4, False, "upscale_in_train", 0, False + add_9, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -927,6 +654,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_10 = paddle._C_ops.add(layer_norm_3, dropout_6) + del dropout_6, layer_norm_3 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_6, layer_norm_7, layer_norm_8 = (lambda x, f: f(x))( @@ -935,7 +663,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_370, parameter_371 + del add_10, parameter_370, parameter_371 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_8 = paddle._C_ops.matmul(layer_norm_6, parameter_369, False, False) @@ -943,10 +671,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_11 = paddle._C_ops.add(matmul_8, parameter_368) - del parameter_368 + del matmul_8, parameter_368 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_4 = paddle._C_ops.reshape(add_11, full_int_array_1) + del add_11 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_4 = paddle._C_ops.transpose(reshape_4, [0, 2, 1, 3]) @@ -958,7 +687,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_12 = paddle._C_ops.add(matmul_9, parameter_366) - del parameter_366 + del matmul_9, parameter_366 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_10 = paddle._C_ops.matmul(layer_norm_6, parameter_365, False, False) @@ -966,10 +695,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_13 = paddle._C_ops.add(matmul_10, parameter_364) - del parameter_364 + del matmul_10, parameter_364 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_5 = paddle._C_ops.reshape(add_12, full_int_array_1) + del add_12 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_5 = paddle._C_ops.transpose(reshape_5, [0, 2, 1, 3]) @@ -977,6 +707,7 @@ def forward( # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_6 = paddle._C_ops.reshape(add_13, full_int_array_1) + del add_13 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_6 = paddle._C_ops.transpose(reshape_6, [0, 2, 1, 3]) @@ -988,9 +719,11 @@ def forward( # pd_op.matmul: (1x16x21x21xf32) <- (1x16x21x64xf32, 1x16x21x64xf32) matmul_11 = paddle._C_ops.matmul(scale_2, transpose_5, False, True) + del scale_2, transpose_5 # pd_op.add: (1x16x21x21xf32) <- (1x16x21x21xf32, 1x1x1x21xf32) add_14 = paddle._C_ops.add(matmul_11, unsqueeze_0) + del matmul_11 # pd_op.softmax: (1x16x21x21xf32) <- (1x16x21x21xf32) softmax_1 = paddle._C_ops.softmax(add_14, -1) @@ -999,13 +732,15 @@ def forward( # pd_op.dropout: (1x16x21x21xf32, 1x16x21x21xui8) <- (1x16x21x21xf32, None, 1xf32) dropout_8, dropout_9 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_1, None, full_4, False, "upscale_in_train", 0, False + softmax_1, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_1 # pd_op.matmul: (1x16x21x64xf32) <- (1x16x21x21xf32, 1x16x21x64xf32) matmul_12 = paddle._C_ops.matmul(dropout_8, transpose_6, False, False) + del dropout_8, transpose_6 # pd_op.transpose: (1x21x16x64xf32) <- (1x16x21x64xf32) transpose_7 = paddle._C_ops.transpose(matmul_12, [0, 2, 1, 3]) @@ -1013,19 +748,20 @@ def forward( # pd_op.reshape: (1x21x1024xf32) <- (1x21x16x64xf32, 3xi64) reshape_7 = paddle._C_ops.reshape(transpose_7, full_int_array_2) + del transpose_7 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_13 = paddle._C_ops.matmul(reshape_7, parameter_363, False, False) - del parameter_363 + del parameter_363, reshape_7 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_15 = paddle._C_ops.add(matmul_13, parameter_362) - del parameter_362 + del matmul_13, parameter_362 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_10, dropout_11 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_15, None, full_4, False, "upscale_in_train", 0, False + add_15, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1033,6 +769,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_16 = paddle._C_ops.add(layer_norm_6, dropout_10) + del dropout_10, layer_norm_6 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_9, layer_norm_10, layer_norm_11 = (lambda x, f: f(x))( @@ -1041,7 +778,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_356, parameter_357 + del add_16, parameter_356, parameter_357 # pd_op.matmul: (1x21x4096xf32) <- (1x21x1024xf32, 1024x4096xf32) matmul_14 = paddle._C_ops.matmul(layer_norm_9, parameter_361, False, False) @@ -1049,23 +786,24 @@ def forward( # pd_op.add: (1x21x4096xf32) <- (1x21x4096xf32, 4096xf32) add_17 = paddle._C_ops.add(matmul_14, parameter_360) - del parameter_360 + del matmul_14, parameter_360 # pd_op.gelu: (1x21x4096xf32) <- (1x21x4096xf32) gelu_1 = paddle._C_ops.gelu(add_17, False) + del add_17 # pd_op.matmul: (1x21x1024xf32) <- (1x21x4096xf32, 4096x1024xf32) matmul_15 = paddle._C_ops.matmul(gelu_1, parameter_359, False, False) - del parameter_359 + del gelu_1, parameter_359 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_18 = paddle._C_ops.add(matmul_15, parameter_358) - del parameter_358 + del matmul_15, parameter_358 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_12, dropout_13 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_18, None, full_4, False, "upscale_in_train", 0, False + add_18, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1073,6 +811,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_19 = paddle._C_ops.add(layer_norm_9, dropout_12) + del dropout_12, layer_norm_9 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_12, layer_norm_13, layer_norm_14 = (lambda x, f: f(x))( @@ -1081,7 +820,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_354, parameter_355 + del add_19, parameter_354, parameter_355 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_16 = paddle._C_ops.matmul(layer_norm_12, parameter_353, False, False) @@ -1089,10 +828,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_20 = paddle._C_ops.add(matmul_16, parameter_352) - del parameter_352 + del matmul_16, parameter_352 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_8 = paddle._C_ops.reshape(add_20, full_int_array_1) + del add_20 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_8 = paddle._C_ops.transpose(reshape_8, [0, 2, 1, 3]) @@ -1104,7 +844,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_21 = paddle._C_ops.add(matmul_17, parameter_350) - del parameter_350 + del matmul_17, parameter_350 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_18 = paddle._C_ops.matmul(layer_norm_12, parameter_349, False, False) @@ -1112,10 +852,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_22 = paddle._C_ops.add(matmul_18, parameter_348) - del parameter_348 + del matmul_18, parameter_348 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_9 = paddle._C_ops.reshape(add_21, full_int_array_1) + del add_21 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_9 = paddle._C_ops.transpose(reshape_9, [0, 2, 1, 3]) @@ -1123,6 +864,7 @@ def forward( # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_10 = paddle._C_ops.reshape(add_22, full_int_array_1) + del add_22 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_10 = paddle._C_ops.transpose(reshape_10, [0, 2, 1, 3]) @@ -1134,9 +876,11 @@ def forward( # pd_op.matmul: (1x16x21x21xf32) <- (1x16x21x64xf32, 1x16x21x64xf32) matmul_19 = paddle._C_ops.matmul(scale_3, transpose_9, False, True) + del scale_3, transpose_9 # pd_op.add: (1x16x21x21xf32) <- (1x16x21x21xf32, 1x1x1x21xf32) add_23 = paddle._C_ops.add(matmul_19, unsqueeze_0) + del matmul_19 # pd_op.softmax: (1x16x21x21xf32) <- (1x16x21x21xf32) softmax_2 = paddle._C_ops.softmax(add_23, -1) @@ -1145,13 +889,15 @@ def forward( # pd_op.dropout: (1x16x21x21xf32, 1x16x21x21xui8) <- (1x16x21x21xf32, None, 1xf32) dropout_14, dropout_15 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_2, None, full_4, False, "upscale_in_train", 0, False + softmax_2, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_2 # pd_op.matmul: (1x16x21x64xf32) <- (1x16x21x21xf32, 1x16x21x64xf32) matmul_20 = paddle._C_ops.matmul(dropout_14, transpose_10, False, False) + del dropout_14, transpose_10 # pd_op.transpose: (1x21x16x64xf32) <- (1x16x21x64xf32) transpose_11 = paddle._C_ops.transpose(matmul_20, [0, 2, 1, 3]) @@ -1159,19 +905,20 @@ def forward( # pd_op.reshape: (1x21x1024xf32) <- (1x21x16x64xf32, 3xi64) reshape_11 = paddle._C_ops.reshape(transpose_11, full_int_array_2) + del transpose_11 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_21 = paddle._C_ops.matmul(reshape_11, parameter_347, False, False) - del parameter_347 + del parameter_347, reshape_11 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_24 = paddle._C_ops.add(matmul_21, parameter_346) - del parameter_346 + del matmul_21, parameter_346 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_16, dropout_17 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_24, None, full_4, False, "upscale_in_train", 0, False + add_24, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1179,6 +926,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_25 = paddle._C_ops.add(layer_norm_12, dropout_16) + del dropout_16, layer_norm_12 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_15, layer_norm_16, layer_norm_17 = (lambda x, f: f(x))( @@ -1187,7 +935,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_340, parameter_341 + del add_25, parameter_340, parameter_341 # pd_op.matmul: (1x21x4096xf32) <- (1x21x1024xf32, 1024x4096xf32) matmul_22 = paddle._C_ops.matmul(layer_norm_15, parameter_345, False, False) @@ -1195,23 +943,24 @@ def forward( # pd_op.add: (1x21x4096xf32) <- (1x21x4096xf32, 4096xf32) add_26 = paddle._C_ops.add(matmul_22, parameter_344) - del parameter_344 + del matmul_22, parameter_344 # pd_op.gelu: (1x21x4096xf32) <- (1x21x4096xf32) gelu_2 = paddle._C_ops.gelu(add_26, False) + del add_26 # pd_op.matmul: (1x21x1024xf32) <- (1x21x4096xf32, 4096x1024xf32) matmul_23 = paddle._C_ops.matmul(gelu_2, parameter_343, False, False) - del parameter_343 + del gelu_2, parameter_343 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_27 = paddle._C_ops.add(matmul_23, parameter_342) - del parameter_342 + del matmul_23, parameter_342 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_18, dropout_19 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_27, None, full_4, False, "upscale_in_train", 0, False + add_27, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1219,6 +968,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_28 = paddle._C_ops.add(layer_norm_15, dropout_18) + del dropout_18, layer_norm_15 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_18, layer_norm_19, layer_norm_20 = (lambda x, f: f(x))( @@ -1227,7 +977,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_338, parameter_339 + del add_28, parameter_338, parameter_339 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_24 = paddle._C_ops.matmul(layer_norm_18, parameter_337, False, False) @@ -1235,10 +985,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_29 = paddle._C_ops.add(matmul_24, parameter_336) - del parameter_336 + del matmul_24, parameter_336 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_12 = paddle._C_ops.reshape(add_29, full_int_array_1) + del add_29 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_12 = paddle._C_ops.transpose(reshape_12, [0, 2, 1, 3]) @@ -1250,7 +1001,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_30 = paddle._C_ops.add(matmul_25, parameter_334) - del parameter_334 + del matmul_25, parameter_334 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_26 = paddle._C_ops.matmul(layer_norm_18, parameter_333, False, False) @@ -1258,10 +1009,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_31 = paddle._C_ops.add(matmul_26, parameter_332) - del parameter_332 + del matmul_26, parameter_332 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_13 = paddle._C_ops.reshape(add_30, full_int_array_1) + del add_30 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_13 = paddle._C_ops.transpose(reshape_13, [0, 2, 1, 3]) @@ -1269,6 +1021,7 @@ def forward( # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_14 = paddle._C_ops.reshape(add_31, full_int_array_1) + del add_31 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_14 = paddle._C_ops.transpose(reshape_14, [0, 2, 1, 3]) @@ -1280,9 +1033,11 @@ def forward( # pd_op.matmul: (1x16x21x21xf32) <- (1x16x21x64xf32, 1x16x21x64xf32) matmul_27 = paddle._C_ops.matmul(scale_4, transpose_13, False, True) + del scale_4, transpose_13 # pd_op.add: (1x16x21x21xf32) <- (1x16x21x21xf32, 1x1x1x21xf32) add_32 = paddle._C_ops.add(matmul_27, unsqueeze_0) + del matmul_27 # pd_op.softmax: (1x16x21x21xf32) <- (1x16x21x21xf32) softmax_3 = paddle._C_ops.softmax(add_32, -1) @@ -1291,13 +1046,15 @@ def forward( # pd_op.dropout: (1x16x21x21xf32, 1x16x21x21xui8) <- (1x16x21x21xf32, None, 1xf32) dropout_20, dropout_21 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_3, None, full_4, False, "upscale_in_train", 0, False + softmax_3, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_3 # pd_op.matmul: (1x16x21x64xf32) <- (1x16x21x21xf32, 1x16x21x64xf32) matmul_28 = paddle._C_ops.matmul(dropout_20, transpose_14, False, False) + del dropout_20, transpose_14 # pd_op.transpose: (1x21x16x64xf32) <- (1x16x21x64xf32) transpose_15 = paddle._C_ops.transpose(matmul_28, [0, 2, 1, 3]) @@ -1305,19 +1062,20 @@ def forward( # pd_op.reshape: (1x21x1024xf32) <- (1x21x16x64xf32, 3xi64) reshape_15 = paddle._C_ops.reshape(transpose_15, full_int_array_2) + del transpose_15 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_29 = paddle._C_ops.matmul(reshape_15, parameter_331, False, False) - del parameter_331 + del parameter_331, reshape_15 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_33 = paddle._C_ops.add(matmul_29, parameter_330) - del parameter_330 + del matmul_29, parameter_330 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_22, dropout_23 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_33, None, full_4, False, "upscale_in_train", 0, False + add_33, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1325,6 +1083,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_34 = paddle._C_ops.add(layer_norm_18, dropout_22) + del dropout_22, layer_norm_18 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_21, layer_norm_22, layer_norm_23 = (lambda x, f: f(x))( @@ -1333,7 +1092,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_324, parameter_325 + del add_34, parameter_324, parameter_325 # pd_op.matmul: (1x21x4096xf32) <- (1x21x1024xf32, 1024x4096xf32) matmul_30 = paddle._C_ops.matmul(layer_norm_21, parameter_329, False, False) @@ -1341,23 +1100,24 @@ def forward( # pd_op.add: (1x21x4096xf32) <- (1x21x4096xf32, 4096xf32) add_35 = paddle._C_ops.add(matmul_30, parameter_328) - del parameter_328 + del matmul_30, parameter_328 # pd_op.gelu: (1x21x4096xf32) <- (1x21x4096xf32) gelu_3 = paddle._C_ops.gelu(add_35, False) + del add_35 # pd_op.matmul: (1x21x1024xf32) <- (1x21x4096xf32, 4096x1024xf32) matmul_31 = paddle._C_ops.matmul(gelu_3, parameter_327, False, False) - del parameter_327 + del gelu_3, parameter_327 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_36 = paddle._C_ops.add(matmul_31, parameter_326) - del parameter_326 + del matmul_31, parameter_326 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_24, dropout_25 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_36, None, full_4, False, "upscale_in_train", 0, False + add_36, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1365,6 +1125,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_37 = paddle._C_ops.add(layer_norm_21, dropout_24) + del dropout_24, layer_norm_21 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_24, layer_norm_25, layer_norm_26 = (lambda x, f: f(x))( @@ -1373,7 +1134,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_322, parameter_323 + del add_37, parameter_322, parameter_323 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_32 = paddle._C_ops.matmul(layer_norm_24, parameter_321, False, False) @@ -1381,10 +1142,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_38 = paddle._C_ops.add(matmul_32, parameter_320) - del parameter_320 + del matmul_32, parameter_320 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_16 = paddle._C_ops.reshape(add_38, full_int_array_1) + del add_38 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_16 = paddle._C_ops.transpose(reshape_16, [0, 2, 1, 3]) @@ -1396,7 +1158,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_39 = paddle._C_ops.add(matmul_33, parameter_318) - del parameter_318 + del matmul_33, parameter_318 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_34 = paddle._C_ops.matmul(layer_norm_24, parameter_317, False, False) @@ -1404,10 +1166,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_40 = paddle._C_ops.add(matmul_34, parameter_316) - del parameter_316 + del matmul_34, parameter_316 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_17 = paddle._C_ops.reshape(add_39, full_int_array_1) + del add_39 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_17 = paddle._C_ops.transpose(reshape_17, [0, 2, 1, 3]) @@ -1415,6 +1178,7 @@ def forward( # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_18 = paddle._C_ops.reshape(add_40, full_int_array_1) + del add_40 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_18 = paddle._C_ops.transpose(reshape_18, [0, 2, 1, 3]) @@ -1426,9 +1190,11 @@ def forward( # pd_op.matmul: (1x16x21x21xf32) <- (1x16x21x64xf32, 1x16x21x64xf32) matmul_35 = paddle._C_ops.matmul(scale_5, transpose_17, False, True) + del scale_5, transpose_17 # pd_op.add: (1x16x21x21xf32) <- (1x16x21x21xf32, 1x1x1x21xf32) add_41 = paddle._C_ops.add(matmul_35, unsqueeze_0) + del matmul_35 # pd_op.softmax: (1x16x21x21xf32) <- (1x16x21x21xf32) softmax_4 = paddle._C_ops.softmax(add_41, -1) @@ -1437,13 +1203,15 @@ def forward( # pd_op.dropout: (1x16x21x21xf32, 1x16x21x21xui8) <- (1x16x21x21xf32, None, 1xf32) dropout_26, dropout_27 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_4, None, full_4, False, "upscale_in_train", 0, False + softmax_4, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_4 # pd_op.matmul: (1x16x21x64xf32) <- (1x16x21x21xf32, 1x16x21x64xf32) matmul_36 = paddle._C_ops.matmul(dropout_26, transpose_18, False, False) + del dropout_26, transpose_18 # pd_op.transpose: (1x21x16x64xf32) <- (1x16x21x64xf32) transpose_19 = paddle._C_ops.transpose(matmul_36, [0, 2, 1, 3]) @@ -1451,19 +1219,20 @@ def forward( # pd_op.reshape: (1x21x1024xf32) <- (1x21x16x64xf32, 3xi64) reshape_19 = paddle._C_ops.reshape(transpose_19, full_int_array_2) + del transpose_19 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_37 = paddle._C_ops.matmul(reshape_19, parameter_315, False, False) - del parameter_315 + del parameter_315, reshape_19 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_42 = paddle._C_ops.add(matmul_37, parameter_314) - del parameter_314 + del matmul_37, parameter_314 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_28, dropout_29 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_42, None, full_4, False, "upscale_in_train", 0, False + add_42, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1471,6 +1240,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_43 = paddle._C_ops.add(layer_norm_24, dropout_28) + del dropout_28, layer_norm_24 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_27, layer_norm_28, layer_norm_29 = (lambda x, f: f(x))( @@ -1479,7 +1249,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_308, parameter_309 + del add_43, parameter_308, parameter_309 # pd_op.matmul: (1x21x4096xf32) <- (1x21x1024xf32, 1024x4096xf32) matmul_38 = paddle._C_ops.matmul(layer_norm_27, parameter_313, False, False) @@ -1487,23 +1257,24 @@ def forward( # pd_op.add: (1x21x4096xf32) <- (1x21x4096xf32, 4096xf32) add_44 = paddle._C_ops.add(matmul_38, parameter_312) - del parameter_312 + del matmul_38, parameter_312 # pd_op.gelu: (1x21x4096xf32) <- (1x21x4096xf32) gelu_4 = paddle._C_ops.gelu(add_44, False) + del add_44 # pd_op.matmul: (1x21x1024xf32) <- (1x21x4096xf32, 4096x1024xf32) matmul_39 = paddle._C_ops.matmul(gelu_4, parameter_311, False, False) - del parameter_311 + del gelu_4, parameter_311 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_45 = paddle._C_ops.add(matmul_39, parameter_310) - del parameter_310 + del matmul_39, parameter_310 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_30, dropout_31 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_45, None, full_4, False, "upscale_in_train", 0, False + add_45, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1511,6 +1282,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_46 = paddle._C_ops.add(layer_norm_27, dropout_30) + del dropout_30, layer_norm_27 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_30, layer_norm_31, layer_norm_32 = (lambda x, f: f(x))( @@ -1519,7 +1291,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_306, parameter_307 + del add_46, parameter_306, parameter_307 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_40 = paddle._C_ops.matmul(layer_norm_30, parameter_305, False, False) @@ -1527,10 +1299,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_47 = paddle._C_ops.add(matmul_40, parameter_304) - del parameter_304 + del matmul_40, parameter_304 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_20 = paddle._C_ops.reshape(add_47, full_int_array_1) + del add_47 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_20 = paddle._C_ops.transpose(reshape_20, [0, 2, 1, 3]) @@ -1542,7 +1315,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_48 = paddle._C_ops.add(matmul_41, parameter_302) - del parameter_302 + del matmul_41, parameter_302 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_42 = paddle._C_ops.matmul(layer_norm_30, parameter_301, False, False) @@ -1550,10 +1323,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_49 = paddle._C_ops.add(matmul_42, parameter_300) - del parameter_300 + del matmul_42, parameter_300 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_21 = paddle._C_ops.reshape(add_48, full_int_array_1) + del add_48 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_21 = paddle._C_ops.transpose(reshape_21, [0, 2, 1, 3]) @@ -1561,6 +1335,7 @@ def forward( # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_22 = paddle._C_ops.reshape(add_49, full_int_array_1) + del add_49 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_22 = paddle._C_ops.transpose(reshape_22, [0, 2, 1, 3]) @@ -1572,9 +1347,11 @@ def forward( # pd_op.matmul: (1x16x21x21xf32) <- (1x16x21x64xf32, 1x16x21x64xf32) matmul_43 = paddle._C_ops.matmul(scale_6, transpose_21, False, True) + del scale_6, transpose_21 # pd_op.add: (1x16x21x21xf32) <- (1x16x21x21xf32, 1x1x1x21xf32) add_50 = paddle._C_ops.add(matmul_43, unsqueeze_0) + del matmul_43 # pd_op.softmax: (1x16x21x21xf32) <- (1x16x21x21xf32) softmax_5 = paddle._C_ops.softmax(add_50, -1) @@ -1583,13 +1360,15 @@ def forward( # pd_op.dropout: (1x16x21x21xf32, 1x16x21x21xui8) <- (1x16x21x21xf32, None, 1xf32) dropout_32, dropout_33 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_5, None, full_4, False, "upscale_in_train", 0, False + softmax_5, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_5 # pd_op.matmul: (1x16x21x64xf32) <- (1x16x21x21xf32, 1x16x21x64xf32) matmul_44 = paddle._C_ops.matmul(dropout_32, transpose_22, False, False) + del dropout_32, transpose_22 # pd_op.transpose: (1x21x16x64xf32) <- (1x16x21x64xf32) transpose_23 = paddle._C_ops.transpose(matmul_44, [0, 2, 1, 3]) @@ -1597,19 +1376,20 @@ def forward( # pd_op.reshape: (1x21x1024xf32) <- (1x21x16x64xf32, 3xi64) reshape_23 = paddle._C_ops.reshape(transpose_23, full_int_array_2) + del transpose_23 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_45 = paddle._C_ops.matmul(reshape_23, parameter_299, False, False) - del parameter_299 + del parameter_299, reshape_23 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_51 = paddle._C_ops.add(matmul_45, parameter_298) - del parameter_298 + del matmul_45, parameter_298 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_34, dropout_35 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_51, None, full_4, False, "upscale_in_train", 0, False + add_51, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1617,6 +1397,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_52 = paddle._C_ops.add(layer_norm_30, dropout_34) + del dropout_34, layer_norm_30 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_33, layer_norm_34, layer_norm_35 = (lambda x, f: f(x))( @@ -1625,7 +1406,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_292, parameter_293 + del add_52, parameter_292, parameter_293 # pd_op.matmul: (1x21x4096xf32) <- (1x21x1024xf32, 1024x4096xf32) matmul_46 = paddle._C_ops.matmul(layer_norm_33, parameter_297, False, False) @@ -1633,23 +1414,24 @@ def forward( # pd_op.add: (1x21x4096xf32) <- (1x21x4096xf32, 4096xf32) add_53 = paddle._C_ops.add(matmul_46, parameter_296) - del parameter_296 + del matmul_46, parameter_296 # pd_op.gelu: (1x21x4096xf32) <- (1x21x4096xf32) gelu_5 = paddle._C_ops.gelu(add_53, False) + del add_53 # pd_op.matmul: (1x21x1024xf32) <- (1x21x4096xf32, 4096x1024xf32) matmul_47 = paddle._C_ops.matmul(gelu_5, parameter_295, False, False) - del parameter_295 + del gelu_5, parameter_295 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_54 = paddle._C_ops.add(matmul_47, parameter_294) - del parameter_294 + del matmul_47, parameter_294 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_36, dropout_37 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_54, None, full_4, False, "upscale_in_train", 0, False + add_54, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1657,6 +1439,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_55 = paddle._C_ops.add(layer_norm_33, dropout_36) + del dropout_36, layer_norm_33 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_36, layer_norm_37, layer_norm_38 = (lambda x, f: f(x))( @@ -1665,7 +1448,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_290, parameter_291 + del add_55, parameter_290, parameter_291 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_48 = paddle._C_ops.matmul(layer_norm_36, parameter_289, False, False) @@ -1673,10 +1456,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_56 = paddle._C_ops.add(matmul_48, parameter_288) - del parameter_288 + del matmul_48, parameter_288 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_24 = paddle._C_ops.reshape(add_56, full_int_array_1) + del add_56 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_24 = paddle._C_ops.transpose(reshape_24, [0, 2, 1, 3]) @@ -1688,7 +1472,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_57 = paddle._C_ops.add(matmul_49, parameter_286) - del parameter_286 + del matmul_49, parameter_286 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_50 = paddle._C_ops.matmul(layer_norm_36, parameter_285, False, False) @@ -1696,10 +1480,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_58 = paddle._C_ops.add(matmul_50, parameter_284) - del parameter_284 + del matmul_50, parameter_284 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_25 = paddle._C_ops.reshape(add_57, full_int_array_1) + del add_57 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_25 = paddle._C_ops.transpose(reshape_25, [0, 2, 1, 3]) @@ -1707,6 +1492,7 @@ def forward( # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_26 = paddle._C_ops.reshape(add_58, full_int_array_1) + del add_58 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_26 = paddle._C_ops.transpose(reshape_26, [0, 2, 1, 3]) @@ -1718,9 +1504,11 @@ def forward( # pd_op.matmul: (1x16x21x21xf32) <- (1x16x21x64xf32, 1x16x21x64xf32) matmul_51 = paddle._C_ops.matmul(scale_7, transpose_25, False, True) + del scale_7, transpose_25 # pd_op.add: (1x16x21x21xf32) <- (1x16x21x21xf32, 1x1x1x21xf32) add_59 = paddle._C_ops.add(matmul_51, unsqueeze_0) + del matmul_51 # pd_op.softmax: (1x16x21x21xf32) <- (1x16x21x21xf32) softmax_6 = paddle._C_ops.softmax(add_59, -1) @@ -1729,13 +1517,15 @@ def forward( # pd_op.dropout: (1x16x21x21xf32, 1x16x21x21xui8) <- (1x16x21x21xf32, None, 1xf32) dropout_38, dropout_39 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_6, None, full_4, False, "upscale_in_train", 0, False + softmax_6, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_6 # pd_op.matmul: (1x16x21x64xf32) <- (1x16x21x21xf32, 1x16x21x64xf32) matmul_52 = paddle._C_ops.matmul(dropout_38, transpose_26, False, False) + del dropout_38, transpose_26 # pd_op.transpose: (1x21x16x64xf32) <- (1x16x21x64xf32) transpose_27 = paddle._C_ops.transpose(matmul_52, [0, 2, 1, 3]) @@ -1743,19 +1533,20 @@ def forward( # pd_op.reshape: (1x21x1024xf32) <- (1x21x16x64xf32, 3xi64) reshape_27 = paddle._C_ops.reshape(transpose_27, full_int_array_2) + del transpose_27 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_53 = paddle._C_ops.matmul(reshape_27, parameter_283, False, False) - del parameter_283 + del parameter_283, reshape_27 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_60 = paddle._C_ops.add(matmul_53, parameter_282) - del parameter_282 + del matmul_53, parameter_282 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_40, dropout_41 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_60, None, full_4, False, "upscale_in_train", 0, False + add_60, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1763,6 +1554,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_61 = paddle._C_ops.add(layer_norm_36, dropout_40) + del dropout_40, layer_norm_36 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_39, layer_norm_40, layer_norm_41 = (lambda x, f: f(x))( @@ -1771,7 +1563,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_276, parameter_277 + del add_61, parameter_276, parameter_277 # pd_op.matmul: (1x21x4096xf32) <- (1x21x1024xf32, 1024x4096xf32) matmul_54 = paddle._C_ops.matmul(layer_norm_39, parameter_281, False, False) @@ -1779,23 +1571,24 @@ def forward( # pd_op.add: (1x21x4096xf32) <- (1x21x4096xf32, 4096xf32) add_62 = paddle._C_ops.add(matmul_54, parameter_280) - del parameter_280 + del matmul_54, parameter_280 # pd_op.gelu: (1x21x4096xf32) <- (1x21x4096xf32) gelu_6 = paddle._C_ops.gelu(add_62, False) + del add_62 # pd_op.matmul: (1x21x1024xf32) <- (1x21x4096xf32, 4096x1024xf32) matmul_55 = paddle._C_ops.matmul(gelu_6, parameter_279, False, False) - del parameter_279 + del gelu_6, parameter_279 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_63 = paddle._C_ops.add(matmul_55, parameter_278) - del parameter_278 + del matmul_55, parameter_278 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_42, dropout_43 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_63, None, full_4, False, "upscale_in_train", 0, False + add_63, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1803,6 +1596,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_64 = paddle._C_ops.add(layer_norm_39, dropout_42) + del dropout_42, layer_norm_39 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_42, layer_norm_43, layer_norm_44 = (lambda x, f: f(x))( @@ -1811,7 +1605,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_274, parameter_275 + del add_64, parameter_274, parameter_275 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_56 = paddle._C_ops.matmul(layer_norm_42, parameter_273, False, False) @@ -1819,10 +1613,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_65 = paddle._C_ops.add(matmul_56, parameter_272) - del parameter_272 + del matmul_56, parameter_272 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_28 = paddle._C_ops.reshape(add_65, full_int_array_1) + del add_65 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_28 = paddle._C_ops.transpose(reshape_28, [0, 2, 1, 3]) @@ -1834,7 +1629,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_66 = paddle._C_ops.add(matmul_57, parameter_270) - del parameter_270 + del matmul_57, parameter_270 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_58 = paddle._C_ops.matmul(layer_norm_42, parameter_269, False, False) @@ -1842,10 +1637,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_67 = paddle._C_ops.add(matmul_58, parameter_268) - del parameter_268 + del matmul_58, parameter_268 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_29 = paddle._C_ops.reshape(add_66, full_int_array_1) + del add_66 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_29 = paddle._C_ops.transpose(reshape_29, [0, 2, 1, 3]) @@ -1853,6 +1649,7 @@ def forward( # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_30 = paddle._C_ops.reshape(add_67, full_int_array_1) + del add_67 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_30 = paddle._C_ops.transpose(reshape_30, [0, 2, 1, 3]) @@ -1864,9 +1661,11 @@ def forward( # pd_op.matmul: (1x16x21x21xf32) <- (1x16x21x64xf32, 1x16x21x64xf32) matmul_59 = paddle._C_ops.matmul(scale_8, transpose_29, False, True) + del scale_8, transpose_29 # pd_op.add: (1x16x21x21xf32) <- (1x16x21x21xf32, 1x1x1x21xf32) add_68 = paddle._C_ops.add(matmul_59, unsqueeze_0) + del matmul_59 # pd_op.softmax: (1x16x21x21xf32) <- (1x16x21x21xf32) softmax_7 = paddle._C_ops.softmax(add_68, -1) @@ -1875,13 +1674,15 @@ def forward( # pd_op.dropout: (1x16x21x21xf32, 1x16x21x21xui8) <- (1x16x21x21xf32, None, 1xf32) dropout_44, dropout_45 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_7, None, full_4, False, "upscale_in_train", 0, False + softmax_7, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_7 # pd_op.matmul: (1x16x21x64xf32) <- (1x16x21x21xf32, 1x16x21x64xf32) matmul_60 = paddle._C_ops.matmul(dropout_44, transpose_30, False, False) + del dropout_44, transpose_30 # pd_op.transpose: (1x21x16x64xf32) <- (1x16x21x64xf32) transpose_31 = paddle._C_ops.transpose(matmul_60, [0, 2, 1, 3]) @@ -1889,19 +1690,20 @@ def forward( # pd_op.reshape: (1x21x1024xf32) <- (1x21x16x64xf32, 3xi64) reshape_31 = paddle._C_ops.reshape(transpose_31, full_int_array_2) + del transpose_31 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_61 = paddle._C_ops.matmul(reshape_31, parameter_267, False, False) - del parameter_267 + del parameter_267, reshape_31 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_69 = paddle._C_ops.add(matmul_61, parameter_266) - del parameter_266 + del matmul_61, parameter_266 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_46, dropout_47 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_69, None, full_4, False, "upscale_in_train", 0, False + add_69, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1909,6 +1711,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_70 = paddle._C_ops.add(layer_norm_42, dropout_46) + del dropout_46, layer_norm_42 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_45, layer_norm_46, layer_norm_47 = (lambda x, f: f(x))( @@ -1917,7 +1720,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_260, parameter_261 + del add_70, parameter_260, parameter_261 # pd_op.matmul: (1x21x4096xf32) <- (1x21x1024xf32, 1024x4096xf32) matmul_62 = paddle._C_ops.matmul(layer_norm_45, parameter_265, False, False) @@ -1925,23 +1728,24 @@ def forward( # pd_op.add: (1x21x4096xf32) <- (1x21x4096xf32, 4096xf32) add_71 = paddle._C_ops.add(matmul_62, parameter_264) - del parameter_264 + del matmul_62, parameter_264 # pd_op.gelu: (1x21x4096xf32) <- (1x21x4096xf32) gelu_7 = paddle._C_ops.gelu(add_71, False) + del add_71 # pd_op.matmul: (1x21x1024xf32) <- (1x21x4096xf32, 4096x1024xf32) matmul_63 = paddle._C_ops.matmul(gelu_7, parameter_263, False, False) - del parameter_263 + del gelu_7, parameter_263 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_72 = paddle._C_ops.add(matmul_63, parameter_262) - del parameter_262 + del matmul_63, parameter_262 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_48, dropout_49 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_72, None, full_4, False, "upscale_in_train", 0, False + add_72, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1949,6 +1753,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_73 = paddle._C_ops.add(layer_norm_45, dropout_48) + del dropout_48, layer_norm_45 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_48, layer_norm_49, layer_norm_50 = (lambda x, f: f(x))( @@ -1957,7 +1762,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_258, parameter_259 + del add_73, parameter_258, parameter_259 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_64 = paddle._C_ops.matmul(layer_norm_48, parameter_257, False, False) @@ -1965,10 +1770,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_74 = paddle._C_ops.add(matmul_64, parameter_256) - del parameter_256 + del matmul_64, parameter_256 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_32 = paddle._C_ops.reshape(add_74, full_int_array_1) + del add_74 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_32 = paddle._C_ops.transpose(reshape_32, [0, 2, 1, 3]) @@ -1980,7 +1786,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_75 = paddle._C_ops.add(matmul_65, parameter_254) - del parameter_254 + del matmul_65, parameter_254 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_66 = paddle._C_ops.matmul(layer_norm_48, parameter_253, False, False) @@ -1988,10 +1794,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_76 = paddle._C_ops.add(matmul_66, parameter_252) - del parameter_252 + del matmul_66, parameter_252 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_33 = paddle._C_ops.reshape(add_75, full_int_array_1) + del add_75 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_33 = paddle._C_ops.transpose(reshape_33, [0, 2, 1, 3]) @@ -1999,6 +1806,7 @@ def forward( # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_34 = paddle._C_ops.reshape(add_76, full_int_array_1) + del add_76 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_34 = paddle._C_ops.transpose(reshape_34, [0, 2, 1, 3]) @@ -2010,9 +1818,11 @@ def forward( # pd_op.matmul: (1x16x21x21xf32) <- (1x16x21x64xf32, 1x16x21x64xf32) matmul_67 = paddle._C_ops.matmul(scale_9, transpose_33, False, True) + del scale_9, transpose_33 # pd_op.add: (1x16x21x21xf32) <- (1x16x21x21xf32, 1x1x1x21xf32) add_77 = paddle._C_ops.add(matmul_67, unsqueeze_0) + del matmul_67 # pd_op.softmax: (1x16x21x21xf32) <- (1x16x21x21xf32) softmax_8 = paddle._C_ops.softmax(add_77, -1) @@ -2021,13 +1831,15 @@ def forward( # pd_op.dropout: (1x16x21x21xf32, 1x16x21x21xui8) <- (1x16x21x21xf32, None, 1xf32) dropout_50, dropout_51 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_8, None, full_4, False, "upscale_in_train", 0, False + softmax_8, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_8 # pd_op.matmul: (1x16x21x64xf32) <- (1x16x21x21xf32, 1x16x21x64xf32) matmul_68 = paddle._C_ops.matmul(dropout_50, transpose_34, False, False) + del dropout_50, transpose_34 # pd_op.transpose: (1x21x16x64xf32) <- (1x16x21x64xf32) transpose_35 = paddle._C_ops.transpose(matmul_68, [0, 2, 1, 3]) @@ -2035,19 +1847,20 @@ def forward( # pd_op.reshape: (1x21x1024xf32) <- (1x21x16x64xf32, 3xi64) reshape_35 = paddle._C_ops.reshape(transpose_35, full_int_array_2) + del transpose_35 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_69 = paddle._C_ops.matmul(reshape_35, parameter_251, False, False) - del parameter_251 + del parameter_251, reshape_35 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_78 = paddle._C_ops.add(matmul_69, parameter_250) - del parameter_250 + del matmul_69, parameter_250 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_52, dropout_53 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_78, None, full_4, False, "upscale_in_train", 0, False + add_78, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -2055,6 +1868,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_79 = paddle._C_ops.add(layer_norm_48, dropout_52) + del dropout_52, layer_norm_48 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_51, layer_norm_52, layer_norm_53 = (lambda x, f: f(x))( @@ -2063,7 +1877,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_244, parameter_245 + del add_79, parameter_244, parameter_245 # pd_op.matmul: (1x21x4096xf32) <- (1x21x1024xf32, 1024x4096xf32) matmul_70 = paddle._C_ops.matmul(layer_norm_51, parameter_249, False, False) @@ -2071,23 +1885,24 @@ def forward( # pd_op.add: (1x21x4096xf32) <- (1x21x4096xf32, 4096xf32) add_80 = paddle._C_ops.add(matmul_70, parameter_248) - del parameter_248 + del matmul_70, parameter_248 # pd_op.gelu: (1x21x4096xf32) <- (1x21x4096xf32) gelu_8 = paddle._C_ops.gelu(add_80, False) + del add_80 # pd_op.matmul: (1x21x1024xf32) <- (1x21x4096xf32, 4096x1024xf32) matmul_71 = paddle._C_ops.matmul(gelu_8, parameter_247, False, False) - del parameter_247 + del gelu_8, parameter_247 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_81 = paddle._C_ops.add(matmul_71, parameter_246) - del parameter_246 + del matmul_71, parameter_246 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_54, dropout_55 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_81, None, full_4, False, "upscale_in_train", 0, False + add_81, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -2095,6 +1910,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_82 = paddle._C_ops.add(layer_norm_51, dropout_54) + del dropout_54, layer_norm_51 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_54, layer_norm_55, layer_norm_56 = (lambda x, f: f(x))( @@ -2103,7 +1919,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_242, parameter_243 + del add_82, parameter_242, parameter_243 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_72 = paddle._C_ops.matmul(layer_norm_54, parameter_241, False, False) @@ -2111,10 +1927,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_83 = paddle._C_ops.add(matmul_72, parameter_240) - del parameter_240 + del matmul_72, parameter_240 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_36 = paddle._C_ops.reshape(add_83, full_int_array_1) + del add_83 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_36 = paddle._C_ops.transpose(reshape_36, [0, 2, 1, 3]) @@ -2126,7 +1943,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_84 = paddle._C_ops.add(matmul_73, parameter_238) - del parameter_238 + del matmul_73, parameter_238 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_74 = paddle._C_ops.matmul(layer_norm_54, parameter_237, False, False) @@ -2134,10 +1951,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_85 = paddle._C_ops.add(matmul_74, parameter_236) - del parameter_236 + del matmul_74, parameter_236 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_37 = paddle._C_ops.reshape(add_84, full_int_array_1) + del add_84 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_37 = paddle._C_ops.transpose(reshape_37, [0, 2, 1, 3]) @@ -2145,6 +1963,7 @@ def forward( # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_38 = paddle._C_ops.reshape(add_85, full_int_array_1) + del add_85 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_38 = paddle._C_ops.transpose(reshape_38, [0, 2, 1, 3]) @@ -2156,9 +1975,11 @@ def forward( # pd_op.matmul: (1x16x21x21xf32) <- (1x16x21x64xf32, 1x16x21x64xf32) matmul_75 = paddle._C_ops.matmul(scale_10, transpose_37, False, True) + del scale_10, transpose_37 # pd_op.add: (1x16x21x21xf32) <- (1x16x21x21xf32, 1x1x1x21xf32) add_86 = paddle._C_ops.add(matmul_75, unsqueeze_0) + del matmul_75 # pd_op.softmax: (1x16x21x21xf32) <- (1x16x21x21xf32) softmax_9 = paddle._C_ops.softmax(add_86, -1) @@ -2167,13 +1988,15 @@ def forward( # pd_op.dropout: (1x16x21x21xf32, 1x16x21x21xui8) <- (1x16x21x21xf32, None, 1xf32) dropout_56, dropout_57 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_9, None, full_4, False, "upscale_in_train", 0, False + softmax_9, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_9 # pd_op.matmul: (1x16x21x64xf32) <- (1x16x21x21xf32, 1x16x21x64xf32) matmul_76 = paddle._C_ops.matmul(dropout_56, transpose_38, False, False) + del dropout_56, transpose_38 # pd_op.transpose: (1x21x16x64xf32) <- (1x16x21x64xf32) transpose_39 = paddle._C_ops.transpose(matmul_76, [0, 2, 1, 3]) @@ -2181,19 +2004,20 @@ def forward( # pd_op.reshape: (1x21x1024xf32) <- (1x21x16x64xf32, 3xi64) reshape_39 = paddle._C_ops.reshape(transpose_39, full_int_array_2) + del transpose_39 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_77 = paddle._C_ops.matmul(reshape_39, parameter_235, False, False) - del parameter_235 + del parameter_235, reshape_39 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_87 = paddle._C_ops.add(matmul_77, parameter_234) - del parameter_234 + del matmul_77, parameter_234 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_58, dropout_59 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_87, None, full_4, False, "upscale_in_train", 0, False + add_87, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -2201,6 +2025,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_88 = paddle._C_ops.add(layer_norm_54, dropout_58) + del dropout_58, layer_norm_54 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_57, layer_norm_58, layer_norm_59 = (lambda x, f: f(x))( @@ -2209,7 +2034,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_228, parameter_229 + del add_88, parameter_228, parameter_229 # pd_op.matmul: (1x21x4096xf32) <- (1x21x1024xf32, 1024x4096xf32) matmul_78 = paddle._C_ops.matmul(layer_norm_57, parameter_233, False, False) @@ -2217,23 +2042,24 @@ def forward( # pd_op.add: (1x21x4096xf32) <- (1x21x4096xf32, 4096xf32) add_89 = paddle._C_ops.add(matmul_78, parameter_232) - del parameter_232 + del matmul_78, parameter_232 # pd_op.gelu: (1x21x4096xf32) <- (1x21x4096xf32) gelu_9 = paddle._C_ops.gelu(add_89, False) + del add_89 # pd_op.matmul: (1x21x1024xf32) <- (1x21x4096xf32, 4096x1024xf32) matmul_79 = paddle._C_ops.matmul(gelu_9, parameter_231, False, False) - del parameter_231 + del gelu_9, parameter_231 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_90 = paddle._C_ops.add(matmul_79, parameter_230) - del parameter_230 + del matmul_79, parameter_230 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_60, dropout_61 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_90, None, full_4, False, "upscale_in_train", 0, False + add_90, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -2241,6 +2067,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_91 = paddle._C_ops.add(layer_norm_57, dropout_60) + del dropout_60, layer_norm_57 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_60, layer_norm_61, layer_norm_62 = (lambda x, f: f(x))( @@ -2249,7 +2076,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_226, parameter_227 + del add_91, parameter_226, parameter_227 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_80 = paddle._C_ops.matmul(layer_norm_60, parameter_225, False, False) @@ -2257,10 +2084,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_92 = paddle._C_ops.add(matmul_80, parameter_224) - del parameter_224 + del matmul_80, parameter_224 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_40 = paddle._C_ops.reshape(add_92, full_int_array_1) + del add_92 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_40 = paddle._C_ops.transpose(reshape_40, [0, 2, 1, 3]) @@ -2272,7 +2100,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_93 = paddle._C_ops.add(matmul_81, parameter_222) - del parameter_222 + del matmul_81, parameter_222 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_82 = paddle._C_ops.matmul(layer_norm_60, parameter_221, False, False) @@ -2280,10 +2108,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_94 = paddle._C_ops.add(matmul_82, parameter_220) - del parameter_220 + del matmul_82, parameter_220 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_41 = paddle._C_ops.reshape(add_93, full_int_array_1) + del add_93 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_41 = paddle._C_ops.transpose(reshape_41, [0, 2, 1, 3]) @@ -2291,6 +2120,7 @@ def forward( # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_42 = paddle._C_ops.reshape(add_94, full_int_array_1) + del add_94 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_42 = paddle._C_ops.transpose(reshape_42, [0, 2, 1, 3]) @@ -2302,9 +2132,11 @@ def forward( # pd_op.matmul: (1x16x21x21xf32) <- (1x16x21x64xf32, 1x16x21x64xf32) matmul_83 = paddle._C_ops.matmul(scale_11, transpose_41, False, True) + del scale_11, transpose_41 # pd_op.add: (1x16x21x21xf32) <- (1x16x21x21xf32, 1x1x1x21xf32) add_95 = paddle._C_ops.add(matmul_83, unsqueeze_0) + del matmul_83 # pd_op.softmax: (1x16x21x21xf32) <- (1x16x21x21xf32) softmax_10 = paddle._C_ops.softmax(add_95, -1) @@ -2313,13 +2145,15 @@ def forward( # pd_op.dropout: (1x16x21x21xf32, 1x16x21x21xui8) <- (1x16x21x21xf32, None, 1xf32) dropout_62, dropout_63 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_10, None, full_4, False, "upscale_in_train", 0, False + softmax_10, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_10 # pd_op.matmul: (1x16x21x64xf32) <- (1x16x21x21xf32, 1x16x21x64xf32) matmul_84 = paddle._C_ops.matmul(dropout_62, transpose_42, False, False) + del dropout_62, transpose_42 # pd_op.transpose: (1x21x16x64xf32) <- (1x16x21x64xf32) transpose_43 = paddle._C_ops.transpose(matmul_84, [0, 2, 1, 3]) @@ -2327,19 +2161,20 @@ def forward( # pd_op.reshape: (1x21x1024xf32) <- (1x21x16x64xf32, 3xi64) reshape_43 = paddle._C_ops.reshape(transpose_43, full_int_array_2) + del transpose_43 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_85 = paddle._C_ops.matmul(reshape_43, parameter_219, False, False) - del parameter_219 + del parameter_219, reshape_43 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_96 = paddle._C_ops.add(matmul_85, parameter_218) - del parameter_218 + del matmul_85, parameter_218 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_64, dropout_65 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_96, None, full_4, False, "upscale_in_train", 0, False + add_96, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -2347,6 +2182,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_97 = paddle._C_ops.add(layer_norm_60, dropout_64) + del dropout_64, layer_norm_60 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_63, layer_norm_64, layer_norm_65 = (lambda x, f: f(x))( @@ -2355,7 +2191,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_212, parameter_213 + del add_97, parameter_212, parameter_213 # pd_op.matmul: (1x21x4096xf32) <- (1x21x1024xf32, 1024x4096xf32) matmul_86 = paddle._C_ops.matmul(layer_norm_63, parameter_217, False, False) @@ -2363,23 +2199,24 @@ def forward( # pd_op.add: (1x21x4096xf32) <- (1x21x4096xf32, 4096xf32) add_98 = paddle._C_ops.add(matmul_86, parameter_216) - del parameter_216 + del matmul_86, parameter_216 # pd_op.gelu: (1x21x4096xf32) <- (1x21x4096xf32) gelu_10 = paddle._C_ops.gelu(add_98, False) + del add_98 # pd_op.matmul: (1x21x1024xf32) <- (1x21x4096xf32, 4096x1024xf32) matmul_87 = paddle._C_ops.matmul(gelu_10, parameter_215, False, False) - del parameter_215 + del gelu_10, parameter_215 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_99 = paddle._C_ops.add(matmul_87, parameter_214) - del parameter_214 + del matmul_87, parameter_214 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_66, dropout_67 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_99, None, full_4, False, "upscale_in_train", 0, False + add_99, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -2387,6 +2224,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_100 = paddle._C_ops.add(layer_norm_63, dropout_66) + del dropout_66, layer_norm_63 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_66, layer_norm_67, layer_norm_68 = (lambda x, f: f(x))( @@ -2395,7 +2233,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_210, parameter_211 + del add_100, parameter_210, parameter_211 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_88 = paddle._C_ops.matmul(layer_norm_66, parameter_209, False, False) @@ -2403,10 +2241,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_101 = paddle._C_ops.add(matmul_88, parameter_208) - del parameter_208 + del matmul_88, parameter_208 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_44 = paddle._C_ops.reshape(add_101, full_int_array_1) + del add_101 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_44 = paddle._C_ops.transpose(reshape_44, [0, 2, 1, 3]) @@ -2418,7 +2257,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_102 = paddle._C_ops.add(matmul_89, parameter_206) - del parameter_206 + del matmul_89, parameter_206 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_90 = paddle._C_ops.matmul(layer_norm_66, parameter_205, False, False) @@ -2426,10 +2265,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_103 = paddle._C_ops.add(matmul_90, parameter_204) - del parameter_204 + del matmul_90, parameter_204 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_45 = paddle._C_ops.reshape(add_102, full_int_array_1) + del add_102 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_45 = paddle._C_ops.transpose(reshape_45, [0, 2, 1, 3]) @@ -2437,6 +2277,7 @@ def forward( # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_46 = paddle._C_ops.reshape(add_103, full_int_array_1) + del add_103 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_46 = paddle._C_ops.transpose(reshape_46, [0, 2, 1, 3]) @@ -2448,9 +2289,11 @@ def forward( # pd_op.matmul: (1x16x21x21xf32) <- (1x16x21x64xf32, 1x16x21x64xf32) matmul_91 = paddle._C_ops.matmul(scale_12, transpose_45, False, True) + del scale_12, transpose_45 # pd_op.add: (1x16x21x21xf32) <- (1x16x21x21xf32, 1x1x1x21xf32) add_104 = paddle._C_ops.add(matmul_91, unsqueeze_0) + del matmul_91 # pd_op.softmax: (1x16x21x21xf32) <- (1x16x21x21xf32) softmax_11 = paddle._C_ops.softmax(add_104, -1) @@ -2459,13 +2302,15 @@ def forward( # pd_op.dropout: (1x16x21x21xf32, 1x16x21x21xui8) <- (1x16x21x21xf32, None, 1xf32) dropout_68, dropout_69 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_11, None, full_4, False, "upscale_in_train", 0, False + softmax_11, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_11 # pd_op.matmul: (1x16x21x64xf32) <- (1x16x21x21xf32, 1x16x21x64xf32) matmul_92 = paddle._C_ops.matmul(dropout_68, transpose_46, False, False) + del dropout_68, transpose_46 # pd_op.transpose: (1x21x16x64xf32) <- (1x16x21x64xf32) transpose_47 = paddle._C_ops.transpose(matmul_92, [0, 2, 1, 3]) @@ -2473,19 +2318,20 @@ def forward( # pd_op.reshape: (1x21x1024xf32) <- (1x21x16x64xf32, 3xi64) reshape_47 = paddle._C_ops.reshape(transpose_47, full_int_array_2) + del transpose_47 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_93 = paddle._C_ops.matmul(reshape_47, parameter_203, False, False) - del parameter_203 + del parameter_203, reshape_47 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_105 = paddle._C_ops.add(matmul_93, parameter_202) - del parameter_202 + del matmul_93, parameter_202 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_70, dropout_71 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_105, None, full_4, False, "upscale_in_train", 0, False + add_105, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -2493,6 +2339,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_106 = paddle._C_ops.add(layer_norm_66, dropout_70) + del dropout_70, layer_norm_66 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_69, layer_norm_70, layer_norm_71 = (lambda x, f: f(x))( @@ -2501,7 +2348,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_196, parameter_197 + del add_106, parameter_196, parameter_197 # pd_op.matmul: (1x21x4096xf32) <- (1x21x1024xf32, 1024x4096xf32) matmul_94 = paddle._C_ops.matmul(layer_norm_69, parameter_201, False, False) @@ -2509,23 +2356,24 @@ def forward( # pd_op.add: (1x21x4096xf32) <- (1x21x4096xf32, 4096xf32) add_107 = paddle._C_ops.add(matmul_94, parameter_200) - del parameter_200 + del matmul_94, parameter_200 # pd_op.gelu: (1x21x4096xf32) <- (1x21x4096xf32) gelu_11 = paddle._C_ops.gelu(add_107, False) + del add_107 # pd_op.matmul: (1x21x1024xf32) <- (1x21x4096xf32, 4096x1024xf32) matmul_95 = paddle._C_ops.matmul(gelu_11, parameter_199, False, False) - del parameter_199 + del gelu_11, parameter_199 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_108 = paddle._C_ops.add(matmul_95, parameter_198) - del parameter_198 + del matmul_95, parameter_198 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_72, dropout_73 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_108, None, full_4, False, "upscale_in_train", 0, False + add_108, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -2533,6 +2381,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_109 = paddle._C_ops.add(layer_norm_69, dropout_72) + del dropout_72, layer_norm_69 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_72, layer_norm_73, layer_norm_74 = (lambda x, f: f(x))( @@ -2541,7 +2390,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_194, parameter_195 + del add_109, parameter_194, parameter_195 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_96 = paddle._C_ops.matmul(layer_norm_72, parameter_193, False, False) @@ -2549,10 +2398,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_110 = paddle._C_ops.add(matmul_96, parameter_192) - del parameter_192 + del matmul_96, parameter_192 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_48 = paddle._C_ops.reshape(add_110, full_int_array_1) + del add_110 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_48 = paddle._C_ops.transpose(reshape_48, [0, 2, 1, 3]) @@ -2564,7 +2414,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_111 = paddle._C_ops.add(matmul_97, parameter_190) - del parameter_190 + del matmul_97, parameter_190 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_98 = paddle._C_ops.matmul(layer_norm_72, parameter_189, False, False) @@ -2572,10 +2422,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_112 = paddle._C_ops.add(matmul_98, parameter_188) - del parameter_188 + del matmul_98, parameter_188 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_49 = paddle._C_ops.reshape(add_111, full_int_array_1) + del add_111 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_49 = paddle._C_ops.transpose(reshape_49, [0, 2, 1, 3]) @@ -2583,6 +2434,7 @@ def forward( # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_50 = paddle._C_ops.reshape(add_112, full_int_array_1) + del add_112 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_50 = paddle._C_ops.transpose(reshape_50, [0, 2, 1, 3]) @@ -2594,9 +2446,11 @@ def forward( # pd_op.matmul: (1x16x21x21xf32) <- (1x16x21x64xf32, 1x16x21x64xf32) matmul_99 = paddle._C_ops.matmul(scale_13, transpose_49, False, True) + del scale_13, transpose_49 # pd_op.add: (1x16x21x21xf32) <- (1x16x21x21xf32, 1x1x1x21xf32) add_113 = paddle._C_ops.add(matmul_99, unsqueeze_0) + del matmul_99 # pd_op.softmax: (1x16x21x21xf32) <- (1x16x21x21xf32) softmax_12 = paddle._C_ops.softmax(add_113, -1) @@ -2605,13 +2459,15 @@ def forward( # pd_op.dropout: (1x16x21x21xf32, 1x16x21x21xui8) <- (1x16x21x21xf32, None, 1xf32) dropout_74, dropout_75 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_12, None, full_4, False, "upscale_in_train", 0, False + softmax_12, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_12 # pd_op.matmul: (1x16x21x64xf32) <- (1x16x21x21xf32, 1x16x21x64xf32) matmul_100 = paddle._C_ops.matmul(dropout_74, transpose_50, False, False) + del dropout_74, transpose_50 # pd_op.transpose: (1x21x16x64xf32) <- (1x16x21x64xf32) transpose_51 = paddle._C_ops.transpose(matmul_100, [0, 2, 1, 3]) @@ -2619,19 +2475,20 @@ def forward( # pd_op.reshape: (1x21x1024xf32) <- (1x21x16x64xf32, 3xi64) reshape_51 = paddle._C_ops.reshape(transpose_51, full_int_array_2) + del transpose_51 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_101 = paddle._C_ops.matmul(reshape_51, parameter_187, False, False) - del parameter_187 + del parameter_187, reshape_51 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_114 = paddle._C_ops.add(matmul_101, parameter_186) - del parameter_186 + del matmul_101, parameter_186 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_76, dropout_77 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_114, None, full_4, False, "upscale_in_train", 0, False + add_114, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -2639,6 +2496,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_115 = paddle._C_ops.add(layer_norm_72, dropout_76) + del dropout_76, layer_norm_72 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_75, layer_norm_76, layer_norm_77 = (lambda x, f: f(x))( @@ -2647,7 +2505,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_180, parameter_181 + del add_115, parameter_180, parameter_181 # pd_op.matmul: (1x21x4096xf32) <- (1x21x1024xf32, 1024x4096xf32) matmul_102 = paddle._C_ops.matmul(layer_norm_75, parameter_185, False, False) @@ -2655,23 +2513,24 @@ def forward( # pd_op.add: (1x21x4096xf32) <- (1x21x4096xf32, 4096xf32) add_116 = paddle._C_ops.add(matmul_102, parameter_184) - del parameter_184 + del matmul_102, parameter_184 # pd_op.gelu: (1x21x4096xf32) <- (1x21x4096xf32) gelu_12 = paddle._C_ops.gelu(add_116, False) + del add_116 # pd_op.matmul: (1x21x1024xf32) <- (1x21x4096xf32, 4096x1024xf32) matmul_103 = paddle._C_ops.matmul(gelu_12, parameter_183, False, False) - del parameter_183 + del gelu_12, parameter_183 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_117 = paddle._C_ops.add(matmul_103, parameter_182) - del parameter_182 + del matmul_103, parameter_182 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_78, dropout_79 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_117, None, full_4, False, "upscale_in_train", 0, False + add_117, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -2679,6 +2538,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_118 = paddle._C_ops.add(layer_norm_75, dropout_78) + del dropout_78, layer_norm_75 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_78, layer_norm_79, layer_norm_80 = (lambda x, f: f(x))( @@ -2687,7 +2547,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_178, parameter_179 + del add_118, parameter_178, parameter_179 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_104 = paddle._C_ops.matmul(layer_norm_78, parameter_177, False, False) @@ -2695,10 +2555,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_119 = paddle._C_ops.add(matmul_104, parameter_176) - del parameter_176 + del matmul_104, parameter_176 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_52 = paddle._C_ops.reshape(add_119, full_int_array_1) + del add_119 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_52 = paddle._C_ops.transpose(reshape_52, [0, 2, 1, 3]) @@ -2710,7 +2571,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_120 = paddle._C_ops.add(matmul_105, parameter_174) - del parameter_174 + del matmul_105, parameter_174 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_106 = paddle._C_ops.matmul(layer_norm_78, parameter_173, False, False) @@ -2718,10 +2579,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_121 = paddle._C_ops.add(matmul_106, parameter_172) - del parameter_172 + del matmul_106, parameter_172 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_53 = paddle._C_ops.reshape(add_120, full_int_array_1) + del add_120 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_53 = paddle._C_ops.transpose(reshape_53, [0, 2, 1, 3]) @@ -2729,6 +2591,7 @@ def forward( # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_54 = paddle._C_ops.reshape(add_121, full_int_array_1) + del add_121 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_54 = paddle._C_ops.transpose(reshape_54, [0, 2, 1, 3]) @@ -2740,9 +2603,11 @@ def forward( # pd_op.matmul: (1x16x21x21xf32) <- (1x16x21x64xf32, 1x16x21x64xf32) matmul_107 = paddle._C_ops.matmul(scale_14, transpose_53, False, True) + del scale_14, transpose_53 # pd_op.add: (1x16x21x21xf32) <- (1x16x21x21xf32, 1x1x1x21xf32) add_122 = paddle._C_ops.add(matmul_107, unsqueeze_0) + del matmul_107 # pd_op.softmax: (1x16x21x21xf32) <- (1x16x21x21xf32) softmax_13 = paddle._C_ops.softmax(add_122, -1) @@ -2751,13 +2616,15 @@ def forward( # pd_op.dropout: (1x16x21x21xf32, 1x16x21x21xui8) <- (1x16x21x21xf32, None, 1xf32) dropout_80, dropout_81 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_13, None, full_4, False, "upscale_in_train", 0, False + softmax_13, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_13 # pd_op.matmul: (1x16x21x64xf32) <- (1x16x21x21xf32, 1x16x21x64xf32) matmul_108 = paddle._C_ops.matmul(dropout_80, transpose_54, False, False) + del dropout_80, transpose_54 # pd_op.transpose: (1x21x16x64xf32) <- (1x16x21x64xf32) transpose_55 = paddle._C_ops.transpose(matmul_108, [0, 2, 1, 3]) @@ -2765,19 +2632,20 @@ def forward( # pd_op.reshape: (1x21x1024xf32) <- (1x21x16x64xf32, 3xi64) reshape_55 = paddle._C_ops.reshape(transpose_55, full_int_array_2) + del transpose_55 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_109 = paddle._C_ops.matmul(reshape_55, parameter_171, False, False) - del parameter_171 + del parameter_171, reshape_55 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_123 = paddle._C_ops.add(matmul_109, parameter_170) - del parameter_170 + del matmul_109, parameter_170 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_82, dropout_83 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_123, None, full_4, False, "upscale_in_train", 0, False + add_123, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -2785,6 +2653,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_124 = paddle._C_ops.add(layer_norm_78, dropout_82) + del dropout_82, layer_norm_78 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_81, layer_norm_82, layer_norm_83 = (lambda x, f: f(x))( @@ -2793,7 +2662,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_164, parameter_165 + del add_124, parameter_164, parameter_165 # pd_op.matmul: (1x21x4096xf32) <- (1x21x1024xf32, 1024x4096xf32) matmul_110 = paddle._C_ops.matmul(layer_norm_81, parameter_169, False, False) @@ -2801,23 +2670,24 @@ def forward( # pd_op.add: (1x21x4096xf32) <- (1x21x4096xf32, 4096xf32) add_125 = paddle._C_ops.add(matmul_110, parameter_168) - del parameter_168 + del matmul_110, parameter_168 # pd_op.gelu: (1x21x4096xf32) <- (1x21x4096xf32) gelu_13 = paddle._C_ops.gelu(add_125, False) + del add_125 # pd_op.matmul: (1x21x1024xf32) <- (1x21x4096xf32, 4096x1024xf32) matmul_111 = paddle._C_ops.matmul(gelu_13, parameter_167, False, False) - del parameter_167 + del gelu_13, parameter_167 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_126 = paddle._C_ops.add(matmul_111, parameter_166) - del parameter_166 + del matmul_111, parameter_166 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_84, dropout_85 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_126, None, full_4, False, "upscale_in_train", 0, False + add_126, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -2825,6 +2695,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_127 = paddle._C_ops.add(layer_norm_81, dropout_84) + del dropout_84, layer_norm_81 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_84, layer_norm_85, layer_norm_86 = (lambda x, f: f(x))( @@ -2833,7 +2704,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_162, parameter_163 + del add_127, parameter_162, parameter_163 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_112 = paddle._C_ops.matmul(layer_norm_84, parameter_161, False, False) @@ -2841,10 +2712,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_128 = paddle._C_ops.add(matmul_112, parameter_160) - del parameter_160 + del matmul_112, parameter_160 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_56 = paddle._C_ops.reshape(add_128, full_int_array_1) + del add_128 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_56 = paddle._C_ops.transpose(reshape_56, [0, 2, 1, 3]) @@ -2856,7 +2728,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_129 = paddle._C_ops.add(matmul_113, parameter_158) - del parameter_158 + del matmul_113, parameter_158 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_114 = paddle._C_ops.matmul(layer_norm_84, parameter_157, False, False) @@ -2864,10 +2736,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_130 = paddle._C_ops.add(matmul_114, parameter_156) - del parameter_156 + del matmul_114, parameter_156 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_57 = paddle._C_ops.reshape(add_129, full_int_array_1) + del add_129 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_57 = paddle._C_ops.transpose(reshape_57, [0, 2, 1, 3]) @@ -2875,6 +2748,7 @@ def forward( # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_58 = paddle._C_ops.reshape(add_130, full_int_array_1) + del add_130 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_58 = paddle._C_ops.transpose(reshape_58, [0, 2, 1, 3]) @@ -2886,9 +2760,11 @@ def forward( # pd_op.matmul: (1x16x21x21xf32) <- (1x16x21x64xf32, 1x16x21x64xf32) matmul_115 = paddle._C_ops.matmul(scale_15, transpose_57, False, True) + del scale_15, transpose_57 # pd_op.add: (1x16x21x21xf32) <- (1x16x21x21xf32, 1x1x1x21xf32) add_131 = paddle._C_ops.add(matmul_115, unsqueeze_0) + del matmul_115 # pd_op.softmax: (1x16x21x21xf32) <- (1x16x21x21xf32) softmax_14 = paddle._C_ops.softmax(add_131, -1) @@ -2897,13 +2773,15 @@ def forward( # pd_op.dropout: (1x16x21x21xf32, 1x16x21x21xui8) <- (1x16x21x21xf32, None, 1xf32) dropout_86, dropout_87 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_14, None, full_4, False, "upscale_in_train", 0, False + softmax_14, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_14 # pd_op.matmul: (1x16x21x64xf32) <- (1x16x21x21xf32, 1x16x21x64xf32) matmul_116 = paddle._C_ops.matmul(dropout_86, transpose_58, False, False) + del dropout_86, transpose_58 # pd_op.transpose: (1x21x16x64xf32) <- (1x16x21x64xf32) transpose_59 = paddle._C_ops.transpose(matmul_116, [0, 2, 1, 3]) @@ -2911,19 +2789,20 @@ def forward( # pd_op.reshape: (1x21x1024xf32) <- (1x21x16x64xf32, 3xi64) reshape_59 = paddle._C_ops.reshape(transpose_59, full_int_array_2) + del transpose_59 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_117 = paddle._C_ops.matmul(reshape_59, parameter_155, False, False) - del parameter_155 + del parameter_155, reshape_59 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_132 = paddle._C_ops.add(matmul_117, parameter_154) - del parameter_154 + del matmul_117, parameter_154 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_88, dropout_89 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_132, None, full_4, False, "upscale_in_train", 0, False + add_132, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -2931,6 +2810,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_133 = paddle._C_ops.add(layer_norm_84, dropout_88) + del dropout_88, layer_norm_84 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_87, layer_norm_88, layer_norm_89 = (lambda x, f: f(x))( @@ -2939,7 +2819,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_148, parameter_149 + del add_133, parameter_148, parameter_149 # pd_op.matmul: (1x21x4096xf32) <- (1x21x1024xf32, 1024x4096xf32) matmul_118 = paddle._C_ops.matmul(layer_norm_87, parameter_153, False, False) @@ -2947,23 +2827,24 @@ def forward( # pd_op.add: (1x21x4096xf32) <- (1x21x4096xf32, 4096xf32) add_134 = paddle._C_ops.add(matmul_118, parameter_152) - del parameter_152 + del matmul_118, parameter_152 # pd_op.gelu: (1x21x4096xf32) <- (1x21x4096xf32) gelu_14 = paddle._C_ops.gelu(add_134, False) + del add_134 # pd_op.matmul: (1x21x1024xf32) <- (1x21x4096xf32, 4096x1024xf32) matmul_119 = paddle._C_ops.matmul(gelu_14, parameter_151, False, False) - del parameter_151 + del gelu_14, parameter_151 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_135 = paddle._C_ops.add(matmul_119, parameter_150) - del parameter_150 + del matmul_119, parameter_150 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_90, dropout_91 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_135, None, full_4, False, "upscale_in_train", 0, False + add_135, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -2971,6 +2852,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_136 = paddle._C_ops.add(layer_norm_87, dropout_90) + del dropout_90, layer_norm_87 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_90, layer_norm_91, layer_norm_92 = (lambda x, f: f(x))( @@ -2979,7 +2861,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_146, parameter_147 + del add_136, parameter_146, parameter_147 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_120 = paddle._C_ops.matmul(layer_norm_90, parameter_145, False, False) @@ -2987,10 +2869,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_137 = paddle._C_ops.add(matmul_120, parameter_144) - del parameter_144 + del matmul_120, parameter_144 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_60 = paddle._C_ops.reshape(add_137, full_int_array_1) + del add_137 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_60 = paddle._C_ops.transpose(reshape_60, [0, 2, 1, 3]) @@ -3002,7 +2885,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_138 = paddle._C_ops.add(matmul_121, parameter_142) - del parameter_142 + del matmul_121, parameter_142 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_122 = paddle._C_ops.matmul(layer_norm_90, parameter_141, False, False) @@ -3010,10 +2893,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_139 = paddle._C_ops.add(matmul_122, parameter_140) - del parameter_140 + del matmul_122, parameter_140 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_61 = paddle._C_ops.reshape(add_138, full_int_array_1) + del add_138 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_61 = paddle._C_ops.transpose(reshape_61, [0, 2, 1, 3]) @@ -3021,6 +2905,7 @@ def forward( # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_62 = paddle._C_ops.reshape(add_139, full_int_array_1) + del add_139 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_62 = paddle._C_ops.transpose(reshape_62, [0, 2, 1, 3]) @@ -3032,9 +2917,11 @@ def forward( # pd_op.matmul: (1x16x21x21xf32) <- (1x16x21x64xf32, 1x16x21x64xf32) matmul_123 = paddle._C_ops.matmul(scale_16, transpose_61, False, True) + del scale_16, transpose_61 # pd_op.add: (1x16x21x21xf32) <- (1x16x21x21xf32, 1x1x1x21xf32) add_140 = paddle._C_ops.add(matmul_123, unsqueeze_0) + del matmul_123 # pd_op.softmax: (1x16x21x21xf32) <- (1x16x21x21xf32) softmax_15 = paddle._C_ops.softmax(add_140, -1) @@ -3043,13 +2930,15 @@ def forward( # pd_op.dropout: (1x16x21x21xf32, 1x16x21x21xui8) <- (1x16x21x21xf32, None, 1xf32) dropout_92, dropout_93 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_15, None, full_4, False, "upscale_in_train", 0, False + softmax_15, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_15 # pd_op.matmul: (1x16x21x64xf32) <- (1x16x21x21xf32, 1x16x21x64xf32) matmul_124 = paddle._C_ops.matmul(dropout_92, transpose_62, False, False) + del dropout_92, transpose_62 # pd_op.transpose: (1x21x16x64xf32) <- (1x16x21x64xf32) transpose_63 = paddle._C_ops.transpose(matmul_124, [0, 2, 1, 3]) @@ -3057,19 +2946,20 @@ def forward( # pd_op.reshape: (1x21x1024xf32) <- (1x21x16x64xf32, 3xi64) reshape_63 = paddle._C_ops.reshape(transpose_63, full_int_array_2) + del transpose_63 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_125 = paddle._C_ops.matmul(reshape_63, parameter_139, False, False) - del parameter_139 + del parameter_139, reshape_63 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_141 = paddle._C_ops.add(matmul_125, parameter_138) - del parameter_138 + del matmul_125, parameter_138 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_94, dropout_95 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_141, None, full_4, False, "upscale_in_train", 0, False + add_141, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -3077,6 +2967,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_142 = paddle._C_ops.add(layer_norm_90, dropout_94) + del dropout_94, layer_norm_90 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_93, layer_norm_94, layer_norm_95 = (lambda x, f: f(x))( @@ -3085,7 +2976,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_132, parameter_133 + del add_142, parameter_132, parameter_133 # pd_op.matmul: (1x21x4096xf32) <- (1x21x1024xf32, 1024x4096xf32) matmul_126 = paddle._C_ops.matmul(layer_norm_93, parameter_137, False, False) @@ -3093,23 +2984,24 @@ def forward( # pd_op.add: (1x21x4096xf32) <- (1x21x4096xf32, 4096xf32) add_143 = paddle._C_ops.add(matmul_126, parameter_136) - del parameter_136 + del matmul_126, parameter_136 # pd_op.gelu: (1x21x4096xf32) <- (1x21x4096xf32) gelu_15 = paddle._C_ops.gelu(add_143, False) + del add_143 # pd_op.matmul: (1x21x1024xf32) <- (1x21x4096xf32, 4096x1024xf32) matmul_127 = paddle._C_ops.matmul(gelu_15, parameter_135, False, False) - del parameter_135 + del gelu_15, parameter_135 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_144 = paddle._C_ops.add(matmul_127, parameter_134) - del parameter_134 + del matmul_127, parameter_134 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_96, dropout_97 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_144, None, full_4, False, "upscale_in_train", 0, False + add_144, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -3117,6 +3009,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_145 = paddle._C_ops.add(layer_norm_93, dropout_96) + del dropout_96, layer_norm_93 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_96, layer_norm_97, layer_norm_98 = (lambda x, f: f(x))( @@ -3125,7 +3018,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_130, parameter_131 + del add_145, parameter_130, parameter_131 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_128 = paddle._C_ops.matmul(layer_norm_96, parameter_129, False, False) @@ -3133,10 +3026,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_146 = paddle._C_ops.add(matmul_128, parameter_128) - del parameter_128 + del matmul_128, parameter_128 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_64 = paddle._C_ops.reshape(add_146, full_int_array_1) + del add_146 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_64 = paddle._C_ops.transpose(reshape_64, [0, 2, 1, 3]) @@ -3148,7 +3042,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_147 = paddle._C_ops.add(matmul_129, parameter_126) - del parameter_126 + del matmul_129, parameter_126 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_130 = paddle._C_ops.matmul(layer_norm_96, parameter_125, False, False) @@ -3156,10 +3050,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_148 = paddle._C_ops.add(matmul_130, parameter_124) - del parameter_124 + del matmul_130, parameter_124 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_65 = paddle._C_ops.reshape(add_147, full_int_array_1) + del add_147 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_65 = paddle._C_ops.transpose(reshape_65, [0, 2, 1, 3]) @@ -3167,6 +3062,7 @@ def forward( # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_66 = paddle._C_ops.reshape(add_148, full_int_array_1) + del add_148 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_66 = paddle._C_ops.transpose(reshape_66, [0, 2, 1, 3]) @@ -3178,9 +3074,11 @@ def forward( # pd_op.matmul: (1x16x21x21xf32) <- (1x16x21x64xf32, 1x16x21x64xf32) matmul_131 = paddle._C_ops.matmul(scale_17, transpose_65, False, True) + del scale_17, transpose_65 # pd_op.add: (1x16x21x21xf32) <- (1x16x21x21xf32, 1x1x1x21xf32) add_149 = paddle._C_ops.add(matmul_131, unsqueeze_0) + del matmul_131 # pd_op.softmax: (1x16x21x21xf32) <- (1x16x21x21xf32) softmax_16 = paddle._C_ops.softmax(add_149, -1) @@ -3189,13 +3087,15 @@ def forward( # pd_op.dropout: (1x16x21x21xf32, 1x16x21x21xui8) <- (1x16x21x21xf32, None, 1xf32) dropout_98, dropout_99 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_16, None, full_4, False, "upscale_in_train", 0, False + softmax_16, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_16 # pd_op.matmul: (1x16x21x64xf32) <- (1x16x21x21xf32, 1x16x21x64xf32) matmul_132 = paddle._C_ops.matmul(dropout_98, transpose_66, False, False) + del dropout_98, transpose_66 # pd_op.transpose: (1x21x16x64xf32) <- (1x16x21x64xf32) transpose_67 = paddle._C_ops.transpose(matmul_132, [0, 2, 1, 3]) @@ -3203,19 +3103,20 @@ def forward( # pd_op.reshape: (1x21x1024xf32) <- (1x21x16x64xf32, 3xi64) reshape_67 = paddle._C_ops.reshape(transpose_67, full_int_array_2) + del transpose_67 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_133 = paddle._C_ops.matmul(reshape_67, parameter_123, False, False) - del parameter_123 + del parameter_123, reshape_67 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_150 = paddle._C_ops.add(matmul_133, parameter_122) - del parameter_122 + del matmul_133, parameter_122 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_100, dropout_101 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_150, None, full_4, False, "upscale_in_train", 0, False + add_150, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -3223,6 +3124,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_151 = paddle._C_ops.add(layer_norm_96, dropout_100) + del dropout_100, layer_norm_96 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_99, layer_norm_100, layer_norm_101 = (lambda x, f: f(x))( @@ -3231,7 +3133,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_116, parameter_117 + del add_151, parameter_116, parameter_117 # pd_op.matmul: (1x21x4096xf32) <- (1x21x1024xf32, 1024x4096xf32) matmul_134 = paddle._C_ops.matmul(layer_norm_99, parameter_121, False, False) @@ -3239,23 +3141,24 @@ def forward( # pd_op.add: (1x21x4096xf32) <- (1x21x4096xf32, 4096xf32) add_152 = paddle._C_ops.add(matmul_134, parameter_120) - del parameter_120 + del matmul_134, parameter_120 # pd_op.gelu: (1x21x4096xf32) <- (1x21x4096xf32) gelu_16 = paddle._C_ops.gelu(add_152, False) + del add_152 # pd_op.matmul: (1x21x1024xf32) <- (1x21x4096xf32, 4096x1024xf32) matmul_135 = paddle._C_ops.matmul(gelu_16, parameter_119, False, False) - del parameter_119 + del gelu_16, parameter_119 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_153 = paddle._C_ops.add(matmul_135, parameter_118) - del parameter_118 + del matmul_135, parameter_118 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_102, dropout_103 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_153, None, full_4, False, "upscale_in_train", 0, False + add_153, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -3263,6 +3166,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_154 = paddle._C_ops.add(layer_norm_99, dropout_102) + del dropout_102, layer_norm_99 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_102, layer_norm_103, layer_norm_104 = (lambda x, f: f(x))( @@ -3271,7 +3175,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_114, parameter_115 + del add_154, parameter_114, parameter_115 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_136 = paddle._C_ops.matmul(layer_norm_102, parameter_113, False, False) @@ -3279,10 +3183,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_155 = paddle._C_ops.add(matmul_136, parameter_112) - del parameter_112 + del matmul_136, parameter_112 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_68 = paddle._C_ops.reshape(add_155, full_int_array_1) + del add_155 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_68 = paddle._C_ops.transpose(reshape_68, [0, 2, 1, 3]) @@ -3294,7 +3199,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_156 = paddle._C_ops.add(matmul_137, parameter_110) - del parameter_110 + del matmul_137, parameter_110 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_138 = paddle._C_ops.matmul(layer_norm_102, parameter_109, False, False) @@ -3302,10 +3207,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_157 = paddle._C_ops.add(matmul_138, parameter_108) - del parameter_108 + del matmul_138, parameter_108 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_69 = paddle._C_ops.reshape(add_156, full_int_array_1) + del add_156 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_69 = paddle._C_ops.transpose(reshape_69, [0, 2, 1, 3]) @@ -3313,6 +3219,7 @@ def forward( # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_70 = paddle._C_ops.reshape(add_157, full_int_array_1) + del add_157 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_70 = paddle._C_ops.transpose(reshape_70, [0, 2, 1, 3]) @@ -3324,9 +3231,11 @@ def forward( # pd_op.matmul: (1x16x21x21xf32) <- (1x16x21x64xf32, 1x16x21x64xf32) matmul_139 = paddle._C_ops.matmul(scale_18, transpose_69, False, True) + del scale_18, transpose_69 # pd_op.add: (1x16x21x21xf32) <- (1x16x21x21xf32, 1x1x1x21xf32) add_158 = paddle._C_ops.add(matmul_139, unsqueeze_0) + del matmul_139 # pd_op.softmax: (1x16x21x21xf32) <- (1x16x21x21xf32) softmax_17 = paddle._C_ops.softmax(add_158, -1) @@ -3335,13 +3244,15 @@ def forward( # pd_op.dropout: (1x16x21x21xf32, 1x16x21x21xui8) <- (1x16x21x21xf32, None, 1xf32) dropout_104, dropout_105 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_17, None, full_4, False, "upscale_in_train", 0, False + softmax_17, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_17 # pd_op.matmul: (1x16x21x64xf32) <- (1x16x21x21xf32, 1x16x21x64xf32) matmul_140 = paddle._C_ops.matmul(dropout_104, transpose_70, False, False) + del dropout_104, transpose_70 # pd_op.transpose: (1x21x16x64xf32) <- (1x16x21x64xf32) transpose_71 = paddle._C_ops.transpose(matmul_140, [0, 2, 1, 3]) @@ -3349,19 +3260,20 @@ def forward( # pd_op.reshape: (1x21x1024xf32) <- (1x21x16x64xf32, 3xi64) reshape_71 = paddle._C_ops.reshape(transpose_71, full_int_array_2) + del transpose_71 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_141 = paddle._C_ops.matmul(reshape_71, parameter_107, False, False) - del parameter_107 + del parameter_107, reshape_71 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_159 = paddle._C_ops.add(matmul_141, parameter_106) - del parameter_106 + del matmul_141, parameter_106 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_106, dropout_107 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_159, None, full_4, False, "upscale_in_train", 0, False + add_159, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -3369,6 +3281,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_160 = paddle._C_ops.add(layer_norm_102, dropout_106) + del dropout_106, layer_norm_102 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_105, layer_norm_106, layer_norm_107 = (lambda x, f: f(x))( @@ -3377,7 +3290,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_100, parameter_101 + del add_160, parameter_100, parameter_101 # pd_op.matmul: (1x21x4096xf32) <- (1x21x1024xf32, 1024x4096xf32) matmul_142 = paddle._C_ops.matmul(layer_norm_105, parameter_105, False, False) @@ -3385,23 +3298,24 @@ def forward( # pd_op.add: (1x21x4096xf32) <- (1x21x4096xf32, 4096xf32) add_161 = paddle._C_ops.add(matmul_142, parameter_104) - del parameter_104 + del matmul_142, parameter_104 # pd_op.gelu: (1x21x4096xf32) <- (1x21x4096xf32) gelu_17 = paddle._C_ops.gelu(add_161, False) + del add_161 # pd_op.matmul: (1x21x1024xf32) <- (1x21x4096xf32, 4096x1024xf32) matmul_143 = paddle._C_ops.matmul(gelu_17, parameter_103, False, False) - del parameter_103 + del gelu_17, parameter_103 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_162 = paddle._C_ops.add(matmul_143, parameter_102) - del parameter_102 + del matmul_143, parameter_102 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_108, dropout_109 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_162, None, full_4, False, "upscale_in_train", 0, False + add_162, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -3409,6 +3323,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_163 = paddle._C_ops.add(layer_norm_105, dropout_108) + del dropout_108, layer_norm_105 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_108, layer_norm_109, layer_norm_110 = (lambda x, f: f(x))( @@ -3417,7 +3332,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_98, parameter_99 + del add_163, parameter_98, parameter_99 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_144 = paddle._C_ops.matmul(layer_norm_108, parameter_97, False, False) @@ -3425,10 +3340,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_164 = paddle._C_ops.add(matmul_144, parameter_96) - del parameter_96 + del matmul_144, parameter_96 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_72 = paddle._C_ops.reshape(add_164, full_int_array_1) + del add_164 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_72 = paddle._C_ops.transpose(reshape_72, [0, 2, 1, 3]) @@ -3440,7 +3356,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_165 = paddle._C_ops.add(matmul_145, parameter_94) - del parameter_94 + del matmul_145, parameter_94 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_146 = paddle._C_ops.matmul(layer_norm_108, parameter_93, False, False) @@ -3448,10 +3364,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_166 = paddle._C_ops.add(matmul_146, parameter_92) - del parameter_92 + del matmul_146, parameter_92 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_73 = paddle._C_ops.reshape(add_165, full_int_array_1) + del add_165 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_73 = paddle._C_ops.transpose(reshape_73, [0, 2, 1, 3]) @@ -3459,6 +3376,7 @@ def forward( # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_74 = paddle._C_ops.reshape(add_166, full_int_array_1) + del add_166 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_74 = paddle._C_ops.transpose(reshape_74, [0, 2, 1, 3]) @@ -3470,9 +3388,11 @@ def forward( # pd_op.matmul: (1x16x21x21xf32) <- (1x16x21x64xf32, 1x16x21x64xf32) matmul_147 = paddle._C_ops.matmul(scale_19, transpose_73, False, True) + del scale_19, transpose_73 # pd_op.add: (1x16x21x21xf32) <- (1x16x21x21xf32, 1x1x1x21xf32) add_167 = paddle._C_ops.add(matmul_147, unsqueeze_0) + del matmul_147 # pd_op.softmax: (1x16x21x21xf32) <- (1x16x21x21xf32) softmax_18 = paddle._C_ops.softmax(add_167, -1) @@ -3481,13 +3401,15 @@ def forward( # pd_op.dropout: (1x16x21x21xf32, 1x16x21x21xui8) <- (1x16x21x21xf32, None, 1xf32) dropout_110, dropout_111 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_18, None, full_4, False, "upscale_in_train", 0, False + softmax_18, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_18 # pd_op.matmul: (1x16x21x64xf32) <- (1x16x21x21xf32, 1x16x21x64xf32) matmul_148 = paddle._C_ops.matmul(dropout_110, transpose_74, False, False) + del dropout_110, transpose_74 # pd_op.transpose: (1x21x16x64xf32) <- (1x16x21x64xf32) transpose_75 = paddle._C_ops.transpose(matmul_148, [0, 2, 1, 3]) @@ -3495,19 +3417,20 @@ def forward( # pd_op.reshape: (1x21x1024xf32) <- (1x21x16x64xf32, 3xi64) reshape_75 = paddle._C_ops.reshape(transpose_75, full_int_array_2) + del transpose_75 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_149 = paddle._C_ops.matmul(reshape_75, parameter_91, False, False) - del parameter_91 + del parameter_91, reshape_75 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_168 = paddle._C_ops.add(matmul_149, parameter_90) - del parameter_90 + del matmul_149, parameter_90 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_112, dropout_113 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_168, None, full_4, False, "upscale_in_train", 0, False + add_168, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -3515,6 +3438,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_169 = paddle._C_ops.add(layer_norm_108, dropout_112) + del dropout_112, layer_norm_108 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_111, layer_norm_112, layer_norm_113 = (lambda x, f: f(x))( @@ -3523,7 +3447,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_84, parameter_85 + del add_169, parameter_84, parameter_85 # pd_op.matmul: (1x21x4096xf32) <- (1x21x1024xf32, 1024x4096xf32) matmul_150 = paddle._C_ops.matmul(layer_norm_111, parameter_89, False, False) @@ -3531,23 +3455,24 @@ def forward( # pd_op.add: (1x21x4096xf32) <- (1x21x4096xf32, 4096xf32) add_170 = paddle._C_ops.add(matmul_150, parameter_88) - del parameter_88 + del matmul_150, parameter_88 # pd_op.gelu: (1x21x4096xf32) <- (1x21x4096xf32) gelu_18 = paddle._C_ops.gelu(add_170, False) + del add_170 # pd_op.matmul: (1x21x1024xf32) <- (1x21x4096xf32, 4096x1024xf32) matmul_151 = paddle._C_ops.matmul(gelu_18, parameter_87, False, False) - del parameter_87 + del gelu_18, parameter_87 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_171 = paddle._C_ops.add(matmul_151, parameter_86) - del parameter_86 + del matmul_151, parameter_86 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_114, dropout_115 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_171, None, full_4, False, "upscale_in_train", 0, False + add_171, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -3555,6 +3480,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_172 = paddle._C_ops.add(layer_norm_111, dropout_114) + del dropout_114, layer_norm_111 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_114, layer_norm_115, layer_norm_116 = (lambda x, f: f(x))( @@ -3563,7 +3489,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_82, parameter_83 + del add_172, parameter_82, parameter_83 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_152 = paddle._C_ops.matmul(layer_norm_114, parameter_81, False, False) @@ -3571,10 +3497,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_173 = paddle._C_ops.add(matmul_152, parameter_80) - del parameter_80 + del matmul_152, parameter_80 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_76 = paddle._C_ops.reshape(add_173, full_int_array_1) + del add_173 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_76 = paddle._C_ops.transpose(reshape_76, [0, 2, 1, 3]) @@ -3586,7 +3513,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_174 = paddle._C_ops.add(matmul_153, parameter_78) - del parameter_78 + del matmul_153, parameter_78 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_154 = paddle._C_ops.matmul(layer_norm_114, parameter_77, False, False) @@ -3594,10 +3521,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_175 = paddle._C_ops.add(matmul_154, parameter_76) - del parameter_76 + del matmul_154, parameter_76 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_77 = paddle._C_ops.reshape(add_174, full_int_array_1) + del add_174 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_77 = paddle._C_ops.transpose(reshape_77, [0, 2, 1, 3]) @@ -3605,6 +3533,7 @@ def forward( # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_78 = paddle._C_ops.reshape(add_175, full_int_array_1) + del add_175 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_78 = paddle._C_ops.transpose(reshape_78, [0, 2, 1, 3]) @@ -3616,9 +3545,11 @@ def forward( # pd_op.matmul: (1x16x21x21xf32) <- (1x16x21x64xf32, 1x16x21x64xf32) matmul_155 = paddle._C_ops.matmul(scale_20, transpose_77, False, True) + del scale_20, transpose_77 # pd_op.add: (1x16x21x21xf32) <- (1x16x21x21xf32, 1x1x1x21xf32) add_176 = paddle._C_ops.add(matmul_155, unsqueeze_0) + del matmul_155 # pd_op.softmax: (1x16x21x21xf32) <- (1x16x21x21xf32) softmax_19 = paddle._C_ops.softmax(add_176, -1) @@ -3627,13 +3558,15 @@ def forward( # pd_op.dropout: (1x16x21x21xf32, 1x16x21x21xui8) <- (1x16x21x21xf32, None, 1xf32) dropout_116, dropout_117 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_19, None, full_4, False, "upscale_in_train", 0, False + softmax_19, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_19 # pd_op.matmul: (1x16x21x64xf32) <- (1x16x21x21xf32, 1x16x21x64xf32) matmul_156 = paddle._C_ops.matmul(dropout_116, transpose_78, False, False) + del dropout_116, transpose_78 # pd_op.transpose: (1x21x16x64xf32) <- (1x16x21x64xf32) transpose_79 = paddle._C_ops.transpose(matmul_156, [0, 2, 1, 3]) @@ -3641,19 +3574,20 @@ def forward( # pd_op.reshape: (1x21x1024xf32) <- (1x21x16x64xf32, 3xi64) reshape_79 = paddle._C_ops.reshape(transpose_79, full_int_array_2) + del transpose_79 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_157 = paddle._C_ops.matmul(reshape_79, parameter_75, False, False) - del parameter_75 + del parameter_75, reshape_79 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_177 = paddle._C_ops.add(matmul_157, parameter_74) - del parameter_74 + del matmul_157, parameter_74 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_118, dropout_119 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_177, None, full_4, False, "upscale_in_train", 0, False + add_177, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -3661,6 +3595,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_178 = paddle._C_ops.add(layer_norm_114, dropout_118) + del dropout_118, layer_norm_114 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_117, layer_norm_118, layer_norm_119 = (lambda x, f: f(x))( @@ -3669,7 +3604,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_68, parameter_69 + del add_178, parameter_68, parameter_69 # pd_op.matmul: (1x21x4096xf32) <- (1x21x1024xf32, 1024x4096xf32) matmul_158 = paddle._C_ops.matmul(layer_norm_117, parameter_73, False, False) @@ -3677,23 +3612,24 @@ def forward( # pd_op.add: (1x21x4096xf32) <- (1x21x4096xf32, 4096xf32) add_179 = paddle._C_ops.add(matmul_158, parameter_72) - del parameter_72 + del matmul_158, parameter_72 # pd_op.gelu: (1x21x4096xf32) <- (1x21x4096xf32) gelu_19 = paddle._C_ops.gelu(add_179, False) + del add_179 # pd_op.matmul: (1x21x1024xf32) <- (1x21x4096xf32, 4096x1024xf32) matmul_159 = paddle._C_ops.matmul(gelu_19, parameter_71, False, False) - del parameter_71 + del gelu_19, parameter_71 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_180 = paddle._C_ops.add(matmul_159, parameter_70) - del parameter_70 + del matmul_159, parameter_70 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_120, dropout_121 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_180, None, full_4, False, "upscale_in_train", 0, False + add_180, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -3701,6 +3637,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_181 = paddle._C_ops.add(layer_norm_117, dropout_120) + del dropout_120, layer_norm_117 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_120, layer_norm_121, layer_norm_122 = (lambda x, f: f(x))( @@ -3709,7 +3646,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_66, parameter_67 + del add_181, parameter_66, parameter_67 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_160 = paddle._C_ops.matmul(layer_norm_120, parameter_65, False, False) @@ -3717,10 +3654,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_182 = paddle._C_ops.add(matmul_160, parameter_64) - del parameter_64 + del matmul_160, parameter_64 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_80 = paddle._C_ops.reshape(add_182, full_int_array_1) + del add_182 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_80 = paddle._C_ops.transpose(reshape_80, [0, 2, 1, 3]) @@ -3732,7 +3670,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_183 = paddle._C_ops.add(matmul_161, parameter_62) - del parameter_62 + del matmul_161, parameter_62 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_162 = paddle._C_ops.matmul(layer_norm_120, parameter_61, False, False) @@ -3740,10 +3678,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_184 = paddle._C_ops.add(matmul_162, parameter_60) - del parameter_60 + del matmul_162, parameter_60 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_81 = paddle._C_ops.reshape(add_183, full_int_array_1) + del add_183 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_81 = paddle._C_ops.transpose(reshape_81, [0, 2, 1, 3]) @@ -3751,6 +3690,7 @@ def forward( # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_82 = paddle._C_ops.reshape(add_184, full_int_array_1) + del add_184 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_82 = paddle._C_ops.transpose(reshape_82, [0, 2, 1, 3]) @@ -3762,9 +3702,11 @@ def forward( # pd_op.matmul: (1x16x21x21xf32) <- (1x16x21x64xf32, 1x16x21x64xf32) matmul_163 = paddle._C_ops.matmul(scale_21, transpose_81, False, True) + del scale_21, transpose_81 # pd_op.add: (1x16x21x21xf32) <- (1x16x21x21xf32, 1x1x1x21xf32) add_185 = paddle._C_ops.add(matmul_163, unsqueeze_0) + del matmul_163 # pd_op.softmax: (1x16x21x21xf32) <- (1x16x21x21xf32) softmax_20 = paddle._C_ops.softmax(add_185, -1) @@ -3773,13 +3715,15 @@ def forward( # pd_op.dropout: (1x16x21x21xf32, 1x16x21x21xui8) <- (1x16x21x21xf32, None, 1xf32) dropout_122, dropout_123 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_20, None, full_4, False, "upscale_in_train", 0, False + softmax_20, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_20 # pd_op.matmul: (1x16x21x64xf32) <- (1x16x21x21xf32, 1x16x21x64xf32) matmul_164 = paddle._C_ops.matmul(dropout_122, transpose_82, False, False) + del dropout_122, transpose_82 # pd_op.transpose: (1x21x16x64xf32) <- (1x16x21x64xf32) transpose_83 = paddle._C_ops.transpose(matmul_164, [0, 2, 1, 3]) @@ -3787,19 +3731,20 @@ def forward( # pd_op.reshape: (1x21x1024xf32) <- (1x21x16x64xf32, 3xi64) reshape_83 = paddle._C_ops.reshape(transpose_83, full_int_array_2) + del transpose_83 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_165 = paddle._C_ops.matmul(reshape_83, parameter_59, False, False) - del parameter_59 + del parameter_59, reshape_83 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_186 = paddle._C_ops.add(matmul_165, parameter_58) - del parameter_58 + del matmul_165, parameter_58 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_124, dropout_125 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_186, None, full_4, False, "upscale_in_train", 0, False + add_186, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -3807,6 +3752,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_187 = paddle._C_ops.add(layer_norm_120, dropout_124) + del dropout_124, layer_norm_120 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_123, layer_norm_124, layer_norm_125 = (lambda x, f: f(x))( @@ -3815,7 +3761,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_52, parameter_53 + del add_187, parameter_52, parameter_53 # pd_op.matmul: (1x21x4096xf32) <- (1x21x1024xf32, 1024x4096xf32) matmul_166 = paddle._C_ops.matmul(layer_norm_123, parameter_57, False, False) @@ -3823,23 +3769,24 @@ def forward( # pd_op.add: (1x21x4096xf32) <- (1x21x4096xf32, 4096xf32) add_188 = paddle._C_ops.add(matmul_166, parameter_56) - del parameter_56 + del matmul_166, parameter_56 # pd_op.gelu: (1x21x4096xf32) <- (1x21x4096xf32) gelu_20 = paddle._C_ops.gelu(add_188, False) + del add_188 # pd_op.matmul: (1x21x1024xf32) <- (1x21x4096xf32, 4096x1024xf32) matmul_167 = paddle._C_ops.matmul(gelu_20, parameter_55, False, False) - del parameter_55 + del gelu_20, parameter_55 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_189 = paddle._C_ops.add(matmul_167, parameter_54) - del parameter_54 + del matmul_167, parameter_54 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_126, dropout_127 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_189, None, full_4, False, "upscale_in_train", 0, False + add_189, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -3847,6 +3794,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_190 = paddle._C_ops.add(layer_norm_123, dropout_126) + del dropout_126, layer_norm_123 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_126, layer_norm_127, layer_norm_128 = (lambda x, f: f(x))( @@ -3855,7 +3803,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_50, parameter_51 + del add_190, parameter_50, parameter_51 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_168 = paddle._C_ops.matmul(layer_norm_126, parameter_49, False, False) @@ -3863,10 +3811,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_191 = paddle._C_ops.add(matmul_168, parameter_48) - del parameter_48 + del matmul_168, parameter_48 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_84 = paddle._C_ops.reshape(add_191, full_int_array_1) + del add_191 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_84 = paddle._C_ops.transpose(reshape_84, [0, 2, 1, 3]) @@ -3878,7 +3827,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_192 = paddle._C_ops.add(matmul_169, parameter_46) - del parameter_46 + del matmul_169, parameter_46 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_170 = paddle._C_ops.matmul(layer_norm_126, parameter_45, False, False) @@ -3886,10 +3835,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_193 = paddle._C_ops.add(matmul_170, parameter_44) - del parameter_44 + del matmul_170, parameter_44 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_85 = paddle._C_ops.reshape(add_192, full_int_array_1) + del add_192 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_85 = paddle._C_ops.transpose(reshape_85, [0, 2, 1, 3]) @@ -3897,6 +3847,7 @@ def forward( # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_86 = paddle._C_ops.reshape(add_193, full_int_array_1) + del add_193 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_86 = paddle._C_ops.transpose(reshape_86, [0, 2, 1, 3]) @@ -3908,9 +3859,11 @@ def forward( # pd_op.matmul: (1x16x21x21xf32) <- (1x16x21x64xf32, 1x16x21x64xf32) matmul_171 = paddle._C_ops.matmul(scale_22, transpose_85, False, True) + del scale_22, transpose_85 # pd_op.add: (1x16x21x21xf32) <- (1x16x21x21xf32, 1x1x1x21xf32) add_194 = paddle._C_ops.add(matmul_171, unsqueeze_0) + del matmul_171 # pd_op.softmax: (1x16x21x21xf32) <- (1x16x21x21xf32) softmax_21 = paddle._C_ops.softmax(add_194, -1) @@ -3919,13 +3872,15 @@ def forward( # pd_op.dropout: (1x16x21x21xf32, 1x16x21x21xui8) <- (1x16x21x21xf32, None, 1xf32) dropout_128, dropout_129 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_21, None, full_4, False, "upscale_in_train", 0, False + softmax_21, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_21 # pd_op.matmul: (1x16x21x64xf32) <- (1x16x21x21xf32, 1x16x21x64xf32) matmul_172 = paddle._C_ops.matmul(dropout_128, transpose_86, False, False) + del dropout_128, transpose_86 # pd_op.transpose: (1x21x16x64xf32) <- (1x16x21x64xf32) transpose_87 = paddle._C_ops.transpose(matmul_172, [0, 2, 1, 3]) @@ -3933,19 +3888,20 @@ def forward( # pd_op.reshape: (1x21x1024xf32) <- (1x21x16x64xf32, 3xi64) reshape_87 = paddle._C_ops.reshape(transpose_87, full_int_array_2) + del transpose_87 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_173 = paddle._C_ops.matmul(reshape_87, parameter_43, False, False) - del parameter_43 + del parameter_43, reshape_87 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_195 = paddle._C_ops.add(matmul_173, parameter_42) - del parameter_42 + del matmul_173, parameter_42 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_130, dropout_131 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_195, None, full_4, False, "upscale_in_train", 0, False + add_195, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -3953,6 +3909,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_196 = paddle._C_ops.add(layer_norm_126, dropout_130) + del dropout_130, layer_norm_126 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_129, layer_norm_130, layer_norm_131 = (lambda x, f: f(x))( @@ -3961,7 +3918,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_36, parameter_37 + del add_196, parameter_36, parameter_37 # pd_op.matmul: (1x21x4096xf32) <- (1x21x1024xf32, 1024x4096xf32) matmul_174 = paddle._C_ops.matmul(layer_norm_129, parameter_41, False, False) @@ -3969,23 +3926,24 @@ def forward( # pd_op.add: (1x21x4096xf32) <- (1x21x4096xf32, 4096xf32) add_197 = paddle._C_ops.add(matmul_174, parameter_40) - del parameter_40 + del matmul_174, parameter_40 # pd_op.gelu: (1x21x4096xf32) <- (1x21x4096xf32) gelu_21 = paddle._C_ops.gelu(add_197, False) + del add_197 # pd_op.matmul: (1x21x1024xf32) <- (1x21x4096xf32, 4096x1024xf32) matmul_175 = paddle._C_ops.matmul(gelu_21, parameter_39, False, False) - del parameter_39 + del gelu_21, parameter_39 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_198 = paddle._C_ops.add(matmul_175, parameter_38) - del parameter_38 + del matmul_175, parameter_38 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_132, dropout_133 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_198, None, full_4, False, "upscale_in_train", 0, False + add_198, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -3993,6 +3951,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_199 = paddle._C_ops.add(layer_norm_129, dropout_132) + del dropout_132, layer_norm_129 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_132, layer_norm_133, layer_norm_134 = (lambda x, f: f(x))( @@ -4001,7 +3960,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_34, parameter_35 + del add_199, parameter_34, parameter_35 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_176 = paddle._C_ops.matmul(layer_norm_132, parameter_33, False, False) @@ -4009,10 +3968,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_200 = paddle._C_ops.add(matmul_176, parameter_32) - del parameter_32 + del matmul_176, parameter_32 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_88 = paddle._C_ops.reshape(add_200, full_int_array_1) + del add_200 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_88 = paddle._C_ops.transpose(reshape_88, [0, 2, 1, 3]) @@ -4024,7 +3984,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_201 = paddle._C_ops.add(matmul_177, parameter_30) - del parameter_30 + del matmul_177, parameter_30 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_178 = paddle._C_ops.matmul(layer_norm_132, parameter_29, False, False) @@ -4032,10 +3992,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_202 = paddle._C_ops.add(matmul_178, parameter_28) - del parameter_28 + del matmul_178, parameter_28 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_89 = paddle._C_ops.reshape(add_201, full_int_array_1) + del add_201 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_89 = paddle._C_ops.transpose(reshape_89, [0, 2, 1, 3]) @@ -4043,6 +4004,7 @@ def forward( # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_90 = paddle._C_ops.reshape(add_202, full_int_array_1) + del add_202 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_90 = paddle._C_ops.transpose(reshape_90, [0, 2, 1, 3]) @@ -4054,9 +4016,11 @@ def forward( # pd_op.matmul: (1x16x21x21xf32) <- (1x16x21x64xf32, 1x16x21x64xf32) matmul_179 = paddle._C_ops.matmul(scale_23, transpose_89, False, True) + del scale_23, transpose_89 # pd_op.add: (1x16x21x21xf32) <- (1x16x21x21xf32, 1x1x1x21xf32) add_203 = paddle._C_ops.add(matmul_179, unsqueeze_0) + del matmul_179 # pd_op.softmax: (1x16x21x21xf32) <- (1x16x21x21xf32) softmax_22 = paddle._C_ops.softmax(add_203, -1) @@ -4065,13 +4029,15 @@ def forward( # pd_op.dropout: (1x16x21x21xf32, 1x16x21x21xui8) <- (1x16x21x21xf32, None, 1xf32) dropout_134, dropout_135 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_22, None, full_4, False, "upscale_in_train", 0, False + softmax_22, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_22 # pd_op.matmul: (1x16x21x64xf32) <- (1x16x21x21xf32, 1x16x21x64xf32) matmul_180 = paddle._C_ops.matmul(dropout_134, transpose_90, False, False) + del dropout_134, transpose_90 # pd_op.transpose: (1x21x16x64xf32) <- (1x16x21x64xf32) transpose_91 = paddle._C_ops.transpose(matmul_180, [0, 2, 1, 3]) @@ -4079,19 +4045,20 @@ def forward( # pd_op.reshape: (1x21x1024xf32) <- (1x21x16x64xf32, 3xi64) reshape_91 = paddle._C_ops.reshape(transpose_91, full_int_array_2) + del transpose_91 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_181 = paddle._C_ops.matmul(reshape_91, parameter_27, False, False) - del parameter_27 + del parameter_27, reshape_91 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_204 = paddle._C_ops.add(matmul_181, parameter_26) - del parameter_26 + del matmul_181, parameter_26 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_136, dropout_137 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_204, None, full_4, False, "upscale_in_train", 0, False + add_204, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -4099,6 +4066,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_205 = paddle._C_ops.add(layer_norm_132, dropout_136) + del dropout_136, layer_norm_132 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_135, layer_norm_136, layer_norm_137 = (lambda x, f: f(x))( @@ -4107,7 +4075,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_20, parameter_21 + del add_205, parameter_20, parameter_21 # pd_op.matmul: (1x21x4096xf32) <- (1x21x1024xf32, 1024x4096xf32) matmul_182 = paddle._C_ops.matmul(layer_norm_135, parameter_25, False, False) @@ -4115,23 +4083,24 @@ def forward( # pd_op.add: (1x21x4096xf32) <- (1x21x4096xf32, 4096xf32) add_206 = paddle._C_ops.add(matmul_182, parameter_24) - del parameter_24 + del matmul_182, parameter_24 # pd_op.gelu: (1x21x4096xf32) <- (1x21x4096xf32) gelu_22 = paddle._C_ops.gelu(add_206, False) + del add_206 # pd_op.matmul: (1x21x1024xf32) <- (1x21x4096xf32, 4096x1024xf32) matmul_183 = paddle._C_ops.matmul(gelu_22, parameter_23, False, False) - del parameter_23 + del gelu_22, parameter_23 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_207 = paddle._C_ops.add(matmul_183, parameter_22) - del parameter_22 + del matmul_183, parameter_22 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_138, dropout_139 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_207, None, full_4, False, "upscale_in_train", 0, False + add_207, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -4139,6 +4108,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_208 = paddle._C_ops.add(layer_norm_135, dropout_138) + del dropout_138, layer_norm_135 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_138, layer_norm_139, layer_norm_140 = (lambda x, f: f(x))( @@ -4147,7 +4117,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_18, parameter_19 + del add_208, parameter_18, parameter_19 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_184 = paddle._C_ops.matmul(layer_norm_138, parameter_17, False, False) @@ -4155,10 +4125,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_209 = paddle._C_ops.add(matmul_184, parameter_16) - del parameter_16 + del matmul_184, parameter_16 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_92 = paddle._C_ops.reshape(add_209, full_int_array_1) + del add_209 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_92 = paddle._C_ops.transpose(reshape_92, [0, 2, 1, 3]) @@ -4170,7 +4141,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_210 = paddle._C_ops.add(matmul_185, parameter_14) - del parameter_14 + del matmul_185, parameter_14 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_186 = paddle._C_ops.matmul(layer_norm_138, parameter_13, False, False) @@ -4178,10 +4149,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_211 = paddle._C_ops.add(matmul_186, parameter_12) - del parameter_12 + del matmul_186, parameter_12 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_93 = paddle._C_ops.reshape(add_210, full_int_array_1) + del add_210 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_93 = paddle._C_ops.transpose(reshape_93, [0, 2, 1, 3]) @@ -4189,7 +4161,7 @@ def forward( # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_94 = paddle._C_ops.reshape(add_211, full_int_array_1) - del full_int_array_1 + del add_211, full_int_array_1 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_94 = paddle._C_ops.transpose(reshape_94, [0, 2, 1, 3]) @@ -4197,13 +4169,15 @@ def forward( # pd_op.scale: (1x16x21x64xf32) <- (1x16x21x64xf32, 1xf32) scale_24 = paddle._C_ops.scale(transpose_92, full_5, float("0"), True) - del transpose_92 + del full_5, transpose_92 # pd_op.matmul: (1x16x21x21xf32) <- (1x16x21x64xf32, 1x16x21x64xf32) matmul_187 = paddle._C_ops.matmul(scale_24, transpose_93, False, True) + del scale_24, transpose_93 # pd_op.add: (1x16x21x21xf32) <- (1x16x21x21xf32, 1x1x1x21xf32) add_212 = paddle._C_ops.add(matmul_187, unsqueeze_0) + del matmul_187, unsqueeze_0 # pd_op.softmax: (1x16x21x21xf32) <- (1x16x21x21xf32) softmax_23 = paddle._C_ops.softmax(add_212, -1) @@ -4212,13 +4186,15 @@ def forward( # pd_op.dropout: (1x16x21x21xf32, 1x16x21x21xui8) <- (1x16x21x21xf32, None, 1xf32) dropout_140, dropout_141 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_23, None, full_4, False, "upscale_in_train", 0, False + softmax_23, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_23 # pd_op.matmul: (1x16x21x64xf32) <- (1x16x21x21xf32, 1x16x21x64xf32) matmul_188 = paddle._C_ops.matmul(dropout_140, transpose_94, False, False) + del dropout_140, transpose_94 # pd_op.transpose: (1x21x16x64xf32) <- (1x16x21x64xf32) transpose_95 = paddle._C_ops.transpose(matmul_188, [0, 2, 1, 3]) @@ -4226,20 +4202,20 @@ def forward( # pd_op.reshape: (1x21x1024xf32) <- (1x21x16x64xf32, 3xi64) reshape_95 = paddle._C_ops.reshape(transpose_95, full_int_array_2) - del full_int_array_2 + del full_int_array_2, transpose_95 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_189 = paddle._C_ops.matmul(reshape_95, parameter_11, False, False) - del parameter_11 + del parameter_11, reshape_95 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_213 = paddle._C_ops.add(matmul_189, parameter_10) - del parameter_10 + del matmul_189, parameter_10 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_142, dropout_143 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_213, None, full_4, False, "upscale_in_train", 0, False + add_213, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -4247,6 +4223,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_214 = paddle._C_ops.add(layer_norm_138, dropout_142) + del dropout_142, layer_norm_138 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_141, layer_norm_142, layer_norm_143 = (lambda x, f: f(x))( @@ -4255,7 +4232,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_4, parameter_5 + del add_214, parameter_4, parameter_5 # pd_op.matmul: (1x21x4096xf32) <- (1x21x1024xf32, 1024x4096xf32) matmul_190 = paddle._C_ops.matmul(layer_norm_141, parameter_9, False, False) @@ -4263,30 +4240,32 @@ def forward( # pd_op.add: (1x21x4096xf32) <- (1x21x4096xf32, 4096xf32) add_215 = paddle._C_ops.add(matmul_190, parameter_8) - del parameter_8 + del matmul_190, parameter_8 # pd_op.gelu: (1x21x4096xf32) <- (1x21x4096xf32) gelu_23 = paddle._C_ops.gelu(add_215, False) + del add_215 # pd_op.matmul: (1x21x1024xf32) <- (1x21x4096xf32, 4096x1024xf32) matmul_191 = paddle._C_ops.matmul(gelu_23, parameter_7, False, False) - del parameter_7 + del gelu_23, parameter_7 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_216 = paddle._C_ops.add(matmul_191, parameter_6) - del parameter_6 + del matmul_191, parameter_6 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_144, dropout_145 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_216, None, full_4, False, "upscale_in_train", 0, False + add_216, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) - del add_216 + del add_216, full_4 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_217 = paddle._C_ops.add(layer_norm_141, dropout_144) + del dropout_144, layer_norm_141 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_144, layer_norm_145, layer_norm_146 = (lambda x, f: f(x))( @@ -4295,7 +4274,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_2, parameter_3 + del add_217, parameter_2, parameter_3 # pd_op.full_int_array: (1xi64) <- () full_int_array_3 = [0] @@ -4307,899 +4286,18 @@ def forward( slice_0 = paddle._C_ops.slice( layer_norm_144, [1], full_int_array_3, full_int_array_4, [1], [1] ) + del full_int_array_3, full_int_array_4, layer_norm_144 # pd_op.matmul: (1x1024xf32) <- (1x1024xf32, 1024x1024xf32) matmul_192 = paddle._C_ops.matmul(slice_0, parameter_1, False, False) - del parameter_1 + del parameter_1, slice_0 # pd_op.add: (1x1024xf32) <- (1x1024xf32, 1024xf32) add_218 = paddle._C_ops.add(matmul_192, parameter_0) - del parameter_0 + del matmul_192, parameter_0 # pd_op.tanh: (1x1024xf32) <- (1x1024xf32) tanh_0 = paddle._C_ops.tanh(add_218) - del ( - add_0, - add_1, - add_10, - add_100, - add_101, - add_102, - add_103, - add_106, - add_107, - add_109, - add_11, - add_110, - add_111, - add_112, - add_115, - add_116, - add_118, - add_119, - add_12, - add_120, - add_121, - add_124, - add_125, - add_127, - add_128, - add_129, - add_13, - add_130, - add_133, - add_134, - add_136, - add_137, - add_138, - add_139, - add_142, - add_143, - add_145, - add_146, - add_147, - add_148, - add_151, - add_152, - add_154, - add_155, - add_156, - add_157, - add_16, - add_160, - add_161, - add_163, - add_164, - add_165, - add_166, - add_169, - add_17, - add_170, - add_172, - add_173, - add_174, - add_175, - add_178, - add_179, - add_181, - add_182, - add_183, - add_184, - add_187, - add_188, - add_19, - add_190, - add_191, - add_192, - add_193, - add_196, - add_197, - add_199, - add_2, - add_20, - add_200, - add_201, - add_202, - add_205, - add_206, - add_208, - add_209, - add_21, - add_210, - add_211, - add_214, - add_215, - add_217, - add_218, - add_22, - add_25, - add_26, - add_28, - add_29, - add_3, - add_30, - add_31, - add_34, - add_35, - add_37, - add_38, - add_39, - add_4, - add_40, - add_43, - add_44, - add_46, - add_47, - add_48, - add_49, - add_52, - add_53, - add_55, - add_56, - add_57, - add_58, - add_61, - add_62, - add_64, - add_65, - add_66, - add_67, - add_7, - add_70, - add_71, - add_73, - add_74, - add_75, - add_76, - add_79, - add_8, - add_80, - add_82, - add_83, - add_84, - add_85, - add_88, - add_89, - add_91, - add_92, - add_93, - add_94, - add_97, - add_98, - assign_0, - assign_1, - assign_10, - assign_11, - assign_12, - assign_13, - assign_14, - assign_15, - assign_16, - assign_17, - assign_18, - assign_19, - assign_2, - assign_20, - assign_21, - assign_22, - assign_23, - assign_24, - assign_25, - assign_26, - assign_27, - assign_28, - assign_29, - assign_3, - assign_30, - assign_31, - assign_32, - assign_33, - assign_34, - assign_35, - assign_36, - assign_37, - assign_38, - assign_39, - assign_4, - assign_40, - assign_41, - assign_42, - assign_43, - assign_44, - assign_45, - assign_46, - assign_47, - assign_48, - assign_49, - assign_5, - assign_50, - assign_51, - assign_52, - assign_53, - assign_54, - assign_55, - assign_56, - assign_57, - assign_58, - assign_59, - assign_6, - assign_60, - assign_61, - assign_62, - assign_63, - assign_64, - assign_65, - assign_66, - assign_67, - assign_68, - assign_69, - assign_7, - assign_70, - assign_71, - assign_72, - assign_73, - assign_74, - assign_75, - assign_76, - assign_77, - assign_78, - assign_79, - assign_8, - assign_80, - assign_81, - assign_82, - assign_83, - assign_84, - assign_85, - assign_86, - assign_87, - assign_88, - assign_89, - assign_9, - assign_90, - assign_91, - assign_92, - assign_93, - assign_94, - dropout_0, - dropout_1, - dropout_10, - dropout_100, - dropout_101, - dropout_102, - dropout_103, - dropout_104, - dropout_105, - dropout_106, - dropout_107, - dropout_108, - dropout_109, - dropout_11, - dropout_110, - dropout_111, - dropout_112, - dropout_113, - dropout_114, - dropout_115, - dropout_116, - dropout_117, - dropout_118, - dropout_119, - dropout_12, - dropout_120, - dropout_121, - dropout_122, - dropout_123, - dropout_124, - dropout_125, - dropout_126, - dropout_127, - dropout_128, - dropout_129, - dropout_13, - dropout_130, - dropout_131, - dropout_132, - dropout_133, - dropout_134, - dropout_135, - dropout_136, - dropout_137, - dropout_138, - dropout_139, - dropout_14, - dropout_140, - dropout_141, - dropout_142, - dropout_143, - dropout_144, - dropout_145, - dropout_15, - dropout_16, - dropout_17, - dropout_18, - dropout_19, - dropout_2, - dropout_20, - dropout_21, - dropout_22, - dropout_23, - dropout_24, - dropout_25, - dropout_26, - dropout_27, - dropout_28, - dropout_29, - dropout_3, - dropout_30, - dropout_31, - dropout_32, - dropout_33, - dropout_34, - dropout_35, - dropout_36, - dropout_37, - dropout_38, - dropout_39, - dropout_4, - dropout_40, - dropout_41, - dropout_42, - dropout_43, - dropout_44, - dropout_45, - dropout_46, - dropout_47, - dropout_48, - dropout_49, - dropout_5, - dropout_50, - dropout_51, - dropout_52, - dropout_53, - dropout_54, - dropout_55, - dropout_56, - dropout_57, - dropout_58, - dropout_59, - dropout_6, - dropout_60, - dropout_61, - dropout_62, - dropout_63, - dropout_64, - dropout_65, - dropout_66, - dropout_67, - dropout_68, - dropout_69, - dropout_7, - dropout_70, - dropout_71, - dropout_72, - dropout_73, - dropout_74, - dropout_75, - dropout_76, - dropout_77, - dropout_78, - dropout_79, - dropout_8, - dropout_80, - dropout_81, - dropout_82, - dropout_83, - dropout_84, - dropout_85, - dropout_86, - dropout_87, - dropout_88, - dropout_89, - dropout_9, - dropout_90, - dropout_91, - dropout_92, - dropout_93, - dropout_94, - dropout_95, - dropout_96, - dropout_97, - dropout_98, - dropout_99, - embedding_0, - embedding_1, - embedding_2, - full_4, - full_5, - full_int_array_3, - full_int_array_4, - gelu_0, - gelu_1, - gelu_10, - gelu_11, - gelu_12, - gelu_13, - gelu_14, - gelu_15, - gelu_16, - gelu_17, - gelu_18, - gelu_19, - gelu_2, - gelu_20, - gelu_21, - gelu_22, - gelu_23, - gelu_3, - gelu_4, - gelu_5, - gelu_6, - gelu_7, - gelu_8, - gelu_9, - layer_norm_1, - layer_norm_10, - layer_norm_100, - layer_norm_101, - layer_norm_102, - layer_norm_103, - layer_norm_104, - layer_norm_105, - layer_norm_106, - layer_norm_107, - layer_norm_108, - layer_norm_109, - layer_norm_11, - layer_norm_110, - layer_norm_111, - layer_norm_112, - layer_norm_113, - layer_norm_114, - layer_norm_115, - layer_norm_116, - layer_norm_117, - layer_norm_118, - layer_norm_119, - layer_norm_12, - layer_norm_120, - layer_norm_121, - layer_norm_122, - layer_norm_123, - layer_norm_124, - layer_norm_125, - layer_norm_126, - layer_norm_127, - layer_norm_128, - layer_norm_129, - layer_norm_13, - layer_norm_130, - layer_norm_131, - layer_norm_132, - layer_norm_133, - layer_norm_134, - layer_norm_135, - layer_norm_136, - layer_norm_137, - layer_norm_138, - layer_norm_139, - layer_norm_14, - layer_norm_140, - layer_norm_141, - layer_norm_142, - layer_norm_143, - layer_norm_144, - layer_norm_145, - layer_norm_146, - layer_norm_15, - layer_norm_16, - layer_norm_17, - layer_norm_18, - layer_norm_19, - layer_norm_2, - layer_norm_20, - layer_norm_21, - layer_norm_22, - layer_norm_23, - layer_norm_24, - layer_norm_25, - layer_norm_26, - layer_norm_27, - layer_norm_28, - layer_norm_29, - layer_norm_3, - layer_norm_30, - layer_norm_31, - layer_norm_32, - layer_norm_33, - layer_norm_34, - layer_norm_35, - layer_norm_36, - layer_norm_37, - layer_norm_38, - layer_norm_39, - layer_norm_4, - layer_norm_40, - layer_norm_41, - layer_norm_42, - layer_norm_43, - layer_norm_44, - layer_norm_45, - layer_norm_46, - layer_norm_47, - layer_norm_48, - layer_norm_49, - layer_norm_5, - layer_norm_50, - layer_norm_51, - layer_norm_52, - layer_norm_53, - layer_norm_54, - layer_norm_55, - layer_norm_56, - layer_norm_57, - layer_norm_58, - layer_norm_59, - layer_norm_6, - layer_norm_60, - layer_norm_61, - layer_norm_62, - layer_norm_63, - layer_norm_64, - layer_norm_65, - layer_norm_66, - layer_norm_67, - layer_norm_68, - layer_norm_69, - layer_norm_7, - layer_norm_70, - layer_norm_71, - layer_norm_72, - layer_norm_73, - layer_norm_74, - layer_norm_75, - layer_norm_76, - layer_norm_77, - layer_norm_78, - layer_norm_79, - layer_norm_8, - layer_norm_80, - layer_norm_81, - layer_norm_82, - layer_norm_83, - layer_norm_84, - layer_norm_85, - layer_norm_86, - layer_norm_87, - layer_norm_88, - layer_norm_89, - layer_norm_9, - layer_norm_90, - layer_norm_91, - layer_norm_92, - layer_norm_93, - layer_norm_94, - layer_norm_95, - layer_norm_96, - layer_norm_97, - layer_norm_98, - layer_norm_99, - matmul_0, - matmul_1, - matmul_10, - matmul_101, - matmul_102, - matmul_103, - matmul_104, - matmul_105, - matmul_106, - matmul_107, - matmul_109, - matmul_11, - matmul_110, - matmul_111, - matmul_112, - matmul_113, - matmul_114, - matmul_115, - matmul_117, - matmul_118, - matmul_119, - matmul_120, - matmul_121, - matmul_122, - matmul_123, - matmul_125, - matmul_126, - matmul_127, - matmul_128, - matmul_129, - matmul_13, - matmul_130, - matmul_131, - matmul_133, - matmul_134, - matmul_135, - matmul_136, - matmul_137, - matmul_138, - matmul_139, - matmul_14, - matmul_141, - matmul_142, - matmul_143, - matmul_144, - matmul_145, - matmul_146, - matmul_147, - matmul_149, - matmul_15, - matmul_150, - matmul_151, - matmul_152, - matmul_153, - matmul_154, - matmul_155, - matmul_157, - matmul_158, - matmul_159, - matmul_16, - matmul_160, - matmul_161, - matmul_162, - matmul_163, - matmul_165, - matmul_166, - matmul_167, - matmul_168, - matmul_169, - matmul_17, - matmul_170, - matmul_171, - matmul_173, - matmul_174, - matmul_175, - matmul_176, - matmul_177, - matmul_178, - matmul_179, - matmul_18, - matmul_181, - matmul_182, - matmul_183, - matmul_184, - matmul_185, - matmul_186, - matmul_187, - matmul_189, - matmul_19, - matmul_190, - matmul_191, - matmul_192, - matmul_2, - matmul_21, - matmul_22, - matmul_23, - matmul_24, - matmul_25, - matmul_26, - matmul_27, - matmul_29, - matmul_3, - matmul_30, - matmul_31, - matmul_32, - matmul_33, - matmul_34, - matmul_35, - matmul_37, - matmul_38, - matmul_39, - matmul_40, - matmul_41, - matmul_42, - matmul_43, - matmul_45, - matmul_46, - matmul_47, - matmul_48, - matmul_49, - matmul_5, - matmul_50, - matmul_51, - matmul_53, - matmul_54, - matmul_55, - matmul_56, - matmul_57, - matmul_58, - matmul_59, - matmul_6, - matmul_61, - matmul_62, - matmul_63, - matmul_64, - matmul_65, - matmul_66, - matmul_67, - matmul_69, - matmul_7, - matmul_70, - matmul_71, - matmul_72, - matmul_73, - matmul_74, - matmul_75, - matmul_77, - matmul_78, - matmul_79, - matmul_8, - matmul_80, - matmul_81, - matmul_82, - matmul_83, - matmul_85, - matmul_86, - matmul_87, - matmul_88, - matmul_89, - matmul_9, - matmul_90, - matmul_91, - matmul_93, - matmul_94, - matmul_95, - matmul_96, - matmul_97, - matmul_98, - matmul_99, - reshape_11, - reshape_15, - reshape_19, - reshape_23, - reshape_27, - reshape_3, - reshape_31, - reshape_35, - reshape_39, - reshape_43, - reshape_47, - reshape_51, - reshape_55, - reshape_59, - reshape_63, - reshape_67, - reshape_7, - reshape_71, - reshape_75, - reshape_79, - reshape_83, - reshape_87, - reshape_91, - reshape_95, - scale_1, - scale_10, - scale_11, - scale_12, - scale_13, - scale_14, - scale_15, - scale_16, - scale_17, - scale_18, - scale_19, - scale_2, - scale_20, - scale_21, - scale_22, - scale_23, - scale_24, - scale_3, - scale_4, - scale_5, - scale_6, - scale_7, - scale_8, - scale_9, - slice_0, - softmax_0, - softmax_1, - softmax_10, - softmax_11, - softmax_12, - softmax_13, - softmax_14, - softmax_15, - softmax_16, - softmax_17, - softmax_18, - softmax_19, - softmax_2, - softmax_20, - softmax_21, - softmax_22, - softmax_23, - softmax_3, - softmax_4, - softmax_5, - softmax_6, - softmax_7, - softmax_8, - softmax_9, - subtract_0, - transpose_1, - transpose_10, - transpose_11, - transpose_13, - transpose_14, - transpose_15, - transpose_17, - transpose_18, - transpose_19, - transpose_2, - transpose_21, - transpose_22, - transpose_23, - transpose_25, - transpose_26, - transpose_27, - transpose_29, - transpose_3, - transpose_30, - transpose_31, - transpose_33, - transpose_34, - transpose_35, - transpose_37, - transpose_38, - transpose_39, - transpose_41, - transpose_42, - transpose_43, - transpose_45, - transpose_46, - transpose_47, - transpose_49, - transpose_5, - transpose_50, - transpose_51, - transpose_53, - transpose_54, - transpose_55, - transpose_57, - transpose_58, - transpose_59, - transpose_6, - transpose_61, - transpose_62, - transpose_63, - transpose_65, - transpose_66, - transpose_67, - transpose_69, - transpose_7, - transpose_70, - transpose_71, - transpose_73, - transpose_74, - transpose_75, - transpose_77, - transpose_78, - transpose_79, - transpose_81, - transpose_82, - transpose_83, - transpose_85, - transpose_86, - transpose_87, - transpose_89, - transpose_9, - transpose_90, - transpose_91, - transpose_93, - transpose_94, - transpose_95, - unsqueeze_0, - ) + del add_218 return tanh_0 diff --git a/paddle_samples/PaddleNLP/bert-large-cased/weight_meta.py b/paddle_samples/PaddleNLP/bert-large-cased/weight_meta.py index 8b334348c..9a7cdf569 100644 --- a/paddle_samples/PaddleNLP/bert-large-cased/weight_meta.py +++ b/paddle_samples/PaddleNLP/bert-large-cased/weight_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_parameter_0: name = "parameter_0" + original_name = "linear_509.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0430012") @@ -11,6 +12,7 @@ class Program_weight_tensor_parameter_0: class Program_weight_tensor_parameter_1: name = "parameter_1" + original_name = "linear_509.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.17985") @@ -22,6 +24,7 @@ class Program_weight_tensor_parameter_1: class Program_weight_tensor_parameter_2: name = "parameter_2" + original_name = "layer_norm_173.b_0" shape = [1024] dtype = "float32" min_val = float("-0.908782") @@ -33,6 +36,7 @@ class Program_weight_tensor_parameter_2: class Program_weight_tensor_parameter_3: name = "parameter_3" + original_name = "layer_norm_173.w_0" shape = [1024] dtype = "float32" min_val = float("0.0100344") @@ -44,6 +48,7 @@ class Program_weight_tensor_parameter_3: class Program_weight_tensor_parameter_4: name = "parameter_4" + original_name = "layer_norm_172.b_0" shape = [1024] dtype = "float32" min_val = float("-0.363062") @@ -55,6 +60,7 @@ class Program_weight_tensor_parameter_4: class Program_weight_tensor_parameter_5: name = "parameter_5" + original_name = "layer_norm_172.w_0" shape = [1024] dtype = "float32" min_val = float("0.728466") @@ -66,6 +72,7 @@ class Program_weight_tensor_parameter_5: class Program_weight_tensor_parameter_6: name = "parameter_6" + original_name = "linear_508.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0882202") @@ -77,6 +84,7 @@ class Program_weight_tensor_parameter_6: class Program_weight_tensor_parameter_7: name = "parameter_7" + original_name = "linear_508.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-1.05412") @@ -88,6 +96,7 @@ class Program_weight_tensor_parameter_7: class Program_weight_tensor_parameter_8: name = "parameter_8" + original_name = "linear_507.b_0" shape = [4096] dtype = "float32" min_val = float("-0.316629") @@ -99,6 +108,7 @@ class Program_weight_tensor_parameter_8: class Program_weight_tensor_parameter_9: name = "parameter_9" + original_name = "linear_507.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.31583") @@ -110,6 +120,7 @@ class Program_weight_tensor_parameter_9: class Program_weight_tensor_parameter_10: name = "parameter_10" + original_name = "linear_506.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0712713") @@ -121,6 +132,7 @@ class Program_weight_tensor_parameter_10: class Program_weight_tensor_parameter_11: name = "parameter_11" + original_name = "linear_506.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.22639") @@ -132,6 +144,7 @@ class Program_weight_tensor_parameter_11: class Program_weight_tensor_parameter_12: name = "parameter_12" + original_name = "linear_505.b_0" shape = [1024] dtype = "float32" min_val = float("-0.139421") @@ -143,6 +156,7 @@ class Program_weight_tensor_parameter_12: class Program_weight_tensor_parameter_13: name = "parameter_13" + original_name = "linear_505.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.23454") @@ -154,6 +168,7 @@ class Program_weight_tensor_parameter_13: class Program_weight_tensor_parameter_14: name = "parameter_14" + original_name = "linear_504.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0158624") @@ -165,6 +180,7 @@ class Program_weight_tensor_parameter_14: class Program_weight_tensor_parameter_15: name = "parameter_15" + original_name = "linear_504.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.247036") @@ -176,6 +192,7 @@ class Program_weight_tensor_parameter_15: class Program_weight_tensor_parameter_16: name = "parameter_16" + original_name = "linear_503.b_0" shape = [1024] dtype = "float32" min_val = float("-0.575811") @@ -187,6 +204,7 @@ class Program_weight_tensor_parameter_16: class Program_weight_tensor_parameter_17: name = "parameter_17" + original_name = "linear_503.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.408011") @@ -198,6 +216,7 @@ class Program_weight_tensor_parameter_17: class Program_weight_tensor_parameter_18: name = "parameter_18" + original_name = "layer_norm_171.b_0" shape = [1024] dtype = "float32" min_val = float("-0.248615") @@ -209,6 +228,7 @@ class Program_weight_tensor_parameter_18: class Program_weight_tensor_parameter_19: name = "parameter_19" + original_name = "layer_norm_171.w_0" shape = [1024] dtype = "float32" min_val = float("0.0341817") @@ -220,6 +240,7 @@ class Program_weight_tensor_parameter_19: class Program_weight_tensor_parameter_20: name = "parameter_20" + original_name = "layer_norm_170.b_0" shape = [1024] dtype = "float32" min_val = float("-0.177391") @@ -231,6 +252,7 @@ class Program_weight_tensor_parameter_20: class Program_weight_tensor_parameter_21: name = "parameter_21" + original_name = "layer_norm_170.w_0" shape = [1024] dtype = "float32" min_val = float("0.773281") @@ -242,6 +264,7 @@ class Program_weight_tensor_parameter_21: class Program_weight_tensor_parameter_22: name = "parameter_22" + original_name = "linear_502.b_0" shape = [1024] dtype = "float32" min_val = float("-0.857333") @@ -253,6 +276,7 @@ class Program_weight_tensor_parameter_22: class Program_weight_tensor_parameter_23: name = "parameter_23" + original_name = "linear_502.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-3.72584") @@ -264,6 +288,7 @@ class Program_weight_tensor_parameter_23: class Program_weight_tensor_parameter_24: name = "parameter_24" + original_name = "linear_501.b_0" shape = [4096] dtype = "float32" min_val = float("-0.307395") @@ -275,6 +300,7 @@ class Program_weight_tensor_parameter_24: class Program_weight_tensor_parameter_25: name = "parameter_25" + original_name = "linear_501.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.440792") @@ -286,6 +312,7 @@ class Program_weight_tensor_parameter_25: class Program_weight_tensor_parameter_26: name = "parameter_26" + original_name = "linear_500.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0862491") @@ -297,6 +324,7 @@ class Program_weight_tensor_parameter_26: class Program_weight_tensor_parameter_27: name = "parameter_27" + original_name = "linear_500.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.341821") @@ -308,6 +336,7 @@ class Program_weight_tensor_parameter_27: class Program_weight_tensor_parameter_28: name = "parameter_28" + original_name = "linear_499.b_0" shape = [1024] dtype = "float32" min_val = float("-0.139509") @@ -319,6 +348,7 @@ class Program_weight_tensor_parameter_28: class Program_weight_tensor_parameter_29: name = "parameter_29" + original_name = "linear_499.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.199009") @@ -330,6 +360,7 @@ class Program_weight_tensor_parameter_29: class Program_weight_tensor_parameter_30: name = "parameter_30" + original_name = "linear_498.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0179717") @@ -341,6 +372,7 @@ class Program_weight_tensor_parameter_30: class Program_weight_tensor_parameter_31: name = "parameter_31" + original_name = "linear_498.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.218362") @@ -352,6 +384,7 @@ class Program_weight_tensor_parameter_31: class Program_weight_tensor_parameter_32: name = "parameter_32" + original_name = "linear_497.b_0" shape = [1024] dtype = "float32" min_val = float("-0.426456") @@ -363,6 +396,7 @@ class Program_weight_tensor_parameter_32: class Program_weight_tensor_parameter_33: name = "parameter_33" + original_name = "linear_497.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.535247") @@ -374,6 +408,7 @@ class Program_weight_tensor_parameter_33: class Program_weight_tensor_parameter_34: name = "parameter_34" + original_name = "layer_norm_169.b_0" shape = [1024] dtype = "float32" min_val = float("-0.212666") @@ -385,6 +420,7 @@ class Program_weight_tensor_parameter_34: class Program_weight_tensor_parameter_35: name = "parameter_35" + original_name = "layer_norm_169.w_0" shape = [1024] dtype = "float32" min_val = float("0.0522882") @@ -396,6 +432,7 @@ class Program_weight_tensor_parameter_35: class Program_weight_tensor_parameter_36: name = "parameter_36" + original_name = "layer_norm_168.b_0" shape = [1024] dtype = "float32" min_val = float("-0.266589") @@ -407,6 +444,7 @@ class Program_weight_tensor_parameter_36: class Program_weight_tensor_parameter_37: name = "parameter_37" + original_name = "layer_norm_168.w_0" shape = [1024] dtype = "float32" min_val = float("0.758878") @@ -418,6 +456,7 @@ class Program_weight_tensor_parameter_37: class Program_weight_tensor_parameter_38: name = "parameter_38" + original_name = "linear_496.b_0" shape = [1024] dtype = "float32" min_val = float("-0.511223") @@ -429,6 +468,7 @@ class Program_weight_tensor_parameter_38: class Program_weight_tensor_parameter_39: name = "parameter_39" + original_name = "linear_496.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-5.50502") @@ -440,6 +480,7 @@ class Program_weight_tensor_parameter_39: class Program_weight_tensor_parameter_40: name = "parameter_40" + original_name = "linear_495.b_0" shape = [4096] dtype = "float32" min_val = float("-0.229202") @@ -451,6 +492,7 @@ class Program_weight_tensor_parameter_40: class Program_weight_tensor_parameter_41: name = "parameter_41" + original_name = "linear_495.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-1.45023") @@ -462,6 +504,7 @@ class Program_weight_tensor_parameter_41: class Program_weight_tensor_parameter_42: name = "parameter_42" + original_name = "linear_494.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0949188") @@ -473,6 +516,7 @@ class Program_weight_tensor_parameter_42: class Program_weight_tensor_parameter_43: name = "parameter_43" + original_name = "linear_494.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.330248") @@ -484,6 +528,7 @@ class Program_weight_tensor_parameter_43: class Program_weight_tensor_parameter_44: name = "parameter_44" + original_name = "linear_493.b_0" shape = [1024] dtype = "float32" min_val = float("-0.175518") @@ -495,6 +540,7 @@ class Program_weight_tensor_parameter_44: class Program_weight_tensor_parameter_45: name = "parameter_45" + original_name = "linear_493.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.199822") @@ -506,6 +552,7 @@ class Program_weight_tensor_parameter_45: class Program_weight_tensor_parameter_46: name = "parameter_46" + original_name = "linear_492.b_0" shape = [1024] dtype = "float32" min_val = float("-0.020861") @@ -517,6 +564,7 @@ class Program_weight_tensor_parameter_46: class Program_weight_tensor_parameter_47: name = "parameter_47" + original_name = "linear_492.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.358463") @@ -528,6 +576,7 @@ class Program_weight_tensor_parameter_47: class Program_weight_tensor_parameter_48: name = "parameter_48" + original_name = "linear_491.b_0" shape = [1024] dtype = "float32" min_val = float("-0.543983") @@ -539,6 +588,7 @@ class Program_weight_tensor_parameter_48: class Program_weight_tensor_parameter_49: name = "parameter_49" + original_name = "linear_491.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.386871") @@ -550,6 +600,7 @@ class Program_weight_tensor_parameter_49: class Program_weight_tensor_parameter_50: name = "parameter_50" + original_name = "layer_norm_167.b_0" shape = [1024] dtype = "float32" min_val = float("-0.511743") @@ -561,6 +612,7 @@ class Program_weight_tensor_parameter_50: class Program_weight_tensor_parameter_51: name = "parameter_51" + original_name = "layer_norm_167.w_0" shape = [1024] dtype = "float32" min_val = float("0.0561587") @@ -572,6 +624,7 @@ class Program_weight_tensor_parameter_51: class Program_weight_tensor_parameter_52: name = "parameter_52" + original_name = "layer_norm_166.b_0" shape = [1024] dtype = "float32" min_val = float("-0.555149") @@ -583,6 +636,7 @@ class Program_weight_tensor_parameter_52: class Program_weight_tensor_parameter_53: name = "parameter_53" + original_name = "layer_norm_166.w_0" shape = [1024] dtype = "float32" min_val = float("0.73855") @@ -594,6 +648,7 @@ class Program_weight_tensor_parameter_53: class Program_weight_tensor_parameter_54: name = "parameter_54" + original_name = "linear_490.b_0" shape = [1024] dtype = "float32" min_val = float("-0.551806") @@ -605,6 +660,7 @@ class Program_weight_tensor_parameter_54: class Program_weight_tensor_parameter_55: name = "parameter_55" + original_name = "linear_490.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-4.42406") @@ -616,6 +672,7 @@ class Program_weight_tensor_parameter_55: class Program_weight_tensor_parameter_56: name = "parameter_56" + original_name = "linear_489.b_0" shape = [4096] dtype = "float32" min_val = float("-0.159612") @@ -627,6 +684,7 @@ class Program_weight_tensor_parameter_56: class Program_weight_tensor_parameter_57: name = "parameter_57" + original_name = "linear_489.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.642199") @@ -638,6 +696,7 @@ class Program_weight_tensor_parameter_57: class Program_weight_tensor_parameter_58: name = "parameter_58" + original_name = "linear_488.b_0" shape = [1024] dtype = "float32" min_val = float("-0.100609") @@ -649,6 +708,7 @@ class Program_weight_tensor_parameter_58: class Program_weight_tensor_parameter_59: name = "parameter_59" + original_name = "linear_488.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.211746") @@ -660,6 +720,7 @@ class Program_weight_tensor_parameter_59: class Program_weight_tensor_parameter_60: name = "parameter_60" + original_name = "linear_487.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0772414") @@ -671,6 +732,7 @@ class Program_weight_tensor_parameter_60: class Program_weight_tensor_parameter_61: name = "parameter_61" + original_name = "linear_487.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.198467") @@ -682,6 +744,7 @@ class Program_weight_tensor_parameter_61: class Program_weight_tensor_parameter_62: name = "parameter_62" + original_name = "linear_486.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0204162") @@ -693,6 +756,7 @@ class Program_weight_tensor_parameter_62: class Program_weight_tensor_parameter_63: name = "parameter_63" + original_name = "linear_486.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.342555") @@ -704,6 +768,7 @@ class Program_weight_tensor_parameter_63: class Program_weight_tensor_parameter_64: name = "parameter_64" + original_name = "linear_485.b_0" shape = [1024] dtype = "float32" min_val = float("-0.506437") @@ -715,6 +780,7 @@ class Program_weight_tensor_parameter_64: class Program_weight_tensor_parameter_65: name = "parameter_65" + original_name = "linear_485.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.382656") @@ -726,6 +792,7 @@ class Program_weight_tensor_parameter_65: class Program_weight_tensor_parameter_66: name = "parameter_66" + original_name = "layer_norm_165.b_0" shape = [1024] dtype = "float32" min_val = float("-0.768838") @@ -737,6 +804,7 @@ class Program_weight_tensor_parameter_66: class Program_weight_tensor_parameter_67: name = "parameter_67" + original_name = "layer_norm_165.w_0" shape = [1024] dtype = "float32" min_val = float("0.156808") @@ -748,6 +816,7 @@ class Program_weight_tensor_parameter_67: class Program_weight_tensor_parameter_68: name = "parameter_68" + original_name = "layer_norm_164.b_0" shape = [1024] dtype = "float32" min_val = float("-0.371178") @@ -759,6 +828,7 @@ class Program_weight_tensor_parameter_68: class Program_weight_tensor_parameter_69: name = "parameter_69" + original_name = "layer_norm_164.w_0" shape = [1024] dtype = "float32" min_val = float("0.696904") @@ -770,6 +840,7 @@ class Program_weight_tensor_parameter_69: class Program_weight_tensor_parameter_70: name = "parameter_70" + original_name = "linear_484.b_0" shape = [1024] dtype = "float32" min_val = float("-0.189996") @@ -781,6 +852,7 @@ class Program_weight_tensor_parameter_70: class Program_weight_tensor_parameter_71: name = "parameter_71" + original_name = "linear_484.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-1.99167") @@ -792,6 +864,7 @@ class Program_weight_tensor_parameter_71: class Program_weight_tensor_parameter_72: name = "parameter_72" + original_name = "linear_483.b_0" shape = [4096] dtype = "float32" min_val = float("-0.185716") @@ -803,6 +876,7 @@ class Program_weight_tensor_parameter_72: class Program_weight_tensor_parameter_73: name = "parameter_73" + original_name = "linear_483.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.579345") @@ -814,6 +888,7 @@ class Program_weight_tensor_parameter_73: class Program_weight_tensor_parameter_74: name = "parameter_74" + original_name = "linear_482.b_0" shape = [1024] dtype = "float32" min_val = float("-0.123784") @@ -825,6 +900,7 @@ class Program_weight_tensor_parameter_74: class Program_weight_tensor_parameter_75: name = "parameter_75" + original_name = "linear_482.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.151208") @@ -836,6 +912,7 @@ class Program_weight_tensor_parameter_75: class Program_weight_tensor_parameter_76: name = "parameter_76" + original_name = "linear_481.b_0" shape = [1024] dtype = "float32" min_val = float("-0.254386") @@ -847,6 +924,7 @@ class Program_weight_tensor_parameter_76: class Program_weight_tensor_parameter_77: name = "parameter_77" + original_name = "linear_481.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.156128") @@ -858,6 +936,7 @@ class Program_weight_tensor_parameter_77: class Program_weight_tensor_parameter_78: name = "parameter_78" + original_name = "linear_480.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0337296") @@ -869,6 +948,7 @@ class Program_weight_tensor_parameter_78: class Program_weight_tensor_parameter_79: name = "parameter_79" + original_name = "linear_480.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.600414") @@ -880,6 +960,7 @@ class Program_weight_tensor_parameter_79: class Program_weight_tensor_parameter_80: name = "parameter_80" + original_name = "linear_479.b_0" shape = [1024] dtype = "float32" min_val = float("-0.446516") @@ -891,6 +972,7 @@ class Program_weight_tensor_parameter_80: class Program_weight_tensor_parameter_81: name = "parameter_81" + original_name = "linear_479.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.304146") @@ -902,6 +984,7 @@ class Program_weight_tensor_parameter_81: class Program_weight_tensor_parameter_82: name = "parameter_82" + original_name = "layer_norm_163.b_0" shape = [1024] dtype = "float32" min_val = float("-0.366448") @@ -913,6 +996,7 @@ class Program_weight_tensor_parameter_82: class Program_weight_tensor_parameter_83: name = "parameter_83" + original_name = "layer_norm_163.w_0" shape = [1024] dtype = "float32" min_val = float("0.232306") @@ -924,6 +1008,7 @@ class Program_weight_tensor_parameter_83: class Program_weight_tensor_parameter_84: name = "parameter_84" + original_name = "layer_norm_162.b_0" shape = [1024] dtype = "float32" min_val = float("-0.634311") @@ -935,6 +1020,7 @@ class Program_weight_tensor_parameter_84: class Program_weight_tensor_parameter_85: name = "parameter_85" + original_name = "layer_norm_162.w_0" shape = [1024] dtype = "float32" min_val = float("0.715317") @@ -946,6 +1032,7 @@ class Program_weight_tensor_parameter_85: class Program_weight_tensor_parameter_86: name = "parameter_86" + original_name = "linear_478.b_0" shape = [1024] dtype = "float32" min_val = float("-0.188416") @@ -957,6 +1044,7 @@ class Program_weight_tensor_parameter_86: class Program_weight_tensor_parameter_87: name = "parameter_87" + original_name = "linear_478.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-1.59409") @@ -968,6 +1056,7 @@ class Program_weight_tensor_parameter_87: class Program_weight_tensor_parameter_88: name = "parameter_88" + original_name = "linear_477.b_0" shape = [4096] dtype = "float32" min_val = float("-0.333227") @@ -979,6 +1068,7 @@ class Program_weight_tensor_parameter_88: class Program_weight_tensor_parameter_89: name = "parameter_89" + original_name = "linear_477.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.519271") @@ -990,6 +1080,7 @@ class Program_weight_tensor_parameter_89: class Program_weight_tensor_parameter_90: name = "parameter_90" + original_name = "linear_476.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0858891") @@ -1001,6 +1092,7 @@ class Program_weight_tensor_parameter_90: class Program_weight_tensor_parameter_91: name = "parameter_91" + original_name = "linear_476.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.161561") @@ -1012,6 +1104,7 @@ class Program_weight_tensor_parameter_91: class Program_weight_tensor_parameter_92: name = "parameter_92" + original_name = "linear_475.b_0" shape = [1024] dtype = "float32" min_val = float("-0.129402") @@ -1023,6 +1116,7 @@ class Program_weight_tensor_parameter_92: class Program_weight_tensor_parameter_93: name = "parameter_93" + original_name = "linear_475.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.159483") @@ -1034,6 +1128,7 @@ class Program_weight_tensor_parameter_93: class Program_weight_tensor_parameter_94: name = "parameter_94" + original_name = "linear_474.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0190495") @@ -1045,6 +1140,7 @@ class Program_weight_tensor_parameter_94: class Program_weight_tensor_parameter_95: name = "parameter_95" + original_name = "linear_474.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.34486") @@ -1056,6 +1152,7 @@ class Program_weight_tensor_parameter_95: class Program_weight_tensor_parameter_96: name = "parameter_96" + original_name = "linear_473.b_0" shape = [1024] dtype = "float32" min_val = float("-0.407315") @@ -1067,6 +1164,7 @@ class Program_weight_tensor_parameter_96: class Program_weight_tensor_parameter_97: name = "parameter_97" + original_name = "linear_473.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.221519") @@ -1078,6 +1176,7 @@ class Program_weight_tensor_parameter_97: class Program_weight_tensor_parameter_98: name = "parameter_98" + original_name = "layer_norm_161.b_0" shape = [1024] dtype = "float32" min_val = float("-0.209739") @@ -1089,6 +1188,7 @@ class Program_weight_tensor_parameter_98: class Program_weight_tensor_parameter_99: name = "parameter_99" + original_name = "layer_norm_161.w_0" shape = [1024] dtype = "float32" min_val = float("0.188925") @@ -1100,6 +1200,7 @@ class Program_weight_tensor_parameter_99: class Program_weight_tensor_parameter_100: name = "parameter_100" + original_name = "layer_norm_160.b_0" shape = [1024] dtype = "float32" min_val = float("-0.934211") @@ -1111,6 +1212,7 @@ class Program_weight_tensor_parameter_100: class Program_weight_tensor_parameter_101: name = "parameter_101" + original_name = "layer_norm_160.w_0" shape = [1024] dtype = "float32" min_val = float("0.706977") @@ -1122,6 +1224,7 @@ class Program_weight_tensor_parameter_101: class Program_weight_tensor_parameter_102: name = "parameter_102" + original_name = "linear_472.b_0" shape = [1024] dtype = "float32" min_val = float("-0.258433") @@ -1133,6 +1236,7 @@ class Program_weight_tensor_parameter_102: class Program_weight_tensor_parameter_103: name = "parameter_103" + original_name = "linear_472.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-1.8565") @@ -1144,6 +1248,7 @@ class Program_weight_tensor_parameter_103: class Program_weight_tensor_parameter_104: name = "parameter_104" + original_name = "linear_471.b_0" shape = [4096] dtype = "float32" min_val = float("-0.240596") @@ -1155,6 +1260,7 @@ class Program_weight_tensor_parameter_104: class Program_weight_tensor_parameter_105: name = "parameter_105" + original_name = "linear_471.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.456073") @@ -1166,6 +1272,7 @@ class Program_weight_tensor_parameter_105: class Program_weight_tensor_parameter_106: name = "parameter_106" + original_name = "linear_470.b_0" shape = [1024] dtype = "float32" min_val = float("-0.081796") @@ -1177,6 +1284,7 @@ class Program_weight_tensor_parameter_106: class Program_weight_tensor_parameter_107: name = "parameter_107" + original_name = "linear_470.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.14612") @@ -1188,6 +1296,7 @@ class Program_weight_tensor_parameter_107: class Program_weight_tensor_parameter_108: name = "parameter_108" + original_name = "linear_469.b_0" shape = [1024] dtype = "float32" min_val = float("-0.113752") @@ -1199,6 +1308,7 @@ class Program_weight_tensor_parameter_108: class Program_weight_tensor_parameter_109: name = "parameter_109" + original_name = "linear_469.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.159771") @@ -1210,6 +1320,7 @@ class Program_weight_tensor_parameter_109: class Program_weight_tensor_parameter_110: name = "parameter_110" + original_name = "linear_468.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0228688") @@ -1221,6 +1332,7 @@ class Program_weight_tensor_parameter_110: class Program_weight_tensor_parameter_111: name = "parameter_111" + original_name = "linear_468.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.396104") @@ -1232,6 +1344,7 @@ class Program_weight_tensor_parameter_111: class Program_weight_tensor_parameter_112: name = "parameter_112" + original_name = "linear_467.b_0" shape = [1024] dtype = "float32" min_val = float("-0.391046") @@ -1243,6 +1356,7 @@ class Program_weight_tensor_parameter_112: class Program_weight_tensor_parameter_113: name = "parameter_113" + original_name = "linear_467.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.207937") @@ -1254,6 +1368,7 @@ class Program_weight_tensor_parameter_113: class Program_weight_tensor_parameter_114: name = "parameter_114" + original_name = "layer_norm_159.b_0" shape = [1024] dtype = "float32" min_val = float("-0.131935") @@ -1265,6 +1380,7 @@ class Program_weight_tensor_parameter_114: class Program_weight_tensor_parameter_115: name = "parameter_115" + original_name = "layer_norm_159.w_0" shape = [1024] dtype = "float32" min_val = float("0.166519") @@ -1276,6 +1392,7 @@ class Program_weight_tensor_parameter_115: class Program_weight_tensor_parameter_116: name = "parameter_116" + original_name = "layer_norm_158.b_0" shape = [1024] dtype = "float32" min_val = float("-1.6728") @@ -1287,6 +1404,7 @@ class Program_weight_tensor_parameter_116: class Program_weight_tensor_parameter_117: name = "parameter_117" + original_name = "layer_norm_158.w_0" shape = [1024] dtype = "float32" min_val = float("0.734145") @@ -1298,6 +1416,7 @@ class Program_weight_tensor_parameter_117: class Program_weight_tensor_parameter_118: name = "parameter_118" + original_name = "linear_466.b_0" shape = [1024] dtype = "float32" min_val = float("-0.62458") @@ -1309,6 +1428,7 @@ class Program_weight_tensor_parameter_118: class Program_weight_tensor_parameter_119: name = "parameter_119" + original_name = "linear_466.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-1.49896") @@ -1320,6 +1440,7 @@ class Program_weight_tensor_parameter_119: class Program_weight_tensor_parameter_120: name = "parameter_120" + original_name = "linear_465.b_0" shape = [4096] dtype = "float32" min_val = float("-0.287174") @@ -1331,6 +1452,7 @@ class Program_weight_tensor_parameter_120: class Program_weight_tensor_parameter_121: name = "parameter_121" + original_name = "linear_465.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.226027") @@ -1342,6 +1464,7 @@ class Program_weight_tensor_parameter_121: class Program_weight_tensor_parameter_122: name = "parameter_122" + original_name = "linear_464.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0927782") @@ -1353,6 +1476,7 @@ class Program_weight_tensor_parameter_122: class Program_weight_tensor_parameter_123: name = "parameter_123" + original_name = "linear_464.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.163711") @@ -1364,6 +1488,7 @@ class Program_weight_tensor_parameter_123: class Program_weight_tensor_parameter_124: name = "parameter_124" + original_name = "linear_463.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0744932") @@ -1375,6 +1500,7 @@ class Program_weight_tensor_parameter_124: class Program_weight_tensor_parameter_125: name = "parameter_125" + original_name = "linear_463.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.165767") @@ -1386,6 +1512,7 @@ class Program_weight_tensor_parameter_125: class Program_weight_tensor_parameter_126: name = "parameter_126" + original_name = "linear_462.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0211937") @@ -1397,6 +1524,7 @@ class Program_weight_tensor_parameter_126: class Program_weight_tensor_parameter_127: name = "parameter_127" + original_name = "linear_462.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.353616") @@ -1408,6 +1536,7 @@ class Program_weight_tensor_parameter_127: class Program_weight_tensor_parameter_128: name = "parameter_128" + original_name = "linear_461.b_0" shape = [1024] dtype = "float32" min_val = float("-0.380702") @@ -1419,6 +1548,7 @@ class Program_weight_tensor_parameter_128: class Program_weight_tensor_parameter_129: name = "parameter_129" + original_name = "linear_461.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.225697") @@ -1430,6 +1560,7 @@ class Program_weight_tensor_parameter_129: class Program_weight_tensor_parameter_130: name = "parameter_130" + original_name = "layer_norm_157.b_0" shape = [1024] dtype = "float32" min_val = float("-0.262908") @@ -1441,6 +1572,7 @@ class Program_weight_tensor_parameter_130: class Program_weight_tensor_parameter_131: name = "parameter_131" + original_name = "layer_norm_157.w_0" shape = [1024] dtype = "float32" min_val = float("0.196569") @@ -1452,6 +1584,7 @@ class Program_weight_tensor_parameter_131: class Program_weight_tensor_parameter_132: name = "parameter_132" + original_name = "layer_norm_156.b_0" shape = [1024] dtype = "float32" min_val = float("-1.72108") @@ -1463,6 +1596,7 @@ class Program_weight_tensor_parameter_132: class Program_weight_tensor_parameter_133: name = "parameter_133" + original_name = "layer_norm_156.w_0" shape = [1024] dtype = "float32" min_val = float("0.752003") @@ -1474,6 +1608,7 @@ class Program_weight_tensor_parameter_133: class Program_weight_tensor_parameter_134: name = "parameter_134" + original_name = "linear_460.b_0" shape = [1024] dtype = "float32" min_val = float("-0.980543") @@ -1485,6 +1620,7 @@ class Program_weight_tensor_parameter_134: class Program_weight_tensor_parameter_135: name = "parameter_135" + original_name = "linear_460.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-2.02448") @@ -1496,6 +1632,7 @@ class Program_weight_tensor_parameter_135: class Program_weight_tensor_parameter_136: name = "parameter_136" + original_name = "linear_459.b_0" shape = [4096] dtype = "float32" min_val = float("-0.24289") @@ -1507,6 +1644,7 @@ class Program_weight_tensor_parameter_136: class Program_weight_tensor_parameter_137: name = "parameter_137" + original_name = "linear_459.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.2255") @@ -1518,6 +1656,7 @@ class Program_weight_tensor_parameter_137: class Program_weight_tensor_parameter_138: name = "parameter_138" + original_name = "linear_458.b_0" shape = [1024] dtype = "float32" min_val = float("-0.117877") @@ -1529,6 +1668,7 @@ class Program_weight_tensor_parameter_138: class Program_weight_tensor_parameter_139: name = "parameter_139" + original_name = "linear_458.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.178463") @@ -1540,6 +1680,7 @@ class Program_weight_tensor_parameter_139: class Program_weight_tensor_parameter_140: name = "parameter_140" + original_name = "linear_457.b_0" shape = [1024] dtype = "float32" min_val = float("-0.11065") @@ -1551,6 +1692,7 @@ class Program_weight_tensor_parameter_140: class Program_weight_tensor_parameter_141: name = "parameter_141" + original_name = "linear_457.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.19698") @@ -1562,6 +1704,7 @@ class Program_weight_tensor_parameter_141: class Program_weight_tensor_parameter_142: name = "parameter_142" + original_name = "linear_456.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0159283") @@ -1573,6 +1716,7 @@ class Program_weight_tensor_parameter_142: class Program_weight_tensor_parameter_143: name = "parameter_143" + original_name = "linear_456.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.337245") @@ -1584,6 +1728,7 @@ class Program_weight_tensor_parameter_143: class Program_weight_tensor_parameter_144: name = "parameter_144" + original_name = "linear_455.b_0" shape = [1024] dtype = "float32" min_val = float("-0.331537") @@ -1595,6 +1740,7 @@ class Program_weight_tensor_parameter_144: class Program_weight_tensor_parameter_145: name = "parameter_145" + original_name = "linear_455.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.242136") @@ -1606,6 +1752,7 @@ class Program_weight_tensor_parameter_145: class Program_weight_tensor_parameter_146: name = "parameter_146" + original_name = "layer_norm_155.b_0" shape = [1024] dtype = "float32" min_val = float("-0.560337") @@ -1617,6 +1764,7 @@ class Program_weight_tensor_parameter_146: class Program_weight_tensor_parameter_147: name = "parameter_147" + original_name = "layer_norm_155.w_0" shape = [1024] dtype = "float32" min_val = float("0.203738") @@ -1628,6 +1776,7 @@ class Program_weight_tensor_parameter_147: class Program_weight_tensor_parameter_148: name = "parameter_148" + original_name = "layer_norm_154.b_0" shape = [1024] dtype = "float32" min_val = float("-1.50667") @@ -1639,6 +1788,7 @@ class Program_weight_tensor_parameter_148: class Program_weight_tensor_parameter_149: name = "parameter_149" + original_name = "layer_norm_154.w_0" shape = [1024] dtype = "float32" min_val = float("0.740913") @@ -1650,6 +1800,7 @@ class Program_weight_tensor_parameter_149: class Program_weight_tensor_parameter_150: name = "parameter_150" + original_name = "linear_454.b_0" shape = [1024] dtype = "float32" min_val = float("-1.1079") @@ -1661,6 +1812,7 @@ class Program_weight_tensor_parameter_150: class Program_weight_tensor_parameter_151: name = "parameter_151" + original_name = "linear_454.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-1.90795") @@ -1672,6 +1824,7 @@ class Program_weight_tensor_parameter_151: class Program_weight_tensor_parameter_152: name = "parameter_152" + original_name = "linear_453.b_0" shape = [4096] dtype = "float32" min_val = float("-0.25883") @@ -1683,6 +1836,7 @@ class Program_weight_tensor_parameter_152: class Program_weight_tensor_parameter_153: name = "parameter_153" + original_name = "linear_453.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.217263") @@ -1694,6 +1848,7 @@ class Program_weight_tensor_parameter_153: class Program_weight_tensor_parameter_154: name = "parameter_154" + original_name = "linear_452.b_0" shape = [1024] dtype = "float32" min_val = float("-0.119395") @@ -1705,6 +1860,7 @@ class Program_weight_tensor_parameter_154: class Program_weight_tensor_parameter_155: name = "parameter_155" + original_name = "linear_452.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.319894") @@ -1716,6 +1872,7 @@ class Program_weight_tensor_parameter_155: class Program_weight_tensor_parameter_156: name = "parameter_156" + original_name = "linear_451.b_0" shape = [1024] dtype = "float32" min_val = float("-0.104547") @@ -1727,6 +1884,7 @@ class Program_weight_tensor_parameter_156: class Program_weight_tensor_parameter_157: name = "parameter_157" + original_name = "linear_451.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.150081") @@ -1738,6 +1896,7 @@ class Program_weight_tensor_parameter_157: class Program_weight_tensor_parameter_158: name = "parameter_158" + original_name = "linear_450.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0195877") @@ -1749,6 +1908,7 @@ class Program_weight_tensor_parameter_158: class Program_weight_tensor_parameter_159: name = "parameter_159" + original_name = "linear_450.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.375189") @@ -1760,6 +1920,7 @@ class Program_weight_tensor_parameter_159: class Program_weight_tensor_parameter_160: name = "parameter_160" + original_name = "linear_449.b_0" shape = [1024] dtype = "float32" min_val = float("-0.340232") @@ -1771,6 +1932,7 @@ class Program_weight_tensor_parameter_160: class Program_weight_tensor_parameter_161: name = "parameter_161" + original_name = "linear_449.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.267048") @@ -1782,6 +1944,7 @@ class Program_weight_tensor_parameter_161: class Program_weight_tensor_parameter_162: name = "parameter_162" + original_name = "layer_norm_153.b_0" shape = [1024] dtype = "float32" min_val = float("-0.429308") @@ -1793,6 +1956,7 @@ class Program_weight_tensor_parameter_162: class Program_weight_tensor_parameter_163: name = "parameter_163" + original_name = "layer_norm_153.w_0" shape = [1024] dtype = "float32" min_val = float("0.509424") @@ -1804,6 +1968,7 @@ class Program_weight_tensor_parameter_163: class Program_weight_tensor_parameter_164: name = "parameter_164" + original_name = "layer_norm_152.b_0" shape = [1024] dtype = "float32" min_val = float("-1.42681") @@ -1815,6 +1980,7 @@ class Program_weight_tensor_parameter_164: class Program_weight_tensor_parameter_165: name = "parameter_165" + original_name = "layer_norm_152.w_0" shape = [1024] dtype = "float32" min_val = float("0.754534") @@ -1826,6 +1992,7 @@ class Program_weight_tensor_parameter_165: class Program_weight_tensor_parameter_166: name = "parameter_166" + original_name = "linear_448.b_0" shape = [1024] dtype = "float32" min_val = float("-1.10884") @@ -1837,6 +2004,7 @@ class Program_weight_tensor_parameter_166: class Program_weight_tensor_parameter_167: name = "parameter_167" + original_name = "linear_448.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-2.30417") @@ -1848,6 +2016,7 @@ class Program_weight_tensor_parameter_167: class Program_weight_tensor_parameter_168: name = "parameter_168" + original_name = "linear_447.b_0" shape = [4096] dtype = "float32" min_val = float("-0.233174") @@ -1859,6 +2028,7 @@ class Program_weight_tensor_parameter_168: class Program_weight_tensor_parameter_169: name = "parameter_169" + original_name = "linear_447.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.222362") @@ -1870,6 +2040,7 @@ class Program_weight_tensor_parameter_169: class Program_weight_tensor_parameter_170: name = "parameter_170" + original_name = "linear_446.b_0" shape = [1024] dtype = "float32" min_val = float("-0.127734") @@ -1881,6 +2052,7 @@ class Program_weight_tensor_parameter_170: class Program_weight_tensor_parameter_171: name = "parameter_171" + original_name = "linear_446.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.366456") @@ -1892,6 +2064,7 @@ class Program_weight_tensor_parameter_171: class Program_weight_tensor_parameter_172: name = "parameter_172" + original_name = "linear_445.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0730385") @@ -1903,6 +2076,7 @@ class Program_weight_tensor_parameter_172: class Program_weight_tensor_parameter_173: name = "parameter_173" + original_name = "linear_445.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.166437") @@ -1914,6 +2088,7 @@ class Program_weight_tensor_parameter_173: class Program_weight_tensor_parameter_174: name = "parameter_174" + original_name = "linear_444.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0159269") @@ -1925,6 +2100,7 @@ class Program_weight_tensor_parameter_174: class Program_weight_tensor_parameter_175: name = "parameter_175" + original_name = "linear_444.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.377275") @@ -1936,6 +2112,7 @@ class Program_weight_tensor_parameter_175: class Program_weight_tensor_parameter_176: name = "parameter_176" + original_name = "linear_443.b_0" shape = [1024] dtype = "float32" min_val = float("-0.267621") @@ -1947,6 +2124,7 @@ class Program_weight_tensor_parameter_176: class Program_weight_tensor_parameter_177: name = "parameter_177" + original_name = "linear_443.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.240734") @@ -1958,6 +2136,7 @@ class Program_weight_tensor_parameter_177: class Program_weight_tensor_parameter_178: name = "parameter_178" + original_name = "layer_norm_151.b_0" shape = [1024] dtype = "float32" min_val = float("-0.319129") @@ -1969,6 +2148,7 @@ class Program_weight_tensor_parameter_178: class Program_weight_tensor_parameter_179: name = "parameter_179" + original_name = "layer_norm_151.w_0" shape = [1024] dtype = "float32" min_val = float("0.531219") @@ -1980,6 +2160,7 @@ class Program_weight_tensor_parameter_179: class Program_weight_tensor_parameter_180: name = "parameter_180" + original_name = "layer_norm_150.b_0" shape = [1024] dtype = "float32" min_val = float("-1.24151") @@ -1991,6 +2172,7 @@ class Program_weight_tensor_parameter_180: class Program_weight_tensor_parameter_181: name = "parameter_181" + original_name = "layer_norm_150.w_0" shape = [1024] dtype = "float32" min_val = float("0.760164") @@ -2002,6 +2184,7 @@ class Program_weight_tensor_parameter_181: class Program_weight_tensor_parameter_182: name = "parameter_182" + original_name = "linear_442.b_0" shape = [1024] dtype = "float32" min_val = float("-0.751521") @@ -2013,6 +2196,7 @@ class Program_weight_tensor_parameter_182: class Program_weight_tensor_parameter_183: name = "parameter_183" + original_name = "linear_442.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-2.57555") @@ -2024,6 +2208,7 @@ class Program_weight_tensor_parameter_183: class Program_weight_tensor_parameter_184: name = "parameter_184" + original_name = "linear_441.b_0" shape = [4096] dtype = "float32" min_val = float("-0.205014") @@ -2035,6 +2220,7 @@ class Program_weight_tensor_parameter_184: class Program_weight_tensor_parameter_185: name = "parameter_185" + original_name = "linear_441.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.223437") @@ -2046,6 +2232,7 @@ class Program_weight_tensor_parameter_185: class Program_weight_tensor_parameter_186: name = "parameter_186" + original_name = "linear_440.b_0" shape = [1024] dtype = "float32" min_val = float("-0.115155") @@ -2057,6 +2244,7 @@ class Program_weight_tensor_parameter_186: class Program_weight_tensor_parameter_187: name = "parameter_187" + original_name = "linear_440.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.303882") @@ -2068,6 +2256,7 @@ class Program_weight_tensor_parameter_187: class Program_weight_tensor_parameter_188: name = "parameter_188" + original_name = "linear_439.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0577186") @@ -2079,6 +2268,7 @@ class Program_weight_tensor_parameter_188: class Program_weight_tensor_parameter_189: name = "parameter_189" + original_name = "linear_439.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.149522") @@ -2090,6 +2280,7 @@ class Program_weight_tensor_parameter_189: class Program_weight_tensor_parameter_190: name = "parameter_190" + original_name = "linear_438.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0225848") @@ -2101,6 +2292,7 @@ class Program_weight_tensor_parameter_190: class Program_weight_tensor_parameter_191: name = "parameter_191" + original_name = "linear_438.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.400763") @@ -2112,6 +2304,7 @@ class Program_weight_tensor_parameter_191: class Program_weight_tensor_parameter_192: name = "parameter_192" + original_name = "linear_437.b_0" shape = [1024] dtype = "float32" min_val = float("-0.337864") @@ -2123,6 +2316,7 @@ class Program_weight_tensor_parameter_192: class Program_weight_tensor_parameter_193: name = "parameter_193" + original_name = "linear_437.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.252943") @@ -2134,6 +2328,7 @@ class Program_weight_tensor_parameter_193: class Program_weight_tensor_parameter_194: name = "parameter_194" + original_name = "layer_norm_149.b_0" shape = [1024] dtype = "float32" min_val = float("-0.242688") @@ -2145,6 +2340,7 @@ class Program_weight_tensor_parameter_194: class Program_weight_tensor_parameter_195: name = "parameter_195" + original_name = "layer_norm_149.w_0" shape = [1024] dtype = "float32" min_val = float("0.540136") @@ -2156,6 +2352,7 @@ class Program_weight_tensor_parameter_195: class Program_weight_tensor_parameter_196: name = "parameter_196" + original_name = "layer_norm_148.b_0" shape = [1024] dtype = "float32" min_val = float("-1.43639") @@ -2167,6 +2364,7 @@ class Program_weight_tensor_parameter_196: class Program_weight_tensor_parameter_197: name = "parameter_197" + original_name = "layer_norm_148.w_0" shape = [1024] dtype = "float32" min_val = float("0.776947") @@ -2178,6 +2376,7 @@ class Program_weight_tensor_parameter_197: class Program_weight_tensor_parameter_198: name = "parameter_198" + original_name = "linear_436.b_0" shape = [1024] dtype = "float32" min_val = float("-0.472831") @@ -2189,6 +2388,7 @@ class Program_weight_tensor_parameter_198: class Program_weight_tensor_parameter_199: name = "parameter_199" + original_name = "linear_436.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-1.24563") @@ -2200,6 +2400,7 @@ class Program_weight_tensor_parameter_199: class Program_weight_tensor_parameter_200: name = "parameter_200" + original_name = "linear_435.b_0" shape = [4096] dtype = "float32" min_val = float("-0.21636") @@ -2211,6 +2412,7 @@ class Program_weight_tensor_parameter_200: class Program_weight_tensor_parameter_201: name = "parameter_201" + original_name = "linear_435.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.277779") @@ -2222,6 +2424,7 @@ class Program_weight_tensor_parameter_201: class Program_weight_tensor_parameter_202: name = "parameter_202" + original_name = "linear_434.b_0" shape = [1024] dtype = "float32" min_val = float("-0.10921") @@ -2233,6 +2436,7 @@ class Program_weight_tensor_parameter_202: class Program_weight_tensor_parameter_203: name = "parameter_203" + original_name = "linear_434.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.292359") @@ -2244,6 +2448,7 @@ class Program_weight_tensor_parameter_203: class Program_weight_tensor_parameter_204: name = "parameter_204" + original_name = "linear_433.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0777452") @@ -2255,6 +2460,7 @@ class Program_weight_tensor_parameter_204: class Program_weight_tensor_parameter_205: name = "parameter_205" + original_name = "linear_433.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.15248") @@ -2266,6 +2472,7 @@ class Program_weight_tensor_parameter_205: class Program_weight_tensor_parameter_206: name = "parameter_206" + original_name = "linear_432.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0170467") @@ -2277,6 +2484,7 @@ class Program_weight_tensor_parameter_206: class Program_weight_tensor_parameter_207: name = "parameter_207" + original_name = "linear_432.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.420545") @@ -2288,6 +2496,7 @@ class Program_weight_tensor_parameter_207: class Program_weight_tensor_parameter_208: name = "parameter_208" + original_name = "linear_431.b_0" shape = [1024] dtype = "float32" min_val = float("-0.390785") @@ -2299,6 +2508,7 @@ class Program_weight_tensor_parameter_208: class Program_weight_tensor_parameter_209: name = "parameter_209" + original_name = "linear_431.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.282042") @@ -2310,6 +2520,7 @@ class Program_weight_tensor_parameter_209: class Program_weight_tensor_parameter_210: name = "parameter_210" + original_name = "layer_norm_147.b_0" shape = [1024] dtype = "float32" min_val = float("-0.389218") @@ -2321,6 +2532,7 @@ class Program_weight_tensor_parameter_210: class Program_weight_tensor_parameter_211: name = "parameter_211" + original_name = "layer_norm_147.w_0" shape = [1024] dtype = "float32" min_val = float("0.288372") @@ -2332,6 +2544,7 @@ class Program_weight_tensor_parameter_211: class Program_weight_tensor_parameter_212: name = "parameter_212" + original_name = "layer_norm_146.b_0" shape = [1024] dtype = "float32" min_val = float("-1.98954") @@ -2343,6 +2556,7 @@ class Program_weight_tensor_parameter_212: class Program_weight_tensor_parameter_213: name = "parameter_213" + original_name = "layer_norm_146.w_0" shape = [1024] dtype = "float32" min_val = float("0.795337") @@ -2354,6 +2568,7 @@ class Program_weight_tensor_parameter_213: class Program_weight_tensor_parameter_214: name = "parameter_214" + original_name = "linear_430.b_0" shape = [1024] dtype = "float32" min_val = float("-0.186844") @@ -2365,6 +2580,7 @@ class Program_weight_tensor_parameter_214: class Program_weight_tensor_parameter_215: name = "parameter_215" + original_name = "linear_430.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-2.33454") @@ -2376,6 +2592,7 @@ class Program_weight_tensor_parameter_215: class Program_weight_tensor_parameter_216: name = "parameter_216" + original_name = "linear_429.b_0" shape = [4096] dtype = "float32" min_val = float("-0.212245") @@ -2387,6 +2604,7 @@ class Program_weight_tensor_parameter_216: class Program_weight_tensor_parameter_217: name = "parameter_217" + original_name = "linear_429.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.278712") @@ -2398,6 +2616,7 @@ class Program_weight_tensor_parameter_217: class Program_weight_tensor_parameter_218: name = "parameter_218" + original_name = "linear_428.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0663037") @@ -2409,6 +2628,7 @@ class Program_weight_tensor_parameter_218: class Program_weight_tensor_parameter_219: name = "parameter_219" + original_name = "linear_428.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.242517") @@ -2420,6 +2640,7 @@ class Program_weight_tensor_parameter_219: class Program_weight_tensor_parameter_220: name = "parameter_220" + original_name = "linear_427.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0440177") @@ -2431,6 +2652,7 @@ class Program_weight_tensor_parameter_220: class Program_weight_tensor_parameter_221: name = "parameter_221" + original_name = "linear_427.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.146316") @@ -2442,6 +2664,7 @@ class Program_weight_tensor_parameter_221: class Program_weight_tensor_parameter_222: name = "parameter_222" + original_name = "linear_426.b_0" shape = [1024] dtype = "float32" min_val = float("-0.017582") @@ -2453,6 +2676,7 @@ class Program_weight_tensor_parameter_222: class Program_weight_tensor_parameter_223: name = "parameter_223" + original_name = "linear_426.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.266794") @@ -2464,6 +2688,7 @@ class Program_weight_tensor_parameter_223: class Program_weight_tensor_parameter_224: name = "parameter_224" + original_name = "linear_425.b_0" shape = [1024] dtype = "float32" min_val = float("-0.450888") @@ -2475,6 +2700,7 @@ class Program_weight_tensor_parameter_224: class Program_weight_tensor_parameter_225: name = "parameter_225" + original_name = "linear_425.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.280539") @@ -2486,6 +2712,7 @@ class Program_weight_tensor_parameter_225: class Program_weight_tensor_parameter_226: name = "parameter_226" + original_name = "layer_norm_145.b_0" shape = [1024] dtype = "float32" min_val = float("-1.33693") @@ -2497,6 +2724,7 @@ class Program_weight_tensor_parameter_226: class Program_weight_tensor_parameter_227: name = "parameter_227" + original_name = "layer_norm_145.w_0" shape = [1024] dtype = "float32" min_val = float("0.274016") @@ -2508,6 +2736,7 @@ class Program_weight_tensor_parameter_227: class Program_weight_tensor_parameter_228: name = "parameter_228" + original_name = "layer_norm_144.b_0" shape = [1024] dtype = "float32" min_val = float("-1.68729") @@ -2519,6 +2748,7 @@ class Program_weight_tensor_parameter_228: class Program_weight_tensor_parameter_229: name = "parameter_229" + original_name = "layer_norm_144.w_0" shape = [1024] dtype = "float32" min_val = float("0.791633") @@ -2530,6 +2760,7 @@ class Program_weight_tensor_parameter_229: class Program_weight_tensor_parameter_230: name = "parameter_230" + original_name = "linear_424.b_0" shape = [1024] dtype = "float32" min_val = float("-0.187463") @@ -2541,6 +2772,7 @@ class Program_weight_tensor_parameter_230: class Program_weight_tensor_parameter_231: name = "parameter_231" + original_name = "linear_424.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-3.22856") @@ -2552,6 +2784,7 @@ class Program_weight_tensor_parameter_231: class Program_weight_tensor_parameter_232: name = "parameter_232" + original_name = "linear_423.b_0" shape = [4096] dtype = "float32" min_val = float("-0.177615") @@ -2563,6 +2796,7 @@ class Program_weight_tensor_parameter_232: class Program_weight_tensor_parameter_233: name = "parameter_233" + original_name = "linear_423.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.268281") @@ -2574,6 +2808,7 @@ class Program_weight_tensor_parameter_233: class Program_weight_tensor_parameter_234: name = "parameter_234" + original_name = "linear_422.b_0" shape = [1024] dtype = "float32" min_val = float("-0.120136") @@ -2585,6 +2820,7 @@ class Program_weight_tensor_parameter_234: class Program_weight_tensor_parameter_235: name = "parameter_235" + original_name = "linear_422.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.213591") @@ -2596,6 +2832,7 @@ class Program_weight_tensor_parameter_235: class Program_weight_tensor_parameter_236: name = "parameter_236" + original_name = "linear_421.b_0" shape = [1024] dtype = "float32" min_val = float("-0.034902") @@ -2607,6 +2844,7 @@ class Program_weight_tensor_parameter_236: class Program_weight_tensor_parameter_237: name = "parameter_237" + original_name = "linear_421.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.14413") @@ -2618,6 +2856,7 @@ class Program_weight_tensor_parameter_237: class Program_weight_tensor_parameter_238: name = "parameter_238" + original_name = "linear_420.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0148175") @@ -2629,6 +2868,7 @@ class Program_weight_tensor_parameter_238: class Program_weight_tensor_parameter_239: name = "parameter_239" + original_name = "linear_420.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.296475") @@ -2640,6 +2880,7 @@ class Program_weight_tensor_parameter_239: class Program_weight_tensor_parameter_240: name = "parameter_240" + original_name = "linear_419.b_0" shape = [1024] dtype = "float32" min_val = float("-0.320084") @@ -2651,6 +2892,7 @@ class Program_weight_tensor_parameter_240: class Program_weight_tensor_parameter_241: name = "parameter_241" + original_name = "linear_419.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.395196") @@ -2662,6 +2904,7 @@ class Program_weight_tensor_parameter_241: class Program_weight_tensor_parameter_242: name = "parameter_242" + original_name = "layer_norm_143.b_0" shape = [1024] dtype = "float32" min_val = float("-1.11657") @@ -2673,6 +2916,7 @@ class Program_weight_tensor_parameter_242: class Program_weight_tensor_parameter_243: name = "parameter_243" + original_name = "layer_norm_143.w_0" shape = [1024] dtype = "float32" min_val = float("0.325422") @@ -2684,6 +2928,7 @@ class Program_weight_tensor_parameter_243: class Program_weight_tensor_parameter_244: name = "parameter_244" + original_name = "layer_norm_142.b_0" shape = [1024] dtype = "float32" min_val = float("-1.54046") @@ -2695,6 +2940,7 @@ class Program_weight_tensor_parameter_244: class Program_weight_tensor_parameter_245: name = "parameter_245" + original_name = "layer_norm_142.w_0" shape = [1024] dtype = "float32" min_val = float("0.769122") @@ -2706,6 +2952,7 @@ class Program_weight_tensor_parameter_245: class Program_weight_tensor_parameter_246: name = "parameter_246" + original_name = "linear_418.b_0" shape = [1024] dtype = "float32" min_val = float("-0.179937") @@ -2717,6 +2964,7 @@ class Program_weight_tensor_parameter_246: class Program_weight_tensor_parameter_247: name = "parameter_247" + original_name = "linear_418.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-2.252") @@ -2728,6 +2976,7 @@ class Program_weight_tensor_parameter_247: class Program_weight_tensor_parameter_248: name = "parameter_248" + original_name = "linear_417.b_0" shape = [4096] dtype = "float32" min_val = float("-0.175273") @@ -2739,6 +2988,7 @@ class Program_weight_tensor_parameter_248: class Program_weight_tensor_parameter_249: name = "parameter_249" + original_name = "linear_417.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.20667") @@ -2750,6 +3000,7 @@ class Program_weight_tensor_parameter_249: class Program_weight_tensor_parameter_250: name = "parameter_250" + original_name = "linear_416.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0976096") @@ -2761,6 +3012,7 @@ class Program_weight_tensor_parameter_250: class Program_weight_tensor_parameter_251: name = "parameter_251" + original_name = "linear_416.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.21447") @@ -2772,6 +3024,7 @@ class Program_weight_tensor_parameter_251: class Program_weight_tensor_parameter_252: name = "parameter_252" + original_name = "linear_415.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0469114") @@ -2783,6 +3036,7 @@ class Program_weight_tensor_parameter_252: class Program_weight_tensor_parameter_253: name = "parameter_253" + original_name = "linear_415.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.151922") @@ -2794,6 +3048,7 @@ class Program_weight_tensor_parameter_253: class Program_weight_tensor_parameter_254: name = "parameter_254" + original_name = "linear_414.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0119037") @@ -2805,6 +3060,7 @@ class Program_weight_tensor_parameter_254: class Program_weight_tensor_parameter_255: name = "parameter_255" + original_name = "linear_414.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.333322") @@ -2816,6 +3072,7 @@ class Program_weight_tensor_parameter_255: class Program_weight_tensor_parameter_256: name = "parameter_256" + original_name = "linear_413.b_0" shape = [1024] dtype = "float32" min_val = float("-0.250271") @@ -2827,6 +3084,7 @@ class Program_weight_tensor_parameter_256: class Program_weight_tensor_parameter_257: name = "parameter_257" + original_name = "linear_413.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.30161") @@ -2838,6 +3096,7 @@ class Program_weight_tensor_parameter_257: class Program_weight_tensor_parameter_258: name = "parameter_258" + original_name = "layer_norm_141.b_0" shape = [1024] dtype = "float32" min_val = float("-0.377568") @@ -2849,6 +3108,7 @@ class Program_weight_tensor_parameter_258: class Program_weight_tensor_parameter_259: name = "parameter_259" + original_name = "layer_norm_141.w_0" shape = [1024] dtype = "float32" min_val = float("0.0786076") @@ -2860,6 +3120,7 @@ class Program_weight_tensor_parameter_259: class Program_weight_tensor_parameter_260: name = "parameter_260" + original_name = "layer_norm_140.b_0" shape = [1024] dtype = "float32" min_val = float("-1.45184") @@ -2871,6 +3132,7 @@ class Program_weight_tensor_parameter_260: class Program_weight_tensor_parameter_261: name = "parameter_261" + original_name = "layer_norm_140.w_0" shape = [1024] dtype = "float32" min_val = float("0.795311") @@ -2882,6 +3144,7 @@ class Program_weight_tensor_parameter_261: class Program_weight_tensor_parameter_262: name = "parameter_262" + original_name = "linear_412.b_0" shape = [1024] dtype = "float32" min_val = float("-0.304321") @@ -2893,6 +3156,7 @@ class Program_weight_tensor_parameter_262: class Program_weight_tensor_parameter_263: name = "parameter_263" + original_name = "linear_412.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-1.82574") @@ -2904,6 +3168,7 @@ class Program_weight_tensor_parameter_263: class Program_weight_tensor_parameter_264: name = "parameter_264" + original_name = "linear_411.b_0" shape = [4096] dtype = "float32" min_val = float("-0.150745") @@ -2915,6 +3180,7 @@ class Program_weight_tensor_parameter_264: class Program_weight_tensor_parameter_265: name = "parameter_265" + original_name = "linear_411.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.251007") @@ -2926,6 +3192,7 @@ class Program_weight_tensor_parameter_265: class Program_weight_tensor_parameter_266: name = "parameter_266" + original_name = "linear_410.b_0" shape = [1024] dtype = "float32" min_val = float("-0.179017") @@ -2937,6 +3204,7 @@ class Program_weight_tensor_parameter_266: class Program_weight_tensor_parameter_267: name = "parameter_267" + original_name = "linear_410.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.263997") @@ -2948,6 +3216,7 @@ class Program_weight_tensor_parameter_267: class Program_weight_tensor_parameter_268: name = "parameter_268" + original_name = "linear_409.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0497341") @@ -2959,6 +3228,7 @@ class Program_weight_tensor_parameter_268: class Program_weight_tensor_parameter_269: name = "parameter_269" + original_name = "linear_409.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.141482") @@ -2970,6 +3240,7 @@ class Program_weight_tensor_parameter_269: class Program_weight_tensor_parameter_270: name = "parameter_270" + original_name = "linear_408.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0200129") @@ -2981,6 +3252,7 @@ class Program_weight_tensor_parameter_270: class Program_weight_tensor_parameter_271: name = "parameter_271" + original_name = "linear_408.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.369548") @@ -2992,6 +3264,7 @@ class Program_weight_tensor_parameter_271: class Program_weight_tensor_parameter_272: name = "parameter_272" + original_name = "linear_407.b_0" shape = [1024] dtype = "float32" min_val = float("-0.347314") @@ -3003,6 +3276,7 @@ class Program_weight_tensor_parameter_272: class Program_weight_tensor_parameter_273: name = "parameter_273" + original_name = "linear_407.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.350792") @@ -3014,6 +3288,7 @@ class Program_weight_tensor_parameter_273: class Program_weight_tensor_parameter_274: name = "parameter_274" + original_name = "layer_norm_139.b_0" shape = [1024] dtype = "float32" min_val = float("-0.226646") @@ -3025,6 +3300,7 @@ class Program_weight_tensor_parameter_274: class Program_weight_tensor_parameter_275: name = "parameter_275" + original_name = "layer_norm_139.w_0" shape = [1024] dtype = "float32" min_val = float("0.201182") @@ -3036,6 +3312,7 @@ class Program_weight_tensor_parameter_275: class Program_weight_tensor_parameter_276: name = "parameter_276" + original_name = "layer_norm_138.b_0" shape = [1024] dtype = "float32" min_val = float("-1.3274") @@ -3047,6 +3324,7 @@ class Program_weight_tensor_parameter_276: class Program_weight_tensor_parameter_277: name = "parameter_277" + original_name = "layer_norm_138.w_0" shape = [1024] dtype = "float32" min_val = float("0.764487") @@ -3058,6 +3336,7 @@ class Program_weight_tensor_parameter_277: class Program_weight_tensor_parameter_278: name = "parameter_278" + original_name = "linear_406.b_0" shape = [1024] dtype = "float32" min_val = float("-0.109975") @@ -3069,6 +3348,7 @@ class Program_weight_tensor_parameter_278: class Program_weight_tensor_parameter_279: name = "parameter_279" + original_name = "linear_406.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-2.26317") @@ -3080,6 +3360,7 @@ class Program_weight_tensor_parameter_279: class Program_weight_tensor_parameter_280: name = "parameter_280" + original_name = "linear_405.b_0" shape = [4096] dtype = "float32" min_val = float("-0.185386") @@ -3091,6 +3372,7 @@ class Program_weight_tensor_parameter_280: class Program_weight_tensor_parameter_281: name = "parameter_281" + original_name = "linear_405.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.246557") @@ -3102,6 +3384,7 @@ class Program_weight_tensor_parameter_281: class Program_weight_tensor_parameter_282: name = "parameter_282" + original_name = "linear_404.b_0" shape = [1024] dtype = "float32" min_val = float("-0.121071") @@ -3113,6 +3396,7 @@ class Program_weight_tensor_parameter_282: class Program_weight_tensor_parameter_283: name = "parameter_283" + original_name = "linear_404.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.197418") @@ -3124,6 +3408,7 @@ class Program_weight_tensor_parameter_283: class Program_weight_tensor_parameter_284: name = "parameter_284" + original_name = "linear_403.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0880857") @@ -3135,6 +3420,7 @@ class Program_weight_tensor_parameter_284: class Program_weight_tensor_parameter_285: name = "parameter_285" + original_name = "linear_403.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.13982") @@ -3146,6 +3432,7 @@ class Program_weight_tensor_parameter_285: class Program_weight_tensor_parameter_286: name = "parameter_286" + original_name = "linear_402.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0158414") @@ -3157,6 +3444,7 @@ class Program_weight_tensor_parameter_286: class Program_weight_tensor_parameter_287: name = "parameter_287" + original_name = "linear_402.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.346159") @@ -3168,6 +3456,7 @@ class Program_weight_tensor_parameter_287: class Program_weight_tensor_parameter_288: name = "parameter_288" + original_name = "linear_401.b_0" shape = [1024] dtype = "float32" min_val = float("-0.209842") @@ -3179,6 +3468,7 @@ class Program_weight_tensor_parameter_288: class Program_weight_tensor_parameter_289: name = "parameter_289" + original_name = "linear_401.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.272204") @@ -3190,6 +3480,7 @@ class Program_weight_tensor_parameter_289: class Program_weight_tensor_parameter_290: name = "parameter_290" + original_name = "layer_norm_137.b_0" shape = [1024] dtype = "float32" min_val = float("-0.21713") @@ -3201,6 +3492,7 @@ class Program_weight_tensor_parameter_290: class Program_weight_tensor_parameter_291: name = "parameter_291" + original_name = "layer_norm_137.w_0" shape = [1024] dtype = "float32" min_val = float("0.381031") @@ -3212,6 +3504,7 @@ class Program_weight_tensor_parameter_291: class Program_weight_tensor_parameter_292: name = "parameter_292" + original_name = "layer_norm_136.b_0" shape = [1024] dtype = "float32" min_val = float("-1.46039") @@ -3223,6 +3516,7 @@ class Program_weight_tensor_parameter_292: class Program_weight_tensor_parameter_293: name = "parameter_293" + original_name = "layer_norm_136.w_0" shape = [1024] dtype = "float32" min_val = float("0.792921") @@ -3234,6 +3528,7 @@ class Program_weight_tensor_parameter_293: class Program_weight_tensor_parameter_294: name = "parameter_294" + original_name = "linear_400.b_0" shape = [1024] dtype = "float32" min_val = float("-0.268236") @@ -3245,6 +3540,7 @@ class Program_weight_tensor_parameter_294: class Program_weight_tensor_parameter_295: name = "parameter_295" + original_name = "linear_400.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-1.18037") @@ -3256,6 +3552,7 @@ class Program_weight_tensor_parameter_295: class Program_weight_tensor_parameter_296: name = "parameter_296" + original_name = "linear_399.b_0" shape = [4096] dtype = "float32" min_val = float("-0.160545") @@ -3267,6 +3564,7 @@ class Program_weight_tensor_parameter_296: class Program_weight_tensor_parameter_297: name = "parameter_297" + original_name = "linear_399.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.288028") @@ -3278,6 +3576,7 @@ class Program_weight_tensor_parameter_297: class Program_weight_tensor_parameter_298: name = "parameter_298" + original_name = "linear_398.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0821837") @@ -3289,6 +3588,7 @@ class Program_weight_tensor_parameter_298: class Program_weight_tensor_parameter_299: name = "parameter_299" + original_name = "linear_398.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.202733") @@ -3300,6 +3600,7 @@ class Program_weight_tensor_parameter_299: class Program_weight_tensor_parameter_300: name = "parameter_300" + original_name = "linear_397.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0594004") @@ -3311,6 +3612,7 @@ class Program_weight_tensor_parameter_300: class Program_weight_tensor_parameter_301: name = "parameter_301" + original_name = "linear_397.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.143204") @@ -3322,6 +3624,7 @@ class Program_weight_tensor_parameter_301: class Program_weight_tensor_parameter_302: name = "parameter_302" + original_name = "linear_396.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0132827") @@ -3333,6 +3636,7 @@ class Program_weight_tensor_parameter_302: class Program_weight_tensor_parameter_303: name = "parameter_303" + original_name = "linear_396.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.247768") @@ -3344,6 +3648,7 @@ class Program_weight_tensor_parameter_303: class Program_weight_tensor_parameter_304: name = "parameter_304" + original_name = "linear_395.b_0" shape = [1024] dtype = "float32" min_val = float("-0.272052") @@ -3355,6 +3660,7 @@ class Program_weight_tensor_parameter_304: class Program_weight_tensor_parameter_305: name = "parameter_305" + original_name = "linear_395.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.262866") @@ -3366,6 +3672,7 @@ class Program_weight_tensor_parameter_305: class Program_weight_tensor_parameter_306: name = "parameter_306" + original_name = "layer_norm_135.b_0" shape = [1024] dtype = "float32" min_val = float("-0.27981") @@ -3377,6 +3684,7 @@ class Program_weight_tensor_parameter_306: class Program_weight_tensor_parameter_307: name = "parameter_307" + original_name = "layer_norm_135.w_0" shape = [1024] dtype = "float32" min_val = float("0.394217") @@ -3388,6 +3696,7 @@ class Program_weight_tensor_parameter_307: class Program_weight_tensor_parameter_308: name = "parameter_308" + original_name = "layer_norm_134.b_0" shape = [1024] dtype = "float32" min_val = float("-1.68968") @@ -3399,6 +3708,7 @@ class Program_weight_tensor_parameter_308: class Program_weight_tensor_parameter_309: name = "parameter_309" + original_name = "layer_norm_134.w_0" shape = [1024] dtype = "float32" min_val = float("0.663221") @@ -3410,6 +3720,7 @@ class Program_weight_tensor_parameter_309: class Program_weight_tensor_parameter_310: name = "parameter_310" + original_name = "linear_394.b_0" shape = [1024] dtype = "float32" min_val = float("-0.147726") @@ -3421,6 +3732,7 @@ class Program_weight_tensor_parameter_310: class Program_weight_tensor_parameter_311: name = "parameter_311" + original_name = "linear_394.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-1.40022") @@ -3432,6 +3744,7 @@ class Program_weight_tensor_parameter_311: class Program_weight_tensor_parameter_312: name = "parameter_312" + original_name = "linear_393.b_0" shape = [4096] dtype = "float32" min_val = float("-0.139311") @@ -3443,6 +3756,7 @@ class Program_weight_tensor_parameter_312: class Program_weight_tensor_parameter_313: name = "parameter_313" + original_name = "linear_393.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.326428") @@ -3454,6 +3768,7 @@ class Program_weight_tensor_parameter_313: class Program_weight_tensor_parameter_314: name = "parameter_314" + original_name = "linear_392.b_0" shape = [1024] dtype = "float32" min_val = float("-0.309887") @@ -3465,6 +3780,7 @@ class Program_weight_tensor_parameter_314: class Program_weight_tensor_parameter_315: name = "parameter_315" + original_name = "linear_392.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.34262") @@ -3476,6 +3792,7 @@ class Program_weight_tensor_parameter_315: class Program_weight_tensor_parameter_316: name = "parameter_316" + original_name = "linear_391.b_0" shape = [1024] dtype = "float32" min_val = float("-0.151588") @@ -3487,6 +3804,7 @@ class Program_weight_tensor_parameter_316: class Program_weight_tensor_parameter_317: name = "parameter_317" + original_name = "linear_391.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.144687") @@ -3498,6 +3816,7 @@ class Program_weight_tensor_parameter_317: class Program_weight_tensor_parameter_318: name = "parameter_318" + original_name = "linear_390.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0325759") @@ -3509,6 +3828,7 @@ class Program_weight_tensor_parameter_318: class Program_weight_tensor_parameter_319: name = "parameter_319" + original_name = "linear_390.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.291394") @@ -3520,6 +3840,7 @@ class Program_weight_tensor_parameter_319: class Program_weight_tensor_parameter_320: name = "parameter_320" + original_name = "linear_389.b_0" shape = [1024] dtype = "float32" min_val = float("-0.382438") @@ -3531,6 +3852,7 @@ class Program_weight_tensor_parameter_320: class Program_weight_tensor_parameter_321: name = "parameter_321" + original_name = "linear_389.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.26093") @@ -3542,6 +3864,7 @@ class Program_weight_tensor_parameter_321: class Program_weight_tensor_parameter_322: name = "parameter_322" + original_name = "layer_norm_133.b_0" shape = [1024] dtype = "float32" min_val = float("-0.403027") @@ -3553,6 +3876,7 @@ class Program_weight_tensor_parameter_322: class Program_weight_tensor_parameter_323: name = "parameter_323" + original_name = "layer_norm_133.w_0" shape = [1024] dtype = "float32" min_val = float("0.28315") @@ -3564,6 +3888,7 @@ class Program_weight_tensor_parameter_323: class Program_weight_tensor_parameter_324: name = "parameter_324" + original_name = "layer_norm_132.b_0" shape = [1024] dtype = "float32" min_val = float("-1.74371") @@ -3575,6 +3900,7 @@ class Program_weight_tensor_parameter_324: class Program_weight_tensor_parameter_325: name = "parameter_325" + original_name = "layer_norm_132.w_0" shape = [1024] dtype = "float32" min_val = float("0.789938") @@ -3586,6 +3912,7 @@ class Program_weight_tensor_parameter_325: class Program_weight_tensor_parameter_326: name = "parameter_326" + original_name = "linear_388.b_0" shape = [1024] dtype = "float32" min_val = float("-0.232719") @@ -3597,6 +3924,7 @@ class Program_weight_tensor_parameter_326: class Program_weight_tensor_parameter_327: name = "parameter_327" + original_name = "linear_388.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-1.2481") @@ -3608,6 +3936,7 @@ class Program_weight_tensor_parameter_327: class Program_weight_tensor_parameter_328: name = "parameter_328" + original_name = "linear_387.b_0" shape = [4096] dtype = "float32" min_val = float("-0.142344") @@ -3619,6 +3948,7 @@ class Program_weight_tensor_parameter_328: class Program_weight_tensor_parameter_329: name = "parameter_329" + original_name = "linear_387.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.327719") @@ -3630,6 +3960,7 @@ class Program_weight_tensor_parameter_329: class Program_weight_tensor_parameter_330: name = "parameter_330" + original_name = "linear_386.b_0" shape = [1024] dtype = "float32" min_val = float("-0.369256") @@ -3641,6 +3972,7 @@ class Program_weight_tensor_parameter_330: class Program_weight_tensor_parameter_331: name = "parameter_331" + original_name = "linear_386.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.309874") @@ -3652,6 +3984,7 @@ class Program_weight_tensor_parameter_331: class Program_weight_tensor_parameter_332: name = "parameter_332" + original_name = "linear_385.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0965048") @@ -3663,6 +3996,7 @@ class Program_weight_tensor_parameter_332: class Program_weight_tensor_parameter_333: name = "parameter_333" + original_name = "linear_385.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.116724") @@ -3674,6 +4008,7 @@ class Program_weight_tensor_parameter_333: class Program_weight_tensor_parameter_334: name = "parameter_334" + original_name = "linear_384.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0181679") @@ -3685,6 +4020,7 @@ class Program_weight_tensor_parameter_334: class Program_weight_tensor_parameter_335: name = "parameter_335" + original_name = "linear_384.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.324206") @@ -3696,6 +4032,7 @@ class Program_weight_tensor_parameter_335: class Program_weight_tensor_parameter_336: name = "parameter_336" + original_name = "linear_383.b_0" shape = [1024] dtype = "float32" min_val = float("-0.245357") @@ -3707,6 +4044,7 @@ class Program_weight_tensor_parameter_336: class Program_weight_tensor_parameter_337: name = "parameter_337" + original_name = "linear_383.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.29031") @@ -3718,6 +4056,7 @@ class Program_weight_tensor_parameter_337: class Program_weight_tensor_parameter_338: name = "parameter_338" + original_name = "layer_norm_131.b_0" shape = [1024] dtype = "float32" min_val = float("-0.624823") @@ -3729,6 +4068,7 @@ class Program_weight_tensor_parameter_338: class Program_weight_tensor_parameter_339: name = "parameter_339" + original_name = "layer_norm_131.w_0" shape = [1024] dtype = "float32" min_val = float("0.253731") @@ -3740,6 +4080,7 @@ class Program_weight_tensor_parameter_339: class Program_weight_tensor_parameter_340: name = "parameter_340" + original_name = "layer_norm_130.b_0" shape = [1024] dtype = "float32" min_val = float("-1.79613") @@ -3751,6 +4092,7 @@ class Program_weight_tensor_parameter_340: class Program_weight_tensor_parameter_341: name = "parameter_341" + original_name = "layer_norm_130.w_0" shape = [1024] dtype = "float32" min_val = float("0.798003") @@ -3762,6 +4104,7 @@ class Program_weight_tensor_parameter_341: class Program_weight_tensor_parameter_342: name = "parameter_342" + original_name = "linear_382.b_0" shape = [1024] dtype = "float32" min_val = float("-0.457782") @@ -3773,6 +4116,7 @@ class Program_weight_tensor_parameter_342: class Program_weight_tensor_parameter_343: name = "parameter_343" + original_name = "linear_382.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-1.23703") @@ -3784,6 +4128,7 @@ class Program_weight_tensor_parameter_343: class Program_weight_tensor_parameter_344: name = "parameter_344" + original_name = "linear_381.b_0" shape = [4096] dtype = "float32" min_val = float("-0.132738") @@ -3795,6 +4140,7 @@ class Program_weight_tensor_parameter_344: class Program_weight_tensor_parameter_345: name = "parameter_345" + original_name = "linear_381.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.234272") @@ -3806,6 +4152,7 @@ class Program_weight_tensor_parameter_345: class Program_weight_tensor_parameter_346: name = "parameter_346" + original_name = "linear_380.b_0" shape = [1024] dtype = "float32" min_val = float("-0.218494") @@ -3817,6 +4164,7 @@ class Program_weight_tensor_parameter_346: class Program_weight_tensor_parameter_347: name = "parameter_347" + original_name = "linear_380.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.861062") @@ -3828,6 +4176,7 @@ class Program_weight_tensor_parameter_347: class Program_weight_tensor_parameter_348: name = "parameter_348" + original_name = "linear_379.b_0" shape = [1024] dtype = "float32" min_val = float("-0.037819") @@ -3839,6 +4188,7 @@ class Program_weight_tensor_parameter_348: class Program_weight_tensor_parameter_349: name = "parameter_349" + original_name = "linear_379.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.15871") @@ -3850,6 +4200,7 @@ class Program_weight_tensor_parameter_349: class Program_weight_tensor_parameter_350: name = "parameter_350" + original_name = "linear_378.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0131982") @@ -3861,6 +4212,7 @@ class Program_weight_tensor_parameter_350: class Program_weight_tensor_parameter_351: name = "parameter_351" + original_name = "linear_378.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.21236") @@ -3872,6 +4224,7 @@ class Program_weight_tensor_parameter_351: class Program_weight_tensor_parameter_352: name = "parameter_352" + original_name = "linear_377.b_0" shape = [1024] dtype = "float32" min_val = float("-0.329922") @@ -3883,6 +4236,7 @@ class Program_weight_tensor_parameter_352: class Program_weight_tensor_parameter_353: name = "parameter_353" + original_name = "linear_377.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.251353") @@ -3894,6 +4248,7 @@ class Program_weight_tensor_parameter_353: class Program_weight_tensor_parameter_354: name = "parameter_354" + original_name = "layer_norm_129.b_0" shape = [1024] dtype = "float32" min_val = float("-0.722776") @@ -3905,6 +4260,7 @@ class Program_weight_tensor_parameter_354: class Program_weight_tensor_parameter_355: name = "parameter_355" + original_name = "layer_norm_129.w_0" shape = [1024] dtype = "float32" min_val = float("0.362818") @@ -3916,6 +4272,7 @@ class Program_weight_tensor_parameter_355: class Program_weight_tensor_parameter_356: name = "parameter_356" + original_name = "layer_norm_128.b_0" shape = [1024] dtype = "float32" min_val = float("-2.16862") @@ -3927,6 +4284,7 @@ class Program_weight_tensor_parameter_356: class Program_weight_tensor_parameter_357: name = "parameter_357" + original_name = "layer_norm_128.w_0" shape = [1024] dtype = "float32" min_val = float("0.867525") @@ -3938,6 +4296,7 @@ class Program_weight_tensor_parameter_357: class Program_weight_tensor_parameter_358: name = "parameter_358" + original_name = "linear_376.b_0" shape = [1024] dtype = "float32" min_val = float("-0.381036") @@ -3949,6 +4308,7 @@ class Program_weight_tensor_parameter_358: class Program_weight_tensor_parameter_359: name = "parameter_359" + original_name = "linear_376.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-2.28357") @@ -3960,6 +4320,7 @@ class Program_weight_tensor_parameter_359: class Program_weight_tensor_parameter_360: name = "parameter_360" + original_name = "linear_375.b_0" shape = [4096] dtype = "float32" min_val = float("-0.107778") @@ -3971,6 +4332,7 @@ class Program_weight_tensor_parameter_360: class Program_weight_tensor_parameter_361: name = "parameter_361" + original_name = "linear_375.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.29725") @@ -3982,6 +4344,7 @@ class Program_weight_tensor_parameter_361: class Program_weight_tensor_parameter_362: name = "parameter_362" + original_name = "linear_374.b_0" shape = [1024] dtype = "float32" min_val = float("-0.319368") @@ -3993,6 +4356,7 @@ class Program_weight_tensor_parameter_362: class Program_weight_tensor_parameter_363: name = "parameter_363" + original_name = "linear_374.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.842125") @@ -4004,6 +4368,7 @@ class Program_weight_tensor_parameter_363: class Program_weight_tensor_parameter_364: name = "parameter_364" + original_name = "linear_373.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0387209") @@ -4015,6 +4380,7 @@ class Program_weight_tensor_parameter_364: class Program_weight_tensor_parameter_365: name = "parameter_365" + original_name = "linear_373.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.148741") @@ -4026,6 +4392,7 @@ class Program_weight_tensor_parameter_365: class Program_weight_tensor_parameter_366: name = "parameter_366" + original_name = "linear_372.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0115399") @@ -4037,6 +4404,7 @@ class Program_weight_tensor_parameter_366: class Program_weight_tensor_parameter_367: name = "parameter_367" + original_name = "linear_372.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.349842") @@ -4048,6 +4416,7 @@ class Program_weight_tensor_parameter_367: class Program_weight_tensor_parameter_368: name = "parameter_368" + original_name = "linear_371.b_0" shape = [1024] dtype = "float32" min_val = float("-0.288008") @@ -4059,6 +4428,7 @@ class Program_weight_tensor_parameter_368: class Program_weight_tensor_parameter_369: name = "parameter_369" + original_name = "linear_371.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.355076") @@ -4070,6 +4440,7 @@ class Program_weight_tensor_parameter_369: class Program_weight_tensor_parameter_370: name = "parameter_370" + original_name = "layer_norm_127.b_0" shape = [1024] dtype = "float32" min_val = float("-1.05409") @@ -4081,6 +4452,7 @@ class Program_weight_tensor_parameter_370: class Program_weight_tensor_parameter_371: name = "parameter_371" + original_name = "layer_norm_127.w_0" shape = [1024] dtype = "float32" min_val = float("0.266859") @@ -4092,6 +4464,7 @@ class Program_weight_tensor_parameter_371: class Program_weight_tensor_parameter_372: name = "parameter_372" + original_name = "layer_norm_126.b_0" shape = [1024] dtype = "float32" min_val = float("-3.75639") @@ -4103,6 +4476,7 @@ class Program_weight_tensor_parameter_372: class Program_weight_tensor_parameter_373: name = "parameter_373" + original_name = "layer_norm_126.w_0" shape = [1024] dtype = "float32" min_val = float("0.710459") @@ -4114,6 +4488,7 @@ class Program_weight_tensor_parameter_373: class Program_weight_tensor_parameter_374: name = "parameter_374" + original_name = "linear_370.b_0" shape = [1024] dtype = "float32" min_val = float("-0.548768") @@ -4125,6 +4500,7 @@ class Program_weight_tensor_parameter_374: class Program_weight_tensor_parameter_375: name = "parameter_375" + original_name = "linear_370.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-2.85999") @@ -4136,6 +4512,7 @@ class Program_weight_tensor_parameter_375: class Program_weight_tensor_parameter_376: name = "parameter_376" + original_name = "linear_369.b_0" shape = [4096] dtype = "float32" min_val = float("-0.217368") @@ -4147,6 +4524,7 @@ class Program_weight_tensor_parameter_376: class Program_weight_tensor_parameter_377: name = "parameter_377" + original_name = "linear_369.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.365614") @@ -4158,6 +4536,7 @@ class Program_weight_tensor_parameter_377: class Program_weight_tensor_parameter_378: name = "parameter_378" + original_name = "linear_368.b_0" shape = [1024] dtype = "float32" min_val = float("-0.634452") @@ -4169,6 +4548,7 @@ class Program_weight_tensor_parameter_378: class Program_weight_tensor_parameter_379: name = "parameter_379" + original_name = "linear_368.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.470766") @@ -4180,6 +4560,7 @@ class Program_weight_tensor_parameter_379: class Program_weight_tensor_parameter_380: name = "parameter_380" + original_name = "linear_367.b_0" shape = [1024] dtype = "float32" min_val = float("-0.138042") @@ -4191,6 +4572,7 @@ class Program_weight_tensor_parameter_380: class Program_weight_tensor_parameter_381: name = "parameter_381" + original_name = "linear_367.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.199258") @@ -4202,6 +4584,7 @@ class Program_weight_tensor_parameter_381: class Program_weight_tensor_parameter_382: name = "parameter_382" + original_name = "linear_366.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0179152") @@ -4213,6 +4596,7 @@ class Program_weight_tensor_parameter_382: class Program_weight_tensor_parameter_383: name = "parameter_383" + original_name = "linear_366.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.55046") @@ -4224,6 +4608,7 @@ class Program_weight_tensor_parameter_383: class Program_weight_tensor_parameter_384: name = "parameter_384" + original_name = "linear_365.b_0" shape = [1024] dtype = "float32" min_val = float("-0.582036") @@ -4235,6 +4620,7 @@ class Program_weight_tensor_parameter_384: class Program_weight_tensor_parameter_385: name = "parameter_385" + original_name = "linear_365.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.352677") @@ -4246,6 +4632,7 @@ class Program_weight_tensor_parameter_385: class Program_weight_tensor_parameter_386: name = "parameter_386" + original_name = "layer_norm_125.b_0" shape = [1024] dtype = "float32" min_val = float("-0.206363") @@ -4257,6 +4644,7 @@ class Program_weight_tensor_parameter_386: class Program_weight_tensor_parameter_387: name = "parameter_387" + original_name = "layer_norm_125.w_0" shape = [1024] dtype = "float32" min_val = float("0.0609985") @@ -4268,6 +4656,7 @@ class Program_weight_tensor_parameter_387: class Program_weight_tensor_parameter_388: name = "parameter_388" + original_name = "embedding_17.w_0" shape = [2, 1024] dtype = "float32" min_val = float("-0.774461") @@ -4279,6 +4668,7 @@ class Program_weight_tensor_parameter_388: class Program_weight_tensor_parameter_389: name = "parameter_389" + original_name = "embedding_16.w_0" shape = [512, 1024] dtype = "float32" min_val = float("-0.817655") @@ -4290,6 +4680,7 @@ class Program_weight_tensor_parameter_389: class Program_weight_tensor_parameter_390: name = "parameter_390" + original_name = "embedding_15.w_0" shape = [28996, 1024] dtype = "float32" min_val = float("-1.45525") diff --git a/paddle_samples/PaddleNLP/ernie-1.0/input_meta.py b/paddle_samples/PaddleNLP/ernie-1.0/input_meta.py index 2ad3baf05..4ec838a4a 100644 --- a/paddle_samples/PaddleNLP/ernie-1.0/input_meta.py +++ b/paddle_samples/PaddleNLP/ernie-1.0/input_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_data_0: name = "data_0" + original_name = "input_ids" shape = [1, 26] dtype = "int64" data = [ @@ -34,6 +35,7 @@ class Program_weight_tensor_data_0: class Program_weight_tensor_data_1: name = "data_1" + original_name = "token_type_ids" shape = [1, 26] dtype = "int64" data = [ diff --git a/paddle_samples/PaddleNLP/ernie-1.0/model.py b/paddle_samples/PaddleNLP/ernie-1.0/model.py index 229d4cc02..9de277d74 100644 --- a/paddle_samples/PaddleNLP/ernie-1.0/model.py +++ b/paddle_samples/PaddleNLP/ernie-1.0/model.py @@ -265,10 +265,11 @@ def forward( # pd_op.embedding: (1x26x768xf32) <- (1x26xi64, 513x768xf32) embedding_1 = paddle._C_ops.embedding(subtract_0, parameter_197, -1, False) - del parameter_197 + del parameter_197, subtract_0 # pd_op.add: (1x26x768xf32) <- (1x26x768xf32, 1x26x768xf32) add_0 = paddle._C_ops.add(embedding_0, embedding_1) + del embedding_0, embedding_1 # pd_op.embedding: (1x26x768xf32) <- (1x26xi64, 2x768xf32) embedding_2 = paddle._C_ops.embedding(data_1, parameter_196, -1, False) @@ -276,6 +277,7 @@ def forward( # pd_op.add: (1x26x768xf32) <- (1x26x768xf32, 1x26x768xf32) add_1 = paddle._C_ops.add(add_0, embedding_2) + del add_0, embedding_2 # pd_op.layer_norm: (1x26x768xf32, 1x26xf32, 1x26xf32) <- (1x26x768xf32, 768xf32, 768xf32) layer_norm_0, layer_norm_1, layer_norm_2 = (lambda x, f: f(x))( @@ -284,125 +286,17 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_194, parameter_195 + del add_1, parameter_194, parameter_195 # pd_op.full: (1xf32) <- () full_4 = paddle._C_ops.full( [1], float("0.1"), paddle.float32, paddle.core.CPUPlace() ) - # pd_op.assign: (1xf32) <- (1xf32) - assign_0 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_1 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_2 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_3 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_4 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_5 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_6 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_7 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_8 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_9 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_10 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_11 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_12 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_13 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_14 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_15 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_16 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_17 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_18 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_19 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_20 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_21 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_22 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_23 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_24 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_25 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_26 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_27 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_28 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_29 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_30 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_31 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_32 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_33 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_34 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_35 = full_4 - # pd_op.dropout: (1x26x768xf32, 1x26x768xui8) <- (1x26x768xf32, None, 1xf32) dropout_0, dropout_1 = (lambda x, f: f(x))( paddle._C_ops.dropout( - layer_norm_0, None, full_4, False, "upscale_in_train", 0, False + layer_norm_0, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -414,13 +308,14 @@ def forward( # pd_op.add: (1x26x768xf32) <- (1x26x768xf32, 768xf32) add_2 = paddle._C_ops.add(matmul_0, parameter_192) - del parameter_192 + del matmul_0, parameter_192 # pd_op.full_int_array: (4xi64) <- () full_int_array_1 = [0, 0, 12, 64] # pd_op.reshape: (1x26x12x64xf32) <- (1x26x768xf32, 4xi64) reshape_0 = paddle._C_ops.reshape(add_2, full_int_array_1) + del add_2 # pd_op.transpose: (1x12x26x64xf32) <- (1x26x12x64xf32) transpose_0 = paddle._C_ops.transpose(reshape_0, [0, 2, 1, 3]) @@ -432,7 +327,7 @@ def forward( # pd_op.add: (1x26x768xf32) <- (1x26x768xf32, 768xf32) add_3 = paddle._C_ops.add(matmul_1, parameter_190) - del parameter_190 + del matmul_1, parameter_190 # pd_op.matmul: (1x26x768xf32) <- (1x26x768xf32, 768x768xf32) matmul_2 = paddle._C_ops.matmul(dropout_0, parameter_189, False, False) @@ -440,10 +335,11 @@ def forward( # pd_op.add: (1x26x768xf32) <- (1x26x768xf32, 768xf32) add_4 = paddle._C_ops.add(matmul_2, parameter_188) - del parameter_188 + del matmul_2, parameter_188 # pd_op.reshape: (1x26x12x64xf32) <- (1x26x768xf32, 4xi64) reshape_1 = paddle._C_ops.reshape(add_3, full_int_array_1) + del add_3 # pd_op.transpose: (1x12x26x64xf32) <- (1x26x12x64xf32) transpose_1 = paddle._C_ops.transpose(reshape_1, [0, 2, 1, 3]) @@ -451,6 +347,7 @@ def forward( # pd_op.reshape: (1x26x12x64xf32) <- (1x26x768xf32, 4xi64) reshape_2 = paddle._C_ops.reshape(add_4, full_int_array_1) + del add_4 # pd_op.transpose: (1x12x26x64xf32) <- (1x26x12x64xf32) transpose_2 = paddle._C_ops.transpose(reshape_2, [0, 2, 1, 3]) @@ -461,48 +358,17 @@ def forward( [1], float("0.125"), paddle.float32, paddle.core.CPUPlace() ) - # pd_op.assign: (1xf32) <- (1xf32) - assign_36 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_37 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_38 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_39 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_40 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_41 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_42 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_43 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_44 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_45 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_46 = full_5 - # pd_op.scale: (1x12x26x64xf32) <- (1x12x26x64xf32, 1xf32) scale_1 = paddle._C_ops.scale(transpose_0, full_5, float("0"), True) del transpose_0 # pd_op.matmul: (1x12x26x26xf32) <- (1x12x26x64xf32, 1x12x26x64xf32) matmul_3 = paddle._C_ops.matmul(scale_1, transpose_1, False, True) + del scale_1, transpose_1 # pd_op.add: (1x12x26x26xf32) <- (1x12x26x26xf32, 1x1x1x26xf32) add_5 = paddle._C_ops.add(matmul_3, unsqueeze_0) + del matmul_3 # pd_op.softmax: (1x12x26x26xf32) <- (1x12x26x26xf32) softmax_0 = paddle._C_ops.softmax(add_5, -1) @@ -511,13 +377,15 @@ def forward( # pd_op.dropout: (1x12x26x26xf32, 1x12x26x26xui8) <- (1x12x26x26xf32, None, 1xf32) dropout_2, dropout_3 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_0, None, full_4, False, "upscale_in_train", 0, False + softmax_0, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_0 # pd_op.matmul: (1x12x26x64xf32) <- (1x12x26x26xf32, 1x12x26x64xf32) matmul_4 = paddle._C_ops.matmul(dropout_2, transpose_2, False, False) + del dropout_2, transpose_2 # pd_op.transpose: (1x26x12x64xf32) <- (1x12x26x64xf32) transpose_3 = paddle._C_ops.transpose(matmul_4, [0, 2, 1, 3]) @@ -528,19 +396,20 @@ def forward( # pd_op.reshape: (1x26x768xf32) <- (1x26x12x64xf32, 3xi64) reshape_3 = paddle._C_ops.reshape(transpose_3, full_int_array_2) + del transpose_3 # pd_op.matmul: (1x26x768xf32) <- (1x26x768xf32, 768x768xf32) matmul_5 = paddle._C_ops.matmul(reshape_3, parameter_187, False, False) - del parameter_187 + del parameter_187, reshape_3 # pd_op.add: (1x26x768xf32) <- (1x26x768xf32, 768xf32) add_6 = paddle._C_ops.add(matmul_5, parameter_186) - del parameter_186 + del matmul_5, parameter_186 # pd_op.dropout: (1x26x768xf32, 1x26x768xui8) <- (1x26x768xf32, None, 1xf32) dropout_4, dropout_5 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_6, None, full_4, False, "upscale_in_train", 0, False + add_6, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -548,6 +417,7 @@ def forward( # pd_op.add: (1x26x768xf32) <- (1x26x768xf32, 1x26x768xf32) add_7 = paddle._C_ops.add(dropout_0, dropout_4) + del dropout_0, dropout_4 # pd_op.layer_norm: (1x26x768xf32, 1x26xf32, 1x26xf32) <- (1x26x768xf32, 768xf32, 768xf32) layer_norm_3, layer_norm_4, layer_norm_5 = (lambda x, f: f(x))( @@ -556,7 +426,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_180, parameter_181 + del add_7, parameter_180, parameter_181 # pd_op.matmul: (1x26x3072xf32) <- (1x26x768xf32, 768x3072xf32) matmul_6 = paddle._C_ops.matmul(layer_norm_3, parameter_185, False, False) @@ -564,7 +434,7 @@ def forward( # pd_op.add: (1x26x3072xf32) <- (1x26x3072xf32, 3072xf32) add_8 = paddle._C_ops.add(matmul_6, parameter_184) - del parameter_184 + del matmul_6, parameter_184 # pd_op.relu: (1x26x3072xf32) <- (1x26x3072xf32) relu_0 = paddle._C_ops.relu(add_8) @@ -572,16 +442,16 @@ def forward( # pd_op.matmul: (1x26x768xf32) <- (1x26x3072xf32, 3072x768xf32) matmul_7 = paddle._C_ops.matmul(relu_0, parameter_183, False, False) - del parameter_183 + del parameter_183, relu_0 # pd_op.add: (1x26x768xf32) <- (1x26x768xf32, 768xf32) add_9 = paddle._C_ops.add(matmul_7, parameter_182) - del parameter_182 + del matmul_7, parameter_182 # pd_op.dropout: (1x26x768xf32, 1x26x768xui8) <- (1x26x768xf32, None, 1xf32) dropout_6, dropout_7 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_9, None, full_4, False, "upscale_in_train", 0, False + add_9, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -589,6 +459,7 @@ def forward( # pd_op.add: (1x26x768xf32) <- (1x26x768xf32, 1x26x768xf32) add_10 = paddle._C_ops.add(layer_norm_3, dropout_6) + del dropout_6, layer_norm_3 # pd_op.layer_norm: (1x26x768xf32, 1x26xf32, 1x26xf32) <- (1x26x768xf32, 768xf32, 768xf32) layer_norm_6, layer_norm_7, layer_norm_8 = (lambda x, f: f(x))( @@ -597,7 +468,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_178, parameter_179 + del add_10, parameter_178, parameter_179 # pd_op.matmul: (1x26x768xf32) <- (1x26x768xf32, 768x768xf32) matmul_8 = paddle._C_ops.matmul(layer_norm_6, parameter_177, False, False) @@ -605,10 +476,11 @@ def forward( # pd_op.add: (1x26x768xf32) <- (1x26x768xf32, 768xf32) add_11 = paddle._C_ops.add(matmul_8, parameter_176) - del parameter_176 + del matmul_8, parameter_176 # pd_op.reshape: (1x26x12x64xf32) <- (1x26x768xf32, 4xi64) reshape_4 = paddle._C_ops.reshape(add_11, full_int_array_1) + del add_11 # pd_op.transpose: (1x12x26x64xf32) <- (1x26x12x64xf32) transpose_4 = paddle._C_ops.transpose(reshape_4, [0, 2, 1, 3]) @@ -620,7 +492,7 @@ def forward( # pd_op.add: (1x26x768xf32) <- (1x26x768xf32, 768xf32) add_12 = paddle._C_ops.add(matmul_9, parameter_174) - del parameter_174 + del matmul_9, parameter_174 # pd_op.matmul: (1x26x768xf32) <- (1x26x768xf32, 768x768xf32) matmul_10 = paddle._C_ops.matmul(layer_norm_6, parameter_173, False, False) @@ -628,10 +500,11 @@ def forward( # pd_op.add: (1x26x768xf32) <- (1x26x768xf32, 768xf32) add_13 = paddle._C_ops.add(matmul_10, parameter_172) - del parameter_172 + del matmul_10, parameter_172 # pd_op.reshape: (1x26x12x64xf32) <- (1x26x768xf32, 4xi64) reshape_5 = paddle._C_ops.reshape(add_12, full_int_array_1) + del add_12 # pd_op.transpose: (1x12x26x64xf32) <- (1x26x12x64xf32) transpose_5 = paddle._C_ops.transpose(reshape_5, [0, 2, 1, 3]) @@ -639,6 +512,7 @@ def forward( # pd_op.reshape: (1x26x12x64xf32) <- (1x26x768xf32, 4xi64) reshape_6 = paddle._C_ops.reshape(add_13, full_int_array_1) + del add_13 # pd_op.transpose: (1x12x26x64xf32) <- (1x26x12x64xf32) transpose_6 = paddle._C_ops.transpose(reshape_6, [0, 2, 1, 3]) @@ -650,9 +524,11 @@ def forward( # pd_op.matmul: (1x12x26x26xf32) <- (1x12x26x64xf32, 1x12x26x64xf32) matmul_11 = paddle._C_ops.matmul(scale_2, transpose_5, False, True) + del scale_2, transpose_5 # pd_op.add: (1x12x26x26xf32) <- (1x12x26x26xf32, 1x1x1x26xf32) add_14 = paddle._C_ops.add(matmul_11, unsqueeze_0) + del matmul_11 # pd_op.softmax: (1x12x26x26xf32) <- (1x12x26x26xf32) softmax_1 = paddle._C_ops.softmax(add_14, -1) @@ -661,13 +537,15 @@ def forward( # pd_op.dropout: (1x12x26x26xf32, 1x12x26x26xui8) <- (1x12x26x26xf32, None, 1xf32) dropout_8, dropout_9 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_1, None, full_4, False, "upscale_in_train", 0, False + softmax_1, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_1 # pd_op.matmul: (1x12x26x64xf32) <- (1x12x26x26xf32, 1x12x26x64xf32) matmul_12 = paddle._C_ops.matmul(dropout_8, transpose_6, False, False) + del dropout_8, transpose_6 # pd_op.transpose: (1x26x12x64xf32) <- (1x12x26x64xf32) transpose_7 = paddle._C_ops.transpose(matmul_12, [0, 2, 1, 3]) @@ -675,19 +553,20 @@ def forward( # pd_op.reshape: (1x26x768xf32) <- (1x26x12x64xf32, 3xi64) reshape_7 = paddle._C_ops.reshape(transpose_7, full_int_array_2) + del transpose_7 # pd_op.matmul: (1x26x768xf32) <- (1x26x768xf32, 768x768xf32) matmul_13 = paddle._C_ops.matmul(reshape_7, parameter_171, False, False) - del parameter_171 + del parameter_171, reshape_7 # pd_op.add: (1x26x768xf32) <- (1x26x768xf32, 768xf32) add_15 = paddle._C_ops.add(matmul_13, parameter_170) - del parameter_170 + del matmul_13, parameter_170 # pd_op.dropout: (1x26x768xf32, 1x26x768xui8) <- (1x26x768xf32, None, 1xf32) dropout_10, dropout_11 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_15, None, full_4, False, "upscale_in_train", 0, False + add_15, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -695,6 +574,7 @@ def forward( # pd_op.add: (1x26x768xf32) <- (1x26x768xf32, 1x26x768xf32) add_16 = paddle._C_ops.add(layer_norm_6, dropout_10) + del dropout_10, layer_norm_6 # pd_op.layer_norm: (1x26x768xf32, 1x26xf32, 1x26xf32) <- (1x26x768xf32, 768xf32, 768xf32) layer_norm_9, layer_norm_10, layer_norm_11 = (lambda x, f: f(x))( @@ -703,7 +583,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_164, parameter_165 + del add_16, parameter_164, parameter_165 # pd_op.matmul: (1x26x3072xf32) <- (1x26x768xf32, 768x3072xf32) matmul_14 = paddle._C_ops.matmul(layer_norm_9, parameter_169, False, False) @@ -711,7 +591,7 @@ def forward( # pd_op.add: (1x26x3072xf32) <- (1x26x3072xf32, 3072xf32) add_17 = paddle._C_ops.add(matmul_14, parameter_168) - del parameter_168 + del matmul_14, parameter_168 # pd_op.relu: (1x26x3072xf32) <- (1x26x3072xf32) relu_1 = paddle._C_ops.relu(add_17) @@ -719,16 +599,16 @@ def forward( # pd_op.matmul: (1x26x768xf32) <- (1x26x3072xf32, 3072x768xf32) matmul_15 = paddle._C_ops.matmul(relu_1, parameter_167, False, False) - del parameter_167 + del parameter_167, relu_1 # pd_op.add: (1x26x768xf32) <- (1x26x768xf32, 768xf32) add_18 = paddle._C_ops.add(matmul_15, parameter_166) - del parameter_166 + del matmul_15, parameter_166 # pd_op.dropout: (1x26x768xf32, 1x26x768xui8) <- (1x26x768xf32, None, 1xf32) dropout_12, dropout_13 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_18, None, full_4, False, "upscale_in_train", 0, False + add_18, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -736,6 +616,7 @@ def forward( # pd_op.add: (1x26x768xf32) <- (1x26x768xf32, 1x26x768xf32) add_19 = paddle._C_ops.add(layer_norm_9, dropout_12) + del dropout_12, layer_norm_9 # pd_op.layer_norm: (1x26x768xf32, 1x26xf32, 1x26xf32) <- (1x26x768xf32, 768xf32, 768xf32) layer_norm_12, layer_norm_13, layer_norm_14 = (lambda x, f: f(x))( @@ -744,7 +625,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_162, parameter_163 + del add_19, parameter_162, parameter_163 # pd_op.matmul: (1x26x768xf32) <- (1x26x768xf32, 768x768xf32) matmul_16 = paddle._C_ops.matmul(layer_norm_12, parameter_161, False, False) @@ -752,10 +633,11 @@ def forward( # pd_op.add: (1x26x768xf32) <- (1x26x768xf32, 768xf32) add_20 = paddle._C_ops.add(matmul_16, parameter_160) - del parameter_160 + del matmul_16, parameter_160 # pd_op.reshape: (1x26x12x64xf32) <- (1x26x768xf32, 4xi64) reshape_8 = paddle._C_ops.reshape(add_20, full_int_array_1) + del add_20 # pd_op.transpose: (1x12x26x64xf32) <- (1x26x12x64xf32) transpose_8 = paddle._C_ops.transpose(reshape_8, [0, 2, 1, 3]) @@ -767,7 +649,7 @@ def forward( # pd_op.add: (1x26x768xf32) <- (1x26x768xf32, 768xf32) add_21 = paddle._C_ops.add(matmul_17, parameter_158) - del parameter_158 + del matmul_17, parameter_158 # pd_op.matmul: (1x26x768xf32) <- (1x26x768xf32, 768x768xf32) matmul_18 = paddle._C_ops.matmul(layer_norm_12, parameter_157, False, False) @@ -775,10 +657,11 @@ def forward( # pd_op.add: (1x26x768xf32) <- (1x26x768xf32, 768xf32) add_22 = paddle._C_ops.add(matmul_18, parameter_156) - del parameter_156 + del matmul_18, parameter_156 # pd_op.reshape: (1x26x12x64xf32) <- (1x26x768xf32, 4xi64) reshape_9 = paddle._C_ops.reshape(add_21, full_int_array_1) + del add_21 # pd_op.transpose: (1x12x26x64xf32) <- (1x26x12x64xf32) transpose_9 = paddle._C_ops.transpose(reshape_9, [0, 2, 1, 3]) @@ -786,6 +669,7 @@ def forward( # pd_op.reshape: (1x26x12x64xf32) <- (1x26x768xf32, 4xi64) reshape_10 = paddle._C_ops.reshape(add_22, full_int_array_1) + del add_22 # pd_op.transpose: (1x12x26x64xf32) <- (1x26x12x64xf32) transpose_10 = paddle._C_ops.transpose(reshape_10, [0, 2, 1, 3]) @@ -797,9 +681,11 @@ def forward( # pd_op.matmul: (1x12x26x26xf32) <- (1x12x26x64xf32, 1x12x26x64xf32) matmul_19 = paddle._C_ops.matmul(scale_3, transpose_9, False, True) + del scale_3, transpose_9 # pd_op.add: (1x12x26x26xf32) <- (1x12x26x26xf32, 1x1x1x26xf32) add_23 = paddle._C_ops.add(matmul_19, unsqueeze_0) + del matmul_19 # pd_op.softmax: (1x12x26x26xf32) <- (1x12x26x26xf32) softmax_2 = paddle._C_ops.softmax(add_23, -1) @@ -808,13 +694,15 @@ def forward( # pd_op.dropout: (1x12x26x26xf32, 1x12x26x26xui8) <- (1x12x26x26xf32, None, 1xf32) dropout_14, dropout_15 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_2, None, full_4, False, "upscale_in_train", 0, False + softmax_2, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_2 # pd_op.matmul: (1x12x26x64xf32) <- (1x12x26x26xf32, 1x12x26x64xf32) matmul_20 = paddle._C_ops.matmul(dropout_14, transpose_10, False, False) + del dropout_14, transpose_10 # pd_op.transpose: (1x26x12x64xf32) <- (1x12x26x64xf32) transpose_11 = paddle._C_ops.transpose(matmul_20, [0, 2, 1, 3]) @@ -822,19 +710,20 @@ def forward( # pd_op.reshape: (1x26x768xf32) <- (1x26x12x64xf32, 3xi64) reshape_11 = paddle._C_ops.reshape(transpose_11, full_int_array_2) + del transpose_11 # pd_op.matmul: (1x26x768xf32) <- (1x26x768xf32, 768x768xf32) matmul_21 = paddle._C_ops.matmul(reshape_11, parameter_155, False, False) - del parameter_155 + del parameter_155, reshape_11 # pd_op.add: (1x26x768xf32) <- (1x26x768xf32, 768xf32) add_24 = paddle._C_ops.add(matmul_21, parameter_154) - del parameter_154 + del matmul_21, parameter_154 # pd_op.dropout: (1x26x768xf32, 1x26x768xui8) <- (1x26x768xf32, None, 1xf32) dropout_16, dropout_17 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_24, None, full_4, False, "upscale_in_train", 0, False + add_24, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -842,6 +731,7 @@ def forward( # pd_op.add: (1x26x768xf32) <- (1x26x768xf32, 1x26x768xf32) add_25 = paddle._C_ops.add(layer_norm_12, dropout_16) + del dropout_16, layer_norm_12 # pd_op.layer_norm: (1x26x768xf32, 1x26xf32, 1x26xf32) <- (1x26x768xf32, 768xf32, 768xf32) layer_norm_15, layer_norm_16, layer_norm_17 = (lambda x, f: f(x))( @@ -850,7 +740,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_148, parameter_149 + del add_25, parameter_148, parameter_149 # pd_op.matmul: (1x26x3072xf32) <- (1x26x768xf32, 768x3072xf32) matmul_22 = paddle._C_ops.matmul(layer_norm_15, parameter_153, False, False) @@ -858,7 +748,7 @@ def forward( # pd_op.add: (1x26x3072xf32) <- (1x26x3072xf32, 3072xf32) add_26 = paddle._C_ops.add(matmul_22, parameter_152) - del parameter_152 + del matmul_22, parameter_152 # pd_op.relu: (1x26x3072xf32) <- (1x26x3072xf32) relu_2 = paddle._C_ops.relu(add_26) @@ -866,16 +756,16 @@ def forward( # pd_op.matmul: (1x26x768xf32) <- (1x26x3072xf32, 3072x768xf32) matmul_23 = paddle._C_ops.matmul(relu_2, parameter_151, False, False) - del parameter_151 + del parameter_151, relu_2 # pd_op.add: (1x26x768xf32) <- (1x26x768xf32, 768xf32) add_27 = paddle._C_ops.add(matmul_23, parameter_150) - del parameter_150 + del matmul_23, parameter_150 # pd_op.dropout: (1x26x768xf32, 1x26x768xui8) <- (1x26x768xf32, None, 1xf32) dropout_18, dropout_19 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_27, None, full_4, False, "upscale_in_train", 0, False + add_27, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -883,6 +773,7 @@ def forward( # pd_op.add: (1x26x768xf32) <- (1x26x768xf32, 1x26x768xf32) add_28 = paddle._C_ops.add(layer_norm_15, dropout_18) + del dropout_18, layer_norm_15 # pd_op.layer_norm: (1x26x768xf32, 1x26xf32, 1x26xf32) <- (1x26x768xf32, 768xf32, 768xf32) layer_norm_18, layer_norm_19, layer_norm_20 = (lambda x, f: f(x))( @@ -891,7 +782,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_146, parameter_147 + del add_28, parameter_146, parameter_147 # pd_op.matmul: (1x26x768xf32) <- (1x26x768xf32, 768x768xf32) matmul_24 = paddle._C_ops.matmul(layer_norm_18, parameter_145, False, False) @@ -899,10 +790,11 @@ def forward( # pd_op.add: (1x26x768xf32) <- (1x26x768xf32, 768xf32) add_29 = paddle._C_ops.add(matmul_24, parameter_144) - del parameter_144 + del matmul_24, parameter_144 # pd_op.reshape: (1x26x12x64xf32) <- (1x26x768xf32, 4xi64) reshape_12 = paddle._C_ops.reshape(add_29, full_int_array_1) + del add_29 # pd_op.transpose: (1x12x26x64xf32) <- (1x26x12x64xf32) transpose_12 = paddle._C_ops.transpose(reshape_12, [0, 2, 1, 3]) @@ -914,7 +806,7 @@ def forward( # pd_op.add: (1x26x768xf32) <- (1x26x768xf32, 768xf32) add_30 = paddle._C_ops.add(matmul_25, parameter_142) - del parameter_142 + del matmul_25, parameter_142 # pd_op.matmul: (1x26x768xf32) <- (1x26x768xf32, 768x768xf32) matmul_26 = paddle._C_ops.matmul(layer_norm_18, parameter_141, False, False) @@ -922,10 +814,11 @@ def forward( # pd_op.add: (1x26x768xf32) <- (1x26x768xf32, 768xf32) add_31 = paddle._C_ops.add(matmul_26, parameter_140) - del parameter_140 + del matmul_26, parameter_140 # pd_op.reshape: (1x26x12x64xf32) <- (1x26x768xf32, 4xi64) reshape_13 = paddle._C_ops.reshape(add_30, full_int_array_1) + del add_30 # pd_op.transpose: (1x12x26x64xf32) <- (1x26x12x64xf32) transpose_13 = paddle._C_ops.transpose(reshape_13, [0, 2, 1, 3]) @@ -933,6 +826,7 @@ def forward( # pd_op.reshape: (1x26x12x64xf32) <- (1x26x768xf32, 4xi64) reshape_14 = paddle._C_ops.reshape(add_31, full_int_array_1) + del add_31 # pd_op.transpose: (1x12x26x64xf32) <- (1x26x12x64xf32) transpose_14 = paddle._C_ops.transpose(reshape_14, [0, 2, 1, 3]) @@ -944,9 +838,11 @@ def forward( # pd_op.matmul: (1x12x26x26xf32) <- (1x12x26x64xf32, 1x12x26x64xf32) matmul_27 = paddle._C_ops.matmul(scale_4, transpose_13, False, True) + del scale_4, transpose_13 # pd_op.add: (1x12x26x26xf32) <- (1x12x26x26xf32, 1x1x1x26xf32) add_32 = paddle._C_ops.add(matmul_27, unsqueeze_0) + del matmul_27 # pd_op.softmax: (1x12x26x26xf32) <- (1x12x26x26xf32) softmax_3 = paddle._C_ops.softmax(add_32, -1) @@ -955,13 +851,15 @@ def forward( # pd_op.dropout: (1x12x26x26xf32, 1x12x26x26xui8) <- (1x12x26x26xf32, None, 1xf32) dropout_20, dropout_21 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_3, None, full_4, False, "upscale_in_train", 0, False + softmax_3, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_3 # pd_op.matmul: (1x12x26x64xf32) <- (1x12x26x26xf32, 1x12x26x64xf32) matmul_28 = paddle._C_ops.matmul(dropout_20, transpose_14, False, False) + del dropout_20, transpose_14 # pd_op.transpose: (1x26x12x64xf32) <- (1x12x26x64xf32) transpose_15 = paddle._C_ops.transpose(matmul_28, [0, 2, 1, 3]) @@ -969,19 +867,20 @@ def forward( # pd_op.reshape: (1x26x768xf32) <- (1x26x12x64xf32, 3xi64) reshape_15 = paddle._C_ops.reshape(transpose_15, full_int_array_2) + del transpose_15 # pd_op.matmul: (1x26x768xf32) <- (1x26x768xf32, 768x768xf32) matmul_29 = paddle._C_ops.matmul(reshape_15, parameter_139, False, False) - del parameter_139 + del parameter_139, reshape_15 # pd_op.add: (1x26x768xf32) <- (1x26x768xf32, 768xf32) add_33 = paddle._C_ops.add(matmul_29, parameter_138) - del parameter_138 + del matmul_29, parameter_138 # pd_op.dropout: (1x26x768xf32, 1x26x768xui8) <- (1x26x768xf32, None, 1xf32) dropout_22, dropout_23 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_33, None, full_4, False, "upscale_in_train", 0, False + add_33, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -989,6 +888,7 @@ def forward( # pd_op.add: (1x26x768xf32) <- (1x26x768xf32, 1x26x768xf32) add_34 = paddle._C_ops.add(layer_norm_18, dropout_22) + del dropout_22, layer_norm_18 # pd_op.layer_norm: (1x26x768xf32, 1x26xf32, 1x26xf32) <- (1x26x768xf32, 768xf32, 768xf32) layer_norm_21, layer_norm_22, layer_norm_23 = (lambda x, f: f(x))( @@ -997,7 +897,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_132, parameter_133 + del add_34, parameter_132, parameter_133 # pd_op.matmul: (1x26x3072xf32) <- (1x26x768xf32, 768x3072xf32) matmul_30 = paddle._C_ops.matmul(layer_norm_21, parameter_137, False, False) @@ -1005,7 +905,7 @@ def forward( # pd_op.add: (1x26x3072xf32) <- (1x26x3072xf32, 3072xf32) add_35 = paddle._C_ops.add(matmul_30, parameter_136) - del parameter_136 + del matmul_30, parameter_136 # pd_op.relu: (1x26x3072xf32) <- (1x26x3072xf32) relu_3 = paddle._C_ops.relu(add_35) @@ -1013,16 +913,16 @@ def forward( # pd_op.matmul: (1x26x768xf32) <- (1x26x3072xf32, 3072x768xf32) matmul_31 = paddle._C_ops.matmul(relu_3, parameter_135, False, False) - del parameter_135 + del parameter_135, relu_3 # pd_op.add: (1x26x768xf32) <- (1x26x768xf32, 768xf32) add_36 = paddle._C_ops.add(matmul_31, parameter_134) - del parameter_134 + del matmul_31, parameter_134 # pd_op.dropout: (1x26x768xf32, 1x26x768xui8) <- (1x26x768xf32, None, 1xf32) dropout_24, dropout_25 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_36, None, full_4, False, "upscale_in_train", 0, False + add_36, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1030,6 +930,7 @@ def forward( # pd_op.add: (1x26x768xf32) <- (1x26x768xf32, 1x26x768xf32) add_37 = paddle._C_ops.add(layer_norm_21, dropout_24) + del dropout_24, layer_norm_21 # pd_op.layer_norm: (1x26x768xf32, 1x26xf32, 1x26xf32) <- (1x26x768xf32, 768xf32, 768xf32) layer_norm_24, layer_norm_25, layer_norm_26 = (lambda x, f: f(x))( @@ -1038,7 +939,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_130, parameter_131 + del add_37, parameter_130, parameter_131 # pd_op.matmul: (1x26x768xf32) <- (1x26x768xf32, 768x768xf32) matmul_32 = paddle._C_ops.matmul(layer_norm_24, parameter_129, False, False) @@ -1046,10 +947,11 @@ def forward( # pd_op.add: (1x26x768xf32) <- (1x26x768xf32, 768xf32) add_38 = paddle._C_ops.add(matmul_32, parameter_128) - del parameter_128 + del matmul_32, parameter_128 # pd_op.reshape: (1x26x12x64xf32) <- (1x26x768xf32, 4xi64) reshape_16 = paddle._C_ops.reshape(add_38, full_int_array_1) + del add_38 # pd_op.transpose: (1x12x26x64xf32) <- (1x26x12x64xf32) transpose_16 = paddle._C_ops.transpose(reshape_16, [0, 2, 1, 3]) @@ -1061,7 +963,7 @@ def forward( # pd_op.add: (1x26x768xf32) <- (1x26x768xf32, 768xf32) add_39 = paddle._C_ops.add(matmul_33, parameter_126) - del parameter_126 + del matmul_33, parameter_126 # pd_op.matmul: (1x26x768xf32) <- (1x26x768xf32, 768x768xf32) matmul_34 = paddle._C_ops.matmul(layer_norm_24, parameter_125, False, False) @@ -1069,10 +971,11 @@ def forward( # pd_op.add: (1x26x768xf32) <- (1x26x768xf32, 768xf32) add_40 = paddle._C_ops.add(matmul_34, parameter_124) - del parameter_124 + del matmul_34, parameter_124 # pd_op.reshape: (1x26x12x64xf32) <- (1x26x768xf32, 4xi64) reshape_17 = paddle._C_ops.reshape(add_39, full_int_array_1) + del add_39 # pd_op.transpose: (1x12x26x64xf32) <- (1x26x12x64xf32) transpose_17 = paddle._C_ops.transpose(reshape_17, [0, 2, 1, 3]) @@ -1080,6 +983,7 @@ def forward( # pd_op.reshape: (1x26x12x64xf32) <- (1x26x768xf32, 4xi64) reshape_18 = paddle._C_ops.reshape(add_40, full_int_array_1) + del add_40 # pd_op.transpose: (1x12x26x64xf32) <- (1x26x12x64xf32) transpose_18 = paddle._C_ops.transpose(reshape_18, [0, 2, 1, 3]) @@ -1091,9 +995,11 @@ def forward( # pd_op.matmul: (1x12x26x26xf32) <- (1x12x26x64xf32, 1x12x26x64xf32) matmul_35 = paddle._C_ops.matmul(scale_5, transpose_17, False, True) + del scale_5, transpose_17 # pd_op.add: (1x12x26x26xf32) <- (1x12x26x26xf32, 1x1x1x26xf32) add_41 = paddle._C_ops.add(matmul_35, unsqueeze_0) + del matmul_35 # pd_op.softmax: (1x12x26x26xf32) <- (1x12x26x26xf32) softmax_4 = paddle._C_ops.softmax(add_41, -1) @@ -1102,13 +1008,15 @@ def forward( # pd_op.dropout: (1x12x26x26xf32, 1x12x26x26xui8) <- (1x12x26x26xf32, None, 1xf32) dropout_26, dropout_27 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_4, None, full_4, False, "upscale_in_train", 0, False + softmax_4, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_4 # pd_op.matmul: (1x12x26x64xf32) <- (1x12x26x26xf32, 1x12x26x64xf32) matmul_36 = paddle._C_ops.matmul(dropout_26, transpose_18, False, False) + del dropout_26, transpose_18 # pd_op.transpose: (1x26x12x64xf32) <- (1x12x26x64xf32) transpose_19 = paddle._C_ops.transpose(matmul_36, [0, 2, 1, 3]) @@ -1116,19 +1024,20 @@ def forward( # pd_op.reshape: (1x26x768xf32) <- (1x26x12x64xf32, 3xi64) reshape_19 = paddle._C_ops.reshape(transpose_19, full_int_array_2) + del transpose_19 # pd_op.matmul: (1x26x768xf32) <- (1x26x768xf32, 768x768xf32) matmul_37 = paddle._C_ops.matmul(reshape_19, parameter_123, False, False) - del parameter_123 + del parameter_123, reshape_19 # pd_op.add: (1x26x768xf32) <- (1x26x768xf32, 768xf32) add_42 = paddle._C_ops.add(matmul_37, parameter_122) - del parameter_122 + del matmul_37, parameter_122 # pd_op.dropout: (1x26x768xf32, 1x26x768xui8) <- (1x26x768xf32, None, 1xf32) dropout_28, dropout_29 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_42, None, full_4, False, "upscale_in_train", 0, False + add_42, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1136,6 +1045,7 @@ def forward( # pd_op.add: (1x26x768xf32) <- (1x26x768xf32, 1x26x768xf32) add_43 = paddle._C_ops.add(layer_norm_24, dropout_28) + del dropout_28, layer_norm_24 # pd_op.layer_norm: (1x26x768xf32, 1x26xf32, 1x26xf32) <- (1x26x768xf32, 768xf32, 768xf32) layer_norm_27, layer_norm_28, layer_norm_29 = (lambda x, f: f(x))( @@ -1144,7 +1054,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_116, parameter_117 + del add_43, parameter_116, parameter_117 # pd_op.matmul: (1x26x3072xf32) <- (1x26x768xf32, 768x3072xf32) matmul_38 = paddle._C_ops.matmul(layer_norm_27, parameter_121, False, False) @@ -1152,7 +1062,7 @@ def forward( # pd_op.add: (1x26x3072xf32) <- (1x26x3072xf32, 3072xf32) add_44 = paddle._C_ops.add(matmul_38, parameter_120) - del parameter_120 + del matmul_38, parameter_120 # pd_op.relu: (1x26x3072xf32) <- (1x26x3072xf32) relu_4 = paddle._C_ops.relu(add_44) @@ -1160,16 +1070,16 @@ def forward( # pd_op.matmul: (1x26x768xf32) <- (1x26x3072xf32, 3072x768xf32) matmul_39 = paddle._C_ops.matmul(relu_4, parameter_119, False, False) - del parameter_119 + del parameter_119, relu_4 # pd_op.add: (1x26x768xf32) <- (1x26x768xf32, 768xf32) add_45 = paddle._C_ops.add(matmul_39, parameter_118) - del parameter_118 + del matmul_39, parameter_118 # pd_op.dropout: (1x26x768xf32, 1x26x768xui8) <- (1x26x768xf32, None, 1xf32) dropout_30, dropout_31 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_45, None, full_4, False, "upscale_in_train", 0, False + add_45, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1177,6 +1087,7 @@ def forward( # pd_op.add: (1x26x768xf32) <- (1x26x768xf32, 1x26x768xf32) add_46 = paddle._C_ops.add(layer_norm_27, dropout_30) + del dropout_30, layer_norm_27 # pd_op.layer_norm: (1x26x768xf32, 1x26xf32, 1x26xf32) <- (1x26x768xf32, 768xf32, 768xf32) layer_norm_30, layer_norm_31, layer_norm_32 = (lambda x, f: f(x))( @@ -1185,7 +1096,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_114, parameter_115 + del add_46, parameter_114, parameter_115 # pd_op.matmul: (1x26x768xf32) <- (1x26x768xf32, 768x768xf32) matmul_40 = paddle._C_ops.matmul(layer_norm_30, parameter_113, False, False) @@ -1193,10 +1104,11 @@ def forward( # pd_op.add: (1x26x768xf32) <- (1x26x768xf32, 768xf32) add_47 = paddle._C_ops.add(matmul_40, parameter_112) - del parameter_112 + del matmul_40, parameter_112 # pd_op.reshape: (1x26x12x64xf32) <- (1x26x768xf32, 4xi64) reshape_20 = paddle._C_ops.reshape(add_47, full_int_array_1) + del add_47 # pd_op.transpose: (1x12x26x64xf32) <- (1x26x12x64xf32) transpose_20 = paddle._C_ops.transpose(reshape_20, [0, 2, 1, 3]) @@ -1208,7 +1120,7 @@ def forward( # pd_op.add: (1x26x768xf32) <- (1x26x768xf32, 768xf32) add_48 = paddle._C_ops.add(matmul_41, parameter_110) - del parameter_110 + del matmul_41, parameter_110 # pd_op.matmul: (1x26x768xf32) <- (1x26x768xf32, 768x768xf32) matmul_42 = paddle._C_ops.matmul(layer_norm_30, parameter_109, False, False) @@ -1216,10 +1128,11 @@ def forward( # pd_op.add: (1x26x768xf32) <- (1x26x768xf32, 768xf32) add_49 = paddle._C_ops.add(matmul_42, parameter_108) - del parameter_108 + del matmul_42, parameter_108 # pd_op.reshape: (1x26x12x64xf32) <- (1x26x768xf32, 4xi64) reshape_21 = paddle._C_ops.reshape(add_48, full_int_array_1) + del add_48 # pd_op.transpose: (1x12x26x64xf32) <- (1x26x12x64xf32) transpose_21 = paddle._C_ops.transpose(reshape_21, [0, 2, 1, 3]) @@ -1227,6 +1140,7 @@ def forward( # pd_op.reshape: (1x26x12x64xf32) <- (1x26x768xf32, 4xi64) reshape_22 = paddle._C_ops.reshape(add_49, full_int_array_1) + del add_49 # pd_op.transpose: (1x12x26x64xf32) <- (1x26x12x64xf32) transpose_22 = paddle._C_ops.transpose(reshape_22, [0, 2, 1, 3]) @@ -1238,9 +1152,11 @@ def forward( # pd_op.matmul: (1x12x26x26xf32) <- (1x12x26x64xf32, 1x12x26x64xf32) matmul_43 = paddle._C_ops.matmul(scale_6, transpose_21, False, True) + del scale_6, transpose_21 # pd_op.add: (1x12x26x26xf32) <- (1x12x26x26xf32, 1x1x1x26xf32) add_50 = paddle._C_ops.add(matmul_43, unsqueeze_0) + del matmul_43 # pd_op.softmax: (1x12x26x26xf32) <- (1x12x26x26xf32) softmax_5 = paddle._C_ops.softmax(add_50, -1) @@ -1249,13 +1165,15 @@ def forward( # pd_op.dropout: (1x12x26x26xf32, 1x12x26x26xui8) <- (1x12x26x26xf32, None, 1xf32) dropout_32, dropout_33 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_5, None, full_4, False, "upscale_in_train", 0, False + softmax_5, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_5 # pd_op.matmul: (1x12x26x64xf32) <- (1x12x26x26xf32, 1x12x26x64xf32) matmul_44 = paddle._C_ops.matmul(dropout_32, transpose_22, False, False) + del dropout_32, transpose_22 # pd_op.transpose: (1x26x12x64xf32) <- (1x12x26x64xf32) transpose_23 = paddle._C_ops.transpose(matmul_44, [0, 2, 1, 3]) @@ -1263,19 +1181,20 @@ def forward( # pd_op.reshape: (1x26x768xf32) <- (1x26x12x64xf32, 3xi64) reshape_23 = paddle._C_ops.reshape(transpose_23, full_int_array_2) + del transpose_23 # pd_op.matmul: (1x26x768xf32) <- (1x26x768xf32, 768x768xf32) matmul_45 = paddle._C_ops.matmul(reshape_23, parameter_107, False, False) - del parameter_107 + del parameter_107, reshape_23 # pd_op.add: (1x26x768xf32) <- (1x26x768xf32, 768xf32) add_51 = paddle._C_ops.add(matmul_45, parameter_106) - del parameter_106 + del matmul_45, parameter_106 # pd_op.dropout: (1x26x768xf32, 1x26x768xui8) <- (1x26x768xf32, None, 1xf32) dropout_34, dropout_35 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_51, None, full_4, False, "upscale_in_train", 0, False + add_51, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1283,6 +1202,7 @@ def forward( # pd_op.add: (1x26x768xf32) <- (1x26x768xf32, 1x26x768xf32) add_52 = paddle._C_ops.add(layer_norm_30, dropout_34) + del dropout_34, layer_norm_30 # pd_op.layer_norm: (1x26x768xf32, 1x26xf32, 1x26xf32) <- (1x26x768xf32, 768xf32, 768xf32) layer_norm_33, layer_norm_34, layer_norm_35 = (lambda x, f: f(x))( @@ -1291,7 +1211,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_100, parameter_101 + del add_52, parameter_100, parameter_101 # pd_op.matmul: (1x26x3072xf32) <- (1x26x768xf32, 768x3072xf32) matmul_46 = paddle._C_ops.matmul(layer_norm_33, parameter_105, False, False) @@ -1299,7 +1219,7 @@ def forward( # pd_op.add: (1x26x3072xf32) <- (1x26x3072xf32, 3072xf32) add_53 = paddle._C_ops.add(matmul_46, parameter_104) - del parameter_104 + del matmul_46, parameter_104 # pd_op.relu: (1x26x3072xf32) <- (1x26x3072xf32) relu_5 = paddle._C_ops.relu(add_53) @@ -1307,16 +1227,16 @@ def forward( # pd_op.matmul: (1x26x768xf32) <- (1x26x3072xf32, 3072x768xf32) matmul_47 = paddle._C_ops.matmul(relu_5, parameter_103, False, False) - del parameter_103 + del parameter_103, relu_5 # pd_op.add: (1x26x768xf32) <- (1x26x768xf32, 768xf32) add_54 = paddle._C_ops.add(matmul_47, parameter_102) - del parameter_102 + del matmul_47, parameter_102 # pd_op.dropout: (1x26x768xf32, 1x26x768xui8) <- (1x26x768xf32, None, 1xf32) dropout_36, dropout_37 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_54, None, full_4, False, "upscale_in_train", 0, False + add_54, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1324,6 +1244,7 @@ def forward( # pd_op.add: (1x26x768xf32) <- (1x26x768xf32, 1x26x768xf32) add_55 = paddle._C_ops.add(layer_norm_33, dropout_36) + del dropout_36, layer_norm_33 # pd_op.layer_norm: (1x26x768xf32, 1x26xf32, 1x26xf32) <- (1x26x768xf32, 768xf32, 768xf32) layer_norm_36, layer_norm_37, layer_norm_38 = (lambda x, f: f(x))( @@ -1332,7 +1253,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_98, parameter_99 + del add_55, parameter_98, parameter_99 # pd_op.matmul: (1x26x768xf32) <- (1x26x768xf32, 768x768xf32) matmul_48 = paddle._C_ops.matmul(layer_norm_36, parameter_97, False, False) @@ -1340,10 +1261,11 @@ def forward( # pd_op.add: (1x26x768xf32) <- (1x26x768xf32, 768xf32) add_56 = paddle._C_ops.add(matmul_48, parameter_96) - del parameter_96 + del matmul_48, parameter_96 # pd_op.reshape: (1x26x12x64xf32) <- (1x26x768xf32, 4xi64) reshape_24 = paddle._C_ops.reshape(add_56, full_int_array_1) + del add_56 # pd_op.transpose: (1x12x26x64xf32) <- (1x26x12x64xf32) transpose_24 = paddle._C_ops.transpose(reshape_24, [0, 2, 1, 3]) @@ -1355,7 +1277,7 @@ def forward( # pd_op.add: (1x26x768xf32) <- (1x26x768xf32, 768xf32) add_57 = paddle._C_ops.add(matmul_49, parameter_94) - del parameter_94 + del matmul_49, parameter_94 # pd_op.matmul: (1x26x768xf32) <- (1x26x768xf32, 768x768xf32) matmul_50 = paddle._C_ops.matmul(layer_norm_36, parameter_93, False, False) @@ -1363,10 +1285,11 @@ def forward( # pd_op.add: (1x26x768xf32) <- (1x26x768xf32, 768xf32) add_58 = paddle._C_ops.add(matmul_50, parameter_92) - del parameter_92 + del matmul_50, parameter_92 # pd_op.reshape: (1x26x12x64xf32) <- (1x26x768xf32, 4xi64) reshape_25 = paddle._C_ops.reshape(add_57, full_int_array_1) + del add_57 # pd_op.transpose: (1x12x26x64xf32) <- (1x26x12x64xf32) transpose_25 = paddle._C_ops.transpose(reshape_25, [0, 2, 1, 3]) @@ -1374,6 +1297,7 @@ def forward( # pd_op.reshape: (1x26x12x64xf32) <- (1x26x768xf32, 4xi64) reshape_26 = paddle._C_ops.reshape(add_58, full_int_array_1) + del add_58 # pd_op.transpose: (1x12x26x64xf32) <- (1x26x12x64xf32) transpose_26 = paddle._C_ops.transpose(reshape_26, [0, 2, 1, 3]) @@ -1385,9 +1309,11 @@ def forward( # pd_op.matmul: (1x12x26x26xf32) <- (1x12x26x64xf32, 1x12x26x64xf32) matmul_51 = paddle._C_ops.matmul(scale_7, transpose_25, False, True) + del scale_7, transpose_25 # pd_op.add: (1x12x26x26xf32) <- (1x12x26x26xf32, 1x1x1x26xf32) add_59 = paddle._C_ops.add(matmul_51, unsqueeze_0) + del matmul_51 # pd_op.softmax: (1x12x26x26xf32) <- (1x12x26x26xf32) softmax_6 = paddle._C_ops.softmax(add_59, -1) @@ -1396,13 +1322,15 @@ def forward( # pd_op.dropout: (1x12x26x26xf32, 1x12x26x26xui8) <- (1x12x26x26xf32, None, 1xf32) dropout_38, dropout_39 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_6, None, full_4, False, "upscale_in_train", 0, False + softmax_6, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_6 # pd_op.matmul: (1x12x26x64xf32) <- (1x12x26x26xf32, 1x12x26x64xf32) matmul_52 = paddle._C_ops.matmul(dropout_38, transpose_26, False, False) + del dropout_38, transpose_26 # pd_op.transpose: (1x26x12x64xf32) <- (1x12x26x64xf32) transpose_27 = paddle._C_ops.transpose(matmul_52, [0, 2, 1, 3]) @@ -1410,19 +1338,20 @@ def forward( # pd_op.reshape: (1x26x768xf32) <- (1x26x12x64xf32, 3xi64) reshape_27 = paddle._C_ops.reshape(transpose_27, full_int_array_2) + del transpose_27 # pd_op.matmul: (1x26x768xf32) <- (1x26x768xf32, 768x768xf32) matmul_53 = paddle._C_ops.matmul(reshape_27, parameter_91, False, False) - del parameter_91 + del parameter_91, reshape_27 # pd_op.add: (1x26x768xf32) <- (1x26x768xf32, 768xf32) add_60 = paddle._C_ops.add(matmul_53, parameter_90) - del parameter_90 + del matmul_53, parameter_90 # pd_op.dropout: (1x26x768xf32, 1x26x768xui8) <- (1x26x768xf32, None, 1xf32) dropout_40, dropout_41 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_60, None, full_4, False, "upscale_in_train", 0, False + add_60, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1430,6 +1359,7 @@ def forward( # pd_op.add: (1x26x768xf32) <- (1x26x768xf32, 1x26x768xf32) add_61 = paddle._C_ops.add(layer_norm_36, dropout_40) + del dropout_40, layer_norm_36 # pd_op.layer_norm: (1x26x768xf32, 1x26xf32, 1x26xf32) <- (1x26x768xf32, 768xf32, 768xf32) layer_norm_39, layer_norm_40, layer_norm_41 = (lambda x, f: f(x))( @@ -1438,7 +1368,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_84, parameter_85 + del add_61, parameter_84, parameter_85 # pd_op.matmul: (1x26x3072xf32) <- (1x26x768xf32, 768x3072xf32) matmul_54 = paddle._C_ops.matmul(layer_norm_39, parameter_89, False, False) @@ -1446,7 +1376,7 @@ def forward( # pd_op.add: (1x26x3072xf32) <- (1x26x3072xf32, 3072xf32) add_62 = paddle._C_ops.add(matmul_54, parameter_88) - del parameter_88 + del matmul_54, parameter_88 # pd_op.relu: (1x26x3072xf32) <- (1x26x3072xf32) relu_6 = paddle._C_ops.relu(add_62) @@ -1454,16 +1384,16 @@ def forward( # pd_op.matmul: (1x26x768xf32) <- (1x26x3072xf32, 3072x768xf32) matmul_55 = paddle._C_ops.matmul(relu_6, parameter_87, False, False) - del parameter_87 + del parameter_87, relu_6 # pd_op.add: (1x26x768xf32) <- (1x26x768xf32, 768xf32) add_63 = paddle._C_ops.add(matmul_55, parameter_86) - del parameter_86 + del matmul_55, parameter_86 # pd_op.dropout: (1x26x768xf32, 1x26x768xui8) <- (1x26x768xf32, None, 1xf32) dropout_42, dropout_43 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_63, None, full_4, False, "upscale_in_train", 0, False + add_63, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1471,6 +1401,7 @@ def forward( # pd_op.add: (1x26x768xf32) <- (1x26x768xf32, 1x26x768xf32) add_64 = paddle._C_ops.add(layer_norm_39, dropout_42) + del dropout_42, layer_norm_39 # pd_op.layer_norm: (1x26x768xf32, 1x26xf32, 1x26xf32) <- (1x26x768xf32, 768xf32, 768xf32) layer_norm_42, layer_norm_43, layer_norm_44 = (lambda x, f: f(x))( @@ -1479,7 +1410,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_82, parameter_83 + del add_64, parameter_82, parameter_83 # pd_op.matmul: (1x26x768xf32) <- (1x26x768xf32, 768x768xf32) matmul_56 = paddle._C_ops.matmul(layer_norm_42, parameter_81, False, False) @@ -1487,10 +1418,11 @@ def forward( # pd_op.add: (1x26x768xf32) <- (1x26x768xf32, 768xf32) add_65 = paddle._C_ops.add(matmul_56, parameter_80) - del parameter_80 + del matmul_56, parameter_80 # pd_op.reshape: (1x26x12x64xf32) <- (1x26x768xf32, 4xi64) reshape_28 = paddle._C_ops.reshape(add_65, full_int_array_1) + del add_65 # pd_op.transpose: (1x12x26x64xf32) <- (1x26x12x64xf32) transpose_28 = paddle._C_ops.transpose(reshape_28, [0, 2, 1, 3]) @@ -1502,7 +1434,7 @@ def forward( # pd_op.add: (1x26x768xf32) <- (1x26x768xf32, 768xf32) add_66 = paddle._C_ops.add(matmul_57, parameter_78) - del parameter_78 + del matmul_57, parameter_78 # pd_op.matmul: (1x26x768xf32) <- (1x26x768xf32, 768x768xf32) matmul_58 = paddle._C_ops.matmul(layer_norm_42, parameter_77, False, False) @@ -1510,10 +1442,11 @@ def forward( # pd_op.add: (1x26x768xf32) <- (1x26x768xf32, 768xf32) add_67 = paddle._C_ops.add(matmul_58, parameter_76) - del parameter_76 + del matmul_58, parameter_76 # pd_op.reshape: (1x26x12x64xf32) <- (1x26x768xf32, 4xi64) reshape_29 = paddle._C_ops.reshape(add_66, full_int_array_1) + del add_66 # pd_op.transpose: (1x12x26x64xf32) <- (1x26x12x64xf32) transpose_29 = paddle._C_ops.transpose(reshape_29, [0, 2, 1, 3]) @@ -1521,6 +1454,7 @@ def forward( # pd_op.reshape: (1x26x12x64xf32) <- (1x26x768xf32, 4xi64) reshape_30 = paddle._C_ops.reshape(add_67, full_int_array_1) + del add_67 # pd_op.transpose: (1x12x26x64xf32) <- (1x26x12x64xf32) transpose_30 = paddle._C_ops.transpose(reshape_30, [0, 2, 1, 3]) @@ -1532,9 +1466,11 @@ def forward( # pd_op.matmul: (1x12x26x26xf32) <- (1x12x26x64xf32, 1x12x26x64xf32) matmul_59 = paddle._C_ops.matmul(scale_8, transpose_29, False, True) + del scale_8, transpose_29 # pd_op.add: (1x12x26x26xf32) <- (1x12x26x26xf32, 1x1x1x26xf32) add_68 = paddle._C_ops.add(matmul_59, unsqueeze_0) + del matmul_59 # pd_op.softmax: (1x12x26x26xf32) <- (1x12x26x26xf32) softmax_7 = paddle._C_ops.softmax(add_68, -1) @@ -1543,13 +1479,15 @@ def forward( # pd_op.dropout: (1x12x26x26xf32, 1x12x26x26xui8) <- (1x12x26x26xf32, None, 1xf32) dropout_44, dropout_45 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_7, None, full_4, False, "upscale_in_train", 0, False + softmax_7, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_7 # pd_op.matmul: (1x12x26x64xf32) <- (1x12x26x26xf32, 1x12x26x64xf32) matmul_60 = paddle._C_ops.matmul(dropout_44, transpose_30, False, False) + del dropout_44, transpose_30 # pd_op.transpose: (1x26x12x64xf32) <- (1x12x26x64xf32) transpose_31 = paddle._C_ops.transpose(matmul_60, [0, 2, 1, 3]) @@ -1557,19 +1495,20 @@ def forward( # pd_op.reshape: (1x26x768xf32) <- (1x26x12x64xf32, 3xi64) reshape_31 = paddle._C_ops.reshape(transpose_31, full_int_array_2) + del transpose_31 # pd_op.matmul: (1x26x768xf32) <- (1x26x768xf32, 768x768xf32) matmul_61 = paddle._C_ops.matmul(reshape_31, parameter_75, False, False) - del parameter_75 + del parameter_75, reshape_31 # pd_op.add: (1x26x768xf32) <- (1x26x768xf32, 768xf32) add_69 = paddle._C_ops.add(matmul_61, parameter_74) - del parameter_74 + del matmul_61, parameter_74 # pd_op.dropout: (1x26x768xf32, 1x26x768xui8) <- (1x26x768xf32, None, 1xf32) dropout_46, dropout_47 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_69, None, full_4, False, "upscale_in_train", 0, False + add_69, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1577,6 +1516,7 @@ def forward( # pd_op.add: (1x26x768xf32) <- (1x26x768xf32, 1x26x768xf32) add_70 = paddle._C_ops.add(layer_norm_42, dropout_46) + del dropout_46, layer_norm_42 # pd_op.layer_norm: (1x26x768xf32, 1x26xf32, 1x26xf32) <- (1x26x768xf32, 768xf32, 768xf32) layer_norm_45, layer_norm_46, layer_norm_47 = (lambda x, f: f(x))( @@ -1585,7 +1525,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_68, parameter_69 + del add_70, parameter_68, parameter_69 # pd_op.matmul: (1x26x3072xf32) <- (1x26x768xf32, 768x3072xf32) matmul_62 = paddle._C_ops.matmul(layer_norm_45, parameter_73, False, False) @@ -1593,7 +1533,7 @@ def forward( # pd_op.add: (1x26x3072xf32) <- (1x26x3072xf32, 3072xf32) add_71 = paddle._C_ops.add(matmul_62, parameter_72) - del parameter_72 + del matmul_62, parameter_72 # pd_op.relu: (1x26x3072xf32) <- (1x26x3072xf32) relu_7 = paddle._C_ops.relu(add_71) @@ -1601,16 +1541,16 @@ def forward( # pd_op.matmul: (1x26x768xf32) <- (1x26x3072xf32, 3072x768xf32) matmul_63 = paddle._C_ops.matmul(relu_7, parameter_71, False, False) - del parameter_71 + del parameter_71, relu_7 # pd_op.add: (1x26x768xf32) <- (1x26x768xf32, 768xf32) add_72 = paddle._C_ops.add(matmul_63, parameter_70) - del parameter_70 + del matmul_63, parameter_70 # pd_op.dropout: (1x26x768xf32, 1x26x768xui8) <- (1x26x768xf32, None, 1xf32) dropout_48, dropout_49 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_72, None, full_4, False, "upscale_in_train", 0, False + add_72, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1618,6 +1558,7 @@ def forward( # pd_op.add: (1x26x768xf32) <- (1x26x768xf32, 1x26x768xf32) add_73 = paddle._C_ops.add(layer_norm_45, dropout_48) + del dropout_48, layer_norm_45 # pd_op.layer_norm: (1x26x768xf32, 1x26xf32, 1x26xf32) <- (1x26x768xf32, 768xf32, 768xf32) layer_norm_48, layer_norm_49, layer_norm_50 = (lambda x, f: f(x))( @@ -1626,7 +1567,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_66, parameter_67 + del add_73, parameter_66, parameter_67 # pd_op.matmul: (1x26x768xf32) <- (1x26x768xf32, 768x768xf32) matmul_64 = paddle._C_ops.matmul(layer_norm_48, parameter_65, False, False) @@ -1634,10 +1575,11 @@ def forward( # pd_op.add: (1x26x768xf32) <- (1x26x768xf32, 768xf32) add_74 = paddle._C_ops.add(matmul_64, parameter_64) - del parameter_64 + del matmul_64, parameter_64 # pd_op.reshape: (1x26x12x64xf32) <- (1x26x768xf32, 4xi64) reshape_32 = paddle._C_ops.reshape(add_74, full_int_array_1) + del add_74 # pd_op.transpose: (1x12x26x64xf32) <- (1x26x12x64xf32) transpose_32 = paddle._C_ops.transpose(reshape_32, [0, 2, 1, 3]) @@ -1649,7 +1591,7 @@ def forward( # pd_op.add: (1x26x768xf32) <- (1x26x768xf32, 768xf32) add_75 = paddle._C_ops.add(matmul_65, parameter_62) - del parameter_62 + del matmul_65, parameter_62 # pd_op.matmul: (1x26x768xf32) <- (1x26x768xf32, 768x768xf32) matmul_66 = paddle._C_ops.matmul(layer_norm_48, parameter_61, False, False) @@ -1657,10 +1599,11 @@ def forward( # pd_op.add: (1x26x768xf32) <- (1x26x768xf32, 768xf32) add_76 = paddle._C_ops.add(matmul_66, parameter_60) - del parameter_60 + del matmul_66, parameter_60 # pd_op.reshape: (1x26x12x64xf32) <- (1x26x768xf32, 4xi64) reshape_33 = paddle._C_ops.reshape(add_75, full_int_array_1) + del add_75 # pd_op.transpose: (1x12x26x64xf32) <- (1x26x12x64xf32) transpose_33 = paddle._C_ops.transpose(reshape_33, [0, 2, 1, 3]) @@ -1668,6 +1611,7 @@ def forward( # pd_op.reshape: (1x26x12x64xf32) <- (1x26x768xf32, 4xi64) reshape_34 = paddle._C_ops.reshape(add_76, full_int_array_1) + del add_76 # pd_op.transpose: (1x12x26x64xf32) <- (1x26x12x64xf32) transpose_34 = paddle._C_ops.transpose(reshape_34, [0, 2, 1, 3]) @@ -1679,9 +1623,11 @@ def forward( # pd_op.matmul: (1x12x26x26xf32) <- (1x12x26x64xf32, 1x12x26x64xf32) matmul_67 = paddle._C_ops.matmul(scale_9, transpose_33, False, True) + del scale_9, transpose_33 # pd_op.add: (1x12x26x26xf32) <- (1x12x26x26xf32, 1x1x1x26xf32) add_77 = paddle._C_ops.add(matmul_67, unsqueeze_0) + del matmul_67 # pd_op.softmax: (1x12x26x26xf32) <- (1x12x26x26xf32) softmax_8 = paddle._C_ops.softmax(add_77, -1) @@ -1690,13 +1636,15 @@ def forward( # pd_op.dropout: (1x12x26x26xf32, 1x12x26x26xui8) <- (1x12x26x26xf32, None, 1xf32) dropout_50, dropout_51 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_8, None, full_4, False, "upscale_in_train", 0, False + softmax_8, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_8 # pd_op.matmul: (1x12x26x64xf32) <- (1x12x26x26xf32, 1x12x26x64xf32) matmul_68 = paddle._C_ops.matmul(dropout_50, transpose_34, False, False) + del dropout_50, transpose_34 # pd_op.transpose: (1x26x12x64xf32) <- (1x12x26x64xf32) transpose_35 = paddle._C_ops.transpose(matmul_68, [0, 2, 1, 3]) @@ -1704,19 +1652,20 @@ def forward( # pd_op.reshape: (1x26x768xf32) <- (1x26x12x64xf32, 3xi64) reshape_35 = paddle._C_ops.reshape(transpose_35, full_int_array_2) + del transpose_35 # pd_op.matmul: (1x26x768xf32) <- (1x26x768xf32, 768x768xf32) matmul_69 = paddle._C_ops.matmul(reshape_35, parameter_59, False, False) - del parameter_59 + del parameter_59, reshape_35 # pd_op.add: (1x26x768xf32) <- (1x26x768xf32, 768xf32) add_78 = paddle._C_ops.add(matmul_69, parameter_58) - del parameter_58 + del matmul_69, parameter_58 # pd_op.dropout: (1x26x768xf32, 1x26x768xui8) <- (1x26x768xf32, None, 1xf32) dropout_52, dropout_53 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_78, None, full_4, False, "upscale_in_train", 0, False + add_78, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1724,6 +1673,7 @@ def forward( # pd_op.add: (1x26x768xf32) <- (1x26x768xf32, 1x26x768xf32) add_79 = paddle._C_ops.add(layer_norm_48, dropout_52) + del dropout_52, layer_norm_48 # pd_op.layer_norm: (1x26x768xf32, 1x26xf32, 1x26xf32) <- (1x26x768xf32, 768xf32, 768xf32) layer_norm_51, layer_norm_52, layer_norm_53 = (lambda x, f: f(x))( @@ -1732,7 +1682,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_52, parameter_53 + del add_79, parameter_52, parameter_53 # pd_op.matmul: (1x26x3072xf32) <- (1x26x768xf32, 768x3072xf32) matmul_70 = paddle._C_ops.matmul(layer_norm_51, parameter_57, False, False) @@ -1740,7 +1690,7 @@ def forward( # pd_op.add: (1x26x3072xf32) <- (1x26x3072xf32, 3072xf32) add_80 = paddle._C_ops.add(matmul_70, parameter_56) - del parameter_56 + del matmul_70, parameter_56 # pd_op.relu: (1x26x3072xf32) <- (1x26x3072xf32) relu_8 = paddle._C_ops.relu(add_80) @@ -1748,16 +1698,16 @@ def forward( # pd_op.matmul: (1x26x768xf32) <- (1x26x3072xf32, 3072x768xf32) matmul_71 = paddle._C_ops.matmul(relu_8, parameter_55, False, False) - del parameter_55 + del parameter_55, relu_8 # pd_op.add: (1x26x768xf32) <- (1x26x768xf32, 768xf32) add_81 = paddle._C_ops.add(matmul_71, parameter_54) - del parameter_54 + del matmul_71, parameter_54 # pd_op.dropout: (1x26x768xf32, 1x26x768xui8) <- (1x26x768xf32, None, 1xf32) dropout_54, dropout_55 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_81, None, full_4, False, "upscale_in_train", 0, False + add_81, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1765,6 +1715,7 @@ def forward( # pd_op.add: (1x26x768xf32) <- (1x26x768xf32, 1x26x768xf32) add_82 = paddle._C_ops.add(layer_norm_51, dropout_54) + del dropout_54, layer_norm_51 # pd_op.layer_norm: (1x26x768xf32, 1x26xf32, 1x26xf32) <- (1x26x768xf32, 768xf32, 768xf32) layer_norm_54, layer_norm_55, layer_norm_56 = (lambda x, f: f(x))( @@ -1773,7 +1724,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_50, parameter_51 + del add_82, parameter_50, parameter_51 # pd_op.matmul: (1x26x768xf32) <- (1x26x768xf32, 768x768xf32) matmul_72 = paddle._C_ops.matmul(layer_norm_54, parameter_49, False, False) @@ -1781,10 +1732,11 @@ def forward( # pd_op.add: (1x26x768xf32) <- (1x26x768xf32, 768xf32) add_83 = paddle._C_ops.add(matmul_72, parameter_48) - del parameter_48 + del matmul_72, parameter_48 # pd_op.reshape: (1x26x12x64xf32) <- (1x26x768xf32, 4xi64) reshape_36 = paddle._C_ops.reshape(add_83, full_int_array_1) + del add_83 # pd_op.transpose: (1x12x26x64xf32) <- (1x26x12x64xf32) transpose_36 = paddle._C_ops.transpose(reshape_36, [0, 2, 1, 3]) @@ -1796,7 +1748,7 @@ def forward( # pd_op.add: (1x26x768xf32) <- (1x26x768xf32, 768xf32) add_84 = paddle._C_ops.add(matmul_73, parameter_46) - del parameter_46 + del matmul_73, parameter_46 # pd_op.matmul: (1x26x768xf32) <- (1x26x768xf32, 768x768xf32) matmul_74 = paddle._C_ops.matmul(layer_norm_54, parameter_45, False, False) @@ -1804,10 +1756,11 @@ def forward( # pd_op.add: (1x26x768xf32) <- (1x26x768xf32, 768xf32) add_85 = paddle._C_ops.add(matmul_74, parameter_44) - del parameter_44 + del matmul_74, parameter_44 # pd_op.reshape: (1x26x12x64xf32) <- (1x26x768xf32, 4xi64) reshape_37 = paddle._C_ops.reshape(add_84, full_int_array_1) + del add_84 # pd_op.transpose: (1x12x26x64xf32) <- (1x26x12x64xf32) transpose_37 = paddle._C_ops.transpose(reshape_37, [0, 2, 1, 3]) @@ -1815,6 +1768,7 @@ def forward( # pd_op.reshape: (1x26x12x64xf32) <- (1x26x768xf32, 4xi64) reshape_38 = paddle._C_ops.reshape(add_85, full_int_array_1) + del add_85 # pd_op.transpose: (1x12x26x64xf32) <- (1x26x12x64xf32) transpose_38 = paddle._C_ops.transpose(reshape_38, [0, 2, 1, 3]) @@ -1826,9 +1780,11 @@ def forward( # pd_op.matmul: (1x12x26x26xf32) <- (1x12x26x64xf32, 1x12x26x64xf32) matmul_75 = paddle._C_ops.matmul(scale_10, transpose_37, False, True) + del scale_10, transpose_37 # pd_op.add: (1x12x26x26xf32) <- (1x12x26x26xf32, 1x1x1x26xf32) add_86 = paddle._C_ops.add(matmul_75, unsqueeze_0) + del matmul_75 # pd_op.softmax: (1x12x26x26xf32) <- (1x12x26x26xf32) softmax_9 = paddle._C_ops.softmax(add_86, -1) @@ -1837,13 +1793,15 @@ def forward( # pd_op.dropout: (1x12x26x26xf32, 1x12x26x26xui8) <- (1x12x26x26xf32, None, 1xf32) dropout_56, dropout_57 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_9, None, full_4, False, "upscale_in_train", 0, False + softmax_9, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_9 # pd_op.matmul: (1x12x26x64xf32) <- (1x12x26x26xf32, 1x12x26x64xf32) matmul_76 = paddle._C_ops.matmul(dropout_56, transpose_38, False, False) + del dropout_56, transpose_38 # pd_op.transpose: (1x26x12x64xf32) <- (1x12x26x64xf32) transpose_39 = paddle._C_ops.transpose(matmul_76, [0, 2, 1, 3]) @@ -1851,19 +1809,20 @@ def forward( # pd_op.reshape: (1x26x768xf32) <- (1x26x12x64xf32, 3xi64) reshape_39 = paddle._C_ops.reshape(transpose_39, full_int_array_2) + del transpose_39 # pd_op.matmul: (1x26x768xf32) <- (1x26x768xf32, 768x768xf32) matmul_77 = paddle._C_ops.matmul(reshape_39, parameter_43, False, False) - del parameter_43 + del parameter_43, reshape_39 # pd_op.add: (1x26x768xf32) <- (1x26x768xf32, 768xf32) add_87 = paddle._C_ops.add(matmul_77, parameter_42) - del parameter_42 + del matmul_77, parameter_42 # pd_op.dropout: (1x26x768xf32, 1x26x768xui8) <- (1x26x768xf32, None, 1xf32) dropout_58, dropout_59 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_87, None, full_4, False, "upscale_in_train", 0, False + add_87, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1871,6 +1830,7 @@ def forward( # pd_op.add: (1x26x768xf32) <- (1x26x768xf32, 1x26x768xf32) add_88 = paddle._C_ops.add(layer_norm_54, dropout_58) + del dropout_58, layer_norm_54 # pd_op.layer_norm: (1x26x768xf32, 1x26xf32, 1x26xf32) <- (1x26x768xf32, 768xf32, 768xf32) layer_norm_57, layer_norm_58, layer_norm_59 = (lambda x, f: f(x))( @@ -1879,7 +1839,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_36, parameter_37 + del add_88, parameter_36, parameter_37 # pd_op.matmul: (1x26x3072xf32) <- (1x26x768xf32, 768x3072xf32) matmul_78 = paddle._C_ops.matmul(layer_norm_57, parameter_41, False, False) @@ -1887,7 +1847,7 @@ def forward( # pd_op.add: (1x26x3072xf32) <- (1x26x3072xf32, 3072xf32) add_89 = paddle._C_ops.add(matmul_78, parameter_40) - del parameter_40 + del matmul_78, parameter_40 # pd_op.relu: (1x26x3072xf32) <- (1x26x3072xf32) relu_9 = paddle._C_ops.relu(add_89) @@ -1895,16 +1855,16 @@ def forward( # pd_op.matmul: (1x26x768xf32) <- (1x26x3072xf32, 3072x768xf32) matmul_79 = paddle._C_ops.matmul(relu_9, parameter_39, False, False) - del parameter_39 + del parameter_39, relu_9 # pd_op.add: (1x26x768xf32) <- (1x26x768xf32, 768xf32) add_90 = paddle._C_ops.add(matmul_79, parameter_38) - del parameter_38 + del matmul_79, parameter_38 # pd_op.dropout: (1x26x768xf32, 1x26x768xui8) <- (1x26x768xf32, None, 1xf32) dropout_60, dropout_61 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_90, None, full_4, False, "upscale_in_train", 0, False + add_90, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1912,6 +1872,7 @@ def forward( # pd_op.add: (1x26x768xf32) <- (1x26x768xf32, 1x26x768xf32) add_91 = paddle._C_ops.add(layer_norm_57, dropout_60) + del dropout_60, layer_norm_57 # pd_op.layer_norm: (1x26x768xf32, 1x26xf32, 1x26xf32) <- (1x26x768xf32, 768xf32, 768xf32) layer_norm_60, layer_norm_61, layer_norm_62 = (lambda x, f: f(x))( @@ -1920,7 +1881,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_34, parameter_35 + del add_91, parameter_34, parameter_35 # pd_op.matmul: (1x26x768xf32) <- (1x26x768xf32, 768x768xf32) matmul_80 = paddle._C_ops.matmul(layer_norm_60, parameter_33, False, False) @@ -1928,10 +1889,11 @@ def forward( # pd_op.add: (1x26x768xf32) <- (1x26x768xf32, 768xf32) add_92 = paddle._C_ops.add(matmul_80, parameter_32) - del parameter_32 + del matmul_80, parameter_32 # pd_op.reshape: (1x26x12x64xf32) <- (1x26x768xf32, 4xi64) reshape_40 = paddle._C_ops.reshape(add_92, full_int_array_1) + del add_92 # pd_op.transpose: (1x12x26x64xf32) <- (1x26x12x64xf32) transpose_40 = paddle._C_ops.transpose(reshape_40, [0, 2, 1, 3]) @@ -1943,7 +1905,7 @@ def forward( # pd_op.add: (1x26x768xf32) <- (1x26x768xf32, 768xf32) add_93 = paddle._C_ops.add(matmul_81, parameter_30) - del parameter_30 + del matmul_81, parameter_30 # pd_op.matmul: (1x26x768xf32) <- (1x26x768xf32, 768x768xf32) matmul_82 = paddle._C_ops.matmul(layer_norm_60, parameter_29, False, False) @@ -1951,10 +1913,11 @@ def forward( # pd_op.add: (1x26x768xf32) <- (1x26x768xf32, 768xf32) add_94 = paddle._C_ops.add(matmul_82, parameter_28) - del parameter_28 + del matmul_82, parameter_28 # pd_op.reshape: (1x26x12x64xf32) <- (1x26x768xf32, 4xi64) reshape_41 = paddle._C_ops.reshape(add_93, full_int_array_1) + del add_93 # pd_op.transpose: (1x12x26x64xf32) <- (1x26x12x64xf32) transpose_41 = paddle._C_ops.transpose(reshape_41, [0, 2, 1, 3]) @@ -1962,6 +1925,7 @@ def forward( # pd_op.reshape: (1x26x12x64xf32) <- (1x26x768xf32, 4xi64) reshape_42 = paddle._C_ops.reshape(add_94, full_int_array_1) + del add_94 # pd_op.transpose: (1x12x26x64xf32) <- (1x26x12x64xf32) transpose_42 = paddle._C_ops.transpose(reshape_42, [0, 2, 1, 3]) @@ -1973,9 +1937,11 @@ def forward( # pd_op.matmul: (1x12x26x26xf32) <- (1x12x26x64xf32, 1x12x26x64xf32) matmul_83 = paddle._C_ops.matmul(scale_11, transpose_41, False, True) + del scale_11, transpose_41 # pd_op.add: (1x12x26x26xf32) <- (1x12x26x26xf32, 1x1x1x26xf32) add_95 = paddle._C_ops.add(matmul_83, unsqueeze_0) + del matmul_83 # pd_op.softmax: (1x12x26x26xf32) <- (1x12x26x26xf32) softmax_10 = paddle._C_ops.softmax(add_95, -1) @@ -1984,13 +1950,15 @@ def forward( # pd_op.dropout: (1x12x26x26xf32, 1x12x26x26xui8) <- (1x12x26x26xf32, None, 1xf32) dropout_62, dropout_63 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_10, None, full_4, False, "upscale_in_train", 0, False + softmax_10, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_10 # pd_op.matmul: (1x12x26x64xf32) <- (1x12x26x26xf32, 1x12x26x64xf32) matmul_84 = paddle._C_ops.matmul(dropout_62, transpose_42, False, False) + del dropout_62, transpose_42 # pd_op.transpose: (1x26x12x64xf32) <- (1x12x26x64xf32) transpose_43 = paddle._C_ops.transpose(matmul_84, [0, 2, 1, 3]) @@ -1998,19 +1966,20 @@ def forward( # pd_op.reshape: (1x26x768xf32) <- (1x26x12x64xf32, 3xi64) reshape_43 = paddle._C_ops.reshape(transpose_43, full_int_array_2) + del transpose_43 # pd_op.matmul: (1x26x768xf32) <- (1x26x768xf32, 768x768xf32) matmul_85 = paddle._C_ops.matmul(reshape_43, parameter_27, False, False) - del parameter_27 + del parameter_27, reshape_43 # pd_op.add: (1x26x768xf32) <- (1x26x768xf32, 768xf32) add_96 = paddle._C_ops.add(matmul_85, parameter_26) - del parameter_26 + del matmul_85, parameter_26 # pd_op.dropout: (1x26x768xf32, 1x26x768xui8) <- (1x26x768xf32, None, 1xf32) dropout_64, dropout_65 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_96, None, full_4, False, "upscale_in_train", 0, False + add_96, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -2018,6 +1987,7 @@ def forward( # pd_op.add: (1x26x768xf32) <- (1x26x768xf32, 1x26x768xf32) add_97 = paddle._C_ops.add(layer_norm_60, dropout_64) + del dropout_64, layer_norm_60 # pd_op.layer_norm: (1x26x768xf32, 1x26xf32, 1x26xf32) <- (1x26x768xf32, 768xf32, 768xf32) layer_norm_63, layer_norm_64, layer_norm_65 = (lambda x, f: f(x))( @@ -2026,7 +1996,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_20, parameter_21 + del add_97, parameter_20, parameter_21 # pd_op.matmul: (1x26x3072xf32) <- (1x26x768xf32, 768x3072xf32) matmul_86 = paddle._C_ops.matmul(layer_norm_63, parameter_25, False, False) @@ -2034,7 +2004,7 @@ def forward( # pd_op.add: (1x26x3072xf32) <- (1x26x3072xf32, 3072xf32) add_98 = paddle._C_ops.add(matmul_86, parameter_24) - del parameter_24 + del matmul_86, parameter_24 # pd_op.relu: (1x26x3072xf32) <- (1x26x3072xf32) relu_10 = paddle._C_ops.relu(add_98) @@ -2042,16 +2012,16 @@ def forward( # pd_op.matmul: (1x26x768xf32) <- (1x26x3072xf32, 3072x768xf32) matmul_87 = paddle._C_ops.matmul(relu_10, parameter_23, False, False) - del parameter_23 + del parameter_23, relu_10 # pd_op.add: (1x26x768xf32) <- (1x26x768xf32, 768xf32) add_99 = paddle._C_ops.add(matmul_87, parameter_22) - del parameter_22 + del matmul_87, parameter_22 # pd_op.dropout: (1x26x768xf32, 1x26x768xui8) <- (1x26x768xf32, None, 1xf32) dropout_66, dropout_67 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_99, None, full_4, False, "upscale_in_train", 0, False + add_99, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -2059,6 +2029,7 @@ def forward( # pd_op.add: (1x26x768xf32) <- (1x26x768xf32, 1x26x768xf32) add_100 = paddle._C_ops.add(layer_norm_63, dropout_66) + del dropout_66, layer_norm_63 # pd_op.layer_norm: (1x26x768xf32, 1x26xf32, 1x26xf32) <- (1x26x768xf32, 768xf32, 768xf32) layer_norm_66, layer_norm_67, layer_norm_68 = (lambda x, f: f(x))( @@ -2067,7 +2038,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_18, parameter_19 + del add_100, parameter_18, parameter_19 # pd_op.matmul: (1x26x768xf32) <- (1x26x768xf32, 768x768xf32) matmul_88 = paddle._C_ops.matmul(layer_norm_66, parameter_17, False, False) @@ -2075,10 +2046,11 @@ def forward( # pd_op.add: (1x26x768xf32) <- (1x26x768xf32, 768xf32) add_101 = paddle._C_ops.add(matmul_88, parameter_16) - del parameter_16 + del matmul_88, parameter_16 # pd_op.reshape: (1x26x12x64xf32) <- (1x26x768xf32, 4xi64) reshape_44 = paddle._C_ops.reshape(add_101, full_int_array_1) + del add_101 # pd_op.transpose: (1x12x26x64xf32) <- (1x26x12x64xf32) transpose_44 = paddle._C_ops.transpose(reshape_44, [0, 2, 1, 3]) @@ -2090,7 +2062,7 @@ def forward( # pd_op.add: (1x26x768xf32) <- (1x26x768xf32, 768xf32) add_102 = paddle._C_ops.add(matmul_89, parameter_14) - del parameter_14 + del matmul_89, parameter_14 # pd_op.matmul: (1x26x768xf32) <- (1x26x768xf32, 768x768xf32) matmul_90 = paddle._C_ops.matmul(layer_norm_66, parameter_13, False, False) @@ -2098,10 +2070,11 @@ def forward( # pd_op.add: (1x26x768xf32) <- (1x26x768xf32, 768xf32) add_103 = paddle._C_ops.add(matmul_90, parameter_12) - del parameter_12 + del matmul_90, parameter_12 # pd_op.reshape: (1x26x12x64xf32) <- (1x26x768xf32, 4xi64) reshape_45 = paddle._C_ops.reshape(add_102, full_int_array_1) + del add_102 # pd_op.transpose: (1x12x26x64xf32) <- (1x26x12x64xf32) transpose_45 = paddle._C_ops.transpose(reshape_45, [0, 2, 1, 3]) @@ -2109,7 +2082,7 @@ def forward( # pd_op.reshape: (1x26x12x64xf32) <- (1x26x768xf32, 4xi64) reshape_46 = paddle._C_ops.reshape(add_103, full_int_array_1) - del full_int_array_1 + del add_103, full_int_array_1 # pd_op.transpose: (1x12x26x64xf32) <- (1x26x12x64xf32) transpose_46 = paddle._C_ops.transpose(reshape_46, [0, 2, 1, 3]) @@ -2117,13 +2090,15 @@ def forward( # pd_op.scale: (1x12x26x64xf32) <- (1x12x26x64xf32, 1xf32) scale_12 = paddle._C_ops.scale(transpose_44, full_5, float("0"), True) - del transpose_44 + del full_5, transpose_44 # pd_op.matmul: (1x12x26x26xf32) <- (1x12x26x64xf32, 1x12x26x64xf32) matmul_91 = paddle._C_ops.matmul(scale_12, transpose_45, False, True) + del scale_12, transpose_45 # pd_op.add: (1x12x26x26xf32) <- (1x12x26x26xf32, 1x1x1x26xf32) add_104 = paddle._C_ops.add(matmul_91, unsqueeze_0) + del matmul_91, unsqueeze_0 # pd_op.softmax: (1x12x26x26xf32) <- (1x12x26x26xf32) softmax_11 = paddle._C_ops.softmax(add_104, -1) @@ -2132,13 +2107,15 @@ def forward( # pd_op.dropout: (1x12x26x26xf32, 1x12x26x26xui8) <- (1x12x26x26xf32, None, 1xf32) dropout_68, dropout_69 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_11, None, full_4, False, "upscale_in_train", 0, False + softmax_11, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_11 # pd_op.matmul: (1x12x26x64xf32) <- (1x12x26x26xf32, 1x12x26x64xf32) matmul_92 = paddle._C_ops.matmul(dropout_68, transpose_46, False, False) + del dropout_68, transpose_46 # pd_op.transpose: (1x26x12x64xf32) <- (1x12x26x64xf32) transpose_47 = paddle._C_ops.transpose(matmul_92, [0, 2, 1, 3]) @@ -2146,20 +2123,20 @@ def forward( # pd_op.reshape: (1x26x768xf32) <- (1x26x12x64xf32, 3xi64) reshape_47 = paddle._C_ops.reshape(transpose_47, full_int_array_2) - del full_int_array_2 + del full_int_array_2, transpose_47 # pd_op.matmul: (1x26x768xf32) <- (1x26x768xf32, 768x768xf32) matmul_93 = paddle._C_ops.matmul(reshape_47, parameter_11, False, False) - del parameter_11 + del parameter_11, reshape_47 # pd_op.add: (1x26x768xf32) <- (1x26x768xf32, 768xf32) add_105 = paddle._C_ops.add(matmul_93, parameter_10) - del parameter_10 + del matmul_93, parameter_10 # pd_op.dropout: (1x26x768xf32, 1x26x768xui8) <- (1x26x768xf32, None, 1xf32) dropout_70, dropout_71 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_105, None, full_4, False, "upscale_in_train", 0, False + add_105, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -2167,6 +2144,7 @@ def forward( # pd_op.add: (1x26x768xf32) <- (1x26x768xf32, 1x26x768xf32) add_106 = paddle._C_ops.add(layer_norm_66, dropout_70) + del dropout_70, layer_norm_66 # pd_op.layer_norm: (1x26x768xf32, 1x26xf32, 1x26xf32) <- (1x26x768xf32, 768xf32, 768xf32) layer_norm_69, layer_norm_70, layer_norm_71 = (lambda x, f: f(x))( @@ -2175,7 +2153,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_4, parameter_5 + del add_106, parameter_4, parameter_5 # pd_op.matmul: (1x26x3072xf32) <- (1x26x768xf32, 768x3072xf32) matmul_94 = paddle._C_ops.matmul(layer_norm_69, parameter_9, False, False) @@ -2183,7 +2161,7 @@ def forward( # pd_op.add: (1x26x3072xf32) <- (1x26x3072xf32, 3072xf32) add_107 = paddle._C_ops.add(matmul_94, parameter_8) - del parameter_8 + del matmul_94, parameter_8 # pd_op.relu: (1x26x3072xf32) <- (1x26x3072xf32) relu_11 = paddle._C_ops.relu(add_107) @@ -2191,23 +2169,24 @@ def forward( # pd_op.matmul: (1x26x768xf32) <- (1x26x3072xf32, 3072x768xf32) matmul_95 = paddle._C_ops.matmul(relu_11, parameter_7, False, False) - del parameter_7 + del parameter_7, relu_11 # pd_op.add: (1x26x768xf32) <- (1x26x768xf32, 768xf32) add_108 = paddle._C_ops.add(matmul_95, parameter_6) - del parameter_6 + del matmul_95, parameter_6 # pd_op.dropout: (1x26x768xf32, 1x26x768xui8) <- (1x26x768xf32, None, 1xf32) dropout_72, dropout_73 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_108, None, full_4, False, "upscale_in_train", 0, False + add_108, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) - del add_108 + del add_108, full_4 # pd_op.add: (1x26x768xf32) <- (1x26x768xf32, 1x26x768xf32) add_109 = paddle._C_ops.add(layer_norm_69, dropout_72) + del dropout_72, layer_norm_69 # pd_op.layer_norm: (1x26x768xf32, 1x26xf32, 1x26xf32) <- (1x26x768xf32, 768xf32, 768xf32) layer_norm_72, layer_norm_73, layer_norm_74 = (lambda x, f: f(x))( @@ -2216,7 +2195,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_2, parameter_3 + del add_109, parameter_2, parameter_3 # pd_op.full_int_array: (1xi64) <- () full_int_array_3 = [0] @@ -2228,455 +2207,18 @@ def forward( slice_0 = paddle._C_ops.slice( layer_norm_72, [1], full_int_array_3, full_int_array_4, [1], [1] ) + del full_int_array_3, full_int_array_4, layer_norm_72 # pd_op.matmul: (1x768xf32) <- (1x768xf32, 768x768xf32) matmul_96 = paddle._C_ops.matmul(slice_0, parameter_1, False, False) - del parameter_1 + del parameter_1, slice_0 # pd_op.add: (1x768xf32) <- (1x768xf32, 768xf32) add_110 = paddle._C_ops.add(matmul_96, parameter_0) - del parameter_0 + del matmul_96, parameter_0 # pd_op.tanh: (1x768xf32) <- (1x768xf32) tanh_0 = paddle._C_ops.tanh(add_110) - del ( - add_0, - add_1, - add_10, - add_100, - add_101, - add_102, - add_103, - add_106, - add_109, - add_11, - add_110, - add_12, - add_13, - add_16, - add_19, - add_2, - add_20, - add_21, - add_22, - add_25, - add_28, - add_29, - add_3, - add_30, - add_31, - add_34, - add_37, - add_38, - add_39, - add_4, - add_40, - add_43, - add_46, - add_47, - add_48, - add_49, - add_52, - add_55, - add_56, - add_57, - add_58, - add_61, - add_64, - add_65, - add_66, - add_67, - add_7, - add_70, - add_73, - add_74, - add_75, - add_76, - add_79, - add_82, - add_83, - add_84, - add_85, - add_88, - add_91, - add_92, - add_93, - add_94, - add_97, - assign_0, - assign_1, - assign_10, - assign_11, - assign_12, - assign_13, - assign_14, - assign_15, - assign_16, - assign_17, - assign_18, - assign_19, - assign_2, - assign_20, - assign_21, - assign_22, - assign_23, - assign_24, - assign_25, - assign_26, - assign_27, - assign_28, - assign_29, - assign_3, - assign_30, - assign_31, - assign_32, - assign_33, - assign_34, - assign_35, - assign_36, - assign_37, - assign_38, - assign_39, - assign_4, - assign_40, - assign_41, - assign_42, - assign_43, - assign_44, - assign_45, - assign_46, - assign_5, - assign_6, - assign_7, - assign_8, - assign_9, - dropout_0, - dropout_1, - dropout_10, - dropout_11, - dropout_12, - dropout_13, - dropout_14, - dropout_15, - dropout_16, - dropout_17, - dropout_18, - dropout_19, - dropout_2, - dropout_20, - dropout_21, - dropout_22, - dropout_23, - dropout_24, - dropout_25, - dropout_26, - dropout_27, - dropout_28, - dropout_29, - dropout_3, - dropout_30, - dropout_31, - dropout_32, - dropout_33, - dropout_34, - dropout_35, - dropout_36, - dropout_37, - dropout_38, - dropout_39, - dropout_4, - dropout_40, - dropout_41, - dropout_42, - dropout_43, - dropout_44, - dropout_45, - dropout_46, - dropout_47, - dropout_48, - dropout_49, - dropout_5, - dropout_50, - dropout_51, - dropout_52, - dropout_53, - dropout_54, - dropout_55, - dropout_56, - dropout_57, - dropout_58, - dropout_59, - dropout_6, - dropout_60, - dropout_61, - dropout_62, - dropout_63, - dropout_64, - dropout_65, - dropout_66, - dropout_67, - dropout_68, - dropout_69, - dropout_7, - dropout_70, - dropout_71, - dropout_72, - dropout_73, - dropout_8, - dropout_9, - embedding_0, - embedding_1, - embedding_2, - full_4, - full_5, - full_int_array_3, - full_int_array_4, - layer_norm_1, - layer_norm_10, - layer_norm_11, - layer_norm_12, - layer_norm_13, - layer_norm_14, - layer_norm_15, - layer_norm_16, - layer_norm_17, - layer_norm_18, - layer_norm_19, - layer_norm_2, - layer_norm_20, - layer_norm_21, - layer_norm_22, - layer_norm_23, - layer_norm_24, - layer_norm_25, - layer_norm_26, - layer_norm_27, - layer_norm_28, - layer_norm_29, - layer_norm_3, - layer_norm_30, - layer_norm_31, - layer_norm_32, - layer_norm_33, - layer_norm_34, - layer_norm_35, - layer_norm_36, - layer_norm_37, - layer_norm_38, - layer_norm_39, - layer_norm_4, - layer_norm_40, - layer_norm_41, - layer_norm_42, - layer_norm_43, - layer_norm_44, - layer_norm_45, - layer_norm_46, - layer_norm_47, - layer_norm_48, - layer_norm_49, - layer_norm_5, - layer_norm_50, - layer_norm_51, - layer_norm_52, - layer_norm_53, - layer_norm_54, - layer_norm_55, - layer_norm_56, - layer_norm_57, - layer_norm_58, - layer_norm_59, - layer_norm_6, - layer_norm_60, - layer_norm_61, - layer_norm_62, - layer_norm_63, - layer_norm_64, - layer_norm_65, - layer_norm_66, - layer_norm_67, - layer_norm_68, - layer_norm_69, - layer_norm_7, - layer_norm_70, - layer_norm_71, - layer_norm_72, - layer_norm_73, - layer_norm_74, - layer_norm_8, - layer_norm_9, - matmul_0, - matmul_1, - matmul_10, - matmul_11, - matmul_13, - matmul_14, - matmul_15, - matmul_16, - matmul_17, - matmul_18, - matmul_19, - matmul_2, - matmul_21, - matmul_22, - matmul_23, - matmul_24, - matmul_25, - matmul_26, - matmul_27, - matmul_29, - matmul_3, - matmul_30, - matmul_31, - matmul_32, - matmul_33, - matmul_34, - matmul_35, - matmul_37, - matmul_38, - matmul_39, - matmul_40, - matmul_41, - matmul_42, - matmul_43, - matmul_45, - matmul_46, - matmul_47, - matmul_48, - matmul_49, - matmul_5, - matmul_50, - matmul_51, - matmul_53, - matmul_54, - matmul_55, - matmul_56, - matmul_57, - matmul_58, - matmul_59, - matmul_6, - matmul_61, - matmul_62, - matmul_63, - matmul_64, - matmul_65, - matmul_66, - matmul_67, - matmul_69, - matmul_7, - matmul_70, - matmul_71, - matmul_72, - matmul_73, - matmul_74, - matmul_75, - matmul_77, - matmul_78, - matmul_79, - matmul_8, - matmul_80, - matmul_81, - matmul_82, - matmul_83, - matmul_85, - matmul_86, - matmul_87, - matmul_88, - matmul_89, - matmul_9, - matmul_90, - matmul_91, - matmul_93, - matmul_94, - matmul_95, - matmul_96, - relu_0, - relu_1, - relu_10, - relu_11, - relu_2, - relu_3, - relu_4, - relu_5, - relu_6, - relu_7, - relu_8, - relu_9, - reshape_11, - reshape_15, - reshape_19, - reshape_23, - reshape_27, - reshape_3, - reshape_31, - reshape_35, - reshape_39, - reshape_43, - reshape_47, - reshape_7, - scale_1, - scale_10, - scale_11, - scale_12, - scale_2, - scale_3, - scale_4, - scale_5, - scale_6, - scale_7, - scale_8, - scale_9, - slice_0, - softmax_0, - softmax_1, - softmax_10, - softmax_11, - softmax_2, - softmax_3, - softmax_4, - softmax_5, - softmax_6, - softmax_7, - softmax_8, - softmax_9, - subtract_0, - transpose_1, - transpose_10, - transpose_11, - transpose_13, - transpose_14, - transpose_15, - transpose_17, - transpose_18, - transpose_19, - transpose_2, - transpose_21, - transpose_22, - transpose_23, - transpose_25, - transpose_26, - transpose_27, - transpose_29, - transpose_3, - transpose_30, - transpose_31, - transpose_33, - transpose_34, - transpose_35, - transpose_37, - transpose_38, - transpose_39, - transpose_41, - transpose_42, - transpose_43, - transpose_45, - transpose_46, - transpose_47, - transpose_5, - transpose_6, - transpose_7, - transpose_9, - unsqueeze_0, - ) + del add_110 return tanh_0 diff --git a/paddle_samples/PaddleNLP/ernie-1.0/weight_meta.py b/paddle_samples/PaddleNLP/ernie-1.0/weight_meta.py index ee5053f89..48e295a7a 100644 --- a/paddle_samples/PaddleNLP/ernie-1.0/weight_meta.py +++ b/paddle_samples/PaddleNLP/ernie-1.0/weight_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_parameter_0: name = "parameter_0" + original_name = "linear_363.b_0" shape = [768] dtype = "float32" min_val = float("-0.480411") @@ -11,6 +12,7 @@ class Program_weight_tensor_parameter_0: class Program_weight_tensor_parameter_1: name = "parameter_1" + original_name = "linear_363.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.324589") @@ -22,6 +24,7 @@ class Program_weight_tensor_parameter_1: class Program_weight_tensor_parameter_2: name = "parameter_2" + original_name = "layer_norm_123.b_0" shape = [768] dtype = "float32" min_val = float("-2.05149") @@ -33,6 +36,7 @@ class Program_weight_tensor_parameter_2: class Program_weight_tensor_parameter_3: name = "parameter_3" + original_name = "layer_norm_123.w_0" shape = [768] dtype = "float32" min_val = float("0.560708") @@ -44,6 +48,7 @@ class Program_weight_tensor_parameter_3: class Program_weight_tensor_parameter_4: name = "parameter_4" + original_name = "layer_norm_122.b_0" shape = [768] dtype = "float32" min_val = float("-2.93306") @@ -55,6 +60,7 @@ class Program_weight_tensor_parameter_4: class Program_weight_tensor_parameter_5: name = "parameter_5" + original_name = "layer_norm_122.w_0" shape = [768] dtype = "float32" min_val = float("0.224299") @@ -66,6 +72,7 @@ class Program_weight_tensor_parameter_5: class Program_weight_tensor_parameter_6: name = "parameter_6" + original_name = "linear_362.b_0" shape = [768] dtype = "float32" min_val = float("-0.170263") @@ -77,6 +84,7 @@ class Program_weight_tensor_parameter_6: class Program_weight_tensor_parameter_7: name = "parameter_7" + original_name = "linear_362.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-0.308771") @@ -88,6 +96,7 @@ class Program_weight_tensor_parameter_7: class Program_weight_tensor_parameter_8: name = "parameter_8" + original_name = "linear_361.b_0" shape = [3072] dtype = "float32" min_val = float("-2.41867") @@ -99,6 +108,7 @@ class Program_weight_tensor_parameter_8: class Program_weight_tensor_parameter_9: name = "parameter_9" + original_name = "linear_361.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.840451") @@ -110,6 +120,7 @@ class Program_weight_tensor_parameter_9: class Program_weight_tensor_parameter_10: name = "parameter_10" + original_name = "linear_360.b_0" shape = [768] dtype = "float32" min_val = float("-0.216988") @@ -121,6 +132,7 @@ class Program_weight_tensor_parameter_10: class Program_weight_tensor_parameter_11: name = "parameter_11" + original_name = "linear_360.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.665468") @@ -132,6 +144,7 @@ class Program_weight_tensor_parameter_11: class Program_weight_tensor_parameter_12: name = "parameter_12" + original_name = "linear_359.b_0" shape = [768] dtype = "float32" min_val = float("-0.531242") @@ -143,6 +156,7 @@ class Program_weight_tensor_parameter_12: class Program_weight_tensor_parameter_13: name = "parameter_13" + original_name = "linear_359.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.294358") @@ -154,6 +168,7 @@ class Program_weight_tensor_parameter_13: class Program_weight_tensor_parameter_14: name = "parameter_14" + original_name = "linear_358.b_0" shape = [768] dtype = "float32" min_val = float("-3.36714") @@ -165,6 +180,7 @@ class Program_weight_tensor_parameter_14: class Program_weight_tensor_parameter_15: name = "parameter_15" + original_name = "linear_358.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.825853") @@ -176,6 +192,7 @@ class Program_weight_tensor_parameter_15: class Program_weight_tensor_parameter_16: name = "parameter_16" + original_name = "linear_357.b_0" shape = [768] dtype = "float32" min_val = float("-2.44784") @@ -187,6 +204,7 @@ class Program_weight_tensor_parameter_16: class Program_weight_tensor_parameter_17: name = "parameter_17" + original_name = "linear_357.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.723588") @@ -198,6 +216,7 @@ class Program_weight_tensor_parameter_17: class Program_weight_tensor_parameter_18: name = "parameter_18" + original_name = "layer_norm_121.b_0" shape = [768] dtype = "float32" min_val = float("-2.2183") @@ -209,6 +228,7 @@ class Program_weight_tensor_parameter_18: class Program_weight_tensor_parameter_19: name = "parameter_19" + original_name = "layer_norm_121.w_0" shape = [768] dtype = "float32" min_val = float("0.138721") @@ -220,6 +240,7 @@ class Program_weight_tensor_parameter_19: class Program_weight_tensor_parameter_20: name = "parameter_20" + original_name = "layer_norm_120.b_0" shape = [768] dtype = "float32" min_val = float("-4.26608") @@ -231,6 +252,7 @@ class Program_weight_tensor_parameter_20: class Program_weight_tensor_parameter_21: name = "parameter_21" + original_name = "layer_norm_120.w_0" shape = [768] dtype = "float32" min_val = float("0.484979") @@ -242,6 +264,7 @@ class Program_weight_tensor_parameter_21: class Program_weight_tensor_parameter_22: name = "parameter_22" + original_name = "linear_356.b_0" shape = [768] dtype = "float32" min_val = float("-0.31975") @@ -253,6 +276,7 @@ class Program_weight_tensor_parameter_22: class Program_weight_tensor_parameter_23: name = "parameter_23" + original_name = "linear_356.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-1.15468") @@ -264,6 +288,7 @@ class Program_weight_tensor_parameter_23: class Program_weight_tensor_parameter_24: name = "parameter_24" + original_name = "linear_355.b_0" shape = [3072] dtype = "float32" min_val = float("-1.84535") @@ -275,6 +300,7 @@ class Program_weight_tensor_parameter_24: class Program_weight_tensor_parameter_25: name = "parameter_25" + original_name = "linear_355.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.706796") @@ -286,6 +312,7 @@ class Program_weight_tensor_parameter_25: class Program_weight_tensor_parameter_26: name = "parameter_26" + original_name = "linear_354.b_0" shape = [768] dtype = "float32" min_val = float("-0.174765") @@ -297,6 +324,7 @@ class Program_weight_tensor_parameter_26: class Program_weight_tensor_parameter_27: name = "parameter_27" + original_name = "linear_354.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.261259") @@ -308,6 +336,7 @@ class Program_weight_tensor_parameter_27: class Program_weight_tensor_parameter_28: name = "parameter_28" + original_name = "linear_353.b_0" shape = [768] dtype = "float32" min_val = float("-0.305205") @@ -319,6 +348,7 @@ class Program_weight_tensor_parameter_28: class Program_weight_tensor_parameter_29: name = "parameter_29" + original_name = "linear_353.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.367948") @@ -330,6 +360,7 @@ class Program_weight_tensor_parameter_29: class Program_weight_tensor_parameter_30: name = "parameter_30" + original_name = "linear_352.b_0" shape = [768] dtype = "float32" min_val = float("-4.90769") @@ -341,6 +372,7 @@ class Program_weight_tensor_parameter_30: class Program_weight_tensor_parameter_31: name = "parameter_31" + original_name = "linear_352.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.688865") @@ -352,6 +384,7 @@ class Program_weight_tensor_parameter_31: class Program_weight_tensor_parameter_32: name = "parameter_32" + original_name = "linear_351.b_0" shape = [768] dtype = "float32" min_val = float("-2.55582") @@ -363,6 +396,7 @@ class Program_weight_tensor_parameter_32: class Program_weight_tensor_parameter_33: name = "parameter_33" + original_name = "linear_351.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.509483") @@ -374,6 +408,7 @@ class Program_weight_tensor_parameter_33: class Program_weight_tensor_parameter_34: name = "parameter_34" + original_name = "layer_norm_119.b_0" shape = [768] dtype = "float32" min_val = float("-0.189531") @@ -385,6 +420,7 @@ class Program_weight_tensor_parameter_34: class Program_weight_tensor_parameter_35: name = "parameter_35" + original_name = "layer_norm_119.w_0" shape = [768] dtype = "float32" min_val = float("0.158298") @@ -396,6 +432,7 @@ class Program_weight_tensor_parameter_35: class Program_weight_tensor_parameter_36: name = "parameter_36" + original_name = "layer_norm_118.b_0" shape = [768] dtype = "float32" min_val = float("-4.9984") @@ -407,6 +444,7 @@ class Program_weight_tensor_parameter_36: class Program_weight_tensor_parameter_37: name = "parameter_37" + original_name = "layer_norm_118.w_0" shape = [768] dtype = "float32" min_val = float("0.387073") @@ -418,6 +456,7 @@ class Program_weight_tensor_parameter_37: class Program_weight_tensor_parameter_38: name = "parameter_38" + original_name = "linear_350.b_0" shape = [768] dtype = "float32" min_val = float("-1.01426") @@ -429,6 +468,7 @@ class Program_weight_tensor_parameter_38: class Program_weight_tensor_parameter_39: name = "parameter_39" + original_name = "linear_350.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-0.589946") @@ -440,6 +480,7 @@ class Program_weight_tensor_parameter_39: class Program_weight_tensor_parameter_40: name = "parameter_40" + original_name = "linear_349.b_0" shape = [3072] dtype = "float32" min_val = float("-1.074") @@ -451,6 +492,7 @@ class Program_weight_tensor_parameter_40: class Program_weight_tensor_parameter_41: name = "parameter_41" + original_name = "linear_349.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.831262") @@ -462,6 +504,7 @@ class Program_weight_tensor_parameter_41: class Program_weight_tensor_parameter_42: name = "parameter_42" + original_name = "linear_348.b_0" shape = [768] dtype = "float32" min_val = float("-0.214419") @@ -473,6 +516,7 @@ class Program_weight_tensor_parameter_42: class Program_weight_tensor_parameter_43: name = "parameter_43" + original_name = "linear_348.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.302823") @@ -484,6 +528,7 @@ class Program_weight_tensor_parameter_43: class Program_weight_tensor_parameter_44: name = "parameter_44" + original_name = "linear_347.b_0" shape = [768] dtype = "float32" min_val = float("-0.861201") @@ -495,6 +540,7 @@ class Program_weight_tensor_parameter_44: class Program_weight_tensor_parameter_45: name = "parameter_45" + original_name = "linear_347.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.467219") @@ -506,6 +552,7 @@ class Program_weight_tensor_parameter_45: class Program_weight_tensor_parameter_46: name = "parameter_46" + original_name = "linear_346.b_0" shape = [768] dtype = "float32" min_val = float("-3.55039") @@ -517,6 +564,7 @@ class Program_weight_tensor_parameter_46: class Program_weight_tensor_parameter_47: name = "parameter_47" + original_name = "linear_346.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.60493") @@ -528,6 +576,7 @@ class Program_weight_tensor_parameter_47: class Program_weight_tensor_parameter_48: name = "parameter_48" + original_name = "linear_345.b_0" shape = [768] dtype = "float32" min_val = float("-2.24882") @@ -539,6 +588,7 @@ class Program_weight_tensor_parameter_48: class Program_weight_tensor_parameter_49: name = "parameter_49" + original_name = "linear_345.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.597795") @@ -550,6 +600,7 @@ class Program_weight_tensor_parameter_49: class Program_weight_tensor_parameter_50: name = "parameter_50" + original_name = "layer_norm_117.b_0" shape = [768] dtype = "float32" min_val = float("-0.42158") @@ -561,6 +612,7 @@ class Program_weight_tensor_parameter_50: class Program_weight_tensor_parameter_51: name = "parameter_51" + original_name = "layer_norm_117.w_0" shape = [768] dtype = "float32" min_val = float("0.072916") @@ -572,6 +624,7 @@ class Program_weight_tensor_parameter_51: class Program_weight_tensor_parameter_52: name = "parameter_52" + original_name = "layer_norm_116.b_0" shape = [768] dtype = "float32" min_val = float("-5.34463") @@ -583,6 +636,7 @@ class Program_weight_tensor_parameter_52: class Program_weight_tensor_parameter_53: name = "parameter_53" + original_name = "layer_norm_116.w_0" shape = [768] dtype = "float32" min_val = float("0.334325") @@ -594,6 +648,7 @@ class Program_weight_tensor_parameter_53: class Program_weight_tensor_parameter_54: name = "parameter_54" + original_name = "linear_344.b_0" shape = [768] dtype = "float32" min_val = float("-0.859261") @@ -605,6 +660,7 @@ class Program_weight_tensor_parameter_54: class Program_weight_tensor_parameter_55: name = "parameter_55" + original_name = "linear_344.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-2.79176") @@ -616,6 +672,7 @@ class Program_weight_tensor_parameter_55: class Program_weight_tensor_parameter_56: name = "parameter_56" + original_name = "linear_343.b_0" shape = [3072] dtype = "float32" min_val = float("-1.14441") @@ -627,6 +684,7 @@ class Program_weight_tensor_parameter_56: class Program_weight_tensor_parameter_57: name = "parameter_57" + original_name = "linear_343.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.734106") @@ -638,6 +696,7 @@ class Program_weight_tensor_parameter_57: class Program_weight_tensor_parameter_58: name = "parameter_58" + original_name = "linear_342.b_0" shape = [768] dtype = "float32" min_val = float("-0.166051") @@ -649,6 +708,7 @@ class Program_weight_tensor_parameter_58: class Program_weight_tensor_parameter_59: name = "parameter_59" + original_name = "linear_342.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.357285") @@ -660,6 +720,7 @@ class Program_weight_tensor_parameter_59: class Program_weight_tensor_parameter_60: name = "parameter_60" + original_name = "linear_341.b_0" shape = [768] dtype = "float32" min_val = float("-0.50267") @@ -671,6 +732,7 @@ class Program_weight_tensor_parameter_60: class Program_weight_tensor_parameter_61: name = "parameter_61" + original_name = "linear_341.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.298987") @@ -682,6 +744,7 @@ class Program_weight_tensor_parameter_61: class Program_weight_tensor_parameter_62: name = "parameter_62" + original_name = "linear_340.b_0" shape = [768] dtype = "float32" min_val = float("-3.13558") @@ -693,6 +756,7 @@ class Program_weight_tensor_parameter_62: class Program_weight_tensor_parameter_63: name = "parameter_63" + original_name = "linear_340.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.709055") @@ -704,6 +768,7 @@ class Program_weight_tensor_parameter_63: class Program_weight_tensor_parameter_64: name = "parameter_64" + original_name = "linear_339.b_0" shape = [768] dtype = "float32" min_val = float("-2.34578") @@ -715,6 +780,7 @@ class Program_weight_tensor_parameter_64: class Program_weight_tensor_parameter_65: name = "parameter_65" + original_name = "linear_339.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.807434") @@ -726,6 +792,7 @@ class Program_weight_tensor_parameter_65: class Program_weight_tensor_parameter_66: name = "parameter_66" + original_name = "layer_norm_115.b_0" shape = [768] dtype = "float32" min_val = float("-0.402449") @@ -737,6 +804,7 @@ class Program_weight_tensor_parameter_66: class Program_weight_tensor_parameter_67: name = "parameter_67" + original_name = "layer_norm_115.w_0" shape = [768] dtype = "float32" min_val = float("0.0818056") @@ -748,6 +816,7 @@ class Program_weight_tensor_parameter_67: class Program_weight_tensor_parameter_68: name = "parameter_68" + original_name = "layer_norm_114.b_0" shape = [768] dtype = "float32" min_val = float("-5.17545") @@ -759,6 +828,7 @@ class Program_weight_tensor_parameter_68: class Program_weight_tensor_parameter_69: name = "parameter_69" + original_name = "layer_norm_114.w_0" shape = [768] dtype = "float32" min_val = float("0.252543") @@ -770,6 +840,7 @@ class Program_weight_tensor_parameter_69: class Program_weight_tensor_parameter_70: name = "parameter_70" + original_name = "linear_338.b_0" shape = [768] dtype = "float32" min_val = float("-0.972878") @@ -781,6 +852,7 @@ class Program_weight_tensor_parameter_70: class Program_weight_tensor_parameter_71: name = "parameter_71" + original_name = "linear_338.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-1.69599") @@ -792,6 +864,7 @@ class Program_weight_tensor_parameter_71: class Program_weight_tensor_parameter_72: name = "parameter_72" + original_name = "linear_337.b_0" shape = [3072] dtype = "float32" min_val = float("-1.26354") @@ -803,6 +876,7 @@ class Program_weight_tensor_parameter_72: class Program_weight_tensor_parameter_73: name = "parameter_73" + original_name = "linear_337.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.78568") @@ -814,6 +888,7 @@ class Program_weight_tensor_parameter_73: class Program_weight_tensor_parameter_74: name = "parameter_74" + original_name = "linear_336.b_0" shape = [768] dtype = "float32" min_val = float("-0.233705") @@ -825,6 +900,7 @@ class Program_weight_tensor_parameter_74: class Program_weight_tensor_parameter_75: name = "parameter_75" + original_name = "linear_336.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.493672") @@ -836,6 +912,7 @@ class Program_weight_tensor_parameter_75: class Program_weight_tensor_parameter_76: name = "parameter_76" + original_name = "linear_335.b_0" shape = [768] dtype = "float32" min_val = float("-0.733703") @@ -847,6 +924,7 @@ class Program_weight_tensor_parameter_76: class Program_weight_tensor_parameter_77: name = "parameter_77" + original_name = "linear_335.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.348997") @@ -858,6 +936,7 @@ class Program_weight_tensor_parameter_77: class Program_weight_tensor_parameter_78: name = "parameter_78" + original_name = "linear_334.b_0" shape = [768] dtype = "float32" min_val = float("-2.21458") @@ -869,6 +948,7 @@ class Program_weight_tensor_parameter_78: class Program_weight_tensor_parameter_79: name = "parameter_79" + original_name = "linear_334.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.79912") @@ -880,6 +960,7 @@ class Program_weight_tensor_parameter_79: class Program_weight_tensor_parameter_80: name = "parameter_80" + original_name = "linear_333.b_0" shape = [768] dtype = "float32" min_val = float("-1.76664") @@ -891,6 +972,7 @@ class Program_weight_tensor_parameter_80: class Program_weight_tensor_parameter_81: name = "parameter_81" + original_name = "linear_333.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.82126") @@ -902,6 +984,7 @@ class Program_weight_tensor_parameter_81: class Program_weight_tensor_parameter_82: name = "parameter_82" + original_name = "layer_norm_113.b_0" shape = [768] dtype = "float32" min_val = float("-0.509987") @@ -913,6 +996,7 @@ class Program_weight_tensor_parameter_82: class Program_weight_tensor_parameter_83: name = "parameter_83" + original_name = "layer_norm_113.w_0" shape = [768] dtype = "float32" min_val = float("0.0598216") @@ -924,6 +1008,7 @@ class Program_weight_tensor_parameter_83: class Program_weight_tensor_parameter_84: name = "parameter_84" + original_name = "layer_norm_112.b_0" shape = [768] dtype = "float32" min_val = float("-5.78951") @@ -935,6 +1020,7 @@ class Program_weight_tensor_parameter_84: class Program_weight_tensor_parameter_85: name = "parameter_85" + original_name = "layer_norm_112.w_0" shape = [768] dtype = "float32" min_val = float("0.202463") @@ -946,6 +1032,7 @@ class Program_weight_tensor_parameter_85: class Program_weight_tensor_parameter_86: name = "parameter_86" + original_name = "linear_332.b_0" shape = [768] dtype = "float32" min_val = float("-0.736747") @@ -957,6 +1044,7 @@ class Program_weight_tensor_parameter_86: class Program_weight_tensor_parameter_87: name = "parameter_87" + original_name = "linear_332.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-2.16706") @@ -968,6 +1056,7 @@ class Program_weight_tensor_parameter_87: class Program_weight_tensor_parameter_88: name = "parameter_88" + original_name = "linear_331.b_0" shape = [3072] dtype = "float32" min_val = float("-0.966602") @@ -979,6 +1068,7 @@ class Program_weight_tensor_parameter_88: class Program_weight_tensor_parameter_89: name = "parameter_89" + original_name = "linear_331.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.79989") @@ -990,6 +1080,7 @@ class Program_weight_tensor_parameter_89: class Program_weight_tensor_parameter_90: name = "parameter_90" + original_name = "linear_330.b_0" shape = [768] dtype = "float32" min_val = float("-0.214107") @@ -1001,6 +1092,7 @@ class Program_weight_tensor_parameter_90: class Program_weight_tensor_parameter_91: name = "parameter_91" + original_name = "linear_330.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.36874") @@ -1012,6 +1104,7 @@ class Program_weight_tensor_parameter_91: class Program_weight_tensor_parameter_92: name = "parameter_92" + original_name = "linear_329.b_0" shape = [768] dtype = "float32" min_val = float("-0.50265") @@ -1023,6 +1116,7 @@ class Program_weight_tensor_parameter_92: class Program_weight_tensor_parameter_93: name = "parameter_93" + original_name = "linear_329.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.345015") @@ -1034,6 +1128,7 @@ class Program_weight_tensor_parameter_93: class Program_weight_tensor_parameter_94: name = "parameter_94" + original_name = "linear_328.b_0" shape = [768] dtype = "float32" min_val = float("-0.829503") @@ -1045,6 +1140,7 @@ class Program_weight_tensor_parameter_94: class Program_weight_tensor_parameter_95: name = "parameter_95" + original_name = "linear_328.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.703511") @@ -1056,6 +1152,7 @@ class Program_weight_tensor_parameter_95: class Program_weight_tensor_parameter_96: name = "parameter_96" + original_name = "linear_327.b_0" shape = [768] dtype = "float32" min_val = float("-2.41948") @@ -1067,6 +1164,7 @@ class Program_weight_tensor_parameter_96: class Program_weight_tensor_parameter_97: name = "parameter_97" + original_name = "linear_327.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.634525") @@ -1078,6 +1176,7 @@ class Program_weight_tensor_parameter_97: class Program_weight_tensor_parameter_98: name = "parameter_98" + original_name = "layer_norm_111.b_0" shape = [768] dtype = "float32" min_val = float("-0.660182") @@ -1089,6 +1188,7 @@ class Program_weight_tensor_parameter_98: class Program_weight_tensor_parameter_99: name = "parameter_99" + original_name = "layer_norm_111.w_0" shape = [768] dtype = "float32" min_val = float("0.0729106") @@ -1100,6 +1200,7 @@ class Program_weight_tensor_parameter_99: class Program_weight_tensor_parameter_100: name = "parameter_100" + original_name = "layer_norm_110.b_0" shape = [768] dtype = "float32" min_val = float("-6.65866") @@ -1111,6 +1212,7 @@ class Program_weight_tensor_parameter_100: class Program_weight_tensor_parameter_101: name = "parameter_101" + original_name = "layer_norm_110.w_0" shape = [768] dtype = "float32" min_val = float("0.186508") @@ -1122,6 +1224,7 @@ class Program_weight_tensor_parameter_101: class Program_weight_tensor_parameter_102: name = "parameter_102" + original_name = "linear_326.b_0" shape = [768] dtype = "float32" min_val = float("-0.840817") @@ -1133,6 +1236,7 @@ class Program_weight_tensor_parameter_102: class Program_weight_tensor_parameter_103: name = "parameter_103" + original_name = "linear_326.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-1.45656") @@ -1144,6 +1248,7 @@ class Program_weight_tensor_parameter_103: class Program_weight_tensor_parameter_104: name = "parameter_104" + original_name = "linear_325.b_0" shape = [3072] dtype = "float32" min_val = float("-0.869612") @@ -1155,6 +1260,7 @@ class Program_weight_tensor_parameter_104: class Program_weight_tensor_parameter_105: name = "parameter_105" + original_name = "linear_325.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-1.61958") @@ -1166,6 +1272,7 @@ class Program_weight_tensor_parameter_105: class Program_weight_tensor_parameter_106: name = "parameter_106" + original_name = "linear_324.b_0" shape = [768] dtype = "float32" min_val = float("-0.204687") @@ -1177,6 +1284,7 @@ class Program_weight_tensor_parameter_106: class Program_weight_tensor_parameter_107: name = "parameter_107" + original_name = "linear_324.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.367927") @@ -1188,6 +1296,7 @@ class Program_weight_tensor_parameter_107: class Program_weight_tensor_parameter_108: name = "parameter_108" + original_name = "linear_323.b_0" shape = [768] dtype = "float32" min_val = float("-0.622309") @@ -1199,6 +1308,7 @@ class Program_weight_tensor_parameter_108: class Program_weight_tensor_parameter_109: name = "parameter_109" + original_name = "linear_323.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.46596") @@ -1210,6 +1320,7 @@ class Program_weight_tensor_parameter_109: class Program_weight_tensor_parameter_110: name = "parameter_110" + original_name = "linear_322.b_0" shape = [768] dtype = "float32" min_val = float("-0.677549") @@ -1221,6 +1332,7 @@ class Program_weight_tensor_parameter_110: class Program_weight_tensor_parameter_111: name = "parameter_111" + original_name = "linear_322.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.703237") @@ -1232,6 +1344,7 @@ class Program_weight_tensor_parameter_111: class Program_weight_tensor_parameter_112: name = "parameter_112" + original_name = "linear_321.b_0" shape = [768] dtype = "float32" min_val = float("-1.87817") @@ -1243,6 +1356,7 @@ class Program_weight_tensor_parameter_112: class Program_weight_tensor_parameter_113: name = "parameter_113" + original_name = "linear_321.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.579951") @@ -1254,6 +1368,7 @@ class Program_weight_tensor_parameter_113: class Program_weight_tensor_parameter_114: name = "parameter_114" + original_name = "layer_norm_109.b_0" shape = [768] dtype = "float32" min_val = float("-0.652849") @@ -1265,6 +1380,7 @@ class Program_weight_tensor_parameter_114: class Program_weight_tensor_parameter_115: name = "parameter_115" + original_name = "layer_norm_109.w_0" shape = [768] dtype = "float32" min_val = float("0.0623647") @@ -1276,6 +1392,7 @@ class Program_weight_tensor_parameter_115: class Program_weight_tensor_parameter_116: name = "parameter_116" + original_name = "layer_norm_108.b_0" shape = [768] dtype = "float32" min_val = float("-6.09326") @@ -1287,6 +1404,7 @@ class Program_weight_tensor_parameter_116: class Program_weight_tensor_parameter_117: name = "parameter_117" + original_name = "layer_norm_108.w_0" shape = [768] dtype = "float32" min_val = float("0.375049") @@ -1298,6 +1416,7 @@ class Program_weight_tensor_parameter_117: class Program_weight_tensor_parameter_118: name = "parameter_118" + original_name = "linear_320.b_0" shape = [768] dtype = "float32" min_val = float("-0.817283") @@ -1309,6 +1428,7 @@ class Program_weight_tensor_parameter_118: class Program_weight_tensor_parameter_119: name = "parameter_119" + original_name = "linear_320.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-2.52549") @@ -1320,6 +1440,7 @@ class Program_weight_tensor_parameter_119: class Program_weight_tensor_parameter_120: name = "parameter_120" + original_name = "linear_319.b_0" shape = [3072] dtype = "float32" min_val = float("-0.775353") @@ -1331,6 +1452,7 @@ class Program_weight_tensor_parameter_120: class Program_weight_tensor_parameter_121: name = "parameter_121" + original_name = "linear_319.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.700416") @@ -1342,6 +1464,7 @@ class Program_weight_tensor_parameter_121: class Program_weight_tensor_parameter_122: name = "parameter_122" + original_name = "linear_318.b_0" shape = [768] dtype = "float32" min_val = float("-0.312836") @@ -1353,6 +1476,7 @@ class Program_weight_tensor_parameter_122: class Program_weight_tensor_parameter_123: name = "parameter_123" + original_name = "linear_318.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.379891") @@ -1364,6 +1488,7 @@ class Program_weight_tensor_parameter_123: class Program_weight_tensor_parameter_124: name = "parameter_124" + original_name = "linear_317.b_0" shape = [768] dtype = "float32" min_val = float("-0.467662") @@ -1375,6 +1500,7 @@ class Program_weight_tensor_parameter_124: class Program_weight_tensor_parameter_125: name = "parameter_125" + original_name = "linear_317.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.482273") @@ -1386,6 +1512,7 @@ class Program_weight_tensor_parameter_125: class Program_weight_tensor_parameter_126: name = "parameter_126" + original_name = "linear_316.b_0" shape = [768] dtype = "float32" min_val = float("-0.649837") @@ -1397,6 +1524,7 @@ class Program_weight_tensor_parameter_126: class Program_weight_tensor_parameter_127: name = "parameter_127" + original_name = "linear_316.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.721084") @@ -1408,6 +1536,7 @@ class Program_weight_tensor_parameter_127: class Program_weight_tensor_parameter_128: name = "parameter_128" + original_name = "linear_315.b_0" shape = [768] dtype = "float32" min_val = float("-2.56956") @@ -1419,6 +1548,7 @@ class Program_weight_tensor_parameter_128: class Program_weight_tensor_parameter_129: name = "parameter_129" + original_name = "linear_315.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.660104") @@ -1430,6 +1560,7 @@ class Program_weight_tensor_parameter_129: class Program_weight_tensor_parameter_130: name = "parameter_130" + original_name = "layer_norm_107.b_0" shape = [768] dtype = "float32" min_val = float("-0.696733") @@ -1441,6 +1572,7 @@ class Program_weight_tensor_parameter_130: class Program_weight_tensor_parameter_131: name = "parameter_131" + original_name = "layer_norm_107.w_0" shape = [768] dtype = "float32" min_val = float("0.156849") @@ -1452,6 +1584,7 @@ class Program_weight_tensor_parameter_131: class Program_weight_tensor_parameter_132: name = "parameter_132" + original_name = "layer_norm_106.b_0" shape = [768] dtype = "float32" min_val = float("-5.54538") @@ -1463,6 +1596,7 @@ class Program_weight_tensor_parameter_132: class Program_weight_tensor_parameter_133: name = "parameter_133" + original_name = "layer_norm_106.w_0" shape = [768] dtype = "float32" min_val = float("0.527438") @@ -1474,6 +1608,7 @@ class Program_weight_tensor_parameter_133: class Program_weight_tensor_parameter_134: name = "parameter_134" + original_name = "linear_314.b_0" shape = [768] dtype = "float32" min_val = float("-0.777422") @@ -1485,6 +1620,7 @@ class Program_weight_tensor_parameter_134: class Program_weight_tensor_parameter_135: name = "parameter_135" + original_name = "linear_314.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-2.50948") @@ -1496,6 +1632,7 @@ class Program_weight_tensor_parameter_135: class Program_weight_tensor_parameter_136: name = "parameter_136" + original_name = "linear_313.b_0" shape = [3072] dtype = "float32" min_val = float("-0.990975") @@ -1507,6 +1644,7 @@ class Program_weight_tensor_parameter_136: class Program_weight_tensor_parameter_137: name = "parameter_137" + original_name = "linear_313.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.750204") @@ -1518,6 +1656,7 @@ class Program_weight_tensor_parameter_137: class Program_weight_tensor_parameter_138: name = "parameter_138" + original_name = "linear_312.b_0" shape = [768] dtype = "float32" min_val = float("-0.200304") @@ -1529,6 +1668,7 @@ class Program_weight_tensor_parameter_138: class Program_weight_tensor_parameter_139: name = "parameter_139" + original_name = "linear_312.w_0" shape = [768, 768] dtype = "float32" min_val = float("-1.10808") @@ -1540,6 +1680,7 @@ class Program_weight_tensor_parameter_139: class Program_weight_tensor_parameter_140: name = "parameter_140" + original_name = "linear_311.b_0" shape = [768] dtype = "float32" min_val = float("-0.586837") @@ -1551,6 +1692,7 @@ class Program_weight_tensor_parameter_140: class Program_weight_tensor_parameter_141: name = "parameter_141" + original_name = "linear_311.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.341552") @@ -1562,6 +1704,7 @@ class Program_weight_tensor_parameter_141: class Program_weight_tensor_parameter_142: name = "parameter_142" + original_name = "linear_310.b_0" shape = [768] dtype = "float32" min_val = float("-0.564123") @@ -1573,6 +1716,7 @@ class Program_weight_tensor_parameter_142: class Program_weight_tensor_parameter_143: name = "parameter_143" + original_name = "linear_310.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.821062") @@ -1584,6 +1728,7 @@ class Program_weight_tensor_parameter_143: class Program_weight_tensor_parameter_144: name = "parameter_144" + original_name = "linear_309.b_0" shape = [768] dtype = "float32" min_val = float("-1.99221") @@ -1595,6 +1740,7 @@ class Program_weight_tensor_parameter_144: class Program_weight_tensor_parameter_145: name = "parameter_145" + original_name = "linear_309.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.684267") @@ -1606,6 +1752,7 @@ class Program_weight_tensor_parameter_145: class Program_weight_tensor_parameter_146: name = "parameter_146" + original_name = "layer_norm_105.b_0" shape = [768] dtype = "float32" min_val = float("-1.39937") @@ -1617,6 +1764,7 @@ class Program_weight_tensor_parameter_146: class Program_weight_tensor_parameter_147: name = "parameter_147" + original_name = "layer_norm_105.w_0" shape = [768] dtype = "float32" min_val = float("0.0842406") @@ -1628,6 +1776,7 @@ class Program_weight_tensor_parameter_147: class Program_weight_tensor_parameter_148: name = "parameter_148" + original_name = "layer_norm_104.b_0" shape = [768] dtype = "float32" min_val = float("-4.76779") @@ -1639,6 +1788,7 @@ class Program_weight_tensor_parameter_148: class Program_weight_tensor_parameter_149: name = "parameter_149" + original_name = "layer_norm_104.w_0" shape = [768] dtype = "float32" min_val = float("0.664356") @@ -1650,6 +1800,7 @@ class Program_weight_tensor_parameter_149: class Program_weight_tensor_parameter_150: name = "parameter_150" + original_name = "linear_308.b_0" shape = [768] dtype = "float32" min_val = float("-0.521509") @@ -1661,6 +1812,7 @@ class Program_weight_tensor_parameter_150: class Program_weight_tensor_parameter_151: name = "parameter_151" + original_name = "linear_308.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-3.5915") @@ -1672,6 +1824,7 @@ class Program_weight_tensor_parameter_151: class Program_weight_tensor_parameter_152: name = "parameter_152" + original_name = "linear_307.b_0" shape = [3072] dtype = "float32" min_val = float("-0.554721") @@ -1683,6 +1836,7 @@ class Program_weight_tensor_parameter_152: class Program_weight_tensor_parameter_153: name = "parameter_153" + original_name = "linear_307.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.569908") @@ -1694,6 +1848,7 @@ class Program_weight_tensor_parameter_153: class Program_weight_tensor_parameter_154: name = "parameter_154" + original_name = "linear_306.b_0" shape = [768] dtype = "float32" min_val = float("-0.199828") @@ -1705,6 +1860,7 @@ class Program_weight_tensor_parameter_154: class Program_weight_tensor_parameter_155: name = "parameter_155" + original_name = "linear_306.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.390414") @@ -1716,6 +1872,7 @@ class Program_weight_tensor_parameter_155: class Program_weight_tensor_parameter_156: name = "parameter_156" + original_name = "linear_305.b_0" shape = [768] dtype = "float32" min_val = float("-0.343137") @@ -1727,6 +1884,7 @@ class Program_weight_tensor_parameter_156: class Program_weight_tensor_parameter_157: name = "parameter_157" + original_name = "linear_305.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.289256") @@ -1738,6 +1896,7 @@ class Program_weight_tensor_parameter_157: class Program_weight_tensor_parameter_158: name = "parameter_158" + original_name = "linear_304.b_0" shape = [768] dtype = "float32" min_val = float("-0.647247") @@ -1749,6 +1908,7 @@ class Program_weight_tensor_parameter_158: class Program_weight_tensor_parameter_159: name = "parameter_159" + original_name = "linear_304.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.82583") @@ -1760,6 +1920,7 @@ class Program_weight_tensor_parameter_159: class Program_weight_tensor_parameter_160: name = "parameter_160" + original_name = "linear_303.b_0" shape = [768] dtype = "float32" min_val = float("-2.40788") @@ -1771,6 +1932,7 @@ class Program_weight_tensor_parameter_160: class Program_weight_tensor_parameter_161: name = "parameter_161" + original_name = "linear_303.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.505435") @@ -1782,6 +1944,7 @@ class Program_weight_tensor_parameter_161: class Program_weight_tensor_parameter_162: name = "parameter_162" + original_name = "layer_norm_103.b_0" shape = [768] dtype = "float32" min_val = float("-0.486846") @@ -1793,6 +1956,7 @@ class Program_weight_tensor_parameter_162: class Program_weight_tensor_parameter_163: name = "parameter_163" + original_name = "layer_norm_103.w_0" shape = [768] dtype = "float32" min_val = float("0.15501") @@ -1804,6 +1968,7 @@ class Program_weight_tensor_parameter_163: class Program_weight_tensor_parameter_164: name = "parameter_164" + original_name = "layer_norm_102.b_0" shape = [768] dtype = "float32" min_val = float("-4.58695") @@ -1815,6 +1980,7 @@ class Program_weight_tensor_parameter_164: class Program_weight_tensor_parameter_165: name = "parameter_165" + original_name = "layer_norm_102.w_0" shape = [768] dtype = "float32" min_val = float("0.661786") @@ -1826,6 +1992,7 @@ class Program_weight_tensor_parameter_165: class Program_weight_tensor_parameter_166: name = "parameter_166" + original_name = "linear_302.b_0" shape = [768] dtype = "float32" min_val = float("-0.553183") @@ -1837,6 +2004,7 @@ class Program_weight_tensor_parameter_166: class Program_weight_tensor_parameter_167: name = "parameter_167" + original_name = "linear_302.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-6.04611") @@ -1848,6 +2016,7 @@ class Program_weight_tensor_parameter_167: class Program_weight_tensor_parameter_168: name = "parameter_168" + original_name = "linear_301.b_0" shape = [3072] dtype = "float32" min_val = float("-0.532987") @@ -1859,6 +2028,7 @@ class Program_weight_tensor_parameter_168: class Program_weight_tensor_parameter_169: name = "parameter_169" + original_name = "linear_301.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.728391") @@ -1870,6 +2040,7 @@ class Program_weight_tensor_parameter_169: class Program_weight_tensor_parameter_170: name = "parameter_170" + original_name = "linear_300.b_0" shape = [768] dtype = "float32" min_val = float("-0.260247") @@ -1881,6 +2052,7 @@ class Program_weight_tensor_parameter_170: class Program_weight_tensor_parameter_171: name = "parameter_171" + original_name = "linear_300.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.423627") @@ -1892,6 +2064,7 @@ class Program_weight_tensor_parameter_171: class Program_weight_tensor_parameter_172: name = "parameter_172" + original_name = "linear_299.b_0" shape = [768] dtype = "float32" min_val = float("-0.22359") @@ -1903,6 +2076,7 @@ class Program_weight_tensor_parameter_172: class Program_weight_tensor_parameter_173: name = "parameter_173" + original_name = "linear_299.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.409747") @@ -1914,6 +2088,7 @@ class Program_weight_tensor_parameter_173: class Program_weight_tensor_parameter_174: name = "parameter_174" + original_name = "linear_298.b_0" shape = [768] dtype = "float32" min_val = float("-0.495849") @@ -1925,6 +2100,7 @@ class Program_weight_tensor_parameter_174: class Program_weight_tensor_parameter_175: name = "parameter_175" + original_name = "linear_298.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.793595") @@ -1936,6 +2112,7 @@ class Program_weight_tensor_parameter_175: class Program_weight_tensor_parameter_176: name = "parameter_176" + original_name = "linear_297.b_0" shape = [768] dtype = "float32" min_val = float("-2.09018") @@ -1947,6 +2124,7 @@ class Program_weight_tensor_parameter_176: class Program_weight_tensor_parameter_177: name = "parameter_177" + original_name = "linear_297.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.727936") @@ -1958,6 +2136,7 @@ class Program_weight_tensor_parameter_177: class Program_weight_tensor_parameter_178: name = "parameter_178" + original_name = "layer_norm_101.b_0" shape = [768] dtype = "float32" min_val = float("-0.447295") @@ -1969,6 +2148,7 @@ class Program_weight_tensor_parameter_178: class Program_weight_tensor_parameter_179: name = "parameter_179" + original_name = "layer_norm_101.w_0" shape = [768] dtype = "float32" min_val = float("0.125799") @@ -1980,6 +2160,7 @@ class Program_weight_tensor_parameter_179: class Program_weight_tensor_parameter_180: name = "parameter_180" + original_name = "layer_norm_100.b_0" shape = [768] dtype = "float32" min_val = float("-7.89271") @@ -1991,6 +2172,7 @@ class Program_weight_tensor_parameter_180: class Program_weight_tensor_parameter_181: name = "parameter_181" + original_name = "layer_norm_100.w_0" shape = [768] dtype = "float32" min_val = float("0.59") @@ -2002,6 +2184,7 @@ class Program_weight_tensor_parameter_181: class Program_weight_tensor_parameter_182: name = "parameter_182" + original_name = "linear_296.b_0" shape = [768] dtype = "float32" min_val = float("-0.60899") @@ -2013,6 +2196,7 @@ class Program_weight_tensor_parameter_182: class Program_weight_tensor_parameter_183: name = "parameter_183" + original_name = "linear_296.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-6.6023") @@ -2024,6 +2208,7 @@ class Program_weight_tensor_parameter_183: class Program_weight_tensor_parameter_184: name = "parameter_184" + original_name = "linear_295.b_0" shape = [3072] dtype = "float32" min_val = float("-0.871608") @@ -2035,6 +2220,7 @@ class Program_weight_tensor_parameter_184: class Program_weight_tensor_parameter_185: name = "parameter_185" + original_name = "linear_295.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.781975") @@ -2046,6 +2232,7 @@ class Program_weight_tensor_parameter_185: class Program_weight_tensor_parameter_186: name = "parameter_186" + original_name = "linear_294.b_0" shape = [768] dtype = "float32" min_val = float("-0.488488") @@ -2057,6 +2244,7 @@ class Program_weight_tensor_parameter_186: class Program_weight_tensor_parameter_187: name = "parameter_187" + original_name = "linear_294.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.55619") @@ -2068,6 +2256,7 @@ class Program_weight_tensor_parameter_187: class Program_weight_tensor_parameter_188: name = "parameter_188" + original_name = "linear_293.b_0" shape = [768] dtype = "float32" min_val = float("-0.933189") @@ -2079,6 +2268,7 @@ class Program_weight_tensor_parameter_188: class Program_weight_tensor_parameter_189: name = "parameter_189" + original_name = "linear_293.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.267799") @@ -2090,6 +2280,7 @@ class Program_weight_tensor_parameter_189: class Program_weight_tensor_parameter_190: name = "parameter_190" + original_name = "linear_292.b_0" shape = [768] dtype = "float32" min_val = float("-0.346125") @@ -2101,6 +2292,7 @@ class Program_weight_tensor_parameter_190: class Program_weight_tensor_parameter_191: name = "parameter_191" + original_name = "linear_292.w_0" shape = [768, 768] dtype = "float32" min_val = float("-1.24565") @@ -2112,6 +2304,7 @@ class Program_weight_tensor_parameter_191: class Program_weight_tensor_parameter_192: name = "parameter_192" + original_name = "linear_291.b_0" shape = [768] dtype = "float32" min_val = float("-2.29807") @@ -2123,6 +2316,7 @@ class Program_weight_tensor_parameter_192: class Program_weight_tensor_parameter_193: name = "parameter_193" + original_name = "linear_291.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.992284") @@ -2134,6 +2328,7 @@ class Program_weight_tensor_parameter_193: class Program_weight_tensor_parameter_194: name = "parameter_194" + original_name = "layer_norm_99.b_0" shape = [768] dtype = "float32" min_val = float("-0.390783") @@ -2145,6 +2340,7 @@ class Program_weight_tensor_parameter_194: class Program_weight_tensor_parameter_195: name = "parameter_195" + original_name = "layer_norm_99.w_0" shape = [768] dtype = "float32" min_val = float("0.0974017") @@ -2156,6 +2352,7 @@ class Program_weight_tensor_parameter_195: class Program_weight_tensor_parameter_196: name = "parameter_196" + original_name = "embedding_12.w_0" shape = [2, 768] dtype = "float32" min_val = float("-1.32665") @@ -2167,6 +2364,7 @@ class Program_weight_tensor_parameter_196: class Program_weight_tensor_parameter_197: name = "parameter_197" + original_name = "embedding_11.w_0" shape = [513, 768] dtype = "float32" min_val = float("-0.587717") @@ -2178,6 +2376,7 @@ class Program_weight_tensor_parameter_197: class Program_weight_tensor_parameter_198: name = "parameter_198" + original_name = "embedding_10.w_0" shape = [18000, 768] dtype = "float32" min_val = float("-2.02907") diff --git a/paddle_samples/PaddleNLP/ernie-2.0-base-zh/input_meta.py b/paddle_samples/PaddleNLP/ernie-2.0-base-zh/input_meta.py index 97eb8a799..a4123e481 100644 --- a/paddle_samples/PaddleNLP/ernie-2.0-base-zh/input_meta.py +++ b/paddle_samples/PaddleNLP/ernie-2.0-base-zh/input_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_data_0: name = "data_0" + original_name = "input_ids" shape = [1, 11] dtype = "int64" data = [1, 811, 1257, 175, 29, 502, 130, 706, 3619, 12046, 2] @@ -7,6 +8,7 @@ class Program_weight_tensor_data_0: class Program_weight_tensor_data_1: name = "data_1" + original_name = "token_type_ids" shape = [1, 11] dtype = "int64" data = [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0] diff --git a/paddle_samples/PaddleNLP/ernie-2.0-base-zh/model.py b/paddle_samples/PaddleNLP/ernie-2.0-base-zh/model.py index 9151d74c4..9b1061e88 100644 --- a/paddle_samples/PaddleNLP/ernie-2.0-base-zh/model.py +++ b/paddle_samples/PaddleNLP/ernie-2.0-base-zh/model.py @@ -265,10 +265,11 @@ def forward( # pd_op.embedding: (1x11x768xf32) <- (1x11xi64, 513x768xf32) embedding_1 = paddle._C_ops.embedding(subtract_0, parameter_197, -1, False) - del parameter_197 + del parameter_197, subtract_0 # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 1x11x768xf32) add_0 = paddle._C_ops.add(embedding_0, embedding_1) + del embedding_0, embedding_1 # pd_op.embedding: (1x11x768xf32) <- (1x11xi64, 4x768xf32) embedding_2 = paddle._C_ops.embedding(data_1, parameter_196, -1, False) @@ -276,6 +277,7 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 1x11x768xf32) add_1 = paddle._C_ops.add(add_0, embedding_2) + del add_0, embedding_2 # pd_op.layer_norm: (1x11x768xf32, 1x11xf32, 1x11xf32) <- (1x11x768xf32, 768xf32, 768xf32) layer_norm_0, layer_norm_1, layer_norm_2 = (lambda x, f: f(x))( @@ -284,125 +286,17 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_194, parameter_195 + del add_1, parameter_194, parameter_195 # pd_op.full: (1xf32) <- () full_4 = paddle._C_ops.full( [1], float("0.1"), paddle.float32, paddle.core.CPUPlace() ) - # pd_op.assign: (1xf32) <- (1xf32) - assign_0 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_1 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_2 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_3 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_4 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_5 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_6 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_7 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_8 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_9 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_10 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_11 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_12 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_13 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_14 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_15 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_16 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_17 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_18 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_19 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_20 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_21 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_22 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_23 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_24 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_25 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_26 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_27 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_28 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_29 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_30 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_31 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_32 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_33 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_34 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_35 = full_4 - # pd_op.dropout: (1x11x768xf32, 1x11x768xui8) <- (1x11x768xf32, None, 1xf32) dropout_0, dropout_1 = (lambda x, f: f(x))( paddle._C_ops.dropout( - layer_norm_0, None, full_4, False, "upscale_in_train", 0, False + layer_norm_0, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -414,13 +308,14 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_2 = paddle._C_ops.add(matmul_0, parameter_192) - del parameter_192 + del matmul_0, parameter_192 # pd_op.full_int_array: (4xi64) <- () full_int_array_1 = [0, 0, 12, 64] # pd_op.reshape: (1x11x12x64xf32) <- (1x11x768xf32, 4xi64) reshape_0 = paddle._C_ops.reshape(add_2, full_int_array_1) + del add_2 # pd_op.transpose: (1x12x11x64xf32) <- (1x11x12x64xf32) transpose_0 = paddle._C_ops.transpose(reshape_0, [0, 2, 1, 3]) @@ -432,7 +327,7 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_3 = paddle._C_ops.add(matmul_1, parameter_190) - del parameter_190 + del matmul_1, parameter_190 # pd_op.matmul: (1x11x768xf32) <- (1x11x768xf32, 768x768xf32) matmul_2 = paddle._C_ops.matmul(dropout_0, parameter_189, False, False) @@ -440,10 +335,11 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_4 = paddle._C_ops.add(matmul_2, parameter_188) - del parameter_188 + del matmul_2, parameter_188 # pd_op.reshape: (1x11x12x64xf32) <- (1x11x768xf32, 4xi64) reshape_1 = paddle._C_ops.reshape(add_3, full_int_array_1) + del add_3 # pd_op.transpose: (1x12x11x64xf32) <- (1x11x12x64xf32) transpose_1 = paddle._C_ops.transpose(reshape_1, [0, 2, 1, 3]) @@ -451,6 +347,7 @@ def forward( # pd_op.reshape: (1x11x12x64xf32) <- (1x11x768xf32, 4xi64) reshape_2 = paddle._C_ops.reshape(add_4, full_int_array_1) + del add_4 # pd_op.transpose: (1x12x11x64xf32) <- (1x11x12x64xf32) transpose_2 = paddle._C_ops.transpose(reshape_2, [0, 2, 1, 3]) @@ -461,48 +358,17 @@ def forward( [1], float("0.125"), paddle.float32, paddle.core.CPUPlace() ) - # pd_op.assign: (1xf32) <- (1xf32) - assign_36 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_37 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_38 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_39 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_40 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_41 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_42 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_43 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_44 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_45 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_46 = full_5 - # pd_op.scale: (1x12x11x64xf32) <- (1x12x11x64xf32, 1xf32) scale_1 = paddle._C_ops.scale(transpose_0, full_5, float("0"), True) del transpose_0 # pd_op.matmul: (1x12x11x11xf32) <- (1x12x11x64xf32, 1x12x11x64xf32) matmul_3 = paddle._C_ops.matmul(scale_1, transpose_1, False, True) + del scale_1, transpose_1 # pd_op.add: (1x12x11x11xf32) <- (1x12x11x11xf32, 1x1x1x11xf32) add_5 = paddle._C_ops.add(matmul_3, unsqueeze_0) + del matmul_3 # pd_op.softmax: (1x12x11x11xf32) <- (1x12x11x11xf32) softmax_0 = paddle._C_ops.softmax(add_5, -1) @@ -511,13 +377,15 @@ def forward( # pd_op.dropout: (1x12x11x11xf32, 1x12x11x11xui8) <- (1x12x11x11xf32, None, 1xf32) dropout_2, dropout_3 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_0, None, full_4, False, "upscale_in_train", 0, False + softmax_0, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_0 # pd_op.matmul: (1x12x11x64xf32) <- (1x12x11x11xf32, 1x12x11x64xf32) matmul_4 = paddle._C_ops.matmul(dropout_2, transpose_2, False, False) + del dropout_2, transpose_2 # pd_op.transpose: (1x11x12x64xf32) <- (1x12x11x64xf32) transpose_3 = paddle._C_ops.transpose(matmul_4, [0, 2, 1, 3]) @@ -528,19 +396,20 @@ def forward( # pd_op.reshape: (1x11x768xf32) <- (1x11x12x64xf32, 3xi64) reshape_3 = paddle._C_ops.reshape(transpose_3, full_int_array_2) + del transpose_3 # pd_op.matmul: (1x11x768xf32) <- (1x11x768xf32, 768x768xf32) matmul_5 = paddle._C_ops.matmul(reshape_3, parameter_187, False, False) - del parameter_187 + del parameter_187, reshape_3 # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_6 = paddle._C_ops.add(matmul_5, parameter_186) - del parameter_186 + del matmul_5, parameter_186 # pd_op.dropout: (1x11x768xf32, 1x11x768xui8) <- (1x11x768xf32, None, 1xf32) dropout_4, dropout_5 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_6, None, full_4, False, "upscale_in_train", 0, False + add_6, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -548,6 +417,7 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 1x11x768xf32) add_7 = paddle._C_ops.add(dropout_0, dropout_4) + del dropout_0, dropout_4 # pd_op.layer_norm: (1x11x768xf32, 1x11xf32, 1x11xf32) <- (1x11x768xf32, 768xf32, 768xf32) layer_norm_3, layer_norm_4, layer_norm_5 = (lambda x, f: f(x))( @@ -556,7 +426,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_180, parameter_181 + del add_7, parameter_180, parameter_181 # pd_op.matmul: (1x11x3072xf32) <- (1x11x768xf32, 768x3072xf32) matmul_6 = paddle._C_ops.matmul(layer_norm_3, parameter_185, False, False) @@ -564,7 +434,7 @@ def forward( # pd_op.add: (1x11x3072xf32) <- (1x11x3072xf32, 3072xf32) add_8 = paddle._C_ops.add(matmul_6, parameter_184) - del parameter_184 + del matmul_6, parameter_184 # pd_op.relu: (1x11x3072xf32) <- (1x11x3072xf32) relu_0 = paddle._C_ops.relu(add_8) @@ -572,16 +442,16 @@ def forward( # pd_op.matmul: (1x11x768xf32) <- (1x11x3072xf32, 3072x768xf32) matmul_7 = paddle._C_ops.matmul(relu_0, parameter_183, False, False) - del parameter_183 + del parameter_183, relu_0 # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_9 = paddle._C_ops.add(matmul_7, parameter_182) - del parameter_182 + del matmul_7, parameter_182 # pd_op.dropout: (1x11x768xf32, 1x11x768xui8) <- (1x11x768xf32, None, 1xf32) dropout_6, dropout_7 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_9, None, full_4, False, "upscale_in_train", 0, False + add_9, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -589,6 +459,7 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 1x11x768xf32) add_10 = paddle._C_ops.add(layer_norm_3, dropout_6) + del dropout_6, layer_norm_3 # pd_op.layer_norm: (1x11x768xf32, 1x11xf32, 1x11xf32) <- (1x11x768xf32, 768xf32, 768xf32) layer_norm_6, layer_norm_7, layer_norm_8 = (lambda x, f: f(x))( @@ -597,7 +468,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_178, parameter_179 + del add_10, parameter_178, parameter_179 # pd_op.matmul: (1x11x768xf32) <- (1x11x768xf32, 768x768xf32) matmul_8 = paddle._C_ops.matmul(layer_norm_6, parameter_177, False, False) @@ -605,10 +476,11 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_11 = paddle._C_ops.add(matmul_8, parameter_176) - del parameter_176 + del matmul_8, parameter_176 # pd_op.reshape: (1x11x12x64xf32) <- (1x11x768xf32, 4xi64) reshape_4 = paddle._C_ops.reshape(add_11, full_int_array_1) + del add_11 # pd_op.transpose: (1x12x11x64xf32) <- (1x11x12x64xf32) transpose_4 = paddle._C_ops.transpose(reshape_4, [0, 2, 1, 3]) @@ -620,7 +492,7 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_12 = paddle._C_ops.add(matmul_9, parameter_174) - del parameter_174 + del matmul_9, parameter_174 # pd_op.matmul: (1x11x768xf32) <- (1x11x768xf32, 768x768xf32) matmul_10 = paddle._C_ops.matmul(layer_norm_6, parameter_173, False, False) @@ -628,10 +500,11 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_13 = paddle._C_ops.add(matmul_10, parameter_172) - del parameter_172 + del matmul_10, parameter_172 # pd_op.reshape: (1x11x12x64xf32) <- (1x11x768xf32, 4xi64) reshape_5 = paddle._C_ops.reshape(add_12, full_int_array_1) + del add_12 # pd_op.transpose: (1x12x11x64xf32) <- (1x11x12x64xf32) transpose_5 = paddle._C_ops.transpose(reshape_5, [0, 2, 1, 3]) @@ -639,6 +512,7 @@ def forward( # pd_op.reshape: (1x11x12x64xf32) <- (1x11x768xf32, 4xi64) reshape_6 = paddle._C_ops.reshape(add_13, full_int_array_1) + del add_13 # pd_op.transpose: (1x12x11x64xf32) <- (1x11x12x64xf32) transpose_6 = paddle._C_ops.transpose(reshape_6, [0, 2, 1, 3]) @@ -650,9 +524,11 @@ def forward( # pd_op.matmul: (1x12x11x11xf32) <- (1x12x11x64xf32, 1x12x11x64xf32) matmul_11 = paddle._C_ops.matmul(scale_2, transpose_5, False, True) + del scale_2, transpose_5 # pd_op.add: (1x12x11x11xf32) <- (1x12x11x11xf32, 1x1x1x11xf32) add_14 = paddle._C_ops.add(matmul_11, unsqueeze_0) + del matmul_11 # pd_op.softmax: (1x12x11x11xf32) <- (1x12x11x11xf32) softmax_1 = paddle._C_ops.softmax(add_14, -1) @@ -661,13 +537,15 @@ def forward( # pd_op.dropout: (1x12x11x11xf32, 1x12x11x11xui8) <- (1x12x11x11xf32, None, 1xf32) dropout_8, dropout_9 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_1, None, full_4, False, "upscale_in_train", 0, False + softmax_1, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_1 # pd_op.matmul: (1x12x11x64xf32) <- (1x12x11x11xf32, 1x12x11x64xf32) matmul_12 = paddle._C_ops.matmul(dropout_8, transpose_6, False, False) + del dropout_8, transpose_6 # pd_op.transpose: (1x11x12x64xf32) <- (1x12x11x64xf32) transpose_7 = paddle._C_ops.transpose(matmul_12, [0, 2, 1, 3]) @@ -675,19 +553,20 @@ def forward( # pd_op.reshape: (1x11x768xf32) <- (1x11x12x64xf32, 3xi64) reshape_7 = paddle._C_ops.reshape(transpose_7, full_int_array_2) + del transpose_7 # pd_op.matmul: (1x11x768xf32) <- (1x11x768xf32, 768x768xf32) matmul_13 = paddle._C_ops.matmul(reshape_7, parameter_171, False, False) - del parameter_171 + del parameter_171, reshape_7 # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_15 = paddle._C_ops.add(matmul_13, parameter_170) - del parameter_170 + del matmul_13, parameter_170 # pd_op.dropout: (1x11x768xf32, 1x11x768xui8) <- (1x11x768xf32, None, 1xf32) dropout_10, dropout_11 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_15, None, full_4, False, "upscale_in_train", 0, False + add_15, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -695,6 +574,7 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 1x11x768xf32) add_16 = paddle._C_ops.add(layer_norm_6, dropout_10) + del dropout_10, layer_norm_6 # pd_op.layer_norm: (1x11x768xf32, 1x11xf32, 1x11xf32) <- (1x11x768xf32, 768xf32, 768xf32) layer_norm_9, layer_norm_10, layer_norm_11 = (lambda x, f: f(x))( @@ -703,7 +583,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_164, parameter_165 + del add_16, parameter_164, parameter_165 # pd_op.matmul: (1x11x3072xf32) <- (1x11x768xf32, 768x3072xf32) matmul_14 = paddle._C_ops.matmul(layer_norm_9, parameter_169, False, False) @@ -711,7 +591,7 @@ def forward( # pd_op.add: (1x11x3072xf32) <- (1x11x3072xf32, 3072xf32) add_17 = paddle._C_ops.add(matmul_14, parameter_168) - del parameter_168 + del matmul_14, parameter_168 # pd_op.relu: (1x11x3072xf32) <- (1x11x3072xf32) relu_1 = paddle._C_ops.relu(add_17) @@ -719,16 +599,16 @@ def forward( # pd_op.matmul: (1x11x768xf32) <- (1x11x3072xf32, 3072x768xf32) matmul_15 = paddle._C_ops.matmul(relu_1, parameter_167, False, False) - del parameter_167 + del parameter_167, relu_1 # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_18 = paddle._C_ops.add(matmul_15, parameter_166) - del parameter_166 + del matmul_15, parameter_166 # pd_op.dropout: (1x11x768xf32, 1x11x768xui8) <- (1x11x768xf32, None, 1xf32) dropout_12, dropout_13 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_18, None, full_4, False, "upscale_in_train", 0, False + add_18, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -736,6 +616,7 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 1x11x768xf32) add_19 = paddle._C_ops.add(layer_norm_9, dropout_12) + del dropout_12, layer_norm_9 # pd_op.layer_norm: (1x11x768xf32, 1x11xf32, 1x11xf32) <- (1x11x768xf32, 768xf32, 768xf32) layer_norm_12, layer_norm_13, layer_norm_14 = (lambda x, f: f(x))( @@ -744,7 +625,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_162, parameter_163 + del add_19, parameter_162, parameter_163 # pd_op.matmul: (1x11x768xf32) <- (1x11x768xf32, 768x768xf32) matmul_16 = paddle._C_ops.matmul(layer_norm_12, parameter_161, False, False) @@ -752,10 +633,11 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_20 = paddle._C_ops.add(matmul_16, parameter_160) - del parameter_160 + del matmul_16, parameter_160 # pd_op.reshape: (1x11x12x64xf32) <- (1x11x768xf32, 4xi64) reshape_8 = paddle._C_ops.reshape(add_20, full_int_array_1) + del add_20 # pd_op.transpose: (1x12x11x64xf32) <- (1x11x12x64xf32) transpose_8 = paddle._C_ops.transpose(reshape_8, [0, 2, 1, 3]) @@ -767,7 +649,7 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_21 = paddle._C_ops.add(matmul_17, parameter_158) - del parameter_158 + del matmul_17, parameter_158 # pd_op.matmul: (1x11x768xf32) <- (1x11x768xf32, 768x768xf32) matmul_18 = paddle._C_ops.matmul(layer_norm_12, parameter_157, False, False) @@ -775,10 +657,11 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_22 = paddle._C_ops.add(matmul_18, parameter_156) - del parameter_156 + del matmul_18, parameter_156 # pd_op.reshape: (1x11x12x64xf32) <- (1x11x768xf32, 4xi64) reshape_9 = paddle._C_ops.reshape(add_21, full_int_array_1) + del add_21 # pd_op.transpose: (1x12x11x64xf32) <- (1x11x12x64xf32) transpose_9 = paddle._C_ops.transpose(reshape_9, [0, 2, 1, 3]) @@ -786,6 +669,7 @@ def forward( # pd_op.reshape: (1x11x12x64xf32) <- (1x11x768xf32, 4xi64) reshape_10 = paddle._C_ops.reshape(add_22, full_int_array_1) + del add_22 # pd_op.transpose: (1x12x11x64xf32) <- (1x11x12x64xf32) transpose_10 = paddle._C_ops.transpose(reshape_10, [0, 2, 1, 3]) @@ -797,9 +681,11 @@ def forward( # pd_op.matmul: (1x12x11x11xf32) <- (1x12x11x64xf32, 1x12x11x64xf32) matmul_19 = paddle._C_ops.matmul(scale_3, transpose_9, False, True) + del scale_3, transpose_9 # pd_op.add: (1x12x11x11xf32) <- (1x12x11x11xf32, 1x1x1x11xf32) add_23 = paddle._C_ops.add(matmul_19, unsqueeze_0) + del matmul_19 # pd_op.softmax: (1x12x11x11xf32) <- (1x12x11x11xf32) softmax_2 = paddle._C_ops.softmax(add_23, -1) @@ -808,13 +694,15 @@ def forward( # pd_op.dropout: (1x12x11x11xf32, 1x12x11x11xui8) <- (1x12x11x11xf32, None, 1xf32) dropout_14, dropout_15 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_2, None, full_4, False, "upscale_in_train", 0, False + softmax_2, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_2 # pd_op.matmul: (1x12x11x64xf32) <- (1x12x11x11xf32, 1x12x11x64xf32) matmul_20 = paddle._C_ops.matmul(dropout_14, transpose_10, False, False) + del dropout_14, transpose_10 # pd_op.transpose: (1x11x12x64xf32) <- (1x12x11x64xf32) transpose_11 = paddle._C_ops.transpose(matmul_20, [0, 2, 1, 3]) @@ -822,19 +710,20 @@ def forward( # pd_op.reshape: (1x11x768xf32) <- (1x11x12x64xf32, 3xi64) reshape_11 = paddle._C_ops.reshape(transpose_11, full_int_array_2) + del transpose_11 # pd_op.matmul: (1x11x768xf32) <- (1x11x768xf32, 768x768xf32) matmul_21 = paddle._C_ops.matmul(reshape_11, parameter_155, False, False) - del parameter_155 + del parameter_155, reshape_11 # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_24 = paddle._C_ops.add(matmul_21, parameter_154) - del parameter_154 + del matmul_21, parameter_154 # pd_op.dropout: (1x11x768xf32, 1x11x768xui8) <- (1x11x768xf32, None, 1xf32) dropout_16, dropout_17 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_24, None, full_4, False, "upscale_in_train", 0, False + add_24, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -842,6 +731,7 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 1x11x768xf32) add_25 = paddle._C_ops.add(layer_norm_12, dropout_16) + del dropout_16, layer_norm_12 # pd_op.layer_norm: (1x11x768xf32, 1x11xf32, 1x11xf32) <- (1x11x768xf32, 768xf32, 768xf32) layer_norm_15, layer_norm_16, layer_norm_17 = (lambda x, f: f(x))( @@ -850,7 +740,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_148, parameter_149 + del add_25, parameter_148, parameter_149 # pd_op.matmul: (1x11x3072xf32) <- (1x11x768xf32, 768x3072xf32) matmul_22 = paddle._C_ops.matmul(layer_norm_15, parameter_153, False, False) @@ -858,7 +748,7 @@ def forward( # pd_op.add: (1x11x3072xf32) <- (1x11x3072xf32, 3072xf32) add_26 = paddle._C_ops.add(matmul_22, parameter_152) - del parameter_152 + del matmul_22, parameter_152 # pd_op.relu: (1x11x3072xf32) <- (1x11x3072xf32) relu_2 = paddle._C_ops.relu(add_26) @@ -866,16 +756,16 @@ def forward( # pd_op.matmul: (1x11x768xf32) <- (1x11x3072xf32, 3072x768xf32) matmul_23 = paddle._C_ops.matmul(relu_2, parameter_151, False, False) - del parameter_151 + del parameter_151, relu_2 # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_27 = paddle._C_ops.add(matmul_23, parameter_150) - del parameter_150 + del matmul_23, parameter_150 # pd_op.dropout: (1x11x768xf32, 1x11x768xui8) <- (1x11x768xf32, None, 1xf32) dropout_18, dropout_19 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_27, None, full_4, False, "upscale_in_train", 0, False + add_27, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -883,6 +773,7 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 1x11x768xf32) add_28 = paddle._C_ops.add(layer_norm_15, dropout_18) + del dropout_18, layer_norm_15 # pd_op.layer_norm: (1x11x768xf32, 1x11xf32, 1x11xf32) <- (1x11x768xf32, 768xf32, 768xf32) layer_norm_18, layer_norm_19, layer_norm_20 = (lambda x, f: f(x))( @@ -891,7 +782,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_146, parameter_147 + del add_28, parameter_146, parameter_147 # pd_op.matmul: (1x11x768xf32) <- (1x11x768xf32, 768x768xf32) matmul_24 = paddle._C_ops.matmul(layer_norm_18, parameter_145, False, False) @@ -899,10 +790,11 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_29 = paddle._C_ops.add(matmul_24, parameter_144) - del parameter_144 + del matmul_24, parameter_144 # pd_op.reshape: (1x11x12x64xf32) <- (1x11x768xf32, 4xi64) reshape_12 = paddle._C_ops.reshape(add_29, full_int_array_1) + del add_29 # pd_op.transpose: (1x12x11x64xf32) <- (1x11x12x64xf32) transpose_12 = paddle._C_ops.transpose(reshape_12, [0, 2, 1, 3]) @@ -914,7 +806,7 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_30 = paddle._C_ops.add(matmul_25, parameter_142) - del parameter_142 + del matmul_25, parameter_142 # pd_op.matmul: (1x11x768xf32) <- (1x11x768xf32, 768x768xf32) matmul_26 = paddle._C_ops.matmul(layer_norm_18, parameter_141, False, False) @@ -922,10 +814,11 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_31 = paddle._C_ops.add(matmul_26, parameter_140) - del parameter_140 + del matmul_26, parameter_140 # pd_op.reshape: (1x11x12x64xf32) <- (1x11x768xf32, 4xi64) reshape_13 = paddle._C_ops.reshape(add_30, full_int_array_1) + del add_30 # pd_op.transpose: (1x12x11x64xf32) <- (1x11x12x64xf32) transpose_13 = paddle._C_ops.transpose(reshape_13, [0, 2, 1, 3]) @@ -933,6 +826,7 @@ def forward( # pd_op.reshape: (1x11x12x64xf32) <- (1x11x768xf32, 4xi64) reshape_14 = paddle._C_ops.reshape(add_31, full_int_array_1) + del add_31 # pd_op.transpose: (1x12x11x64xf32) <- (1x11x12x64xf32) transpose_14 = paddle._C_ops.transpose(reshape_14, [0, 2, 1, 3]) @@ -944,9 +838,11 @@ def forward( # pd_op.matmul: (1x12x11x11xf32) <- (1x12x11x64xf32, 1x12x11x64xf32) matmul_27 = paddle._C_ops.matmul(scale_4, transpose_13, False, True) + del scale_4, transpose_13 # pd_op.add: (1x12x11x11xf32) <- (1x12x11x11xf32, 1x1x1x11xf32) add_32 = paddle._C_ops.add(matmul_27, unsqueeze_0) + del matmul_27 # pd_op.softmax: (1x12x11x11xf32) <- (1x12x11x11xf32) softmax_3 = paddle._C_ops.softmax(add_32, -1) @@ -955,13 +851,15 @@ def forward( # pd_op.dropout: (1x12x11x11xf32, 1x12x11x11xui8) <- (1x12x11x11xf32, None, 1xf32) dropout_20, dropout_21 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_3, None, full_4, False, "upscale_in_train", 0, False + softmax_3, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_3 # pd_op.matmul: (1x12x11x64xf32) <- (1x12x11x11xf32, 1x12x11x64xf32) matmul_28 = paddle._C_ops.matmul(dropout_20, transpose_14, False, False) + del dropout_20, transpose_14 # pd_op.transpose: (1x11x12x64xf32) <- (1x12x11x64xf32) transpose_15 = paddle._C_ops.transpose(matmul_28, [0, 2, 1, 3]) @@ -969,19 +867,20 @@ def forward( # pd_op.reshape: (1x11x768xf32) <- (1x11x12x64xf32, 3xi64) reshape_15 = paddle._C_ops.reshape(transpose_15, full_int_array_2) + del transpose_15 # pd_op.matmul: (1x11x768xf32) <- (1x11x768xf32, 768x768xf32) matmul_29 = paddle._C_ops.matmul(reshape_15, parameter_139, False, False) - del parameter_139 + del parameter_139, reshape_15 # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_33 = paddle._C_ops.add(matmul_29, parameter_138) - del parameter_138 + del matmul_29, parameter_138 # pd_op.dropout: (1x11x768xf32, 1x11x768xui8) <- (1x11x768xf32, None, 1xf32) dropout_22, dropout_23 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_33, None, full_4, False, "upscale_in_train", 0, False + add_33, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -989,6 +888,7 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 1x11x768xf32) add_34 = paddle._C_ops.add(layer_norm_18, dropout_22) + del dropout_22, layer_norm_18 # pd_op.layer_norm: (1x11x768xf32, 1x11xf32, 1x11xf32) <- (1x11x768xf32, 768xf32, 768xf32) layer_norm_21, layer_norm_22, layer_norm_23 = (lambda x, f: f(x))( @@ -997,7 +897,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_132, parameter_133 + del add_34, parameter_132, parameter_133 # pd_op.matmul: (1x11x3072xf32) <- (1x11x768xf32, 768x3072xf32) matmul_30 = paddle._C_ops.matmul(layer_norm_21, parameter_137, False, False) @@ -1005,7 +905,7 @@ def forward( # pd_op.add: (1x11x3072xf32) <- (1x11x3072xf32, 3072xf32) add_35 = paddle._C_ops.add(matmul_30, parameter_136) - del parameter_136 + del matmul_30, parameter_136 # pd_op.relu: (1x11x3072xf32) <- (1x11x3072xf32) relu_3 = paddle._C_ops.relu(add_35) @@ -1013,16 +913,16 @@ def forward( # pd_op.matmul: (1x11x768xf32) <- (1x11x3072xf32, 3072x768xf32) matmul_31 = paddle._C_ops.matmul(relu_3, parameter_135, False, False) - del parameter_135 + del parameter_135, relu_3 # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_36 = paddle._C_ops.add(matmul_31, parameter_134) - del parameter_134 + del matmul_31, parameter_134 # pd_op.dropout: (1x11x768xf32, 1x11x768xui8) <- (1x11x768xf32, None, 1xf32) dropout_24, dropout_25 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_36, None, full_4, False, "upscale_in_train", 0, False + add_36, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1030,6 +930,7 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 1x11x768xf32) add_37 = paddle._C_ops.add(layer_norm_21, dropout_24) + del dropout_24, layer_norm_21 # pd_op.layer_norm: (1x11x768xf32, 1x11xf32, 1x11xf32) <- (1x11x768xf32, 768xf32, 768xf32) layer_norm_24, layer_norm_25, layer_norm_26 = (lambda x, f: f(x))( @@ -1038,7 +939,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_130, parameter_131 + del add_37, parameter_130, parameter_131 # pd_op.matmul: (1x11x768xf32) <- (1x11x768xf32, 768x768xf32) matmul_32 = paddle._C_ops.matmul(layer_norm_24, parameter_129, False, False) @@ -1046,10 +947,11 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_38 = paddle._C_ops.add(matmul_32, parameter_128) - del parameter_128 + del matmul_32, parameter_128 # pd_op.reshape: (1x11x12x64xf32) <- (1x11x768xf32, 4xi64) reshape_16 = paddle._C_ops.reshape(add_38, full_int_array_1) + del add_38 # pd_op.transpose: (1x12x11x64xf32) <- (1x11x12x64xf32) transpose_16 = paddle._C_ops.transpose(reshape_16, [0, 2, 1, 3]) @@ -1061,7 +963,7 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_39 = paddle._C_ops.add(matmul_33, parameter_126) - del parameter_126 + del matmul_33, parameter_126 # pd_op.matmul: (1x11x768xf32) <- (1x11x768xf32, 768x768xf32) matmul_34 = paddle._C_ops.matmul(layer_norm_24, parameter_125, False, False) @@ -1069,10 +971,11 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_40 = paddle._C_ops.add(matmul_34, parameter_124) - del parameter_124 + del matmul_34, parameter_124 # pd_op.reshape: (1x11x12x64xf32) <- (1x11x768xf32, 4xi64) reshape_17 = paddle._C_ops.reshape(add_39, full_int_array_1) + del add_39 # pd_op.transpose: (1x12x11x64xf32) <- (1x11x12x64xf32) transpose_17 = paddle._C_ops.transpose(reshape_17, [0, 2, 1, 3]) @@ -1080,6 +983,7 @@ def forward( # pd_op.reshape: (1x11x12x64xf32) <- (1x11x768xf32, 4xi64) reshape_18 = paddle._C_ops.reshape(add_40, full_int_array_1) + del add_40 # pd_op.transpose: (1x12x11x64xf32) <- (1x11x12x64xf32) transpose_18 = paddle._C_ops.transpose(reshape_18, [0, 2, 1, 3]) @@ -1091,9 +995,11 @@ def forward( # pd_op.matmul: (1x12x11x11xf32) <- (1x12x11x64xf32, 1x12x11x64xf32) matmul_35 = paddle._C_ops.matmul(scale_5, transpose_17, False, True) + del scale_5, transpose_17 # pd_op.add: (1x12x11x11xf32) <- (1x12x11x11xf32, 1x1x1x11xf32) add_41 = paddle._C_ops.add(matmul_35, unsqueeze_0) + del matmul_35 # pd_op.softmax: (1x12x11x11xf32) <- (1x12x11x11xf32) softmax_4 = paddle._C_ops.softmax(add_41, -1) @@ -1102,13 +1008,15 @@ def forward( # pd_op.dropout: (1x12x11x11xf32, 1x12x11x11xui8) <- (1x12x11x11xf32, None, 1xf32) dropout_26, dropout_27 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_4, None, full_4, False, "upscale_in_train", 0, False + softmax_4, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_4 # pd_op.matmul: (1x12x11x64xf32) <- (1x12x11x11xf32, 1x12x11x64xf32) matmul_36 = paddle._C_ops.matmul(dropout_26, transpose_18, False, False) + del dropout_26, transpose_18 # pd_op.transpose: (1x11x12x64xf32) <- (1x12x11x64xf32) transpose_19 = paddle._C_ops.transpose(matmul_36, [0, 2, 1, 3]) @@ -1116,19 +1024,20 @@ def forward( # pd_op.reshape: (1x11x768xf32) <- (1x11x12x64xf32, 3xi64) reshape_19 = paddle._C_ops.reshape(transpose_19, full_int_array_2) + del transpose_19 # pd_op.matmul: (1x11x768xf32) <- (1x11x768xf32, 768x768xf32) matmul_37 = paddle._C_ops.matmul(reshape_19, parameter_123, False, False) - del parameter_123 + del parameter_123, reshape_19 # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_42 = paddle._C_ops.add(matmul_37, parameter_122) - del parameter_122 + del matmul_37, parameter_122 # pd_op.dropout: (1x11x768xf32, 1x11x768xui8) <- (1x11x768xf32, None, 1xf32) dropout_28, dropout_29 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_42, None, full_4, False, "upscale_in_train", 0, False + add_42, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1136,6 +1045,7 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 1x11x768xf32) add_43 = paddle._C_ops.add(layer_norm_24, dropout_28) + del dropout_28, layer_norm_24 # pd_op.layer_norm: (1x11x768xf32, 1x11xf32, 1x11xf32) <- (1x11x768xf32, 768xf32, 768xf32) layer_norm_27, layer_norm_28, layer_norm_29 = (lambda x, f: f(x))( @@ -1144,7 +1054,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_116, parameter_117 + del add_43, parameter_116, parameter_117 # pd_op.matmul: (1x11x3072xf32) <- (1x11x768xf32, 768x3072xf32) matmul_38 = paddle._C_ops.matmul(layer_norm_27, parameter_121, False, False) @@ -1152,7 +1062,7 @@ def forward( # pd_op.add: (1x11x3072xf32) <- (1x11x3072xf32, 3072xf32) add_44 = paddle._C_ops.add(matmul_38, parameter_120) - del parameter_120 + del matmul_38, parameter_120 # pd_op.relu: (1x11x3072xf32) <- (1x11x3072xf32) relu_4 = paddle._C_ops.relu(add_44) @@ -1160,16 +1070,16 @@ def forward( # pd_op.matmul: (1x11x768xf32) <- (1x11x3072xf32, 3072x768xf32) matmul_39 = paddle._C_ops.matmul(relu_4, parameter_119, False, False) - del parameter_119 + del parameter_119, relu_4 # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_45 = paddle._C_ops.add(matmul_39, parameter_118) - del parameter_118 + del matmul_39, parameter_118 # pd_op.dropout: (1x11x768xf32, 1x11x768xui8) <- (1x11x768xf32, None, 1xf32) dropout_30, dropout_31 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_45, None, full_4, False, "upscale_in_train", 0, False + add_45, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1177,6 +1087,7 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 1x11x768xf32) add_46 = paddle._C_ops.add(layer_norm_27, dropout_30) + del dropout_30, layer_norm_27 # pd_op.layer_norm: (1x11x768xf32, 1x11xf32, 1x11xf32) <- (1x11x768xf32, 768xf32, 768xf32) layer_norm_30, layer_norm_31, layer_norm_32 = (lambda x, f: f(x))( @@ -1185,7 +1096,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_114, parameter_115 + del add_46, parameter_114, parameter_115 # pd_op.matmul: (1x11x768xf32) <- (1x11x768xf32, 768x768xf32) matmul_40 = paddle._C_ops.matmul(layer_norm_30, parameter_113, False, False) @@ -1193,10 +1104,11 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_47 = paddle._C_ops.add(matmul_40, parameter_112) - del parameter_112 + del matmul_40, parameter_112 # pd_op.reshape: (1x11x12x64xf32) <- (1x11x768xf32, 4xi64) reshape_20 = paddle._C_ops.reshape(add_47, full_int_array_1) + del add_47 # pd_op.transpose: (1x12x11x64xf32) <- (1x11x12x64xf32) transpose_20 = paddle._C_ops.transpose(reshape_20, [0, 2, 1, 3]) @@ -1208,7 +1120,7 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_48 = paddle._C_ops.add(matmul_41, parameter_110) - del parameter_110 + del matmul_41, parameter_110 # pd_op.matmul: (1x11x768xf32) <- (1x11x768xf32, 768x768xf32) matmul_42 = paddle._C_ops.matmul(layer_norm_30, parameter_109, False, False) @@ -1216,10 +1128,11 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_49 = paddle._C_ops.add(matmul_42, parameter_108) - del parameter_108 + del matmul_42, parameter_108 # pd_op.reshape: (1x11x12x64xf32) <- (1x11x768xf32, 4xi64) reshape_21 = paddle._C_ops.reshape(add_48, full_int_array_1) + del add_48 # pd_op.transpose: (1x12x11x64xf32) <- (1x11x12x64xf32) transpose_21 = paddle._C_ops.transpose(reshape_21, [0, 2, 1, 3]) @@ -1227,6 +1140,7 @@ def forward( # pd_op.reshape: (1x11x12x64xf32) <- (1x11x768xf32, 4xi64) reshape_22 = paddle._C_ops.reshape(add_49, full_int_array_1) + del add_49 # pd_op.transpose: (1x12x11x64xf32) <- (1x11x12x64xf32) transpose_22 = paddle._C_ops.transpose(reshape_22, [0, 2, 1, 3]) @@ -1238,9 +1152,11 @@ def forward( # pd_op.matmul: (1x12x11x11xf32) <- (1x12x11x64xf32, 1x12x11x64xf32) matmul_43 = paddle._C_ops.matmul(scale_6, transpose_21, False, True) + del scale_6, transpose_21 # pd_op.add: (1x12x11x11xf32) <- (1x12x11x11xf32, 1x1x1x11xf32) add_50 = paddle._C_ops.add(matmul_43, unsqueeze_0) + del matmul_43 # pd_op.softmax: (1x12x11x11xf32) <- (1x12x11x11xf32) softmax_5 = paddle._C_ops.softmax(add_50, -1) @@ -1249,13 +1165,15 @@ def forward( # pd_op.dropout: (1x12x11x11xf32, 1x12x11x11xui8) <- (1x12x11x11xf32, None, 1xf32) dropout_32, dropout_33 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_5, None, full_4, False, "upscale_in_train", 0, False + softmax_5, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_5 # pd_op.matmul: (1x12x11x64xf32) <- (1x12x11x11xf32, 1x12x11x64xf32) matmul_44 = paddle._C_ops.matmul(dropout_32, transpose_22, False, False) + del dropout_32, transpose_22 # pd_op.transpose: (1x11x12x64xf32) <- (1x12x11x64xf32) transpose_23 = paddle._C_ops.transpose(matmul_44, [0, 2, 1, 3]) @@ -1263,19 +1181,20 @@ def forward( # pd_op.reshape: (1x11x768xf32) <- (1x11x12x64xf32, 3xi64) reshape_23 = paddle._C_ops.reshape(transpose_23, full_int_array_2) + del transpose_23 # pd_op.matmul: (1x11x768xf32) <- (1x11x768xf32, 768x768xf32) matmul_45 = paddle._C_ops.matmul(reshape_23, parameter_107, False, False) - del parameter_107 + del parameter_107, reshape_23 # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_51 = paddle._C_ops.add(matmul_45, parameter_106) - del parameter_106 + del matmul_45, parameter_106 # pd_op.dropout: (1x11x768xf32, 1x11x768xui8) <- (1x11x768xf32, None, 1xf32) dropout_34, dropout_35 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_51, None, full_4, False, "upscale_in_train", 0, False + add_51, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1283,6 +1202,7 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 1x11x768xf32) add_52 = paddle._C_ops.add(layer_norm_30, dropout_34) + del dropout_34, layer_norm_30 # pd_op.layer_norm: (1x11x768xf32, 1x11xf32, 1x11xf32) <- (1x11x768xf32, 768xf32, 768xf32) layer_norm_33, layer_norm_34, layer_norm_35 = (lambda x, f: f(x))( @@ -1291,7 +1211,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_100, parameter_101 + del add_52, parameter_100, parameter_101 # pd_op.matmul: (1x11x3072xf32) <- (1x11x768xf32, 768x3072xf32) matmul_46 = paddle._C_ops.matmul(layer_norm_33, parameter_105, False, False) @@ -1299,7 +1219,7 @@ def forward( # pd_op.add: (1x11x3072xf32) <- (1x11x3072xf32, 3072xf32) add_53 = paddle._C_ops.add(matmul_46, parameter_104) - del parameter_104 + del matmul_46, parameter_104 # pd_op.relu: (1x11x3072xf32) <- (1x11x3072xf32) relu_5 = paddle._C_ops.relu(add_53) @@ -1307,16 +1227,16 @@ def forward( # pd_op.matmul: (1x11x768xf32) <- (1x11x3072xf32, 3072x768xf32) matmul_47 = paddle._C_ops.matmul(relu_5, parameter_103, False, False) - del parameter_103 + del parameter_103, relu_5 # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_54 = paddle._C_ops.add(matmul_47, parameter_102) - del parameter_102 + del matmul_47, parameter_102 # pd_op.dropout: (1x11x768xf32, 1x11x768xui8) <- (1x11x768xf32, None, 1xf32) dropout_36, dropout_37 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_54, None, full_4, False, "upscale_in_train", 0, False + add_54, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1324,6 +1244,7 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 1x11x768xf32) add_55 = paddle._C_ops.add(layer_norm_33, dropout_36) + del dropout_36, layer_norm_33 # pd_op.layer_norm: (1x11x768xf32, 1x11xf32, 1x11xf32) <- (1x11x768xf32, 768xf32, 768xf32) layer_norm_36, layer_norm_37, layer_norm_38 = (lambda x, f: f(x))( @@ -1332,7 +1253,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_98, parameter_99 + del add_55, parameter_98, parameter_99 # pd_op.matmul: (1x11x768xf32) <- (1x11x768xf32, 768x768xf32) matmul_48 = paddle._C_ops.matmul(layer_norm_36, parameter_97, False, False) @@ -1340,10 +1261,11 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_56 = paddle._C_ops.add(matmul_48, parameter_96) - del parameter_96 + del matmul_48, parameter_96 # pd_op.reshape: (1x11x12x64xf32) <- (1x11x768xf32, 4xi64) reshape_24 = paddle._C_ops.reshape(add_56, full_int_array_1) + del add_56 # pd_op.transpose: (1x12x11x64xf32) <- (1x11x12x64xf32) transpose_24 = paddle._C_ops.transpose(reshape_24, [0, 2, 1, 3]) @@ -1355,7 +1277,7 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_57 = paddle._C_ops.add(matmul_49, parameter_94) - del parameter_94 + del matmul_49, parameter_94 # pd_op.matmul: (1x11x768xf32) <- (1x11x768xf32, 768x768xf32) matmul_50 = paddle._C_ops.matmul(layer_norm_36, parameter_93, False, False) @@ -1363,10 +1285,11 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_58 = paddle._C_ops.add(matmul_50, parameter_92) - del parameter_92 + del matmul_50, parameter_92 # pd_op.reshape: (1x11x12x64xf32) <- (1x11x768xf32, 4xi64) reshape_25 = paddle._C_ops.reshape(add_57, full_int_array_1) + del add_57 # pd_op.transpose: (1x12x11x64xf32) <- (1x11x12x64xf32) transpose_25 = paddle._C_ops.transpose(reshape_25, [0, 2, 1, 3]) @@ -1374,6 +1297,7 @@ def forward( # pd_op.reshape: (1x11x12x64xf32) <- (1x11x768xf32, 4xi64) reshape_26 = paddle._C_ops.reshape(add_58, full_int_array_1) + del add_58 # pd_op.transpose: (1x12x11x64xf32) <- (1x11x12x64xf32) transpose_26 = paddle._C_ops.transpose(reshape_26, [0, 2, 1, 3]) @@ -1385,9 +1309,11 @@ def forward( # pd_op.matmul: (1x12x11x11xf32) <- (1x12x11x64xf32, 1x12x11x64xf32) matmul_51 = paddle._C_ops.matmul(scale_7, transpose_25, False, True) + del scale_7, transpose_25 # pd_op.add: (1x12x11x11xf32) <- (1x12x11x11xf32, 1x1x1x11xf32) add_59 = paddle._C_ops.add(matmul_51, unsqueeze_0) + del matmul_51 # pd_op.softmax: (1x12x11x11xf32) <- (1x12x11x11xf32) softmax_6 = paddle._C_ops.softmax(add_59, -1) @@ -1396,13 +1322,15 @@ def forward( # pd_op.dropout: (1x12x11x11xf32, 1x12x11x11xui8) <- (1x12x11x11xf32, None, 1xf32) dropout_38, dropout_39 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_6, None, full_4, False, "upscale_in_train", 0, False + softmax_6, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_6 # pd_op.matmul: (1x12x11x64xf32) <- (1x12x11x11xf32, 1x12x11x64xf32) matmul_52 = paddle._C_ops.matmul(dropout_38, transpose_26, False, False) + del dropout_38, transpose_26 # pd_op.transpose: (1x11x12x64xf32) <- (1x12x11x64xf32) transpose_27 = paddle._C_ops.transpose(matmul_52, [0, 2, 1, 3]) @@ -1410,19 +1338,20 @@ def forward( # pd_op.reshape: (1x11x768xf32) <- (1x11x12x64xf32, 3xi64) reshape_27 = paddle._C_ops.reshape(transpose_27, full_int_array_2) + del transpose_27 # pd_op.matmul: (1x11x768xf32) <- (1x11x768xf32, 768x768xf32) matmul_53 = paddle._C_ops.matmul(reshape_27, parameter_91, False, False) - del parameter_91 + del parameter_91, reshape_27 # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_60 = paddle._C_ops.add(matmul_53, parameter_90) - del parameter_90 + del matmul_53, parameter_90 # pd_op.dropout: (1x11x768xf32, 1x11x768xui8) <- (1x11x768xf32, None, 1xf32) dropout_40, dropout_41 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_60, None, full_4, False, "upscale_in_train", 0, False + add_60, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1430,6 +1359,7 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 1x11x768xf32) add_61 = paddle._C_ops.add(layer_norm_36, dropout_40) + del dropout_40, layer_norm_36 # pd_op.layer_norm: (1x11x768xf32, 1x11xf32, 1x11xf32) <- (1x11x768xf32, 768xf32, 768xf32) layer_norm_39, layer_norm_40, layer_norm_41 = (lambda x, f: f(x))( @@ -1438,7 +1368,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_84, parameter_85 + del add_61, parameter_84, parameter_85 # pd_op.matmul: (1x11x3072xf32) <- (1x11x768xf32, 768x3072xf32) matmul_54 = paddle._C_ops.matmul(layer_norm_39, parameter_89, False, False) @@ -1446,7 +1376,7 @@ def forward( # pd_op.add: (1x11x3072xf32) <- (1x11x3072xf32, 3072xf32) add_62 = paddle._C_ops.add(matmul_54, parameter_88) - del parameter_88 + del matmul_54, parameter_88 # pd_op.relu: (1x11x3072xf32) <- (1x11x3072xf32) relu_6 = paddle._C_ops.relu(add_62) @@ -1454,16 +1384,16 @@ def forward( # pd_op.matmul: (1x11x768xf32) <- (1x11x3072xf32, 3072x768xf32) matmul_55 = paddle._C_ops.matmul(relu_6, parameter_87, False, False) - del parameter_87 + del parameter_87, relu_6 # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_63 = paddle._C_ops.add(matmul_55, parameter_86) - del parameter_86 + del matmul_55, parameter_86 # pd_op.dropout: (1x11x768xf32, 1x11x768xui8) <- (1x11x768xf32, None, 1xf32) dropout_42, dropout_43 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_63, None, full_4, False, "upscale_in_train", 0, False + add_63, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1471,6 +1401,7 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 1x11x768xf32) add_64 = paddle._C_ops.add(layer_norm_39, dropout_42) + del dropout_42, layer_norm_39 # pd_op.layer_norm: (1x11x768xf32, 1x11xf32, 1x11xf32) <- (1x11x768xf32, 768xf32, 768xf32) layer_norm_42, layer_norm_43, layer_norm_44 = (lambda x, f: f(x))( @@ -1479,7 +1410,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_82, parameter_83 + del add_64, parameter_82, parameter_83 # pd_op.matmul: (1x11x768xf32) <- (1x11x768xf32, 768x768xf32) matmul_56 = paddle._C_ops.matmul(layer_norm_42, parameter_81, False, False) @@ -1487,10 +1418,11 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_65 = paddle._C_ops.add(matmul_56, parameter_80) - del parameter_80 + del matmul_56, parameter_80 # pd_op.reshape: (1x11x12x64xf32) <- (1x11x768xf32, 4xi64) reshape_28 = paddle._C_ops.reshape(add_65, full_int_array_1) + del add_65 # pd_op.transpose: (1x12x11x64xf32) <- (1x11x12x64xf32) transpose_28 = paddle._C_ops.transpose(reshape_28, [0, 2, 1, 3]) @@ -1502,7 +1434,7 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_66 = paddle._C_ops.add(matmul_57, parameter_78) - del parameter_78 + del matmul_57, parameter_78 # pd_op.matmul: (1x11x768xf32) <- (1x11x768xf32, 768x768xf32) matmul_58 = paddle._C_ops.matmul(layer_norm_42, parameter_77, False, False) @@ -1510,10 +1442,11 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_67 = paddle._C_ops.add(matmul_58, parameter_76) - del parameter_76 + del matmul_58, parameter_76 # pd_op.reshape: (1x11x12x64xf32) <- (1x11x768xf32, 4xi64) reshape_29 = paddle._C_ops.reshape(add_66, full_int_array_1) + del add_66 # pd_op.transpose: (1x12x11x64xf32) <- (1x11x12x64xf32) transpose_29 = paddle._C_ops.transpose(reshape_29, [0, 2, 1, 3]) @@ -1521,6 +1454,7 @@ def forward( # pd_op.reshape: (1x11x12x64xf32) <- (1x11x768xf32, 4xi64) reshape_30 = paddle._C_ops.reshape(add_67, full_int_array_1) + del add_67 # pd_op.transpose: (1x12x11x64xf32) <- (1x11x12x64xf32) transpose_30 = paddle._C_ops.transpose(reshape_30, [0, 2, 1, 3]) @@ -1532,9 +1466,11 @@ def forward( # pd_op.matmul: (1x12x11x11xf32) <- (1x12x11x64xf32, 1x12x11x64xf32) matmul_59 = paddle._C_ops.matmul(scale_8, transpose_29, False, True) + del scale_8, transpose_29 # pd_op.add: (1x12x11x11xf32) <- (1x12x11x11xf32, 1x1x1x11xf32) add_68 = paddle._C_ops.add(matmul_59, unsqueeze_0) + del matmul_59 # pd_op.softmax: (1x12x11x11xf32) <- (1x12x11x11xf32) softmax_7 = paddle._C_ops.softmax(add_68, -1) @@ -1543,13 +1479,15 @@ def forward( # pd_op.dropout: (1x12x11x11xf32, 1x12x11x11xui8) <- (1x12x11x11xf32, None, 1xf32) dropout_44, dropout_45 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_7, None, full_4, False, "upscale_in_train", 0, False + softmax_7, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_7 # pd_op.matmul: (1x12x11x64xf32) <- (1x12x11x11xf32, 1x12x11x64xf32) matmul_60 = paddle._C_ops.matmul(dropout_44, transpose_30, False, False) + del dropout_44, transpose_30 # pd_op.transpose: (1x11x12x64xf32) <- (1x12x11x64xf32) transpose_31 = paddle._C_ops.transpose(matmul_60, [0, 2, 1, 3]) @@ -1557,19 +1495,20 @@ def forward( # pd_op.reshape: (1x11x768xf32) <- (1x11x12x64xf32, 3xi64) reshape_31 = paddle._C_ops.reshape(transpose_31, full_int_array_2) + del transpose_31 # pd_op.matmul: (1x11x768xf32) <- (1x11x768xf32, 768x768xf32) matmul_61 = paddle._C_ops.matmul(reshape_31, parameter_75, False, False) - del parameter_75 + del parameter_75, reshape_31 # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_69 = paddle._C_ops.add(matmul_61, parameter_74) - del parameter_74 + del matmul_61, parameter_74 # pd_op.dropout: (1x11x768xf32, 1x11x768xui8) <- (1x11x768xf32, None, 1xf32) dropout_46, dropout_47 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_69, None, full_4, False, "upscale_in_train", 0, False + add_69, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1577,6 +1516,7 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 1x11x768xf32) add_70 = paddle._C_ops.add(layer_norm_42, dropout_46) + del dropout_46, layer_norm_42 # pd_op.layer_norm: (1x11x768xf32, 1x11xf32, 1x11xf32) <- (1x11x768xf32, 768xf32, 768xf32) layer_norm_45, layer_norm_46, layer_norm_47 = (lambda x, f: f(x))( @@ -1585,7 +1525,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_68, parameter_69 + del add_70, parameter_68, parameter_69 # pd_op.matmul: (1x11x3072xf32) <- (1x11x768xf32, 768x3072xf32) matmul_62 = paddle._C_ops.matmul(layer_norm_45, parameter_73, False, False) @@ -1593,7 +1533,7 @@ def forward( # pd_op.add: (1x11x3072xf32) <- (1x11x3072xf32, 3072xf32) add_71 = paddle._C_ops.add(matmul_62, parameter_72) - del parameter_72 + del matmul_62, parameter_72 # pd_op.relu: (1x11x3072xf32) <- (1x11x3072xf32) relu_7 = paddle._C_ops.relu(add_71) @@ -1601,16 +1541,16 @@ def forward( # pd_op.matmul: (1x11x768xf32) <- (1x11x3072xf32, 3072x768xf32) matmul_63 = paddle._C_ops.matmul(relu_7, parameter_71, False, False) - del parameter_71 + del parameter_71, relu_7 # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_72 = paddle._C_ops.add(matmul_63, parameter_70) - del parameter_70 + del matmul_63, parameter_70 # pd_op.dropout: (1x11x768xf32, 1x11x768xui8) <- (1x11x768xf32, None, 1xf32) dropout_48, dropout_49 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_72, None, full_4, False, "upscale_in_train", 0, False + add_72, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1618,6 +1558,7 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 1x11x768xf32) add_73 = paddle._C_ops.add(layer_norm_45, dropout_48) + del dropout_48, layer_norm_45 # pd_op.layer_norm: (1x11x768xf32, 1x11xf32, 1x11xf32) <- (1x11x768xf32, 768xf32, 768xf32) layer_norm_48, layer_norm_49, layer_norm_50 = (lambda x, f: f(x))( @@ -1626,7 +1567,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_66, parameter_67 + del add_73, parameter_66, parameter_67 # pd_op.matmul: (1x11x768xf32) <- (1x11x768xf32, 768x768xf32) matmul_64 = paddle._C_ops.matmul(layer_norm_48, parameter_65, False, False) @@ -1634,10 +1575,11 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_74 = paddle._C_ops.add(matmul_64, parameter_64) - del parameter_64 + del matmul_64, parameter_64 # pd_op.reshape: (1x11x12x64xf32) <- (1x11x768xf32, 4xi64) reshape_32 = paddle._C_ops.reshape(add_74, full_int_array_1) + del add_74 # pd_op.transpose: (1x12x11x64xf32) <- (1x11x12x64xf32) transpose_32 = paddle._C_ops.transpose(reshape_32, [0, 2, 1, 3]) @@ -1649,7 +1591,7 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_75 = paddle._C_ops.add(matmul_65, parameter_62) - del parameter_62 + del matmul_65, parameter_62 # pd_op.matmul: (1x11x768xf32) <- (1x11x768xf32, 768x768xf32) matmul_66 = paddle._C_ops.matmul(layer_norm_48, parameter_61, False, False) @@ -1657,10 +1599,11 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_76 = paddle._C_ops.add(matmul_66, parameter_60) - del parameter_60 + del matmul_66, parameter_60 # pd_op.reshape: (1x11x12x64xf32) <- (1x11x768xf32, 4xi64) reshape_33 = paddle._C_ops.reshape(add_75, full_int_array_1) + del add_75 # pd_op.transpose: (1x12x11x64xf32) <- (1x11x12x64xf32) transpose_33 = paddle._C_ops.transpose(reshape_33, [0, 2, 1, 3]) @@ -1668,6 +1611,7 @@ def forward( # pd_op.reshape: (1x11x12x64xf32) <- (1x11x768xf32, 4xi64) reshape_34 = paddle._C_ops.reshape(add_76, full_int_array_1) + del add_76 # pd_op.transpose: (1x12x11x64xf32) <- (1x11x12x64xf32) transpose_34 = paddle._C_ops.transpose(reshape_34, [0, 2, 1, 3]) @@ -1679,9 +1623,11 @@ def forward( # pd_op.matmul: (1x12x11x11xf32) <- (1x12x11x64xf32, 1x12x11x64xf32) matmul_67 = paddle._C_ops.matmul(scale_9, transpose_33, False, True) + del scale_9, transpose_33 # pd_op.add: (1x12x11x11xf32) <- (1x12x11x11xf32, 1x1x1x11xf32) add_77 = paddle._C_ops.add(matmul_67, unsqueeze_0) + del matmul_67 # pd_op.softmax: (1x12x11x11xf32) <- (1x12x11x11xf32) softmax_8 = paddle._C_ops.softmax(add_77, -1) @@ -1690,13 +1636,15 @@ def forward( # pd_op.dropout: (1x12x11x11xf32, 1x12x11x11xui8) <- (1x12x11x11xf32, None, 1xf32) dropout_50, dropout_51 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_8, None, full_4, False, "upscale_in_train", 0, False + softmax_8, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_8 # pd_op.matmul: (1x12x11x64xf32) <- (1x12x11x11xf32, 1x12x11x64xf32) matmul_68 = paddle._C_ops.matmul(dropout_50, transpose_34, False, False) + del dropout_50, transpose_34 # pd_op.transpose: (1x11x12x64xf32) <- (1x12x11x64xf32) transpose_35 = paddle._C_ops.transpose(matmul_68, [0, 2, 1, 3]) @@ -1704,19 +1652,20 @@ def forward( # pd_op.reshape: (1x11x768xf32) <- (1x11x12x64xf32, 3xi64) reshape_35 = paddle._C_ops.reshape(transpose_35, full_int_array_2) + del transpose_35 # pd_op.matmul: (1x11x768xf32) <- (1x11x768xf32, 768x768xf32) matmul_69 = paddle._C_ops.matmul(reshape_35, parameter_59, False, False) - del parameter_59 + del parameter_59, reshape_35 # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_78 = paddle._C_ops.add(matmul_69, parameter_58) - del parameter_58 + del matmul_69, parameter_58 # pd_op.dropout: (1x11x768xf32, 1x11x768xui8) <- (1x11x768xf32, None, 1xf32) dropout_52, dropout_53 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_78, None, full_4, False, "upscale_in_train", 0, False + add_78, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1724,6 +1673,7 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 1x11x768xf32) add_79 = paddle._C_ops.add(layer_norm_48, dropout_52) + del dropout_52, layer_norm_48 # pd_op.layer_norm: (1x11x768xf32, 1x11xf32, 1x11xf32) <- (1x11x768xf32, 768xf32, 768xf32) layer_norm_51, layer_norm_52, layer_norm_53 = (lambda x, f: f(x))( @@ -1732,7 +1682,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_52, parameter_53 + del add_79, parameter_52, parameter_53 # pd_op.matmul: (1x11x3072xf32) <- (1x11x768xf32, 768x3072xf32) matmul_70 = paddle._C_ops.matmul(layer_norm_51, parameter_57, False, False) @@ -1740,7 +1690,7 @@ def forward( # pd_op.add: (1x11x3072xf32) <- (1x11x3072xf32, 3072xf32) add_80 = paddle._C_ops.add(matmul_70, parameter_56) - del parameter_56 + del matmul_70, parameter_56 # pd_op.relu: (1x11x3072xf32) <- (1x11x3072xf32) relu_8 = paddle._C_ops.relu(add_80) @@ -1748,16 +1698,16 @@ def forward( # pd_op.matmul: (1x11x768xf32) <- (1x11x3072xf32, 3072x768xf32) matmul_71 = paddle._C_ops.matmul(relu_8, parameter_55, False, False) - del parameter_55 + del parameter_55, relu_8 # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_81 = paddle._C_ops.add(matmul_71, parameter_54) - del parameter_54 + del matmul_71, parameter_54 # pd_op.dropout: (1x11x768xf32, 1x11x768xui8) <- (1x11x768xf32, None, 1xf32) dropout_54, dropout_55 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_81, None, full_4, False, "upscale_in_train", 0, False + add_81, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1765,6 +1715,7 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 1x11x768xf32) add_82 = paddle._C_ops.add(layer_norm_51, dropout_54) + del dropout_54, layer_norm_51 # pd_op.layer_norm: (1x11x768xf32, 1x11xf32, 1x11xf32) <- (1x11x768xf32, 768xf32, 768xf32) layer_norm_54, layer_norm_55, layer_norm_56 = (lambda x, f: f(x))( @@ -1773,7 +1724,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_50, parameter_51 + del add_82, parameter_50, parameter_51 # pd_op.matmul: (1x11x768xf32) <- (1x11x768xf32, 768x768xf32) matmul_72 = paddle._C_ops.matmul(layer_norm_54, parameter_49, False, False) @@ -1781,10 +1732,11 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_83 = paddle._C_ops.add(matmul_72, parameter_48) - del parameter_48 + del matmul_72, parameter_48 # pd_op.reshape: (1x11x12x64xf32) <- (1x11x768xf32, 4xi64) reshape_36 = paddle._C_ops.reshape(add_83, full_int_array_1) + del add_83 # pd_op.transpose: (1x12x11x64xf32) <- (1x11x12x64xf32) transpose_36 = paddle._C_ops.transpose(reshape_36, [0, 2, 1, 3]) @@ -1796,7 +1748,7 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_84 = paddle._C_ops.add(matmul_73, parameter_46) - del parameter_46 + del matmul_73, parameter_46 # pd_op.matmul: (1x11x768xf32) <- (1x11x768xf32, 768x768xf32) matmul_74 = paddle._C_ops.matmul(layer_norm_54, parameter_45, False, False) @@ -1804,10 +1756,11 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_85 = paddle._C_ops.add(matmul_74, parameter_44) - del parameter_44 + del matmul_74, parameter_44 # pd_op.reshape: (1x11x12x64xf32) <- (1x11x768xf32, 4xi64) reshape_37 = paddle._C_ops.reshape(add_84, full_int_array_1) + del add_84 # pd_op.transpose: (1x12x11x64xf32) <- (1x11x12x64xf32) transpose_37 = paddle._C_ops.transpose(reshape_37, [0, 2, 1, 3]) @@ -1815,6 +1768,7 @@ def forward( # pd_op.reshape: (1x11x12x64xf32) <- (1x11x768xf32, 4xi64) reshape_38 = paddle._C_ops.reshape(add_85, full_int_array_1) + del add_85 # pd_op.transpose: (1x12x11x64xf32) <- (1x11x12x64xf32) transpose_38 = paddle._C_ops.transpose(reshape_38, [0, 2, 1, 3]) @@ -1826,9 +1780,11 @@ def forward( # pd_op.matmul: (1x12x11x11xf32) <- (1x12x11x64xf32, 1x12x11x64xf32) matmul_75 = paddle._C_ops.matmul(scale_10, transpose_37, False, True) + del scale_10, transpose_37 # pd_op.add: (1x12x11x11xf32) <- (1x12x11x11xf32, 1x1x1x11xf32) add_86 = paddle._C_ops.add(matmul_75, unsqueeze_0) + del matmul_75 # pd_op.softmax: (1x12x11x11xf32) <- (1x12x11x11xf32) softmax_9 = paddle._C_ops.softmax(add_86, -1) @@ -1837,13 +1793,15 @@ def forward( # pd_op.dropout: (1x12x11x11xf32, 1x12x11x11xui8) <- (1x12x11x11xf32, None, 1xf32) dropout_56, dropout_57 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_9, None, full_4, False, "upscale_in_train", 0, False + softmax_9, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_9 # pd_op.matmul: (1x12x11x64xf32) <- (1x12x11x11xf32, 1x12x11x64xf32) matmul_76 = paddle._C_ops.matmul(dropout_56, transpose_38, False, False) + del dropout_56, transpose_38 # pd_op.transpose: (1x11x12x64xf32) <- (1x12x11x64xf32) transpose_39 = paddle._C_ops.transpose(matmul_76, [0, 2, 1, 3]) @@ -1851,19 +1809,20 @@ def forward( # pd_op.reshape: (1x11x768xf32) <- (1x11x12x64xf32, 3xi64) reshape_39 = paddle._C_ops.reshape(transpose_39, full_int_array_2) + del transpose_39 # pd_op.matmul: (1x11x768xf32) <- (1x11x768xf32, 768x768xf32) matmul_77 = paddle._C_ops.matmul(reshape_39, parameter_43, False, False) - del parameter_43 + del parameter_43, reshape_39 # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_87 = paddle._C_ops.add(matmul_77, parameter_42) - del parameter_42 + del matmul_77, parameter_42 # pd_op.dropout: (1x11x768xf32, 1x11x768xui8) <- (1x11x768xf32, None, 1xf32) dropout_58, dropout_59 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_87, None, full_4, False, "upscale_in_train", 0, False + add_87, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1871,6 +1830,7 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 1x11x768xf32) add_88 = paddle._C_ops.add(layer_norm_54, dropout_58) + del dropout_58, layer_norm_54 # pd_op.layer_norm: (1x11x768xf32, 1x11xf32, 1x11xf32) <- (1x11x768xf32, 768xf32, 768xf32) layer_norm_57, layer_norm_58, layer_norm_59 = (lambda x, f: f(x))( @@ -1879,7 +1839,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_36, parameter_37 + del add_88, parameter_36, parameter_37 # pd_op.matmul: (1x11x3072xf32) <- (1x11x768xf32, 768x3072xf32) matmul_78 = paddle._C_ops.matmul(layer_norm_57, parameter_41, False, False) @@ -1887,7 +1847,7 @@ def forward( # pd_op.add: (1x11x3072xf32) <- (1x11x3072xf32, 3072xf32) add_89 = paddle._C_ops.add(matmul_78, parameter_40) - del parameter_40 + del matmul_78, parameter_40 # pd_op.relu: (1x11x3072xf32) <- (1x11x3072xf32) relu_9 = paddle._C_ops.relu(add_89) @@ -1895,16 +1855,16 @@ def forward( # pd_op.matmul: (1x11x768xf32) <- (1x11x3072xf32, 3072x768xf32) matmul_79 = paddle._C_ops.matmul(relu_9, parameter_39, False, False) - del parameter_39 + del parameter_39, relu_9 # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_90 = paddle._C_ops.add(matmul_79, parameter_38) - del parameter_38 + del matmul_79, parameter_38 # pd_op.dropout: (1x11x768xf32, 1x11x768xui8) <- (1x11x768xf32, None, 1xf32) dropout_60, dropout_61 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_90, None, full_4, False, "upscale_in_train", 0, False + add_90, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1912,6 +1872,7 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 1x11x768xf32) add_91 = paddle._C_ops.add(layer_norm_57, dropout_60) + del dropout_60, layer_norm_57 # pd_op.layer_norm: (1x11x768xf32, 1x11xf32, 1x11xf32) <- (1x11x768xf32, 768xf32, 768xf32) layer_norm_60, layer_norm_61, layer_norm_62 = (lambda x, f: f(x))( @@ -1920,7 +1881,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_34, parameter_35 + del add_91, parameter_34, parameter_35 # pd_op.matmul: (1x11x768xf32) <- (1x11x768xf32, 768x768xf32) matmul_80 = paddle._C_ops.matmul(layer_norm_60, parameter_33, False, False) @@ -1928,10 +1889,11 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_92 = paddle._C_ops.add(matmul_80, parameter_32) - del parameter_32 + del matmul_80, parameter_32 # pd_op.reshape: (1x11x12x64xf32) <- (1x11x768xf32, 4xi64) reshape_40 = paddle._C_ops.reshape(add_92, full_int_array_1) + del add_92 # pd_op.transpose: (1x12x11x64xf32) <- (1x11x12x64xf32) transpose_40 = paddle._C_ops.transpose(reshape_40, [0, 2, 1, 3]) @@ -1943,7 +1905,7 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_93 = paddle._C_ops.add(matmul_81, parameter_30) - del parameter_30 + del matmul_81, parameter_30 # pd_op.matmul: (1x11x768xf32) <- (1x11x768xf32, 768x768xf32) matmul_82 = paddle._C_ops.matmul(layer_norm_60, parameter_29, False, False) @@ -1951,10 +1913,11 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_94 = paddle._C_ops.add(matmul_82, parameter_28) - del parameter_28 + del matmul_82, parameter_28 # pd_op.reshape: (1x11x12x64xf32) <- (1x11x768xf32, 4xi64) reshape_41 = paddle._C_ops.reshape(add_93, full_int_array_1) + del add_93 # pd_op.transpose: (1x12x11x64xf32) <- (1x11x12x64xf32) transpose_41 = paddle._C_ops.transpose(reshape_41, [0, 2, 1, 3]) @@ -1962,6 +1925,7 @@ def forward( # pd_op.reshape: (1x11x12x64xf32) <- (1x11x768xf32, 4xi64) reshape_42 = paddle._C_ops.reshape(add_94, full_int_array_1) + del add_94 # pd_op.transpose: (1x12x11x64xf32) <- (1x11x12x64xf32) transpose_42 = paddle._C_ops.transpose(reshape_42, [0, 2, 1, 3]) @@ -1973,9 +1937,11 @@ def forward( # pd_op.matmul: (1x12x11x11xf32) <- (1x12x11x64xf32, 1x12x11x64xf32) matmul_83 = paddle._C_ops.matmul(scale_11, transpose_41, False, True) + del scale_11, transpose_41 # pd_op.add: (1x12x11x11xf32) <- (1x12x11x11xf32, 1x1x1x11xf32) add_95 = paddle._C_ops.add(matmul_83, unsqueeze_0) + del matmul_83 # pd_op.softmax: (1x12x11x11xf32) <- (1x12x11x11xf32) softmax_10 = paddle._C_ops.softmax(add_95, -1) @@ -1984,13 +1950,15 @@ def forward( # pd_op.dropout: (1x12x11x11xf32, 1x12x11x11xui8) <- (1x12x11x11xf32, None, 1xf32) dropout_62, dropout_63 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_10, None, full_4, False, "upscale_in_train", 0, False + softmax_10, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_10 # pd_op.matmul: (1x12x11x64xf32) <- (1x12x11x11xf32, 1x12x11x64xf32) matmul_84 = paddle._C_ops.matmul(dropout_62, transpose_42, False, False) + del dropout_62, transpose_42 # pd_op.transpose: (1x11x12x64xf32) <- (1x12x11x64xf32) transpose_43 = paddle._C_ops.transpose(matmul_84, [0, 2, 1, 3]) @@ -1998,19 +1966,20 @@ def forward( # pd_op.reshape: (1x11x768xf32) <- (1x11x12x64xf32, 3xi64) reshape_43 = paddle._C_ops.reshape(transpose_43, full_int_array_2) + del transpose_43 # pd_op.matmul: (1x11x768xf32) <- (1x11x768xf32, 768x768xf32) matmul_85 = paddle._C_ops.matmul(reshape_43, parameter_27, False, False) - del parameter_27 + del parameter_27, reshape_43 # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_96 = paddle._C_ops.add(matmul_85, parameter_26) - del parameter_26 + del matmul_85, parameter_26 # pd_op.dropout: (1x11x768xf32, 1x11x768xui8) <- (1x11x768xf32, None, 1xf32) dropout_64, dropout_65 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_96, None, full_4, False, "upscale_in_train", 0, False + add_96, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -2018,6 +1987,7 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 1x11x768xf32) add_97 = paddle._C_ops.add(layer_norm_60, dropout_64) + del dropout_64, layer_norm_60 # pd_op.layer_norm: (1x11x768xf32, 1x11xf32, 1x11xf32) <- (1x11x768xf32, 768xf32, 768xf32) layer_norm_63, layer_norm_64, layer_norm_65 = (lambda x, f: f(x))( @@ -2026,7 +1996,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_20, parameter_21 + del add_97, parameter_20, parameter_21 # pd_op.matmul: (1x11x3072xf32) <- (1x11x768xf32, 768x3072xf32) matmul_86 = paddle._C_ops.matmul(layer_norm_63, parameter_25, False, False) @@ -2034,7 +2004,7 @@ def forward( # pd_op.add: (1x11x3072xf32) <- (1x11x3072xf32, 3072xf32) add_98 = paddle._C_ops.add(matmul_86, parameter_24) - del parameter_24 + del matmul_86, parameter_24 # pd_op.relu: (1x11x3072xf32) <- (1x11x3072xf32) relu_10 = paddle._C_ops.relu(add_98) @@ -2042,16 +2012,16 @@ def forward( # pd_op.matmul: (1x11x768xf32) <- (1x11x3072xf32, 3072x768xf32) matmul_87 = paddle._C_ops.matmul(relu_10, parameter_23, False, False) - del parameter_23 + del parameter_23, relu_10 # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_99 = paddle._C_ops.add(matmul_87, parameter_22) - del parameter_22 + del matmul_87, parameter_22 # pd_op.dropout: (1x11x768xf32, 1x11x768xui8) <- (1x11x768xf32, None, 1xf32) dropout_66, dropout_67 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_99, None, full_4, False, "upscale_in_train", 0, False + add_99, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -2059,6 +2029,7 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 1x11x768xf32) add_100 = paddle._C_ops.add(layer_norm_63, dropout_66) + del dropout_66, layer_norm_63 # pd_op.layer_norm: (1x11x768xf32, 1x11xf32, 1x11xf32) <- (1x11x768xf32, 768xf32, 768xf32) layer_norm_66, layer_norm_67, layer_norm_68 = (lambda x, f: f(x))( @@ -2067,7 +2038,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_18, parameter_19 + del add_100, parameter_18, parameter_19 # pd_op.matmul: (1x11x768xf32) <- (1x11x768xf32, 768x768xf32) matmul_88 = paddle._C_ops.matmul(layer_norm_66, parameter_17, False, False) @@ -2075,10 +2046,11 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_101 = paddle._C_ops.add(matmul_88, parameter_16) - del parameter_16 + del matmul_88, parameter_16 # pd_op.reshape: (1x11x12x64xf32) <- (1x11x768xf32, 4xi64) reshape_44 = paddle._C_ops.reshape(add_101, full_int_array_1) + del add_101 # pd_op.transpose: (1x12x11x64xf32) <- (1x11x12x64xf32) transpose_44 = paddle._C_ops.transpose(reshape_44, [0, 2, 1, 3]) @@ -2090,7 +2062,7 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_102 = paddle._C_ops.add(matmul_89, parameter_14) - del parameter_14 + del matmul_89, parameter_14 # pd_op.matmul: (1x11x768xf32) <- (1x11x768xf32, 768x768xf32) matmul_90 = paddle._C_ops.matmul(layer_norm_66, parameter_13, False, False) @@ -2098,10 +2070,11 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_103 = paddle._C_ops.add(matmul_90, parameter_12) - del parameter_12 + del matmul_90, parameter_12 # pd_op.reshape: (1x11x12x64xf32) <- (1x11x768xf32, 4xi64) reshape_45 = paddle._C_ops.reshape(add_102, full_int_array_1) + del add_102 # pd_op.transpose: (1x12x11x64xf32) <- (1x11x12x64xf32) transpose_45 = paddle._C_ops.transpose(reshape_45, [0, 2, 1, 3]) @@ -2109,7 +2082,7 @@ def forward( # pd_op.reshape: (1x11x12x64xf32) <- (1x11x768xf32, 4xi64) reshape_46 = paddle._C_ops.reshape(add_103, full_int_array_1) - del full_int_array_1 + del add_103, full_int_array_1 # pd_op.transpose: (1x12x11x64xf32) <- (1x11x12x64xf32) transpose_46 = paddle._C_ops.transpose(reshape_46, [0, 2, 1, 3]) @@ -2117,13 +2090,15 @@ def forward( # pd_op.scale: (1x12x11x64xf32) <- (1x12x11x64xf32, 1xf32) scale_12 = paddle._C_ops.scale(transpose_44, full_5, float("0"), True) - del transpose_44 + del full_5, transpose_44 # pd_op.matmul: (1x12x11x11xf32) <- (1x12x11x64xf32, 1x12x11x64xf32) matmul_91 = paddle._C_ops.matmul(scale_12, transpose_45, False, True) + del scale_12, transpose_45 # pd_op.add: (1x12x11x11xf32) <- (1x12x11x11xf32, 1x1x1x11xf32) add_104 = paddle._C_ops.add(matmul_91, unsqueeze_0) + del matmul_91, unsqueeze_0 # pd_op.softmax: (1x12x11x11xf32) <- (1x12x11x11xf32) softmax_11 = paddle._C_ops.softmax(add_104, -1) @@ -2132,13 +2107,15 @@ def forward( # pd_op.dropout: (1x12x11x11xf32, 1x12x11x11xui8) <- (1x12x11x11xf32, None, 1xf32) dropout_68, dropout_69 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_11, None, full_4, False, "upscale_in_train", 0, False + softmax_11, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_11 # pd_op.matmul: (1x12x11x64xf32) <- (1x12x11x11xf32, 1x12x11x64xf32) matmul_92 = paddle._C_ops.matmul(dropout_68, transpose_46, False, False) + del dropout_68, transpose_46 # pd_op.transpose: (1x11x12x64xf32) <- (1x12x11x64xf32) transpose_47 = paddle._C_ops.transpose(matmul_92, [0, 2, 1, 3]) @@ -2146,20 +2123,20 @@ def forward( # pd_op.reshape: (1x11x768xf32) <- (1x11x12x64xf32, 3xi64) reshape_47 = paddle._C_ops.reshape(transpose_47, full_int_array_2) - del full_int_array_2 + del full_int_array_2, transpose_47 # pd_op.matmul: (1x11x768xf32) <- (1x11x768xf32, 768x768xf32) matmul_93 = paddle._C_ops.matmul(reshape_47, parameter_11, False, False) - del parameter_11 + del parameter_11, reshape_47 # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_105 = paddle._C_ops.add(matmul_93, parameter_10) - del parameter_10 + del matmul_93, parameter_10 # pd_op.dropout: (1x11x768xf32, 1x11x768xui8) <- (1x11x768xf32, None, 1xf32) dropout_70, dropout_71 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_105, None, full_4, False, "upscale_in_train", 0, False + add_105, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -2167,6 +2144,7 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 1x11x768xf32) add_106 = paddle._C_ops.add(layer_norm_66, dropout_70) + del dropout_70, layer_norm_66 # pd_op.layer_norm: (1x11x768xf32, 1x11xf32, 1x11xf32) <- (1x11x768xf32, 768xf32, 768xf32) layer_norm_69, layer_norm_70, layer_norm_71 = (lambda x, f: f(x))( @@ -2175,7 +2153,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_4, parameter_5 + del add_106, parameter_4, parameter_5 # pd_op.matmul: (1x11x3072xf32) <- (1x11x768xf32, 768x3072xf32) matmul_94 = paddle._C_ops.matmul(layer_norm_69, parameter_9, False, False) @@ -2183,7 +2161,7 @@ def forward( # pd_op.add: (1x11x3072xf32) <- (1x11x3072xf32, 3072xf32) add_107 = paddle._C_ops.add(matmul_94, parameter_8) - del parameter_8 + del matmul_94, parameter_8 # pd_op.relu: (1x11x3072xf32) <- (1x11x3072xf32) relu_11 = paddle._C_ops.relu(add_107) @@ -2191,23 +2169,24 @@ def forward( # pd_op.matmul: (1x11x768xf32) <- (1x11x3072xf32, 3072x768xf32) matmul_95 = paddle._C_ops.matmul(relu_11, parameter_7, False, False) - del parameter_7 + del parameter_7, relu_11 # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_108 = paddle._C_ops.add(matmul_95, parameter_6) - del parameter_6 + del matmul_95, parameter_6 # pd_op.dropout: (1x11x768xf32, 1x11x768xui8) <- (1x11x768xf32, None, 1xf32) dropout_72, dropout_73 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_108, None, full_4, False, "upscale_in_train", 0, False + add_108, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) - del add_108 + del add_108, full_4 # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 1x11x768xf32) add_109 = paddle._C_ops.add(layer_norm_69, dropout_72) + del dropout_72, layer_norm_69 # pd_op.layer_norm: (1x11x768xf32, 1x11xf32, 1x11xf32) <- (1x11x768xf32, 768xf32, 768xf32) layer_norm_72, layer_norm_73, layer_norm_74 = (lambda x, f: f(x))( @@ -2216,7 +2195,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_2, parameter_3 + del add_109, parameter_2, parameter_3 # pd_op.full_int_array: (1xi64) <- () full_int_array_3 = [0] @@ -2228,455 +2207,18 @@ def forward( slice_0 = paddle._C_ops.slice( layer_norm_72, [1], full_int_array_3, full_int_array_4, [1], [1] ) + del full_int_array_3, full_int_array_4, layer_norm_72 # pd_op.matmul: (1x768xf32) <- (1x768xf32, 768x768xf32) matmul_96 = paddle._C_ops.matmul(slice_0, parameter_1, False, False) - del parameter_1 + del parameter_1, slice_0 # pd_op.add: (1x768xf32) <- (1x768xf32, 768xf32) add_110 = paddle._C_ops.add(matmul_96, parameter_0) - del parameter_0 + del matmul_96, parameter_0 # pd_op.tanh: (1x768xf32) <- (1x768xf32) tanh_0 = paddle._C_ops.tanh(add_110) - del ( - add_0, - add_1, - add_10, - add_100, - add_101, - add_102, - add_103, - add_106, - add_109, - add_11, - add_110, - add_12, - add_13, - add_16, - add_19, - add_2, - add_20, - add_21, - add_22, - add_25, - add_28, - add_29, - add_3, - add_30, - add_31, - add_34, - add_37, - add_38, - add_39, - add_4, - add_40, - add_43, - add_46, - add_47, - add_48, - add_49, - add_52, - add_55, - add_56, - add_57, - add_58, - add_61, - add_64, - add_65, - add_66, - add_67, - add_7, - add_70, - add_73, - add_74, - add_75, - add_76, - add_79, - add_82, - add_83, - add_84, - add_85, - add_88, - add_91, - add_92, - add_93, - add_94, - add_97, - assign_0, - assign_1, - assign_10, - assign_11, - assign_12, - assign_13, - assign_14, - assign_15, - assign_16, - assign_17, - assign_18, - assign_19, - assign_2, - assign_20, - assign_21, - assign_22, - assign_23, - assign_24, - assign_25, - assign_26, - assign_27, - assign_28, - assign_29, - assign_3, - assign_30, - assign_31, - assign_32, - assign_33, - assign_34, - assign_35, - assign_36, - assign_37, - assign_38, - assign_39, - assign_4, - assign_40, - assign_41, - assign_42, - assign_43, - assign_44, - assign_45, - assign_46, - assign_5, - assign_6, - assign_7, - assign_8, - assign_9, - dropout_0, - dropout_1, - dropout_10, - dropout_11, - dropout_12, - dropout_13, - dropout_14, - dropout_15, - dropout_16, - dropout_17, - dropout_18, - dropout_19, - dropout_2, - dropout_20, - dropout_21, - dropout_22, - dropout_23, - dropout_24, - dropout_25, - dropout_26, - dropout_27, - dropout_28, - dropout_29, - dropout_3, - dropout_30, - dropout_31, - dropout_32, - dropout_33, - dropout_34, - dropout_35, - dropout_36, - dropout_37, - dropout_38, - dropout_39, - dropout_4, - dropout_40, - dropout_41, - dropout_42, - dropout_43, - dropout_44, - dropout_45, - dropout_46, - dropout_47, - dropout_48, - dropout_49, - dropout_5, - dropout_50, - dropout_51, - dropout_52, - dropout_53, - dropout_54, - dropout_55, - dropout_56, - dropout_57, - dropout_58, - dropout_59, - dropout_6, - dropout_60, - dropout_61, - dropout_62, - dropout_63, - dropout_64, - dropout_65, - dropout_66, - dropout_67, - dropout_68, - dropout_69, - dropout_7, - dropout_70, - dropout_71, - dropout_72, - dropout_73, - dropout_8, - dropout_9, - embedding_0, - embedding_1, - embedding_2, - full_4, - full_5, - full_int_array_3, - full_int_array_4, - layer_norm_1, - layer_norm_10, - layer_norm_11, - layer_norm_12, - layer_norm_13, - layer_norm_14, - layer_norm_15, - layer_norm_16, - layer_norm_17, - layer_norm_18, - layer_norm_19, - layer_norm_2, - layer_norm_20, - layer_norm_21, - layer_norm_22, - layer_norm_23, - layer_norm_24, - layer_norm_25, - layer_norm_26, - layer_norm_27, - layer_norm_28, - layer_norm_29, - layer_norm_3, - layer_norm_30, - layer_norm_31, - layer_norm_32, - layer_norm_33, - layer_norm_34, - layer_norm_35, - layer_norm_36, - layer_norm_37, - layer_norm_38, - layer_norm_39, - layer_norm_4, - layer_norm_40, - layer_norm_41, - layer_norm_42, - layer_norm_43, - layer_norm_44, - layer_norm_45, - layer_norm_46, - layer_norm_47, - layer_norm_48, - layer_norm_49, - layer_norm_5, - layer_norm_50, - layer_norm_51, - layer_norm_52, - layer_norm_53, - layer_norm_54, - layer_norm_55, - layer_norm_56, - layer_norm_57, - layer_norm_58, - layer_norm_59, - layer_norm_6, - layer_norm_60, - layer_norm_61, - layer_norm_62, - layer_norm_63, - layer_norm_64, - layer_norm_65, - layer_norm_66, - layer_norm_67, - layer_norm_68, - layer_norm_69, - layer_norm_7, - layer_norm_70, - layer_norm_71, - layer_norm_72, - layer_norm_73, - layer_norm_74, - layer_norm_8, - layer_norm_9, - matmul_0, - matmul_1, - matmul_10, - matmul_11, - matmul_13, - matmul_14, - matmul_15, - matmul_16, - matmul_17, - matmul_18, - matmul_19, - matmul_2, - matmul_21, - matmul_22, - matmul_23, - matmul_24, - matmul_25, - matmul_26, - matmul_27, - matmul_29, - matmul_3, - matmul_30, - matmul_31, - matmul_32, - matmul_33, - matmul_34, - matmul_35, - matmul_37, - matmul_38, - matmul_39, - matmul_40, - matmul_41, - matmul_42, - matmul_43, - matmul_45, - matmul_46, - matmul_47, - matmul_48, - matmul_49, - matmul_5, - matmul_50, - matmul_51, - matmul_53, - matmul_54, - matmul_55, - matmul_56, - matmul_57, - matmul_58, - matmul_59, - matmul_6, - matmul_61, - matmul_62, - matmul_63, - matmul_64, - matmul_65, - matmul_66, - matmul_67, - matmul_69, - matmul_7, - matmul_70, - matmul_71, - matmul_72, - matmul_73, - matmul_74, - matmul_75, - matmul_77, - matmul_78, - matmul_79, - matmul_8, - matmul_80, - matmul_81, - matmul_82, - matmul_83, - matmul_85, - matmul_86, - matmul_87, - matmul_88, - matmul_89, - matmul_9, - matmul_90, - matmul_91, - matmul_93, - matmul_94, - matmul_95, - matmul_96, - relu_0, - relu_1, - relu_10, - relu_11, - relu_2, - relu_3, - relu_4, - relu_5, - relu_6, - relu_7, - relu_8, - relu_9, - reshape_11, - reshape_15, - reshape_19, - reshape_23, - reshape_27, - reshape_3, - reshape_31, - reshape_35, - reshape_39, - reshape_43, - reshape_47, - reshape_7, - scale_1, - scale_10, - scale_11, - scale_12, - scale_2, - scale_3, - scale_4, - scale_5, - scale_6, - scale_7, - scale_8, - scale_9, - slice_0, - softmax_0, - softmax_1, - softmax_10, - softmax_11, - softmax_2, - softmax_3, - softmax_4, - softmax_5, - softmax_6, - softmax_7, - softmax_8, - softmax_9, - subtract_0, - transpose_1, - transpose_10, - transpose_11, - transpose_13, - transpose_14, - transpose_15, - transpose_17, - transpose_18, - transpose_19, - transpose_2, - transpose_21, - transpose_22, - transpose_23, - transpose_25, - transpose_26, - transpose_27, - transpose_29, - transpose_3, - transpose_30, - transpose_31, - transpose_33, - transpose_34, - transpose_35, - transpose_37, - transpose_38, - transpose_39, - transpose_41, - transpose_42, - transpose_43, - transpose_45, - transpose_46, - transpose_47, - transpose_5, - transpose_6, - transpose_7, - transpose_9, - unsqueeze_0, - ) + del add_110 return tanh_0 diff --git a/paddle_samples/PaddleNLP/ernie-2.0-base-zh/weight_meta.py b/paddle_samples/PaddleNLP/ernie-2.0-base-zh/weight_meta.py index c8534568b..0e5812bd3 100644 --- a/paddle_samples/PaddleNLP/ernie-2.0-base-zh/weight_meta.py +++ b/paddle_samples/PaddleNLP/ernie-2.0-base-zh/weight_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_parameter_0: name = "parameter_0" + original_name = "linear_436.b_0" shape = [768] dtype = "float32" min_val = float("-0.512987") @@ -11,6 +12,7 @@ class Program_weight_tensor_parameter_0: class Program_weight_tensor_parameter_1: name = "parameter_1" + original_name = "linear_436.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.32332") @@ -22,6 +24,7 @@ class Program_weight_tensor_parameter_1: class Program_weight_tensor_parameter_2: name = "parameter_2" + original_name = "layer_norm_148.b_0" shape = [768] dtype = "float32" min_val = float("-0.981212") @@ -33,6 +36,7 @@ class Program_weight_tensor_parameter_2: class Program_weight_tensor_parameter_3: name = "parameter_3" + original_name = "layer_norm_148.w_0" shape = [768] dtype = "float32" min_val = float("0.0943044") @@ -44,6 +48,7 @@ class Program_weight_tensor_parameter_3: class Program_weight_tensor_parameter_4: name = "parameter_4" + original_name = "layer_norm_147.b_0" shape = [768] dtype = "float32" min_val = float("-2.57665") @@ -55,6 +60,7 @@ class Program_weight_tensor_parameter_4: class Program_weight_tensor_parameter_5: name = "parameter_5" + original_name = "layer_norm_147.w_0" shape = [768] dtype = "float32" min_val = float("0.172685") @@ -66,6 +72,7 @@ class Program_weight_tensor_parameter_5: class Program_weight_tensor_parameter_6: name = "parameter_6" + original_name = "linear_435.b_0" shape = [768] dtype = "float32" min_val = float("-1.06961") @@ -77,6 +84,7 @@ class Program_weight_tensor_parameter_6: class Program_weight_tensor_parameter_7: name = "parameter_7" + original_name = "linear_435.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-1.98247") @@ -88,6 +96,7 @@ class Program_weight_tensor_parameter_7: class Program_weight_tensor_parameter_8: name = "parameter_8" + original_name = "linear_434.b_0" shape = [3072] dtype = "float32" min_val = float("-2.43092") @@ -99,6 +108,7 @@ class Program_weight_tensor_parameter_8: class Program_weight_tensor_parameter_9: name = "parameter_9" + original_name = "linear_434.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.663976") @@ -110,6 +120,7 @@ class Program_weight_tensor_parameter_9: class Program_weight_tensor_parameter_10: name = "parameter_10" + original_name = "linear_433.b_0" shape = [768] dtype = "float32" min_val = float("-0.428988") @@ -121,6 +132,7 @@ class Program_weight_tensor_parameter_10: class Program_weight_tensor_parameter_11: name = "parameter_11" + original_name = "linear_433.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.735509") @@ -132,6 +144,7 @@ class Program_weight_tensor_parameter_11: class Program_weight_tensor_parameter_12: name = "parameter_12" + original_name = "linear_432.b_0" shape = [768] dtype = "float32" min_val = float("-0.907722") @@ -143,6 +156,7 @@ class Program_weight_tensor_parameter_12: class Program_weight_tensor_parameter_13: name = "parameter_13" + original_name = "linear_432.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.263635") @@ -154,6 +168,7 @@ class Program_weight_tensor_parameter_13: class Program_weight_tensor_parameter_14: name = "parameter_14" + original_name = "linear_431.b_0" shape = [768] dtype = "float32" min_val = float("-14.6171") @@ -165,6 +180,7 @@ class Program_weight_tensor_parameter_14: class Program_weight_tensor_parameter_15: name = "parameter_15" + original_name = "linear_431.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.648596") @@ -176,6 +192,7 @@ class Program_weight_tensor_parameter_15: class Program_weight_tensor_parameter_16: name = "parameter_16" + original_name = "linear_430.b_0" shape = [768] dtype = "float32" min_val = float("-3.13484") @@ -187,6 +204,7 @@ class Program_weight_tensor_parameter_16: class Program_weight_tensor_parameter_17: name = "parameter_17" + original_name = "linear_430.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.706931") @@ -198,6 +216,7 @@ class Program_weight_tensor_parameter_17: class Program_weight_tensor_parameter_18: name = "parameter_18" + original_name = "layer_norm_146.b_0" shape = [768] dtype = "float32" min_val = float("-3.11886") @@ -209,6 +228,7 @@ class Program_weight_tensor_parameter_18: class Program_weight_tensor_parameter_19: name = "parameter_19" + original_name = "layer_norm_146.w_0" shape = [768] dtype = "float32" min_val = float("0.164392") @@ -220,6 +240,7 @@ class Program_weight_tensor_parameter_19: class Program_weight_tensor_parameter_20: name = "parameter_20" + original_name = "layer_norm_145.b_0" shape = [768] dtype = "float32" min_val = float("-5.15438") @@ -231,6 +252,7 @@ class Program_weight_tensor_parameter_20: class Program_weight_tensor_parameter_21: name = "parameter_21" + original_name = "layer_norm_145.w_0" shape = [768] dtype = "float32" min_val = float("0.587894") @@ -242,6 +264,7 @@ class Program_weight_tensor_parameter_21: class Program_weight_tensor_parameter_22: name = "parameter_22" + original_name = "linear_429.b_0" shape = [768] dtype = "float32" min_val = float("-0.33617") @@ -253,6 +276,7 @@ class Program_weight_tensor_parameter_22: class Program_weight_tensor_parameter_23: name = "parameter_23" + original_name = "linear_429.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-0.803004") @@ -264,6 +288,7 @@ class Program_weight_tensor_parameter_23: class Program_weight_tensor_parameter_24: name = "parameter_24" + original_name = "linear_428.b_0" shape = [3072] dtype = "float32" min_val = float("-2.14237") @@ -275,6 +300,7 @@ class Program_weight_tensor_parameter_24: class Program_weight_tensor_parameter_25: name = "parameter_25" + original_name = "linear_428.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.499519") @@ -286,6 +312,7 @@ class Program_weight_tensor_parameter_25: class Program_weight_tensor_parameter_26: name = "parameter_26" + original_name = "linear_427.b_0" shape = [768] dtype = "float32" min_val = float("-0.2218") @@ -297,6 +324,7 @@ class Program_weight_tensor_parameter_26: class Program_weight_tensor_parameter_27: name = "parameter_27" + original_name = "linear_427.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.297697") @@ -308,6 +336,7 @@ class Program_weight_tensor_parameter_27: class Program_weight_tensor_parameter_28: name = "parameter_28" + original_name = "linear_426.b_0" shape = [768] dtype = "float32" min_val = float("-0.49257") @@ -319,6 +348,7 @@ class Program_weight_tensor_parameter_28: class Program_weight_tensor_parameter_29: name = "parameter_29" + original_name = "linear_426.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.411621") @@ -330,6 +360,7 @@ class Program_weight_tensor_parameter_29: class Program_weight_tensor_parameter_30: name = "parameter_30" + original_name = "linear_425.b_0" shape = [768] dtype = "float32" min_val = float("-17.7545") @@ -341,6 +372,7 @@ class Program_weight_tensor_parameter_30: class Program_weight_tensor_parameter_31: name = "parameter_31" + original_name = "linear_425.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.440984") @@ -352,6 +384,7 @@ class Program_weight_tensor_parameter_31: class Program_weight_tensor_parameter_32: name = "parameter_32" + original_name = "linear_424.b_0" shape = [768] dtype = "float32" min_val = float("-3.00254") @@ -363,6 +396,7 @@ class Program_weight_tensor_parameter_32: class Program_weight_tensor_parameter_33: name = "parameter_33" + original_name = "linear_424.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.697498") @@ -374,6 +408,7 @@ class Program_weight_tensor_parameter_33: class Program_weight_tensor_parameter_34: name = "parameter_34" + original_name = "layer_norm_144.b_0" shape = [768] dtype = "float32" min_val = float("-0.332045") @@ -385,6 +420,7 @@ class Program_weight_tensor_parameter_34: class Program_weight_tensor_parameter_35: name = "parameter_35" + original_name = "layer_norm_144.w_0" shape = [768] dtype = "float32" min_val = float("0.229145") @@ -396,6 +432,7 @@ class Program_weight_tensor_parameter_35: class Program_weight_tensor_parameter_36: name = "parameter_36" + original_name = "layer_norm_143.b_0" shape = [768] dtype = "float32" min_val = float("-5.60329") @@ -407,6 +444,7 @@ class Program_weight_tensor_parameter_36: class Program_weight_tensor_parameter_37: name = "parameter_37" + original_name = "layer_norm_143.w_0" shape = [768] dtype = "float32" min_val = float("0.397277") @@ -418,6 +456,7 @@ class Program_weight_tensor_parameter_37: class Program_weight_tensor_parameter_38: name = "parameter_38" + original_name = "linear_423.b_0" shape = [768] dtype = "float32" min_val = float("-1.62406") @@ -429,6 +468,7 @@ class Program_weight_tensor_parameter_38: class Program_weight_tensor_parameter_39: name = "parameter_39" + original_name = "linear_423.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-0.820349") @@ -440,6 +480,7 @@ class Program_weight_tensor_parameter_39: class Program_weight_tensor_parameter_40: name = "parameter_40" + original_name = "linear_422.b_0" shape = [3072] dtype = "float32" min_val = float("-1.35991") @@ -451,6 +492,7 @@ class Program_weight_tensor_parameter_40: class Program_weight_tensor_parameter_41: name = "parameter_41" + original_name = "linear_422.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.802423") @@ -462,6 +504,7 @@ class Program_weight_tensor_parameter_41: class Program_weight_tensor_parameter_42: name = "parameter_42" + original_name = "linear_421.b_0" shape = [768] dtype = "float32" min_val = float("-0.258821") @@ -473,6 +516,7 @@ class Program_weight_tensor_parameter_42: class Program_weight_tensor_parameter_43: name = "parameter_43" + original_name = "linear_421.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.266023") @@ -484,6 +528,7 @@ class Program_weight_tensor_parameter_43: class Program_weight_tensor_parameter_44: name = "parameter_44" + original_name = "linear_420.b_0" shape = [768] dtype = "float32" min_val = float("-0.855859") @@ -495,6 +540,7 @@ class Program_weight_tensor_parameter_44: class Program_weight_tensor_parameter_45: name = "parameter_45" + original_name = "linear_420.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.44141") @@ -506,6 +552,7 @@ class Program_weight_tensor_parameter_45: class Program_weight_tensor_parameter_46: name = "parameter_46" + original_name = "linear_419.b_0" shape = [768] dtype = "float32" min_val = float("-14.812") @@ -517,6 +564,7 @@ class Program_weight_tensor_parameter_46: class Program_weight_tensor_parameter_47: name = "parameter_47" + original_name = "linear_419.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.474379") @@ -528,6 +576,7 @@ class Program_weight_tensor_parameter_47: class Program_weight_tensor_parameter_48: name = "parameter_48" + original_name = "linear_418.b_0" shape = [768] dtype = "float32" min_val = float("-2.84249") @@ -539,6 +588,7 @@ class Program_weight_tensor_parameter_48: class Program_weight_tensor_parameter_49: name = "parameter_49" + original_name = "linear_418.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.430674") @@ -550,6 +600,7 @@ class Program_weight_tensor_parameter_49: class Program_weight_tensor_parameter_50: name = "parameter_50" + original_name = "layer_norm_142.b_0" shape = [768] dtype = "float32" min_val = float("-0.652498") @@ -561,6 +612,7 @@ class Program_weight_tensor_parameter_50: class Program_weight_tensor_parameter_51: name = "parameter_51" + original_name = "layer_norm_142.w_0" shape = [768] dtype = "float32" min_val = float("0.119963") @@ -572,6 +624,7 @@ class Program_weight_tensor_parameter_51: class Program_weight_tensor_parameter_52: name = "parameter_52" + original_name = "layer_norm_141.b_0" shape = [768] dtype = "float32" min_val = float("-7.68663") @@ -583,6 +636,7 @@ class Program_weight_tensor_parameter_52: class Program_weight_tensor_parameter_53: name = "parameter_53" + original_name = "layer_norm_141.w_0" shape = [768] dtype = "float32" min_val = float("0.472672") @@ -594,6 +648,7 @@ class Program_weight_tensor_parameter_53: class Program_weight_tensor_parameter_54: name = "parameter_54" + original_name = "linear_417.b_0" shape = [768] dtype = "float32" min_val = float("-0.919117") @@ -605,6 +660,7 @@ class Program_weight_tensor_parameter_54: class Program_weight_tensor_parameter_55: name = "parameter_55" + original_name = "linear_417.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-2.07938") @@ -616,6 +672,7 @@ class Program_weight_tensor_parameter_55: class Program_weight_tensor_parameter_56: name = "parameter_56" + original_name = "linear_416.b_0" shape = [3072] dtype = "float32" min_val = float("-1.36799") @@ -627,6 +684,7 @@ class Program_weight_tensor_parameter_56: class Program_weight_tensor_parameter_57: name = "parameter_57" + original_name = "linear_416.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.553749") @@ -638,6 +696,7 @@ class Program_weight_tensor_parameter_57: class Program_weight_tensor_parameter_58: name = "parameter_58" + original_name = "linear_415.b_0" shape = [768] dtype = "float32" min_val = float("-0.205047") @@ -649,6 +708,7 @@ class Program_weight_tensor_parameter_58: class Program_weight_tensor_parameter_59: name = "parameter_59" + original_name = "linear_415.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.294118") @@ -660,6 +720,7 @@ class Program_weight_tensor_parameter_59: class Program_weight_tensor_parameter_60: name = "parameter_60" + original_name = "linear_414.b_0" shape = [768] dtype = "float32" min_val = float("-0.623821") @@ -671,6 +732,7 @@ class Program_weight_tensor_parameter_60: class Program_weight_tensor_parameter_61: name = "parameter_61" + original_name = "linear_414.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.304894") @@ -682,6 +744,7 @@ class Program_weight_tensor_parameter_61: class Program_weight_tensor_parameter_62: name = "parameter_62" + original_name = "linear_413.b_0" shape = [768] dtype = "float32" min_val = float("-12.2843") @@ -693,6 +756,7 @@ class Program_weight_tensor_parameter_62: class Program_weight_tensor_parameter_63: name = "parameter_63" + original_name = "linear_413.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.558544") @@ -704,6 +768,7 @@ class Program_weight_tensor_parameter_63: class Program_weight_tensor_parameter_64: name = "parameter_64" + original_name = "linear_412.b_0" shape = [768] dtype = "float32" min_val = float("-2.97359") @@ -715,6 +780,7 @@ class Program_weight_tensor_parameter_64: class Program_weight_tensor_parameter_65: name = "parameter_65" + original_name = "linear_412.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.577896") @@ -726,6 +792,7 @@ class Program_weight_tensor_parameter_65: class Program_weight_tensor_parameter_66: name = "parameter_66" + original_name = "layer_norm_140.b_0" shape = [768] dtype = "float32" min_val = float("-0.641154") @@ -737,6 +804,7 @@ class Program_weight_tensor_parameter_66: class Program_weight_tensor_parameter_67: name = "parameter_67" + original_name = "layer_norm_140.w_0" shape = [768] dtype = "float32" min_val = float("0.132489") @@ -748,6 +816,7 @@ class Program_weight_tensor_parameter_67: class Program_weight_tensor_parameter_68: name = "parameter_68" + original_name = "layer_norm_139.b_0" shape = [768] dtype = "float32" min_val = float("-6.90348") @@ -759,6 +828,7 @@ class Program_weight_tensor_parameter_68: class Program_weight_tensor_parameter_69: name = "parameter_69" + original_name = "layer_norm_139.w_0" shape = [768] dtype = "float32" min_val = float("0.2361") @@ -770,6 +840,7 @@ class Program_weight_tensor_parameter_69: class Program_weight_tensor_parameter_70: name = "parameter_70" + original_name = "linear_411.b_0" shape = [768] dtype = "float32" min_val = float("-0.812096") @@ -781,6 +852,7 @@ class Program_weight_tensor_parameter_70: class Program_weight_tensor_parameter_71: name = "parameter_71" + original_name = "linear_411.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-1.77185") @@ -792,6 +864,7 @@ class Program_weight_tensor_parameter_71: class Program_weight_tensor_parameter_72: name = "parameter_72" + original_name = "linear_410.b_0" shape = [3072] dtype = "float32" min_val = float("-1.46166") @@ -803,6 +876,7 @@ class Program_weight_tensor_parameter_72: class Program_weight_tensor_parameter_73: name = "parameter_73" + original_name = "linear_410.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.835728") @@ -814,6 +888,7 @@ class Program_weight_tensor_parameter_73: class Program_weight_tensor_parameter_74: name = "parameter_74" + original_name = "linear_409.b_0" shape = [768] dtype = "float32" min_val = float("-0.289332") @@ -825,6 +900,7 @@ class Program_weight_tensor_parameter_74: class Program_weight_tensor_parameter_75: name = "parameter_75" + original_name = "linear_409.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.571138") @@ -836,6 +912,7 @@ class Program_weight_tensor_parameter_75: class Program_weight_tensor_parameter_76: name = "parameter_76" + original_name = "linear_408.b_0" shape = [768] dtype = "float32" min_val = float("-0.867828") @@ -847,6 +924,7 @@ class Program_weight_tensor_parameter_76: class Program_weight_tensor_parameter_77: name = "parameter_77" + original_name = "linear_408.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.276526") @@ -858,6 +936,7 @@ class Program_weight_tensor_parameter_77: class Program_weight_tensor_parameter_78: name = "parameter_78" + original_name = "linear_407.b_0" shape = [768] dtype = "float32" min_val = float("-7.56818") @@ -869,6 +948,7 @@ class Program_weight_tensor_parameter_78: class Program_weight_tensor_parameter_79: name = "parameter_79" + original_name = "linear_407.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.61523") @@ -880,6 +960,7 @@ class Program_weight_tensor_parameter_79: class Program_weight_tensor_parameter_80: name = "parameter_80" + original_name = "linear_406.b_0" shape = [768] dtype = "float32" min_val = float("-2.3071") @@ -891,6 +972,7 @@ class Program_weight_tensor_parameter_80: class Program_weight_tensor_parameter_81: name = "parameter_81" + original_name = "linear_406.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.573666") @@ -902,6 +984,7 @@ class Program_weight_tensor_parameter_81: class Program_weight_tensor_parameter_82: name = "parameter_82" + original_name = "layer_norm_138.b_0" shape = [768] dtype = "float32" min_val = float("-0.921464") @@ -913,6 +996,7 @@ class Program_weight_tensor_parameter_82: class Program_weight_tensor_parameter_83: name = "parameter_83" + original_name = "layer_norm_138.w_0" shape = [768] dtype = "float32" min_val = float("0.106892") @@ -924,6 +1008,7 @@ class Program_weight_tensor_parameter_83: class Program_weight_tensor_parameter_84: name = "parameter_84" + original_name = "layer_norm_137.b_0" shape = [768] dtype = "float32" min_val = float("-7.86422") @@ -935,6 +1020,7 @@ class Program_weight_tensor_parameter_84: class Program_weight_tensor_parameter_85: name = "parameter_85" + original_name = "layer_norm_137.w_0" shape = [768] dtype = "float32" min_val = float("0.227837") @@ -946,6 +1032,7 @@ class Program_weight_tensor_parameter_85: class Program_weight_tensor_parameter_86: name = "parameter_86" + original_name = "linear_405.b_0" shape = [768] dtype = "float32" min_val = float("-0.669091") @@ -957,6 +1044,7 @@ class Program_weight_tensor_parameter_86: class Program_weight_tensor_parameter_87: name = "parameter_87" + original_name = "linear_405.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-2.03318") @@ -968,6 +1056,7 @@ class Program_weight_tensor_parameter_87: class Program_weight_tensor_parameter_88: name = "parameter_88" + original_name = "linear_404.b_0" shape = [3072] dtype = "float32" min_val = float("-1.15983") @@ -979,6 +1068,7 @@ class Program_weight_tensor_parameter_88: class Program_weight_tensor_parameter_89: name = "parameter_89" + original_name = "linear_404.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.697454") @@ -990,6 +1080,7 @@ class Program_weight_tensor_parameter_89: class Program_weight_tensor_parameter_90: name = "parameter_90" + original_name = "linear_403.b_0" shape = [768] dtype = "float32" min_val = float("-0.213032") @@ -1001,6 +1092,7 @@ class Program_weight_tensor_parameter_90: class Program_weight_tensor_parameter_91: name = "parameter_91" + original_name = "linear_403.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.610191") @@ -1012,6 +1104,7 @@ class Program_weight_tensor_parameter_91: class Program_weight_tensor_parameter_92: name = "parameter_92" + original_name = "linear_402.b_0" shape = [768] dtype = "float32" min_val = float("-0.5942") @@ -1023,6 +1116,7 @@ class Program_weight_tensor_parameter_92: class Program_weight_tensor_parameter_93: name = "parameter_93" + original_name = "linear_402.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.318666") @@ -1034,6 +1128,7 @@ class Program_weight_tensor_parameter_93: class Program_weight_tensor_parameter_94: name = "parameter_94" + original_name = "linear_401.b_0" shape = [768] dtype = "float32" min_val = float("-4.98363") @@ -1045,6 +1140,7 @@ class Program_weight_tensor_parameter_94: class Program_weight_tensor_parameter_95: name = "parameter_95" + original_name = "linear_401.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.518714") @@ -1056,6 +1152,7 @@ class Program_weight_tensor_parameter_95: class Program_weight_tensor_parameter_96: name = "parameter_96" + original_name = "linear_400.b_0" shape = [768] dtype = "float32" min_val = float("-3.01381") @@ -1067,6 +1164,7 @@ class Program_weight_tensor_parameter_96: class Program_weight_tensor_parameter_97: name = "parameter_97" + original_name = "linear_400.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.567549") @@ -1078,6 +1176,7 @@ class Program_weight_tensor_parameter_97: class Program_weight_tensor_parameter_98: name = "parameter_98" + original_name = "layer_norm_136.b_0" shape = [768] dtype = "float32" min_val = float("-1.18627") @@ -1089,6 +1188,7 @@ class Program_weight_tensor_parameter_98: class Program_weight_tensor_parameter_99: name = "parameter_99" + original_name = "layer_norm_136.w_0" shape = [768] dtype = "float32" min_val = float("0.106247") @@ -1100,6 +1200,7 @@ class Program_weight_tensor_parameter_99: class Program_weight_tensor_parameter_100: name = "parameter_100" + original_name = "layer_norm_135.b_0" shape = [768] dtype = "float32" min_val = float("-8.56548") @@ -1111,6 +1212,7 @@ class Program_weight_tensor_parameter_100: class Program_weight_tensor_parameter_101: name = "parameter_101" + original_name = "layer_norm_135.w_0" shape = [768] dtype = "float32" min_val = float("0.075071") @@ -1122,6 +1224,7 @@ class Program_weight_tensor_parameter_101: class Program_weight_tensor_parameter_102: name = "parameter_102" + original_name = "linear_399.b_0" shape = [768] dtype = "float32" min_val = float("-0.549237") @@ -1133,6 +1236,7 @@ class Program_weight_tensor_parameter_102: class Program_weight_tensor_parameter_103: name = "parameter_103" + original_name = "linear_399.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-1.4705") @@ -1144,6 +1248,7 @@ class Program_weight_tensor_parameter_103: class Program_weight_tensor_parameter_104: name = "parameter_104" + original_name = "linear_398.b_0" shape = [3072] dtype = "float32" min_val = float("-1.05818") @@ -1155,6 +1260,7 @@ class Program_weight_tensor_parameter_104: class Program_weight_tensor_parameter_105: name = "parameter_105" + original_name = "linear_398.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-1.07122") @@ -1166,6 +1272,7 @@ class Program_weight_tensor_parameter_105: class Program_weight_tensor_parameter_106: name = "parameter_106" + original_name = "linear_397.b_0" shape = [768] dtype = "float32" min_val = float("-0.172784") @@ -1177,6 +1284,7 @@ class Program_weight_tensor_parameter_106: class Program_weight_tensor_parameter_107: name = "parameter_107" + original_name = "linear_397.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.657752") @@ -1188,6 +1296,7 @@ class Program_weight_tensor_parameter_107: class Program_weight_tensor_parameter_108: name = "parameter_108" + original_name = "linear_396.b_0" shape = [768] dtype = "float32" min_val = float("-0.78249") @@ -1199,6 +1308,7 @@ class Program_weight_tensor_parameter_108: class Program_weight_tensor_parameter_109: name = "parameter_109" + original_name = "linear_396.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.565074") @@ -1210,6 +1320,7 @@ class Program_weight_tensor_parameter_109: class Program_weight_tensor_parameter_110: name = "parameter_110" + original_name = "linear_395.b_0" shape = [768] dtype = "float32" min_val = float("-3.08743") @@ -1221,6 +1332,7 @@ class Program_weight_tensor_parameter_110: class Program_weight_tensor_parameter_111: name = "parameter_111" + original_name = "linear_395.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.527467") @@ -1232,6 +1344,7 @@ class Program_weight_tensor_parameter_111: class Program_weight_tensor_parameter_112: name = "parameter_112" + original_name = "linear_394.b_0" shape = [768] dtype = "float32" min_val = float("-2.46207") @@ -1243,6 +1356,7 @@ class Program_weight_tensor_parameter_112: class Program_weight_tensor_parameter_113: name = "parameter_113" + original_name = "linear_394.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.427855") @@ -1254,6 +1368,7 @@ class Program_weight_tensor_parameter_113: class Program_weight_tensor_parameter_114: name = "parameter_114" + original_name = "layer_norm_134.b_0" shape = [768] dtype = "float32" min_val = float("-1.23216") @@ -1265,6 +1380,7 @@ class Program_weight_tensor_parameter_114: class Program_weight_tensor_parameter_115: name = "parameter_115" + original_name = "layer_norm_134.w_0" shape = [768] dtype = "float32" min_val = float("0.0914138") @@ -1276,6 +1392,7 @@ class Program_weight_tensor_parameter_115: class Program_weight_tensor_parameter_116: name = "parameter_116" + original_name = "layer_norm_133.b_0" shape = [768] dtype = "float32" min_val = float("-7.65946") @@ -1287,6 +1404,7 @@ class Program_weight_tensor_parameter_116: class Program_weight_tensor_parameter_117: name = "parameter_117" + original_name = "layer_norm_133.w_0" shape = [768] dtype = "float32" min_val = float("0.390751") @@ -1298,6 +1416,7 @@ class Program_weight_tensor_parameter_117: class Program_weight_tensor_parameter_118: name = "parameter_118" + original_name = "linear_393.b_0" shape = [768] dtype = "float32" min_val = float("-0.650577") @@ -1309,6 +1428,7 @@ class Program_weight_tensor_parameter_118: class Program_weight_tensor_parameter_119: name = "parameter_119" + original_name = "linear_393.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-2.77351") @@ -1320,6 +1440,7 @@ class Program_weight_tensor_parameter_119: class Program_weight_tensor_parameter_120: name = "parameter_120" + original_name = "linear_392.b_0" shape = [3072] dtype = "float32" min_val = float("-0.98651") @@ -1331,6 +1452,7 @@ class Program_weight_tensor_parameter_120: class Program_weight_tensor_parameter_121: name = "parameter_121" + original_name = "linear_392.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.785151") @@ -1342,6 +1464,7 @@ class Program_weight_tensor_parameter_121: class Program_weight_tensor_parameter_122: name = "parameter_122" + original_name = "linear_391.b_0" shape = [768] dtype = "float32" min_val = float("-0.319532") @@ -1353,6 +1476,7 @@ class Program_weight_tensor_parameter_122: class Program_weight_tensor_parameter_123: name = "parameter_123" + original_name = "linear_391.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.365041") @@ -1364,6 +1488,7 @@ class Program_weight_tensor_parameter_123: class Program_weight_tensor_parameter_124: name = "parameter_124" + original_name = "linear_390.b_0" shape = [768] dtype = "float32" min_val = float("-0.505784") @@ -1375,6 +1500,7 @@ class Program_weight_tensor_parameter_124: class Program_weight_tensor_parameter_125: name = "parameter_125" + original_name = "linear_390.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.584495") @@ -1386,6 +1512,7 @@ class Program_weight_tensor_parameter_125: class Program_weight_tensor_parameter_126: name = "parameter_126" + original_name = "linear_389.b_0" shape = [768] dtype = "float32" min_val = float("-2.66824") @@ -1397,6 +1524,7 @@ class Program_weight_tensor_parameter_126: class Program_weight_tensor_parameter_127: name = "parameter_127" + original_name = "linear_389.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.570038") @@ -1408,6 +1536,7 @@ class Program_weight_tensor_parameter_127: class Program_weight_tensor_parameter_128: name = "parameter_128" + original_name = "linear_388.b_0" shape = [768] dtype = "float32" min_val = float("-3.30395") @@ -1419,6 +1548,7 @@ class Program_weight_tensor_parameter_128: class Program_weight_tensor_parameter_129: name = "parameter_129" + original_name = "linear_388.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.568317") @@ -1430,6 +1560,7 @@ class Program_weight_tensor_parameter_129: class Program_weight_tensor_parameter_130: name = "parameter_130" + original_name = "layer_norm_132.b_0" shape = [768] dtype = "float32" min_val = float("-1.0551") @@ -1441,6 +1572,7 @@ class Program_weight_tensor_parameter_130: class Program_weight_tensor_parameter_131: name = "parameter_131" + original_name = "layer_norm_132.w_0" shape = [768] dtype = "float32" min_val = float("0.215993") @@ -1452,6 +1584,7 @@ class Program_weight_tensor_parameter_131: class Program_weight_tensor_parameter_132: name = "parameter_132" + original_name = "layer_norm_131.b_0" shape = [768] dtype = "float32" min_val = float("-6.72996") @@ -1463,6 +1596,7 @@ class Program_weight_tensor_parameter_132: class Program_weight_tensor_parameter_133: name = "parameter_133" + original_name = "layer_norm_131.w_0" shape = [768] dtype = "float32" min_val = float("0.509023") @@ -1474,6 +1608,7 @@ class Program_weight_tensor_parameter_133: class Program_weight_tensor_parameter_134: name = "parameter_134" + original_name = "linear_387.b_0" shape = [768] dtype = "float32" min_val = float("-0.566459") @@ -1485,6 +1620,7 @@ class Program_weight_tensor_parameter_134: class Program_weight_tensor_parameter_135: name = "parameter_135" + original_name = "linear_387.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-6.51375") @@ -1496,6 +1632,7 @@ class Program_weight_tensor_parameter_135: class Program_weight_tensor_parameter_136: name = "parameter_136" + original_name = "linear_386.b_0" shape = [3072] dtype = "float32" min_val = float("-1.28077") @@ -1507,6 +1644,7 @@ class Program_weight_tensor_parameter_136: class Program_weight_tensor_parameter_137: name = "parameter_137" + original_name = "linear_386.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.728932") @@ -1518,6 +1656,7 @@ class Program_weight_tensor_parameter_137: class Program_weight_tensor_parameter_138: name = "parameter_138" + original_name = "linear_385.b_0" shape = [768] dtype = "float32" min_val = float("-0.294423") @@ -1529,6 +1668,7 @@ class Program_weight_tensor_parameter_138: class Program_weight_tensor_parameter_139: name = "parameter_139" + original_name = "linear_385.w_0" shape = [768, 768] dtype = "float32" min_val = float("-1.08852") @@ -1540,6 +1680,7 @@ class Program_weight_tensor_parameter_139: class Program_weight_tensor_parameter_140: name = "parameter_140" + original_name = "linear_384.b_0" shape = [768] dtype = "float32" min_val = float("-0.822656") @@ -1551,6 +1692,7 @@ class Program_weight_tensor_parameter_140: class Program_weight_tensor_parameter_141: name = "parameter_141" + original_name = "linear_384.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.461696") @@ -1562,6 +1704,7 @@ class Program_weight_tensor_parameter_141: class Program_weight_tensor_parameter_142: name = "parameter_142" + original_name = "linear_383.b_0" shape = [768] dtype = "float32" min_val = float("-1.84711") @@ -1573,6 +1716,7 @@ class Program_weight_tensor_parameter_142: class Program_weight_tensor_parameter_143: name = "parameter_143" + original_name = "linear_383.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.559661") @@ -1584,6 +1728,7 @@ class Program_weight_tensor_parameter_143: class Program_weight_tensor_parameter_144: name = "parameter_144" + original_name = "linear_382.b_0" shape = [768] dtype = "float32" min_val = float("-2.74563") @@ -1595,6 +1740,7 @@ class Program_weight_tensor_parameter_144: class Program_weight_tensor_parameter_145: name = "parameter_145" + original_name = "linear_382.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.60494") @@ -1606,6 +1752,7 @@ class Program_weight_tensor_parameter_145: class Program_weight_tensor_parameter_146: name = "parameter_146" + original_name = "layer_norm_130.b_0" shape = [768] dtype = "float32" min_val = float("-1.23452") @@ -1617,6 +1764,7 @@ class Program_weight_tensor_parameter_146: class Program_weight_tensor_parameter_147: name = "parameter_147" + original_name = "layer_norm_130.w_0" shape = [768] dtype = "float32" min_val = float("0.144665") @@ -1628,6 +1776,7 @@ class Program_weight_tensor_parameter_147: class Program_weight_tensor_parameter_148: name = "parameter_148" + original_name = "layer_norm_129.b_0" shape = [768] dtype = "float32" min_val = float("-6.00353") @@ -1639,6 +1788,7 @@ class Program_weight_tensor_parameter_148: class Program_weight_tensor_parameter_149: name = "parameter_149" + original_name = "layer_norm_129.w_0" shape = [768] dtype = "float32" min_val = float("0.657183") @@ -1650,6 +1800,7 @@ class Program_weight_tensor_parameter_149: class Program_weight_tensor_parameter_150: name = "parameter_150" + original_name = "linear_381.b_0" shape = [768] dtype = "float32" min_val = float("-0.550122") @@ -1661,6 +1812,7 @@ class Program_weight_tensor_parameter_150: class Program_weight_tensor_parameter_151: name = "parameter_151" + original_name = "linear_381.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-1.71372") @@ -1672,6 +1824,7 @@ class Program_weight_tensor_parameter_151: class Program_weight_tensor_parameter_152: name = "parameter_152" + original_name = "linear_380.b_0" shape = [3072] dtype = "float32" min_val = float("-0.659757") @@ -1683,6 +1836,7 @@ class Program_weight_tensor_parameter_152: class Program_weight_tensor_parameter_153: name = "parameter_153" + original_name = "linear_380.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.588486") @@ -1694,6 +1848,7 @@ class Program_weight_tensor_parameter_153: class Program_weight_tensor_parameter_154: name = "parameter_154" + original_name = "linear_379.b_0" shape = [768] dtype = "float32" min_val = float("-0.271344") @@ -1705,6 +1860,7 @@ class Program_weight_tensor_parameter_154: class Program_weight_tensor_parameter_155: name = "parameter_155" + original_name = "linear_379.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.538396") @@ -1716,6 +1872,7 @@ class Program_weight_tensor_parameter_155: class Program_weight_tensor_parameter_156: name = "parameter_156" + original_name = "linear_378.b_0" shape = [768] dtype = "float32" min_val = float("-0.733249") @@ -1727,6 +1884,7 @@ class Program_weight_tensor_parameter_156: class Program_weight_tensor_parameter_157: name = "parameter_157" + original_name = "linear_378.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.337595") @@ -1738,6 +1896,7 @@ class Program_weight_tensor_parameter_157: class Program_weight_tensor_parameter_158: name = "parameter_158" + original_name = "linear_377.b_0" shape = [768] dtype = "float32" min_val = float("-2.65876") @@ -1749,6 +1908,7 @@ class Program_weight_tensor_parameter_158: class Program_weight_tensor_parameter_159: name = "parameter_159" + original_name = "linear_377.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.580315") @@ -1760,6 +1920,7 @@ class Program_weight_tensor_parameter_159: class Program_weight_tensor_parameter_160: name = "parameter_160" + original_name = "linear_376.b_0" shape = [768] dtype = "float32" min_val = float("-3.11082") @@ -1771,6 +1932,7 @@ class Program_weight_tensor_parameter_160: class Program_weight_tensor_parameter_161: name = "parameter_161" + original_name = "linear_376.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.39365") @@ -1782,6 +1944,7 @@ class Program_weight_tensor_parameter_161: class Program_weight_tensor_parameter_162: name = "parameter_162" + original_name = "layer_norm_128.b_0" shape = [768] dtype = "float32" min_val = float("-0.74969") @@ -1793,6 +1956,7 @@ class Program_weight_tensor_parameter_162: class Program_weight_tensor_parameter_163: name = "parameter_163" + original_name = "layer_norm_128.w_0" shape = [768] dtype = "float32" min_val = float("0.149365") @@ -1804,6 +1968,7 @@ class Program_weight_tensor_parameter_163: class Program_weight_tensor_parameter_164: name = "parameter_164" + original_name = "layer_norm_127.b_0" shape = [768] dtype = "float32" min_val = float("-5.2366") @@ -1815,6 +1980,7 @@ class Program_weight_tensor_parameter_164: class Program_weight_tensor_parameter_165: name = "parameter_165" + original_name = "layer_norm_127.w_0" shape = [768] dtype = "float32" min_val = float("0.599738") @@ -1826,6 +1992,7 @@ class Program_weight_tensor_parameter_165: class Program_weight_tensor_parameter_166: name = "parameter_166" + original_name = "linear_375.b_0" shape = [768] dtype = "float32" min_val = float("-0.573859") @@ -1837,6 +2004,7 @@ class Program_weight_tensor_parameter_166: class Program_weight_tensor_parameter_167: name = "parameter_167" + original_name = "linear_375.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-5.28226") @@ -1848,6 +2016,7 @@ class Program_weight_tensor_parameter_167: class Program_weight_tensor_parameter_168: name = "parameter_168" + original_name = "linear_374.b_0" shape = [3072] dtype = "float32" min_val = float("-0.903374") @@ -1859,6 +2028,7 @@ class Program_weight_tensor_parameter_168: class Program_weight_tensor_parameter_169: name = "parameter_169" + original_name = "linear_374.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.608393") @@ -1870,6 +2040,7 @@ class Program_weight_tensor_parameter_169: class Program_weight_tensor_parameter_170: name = "parameter_170" + original_name = "linear_373.b_0" shape = [768] dtype = "float32" min_val = float("-0.291032") @@ -1881,6 +2052,7 @@ class Program_weight_tensor_parameter_170: class Program_weight_tensor_parameter_171: name = "parameter_171" + original_name = "linear_373.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.46967") @@ -1892,6 +2064,7 @@ class Program_weight_tensor_parameter_171: class Program_weight_tensor_parameter_172: name = "parameter_172" + original_name = "linear_372.b_0" shape = [768] dtype = "float32" min_val = float("-0.388822") @@ -1903,6 +2076,7 @@ class Program_weight_tensor_parameter_172: class Program_weight_tensor_parameter_173: name = "parameter_173" + original_name = "linear_372.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.436494") @@ -1914,6 +2088,7 @@ class Program_weight_tensor_parameter_173: class Program_weight_tensor_parameter_174: name = "parameter_174" + original_name = "linear_371.b_0" shape = [768] dtype = "float32" min_val = float("-3.12397") @@ -1925,6 +2100,7 @@ class Program_weight_tensor_parameter_174: class Program_weight_tensor_parameter_175: name = "parameter_175" + original_name = "linear_371.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.553966") @@ -1936,6 +2112,7 @@ class Program_weight_tensor_parameter_175: class Program_weight_tensor_parameter_176: name = "parameter_176" + original_name = "linear_370.b_0" shape = [768] dtype = "float32" min_val = float("-2.58118") @@ -1947,6 +2124,7 @@ class Program_weight_tensor_parameter_176: class Program_weight_tensor_parameter_177: name = "parameter_177" + original_name = "linear_370.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.453115") @@ -1958,6 +2136,7 @@ class Program_weight_tensor_parameter_177: class Program_weight_tensor_parameter_178: name = "parameter_178" + original_name = "layer_norm_126.b_0" shape = [768] dtype = "float32" min_val = float("-0.58619") @@ -1969,6 +2148,7 @@ class Program_weight_tensor_parameter_178: class Program_weight_tensor_parameter_179: name = "parameter_179" + original_name = "layer_norm_126.w_0" shape = [768] dtype = "float32" min_val = float("0.0981678") @@ -1980,6 +2160,7 @@ class Program_weight_tensor_parameter_179: class Program_weight_tensor_parameter_180: name = "parameter_180" + original_name = "layer_norm_125.b_0" shape = [768] dtype = "float32" min_val = float("-7.9122") @@ -1991,6 +2172,7 @@ class Program_weight_tensor_parameter_180: class Program_weight_tensor_parameter_181: name = "parameter_181" + original_name = "layer_norm_125.w_0" shape = [768] dtype = "float32" min_val = float("0.342963") @@ -2002,6 +2184,7 @@ class Program_weight_tensor_parameter_181: class Program_weight_tensor_parameter_182: name = "parameter_182" + original_name = "linear_369.b_0" shape = [768] dtype = "float32" min_val = float("-0.565859") @@ -2013,6 +2196,7 @@ class Program_weight_tensor_parameter_182: class Program_weight_tensor_parameter_183: name = "parameter_183" + original_name = "linear_369.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-5.12289") @@ -2024,6 +2208,7 @@ class Program_weight_tensor_parameter_183: class Program_weight_tensor_parameter_184: name = "parameter_184" + original_name = "linear_368.b_0" shape = [3072] dtype = "float32" min_val = float("-1.4118") @@ -2035,6 +2220,7 @@ class Program_weight_tensor_parameter_184: class Program_weight_tensor_parameter_185: name = "parameter_185" + original_name = "linear_368.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.645851") @@ -2046,6 +2232,7 @@ class Program_weight_tensor_parameter_185: class Program_weight_tensor_parameter_186: name = "parameter_186" + original_name = "linear_367.b_0" shape = [768] dtype = "float32" min_val = float("-0.612196") @@ -2057,6 +2244,7 @@ class Program_weight_tensor_parameter_186: class Program_weight_tensor_parameter_187: name = "parameter_187" + original_name = "linear_367.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.446235") @@ -2068,6 +2256,7 @@ class Program_weight_tensor_parameter_187: class Program_weight_tensor_parameter_188: name = "parameter_188" + original_name = "linear_366.b_0" shape = [768] dtype = "float32" min_val = float("-1.39206") @@ -2079,6 +2268,7 @@ class Program_weight_tensor_parameter_188: class Program_weight_tensor_parameter_189: name = "parameter_189" + original_name = "linear_366.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.321918") @@ -2090,6 +2280,7 @@ class Program_weight_tensor_parameter_189: class Program_weight_tensor_parameter_190: name = "parameter_190" + original_name = "linear_365.b_0" shape = [768] dtype = "float32" min_val = float("-1.59772") @@ -2101,6 +2292,7 @@ class Program_weight_tensor_parameter_190: class Program_weight_tensor_parameter_191: name = "parameter_191" + original_name = "linear_365.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.534877") @@ -2112,6 +2304,7 @@ class Program_weight_tensor_parameter_191: class Program_weight_tensor_parameter_192: name = "parameter_192" + original_name = "linear_364.b_0" shape = [768] dtype = "float32" min_val = float("-2.82965") @@ -2123,6 +2316,7 @@ class Program_weight_tensor_parameter_192: class Program_weight_tensor_parameter_193: name = "parameter_193" + original_name = "linear_364.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.564355") @@ -2134,6 +2328,7 @@ class Program_weight_tensor_parameter_193: class Program_weight_tensor_parameter_194: name = "parameter_194" + original_name = "layer_norm_124.b_0" shape = [768] dtype = "float32" min_val = float("-0.454256") @@ -2145,6 +2340,7 @@ class Program_weight_tensor_parameter_194: class Program_weight_tensor_parameter_195: name = "parameter_195" + original_name = "layer_norm_124.w_0" shape = [768] dtype = "float32" min_val = float("0.108537") @@ -2156,6 +2352,7 @@ class Program_weight_tensor_parameter_195: class Program_weight_tensor_parameter_196: name = "parameter_196" + original_name = "embedding_15.w_0" shape = [4, 768] dtype = "float32" min_val = float("-0.202293") @@ -2167,6 +2364,7 @@ class Program_weight_tensor_parameter_196: class Program_weight_tensor_parameter_197: name = "parameter_197" + original_name = "embedding_14.w_0" shape = [513, 768] dtype = "float32" min_val = float("-0.203972") @@ -2178,6 +2376,7 @@ class Program_weight_tensor_parameter_197: class Program_weight_tensor_parameter_198: name = "parameter_198" + original_name = "embedding_13.w_0" shape = [18000, 768] dtype = "float32" min_val = float("-1.07755") diff --git a/paddle_samples/PaddleNLP/ernie-2.0-large-zh/input_meta.py b/paddle_samples/PaddleNLP/ernie-2.0-large-zh/input_meta.py index 97eb8a799..a4123e481 100644 --- a/paddle_samples/PaddleNLP/ernie-2.0-large-zh/input_meta.py +++ b/paddle_samples/PaddleNLP/ernie-2.0-large-zh/input_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_data_0: name = "data_0" + original_name = "input_ids" shape = [1, 11] dtype = "int64" data = [1, 811, 1257, 175, 29, 502, 130, 706, 3619, 12046, 2] @@ -7,6 +8,7 @@ class Program_weight_tensor_data_0: class Program_weight_tensor_data_1: name = "data_1" + original_name = "token_type_ids" shape = [1, 11] dtype = "int64" data = [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0] diff --git a/paddle_samples/PaddleNLP/ernie-2.0-large-zh/model.py b/paddle_samples/PaddleNLP/ernie-2.0-large-zh/model.py index 892e22c63..00fe7c714 100644 --- a/paddle_samples/PaddleNLP/ernie-2.0-large-zh/model.py +++ b/paddle_samples/PaddleNLP/ernie-2.0-large-zh/model.py @@ -457,10 +457,11 @@ def forward( # pd_op.embedding: (1x11x1024xf32) <- (1x11xi64, 512x1024xf32) embedding_1 = paddle._C_ops.embedding(subtract_0, parameter_389, -1, False) - del parameter_389 + del parameter_389, subtract_0 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_0 = paddle._C_ops.add(embedding_0, embedding_1) + del embedding_0, embedding_1 # pd_op.embedding: (1x11x1024xf32) <- (1x11xi64, 4x1024xf32) embedding_2 = paddle._C_ops.embedding(data_1, parameter_388, -1, False) @@ -468,6 +469,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_1 = paddle._C_ops.add(add_0, embedding_2) + del add_0, embedding_2 # pd_op.layer_norm: (1x11x1024xf32, 1x11xf32, 1x11xf32) <- (1x11x1024xf32, 1024xf32, 1024xf32) layer_norm_0, layer_norm_1, layer_norm_2 = (lambda x, f: f(x))( @@ -476,233 +478,17 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_386, parameter_387 + del add_1, parameter_386, parameter_387 # pd_op.full: (1xf32) <- () full_4 = paddle._C_ops.full( [1], float("0.1"), paddle.float32, paddle.core.CPUPlace() ) - # pd_op.assign: (1xf32) <- (1xf32) - assign_0 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_1 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_2 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_3 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_4 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_5 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_6 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_7 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_8 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_9 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_10 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_11 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_12 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_13 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_14 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_15 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_16 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_17 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_18 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_19 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_20 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_21 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_22 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_23 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_24 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_25 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_26 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_27 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_28 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_29 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_30 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_31 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_32 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_33 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_34 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_35 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_36 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_37 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_38 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_39 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_40 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_41 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_42 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_43 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_44 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_45 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_46 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_47 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_48 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_49 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_50 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_51 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_52 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_53 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_54 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_55 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_56 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_57 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_58 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_59 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_60 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_61 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_62 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_63 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_64 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_65 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_66 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_67 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_68 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_69 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_70 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_71 = full_4 - # pd_op.dropout: (1x11x1024xf32, 1x11x1024xui8) <- (1x11x1024xf32, None, 1xf32) dropout_0, dropout_1 = (lambda x, f: f(x))( paddle._C_ops.dropout( - layer_norm_0, None, full_4, False, "upscale_in_train", 0, False + layer_norm_0, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -714,13 +500,14 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_2 = paddle._C_ops.add(matmul_0, parameter_384) - del parameter_384 + del matmul_0, parameter_384 # pd_op.full_int_array: (4xi64) <- () full_int_array_1 = [0, 0, 16, 64] # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_0 = paddle._C_ops.reshape(add_2, full_int_array_1) + del add_2 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_0 = paddle._C_ops.transpose(reshape_0, [0, 2, 1, 3]) @@ -732,7 +519,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_3 = paddle._C_ops.add(matmul_1, parameter_382) - del parameter_382 + del matmul_1, parameter_382 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_2 = paddle._C_ops.matmul(dropout_0, parameter_381, False, False) @@ -740,10 +527,11 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_4 = paddle._C_ops.add(matmul_2, parameter_380) - del parameter_380 + del matmul_2, parameter_380 # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_1 = paddle._C_ops.reshape(add_3, full_int_array_1) + del add_3 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_1 = paddle._C_ops.transpose(reshape_1, [0, 2, 1, 3]) @@ -751,6 +539,7 @@ def forward( # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_2 = paddle._C_ops.reshape(add_4, full_int_array_1) + del add_4 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_2 = paddle._C_ops.transpose(reshape_2, [0, 2, 1, 3]) @@ -761,84 +550,17 @@ def forward( [1], float("0.125"), paddle.float32, paddle.core.CPUPlace() ) - # pd_op.assign: (1xf32) <- (1xf32) - assign_72 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_73 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_74 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_75 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_76 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_77 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_78 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_79 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_80 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_81 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_82 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_83 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_84 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_85 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_86 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_87 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_88 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_89 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_90 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_91 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_92 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_93 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_94 = full_5 - # pd_op.scale: (1x16x11x64xf32) <- (1x16x11x64xf32, 1xf32) scale_1 = paddle._C_ops.scale(transpose_0, full_5, float("0"), True) del transpose_0 # pd_op.matmul: (1x16x11x11xf32) <- (1x16x11x64xf32, 1x16x11x64xf32) matmul_3 = paddle._C_ops.matmul(scale_1, transpose_1, False, True) + del scale_1, transpose_1 # pd_op.add: (1x16x11x11xf32) <- (1x16x11x11xf32, 1x1x1x11xf32) add_5 = paddle._C_ops.add(matmul_3, unsqueeze_0) + del matmul_3 # pd_op.softmax: (1x16x11x11xf32) <- (1x16x11x11xf32) softmax_0 = paddle._C_ops.softmax(add_5, -1) @@ -847,13 +569,15 @@ def forward( # pd_op.dropout: (1x16x11x11xf32, 1x16x11x11xui8) <- (1x16x11x11xf32, None, 1xf32) dropout_2, dropout_3 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_0, None, full_4, False, "upscale_in_train", 0, False + softmax_0, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_0 # pd_op.matmul: (1x16x11x64xf32) <- (1x16x11x11xf32, 1x16x11x64xf32) matmul_4 = paddle._C_ops.matmul(dropout_2, transpose_2, False, False) + del dropout_2, transpose_2 # pd_op.transpose: (1x11x16x64xf32) <- (1x16x11x64xf32) transpose_3 = paddle._C_ops.transpose(matmul_4, [0, 2, 1, 3]) @@ -864,19 +588,20 @@ def forward( # pd_op.reshape: (1x11x1024xf32) <- (1x11x16x64xf32, 3xi64) reshape_3 = paddle._C_ops.reshape(transpose_3, full_int_array_2) + del transpose_3 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_5 = paddle._C_ops.matmul(reshape_3, parameter_379, False, False) - del parameter_379 + del parameter_379, reshape_3 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_6 = paddle._C_ops.add(matmul_5, parameter_378) - del parameter_378 + del matmul_5, parameter_378 # pd_op.dropout: (1x11x1024xf32, 1x11x1024xui8) <- (1x11x1024xf32, None, 1xf32) dropout_4, dropout_5 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_6, None, full_4, False, "upscale_in_train", 0, False + add_6, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -884,6 +609,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_7 = paddle._C_ops.add(dropout_0, dropout_4) + del dropout_0, dropout_4 # pd_op.layer_norm: (1x11x1024xf32, 1x11xf32, 1x11xf32) <- (1x11x1024xf32, 1024xf32, 1024xf32) layer_norm_3, layer_norm_4, layer_norm_5 = (lambda x, f: f(x))( @@ -892,7 +618,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_372, parameter_373 + del add_7, parameter_372, parameter_373 # pd_op.matmul: (1x11x4096xf32) <- (1x11x1024xf32, 1024x4096xf32) matmul_6 = paddle._C_ops.matmul(layer_norm_3, parameter_377, False, False) @@ -900,7 +626,7 @@ def forward( # pd_op.add: (1x11x4096xf32) <- (1x11x4096xf32, 4096xf32) add_8 = paddle._C_ops.add(matmul_6, parameter_376) - del parameter_376 + del matmul_6, parameter_376 # pd_op.relu: (1x11x4096xf32) <- (1x11x4096xf32) relu_0 = paddle._C_ops.relu(add_8) @@ -908,16 +634,16 @@ def forward( # pd_op.matmul: (1x11x1024xf32) <- (1x11x4096xf32, 4096x1024xf32) matmul_7 = paddle._C_ops.matmul(relu_0, parameter_375, False, False) - del parameter_375 + del parameter_375, relu_0 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_9 = paddle._C_ops.add(matmul_7, parameter_374) - del parameter_374 + del matmul_7, parameter_374 # pd_op.dropout: (1x11x1024xf32, 1x11x1024xui8) <- (1x11x1024xf32, None, 1xf32) dropout_6, dropout_7 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_9, None, full_4, False, "upscale_in_train", 0, False + add_9, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -925,6 +651,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_10 = paddle._C_ops.add(layer_norm_3, dropout_6) + del dropout_6, layer_norm_3 # pd_op.layer_norm: (1x11x1024xf32, 1x11xf32, 1x11xf32) <- (1x11x1024xf32, 1024xf32, 1024xf32) layer_norm_6, layer_norm_7, layer_norm_8 = (lambda x, f: f(x))( @@ -933,7 +660,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_370, parameter_371 + del add_10, parameter_370, parameter_371 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_8 = paddle._C_ops.matmul(layer_norm_6, parameter_369, False, False) @@ -941,10 +668,11 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_11 = paddle._C_ops.add(matmul_8, parameter_368) - del parameter_368 + del matmul_8, parameter_368 # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_4 = paddle._C_ops.reshape(add_11, full_int_array_1) + del add_11 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_4 = paddle._C_ops.transpose(reshape_4, [0, 2, 1, 3]) @@ -956,7 +684,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_12 = paddle._C_ops.add(matmul_9, parameter_366) - del parameter_366 + del matmul_9, parameter_366 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_10 = paddle._C_ops.matmul(layer_norm_6, parameter_365, False, False) @@ -964,10 +692,11 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_13 = paddle._C_ops.add(matmul_10, parameter_364) - del parameter_364 + del matmul_10, parameter_364 # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_5 = paddle._C_ops.reshape(add_12, full_int_array_1) + del add_12 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_5 = paddle._C_ops.transpose(reshape_5, [0, 2, 1, 3]) @@ -975,6 +704,7 @@ def forward( # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_6 = paddle._C_ops.reshape(add_13, full_int_array_1) + del add_13 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_6 = paddle._C_ops.transpose(reshape_6, [0, 2, 1, 3]) @@ -986,9 +716,11 @@ def forward( # pd_op.matmul: (1x16x11x11xf32) <- (1x16x11x64xf32, 1x16x11x64xf32) matmul_11 = paddle._C_ops.matmul(scale_2, transpose_5, False, True) + del scale_2, transpose_5 # pd_op.add: (1x16x11x11xf32) <- (1x16x11x11xf32, 1x1x1x11xf32) add_14 = paddle._C_ops.add(matmul_11, unsqueeze_0) + del matmul_11 # pd_op.softmax: (1x16x11x11xf32) <- (1x16x11x11xf32) softmax_1 = paddle._C_ops.softmax(add_14, -1) @@ -997,13 +729,15 @@ def forward( # pd_op.dropout: (1x16x11x11xf32, 1x16x11x11xui8) <- (1x16x11x11xf32, None, 1xf32) dropout_8, dropout_9 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_1, None, full_4, False, "upscale_in_train", 0, False + softmax_1, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_1 # pd_op.matmul: (1x16x11x64xf32) <- (1x16x11x11xf32, 1x16x11x64xf32) matmul_12 = paddle._C_ops.matmul(dropout_8, transpose_6, False, False) + del dropout_8, transpose_6 # pd_op.transpose: (1x11x16x64xf32) <- (1x16x11x64xf32) transpose_7 = paddle._C_ops.transpose(matmul_12, [0, 2, 1, 3]) @@ -1011,19 +745,20 @@ def forward( # pd_op.reshape: (1x11x1024xf32) <- (1x11x16x64xf32, 3xi64) reshape_7 = paddle._C_ops.reshape(transpose_7, full_int_array_2) + del transpose_7 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_13 = paddle._C_ops.matmul(reshape_7, parameter_363, False, False) - del parameter_363 + del parameter_363, reshape_7 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_15 = paddle._C_ops.add(matmul_13, parameter_362) - del parameter_362 + del matmul_13, parameter_362 # pd_op.dropout: (1x11x1024xf32, 1x11x1024xui8) <- (1x11x1024xf32, None, 1xf32) dropout_10, dropout_11 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_15, None, full_4, False, "upscale_in_train", 0, False + add_15, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1031,6 +766,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_16 = paddle._C_ops.add(layer_norm_6, dropout_10) + del dropout_10, layer_norm_6 # pd_op.layer_norm: (1x11x1024xf32, 1x11xf32, 1x11xf32) <- (1x11x1024xf32, 1024xf32, 1024xf32) layer_norm_9, layer_norm_10, layer_norm_11 = (lambda x, f: f(x))( @@ -1039,7 +775,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_356, parameter_357 + del add_16, parameter_356, parameter_357 # pd_op.matmul: (1x11x4096xf32) <- (1x11x1024xf32, 1024x4096xf32) matmul_14 = paddle._C_ops.matmul(layer_norm_9, parameter_361, False, False) @@ -1047,7 +783,7 @@ def forward( # pd_op.add: (1x11x4096xf32) <- (1x11x4096xf32, 4096xf32) add_17 = paddle._C_ops.add(matmul_14, parameter_360) - del parameter_360 + del matmul_14, parameter_360 # pd_op.relu: (1x11x4096xf32) <- (1x11x4096xf32) relu_1 = paddle._C_ops.relu(add_17) @@ -1055,16 +791,16 @@ def forward( # pd_op.matmul: (1x11x1024xf32) <- (1x11x4096xf32, 4096x1024xf32) matmul_15 = paddle._C_ops.matmul(relu_1, parameter_359, False, False) - del parameter_359 + del parameter_359, relu_1 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_18 = paddle._C_ops.add(matmul_15, parameter_358) - del parameter_358 + del matmul_15, parameter_358 # pd_op.dropout: (1x11x1024xf32, 1x11x1024xui8) <- (1x11x1024xf32, None, 1xf32) dropout_12, dropout_13 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_18, None, full_4, False, "upscale_in_train", 0, False + add_18, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1072,6 +808,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_19 = paddle._C_ops.add(layer_norm_9, dropout_12) + del dropout_12, layer_norm_9 # pd_op.layer_norm: (1x11x1024xf32, 1x11xf32, 1x11xf32) <- (1x11x1024xf32, 1024xf32, 1024xf32) layer_norm_12, layer_norm_13, layer_norm_14 = (lambda x, f: f(x))( @@ -1080,7 +817,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_354, parameter_355 + del add_19, parameter_354, parameter_355 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_16 = paddle._C_ops.matmul(layer_norm_12, parameter_353, False, False) @@ -1088,10 +825,11 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_20 = paddle._C_ops.add(matmul_16, parameter_352) - del parameter_352 + del matmul_16, parameter_352 # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_8 = paddle._C_ops.reshape(add_20, full_int_array_1) + del add_20 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_8 = paddle._C_ops.transpose(reshape_8, [0, 2, 1, 3]) @@ -1103,7 +841,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_21 = paddle._C_ops.add(matmul_17, parameter_350) - del parameter_350 + del matmul_17, parameter_350 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_18 = paddle._C_ops.matmul(layer_norm_12, parameter_349, False, False) @@ -1111,10 +849,11 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_22 = paddle._C_ops.add(matmul_18, parameter_348) - del parameter_348 + del matmul_18, parameter_348 # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_9 = paddle._C_ops.reshape(add_21, full_int_array_1) + del add_21 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_9 = paddle._C_ops.transpose(reshape_9, [0, 2, 1, 3]) @@ -1122,6 +861,7 @@ def forward( # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_10 = paddle._C_ops.reshape(add_22, full_int_array_1) + del add_22 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_10 = paddle._C_ops.transpose(reshape_10, [0, 2, 1, 3]) @@ -1133,9 +873,11 @@ def forward( # pd_op.matmul: (1x16x11x11xf32) <- (1x16x11x64xf32, 1x16x11x64xf32) matmul_19 = paddle._C_ops.matmul(scale_3, transpose_9, False, True) + del scale_3, transpose_9 # pd_op.add: (1x16x11x11xf32) <- (1x16x11x11xf32, 1x1x1x11xf32) add_23 = paddle._C_ops.add(matmul_19, unsqueeze_0) + del matmul_19 # pd_op.softmax: (1x16x11x11xf32) <- (1x16x11x11xf32) softmax_2 = paddle._C_ops.softmax(add_23, -1) @@ -1144,13 +886,15 @@ def forward( # pd_op.dropout: (1x16x11x11xf32, 1x16x11x11xui8) <- (1x16x11x11xf32, None, 1xf32) dropout_14, dropout_15 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_2, None, full_4, False, "upscale_in_train", 0, False + softmax_2, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_2 # pd_op.matmul: (1x16x11x64xf32) <- (1x16x11x11xf32, 1x16x11x64xf32) matmul_20 = paddle._C_ops.matmul(dropout_14, transpose_10, False, False) + del dropout_14, transpose_10 # pd_op.transpose: (1x11x16x64xf32) <- (1x16x11x64xf32) transpose_11 = paddle._C_ops.transpose(matmul_20, [0, 2, 1, 3]) @@ -1158,19 +902,20 @@ def forward( # pd_op.reshape: (1x11x1024xf32) <- (1x11x16x64xf32, 3xi64) reshape_11 = paddle._C_ops.reshape(transpose_11, full_int_array_2) + del transpose_11 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_21 = paddle._C_ops.matmul(reshape_11, parameter_347, False, False) - del parameter_347 + del parameter_347, reshape_11 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_24 = paddle._C_ops.add(matmul_21, parameter_346) - del parameter_346 + del matmul_21, parameter_346 # pd_op.dropout: (1x11x1024xf32, 1x11x1024xui8) <- (1x11x1024xf32, None, 1xf32) dropout_16, dropout_17 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_24, None, full_4, False, "upscale_in_train", 0, False + add_24, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1178,6 +923,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_25 = paddle._C_ops.add(layer_norm_12, dropout_16) + del dropout_16, layer_norm_12 # pd_op.layer_norm: (1x11x1024xf32, 1x11xf32, 1x11xf32) <- (1x11x1024xf32, 1024xf32, 1024xf32) layer_norm_15, layer_norm_16, layer_norm_17 = (lambda x, f: f(x))( @@ -1186,7 +932,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_340, parameter_341 + del add_25, parameter_340, parameter_341 # pd_op.matmul: (1x11x4096xf32) <- (1x11x1024xf32, 1024x4096xf32) matmul_22 = paddle._C_ops.matmul(layer_norm_15, parameter_345, False, False) @@ -1194,7 +940,7 @@ def forward( # pd_op.add: (1x11x4096xf32) <- (1x11x4096xf32, 4096xf32) add_26 = paddle._C_ops.add(matmul_22, parameter_344) - del parameter_344 + del matmul_22, parameter_344 # pd_op.relu: (1x11x4096xf32) <- (1x11x4096xf32) relu_2 = paddle._C_ops.relu(add_26) @@ -1202,16 +948,16 @@ def forward( # pd_op.matmul: (1x11x1024xf32) <- (1x11x4096xf32, 4096x1024xf32) matmul_23 = paddle._C_ops.matmul(relu_2, parameter_343, False, False) - del parameter_343 + del parameter_343, relu_2 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_27 = paddle._C_ops.add(matmul_23, parameter_342) - del parameter_342 + del matmul_23, parameter_342 # pd_op.dropout: (1x11x1024xf32, 1x11x1024xui8) <- (1x11x1024xf32, None, 1xf32) dropout_18, dropout_19 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_27, None, full_4, False, "upscale_in_train", 0, False + add_27, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1219,6 +965,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_28 = paddle._C_ops.add(layer_norm_15, dropout_18) + del dropout_18, layer_norm_15 # pd_op.layer_norm: (1x11x1024xf32, 1x11xf32, 1x11xf32) <- (1x11x1024xf32, 1024xf32, 1024xf32) layer_norm_18, layer_norm_19, layer_norm_20 = (lambda x, f: f(x))( @@ -1227,7 +974,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_338, parameter_339 + del add_28, parameter_338, parameter_339 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_24 = paddle._C_ops.matmul(layer_norm_18, parameter_337, False, False) @@ -1235,10 +982,11 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_29 = paddle._C_ops.add(matmul_24, parameter_336) - del parameter_336 + del matmul_24, parameter_336 # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_12 = paddle._C_ops.reshape(add_29, full_int_array_1) + del add_29 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_12 = paddle._C_ops.transpose(reshape_12, [0, 2, 1, 3]) @@ -1250,7 +998,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_30 = paddle._C_ops.add(matmul_25, parameter_334) - del parameter_334 + del matmul_25, parameter_334 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_26 = paddle._C_ops.matmul(layer_norm_18, parameter_333, False, False) @@ -1258,10 +1006,11 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_31 = paddle._C_ops.add(matmul_26, parameter_332) - del parameter_332 + del matmul_26, parameter_332 # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_13 = paddle._C_ops.reshape(add_30, full_int_array_1) + del add_30 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_13 = paddle._C_ops.transpose(reshape_13, [0, 2, 1, 3]) @@ -1269,6 +1018,7 @@ def forward( # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_14 = paddle._C_ops.reshape(add_31, full_int_array_1) + del add_31 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_14 = paddle._C_ops.transpose(reshape_14, [0, 2, 1, 3]) @@ -1280,9 +1030,11 @@ def forward( # pd_op.matmul: (1x16x11x11xf32) <- (1x16x11x64xf32, 1x16x11x64xf32) matmul_27 = paddle._C_ops.matmul(scale_4, transpose_13, False, True) + del scale_4, transpose_13 # pd_op.add: (1x16x11x11xf32) <- (1x16x11x11xf32, 1x1x1x11xf32) add_32 = paddle._C_ops.add(matmul_27, unsqueeze_0) + del matmul_27 # pd_op.softmax: (1x16x11x11xf32) <- (1x16x11x11xf32) softmax_3 = paddle._C_ops.softmax(add_32, -1) @@ -1291,13 +1043,15 @@ def forward( # pd_op.dropout: (1x16x11x11xf32, 1x16x11x11xui8) <- (1x16x11x11xf32, None, 1xf32) dropout_20, dropout_21 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_3, None, full_4, False, "upscale_in_train", 0, False + softmax_3, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_3 # pd_op.matmul: (1x16x11x64xf32) <- (1x16x11x11xf32, 1x16x11x64xf32) matmul_28 = paddle._C_ops.matmul(dropout_20, transpose_14, False, False) + del dropout_20, transpose_14 # pd_op.transpose: (1x11x16x64xf32) <- (1x16x11x64xf32) transpose_15 = paddle._C_ops.transpose(matmul_28, [0, 2, 1, 3]) @@ -1305,19 +1059,20 @@ def forward( # pd_op.reshape: (1x11x1024xf32) <- (1x11x16x64xf32, 3xi64) reshape_15 = paddle._C_ops.reshape(transpose_15, full_int_array_2) + del transpose_15 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_29 = paddle._C_ops.matmul(reshape_15, parameter_331, False, False) - del parameter_331 + del parameter_331, reshape_15 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_33 = paddle._C_ops.add(matmul_29, parameter_330) - del parameter_330 + del matmul_29, parameter_330 # pd_op.dropout: (1x11x1024xf32, 1x11x1024xui8) <- (1x11x1024xf32, None, 1xf32) dropout_22, dropout_23 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_33, None, full_4, False, "upscale_in_train", 0, False + add_33, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1325,6 +1080,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_34 = paddle._C_ops.add(layer_norm_18, dropout_22) + del dropout_22, layer_norm_18 # pd_op.layer_norm: (1x11x1024xf32, 1x11xf32, 1x11xf32) <- (1x11x1024xf32, 1024xf32, 1024xf32) layer_norm_21, layer_norm_22, layer_norm_23 = (lambda x, f: f(x))( @@ -1333,7 +1089,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_324, parameter_325 + del add_34, parameter_324, parameter_325 # pd_op.matmul: (1x11x4096xf32) <- (1x11x1024xf32, 1024x4096xf32) matmul_30 = paddle._C_ops.matmul(layer_norm_21, parameter_329, False, False) @@ -1341,7 +1097,7 @@ def forward( # pd_op.add: (1x11x4096xf32) <- (1x11x4096xf32, 4096xf32) add_35 = paddle._C_ops.add(matmul_30, parameter_328) - del parameter_328 + del matmul_30, parameter_328 # pd_op.relu: (1x11x4096xf32) <- (1x11x4096xf32) relu_3 = paddle._C_ops.relu(add_35) @@ -1349,16 +1105,16 @@ def forward( # pd_op.matmul: (1x11x1024xf32) <- (1x11x4096xf32, 4096x1024xf32) matmul_31 = paddle._C_ops.matmul(relu_3, parameter_327, False, False) - del parameter_327 + del parameter_327, relu_3 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_36 = paddle._C_ops.add(matmul_31, parameter_326) - del parameter_326 + del matmul_31, parameter_326 # pd_op.dropout: (1x11x1024xf32, 1x11x1024xui8) <- (1x11x1024xf32, None, 1xf32) dropout_24, dropout_25 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_36, None, full_4, False, "upscale_in_train", 0, False + add_36, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1366,6 +1122,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_37 = paddle._C_ops.add(layer_norm_21, dropout_24) + del dropout_24, layer_norm_21 # pd_op.layer_norm: (1x11x1024xf32, 1x11xf32, 1x11xf32) <- (1x11x1024xf32, 1024xf32, 1024xf32) layer_norm_24, layer_norm_25, layer_norm_26 = (lambda x, f: f(x))( @@ -1374,7 +1131,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_322, parameter_323 + del add_37, parameter_322, parameter_323 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_32 = paddle._C_ops.matmul(layer_norm_24, parameter_321, False, False) @@ -1382,10 +1139,11 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_38 = paddle._C_ops.add(matmul_32, parameter_320) - del parameter_320 + del matmul_32, parameter_320 # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_16 = paddle._C_ops.reshape(add_38, full_int_array_1) + del add_38 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_16 = paddle._C_ops.transpose(reshape_16, [0, 2, 1, 3]) @@ -1397,7 +1155,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_39 = paddle._C_ops.add(matmul_33, parameter_318) - del parameter_318 + del matmul_33, parameter_318 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_34 = paddle._C_ops.matmul(layer_norm_24, parameter_317, False, False) @@ -1405,10 +1163,11 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_40 = paddle._C_ops.add(matmul_34, parameter_316) - del parameter_316 + del matmul_34, parameter_316 # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_17 = paddle._C_ops.reshape(add_39, full_int_array_1) + del add_39 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_17 = paddle._C_ops.transpose(reshape_17, [0, 2, 1, 3]) @@ -1416,6 +1175,7 @@ def forward( # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_18 = paddle._C_ops.reshape(add_40, full_int_array_1) + del add_40 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_18 = paddle._C_ops.transpose(reshape_18, [0, 2, 1, 3]) @@ -1427,9 +1187,11 @@ def forward( # pd_op.matmul: (1x16x11x11xf32) <- (1x16x11x64xf32, 1x16x11x64xf32) matmul_35 = paddle._C_ops.matmul(scale_5, transpose_17, False, True) + del scale_5, transpose_17 # pd_op.add: (1x16x11x11xf32) <- (1x16x11x11xf32, 1x1x1x11xf32) add_41 = paddle._C_ops.add(matmul_35, unsqueeze_0) + del matmul_35 # pd_op.softmax: (1x16x11x11xf32) <- (1x16x11x11xf32) softmax_4 = paddle._C_ops.softmax(add_41, -1) @@ -1438,13 +1200,15 @@ def forward( # pd_op.dropout: (1x16x11x11xf32, 1x16x11x11xui8) <- (1x16x11x11xf32, None, 1xf32) dropout_26, dropout_27 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_4, None, full_4, False, "upscale_in_train", 0, False + softmax_4, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_4 # pd_op.matmul: (1x16x11x64xf32) <- (1x16x11x11xf32, 1x16x11x64xf32) matmul_36 = paddle._C_ops.matmul(dropout_26, transpose_18, False, False) + del dropout_26, transpose_18 # pd_op.transpose: (1x11x16x64xf32) <- (1x16x11x64xf32) transpose_19 = paddle._C_ops.transpose(matmul_36, [0, 2, 1, 3]) @@ -1452,19 +1216,20 @@ def forward( # pd_op.reshape: (1x11x1024xf32) <- (1x11x16x64xf32, 3xi64) reshape_19 = paddle._C_ops.reshape(transpose_19, full_int_array_2) + del transpose_19 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_37 = paddle._C_ops.matmul(reshape_19, parameter_315, False, False) - del parameter_315 + del parameter_315, reshape_19 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_42 = paddle._C_ops.add(matmul_37, parameter_314) - del parameter_314 + del matmul_37, parameter_314 # pd_op.dropout: (1x11x1024xf32, 1x11x1024xui8) <- (1x11x1024xf32, None, 1xf32) dropout_28, dropout_29 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_42, None, full_4, False, "upscale_in_train", 0, False + add_42, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1472,6 +1237,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_43 = paddle._C_ops.add(layer_norm_24, dropout_28) + del dropout_28, layer_norm_24 # pd_op.layer_norm: (1x11x1024xf32, 1x11xf32, 1x11xf32) <- (1x11x1024xf32, 1024xf32, 1024xf32) layer_norm_27, layer_norm_28, layer_norm_29 = (lambda x, f: f(x))( @@ -1480,7 +1246,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_308, parameter_309 + del add_43, parameter_308, parameter_309 # pd_op.matmul: (1x11x4096xf32) <- (1x11x1024xf32, 1024x4096xf32) matmul_38 = paddle._C_ops.matmul(layer_norm_27, parameter_313, False, False) @@ -1488,7 +1254,7 @@ def forward( # pd_op.add: (1x11x4096xf32) <- (1x11x4096xf32, 4096xf32) add_44 = paddle._C_ops.add(matmul_38, parameter_312) - del parameter_312 + del matmul_38, parameter_312 # pd_op.relu: (1x11x4096xf32) <- (1x11x4096xf32) relu_4 = paddle._C_ops.relu(add_44) @@ -1496,16 +1262,16 @@ def forward( # pd_op.matmul: (1x11x1024xf32) <- (1x11x4096xf32, 4096x1024xf32) matmul_39 = paddle._C_ops.matmul(relu_4, parameter_311, False, False) - del parameter_311 + del parameter_311, relu_4 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_45 = paddle._C_ops.add(matmul_39, parameter_310) - del parameter_310 + del matmul_39, parameter_310 # pd_op.dropout: (1x11x1024xf32, 1x11x1024xui8) <- (1x11x1024xf32, None, 1xf32) dropout_30, dropout_31 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_45, None, full_4, False, "upscale_in_train", 0, False + add_45, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1513,6 +1279,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_46 = paddle._C_ops.add(layer_norm_27, dropout_30) + del dropout_30, layer_norm_27 # pd_op.layer_norm: (1x11x1024xf32, 1x11xf32, 1x11xf32) <- (1x11x1024xf32, 1024xf32, 1024xf32) layer_norm_30, layer_norm_31, layer_norm_32 = (lambda x, f: f(x))( @@ -1521,7 +1288,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_306, parameter_307 + del add_46, parameter_306, parameter_307 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_40 = paddle._C_ops.matmul(layer_norm_30, parameter_305, False, False) @@ -1529,10 +1296,11 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_47 = paddle._C_ops.add(matmul_40, parameter_304) - del parameter_304 + del matmul_40, parameter_304 # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_20 = paddle._C_ops.reshape(add_47, full_int_array_1) + del add_47 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_20 = paddle._C_ops.transpose(reshape_20, [0, 2, 1, 3]) @@ -1544,7 +1312,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_48 = paddle._C_ops.add(matmul_41, parameter_302) - del parameter_302 + del matmul_41, parameter_302 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_42 = paddle._C_ops.matmul(layer_norm_30, parameter_301, False, False) @@ -1552,10 +1320,11 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_49 = paddle._C_ops.add(matmul_42, parameter_300) - del parameter_300 + del matmul_42, parameter_300 # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_21 = paddle._C_ops.reshape(add_48, full_int_array_1) + del add_48 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_21 = paddle._C_ops.transpose(reshape_21, [0, 2, 1, 3]) @@ -1563,6 +1332,7 @@ def forward( # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_22 = paddle._C_ops.reshape(add_49, full_int_array_1) + del add_49 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_22 = paddle._C_ops.transpose(reshape_22, [0, 2, 1, 3]) @@ -1574,9 +1344,11 @@ def forward( # pd_op.matmul: (1x16x11x11xf32) <- (1x16x11x64xf32, 1x16x11x64xf32) matmul_43 = paddle._C_ops.matmul(scale_6, transpose_21, False, True) + del scale_6, transpose_21 # pd_op.add: (1x16x11x11xf32) <- (1x16x11x11xf32, 1x1x1x11xf32) add_50 = paddle._C_ops.add(matmul_43, unsqueeze_0) + del matmul_43 # pd_op.softmax: (1x16x11x11xf32) <- (1x16x11x11xf32) softmax_5 = paddle._C_ops.softmax(add_50, -1) @@ -1585,13 +1357,15 @@ def forward( # pd_op.dropout: (1x16x11x11xf32, 1x16x11x11xui8) <- (1x16x11x11xf32, None, 1xf32) dropout_32, dropout_33 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_5, None, full_4, False, "upscale_in_train", 0, False + softmax_5, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_5 # pd_op.matmul: (1x16x11x64xf32) <- (1x16x11x11xf32, 1x16x11x64xf32) matmul_44 = paddle._C_ops.matmul(dropout_32, transpose_22, False, False) + del dropout_32, transpose_22 # pd_op.transpose: (1x11x16x64xf32) <- (1x16x11x64xf32) transpose_23 = paddle._C_ops.transpose(matmul_44, [0, 2, 1, 3]) @@ -1599,19 +1373,20 @@ def forward( # pd_op.reshape: (1x11x1024xf32) <- (1x11x16x64xf32, 3xi64) reshape_23 = paddle._C_ops.reshape(transpose_23, full_int_array_2) + del transpose_23 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_45 = paddle._C_ops.matmul(reshape_23, parameter_299, False, False) - del parameter_299 + del parameter_299, reshape_23 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_51 = paddle._C_ops.add(matmul_45, parameter_298) - del parameter_298 + del matmul_45, parameter_298 # pd_op.dropout: (1x11x1024xf32, 1x11x1024xui8) <- (1x11x1024xf32, None, 1xf32) dropout_34, dropout_35 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_51, None, full_4, False, "upscale_in_train", 0, False + add_51, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1619,6 +1394,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_52 = paddle._C_ops.add(layer_norm_30, dropout_34) + del dropout_34, layer_norm_30 # pd_op.layer_norm: (1x11x1024xf32, 1x11xf32, 1x11xf32) <- (1x11x1024xf32, 1024xf32, 1024xf32) layer_norm_33, layer_norm_34, layer_norm_35 = (lambda x, f: f(x))( @@ -1627,7 +1403,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_292, parameter_293 + del add_52, parameter_292, parameter_293 # pd_op.matmul: (1x11x4096xf32) <- (1x11x1024xf32, 1024x4096xf32) matmul_46 = paddle._C_ops.matmul(layer_norm_33, parameter_297, False, False) @@ -1635,7 +1411,7 @@ def forward( # pd_op.add: (1x11x4096xf32) <- (1x11x4096xf32, 4096xf32) add_53 = paddle._C_ops.add(matmul_46, parameter_296) - del parameter_296 + del matmul_46, parameter_296 # pd_op.relu: (1x11x4096xf32) <- (1x11x4096xf32) relu_5 = paddle._C_ops.relu(add_53) @@ -1643,16 +1419,16 @@ def forward( # pd_op.matmul: (1x11x1024xf32) <- (1x11x4096xf32, 4096x1024xf32) matmul_47 = paddle._C_ops.matmul(relu_5, parameter_295, False, False) - del parameter_295 + del parameter_295, relu_5 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_54 = paddle._C_ops.add(matmul_47, parameter_294) - del parameter_294 + del matmul_47, parameter_294 # pd_op.dropout: (1x11x1024xf32, 1x11x1024xui8) <- (1x11x1024xf32, None, 1xf32) dropout_36, dropout_37 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_54, None, full_4, False, "upscale_in_train", 0, False + add_54, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1660,6 +1436,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_55 = paddle._C_ops.add(layer_norm_33, dropout_36) + del dropout_36, layer_norm_33 # pd_op.layer_norm: (1x11x1024xf32, 1x11xf32, 1x11xf32) <- (1x11x1024xf32, 1024xf32, 1024xf32) layer_norm_36, layer_norm_37, layer_norm_38 = (lambda x, f: f(x))( @@ -1668,7 +1445,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_290, parameter_291 + del add_55, parameter_290, parameter_291 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_48 = paddle._C_ops.matmul(layer_norm_36, parameter_289, False, False) @@ -1676,10 +1453,11 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_56 = paddle._C_ops.add(matmul_48, parameter_288) - del parameter_288 + del matmul_48, parameter_288 # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_24 = paddle._C_ops.reshape(add_56, full_int_array_1) + del add_56 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_24 = paddle._C_ops.transpose(reshape_24, [0, 2, 1, 3]) @@ -1691,7 +1469,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_57 = paddle._C_ops.add(matmul_49, parameter_286) - del parameter_286 + del matmul_49, parameter_286 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_50 = paddle._C_ops.matmul(layer_norm_36, parameter_285, False, False) @@ -1699,10 +1477,11 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_58 = paddle._C_ops.add(matmul_50, parameter_284) - del parameter_284 + del matmul_50, parameter_284 # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_25 = paddle._C_ops.reshape(add_57, full_int_array_1) + del add_57 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_25 = paddle._C_ops.transpose(reshape_25, [0, 2, 1, 3]) @@ -1710,6 +1489,7 @@ def forward( # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_26 = paddle._C_ops.reshape(add_58, full_int_array_1) + del add_58 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_26 = paddle._C_ops.transpose(reshape_26, [0, 2, 1, 3]) @@ -1721,9 +1501,11 @@ def forward( # pd_op.matmul: (1x16x11x11xf32) <- (1x16x11x64xf32, 1x16x11x64xf32) matmul_51 = paddle._C_ops.matmul(scale_7, transpose_25, False, True) + del scale_7, transpose_25 # pd_op.add: (1x16x11x11xf32) <- (1x16x11x11xf32, 1x1x1x11xf32) add_59 = paddle._C_ops.add(matmul_51, unsqueeze_0) + del matmul_51 # pd_op.softmax: (1x16x11x11xf32) <- (1x16x11x11xf32) softmax_6 = paddle._C_ops.softmax(add_59, -1) @@ -1732,13 +1514,15 @@ def forward( # pd_op.dropout: (1x16x11x11xf32, 1x16x11x11xui8) <- (1x16x11x11xf32, None, 1xf32) dropout_38, dropout_39 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_6, None, full_4, False, "upscale_in_train", 0, False + softmax_6, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_6 # pd_op.matmul: (1x16x11x64xf32) <- (1x16x11x11xf32, 1x16x11x64xf32) matmul_52 = paddle._C_ops.matmul(dropout_38, transpose_26, False, False) + del dropout_38, transpose_26 # pd_op.transpose: (1x11x16x64xf32) <- (1x16x11x64xf32) transpose_27 = paddle._C_ops.transpose(matmul_52, [0, 2, 1, 3]) @@ -1746,19 +1530,20 @@ def forward( # pd_op.reshape: (1x11x1024xf32) <- (1x11x16x64xf32, 3xi64) reshape_27 = paddle._C_ops.reshape(transpose_27, full_int_array_2) + del transpose_27 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_53 = paddle._C_ops.matmul(reshape_27, parameter_283, False, False) - del parameter_283 + del parameter_283, reshape_27 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_60 = paddle._C_ops.add(matmul_53, parameter_282) - del parameter_282 + del matmul_53, parameter_282 # pd_op.dropout: (1x11x1024xf32, 1x11x1024xui8) <- (1x11x1024xf32, None, 1xf32) dropout_40, dropout_41 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_60, None, full_4, False, "upscale_in_train", 0, False + add_60, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1766,6 +1551,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_61 = paddle._C_ops.add(layer_norm_36, dropout_40) + del dropout_40, layer_norm_36 # pd_op.layer_norm: (1x11x1024xf32, 1x11xf32, 1x11xf32) <- (1x11x1024xf32, 1024xf32, 1024xf32) layer_norm_39, layer_norm_40, layer_norm_41 = (lambda x, f: f(x))( @@ -1774,7 +1560,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_276, parameter_277 + del add_61, parameter_276, parameter_277 # pd_op.matmul: (1x11x4096xf32) <- (1x11x1024xf32, 1024x4096xf32) matmul_54 = paddle._C_ops.matmul(layer_norm_39, parameter_281, False, False) @@ -1782,7 +1568,7 @@ def forward( # pd_op.add: (1x11x4096xf32) <- (1x11x4096xf32, 4096xf32) add_62 = paddle._C_ops.add(matmul_54, parameter_280) - del parameter_280 + del matmul_54, parameter_280 # pd_op.relu: (1x11x4096xf32) <- (1x11x4096xf32) relu_6 = paddle._C_ops.relu(add_62) @@ -1790,16 +1576,16 @@ def forward( # pd_op.matmul: (1x11x1024xf32) <- (1x11x4096xf32, 4096x1024xf32) matmul_55 = paddle._C_ops.matmul(relu_6, parameter_279, False, False) - del parameter_279 + del parameter_279, relu_6 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_63 = paddle._C_ops.add(matmul_55, parameter_278) - del parameter_278 + del matmul_55, parameter_278 # pd_op.dropout: (1x11x1024xf32, 1x11x1024xui8) <- (1x11x1024xf32, None, 1xf32) dropout_42, dropout_43 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_63, None, full_4, False, "upscale_in_train", 0, False + add_63, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1807,6 +1593,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_64 = paddle._C_ops.add(layer_norm_39, dropout_42) + del dropout_42, layer_norm_39 # pd_op.layer_norm: (1x11x1024xf32, 1x11xf32, 1x11xf32) <- (1x11x1024xf32, 1024xf32, 1024xf32) layer_norm_42, layer_norm_43, layer_norm_44 = (lambda x, f: f(x))( @@ -1815,7 +1602,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_274, parameter_275 + del add_64, parameter_274, parameter_275 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_56 = paddle._C_ops.matmul(layer_norm_42, parameter_273, False, False) @@ -1823,10 +1610,11 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_65 = paddle._C_ops.add(matmul_56, parameter_272) - del parameter_272 + del matmul_56, parameter_272 # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_28 = paddle._C_ops.reshape(add_65, full_int_array_1) + del add_65 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_28 = paddle._C_ops.transpose(reshape_28, [0, 2, 1, 3]) @@ -1838,7 +1626,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_66 = paddle._C_ops.add(matmul_57, parameter_270) - del parameter_270 + del matmul_57, parameter_270 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_58 = paddle._C_ops.matmul(layer_norm_42, parameter_269, False, False) @@ -1846,10 +1634,11 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_67 = paddle._C_ops.add(matmul_58, parameter_268) - del parameter_268 + del matmul_58, parameter_268 # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_29 = paddle._C_ops.reshape(add_66, full_int_array_1) + del add_66 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_29 = paddle._C_ops.transpose(reshape_29, [0, 2, 1, 3]) @@ -1857,6 +1646,7 @@ def forward( # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_30 = paddle._C_ops.reshape(add_67, full_int_array_1) + del add_67 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_30 = paddle._C_ops.transpose(reshape_30, [0, 2, 1, 3]) @@ -1868,9 +1658,11 @@ def forward( # pd_op.matmul: (1x16x11x11xf32) <- (1x16x11x64xf32, 1x16x11x64xf32) matmul_59 = paddle._C_ops.matmul(scale_8, transpose_29, False, True) + del scale_8, transpose_29 # pd_op.add: (1x16x11x11xf32) <- (1x16x11x11xf32, 1x1x1x11xf32) add_68 = paddle._C_ops.add(matmul_59, unsqueeze_0) + del matmul_59 # pd_op.softmax: (1x16x11x11xf32) <- (1x16x11x11xf32) softmax_7 = paddle._C_ops.softmax(add_68, -1) @@ -1879,13 +1671,15 @@ def forward( # pd_op.dropout: (1x16x11x11xf32, 1x16x11x11xui8) <- (1x16x11x11xf32, None, 1xf32) dropout_44, dropout_45 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_7, None, full_4, False, "upscale_in_train", 0, False + softmax_7, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_7 # pd_op.matmul: (1x16x11x64xf32) <- (1x16x11x11xf32, 1x16x11x64xf32) matmul_60 = paddle._C_ops.matmul(dropout_44, transpose_30, False, False) + del dropout_44, transpose_30 # pd_op.transpose: (1x11x16x64xf32) <- (1x16x11x64xf32) transpose_31 = paddle._C_ops.transpose(matmul_60, [0, 2, 1, 3]) @@ -1893,19 +1687,20 @@ def forward( # pd_op.reshape: (1x11x1024xf32) <- (1x11x16x64xf32, 3xi64) reshape_31 = paddle._C_ops.reshape(transpose_31, full_int_array_2) + del transpose_31 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_61 = paddle._C_ops.matmul(reshape_31, parameter_267, False, False) - del parameter_267 + del parameter_267, reshape_31 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_69 = paddle._C_ops.add(matmul_61, parameter_266) - del parameter_266 + del matmul_61, parameter_266 # pd_op.dropout: (1x11x1024xf32, 1x11x1024xui8) <- (1x11x1024xf32, None, 1xf32) dropout_46, dropout_47 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_69, None, full_4, False, "upscale_in_train", 0, False + add_69, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1913,6 +1708,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_70 = paddle._C_ops.add(layer_norm_42, dropout_46) + del dropout_46, layer_norm_42 # pd_op.layer_norm: (1x11x1024xf32, 1x11xf32, 1x11xf32) <- (1x11x1024xf32, 1024xf32, 1024xf32) layer_norm_45, layer_norm_46, layer_norm_47 = (lambda x, f: f(x))( @@ -1921,7 +1717,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_260, parameter_261 + del add_70, parameter_260, parameter_261 # pd_op.matmul: (1x11x4096xf32) <- (1x11x1024xf32, 1024x4096xf32) matmul_62 = paddle._C_ops.matmul(layer_norm_45, parameter_265, False, False) @@ -1929,7 +1725,7 @@ def forward( # pd_op.add: (1x11x4096xf32) <- (1x11x4096xf32, 4096xf32) add_71 = paddle._C_ops.add(matmul_62, parameter_264) - del parameter_264 + del matmul_62, parameter_264 # pd_op.relu: (1x11x4096xf32) <- (1x11x4096xf32) relu_7 = paddle._C_ops.relu(add_71) @@ -1937,16 +1733,16 @@ def forward( # pd_op.matmul: (1x11x1024xf32) <- (1x11x4096xf32, 4096x1024xf32) matmul_63 = paddle._C_ops.matmul(relu_7, parameter_263, False, False) - del parameter_263 + del parameter_263, relu_7 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_72 = paddle._C_ops.add(matmul_63, parameter_262) - del parameter_262 + del matmul_63, parameter_262 # pd_op.dropout: (1x11x1024xf32, 1x11x1024xui8) <- (1x11x1024xf32, None, 1xf32) dropout_48, dropout_49 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_72, None, full_4, False, "upscale_in_train", 0, False + add_72, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1954,6 +1750,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_73 = paddle._C_ops.add(layer_norm_45, dropout_48) + del dropout_48, layer_norm_45 # pd_op.layer_norm: (1x11x1024xf32, 1x11xf32, 1x11xf32) <- (1x11x1024xf32, 1024xf32, 1024xf32) layer_norm_48, layer_norm_49, layer_norm_50 = (lambda x, f: f(x))( @@ -1962,7 +1759,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_258, parameter_259 + del add_73, parameter_258, parameter_259 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_64 = paddle._C_ops.matmul(layer_norm_48, parameter_257, False, False) @@ -1970,10 +1767,11 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_74 = paddle._C_ops.add(matmul_64, parameter_256) - del parameter_256 + del matmul_64, parameter_256 # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_32 = paddle._C_ops.reshape(add_74, full_int_array_1) + del add_74 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_32 = paddle._C_ops.transpose(reshape_32, [0, 2, 1, 3]) @@ -1985,7 +1783,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_75 = paddle._C_ops.add(matmul_65, parameter_254) - del parameter_254 + del matmul_65, parameter_254 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_66 = paddle._C_ops.matmul(layer_norm_48, parameter_253, False, False) @@ -1993,10 +1791,11 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_76 = paddle._C_ops.add(matmul_66, parameter_252) - del parameter_252 + del matmul_66, parameter_252 # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_33 = paddle._C_ops.reshape(add_75, full_int_array_1) + del add_75 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_33 = paddle._C_ops.transpose(reshape_33, [0, 2, 1, 3]) @@ -2004,6 +1803,7 @@ def forward( # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_34 = paddle._C_ops.reshape(add_76, full_int_array_1) + del add_76 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_34 = paddle._C_ops.transpose(reshape_34, [0, 2, 1, 3]) @@ -2015,9 +1815,11 @@ def forward( # pd_op.matmul: (1x16x11x11xf32) <- (1x16x11x64xf32, 1x16x11x64xf32) matmul_67 = paddle._C_ops.matmul(scale_9, transpose_33, False, True) + del scale_9, transpose_33 # pd_op.add: (1x16x11x11xf32) <- (1x16x11x11xf32, 1x1x1x11xf32) add_77 = paddle._C_ops.add(matmul_67, unsqueeze_0) + del matmul_67 # pd_op.softmax: (1x16x11x11xf32) <- (1x16x11x11xf32) softmax_8 = paddle._C_ops.softmax(add_77, -1) @@ -2026,13 +1828,15 @@ def forward( # pd_op.dropout: (1x16x11x11xf32, 1x16x11x11xui8) <- (1x16x11x11xf32, None, 1xf32) dropout_50, dropout_51 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_8, None, full_4, False, "upscale_in_train", 0, False + softmax_8, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_8 # pd_op.matmul: (1x16x11x64xf32) <- (1x16x11x11xf32, 1x16x11x64xf32) matmul_68 = paddle._C_ops.matmul(dropout_50, transpose_34, False, False) + del dropout_50, transpose_34 # pd_op.transpose: (1x11x16x64xf32) <- (1x16x11x64xf32) transpose_35 = paddle._C_ops.transpose(matmul_68, [0, 2, 1, 3]) @@ -2040,19 +1844,20 @@ def forward( # pd_op.reshape: (1x11x1024xf32) <- (1x11x16x64xf32, 3xi64) reshape_35 = paddle._C_ops.reshape(transpose_35, full_int_array_2) + del transpose_35 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_69 = paddle._C_ops.matmul(reshape_35, parameter_251, False, False) - del parameter_251 + del parameter_251, reshape_35 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_78 = paddle._C_ops.add(matmul_69, parameter_250) - del parameter_250 + del matmul_69, parameter_250 # pd_op.dropout: (1x11x1024xf32, 1x11x1024xui8) <- (1x11x1024xf32, None, 1xf32) dropout_52, dropout_53 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_78, None, full_4, False, "upscale_in_train", 0, False + add_78, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -2060,6 +1865,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_79 = paddle._C_ops.add(layer_norm_48, dropout_52) + del dropout_52, layer_norm_48 # pd_op.layer_norm: (1x11x1024xf32, 1x11xf32, 1x11xf32) <- (1x11x1024xf32, 1024xf32, 1024xf32) layer_norm_51, layer_norm_52, layer_norm_53 = (lambda x, f: f(x))( @@ -2068,7 +1874,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_244, parameter_245 + del add_79, parameter_244, parameter_245 # pd_op.matmul: (1x11x4096xf32) <- (1x11x1024xf32, 1024x4096xf32) matmul_70 = paddle._C_ops.matmul(layer_norm_51, parameter_249, False, False) @@ -2076,7 +1882,7 @@ def forward( # pd_op.add: (1x11x4096xf32) <- (1x11x4096xf32, 4096xf32) add_80 = paddle._C_ops.add(matmul_70, parameter_248) - del parameter_248 + del matmul_70, parameter_248 # pd_op.relu: (1x11x4096xf32) <- (1x11x4096xf32) relu_8 = paddle._C_ops.relu(add_80) @@ -2084,16 +1890,16 @@ def forward( # pd_op.matmul: (1x11x1024xf32) <- (1x11x4096xf32, 4096x1024xf32) matmul_71 = paddle._C_ops.matmul(relu_8, parameter_247, False, False) - del parameter_247 + del parameter_247, relu_8 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_81 = paddle._C_ops.add(matmul_71, parameter_246) - del parameter_246 + del matmul_71, parameter_246 # pd_op.dropout: (1x11x1024xf32, 1x11x1024xui8) <- (1x11x1024xf32, None, 1xf32) dropout_54, dropout_55 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_81, None, full_4, False, "upscale_in_train", 0, False + add_81, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -2101,6 +1907,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_82 = paddle._C_ops.add(layer_norm_51, dropout_54) + del dropout_54, layer_norm_51 # pd_op.layer_norm: (1x11x1024xf32, 1x11xf32, 1x11xf32) <- (1x11x1024xf32, 1024xf32, 1024xf32) layer_norm_54, layer_norm_55, layer_norm_56 = (lambda x, f: f(x))( @@ -2109,7 +1916,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_242, parameter_243 + del add_82, parameter_242, parameter_243 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_72 = paddle._C_ops.matmul(layer_norm_54, parameter_241, False, False) @@ -2117,10 +1924,11 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_83 = paddle._C_ops.add(matmul_72, parameter_240) - del parameter_240 + del matmul_72, parameter_240 # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_36 = paddle._C_ops.reshape(add_83, full_int_array_1) + del add_83 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_36 = paddle._C_ops.transpose(reshape_36, [0, 2, 1, 3]) @@ -2132,7 +1940,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_84 = paddle._C_ops.add(matmul_73, parameter_238) - del parameter_238 + del matmul_73, parameter_238 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_74 = paddle._C_ops.matmul(layer_norm_54, parameter_237, False, False) @@ -2140,10 +1948,11 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_85 = paddle._C_ops.add(matmul_74, parameter_236) - del parameter_236 + del matmul_74, parameter_236 # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_37 = paddle._C_ops.reshape(add_84, full_int_array_1) + del add_84 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_37 = paddle._C_ops.transpose(reshape_37, [0, 2, 1, 3]) @@ -2151,6 +1960,7 @@ def forward( # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_38 = paddle._C_ops.reshape(add_85, full_int_array_1) + del add_85 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_38 = paddle._C_ops.transpose(reshape_38, [0, 2, 1, 3]) @@ -2162,9 +1972,11 @@ def forward( # pd_op.matmul: (1x16x11x11xf32) <- (1x16x11x64xf32, 1x16x11x64xf32) matmul_75 = paddle._C_ops.matmul(scale_10, transpose_37, False, True) + del scale_10, transpose_37 # pd_op.add: (1x16x11x11xf32) <- (1x16x11x11xf32, 1x1x1x11xf32) add_86 = paddle._C_ops.add(matmul_75, unsqueeze_0) + del matmul_75 # pd_op.softmax: (1x16x11x11xf32) <- (1x16x11x11xf32) softmax_9 = paddle._C_ops.softmax(add_86, -1) @@ -2173,13 +1985,15 @@ def forward( # pd_op.dropout: (1x16x11x11xf32, 1x16x11x11xui8) <- (1x16x11x11xf32, None, 1xf32) dropout_56, dropout_57 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_9, None, full_4, False, "upscale_in_train", 0, False + softmax_9, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_9 # pd_op.matmul: (1x16x11x64xf32) <- (1x16x11x11xf32, 1x16x11x64xf32) matmul_76 = paddle._C_ops.matmul(dropout_56, transpose_38, False, False) + del dropout_56, transpose_38 # pd_op.transpose: (1x11x16x64xf32) <- (1x16x11x64xf32) transpose_39 = paddle._C_ops.transpose(matmul_76, [0, 2, 1, 3]) @@ -2187,19 +2001,20 @@ def forward( # pd_op.reshape: (1x11x1024xf32) <- (1x11x16x64xf32, 3xi64) reshape_39 = paddle._C_ops.reshape(transpose_39, full_int_array_2) + del transpose_39 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_77 = paddle._C_ops.matmul(reshape_39, parameter_235, False, False) - del parameter_235 + del parameter_235, reshape_39 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_87 = paddle._C_ops.add(matmul_77, parameter_234) - del parameter_234 + del matmul_77, parameter_234 # pd_op.dropout: (1x11x1024xf32, 1x11x1024xui8) <- (1x11x1024xf32, None, 1xf32) dropout_58, dropout_59 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_87, None, full_4, False, "upscale_in_train", 0, False + add_87, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -2207,6 +2022,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_88 = paddle._C_ops.add(layer_norm_54, dropout_58) + del dropout_58, layer_norm_54 # pd_op.layer_norm: (1x11x1024xf32, 1x11xf32, 1x11xf32) <- (1x11x1024xf32, 1024xf32, 1024xf32) layer_norm_57, layer_norm_58, layer_norm_59 = (lambda x, f: f(x))( @@ -2215,7 +2031,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_228, parameter_229 + del add_88, parameter_228, parameter_229 # pd_op.matmul: (1x11x4096xf32) <- (1x11x1024xf32, 1024x4096xf32) matmul_78 = paddle._C_ops.matmul(layer_norm_57, parameter_233, False, False) @@ -2223,7 +2039,7 @@ def forward( # pd_op.add: (1x11x4096xf32) <- (1x11x4096xf32, 4096xf32) add_89 = paddle._C_ops.add(matmul_78, parameter_232) - del parameter_232 + del matmul_78, parameter_232 # pd_op.relu: (1x11x4096xf32) <- (1x11x4096xf32) relu_9 = paddle._C_ops.relu(add_89) @@ -2231,16 +2047,16 @@ def forward( # pd_op.matmul: (1x11x1024xf32) <- (1x11x4096xf32, 4096x1024xf32) matmul_79 = paddle._C_ops.matmul(relu_9, parameter_231, False, False) - del parameter_231 + del parameter_231, relu_9 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_90 = paddle._C_ops.add(matmul_79, parameter_230) - del parameter_230 + del matmul_79, parameter_230 # pd_op.dropout: (1x11x1024xf32, 1x11x1024xui8) <- (1x11x1024xf32, None, 1xf32) dropout_60, dropout_61 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_90, None, full_4, False, "upscale_in_train", 0, False + add_90, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -2248,6 +2064,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_91 = paddle._C_ops.add(layer_norm_57, dropout_60) + del dropout_60, layer_norm_57 # pd_op.layer_norm: (1x11x1024xf32, 1x11xf32, 1x11xf32) <- (1x11x1024xf32, 1024xf32, 1024xf32) layer_norm_60, layer_norm_61, layer_norm_62 = (lambda x, f: f(x))( @@ -2256,7 +2073,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_226, parameter_227 + del add_91, parameter_226, parameter_227 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_80 = paddle._C_ops.matmul(layer_norm_60, parameter_225, False, False) @@ -2264,10 +2081,11 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_92 = paddle._C_ops.add(matmul_80, parameter_224) - del parameter_224 + del matmul_80, parameter_224 # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_40 = paddle._C_ops.reshape(add_92, full_int_array_1) + del add_92 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_40 = paddle._C_ops.transpose(reshape_40, [0, 2, 1, 3]) @@ -2279,7 +2097,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_93 = paddle._C_ops.add(matmul_81, parameter_222) - del parameter_222 + del matmul_81, parameter_222 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_82 = paddle._C_ops.matmul(layer_norm_60, parameter_221, False, False) @@ -2287,10 +2105,11 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_94 = paddle._C_ops.add(matmul_82, parameter_220) - del parameter_220 + del matmul_82, parameter_220 # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_41 = paddle._C_ops.reshape(add_93, full_int_array_1) + del add_93 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_41 = paddle._C_ops.transpose(reshape_41, [0, 2, 1, 3]) @@ -2298,6 +2117,7 @@ def forward( # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_42 = paddle._C_ops.reshape(add_94, full_int_array_1) + del add_94 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_42 = paddle._C_ops.transpose(reshape_42, [0, 2, 1, 3]) @@ -2309,9 +2129,11 @@ def forward( # pd_op.matmul: (1x16x11x11xf32) <- (1x16x11x64xf32, 1x16x11x64xf32) matmul_83 = paddle._C_ops.matmul(scale_11, transpose_41, False, True) + del scale_11, transpose_41 # pd_op.add: (1x16x11x11xf32) <- (1x16x11x11xf32, 1x1x1x11xf32) add_95 = paddle._C_ops.add(matmul_83, unsqueeze_0) + del matmul_83 # pd_op.softmax: (1x16x11x11xf32) <- (1x16x11x11xf32) softmax_10 = paddle._C_ops.softmax(add_95, -1) @@ -2320,13 +2142,15 @@ def forward( # pd_op.dropout: (1x16x11x11xf32, 1x16x11x11xui8) <- (1x16x11x11xf32, None, 1xf32) dropout_62, dropout_63 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_10, None, full_4, False, "upscale_in_train", 0, False + softmax_10, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_10 # pd_op.matmul: (1x16x11x64xf32) <- (1x16x11x11xf32, 1x16x11x64xf32) matmul_84 = paddle._C_ops.matmul(dropout_62, transpose_42, False, False) + del dropout_62, transpose_42 # pd_op.transpose: (1x11x16x64xf32) <- (1x16x11x64xf32) transpose_43 = paddle._C_ops.transpose(matmul_84, [0, 2, 1, 3]) @@ -2334,19 +2158,20 @@ def forward( # pd_op.reshape: (1x11x1024xf32) <- (1x11x16x64xf32, 3xi64) reshape_43 = paddle._C_ops.reshape(transpose_43, full_int_array_2) + del transpose_43 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_85 = paddle._C_ops.matmul(reshape_43, parameter_219, False, False) - del parameter_219 + del parameter_219, reshape_43 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_96 = paddle._C_ops.add(matmul_85, parameter_218) - del parameter_218 + del matmul_85, parameter_218 # pd_op.dropout: (1x11x1024xf32, 1x11x1024xui8) <- (1x11x1024xf32, None, 1xf32) dropout_64, dropout_65 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_96, None, full_4, False, "upscale_in_train", 0, False + add_96, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -2354,6 +2179,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_97 = paddle._C_ops.add(layer_norm_60, dropout_64) + del dropout_64, layer_norm_60 # pd_op.layer_norm: (1x11x1024xf32, 1x11xf32, 1x11xf32) <- (1x11x1024xf32, 1024xf32, 1024xf32) layer_norm_63, layer_norm_64, layer_norm_65 = (lambda x, f: f(x))( @@ -2362,7 +2188,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_212, parameter_213 + del add_97, parameter_212, parameter_213 # pd_op.matmul: (1x11x4096xf32) <- (1x11x1024xf32, 1024x4096xf32) matmul_86 = paddle._C_ops.matmul(layer_norm_63, parameter_217, False, False) @@ -2370,7 +2196,7 @@ def forward( # pd_op.add: (1x11x4096xf32) <- (1x11x4096xf32, 4096xf32) add_98 = paddle._C_ops.add(matmul_86, parameter_216) - del parameter_216 + del matmul_86, parameter_216 # pd_op.relu: (1x11x4096xf32) <- (1x11x4096xf32) relu_10 = paddle._C_ops.relu(add_98) @@ -2378,16 +2204,16 @@ def forward( # pd_op.matmul: (1x11x1024xf32) <- (1x11x4096xf32, 4096x1024xf32) matmul_87 = paddle._C_ops.matmul(relu_10, parameter_215, False, False) - del parameter_215 + del parameter_215, relu_10 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_99 = paddle._C_ops.add(matmul_87, parameter_214) - del parameter_214 + del matmul_87, parameter_214 # pd_op.dropout: (1x11x1024xf32, 1x11x1024xui8) <- (1x11x1024xf32, None, 1xf32) dropout_66, dropout_67 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_99, None, full_4, False, "upscale_in_train", 0, False + add_99, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -2395,6 +2221,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_100 = paddle._C_ops.add(layer_norm_63, dropout_66) + del dropout_66, layer_norm_63 # pd_op.layer_norm: (1x11x1024xf32, 1x11xf32, 1x11xf32) <- (1x11x1024xf32, 1024xf32, 1024xf32) layer_norm_66, layer_norm_67, layer_norm_68 = (lambda x, f: f(x))( @@ -2403,7 +2230,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_210, parameter_211 + del add_100, parameter_210, parameter_211 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_88 = paddle._C_ops.matmul(layer_norm_66, parameter_209, False, False) @@ -2411,10 +2238,11 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_101 = paddle._C_ops.add(matmul_88, parameter_208) - del parameter_208 + del matmul_88, parameter_208 # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_44 = paddle._C_ops.reshape(add_101, full_int_array_1) + del add_101 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_44 = paddle._C_ops.transpose(reshape_44, [0, 2, 1, 3]) @@ -2426,7 +2254,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_102 = paddle._C_ops.add(matmul_89, parameter_206) - del parameter_206 + del matmul_89, parameter_206 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_90 = paddle._C_ops.matmul(layer_norm_66, parameter_205, False, False) @@ -2434,10 +2262,11 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_103 = paddle._C_ops.add(matmul_90, parameter_204) - del parameter_204 + del matmul_90, parameter_204 # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_45 = paddle._C_ops.reshape(add_102, full_int_array_1) + del add_102 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_45 = paddle._C_ops.transpose(reshape_45, [0, 2, 1, 3]) @@ -2445,6 +2274,7 @@ def forward( # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_46 = paddle._C_ops.reshape(add_103, full_int_array_1) + del add_103 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_46 = paddle._C_ops.transpose(reshape_46, [0, 2, 1, 3]) @@ -2456,9 +2286,11 @@ def forward( # pd_op.matmul: (1x16x11x11xf32) <- (1x16x11x64xf32, 1x16x11x64xf32) matmul_91 = paddle._C_ops.matmul(scale_12, transpose_45, False, True) + del scale_12, transpose_45 # pd_op.add: (1x16x11x11xf32) <- (1x16x11x11xf32, 1x1x1x11xf32) add_104 = paddle._C_ops.add(matmul_91, unsqueeze_0) + del matmul_91 # pd_op.softmax: (1x16x11x11xf32) <- (1x16x11x11xf32) softmax_11 = paddle._C_ops.softmax(add_104, -1) @@ -2467,13 +2299,15 @@ def forward( # pd_op.dropout: (1x16x11x11xf32, 1x16x11x11xui8) <- (1x16x11x11xf32, None, 1xf32) dropout_68, dropout_69 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_11, None, full_4, False, "upscale_in_train", 0, False + softmax_11, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_11 # pd_op.matmul: (1x16x11x64xf32) <- (1x16x11x11xf32, 1x16x11x64xf32) matmul_92 = paddle._C_ops.matmul(dropout_68, transpose_46, False, False) + del dropout_68, transpose_46 # pd_op.transpose: (1x11x16x64xf32) <- (1x16x11x64xf32) transpose_47 = paddle._C_ops.transpose(matmul_92, [0, 2, 1, 3]) @@ -2481,19 +2315,20 @@ def forward( # pd_op.reshape: (1x11x1024xf32) <- (1x11x16x64xf32, 3xi64) reshape_47 = paddle._C_ops.reshape(transpose_47, full_int_array_2) + del transpose_47 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_93 = paddle._C_ops.matmul(reshape_47, parameter_203, False, False) - del parameter_203 + del parameter_203, reshape_47 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_105 = paddle._C_ops.add(matmul_93, parameter_202) - del parameter_202 + del matmul_93, parameter_202 # pd_op.dropout: (1x11x1024xf32, 1x11x1024xui8) <- (1x11x1024xf32, None, 1xf32) dropout_70, dropout_71 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_105, None, full_4, False, "upscale_in_train", 0, False + add_105, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -2501,6 +2336,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_106 = paddle._C_ops.add(layer_norm_66, dropout_70) + del dropout_70, layer_norm_66 # pd_op.layer_norm: (1x11x1024xf32, 1x11xf32, 1x11xf32) <- (1x11x1024xf32, 1024xf32, 1024xf32) layer_norm_69, layer_norm_70, layer_norm_71 = (lambda x, f: f(x))( @@ -2509,7 +2345,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_196, parameter_197 + del add_106, parameter_196, parameter_197 # pd_op.matmul: (1x11x4096xf32) <- (1x11x1024xf32, 1024x4096xf32) matmul_94 = paddle._C_ops.matmul(layer_norm_69, parameter_201, False, False) @@ -2517,7 +2353,7 @@ def forward( # pd_op.add: (1x11x4096xf32) <- (1x11x4096xf32, 4096xf32) add_107 = paddle._C_ops.add(matmul_94, parameter_200) - del parameter_200 + del matmul_94, parameter_200 # pd_op.relu: (1x11x4096xf32) <- (1x11x4096xf32) relu_11 = paddle._C_ops.relu(add_107) @@ -2525,16 +2361,16 @@ def forward( # pd_op.matmul: (1x11x1024xf32) <- (1x11x4096xf32, 4096x1024xf32) matmul_95 = paddle._C_ops.matmul(relu_11, parameter_199, False, False) - del parameter_199 + del parameter_199, relu_11 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_108 = paddle._C_ops.add(matmul_95, parameter_198) - del parameter_198 + del matmul_95, parameter_198 # pd_op.dropout: (1x11x1024xf32, 1x11x1024xui8) <- (1x11x1024xf32, None, 1xf32) dropout_72, dropout_73 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_108, None, full_4, False, "upscale_in_train", 0, False + add_108, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -2542,6 +2378,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_109 = paddle._C_ops.add(layer_norm_69, dropout_72) + del dropout_72, layer_norm_69 # pd_op.layer_norm: (1x11x1024xf32, 1x11xf32, 1x11xf32) <- (1x11x1024xf32, 1024xf32, 1024xf32) layer_norm_72, layer_norm_73, layer_norm_74 = (lambda x, f: f(x))( @@ -2550,7 +2387,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_194, parameter_195 + del add_109, parameter_194, parameter_195 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_96 = paddle._C_ops.matmul(layer_norm_72, parameter_193, False, False) @@ -2558,10 +2395,11 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_110 = paddle._C_ops.add(matmul_96, parameter_192) - del parameter_192 + del matmul_96, parameter_192 # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_48 = paddle._C_ops.reshape(add_110, full_int_array_1) + del add_110 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_48 = paddle._C_ops.transpose(reshape_48, [0, 2, 1, 3]) @@ -2573,7 +2411,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_111 = paddle._C_ops.add(matmul_97, parameter_190) - del parameter_190 + del matmul_97, parameter_190 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_98 = paddle._C_ops.matmul(layer_norm_72, parameter_189, False, False) @@ -2581,10 +2419,11 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_112 = paddle._C_ops.add(matmul_98, parameter_188) - del parameter_188 + del matmul_98, parameter_188 # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_49 = paddle._C_ops.reshape(add_111, full_int_array_1) + del add_111 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_49 = paddle._C_ops.transpose(reshape_49, [0, 2, 1, 3]) @@ -2592,6 +2431,7 @@ def forward( # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_50 = paddle._C_ops.reshape(add_112, full_int_array_1) + del add_112 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_50 = paddle._C_ops.transpose(reshape_50, [0, 2, 1, 3]) @@ -2603,9 +2443,11 @@ def forward( # pd_op.matmul: (1x16x11x11xf32) <- (1x16x11x64xf32, 1x16x11x64xf32) matmul_99 = paddle._C_ops.matmul(scale_13, transpose_49, False, True) + del scale_13, transpose_49 # pd_op.add: (1x16x11x11xf32) <- (1x16x11x11xf32, 1x1x1x11xf32) add_113 = paddle._C_ops.add(matmul_99, unsqueeze_0) + del matmul_99 # pd_op.softmax: (1x16x11x11xf32) <- (1x16x11x11xf32) softmax_12 = paddle._C_ops.softmax(add_113, -1) @@ -2614,13 +2456,15 @@ def forward( # pd_op.dropout: (1x16x11x11xf32, 1x16x11x11xui8) <- (1x16x11x11xf32, None, 1xf32) dropout_74, dropout_75 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_12, None, full_4, False, "upscale_in_train", 0, False + softmax_12, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_12 # pd_op.matmul: (1x16x11x64xf32) <- (1x16x11x11xf32, 1x16x11x64xf32) matmul_100 = paddle._C_ops.matmul(dropout_74, transpose_50, False, False) + del dropout_74, transpose_50 # pd_op.transpose: (1x11x16x64xf32) <- (1x16x11x64xf32) transpose_51 = paddle._C_ops.transpose(matmul_100, [0, 2, 1, 3]) @@ -2628,19 +2472,20 @@ def forward( # pd_op.reshape: (1x11x1024xf32) <- (1x11x16x64xf32, 3xi64) reshape_51 = paddle._C_ops.reshape(transpose_51, full_int_array_2) + del transpose_51 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_101 = paddle._C_ops.matmul(reshape_51, parameter_187, False, False) - del parameter_187 + del parameter_187, reshape_51 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_114 = paddle._C_ops.add(matmul_101, parameter_186) - del parameter_186 + del matmul_101, parameter_186 # pd_op.dropout: (1x11x1024xf32, 1x11x1024xui8) <- (1x11x1024xf32, None, 1xf32) dropout_76, dropout_77 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_114, None, full_4, False, "upscale_in_train", 0, False + add_114, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -2648,6 +2493,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_115 = paddle._C_ops.add(layer_norm_72, dropout_76) + del dropout_76, layer_norm_72 # pd_op.layer_norm: (1x11x1024xf32, 1x11xf32, 1x11xf32) <- (1x11x1024xf32, 1024xf32, 1024xf32) layer_norm_75, layer_norm_76, layer_norm_77 = (lambda x, f: f(x))( @@ -2656,7 +2502,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_180, parameter_181 + del add_115, parameter_180, parameter_181 # pd_op.matmul: (1x11x4096xf32) <- (1x11x1024xf32, 1024x4096xf32) matmul_102 = paddle._C_ops.matmul(layer_norm_75, parameter_185, False, False) @@ -2664,7 +2510,7 @@ def forward( # pd_op.add: (1x11x4096xf32) <- (1x11x4096xf32, 4096xf32) add_116 = paddle._C_ops.add(matmul_102, parameter_184) - del parameter_184 + del matmul_102, parameter_184 # pd_op.relu: (1x11x4096xf32) <- (1x11x4096xf32) relu_12 = paddle._C_ops.relu(add_116) @@ -2672,16 +2518,16 @@ def forward( # pd_op.matmul: (1x11x1024xf32) <- (1x11x4096xf32, 4096x1024xf32) matmul_103 = paddle._C_ops.matmul(relu_12, parameter_183, False, False) - del parameter_183 + del parameter_183, relu_12 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_117 = paddle._C_ops.add(matmul_103, parameter_182) - del parameter_182 + del matmul_103, parameter_182 # pd_op.dropout: (1x11x1024xf32, 1x11x1024xui8) <- (1x11x1024xf32, None, 1xf32) dropout_78, dropout_79 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_117, None, full_4, False, "upscale_in_train", 0, False + add_117, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -2689,6 +2535,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_118 = paddle._C_ops.add(layer_norm_75, dropout_78) + del dropout_78, layer_norm_75 # pd_op.layer_norm: (1x11x1024xf32, 1x11xf32, 1x11xf32) <- (1x11x1024xf32, 1024xf32, 1024xf32) layer_norm_78, layer_norm_79, layer_norm_80 = (lambda x, f: f(x))( @@ -2697,7 +2544,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_178, parameter_179 + del add_118, parameter_178, parameter_179 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_104 = paddle._C_ops.matmul(layer_norm_78, parameter_177, False, False) @@ -2705,10 +2552,11 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_119 = paddle._C_ops.add(matmul_104, parameter_176) - del parameter_176 + del matmul_104, parameter_176 # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_52 = paddle._C_ops.reshape(add_119, full_int_array_1) + del add_119 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_52 = paddle._C_ops.transpose(reshape_52, [0, 2, 1, 3]) @@ -2720,7 +2568,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_120 = paddle._C_ops.add(matmul_105, parameter_174) - del parameter_174 + del matmul_105, parameter_174 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_106 = paddle._C_ops.matmul(layer_norm_78, parameter_173, False, False) @@ -2728,10 +2576,11 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_121 = paddle._C_ops.add(matmul_106, parameter_172) - del parameter_172 + del matmul_106, parameter_172 # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_53 = paddle._C_ops.reshape(add_120, full_int_array_1) + del add_120 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_53 = paddle._C_ops.transpose(reshape_53, [0, 2, 1, 3]) @@ -2739,6 +2588,7 @@ def forward( # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_54 = paddle._C_ops.reshape(add_121, full_int_array_1) + del add_121 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_54 = paddle._C_ops.transpose(reshape_54, [0, 2, 1, 3]) @@ -2750,9 +2600,11 @@ def forward( # pd_op.matmul: (1x16x11x11xf32) <- (1x16x11x64xf32, 1x16x11x64xf32) matmul_107 = paddle._C_ops.matmul(scale_14, transpose_53, False, True) + del scale_14, transpose_53 # pd_op.add: (1x16x11x11xf32) <- (1x16x11x11xf32, 1x1x1x11xf32) add_122 = paddle._C_ops.add(matmul_107, unsqueeze_0) + del matmul_107 # pd_op.softmax: (1x16x11x11xf32) <- (1x16x11x11xf32) softmax_13 = paddle._C_ops.softmax(add_122, -1) @@ -2761,13 +2613,15 @@ def forward( # pd_op.dropout: (1x16x11x11xf32, 1x16x11x11xui8) <- (1x16x11x11xf32, None, 1xf32) dropout_80, dropout_81 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_13, None, full_4, False, "upscale_in_train", 0, False + softmax_13, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_13 # pd_op.matmul: (1x16x11x64xf32) <- (1x16x11x11xf32, 1x16x11x64xf32) matmul_108 = paddle._C_ops.matmul(dropout_80, transpose_54, False, False) + del dropout_80, transpose_54 # pd_op.transpose: (1x11x16x64xf32) <- (1x16x11x64xf32) transpose_55 = paddle._C_ops.transpose(matmul_108, [0, 2, 1, 3]) @@ -2775,19 +2629,20 @@ def forward( # pd_op.reshape: (1x11x1024xf32) <- (1x11x16x64xf32, 3xi64) reshape_55 = paddle._C_ops.reshape(transpose_55, full_int_array_2) + del transpose_55 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_109 = paddle._C_ops.matmul(reshape_55, parameter_171, False, False) - del parameter_171 + del parameter_171, reshape_55 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_123 = paddle._C_ops.add(matmul_109, parameter_170) - del parameter_170 + del matmul_109, parameter_170 # pd_op.dropout: (1x11x1024xf32, 1x11x1024xui8) <- (1x11x1024xf32, None, 1xf32) dropout_82, dropout_83 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_123, None, full_4, False, "upscale_in_train", 0, False + add_123, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -2795,6 +2650,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_124 = paddle._C_ops.add(layer_norm_78, dropout_82) + del dropout_82, layer_norm_78 # pd_op.layer_norm: (1x11x1024xf32, 1x11xf32, 1x11xf32) <- (1x11x1024xf32, 1024xf32, 1024xf32) layer_norm_81, layer_norm_82, layer_norm_83 = (lambda x, f: f(x))( @@ -2803,7 +2659,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_164, parameter_165 + del add_124, parameter_164, parameter_165 # pd_op.matmul: (1x11x4096xf32) <- (1x11x1024xf32, 1024x4096xf32) matmul_110 = paddle._C_ops.matmul(layer_norm_81, parameter_169, False, False) @@ -2811,7 +2667,7 @@ def forward( # pd_op.add: (1x11x4096xf32) <- (1x11x4096xf32, 4096xf32) add_125 = paddle._C_ops.add(matmul_110, parameter_168) - del parameter_168 + del matmul_110, parameter_168 # pd_op.relu: (1x11x4096xf32) <- (1x11x4096xf32) relu_13 = paddle._C_ops.relu(add_125) @@ -2819,16 +2675,16 @@ def forward( # pd_op.matmul: (1x11x1024xf32) <- (1x11x4096xf32, 4096x1024xf32) matmul_111 = paddle._C_ops.matmul(relu_13, parameter_167, False, False) - del parameter_167 + del parameter_167, relu_13 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_126 = paddle._C_ops.add(matmul_111, parameter_166) - del parameter_166 + del matmul_111, parameter_166 # pd_op.dropout: (1x11x1024xf32, 1x11x1024xui8) <- (1x11x1024xf32, None, 1xf32) dropout_84, dropout_85 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_126, None, full_4, False, "upscale_in_train", 0, False + add_126, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -2836,6 +2692,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_127 = paddle._C_ops.add(layer_norm_81, dropout_84) + del dropout_84, layer_norm_81 # pd_op.layer_norm: (1x11x1024xf32, 1x11xf32, 1x11xf32) <- (1x11x1024xf32, 1024xf32, 1024xf32) layer_norm_84, layer_norm_85, layer_norm_86 = (lambda x, f: f(x))( @@ -2844,7 +2701,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_162, parameter_163 + del add_127, parameter_162, parameter_163 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_112 = paddle._C_ops.matmul(layer_norm_84, parameter_161, False, False) @@ -2852,10 +2709,11 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_128 = paddle._C_ops.add(matmul_112, parameter_160) - del parameter_160 + del matmul_112, parameter_160 # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_56 = paddle._C_ops.reshape(add_128, full_int_array_1) + del add_128 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_56 = paddle._C_ops.transpose(reshape_56, [0, 2, 1, 3]) @@ -2867,7 +2725,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_129 = paddle._C_ops.add(matmul_113, parameter_158) - del parameter_158 + del matmul_113, parameter_158 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_114 = paddle._C_ops.matmul(layer_norm_84, parameter_157, False, False) @@ -2875,10 +2733,11 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_130 = paddle._C_ops.add(matmul_114, parameter_156) - del parameter_156 + del matmul_114, parameter_156 # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_57 = paddle._C_ops.reshape(add_129, full_int_array_1) + del add_129 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_57 = paddle._C_ops.transpose(reshape_57, [0, 2, 1, 3]) @@ -2886,6 +2745,7 @@ def forward( # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_58 = paddle._C_ops.reshape(add_130, full_int_array_1) + del add_130 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_58 = paddle._C_ops.transpose(reshape_58, [0, 2, 1, 3]) @@ -2897,9 +2757,11 @@ def forward( # pd_op.matmul: (1x16x11x11xf32) <- (1x16x11x64xf32, 1x16x11x64xf32) matmul_115 = paddle._C_ops.matmul(scale_15, transpose_57, False, True) + del scale_15, transpose_57 # pd_op.add: (1x16x11x11xf32) <- (1x16x11x11xf32, 1x1x1x11xf32) add_131 = paddle._C_ops.add(matmul_115, unsqueeze_0) + del matmul_115 # pd_op.softmax: (1x16x11x11xf32) <- (1x16x11x11xf32) softmax_14 = paddle._C_ops.softmax(add_131, -1) @@ -2908,13 +2770,15 @@ def forward( # pd_op.dropout: (1x16x11x11xf32, 1x16x11x11xui8) <- (1x16x11x11xf32, None, 1xf32) dropout_86, dropout_87 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_14, None, full_4, False, "upscale_in_train", 0, False + softmax_14, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_14 # pd_op.matmul: (1x16x11x64xf32) <- (1x16x11x11xf32, 1x16x11x64xf32) matmul_116 = paddle._C_ops.matmul(dropout_86, transpose_58, False, False) + del dropout_86, transpose_58 # pd_op.transpose: (1x11x16x64xf32) <- (1x16x11x64xf32) transpose_59 = paddle._C_ops.transpose(matmul_116, [0, 2, 1, 3]) @@ -2922,19 +2786,20 @@ def forward( # pd_op.reshape: (1x11x1024xf32) <- (1x11x16x64xf32, 3xi64) reshape_59 = paddle._C_ops.reshape(transpose_59, full_int_array_2) + del transpose_59 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_117 = paddle._C_ops.matmul(reshape_59, parameter_155, False, False) - del parameter_155 + del parameter_155, reshape_59 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_132 = paddle._C_ops.add(matmul_117, parameter_154) - del parameter_154 + del matmul_117, parameter_154 # pd_op.dropout: (1x11x1024xf32, 1x11x1024xui8) <- (1x11x1024xf32, None, 1xf32) dropout_88, dropout_89 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_132, None, full_4, False, "upscale_in_train", 0, False + add_132, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -2942,6 +2807,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_133 = paddle._C_ops.add(layer_norm_84, dropout_88) + del dropout_88, layer_norm_84 # pd_op.layer_norm: (1x11x1024xf32, 1x11xf32, 1x11xf32) <- (1x11x1024xf32, 1024xf32, 1024xf32) layer_norm_87, layer_norm_88, layer_norm_89 = (lambda x, f: f(x))( @@ -2950,7 +2816,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_148, parameter_149 + del add_133, parameter_148, parameter_149 # pd_op.matmul: (1x11x4096xf32) <- (1x11x1024xf32, 1024x4096xf32) matmul_118 = paddle._C_ops.matmul(layer_norm_87, parameter_153, False, False) @@ -2958,7 +2824,7 @@ def forward( # pd_op.add: (1x11x4096xf32) <- (1x11x4096xf32, 4096xf32) add_134 = paddle._C_ops.add(matmul_118, parameter_152) - del parameter_152 + del matmul_118, parameter_152 # pd_op.relu: (1x11x4096xf32) <- (1x11x4096xf32) relu_14 = paddle._C_ops.relu(add_134) @@ -2966,16 +2832,16 @@ def forward( # pd_op.matmul: (1x11x1024xf32) <- (1x11x4096xf32, 4096x1024xf32) matmul_119 = paddle._C_ops.matmul(relu_14, parameter_151, False, False) - del parameter_151 + del parameter_151, relu_14 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_135 = paddle._C_ops.add(matmul_119, parameter_150) - del parameter_150 + del matmul_119, parameter_150 # pd_op.dropout: (1x11x1024xf32, 1x11x1024xui8) <- (1x11x1024xf32, None, 1xf32) dropout_90, dropout_91 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_135, None, full_4, False, "upscale_in_train", 0, False + add_135, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -2983,6 +2849,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_136 = paddle._C_ops.add(layer_norm_87, dropout_90) + del dropout_90, layer_norm_87 # pd_op.layer_norm: (1x11x1024xf32, 1x11xf32, 1x11xf32) <- (1x11x1024xf32, 1024xf32, 1024xf32) layer_norm_90, layer_norm_91, layer_norm_92 = (lambda x, f: f(x))( @@ -2991,7 +2858,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_146, parameter_147 + del add_136, parameter_146, parameter_147 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_120 = paddle._C_ops.matmul(layer_norm_90, parameter_145, False, False) @@ -2999,10 +2866,11 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_137 = paddle._C_ops.add(matmul_120, parameter_144) - del parameter_144 + del matmul_120, parameter_144 # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_60 = paddle._C_ops.reshape(add_137, full_int_array_1) + del add_137 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_60 = paddle._C_ops.transpose(reshape_60, [0, 2, 1, 3]) @@ -3014,7 +2882,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_138 = paddle._C_ops.add(matmul_121, parameter_142) - del parameter_142 + del matmul_121, parameter_142 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_122 = paddle._C_ops.matmul(layer_norm_90, parameter_141, False, False) @@ -3022,10 +2890,11 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_139 = paddle._C_ops.add(matmul_122, parameter_140) - del parameter_140 + del matmul_122, parameter_140 # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_61 = paddle._C_ops.reshape(add_138, full_int_array_1) + del add_138 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_61 = paddle._C_ops.transpose(reshape_61, [0, 2, 1, 3]) @@ -3033,6 +2902,7 @@ def forward( # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_62 = paddle._C_ops.reshape(add_139, full_int_array_1) + del add_139 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_62 = paddle._C_ops.transpose(reshape_62, [0, 2, 1, 3]) @@ -3044,9 +2914,11 @@ def forward( # pd_op.matmul: (1x16x11x11xf32) <- (1x16x11x64xf32, 1x16x11x64xf32) matmul_123 = paddle._C_ops.matmul(scale_16, transpose_61, False, True) + del scale_16, transpose_61 # pd_op.add: (1x16x11x11xf32) <- (1x16x11x11xf32, 1x1x1x11xf32) add_140 = paddle._C_ops.add(matmul_123, unsqueeze_0) + del matmul_123 # pd_op.softmax: (1x16x11x11xf32) <- (1x16x11x11xf32) softmax_15 = paddle._C_ops.softmax(add_140, -1) @@ -3055,13 +2927,15 @@ def forward( # pd_op.dropout: (1x16x11x11xf32, 1x16x11x11xui8) <- (1x16x11x11xf32, None, 1xf32) dropout_92, dropout_93 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_15, None, full_4, False, "upscale_in_train", 0, False + softmax_15, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_15 # pd_op.matmul: (1x16x11x64xf32) <- (1x16x11x11xf32, 1x16x11x64xf32) matmul_124 = paddle._C_ops.matmul(dropout_92, transpose_62, False, False) + del dropout_92, transpose_62 # pd_op.transpose: (1x11x16x64xf32) <- (1x16x11x64xf32) transpose_63 = paddle._C_ops.transpose(matmul_124, [0, 2, 1, 3]) @@ -3069,19 +2943,20 @@ def forward( # pd_op.reshape: (1x11x1024xf32) <- (1x11x16x64xf32, 3xi64) reshape_63 = paddle._C_ops.reshape(transpose_63, full_int_array_2) + del transpose_63 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_125 = paddle._C_ops.matmul(reshape_63, parameter_139, False, False) - del parameter_139 + del parameter_139, reshape_63 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_141 = paddle._C_ops.add(matmul_125, parameter_138) - del parameter_138 + del matmul_125, parameter_138 # pd_op.dropout: (1x11x1024xf32, 1x11x1024xui8) <- (1x11x1024xf32, None, 1xf32) dropout_94, dropout_95 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_141, None, full_4, False, "upscale_in_train", 0, False + add_141, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -3089,6 +2964,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_142 = paddle._C_ops.add(layer_norm_90, dropout_94) + del dropout_94, layer_norm_90 # pd_op.layer_norm: (1x11x1024xf32, 1x11xf32, 1x11xf32) <- (1x11x1024xf32, 1024xf32, 1024xf32) layer_norm_93, layer_norm_94, layer_norm_95 = (lambda x, f: f(x))( @@ -3097,7 +2973,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_132, parameter_133 + del add_142, parameter_132, parameter_133 # pd_op.matmul: (1x11x4096xf32) <- (1x11x1024xf32, 1024x4096xf32) matmul_126 = paddle._C_ops.matmul(layer_norm_93, parameter_137, False, False) @@ -3105,7 +2981,7 @@ def forward( # pd_op.add: (1x11x4096xf32) <- (1x11x4096xf32, 4096xf32) add_143 = paddle._C_ops.add(matmul_126, parameter_136) - del parameter_136 + del matmul_126, parameter_136 # pd_op.relu: (1x11x4096xf32) <- (1x11x4096xf32) relu_15 = paddle._C_ops.relu(add_143) @@ -3113,16 +2989,16 @@ def forward( # pd_op.matmul: (1x11x1024xf32) <- (1x11x4096xf32, 4096x1024xf32) matmul_127 = paddle._C_ops.matmul(relu_15, parameter_135, False, False) - del parameter_135 + del parameter_135, relu_15 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_144 = paddle._C_ops.add(matmul_127, parameter_134) - del parameter_134 + del matmul_127, parameter_134 # pd_op.dropout: (1x11x1024xf32, 1x11x1024xui8) <- (1x11x1024xf32, None, 1xf32) dropout_96, dropout_97 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_144, None, full_4, False, "upscale_in_train", 0, False + add_144, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -3130,6 +3006,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_145 = paddle._C_ops.add(layer_norm_93, dropout_96) + del dropout_96, layer_norm_93 # pd_op.layer_norm: (1x11x1024xf32, 1x11xf32, 1x11xf32) <- (1x11x1024xf32, 1024xf32, 1024xf32) layer_norm_96, layer_norm_97, layer_norm_98 = (lambda x, f: f(x))( @@ -3138,7 +3015,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_130, parameter_131 + del add_145, parameter_130, parameter_131 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_128 = paddle._C_ops.matmul(layer_norm_96, parameter_129, False, False) @@ -3146,10 +3023,11 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_146 = paddle._C_ops.add(matmul_128, parameter_128) - del parameter_128 + del matmul_128, parameter_128 # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_64 = paddle._C_ops.reshape(add_146, full_int_array_1) + del add_146 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_64 = paddle._C_ops.transpose(reshape_64, [0, 2, 1, 3]) @@ -3161,7 +3039,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_147 = paddle._C_ops.add(matmul_129, parameter_126) - del parameter_126 + del matmul_129, parameter_126 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_130 = paddle._C_ops.matmul(layer_norm_96, parameter_125, False, False) @@ -3169,10 +3047,11 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_148 = paddle._C_ops.add(matmul_130, parameter_124) - del parameter_124 + del matmul_130, parameter_124 # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_65 = paddle._C_ops.reshape(add_147, full_int_array_1) + del add_147 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_65 = paddle._C_ops.transpose(reshape_65, [0, 2, 1, 3]) @@ -3180,6 +3059,7 @@ def forward( # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_66 = paddle._C_ops.reshape(add_148, full_int_array_1) + del add_148 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_66 = paddle._C_ops.transpose(reshape_66, [0, 2, 1, 3]) @@ -3191,9 +3071,11 @@ def forward( # pd_op.matmul: (1x16x11x11xf32) <- (1x16x11x64xf32, 1x16x11x64xf32) matmul_131 = paddle._C_ops.matmul(scale_17, transpose_65, False, True) + del scale_17, transpose_65 # pd_op.add: (1x16x11x11xf32) <- (1x16x11x11xf32, 1x1x1x11xf32) add_149 = paddle._C_ops.add(matmul_131, unsqueeze_0) + del matmul_131 # pd_op.softmax: (1x16x11x11xf32) <- (1x16x11x11xf32) softmax_16 = paddle._C_ops.softmax(add_149, -1) @@ -3202,13 +3084,15 @@ def forward( # pd_op.dropout: (1x16x11x11xf32, 1x16x11x11xui8) <- (1x16x11x11xf32, None, 1xf32) dropout_98, dropout_99 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_16, None, full_4, False, "upscale_in_train", 0, False + softmax_16, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_16 # pd_op.matmul: (1x16x11x64xf32) <- (1x16x11x11xf32, 1x16x11x64xf32) matmul_132 = paddle._C_ops.matmul(dropout_98, transpose_66, False, False) + del dropout_98, transpose_66 # pd_op.transpose: (1x11x16x64xf32) <- (1x16x11x64xf32) transpose_67 = paddle._C_ops.transpose(matmul_132, [0, 2, 1, 3]) @@ -3216,19 +3100,20 @@ def forward( # pd_op.reshape: (1x11x1024xf32) <- (1x11x16x64xf32, 3xi64) reshape_67 = paddle._C_ops.reshape(transpose_67, full_int_array_2) + del transpose_67 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_133 = paddle._C_ops.matmul(reshape_67, parameter_123, False, False) - del parameter_123 + del parameter_123, reshape_67 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_150 = paddle._C_ops.add(matmul_133, parameter_122) - del parameter_122 + del matmul_133, parameter_122 # pd_op.dropout: (1x11x1024xf32, 1x11x1024xui8) <- (1x11x1024xf32, None, 1xf32) dropout_100, dropout_101 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_150, None, full_4, False, "upscale_in_train", 0, False + add_150, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -3236,6 +3121,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_151 = paddle._C_ops.add(layer_norm_96, dropout_100) + del dropout_100, layer_norm_96 # pd_op.layer_norm: (1x11x1024xf32, 1x11xf32, 1x11xf32) <- (1x11x1024xf32, 1024xf32, 1024xf32) layer_norm_99, layer_norm_100, layer_norm_101 = (lambda x, f: f(x))( @@ -3244,7 +3130,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_116, parameter_117 + del add_151, parameter_116, parameter_117 # pd_op.matmul: (1x11x4096xf32) <- (1x11x1024xf32, 1024x4096xf32) matmul_134 = paddle._C_ops.matmul(layer_norm_99, parameter_121, False, False) @@ -3252,7 +3138,7 @@ def forward( # pd_op.add: (1x11x4096xf32) <- (1x11x4096xf32, 4096xf32) add_152 = paddle._C_ops.add(matmul_134, parameter_120) - del parameter_120 + del matmul_134, parameter_120 # pd_op.relu: (1x11x4096xf32) <- (1x11x4096xf32) relu_16 = paddle._C_ops.relu(add_152) @@ -3260,16 +3146,16 @@ def forward( # pd_op.matmul: (1x11x1024xf32) <- (1x11x4096xf32, 4096x1024xf32) matmul_135 = paddle._C_ops.matmul(relu_16, parameter_119, False, False) - del parameter_119 + del parameter_119, relu_16 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_153 = paddle._C_ops.add(matmul_135, parameter_118) - del parameter_118 + del matmul_135, parameter_118 # pd_op.dropout: (1x11x1024xf32, 1x11x1024xui8) <- (1x11x1024xf32, None, 1xf32) dropout_102, dropout_103 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_153, None, full_4, False, "upscale_in_train", 0, False + add_153, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -3277,6 +3163,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_154 = paddle._C_ops.add(layer_norm_99, dropout_102) + del dropout_102, layer_norm_99 # pd_op.layer_norm: (1x11x1024xf32, 1x11xf32, 1x11xf32) <- (1x11x1024xf32, 1024xf32, 1024xf32) layer_norm_102, layer_norm_103, layer_norm_104 = (lambda x, f: f(x))( @@ -3285,7 +3172,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_114, parameter_115 + del add_154, parameter_114, parameter_115 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_136 = paddle._C_ops.matmul(layer_norm_102, parameter_113, False, False) @@ -3293,10 +3180,11 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_155 = paddle._C_ops.add(matmul_136, parameter_112) - del parameter_112 + del matmul_136, parameter_112 # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_68 = paddle._C_ops.reshape(add_155, full_int_array_1) + del add_155 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_68 = paddle._C_ops.transpose(reshape_68, [0, 2, 1, 3]) @@ -3308,7 +3196,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_156 = paddle._C_ops.add(matmul_137, parameter_110) - del parameter_110 + del matmul_137, parameter_110 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_138 = paddle._C_ops.matmul(layer_norm_102, parameter_109, False, False) @@ -3316,10 +3204,11 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_157 = paddle._C_ops.add(matmul_138, parameter_108) - del parameter_108 + del matmul_138, parameter_108 # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_69 = paddle._C_ops.reshape(add_156, full_int_array_1) + del add_156 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_69 = paddle._C_ops.transpose(reshape_69, [0, 2, 1, 3]) @@ -3327,6 +3216,7 @@ def forward( # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_70 = paddle._C_ops.reshape(add_157, full_int_array_1) + del add_157 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_70 = paddle._C_ops.transpose(reshape_70, [0, 2, 1, 3]) @@ -3338,9 +3228,11 @@ def forward( # pd_op.matmul: (1x16x11x11xf32) <- (1x16x11x64xf32, 1x16x11x64xf32) matmul_139 = paddle._C_ops.matmul(scale_18, transpose_69, False, True) + del scale_18, transpose_69 # pd_op.add: (1x16x11x11xf32) <- (1x16x11x11xf32, 1x1x1x11xf32) add_158 = paddle._C_ops.add(matmul_139, unsqueeze_0) + del matmul_139 # pd_op.softmax: (1x16x11x11xf32) <- (1x16x11x11xf32) softmax_17 = paddle._C_ops.softmax(add_158, -1) @@ -3349,13 +3241,15 @@ def forward( # pd_op.dropout: (1x16x11x11xf32, 1x16x11x11xui8) <- (1x16x11x11xf32, None, 1xf32) dropout_104, dropout_105 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_17, None, full_4, False, "upscale_in_train", 0, False + softmax_17, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_17 # pd_op.matmul: (1x16x11x64xf32) <- (1x16x11x11xf32, 1x16x11x64xf32) matmul_140 = paddle._C_ops.matmul(dropout_104, transpose_70, False, False) + del dropout_104, transpose_70 # pd_op.transpose: (1x11x16x64xf32) <- (1x16x11x64xf32) transpose_71 = paddle._C_ops.transpose(matmul_140, [0, 2, 1, 3]) @@ -3363,19 +3257,20 @@ def forward( # pd_op.reshape: (1x11x1024xf32) <- (1x11x16x64xf32, 3xi64) reshape_71 = paddle._C_ops.reshape(transpose_71, full_int_array_2) + del transpose_71 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_141 = paddle._C_ops.matmul(reshape_71, parameter_107, False, False) - del parameter_107 + del parameter_107, reshape_71 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_159 = paddle._C_ops.add(matmul_141, parameter_106) - del parameter_106 + del matmul_141, parameter_106 # pd_op.dropout: (1x11x1024xf32, 1x11x1024xui8) <- (1x11x1024xf32, None, 1xf32) dropout_106, dropout_107 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_159, None, full_4, False, "upscale_in_train", 0, False + add_159, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -3383,6 +3278,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_160 = paddle._C_ops.add(layer_norm_102, dropout_106) + del dropout_106, layer_norm_102 # pd_op.layer_norm: (1x11x1024xf32, 1x11xf32, 1x11xf32) <- (1x11x1024xf32, 1024xf32, 1024xf32) layer_norm_105, layer_norm_106, layer_norm_107 = (lambda x, f: f(x))( @@ -3391,7 +3287,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_100, parameter_101 + del add_160, parameter_100, parameter_101 # pd_op.matmul: (1x11x4096xf32) <- (1x11x1024xf32, 1024x4096xf32) matmul_142 = paddle._C_ops.matmul(layer_norm_105, parameter_105, False, False) @@ -3399,7 +3295,7 @@ def forward( # pd_op.add: (1x11x4096xf32) <- (1x11x4096xf32, 4096xf32) add_161 = paddle._C_ops.add(matmul_142, parameter_104) - del parameter_104 + del matmul_142, parameter_104 # pd_op.relu: (1x11x4096xf32) <- (1x11x4096xf32) relu_17 = paddle._C_ops.relu(add_161) @@ -3407,16 +3303,16 @@ def forward( # pd_op.matmul: (1x11x1024xf32) <- (1x11x4096xf32, 4096x1024xf32) matmul_143 = paddle._C_ops.matmul(relu_17, parameter_103, False, False) - del parameter_103 + del parameter_103, relu_17 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_162 = paddle._C_ops.add(matmul_143, parameter_102) - del parameter_102 + del matmul_143, parameter_102 # pd_op.dropout: (1x11x1024xf32, 1x11x1024xui8) <- (1x11x1024xf32, None, 1xf32) dropout_108, dropout_109 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_162, None, full_4, False, "upscale_in_train", 0, False + add_162, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -3424,6 +3320,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_163 = paddle._C_ops.add(layer_norm_105, dropout_108) + del dropout_108, layer_norm_105 # pd_op.layer_norm: (1x11x1024xf32, 1x11xf32, 1x11xf32) <- (1x11x1024xf32, 1024xf32, 1024xf32) layer_norm_108, layer_norm_109, layer_norm_110 = (lambda x, f: f(x))( @@ -3432,7 +3329,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_98, parameter_99 + del add_163, parameter_98, parameter_99 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_144 = paddle._C_ops.matmul(layer_norm_108, parameter_97, False, False) @@ -3440,10 +3337,11 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_164 = paddle._C_ops.add(matmul_144, parameter_96) - del parameter_96 + del matmul_144, parameter_96 # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_72 = paddle._C_ops.reshape(add_164, full_int_array_1) + del add_164 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_72 = paddle._C_ops.transpose(reshape_72, [0, 2, 1, 3]) @@ -3455,7 +3353,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_165 = paddle._C_ops.add(matmul_145, parameter_94) - del parameter_94 + del matmul_145, parameter_94 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_146 = paddle._C_ops.matmul(layer_norm_108, parameter_93, False, False) @@ -3463,10 +3361,11 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_166 = paddle._C_ops.add(matmul_146, parameter_92) - del parameter_92 + del matmul_146, parameter_92 # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_73 = paddle._C_ops.reshape(add_165, full_int_array_1) + del add_165 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_73 = paddle._C_ops.transpose(reshape_73, [0, 2, 1, 3]) @@ -3474,6 +3373,7 @@ def forward( # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_74 = paddle._C_ops.reshape(add_166, full_int_array_1) + del add_166 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_74 = paddle._C_ops.transpose(reshape_74, [0, 2, 1, 3]) @@ -3485,9 +3385,11 @@ def forward( # pd_op.matmul: (1x16x11x11xf32) <- (1x16x11x64xf32, 1x16x11x64xf32) matmul_147 = paddle._C_ops.matmul(scale_19, transpose_73, False, True) + del scale_19, transpose_73 # pd_op.add: (1x16x11x11xf32) <- (1x16x11x11xf32, 1x1x1x11xf32) add_167 = paddle._C_ops.add(matmul_147, unsqueeze_0) + del matmul_147 # pd_op.softmax: (1x16x11x11xf32) <- (1x16x11x11xf32) softmax_18 = paddle._C_ops.softmax(add_167, -1) @@ -3496,13 +3398,15 @@ def forward( # pd_op.dropout: (1x16x11x11xf32, 1x16x11x11xui8) <- (1x16x11x11xf32, None, 1xf32) dropout_110, dropout_111 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_18, None, full_4, False, "upscale_in_train", 0, False + softmax_18, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_18 # pd_op.matmul: (1x16x11x64xf32) <- (1x16x11x11xf32, 1x16x11x64xf32) matmul_148 = paddle._C_ops.matmul(dropout_110, transpose_74, False, False) + del dropout_110, transpose_74 # pd_op.transpose: (1x11x16x64xf32) <- (1x16x11x64xf32) transpose_75 = paddle._C_ops.transpose(matmul_148, [0, 2, 1, 3]) @@ -3510,19 +3414,20 @@ def forward( # pd_op.reshape: (1x11x1024xf32) <- (1x11x16x64xf32, 3xi64) reshape_75 = paddle._C_ops.reshape(transpose_75, full_int_array_2) + del transpose_75 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_149 = paddle._C_ops.matmul(reshape_75, parameter_91, False, False) - del parameter_91 + del parameter_91, reshape_75 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_168 = paddle._C_ops.add(matmul_149, parameter_90) - del parameter_90 + del matmul_149, parameter_90 # pd_op.dropout: (1x11x1024xf32, 1x11x1024xui8) <- (1x11x1024xf32, None, 1xf32) dropout_112, dropout_113 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_168, None, full_4, False, "upscale_in_train", 0, False + add_168, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -3530,6 +3435,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_169 = paddle._C_ops.add(layer_norm_108, dropout_112) + del dropout_112, layer_norm_108 # pd_op.layer_norm: (1x11x1024xf32, 1x11xf32, 1x11xf32) <- (1x11x1024xf32, 1024xf32, 1024xf32) layer_norm_111, layer_norm_112, layer_norm_113 = (lambda x, f: f(x))( @@ -3538,7 +3444,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_84, parameter_85 + del add_169, parameter_84, parameter_85 # pd_op.matmul: (1x11x4096xf32) <- (1x11x1024xf32, 1024x4096xf32) matmul_150 = paddle._C_ops.matmul(layer_norm_111, parameter_89, False, False) @@ -3546,7 +3452,7 @@ def forward( # pd_op.add: (1x11x4096xf32) <- (1x11x4096xf32, 4096xf32) add_170 = paddle._C_ops.add(matmul_150, parameter_88) - del parameter_88 + del matmul_150, parameter_88 # pd_op.relu: (1x11x4096xf32) <- (1x11x4096xf32) relu_18 = paddle._C_ops.relu(add_170) @@ -3554,16 +3460,16 @@ def forward( # pd_op.matmul: (1x11x1024xf32) <- (1x11x4096xf32, 4096x1024xf32) matmul_151 = paddle._C_ops.matmul(relu_18, parameter_87, False, False) - del parameter_87 + del parameter_87, relu_18 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_171 = paddle._C_ops.add(matmul_151, parameter_86) - del parameter_86 + del matmul_151, parameter_86 # pd_op.dropout: (1x11x1024xf32, 1x11x1024xui8) <- (1x11x1024xf32, None, 1xf32) dropout_114, dropout_115 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_171, None, full_4, False, "upscale_in_train", 0, False + add_171, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -3571,6 +3477,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_172 = paddle._C_ops.add(layer_norm_111, dropout_114) + del dropout_114, layer_norm_111 # pd_op.layer_norm: (1x11x1024xf32, 1x11xf32, 1x11xf32) <- (1x11x1024xf32, 1024xf32, 1024xf32) layer_norm_114, layer_norm_115, layer_norm_116 = (lambda x, f: f(x))( @@ -3579,7 +3486,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_82, parameter_83 + del add_172, parameter_82, parameter_83 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_152 = paddle._C_ops.matmul(layer_norm_114, parameter_81, False, False) @@ -3587,10 +3494,11 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_173 = paddle._C_ops.add(matmul_152, parameter_80) - del parameter_80 + del matmul_152, parameter_80 # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_76 = paddle._C_ops.reshape(add_173, full_int_array_1) + del add_173 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_76 = paddle._C_ops.transpose(reshape_76, [0, 2, 1, 3]) @@ -3602,7 +3510,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_174 = paddle._C_ops.add(matmul_153, parameter_78) - del parameter_78 + del matmul_153, parameter_78 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_154 = paddle._C_ops.matmul(layer_norm_114, parameter_77, False, False) @@ -3610,10 +3518,11 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_175 = paddle._C_ops.add(matmul_154, parameter_76) - del parameter_76 + del matmul_154, parameter_76 # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_77 = paddle._C_ops.reshape(add_174, full_int_array_1) + del add_174 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_77 = paddle._C_ops.transpose(reshape_77, [0, 2, 1, 3]) @@ -3621,6 +3530,7 @@ def forward( # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_78 = paddle._C_ops.reshape(add_175, full_int_array_1) + del add_175 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_78 = paddle._C_ops.transpose(reshape_78, [0, 2, 1, 3]) @@ -3632,9 +3542,11 @@ def forward( # pd_op.matmul: (1x16x11x11xf32) <- (1x16x11x64xf32, 1x16x11x64xf32) matmul_155 = paddle._C_ops.matmul(scale_20, transpose_77, False, True) + del scale_20, transpose_77 # pd_op.add: (1x16x11x11xf32) <- (1x16x11x11xf32, 1x1x1x11xf32) add_176 = paddle._C_ops.add(matmul_155, unsqueeze_0) + del matmul_155 # pd_op.softmax: (1x16x11x11xf32) <- (1x16x11x11xf32) softmax_19 = paddle._C_ops.softmax(add_176, -1) @@ -3643,13 +3555,15 @@ def forward( # pd_op.dropout: (1x16x11x11xf32, 1x16x11x11xui8) <- (1x16x11x11xf32, None, 1xf32) dropout_116, dropout_117 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_19, None, full_4, False, "upscale_in_train", 0, False + softmax_19, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_19 # pd_op.matmul: (1x16x11x64xf32) <- (1x16x11x11xf32, 1x16x11x64xf32) matmul_156 = paddle._C_ops.matmul(dropout_116, transpose_78, False, False) + del dropout_116, transpose_78 # pd_op.transpose: (1x11x16x64xf32) <- (1x16x11x64xf32) transpose_79 = paddle._C_ops.transpose(matmul_156, [0, 2, 1, 3]) @@ -3657,19 +3571,20 @@ def forward( # pd_op.reshape: (1x11x1024xf32) <- (1x11x16x64xf32, 3xi64) reshape_79 = paddle._C_ops.reshape(transpose_79, full_int_array_2) + del transpose_79 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_157 = paddle._C_ops.matmul(reshape_79, parameter_75, False, False) - del parameter_75 + del parameter_75, reshape_79 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_177 = paddle._C_ops.add(matmul_157, parameter_74) - del parameter_74 + del matmul_157, parameter_74 # pd_op.dropout: (1x11x1024xf32, 1x11x1024xui8) <- (1x11x1024xf32, None, 1xf32) dropout_118, dropout_119 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_177, None, full_4, False, "upscale_in_train", 0, False + add_177, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -3677,6 +3592,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_178 = paddle._C_ops.add(layer_norm_114, dropout_118) + del dropout_118, layer_norm_114 # pd_op.layer_norm: (1x11x1024xf32, 1x11xf32, 1x11xf32) <- (1x11x1024xf32, 1024xf32, 1024xf32) layer_norm_117, layer_norm_118, layer_norm_119 = (lambda x, f: f(x))( @@ -3685,7 +3601,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_68, parameter_69 + del add_178, parameter_68, parameter_69 # pd_op.matmul: (1x11x4096xf32) <- (1x11x1024xf32, 1024x4096xf32) matmul_158 = paddle._C_ops.matmul(layer_norm_117, parameter_73, False, False) @@ -3693,7 +3609,7 @@ def forward( # pd_op.add: (1x11x4096xf32) <- (1x11x4096xf32, 4096xf32) add_179 = paddle._C_ops.add(matmul_158, parameter_72) - del parameter_72 + del matmul_158, parameter_72 # pd_op.relu: (1x11x4096xf32) <- (1x11x4096xf32) relu_19 = paddle._C_ops.relu(add_179) @@ -3701,16 +3617,16 @@ def forward( # pd_op.matmul: (1x11x1024xf32) <- (1x11x4096xf32, 4096x1024xf32) matmul_159 = paddle._C_ops.matmul(relu_19, parameter_71, False, False) - del parameter_71 + del parameter_71, relu_19 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_180 = paddle._C_ops.add(matmul_159, parameter_70) - del parameter_70 + del matmul_159, parameter_70 # pd_op.dropout: (1x11x1024xf32, 1x11x1024xui8) <- (1x11x1024xf32, None, 1xf32) dropout_120, dropout_121 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_180, None, full_4, False, "upscale_in_train", 0, False + add_180, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -3718,6 +3634,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_181 = paddle._C_ops.add(layer_norm_117, dropout_120) + del dropout_120, layer_norm_117 # pd_op.layer_norm: (1x11x1024xf32, 1x11xf32, 1x11xf32) <- (1x11x1024xf32, 1024xf32, 1024xf32) layer_norm_120, layer_norm_121, layer_norm_122 = (lambda x, f: f(x))( @@ -3726,7 +3643,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_66, parameter_67 + del add_181, parameter_66, parameter_67 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_160 = paddle._C_ops.matmul(layer_norm_120, parameter_65, False, False) @@ -3734,10 +3651,11 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_182 = paddle._C_ops.add(matmul_160, parameter_64) - del parameter_64 + del matmul_160, parameter_64 # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_80 = paddle._C_ops.reshape(add_182, full_int_array_1) + del add_182 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_80 = paddle._C_ops.transpose(reshape_80, [0, 2, 1, 3]) @@ -3749,7 +3667,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_183 = paddle._C_ops.add(matmul_161, parameter_62) - del parameter_62 + del matmul_161, parameter_62 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_162 = paddle._C_ops.matmul(layer_norm_120, parameter_61, False, False) @@ -3757,10 +3675,11 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_184 = paddle._C_ops.add(matmul_162, parameter_60) - del parameter_60 + del matmul_162, parameter_60 # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_81 = paddle._C_ops.reshape(add_183, full_int_array_1) + del add_183 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_81 = paddle._C_ops.transpose(reshape_81, [0, 2, 1, 3]) @@ -3768,6 +3687,7 @@ def forward( # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_82 = paddle._C_ops.reshape(add_184, full_int_array_1) + del add_184 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_82 = paddle._C_ops.transpose(reshape_82, [0, 2, 1, 3]) @@ -3779,9 +3699,11 @@ def forward( # pd_op.matmul: (1x16x11x11xf32) <- (1x16x11x64xf32, 1x16x11x64xf32) matmul_163 = paddle._C_ops.matmul(scale_21, transpose_81, False, True) + del scale_21, transpose_81 # pd_op.add: (1x16x11x11xf32) <- (1x16x11x11xf32, 1x1x1x11xf32) add_185 = paddle._C_ops.add(matmul_163, unsqueeze_0) + del matmul_163 # pd_op.softmax: (1x16x11x11xf32) <- (1x16x11x11xf32) softmax_20 = paddle._C_ops.softmax(add_185, -1) @@ -3790,13 +3712,15 @@ def forward( # pd_op.dropout: (1x16x11x11xf32, 1x16x11x11xui8) <- (1x16x11x11xf32, None, 1xf32) dropout_122, dropout_123 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_20, None, full_4, False, "upscale_in_train", 0, False + softmax_20, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_20 # pd_op.matmul: (1x16x11x64xf32) <- (1x16x11x11xf32, 1x16x11x64xf32) matmul_164 = paddle._C_ops.matmul(dropout_122, transpose_82, False, False) + del dropout_122, transpose_82 # pd_op.transpose: (1x11x16x64xf32) <- (1x16x11x64xf32) transpose_83 = paddle._C_ops.transpose(matmul_164, [0, 2, 1, 3]) @@ -3804,19 +3728,20 @@ def forward( # pd_op.reshape: (1x11x1024xf32) <- (1x11x16x64xf32, 3xi64) reshape_83 = paddle._C_ops.reshape(transpose_83, full_int_array_2) + del transpose_83 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_165 = paddle._C_ops.matmul(reshape_83, parameter_59, False, False) - del parameter_59 + del parameter_59, reshape_83 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_186 = paddle._C_ops.add(matmul_165, parameter_58) - del parameter_58 + del matmul_165, parameter_58 # pd_op.dropout: (1x11x1024xf32, 1x11x1024xui8) <- (1x11x1024xf32, None, 1xf32) dropout_124, dropout_125 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_186, None, full_4, False, "upscale_in_train", 0, False + add_186, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -3824,6 +3749,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_187 = paddle._C_ops.add(layer_norm_120, dropout_124) + del dropout_124, layer_norm_120 # pd_op.layer_norm: (1x11x1024xf32, 1x11xf32, 1x11xf32) <- (1x11x1024xf32, 1024xf32, 1024xf32) layer_norm_123, layer_norm_124, layer_norm_125 = (lambda x, f: f(x))( @@ -3832,7 +3758,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_52, parameter_53 + del add_187, parameter_52, parameter_53 # pd_op.matmul: (1x11x4096xf32) <- (1x11x1024xf32, 1024x4096xf32) matmul_166 = paddle._C_ops.matmul(layer_norm_123, parameter_57, False, False) @@ -3840,7 +3766,7 @@ def forward( # pd_op.add: (1x11x4096xf32) <- (1x11x4096xf32, 4096xf32) add_188 = paddle._C_ops.add(matmul_166, parameter_56) - del parameter_56 + del matmul_166, parameter_56 # pd_op.relu: (1x11x4096xf32) <- (1x11x4096xf32) relu_20 = paddle._C_ops.relu(add_188) @@ -3848,16 +3774,16 @@ def forward( # pd_op.matmul: (1x11x1024xf32) <- (1x11x4096xf32, 4096x1024xf32) matmul_167 = paddle._C_ops.matmul(relu_20, parameter_55, False, False) - del parameter_55 + del parameter_55, relu_20 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_189 = paddle._C_ops.add(matmul_167, parameter_54) - del parameter_54 + del matmul_167, parameter_54 # pd_op.dropout: (1x11x1024xf32, 1x11x1024xui8) <- (1x11x1024xf32, None, 1xf32) dropout_126, dropout_127 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_189, None, full_4, False, "upscale_in_train", 0, False + add_189, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -3865,6 +3791,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_190 = paddle._C_ops.add(layer_norm_123, dropout_126) + del dropout_126, layer_norm_123 # pd_op.layer_norm: (1x11x1024xf32, 1x11xf32, 1x11xf32) <- (1x11x1024xf32, 1024xf32, 1024xf32) layer_norm_126, layer_norm_127, layer_norm_128 = (lambda x, f: f(x))( @@ -3873,7 +3800,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_50, parameter_51 + del add_190, parameter_50, parameter_51 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_168 = paddle._C_ops.matmul(layer_norm_126, parameter_49, False, False) @@ -3881,10 +3808,11 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_191 = paddle._C_ops.add(matmul_168, parameter_48) - del parameter_48 + del matmul_168, parameter_48 # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_84 = paddle._C_ops.reshape(add_191, full_int_array_1) + del add_191 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_84 = paddle._C_ops.transpose(reshape_84, [0, 2, 1, 3]) @@ -3896,7 +3824,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_192 = paddle._C_ops.add(matmul_169, parameter_46) - del parameter_46 + del matmul_169, parameter_46 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_170 = paddle._C_ops.matmul(layer_norm_126, parameter_45, False, False) @@ -3904,10 +3832,11 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_193 = paddle._C_ops.add(matmul_170, parameter_44) - del parameter_44 + del matmul_170, parameter_44 # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_85 = paddle._C_ops.reshape(add_192, full_int_array_1) + del add_192 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_85 = paddle._C_ops.transpose(reshape_85, [0, 2, 1, 3]) @@ -3915,6 +3844,7 @@ def forward( # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_86 = paddle._C_ops.reshape(add_193, full_int_array_1) + del add_193 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_86 = paddle._C_ops.transpose(reshape_86, [0, 2, 1, 3]) @@ -3926,9 +3856,11 @@ def forward( # pd_op.matmul: (1x16x11x11xf32) <- (1x16x11x64xf32, 1x16x11x64xf32) matmul_171 = paddle._C_ops.matmul(scale_22, transpose_85, False, True) + del scale_22, transpose_85 # pd_op.add: (1x16x11x11xf32) <- (1x16x11x11xf32, 1x1x1x11xf32) add_194 = paddle._C_ops.add(matmul_171, unsqueeze_0) + del matmul_171 # pd_op.softmax: (1x16x11x11xf32) <- (1x16x11x11xf32) softmax_21 = paddle._C_ops.softmax(add_194, -1) @@ -3937,13 +3869,15 @@ def forward( # pd_op.dropout: (1x16x11x11xf32, 1x16x11x11xui8) <- (1x16x11x11xf32, None, 1xf32) dropout_128, dropout_129 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_21, None, full_4, False, "upscale_in_train", 0, False + softmax_21, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_21 # pd_op.matmul: (1x16x11x64xf32) <- (1x16x11x11xf32, 1x16x11x64xf32) matmul_172 = paddle._C_ops.matmul(dropout_128, transpose_86, False, False) + del dropout_128, transpose_86 # pd_op.transpose: (1x11x16x64xf32) <- (1x16x11x64xf32) transpose_87 = paddle._C_ops.transpose(matmul_172, [0, 2, 1, 3]) @@ -3951,19 +3885,20 @@ def forward( # pd_op.reshape: (1x11x1024xf32) <- (1x11x16x64xf32, 3xi64) reshape_87 = paddle._C_ops.reshape(transpose_87, full_int_array_2) + del transpose_87 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_173 = paddle._C_ops.matmul(reshape_87, parameter_43, False, False) - del parameter_43 + del parameter_43, reshape_87 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_195 = paddle._C_ops.add(matmul_173, parameter_42) - del parameter_42 + del matmul_173, parameter_42 # pd_op.dropout: (1x11x1024xf32, 1x11x1024xui8) <- (1x11x1024xf32, None, 1xf32) dropout_130, dropout_131 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_195, None, full_4, False, "upscale_in_train", 0, False + add_195, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -3971,6 +3906,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_196 = paddle._C_ops.add(layer_norm_126, dropout_130) + del dropout_130, layer_norm_126 # pd_op.layer_norm: (1x11x1024xf32, 1x11xf32, 1x11xf32) <- (1x11x1024xf32, 1024xf32, 1024xf32) layer_norm_129, layer_norm_130, layer_norm_131 = (lambda x, f: f(x))( @@ -3979,7 +3915,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_36, parameter_37 + del add_196, parameter_36, parameter_37 # pd_op.matmul: (1x11x4096xf32) <- (1x11x1024xf32, 1024x4096xf32) matmul_174 = paddle._C_ops.matmul(layer_norm_129, parameter_41, False, False) @@ -3987,7 +3923,7 @@ def forward( # pd_op.add: (1x11x4096xf32) <- (1x11x4096xf32, 4096xf32) add_197 = paddle._C_ops.add(matmul_174, parameter_40) - del parameter_40 + del matmul_174, parameter_40 # pd_op.relu: (1x11x4096xf32) <- (1x11x4096xf32) relu_21 = paddle._C_ops.relu(add_197) @@ -3995,16 +3931,16 @@ def forward( # pd_op.matmul: (1x11x1024xf32) <- (1x11x4096xf32, 4096x1024xf32) matmul_175 = paddle._C_ops.matmul(relu_21, parameter_39, False, False) - del parameter_39 + del parameter_39, relu_21 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_198 = paddle._C_ops.add(matmul_175, parameter_38) - del parameter_38 + del matmul_175, parameter_38 # pd_op.dropout: (1x11x1024xf32, 1x11x1024xui8) <- (1x11x1024xf32, None, 1xf32) dropout_132, dropout_133 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_198, None, full_4, False, "upscale_in_train", 0, False + add_198, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -4012,6 +3948,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_199 = paddle._C_ops.add(layer_norm_129, dropout_132) + del dropout_132, layer_norm_129 # pd_op.layer_norm: (1x11x1024xf32, 1x11xf32, 1x11xf32) <- (1x11x1024xf32, 1024xf32, 1024xf32) layer_norm_132, layer_norm_133, layer_norm_134 = (lambda x, f: f(x))( @@ -4020,7 +3957,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_34, parameter_35 + del add_199, parameter_34, parameter_35 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_176 = paddle._C_ops.matmul(layer_norm_132, parameter_33, False, False) @@ -4028,10 +3965,11 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_200 = paddle._C_ops.add(matmul_176, parameter_32) - del parameter_32 + del matmul_176, parameter_32 # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_88 = paddle._C_ops.reshape(add_200, full_int_array_1) + del add_200 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_88 = paddle._C_ops.transpose(reshape_88, [0, 2, 1, 3]) @@ -4043,7 +3981,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_201 = paddle._C_ops.add(matmul_177, parameter_30) - del parameter_30 + del matmul_177, parameter_30 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_178 = paddle._C_ops.matmul(layer_norm_132, parameter_29, False, False) @@ -4051,10 +3989,11 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_202 = paddle._C_ops.add(matmul_178, parameter_28) - del parameter_28 + del matmul_178, parameter_28 # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_89 = paddle._C_ops.reshape(add_201, full_int_array_1) + del add_201 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_89 = paddle._C_ops.transpose(reshape_89, [0, 2, 1, 3]) @@ -4062,6 +4001,7 @@ def forward( # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_90 = paddle._C_ops.reshape(add_202, full_int_array_1) + del add_202 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_90 = paddle._C_ops.transpose(reshape_90, [0, 2, 1, 3]) @@ -4073,9 +4013,11 @@ def forward( # pd_op.matmul: (1x16x11x11xf32) <- (1x16x11x64xf32, 1x16x11x64xf32) matmul_179 = paddle._C_ops.matmul(scale_23, transpose_89, False, True) + del scale_23, transpose_89 # pd_op.add: (1x16x11x11xf32) <- (1x16x11x11xf32, 1x1x1x11xf32) add_203 = paddle._C_ops.add(matmul_179, unsqueeze_0) + del matmul_179 # pd_op.softmax: (1x16x11x11xf32) <- (1x16x11x11xf32) softmax_22 = paddle._C_ops.softmax(add_203, -1) @@ -4084,13 +4026,15 @@ def forward( # pd_op.dropout: (1x16x11x11xf32, 1x16x11x11xui8) <- (1x16x11x11xf32, None, 1xf32) dropout_134, dropout_135 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_22, None, full_4, False, "upscale_in_train", 0, False + softmax_22, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_22 # pd_op.matmul: (1x16x11x64xf32) <- (1x16x11x11xf32, 1x16x11x64xf32) matmul_180 = paddle._C_ops.matmul(dropout_134, transpose_90, False, False) + del dropout_134, transpose_90 # pd_op.transpose: (1x11x16x64xf32) <- (1x16x11x64xf32) transpose_91 = paddle._C_ops.transpose(matmul_180, [0, 2, 1, 3]) @@ -4098,19 +4042,20 @@ def forward( # pd_op.reshape: (1x11x1024xf32) <- (1x11x16x64xf32, 3xi64) reshape_91 = paddle._C_ops.reshape(transpose_91, full_int_array_2) + del transpose_91 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_181 = paddle._C_ops.matmul(reshape_91, parameter_27, False, False) - del parameter_27 + del parameter_27, reshape_91 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_204 = paddle._C_ops.add(matmul_181, parameter_26) - del parameter_26 + del matmul_181, parameter_26 # pd_op.dropout: (1x11x1024xf32, 1x11x1024xui8) <- (1x11x1024xf32, None, 1xf32) dropout_136, dropout_137 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_204, None, full_4, False, "upscale_in_train", 0, False + add_204, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -4118,6 +4063,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_205 = paddle._C_ops.add(layer_norm_132, dropout_136) + del dropout_136, layer_norm_132 # pd_op.layer_norm: (1x11x1024xf32, 1x11xf32, 1x11xf32) <- (1x11x1024xf32, 1024xf32, 1024xf32) layer_norm_135, layer_norm_136, layer_norm_137 = (lambda x, f: f(x))( @@ -4126,7 +4072,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_20, parameter_21 + del add_205, parameter_20, parameter_21 # pd_op.matmul: (1x11x4096xf32) <- (1x11x1024xf32, 1024x4096xf32) matmul_182 = paddle._C_ops.matmul(layer_norm_135, parameter_25, False, False) @@ -4134,7 +4080,7 @@ def forward( # pd_op.add: (1x11x4096xf32) <- (1x11x4096xf32, 4096xf32) add_206 = paddle._C_ops.add(matmul_182, parameter_24) - del parameter_24 + del matmul_182, parameter_24 # pd_op.relu: (1x11x4096xf32) <- (1x11x4096xf32) relu_22 = paddle._C_ops.relu(add_206) @@ -4142,16 +4088,16 @@ def forward( # pd_op.matmul: (1x11x1024xf32) <- (1x11x4096xf32, 4096x1024xf32) matmul_183 = paddle._C_ops.matmul(relu_22, parameter_23, False, False) - del parameter_23 + del parameter_23, relu_22 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_207 = paddle._C_ops.add(matmul_183, parameter_22) - del parameter_22 + del matmul_183, parameter_22 # pd_op.dropout: (1x11x1024xf32, 1x11x1024xui8) <- (1x11x1024xf32, None, 1xf32) dropout_138, dropout_139 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_207, None, full_4, False, "upscale_in_train", 0, False + add_207, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -4159,6 +4105,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_208 = paddle._C_ops.add(layer_norm_135, dropout_138) + del dropout_138, layer_norm_135 # pd_op.layer_norm: (1x11x1024xf32, 1x11xf32, 1x11xf32) <- (1x11x1024xf32, 1024xf32, 1024xf32) layer_norm_138, layer_norm_139, layer_norm_140 = (lambda x, f: f(x))( @@ -4167,7 +4114,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_18, parameter_19 + del add_208, parameter_18, parameter_19 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_184 = paddle._C_ops.matmul(layer_norm_138, parameter_17, False, False) @@ -4175,10 +4122,11 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_209 = paddle._C_ops.add(matmul_184, parameter_16) - del parameter_16 + del matmul_184, parameter_16 # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_92 = paddle._C_ops.reshape(add_209, full_int_array_1) + del add_209 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_92 = paddle._C_ops.transpose(reshape_92, [0, 2, 1, 3]) @@ -4190,7 +4138,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_210 = paddle._C_ops.add(matmul_185, parameter_14) - del parameter_14 + del matmul_185, parameter_14 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_186 = paddle._C_ops.matmul(layer_norm_138, parameter_13, False, False) @@ -4198,10 +4146,11 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_211 = paddle._C_ops.add(matmul_186, parameter_12) - del parameter_12 + del matmul_186, parameter_12 # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_93 = paddle._C_ops.reshape(add_210, full_int_array_1) + del add_210 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_93 = paddle._C_ops.transpose(reshape_93, [0, 2, 1, 3]) @@ -4209,7 +4158,7 @@ def forward( # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_94 = paddle._C_ops.reshape(add_211, full_int_array_1) - del full_int_array_1 + del add_211, full_int_array_1 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_94 = paddle._C_ops.transpose(reshape_94, [0, 2, 1, 3]) @@ -4217,13 +4166,15 @@ def forward( # pd_op.scale: (1x16x11x64xf32) <- (1x16x11x64xf32, 1xf32) scale_24 = paddle._C_ops.scale(transpose_92, full_5, float("0"), True) - del transpose_92 + del full_5, transpose_92 # pd_op.matmul: (1x16x11x11xf32) <- (1x16x11x64xf32, 1x16x11x64xf32) matmul_187 = paddle._C_ops.matmul(scale_24, transpose_93, False, True) + del scale_24, transpose_93 # pd_op.add: (1x16x11x11xf32) <- (1x16x11x11xf32, 1x1x1x11xf32) add_212 = paddle._C_ops.add(matmul_187, unsqueeze_0) + del matmul_187, unsqueeze_0 # pd_op.softmax: (1x16x11x11xf32) <- (1x16x11x11xf32) softmax_23 = paddle._C_ops.softmax(add_212, -1) @@ -4232,13 +4183,15 @@ def forward( # pd_op.dropout: (1x16x11x11xf32, 1x16x11x11xui8) <- (1x16x11x11xf32, None, 1xf32) dropout_140, dropout_141 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_23, None, full_4, False, "upscale_in_train", 0, False + softmax_23, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_23 # pd_op.matmul: (1x16x11x64xf32) <- (1x16x11x11xf32, 1x16x11x64xf32) matmul_188 = paddle._C_ops.matmul(dropout_140, transpose_94, False, False) + del dropout_140, transpose_94 # pd_op.transpose: (1x11x16x64xf32) <- (1x16x11x64xf32) transpose_95 = paddle._C_ops.transpose(matmul_188, [0, 2, 1, 3]) @@ -4246,20 +4199,20 @@ def forward( # pd_op.reshape: (1x11x1024xf32) <- (1x11x16x64xf32, 3xi64) reshape_95 = paddle._C_ops.reshape(transpose_95, full_int_array_2) - del full_int_array_2 + del full_int_array_2, transpose_95 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_189 = paddle._C_ops.matmul(reshape_95, parameter_11, False, False) - del parameter_11 + del parameter_11, reshape_95 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_213 = paddle._C_ops.add(matmul_189, parameter_10) - del parameter_10 + del matmul_189, parameter_10 # pd_op.dropout: (1x11x1024xf32, 1x11x1024xui8) <- (1x11x1024xf32, None, 1xf32) dropout_142, dropout_143 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_213, None, full_4, False, "upscale_in_train", 0, False + add_213, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -4267,6 +4220,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_214 = paddle._C_ops.add(layer_norm_138, dropout_142) + del dropout_142, layer_norm_138 # pd_op.layer_norm: (1x11x1024xf32, 1x11xf32, 1x11xf32) <- (1x11x1024xf32, 1024xf32, 1024xf32) layer_norm_141, layer_norm_142, layer_norm_143 = (lambda x, f: f(x))( @@ -4275,7 +4229,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_4, parameter_5 + del add_214, parameter_4, parameter_5 # pd_op.matmul: (1x11x4096xf32) <- (1x11x1024xf32, 1024x4096xf32) matmul_190 = paddle._C_ops.matmul(layer_norm_141, parameter_9, False, False) @@ -4283,7 +4237,7 @@ def forward( # pd_op.add: (1x11x4096xf32) <- (1x11x4096xf32, 4096xf32) add_215 = paddle._C_ops.add(matmul_190, parameter_8) - del parameter_8 + del matmul_190, parameter_8 # pd_op.relu: (1x11x4096xf32) <- (1x11x4096xf32) relu_23 = paddle._C_ops.relu(add_215) @@ -4291,23 +4245,24 @@ def forward( # pd_op.matmul: (1x11x1024xf32) <- (1x11x4096xf32, 4096x1024xf32) matmul_191 = paddle._C_ops.matmul(relu_23, parameter_7, False, False) - del parameter_7 + del parameter_7, relu_23 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_216 = paddle._C_ops.add(matmul_191, parameter_6) - del parameter_6 + del matmul_191, parameter_6 # pd_op.dropout: (1x11x1024xf32, 1x11x1024xui8) <- (1x11x1024xf32, None, 1xf32) dropout_144, dropout_145 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_216, None, full_4, False, "upscale_in_train", 0, False + add_216, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) - del add_216 + del add_216, full_4 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_217 = paddle._C_ops.add(layer_norm_141, dropout_144) + del dropout_144, layer_norm_141 # pd_op.layer_norm: (1x11x1024xf32, 1x11xf32, 1x11xf32) <- (1x11x1024xf32, 1024xf32, 1024xf32) layer_norm_144, layer_norm_145, layer_norm_146 = (lambda x, f: f(x))( @@ -4316,7 +4271,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_2, parameter_3 + del add_217, parameter_2, parameter_3 # pd_op.full_int_array: (1xi64) <- () full_int_array_3 = [0] @@ -4328,875 +4283,18 @@ def forward( slice_0 = paddle._C_ops.slice( layer_norm_144, [1], full_int_array_3, full_int_array_4, [1], [1] ) + del full_int_array_3, full_int_array_4, layer_norm_144 # pd_op.matmul: (1x1024xf32) <- (1x1024xf32, 1024x1024xf32) matmul_192 = paddle._C_ops.matmul(slice_0, parameter_1, False, False) - del parameter_1 + del parameter_1, slice_0 # pd_op.add: (1x1024xf32) <- (1x1024xf32, 1024xf32) add_218 = paddle._C_ops.add(matmul_192, parameter_0) - del parameter_0 + del matmul_192, parameter_0 # pd_op.tanh: (1x1024xf32) <- (1x1024xf32) tanh_0 = paddle._C_ops.tanh(add_218) - del ( - add_0, - add_1, - add_10, - add_100, - add_101, - add_102, - add_103, - add_106, - add_109, - add_11, - add_110, - add_111, - add_112, - add_115, - add_118, - add_119, - add_12, - add_120, - add_121, - add_124, - add_127, - add_128, - add_129, - add_13, - add_130, - add_133, - add_136, - add_137, - add_138, - add_139, - add_142, - add_145, - add_146, - add_147, - add_148, - add_151, - add_154, - add_155, - add_156, - add_157, - add_16, - add_160, - add_163, - add_164, - add_165, - add_166, - add_169, - add_172, - add_173, - add_174, - add_175, - add_178, - add_181, - add_182, - add_183, - add_184, - add_187, - add_19, - add_190, - add_191, - add_192, - add_193, - add_196, - add_199, - add_2, - add_20, - add_200, - add_201, - add_202, - add_205, - add_208, - add_209, - add_21, - add_210, - add_211, - add_214, - add_217, - add_218, - add_22, - add_25, - add_28, - add_29, - add_3, - add_30, - add_31, - add_34, - add_37, - add_38, - add_39, - add_4, - add_40, - add_43, - add_46, - add_47, - add_48, - add_49, - add_52, - add_55, - add_56, - add_57, - add_58, - add_61, - add_64, - add_65, - add_66, - add_67, - add_7, - add_70, - add_73, - add_74, - add_75, - add_76, - add_79, - add_82, - add_83, - add_84, - add_85, - add_88, - add_91, - add_92, - add_93, - add_94, - add_97, - assign_0, - assign_1, - assign_10, - assign_11, - assign_12, - assign_13, - assign_14, - assign_15, - assign_16, - assign_17, - assign_18, - assign_19, - assign_2, - assign_20, - assign_21, - assign_22, - assign_23, - assign_24, - assign_25, - assign_26, - assign_27, - assign_28, - assign_29, - assign_3, - assign_30, - assign_31, - assign_32, - assign_33, - assign_34, - assign_35, - assign_36, - assign_37, - assign_38, - assign_39, - assign_4, - assign_40, - assign_41, - assign_42, - assign_43, - assign_44, - assign_45, - assign_46, - assign_47, - assign_48, - assign_49, - assign_5, - assign_50, - assign_51, - assign_52, - assign_53, - assign_54, - assign_55, - assign_56, - assign_57, - assign_58, - assign_59, - assign_6, - assign_60, - assign_61, - assign_62, - assign_63, - assign_64, - assign_65, - assign_66, - assign_67, - assign_68, - assign_69, - assign_7, - assign_70, - assign_71, - assign_72, - assign_73, - assign_74, - assign_75, - assign_76, - assign_77, - assign_78, - assign_79, - assign_8, - assign_80, - assign_81, - assign_82, - assign_83, - assign_84, - assign_85, - assign_86, - assign_87, - assign_88, - assign_89, - assign_9, - assign_90, - assign_91, - assign_92, - assign_93, - assign_94, - dropout_0, - dropout_1, - dropout_10, - dropout_100, - dropout_101, - dropout_102, - dropout_103, - dropout_104, - dropout_105, - dropout_106, - dropout_107, - dropout_108, - dropout_109, - dropout_11, - dropout_110, - dropout_111, - dropout_112, - dropout_113, - dropout_114, - dropout_115, - dropout_116, - dropout_117, - dropout_118, - dropout_119, - dropout_12, - dropout_120, - dropout_121, - dropout_122, - dropout_123, - dropout_124, - dropout_125, - dropout_126, - dropout_127, - dropout_128, - dropout_129, - dropout_13, - dropout_130, - dropout_131, - dropout_132, - dropout_133, - dropout_134, - dropout_135, - dropout_136, - dropout_137, - dropout_138, - dropout_139, - dropout_14, - dropout_140, - dropout_141, - dropout_142, - dropout_143, - dropout_144, - dropout_145, - dropout_15, - dropout_16, - dropout_17, - dropout_18, - dropout_19, - dropout_2, - dropout_20, - dropout_21, - dropout_22, - dropout_23, - dropout_24, - dropout_25, - dropout_26, - dropout_27, - dropout_28, - dropout_29, - dropout_3, - dropout_30, - dropout_31, - dropout_32, - dropout_33, - dropout_34, - dropout_35, - dropout_36, - dropout_37, - dropout_38, - dropout_39, - dropout_4, - dropout_40, - dropout_41, - dropout_42, - dropout_43, - dropout_44, - dropout_45, - dropout_46, - dropout_47, - dropout_48, - dropout_49, - dropout_5, - dropout_50, - dropout_51, - dropout_52, - dropout_53, - dropout_54, - dropout_55, - dropout_56, - dropout_57, - dropout_58, - dropout_59, - dropout_6, - dropout_60, - dropout_61, - dropout_62, - dropout_63, - dropout_64, - dropout_65, - dropout_66, - dropout_67, - dropout_68, - dropout_69, - dropout_7, - dropout_70, - dropout_71, - dropout_72, - dropout_73, - dropout_74, - dropout_75, - dropout_76, - dropout_77, - dropout_78, - dropout_79, - dropout_8, - dropout_80, - dropout_81, - dropout_82, - dropout_83, - dropout_84, - dropout_85, - dropout_86, - dropout_87, - dropout_88, - dropout_89, - dropout_9, - dropout_90, - dropout_91, - dropout_92, - dropout_93, - dropout_94, - dropout_95, - dropout_96, - dropout_97, - dropout_98, - dropout_99, - embedding_0, - embedding_1, - embedding_2, - full_4, - full_5, - full_int_array_3, - full_int_array_4, - layer_norm_1, - layer_norm_10, - layer_norm_100, - layer_norm_101, - layer_norm_102, - layer_norm_103, - layer_norm_104, - layer_norm_105, - layer_norm_106, - layer_norm_107, - layer_norm_108, - layer_norm_109, - layer_norm_11, - layer_norm_110, - layer_norm_111, - layer_norm_112, - layer_norm_113, - layer_norm_114, - layer_norm_115, - layer_norm_116, - layer_norm_117, - layer_norm_118, - layer_norm_119, - layer_norm_12, - layer_norm_120, - layer_norm_121, - layer_norm_122, - layer_norm_123, - layer_norm_124, - layer_norm_125, - layer_norm_126, - layer_norm_127, - layer_norm_128, - layer_norm_129, - layer_norm_13, - layer_norm_130, - layer_norm_131, - layer_norm_132, - layer_norm_133, - layer_norm_134, - layer_norm_135, - layer_norm_136, - layer_norm_137, - layer_norm_138, - layer_norm_139, - layer_norm_14, - layer_norm_140, - layer_norm_141, - layer_norm_142, - layer_norm_143, - layer_norm_144, - layer_norm_145, - layer_norm_146, - layer_norm_15, - layer_norm_16, - layer_norm_17, - layer_norm_18, - layer_norm_19, - layer_norm_2, - layer_norm_20, - layer_norm_21, - layer_norm_22, - layer_norm_23, - layer_norm_24, - layer_norm_25, - layer_norm_26, - layer_norm_27, - layer_norm_28, - layer_norm_29, - layer_norm_3, - layer_norm_30, - layer_norm_31, - layer_norm_32, - layer_norm_33, - layer_norm_34, - layer_norm_35, - layer_norm_36, - layer_norm_37, - layer_norm_38, - layer_norm_39, - layer_norm_4, - layer_norm_40, - layer_norm_41, - layer_norm_42, - layer_norm_43, - layer_norm_44, - layer_norm_45, - layer_norm_46, - layer_norm_47, - layer_norm_48, - layer_norm_49, - layer_norm_5, - layer_norm_50, - layer_norm_51, - layer_norm_52, - layer_norm_53, - layer_norm_54, - layer_norm_55, - layer_norm_56, - layer_norm_57, - layer_norm_58, - layer_norm_59, - layer_norm_6, - layer_norm_60, - layer_norm_61, - layer_norm_62, - layer_norm_63, - layer_norm_64, - layer_norm_65, - layer_norm_66, - layer_norm_67, - layer_norm_68, - layer_norm_69, - layer_norm_7, - layer_norm_70, - layer_norm_71, - layer_norm_72, - layer_norm_73, - layer_norm_74, - layer_norm_75, - layer_norm_76, - layer_norm_77, - layer_norm_78, - layer_norm_79, - layer_norm_8, - layer_norm_80, - layer_norm_81, - layer_norm_82, - layer_norm_83, - layer_norm_84, - layer_norm_85, - layer_norm_86, - layer_norm_87, - layer_norm_88, - layer_norm_89, - layer_norm_9, - layer_norm_90, - layer_norm_91, - layer_norm_92, - layer_norm_93, - layer_norm_94, - layer_norm_95, - layer_norm_96, - layer_norm_97, - layer_norm_98, - layer_norm_99, - matmul_0, - matmul_1, - matmul_10, - matmul_101, - matmul_102, - matmul_103, - matmul_104, - matmul_105, - matmul_106, - matmul_107, - matmul_109, - matmul_11, - matmul_110, - matmul_111, - matmul_112, - matmul_113, - matmul_114, - matmul_115, - matmul_117, - matmul_118, - matmul_119, - matmul_120, - matmul_121, - matmul_122, - matmul_123, - matmul_125, - matmul_126, - matmul_127, - matmul_128, - matmul_129, - matmul_13, - matmul_130, - matmul_131, - matmul_133, - matmul_134, - matmul_135, - matmul_136, - matmul_137, - matmul_138, - matmul_139, - matmul_14, - matmul_141, - matmul_142, - matmul_143, - matmul_144, - matmul_145, - matmul_146, - matmul_147, - matmul_149, - matmul_15, - matmul_150, - matmul_151, - matmul_152, - matmul_153, - matmul_154, - matmul_155, - matmul_157, - matmul_158, - matmul_159, - matmul_16, - matmul_160, - matmul_161, - matmul_162, - matmul_163, - matmul_165, - matmul_166, - matmul_167, - matmul_168, - matmul_169, - matmul_17, - matmul_170, - matmul_171, - matmul_173, - matmul_174, - matmul_175, - matmul_176, - matmul_177, - matmul_178, - matmul_179, - matmul_18, - matmul_181, - matmul_182, - matmul_183, - matmul_184, - matmul_185, - matmul_186, - matmul_187, - matmul_189, - matmul_19, - matmul_190, - matmul_191, - matmul_192, - matmul_2, - matmul_21, - matmul_22, - matmul_23, - matmul_24, - matmul_25, - matmul_26, - matmul_27, - matmul_29, - matmul_3, - matmul_30, - matmul_31, - matmul_32, - matmul_33, - matmul_34, - matmul_35, - matmul_37, - matmul_38, - matmul_39, - matmul_40, - matmul_41, - matmul_42, - matmul_43, - matmul_45, - matmul_46, - matmul_47, - matmul_48, - matmul_49, - matmul_5, - matmul_50, - matmul_51, - matmul_53, - matmul_54, - matmul_55, - matmul_56, - matmul_57, - matmul_58, - matmul_59, - matmul_6, - matmul_61, - matmul_62, - matmul_63, - matmul_64, - matmul_65, - matmul_66, - matmul_67, - matmul_69, - matmul_7, - matmul_70, - matmul_71, - matmul_72, - matmul_73, - matmul_74, - matmul_75, - matmul_77, - matmul_78, - matmul_79, - matmul_8, - matmul_80, - matmul_81, - matmul_82, - matmul_83, - matmul_85, - matmul_86, - matmul_87, - matmul_88, - matmul_89, - matmul_9, - matmul_90, - matmul_91, - matmul_93, - matmul_94, - matmul_95, - matmul_96, - matmul_97, - matmul_98, - matmul_99, - relu_0, - relu_1, - relu_10, - relu_11, - relu_12, - relu_13, - relu_14, - relu_15, - relu_16, - relu_17, - relu_18, - relu_19, - relu_2, - relu_20, - relu_21, - relu_22, - relu_23, - relu_3, - relu_4, - relu_5, - relu_6, - relu_7, - relu_8, - relu_9, - reshape_11, - reshape_15, - reshape_19, - reshape_23, - reshape_27, - reshape_3, - reshape_31, - reshape_35, - reshape_39, - reshape_43, - reshape_47, - reshape_51, - reshape_55, - reshape_59, - reshape_63, - reshape_67, - reshape_7, - reshape_71, - reshape_75, - reshape_79, - reshape_83, - reshape_87, - reshape_91, - reshape_95, - scale_1, - scale_10, - scale_11, - scale_12, - scale_13, - scale_14, - scale_15, - scale_16, - scale_17, - scale_18, - scale_19, - scale_2, - scale_20, - scale_21, - scale_22, - scale_23, - scale_24, - scale_3, - scale_4, - scale_5, - scale_6, - scale_7, - scale_8, - scale_9, - slice_0, - softmax_0, - softmax_1, - softmax_10, - softmax_11, - softmax_12, - softmax_13, - softmax_14, - softmax_15, - softmax_16, - softmax_17, - softmax_18, - softmax_19, - softmax_2, - softmax_20, - softmax_21, - softmax_22, - softmax_23, - softmax_3, - softmax_4, - softmax_5, - softmax_6, - softmax_7, - softmax_8, - softmax_9, - subtract_0, - transpose_1, - transpose_10, - transpose_11, - transpose_13, - transpose_14, - transpose_15, - transpose_17, - transpose_18, - transpose_19, - transpose_2, - transpose_21, - transpose_22, - transpose_23, - transpose_25, - transpose_26, - transpose_27, - transpose_29, - transpose_3, - transpose_30, - transpose_31, - transpose_33, - transpose_34, - transpose_35, - transpose_37, - transpose_38, - transpose_39, - transpose_41, - transpose_42, - transpose_43, - transpose_45, - transpose_46, - transpose_47, - transpose_49, - transpose_5, - transpose_50, - transpose_51, - transpose_53, - transpose_54, - transpose_55, - transpose_57, - transpose_58, - transpose_59, - transpose_6, - transpose_61, - transpose_62, - transpose_63, - transpose_65, - transpose_66, - transpose_67, - transpose_69, - transpose_7, - transpose_70, - transpose_71, - transpose_73, - transpose_74, - transpose_75, - transpose_77, - transpose_78, - transpose_79, - transpose_81, - transpose_82, - transpose_83, - transpose_85, - transpose_86, - transpose_87, - transpose_89, - transpose_9, - transpose_90, - transpose_91, - transpose_93, - transpose_94, - transpose_95, - unsqueeze_0, - ) + del add_218 return tanh_0 diff --git a/paddle_samples/PaddleNLP/ernie-2.0-large-zh/weight_meta.py b/paddle_samples/PaddleNLP/ernie-2.0-large-zh/weight_meta.py index 5afc30d63..c247a643b 100644 --- a/paddle_samples/PaddleNLP/ernie-2.0-large-zh/weight_meta.py +++ b/paddle_samples/PaddleNLP/ernie-2.0-large-zh/weight_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_parameter_0: name = "parameter_0" + original_name = "linear_581.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0744659") @@ -11,6 +12,7 @@ class Program_weight_tensor_parameter_0: class Program_weight_tensor_parameter_1: name = "parameter_1" + original_name = "linear_581.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.185181") @@ -22,6 +24,7 @@ class Program_weight_tensor_parameter_1: class Program_weight_tensor_parameter_2: name = "parameter_2" + original_name = "layer_norm_197.b_0" shape = [1024] dtype = "float32" min_val = float("-0.261899") @@ -33,6 +36,7 @@ class Program_weight_tensor_parameter_2: class Program_weight_tensor_parameter_3: name = "parameter_3" + original_name = "layer_norm_197.w_0" shape = [1024] dtype = "float32" min_val = float("0.264642") @@ -44,6 +48,7 @@ class Program_weight_tensor_parameter_3: class Program_weight_tensor_parameter_4: name = "parameter_4" + original_name = "layer_norm_196.b_0" shape = [1024] dtype = "float32" min_val = float("-0.335826") @@ -55,6 +60,7 @@ class Program_weight_tensor_parameter_4: class Program_weight_tensor_parameter_5: name = "parameter_5" + original_name = "layer_norm_196.w_0" shape = [1024] dtype = "float32" min_val = float("0.785752") @@ -66,6 +72,7 @@ class Program_weight_tensor_parameter_5: class Program_weight_tensor_parameter_6: name = "parameter_6" + original_name = "linear_580.b_0" shape = [1024] dtype = "float32" min_val = float("-0.109943") @@ -77,6 +84,7 @@ class Program_weight_tensor_parameter_6: class Program_weight_tensor_parameter_7: name = "parameter_7" + original_name = "linear_580.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-0.29874") @@ -88,6 +96,7 @@ class Program_weight_tensor_parameter_7: class Program_weight_tensor_parameter_8: name = "parameter_8" + original_name = "linear_579.b_0" shape = [4096] dtype = "float32" min_val = float("-0.358759") @@ -99,6 +108,7 @@ class Program_weight_tensor_parameter_8: class Program_weight_tensor_parameter_9: name = "parameter_9" + original_name = "linear_579.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.288226") @@ -110,6 +120,7 @@ class Program_weight_tensor_parameter_9: class Program_weight_tensor_parameter_10: name = "parameter_10" + original_name = "linear_578.b_0" shape = [1024] dtype = "float32" min_val = float("-0.144674") @@ -121,6 +132,7 @@ class Program_weight_tensor_parameter_10: class Program_weight_tensor_parameter_11: name = "parameter_11" + original_name = "linear_578.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.161795") @@ -132,6 +144,7 @@ class Program_weight_tensor_parameter_11: class Program_weight_tensor_parameter_12: name = "parameter_12" + original_name = "linear_577.b_0" shape = [1024] dtype = "float32" min_val = float("-0.12454") @@ -143,6 +156,7 @@ class Program_weight_tensor_parameter_12: class Program_weight_tensor_parameter_13: name = "parameter_13" + original_name = "linear_577.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.184797") @@ -154,6 +168,7 @@ class Program_weight_tensor_parameter_13: class Program_weight_tensor_parameter_14: name = "parameter_14" + original_name = "linear_576.b_0" shape = [1024] dtype = "float32" min_val = float("-19.9997") @@ -165,6 +180,7 @@ class Program_weight_tensor_parameter_14: class Program_weight_tensor_parameter_15: name = "parameter_15" + original_name = "linear_576.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.267094") @@ -176,6 +192,7 @@ class Program_weight_tensor_parameter_15: class Program_weight_tensor_parameter_16: name = "parameter_16" + original_name = "linear_575.b_0" shape = [1024] dtype = "float32" min_val = float("-0.480395") @@ -187,6 +204,7 @@ class Program_weight_tensor_parameter_16: class Program_weight_tensor_parameter_17: name = "parameter_17" + original_name = "linear_575.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.635012") @@ -198,6 +216,7 @@ class Program_weight_tensor_parameter_17: class Program_weight_tensor_parameter_18: name = "parameter_18" + original_name = "layer_norm_195.b_0" shape = [1024] dtype = "float32" min_val = float("-0.320861") @@ -209,6 +228,7 @@ class Program_weight_tensor_parameter_18: class Program_weight_tensor_parameter_19: name = "parameter_19" + original_name = "layer_norm_195.w_0" shape = [1024] dtype = "float32" min_val = float("0.514658") @@ -220,6 +240,7 @@ class Program_weight_tensor_parameter_19: class Program_weight_tensor_parameter_20: name = "parameter_20" + original_name = "layer_norm_194.b_0" shape = [1024] dtype = "float32" min_val = float("-0.495663") @@ -231,6 +252,7 @@ class Program_weight_tensor_parameter_20: class Program_weight_tensor_parameter_21: name = "parameter_21" + original_name = "layer_norm_194.w_0" shape = [1024] dtype = "float32" min_val = float("0.876996") @@ -242,6 +264,7 @@ class Program_weight_tensor_parameter_21: class Program_weight_tensor_parameter_22: name = "parameter_22" + original_name = "linear_574.b_0" shape = [1024] dtype = "float32" min_val = float("-0.239644") @@ -253,6 +276,7 @@ class Program_weight_tensor_parameter_22: class Program_weight_tensor_parameter_23: name = "parameter_23" + original_name = "linear_574.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-0.301115") @@ -264,6 +288,7 @@ class Program_weight_tensor_parameter_23: class Program_weight_tensor_parameter_24: name = "parameter_24" + original_name = "linear_573.b_0" shape = [4096] dtype = "float32" min_val = float("-0.249649") @@ -275,6 +300,7 @@ class Program_weight_tensor_parameter_24: class Program_weight_tensor_parameter_25: name = "parameter_25" + original_name = "linear_573.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.234421") @@ -286,6 +312,7 @@ class Program_weight_tensor_parameter_25: class Program_weight_tensor_parameter_26: name = "parameter_26" + original_name = "linear_572.b_0" shape = [1024] dtype = "float32" min_val = float("-0.13388") @@ -297,6 +324,7 @@ class Program_weight_tensor_parameter_26: class Program_weight_tensor_parameter_27: name = "parameter_27" + original_name = "linear_572.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.150447") @@ -308,6 +336,7 @@ class Program_weight_tensor_parameter_27: class Program_weight_tensor_parameter_28: name = "parameter_28" + original_name = "linear_571.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0590133") @@ -319,6 +348,7 @@ class Program_weight_tensor_parameter_28: class Program_weight_tensor_parameter_29: name = "parameter_29" + original_name = "linear_571.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.169408") @@ -330,6 +360,7 @@ class Program_weight_tensor_parameter_29: class Program_weight_tensor_parameter_30: name = "parameter_30" + original_name = "linear_570.b_0" shape = [1024] dtype = "float32" min_val = float("-23.9611") @@ -341,6 +372,7 @@ class Program_weight_tensor_parameter_30: class Program_weight_tensor_parameter_31: name = "parameter_31" + original_name = "linear_570.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.287061") @@ -352,6 +384,7 @@ class Program_weight_tensor_parameter_31: class Program_weight_tensor_parameter_32: name = "parameter_32" + original_name = "linear_569.b_0" shape = [1024] dtype = "float32" min_val = float("-0.517494") @@ -363,6 +396,7 @@ class Program_weight_tensor_parameter_32: class Program_weight_tensor_parameter_33: name = "parameter_33" + original_name = "linear_569.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.587797") @@ -374,6 +408,7 @@ class Program_weight_tensor_parameter_33: class Program_weight_tensor_parameter_34: name = "parameter_34" + original_name = "layer_norm_193.b_0" shape = [1024] dtype = "float32" min_val = float("-0.356986") @@ -385,6 +420,7 @@ class Program_weight_tensor_parameter_34: class Program_weight_tensor_parameter_35: name = "parameter_35" + original_name = "layer_norm_193.w_0" shape = [1024] dtype = "float32" min_val = float("0.540379") @@ -396,6 +432,7 @@ class Program_weight_tensor_parameter_35: class Program_weight_tensor_parameter_36: name = "parameter_36" + original_name = "layer_norm_192.b_0" shape = [1024] dtype = "float32" min_val = float("-1.16107") @@ -407,6 +444,7 @@ class Program_weight_tensor_parameter_36: class Program_weight_tensor_parameter_37: name = "parameter_37" + original_name = "layer_norm_192.w_0" shape = [1024] dtype = "float32" min_val = float("0.895862") @@ -418,6 +456,7 @@ class Program_weight_tensor_parameter_37: class Program_weight_tensor_parameter_38: name = "parameter_38" + original_name = "linear_568.b_0" shape = [1024] dtype = "float32" min_val = float("-0.300437") @@ -429,6 +468,7 @@ class Program_weight_tensor_parameter_38: class Program_weight_tensor_parameter_39: name = "parameter_39" + original_name = "linear_568.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-0.534301") @@ -440,6 +480,7 @@ class Program_weight_tensor_parameter_39: class Program_weight_tensor_parameter_40: name = "parameter_40" + original_name = "linear_567.b_0" shape = [4096] dtype = "float32" min_val = float("-0.167803") @@ -451,6 +492,7 @@ class Program_weight_tensor_parameter_40: class Program_weight_tensor_parameter_41: name = "parameter_41" + original_name = "linear_567.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.42634") @@ -462,6 +504,7 @@ class Program_weight_tensor_parameter_41: class Program_weight_tensor_parameter_42: name = "parameter_42" + original_name = "linear_566.b_0" shape = [1024] dtype = "float32" min_val = float("-0.186638") @@ -473,6 +516,7 @@ class Program_weight_tensor_parameter_42: class Program_weight_tensor_parameter_43: name = "parameter_43" + original_name = "linear_566.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.271789") @@ -484,6 +528,7 @@ class Program_weight_tensor_parameter_43: class Program_weight_tensor_parameter_44: name = "parameter_44" + original_name = "linear_565.b_0" shape = [1024] dtype = "float32" min_val = float("-0.162104") @@ -495,6 +540,7 @@ class Program_weight_tensor_parameter_44: class Program_weight_tensor_parameter_45: name = "parameter_45" + original_name = "linear_565.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.155829") @@ -506,6 +552,7 @@ class Program_weight_tensor_parameter_45: class Program_weight_tensor_parameter_46: name = "parameter_46" + original_name = "linear_564.b_0" shape = [1024] dtype = "float32" min_val = float("-17.5883") @@ -517,6 +564,7 @@ class Program_weight_tensor_parameter_46: class Program_weight_tensor_parameter_47: name = "parameter_47" + original_name = "linear_564.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.222615") @@ -528,6 +576,7 @@ class Program_weight_tensor_parameter_47: class Program_weight_tensor_parameter_48: name = "parameter_48" + original_name = "linear_563.b_0" shape = [1024] dtype = "float32" min_val = float("-0.325444") @@ -539,6 +588,7 @@ class Program_weight_tensor_parameter_48: class Program_weight_tensor_parameter_49: name = "parameter_49" + original_name = "linear_563.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.489128") @@ -550,6 +600,7 @@ class Program_weight_tensor_parameter_49: class Program_weight_tensor_parameter_50: name = "parameter_50" + original_name = "layer_norm_191.b_0" shape = [1024] dtype = "float32" min_val = float("-0.478087") @@ -561,6 +612,7 @@ class Program_weight_tensor_parameter_50: class Program_weight_tensor_parameter_51: name = "parameter_51" + original_name = "layer_norm_191.w_0" shape = [1024] dtype = "float32" min_val = float("0.417311") @@ -572,6 +624,7 @@ class Program_weight_tensor_parameter_51: class Program_weight_tensor_parameter_52: name = "parameter_52" + original_name = "layer_norm_190.b_0" shape = [1024] dtype = "float32" min_val = float("-1.36018") @@ -583,6 +636,7 @@ class Program_weight_tensor_parameter_52: class Program_weight_tensor_parameter_53: name = "parameter_53" + original_name = "layer_norm_190.w_0" shape = [1024] dtype = "float32" min_val = float("0.895814") @@ -594,6 +648,7 @@ class Program_weight_tensor_parameter_53: class Program_weight_tensor_parameter_54: name = "parameter_54" + original_name = "linear_562.b_0" shape = [1024] dtype = "float32" min_val = float("-0.40168") @@ -605,6 +660,7 @@ class Program_weight_tensor_parameter_54: class Program_weight_tensor_parameter_55: name = "parameter_55" + original_name = "linear_562.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-0.45261") @@ -616,6 +672,7 @@ class Program_weight_tensor_parameter_55: class Program_weight_tensor_parameter_56: name = "parameter_56" + original_name = "linear_561.b_0" shape = [4096] dtype = "float32" min_val = float("-0.158579") @@ -627,6 +684,7 @@ class Program_weight_tensor_parameter_56: class Program_weight_tensor_parameter_57: name = "parameter_57" + original_name = "linear_561.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.63838") @@ -638,6 +696,7 @@ class Program_weight_tensor_parameter_57: class Program_weight_tensor_parameter_58: name = "parameter_58" + original_name = "linear_560.b_0" shape = [1024] dtype = "float32" min_val = float("-0.180014") @@ -649,6 +708,7 @@ class Program_weight_tensor_parameter_58: class Program_weight_tensor_parameter_59: name = "parameter_59" + original_name = "linear_560.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.235644") @@ -660,6 +720,7 @@ class Program_weight_tensor_parameter_59: class Program_weight_tensor_parameter_60: name = "parameter_60" + original_name = "linear_559.b_0" shape = [1024] dtype = "float32" min_val = float("-0.121848") @@ -671,6 +732,7 @@ class Program_weight_tensor_parameter_60: class Program_weight_tensor_parameter_61: name = "parameter_61" + original_name = "linear_559.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.169137") @@ -682,6 +744,7 @@ class Program_weight_tensor_parameter_61: class Program_weight_tensor_parameter_62: name = "parameter_62" + original_name = "linear_558.b_0" shape = [1024] dtype = "float32" min_val = float("-15.7793") @@ -693,6 +756,7 @@ class Program_weight_tensor_parameter_62: class Program_weight_tensor_parameter_63: name = "parameter_63" + original_name = "linear_558.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.821348") @@ -704,6 +768,7 @@ class Program_weight_tensor_parameter_63: class Program_weight_tensor_parameter_64: name = "parameter_64" + original_name = "linear_557.b_0" shape = [1024] dtype = "float32" min_val = float("-0.294335") @@ -715,6 +780,7 @@ class Program_weight_tensor_parameter_64: class Program_weight_tensor_parameter_65: name = "parameter_65" + original_name = "linear_557.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.369295") @@ -726,6 +792,7 @@ class Program_weight_tensor_parameter_65: class Program_weight_tensor_parameter_66: name = "parameter_66" + original_name = "layer_norm_189.b_0" shape = [1024] dtype = "float32" min_val = float("-0.589924") @@ -737,6 +804,7 @@ class Program_weight_tensor_parameter_66: class Program_weight_tensor_parameter_67: name = "parameter_67" + original_name = "layer_norm_189.w_0" shape = [1024] dtype = "float32" min_val = float("0.373627") @@ -748,6 +816,7 @@ class Program_weight_tensor_parameter_67: class Program_weight_tensor_parameter_68: name = "parameter_68" + original_name = "layer_norm_188.b_0" shape = [1024] dtype = "float32" min_val = float("-1.66511") @@ -759,6 +828,7 @@ class Program_weight_tensor_parameter_68: class Program_weight_tensor_parameter_69: name = "parameter_69" + original_name = "layer_norm_188.w_0" shape = [1024] dtype = "float32" min_val = float("0.872851") @@ -770,6 +840,7 @@ class Program_weight_tensor_parameter_69: class Program_weight_tensor_parameter_70: name = "parameter_70" + original_name = "linear_556.b_0" shape = [1024] dtype = "float32" min_val = float("-0.84757") @@ -781,6 +852,7 @@ class Program_weight_tensor_parameter_70: class Program_weight_tensor_parameter_71: name = "parameter_71" + original_name = "linear_556.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-0.720292") @@ -792,6 +864,7 @@ class Program_weight_tensor_parameter_71: class Program_weight_tensor_parameter_72: name = "parameter_72" + original_name = "linear_555.b_0" shape = [4096] dtype = "float32" min_val = float("-0.183701") @@ -803,6 +876,7 @@ class Program_weight_tensor_parameter_72: class Program_weight_tensor_parameter_73: name = "parameter_73" + original_name = "linear_555.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.595906") @@ -814,6 +888,7 @@ class Program_weight_tensor_parameter_73: class Program_weight_tensor_parameter_74: name = "parameter_74" + original_name = "linear_554.b_0" shape = [1024] dtype = "float32" min_val = float("-0.114156") @@ -825,6 +900,7 @@ class Program_weight_tensor_parameter_74: class Program_weight_tensor_parameter_75: name = "parameter_75" + original_name = "linear_554.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.271976") @@ -836,6 +912,7 @@ class Program_weight_tensor_parameter_75: class Program_weight_tensor_parameter_76: name = "parameter_76" + original_name = "linear_553.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0721884") @@ -847,6 +924,7 @@ class Program_weight_tensor_parameter_76: class Program_weight_tensor_parameter_77: name = "parameter_77" + original_name = "linear_553.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.145728") @@ -858,6 +936,7 @@ class Program_weight_tensor_parameter_77: class Program_weight_tensor_parameter_78: name = "parameter_78" + original_name = "linear_552.b_0" shape = [1024] dtype = "float32" min_val = float("-19.7769") @@ -869,6 +948,7 @@ class Program_weight_tensor_parameter_78: class Program_weight_tensor_parameter_79: name = "parameter_79" + original_name = "linear_552.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.800982") @@ -880,6 +960,7 @@ class Program_weight_tensor_parameter_79: class Program_weight_tensor_parameter_80: name = "parameter_80" + original_name = "linear_551.b_0" shape = [1024] dtype = "float32" min_val = float("-0.34265") @@ -891,6 +972,7 @@ class Program_weight_tensor_parameter_80: class Program_weight_tensor_parameter_81: name = "parameter_81" + original_name = "linear_551.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.264385") @@ -902,6 +984,7 @@ class Program_weight_tensor_parameter_81: class Program_weight_tensor_parameter_82: name = "parameter_82" + original_name = "layer_norm_187.b_0" shape = [1024] dtype = "float32" min_val = float("-0.763826") @@ -913,6 +996,7 @@ class Program_weight_tensor_parameter_82: class Program_weight_tensor_parameter_83: name = "parameter_83" + original_name = "layer_norm_187.w_0" shape = [1024] dtype = "float32" min_val = float("0.587987") @@ -924,6 +1008,7 @@ class Program_weight_tensor_parameter_83: class Program_weight_tensor_parameter_84: name = "parameter_84" + original_name = "layer_norm_186.b_0" shape = [1024] dtype = "float32" min_val = float("-1.7926") @@ -935,6 +1020,7 @@ class Program_weight_tensor_parameter_84: class Program_weight_tensor_parameter_85: name = "parameter_85" + original_name = "layer_norm_186.w_0" shape = [1024] dtype = "float32" min_val = float("0.870486") @@ -946,6 +1032,7 @@ class Program_weight_tensor_parameter_85: class Program_weight_tensor_parameter_86: name = "parameter_86" + original_name = "linear_550.b_0" shape = [1024] dtype = "float32" min_val = float("-1.05862") @@ -957,6 +1044,7 @@ class Program_weight_tensor_parameter_86: class Program_weight_tensor_parameter_87: name = "parameter_87" + original_name = "linear_550.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-0.825241") @@ -968,6 +1056,7 @@ class Program_weight_tensor_parameter_87: class Program_weight_tensor_parameter_88: name = "parameter_88" + original_name = "linear_549.b_0" shape = [4096] dtype = "float32" min_val = float("-0.175247") @@ -979,6 +1068,7 @@ class Program_weight_tensor_parameter_88: class Program_weight_tensor_parameter_89: name = "parameter_89" + original_name = "linear_549.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.365279") @@ -990,6 +1080,7 @@ class Program_weight_tensor_parameter_89: class Program_weight_tensor_parameter_90: name = "parameter_90" + original_name = "linear_548.b_0" shape = [1024] dtype = "float32" min_val = float("-0.140304") @@ -1001,6 +1092,7 @@ class Program_weight_tensor_parameter_90: class Program_weight_tensor_parameter_91: name = "parameter_91" + original_name = "linear_548.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.172214") @@ -1012,6 +1104,7 @@ class Program_weight_tensor_parameter_91: class Program_weight_tensor_parameter_92: name = "parameter_92" + original_name = "linear_547.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0943922") @@ -1023,6 +1116,7 @@ class Program_weight_tensor_parameter_92: class Program_weight_tensor_parameter_93: name = "parameter_93" + original_name = "linear_547.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.15612") @@ -1034,6 +1128,7 @@ class Program_weight_tensor_parameter_93: class Program_weight_tensor_parameter_94: name = "parameter_94" + original_name = "linear_546.b_0" shape = [1024] dtype = "float32" min_val = float("-11.7468") @@ -1045,6 +1140,7 @@ class Program_weight_tensor_parameter_94: class Program_weight_tensor_parameter_95: name = "parameter_95" + original_name = "linear_546.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.512125") @@ -1056,6 +1152,7 @@ class Program_weight_tensor_parameter_95: class Program_weight_tensor_parameter_96: name = "parameter_96" + original_name = "linear_545.b_0" shape = [1024] dtype = "float32" min_val = float("-0.271217") @@ -1067,6 +1164,7 @@ class Program_weight_tensor_parameter_96: class Program_weight_tensor_parameter_97: name = "parameter_97" + original_name = "linear_545.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.318624") @@ -1078,6 +1176,7 @@ class Program_weight_tensor_parameter_97: class Program_weight_tensor_parameter_98: name = "parameter_98" + original_name = "layer_norm_185.b_0" shape = [1024] dtype = "float32" min_val = float("-0.832476") @@ -1089,6 +1188,7 @@ class Program_weight_tensor_parameter_98: class Program_weight_tensor_parameter_99: name = "parameter_99" + original_name = "layer_norm_185.w_0" shape = [1024] dtype = "float32" min_val = float("0.619347") @@ -1100,6 +1200,7 @@ class Program_weight_tensor_parameter_99: class Program_weight_tensor_parameter_100: name = "parameter_100" + original_name = "layer_norm_184.b_0" shape = [1024] dtype = "float32" min_val = float("-1.85695") @@ -1111,6 +1212,7 @@ class Program_weight_tensor_parameter_100: class Program_weight_tensor_parameter_101: name = "parameter_101" + original_name = "layer_norm_184.w_0" shape = [1024] dtype = "float32" min_val = float("0.872845") @@ -1122,6 +1224,7 @@ class Program_weight_tensor_parameter_101: class Program_weight_tensor_parameter_102: name = "parameter_102" + original_name = "linear_544.b_0" shape = [1024] dtype = "float32" min_val = float("-1.19279") @@ -1133,6 +1236,7 @@ class Program_weight_tensor_parameter_102: class Program_weight_tensor_parameter_103: name = "parameter_103" + original_name = "linear_544.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-0.670875") @@ -1144,6 +1248,7 @@ class Program_weight_tensor_parameter_103: class Program_weight_tensor_parameter_104: name = "parameter_104" + original_name = "linear_543.b_0" shape = [4096] dtype = "float32" min_val = float("-0.158382") @@ -1155,6 +1260,7 @@ class Program_weight_tensor_parameter_104: class Program_weight_tensor_parameter_105: name = "parameter_105" + original_name = "linear_543.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.301033") @@ -1166,6 +1272,7 @@ class Program_weight_tensor_parameter_105: class Program_weight_tensor_parameter_106: name = "parameter_106" + original_name = "linear_542.b_0" shape = [1024] dtype = "float32" min_val = float("-0.185905") @@ -1177,6 +1284,7 @@ class Program_weight_tensor_parameter_106: class Program_weight_tensor_parameter_107: name = "parameter_107" + original_name = "linear_542.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.192572") @@ -1188,6 +1296,7 @@ class Program_weight_tensor_parameter_107: class Program_weight_tensor_parameter_108: name = "parameter_108" + original_name = "linear_541.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0722268") @@ -1199,6 +1308,7 @@ class Program_weight_tensor_parameter_108: class Program_weight_tensor_parameter_109: name = "parameter_109" + original_name = "linear_541.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.157104") @@ -1210,6 +1320,7 @@ class Program_weight_tensor_parameter_109: class Program_weight_tensor_parameter_110: name = "parameter_110" + original_name = "linear_540.b_0" shape = [1024] dtype = "float32" min_val = float("-10.0645") @@ -1221,6 +1332,7 @@ class Program_weight_tensor_parameter_110: class Program_weight_tensor_parameter_111: name = "parameter_111" + original_name = "linear_540.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.469493") @@ -1232,6 +1344,7 @@ class Program_weight_tensor_parameter_111: class Program_weight_tensor_parameter_112: name = "parameter_112" + original_name = "linear_539.b_0" shape = [1024] dtype = "float32" min_val = float("-0.314037") @@ -1243,6 +1356,7 @@ class Program_weight_tensor_parameter_112: class Program_weight_tensor_parameter_113: name = "parameter_113" + original_name = "linear_539.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.255523") @@ -1254,6 +1368,7 @@ class Program_weight_tensor_parameter_113: class Program_weight_tensor_parameter_114: name = "parameter_114" + original_name = "layer_norm_183.b_0" shape = [1024] dtype = "float32" min_val = float("-0.76991") @@ -1265,6 +1380,7 @@ class Program_weight_tensor_parameter_114: class Program_weight_tensor_parameter_115: name = "parameter_115" + original_name = "layer_norm_183.w_0" shape = [1024] dtype = "float32" min_val = float("0.626067") @@ -1276,6 +1392,7 @@ class Program_weight_tensor_parameter_115: class Program_weight_tensor_parameter_116: name = "parameter_116" + original_name = "layer_norm_182.b_0" shape = [1024] dtype = "float32" min_val = float("-1.56406") @@ -1287,6 +1404,7 @@ class Program_weight_tensor_parameter_116: class Program_weight_tensor_parameter_117: name = "parameter_117" + original_name = "layer_norm_182.w_0" shape = [1024] dtype = "float32" min_val = float("0.856579") @@ -1298,6 +1416,7 @@ class Program_weight_tensor_parameter_117: class Program_weight_tensor_parameter_118: name = "parameter_118" + original_name = "linear_538.b_0" shape = [1024] dtype = "float32" min_val = float("-1.15514") @@ -1309,6 +1428,7 @@ class Program_weight_tensor_parameter_118: class Program_weight_tensor_parameter_119: name = "parameter_119" + original_name = "linear_538.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-1.06826") @@ -1320,6 +1440,7 @@ class Program_weight_tensor_parameter_119: class Program_weight_tensor_parameter_120: name = "parameter_120" + original_name = "linear_537.b_0" shape = [4096] dtype = "float32" min_val = float("-0.135786") @@ -1331,6 +1452,7 @@ class Program_weight_tensor_parameter_120: class Program_weight_tensor_parameter_121: name = "parameter_121" + original_name = "linear_537.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.280879") @@ -1342,6 +1464,7 @@ class Program_weight_tensor_parameter_121: class Program_weight_tensor_parameter_122: name = "parameter_122" + original_name = "linear_536.b_0" shape = [1024] dtype = "float32" min_val = float("-0.145572") @@ -1353,6 +1476,7 @@ class Program_weight_tensor_parameter_122: class Program_weight_tensor_parameter_123: name = "parameter_123" + original_name = "linear_536.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.217586") @@ -1364,6 +1488,7 @@ class Program_weight_tensor_parameter_123: class Program_weight_tensor_parameter_124: name = "parameter_124" + original_name = "linear_535.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0559941") @@ -1375,6 +1500,7 @@ class Program_weight_tensor_parameter_124: class Program_weight_tensor_parameter_125: name = "parameter_125" + original_name = "linear_535.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.151624") @@ -1386,6 +1512,7 @@ class Program_weight_tensor_parameter_125: class Program_weight_tensor_parameter_126: name = "parameter_126" + original_name = "linear_534.b_0" shape = [1024] dtype = "float32" min_val = float("-7.49084") @@ -1397,6 +1524,7 @@ class Program_weight_tensor_parameter_126: class Program_weight_tensor_parameter_127: name = "parameter_127" + original_name = "linear_534.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.572659") @@ -1408,6 +1536,7 @@ class Program_weight_tensor_parameter_127: class Program_weight_tensor_parameter_128: name = "parameter_128" + original_name = "linear_533.b_0" shape = [1024] dtype = "float32" min_val = float("-0.242108") @@ -1419,6 +1548,7 @@ class Program_weight_tensor_parameter_128: class Program_weight_tensor_parameter_129: name = "parameter_129" + original_name = "linear_533.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.267373") @@ -1430,6 +1560,7 @@ class Program_weight_tensor_parameter_129: class Program_weight_tensor_parameter_130: name = "parameter_130" + original_name = "layer_norm_181.b_0" shape = [1024] dtype = "float32" min_val = float("-0.689925") @@ -1441,6 +1572,7 @@ class Program_weight_tensor_parameter_130: class Program_weight_tensor_parameter_131: name = "parameter_131" + original_name = "layer_norm_181.w_0" shape = [1024] dtype = "float32" min_val = float("0.665175") @@ -1452,6 +1584,7 @@ class Program_weight_tensor_parameter_131: class Program_weight_tensor_parameter_132: name = "parameter_132" + original_name = "layer_norm_180.b_0" shape = [1024] dtype = "float32" min_val = float("-0.910117") @@ -1463,6 +1596,7 @@ class Program_weight_tensor_parameter_132: class Program_weight_tensor_parameter_133: name = "parameter_133" + original_name = "layer_norm_180.w_0" shape = [1024] dtype = "float32" min_val = float("0.83987") @@ -1474,6 +1608,7 @@ class Program_weight_tensor_parameter_133: class Program_weight_tensor_parameter_134: name = "parameter_134" + original_name = "linear_532.b_0" shape = [1024] dtype = "float32" min_val = float("-1.23652") @@ -1485,6 +1620,7 @@ class Program_weight_tensor_parameter_134: class Program_weight_tensor_parameter_135: name = "parameter_135" + original_name = "linear_532.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-1.37841") @@ -1496,6 +1632,7 @@ class Program_weight_tensor_parameter_135: class Program_weight_tensor_parameter_136: name = "parameter_136" + original_name = "linear_531.b_0" shape = [4096] dtype = "float32" min_val = float("-0.149571") @@ -1507,6 +1644,7 @@ class Program_weight_tensor_parameter_136: class Program_weight_tensor_parameter_137: name = "parameter_137" + original_name = "linear_531.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.297557") @@ -1518,6 +1656,7 @@ class Program_weight_tensor_parameter_137: class Program_weight_tensor_parameter_138: name = "parameter_138" + original_name = "linear_530.b_0" shape = [1024] dtype = "float32" min_val = float("-0.147255") @@ -1529,6 +1668,7 @@ class Program_weight_tensor_parameter_138: class Program_weight_tensor_parameter_139: name = "parameter_139" + original_name = "linear_530.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.413871") @@ -1540,6 +1680,7 @@ class Program_weight_tensor_parameter_139: class Program_weight_tensor_parameter_140: name = "parameter_140" + original_name = "linear_529.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0513678") @@ -1551,6 +1692,7 @@ class Program_weight_tensor_parameter_140: class Program_weight_tensor_parameter_141: name = "parameter_141" + original_name = "linear_529.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.142769") @@ -1562,6 +1704,7 @@ class Program_weight_tensor_parameter_141: class Program_weight_tensor_parameter_142: name = "parameter_142" + original_name = "linear_528.b_0" shape = [1024] dtype = "float32" min_val = float("-12.6827") @@ -1573,6 +1716,7 @@ class Program_weight_tensor_parameter_142: class Program_weight_tensor_parameter_143: name = "parameter_143" + original_name = "linear_528.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.775072") @@ -1584,6 +1728,7 @@ class Program_weight_tensor_parameter_143: class Program_weight_tensor_parameter_144: name = "parameter_144" + original_name = "linear_527.b_0" shape = [1024] dtype = "float32" min_val = float("-0.186946") @@ -1595,6 +1740,7 @@ class Program_weight_tensor_parameter_144: class Program_weight_tensor_parameter_145: name = "parameter_145" + original_name = "linear_527.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.285869") @@ -1606,6 +1752,7 @@ class Program_weight_tensor_parameter_145: class Program_weight_tensor_parameter_146: name = "parameter_146" + original_name = "layer_norm_179.b_0" shape = [1024] dtype = "float32" min_val = float("-0.366065") @@ -1617,6 +1764,7 @@ class Program_weight_tensor_parameter_146: class Program_weight_tensor_parameter_147: name = "parameter_147" + original_name = "layer_norm_179.w_0" shape = [1024] dtype = "float32" min_val = float("0.737132") @@ -1628,6 +1776,7 @@ class Program_weight_tensor_parameter_147: class Program_weight_tensor_parameter_148: name = "parameter_148" + original_name = "layer_norm_178.b_0" shape = [1024] dtype = "float32" min_val = float("-0.517494") @@ -1639,6 +1788,7 @@ class Program_weight_tensor_parameter_148: class Program_weight_tensor_parameter_149: name = "parameter_149" + original_name = "layer_norm_178.w_0" shape = [1024] dtype = "float32" min_val = float("0.867884") @@ -1650,6 +1800,7 @@ class Program_weight_tensor_parameter_149: class Program_weight_tensor_parameter_150: name = "parameter_150" + original_name = "linear_526.b_0" shape = [1024] dtype = "float32" min_val = float("-1.29669") @@ -1661,6 +1812,7 @@ class Program_weight_tensor_parameter_150: class Program_weight_tensor_parameter_151: name = "parameter_151" + original_name = "linear_526.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-1.146") @@ -1672,6 +1824,7 @@ class Program_weight_tensor_parameter_151: class Program_weight_tensor_parameter_152: name = "parameter_152" + original_name = "linear_525.b_0" shape = [4096] dtype = "float32" min_val = float("-0.146528") @@ -1683,6 +1836,7 @@ class Program_weight_tensor_parameter_152: class Program_weight_tensor_parameter_153: name = "parameter_153" + original_name = "linear_525.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.215702") @@ -1694,6 +1848,7 @@ class Program_weight_tensor_parameter_153: class Program_weight_tensor_parameter_154: name = "parameter_154" + original_name = "linear_524.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0985054") @@ -1705,6 +1860,7 @@ class Program_weight_tensor_parameter_154: class Program_weight_tensor_parameter_155: name = "parameter_155" + original_name = "linear_524.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.261661") @@ -1716,6 +1872,7 @@ class Program_weight_tensor_parameter_155: class Program_weight_tensor_parameter_156: name = "parameter_156" + original_name = "linear_523.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0604655") @@ -1727,6 +1884,7 @@ class Program_weight_tensor_parameter_156: class Program_weight_tensor_parameter_157: name = "parameter_157" + original_name = "linear_523.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.163426") @@ -1738,6 +1896,7 @@ class Program_weight_tensor_parameter_157: class Program_weight_tensor_parameter_158: name = "parameter_158" + original_name = "linear_522.b_0" shape = [1024] dtype = "float32" min_val = float("-5.83798") @@ -1749,6 +1908,7 @@ class Program_weight_tensor_parameter_158: class Program_weight_tensor_parameter_159: name = "parameter_159" + original_name = "linear_522.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.285799") @@ -1760,6 +1920,7 @@ class Program_weight_tensor_parameter_159: class Program_weight_tensor_parameter_160: name = "parameter_160" + original_name = "linear_521.b_0" shape = [1024] dtype = "float32" min_val = float("-0.207403") @@ -1771,6 +1932,7 @@ class Program_weight_tensor_parameter_160: class Program_weight_tensor_parameter_161: name = "parameter_161" + original_name = "linear_521.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.27673") @@ -1782,6 +1944,7 @@ class Program_weight_tensor_parameter_161: class Program_weight_tensor_parameter_162: name = "parameter_162" + original_name = "layer_norm_177.b_0" shape = [1024] dtype = "float32" min_val = float("-0.143426") @@ -1793,6 +1956,7 @@ class Program_weight_tensor_parameter_162: class Program_weight_tensor_parameter_163: name = "parameter_163" + original_name = "layer_norm_177.w_0" shape = [1024] dtype = "float32" min_val = float("0.541709") @@ -1804,6 +1968,7 @@ class Program_weight_tensor_parameter_163: class Program_weight_tensor_parameter_164: name = "parameter_164" + original_name = "layer_norm_176.b_0" shape = [1024] dtype = "float32" min_val = float("-0.426408") @@ -1815,6 +1980,7 @@ class Program_weight_tensor_parameter_164: class Program_weight_tensor_parameter_165: name = "parameter_165" + original_name = "layer_norm_176.w_0" shape = [1024] dtype = "float32" min_val = float("0.85969") @@ -1826,6 +1992,7 @@ class Program_weight_tensor_parameter_165: class Program_weight_tensor_parameter_166: name = "parameter_166" + original_name = "linear_520.b_0" shape = [1024] dtype = "float32" min_val = float("-1.24671") @@ -1837,6 +2004,7 @@ class Program_weight_tensor_parameter_166: class Program_weight_tensor_parameter_167: name = "parameter_167" + original_name = "linear_520.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-0.789178") @@ -1848,6 +2016,7 @@ class Program_weight_tensor_parameter_167: class Program_weight_tensor_parameter_168: name = "parameter_168" + original_name = "linear_519.b_0" shape = [4096] dtype = "float32" min_val = float("-0.113231") @@ -1859,6 +2028,7 @@ class Program_weight_tensor_parameter_168: class Program_weight_tensor_parameter_169: name = "parameter_169" + original_name = "linear_519.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.216621") @@ -1870,6 +2040,7 @@ class Program_weight_tensor_parameter_169: class Program_weight_tensor_parameter_170: name = "parameter_170" + original_name = "linear_518.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0740375") @@ -1881,6 +2052,7 @@ class Program_weight_tensor_parameter_170: class Program_weight_tensor_parameter_171: name = "parameter_171" + original_name = "linear_518.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.20176") @@ -1892,6 +2064,7 @@ class Program_weight_tensor_parameter_171: class Program_weight_tensor_parameter_172: name = "parameter_172" + original_name = "linear_517.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0515267") @@ -1903,6 +2076,7 @@ class Program_weight_tensor_parameter_172: class Program_weight_tensor_parameter_173: name = "parameter_173" + original_name = "linear_517.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.183603") @@ -1914,6 +2088,7 @@ class Program_weight_tensor_parameter_173: class Program_weight_tensor_parameter_174: name = "parameter_174" + original_name = "linear_516.b_0" shape = [1024] dtype = "float32" min_val = float("-6.43269") @@ -1925,6 +2100,7 @@ class Program_weight_tensor_parameter_174: class Program_weight_tensor_parameter_175: name = "parameter_175" + original_name = "linear_516.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.267283") @@ -1936,6 +2112,7 @@ class Program_weight_tensor_parameter_175: class Program_weight_tensor_parameter_176: name = "parameter_176" + original_name = "linear_515.b_0" shape = [1024] dtype = "float32" min_val = float("-0.197745") @@ -1947,6 +2124,7 @@ class Program_weight_tensor_parameter_176: class Program_weight_tensor_parameter_177: name = "parameter_177" + original_name = "linear_515.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.240705") @@ -1958,6 +2136,7 @@ class Program_weight_tensor_parameter_177: class Program_weight_tensor_parameter_178: name = "parameter_178" + original_name = "layer_norm_175.b_0" shape = [1024] dtype = "float32" min_val = float("-0.213067") @@ -1969,6 +2148,7 @@ class Program_weight_tensor_parameter_178: class Program_weight_tensor_parameter_179: name = "parameter_179" + original_name = "layer_norm_175.w_0" shape = [1024] dtype = "float32" min_val = float("0.388813") @@ -1980,6 +2160,7 @@ class Program_weight_tensor_parameter_179: class Program_weight_tensor_parameter_180: name = "parameter_180" + original_name = "layer_norm_174.b_0" shape = [1024] dtype = "float32" min_val = float("-0.769902") @@ -1991,6 +2172,7 @@ class Program_weight_tensor_parameter_180: class Program_weight_tensor_parameter_181: name = "parameter_181" + original_name = "layer_norm_174.w_0" shape = [1024] dtype = "float32" min_val = float("0.865521") @@ -2002,6 +2184,7 @@ class Program_weight_tensor_parameter_181: class Program_weight_tensor_parameter_182: name = "parameter_182" + original_name = "linear_514.b_0" shape = [1024] dtype = "float32" min_val = float("-1.20712") @@ -2013,6 +2196,7 @@ class Program_weight_tensor_parameter_182: class Program_weight_tensor_parameter_183: name = "parameter_183" + original_name = "linear_514.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-0.563351") @@ -2024,6 +2208,7 @@ class Program_weight_tensor_parameter_183: class Program_weight_tensor_parameter_184: name = "parameter_184" + original_name = "linear_513.b_0" shape = [4096] dtype = "float32" min_val = float("-0.106842") @@ -2035,6 +2220,7 @@ class Program_weight_tensor_parameter_184: class Program_weight_tensor_parameter_185: name = "parameter_185" + original_name = "linear_513.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.241389") @@ -2046,6 +2232,7 @@ class Program_weight_tensor_parameter_185: class Program_weight_tensor_parameter_186: name = "parameter_186" + original_name = "linear_512.b_0" shape = [1024] dtype = "float32" min_val = float("-0.11879") @@ -2057,6 +2244,7 @@ class Program_weight_tensor_parameter_186: class Program_weight_tensor_parameter_187: name = "parameter_187" + original_name = "linear_512.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.175099") @@ -2068,6 +2256,7 @@ class Program_weight_tensor_parameter_187: class Program_weight_tensor_parameter_188: name = "parameter_188" + original_name = "linear_511.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0507553") @@ -2079,6 +2268,7 @@ class Program_weight_tensor_parameter_188: class Program_weight_tensor_parameter_189: name = "parameter_189" + original_name = "linear_511.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.15569") @@ -2090,6 +2280,7 @@ class Program_weight_tensor_parameter_189: class Program_weight_tensor_parameter_190: name = "parameter_190" + original_name = "linear_510.b_0" shape = [1024] dtype = "float32" min_val = float("-4.65363") @@ -2101,6 +2292,7 @@ class Program_weight_tensor_parameter_190: class Program_weight_tensor_parameter_191: name = "parameter_191" + original_name = "linear_510.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.216398") @@ -2112,6 +2304,7 @@ class Program_weight_tensor_parameter_191: class Program_weight_tensor_parameter_192: name = "parameter_192" + original_name = "linear_509.b_0" shape = [1024] dtype = "float32" min_val = float("-0.223197") @@ -2123,6 +2316,7 @@ class Program_weight_tensor_parameter_192: class Program_weight_tensor_parameter_193: name = "parameter_193" + original_name = "linear_509.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.181706") @@ -2134,6 +2328,7 @@ class Program_weight_tensor_parameter_193: class Program_weight_tensor_parameter_194: name = "parameter_194" + original_name = "layer_norm_173.b_0" shape = [1024] dtype = "float32" min_val = float("-0.212204") @@ -2145,6 +2340,7 @@ class Program_weight_tensor_parameter_194: class Program_weight_tensor_parameter_195: name = "parameter_195" + original_name = "layer_norm_173.w_0" shape = [1024] dtype = "float32" min_val = float("0.236857") @@ -2156,6 +2352,7 @@ class Program_weight_tensor_parameter_195: class Program_weight_tensor_parameter_196: name = "parameter_196" + original_name = "layer_norm_172.b_0" shape = [1024] dtype = "float32" min_val = float("-1.07289") @@ -2167,6 +2364,7 @@ class Program_weight_tensor_parameter_196: class Program_weight_tensor_parameter_197: name = "parameter_197" + original_name = "layer_norm_172.w_0" shape = [1024] dtype = "float32" min_val = float("0.875759") @@ -2178,6 +2376,7 @@ class Program_weight_tensor_parameter_197: class Program_weight_tensor_parameter_198: name = "parameter_198" + original_name = "linear_508.b_0" shape = [1024] dtype = "float32" min_val = float("-0.829298") @@ -2189,6 +2388,7 @@ class Program_weight_tensor_parameter_198: class Program_weight_tensor_parameter_199: name = "parameter_199" + original_name = "linear_508.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-0.684168") @@ -2200,6 +2400,7 @@ class Program_weight_tensor_parameter_199: class Program_weight_tensor_parameter_200: name = "parameter_200" + original_name = "linear_507.b_0" shape = [4096] dtype = "float32" min_val = float("-0.0975706") @@ -2211,6 +2412,7 @@ class Program_weight_tensor_parameter_200: class Program_weight_tensor_parameter_201: name = "parameter_201" + original_name = "linear_507.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.222731") @@ -2222,6 +2424,7 @@ class Program_weight_tensor_parameter_201: class Program_weight_tensor_parameter_202: name = "parameter_202" + original_name = "linear_506.b_0" shape = [1024] dtype = "float32" min_val = float("-0.2304") @@ -2233,6 +2436,7 @@ class Program_weight_tensor_parameter_202: class Program_weight_tensor_parameter_203: name = "parameter_203" + original_name = "linear_506.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.150724") @@ -2244,6 +2448,7 @@ class Program_weight_tensor_parameter_203: class Program_weight_tensor_parameter_204: name = "parameter_204" + original_name = "linear_505.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0400974") @@ -2255,6 +2460,7 @@ class Program_weight_tensor_parameter_204: class Program_weight_tensor_parameter_205: name = "parameter_205" + original_name = "linear_505.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.151239") @@ -2266,6 +2472,7 @@ class Program_weight_tensor_parameter_205: class Program_weight_tensor_parameter_206: name = "parameter_206" + original_name = "linear_504.b_0" shape = [1024] dtype = "float32" min_val = float("-5.75159") @@ -2277,6 +2484,7 @@ class Program_weight_tensor_parameter_206: class Program_weight_tensor_parameter_207: name = "parameter_207" + original_name = "linear_504.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.301393") @@ -2288,6 +2496,7 @@ class Program_weight_tensor_parameter_207: class Program_weight_tensor_parameter_208: name = "parameter_208" + original_name = "linear_503.b_0" shape = [1024] dtype = "float32" min_val = float("-0.198194") @@ -2299,6 +2508,7 @@ class Program_weight_tensor_parameter_208: class Program_weight_tensor_parameter_209: name = "parameter_209" + original_name = "linear_503.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.188228") @@ -2310,6 +2520,7 @@ class Program_weight_tensor_parameter_209: class Program_weight_tensor_parameter_210: name = "parameter_210" + original_name = "layer_norm_171.b_0" shape = [1024] dtype = "float32" min_val = float("-0.244222") @@ -2321,6 +2532,7 @@ class Program_weight_tensor_parameter_210: class Program_weight_tensor_parameter_211: name = "parameter_211" + original_name = "layer_norm_171.w_0" shape = [1024] dtype = "float32" min_val = float("0.292687") @@ -2332,6 +2544,7 @@ class Program_weight_tensor_parameter_211: class Program_weight_tensor_parameter_212: name = "parameter_212" + original_name = "layer_norm_170.b_0" shape = [1024] dtype = "float32" min_val = float("-1.21241") @@ -2343,6 +2556,7 @@ class Program_weight_tensor_parameter_212: class Program_weight_tensor_parameter_213: name = "parameter_213" + original_name = "layer_norm_170.w_0" shape = [1024] dtype = "float32" min_val = float("0.847651") @@ -2354,6 +2568,7 @@ class Program_weight_tensor_parameter_213: class Program_weight_tensor_parameter_214: name = "parameter_214" + original_name = "linear_502.b_0" shape = [1024] dtype = "float32" min_val = float("-0.536311") @@ -2365,6 +2580,7 @@ class Program_weight_tensor_parameter_214: class Program_weight_tensor_parameter_215: name = "parameter_215" + original_name = "linear_502.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-0.695161") @@ -2376,6 +2592,7 @@ class Program_weight_tensor_parameter_215: class Program_weight_tensor_parameter_216: name = "parameter_216" + original_name = "linear_501.b_0" shape = [4096] dtype = "float32" min_val = float("-0.10222") @@ -2387,6 +2604,7 @@ class Program_weight_tensor_parameter_216: class Program_weight_tensor_parameter_217: name = "parameter_217" + original_name = "linear_501.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.303235") @@ -2398,6 +2616,7 @@ class Program_weight_tensor_parameter_217: class Program_weight_tensor_parameter_218: name = "parameter_218" + original_name = "linear_500.b_0" shape = [1024] dtype = "float32" min_val = float("-0.234338") @@ -2409,6 +2628,7 @@ class Program_weight_tensor_parameter_218: class Program_weight_tensor_parameter_219: name = "parameter_219" + original_name = "linear_500.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.227848") @@ -2420,6 +2640,7 @@ class Program_weight_tensor_parameter_219: class Program_weight_tensor_parameter_220: name = "parameter_220" + original_name = "linear_499.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0317948") @@ -2431,6 +2652,7 @@ class Program_weight_tensor_parameter_220: class Program_weight_tensor_parameter_221: name = "parameter_221" + original_name = "linear_499.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.173446") @@ -2442,6 +2664,7 @@ class Program_weight_tensor_parameter_221: class Program_weight_tensor_parameter_222: name = "parameter_222" + original_name = "linear_498.b_0" shape = [1024] dtype = "float32" min_val = float("-4.40276") @@ -2453,6 +2676,7 @@ class Program_weight_tensor_parameter_222: class Program_weight_tensor_parameter_223: name = "parameter_223" + original_name = "linear_498.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.510567") @@ -2464,6 +2688,7 @@ class Program_weight_tensor_parameter_223: class Program_weight_tensor_parameter_224: name = "parameter_224" + original_name = "linear_497.b_0" shape = [1024] dtype = "float32" min_val = float("-0.230167") @@ -2475,6 +2700,7 @@ class Program_weight_tensor_parameter_224: class Program_weight_tensor_parameter_225: name = "parameter_225" + original_name = "linear_497.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.207081") @@ -2486,6 +2712,7 @@ class Program_weight_tensor_parameter_225: class Program_weight_tensor_parameter_226: name = "parameter_226" + original_name = "layer_norm_169.b_0" shape = [1024] dtype = "float32" min_val = float("-0.269752") @@ -2497,6 +2724,7 @@ class Program_weight_tensor_parameter_226: class Program_weight_tensor_parameter_227: name = "parameter_227" + original_name = "layer_norm_169.w_0" shape = [1024] dtype = "float32" min_val = float("0.449855") @@ -2508,6 +2736,7 @@ class Program_weight_tensor_parameter_227: class Program_weight_tensor_parameter_228: name = "parameter_228" + original_name = "layer_norm_168.b_0" shape = [1024] dtype = "float32" min_val = float("-1.08149") @@ -2519,6 +2748,7 @@ class Program_weight_tensor_parameter_228: class Program_weight_tensor_parameter_229: name = "parameter_229" + original_name = "layer_norm_168.w_0" shape = [1024] dtype = "float32" min_val = float("0.891385") @@ -2530,6 +2760,7 @@ class Program_weight_tensor_parameter_229: class Program_weight_tensor_parameter_230: name = "parameter_230" + original_name = "linear_496.b_0" shape = [1024] dtype = "float32" min_val = float("-0.357139") @@ -2541,6 +2772,7 @@ class Program_weight_tensor_parameter_230: class Program_weight_tensor_parameter_231: name = "parameter_231" + original_name = "linear_496.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-1.26729") @@ -2552,6 +2784,7 @@ class Program_weight_tensor_parameter_231: class Program_weight_tensor_parameter_232: name = "parameter_232" + original_name = "linear_495.b_0" shape = [4096] dtype = "float32" min_val = float("-0.102234") @@ -2563,6 +2796,7 @@ class Program_weight_tensor_parameter_232: class Program_weight_tensor_parameter_233: name = "parameter_233" + original_name = "linear_495.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.293126") @@ -2574,6 +2808,7 @@ class Program_weight_tensor_parameter_233: class Program_weight_tensor_parameter_234: name = "parameter_234" + original_name = "linear_494.b_0" shape = [1024] dtype = "float32" min_val = float("-0.212722") @@ -2585,6 +2820,7 @@ class Program_weight_tensor_parameter_234: class Program_weight_tensor_parameter_235: name = "parameter_235" + original_name = "linear_494.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.228191") @@ -2596,6 +2832,7 @@ class Program_weight_tensor_parameter_235: class Program_weight_tensor_parameter_236: name = "parameter_236" + original_name = "linear_493.b_0" shape = [1024] dtype = "float32" min_val = float("-0.041069") @@ -2607,6 +2844,7 @@ class Program_weight_tensor_parameter_236: class Program_weight_tensor_parameter_237: name = "parameter_237" + original_name = "linear_493.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.185611") @@ -2618,6 +2856,7 @@ class Program_weight_tensor_parameter_237: class Program_weight_tensor_parameter_238: name = "parameter_238" + original_name = "linear_492.b_0" shape = [1024] dtype = "float32" min_val = float("-3.91512") @@ -2629,6 +2868,7 @@ class Program_weight_tensor_parameter_238: class Program_weight_tensor_parameter_239: name = "parameter_239" + original_name = "linear_492.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.371068") @@ -2640,6 +2880,7 @@ class Program_weight_tensor_parameter_239: class Program_weight_tensor_parameter_240: name = "parameter_240" + original_name = "linear_491.b_0" shape = [1024] dtype = "float32" min_val = float("-0.24378") @@ -2651,6 +2892,7 @@ class Program_weight_tensor_parameter_240: class Program_weight_tensor_parameter_241: name = "parameter_241" + original_name = "linear_491.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.220385") @@ -2662,6 +2904,7 @@ class Program_weight_tensor_parameter_241: class Program_weight_tensor_parameter_242: name = "parameter_242" + original_name = "layer_norm_167.b_0" shape = [1024] dtype = "float32" min_val = float("-0.235083") @@ -2673,6 +2916,7 @@ class Program_weight_tensor_parameter_242: class Program_weight_tensor_parameter_243: name = "parameter_243" + original_name = "layer_norm_167.w_0" shape = [1024] dtype = "float32" min_val = float("0.456456") @@ -2684,6 +2928,7 @@ class Program_weight_tensor_parameter_243: class Program_weight_tensor_parameter_244: name = "parameter_244" + original_name = "layer_norm_166.b_0" shape = [1024] dtype = "float32" min_val = float("-1.08267") @@ -2695,6 +2940,7 @@ class Program_weight_tensor_parameter_244: class Program_weight_tensor_parameter_245: name = "parameter_245" + original_name = "layer_norm_166.w_0" shape = [1024] dtype = "float32" min_val = float("0.862381") @@ -2706,6 +2952,7 @@ class Program_weight_tensor_parameter_245: class Program_weight_tensor_parameter_246: name = "parameter_246" + original_name = "linear_490.b_0" shape = [1024] dtype = "float32" min_val = float("-0.347398") @@ -2717,6 +2964,7 @@ class Program_weight_tensor_parameter_246: class Program_weight_tensor_parameter_247: name = "parameter_247" + original_name = "linear_490.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-1.68285") @@ -2728,6 +2976,7 @@ class Program_weight_tensor_parameter_247: class Program_weight_tensor_parameter_248: name = "parameter_248" + original_name = "linear_489.b_0" shape = [4096] dtype = "float32" min_val = float("-0.0965087") @@ -2739,6 +2988,7 @@ class Program_weight_tensor_parameter_248: class Program_weight_tensor_parameter_249: name = "parameter_249" + original_name = "linear_489.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.298111") @@ -2750,6 +3000,7 @@ class Program_weight_tensor_parameter_249: class Program_weight_tensor_parameter_250: name = "parameter_250" + original_name = "linear_488.b_0" shape = [1024] dtype = "float32" min_val = float("-0.139458") @@ -2761,6 +3012,7 @@ class Program_weight_tensor_parameter_250: class Program_weight_tensor_parameter_251: name = "parameter_251" + original_name = "linear_488.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.219875") @@ -2772,6 +3024,7 @@ class Program_weight_tensor_parameter_251: class Program_weight_tensor_parameter_252: name = "parameter_252" + original_name = "linear_487.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0448741") @@ -2783,6 +3036,7 @@ class Program_weight_tensor_parameter_252: class Program_weight_tensor_parameter_253: name = "parameter_253" + original_name = "linear_487.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.149013") @@ -2794,6 +3048,7 @@ class Program_weight_tensor_parameter_253: class Program_weight_tensor_parameter_254: name = "parameter_254" + original_name = "linear_486.b_0" shape = [1024] dtype = "float32" min_val = float("-5.02332") @@ -2805,6 +3060,7 @@ class Program_weight_tensor_parameter_254: class Program_weight_tensor_parameter_255: name = "parameter_255" + original_name = "linear_486.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.495549") @@ -2816,6 +3072,7 @@ class Program_weight_tensor_parameter_255: class Program_weight_tensor_parameter_256: name = "parameter_256" + original_name = "linear_485.b_0" shape = [1024] dtype = "float32" min_val = float("-0.193185") @@ -2827,6 +3084,7 @@ class Program_weight_tensor_parameter_256: class Program_weight_tensor_parameter_257: name = "parameter_257" + original_name = "linear_485.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.302834") @@ -2838,6 +3096,7 @@ class Program_weight_tensor_parameter_257: class Program_weight_tensor_parameter_258: name = "parameter_258" + original_name = "layer_norm_165.b_0" shape = [1024] dtype = "float32" min_val = float("-0.278043") @@ -2849,6 +3108,7 @@ class Program_weight_tensor_parameter_258: class Program_weight_tensor_parameter_259: name = "parameter_259" + original_name = "layer_norm_165.w_0" shape = [1024] dtype = "float32" min_val = float("0.484628") @@ -2860,6 +3120,7 @@ class Program_weight_tensor_parameter_259: class Program_weight_tensor_parameter_260: name = "parameter_260" + original_name = "layer_norm_164.b_0" shape = [1024] dtype = "float32" min_val = float("-1.16772") @@ -2871,6 +3132,7 @@ class Program_weight_tensor_parameter_260: class Program_weight_tensor_parameter_261: name = "parameter_261" + original_name = "layer_norm_164.w_0" shape = [1024] dtype = "float32" min_val = float("0.811656") @@ -2882,6 +3144,7 @@ class Program_weight_tensor_parameter_261: class Program_weight_tensor_parameter_262: name = "parameter_262" + original_name = "linear_484.b_0" shape = [1024] dtype = "float32" min_val = float("-0.365701") @@ -2893,6 +3156,7 @@ class Program_weight_tensor_parameter_262: class Program_weight_tensor_parameter_263: name = "parameter_263" + original_name = "linear_484.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-1.95919") @@ -2904,6 +3168,7 @@ class Program_weight_tensor_parameter_263: class Program_weight_tensor_parameter_264: name = "parameter_264" + original_name = "linear_483.b_0" shape = [4096] dtype = "float32" min_val = float("-0.0913696") @@ -2915,6 +3180,7 @@ class Program_weight_tensor_parameter_264: class Program_weight_tensor_parameter_265: name = "parameter_265" + original_name = "linear_483.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.228136") @@ -2926,6 +3192,7 @@ class Program_weight_tensor_parameter_265: class Program_weight_tensor_parameter_266: name = "parameter_266" + original_name = "linear_482.b_0" shape = [1024] dtype = "float32" min_val = float("-0.11789") @@ -2937,6 +3204,7 @@ class Program_weight_tensor_parameter_266: class Program_weight_tensor_parameter_267: name = "parameter_267" + original_name = "linear_482.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.315621") @@ -2948,6 +3216,7 @@ class Program_weight_tensor_parameter_267: class Program_weight_tensor_parameter_268: name = "parameter_268" + original_name = "linear_481.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0475283") @@ -2959,6 +3228,7 @@ class Program_weight_tensor_parameter_268: class Program_weight_tensor_parameter_269: name = "parameter_269" + original_name = "linear_481.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.281779") @@ -2970,6 +3240,7 @@ class Program_weight_tensor_parameter_269: class Program_weight_tensor_parameter_270: name = "parameter_270" + original_name = "linear_480.b_0" shape = [1024] dtype = "float32" min_val = float("-4.6523") @@ -2981,6 +3252,7 @@ class Program_weight_tensor_parameter_270: class Program_weight_tensor_parameter_271: name = "parameter_271" + original_name = "linear_480.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.298045") @@ -2992,6 +3264,7 @@ class Program_weight_tensor_parameter_271: class Program_weight_tensor_parameter_272: name = "parameter_272" + original_name = "linear_479.b_0" shape = [1024] dtype = "float32" min_val = float("-0.166597") @@ -3003,6 +3276,7 @@ class Program_weight_tensor_parameter_272: class Program_weight_tensor_parameter_273: name = "parameter_273" + original_name = "linear_479.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.286187") @@ -3014,6 +3288,7 @@ class Program_weight_tensor_parameter_273: class Program_weight_tensor_parameter_274: name = "parameter_274" + original_name = "layer_norm_163.b_0" shape = [1024] dtype = "float32" min_val = float("-0.381175") @@ -3025,6 +3300,7 @@ class Program_weight_tensor_parameter_274: class Program_weight_tensor_parameter_275: name = "parameter_275" + original_name = "layer_norm_163.w_0" shape = [1024] dtype = "float32" min_val = float("0.567789") @@ -3036,6 +3312,7 @@ class Program_weight_tensor_parameter_275: class Program_weight_tensor_parameter_276: name = "parameter_276" + original_name = "layer_norm_162.b_0" shape = [1024] dtype = "float32" min_val = float("-1.22967") @@ -3047,6 +3324,7 @@ class Program_weight_tensor_parameter_276: class Program_weight_tensor_parameter_277: name = "parameter_277" + original_name = "layer_norm_162.w_0" shape = [1024] dtype = "float32" min_val = float("0.823676") @@ -3058,6 +3336,7 @@ class Program_weight_tensor_parameter_277: class Program_weight_tensor_parameter_278: name = "parameter_278" + original_name = "linear_478.b_0" shape = [1024] dtype = "float32" min_val = float("-0.380564") @@ -3069,6 +3348,7 @@ class Program_weight_tensor_parameter_278: class Program_weight_tensor_parameter_279: name = "parameter_279" + original_name = "linear_478.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-1.88637") @@ -3080,6 +3360,7 @@ class Program_weight_tensor_parameter_279: class Program_weight_tensor_parameter_280: name = "parameter_280" + original_name = "linear_477.b_0" shape = [4096] dtype = "float32" min_val = float("-0.109606") @@ -3091,6 +3372,7 @@ class Program_weight_tensor_parameter_280: class Program_weight_tensor_parameter_281: name = "parameter_281" + original_name = "linear_477.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.211759") @@ -3102,6 +3384,7 @@ class Program_weight_tensor_parameter_281: class Program_weight_tensor_parameter_282: name = "parameter_282" + original_name = "linear_476.b_0" shape = [1024] dtype = "float32" min_val = float("-0.140376") @@ -3113,6 +3396,7 @@ class Program_weight_tensor_parameter_282: class Program_weight_tensor_parameter_283: name = "parameter_283" + original_name = "linear_476.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.383534") @@ -3124,6 +3408,7 @@ class Program_weight_tensor_parameter_283: class Program_weight_tensor_parameter_284: name = "parameter_284" + original_name = "linear_475.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0447765") @@ -3135,6 +3420,7 @@ class Program_weight_tensor_parameter_284: class Program_weight_tensor_parameter_285: name = "parameter_285" + original_name = "linear_475.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.153696") @@ -3146,6 +3432,7 @@ class Program_weight_tensor_parameter_285: class Program_weight_tensor_parameter_286: name = "parameter_286" + original_name = "linear_474.b_0" shape = [1024] dtype = "float32" min_val = float("-4.84247") @@ -3157,6 +3444,7 @@ class Program_weight_tensor_parameter_286: class Program_weight_tensor_parameter_287: name = "parameter_287" + original_name = "linear_474.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.303594") @@ -3168,6 +3456,7 @@ class Program_weight_tensor_parameter_287: class Program_weight_tensor_parameter_288: name = "parameter_288" + original_name = "linear_473.b_0" shape = [1024] dtype = "float32" min_val = float("-0.190927") @@ -3179,6 +3468,7 @@ class Program_weight_tensor_parameter_288: class Program_weight_tensor_parameter_289: name = "parameter_289" + original_name = "linear_473.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.278471") @@ -3190,6 +3480,7 @@ class Program_weight_tensor_parameter_289: class Program_weight_tensor_parameter_290: name = "parameter_290" + original_name = "layer_norm_161.b_0" shape = [1024] dtype = "float32" min_val = float("-0.654247") @@ -3201,6 +3492,7 @@ class Program_weight_tensor_parameter_290: class Program_weight_tensor_parameter_291: name = "parameter_291" + original_name = "layer_norm_161.w_0" shape = [1024] dtype = "float32" min_val = float("0.329359") @@ -3212,6 +3504,7 @@ class Program_weight_tensor_parameter_291: class Program_weight_tensor_parameter_292: name = "parameter_292" + original_name = "layer_norm_160.b_0" shape = [1024] dtype = "float32" min_val = float("-1.23746") @@ -3223,6 +3516,7 @@ class Program_weight_tensor_parameter_292: class Program_weight_tensor_parameter_293: name = "parameter_293" + original_name = "layer_norm_160.w_0" shape = [1024] dtype = "float32" min_val = float("0.846369") @@ -3234,6 +3528,7 @@ class Program_weight_tensor_parameter_293: class Program_weight_tensor_parameter_294: name = "parameter_294" + original_name = "linear_472.b_0" shape = [1024] dtype = "float32" min_val = float("-0.421182") @@ -3245,6 +3540,7 @@ class Program_weight_tensor_parameter_294: class Program_weight_tensor_parameter_295: name = "parameter_295" + original_name = "linear_472.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-3.41105") @@ -3256,6 +3552,7 @@ class Program_weight_tensor_parameter_295: class Program_weight_tensor_parameter_296: name = "parameter_296" + original_name = "linear_471.b_0" shape = [4096] dtype = "float32" min_val = float("-0.081601") @@ -3267,6 +3564,7 @@ class Program_weight_tensor_parameter_296: class Program_weight_tensor_parameter_297: name = "parameter_297" + original_name = "linear_471.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.212987") @@ -3278,6 +3576,7 @@ class Program_weight_tensor_parameter_297: class Program_weight_tensor_parameter_298: name = "parameter_298" + original_name = "linear_470.b_0" shape = [1024] dtype = "float32" min_val = float("-0.20923") @@ -3289,6 +3588,7 @@ class Program_weight_tensor_parameter_298: class Program_weight_tensor_parameter_299: name = "parameter_299" + original_name = "linear_470.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.571198") @@ -3300,6 +3600,7 @@ class Program_weight_tensor_parameter_299: class Program_weight_tensor_parameter_300: name = "parameter_300" + original_name = "linear_469.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0374245") @@ -3311,6 +3612,7 @@ class Program_weight_tensor_parameter_300: class Program_weight_tensor_parameter_301: name = "parameter_301" + original_name = "linear_469.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.142102") @@ -3322,6 +3624,7 @@ class Program_weight_tensor_parameter_301: class Program_weight_tensor_parameter_302: name = "parameter_302" + original_name = "linear_468.b_0" shape = [1024] dtype = "float32" min_val = float("-5.39367") @@ -3333,6 +3636,7 @@ class Program_weight_tensor_parameter_302: class Program_weight_tensor_parameter_303: name = "parameter_303" + original_name = "linear_468.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.254816") @@ -3344,6 +3648,7 @@ class Program_weight_tensor_parameter_303: class Program_weight_tensor_parameter_304: name = "parameter_304" + original_name = "linear_467.b_0" shape = [1024] dtype = "float32" min_val = float("-0.18004") @@ -3355,6 +3660,7 @@ class Program_weight_tensor_parameter_304: class Program_weight_tensor_parameter_305: name = "parameter_305" + original_name = "linear_467.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.310694") @@ -3366,6 +3672,7 @@ class Program_weight_tensor_parameter_305: class Program_weight_tensor_parameter_306: name = "parameter_306" + original_name = "layer_norm_159.b_0" shape = [1024] dtype = "float32" min_val = float("-0.540843") @@ -3377,6 +3684,7 @@ class Program_weight_tensor_parameter_306: class Program_weight_tensor_parameter_307: name = "parameter_307" + original_name = "layer_norm_159.w_0" shape = [1024] dtype = "float32" min_val = float("0.300119") @@ -3388,6 +3696,7 @@ class Program_weight_tensor_parameter_307: class Program_weight_tensor_parameter_308: name = "parameter_308" + original_name = "layer_norm_158.b_0" shape = [1024] dtype = "float32" min_val = float("-1.458") @@ -3399,6 +3708,7 @@ class Program_weight_tensor_parameter_308: class Program_weight_tensor_parameter_309: name = "parameter_309" + original_name = "layer_norm_158.w_0" shape = [1024] dtype = "float32" min_val = float("0.819062") @@ -3410,6 +3720,7 @@ class Program_weight_tensor_parameter_309: class Program_weight_tensor_parameter_310: name = "parameter_310" + original_name = "linear_466.b_0" shape = [1024] dtype = "float32" min_val = float("-0.580768") @@ -3421,6 +3732,7 @@ class Program_weight_tensor_parameter_310: class Program_weight_tensor_parameter_311: name = "parameter_311" + original_name = "linear_466.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-3.7447") @@ -3432,6 +3744,7 @@ class Program_weight_tensor_parameter_311: class Program_weight_tensor_parameter_312: name = "parameter_312" + original_name = "linear_465.b_0" shape = [4096] dtype = "float32" min_val = float("-0.0931396") @@ -3443,6 +3756,7 @@ class Program_weight_tensor_parameter_312: class Program_weight_tensor_parameter_313: name = "parameter_313" + original_name = "linear_465.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.239162") @@ -3454,6 +3768,7 @@ class Program_weight_tensor_parameter_313: class Program_weight_tensor_parameter_314: name = "parameter_314" + original_name = "linear_464.b_0" shape = [1024] dtype = "float32" min_val = float("-0.180578") @@ -3465,6 +3780,7 @@ class Program_weight_tensor_parameter_314: class Program_weight_tensor_parameter_315: name = "parameter_315" + original_name = "linear_464.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.516373") @@ -3476,6 +3792,7 @@ class Program_weight_tensor_parameter_315: class Program_weight_tensor_parameter_316: name = "parameter_316" + original_name = "linear_463.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0597177") @@ -3487,6 +3804,7 @@ class Program_weight_tensor_parameter_316: class Program_weight_tensor_parameter_317: name = "parameter_317" + original_name = "linear_463.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.137833") @@ -3498,6 +3816,7 @@ class Program_weight_tensor_parameter_317: class Program_weight_tensor_parameter_318: name = "parameter_318" + original_name = "linear_462.b_0" shape = [1024] dtype = "float32" min_val = float("-5.61411") @@ -3509,6 +3828,7 @@ class Program_weight_tensor_parameter_318: class Program_weight_tensor_parameter_319: name = "parameter_319" + original_name = "linear_462.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.298204") @@ -3520,6 +3840,7 @@ class Program_weight_tensor_parameter_319: class Program_weight_tensor_parameter_320: name = "parameter_320" + original_name = "linear_461.b_0" shape = [1024] dtype = "float32" min_val = float("-0.237091") @@ -3531,6 +3852,7 @@ class Program_weight_tensor_parameter_320: class Program_weight_tensor_parameter_321: name = "parameter_321" + original_name = "linear_461.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.221075") @@ -3542,6 +3864,7 @@ class Program_weight_tensor_parameter_321: class Program_weight_tensor_parameter_322: name = "parameter_322" + original_name = "layer_norm_157.b_0" shape = [1024] dtype = "float32" min_val = float("-0.788091") @@ -3553,6 +3876,7 @@ class Program_weight_tensor_parameter_322: class Program_weight_tensor_parameter_323: name = "parameter_323" + original_name = "layer_norm_157.w_0" shape = [1024] dtype = "float32" min_val = float("0.487084") @@ -3564,6 +3888,7 @@ class Program_weight_tensor_parameter_323: class Program_weight_tensor_parameter_324: name = "parameter_324" + original_name = "layer_norm_156.b_0" shape = [1024] dtype = "float32" min_val = float("-1.48341") @@ -3575,6 +3900,7 @@ class Program_weight_tensor_parameter_324: class Program_weight_tensor_parameter_325: name = "parameter_325" + original_name = "layer_norm_156.w_0" shape = [1024] dtype = "float32" min_val = float("0.848869") @@ -3586,6 +3912,7 @@ class Program_weight_tensor_parameter_325: class Program_weight_tensor_parameter_326: name = "parameter_326" + original_name = "linear_460.b_0" shape = [1024] dtype = "float32" min_val = float("-0.74917") @@ -3597,6 +3924,7 @@ class Program_weight_tensor_parameter_326: class Program_weight_tensor_parameter_327: name = "parameter_327" + original_name = "linear_460.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-2.88641") @@ -3608,6 +3936,7 @@ class Program_weight_tensor_parameter_327: class Program_weight_tensor_parameter_328: name = "parameter_328" + original_name = "linear_459.b_0" shape = [4096] dtype = "float32" min_val = float("-0.0960693") @@ -3619,6 +3948,7 @@ class Program_weight_tensor_parameter_328: class Program_weight_tensor_parameter_329: name = "parameter_329" + original_name = "linear_459.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.239208") @@ -3630,6 +3960,7 @@ class Program_weight_tensor_parameter_329: class Program_weight_tensor_parameter_330: name = "parameter_330" + original_name = "linear_458.b_0" shape = [1024] dtype = "float32" min_val = float("-0.183249") @@ -3641,6 +3972,7 @@ class Program_weight_tensor_parameter_330: class Program_weight_tensor_parameter_331: name = "parameter_331" + original_name = "linear_458.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.236141") @@ -3652,6 +3984,7 @@ class Program_weight_tensor_parameter_331: class Program_weight_tensor_parameter_332: name = "parameter_332" + original_name = "linear_457.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0546188") @@ -3663,6 +3996,7 @@ class Program_weight_tensor_parameter_332: class Program_weight_tensor_parameter_333: name = "parameter_333" + original_name = "linear_457.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.163183") @@ -3674,6 +4008,7 @@ class Program_weight_tensor_parameter_333: class Program_weight_tensor_parameter_334: name = "parameter_334" + original_name = "linear_456.b_0" shape = [1024] dtype = "float32" min_val = float("-4.52185") @@ -3685,6 +4020,7 @@ class Program_weight_tensor_parameter_334: class Program_weight_tensor_parameter_335: name = "parameter_335" + original_name = "linear_456.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.337653") @@ -3696,6 +4032,7 @@ class Program_weight_tensor_parameter_335: class Program_weight_tensor_parameter_336: name = "parameter_336" + original_name = "linear_455.b_0" shape = [1024] dtype = "float32" min_val = float("-0.203319") @@ -3707,6 +4044,7 @@ class Program_weight_tensor_parameter_336: class Program_weight_tensor_parameter_337: name = "parameter_337" + original_name = "linear_455.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.234942") @@ -3718,6 +4056,7 @@ class Program_weight_tensor_parameter_337: class Program_weight_tensor_parameter_338: name = "parameter_338" + original_name = "layer_norm_155.b_0" shape = [1024] dtype = "float32" min_val = float("-0.962042") @@ -3729,6 +4068,7 @@ class Program_weight_tensor_parameter_338: class Program_weight_tensor_parameter_339: name = "parameter_339" + original_name = "layer_norm_155.w_0" shape = [1024] dtype = "float32" min_val = float("0.575874") @@ -3740,6 +4080,7 @@ class Program_weight_tensor_parameter_339: class Program_weight_tensor_parameter_340: name = "parameter_340" + original_name = "layer_norm_154.b_0" shape = [1024] dtype = "float32" min_val = float("-1.57202") @@ -3751,6 +4092,7 @@ class Program_weight_tensor_parameter_340: class Program_weight_tensor_parameter_341: name = "parameter_341" + original_name = "layer_norm_154.w_0" shape = [1024] dtype = "float32" min_val = float("0.847908") @@ -3762,6 +4104,7 @@ class Program_weight_tensor_parameter_341: class Program_weight_tensor_parameter_342: name = "parameter_342" + original_name = "linear_454.b_0" shape = [1024] dtype = "float32" min_val = float("-0.800984") @@ -3773,6 +4116,7 @@ class Program_weight_tensor_parameter_342: class Program_weight_tensor_parameter_343: name = "parameter_343" + original_name = "linear_454.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-2.22345") @@ -3784,6 +4128,7 @@ class Program_weight_tensor_parameter_343: class Program_weight_tensor_parameter_344: name = "parameter_344" + original_name = "linear_453.b_0" shape = [4096] dtype = "float32" min_val = float("-0.0926698") @@ -3795,6 +4140,7 @@ class Program_weight_tensor_parameter_344: class Program_weight_tensor_parameter_345: name = "parameter_345" + original_name = "linear_453.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.208188") @@ -3806,6 +4152,7 @@ class Program_weight_tensor_parameter_345: class Program_weight_tensor_parameter_346: name = "parameter_346" + original_name = "linear_452.b_0" shape = [1024] dtype = "float32" min_val = float("-0.188814") @@ -3817,6 +4164,7 @@ class Program_weight_tensor_parameter_346: class Program_weight_tensor_parameter_347: name = "parameter_347" + original_name = "linear_452.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.276719") @@ -3828,6 +4176,7 @@ class Program_weight_tensor_parameter_347: class Program_weight_tensor_parameter_348: name = "parameter_348" + original_name = "linear_451.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0673743") @@ -3839,6 +4188,7 @@ class Program_weight_tensor_parameter_348: class Program_weight_tensor_parameter_349: name = "parameter_349" + original_name = "linear_451.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.136648") @@ -3850,6 +4200,7 @@ class Program_weight_tensor_parameter_349: class Program_weight_tensor_parameter_350: name = "parameter_350" + original_name = "linear_450.b_0" shape = [1024] dtype = "float32" min_val = float("-4.58268") @@ -3861,6 +4212,7 @@ class Program_weight_tensor_parameter_350: class Program_weight_tensor_parameter_351: name = "parameter_351" + original_name = "linear_450.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.354876") @@ -3872,6 +4224,7 @@ class Program_weight_tensor_parameter_351: class Program_weight_tensor_parameter_352: name = "parameter_352" + original_name = "linear_449.b_0" shape = [1024] dtype = "float32" min_val = float("-0.186255") @@ -3883,6 +4236,7 @@ class Program_weight_tensor_parameter_352: class Program_weight_tensor_parameter_353: name = "parameter_353" + original_name = "linear_449.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.232483") @@ -3894,6 +4248,7 @@ class Program_weight_tensor_parameter_353: class Program_weight_tensor_parameter_354: name = "parameter_354" + original_name = "layer_norm_153.b_0" shape = [1024] dtype = "float32" min_val = float("-1.04033") @@ -3905,6 +4260,7 @@ class Program_weight_tensor_parameter_354: class Program_weight_tensor_parameter_355: name = "parameter_355" + original_name = "layer_norm_153.w_0" shape = [1024] dtype = "float32" min_val = float("0.566165") @@ -3916,6 +4272,7 @@ class Program_weight_tensor_parameter_355: class Program_weight_tensor_parameter_356: name = "parameter_356" + original_name = "layer_norm_152.b_0" shape = [1024] dtype = "float32" min_val = float("-1.53316") @@ -3927,6 +4284,7 @@ class Program_weight_tensor_parameter_356: class Program_weight_tensor_parameter_357: name = "parameter_357" + original_name = "layer_norm_152.w_0" shape = [1024] dtype = "float32" min_val = float("0.856517") @@ -3938,6 +4296,7 @@ class Program_weight_tensor_parameter_357: class Program_weight_tensor_parameter_358: name = "parameter_358" + original_name = "linear_448.b_0" shape = [1024] dtype = "float32" min_val = float("-0.761929") @@ -3949,6 +4308,7 @@ class Program_weight_tensor_parameter_358: class Program_weight_tensor_parameter_359: name = "parameter_359" + original_name = "linear_448.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-1.66171") @@ -3960,6 +4320,7 @@ class Program_weight_tensor_parameter_359: class Program_weight_tensor_parameter_360: name = "parameter_360" + original_name = "linear_447.b_0" shape = [4096] dtype = "float32" min_val = float("-0.095568") @@ -3971,6 +4332,7 @@ class Program_weight_tensor_parameter_360: class Program_weight_tensor_parameter_361: name = "parameter_361" + original_name = "linear_447.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.214982") @@ -3982,6 +4344,7 @@ class Program_weight_tensor_parameter_361: class Program_weight_tensor_parameter_362: name = "parameter_362" + original_name = "linear_446.b_0" shape = [1024] dtype = "float32" min_val = float("-0.257232") @@ -3993,6 +4356,7 @@ class Program_weight_tensor_parameter_362: class Program_weight_tensor_parameter_363: name = "parameter_363" + original_name = "linear_446.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.883061") @@ -4004,6 +4368,7 @@ class Program_weight_tensor_parameter_363: class Program_weight_tensor_parameter_364: name = "parameter_364" + original_name = "linear_445.b_0" shape = [1024] dtype = "float32" min_val = float("-0.128426") @@ -4015,6 +4380,7 @@ class Program_weight_tensor_parameter_364: class Program_weight_tensor_parameter_365: name = "parameter_365" + original_name = "linear_445.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.13363") @@ -4026,6 +4392,7 @@ class Program_weight_tensor_parameter_365: class Program_weight_tensor_parameter_366: name = "parameter_366" + original_name = "linear_444.b_0" shape = [1024] dtype = "float32" min_val = float("-4.79968") @@ -4037,6 +4404,7 @@ class Program_weight_tensor_parameter_366: class Program_weight_tensor_parameter_367: name = "parameter_367" + original_name = "linear_444.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.255794") @@ -4048,6 +4416,7 @@ class Program_weight_tensor_parameter_367: class Program_weight_tensor_parameter_368: name = "parameter_368" + original_name = "linear_443.b_0" shape = [1024] dtype = "float32" min_val = float("-0.464429") @@ -4059,6 +4428,7 @@ class Program_weight_tensor_parameter_368: class Program_weight_tensor_parameter_369: name = "parameter_369" + original_name = "linear_443.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.290613") @@ -4070,6 +4440,7 @@ class Program_weight_tensor_parameter_369: class Program_weight_tensor_parameter_370: name = "parameter_370" + original_name = "layer_norm_151.b_0" shape = [1024] dtype = "float32" min_val = float("-0.988759") @@ -4081,6 +4452,7 @@ class Program_weight_tensor_parameter_370: class Program_weight_tensor_parameter_371: name = "parameter_371" + original_name = "layer_norm_151.w_0" shape = [1024] dtype = "float32" min_val = float("0.325382") @@ -4092,6 +4464,7 @@ class Program_weight_tensor_parameter_371: class Program_weight_tensor_parameter_372: name = "parameter_372" + original_name = "layer_norm_150.b_0" shape = [1024] dtype = "float32" min_val = float("-2.02986") @@ -4103,6 +4476,7 @@ class Program_weight_tensor_parameter_372: class Program_weight_tensor_parameter_373: name = "parameter_373" + original_name = "layer_norm_150.w_0" shape = [1024] dtype = "float32" min_val = float("0.865773") @@ -4114,6 +4488,7 @@ class Program_weight_tensor_parameter_373: class Program_weight_tensor_parameter_374: name = "parameter_374" + original_name = "linear_442.b_0" shape = [1024] dtype = "float32" min_val = float("-0.669053") @@ -4125,6 +4500,7 @@ class Program_weight_tensor_parameter_374: class Program_weight_tensor_parameter_375: name = "parameter_375" + original_name = "linear_442.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-1.53426") @@ -4136,6 +4512,7 @@ class Program_weight_tensor_parameter_375: class Program_weight_tensor_parameter_376: name = "parameter_376" + original_name = "linear_441.b_0" shape = [4096] dtype = "float32" min_val = float("-0.263098") @@ -4147,6 +4524,7 @@ class Program_weight_tensor_parameter_376: class Program_weight_tensor_parameter_377: name = "parameter_377" + original_name = "linear_441.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.286639") @@ -4158,6 +4536,7 @@ class Program_weight_tensor_parameter_377: class Program_weight_tensor_parameter_378: name = "parameter_378" + original_name = "linear_440.b_0" shape = [1024] dtype = "float32" min_val = float("-0.420357") @@ -4169,6 +4548,7 @@ class Program_weight_tensor_parameter_378: class Program_weight_tensor_parameter_379: name = "parameter_379" + original_name = "linear_440.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.895622") @@ -4180,6 +4560,7 @@ class Program_weight_tensor_parameter_379: class Program_weight_tensor_parameter_380: name = "parameter_380" + original_name = "linear_439.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0901028") @@ -4191,6 +4572,7 @@ class Program_weight_tensor_parameter_380: class Program_weight_tensor_parameter_381: name = "parameter_381" + original_name = "linear_439.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.117307") @@ -4202,6 +4584,7 @@ class Program_weight_tensor_parameter_381: class Program_weight_tensor_parameter_382: name = "parameter_382" + original_name = "linear_438.b_0" shape = [1024] dtype = "float32" min_val = float("-2.15888") @@ -4213,6 +4596,7 @@ class Program_weight_tensor_parameter_382: class Program_weight_tensor_parameter_383: name = "parameter_383" + original_name = "linear_438.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.180026") @@ -4224,6 +4608,7 @@ class Program_weight_tensor_parameter_383: class Program_weight_tensor_parameter_384: name = "parameter_384" + original_name = "linear_437.b_0" shape = [1024] dtype = "float32" min_val = float("-0.544662") @@ -4235,6 +4620,7 @@ class Program_weight_tensor_parameter_384: class Program_weight_tensor_parameter_385: name = "parameter_385" + original_name = "linear_437.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.201326") @@ -4246,6 +4632,7 @@ class Program_weight_tensor_parameter_385: class Program_weight_tensor_parameter_386: name = "parameter_386" + original_name = "layer_norm_149.b_0" shape = [1024] dtype = "float32" min_val = float("-0.762611") @@ -4257,6 +4644,7 @@ class Program_weight_tensor_parameter_386: class Program_weight_tensor_parameter_387: name = "parameter_387" + original_name = "layer_norm_149.w_0" shape = [1024] dtype = "float32" min_val = float("0.488272") @@ -4268,6 +4656,7 @@ class Program_weight_tensor_parameter_387: class Program_weight_tensor_parameter_388: name = "parameter_388" + original_name = "embedding_18.w_0" shape = [4, 1024] dtype = "float32" min_val = float("-0.0415497") @@ -4279,6 +4668,7 @@ class Program_weight_tensor_parameter_388: class Program_weight_tensor_parameter_389: name = "parameter_389" + original_name = "embedding_17.w_0" shape = [512, 1024] dtype = "float32" min_val = float("-0.776407") @@ -4290,6 +4680,7 @@ class Program_weight_tensor_parameter_389: class Program_weight_tensor_parameter_390: name = "parameter_390" + original_name = "embedding_16.w_0" shape = [12800, 1024] dtype = "float32" min_val = float("-0.775013") diff --git a/paddle_samples/PaddleNLP/ernie-3.0-base-zh/input_meta.py b/paddle_samples/PaddleNLP/ernie-3.0-base-zh/input_meta.py index 97eb8a799..a4123e481 100644 --- a/paddle_samples/PaddleNLP/ernie-3.0-base-zh/input_meta.py +++ b/paddle_samples/PaddleNLP/ernie-3.0-base-zh/input_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_data_0: name = "data_0" + original_name = "input_ids" shape = [1, 11] dtype = "int64" data = [1, 811, 1257, 175, 29, 502, 130, 706, 3619, 12046, 2] @@ -7,6 +8,7 @@ class Program_weight_tensor_data_0: class Program_weight_tensor_data_1: name = "data_1" + original_name = "token_type_ids" shape = [1, 11] dtype = "int64" data = [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0] diff --git a/paddle_samples/PaddleNLP/ernie-3.0-base-zh/model.py b/paddle_samples/PaddleNLP/ernie-3.0-base-zh/model.py index d45118d27..6c91a32b3 100644 --- a/paddle_samples/PaddleNLP/ernie-3.0-base-zh/model.py +++ b/paddle_samples/PaddleNLP/ernie-3.0-base-zh/model.py @@ -266,10 +266,11 @@ def forward( # pd_op.embedding: (1x11x768xf32) <- (1x11xi64, 2048x768xf32) embedding_1 = paddle._C_ops.embedding(subtract_0, parameter_198, -1, False) - del parameter_198 + del parameter_198, subtract_0 # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 1x11x768xf32) add_0 = paddle._C_ops.add(embedding_0, embedding_1) + del embedding_0, embedding_1 # pd_op.embedding: (1x11x768xf32) <- (1x11xi64, 4x768xf32) embedding_2 = paddle._C_ops.embedding(data_1, parameter_197, -1, False) @@ -277,6 +278,7 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 1x11x768xf32) add_1 = paddle._C_ops.add(add_0, embedding_2) + del add_0, embedding_2 # pd_op.full: (1xf32) <- () full_4 = paddle._C_ops.full( @@ -289,10 +291,11 @@ def forward( # pd_op.embedding: (1x11x768xf32) <- (1x11xi64, 3x768xf32) embedding_3 = paddle._C_ops.embedding(scale_1, parameter_196, -1, False) - del parameter_196 + del parameter_196, scale_1 # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 1x11x768xf32) add_2 = paddle._C_ops.add(add_1, embedding_3) + del add_1, embedding_3 # pd_op.layer_norm: (1x11x768xf32, 1x11xf32, 1x11xf32) <- (1x11x768xf32, 768xf32, 768xf32) layer_norm_0, layer_norm_1, layer_norm_2 = (lambda x, f: f(x))( @@ -301,125 +304,17 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_194, parameter_195 + del add_2, parameter_194, parameter_195 # pd_op.full: (1xf32) <- () full_5 = paddle._C_ops.full( [1], float("0.1"), paddle.float32, paddle.core.CPUPlace() ) - # pd_op.assign: (1xf32) <- (1xf32) - assign_0 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_1 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_2 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_3 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_4 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_5 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_6 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_7 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_8 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_9 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_10 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_11 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_12 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_13 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_14 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_15 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_16 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_17 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_18 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_19 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_20 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_21 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_22 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_23 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_24 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_25 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_26 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_27 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_28 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_29 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_30 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_31 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_32 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_33 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_34 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_35 = full_5 - # pd_op.dropout: (1x11x768xf32, 1x11x768xui8) <- (1x11x768xf32, None, 1xf32) dropout_0, dropout_1 = (lambda x, f: f(x))( paddle._C_ops.dropout( - layer_norm_0, None, full_5, False, "upscale_in_train", 0, False + layer_norm_0, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -431,13 +326,14 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_3 = paddle._C_ops.add(matmul_0, parameter_192) - del parameter_192 + del matmul_0, parameter_192 # pd_op.full_int_array: (4xi64) <- () full_int_array_1 = [0, 0, 12, 64] # pd_op.reshape: (1x11x12x64xf32) <- (1x11x768xf32, 4xi64) reshape_0 = paddle._C_ops.reshape(add_3, full_int_array_1) + del add_3 # pd_op.transpose: (1x12x11x64xf32) <- (1x11x12x64xf32) transpose_0 = paddle._C_ops.transpose(reshape_0, [0, 2, 1, 3]) @@ -449,7 +345,7 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_4 = paddle._C_ops.add(matmul_1, parameter_190) - del parameter_190 + del matmul_1, parameter_190 # pd_op.matmul: (1x11x768xf32) <- (1x11x768xf32, 768x768xf32) matmul_2 = paddle._C_ops.matmul(dropout_0, parameter_189, False, False) @@ -457,10 +353,11 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_5 = paddle._C_ops.add(matmul_2, parameter_188) - del parameter_188 + del matmul_2, parameter_188 # pd_op.reshape: (1x11x12x64xf32) <- (1x11x768xf32, 4xi64) reshape_1 = paddle._C_ops.reshape(add_4, full_int_array_1) + del add_4 # pd_op.transpose: (1x12x11x64xf32) <- (1x11x12x64xf32) transpose_1 = paddle._C_ops.transpose(reshape_1, [0, 2, 1, 3]) @@ -468,6 +365,7 @@ def forward( # pd_op.reshape: (1x11x12x64xf32) <- (1x11x768xf32, 4xi64) reshape_2 = paddle._C_ops.reshape(add_5, full_int_array_1) + del add_5 # pd_op.transpose: (1x12x11x64xf32) <- (1x11x12x64xf32) transpose_2 = paddle._C_ops.transpose(reshape_2, [0, 2, 1, 3]) @@ -478,48 +376,17 @@ def forward( [1], float("0.125"), paddle.float32, paddle.core.CPUPlace() ) - # pd_op.assign: (1xf32) <- (1xf32) - assign_36 = full_6 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_37 = full_6 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_38 = full_6 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_39 = full_6 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_40 = full_6 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_41 = full_6 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_42 = full_6 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_43 = full_6 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_44 = full_6 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_45 = full_6 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_46 = full_6 - # pd_op.scale: (1x12x11x64xf32) <- (1x12x11x64xf32, 1xf32) scale_2 = paddle._C_ops.scale(transpose_0, full_6, float("0"), True) del transpose_0 # pd_op.matmul: (1x12x11x11xf32) <- (1x12x11x64xf32, 1x12x11x64xf32) matmul_3 = paddle._C_ops.matmul(scale_2, transpose_1, False, True) + del scale_2, transpose_1 # pd_op.add: (1x12x11x11xf32) <- (1x12x11x11xf32, 1x1x1x11xf32) add_6 = paddle._C_ops.add(matmul_3, unsqueeze_0) + del matmul_3 # pd_op.softmax: (1x12x11x11xf32) <- (1x12x11x11xf32) softmax_0 = paddle._C_ops.softmax(add_6, -1) @@ -528,13 +395,15 @@ def forward( # pd_op.dropout: (1x12x11x11xf32, 1x12x11x11xui8) <- (1x12x11x11xf32, None, 1xf32) dropout_2, dropout_3 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_0, None, full_5, False, "upscale_in_train", 0, False + softmax_0, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_0 # pd_op.matmul: (1x12x11x64xf32) <- (1x12x11x11xf32, 1x12x11x64xf32) matmul_4 = paddle._C_ops.matmul(dropout_2, transpose_2, False, False) + del dropout_2, transpose_2 # pd_op.transpose: (1x11x12x64xf32) <- (1x12x11x64xf32) transpose_3 = paddle._C_ops.transpose(matmul_4, [0, 2, 1, 3]) @@ -545,19 +414,20 @@ def forward( # pd_op.reshape: (1x11x768xf32) <- (1x11x12x64xf32, 3xi64) reshape_3 = paddle._C_ops.reshape(transpose_3, full_int_array_2) + del transpose_3 # pd_op.matmul: (1x11x768xf32) <- (1x11x768xf32, 768x768xf32) matmul_5 = paddle._C_ops.matmul(reshape_3, parameter_187, False, False) - del parameter_187 + del parameter_187, reshape_3 # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_7 = paddle._C_ops.add(matmul_5, parameter_186) - del parameter_186 + del matmul_5, parameter_186 # pd_op.dropout: (1x11x768xf32, 1x11x768xui8) <- (1x11x768xf32, None, 1xf32) dropout_4, dropout_5 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_7, None, full_5, False, "upscale_in_train", 0, False + add_7, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -565,6 +435,7 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 1x11x768xf32) add_8 = paddle._C_ops.add(dropout_0, dropout_4) + del dropout_0, dropout_4 # pd_op.layer_norm: (1x11x768xf32, 1x11xf32, 1x11xf32) <- (1x11x768xf32, 768xf32, 768xf32) layer_norm_3, layer_norm_4, layer_norm_5 = (lambda x, f: f(x))( @@ -573,7 +444,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_180, parameter_181 + del add_8, parameter_180, parameter_181 # pd_op.matmul: (1x11x3072xf32) <- (1x11x768xf32, 768x3072xf32) matmul_6 = paddle._C_ops.matmul(layer_norm_3, parameter_185, False, False) @@ -581,23 +452,24 @@ def forward( # pd_op.add: (1x11x3072xf32) <- (1x11x3072xf32, 3072xf32) add_9 = paddle._C_ops.add(matmul_6, parameter_184) - del parameter_184 + del matmul_6, parameter_184 # pd_op.gelu: (1x11x3072xf32) <- (1x11x3072xf32) gelu_0 = paddle._C_ops.gelu(add_9, False) + del add_9 # pd_op.matmul: (1x11x768xf32) <- (1x11x3072xf32, 3072x768xf32) matmul_7 = paddle._C_ops.matmul(gelu_0, parameter_183, False, False) - del parameter_183 + del gelu_0, parameter_183 # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_10 = paddle._C_ops.add(matmul_7, parameter_182) - del parameter_182 + del matmul_7, parameter_182 # pd_op.dropout: (1x11x768xf32, 1x11x768xui8) <- (1x11x768xf32, None, 1xf32) dropout_6, dropout_7 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_10, None, full_5, False, "upscale_in_train", 0, False + add_10, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -605,6 +477,7 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 1x11x768xf32) add_11 = paddle._C_ops.add(layer_norm_3, dropout_6) + del dropout_6, layer_norm_3 # pd_op.layer_norm: (1x11x768xf32, 1x11xf32, 1x11xf32) <- (1x11x768xf32, 768xf32, 768xf32) layer_norm_6, layer_norm_7, layer_norm_8 = (lambda x, f: f(x))( @@ -613,7 +486,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_178, parameter_179 + del add_11, parameter_178, parameter_179 # pd_op.matmul: (1x11x768xf32) <- (1x11x768xf32, 768x768xf32) matmul_8 = paddle._C_ops.matmul(layer_norm_6, parameter_177, False, False) @@ -621,10 +494,11 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_12 = paddle._C_ops.add(matmul_8, parameter_176) - del parameter_176 + del matmul_8, parameter_176 # pd_op.reshape: (1x11x12x64xf32) <- (1x11x768xf32, 4xi64) reshape_4 = paddle._C_ops.reshape(add_12, full_int_array_1) + del add_12 # pd_op.transpose: (1x12x11x64xf32) <- (1x11x12x64xf32) transpose_4 = paddle._C_ops.transpose(reshape_4, [0, 2, 1, 3]) @@ -636,7 +510,7 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_13 = paddle._C_ops.add(matmul_9, parameter_174) - del parameter_174 + del matmul_9, parameter_174 # pd_op.matmul: (1x11x768xf32) <- (1x11x768xf32, 768x768xf32) matmul_10 = paddle._C_ops.matmul(layer_norm_6, parameter_173, False, False) @@ -644,10 +518,11 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_14 = paddle._C_ops.add(matmul_10, parameter_172) - del parameter_172 + del matmul_10, parameter_172 # pd_op.reshape: (1x11x12x64xf32) <- (1x11x768xf32, 4xi64) reshape_5 = paddle._C_ops.reshape(add_13, full_int_array_1) + del add_13 # pd_op.transpose: (1x12x11x64xf32) <- (1x11x12x64xf32) transpose_5 = paddle._C_ops.transpose(reshape_5, [0, 2, 1, 3]) @@ -655,6 +530,7 @@ def forward( # pd_op.reshape: (1x11x12x64xf32) <- (1x11x768xf32, 4xi64) reshape_6 = paddle._C_ops.reshape(add_14, full_int_array_1) + del add_14 # pd_op.transpose: (1x12x11x64xf32) <- (1x11x12x64xf32) transpose_6 = paddle._C_ops.transpose(reshape_6, [0, 2, 1, 3]) @@ -666,9 +542,11 @@ def forward( # pd_op.matmul: (1x12x11x11xf32) <- (1x12x11x64xf32, 1x12x11x64xf32) matmul_11 = paddle._C_ops.matmul(scale_3, transpose_5, False, True) + del scale_3, transpose_5 # pd_op.add: (1x12x11x11xf32) <- (1x12x11x11xf32, 1x1x1x11xf32) add_15 = paddle._C_ops.add(matmul_11, unsqueeze_0) + del matmul_11 # pd_op.softmax: (1x12x11x11xf32) <- (1x12x11x11xf32) softmax_1 = paddle._C_ops.softmax(add_15, -1) @@ -677,13 +555,15 @@ def forward( # pd_op.dropout: (1x12x11x11xf32, 1x12x11x11xui8) <- (1x12x11x11xf32, None, 1xf32) dropout_8, dropout_9 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_1, None, full_5, False, "upscale_in_train", 0, False + softmax_1, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_1 # pd_op.matmul: (1x12x11x64xf32) <- (1x12x11x11xf32, 1x12x11x64xf32) matmul_12 = paddle._C_ops.matmul(dropout_8, transpose_6, False, False) + del dropout_8, transpose_6 # pd_op.transpose: (1x11x12x64xf32) <- (1x12x11x64xf32) transpose_7 = paddle._C_ops.transpose(matmul_12, [0, 2, 1, 3]) @@ -691,19 +571,20 @@ def forward( # pd_op.reshape: (1x11x768xf32) <- (1x11x12x64xf32, 3xi64) reshape_7 = paddle._C_ops.reshape(transpose_7, full_int_array_2) + del transpose_7 # pd_op.matmul: (1x11x768xf32) <- (1x11x768xf32, 768x768xf32) matmul_13 = paddle._C_ops.matmul(reshape_7, parameter_171, False, False) - del parameter_171 + del parameter_171, reshape_7 # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_16 = paddle._C_ops.add(matmul_13, parameter_170) - del parameter_170 + del matmul_13, parameter_170 # pd_op.dropout: (1x11x768xf32, 1x11x768xui8) <- (1x11x768xf32, None, 1xf32) dropout_10, dropout_11 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_16, None, full_5, False, "upscale_in_train", 0, False + add_16, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -711,6 +592,7 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 1x11x768xf32) add_17 = paddle._C_ops.add(layer_norm_6, dropout_10) + del dropout_10, layer_norm_6 # pd_op.layer_norm: (1x11x768xf32, 1x11xf32, 1x11xf32) <- (1x11x768xf32, 768xf32, 768xf32) layer_norm_9, layer_norm_10, layer_norm_11 = (lambda x, f: f(x))( @@ -719,7 +601,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_164, parameter_165 + del add_17, parameter_164, parameter_165 # pd_op.matmul: (1x11x3072xf32) <- (1x11x768xf32, 768x3072xf32) matmul_14 = paddle._C_ops.matmul(layer_norm_9, parameter_169, False, False) @@ -727,23 +609,24 @@ def forward( # pd_op.add: (1x11x3072xf32) <- (1x11x3072xf32, 3072xf32) add_18 = paddle._C_ops.add(matmul_14, parameter_168) - del parameter_168 + del matmul_14, parameter_168 # pd_op.gelu: (1x11x3072xf32) <- (1x11x3072xf32) gelu_1 = paddle._C_ops.gelu(add_18, False) + del add_18 # pd_op.matmul: (1x11x768xf32) <- (1x11x3072xf32, 3072x768xf32) matmul_15 = paddle._C_ops.matmul(gelu_1, parameter_167, False, False) - del parameter_167 + del gelu_1, parameter_167 # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_19 = paddle._C_ops.add(matmul_15, parameter_166) - del parameter_166 + del matmul_15, parameter_166 # pd_op.dropout: (1x11x768xf32, 1x11x768xui8) <- (1x11x768xf32, None, 1xf32) dropout_12, dropout_13 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_19, None, full_5, False, "upscale_in_train", 0, False + add_19, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -751,6 +634,7 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 1x11x768xf32) add_20 = paddle._C_ops.add(layer_norm_9, dropout_12) + del dropout_12, layer_norm_9 # pd_op.layer_norm: (1x11x768xf32, 1x11xf32, 1x11xf32) <- (1x11x768xf32, 768xf32, 768xf32) layer_norm_12, layer_norm_13, layer_norm_14 = (lambda x, f: f(x))( @@ -759,7 +643,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_162, parameter_163 + del add_20, parameter_162, parameter_163 # pd_op.matmul: (1x11x768xf32) <- (1x11x768xf32, 768x768xf32) matmul_16 = paddle._C_ops.matmul(layer_norm_12, parameter_161, False, False) @@ -767,10 +651,11 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_21 = paddle._C_ops.add(matmul_16, parameter_160) - del parameter_160 + del matmul_16, parameter_160 # pd_op.reshape: (1x11x12x64xf32) <- (1x11x768xf32, 4xi64) reshape_8 = paddle._C_ops.reshape(add_21, full_int_array_1) + del add_21 # pd_op.transpose: (1x12x11x64xf32) <- (1x11x12x64xf32) transpose_8 = paddle._C_ops.transpose(reshape_8, [0, 2, 1, 3]) @@ -782,7 +667,7 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_22 = paddle._C_ops.add(matmul_17, parameter_158) - del parameter_158 + del matmul_17, parameter_158 # pd_op.matmul: (1x11x768xf32) <- (1x11x768xf32, 768x768xf32) matmul_18 = paddle._C_ops.matmul(layer_norm_12, parameter_157, False, False) @@ -790,10 +675,11 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_23 = paddle._C_ops.add(matmul_18, parameter_156) - del parameter_156 + del matmul_18, parameter_156 # pd_op.reshape: (1x11x12x64xf32) <- (1x11x768xf32, 4xi64) reshape_9 = paddle._C_ops.reshape(add_22, full_int_array_1) + del add_22 # pd_op.transpose: (1x12x11x64xf32) <- (1x11x12x64xf32) transpose_9 = paddle._C_ops.transpose(reshape_9, [0, 2, 1, 3]) @@ -801,6 +687,7 @@ def forward( # pd_op.reshape: (1x11x12x64xf32) <- (1x11x768xf32, 4xi64) reshape_10 = paddle._C_ops.reshape(add_23, full_int_array_1) + del add_23 # pd_op.transpose: (1x12x11x64xf32) <- (1x11x12x64xf32) transpose_10 = paddle._C_ops.transpose(reshape_10, [0, 2, 1, 3]) @@ -812,9 +699,11 @@ def forward( # pd_op.matmul: (1x12x11x11xf32) <- (1x12x11x64xf32, 1x12x11x64xf32) matmul_19 = paddle._C_ops.matmul(scale_4, transpose_9, False, True) + del scale_4, transpose_9 # pd_op.add: (1x12x11x11xf32) <- (1x12x11x11xf32, 1x1x1x11xf32) add_24 = paddle._C_ops.add(matmul_19, unsqueeze_0) + del matmul_19 # pd_op.softmax: (1x12x11x11xf32) <- (1x12x11x11xf32) softmax_2 = paddle._C_ops.softmax(add_24, -1) @@ -823,13 +712,15 @@ def forward( # pd_op.dropout: (1x12x11x11xf32, 1x12x11x11xui8) <- (1x12x11x11xf32, None, 1xf32) dropout_14, dropout_15 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_2, None, full_5, False, "upscale_in_train", 0, False + softmax_2, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_2 # pd_op.matmul: (1x12x11x64xf32) <- (1x12x11x11xf32, 1x12x11x64xf32) matmul_20 = paddle._C_ops.matmul(dropout_14, transpose_10, False, False) + del dropout_14, transpose_10 # pd_op.transpose: (1x11x12x64xf32) <- (1x12x11x64xf32) transpose_11 = paddle._C_ops.transpose(matmul_20, [0, 2, 1, 3]) @@ -837,19 +728,20 @@ def forward( # pd_op.reshape: (1x11x768xf32) <- (1x11x12x64xf32, 3xi64) reshape_11 = paddle._C_ops.reshape(transpose_11, full_int_array_2) + del transpose_11 # pd_op.matmul: (1x11x768xf32) <- (1x11x768xf32, 768x768xf32) matmul_21 = paddle._C_ops.matmul(reshape_11, parameter_155, False, False) - del parameter_155 + del parameter_155, reshape_11 # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_25 = paddle._C_ops.add(matmul_21, parameter_154) - del parameter_154 + del matmul_21, parameter_154 # pd_op.dropout: (1x11x768xf32, 1x11x768xui8) <- (1x11x768xf32, None, 1xf32) dropout_16, dropout_17 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_25, None, full_5, False, "upscale_in_train", 0, False + add_25, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -857,6 +749,7 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 1x11x768xf32) add_26 = paddle._C_ops.add(layer_norm_12, dropout_16) + del dropout_16, layer_norm_12 # pd_op.layer_norm: (1x11x768xf32, 1x11xf32, 1x11xf32) <- (1x11x768xf32, 768xf32, 768xf32) layer_norm_15, layer_norm_16, layer_norm_17 = (lambda x, f: f(x))( @@ -865,7 +758,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_148, parameter_149 + del add_26, parameter_148, parameter_149 # pd_op.matmul: (1x11x3072xf32) <- (1x11x768xf32, 768x3072xf32) matmul_22 = paddle._C_ops.matmul(layer_norm_15, parameter_153, False, False) @@ -873,23 +766,24 @@ def forward( # pd_op.add: (1x11x3072xf32) <- (1x11x3072xf32, 3072xf32) add_27 = paddle._C_ops.add(matmul_22, parameter_152) - del parameter_152 + del matmul_22, parameter_152 # pd_op.gelu: (1x11x3072xf32) <- (1x11x3072xf32) gelu_2 = paddle._C_ops.gelu(add_27, False) + del add_27 # pd_op.matmul: (1x11x768xf32) <- (1x11x3072xf32, 3072x768xf32) matmul_23 = paddle._C_ops.matmul(gelu_2, parameter_151, False, False) - del parameter_151 + del gelu_2, parameter_151 # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_28 = paddle._C_ops.add(matmul_23, parameter_150) - del parameter_150 + del matmul_23, parameter_150 # pd_op.dropout: (1x11x768xf32, 1x11x768xui8) <- (1x11x768xf32, None, 1xf32) dropout_18, dropout_19 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_28, None, full_5, False, "upscale_in_train", 0, False + add_28, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -897,6 +791,7 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 1x11x768xf32) add_29 = paddle._C_ops.add(layer_norm_15, dropout_18) + del dropout_18, layer_norm_15 # pd_op.layer_norm: (1x11x768xf32, 1x11xf32, 1x11xf32) <- (1x11x768xf32, 768xf32, 768xf32) layer_norm_18, layer_norm_19, layer_norm_20 = (lambda x, f: f(x))( @@ -905,7 +800,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_146, parameter_147 + del add_29, parameter_146, parameter_147 # pd_op.matmul: (1x11x768xf32) <- (1x11x768xf32, 768x768xf32) matmul_24 = paddle._C_ops.matmul(layer_norm_18, parameter_145, False, False) @@ -913,10 +808,11 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_30 = paddle._C_ops.add(matmul_24, parameter_144) - del parameter_144 + del matmul_24, parameter_144 # pd_op.reshape: (1x11x12x64xf32) <- (1x11x768xf32, 4xi64) reshape_12 = paddle._C_ops.reshape(add_30, full_int_array_1) + del add_30 # pd_op.transpose: (1x12x11x64xf32) <- (1x11x12x64xf32) transpose_12 = paddle._C_ops.transpose(reshape_12, [0, 2, 1, 3]) @@ -928,7 +824,7 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_31 = paddle._C_ops.add(matmul_25, parameter_142) - del parameter_142 + del matmul_25, parameter_142 # pd_op.matmul: (1x11x768xf32) <- (1x11x768xf32, 768x768xf32) matmul_26 = paddle._C_ops.matmul(layer_norm_18, parameter_141, False, False) @@ -936,10 +832,11 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_32 = paddle._C_ops.add(matmul_26, parameter_140) - del parameter_140 + del matmul_26, parameter_140 # pd_op.reshape: (1x11x12x64xf32) <- (1x11x768xf32, 4xi64) reshape_13 = paddle._C_ops.reshape(add_31, full_int_array_1) + del add_31 # pd_op.transpose: (1x12x11x64xf32) <- (1x11x12x64xf32) transpose_13 = paddle._C_ops.transpose(reshape_13, [0, 2, 1, 3]) @@ -947,6 +844,7 @@ def forward( # pd_op.reshape: (1x11x12x64xf32) <- (1x11x768xf32, 4xi64) reshape_14 = paddle._C_ops.reshape(add_32, full_int_array_1) + del add_32 # pd_op.transpose: (1x12x11x64xf32) <- (1x11x12x64xf32) transpose_14 = paddle._C_ops.transpose(reshape_14, [0, 2, 1, 3]) @@ -958,9 +856,11 @@ def forward( # pd_op.matmul: (1x12x11x11xf32) <- (1x12x11x64xf32, 1x12x11x64xf32) matmul_27 = paddle._C_ops.matmul(scale_5, transpose_13, False, True) + del scale_5, transpose_13 # pd_op.add: (1x12x11x11xf32) <- (1x12x11x11xf32, 1x1x1x11xf32) add_33 = paddle._C_ops.add(matmul_27, unsqueeze_0) + del matmul_27 # pd_op.softmax: (1x12x11x11xf32) <- (1x12x11x11xf32) softmax_3 = paddle._C_ops.softmax(add_33, -1) @@ -969,13 +869,15 @@ def forward( # pd_op.dropout: (1x12x11x11xf32, 1x12x11x11xui8) <- (1x12x11x11xf32, None, 1xf32) dropout_20, dropout_21 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_3, None, full_5, False, "upscale_in_train", 0, False + softmax_3, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_3 # pd_op.matmul: (1x12x11x64xf32) <- (1x12x11x11xf32, 1x12x11x64xf32) matmul_28 = paddle._C_ops.matmul(dropout_20, transpose_14, False, False) + del dropout_20, transpose_14 # pd_op.transpose: (1x11x12x64xf32) <- (1x12x11x64xf32) transpose_15 = paddle._C_ops.transpose(matmul_28, [0, 2, 1, 3]) @@ -983,19 +885,20 @@ def forward( # pd_op.reshape: (1x11x768xf32) <- (1x11x12x64xf32, 3xi64) reshape_15 = paddle._C_ops.reshape(transpose_15, full_int_array_2) + del transpose_15 # pd_op.matmul: (1x11x768xf32) <- (1x11x768xf32, 768x768xf32) matmul_29 = paddle._C_ops.matmul(reshape_15, parameter_139, False, False) - del parameter_139 + del parameter_139, reshape_15 # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_34 = paddle._C_ops.add(matmul_29, parameter_138) - del parameter_138 + del matmul_29, parameter_138 # pd_op.dropout: (1x11x768xf32, 1x11x768xui8) <- (1x11x768xf32, None, 1xf32) dropout_22, dropout_23 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_34, None, full_5, False, "upscale_in_train", 0, False + add_34, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1003,6 +906,7 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 1x11x768xf32) add_35 = paddle._C_ops.add(layer_norm_18, dropout_22) + del dropout_22, layer_norm_18 # pd_op.layer_norm: (1x11x768xf32, 1x11xf32, 1x11xf32) <- (1x11x768xf32, 768xf32, 768xf32) layer_norm_21, layer_norm_22, layer_norm_23 = (lambda x, f: f(x))( @@ -1011,7 +915,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_132, parameter_133 + del add_35, parameter_132, parameter_133 # pd_op.matmul: (1x11x3072xf32) <- (1x11x768xf32, 768x3072xf32) matmul_30 = paddle._C_ops.matmul(layer_norm_21, parameter_137, False, False) @@ -1019,23 +923,24 @@ def forward( # pd_op.add: (1x11x3072xf32) <- (1x11x3072xf32, 3072xf32) add_36 = paddle._C_ops.add(matmul_30, parameter_136) - del parameter_136 + del matmul_30, parameter_136 # pd_op.gelu: (1x11x3072xf32) <- (1x11x3072xf32) gelu_3 = paddle._C_ops.gelu(add_36, False) + del add_36 # pd_op.matmul: (1x11x768xf32) <- (1x11x3072xf32, 3072x768xf32) matmul_31 = paddle._C_ops.matmul(gelu_3, parameter_135, False, False) - del parameter_135 + del gelu_3, parameter_135 # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_37 = paddle._C_ops.add(matmul_31, parameter_134) - del parameter_134 + del matmul_31, parameter_134 # pd_op.dropout: (1x11x768xf32, 1x11x768xui8) <- (1x11x768xf32, None, 1xf32) dropout_24, dropout_25 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_37, None, full_5, False, "upscale_in_train", 0, False + add_37, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1043,6 +948,7 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 1x11x768xf32) add_38 = paddle._C_ops.add(layer_norm_21, dropout_24) + del dropout_24, layer_norm_21 # pd_op.layer_norm: (1x11x768xf32, 1x11xf32, 1x11xf32) <- (1x11x768xf32, 768xf32, 768xf32) layer_norm_24, layer_norm_25, layer_norm_26 = (lambda x, f: f(x))( @@ -1051,7 +957,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_130, parameter_131 + del add_38, parameter_130, parameter_131 # pd_op.matmul: (1x11x768xf32) <- (1x11x768xf32, 768x768xf32) matmul_32 = paddle._C_ops.matmul(layer_norm_24, parameter_129, False, False) @@ -1059,10 +965,11 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_39 = paddle._C_ops.add(matmul_32, parameter_128) - del parameter_128 + del matmul_32, parameter_128 # pd_op.reshape: (1x11x12x64xf32) <- (1x11x768xf32, 4xi64) reshape_16 = paddle._C_ops.reshape(add_39, full_int_array_1) + del add_39 # pd_op.transpose: (1x12x11x64xf32) <- (1x11x12x64xf32) transpose_16 = paddle._C_ops.transpose(reshape_16, [0, 2, 1, 3]) @@ -1074,7 +981,7 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_40 = paddle._C_ops.add(matmul_33, parameter_126) - del parameter_126 + del matmul_33, parameter_126 # pd_op.matmul: (1x11x768xf32) <- (1x11x768xf32, 768x768xf32) matmul_34 = paddle._C_ops.matmul(layer_norm_24, parameter_125, False, False) @@ -1082,10 +989,11 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_41 = paddle._C_ops.add(matmul_34, parameter_124) - del parameter_124 + del matmul_34, parameter_124 # pd_op.reshape: (1x11x12x64xf32) <- (1x11x768xf32, 4xi64) reshape_17 = paddle._C_ops.reshape(add_40, full_int_array_1) + del add_40 # pd_op.transpose: (1x12x11x64xf32) <- (1x11x12x64xf32) transpose_17 = paddle._C_ops.transpose(reshape_17, [0, 2, 1, 3]) @@ -1093,6 +1001,7 @@ def forward( # pd_op.reshape: (1x11x12x64xf32) <- (1x11x768xf32, 4xi64) reshape_18 = paddle._C_ops.reshape(add_41, full_int_array_1) + del add_41 # pd_op.transpose: (1x12x11x64xf32) <- (1x11x12x64xf32) transpose_18 = paddle._C_ops.transpose(reshape_18, [0, 2, 1, 3]) @@ -1104,9 +1013,11 @@ def forward( # pd_op.matmul: (1x12x11x11xf32) <- (1x12x11x64xf32, 1x12x11x64xf32) matmul_35 = paddle._C_ops.matmul(scale_6, transpose_17, False, True) + del scale_6, transpose_17 # pd_op.add: (1x12x11x11xf32) <- (1x12x11x11xf32, 1x1x1x11xf32) add_42 = paddle._C_ops.add(matmul_35, unsqueeze_0) + del matmul_35 # pd_op.softmax: (1x12x11x11xf32) <- (1x12x11x11xf32) softmax_4 = paddle._C_ops.softmax(add_42, -1) @@ -1115,13 +1026,15 @@ def forward( # pd_op.dropout: (1x12x11x11xf32, 1x12x11x11xui8) <- (1x12x11x11xf32, None, 1xf32) dropout_26, dropout_27 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_4, None, full_5, False, "upscale_in_train", 0, False + softmax_4, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_4 # pd_op.matmul: (1x12x11x64xf32) <- (1x12x11x11xf32, 1x12x11x64xf32) matmul_36 = paddle._C_ops.matmul(dropout_26, transpose_18, False, False) + del dropout_26, transpose_18 # pd_op.transpose: (1x11x12x64xf32) <- (1x12x11x64xf32) transpose_19 = paddle._C_ops.transpose(matmul_36, [0, 2, 1, 3]) @@ -1129,19 +1042,20 @@ def forward( # pd_op.reshape: (1x11x768xf32) <- (1x11x12x64xf32, 3xi64) reshape_19 = paddle._C_ops.reshape(transpose_19, full_int_array_2) + del transpose_19 # pd_op.matmul: (1x11x768xf32) <- (1x11x768xf32, 768x768xf32) matmul_37 = paddle._C_ops.matmul(reshape_19, parameter_123, False, False) - del parameter_123 + del parameter_123, reshape_19 # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_43 = paddle._C_ops.add(matmul_37, parameter_122) - del parameter_122 + del matmul_37, parameter_122 # pd_op.dropout: (1x11x768xf32, 1x11x768xui8) <- (1x11x768xf32, None, 1xf32) dropout_28, dropout_29 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_43, None, full_5, False, "upscale_in_train", 0, False + add_43, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1149,6 +1063,7 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 1x11x768xf32) add_44 = paddle._C_ops.add(layer_norm_24, dropout_28) + del dropout_28, layer_norm_24 # pd_op.layer_norm: (1x11x768xf32, 1x11xf32, 1x11xf32) <- (1x11x768xf32, 768xf32, 768xf32) layer_norm_27, layer_norm_28, layer_norm_29 = (lambda x, f: f(x))( @@ -1157,7 +1072,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_116, parameter_117 + del add_44, parameter_116, parameter_117 # pd_op.matmul: (1x11x3072xf32) <- (1x11x768xf32, 768x3072xf32) matmul_38 = paddle._C_ops.matmul(layer_norm_27, parameter_121, False, False) @@ -1165,23 +1080,24 @@ def forward( # pd_op.add: (1x11x3072xf32) <- (1x11x3072xf32, 3072xf32) add_45 = paddle._C_ops.add(matmul_38, parameter_120) - del parameter_120 + del matmul_38, parameter_120 # pd_op.gelu: (1x11x3072xf32) <- (1x11x3072xf32) gelu_4 = paddle._C_ops.gelu(add_45, False) + del add_45 # pd_op.matmul: (1x11x768xf32) <- (1x11x3072xf32, 3072x768xf32) matmul_39 = paddle._C_ops.matmul(gelu_4, parameter_119, False, False) - del parameter_119 + del gelu_4, parameter_119 # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_46 = paddle._C_ops.add(matmul_39, parameter_118) - del parameter_118 + del matmul_39, parameter_118 # pd_op.dropout: (1x11x768xf32, 1x11x768xui8) <- (1x11x768xf32, None, 1xf32) dropout_30, dropout_31 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_46, None, full_5, False, "upscale_in_train", 0, False + add_46, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1189,6 +1105,7 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 1x11x768xf32) add_47 = paddle._C_ops.add(layer_norm_27, dropout_30) + del dropout_30, layer_norm_27 # pd_op.layer_norm: (1x11x768xf32, 1x11xf32, 1x11xf32) <- (1x11x768xf32, 768xf32, 768xf32) layer_norm_30, layer_norm_31, layer_norm_32 = (lambda x, f: f(x))( @@ -1197,7 +1114,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_114, parameter_115 + del add_47, parameter_114, parameter_115 # pd_op.matmul: (1x11x768xf32) <- (1x11x768xf32, 768x768xf32) matmul_40 = paddle._C_ops.matmul(layer_norm_30, parameter_113, False, False) @@ -1205,10 +1122,11 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_48 = paddle._C_ops.add(matmul_40, parameter_112) - del parameter_112 + del matmul_40, parameter_112 # pd_op.reshape: (1x11x12x64xf32) <- (1x11x768xf32, 4xi64) reshape_20 = paddle._C_ops.reshape(add_48, full_int_array_1) + del add_48 # pd_op.transpose: (1x12x11x64xf32) <- (1x11x12x64xf32) transpose_20 = paddle._C_ops.transpose(reshape_20, [0, 2, 1, 3]) @@ -1220,7 +1138,7 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_49 = paddle._C_ops.add(matmul_41, parameter_110) - del parameter_110 + del matmul_41, parameter_110 # pd_op.matmul: (1x11x768xf32) <- (1x11x768xf32, 768x768xf32) matmul_42 = paddle._C_ops.matmul(layer_norm_30, parameter_109, False, False) @@ -1228,10 +1146,11 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_50 = paddle._C_ops.add(matmul_42, parameter_108) - del parameter_108 + del matmul_42, parameter_108 # pd_op.reshape: (1x11x12x64xf32) <- (1x11x768xf32, 4xi64) reshape_21 = paddle._C_ops.reshape(add_49, full_int_array_1) + del add_49 # pd_op.transpose: (1x12x11x64xf32) <- (1x11x12x64xf32) transpose_21 = paddle._C_ops.transpose(reshape_21, [0, 2, 1, 3]) @@ -1239,6 +1158,7 @@ def forward( # pd_op.reshape: (1x11x12x64xf32) <- (1x11x768xf32, 4xi64) reshape_22 = paddle._C_ops.reshape(add_50, full_int_array_1) + del add_50 # pd_op.transpose: (1x12x11x64xf32) <- (1x11x12x64xf32) transpose_22 = paddle._C_ops.transpose(reshape_22, [0, 2, 1, 3]) @@ -1250,9 +1170,11 @@ def forward( # pd_op.matmul: (1x12x11x11xf32) <- (1x12x11x64xf32, 1x12x11x64xf32) matmul_43 = paddle._C_ops.matmul(scale_7, transpose_21, False, True) + del scale_7, transpose_21 # pd_op.add: (1x12x11x11xf32) <- (1x12x11x11xf32, 1x1x1x11xf32) add_51 = paddle._C_ops.add(matmul_43, unsqueeze_0) + del matmul_43 # pd_op.softmax: (1x12x11x11xf32) <- (1x12x11x11xf32) softmax_5 = paddle._C_ops.softmax(add_51, -1) @@ -1261,13 +1183,15 @@ def forward( # pd_op.dropout: (1x12x11x11xf32, 1x12x11x11xui8) <- (1x12x11x11xf32, None, 1xf32) dropout_32, dropout_33 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_5, None, full_5, False, "upscale_in_train", 0, False + softmax_5, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_5 # pd_op.matmul: (1x12x11x64xf32) <- (1x12x11x11xf32, 1x12x11x64xf32) matmul_44 = paddle._C_ops.matmul(dropout_32, transpose_22, False, False) + del dropout_32, transpose_22 # pd_op.transpose: (1x11x12x64xf32) <- (1x12x11x64xf32) transpose_23 = paddle._C_ops.transpose(matmul_44, [0, 2, 1, 3]) @@ -1275,19 +1199,20 @@ def forward( # pd_op.reshape: (1x11x768xf32) <- (1x11x12x64xf32, 3xi64) reshape_23 = paddle._C_ops.reshape(transpose_23, full_int_array_2) + del transpose_23 # pd_op.matmul: (1x11x768xf32) <- (1x11x768xf32, 768x768xf32) matmul_45 = paddle._C_ops.matmul(reshape_23, parameter_107, False, False) - del parameter_107 + del parameter_107, reshape_23 # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_52 = paddle._C_ops.add(matmul_45, parameter_106) - del parameter_106 + del matmul_45, parameter_106 # pd_op.dropout: (1x11x768xf32, 1x11x768xui8) <- (1x11x768xf32, None, 1xf32) dropout_34, dropout_35 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_52, None, full_5, False, "upscale_in_train", 0, False + add_52, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1295,6 +1220,7 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 1x11x768xf32) add_53 = paddle._C_ops.add(layer_norm_30, dropout_34) + del dropout_34, layer_norm_30 # pd_op.layer_norm: (1x11x768xf32, 1x11xf32, 1x11xf32) <- (1x11x768xf32, 768xf32, 768xf32) layer_norm_33, layer_norm_34, layer_norm_35 = (lambda x, f: f(x))( @@ -1303,7 +1229,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_100, parameter_101 + del add_53, parameter_100, parameter_101 # pd_op.matmul: (1x11x3072xf32) <- (1x11x768xf32, 768x3072xf32) matmul_46 = paddle._C_ops.matmul(layer_norm_33, parameter_105, False, False) @@ -1311,23 +1237,24 @@ def forward( # pd_op.add: (1x11x3072xf32) <- (1x11x3072xf32, 3072xf32) add_54 = paddle._C_ops.add(matmul_46, parameter_104) - del parameter_104 + del matmul_46, parameter_104 # pd_op.gelu: (1x11x3072xf32) <- (1x11x3072xf32) gelu_5 = paddle._C_ops.gelu(add_54, False) + del add_54 # pd_op.matmul: (1x11x768xf32) <- (1x11x3072xf32, 3072x768xf32) matmul_47 = paddle._C_ops.matmul(gelu_5, parameter_103, False, False) - del parameter_103 + del gelu_5, parameter_103 # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_55 = paddle._C_ops.add(matmul_47, parameter_102) - del parameter_102 + del matmul_47, parameter_102 # pd_op.dropout: (1x11x768xf32, 1x11x768xui8) <- (1x11x768xf32, None, 1xf32) dropout_36, dropout_37 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_55, None, full_5, False, "upscale_in_train", 0, False + add_55, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1335,6 +1262,7 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 1x11x768xf32) add_56 = paddle._C_ops.add(layer_norm_33, dropout_36) + del dropout_36, layer_norm_33 # pd_op.layer_norm: (1x11x768xf32, 1x11xf32, 1x11xf32) <- (1x11x768xf32, 768xf32, 768xf32) layer_norm_36, layer_norm_37, layer_norm_38 = (lambda x, f: f(x))( @@ -1343,7 +1271,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_98, parameter_99 + del add_56, parameter_98, parameter_99 # pd_op.matmul: (1x11x768xf32) <- (1x11x768xf32, 768x768xf32) matmul_48 = paddle._C_ops.matmul(layer_norm_36, parameter_97, False, False) @@ -1351,10 +1279,11 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_57 = paddle._C_ops.add(matmul_48, parameter_96) - del parameter_96 + del matmul_48, parameter_96 # pd_op.reshape: (1x11x12x64xf32) <- (1x11x768xf32, 4xi64) reshape_24 = paddle._C_ops.reshape(add_57, full_int_array_1) + del add_57 # pd_op.transpose: (1x12x11x64xf32) <- (1x11x12x64xf32) transpose_24 = paddle._C_ops.transpose(reshape_24, [0, 2, 1, 3]) @@ -1366,7 +1295,7 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_58 = paddle._C_ops.add(matmul_49, parameter_94) - del parameter_94 + del matmul_49, parameter_94 # pd_op.matmul: (1x11x768xf32) <- (1x11x768xf32, 768x768xf32) matmul_50 = paddle._C_ops.matmul(layer_norm_36, parameter_93, False, False) @@ -1374,10 +1303,11 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_59 = paddle._C_ops.add(matmul_50, parameter_92) - del parameter_92 + del matmul_50, parameter_92 # pd_op.reshape: (1x11x12x64xf32) <- (1x11x768xf32, 4xi64) reshape_25 = paddle._C_ops.reshape(add_58, full_int_array_1) + del add_58 # pd_op.transpose: (1x12x11x64xf32) <- (1x11x12x64xf32) transpose_25 = paddle._C_ops.transpose(reshape_25, [0, 2, 1, 3]) @@ -1385,6 +1315,7 @@ def forward( # pd_op.reshape: (1x11x12x64xf32) <- (1x11x768xf32, 4xi64) reshape_26 = paddle._C_ops.reshape(add_59, full_int_array_1) + del add_59 # pd_op.transpose: (1x12x11x64xf32) <- (1x11x12x64xf32) transpose_26 = paddle._C_ops.transpose(reshape_26, [0, 2, 1, 3]) @@ -1396,9 +1327,11 @@ def forward( # pd_op.matmul: (1x12x11x11xf32) <- (1x12x11x64xf32, 1x12x11x64xf32) matmul_51 = paddle._C_ops.matmul(scale_8, transpose_25, False, True) + del scale_8, transpose_25 # pd_op.add: (1x12x11x11xf32) <- (1x12x11x11xf32, 1x1x1x11xf32) add_60 = paddle._C_ops.add(matmul_51, unsqueeze_0) + del matmul_51 # pd_op.softmax: (1x12x11x11xf32) <- (1x12x11x11xf32) softmax_6 = paddle._C_ops.softmax(add_60, -1) @@ -1407,13 +1340,15 @@ def forward( # pd_op.dropout: (1x12x11x11xf32, 1x12x11x11xui8) <- (1x12x11x11xf32, None, 1xf32) dropout_38, dropout_39 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_6, None, full_5, False, "upscale_in_train", 0, False + softmax_6, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_6 # pd_op.matmul: (1x12x11x64xf32) <- (1x12x11x11xf32, 1x12x11x64xf32) matmul_52 = paddle._C_ops.matmul(dropout_38, transpose_26, False, False) + del dropout_38, transpose_26 # pd_op.transpose: (1x11x12x64xf32) <- (1x12x11x64xf32) transpose_27 = paddle._C_ops.transpose(matmul_52, [0, 2, 1, 3]) @@ -1421,19 +1356,20 @@ def forward( # pd_op.reshape: (1x11x768xf32) <- (1x11x12x64xf32, 3xi64) reshape_27 = paddle._C_ops.reshape(transpose_27, full_int_array_2) + del transpose_27 # pd_op.matmul: (1x11x768xf32) <- (1x11x768xf32, 768x768xf32) matmul_53 = paddle._C_ops.matmul(reshape_27, parameter_91, False, False) - del parameter_91 + del parameter_91, reshape_27 # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_61 = paddle._C_ops.add(matmul_53, parameter_90) - del parameter_90 + del matmul_53, parameter_90 # pd_op.dropout: (1x11x768xf32, 1x11x768xui8) <- (1x11x768xf32, None, 1xf32) dropout_40, dropout_41 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_61, None, full_5, False, "upscale_in_train", 0, False + add_61, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1441,6 +1377,7 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 1x11x768xf32) add_62 = paddle._C_ops.add(layer_norm_36, dropout_40) + del dropout_40, layer_norm_36 # pd_op.layer_norm: (1x11x768xf32, 1x11xf32, 1x11xf32) <- (1x11x768xf32, 768xf32, 768xf32) layer_norm_39, layer_norm_40, layer_norm_41 = (lambda x, f: f(x))( @@ -1449,7 +1386,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_84, parameter_85 + del add_62, parameter_84, parameter_85 # pd_op.matmul: (1x11x3072xf32) <- (1x11x768xf32, 768x3072xf32) matmul_54 = paddle._C_ops.matmul(layer_norm_39, parameter_89, False, False) @@ -1457,23 +1394,24 @@ def forward( # pd_op.add: (1x11x3072xf32) <- (1x11x3072xf32, 3072xf32) add_63 = paddle._C_ops.add(matmul_54, parameter_88) - del parameter_88 + del matmul_54, parameter_88 # pd_op.gelu: (1x11x3072xf32) <- (1x11x3072xf32) gelu_6 = paddle._C_ops.gelu(add_63, False) + del add_63 # pd_op.matmul: (1x11x768xf32) <- (1x11x3072xf32, 3072x768xf32) matmul_55 = paddle._C_ops.matmul(gelu_6, parameter_87, False, False) - del parameter_87 + del gelu_6, parameter_87 # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_64 = paddle._C_ops.add(matmul_55, parameter_86) - del parameter_86 + del matmul_55, parameter_86 # pd_op.dropout: (1x11x768xf32, 1x11x768xui8) <- (1x11x768xf32, None, 1xf32) dropout_42, dropout_43 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_64, None, full_5, False, "upscale_in_train", 0, False + add_64, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1481,6 +1419,7 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 1x11x768xf32) add_65 = paddle._C_ops.add(layer_norm_39, dropout_42) + del dropout_42, layer_norm_39 # pd_op.layer_norm: (1x11x768xf32, 1x11xf32, 1x11xf32) <- (1x11x768xf32, 768xf32, 768xf32) layer_norm_42, layer_norm_43, layer_norm_44 = (lambda x, f: f(x))( @@ -1489,7 +1428,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_82, parameter_83 + del add_65, parameter_82, parameter_83 # pd_op.matmul: (1x11x768xf32) <- (1x11x768xf32, 768x768xf32) matmul_56 = paddle._C_ops.matmul(layer_norm_42, parameter_81, False, False) @@ -1497,10 +1436,11 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_66 = paddle._C_ops.add(matmul_56, parameter_80) - del parameter_80 + del matmul_56, parameter_80 # pd_op.reshape: (1x11x12x64xf32) <- (1x11x768xf32, 4xi64) reshape_28 = paddle._C_ops.reshape(add_66, full_int_array_1) + del add_66 # pd_op.transpose: (1x12x11x64xf32) <- (1x11x12x64xf32) transpose_28 = paddle._C_ops.transpose(reshape_28, [0, 2, 1, 3]) @@ -1512,7 +1452,7 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_67 = paddle._C_ops.add(matmul_57, parameter_78) - del parameter_78 + del matmul_57, parameter_78 # pd_op.matmul: (1x11x768xf32) <- (1x11x768xf32, 768x768xf32) matmul_58 = paddle._C_ops.matmul(layer_norm_42, parameter_77, False, False) @@ -1520,10 +1460,11 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_68 = paddle._C_ops.add(matmul_58, parameter_76) - del parameter_76 + del matmul_58, parameter_76 # pd_op.reshape: (1x11x12x64xf32) <- (1x11x768xf32, 4xi64) reshape_29 = paddle._C_ops.reshape(add_67, full_int_array_1) + del add_67 # pd_op.transpose: (1x12x11x64xf32) <- (1x11x12x64xf32) transpose_29 = paddle._C_ops.transpose(reshape_29, [0, 2, 1, 3]) @@ -1531,6 +1472,7 @@ def forward( # pd_op.reshape: (1x11x12x64xf32) <- (1x11x768xf32, 4xi64) reshape_30 = paddle._C_ops.reshape(add_68, full_int_array_1) + del add_68 # pd_op.transpose: (1x12x11x64xf32) <- (1x11x12x64xf32) transpose_30 = paddle._C_ops.transpose(reshape_30, [0, 2, 1, 3]) @@ -1542,9 +1484,11 @@ def forward( # pd_op.matmul: (1x12x11x11xf32) <- (1x12x11x64xf32, 1x12x11x64xf32) matmul_59 = paddle._C_ops.matmul(scale_9, transpose_29, False, True) + del scale_9, transpose_29 # pd_op.add: (1x12x11x11xf32) <- (1x12x11x11xf32, 1x1x1x11xf32) add_69 = paddle._C_ops.add(matmul_59, unsqueeze_0) + del matmul_59 # pd_op.softmax: (1x12x11x11xf32) <- (1x12x11x11xf32) softmax_7 = paddle._C_ops.softmax(add_69, -1) @@ -1553,13 +1497,15 @@ def forward( # pd_op.dropout: (1x12x11x11xf32, 1x12x11x11xui8) <- (1x12x11x11xf32, None, 1xf32) dropout_44, dropout_45 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_7, None, full_5, False, "upscale_in_train", 0, False + softmax_7, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_7 # pd_op.matmul: (1x12x11x64xf32) <- (1x12x11x11xf32, 1x12x11x64xf32) matmul_60 = paddle._C_ops.matmul(dropout_44, transpose_30, False, False) + del dropout_44, transpose_30 # pd_op.transpose: (1x11x12x64xf32) <- (1x12x11x64xf32) transpose_31 = paddle._C_ops.transpose(matmul_60, [0, 2, 1, 3]) @@ -1567,19 +1513,20 @@ def forward( # pd_op.reshape: (1x11x768xf32) <- (1x11x12x64xf32, 3xi64) reshape_31 = paddle._C_ops.reshape(transpose_31, full_int_array_2) + del transpose_31 # pd_op.matmul: (1x11x768xf32) <- (1x11x768xf32, 768x768xf32) matmul_61 = paddle._C_ops.matmul(reshape_31, parameter_75, False, False) - del parameter_75 + del parameter_75, reshape_31 # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_70 = paddle._C_ops.add(matmul_61, parameter_74) - del parameter_74 + del matmul_61, parameter_74 # pd_op.dropout: (1x11x768xf32, 1x11x768xui8) <- (1x11x768xf32, None, 1xf32) dropout_46, dropout_47 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_70, None, full_5, False, "upscale_in_train", 0, False + add_70, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1587,6 +1534,7 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 1x11x768xf32) add_71 = paddle._C_ops.add(layer_norm_42, dropout_46) + del dropout_46, layer_norm_42 # pd_op.layer_norm: (1x11x768xf32, 1x11xf32, 1x11xf32) <- (1x11x768xf32, 768xf32, 768xf32) layer_norm_45, layer_norm_46, layer_norm_47 = (lambda x, f: f(x))( @@ -1595,7 +1543,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_68, parameter_69 + del add_71, parameter_68, parameter_69 # pd_op.matmul: (1x11x3072xf32) <- (1x11x768xf32, 768x3072xf32) matmul_62 = paddle._C_ops.matmul(layer_norm_45, parameter_73, False, False) @@ -1603,23 +1551,24 @@ def forward( # pd_op.add: (1x11x3072xf32) <- (1x11x3072xf32, 3072xf32) add_72 = paddle._C_ops.add(matmul_62, parameter_72) - del parameter_72 + del matmul_62, parameter_72 # pd_op.gelu: (1x11x3072xf32) <- (1x11x3072xf32) gelu_7 = paddle._C_ops.gelu(add_72, False) + del add_72 # pd_op.matmul: (1x11x768xf32) <- (1x11x3072xf32, 3072x768xf32) matmul_63 = paddle._C_ops.matmul(gelu_7, parameter_71, False, False) - del parameter_71 + del gelu_7, parameter_71 # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_73 = paddle._C_ops.add(matmul_63, parameter_70) - del parameter_70 + del matmul_63, parameter_70 # pd_op.dropout: (1x11x768xf32, 1x11x768xui8) <- (1x11x768xf32, None, 1xf32) dropout_48, dropout_49 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_73, None, full_5, False, "upscale_in_train", 0, False + add_73, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1627,6 +1576,7 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 1x11x768xf32) add_74 = paddle._C_ops.add(layer_norm_45, dropout_48) + del dropout_48, layer_norm_45 # pd_op.layer_norm: (1x11x768xf32, 1x11xf32, 1x11xf32) <- (1x11x768xf32, 768xf32, 768xf32) layer_norm_48, layer_norm_49, layer_norm_50 = (lambda x, f: f(x))( @@ -1635,7 +1585,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_66, parameter_67 + del add_74, parameter_66, parameter_67 # pd_op.matmul: (1x11x768xf32) <- (1x11x768xf32, 768x768xf32) matmul_64 = paddle._C_ops.matmul(layer_norm_48, parameter_65, False, False) @@ -1643,10 +1593,11 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_75 = paddle._C_ops.add(matmul_64, parameter_64) - del parameter_64 + del matmul_64, parameter_64 # pd_op.reshape: (1x11x12x64xf32) <- (1x11x768xf32, 4xi64) reshape_32 = paddle._C_ops.reshape(add_75, full_int_array_1) + del add_75 # pd_op.transpose: (1x12x11x64xf32) <- (1x11x12x64xf32) transpose_32 = paddle._C_ops.transpose(reshape_32, [0, 2, 1, 3]) @@ -1658,7 +1609,7 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_76 = paddle._C_ops.add(matmul_65, parameter_62) - del parameter_62 + del matmul_65, parameter_62 # pd_op.matmul: (1x11x768xf32) <- (1x11x768xf32, 768x768xf32) matmul_66 = paddle._C_ops.matmul(layer_norm_48, parameter_61, False, False) @@ -1666,10 +1617,11 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_77 = paddle._C_ops.add(matmul_66, parameter_60) - del parameter_60 + del matmul_66, parameter_60 # pd_op.reshape: (1x11x12x64xf32) <- (1x11x768xf32, 4xi64) reshape_33 = paddle._C_ops.reshape(add_76, full_int_array_1) + del add_76 # pd_op.transpose: (1x12x11x64xf32) <- (1x11x12x64xf32) transpose_33 = paddle._C_ops.transpose(reshape_33, [0, 2, 1, 3]) @@ -1677,6 +1629,7 @@ def forward( # pd_op.reshape: (1x11x12x64xf32) <- (1x11x768xf32, 4xi64) reshape_34 = paddle._C_ops.reshape(add_77, full_int_array_1) + del add_77 # pd_op.transpose: (1x12x11x64xf32) <- (1x11x12x64xf32) transpose_34 = paddle._C_ops.transpose(reshape_34, [0, 2, 1, 3]) @@ -1688,9 +1641,11 @@ def forward( # pd_op.matmul: (1x12x11x11xf32) <- (1x12x11x64xf32, 1x12x11x64xf32) matmul_67 = paddle._C_ops.matmul(scale_10, transpose_33, False, True) + del scale_10, transpose_33 # pd_op.add: (1x12x11x11xf32) <- (1x12x11x11xf32, 1x1x1x11xf32) add_78 = paddle._C_ops.add(matmul_67, unsqueeze_0) + del matmul_67 # pd_op.softmax: (1x12x11x11xf32) <- (1x12x11x11xf32) softmax_8 = paddle._C_ops.softmax(add_78, -1) @@ -1699,13 +1654,15 @@ def forward( # pd_op.dropout: (1x12x11x11xf32, 1x12x11x11xui8) <- (1x12x11x11xf32, None, 1xf32) dropout_50, dropout_51 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_8, None, full_5, False, "upscale_in_train", 0, False + softmax_8, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_8 # pd_op.matmul: (1x12x11x64xf32) <- (1x12x11x11xf32, 1x12x11x64xf32) matmul_68 = paddle._C_ops.matmul(dropout_50, transpose_34, False, False) + del dropout_50, transpose_34 # pd_op.transpose: (1x11x12x64xf32) <- (1x12x11x64xf32) transpose_35 = paddle._C_ops.transpose(matmul_68, [0, 2, 1, 3]) @@ -1713,19 +1670,20 @@ def forward( # pd_op.reshape: (1x11x768xf32) <- (1x11x12x64xf32, 3xi64) reshape_35 = paddle._C_ops.reshape(transpose_35, full_int_array_2) + del transpose_35 # pd_op.matmul: (1x11x768xf32) <- (1x11x768xf32, 768x768xf32) matmul_69 = paddle._C_ops.matmul(reshape_35, parameter_59, False, False) - del parameter_59 + del parameter_59, reshape_35 # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_79 = paddle._C_ops.add(matmul_69, parameter_58) - del parameter_58 + del matmul_69, parameter_58 # pd_op.dropout: (1x11x768xf32, 1x11x768xui8) <- (1x11x768xf32, None, 1xf32) dropout_52, dropout_53 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_79, None, full_5, False, "upscale_in_train", 0, False + add_79, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1733,6 +1691,7 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 1x11x768xf32) add_80 = paddle._C_ops.add(layer_norm_48, dropout_52) + del dropout_52, layer_norm_48 # pd_op.layer_norm: (1x11x768xf32, 1x11xf32, 1x11xf32) <- (1x11x768xf32, 768xf32, 768xf32) layer_norm_51, layer_norm_52, layer_norm_53 = (lambda x, f: f(x))( @@ -1741,7 +1700,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_52, parameter_53 + del add_80, parameter_52, parameter_53 # pd_op.matmul: (1x11x3072xf32) <- (1x11x768xf32, 768x3072xf32) matmul_70 = paddle._C_ops.matmul(layer_norm_51, parameter_57, False, False) @@ -1749,23 +1708,24 @@ def forward( # pd_op.add: (1x11x3072xf32) <- (1x11x3072xf32, 3072xf32) add_81 = paddle._C_ops.add(matmul_70, parameter_56) - del parameter_56 + del matmul_70, parameter_56 # pd_op.gelu: (1x11x3072xf32) <- (1x11x3072xf32) gelu_8 = paddle._C_ops.gelu(add_81, False) + del add_81 # pd_op.matmul: (1x11x768xf32) <- (1x11x3072xf32, 3072x768xf32) matmul_71 = paddle._C_ops.matmul(gelu_8, parameter_55, False, False) - del parameter_55 + del gelu_8, parameter_55 # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_82 = paddle._C_ops.add(matmul_71, parameter_54) - del parameter_54 + del matmul_71, parameter_54 # pd_op.dropout: (1x11x768xf32, 1x11x768xui8) <- (1x11x768xf32, None, 1xf32) dropout_54, dropout_55 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_82, None, full_5, False, "upscale_in_train", 0, False + add_82, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1773,6 +1733,7 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 1x11x768xf32) add_83 = paddle._C_ops.add(layer_norm_51, dropout_54) + del dropout_54, layer_norm_51 # pd_op.layer_norm: (1x11x768xf32, 1x11xf32, 1x11xf32) <- (1x11x768xf32, 768xf32, 768xf32) layer_norm_54, layer_norm_55, layer_norm_56 = (lambda x, f: f(x))( @@ -1781,7 +1742,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_50, parameter_51 + del add_83, parameter_50, parameter_51 # pd_op.matmul: (1x11x768xf32) <- (1x11x768xf32, 768x768xf32) matmul_72 = paddle._C_ops.matmul(layer_norm_54, parameter_49, False, False) @@ -1789,10 +1750,11 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_84 = paddle._C_ops.add(matmul_72, parameter_48) - del parameter_48 + del matmul_72, parameter_48 # pd_op.reshape: (1x11x12x64xf32) <- (1x11x768xf32, 4xi64) reshape_36 = paddle._C_ops.reshape(add_84, full_int_array_1) + del add_84 # pd_op.transpose: (1x12x11x64xf32) <- (1x11x12x64xf32) transpose_36 = paddle._C_ops.transpose(reshape_36, [0, 2, 1, 3]) @@ -1804,7 +1766,7 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_85 = paddle._C_ops.add(matmul_73, parameter_46) - del parameter_46 + del matmul_73, parameter_46 # pd_op.matmul: (1x11x768xf32) <- (1x11x768xf32, 768x768xf32) matmul_74 = paddle._C_ops.matmul(layer_norm_54, parameter_45, False, False) @@ -1812,10 +1774,11 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_86 = paddle._C_ops.add(matmul_74, parameter_44) - del parameter_44 + del matmul_74, parameter_44 # pd_op.reshape: (1x11x12x64xf32) <- (1x11x768xf32, 4xi64) reshape_37 = paddle._C_ops.reshape(add_85, full_int_array_1) + del add_85 # pd_op.transpose: (1x12x11x64xf32) <- (1x11x12x64xf32) transpose_37 = paddle._C_ops.transpose(reshape_37, [0, 2, 1, 3]) @@ -1823,6 +1786,7 @@ def forward( # pd_op.reshape: (1x11x12x64xf32) <- (1x11x768xf32, 4xi64) reshape_38 = paddle._C_ops.reshape(add_86, full_int_array_1) + del add_86 # pd_op.transpose: (1x12x11x64xf32) <- (1x11x12x64xf32) transpose_38 = paddle._C_ops.transpose(reshape_38, [0, 2, 1, 3]) @@ -1834,9 +1798,11 @@ def forward( # pd_op.matmul: (1x12x11x11xf32) <- (1x12x11x64xf32, 1x12x11x64xf32) matmul_75 = paddle._C_ops.matmul(scale_11, transpose_37, False, True) + del scale_11, transpose_37 # pd_op.add: (1x12x11x11xf32) <- (1x12x11x11xf32, 1x1x1x11xf32) add_87 = paddle._C_ops.add(matmul_75, unsqueeze_0) + del matmul_75 # pd_op.softmax: (1x12x11x11xf32) <- (1x12x11x11xf32) softmax_9 = paddle._C_ops.softmax(add_87, -1) @@ -1845,13 +1811,15 @@ def forward( # pd_op.dropout: (1x12x11x11xf32, 1x12x11x11xui8) <- (1x12x11x11xf32, None, 1xf32) dropout_56, dropout_57 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_9, None, full_5, False, "upscale_in_train", 0, False + softmax_9, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_9 # pd_op.matmul: (1x12x11x64xf32) <- (1x12x11x11xf32, 1x12x11x64xf32) matmul_76 = paddle._C_ops.matmul(dropout_56, transpose_38, False, False) + del dropout_56, transpose_38 # pd_op.transpose: (1x11x12x64xf32) <- (1x12x11x64xf32) transpose_39 = paddle._C_ops.transpose(matmul_76, [0, 2, 1, 3]) @@ -1859,19 +1827,20 @@ def forward( # pd_op.reshape: (1x11x768xf32) <- (1x11x12x64xf32, 3xi64) reshape_39 = paddle._C_ops.reshape(transpose_39, full_int_array_2) + del transpose_39 # pd_op.matmul: (1x11x768xf32) <- (1x11x768xf32, 768x768xf32) matmul_77 = paddle._C_ops.matmul(reshape_39, parameter_43, False, False) - del parameter_43 + del parameter_43, reshape_39 # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_88 = paddle._C_ops.add(matmul_77, parameter_42) - del parameter_42 + del matmul_77, parameter_42 # pd_op.dropout: (1x11x768xf32, 1x11x768xui8) <- (1x11x768xf32, None, 1xf32) dropout_58, dropout_59 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_88, None, full_5, False, "upscale_in_train", 0, False + add_88, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1879,6 +1848,7 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 1x11x768xf32) add_89 = paddle._C_ops.add(layer_norm_54, dropout_58) + del dropout_58, layer_norm_54 # pd_op.layer_norm: (1x11x768xf32, 1x11xf32, 1x11xf32) <- (1x11x768xf32, 768xf32, 768xf32) layer_norm_57, layer_norm_58, layer_norm_59 = (lambda x, f: f(x))( @@ -1887,7 +1857,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_36, parameter_37 + del add_89, parameter_36, parameter_37 # pd_op.matmul: (1x11x3072xf32) <- (1x11x768xf32, 768x3072xf32) matmul_78 = paddle._C_ops.matmul(layer_norm_57, parameter_41, False, False) @@ -1895,23 +1865,24 @@ def forward( # pd_op.add: (1x11x3072xf32) <- (1x11x3072xf32, 3072xf32) add_90 = paddle._C_ops.add(matmul_78, parameter_40) - del parameter_40 + del matmul_78, parameter_40 # pd_op.gelu: (1x11x3072xf32) <- (1x11x3072xf32) gelu_9 = paddle._C_ops.gelu(add_90, False) + del add_90 # pd_op.matmul: (1x11x768xf32) <- (1x11x3072xf32, 3072x768xf32) matmul_79 = paddle._C_ops.matmul(gelu_9, parameter_39, False, False) - del parameter_39 + del gelu_9, parameter_39 # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_91 = paddle._C_ops.add(matmul_79, parameter_38) - del parameter_38 + del matmul_79, parameter_38 # pd_op.dropout: (1x11x768xf32, 1x11x768xui8) <- (1x11x768xf32, None, 1xf32) dropout_60, dropout_61 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_91, None, full_5, False, "upscale_in_train", 0, False + add_91, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1919,6 +1890,7 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 1x11x768xf32) add_92 = paddle._C_ops.add(layer_norm_57, dropout_60) + del dropout_60, layer_norm_57 # pd_op.layer_norm: (1x11x768xf32, 1x11xf32, 1x11xf32) <- (1x11x768xf32, 768xf32, 768xf32) layer_norm_60, layer_norm_61, layer_norm_62 = (lambda x, f: f(x))( @@ -1927,7 +1899,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_34, parameter_35 + del add_92, parameter_34, parameter_35 # pd_op.matmul: (1x11x768xf32) <- (1x11x768xf32, 768x768xf32) matmul_80 = paddle._C_ops.matmul(layer_norm_60, parameter_33, False, False) @@ -1935,10 +1907,11 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_93 = paddle._C_ops.add(matmul_80, parameter_32) - del parameter_32 + del matmul_80, parameter_32 # pd_op.reshape: (1x11x12x64xf32) <- (1x11x768xf32, 4xi64) reshape_40 = paddle._C_ops.reshape(add_93, full_int_array_1) + del add_93 # pd_op.transpose: (1x12x11x64xf32) <- (1x11x12x64xf32) transpose_40 = paddle._C_ops.transpose(reshape_40, [0, 2, 1, 3]) @@ -1950,7 +1923,7 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_94 = paddle._C_ops.add(matmul_81, parameter_30) - del parameter_30 + del matmul_81, parameter_30 # pd_op.matmul: (1x11x768xf32) <- (1x11x768xf32, 768x768xf32) matmul_82 = paddle._C_ops.matmul(layer_norm_60, parameter_29, False, False) @@ -1958,10 +1931,11 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_95 = paddle._C_ops.add(matmul_82, parameter_28) - del parameter_28 + del matmul_82, parameter_28 # pd_op.reshape: (1x11x12x64xf32) <- (1x11x768xf32, 4xi64) reshape_41 = paddle._C_ops.reshape(add_94, full_int_array_1) + del add_94 # pd_op.transpose: (1x12x11x64xf32) <- (1x11x12x64xf32) transpose_41 = paddle._C_ops.transpose(reshape_41, [0, 2, 1, 3]) @@ -1969,6 +1943,7 @@ def forward( # pd_op.reshape: (1x11x12x64xf32) <- (1x11x768xf32, 4xi64) reshape_42 = paddle._C_ops.reshape(add_95, full_int_array_1) + del add_95 # pd_op.transpose: (1x12x11x64xf32) <- (1x11x12x64xf32) transpose_42 = paddle._C_ops.transpose(reshape_42, [0, 2, 1, 3]) @@ -1980,9 +1955,11 @@ def forward( # pd_op.matmul: (1x12x11x11xf32) <- (1x12x11x64xf32, 1x12x11x64xf32) matmul_83 = paddle._C_ops.matmul(scale_12, transpose_41, False, True) + del scale_12, transpose_41 # pd_op.add: (1x12x11x11xf32) <- (1x12x11x11xf32, 1x1x1x11xf32) add_96 = paddle._C_ops.add(matmul_83, unsqueeze_0) + del matmul_83 # pd_op.softmax: (1x12x11x11xf32) <- (1x12x11x11xf32) softmax_10 = paddle._C_ops.softmax(add_96, -1) @@ -1991,13 +1968,15 @@ def forward( # pd_op.dropout: (1x12x11x11xf32, 1x12x11x11xui8) <- (1x12x11x11xf32, None, 1xf32) dropout_62, dropout_63 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_10, None, full_5, False, "upscale_in_train", 0, False + softmax_10, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_10 # pd_op.matmul: (1x12x11x64xf32) <- (1x12x11x11xf32, 1x12x11x64xf32) matmul_84 = paddle._C_ops.matmul(dropout_62, transpose_42, False, False) + del dropout_62, transpose_42 # pd_op.transpose: (1x11x12x64xf32) <- (1x12x11x64xf32) transpose_43 = paddle._C_ops.transpose(matmul_84, [0, 2, 1, 3]) @@ -2005,19 +1984,20 @@ def forward( # pd_op.reshape: (1x11x768xf32) <- (1x11x12x64xf32, 3xi64) reshape_43 = paddle._C_ops.reshape(transpose_43, full_int_array_2) + del transpose_43 # pd_op.matmul: (1x11x768xf32) <- (1x11x768xf32, 768x768xf32) matmul_85 = paddle._C_ops.matmul(reshape_43, parameter_27, False, False) - del parameter_27 + del parameter_27, reshape_43 # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_97 = paddle._C_ops.add(matmul_85, parameter_26) - del parameter_26 + del matmul_85, parameter_26 # pd_op.dropout: (1x11x768xf32, 1x11x768xui8) <- (1x11x768xf32, None, 1xf32) dropout_64, dropout_65 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_97, None, full_5, False, "upscale_in_train", 0, False + add_97, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -2025,6 +2005,7 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 1x11x768xf32) add_98 = paddle._C_ops.add(layer_norm_60, dropout_64) + del dropout_64, layer_norm_60 # pd_op.layer_norm: (1x11x768xf32, 1x11xf32, 1x11xf32) <- (1x11x768xf32, 768xf32, 768xf32) layer_norm_63, layer_norm_64, layer_norm_65 = (lambda x, f: f(x))( @@ -2033,7 +2014,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_20, parameter_21 + del add_98, parameter_20, parameter_21 # pd_op.matmul: (1x11x3072xf32) <- (1x11x768xf32, 768x3072xf32) matmul_86 = paddle._C_ops.matmul(layer_norm_63, parameter_25, False, False) @@ -2041,23 +2022,24 @@ def forward( # pd_op.add: (1x11x3072xf32) <- (1x11x3072xf32, 3072xf32) add_99 = paddle._C_ops.add(matmul_86, parameter_24) - del parameter_24 + del matmul_86, parameter_24 # pd_op.gelu: (1x11x3072xf32) <- (1x11x3072xf32) gelu_10 = paddle._C_ops.gelu(add_99, False) + del add_99 # pd_op.matmul: (1x11x768xf32) <- (1x11x3072xf32, 3072x768xf32) matmul_87 = paddle._C_ops.matmul(gelu_10, parameter_23, False, False) - del parameter_23 + del gelu_10, parameter_23 # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_100 = paddle._C_ops.add(matmul_87, parameter_22) - del parameter_22 + del matmul_87, parameter_22 # pd_op.dropout: (1x11x768xf32, 1x11x768xui8) <- (1x11x768xf32, None, 1xf32) dropout_66, dropout_67 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_100, None, full_5, False, "upscale_in_train", 0, False + add_100, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -2065,6 +2047,7 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 1x11x768xf32) add_101 = paddle._C_ops.add(layer_norm_63, dropout_66) + del dropout_66, layer_norm_63 # pd_op.layer_norm: (1x11x768xf32, 1x11xf32, 1x11xf32) <- (1x11x768xf32, 768xf32, 768xf32) layer_norm_66, layer_norm_67, layer_norm_68 = (lambda x, f: f(x))( @@ -2073,7 +2056,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_18, parameter_19 + del add_101, parameter_18, parameter_19 # pd_op.matmul: (1x11x768xf32) <- (1x11x768xf32, 768x768xf32) matmul_88 = paddle._C_ops.matmul(layer_norm_66, parameter_17, False, False) @@ -2081,10 +2064,11 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_102 = paddle._C_ops.add(matmul_88, parameter_16) - del parameter_16 + del matmul_88, parameter_16 # pd_op.reshape: (1x11x12x64xf32) <- (1x11x768xf32, 4xi64) reshape_44 = paddle._C_ops.reshape(add_102, full_int_array_1) + del add_102 # pd_op.transpose: (1x12x11x64xf32) <- (1x11x12x64xf32) transpose_44 = paddle._C_ops.transpose(reshape_44, [0, 2, 1, 3]) @@ -2096,7 +2080,7 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_103 = paddle._C_ops.add(matmul_89, parameter_14) - del parameter_14 + del matmul_89, parameter_14 # pd_op.matmul: (1x11x768xf32) <- (1x11x768xf32, 768x768xf32) matmul_90 = paddle._C_ops.matmul(layer_norm_66, parameter_13, False, False) @@ -2104,10 +2088,11 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_104 = paddle._C_ops.add(matmul_90, parameter_12) - del parameter_12 + del matmul_90, parameter_12 # pd_op.reshape: (1x11x12x64xf32) <- (1x11x768xf32, 4xi64) reshape_45 = paddle._C_ops.reshape(add_103, full_int_array_1) + del add_103 # pd_op.transpose: (1x12x11x64xf32) <- (1x11x12x64xf32) transpose_45 = paddle._C_ops.transpose(reshape_45, [0, 2, 1, 3]) @@ -2115,7 +2100,7 @@ def forward( # pd_op.reshape: (1x11x12x64xf32) <- (1x11x768xf32, 4xi64) reshape_46 = paddle._C_ops.reshape(add_104, full_int_array_1) - del full_int_array_1 + del add_104, full_int_array_1 # pd_op.transpose: (1x12x11x64xf32) <- (1x11x12x64xf32) transpose_46 = paddle._C_ops.transpose(reshape_46, [0, 2, 1, 3]) @@ -2123,13 +2108,15 @@ def forward( # pd_op.scale: (1x12x11x64xf32) <- (1x12x11x64xf32, 1xf32) scale_13 = paddle._C_ops.scale(transpose_44, full_6, float("0"), True) - del transpose_44 + del full_6, transpose_44 # pd_op.matmul: (1x12x11x11xf32) <- (1x12x11x64xf32, 1x12x11x64xf32) matmul_91 = paddle._C_ops.matmul(scale_13, transpose_45, False, True) + del scale_13, transpose_45 # pd_op.add: (1x12x11x11xf32) <- (1x12x11x11xf32, 1x1x1x11xf32) add_105 = paddle._C_ops.add(matmul_91, unsqueeze_0) + del matmul_91, unsqueeze_0 # pd_op.softmax: (1x12x11x11xf32) <- (1x12x11x11xf32) softmax_11 = paddle._C_ops.softmax(add_105, -1) @@ -2138,13 +2125,15 @@ def forward( # pd_op.dropout: (1x12x11x11xf32, 1x12x11x11xui8) <- (1x12x11x11xf32, None, 1xf32) dropout_68, dropout_69 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_11, None, full_5, False, "upscale_in_train", 0, False + softmax_11, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_11 # pd_op.matmul: (1x12x11x64xf32) <- (1x12x11x11xf32, 1x12x11x64xf32) matmul_92 = paddle._C_ops.matmul(dropout_68, transpose_46, False, False) + del dropout_68, transpose_46 # pd_op.transpose: (1x11x12x64xf32) <- (1x12x11x64xf32) transpose_47 = paddle._C_ops.transpose(matmul_92, [0, 2, 1, 3]) @@ -2152,20 +2141,20 @@ def forward( # pd_op.reshape: (1x11x768xf32) <- (1x11x12x64xf32, 3xi64) reshape_47 = paddle._C_ops.reshape(transpose_47, full_int_array_2) - del full_int_array_2 + del full_int_array_2, transpose_47 # pd_op.matmul: (1x11x768xf32) <- (1x11x768xf32, 768x768xf32) matmul_93 = paddle._C_ops.matmul(reshape_47, parameter_11, False, False) - del parameter_11 + del parameter_11, reshape_47 # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_106 = paddle._C_ops.add(matmul_93, parameter_10) - del parameter_10 + del matmul_93, parameter_10 # pd_op.dropout: (1x11x768xf32, 1x11x768xui8) <- (1x11x768xf32, None, 1xf32) dropout_70, dropout_71 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_106, None, full_5, False, "upscale_in_train", 0, False + add_106, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -2173,6 +2162,7 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 1x11x768xf32) add_107 = paddle._C_ops.add(layer_norm_66, dropout_70) + del dropout_70, layer_norm_66 # pd_op.layer_norm: (1x11x768xf32, 1x11xf32, 1x11xf32) <- (1x11x768xf32, 768xf32, 768xf32) layer_norm_69, layer_norm_70, layer_norm_71 = (lambda x, f: f(x))( @@ -2181,7 +2171,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_4, parameter_5 + del add_107, parameter_4, parameter_5 # pd_op.matmul: (1x11x3072xf32) <- (1x11x768xf32, 768x3072xf32) matmul_94 = paddle._C_ops.matmul(layer_norm_69, parameter_9, False, False) @@ -2189,30 +2179,32 @@ def forward( # pd_op.add: (1x11x3072xf32) <- (1x11x3072xf32, 3072xf32) add_108 = paddle._C_ops.add(matmul_94, parameter_8) - del parameter_8 + del matmul_94, parameter_8 # pd_op.gelu: (1x11x3072xf32) <- (1x11x3072xf32) gelu_11 = paddle._C_ops.gelu(add_108, False) + del add_108 # pd_op.matmul: (1x11x768xf32) <- (1x11x3072xf32, 3072x768xf32) matmul_95 = paddle._C_ops.matmul(gelu_11, parameter_7, False, False) - del parameter_7 + del gelu_11, parameter_7 # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_109 = paddle._C_ops.add(matmul_95, parameter_6) - del parameter_6 + del matmul_95, parameter_6 # pd_op.dropout: (1x11x768xf32, 1x11x768xui8) <- (1x11x768xf32, None, 1xf32) dropout_72, dropout_73 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_109, None, full_5, False, "upscale_in_train", 0, False + add_109, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) - del add_109 + del add_109, full_5 # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 1x11x768xf32) add_110 = paddle._C_ops.add(layer_norm_69, dropout_72) + del dropout_72, layer_norm_69 # pd_op.layer_norm: (1x11x768xf32, 1x11xf32, 1x11xf32) <- (1x11x768xf32, 768xf32, 768xf32) layer_norm_72, layer_norm_73, layer_norm_74 = (lambda x, f: f(x))( @@ -2221,7 +2213,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_2, parameter_3 + del add_110, parameter_2, parameter_3 # pd_op.full_int_array: (1xi64) <- () full_int_array_3 = [0] @@ -2233,470 +2225,18 @@ def forward( slice_0 = paddle._C_ops.slice( layer_norm_72, [1], full_int_array_3, full_int_array_4, [1], [1] ) + del full_int_array_3, full_int_array_4, layer_norm_72 # pd_op.matmul: (1x768xf32) <- (1x768xf32, 768x768xf32) matmul_96 = paddle._C_ops.matmul(slice_0, parameter_1, False, False) - del parameter_1 + del parameter_1, slice_0 # pd_op.add: (1x768xf32) <- (1x768xf32, 768xf32) add_111 = paddle._C_ops.add(matmul_96, parameter_0) - del parameter_0 + del matmul_96, parameter_0 # pd_op.tanh: (1x768xf32) <- (1x768xf32) tanh_0 = paddle._C_ops.tanh(add_111) - del ( - add_0, - add_1, - add_101, - add_102, - add_103, - add_104, - add_107, - add_108, - add_11, - add_110, - add_111, - add_12, - add_13, - add_14, - add_17, - add_18, - add_2, - add_20, - add_21, - add_22, - add_23, - add_26, - add_27, - add_29, - add_3, - add_30, - add_31, - add_32, - add_35, - add_36, - add_38, - add_39, - add_4, - add_40, - add_41, - add_44, - add_45, - add_47, - add_48, - add_49, - add_5, - add_50, - add_53, - add_54, - add_56, - add_57, - add_58, - add_59, - add_62, - add_63, - add_65, - add_66, - add_67, - add_68, - add_71, - add_72, - add_74, - add_75, - add_76, - add_77, - add_8, - add_80, - add_81, - add_83, - add_84, - add_85, - add_86, - add_89, - add_9, - add_90, - add_92, - add_93, - add_94, - add_95, - add_98, - add_99, - assign_0, - assign_1, - assign_10, - assign_11, - assign_12, - assign_13, - assign_14, - assign_15, - assign_16, - assign_17, - assign_18, - assign_19, - assign_2, - assign_20, - assign_21, - assign_22, - assign_23, - assign_24, - assign_25, - assign_26, - assign_27, - assign_28, - assign_29, - assign_3, - assign_30, - assign_31, - assign_32, - assign_33, - assign_34, - assign_35, - assign_36, - assign_37, - assign_38, - assign_39, - assign_4, - assign_40, - assign_41, - assign_42, - assign_43, - assign_44, - assign_45, - assign_46, - assign_5, - assign_6, - assign_7, - assign_8, - assign_9, - dropout_0, - dropout_1, - dropout_10, - dropout_11, - dropout_12, - dropout_13, - dropout_14, - dropout_15, - dropout_16, - dropout_17, - dropout_18, - dropout_19, - dropout_2, - dropout_20, - dropout_21, - dropout_22, - dropout_23, - dropout_24, - dropout_25, - dropout_26, - dropout_27, - dropout_28, - dropout_29, - dropout_3, - dropout_30, - dropout_31, - dropout_32, - dropout_33, - dropout_34, - dropout_35, - dropout_36, - dropout_37, - dropout_38, - dropout_39, - dropout_4, - dropout_40, - dropout_41, - dropout_42, - dropout_43, - dropout_44, - dropout_45, - dropout_46, - dropout_47, - dropout_48, - dropout_49, - dropout_5, - dropout_50, - dropout_51, - dropout_52, - dropout_53, - dropout_54, - dropout_55, - dropout_56, - dropout_57, - dropout_58, - dropout_59, - dropout_6, - dropout_60, - dropout_61, - dropout_62, - dropout_63, - dropout_64, - dropout_65, - dropout_66, - dropout_67, - dropout_68, - dropout_69, - dropout_7, - dropout_70, - dropout_71, - dropout_72, - dropout_73, - dropout_8, - dropout_9, - embedding_0, - embedding_1, - embedding_2, - embedding_3, - full_5, - full_6, - full_int_array_3, - full_int_array_4, - gelu_0, - gelu_1, - gelu_10, - gelu_11, - gelu_2, - gelu_3, - gelu_4, - gelu_5, - gelu_6, - gelu_7, - gelu_8, - gelu_9, - layer_norm_1, - layer_norm_10, - layer_norm_11, - layer_norm_12, - layer_norm_13, - layer_norm_14, - layer_norm_15, - layer_norm_16, - layer_norm_17, - layer_norm_18, - layer_norm_19, - layer_norm_2, - layer_norm_20, - layer_norm_21, - layer_norm_22, - layer_norm_23, - layer_norm_24, - layer_norm_25, - layer_norm_26, - layer_norm_27, - layer_norm_28, - layer_norm_29, - layer_norm_3, - layer_norm_30, - layer_norm_31, - layer_norm_32, - layer_norm_33, - layer_norm_34, - layer_norm_35, - layer_norm_36, - layer_norm_37, - layer_norm_38, - layer_norm_39, - layer_norm_4, - layer_norm_40, - layer_norm_41, - layer_norm_42, - layer_norm_43, - layer_norm_44, - layer_norm_45, - layer_norm_46, - layer_norm_47, - layer_norm_48, - layer_norm_49, - layer_norm_5, - layer_norm_50, - layer_norm_51, - layer_norm_52, - layer_norm_53, - layer_norm_54, - layer_norm_55, - layer_norm_56, - layer_norm_57, - layer_norm_58, - layer_norm_59, - layer_norm_6, - layer_norm_60, - layer_norm_61, - layer_norm_62, - layer_norm_63, - layer_norm_64, - layer_norm_65, - layer_norm_66, - layer_norm_67, - layer_norm_68, - layer_norm_69, - layer_norm_7, - layer_norm_70, - layer_norm_71, - layer_norm_72, - layer_norm_73, - layer_norm_74, - layer_norm_8, - layer_norm_9, - matmul_0, - matmul_1, - matmul_10, - matmul_11, - matmul_13, - matmul_14, - matmul_15, - matmul_16, - matmul_17, - matmul_18, - matmul_19, - matmul_2, - matmul_21, - matmul_22, - matmul_23, - matmul_24, - matmul_25, - matmul_26, - matmul_27, - matmul_29, - matmul_3, - matmul_30, - matmul_31, - matmul_32, - matmul_33, - matmul_34, - matmul_35, - matmul_37, - matmul_38, - matmul_39, - matmul_40, - matmul_41, - matmul_42, - matmul_43, - matmul_45, - matmul_46, - matmul_47, - matmul_48, - matmul_49, - matmul_5, - matmul_50, - matmul_51, - matmul_53, - matmul_54, - matmul_55, - matmul_56, - matmul_57, - matmul_58, - matmul_59, - matmul_6, - matmul_61, - matmul_62, - matmul_63, - matmul_64, - matmul_65, - matmul_66, - matmul_67, - matmul_69, - matmul_7, - matmul_70, - matmul_71, - matmul_72, - matmul_73, - matmul_74, - matmul_75, - matmul_77, - matmul_78, - matmul_79, - matmul_8, - matmul_80, - matmul_81, - matmul_82, - matmul_83, - matmul_85, - matmul_86, - matmul_87, - matmul_88, - matmul_89, - matmul_9, - matmul_90, - matmul_91, - matmul_93, - matmul_94, - matmul_95, - matmul_96, - reshape_11, - reshape_15, - reshape_19, - reshape_23, - reshape_27, - reshape_3, - reshape_31, - reshape_35, - reshape_39, - reshape_43, - reshape_47, - reshape_7, - scale_1, - scale_10, - scale_11, - scale_12, - scale_13, - scale_2, - scale_3, - scale_4, - scale_5, - scale_6, - scale_7, - scale_8, - scale_9, - slice_0, - softmax_0, - softmax_1, - softmax_10, - softmax_11, - softmax_2, - softmax_3, - softmax_4, - softmax_5, - softmax_6, - softmax_7, - softmax_8, - softmax_9, - subtract_0, - transpose_1, - transpose_10, - transpose_11, - transpose_13, - transpose_14, - transpose_15, - transpose_17, - transpose_18, - transpose_19, - transpose_2, - transpose_21, - transpose_22, - transpose_23, - transpose_25, - transpose_26, - transpose_27, - transpose_29, - transpose_3, - transpose_30, - transpose_31, - transpose_33, - transpose_34, - transpose_35, - transpose_37, - transpose_38, - transpose_39, - transpose_41, - transpose_42, - transpose_43, - transpose_45, - transpose_46, - transpose_47, - transpose_5, - transpose_6, - transpose_7, - transpose_9, - unsqueeze_0, - ) + del add_111 return tanh_0 diff --git a/paddle_samples/PaddleNLP/ernie-3.0-base-zh/weight_meta.py b/paddle_samples/PaddleNLP/ernie-3.0-base-zh/weight_meta.py index 1e7dafcce..2340752d2 100644 --- a/paddle_samples/PaddleNLP/ernie-3.0-base-zh/weight_meta.py +++ b/paddle_samples/PaddleNLP/ernie-3.0-base-zh/weight_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_parameter_0: name = "parameter_0" + original_name = "linear_654.b_0" shape = [768] dtype = "float32" min_val = float("-0.498349") @@ -11,6 +12,7 @@ class Program_weight_tensor_parameter_0: class Program_weight_tensor_parameter_1: name = "parameter_1" + original_name = "linear_654.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.327039") @@ -22,6 +24,7 @@ class Program_weight_tensor_parameter_1: class Program_weight_tensor_parameter_2: name = "parameter_2" + original_name = "layer_norm_222.b_0" shape = [768] dtype = "float32" min_val = float("-0.866301") @@ -33,6 +36,7 @@ class Program_weight_tensor_parameter_2: class Program_weight_tensor_parameter_3: name = "parameter_3" + original_name = "layer_norm_222.w_0" shape = [768] dtype = "float32" min_val = float("0.0917045") @@ -44,6 +48,7 @@ class Program_weight_tensor_parameter_3: class Program_weight_tensor_parameter_4: name = "parameter_4" + original_name = "layer_norm_221.b_0" shape = [768] dtype = "float32" min_val = float("-2.51493") @@ -55,6 +60,7 @@ class Program_weight_tensor_parameter_4: class Program_weight_tensor_parameter_5: name = "parameter_5" + original_name = "layer_norm_221.w_0" shape = [768] dtype = "float32" min_val = float("0.126376") @@ -66,6 +72,7 @@ class Program_weight_tensor_parameter_5: class Program_weight_tensor_parameter_6: name = "parameter_6" + original_name = "linear_653.b_0" shape = [768] dtype = "float32" min_val = float("-1.63512") @@ -77,6 +84,7 @@ class Program_weight_tensor_parameter_6: class Program_weight_tensor_parameter_7: name = "parameter_7" + original_name = "linear_653.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-1.39808") @@ -88,6 +96,7 @@ class Program_weight_tensor_parameter_7: class Program_weight_tensor_parameter_8: name = "parameter_8" + original_name = "linear_652.b_0" shape = [3072] dtype = "float32" min_val = float("-2.39685") @@ -99,6 +108,7 @@ class Program_weight_tensor_parameter_8: class Program_weight_tensor_parameter_9: name = "parameter_9" + original_name = "linear_652.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.60895") @@ -110,6 +120,7 @@ class Program_weight_tensor_parameter_9: class Program_weight_tensor_parameter_10: name = "parameter_10" + original_name = "linear_651.b_0" shape = [768] dtype = "float32" min_val = float("-0.813321") @@ -121,6 +132,7 @@ class Program_weight_tensor_parameter_10: class Program_weight_tensor_parameter_11: name = "parameter_11" + original_name = "linear_651.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.648102") @@ -132,6 +144,7 @@ class Program_weight_tensor_parameter_11: class Program_weight_tensor_parameter_12: name = "parameter_12" + original_name = "linear_650.b_0" shape = [768] dtype = "float32" min_val = float("-0.81287") @@ -143,6 +156,7 @@ class Program_weight_tensor_parameter_12: class Program_weight_tensor_parameter_13: name = "parameter_13" + original_name = "linear_650.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.312481") @@ -154,6 +168,7 @@ class Program_weight_tensor_parameter_13: class Program_weight_tensor_parameter_14: name = "parameter_14" + original_name = "linear_649.b_0" shape = [768] dtype = "float32" min_val = float("-16.6109") @@ -165,6 +180,7 @@ class Program_weight_tensor_parameter_14: class Program_weight_tensor_parameter_15: name = "parameter_15" + original_name = "linear_649.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.653401") @@ -176,6 +192,7 @@ class Program_weight_tensor_parameter_15: class Program_weight_tensor_parameter_16: name = "parameter_16" + original_name = "linear_648.b_0" shape = [768] dtype = "float32" min_val = float("-3.43079") @@ -187,6 +204,7 @@ class Program_weight_tensor_parameter_16: class Program_weight_tensor_parameter_17: name = "parameter_17" + original_name = "linear_648.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.697798") @@ -198,6 +216,7 @@ class Program_weight_tensor_parameter_17: class Program_weight_tensor_parameter_18: name = "parameter_18" + original_name = "layer_norm_220.b_0" shape = [768] dtype = "float32" min_val = float("-3.47074") @@ -209,6 +228,7 @@ class Program_weight_tensor_parameter_18: class Program_weight_tensor_parameter_19: name = "parameter_19" + original_name = "layer_norm_220.w_0" shape = [768] dtype = "float32" min_val = float("0.180939") @@ -220,6 +240,7 @@ class Program_weight_tensor_parameter_19: class Program_weight_tensor_parameter_20: name = "parameter_20" + original_name = "layer_norm_219.b_0" shape = [768] dtype = "float32" min_val = float("-5.17208") @@ -231,6 +252,7 @@ class Program_weight_tensor_parameter_20: class Program_weight_tensor_parameter_21: name = "parameter_21" + original_name = "layer_norm_219.w_0" shape = [768] dtype = "float32" min_val = float("0.50495") @@ -242,6 +264,7 @@ class Program_weight_tensor_parameter_21: class Program_weight_tensor_parameter_22: name = "parameter_22" + original_name = "linear_647.b_0" shape = [768] dtype = "float32" min_val = float("-0.342883") @@ -253,6 +276,7 @@ class Program_weight_tensor_parameter_22: class Program_weight_tensor_parameter_23: name = "parameter_23" + original_name = "linear_647.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-0.964307") @@ -264,6 +288,7 @@ class Program_weight_tensor_parameter_23: class Program_weight_tensor_parameter_24: name = "parameter_24" + original_name = "linear_646.b_0" shape = [3072] dtype = "float32" min_val = float("-2.17151") @@ -275,6 +300,7 @@ class Program_weight_tensor_parameter_24: class Program_weight_tensor_parameter_25: name = "parameter_25" + original_name = "linear_646.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.599831") @@ -286,6 +312,7 @@ class Program_weight_tensor_parameter_25: class Program_weight_tensor_parameter_26: name = "parameter_26" + original_name = "linear_645.b_0" shape = [768] dtype = "float32" min_val = float("-0.246663") @@ -297,6 +324,7 @@ class Program_weight_tensor_parameter_26: class Program_weight_tensor_parameter_27: name = "parameter_27" + original_name = "linear_645.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.895001") @@ -308,6 +336,7 @@ class Program_weight_tensor_parameter_27: class Program_weight_tensor_parameter_28: name = "parameter_28" + original_name = "linear_644.b_0" shape = [768] dtype = "float32" min_val = float("-0.527457") @@ -319,6 +348,7 @@ class Program_weight_tensor_parameter_28: class Program_weight_tensor_parameter_29: name = "parameter_29" + original_name = "linear_644.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.357094") @@ -330,6 +360,7 @@ class Program_weight_tensor_parameter_29: class Program_weight_tensor_parameter_30: name = "parameter_30" + original_name = "linear_643.b_0" shape = [768] dtype = "float32" min_val = float("-18.3291") @@ -341,6 +372,7 @@ class Program_weight_tensor_parameter_30: class Program_weight_tensor_parameter_31: name = "parameter_31" + original_name = "linear_643.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.46863") @@ -352,6 +384,7 @@ class Program_weight_tensor_parameter_31: class Program_weight_tensor_parameter_32: name = "parameter_32" + original_name = "linear_642.b_0" shape = [768] dtype = "float32" min_val = float("-3.22364") @@ -363,6 +396,7 @@ class Program_weight_tensor_parameter_32: class Program_weight_tensor_parameter_33: name = "parameter_33" + original_name = "linear_642.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.516702") @@ -374,6 +408,7 @@ class Program_weight_tensor_parameter_33: class Program_weight_tensor_parameter_34: name = "parameter_34" + original_name = "layer_norm_218.b_0" shape = [768] dtype = "float32" min_val = float("-0.530658") @@ -385,6 +420,7 @@ class Program_weight_tensor_parameter_34: class Program_weight_tensor_parameter_35: name = "parameter_35" + original_name = "layer_norm_218.w_0" shape = [768] dtype = "float32" min_val = float("0.233655") @@ -396,6 +432,7 @@ class Program_weight_tensor_parameter_35: class Program_weight_tensor_parameter_36: name = "parameter_36" + original_name = "layer_norm_217.b_0" shape = [768] dtype = "float32" min_val = float("-5.74696") @@ -407,6 +444,7 @@ class Program_weight_tensor_parameter_36: class Program_weight_tensor_parameter_37: name = "parameter_37" + original_name = "layer_norm_217.w_0" shape = [768] dtype = "float32" min_val = float("0.476148") @@ -418,6 +456,7 @@ class Program_weight_tensor_parameter_37: class Program_weight_tensor_parameter_38: name = "parameter_38" + original_name = "linear_641.b_0" shape = [768] dtype = "float32" min_val = float("-1.21558") @@ -429,6 +468,7 @@ class Program_weight_tensor_parameter_38: class Program_weight_tensor_parameter_39: name = "parameter_39" + original_name = "linear_641.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-1.20498") @@ -440,6 +480,7 @@ class Program_weight_tensor_parameter_39: class Program_weight_tensor_parameter_40: name = "parameter_40" + original_name = "linear_640.b_0" shape = [3072] dtype = "float32" min_val = float("-1.41609") @@ -451,6 +492,7 @@ class Program_weight_tensor_parameter_40: class Program_weight_tensor_parameter_41: name = "parameter_41" + original_name = "linear_640.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.602099") @@ -462,6 +504,7 @@ class Program_weight_tensor_parameter_41: class Program_weight_tensor_parameter_42: name = "parameter_42" + original_name = "linear_639.b_0" shape = [768] dtype = "float32" min_val = float("-0.239171") @@ -473,6 +516,7 @@ class Program_weight_tensor_parameter_42: class Program_weight_tensor_parameter_43: name = "parameter_43" + original_name = "linear_639.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.813144") @@ -484,6 +528,7 @@ class Program_weight_tensor_parameter_43: class Program_weight_tensor_parameter_44: name = "parameter_44" + original_name = "linear_638.b_0" shape = [768] dtype = "float32" min_val = float("-0.808215") @@ -495,6 +540,7 @@ class Program_weight_tensor_parameter_44: class Program_weight_tensor_parameter_45: name = "parameter_45" + original_name = "linear_638.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.459519") @@ -506,6 +552,7 @@ class Program_weight_tensor_parameter_45: class Program_weight_tensor_parameter_46: name = "parameter_46" + original_name = "linear_637.b_0" shape = [768] dtype = "float32" min_val = float("-16.4834") @@ -517,6 +564,7 @@ class Program_weight_tensor_parameter_46: class Program_weight_tensor_parameter_47: name = "parameter_47" + original_name = "linear_637.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.455678") @@ -528,6 +576,7 @@ class Program_weight_tensor_parameter_47: class Program_weight_tensor_parameter_48: name = "parameter_48" + original_name = "linear_636.b_0" shape = [768] dtype = "float32" min_val = float("-3.05984") @@ -539,6 +588,7 @@ class Program_weight_tensor_parameter_48: class Program_weight_tensor_parameter_49: name = "parameter_49" + original_name = "linear_636.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.417866") @@ -550,6 +600,7 @@ class Program_weight_tensor_parameter_49: class Program_weight_tensor_parameter_50: name = "parameter_50" + original_name = "layer_norm_216.b_0" shape = [768] dtype = "float32" min_val = float("-0.693036") @@ -561,6 +612,7 @@ class Program_weight_tensor_parameter_50: class Program_weight_tensor_parameter_51: name = "parameter_51" + original_name = "layer_norm_216.w_0" shape = [768] dtype = "float32" min_val = float("0.115743") @@ -572,6 +624,7 @@ class Program_weight_tensor_parameter_51: class Program_weight_tensor_parameter_52: name = "parameter_52" + original_name = "layer_norm_215.b_0" shape = [768] dtype = "float32" min_val = float("-8.19627") @@ -583,6 +636,7 @@ class Program_weight_tensor_parameter_52: class Program_weight_tensor_parameter_53: name = "parameter_53" + original_name = "layer_norm_215.w_0" shape = [768] dtype = "float32" min_val = float("0.477878") @@ -594,6 +648,7 @@ class Program_weight_tensor_parameter_53: class Program_weight_tensor_parameter_54: name = "parameter_54" + original_name = "linear_635.b_0" shape = [768] dtype = "float32" min_val = float("-0.707393") @@ -605,6 +660,7 @@ class Program_weight_tensor_parameter_54: class Program_weight_tensor_parameter_55: name = "parameter_55" + original_name = "linear_635.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-1.92492") @@ -616,6 +672,7 @@ class Program_weight_tensor_parameter_55: class Program_weight_tensor_parameter_56: name = "parameter_56" + original_name = "linear_634.b_0" shape = [3072] dtype = "float32" min_val = float("-1.41527") @@ -627,6 +684,7 @@ class Program_weight_tensor_parameter_56: class Program_weight_tensor_parameter_57: name = "parameter_57" + original_name = "linear_634.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.552465") @@ -638,6 +696,7 @@ class Program_weight_tensor_parameter_57: class Program_weight_tensor_parameter_58: name = "parameter_58" + original_name = "linear_633.b_0" shape = [768] dtype = "float32" min_val = float("-0.158871") @@ -649,6 +708,7 @@ class Program_weight_tensor_parameter_58: class Program_weight_tensor_parameter_59: name = "parameter_59" + original_name = "linear_633.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.369395") @@ -660,6 +720,7 @@ class Program_weight_tensor_parameter_59: class Program_weight_tensor_parameter_60: name = "parameter_60" + original_name = "linear_632.b_0" shape = [768] dtype = "float32" min_val = float("-0.523862") @@ -671,6 +732,7 @@ class Program_weight_tensor_parameter_60: class Program_weight_tensor_parameter_61: name = "parameter_61" + original_name = "linear_632.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.315869") @@ -682,6 +744,7 @@ class Program_weight_tensor_parameter_61: class Program_weight_tensor_parameter_62: name = "parameter_62" + original_name = "linear_631.b_0" shape = [768] dtype = "float32" min_val = float("-13.719") @@ -693,6 +756,7 @@ class Program_weight_tensor_parameter_62: class Program_weight_tensor_parameter_63: name = "parameter_63" + original_name = "linear_631.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.503088") @@ -704,6 +768,7 @@ class Program_weight_tensor_parameter_63: class Program_weight_tensor_parameter_64: name = "parameter_64" + original_name = "linear_630.b_0" shape = [768] dtype = "float32" min_val = float("-3.20893") @@ -715,6 +780,7 @@ class Program_weight_tensor_parameter_64: class Program_weight_tensor_parameter_65: name = "parameter_65" + original_name = "linear_630.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.395917") @@ -726,6 +792,7 @@ class Program_weight_tensor_parameter_65: class Program_weight_tensor_parameter_66: name = "parameter_66" + original_name = "layer_norm_214.b_0" shape = [768] dtype = "float32" min_val = float("-0.612194") @@ -737,6 +804,7 @@ class Program_weight_tensor_parameter_66: class Program_weight_tensor_parameter_67: name = "parameter_67" + original_name = "layer_norm_214.w_0" shape = [768] dtype = "float32" min_val = float("0.148656") @@ -748,6 +816,7 @@ class Program_weight_tensor_parameter_67: class Program_weight_tensor_parameter_68: name = "parameter_68" + original_name = "layer_norm_213.b_0" shape = [768] dtype = "float32" min_val = float("-7.50134") @@ -759,6 +828,7 @@ class Program_weight_tensor_parameter_68: class Program_weight_tensor_parameter_69: name = "parameter_69" + original_name = "layer_norm_213.w_0" shape = [768] dtype = "float32" min_val = float("0.333985") @@ -770,6 +840,7 @@ class Program_weight_tensor_parameter_69: class Program_weight_tensor_parameter_70: name = "parameter_70" + original_name = "linear_629.b_0" shape = [768] dtype = "float32" min_val = float("-0.675722") @@ -781,6 +852,7 @@ class Program_weight_tensor_parameter_70: class Program_weight_tensor_parameter_71: name = "parameter_71" + original_name = "linear_629.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-1.56915") @@ -792,6 +864,7 @@ class Program_weight_tensor_parameter_71: class Program_weight_tensor_parameter_72: name = "parameter_72" + original_name = "linear_628.b_0" shape = [3072] dtype = "float32" min_val = float("-1.51892") @@ -803,6 +876,7 @@ class Program_weight_tensor_parameter_72: class Program_weight_tensor_parameter_73: name = "parameter_73" + original_name = "linear_628.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.760946") @@ -814,6 +888,7 @@ class Program_weight_tensor_parameter_73: class Program_weight_tensor_parameter_74: name = "parameter_74" + original_name = "linear_627.b_0" shape = [768] dtype = "float32" min_val = float("-0.249452") @@ -825,6 +900,7 @@ class Program_weight_tensor_parameter_74: class Program_weight_tensor_parameter_75: name = "parameter_75" + original_name = "linear_627.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.385008") @@ -836,6 +912,7 @@ class Program_weight_tensor_parameter_75: class Program_weight_tensor_parameter_76: name = "parameter_76" + original_name = "linear_626.b_0" shape = [768] dtype = "float32" min_val = float("-0.666143") @@ -847,6 +924,7 @@ class Program_weight_tensor_parameter_76: class Program_weight_tensor_parameter_77: name = "parameter_77" + original_name = "linear_626.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.28779") @@ -858,6 +936,7 @@ class Program_weight_tensor_parameter_77: class Program_weight_tensor_parameter_78: name = "parameter_78" + original_name = "linear_625.b_0" shape = [768] dtype = "float32" min_val = float("-8.67944") @@ -869,6 +948,7 @@ class Program_weight_tensor_parameter_78: class Program_weight_tensor_parameter_79: name = "parameter_79" + original_name = "linear_625.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.552326") @@ -880,6 +960,7 @@ class Program_weight_tensor_parameter_79: class Program_weight_tensor_parameter_80: name = "parameter_80" + original_name = "linear_624.b_0" shape = [768] dtype = "float32" min_val = float("-2.53418") @@ -891,6 +972,7 @@ class Program_weight_tensor_parameter_80: class Program_weight_tensor_parameter_81: name = "parameter_81" + original_name = "linear_624.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.494573") @@ -902,6 +984,7 @@ class Program_weight_tensor_parameter_81: class Program_weight_tensor_parameter_82: name = "parameter_82" + original_name = "layer_norm_212.b_0" shape = [768] dtype = "float32" min_val = float("-0.894087") @@ -913,6 +996,7 @@ class Program_weight_tensor_parameter_82: class Program_weight_tensor_parameter_83: name = "parameter_83" + original_name = "layer_norm_212.w_0" shape = [768] dtype = "float32" min_val = float("0.118243") @@ -924,6 +1008,7 @@ class Program_weight_tensor_parameter_83: class Program_weight_tensor_parameter_84: name = "parameter_84" + original_name = "layer_norm_211.b_0" shape = [768] dtype = "float32" min_val = float("-8.49471") @@ -935,6 +1020,7 @@ class Program_weight_tensor_parameter_84: class Program_weight_tensor_parameter_85: name = "parameter_85" + original_name = "layer_norm_211.w_0" shape = [768] dtype = "float32" min_val = float("0.3277") @@ -946,6 +1032,7 @@ class Program_weight_tensor_parameter_85: class Program_weight_tensor_parameter_86: name = "parameter_86" + original_name = "linear_623.b_0" shape = [768] dtype = "float32" min_val = float("-0.498074") @@ -957,6 +1044,7 @@ class Program_weight_tensor_parameter_86: class Program_weight_tensor_parameter_87: name = "parameter_87" + original_name = "linear_623.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-1.64563") @@ -968,6 +1056,7 @@ class Program_weight_tensor_parameter_87: class Program_weight_tensor_parameter_88: name = "parameter_88" + original_name = "linear_622.b_0" shape = [3072] dtype = "float32" min_val = float("-1.21876") @@ -979,6 +1068,7 @@ class Program_weight_tensor_parameter_88: class Program_weight_tensor_parameter_89: name = "parameter_89" + original_name = "linear_622.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.602838") @@ -990,6 +1080,7 @@ class Program_weight_tensor_parameter_89: class Program_weight_tensor_parameter_90: name = "parameter_90" + original_name = "linear_621.b_0" shape = [768] dtype = "float32" min_val = float("-0.364125") @@ -1001,6 +1092,7 @@ class Program_weight_tensor_parameter_90: class Program_weight_tensor_parameter_91: name = "parameter_91" + original_name = "linear_621.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.483536") @@ -1012,6 +1104,7 @@ class Program_weight_tensor_parameter_91: class Program_weight_tensor_parameter_92: name = "parameter_92" + original_name = "linear_620.b_0" shape = [768] dtype = "float32" min_val = float("-0.456832") @@ -1023,6 +1116,7 @@ class Program_weight_tensor_parameter_92: class Program_weight_tensor_parameter_93: name = "parameter_93" + original_name = "linear_620.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.334534") @@ -1034,6 +1128,7 @@ class Program_weight_tensor_parameter_93: class Program_weight_tensor_parameter_94: name = "parameter_94" + original_name = "linear_619.b_0" shape = [768] dtype = "float32" min_val = float("-5.33118") @@ -1045,6 +1140,7 @@ class Program_weight_tensor_parameter_94: class Program_weight_tensor_parameter_95: name = "parameter_95" + original_name = "linear_619.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.565856") @@ -1056,6 +1152,7 @@ class Program_weight_tensor_parameter_95: class Program_weight_tensor_parameter_96: name = "parameter_96" + original_name = "linear_618.b_0" shape = [768] dtype = "float32" min_val = float("-3.20808") @@ -1067,6 +1164,7 @@ class Program_weight_tensor_parameter_96: class Program_weight_tensor_parameter_97: name = "parameter_97" + original_name = "linear_618.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.619657") @@ -1078,6 +1176,7 @@ class Program_weight_tensor_parameter_97: class Program_weight_tensor_parameter_98: name = "parameter_98" + original_name = "layer_norm_210.b_0" shape = [768] dtype = "float32" min_val = float("-1.16209") @@ -1089,6 +1188,7 @@ class Program_weight_tensor_parameter_98: class Program_weight_tensor_parameter_99: name = "parameter_99" + original_name = "layer_norm_210.w_0" shape = [768] dtype = "float32" min_val = float("0.105108") @@ -1100,6 +1200,7 @@ class Program_weight_tensor_parameter_99: class Program_weight_tensor_parameter_100: name = "parameter_100" + original_name = "layer_norm_209.b_0" shape = [768] dtype = "float32" min_val = float("-9.2912") @@ -1111,6 +1212,7 @@ class Program_weight_tensor_parameter_100: class Program_weight_tensor_parameter_101: name = "parameter_101" + original_name = "layer_norm_209.w_0" shape = [768] dtype = "float32" min_val = float("0.0317146") @@ -1122,6 +1224,7 @@ class Program_weight_tensor_parameter_101: class Program_weight_tensor_parameter_102: name = "parameter_102" + original_name = "linear_617.b_0" shape = [768] dtype = "float32" min_val = float("-0.40929") @@ -1133,6 +1236,7 @@ class Program_weight_tensor_parameter_102: class Program_weight_tensor_parameter_103: name = "parameter_103" + original_name = "linear_617.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-1.45875") @@ -1144,6 +1248,7 @@ class Program_weight_tensor_parameter_103: class Program_weight_tensor_parameter_104: name = "parameter_104" + original_name = "linear_616.b_0" shape = [3072] dtype = "float32" min_val = float("-1.21348") @@ -1155,6 +1260,7 @@ class Program_weight_tensor_parameter_104: class Program_weight_tensor_parameter_105: name = "parameter_105" + original_name = "linear_616.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-1.1368") @@ -1166,6 +1272,7 @@ class Program_weight_tensor_parameter_105: class Program_weight_tensor_parameter_106: name = "parameter_106" + original_name = "linear_615.b_0" shape = [768] dtype = "float32" min_val = float("-0.227283") @@ -1177,6 +1284,7 @@ class Program_weight_tensor_parameter_106: class Program_weight_tensor_parameter_107: name = "parameter_107" + original_name = "linear_615.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.803245") @@ -1188,6 +1296,7 @@ class Program_weight_tensor_parameter_107: class Program_weight_tensor_parameter_108: name = "parameter_108" + original_name = "linear_614.b_0" shape = [768] dtype = "float32" min_val = float("-0.504589") @@ -1199,6 +1308,7 @@ class Program_weight_tensor_parameter_108: class Program_weight_tensor_parameter_109: name = "parameter_109" + original_name = "linear_614.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.577865") @@ -1210,6 +1320,7 @@ class Program_weight_tensor_parameter_109: class Program_weight_tensor_parameter_110: name = "parameter_110" + original_name = "linear_613.b_0" shape = [768] dtype = "float32" min_val = float("-3.32848") @@ -1221,6 +1332,7 @@ class Program_weight_tensor_parameter_110: class Program_weight_tensor_parameter_111: name = "parameter_111" + original_name = "linear_613.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.478512") @@ -1232,6 +1344,7 @@ class Program_weight_tensor_parameter_111: class Program_weight_tensor_parameter_112: name = "parameter_112" + original_name = "linear_612.b_0" shape = [768] dtype = "float32" min_val = float("-2.6096") @@ -1243,6 +1356,7 @@ class Program_weight_tensor_parameter_112: class Program_weight_tensor_parameter_113: name = "parameter_113" + original_name = "linear_612.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.497878") @@ -1254,6 +1368,7 @@ class Program_weight_tensor_parameter_113: class Program_weight_tensor_parameter_114: name = "parameter_114" + original_name = "layer_norm_208.b_0" shape = [768] dtype = "float32" min_val = float("-1.30834") @@ -1265,6 +1380,7 @@ class Program_weight_tensor_parameter_114: class Program_weight_tensor_parameter_115: name = "parameter_115" + original_name = "layer_norm_208.w_0" shape = [768] dtype = "float32" min_val = float("0.0966392") @@ -1276,6 +1392,7 @@ class Program_weight_tensor_parameter_115: class Program_weight_tensor_parameter_116: name = "parameter_116" + original_name = "layer_norm_207.b_0" shape = [768] dtype = "float32" min_val = float("-8.08704") @@ -1287,6 +1404,7 @@ class Program_weight_tensor_parameter_116: class Program_weight_tensor_parameter_117: name = "parameter_117" + original_name = "layer_norm_207.w_0" shape = [768] dtype = "float32" min_val = float("0.445565") @@ -1298,6 +1416,7 @@ class Program_weight_tensor_parameter_117: class Program_weight_tensor_parameter_118: name = "parameter_118" + original_name = "linear_611.b_0" shape = [768] dtype = "float32" min_val = float("-0.648921") @@ -1309,6 +1428,7 @@ class Program_weight_tensor_parameter_118: class Program_weight_tensor_parameter_119: name = "parameter_119" + original_name = "linear_611.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-2.18504") @@ -1320,6 +1440,7 @@ class Program_weight_tensor_parameter_119: class Program_weight_tensor_parameter_120: name = "parameter_120" + original_name = "linear_610.b_0" shape = [3072] dtype = "float32" min_val = float("-1.04488") @@ -1331,6 +1452,7 @@ class Program_weight_tensor_parameter_120: class Program_weight_tensor_parameter_121: name = "parameter_121" + original_name = "linear_610.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.87387") @@ -1342,6 +1464,7 @@ class Program_weight_tensor_parameter_121: class Program_weight_tensor_parameter_122: name = "parameter_122" + original_name = "linear_609.b_0" shape = [768] dtype = "float32" min_val = float("-0.406648") @@ -1353,6 +1476,7 @@ class Program_weight_tensor_parameter_122: class Program_weight_tensor_parameter_123: name = "parameter_123" + original_name = "linear_609.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.325325") @@ -1364,6 +1488,7 @@ class Program_weight_tensor_parameter_123: class Program_weight_tensor_parameter_124: name = "parameter_124" + original_name = "linear_608.b_0" shape = [768] dtype = "float32" min_val = float("-0.475245") @@ -1375,6 +1500,7 @@ class Program_weight_tensor_parameter_124: class Program_weight_tensor_parameter_125: name = "parameter_125" + original_name = "linear_608.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.608167") @@ -1386,6 +1512,7 @@ class Program_weight_tensor_parameter_125: class Program_weight_tensor_parameter_126: name = "parameter_126" + original_name = "linear_607.b_0" shape = [768] dtype = "float32" min_val = float("-2.95977") @@ -1397,6 +1524,7 @@ class Program_weight_tensor_parameter_126: class Program_weight_tensor_parameter_127: name = "parameter_127" + original_name = "linear_607.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.476853") @@ -1408,6 +1536,7 @@ class Program_weight_tensor_parameter_127: class Program_weight_tensor_parameter_128: name = "parameter_128" + original_name = "linear_606.b_0" shape = [768] dtype = "float32" min_val = float("-3.56482") @@ -1419,6 +1548,7 @@ class Program_weight_tensor_parameter_128: class Program_weight_tensor_parameter_129: name = "parameter_129" + original_name = "linear_606.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.436752") @@ -1430,6 +1560,7 @@ class Program_weight_tensor_parameter_129: class Program_weight_tensor_parameter_130: name = "parameter_130" + original_name = "layer_norm_206.b_0" shape = [768] dtype = "float32" min_val = float("-1.00721") @@ -1441,6 +1572,7 @@ class Program_weight_tensor_parameter_130: class Program_weight_tensor_parameter_131: name = "parameter_131" + original_name = "layer_norm_206.w_0" shape = [768] dtype = "float32" min_val = float("0.23472") @@ -1452,6 +1584,7 @@ class Program_weight_tensor_parameter_131: class Program_weight_tensor_parameter_132: name = "parameter_132" + original_name = "layer_norm_205.b_0" shape = [768] dtype = "float32" min_val = float("-7.10482") @@ -1463,6 +1596,7 @@ class Program_weight_tensor_parameter_132: class Program_weight_tensor_parameter_133: name = "parameter_133" + original_name = "layer_norm_205.w_0" shape = [768] dtype = "float32" min_val = float("0.538137") @@ -1474,6 +1608,7 @@ class Program_weight_tensor_parameter_133: class Program_weight_tensor_parameter_134: name = "parameter_134" + original_name = "linear_605.b_0" shape = [768] dtype = "float32" min_val = float("-0.539944") @@ -1485,6 +1620,7 @@ class Program_weight_tensor_parameter_134: class Program_weight_tensor_parameter_135: name = "parameter_135" + original_name = "linear_605.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-4.45508") @@ -1496,6 +1632,7 @@ class Program_weight_tensor_parameter_135: class Program_weight_tensor_parameter_136: name = "parameter_136" + original_name = "linear_604.b_0" shape = [3072] dtype = "float32" min_val = float("-1.31049") @@ -1507,6 +1644,7 @@ class Program_weight_tensor_parameter_136: class Program_weight_tensor_parameter_137: name = "parameter_137" + original_name = "linear_604.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.640963") @@ -1518,6 +1656,7 @@ class Program_weight_tensor_parameter_137: class Program_weight_tensor_parameter_138: name = "parameter_138" + original_name = "linear_603.b_0" shape = [768] dtype = "float32" min_val = float("-0.306928") @@ -1529,6 +1668,7 @@ class Program_weight_tensor_parameter_138: class Program_weight_tensor_parameter_139: name = "parameter_139" + original_name = "linear_603.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.69083") @@ -1540,6 +1680,7 @@ class Program_weight_tensor_parameter_139: class Program_weight_tensor_parameter_140: name = "parameter_140" + original_name = "linear_602.b_0" shape = [768] dtype = "float32" min_val = float("-0.48088") @@ -1551,6 +1692,7 @@ class Program_weight_tensor_parameter_140: class Program_weight_tensor_parameter_141: name = "parameter_141" + original_name = "linear_602.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.480303") @@ -1562,6 +1704,7 @@ class Program_weight_tensor_parameter_141: class Program_weight_tensor_parameter_142: name = "parameter_142" + original_name = "linear_601.b_0" shape = [768] dtype = "float32" min_val = float("-1.91661") @@ -1573,6 +1716,7 @@ class Program_weight_tensor_parameter_142: class Program_weight_tensor_parameter_143: name = "parameter_143" + original_name = "linear_601.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.5189") @@ -1584,6 +1728,7 @@ class Program_weight_tensor_parameter_143: class Program_weight_tensor_parameter_144: name = "parameter_144" + original_name = "linear_600.b_0" shape = [768] dtype = "float32" min_val = float("-2.94715") @@ -1595,6 +1740,7 @@ class Program_weight_tensor_parameter_144: class Program_weight_tensor_parameter_145: name = "parameter_145" + original_name = "linear_600.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.586879") @@ -1606,6 +1752,7 @@ class Program_weight_tensor_parameter_145: class Program_weight_tensor_parameter_146: name = "parameter_146" + original_name = "layer_norm_204.b_0" shape = [768] dtype = "float32" min_val = float("-1.4412") @@ -1617,6 +1764,7 @@ class Program_weight_tensor_parameter_146: class Program_weight_tensor_parameter_147: name = "parameter_147" + original_name = "layer_norm_204.w_0" shape = [768] dtype = "float32" min_val = float("0.162569") @@ -1628,6 +1776,7 @@ class Program_weight_tensor_parameter_147: class Program_weight_tensor_parameter_148: name = "parameter_148" + original_name = "layer_norm_203.b_0" shape = [768] dtype = "float32" min_val = float("-6.17391") @@ -1639,6 +1788,7 @@ class Program_weight_tensor_parameter_148: class Program_weight_tensor_parameter_149: name = "parameter_149" + original_name = "layer_norm_203.w_0" shape = [768] dtype = "float32" min_val = float("0.671115") @@ -1650,6 +1800,7 @@ class Program_weight_tensor_parameter_149: class Program_weight_tensor_parameter_150: name = "parameter_150" + original_name = "linear_599.b_0" shape = [768] dtype = "float32" min_val = float("-0.621028") @@ -1661,6 +1812,7 @@ class Program_weight_tensor_parameter_150: class Program_weight_tensor_parameter_151: name = "parameter_151" + original_name = "linear_599.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-1.61366") @@ -1672,6 +1824,7 @@ class Program_weight_tensor_parameter_151: class Program_weight_tensor_parameter_152: name = "parameter_152" + original_name = "linear_598.b_0" shape = [3072] dtype = "float32" min_val = float("-0.704459") @@ -1683,6 +1836,7 @@ class Program_weight_tensor_parameter_152: class Program_weight_tensor_parameter_153: name = "parameter_153" + original_name = "linear_598.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.629047") @@ -1694,6 +1848,7 @@ class Program_weight_tensor_parameter_153: class Program_weight_tensor_parameter_154: name = "parameter_154" + original_name = "linear_597.b_0" shape = [768] dtype = "float32" min_val = float("-0.264184") @@ -1705,6 +1860,7 @@ class Program_weight_tensor_parameter_154: class Program_weight_tensor_parameter_155: name = "parameter_155" + original_name = "linear_597.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.604281") @@ -1716,6 +1872,7 @@ class Program_weight_tensor_parameter_155: class Program_weight_tensor_parameter_156: name = "parameter_156" + original_name = "linear_596.b_0" shape = [768] dtype = "float32" min_val = float("-0.310892") @@ -1727,6 +1884,7 @@ class Program_weight_tensor_parameter_156: class Program_weight_tensor_parameter_157: name = "parameter_157" + original_name = "linear_596.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.381732") @@ -1738,6 +1896,7 @@ class Program_weight_tensor_parameter_157: class Program_weight_tensor_parameter_158: name = "parameter_158" + original_name = "linear_595.b_0" shape = [768] dtype = "float32" min_val = float("-2.67565") @@ -1749,6 +1908,7 @@ class Program_weight_tensor_parameter_158: class Program_weight_tensor_parameter_159: name = "parameter_159" + original_name = "linear_595.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.486523") @@ -1760,6 +1920,7 @@ class Program_weight_tensor_parameter_159: class Program_weight_tensor_parameter_160: name = "parameter_160" + original_name = "linear_594.b_0" shape = [768] dtype = "float32" min_val = float("-3.28191") @@ -1771,6 +1932,7 @@ class Program_weight_tensor_parameter_160: class Program_weight_tensor_parameter_161: name = "parameter_161" + original_name = "linear_594.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.440847") @@ -1782,6 +1944,7 @@ class Program_weight_tensor_parameter_161: class Program_weight_tensor_parameter_162: name = "parameter_162" + original_name = "layer_norm_202.b_0" shape = [768] dtype = "float32" min_val = float("-0.92585") @@ -1793,6 +1956,7 @@ class Program_weight_tensor_parameter_162: class Program_weight_tensor_parameter_163: name = "parameter_163" + original_name = "layer_norm_202.w_0" shape = [768] dtype = "float32" min_val = float("0.133932") @@ -1804,6 +1968,7 @@ class Program_weight_tensor_parameter_163: class Program_weight_tensor_parameter_164: name = "parameter_164" + original_name = "layer_norm_201.b_0" shape = [768] dtype = "float32" min_val = float("-5.39594") @@ -1815,6 +1980,7 @@ class Program_weight_tensor_parameter_164: class Program_weight_tensor_parameter_165: name = "parameter_165" + original_name = "layer_norm_201.w_0" shape = [768] dtype = "float32" min_val = float("0.592799") @@ -1826,6 +1992,7 @@ class Program_weight_tensor_parameter_165: class Program_weight_tensor_parameter_166: name = "parameter_166" + original_name = "linear_593.b_0" shape = [768] dtype = "float32" min_val = float("-0.739173") @@ -1837,6 +2004,7 @@ class Program_weight_tensor_parameter_166: class Program_weight_tensor_parameter_167: name = "parameter_167" + original_name = "linear_593.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-3.81737") @@ -1848,6 +2016,7 @@ class Program_weight_tensor_parameter_167: class Program_weight_tensor_parameter_168: name = "parameter_168" + original_name = "linear_592.b_0" shape = [3072] dtype = "float32" min_val = float("-0.988765") @@ -1859,6 +2028,7 @@ class Program_weight_tensor_parameter_168: class Program_weight_tensor_parameter_169: name = "parameter_169" + original_name = "linear_592.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.58617") @@ -1870,6 +2040,7 @@ class Program_weight_tensor_parameter_169: class Program_weight_tensor_parameter_170: name = "parameter_170" + original_name = "linear_591.b_0" shape = [768] dtype = "float32" min_val = float("-0.246057") @@ -1881,6 +2052,7 @@ class Program_weight_tensor_parameter_170: class Program_weight_tensor_parameter_171: name = "parameter_171" + original_name = "linear_591.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.366085") @@ -1892,6 +2064,7 @@ class Program_weight_tensor_parameter_171: class Program_weight_tensor_parameter_172: name = "parameter_172" + original_name = "linear_590.b_0" shape = [768] dtype = "float32" min_val = float("-0.36641") @@ -1903,6 +2076,7 @@ class Program_weight_tensor_parameter_172: class Program_weight_tensor_parameter_173: name = "parameter_173" + original_name = "linear_590.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.425146") @@ -1914,6 +2088,7 @@ class Program_weight_tensor_parameter_173: class Program_weight_tensor_parameter_174: name = "parameter_174" + original_name = "linear_589.b_0" shape = [768] dtype = "float32" min_val = float("-3.24228") @@ -1925,6 +2100,7 @@ class Program_weight_tensor_parameter_174: class Program_weight_tensor_parameter_175: name = "parameter_175" + original_name = "linear_589.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.615107") @@ -1936,6 +2112,7 @@ class Program_weight_tensor_parameter_175: class Program_weight_tensor_parameter_176: name = "parameter_176" + original_name = "linear_588.b_0" shape = [768] dtype = "float32" min_val = float("-2.7673") @@ -1947,6 +2124,7 @@ class Program_weight_tensor_parameter_176: class Program_weight_tensor_parameter_177: name = "parameter_177" + original_name = "linear_588.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.399855") @@ -1958,6 +2136,7 @@ class Program_weight_tensor_parameter_177: class Program_weight_tensor_parameter_178: name = "parameter_178" + original_name = "layer_norm_200.b_0" shape = [768] dtype = "float32" min_val = float("-0.623707") @@ -1969,6 +2148,7 @@ class Program_weight_tensor_parameter_178: class Program_weight_tensor_parameter_179: name = "parameter_179" + original_name = "layer_norm_200.w_0" shape = [768] dtype = "float32" min_val = float("0.100484") @@ -1980,6 +2160,7 @@ class Program_weight_tensor_parameter_179: class Program_weight_tensor_parameter_180: name = "parameter_180" + original_name = "layer_norm_199.b_0" shape = [768] dtype = "float32" min_val = float("-7.94408") @@ -1991,6 +2172,7 @@ class Program_weight_tensor_parameter_180: class Program_weight_tensor_parameter_181: name = "parameter_181" + original_name = "layer_norm_199.w_0" shape = [768] dtype = "float32" min_val = float("0.347752") @@ -2002,6 +2184,7 @@ class Program_weight_tensor_parameter_181: class Program_weight_tensor_parameter_182: name = "parameter_182" + original_name = "linear_587.b_0" shape = [768] dtype = "float32" min_val = float("-0.612331") @@ -2013,6 +2196,7 @@ class Program_weight_tensor_parameter_182: class Program_weight_tensor_parameter_183: name = "parameter_183" + original_name = "linear_587.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-3.55145") @@ -2024,6 +2208,7 @@ class Program_weight_tensor_parameter_183: class Program_weight_tensor_parameter_184: name = "parameter_184" + original_name = "linear_586.b_0" shape = [3072] dtype = "float32" min_val = float("-1.50494") @@ -2035,6 +2220,7 @@ class Program_weight_tensor_parameter_184: class Program_weight_tensor_parameter_185: name = "parameter_185" + original_name = "linear_586.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.602464") @@ -2046,6 +2232,7 @@ class Program_weight_tensor_parameter_185: class Program_weight_tensor_parameter_186: name = "parameter_186" + original_name = "linear_585.b_0" shape = [768] dtype = "float32" min_val = float("-0.593266") @@ -2057,6 +2244,7 @@ class Program_weight_tensor_parameter_186: class Program_weight_tensor_parameter_187: name = "parameter_187" + original_name = "linear_585.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.387033") @@ -2068,6 +2256,7 @@ class Program_weight_tensor_parameter_187: class Program_weight_tensor_parameter_188: name = "parameter_188" + original_name = "linear_584.b_0" shape = [768] dtype = "float32" min_val = float("-1.11233") @@ -2079,6 +2268,7 @@ class Program_weight_tensor_parameter_188: class Program_weight_tensor_parameter_189: name = "parameter_189" + original_name = "linear_584.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.515961") @@ -2090,6 +2280,7 @@ class Program_weight_tensor_parameter_189: class Program_weight_tensor_parameter_190: name = "parameter_190" + original_name = "linear_583.b_0" shape = [768] dtype = "float32" min_val = float("-1.64709") @@ -2101,6 +2292,7 @@ class Program_weight_tensor_parameter_190: class Program_weight_tensor_parameter_191: name = "parameter_191" + original_name = "linear_583.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.452388") @@ -2112,6 +2304,7 @@ class Program_weight_tensor_parameter_191: class Program_weight_tensor_parameter_192: name = "parameter_192" + original_name = "linear_582.b_0" shape = [768] dtype = "float32" min_val = float("-3.03019") @@ -2123,6 +2316,7 @@ class Program_weight_tensor_parameter_192: class Program_weight_tensor_parameter_193: name = "parameter_193" + original_name = "linear_582.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.335856") @@ -2134,6 +2328,7 @@ class Program_weight_tensor_parameter_193: class Program_weight_tensor_parameter_194: name = "parameter_194" + original_name = "layer_norm_198.b_0" shape = [768] dtype = "float32" min_val = float("-0.559598") @@ -2145,6 +2340,7 @@ class Program_weight_tensor_parameter_194: class Program_weight_tensor_parameter_195: name = "parameter_195" + original_name = "layer_norm_198.w_0" shape = [768] dtype = "float32" min_val = float("0.133497") @@ -2156,6 +2352,7 @@ class Program_weight_tensor_parameter_195: class Program_weight_tensor_parameter_196: name = "parameter_196" + original_name = "embedding_22.w_0" shape = [3, 768] dtype = "float32" min_val = float("-0.195099") @@ -2167,6 +2364,7 @@ class Program_weight_tensor_parameter_196: class Program_weight_tensor_parameter_197: name = "parameter_197" + original_name = "embedding_21.w_0" shape = [4, 768] dtype = "float32" min_val = float("-0.340029") @@ -2178,6 +2376,7 @@ class Program_weight_tensor_parameter_197: class Program_weight_tensor_parameter_198: name = "parameter_198" + original_name = "embedding_20.w_0" shape = [2048, 768] dtype = "float32" min_val = float("-0.208018") @@ -2189,6 +2388,7 @@ class Program_weight_tensor_parameter_198: class Program_weight_tensor_parameter_199: name = "parameter_199" + original_name = "embedding_19.w_0" shape = [40000, 768] dtype = "float32" min_val = float("-0.869171") diff --git a/paddle_samples/PaddleNLP/ernie-3.0-medium-zh/input_meta.py b/paddle_samples/PaddleNLP/ernie-3.0-medium-zh/input_meta.py index 97eb8a799..a4123e481 100644 --- a/paddle_samples/PaddleNLP/ernie-3.0-medium-zh/input_meta.py +++ b/paddle_samples/PaddleNLP/ernie-3.0-medium-zh/input_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_data_0: name = "data_0" + original_name = "input_ids" shape = [1, 11] dtype = "int64" data = [1, 811, 1257, 175, 29, 502, 130, 706, 3619, 12046, 2] @@ -7,6 +8,7 @@ class Program_weight_tensor_data_0: class Program_weight_tensor_data_1: name = "data_1" + original_name = "token_type_ids" shape = [1, 11] dtype = "int64" data = [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0] diff --git a/paddle_samples/PaddleNLP/ernie-3.0-medium-zh/model.py b/paddle_samples/PaddleNLP/ernie-3.0-medium-zh/model.py index edcf4dfee..2cc06c2be 100644 --- a/paddle_samples/PaddleNLP/ernie-3.0-medium-zh/model.py +++ b/paddle_samples/PaddleNLP/ernie-3.0-medium-zh/model.py @@ -170,10 +170,11 @@ def forward( # pd_op.embedding: (1x11x768xf32) <- (1x11xi64, 2048x768xf32) embedding_1 = paddle._C_ops.embedding(subtract_0, parameter_102, -1, False) - del parameter_102 + del parameter_102, subtract_0 # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 1x11x768xf32) add_0 = paddle._C_ops.add(embedding_0, embedding_1) + del embedding_0, embedding_1 # pd_op.embedding: (1x11x768xf32) <- (1x11xi64, 4x768xf32) embedding_2 = paddle._C_ops.embedding(data_1, parameter_101, -1, False) @@ -181,6 +182,7 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 1x11x768xf32) add_1 = paddle._C_ops.add(add_0, embedding_2) + del add_0, embedding_2 # pd_op.full: (1xf32) <- () full_4 = paddle._C_ops.full( @@ -193,10 +195,11 @@ def forward( # pd_op.embedding: (1x11x768xf32) <- (1x11xi64, 16x768xf32) embedding_3 = paddle._C_ops.embedding(scale_1, parameter_100, -1, False) - del parameter_100 + del parameter_100, scale_1 # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 1x11x768xf32) add_2 = paddle._C_ops.add(add_1, embedding_3) + del add_1, embedding_3 # pd_op.layer_norm: (1x11x768xf32, 1x11xf32, 1x11xf32) <- (1x11x768xf32, 768xf32, 768xf32) layer_norm_0, layer_norm_1, layer_norm_2 = (lambda x, f: f(x))( @@ -205,71 +208,17 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_98, parameter_99 + del add_2, parameter_98, parameter_99 # pd_op.full: (1xf32) <- () full_5 = paddle._C_ops.full( [1], float("0.1"), paddle.float32, paddle.core.CPUPlace() ) - # pd_op.assign: (1xf32) <- (1xf32) - assign_0 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_1 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_2 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_3 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_4 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_5 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_6 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_7 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_8 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_9 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_10 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_11 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_12 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_13 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_14 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_15 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_16 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_17 = full_5 - # pd_op.dropout: (1x11x768xf32, 1x11x768xui8) <- (1x11x768xf32, None, 1xf32) dropout_0, dropout_1 = (lambda x, f: f(x))( paddle._C_ops.dropout( - layer_norm_0, None, full_5, False, "upscale_in_train", 0, False + layer_norm_0, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -281,13 +230,14 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_3 = paddle._C_ops.add(matmul_0, parameter_96) - del parameter_96 + del matmul_0, parameter_96 # pd_op.full_int_array: (4xi64) <- () full_int_array_1 = [0, 0, 12, 64] # pd_op.reshape: (1x11x12x64xf32) <- (1x11x768xf32, 4xi64) reshape_0 = paddle._C_ops.reshape(add_3, full_int_array_1) + del add_3 # pd_op.transpose: (1x12x11x64xf32) <- (1x11x12x64xf32) transpose_0 = paddle._C_ops.transpose(reshape_0, [0, 2, 1, 3]) @@ -299,7 +249,7 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_4 = paddle._C_ops.add(matmul_1, parameter_94) - del parameter_94 + del matmul_1, parameter_94 # pd_op.matmul: (1x11x768xf32) <- (1x11x768xf32, 768x768xf32) matmul_2 = paddle._C_ops.matmul(dropout_0, parameter_93, False, False) @@ -307,10 +257,11 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_5 = paddle._C_ops.add(matmul_2, parameter_92) - del parameter_92 + del matmul_2, parameter_92 # pd_op.reshape: (1x11x12x64xf32) <- (1x11x768xf32, 4xi64) reshape_1 = paddle._C_ops.reshape(add_4, full_int_array_1) + del add_4 # pd_op.transpose: (1x12x11x64xf32) <- (1x11x12x64xf32) transpose_1 = paddle._C_ops.transpose(reshape_1, [0, 2, 1, 3]) @@ -318,6 +269,7 @@ def forward( # pd_op.reshape: (1x11x12x64xf32) <- (1x11x768xf32, 4xi64) reshape_2 = paddle._C_ops.reshape(add_5, full_int_array_1) + del add_5 # pd_op.transpose: (1x12x11x64xf32) <- (1x11x12x64xf32) transpose_2 = paddle._C_ops.transpose(reshape_2, [0, 2, 1, 3]) @@ -328,30 +280,17 @@ def forward( [1], float("0.125"), paddle.float32, paddle.core.CPUPlace() ) - # pd_op.assign: (1xf32) <- (1xf32) - assign_18 = full_6 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_19 = full_6 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_20 = full_6 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_21 = full_6 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_22 = full_6 - # pd_op.scale: (1x12x11x64xf32) <- (1x12x11x64xf32, 1xf32) scale_2 = paddle._C_ops.scale(transpose_0, full_6, float("0"), True) del transpose_0 # pd_op.matmul: (1x12x11x11xf32) <- (1x12x11x64xf32, 1x12x11x64xf32) matmul_3 = paddle._C_ops.matmul(scale_2, transpose_1, False, True) + del scale_2, transpose_1 # pd_op.add: (1x12x11x11xf32) <- (1x12x11x11xf32, 1x1x1x11xf32) add_6 = paddle._C_ops.add(matmul_3, unsqueeze_0) + del matmul_3 # pd_op.softmax: (1x12x11x11xf32) <- (1x12x11x11xf32) softmax_0 = paddle._C_ops.softmax(add_6, -1) @@ -360,13 +299,15 @@ def forward( # pd_op.dropout: (1x12x11x11xf32, 1x12x11x11xui8) <- (1x12x11x11xf32, None, 1xf32) dropout_2, dropout_3 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_0, None, full_5, False, "upscale_in_train", 0, False + softmax_0, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_0 # pd_op.matmul: (1x12x11x64xf32) <- (1x12x11x11xf32, 1x12x11x64xf32) matmul_4 = paddle._C_ops.matmul(dropout_2, transpose_2, False, False) + del dropout_2, transpose_2 # pd_op.transpose: (1x11x12x64xf32) <- (1x12x11x64xf32) transpose_3 = paddle._C_ops.transpose(matmul_4, [0, 2, 1, 3]) @@ -377,19 +318,20 @@ def forward( # pd_op.reshape: (1x11x768xf32) <- (1x11x12x64xf32, 3xi64) reshape_3 = paddle._C_ops.reshape(transpose_3, full_int_array_2) + del transpose_3 # pd_op.matmul: (1x11x768xf32) <- (1x11x768xf32, 768x768xf32) matmul_5 = paddle._C_ops.matmul(reshape_3, parameter_91, False, False) - del parameter_91 + del parameter_91, reshape_3 # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_7 = paddle._C_ops.add(matmul_5, parameter_90) - del parameter_90 + del matmul_5, parameter_90 # pd_op.dropout: (1x11x768xf32, 1x11x768xui8) <- (1x11x768xf32, None, 1xf32) dropout_4, dropout_5 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_7, None, full_5, False, "upscale_in_train", 0, False + add_7, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -397,6 +339,7 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 1x11x768xf32) add_8 = paddle._C_ops.add(dropout_0, dropout_4) + del dropout_0, dropout_4 # pd_op.layer_norm: (1x11x768xf32, 1x11xf32, 1x11xf32) <- (1x11x768xf32, 768xf32, 768xf32) layer_norm_3, layer_norm_4, layer_norm_5 = (lambda x, f: f(x))( @@ -405,7 +348,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_84, parameter_85 + del add_8, parameter_84, parameter_85 # pd_op.matmul: (1x11x3072xf32) <- (1x11x768xf32, 768x3072xf32) matmul_6 = paddle._C_ops.matmul(layer_norm_3, parameter_89, False, False) @@ -413,23 +356,24 @@ def forward( # pd_op.add: (1x11x3072xf32) <- (1x11x3072xf32, 3072xf32) add_9 = paddle._C_ops.add(matmul_6, parameter_88) - del parameter_88 + del matmul_6, parameter_88 # pd_op.gelu: (1x11x3072xf32) <- (1x11x3072xf32) gelu_0 = paddle._C_ops.gelu(add_9, False) + del add_9 # pd_op.matmul: (1x11x768xf32) <- (1x11x3072xf32, 3072x768xf32) matmul_7 = paddle._C_ops.matmul(gelu_0, parameter_87, False, False) - del parameter_87 + del gelu_0, parameter_87 # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_10 = paddle._C_ops.add(matmul_7, parameter_86) - del parameter_86 + del matmul_7, parameter_86 # pd_op.dropout: (1x11x768xf32, 1x11x768xui8) <- (1x11x768xf32, None, 1xf32) dropout_6, dropout_7 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_10, None, full_5, False, "upscale_in_train", 0, False + add_10, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -437,6 +381,7 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 1x11x768xf32) add_11 = paddle._C_ops.add(layer_norm_3, dropout_6) + del dropout_6, layer_norm_3 # pd_op.layer_norm: (1x11x768xf32, 1x11xf32, 1x11xf32) <- (1x11x768xf32, 768xf32, 768xf32) layer_norm_6, layer_norm_7, layer_norm_8 = (lambda x, f: f(x))( @@ -445,7 +390,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_82, parameter_83 + del add_11, parameter_82, parameter_83 # pd_op.matmul: (1x11x768xf32) <- (1x11x768xf32, 768x768xf32) matmul_8 = paddle._C_ops.matmul(layer_norm_6, parameter_81, False, False) @@ -453,10 +398,11 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_12 = paddle._C_ops.add(matmul_8, parameter_80) - del parameter_80 + del matmul_8, parameter_80 # pd_op.reshape: (1x11x12x64xf32) <- (1x11x768xf32, 4xi64) reshape_4 = paddle._C_ops.reshape(add_12, full_int_array_1) + del add_12 # pd_op.transpose: (1x12x11x64xf32) <- (1x11x12x64xf32) transpose_4 = paddle._C_ops.transpose(reshape_4, [0, 2, 1, 3]) @@ -468,7 +414,7 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_13 = paddle._C_ops.add(matmul_9, parameter_78) - del parameter_78 + del matmul_9, parameter_78 # pd_op.matmul: (1x11x768xf32) <- (1x11x768xf32, 768x768xf32) matmul_10 = paddle._C_ops.matmul(layer_norm_6, parameter_77, False, False) @@ -476,10 +422,11 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_14 = paddle._C_ops.add(matmul_10, parameter_76) - del parameter_76 + del matmul_10, parameter_76 # pd_op.reshape: (1x11x12x64xf32) <- (1x11x768xf32, 4xi64) reshape_5 = paddle._C_ops.reshape(add_13, full_int_array_1) + del add_13 # pd_op.transpose: (1x12x11x64xf32) <- (1x11x12x64xf32) transpose_5 = paddle._C_ops.transpose(reshape_5, [0, 2, 1, 3]) @@ -487,6 +434,7 @@ def forward( # pd_op.reshape: (1x11x12x64xf32) <- (1x11x768xf32, 4xi64) reshape_6 = paddle._C_ops.reshape(add_14, full_int_array_1) + del add_14 # pd_op.transpose: (1x12x11x64xf32) <- (1x11x12x64xf32) transpose_6 = paddle._C_ops.transpose(reshape_6, [0, 2, 1, 3]) @@ -498,9 +446,11 @@ def forward( # pd_op.matmul: (1x12x11x11xf32) <- (1x12x11x64xf32, 1x12x11x64xf32) matmul_11 = paddle._C_ops.matmul(scale_3, transpose_5, False, True) + del scale_3, transpose_5 # pd_op.add: (1x12x11x11xf32) <- (1x12x11x11xf32, 1x1x1x11xf32) add_15 = paddle._C_ops.add(matmul_11, unsqueeze_0) + del matmul_11 # pd_op.softmax: (1x12x11x11xf32) <- (1x12x11x11xf32) softmax_1 = paddle._C_ops.softmax(add_15, -1) @@ -509,13 +459,15 @@ def forward( # pd_op.dropout: (1x12x11x11xf32, 1x12x11x11xui8) <- (1x12x11x11xf32, None, 1xf32) dropout_8, dropout_9 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_1, None, full_5, False, "upscale_in_train", 0, False + softmax_1, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_1 # pd_op.matmul: (1x12x11x64xf32) <- (1x12x11x11xf32, 1x12x11x64xf32) matmul_12 = paddle._C_ops.matmul(dropout_8, transpose_6, False, False) + del dropout_8, transpose_6 # pd_op.transpose: (1x11x12x64xf32) <- (1x12x11x64xf32) transpose_7 = paddle._C_ops.transpose(matmul_12, [0, 2, 1, 3]) @@ -523,19 +475,20 @@ def forward( # pd_op.reshape: (1x11x768xf32) <- (1x11x12x64xf32, 3xi64) reshape_7 = paddle._C_ops.reshape(transpose_7, full_int_array_2) + del transpose_7 # pd_op.matmul: (1x11x768xf32) <- (1x11x768xf32, 768x768xf32) matmul_13 = paddle._C_ops.matmul(reshape_7, parameter_75, False, False) - del parameter_75 + del parameter_75, reshape_7 # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_16 = paddle._C_ops.add(matmul_13, parameter_74) - del parameter_74 + del matmul_13, parameter_74 # pd_op.dropout: (1x11x768xf32, 1x11x768xui8) <- (1x11x768xf32, None, 1xf32) dropout_10, dropout_11 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_16, None, full_5, False, "upscale_in_train", 0, False + add_16, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -543,6 +496,7 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 1x11x768xf32) add_17 = paddle._C_ops.add(layer_norm_6, dropout_10) + del dropout_10, layer_norm_6 # pd_op.layer_norm: (1x11x768xf32, 1x11xf32, 1x11xf32) <- (1x11x768xf32, 768xf32, 768xf32) layer_norm_9, layer_norm_10, layer_norm_11 = (lambda x, f: f(x))( @@ -551,7 +505,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_68, parameter_69 + del add_17, parameter_68, parameter_69 # pd_op.matmul: (1x11x3072xf32) <- (1x11x768xf32, 768x3072xf32) matmul_14 = paddle._C_ops.matmul(layer_norm_9, parameter_73, False, False) @@ -559,23 +513,24 @@ def forward( # pd_op.add: (1x11x3072xf32) <- (1x11x3072xf32, 3072xf32) add_18 = paddle._C_ops.add(matmul_14, parameter_72) - del parameter_72 + del matmul_14, parameter_72 # pd_op.gelu: (1x11x3072xf32) <- (1x11x3072xf32) gelu_1 = paddle._C_ops.gelu(add_18, False) + del add_18 # pd_op.matmul: (1x11x768xf32) <- (1x11x3072xf32, 3072x768xf32) matmul_15 = paddle._C_ops.matmul(gelu_1, parameter_71, False, False) - del parameter_71 + del gelu_1, parameter_71 # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_19 = paddle._C_ops.add(matmul_15, parameter_70) - del parameter_70 + del matmul_15, parameter_70 # pd_op.dropout: (1x11x768xf32, 1x11x768xui8) <- (1x11x768xf32, None, 1xf32) dropout_12, dropout_13 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_19, None, full_5, False, "upscale_in_train", 0, False + add_19, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -583,6 +538,7 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 1x11x768xf32) add_20 = paddle._C_ops.add(layer_norm_9, dropout_12) + del dropout_12, layer_norm_9 # pd_op.layer_norm: (1x11x768xf32, 1x11xf32, 1x11xf32) <- (1x11x768xf32, 768xf32, 768xf32) layer_norm_12, layer_norm_13, layer_norm_14 = (lambda x, f: f(x))( @@ -591,7 +547,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_66, parameter_67 + del add_20, parameter_66, parameter_67 # pd_op.matmul: (1x11x768xf32) <- (1x11x768xf32, 768x768xf32) matmul_16 = paddle._C_ops.matmul(layer_norm_12, parameter_65, False, False) @@ -599,10 +555,11 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_21 = paddle._C_ops.add(matmul_16, parameter_64) - del parameter_64 + del matmul_16, parameter_64 # pd_op.reshape: (1x11x12x64xf32) <- (1x11x768xf32, 4xi64) reshape_8 = paddle._C_ops.reshape(add_21, full_int_array_1) + del add_21 # pd_op.transpose: (1x12x11x64xf32) <- (1x11x12x64xf32) transpose_8 = paddle._C_ops.transpose(reshape_8, [0, 2, 1, 3]) @@ -614,7 +571,7 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_22 = paddle._C_ops.add(matmul_17, parameter_62) - del parameter_62 + del matmul_17, parameter_62 # pd_op.matmul: (1x11x768xf32) <- (1x11x768xf32, 768x768xf32) matmul_18 = paddle._C_ops.matmul(layer_norm_12, parameter_61, False, False) @@ -622,10 +579,11 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_23 = paddle._C_ops.add(matmul_18, parameter_60) - del parameter_60 + del matmul_18, parameter_60 # pd_op.reshape: (1x11x12x64xf32) <- (1x11x768xf32, 4xi64) reshape_9 = paddle._C_ops.reshape(add_22, full_int_array_1) + del add_22 # pd_op.transpose: (1x12x11x64xf32) <- (1x11x12x64xf32) transpose_9 = paddle._C_ops.transpose(reshape_9, [0, 2, 1, 3]) @@ -633,6 +591,7 @@ def forward( # pd_op.reshape: (1x11x12x64xf32) <- (1x11x768xf32, 4xi64) reshape_10 = paddle._C_ops.reshape(add_23, full_int_array_1) + del add_23 # pd_op.transpose: (1x12x11x64xf32) <- (1x11x12x64xf32) transpose_10 = paddle._C_ops.transpose(reshape_10, [0, 2, 1, 3]) @@ -644,9 +603,11 @@ def forward( # pd_op.matmul: (1x12x11x11xf32) <- (1x12x11x64xf32, 1x12x11x64xf32) matmul_19 = paddle._C_ops.matmul(scale_4, transpose_9, False, True) + del scale_4, transpose_9 # pd_op.add: (1x12x11x11xf32) <- (1x12x11x11xf32, 1x1x1x11xf32) add_24 = paddle._C_ops.add(matmul_19, unsqueeze_0) + del matmul_19 # pd_op.softmax: (1x12x11x11xf32) <- (1x12x11x11xf32) softmax_2 = paddle._C_ops.softmax(add_24, -1) @@ -655,13 +616,15 @@ def forward( # pd_op.dropout: (1x12x11x11xf32, 1x12x11x11xui8) <- (1x12x11x11xf32, None, 1xf32) dropout_14, dropout_15 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_2, None, full_5, False, "upscale_in_train", 0, False + softmax_2, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_2 # pd_op.matmul: (1x12x11x64xf32) <- (1x12x11x11xf32, 1x12x11x64xf32) matmul_20 = paddle._C_ops.matmul(dropout_14, transpose_10, False, False) + del dropout_14, transpose_10 # pd_op.transpose: (1x11x12x64xf32) <- (1x12x11x64xf32) transpose_11 = paddle._C_ops.transpose(matmul_20, [0, 2, 1, 3]) @@ -669,19 +632,20 @@ def forward( # pd_op.reshape: (1x11x768xf32) <- (1x11x12x64xf32, 3xi64) reshape_11 = paddle._C_ops.reshape(transpose_11, full_int_array_2) + del transpose_11 # pd_op.matmul: (1x11x768xf32) <- (1x11x768xf32, 768x768xf32) matmul_21 = paddle._C_ops.matmul(reshape_11, parameter_59, False, False) - del parameter_59 + del parameter_59, reshape_11 # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_25 = paddle._C_ops.add(matmul_21, parameter_58) - del parameter_58 + del matmul_21, parameter_58 # pd_op.dropout: (1x11x768xf32, 1x11x768xui8) <- (1x11x768xf32, None, 1xf32) dropout_16, dropout_17 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_25, None, full_5, False, "upscale_in_train", 0, False + add_25, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -689,6 +653,7 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 1x11x768xf32) add_26 = paddle._C_ops.add(layer_norm_12, dropout_16) + del dropout_16, layer_norm_12 # pd_op.layer_norm: (1x11x768xf32, 1x11xf32, 1x11xf32) <- (1x11x768xf32, 768xf32, 768xf32) layer_norm_15, layer_norm_16, layer_norm_17 = (lambda x, f: f(x))( @@ -697,7 +662,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_52, parameter_53 + del add_26, parameter_52, parameter_53 # pd_op.matmul: (1x11x3072xf32) <- (1x11x768xf32, 768x3072xf32) matmul_22 = paddle._C_ops.matmul(layer_norm_15, parameter_57, False, False) @@ -705,23 +670,24 @@ def forward( # pd_op.add: (1x11x3072xf32) <- (1x11x3072xf32, 3072xf32) add_27 = paddle._C_ops.add(matmul_22, parameter_56) - del parameter_56 + del matmul_22, parameter_56 # pd_op.gelu: (1x11x3072xf32) <- (1x11x3072xf32) gelu_2 = paddle._C_ops.gelu(add_27, False) + del add_27 # pd_op.matmul: (1x11x768xf32) <- (1x11x3072xf32, 3072x768xf32) matmul_23 = paddle._C_ops.matmul(gelu_2, parameter_55, False, False) - del parameter_55 + del gelu_2, parameter_55 # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_28 = paddle._C_ops.add(matmul_23, parameter_54) - del parameter_54 + del matmul_23, parameter_54 # pd_op.dropout: (1x11x768xf32, 1x11x768xui8) <- (1x11x768xf32, None, 1xf32) dropout_18, dropout_19 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_28, None, full_5, False, "upscale_in_train", 0, False + add_28, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -729,6 +695,7 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 1x11x768xf32) add_29 = paddle._C_ops.add(layer_norm_15, dropout_18) + del dropout_18, layer_norm_15 # pd_op.layer_norm: (1x11x768xf32, 1x11xf32, 1x11xf32) <- (1x11x768xf32, 768xf32, 768xf32) layer_norm_18, layer_norm_19, layer_norm_20 = (lambda x, f: f(x))( @@ -737,7 +704,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_50, parameter_51 + del add_29, parameter_50, parameter_51 # pd_op.matmul: (1x11x768xf32) <- (1x11x768xf32, 768x768xf32) matmul_24 = paddle._C_ops.matmul(layer_norm_18, parameter_49, False, False) @@ -745,10 +712,11 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_30 = paddle._C_ops.add(matmul_24, parameter_48) - del parameter_48 + del matmul_24, parameter_48 # pd_op.reshape: (1x11x12x64xf32) <- (1x11x768xf32, 4xi64) reshape_12 = paddle._C_ops.reshape(add_30, full_int_array_1) + del add_30 # pd_op.transpose: (1x12x11x64xf32) <- (1x11x12x64xf32) transpose_12 = paddle._C_ops.transpose(reshape_12, [0, 2, 1, 3]) @@ -760,7 +728,7 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_31 = paddle._C_ops.add(matmul_25, parameter_46) - del parameter_46 + del matmul_25, parameter_46 # pd_op.matmul: (1x11x768xf32) <- (1x11x768xf32, 768x768xf32) matmul_26 = paddle._C_ops.matmul(layer_norm_18, parameter_45, False, False) @@ -768,10 +736,11 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_32 = paddle._C_ops.add(matmul_26, parameter_44) - del parameter_44 + del matmul_26, parameter_44 # pd_op.reshape: (1x11x12x64xf32) <- (1x11x768xf32, 4xi64) reshape_13 = paddle._C_ops.reshape(add_31, full_int_array_1) + del add_31 # pd_op.transpose: (1x12x11x64xf32) <- (1x11x12x64xf32) transpose_13 = paddle._C_ops.transpose(reshape_13, [0, 2, 1, 3]) @@ -779,6 +748,7 @@ def forward( # pd_op.reshape: (1x11x12x64xf32) <- (1x11x768xf32, 4xi64) reshape_14 = paddle._C_ops.reshape(add_32, full_int_array_1) + del add_32 # pd_op.transpose: (1x12x11x64xf32) <- (1x11x12x64xf32) transpose_14 = paddle._C_ops.transpose(reshape_14, [0, 2, 1, 3]) @@ -790,9 +760,11 @@ def forward( # pd_op.matmul: (1x12x11x11xf32) <- (1x12x11x64xf32, 1x12x11x64xf32) matmul_27 = paddle._C_ops.matmul(scale_5, transpose_13, False, True) + del scale_5, transpose_13 # pd_op.add: (1x12x11x11xf32) <- (1x12x11x11xf32, 1x1x1x11xf32) add_33 = paddle._C_ops.add(matmul_27, unsqueeze_0) + del matmul_27 # pd_op.softmax: (1x12x11x11xf32) <- (1x12x11x11xf32) softmax_3 = paddle._C_ops.softmax(add_33, -1) @@ -801,13 +773,15 @@ def forward( # pd_op.dropout: (1x12x11x11xf32, 1x12x11x11xui8) <- (1x12x11x11xf32, None, 1xf32) dropout_20, dropout_21 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_3, None, full_5, False, "upscale_in_train", 0, False + softmax_3, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_3 # pd_op.matmul: (1x12x11x64xf32) <- (1x12x11x11xf32, 1x12x11x64xf32) matmul_28 = paddle._C_ops.matmul(dropout_20, transpose_14, False, False) + del dropout_20, transpose_14 # pd_op.transpose: (1x11x12x64xf32) <- (1x12x11x64xf32) transpose_15 = paddle._C_ops.transpose(matmul_28, [0, 2, 1, 3]) @@ -815,19 +789,20 @@ def forward( # pd_op.reshape: (1x11x768xf32) <- (1x11x12x64xf32, 3xi64) reshape_15 = paddle._C_ops.reshape(transpose_15, full_int_array_2) + del transpose_15 # pd_op.matmul: (1x11x768xf32) <- (1x11x768xf32, 768x768xf32) matmul_29 = paddle._C_ops.matmul(reshape_15, parameter_43, False, False) - del parameter_43 + del parameter_43, reshape_15 # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_34 = paddle._C_ops.add(matmul_29, parameter_42) - del parameter_42 + del matmul_29, parameter_42 # pd_op.dropout: (1x11x768xf32, 1x11x768xui8) <- (1x11x768xf32, None, 1xf32) dropout_22, dropout_23 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_34, None, full_5, False, "upscale_in_train", 0, False + add_34, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -835,6 +810,7 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 1x11x768xf32) add_35 = paddle._C_ops.add(layer_norm_18, dropout_22) + del dropout_22, layer_norm_18 # pd_op.layer_norm: (1x11x768xf32, 1x11xf32, 1x11xf32) <- (1x11x768xf32, 768xf32, 768xf32) layer_norm_21, layer_norm_22, layer_norm_23 = (lambda x, f: f(x))( @@ -843,7 +819,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_36, parameter_37 + del add_35, parameter_36, parameter_37 # pd_op.matmul: (1x11x3072xf32) <- (1x11x768xf32, 768x3072xf32) matmul_30 = paddle._C_ops.matmul(layer_norm_21, parameter_41, False, False) @@ -851,23 +827,24 @@ def forward( # pd_op.add: (1x11x3072xf32) <- (1x11x3072xf32, 3072xf32) add_36 = paddle._C_ops.add(matmul_30, parameter_40) - del parameter_40 + del matmul_30, parameter_40 # pd_op.gelu: (1x11x3072xf32) <- (1x11x3072xf32) gelu_3 = paddle._C_ops.gelu(add_36, False) + del add_36 # pd_op.matmul: (1x11x768xf32) <- (1x11x3072xf32, 3072x768xf32) matmul_31 = paddle._C_ops.matmul(gelu_3, parameter_39, False, False) - del parameter_39 + del gelu_3, parameter_39 # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_37 = paddle._C_ops.add(matmul_31, parameter_38) - del parameter_38 + del matmul_31, parameter_38 # pd_op.dropout: (1x11x768xf32, 1x11x768xui8) <- (1x11x768xf32, None, 1xf32) dropout_24, dropout_25 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_37, None, full_5, False, "upscale_in_train", 0, False + add_37, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -875,6 +852,7 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 1x11x768xf32) add_38 = paddle._C_ops.add(layer_norm_21, dropout_24) + del dropout_24, layer_norm_21 # pd_op.layer_norm: (1x11x768xf32, 1x11xf32, 1x11xf32) <- (1x11x768xf32, 768xf32, 768xf32) layer_norm_24, layer_norm_25, layer_norm_26 = (lambda x, f: f(x))( @@ -883,7 +861,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_34, parameter_35 + del add_38, parameter_34, parameter_35 # pd_op.matmul: (1x11x768xf32) <- (1x11x768xf32, 768x768xf32) matmul_32 = paddle._C_ops.matmul(layer_norm_24, parameter_33, False, False) @@ -891,10 +869,11 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_39 = paddle._C_ops.add(matmul_32, parameter_32) - del parameter_32 + del matmul_32, parameter_32 # pd_op.reshape: (1x11x12x64xf32) <- (1x11x768xf32, 4xi64) reshape_16 = paddle._C_ops.reshape(add_39, full_int_array_1) + del add_39 # pd_op.transpose: (1x12x11x64xf32) <- (1x11x12x64xf32) transpose_16 = paddle._C_ops.transpose(reshape_16, [0, 2, 1, 3]) @@ -906,7 +885,7 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_40 = paddle._C_ops.add(matmul_33, parameter_30) - del parameter_30 + del matmul_33, parameter_30 # pd_op.matmul: (1x11x768xf32) <- (1x11x768xf32, 768x768xf32) matmul_34 = paddle._C_ops.matmul(layer_norm_24, parameter_29, False, False) @@ -914,10 +893,11 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_41 = paddle._C_ops.add(matmul_34, parameter_28) - del parameter_28 + del matmul_34, parameter_28 # pd_op.reshape: (1x11x12x64xf32) <- (1x11x768xf32, 4xi64) reshape_17 = paddle._C_ops.reshape(add_40, full_int_array_1) + del add_40 # pd_op.transpose: (1x12x11x64xf32) <- (1x11x12x64xf32) transpose_17 = paddle._C_ops.transpose(reshape_17, [0, 2, 1, 3]) @@ -925,6 +905,7 @@ def forward( # pd_op.reshape: (1x11x12x64xf32) <- (1x11x768xf32, 4xi64) reshape_18 = paddle._C_ops.reshape(add_41, full_int_array_1) + del add_41 # pd_op.transpose: (1x12x11x64xf32) <- (1x11x12x64xf32) transpose_18 = paddle._C_ops.transpose(reshape_18, [0, 2, 1, 3]) @@ -936,9 +917,11 @@ def forward( # pd_op.matmul: (1x12x11x11xf32) <- (1x12x11x64xf32, 1x12x11x64xf32) matmul_35 = paddle._C_ops.matmul(scale_6, transpose_17, False, True) + del scale_6, transpose_17 # pd_op.add: (1x12x11x11xf32) <- (1x12x11x11xf32, 1x1x1x11xf32) add_42 = paddle._C_ops.add(matmul_35, unsqueeze_0) + del matmul_35 # pd_op.softmax: (1x12x11x11xf32) <- (1x12x11x11xf32) softmax_4 = paddle._C_ops.softmax(add_42, -1) @@ -947,13 +930,15 @@ def forward( # pd_op.dropout: (1x12x11x11xf32, 1x12x11x11xui8) <- (1x12x11x11xf32, None, 1xf32) dropout_26, dropout_27 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_4, None, full_5, False, "upscale_in_train", 0, False + softmax_4, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_4 # pd_op.matmul: (1x12x11x64xf32) <- (1x12x11x11xf32, 1x12x11x64xf32) matmul_36 = paddle._C_ops.matmul(dropout_26, transpose_18, False, False) + del dropout_26, transpose_18 # pd_op.transpose: (1x11x12x64xf32) <- (1x12x11x64xf32) transpose_19 = paddle._C_ops.transpose(matmul_36, [0, 2, 1, 3]) @@ -961,19 +946,20 @@ def forward( # pd_op.reshape: (1x11x768xf32) <- (1x11x12x64xf32, 3xi64) reshape_19 = paddle._C_ops.reshape(transpose_19, full_int_array_2) + del transpose_19 # pd_op.matmul: (1x11x768xf32) <- (1x11x768xf32, 768x768xf32) matmul_37 = paddle._C_ops.matmul(reshape_19, parameter_27, False, False) - del parameter_27 + del parameter_27, reshape_19 # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_43 = paddle._C_ops.add(matmul_37, parameter_26) - del parameter_26 + del matmul_37, parameter_26 # pd_op.dropout: (1x11x768xf32, 1x11x768xui8) <- (1x11x768xf32, None, 1xf32) dropout_28, dropout_29 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_43, None, full_5, False, "upscale_in_train", 0, False + add_43, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -981,6 +967,7 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 1x11x768xf32) add_44 = paddle._C_ops.add(layer_norm_24, dropout_28) + del dropout_28, layer_norm_24 # pd_op.layer_norm: (1x11x768xf32, 1x11xf32, 1x11xf32) <- (1x11x768xf32, 768xf32, 768xf32) layer_norm_27, layer_norm_28, layer_norm_29 = (lambda x, f: f(x))( @@ -989,7 +976,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_20, parameter_21 + del add_44, parameter_20, parameter_21 # pd_op.matmul: (1x11x3072xf32) <- (1x11x768xf32, 768x3072xf32) matmul_38 = paddle._C_ops.matmul(layer_norm_27, parameter_25, False, False) @@ -997,23 +984,24 @@ def forward( # pd_op.add: (1x11x3072xf32) <- (1x11x3072xf32, 3072xf32) add_45 = paddle._C_ops.add(matmul_38, parameter_24) - del parameter_24 + del matmul_38, parameter_24 # pd_op.gelu: (1x11x3072xf32) <- (1x11x3072xf32) gelu_4 = paddle._C_ops.gelu(add_45, False) + del add_45 # pd_op.matmul: (1x11x768xf32) <- (1x11x3072xf32, 3072x768xf32) matmul_39 = paddle._C_ops.matmul(gelu_4, parameter_23, False, False) - del parameter_23 + del gelu_4, parameter_23 # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_46 = paddle._C_ops.add(matmul_39, parameter_22) - del parameter_22 + del matmul_39, parameter_22 # pd_op.dropout: (1x11x768xf32, 1x11x768xui8) <- (1x11x768xf32, None, 1xf32) dropout_30, dropout_31 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_46, None, full_5, False, "upscale_in_train", 0, False + add_46, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1021,6 +1009,7 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 1x11x768xf32) add_47 = paddle._C_ops.add(layer_norm_27, dropout_30) + del dropout_30, layer_norm_27 # pd_op.layer_norm: (1x11x768xf32, 1x11xf32, 1x11xf32) <- (1x11x768xf32, 768xf32, 768xf32) layer_norm_30, layer_norm_31, layer_norm_32 = (lambda x, f: f(x))( @@ -1029,7 +1018,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_18, parameter_19 + del add_47, parameter_18, parameter_19 # pd_op.matmul: (1x11x768xf32) <- (1x11x768xf32, 768x768xf32) matmul_40 = paddle._C_ops.matmul(layer_norm_30, parameter_17, False, False) @@ -1037,10 +1026,11 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_48 = paddle._C_ops.add(matmul_40, parameter_16) - del parameter_16 + del matmul_40, parameter_16 # pd_op.reshape: (1x11x12x64xf32) <- (1x11x768xf32, 4xi64) reshape_20 = paddle._C_ops.reshape(add_48, full_int_array_1) + del add_48 # pd_op.transpose: (1x12x11x64xf32) <- (1x11x12x64xf32) transpose_20 = paddle._C_ops.transpose(reshape_20, [0, 2, 1, 3]) @@ -1052,7 +1042,7 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_49 = paddle._C_ops.add(matmul_41, parameter_14) - del parameter_14 + del matmul_41, parameter_14 # pd_op.matmul: (1x11x768xf32) <- (1x11x768xf32, 768x768xf32) matmul_42 = paddle._C_ops.matmul(layer_norm_30, parameter_13, False, False) @@ -1060,10 +1050,11 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_50 = paddle._C_ops.add(matmul_42, parameter_12) - del parameter_12 + del matmul_42, parameter_12 # pd_op.reshape: (1x11x12x64xf32) <- (1x11x768xf32, 4xi64) reshape_21 = paddle._C_ops.reshape(add_49, full_int_array_1) + del add_49 # pd_op.transpose: (1x12x11x64xf32) <- (1x11x12x64xf32) transpose_21 = paddle._C_ops.transpose(reshape_21, [0, 2, 1, 3]) @@ -1071,7 +1062,7 @@ def forward( # pd_op.reshape: (1x11x12x64xf32) <- (1x11x768xf32, 4xi64) reshape_22 = paddle._C_ops.reshape(add_50, full_int_array_1) - del full_int_array_1 + del add_50, full_int_array_1 # pd_op.transpose: (1x12x11x64xf32) <- (1x11x12x64xf32) transpose_22 = paddle._C_ops.transpose(reshape_22, [0, 2, 1, 3]) @@ -1079,13 +1070,15 @@ def forward( # pd_op.scale: (1x12x11x64xf32) <- (1x12x11x64xf32, 1xf32) scale_7 = paddle._C_ops.scale(transpose_20, full_6, float("0"), True) - del transpose_20 + del full_6, transpose_20 # pd_op.matmul: (1x12x11x11xf32) <- (1x12x11x64xf32, 1x12x11x64xf32) matmul_43 = paddle._C_ops.matmul(scale_7, transpose_21, False, True) + del scale_7, transpose_21 # pd_op.add: (1x12x11x11xf32) <- (1x12x11x11xf32, 1x1x1x11xf32) add_51 = paddle._C_ops.add(matmul_43, unsqueeze_0) + del matmul_43, unsqueeze_0 # pd_op.softmax: (1x12x11x11xf32) <- (1x12x11x11xf32) softmax_5 = paddle._C_ops.softmax(add_51, -1) @@ -1094,13 +1087,15 @@ def forward( # pd_op.dropout: (1x12x11x11xf32, 1x12x11x11xui8) <- (1x12x11x11xf32, None, 1xf32) dropout_32, dropout_33 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_5, None, full_5, False, "upscale_in_train", 0, False + softmax_5, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_5 # pd_op.matmul: (1x12x11x64xf32) <- (1x12x11x11xf32, 1x12x11x64xf32) matmul_44 = paddle._C_ops.matmul(dropout_32, transpose_22, False, False) + del dropout_32, transpose_22 # pd_op.transpose: (1x11x12x64xf32) <- (1x12x11x64xf32) transpose_23 = paddle._C_ops.transpose(matmul_44, [0, 2, 1, 3]) @@ -1108,20 +1103,20 @@ def forward( # pd_op.reshape: (1x11x768xf32) <- (1x11x12x64xf32, 3xi64) reshape_23 = paddle._C_ops.reshape(transpose_23, full_int_array_2) - del full_int_array_2 + del full_int_array_2, transpose_23 # pd_op.matmul: (1x11x768xf32) <- (1x11x768xf32, 768x768xf32) matmul_45 = paddle._C_ops.matmul(reshape_23, parameter_11, False, False) - del parameter_11 + del parameter_11, reshape_23 # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_52 = paddle._C_ops.add(matmul_45, parameter_10) - del parameter_10 + del matmul_45, parameter_10 # pd_op.dropout: (1x11x768xf32, 1x11x768xui8) <- (1x11x768xf32, None, 1xf32) dropout_34, dropout_35 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_52, None, full_5, False, "upscale_in_train", 0, False + add_52, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1129,6 +1124,7 @@ def forward( # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 1x11x768xf32) add_53 = paddle._C_ops.add(layer_norm_30, dropout_34) + del dropout_34, layer_norm_30 # pd_op.layer_norm: (1x11x768xf32, 1x11xf32, 1x11xf32) <- (1x11x768xf32, 768xf32, 768xf32) layer_norm_33, layer_norm_34, layer_norm_35 = (lambda x, f: f(x))( @@ -1137,7 +1133,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_4, parameter_5 + del add_53, parameter_4, parameter_5 # pd_op.matmul: (1x11x3072xf32) <- (1x11x768xf32, 768x3072xf32) matmul_46 = paddle._C_ops.matmul(layer_norm_33, parameter_9, False, False) @@ -1145,30 +1141,32 @@ def forward( # pd_op.add: (1x11x3072xf32) <- (1x11x3072xf32, 3072xf32) add_54 = paddle._C_ops.add(matmul_46, parameter_8) - del parameter_8 + del matmul_46, parameter_8 # pd_op.gelu: (1x11x3072xf32) <- (1x11x3072xf32) gelu_5 = paddle._C_ops.gelu(add_54, False) + del add_54 # pd_op.matmul: (1x11x768xf32) <- (1x11x3072xf32, 3072x768xf32) matmul_47 = paddle._C_ops.matmul(gelu_5, parameter_7, False, False) - del parameter_7 + del gelu_5, parameter_7 # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 768xf32) add_55 = paddle._C_ops.add(matmul_47, parameter_6) - del parameter_6 + del matmul_47, parameter_6 # pd_op.dropout: (1x11x768xf32, 1x11x768xui8) <- (1x11x768xf32, None, 1xf32) dropout_36, dropout_37 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_55, None, full_5, False, "upscale_in_train", 0, False + add_55, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) - del add_55 + del add_55, full_5 # pd_op.add: (1x11x768xf32) <- (1x11x768xf32, 1x11x768xf32) add_56 = paddle._C_ops.add(layer_norm_33, dropout_36) + del dropout_36, layer_norm_33 # pd_op.layer_norm: (1x11x768xf32, 1x11xf32, 1x11xf32) <- (1x11x768xf32, 768xf32, 768xf32) layer_norm_36, layer_norm_37, layer_norm_38 = (lambda x, f: f(x))( @@ -1177,7 +1175,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_2, parameter_3 + del add_56, parameter_2, parameter_3 # pd_op.full_int_array: (1xi64) <- () full_int_array_3 = [0] @@ -1189,254 +1187,18 @@ def forward( slice_0 = paddle._C_ops.slice( layer_norm_36, [1], full_int_array_3, full_int_array_4, [1], [1] ) + del full_int_array_3, full_int_array_4, layer_norm_36 # pd_op.matmul: (1x768xf32) <- (1x768xf32, 768x768xf32) matmul_48 = paddle._C_ops.matmul(slice_0, parameter_1, False, False) - del parameter_1 + del parameter_1, slice_0 # pd_op.add: (1x768xf32) <- (1x768xf32, 768xf32) add_57 = paddle._C_ops.add(matmul_48, parameter_0) - del parameter_0 + del matmul_48, parameter_0 # pd_op.tanh: (1x768xf32) <- (1x768xf32) tanh_0 = paddle._C_ops.tanh(add_57) - del ( - add_0, - add_1, - add_11, - add_12, - add_13, - add_14, - add_17, - add_18, - add_2, - add_20, - add_21, - add_22, - add_23, - add_26, - add_27, - add_29, - add_3, - add_30, - add_31, - add_32, - add_35, - add_36, - add_38, - add_39, - add_4, - add_40, - add_41, - add_44, - add_45, - add_47, - add_48, - add_49, - add_5, - add_50, - add_53, - add_54, - add_56, - add_57, - add_8, - add_9, - assign_0, - assign_1, - assign_10, - assign_11, - assign_12, - assign_13, - assign_14, - assign_15, - assign_16, - assign_17, - assign_18, - assign_19, - assign_2, - assign_20, - assign_21, - assign_22, - assign_3, - assign_4, - assign_5, - assign_6, - assign_7, - assign_8, - assign_9, - dropout_0, - dropout_1, - dropout_10, - dropout_11, - dropout_12, - dropout_13, - dropout_14, - dropout_15, - dropout_16, - dropout_17, - dropout_18, - dropout_19, - dropout_2, - dropout_20, - dropout_21, - dropout_22, - dropout_23, - dropout_24, - dropout_25, - dropout_26, - dropout_27, - dropout_28, - dropout_29, - dropout_3, - dropout_30, - dropout_31, - dropout_32, - dropout_33, - dropout_34, - dropout_35, - dropout_36, - dropout_37, - dropout_4, - dropout_5, - dropout_6, - dropout_7, - dropout_8, - dropout_9, - embedding_0, - embedding_1, - embedding_2, - embedding_3, - full_5, - full_6, - full_int_array_3, - full_int_array_4, - gelu_0, - gelu_1, - gelu_2, - gelu_3, - gelu_4, - gelu_5, - layer_norm_1, - layer_norm_10, - layer_norm_11, - layer_norm_12, - layer_norm_13, - layer_norm_14, - layer_norm_15, - layer_norm_16, - layer_norm_17, - layer_norm_18, - layer_norm_19, - layer_norm_2, - layer_norm_20, - layer_norm_21, - layer_norm_22, - layer_norm_23, - layer_norm_24, - layer_norm_25, - layer_norm_26, - layer_norm_27, - layer_norm_28, - layer_norm_29, - layer_norm_3, - layer_norm_30, - layer_norm_31, - layer_norm_32, - layer_norm_33, - layer_norm_34, - layer_norm_35, - layer_norm_36, - layer_norm_37, - layer_norm_38, - layer_norm_4, - layer_norm_5, - layer_norm_6, - layer_norm_7, - layer_norm_8, - layer_norm_9, - matmul_0, - matmul_1, - matmul_10, - matmul_11, - matmul_13, - matmul_14, - matmul_15, - matmul_16, - matmul_17, - matmul_18, - matmul_19, - matmul_2, - matmul_21, - matmul_22, - matmul_23, - matmul_24, - matmul_25, - matmul_26, - matmul_27, - matmul_29, - matmul_3, - matmul_30, - matmul_31, - matmul_32, - matmul_33, - matmul_34, - matmul_35, - matmul_37, - matmul_38, - matmul_39, - matmul_40, - matmul_41, - matmul_42, - matmul_43, - matmul_45, - matmul_46, - matmul_47, - matmul_48, - matmul_5, - matmul_6, - matmul_7, - matmul_8, - matmul_9, - reshape_11, - reshape_15, - reshape_19, - reshape_23, - reshape_3, - reshape_7, - scale_1, - scale_2, - scale_3, - scale_4, - scale_5, - scale_6, - scale_7, - slice_0, - softmax_0, - softmax_1, - softmax_2, - softmax_3, - softmax_4, - softmax_5, - subtract_0, - transpose_1, - transpose_10, - transpose_11, - transpose_13, - transpose_14, - transpose_15, - transpose_17, - transpose_18, - transpose_19, - transpose_2, - transpose_21, - transpose_22, - transpose_23, - transpose_3, - transpose_5, - transpose_6, - transpose_7, - transpose_9, - unsqueeze_0, - ) + del add_57 return tanh_0 diff --git a/paddle_samples/PaddleNLP/ernie-3.0-medium-zh/weight_meta.py b/paddle_samples/PaddleNLP/ernie-3.0-medium-zh/weight_meta.py index 2c2ed1604..7cc6ba3a2 100644 --- a/paddle_samples/PaddleNLP/ernie-3.0-medium-zh/weight_meta.py +++ b/paddle_samples/PaddleNLP/ernie-3.0-medium-zh/weight_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_parameter_0: name = "parameter_0" + original_name = "linear_691.b_0" shape = [768] dtype = "float32" data = None @@ -7,6 +8,7 @@ class Program_weight_tensor_parameter_0: class Program_weight_tensor_parameter_1: name = "parameter_1" + original_name = "linear_691.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.0994008") @@ -18,6 +20,7 @@ class Program_weight_tensor_parameter_1: class Program_weight_tensor_parameter_2: name = "parameter_2" + original_name = "layer_norm_235.b_0" shape = [768] dtype = "float32" min_val = float("-0.613629") @@ -29,6 +32,7 @@ class Program_weight_tensor_parameter_2: class Program_weight_tensor_parameter_3: name = "parameter_3" + original_name = "layer_norm_235.w_0" shape = [768] dtype = "float32" min_val = float("0.445939") @@ -40,6 +44,7 @@ class Program_weight_tensor_parameter_3: class Program_weight_tensor_parameter_4: name = "parameter_4" + original_name = "layer_norm_234.b_0" shape = [768] dtype = "float32" min_val = float("-0.78441") @@ -51,6 +56,7 @@ class Program_weight_tensor_parameter_4: class Program_weight_tensor_parameter_5: name = "parameter_5" + original_name = "layer_norm_234.w_0" shape = [768] dtype = "float32" min_val = float("0.216457") @@ -62,6 +68,7 @@ class Program_weight_tensor_parameter_5: class Program_weight_tensor_parameter_6: name = "parameter_6" + original_name = "linear_690.b_0" shape = [768] dtype = "float32" min_val = float("-0.21587") @@ -73,6 +80,7 @@ class Program_weight_tensor_parameter_6: class Program_weight_tensor_parameter_7: name = "parameter_7" + original_name = "linear_690.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-1.78887") @@ -84,6 +92,7 @@ class Program_weight_tensor_parameter_7: class Program_weight_tensor_parameter_8: name = "parameter_8" + original_name = "linear_689.b_0" shape = [3072] dtype = "float32" min_val = float("-0.386474") @@ -95,6 +104,7 @@ class Program_weight_tensor_parameter_8: class Program_weight_tensor_parameter_9: name = "parameter_9" + original_name = "linear_689.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.421915") @@ -106,6 +116,7 @@ class Program_weight_tensor_parameter_9: class Program_weight_tensor_parameter_10: name = "parameter_10" + original_name = "linear_688.b_0" shape = [768] dtype = "float32" min_val = float("-0.928453") @@ -117,6 +128,7 @@ class Program_weight_tensor_parameter_10: class Program_weight_tensor_parameter_11: name = "parameter_11" + original_name = "linear_688.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.560365") @@ -128,6 +140,7 @@ class Program_weight_tensor_parameter_11: class Program_weight_tensor_parameter_12: name = "parameter_12" + original_name = "linear_687.b_0" shape = [768] dtype = "float32" min_val = float("-0.15698") @@ -139,6 +152,7 @@ class Program_weight_tensor_parameter_12: class Program_weight_tensor_parameter_13: name = "parameter_13" + original_name = "linear_687.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.272663") @@ -150,6 +164,7 @@ class Program_weight_tensor_parameter_13: class Program_weight_tensor_parameter_14: name = "parameter_14" + original_name = "linear_686.b_0" shape = [768] dtype = "float32" min_val = float("-0.0212569") @@ -161,6 +176,7 @@ class Program_weight_tensor_parameter_14: class Program_weight_tensor_parameter_15: name = "parameter_15" + original_name = "linear_686.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.37475") @@ -172,6 +188,7 @@ class Program_weight_tensor_parameter_15: class Program_weight_tensor_parameter_16: name = "parameter_16" + original_name = "linear_685.b_0" shape = [768] dtype = "float32" min_val = float("-0.383346") @@ -183,6 +200,7 @@ class Program_weight_tensor_parameter_16: class Program_weight_tensor_parameter_17: name = "parameter_17" + original_name = "linear_685.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.262675") @@ -194,6 +212,7 @@ class Program_weight_tensor_parameter_17: class Program_weight_tensor_parameter_18: name = "parameter_18" + original_name = "layer_norm_233.b_0" shape = [768] dtype = "float32" min_val = float("-0.712593") @@ -205,6 +224,7 @@ class Program_weight_tensor_parameter_18: class Program_weight_tensor_parameter_19: name = "parameter_19" + original_name = "layer_norm_233.w_0" shape = [768] dtype = "float32" min_val = float("0.652106") @@ -216,6 +236,7 @@ class Program_weight_tensor_parameter_19: class Program_weight_tensor_parameter_20: name = "parameter_20" + original_name = "layer_norm_232.b_0" shape = [768] dtype = "float32" min_val = float("-0.600116") @@ -227,6 +248,7 @@ class Program_weight_tensor_parameter_20: class Program_weight_tensor_parameter_21: name = "parameter_21" + original_name = "layer_norm_232.w_0" shape = [768] dtype = "float32" min_val = float("0.404644") @@ -238,6 +260,7 @@ class Program_weight_tensor_parameter_21: class Program_weight_tensor_parameter_22: name = "parameter_22" + original_name = "linear_684.b_0" shape = [768] dtype = "float32" min_val = float("-0.23135") @@ -249,6 +272,7 @@ class Program_weight_tensor_parameter_22: class Program_weight_tensor_parameter_23: name = "parameter_23" + original_name = "linear_684.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-0.576683") @@ -260,6 +284,7 @@ class Program_weight_tensor_parameter_23: class Program_weight_tensor_parameter_24: name = "parameter_24" + original_name = "linear_683.b_0" shape = [3072] dtype = "float32" min_val = float("-0.434721") @@ -271,6 +296,7 @@ class Program_weight_tensor_parameter_24: class Program_weight_tensor_parameter_25: name = "parameter_25" + original_name = "linear_683.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.461051") @@ -282,6 +308,7 @@ class Program_weight_tensor_parameter_25: class Program_weight_tensor_parameter_26: name = "parameter_26" + original_name = "linear_682.b_0" shape = [768] dtype = "float32" min_val = float("-0.776673") @@ -293,6 +320,7 @@ class Program_weight_tensor_parameter_26: class Program_weight_tensor_parameter_27: name = "parameter_27" + original_name = "linear_682.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.587545") @@ -304,6 +332,7 @@ class Program_weight_tensor_parameter_27: class Program_weight_tensor_parameter_28: name = "parameter_28" + original_name = "linear_681.b_0" shape = [768] dtype = "float32" min_val = float("-0.121946") @@ -315,6 +344,7 @@ class Program_weight_tensor_parameter_28: class Program_weight_tensor_parameter_29: name = "parameter_29" + original_name = "linear_681.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.355221") @@ -326,6 +356,7 @@ class Program_weight_tensor_parameter_29: class Program_weight_tensor_parameter_30: name = "parameter_30" + original_name = "linear_680.b_0" shape = [768] dtype = "float32" min_val = float("-0.00326389") @@ -337,6 +368,7 @@ class Program_weight_tensor_parameter_30: class Program_weight_tensor_parameter_31: name = "parameter_31" + original_name = "linear_680.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.322173") @@ -348,6 +380,7 @@ class Program_weight_tensor_parameter_31: class Program_weight_tensor_parameter_32: name = "parameter_32" + original_name = "linear_679.b_0" shape = [768] dtype = "float32" min_val = float("-0.517067") @@ -359,6 +392,7 @@ class Program_weight_tensor_parameter_32: class Program_weight_tensor_parameter_33: name = "parameter_33" + original_name = "linear_679.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.247652") @@ -370,6 +404,7 @@ class Program_weight_tensor_parameter_33: class Program_weight_tensor_parameter_34: name = "parameter_34" + original_name = "layer_norm_231.b_0" shape = [768] dtype = "float32" min_val = float("-0.50621") @@ -381,6 +416,7 @@ class Program_weight_tensor_parameter_34: class Program_weight_tensor_parameter_35: name = "parameter_35" + original_name = "layer_norm_231.w_0" shape = [768] dtype = "float32" min_val = float("0.695902") @@ -392,6 +428,7 @@ class Program_weight_tensor_parameter_35: class Program_weight_tensor_parameter_36: name = "parameter_36" + original_name = "layer_norm_230.b_0" shape = [768] dtype = "float32" min_val = float("-0.741106") @@ -403,6 +440,7 @@ class Program_weight_tensor_parameter_36: class Program_weight_tensor_parameter_37: name = "parameter_37" + original_name = "layer_norm_230.w_0" shape = [768] dtype = "float32" min_val = float("0.47554") @@ -414,6 +452,7 @@ class Program_weight_tensor_parameter_37: class Program_weight_tensor_parameter_38: name = "parameter_38" + original_name = "linear_678.b_0" shape = [768] dtype = "float32" min_val = float("-0.208921") @@ -425,6 +464,7 @@ class Program_weight_tensor_parameter_38: class Program_weight_tensor_parameter_39: name = "parameter_39" + original_name = "linear_678.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-1.401") @@ -436,6 +476,7 @@ class Program_weight_tensor_parameter_39: class Program_weight_tensor_parameter_40: name = "parameter_40" + original_name = "linear_677.b_0" shape = [3072] dtype = "float32" min_val = float("-0.479829") @@ -447,6 +488,7 @@ class Program_weight_tensor_parameter_40: class Program_weight_tensor_parameter_41: name = "parameter_41" + original_name = "linear_677.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.4543") @@ -458,6 +500,7 @@ class Program_weight_tensor_parameter_41: class Program_weight_tensor_parameter_42: name = "parameter_42" + original_name = "linear_676.b_0" shape = [768] dtype = "float32" min_val = float("-0.560588") @@ -469,6 +512,7 @@ class Program_weight_tensor_parameter_42: class Program_weight_tensor_parameter_43: name = "parameter_43" + original_name = "linear_676.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.360816") @@ -480,6 +524,7 @@ class Program_weight_tensor_parameter_43: class Program_weight_tensor_parameter_44: name = "parameter_44" + original_name = "linear_675.b_0" shape = [768] dtype = "float32" min_val = float("-0.158797") @@ -491,6 +536,7 @@ class Program_weight_tensor_parameter_44: class Program_weight_tensor_parameter_45: name = "parameter_45" + original_name = "linear_675.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.578496") @@ -502,6 +548,7 @@ class Program_weight_tensor_parameter_45: class Program_weight_tensor_parameter_46: name = "parameter_46" + original_name = "linear_674.b_0" shape = [768] dtype = "float32" min_val = float("-0.00288625") @@ -513,6 +560,7 @@ class Program_weight_tensor_parameter_46: class Program_weight_tensor_parameter_47: name = "parameter_47" + original_name = "linear_674.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.499791") @@ -524,6 +572,7 @@ class Program_weight_tensor_parameter_47: class Program_weight_tensor_parameter_48: name = "parameter_48" + original_name = "linear_673.b_0" shape = [768] dtype = "float32" min_val = float("-0.402486") @@ -535,6 +584,7 @@ class Program_weight_tensor_parameter_48: class Program_weight_tensor_parameter_49: name = "parameter_49" + original_name = "linear_673.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.305264") @@ -546,6 +596,7 @@ class Program_weight_tensor_parameter_49: class Program_weight_tensor_parameter_50: name = "parameter_50" + original_name = "layer_norm_229.b_0" shape = [768] dtype = "float32" min_val = float("-0.624001") @@ -557,6 +608,7 @@ class Program_weight_tensor_parameter_50: class Program_weight_tensor_parameter_51: name = "parameter_51" + original_name = "layer_norm_229.w_0" shape = [768] dtype = "float32" min_val = float("0.80285") @@ -568,6 +620,7 @@ class Program_weight_tensor_parameter_51: class Program_weight_tensor_parameter_52: name = "parameter_52" + original_name = "layer_norm_228.b_0" shape = [768] dtype = "float32" min_val = float("-0.773582") @@ -579,6 +632,7 @@ class Program_weight_tensor_parameter_52: class Program_weight_tensor_parameter_53: name = "parameter_53" + original_name = "layer_norm_228.w_0" shape = [768] dtype = "float32" min_val = float("0.473972") @@ -590,6 +644,7 @@ class Program_weight_tensor_parameter_53: class Program_weight_tensor_parameter_54: name = "parameter_54" + original_name = "linear_672.b_0" shape = [768] dtype = "float32" min_val = float("-0.210172") @@ -601,6 +656,7 @@ class Program_weight_tensor_parameter_54: class Program_weight_tensor_parameter_55: name = "parameter_55" + original_name = "linear_672.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-1.78661") @@ -612,6 +668,7 @@ class Program_weight_tensor_parameter_55: class Program_weight_tensor_parameter_56: name = "parameter_56" + original_name = "linear_671.b_0" shape = [3072] dtype = "float32" min_val = float("-0.542027") @@ -623,6 +680,7 @@ class Program_weight_tensor_parameter_56: class Program_weight_tensor_parameter_57: name = "parameter_57" + original_name = "linear_671.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.55249") @@ -634,6 +692,7 @@ class Program_weight_tensor_parameter_57: class Program_weight_tensor_parameter_58: name = "parameter_58" + original_name = "linear_670.b_0" shape = [768] dtype = "float32" min_val = float("-0.415902") @@ -645,6 +704,7 @@ class Program_weight_tensor_parameter_58: class Program_weight_tensor_parameter_59: name = "parameter_59" + original_name = "linear_670.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.369138") @@ -656,6 +716,7 @@ class Program_weight_tensor_parameter_59: class Program_weight_tensor_parameter_60: name = "parameter_60" + original_name = "linear_669.b_0" shape = [768] dtype = "float32" min_val = float("-0.151438") @@ -667,6 +728,7 @@ class Program_weight_tensor_parameter_60: class Program_weight_tensor_parameter_61: name = "parameter_61" + original_name = "linear_669.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.322518") @@ -678,6 +740,7 @@ class Program_weight_tensor_parameter_61: class Program_weight_tensor_parameter_62: name = "parameter_62" + original_name = "linear_668.b_0" shape = [768] dtype = "float32" min_val = float("-0.00115705") @@ -689,6 +752,7 @@ class Program_weight_tensor_parameter_62: class Program_weight_tensor_parameter_63: name = "parameter_63" + original_name = "linear_668.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.589469") @@ -700,6 +764,7 @@ class Program_weight_tensor_parameter_63: class Program_weight_tensor_parameter_64: name = "parameter_64" + original_name = "linear_667.b_0" shape = [768] dtype = "float32" min_val = float("-0.560731") @@ -711,6 +776,7 @@ class Program_weight_tensor_parameter_64: class Program_weight_tensor_parameter_65: name = "parameter_65" + original_name = "linear_667.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.572933") @@ -722,6 +788,7 @@ class Program_weight_tensor_parameter_65: class Program_weight_tensor_parameter_66: name = "parameter_66" + original_name = "layer_norm_227.b_0" shape = [768] dtype = "float32" min_val = float("-0.813942") @@ -733,6 +800,7 @@ class Program_weight_tensor_parameter_66: class Program_weight_tensor_parameter_67: name = "parameter_67" + original_name = "layer_norm_227.w_0" shape = [768] dtype = "float32" min_val = float("0.81998") @@ -744,6 +812,7 @@ class Program_weight_tensor_parameter_67: class Program_weight_tensor_parameter_68: name = "parameter_68" + original_name = "layer_norm_226.b_0" shape = [768] dtype = "float32" min_val = float("-1.08923") @@ -755,6 +824,7 @@ class Program_weight_tensor_parameter_68: class Program_weight_tensor_parameter_69: name = "parameter_69" + original_name = "layer_norm_226.w_0" shape = [768] dtype = "float32" min_val = float("0.529043") @@ -766,6 +836,7 @@ class Program_weight_tensor_parameter_69: class Program_weight_tensor_parameter_70: name = "parameter_70" + original_name = "linear_666.b_0" shape = [768] dtype = "float32" min_val = float("-0.246703") @@ -777,6 +848,7 @@ class Program_weight_tensor_parameter_70: class Program_weight_tensor_parameter_71: name = "parameter_71" + original_name = "linear_666.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-2.91331") @@ -788,6 +860,7 @@ class Program_weight_tensor_parameter_71: class Program_weight_tensor_parameter_72: name = "parameter_72" + original_name = "linear_665.b_0" shape = [3072] dtype = "float32" min_val = float("-0.419512") @@ -799,6 +872,7 @@ class Program_weight_tensor_parameter_72: class Program_weight_tensor_parameter_73: name = "parameter_73" + original_name = "linear_665.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.423732") @@ -810,6 +884,7 @@ class Program_weight_tensor_parameter_73: class Program_weight_tensor_parameter_74: name = "parameter_74" + original_name = "linear_664.b_0" shape = [768] dtype = "float32" min_val = float("-0.31092") @@ -821,6 +896,7 @@ class Program_weight_tensor_parameter_74: class Program_weight_tensor_parameter_75: name = "parameter_75" + original_name = "linear_664.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.355739") @@ -832,6 +908,7 @@ class Program_weight_tensor_parameter_75: class Program_weight_tensor_parameter_76: name = "parameter_76" + original_name = "linear_663.b_0" shape = [768] dtype = "float32" min_val = float("-0.104585") @@ -843,6 +920,7 @@ class Program_weight_tensor_parameter_76: class Program_weight_tensor_parameter_77: name = "parameter_77" + original_name = "linear_663.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.235601") @@ -854,6 +932,7 @@ class Program_weight_tensor_parameter_77: class Program_weight_tensor_parameter_78: name = "parameter_78" + original_name = "linear_662.b_0" shape = [768] dtype = "float32" min_val = float("-0.00219675") @@ -865,6 +944,7 @@ class Program_weight_tensor_parameter_78: class Program_weight_tensor_parameter_79: name = "parameter_79" + original_name = "linear_662.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.553803") @@ -876,6 +956,7 @@ class Program_weight_tensor_parameter_79: class Program_weight_tensor_parameter_80: name = "parameter_80" + original_name = "linear_661.b_0" shape = [768] dtype = "float32" min_val = float("-0.704081") @@ -887,6 +968,7 @@ class Program_weight_tensor_parameter_80: class Program_weight_tensor_parameter_81: name = "parameter_81" + original_name = "linear_661.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.337411") @@ -898,6 +980,7 @@ class Program_weight_tensor_parameter_81: class Program_weight_tensor_parameter_82: name = "parameter_82" + original_name = "layer_norm_225.b_0" shape = [768] dtype = "float32" min_val = float("-1.18989") @@ -909,6 +992,7 @@ class Program_weight_tensor_parameter_82: class Program_weight_tensor_parameter_83: name = "parameter_83" + original_name = "layer_norm_225.w_0" shape = [768] dtype = "float32" min_val = float("0.751453") @@ -920,6 +1004,7 @@ class Program_weight_tensor_parameter_83: class Program_weight_tensor_parameter_84: name = "parameter_84" + original_name = "layer_norm_224.b_0" shape = [768] dtype = "float32" min_val = float("-2.65911") @@ -931,6 +1016,7 @@ class Program_weight_tensor_parameter_84: class Program_weight_tensor_parameter_85: name = "parameter_85" + original_name = "layer_norm_224.w_0" shape = [768] dtype = "float32" min_val = float("0.127841") @@ -942,6 +1028,7 @@ class Program_weight_tensor_parameter_85: class Program_weight_tensor_parameter_86: name = "parameter_86" + original_name = "linear_660.b_0" shape = [768] dtype = "float32" min_val = float("-0.249176") @@ -953,6 +1040,7 @@ class Program_weight_tensor_parameter_86: class Program_weight_tensor_parameter_87: name = "parameter_87" + original_name = "linear_660.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-6.36231") @@ -964,6 +1052,7 @@ class Program_weight_tensor_parameter_87: class Program_weight_tensor_parameter_88: name = "parameter_88" + original_name = "linear_659.b_0" shape = [3072] dtype = "float32" min_val = float("-0.471154") @@ -975,6 +1064,7 @@ class Program_weight_tensor_parameter_88: class Program_weight_tensor_parameter_89: name = "parameter_89" + original_name = "linear_659.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.957504") @@ -986,6 +1076,7 @@ class Program_weight_tensor_parameter_89: class Program_weight_tensor_parameter_90: name = "parameter_90" + original_name = "linear_658.b_0" shape = [768] dtype = "float32" min_val = float("-0.241205") @@ -997,6 +1088,7 @@ class Program_weight_tensor_parameter_90: class Program_weight_tensor_parameter_91: name = "parameter_91" + original_name = "linear_658.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.714421") @@ -1008,6 +1100,7 @@ class Program_weight_tensor_parameter_91: class Program_weight_tensor_parameter_92: name = "parameter_92" + original_name = "linear_657.b_0" shape = [768] dtype = "float32" min_val = float("-0.422264") @@ -1019,6 +1112,7 @@ class Program_weight_tensor_parameter_92: class Program_weight_tensor_parameter_93: name = "parameter_93" + original_name = "linear_657.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.284129") @@ -1030,6 +1124,7 @@ class Program_weight_tensor_parameter_93: class Program_weight_tensor_parameter_94: name = "parameter_94" + original_name = "linear_656.b_0" shape = [768] dtype = "float32" min_val = float("-0.000517314") @@ -1041,6 +1136,7 @@ class Program_weight_tensor_parameter_94: class Program_weight_tensor_parameter_95: name = "parameter_95" + original_name = "linear_656.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.325635") @@ -1052,6 +1148,7 @@ class Program_weight_tensor_parameter_95: class Program_weight_tensor_parameter_96: name = "parameter_96" + original_name = "linear_655.b_0" shape = [768] dtype = "float32" min_val = float("-1.0396") @@ -1063,6 +1160,7 @@ class Program_weight_tensor_parameter_96: class Program_weight_tensor_parameter_97: name = "parameter_97" + original_name = "linear_655.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.42285") @@ -1074,6 +1172,7 @@ class Program_weight_tensor_parameter_97: class Program_weight_tensor_parameter_98: name = "parameter_98" + original_name = "layer_norm_223.b_0" shape = [768] dtype = "float32" min_val = float("-3.10219") @@ -1085,6 +1184,7 @@ class Program_weight_tensor_parameter_98: class Program_weight_tensor_parameter_99: name = "parameter_99" + original_name = "layer_norm_223.w_0" shape = [768] dtype = "float32" min_val = float("0.079") @@ -1096,6 +1196,7 @@ class Program_weight_tensor_parameter_99: class Program_weight_tensor_parameter_100: name = "parameter_100" + original_name = "embedding_26.w_0" shape = [16, 768] dtype = "float32" min_val = float("-0.0339038") @@ -1107,6 +1208,7 @@ class Program_weight_tensor_parameter_100: class Program_weight_tensor_parameter_101: name = "parameter_101" + original_name = "embedding_25.w_0" shape = [4, 768] dtype = "float32" min_val = float("-0.0726872") @@ -1118,6 +1220,7 @@ class Program_weight_tensor_parameter_101: class Program_weight_tensor_parameter_102: name = "parameter_102" + original_name = "embedding_24.w_0" shape = [2048, 768] dtype = "float32" min_val = float("-0.85655") @@ -1129,6 +1232,7 @@ class Program_weight_tensor_parameter_102: class Program_weight_tensor_parameter_103: name = "parameter_103" + original_name = "embedding_23.w_0" shape = [40000, 768] dtype = "float32" min_val = float("-1.14225") diff --git a/paddle_samples/PaddleNLP/ernie-3.0-micro-zh/input_meta.py b/paddle_samples/PaddleNLP/ernie-3.0-micro-zh/input_meta.py index 97eb8a799..a4123e481 100644 --- a/paddle_samples/PaddleNLP/ernie-3.0-micro-zh/input_meta.py +++ b/paddle_samples/PaddleNLP/ernie-3.0-micro-zh/input_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_data_0: name = "data_0" + original_name = "input_ids" shape = [1, 11] dtype = "int64" data = [1, 811, 1257, 175, 29, 502, 130, 706, 3619, 12046, 2] @@ -7,6 +8,7 @@ class Program_weight_tensor_data_0: class Program_weight_tensor_data_1: name = "data_1" + original_name = "token_type_ids" shape = [1, 11] dtype = "int64" data = [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0] diff --git a/paddle_samples/PaddleNLP/ernie-3.0-micro-zh/model.py b/paddle_samples/PaddleNLP/ernie-3.0-micro-zh/model.py index 0cff6c45d..4b23cf3f8 100644 --- a/paddle_samples/PaddleNLP/ernie-3.0-micro-zh/model.py +++ b/paddle_samples/PaddleNLP/ernie-3.0-micro-zh/model.py @@ -138,10 +138,11 @@ def forward( # pd_op.embedding: (1x11x384xf32) <- (1x11xi64, 2048x384xf32) embedding_1 = paddle._C_ops.embedding(subtract_0, parameter_70, -1, False) - del parameter_70 + del parameter_70, subtract_0 # pd_op.add: (1x11x384xf32) <- (1x11x384xf32, 1x11x384xf32) add_0 = paddle._C_ops.add(embedding_0, embedding_1) + del embedding_0, embedding_1 # pd_op.embedding: (1x11x384xf32) <- (1x11xi64, 4x384xf32) embedding_2 = paddle._C_ops.embedding(data_1, parameter_69, -1, False) @@ -149,6 +150,7 @@ def forward( # pd_op.add: (1x11x384xf32) <- (1x11x384xf32, 1x11x384xf32) add_1 = paddle._C_ops.add(add_0, embedding_2) + del add_0, embedding_2 # pd_op.full: (1xf32) <- () full_4 = paddle._C_ops.full( @@ -161,10 +163,11 @@ def forward( # pd_op.embedding: (1x11x384xf32) <- (1x11xi64, 16x384xf32) embedding_3 = paddle._C_ops.embedding(scale_1, parameter_68, -1, False) - del parameter_68 + del parameter_68, scale_1 # pd_op.add: (1x11x384xf32) <- (1x11x384xf32, 1x11x384xf32) add_2 = paddle._C_ops.add(add_1, embedding_3) + del add_1, embedding_3 # pd_op.layer_norm: (1x11x384xf32, 1x11xf32, 1x11xf32) <- (1x11x384xf32, 384xf32, 384xf32) layer_norm_0, layer_norm_1, layer_norm_2 = (lambda x, f: f(x))( @@ -173,53 +176,17 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_66, parameter_67 + del add_2, parameter_66, parameter_67 # pd_op.full: (1xf32) <- () full_5 = paddle._C_ops.full( [1], float("0.1"), paddle.float32, paddle.core.CPUPlace() ) - # pd_op.assign: (1xf32) <- (1xf32) - assign_0 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_1 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_2 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_3 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_4 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_5 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_6 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_7 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_8 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_9 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_10 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_11 = full_5 - # pd_op.dropout: (1x11x384xf32, 1x11x384xui8) <- (1x11x384xf32, None, 1xf32) dropout_0, dropout_1 = (lambda x, f: f(x))( paddle._C_ops.dropout( - layer_norm_0, None, full_5, False, "upscale_in_train", 0, False + layer_norm_0, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -231,13 +198,14 @@ def forward( # pd_op.add: (1x11x384xf32) <- (1x11x384xf32, 384xf32) add_3 = paddle._C_ops.add(matmul_0, parameter_64) - del parameter_64 + del matmul_0, parameter_64 # pd_op.full_int_array: (4xi64) <- () full_int_array_1 = [0, 0, 12, 32] # pd_op.reshape: (1x11x12x32xf32) <- (1x11x384xf32, 4xi64) reshape_0 = paddle._C_ops.reshape(add_3, full_int_array_1) + del add_3 # pd_op.transpose: (1x12x11x32xf32) <- (1x11x12x32xf32) transpose_0 = paddle._C_ops.transpose(reshape_0, [0, 2, 1, 3]) @@ -249,7 +217,7 @@ def forward( # pd_op.add: (1x11x384xf32) <- (1x11x384xf32, 384xf32) add_4 = paddle._C_ops.add(matmul_1, parameter_62) - del parameter_62 + del matmul_1, parameter_62 # pd_op.matmul: (1x11x384xf32) <- (1x11x384xf32, 384x384xf32) matmul_2 = paddle._C_ops.matmul(dropout_0, parameter_61, False, False) @@ -257,10 +225,11 @@ def forward( # pd_op.add: (1x11x384xf32) <- (1x11x384xf32, 384xf32) add_5 = paddle._C_ops.add(matmul_2, parameter_60) - del parameter_60 + del matmul_2, parameter_60 # pd_op.reshape: (1x11x12x32xf32) <- (1x11x384xf32, 4xi64) reshape_1 = paddle._C_ops.reshape(add_4, full_int_array_1) + del add_4 # pd_op.transpose: (1x12x11x32xf32) <- (1x11x12x32xf32) transpose_1 = paddle._C_ops.transpose(reshape_1, [0, 2, 1, 3]) @@ -268,6 +237,7 @@ def forward( # pd_op.reshape: (1x11x12x32xf32) <- (1x11x384xf32, 4xi64) reshape_2 = paddle._C_ops.reshape(add_5, full_int_array_1) + del add_5 # pd_op.transpose: (1x12x11x32xf32) <- (1x11x12x32xf32) transpose_2 = paddle._C_ops.transpose(reshape_2, [0, 2, 1, 3]) @@ -278,24 +248,17 @@ def forward( [1], float("0.176777"), paddle.float32, paddle.core.CPUPlace() ) - # pd_op.assign: (1xf32) <- (1xf32) - assign_12 = full_6 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_13 = full_6 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_14 = full_6 - # pd_op.scale: (1x12x11x32xf32) <- (1x12x11x32xf32, 1xf32) scale_2 = paddle._C_ops.scale(transpose_0, full_6, float("0"), True) del transpose_0 # pd_op.matmul: (1x12x11x11xf32) <- (1x12x11x32xf32, 1x12x11x32xf32) matmul_3 = paddle._C_ops.matmul(scale_2, transpose_1, False, True) + del scale_2, transpose_1 # pd_op.add: (1x12x11x11xf32) <- (1x12x11x11xf32, 1x1x1x11xf32) add_6 = paddle._C_ops.add(matmul_3, unsqueeze_0) + del matmul_3 # pd_op.softmax: (1x12x11x11xf32) <- (1x12x11x11xf32) softmax_0 = paddle._C_ops.softmax(add_6, -1) @@ -304,13 +267,15 @@ def forward( # pd_op.dropout: (1x12x11x11xf32, 1x12x11x11xui8) <- (1x12x11x11xf32, None, 1xf32) dropout_2, dropout_3 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_0, None, full_5, False, "upscale_in_train", 0, False + softmax_0, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_0 # pd_op.matmul: (1x12x11x32xf32) <- (1x12x11x11xf32, 1x12x11x32xf32) matmul_4 = paddle._C_ops.matmul(dropout_2, transpose_2, False, False) + del dropout_2, transpose_2 # pd_op.transpose: (1x11x12x32xf32) <- (1x12x11x32xf32) transpose_3 = paddle._C_ops.transpose(matmul_4, [0, 2, 1, 3]) @@ -321,19 +286,20 @@ def forward( # pd_op.reshape: (1x11x384xf32) <- (1x11x12x32xf32, 3xi64) reshape_3 = paddle._C_ops.reshape(transpose_3, full_int_array_2) + del transpose_3 # pd_op.matmul: (1x11x384xf32) <- (1x11x384xf32, 384x384xf32) matmul_5 = paddle._C_ops.matmul(reshape_3, parameter_59, False, False) - del parameter_59 + del parameter_59, reshape_3 # pd_op.add: (1x11x384xf32) <- (1x11x384xf32, 384xf32) add_7 = paddle._C_ops.add(matmul_5, parameter_58) - del parameter_58 + del matmul_5, parameter_58 # pd_op.dropout: (1x11x384xf32, 1x11x384xui8) <- (1x11x384xf32, None, 1xf32) dropout_4, dropout_5 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_7, None, full_5, False, "upscale_in_train", 0, False + add_7, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -341,6 +307,7 @@ def forward( # pd_op.add: (1x11x384xf32) <- (1x11x384xf32, 1x11x384xf32) add_8 = paddle._C_ops.add(dropout_0, dropout_4) + del dropout_0, dropout_4 # pd_op.layer_norm: (1x11x384xf32, 1x11xf32, 1x11xf32) <- (1x11x384xf32, 384xf32, 384xf32) layer_norm_3, layer_norm_4, layer_norm_5 = (lambda x, f: f(x))( @@ -349,7 +316,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_52, parameter_53 + del add_8, parameter_52, parameter_53 # pd_op.matmul: (1x11x1536xf32) <- (1x11x384xf32, 384x1536xf32) matmul_6 = paddle._C_ops.matmul(layer_norm_3, parameter_57, False, False) @@ -357,23 +324,24 @@ def forward( # pd_op.add: (1x11x1536xf32) <- (1x11x1536xf32, 1536xf32) add_9 = paddle._C_ops.add(matmul_6, parameter_56) - del parameter_56 + del matmul_6, parameter_56 # pd_op.gelu: (1x11x1536xf32) <- (1x11x1536xf32) gelu_0 = paddle._C_ops.gelu(add_9, False) + del add_9 # pd_op.matmul: (1x11x384xf32) <- (1x11x1536xf32, 1536x384xf32) matmul_7 = paddle._C_ops.matmul(gelu_0, parameter_55, False, False) - del parameter_55 + del gelu_0, parameter_55 # pd_op.add: (1x11x384xf32) <- (1x11x384xf32, 384xf32) add_10 = paddle._C_ops.add(matmul_7, parameter_54) - del parameter_54 + del matmul_7, parameter_54 # pd_op.dropout: (1x11x384xf32, 1x11x384xui8) <- (1x11x384xf32, None, 1xf32) dropout_6, dropout_7 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_10, None, full_5, False, "upscale_in_train", 0, False + add_10, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -381,6 +349,7 @@ def forward( # pd_op.add: (1x11x384xf32) <- (1x11x384xf32, 1x11x384xf32) add_11 = paddle._C_ops.add(layer_norm_3, dropout_6) + del dropout_6, layer_norm_3 # pd_op.layer_norm: (1x11x384xf32, 1x11xf32, 1x11xf32) <- (1x11x384xf32, 384xf32, 384xf32) layer_norm_6, layer_norm_7, layer_norm_8 = (lambda x, f: f(x))( @@ -389,7 +358,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_50, parameter_51 + del add_11, parameter_50, parameter_51 # pd_op.matmul: (1x11x384xf32) <- (1x11x384xf32, 384x384xf32) matmul_8 = paddle._C_ops.matmul(layer_norm_6, parameter_49, False, False) @@ -397,10 +366,11 @@ def forward( # pd_op.add: (1x11x384xf32) <- (1x11x384xf32, 384xf32) add_12 = paddle._C_ops.add(matmul_8, parameter_48) - del parameter_48 + del matmul_8, parameter_48 # pd_op.reshape: (1x11x12x32xf32) <- (1x11x384xf32, 4xi64) reshape_4 = paddle._C_ops.reshape(add_12, full_int_array_1) + del add_12 # pd_op.transpose: (1x12x11x32xf32) <- (1x11x12x32xf32) transpose_4 = paddle._C_ops.transpose(reshape_4, [0, 2, 1, 3]) @@ -412,7 +382,7 @@ def forward( # pd_op.add: (1x11x384xf32) <- (1x11x384xf32, 384xf32) add_13 = paddle._C_ops.add(matmul_9, parameter_46) - del parameter_46 + del matmul_9, parameter_46 # pd_op.matmul: (1x11x384xf32) <- (1x11x384xf32, 384x384xf32) matmul_10 = paddle._C_ops.matmul(layer_norm_6, parameter_45, False, False) @@ -420,10 +390,11 @@ def forward( # pd_op.add: (1x11x384xf32) <- (1x11x384xf32, 384xf32) add_14 = paddle._C_ops.add(matmul_10, parameter_44) - del parameter_44 + del matmul_10, parameter_44 # pd_op.reshape: (1x11x12x32xf32) <- (1x11x384xf32, 4xi64) reshape_5 = paddle._C_ops.reshape(add_13, full_int_array_1) + del add_13 # pd_op.transpose: (1x12x11x32xf32) <- (1x11x12x32xf32) transpose_5 = paddle._C_ops.transpose(reshape_5, [0, 2, 1, 3]) @@ -431,6 +402,7 @@ def forward( # pd_op.reshape: (1x11x12x32xf32) <- (1x11x384xf32, 4xi64) reshape_6 = paddle._C_ops.reshape(add_14, full_int_array_1) + del add_14 # pd_op.transpose: (1x12x11x32xf32) <- (1x11x12x32xf32) transpose_6 = paddle._C_ops.transpose(reshape_6, [0, 2, 1, 3]) @@ -442,9 +414,11 @@ def forward( # pd_op.matmul: (1x12x11x11xf32) <- (1x12x11x32xf32, 1x12x11x32xf32) matmul_11 = paddle._C_ops.matmul(scale_3, transpose_5, False, True) + del scale_3, transpose_5 # pd_op.add: (1x12x11x11xf32) <- (1x12x11x11xf32, 1x1x1x11xf32) add_15 = paddle._C_ops.add(matmul_11, unsqueeze_0) + del matmul_11 # pd_op.softmax: (1x12x11x11xf32) <- (1x12x11x11xf32) softmax_1 = paddle._C_ops.softmax(add_15, -1) @@ -453,13 +427,15 @@ def forward( # pd_op.dropout: (1x12x11x11xf32, 1x12x11x11xui8) <- (1x12x11x11xf32, None, 1xf32) dropout_8, dropout_9 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_1, None, full_5, False, "upscale_in_train", 0, False + softmax_1, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_1 # pd_op.matmul: (1x12x11x32xf32) <- (1x12x11x11xf32, 1x12x11x32xf32) matmul_12 = paddle._C_ops.matmul(dropout_8, transpose_6, False, False) + del dropout_8, transpose_6 # pd_op.transpose: (1x11x12x32xf32) <- (1x12x11x32xf32) transpose_7 = paddle._C_ops.transpose(matmul_12, [0, 2, 1, 3]) @@ -467,19 +443,20 @@ def forward( # pd_op.reshape: (1x11x384xf32) <- (1x11x12x32xf32, 3xi64) reshape_7 = paddle._C_ops.reshape(transpose_7, full_int_array_2) + del transpose_7 # pd_op.matmul: (1x11x384xf32) <- (1x11x384xf32, 384x384xf32) matmul_13 = paddle._C_ops.matmul(reshape_7, parameter_43, False, False) - del parameter_43 + del parameter_43, reshape_7 # pd_op.add: (1x11x384xf32) <- (1x11x384xf32, 384xf32) add_16 = paddle._C_ops.add(matmul_13, parameter_42) - del parameter_42 + del matmul_13, parameter_42 # pd_op.dropout: (1x11x384xf32, 1x11x384xui8) <- (1x11x384xf32, None, 1xf32) dropout_10, dropout_11 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_16, None, full_5, False, "upscale_in_train", 0, False + add_16, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -487,6 +464,7 @@ def forward( # pd_op.add: (1x11x384xf32) <- (1x11x384xf32, 1x11x384xf32) add_17 = paddle._C_ops.add(layer_norm_6, dropout_10) + del dropout_10, layer_norm_6 # pd_op.layer_norm: (1x11x384xf32, 1x11xf32, 1x11xf32) <- (1x11x384xf32, 384xf32, 384xf32) layer_norm_9, layer_norm_10, layer_norm_11 = (lambda x, f: f(x))( @@ -495,7 +473,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_36, parameter_37 + del add_17, parameter_36, parameter_37 # pd_op.matmul: (1x11x1536xf32) <- (1x11x384xf32, 384x1536xf32) matmul_14 = paddle._C_ops.matmul(layer_norm_9, parameter_41, False, False) @@ -503,23 +481,24 @@ def forward( # pd_op.add: (1x11x1536xf32) <- (1x11x1536xf32, 1536xf32) add_18 = paddle._C_ops.add(matmul_14, parameter_40) - del parameter_40 + del matmul_14, parameter_40 # pd_op.gelu: (1x11x1536xf32) <- (1x11x1536xf32) gelu_1 = paddle._C_ops.gelu(add_18, False) + del add_18 # pd_op.matmul: (1x11x384xf32) <- (1x11x1536xf32, 1536x384xf32) matmul_15 = paddle._C_ops.matmul(gelu_1, parameter_39, False, False) - del parameter_39 + del gelu_1, parameter_39 # pd_op.add: (1x11x384xf32) <- (1x11x384xf32, 384xf32) add_19 = paddle._C_ops.add(matmul_15, parameter_38) - del parameter_38 + del matmul_15, parameter_38 # pd_op.dropout: (1x11x384xf32, 1x11x384xui8) <- (1x11x384xf32, None, 1xf32) dropout_12, dropout_13 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_19, None, full_5, False, "upscale_in_train", 0, False + add_19, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -527,6 +506,7 @@ def forward( # pd_op.add: (1x11x384xf32) <- (1x11x384xf32, 1x11x384xf32) add_20 = paddle._C_ops.add(layer_norm_9, dropout_12) + del dropout_12, layer_norm_9 # pd_op.layer_norm: (1x11x384xf32, 1x11xf32, 1x11xf32) <- (1x11x384xf32, 384xf32, 384xf32) layer_norm_12, layer_norm_13, layer_norm_14 = (lambda x, f: f(x))( @@ -535,7 +515,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_34, parameter_35 + del add_20, parameter_34, parameter_35 # pd_op.matmul: (1x11x384xf32) <- (1x11x384xf32, 384x384xf32) matmul_16 = paddle._C_ops.matmul(layer_norm_12, parameter_33, False, False) @@ -543,10 +523,11 @@ def forward( # pd_op.add: (1x11x384xf32) <- (1x11x384xf32, 384xf32) add_21 = paddle._C_ops.add(matmul_16, parameter_32) - del parameter_32 + del matmul_16, parameter_32 # pd_op.reshape: (1x11x12x32xf32) <- (1x11x384xf32, 4xi64) reshape_8 = paddle._C_ops.reshape(add_21, full_int_array_1) + del add_21 # pd_op.transpose: (1x12x11x32xf32) <- (1x11x12x32xf32) transpose_8 = paddle._C_ops.transpose(reshape_8, [0, 2, 1, 3]) @@ -558,7 +539,7 @@ def forward( # pd_op.add: (1x11x384xf32) <- (1x11x384xf32, 384xf32) add_22 = paddle._C_ops.add(matmul_17, parameter_30) - del parameter_30 + del matmul_17, parameter_30 # pd_op.matmul: (1x11x384xf32) <- (1x11x384xf32, 384x384xf32) matmul_18 = paddle._C_ops.matmul(layer_norm_12, parameter_29, False, False) @@ -566,10 +547,11 @@ def forward( # pd_op.add: (1x11x384xf32) <- (1x11x384xf32, 384xf32) add_23 = paddle._C_ops.add(matmul_18, parameter_28) - del parameter_28 + del matmul_18, parameter_28 # pd_op.reshape: (1x11x12x32xf32) <- (1x11x384xf32, 4xi64) reshape_9 = paddle._C_ops.reshape(add_22, full_int_array_1) + del add_22 # pd_op.transpose: (1x12x11x32xf32) <- (1x11x12x32xf32) transpose_9 = paddle._C_ops.transpose(reshape_9, [0, 2, 1, 3]) @@ -577,6 +559,7 @@ def forward( # pd_op.reshape: (1x11x12x32xf32) <- (1x11x384xf32, 4xi64) reshape_10 = paddle._C_ops.reshape(add_23, full_int_array_1) + del add_23 # pd_op.transpose: (1x12x11x32xf32) <- (1x11x12x32xf32) transpose_10 = paddle._C_ops.transpose(reshape_10, [0, 2, 1, 3]) @@ -588,9 +571,11 @@ def forward( # pd_op.matmul: (1x12x11x11xf32) <- (1x12x11x32xf32, 1x12x11x32xf32) matmul_19 = paddle._C_ops.matmul(scale_4, transpose_9, False, True) + del scale_4, transpose_9 # pd_op.add: (1x12x11x11xf32) <- (1x12x11x11xf32, 1x1x1x11xf32) add_24 = paddle._C_ops.add(matmul_19, unsqueeze_0) + del matmul_19 # pd_op.softmax: (1x12x11x11xf32) <- (1x12x11x11xf32) softmax_2 = paddle._C_ops.softmax(add_24, -1) @@ -599,13 +584,15 @@ def forward( # pd_op.dropout: (1x12x11x11xf32, 1x12x11x11xui8) <- (1x12x11x11xf32, None, 1xf32) dropout_14, dropout_15 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_2, None, full_5, False, "upscale_in_train", 0, False + softmax_2, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_2 # pd_op.matmul: (1x12x11x32xf32) <- (1x12x11x11xf32, 1x12x11x32xf32) matmul_20 = paddle._C_ops.matmul(dropout_14, transpose_10, False, False) + del dropout_14, transpose_10 # pd_op.transpose: (1x11x12x32xf32) <- (1x12x11x32xf32) transpose_11 = paddle._C_ops.transpose(matmul_20, [0, 2, 1, 3]) @@ -613,19 +600,20 @@ def forward( # pd_op.reshape: (1x11x384xf32) <- (1x11x12x32xf32, 3xi64) reshape_11 = paddle._C_ops.reshape(transpose_11, full_int_array_2) + del transpose_11 # pd_op.matmul: (1x11x384xf32) <- (1x11x384xf32, 384x384xf32) matmul_21 = paddle._C_ops.matmul(reshape_11, parameter_27, False, False) - del parameter_27 + del parameter_27, reshape_11 # pd_op.add: (1x11x384xf32) <- (1x11x384xf32, 384xf32) add_25 = paddle._C_ops.add(matmul_21, parameter_26) - del parameter_26 + del matmul_21, parameter_26 # pd_op.dropout: (1x11x384xf32, 1x11x384xui8) <- (1x11x384xf32, None, 1xf32) dropout_16, dropout_17 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_25, None, full_5, False, "upscale_in_train", 0, False + add_25, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -633,6 +621,7 @@ def forward( # pd_op.add: (1x11x384xf32) <- (1x11x384xf32, 1x11x384xf32) add_26 = paddle._C_ops.add(layer_norm_12, dropout_16) + del dropout_16, layer_norm_12 # pd_op.layer_norm: (1x11x384xf32, 1x11xf32, 1x11xf32) <- (1x11x384xf32, 384xf32, 384xf32) layer_norm_15, layer_norm_16, layer_norm_17 = (lambda x, f: f(x))( @@ -641,7 +630,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_20, parameter_21 + del add_26, parameter_20, parameter_21 # pd_op.matmul: (1x11x1536xf32) <- (1x11x384xf32, 384x1536xf32) matmul_22 = paddle._C_ops.matmul(layer_norm_15, parameter_25, False, False) @@ -649,23 +638,24 @@ def forward( # pd_op.add: (1x11x1536xf32) <- (1x11x1536xf32, 1536xf32) add_27 = paddle._C_ops.add(matmul_22, parameter_24) - del parameter_24 + del matmul_22, parameter_24 # pd_op.gelu: (1x11x1536xf32) <- (1x11x1536xf32) gelu_2 = paddle._C_ops.gelu(add_27, False) + del add_27 # pd_op.matmul: (1x11x384xf32) <- (1x11x1536xf32, 1536x384xf32) matmul_23 = paddle._C_ops.matmul(gelu_2, parameter_23, False, False) - del parameter_23 + del gelu_2, parameter_23 # pd_op.add: (1x11x384xf32) <- (1x11x384xf32, 384xf32) add_28 = paddle._C_ops.add(matmul_23, parameter_22) - del parameter_22 + del matmul_23, parameter_22 # pd_op.dropout: (1x11x384xf32, 1x11x384xui8) <- (1x11x384xf32, None, 1xf32) dropout_18, dropout_19 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_28, None, full_5, False, "upscale_in_train", 0, False + add_28, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -673,6 +663,7 @@ def forward( # pd_op.add: (1x11x384xf32) <- (1x11x384xf32, 1x11x384xf32) add_29 = paddle._C_ops.add(layer_norm_15, dropout_18) + del dropout_18, layer_norm_15 # pd_op.layer_norm: (1x11x384xf32, 1x11xf32, 1x11xf32) <- (1x11x384xf32, 384xf32, 384xf32) layer_norm_18, layer_norm_19, layer_norm_20 = (lambda x, f: f(x))( @@ -681,7 +672,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_18, parameter_19 + del add_29, parameter_18, parameter_19 # pd_op.matmul: (1x11x384xf32) <- (1x11x384xf32, 384x384xf32) matmul_24 = paddle._C_ops.matmul(layer_norm_18, parameter_17, False, False) @@ -689,10 +680,11 @@ def forward( # pd_op.add: (1x11x384xf32) <- (1x11x384xf32, 384xf32) add_30 = paddle._C_ops.add(matmul_24, parameter_16) - del parameter_16 + del matmul_24, parameter_16 # pd_op.reshape: (1x11x12x32xf32) <- (1x11x384xf32, 4xi64) reshape_12 = paddle._C_ops.reshape(add_30, full_int_array_1) + del add_30 # pd_op.transpose: (1x12x11x32xf32) <- (1x11x12x32xf32) transpose_12 = paddle._C_ops.transpose(reshape_12, [0, 2, 1, 3]) @@ -704,7 +696,7 @@ def forward( # pd_op.add: (1x11x384xf32) <- (1x11x384xf32, 384xf32) add_31 = paddle._C_ops.add(matmul_25, parameter_14) - del parameter_14 + del matmul_25, parameter_14 # pd_op.matmul: (1x11x384xf32) <- (1x11x384xf32, 384x384xf32) matmul_26 = paddle._C_ops.matmul(layer_norm_18, parameter_13, False, False) @@ -712,10 +704,11 @@ def forward( # pd_op.add: (1x11x384xf32) <- (1x11x384xf32, 384xf32) add_32 = paddle._C_ops.add(matmul_26, parameter_12) - del parameter_12 + del matmul_26, parameter_12 # pd_op.reshape: (1x11x12x32xf32) <- (1x11x384xf32, 4xi64) reshape_13 = paddle._C_ops.reshape(add_31, full_int_array_1) + del add_31 # pd_op.transpose: (1x12x11x32xf32) <- (1x11x12x32xf32) transpose_13 = paddle._C_ops.transpose(reshape_13, [0, 2, 1, 3]) @@ -723,7 +716,7 @@ def forward( # pd_op.reshape: (1x11x12x32xf32) <- (1x11x384xf32, 4xi64) reshape_14 = paddle._C_ops.reshape(add_32, full_int_array_1) - del full_int_array_1 + del add_32, full_int_array_1 # pd_op.transpose: (1x12x11x32xf32) <- (1x11x12x32xf32) transpose_14 = paddle._C_ops.transpose(reshape_14, [0, 2, 1, 3]) @@ -731,13 +724,15 @@ def forward( # pd_op.scale: (1x12x11x32xf32) <- (1x12x11x32xf32, 1xf32) scale_5 = paddle._C_ops.scale(transpose_12, full_6, float("0"), True) - del transpose_12 + del full_6, transpose_12 # pd_op.matmul: (1x12x11x11xf32) <- (1x12x11x32xf32, 1x12x11x32xf32) matmul_27 = paddle._C_ops.matmul(scale_5, transpose_13, False, True) + del scale_5, transpose_13 # pd_op.add: (1x12x11x11xf32) <- (1x12x11x11xf32, 1x1x1x11xf32) add_33 = paddle._C_ops.add(matmul_27, unsqueeze_0) + del matmul_27, unsqueeze_0 # pd_op.softmax: (1x12x11x11xf32) <- (1x12x11x11xf32) softmax_3 = paddle._C_ops.softmax(add_33, -1) @@ -746,13 +741,15 @@ def forward( # pd_op.dropout: (1x12x11x11xf32, 1x12x11x11xui8) <- (1x12x11x11xf32, None, 1xf32) dropout_20, dropout_21 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_3, None, full_5, False, "upscale_in_train", 0, False + softmax_3, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_3 # pd_op.matmul: (1x12x11x32xf32) <- (1x12x11x11xf32, 1x12x11x32xf32) matmul_28 = paddle._C_ops.matmul(dropout_20, transpose_14, False, False) + del dropout_20, transpose_14 # pd_op.transpose: (1x11x12x32xf32) <- (1x12x11x32xf32) transpose_15 = paddle._C_ops.transpose(matmul_28, [0, 2, 1, 3]) @@ -760,20 +757,20 @@ def forward( # pd_op.reshape: (1x11x384xf32) <- (1x11x12x32xf32, 3xi64) reshape_15 = paddle._C_ops.reshape(transpose_15, full_int_array_2) - del full_int_array_2 + del full_int_array_2, transpose_15 # pd_op.matmul: (1x11x384xf32) <- (1x11x384xf32, 384x384xf32) matmul_29 = paddle._C_ops.matmul(reshape_15, parameter_11, False, False) - del parameter_11 + del parameter_11, reshape_15 # pd_op.add: (1x11x384xf32) <- (1x11x384xf32, 384xf32) add_34 = paddle._C_ops.add(matmul_29, parameter_10) - del parameter_10 + del matmul_29, parameter_10 # pd_op.dropout: (1x11x384xf32, 1x11x384xui8) <- (1x11x384xf32, None, 1xf32) dropout_22, dropout_23 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_34, None, full_5, False, "upscale_in_train", 0, False + add_34, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -781,6 +778,7 @@ def forward( # pd_op.add: (1x11x384xf32) <- (1x11x384xf32, 1x11x384xf32) add_35 = paddle._C_ops.add(layer_norm_18, dropout_22) + del dropout_22, layer_norm_18 # pd_op.layer_norm: (1x11x384xf32, 1x11xf32, 1x11xf32) <- (1x11x384xf32, 384xf32, 384xf32) layer_norm_21, layer_norm_22, layer_norm_23 = (lambda x, f: f(x))( @@ -789,7 +787,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_4, parameter_5 + del add_35, parameter_4, parameter_5 # pd_op.matmul: (1x11x1536xf32) <- (1x11x384xf32, 384x1536xf32) matmul_30 = paddle._C_ops.matmul(layer_norm_21, parameter_9, False, False) @@ -797,30 +795,32 @@ def forward( # pd_op.add: (1x11x1536xf32) <- (1x11x1536xf32, 1536xf32) add_36 = paddle._C_ops.add(matmul_30, parameter_8) - del parameter_8 + del matmul_30, parameter_8 # pd_op.gelu: (1x11x1536xf32) <- (1x11x1536xf32) gelu_3 = paddle._C_ops.gelu(add_36, False) + del add_36 # pd_op.matmul: (1x11x384xf32) <- (1x11x1536xf32, 1536x384xf32) matmul_31 = paddle._C_ops.matmul(gelu_3, parameter_7, False, False) - del parameter_7 + del gelu_3, parameter_7 # pd_op.add: (1x11x384xf32) <- (1x11x384xf32, 384xf32) add_37 = paddle._C_ops.add(matmul_31, parameter_6) - del parameter_6 + del matmul_31, parameter_6 # pd_op.dropout: (1x11x384xf32, 1x11x384xui8) <- (1x11x384xf32, None, 1xf32) dropout_24, dropout_25 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_37, None, full_5, False, "upscale_in_train", 0, False + add_37, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) - del add_37 + del add_37, full_5 # pd_op.add: (1x11x384xf32) <- (1x11x384xf32, 1x11x384xf32) add_38 = paddle._C_ops.add(layer_norm_21, dropout_24) + del dropout_24, layer_norm_21 # pd_op.layer_norm: (1x11x384xf32, 1x11xf32, 1x11xf32) <- (1x11x384xf32, 384xf32, 384xf32) layer_norm_24, layer_norm_25, layer_norm_26 = (lambda x, f: f(x))( @@ -829,7 +829,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_2, parameter_3 + del add_38, parameter_2, parameter_3 # pd_op.full_int_array: (1xi64) <- () full_int_array_3 = [0] @@ -841,182 +841,18 @@ def forward( slice_0 = paddle._C_ops.slice( layer_norm_24, [1], full_int_array_3, full_int_array_4, [1], [1] ) + del full_int_array_3, full_int_array_4, layer_norm_24 # pd_op.matmul: (1x384xf32) <- (1x384xf32, 384x384xf32) matmul_32 = paddle._C_ops.matmul(slice_0, parameter_1, False, False) - del parameter_1 + del parameter_1, slice_0 # pd_op.add: (1x384xf32) <- (1x384xf32, 384xf32) add_39 = paddle._C_ops.add(matmul_32, parameter_0) - del parameter_0 + del matmul_32, parameter_0 # pd_op.tanh: (1x384xf32) <- (1x384xf32) tanh_0 = paddle._C_ops.tanh(add_39) - del ( - add_0, - add_1, - add_11, - add_12, - add_13, - add_14, - add_17, - add_18, - add_2, - add_20, - add_21, - add_22, - add_23, - add_26, - add_27, - add_29, - add_3, - add_30, - add_31, - add_32, - add_35, - add_36, - add_38, - add_39, - add_4, - add_5, - add_8, - add_9, - assign_0, - assign_1, - assign_10, - assign_11, - assign_12, - assign_13, - assign_14, - assign_2, - assign_3, - assign_4, - assign_5, - assign_6, - assign_7, - assign_8, - assign_9, - dropout_0, - dropout_1, - dropout_10, - dropout_11, - dropout_12, - dropout_13, - dropout_14, - dropout_15, - dropout_16, - dropout_17, - dropout_18, - dropout_19, - dropout_2, - dropout_20, - dropout_21, - dropout_22, - dropout_23, - dropout_24, - dropout_25, - dropout_3, - dropout_4, - dropout_5, - dropout_6, - dropout_7, - dropout_8, - dropout_9, - embedding_0, - embedding_1, - embedding_2, - embedding_3, - full_5, - full_6, - full_int_array_3, - full_int_array_4, - gelu_0, - gelu_1, - gelu_2, - gelu_3, - layer_norm_1, - layer_norm_10, - layer_norm_11, - layer_norm_12, - layer_norm_13, - layer_norm_14, - layer_norm_15, - layer_norm_16, - layer_norm_17, - layer_norm_18, - layer_norm_19, - layer_norm_2, - layer_norm_20, - layer_norm_21, - layer_norm_22, - layer_norm_23, - layer_norm_24, - layer_norm_25, - layer_norm_26, - layer_norm_3, - layer_norm_4, - layer_norm_5, - layer_norm_6, - layer_norm_7, - layer_norm_8, - layer_norm_9, - matmul_0, - matmul_1, - matmul_10, - matmul_11, - matmul_13, - matmul_14, - matmul_15, - matmul_16, - matmul_17, - matmul_18, - matmul_19, - matmul_2, - matmul_21, - matmul_22, - matmul_23, - matmul_24, - matmul_25, - matmul_26, - matmul_27, - matmul_29, - matmul_3, - matmul_30, - matmul_31, - matmul_32, - matmul_5, - matmul_6, - matmul_7, - matmul_8, - matmul_9, - reshape_11, - reshape_15, - reshape_3, - reshape_7, - scale_1, - scale_2, - scale_3, - scale_4, - scale_5, - slice_0, - softmax_0, - softmax_1, - softmax_2, - softmax_3, - subtract_0, - transpose_1, - transpose_10, - transpose_11, - transpose_13, - transpose_14, - transpose_15, - transpose_2, - transpose_3, - transpose_5, - transpose_6, - transpose_7, - transpose_9, - unsqueeze_0, - ) + del add_39 return tanh_0 diff --git a/paddle_samples/PaddleNLP/ernie-3.0-micro-zh/weight_meta.py b/paddle_samples/PaddleNLP/ernie-3.0-micro-zh/weight_meta.py index e58922255..9a13d463e 100644 --- a/paddle_samples/PaddleNLP/ernie-3.0-micro-zh/weight_meta.py +++ b/paddle_samples/PaddleNLP/ernie-3.0-micro-zh/weight_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_parameter_0: name = "parameter_0" + original_name = "linear_716.b_0" shape = [384] dtype = "float32" data = None @@ -7,6 +8,7 @@ class Program_weight_tensor_parameter_0: class Program_weight_tensor_parameter_1: name = "parameter_1" + original_name = "linear_716.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.0855754") @@ -18,6 +20,7 @@ class Program_weight_tensor_parameter_1: class Program_weight_tensor_parameter_2: name = "parameter_2" + original_name = "layer_norm_244.b_0" shape = [384] dtype = "float32" min_val = float("-0.527264") @@ -29,6 +32,7 @@ class Program_weight_tensor_parameter_2: class Program_weight_tensor_parameter_3: name = "parameter_3" + original_name = "layer_norm_244.w_0" shape = [384] dtype = "float32" min_val = float("0.689122") @@ -40,6 +44,7 @@ class Program_weight_tensor_parameter_3: class Program_weight_tensor_parameter_4: name = "parameter_4" + original_name = "layer_norm_243.b_0" shape = [384] dtype = "float32" min_val = float("-1.04623") @@ -51,6 +56,7 @@ class Program_weight_tensor_parameter_4: class Program_weight_tensor_parameter_5: name = "parameter_5" + original_name = "layer_norm_243.w_0" shape = [384] dtype = "float32" min_val = float("0.346517") @@ -62,6 +68,7 @@ class Program_weight_tensor_parameter_5: class Program_weight_tensor_parameter_6: name = "parameter_6" + original_name = "linear_715.b_0" shape = [384] dtype = "float32" min_val = float("-0.252589") @@ -73,6 +80,7 @@ class Program_weight_tensor_parameter_6: class Program_weight_tensor_parameter_7: name = "parameter_7" + original_name = "linear_715.w_0" shape = [1536, 384] dtype = "float32" min_val = float("-1.22177") @@ -84,6 +92,7 @@ class Program_weight_tensor_parameter_7: class Program_weight_tensor_parameter_8: name = "parameter_8" + original_name = "linear_714.b_0" shape = [1536] dtype = "float32" min_val = float("-0.839606") @@ -95,6 +104,7 @@ class Program_weight_tensor_parameter_8: class Program_weight_tensor_parameter_9: name = "parameter_9" + original_name = "linear_714.w_0" shape = [384, 1536] dtype = "float32" min_val = float("-0.50079") @@ -106,6 +116,7 @@ class Program_weight_tensor_parameter_9: class Program_weight_tensor_parameter_10: name = "parameter_10" + original_name = "linear_713.b_0" shape = [384] dtype = "float32" min_val = float("-0.449546") @@ -117,6 +128,7 @@ class Program_weight_tensor_parameter_10: class Program_weight_tensor_parameter_11: name = "parameter_11" + original_name = "linear_713.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.57316") @@ -128,6 +140,7 @@ class Program_weight_tensor_parameter_11: class Program_weight_tensor_parameter_12: name = "parameter_12" + original_name = "linear_712.b_0" shape = [384] dtype = "float32" min_val = float("-0.168006") @@ -139,6 +152,7 @@ class Program_weight_tensor_parameter_12: class Program_weight_tensor_parameter_13: name = "parameter_13" + original_name = "linear_712.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.357998") @@ -150,6 +164,7 @@ class Program_weight_tensor_parameter_13: class Program_weight_tensor_parameter_14: name = "parameter_14" + original_name = "linear_711.b_0" shape = [384] dtype = "float32" min_val = float("-0.0239274") @@ -161,6 +176,7 @@ class Program_weight_tensor_parameter_14: class Program_weight_tensor_parameter_15: name = "parameter_15" + original_name = "linear_711.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.411845") @@ -172,6 +188,7 @@ class Program_weight_tensor_parameter_15: class Program_weight_tensor_parameter_16: name = "parameter_16" + original_name = "linear_710.b_0" shape = [384] dtype = "float32" min_val = float("-0.84983") @@ -183,6 +200,7 @@ class Program_weight_tensor_parameter_16: class Program_weight_tensor_parameter_17: name = "parameter_17" + original_name = "linear_710.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.340863") @@ -194,6 +212,7 @@ class Program_weight_tensor_parameter_17: class Program_weight_tensor_parameter_18: name = "parameter_18" + original_name = "layer_norm_242.b_0" shape = [384] dtype = "float32" min_val = float("-0.798408") @@ -205,6 +224,7 @@ class Program_weight_tensor_parameter_18: class Program_weight_tensor_parameter_19: name = "parameter_19" + original_name = "layer_norm_242.w_0" shape = [384] dtype = "float32" min_val = float("0.585196") @@ -216,6 +236,7 @@ class Program_weight_tensor_parameter_19: class Program_weight_tensor_parameter_20: name = "parameter_20" + original_name = "layer_norm_241.b_0" shape = [384] dtype = "float32" min_val = float("-1.32843") @@ -227,6 +248,7 @@ class Program_weight_tensor_parameter_20: class Program_weight_tensor_parameter_21: name = "parameter_21" + original_name = "layer_norm_241.w_0" shape = [384] dtype = "float32" min_val = float("0.584004") @@ -238,6 +260,7 @@ class Program_weight_tensor_parameter_21: class Program_weight_tensor_parameter_22: name = "parameter_22" + original_name = "linear_709.b_0" shape = [384] dtype = "float32" min_val = float("-0.221448") @@ -249,6 +272,7 @@ class Program_weight_tensor_parameter_22: class Program_weight_tensor_parameter_23: name = "parameter_23" + original_name = "linear_709.w_0" shape = [1536, 384] dtype = "float32" min_val = float("-3.03005") @@ -260,6 +284,7 @@ class Program_weight_tensor_parameter_23: class Program_weight_tensor_parameter_24: name = "parameter_24" + original_name = "linear_708.b_0" shape = [1536] dtype = "float32" min_val = float("-0.79632") @@ -271,6 +296,7 @@ class Program_weight_tensor_parameter_24: class Program_weight_tensor_parameter_25: name = "parameter_25" + original_name = "linear_708.w_0" shape = [384, 1536] dtype = "float32" min_val = float("-0.71613") @@ -282,6 +308,7 @@ class Program_weight_tensor_parameter_25: class Program_weight_tensor_parameter_26: name = "parameter_26" + original_name = "linear_707.b_0" shape = [384] dtype = "float32" min_val = float("-0.43232") @@ -293,6 +320,7 @@ class Program_weight_tensor_parameter_26: class Program_weight_tensor_parameter_27: name = "parameter_27" + original_name = "linear_707.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.400469") @@ -304,6 +332,7 @@ class Program_weight_tensor_parameter_27: class Program_weight_tensor_parameter_28: name = "parameter_28" + original_name = "linear_706.b_0" shape = [384] dtype = "float32" min_val = float("-0.225714") @@ -315,6 +344,7 @@ class Program_weight_tensor_parameter_28: class Program_weight_tensor_parameter_29: name = "parameter_29" + original_name = "linear_706.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.338028") @@ -326,6 +356,7 @@ class Program_weight_tensor_parameter_29: class Program_weight_tensor_parameter_30: name = "parameter_30" + original_name = "linear_705.b_0" shape = [384] dtype = "float32" min_val = float("-0.00667432") @@ -337,6 +368,7 @@ class Program_weight_tensor_parameter_30: class Program_weight_tensor_parameter_31: name = "parameter_31" + original_name = "linear_705.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.486352") @@ -348,6 +380,7 @@ class Program_weight_tensor_parameter_31: class Program_weight_tensor_parameter_32: name = "parameter_32" + original_name = "linear_704.b_0" shape = [384] dtype = "float32" min_val = float("-0.70384") @@ -359,6 +392,7 @@ class Program_weight_tensor_parameter_32: class Program_weight_tensor_parameter_33: name = "parameter_33" + original_name = "linear_704.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.660646") @@ -370,6 +404,7 @@ class Program_weight_tensor_parameter_33: class Program_weight_tensor_parameter_34: name = "parameter_34" + original_name = "layer_norm_240.b_0" shape = [384] dtype = "float32" min_val = float("-1.0285") @@ -381,6 +416,7 @@ class Program_weight_tensor_parameter_34: class Program_weight_tensor_parameter_35: name = "parameter_35" + original_name = "layer_norm_240.w_0" shape = [384] dtype = "float32" min_val = float("0.946841") @@ -392,6 +428,7 @@ class Program_weight_tensor_parameter_35: class Program_weight_tensor_parameter_36: name = "parameter_36" + original_name = "layer_norm_239.b_0" shape = [384] dtype = "float32" min_val = float("-1.21336") @@ -403,6 +440,7 @@ class Program_weight_tensor_parameter_36: class Program_weight_tensor_parameter_37: name = "parameter_37" + original_name = "layer_norm_239.w_0" shape = [384] dtype = "float32" min_val = float("0.676564") @@ -414,6 +452,7 @@ class Program_weight_tensor_parameter_37: class Program_weight_tensor_parameter_38: name = "parameter_38" + original_name = "linear_703.b_0" shape = [384] dtype = "float32" min_val = float("-0.241318") @@ -425,6 +464,7 @@ class Program_weight_tensor_parameter_38: class Program_weight_tensor_parameter_39: name = "parameter_39" + original_name = "linear_703.w_0" shape = [1536, 384] dtype = "float32" min_val = float("-4.23129") @@ -436,6 +476,7 @@ class Program_weight_tensor_parameter_39: class Program_weight_tensor_parameter_40: name = "parameter_40" + original_name = "linear_702.b_0" shape = [1536] dtype = "float32" min_val = float("-0.791123") @@ -447,6 +488,7 @@ class Program_weight_tensor_parameter_40: class Program_weight_tensor_parameter_41: name = "parameter_41" + original_name = "linear_702.w_0" shape = [384, 1536] dtype = "float32" min_val = float("-1.1258") @@ -458,6 +500,7 @@ class Program_weight_tensor_parameter_41: class Program_weight_tensor_parameter_42: name = "parameter_42" + original_name = "linear_701.b_0" shape = [384] dtype = "float32" min_val = float("-0.403679") @@ -469,6 +512,7 @@ class Program_weight_tensor_parameter_42: class Program_weight_tensor_parameter_43: name = "parameter_43" + original_name = "linear_701.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.580382") @@ -480,6 +524,7 @@ class Program_weight_tensor_parameter_43: class Program_weight_tensor_parameter_44: name = "parameter_44" + original_name = "linear_700.b_0" shape = [384] dtype = "float32" min_val = float("-0.201122") @@ -491,6 +536,7 @@ class Program_weight_tensor_parameter_44: class Program_weight_tensor_parameter_45: name = "parameter_45" + original_name = "linear_700.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.31056") @@ -502,6 +548,7 @@ class Program_weight_tensor_parameter_45: class Program_weight_tensor_parameter_46: name = "parameter_46" + original_name = "linear_699.b_0" shape = [384] dtype = "float32" min_val = float("-0.00252961") @@ -513,6 +560,7 @@ class Program_weight_tensor_parameter_46: class Program_weight_tensor_parameter_47: name = "parameter_47" + original_name = "linear_699.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.51782") @@ -524,6 +572,7 @@ class Program_weight_tensor_parameter_47: class Program_weight_tensor_parameter_48: name = "parameter_48" + original_name = "linear_698.b_0" shape = [384] dtype = "float32" min_val = float("-0.561087") @@ -535,6 +584,7 @@ class Program_weight_tensor_parameter_48: class Program_weight_tensor_parameter_49: name = "parameter_49" + original_name = "linear_698.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.450637") @@ -546,6 +596,7 @@ class Program_weight_tensor_parameter_49: class Program_weight_tensor_parameter_50: name = "parameter_50" + original_name = "layer_norm_238.b_0" shape = [384] dtype = "float32" min_val = float("-0.966331") @@ -557,6 +608,7 @@ class Program_weight_tensor_parameter_50: class Program_weight_tensor_parameter_51: name = "parameter_51" + original_name = "layer_norm_238.w_0" shape = [384] dtype = "float32" min_val = float("0.82987") @@ -568,6 +620,7 @@ class Program_weight_tensor_parameter_51: class Program_weight_tensor_parameter_52: name = "parameter_52" + original_name = "layer_norm_237.b_0" shape = [384] dtype = "float32" min_val = float("-2.82681") @@ -579,6 +632,7 @@ class Program_weight_tensor_parameter_52: class Program_weight_tensor_parameter_53: name = "parameter_53" + original_name = "layer_norm_237.w_0" shape = [384] dtype = "float32" min_val = float("0.569315") @@ -590,6 +644,7 @@ class Program_weight_tensor_parameter_53: class Program_weight_tensor_parameter_54: name = "parameter_54" + original_name = "linear_697.b_0" shape = [384] dtype = "float32" min_val = float("-0.234367") @@ -601,6 +656,7 @@ class Program_weight_tensor_parameter_54: class Program_weight_tensor_parameter_55: name = "parameter_55" + original_name = "linear_697.w_0" shape = [1536, 384] dtype = "float32" min_val = float("-1.31668") @@ -612,6 +668,7 @@ class Program_weight_tensor_parameter_55: class Program_weight_tensor_parameter_56: name = "parameter_56" + original_name = "linear_696.b_0" shape = [1536] dtype = "float32" min_val = float("-0.647786") @@ -623,6 +680,7 @@ class Program_weight_tensor_parameter_56: class Program_weight_tensor_parameter_57: name = "parameter_57" + original_name = "linear_696.w_0" shape = [384, 1536] dtype = "float32" min_val = float("-0.527084") @@ -634,6 +692,7 @@ class Program_weight_tensor_parameter_57: class Program_weight_tensor_parameter_58: name = "parameter_58" + original_name = "linear_695.b_0" shape = [384] dtype = "float32" min_val = float("-0.317365") @@ -645,6 +704,7 @@ class Program_weight_tensor_parameter_58: class Program_weight_tensor_parameter_59: name = "parameter_59" + original_name = "linear_695.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.297335") @@ -656,6 +716,7 @@ class Program_weight_tensor_parameter_59: class Program_weight_tensor_parameter_60: name = "parameter_60" + original_name = "linear_694.b_0" shape = [384] dtype = "float32" min_val = float("-0.375051") @@ -667,6 +728,7 @@ class Program_weight_tensor_parameter_60: class Program_weight_tensor_parameter_61: name = "parameter_61" + original_name = "linear_694.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.270746") @@ -678,6 +740,7 @@ class Program_weight_tensor_parameter_61: class Program_weight_tensor_parameter_62: name = "parameter_62" + original_name = "linear_693.b_0" shape = [384] dtype = "float32" min_val = float("-0.000782185") @@ -689,6 +752,7 @@ class Program_weight_tensor_parameter_62: class Program_weight_tensor_parameter_63: name = "parameter_63" + original_name = "linear_693.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.545894") @@ -700,6 +764,7 @@ class Program_weight_tensor_parameter_63: class Program_weight_tensor_parameter_64: name = "parameter_64" + original_name = "linear_692.b_0" shape = [384] dtype = "float32" min_val = float("-1.10137") @@ -711,6 +776,7 @@ class Program_weight_tensor_parameter_64: class Program_weight_tensor_parameter_65: name = "parameter_65" + original_name = "linear_692.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.562136") @@ -722,6 +788,7 @@ class Program_weight_tensor_parameter_65: class Program_weight_tensor_parameter_66: name = "parameter_66" + original_name = "layer_norm_236.b_0" shape = [384] dtype = "float32" min_val = float("-2.06395") @@ -733,6 +800,7 @@ class Program_weight_tensor_parameter_66: class Program_weight_tensor_parameter_67: name = "parameter_67" + original_name = "layer_norm_236.w_0" shape = [384] dtype = "float32" min_val = float("0.300635") @@ -744,6 +812,7 @@ class Program_weight_tensor_parameter_67: class Program_weight_tensor_parameter_68: name = "parameter_68" + original_name = "embedding_30.w_0" shape = [16, 384] dtype = "float32" min_val = float("-0.277551") @@ -755,6 +824,7 @@ class Program_weight_tensor_parameter_68: class Program_weight_tensor_parameter_69: name = "parameter_69" + original_name = "embedding_29.w_0" shape = [4, 384] dtype = "float32" min_val = float("-0.185101") @@ -766,6 +836,7 @@ class Program_weight_tensor_parameter_69: class Program_weight_tensor_parameter_70: name = "parameter_70" + original_name = "embedding_28.w_0" shape = [2048, 384] dtype = "float32" min_val = float("-0.477353") @@ -777,6 +848,7 @@ class Program_weight_tensor_parameter_70: class Program_weight_tensor_parameter_71: name = "parameter_71" + original_name = "embedding_27.w_0" shape = [40000, 384] dtype = "float32" min_val = float("-0.78961") diff --git a/paddle_samples/PaddleNLP/ernie-3.0-xbase-zh/input_meta.py b/paddle_samples/PaddleNLP/ernie-3.0-xbase-zh/input_meta.py index 97eb8a799..a4123e481 100644 --- a/paddle_samples/PaddleNLP/ernie-3.0-xbase-zh/input_meta.py +++ b/paddle_samples/PaddleNLP/ernie-3.0-xbase-zh/input_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_data_0: name = "data_0" + original_name = "input_ids" shape = [1, 11] dtype = "int64" data = [1, 811, 1257, 175, 29, 502, 130, 706, 3619, 12046, 2] @@ -7,6 +8,7 @@ class Program_weight_tensor_data_0: class Program_weight_tensor_data_1: name = "data_1" + original_name = "token_type_ids" shape = [1, 11] dtype = "int64" data = [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0] diff --git a/paddle_samples/PaddleNLP/ernie-3.0-xbase-zh/model.py b/paddle_samples/PaddleNLP/ernie-3.0-xbase-zh/model.py index 777a41a3c..77685c19b 100644 --- a/paddle_samples/PaddleNLP/ernie-3.0-xbase-zh/model.py +++ b/paddle_samples/PaddleNLP/ernie-3.0-xbase-zh/model.py @@ -394,10 +394,11 @@ def forward( # pd_op.embedding: (1x11x1024xf32) <- (1x11xi64, 2048x1024xf32) embedding_1 = paddle._C_ops.embedding(subtract_0, parameter_326, -1, False) - del parameter_326 + del parameter_326, subtract_0 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_0 = paddle._C_ops.add(embedding_0, embedding_1) + del embedding_0, embedding_1 # pd_op.embedding: (1x11x1024xf32) <- (1x11xi64, 4x1024xf32) embedding_2 = paddle._C_ops.embedding(data_1, parameter_325, -1, False) @@ -405,6 +406,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_1 = paddle._C_ops.add(add_0, embedding_2) + del add_0, embedding_2 # pd_op.full: (1xf32) <- () full_4 = paddle._C_ops.full( @@ -417,10 +419,11 @@ def forward( # pd_op.embedding: (1x11x1024xf32) <- (1x11xi64, 16x1024xf32) embedding_3 = paddle._C_ops.embedding(scale_1, parameter_324, -1, False) - del parameter_324 + del parameter_324, scale_1 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_2 = paddle._C_ops.add(add_1, embedding_3) + del add_1, embedding_3 # pd_op.layer_norm: (1x11x1024xf32, 1x11xf32, 1x11xf32) <- (1x11x1024xf32, 1024xf32, 1024xf32) layer_norm_0, layer_norm_1, layer_norm_2 = (lambda x, f: f(x))( @@ -429,197 +432,17 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_322, parameter_323 + del add_2, parameter_322, parameter_323 # pd_op.full: (1xf32) <- () full_5 = paddle._C_ops.full( [1], float("0.1"), paddle.float32, paddle.core.CPUPlace() ) - # pd_op.assign: (1xf32) <- (1xf32) - assign_0 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_1 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_2 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_3 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_4 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_5 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_6 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_7 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_8 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_9 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_10 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_11 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_12 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_13 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_14 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_15 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_16 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_17 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_18 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_19 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_20 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_21 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_22 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_23 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_24 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_25 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_26 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_27 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_28 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_29 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_30 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_31 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_32 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_33 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_34 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_35 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_36 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_37 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_38 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_39 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_40 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_41 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_42 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_43 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_44 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_45 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_46 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_47 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_48 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_49 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_50 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_51 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_52 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_53 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_54 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_55 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_56 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_57 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_58 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_59 = full_5 - # pd_op.dropout: (1x11x1024xf32, 1x11x1024xui8) <- (1x11x1024xf32, None, 1xf32) dropout_0, dropout_1 = (lambda x, f: f(x))( paddle._C_ops.dropout( - layer_norm_0, None, full_5, False, "upscale_in_train", 0, False + layer_norm_0, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -631,13 +454,14 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_3 = paddle._C_ops.add(matmul_0, parameter_320) - del parameter_320 + del matmul_0, parameter_320 # pd_op.full_int_array: (4xi64) <- () full_int_array_1 = [0, 0, 16, 64] # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_0 = paddle._C_ops.reshape(add_3, full_int_array_1) + del add_3 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_0 = paddle._C_ops.transpose(reshape_0, [0, 2, 1, 3]) @@ -649,7 +473,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_4 = paddle._C_ops.add(matmul_1, parameter_318) - del parameter_318 + del matmul_1, parameter_318 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_2 = paddle._C_ops.matmul(dropout_0, parameter_317, False, False) @@ -657,10 +481,11 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_5 = paddle._C_ops.add(matmul_2, parameter_316) - del parameter_316 + del matmul_2, parameter_316 # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_1 = paddle._C_ops.reshape(add_4, full_int_array_1) + del add_4 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_1 = paddle._C_ops.transpose(reshape_1, [0, 2, 1, 3]) @@ -668,6 +493,7 @@ def forward( # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_2 = paddle._C_ops.reshape(add_5, full_int_array_1) + del add_5 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_2 = paddle._C_ops.transpose(reshape_2, [0, 2, 1, 3]) @@ -678,72 +504,17 @@ def forward( [1], float("0.125"), paddle.float32, paddle.core.CPUPlace() ) - # pd_op.assign: (1xf32) <- (1xf32) - assign_60 = full_6 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_61 = full_6 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_62 = full_6 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_63 = full_6 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_64 = full_6 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_65 = full_6 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_66 = full_6 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_67 = full_6 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_68 = full_6 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_69 = full_6 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_70 = full_6 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_71 = full_6 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_72 = full_6 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_73 = full_6 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_74 = full_6 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_75 = full_6 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_76 = full_6 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_77 = full_6 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_78 = full_6 - # pd_op.scale: (1x16x11x64xf32) <- (1x16x11x64xf32, 1xf32) scale_2 = paddle._C_ops.scale(transpose_0, full_6, float("0"), True) del transpose_0 # pd_op.matmul: (1x16x11x11xf32) <- (1x16x11x64xf32, 1x16x11x64xf32) matmul_3 = paddle._C_ops.matmul(scale_2, transpose_1, False, True) + del scale_2, transpose_1 # pd_op.add: (1x16x11x11xf32) <- (1x16x11x11xf32, 1x1x1x11xf32) add_6 = paddle._C_ops.add(matmul_3, unsqueeze_0) + del matmul_3 # pd_op.softmax: (1x16x11x11xf32) <- (1x16x11x11xf32) softmax_0 = paddle._C_ops.softmax(add_6, -1) @@ -752,13 +523,15 @@ def forward( # pd_op.dropout: (1x16x11x11xf32, 1x16x11x11xui8) <- (1x16x11x11xf32, None, 1xf32) dropout_2, dropout_3 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_0, None, full_5, False, "upscale_in_train", 0, False + softmax_0, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_0 # pd_op.matmul: (1x16x11x64xf32) <- (1x16x11x11xf32, 1x16x11x64xf32) matmul_4 = paddle._C_ops.matmul(dropout_2, transpose_2, False, False) + del dropout_2, transpose_2 # pd_op.transpose: (1x11x16x64xf32) <- (1x16x11x64xf32) transpose_3 = paddle._C_ops.transpose(matmul_4, [0, 2, 1, 3]) @@ -769,19 +542,20 @@ def forward( # pd_op.reshape: (1x11x1024xf32) <- (1x11x16x64xf32, 3xi64) reshape_3 = paddle._C_ops.reshape(transpose_3, full_int_array_2) + del transpose_3 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_5 = paddle._C_ops.matmul(reshape_3, parameter_315, False, False) - del parameter_315 + del parameter_315, reshape_3 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_7 = paddle._C_ops.add(matmul_5, parameter_314) - del parameter_314 + del matmul_5, parameter_314 # pd_op.dropout: (1x11x1024xf32, 1x11x1024xui8) <- (1x11x1024xf32, None, 1xf32) dropout_4, dropout_5 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_7, None, full_5, False, "upscale_in_train", 0, False + add_7, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -789,6 +563,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_8 = paddle._C_ops.add(dropout_0, dropout_4) + del dropout_0, dropout_4 # pd_op.layer_norm: (1x11x1024xf32, 1x11xf32, 1x11xf32) <- (1x11x1024xf32, 1024xf32, 1024xf32) layer_norm_3, layer_norm_4, layer_norm_5 = (lambda x, f: f(x))( @@ -797,7 +572,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_308, parameter_309 + del add_8, parameter_308, parameter_309 # pd_op.matmul: (1x11x4096xf32) <- (1x11x1024xf32, 1024x4096xf32) matmul_6 = paddle._C_ops.matmul(layer_norm_3, parameter_313, False, False) @@ -805,23 +580,24 @@ def forward( # pd_op.add: (1x11x4096xf32) <- (1x11x4096xf32, 4096xf32) add_9 = paddle._C_ops.add(matmul_6, parameter_312) - del parameter_312 + del matmul_6, parameter_312 # pd_op.gelu: (1x11x4096xf32) <- (1x11x4096xf32) gelu_0 = paddle._C_ops.gelu(add_9, False) + del add_9 # pd_op.matmul: (1x11x1024xf32) <- (1x11x4096xf32, 4096x1024xf32) matmul_7 = paddle._C_ops.matmul(gelu_0, parameter_311, False, False) - del parameter_311 + del gelu_0, parameter_311 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_10 = paddle._C_ops.add(matmul_7, parameter_310) - del parameter_310 + del matmul_7, parameter_310 # pd_op.dropout: (1x11x1024xf32, 1x11x1024xui8) <- (1x11x1024xf32, None, 1xf32) dropout_6, dropout_7 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_10, None, full_5, False, "upscale_in_train", 0, False + add_10, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -829,6 +605,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_11 = paddle._C_ops.add(layer_norm_3, dropout_6) + del dropout_6, layer_norm_3 # pd_op.layer_norm: (1x11x1024xf32, 1x11xf32, 1x11xf32) <- (1x11x1024xf32, 1024xf32, 1024xf32) layer_norm_6, layer_norm_7, layer_norm_8 = (lambda x, f: f(x))( @@ -837,7 +614,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_306, parameter_307 + del add_11, parameter_306, parameter_307 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_8 = paddle._C_ops.matmul(layer_norm_6, parameter_305, False, False) @@ -845,10 +622,11 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_12 = paddle._C_ops.add(matmul_8, parameter_304) - del parameter_304 + del matmul_8, parameter_304 # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_4 = paddle._C_ops.reshape(add_12, full_int_array_1) + del add_12 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_4 = paddle._C_ops.transpose(reshape_4, [0, 2, 1, 3]) @@ -860,7 +638,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_13 = paddle._C_ops.add(matmul_9, parameter_302) - del parameter_302 + del matmul_9, parameter_302 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_10 = paddle._C_ops.matmul(layer_norm_6, parameter_301, False, False) @@ -868,10 +646,11 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_14 = paddle._C_ops.add(matmul_10, parameter_300) - del parameter_300 + del matmul_10, parameter_300 # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_5 = paddle._C_ops.reshape(add_13, full_int_array_1) + del add_13 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_5 = paddle._C_ops.transpose(reshape_5, [0, 2, 1, 3]) @@ -879,6 +658,7 @@ def forward( # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_6 = paddle._C_ops.reshape(add_14, full_int_array_1) + del add_14 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_6 = paddle._C_ops.transpose(reshape_6, [0, 2, 1, 3]) @@ -890,9 +670,11 @@ def forward( # pd_op.matmul: (1x16x11x11xf32) <- (1x16x11x64xf32, 1x16x11x64xf32) matmul_11 = paddle._C_ops.matmul(scale_3, transpose_5, False, True) + del scale_3, transpose_5 # pd_op.add: (1x16x11x11xf32) <- (1x16x11x11xf32, 1x1x1x11xf32) add_15 = paddle._C_ops.add(matmul_11, unsqueeze_0) + del matmul_11 # pd_op.softmax: (1x16x11x11xf32) <- (1x16x11x11xf32) softmax_1 = paddle._C_ops.softmax(add_15, -1) @@ -901,13 +683,15 @@ def forward( # pd_op.dropout: (1x16x11x11xf32, 1x16x11x11xui8) <- (1x16x11x11xf32, None, 1xf32) dropout_8, dropout_9 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_1, None, full_5, False, "upscale_in_train", 0, False + softmax_1, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_1 # pd_op.matmul: (1x16x11x64xf32) <- (1x16x11x11xf32, 1x16x11x64xf32) matmul_12 = paddle._C_ops.matmul(dropout_8, transpose_6, False, False) + del dropout_8, transpose_6 # pd_op.transpose: (1x11x16x64xf32) <- (1x16x11x64xf32) transpose_7 = paddle._C_ops.transpose(matmul_12, [0, 2, 1, 3]) @@ -915,19 +699,20 @@ def forward( # pd_op.reshape: (1x11x1024xf32) <- (1x11x16x64xf32, 3xi64) reshape_7 = paddle._C_ops.reshape(transpose_7, full_int_array_2) + del transpose_7 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_13 = paddle._C_ops.matmul(reshape_7, parameter_299, False, False) - del parameter_299 + del parameter_299, reshape_7 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_16 = paddle._C_ops.add(matmul_13, parameter_298) - del parameter_298 + del matmul_13, parameter_298 # pd_op.dropout: (1x11x1024xf32, 1x11x1024xui8) <- (1x11x1024xf32, None, 1xf32) dropout_10, dropout_11 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_16, None, full_5, False, "upscale_in_train", 0, False + add_16, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -935,6 +720,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_17 = paddle._C_ops.add(layer_norm_6, dropout_10) + del dropout_10, layer_norm_6 # pd_op.layer_norm: (1x11x1024xf32, 1x11xf32, 1x11xf32) <- (1x11x1024xf32, 1024xf32, 1024xf32) layer_norm_9, layer_norm_10, layer_norm_11 = (lambda x, f: f(x))( @@ -943,7 +729,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_292, parameter_293 + del add_17, parameter_292, parameter_293 # pd_op.matmul: (1x11x4096xf32) <- (1x11x1024xf32, 1024x4096xf32) matmul_14 = paddle._C_ops.matmul(layer_norm_9, parameter_297, False, False) @@ -951,23 +737,24 @@ def forward( # pd_op.add: (1x11x4096xf32) <- (1x11x4096xf32, 4096xf32) add_18 = paddle._C_ops.add(matmul_14, parameter_296) - del parameter_296 + del matmul_14, parameter_296 # pd_op.gelu: (1x11x4096xf32) <- (1x11x4096xf32) gelu_1 = paddle._C_ops.gelu(add_18, False) + del add_18 # pd_op.matmul: (1x11x1024xf32) <- (1x11x4096xf32, 4096x1024xf32) matmul_15 = paddle._C_ops.matmul(gelu_1, parameter_295, False, False) - del parameter_295 + del gelu_1, parameter_295 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_19 = paddle._C_ops.add(matmul_15, parameter_294) - del parameter_294 + del matmul_15, parameter_294 # pd_op.dropout: (1x11x1024xf32, 1x11x1024xui8) <- (1x11x1024xf32, None, 1xf32) dropout_12, dropout_13 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_19, None, full_5, False, "upscale_in_train", 0, False + add_19, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -975,6 +762,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_20 = paddle._C_ops.add(layer_norm_9, dropout_12) + del dropout_12, layer_norm_9 # pd_op.layer_norm: (1x11x1024xf32, 1x11xf32, 1x11xf32) <- (1x11x1024xf32, 1024xf32, 1024xf32) layer_norm_12, layer_norm_13, layer_norm_14 = (lambda x, f: f(x))( @@ -983,7 +771,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_290, parameter_291 + del add_20, parameter_290, parameter_291 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_16 = paddle._C_ops.matmul(layer_norm_12, parameter_289, False, False) @@ -991,10 +779,11 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_21 = paddle._C_ops.add(matmul_16, parameter_288) - del parameter_288 + del matmul_16, parameter_288 # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_8 = paddle._C_ops.reshape(add_21, full_int_array_1) + del add_21 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_8 = paddle._C_ops.transpose(reshape_8, [0, 2, 1, 3]) @@ -1006,7 +795,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_22 = paddle._C_ops.add(matmul_17, parameter_286) - del parameter_286 + del matmul_17, parameter_286 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_18 = paddle._C_ops.matmul(layer_norm_12, parameter_285, False, False) @@ -1014,10 +803,11 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_23 = paddle._C_ops.add(matmul_18, parameter_284) - del parameter_284 + del matmul_18, parameter_284 # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_9 = paddle._C_ops.reshape(add_22, full_int_array_1) + del add_22 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_9 = paddle._C_ops.transpose(reshape_9, [0, 2, 1, 3]) @@ -1025,6 +815,7 @@ def forward( # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_10 = paddle._C_ops.reshape(add_23, full_int_array_1) + del add_23 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_10 = paddle._C_ops.transpose(reshape_10, [0, 2, 1, 3]) @@ -1036,9 +827,11 @@ def forward( # pd_op.matmul: (1x16x11x11xf32) <- (1x16x11x64xf32, 1x16x11x64xf32) matmul_19 = paddle._C_ops.matmul(scale_4, transpose_9, False, True) + del scale_4, transpose_9 # pd_op.add: (1x16x11x11xf32) <- (1x16x11x11xf32, 1x1x1x11xf32) add_24 = paddle._C_ops.add(matmul_19, unsqueeze_0) + del matmul_19 # pd_op.softmax: (1x16x11x11xf32) <- (1x16x11x11xf32) softmax_2 = paddle._C_ops.softmax(add_24, -1) @@ -1047,13 +840,15 @@ def forward( # pd_op.dropout: (1x16x11x11xf32, 1x16x11x11xui8) <- (1x16x11x11xf32, None, 1xf32) dropout_14, dropout_15 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_2, None, full_5, False, "upscale_in_train", 0, False + softmax_2, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_2 # pd_op.matmul: (1x16x11x64xf32) <- (1x16x11x11xf32, 1x16x11x64xf32) matmul_20 = paddle._C_ops.matmul(dropout_14, transpose_10, False, False) + del dropout_14, transpose_10 # pd_op.transpose: (1x11x16x64xf32) <- (1x16x11x64xf32) transpose_11 = paddle._C_ops.transpose(matmul_20, [0, 2, 1, 3]) @@ -1061,19 +856,20 @@ def forward( # pd_op.reshape: (1x11x1024xf32) <- (1x11x16x64xf32, 3xi64) reshape_11 = paddle._C_ops.reshape(transpose_11, full_int_array_2) + del transpose_11 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_21 = paddle._C_ops.matmul(reshape_11, parameter_283, False, False) - del parameter_283 + del parameter_283, reshape_11 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_25 = paddle._C_ops.add(matmul_21, parameter_282) - del parameter_282 + del matmul_21, parameter_282 # pd_op.dropout: (1x11x1024xf32, 1x11x1024xui8) <- (1x11x1024xf32, None, 1xf32) dropout_16, dropout_17 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_25, None, full_5, False, "upscale_in_train", 0, False + add_25, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1081,6 +877,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_26 = paddle._C_ops.add(layer_norm_12, dropout_16) + del dropout_16, layer_norm_12 # pd_op.layer_norm: (1x11x1024xf32, 1x11xf32, 1x11xf32) <- (1x11x1024xf32, 1024xf32, 1024xf32) layer_norm_15, layer_norm_16, layer_norm_17 = (lambda x, f: f(x))( @@ -1089,7 +886,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_276, parameter_277 + del add_26, parameter_276, parameter_277 # pd_op.matmul: (1x11x4096xf32) <- (1x11x1024xf32, 1024x4096xf32) matmul_22 = paddle._C_ops.matmul(layer_norm_15, parameter_281, False, False) @@ -1097,23 +894,24 @@ def forward( # pd_op.add: (1x11x4096xf32) <- (1x11x4096xf32, 4096xf32) add_27 = paddle._C_ops.add(matmul_22, parameter_280) - del parameter_280 + del matmul_22, parameter_280 # pd_op.gelu: (1x11x4096xf32) <- (1x11x4096xf32) gelu_2 = paddle._C_ops.gelu(add_27, False) + del add_27 # pd_op.matmul: (1x11x1024xf32) <- (1x11x4096xf32, 4096x1024xf32) matmul_23 = paddle._C_ops.matmul(gelu_2, parameter_279, False, False) - del parameter_279 + del gelu_2, parameter_279 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_28 = paddle._C_ops.add(matmul_23, parameter_278) - del parameter_278 + del matmul_23, parameter_278 # pd_op.dropout: (1x11x1024xf32, 1x11x1024xui8) <- (1x11x1024xf32, None, 1xf32) dropout_18, dropout_19 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_28, None, full_5, False, "upscale_in_train", 0, False + add_28, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1121,6 +919,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_29 = paddle._C_ops.add(layer_norm_15, dropout_18) + del dropout_18, layer_norm_15 # pd_op.layer_norm: (1x11x1024xf32, 1x11xf32, 1x11xf32) <- (1x11x1024xf32, 1024xf32, 1024xf32) layer_norm_18, layer_norm_19, layer_norm_20 = (lambda x, f: f(x))( @@ -1129,7 +928,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_274, parameter_275 + del add_29, parameter_274, parameter_275 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_24 = paddle._C_ops.matmul(layer_norm_18, parameter_273, False, False) @@ -1137,10 +936,11 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_30 = paddle._C_ops.add(matmul_24, parameter_272) - del parameter_272 + del matmul_24, parameter_272 # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_12 = paddle._C_ops.reshape(add_30, full_int_array_1) + del add_30 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_12 = paddle._C_ops.transpose(reshape_12, [0, 2, 1, 3]) @@ -1152,7 +952,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_31 = paddle._C_ops.add(matmul_25, parameter_270) - del parameter_270 + del matmul_25, parameter_270 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_26 = paddle._C_ops.matmul(layer_norm_18, parameter_269, False, False) @@ -1160,10 +960,11 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_32 = paddle._C_ops.add(matmul_26, parameter_268) - del parameter_268 + del matmul_26, parameter_268 # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_13 = paddle._C_ops.reshape(add_31, full_int_array_1) + del add_31 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_13 = paddle._C_ops.transpose(reshape_13, [0, 2, 1, 3]) @@ -1171,6 +972,7 @@ def forward( # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_14 = paddle._C_ops.reshape(add_32, full_int_array_1) + del add_32 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_14 = paddle._C_ops.transpose(reshape_14, [0, 2, 1, 3]) @@ -1182,9 +984,11 @@ def forward( # pd_op.matmul: (1x16x11x11xf32) <- (1x16x11x64xf32, 1x16x11x64xf32) matmul_27 = paddle._C_ops.matmul(scale_5, transpose_13, False, True) + del scale_5, transpose_13 # pd_op.add: (1x16x11x11xf32) <- (1x16x11x11xf32, 1x1x1x11xf32) add_33 = paddle._C_ops.add(matmul_27, unsqueeze_0) + del matmul_27 # pd_op.softmax: (1x16x11x11xf32) <- (1x16x11x11xf32) softmax_3 = paddle._C_ops.softmax(add_33, -1) @@ -1193,13 +997,15 @@ def forward( # pd_op.dropout: (1x16x11x11xf32, 1x16x11x11xui8) <- (1x16x11x11xf32, None, 1xf32) dropout_20, dropout_21 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_3, None, full_5, False, "upscale_in_train", 0, False + softmax_3, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_3 # pd_op.matmul: (1x16x11x64xf32) <- (1x16x11x11xf32, 1x16x11x64xf32) matmul_28 = paddle._C_ops.matmul(dropout_20, transpose_14, False, False) + del dropout_20, transpose_14 # pd_op.transpose: (1x11x16x64xf32) <- (1x16x11x64xf32) transpose_15 = paddle._C_ops.transpose(matmul_28, [0, 2, 1, 3]) @@ -1207,19 +1013,20 @@ def forward( # pd_op.reshape: (1x11x1024xf32) <- (1x11x16x64xf32, 3xi64) reshape_15 = paddle._C_ops.reshape(transpose_15, full_int_array_2) + del transpose_15 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_29 = paddle._C_ops.matmul(reshape_15, parameter_267, False, False) - del parameter_267 + del parameter_267, reshape_15 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_34 = paddle._C_ops.add(matmul_29, parameter_266) - del parameter_266 + del matmul_29, parameter_266 # pd_op.dropout: (1x11x1024xf32, 1x11x1024xui8) <- (1x11x1024xf32, None, 1xf32) dropout_22, dropout_23 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_34, None, full_5, False, "upscale_in_train", 0, False + add_34, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1227,6 +1034,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_35 = paddle._C_ops.add(layer_norm_18, dropout_22) + del dropout_22, layer_norm_18 # pd_op.layer_norm: (1x11x1024xf32, 1x11xf32, 1x11xf32) <- (1x11x1024xf32, 1024xf32, 1024xf32) layer_norm_21, layer_norm_22, layer_norm_23 = (lambda x, f: f(x))( @@ -1235,7 +1043,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_260, parameter_261 + del add_35, parameter_260, parameter_261 # pd_op.matmul: (1x11x4096xf32) <- (1x11x1024xf32, 1024x4096xf32) matmul_30 = paddle._C_ops.matmul(layer_norm_21, parameter_265, False, False) @@ -1243,23 +1051,24 @@ def forward( # pd_op.add: (1x11x4096xf32) <- (1x11x4096xf32, 4096xf32) add_36 = paddle._C_ops.add(matmul_30, parameter_264) - del parameter_264 + del matmul_30, parameter_264 # pd_op.gelu: (1x11x4096xf32) <- (1x11x4096xf32) gelu_3 = paddle._C_ops.gelu(add_36, False) + del add_36 # pd_op.matmul: (1x11x1024xf32) <- (1x11x4096xf32, 4096x1024xf32) matmul_31 = paddle._C_ops.matmul(gelu_3, parameter_263, False, False) - del parameter_263 + del gelu_3, parameter_263 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_37 = paddle._C_ops.add(matmul_31, parameter_262) - del parameter_262 + del matmul_31, parameter_262 # pd_op.dropout: (1x11x1024xf32, 1x11x1024xui8) <- (1x11x1024xf32, None, 1xf32) dropout_24, dropout_25 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_37, None, full_5, False, "upscale_in_train", 0, False + add_37, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1267,6 +1076,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_38 = paddle._C_ops.add(layer_norm_21, dropout_24) + del dropout_24, layer_norm_21 # pd_op.layer_norm: (1x11x1024xf32, 1x11xf32, 1x11xf32) <- (1x11x1024xf32, 1024xf32, 1024xf32) layer_norm_24, layer_norm_25, layer_norm_26 = (lambda x, f: f(x))( @@ -1275,7 +1085,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_258, parameter_259 + del add_38, parameter_258, parameter_259 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_32 = paddle._C_ops.matmul(layer_norm_24, parameter_257, False, False) @@ -1283,10 +1093,11 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_39 = paddle._C_ops.add(matmul_32, parameter_256) - del parameter_256 + del matmul_32, parameter_256 # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_16 = paddle._C_ops.reshape(add_39, full_int_array_1) + del add_39 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_16 = paddle._C_ops.transpose(reshape_16, [0, 2, 1, 3]) @@ -1298,7 +1109,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_40 = paddle._C_ops.add(matmul_33, parameter_254) - del parameter_254 + del matmul_33, parameter_254 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_34 = paddle._C_ops.matmul(layer_norm_24, parameter_253, False, False) @@ -1306,10 +1117,11 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_41 = paddle._C_ops.add(matmul_34, parameter_252) - del parameter_252 + del matmul_34, parameter_252 # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_17 = paddle._C_ops.reshape(add_40, full_int_array_1) + del add_40 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_17 = paddle._C_ops.transpose(reshape_17, [0, 2, 1, 3]) @@ -1317,6 +1129,7 @@ def forward( # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_18 = paddle._C_ops.reshape(add_41, full_int_array_1) + del add_41 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_18 = paddle._C_ops.transpose(reshape_18, [0, 2, 1, 3]) @@ -1328,9 +1141,11 @@ def forward( # pd_op.matmul: (1x16x11x11xf32) <- (1x16x11x64xf32, 1x16x11x64xf32) matmul_35 = paddle._C_ops.matmul(scale_6, transpose_17, False, True) + del scale_6, transpose_17 # pd_op.add: (1x16x11x11xf32) <- (1x16x11x11xf32, 1x1x1x11xf32) add_42 = paddle._C_ops.add(matmul_35, unsqueeze_0) + del matmul_35 # pd_op.softmax: (1x16x11x11xf32) <- (1x16x11x11xf32) softmax_4 = paddle._C_ops.softmax(add_42, -1) @@ -1339,13 +1154,15 @@ def forward( # pd_op.dropout: (1x16x11x11xf32, 1x16x11x11xui8) <- (1x16x11x11xf32, None, 1xf32) dropout_26, dropout_27 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_4, None, full_5, False, "upscale_in_train", 0, False + softmax_4, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_4 # pd_op.matmul: (1x16x11x64xf32) <- (1x16x11x11xf32, 1x16x11x64xf32) matmul_36 = paddle._C_ops.matmul(dropout_26, transpose_18, False, False) + del dropout_26, transpose_18 # pd_op.transpose: (1x11x16x64xf32) <- (1x16x11x64xf32) transpose_19 = paddle._C_ops.transpose(matmul_36, [0, 2, 1, 3]) @@ -1353,19 +1170,20 @@ def forward( # pd_op.reshape: (1x11x1024xf32) <- (1x11x16x64xf32, 3xi64) reshape_19 = paddle._C_ops.reshape(transpose_19, full_int_array_2) + del transpose_19 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_37 = paddle._C_ops.matmul(reshape_19, parameter_251, False, False) - del parameter_251 + del parameter_251, reshape_19 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_43 = paddle._C_ops.add(matmul_37, parameter_250) - del parameter_250 + del matmul_37, parameter_250 # pd_op.dropout: (1x11x1024xf32, 1x11x1024xui8) <- (1x11x1024xf32, None, 1xf32) dropout_28, dropout_29 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_43, None, full_5, False, "upscale_in_train", 0, False + add_43, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1373,6 +1191,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_44 = paddle._C_ops.add(layer_norm_24, dropout_28) + del dropout_28, layer_norm_24 # pd_op.layer_norm: (1x11x1024xf32, 1x11xf32, 1x11xf32) <- (1x11x1024xf32, 1024xf32, 1024xf32) layer_norm_27, layer_norm_28, layer_norm_29 = (lambda x, f: f(x))( @@ -1381,7 +1200,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_244, parameter_245 + del add_44, parameter_244, parameter_245 # pd_op.matmul: (1x11x4096xf32) <- (1x11x1024xf32, 1024x4096xf32) matmul_38 = paddle._C_ops.matmul(layer_norm_27, parameter_249, False, False) @@ -1389,23 +1208,24 @@ def forward( # pd_op.add: (1x11x4096xf32) <- (1x11x4096xf32, 4096xf32) add_45 = paddle._C_ops.add(matmul_38, parameter_248) - del parameter_248 + del matmul_38, parameter_248 # pd_op.gelu: (1x11x4096xf32) <- (1x11x4096xf32) gelu_4 = paddle._C_ops.gelu(add_45, False) + del add_45 # pd_op.matmul: (1x11x1024xf32) <- (1x11x4096xf32, 4096x1024xf32) matmul_39 = paddle._C_ops.matmul(gelu_4, parameter_247, False, False) - del parameter_247 + del gelu_4, parameter_247 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_46 = paddle._C_ops.add(matmul_39, parameter_246) - del parameter_246 + del matmul_39, parameter_246 # pd_op.dropout: (1x11x1024xf32, 1x11x1024xui8) <- (1x11x1024xf32, None, 1xf32) dropout_30, dropout_31 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_46, None, full_5, False, "upscale_in_train", 0, False + add_46, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1413,6 +1233,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_47 = paddle._C_ops.add(layer_norm_27, dropout_30) + del dropout_30, layer_norm_27 # pd_op.layer_norm: (1x11x1024xf32, 1x11xf32, 1x11xf32) <- (1x11x1024xf32, 1024xf32, 1024xf32) layer_norm_30, layer_norm_31, layer_norm_32 = (lambda x, f: f(x))( @@ -1421,7 +1242,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_242, parameter_243 + del add_47, parameter_242, parameter_243 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_40 = paddle._C_ops.matmul(layer_norm_30, parameter_241, False, False) @@ -1429,10 +1250,11 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_48 = paddle._C_ops.add(matmul_40, parameter_240) - del parameter_240 + del matmul_40, parameter_240 # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_20 = paddle._C_ops.reshape(add_48, full_int_array_1) + del add_48 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_20 = paddle._C_ops.transpose(reshape_20, [0, 2, 1, 3]) @@ -1444,7 +1266,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_49 = paddle._C_ops.add(matmul_41, parameter_238) - del parameter_238 + del matmul_41, parameter_238 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_42 = paddle._C_ops.matmul(layer_norm_30, parameter_237, False, False) @@ -1452,10 +1274,11 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_50 = paddle._C_ops.add(matmul_42, parameter_236) - del parameter_236 + del matmul_42, parameter_236 # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_21 = paddle._C_ops.reshape(add_49, full_int_array_1) + del add_49 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_21 = paddle._C_ops.transpose(reshape_21, [0, 2, 1, 3]) @@ -1463,6 +1286,7 @@ def forward( # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_22 = paddle._C_ops.reshape(add_50, full_int_array_1) + del add_50 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_22 = paddle._C_ops.transpose(reshape_22, [0, 2, 1, 3]) @@ -1474,9 +1298,11 @@ def forward( # pd_op.matmul: (1x16x11x11xf32) <- (1x16x11x64xf32, 1x16x11x64xf32) matmul_43 = paddle._C_ops.matmul(scale_7, transpose_21, False, True) + del scale_7, transpose_21 # pd_op.add: (1x16x11x11xf32) <- (1x16x11x11xf32, 1x1x1x11xf32) add_51 = paddle._C_ops.add(matmul_43, unsqueeze_0) + del matmul_43 # pd_op.softmax: (1x16x11x11xf32) <- (1x16x11x11xf32) softmax_5 = paddle._C_ops.softmax(add_51, -1) @@ -1485,13 +1311,15 @@ def forward( # pd_op.dropout: (1x16x11x11xf32, 1x16x11x11xui8) <- (1x16x11x11xf32, None, 1xf32) dropout_32, dropout_33 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_5, None, full_5, False, "upscale_in_train", 0, False + softmax_5, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_5 # pd_op.matmul: (1x16x11x64xf32) <- (1x16x11x11xf32, 1x16x11x64xf32) matmul_44 = paddle._C_ops.matmul(dropout_32, transpose_22, False, False) + del dropout_32, transpose_22 # pd_op.transpose: (1x11x16x64xf32) <- (1x16x11x64xf32) transpose_23 = paddle._C_ops.transpose(matmul_44, [0, 2, 1, 3]) @@ -1499,19 +1327,20 @@ def forward( # pd_op.reshape: (1x11x1024xf32) <- (1x11x16x64xf32, 3xi64) reshape_23 = paddle._C_ops.reshape(transpose_23, full_int_array_2) + del transpose_23 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_45 = paddle._C_ops.matmul(reshape_23, parameter_235, False, False) - del parameter_235 + del parameter_235, reshape_23 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_52 = paddle._C_ops.add(matmul_45, parameter_234) - del parameter_234 + del matmul_45, parameter_234 # pd_op.dropout: (1x11x1024xf32, 1x11x1024xui8) <- (1x11x1024xf32, None, 1xf32) dropout_34, dropout_35 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_52, None, full_5, False, "upscale_in_train", 0, False + add_52, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1519,6 +1348,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_53 = paddle._C_ops.add(layer_norm_30, dropout_34) + del dropout_34, layer_norm_30 # pd_op.layer_norm: (1x11x1024xf32, 1x11xf32, 1x11xf32) <- (1x11x1024xf32, 1024xf32, 1024xf32) layer_norm_33, layer_norm_34, layer_norm_35 = (lambda x, f: f(x))( @@ -1527,7 +1357,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_228, parameter_229 + del add_53, parameter_228, parameter_229 # pd_op.matmul: (1x11x4096xf32) <- (1x11x1024xf32, 1024x4096xf32) matmul_46 = paddle._C_ops.matmul(layer_norm_33, parameter_233, False, False) @@ -1535,23 +1365,24 @@ def forward( # pd_op.add: (1x11x4096xf32) <- (1x11x4096xf32, 4096xf32) add_54 = paddle._C_ops.add(matmul_46, parameter_232) - del parameter_232 + del matmul_46, parameter_232 # pd_op.gelu: (1x11x4096xf32) <- (1x11x4096xf32) gelu_5 = paddle._C_ops.gelu(add_54, False) + del add_54 # pd_op.matmul: (1x11x1024xf32) <- (1x11x4096xf32, 4096x1024xf32) matmul_47 = paddle._C_ops.matmul(gelu_5, parameter_231, False, False) - del parameter_231 + del gelu_5, parameter_231 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_55 = paddle._C_ops.add(matmul_47, parameter_230) - del parameter_230 + del matmul_47, parameter_230 # pd_op.dropout: (1x11x1024xf32, 1x11x1024xui8) <- (1x11x1024xf32, None, 1xf32) dropout_36, dropout_37 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_55, None, full_5, False, "upscale_in_train", 0, False + add_55, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1559,6 +1390,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_56 = paddle._C_ops.add(layer_norm_33, dropout_36) + del dropout_36, layer_norm_33 # pd_op.layer_norm: (1x11x1024xf32, 1x11xf32, 1x11xf32) <- (1x11x1024xf32, 1024xf32, 1024xf32) layer_norm_36, layer_norm_37, layer_norm_38 = (lambda x, f: f(x))( @@ -1567,7 +1399,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_226, parameter_227 + del add_56, parameter_226, parameter_227 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_48 = paddle._C_ops.matmul(layer_norm_36, parameter_225, False, False) @@ -1575,10 +1407,11 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_57 = paddle._C_ops.add(matmul_48, parameter_224) - del parameter_224 + del matmul_48, parameter_224 # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_24 = paddle._C_ops.reshape(add_57, full_int_array_1) + del add_57 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_24 = paddle._C_ops.transpose(reshape_24, [0, 2, 1, 3]) @@ -1590,7 +1423,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_58 = paddle._C_ops.add(matmul_49, parameter_222) - del parameter_222 + del matmul_49, parameter_222 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_50 = paddle._C_ops.matmul(layer_norm_36, parameter_221, False, False) @@ -1598,10 +1431,11 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_59 = paddle._C_ops.add(matmul_50, parameter_220) - del parameter_220 + del matmul_50, parameter_220 # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_25 = paddle._C_ops.reshape(add_58, full_int_array_1) + del add_58 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_25 = paddle._C_ops.transpose(reshape_25, [0, 2, 1, 3]) @@ -1609,6 +1443,7 @@ def forward( # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_26 = paddle._C_ops.reshape(add_59, full_int_array_1) + del add_59 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_26 = paddle._C_ops.transpose(reshape_26, [0, 2, 1, 3]) @@ -1620,9 +1455,11 @@ def forward( # pd_op.matmul: (1x16x11x11xf32) <- (1x16x11x64xf32, 1x16x11x64xf32) matmul_51 = paddle._C_ops.matmul(scale_8, transpose_25, False, True) + del scale_8, transpose_25 # pd_op.add: (1x16x11x11xf32) <- (1x16x11x11xf32, 1x1x1x11xf32) add_60 = paddle._C_ops.add(matmul_51, unsqueeze_0) + del matmul_51 # pd_op.softmax: (1x16x11x11xf32) <- (1x16x11x11xf32) softmax_6 = paddle._C_ops.softmax(add_60, -1) @@ -1631,13 +1468,15 @@ def forward( # pd_op.dropout: (1x16x11x11xf32, 1x16x11x11xui8) <- (1x16x11x11xf32, None, 1xf32) dropout_38, dropout_39 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_6, None, full_5, False, "upscale_in_train", 0, False + softmax_6, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_6 # pd_op.matmul: (1x16x11x64xf32) <- (1x16x11x11xf32, 1x16x11x64xf32) matmul_52 = paddle._C_ops.matmul(dropout_38, transpose_26, False, False) + del dropout_38, transpose_26 # pd_op.transpose: (1x11x16x64xf32) <- (1x16x11x64xf32) transpose_27 = paddle._C_ops.transpose(matmul_52, [0, 2, 1, 3]) @@ -1645,19 +1484,20 @@ def forward( # pd_op.reshape: (1x11x1024xf32) <- (1x11x16x64xf32, 3xi64) reshape_27 = paddle._C_ops.reshape(transpose_27, full_int_array_2) + del transpose_27 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_53 = paddle._C_ops.matmul(reshape_27, parameter_219, False, False) - del parameter_219 + del parameter_219, reshape_27 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_61 = paddle._C_ops.add(matmul_53, parameter_218) - del parameter_218 + del matmul_53, parameter_218 # pd_op.dropout: (1x11x1024xf32, 1x11x1024xui8) <- (1x11x1024xf32, None, 1xf32) dropout_40, dropout_41 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_61, None, full_5, False, "upscale_in_train", 0, False + add_61, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1665,6 +1505,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_62 = paddle._C_ops.add(layer_norm_36, dropout_40) + del dropout_40, layer_norm_36 # pd_op.layer_norm: (1x11x1024xf32, 1x11xf32, 1x11xf32) <- (1x11x1024xf32, 1024xf32, 1024xf32) layer_norm_39, layer_norm_40, layer_norm_41 = (lambda x, f: f(x))( @@ -1673,7 +1514,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_212, parameter_213 + del add_62, parameter_212, parameter_213 # pd_op.matmul: (1x11x4096xf32) <- (1x11x1024xf32, 1024x4096xf32) matmul_54 = paddle._C_ops.matmul(layer_norm_39, parameter_217, False, False) @@ -1681,23 +1522,24 @@ def forward( # pd_op.add: (1x11x4096xf32) <- (1x11x4096xf32, 4096xf32) add_63 = paddle._C_ops.add(matmul_54, parameter_216) - del parameter_216 + del matmul_54, parameter_216 # pd_op.gelu: (1x11x4096xf32) <- (1x11x4096xf32) gelu_6 = paddle._C_ops.gelu(add_63, False) + del add_63 # pd_op.matmul: (1x11x1024xf32) <- (1x11x4096xf32, 4096x1024xf32) matmul_55 = paddle._C_ops.matmul(gelu_6, parameter_215, False, False) - del parameter_215 + del gelu_6, parameter_215 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_64 = paddle._C_ops.add(matmul_55, parameter_214) - del parameter_214 + del matmul_55, parameter_214 # pd_op.dropout: (1x11x1024xf32, 1x11x1024xui8) <- (1x11x1024xf32, None, 1xf32) dropout_42, dropout_43 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_64, None, full_5, False, "upscale_in_train", 0, False + add_64, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1705,6 +1547,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_65 = paddle._C_ops.add(layer_norm_39, dropout_42) + del dropout_42, layer_norm_39 # pd_op.layer_norm: (1x11x1024xf32, 1x11xf32, 1x11xf32) <- (1x11x1024xf32, 1024xf32, 1024xf32) layer_norm_42, layer_norm_43, layer_norm_44 = (lambda x, f: f(x))( @@ -1713,7 +1556,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_210, parameter_211 + del add_65, parameter_210, parameter_211 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_56 = paddle._C_ops.matmul(layer_norm_42, parameter_209, False, False) @@ -1721,10 +1564,11 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_66 = paddle._C_ops.add(matmul_56, parameter_208) - del parameter_208 + del matmul_56, parameter_208 # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_28 = paddle._C_ops.reshape(add_66, full_int_array_1) + del add_66 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_28 = paddle._C_ops.transpose(reshape_28, [0, 2, 1, 3]) @@ -1736,7 +1580,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_67 = paddle._C_ops.add(matmul_57, parameter_206) - del parameter_206 + del matmul_57, parameter_206 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_58 = paddle._C_ops.matmul(layer_norm_42, parameter_205, False, False) @@ -1744,10 +1588,11 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_68 = paddle._C_ops.add(matmul_58, parameter_204) - del parameter_204 + del matmul_58, parameter_204 # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_29 = paddle._C_ops.reshape(add_67, full_int_array_1) + del add_67 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_29 = paddle._C_ops.transpose(reshape_29, [0, 2, 1, 3]) @@ -1755,6 +1600,7 @@ def forward( # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_30 = paddle._C_ops.reshape(add_68, full_int_array_1) + del add_68 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_30 = paddle._C_ops.transpose(reshape_30, [0, 2, 1, 3]) @@ -1766,9 +1612,11 @@ def forward( # pd_op.matmul: (1x16x11x11xf32) <- (1x16x11x64xf32, 1x16x11x64xf32) matmul_59 = paddle._C_ops.matmul(scale_9, transpose_29, False, True) + del scale_9, transpose_29 # pd_op.add: (1x16x11x11xf32) <- (1x16x11x11xf32, 1x1x1x11xf32) add_69 = paddle._C_ops.add(matmul_59, unsqueeze_0) + del matmul_59 # pd_op.softmax: (1x16x11x11xf32) <- (1x16x11x11xf32) softmax_7 = paddle._C_ops.softmax(add_69, -1) @@ -1777,13 +1625,15 @@ def forward( # pd_op.dropout: (1x16x11x11xf32, 1x16x11x11xui8) <- (1x16x11x11xf32, None, 1xf32) dropout_44, dropout_45 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_7, None, full_5, False, "upscale_in_train", 0, False + softmax_7, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_7 # pd_op.matmul: (1x16x11x64xf32) <- (1x16x11x11xf32, 1x16x11x64xf32) matmul_60 = paddle._C_ops.matmul(dropout_44, transpose_30, False, False) + del dropout_44, transpose_30 # pd_op.transpose: (1x11x16x64xf32) <- (1x16x11x64xf32) transpose_31 = paddle._C_ops.transpose(matmul_60, [0, 2, 1, 3]) @@ -1791,19 +1641,20 @@ def forward( # pd_op.reshape: (1x11x1024xf32) <- (1x11x16x64xf32, 3xi64) reshape_31 = paddle._C_ops.reshape(transpose_31, full_int_array_2) + del transpose_31 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_61 = paddle._C_ops.matmul(reshape_31, parameter_203, False, False) - del parameter_203 + del parameter_203, reshape_31 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_70 = paddle._C_ops.add(matmul_61, parameter_202) - del parameter_202 + del matmul_61, parameter_202 # pd_op.dropout: (1x11x1024xf32, 1x11x1024xui8) <- (1x11x1024xf32, None, 1xf32) dropout_46, dropout_47 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_70, None, full_5, False, "upscale_in_train", 0, False + add_70, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1811,6 +1662,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_71 = paddle._C_ops.add(layer_norm_42, dropout_46) + del dropout_46, layer_norm_42 # pd_op.layer_norm: (1x11x1024xf32, 1x11xf32, 1x11xf32) <- (1x11x1024xf32, 1024xf32, 1024xf32) layer_norm_45, layer_norm_46, layer_norm_47 = (lambda x, f: f(x))( @@ -1819,7 +1671,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_196, parameter_197 + del add_71, parameter_196, parameter_197 # pd_op.matmul: (1x11x4096xf32) <- (1x11x1024xf32, 1024x4096xf32) matmul_62 = paddle._C_ops.matmul(layer_norm_45, parameter_201, False, False) @@ -1827,23 +1679,24 @@ def forward( # pd_op.add: (1x11x4096xf32) <- (1x11x4096xf32, 4096xf32) add_72 = paddle._C_ops.add(matmul_62, parameter_200) - del parameter_200 + del matmul_62, parameter_200 # pd_op.gelu: (1x11x4096xf32) <- (1x11x4096xf32) gelu_7 = paddle._C_ops.gelu(add_72, False) + del add_72 # pd_op.matmul: (1x11x1024xf32) <- (1x11x4096xf32, 4096x1024xf32) matmul_63 = paddle._C_ops.matmul(gelu_7, parameter_199, False, False) - del parameter_199 + del gelu_7, parameter_199 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_73 = paddle._C_ops.add(matmul_63, parameter_198) - del parameter_198 + del matmul_63, parameter_198 # pd_op.dropout: (1x11x1024xf32, 1x11x1024xui8) <- (1x11x1024xf32, None, 1xf32) dropout_48, dropout_49 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_73, None, full_5, False, "upscale_in_train", 0, False + add_73, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1851,6 +1704,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_74 = paddle._C_ops.add(layer_norm_45, dropout_48) + del dropout_48, layer_norm_45 # pd_op.layer_norm: (1x11x1024xf32, 1x11xf32, 1x11xf32) <- (1x11x1024xf32, 1024xf32, 1024xf32) layer_norm_48, layer_norm_49, layer_norm_50 = (lambda x, f: f(x))( @@ -1859,7 +1713,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_194, parameter_195 + del add_74, parameter_194, parameter_195 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_64 = paddle._C_ops.matmul(layer_norm_48, parameter_193, False, False) @@ -1867,10 +1721,11 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_75 = paddle._C_ops.add(matmul_64, parameter_192) - del parameter_192 + del matmul_64, parameter_192 # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_32 = paddle._C_ops.reshape(add_75, full_int_array_1) + del add_75 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_32 = paddle._C_ops.transpose(reshape_32, [0, 2, 1, 3]) @@ -1882,7 +1737,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_76 = paddle._C_ops.add(matmul_65, parameter_190) - del parameter_190 + del matmul_65, parameter_190 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_66 = paddle._C_ops.matmul(layer_norm_48, parameter_189, False, False) @@ -1890,10 +1745,11 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_77 = paddle._C_ops.add(matmul_66, parameter_188) - del parameter_188 + del matmul_66, parameter_188 # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_33 = paddle._C_ops.reshape(add_76, full_int_array_1) + del add_76 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_33 = paddle._C_ops.transpose(reshape_33, [0, 2, 1, 3]) @@ -1901,6 +1757,7 @@ def forward( # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_34 = paddle._C_ops.reshape(add_77, full_int_array_1) + del add_77 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_34 = paddle._C_ops.transpose(reshape_34, [0, 2, 1, 3]) @@ -1912,9 +1769,11 @@ def forward( # pd_op.matmul: (1x16x11x11xf32) <- (1x16x11x64xf32, 1x16x11x64xf32) matmul_67 = paddle._C_ops.matmul(scale_10, transpose_33, False, True) + del scale_10, transpose_33 # pd_op.add: (1x16x11x11xf32) <- (1x16x11x11xf32, 1x1x1x11xf32) add_78 = paddle._C_ops.add(matmul_67, unsqueeze_0) + del matmul_67 # pd_op.softmax: (1x16x11x11xf32) <- (1x16x11x11xf32) softmax_8 = paddle._C_ops.softmax(add_78, -1) @@ -1923,13 +1782,15 @@ def forward( # pd_op.dropout: (1x16x11x11xf32, 1x16x11x11xui8) <- (1x16x11x11xf32, None, 1xf32) dropout_50, dropout_51 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_8, None, full_5, False, "upscale_in_train", 0, False + softmax_8, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_8 # pd_op.matmul: (1x16x11x64xf32) <- (1x16x11x11xf32, 1x16x11x64xf32) matmul_68 = paddle._C_ops.matmul(dropout_50, transpose_34, False, False) + del dropout_50, transpose_34 # pd_op.transpose: (1x11x16x64xf32) <- (1x16x11x64xf32) transpose_35 = paddle._C_ops.transpose(matmul_68, [0, 2, 1, 3]) @@ -1937,19 +1798,20 @@ def forward( # pd_op.reshape: (1x11x1024xf32) <- (1x11x16x64xf32, 3xi64) reshape_35 = paddle._C_ops.reshape(transpose_35, full_int_array_2) + del transpose_35 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_69 = paddle._C_ops.matmul(reshape_35, parameter_187, False, False) - del parameter_187 + del parameter_187, reshape_35 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_79 = paddle._C_ops.add(matmul_69, parameter_186) - del parameter_186 + del matmul_69, parameter_186 # pd_op.dropout: (1x11x1024xf32, 1x11x1024xui8) <- (1x11x1024xf32, None, 1xf32) dropout_52, dropout_53 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_79, None, full_5, False, "upscale_in_train", 0, False + add_79, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1957,6 +1819,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_80 = paddle._C_ops.add(layer_norm_48, dropout_52) + del dropout_52, layer_norm_48 # pd_op.layer_norm: (1x11x1024xf32, 1x11xf32, 1x11xf32) <- (1x11x1024xf32, 1024xf32, 1024xf32) layer_norm_51, layer_norm_52, layer_norm_53 = (lambda x, f: f(x))( @@ -1965,7 +1828,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_180, parameter_181 + del add_80, parameter_180, parameter_181 # pd_op.matmul: (1x11x4096xf32) <- (1x11x1024xf32, 1024x4096xf32) matmul_70 = paddle._C_ops.matmul(layer_norm_51, parameter_185, False, False) @@ -1973,23 +1836,24 @@ def forward( # pd_op.add: (1x11x4096xf32) <- (1x11x4096xf32, 4096xf32) add_81 = paddle._C_ops.add(matmul_70, parameter_184) - del parameter_184 + del matmul_70, parameter_184 # pd_op.gelu: (1x11x4096xf32) <- (1x11x4096xf32) gelu_8 = paddle._C_ops.gelu(add_81, False) + del add_81 # pd_op.matmul: (1x11x1024xf32) <- (1x11x4096xf32, 4096x1024xf32) matmul_71 = paddle._C_ops.matmul(gelu_8, parameter_183, False, False) - del parameter_183 + del gelu_8, parameter_183 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_82 = paddle._C_ops.add(matmul_71, parameter_182) - del parameter_182 + del matmul_71, parameter_182 # pd_op.dropout: (1x11x1024xf32, 1x11x1024xui8) <- (1x11x1024xf32, None, 1xf32) dropout_54, dropout_55 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_82, None, full_5, False, "upscale_in_train", 0, False + add_82, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1997,6 +1861,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_83 = paddle._C_ops.add(layer_norm_51, dropout_54) + del dropout_54, layer_norm_51 # pd_op.layer_norm: (1x11x1024xf32, 1x11xf32, 1x11xf32) <- (1x11x1024xf32, 1024xf32, 1024xf32) layer_norm_54, layer_norm_55, layer_norm_56 = (lambda x, f: f(x))( @@ -2005,7 +1870,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_178, parameter_179 + del add_83, parameter_178, parameter_179 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_72 = paddle._C_ops.matmul(layer_norm_54, parameter_177, False, False) @@ -2013,10 +1878,11 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_84 = paddle._C_ops.add(matmul_72, parameter_176) - del parameter_176 + del matmul_72, parameter_176 # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_36 = paddle._C_ops.reshape(add_84, full_int_array_1) + del add_84 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_36 = paddle._C_ops.transpose(reshape_36, [0, 2, 1, 3]) @@ -2028,7 +1894,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_85 = paddle._C_ops.add(matmul_73, parameter_174) - del parameter_174 + del matmul_73, parameter_174 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_74 = paddle._C_ops.matmul(layer_norm_54, parameter_173, False, False) @@ -2036,10 +1902,11 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_86 = paddle._C_ops.add(matmul_74, parameter_172) - del parameter_172 + del matmul_74, parameter_172 # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_37 = paddle._C_ops.reshape(add_85, full_int_array_1) + del add_85 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_37 = paddle._C_ops.transpose(reshape_37, [0, 2, 1, 3]) @@ -2047,6 +1914,7 @@ def forward( # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_38 = paddle._C_ops.reshape(add_86, full_int_array_1) + del add_86 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_38 = paddle._C_ops.transpose(reshape_38, [0, 2, 1, 3]) @@ -2058,9 +1926,11 @@ def forward( # pd_op.matmul: (1x16x11x11xf32) <- (1x16x11x64xf32, 1x16x11x64xf32) matmul_75 = paddle._C_ops.matmul(scale_11, transpose_37, False, True) + del scale_11, transpose_37 # pd_op.add: (1x16x11x11xf32) <- (1x16x11x11xf32, 1x1x1x11xf32) add_87 = paddle._C_ops.add(matmul_75, unsqueeze_0) + del matmul_75 # pd_op.softmax: (1x16x11x11xf32) <- (1x16x11x11xf32) softmax_9 = paddle._C_ops.softmax(add_87, -1) @@ -2069,13 +1939,15 @@ def forward( # pd_op.dropout: (1x16x11x11xf32, 1x16x11x11xui8) <- (1x16x11x11xf32, None, 1xf32) dropout_56, dropout_57 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_9, None, full_5, False, "upscale_in_train", 0, False + softmax_9, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_9 # pd_op.matmul: (1x16x11x64xf32) <- (1x16x11x11xf32, 1x16x11x64xf32) matmul_76 = paddle._C_ops.matmul(dropout_56, transpose_38, False, False) + del dropout_56, transpose_38 # pd_op.transpose: (1x11x16x64xf32) <- (1x16x11x64xf32) transpose_39 = paddle._C_ops.transpose(matmul_76, [0, 2, 1, 3]) @@ -2083,19 +1955,20 @@ def forward( # pd_op.reshape: (1x11x1024xf32) <- (1x11x16x64xf32, 3xi64) reshape_39 = paddle._C_ops.reshape(transpose_39, full_int_array_2) + del transpose_39 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_77 = paddle._C_ops.matmul(reshape_39, parameter_171, False, False) - del parameter_171 + del parameter_171, reshape_39 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_88 = paddle._C_ops.add(matmul_77, parameter_170) - del parameter_170 + del matmul_77, parameter_170 # pd_op.dropout: (1x11x1024xf32, 1x11x1024xui8) <- (1x11x1024xf32, None, 1xf32) dropout_58, dropout_59 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_88, None, full_5, False, "upscale_in_train", 0, False + add_88, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -2103,6 +1976,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_89 = paddle._C_ops.add(layer_norm_54, dropout_58) + del dropout_58, layer_norm_54 # pd_op.layer_norm: (1x11x1024xf32, 1x11xf32, 1x11xf32) <- (1x11x1024xf32, 1024xf32, 1024xf32) layer_norm_57, layer_norm_58, layer_norm_59 = (lambda x, f: f(x))( @@ -2111,7 +1985,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_164, parameter_165 + del add_89, parameter_164, parameter_165 # pd_op.matmul: (1x11x4096xf32) <- (1x11x1024xf32, 1024x4096xf32) matmul_78 = paddle._C_ops.matmul(layer_norm_57, parameter_169, False, False) @@ -2119,23 +1993,24 @@ def forward( # pd_op.add: (1x11x4096xf32) <- (1x11x4096xf32, 4096xf32) add_90 = paddle._C_ops.add(matmul_78, parameter_168) - del parameter_168 + del matmul_78, parameter_168 # pd_op.gelu: (1x11x4096xf32) <- (1x11x4096xf32) gelu_9 = paddle._C_ops.gelu(add_90, False) + del add_90 # pd_op.matmul: (1x11x1024xf32) <- (1x11x4096xf32, 4096x1024xf32) matmul_79 = paddle._C_ops.matmul(gelu_9, parameter_167, False, False) - del parameter_167 + del gelu_9, parameter_167 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_91 = paddle._C_ops.add(matmul_79, parameter_166) - del parameter_166 + del matmul_79, parameter_166 # pd_op.dropout: (1x11x1024xf32, 1x11x1024xui8) <- (1x11x1024xf32, None, 1xf32) dropout_60, dropout_61 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_91, None, full_5, False, "upscale_in_train", 0, False + add_91, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -2143,6 +2018,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_92 = paddle._C_ops.add(layer_norm_57, dropout_60) + del dropout_60, layer_norm_57 # pd_op.layer_norm: (1x11x1024xf32, 1x11xf32, 1x11xf32) <- (1x11x1024xf32, 1024xf32, 1024xf32) layer_norm_60, layer_norm_61, layer_norm_62 = (lambda x, f: f(x))( @@ -2151,7 +2027,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_162, parameter_163 + del add_92, parameter_162, parameter_163 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_80 = paddle._C_ops.matmul(layer_norm_60, parameter_161, False, False) @@ -2159,10 +2035,11 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_93 = paddle._C_ops.add(matmul_80, parameter_160) - del parameter_160 + del matmul_80, parameter_160 # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_40 = paddle._C_ops.reshape(add_93, full_int_array_1) + del add_93 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_40 = paddle._C_ops.transpose(reshape_40, [0, 2, 1, 3]) @@ -2174,7 +2051,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_94 = paddle._C_ops.add(matmul_81, parameter_158) - del parameter_158 + del matmul_81, parameter_158 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_82 = paddle._C_ops.matmul(layer_norm_60, parameter_157, False, False) @@ -2182,10 +2059,11 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_95 = paddle._C_ops.add(matmul_82, parameter_156) - del parameter_156 + del matmul_82, parameter_156 # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_41 = paddle._C_ops.reshape(add_94, full_int_array_1) + del add_94 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_41 = paddle._C_ops.transpose(reshape_41, [0, 2, 1, 3]) @@ -2193,6 +2071,7 @@ def forward( # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_42 = paddle._C_ops.reshape(add_95, full_int_array_1) + del add_95 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_42 = paddle._C_ops.transpose(reshape_42, [0, 2, 1, 3]) @@ -2204,9 +2083,11 @@ def forward( # pd_op.matmul: (1x16x11x11xf32) <- (1x16x11x64xf32, 1x16x11x64xf32) matmul_83 = paddle._C_ops.matmul(scale_12, transpose_41, False, True) + del scale_12, transpose_41 # pd_op.add: (1x16x11x11xf32) <- (1x16x11x11xf32, 1x1x1x11xf32) add_96 = paddle._C_ops.add(matmul_83, unsqueeze_0) + del matmul_83 # pd_op.softmax: (1x16x11x11xf32) <- (1x16x11x11xf32) softmax_10 = paddle._C_ops.softmax(add_96, -1) @@ -2215,13 +2096,15 @@ def forward( # pd_op.dropout: (1x16x11x11xf32, 1x16x11x11xui8) <- (1x16x11x11xf32, None, 1xf32) dropout_62, dropout_63 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_10, None, full_5, False, "upscale_in_train", 0, False + softmax_10, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_10 # pd_op.matmul: (1x16x11x64xf32) <- (1x16x11x11xf32, 1x16x11x64xf32) matmul_84 = paddle._C_ops.matmul(dropout_62, transpose_42, False, False) + del dropout_62, transpose_42 # pd_op.transpose: (1x11x16x64xf32) <- (1x16x11x64xf32) transpose_43 = paddle._C_ops.transpose(matmul_84, [0, 2, 1, 3]) @@ -2229,19 +2112,20 @@ def forward( # pd_op.reshape: (1x11x1024xf32) <- (1x11x16x64xf32, 3xi64) reshape_43 = paddle._C_ops.reshape(transpose_43, full_int_array_2) + del transpose_43 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_85 = paddle._C_ops.matmul(reshape_43, parameter_155, False, False) - del parameter_155 + del parameter_155, reshape_43 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_97 = paddle._C_ops.add(matmul_85, parameter_154) - del parameter_154 + del matmul_85, parameter_154 # pd_op.dropout: (1x11x1024xf32, 1x11x1024xui8) <- (1x11x1024xf32, None, 1xf32) dropout_64, dropout_65 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_97, None, full_5, False, "upscale_in_train", 0, False + add_97, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -2249,6 +2133,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_98 = paddle._C_ops.add(layer_norm_60, dropout_64) + del dropout_64, layer_norm_60 # pd_op.layer_norm: (1x11x1024xf32, 1x11xf32, 1x11xf32) <- (1x11x1024xf32, 1024xf32, 1024xf32) layer_norm_63, layer_norm_64, layer_norm_65 = (lambda x, f: f(x))( @@ -2257,7 +2142,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_148, parameter_149 + del add_98, parameter_148, parameter_149 # pd_op.matmul: (1x11x4096xf32) <- (1x11x1024xf32, 1024x4096xf32) matmul_86 = paddle._C_ops.matmul(layer_norm_63, parameter_153, False, False) @@ -2265,23 +2150,24 @@ def forward( # pd_op.add: (1x11x4096xf32) <- (1x11x4096xf32, 4096xf32) add_99 = paddle._C_ops.add(matmul_86, parameter_152) - del parameter_152 + del matmul_86, parameter_152 # pd_op.gelu: (1x11x4096xf32) <- (1x11x4096xf32) gelu_10 = paddle._C_ops.gelu(add_99, False) + del add_99 # pd_op.matmul: (1x11x1024xf32) <- (1x11x4096xf32, 4096x1024xf32) matmul_87 = paddle._C_ops.matmul(gelu_10, parameter_151, False, False) - del parameter_151 + del gelu_10, parameter_151 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_100 = paddle._C_ops.add(matmul_87, parameter_150) - del parameter_150 + del matmul_87, parameter_150 # pd_op.dropout: (1x11x1024xf32, 1x11x1024xui8) <- (1x11x1024xf32, None, 1xf32) dropout_66, dropout_67 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_100, None, full_5, False, "upscale_in_train", 0, False + add_100, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -2289,6 +2175,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_101 = paddle._C_ops.add(layer_norm_63, dropout_66) + del dropout_66, layer_norm_63 # pd_op.layer_norm: (1x11x1024xf32, 1x11xf32, 1x11xf32) <- (1x11x1024xf32, 1024xf32, 1024xf32) layer_norm_66, layer_norm_67, layer_norm_68 = (lambda x, f: f(x))( @@ -2297,7 +2184,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_146, parameter_147 + del add_101, parameter_146, parameter_147 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_88 = paddle._C_ops.matmul(layer_norm_66, parameter_145, False, False) @@ -2305,10 +2192,11 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_102 = paddle._C_ops.add(matmul_88, parameter_144) - del parameter_144 + del matmul_88, parameter_144 # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_44 = paddle._C_ops.reshape(add_102, full_int_array_1) + del add_102 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_44 = paddle._C_ops.transpose(reshape_44, [0, 2, 1, 3]) @@ -2320,7 +2208,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_103 = paddle._C_ops.add(matmul_89, parameter_142) - del parameter_142 + del matmul_89, parameter_142 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_90 = paddle._C_ops.matmul(layer_norm_66, parameter_141, False, False) @@ -2328,10 +2216,11 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_104 = paddle._C_ops.add(matmul_90, parameter_140) - del parameter_140 + del matmul_90, parameter_140 # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_45 = paddle._C_ops.reshape(add_103, full_int_array_1) + del add_103 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_45 = paddle._C_ops.transpose(reshape_45, [0, 2, 1, 3]) @@ -2339,6 +2228,7 @@ def forward( # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_46 = paddle._C_ops.reshape(add_104, full_int_array_1) + del add_104 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_46 = paddle._C_ops.transpose(reshape_46, [0, 2, 1, 3]) @@ -2350,9 +2240,11 @@ def forward( # pd_op.matmul: (1x16x11x11xf32) <- (1x16x11x64xf32, 1x16x11x64xf32) matmul_91 = paddle._C_ops.matmul(scale_13, transpose_45, False, True) + del scale_13, transpose_45 # pd_op.add: (1x16x11x11xf32) <- (1x16x11x11xf32, 1x1x1x11xf32) add_105 = paddle._C_ops.add(matmul_91, unsqueeze_0) + del matmul_91 # pd_op.softmax: (1x16x11x11xf32) <- (1x16x11x11xf32) softmax_11 = paddle._C_ops.softmax(add_105, -1) @@ -2361,13 +2253,15 @@ def forward( # pd_op.dropout: (1x16x11x11xf32, 1x16x11x11xui8) <- (1x16x11x11xf32, None, 1xf32) dropout_68, dropout_69 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_11, None, full_5, False, "upscale_in_train", 0, False + softmax_11, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_11 # pd_op.matmul: (1x16x11x64xf32) <- (1x16x11x11xf32, 1x16x11x64xf32) matmul_92 = paddle._C_ops.matmul(dropout_68, transpose_46, False, False) + del dropout_68, transpose_46 # pd_op.transpose: (1x11x16x64xf32) <- (1x16x11x64xf32) transpose_47 = paddle._C_ops.transpose(matmul_92, [0, 2, 1, 3]) @@ -2375,19 +2269,20 @@ def forward( # pd_op.reshape: (1x11x1024xf32) <- (1x11x16x64xf32, 3xi64) reshape_47 = paddle._C_ops.reshape(transpose_47, full_int_array_2) + del transpose_47 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_93 = paddle._C_ops.matmul(reshape_47, parameter_139, False, False) - del parameter_139 + del parameter_139, reshape_47 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_106 = paddle._C_ops.add(matmul_93, parameter_138) - del parameter_138 + del matmul_93, parameter_138 # pd_op.dropout: (1x11x1024xf32, 1x11x1024xui8) <- (1x11x1024xf32, None, 1xf32) dropout_70, dropout_71 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_106, None, full_5, False, "upscale_in_train", 0, False + add_106, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -2395,6 +2290,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_107 = paddle._C_ops.add(layer_norm_66, dropout_70) + del dropout_70, layer_norm_66 # pd_op.layer_norm: (1x11x1024xf32, 1x11xf32, 1x11xf32) <- (1x11x1024xf32, 1024xf32, 1024xf32) layer_norm_69, layer_norm_70, layer_norm_71 = (lambda x, f: f(x))( @@ -2403,7 +2299,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_132, parameter_133 + del add_107, parameter_132, parameter_133 # pd_op.matmul: (1x11x4096xf32) <- (1x11x1024xf32, 1024x4096xf32) matmul_94 = paddle._C_ops.matmul(layer_norm_69, parameter_137, False, False) @@ -2411,23 +2307,24 @@ def forward( # pd_op.add: (1x11x4096xf32) <- (1x11x4096xf32, 4096xf32) add_108 = paddle._C_ops.add(matmul_94, parameter_136) - del parameter_136 + del matmul_94, parameter_136 # pd_op.gelu: (1x11x4096xf32) <- (1x11x4096xf32) gelu_11 = paddle._C_ops.gelu(add_108, False) + del add_108 # pd_op.matmul: (1x11x1024xf32) <- (1x11x4096xf32, 4096x1024xf32) matmul_95 = paddle._C_ops.matmul(gelu_11, parameter_135, False, False) - del parameter_135 + del gelu_11, parameter_135 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_109 = paddle._C_ops.add(matmul_95, parameter_134) - del parameter_134 + del matmul_95, parameter_134 # pd_op.dropout: (1x11x1024xf32, 1x11x1024xui8) <- (1x11x1024xf32, None, 1xf32) dropout_72, dropout_73 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_109, None, full_5, False, "upscale_in_train", 0, False + add_109, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -2435,6 +2332,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_110 = paddle._C_ops.add(layer_norm_69, dropout_72) + del dropout_72, layer_norm_69 # pd_op.layer_norm: (1x11x1024xf32, 1x11xf32, 1x11xf32) <- (1x11x1024xf32, 1024xf32, 1024xf32) layer_norm_72, layer_norm_73, layer_norm_74 = (lambda x, f: f(x))( @@ -2443,7 +2341,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_130, parameter_131 + del add_110, parameter_130, parameter_131 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_96 = paddle._C_ops.matmul(layer_norm_72, parameter_129, False, False) @@ -2451,10 +2349,11 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_111 = paddle._C_ops.add(matmul_96, parameter_128) - del parameter_128 + del matmul_96, parameter_128 # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_48 = paddle._C_ops.reshape(add_111, full_int_array_1) + del add_111 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_48 = paddle._C_ops.transpose(reshape_48, [0, 2, 1, 3]) @@ -2466,7 +2365,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_112 = paddle._C_ops.add(matmul_97, parameter_126) - del parameter_126 + del matmul_97, parameter_126 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_98 = paddle._C_ops.matmul(layer_norm_72, parameter_125, False, False) @@ -2474,10 +2373,11 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_113 = paddle._C_ops.add(matmul_98, parameter_124) - del parameter_124 + del matmul_98, parameter_124 # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_49 = paddle._C_ops.reshape(add_112, full_int_array_1) + del add_112 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_49 = paddle._C_ops.transpose(reshape_49, [0, 2, 1, 3]) @@ -2485,6 +2385,7 @@ def forward( # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_50 = paddle._C_ops.reshape(add_113, full_int_array_1) + del add_113 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_50 = paddle._C_ops.transpose(reshape_50, [0, 2, 1, 3]) @@ -2496,9 +2397,11 @@ def forward( # pd_op.matmul: (1x16x11x11xf32) <- (1x16x11x64xf32, 1x16x11x64xf32) matmul_99 = paddle._C_ops.matmul(scale_14, transpose_49, False, True) + del scale_14, transpose_49 # pd_op.add: (1x16x11x11xf32) <- (1x16x11x11xf32, 1x1x1x11xf32) add_114 = paddle._C_ops.add(matmul_99, unsqueeze_0) + del matmul_99 # pd_op.softmax: (1x16x11x11xf32) <- (1x16x11x11xf32) softmax_12 = paddle._C_ops.softmax(add_114, -1) @@ -2507,13 +2410,15 @@ def forward( # pd_op.dropout: (1x16x11x11xf32, 1x16x11x11xui8) <- (1x16x11x11xf32, None, 1xf32) dropout_74, dropout_75 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_12, None, full_5, False, "upscale_in_train", 0, False + softmax_12, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_12 # pd_op.matmul: (1x16x11x64xf32) <- (1x16x11x11xf32, 1x16x11x64xf32) matmul_100 = paddle._C_ops.matmul(dropout_74, transpose_50, False, False) + del dropout_74, transpose_50 # pd_op.transpose: (1x11x16x64xf32) <- (1x16x11x64xf32) transpose_51 = paddle._C_ops.transpose(matmul_100, [0, 2, 1, 3]) @@ -2521,19 +2426,20 @@ def forward( # pd_op.reshape: (1x11x1024xf32) <- (1x11x16x64xf32, 3xi64) reshape_51 = paddle._C_ops.reshape(transpose_51, full_int_array_2) + del transpose_51 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_101 = paddle._C_ops.matmul(reshape_51, parameter_123, False, False) - del parameter_123 + del parameter_123, reshape_51 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_115 = paddle._C_ops.add(matmul_101, parameter_122) - del parameter_122 + del matmul_101, parameter_122 # pd_op.dropout: (1x11x1024xf32, 1x11x1024xui8) <- (1x11x1024xf32, None, 1xf32) dropout_76, dropout_77 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_115, None, full_5, False, "upscale_in_train", 0, False + add_115, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -2541,6 +2447,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_116 = paddle._C_ops.add(layer_norm_72, dropout_76) + del dropout_76, layer_norm_72 # pd_op.layer_norm: (1x11x1024xf32, 1x11xf32, 1x11xf32) <- (1x11x1024xf32, 1024xf32, 1024xf32) layer_norm_75, layer_norm_76, layer_norm_77 = (lambda x, f: f(x))( @@ -2549,7 +2456,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_116, parameter_117 + del add_116, parameter_116, parameter_117 # pd_op.matmul: (1x11x4096xf32) <- (1x11x1024xf32, 1024x4096xf32) matmul_102 = paddle._C_ops.matmul(layer_norm_75, parameter_121, False, False) @@ -2557,23 +2464,24 @@ def forward( # pd_op.add: (1x11x4096xf32) <- (1x11x4096xf32, 4096xf32) add_117 = paddle._C_ops.add(matmul_102, parameter_120) - del parameter_120 + del matmul_102, parameter_120 # pd_op.gelu: (1x11x4096xf32) <- (1x11x4096xf32) gelu_12 = paddle._C_ops.gelu(add_117, False) + del add_117 # pd_op.matmul: (1x11x1024xf32) <- (1x11x4096xf32, 4096x1024xf32) matmul_103 = paddle._C_ops.matmul(gelu_12, parameter_119, False, False) - del parameter_119 + del gelu_12, parameter_119 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_118 = paddle._C_ops.add(matmul_103, parameter_118) - del parameter_118 + del matmul_103, parameter_118 # pd_op.dropout: (1x11x1024xf32, 1x11x1024xui8) <- (1x11x1024xf32, None, 1xf32) dropout_78, dropout_79 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_118, None, full_5, False, "upscale_in_train", 0, False + add_118, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -2581,6 +2489,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_119 = paddle._C_ops.add(layer_norm_75, dropout_78) + del dropout_78, layer_norm_75 # pd_op.layer_norm: (1x11x1024xf32, 1x11xf32, 1x11xf32) <- (1x11x1024xf32, 1024xf32, 1024xf32) layer_norm_78, layer_norm_79, layer_norm_80 = (lambda x, f: f(x))( @@ -2589,7 +2498,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_114, parameter_115 + del add_119, parameter_114, parameter_115 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_104 = paddle._C_ops.matmul(layer_norm_78, parameter_113, False, False) @@ -2597,10 +2506,11 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_120 = paddle._C_ops.add(matmul_104, parameter_112) - del parameter_112 + del matmul_104, parameter_112 # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_52 = paddle._C_ops.reshape(add_120, full_int_array_1) + del add_120 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_52 = paddle._C_ops.transpose(reshape_52, [0, 2, 1, 3]) @@ -2612,7 +2522,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_121 = paddle._C_ops.add(matmul_105, parameter_110) - del parameter_110 + del matmul_105, parameter_110 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_106 = paddle._C_ops.matmul(layer_norm_78, parameter_109, False, False) @@ -2620,10 +2530,11 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_122 = paddle._C_ops.add(matmul_106, parameter_108) - del parameter_108 + del matmul_106, parameter_108 # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_53 = paddle._C_ops.reshape(add_121, full_int_array_1) + del add_121 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_53 = paddle._C_ops.transpose(reshape_53, [0, 2, 1, 3]) @@ -2631,6 +2542,7 @@ def forward( # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_54 = paddle._C_ops.reshape(add_122, full_int_array_1) + del add_122 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_54 = paddle._C_ops.transpose(reshape_54, [0, 2, 1, 3]) @@ -2642,9 +2554,11 @@ def forward( # pd_op.matmul: (1x16x11x11xf32) <- (1x16x11x64xf32, 1x16x11x64xf32) matmul_107 = paddle._C_ops.matmul(scale_15, transpose_53, False, True) + del scale_15, transpose_53 # pd_op.add: (1x16x11x11xf32) <- (1x16x11x11xf32, 1x1x1x11xf32) add_123 = paddle._C_ops.add(matmul_107, unsqueeze_0) + del matmul_107 # pd_op.softmax: (1x16x11x11xf32) <- (1x16x11x11xf32) softmax_13 = paddle._C_ops.softmax(add_123, -1) @@ -2653,13 +2567,15 @@ def forward( # pd_op.dropout: (1x16x11x11xf32, 1x16x11x11xui8) <- (1x16x11x11xf32, None, 1xf32) dropout_80, dropout_81 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_13, None, full_5, False, "upscale_in_train", 0, False + softmax_13, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_13 # pd_op.matmul: (1x16x11x64xf32) <- (1x16x11x11xf32, 1x16x11x64xf32) matmul_108 = paddle._C_ops.matmul(dropout_80, transpose_54, False, False) + del dropout_80, transpose_54 # pd_op.transpose: (1x11x16x64xf32) <- (1x16x11x64xf32) transpose_55 = paddle._C_ops.transpose(matmul_108, [0, 2, 1, 3]) @@ -2667,19 +2583,20 @@ def forward( # pd_op.reshape: (1x11x1024xf32) <- (1x11x16x64xf32, 3xi64) reshape_55 = paddle._C_ops.reshape(transpose_55, full_int_array_2) + del transpose_55 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_109 = paddle._C_ops.matmul(reshape_55, parameter_107, False, False) - del parameter_107 + del parameter_107, reshape_55 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_124 = paddle._C_ops.add(matmul_109, parameter_106) - del parameter_106 + del matmul_109, parameter_106 # pd_op.dropout: (1x11x1024xf32, 1x11x1024xui8) <- (1x11x1024xf32, None, 1xf32) dropout_82, dropout_83 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_124, None, full_5, False, "upscale_in_train", 0, False + add_124, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -2687,6 +2604,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_125 = paddle._C_ops.add(layer_norm_78, dropout_82) + del dropout_82, layer_norm_78 # pd_op.layer_norm: (1x11x1024xf32, 1x11xf32, 1x11xf32) <- (1x11x1024xf32, 1024xf32, 1024xf32) layer_norm_81, layer_norm_82, layer_norm_83 = (lambda x, f: f(x))( @@ -2695,7 +2613,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_100, parameter_101 + del add_125, parameter_100, parameter_101 # pd_op.matmul: (1x11x4096xf32) <- (1x11x1024xf32, 1024x4096xf32) matmul_110 = paddle._C_ops.matmul(layer_norm_81, parameter_105, False, False) @@ -2703,23 +2621,24 @@ def forward( # pd_op.add: (1x11x4096xf32) <- (1x11x4096xf32, 4096xf32) add_126 = paddle._C_ops.add(matmul_110, parameter_104) - del parameter_104 + del matmul_110, parameter_104 # pd_op.gelu: (1x11x4096xf32) <- (1x11x4096xf32) gelu_13 = paddle._C_ops.gelu(add_126, False) + del add_126 # pd_op.matmul: (1x11x1024xf32) <- (1x11x4096xf32, 4096x1024xf32) matmul_111 = paddle._C_ops.matmul(gelu_13, parameter_103, False, False) - del parameter_103 + del gelu_13, parameter_103 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_127 = paddle._C_ops.add(matmul_111, parameter_102) - del parameter_102 + del matmul_111, parameter_102 # pd_op.dropout: (1x11x1024xf32, 1x11x1024xui8) <- (1x11x1024xf32, None, 1xf32) dropout_84, dropout_85 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_127, None, full_5, False, "upscale_in_train", 0, False + add_127, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -2727,6 +2646,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_128 = paddle._C_ops.add(layer_norm_81, dropout_84) + del dropout_84, layer_norm_81 # pd_op.layer_norm: (1x11x1024xf32, 1x11xf32, 1x11xf32) <- (1x11x1024xf32, 1024xf32, 1024xf32) layer_norm_84, layer_norm_85, layer_norm_86 = (lambda x, f: f(x))( @@ -2735,7 +2655,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_98, parameter_99 + del add_128, parameter_98, parameter_99 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_112 = paddle._C_ops.matmul(layer_norm_84, parameter_97, False, False) @@ -2743,10 +2663,11 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_129 = paddle._C_ops.add(matmul_112, parameter_96) - del parameter_96 + del matmul_112, parameter_96 # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_56 = paddle._C_ops.reshape(add_129, full_int_array_1) + del add_129 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_56 = paddle._C_ops.transpose(reshape_56, [0, 2, 1, 3]) @@ -2758,7 +2679,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_130 = paddle._C_ops.add(matmul_113, parameter_94) - del parameter_94 + del matmul_113, parameter_94 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_114 = paddle._C_ops.matmul(layer_norm_84, parameter_93, False, False) @@ -2766,10 +2687,11 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_131 = paddle._C_ops.add(matmul_114, parameter_92) - del parameter_92 + del matmul_114, parameter_92 # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_57 = paddle._C_ops.reshape(add_130, full_int_array_1) + del add_130 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_57 = paddle._C_ops.transpose(reshape_57, [0, 2, 1, 3]) @@ -2777,6 +2699,7 @@ def forward( # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_58 = paddle._C_ops.reshape(add_131, full_int_array_1) + del add_131 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_58 = paddle._C_ops.transpose(reshape_58, [0, 2, 1, 3]) @@ -2788,9 +2711,11 @@ def forward( # pd_op.matmul: (1x16x11x11xf32) <- (1x16x11x64xf32, 1x16x11x64xf32) matmul_115 = paddle._C_ops.matmul(scale_16, transpose_57, False, True) + del scale_16, transpose_57 # pd_op.add: (1x16x11x11xf32) <- (1x16x11x11xf32, 1x1x1x11xf32) add_132 = paddle._C_ops.add(matmul_115, unsqueeze_0) + del matmul_115 # pd_op.softmax: (1x16x11x11xf32) <- (1x16x11x11xf32) softmax_14 = paddle._C_ops.softmax(add_132, -1) @@ -2799,13 +2724,15 @@ def forward( # pd_op.dropout: (1x16x11x11xf32, 1x16x11x11xui8) <- (1x16x11x11xf32, None, 1xf32) dropout_86, dropout_87 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_14, None, full_5, False, "upscale_in_train", 0, False + softmax_14, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_14 # pd_op.matmul: (1x16x11x64xf32) <- (1x16x11x11xf32, 1x16x11x64xf32) matmul_116 = paddle._C_ops.matmul(dropout_86, transpose_58, False, False) + del dropout_86, transpose_58 # pd_op.transpose: (1x11x16x64xf32) <- (1x16x11x64xf32) transpose_59 = paddle._C_ops.transpose(matmul_116, [0, 2, 1, 3]) @@ -2813,19 +2740,20 @@ def forward( # pd_op.reshape: (1x11x1024xf32) <- (1x11x16x64xf32, 3xi64) reshape_59 = paddle._C_ops.reshape(transpose_59, full_int_array_2) + del transpose_59 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_117 = paddle._C_ops.matmul(reshape_59, parameter_91, False, False) - del parameter_91 + del parameter_91, reshape_59 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_133 = paddle._C_ops.add(matmul_117, parameter_90) - del parameter_90 + del matmul_117, parameter_90 # pd_op.dropout: (1x11x1024xf32, 1x11x1024xui8) <- (1x11x1024xf32, None, 1xf32) dropout_88, dropout_89 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_133, None, full_5, False, "upscale_in_train", 0, False + add_133, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -2833,6 +2761,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_134 = paddle._C_ops.add(layer_norm_84, dropout_88) + del dropout_88, layer_norm_84 # pd_op.layer_norm: (1x11x1024xf32, 1x11xf32, 1x11xf32) <- (1x11x1024xf32, 1024xf32, 1024xf32) layer_norm_87, layer_norm_88, layer_norm_89 = (lambda x, f: f(x))( @@ -2841,7 +2770,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_84, parameter_85 + del add_134, parameter_84, parameter_85 # pd_op.matmul: (1x11x4096xf32) <- (1x11x1024xf32, 1024x4096xf32) matmul_118 = paddle._C_ops.matmul(layer_norm_87, parameter_89, False, False) @@ -2849,23 +2778,24 @@ def forward( # pd_op.add: (1x11x4096xf32) <- (1x11x4096xf32, 4096xf32) add_135 = paddle._C_ops.add(matmul_118, parameter_88) - del parameter_88 + del matmul_118, parameter_88 # pd_op.gelu: (1x11x4096xf32) <- (1x11x4096xf32) gelu_14 = paddle._C_ops.gelu(add_135, False) + del add_135 # pd_op.matmul: (1x11x1024xf32) <- (1x11x4096xf32, 4096x1024xf32) matmul_119 = paddle._C_ops.matmul(gelu_14, parameter_87, False, False) - del parameter_87 + del gelu_14, parameter_87 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_136 = paddle._C_ops.add(matmul_119, parameter_86) - del parameter_86 + del matmul_119, parameter_86 # pd_op.dropout: (1x11x1024xf32, 1x11x1024xui8) <- (1x11x1024xf32, None, 1xf32) dropout_90, dropout_91 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_136, None, full_5, False, "upscale_in_train", 0, False + add_136, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -2873,6 +2803,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_137 = paddle._C_ops.add(layer_norm_87, dropout_90) + del dropout_90, layer_norm_87 # pd_op.layer_norm: (1x11x1024xf32, 1x11xf32, 1x11xf32) <- (1x11x1024xf32, 1024xf32, 1024xf32) layer_norm_90, layer_norm_91, layer_norm_92 = (lambda x, f: f(x))( @@ -2881,7 +2812,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_82, parameter_83 + del add_137, parameter_82, parameter_83 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_120 = paddle._C_ops.matmul(layer_norm_90, parameter_81, False, False) @@ -2889,10 +2820,11 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_138 = paddle._C_ops.add(matmul_120, parameter_80) - del parameter_80 + del matmul_120, parameter_80 # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_60 = paddle._C_ops.reshape(add_138, full_int_array_1) + del add_138 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_60 = paddle._C_ops.transpose(reshape_60, [0, 2, 1, 3]) @@ -2904,7 +2836,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_139 = paddle._C_ops.add(matmul_121, parameter_78) - del parameter_78 + del matmul_121, parameter_78 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_122 = paddle._C_ops.matmul(layer_norm_90, parameter_77, False, False) @@ -2912,10 +2844,11 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_140 = paddle._C_ops.add(matmul_122, parameter_76) - del parameter_76 + del matmul_122, parameter_76 # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_61 = paddle._C_ops.reshape(add_139, full_int_array_1) + del add_139 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_61 = paddle._C_ops.transpose(reshape_61, [0, 2, 1, 3]) @@ -2923,6 +2856,7 @@ def forward( # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_62 = paddle._C_ops.reshape(add_140, full_int_array_1) + del add_140 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_62 = paddle._C_ops.transpose(reshape_62, [0, 2, 1, 3]) @@ -2934,9 +2868,11 @@ def forward( # pd_op.matmul: (1x16x11x11xf32) <- (1x16x11x64xf32, 1x16x11x64xf32) matmul_123 = paddle._C_ops.matmul(scale_17, transpose_61, False, True) + del scale_17, transpose_61 # pd_op.add: (1x16x11x11xf32) <- (1x16x11x11xf32, 1x1x1x11xf32) add_141 = paddle._C_ops.add(matmul_123, unsqueeze_0) + del matmul_123 # pd_op.softmax: (1x16x11x11xf32) <- (1x16x11x11xf32) softmax_15 = paddle._C_ops.softmax(add_141, -1) @@ -2945,13 +2881,15 @@ def forward( # pd_op.dropout: (1x16x11x11xf32, 1x16x11x11xui8) <- (1x16x11x11xf32, None, 1xf32) dropout_92, dropout_93 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_15, None, full_5, False, "upscale_in_train", 0, False + softmax_15, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_15 # pd_op.matmul: (1x16x11x64xf32) <- (1x16x11x11xf32, 1x16x11x64xf32) matmul_124 = paddle._C_ops.matmul(dropout_92, transpose_62, False, False) + del dropout_92, transpose_62 # pd_op.transpose: (1x11x16x64xf32) <- (1x16x11x64xf32) transpose_63 = paddle._C_ops.transpose(matmul_124, [0, 2, 1, 3]) @@ -2959,19 +2897,20 @@ def forward( # pd_op.reshape: (1x11x1024xf32) <- (1x11x16x64xf32, 3xi64) reshape_63 = paddle._C_ops.reshape(transpose_63, full_int_array_2) + del transpose_63 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_125 = paddle._C_ops.matmul(reshape_63, parameter_75, False, False) - del parameter_75 + del parameter_75, reshape_63 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_142 = paddle._C_ops.add(matmul_125, parameter_74) - del parameter_74 + del matmul_125, parameter_74 # pd_op.dropout: (1x11x1024xf32, 1x11x1024xui8) <- (1x11x1024xf32, None, 1xf32) dropout_94, dropout_95 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_142, None, full_5, False, "upscale_in_train", 0, False + add_142, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -2979,6 +2918,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_143 = paddle._C_ops.add(layer_norm_90, dropout_94) + del dropout_94, layer_norm_90 # pd_op.layer_norm: (1x11x1024xf32, 1x11xf32, 1x11xf32) <- (1x11x1024xf32, 1024xf32, 1024xf32) layer_norm_93, layer_norm_94, layer_norm_95 = (lambda x, f: f(x))( @@ -2987,7 +2927,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_68, parameter_69 + del add_143, parameter_68, parameter_69 # pd_op.matmul: (1x11x4096xf32) <- (1x11x1024xf32, 1024x4096xf32) matmul_126 = paddle._C_ops.matmul(layer_norm_93, parameter_73, False, False) @@ -2995,23 +2935,24 @@ def forward( # pd_op.add: (1x11x4096xf32) <- (1x11x4096xf32, 4096xf32) add_144 = paddle._C_ops.add(matmul_126, parameter_72) - del parameter_72 + del matmul_126, parameter_72 # pd_op.gelu: (1x11x4096xf32) <- (1x11x4096xf32) gelu_15 = paddle._C_ops.gelu(add_144, False) + del add_144 # pd_op.matmul: (1x11x1024xf32) <- (1x11x4096xf32, 4096x1024xf32) matmul_127 = paddle._C_ops.matmul(gelu_15, parameter_71, False, False) - del parameter_71 + del gelu_15, parameter_71 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_145 = paddle._C_ops.add(matmul_127, parameter_70) - del parameter_70 + del matmul_127, parameter_70 # pd_op.dropout: (1x11x1024xf32, 1x11x1024xui8) <- (1x11x1024xf32, None, 1xf32) dropout_96, dropout_97 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_145, None, full_5, False, "upscale_in_train", 0, False + add_145, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -3019,6 +2960,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_146 = paddle._C_ops.add(layer_norm_93, dropout_96) + del dropout_96, layer_norm_93 # pd_op.layer_norm: (1x11x1024xf32, 1x11xf32, 1x11xf32) <- (1x11x1024xf32, 1024xf32, 1024xf32) layer_norm_96, layer_norm_97, layer_norm_98 = (lambda x, f: f(x))( @@ -3027,7 +2969,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_66, parameter_67 + del add_146, parameter_66, parameter_67 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_128 = paddle._C_ops.matmul(layer_norm_96, parameter_65, False, False) @@ -3035,10 +2977,11 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_147 = paddle._C_ops.add(matmul_128, parameter_64) - del parameter_64 + del matmul_128, parameter_64 # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_64 = paddle._C_ops.reshape(add_147, full_int_array_1) + del add_147 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_64 = paddle._C_ops.transpose(reshape_64, [0, 2, 1, 3]) @@ -3050,7 +2993,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_148 = paddle._C_ops.add(matmul_129, parameter_62) - del parameter_62 + del matmul_129, parameter_62 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_130 = paddle._C_ops.matmul(layer_norm_96, parameter_61, False, False) @@ -3058,10 +3001,11 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_149 = paddle._C_ops.add(matmul_130, parameter_60) - del parameter_60 + del matmul_130, parameter_60 # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_65 = paddle._C_ops.reshape(add_148, full_int_array_1) + del add_148 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_65 = paddle._C_ops.transpose(reshape_65, [0, 2, 1, 3]) @@ -3069,6 +3013,7 @@ def forward( # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_66 = paddle._C_ops.reshape(add_149, full_int_array_1) + del add_149 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_66 = paddle._C_ops.transpose(reshape_66, [0, 2, 1, 3]) @@ -3080,9 +3025,11 @@ def forward( # pd_op.matmul: (1x16x11x11xf32) <- (1x16x11x64xf32, 1x16x11x64xf32) matmul_131 = paddle._C_ops.matmul(scale_18, transpose_65, False, True) + del scale_18, transpose_65 # pd_op.add: (1x16x11x11xf32) <- (1x16x11x11xf32, 1x1x1x11xf32) add_150 = paddle._C_ops.add(matmul_131, unsqueeze_0) + del matmul_131 # pd_op.softmax: (1x16x11x11xf32) <- (1x16x11x11xf32) softmax_16 = paddle._C_ops.softmax(add_150, -1) @@ -3091,13 +3038,15 @@ def forward( # pd_op.dropout: (1x16x11x11xf32, 1x16x11x11xui8) <- (1x16x11x11xf32, None, 1xf32) dropout_98, dropout_99 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_16, None, full_5, False, "upscale_in_train", 0, False + softmax_16, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_16 # pd_op.matmul: (1x16x11x64xf32) <- (1x16x11x11xf32, 1x16x11x64xf32) matmul_132 = paddle._C_ops.matmul(dropout_98, transpose_66, False, False) + del dropout_98, transpose_66 # pd_op.transpose: (1x11x16x64xf32) <- (1x16x11x64xf32) transpose_67 = paddle._C_ops.transpose(matmul_132, [0, 2, 1, 3]) @@ -3105,19 +3054,20 @@ def forward( # pd_op.reshape: (1x11x1024xf32) <- (1x11x16x64xf32, 3xi64) reshape_67 = paddle._C_ops.reshape(transpose_67, full_int_array_2) + del transpose_67 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_133 = paddle._C_ops.matmul(reshape_67, parameter_59, False, False) - del parameter_59 + del parameter_59, reshape_67 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_151 = paddle._C_ops.add(matmul_133, parameter_58) - del parameter_58 + del matmul_133, parameter_58 # pd_op.dropout: (1x11x1024xf32, 1x11x1024xui8) <- (1x11x1024xf32, None, 1xf32) dropout_100, dropout_101 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_151, None, full_5, False, "upscale_in_train", 0, False + add_151, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -3125,6 +3075,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_152 = paddle._C_ops.add(layer_norm_96, dropout_100) + del dropout_100, layer_norm_96 # pd_op.layer_norm: (1x11x1024xf32, 1x11xf32, 1x11xf32) <- (1x11x1024xf32, 1024xf32, 1024xf32) layer_norm_99, layer_norm_100, layer_norm_101 = (lambda x, f: f(x))( @@ -3133,7 +3084,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_52, parameter_53 + del add_152, parameter_52, parameter_53 # pd_op.matmul: (1x11x4096xf32) <- (1x11x1024xf32, 1024x4096xf32) matmul_134 = paddle._C_ops.matmul(layer_norm_99, parameter_57, False, False) @@ -3141,23 +3092,24 @@ def forward( # pd_op.add: (1x11x4096xf32) <- (1x11x4096xf32, 4096xf32) add_153 = paddle._C_ops.add(matmul_134, parameter_56) - del parameter_56 + del matmul_134, parameter_56 # pd_op.gelu: (1x11x4096xf32) <- (1x11x4096xf32) gelu_16 = paddle._C_ops.gelu(add_153, False) + del add_153 # pd_op.matmul: (1x11x1024xf32) <- (1x11x4096xf32, 4096x1024xf32) matmul_135 = paddle._C_ops.matmul(gelu_16, parameter_55, False, False) - del parameter_55 + del gelu_16, parameter_55 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_154 = paddle._C_ops.add(matmul_135, parameter_54) - del parameter_54 + del matmul_135, parameter_54 # pd_op.dropout: (1x11x1024xf32, 1x11x1024xui8) <- (1x11x1024xf32, None, 1xf32) dropout_102, dropout_103 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_154, None, full_5, False, "upscale_in_train", 0, False + add_154, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -3165,6 +3117,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_155 = paddle._C_ops.add(layer_norm_99, dropout_102) + del dropout_102, layer_norm_99 # pd_op.layer_norm: (1x11x1024xf32, 1x11xf32, 1x11xf32) <- (1x11x1024xf32, 1024xf32, 1024xf32) layer_norm_102, layer_norm_103, layer_norm_104 = (lambda x, f: f(x))( @@ -3173,7 +3126,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_50, parameter_51 + del add_155, parameter_50, parameter_51 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_136 = paddle._C_ops.matmul(layer_norm_102, parameter_49, False, False) @@ -3181,10 +3134,11 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_156 = paddle._C_ops.add(matmul_136, parameter_48) - del parameter_48 + del matmul_136, parameter_48 # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_68 = paddle._C_ops.reshape(add_156, full_int_array_1) + del add_156 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_68 = paddle._C_ops.transpose(reshape_68, [0, 2, 1, 3]) @@ -3196,7 +3150,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_157 = paddle._C_ops.add(matmul_137, parameter_46) - del parameter_46 + del matmul_137, parameter_46 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_138 = paddle._C_ops.matmul(layer_norm_102, parameter_45, False, False) @@ -3204,10 +3158,11 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_158 = paddle._C_ops.add(matmul_138, parameter_44) - del parameter_44 + del matmul_138, parameter_44 # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_69 = paddle._C_ops.reshape(add_157, full_int_array_1) + del add_157 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_69 = paddle._C_ops.transpose(reshape_69, [0, 2, 1, 3]) @@ -3215,6 +3170,7 @@ def forward( # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_70 = paddle._C_ops.reshape(add_158, full_int_array_1) + del add_158 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_70 = paddle._C_ops.transpose(reshape_70, [0, 2, 1, 3]) @@ -3226,9 +3182,11 @@ def forward( # pd_op.matmul: (1x16x11x11xf32) <- (1x16x11x64xf32, 1x16x11x64xf32) matmul_139 = paddle._C_ops.matmul(scale_19, transpose_69, False, True) + del scale_19, transpose_69 # pd_op.add: (1x16x11x11xf32) <- (1x16x11x11xf32, 1x1x1x11xf32) add_159 = paddle._C_ops.add(matmul_139, unsqueeze_0) + del matmul_139 # pd_op.softmax: (1x16x11x11xf32) <- (1x16x11x11xf32) softmax_17 = paddle._C_ops.softmax(add_159, -1) @@ -3237,13 +3195,15 @@ def forward( # pd_op.dropout: (1x16x11x11xf32, 1x16x11x11xui8) <- (1x16x11x11xf32, None, 1xf32) dropout_104, dropout_105 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_17, None, full_5, False, "upscale_in_train", 0, False + softmax_17, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_17 # pd_op.matmul: (1x16x11x64xf32) <- (1x16x11x11xf32, 1x16x11x64xf32) matmul_140 = paddle._C_ops.matmul(dropout_104, transpose_70, False, False) + del dropout_104, transpose_70 # pd_op.transpose: (1x11x16x64xf32) <- (1x16x11x64xf32) transpose_71 = paddle._C_ops.transpose(matmul_140, [0, 2, 1, 3]) @@ -3251,19 +3211,20 @@ def forward( # pd_op.reshape: (1x11x1024xf32) <- (1x11x16x64xf32, 3xi64) reshape_71 = paddle._C_ops.reshape(transpose_71, full_int_array_2) + del transpose_71 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_141 = paddle._C_ops.matmul(reshape_71, parameter_43, False, False) - del parameter_43 + del parameter_43, reshape_71 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_160 = paddle._C_ops.add(matmul_141, parameter_42) - del parameter_42 + del matmul_141, parameter_42 # pd_op.dropout: (1x11x1024xf32, 1x11x1024xui8) <- (1x11x1024xf32, None, 1xf32) dropout_106, dropout_107 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_160, None, full_5, False, "upscale_in_train", 0, False + add_160, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -3271,6 +3232,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_161 = paddle._C_ops.add(layer_norm_102, dropout_106) + del dropout_106, layer_norm_102 # pd_op.layer_norm: (1x11x1024xf32, 1x11xf32, 1x11xf32) <- (1x11x1024xf32, 1024xf32, 1024xf32) layer_norm_105, layer_norm_106, layer_norm_107 = (lambda x, f: f(x))( @@ -3279,7 +3241,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_36, parameter_37 + del add_161, parameter_36, parameter_37 # pd_op.matmul: (1x11x4096xf32) <- (1x11x1024xf32, 1024x4096xf32) matmul_142 = paddle._C_ops.matmul(layer_norm_105, parameter_41, False, False) @@ -3287,23 +3249,24 @@ def forward( # pd_op.add: (1x11x4096xf32) <- (1x11x4096xf32, 4096xf32) add_162 = paddle._C_ops.add(matmul_142, parameter_40) - del parameter_40 + del matmul_142, parameter_40 # pd_op.gelu: (1x11x4096xf32) <- (1x11x4096xf32) gelu_17 = paddle._C_ops.gelu(add_162, False) + del add_162 # pd_op.matmul: (1x11x1024xf32) <- (1x11x4096xf32, 4096x1024xf32) matmul_143 = paddle._C_ops.matmul(gelu_17, parameter_39, False, False) - del parameter_39 + del gelu_17, parameter_39 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_163 = paddle._C_ops.add(matmul_143, parameter_38) - del parameter_38 + del matmul_143, parameter_38 # pd_op.dropout: (1x11x1024xf32, 1x11x1024xui8) <- (1x11x1024xf32, None, 1xf32) dropout_108, dropout_109 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_163, None, full_5, False, "upscale_in_train", 0, False + add_163, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -3311,6 +3274,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_164 = paddle._C_ops.add(layer_norm_105, dropout_108) + del dropout_108, layer_norm_105 # pd_op.layer_norm: (1x11x1024xf32, 1x11xf32, 1x11xf32) <- (1x11x1024xf32, 1024xf32, 1024xf32) layer_norm_108, layer_norm_109, layer_norm_110 = (lambda x, f: f(x))( @@ -3319,7 +3283,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_34, parameter_35 + del add_164, parameter_34, parameter_35 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_144 = paddle._C_ops.matmul(layer_norm_108, parameter_33, False, False) @@ -3327,10 +3291,11 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_165 = paddle._C_ops.add(matmul_144, parameter_32) - del parameter_32 + del matmul_144, parameter_32 # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_72 = paddle._C_ops.reshape(add_165, full_int_array_1) + del add_165 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_72 = paddle._C_ops.transpose(reshape_72, [0, 2, 1, 3]) @@ -3342,7 +3307,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_166 = paddle._C_ops.add(matmul_145, parameter_30) - del parameter_30 + del matmul_145, parameter_30 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_146 = paddle._C_ops.matmul(layer_norm_108, parameter_29, False, False) @@ -3350,10 +3315,11 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_167 = paddle._C_ops.add(matmul_146, parameter_28) - del parameter_28 + del matmul_146, parameter_28 # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_73 = paddle._C_ops.reshape(add_166, full_int_array_1) + del add_166 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_73 = paddle._C_ops.transpose(reshape_73, [0, 2, 1, 3]) @@ -3361,6 +3327,7 @@ def forward( # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_74 = paddle._C_ops.reshape(add_167, full_int_array_1) + del add_167 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_74 = paddle._C_ops.transpose(reshape_74, [0, 2, 1, 3]) @@ -3372,9 +3339,11 @@ def forward( # pd_op.matmul: (1x16x11x11xf32) <- (1x16x11x64xf32, 1x16x11x64xf32) matmul_147 = paddle._C_ops.matmul(scale_20, transpose_73, False, True) + del scale_20, transpose_73 # pd_op.add: (1x16x11x11xf32) <- (1x16x11x11xf32, 1x1x1x11xf32) add_168 = paddle._C_ops.add(matmul_147, unsqueeze_0) + del matmul_147 # pd_op.softmax: (1x16x11x11xf32) <- (1x16x11x11xf32) softmax_18 = paddle._C_ops.softmax(add_168, -1) @@ -3383,13 +3352,15 @@ def forward( # pd_op.dropout: (1x16x11x11xf32, 1x16x11x11xui8) <- (1x16x11x11xf32, None, 1xf32) dropout_110, dropout_111 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_18, None, full_5, False, "upscale_in_train", 0, False + softmax_18, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_18 # pd_op.matmul: (1x16x11x64xf32) <- (1x16x11x11xf32, 1x16x11x64xf32) matmul_148 = paddle._C_ops.matmul(dropout_110, transpose_74, False, False) + del dropout_110, transpose_74 # pd_op.transpose: (1x11x16x64xf32) <- (1x16x11x64xf32) transpose_75 = paddle._C_ops.transpose(matmul_148, [0, 2, 1, 3]) @@ -3397,19 +3368,20 @@ def forward( # pd_op.reshape: (1x11x1024xf32) <- (1x11x16x64xf32, 3xi64) reshape_75 = paddle._C_ops.reshape(transpose_75, full_int_array_2) + del transpose_75 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_149 = paddle._C_ops.matmul(reshape_75, parameter_27, False, False) - del parameter_27 + del parameter_27, reshape_75 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_169 = paddle._C_ops.add(matmul_149, parameter_26) - del parameter_26 + del matmul_149, parameter_26 # pd_op.dropout: (1x11x1024xf32, 1x11x1024xui8) <- (1x11x1024xf32, None, 1xf32) dropout_112, dropout_113 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_169, None, full_5, False, "upscale_in_train", 0, False + add_169, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -3417,6 +3389,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_170 = paddle._C_ops.add(layer_norm_108, dropout_112) + del dropout_112, layer_norm_108 # pd_op.layer_norm: (1x11x1024xf32, 1x11xf32, 1x11xf32) <- (1x11x1024xf32, 1024xf32, 1024xf32) layer_norm_111, layer_norm_112, layer_norm_113 = (lambda x, f: f(x))( @@ -3425,7 +3398,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_20, parameter_21 + del add_170, parameter_20, parameter_21 # pd_op.matmul: (1x11x4096xf32) <- (1x11x1024xf32, 1024x4096xf32) matmul_150 = paddle._C_ops.matmul(layer_norm_111, parameter_25, False, False) @@ -3433,23 +3406,24 @@ def forward( # pd_op.add: (1x11x4096xf32) <- (1x11x4096xf32, 4096xf32) add_171 = paddle._C_ops.add(matmul_150, parameter_24) - del parameter_24 + del matmul_150, parameter_24 # pd_op.gelu: (1x11x4096xf32) <- (1x11x4096xf32) gelu_18 = paddle._C_ops.gelu(add_171, False) + del add_171 # pd_op.matmul: (1x11x1024xf32) <- (1x11x4096xf32, 4096x1024xf32) matmul_151 = paddle._C_ops.matmul(gelu_18, parameter_23, False, False) - del parameter_23 + del gelu_18, parameter_23 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_172 = paddle._C_ops.add(matmul_151, parameter_22) - del parameter_22 + del matmul_151, parameter_22 # pd_op.dropout: (1x11x1024xf32, 1x11x1024xui8) <- (1x11x1024xf32, None, 1xf32) dropout_114, dropout_115 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_172, None, full_5, False, "upscale_in_train", 0, False + add_172, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -3457,6 +3431,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_173 = paddle._C_ops.add(layer_norm_111, dropout_114) + del dropout_114, layer_norm_111 # pd_op.layer_norm: (1x11x1024xf32, 1x11xf32, 1x11xf32) <- (1x11x1024xf32, 1024xf32, 1024xf32) layer_norm_114, layer_norm_115, layer_norm_116 = (lambda x, f: f(x))( @@ -3465,7 +3440,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_18, parameter_19 + del add_173, parameter_18, parameter_19 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_152 = paddle._C_ops.matmul(layer_norm_114, parameter_17, False, False) @@ -3473,10 +3448,11 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_174 = paddle._C_ops.add(matmul_152, parameter_16) - del parameter_16 + del matmul_152, parameter_16 # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_76 = paddle._C_ops.reshape(add_174, full_int_array_1) + del add_174 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_76 = paddle._C_ops.transpose(reshape_76, [0, 2, 1, 3]) @@ -3488,7 +3464,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_175 = paddle._C_ops.add(matmul_153, parameter_14) - del parameter_14 + del matmul_153, parameter_14 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_154 = paddle._C_ops.matmul(layer_norm_114, parameter_13, False, False) @@ -3496,10 +3472,11 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_176 = paddle._C_ops.add(matmul_154, parameter_12) - del parameter_12 + del matmul_154, parameter_12 # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_77 = paddle._C_ops.reshape(add_175, full_int_array_1) + del add_175 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_77 = paddle._C_ops.transpose(reshape_77, [0, 2, 1, 3]) @@ -3507,7 +3484,7 @@ def forward( # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_78 = paddle._C_ops.reshape(add_176, full_int_array_1) - del full_int_array_1 + del add_176, full_int_array_1 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_78 = paddle._C_ops.transpose(reshape_78, [0, 2, 1, 3]) @@ -3515,13 +3492,15 @@ def forward( # pd_op.scale: (1x16x11x64xf32) <- (1x16x11x64xf32, 1xf32) scale_21 = paddle._C_ops.scale(transpose_76, full_6, float("0"), True) - del transpose_76 + del full_6, transpose_76 # pd_op.matmul: (1x16x11x11xf32) <- (1x16x11x64xf32, 1x16x11x64xf32) matmul_155 = paddle._C_ops.matmul(scale_21, transpose_77, False, True) + del scale_21, transpose_77 # pd_op.add: (1x16x11x11xf32) <- (1x16x11x11xf32, 1x1x1x11xf32) add_177 = paddle._C_ops.add(matmul_155, unsqueeze_0) + del matmul_155, unsqueeze_0 # pd_op.softmax: (1x16x11x11xf32) <- (1x16x11x11xf32) softmax_19 = paddle._C_ops.softmax(add_177, -1) @@ -3530,13 +3509,15 @@ def forward( # pd_op.dropout: (1x16x11x11xf32, 1x16x11x11xui8) <- (1x16x11x11xf32, None, 1xf32) dropout_116, dropout_117 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_19, None, full_5, False, "upscale_in_train", 0, False + softmax_19, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_19 # pd_op.matmul: (1x16x11x64xf32) <- (1x16x11x11xf32, 1x16x11x64xf32) matmul_156 = paddle._C_ops.matmul(dropout_116, transpose_78, False, False) + del dropout_116, transpose_78 # pd_op.transpose: (1x11x16x64xf32) <- (1x16x11x64xf32) transpose_79 = paddle._C_ops.transpose(matmul_156, [0, 2, 1, 3]) @@ -3544,20 +3525,20 @@ def forward( # pd_op.reshape: (1x11x1024xf32) <- (1x11x16x64xf32, 3xi64) reshape_79 = paddle._C_ops.reshape(transpose_79, full_int_array_2) - del full_int_array_2 + del full_int_array_2, transpose_79 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_157 = paddle._C_ops.matmul(reshape_79, parameter_11, False, False) - del parameter_11 + del parameter_11, reshape_79 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_178 = paddle._C_ops.add(matmul_157, parameter_10) - del parameter_10 + del matmul_157, parameter_10 # pd_op.dropout: (1x11x1024xf32, 1x11x1024xui8) <- (1x11x1024xf32, None, 1xf32) dropout_118, dropout_119 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_178, None, full_5, False, "upscale_in_train", 0, False + add_178, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -3565,6 +3546,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_179 = paddle._C_ops.add(layer_norm_114, dropout_118) + del dropout_118, layer_norm_114 # pd_op.layer_norm: (1x11x1024xf32, 1x11xf32, 1x11xf32) <- (1x11x1024xf32, 1024xf32, 1024xf32) layer_norm_117, layer_norm_118, layer_norm_119 = (lambda x, f: f(x))( @@ -3573,7 +3555,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_4, parameter_5 + del add_179, parameter_4, parameter_5 # pd_op.matmul: (1x11x4096xf32) <- (1x11x1024xf32, 1024x4096xf32) matmul_158 = paddle._C_ops.matmul(layer_norm_117, parameter_9, False, False) @@ -3581,30 +3563,32 @@ def forward( # pd_op.add: (1x11x4096xf32) <- (1x11x4096xf32, 4096xf32) add_180 = paddle._C_ops.add(matmul_158, parameter_8) - del parameter_8 + del matmul_158, parameter_8 # pd_op.gelu: (1x11x4096xf32) <- (1x11x4096xf32) gelu_19 = paddle._C_ops.gelu(add_180, False) + del add_180 # pd_op.matmul: (1x11x1024xf32) <- (1x11x4096xf32, 4096x1024xf32) matmul_159 = paddle._C_ops.matmul(gelu_19, parameter_7, False, False) - del parameter_7 + del gelu_19, parameter_7 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_181 = paddle._C_ops.add(matmul_159, parameter_6) - del parameter_6 + del matmul_159, parameter_6 # pd_op.dropout: (1x11x1024xf32, 1x11x1024xui8) <- (1x11x1024xf32, None, 1xf32) dropout_120, dropout_121 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_181, None, full_5, False, "upscale_in_train", 0, False + add_181, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) - del add_181 + del add_181, full_5 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_182 = paddle._C_ops.add(layer_norm_117, dropout_120) + del dropout_120, layer_norm_117 # pd_op.layer_norm: (1x11x1024xf32, 1x11xf32, 1x11xf32) <- (1x11x1024xf32, 1024xf32, 1024xf32) layer_norm_120, layer_norm_121, layer_norm_122 = (lambda x, f: f(x))( @@ -3613,7 +3597,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_2, parameter_3 + del add_182, parameter_2, parameter_3 # pd_op.full_int_array: (1xi64) <- () full_int_array_3 = [0] @@ -3625,758 +3609,18 @@ def forward( slice_0 = paddle._C_ops.slice( layer_norm_120, [1], full_int_array_3, full_int_array_4, [1], [1] ) + del full_int_array_3, full_int_array_4, layer_norm_120 # pd_op.matmul: (1x1024xf32) <- (1x1024xf32, 1024x1024xf32) matmul_160 = paddle._C_ops.matmul(slice_0, parameter_1, False, False) - del parameter_1 + del parameter_1, slice_0 # pd_op.add: (1x1024xf32) <- (1x1024xf32, 1024xf32) add_183 = paddle._C_ops.add(matmul_160, parameter_0) - del parameter_0 + del matmul_160, parameter_0 # pd_op.tanh: (1x1024xf32) <- (1x1024xf32) tanh_0 = paddle._C_ops.tanh(add_183) - del ( - add_0, - add_1, - add_101, - add_102, - add_103, - add_104, - add_107, - add_108, - add_11, - add_110, - add_111, - add_112, - add_113, - add_116, - add_117, - add_119, - add_12, - add_120, - add_121, - add_122, - add_125, - add_126, - add_128, - add_129, - add_13, - add_130, - add_131, - add_134, - add_135, - add_137, - add_138, - add_139, - add_14, - add_140, - add_143, - add_144, - add_146, - add_147, - add_148, - add_149, - add_152, - add_153, - add_155, - add_156, - add_157, - add_158, - add_161, - add_162, - add_164, - add_165, - add_166, - add_167, - add_17, - add_170, - add_171, - add_173, - add_174, - add_175, - add_176, - add_179, - add_18, - add_180, - add_182, - add_183, - add_2, - add_20, - add_21, - add_22, - add_23, - add_26, - add_27, - add_29, - add_3, - add_30, - add_31, - add_32, - add_35, - add_36, - add_38, - add_39, - add_4, - add_40, - add_41, - add_44, - add_45, - add_47, - add_48, - add_49, - add_5, - add_50, - add_53, - add_54, - add_56, - add_57, - add_58, - add_59, - add_62, - add_63, - add_65, - add_66, - add_67, - add_68, - add_71, - add_72, - add_74, - add_75, - add_76, - add_77, - add_8, - add_80, - add_81, - add_83, - add_84, - add_85, - add_86, - add_89, - add_9, - add_90, - add_92, - add_93, - add_94, - add_95, - add_98, - add_99, - assign_0, - assign_1, - assign_10, - assign_11, - assign_12, - assign_13, - assign_14, - assign_15, - assign_16, - assign_17, - assign_18, - assign_19, - assign_2, - assign_20, - assign_21, - assign_22, - assign_23, - assign_24, - assign_25, - assign_26, - assign_27, - assign_28, - assign_29, - assign_3, - assign_30, - assign_31, - assign_32, - assign_33, - assign_34, - assign_35, - assign_36, - assign_37, - assign_38, - assign_39, - assign_4, - assign_40, - assign_41, - assign_42, - assign_43, - assign_44, - assign_45, - assign_46, - assign_47, - assign_48, - assign_49, - assign_5, - assign_50, - assign_51, - assign_52, - assign_53, - assign_54, - assign_55, - assign_56, - assign_57, - assign_58, - assign_59, - assign_6, - assign_60, - assign_61, - assign_62, - assign_63, - assign_64, - assign_65, - assign_66, - assign_67, - assign_68, - assign_69, - assign_7, - assign_70, - assign_71, - assign_72, - assign_73, - assign_74, - assign_75, - assign_76, - assign_77, - assign_78, - assign_8, - assign_9, - dropout_0, - dropout_1, - dropout_10, - dropout_100, - dropout_101, - dropout_102, - dropout_103, - dropout_104, - dropout_105, - dropout_106, - dropout_107, - dropout_108, - dropout_109, - dropout_11, - dropout_110, - dropout_111, - dropout_112, - dropout_113, - dropout_114, - dropout_115, - dropout_116, - dropout_117, - dropout_118, - dropout_119, - dropout_12, - dropout_120, - dropout_121, - dropout_13, - dropout_14, - dropout_15, - dropout_16, - dropout_17, - dropout_18, - dropout_19, - dropout_2, - dropout_20, - dropout_21, - dropout_22, - dropout_23, - dropout_24, - dropout_25, - dropout_26, - dropout_27, - dropout_28, - dropout_29, - dropout_3, - dropout_30, - dropout_31, - dropout_32, - dropout_33, - dropout_34, - dropout_35, - dropout_36, - dropout_37, - dropout_38, - dropout_39, - dropout_4, - dropout_40, - dropout_41, - dropout_42, - dropout_43, - dropout_44, - dropout_45, - dropout_46, - dropout_47, - dropout_48, - dropout_49, - dropout_5, - dropout_50, - dropout_51, - dropout_52, - dropout_53, - dropout_54, - dropout_55, - dropout_56, - dropout_57, - dropout_58, - dropout_59, - dropout_6, - dropout_60, - dropout_61, - dropout_62, - dropout_63, - dropout_64, - dropout_65, - dropout_66, - dropout_67, - dropout_68, - dropout_69, - dropout_7, - dropout_70, - dropout_71, - dropout_72, - dropout_73, - dropout_74, - dropout_75, - dropout_76, - dropout_77, - dropout_78, - dropout_79, - dropout_8, - dropout_80, - dropout_81, - dropout_82, - dropout_83, - dropout_84, - dropout_85, - dropout_86, - dropout_87, - dropout_88, - dropout_89, - dropout_9, - dropout_90, - dropout_91, - dropout_92, - dropout_93, - dropout_94, - dropout_95, - dropout_96, - dropout_97, - dropout_98, - dropout_99, - embedding_0, - embedding_1, - embedding_2, - embedding_3, - full_5, - full_6, - full_int_array_3, - full_int_array_4, - gelu_0, - gelu_1, - gelu_10, - gelu_11, - gelu_12, - gelu_13, - gelu_14, - gelu_15, - gelu_16, - gelu_17, - gelu_18, - gelu_19, - gelu_2, - gelu_3, - gelu_4, - gelu_5, - gelu_6, - gelu_7, - gelu_8, - gelu_9, - layer_norm_1, - layer_norm_10, - layer_norm_100, - layer_norm_101, - layer_norm_102, - layer_norm_103, - layer_norm_104, - layer_norm_105, - layer_norm_106, - layer_norm_107, - layer_norm_108, - layer_norm_109, - layer_norm_11, - layer_norm_110, - layer_norm_111, - layer_norm_112, - layer_norm_113, - layer_norm_114, - layer_norm_115, - layer_norm_116, - layer_norm_117, - layer_norm_118, - layer_norm_119, - layer_norm_12, - layer_norm_120, - layer_norm_121, - layer_norm_122, - layer_norm_13, - layer_norm_14, - layer_norm_15, - layer_norm_16, - layer_norm_17, - layer_norm_18, - layer_norm_19, - layer_norm_2, - layer_norm_20, - layer_norm_21, - layer_norm_22, - layer_norm_23, - layer_norm_24, - layer_norm_25, - layer_norm_26, - layer_norm_27, - layer_norm_28, - layer_norm_29, - layer_norm_3, - layer_norm_30, - layer_norm_31, - layer_norm_32, - layer_norm_33, - layer_norm_34, - layer_norm_35, - layer_norm_36, - layer_norm_37, - layer_norm_38, - layer_norm_39, - layer_norm_4, - layer_norm_40, - layer_norm_41, - layer_norm_42, - layer_norm_43, - layer_norm_44, - layer_norm_45, - layer_norm_46, - layer_norm_47, - layer_norm_48, - layer_norm_49, - layer_norm_5, - layer_norm_50, - layer_norm_51, - layer_norm_52, - layer_norm_53, - layer_norm_54, - layer_norm_55, - layer_norm_56, - layer_norm_57, - layer_norm_58, - layer_norm_59, - layer_norm_6, - layer_norm_60, - layer_norm_61, - layer_norm_62, - layer_norm_63, - layer_norm_64, - layer_norm_65, - layer_norm_66, - layer_norm_67, - layer_norm_68, - layer_norm_69, - layer_norm_7, - layer_norm_70, - layer_norm_71, - layer_norm_72, - layer_norm_73, - layer_norm_74, - layer_norm_75, - layer_norm_76, - layer_norm_77, - layer_norm_78, - layer_norm_79, - layer_norm_8, - layer_norm_80, - layer_norm_81, - layer_norm_82, - layer_norm_83, - layer_norm_84, - layer_norm_85, - layer_norm_86, - layer_norm_87, - layer_norm_88, - layer_norm_89, - layer_norm_9, - layer_norm_90, - layer_norm_91, - layer_norm_92, - layer_norm_93, - layer_norm_94, - layer_norm_95, - layer_norm_96, - layer_norm_97, - layer_norm_98, - layer_norm_99, - matmul_0, - matmul_1, - matmul_10, - matmul_101, - matmul_102, - matmul_103, - matmul_104, - matmul_105, - matmul_106, - matmul_107, - matmul_109, - matmul_11, - matmul_110, - matmul_111, - matmul_112, - matmul_113, - matmul_114, - matmul_115, - matmul_117, - matmul_118, - matmul_119, - matmul_120, - matmul_121, - matmul_122, - matmul_123, - matmul_125, - matmul_126, - matmul_127, - matmul_128, - matmul_129, - matmul_13, - matmul_130, - matmul_131, - matmul_133, - matmul_134, - matmul_135, - matmul_136, - matmul_137, - matmul_138, - matmul_139, - matmul_14, - matmul_141, - matmul_142, - matmul_143, - matmul_144, - matmul_145, - matmul_146, - matmul_147, - matmul_149, - matmul_15, - matmul_150, - matmul_151, - matmul_152, - matmul_153, - matmul_154, - matmul_155, - matmul_157, - matmul_158, - matmul_159, - matmul_16, - matmul_160, - matmul_17, - matmul_18, - matmul_19, - matmul_2, - matmul_21, - matmul_22, - matmul_23, - matmul_24, - matmul_25, - matmul_26, - matmul_27, - matmul_29, - matmul_3, - matmul_30, - matmul_31, - matmul_32, - matmul_33, - matmul_34, - matmul_35, - matmul_37, - matmul_38, - matmul_39, - matmul_40, - matmul_41, - matmul_42, - matmul_43, - matmul_45, - matmul_46, - matmul_47, - matmul_48, - matmul_49, - matmul_5, - matmul_50, - matmul_51, - matmul_53, - matmul_54, - matmul_55, - matmul_56, - matmul_57, - matmul_58, - matmul_59, - matmul_6, - matmul_61, - matmul_62, - matmul_63, - matmul_64, - matmul_65, - matmul_66, - matmul_67, - matmul_69, - matmul_7, - matmul_70, - matmul_71, - matmul_72, - matmul_73, - matmul_74, - matmul_75, - matmul_77, - matmul_78, - matmul_79, - matmul_8, - matmul_80, - matmul_81, - matmul_82, - matmul_83, - matmul_85, - matmul_86, - matmul_87, - matmul_88, - matmul_89, - matmul_9, - matmul_90, - matmul_91, - matmul_93, - matmul_94, - matmul_95, - matmul_96, - matmul_97, - matmul_98, - matmul_99, - reshape_11, - reshape_15, - reshape_19, - reshape_23, - reshape_27, - reshape_3, - reshape_31, - reshape_35, - reshape_39, - reshape_43, - reshape_47, - reshape_51, - reshape_55, - reshape_59, - reshape_63, - reshape_67, - reshape_7, - reshape_71, - reshape_75, - reshape_79, - scale_1, - scale_10, - scale_11, - scale_12, - scale_13, - scale_14, - scale_15, - scale_16, - scale_17, - scale_18, - scale_19, - scale_2, - scale_20, - scale_21, - scale_3, - scale_4, - scale_5, - scale_6, - scale_7, - scale_8, - scale_9, - slice_0, - softmax_0, - softmax_1, - softmax_10, - softmax_11, - softmax_12, - softmax_13, - softmax_14, - softmax_15, - softmax_16, - softmax_17, - softmax_18, - softmax_19, - softmax_2, - softmax_3, - softmax_4, - softmax_5, - softmax_6, - softmax_7, - softmax_8, - softmax_9, - subtract_0, - transpose_1, - transpose_10, - transpose_11, - transpose_13, - transpose_14, - transpose_15, - transpose_17, - transpose_18, - transpose_19, - transpose_2, - transpose_21, - transpose_22, - transpose_23, - transpose_25, - transpose_26, - transpose_27, - transpose_29, - transpose_3, - transpose_30, - transpose_31, - transpose_33, - transpose_34, - transpose_35, - transpose_37, - transpose_38, - transpose_39, - transpose_41, - transpose_42, - transpose_43, - transpose_45, - transpose_46, - transpose_47, - transpose_49, - transpose_5, - transpose_50, - transpose_51, - transpose_53, - transpose_54, - transpose_55, - transpose_57, - transpose_58, - transpose_59, - transpose_6, - transpose_61, - transpose_62, - transpose_63, - transpose_65, - transpose_66, - transpose_67, - transpose_69, - transpose_7, - transpose_70, - transpose_71, - transpose_73, - transpose_74, - transpose_75, - transpose_77, - transpose_78, - transpose_79, - transpose_9, - unsqueeze_0, - ) + del add_183 return tanh_0 diff --git a/paddle_samples/PaddleNLP/ernie-3.0-xbase-zh/weight_meta.py b/paddle_samples/PaddleNLP/ernie-3.0-xbase-zh/weight_meta.py index 67df484ac..68038ab15 100644 --- a/paddle_samples/PaddleNLP/ernie-3.0-xbase-zh/weight_meta.py +++ b/paddle_samples/PaddleNLP/ernie-3.0-xbase-zh/weight_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_parameter_0: name = "parameter_0" + original_name = "linear_1312.b_0" shape = [1024] dtype = "float32" min_val = float("-0.109667") @@ -11,6 +12,7 @@ class Program_weight_tensor_parameter_0: class Program_weight_tensor_parameter_1: name = "parameter_1" + original_name = "linear_1312.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.228068") @@ -22,6 +24,7 @@ class Program_weight_tensor_parameter_1: class Program_weight_tensor_parameter_2: name = "parameter_2" + original_name = "layer_norm_452.b_0" shape = [1024] dtype = "float32" min_val = float("-0.827796") @@ -33,6 +36,7 @@ class Program_weight_tensor_parameter_2: class Program_weight_tensor_parameter_3: name = "parameter_3" + original_name = "layer_norm_452.w_0" shape = [1024] dtype = "float32" min_val = float("0.125665") @@ -44,6 +48,7 @@ class Program_weight_tensor_parameter_3: class Program_weight_tensor_parameter_4: name = "parameter_4" + original_name = "layer_norm_451.b_0" shape = [1024] dtype = "float32" min_val = float("-1.43386") @@ -55,6 +60,7 @@ class Program_weight_tensor_parameter_4: class Program_weight_tensor_parameter_5: name = "parameter_5" + original_name = "layer_norm_451.w_0" shape = [1024] dtype = "float32" min_val = float("0.764603") @@ -66,6 +72,7 @@ class Program_weight_tensor_parameter_5: class Program_weight_tensor_parameter_6: name = "parameter_6" + original_name = "linear_1311.b_0" shape = [1024] dtype = "float32" min_val = float("-0.225185") @@ -77,6 +84,7 @@ class Program_weight_tensor_parameter_6: class Program_weight_tensor_parameter_7: name = "parameter_7" + original_name = "linear_1311.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-0.960119") @@ -88,6 +96,7 @@ class Program_weight_tensor_parameter_7: class Program_weight_tensor_parameter_8: name = "parameter_8" + original_name = "linear_1310.b_0" shape = [4096] dtype = "float32" min_val = float("-0.287265") @@ -99,6 +108,7 @@ class Program_weight_tensor_parameter_8: class Program_weight_tensor_parameter_9: name = "parameter_9" + original_name = "linear_1310.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.50612") @@ -110,6 +120,7 @@ class Program_weight_tensor_parameter_9: class Program_weight_tensor_parameter_10: name = "parameter_10" + original_name = "linear_1309.b_0" shape = [1024] dtype = "float32" min_val = float("-0.667739") @@ -121,6 +132,7 @@ class Program_weight_tensor_parameter_10: class Program_weight_tensor_parameter_11: name = "parameter_11" + original_name = "linear_1309.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-1.02207") @@ -132,6 +144,7 @@ class Program_weight_tensor_parameter_11: class Program_weight_tensor_parameter_12: name = "parameter_12" + original_name = "linear_1308.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0927642") @@ -143,6 +156,7 @@ class Program_weight_tensor_parameter_12: class Program_weight_tensor_parameter_13: name = "parameter_13" + original_name = "linear_1308.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.290383") @@ -154,6 +168,7 @@ class Program_weight_tensor_parameter_13: class Program_weight_tensor_parameter_14: name = "parameter_14" + original_name = "linear_1307.b_0" shape = [1024] dtype = "float32" min_val = float("-20.2271") @@ -165,6 +180,7 @@ class Program_weight_tensor_parameter_14: class Program_weight_tensor_parameter_15: name = "parameter_15" + original_name = "linear_1307.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.563203") @@ -176,6 +192,7 @@ class Program_weight_tensor_parameter_15: class Program_weight_tensor_parameter_16: name = "parameter_16" + original_name = "linear_1306.b_0" shape = [1024] dtype = "float32" min_val = float("-0.568509") @@ -187,6 +204,7 @@ class Program_weight_tensor_parameter_16: class Program_weight_tensor_parameter_17: name = "parameter_17" + original_name = "linear_1306.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.27919") @@ -198,6 +216,7 @@ class Program_weight_tensor_parameter_17: class Program_weight_tensor_parameter_18: name = "parameter_18" + original_name = "layer_norm_450.b_0" shape = [1024] dtype = "float32" min_val = float("-1.15396") @@ -209,6 +228,7 @@ class Program_weight_tensor_parameter_18: class Program_weight_tensor_parameter_19: name = "parameter_19" + original_name = "layer_norm_450.w_0" shape = [1024] dtype = "float32" min_val = float("0.301882") @@ -220,6 +240,7 @@ class Program_weight_tensor_parameter_19: class Program_weight_tensor_parameter_20: name = "parameter_20" + original_name = "layer_norm_449.b_0" shape = [1024] dtype = "float32" min_val = float("-1.89985") @@ -231,6 +252,7 @@ class Program_weight_tensor_parameter_20: class Program_weight_tensor_parameter_21: name = "parameter_21" + original_name = "layer_norm_449.w_0" shape = [1024] dtype = "float32" min_val = float("0.73653") @@ -242,6 +264,7 @@ class Program_weight_tensor_parameter_21: class Program_weight_tensor_parameter_22: name = "parameter_22" + original_name = "linear_1305.b_0" shape = [1024] dtype = "float32" min_val = float("-0.329541") @@ -253,6 +276,7 @@ class Program_weight_tensor_parameter_22: class Program_weight_tensor_parameter_23: name = "parameter_23" + original_name = "linear_1305.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-0.762924") @@ -264,6 +288,7 @@ class Program_weight_tensor_parameter_23: class Program_weight_tensor_parameter_24: name = "parameter_24" + original_name = "linear_1304.b_0" shape = [4096] dtype = "float32" min_val = float("-0.267895") @@ -275,6 +300,7 @@ class Program_weight_tensor_parameter_24: class Program_weight_tensor_parameter_25: name = "parameter_25" + original_name = "linear_1304.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.567133") @@ -286,6 +312,7 @@ class Program_weight_tensor_parameter_25: class Program_weight_tensor_parameter_26: name = "parameter_26" + original_name = "linear_1303.b_0" shape = [1024] dtype = "float32" min_val = float("-0.791952") @@ -297,6 +324,7 @@ class Program_weight_tensor_parameter_26: class Program_weight_tensor_parameter_27: name = "parameter_27" + original_name = "linear_1303.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.525997") @@ -308,6 +336,7 @@ class Program_weight_tensor_parameter_27: class Program_weight_tensor_parameter_28: name = "parameter_28" + original_name = "linear_1302.b_0" shape = [1024] dtype = "float32" min_val = float("-0.140775") @@ -319,6 +348,7 @@ class Program_weight_tensor_parameter_28: class Program_weight_tensor_parameter_29: name = "parameter_29" + original_name = "linear_1302.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.20009") @@ -330,6 +360,7 @@ class Program_weight_tensor_parameter_29: class Program_weight_tensor_parameter_30: name = "parameter_30" + original_name = "linear_1301.b_0" shape = [1024] dtype = "float32" min_val = float("-13.1723") @@ -341,6 +372,7 @@ class Program_weight_tensor_parameter_30: class Program_weight_tensor_parameter_31: name = "parameter_31" + original_name = "linear_1301.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.468359") @@ -352,6 +384,7 @@ class Program_weight_tensor_parameter_31: class Program_weight_tensor_parameter_32: name = "parameter_32" + original_name = "linear_1300.b_0" shape = [1024] dtype = "float32" min_val = float("-0.507779") @@ -363,6 +396,7 @@ class Program_weight_tensor_parameter_32: class Program_weight_tensor_parameter_33: name = "parameter_33" + original_name = "linear_1300.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.267784") @@ -374,6 +408,7 @@ class Program_weight_tensor_parameter_33: class Program_weight_tensor_parameter_34: name = "parameter_34" + original_name = "layer_norm_448.b_0" shape = [1024] dtype = "float32" min_val = float("-1.09078") @@ -385,6 +420,7 @@ class Program_weight_tensor_parameter_34: class Program_weight_tensor_parameter_35: name = "parameter_35" + original_name = "layer_norm_448.w_0" shape = [1024] dtype = "float32" min_val = float("0.40153") @@ -396,6 +432,7 @@ class Program_weight_tensor_parameter_35: class Program_weight_tensor_parameter_36: name = "parameter_36" + original_name = "layer_norm_447.b_0" shape = [1024] dtype = "float32" min_val = float("-2.04763") @@ -407,6 +444,7 @@ class Program_weight_tensor_parameter_36: class Program_weight_tensor_parameter_37: name = "parameter_37" + original_name = "layer_norm_447.w_0" shape = [1024] dtype = "float32" min_val = float("0.767486") @@ -418,6 +456,7 @@ class Program_weight_tensor_parameter_37: class Program_weight_tensor_parameter_38: name = "parameter_38" + original_name = "linear_1299.b_0" shape = [1024] dtype = "float32" min_val = float("-0.461443") @@ -429,6 +468,7 @@ class Program_weight_tensor_parameter_38: class Program_weight_tensor_parameter_39: name = "parameter_39" + original_name = "linear_1299.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-0.671617") @@ -440,6 +480,7 @@ class Program_weight_tensor_parameter_39: class Program_weight_tensor_parameter_40: name = "parameter_40" + original_name = "linear_1298.b_0" shape = [4096] dtype = "float32" min_val = float("-0.24996") @@ -451,6 +492,7 @@ class Program_weight_tensor_parameter_40: class Program_weight_tensor_parameter_41: name = "parameter_41" + original_name = "linear_1298.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.522052") @@ -462,6 +504,7 @@ class Program_weight_tensor_parameter_41: class Program_weight_tensor_parameter_42: name = "parameter_42" + original_name = "linear_1297.b_0" shape = [1024] dtype = "float32" min_val = float("-0.112145") @@ -473,6 +516,7 @@ class Program_weight_tensor_parameter_42: class Program_weight_tensor_parameter_43: name = "parameter_43" + original_name = "linear_1297.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.67848") @@ -484,6 +528,7 @@ class Program_weight_tensor_parameter_43: class Program_weight_tensor_parameter_44: name = "parameter_44" + original_name = "linear_1296.b_0" shape = [1024] dtype = "float32" min_val = float("-0.117345") @@ -495,6 +540,7 @@ class Program_weight_tensor_parameter_44: class Program_weight_tensor_parameter_45: name = "parameter_45" + original_name = "linear_1296.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.210359") @@ -506,6 +552,7 @@ class Program_weight_tensor_parameter_45: class Program_weight_tensor_parameter_46: name = "parameter_46" + original_name = "linear_1295.b_0" shape = [1024] dtype = "float32" min_val = float("-10.8041") @@ -517,6 +564,7 @@ class Program_weight_tensor_parameter_46: class Program_weight_tensor_parameter_47: name = "parameter_47" + original_name = "linear_1295.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.470647") @@ -528,6 +576,7 @@ class Program_weight_tensor_parameter_47: class Program_weight_tensor_parameter_48: name = "parameter_48" + original_name = "linear_1294.b_0" shape = [1024] dtype = "float32" min_val = float("-0.563479") @@ -539,6 +588,7 @@ class Program_weight_tensor_parameter_48: class Program_weight_tensor_parameter_49: name = "parameter_49" + original_name = "linear_1294.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.253513") @@ -550,6 +600,7 @@ class Program_weight_tensor_parameter_49: class Program_weight_tensor_parameter_50: name = "parameter_50" + original_name = "layer_norm_446.b_0" shape = [1024] dtype = "float32" min_val = float("-0.729888") @@ -561,6 +612,7 @@ class Program_weight_tensor_parameter_50: class Program_weight_tensor_parameter_51: name = "parameter_51" + original_name = "layer_norm_446.w_0" shape = [1024] dtype = "float32" min_val = float("0.412742") @@ -572,6 +624,7 @@ class Program_weight_tensor_parameter_51: class Program_weight_tensor_parameter_52: name = "parameter_52" + original_name = "layer_norm_445.b_0" shape = [1024] dtype = "float32" min_val = float("-1.9443") @@ -583,6 +636,7 @@ class Program_weight_tensor_parameter_52: class Program_weight_tensor_parameter_53: name = "parameter_53" + original_name = "layer_norm_445.w_0" shape = [1024] dtype = "float32" min_val = float("0.73201") @@ -594,6 +648,7 @@ class Program_weight_tensor_parameter_53: class Program_weight_tensor_parameter_54: name = "parameter_54" + original_name = "linear_1293.b_0" shape = [1024] dtype = "float32" min_val = float("-0.407796") @@ -605,6 +660,7 @@ class Program_weight_tensor_parameter_54: class Program_weight_tensor_parameter_55: name = "parameter_55" + original_name = "linear_1293.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-1.44174") @@ -616,6 +672,7 @@ class Program_weight_tensor_parameter_55: class Program_weight_tensor_parameter_56: name = "parameter_56" + original_name = "linear_1292.b_0" shape = [4096] dtype = "float32" min_val = float("-0.210009") @@ -627,6 +684,7 @@ class Program_weight_tensor_parameter_56: class Program_weight_tensor_parameter_57: name = "parameter_57" + original_name = "linear_1292.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.706008") @@ -638,6 +696,7 @@ class Program_weight_tensor_parameter_57: class Program_weight_tensor_parameter_58: name = "parameter_58" + original_name = "linear_1291.b_0" shape = [1024] dtype = "float32" min_val = float("-0.135633") @@ -649,6 +708,7 @@ class Program_weight_tensor_parameter_58: class Program_weight_tensor_parameter_59: name = "parameter_59" + original_name = "linear_1291.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.463214") @@ -660,6 +720,7 @@ class Program_weight_tensor_parameter_59: class Program_weight_tensor_parameter_60: name = "parameter_60" + original_name = "linear_1290.b_0" shape = [1024] dtype = "float32" min_val = float("-0.090059") @@ -671,6 +732,7 @@ class Program_weight_tensor_parameter_60: class Program_weight_tensor_parameter_61: name = "parameter_61" + original_name = "linear_1290.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.225052") @@ -682,6 +744,7 @@ class Program_weight_tensor_parameter_61: class Program_weight_tensor_parameter_62: name = "parameter_62" + original_name = "linear_1289.b_0" shape = [1024] dtype = "float32" min_val = float("-8.23645") @@ -693,6 +756,7 @@ class Program_weight_tensor_parameter_62: class Program_weight_tensor_parameter_63: name = "parameter_63" + original_name = "linear_1289.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.46628") @@ -704,6 +768,7 @@ class Program_weight_tensor_parameter_63: class Program_weight_tensor_parameter_64: name = "parameter_64" + original_name = "linear_1288.b_0" shape = [1024] dtype = "float32" min_val = float("-0.407083") @@ -715,6 +780,7 @@ class Program_weight_tensor_parameter_64: class Program_weight_tensor_parameter_65: name = "parameter_65" + original_name = "linear_1288.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.341312") @@ -726,6 +792,7 @@ class Program_weight_tensor_parameter_65: class Program_weight_tensor_parameter_66: name = "parameter_66" + original_name = "layer_norm_444.b_0" shape = [1024] dtype = "float32" min_val = float("-0.646623") @@ -737,6 +804,7 @@ class Program_weight_tensor_parameter_66: class Program_weight_tensor_parameter_67: name = "parameter_67" + original_name = "layer_norm_444.w_0" shape = [1024] dtype = "float32" min_val = float("0.435432") @@ -748,6 +816,7 @@ class Program_weight_tensor_parameter_67: class Program_weight_tensor_parameter_68: name = "parameter_68" + original_name = "layer_norm_443.b_0" shape = [1024] dtype = "float32" min_val = float("-1.26834") @@ -759,6 +828,7 @@ class Program_weight_tensor_parameter_68: class Program_weight_tensor_parameter_69: name = "parameter_69" + original_name = "layer_norm_443.w_0" shape = [1024] dtype = "float32" min_val = float("0.780996") @@ -770,6 +840,7 @@ class Program_weight_tensor_parameter_69: class Program_weight_tensor_parameter_70: name = "parameter_70" + original_name = "linear_1287.b_0" shape = [1024] dtype = "float32" min_val = float("-0.505681") @@ -781,6 +852,7 @@ class Program_weight_tensor_parameter_70: class Program_weight_tensor_parameter_71: name = "parameter_71" + original_name = "linear_1287.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-1.09293") @@ -792,6 +864,7 @@ class Program_weight_tensor_parameter_71: class Program_weight_tensor_parameter_72: name = "parameter_72" + original_name = "linear_1286.b_0" shape = [4096] dtype = "float32" min_val = float("-0.211558") @@ -803,6 +876,7 @@ class Program_weight_tensor_parameter_72: class Program_weight_tensor_parameter_73: name = "parameter_73" + original_name = "linear_1286.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.585366") @@ -814,6 +888,7 @@ class Program_weight_tensor_parameter_73: class Program_weight_tensor_parameter_74: name = "parameter_74" + original_name = "linear_1285.b_0" shape = [1024] dtype = "float32" min_val = float("-0.1909") @@ -825,6 +900,7 @@ class Program_weight_tensor_parameter_74: class Program_weight_tensor_parameter_75: name = "parameter_75" + original_name = "linear_1285.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.787407") @@ -836,6 +912,7 @@ class Program_weight_tensor_parameter_75: class Program_weight_tensor_parameter_76: name = "parameter_76" + original_name = "linear_1284.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0774619") @@ -847,6 +924,7 @@ class Program_weight_tensor_parameter_76: class Program_weight_tensor_parameter_77: name = "parameter_77" + original_name = "linear_1284.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.205476") @@ -858,6 +936,7 @@ class Program_weight_tensor_parameter_77: class Program_weight_tensor_parameter_78: name = "parameter_78" + original_name = "linear_1283.b_0" shape = [1024] dtype = "float32" min_val = float("-15.9508") @@ -869,6 +948,7 @@ class Program_weight_tensor_parameter_78: class Program_weight_tensor_parameter_79: name = "parameter_79" + original_name = "linear_1283.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.543271") @@ -880,6 +960,7 @@ class Program_weight_tensor_parameter_79: class Program_weight_tensor_parameter_80: name = "parameter_80" + original_name = "linear_1282.b_0" shape = [1024] dtype = "float32" min_val = float("-0.328526") @@ -891,6 +972,7 @@ class Program_weight_tensor_parameter_80: class Program_weight_tensor_parameter_81: name = "parameter_81" + original_name = "linear_1282.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.34476") @@ -902,6 +984,7 @@ class Program_weight_tensor_parameter_81: class Program_weight_tensor_parameter_82: name = "parameter_82" + original_name = "layer_norm_442.b_0" shape = [1024] dtype = "float32" min_val = float("-0.347675") @@ -913,6 +996,7 @@ class Program_weight_tensor_parameter_82: class Program_weight_tensor_parameter_83: name = "parameter_83" + original_name = "layer_norm_442.w_0" shape = [1024] dtype = "float32" min_val = float("0.465036") @@ -924,6 +1008,7 @@ class Program_weight_tensor_parameter_83: class Program_weight_tensor_parameter_84: name = "parameter_84" + original_name = "layer_norm_441.b_0" shape = [1024] dtype = "float32" min_val = float("-0.871286") @@ -935,6 +1020,7 @@ class Program_weight_tensor_parameter_84: class Program_weight_tensor_parameter_85: name = "parameter_85" + original_name = "layer_norm_441.w_0" shape = [1024] dtype = "float32" min_val = float("0.748859") @@ -946,6 +1032,7 @@ class Program_weight_tensor_parameter_85: class Program_weight_tensor_parameter_86: name = "parameter_86" + original_name = "linear_1281.b_0" shape = [1024] dtype = "float32" min_val = float("-0.322427") @@ -957,6 +1044,7 @@ class Program_weight_tensor_parameter_86: class Program_weight_tensor_parameter_87: name = "parameter_87" + original_name = "linear_1281.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-0.976611") @@ -968,6 +1056,7 @@ class Program_weight_tensor_parameter_87: class Program_weight_tensor_parameter_88: name = "parameter_88" + original_name = "linear_1280.b_0" shape = [4096] dtype = "float32" min_val = float("-0.246578") @@ -979,6 +1068,7 @@ class Program_weight_tensor_parameter_88: class Program_weight_tensor_parameter_89: name = "parameter_89" + original_name = "linear_1280.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.463362") @@ -990,6 +1080,7 @@ class Program_weight_tensor_parameter_89: class Program_weight_tensor_parameter_90: name = "parameter_90" + original_name = "linear_1279.b_0" shape = [1024] dtype = "float32" min_val = float("-0.219697") @@ -1001,6 +1092,7 @@ class Program_weight_tensor_parameter_90: class Program_weight_tensor_parameter_91: name = "parameter_91" + original_name = "linear_1279.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.59404") @@ -1012,6 +1104,7 @@ class Program_weight_tensor_parameter_91: class Program_weight_tensor_parameter_92: name = "parameter_92" + original_name = "linear_1278.b_0" shape = [1024] dtype = "float32" min_val = float("-0.101392") @@ -1023,6 +1116,7 @@ class Program_weight_tensor_parameter_92: class Program_weight_tensor_parameter_93: name = "parameter_93" + original_name = "linear_1278.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.203777") @@ -1034,6 +1128,7 @@ class Program_weight_tensor_parameter_93: class Program_weight_tensor_parameter_94: name = "parameter_94" + original_name = "linear_1277.b_0" shape = [1024] dtype = "float32" min_val = float("-6.20397") @@ -1045,6 +1140,7 @@ class Program_weight_tensor_parameter_94: class Program_weight_tensor_parameter_95: name = "parameter_95" + original_name = "linear_1277.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.391353") @@ -1056,6 +1152,7 @@ class Program_weight_tensor_parameter_95: class Program_weight_tensor_parameter_96: name = "parameter_96" + original_name = "linear_1276.b_0" shape = [1024] dtype = "float32" min_val = float("-0.426534") @@ -1067,6 +1164,7 @@ class Program_weight_tensor_parameter_96: class Program_weight_tensor_parameter_97: name = "parameter_97" + original_name = "linear_1276.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.350518") @@ -1078,6 +1176,7 @@ class Program_weight_tensor_parameter_97: class Program_weight_tensor_parameter_98: name = "parameter_98" + original_name = "layer_norm_440.b_0" shape = [1024] dtype = "float32" min_val = float("-0.124209") @@ -1089,6 +1188,7 @@ class Program_weight_tensor_parameter_98: class Program_weight_tensor_parameter_99: name = "parameter_99" + original_name = "layer_norm_440.w_0" shape = [1024] dtype = "float32" min_val = float("0.522723") @@ -1100,6 +1200,7 @@ class Program_weight_tensor_parameter_99: class Program_weight_tensor_parameter_100: name = "parameter_100" + original_name = "layer_norm_439.b_0" shape = [1024] dtype = "float32" min_val = float("-0.63878") @@ -1111,6 +1212,7 @@ class Program_weight_tensor_parameter_100: class Program_weight_tensor_parameter_101: name = "parameter_101" + original_name = "layer_norm_439.w_0" shape = [1024] dtype = "float32" min_val = float("0.761469") @@ -1122,6 +1224,7 @@ class Program_weight_tensor_parameter_101: class Program_weight_tensor_parameter_102: name = "parameter_102" + original_name = "linear_1275.b_0" shape = [1024] dtype = "float32" min_val = float("-0.404277") @@ -1133,6 +1236,7 @@ class Program_weight_tensor_parameter_102: class Program_weight_tensor_parameter_103: name = "parameter_103" + original_name = "linear_1275.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-1.0183") @@ -1144,6 +1248,7 @@ class Program_weight_tensor_parameter_103: class Program_weight_tensor_parameter_104: name = "parameter_104" + original_name = "linear_1274.b_0" shape = [4096] dtype = "float32" min_val = float("-0.25224") @@ -1155,6 +1260,7 @@ class Program_weight_tensor_parameter_104: class Program_weight_tensor_parameter_105: name = "parameter_105" + original_name = "linear_1274.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.549129") @@ -1166,6 +1272,7 @@ class Program_weight_tensor_parameter_105: class Program_weight_tensor_parameter_106: name = "parameter_106" + original_name = "linear_1273.b_0" shape = [1024] dtype = "float32" min_val = float("-0.141216") @@ -1177,6 +1284,7 @@ class Program_weight_tensor_parameter_106: class Program_weight_tensor_parameter_107: name = "parameter_107" + original_name = "linear_1273.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.424291") @@ -1188,6 +1296,7 @@ class Program_weight_tensor_parameter_107: class Program_weight_tensor_parameter_108: name = "parameter_108" + original_name = "linear_1272.b_0" shape = [1024] dtype = "float32" min_val = float("-0.187029") @@ -1199,6 +1308,7 @@ class Program_weight_tensor_parameter_108: class Program_weight_tensor_parameter_109: name = "parameter_109" + original_name = "linear_1272.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.217074") @@ -1210,6 +1320,7 @@ class Program_weight_tensor_parameter_109: class Program_weight_tensor_parameter_110: name = "parameter_110" + original_name = "linear_1271.b_0" shape = [1024] dtype = "float32" min_val = float("-6.96271") @@ -1221,6 +1332,7 @@ class Program_weight_tensor_parameter_110: class Program_weight_tensor_parameter_111: name = "parameter_111" + original_name = "linear_1271.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.347538") @@ -1232,6 +1344,7 @@ class Program_weight_tensor_parameter_111: class Program_weight_tensor_parameter_112: name = "parameter_112" + original_name = "linear_1270.b_0" shape = [1024] dtype = "float32" min_val = float("-0.404909") @@ -1243,6 +1356,7 @@ class Program_weight_tensor_parameter_112: class Program_weight_tensor_parameter_113: name = "parameter_113" + original_name = "linear_1270.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.288851") @@ -1254,6 +1368,7 @@ class Program_weight_tensor_parameter_113: class Program_weight_tensor_parameter_114: name = "parameter_114" + original_name = "layer_norm_438.b_0" shape = [1024] dtype = "float32" min_val = float("-0.216554") @@ -1265,6 +1380,7 @@ class Program_weight_tensor_parameter_114: class Program_weight_tensor_parameter_115: name = "parameter_115" + original_name = "layer_norm_438.w_0" shape = [1024] dtype = "float32" min_val = float("0.352762") @@ -1276,6 +1392,7 @@ class Program_weight_tensor_parameter_115: class Program_weight_tensor_parameter_116: name = "parameter_116" + original_name = "layer_norm_437.b_0" shape = [1024] dtype = "float32" min_val = float("-0.981913") @@ -1287,6 +1404,7 @@ class Program_weight_tensor_parameter_116: class Program_weight_tensor_parameter_117: name = "parameter_117" + original_name = "layer_norm_437.w_0" shape = [1024] dtype = "float32" min_val = float("0.772004") @@ -1298,6 +1416,7 @@ class Program_weight_tensor_parameter_117: class Program_weight_tensor_parameter_118: name = "parameter_118" + original_name = "linear_1269.b_0" shape = [1024] dtype = "float32" min_val = float("-0.45557") @@ -1309,6 +1428,7 @@ class Program_weight_tensor_parameter_118: class Program_weight_tensor_parameter_119: name = "parameter_119" + original_name = "linear_1269.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-0.599151") @@ -1320,6 +1440,7 @@ class Program_weight_tensor_parameter_119: class Program_weight_tensor_parameter_120: name = "parameter_120" + original_name = "linear_1268.b_0" shape = [4096] dtype = "float32" min_val = float("-0.316903") @@ -1331,6 +1452,7 @@ class Program_weight_tensor_parameter_120: class Program_weight_tensor_parameter_121: name = "parameter_121" + original_name = "linear_1268.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.358875") @@ -1342,6 +1464,7 @@ class Program_weight_tensor_parameter_121: class Program_weight_tensor_parameter_122: name = "parameter_122" + original_name = "linear_1267.b_0" shape = [1024] dtype = "float32" min_val = float("-0.371451") @@ -1353,6 +1476,7 @@ class Program_weight_tensor_parameter_122: class Program_weight_tensor_parameter_123: name = "parameter_123" + original_name = "linear_1267.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.290732") @@ -1364,6 +1488,7 @@ class Program_weight_tensor_parameter_123: class Program_weight_tensor_parameter_124: name = "parameter_124" + original_name = "linear_1266.b_0" shape = [1024] dtype = "float32" min_val = float("-0.081295") @@ -1375,6 +1500,7 @@ class Program_weight_tensor_parameter_124: class Program_weight_tensor_parameter_125: name = "parameter_125" + original_name = "linear_1266.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.211307") @@ -1386,6 +1512,7 @@ class Program_weight_tensor_parameter_125: class Program_weight_tensor_parameter_126: name = "parameter_126" + original_name = "linear_1265.b_0" shape = [1024] dtype = "float32" min_val = float("-4.83172") @@ -1397,6 +1524,7 @@ class Program_weight_tensor_parameter_126: class Program_weight_tensor_parameter_127: name = "parameter_127" + original_name = "linear_1265.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.31883") @@ -1408,6 +1536,7 @@ class Program_weight_tensor_parameter_127: class Program_weight_tensor_parameter_128: name = "parameter_128" + original_name = "linear_1264.b_0" shape = [1024] dtype = "float32" min_val = float("-0.493883") @@ -1419,6 +1548,7 @@ class Program_weight_tensor_parameter_128: class Program_weight_tensor_parameter_129: name = "parameter_129" + original_name = "linear_1264.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.305872") @@ -1430,6 +1560,7 @@ class Program_weight_tensor_parameter_129: class Program_weight_tensor_parameter_130: name = "parameter_130" + original_name = "layer_norm_436.b_0" shape = [1024] dtype = "float32" min_val = float("-0.192699") @@ -1441,6 +1572,7 @@ class Program_weight_tensor_parameter_130: class Program_weight_tensor_parameter_131: name = "parameter_131" + original_name = "layer_norm_436.w_0" shape = [1024] dtype = "float32" min_val = float("0.189962") @@ -1452,6 +1584,7 @@ class Program_weight_tensor_parameter_131: class Program_weight_tensor_parameter_132: name = "parameter_132" + original_name = "layer_norm_435.b_0" shape = [1024] dtype = "float32" min_val = float("-1.61714") @@ -1463,6 +1596,7 @@ class Program_weight_tensor_parameter_132: class Program_weight_tensor_parameter_133: name = "parameter_133" + original_name = "layer_norm_435.w_0" shape = [1024] dtype = "float32" min_val = float("0.709626") @@ -1474,6 +1608,7 @@ class Program_weight_tensor_parameter_133: class Program_weight_tensor_parameter_134: name = "parameter_134" + original_name = "linear_1263.b_0" shape = [1024] dtype = "float32" min_val = float("-0.247222") @@ -1485,6 +1620,7 @@ class Program_weight_tensor_parameter_134: class Program_weight_tensor_parameter_135: name = "parameter_135" + original_name = "linear_1263.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-0.763152") @@ -1496,6 +1632,7 @@ class Program_weight_tensor_parameter_135: class Program_weight_tensor_parameter_136: name = "parameter_136" + original_name = "linear_1262.b_0" shape = [4096] dtype = "float32" min_val = float("-0.232526") @@ -1507,6 +1644,7 @@ class Program_weight_tensor_parameter_136: class Program_weight_tensor_parameter_137: name = "parameter_137" + original_name = "linear_1262.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.409659") @@ -1518,6 +1656,7 @@ class Program_weight_tensor_parameter_137: class Program_weight_tensor_parameter_138: name = "parameter_138" + original_name = "linear_1261.b_0" shape = [1024] dtype = "float32" min_val = float("-0.723453") @@ -1529,6 +1668,7 @@ class Program_weight_tensor_parameter_138: class Program_weight_tensor_parameter_139: name = "parameter_139" + original_name = "linear_1261.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.381667") @@ -1540,6 +1680,7 @@ class Program_weight_tensor_parameter_139: class Program_weight_tensor_parameter_140: name = "parameter_140" + original_name = "linear_1260.b_0" shape = [1024] dtype = "float32" min_val = float("-0.112433") @@ -1551,6 +1692,7 @@ class Program_weight_tensor_parameter_140: class Program_weight_tensor_parameter_141: name = "parameter_141" + original_name = "linear_1260.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.313983") @@ -1562,6 +1704,7 @@ class Program_weight_tensor_parameter_141: class Program_weight_tensor_parameter_142: name = "parameter_142" + original_name = "linear_1259.b_0" shape = [1024] dtype = "float32" min_val = float("-6.00247") @@ -1573,6 +1716,7 @@ class Program_weight_tensor_parameter_142: class Program_weight_tensor_parameter_143: name = "parameter_143" + original_name = "linear_1259.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.312379") @@ -1584,6 +1728,7 @@ class Program_weight_tensor_parameter_143: class Program_weight_tensor_parameter_144: name = "parameter_144" + original_name = "linear_1258.b_0" shape = [1024] dtype = "float32" min_val = float("-0.518143") @@ -1595,6 +1740,7 @@ class Program_weight_tensor_parameter_144: class Program_weight_tensor_parameter_145: name = "parameter_145" + original_name = "linear_1258.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.320926") @@ -1606,6 +1752,7 @@ class Program_weight_tensor_parameter_145: class Program_weight_tensor_parameter_146: name = "parameter_146" + original_name = "layer_norm_434.b_0" shape = [1024] dtype = "float32" min_val = float("-0.23919") @@ -1617,6 +1764,7 @@ class Program_weight_tensor_parameter_146: class Program_weight_tensor_parameter_147: name = "parameter_147" + original_name = "layer_norm_434.w_0" shape = [1024] dtype = "float32" min_val = float("0.20798") @@ -1628,6 +1776,7 @@ class Program_weight_tensor_parameter_147: class Program_weight_tensor_parameter_148: name = "parameter_148" + original_name = "layer_norm_433.b_0" shape = [1024] dtype = "float32" min_val = float("-2.06876") @@ -1639,6 +1788,7 @@ class Program_weight_tensor_parameter_148: class Program_weight_tensor_parameter_149: name = "parameter_149" + original_name = "layer_norm_433.w_0" shape = [1024] dtype = "float32" min_val = float("0.767997") @@ -1650,6 +1800,7 @@ class Program_weight_tensor_parameter_149: class Program_weight_tensor_parameter_150: name = "parameter_150" + original_name = "linear_1257.b_0" shape = [1024] dtype = "float32" min_val = float("-0.310583") @@ -1661,6 +1812,7 @@ class Program_weight_tensor_parameter_150: class Program_weight_tensor_parameter_151: name = "parameter_151" + original_name = "linear_1257.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-1.3333") @@ -1672,6 +1824,7 @@ class Program_weight_tensor_parameter_151: class Program_weight_tensor_parameter_152: name = "parameter_152" + original_name = "linear_1256.b_0" shape = [4096] dtype = "float32" min_val = float("-0.194196") @@ -1683,6 +1836,7 @@ class Program_weight_tensor_parameter_152: class Program_weight_tensor_parameter_153: name = "parameter_153" + original_name = "linear_1256.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.517372") @@ -1694,6 +1848,7 @@ class Program_weight_tensor_parameter_153: class Program_weight_tensor_parameter_154: name = "parameter_154" + original_name = "linear_1255.b_0" shape = [1024] dtype = "float32" min_val = float("-0.838094") @@ -1705,6 +1860,7 @@ class Program_weight_tensor_parameter_154: class Program_weight_tensor_parameter_155: name = "parameter_155" + original_name = "linear_1255.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.388766") @@ -1716,6 +1872,7 @@ class Program_weight_tensor_parameter_155: class Program_weight_tensor_parameter_156: name = "parameter_156" + original_name = "linear_1254.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0665326") @@ -1727,6 +1884,7 @@ class Program_weight_tensor_parameter_156: class Program_weight_tensor_parameter_157: name = "parameter_157" + original_name = "linear_1254.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.19643") @@ -1738,6 +1896,7 @@ class Program_weight_tensor_parameter_157: class Program_weight_tensor_parameter_158: name = "parameter_158" + original_name = "linear_1253.b_0" shape = [1024] dtype = "float32" min_val = float("-4.46037") @@ -1749,6 +1908,7 @@ class Program_weight_tensor_parameter_158: class Program_weight_tensor_parameter_159: name = "parameter_159" + original_name = "linear_1253.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.539263") @@ -1760,6 +1920,7 @@ class Program_weight_tensor_parameter_159: class Program_weight_tensor_parameter_160: name = "parameter_160" + original_name = "linear_1252.b_0" shape = [1024] dtype = "float32" min_val = float("-0.435501") @@ -1771,6 +1932,7 @@ class Program_weight_tensor_parameter_160: class Program_weight_tensor_parameter_161: name = "parameter_161" + original_name = "linear_1252.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.308494") @@ -1782,6 +1944,7 @@ class Program_weight_tensor_parameter_161: class Program_weight_tensor_parameter_162: name = "parameter_162" + original_name = "layer_norm_432.b_0" shape = [1024] dtype = "float32" min_val = float("-0.305387") @@ -1793,6 +1956,7 @@ class Program_weight_tensor_parameter_162: class Program_weight_tensor_parameter_163: name = "parameter_163" + original_name = "layer_norm_432.w_0" shape = [1024] dtype = "float32" min_val = float("0.382061") @@ -1804,6 +1968,7 @@ class Program_weight_tensor_parameter_163: class Program_weight_tensor_parameter_164: name = "parameter_164" + original_name = "layer_norm_431.b_0" shape = [1024] dtype = "float32" min_val = float("-1.92787") @@ -1815,6 +1980,7 @@ class Program_weight_tensor_parameter_164: class Program_weight_tensor_parameter_165: name = "parameter_165" + original_name = "layer_norm_431.w_0" shape = [1024] dtype = "float32" min_val = float("0.823414") @@ -1826,6 +1992,7 @@ class Program_weight_tensor_parameter_165: class Program_weight_tensor_parameter_166: name = "parameter_166" + original_name = "linear_1251.b_0" shape = [1024] dtype = "float32" min_val = float("-0.503221") @@ -1837,6 +2004,7 @@ class Program_weight_tensor_parameter_166: class Program_weight_tensor_parameter_167: name = "parameter_167" + original_name = "linear_1251.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-1.84512") @@ -1848,6 +2016,7 @@ class Program_weight_tensor_parameter_167: class Program_weight_tensor_parameter_168: name = "parameter_168" + original_name = "linear_1250.b_0" shape = [4096] dtype = "float32" min_val = float("-0.171969") @@ -1859,6 +2028,7 @@ class Program_weight_tensor_parameter_168: class Program_weight_tensor_parameter_169: name = "parameter_169" + original_name = "linear_1250.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.453581") @@ -1870,6 +2040,7 @@ class Program_weight_tensor_parameter_169: class Program_weight_tensor_parameter_170: name = "parameter_170" + original_name = "linear_1249.b_0" shape = [1024] dtype = "float32" min_val = float("-0.692595") @@ -1881,6 +2052,7 @@ class Program_weight_tensor_parameter_170: class Program_weight_tensor_parameter_171: name = "parameter_171" + original_name = "linear_1249.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.366132") @@ -1892,6 +2064,7 @@ class Program_weight_tensor_parameter_171: class Program_weight_tensor_parameter_172: name = "parameter_172" + original_name = "linear_1248.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0844584") @@ -1903,6 +2076,7 @@ class Program_weight_tensor_parameter_172: class Program_weight_tensor_parameter_173: name = "parameter_173" + original_name = "linear_1248.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.247418") @@ -1914,6 +2088,7 @@ class Program_weight_tensor_parameter_173: class Program_weight_tensor_parameter_174: name = "parameter_174" + original_name = "linear_1247.b_0" shape = [1024] dtype = "float32" min_val = float("-4.02521") @@ -1925,6 +2100,7 @@ class Program_weight_tensor_parameter_174: class Program_weight_tensor_parameter_175: name = "parameter_175" + original_name = "linear_1247.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.348556") @@ -1936,6 +2112,7 @@ class Program_weight_tensor_parameter_175: class Program_weight_tensor_parameter_176: name = "parameter_176" + original_name = "linear_1246.b_0" shape = [1024] dtype = "float32" min_val = float("-0.550233") @@ -1947,6 +2124,7 @@ class Program_weight_tensor_parameter_176: class Program_weight_tensor_parameter_177: name = "parameter_177" + original_name = "linear_1246.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.294902") @@ -1958,6 +2136,7 @@ class Program_weight_tensor_parameter_177: class Program_weight_tensor_parameter_178: name = "parameter_178" + original_name = "layer_norm_430.b_0" shape = [1024] dtype = "float32" min_val = float("-0.224808") @@ -1969,6 +2148,7 @@ class Program_weight_tensor_parameter_178: class Program_weight_tensor_parameter_179: name = "parameter_179" + original_name = "layer_norm_430.w_0" shape = [1024] dtype = "float32" min_val = float("0.384355") @@ -1980,6 +2160,7 @@ class Program_weight_tensor_parameter_179: class Program_weight_tensor_parameter_180: name = "parameter_180" + original_name = "layer_norm_429.b_0" shape = [1024] dtype = "float32" min_val = float("-1.87549") @@ -1991,6 +2172,7 @@ class Program_weight_tensor_parameter_180: class Program_weight_tensor_parameter_181: name = "parameter_181" + original_name = "layer_norm_429.w_0" shape = [1024] dtype = "float32" min_val = float("0.844714") @@ -2002,6 +2184,7 @@ class Program_weight_tensor_parameter_181: class Program_weight_tensor_parameter_182: name = "parameter_182" + original_name = "linear_1245.b_0" shape = [1024] dtype = "float32" min_val = float("-0.649267") @@ -2013,6 +2196,7 @@ class Program_weight_tensor_parameter_182: class Program_weight_tensor_parameter_183: name = "parameter_183" + original_name = "linear_1245.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-2.21908") @@ -2024,6 +2208,7 @@ class Program_weight_tensor_parameter_183: class Program_weight_tensor_parameter_184: name = "parameter_184" + original_name = "linear_1244.b_0" shape = [4096] dtype = "float32" min_val = float("-0.206103") @@ -2035,6 +2220,7 @@ class Program_weight_tensor_parameter_184: class Program_weight_tensor_parameter_185: name = "parameter_185" + original_name = "linear_1244.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.395661") @@ -2046,6 +2232,7 @@ class Program_weight_tensor_parameter_185: class Program_weight_tensor_parameter_186: name = "parameter_186" + original_name = "linear_1243.b_0" shape = [1024] dtype = "float32" min_val = float("-0.42151") @@ -2057,6 +2244,7 @@ class Program_weight_tensor_parameter_186: class Program_weight_tensor_parameter_187: name = "parameter_187" + original_name = "linear_1243.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.300479") @@ -2068,6 +2256,7 @@ class Program_weight_tensor_parameter_187: class Program_weight_tensor_parameter_188: name = "parameter_188" + original_name = "linear_1242.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0896553") @@ -2079,6 +2268,7 @@ class Program_weight_tensor_parameter_188: class Program_weight_tensor_parameter_189: name = "parameter_189" + original_name = "linear_1242.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.224958") @@ -2090,6 +2280,7 @@ class Program_weight_tensor_parameter_189: class Program_weight_tensor_parameter_190: name = "parameter_190" + original_name = "linear_1241.b_0" shape = [1024] dtype = "float32" min_val = float("-5.25712") @@ -2101,6 +2292,7 @@ class Program_weight_tensor_parameter_190: class Program_weight_tensor_parameter_191: name = "parameter_191" + original_name = "linear_1241.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.735617") @@ -2112,6 +2304,7 @@ class Program_weight_tensor_parameter_191: class Program_weight_tensor_parameter_192: name = "parameter_192" + original_name = "linear_1240.b_0" shape = [1024] dtype = "float32" min_val = float("-0.406131") @@ -2123,6 +2316,7 @@ class Program_weight_tensor_parameter_192: class Program_weight_tensor_parameter_193: name = "parameter_193" + original_name = "linear_1240.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.391997") @@ -2134,6 +2328,7 @@ class Program_weight_tensor_parameter_193: class Program_weight_tensor_parameter_194: name = "parameter_194" + original_name = "layer_norm_428.b_0" shape = [1024] dtype = "float32" min_val = float("-0.362033") @@ -2145,6 +2340,7 @@ class Program_weight_tensor_parameter_194: class Program_weight_tensor_parameter_195: name = "parameter_195" + original_name = "layer_norm_428.w_0" shape = [1024] dtype = "float32" min_val = float("0.400483") @@ -2156,6 +2352,7 @@ class Program_weight_tensor_parameter_195: class Program_weight_tensor_parameter_196: name = "parameter_196" + original_name = "layer_norm_427.b_0" shape = [1024] dtype = "float32" min_val = float("-2.02893") @@ -2167,6 +2364,7 @@ class Program_weight_tensor_parameter_196: class Program_weight_tensor_parameter_197: name = "parameter_197" + original_name = "layer_norm_427.w_0" shape = [1024] dtype = "float32" min_val = float("0.816563") @@ -2178,6 +2376,7 @@ class Program_weight_tensor_parameter_197: class Program_weight_tensor_parameter_198: name = "parameter_198" + original_name = "linear_1239.b_0" shape = [1024] dtype = "float32" min_val = float("-0.581401") @@ -2189,6 +2388,7 @@ class Program_weight_tensor_parameter_198: class Program_weight_tensor_parameter_199: name = "parameter_199" + original_name = "linear_1239.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-2.12571") @@ -2200,6 +2400,7 @@ class Program_weight_tensor_parameter_199: class Program_weight_tensor_parameter_200: name = "parameter_200" + original_name = "linear_1238.b_0" shape = [4096] dtype = "float32" min_val = float("-0.147417") @@ -2211,6 +2412,7 @@ class Program_weight_tensor_parameter_200: class Program_weight_tensor_parameter_201: name = "parameter_201" + original_name = "linear_1238.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.341821") @@ -2222,6 +2424,7 @@ class Program_weight_tensor_parameter_201: class Program_weight_tensor_parameter_202: name = "parameter_202" + original_name = "linear_1237.b_0" shape = [1024] dtype = "float32" min_val = float("-0.601638") @@ -2233,6 +2436,7 @@ class Program_weight_tensor_parameter_202: class Program_weight_tensor_parameter_203: name = "parameter_203" + original_name = "linear_1237.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.344027") @@ -2244,6 +2448,7 @@ class Program_weight_tensor_parameter_203: class Program_weight_tensor_parameter_204: name = "parameter_204" + original_name = "linear_1236.b_0" shape = [1024] dtype = "float32" min_val = float("-0.109856") @@ -2255,6 +2460,7 @@ class Program_weight_tensor_parameter_204: class Program_weight_tensor_parameter_205: name = "parameter_205" + original_name = "linear_1236.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.330759") @@ -2266,6 +2472,7 @@ class Program_weight_tensor_parameter_205: class Program_weight_tensor_parameter_206: name = "parameter_206" + original_name = "linear_1235.b_0" shape = [1024] dtype = "float32" min_val = float("-4.92922") @@ -2277,6 +2484,7 @@ class Program_weight_tensor_parameter_206: class Program_weight_tensor_parameter_207: name = "parameter_207" + original_name = "linear_1235.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.462936") @@ -2288,6 +2496,7 @@ class Program_weight_tensor_parameter_207: class Program_weight_tensor_parameter_208: name = "parameter_208" + original_name = "linear_1234.b_0" shape = [1024] dtype = "float32" min_val = float("-0.357544") @@ -2299,6 +2508,7 @@ class Program_weight_tensor_parameter_208: class Program_weight_tensor_parameter_209: name = "parameter_209" + original_name = "linear_1234.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.333876") @@ -2310,6 +2520,7 @@ class Program_weight_tensor_parameter_209: class Program_weight_tensor_parameter_210: name = "parameter_210" + original_name = "layer_norm_426.b_0" shape = [1024] dtype = "float32" min_val = float("-0.474219") @@ -2321,6 +2532,7 @@ class Program_weight_tensor_parameter_210: class Program_weight_tensor_parameter_211: name = "parameter_211" + original_name = "layer_norm_426.w_0" shape = [1024] dtype = "float32" min_val = float("0.494691") @@ -2332,6 +2544,7 @@ class Program_weight_tensor_parameter_211: class Program_weight_tensor_parameter_212: name = "parameter_212" + original_name = "layer_norm_425.b_0" shape = [1024] dtype = "float32" min_val = float("-1.97208") @@ -2343,6 +2556,7 @@ class Program_weight_tensor_parameter_212: class Program_weight_tensor_parameter_213: name = "parameter_213" + original_name = "layer_norm_425.w_0" shape = [1024] dtype = "float32" min_val = float("0.846155") @@ -2354,6 +2568,7 @@ class Program_weight_tensor_parameter_213: class Program_weight_tensor_parameter_214: name = "parameter_214" + original_name = "linear_1233.b_0" shape = [1024] dtype = "float32" min_val = float("-0.401295") @@ -2365,6 +2580,7 @@ class Program_weight_tensor_parameter_214: class Program_weight_tensor_parameter_215: name = "parameter_215" + original_name = "linear_1233.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-2.42523") @@ -2376,6 +2592,7 @@ class Program_weight_tensor_parameter_215: class Program_weight_tensor_parameter_216: name = "parameter_216" + original_name = "linear_1232.b_0" shape = [4096] dtype = "float32" min_val = float("-0.183465") @@ -2387,6 +2604,7 @@ class Program_weight_tensor_parameter_216: class Program_weight_tensor_parameter_217: name = "parameter_217" + original_name = "linear_1232.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.33174") @@ -2398,6 +2616,7 @@ class Program_weight_tensor_parameter_217: class Program_weight_tensor_parameter_218: name = "parameter_218" + original_name = "linear_1231.b_0" shape = [1024] dtype = "float32" min_val = float("-0.404088") @@ -2409,6 +2628,7 @@ class Program_weight_tensor_parameter_218: class Program_weight_tensor_parameter_219: name = "parameter_219" + original_name = "linear_1231.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.349317") @@ -2420,6 +2640,7 @@ class Program_weight_tensor_parameter_219: class Program_weight_tensor_parameter_220: name = "parameter_220" + original_name = "linear_1230.b_0" shape = [1024] dtype = "float32" min_val = float("-0.101726") @@ -2431,6 +2652,7 @@ class Program_weight_tensor_parameter_220: class Program_weight_tensor_parameter_221: name = "parameter_221" + original_name = "linear_1230.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.208427") @@ -2442,6 +2664,7 @@ class Program_weight_tensor_parameter_221: class Program_weight_tensor_parameter_222: name = "parameter_222" + original_name = "linear_1229.b_0" shape = [1024] dtype = "float32" min_val = float("-5.03153") @@ -2453,6 +2676,7 @@ class Program_weight_tensor_parameter_222: class Program_weight_tensor_parameter_223: name = "parameter_223" + original_name = "linear_1229.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.528424") @@ -2464,6 +2688,7 @@ class Program_weight_tensor_parameter_223: class Program_weight_tensor_parameter_224: name = "parameter_224" + original_name = "linear_1228.b_0" shape = [1024] dtype = "float32" min_val = float("-0.457843") @@ -2475,6 +2700,7 @@ class Program_weight_tensor_parameter_224: class Program_weight_tensor_parameter_225: name = "parameter_225" + original_name = "linear_1228.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.369502") @@ -2486,6 +2712,7 @@ class Program_weight_tensor_parameter_225: class Program_weight_tensor_parameter_226: name = "parameter_226" + original_name = "layer_norm_424.b_0" shape = [1024] dtype = "float32" min_val = float("-0.803124") @@ -2497,6 +2724,7 @@ class Program_weight_tensor_parameter_226: class Program_weight_tensor_parameter_227: name = "parameter_227" + original_name = "layer_norm_424.w_0" shape = [1024] dtype = "float32" min_val = float("0.376918") @@ -2508,6 +2736,7 @@ class Program_weight_tensor_parameter_227: class Program_weight_tensor_parameter_228: name = "parameter_228" + original_name = "layer_norm_423.b_0" shape = [1024] dtype = "float32" min_val = float("-1.98508") @@ -2519,6 +2748,7 @@ class Program_weight_tensor_parameter_228: class Program_weight_tensor_parameter_229: name = "parameter_229" + original_name = "layer_norm_423.w_0" shape = [1024] dtype = "float32" min_val = float("0.835429") @@ -2530,6 +2760,7 @@ class Program_weight_tensor_parameter_229: class Program_weight_tensor_parameter_230: name = "parameter_230" + original_name = "linear_1227.b_0" shape = [1024] dtype = "float32" min_val = float("-0.417775") @@ -2541,6 +2772,7 @@ class Program_weight_tensor_parameter_230: class Program_weight_tensor_parameter_231: name = "parameter_231" + original_name = "linear_1227.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-3.80871") @@ -2552,6 +2784,7 @@ class Program_weight_tensor_parameter_231: class Program_weight_tensor_parameter_232: name = "parameter_232" + original_name = "linear_1226.b_0" shape = [4096] dtype = "float32" min_val = float("-0.140355") @@ -2563,6 +2796,7 @@ class Program_weight_tensor_parameter_232: class Program_weight_tensor_parameter_233: name = "parameter_233" + original_name = "linear_1226.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.653851") @@ -2574,6 +2808,7 @@ class Program_weight_tensor_parameter_233: class Program_weight_tensor_parameter_234: name = "parameter_234" + original_name = "linear_1225.b_0" shape = [1024] dtype = "float32" min_val = float("-0.68959") @@ -2585,6 +2820,7 @@ class Program_weight_tensor_parameter_234: class Program_weight_tensor_parameter_235: name = "parameter_235" + original_name = "linear_1225.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.472615") @@ -2596,6 +2832,7 @@ class Program_weight_tensor_parameter_235: class Program_weight_tensor_parameter_236: name = "parameter_236" + original_name = "linear_1224.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0636851") @@ -2607,6 +2844,7 @@ class Program_weight_tensor_parameter_236: class Program_weight_tensor_parameter_237: name = "parameter_237" + original_name = "linear_1224.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.233787") @@ -2618,6 +2856,7 @@ class Program_weight_tensor_parameter_237: class Program_weight_tensor_parameter_238: name = "parameter_238" + original_name = "linear_1223.b_0" shape = [1024] dtype = "float32" min_val = float("-5.60581") @@ -2629,6 +2868,7 @@ class Program_weight_tensor_parameter_238: class Program_weight_tensor_parameter_239: name = "parameter_239" + original_name = "linear_1223.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.381037") @@ -2640,6 +2880,7 @@ class Program_weight_tensor_parameter_239: class Program_weight_tensor_parameter_240: name = "parameter_240" + original_name = "linear_1222.b_0" shape = [1024] dtype = "float32" min_val = float("-0.448509") @@ -2651,6 +2892,7 @@ class Program_weight_tensor_parameter_240: class Program_weight_tensor_parameter_241: name = "parameter_241" + original_name = "linear_1222.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.416461") @@ -2662,6 +2904,7 @@ class Program_weight_tensor_parameter_241: class Program_weight_tensor_parameter_242: name = "parameter_242" + original_name = "layer_norm_422.b_0" shape = [1024] dtype = "float32" min_val = float("-0.934927") @@ -2673,6 +2916,7 @@ class Program_weight_tensor_parameter_242: class Program_weight_tensor_parameter_243: name = "parameter_243" + original_name = "layer_norm_422.w_0" shape = [1024] dtype = "float32" min_val = float("0.139467") @@ -2684,6 +2928,7 @@ class Program_weight_tensor_parameter_243: class Program_weight_tensor_parameter_244: name = "parameter_244" + original_name = "layer_norm_421.b_0" shape = [1024] dtype = "float32" min_val = float("-2.0933") @@ -2695,6 +2940,7 @@ class Program_weight_tensor_parameter_244: class Program_weight_tensor_parameter_245: name = "parameter_245" + original_name = "layer_norm_421.w_0" shape = [1024] dtype = "float32" min_val = float("0.785764") @@ -2706,6 +2952,7 @@ class Program_weight_tensor_parameter_245: class Program_weight_tensor_parameter_246: name = "parameter_246" + original_name = "linear_1221.b_0" shape = [1024] dtype = "float32" min_val = float("-0.410728") @@ -2717,6 +2964,7 @@ class Program_weight_tensor_parameter_246: class Program_weight_tensor_parameter_247: name = "parameter_247" + original_name = "linear_1221.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-4.48571") @@ -2728,6 +2976,7 @@ class Program_weight_tensor_parameter_247: class Program_weight_tensor_parameter_248: name = "parameter_248" + original_name = "linear_1220.b_0" shape = [4096] dtype = "float32" min_val = float("-0.132829") @@ -2739,6 +2988,7 @@ class Program_weight_tensor_parameter_248: class Program_weight_tensor_parameter_249: name = "parameter_249" + original_name = "linear_1220.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.534071") @@ -2750,6 +3000,7 @@ class Program_weight_tensor_parameter_249: class Program_weight_tensor_parameter_250: name = "parameter_250" + original_name = "linear_1219.b_0" shape = [1024] dtype = "float32" min_val = float("-0.454935") @@ -2761,6 +3012,7 @@ class Program_weight_tensor_parameter_250: class Program_weight_tensor_parameter_251: name = "parameter_251" + original_name = "linear_1219.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.530523") @@ -2772,6 +3024,7 @@ class Program_weight_tensor_parameter_251: class Program_weight_tensor_parameter_252: name = "parameter_252" + original_name = "linear_1218.b_0" shape = [1024] dtype = "float32" min_val = float("-0.122822") @@ -2783,6 +3036,7 @@ class Program_weight_tensor_parameter_252: class Program_weight_tensor_parameter_253: name = "parameter_253" + original_name = "linear_1218.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.1901") @@ -2794,6 +3048,7 @@ class Program_weight_tensor_parameter_253: class Program_weight_tensor_parameter_254: name = "parameter_254" + original_name = "linear_1217.b_0" shape = [1024] dtype = "float32" min_val = float("-5.89314") @@ -2805,6 +3060,7 @@ class Program_weight_tensor_parameter_254: class Program_weight_tensor_parameter_255: name = "parameter_255" + original_name = "linear_1217.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.352211") @@ -2816,6 +3072,7 @@ class Program_weight_tensor_parameter_255: class Program_weight_tensor_parameter_256: name = "parameter_256" + original_name = "linear_1216.b_0" shape = [1024] dtype = "float32" min_val = float("-0.460587") @@ -2827,6 +3084,7 @@ class Program_weight_tensor_parameter_256: class Program_weight_tensor_parameter_257: name = "parameter_257" + original_name = "linear_1216.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.310441") @@ -2838,6 +3096,7 @@ class Program_weight_tensor_parameter_257: class Program_weight_tensor_parameter_258: name = "parameter_258" + original_name = "layer_norm_420.b_0" shape = [1024] dtype = "float32" min_val = float("-0.786415") @@ -2849,6 +3108,7 @@ class Program_weight_tensor_parameter_258: class Program_weight_tensor_parameter_259: name = "parameter_259" + original_name = "layer_norm_420.w_0" shape = [1024] dtype = "float32" min_val = float("0.262968") @@ -2860,6 +3120,7 @@ class Program_weight_tensor_parameter_259: class Program_weight_tensor_parameter_260: name = "parameter_260" + original_name = "layer_norm_419.b_0" shape = [1024] dtype = "float32" min_val = float("-1.89774") @@ -2871,6 +3132,7 @@ class Program_weight_tensor_parameter_260: class Program_weight_tensor_parameter_261: name = "parameter_261" + original_name = "layer_norm_419.w_0" shape = [1024] dtype = "float32" min_val = float("0.772355") @@ -2882,6 +3144,7 @@ class Program_weight_tensor_parameter_261: class Program_weight_tensor_parameter_262: name = "parameter_262" + original_name = "linear_1215.b_0" shape = [1024] dtype = "float32" min_val = float("-0.388127") @@ -2893,6 +3156,7 @@ class Program_weight_tensor_parameter_262: class Program_weight_tensor_parameter_263: name = "parameter_263" + original_name = "linear_1215.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-3.55579") @@ -2904,6 +3168,7 @@ class Program_weight_tensor_parameter_263: class Program_weight_tensor_parameter_264: name = "parameter_264" + original_name = "linear_1214.b_0" shape = [4096] dtype = "float32" min_val = float("-0.150666") @@ -2915,6 +3180,7 @@ class Program_weight_tensor_parameter_264: class Program_weight_tensor_parameter_265: name = "parameter_265" + original_name = "linear_1214.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.619748") @@ -2926,6 +3192,7 @@ class Program_weight_tensor_parameter_265: class Program_weight_tensor_parameter_266: name = "parameter_266" + original_name = "linear_1213.b_0" shape = [1024] dtype = "float32" min_val = float("-0.527334") @@ -2937,6 +3204,7 @@ class Program_weight_tensor_parameter_266: class Program_weight_tensor_parameter_267: name = "parameter_267" + original_name = "linear_1213.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.510171") @@ -2948,6 +3216,7 @@ class Program_weight_tensor_parameter_267: class Program_weight_tensor_parameter_268: name = "parameter_268" + original_name = "linear_1212.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0952708") @@ -2959,6 +3228,7 @@ class Program_weight_tensor_parameter_268: class Program_weight_tensor_parameter_269: name = "parameter_269" + original_name = "linear_1212.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.250617") @@ -2970,6 +3240,7 @@ class Program_weight_tensor_parameter_269: class Program_weight_tensor_parameter_270: name = "parameter_270" + original_name = "linear_1211.b_0" shape = [1024] dtype = "float32" min_val = float("-4.6147") @@ -2981,6 +3252,7 @@ class Program_weight_tensor_parameter_270: class Program_weight_tensor_parameter_271: name = "parameter_271" + original_name = "linear_1211.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.440693") @@ -2992,6 +3264,7 @@ class Program_weight_tensor_parameter_271: class Program_weight_tensor_parameter_272: name = "parameter_272" + original_name = "linear_1210.b_0" shape = [1024] dtype = "float32" min_val = float("-0.357678") @@ -3003,6 +3276,7 @@ class Program_weight_tensor_parameter_272: class Program_weight_tensor_parameter_273: name = "parameter_273" + original_name = "linear_1210.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.297448") @@ -3014,6 +3288,7 @@ class Program_weight_tensor_parameter_273: class Program_weight_tensor_parameter_274: name = "parameter_274" + original_name = "layer_norm_418.b_0" shape = [1024] dtype = "float32" min_val = float("-0.942994") @@ -3025,6 +3300,7 @@ class Program_weight_tensor_parameter_274: class Program_weight_tensor_parameter_275: name = "parameter_275" + original_name = "layer_norm_418.w_0" shape = [1024] dtype = "float32" min_val = float("0.401152") @@ -3036,6 +3312,7 @@ class Program_weight_tensor_parameter_275: class Program_weight_tensor_parameter_276: name = "parameter_276" + original_name = "layer_norm_417.b_0" shape = [1024] dtype = "float32" min_val = float("-1.78239") @@ -3047,6 +3324,7 @@ class Program_weight_tensor_parameter_276: class Program_weight_tensor_parameter_277: name = "parameter_277" + original_name = "layer_norm_417.w_0" shape = [1024] dtype = "float32" min_val = float("0.820196") @@ -3058,6 +3336,7 @@ class Program_weight_tensor_parameter_277: class Program_weight_tensor_parameter_278: name = "parameter_278" + original_name = "linear_1209.b_0" shape = [1024] dtype = "float32" min_val = float("-0.447711") @@ -3069,6 +3348,7 @@ class Program_weight_tensor_parameter_278: class Program_weight_tensor_parameter_279: name = "parameter_279" + original_name = "linear_1209.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-1.67427") @@ -3080,6 +3360,7 @@ class Program_weight_tensor_parameter_279: class Program_weight_tensor_parameter_280: name = "parameter_280" + original_name = "linear_1208.b_0" shape = [4096] dtype = "float32" min_val = float("-0.128332") @@ -3091,6 +3372,7 @@ class Program_weight_tensor_parameter_280: class Program_weight_tensor_parameter_281: name = "parameter_281" + original_name = "linear_1208.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.337849") @@ -3102,6 +3384,7 @@ class Program_weight_tensor_parameter_281: class Program_weight_tensor_parameter_282: name = "parameter_282" + original_name = "linear_1207.b_0" shape = [1024] dtype = "float32" min_val = float("-0.304858") @@ -3113,6 +3396,7 @@ class Program_weight_tensor_parameter_282: class Program_weight_tensor_parameter_283: name = "parameter_283" + original_name = "linear_1207.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.284752") @@ -3124,6 +3408,7 @@ class Program_weight_tensor_parameter_283: class Program_weight_tensor_parameter_284: name = "parameter_284" + original_name = "linear_1206.b_0" shape = [1024] dtype = "float32" min_val = float("-0.102442") @@ -3135,6 +3420,7 @@ class Program_weight_tensor_parameter_284: class Program_weight_tensor_parameter_285: name = "parameter_285" + original_name = "linear_1206.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.194304") @@ -3146,6 +3432,7 @@ class Program_weight_tensor_parameter_285: class Program_weight_tensor_parameter_286: name = "parameter_286" + original_name = "linear_1205.b_0" shape = [1024] dtype = "float32" min_val = float("-4.72085") @@ -3157,6 +3444,7 @@ class Program_weight_tensor_parameter_286: class Program_weight_tensor_parameter_287: name = "parameter_287" + original_name = "linear_1205.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.41063") @@ -3168,6 +3456,7 @@ class Program_weight_tensor_parameter_287: class Program_weight_tensor_parameter_288: name = "parameter_288" + original_name = "linear_1204.b_0" shape = [1024] dtype = "float32" min_val = float("-0.446432") @@ -3179,6 +3468,7 @@ class Program_weight_tensor_parameter_288: class Program_weight_tensor_parameter_289: name = "parameter_289" + original_name = "linear_1204.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.315006") @@ -3190,6 +3480,7 @@ class Program_weight_tensor_parameter_289: class Program_weight_tensor_parameter_290: name = "parameter_290" + original_name = "layer_norm_416.b_0" shape = [1024] dtype = "float32" min_val = float("-0.834191") @@ -3201,6 +3492,7 @@ class Program_weight_tensor_parameter_290: class Program_weight_tensor_parameter_291: name = "parameter_291" + original_name = "layer_norm_416.w_0" shape = [1024] dtype = "float32" min_val = float("0.452877") @@ -3212,6 +3504,7 @@ class Program_weight_tensor_parameter_291: class Program_weight_tensor_parameter_292: name = "parameter_292" + original_name = "layer_norm_415.b_0" shape = [1024] dtype = "float32" min_val = float("-1.56987") @@ -3223,6 +3516,7 @@ class Program_weight_tensor_parameter_292: class Program_weight_tensor_parameter_293: name = "parameter_293" + original_name = "layer_norm_415.w_0" shape = [1024] dtype = "float32" min_val = float("0.828276") @@ -3234,6 +3528,7 @@ class Program_weight_tensor_parameter_293: class Program_weight_tensor_parameter_294: name = "parameter_294" + original_name = "linear_1203.b_0" shape = [1024] dtype = "float32" min_val = float("-0.361594") @@ -3245,6 +3540,7 @@ class Program_weight_tensor_parameter_294: class Program_weight_tensor_parameter_295: name = "parameter_295" + original_name = "linear_1203.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-1.08239") @@ -3256,6 +3552,7 @@ class Program_weight_tensor_parameter_295: class Program_weight_tensor_parameter_296: name = "parameter_296" + original_name = "linear_1202.b_0" shape = [4096] dtype = "float32" min_val = float("-0.124219") @@ -3267,6 +3564,7 @@ class Program_weight_tensor_parameter_296: class Program_weight_tensor_parameter_297: name = "parameter_297" + original_name = "linear_1202.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.319696") @@ -3278,6 +3576,7 @@ class Program_weight_tensor_parameter_297: class Program_weight_tensor_parameter_298: name = "parameter_298" + original_name = "linear_1201.b_0" shape = [1024] dtype = "float32" min_val = float("-0.31478") @@ -3289,6 +3588,7 @@ class Program_weight_tensor_parameter_298: class Program_weight_tensor_parameter_299: name = "parameter_299" + original_name = "linear_1201.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.885895") @@ -3300,6 +3600,7 @@ class Program_weight_tensor_parameter_299: class Program_weight_tensor_parameter_300: name = "parameter_300" + original_name = "linear_1200.b_0" shape = [1024] dtype = "float32" min_val = float("-0.159798") @@ -3311,6 +3612,7 @@ class Program_weight_tensor_parameter_300: class Program_weight_tensor_parameter_301: name = "parameter_301" + original_name = "linear_1200.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.209054") @@ -3322,6 +3624,7 @@ class Program_weight_tensor_parameter_301: class Program_weight_tensor_parameter_302: name = "parameter_302" + original_name = "linear_1199.b_0" shape = [1024] dtype = "float32" min_val = float("-5.15342") @@ -3333,6 +3636,7 @@ class Program_weight_tensor_parameter_302: class Program_weight_tensor_parameter_303: name = "parameter_303" + original_name = "linear_1199.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.391594") @@ -3344,6 +3648,7 @@ class Program_weight_tensor_parameter_303: class Program_weight_tensor_parameter_304: name = "parameter_304" + original_name = "linear_1198.b_0" shape = [1024] dtype = "float32" min_val = float("-0.777951") @@ -3355,6 +3660,7 @@ class Program_weight_tensor_parameter_304: class Program_weight_tensor_parameter_305: name = "parameter_305" + original_name = "linear_1198.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.377895") @@ -3366,6 +3672,7 @@ class Program_weight_tensor_parameter_305: class Program_weight_tensor_parameter_306: name = "parameter_306" + original_name = "layer_norm_414.b_0" shape = [1024] dtype = "float32" min_val = float("-0.659548") @@ -3377,6 +3684,7 @@ class Program_weight_tensor_parameter_306: class Program_weight_tensor_parameter_307: name = "parameter_307" + original_name = "layer_norm_414.w_0" shape = [1024] dtype = "float32" min_val = float("0.336522") @@ -3388,6 +3696,7 @@ class Program_weight_tensor_parameter_307: class Program_weight_tensor_parameter_308: name = "parameter_308" + original_name = "layer_norm_413.b_0" shape = [1024] dtype = "float32" min_val = float("-1.8877") @@ -3399,6 +3708,7 @@ class Program_weight_tensor_parameter_308: class Program_weight_tensor_parameter_309: name = "parameter_309" + original_name = "layer_norm_413.w_0" shape = [1024] dtype = "float32" min_val = float("0.863356") @@ -3410,6 +3720,7 @@ class Program_weight_tensor_parameter_309: class Program_weight_tensor_parameter_310: name = "parameter_310" + original_name = "linear_1197.b_0" shape = [1024] dtype = "float32" min_val = float("-0.450982") @@ -3421,6 +3732,7 @@ class Program_weight_tensor_parameter_310: class Program_weight_tensor_parameter_311: name = "parameter_311" + original_name = "linear_1197.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-1.01268") @@ -3432,6 +3744,7 @@ class Program_weight_tensor_parameter_311: class Program_weight_tensor_parameter_312: name = "parameter_312" + original_name = "linear_1196.b_0" shape = [4096] dtype = "float32" min_val = float("-0.429108") @@ -3443,6 +3756,7 @@ class Program_weight_tensor_parameter_312: class Program_weight_tensor_parameter_313: name = "parameter_313" + original_name = "linear_1196.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.507511") @@ -3454,6 +3768,7 @@ class Program_weight_tensor_parameter_313: class Program_weight_tensor_parameter_314: name = "parameter_314" + original_name = "linear_1195.b_0" shape = [1024] dtype = "float32" min_val = float("-0.304311") @@ -3465,6 +3780,7 @@ class Program_weight_tensor_parameter_314: class Program_weight_tensor_parameter_315: name = "parameter_315" + original_name = "linear_1195.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.614152") @@ -3476,6 +3792,7 @@ class Program_weight_tensor_parameter_315: class Program_weight_tensor_parameter_316: name = "parameter_316" + original_name = "linear_1194.b_0" shape = [1024] dtype = "float32" min_val = float("-0.114353") @@ -3487,6 +3804,7 @@ class Program_weight_tensor_parameter_316: class Program_weight_tensor_parameter_317: name = "parameter_317" + original_name = "linear_1194.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.184489") @@ -3498,6 +3816,7 @@ class Program_weight_tensor_parameter_317: class Program_weight_tensor_parameter_318: name = "parameter_318" + original_name = "linear_1193.b_0" shape = [1024] dtype = "float32" min_val = float("-2.2891") @@ -3509,6 +3828,7 @@ class Program_weight_tensor_parameter_318: class Program_weight_tensor_parameter_319: name = "parameter_319" + original_name = "linear_1193.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.31461") @@ -3520,6 +3840,7 @@ class Program_weight_tensor_parameter_319: class Program_weight_tensor_parameter_320: name = "parameter_320" + original_name = "linear_1192.b_0" shape = [1024] dtype = "float32" min_val = float("-0.620979") @@ -3531,6 +3852,7 @@ class Program_weight_tensor_parameter_320: class Program_weight_tensor_parameter_321: name = "parameter_321" + original_name = "linear_1192.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.242254") @@ -3542,6 +3864,7 @@ class Program_weight_tensor_parameter_321: class Program_weight_tensor_parameter_322: name = "parameter_322" + original_name = "layer_norm_412.b_0" shape = [1024] dtype = "float32" min_val = float("-1.18189") @@ -3553,6 +3876,7 @@ class Program_weight_tensor_parameter_322: class Program_weight_tensor_parameter_323: name = "parameter_323" + original_name = "layer_norm_412.w_0" shape = [1024] dtype = "float32" min_val = float("0.280754") @@ -3564,6 +3888,7 @@ class Program_weight_tensor_parameter_323: class Program_weight_tensor_parameter_324: name = "parameter_324" + original_name = "embedding_80.w_0" shape = [16, 1024] dtype = "float32" min_val = float("-0.0753294") @@ -3575,6 +3900,7 @@ class Program_weight_tensor_parameter_324: class Program_weight_tensor_parameter_325: name = "parameter_325" + original_name = "embedding_79.w_0" shape = [4, 1024] dtype = "float32" min_val = float("-0.237642") @@ -3586,6 +3912,7 @@ class Program_weight_tensor_parameter_325: class Program_weight_tensor_parameter_326: name = "parameter_326" + original_name = "embedding_78.w_0" shape = [2048, 1024] dtype = "float32" min_val = float("-0.781257") @@ -3597,6 +3924,7 @@ class Program_weight_tensor_parameter_326: class Program_weight_tensor_parameter_327: name = "parameter_327" + original_name = "embedding_77.w_0" shape = [40000, 1024] dtype = "float32" min_val = float("-1.04414") diff --git a/paddle_samples/PaddleNLP/ernie-m-large/input_meta.py b/paddle_samples/PaddleNLP/ernie-m-large/input_meta.py index 56e34626a..fcd71feb6 100644 --- a/paddle_samples/PaddleNLP/ernie-m-large/input_meta.py +++ b/paddle_samples/PaddleNLP/ernie-m-large/input_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_data_0: name = "data_0" + original_name = "input_ids" shape = [1, 21] dtype = "int64" data = [ diff --git a/paddle_samples/PaddleNLP/ernie-m-large/model.py b/paddle_samples/PaddleNLP/ernie-m-large/model.py index e1b1746a0..99ea3cc44 100644 --- a/paddle_samples/PaddleNLP/ernie-m-large/model.py +++ b/paddle_samples/PaddleNLP/ernie-m-large/model.py @@ -464,10 +464,11 @@ def forward( # pd_op.embedding: (1x21x1024xf32) <- (1x21xi64, 514x1024xf32) embedding_1 = paddle._C_ops.embedding(scale_1, parameter_388, -1, False) - del parameter_388 + del parameter_388, scale_1 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_0 = paddle._C_ops.add(embedding_0, embedding_1) + del embedding_0, embedding_1 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_0, layer_norm_1, layer_norm_2 = (lambda x, f: f(x))( @@ -476,233 +477,17 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_386, parameter_387 + del add_0, parameter_386, parameter_387 # pd_op.full: (1xf32) <- () full_5 = paddle._C_ops.full( [1], float("0.1"), paddle.float32, paddle.core.CPUPlace() ) - # pd_op.assign: (1xf32) <- (1xf32) - assign_0 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_1 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_2 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_3 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_4 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_5 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_6 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_7 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_8 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_9 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_10 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_11 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_12 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_13 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_14 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_15 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_16 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_17 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_18 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_19 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_20 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_21 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_22 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_23 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_24 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_25 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_26 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_27 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_28 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_29 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_30 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_31 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_32 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_33 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_34 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_35 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_36 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_37 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_38 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_39 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_40 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_41 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_42 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_43 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_44 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_45 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_46 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_47 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_48 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_49 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_50 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_51 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_52 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_53 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_54 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_55 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_56 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_57 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_58 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_59 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_60 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_61 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_62 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_63 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_64 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_65 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_66 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_67 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_68 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_69 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_70 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_71 = full_5 - # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_0, dropout_1 = (lambda x, f: f(x))( paddle._C_ops.dropout( - layer_norm_0, None, full_5, False, "upscale_in_train", 0, False + layer_norm_0, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -714,13 +499,14 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_1 = paddle._C_ops.add(matmul_0, parameter_384) - del parameter_384 + del matmul_0, parameter_384 # pd_op.full_int_array: (4xi64) <- () full_int_array_1 = [0, 0, 16, 64] # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_0 = paddle._C_ops.reshape(add_1, full_int_array_1) + del add_1 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_0 = paddle._C_ops.transpose(reshape_0, [0, 2, 1, 3]) @@ -732,7 +518,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_2 = paddle._C_ops.add(matmul_1, parameter_382) - del parameter_382 + del matmul_1, parameter_382 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_2 = paddle._C_ops.matmul(dropout_0, parameter_381, False, False) @@ -740,10 +526,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_3 = paddle._C_ops.add(matmul_2, parameter_380) - del parameter_380 + del matmul_2, parameter_380 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_1 = paddle._C_ops.reshape(add_2, full_int_array_1) + del add_2 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_1 = paddle._C_ops.transpose(reshape_1, [0, 2, 1, 3]) @@ -751,6 +538,7 @@ def forward( # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_2 = paddle._C_ops.reshape(add_3, full_int_array_1) + del add_3 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_2 = paddle._C_ops.transpose(reshape_2, [0, 2, 1, 3]) @@ -761,84 +549,17 @@ def forward( [1], float("0.125"), paddle.float32, paddle.core.CPUPlace() ) - # pd_op.assign: (1xf32) <- (1xf32) - assign_72 = full_6 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_73 = full_6 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_74 = full_6 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_75 = full_6 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_76 = full_6 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_77 = full_6 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_78 = full_6 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_79 = full_6 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_80 = full_6 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_81 = full_6 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_82 = full_6 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_83 = full_6 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_84 = full_6 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_85 = full_6 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_86 = full_6 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_87 = full_6 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_88 = full_6 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_89 = full_6 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_90 = full_6 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_91 = full_6 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_92 = full_6 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_93 = full_6 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_94 = full_6 - # pd_op.scale: (1x16x21x64xf32) <- (1x16x21x64xf32, 1xf32) scale_2 = paddle._C_ops.scale(transpose_0, full_6, float("0"), True) del transpose_0 # pd_op.matmul: (1x16x21x21xf32) <- (1x16x21x64xf32, 1x16x21x64xf32) matmul_3 = paddle._C_ops.matmul(scale_2, transpose_1, False, True) + del scale_2, transpose_1 # pd_op.add: (1x16x21x21xf32) <- (1x16x21x21xf32, 1x1x1x21xf32) add_4 = paddle._C_ops.add(matmul_3, unsqueeze_0) + del matmul_3 # pd_op.softmax: (1x16x21x21xf32) <- (1x16x21x21xf32) softmax_0 = paddle._C_ops.softmax(add_4, -1) @@ -847,13 +568,15 @@ def forward( # pd_op.dropout: (1x16x21x21xf32, 1x16x21x21xui8) <- (1x16x21x21xf32, None, 1xf32) dropout_2, dropout_3 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_0, None, full_5, False, "upscale_in_train", 0, False + softmax_0, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_0 # pd_op.matmul: (1x16x21x64xf32) <- (1x16x21x21xf32, 1x16x21x64xf32) matmul_4 = paddle._C_ops.matmul(dropout_2, transpose_2, False, False) + del dropout_2, transpose_2 # pd_op.transpose: (1x21x16x64xf32) <- (1x16x21x64xf32) transpose_3 = paddle._C_ops.transpose(matmul_4, [0, 2, 1, 3]) @@ -864,19 +587,20 @@ def forward( # pd_op.reshape: (1x21x1024xf32) <- (1x21x16x64xf32, 3xi64) reshape_3 = paddle._C_ops.reshape(transpose_3, full_int_array_2) + del transpose_3 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_5 = paddle._C_ops.matmul(reshape_3, parameter_379, False, False) - del parameter_379 + del parameter_379, reshape_3 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_5 = paddle._C_ops.add(matmul_5, parameter_378) - del parameter_378 + del matmul_5, parameter_378 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_4, dropout_5 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_5, None, full_5, False, "upscale_in_train", 0, False + add_5, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -884,6 +608,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_6 = paddle._C_ops.add(dropout_0, dropout_4) + del dropout_0, dropout_4 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_3, layer_norm_4, layer_norm_5 = (lambda x, f: f(x))( @@ -892,7 +617,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_372, parameter_373 + del add_6, parameter_372, parameter_373 # pd_op.matmul: (1x21x4096xf32) <- (1x21x1024xf32, 1024x4096xf32) matmul_6 = paddle._C_ops.matmul(layer_norm_3, parameter_377, False, False) @@ -900,23 +625,24 @@ def forward( # pd_op.add: (1x21x4096xf32) <- (1x21x4096xf32, 4096xf32) add_7 = paddle._C_ops.add(matmul_6, parameter_376) - del parameter_376 + del matmul_6, parameter_376 # pd_op.gelu: (1x21x4096xf32) <- (1x21x4096xf32) gelu_0 = paddle._C_ops.gelu(add_7, False) + del add_7 # pd_op.matmul: (1x21x1024xf32) <- (1x21x4096xf32, 4096x1024xf32) matmul_7 = paddle._C_ops.matmul(gelu_0, parameter_375, False, False) - del parameter_375 + del gelu_0, parameter_375 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_8 = paddle._C_ops.add(matmul_7, parameter_374) - del parameter_374 + del matmul_7, parameter_374 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_6, dropout_7 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_8, None, full_5, False, "upscale_in_train", 0, False + add_8, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -924,6 +650,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_9 = paddle._C_ops.add(layer_norm_3, dropout_6) + del dropout_6, layer_norm_3 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_6, layer_norm_7, layer_norm_8 = (lambda x, f: f(x))( @@ -932,7 +659,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_370, parameter_371 + del add_9, parameter_370, parameter_371 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_8 = paddle._C_ops.matmul(layer_norm_6, parameter_369, False, False) @@ -940,10 +667,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_10 = paddle._C_ops.add(matmul_8, parameter_368) - del parameter_368 + del matmul_8, parameter_368 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_4 = paddle._C_ops.reshape(add_10, full_int_array_1) + del add_10 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_4 = paddle._C_ops.transpose(reshape_4, [0, 2, 1, 3]) @@ -955,7 +683,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_11 = paddle._C_ops.add(matmul_9, parameter_366) - del parameter_366 + del matmul_9, parameter_366 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_10 = paddle._C_ops.matmul(layer_norm_6, parameter_365, False, False) @@ -963,10 +691,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_12 = paddle._C_ops.add(matmul_10, parameter_364) - del parameter_364 + del matmul_10, parameter_364 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_5 = paddle._C_ops.reshape(add_11, full_int_array_1) + del add_11 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_5 = paddle._C_ops.transpose(reshape_5, [0, 2, 1, 3]) @@ -974,6 +703,7 @@ def forward( # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_6 = paddle._C_ops.reshape(add_12, full_int_array_1) + del add_12 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_6 = paddle._C_ops.transpose(reshape_6, [0, 2, 1, 3]) @@ -985,9 +715,11 @@ def forward( # pd_op.matmul: (1x16x21x21xf32) <- (1x16x21x64xf32, 1x16x21x64xf32) matmul_11 = paddle._C_ops.matmul(scale_3, transpose_5, False, True) + del scale_3, transpose_5 # pd_op.add: (1x16x21x21xf32) <- (1x16x21x21xf32, 1x1x1x21xf32) add_13 = paddle._C_ops.add(matmul_11, unsqueeze_0) + del matmul_11 # pd_op.softmax: (1x16x21x21xf32) <- (1x16x21x21xf32) softmax_1 = paddle._C_ops.softmax(add_13, -1) @@ -996,13 +728,15 @@ def forward( # pd_op.dropout: (1x16x21x21xf32, 1x16x21x21xui8) <- (1x16x21x21xf32, None, 1xf32) dropout_8, dropout_9 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_1, None, full_5, False, "upscale_in_train", 0, False + softmax_1, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_1 # pd_op.matmul: (1x16x21x64xf32) <- (1x16x21x21xf32, 1x16x21x64xf32) matmul_12 = paddle._C_ops.matmul(dropout_8, transpose_6, False, False) + del dropout_8, transpose_6 # pd_op.transpose: (1x21x16x64xf32) <- (1x16x21x64xf32) transpose_7 = paddle._C_ops.transpose(matmul_12, [0, 2, 1, 3]) @@ -1010,19 +744,20 @@ def forward( # pd_op.reshape: (1x21x1024xf32) <- (1x21x16x64xf32, 3xi64) reshape_7 = paddle._C_ops.reshape(transpose_7, full_int_array_2) + del transpose_7 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_13 = paddle._C_ops.matmul(reshape_7, parameter_363, False, False) - del parameter_363 + del parameter_363, reshape_7 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_14 = paddle._C_ops.add(matmul_13, parameter_362) - del parameter_362 + del matmul_13, parameter_362 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_10, dropout_11 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_14, None, full_5, False, "upscale_in_train", 0, False + add_14, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1030,6 +765,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_15 = paddle._C_ops.add(layer_norm_6, dropout_10) + del dropout_10, layer_norm_6 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_9, layer_norm_10, layer_norm_11 = (lambda x, f: f(x))( @@ -1038,7 +774,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_356, parameter_357 + del add_15, parameter_356, parameter_357 # pd_op.matmul: (1x21x4096xf32) <- (1x21x1024xf32, 1024x4096xf32) matmul_14 = paddle._C_ops.matmul(layer_norm_9, parameter_361, False, False) @@ -1046,23 +782,24 @@ def forward( # pd_op.add: (1x21x4096xf32) <- (1x21x4096xf32, 4096xf32) add_16 = paddle._C_ops.add(matmul_14, parameter_360) - del parameter_360 + del matmul_14, parameter_360 # pd_op.gelu: (1x21x4096xf32) <- (1x21x4096xf32) gelu_1 = paddle._C_ops.gelu(add_16, False) + del add_16 # pd_op.matmul: (1x21x1024xf32) <- (1x21x4096xf32, 4096x1024xf32) matmul_15 = paddle._C_ops.matmul(gelu_1, parameter_359, False, False) - del parameter_359 + del gelu_1, parameter_359 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_17 = paddle._C_ops.add(matmul_15, parameter_358) - del parameter_358 + del matmul_15, parameter_358 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_12, dropout_13 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_17, None, full_5, False, "upscale_in_train", 0, False + add_17, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1070,6 +807,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_18 = paddle._C_ops.add(layer_norm_9, dropout_12) + del dropout_12, layer_norm_9 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_12, layer_norm_13, layer_norm_14 = (lambda x, f: f(x))( @@ -1078,7 +816,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_354, parameter_355 + del add_18, parameter_354, parameter_355 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_16 = paddle._C_ops.matmul(layer_norm_12, parameter_353, False, False) @@ -1086,10 +824,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_19 = paddle._C_ops.add(matmul_16, parameter_352) - del parameter_352 + del matmul_16, parameter_352 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_8 = paddle._C_ops.reshape(add_19, full_int_array_1) + del add_19 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_8 = paddle._C_ops.transpose(reshape_8, [0, 2, 1, 3]) @@ -1101,7 +840,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_20 = paddle._C_ops.add(matmul_17, parameter_350) - del parameter_350 + del matmul_17, parameter_350 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_18 = paddle._C_ops.matmul(layer_norm_12, parameter_349, False, False) @@ -1109,10 +848,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_21 = paddle._C_ops.add(matmul_18, parameter_348) - del parameter_348 + del matmul_18, parameter_348 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_9 = paddle._C_ops.reshape(add_20, full_int_array_1) + del add_20 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_9 = paddle._C_ops.transpose(reshape_9, [0, 2, 1, 3]) @@ -1120,6 +860,7 @@ def forward( # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_10 = paddle._C_ops.reshape(add_21, full_int_array_1) + del add_21 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_10 = paddle._C_ops.transpose(reshape_10, [0, 2, 1, 3]) @@ -1131,9 +872,11 @@ def forward( # pd_op.matmul: (1x16x21x21xf32) <- (1x16x21x64xf32, 1x16x21x64xf32) matmul_19 = paddle._C_ops.matmul(scale_4, transpose_9, False, True) + del scale_4, transpose_9 # pd_op.add: (1x16x21x21xf32) <- (1x16x21x21xf32, 1x1x1x21xf32) add_22 = paddle._C_ops.add(matmul_19, unsqueeze_0) + del matmul_19 # pd_op.softmax: (1x16x21x21xf32) <- (1x16x21x21xf32) softmax_2 = paddle._C_ops.softmax(add_22, -1) @@ -1142,13 +885,15 @@ def forward( # pd_op.dropout: (1x16x21x21xf32, 1x16x21x21xui8) <- (1x16x21x21xf32, None, 1xf32) dropout_14, dropout_15 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_2, None, full_5, False, "upscale_in_train", 0, False + softmax_2, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_2 # pd_op.matmul: (1x16x21x64xf32) <- (1x16x21x21xf32, 1x16x21x64xf32) matmul_20 = paddle._C_ops.matmul(dropout_14, transpose_10, False, False) + del dropout_14, transpose_10 # pd_op.transpose: (1x21x16x64xf32) <- (1x16x21x64xf32) transpose_11 = paddle._C_ops.transpose(matmul_20, [0, 2, 1, 3]) @@ -1156,19 +901,20 @@ def forward( # pd_op.reshape: (1x21x1024xf32) <- (1x21x16x64xf32, 3xi64) reshape_11 = paddle._C_ops.reshape(transpose_11, full_int_array_2) + del transpose_11 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_21 = paddle._C_ops.matmul(reshape_11, parameter_347, False, False) - del parameter_347 + del parameter_347, reshape_11 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_23 = paddle._C_ops.add(matmul_21, parameter_346) - del parameter_346 + del matmul_21, parameter_346 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_16, dropout_17 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_23, None, full_5, False, "upscale_in_train", 0, False + add_23, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1176,6 +922,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_24 = paddle._C_ops.add(layer_norm_12, dropout_16) + del dropout_16, layer_norm_12 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_15, layer_norm_16, layer_norm_17 = (lambda x, f: f(x))( @@ -1184,7 +931,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_340, parameter_341 + del add_24, parameter_340, parameter_341 # pd_op.matmul: (1x21x4096xf32) <- (1x21x1024xf32, 1024x4096xf32) matmul_22 = paddle._C_ops.matmul(layer_norm_15, parameter_345, False, False) @@ -1192,23 +939,24 @@ def forward( # pd_op.add: (1x21x4096xf32) <- (1x21x4096xf32, 4096xf32) add_25 = paddle._C_ops.add(matmul_22, parameter_344) - del parameter_344 + del matmul_22, parameter_344 # pd_op.gelu: (1x21x4096xf32) <- (1x21x4096xf32) gelu_2 = paddle._C_ops.gelu(add_25, False) + del add_25 # pd_op.matmul: (1x21x1024xf32) <- (1x21x4096xf32, 4096x1024xf32) matmul_23 = paddle._C_ops.matmul(gelu_2, parameter_343, False, False) - del parameter_343 + del gelu_2, parameter_343 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_26 = paddle._C_ops.add(matmul_23, parameter_342) - del parameter_342 + del matmul_23, parameter_342 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_18, dropout_19 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_26, None, full_5, False, "upscale_in_train", 0, False + add_26, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1216,6 +964,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_27 = paddle._C_ops.add(layer_norm_15, dropout_18) + del dropout_18, layer_norm_15 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_18, layer_norm_19, layer_norm_20 = (lambda x, f: f(x))( @@ -1224,7 +973,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_338, parameter_339 + del add_27, parameter_338, parameter_339 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_24 = paddle._C_ops.matmul(layer_norm_18, parameter_337, False, False) @@ -1232,10 +981,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_28 = paddle._C_ops.add(matmul_24, parameter_336) - del parameter_336 + del matmul_24, parameter_336 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_12 = paddle._C_ops.reshape(add_28, full_int_array_1) + del add_28 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_12 = paddle._C_ops.transpose(reshape_12, [0, 2, 1, 3]) @@ -1247,7 +997,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_29 = paddle._C_ops.add(matmul_25, parameter_334) - del parameter_334 + del matmul_25, parameter_334 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_26 = paddle._C_ops.matmul(layer_norm_18, parameter_333, False, False) @@ -1255,10 +1005,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_30 = paddle._C_ops.add(matmul_26, parameter_332) - del parameter_332 + del matmul_26, parameter_332 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_13 = paddle._C_ops.reshape(add_29, full_int_array_1) + del add_29 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_13 = paddle._C_ops.transpose(reshape_13, [0, 2, 1, 3]) @@ -1266,6 +1017,7 @@ def forward( # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_14 = paddle._C_ops.reshape(add_30, full_int_array_1) + del add_30 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_14 = paddle._C_ops.transpose(reshape_14, [0, 2, 1, 3]) @@ -1277,9 +1029,11 @@ def forward( # pd_op.matmul: (1x16x21x21xf32) <- (1x16x21x64xf32, 1x16x21x64xf32) matmul_27 = paddle._C_ops.matmul(scale_5, transpose_13, False, True) + del scale_5, transpose_13 # pd_op.add: (1x16x21x21xf32) <- (1x16x21x21xf32, 1x1x1x21xf32) add_31 = paddle._C_ops.add(matmul_27, unsqueeze_0) + del matmul_27 # pd_op.softmax: (1x16x21x21xf32) <- (1x16x21x21xf32) softmax_3 = paddle._C_ops.softmax(add_31, -1) @@ -1288,13 +1042,15 @@ def forward( # pd_op.dropout: (1x16x21x21xf32, 1x16x21x21xui8) <- (1x16x21x21xf32, None, 1xf32) dropout_20, dropout_21 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_3, None, full_5, False, "upscale_in_train", 0, False + softmax_3, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_3 # pd_op.matmul: (1x16x21x64xf32) <- (1x16x21x21xf32, 1x16x21x64xf32) matmul_28 = paddle._C_ops.matmul(dropout_20, transpose_14, False, False) + del dropout_20, transpose_14 # pd_op.transpose: (1x21x16x64xf32) <- (1x16x21x64xf32) transpose_15 = paddle._C_ops.transpose(matmul_28, [0, 2, 1, 3]) @@ -1302,19 +1058,20 @@ def forward( # pd_op.reshape: (1x21x1024xf32) <- (1x21x16x64xf32, 3xi64) reshape_15 = paddle._C_ops.reshape(transpose_15, full_int_array_2) + del transpose_15 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_29 = paddle._C_ops.matmul(reshape_15, parameter_331, False, False) - del parameter_331 + del parameter_331, reshape_15 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_32 = paddle._C_ops.add(matmul_29, parameter_330) - del parameter_330 + del matmul_29, parameter_330 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_22, dropout_23 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_32, None, full_5, False, "upscale_in_train", 0, False + add_32, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1322,6 +1079,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_33 = paddle._C_ops.add(layer_norm_18, dropout_22) + del dropout_22, layer_norm_18 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_21, layer_norm_22, layer_norm_23 = (lambda x, f: f(x))( @@ -1330,7 +1088,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_324, parameter_325 + del add_33, parameter_324, parameter_325 # pd_op.matmul: (1x21x4096xf32) <- (1x21x1024xf32, 1024x4096xf32) matmul_30 = paddle._C_ops.matmul(layer_norm_21, parameter_329, False, False) @@ -1338,23 +1096,24 @@ def forward( # pd_op.add: (1x21x4096xf32) <- (1x21x4096xf32, 4096xf32) add_34 = paddle._C_ops.add(matmul_30, parameter_328) - del parameter_328 + del matmul_30, parameter_328 # pd_op.gelu: (1x21x4096xf32) <- (1x21x4096xf32) gelu_3 = paddle._C_ops.gelu(add_34, False) + del add_34 # pd_op.matmul: (1x21x1024xf32) <- (1x21x4096xf32, 4096x1024xf32) matmul_31 = paddle._C_ops.matmul(gelu_3, parameter_327, False, False) - del parameter_327 + del gelu_3, parameter_327 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_35 = paddle._C_ops.add(matmul_31, parameter_326) - del parameter_326 + del matmul_31, parameter_326 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_24, dropout_25 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_35, None, full_5, False, "upscale_in_train", 0, False + add_35, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1362,6 +1121,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_36 = paddle._C_ops.add(layer_norm_21, dropout_24) + del dropout_24, layer_norm_21 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_24, layer_norm_25, layer_norm_26 = (lambda x, f: f(x))( @@ -1370,7 +1130,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_322, parameter_323 + del add_36, parameter_322, parameter_323 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_32 = paddle._C_ops.matmul(layer_norm_24, parameter_321, False, False) @@ -1378,10 +1138,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_37 = paddle._C_ops.add(matmul_32, parameter_320) - del parameter_320 + del matmul_32, parameter_320 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_16 = paddle._C_ops.reshape(add_37, full_int_array_1) + del add_37 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_16 = paddle._C_ops.transpose(reshape_16, [0, 2, 1, 3]) @@ -1393,7 +1154,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_38 = paddle._C_ops.add(matmul_33, parameter_318) - del parameter_318 + del matmul_33, parameter_318 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_34 = paddle._C_ops.matmul(layer_norm_24, parameter_317, False, False) @@ -1401,10 +1162,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_39 = paddle._C_ops.add(matmul_34, parameter_316) - del parameter_316 + del matmul_34, parameter_316 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_17 = paddle._C_ops.reshape(add_38, full_int_array_1) + del add_38 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_17 = paddle._C_ops.transpose(reshape_17, [0, 2, 1, 3]) @@ -1412,6 +1174,7 @@ def forward( # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_18 = paddle._C_ops.reshape(add_39, full_int_array_1) + del add_39 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_18 = paddle._C_ops.transpose(reshape_18, [0, 2, 1, 3]) @@ -1423,9 +1186,11 @@ def forward( # pd_op.matmul: (1x16x21x21xf32) <- (1x16x21x64xf32, 1x16x21x64xf32) matmul_35 = paddle._C_ops.matmul(scale_6, transpose_17, False, True) + del scale_6, transpose_17 # pd_op.add: (1x16x21x21xf32) <- (1x16x21x21xf32, 1x1x1x21xf32) add_40 = paddle._C_ops.add(matmul_35, unsqueeze_0) + del matmul_35 # pd_op.softmax: (1x16x21x21xf32) <- (1x16x21x21xf32) softmax_4 = paddle._C_ops.softmax(add_40, -1) @@ -1434,13 +1199,15 @@ def forward( # pd_op.dropout: (1x16x21x21xf32, 1x16x21x21xui8) <- (1x16x21x21xf32, None, 1xf32) dropout_26, dropout_27 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_4, None, full_5, False, "upscale_in_train", 0, False + softmax_4, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_4 # pd_op.matmul: (1x16x21x64xf32) <- (1x16x21x21xf32, 1x16x21x64xf32) matmul_36 = paddle._C_ops.matmul(dropout_26, transpose_18, False, False) + del dropout_26, transpose_18 # pd_op.transpose: (1x21x16x64xf32) <- (1x16x21x64xf32) transpose_19 = paddle._C_ops.transpose(matmul_36, [0, 2, 1, 3]) @@ -1448,19 +1215,20 @@ def forward( # pd_op.reshape: (1x21x1024xf32) <- (1x21x16x64xf32, 3xi64) reshape_19 = paddle._C_ops.reshape(transpose_19, full_int_array_2) + del transpose_19 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_37 = paddle._C_ops.matmul(reshape_19, parameter_315, False, False) - del parameter_315 + del parameter_315, reshape_19 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_41 = paddle._C_ops.add(matmul_37, parameter_314) - del parameter_314 + del matmul_37, parameter_314 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_28, dropout_29 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_41, None, full_5, False, "upscale_in_train", 0, False + add_41, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1468,6 +1236,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_42 = paddle._C_ops.add(layer_norm_24, dropout_28) + del dropout_28, layer_norm_24 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_27, layer_norm_28, layer_norm_29 = (lambda x, f: f(x))( @@ -1476,7 +1245,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_308, parameter_309 + del add_42, parameter_308, parameter_309 # pd_op.matmul: (1x21x4096xf32) <- (1x21x1024xf32, 1024x4096xf32) matmul_38 = paddle._C_ops.matmul(layer_norm_27, parameter_313, False, False) @@ -1484,23 +1253,24 @@ def forward( # pd_op.add: (1x21x4096xf32) <- (1x21x4096xf32, 4096xf32) add_43 = paddle._C_ops.add(matmul_38, parameter_312) - del parameter_312 + del matmul_38, parameter_312 # pd_op.gelu: (1x21x4096xf32) <- (1x21x4096xf32) gelu_4 = paddle._C_ops.gelu(add_43, False) + del add_43 # pd_op.matmul: (1x21x1024xf32) <- (1x21x4096xf32, 4096x1024xf32) matmul_39 = paddle._C_ops.matmul(gelu_4, parameter_311, False, False) - del parameter_311 + del gelu_4, parameter_311 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_44 = paddle._C_ops.add(matmul_39, parameter_310) - del parameter_310 + del matmul_39, parameter_310 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_30, dropout_31 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_44, None, full_5, False, "upscale_in_train", 0, False + add_44, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1508,6 +1278,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_45 = paddle._C_ops.add(layer_norm_27, dropout_30) + del dropout_30, layer_norm_27 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_30, layer_norm_31, layer_norm_32 = (lambda x, f: f(x))( @@ -1516,7 +1287,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_306, parameter_307 + del add_45, parameter_306, parameter_307 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_40 = paddle._C_ops.matmul(layer_norm_30, parameter_305, False, False) @@ -1524,10 +1295,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_46 = paddle._C_ops.add(matmul_40, parameter_304) - del parameter_304 + del matmul_40, parameter_304 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_20 = paddle._C_ops.reshape(add_46, full_int_array_1) + del add_46 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_20 = paddle._C_ops.transpose(reshape_20, [0, 2, 1, 3]) @@ -1539,7 +1311,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_47 = paddle._C_ops.add(matmul_41, parameter_302) - del parameter_302 + del matmul_41, parameter_302 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_42 = paddle._C_ops.matmul(layer_norm_30, parameter_301, False, False) @@ -1547,10 +1319,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_48 = paddle._C_ops.add(matmul_42, parameter_300) - del parameter_300 + del matmul_42, parameter_300 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_21 = paddle._C_ops.reshape(add_47, full_int_array_1) + del add_47 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_21 = paddle._C_ops.transpose(reshape_21, [0, 2, 1, 3]) @@ -1558,6 +1331,7 @@ def forward( # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_22 = paddle._C_ops.reshape(add_48, full_int_array_1) + del add_48 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_22 = paddle._C_ops.transpose(reshape_22, [0, 2, 1, 3]) @@ -1569,9 +1343,11 @@ def forward( # pd_op.matmul: (1x16x21x21xf32) <- (1x16x21x64xf32, 1x16x21x64xf32) matmul_43 = paddle._C_ops.matmul(scale_7, transpose_21, False, True) + del scale_7, transpose_21 # pd_op.add: (1x16x21x21xf32) <- (1x16x21x21xf32, 1x1x1x21xf32) add_49 = paddle._C_ops.add(matmul_43, unsqueeze_0) + del matmul_43 # pd_op.softmax: (1x16x21x21xf32) <- (1x16x21x21xf32) softmax_5 = paddle._C_ops.softmax(add_49, -1) @@ -1580,13 +1356,15 @@ def forward( # pd_op.dropout: (1x16x21x21xf32, 1x16x21x21xui8) <- (1x16x21x21xf32, None, 1xf32) dropout_32, dropout_33 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_5, None, full_5, False, "upscale_in_train", 0, False + softmax_5, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_5 # pd_op.matmul: (1x16x21x64xf32) <- (1x16x21x21xf32, 1x16x21x64xf32) matmul_44 = paddle._C_ops.matmul(dropout_32, transpose_22, False, False) + del dropout_32, transpose_22 # pd_op.transpose: (1x21x16x64xf32) <- (1x16x21x64xf32) transpose_23 = paddle._C_ops.transpose(matmul_44, [0, 2, 1, 3]) @@ -1594,19 +1372,20 @@ def forward( # pd_op.reshape: (1x21x1024xf32) <- (1x21x16x64xf32, 3xi64) reshape_23 = paddle._C_ops.reshape(transpose_23, full_int_array_2) + del transpose_23 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_45 = paddle._C_ops.matmul(reshape_23, parameter_299, False, False) - del parameter_299 + del parameter_299, reshape_23 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_50 = paddle._C_ops.add(matmul_45, parameter_298) - del parameter_298 + del matmul_45, parameter_298 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_34, dropout_35 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_50, None, full_5, False, "upscale_in_train", 0, False + add_50, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1614,6 +1393,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_51 = paddle._C_ops.add(layer_norm_30, dropout_34) + del dropout_34, layer_norm_30 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_33, layer_norm_34, layer_norm_35 = (lambda x, f: f(x))( @@ -1622,7 +1402,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_292, parameter_293 + del add_51, parameter_292, parameter_293 # pd_op.matmul: (1x21x4096xf32) <- (1x21x1024xf32, 1024x4096xf32) matmul_46 = paddle._C_ops.matmul(layer_norm_33, parameter_297, False, False) @@ -1630,23 +1410,24 @@ def forward( # pd_op.add: (1x21x4096xf32) <- (1x21x4096xf32, 4096xf32) add_52 = paddle._C_ops.add(matmul_46, parameter_296) - del parameter_296 + del matmul_46, parameter_296 # pd_op.gelu: (1x21x4096xf32) <- (1x21x4096xf32) gelu_5 = paddle._C_ops.gelu(add_52, False) + del add_52 # pd_op.matmul: (1x21x1024xf32) <- (1x21x4096xf32, 4096x1024xf32) matmul_47 = paddle._C_ops.matmul(gelu_5, parameter_295, False, False) - del parameter_295 + del gelu_5, parameter_295 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_53 = paddle._C_ops.add(matmul_47, parameter_294) - del parameter_294 + del matmul_47, parameter_294 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_36, dropout_37 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_53, None, full_5, False, "upscale_in_train", 0, False + add_53, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1654,6 +1435,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_54 = paddle._C_ops.add(layer_norm_33, dropout_36) + del dropout_36, layer_norm_33 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_36, layer_norm_37, layer_norm_38 = (lambda x, f: f(x))( @@ -1662,7 +1444,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_290, parameter_291 + del add_54, parameter_290, parameter_291 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_48 = paddle._C_ops.matmul(layer_norm_36, parameter_289, False, False) @@ -1670,10 +1452,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_55 = paddle._C_ops.add(matmul_48, parameter_288) - del parameter_288 + del matmul_48, parameter_288 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_24 = paddle._C_ops.reshape(add_55, full_int_array_1) + del add_55 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_24 = paddle._C_ops.transpose(reshape_24, [0, 2, 1, 3]) @@ -1685,7 +1468,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_56 = paddle._C_ops.add(matmul_49, parameter_286) - del parameter_286 + del matmul_49, parameter_286 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_50 = paddle._C_ops.matmul(layer_norm_36, parameter_285, False, False) @@ -1693,10 +1476,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_57 = paddle._C_ops.add(matmul_50, parameter_284) - del parameter_284 + del matmul_50, parameter_284 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_25 = paddle._C_ops.reshape(add_56, full_int_array_1) + del add_56 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_25 = paddle._C_ops.transpose(reshape_25, [0, 2, 1, 3]) @@ -1704,6 +1488,7 @@ def forward( # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_26 = paddle._C_ops.reshape(add_57, full_int_array_1) + del add_57 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_26 = paddle._C_ops.transpose(reshape_26, [0, 2, 1, 3]) @@ -1715,9 +1500,11 @@ def forward( # pd_op.matmul: (1x16x21x21xf32) <- (1x16x21x64xf32, 1x16x21x64xf32) matmul_51 = paddle._C_ops.matmul(scale_8, transpose_25, False, True) + del scale_8, transpose_25 # pd_op.add: (1x16x21x21xf32) <- (1x16x21x21xf32, 1x1x1x21xf32) add_58 = paddle._C_ops.add(matmul_51, unsqueeze_0) + del matmul_51 # pd_op.softmax: (1x16x21x21xf32) <- (1x16x21x21xf32) softmax_6 = paddle._C_ops.softmax(add_58, -1) @@ -1726,13 +1513,15 @@ def forward( # pd_op.dropout: (1x16x21x21xf32, 1x16x21x21xui8) <- (1x16x21x21xf32, None, 1xf32) dropout_38, dropout_39 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_6, None, full_5, False, "upscale_in_train", 0, False + softmax_6, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_6 # pd_op.matmul: (1x16x21x64xf32) <- (1x16x21x21xf32, 1x16x21x64xf32) matmul_52 = paddle._C_ops.matmul(dropout_38, transpose_26, False, False) + del dropout_38, transpose_26 # pd_op.transpose: (1x21x16x64xf32) <- (1x16x21x64xf32) transpose_27 = paddle._C_ops.transpose(matmul_52, [0, 2, 1, 3]) @@ -1740,19 +1529,20 @@ def forward( # pd_op.reshape: (1x21x1024xf32) <- (1x21x16x64xf32, 3xi64) reshape_27 = paddle._C_ops.reshape(transpose_27, full_int_array_2) + del transpose_27 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_53 = paddle._C_ops.matmul(reshape_27, parameter_283, False, False) - del parameter_283 + del parameter_283, reshape_27 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_59 = paddle._C_ops.add(matmul_53, parameter_282) - del parameter_282 + del matmul_53, parameter_282 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_40, dropout_41 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_59, None, full_5, False, "upscale_in_train", 0, False + add_59, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1760,6 +1550,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_60 = paddle._C_ops.add(layer_norm_36, dropout_40) + del dropout_40, layer_norm_36 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_39, layer_norm_40, layer_norm_41 = (lambda x, f: f(x))( @@ -1768,7 +1559,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_276, parameter_277 + del add_60, parameter_276, parameter_277 # pd_op.matmul: (1x21x4096xf32) <- (1x21x1024xf32, 1024x4096xf32) matmul_54 = paddle._C_ops.matmul(layer_norm_39, parameter_281, False, False) @@ -1776,23 +1567,24 @@ def forward( # pd_op.add: (1x21x4096xf32) <- (1x21x4096xf32, 4096xf32) add_61 = paddle._C_ops.add(matmul_54, parameter_280) - del parameter_280 + del matmul_54, parameter_280 # pd_op.gelu: (1x21x4096xf32) <- (1x21x4096xf32) gelu_6 = paddle._C_ops.gelu(add_61, False) + del add_61 # pd_op.matmul: (1x21x1024xf32) <- (1x21x4096xf32, 4096x1024xf32) matmul_55 = paddle._C_ops.matmul(gelu_6, parameter_279, False, False) - del parameter_279 + del gelu_6, parameter_279 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_62 = paddle._C_ops.add(matmul_55, parameter_278) - del parameter_278 + del matmul_55, parameter_278 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_42, dropout_43 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_62, None, full_5, False, "upscale_in_train", 0, False + add_62, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1800,6 +1592,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_63 = paddle._C_ops.add(layer_norm_39, dropout_42) + del dropout_42, layer_norm_39 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_42, layer_norm_43, layer_norm_44 = (lambda x, f: f(x))( @@ -1808,7 +1601,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_274, parameter_275 + del add_63, parameter_274, parameter_275 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_56 = paddle._C_ops.matmul(layer_norm_42, parameter_273, False, False) @@ -1816,10 +1609,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_64 = paddle._C_ops.add(matmul_56, parameter_272) - del parameter_272 + del matmul_56, parameter_272 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_28 = paddle._C_ops.reshape(add_64, full_int_array_1) + del add_64 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_28 = paddle._C_ops.transpose(reshape_28, [0, 2, 1, 3]) @@ -1831,7 +1625,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_65 = paddle._C_ops.add(matmul_57, parameter_270) - del parameter_270 + del matmul_57, parameter_270 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_58 = paddle._C_ops.matmul(layer_norm_42, parameter_269, False, False) @@ -1839,10 +1633,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_66 = paddle._C_ops.add(matmul_58, parameter_268) - del parameter_268 + del matmul_58, parameter_268 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_29 = paddle._C_ops.reshape(add_65, full_int_array_1) + del add_65 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_29 = paddle._C_ops.transpose(reshape_29, [0, 2, 1, 3]) @@ -1850,6 +1645,7 @@ def forward( # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_30 = paddle._C_ops.reshape(add_66, full_int_array_1) + del add_66 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_30 = paddle._C_ops.transpose(reshape_30, [0, 2, 1, 3]) @@ -1861,9 +1657,11 @@ def forward( # pd_op.matmul: (1x16x21x21xf32) <- (1x16x21x64xf32, 1x16x21x64xf32) matmul_59 = paddle._C_ops.matmul(scale_9, transpose_29, False, True) + del scale_9, transpose_29 # pd_op.add: (1x16x21x21xf32) <- (1x16x21x21xf32, 1x1x1x21xf32) add_67 = paddle._C_ops.add(matmul_59, unsqueeze_0) + del matmul_59 # pd_op.softmax: (1x16x21x21xf32) <- (1x16x21x21xf32) softmax_7 = paddle._C_ops.softmax(add_67, -1) @@ -1872,13 +1670,15 @@ def forward( # pd_op.dropout: (1x16x21x21xf32, 1x16x21x21xui8) <- (1x16x21x21xf32, None, 1xf32) dropout_44, dropout_45 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_7, None, full_5, False, "upscale_in_train", 0, False + softmax_7, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_7 # pd_op.matmul: (1x16x21x64xf32) <- (1x16x21x21xf32, 1x16x21x64xf32) matmul_60 = paddle._C_ops.matmul(dropout_44, transpose_30, False, False) + del dropout_44, transpose_30 # pd_op.transpose: (1x21x16x64xf32) <- (1x16x21x64xf32) transpose_31 = paddle._C_ops.transpose(matmul_60, [0, 2, 1, 3]) @@ -1886,19 +1686,20 @@ def forward( # pd_op.reshape: (1x21x1024xf32) <- (1x21x16x64xf32, 3xi64) reshape_31 = paddle._C_ops.reshape(transpose_31, full_int_array_2) + del transpose_31 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_61 = paddle._C_ops.matmul(reshape_31, parameter_267, False, False) - del parameter_267 + del parameter_267, reshape_31 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_68 = paddle._C_ops.add(matmul_61, parameter_266) - del parameter_266 + del matmul_61, parameter_266 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_46, dropout_47 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_68, None, full_5, False, "upscale_in_train", 0, False + add_68, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1906,6 +1707,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_69 = paddle._C_ops.add(layer_norm_42, dropout_46) + del dropout_46, layer_norm_42 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_45, layer_norm_46, layer_norm_47 = (lambda x, f: f(x))( @@ -1914,7 +1716,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_260, parameter_261 + del add_69, parameter_260, parameter_261 # pd_op.matmul: (1x21x4096xf32) <- (1x21x1024xf32, 1024x4096xf32) matmul_62 = paddle._C_ops.matmul(layer_norm_45, parameter_265, False, False) @@ -1922,23 +1724,24 @@ def forward( # pd_op.add: (1x21x4096xf32) <- (1x21x4096xf32, 4096xf32) add_70 = paddle._C_ops.add(matmul_62, parameter_264) - del parameter_264 + del matmul_62, parameter_264 # pd_op.gelu: (1x21x4096xf32) <- (1x21x4096xf32) gelu_7 = paddle._C_ops.gelu(add_70, False) + del add_70 # pd_op.matmul: (1x21x1024xf32) <- (1x21x4096xf32, 4096x1024xf32) matmul_63 = paddle._C_ops.matmul(gelu_7, parameter_263, False, False) - del parameter_263 + del gelu_7, parameter_263 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_71 = paddle._C_ops.add(matmul_63, parameter_262) - del parameter_262 + del matmul_63, parameter_262 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_48, dropout_49 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_71, None, full_5, False, "upscale_in_train", 0, False + add_71, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1946,6 +1749,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_72 = paddle._C_ops.add(layer_norm_45, dropout_48) + del dropout_48, layer_norm_45 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_48, layer_norm_49, layer_norm_50 = (lambda x, f: f(x))( @@ -1954,7 +1758,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_258, parameter_259 + del add_72, parameter_258, parameter_259 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_64 = paddle._C_ops.matmul(layer_norm_48, parameter_257, False, False) @@ -1962,10 +1766,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_73 = paddle._C_ops.add(matmul_64, parameter_256) - del parameter_256 + del matmul_64, parameter_256 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_32 = paddle._C_ops.reshape(add_73, full_int_array_1) + del add_73 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_32 = paddle._C_ops.transpose(reshape_32, [0, 2, 1, 3]) @@ -1977,7 +1782,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_74 = paddle._C_ops.add(matmul_65, parameter_254) - del parameter_254 + del matmul_65, parameter_254 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_66 = paddle._C_ops.matmul(layer_norm_48, parameter_253, False, False) @@ -1985,10 +1790,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_75 = paddle._C_ops.add(matmul_66, parameter_252) - del parameter_252 + del matmul_66, parameter_252 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_33 = paddle._C_ops.reshape(add_74, full_int_array_1) + del add_74 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_33 = paddle._C_ops.transpose(reshape_33, [0, 2, 1, 3]) @@ -1996,6 +1802,7 @@ def forward( # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_34 = paddle._C_ops.reshape(add_75, full_int_array_1) + del add_75 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_34 = paddle._C_ops.transpose(reshape_34, [0, 2, 1, 3]) @@ -2007,9 +1814,11 @@ def forward( # pd_op.matmul: (1x16x21x21xf32) <- (1x16x21x64xf32, 1x16x21x64xf32) matmul_67 = paddle._C_ops.matmul(scale_10, transpose_33, False, True) + del scale_10, transpose_33 # pd_op.add: (1x16x21x21xf32) <- (1x16x21x21xf32, 1x1x1x21xf32) add_76 = paddle._C_ops.add(matmul_67, unsqueeze_0) + del matmul_67 # pd_op.softmax: (1x16x21x21xf32) <- (1x16x21x21xf32) softmax_8 = paddle._C_ops.softmax(add_76, -1) @@ -2018,13 +1827,15 @@ def forward( # pd_op.dropout: (1x16x21x21xf32, 1x16x21x21xui8) <- (1x16x21x21xf32, None, 1xf32) dropout_50, dropout_51 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_8, None, full_5, False, "upscale_in_train", 0, False + softmax_8, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_8 # pd_op.matmul: (1x16x21x64xf32) <- (1x16x21x21xf32, 1x16x21x64xf32) matmul_68 = paddle._C_ops.matmul(dropout_50, transpose_34, False, False) + del dropout_50, transpose_34 # pd_op.transpose: (1x21x16x64xf32) <- (1x16x21x64xf32) transpose_35 = paddle._C_ops.transpose(matmul_68, [0, 2, 1, 3]) @@ -2032,19 +1843,20 @@ def forward( # pd_op.reshape: (1x21x1024xf32) <- (1x21x16x64xf32, 3xi64) reshape_35 = paddle._C_ops.reshape(transpose_35, full_int_array_2) + del transpose_35 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_69 = paddle._C_ops.matmul(reshape_35, parameter_251, False, False) - del parameter_251 + del parameter_251, reshape_35 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_77 = paddle._C_ops.add(matmul_69, parameter_250) - del parameter_250 + del matmul_69, parameter_250 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_52, dropout_53 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_77, None, full_5, False, "upscale_in_train", 0, False + add_77, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -2052,6 +1864,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_78 = paddle._C_ops.add(layer_norm_48, dropout_52) + del dropout_52, layer_norm_48 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_51, layer_norm_52, layer_norm_53 = (lambda x, f: f(x))( @@ -2060,7 +1873,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_244, parameter_245 + del add_78, parameter_244, parameter_245 # pd_op.matmul: (1x21x4096xf32) <- (1x21x1024xf32, 1024x4096xf32) matmul_70 = paddle._C_ops.matmul(layer_norm_51, parameter_249, False, False) @@ -2068,23 +1881,24 @@ def forward( # pd_op.add: (1x21x4096xf32) <- (1x21x4096xf32, 4096xf32) add_79 = paddle._C_ops.add(matmul_70, parameter_248) - del parameter_248 + del matmul_70, parameter_248 # pd_op.gelu: (1x21x4096xf32) <- (1x21x4096xf32) gelu_8 = paddle._C_ops.gelu(add_79, False) + del add_79 # pd_op.matmul: (1x21x1024xf32) <- (1x21x4096xf32, 4096x1024xf32) matmul_71 = paddle._C_ops.matmul(gelu_8, parameter_247, False, False) - del parameter_247 + del gelu_8, parameter_247 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_80 = paddle._C_ops.add(matmul_71, parameter_246) - del parameter_246 + del matmul_71, parameter_246 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_54, dropout_55 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_80, None, full_5, False, "upscale_in_train", 0, False + add_80, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -2092,6 +1906,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_81 = paddle._C_ops.add(layer_norm_51, dropout_54) + del dropout_54, layer_norm_51 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_54, layer_norm_55, layer_norm_56 = (lambda x, f: f(x))( @@ -2100,7 +1915,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_242, parameter_243 + del add_81, parameter_242, parameter_243 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_72 = paddle._C_ops.matmul(layer_norm_54, parameter_241, False, False) @@ -2108,10 +1923,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_82 = paddle._C_ops.add(matmul_72, parameter_240) - del parameter_240 + del matmul_72, parameter_240 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_36 = paddle._C_ops.reshape(add_82, full_int_array_1) + del add_82 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_36 = paddle._C_ops.transpose(reshape_36, [0, 2, 1, 3]) @@ -2123,7 +1939,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_83 = paddle._C_ops.add(matmul_73, parameter_238) - del parameter_238 + del matmul_73, parameter_238 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_74 = paddle._C_ops.matmul(layer_norm_54, parameter_237, False, False) @@ -2131,10 +1947,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_84 = paddle._C_ops.add(matmul_74, parameter_236) - del parameter_236 + del matmul_74, parameter_236 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_37 = paddle._C_ops.reshape(add_83, full_int_array_1) + del add_83 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_37 = paddle._C_ops.transpose(reshape_37, [0, 2, 1, 3]) @@ -2142,6 +1959,7 @@ def forward( # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_38 = paddle._C_ops.reshape(add_84, full_int_array_1) + del add_84 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_38 = paddle._C_ops.transpose(reshape_38, [0, 2, 1, 3]) @@ -2153,9 +1971,11 @@ def forward( # pd_op.matmul: (1x16x21x21xf32) <- (1x16x21x64xf32, 1x16x21x64xf32) matmul_75 = paddle._C_ops.matmul(scale_11, transpose_37, False, True) + del scale_11, transpose_37 # pd_op.add: (1x16x21x21xf32) <- (1x16x21x21xf32, 1x1x1x21xf32) add_85 = paddle._C_ops.add(matmul_75, unsqueeze_0) + del matmul_75 # pd_op.softmax: (1x16x21x21xf32) <- (1x16x21x21xf32) softmax_9 = paddle._C_ops.softmax(add_85, -1) @@ -2164,13 +1984,15 @@ def forward( # pd_op.dropout: (1x16x21x21xf32, 1x16x21x21xui8) <- (1x16x21x21xf32, None, 1xf32) dropout_56, dropout_57 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_9, None, full_5, False, "upscale_in_train", 0, False + softmax_9, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_9 # pd_op.matmul: (1x16x21x64xf32) <- (1x16x21x21xf32, 1x16x21x64xf32) matmul_76 = paddle._C_ops.matmul(dropout_56, transpose_38, False, False) + del dropout_56, transpose_38 # pd_op.transpose: (1x21x16x64xf32) <- (1x16x21x64xf32) transpose_39 = paddle._C_ops.transpose(matmul_76, [0, 2, 1, 3]) @@ -2178,19 +2000,20 @@ def forward( # pd_op.reshape: (1x21x1024xf32) <- (1x21x16x64xf32, 3xi64) reshape_39 = paddle._C_ops.reshape(transpose_39, full_int_array_2) + del transpose_39 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_77 = paddle._C_ops.matmul(reshape_39, parameter_235, False, False) - del parameter_235 + del parameter_235, reshape_39 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_86 = paddle._C_ops.add(matmul_77, parameter_234) - del parameter_234 + del matmul_77, parameter_234 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_58, dropout_59 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_86, None, full_5, False, "upscale_in_train", 0, False + add_86, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -2198,6 +2021,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_87 = paddle._C_ops.add(layer_norm_54, dropout_58) + del dropout_58, layer_norm_54 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_57, layer_norm_58, layer_norm_59 = (lambda x, f: f(x))( @@ -2206,7 +2030,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_228, parameter_229 + del add_87, parameter_228, parameter_229 # pd_op.matmul: (1x21x4096xf32) <- (1x21x1024xf32, 1024x4096xf32) matmul_78 = paddle._C_ops.matmul(layer_norm_57, parameter_233, False, False) @@ -2214,23 +2038,24 @@ def forward( # pd_op.add: (1x21x4096xf32) <- (1x21x4096xf32, 4096xf32) add_88 = paddle._C_ops.add(matmul_78, parameter_232) - del parameter_232 + del matmul_78, parameter_232 # pd_op.gelu: (1x21x4096xf32) <- (1x21x4096xf32) gelu_9 = paddle._C_ops.gelu(add_88, False) + del add_88 # pd_op.matmul: (1x21x1024xf32) <- (1x21x4096xf32, 4096x1024xf32) matmul_79 = paddle._C_ops.matmul(gelu_9, parameter_231, False, False) - del parameter_231 + del gelu_9, parameter_231 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_89 = paddle._C_ops.add(matmul_79, parameter_230) - del parameter_230 + del matmul_79, parameter_230 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_60, dropout_61 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_89, None, full_5, False, "upscale_in_train", 0, False + add_89, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -2238,6 +2063,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_90 = paddle._C_ops.add(layer_norm_57, dropout_60) + del dropout_60, layer_norm_57 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_60, layer_norm_61, layer_norm_62 = (lambda x, f: f(x))( @@ -2246,7 +2072,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_226, parameter_227 + del add_90, parameter_226, parameter_227 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_80 = paddle._C_ops.matmul(layer_norm_60, parameter_225, False, False) @@ -2254,10 +2080,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_91 = paddle._C_ops.add(matmul_80, parameter_224) - del parameter_224 + del matmul_80, parameter_224 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_40 = paddle._C_ops.reshape(add_91, full_int_array_1) + del add_91 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_40 = paddle._C_ops.transpose(reshape_40, [0, 2, 1, 3]) @@ -2269,7 +2096,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_92 = paddle._C_ops.add(matmul_81, parameter_222) - del parameter_222 + del matmul_81, parameter_222 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_82 = paddle._C_ops.matmul(layer_norm_60, parameter_221, False, False) @@ -2277,10 +2104,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_93 = paddle._C_ops.add(matmul_82, parameter_220) - del parameter_220 + del matmul_82, parameter_220 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_41 = paddle._C_ops.reshape(add_92, full_int_array_1) + del add_92 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_41 = paddle._C_ops.transpose(reshape_41, [0, 2, 1, 3]) @@ -2288,6 +2116,7 @@ def forward( # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_42 = paddle._C_ops.reshape(add_93, full_int_array_1) + del add_93 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_42 = paddle._C_ops.transpose(reshape_42, [0, 2, 1, 3]) @@ -2299,9 +2128,11 @@ def forward( # pd_op.matmul: (1x16x21x21xf32) <- (1x16x21x64xf32, 1x16x21x64xf32) matmul_83 = paddle._C_ops.matmul(scale_12, transpose_41, False, True) + del scale_12, transpose_41 # pd_op.add: (1x16x21x21xf32) <- (1x16x21x21xf32, 1x1x1x21xf32) add_94 = paddle._C_ops.add(matmul_83, unsqueeze_0) + del matmul_83 # pd_op.softmax: (1x16x21x21xf32) <- (1x16x21x21xf32) softmax_10 = paddle._C_ops.softmax(add_94, -1) @@ -2310,13 +2141,15 @@ def forward( # pd_op.dropout: (1x16x21x21xf32, 1x16x21x21xui8) <- (1x16x21x21xf32, None, 1xf32) dropout_62, dropout_63 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_10, None, full_5, False, "upscale_in_train", 0, False + softmax_10, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_10 # pd_op.matmul: (1x16x21x64xf32) <- (1x16x21x21xf32, 1x16x21x64xf32) matmul_84 = paddle._C_ops.matmul(dropout_62, transpose_42, False, False) + del dropout_62, transpose_42 # pd_op.transpose: (1x21x16x64xf32) <- (1x16x21x64xf32) transpose_43 = paddle._C_ops.transpose(matmul_84, [0, 2, 1, 3]) @@ -2324,19 +2157,20 @@ def forward( # pd_op.reshape: (1x21x1024xf32) <- (1x21x16x64xf32, 3xi64) reshape_43 = paddle._C_ops.reshape(transpose_43, full_int_array_2) + del transpose_43 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_85 = paddle._C_ops.matmul(reshape_43, parameter_219, False, False) - del parameter_219 + del parameter_219, reshape_43 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_95 = paddle._C_ops.add(matmul_85, parameter_218) - del parameter_218 + del matmul_85, parameter_218 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_64, dropout_65 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_95, None, full_5, False, "upscale_in_train", 0, False + add_95, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -2344,6 +2178,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_96 = paddle._C_ops.add(layer_norm_60, dropout_64) + del dropout_64, layer_norm_60 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_63, layer_norm_64, layer_norm_65 = (lambda x, f: f(x))( @@ -2352,7 +2187,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_212, parameter_213 + del add_96, parameter_212, parameter_213 # pd_op.matmul: (1x21x4096xf32) <- (1x21x1024xf32, 1024x4096xf32) matmul_86 = paddle._C_ops.matmul(layer_norm_63, parameter_217, False, False) @@ -2360,23 +2195,24 @@ def forward( # pd_op.add: (1x21x4096xf32) <- (1x21x4096xf32, 4096xf32) add_97 = paddle._C_ops.add(matmul_86, parameter_216) - del parameter_216 + del matmul_86, parameter_216 # pd_op.gelu: (1x21x4096xf32) <- (1x21x4096xf32) gelu_10 = paddle._C_ops.gelu(add_97, False) + del add_97 # pd_op.matmul: (1x21x1024xf32) <- (1x21x4096xf32, 4096x1024xf32) matmul_87 = paddle._C_ops.matmul(gelu_10, parameter_215, False, False) - del parameter_215 + del gelu_10, parameter_215 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_98 = paddle._C_ops.add(matmul_87, parameter_214) - del parameter_214 + del matmul_87, parameter_214 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_66, dropout_67 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_98, None, full_5, False, "upscale_in_train", 0, False + add_98, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -2384,6 +2220,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_99 = paddle._C_ops.add(layer_norm_63, dropout_66) + del dropout_66, layer_norm_63 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_66, layer_norm_67, layer_norm_68 = (lambda x, f: f(x))( @@ -2392,7 +2229,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_210, parameter_211 + del add_99, parameter_210, parameter_211 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_88 = paddle._C_ops.matmul(layer_norm_66, parameter_209, False, False) @@ -2400,10 +2237,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_100 = paddle._C_ops.add(matmul_88, parameter_208) - del parameter_208 + del matmul_88, parameter_208 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_44 = paddle._C_ops.reshape(add_100, full_int_array_1) + del add_100 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_44 = paddle._C_ops.transpose(reshape_44, [0, 2, 1, 3]) @@ -2415,7 +2253,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_101 = paddle._C_ops.add(matmul_89, parameter_206) - del parameter_206 + del matmul_89, parameter_206 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_90 = paddle._C_ops.matmul(layer_norm_66, parameter_205, False, False) @@ -2423,10 +2261,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_102 = paddle._C_ops.add(matmul_90, parameter_204) - del parameter_204 + del matmul_90, parameter_204 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_45 = paddle._C_ops.reshape(add_101, full_int_array_1) + del add_101 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_45 = paddle._C_ops.transpose(reshape_45, [0, 2, 1, 3]) @@ -2434,6 +2273,7 @@ def forward( # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_46 = paddle._C_ops.reshape(add_102, full_int_array_1) + del add_102 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_46 = paddle._C_ops.transpose(reshape_46, [0, 2, 1, 3]) @@ -2445,9 +2285,11 @@ def forward( # pd_op.matmul: (1x16x21x21xf32) <- (1x16x21x64xf32, 1x16x21x64xf32) matmul_91 = paddle._C_ops.matmul(scale_13, transpose_45, False, True) + del scale_13, transpose_45 # pd_op.add: (1x16x21x21xf32) <- (1x16x21x21xf32, 1x1x1x21xf32) add_103 = paddle._C_ops.add(matmul_91, unsqueeze_0) + del matmul_91 # pd_op.softmax: (1x16x21x21xf32) <- (1x16x21x21xf32) softmax_11 = paddle._C_ops.softmax(add_103, -1) @@ -2456,13 +2298,15 @@ def forward( # pd_op.dropout: (1x16x21x21xf32, 1x16x21x21xui8) <- (1x16x21x21xf32, None, 1xf32) dropout_68, dropout_69 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_11, None, full_5, False, "upscale_in_train", 0, False + softmax_11, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_11 # pd_op.matmul: (1x16x21x64xf32) <- (1x16x21x21xf32, 1x16x21x64xf32) matmul_92 = paddle._C_ops.matmul(dropout_68, transpose_46, False, False) + del dropout_68, transpose_46 # pd_op.transpose: (1x21x16x64xf32) <- (1x16x21x64xf32) transpose_47 = paddle._C_ops.transpose(matmul_92, [0, 2, 1, 3]) @@ -2470,19 +2314,20 @@ def forward( # pd_op.reshape: (1x21x1024xf32) <- (1x21x16x64xf32, 3xi64) reshape_47 = paddle._C_ops.reshape(transpose_47, full_int_array_2) + del transpose_47 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_93 = paddle._C_ops.matmul(reshape_47, parameter_203, False, False) - del parameter_203 + del parameter_203, reshape_47 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_104 = paddle._C_ops.add(matmul_93, parameter_202) - del parameter_202 + del matmul_93, parameter_202 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_70, dropout_71 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_104, None, full_5, False, "upscale_in_train", 0, False + add_104, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -2490,6 +2335,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_105 = paddle._C_ops.add(layer_norm_66, dropout_70) + del dropout_70, layer_norm_66 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_69, layer_norm_70, layer_norm_71 = (lambda x, f: f(x))( @@ -2498,7 +2344,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_196, parameter_197 + del add_105, parameter_196, parameter_197 # pd_op.matmul: (1x21x4096xf32) <- (1x21x1024xf32, 1024x4096xf32) matmul_94 = paddle._C_ops.matmul(layer_norm_69, parameter_201, False, False) @@ -2506,23 +2352,24 @@ def forward( # pd_op.add: (1x21x4096xf32) <- (1x21x4096xf32, 4096xf32) add_106 = paddle._C_ops.add(matmul_94, parameter_200) - del parameter_200 + del matmul_94, parameter_200 # pd_op.gelu: (1x21x4096xf32) <- (1x21x4096xf32) gelu_11 = paddle._C_ops.gelu(add_106, False) + del add_106 # pd_op.matmul: (1x21x1024xf32) <- (1x21x4096xf32, 4096x1024xf32) matmul_95 = paddle._C_ops.matmul(gelu_11, parameter_199, False, False) - del parameter_199 + del gelu_11, parameter_199 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_107 = paddle._C_ops.add(matmul_95, parameter_198) - del parameter_198 + del matmul_95, parameter_198 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_72, dropout_73 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_107, None, full_5, False, "upscale_in_train", 0, False + add_107, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -2530,6 +2377,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_108 = paddle._C_ops.add(layer_norm_69, dropout_72) + del dropout_72, layer_norm_69 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_72, layer_norm_73, layer_norm_74 = (lambda x, f: f(x))( @@ -2538,7 +2386,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_194, parameter_195 + del add_108, parameter_194, parameter_195 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_96 = paddle._C_ops.matmul(layer_norm_72, parameter_193, False, False) @@ -2546,10 +2394,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_109 = paddle._C_ops.add(matmul_96, parameter_192) - del parameter_192 + del matmul_96, parameter_192 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_48 = paddle._C_ops.reshape(add_109, full_int_array_1) + del add_109 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_48 = paddle._C_ops.transpose(reshape_48, [0, 2, 1, 3]) @@ -2561,7 +2410,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_110 = paddle._C_ops.add(matmul_97, parameter_190) - del parameter_190 + del matmul_97, parameter_190 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_98 = paddle._C_ops.matmul(layer_norm_72, parameter_189, False, False) @@ -2569,10 +2418,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_111 = paddle._C_ops.add(matmul_98, parameter_188) - del parameter_188 + del matmul_98, parameter_188 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_49 = paddle._C_ops.reshape(add_110, full_int_array_1) + del add_110 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_49 = paddle._C_ops.transpose(reshape_49, [0, 2, 1, 3]) @@ -2580,6 +2430,7 @@ def forward( # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_50 = paddle._C_ops.reshape(add_111, full_int_array_1) + del add_111 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_50 = paddle._C_ops.transpose(reshape_50, [0, 2, 1, 3]) @@ -2591,9 +2442,11 @@ def forward( # pd_op.matmul: (1x16x21x21xf32) <- (1x16x21x64xf32, 1x16x21x64xf32) matmul_99 = paddle._C_ops.matmul(scale_14, transpose_49, False, True) + del scale_14, transpose_49 # pd_op.add: (1x16x21x21xf32) <- (1x16x21x21xf32, 1x1x1x21xf32) add_112 = paddle._C_ops.add(matmul_99, unsqueeze_0) + del matmul_99 # pd_op.softmax: (1x16x21x21xf32) <- (1x16x21x21xf32) softmax_12 = paddle._C_ops.softmax(add_112, -1) @@ -2602,13 +2455,15 @@ def forward( # pd_op.dropout: (1x16x21x21xf32, 1x16x21x21xui8) <- (1x16x21x21xf32, None, 1xf32) dropout_74, dropout_75 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_12, None, full_5, False, "upscale_in_train", 0, False + softmax_12, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_12 # pd_op.matmul: (1x16x21x64xf32) <- (1x16x21x21xf32, 1x16x21x64xf32) matmul_100 = paddle._C_ops.matmul(dropout_74, transpose_50, False, False) + del dropout_74, transpose_50 # pd_op.transpose: (1x21x16x64xf32) <- (1x16x21x64xf32) transpose_51 = paddle._C_ops.transpose(matmul_100, [0, 2, 1, 3]) @@ -2616,19 +2471,20 @@ def forward( # pd_op.reshape: (1x21x1024xf32) <- (1x21x16x64xf32, 3xi64) reshape_51 = paddle._C_ops.reshape(transpose_51, full_int_array_2) + del transpose_51 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_101 = paddle._C_ops.matmul(reshape_51, parameter_187, False, False) - del parameter_187 + del parameter_187, reshape_51 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_113 = paddle._C_ops.add(matmul_101, parameter_186) - del parameter_186 + del matmul_101, parameter_186 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_76, dropout_77 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_113, None, full_5, False, "upscale_in_train", 0, False + add_113, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -2636,6 +2492,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_114 = paddle._C_ops.add(layer_norm_72, dropout_76) + del dropout_76, layer_norm_72 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_75, layer_norm_76, layer_norm_77 = (lambda x, f: f(x))( @@ -2644,7 +2501,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_180, parameter_181 + del add_114, parameter_180, parameter_181 # pd_op.matmul: (1x21x4096xf32) <- (1x21x1024xf32, 1024x4096xf32) matmul_102 = paddle._C_ops.matmul(layer_norm_75, parameter_185, False, False) @@ -2652,23 +2509,24 @@ def forward( # pd_op.add: (1x21x4096xf32) <- (1x21x4096xf32, 4096xf32) add_115 = paddle._C_ops.add(matmul_102, parameter_184) - del parameter_184 + del matmul_102, parameter_184 # pd_op.gelu: (1x21x4096xf32) <- (1x21x4096xf32) gelu_12 = paddle._C_ops.gelu(add_115, False) + del add_115 # pd_op.matmul: (1x21x1024xf32) <- (1x21x4096xf32, 4096x1024xf32) matmul_103 = paddle._C_ops.matmul(gelu_12, parameter_183, False, False) - del parameter_183 + del gelu_12, parameter_183 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_116 = paddle._C_ops.add(matmul_103, parameter_182) - del parameter_182 + del matmul_103, parameter_182 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_78, dropout_79 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_116, None, full_5, False, "upscale_in_train", 0, False + add_116, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -2676,6 +2534,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_117 = paddle._C_ops.add(layer_norm_75, dropout_78) + del dropout_78, layer_norm_75 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_78, layer_norm_79, layer_norm_80 = (lambda x, f: f(x))( @@ -2684,7 +2543,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_178, parameter_179 + del add_117, parameter_178, parameter_179 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_104 = paddle._C_ops.matmul(layer_norm_78, parameter_177, False, False) @@ -2692,10 +2551,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_118 = paddle._C_ops.add(matmul_104, parameter_176) - del parameter_176 + del matmul_104, parameter_176 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_52 = paddle._C_ops.reshape(add_118, full_int_array_1) + del add_118 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_52 = paddle._C_ops.transpose(reshape_52, [0, 2, 1, 3]) @@ -2707,7 +2567,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_119 = paddle._C_ops.add(matmul_105, parameter_174) - del parameter_174 + del matmul_105, parameter_174 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_106 = paddle._C_ops.matmul(layer_norm_78, parameter_173, False, False) @@ -2715,10 +2575,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_120 = paddle._C_ops.add(matmul_106, parameter_172) - del parameter_172 + del matmul_106, parameter_172 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_53 = paddle._C_ops.reshape(add_119, full_int_array_1) + del add_119 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_53 = paddle._C_ops.transpose(reshape_53, [0, 2, 1, 3]) @@ -2726,6 +2587,7 @@ def forward( # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_54 = paddle._C_ops.reshape(add_120, full_int_array_1) + del add_120 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_54 = paddle._C_ops.transpose(reshape_54, [0, 2, 1, 3]) @@ -2737,9 +2599,11 @@ def forward( # pd_op.matmul: (1x16x21x21xf32) <- (1x16x21x64xf32, 1x16x21x64xf32) matmul_107 = paddle._C_ops.matmul(scale_15, transpose_53, False, True) + del scale_15, transpose_53 # pd_op.add: (1x16x21x21xf32) <- (1x16x21x21xf32, 1x1x1x21xf32) add_121 = paddle._C_ops.add(matmul_107, unsqueeze_0) + del matmul_107 # pd_op.softmax: (1x16x21x21xf32) <- (1x16x21x21xf32) softmax_13 = paddle._C_ops.softmax(add_121, -1) @@ -2748,13 +2612,15 @@ def forward( # pd_op.dropout: (1x16x21x21xf32, 1x16x21x21xui8) <- (1x16x21x21xf32, None, 1xf32) dropout_80, dropout_81 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_13, None, full_5, False, "upscale_in_train", 0, False + softmax_13, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_13 # pd_op.matmul: (1x16x21x64xf32) <- (1x16x21x21xf32, 1x16x21x64xf32) matmul_108 = paddle._C_ops.matmul(dropout_80, transpose_54, False, False) + del dropout_80, transpose_54 # pd_op.transpose: (1x21x16x64xf32) <- (1x16x21x64xf32) transpose_55 = paddle._C_ops.transpose(matmul_108, [0, 2, 1, 3]) @@ -2762,19 +2628,20 @@ def forward( # pd_op.reshape: (1x21x1024xf32) <- (1x21x16x64xf32, 3xi64) reshape_55 = paddle._C_ops.reshape(transpose_55, full_int_array_2) + del transpose_55 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_109 = paddle._C_ops.matmul(reshape_55, parameter_171, False, False) - del parameter_171 + del parameter_171, reshape_55 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_122 = paddle._C_ops.add(matmul_109, parameter_170) - del parameter_170 + del matmul_109, parameter_170 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_82, dropout_83 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_122, None, full_5, False, "upscale_in_train", 0, False + add_122, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -2782,6 +2649,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_123 = paddle._C_ops.add(layer_norm_78, dropout_82) + del dropout_82, layer_norm_78 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_81, layer_norm_82, layer_norm_83 = (lambda x, f: f(x))( @@ -2790,7 +2658,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_164, parameter_165 + del add_123, parameter_164, parameter_165 # pd_op.matmul: (1x21x4096xf32) <- (1x21x1024xf32, 1024x4096xf32) matmul_110 = paddle._C_ops.matmul(layer_norm_81, parameter_169, False, False) @@ -2798,23 +2666,24 @@ def forward( # pd_op.add: (1x21x4096xf32) <- (1x21x4096xf32, 4096xf32) add_124 = paddle._C_ops.add(matmul_110, parameter_168) - del parameter_168 + del matmul_110, parameter_168 # pd_op.gelu: (1x21x4096xf32) <- (1x21x4096xf32) gelu_13 = paddle._C_ops.gelu(add_124, False) + del add_124 # pd_op.matmul: (1x21x1024xf32) <- (1x21x4096xf32, 4096x1024xf32) matmul_111 = paddle._C_ops.matmul(gelu_13, parameter_167, False, False) - del parameter_167 + del gelu_13, parameter_167 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_125 = paddle._C_ops.add(matmul_111, parameter_166) - del parameter_166 + del matmul_111, parameter_166 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_84, dropout_85 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_125, None, full_5, False, "upscale_in_train", 0, False + add_125, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -2822,6 +2691,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_126 = paddle._C_ops.add(layer_norm_81, dropout_84) + del dropout_84, layer_norm_81 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_84, layer_norm_85, layer_norm_86 = (lambda x, f: f(x))( @@ -2830,7 +2700,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_162, parameter_163 + del add_126, parameter_162, parameter_163 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_112 = paddle._C_ops.matmul(layer_norm_84, parameter_161, False, False) @@ -2838,10 +2708,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_127 = paddle._C_ops.add(matmul_112, parameter_160) - del parameter_160 + del matmul_112, parameter_160 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_56 = paddle._C_ops.reshape(add_127, full_int_array_1) + del add_127 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_56 = paddle._C_ops.transpose(reshape_56, [0, 2, 1, 3]) @@ -2853,7 +2724,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_128 = paddle._C_ops.add(matmul_113, parameter_158) - del parameter_158 + del matmul_113, parameter_158 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_114 = paddle._C_ops.matmul(layer_norm_84, parameter_157, False, False) @@ -2861,10 +2732,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_129 = paddle._C_ops.add(matmul_114, parameter_156) - del parameter_156 + del matmul_114, parameter_156 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_57 = paddle._C_ops.reshape(add_128, full_int_array_1) + del add_128 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_57 = paddle._C_ops.transpose(reshape_57, [0, 2, 1, 3]) @@ -2872,6 +2744,7 @@ def forward( # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_58 = paddle._C_ops.reshape(add_129, full_int_array_1) + del add_129 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_58 = paddle._C_ops.transpose(reshape_58, [0, 2, 1, 3]) @@ -2883,9 +2756,11 @@ def forward( # pd_op.matmul: (1x16x21x21xf32) <- (1x16x21x64xf32, 1x16x21x64xf32) matmul_115 = paddle._C_ops.matmul(scale_16, transpose_57, False, True) + del scale_16, transpose_57 # pd_op.add: (1x16x21x21xf32) <- (1x16x21x21xf32, 1x1x1x21xf32) add_130 = paddle._C_ops.add(matmul_115, unsqueeze_0) + del matmul_115 # pd_op.softmax: (1x16x21x21xf32) <- (1x16x21x21xf32) softmax_14 = paddle._C_ops.softmax(add_130, -1) @@ -2894,13 +2769,15 @@ def forward( # pd_op.dropout: (1x16x21x21xf32, 1x16x21x21xui8) <- (1x16x21x21xf32, None, 1xf32) dropout_86, dropout_87 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_14, None, full_5, False, "upscale_in_train", 0, False + softmax_14, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_14 # pd_op.matmul: (1x16x21x64xf32) <- (1x16x21x21xf32, 1x16x21x64xf32) matmul_116 = paddle._C_ops.matmul(dropout_86, transpose_58, False, False) + del dropout_86, transpose_58 # pd_op.transpose: (1x21x16x64xf32) <- (1x16x21x64xf32) transpose_59 = paddle._C_ops.transpose(matmul_116, [0, 2, 1, 3]) @@ -2908,19 +2785,20 @@ def forward( # pd_op.reshape: (1x21x1024xf32) <- (1x21x16x64xf32, 3xi64) reshape_59 = paddle._C_ops.reshape(transpose_59, full_int_array_2) + del transpose_59 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_117 = paddle._C_ops.matmul(reshape_59, parameter_155, False, False) - del parameter_155 + del parameter_155, reshape_59 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_131 = paddle._C_ops.add(matmul_117, parameter_154) - del parameter_154 + del matmul_117, parameter_154 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_88, dropout_89 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_131, None, full_5, False, "upscale_in_train", 0, False + add_131, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -2928,6 +2806,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_132 = paddle._C_ops.add(layer_norm_84, dropout_88) + del dropout_88, layer_norm_84 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_87, layer_norm_88, layer_norm_89 = (lambda x, f: f(x))( @@ -2936,7 +2815,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_148, parameter_149 + del add_132, parameter_148, parameter_149 # pd_op.matmul: (1x21x4096xf32) <- (1x21x1024xf32, 1024x4096xf32) matmul_118 = paddle._C_ops.matmul(layer_norm_87, parameter_153, False, False) @@ -2944,23 +2823,24 @@ def forward( # pd_op.add: (1x21x4096xf32) <- (1x21x4096xf32, 4096xf32) add_133 = paddle._C_ops.add(matmul_118, parameter_152) - del parameter_152 + del matmul_118, parameter_152 # pd_op.gelu: (1x21x4096xf32) <- (1x21x4096xf32) gelu_14 = paddle._C_ops.gelu(add_133, False) + del add_133 # pd_op.matmul: (1x21x1024xf32) <- (1x21x4096xf32, 4096x1024xf32) matmul_119 = paddle._C_ops.matmul(gelu_14, parameter_151, False, False) - del parameter_151 + del gelu_14, parameter_151 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_134 = paddle._C_ops.add(matmul_119, parameter_150) - del parameter_150 + del matmul_119, parameter_150 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_90, dropout_91 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_134, None, full_5, False, "upscale_in_train", 0, False + add_134, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -2968,6 +2848,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_135 = paddle._C_ops.add(layer_norm_87, dropout_90) + del dropout_90, layer_norm_87 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_90, layer_norm_91, layer_norm_92 = (lambda x, f: f(x))( @@ -2976,7 +2857,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_146, parameter_147 + del add_135, parameter_146, parameter_147 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_120 = paddle._C_ops.matmul(layer_norm_90, parameter_145, False, False) @@ -2984,10 +2865,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_136 = paddle._C_ops.add(matmul_120, parameter_144) - del parameter_144 + del matmul_120, parameter_144 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_60 = paddle._C_ops.reshape(add_136, full_int_array_1) + del add_136 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_60 = paddle._C_ops.transpose(reshape_60, [0, 2, 1, 3]) @@ -2999,7 +2881,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_137 = paddle._C_ops.add(matmul_121, parameter_142) - del parameter_142 + del matmul_121, parameter_142 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_122 = paddle._C_ops.matmul(layer_norm_90, parameter_141, False, False) @@ -3007,10 +2889,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_138 = paddle._C_ops.add(matmul_122, parameter_140) - del parameter_140 + del matmul_122, parameter_140 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_61 = paddle._C_ops.reshape(add_137, full_int_array_1) + del add_137 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_61 = paddle._C_ops.transpose(reshape_61, [0, 2, 1, 3]) @@ -3018,6 +2901,7 @@ def forward( # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_62 = paddle._C_ops.reshape(add_138, full_int_array_1) + del add_138 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_62 = paddle._C_ops.transpose(reshape_62, [0, 2, 1, 3]) @@ -3029,9 +2913,11 @@ def forward( # pd_op.matmul: (1x16x21x21xf32) <- (1x16x21x64xf32, 1x16x21x64xf32) matmul_123 = paddle._C_ops.matmul(scale_17, transpose_61, False, True) + del scale_17, transpose_61 # pd_op.add: (1x16x21x21xf32) <- (1x16x21x21xf32, 1x1x1x21xf32) add_139 = paddle._C_ops.add(matmul_123, unsqueeze_0) + del matmul_123 # pd_op.softmax: (1x16x21x21xf32) <- (1x16x21x21xf32) softmax_15 = paddle._C_ops.softmax(add_139, -1) @@ -3040,13 +2926,15 @@ def forward( # pd_op.dropout: (1x16x21x21xf32, 1x16x21x21xui8) <- (1x16x21x21xf32, None, 1xf32) dropout_92, dropout_93 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_15, None, full_5, False, "upscale_in_train", 0, False + softmax_15, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_15 # pd_op.matmul: (1x16x21x64xf32) <- (1x16x21x21xf32, 1x16x21x64xf32) matmul_124 = paddle._C_ops.matmul(dropout_92, transpose_62, False, False) + del dropout_92, transpose_62 # pd_op.transpose: (1x21x16x64xf32) <- (1x16x21x64xf32) transpose_63 = paddle._C_ops.transpose(matmul_124, [0, 2, 1, 3]) @@ -3054,19 +2942,20 @@ def forward( # pd_op.reshape: (1x21x1024xf32) <- (1x21x16x64xf32, 3xi64) reshape_63 = paddle._C_ops.reshape(transpose_63, full_int_array_2) + del transpose_63 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_125 = paddle._C_ops.matmul(reshape_63, parameter_139, False, False) - del parameter_139 + del parameter_139, reshape_63 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_140 = paddle._C_ops.add(matmul_125, parameter_138) - del parameter_138 + del matmul_125, parameter_138 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_94, dropout_95 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_140, None, full_5, False, "upscale_in_train", 0, False + add_140, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -3074,6 +2963,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_141 = paddle._C_ops.add(layer_norm_90, dropout_94) + del dropout_94, layer_norm_90 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_93, layer_norm_94, layer_norm_95 = (lambda x, f: f(x))( @@ -3082,7 +2972,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_132, parameter_133 + del add_141, parameter_132, parameter_133 # pd_op.matmul: (1x21x4096xf32) <- (1x21x1024xf32, 1024x4096xf32) matmul_126 = paddle._C_ops.matmul(layer_norm_93, parameter_137, False, False) @@ -3090,23 +2980,24 @@ def forward( # pd_op.add: (1x21x4096xf32) <- (1x21x4096xf32, 4096xf32) add_142 = paddle._C_ops.add(matmul_126, parameter_136) - del parameter_136 + del matmul_126, parameter_136 # pd_op.gelu: (1x21x4096xf32) <- (1x21x4096xf32) gelu_15 = paddle._C_ops.gelu(add_142, False) + del add_142 # pd_op.matmul: (1x21x1024xf32) <- (1x21x4096xf32, 4096x1024xf32) matmul_127 = paddle._C_ops.matmul(gelu_15, parameter_135, False, False) - del parameter_135 + del gelu_15, parameter_135 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_143 = paddle._C_ops.add(matmul_127, parameter_134) - del parameter_134 + del matmul_127, parameter_134 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_96, dropout_97 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_143, None, full_5, False, "upscale_in_train", 0, False + add_143, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -3114,6 +3005,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_144 = paddle._C_ops.add(layer_norm_93, dropout_96) + del dropout_96, layer_norm_93 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_96, layer_norm_97, layer_norm_98 = (lambda x, f: f(x))( @@ -3122,7 +3014,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_130, parameter_131 + del add_144, parameter_130, parameter_131 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_128 = paddle._C_ops.matmul(layer_norm_96, parameter_129, False, False) @@ -3130,10 +3022,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_145 = paddle._C_ops.add(matmul_128, parameter_128) - del parameter_128 + del matmul_128, parameter_128 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_64 = paddle._C_ops.reshape(add_145, full_int_array_1) + del add_145 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_64 = paddle._C_ops.transpose(reshape_64, [0, 2, 1, 3]) @@ -3145,7 +3038,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_146 = paddle._C_ops.add(matmul_129, parameter_126) - del parameter_126 + del matmul_129, parameter_126 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_130 = paddle._C_ops.matmul(layer_norm_96, parameter_125, False, False) @@ -3153,10 +3046,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_147 = paddle._C_ops.add(matmul_130, parameter_124) - del parameter_124 + del matmul_130, parameter_124 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_65 = paddle._C_ops.reshape(add_146, full_int_array_1) + del add_146 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_65 = paddle._C_ops.transpose(reshape_65, [0, 2, 1, 3]) @@ -3164,6 +3058,7 @@ def forward( # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_66 = paddle._C_ops.reshape(add_147, full_int_array_1) + del add_147 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_66 = paddle._C_ops.transpose(reshape_66, [0, 2, 1, 3]) @@ -3175,9 +3070,11 @@ def forward( # pd_op.matmul: (1x16x21x21xf32) <- (1x16x21x64xf32, 1x16x21x64xf32) matmul_131 = paddle._C_ops.matmul(scale_18, transpose_65, False, True) + del scale_18, transpose_65 # pd_op.add: (1x16x21x21xf32) <- (1x16x21x21xf32, 1x1x1x21xf32) add_148 = paddle._C_ops.add(matmul_131, unsqueeze_0) + del matmul_131 # pd_op.softmax: (1x16x21x21xf32) <- (1x16x21x21xf32) softmax_16 = paddle._C_ops.softmax(add_148, -1) @@ -3186,13 +3083,15 @@ def forward( # pd_op.dropout: (1x16x21x21xf32, 1x16x21x21xui8) <- (1x16x21x21xf32, None, 1xf32) dropout_98, dropout_99 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_16, None, full_5, False, "upscale_in_train", 0, False + softmax_16, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_16 # pd_op.matmul: (1x16x21x64xf32) <- (1x16x21x21xf32, 1x16x21x64xf32) matmul_132 = paddle._C_ops.matmul(dropout_98, transpose_66, False, False) + del dropout_98, transpose_66 # pd_op.transpose: (1x21x16x64xf32) <- (1x16x21x64xf32) transpose_67 = paddle._C_ops.transpose(matmul_132, [0, 2, 1, 3]) @@ -3200,19 +3099,20 @@ def forward( # pd_op.reshape: (1x21x1024xf32) <- (1x21x16x64xf32, 3xi64) reshape_67 = paddle._C_ops.reshape(transpose_67, full_int_array_2) + del transpose_67 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_133 = paddle._C_ops.matmul(reshape_67, parameter_123, False, False) - del parameter_123 + del parameter_123, reshape_67 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_149 = paddle._C_ops.add(matmul_133, parameter_122) - del parameter_122 + del matmul_133, parameter_122 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_100, dropout_101 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_149, None, full_5, False, "upscale_in_train", 0, False + add_149, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -3220,6 +3120,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_150 = paddle._C_ops.add(layer_norm_96, dropout_100) + del dropout_100, layer_norm_96 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_99, layer_norm_100, layer_norm_101 = (lambda x, f: f(x))( @@ -3228,7 +3129,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_116, parameter_117 + del add_150, parameter_116, parameter_117 # pd_op.matmul: (1x21x4096xf32) <- (1x21x1024xf32, 1024x4096xf32) matmul_134 = paddle._C_ops.matmul(layer_norm_99, parameter_121, False, False) @@ -3236,23 +3137,24 @@ def forward( # pd_op.add: (1x21x4096xf32) <- (1x21x4096xf32, 4096xf32) add_151 = paddle._C_ops.add(matmul_134, parameter_120) - del parameter_120 + del matmul_134, parameter_120 # pd_op.gelu: (1x21x4096xf32) <- (1x21x4096xf32) gelu_16 = paddle._C_ops.gelu(add_151, False) + del add_151 # pd_op.matmul: (1x21x1024xf32) <- (1x21x4096xf32, 4096x1024xf32) matmul_135 = paddle._C_ops.matmul(gelu_16, parameter_119, False, False) - del parameter_119 + del gelu_16, parameter_119 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_152 = paddle._C_ops.add(matmul_135, parameter_118) - del parameter_118 + del matmul_135, parameter_118 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_102, dropout_103 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_152, None, full_5, False, "upscale_in_train", 0, False + add_152, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -3260,6 +3162,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_153 = paddle._C_ops.add(layer_norm_99, dropout_102) + del dropout_102, layer_norm_99 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_102, layer_norm_103, layer_norm_104 = (lambda x, f: f(x))( @@ -3268,7 +3171,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_114, parameter_115 + del add_153, parameter_114, parameter_115 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_136 = paddle._C_ops.matmul(layer_norm_102, parameter_113, False, False) @@ -3276,10 +3179,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_154 = paddle._C_ops.add(matmul_136, parameter_112) - del parameter_112 + del matmul_136, parameter_112 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_68 = paddle._C_ops.reshape(add_154, full_int_array_1) + del add_154 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_68 = paddle._C_ops.transpose(reshape_68, [0, 2, 1, 3]) @@ -3291,7 +3195,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_155 = paddle._C_ops.add(matmul_137, parameter_110) - del parameter_110 + del matmul_137, parameter_110 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_138 = paddle._C_ops.matmul(layer_norm_102, parameter_109, False, False) @@ -3299,10 +3203,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_156 = paddle._C_ops.add(matmul_138, parameter_108) - del parameter_108 + del matmul_138, parameter_108 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_69 = paddle._C_ops.reshape(add_155, full_int_array_1) + del add_155 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_69 = paddle._C_ops.transpose(reshape_69, [0, 2, 1, 3]) @@ -3310,6 +3215,7 @@ def forward( # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_70 = paddle._C_ops.reshape(add_156, full_int_array_1) + del add_156 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_70 = paddle._C_ops.transpose(reshape_70, [0, 2, 1, 3]) @@ -3321,9 +3227,11 @@ def forward( # pd_op.matmul: (1x16x21x21xf32) <- (1x16x21x64xf32, 1x16x21x64xf32) matmul_139 = paddle._C_ops.matmul(scale_19, transpose_69, False, True) + del scale_19, transpose_69 # pd_op.add: (1x16x21x21xf32) <- (1x16x21x21xf32, 1x1x1x21xf32) add_157 = paddle._C_ops.add(matmul_139, unsqueeze_0) + del matmul_139 # pd_op.softmax: (1x16x21x21xf32) <- (1x16x21x21xf32) softmax_17 = paddle._C_ops.softmax(add_157, -1) @@ -3332,13 +3240,15 @@ def forward( # pd_op.dropout: (1x16x21x21xf32, 1x16x21x21xui8) <- (1x16x21x21xf32, None, 1xf32) dropout_104, dropout_105 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_17, None, full_5, False, "upscale_in_train", 0, False + softmax_17, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_17 # pd_op.matmul: (1x16x21x64xf32) <- (1x16x21x21xf32, 1x16x21x64xf32) matmul_140 = paddle._C_ops.matmul(dropout_104, transpose_70, False, False) + del dropout_104, transpose_70 # pd_op.transpose: (1x21x16x64xf32) <- (1x16x21x64xf32) transpose_71 = paddle._C_ops.transpose(matmul_140, [0, 2, 1, 3]) @@ -3346,19 +3256,20 @@ def forward( # pd_op.reshape: (1x21x1024xf32) <- (1x21x16x64xf32, 3xi64) reshape_71 = paddle._C_ops.reshape(transpose_71, full_int_array_2) + del transpose_71 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_141 = paddle._C_ops.matmul(reshape_71, parameter_107, False, False) - del parameter_107 + del parameter_107, reshape_71 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_158 = paddle._C_ops.add(matmul_141, parameter_106) - del parameter_106 + del matmul_141, parameter_106 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_106, dropout_107 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_158, None, full_5, False, "upscale_in_train", 0, False + add_158, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -3366,6 +3277,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_159 = paddle._C_ops.add(layer_norm_102, dropout_106) + del dropout_106, layer_norm_102 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_105, layer_norm_106, layer_norm_107 = (lambda x, f: f(x))( @@ -3374,7 +3286,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_100, parameter_101 + del add_159, parameter_100, parameter_101 # pd_op.matmul: (1x21x4096xf32) <- (1x21x1024xf32, 1024x4096xf32) matmul_142 = paddle._C_ops.matmul(layer_norm_105, parameter_105, False, False) @@ -3382,23 +3294,24 @@ def forward( # pd_op.add: (1x21x4096xf32) <- (1x21x4096xf32, 4096xf32) add_160 = paddle._C_ops.add(matmul_142, parameter_104) - del parameter_104 + del matmul_142, parameter_104 # pd_op.gelu: (1x21x4096xf32) <- (1x21x4096xf32) gelu_17 = paddle._C_ops.gelu(add_160, False) + del add_160 # pd_op.matmul: (1x21x1024xf32) <- (1x21x4096xf32, 4096x1024xf32) matmul_143 = paddle._C_ops.matmul(gelu_17, parameter_103, False, False) - del parameter_103 + del gelu_17, parameter_103 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_161 = paddle._C_ops.add(matmul_143, parameter_102) - del parameter_102 + del matmul_143, parameter_102 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_108, dropout_109 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_161, None, full_5, False, "upscale_in_train", 0, False + add_161, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -3406,6 +3319,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_162 = paddle._C_ops.add(layer_norm_105, dropout_108) + del dropout_108, layer_norm_105 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_108, layer_norm_109, layer_norm_110 = (lambda x, f: f(x))( @@ -3414,7 +3328,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_98, parameter_99 + del add_162, parameter_98, parameter_99 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_144 = paddle._C_ops.matmul(layer_norm_108, parameter_97, False, False) @@ -3422,10 +3336,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_163 = paddle._C_ops.add(matmul_144, parameter_96) - del parameter_96 + del matmul_144, parameter_96 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_72 = paddle._C_ops.reshape(add_163, full_int_array_1) + del add_163 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_72 = paddle._C_ops.transpose(reshape_72, [0, 2, 1, 3]) @@ -3437,7 +3352,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_164 = paddle._C_ops.add(matmul_145, parameter_94) - del parameter_94 + del matmul_145, parameter_94 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_146 = paddle._C_ops.matmul(layer_norm_108, parameter_93, False, False) @@ -3445,10 +3360,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_165 = paddle._C_ops.add(matmul_146, parameter_92) - del parameter_92 + del matmul_146, parameter_92 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_73 = paddle._C_ops.reshape(add_164, full_int_array_1) + del add_164 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_73 = paddle._C_ops.transpose(reshape_73, [0, 2, 1, 3]) @@ -3456,6 +3372,7 @@ def forward( # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_74 = paddle._C_ops.reshape(add_165, full_int_array_1) + del add_165 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_74 = paddle._C_ops.transpose(reshape_74, [0, 2, 1, 3]) @@ -3467,9 +3384,11 @@ def forward( # pd_op.matmul: (1x16x21x21xf32) <- (1x16x21x64xf32, 1x16x21x64xf32) matmul_147 = paddle._C_ops.matmul(scale_20, transpose_73, False, True) + del scale_20, transpose_73 # pd_op.add: (1x16x21x21xf32) <- (1x16x21x21xf32, 1x1x1x21xf32) add_166 = paddle._C_ops.add(matmul_147, unsqueeze_0) + del matmul_147 # pd_op.softmax: (1x16x21x21xf32) <- (1x16x21x21xf32) softmax_18 = paddle._C_ops.softmax(add_166, -1) @@ -3478,13 +3397,15 @@ def forward( # pd_op.dropout: (1x16x21x21xf32, 1x16x21x21xui8) <- (1x16x21x21xf32, None, 1xf32) dropout_110, dropout_111 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_18, None, full_5, False, "upscale_in_train", 0, False + softmax_18, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_18 # pd_op.matmul: (1x16x21x64xf32) <- (1x16x21x21xf32, 1x16x21x64xf32) matmul_148 = paddle._C_ops.matmul(dropout_110, transpose_74, False, False) + del dropout_110, transpose_74 # pd_op.transpose: (1x21x16x64xf32) <- (1x16x21x64xf32) transpose_75 = paddle._C_ops.transpose(matmul_148, [0, 2, 1, 3]) @@ -3492,19 +3413,20 @@ def forward( # pd_op.reshape: (1x21x1024xf32) <- (1x21x16x64xf32, 3xi64) reshape_75 = paddle._C_ops.reshape(transpose_75, full_int_array_2) + del transpose_75 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_149 = paddle._C_ops.matmul(reshape_75, parameter_91, False, False) - del parameter_91 + del parameter_91, reshape_75 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_167 = paddle._C_ops.add(matmul_149, parameter_90) - del parameter_90 + del matmul_149, parameter_90 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_112, dropout_113 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_167, None, full_5, False, "upscale_in_train", 0, False + add_167, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -3512,6 +3434,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_168 = paddle._C_ops.add(layer_norm_108, dropout_112) + del dropout_112, layer_norm_108 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_111, layer_norm_112, layer_norm_113 = (lambda x, f: f(x))( @@ -3520,7 +3443,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_84, parameter_85 + del add_168, parameter_84, parameter_85 # pd_op.matmul: (1x21x4096xf32) <- (1x21x1024xf32, 1024x4096xf32) matmul_150 = paddle._C_ops.matmul(layer_norm_111, parameter_89, False, False) @@ -3528,23 +3451,24 @@ def forward( # pd_op.add: (1x21x4096xf32) <- (1x21x4096xf32, 4096xf32) add_169 = paddle._C_ops.add(matmul_150, parameter_88) - del parameter_88 + del matmul_150, parameter_88 # pd_op.gelu: (1x21x4096xf32) <- (1x21x4096xf32) gelu_18 = paddle._C_ops.gelu(add_169, False) + del add_169 # pd_op.matmul: (1x21x1024xf32) <- (1x21x4096xf32, 4096x1024xf32) matmul_151 = paddle._C_ops.matmul(gelu_18, parameter_87, False, False) - del parameter_87 + del gelu_18, parameter_87 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_170 = paddle._C_ops.add(matmul_151, parameter_86) - del parameter_86 + del matmul_151, parameter_86 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_114, dropout_115 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_170, None, full_5, False, "upscale_in_train", 0, False + add_170, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -3552,6 +3476,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_171 = paddle._C_ops.add(layer_norm_111, dropout_114) + del dropout_114, layer_norm_111 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_114, layer_norm_115, layer_norm_116 = (lambda x, f: f(x))( @@ -3560,7 +3485,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_82, parameter_83 + del add_171, parameter_82, parameter_83 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_152 = paddle._C_ops.matmul(layer_norm_114, parameter_81, False, False) @@ -3568,10 +3493,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_172 = paddle._C_ops.add(matmul_152, parameter_80) - del parameter_80 + del matmul_152, parameter_80 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_76 = paddle._C_ops.reshape(add_172, full_int_array_1) + del add_172 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_76 = paddle._C_ops.transpose(reshape_76, [0, 2, 1, 3]) @@ -3583,7 +3509,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_173 = paddle._C_ops.add(matmul_153, parameter_78) - del parameter_78 + del matmul_153, parameter_78 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_154 = paddle._C_ops.matmul(layer_norm_114, parameter_77, False, False) @@ -3591,10 +3517,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_174 = paddle._C_ops.add(matmul_154, parameter_76) - del parameter_76 + del matmul_154, parameter_76 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_77 = paddle._C_ops.reshape(add_173, full_int_array_1) + del add_173 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_77 = paddle._C_ops.transpose(reshape_77, [0, 2, 1, 3]) @@ -3602,6 +3529,7 @@ def forward( # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_78 = paddle._C_ops.reshape(add_174, full_int_array_1) + del add_174 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_78 = paddle._C_ops.transpose(reshape_78, [0, 2, 1, 3]) @@ -3613,9 +3541,11 @@ def forward( # pd_op.matmul: (1x16x21x21xf32) <- (1x16x21x64xf32, 1x16x21x64xf32) matmul_155 = paddle._C_ops.matmul(scale_21, transpose_77, False, True) + del scale_21, transpose_77 # pd_op.add: (1x16x21x21xf32) <- (1x16x21x21xf32, 1x1x1x21xf32) add_175 = paddle._C_ops.add(matmul_155, unsqueeze_0) + del matmul_155 # pd_op.softmax: (1x16x21x21xf32) <- (1x16x21x21xf32) softmax_19 = paddle._C_ops.softmax(add_175, -1) @@ -3624,13 +3554,15 @@ def forward( # pd_op.dropout: (1x16x21x21xf32, 1x16x21x21xui8) <- (1x16x21x21xf32, None, 1xf32) dropout_116, dropout_117 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_19, None, full_5, False, "upscale_in_train", 0, False + softmax_19, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_19 # pd_op.matmul: (1x16x21x64xf32) <- (1x16x21x21xf32, 1x16x21x64xf32) matmul_156 = paddle._C_ops.matmul(dropout_116, transpose_78, False, False) + del dropout_116, transpose_78 # pd_op.transpose: (1x21x16x64xf32) <- (1x16x21x64xf32) transpose_79 = paddle._C_ops.transpose(matmul_156, [0, 2, 1, 3]) @@ -3638,19 +3570,20 @@ def forward( # pd_op.reshape: (1x21x1024xf32) <- (1x21x16x64xf32, 3xi64) reshape_79 = paddle._C_ops.reshape(transpose_79, full_int_array_2) + del transpose_79 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_157 = paddle._C_ops.matmul(reshape_79, parameter_75, False, False) - del parameter_75 + del parameter_75, reshape_79 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_176 = paddle._C_ops.add(matmul_157, parameter_74) - del parameter_74 + del matmul_157, parameter_74 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_118, dropout_119 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_176, None, full_5, False, "upscale_in_train", 0, False + add_176, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -3658,6 +3591,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_177 = paddle._C_ops.add(layer_norm_114, dropout_118) + del dropout_118, layer_norm_114 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_117, layer_norm_118, layer_norm_119 = (lambda x, f: f(x))( @@ -3666,7 +3600,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_68, parameter_69 + del add_177, parameter_68, parameter_69 # pd_op.matmul: (1x21x4096xf32) <- (1x21x1024xf32, 1024x4096xf32) matmul_158 = paddle._C_ops.matmul(layer_norm_117, parameter_73, False, False) @@ -3674,23 +3608,24 @@ def forward( # pd_op.add: (1x21x4096xf32) <- (1x21x4096xf32, 4096xf32) add_178 = paddle._C_ops.add(matmul_158, parameter_72) - del parameter_72 + del matmul_158, parameter_72 # pd_op.gelu: (1x21x4096xf32) <- (1x21x4096xf32) gelu_19 = paddle._C_ops.gelu(add_178, False) + del add_178 # pd_op.matmul: (1x21x1024xf32) <- (1x21x4096xf32, 4096x1024xf32) matmul_159 = paddle._C_ops.matmul(gelu_19, parameter_71, False, False) - del parameter_71 + del gelu_19, parameter_71 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_179 = paddle._C_ops.add(matmul_159, parameter_70) - del parameter_70 + del matmul_159, parameter_70 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_120, dropout_121 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_179, None, full_5, False, "upscale_in_train", 0, False + add_179, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -3698,6 +3633,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_180 = paddle._C_ops.add(layer_norm_117, dropout_120) + del dropout_120, layer_norm_117 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_120, layer_norm_121, layer_norm_122 = (lambda x, f: f(x))( @@ -3706,7 +3642,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_66, parameter_67 + del add_180, parameter_66, parameter_67 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_160 = paddle._C_ops.matmul(layer_norm_120, parameter_65, False, False) @@ -3714,10 +3650,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_181 = paddle._C_ops.add(matmul_160, parameter_64) - del parameter_64 + del matmul_160, parameter_64 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_80 = paddle._C_ops.reshape(add_181, full_int_array_1) + del add_181 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_80 = paddle._C_ops.transpose(reshape_80, [0, 2, 1, 3]) @@ -3729,7 +3666,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_182 = paddle._C_ops.add(matmul_161, parameter_62) - del parameter_62 + del matmul_161, parameter_62 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_162 = paddle._C_ops.matmul(layer_norm_120, parameter_61, False, False) @@ -3737,10 +3674,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_183 = paddle._C_ops.add(matmul_162, parameter_60) - del parameter_60 + del matmul_162, parameter_60 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_81 = paddle._C_ops.reshape(add_182, full_int_array_1) + del add_182 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_81 = paddle._C_ops.transpose(reshape_81, [0, 2, 1, 3]) @@ -3748,6 +3686,7 @@ def forward( # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_82 = paddle._C_ops.reshape(add_183, full_int_array_1) + del add_183 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_82 = paddle._C_ops.transpose(reshape_82, [0, 2, 1, 3]) @@ -3759,9 +3698,11 @@ def forward( # pd_op.matmul: (1x16x21x21xf32) <- (1x16x21x64xf32, 1x16x21x64xf32) matmul_163 = paddle._C_ops.matmul(scale_22, transpose_81, False, True) + del scale_22, transpose_81 # pd_op.add: (1x16x21x21xf32) <- (1x16x21x21xf32, 1x1x1x21xf32) add_184 = paddle._C_ops.add(matmul_163, unsqueeze_0) + del matmul_163 # pd_op.softmax: (1x16x21x21xf32) <- (1x16x21x21xf32) softmax_20 = paddle._C_ops.softmax(add_184, -1) @@ -3770,13 +3711,15 @@ def forward( # pd_op.dropout: (1x16x21x21xf32, 1x16x21x21xui8) <- (1x16x21x21xf32, None, 1xf32) dropout_122, dropout_123 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_20, None, full_5, False, "upscale_in_train", 0, False + softmax_20, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_20 # pd_op.matmul: (1x16x21x64xf32) <- (1x16x21x21xf32, 1x16x21x64xf32) matmul_164 = paddle._C_ops.matmul(dropout_122, transpose_82, False, False) + del dropout_122, transpose_82 # pd_op.transpose: (1x21x16x64xf32) <- (1x16x21x64xf32) transpose_83 = paddle._C_ops.transpose(matmul_164, [0, 2, 1, 3]) @@ -3784,19 +3727,20 @@ def forward( # pd_op.reshape: (1x21x1024xf32) <- (1x21x16x64xf32, 3xi64) reshape_83 = paddle._C_ops.reshape(transpose_83, full_int_array_2) + del transpose_83 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_165 = paddle._C_ops.matmul(reshape_83, parameter_59, False, False) - del parameter_59 + del parameter_59, reshape_83 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_185 = paddle._C_ops.add(matmul_165, parameter_58) - del parameter_58 + del matmul_165, parameter_58 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_124, dropout_125 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_185, None, full_5, False, "upscale_in_train", 0, False + add_185, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -3804,6 +3748,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_186 = paddle._C_ops.add(layer_norm_120, dropout_124) + del dropout_124, layer_norm_120 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_123, layer_norm_124, layer_norm_125 = (lambda x, f: f(x))( @@ -3812,7 +3757,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_52, parameter_53 + del add_186, parameter_52, parameter_53 # pd_op.matmul: (1x21x4096xf32) <- (1x21x1024xf32, 1024x4096xf32) matmul_166 = paddle._C_ops.matmul(layer_norm_123, parameter_57, False, False) @@ -3820,23 +3765,24 @@ def forward( # pd_op.add: (1x21x4096xf32) <- (1x21x4096xf32, 4096xf32) add_187 = paddle._C_ops.add(matmul_166, parameter_56) - del parameter_56 + del matmul_166, parameter_56 # pd_op.gelu: (1x21x4096xf32) <- (1x21x4096xf32) gelu_20 = paddle._C_ops.gelu(add_187, False) + del add_187 # pd_op.matmul: (1x21x1024xf32) <- (1x21x4096xf32, 4096x1024xf32) matmul_167 = paddle._C_ops.matmul(gelu_20, parameter_55, False, False) - del parameter_55 + del gelu_20, parameter_55 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_188 = paddle._C_ops.add(matmul_167, parameter_54) - del parameter_54 + del matmul_167, parameter_54 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_126, dropout_127 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_188, None, full_5, False, "upscale_in_train", 0, False + add_188, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -3844,6 +3790,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_189 = paddle._C_ops.add(layer_norm_123, dropout_126) + del dropout_126, layer_norm_123 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_126, layer_norm_127, layer_norm_128 = (lambda x, f: f(x))( @@ -3852,7 +3799,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_50, parameter_51 + del add_189, parameter_50, parameter_51 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_168 = paddle._C_ops.matmul(layer_norm_126, parameter_49, False, False) @@ -3860,10 +3807,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_190 = paddle._C_ops.add(matmul_168, parameter_48) - del parameter_48 + del matmul_168, parameter_48 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_84 = paddle._C_ops.reshape(add_190, full_int_array_1) + del add_190 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_84 = paddle._C_ops.transpose(reshape_84, [0, 2, 1, 3]) @@ -3875,7 +3823,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_191 = paddle._C_ops.add(matmul_169, parameter_46) - del parameter_46 + del matmul_169, parameter_46 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_170 = paddle._C_ops.matmul(layer_norm_126, parameter_45, False, False) @@ -3883,10 +3831,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_192 = paddle._C_ops.add(matmul_170, parameter_44) - del parameter_44 + del matmul_170, parameter_44 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_85 = paddle._C_ops.reshape(add_191, full_int_array_1) + del add_191 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_85 = paddle._C_ops.transpose(reshape_85, [0, 2, 1, 3]) @@ -3894,6 +3843,7 @@ def forward( # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_86 = paddle._C_ops.reshape(add_192, full_int_array_1) + del add_192 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_86 = paddle._C_ops.transpose(reshape_86, [0, 2, 1, 3]) @@ -3905,9 +3855,11 @@ def forward( # pd_op.matmul: (1x16x21x21xf32) <- (1x16x21x64xf32, 1x16x21x64xf32) matmul_171 = paddle._C_ops.matmul(scale_23, transpose_85, False, True) + del scale_23, transpose_85 # pd_op.add: (1x16x21x21xf32) <- (1x16x21x21xf32, 1x1x1x21xf32) add_193 = paddle._C_ops.add(matmul_171, unsqueeze_0) + del matmul_171 # pd_op.softmax: (1x16x21x21xf32) <- (1x16x21x21xf32) softmax_21 = paddle._C_ops.softmax(add_193, -1) @@ -3916,13 +3868,15 @@ def forward( # pd_op.dropout: (1x16x21x21xf32, 1x16x21x21xui8) <- (1x16x21x21xf32, None, 1xf32) dropout_128, dropout_129 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_21, None, full_5, False, "upscale_in_train", 0, False + softmax_21, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_21 # pd_op.matmul: (1x16x21x64xf32) <- (1x16x21x21xf32, 1x16x21x64xf32) matmul_172 = paddle._C_ops.matmul(dropout_128, transpose_86, False, False) + del dropout_128, transpose_86 # pd_op.transpose: (1x21x16x64xf32) <- (1x16x21x64xf32) transpose_87 = paddle._C_ops.transpose(matmul_172, [0, 2, 1, 3]) @@ -3930,19 +3884,20 @@ def forward( # pd_op.reshape: (1x21x1024xf32) <- (1x21x16x64xf32, 3xi64) reshape_87 = paddle._C_ops.reshape(transpose_87, full_int_array_2) + del transpose_87 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_173 = paddle._C_ops.matmul(reshape_87, parameter_43, False, False) - del parameter_43 + del parameter_43, reshape_87 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_194 = paddle._C_ops.add(matmul_173, parameter_42) - del parameter_42 + del matmul_173, parameter_42 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_130, dropout_131 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_194, None, full_5, False, "upscale_in_train", 0, False + add_194, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -3950,6 +3905,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_195 = paddle._C_ops.add(layer_norm_126, dropout_130) + del dropout_130, layer_norm_126 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_129, layer_norm_130, layer_norm_131 = (lambda x, f: f(x))( @@ -3958,7 +3914,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_36, parameter_37 + del add_195, parameter_36, parameter_37 # pd_op.matmul: (1x21x4096xf32) <- (1x21x1024xf32, 1024x4096xf32) matmul_174 = paddle._C_ops.matmul(layer_norm_129, parameter_41, False, False) @@ -3966,23 +3922,24 @@ def forward( # pd_op.add: (1x21x4096xf32) <- (1x21x4096xf32, 4096xf32) add_196 = paddle._C_ops.add(matmul_174, parameter_40) - del parameter_40 + del matmul_174, parameter_40 # pd_op.gelu: (1x21x4096xf32) <- (1x21x4096xf32) gelu_21 = paddle._C_ops.gelu(add_196, False) + del add_196 # pd_op.matmul: (1x21x1024xf32) <- (1x21x4096xf32, 4096x1024xf32) matmul_175 = paddle._C_ops.matmul(gelu_21, parameter_39, False, False) - del parameter_39 + del gelu_21, parameter_39 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_197 = paddle._C_ops.add(matmul_175, parameter_38) - del parameter_38 + del matmul_175, parameter_38 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_132, dropout_133 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_197, None, full_5, False, "upscale_in_train", 0, False + add_197, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -3990,6 +3947,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_198 = paddle._C_ops.add(layer_norm_129, dropout_132) + del dropout_132, layer_norm_129 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_132, layer_norm_133, layer_norm_134 = (lambda x, f: f(x))( @@ -3998,7 +3956,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_34, parameter_35 + del add_198, parameter_34, parameter_35 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_176 = paddle._C_ops.matmul(layer_norm_132, parameter_33, False, False) @@ -4006,10 +3964,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_199 = paddle._C_ops.add(matmul_176, parameter_32) - del parameter_32 + del matmul_176, parameter_32 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_88 = paddle._C_ops.reshape(add_199, full_int_array_1) + del add_199 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_88 = paddle._C_ops.transpose(reshape_88, [0, 2, 1, 3]) @@ -4021,7 +3980,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_200 = paddle._C_ops.add(matmul_177, parameter_30) - del parameter_30 + del matmul_177, parameter_30 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_178 = paddle._C_ops.matmul(layer_norm_132, parameter_29, False, False) @@ -4029,10 +3988,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_201 = paddle._C_ops.add(matmul_178, parameter_28) - del parameter_28 + del matmul_178, parameter_28 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_89 = paddle._C_ops.reshape(add_200, full_int_array_1) + del add_200 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_89 = paddle._C_ops.transpose(reshape_89, [0, 2, 1, 3]) @@ -4040,6 +4000,7 @@ def forward( # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_90 = paddle._C_ops.reshape(add_201, full_int_array_1) + del add_201 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_90 = paddle._C_ops.transpose(reshape_90, [0, 2, 1, 3]) @@ -4051,9 +4012,11 @@ def forward( # pd_op.matmul: (1x16x21x21xf32) <- (1x16x21x64xf32, 1x16x21x64xf32) matmul_179 = paddle._C_ops.matmul(scale_24, transpose_89, False, True) + del scale_24, transpose_89 # pd_op.add: (1x16x21x21xf32) <- (1x16x21x21xf32, 1x1x1x21xf32) add_202 = paddle._C_ops.add(matmul_179, unsqueeze_0) + del matmul_179 # pd_op.softmax: (1x16x21x21xf32) <- (1x16x21x21xf32) softmax_22 = paddle._C_ops.softmax(add_202, -1) @@ -4062,13 +4025,15 @@ def forward( # pd_op.dropout: (1x16x21x21xf32, 1x16x21x21xui8) <- (1x16x21x21xf32, None, 1xf32) dropout_134, dropout_135 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_22, None, full_5, False, "upscale_in_train", 0, False + softmax_22, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_22 # pd_op.matmul: (1x16x21x64xf32) <- (1x16x21x21xf32, 1x16x21x64xf32) matmul_180 = paddle._C_ops.matmul(dropout_134, transpose_90, False, False) + del dropout_134, transpose_90 # pd_op.transpose: (1x21x16x64xf32) <- (1x16x21x64xf32) transpose_91 = paddle._C_ops.transpose(matmul_180, [0, 2, 1, 3]) @@ -4076,19 +4041,20 @@ def forward( # pd_op.reshape: (1x21x1024xf32) <- (1x21x16x64xf32, 3xi64) reshape_91 = paddle._C_ops.reshape(transpose_91, full_int_array_2) + del transpose_91 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_181 = paddle._C_ops.matmul(reshape_91, parameter_27, False, False) - del parameter_27 + del parameter_27, reshape_91 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_203 = paddle._C_ops.add(matmul_181, parameter_26) - del parameter_26 + del matmul_181, parameter_26 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_136, dropout_137 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_203, None, full_5, False, "upscale_in_train", 0, False + add_203, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -4096,6 +4062,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_204 = paddle._C_ops.add(layer_norm_132, dropout_136) + del dropout_136, layer_norm_132 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_135, layer_norm_136, layer_norm_137 = (lambda x, f: f(x))( @@ -4104,7 +4071,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_20, parameter_21 + del add_204, parameter_20, parameter_21 # pd_op.matmul: (1x21x4096xf32) <- (1x21x1024xf32, 1024x4096xf32) matmul_182 = paddle._C_ops.matmul(layer_norm_135, parameter_25, False, False) @@ -4112,23 +4079,24 @@ def forward( # pd_op.add: (1x21x4096xf32) <- (1x21x4096xf32, 4096xf32) add_205 = paddle._C_ops.add(matmul_182, parameter_24) - del parameter_24 + del matmul_182, parameter_24 # pd_op.gelu: (1x21x4096xf32) <- (1x21x4096xf32) gelu_22 = paddle._C_ops.gelu(add_205, False) + del add_205 # pd_op.matmul: (1x21x1024xf32) <- (1x21x4096xf32, 4096x1024xf32) matmul_183 = paddle._C_ops.matmul(gelu_22, parameter_23, False, False) - del parameter_23 + del gelu_22, parameter_23 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_206 = paddle._C_ops.add(matmul_183, parameter_22) - del parameter_22 + del matmul_183, parameter_22 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_138, dropout_139 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_206, None, full_5, False, "upscale_in_train", 0, False + add_206, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -4136,6 +4104,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_207 = paddle._C_ops.add(layer_norm_135, dropout_138) + del dropout_138, layer_norm_135 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_138, layer_norm_139, layer_norm_140 = (lambda x, f: f(x))( @@ -4144,7 +4113,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_18, parameter_19 + del add_207, parameter_18, parameter_19 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_184 = paddle._C_ops.matmul(layer_norm_138, parameter_17, False, False) @@ -4152,10 +4121,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_208 = paddle._C_ops.add(matmul_184, parameter_16) - del parameter_16 + del matmul_184, parameter_16 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_92 = paddle._C_ops.reshape(add_208, full_int_array_1) + del add_208 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_92 = paddle._C_ops.transpose(reshape_92, [0, 2, 1, 3]) @@ -4167,7 +4137,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_209 = paddle._C_ops.add(matmul_185, parameter_14) - del parameter_14 + del matmul_185, parameter_14 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_186 = paddle._C_ops.matmul(layer_norm_138, parameter_13, False, False) @@ -4175,10 +4145,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_210 = paddle._C_ops.add(matmul_186, parameter_12) - del parameter_12 + del matmul_186, parameter_12 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_93 = paddle._C_ops.reshape(add_209, full_int_array_1) + del add_209 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_93 = paddle._C_ops.transpose(reshape_93, [0, 2, 1, 3]) @@ -4186,7 +4157,7 @@ def forward( # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_94 = paddle._C_ops.reshape(add_210, full_int_array_1) - del full_int_array_1 + del add_210, full_int_array_1 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_94 = paddle._C_ops.transpose(reshape_94, [0, 2, 1, 3]) @@ -4194,13 +4165,15 @@ def forward( # pd_op.scale: (1x16x21x64xf32) <- (1x16x21x64xf32, 1xf32) scale_25 = paddle._C_ops.scale(transpose_92, full_6, float("0"), True) - del transpose_92 + del full_6, transpose_92 # pd_op.matmul: (1x16x21x21xf32) <- (1x16x21x64xf32, 1x16x21x64xf32) matmul_187 = paddle._C_ops.matmul(scale_25, transpose_93, False, True) + del scale_25, transpose_93 # pd_op.add: (1x16x21x21xf32) <- (1x16x21x21xf32, 1x1x1x21xf32) add_211 = paddle._C_ops.add(matmul_187, unsqueeze_0) + del matmul_187, unsqueeze_0 # pd_op.softmax: (1x16x21x21xf32) <- (1x16x21x21xf32) softmax_23 = paddle._C_ops.softmax(add_211, -1) @@ -4209,13 +4182,15 @@ def forward( # pd_op.dropout: (1x16x21x21xf32, 1x16x21x21xui8) <- (1x16x21x21xf32, None, 1xf32) dropout_140, dropout_141 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_23, None, full_5, False, "upscale_in_train", 0, False + softmax_23, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_23 # pd_op.matmul: (1x16x21x64xf32) <- (1x16x21x21xf32, 1x16x21x64xf32) matmul_188 = paddle._C_ops.matmul(dropout_140, transpose_94, False, False) + del dropout_140, transpose_94 # pd_op.transpose: (1x21x16x64xf32) <- (1x16x21x64xf32) transpose_95 = paddle._C_ops.transpose(matmul_188, [0, 2, 1, 3]) @@ -4223,20 +4198,20 @@ def forward( # pd_op.reshape: (1x21x1024xf32) <- (1x21x16x64xf32, 3xi64) reshape_95 = paddle._C_ops.reshape(transpose_95, full_int_array_2) - del full_int_array_2 + del full_int_array_2, transpose_95 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_189 = paddle._C_ops.matmul(reshape_95, parameter_11, False, False) - del parameter_11 + del parameter_11, reshape_95 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_212 = paddle._C_ops.add(matmul_189, parameter_10) - del parameter_10 + del matmul_189, parameter_10 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_142, dropout_143 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_212, None, full_5, False, "upscale_in_train", 0, False + add_212, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -4244,6 +4219,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_213 = paddle._C_ops.add(layer_norm_138, dropout_142) + del dropout_142, layer_norm_138 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_141, layer_norm_142, layer_norm_143 = (lambda x, f: f(x))( @@ -4252,7 +4228,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_4, parameter_5 + del add_213, parameter_4, parameter_5 # pd_op.matmul: (1x21x4096xf32) <- (1x21x1024xf32, 1024x4096xf32) matmul_190 = paddle._C_ops.matmul(layer_norm_141, parameter_9, False, False) @@ -4260,30 +4236,32 @@ def forward( # pd_op.add: (1x21x4096xf32) <- (1x21x4096xf32, 4096xf32) add_214 = paddle._C_ops.add(matmul_190, parameter_8) - del parameter_8 + del matmul_190, parameter_8 # pd_op.gelu: (1x21x4096xf32) <- (1x21x4096xf32) gelu_23 = paddle._C_ops.gelu(add_214, False) + del add_214 # pd_op.matmul: (1x21x1024xf32) <- (1x21x4096xf32, 4096x1024xf32) matmul_191 = paddle._C_ops.matmul(gelu_23, parameter_7, False, False) - del parameter_7 + del gelu_23, parameter_7 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_215 = paddle._C_ops.add(matmul_191, parameter_6) - del parameter_6 + del matmul_191, parameter_6 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_144, dropout_145 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_215, None, full_5, False, "upscale_in_train", 0, False + add_215, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) - del add_215 + del add_215, full_5 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_216 = paddle._C_ops.add(layer_norm_141, dropout_144) + del dropout_144, layer_norm_141 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_144, layer_norm_145, layer_norm_146 = (lambda x, f: f(x))( @@ -4292,7 +4270,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_2, parameter_3 + del add_216, parameter_2, parameter_3 # pd_op.full_int_array: (1xi64) <- () full_int_array_3 = [0] @@ -4304,897 +4282,18 @@ def forward( slice_0 = paddle._C_ops.slice( layer_norm_144, [1], full_int_array_3, full_int_array_4, [1], [1] ) + del full_int_array_3, full_int_array_4, layer_norm_144 # pd_op.matmul: (1x1024xf32) <- (1x1024xf32, 1024x1024xf32) matmul_192 = paddle._C_ops.matmul(slice_0, parameter_1, False, False) - del parameter_1 + del parameter_1, slice_0 # pd_op.add: (1x1024xf32) <- (1x1024xf32, 1024xf32) add_217 = paddle._C_ops.add(matmul_192, parameter_0) - del parameter_0 + del matmul_192, parameter_0 # pd_op.tanh: (1x1024xf32) <- (1x1024xf32) tanh_0 = paddle._C_ops.tanh(add_217) - del ( - add_0, - add_1, - add_10, - add_100, - add_101, - add_102, - add_105, - add_106, - add_108, - add_109, - add_11, - add_110, - add_111, - add_114, - add_115, - add_117, - add_118, - add_119, - add_12, - add_120, - add_123, - add_124, - add_126, - add_127, - add_128, - add_129, - add_132, - add_133, - add_135, - add_136, - add_137, - add_138, - add_141, - add_142, - add_144, - add_145, - add_146, - add_147, - add_15, - add_150, - add_151, - add_153, - add_154, - add_155, - add_156, - add_159, - add_16, - add_160, - add_162, - add_163, - add_164, - add_165, - add_168, - add_169, - add_171, - add_172, - add_173, - add_174, - add_177, - add_178, - add_18, - add_180, - add_181, - add_182, - add_183, - add_186, - add_187, - add_189, - add_19, - add_190, - add_191, - add_192, - add_195, - add_196, - add_198, - add_199, - add_2, - add_20, - add_200, - add_201, - add_204, - add_205, - add_207, - add_208, - add_209, - add_21, - add_210, - add_213, - add_214, - add_216, - add_217, - add_24, - add_25, - add_27, - add_28, - add_29, - add_3, - add_30, - add_33, - add_34, - add_36, - add_37, - add_38, - add_39, - add_42, - add_43, - add_45, - add_46, - add_47, - add_48, - add_51, - add_52, - add_54, - add_55, - add_56, - add_57, - add_6, - add_60, - add_61, - add_63, - add_64, - add_65, - add_66, - add_69, - add_7, - add_70, - add_72, - add_73, - add_74, - add_75, - add_78, - add_79, - add_81, - add_82, - add_83, - add_84, - add_87, - add_88, - add_9, - add_90, - add_91, - add_92, - add_93, - add_96, - add_97, - add_99, - assign_0, - assign_1, - assign_10, - assign_11, - assign_12, - assign_13, - assign_14, - assign_15, - assign_16, - assign_17, - assign_18, - assign_19, - assign_2, - assign_20, - assign_21, - assign_22, - assign_23, - assign_24, - assign_25, - assign_26, - assign_27, - assign_28, - assign_29, - assign_3, - assign_30, - assign_31, - assign_32, - assign_33, - assign_34, - assign_35, - assign_36, - assign_37, - assign_38, - assign_39, - assign_4, - assign_40, - assign_41, - assign_42, - assign_43, - assign_44, - assign_45, - assign_46, - assign_47, - assign_48, - assign_49, - assign_5, - assign_50, - assign_51, - assign_52, - assign_53, - assign_54, - assign_55, - assign_56, - assign_57, - assign_58, - assign_59, - assign_6, - assign_60, - assign_61, - assign_62, - assign_63, - assign_64, - assign_65, - assign_66, - assign_67, - assign_68, - assign_69, - assign_7, - assign_70, - assign_71, - assign_72, - assign_73, - assign_74, - assign_75, - assign_76, - assign_77, - assign_78, - assign_79, - assign_8, - assign_80, - assign_81, - assign_82, - assign_83, - assign_84, - assign_85, - assign_86, - assign_87, - assign_88, - assign_89, - assign_9, - assign_90, - assign_91, - assign_92, - assign_93, - assign_94, - dropout_0, - dropout_1, - dropout_10, - dropout_100, - dropout_101, - dropout_102, - dropout_103, - dropout_104, - dropout_105, - dropout_106, - dropout_107, - dropout_108, - dropout_109, - dropout_11, - dropout_110, - dropout_111, - dropout_112, - dropout_113, - dropout_114, - dropout_115, - dropout_116, - dropout_117, - dropout_118, - dropout_119, - dropout_12, - dropout_120, - dropout_121, - dropout_122, - dropout_123, - dropout_124, - dropout_125, - dropout_126, - dropout_127, - dropout_128, - dropout_129, - dropout_13, - dropout_130, - dropout_131, - dropout_132, - dropout_133, - dropout_134, - dropout_135, - dropout_136, - dropout_137, - dropout_138, - dropout_139, - dropout_14, - dropout_140, - dropout_141, - dropout_142, - dropout_143, - dropout_144, - dropout_145, - dropout_15, - dropout_16, - dropout_17, - dropout_18, - dropout_19, - dropout_2, - dropout_20, - dropout_21, - dropout_22, - dropout_23, - dropout_24, - dropout_25, - dropout_26, - dropout_27, - dropout_28, - dropout_29, - dropout_3, - dropout_30, - dropout_31, - dropout_32, - dropout_33, - dropout_34, - dropout_35, - dropout_36, - dropout_37, - dropout_38, - dropout_39, - dropout_4, - dropout_40, - dropout_41, - dropout_42, - dropout_43, - dropout_44, - dropout_45, - dropout_46, - dropout_47, - dropout_48, - dropout_49, - dropout_5, - dropout_50, - dropout_51, - dropout_52, - dropout_53, - dropout_54, - dropout_55, - dropout_56, - dropout_57, - dropout_58, - dropout_59, - dropout_6, - dropout_60, - dropout_61, - dropout_62, - dropout_63, - dropout_64, - dropout_65, - dropout_66, - dropout_67, - dropout_68, - dropout_69, - dropout_7, - dropout_70, - dropout_71, - dropout_72, - dropout_73, - dropout_74, - dropout_75, - dropout_76, - dropout_77, - dropout_78, - dropout_79, - dropout_8, - dropout_80, - dropout_81, - dropout_82, - dropout_83, - dropout_84, - dropout_85, - dropout_86, - dropout_87, - dropout_88, - dropout_89, - dropout_9, - dropout_90, - dropout_91, - dropout_92, - dropout_93, - dropout_94, - dropout_95, - dropout_96, - dropout_97, - dropout_98, - dropout_99, - embedding_0, - embedding_1, - full_5, - full_6, - full_int_array_3, - full_int_array_4, - gelu_0, - gelu_1, - gelu_10, - gelu_11, - gelu_12, - gelu_13, - gelu_14, - gelu_15, - gelu_16, - gelu_17, - gelu_18, - gelu_19, - gelu_2, - gelu_20, - gelu_21, - gelu_22, - gelu_23, - gelu_3, - gelu_4, - gelu_5, - gelu_6, - gelu_7, - gelu_8, - gelu_9, - layer_norm_1, - layer_norm_10, - layer_norm_100, - layer_norm_101, - layer_norm_102, - layer_norm_103, - layer_norm_104, - layer_norm_105, - layer_norm_106, - layer_norm_107, - layer_norm_108, - layer_norm_109, - layer_norm_11, - layer_norm_110, - layer_norm_111, - layer_norm_112, - layer_norm_113, - layer_norm_114, - layer_norm_115, - layer_norm_116, - layer_norm_117, - layer_norm_118, - layer_norm_119, - layer_norm_12, - layer_norm_120, - layer_norm_121, - layer_norm_122, - layer_norm_123, - layer_norm_124, - layer_norm_125, - layer_norm_126, - layer_norm_127, - layer_norm_128, - layer_norm_129, - layer_norm_13, - layer_norm_130, - layer_norm_131, - layer_norm_132, - layer_norm_133, - layer_norm_134, - layer_norm_135, - layer_norm_136, - layer_norm_137, - layer_norm_138, - layer_norm_139, - layer_norm_14, - layer_norm_140, - layer_norm_141, - layer_norm_142, - layer_norm_143, - layer_norm_144, - layer_norm_145, - layer_norm_146, - layer_norm_15, - layer_norm_16, - layer_norm_17, - layer_norm_18, - layer_norm_19, - layer_norm_2, - layer_norm_20, - layer_norm_21, - layer_norm_22, - layer_norm_23, - layer_norm_24, - layer_norm_25, - layer_norm_26, - layer_norm_27, - layer_norm_28, - layer_norm_29, - layer_norm_3, - layer_norm_30, - layer_norm_31, - layer_norm_32, - layer_norm_33, - layer_norm_34, - layer_norm_35, - layer_norm_36, - layer_norm_37, - layer_norm_38, - layer_norm_39, - layer_norm_4, - layer_norm_40, - layer_norm_41, - layer_norm_42, - layer_norm_43, - layer_norm_44, - layer_norm_45, - layer_norm_46, - layer_norm_47, - layer_norm_48, - layer_norm_49, - layer_norm_5, - layer_norm_50, - layer_norm_51, - layer_norm_52, - layer_norm_53, - layer_norm_54, - layer_norm_55, - layer_norm_56, - layer_norm_57, - layer_norm_58, - layer_norm_59, - layer_norm_6, - layer_norm_60, - layer_norm_61, - layer_norm_62, - layer_norm_63, - layer_norm_64, - layer_norm_65, - layer_norm_66, - layer_norm_67, - layer_norm_68, - layer_norm_69, - layer_norm_7, - layer_norm_70, - layer_norm_71, - layer_norm_72, - layer_norm_73, - layer_norm_74, - layer_norm_75, - layer_norm_76, - layer_norm_77, - layer_norm_78, - layer_norm_79, - layer_norm_8, - layer_norm_80, - layer_norm_81, - layer_norm_82, - layer_norm_83, - layer_norm_84, - layer_norm_85, - layer_norm_86, - layer_norm_87, - layer_norm_88, - layer_norm_89, - layer_norm_9, - layer_norm_90, - layer_norm_91, - layer_norm_92, - layer_norm_93, - layer_norm_94, - layer_norm_95, - layer_norm_96, - layer_norm_97, - layer_norm_98, - layer_norm_99, - matmul_0, - matmul_1, - matmul_10, - matmul_101, - matmul_102, - matmul_103, - matmul_104, - matmul_105, - matmul_106, - matmul_107, - matmul_109, - matmul_11, - matmul_110, - matmul_111, - matmul_112, - matmul_113, - matmul_114, - matmul_115, - matmul_117, - matmul_118, - matmul_119, - matmul_120, - matmul_121, - matmul_122, - matmul_123, - matmul_125, - matmul_126, - matmul_127, - matmul_128, - matmul_129, - matmul_13, - matmul_130, - matmul_131, - matmul_133, - matmul_134, - matmul_135, - matmul_136, - matmul_137, - matmul_138, - matmul_139, - matmul_14, - matmul_141, - matmul_142, - matmul_143, - matmul_144, - matmul_145, - matmul_146, - matmul_147, - matmul_149, - matmul_15, - matmul_150, - matmul_151, - matmul_152, - matmul_153, - matmul_154, - matmul_155, - matmul_157, - matmul_158, - matmul_159, - matmul_16, - matmul_160, - matmul_161, - matmul_162, - matmul_163, - matmul_165, - matmul_166, - matmul_167, - matmul_168, - matmul_169, - matmul_17, - matmul_170, - matmul_171, - matmul_173, - matmul_174, - matmul_175, - matmul_176, - matmul_177, - matmul_178, - matmul_179, - matmul_18, - matmul_181, - matmul_182, - matmul_183, - matmul_184, - matmul_185, - matmul_186, - matmul_187, - matmul_189, - matmul_19, - matmul_190, - matmul_191, - matmul_192, - matmul_2, - matmul_21, - matmul_22, - matmul_23, - matmul_24, - matmul_25, - matmul_26, - matmul_27, - matmul_29, - matmul_3, - matmul_30, - matmul_31, - matmul_32, - matmul_33, - matmul_34, - matmul_35, - matmul_37, - matmul_38, - matmul_39, - matmul_40, - matmul_41, - matmul_42, - matmul_43, - matmul_45, - matmul_46, - matmul_47, - matmul_48, - matmul_49, - matmul_5, - matmul_50, - matmul_51, - matmul_53, - matmul_54, - matmul_55, - matmul_56, - matmul_57, - matmul_58, - matmul_59, - matmul_6, - matmul_61, - matmul_62, - matmul_63, - matmul_64, - matmul_65, - matmul_66, - matmul_67, - matmul_69, - matmul_7, - matmul_70, - matmul_71, - matmul_72, - matmul_73, - matmul_74, - matmul_75, - matmul_77, - matmul_78, - matmul_79, - matmul_8, - matmul_80, - matmul_81, - matmul_82, - matmul_83, - matmul_85, - matmul_86, - matmul_87, - matmul_88, - matmul_89, - matmul_9, - matmul_90, - matmul_91, - matmul_93, - matmul_94, - matmul_95, - matmul_96, - matmul_97, - matmul_98, - matmul_99, - reshape_11, - reshape_15, - reshape_19, - reshape_23, - reshape_27, - reshape_3, - reshape_31, - reshape_35, - reshape_39, - reshape_43, - reshape_47, - reshape_51, - reshape_55, - reshape_59, - reshape_63, - reshape_67, - reshape_7, - reshape_71, - reshape_75, - reshape_79, - reshape_83, - reshape_87, - reshape_91, - reshape_95, - scale_1, - scale_10, - scale_11, - scale_12, - scale_13, - scale_14, - scale_15, - scale_16, - scale_17, - scale_18, - scale_19, - scale_2, - scale_20, - scale_21, - scale_22, - scale_23, - scale_24, - scale_25, - scale_3, - scale_4, - scale_5, - scale_6, - scale_7, - scale_8, - scale_9, - slice_0, - softmax_0, - softmax_1, - softmax_10, - softmax_11, - softmax_12, - softmax_13, - softmax_14, - softmax_15, - softmax_16, - softmax_17, - softmax_18, - softmax_19, - softmax_2, - softmax_20, - softmax_21, - softmax_22, - softmax_23, - softmax_3, - softmax_4, - softmax_5, - softmax_6, - softmax_7, - softmax_8, - softmax_9, - transpose_1, - transpose_10, - transpose_11, - transpose_13, - transpose_14, - transpose_15, - transpose_17, - transpose_18, - transpose_19, - transpose_2, - transpose_21, - transpose_22, - transpose_23, - transpose_25, - transpose_26, - transpose_27, - transpose_29, - transpose_3, - transpose_30, - transpose_31, - transpose_33, - transpose_34, - transpose_35, - transpose_37, - transpose_38, - transpose_39, - transpose_41, - transpose_42, - transpose_43, - transpose_45, - transpose_46, - transpose_47, - transpose_49, - transpose_5, - transpose_50, - transpose_51, - transpose_53, - transpose_54, - transpose_55, - transpose_57, - transpose_58, - transpose_59, - transpose_6, - transpose_61, - transpose_62, - transpose_63, - transpose_65, - transpose_66, - transpose_67, - transpose_69, - transpose_7, - transpose_70, - transpose_71, - transpose_73, - transpose_74, - transpose_75, - transpose_77, - transpose_78, - transpose_79, - transpose_81, - transpose_82, - transpose_83, - transpose_85, - transpose_86, - transpose_87, - transpose_89, - transpose_9, - transpose_90, - transpose_91, - transpose_93, - transpose_94, - transpose_95, - unsqueeze_0, - ) + del add_217 return tanh_0 diff --git a/paddle_samples/PaddleNLP/ernie-m-large/weight_meta.py b/paddle_samples/PaddleNLP/ernie-m-large/weight_meta.py index ceaf57eec..5b3ea7c14 100644 --- a/paddle_samples/PaddleNLP/ernie-m-large/weight_meta.py +++ b/paddle_samples/PaddleNLP/ernie-m-large/weight_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_parameter_0: name = "parameter_0" + original_name = "linear_3891.b_0" shape = [1024] dtype = "float32" data = None @@ -7,6 +8,7 @@ class Program_weight_tensor_parameter_0: class Program_weight_tensor_parameter_1: name = "parameter_1" + original_name = "linear_3891.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.103663") @@ -18,6 +20,7 @@ class Program_weight_tensor_parameter_1: class Program_weight_tensor_parameter_2: name = "parameter_2" + original_name = "layer_norm_1343.b_0" shape = [1024] dtype = "float32" min_val = float("-0.176081") @@ -29,6 +32,7 @@ class Program_weight_tensor_parameter_2: class Program_weight_tensor_parameter_3: name = "parameter_3" + original_name = "layer_norm_1343.w_0" shape = [1024] dtype = "float32" min_val = float("0.479965") @@ -40,6 +44,7 @@ class Program_weight_tensor_parameter_3: class Program_weight_tensor_parameter_4: name = "parameter_4" + original_name = "layer_norm_1342.b_0" shape = [1024] dtype = "float32" min_val = float("-0.386177") @@ -51,6 +56,7 @@ class Program_weight_tensor_parameter_4: class Program_weight_tensor_parameter_5: name = "parameter_5" + original_name = "layer_norm_1342.w_0" shape = [1024] dtype = "float32" min_val = float("0.158595") @@ -62,6 +68,7 @@ class Program_weight_tensor_parameter_5: class Program_weight_tensor_parameter_6: name = "parameter_6" + original_name = "linear_3890.b_0" shape = [1024] dtype = "float32" min_val = float("-1.11645") @@ -73,6 +80,7 @@ class Program_weight_tensor_parameter_6: class Program_weight_tensor_parameter_7: name = "parameter_7" + original_name = "linear_3890.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-2.04274") @@ -84,6 +92,7 @@ class Program_weight_tensor_parameter_7: class Program_weight_tensor_parameter_8: name = "parameter_8" + original_name = "linear_3889.b_0" shape = [4096] dtype = "float32" min_val = float("-0.340223") @@ -95,6 +104,7 @@ class Program_weight_tensor_parameter_8: class Program_weight_tensor_parameter_9: name = "parameter_9" + original_name = "linear_3889.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.403656") @@ -106,6 +116,7 @@ class Program_weight_tensor_parameter_9: class Program_weight_tensor_parameter_10: name = "parameter_10" + original_name = "linear_3888.b_0" shape = [1024] dtype = "float32" min_val = float("-0.422438") @@ -117,6 +128,7 @@ class Program_weight_tensor_parameter_10: class Program_weight_tensor_parameter_11: name = "parameter_11" + original_name = "linear_3888.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.656597") @@ -128,6 +140,7 @@ class Program_weight_tensor_parameter_11: class Program_weight_tensor_parameter_12: name = "parameter_12" + original_name = "linear_3887.b_0" shape = [1024] dtype = "float32" min_val = float("-0.168314") @@ -139,6 +152,7 @@ class Program_weight_tensor_parameter_12: class Program_weight_tensor_parameter_13: name = "parameter_13" + original_name = "linear_3887.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.303913") @@ -150,6 +164,7 @@ class Program_weight_tensor_parameter_13: class Program_weight_tensor_parameter_14: name = "parameter_14" + original_name = "linear_3886.b_0" shape = [1024] dtype = "float32" min_val = float("-0.779785") @@ -161,6 +176,7 @@ class Program_weight_tensor_parameter_14: class Program_weight_tensor_parameter_15: name = "parameter_15" + original_name = "linear_3886.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.517846") @@ -172,6 +188,7 @@ class Program_weight_tensor_parameter_15: class Program_weight_tensor_parameter_16: name = "parameter_16" + original_name = "linear_3885.b_0" shape = [1024] dtype = "float32" min_val = float("-0.560007") @@ -183,6 +200,7 @@ class Program_weight_tensor_parameter_16: class Program_weight_tensor_parameter_17: name = "parameter_17" + original_name = "linear_3885.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.527294") @@ -194,6 +212,7 @@ class Program_weight_tensor_parameter_17: class Program_weight_tensor_parameter_18: name = "parameter_18" + original_name = "layer_norm_1341.b_0" shape = [1024] dtype = "float32" min_val = float("-1.20589") @@ -205,6 +224,7 @@ class Program_weight_tensor_parameter_18: class Program_weight_tensor_parameter_19: name = "parameter_19" + original_name = "layer_norm_1341.w_0" shape = [1024] dtype = "float32" min_val = float("0.197786") @@ -216,6 +236,7 @@ class Program_weight_tensor_parameter_19: class Program_weight_tensor_parameter_20: name = "parameter_20" + original_name = "layer_norm_1340.b_0" shape = [1024] dtype = "float32" min_val = float("-0.693641") @@ -227,6 +248,7 @@ class Program_weight_tensor_parameter_20: class Program_weight_tensor_parameter_21: name = "parameter_21" + original_name = "layer_norm_1340.w_0" shape = [1024] dtype = "float32" min_val = float("0.799307") @@ -238,6 +260,7 @@ class Program_weight_tensor_parameter_21: class Program_weight_tensor_parameter_22: name = "parameter_22" + original_name = "linear_3884.b_0" shape = [1024] dtype = "float32" min_val = float("-0.398759") @@ -249,6 +272,7 @@ class Program_weight_tensor_parameter_22: class Program_weight_tensor_parameter_23: name = "parameter_23" + original_name = "linear_3884.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-1.37008") @@ -260,6 +284,7 @@ class Program_weight_tensor_parameter_23: class Program_weight_tensor_parameter_24: name = "parameter_24" + original_name = "linear_3883.b_0" shape = [4096] dtype = "float32" min_val = float("-0.51396") @@ -271,6 +296,7 @@ class Program_weight_tensor_parameter_24: class Program_weight_tensor_parameter_25: name = "parameter_25" + original_name = "linear_3883.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.513917") @@ -282,6 +308,7 @@ class Program_weight_tensor_parameter_25: class Program_weight_tensor_parameter_26: name = "parameter_26" + original_name = "linear_3882.b_0" shape = [1024] dtype = "float32" min_val = float("-0.237958") @@ -293,6 +320,7 @@ class Program_weight_tensor_parameter_26: class Program_weight_tensor_parameter_27: name = "parameter_27" + original_name = "linear_3882.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.619973") @@ -304,6 +332,7 @@ class Program_weight_tensor_parameter_27: class Program_weight_tensor_parameter_28: name = "parameter_28" + original_name = "linear_3881.b_0" shape = [1024] dtype = "float32" min_val = float("-0.232344") @@ -315,6 +344,7 @@ class Program_weight_tensor_parameter_28: class Program_weight_tensor_parameter_29: name = "parameter_29" + original_name = "linear_3881.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.448926") @@ -326,6 +356,7 @@ class Program_weight_tensor_parameter_29: class Program_weight_tensor_parameter_30: name = "parameter_30" + original_name = "linear_3880.b_0" shape = [1024] dtype = "float32" min_val = float("-0.655762") @@ -337,6 +368,7 @@ class Program_weight_tensor_parameter_30: class Program_weight_tensor_parameter_31: name = "parameter_31" + original_name = "linear_3880.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.591113") @@ -348,6 +380,7 @@ class Program_weight_tensor_parameter_31: class Program_weight_tensor_parameter_32: name = "parameter_32" + original_name = "linear_3879.b_0" shape = [1024] dtype = "float32" min_val = float("-0.518921") @@ -359,6 +392,7 @@ class Program_weight_tensor_parameter_32: class Program_weight_tensor_parameter_33: name = "parameter_33" + original_name = "linear_3879.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.533436") @@ -370,6 +404,7 @@ class Program_weight_tensor_parameter_33: class Program_weight_tensor_parameter_34: name = "parameter_34" + original_name = "layer_norm_1339.b_0" shape = [1024] dtype = "float32" min_val = float("-1.01846") @@ -381,6 +416,7 @@ class Program_weight_tensor_parameter_34: class Program_weight_tensor_parameter_35: name = "parameter_35" + original_name = "layer_norm_1339.w_0" shape = [1024] dtype = "float32" min_val = float("0.157124") @@ -392,6 +428,7 @@ class Program_weight_tensor_parameter_35: class Program_weight_tensor_parameter_36: name = "parameter_36" + original_name = "layer_norm_1338.b_0" shape = [1024] dtype = "float32" min_val = float("-0.671069") @@ -403,6 +440,7 @@ class Program_weight_tensor_parameter_36: class Program_weight_tensor_parameter_37: name = "parameter_37" + original_name = "layer_norm_1338.w_0" shape = [1024] dtype = "float32" min_val = float("0.663683") @@ -414,6 +452,7 @@ class Program_weight_tensor_parameter_37: class Program_weight_tensor_parameter_38: name = "parameter_38" + original_name = "linear_3878.b_0" shape = [1024] dtype = "float32" min_val = float("-0.385779") @@ -425,6 +464,7 @@ class Program_weight_tensor_parameter_38: class Program_weight_tensor_parameter_39: name = "parameter_39" + original_name = "linear_3878.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-1.69639") @@ -436,6 +476,7 @@ class Program_weight_tensor_parameter_39: class Program_weight_tensor_parameter_40: name = "parameter_40" + original_name = "linear_3877.b_0" shape = [4096] dtype = "float32" min_val = float("-0.417964") @@ -447,6 +488,7 @@ class Program_weight_tensor_parameter_40: class Program_weight_tensor_parameter_41: name = "parameter_41" + original_name = "linear_3877.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.538909") @@ -458,6 +500,7 @@ class Program_weight_tensor_parameter_41: class Program_weight_tensor_parameter_42: name = "parameter_42" + original_name = "linear_3876.b_0" shape = [1024] dtype = "float32" min_val = float("-0.30224") @@ -469,6 +512,7 @@ class Program_weight_tensor_parameter_42: class Program_weight_tensor_parameter_43: name = "parameter_43" + original_name = "linear_3876.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.713637") @@ -480,6 +524,7 @@ class Program_weight_tensor_parameter_43: class Program_weight_tensor_parameter_44: name = "parameter_44" + original_name = "linear_3875.b_0" shape = [1024] dtype = "float32" min_val = float("-0.164097") @@ -491,6 +536,7 @@ class Program_weight_tensor_parameter_44: class Program_weight_tensor_parameter_45: name = "parameter_45" + original_name = "linear_3875.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.287109") @@ -502,6 +548,7 @@ class Program_weight_tensor_parameter_45: class Program_weight_tensor_parameter_46: name = "parameter_46" + original_name = "linear_3874.b_0" shape = [1024] dtype = "float32" min_val = float("-0.693359") @@ -513,6 +560,7 @@ class Program_weight_tensor_parameter_46: class Program_weight_tensor_parameter_47: name = "parameter_47" + original_name = "linear_3874.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.535281") @@ -524,6 +572,7 @@ class Program_weight_tensor_parameter_47: class Program_weight_tensor_parameter_48: name = "parameter_48" + original_name = "linear_3873.b_0" shape = [1024] dtype = "float32" min_val = float("-0.466281") @@ -535,6 +584,7 @@ class Program_weight_tensor_parameter_48: class Program_weight_tensor_parameter_49: name = "parameter_49" + original_name = "linear_3873.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.535577") @@ -546,6 +596,7 @@ class Program_weight_tensor_parameter_49: class Program_weight_tensor_parameter_50: name = "parameter_50" + original_name = "layer_norm_1337.b_0" shape = [1024] dtype = "float32" min_val = float("-0.933555") @@ -557,6 +608,7 @@ class Program_weight_tensor_parameter_50: class Program_weight_tensor_parameter_51: name = "parameter_51" + original_name = "layer_norm_1337.w_0" shape = [1024] dtype = "float32" min_val = float("0.363444") @@ -568,6 +620,7 @@ class Program_weight_tensor_parameter_51: class Program_weight_tensor_parameter_52: name = "parameter_52" + original_name = "layer_norm_1336.b_0" shape = [1024] dtype = "float32" min_val = float("-0.678259") @@ -579,6 +632,7 @@ class Program_weight_tensor_parameter_52: class Program_weight_tensor_parameter_53: name = "parameter_53" + original_name = "layer_norm_1336.w_0" shape = [1024] dtype = "float32" min_val = float("0.684749") @@ -590,6 +644,7 @@ class Program_weight_tensor_parameter_53: class Program_weight_tensor_parameter_54: name = "parameter_54" + original_name = "linear_3872.b_0" shape = [1024] dtype = "float32" min_val = float("-0.401758") @@ -601,6 +656,7 @@ class Program_weight_tensor_parameter_54: class Program_weight_tensor_parameter_55: name = "parameter_55" + original_name = "linear_3872.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-3.56209") @@ -612,6 +668,7 @@ class Program_weight_tensor_parameter_55: class Program_weight_tensor_parameter_56: name = "parameter_56" + original_name = "linear_3871.b_0" shape = [4096] dtype = "float32" min_val = float("-0.446471") @@ -623,6 +680,7 @@ class Program_weight_tensor_parameter_56: class Program_weight_tensor_parameter_57: name = "parameter_57" + original_name = "linear_3871.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.687966") @@ -634,6 +692,7 @@ class Program_weight_tensor_parameter_57: class Program_weight_tensor_parameter_58: name = "parameter_58" + original_name = "linear_3870.b_0" shape = [1024] dtype = "float32" min_val = float("-0.309769") @@ -645,6 +704,7 @@ class Program_weight_tensor_parameter_58: class Program_weight_tensor_parameter_59: name = "parameter_59" + original_name = "linear_3870.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.688203") @@ -656,6 +716,7 @@ class Program_weight_tensor_parameter_59: class Program_weight_tensor_parameter_60: name = "parameter_60" + original_name = "linear_3869.b_0" shape = [1024] dtype = "float32" min_val = float("-0.194799") @@ -667,6 +728,7 @@ class Program_weight_tensor_parameter_60: class Program_weight_tensor_parameter_61: name = "parameter_61" + original_name = "linear_3869.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.281976") @@ -678,6 +740,7 @@ class Program_weight_tensor_parameter_61: class Program_weight_tensor_parameter_62: name = "parameter_62" + original_name = "linear_3868.b_0" shape = [1024] dtype = "float32" min_val = float("-0.374034") @@ -689,6 +752,7 @@ class Program_weight_tensor_parameter_62: class Program_weight_tensor_parameter_63: name = "parameter_63" + original_name = "linear_3868.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.549631") @@ -700,6 +764,7 @@ class Program_weight_tensor_parameter_63: class Program_weight_tensor_parameter_64: name = "parameter_64" + original_name = "linear_3867.b_0" shape = [1024] dtype = "float32" min_val = float("-0.489822") @@ -711,6 +776,7 @@ class Program_weight_tensor_parameter_64: class Program_weight_tensor_parameter_65: name = "parameter_65" + original_name = "linear_3867.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.532223") @@ -722,6 +788,7 @@ class Program_weight_tensor_parameter_65: class Program_weight_tensor_parameter_66: name = "parameter_66" + original_name = "layer_norm_1335.b_0" shape = [1024] dtype = "float32" min_val = float("-2.28326") @@ -733,6 +800,7 @@ class Program_weight_tensor_parameter_66: class Program_weight_tensor_parameter_67: name = "parameter_67" + original_name = "layer_norm_1335.w_0" shape = [1024] dtype = "float32" min_val = float("0.313085") @@ -744,6 +812,7 @@ class Program_weight_tensor_parameter_67: class Program_weight_tensor_parameter_68: name = "parameter_68" + original_name = "layer_norm_1334.b_0" shape = [1024] dtype = "float32" min_val = float("-0.970182") @@ -755,6 +824,7 @@ class Program_weight_tensor_parameter_68: class Program_weight_tensor_parameter_69: name = "parameter_69" + original_name = "layer_norm_1334.w_0" shape = [1024] dtype = "float32" min_val = float("0.632709") @@ -766,6 +836,7 @@ class Program_weight_tensor_parameter_69: class Program_weight_tensor_parameter_70: name = "parameter_70" + original_name = "linear_3866.b_0" shape = [1024] dtype = "float32" min_val = float("-0.497513") @@ -777,6 +848,7 @@ class Program_weight_tensor_parameter_70: class Program_weight_tensor_parameter_71: name = "parameter_71" + original_name = "linear_3866.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-2.53108") @@ -788,6 +860,7 @@ class Program_weight_tensor_parameter_71: class Program_weight_tensor_parameter_72: name = "parameter_72" + original_name = "linear_3865.b_0" shape = [4096] dtype = "float32" min_val = float("-0.350661") @@ -799,6 +872,7 @@ class Program_weight_tensor_parameter_72: class Program_weight_tensor_parameter_73: name = "parameter_73" + original_name = "linear_3865.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.688573") @@ -810,6 +884,7 @@ class Program_weight_tensor_parameter_73: class Program_weight_tensor_parameter_74: name = "parameter_74" + original_name = "linear_3864.b_0" shape = [1024] dtype = "float32" min_val = float("-0.339246") @@ -821,6 +896,7 @@ class Program_weight_tensor_parameter_74: class Program_weight_tensor_parameter_75: name = "parameter_75" + original_name = "linear_3864.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.772574") @@ -832,6 +908,7 @@ class Program_weight_tensor_parameter_75: class Program_weight_tensor_parameter_76: name = "parameter_76" + original_name = "linear_3863.b_0" shape = [1024] dtype = "float32" min_val = float("-0.212428") @@ -843,6 +920,7 @@ class Program_weight_tensor_parameter_76: class Program_weight_tensor_parameter_77: name = "parameter_77" + original_name = "linear_3863.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.415737") @@ -854,6 +932,7 @@ class Program_weight_tensor_parameter_77: class Program_weight_tensor_parameter_78: name = "parameter_78" + original_name = "linear_3862.b_0" shape = [1024] dtype = "float32" min_val = float("-0.689941") @@ -865,6 +944,7 @@ class Program_weight_tensor_parameter_78: class Program_weight_tensor_parameter_79: name = "parameter_79" + original_name = "linear_3862.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.546115") @@ -876,6 +956,7 @@ class Program_weight_tensor_parameter_79: class Program_weight_tensor_parameter_80: name = "parameter_80" + original_name = "linear_3861.b_0" shape = [1024] dtype = "float32" min_val = float("-0.527477") @@ -887,6 +968,7 @@ class Program_weight_tensor_parameter_80: class Program_weight_tensor_parameter_81: name = "parameter_81" + original_name = "linear_3861.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.51254") @@ -898,6 +980,7 @@ class Program_weight_tensor_parameter_81: class Program_weight_tensor_parameter_82: name = "parameter_82" + original_name = "layer_norm_1333.b_0" shape = [1024] dtype = "float32" min_val = float("-1.41298") @@ -909,6 +992,7 @@ class Program_weight_tensor_parameter_82: class Program_weight_tensor_parameter_83: name = "parameter_83" + original_name = "layer_norm_1333.w_0" shape = [1024] dtype = "float32" min_val = float("0.250192") @@ -920,6 +1004,7 @@ class Program_weight_tensor_parameter_83: class Program_weight_tensor_parameter_84: name = "parameter_84" + original_name = "layer_norm_1332.b_0" shape = [1024] dtype = "float32" min_val = float("-1.27474") @@ -931,6 +1016,7 @@ class Program_weight_tensor_parameter_84: class Program_weight_tensor_parameter_85: name = "parameter_85" + original_name = "layer_norm_1332.w_0" shape = [1024] dtype = "float32" min_val = float("0.643315") @@ -942,6 +1028,7 @@ class Program_weight_tensor_parameter_85: class Program_weight_tensor_parameter_86: name = "parameter_86" + original_name = "linear_3860.b_0" shape = [1024] dtype = "float32" min_val = float("-0.599623") @@ -953,6 +1040,7 @@ class Program_weight_tensor_parameter_86: class Program_weight_tensor_parameter_87: name = "parameter_87" + original_name = "linear_3860.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-2.63931") @@ -964,6 +1052,7 @@ class Program_weight_tensor_parameter_87: class Program_weight_tensor_parameter_88: name = "parameter_88" + original_name = "linear_3859.b_0" shape = [4096] dtype = "float32" min_val = float("-0.320866") @@ -975,6 +1064,7 @@ class Program_weight_tensor_parameter_88: class Program_weight_tensor_parameter_89: name = "parameter_89" + original_name = "linear_3859.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.616851") @@ -986,6 +1076,7 @@ class Program_weight_tensor_parameter_89: class Program_weight_tensor_parameter_90: name = "parameter_90" + original_name = "linear_3858.b_0" shape = [1024] dtype = "float32" min_val = float("-0.286181") @@ -997,6 +1088,7 @@ class Program_weight_tensor_parameter_90: class Program_weight_tensor_parameter_91: name = "parameter_91" + original_name = "linear_3858.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.630289") @@ -1008,6 +1100,7 @@ class Program_weight_tensor_parameter_91: class Program_weight_tensor_parameter_92: name = "parameter_92" + original_name = "linear_3857.b_0" shape = [1024] dtype = "float32" min_val = float("-0.159188") @@ -1019,6 +1112,7 @@ class Program_weight_tensor_parameter_92: class Program_weight_tensor_parameter_93: name = "parameter_93" + original_name = "linear_3857.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.320606") @@ -1030,6 +1124,7 @@ class Program_weight_tensor_parameter_93: class Program_weight_tensor_parameter_94: name = "parameter_94" + original_name = "linear_3856.b_0" shape = [1024] dtype = "float32" min_val = float("-0.722168") @@ -1041,6 +1136,7 @@ class Program_weight_tensor_parameter_94: class Program_weight_tensor_parameter_95: name = "parameter_95" + original_name = "linear_3856.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.544046") @@ -1052,6 +1148,7 @@ class Program_weight_tensor_parameter_95: class Program_weight_tensor_parameter_96: name = "parameter_96" + original_name = "linear_3855.b_0" shape = [1024] dtype = "float32" min_val = float("-0.330498") @@ -1063,6 +1160,7 @@ class Program_weight_tensor_parameter_96: class Program_weight_tensor_parameter_97: name = "parameter_97" + original_name = "linear_3855.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.530946") @@ -1074,6 +1172,7 @@ class Program_weight_tensor_parameter_97: class Program_weight_tensor_parameter_98: name = "parameter_98" + original_name = "layer_norm_1331.b_0" shape = [1024] dtype = "float32" min_val = float("-0.614367") @@ -1085,6 +1184,7 @@ class Program_weight_tensor_parameter_98: class Program_weight_tensor_parameter_99: name = "parameter_99" + original_name = "layer_norm_1331.w_0" shape = [1024] dtype = "float32" min_val = float("0.673175") @@ -1096,6 +1196,7 @@ class Program_weight_tensor_parameter_99: class Program_weight_tensor_parameter_100: name = "parameter_100" + original_name = "layer_norm_1330.b_0" shape = [1024] dtype = "float32" min_val = float("-1.20986") @@ -1107,6 +1208,7 @@ class Program_weight_tensor_parameter_100: class Program_weight_tensor_parameter_101: name = "parameter_101" + original_name = "layer_norm_1330.w_0" shape = [1024] dtype = "float32" min_val = float("0.69822") @@ -1118,6 +1220,7 @@ class Program_weight_tensor_parameter_101: class Program_weight_tensor_parameter_102: name = "parameter_102" + original_name = "linear_3854.b_0" shape = [1024] dtype = "float32" min_val = float("-0.425355") @@ -1129,6 +1232,7 @@ class Program_weight_tensor_parameter_102: class Program_weight_tensor_parameter_103: name = "parameter_103" + original_name = "linear_3854.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-1.56473") @@ -1140,6 +1244,7 @@ class Program_weight_tensor_parameter_103: class Program_weight_tensor_parameter_104: name = "parameter_104" + original_name = "linear_3853.b_0" shape = [4096] dtype = "float32" min_val = float("-0.266531") @@ -1151,6 +1256,7 @@ class Program_weight_tensor_parameter_104: class Program_weight_tensor_parameter_105: name = "parameter_105" + original_name = "linear_3853.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.613365") @@ -1162,6 +1268,7 @@ class Program_weight_tensor_parameter_105: class Program_weight_tensor_parameter_106: name = "parameter_106" + original_name = "linear_3852.b_0" shape = [1024] dtype = "float32" min_val = float("-0.289766") @@ -1173,6 +1280,7 @@ class Program_weight_tensor_parameter_106: class Program_weight_tensor_parameter_107: name = "parameter_107" + original_name = "linear_3852.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.579557") @@ -1184,6 +1292,7 @@ class Program_weight_tensor_parameter_107: class Program_weight_tensor_parameter_108: name = "parameter_108" + original_name = "linear_3851.b_0" shape = [1024] dtype = "float32" min_val = float("-0.128746") @@ -1195,6 +1304,7 @@ class Program_weight_tensor_parameter_108: class Program_weight_tensor_parameter_109: name = "parameter_109" + original_name = "linear_3851.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.466527") @@ -1206,6 +1316,7 @@ class Program_weight_tensor_parameter_109: class Program_weight_tensor_parameter_110: name = "parameter_110" + original_name = "linear_3850.b_0" shape = [1024] dtype = "float32" min_val = float("-0.434815") @@ -1217,6 +1328,7 @@ class Program_weight_tensor_parameter_110: class Program_weight_tensor_parameter_111: name = "parameter_111" + original_name = "linear_3850.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.527818") @@ -1228,6 +1340,7 @@ class Program_weight_tensor_parameter_111: class Program_weight_tensor_parameter_112: name = "parameter_112" + original_name = "linear_3849.b_0" shape = [1024] dtype = "float32" min_val = float("-0.359493") @@ -1239,6 +1352,7 @@ class Program_weight_tensor_parameter_112: class Program_weight_tensor_parameter_113: name = "parameter_113" + original_name = "linear_3849.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.52973") @@ -1250,6 +1364,7 @@ class Program_weight_tensor_parameter_113: class Program_weight_tensor_parameter_114: name = "parameter_114" + original_name = "layer_norm_1329.b_0" shape = [1024] dtype = "float32" min_val = float("-0.578878") @@ -1261,6 +1376,7 @@ class Program_weight_tensor_parameter_114: class Program_weight_tensor_parameter_115: name = "parameter_115" + original_name = "layer_norm_1329.w_0" shape = [1024] dtype = "float32" min_val = float("0.476504") @@ -1272,6 +1388,7 @@ class Program_weight_tensor_parameter_115: class Program_weight_tensor_parameter_116: name = "parameter_116" + original_name = "layer_norm_1328.b_0" shape = [1024] dtype = "float32" min_val = float("-1.25181") @@ -1283,6 +1400,7 @@ class Program_weight_tensor_parameter_116: class Program_weight_tensor_parameter_117: name = "parameter_117" + original_name = "layer_norm_1328.w_0" shape = [1024] dtype = "float32" min_val = float("0.686701") @@ -1294,6 +1412,7 @@ class Program_weight_tensor_parameter_117: class Program_weight_tensor_parameter_118: name = "parameter_118" + original_name = "linear_3848.b_0" shape = [1024] dtype = "float32" min_val = float("-0.315595") @@ -1305,6 +1424,7 @@ class Program_weight_tensor_parameter_118: class Program_weight_tensor_parameter_119: name = "parameter_119" + original_name = "linear_3848.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-1.36953") @@ -1316,6 +1436,7 @@ class Program_weight_tensor_parameter_119: class Program_weight_tensor_parameter_120: name = "parameter_120" + original_name = "linear_3847.b_0" shape = [4096] dtype = "float32" min_val = float("-0.386791") @@ -1327,6 +1448,7 @@ class Program_weight_tensor_parameter_120: class Program_weight_tensor_parameter_121: name = "parameter_121" + original_name = "linear_3847.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.635358") @@ -1338,6 +1460,7 @@ class Program_weight_tensor_parameter_121: class Program_weight_tensor_parameter_122: name = "parameter_122" + original_name = "linear_3846.b_0" shape = [1024] dtype = "float32" min_val = float("-0.209918") @@ -1349,6 +1472,7 @@ class Program_weight_tensor_parameter_122: class Program_weight_tensor_parameter_123: name = "parameter_123" + original_name = "linear_3846.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.639031") @@ -1360,6 +1484,7 @@ class Program_weight_tensor_parameter_123: class Program_weight_tensor_parameter_124: name = "parameter_124" + original_name = "linear_3845.b_0" shape = [1024] dtype = "float32" min_val = float("-0.126776") @@ -1371,6 +1496,7 @@ class Program_weight_tensor_parameter_124: class Program_weight_tensor_parameter_125: name = "parameter_125" + original_name = "linear_3845.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.32147") @@ -1382,6 +1508,7 @@ class Program_weight_tensor_parameter_125: class Program_weight_tensor_parameter_126: name = "parameter_126" + original_name = "linear_3844.b_0" shape = [1024] dtype = "float32" min_val = float("-0.468262") @@ -1393,6 +1520,7 @@ class Program_weight_tensor_parameter_126: class Program_weight_tensor_parameter_127: name = "parameter_127" + original_name = "linear_3844.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.585917") @@ -1404,6 +1532,7 @@ class Program_weight_tensor_parameter_127: class Program_weight_tensor_parameter_128: name = "parameter_128" + original_name = "linear_3843.b_0" shape = [1024] dtype = "float32" min_val = float("-0.342448") @@ -1415,6 +1544,7 @@ class Program_weight_tensor_parameter_128: class Program_weight_tensor_parameter_129: name = "parameter_129" + original_name = "linear_3843.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.580423") @@ -1426,6 +1556,7 @@ class Program_weight_tensor_parameter_129: class Program_weight_tensor_parameter_130: name = "parameter_130" + original_name = "layer_norm_1327.b_0" shape = [1024] dtype = "float32" min_val = float("-0.455857") @@ -1437,6 +1568,7 @@ class Program_weight_tensor_parameter_130: class Program_weight_tensor_parameter_131: name = "parameter_131" + original_name = "layer_norm_1327.w_0" shape = [1024] dtype = "float32" min_val = float("0.293481") @@ -1448,6 +1580,7 @@ class Program_weight_tensor_parameter_131: class Program_weight_tensor_parameter_132: name = "parameter_132" + original_name = "layer_norm_1326.b_0" shape = [1024] dtype = "float32" min_val = float("-1.05344") @@ -1459,6 +1592,7 @@ class Program_weight_tensor_parameter_132: class Program_weight_tensor_parameter_133: name = "parameter_133" + original_name = "layer_norm_1326.w_0" shape = [1024] dtype = "float32" min_val = float("0.715948") @@ -1470,6 +1604,7 @@ class Program_weight_tensor_parameter_133: class Program_weight_tensor_parameter_134: name = "parameter_134" + original_name = "linear_3842.b_0" shape = [1024] dtype = "float32" min_val = float("-0.465619") @@ -1481,6 +1616,7 @@ class Program_weight_tensor_parameter_134: class Program_weight_tensor_parameter_135: name = "parameter_135" + original_name = "linear_3842.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-1.43547") @@ -1492,6 +1628,7 @@ class Program_weight_tensor_parameter_135: class Program_weight_tensor_parameter_136: name = "parameter_136" + original_name = "linear_3841.b_0" shape = [4096] dtype = "float32" min_val = float("-0.439239") @@ -1503,6 +1640,7 @@ class Program_weight_tensor_parameter_136: class Program_weight_tensor_parameter_137: name = "parameter_137" + original_name = "linear_3841.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.631935") @@ -1514,6 +1652,7 @@ class Program_weight_tensor_parameter_137: class Program_weight_tensor_parameter_138: name = "parameter_138" + original_name = "linear_3840.b_0" shape = [1024] dtype = "float32" min_val = float("-0.345192") @@ -1525,6 +1664,7 @@ class Program_weight_tensor_parameter_138: class Program_weight_tensor_parameter_139: name = "parameter_139" + original_name = "linear_3840.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.537214") @@ -1536,6 +1676,7 @@ class Program_weight_tensor_parameter_139: class Program_weight_tensor_parameter_140: name = "parameter_140" + original_name = "linear_3839.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0938327") @@ -1547,6 +1688,7 @@ class Program_weight_tensor_parameter_140: class Program_weight_tensor_parameter_141: name = "parameter_141" + original_name = "linear_3839.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.31904") @@ -1558,6 +1700,7 @@ class Program_weight_tensor_parameter_141: class Program_weight_tensor_parameter_142: name = "parameter_142" + original_name = "linear_3838.b_0" shape = [1024] dtype = "float32" min_val = float("-0.349853") @@ -1569,6 +1712,7 @@ class Program_weight_tensor_parameter_142: class Program_weight_tensor_parameter_143: name = "parameter_143" + original_name = "linear_3838.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.550114") @@ -1580,6 +1724,7 @@ class Program_weight_tensor_parameter_143: class Program_weight_tensor_parameter_144: name = "parameter_144" + original_name = "linear_3837.b_0" shape = [1024] dtype = "float32" min_val = float("-0.474724") @@ -1591,6 +1736,7 @@ class Program_weight_tensor_parameter_144: class Program_weight_tensor_parameter_145: name = "parameter_145" + original_name = "linear_3837.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.539039") @@ -1602,6 +1748,7 @@ class Program_weight_tensor_parameter_145: class Program_weight_tensor_parameter_146: name = "parameter_146" + original_name = "layer_norm_1325.b_0" shape = [1024] dtype = "float32" min_val = float("-0.331598") @@ -1613,6 +1760,7 @@ class Program_weight_tensor_parameter_146: class Program_weight_tensor_parameter_147: name = "parameter_147" + original_name = "layer_norm_1325.w_0" shape = [1024] dtype = "float32" min_val = float("0.511606") @@ -1624,6 +1772,7 @@ class Program_weight_tensor_parameter_147: class Program_weight_tensor_parameter_148: name = "parameter_148" + original_name = "layer_norm_1324.b_0" shape = [1024] dtype = "float32" min_val = float("-0.378427") @@ -1635,6 +1784,7 @@ class Program_weight_tensor_parameter_148: class Program_weight_tensor_parameter_149: name = "parameter_149" + original_name = "layer_norm_1324.w_0" shape = [1024] dtype = "float32" min_val = float("0.701753") @@ -1646,6 +1796,7 @@ class Program_weight_tensor_parameter_149: class Program_weight_tensor_parameter_150: name = "parameter_150" + original_name = "linear_3836.b_0" shape = [1024] dtype = "float32" min_val = float("-0.356785") @@ -1657,6 +1808,7 @@ class Program_weight_tensor_parameter_150: class Program_weight_tensor_parameter_151: name = "parameter_151" + original_name = "linear_3836.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-1.23928") @@ -1668,6 +1820,7 @@ class Program_weight_tensor_parameter_151: class Program_weight_tensor_parameter_152: name = "parameter_152" + original_name = "linear_3835.b_0" shape = [4096] dtype = "float32" min_val = float("-0.357434") @@ -1679,6 +1832,7 @@ class Program_weight_tensor_parameter_152: class Program_weight_tensor_parameter_153: name = "parameter_153" + original_name = "linear_3835.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.533232") @@ -1690,6 +1844,7 @@ class Program_weight_tensor_parameter_153: class Program_weight_tensor_parameter_154: name = "parameter_154" + original_name = "linear_3834.b_0" shape = [1024] dtype = "float32" min_val = float("-0.148012") @@ -1701,6 +1856,7 @@ class Program_weight_tensor_parameter_154: class Program_weight_tensor_parameter_155: name = "parameter_155" + original_name = "linear_3834.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.597103") @@ -1712,6 +1868,7 @@ class Program_weight_tensor_parameter_155: class Program_weight_tensor_parameter_156: name = "parameter_156" + original_name = "linear_3833.b_0" shape = [1024] dtype = "float32" min_val = float("-0.206594") @@ -1723,6 +1880,7 @@ class Program_weight_tensor_parameter_156: class Program_weight_tensor_parameter_157: name = "parameter_157" + original_name = "linear_3833.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.293887") @@ -1734,6 +1892,7 @@ class Program_weight_tensor_parameter_157: class Program_weight_tensor_parameter_158: name = "parameter_158" + original_name = "linear_3832.b_0" shape = [1024] dtype = "float32" min_val = float("-0.35498") @@ -1745,6 +1904,7 @@ class Program_weight_tensor_parameter_158: class Program_weight_tensor_parameter_159: name = "parameter_159" + original_name = "linear_3832.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.525322") @@ -1756,6 +1916,7 @@ class Program_weight_tensor_parameter_159: class Program_weight_tensor_parameter_160: name = "parameter_160" + original_name = "linear_3831.b_0" shape = [1024] dtype = "float32" min_val = float("-0.410748") @@ -1767,6 +1928,7 @@ class Program_weight_tensor_parameter_160: class Program_weight_tensor_parameter_161: name = "parameter_161" + original_name = "linear_3831.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.517217") @@ -1778,6 +1940,7 @@ class Program_weight_tensor_parameter_161: class Program_weight_tensor_parameter_162: name = "parameter_162" + original_name = "layer_norm_1323.b_0" shape = [1024] dtype = "float32" min_val = float("-0.368654") @@ -1789,6 +1952,7 @@ class Program_weight_tensor_parameter_162: class Program_weight_tensor_parameter_163: name = "parameter_163" + original_name = "layer_norm_1323.w_0" shape = [1024] dtype = "float32" min_val = float("-0.0190548") @@ -1800,6 +1964,7 @@ class Program_weight_tensor_parameter_163: class Program_weight_tensor_parameter_164: name = "parameter_164" + original_name = "layer_norm_1322.b_0" shape = [1024] dtype = "float32" min_val = float("-0.378983") @@ -1811,6 +1976,7 @@ class Program_weight_tensor_parameter_164: class Program_weight_tensor_parameter_165: name = "parameter_165" + original_name = "layer_norm_1322.w_0" shape = [1024] dtype = "float32" min_val = float("0.715755") @@ -1822,6 +1988,7 @@ class Program_weight_tensor_parameter_165: class Program_weight_tensor_parameter_166: name = "parameter_166" + original_name = "linear_3830.b_0" shape = [1024] dtype = "float32" min_val = float("-0.453021") @@ -1833,6 +2000,7 @@ class Program_weight_tensor_parameter_166: class Program_weight_tensor_parameter_167: name = "parameter_167" + original_name = "linear_3830.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-1.607") @@ -1844,6 +2012,7 @@ class Program_weight_tensor_parameter_167: class Program_weight_tensor_parameter_168: name = "parameter_168" + original_name = "linear_3829.b_0" shape = [4096] dtype = "float32" min_val = float("-0.359887") @@ -1855,6 +2024,7 @@ class Program_weight_tensor_parameter_168: class Program_weight_tensor_parameter_169: name = "parameter_169" + original_name = "linear_3829.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.558074") @@ -1866,6 +2036,7 @@ class Program_weight_tensor_parameter_169: class Program_weight_tensor_parameter_170: name = "parameter_170" + original_name = "linear_3828.b_0" shape = [1024] dtype = "float32" min_val = float("-0.345319") @@ -1877,6 +2048,7 @@ class Program_weight_tensor_parameter_170: class Program_weight_tensor_parameter_171: name = "parameter_171" + original_name = "linear_3828.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.516601") @@ -1888,6 +2060,7 @@ class Program_weight_tensor_parameter_171: class Program_weight_tensor_parameter_172: name = "parameter_172" + original_name = "linear_3827.b_0" shape = [1024] dtype = "float32" min_val = float("-0.119613") @@ -1899,6 +2072,7 @@ class Program_weight_tensor_parameter_172: class Program_weight_tensor_parameter_173: name = "parameter_173" + original_name = "linear_3827.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.355801") @@ -1910,6 +2084,7 @@ class Program_weight_tensor_parameter_173: class Program_weight_tensor_parameter_174: name = "parameter_174" + original_name = "linear_3826.b_0" shape = [1024] dtype = "float32" min_val = float("-0.22473") @@ -1921,6 +2096,7 @@ class Program_weight_tensor_parameter_174: class Program_weight_tensor_parameter_175: name = "parameter_175" + original_name = "linear_3826.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.5516") @@ -1932,6 +2108,7 @@ class Program_weight_tensor_parameter_175: class Program_weight_tensor_parameter_176: name = "parameter_176" + original_name = "linear_3825.b_0" shape = [1024] dtype = "float32" min_val = float("-0.368859") @@ -1943,6 +2120,7 @@ class Program_weight_tensor_parameter_176: class Program_weight_tensor_parameter_177: name = "parameter_177" + original_name = "linear_3825.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.526374") @@ -1954,6 +2132,7 @@ class Program_weight_tensor_parameter_177: class Program_weight_tensor_parameter_178: name = "parameter_178" + original_name = "layer_norm_1321.b_0" shape = [1024] dtype = "float32" min_val = float("-0.38698") @@ -1965,6 +2144,7 @@ class Program_weight_tensor_parameter_178: class Program_weight_tensor_parameter_179: name = "parameter_179" + original_name = "layer_norm_1321.w_0" shape = [1024] dtype = "float32" min_val = float("0.000652055") @@ -1976,6 +2156,7 @@ class Program_weight_tensor_parameter_179: class Program_weight_tensor_parameter_180: name = "parameter_180" + original_name = "layer_norm_1320.b_0" shape = [1024] dtype = "float32" min_val = float("-0.410956") @@ -1987,6 +2168,7 @@ class Program_weight_tensor_parameter_180: class Program_weight_tensor_parameter_181: name = "parameter_181" + original_name = "layer_norm_1320.w_0" shape = [1024] dtype = "float32" min_val = float("0.679958") @@ -1998,6 +2180,7 @@ class Program_weight_tensor_parameter_181: class Program_weight_tensor_parameter_182: name = "parameter_182" + original_name = "linear_3824.b_0" shape = [1024] dtype = "float32" min_val = float("-0.56971") @@ -2009,6 +2192,7 @@ class Program_weight_tensor_parameter_182: class Program_weight_tensor_parameter_183: name = "parameter_183" + original_name = "linear_3824.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-1.64888") @@ -2020,6 +2204,7 @@ class Program_weight_tensor_parameter_183: class Program_weight_tensor_parameter_184: name = "parameter_184" + original_name = "linear_3823.b_0" shape = [4096] dtype = "float32" min_val = float("-0.368906") @@ -2031,6 +2216,7 @@ class Program_weight_tensor_parameter_184: class Program_weight_tensor_parameter_185: name = "parameter_185" + original_name = "linear_3823.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.547558") @@ -2042,6 +2228,7 @@ class Program_weight_tensor_parameter_185: class Program_weight_tensor_parameter_186: name = "parameter_186" + original_name = "linear_3822.b_0" shape = [1024] dtype = "float32" min_val = float("-0.361089") @@ -2053,6 +2240,7 @@ class Program_weight_tensor_parameter_186: class Program_weight_tensor_parameter_187: name = "parameter_187" + original_name = "linear_3822.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.501863") @@ -2064,6 +2252,7 @@ class Program_weight_tensor_parameter_187: class Program_weight_tensor_parameter_188: name = "parameter_188" + original_name = "linear_3821.b_0" shape = [1024] dtype = "float32" min_val = float("-0.146981") @@ -2075,6 +2264,7 @@ class Program_weight_tensor_parameter_188: class Program_weight_tensor_parameter_189: name = "parameter_189" + original_name = "linear_3821.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.371576") @@ -2086,6 +2276,7 @@ class Program_weight_tensor_parameter_189: class Program_weight_tensor_parameter_190: name = "parameter_190" + original_name = "linear_3820.b_0" shape = [1024] dtype = "float32" min_val = float("-0.319581") @@ -2097,6 +2288,7 @@ class Program_weight_tensor_parameter_190: class Program_weight_tensor_parameter_191: name = "parameter_191" + original_name = "linear_3820.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.568245") @@ -2108,6 +2300,7 @@ class Program_weight_tensor_parameter_191: class Program_weight_tensor_parameter_192: name = "parameter_192" + original_name = "linear_3819.b_0" shape = [1024] dtype = "float32" min_val = float("-0.443798") @@ -2119,6 +2312,7 @@ class Program_weight_tensor_parameter_192: class Program_weight_tensor_parameter_193: name = "parameter_193" + original_name = "linear_3819.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.550512") @@ -2130,6 +2324,7 @@ class Program_weight_tensor_parameter_193: class Program_weight_tensor_parameter_194: name = "parameter_194" + original_name = "layer_norm_1319.b_0" shape = [1024] dtype = "float32" min_val = float("-0.259008") @@ -2141,6 +2336,7 @@ class Program_weight_tensor_parameter_194: class Program_weight_tensor_parameter_195: name = "parameter_195" + original_name = "layer_norm_1319.w_0" shape = [1024] dtype = "float32" min_val = float("0.593632") @@ -2152,6 +2348,7 @@ class Program_weight_tensor_parameter_195: class Program_weight_tensor_parameter_196: name = "parameter_196" + original_name = "layer_norm_1318.b_0" shape = [1024] dtype = "float32" min_val = float("-0.34147") @@ -2163,6 +2360,7 @@ class Program_weight_tensor_parameter_196: class Program_weight_tensor_parameter_197: name = "parameter_197" + original_name = "layer_norm_1318.w_0" shape = [1024] dtype = "float32" min_val = float("0.642908") @@ -2174,6 +2372,7 @@ class Program_weight_tensor_parameter_197: class Program_weight_tensor_parameter_198: name = "parameter_198" + original_name = "linear_3818.b_0" shape = [1024] dtype = "float32" min_val = float("-0.248538") @@ -2185,6 +2384,7 @@ class Program_weight_tensor_parameter_198: class Program_weight_tensor_parameter_199: name = "parameter_199" + original_name = "linear_3818.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-1.41499") @@ -2196,6 +2396,7 @@ class Program_weight_tensor_parameter_199: class Program_weight_tensor_parameter_200: name = "parameter_200" + original_name = "linear_3817.b_0" shape = [4096] dtype = "float32" min_val = float("-0.508465") @@ -2207,6 +2408,7 @@ class Program_weight_tensor_parameter_200: class Program_weight_tensor_parameter_201: name = "parameter_201" + original_name = "linear_3817.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.60016") @@ -2218,6 +2420,7 @@ class Program_weight_tensor_parameter_201: class Program_weight_tensor_parameter_202: name = "parameter_202" + original_name = "linear_3816.b_0" shape = [1024] dtype = "float32" min_val = float("-0.266421") @@ -2229,6 +2432,7 @@ class Program_weight_tensor_parameter_202: class Program_weight_tensor_parameter_203: name = "parameter_203" + original_name = "linear_3816.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.457188") @@ -2240,6 +2444,7 @@ class Program_weight_tensor_parameter_203: class Program_weight_tensor_parameter_204: name = "parameter_204" + original_name = "linear_3815.b_0" shape = [1024] dtype = "float32" min_val = float("-0.187916") @@ -2251,6 +2456,7 @@ class Program_weight_tensor_parameter_204: class Program_weight_tensor_parameter_205: name = "parameter_205" + original_name = "linear_3815.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.346934") @@ -2262,6 +2468,7 @@ class Program_weight_tensor_parameter_205: class Program_weight_tensor_parameter_206: name = "parameter_206" + original_name = "linear_3814.b_0" shape = [1024] dtype = "float32" min_val = float("-0.336669") @@ -2273,6 +2480,7 @@ class Program_weight_tensor_parameter_206: class Program_weight_tensor_parameter_207: name = "parameter_207" + original_name = "linear_3814.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.551304") @@ -2284,6 +2492,7 @@ class Program_weight_tensor_parameter_207: class Program_weight_tensor_parameter_208: name = "parameter_208" + original_name = "linear_3813.b_0" shape = [1024] dtype = "float32" min_val = float("-0.435782") @@ -2295,6 +2504,7 @@ class Program_weight_tensor_parameter_208: class Program_weight_tensor_parameter_209: name = "parameter_209" + original_name = "linear_3813.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.545594") @@ -2306,6 +2516,7 @@ class Program_weight_tensor_parameter_209: class Program_weight_tensor_parameter_210: name = "parameter_210" + original_name = "layer_norm_1317.b_0" shape = [1024] dtype = "float32" min_val = float("-0.253758") @@ -2317,6 +2528,7 @@ class Program_weight_tensor_parameter_210: class Program_weight_tensor_parameter_211: name = "parameter_211" + original_name = "layer_norm_1317.w_0" shape = [1024] dtype = "float32" min_val = float("0.641647") @@ -2328,6 +2540,7 @@ class Program_weight_tensor_parameter_211: class Program_weight_tensor_parameter_212: name = "parameter_212" + original_name = "layer_norm_1316.b_0" shape = [1024] dtype = "float32" min_val = float("-0.372017") @@ -2339,6 +2552,7 @@ class Program_weight_tensor_parameter_212: class Program_weight_tensor_parameter_213: name = "parameter_213" + original_name = "layer_norm_1316.w_0" shape = [1024] dtype = "float32" min_val = float("0.661455") @@ -2350,6 +2564,7 @@ class Program_weight_tensor_parameter_213: class Program_weight_tensor_parameter_214: name = "parameter_214" + original_name = "linear_3812.b_0" shape = [1024] dtype = "float32" min_val = float("-0.204862") @@ -2361,6 +2576,7 @@ class Program_weight_tensor_parameter_214: class Program_weight_tensor_parameter_215: name = "parameter_215" + original_name = "linear_3812.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-1.91112") @@ -2372,6 +2588,7 @@ class Program_weight_tensor_parameter_215: class Program_weight_tensor_parameter_216: name = "parameter_216" + original_name = "linear_3811.b_0" shape = [4096] dtype = "float32" min_val = float("-0.298144") @@ -2383,6 +2600,7 @@ class Program_weight_tensor_parameter_216: class Program_weight_tensor_parameter_217: name = "parameter_217" + original_name = "linear_3811.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.561885") @@ -2394,6 +2612,7 @@ class Program_weight_tensor_parameter_217: class Program_weight_tensor_parameter_218: name = "parameter_218" + original_name = "linear_3810.b_0" shape = [1024] dtype = "float32" min_val = float("-0.234198") @@ -2405,6 +2624,7 @@ class Program_weight_tensor_parameter_218: class Program_weight_tensor_parameter_219: name = "parameter_219" + original_name = "linear_3810.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.491426") @@ -2416,6 +2636,7 @@ class Program_weight_tensor_parameter_219: class Program_weight_tensor_parameter_220: name = "parameter_220" + original_name = "linear_3809.b_0" shape = [1024] dtype = "float32" min_val = float("-0.136688") @@ -2427,6 +2648,7 @@ class Program_weight_tensor_parameter_220: class Program_weight_tensor_parameter_221: name = "parameter_221" + original_name = "linear_3809.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.295921") @@ -2438,6 +2660,7 @@ class Program_weight_tensor_parameter_221: class Program_weight_tensor_parameter_222: name = "parameter_222" + original_name = "linear_3808.b_0" shape = [1024] dtype = "float32" min_val = float("-0.357164") @@ -2449,6 +2672,7 @@ class Program_weight_tensor_parameter_222: class Program_weight_tensor_parameter_223: name = "parameter_223" + original_name = "linear_3808.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.529938") @@ -2460,6 +2684,7 @@ class Program_weight_tensor_parameter_223: class Program_weight_tensor_parameter_224: name = "parameter_224" + original_name = "linear_3807.b_0" shape = [1024] dtype = "float32" min_val = float("-0.438997") @@ -2471,6 +2696,7 @@ class Program_weight_tensor_parameter_224: class Program_weight_tensor_parameter_225: name = "parameter_225" + original_name = "linear_3807.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.455566") @@ -2482,6 +2708,7 @@ class Program_weight_tensor_parameter_225: class Program_weight_tensor_parameter_226: name = "parameter_226" + original_name = "layer_norm_1315.b_0" shape = [1024] dtype = "float32" min_val = float("-0.21558") @@ -2493,6 +2720,7 @@ class Program_weight_tensor_parameter_226: class Program_weight_tensor_parameter_227: name = "parameter_227" + original_name = "layer_norm_1315.w_0" shape = [1024] dtype = "float32" min_val = float("0.688545") @@ -2504,6 +2732,7 @@ class Program_weight_tensor_parameter_227: class Program_weight_tensor_parameter_228: name = "parameter_228" + original_name = "layer_norm_1314.b_0" shape = [1024] dtype = "float32" min_val = float("-0.415702") @@ -2515,6 +2744,7 @@ class Program_weight_tensor_parameter_228: class Program_weight_tensor_parameter_229: name = "parameter_229" + original_name = "layer_norm_1314.w_0" shape = [1024] dtype = "float32" min_val = float("0.676931") @@ -2526,6 +2756,7 @@ class Program_weight_tensor_parameter_229: class Program_weight_tensor_parameter_230: name = "parameter_230" + original_name = "linear_3806.b_0" shape = [1024] dtype = "float32" min_val = float("-0.252108") @@ -2537,6 +2768,7 @@ class Program_weight_tensor_parameter_230: class Program_weight_tensor_parameter_231: name = "parameter_231" + original_name = "linear_3806.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-0.694639") @@ -2548,6 +2780,7 @@ class Program_weight_tensor_parameter_231: class Program_weight_tensor_parameter_232: name = "parameter_232" + original_name = "linear_3805.b_0" shape = [4096] dtype = "float32" min_val = float("-0.397766") @@ -2559,6 +2792,7 @@ class Program_weight_tensor_parameter_232: class Program_weight_tensor_parameter_233: name = "parameter_233" + original_name = "linear_3805.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.548778") @@ -2570,6 +2804,7 @@ class Program_weight_tensor_parameter_233: class Program_weight_tensor_parameter_234: name = "parameter_234" + original_name = "linear_3804.b_0" shape = [1024] dtype = "float32" min_val = float("-0.244012") @@ -2581,6 +2816,7 @@ class Program_weight_tensor_parameter_234: class Program_weight_tensor_parameter_235: name = "parameter_235" + original_name = "linear_3804.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.489453") @@ -2592,6 +2828,7 @@ class Program_weight_tensor_parameter_235: class Program_weight_tensor_parameter_236: name = "parameter_236" + original_name = "linear_3803.b_0" shape = [1024] dtype = "float32" min_val = float("-0.107183") @@ -2603,6 +2840,7 @@ class Program_weight_tensor_parameter_236: class Program_weight_tensor_parameter_237: name = "parameter_237" + original_name = "linear_3803.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.329806") @@ -2614,6 +2852,7 @@ class Program_weight_tensor_parameter_237: class Program_weight_tensor_parameter_238: name = "parameter_238" + original_name = "linear_3802.b_0" shape = [1024] dtype = "float32" min_val = float("-0.168334") @@ -2625,6 +2864,7 @@ class Program_weight_tensor_parameter_238: class Program_weight_tensor_parameter_239: name = "parameter_239" + original_name = "linear_3802.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.536813") @@ -2636,6 +2876,7 @@ class Program_weight_tensor_parameter_239: class Program_weight_tensor_parameter_240: name = "parameter_240" + original_name = "linear_3801.b_0" shape = [1024] dtype = "float32" min_val = float("-0.484044") @@ -2647,6 +2888,7 @@ class Program_weight_tensor_parameter_240: class Program_weight_tensor_parameter_241: name = "parameter_241" + original_name = "linear_3801.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.524716") @@ -2658,6 +2900,7 @@ class Program_weight_tensor_parameter_241: class Program_weight_tensor_parameter_242: name = "parameter_242" + original_name = "layer_norm_1313.b_0" shape = [1024] dtype = "float32" min_val = float("-0.320334") @@ -2669,6 +2912,7 @@ class Program_weight_tensor_parameter_242: class Program_weight_tensor_parameter_243: name = "parameter_243" + original_name = "layer_norm_1313.w_0" shape = [1024] dtype = "float32" min_val = float("0.67117") @@ -2680,6 +2924,7 @@ class Program_weight_tensor_parameter_243: class Program_weight_tensor_parameter_244: name = "parameter_244" + original_name = "layer_norm_1312.b_0" shape = [1024] dtype = "float32" min_val = float("-0.421445") @@ -2691,6 +2936,7 @@ class Program_weight_tensor_parameter_244: class Program_weight_tensor_parameter_245: name = "parameter_245" + original_name = "layer_norm_1312.w_0" shape = [1024] dtype = "float32" min_val = float("0.651908") @@ -2702,6 +2948,7 @@ class Program_weight_tensor_parameter_245: class Program_weight_tensor_parameter_246: name = "parameter_246" + original_name = "linear_3800.b_0" shape = [1024] dtype = "float32" min_val = float("-0.260013") @@ -2713,6 +2960,7 @@ class Program_weight_tensor_parameter_246: class Program_weight_tensor_parameter_247: name = "parameter_247" + original_name = "linear_3800.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-0.601779") @@ -2724,6 +2972,7 @@ class Program_weight_tensor_parameter_247: class Program_weight_tensor_parameter_248: name = "parameter_248" + original_name = "linear_3799.b_0" shape = [4096] dtype = "float32" min_val = float("-0.514472") @@ -2735,6 +2984,7 @@ class Program_weight_tensor_parameter_248: class Program_weight_tensor_parameter_249: name = "parameter_249" + original_name = "linear_3799.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.804441") @@ -2746,6 +2996,7 @@ class Program_weight_tensor_parameter_249: class Program_weight_tensor_parameter_250: name = "parameter_250" + original_name = "linear_3798.b_0" shape = [1024] dtype = "float32" min_val = float("-0.30967") @@ -2757,6 +3008,7 @@ class Program_weight_tensor_parameter_250: class Program_weight_tensor_parameter_251: name = "parameter_251" + original_name = "linear_3798.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.528956") @@ -2768,6 +3020,7 @@ class Program_weight_tensor_parameter_251: class Program_weight_tensor_parameter_252: name = "parameter_252" + original_name = "linear_3797.b_0" shape = [1024] dtype = "float32" min_val = float("-0.198539") @@ -2779,6 +3032,7 @@ class Program_weight_tensor_parameter_252: class Program_weight_tensor_parameter_253: name = "parameter_253" + original_name = "linear_3797.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.49272") @@ -2790,6 +3044,7 @@ class Program_weight_tensor_parameter_253: class Program_weight_tensor_parameter_254: name = "parameter_254" + original_name = "linear_3796.b_0" shape = [1024] dtype = "float32" min_val = float("-0.160278") @@ -2801,6 +3056,7 @@ class Program_weight_tensor_parameter_254: class Program_weight_tensor_parameter_255: name = "parameter_255" + original_name = "linear_3796.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.556439") @@ -2812,6 +3068,7 @@ class Program_weight_tensor_parameter_255: class Program_weight_tensor_parameter_256: name = "parameter_256" + original_name = "linear_3795.b_0" shape = [1024] dtype = "float32" min_val = float("-0.395766") @@ -2823,6 +3080,7 @@ class Program_weight_tensor_parameter_256: class Program_weight_tensor_parameter_257: name = "parameter_257" + original_name = "linear_3795.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.516337") @@ -2834,6 +3092,7 @@ class Program_weight_tensor_parameter_257: class Program_weight_tensor_parameter_258: name = "parameter_258" + original_name = "layer_norm_1311.b_0" shape = [1024] dtype = "float32" min_val = float("-0.274522") @@ -2845,6 +3104,7 @@ class Program_weight_tensor_parameter_258: class Program_weight_tensor_parameter_259: name = "parameter_259" + original_name = "layer_norm_1311.w_0" shape = [1024] dtype = "float32" min_val = float("0.679088") @@ -2856,6 +3116,7 @@ class Program_weight_tensor_parameter_259: class Program_weight_tensor_parameter_260: name = "parameter_260" + original_name = "layer_norm_1310.b_0" shape = [1024] dtype = "float32" min_val = float("-0.426888") @@ -2867,6 +3128,7 @@ class Program_weight_tensor_parameter_260: class Program_weight_tensor_parameter_261: name = "parameter_261" + original_name = "layer_norm_1310.w_0" shape = [1024] dtype = "float32" min_val = float("0.722768") @@ -2878,6 +3140,7 @@ class Program_weight_tensor_parameter_261: class Program_weight_tensor_parameter_262: name = "parameter_262" + original_name = "linear_3794.b_0" shape = [1024] dtype = "float32" min_val = float("-0.19762") @@ -2889,6 +3152,7 @@ class Program_weight_tensor_parameter_262: class Program_weight_tensor_parameter_263: name = "parameter_263" + original_name = "linear_3794.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-0.713308") @@ -2900,6 +3164,7 @@ class Program_weight_tensor_parameter_263: class Program_weight_tensor_parameter_264: name = "parameter_264" + original_name = "linear_3793.b_0" shape = [4096] dtype = "float32" min_val = float("-0.475189") @@ -2911,6 +3176,7 @@ class Program_weight_tensor_parameter_264: class Program_weight_tensor_parameter_265: name = "parameter_265" + original_name = "linear_3793.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.55513") @@ -2922,6 +3188,7 @@ class Program_weight_tensor_parameter_265: class Program_weight_tensor_parameter_266: name = "parameter_266" + original_name = "linear_3792.b_0" shape = [1024] dtype = "float32" min_val = float("-0.152921") @@ -2933,6 +3200,7 @@ class Program_weight_tensor_parameter_266: class Program_weight_tensor_parameter_267: name = "parameter_267" + original_name = "linear_3792.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.566284") @@ -2944,6 +3212,7 @@ class Program_weight_tensor_parameter_267: class Program_weight_tensor_parameter_268: name = "parameter_268" + original_name = "linear_3791.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0605642") @@ -2955,6 +3224,7 @@ class Program_weight_tensor_parameter_268: class Program_weight_tensor_parameter_269: name = "parameter_269" + original_name = "linear_3791.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.373408") @@ -2966,6 +3236,7 @@ class Program_weight_tensor_parameter_269: class Program_weight_tensor_parameter_270: name = "parameter_270" + original_name = "linear_3790.b_0" shape = [1024] dtype = "float32" min_val = float("-0.181519") @@ -2977,6 +3248,7 @@ class Program_weight_tensor_parameter_270: class Program_weight_tensor_parameter_271: name = "parameter_271" + original_name = "linear_3790.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.533399") @@ -2988,6 +3260,7 @@ class Program_weight_tensor_parameter_271: class Program_weight_tensor_parameter_272: name = "parameter_272" + original_name = "linear_3789.b_0" shape = [1024] dtype = "float32" min_val = float("-0.46073") @@ -2999,6 +3272,7 @@ class Program_weight_tensor_parameter_272: class Program_weight_tensor_parameter_273: name = "parameter_273" + original_name = "linear_3789.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.53223") @@ -3010,6 +3284,7 @@ class Program_weight_tensor_parameter_273: class Program_weight_tensor_parameter_274: name = "parameter_274" + original_name = "layer_norm_1309.b_0" shape = [1024] dtype = "float32" min_val = float("-0.282648") @@ -3021,6 +3296,7 @@ class Program_weight_tensor_parameter_274: class Program_weight_tensor_parameter_275: name = "parameter_275" + original_name = "layer_norm_1309.w_0" shape = [1024] dtype = "float32" min_val = float("0.768463") @@ -3032,6 +3308,7 @@ class Program_weight_tensor_parameter_275: class Program_weight_tensor_parameter_276: name = "parameter_276" + original_name = "layer_norm_1308.b_0" shape = [1024] dtype = "float32" min_val = float("-0.378607") @@ -3043,6 +3320,7 @@ class Program_weight_tensor_parameter_276: class Program_weight_tensor_parameter_277: name = "parameter_277" + original_name = "layer_norm_1308.w_0" shape = [1024] dtype = "float32" min_val = float("0.77338") @@ -3054,6 +3332,7 @@ class Program_weight_tensor_parameter_277: class Program_weight_tensor_parameter_278: name = "parameter_278" + original_name = "linear_3788.b_0" shape = [1024] dtype = "float32" min_val = float("-0.257294") @@ -3065,6 +3344,7 @@ class Program_weight_tensor_parameter_278: class Program_weight_tensor_parameter_279: name = "parameter_279" + original_name = "linear_3788.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-0.726703") @@ -3076,6 +3356,7 @@ class Program_weight_tensor_parameter_279: class Program_weight_tensor_parameter_280: name = "parameter_280" + original_name = "linear_3787.b_0" shape = [4096] dtype = "float32" min_val = float("-0.37105") @@ -3087,6 +3368,7 @@ class Program_weight_tensor_parameter_280: class Program_weight_tensor_parameter_281: name = "parameter_281" + original_name = "linear_3787.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.574747") @@ -3098,6 +3380,7 @@ class Program_weight_tensor_parameter_281: class Program_weight_tensor_parameter_282: name = "parameter_282" + original_name = "linear_3786.b_0" shape = [1024] dtype = "float32" min_val = float("-0.155782") @@ -3109,6 +3392,7 @@ class Program_weight_tensor_parameter_282: class Program_weight_tensor_parameter_283: name = "parameter_283" + original_name = "linear_3786.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.542543") @@ -3120,6 +3404,7 @@ class Program_weight_tensor_parameter_283: class Program_weight_tensor_parameter_284: name = "parameter_284" + original_name = "linear_3785.b_0" shape = [1024] dtype = "float32" min_val = float("-0.104451") @@ -3131,6 +3416,7 @@ class Program_weight_tensor_parameter_284: class Program_weight_tensor_parameter_285: name = "parameter_285" + original_name = "linear_3785.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.317339") @@ -3142,6 +3428,7 @@ class Program_weight_tensor_parameter_285: class Program_weight_tensor_parameter_286: name = "parameter_286" + original_name = "linear_3784.b_0" shape = [1024] dtype = "float32" min_val = float("-0.190186") @@ -3153,6 +3440,7 @@ class Program_weight_tensor_parameter_286: class Program_weight_tensor_parameter_287: name = "parameter_287" + original_name = "linear_3784.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.550028") @@ -3164,6 +3452,7 @@ class Program_weight_tensor_parameter_287: class Program_weight_tensor_parameter_288: name = "parameter_288" + original_name = "linear_3783.b_0" shape = [1024] dtype = "float32" min_val = float("-0.487141") @@ -3175,6 +3464,7 @@ class Program_weight_tensor_parameter_288: class Program_weight_tensor_parameter_289: name = "parameter_289" + original_name = "linear_3783.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.525396") @@ -3186,6 +3476,7 @@ class Program_weight_tensor_parameter_289: class Program_weight_tensor_parameter_290: name = "parameter_290" + original_name = "layer_norm_1307.b_0" shape = [1024] dtype = "float32" min_val = float("-0.284865") @@ -3197,6 +3488,7 @@ class Program_weight_tensor_parameter_290: class Program_weight_tensor_parameter_291: name = "parameter_291" + original_name = "layer_norm_1307.w_0" shape = [1024] dtype = "float32" min_val = float("0.804962") @@ -3208,6 +3500,7 @@ class Program_weight_tensor_parameter_291: class Program_weight_tensor_parameter_292: name = "parameter_292" + original_name = "layer_norm_1306.b_0" shape = [1024] dtype = "float32" min_val = float("-0.433342") @@ -3219,6 +3512,7 @@ class Program_weight_tensor_parameter_292: class Program_weight_tensor_parameter_293: name = "parameter_293" + original_name = "layer_norm_1306.w_0" shape = [1024] dtype = "float32" min_val = float("0.782521") @@ -3230,6 +3524,7 @@ class Program_weight_tensor_parameter_293: class Program_weight_tensor_parameter_294: name = "parameter_294" + original_name = "linear_3782.b_0" shape = [1024] dtype = "float32" min_val = float("-0.253443") @@ -3241,6 +3536,7 @@ class Program_weight_tensor_parameter_294: class Program_weight_tensor_parameter_295: name = "parameter_295" + original_name = "linear_3782.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-0.783588") @@ -3252,6 +3548,7 @@ class Program_weight_tensor_parameter_295: class Program_weight_tensor_parameter_296: name = "parameter_296" + original_name = "linear_3781.b_0" shape = [4096] dtype = "float32" min_val = float("-0.509251") @@ -3263,6 +3560,7 @@ class Program_weight_tensor_parameter_296: class Program_weight_tensor_parameter_297: name = "parameter_297" + original_name = "linear_3781.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.572005") @@ -3274,6 +3572,7 @@ class Program_weight_tensor_parameter_297: class Program_weight_tensor_parameter_298: name = "parameter_298" + original_name = "linear_3780.b_0" shape = [1024] dtype = "float32" min_val = float("-0.106244") @@ -3285,6 +3584,7 @@ class Program_weight_tensor_parameter_298: class Program_weight_tensor_parameter_299: name = "parameter_299" + original_name = "linear_3780.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.628351") @@ -3296,6 +3596,7 @@ class Program_weight_tensor_parameter_299: class Program_weight_tensor_parameter_300: name = "parameter_300" + original_name = "linear_3779.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0576496") @@ -3307,6 +3608,7 @@ class Program_weight_tensor_parameter_300: class Program_weight_tensor_parameter_301: name = "parameter_301" + original_name = "linear_3779.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.331622") @@ -3318,6 +3620,7 @@ class Program_weight_tensor_parameter_301: class Program_weight_tensor_parameter_302: name = "parameter_302" + original_name = "linear_3778.b_0" shape = [1024] dtype = "float32" min_val = float("-0.193359") @@ -3329,6 +3632,7 @@ class Program_weight_tensor_parameter_302: class Program_weight_tensor_parameter_303: name = "parameter_303" + original_name = "linear_3778.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.48255") @@ -3340,6 +3644,7 @@ class Program_weight_tensor_parameter_303: class Program_weight_tensor_parameter_304: name = "parameter_304" + original_name = "linear_3777.b_0" shape = [1024] dtype = "float32" min_val = float("-0.505243") @@ -3351,6 +3656,7 @@ class Program_weight_tensor_parameter_304: class Program_weight_tensor_parameter_305: name = "parameter_305" + original_name = "linear_3777.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.458156") @@ -3362,6 +3668,7 @@ class Program_weight_tensor_parameter_305: class Program_weight_tensor_parameter_306: name = "parameter_306" + original_name = "layer_norm_1305.b_0" shape = [1024] dtype = "float32" min_val = float("-0.278834") @@ -3373,6 +3680,7 @@ class Program_weight_tensor_parameter_306: class Program_weight_tensor_parameter_307: name = "parameter_307" + original_name = "layer_norm_1305.w_0" shape = [1024] dtype = "float32" min_val = float("0.806547") @@ -3384,6 +3692,7 @@ class Program_weight_tensor_parameter_307: class Program_weight_tensor_parameter_308: name = "parameter_308" + original_name = "layer_norm_1304.b_0" shape = [1024] dtype = "float32" min_val = float("-0.354849") @@ -3395,6 +3704,7 @@ class Program_weight_tensor_parameter_308: class Program_weight_tensor_parameter_309: name = "parameter_309" + original_name = "layer_norm_1304.w_0" shape = [1024] dtype = "float32" min_val = float("0.77167") @@ -3406,6 +3716,7 @@ class Program_weight_tensor_parameter_309: class Program_weight_tensor_parameter_310: name = "parameter_310" + original_name = "linear_3776.b_0" shape = [1024] dtype = "float32" min_val = float("-0.222084") @@ -3417,6 +3728,7 @@ class Program_weight_tensor_parameter_310: class Program_weight_tensor_parameter_311: name = "parameter_311" + original_name = "linear_3776.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-0.708304") @@ -3428,6 +3740,7 @@ class Program_weight_tensor_parameter_311: class Program_weight_tensor_parameter_312: name = "parameter_312" + original_name = "linear_3775.b_0" shape = [4096] dtype = "float32" min_val = float("-0.411729") @@ -3439,6 +3752,7 @@ class Program_weight_tensor_parameter_312: class Program_weight_tensor_parameter_313: name = "parameter_313" + original_name = "linear_3775.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.544797") @@ -3450,6 +3764,7 @@ class Program_weight_tensor_parameter_313: class Program_weight_tensor_parameter_314: name = "parameter_314" + original_name = "linear_3774.b_0" shape = [1024] dtype = "float32" min_val = float("-0.210687") @@ -3461,6 +3776,7 @@ class Program_weight_tensor_parameter_314: class Program_weight_tensor_parameter_315: name = "parameter_315" + original_name = "linear_3774.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.806491") @@ -3472,6 +3788,7 @@ class Program_weight_tensor_parameter_315: class Program_weight_tensor_parameter_316: name = "parameter_316" + original_name = "linear_3773.b_0" shape = [1024] dtype = "float32" min_val = float("-0.25632") @@ -3483,6 +3800,7 @@ class Program_weight_tensor_parameter_316: class Program_weight_tensor_parameter_317: name = "parameter_317" + original_name = "linear_3773.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.263863") @@ -3494,6 +3812,7 @@ class Program_weight_tensor_parameter_317: class Program_weight_tensor_parameter_318: name = "parameter_318" + original_name = "linear_3772.b_0" shape = [1024] dtype = "float32" min_val = float("-0.184815") @@ -3505,6 +3824,7 @@ class Program_weight_tensor_parameter_318: class Program_weight_tensor_parameter_319: name = "parameter_319" + original_name = "linear_3772.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.551826") @@ -3516,6 +3836,7 @@ class Program_weight_tensor_parameter_319: class Program_weight_tensor_parameter_320: name = "parameter_320" + original_name = "linear_3771.b_0" shape = [1024] dtype = "float32" min_val = float("-0.477676") @@ -3527,6 +3848,7 @@ class Program_weight_tensor_parameter_320: class Program_weight_tensor_parameter_321: name = "parameter_321" + original_name = "linear_3771.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.533925") @@ -3538,6 +3860,7 @@ class Program_weight_tensor_parameter_321: class Program_weight_tensor_parameter_322: name = "parameter_322" + original_name = "layer_norm_1303.b_0" shape = [1024] dtype = "float32" min_val = float("-0.307493") @@ -3549,6 +3872,7 @@ class Program_weight_tensor_parameter_322: class Program_weight_tensor_parameter_323: name = "parameter_323" + original_name = "layer_norm_1303.w_0" shape = [1024] dtype = "float32" min_val = float("0.760319") @@ -3560,6 +3884,7 @@ class Program_weight_tensor_parameter_323: class Program_weight_tensor_parameter_324: name = "parameter_324" + original_name = "layer_norm_1302.b_0" shape = [1024] dtype = "float32" min_val = float("-0.471947") @@ -3571,6 +3896,7 @@ class Program_weight_tensor_parameter_324: class Program_weight_tensor_parameter_325: name = "parameter_325" + original_name = "layer_norm_1302.w_0" shape = [1024] dtype = "float32" min_val = float("0.702611") @@ -3582,6 +3908,7 @@ class Program_weight_tensor_parameter_325: class Program_weight_tensor_parameter_326: name = "parameter_326" + original_name = "linear_3770.b_0" shape = [1024] dtype = "float32" min_val = float("-0.237399") @@ -3593,6 +3920,7 @@ class Program_weight_tensor_parameter_326: class Program_weight_tensor_parameter_327: name = "parameter_327" + original_name = "linear_3770.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-1.16249") @@ -3604,6 +3932,7 @@ class Program_weight_tensor_parameter_327: class Program_weight_tensor_parameter_328: name = "parameter_328" + original_name = "linear_3769.b_0" shape = [4096] dtype = "float32" min_val = float("-0.529424") @@ -3615,6 +3944,7 @@ class Program_weight_tensor_parameter_328: class Program_weight_tensor_parameter_329: name = "parameter_329" + original_name = "linear_3769.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.611725") @@ -3626,6 +3956,7 @@ class Program_weight_tensor_parameter_329: class Program_weight_tensor_parameter_330: name = "parameter_330" + original_name = "linear_3768.b_0" shape = [1024] dtype = "float32" min_val = float("-0.161454") @@ -3637,6 +3968,7 @@ class Program_weight_tensor_parameter_330: class Program_weight_tensor_parameter_331: name = "parameter_331" + original_name = "linear_3768.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.612582") @@ -3648,6 +3980,7 @@ class Program_weight_tensor_parameter_331: class Program_weight_tensor_parameter_332: name = "parameter_332" + original_name = "linear_3767.b_0" shape = [1024] dtype = "float32" min_val = float("-0.228532") @@ -3659,6 +3992,7 @@ class Program_weight_tensor_parameter_332: class Program_weight_tensor_parameter_333: name = "parameter_333" + original_name = "linear_3767.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.278547") @@ -3670,6 +4004,7 @@ class Program_weight_tensor_parameter_333: class Program_weight_tensor_parameter_334: name = "parameter_334" + original_name = "linear_3766.b_0" shape = [1024] dtype = "float32" min_val = float("-0.17029") @@ -3681,6 +4016,7 @@ class Program_weight_tensor_parameter_334: class Program_weight_tensor_parameter_335: name = "parameter_335" + original_name = "linear_3766.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.541136") @@ -3692,6 +4028,7 @@ class Program_weight_tensor_parameter_335: class Program_weight_tensor_parameter_336: name = "parameter_336" + original_name = "linear_3765.b_0" shape = [1024] dtype = "float32" min_val = float("-0.493535") @@ -3703,6 +4040,7 @@ class Program_weight_tensor_parameter_336: class Program_weight_tensor_parameter_337: name = "parameter_337" + original_name = "linear_3765.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.514264") @@ -3714,6 +4052,7 @@ class Program_weight_tensor_parameter_337: class Program_weight_tensor_parameter_338: name = "parameter_338" + original_name = "layer_norm_1301.b_0" shape = [1024] dtype = "float32" min_val = float("-0.332124") @@ -3725,6 +4064,7 @@ class Program_weight_tensor_parameter_338: class Program_weight_tensor_parameter_339: name = "parameter_339" + original_name = "layer_norm_1301.w_0" shape = [1024] dtype = "float32" min_val = float("0.730478") @@ -3736,6 +4076,7 @@ class Program_weight_tensor_parameter_339: class Program_weight_tensor_parameter_340: name = "parameter_340" + original_name = "layer_norm_1300.b_0" shape = [1024] dtype = "float32" min_val = float("-0.407631") @@ -3747,6 +4088,7 @@ class Program_weight_tensor_parameter_340: class Program_weight_tensor_parameter_341: name = "parameter_341" + original_name = "layer_norm_1300.w_0" shape = [1024] dtype = "float32" min_val = float("0.63354") @@ -3758,6 +4100,7 @@ class Program_weight_tensor_parameter_341: class Program_weight_tensor_parameter_342: name = "parameter_342" + original_name = "linear_3764.b_0" shape = [1024] dtype = "float32" min_val = float("-0.256634") @@ -3769,6 +4112,7 @@ class Program_weight_tensor_parameter_342: class Program_weight_tensor_parameter_343: name = "parameter_343" + original_name = "linear_3764.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-0.711196") @@ -3780,6 +4124,7 @@ class Program_weight_tensor_parameter_343: class Program_weight_tensor_parameter_344: name = "parameter_344" + original_name = "linear_3763.b_0" shape = [4096] dtype = "float32" min_val = float("-0.498922") @@ -3791,6 +4136,7 @@ class Program_weight_tensor_parameter_344: class Program_weight_tensor_parameter_345: name = "parameter_345" + original_name = "linear_3763.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.557151") @@ -3802,6 +4148,7 @@ class Program_weight_tensor_parameter_345: class Program_weight_tensor_parameter_346: name = "parameter_346" + original_name = "linear_3762.b_0" shape = [1024] dtype = "float32" min_val = float("-0.201235") @@ -3813,6 +4160,7 @@ class Program_weight_tensor_parameter_346: class Program_weight_tensor_parameter_347: name = "parameter_347" + original_name = "linear_3762.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.595007") @@ -3824,6 +4172,7 @@ class Program_weight_tensor_parameter_347: class Program_weight_tensor_parameter_348: name = "parameter_348" + original_name = "linear_3761.b_0" shape = [1024] dtype = "float32" min_val = float("-0.218069") @@ -3835,6 +4184,7 @@ class Program_weight_tensor_parameter_348: class Program_weight_tensor_parameter_349: name = "parameter_349" + original_name = "linear_3761.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.298953") @@ -3846,6 +4196,7 @@ class Program_weight_tensor_parameter_349: class Program_weight_tensor_parameter_350: name = "parameter_350" + original_name = "linear_3760.b_0" shape = [1024] dtype = "float32" min_val = float("-0.173457") @@ -3857,6 +4208,7 @@ class Program_weight_tensor_parameter_350: class Program_weight_tensor_parameter_351: name = "parameter_351" + original_name = "linear_3760.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.536147") @@ -3868,6 +4220,7 @@ class Program_weight_tensor_parameter_351: class Program_weight_tensor_parameter_352: name = "parameter_352" + original_name = "linear_3759.b_0" shape = [1024] dtype = "float32" min_val = float("-0.514189") @@ -3879,6 +4232,7 @@ class Program_weight_tensor_parameter_352: class Program_weight_tensor_parameter_353: name = "parameter_353" + original_name = "linear_3759.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.508585") @@ -3890,6 +4244,7 @@ class Program_weight_tensor_parameter_353: class Program_weight_tensor_parameter_354: name = "parameter_354" + original_name = "layer_norm_1299.b_0" shape = [1024] dtype = "float32" min_val = float("-0.569962") @@ -3901,6 +4256,7 @@ class Program_weight_tensor_parameter_354: class Program_weight_tensor_parameter_355: name = "parameter_355" + original_name = "layer_norm_1299.w_0" shape = [1024] dtype = "float32" min_val = float("0.713577") @@ -3912,6 +4268,7 @@ class Program_weight_tensor_parameter_355: class Program_weight_tensor_parameter_356: name = "parameter_356" + original_name = "layer_norm_1298.b_0" shape = [1024] dtype = "float32" min_val = float("-0.558029") @@ -3923,6 +4280,7 @@ class Program_weight_tensor_parameter_356: class Program_weight_tensor_parameter_357: name = "parameter_357" + original_name = "layer_norm_1298.w_0" shape = [1024] dtype = "float32" min_val = float("0.648001") @@ -3934,6 +4292,7 @@ class Program_weight_tensor_parameter_357: class Program_weight_tensor_parameter_358: name = "parameter_358" + original_name = "linear_3758.b_0" shape = [1024] dtype = "float32" min_val = float("-0.163382") @@ -3945,6 +4304,7 @@ class Program_weight_tensor_parameter_358: class Program_weight_tensor_parameter_359: name = "parameter_359" + original_name = "linear_3758.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-0.771041") @@ -3956,6 +4316,7 @@ class Program_weight_tensor_parameter_359: class Program_weight_tensor_parameter_360: name = "parameter_360" + original_name = "linear_3757.b_0" shape = [4096] dtype = "float32" min_val = float("-0.503343") @@ -3967,6 +4328,7 @@ class Program_weight_tensor_parameter_360: class Program_weight_tensor_parameter_361: name = "parameter_361" + original_name = "linear_3757.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.509248") @@ -3978,6 +4340,7 @@ class Program_weight_tensor_parameter_361: class Program_weight_tensor_parameter_362: name = "parameter_362" + original_name = "linear_3756.b_0" shape = [1024] dtype = "float32" min_val = float("-0.313562") @@ -3989,6 +4352,7 @@ class Program_weight_tensor_parameter_362: class Program_weight_tensor_parameter_363: name = "parameter_363" + original_name = "linear_3756.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.602474") @@ -4000,6 +4364,7 @@ class Program_weight_tensor_parameter_363: class Program_weight_tensor_parameter_364: name = "parameter_364" + original_name = "linear_3755.b_0" shape = [1024] dtype = "float32" min_val = float("-0.351547") @@ -4011,6 +4376,7 @@ class Program_weight_tensor_parameter_364: class Program_weight_tensor_parameter_365: name = "parameter_365" + original_name = "linear_3755.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.2425") @@ -4022,6 +4388,7 @@ class Program_weight_tensor_parameter_365: class Program_weight_tensor_parameter_366: name = "parameter_366" + original_name = "linear_3754.b_0" shape = [1024] dtype = "float32" min_val = float("-0.212967") @@ -4033,6 +4400,7 @@ class Program_weight_tensor_parameter_366: class Program_weight_tensor_parameter_367: name = "parameter_367" + original_name = "linear_3754.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.548985") @@ -4044,6 +4412,7 @@ class Program_weight_tensor_parameter_367: class Program_weight_tensor_parameter_368: name = "parameter_368" + original_name = "linear_3753.b_0" shape = [1024] dtype = "float32" min_val = float("-0.539128") @@ -4055,6 +4424,7 @@ class Program_weight_tensor_parameter_368: class Program_weight_tensor_parameter_369: name = "parameter_369" + original_name = "linear_3753.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.515722") @@ -4066,6 +4436,7 @@ class Program_weight_tensor_parameter_369: class Program_weight_tensor_parameter_370: name = "parameter_370" + original_name = "layer_norm_1297.b_0" shape = [1024] dtype = "float32" min_val = float("-0.517964") @@ -4077,6 +4448,7 @@ class Program_weight_tensor_parameter_370: class Program_weight_tensor_parameter_371: name = "parameter_371" + original_name = "layer_norm_1297.w_0" shape = [1024] dtype = "float32" min_val = float("0.474293") @@ -4088,6 +4460,7 @@ class Program_weight_tensor_parameter_371: class Program_weight_tensor_parameter_372: name = "parameter_372" + original_name = "layer_norm_1296.b_0" shape = [1024] dtype = "float32" min_val = float("-0.790177") @@ -4099,6 +4472,7 @@ class Program_weight_tensor_parameter_372: class Program_weight_tensor_parameter_373: name = "parameter_373" + original_name = "layer_norm_1296.w_0" shape = [1024] dtype = "float32" min_val = float("0.739699") @@ -4110,6 +4484,7 @@ class Program_weight_tensor_parameter_373: class Program_weight_tensor_parameter_374: name = "parameter_374" + original_name = "linear_3752.b_0" shape = [1024] dtype = "float32" min_val = float("-0.28176") @@ -4121,6 +4496,7 @@ class Program_weight_tensor_parameter_374: class Program_weight_tensor_parameter_375: name = "parameter_375" + original_name = "linear_3752.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-1.02163") @@ -4132,6 +4508,7 @@ class Program_weight_tensor_parameter_375: class Program_weight_tensor_parameter_376: name = "parameter_376" + original_name = "linear_3751.b_0" shape = [4096] dtype = "float32" min_val = float("-0.513157") @@ -4143,6 +4520,7 @@ class Program_weight_tensor_parameter_376: class Program_weight_tensor_parameter_377: name = "parameter_377" + original_name = "linear_3751.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.499794") @@ -4154,6 +4532,7 @@ class Program_weight_tensor_parameter_377: class Program_weight_tensor_parameter_378: name = "parameter_378" + original_name = "linear_3750.b_0" shape = [1024] dtype = "float32" min_val = float("-0.467788") @@ -4165,6 +4544,7 @@ class Program_weight_tensor_parameter_378: class Program_weight_tensor_parameter_379: name = "parameter_379" + original_name = "linear_3750.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.892194") @@ -4176,6 +4556,7 @@ class Program_weight_tensor_parameter_379: class Program_weight_tensor_parameter_380: name = "parameter_380" + original_name = "linear_3749.b_0" shape = [1024] dtype = "float32" min_val = float("-0.468041") @@ -4187,6 +4568,7 @@ class Program_weight_tensor_parameter_380: class Program_weight_tensor_parameter_381: name = "parameter_381" + original_name = "linear_3749.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.3715") @@ -4198,6 +4580,7 @@ class Program_weight_tensor_parameter_381: class Program_weight_tensor_parameter_382: name = "parameter_382" + original_name = "linear_3748.b_0" shape = [1024] dtype = "float32" min_val = float("-0.110297") @@ -4209,6 +4592,7 @@ class Program_weight_tensor_parameter_382: class Program_weight_tensor_parameter_383: name = "parameter_383" + original_name = "linear_3748.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.529198") @@ -4220,6 +4604,7 @@ class Program_weight_tensor_parameter_383: class Program_weight_tensor_parameter_384: name = "parameter_384" + original_name = "linear_3747.b_0" shape = [1024] dtype = "float32" min_val = float("-0.606511") @@ -4231,6 +4616,7 @@ class Program_weight_tensor_parameter_384: class Program_weight_tensor_parameter_385: name = "parameter_385" + original_name = "linear_3747.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.530734") @@ -4242,6 +4628,7 @@ class Program_weight_tensor_parameter_385: class Program_weight_tensor_parameter_386: name = "parameter_386" + original_name = "layer_norm_1295.b_0" shape = [1024] dtype = "float32" min_val = float("-0.619314") @@ -4253,6 +4640,7 @@ class Program_weight_tensor_parameter_386: class Program_weight_tensor_parameter_387: name = "parameter_387" + original_name = "layer_norm_1295.w_0" shape = [1024] dtype = "float32" min_val = float("0.375533") @@ -4264,6 +4652,7 @@ class Program_weight_tensor_parameter_387: class Program_weight_tensor_parameter_388: name = "parameter_388" + original_name = "embedding_248.w_0" shape = [514, 1024] dtype = "float32" min_val = float("-0.679154") @@ -4275,6 +4664,7 @@ class Program_weight_tensor_parameter_388: class Program_weight_tensor_parameter_389: name = "parameter_389" + original_name = "embedding_247.w_0" shape = [250002, 1024] dtype = "float32" min_val = float("-1.49479") diff --git a/paddle_samples/PaddleNLP/ernie-search-large-cross-encoder-marco-en/input_meta.py b/paddle_samples/PaddleNLP/ernie-search-large-cross-encoder-marco-en/input_meta.py index fd1f7db6f..060f275ed 100644 --- a/paddle_samples/PaddleNLP/ernie-search-large-cross-encoder-marco-en/input_meta.py +++ b/paddle_samples/PaddleNLP/ernie-search-large-cross-encoder-marco-en/input_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_data_0: name = "data_0" + original_name = "input_ids" shape = [1, 21] dtype = "int64" data = [ @@ -29,6 +30,7 @@ class Program_weight_tensor_data_0: class Program_weight_tensor_data_1: name = "data_1" + original_name = "token_type_ids" shape = [1, 21] dtype = "int64" data = [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0] diff --git a/paddle_samples/PaddleNLP/ernie-search-large-cross-encoder-marco-en/model.py b/paddle_samples/PaddleNLP/ernie-search-large-cross-encoder-marco-en/model.py index 43ef76b4a..93d97ed07 100644 --- a/paddle_samples/PaddleNLP/ernie-search-large-cross-encoder-marco-en/model.py +++ b/paddle_samples/PaddleNLP/ernie-search-large-cross-encoder-marco-en/model.py @@ -457,10 +457,11 @@ def forward( # pd_op.embedding: (1x21x1024xf32) <- (1x21xi64, 512x1024xf32) embedding_1 = paddle._C_ops.embedding(subtract_0, parameter_389, -1, False) - del parameter_389 + del parameter_389, subtract_0 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_0 = paddle._C_ops.add(embedding_0, embedding_1) + del embedding_0, embedding_1 # pd_op.embedding: (1x21x1024xf32) <- (1x21xi64, 4x1024xf32) embedding_2 = paddle._C_ops.embedding(data_1, parameter_388, -1, False) @@ -468,6 +469,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_1 = paddle._C_ops.add(add_0, embedding_2) + del add_0, embedding_2 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_0, layer_norm_1, layer_norm_2 = (lambda x, f: f(x))( @@ -476,233 +478,17 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_386, parameter_387 + del add_1, parameter_386, parameter_387 # pd_op.full: (1xf32) <- () full_4 = paddle._C_ops.full( [1], float("0.1"), paddle.float32, paddle.core.CPUPlace() ) - # pd_op.assign: (1xf32) <- (1xf32) - assign_0 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_1 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_2 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_3 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_4 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_5 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_6 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_7 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_8 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_9 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_10 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_11 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_12 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_13 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_14 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_15 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_16 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_17 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_18 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_19 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_20 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_21 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_22 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_23 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_24 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_25 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_26 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_27 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_28 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_29 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_30 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_31 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_32 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_33 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_34 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_35 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_36 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_37 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_38 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_39 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_40 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_41 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_42 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_43 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_44 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_45 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_46 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_47 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_48 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_49 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_50 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_51 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_52 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_53 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_54 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_55 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_56 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_57 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_58 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_59 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_60 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_61 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_62 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_63 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_64 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_65 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_66 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_67 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_68 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_69 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_70 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_71 = full_4 - # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_0, dropout_1 = (lambda x, f: f(x))( paddle._C_ops.dropout( - layer_norm_0, None, full_4, False, "upscale_in_train", 0, False + layer_norm_0, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -714,13 +500,14 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_2 = paddle._C_ops.add(matmul_0, parameter_384) - del parameter_384 + del matmul_0, parameter_384 # pd_op.full_int_array: (4xi64) <- () full_int_array_1 = [0, 0, 16, 64] # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_0 = paddle._C_ops.reshape(add_2, full_int_array_1) + del add_2 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_0 = paddle._C_ops.transpose(reshape_0, [0, 2, 1, 3]) @@ -732,7 +519,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_3 = paddle._C_ops.add(matmul_1, parameter_382) - del parameter_382 + del matmul_1, parameter_382 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_2 = paddle._C_ops.matmul(dropout_0, parameter_381, False, False) @@ -740,10 +527,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_4 = paddle._C_ops.add(matmul_2, parameter_380) - del parameter_380 + del matmul_2, parameter_380 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_1 = paddle._C_ops.reshape(add_3, full_int_array_1) + del add_3 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_1 = paddle._C_ops.transpose(reshape_1, [0, 2, 1, 3]) @@ -751,6 +539,7 @@ def forward( # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_2 = paddle._C_ops.reshape(add_4, full_int_array_1) + del add_4 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_2 = paddle._C_ops.transpose(reshape_2, [0, 2, 1, 3]) @@ -761,84 +550,17 @@ def forward( [1], float("0.125"), paddle.float32, paddle.core.CPUPlace() ) - # pd_op.assign: (1xf32) <- (1xf32) - assign_72 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_73 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_74 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_75 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_76 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_77 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_78 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_79 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_80 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_81 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_82 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_83 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_84 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_85 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_86 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_87 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_88 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_89 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_90 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_91 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_92 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_93 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_94 = full_5 - # pd_op.scale: (1x16x21x64xf32) <- (1x16x21x64xf32, 1xf32) scale_1 = paddle._C_ops.scale(transpose_0, full_5, float("0"), True) del transpose_0 # pd_op.matmul: (1x16x21x21xf32) <- (1x16x21x64xf32, 1x16x21x64xf32) matmul_3 = paddle._C_ops.matmul(scale_1, transpose_1, False, True) + del scale_1, transpose_1 # pd_op.add: (1x16x21x21xf32) <- (1x16x21x21xf32, 1x1x1x21xf32) add_5 = paddle._C_ops.add(matmul_3, unsqueeze_0) + del matmul_3 # pd_op.softmax: (1x16x21x21xf32) <- (1x16x21x21xf32) softmax_0 = paddle._C_ops.softmax(add_5, -1) @@ -847,13 +569,15 @@ def forward( # pd_op.dropout: (1x16x21x21xf32, 1x16x21x21xui8) <- (1x16x21x21xf32, None, 1xf32) dropout_2, dropout_3 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_0, None, full_4, False, "upscale_in_train", 0, False + softmax_0, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_0 # pd_op.matmul: (1x16x21x64xf32) <- (1x16x21x21xf32, 1x16x21x64xf32) matmul_4 = paddle._C_ops.matmul(dropout_2, transpose_2, False, False) + del dropout_2, transpose_2 # pd_op.transpose: (1x21x16x64xf32) <- (1x16x21x64xf32) transpose_3 = paddle._C_ops.transpose(matmul_4, [0, 2, 1, 3]) @@ -864,19 +588,20 @@ def forward( # pd_op.reshape: (1x21x1024xf32) <- (1x21x16x64xf32, 3xi64) reshape_3 = paddle._C_ops.reshape(transpose_3, full_int_array_2) + del transpose_3 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_5 = paddle._C_ops.matmul(reshape_3, parameter_379, False, False) - del parameter_379 + del parameter_379, reshape_3 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_6 = paddle._C_ops.add(matmul_5, parameter_378) - del parameter_378 + del matmul_5, parameter_378 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_4, dropout_5 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_6, None, full_4, False, "upscale_in_train", 0, False + add_6, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -884,6 +609,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_7 = paddle._C_ops.add(dropout_0, dropout_4) + del dropout_0, dropout_4 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_3, layer_norm_4, layer_norm_5 = (lambda x, f: f(x))( @@ -892,7 +618,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_372, parameter_373 + del add_7, parameter_372, parameter_373 # pd_op.matmul: (1x21x4096xf32) <- (1x21x1024xf32, 1024x4096xf32) matmul_6 = paddle._C_ops.matmul(layer_norm_3, parameter_377, False, False) @@ -900,23 +626,24 @@ def forward( # pd_op.add: (1x21x4096xf32) <- (1x21x4096xf32, 4096xf32) add_8 = paddle._C_ops.add(matmul_6, parameter_376) - del parameter_376 + del matmul_6, parameter_376 # pd_op.gelu: (1x21x4096xf32) <- (1x21x4096xf32) gelu_0 = paddle._C_ops.gelu(add_8, False) + del add_8 # pd_op.matmul: (1x21x1024xf32) <- (1x21x4096xf32, 4096x1024xf32) matmul_7 = paddle._C_ops.matmul(gelu_0, parameter_375, False, False) - del parameter_375 + del gelu_0, parameter_375 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_9 = paddle._C_ops.add(matmul_7, parameter_374) - del parameter_374 + del matmul_7, parameter_374 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_6, dropout_7 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_9, None, full_4, False, "upscale_in_train", 0, False + add_9, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -924,6 +651,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_10 = paddle._C_ops.add(layer_norm_3, dropout_6) + del dropout_6, layer_norm_3 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_6, layer_norm_7, layer_norm_8 = (lambda x, f: f(x))( @@ -932,7 +660,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_370, parameter_371 + del add_10, parameter_370, parameter_371 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_8 = paddle._C_ops.matmul(layer_norm_6, parameter_369, False, False) @@ -940,10 +668,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_11 = paddle._C_ops.add(matmul_8, parameter_368) - del parameter_368 + del matmul_8, parameter_368 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_4 = paddle._C_ops.reshape(add_11, full_int_array_1) + del add_11 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_4 = paddle._C_ops.transpose(reshape_4, [0, 2, 1, 3]) @@ -955,7 +684,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_12 = paddle._C_ops.add(matmul_9, parameter_366) - del parameter_366 + del matmul_9, parameter_366 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_10 = paddle._C_ops.matmul(layer_norm_6, parameter_365, False, False) @@ -963,10 +692,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_13 = paddle._C_ops.add(matmul_10, parameter_364) - del parameter_364 + del matmul_10, parameter_364 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_5 = paddle._C_ops.reshape(add_12, full_int_array_1) + del add_12 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_5 = paddle._C_ops.transpose(reshape_5, [0, 2, 1, 3]) @@ -974,6 +704,7 @@ def forward( # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_6 = paddle._C_ops.reshape(add_13, full_int_array_1) + del add_13 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_6 = paddle._C_ops.transpose(reshape_6, [0, 2, 1, 3]) @@ -985,9 +716,11 @@ def forward( # pd_op.matmul: (1x16x21x21xf32) <- (1x16x21x64xf32, 1x16x21x64xf32) matmul_11 = paddle._C_ops.matmul(scale_2, transpose_5, False, True) + del scale_2, transpose_5 # pd_op.add: (1x16x21x21xf32) <- (1x16x21x21xf32, 1x1x1x21xf32) add_14 = paddle._C_ops.add(matmul_11, unsqueeze_0) + del matmul_11 # pd_op.softmax: (1x16x21x21xf32) <- (1x16x21x21xf32) softmax_1 = paddle._C_ops.softmax(add_14, -1) @@ -996,13 +729,15 @@ def forward( # pd_op.dropout: (1x16x21x21xf32, 1x16x21x21xui8) <- (1x16x21x21xf32, None, 1xf32) dropout_8, dropout_9 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_1, None, full_4, False, "upscale_in_train", 0, False + softmax_1, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_1 # pd_op.matmul: (1x16x21x64xf32) <- (1x16x21x21xf32, 1x16x21x64xf32) matmul_12 = paddle._C_ops.matmul(dropout_8, transpose_6, False, False) + del dropout_8, transpose_6 # pd_op.transpose: (1x21x16x64xf32) <- (1x16x21x64xf32) transpose_7 = paddle._C_ops.transpose(matmul_12, [0, 2, 1, 3]) @@ -1010,19 +745,20 @@ def forward( # pd_op.reshape: (1x21x1024xf32) <- (1x21x16x64xf32, 3xi64) reshape_7 = paddle._C_ops.reshape(transpose_7, full_int_array_2) + del transpose_7 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_13 = paddle._C_ops.matmul(reshape_7, parameter_363, False, False) - del parameter_363 + del parameter_363, reshape_7 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_15 = paddle._C_ops.add(matmul_13, parameter_362) - del parameter_362 + del matmul_13, parameter_362 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_10, dropout_11 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_15, None, full_4, False, "upscale_in_train", 0, False + add_15, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1030,6 +766,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_16 = paddle._C_ops.add(layer_norm_6, dropout_10) + del dropout_10, layer_norm_6 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_9, layer_norm_10, layer_norm_11 = (lambda x, f: f(x))( @@ -1038,7 +775,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_356, parameter_357 + del add_16, parameter_356, parameter_357 # pd_op.matmul: (1x21x4096xf32) <- (1x21x1024xf32, 1024x4096xf32) matmul_14 = paddle._C_ops.matmul(layer_norm_9, parameter_361, False, False) @@ -1046,23 +783,24 @@ def forward( # pd_op.add: (1x21x4096xf32) <- (1x21x4096xf32, 4096xf32) add_17 = paddle._C_ops.add(matmul_14, parameter_360) - del parameter_360 + del matmul_14, parameter_360 # pd_op.gelu: (1x21x4096xf32) <- (1x21x4096xf32) gelu_1 = paddle._C_ops.gelu(add_17, False) + del add_17 # pd_op.matmul: (1x21x1024xf32) <- (1x21x4096xf32, 4096x1024xf32) matmul_15 = paddle._C_ops.matmul(gelu_1, parameter_359, False, False) - del parameter_359 + del gelu_1, parameter_359 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_18 = paddle._C_ops.add(matmul_15, parameter_358) - del parameter_358 + del matmul_15, parameter_358 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_12, dropout_13 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_18, None, full_4, False, "upscale_in_train", 0, False + add_18, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1070,6 +808,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_19 = paddle._C_ops.add(layer_norm_9, dropout_12) + del dropout_12, layer_norm_9 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_12, layer_norm_13, layer_norm_14 = (lambda x, f: f(x))( @@ -1078,7 +817,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_354, parameter_355 + del add_19, parameter_354, parameter_355 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_16 = paddle._C_ops.matmul(layer_norm_12, parameter_353, False, False) @@ -1086,10 +825,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_20 = paddle._C_ops.add(matmul_16, parameter_352) - del parameter_352 + del matmul_16, parameter_352 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_8 = paddle._C_ops.reshape(add_20, full_int_array_1) + del add_20 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_8 = paddle._C_ops.transpose(reshape_8, [0, 2, 1, 3]) @@ -1101,7 +841,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_21 = paddle._C_ops.add(matmul_17, parameter_350) - del parameter_350 + del matmul_17, parameter_350 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_18 = paddle._C_ops.matmul(layer_norm_12, parameter_349, False, False) @@ -1109,10 +849,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_22 = paddle._C_ops.add(matmul_18, parameter_348) - del parameter_348 + del matmul_18, parameter_348 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_9 = paddle._C_ops.reshape(add_21, full_int_array_1) + del add_21 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_9 = paddle._C_ops.transpose(reshape_9, [0, 2, 1, 3]) @@ -1120,6 +861,7 @@ def forward( # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_10 = paddle._C_ops.reshape(add_22, full_int_array_1) + del add_22 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_10 = paddle._C_ops.transpose(reshape_10, [0, 2, 1, 3]) @@ -1131,9 +873,11 @@ def forward( # pd_op.matmul: (1x16x21x21xf32) <- (1x16x21x64xf32, 1x16x21x64xf32) matmul_19 = paddle._C_ops.matmul(scale_3, transpose_9, False, True) + del scale_3, transpose_9 # pd_op.add: (1x16x21x21xf32) <- (1x16x21x21xf32, 1x1x1x21xf32) add_23 = paddle._C_ops.add(matmul_19, unsqueeze_0) + del matmul_19 # pd_op.softmax: (1x16x21x21xf32) <- (1x16x21x21xf32) softmax_2 = paddle._C_ops.softmax(add_23, -1) @@ -1142,13 +886,15 @@ def forward( # pd_op.dropout: (1x16x21x21xf32, 1x16x21x21xui8) <- (1x16x21x21xf32, None, 1xf32) dropout_14, dropout_15 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_2, None, full_4, False, "upscale_in_train", 0, False + softmax_2, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_2 # pd_op.matmul: (1x16x21x64xf32) <- (1x16x21x21xf32, 1x16x21x64xf32) matmul_20 = paddle._C_ops.matmul(dropout_14, transpose_10, False, False) + del dropout_14, transpose_10 # pd_op.transpose: (1x21x16x64xf32) <- (1x16x21x64xf32) transpose_11 = paddle._C_ops.transpose(matmul_20, [0, 2, 1, 3]) @@ -1156,19 +902,20 @@ def forward( # pd_op.reshape: (1x21x1024xf32) <- (1x21x16x64xf32, 3xi64) reshape_11 = paddle._C_ops.reshape(transpose_11, full_int_array_2) + del transpose_11 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_21 = paddle._C_ops.matmul(reshape_11, parameter_347, False, False) - del parameter_347 + del parameter_347, reshape_11 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_24 = paddle._C_ops.add(matmul_21, parameter_346) - del parameter_346 + del matmul_21, parameter_346 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_16, dropout_17 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_24, None, full_4, False, "upscale_in_train", 0, False + add_24, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1176,6 +923,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_25 = paddle._C_ops.add(layer_norm_12, dropout_16) + del dropout_16, layer_norm_12 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_15, layer_norm_16, layer_norm_17 = (lambda x, f: f(x))( @@ -1184,7 +932,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_340, parameter_341 + del add_25, parameter_340, parameter_341 # pd_op.matmul: (1x21x4096xf32) <- (1x21x1024xf32, 1024x4096xf32) matmul_22 = paddle._C_ops.matmul(layer_norm_15, parameter_345, False, False) @@ -1192,23 +940,24 @@ def forward( # pd_op.add: (1x21x4096xf32) <- (1x21x4096xf32, 4096xf32) add_26 = paddle._C_ops.add(matmul_22, parameter_344) - del parameter_344 + del matmul_22, parameter_344 # pd_op.gelu: (1x21x4096xf32) <- (1x21x4096xf32) gelu_2 = paddle._C_ops.gelu(add_26, False) + del add_26 # pd_op.matmul: (1x21x1024xf32) <- (1x21x4096xf32, 4096x1024xf32) matmul_23 = paddle._C_ops.matmul(gelu_2, parameter_343, False, False) - del parameter_343 + del gelu_2, parameter_343 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_27 = paddle._C_ops.add(matmul_23, parameter_342) - del parameter_342 + del matmul_23, parameter_342 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_18, dropout_19 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_27, None, full_4, False, "upscale_in_train", 0, False + add_27, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1216,6 +965,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_28 = paddle._C_ops.add(layer_norm_15, dropout_18) + del dropout_18, layer_norm_15 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_18, layer_norm_19, layer_norm_20 = (lambda x, f: f(x))( @@ -1224,7 +974,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_338, parameter_339 + del add_28, parameter_338, parameter_339 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_24 = paddle._C_ops.matmul(layer_norm_18, parameter_337, False, False) @@ -1232,10 +982,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_29 = paddle._C_ops.add(matmul_24, parameter_336) - del parameter_336 + del matmul_24, parameter_336 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_12 = paddle._C_ops.reshape(add_29, full_int_array_1) + del add_29 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_12 = paddle._C_ops.transpose(reshape_12, [0, 2, 1, 3]) @@ -1247,7 +998,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_30 = paddle._C_ops.add(matmul_25, parameter_334) - del parameter_334 + del matmul_25, parameter_334 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_26 = paddle._C_ops.matmul(layer_norm_18, parameter_333, False, False) @@ -1255,10 +1006,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_31 = paddle._C_ops.add(matmul_26, parameter_332) - del parameter_332 + del matmul_26, parameter_332 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_13 = paddle._C_ops.reshape(add_30, full_int_array_1) + del add_30 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_13 = paddle._C_ops.transpose(reshape_13, [0, 2, 1, 3]) @@ -1266,6 +1018,7 @@ def forward( # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_14 = paddle._C_ops.reshape(add_31, full_int_array_1) + del add_31 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_14 = paddle._C_ops.transpose(reshape_14, [0, 2, 1, 3]) @@ -1277,9 +1030,11 @@ def forward( # pd_op.matmul: (1x16x21x21xf32) <- (1x16x21x64xf32, 1x16x21x64xf32) matmul_27 = paddle._C_ops.matmul(scale_4, transpose_13, False, True) + del scale_4, transpose_13 # pd_op.add: (1x16x21x21xf32) <- (1x16x21x21xf32, 1x1x1x21xf32) add_32 = paddle._C_ops.add(matmul_27, unsqueeze_0) + del matmul_27 # pd_op.softmax: (1x16x21x21xf32) <- (1x16x21x21xf32) softmax_3 = paddle._C_ops.softmax(add_32, -1) @@ -1288,13 +1043,15 @@ def forward( # pd_op.dropout: (1x16x21x21xf32, 1x16x21x21xui8) <- (1x16x21x21xf32, None, 1xf32) dropout_20, dropout_21 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_3, None, full_4, False, "upscale_in_train", 0, False + softmax_3, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_3 # pd_op.matmul: (1x16x21x64xf32) <- (1x16x21x21xf32, 1x16x21x64xf32) matmul_28 = paddle._C_ops.matmul(dropout_20, transpose_14, False, False) + del dropout_20, transpose_14 # pd_op.transpose: (1x21x16x64xf32) <- (1x16x21x64xf32) transpose_15 = paddle._C_ops.transpose(matmul_28, [0, 2, 1, 3]) @@ -1302,19 +1059,20 @@ def forward( # pd_op.reshape: (1x21x1024xf32) <- (1x21x16x64xf32, 3xi64) reshape_15 = paddle._C_ops.reshape(transpose_15, full_int_array_2) + del transpose_15 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_29 = paddle._C_ops.matmul(reshape_15, parameter_331, False, False) - del parameter_331 + del parameter_331, reshape_15 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_33 = paddle._C_ops.add(matmul_29, parameter_330) - del parameter_330 + del matmul_29, parameter_330 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_22, dropout_23 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_33, None, full_4, False, "upscale_in_train", 0, False + add_33, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1322,6 +1080,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_34 = paddle._C_ops.add(layer_norm_18, dropout_22) + del dropout_22, layer_norm_18 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_21, layer_norm_22, layer_norm_23 = (lambda x, f: f(x))( @@ -1330,7 +1089,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_324, parameter_325 + del add_34, parameter_324, parameter_325 # pd_op.matmul: (1x21x4096xf32) <- (1x21x1024xf32, 1024x4096xf32) matmul_30 = paddle._C_ops.matmul(layer_norm_21, parameter_329, False, False) @@ -1338,23 +1097,24 @@ def forward( # pd_op.add: (1x21x4096xf32) <- (1x21x4096xf32, 4096xf32) add_35 = paddle._C_ops.add(matmul_30, parameter_328) - del parameter_328 + del matmul_30, parameter_328 # pd_op.gelu: (1x21x4096xf32) <- (1x21x4096xf32) gelu_3 = paddle._C_ops.gelu(add_35, False) + del add_35 # pd_op.matmul: (1x21x1024xf32) <- (1x21x4096xf32, 4096x1024xf32) matmul_31 = paddle._C_ops.matmul(gelu_3, parameter_327, False, False) - del parameter_327 + del gelu_3, parameter_327 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_36 = paddle._C_ops.add(matmul_31, parameter_326) - del parameter_326 + del matmul_31, parameter_326 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_24, dropout_25 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_36, None, full_4, False, "upscale_in_train", 0, False + add_36, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1362,6 +1122,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_37 = paddle._C_ops.add(layer_norm_21, dropout_24) + del dropout_24, layer_norm_21 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_24, layer_norm_25, layer_norm_26 = (lambda x, f: f(x))( @@ -1370,7 +1131,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_322, parameter_323 + del add_37, parameter_322, parameter_323 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_32 = paddle._C_ops.matmul(layer_norm_24, parameter_321, False, False) @@ -1378,10 +1139,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_38 = paddle._C_ops.add(matmul_32, parameter_320) - del parameter_320 + del matmul_32, parameter_320 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_16 = paddle._C_ops.reshape(add_38, full_int_array_1) + del add_38 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_16 = paddle._C_ops.transpose(reshape_16, [0, 2, 1, 3]) @@ -1393,7 +1155,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_39 = paddle._C_ops.add(matmul_33, parameter_318) - del parameter_318 + del matmul_33, parameter_318 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_34 = paddle._C_ops.matmul(layer_norm_24, parameter_317, False, False) @@ -1401,10 +1163,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_40 = paddle._C_ops.add(matmul_34, parameter_316) - del parameter_316 + del matmul_34, parameter_316 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_17 = paddle._C_ops.reshape(add_39, full_int_array_1) + del add_39 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_17 = paddle._C_ops.transpose(reshape_17, [0, 2, 1, 3]) @@ -1412,6 +1175,7 @@ def forward( # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_18 = paddle._C_ops.reshape(add_40, full_int_array_1) + del add_40 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_18 = paddle._C_ops.transpose(reshape_18, [0, 2, 1, 3]) @@ -1423,9 +1187,11 @@ def forward( # pd_op.matmul: (1x16x21x21xf32) <- (1x16x21x64xf32, 1x16x21x64xf32) matmul_35 = paddle._C_ops.matmul(scale_5, transpose_17, False, True) + del scale_5, transpose_17 # pd_op.add: (1x16x21x21xf32) <- (1x16x21x21xf32, 1x1x1x21xf32) add_41 = paddle._C_ops.add(matmul_35, unsqueeze_0) + del matmul_35 # pd_op.softmax: (1x16x21x21xf32) <- (1x16x21x21xf32) softmax_4 = paddle._C_ops.softmax(add_41, -1) @@ -1434,13 +1200,15 @@ def forward( # pd_op.dropout: (1x16x21x21xf32, 1x16x21x21xui8) <- (1x16x21x21xf32, None, 1xf32) dropout_26, dropout_27 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_4, None, full_4, False, "upscale_in_train", 0, False + softmax_4, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_4 # pd_op.matmul: (1x16x21x64xf32) <- (1x16x21x21xf32, 1x16x21x64xf32) matmul_36 = paddle._C_ops.matmul(dropout_26, transpose_18, False, False) + del dropout_26, transpose_18 # pd_op.transpose: (1x21x16x64xf32) <- (1x16x21x64xf32) transpose_19 = paddle._C_ops.transpose(matmul_36, [0, 2, 1, 3]) @@ -1448,19 +1216,20 @@ def forward( # pd_op.reshape: (1x21x1024xf32) <- (1x21x16x64xf32, 3xi64) reshape_19 = paddle._C_ops.reshape(transpose_19, full_int_array_2) + del transpose_19 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_37 = paddle._C_ops.matmul(reshape_19, parameter_315, False, False) - del parameter_315 + del parameter_315, reshape_19 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_42 = paddle._C_ops.add(matmul_37, parameter_314) - del parameter_314 + del matmul_37, parameter_314 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_28, dropout_29 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_42, None, full_4, False, "upscale_in_train", 0, False + add_42, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1468,6 +1237,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_43 = paddle._C_ops.add(layer_norm_24, dropout_28) + del dropout_28, layer_norm_24 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_27, layer_norm_28, layer_norm_29 = (lambda x, f: f(x))( @@ -1476,7 +1246,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_308, parameter_309 + del add_43, parameter_308, parameter_309 # pd_op.matmul: (1x21x4096xf32) <- (1x21x1024xf32, 1024x4096xf32) matmul_38 = paddle._C_ops.matmul(layer_norm_27, parameter_313, False, False) @@ -1484,23 +1254,24 @@ def forward( # pd_op.add: (1x21x4096xf32) <- (1x21x4096xf32, 4096xf32) add_44 = paddle._C_ops.add(matmul_38, parameter_312) - del parameter_312 + del matmul_38, parameter_312 # pd_op.gelu: (1x21x4096xf32) <- (1x21x4096xf32) gelu_4 = paddle._C_ops.gelu(add_44, False) + del add_44 # pd_op.matmul: (1x21x1024xf32) <- (1x21x4096xf32, 4096x1024xf32) matmul_39 = paddle._C_ops.matmul(gelu_4, parameter_311, False, False) - del parameter_311 + del gelu_4, parameter_311 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_45 = paddle._C_ops.add(matmul_39, parameter_310) - del parameter_310 + del matmul_39, parameter_310 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_30, dropout_31 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_45, None, full_4, False, "upscale_in_train", 0, False + add_45, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1508,6 +1279,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_46 = paddle._C_ops.add(layer_norm_27, dropout_30) + del dropout_30, layer_norm_27 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_30, layer_norm_31, layer_norm_32 = (lambda x, f: f(x))( @@ -1516,7 +1288,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_306, parameter_307 + del add_46, parameter_306, parameter_307 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_40 = paddle._C_ops.matmul(layer_norm_30, parameter_305, False, False) @@ -1524,10 +1296,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_47 = paddle._C_ops.add(matmul_40, parameter_304) - del parameter_304 + del matmul_40, parameter_304 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_20 = paddle._C_ops.reshape(add_47, full_int_array_1) + del add_47 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_20 = paddle._C_ops.transpose(reshape_20, [0, 2, 1, 3]) @@ -1539,7 +1312,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_48 = paddle._C_ops.add(matmul_41, parameter_302) - del parameter_302 + del matmul_41, parameter_302 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_42 = paddle._C_ops.matmul(layer_norm_30, parameter_301, False, False) @@ -1547,10 +1320,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_49 = paddle._C_ops.add(matmul_42, parameter_300) - del parameter_300 + del matmul_42, parameter_300 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_21 = paddle._C_ops.reshape(add_48, full_int_array_1) + del add_48 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_21 = paddle._C_ops.transpose(reshape_21, [0, 2, 1, 3]) @@ -1558,6 +1332,7 @@ def forward( # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_22 = paddle._C_ops.reshape(add_49, full_int_array_1) + del add_49 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_22 = paddle._C_ops.transpose(reshape_22, [0, 2, 1, 3]) @@ -1569,9 +1344,11 @@ def forward( # pd_op.matmul: (1x16x21x21xf32) <- (1x16x21x64xf32, 1x16x21x64xf32) matmul_43 = paddle._C_ops.matmul(scale_6, transpose_21, False, True) + del scale_6, transpose_21 # pd_op.add: (1x16x21x21xf32) <- (1x16x21x21xf32, 1x1x1x21xf32) add_50 = paddle._C_ops.add(matmul_43, unsqueeze_0) + del matmul_43 # pd_op.softmax: (1x16x21x21xf32) <- (1x16x21x21xf32) softmax_5 = paddle._C_ops.softmax(add_50, -1) @@ -1580,13 +1357,15 @@ def forward( # pd_op.dropout: (1x16x21x21xf32, 1x16x21x21xui8) <- (1x16x21x21xf32, None, 1xf32) dropout_32, dropout_33 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_5, None, full_4, False, "upscale_in_train", 0, False + softmax_5, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_5 # pd_op.matmul: (1x16x21x64xf32) <- (1x16x21x21xf32, 1x16x21x64xf32) matmul_44 = paddle._C_ops.matmul(dropout_32, transpose_22, False, False) + del dropout_32, transpose_22 # pd_op.transpose: (1x21x16x64xf32) <- (1x16x21x64xf32) transpose_23 = paddle._C_ops.transpose(matmul_44, [0, 2, 1, 3]) @@ -1594,19 +1373,20 @@ def forward( # pd_op.reshape: (1x21x1024xf32) <- (1x21x16x64xf32, 3xi64) reshape_23 = paddle._C_ops.reshape(transpose_23, full_int_array_2) + del transpose_23 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_45 = paddle._C_ops.matmul(reshape_23, parameter_299, False, False) - del parameter_299 + del parameter_299, reshape_23 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_51 = paddle._C_ops.add(matmul_45, parameter_298) - del parameter_298 + del matmul_45, parameter_298 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_34, dropout_35 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_51, None, full_4, False, "upscale_in_train", 0, False + add_51, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1614,6 +1394,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_52 = paddle._C_ops.add(layer_norm_30, dropout_34) + del dropout_34, layer_norm_30 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_33, layer_norm_34, layer_norm_35 = (lambda x, f: f(x))( @@ -1622,7 +1403,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_292, parameter_293 + del add_52, parameter_292, parameter_293 # pd_op.matmul: (1x21x4096xf32) <- (1x21x1024xf32, 1024x4096xf32) matmul_46 = paddle._C_ops.matmul(layer_norm_33, parameter_297, False, False) @@ -1630,23 +1411,24 @@ def forward( # pd_op.add: (1x21x4096xf32) <- (1x21x4096xf32, 4096xf32) add_53 = paddle._C_ops.add(matmul_46, parameter_296) - del parameter_296 + del matmul_46, parameter_296 # pd_op.gelu: (1x21x4096xf32) <- (1x21x4096xf32) gelu_5 = paddle._C_ops.gelu(add_53, False) + del add_53 # pd_op.matmul: (1x21x1024xf32) <- (1x21x4096xf32, 4096x1024xf32) matmul_47 = paddle._C_ops.matmul(gelu_5, parameter_295, False, False) - del parameter_295 + del gelu_5, parameter_295 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_54 = paddle._C_ops.add(matmul_47, parameter_294) - del parameter_294 + del matmul_47, parameter_294 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_36, dropout_37 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_54, None, full_4, False, "upscale_in_train", 0, False + add_54, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1654,6 +1436,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_55 = paddle._C_ops.add(layer_norm_33, dropout_36) + del dropout_36, layer_norm_33 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_36, layer_norm_37, layer_norm_38 = (lambda x, f: f(x))( @@ -1662,7 +1445,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_290, parameter_291 + del add_55, parameter_290, parameter_291 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_48 = paddle._C_ops.matmul(layer_norm_36, parameter_289, False, False) @@ -1670,10 +1453,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_56 = paddle._C_ops.add(matmul_48, parameter_288) - del parameter_288 + del matmul_48, parameter_288 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_24 = paddle._C_ops.reshape(add_56, full_int_array_1) + del add_56 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_24 = paddle._C_ops.transpose(reshape_24, [0, 2, 1, 3]) @@ -1685,7 +1469,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_57 = paddle._C_ops.add(matmul_49, parameter_286) - del parameter_286 + del matmul_49, parameter_286 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_50 = paddle._C_ops.matmul(layer_norm_36, parameter_285, False, False) @@ -1693,10 +1477,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_58 = paddle._C_ops.add(matmul_50, parameter_284) - del parameter_284 + del matmul_50, parameter_284 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_25 = paddle._C_ops.reshape(add_57, full_int_array_1) + del add_57 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_25 = paddle._C_ops.transpose(reshape_25, [0, 2, 1, 3]) @@ -1704,6 +1489,7 @@ def forward( # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_26 = paddle._C_ops.reshape(add_58, full_int_array_1) + del add_58 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_26 = paddle._C_ops.transpose(reshape_26, [0, 2, 1, 3]) @@ -1715,9 +1501,11 @@ def forward( # pd_op.matmul: (1x16x21x21xf32) <- (1x16x21x64xf32, 1x16x21x64xf32) matmul_51 = paddle._C_ops.matmul(scale_7, transpose_25, False, True) + del scale_7, transpose_25 # pd_op.add: (1x16x21x21xf32) <- (1x16x21x21xf32, 1x1x1x21xf32) add_59 = paddle._C_ops.add(matmul_51, unsqueeze_0) + del matmul_51 # pd_op.softmax: (1x16x21x21xf32) <- (1x16x21x21xf32) softmax_6 = paddle._C_ops.softmax(add_59, -1) @@ -1726,13 +1514,15 @@ def forward( # pd_op.dropout: (1x16x21x21xf32, 1x16x21x21xui8) <- (1x16x21x21xf32, None, 1xf32) dropout_38, dropout_39 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_6, None, full_4, False, "upscale_in_train", 0, False + softmax_6, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_6 # pd_op.matmul: (1x16x21x64xf32) <- (1x16x21x21xf32, 1x16x21x64xf32) matmul_52 = paddle._C_ops.matmul(dropout_38, transpose_26, False, False) + del dropout_38, transpose_26 # pd_op.transpose: (1x21x16x64xf32) <- (1x16x21x64xf32) transpose_27 = paddle._C_ops.transpose(matmul_52, [0, 2, 1, 3]) @@ -1740,19 +1530,20 @@ def forward( # pd_op.reshape: (1x21x1024xf32) <- (1x21x16x64xf32, 3xi64) reshape_27 = paddle._C_ops.reshape(transpose_27, full_int_array_2) + del transpose_27 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_53 = paddle._C_ops.matmul(reshape_27, parameter_283, False, False) - del parameter_283 + del parameter_283, reshape_27 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_60 = paddle._C_ops.add(matmul_53, parameter_282) - del parameter_282 + del matmul_53, parameter_282 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_40, dropout_41 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_60, None, full_4, False, "upscale_in_train", 0, False + add_60, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1760,6 +1551,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_61 = paddle._C_ops.add(layer_norm_36, dropout_40) + del dropout_40, layer_norm_36 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_39, layer_norm_40, layer_norm_41 = (lambda x, f: f(x))( @@ -1768,7 +1560,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_276, parameter_277 + del add_61, parameter_276, parameter_277 # pd_op.matmul: (1x21x4096xf32) <- (1x21x1024xf32, 1024x4096xf32) matmul_54 = paddle._C_ops.matmul(layer_norm_39, parameter_281, False, False) @@ -1776,23 +1568,24 @@ def forward( # pd_op.add: (1x21x4096xf32) <- (1x21x4096xf32, 4096xf32) add_62 = paddle._C_ops.add(matmul_54, parameter_280) - del parameter_280 + del matmul_54, parameter_280 # pd_op.gelu: (1x21x4096xf32) <- (1x21x4096xf32) gelu_6 = paddle._C_ops.gelu(add_62, False) + del add_62 # pd_op.matmul: (1x21x1024xf32) <- (1x21x4096xf32, 4096x1024xf32) matmul_55 = paddle._C_ops.matmul(gelu_6, parameter_279, False, False) - del parameter_279 + del gelu_6, parameter_279 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_63 = paddle._C_ops.add(matmul_55, parameter_278) - del parameter_278 + del matmul_55, parameter_278 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_42, dropout_43 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_63, None, full_4, False, "upscale_in_train", 0, False + add_63, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1800,6 +1593,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_64 = paddle._C_ops.add(layer_norm_39, dropout_42) + del dropout_42, layer_norm_39 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_42, layer_norm_43, layer_norm_44 = (lambda x, f: f(x))( @@ -1808,7 +1602,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_274, parameter_275 + del add_64, parameter_274, parameter_275 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_56 = paddle._C_ops.matmul(layer_norm_42, parameter_273, False, False) @@ -1816,10 +1610,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_65 = paddle._C_ops.add(matmul_56, parameter_272) - del parameter_272 + del matmul_56, parameter_272 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_28 = paddle._C_ops.reshape(add_65, full_int_array_1) + del add_65 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_28 = paddle._C_ops.transpose(reshape_28, [0, 2, 1, 3]) @@ -1831,7 +1626,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_66 = paddle._C_ops.add(matmul_57, parameter_270) - del parameter_270 + del matmul_57, parameter_270 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_58 = paddle._C_ops.matmul(layer_norm_42, parameter_269, False, False) @@ -1839,10 +1634,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_67 = paddle._C_ops.add(matmul_58, parameter_268) - del parameter_268 + del matmul_58, parameter_268 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_29 = paddle._C_ops.reshape(add_66, full_int_array_1) + del add_66 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_29 = paddle._C_ops.transpose(reshape_29, [0, 2, 1, 3]) @@ -1850,6 +1646,7 @@ def forward( # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_30 = paddle._C_ops.reshape(add_67, full_int_array_1) + del add_67 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_30 = paddle._C_ops.transpose(reshape_30, [0, 2, 1, 3]) @@ -1861,9 +1658,11 @@ def forward( # pd_op.matmul: (1x16x21x21xf32) <- (1x16x21x64xf32, 1x16x21x64xf32) matmul_59 = paddle._C_ops.matmul(scale_8, transpose_29, False, True) + del scale_8, transpose_29 # pd_op.add: (1x16x21x21xf32) <- (1x16x21x21xf32, 1x1x1x21xf32) add_68 = paddle._C_ops.add(matmul_59, unsqueeze_0) + del matmul_59 # pd_op.softmax: (1x16x21x21xf32) <- (1x16x21x21xf32) softmax_7 = paddle._C_ops.softmax(add_68, -1) @@ -1872,13 +1671,15 @@ def forward( # pd_op.dropout: (1x16x21x21xf32, 1x16x21x21xui8) <- (1x16x21x21xf32, None, 1xf32) dropout_44, dropout_45 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_7, None, full_4, False, "upscale_in_train", 0, False + softmax_7, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_7 # pd_op.matmul: (1x16x21x64xf32) <- (1x16x21x21xf32, 1x16x21x64xf32) matmul_60 = paddle._C_ops.matmul(dropout_44, transpose_30, False, False) + del dropout_44, transpose_30 # pd_op.transpose: (1x21x16x64xf32) <- (1x16x21x64xf32) transpose_31 = paddle._C_ops.transpose(matmul_60, [0, 2, 1, 3]) @@ -1886,19 +1687,20 @@ def forward( # pd_op.reshape: (1x21x1024xf32) <- (1x21x16x64xf32, 3xi64) reshape_31 = paddle._C_ops.reshape(transpose_31, full_int_array_2) + del transpose_31 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_61 = paddle._C_ops.matmul(reshape_31, parameter_267, False, False) - del parameter_267 + del parameter_267, reshape_31 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_69 = paddle._C_ops.add(matmul_61, parameter_266) - del parameter_266 + del matmul_61, parameter_266 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_46, dropout_47 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_69, None, full_4, False, "upscale_in_train", 0, False + add_69, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1906,6 +1708,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_70 = paddle._C_ops.add(layer_norm_42, dropout_46) + del dropout_46, layer_norm_42 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_45, layer_norm_46, layer_norm_47 = (lambda x, f: f(x))( @@ -1914,7 +1717,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_260, parameter_261 + del add_70, parameter_260, parameter_261 # pd_op.matmul: (1x21x4096xf32) <- (1x21x1024xf32, 1024x4096xf32) matmul_62 = paddle._C_ops.matmul(layer_norm_45, parameter_265, False, False) @@ -1922,23 +1725,24 @@ def forward( # pd_op.add: (1x21x4096xf32) <- (1x21x4096xf32, 4096xf32) add_71 = paddle._C_ops.add(matmul_62, parameter_264) - del parameter_264 + del matmul_62, parameter_264 # pd_op.gelu: (1x21x4096xf32) <- (1x21x4096xf32) gelu_7 = paddle._C_ops.gelu(add_71, False) + del add_71 # pd_op.matmul: (1x21x1024xf32) <- (1x21x4096xf32, 4096x1024xf32) matmul_63 = paddle._C_ops.matmul(gelu_7, parameter_263, False, False) - del parameter_263 + del gelu_7, parameter_263 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_72 = paddle._C_ops.add(matmul_63, parameter_262) - del parameter_262 + del matmul_63, parameter_262 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_48, dropout_49 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_72, None, full_4, False, "upscale_in_train", 0, False + add_72, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1946,6 +1750,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_73 = paddle._C_ops.add(layer_norm_45, dropout_48) + del dropout_48, layer_norm_45 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_48, layer_norm_49, layer_norm_50 = (lambda x, f: f(x))( @@ -1954,7 +1759,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_258, parameter_259 + del add_73, parameter_258, parameter_259 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_64 = paddle._C_ops.matmul(layer_norm_48, parameter_257, False, False) @@ -1962,10 +1767,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_74 = paddle._C_ops.add(matmul_64, parameter_256) - del parameter_256 + del matmul_64, parameter_256 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_32 = paddle._C_ops.reshape(add_74, full_int_array_1) + del add_74 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_32 = paddle._C_ops.transpose(reshape_32, [0, 2, 1, 3]) @@ -1977,7 +1783,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_75 = paddle._C_ops.add(matmul_65, parameter_254) - del parameter_254 + del matmul_65, parameter_254 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_66 = paddle._C_ops.matmul(layer_norm_48, parameter_253, False, False) @@ -1985,10 +1791,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_76 = paddle._C_ops.add(matmul_66, parameter_252) - del parameter_252 + del matmul_66, parameter_252 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_33 = paddle._C_ops.reshape(add_75, full_int_array_1) + del add_75 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_33 = paddle._C_ops.transpose(reshape_33, [0, 2, 1, 3]) @@ -1996,6 +1803,7 @@ def forward( # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_34 = paddle._C_ops.reshape(add_76, full_int_array_1) + del add_76 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_34 = paddle._C_ops.transpose(reshape_34, [0, 2, 1, 3]) @@ -2007,9 +1815,11 @@ def forward( # pd_op.matmul: (1x16x21x21xf32) <- (1x16x21x64xf32, 1x16x21x64xf32) matmul_67 = paddle._C_ops.matmul(scale_9, transpose_33, False, True) + del scale_9, transpose_33 # pd_op.add: (1x16x21x21xf32) <- (1x16x21x21xf32, 1x1x1x21xf32) add_77 = paddle._C_ops.add(matmul_67, unsqueeze_0) + del matmul_67 # pd_op.softmax: (1x16x21x21xf32) <- (1x16x21x21xf32) softmax_8 = paddle._C_ops.softmax(add_77, -1) @@ -2018,13 +1828,15 @@ def forward( # pd_op.dropout: (1x16x21x21xf32, 1x16x21x21xui8) <- (1x16x21x21xf32, None, 1xf32) dropout_50, dropout_51 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_8, None, full_4, False, "upscale_in_train", 0, False + softmax_8, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_8 # pd_op.matmul: (1x16x21x64xf32) <- (1x16x21x21xf32, 1x16x21x64xf32) matmul_68 = paddle._C_ops.matmul(dropout_50, transpose_34, False, False) + del dropout_50, transpose_34 # pd_op.transpose: (1x21x16x64xf32) <- (1x16x21x64xf32) transpose_35 = paddle._C_ops.transpose(matmul_68, [0, 2, 1, 3]) @@ -2032,19 +1844,20 @@ def forward( # pd_op.reshape: (1x21x1024xf32) <- (1x21x16x64xf32, 3xi64) reshape_35 = paddle._C_ops.reshape(transpose_35, full_int_array_2) + del transpose_35 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_69 = paddle._C_ops.matmul(reshape_35, parameter_251, False, False) - del parameter_251 + del parameter_251, reshape_35 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_78 = paddle._C_ops.add(matmul_69, parameter_250) - del parameter_250 + del matmul_69, parameter_250 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_52, dropout_53 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_78, None, full_4, False, "upscale_in_train", 0, False + add_78, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -2052,6 +1865,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_79 = paddle._C_ops.add(layer_norm_48, dropout_52) + del dropout_52, layer_norm_48 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_51, layer_norm_52, layer_norm_53 = (lambda x, f: f(x))( @@ -2060,7 +1874,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_244, parameter_245 + del add_79, parameter_244, parameter_245 # pd_op.matmul: (1x21x4096xf32) <- (1x21x1024xf32, 1024x4096xf32) matmul_70 = paddle._C_ops.matmul(layer_norm_51, parameter_249, False, False) @@ -2068,23 +1882,24 @@ def forward( # pd_op.add: (1x21x4096xf32) <- (1x21x4096xf32, 4096xf32) add_80 = paddle._C_ops.add(matmul_70, parameter_248) - del parameter_248 + del matmul_70, parameter_248 # pd_op.gelu: (1x21x4096xf32) <- (1x21x4096xf32) gelu_8 = paddle._C_ops.gelu(add_80, False) + del add_80 # pd_op.matmul: (1x21x1024xf32) <- (1x21x4096xf32, 4096x1024xf32) matmul_71 = paddle._C_ops.matmul(gelu_8, parameter_247, False, False) - del parameter_247 + del gelu_8, parameter_247 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_81 = paddle._C_ops.add(matmul_71, parameter_246) - del parameter_246 + del matmul_71, parameter_246 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_54, dropout_55 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_81, None, full_4, False, "upscale_in_train", 0, False + add_81, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -2092,6 +1907,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_82 = paddle._C_ops.add(layer_norm_51, dropout_54) + del dropout_54, layer_norm_51 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_54, layer_norm_55, layer_norm_56 = (lambda x, f: f(x))( @@ -2100,7 +1916,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_242, parameter_243 + del add_82, parameter_242, parameter_243 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_72 = paddle._C_ops.matmul(layer_norm_54, parameter_241, False, False) @@ -2108,10 +1924,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_83 = paddle._C_ops.add(matmul_72, parameter_240) - del parameter_240 + del matmul_72, parameter_240 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_36 = paddle._C_ops.reshape(add_83, full_int_array_1) + del add_83 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_36 = paddle._C_ops.transpose(reshape_36, [0, 2, 1, 3]) @@ -2123,7 +1940,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_84 = paddle._C_ops.add(matmul_73, parameter_238) - del parameter_238 + del matmul_73, parameter_238 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_74 = paddle._C_ops.matmul(layer_norm_54, parameter_237, False, False) @@ -2131,10 +1948,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_85 = paddle._C_ops.add(matmul_74, parameter_236) - del parameter_236 + del matmul_74, parameter_236 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_37 = paddle._C_ops.reshape(add_84, full_int_array_1) + del add_84 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_37 = paddle._C_ops.transpose(reshape_37, [0, 2, 1, 3]) @@ -2142,6 +1960,7 @@ def forward( # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_38 = paddle._C_ops.reshape(add_85, full_int_array_1) + del add_85 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_38 = paddle._C_ops.transpose(reshape_38, [0, 2, 1, 3]) @@ -2153,9 +1972,11 @@ def forward( # pd_op.matmul: (1x16x21x21xf32) <- (1x16x21x64xf32, 1x16x21x64xf32) matmul_75 = paddle._C_ops.matmul(scale_10, transpose_37, False, True) + del scale_10, transpose_37 # pd_op.add: (1x16x21x21xf32) <- (1x16x21x21xf32, 1x1x1x21xf32) add_86 = paddle._C_ops.add(matmul_75, unsqueeze_0) + del matmul_75 # pd_op.softmax: (1x16x21x21xf32) <- (1x16x21x21xf32) softmax_9 = paddle._C_ops.softmax(add_86, -1) @@ -2164,13 +1985,15 @@ def forward( # pd_op.dropout: (1x16x21x21xf32, 1x16x21x21xui8) <- (1x16x21x21xf32, None, 1xf32) dropout_56, dropout_57 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_9, None, full_4, False, "upscale_in_train", 0, False + softmax_9, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_9 # pd_op.matmul: (1x16x21x64xf32) <- (1x16x21x21xf32, 1x16x21x64xf32) matmul_76 = paddle._C_ops.matmul(dropout_56, transpose_38, False, False) + del dropout_56, transpose_38 # pd_op.transpose: (1x21x16x64xf32) <- (1x16x21x64xf32) transpose_39 = paddle._C_ops.transpose(matmul_76, [0, 2, 1, 3]) @@ -2178,19 +2001,20 @@ def forward( # pd_op.reshape: (1x21x1024xf32) <- (1x21x16x64xf32, 3xi64) reshape_39 = paddle._C_ops.reshape(transpose_39, full_int_array_2) + del transpose_39 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_77 = paddle._C_ops.matmul(reshape_39, parameter_235, False, False) - del parameter_235 + del parameter_235, reshape_39 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_87 = paddle._C_ops.add(matmul_77, parameter_234) - del parameter_234 + del matmul_77, parameter_234 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_58, dropout_59 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_87, None, full_4, False, "upscale_in_train", 0, False + add_87, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -2198,6 +2022,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_88 = paddle._C_ops.add(layer_norm_54, dropout_58) + del dropout_58, layer_norm_54 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_57, layer_norm_58, layer_norm_59 = (lambda x, f: f(x))( @@ -2206,7 +2031,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_228, parameter_229 + del add_88, parameter_228, parameter_229 # pd_op.matmul: (1x21x4096xf32) <- (1x21x1024xf32, 1024x4096xf32) matmul_78 = paddle._C_ops.matmul(layer_norm_57, parameter_233, False, False) @@ -2214,23 +2039,24 @@ def forward( # pd_op.add: (1x21x4096xf32) <- (1x21x4096xf32, 4096xf32) add_89 = paddle._C_ops.add(matmul_78, parameter_232) - del parameter_232 + del matmul_78, parameter_232 # pd_op.gelu: (1x21x4096xf32) <- (1x21x4096xf32) gelu_9 = paddle._C_ops.gelu(add_89, False) + del add_89 # pd_op.matmul: (1x21x1024xf32) <- (1x21x4096xf32, 4096x1024xf32) matmul_79 = paddle._C_ops.matmul(gelu_9, parameter_231, False, False) - del parameter_231 + del gelu_9, parameter_231 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_90 = paddle._C_ops.add(matmul_79, parameter_230) - del parameter_230 + del matmul_79, parameter_230 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_60, dropout_61 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_90, None, full_4, False, "upscale_in_train", 0, False + add_90, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -2238,6 +2064,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_91 = paddle._C_ops.add(layer_norm_57, dropout_60) + del dropout_60, layer_norm_57 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_60, layer_norm_61, layer_norm_62 = (lambda x, f: f(x))( @@ -2246,7 +2073,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_226, parameter_227 + del add_91, parameter_226, parameter_227 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_80 = paddle._C_ops.matmul(layer_norm_60, parameter_225, False, False) @@ -2254,10 +2081,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_92 = paddle._C_ops.add(matmul_80, parameter_224) - del parameter_224 + del matmul_80, parameter_224 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_40 = paddle._C_ops.reshape(add_92, full_int_array_1) + del add_92 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_40 = paddle._C_ops.transpose(reshape_40, [0, 2, 1, 3]) @@ -2269,7 +2097,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_93 = paddle._C_ops.add(matmul_81, parameter_222) - del parameter_222 + del matmul_81, parameter_222 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_82 = paddle._C_ops.matmul(layer_norm_60, parameter_221, False, False) @@ -2277,10 +2105,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_94 = paddle._C_ops.add(matmul_82, parameter_220) - del parameter_220 + del matmul_82, parameter_220 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_41 = paddle._C_ops.reshape(add_93, full_int_array_1) + del add_93 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_41 = paddle._C_ops.transpose(reshape_41, [0, 2, 1, 3]) @@ -2288,6 +2117,7 @@ def forward( # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_42 = paddle._C_ops.reshape(add_94, full_int_array_1) + del add_94 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_42 = paddle._C_ops.transpose(reshape_42, [0, 2, 1, 3]) @@ -2299,9 +2129,11 @@ def forward( # pd_op.matmul: (1x16x21x21xf32) <- (1x16x21x64xf32, 1x16x21x64xf32) matmul_83 = paddle._C_ops.matmul(scale_11, transpose_41, False, True) + del scale_11, transpose_41 # pd_op.add: (1x16x21x21xf32) <- (1x16x21x21xf32, 1x1x1x21xf32) add_95 = paddle._C_ops.add(matmul_83, unsqueeze_0) + del matmul_83 # pd_op.softmax: (1x16x21x21xf32) <- (1x16x21x21xf32) softmax_10 = paddle._C_ops.softmax(add_95, -1) @@ -2310,13 +2142,15 @@ def forward( # pd_op.dropout: (1x16x21x21xf32, 1x16x21x21xui8) <- (1x16x21x21xf32, None, 1xf32) dropout_62, dropout_63 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_10, None, full_4, False, "upscale_in_train", 0, False + softmax_10, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_10 # pd_op.matmul: (1x16x21x64xf32) <- (1x16x21x21xf32, 1x16x21x64xf32) matmul_84 = paddle._C_ops.matmul(dropout_62, transpose_42, False, False) + del dropout_62, transpose_42 # pd_op.transpose: (1x21x16x64xf32) <- (1x16x21x64xf32) transpose_43 = paddle._C_ops.transpose(matmul_84, [0, 2, 1, 3]) @@ -2324,19 +2158,20 @@ def forward( # pd_op.reshape: (1x21x1024xf32) <- (1x21x16x64xf32, 3xi64) reshape_43 = paddle._C_ops.reshape(transpose_43, full_int_array_2) + del transpose_43 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_85 = paddle._C_ops.matmul(reshape_43, parameter_219, False, False) - del parameter_219 + del parameter_219, reshape_43 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_96 = paddle._C_ops.add(matmul_85, parameter_218) - del parameter_218 + del matmul_85, parameter_218 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_64, dropout_65 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_96, None, full_4, False, "upscale_in_train", 0, False + add_96, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -2344,6 +2179,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_97 = paddle._C_ops.add(layer_norm_60, dropout_64) + del dropout_64, layer_norm_60 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_63, layer_norm_64, layer_norm_65 = (lambda x, f: f(x))( @@ -2352,7 +2188,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_212, parameter_213 + del add_97, parameter_212, parameter_213 # pd_op.matmul: (1x21x4096xf32) <- (1x21x1024xf32, 1024x4096xf32) matmul_86 = paddle._C_ops.matmul(layer_norm_63, parameter_217, False, False) @@ -2360,23 +2196,24 @@ def forward( # pd_op.add: (1x21x4096xf32) <- (1x21x4096xf32, 4096xf32) add_98 = paddle._C_ops.add(matmul_86, parameter_216) - del parameter_216 + del matmul_86, parameter_216 # pd_op.gelu: (1x21x4096xf32) <- (1x21x4096xf32) gelu_10 = paddle._C_ops.gelu(add_98, False) + del add_98 # pd_op.matmul: (1x21x1024xf32) <- (1x21x4096xf32, 4096x1024xf32) matmul_87 = paddle._C_ops.matmul(gelu_10, parameter_215, False, False) - del parameter_215 + del gelu_10, parameter_215 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_99 = paddle._C_ops.add(matmul_87, parameter_214) - del parameter_214 + del matmul_87, parameter_214 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_66, dropout_67 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_99, None, full_4, False, "upscale_in_train", 0, False + add_99, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -2384,6 +2221,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_100 = paddle._C_ops.add(layer_norm_63, dropout_66) + del dropout_66, layer_norm_63 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_66, layer_norm_67, layer_norm_68 = (lambda x, f: f(x))( @@ -2392,7 +2230,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_210, parameter_211 + del add_100, parameter_210, parameter_211 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_88 = paddle._C_ops.matmul(layer_norm_66, parameter_209, False, False) @@ -2400,10 +2238,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_101 = paddle._C_ops.add(matmul_88, parameter_208) - del parameter_208 + del matmul_88, parameter_208 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_44 = paddle._C_ops.reshape(add_101, full_int_array_1) + del add_101 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_44 = paddle._C_ops.transpose(reshape_44, [0, 2, 1, 3]) @@ -2415,7 +2254,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_102 = paddle._C_ops.add(matmul_89, parameter_206) - del parameter_206 + del matmul_89, parameter_206 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_90 = paddle._C_ops.matmul(layer_norm_66, parameter_205, False, False) @@ -2423,10 +2262,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_103 = paddle._C_ops.add(matmul_90, parameter_204) - del parameter_204 + del matmul_90, parameter_204 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_45 = paddle._C_ops.reshape(add_102, full_int_array_1) + del add_102 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_45 = paddle._C_ops.transpose(reshape_45, [0, 2, 1, 3]) @@ -2434,6 +2274,7 @@ def forward( # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_46 = paddle._C_ops.reshape(add_103, full_int_array_1) + del add_103 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_46 = paddle._C_ops.transpose(reshape_46, [0, 2, 1, 3]) @@ -2445,9 +2286,11 @@ def forward( # pd_op.matmul: (1x16x21x21xf32) <- (1x16x21x64xf32, 1x16x21x64xf32) matmul_91 = paddle._C_ops.matmul(scale_12, transpose_45, False, True) + del scale_12, transpose_45 # pd_op.add: (1x16x21x21xf32) <- (1x16x21x21xf32, 1x1x1x21xf32) add_104 = paddle._C_ops.add(matmul_91, unsqueeze_0) + del matmul_91 # pd_op.softmax: (1x16x21x21xf32) <- (1x16x21x21xf32) softmax_11 = paddle._C_ops.softmax(add_104, -1) @@ -2456,13 +2299,15 @@ def forward( # pd_op.dropout: (1x16x21x21xf32, 1x16x21x21xui8) <- (1x16x21x21xf32, None, 1xf32) dropout_68, dropout_69 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_11, None, full_4, False, "upscale_in_train", 0, False + softmax_11, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_11 # pd_op.matmul: (1x16x21x64xf32) <- (1x16x21x21xf32, 1x16x21x64xf32) matmul_92 = paddle._C_ops.matmul(dropout_68, transpose_46, False, False) + del dropout_68, transpose_46 # pd_op.transpose: (1x21x16x64xf32) <- (1x16x21x64xf32) transpose_47 = paddle._C_ops.transpose(matmul_92, [0, 2, 1, 3]) @@ -2470,19 +2315,20 @@ def forward( # pd_op.reshape: (1x21x1024xf32) <- (1x21x16x64xf32, 3xi64) reshape_47 = paddle._C_ops.reshape(transpose_47, full_int_array_2) + del transpose_47 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_93 = paddle._C_ops.matmul(reshape_47, parameter_203, False, False) - del parameter_203 + del parameter_203, reshape_47 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_105 = paddle._C_ops.add(matmul_93, parameter_202) - del parameter_202 + del matmul_93, parameter_202 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_70, dropout_71 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_105, None, full_4, False, "upscale_in_train", 0, False + add_105, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -2490,6 +2336,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_106 = paddle._C_ops.add(layer_norm_66, dropout_70) + del dropout_70, layer_norm_66 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_69, layer_norm_70, layer_norm_71 = (lambda x, f: f(x))( @@ -2498,7 +2345,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_196, parameter_197 + del add_106, parameter_196, parameter_197 # pd_op.matmul: (1x21x4096xf32) <- (1x21x1024xf32, 1024x4096xf32) matmul_94 = paddle._C_ops.matmul(layer_norm_69, parameter_201, False, False) @@ -2506,23 +2353,24 @@ def forward( # pd_op.add: (1x21x4096xf32) <- (1x21x4096xf32, 4096xf32) add_107 = paddle._C_ops.add(matmul_94, parameter_200) - del parameter_200 + del matmul_94, parameter_200 # pd_op.gelu: (1x21x4096xf32) <- (1x21x4096xf32) gelu_11 = paddle._C_ops.gelu(add_107, False) + del add_107 # pd_op.matmul: (1x21x1024xf32) <- (1x21x4096xf32, 4096x1024xf32) matmul_95 = paddle._C_ops.matmul(gelu_11, parameter_199, False, False) - del parameter_199 + del gelu_11, parameter_199 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_108 = paddle._C_ops.add(matmul_95, parameter_198) - del parameter_198 + del matmul_95, parameter_198 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_72, dropout_73 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_108, None, full_4, False, "upscale_in_train", 0, False + add_108, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -2530,6 +2378,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_109 = paddle._C_ops.add(layer_norm_69, dropout_72) + del dropout_72, layer_norm_69 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_72, layer_norm_73, layer_norm_74 = (lambda x, f: f(x))( @@ -2538,7 +2387,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_194, parameter_195 + del add_109, parameter_194, parameter_195 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_96 = paddle._C_ops.matmul(layer_norm_72, parameter_193, False, False) @@ -2546,10 +2395,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_110 = paddle._C_ops.add(matmul_96, parameter_192) - del parameter_192 + del matmul_96, parameter_192 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_48 = paddle._C_ops.reshape(add_110, full_int_array_1) + del add_110 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_48 = paddle._C_ops.transpose(reshape_48, [0, 2, 1, 3]) @@ -2561,7 +2411,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_111 = paddle._C_ops.add(matmul_97, parameter_190) - del parameter_190 + del matmul_97, parameter_190 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_98 = paddle._C_ops.matmul(layer_norm_72, parameter_189, False, False) @@ -2569,10 +2419,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_112 = paddle._C_ops.add(matmul_98, parameter_188) - del parameter_188 + del matmul_98, parameter_188 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_49 = paddle._C_ops.reshape(add_111, full_int_array_1) + del add_111 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_49 = paddle._C_ops.transpose(reshape_49, [0, 2, 1, 3]) @@ -2580,6 +2431,7 @@ def forward( # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_50 = paddle._C_ops.reshape(add_112, full_int_array_1) + del add_112 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_50 = paddle._C_ops.transpose(reshape_50, [0, 2, 1, 3]) @@ -2591,9 +2443,11 @@ def forward( # pd_op.matmul: (1x16x21x21xf32) <- (1x16x21x64xf32, 1x16x21x64xf32) matmul_99 = paddle._C_ops.matmul(scale_13, transpose_49, False, True) + del scale_13, transpose_49 # pd_op.add: (1x16x21x21xf32) <- (1x16x21x21xf32, 1x1x1x21xf32) add_113 = paddle._C_ops.add(matmul_99, unsqueeze_0) + del matmul_99 # pd_op.softmax: (1x16x21x21xf32) <- (1x16x21x21xf32) softmax_12 = paddle._C_ops.softmax(add_113, -1) @@ -2602,13 +2456,15 @@ def forward( # pd_op.dropout: (1x16x21x21xf32, 1x16x21x21xui8) <- (1x16x21x21xf32, None, 1xf32) dropout_74, dropout_75 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_12, None, full_4, False, "upscale_in_train", 0, False + softmax_12, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_12 # pd_op.matmul: (1x16x21x64xf32) <- (1x16x21x21xf32, 1x16x21x64xf32) matmul_100 = paddle._C_ops.matmul(dropout_74, transpose_50, False, False) + del dropout_74, transpose_50 # pd_op.transpose: (1x21x16x64xf32) <- (1x16x21x64xf32) transpose_51 = paddle._C_ops.transpose(matmul_100, [0, 2, 1, 3]) @@ -2616,19 +2472,20 @@ def forward( # pd_op.reshape: (1x21x1024xf32) <- (1x21x16x64xf32, 3xi64) reshape_51 = paddle._C_ops.reshape(transpose_51, full_int_array_2) + del transpose_51 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_101 = paddle._C_ops.matmul(reshape_51, parameter_187, False, False) - del parameter_187 + del parameter_187, reshape_51 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_114 = paddle._C_ops.add(matmul_101, parameter_186) - del parameter_186 + del matmul_101, parameter_186 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_76, dropout_77 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_114, None, full_4, False, "upscale_in_train", 0, False + add_114, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -2636,6 +2493,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_115 = paddle._C_ops.add(layer_norm_72, dropout_76) + del dropout_76, layer_norm_72 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_75, layer_norm_76, layer_norm_77 = (lambda x, f: f(x))( @@ -2644,7 +2502,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_180, parameter_181 + del add_115, parameter_180, parameter_181 # pd_op.matmul: (1x21x4096xf32) <- (1x21x1024xf32, 1024x4096xf32) matmul_102 = paddle._C_ops.matmul(layer_norm_75, parameter_185, False, False) @@ -2652,23 +2510,24 @@ def forward( # pd_op.add: (1x21x4096xf32) <- (1x21x4096xf32, 4096xf32) add_116 = paddle._C_ops.add(matmul_102, parameter_184) - del parameter_184 + del matmul_102, parameter_184 # pd_op.gelu: (1x21x4096xf32) <- (1x21x4096xf32) gelu_12 = paddle._C_ops.gelu(add_116, False) + del add_116 # pd_op.matmul: (1x21x1024xf32) <- (1x21x4096xf32, 4096x1024xf32) matmul_103 = paddle._C_ops.matmul(gelu_12, parameter_183, False, False) - del parameter_183 + del gelu_12, parameter_183 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_117 = paddle._C_ops.add(matmul_103, parameter_182) - del parameter_182 + del matmul_103, parameter_182 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_78, dropout_79 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_117, None, full_4, False, "upscale_in_train", 0, False + add_117, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -2676,6 +2535,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_118 = paddle._C_ops.add(layer_norm_75, dropout_78) + del dropout_78, layer_norm_75 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_78, layer_norm_79, layer_norm_80 = (lambda x, f: f(x))( @@ -2684,7 +2544,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_178, parameter_179 + del add_118, parameter_178, parameter_179 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_104 = paddle._C_ops.matmul(layer_norm_78, parameter_177, False, False) @@ -2692,10 +2552,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_119 = paddle._C_ops.add(matmul_104, parameter_176) - del parameter_176 + del matmul_104, parameter_176 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_52 = paddle._C_ops.reshape(add_119, full_int_array_1) + del add_119 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_52 = paddle._C_ops.transpose(reshape_52, [0, 2, 1, 3]) @@ -2707,7 +2568,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_120 = paddle._C_ops.add(matmul_105, parameter_174) - del parameter_174 + del matmul_105, parameter_174 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_106 = paddle._C_ops.matmul(layer_norm_78, parameter_173, False, False) @@ -2715,10 +2576,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_121 = paddle._C_ops.add(matmul_106, parameter_172) - del parameter_172 + del matmul_106, parameter_172 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_53 = paddle._C_ops.reshape(add_120, full_int_array_1) + del add_120 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_53 = paddle._C_ops.transpose(reshape_53, [0, 2, 1, 3]) @@ -2726,6 +2588,7 @@ def forward( # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_54 = paddle._C_ops.reshape(add_121, full_int_array_1) + del add_121 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_54 = paddle._C_ops.transpose(reshape_54, [0, 2, 1, 3]) @@ -2737,9 +2600,11 @@ def forward( # pd_op.matmul: (1x16x21x21xf32) <- (1x16x21x64xf32, 1x16x21x64xf32) matmul_107 = paddle._C_ops.matmul(scale_14, transpose_53, False, True) + del scale_14, transpose_53 # pd_op.add: (1x16x21x21xf32) <- (1x16x21x21xf32, 1x1x1x21xf32) add_122 = paddle._C_ops.add(matmul_107, unsqueeze_0) + del matmul_107 # pd_op.softmax: (1x16x21x21xf32) <- (1x16x21x21xf32) softmax_13 = paddle._C_ops.softmax(add_122, -1) @@ -2748,13 +2613,15 @@ def forward( # pd_op.dropout: (1x16x21x21xf32, 1x16x21x21xui8) <- (1x16x21x21xf32, None, 1xf32) dropout_80, dropout_81 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_13, None, full_4, False, "upscale_in_train", 0, False + softmax_13, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_13 # pd_op.matmul: (1x16x21x64xf32) <- (1x16x21x21xf32, 1x16x21x64xf32) matmul_108 = paddle._C_ops.matmul(dropout_80, transpose_54, False, False) + del dropout_80, transpose_54 # pd_op.transpose: (1x21x16x64xf32) <- (1x16x21x64xf32) transpose_55 = paddle._C_ops.transpose(matmul_108, [0, 2, 1, 3]) @@ -2762,19 +2629,20 @@ def forward( # pd_op.reshape: (1x21x1024xf32) <- (1x21x16x64xf32, 3xi64) reshape_55 = paddle._C_ops.reshape(transpose_55, full_int_array_2) + del transpose_55 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_109 = paddle._C_ops.matmul(reshape_55, parameter_171, False, False) - del parameter_171 + del parameter_171, reshape_55 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_123 = paddle._C_ops.add(matmul_109, parameter_170) - del parameter_170 + del matmul_109, parameter_170 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_82, dropout_83 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_123, None, full_4, False, "upscale_in_train", 0, False + add_123, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -2782,6 +2650,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_124 = paddle._C_ops.add(layer_norm_78, dropout_82) + del dropout_82, layer_norm_78 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_81, layer_norm_82, layer_norm_83 = (lambda x, f: f(x))( @@ -2790,7 +2659,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_164, parameter_165 + del add_124, parameter_164, parameter_165 # pd_op.matmul: (1x21x4096xf32) <- (1x21x1024xf32, 1024x4096xf32) matmul_110 = paddle._C_ops.matmul(layer_norm_81, parameter_169, False, False) @@ -2798,23 +2667,24 @@ def forward( # pd_op.add: (1x21x4096xf32) <- (1x21x4096xf32, 4096xf32) add_125 = paddle._C_ops.add(matmul_110, parameter_168) - del parameter_168 + del matmul_110, parameter_168 # pd_op.gelu: (1x21x4096xf32) <- (1x21x4096xf32) gelu_13 = paddle._C_ops.gelu(add_125, False) + del add_125 # pd_op.matmul: (1x21x1024xf32) <- (1x21x4096xf32, 4096x1024xf32) matmul_111 = paddle._C_ops.matmul(gelu_13, parameter_167, False, False) - del parameter_167 + del gelu_13, parameter_167 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_126 = paddle._C_ops.add(matmul_111, parameter_166) - del parameter_166 + del matmul_111, parameter_166 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_84, dropout_85 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_126, None, full_4, False, "upscale_in_train", 0, False + add_126, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -2822,6 +2692,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_127 = paddle._C_ops.add(layer_norm_81, dropout_84) + del dropout_84, layer_norm_81 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_84, layer_norm_85, layer_norm_86 = (lambda x, f: f(x))( @@ -2830,7 +2701,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_162, parameter_163 + del add_127, parameter_162, parameter_163 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_112 = paddle._C_ops.matmul(layer_norm_84, parameter_161, False, False) @@ -2838,10 +2709,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_128 = paddle._C_ops.add(matmul_112, parameter_160) - del parameter_160 + del matmul_112, parameter_160 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_56 = paddle._C_ops.reshape(add_128, full_int_array_1) + del add_128 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_56 = paddle._C_ops.transpose(reshape_56, [0, 2, 1, 3]) @@ -2853,7 +2725,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_129 = paddle._C_ops.add(matmul_113, parameter_158) - del parameter_158 + del matmul_113, parameter_158 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_114 = paddle._C_ops.matmul(layer_norm_84, parameter_157, False, False) @@ -2861,10 +2733,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_130 = paddle._C_ops.add(matmul_114, parameter_156) - del parameter_156 + del matmul_114, parameter_156 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_57 = paddle._C_ops.reshape(add_129, full_int_array_1) + del add_129 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_57 = paddle._C_ops.transpose(reshape_57, [0, 2, 1, 3]) @@ -2872,6 +2745,7 @@ def forward( # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_58 = paddle._C_ops.reshape(add_130, full_int_array_1) + del add_130 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_58 = paddle._C_ops.transpose(reshape_58, [0, 2, 1, 3]) @@ -2883,9 +2757,11 @@ def forward( # pd_op.matmul: (1x16x21x21xf32) <- (1x16x21x64xf32, 1x16x21x64xf32) matmul_115 = paddle._C_ops.matmul(scale_15, transpose_57, False, True) + del scale_15, transpose_57 # pd_op.add: (1x16x21x21xf32) <- (1x16x21x21xf32, 1x1x1x21xf32) add_131 = paddle._C_ops.add(matmul_115, unsqueeze_0) + del matmul_115 # pd_op.softmax: (1x16x21x21xf32) <- (1x16x21x21xf32) softmax_14 = paddle._C_ops.softmax(add_131, -1) @@ -2894,13 +2770,15 @@ def forward( # pd_op.dropout: (1x16x21x21xf32, 1x16x21x21xui8) <- (1x16x21x21xf32, None, 1xf32) dropout_86, dropout_87 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_14, None, full_4, False, "upscale_in_train", 0, False + softmax_14, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_14 # pd_op.matmul: (1x16x21x64xf32) <- (1x16x21x21xf32, 1x16x21x64xf32) matmul_116 = paddle._C_ops.matmul(dropout_86, transpose_58, False, False) + del dropout_86, transpose_58 # pd_op.transpose: (1x21x16x64xf32) <- (1x16x21x64xf32) transpose_59 = paddle._C_ops.transpose(matmul_116, [0, 2, 1, 3]) @@ -2908,19 +2786,20 @@ def forward( # pd_op.reshape: (1x21x1024xf32) <- (1x21x16x64xf32, 3xi64) reshape_59 = paddle._C_ops.reshape(transpose_59, full_int_array_2) + del transpose_59 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_117 = paddle._C_ops.matmul(reshape_59, parameter_155, False, False) - del parameter_155 + del parameter_155, reshape_59 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_132 = paddle._C_ops.add(matmul_117, parameter_154) - del parameter_154 + del matmul_117, parameter_154 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_88, dropout_89 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_132, None, full_4, False, "upscale_in_train", 0, False + add_132, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -2928,6 +2807,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_133 = paddle._C_ops.add(layer_norm_84, dropout_88) + del dropout_88, layer_norm_84 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_87, layer_norm_88, layer_norm_89 = (lambda x, f: f(x))( @@ -2936,7 +2816,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_148, parameter_149 + del add_133, parameter_148, parameter_149 # pd_op.matmul: (1x21x4096xf32) <- (1x21x1024xf32, 1024x4096xf32) matmul_118 = paddle._C_ops.matmul(layer_norm_87, parameter_153, False, False) @@ -2944,23 +2824,24 @@ def forward( # pd_op.add: (1x21x4096xf32) <- (1x21x4096xf32, 4096xf32) add_134 = paddle._C_ops.add(matmul_118, parameter_152) - del parameter_152 + del matmul_118, parameter_152 # pd_op.gelu: (1x21x4096xf32) <- (1x21x4096xf32) gelu_14 = paddle._C_ops.gelu(add_134, False) + del add_134 # pd_op.matmul: (1x21x1024xf32) <- (1x21x4096xf32, 4096x1024xf32) matmul_119 = paddle._C_ops.matmul(gelu_14, parameter_151, False, False) - del parameter_151 + del gelu_14, parameter_151 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_135 = paddle._C_ops.add(matmul_119, parameter_150) - del parameter_150 + del matmul_119, parameter_150 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_90, dropout_91 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_135, None, full_4, False, "upscale_in_train", 0, False + add_135, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -2968,6 +2849,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_136 = paddle._C_ops.add(layer_norm_87, dropout_90) + del dropout_90, layer_norm_87 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_90, layer_norm_91, layer_norm_92 = (lambda x, f: f(x))( @@ -2976,7 +2858,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_146, parameter_147 + del add_136, parameter_146, parameter_147 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_120 = paddle._C_ops.matmul(layer_norm_90, parameter_145, False, False) @@ -2984,10 +2866,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_137 = paddle._C_ops.add(matmul_120, parameter_144) - del parameter_144 + del matmul_120, parameter_144 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_60 = paddle._C_ops.reshape(add_137, full_int_array_1) + del add_137 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_60 = paddle._C_ops.transpose(reshape_60, [0, 2, 1, 3]) @@ -2999,7 +2882,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_138 = paddle._C_ops.add(matmul_121, parameter_142) - del parameter_142 + del matmul_121, parameter_142 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_122 = paddle._C_ops.matmul(layer_norm_90, parameter_141, False, False) @@ -3007,10 +2890,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_139 = paddle._C_ops.add(matmul_122, parameter_140) - del parameter_140 + del matmul_122, parameter_140 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_61 = paddle._C_ops.reshape(add_138, full_int_array_1) + del add_138 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_61 = paddle._C_ops.transpose(reshape_61, [0, 2, 1, 3]) @@ -3018,6 +2902,7 @@ def forward( # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_62 = paddle._C_ops.reshape(add_139, full_int_array_1) + del add_139 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_62 = paddle._C_ops.transpose(reshape_62, [0, 2, 1, 3]) @@ -3029,9 +2914,11 @@ def forward( # pd_op.matmul: (1x16x21x21xf32) <- (1x16x21x64xf32, 1x16x21x64xf32) matmul_123 = paddle._C_ops.matmul(scale_16, transpose_61, False, True) + del scale_16, transpose_61 # pd_op.add: (1x16x21x21xf32) <- (1x16x21x21xf32, 1x1x1x21xf32) add_140 = paddle._C_ops.add(matmul_123, unsqueeze_0) + del matmul_123 # pd_op.softmax: (1x16x21x21xf32) <- (1x16x21x21xf32) softmax_15 = paddle._C_ops.softmax(add_140, -1) @@ -3040,13 +2927,15 @@ def forward( # pd_op.dropout: (1x16x21x21xf32, 1x16x21x21xui8) <- (1x16x21x21xf32, None, 1xf32) dropout_92, dropout_93 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_15, None, full_4, False, "upscale_in_train", 0, False + softmax_15, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_15 # pd_op.matmul: (1x16x21x64xf32) <- (1x16x21x21xf32, 1x16x21x64xf32) matmul_124 = paddle._C_ops.matmul(dropout_92, transpose_62, False, False) + del dropout_92, transpose_62 # pd_op.transpose: (1x21x16x64xf32) <- (1x16x21x64xf32) transpose_63 = paddle._C_ops.transpose(matmul_124, [0, 2, 1, 3]) @@ -3054,19 +2943,20 @@ def forward( # pd_op.reshape: (1x21x1024xf32) <- (1x21x16x64xf32, 3xi64) reshape_63 = paddle._C_ops.reshape(transpose_63, full_int_array_2) + del transpose_63 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_125 = paddle._C_ops.matmul(reshape_63, parameter_139, False, False) - del parameter_139 + del parameter_139, reshape_63 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_141 = paddle._C_ops.add(matmul_125, parameter_138) - del parameter_138 + del matmul_125, parameter_138 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_94, dropout_95 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_141, None, full_4, False, "upscale_in_train", 0, False + add_141, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -3074,6 +2964,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_142 = paddle._C_ops.add(layer_norm_90, dropout_94) + del dropout_94, layer_norm_90 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_93, layer_norm_94, layer_norm_95 = (lambda x, f: f(x))( @@ -3082,7 +2973,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_132, parameter_133 + del add_142, parameter_132, parameter_133 # pd_op.matmul: (1x21x4096xf32) <- (1x21x1024xf32, 1024x4096xf32) matmul_126 = paddle._C_ops.matmul(layer_norm_93, parameter_137, False, False) @@ -3090,23 +2981,24 @@ def forward( # pd_op.add: (1x21x4096xf32) <- (1x21x4096xf32, 4096xf32) add_143 = paddle._C_ops.add(matmul_126, parameter_136) - del parameter_136 + del matmul_126, parameter_136 # pd_op.gelu: (1x21x4096xf32) <- (1x21x4096xf32) gelu_15 = paddle._C_ops.gelu(add_143, False) + del add_143 # pd_op.matmul: (1x21x1024xf32) <- (1x21x4096xf32, 4096x1024xf32) matmul_127 = paddle._C_ops.matmul(gelu_15, parameter_135, False, False) - del parameter_135 + del gelu_15, parameter_135 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_144 = paddle._C_ops.add(matmul_127, parameter_134) - del parameter_134 + del matmul_127, parameter_134 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_96, dropout_97 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_144, None, full_4, False, "upscale_in_train", 0, False + add_144, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -3114,6 +3006,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_145 = paddle._C_ops.add(layer_norm_93, dropout_96) + del dropout_96, layer_norm_93 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_96, layer_norm_97, layer_norm_98 = (lambda x, f: f(x))( @@ -3122,7 +3015,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_130, parameter_131 + del add_145, parameter_130, parameter_131 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_128 = paddle._C_ops.matmul(layer_norm_96, parameter_129, False, False) @@ -3130,10 +3023,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_146 = paddle._C_ops.add(matmul_128, parameter_128) - del parameter_128 + del matmul_128, parameter_128 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_64 = paddle._C_ops.reshape(add_146, full_int_array_1) + del add_146 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_64 = paddle._C_ops.transpose(reshape_64, [0, 2, 1, 3]) @@ -3145,7 +3039,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_147 = paddle._C_ops.add(matmul_129, parameter_126) - del parameter_126 + del matmul_129, parameter_126 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_130 = paddle._C_ops.matmul(layer_norm_96, parameter_125, False, False) @@ -3153,10 +3047,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_148 = paddle._C_ops.add(matmul_130, parameter_124) - del parameter_124 + del matmul_130, parameter_124 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_65 = paddle._C_ops.reshape(add_147, full_int_array_1) + del add_147 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_65 = paddle._C_ops.transpose(reshape_65, [0, 2, 1, 3]) @@ -3164,6 +3059,7 @@ def forward( # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_66 = paddle._C_ops.reshape(add_148, full_int_array_1) + del add_148 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_66 = paddle._C_ops.transpose(reshape_66, [0, 2, 1, 3]) @@ -3175,9 +3071,11 @@ def forward( # pd_op.matmul: (1x16x21x21xf32) <- (1x16x21x64xf32, 1x16x21x64xf32) matmul_131 = paddle._C_ops.matmul(scale_17, transpose_65, False, True) + del scale_17, transpose_65 # pd_op.add: (1x16x21x21xf32) <- (1x16x21x21xf32, 1x1x1x21xf32) add_149 = paddle._C_ops.add(matmul_131, unsqueeze_0) + del matmul_131 # pd_op.softmax: (1x16x21x21xf32) <- (1x16x21x21xf32) softmax_16 = paddle._C_ops.softmax(add_149, -1) @@ -3186,13 +3084,15 @@ def forward( # pd_op.dropout: (1x16x21x21xf32, 1x16x21x21xui8) <- (1x16x21x21xf32, None, 1xf32) dropout_98, dropout_99 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_16, None, full_4, False, "upscale_in_train", 0, False + softmax_16, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_16 # pd_op.matmul: (1x16x21x64xf32) <- (1x16x21x21xf32, 1x16x21x64xf32) matmul_132 = paddle._C_ops.matmul(dropout_98, transpose_66, False, False) + del dropout_98, transpose_66 # pd_op.transpose: (1x21x16x64xf32) <- (1x16x21x64xf32) transpose_67 = paddle._C_ops.transpose(matmul_132, [0, 2, 1, 3]) @@ -3200,19 +3100,20 @@ def forward( # pd_op.reshape: (1x21x1024xf32) <- (1x21x16x64xf32, 3xi64) reshape_67 = paddle._C_ops.reshape(transpose_67, full_int_array_2) + del transpose_67 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_133 = paddle._C_ops.matmul(reshape_67, parameter_123, False, False) - del parameter_123 + del parameter_123, reshape_67 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_150 = paddle._C_ops.add(matmul_133, parameter_122) - del parameter_122 + del matmul_133, parameter_122 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_100, dropout_101 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_150, None, full_4, False, "upscale_in_train", 0, False + add_150, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -3220,6 +3121,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_151 = paddle._C_ops.add(layer_norm_96, dropout_100) + del dropout_100, layer_norm_96 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_99, layer_norm_100, layer_norm_101 = (lambda x, f: f(x))( @@ -3228,7 +3130,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_116, parameter_117 + del add_151, parameter_116, parameter_117 # pd_op.matmul: (1x21x4096xf32) <- (1x21x1024xf32, 1024x4096xf32) matmul_134 = paddle._C_ops.matmul(layer_norm_99, parameter_121, False, False) @@ -3236,23 +3138,24 @@ def forward( # pd_op.add: (1x21x4096xf32) <- (1x21x4096xf32, 4096xf32) add_152 = paddle._C_ops.add(matmul_134, parameter_120) - del parameter_120 + del matmul_134, parameter_120 # pd_op.gelu: (1x21x4096xf32) <- (1x21x4096xf32) gelu_16 = paddle._C_ops.gelu(add_152, False) + del add_152 # pd_op.matmul: (1x21x1024xf32) <- (1x21x4096xf32, 4096x1024xf32) matmul_135 = paddle._C_ops.matmul(gelu_16, parameter_119, False, False) - del parameter_119 + del gelu_16, parameter_119 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_153 = paddle._C_ops.add(matmul_135, parameter_118) - del parameter_118 + del matmul_135, parameter_118 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_102, dropout_103 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_153, None, full_4, False, "upscale_in_train", 0, False + add_153, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -3260,6 +3163,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_154 = paddle._C_ops.add(layer_norm_99, dropout_102) + del dropout_102, layer_norm_99 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_102, layer_norm_103, layer_norm_104 = (lambda x, f: f(x))( @@ -3268,7 +3172,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_114, parameter_115 + del add_154, parameter_114, parameter_115 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_136 = paddle._C_ops.matmul(layer_norm_102, parameter_113, False, False) @@ -3276,10 +3180,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_155 = paddle._C_ops.add(matmul_136, parameter_112) - del parameter_112 + del matmul_136, parameter_112 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_68 = paddle._C_ops.reshape(add_155, full_int_array_1) + del add_155 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_68 = paddle._C_ops.transpose(reshape_68, [0, 2, 1, 3]) @@ -3291,7 +3196,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_156 = paddle._C_ops.add(matmul_137, parameter_110) - del parameter_110 + del matmul_137, parameter_110 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_138 = paddle._C_ops.matmul(layer_norm_102, parameter_109, False, False) @@ -3299,10 +3204,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_157 = paddle._C_ops.add(matmul_138, parameter_108) - del parameter_108 + del matmul_138, parameter_108 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_69 = paddle._C_ops.reshape(add_156, full_int_array_1) + del add_156 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_69 = paddle._C_ops.transpose(reshape_69, [0, 2, 1, 3]) @@ -3310,6 +3216,7 @@ def forward( # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_70 = paddle._C_ops.reshape(add_157, full_int_array_1) + del add_157 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_70 = paddle._C_ops.transpose(reshape_70, [0, 2, 1, 3]) @@ -3321,9 +3228,11 @@ def forward( # pd_op.matmul: (1x16x21x21xf32) <- (1x16x21x64xf32, 1x16x21x64xf32) matmul_139 = paddle._C_ops.matmul(scale_18, transpose_69, False, True) + del scale_18, transpose_69 # pd_op.add: (1x16x21x21xf32) <- (1x16x21x21xf32, 1x1x1x21xf32) add_158 = paddle._C_ops.add(matmul_139, unsqueeze_0) + del matmul_139 # pd_op.softmax: (1x16x21x21xf32) <- (1x16x21x21xf32) softmax_17 = paddle._C_ops.softmax(add_158, -1) @@ -3332,13 +3241,15 @@ def forward( # pd_op.dropout: (1x16x21x21xf32, 1x16x21x21xui8) <- (1x16x21x21xf32, None, 1xf32) dropout_104, dropout_105 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_17, None, full_4, False, "upscale_in_train", 0, False + softmax_17, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_17 # pd_op.matmul: (1x16x21x64xf32) <- (1x16x21x21xf32, 1x16x21x64xf32) matmul_140 = paddle._C_ops.matmul(dropout_104, transpose_70, False, False) + del dropout_104, transpose_70 # pd_op.transpose: (1x21x16x64xf32) <- (1x16x21x64xf32) transpose_71 = paddle._C_ops.transpose(matmul_140, [0, 2, 1, 3]) @@ -3346,19 +3257,20 @@ def forward( # pd_op.reshape: (1x21x1024xf32) <- (1x21x16x64xf32, 3xi64) reshape_71 = paddle._C_ops.reshape(transpose_71, full_int_array_2) + del transpose_71 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_141 = paddle._C_ops.matmul(reshape_71, parameter_107, False, False) - del parameter_107 + del parameter_107, reshape_71 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_159 = paddle._C_ops.add(matmul_141, parameter_106) - del parameter_106 + del matmul_141, parameter_106 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_106, dropout_107 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_159, None, full_4, False, "upscale_in_train", 0, False + add_159, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -3366,6 +3278,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_160 = paddle._C_ops.add(layer_norm_102, dropout_106) + del dropout_106, layer_norm_102 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_105, layer_norm_106, layer_norm_107 = (lambda x, f: f(x))( @@ -3374,7 +3287,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_100, parameter_101 + del add_160, parameter_100, parameter_101 # pd_op.matmul: (1x21x4096xf32) <- (1x21x1024xf32, 1024x4096xf32) matmul_142 = paddle._C_ops.matmul(layer_norm_105, parameter_105, False, False) @@ -3382,23 +3295,24 @@ def forward( # pd_op.add: (1x21x4096xf32) <- (1x21x4096xf32, 4096xf32) add_161 = paddle._C_ops.add(matmul_142, parameter_104) - del parameter_104 + del matmul_142, parameter_104 # pd_op.gelu: (1x21x4096xf32) <- (1x21x4096xf32) gelu_17 = paddle._C_ops.gelu(add_161, False) + del add_161 # pd_op.matmul: (1x21x1024xf32) <- (1x21x4096xf32, 4096x1024xf32) matmul_143 = paddle._C_ops.matmul(gelu_17, parameter_103, False, False) - del parameter_103 + del gelu_17, parameter_103 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_162 = paddle._C_ops.add(matmul_143, parameter_102) - del parameter_102 + del matmul_143, parameter_102 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_108, dropout_109 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_162, None, full_4, False, "upscale_in_train", 0, False + add_162, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -3406,6 +3320,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_163 = paddle._C_ops.add(layer_norm_105, dropout_108) + del dropout_108, layer_norm_105 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_108, layer_norm_109, layer_norm_110 = (lambda x, f: f(x))( @@ -3414,7 +3329,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_98, parameter_99 + del add_163, parameter_98, parameter_99 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_144 = paddle._C_ops.matmul(layer_norm_108, parameter_97, False, False) @@ -3422,10 +3337,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_164 = paddle._C_ops.add(matmul_144, parameter_96) - del parameter_96 + del matmul_144, parameter_96 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_72 = paddle._C_ops.reshape(add_164, full_int_array_1) + del add_164 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_72 = paddle._C_ops.transpose(reshape_72, [0, 2, 1, 3]) @@ -3437,7 +3353,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_165 = paddle._C_ops.add(matmul_145, parameter_94) - del parameter_94 + del matmul_145, parameter_94 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_146 = paddle._C_ops.matmul(layer_norm_108, parameter_93, False, False) @@ -3445,10 +3361,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_166 = paddle._C_ops.add(matmul_146, parameter_92) - del parameter_92 + del matmul_146, parameter_92 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_73 = paddle._C_ops.reshape(add_165, full_int_array_1) + del add_165 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_73 = paddle._C_ops.transpose(reshape_73, [0, 2, 1, 3]) @@ -3456,6 +3373,7 @@ def forward( # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_74 = paddle._C_ops.reshape(add_166, full_int_array_1) + del add_166 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_74 = paddle._C_ops.transpose(reshape_74, [0, 2, 1, 3]) @@ -3467,9 +3385,11 @@ def forward( # pd_op.matmul: (1x16x21x21xf32) <- (1x16x21x64xf32, 1x16x21x64xf32) matmul_147 = paddle._C_ops.matmul(scale_19, transpose_73, False, True) + del scale_19, transpose_73 # pd_op.add: (1x16x21x21xf32) <- (1x16x21x21xf32, 1x1x1x21xf32) add_167 = paddle._C_ops.add(matmul_147, unsqueeze_0) + del matmul_147 # pd_op.softmax: (1x16x21x21xf32) <- (1x16x21x21xf32) softmax_18 = paddle._C_ops.softmax(add_167, -1) @@ -3478,13 +3398,15 @@ def forward( # pd_op.dropout: (1x16x21x21xf32, 1x16x21x21xui8) <- (1x16x21x21xf32, None, 1xf32) dropout_110, dropout_111 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_18, None, full_4, False, "upscale_in_train", 0, False + softmax_18, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_18 # pd_op.matmul: (1x16x21x64xf32) <- (1x16x21x21xf32, 1x16x21x64xf32) matmul_148 = paddle._C_ops.matmul(dropout_110, transpose_74, False, False) + del dropout_110, transpose_74 # pd_op.transpose: (1x21x16x64xf32) <- (1x16x21x64xf32) transpose_75 = paddle._C_ops.transpose(matmul_148, [0, 2, 1, 3]) @@ -3492,19 +3414,20 @@ def forward( # pd_op.reshape: (1x21x1024xf32) <- (1x21x16x64xf32, 3xi64) reshape_75 = paddle._C_ops.reshape(transpose_75, full_int_array_2) + del transpose_75 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_149 = paddle._C_ops.matmul(reshape_75, parameter_91, False, False) - del parameter_91 + del parameter_91, reshape_75 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_168 = paddle._C_ops.add(matmul_149, parameter_90) - del parameter_90 + del matmul_149, parameter_90 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_112, dropout_113 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_168, None, full_4, False, "upscale_in_train", 0, False + add_168, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -3512,6 +3435,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_169 = paddle._C_ops.add(layer_norm_108, dropout_112) + del dropout_112, layer_norm_108 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_111, layer_norm_112, layer_norm_113 = (lambda x, f: f(x))( @@ -3520,7 +3444,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_84, parameter_85 + del add_169, parameter_84, parameter_85 # pd_op.matmul: (1x21x4096xf32) <- (1x21x1024xf32, 1024x4096xf32) matmul_150 = paddle._C_ops.matmul(layer_norm_111, parameter_89, False, False) @@ -3528,23 +3452,24 @@ def forward( # pd_op.add: (1x21x4096xf32) <- (1x21x4096xf32, 4096xf32) add_170 = paddle._C_ops.add(matmul_150, parameter_88) - del parameter_88 + del matmul_150, parameter_88 # pd_op.gelu: (1x21x4096xf32) <- (1x21x4096xf32) gelu_18 = paddle._C_ops.gelu(add_170, False) + del add_170 # pd_op.matmul: (1x21x1024xf32) <- (1x21x4096xf32, 4096x1024xf32) matmul_151 = paddle._C_ops.matmul(gelu_18, parameter_87, False, False) - del parameter_87 + del gelu_18, parameter_87 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_171 = paddle._C_ops.add(matmul_151, parameter_86) - del parameter_86 + del matmul_151, parameter_86 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_114, dropout_115 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_171, None, full_4, False, "upscale_in_train", 0, False + add_171, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -3552,6 +3477,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_172 = paddle._C_ops.add(layer_norm_111, dropout_114) + del dropout_114, layer_norm_111 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_114, layer_norm_115, layer_norm_116 = (lambda x, f: f(x))( @@ -3560,7 +3486,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_82, parameter_83 + del add_172, parameter_82, parameter_83 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_152 = paddle._C_ops.matmul(layer_norm_114, parameter_81, False, False) @@ -3568,10 +3494,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_173 = paddle._C_ops.add(matmul_152, parameter_80) - del parameter_80 + del matmul_152, parameter_80 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_76 = paddle._C_ops.reshape(add_173, full_int_array_1) + del add_173 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_76 = paddle._C_ops.transpose(reshape_76, [0, 2, 1, 3]) @@ -3583,7 +3510,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_174 = paddle._C_ops.add(matmul_153, parameter_78) - del parameter_78 + del matmul_153, parameter_78 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_154 = paddle._C_ops.matmul(layer_norm_114, parameter_77, False, False) @@ -3591,10 +3518,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_175 = paddle._C_ops.add(matmul_154, parameter_76) - del parameter_76 + del matmul_154, parameter_76 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_77 = paddle._C_ops.reshape(add_174, full_int_array_1) + del add_174 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_77 = paddle._C_ops.transpose(reshape_77, [0, 2, 1, 3]) @@ -3602,6 +3530,7 @@ def forward( # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_78 = paddle._C_ops.reshape(add_175, full_int_array_1) + del add_175 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_78 = paddle._C_ops.transpose(reshape_78, [0, 2, 1, 3]) @@ -3613,9 +3542,11 @@ def forward( # pd_op.matmul: (1x16x21x21xf32) <- (1x16x21x64xf32, 1x16x21x64xf32) matmul_155 = paddle._C_ops.matmul(scale_20, transpose_77, False, True) + del scale_20, transpose_77 # pd_op.add: (1x16x21x21xf32) <- (1x16x21x21xf32, 1x1x1x21xf32) add_176 = paddle._C_ops.add(matmul_155, unsqueeze_0) + del matmul_155 # pd_op.softmax: (1x16x21x21xf32) <- (1x16x21x21xf32) softmax_19 = paddle._C_ops.softmax(add_176, -1) @@ -3624,13 +3555,15 @@ def forward( # pd_op.dropout: (1x16x21x21xf32, 1x16x21x21xui8) <- (1x16x21x21xf32, None, 1xf32) dropout_116, dropout_117 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_19, None, full_4, False, "upscale_in_train", 0, False + softmax_19, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_19 # pd_op.matmul: (1x16x21x64xf32) <- (1x16x21x21xf32, 1x16x21x64xf32) matmul_156 = paddle._C_ops.matmul(dropout_116, transpose_78, False, False) + del dropout_116, transpose_78 # pd_op.transpose: (1x21x16x64xf32) <- (1x16x21x64xf32) transpose_79 = paddle._C_ops.transpose(matmul_156, [0, 2, 1, 3]) @@ -3638,19 +3571,20 @@ def forward( # pd_op.reshape: (1x21x1024xf32) <- (1x21x16x64xf32, 3xi64) reshape_79 = paddle._C_ops.reshape(transpose_79, full_int_array_2) + del transpose_79 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_157 = paddle._C_ops.matmul(reshape_79, parameter_75, False, False) - del parameter_75 + del parameter_75, reshape_79 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_177 = paddle._C_ops.add(matmul_157, parameter_74) - del parameter_74 + del matmul_157, parameter_74 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_118, dropout_119 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_177, None, full_4, False, "upscale_in_train", 0, False + add_177, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -3658,6 +3592,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_178 = paddle._C_ops.add(layer_norm_114, dropout_118) + del dropout_118, layer_norm_114 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_117, layer_norm_118, layer_norm_119 = (lambda x, f: f(x))( @@ -3666,7 +3601,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_68, parameter_69 + del add_178, parameter_68, parameter_69 # pd_op.matmul: (1x21x4096xf32) <- (1x21x1024xf32, 1024x4096xf32) matmul_158 = paddle._C_ops.matmul(layer_norm_117, parameter_73, False, False) @@ -3674,23 +3609,24 @@ def forward( # pd_op.add: (1x21x4096xf32) <- (1x21x4096xf32, 4096xf32) add_179 = paddle._C_ops.add(matmul_158, parameter_72) - del parameter_72 + del matmul_158, parameter_72 # pd_op.gelu: (1x21x4096xf32) <- (1x21x4096xf32) gelu_19 = paddle._C_ops.gelu(add_179, False) + del add_179 # pd_op.matmul: (1x21x1024xf32) <- (1x21x4096xf32, 4096x1024xf32) matmul_159 = paddle._C_ops.matmul(gelu_19, parameter_71, False, False) - del parameter_71 + del gelu_19, parameter_71 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_180 = paddle._C_ops.add(matmul_159, parameter_70) - del parameter_70 + del matmul_159, parameter_70 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_120, dropout_121 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_180, None, full_4, False, "upscale_in_train", 0, False + add_180, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -3698,6 +3634,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_181 = paddle._C_ops.add(layer_norm_117, dropout_120) + del dropout_120, layer_norm_117 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_120, layer_norm_121, layer_norm_122 = (lambda x, f: f(x))( @@ -3706,7 +3643,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_66, parameter_67 + del add_181, parameter_66, parameter_67 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_160 = paddle._C_ops.matmul(layer_norm_120, parameter_65, False, False) @@ -3714,10 +3651,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_182 = paddle._C_ops.add(matmul_160, parameter_64) - del parameter_64 + del matmul_160, parameter_64 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_80 = paddle._C_ops.reshape(add_182, full_int_array_1) + del add_182 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_80 = paddle._C_ops.transpose(reshape_80, [0, 2, 1, 3]) @@ -3729,7 +3667,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_183 = paddle._C_ops.add(matmul_161, parameter_62) - del parameter_62 + del matmul_161, parameter_62 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_162 = paddle._C_ops.matmul(layer_norm_120, parameter_61, False, False) @@ -3737,10 +3675,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_184 = paddle._C_ops.add(matmul_162, parameter_60) - del parameter_60 + del matmul_162, parameter_60 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_81 = paddle._C_ops.reshape(add_183, full_int_array_1) + del add_183 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_81 = paddle._C_ops.transpose(reshape_81, [0, 2, 1, 3]) @@ -3748,6 +3687,7 @@ def forward( # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_82 = paddle._C_ops.reshape(add_184, full_int_array_1) + del add_184 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_82 = paddle._C_ops.transpose(reshape_82, [0, 2, 1, 3]) @@ -3759,9 +3699,11 @@ def forward( # pd_op.matmul: (1x16x21x21xf32) <- (1x16x21x64xf32, 1x16x21x64xf32) matmul_163 = paddle._C_ops.matmul(scale_21, transpose_81, False, True) + del scale_21, transpose_81 # pd_op.add: (1x16x21x21xf32) <- (1x16x21x21xf32, 1x1x1x21xf32) add_185 = paddle._C_ops.add(matmul_163, unsqueeze_0) + del matmul_163 # pd_op.softmax: (1x16x21x21xf32) <- (1x16x21x21xf32) softmax_20 = paddle._C_ops.softmax(add_185, -1) @@ -3770,13 +3712,15 @@ def forward( # pd_op.dropout: (1x16x21x21xf32, 1x16x21x21xui8) <- (1x16x21x21xf32, None, 1xf32) dropout_122, dropout_123 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_20, None, full_4, False, "upscale_in_train", 0, False + softmax_20, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_20 # pd_op.matmul: (1x16x21x64xf32) <- (1x16x21x21xf32, 1x16x21x64xf32) matmul_164 = paddle._C_ops.matmul(dropout_122, transpose_82, False, False) + del dropout_122, transpose_82 # pd_op.transpose: (1x21x16x64xf32) <- (1x16x21x64xf32) transpose_83 = paddle._C_ops.transpose(matmul_164, [0, 2, 1, 3]) @@ -3784,19 +3728,20 @@ def forward( # pd_op.reshape: (1x21x1024xf32) <- (1x21x16x64xf32, 3xi64) reshape_83 = paddle._C_ops.reshape(transpose_83, full_int_array_2) + del transpose_83 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_165 = paddle._C_ops.matmul(reshape_83, parameter_59, False, False) - del parameter_59 + del parameter_59, reshape_83 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_186 = paddle._C_ops.add(matmul_165, parameter_58) - del parameter_58 + del matmul_165, parameter_58 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_124, dropout_125 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_186, None, full_4, False, "upscale_in_train", 0, False + add_186, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -3804,6 +3749,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_187 = paddle._C_ops.add(layer_norm_120, dropout_124) + del dropout_124, layer_norm_120 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_123, layer_norm_124, layer_norm_125 = (lambda x, f: f(x))( @@ -3812,7 +3758,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_52, parameter_53 + del add_187, parameter_52, parameter_53 # pd_op.matmul: (1x21x4096xf32) <- (1x21x1024xf32, 1024x4096xf32) matmul_166 = paddle._C_ops.matmul(layer_norm_123, parameter_57, False, False) @@ -3820,23 +3766,24 @@ def forward( # pd_op.add: (1x21x4096xf32) <- (1x21x4096xf32, 4096xf32) add_188 = paddle._C_ops.add(matmul_166, parameter_56) - del parameter_56 + del matmul_166, parameter_56 # pd_op.gelu: (1x21x4096xf32) <- (1x21x4096xf32) gelu_20 = paddle._C_ops.gelu(add_188, False) + del add_188 # pd_op.matmul: (1x21x1024xf32) <- (1x21x4096xf32, 4096x1024xf32) matmul_167 = paddle._C_ops.matmul(gelu_20, parameter_55, False, False) - del parameter_55 + del gelu_20, parameter_55 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_189 = paddle._C_ops.add(matmul_167, parameter_54) - del parameter_54 + del matmul_167, parameter_54 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_126, dropout_127 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_189, None, full_4, False, "upscale_in_train", 0, False + add_189, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -3844,6 +3791,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_190 = paddle._C_ops.add(layer_norm_123, dropout_126) + del dropout_126, layer_norm_123 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_126, layer_norm_127, layer_norm_128 = (lambda x, f: f(x))( @@ -3852,7 +3800,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_50, parameter_51 + del add_190, parameter_50, parameter_51 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_168 = paddle._C_ops.matmul(layer_norm_126, parameter_49, False, False) @@ -3860,10 +3808,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_191 = paddle._C_ops.add(matmul_168, parameter_48) - del parameter_48 + del matmul_168, parameter_48 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_84 = paddle._C_ops.reshape(add_191, full_int_array_1) + del add_191 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_84 = paddle._C_ops.transpose(reshape_84, [0, 2, 1, 3]) @@ -3875,7 +3824,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_192 = paddle._C_ops.add(matmul_169, parameter_46) - del parameter_46 + del matmul_169, parameter_46 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_170 = paddle._C_ops.matmul(layer_norm_126, parameter_45, False, False) @@ -3883,10 +3832,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_193 = paddle._C_ops.add(matmul_170, parameter_44) - del parameter_44 + del matmul_170, parameter_44 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_85 = paddle._C_ops.reshape(add_192, full_int_array_1) + del add_192 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_85 = paddle._C_ops.transpose(reshape_85, [0, 2, 1, 3]) @@ -3894,6 +3844,7 @@ def forward( # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_86 = paddle._C_ops.reshape(add_193, full_int_array_1) + del add_193 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_86 = paddle._C_ops.transpose(reshape_86, [0, 2, 1, 3]) @@ -3905,9 +3856,11 @@ def forward( # pd_op.matmul: (1x16x21x21xf32) <- (1x16x21x64xf32, 1x16x21x64xf32) matmul_171 = paddle._C_ops.matmul(scale_22, transpose_85, False, True) + del scale_22, transpose_85 # pd_op.add: (1x16x21x21xf32) <- (1x16x21x21xf32, 1x1x1x21xf32) add_194 = paddle._C_ops.add(matmul_171, unsqueeze_0) + del matmul_171 # pd_op.softmax: (1x16x21x21xf32) <- (1x16x21x21xf32) softmax_21 = paddle._C_ops.softmax(add_194, -1) @@ -3916,13 +3869,15 @@ def forward( # pd_op.dropout: (1x16x21x21xf32, 1x16x21x21xui8) <- (1x16x21x21xf32, None, 1xf32) dropout_128, dropout_129 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_21, None, full_4, False, "upscale_in_train", 0, False + softmax_21, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_21 # pd_op.matmul: (1x16x21x64xf32) <- (1x16x21x21xf32, 1x16x21x64xf32) matmul_172 = paddle._C_ops.matmul(dropout_128, transpose_86, False, False) + del dropout_128, transpose_86 # pd_op.transpose: (1x21x16x64xf32) <- (1x16x21x64xf32) transpose_87 = paddle._C_ops.transpose(matmul_172, [0, 2, 1, 3]) @@ -3930,19 +3885,20 @@ def forward( # pd_op.reshape: (1x21x1024xf32) <- (1x21x16x64xf32, 3xi64) reshape_87 = paddle._C_ops.reshape(transpose_87, full_int_array_2) + del transpose_87 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_173 = paddle._C_ops.matmul(reshape_87, parameter_43, False, False) - del parameter_43 + del parameter_43, reshape_87 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_195 = paddle._C_ops.add(matmul_173, parameter_42) - del parameter_42 + del matmul_173, parameter_42 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_130, dropout_131 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_195, None, full_4, False, "upscale_in_train", 0, False + add_195, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -3950,6 +3906,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_196 = paddle._C_ops.add(layer_norm_126, dropout_130) + del dropout_130, layer_norm_126 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_129, layer_norm_130, layer_norm_131 = (lambda x, f: f(x))( @@ -3958,7 +3915,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_36, parameter_37 + del add_196, parameter_36, parameter_37 # pd_op.matmul: (1x21x4096xf32) <- (1x21x1024xf32, 1024x4096xf32) matmul_174 = paddle._C_ops.matmul(layer_norm_129, parameter_41, False, False) @@ -3966,23 +3923,24 @@ def forward( # pd_op.add: (1x21x4096xf32) <- (1x21x4096xf32, 4096xf32) add_197 = paddle._C_ops.add(matmul_174, parameter_40) - del parameter_40 + del matmul_174, parameter_40 # pd_op.gelu: (1x21x4096xf32) <- (1x21x4096xf32) gelu_21 = paddle._C_ops.gelu(add_197, False) + del add_197 # pd_op.matmul: (1x21x1024xf32) <- (1x21x4096xf32, 4096x1024xf32) matmul_175 = paddle._C_ops.matmul(gelu_21, parameter_39, False, False) - del parameter_39 + del gelu_21, parameter_39 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_198 = paddle._C_ops.add(matmul_175, parameter_38) - del parameter_38 + del matmul_175, parameter_38 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_132, dropout_133 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_198, None, full_4, False, "upscale_in_train", 0, False + add_198, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -3990,6 +3948,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_199 = paddle._C_ops.add(layer_norm_129, dropout_132) + del dropout_132, layer_norm_129 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_132, layer_norm_133, layer_norm_134 = (lambda x, f: f(x))( @@ -3998,7 +3957,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_34, parameter_35 + del add_199, parameter_34, parameter_35 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_176 = paddle._C_ops.matmul(layer_norm_132, parameter_33, False, False) @@ -4006,10 +3965,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_200 = paddle._C_ops.add(matmul_176, parameter_32) - del parameter_32 + del matmul_176, parameter_32 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_88 = paddle._C_ops.reshape(add_200, full_int_array_1) + del add_200 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_88 = paddle._C_ops.transpose(reshape_88, [0, 2, 1, 3]) @@ -4021,7 +3981,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_201 = paddle._C_ops.add(matmul_177, parameter_30) - del parameter_30 + del matmul_177, parameter_30 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_178 = paddle._C_ops.matmul(layer_norm_132, parameter_29, False, False) @@ -4029,10 +3989,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_202 = paddle._C_ops.add(matmul_178, parameter_28) - del parameter_28 + del matmul_178, parameter_28 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_89 = paddle._C_ops.reshape(add_201, full_int_array_1) + del add_201 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_89 = paddle._C_ops.transpose(reshape_89, [0, 2, 1, 3]) @@ -4040,6 +4001,7 @@ def forward( # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_90 = paddle._C_ops.reshape(add_202, full_int_array_1) + del add_202 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_90 = paddle._C_ops.transpose(reshape_90, [0, 2, 1, 3]) @@ -4051,9 +4013,11 @@ def forward( # pd_op.matmul: (1x16x21x21xf32) <- (1x16x21x64xf32, 1x16x21x64xf32) matmul_179 = paddle._C_ops.matmul(scale_23, transpose_89, False, True) + del scale_23, transpose_89 # pd_op.add: (1x16x21x21xf32) <- (1x16x21x21xf32, 1x1x1x21xf32) add_203 = paddle._C_ops.add(matmul_179, unsqueeze_0) + del matmul_179 # pd_op.softmax: (1x16x21x21xf32) <- (1x16x21x21xf32) softmax_22 = paddle._C_ops.softmax(add_203, -1) @@ -4062,13 +4026,15 @@ def forward( # pd_op.dropout: (1x16x21x21xf32, 1x16x21x21xui8) <- (1x16x21x21xf32, None, 1xf32) dropout_134, dropout_135 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_22, None, full_4, False, "upscale_in_train", 0, False + softmax_22, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_22 # pd_op.matmul: (1x16x21x64xf32) <- (1x16x21x21xf32, 1x16x21x64xf32) matmul_180 = paddle._C_ops.matmul(dropout_134, transpose_90, False, False) + del dropout_134, transpose_90 # pd_op.transpose: (1x21x16x64xf32) <- (1x16x21x64xf32) transpose_91 = paddle._C_ops.transpose(matmul_180, [0, 2, 1, 3]) @@ -4076,19 +4042,20 @@ def forward( # pd_op.reshape: (1x21x1024xf32) <- (1x21x16x64xf32, 3xi64) reshape_91 = paddle._C_ops.reshape(transpose_91, full_int_array_2) + del transpose_91 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_181 = paddle._C_ops.matmul(reshape_91, parameter_27, False, False) - del parameter_27 + del parameter_27, reshape_91 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_204 = paddle._C_ops.add(matmul_181, parameter_26) - del parameter_26 + del matmul_181, parameter_26 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_136, dropout_137 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_204, None, full_4, False, "upscale_in_train", 0, False + add_204, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -4096,6 +4063,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_205 = paddle._C_ops.add(layer_norm_132, dropout_136) + del dropout_136, layer_norm_132 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_135, layer_norm_136, layer_norm_137 = (lambda x, f: f(x))( @@ -4104,7 +4072,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_20, parameter_21 + del add_205, parameter_20, parameter_21 # pd_op.matmul: (1x21x4096xf32) <- (1x21x1024xf32, 1024x4096xf32) matmul_182 = paddle._C_ops.matmul(layer_norm_135, parameter_25, False, False) @@ -4112,23 +4080,24 @@ def forward( # pd_op.add: (1x21x4096xf32) <- (1x21x4096xf32, 4096xf32) add_206 = paddle._C_ops.add(matmul_182, parameter_24) - del parameter_24 + del matmul_182, parameter_24 # pd_op.gelu: (1x21x4096xf32) <- (1x21x4096xf32) gelu_22 = paddle._C_ops.gelu(add_206, False) + del add_206 # pd_op.matmul: (1x21x1024xf32) <- (1x21x4096xf32, 4096x1024xf32) matmul_183 = paddle._C_ops.matmul(gelu_22, parameter_23, False, False) - del parameter_23 + del gelu_22, parameter_23 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_207 = paddle._C_ops.add(matmul_183, parameter_22) - del parameter_22 + del matmul_183, parameter_22 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_138, dropout_139 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_207, None, full_4, False, "upscale_in_train", 0, False + add_207, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -4136,6 +4105,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_208 = paddle._C_ops.add(layer_norm_135, dropout_138) + del dropout_138, layer_norm_135 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_138, layer_norm_139, layer_norm_140 = (lambda x, f: f(x))( @@ -4144,7 +4114,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_18, parameter_19 + del add_208, parameter_18, parameter_19 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_184 = paddle._C_ops.matmul(layer_norm_138, parameter_17, False, False) @@ -4152,10 +4122,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_209 = paddle._C_ops.add(matmul_184, parameter_16) - del parameter_16 + del matmul_184, parameter_16 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_92 = paddle._C_ops.reshape(add_209, full_int_array_1) + del add_209 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_92 = paddle._C_ops.transpose(reshape_92, [0, 2, 1, 3]) @@ -4167,7 +4138,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_210 = paddle._C_ops.add(matmul_185, parameter_14) - del parameter_14 + del matmul_185, parameter_14 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_186 = paddle._C_ops.matmul(layer_norm_138, parameter_13, False, False) @@ -4175,10 +4146,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_211 = paddle._C_ops.add(matmul_186, parameter_12) - del parameter_12 + del matmul_186, parameter_12 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_93 = paddle._C_ops.reshape(add_210, full_int_array_1) + del add_210 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_93 = paddle._C_ops.transpose(reshape_93, [0, 2, 1, 3]) @@ -4186,7 +4158,7 @@ def forward( # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_94 = paddle._C_ops.reshape(add_211, full_int_array_1) - del full_int_array_1 + del add_211, full_int_array_1 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_94 = paddle._C_ops.transpose(reshape_94, [0, 2, 1, 3]) @@ -4194,13 +4166,15 @@ def forward( # pd_op.scale: (1x16x21x64xf32) <- (1x16x21x64xf32, 1xf32) scale_24 = paddle._C_ops.scale(transpose_92, full_5, float("0"), True) - del transpose_92 + del full_5, transpose_92 # pd_op.matmul: (1x16x21x21xf32) <- (1x16x21x64xf32, 1x16x21x64xf32) matmul_187 = paddle._C_ops.matmul(scale_24, transpose_93, False, True) + del scale_24, transpose_93 # pd_op.add: (1x16x21x21xf32) <- (1x16x21x21xf32, 1x1x1x21xf32) add_212 = paddle._C_ops.add(matmul_187, unsqueeze_0) + del matmul_187, unsqueeze_0 # pd_op.softmax: (1x16x21x21xf32) <- (1x16x21x21xf32) softmax_23 = paddle._C_ops.softmax(add_212, -1) @@ -4209,13 +4183,15 @@ def forward( # pd_op.dropout: (1x16x21x21xf32, 1x16x21x21xui8) <- (1x16x21x21xf32, None, 1xf32) dropout_140, dropout_141 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_23, None, full_4, False, "upscale_in_train", 0, False + softmax_23, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_23 # pd_op.matmul: (1x16x21x64xf32) <- (1x16x21x21xf32, 1x16x21x64xf32) matmul_188 = paddle._C_ops.matmul(dropout_140, transpose_94, False, False) + del dropout_140, transpose_94 # pd_op.transpose: (1x21x16x64xf32) <- (1x16x21x64xf32) transpose_95 = paddle._C_ops.transpose(matmul_188, [0, 2, 1, 3]) @@ -4223,20 +4199,20 @@ def forward( # pd_op.reshape: (1x21x1024xf32) <- (1x21x16x64xf32, 3xi64) reshape_95 = paddle._C_ops.reshape(transpose_95, full_int_array_2) - del full_int_array_2 + del full_int_array_2, transpose_95 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_189 = paddle._C_ops.matmul(reshape_95, parameter_11, False, False) - del parameter_11 + del parameter_11, reshape_95 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_213 = paddle._C_ops.add(matmul_189, parameter_10) - del parameter_10 + del matmul_189, parameter_10 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_142, dropout_143 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_213, None, full_4, False, "upscale_in_train", 0, False + add_213, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -4244,6 +4220,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_214 = paddle._C_ops.add(layer_norm_138, dropout_142) + del dropout_142, layer_norm_138 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_141, layer_norm_142, layer_norm_143 = (lambda x, f: f(x))( @@ -4252,7 +4229,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_4, parameter_5 + del add_214, parameter_4, parameter_5 # pd_op.matmul: (1x21x4096xf32) <- (1x21x1024xf32, 1024x4096xf32) matmul_190 = paddle._C_ops.matmul(layer_norm_141, parameter_9, False, False) @@ -4260,30 +4237,32 @@ def forward( # pd_op.add: (1x21x4096xf32) <- (1x21x4096xf32, 4096xf32) add_215 = paddle._C_ops.add(matmul_190, parameter_8) - del parameter_8 + del matmul_190, parameter_8 # pd_op.gelu: (1x21x4096xf32) <- (1x21x4096xf32) gelu_23 = paddle._C_ops.gelu(add_215, False) + del add_215 # pd_op.matmul: (1x21x1024xf32) <- (1x21x4096xf32, 4096x1024xf32) matmul_191 = paddle._C_ops.matmul(gelu_23, parameter_7, False, False) - del parameter_7 + del gelu_23, parameter_7 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_216 = paddle._C_ops.add(matmul_191, parameter_6) - del parameter_6 + del matmul_191, parameter_6 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_144, dropout_145 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_216, None, full_4, False, "upscale_in_train", 0, False + add_216, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) - del add_216 + del add_216, full_4 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_217 = paddle._C_ops.add(layer_norm_141, dropout_144) + del dropout_144, layer_norm_141 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_144, layer_norm_145, layer_norm_146 = (lambda x, f: f(x))( @@ -4292,7 +4271,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_2, parameter_3 + del add_217, parameter_2, parameter_3 # pd_op.full_int_array: (1xi64) <- () full_int_array_3 = [0] @@ -4304,899 +4283,18 @@ def forward( slice_0 = paddle._C_ops.slice( layer_norm_144, [1], full_int_array_3, full_int_array_4, [1], [1] ) + del full_int_array_3, full_int_array_4, layer_norm_144 # pd_op.matmul: (1x1024xf32) <- (1x1024xf32, 1024x1024xf32) matmul_192 = paddle._C_ops.matmul(slice_0, parameter_1, False, False) - del parameter_1 + del parameter_1, slice_0 # pd_op.add: (1x1024xf32) <- (1x1024xf32, 1024xf32) add_218 = paddle._C_ops.add(matmul_192, parameter_0) - del parameter_0 + del matmul_192, parameter_0 # pd_op.tanh: (1x1024xf32) <- (1x1024xf32) tanh_0 = paddle._C_ops.tanh(add_218) - del ( - add_0, - add_1, - add_10, - add_100, - add_101, - add_102, - add_103, - add_106, - add_107, - add_109, - add_11, - add_110, - add_111, - add_112, - add_115, - add_116, - add_118, - add_119, - add_12, - add_120, - add_121, - add_124, - add_125, - add_127, - add_128, - add_129, - add_13, - add_130, - add_133, - add_134, - add_136, - add_137, - add_138, - add_139, - add_142, - add_143, - add_145, - add_146, - add_147, - add_148, - add_151, - add_152, - add_154, - add_155, - add_156, - add_157, - add_16, - add_160, - add_161, - add_163, - add_164, - add_165, - add_166, - add_169, - add_17, - add_170, - add_172, - add_173, - add_174, - add_175, - add_178, - add_179, - add_181, - add_182, - add_183, - add_184, - add_187, - add_188, - add_19, - add_190, - add_191, - add_192, - add_193, - add_196, - add_197, - add_199, - add_2, - add_20, - add_200, - add_201, - add_202, - add_205, - add_206, - add_208, - add_209, - add_21, - add_210, - add_211, - add_214, - add_215, - add_217, - add_218, - add_22, - add_25, - add_26, - add_28, - add_29, - add_3, - add_30, - add_31, - add_34, - add_35, - add_37, - add_38, - add_39, - add_4, - add_40, - add_43, - add_44, - add_46, - add_47, - add_48, - add_49, - add_52, - add_53, - add_55, - add_56, - add_57, - add_58, - add_61, - add_62, - add_64, - add_65, - add_66, - add_67, - add_7, - add_70, - add_71, - add_73, - add_74, - add_75, - add_76, - add_79, - add_8, - add_80, - add_82, - add_83, - add_84, - add_85, - add_88, - add_89, - add_91, - add_92, - add_93, - add_94, - add_97, - add_98, - assign_0, - assign_1, - assign_10, - assign_11, - assign_12, - assign_13, - assign_14, - assign_15, - assign_16, - assign_17, - assign_18, - assign_19, - assign_2, - assign_20, - assign_21, - assign_22, - assign_23, - assign_24, - assign_25, - assign_26, - assign_27, - assign_28, - assign_29, - assign_3, - assign_30, - assign_31, - assign_32, - assign_33, - assign_34, - assign_35, - assign_36, - assign_37, - assign_38, - assign_39, - assign_4, - assign_40, - assign_41, - assign_42, - assign_43, - assign_44, - assign_45, - assign_46, - assign_47, - assign_48, - assign_49, - assign_5, - assign_50, - assign_51, - assign_52, - assign_53, - assign_54, - assign_55, - assign_56, - assign_57, - assign_58, - assign_59, - assign_6, - assign_60, - assign_61, - assign_62, - assign_63, - assign_64, - assign_65, - assign_66, - assign_67, - assign_68, - assign_69, - assign_7, - assign_70, - assign_71, - assign_72, - assign_73, - assign_74, - assign_75, - assign_76, - assign_77, - assign_78, - assign_79, - assign_8, - assign_80, - assign_81, - assign_82, - assign_83, - assign_84, - assign_85, - assign_86, - assign_87, - assign_88, - assign_89, - assign_9, - assign_90, - assign_91, - assign_92, - assign_93, - assign_94, - dropout_0, - dropout_1, - dropout_10, - dropout_100, - dropout_101, - dropout_102, - dropout_103, - dropout_104, - dropout_105, - dropout_106, - dropout_107, - dropout_108, - dropout_109, - dropout_11, - dropout_110, - dropout_111, - dropout_112, - dropout_113, - dropout_114, - dropout_115, - dropout_116, - dropout_117, - dropout_118, - dropout_119, - dropout_12, - dropout_120, - dropout_121, - dropout_122, - dropout_123, - dropout_124, - dropout_125, - dropout_126, - dropout_127, - dropout_128, - dropout_129, - dropout_13, - dropout_130, - dropout_131, - dropout_132, - dropout_133, - dropout_134, - dropout_135, - dropout_136, - dropout_137, - dropout_138, - dropout_139, - dropout_14, - dropout_140, - dropout_141, - dropout_142, - dropout_143, - dropout_144, - dropout_145, - dropout_15, - dropout_16, - dropout_17, - dropout_18, - dropout_19, - dropout_2, - dropout_20, - dropout_21, - dropout_22, - dropout_23, - dropout_24, - dropout_25, - dropout_26, - dropout_27, - dropout_28, - dropout_29, - dropout_3, - dropout_30, - dropout_31, - dropout_32, - dropout_33, - dropout_34, - dropout_35, - dropout_36, - dropout_37, - dropout_38, - dropout_39, - dropout_4, - dropout_40, - dropout_41, - dropout_42, - dropout_43, - dropout_44, - dropout_45, - dropout_46, - dropout_47, - dropout_48, - dropout_49, - dropout_5, - dropout_50, - dropout_51, - dropout_52, - dropout_53, - dropout_54, - dropout_55, - dropout_56, - dropout_57, - dropout_58, - dropout_59, - dropout_6, - dropout_60, - dropout_61, - dropout_62, - dropout_63, - dropout_64, - dropout_65, - dropout_66, - dropout_67, - dropout_68, - dropout_69, - dropout_7, - dropout_70, - dropout_71, - dropout_72, - dropout_73, - dropout_74, - dropout_75, - dropout_76, - dropout_77, - dropout_78, - dropout_79, - dropout_8, - dropout_80, - dropout_81, - dropout_82, - dropout_83, - dropout_84, - dropout_85, - dropout_86, - dropout_87, - dropout_88, - dropout_89, - dropout_9, - dropout_90, - dropout_91, - dropout_92, - dropout_93, - dropout_94, - dropout_95, - dropout_96, - dropout_97, - dropout_98, - dropout_99, - embedding_0, - embedding_1, - embedding_2, - full_4, - full_5, - full_int_array_3, - full_int_array_4, - gelu_0, - gelu_1, - gelu_10, - gelu_11, - gelu_12, - gelu_13, - gelu_14, - gelu_15, - gelu_16, - gelu_17, - gelu_18, - gelu_19, - gelu_2, - gelu_20, - gelu_21, - gelu_22, - gelu_23, - gelu_3, - gelu_4, - gelu_5, - gelu_6, - gelu_7, - gelu_8, - gelu_9, - layer_norm_1, - layer_norm_10, - layer_norm_100, - layer_norm_101, - layer_norm_102, - layer_norm_103, - layer_norm_104, - layer_norm_105, - layer_norm_106, - layer_norm_107, - layer_norm_108, - layer_norm_109, - layer_norm_11, - layer_norm_110, - layer_norm_111, - layer_norm_112, - layer_norm_113, - layer_norm_114, - layer_norm_115, - layer_norm_116, - layer_norm_117, - layer_norm_118, - layer_norm_119, - layer_norm_12, - layer_norm_120, - layer_norm_121, - layer_norm_122, - layer_norm_123, - layer_norm_124, - layer_norm_125, - layer_norm_126, - layer_norm_127, - layer_norm_128, - layer_norm_129, - layer_norm_13, - layer_norm_130, - layer_norm_131, - layer_norm_132, - layer_norm_133, - layer_norm_134, - layer_norm_135, - layer_norm_136, - layer_norm_137, - layer_norm_138, - layer_norm_139, - layer_norm_14, - layer_norm_140, - layer_norm_141, - layer_norm_142, - layer_norm_143, - layer_norm_144, - layer_norm_145, - layer_norm_146, - layer_norm_15, - layer_norm_16, - layer_norm_17, - layer_norm_18, - layer_norm_19, - layer_norm_2, - layer_norm_20, - layer_norm_21, - layer_norm_22, - layer_norm_23, - layer_norm_24, - layer_norm_25, - layer_norm_26, - layer_norm_27, - layer_norm_28, - layer_norm_29, - layer_norm_3, - layer_norm_30, - layer_norm_31, - layer_norm_32, - layer_norm_33, - layer_norm_34, - layer_norm_35, - layer_norm_36, - layer_norm_37, - layer_norm_38, - layer_norm_39, - layer_norm_4, - layer_norm_40, - layer_norm_41, - layer_norm_42, - layer_norm_43, - layer_norm_44, - layer_norm_45, - layer_norm_46, - layer_norm_47, - layer_norm_48, - layer_norm_49, - layer_norm_5, - layer_norm_50, - layer_norm_51, - layer_norm_52, - layer_norm_53, - layer_norm_54, - layer_norm_55, - layer_norm_56, - layer_norm_57, - layer_norm_58, - layer_norm_59, - layer_norm_6, - layer_norm_60, - layer_norm_61, - layer_norm_62, - layer_norm_63, - layer_norm_64, - layer_norm_65, - layer_norm_66, - layer_norm_67, - layer_norm_68, - layer_norm_69, - layer_norm_7, - layer_norm_70, - layer_norm_71, - layer_norm_72, - layer_norm_73, - layer_norm_74, - layer_norm_75, - layer_norm_76, - layer_norm_77, - layer_norm_78, - layer_norm_79, - layer_norm_8, - layer_norm_80, - layer_norm_81, - layer_norm_82, - layer_norm_83, - layer_norm_84, - layer_norm_85, - layer_norm_86, - layer_norm_87, - layer_norm_88, - layer_norm_89, - layer_norm_9, - layer_norm_90, - layer_norm_91, - layer_norm_92, - layer_norm_93, - layer_norm_94, - layer_norm_95, - layer_norm_96, - layer_norm_97, - layer_norm_98, - layer_norm_99, - matmul_0, - matmul_1, - matmul_10, - matmul_101, - matmul_102, - matmul_103, - matmul_104, - matmul_105, - matmul_106, - matmul_107, - matmul_109, - matmul_11, - matmul_110, - matmul_111, - matmul_112, - matmul_113, - matmul_114, - matmul_115, - matmul_117, - matmul_118, - matmul_119, - matmul_120, - matmul_121, - matmul_122, - matmul_123, - matmul_125, - matmul_126, - matmul_127, - matmul_128, - matmul_129, - matmul_13, - matmul_130, - matmul_131, - matmul_133, - matmul_134, - matmul_135, - matmul_136, - matmul_137, - matmul_138, - matmul_139, - matmul_14, - matmul_141, - matmul_142, - matmul_143, - matmul_144, - matmul_145, - matmul_146, - matmul_147, - matmul_149, - matmul_15, - matmul_150, - matmul_151, - matmul_152, - matmul_153, - matmul_154, - matmul_155, - matmul_157, - matmul_158, - matmul_159, - matmul_16, - matmul_160, - matmul_161, - matmul_162, - matmul_163, - matmul_165, - matmul_166, - matmul_167, - matmul_168, - matmul_169, - matmul_17, - matmul_170, - matmul_171, - matmul_173, - matmul_174, - matmul_175, - matmul_176, - matmul_177, - matmul_178, - matmul_179, - matmul_18, - matmul_181, - matmul_182, - matmul_183, - matmul_184, - matmul_185, - matmul_186, - matmul_187, - matmul_189, - matmul_19, - matmul_190, - matmul_191, - matmul_192, - matmul_2, - matmul_21, - matmul_22, - matmul_23, - matmul_24, - matmul_25, - matmul_26, - matmul_27, - matmul_29, - matmul_3, - matmul_30, - matmul_31, - matmul_32, - matmul_33, - matmul_34, - matmul_35, - matmul_37, - matmul_38, - matmul_39, - matmul_40, - matmul_41, - matmul_42, - matmul_43, - matmul_45, - matmul_46, - matmul_47, - matmul_48, - matmul_49, - matmul_5, - matmul_50, - matmul_51, - matmul_53, - matmul_54, - matmul_55, - matmul_56, - matmul_57, - matmul_58, - matmul_59, - matmul_6, - matmul_61, - matmul_62, - matmul_63, - matmul_64, - matmul_65, - matmul_66, - matmul_67, - matmul_69, - matmul_7, - matmul_70, - matmul_71, - matmul_72, - matmul_73, - matmul_74, - matmul_75, - matmul_77, - matmul_78, - matmul_79, - matmul_8, - matmul_80, - matmul_81, - matmul_82, - matmul_83, - matmul_85, - matmul_86, - matmul_87, - matmul_88, - matmul_89, - matmul_9, - matmul_90, - matmul_91, - matmul_93, - matmul_94, - matmul_95, - matmul_96, - matmul_97, - matmul_98, - matmul_99, - reshape_11, - reshape_15, - reshape_19, - reshape_23, - reshape_27, - reshape_3, - reshape_31, - reshape_35, - reshape_39, - reshape_43, - reshape_47, - reshape_51, - reshape_55, - reshape_59, - reshape_63, - reshape_67, - reshape_7, - reshape_71, - reshape_75, - reshape_79, - reshape_83, - reshape_87, - reshape_91, - reshape_95, - scale_1, - scale_10, - scale_11, - scale_12, - scale_13, - scale_14, - scale_15, - scale_16, - scale_17, - scale_18, - scale_19, - scale_2, - scale_20, - scale_21, - scale_22, - scale_23, - scale_24, - scale_3, - scale_4, - scale_5, - scale_6, - scale_7, - scale_8, - scale_9, - slice_0, - softmax_0, - softmax_1, - softmax_10, - softmax_11, - softmax_12, - softmax_13, - softmax_14, - softmax_15, - softmax_16, - softmax_17, - softmax_18, - softmax_19, - softmax_2, - softmax_20, - softmax_21, - softmax_22, - softmax_23, - softmax_3, - softmax_4, - softmax_5, - softmax_6, - softmax_7, - softmax_8, - softmax_9, - subtract_0, - transpose_1, - transpose_10, - transpose_11, - transpose_13, - transpose_14, - transpose_15, - transpose_17, - transpose_18, - transpose_19, - transpose_2, - transpose_21, - transpose_22, - transpose_23, - transpose_25, - transpose_26, - transpose_27, - transpose_29, - transpose_3, - transpose_30, - transpose_31, - transpose_33, - transpose_34, - transpose_35, - transpose_37, - transpose_38, - transpose_39, - transpose_41, - transpose_42, - transpose_43, - transpose_45, - transpose_46, - transpose_47, - transpose_49, - transpose_5, - transpose_50, - transpose_51, - transpose_53, - transpose_54, - transpose_55, - transpose_57, - transpose_58, - transpose_59, - transpose_6, - transpose_61, - transpose_62, - transpose_63, - transpose_65, - transpose_66, - transpose_67, - transpose_69, - transpose_7, - transpose_70, - transpose_71, - transpose_73, - transpose_74, - transpose_75, - transpose_77, - transpose_78, - transpose_79, - transpose_81, - transpose_82, - transpose_83, - transpose_85, - transpose_86, - transpose_87, - transpose_89, - transpose_9, - transpose_90, - transpose_91, - transpose_93, - transpose_94, - transpose_95, - unsqueeze_0, - ) + del add_218 return tanh_0 diff --git a/paddle_samples/PaddleNLP/ernie-search-large-cross-encoder-marco-en/weight_meta.py b/paddle_samples/PaddleNLP/ernie-search-large-cross-encoder-marco-en/weight_meta.py index 0be003ce3..5c8619454 100644 --- a/paddle_samples/PaddleNLP/ernie-search-large-cross-encoder-marco-en/weight_meta.py +++ b/paddle_samples/PaddleNLP/ernie-search-large-cross-encoder-marco-en/weight_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_parameter_0: name = "parameter_0" + original_name = "linear_1530.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0980704") @@ -11,6 +12,7 @@ class Program_weight_tensor_parameter_0: class Program_weight_tensor_parameter_1: name = "parameter_1" + original_name = "linear_1530.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.204461") @@ -22,6 +24,7 @@ class Program_weight_tensor_parameter_1: class Program_weight_tensor_parameter_2: name = "parameter_2" + original_name = "layer_norm_526.b_0" shape = [1024] dtype = "float32" min_val = float("-0.56166") @@ -33,6 +36,7 @@ class Program_weight_tensor_parameter_2: class Program_weight_tensor_parameter_3: name = "parameter_3" + original_name = "layer_norm_526.w_0" shape = [1024] dtype = "float32" min_val = float("0.279228") @@ -44,6 +48,7 @@ class Program_weight_tensor_parameter_3: class Program_weight_tensor_parameter_4: name = "parameter_4" + original_name = "layer_norm_525.b_0" shape = [1024] dtype = "float32" min_val = float("-0.228217") @@ -55,6 +60,7 @@ class Program_weight_tensor_parameter_4: class Program_weight_tensor_parameter_5: name = "parameter_5" + original_name = "layer_norm_525.w_0" shape = [1024] dtype = "float32" min_val = float("0.755116") @@ -66,6 +72,7 @@ class Program_weight_tensor_parameter_5: class Program_weight_tensor_parameter_6: name = "parameter_6" + original_name = "linear_1529.b_0" shape = [1024] dtype = "float32" min_val = float("-0.187889") @@ -77,6 +84,7 @@ class Program_weight_tensor_parameter_6: class Program_weight_tensor_parameter_7: name = "parameter_7" + original_name = "linear_1529.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-0.849642") @@ -88,6 +96,7 @@ class Program_weight_tensor_parameter_7: class Program_weight_tensor_parameter_8: name = "parameter_8" + original_name = "linear_1528.b_0" shape = [4096] dtype = "float32" min_val = float("-0.371455") @@ -99,6 +108,7 @@ class Program_weight_tensor_parameter_8: class Program_weight_tensor_parameter_9: name = "parameter_9" + original_name = "linear_1528.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.205542") @@ -110,6 +120,7 @@ class Program_weight_tensor_parameter_9: class Program_weight_tensor_parameter_10: name = "parameter_10" + original_name = "linear_1527.b_0" shape = [1024] dtype = "float32" min_val = float("-0.155179") @@ -121,6 +132,7 @@ class Program_weight_tensor_parameter_10: class Program_weight_tensor_parameter_11: name = "parameter_11" + original_name = "linear_1527.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.179967") @@ -132,6 +144,7 @@ class Program_weight_tensor_parameter_11: class Program_weight_tensor_parameter_12: name = "parameter_12" + original_name = "linear_1526.b_0" shape = [1024] dtype = "float32" min_val = float("-0.109387") @@ -143,6 +156,7 @@ class Program_weight_tensor_parameter_12: class Program_weight_tensor_parameter_13: name = "parameter_13" + original_name = "linear_1526.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.206777") @@ -154,6 +168,7 @@ class Program_weight_tensor_parameter_13: class Program_weight_tensor_parameter_14: name = "parameter_14" + original_name = "linear_1525.b_0" shape = [1024] dtype = "float32" min_val = float("-5.89913") @@ -165,6 +180,7 @@ class Program_weight_tensor_parameter_14: class Program_weight_tensor_parameter_15: name = "parameter_15" + original_name = "linear_1525.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.525176") @@ -176,6 +192,7 @@ class Program_weight_tensor_parameter_15: class Program_weight_tensor_parameter_16: name = "parameter_16" + original_name = "linear_1524.b_0" shape = [1024] dtype = "float32" min_val = float("-0.59771") @@ -187,6 +204,7 @@ class Program_weight_tensor_parameter_16: class Program_weight_tensor_parameter_17: name = "parameter_17" + original_name = "linear_1524.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.340598") @@ -198,6 +216,7 @@ class Program_weight_tensor_parameter_17: class Program_weight_tensor_parameter_18: name = "parameter_18" + original_name = "layer_norm_524.b_0" shape = [1024] dtype = "float32" min_val = float("-0.114866") @@ -209,6 +228,7 @@ class Program_weight_tensor_parameter_18: class Program_weight_tensor_parameter_19: name = "parameter_19" + original_name = "layer_norm_524.w_0" shape = [1024] dtype = "float32" min_val = float("0.050755") @@ -220,6 +240,7 @@ class Program_weight_tensor_parameter_19: class Program_weight_tensor_parameter_20: name = "parameter_20" + original_name = "layer_norm_523.b_0" shape = [1024] dtype = "float32" min_val = float("-0.394139") @@ -231,6 +252,7 @@ class Program_weight_tensor_parameter_20: class Program_weight_tensor_parameter_21: name = "parameter_21" + original_name = "layer_norm_523.w_0" shape = [1024] dtype = "float32" min_val = float("0.691805") @@ -242,6 +264,7 @@ class Program_weight_tensor_parameter_21: class Program_weight_tensor_parameter_22: name = "parameter_22" + original_name = "linear_1523.b_0" shape = [1024] dtype = "float32" min_val = float("-0.671355") @@ -253,6 +276,7 @@ class Program_weight_tensor_parameter_22: class Program_weight_tensor_parameter_23: name = "parameter_23" + original_name = "linear_1523.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-5.6009") @@ -264,6 +288,7 @@ class Program_weight_tensor_parameter_23: class Program_weight_tensor_parameter_24: name = "parameter_24" + original_name = "linear_1522.b_0" shape = [4096] dtype = "float32" min_val = float("-0.273587") @@ -275,6 +300,7 @@ class Program_weight_tensor_parameter_24: class Program_weight_tensor_parameter_25: name = "parameter_25" + original_name = "linear_1522.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.91336") @@ -286,6 +312,7 @@ class Program_weight_tensor_parameter_25: class Program_weight_tensor_parameter_26: name = "parameter_26" + original_name = "linear_1521.b_0" shape = [1024] dtype = "float32" min_val = float("-0.226336") @@ -297,6 +324,7 @@ class Program_weight_tensor_parameter_26: class Program_weight_tensor_parameter_27: name = "parameter_27" + original_name = "linear_1521.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.187885") @@ -308,6 +336,7 @@ class Program_weight_tensor_parameter_27: class Program_weight_tensor_parameter_28: name = "parameter_28" + original_name = "linear_1520.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0828999") @@ -319,6 +348,7 @@ class Program_weight_tensor_parameter_28: class Program_weight_tensor_parameter_29: name = "parameter_29" + original_name = "linear_1520.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.212557") @@ -330,6 +360,7 @@ class Program_weight_tensor_parameter_29: class Program_weight_tensor_parameter_30: name = "parameter_30" + original_name = "linear_1519.b_0" shape = [1024] dtype = "float32" min_val = float("-6.07655") @@ -341,6 +372,7 @@ class Program_weight_tensor_parameter_30: class Program_weight_tensor_parameter_31: name = "parameter_31" + original_name = "linear_1519.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.258105") @@ -352,6 +384,7 @@ class Program_weight_tensor_parameter_31: class Program_weight_tensor_parameter_32: name = "parameter_32" + original_name = "linear_1518.b_0" shape = [1024] dtype = "float32" min_val = float("-0.534398") @@ -363,6 +396,7 @@ class Program_weight_tensor_parameter_32: class Program_weight_tensor_parameter_33: name = "parameter_33" + original_name = "linear_1518.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.394386") @@ -374,6 +408,7 @@ class Program_weight_tensor_parameter_33: class Program_weight_tensor_parameter_34: name = "parameter_34" + original_name = "layer_norm_522.b_0" shape = [1024] dtype = "float32" min_val = float("-0.394458") @@ -385,6 +420,7 @@ class Program_weight_tensor_parameter_34: class Program_weight_tensor_parameter_35: name = "parameter_35" + original_name = "layer_norm_522.w_0" shape = [1024] dtype = "float32" min_val = float("0.0692773") @@ -396,6 +432,7 @@ class Program_weight_tensor_parameter_35: class Program_weight_tensor_parameter_36: name = "parameter_36" + original_name = "layer_norm_521.b_0" shape = [1024] dtype = "float32" min_val = float("-0.508843") @@ -407,6 +444,7 @@ class Program_weight_tensor_parameter_36: class Program_weight_tensor_parameter_37: name = "parameter_37" + original_name = "layer_norm_521.w_0" shape = [1024] dtype = "float32" min_val = float("0.633432") @@ -418,6 +456,7 @@ class Program_weight_tensor_parameter_37: class Program_weight_tensor_parameter_38: name = "parameter_38" + original_name = "linear_1517.b_0" shape = [1024] dtype = "float32" min_val = float("-0.552495") @@ -429,6 +468,7 @@ class Program_weight_tensor_parameter_38: class Program_weight_tensor_parameter_39: name = "parameter_39" + original_name = "linear_1517.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-6.76227") @@ -440,6 +480,7 @@ class Program_weight_tensor_parameter_39: class Program_weight_tensor_parameter_40: name = "parameter_40" + original_name = "linear_1516.b_0" shape = [4096] dtype = "float32" min_val = float("-0.163101") @@ -451,6 +492,7 @@ class Program_weight_tensor_parameter_40: class Program_weight_tensor_parameter_41: name = "parameter_41" + original_name = "linear_1516.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-1.51378") @@ -462,6 +504,7 @@ class Program_weight_tensor_parameter_41: class Program_weight_tensor_parameter_42: name = "parameter_42" + original_name = "linear_1515.b_0" shape = [1024] dtype = "float32" min_val = float("-0.281544") @@ -473,6 +516,7 @@ class Program_weight_tensor_parameter_42: class Program_weight_tensor_parameter_43: name = "parameter_43" + original_name = "linear_1515.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.188932") @@ -484,6 +528,7 @@ class Program_weight_tensor_parameter_43: class Program_weight_tensor_parameter_44: name = "parameter_44" + original_name = "linear_1514.b_0" shape = [1024] dtype = "float32" min_val = float("-0.252618") @@ -495,6 +540,7 @@ class Program_weight_tensor_parameter_44: class Program_weight_tensor_parameter_45: name = "parameter_45" + original_name = "linear_1514.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.201451") @@ -506,6 +552,7 @@ class Program_weight_tensor_parameter_45: class Program_weight_tensor_parameter_46: name = "parameter_46" + original_name = "linear_1513.b_0" shape = [1024] dtype = "float32" min_val = float("-6.32712") @@ -517,6 +564,7 @@ class Program_weight_tensor_parameter_46: class Program_weight_tensor_parameter_47: name = "parameter_47" + original_name = "linear_1513.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.431391") @@ -528,6 +576,7 @@ class Program_weight_tensor_parameter_47: class Program_weight_tensor_parameter_48: name = "parameter_48" + original_name = "linear_1512.b_0" shape = [1024] dtype = "float32" min_val = float("-0.734997") @@ -539,6 +588,7 @@ class Program_weight_tensor_parameter_48: class Program_weight_tensor_parameter_49: name = "parameter_49" + original_name = "linear_1512.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.30246") @@ -550,6 +600,7 @@ class Program_weight_tensor_parameter_49: class Program_weight_tensor_parameter_50: name = "parameter_50" + original_name = "layer_norm_520.b_0" shape = [1024] dtype = "float32" min_val = float("-0.185066") @@ -561,6 +612,7 @@ class Program_weight_tensor_parameter_50: class Program_weight_tensor_parameter_51: name = "parameter_51" + original_name = "layer_norm_520.w_0" shape = [1024] dtype = "float32" min_val = float("0.0383176") @@ -572,6 +624,7 @@ class Program_weight_tensor_parameter_51: class Program_weight_tensor_parameter_52: name = "parameter_52" + original_name = "layer_norm_519.b_0" shape = [1024] dtype = "float32" min_val = float("-1.18979") @@ -583,6 +636,7 @@ class Program_weight_tensor_parameter_52: class Program_weight_tensor_parameter_53: name = "parameter_53" + original_name = "layer_norm_519.w_0" shape = [1024] dtype = "float32" min_val = float("0.584633") @@ -594,6 +648,7 @@ class Program_weight_tensor_parameter_53: class Program_weight_tensor_parameter_54: name = "parameter_54" + original_name = "linear_1511.b_0" shape = [1024] dtype = "float32" min_val = float("-0.691968") @@ -605,6 +660,7 @@ class Program_weight_tensor_parameter_54: class Program_weight_tensor_parameter_55: name = "parameter_55" + original_name = "linear_1511.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-5.03474") @@ -616,6 +672,7 @@ class Program_weight_tensor_parameter_55: class Program_weight_tensor_parameter_56: name = "parameter_56" + original_name = "linear_1510.b_0" shape = [4096] dtype = "float32" min_val = float("-0.217195") @@ -627,6 +684,7 @@ class Program_weight_tensor_parameter_56: class Program_weight_tensor_parameter_57: name = "parameter_57" + original_name = "linear_1510.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.44622") @@ -638,6 +696,7 @@ class Program_weight_tensor_parameter_57: class Program_weight_tensor_parameter_58: name = "parameter_58" + original_name = "linear_1509.b_0" shape = [1024] dtype = "float32" min_val = float("-0.328422") @@ -649,6 +708,7 @@ class Program_weight_tensor_parameter_58: class Program_weight_tensor_parameter_59: name = "parameter_59" + original_name = "linear_1509.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.203507") @@ -660,6 +720,7 @@ class Program_weight_tensor_parameter_59: class Program_weight_tensor_parameter_60: name = "parameter_60" + original_name = "linear_1508.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0924463") @@ -671,6 +732,7 @@ class Program_weight_tensor_parameter_60: class Program_weight_tensor_parameter_61: name = "parameter_61" + original_name = "linear_1508.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.211383") @@ -682,6 +744,7 @@ class Program_weight_tensor_parameter_61: class Program_weight_tensor_parameter_62: name = "parameter_62" + original_name = "linear_1507.b_0" shape = [1024] dtype = "float32" min_val = float("-6.67368") @@ -693,6 +756,7 @@ class Program_weight_tensor_parameter_62: class Program_weight_tensor_parameter_63: name = "parameter_63" + original_name = "linear_1507.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-1.95272") @@ -704,6 +768,7 @@ class Program_weight_tensor_parameter_63: class Program_weight_tensor_parameter_64: name = "parameter_64" + original_name = "linear_1506.b_0" shape = [1024] dtype = "float32" min_val = float("-0.743046") @@ -715,6 +780,7 @@ class Program_weight_tensor_parameter_64: class Program_weight_tensor_parameter_65: name = "parameter_65" + original_name = "linear_1506.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.373103") @@ -726,6 +792,7 @@ class Program_weight_tensor_parameter_65: class Program_weight_tensor_parameter_66: name = "parameter_66" + original_name = "layer_norm_518.b_0" shape = [1024] dtype = "float32" min_val = float("-0.687953") @@ -737,6 +804,7 @@ class Program_weight_tensor_parameter_66: class Program_weight_tensor_parameter_67: name = "parameter_67" + original_name = "layer_norm_518.w_0" shape = [1024] dtype = "float32" min_val = float("0.0102745") @@ -748,6 +816,7 @@ class Program_weight_tensor_parameter_67: class Program_weight_tensor_parameter_68: name = "parameter_68" + original_name = "layer_norm_517.b_0" shape = [1024] dtype = "float32" min_val = float("-1.29401") @@ -759,6 +828,7 @@ class Program_weight_tensor_parameter_68: class Program_weight_tensor_parameter_69: name = "parameter_69" + original_name = "layer_norm_517.w_0" shape = [1024] dtype = "float32" min_val = float("0.661984") @@ -770,6 +840,7 @@ class Program_weight_tensor_parameter_69: class Program_weight_tensor_parameter_70: name = "parameter_70" + original_name = "linear_1505.b_0" shape = [1024] dtype = "float32" min_val = float("-0.378371") @@ -781,6 +852,7 @@ class Program_weight_tensor_parameter_70: class Program_weight_tensor_parameter_71: name = "parameter_71" + original_name = "linear_1505.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-5.97201") @@ -792,6 +864,7 @@ class Program_weight_tensor_parameter_71: class Program_weight_tensor_parameter_72: name = "parameter_72" + original_name = "linear_1504.b_0" shape = [4096] dtype = "float32" min_val = float("-0.299378") @@ -803,6 +876,7 @@ class Program_weight_tensor_parameter_72: class Program_weight_tensor_parameter_73: name = "parameter_73" + original_name = "linear_1504.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.591248") @@ -814,6 +888,7 @@ class Program_weight_tensor_parameter_73: class Program_weight_tensor_parameter_74: name = "parameter_74" + original_name = "linear_1503.b_0" shape = [1024] dtype = "float32" min_val = float("-0.268955") @@ -825,6 +900,7 @@ class Program_weight_tensor_parameter_74: class Program_weight_tensor_parameter_75: name = "parameter_75" + original_name = "linear_1503.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.163577") @@ -836,6 +912,7 @@ class Program_weight_tensor_parameter_75: class Program_weight_tensor_parameter_76: name = "parameter_76" + original_name = "linear_1502.b_0" shape = [1024] dtype = "float32" min_val = float("-0.147903") @@ -847,6 +924,7 @@ class Program_weight_tensor_parameter_76: class Program_weight_tensor_parameter_77: name = "parameter_77" + original_name = "linear_1502.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.166745") @@ -858,6 +936,7 @@ class Program_weight_tensor_parameter_77: class Program_weight_tensor_parameter_78: name = "parameter_78" + original_name = "linear_1501.b_0" shape = [1024] dtype = "float32" min_val = float("-3.94946") @@ -869,6 +948,7 @@ class Program_weight_tensor_parameter_78: class Program_weight_tensor_parameter_79: name = "parameter_79" + original_name = "linear_1501.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.312305") @@ -880,6 +960,7 @@ class Program_weight_tensor_parameter_79: class Program_weight_tensor_parameter_80: name = "parameter_80" + original_name = "linear_1500.b_0" shape = [1024] dtype = "float32" min_val = float("-0.534652") @@ -891,6 +972,7 @@ class Program_weight_tensor_parameter_80: class Program_weight_tensor_parameter_81: name = "parameter_81" + original_name = "linear_1500.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.300548") @@ -902,6 +984,7 @@ class Program_weight_tensor_parameter_81: class Program_weight_tensor_parameter_82: name = "parameter_82" + original_name = "layer_norm_516.b_0" shape = [1024] dtype = "float32" min_val = float("-0.383991") @@ -913,6 +996,7 @@ class Program_weight_tensor_parameter_82: class Program_weight_tensor_parameter_83: name = "parameter_83" + original_name = "layer_norm_516.w_0" shape = [1024] dtype = "float32" min_val = float("0.204091") @@ -924,6 +1008,7 @@ class Program_weight_tensor_parameter_83: class Program_weight_tensor_parameter_84: name = "parameter_84" + original_name = "layer_norm_515.b_0" shape = [1024] dtype = "float32" min_val = float("-1.78976") @@ -935,6 +1020,7 @@ class Program_weight_tensor_parameter_84: class Program_weight_tensor_parameter_85: name = "parameter_85" + original_name = "layer_norm_515.w_0" shape = [1024] dtype = "float32" min_val = float("0.660906") @@ -946,6 +1032,7 @@ class Program_weight_tensor_parameter_85: class Program_weight_tensor_parameter_86: name = "parameter_86" + original_name = "linear_1499.b_0" shape = [1024] dtype = "float32" min_val = float("-0.244138") @@ -957,6 +1044,7 @@ class Program_weight_tensor_parameter_86: class Program_weight_tensor_parameter_87: name = "parameter_87" + original_name = "linear_1499.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-5.30464") @@ -968,6 +1056,7 @@ class Program_weight_tensor_parameter_87: class Program_weight_tensor_parameter_88: name = "parameter_88" + original_name = "linear_1498.b_0" shape = [4096] dtype = "float32" min_val = float("-0.297804") @@ -979,6 +1068,7 @@ class Program_weight_tensor_parameter_88: class Program_weight_tensor_parameter_89: name = "parameter_89" + original_name = "linear_1498.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-1.02905") @@ -990,6 +1080,7 @@ class Program_weight_tensor_parameter_89: class Program_weight_tensor_parameter_90: name = "parameter_90" + original_name = "linear_1497.b_0" shape = [1024] dtype = "float32" min_val = float("-0.16034") @@ -1001,6 +1092,7 @@ class Program_weight_tensor_parameter_90: class Program_weight_tensor_parameter_91: name = "parameter_91" + original_name = "linear_1497.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.153524") @@ -1012,6 +1104,7 @@ class Program_weight_tensor_parameter_91: class Program_weight_tensor_parameter_92: name = "parameter_92" + original_name = "linear_1496.b_0" shape = [1024] dtype = "float32" min_val = float("-0.178181") @@ -1023,6 +1116,7 @@ class Program_weight_tensor_parameter_92: class Program_weight_tensor_parameter_93: name = "parameter_93" + original_name = "linear_1496.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.196658") @@ -1034,6 +1128,7 @@ class Program_weight_tensor_parameter_93: class Program_weight_tensor_parameter_94: name = "parameter_94" + original_name = "linear_1495.b_0" shape = [1024] dtype = "float32" min_val = float("-5.66793") @@ -1045,6 +1140,7 @@ class Program_weight_tensor_parameter_94: class Program_weight_tensor_parameter_95: name = "parameter_95" + original_name = "linear_1495.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.28802") @@ -1056,6 +1152,7 @@ class Program_weight_tensor_parameter_95: class Program_weight_tensor_parameter_96: name = "parameter_96" + original_name = "linear_1494.b_0" shape = [1024] dtype = "float32" min_val = float("-0.476285") @@ -1067,6 +1164,7 @@ class Program_weight_tensor_parameter_96: class Program_weight_tensor_parameter_97: name = "parameter_97" + original_name = "linear_1494.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.35282") @@ -1078,6 +1176,7 @@ class Program_weight_tensor_parameter_97: class Program_weight_tensor_parameter_98: name = "parameter_98" + original_name = "layer_norm_514.b_0" shape = [1024] dtype = "float32" min_val = float("-0.619333") @@ -1089,6 +1188,7 @@ class Program_weight_tensor_parameter_98: class Program_weight_tensor_parameter_99: name = "parameter_99" + original_name = "layer_norm_514.w_0" shape = [1024] dtype = "float32" min_val = float("0.271188") @@ -1100,6 +1200,7 @@ class Program_weight_tensor_parameter_99: class Program_weight_tensor_parameter_100: name = "parameter_100" + original_name = "layer_norm_513.b_0" shape = [1024] dtype = "float32" min_val = float("-1.66409") @@ -1111,6 +1212,7 @@ class Program_weight_tensor_parameter_100: class Program_weight_tensor_parameter_101: name = "parameter_101" + original_name = "layer_norm_513.w_0" shape = [1024] dtype = "float32" min_val = float("0.678519") @@ -1122,6 +1224,7 @@ class Program_weight_tensor_parameter_101: class Program_weight_tensor_parameter_102: name = "parameter_102" + original_name = "linear_1493.b_0" shape = [1024] dtype = "float32" min_val = float("-0.660498") @@ -1133,6 +1236,7 @@ class Program_weight_tensor_parameter_102: class Program_weight_tensor_parameter_103: name = "parameter_103" + original_name = "linear_1493.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-4.82978") @@ -1144,6 +1248,7 @@ class Program_weight_tensor_parameter_103: class Program_weight_tensor_parameter_104: name = "parameter_104" + original_name = "linear_1492.b_0" shape = [4096] dtype = "float32" min_val = float("-0.283544") @@ -1155,6 +1260,7 @@ class Program_weight_tensor_parameter_104: class Program_weight_tensor_parameter_105: name = "parameter_105" + original_name = "linear_1492.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.539925") @@ -1166,6 +1272,7 @@ class Program_weight_tensor_parameter_105: class Program_weight_tensor_parameter_106: name = "parameter_106" + original_name = "linear_1491.b_0" shape = [1024] dtype = "float32" min_val = float("-0.129062") @@ -1177,6 +1284,7 @@ class Program_weight_tensor_parameter_106: class Program_weight_tensor_parameter_107: name = "parameter_107" + original_name = "linear_1491.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.177063") @@ -1188,6 +1296,7 @@ class Program_weight_tensor_parameter_107: class Program_weight_tensor_parameter_108: name = "parameter_108" + original_name = "linear_1490.b_0" shape = [1024] dtype = "float32" min_val = float("-0.223686") @@ -1199,6 +1308,7 @@ class Program_weight_tensor_parameter_108: class Program_weight_tensor_parameter_109: name = "parameter_109" + original_name = "linear_1490.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.17397") @@ -1210,6 +1320,7 @@ class Program_weight_tensor_parameter_109: class Program_weight_tensor_parameter_110: name = "parameter_110" + original_name = "linear_1489.b_0" shape = [1024] dtype = "float32" min_val = float("-1.68359") @@ -1221,6 +1332,7 @@ class Program_weight_tensor_parameter_110: class Program_weight_tensor_parameter_111: name = "parameter_111" + original_name = "linear_1489.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.279169") @@ -1232,6 +1344,7 @@ class Program_weight_tensor_parameter_111: class Program_weight_tensor_parameter_112: name = "parameter_112" + original_name = "linear_1488.b_0" shape = [1024] dtype = "float32" min_val = float("-0.450659") @@ -1243,6 +1356,7 @@ class Program_weight_tensor_parameter_112: class Program_weight_tensor_parameter_113: name = "parameter_113" + original_name = "linear_1488.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.245051") @@ -1254,6 +1368,7 @@ class Program_weight_tensor_parameter_113: class Program_weight_tensor_parameter_114: name = "parameter_114" + original_name = "layer_norm_512.b_0" shape = [1024] dtype = "float32" min_val = float("-0.553572") @@ -1265,6 +1380,7 @@ class Program_weight_tensor_parameter_114: class Program_weight_tensor_parameter_115: name = "parameter_115" + original_name = "layer_norm_512.w_0" shape = [1024] dtype = "float32" min_val = float("0.431506") @@ -1276,6 +1392,7 @@ class Program_weight_tensor_parameter_115: class Program_weight_tensor_parameter_116: name = "parameter_116" + original_name = "layer_norm_511.b_0" shape = [1024] dtype = "float32" min_val = float("-1.66846") @@ -1287,6 +1404,7 @@ class Program_weight_tensor_parameter_116: class Program_weight_tensor_parameter_117: name = "parameter_117" + original_name = "layer_norm_511.w_0" shape = [1024] dtype = "float32" min_val = float("0.650845") @@ -1298,6 +1416,7 @@ class Program_weight_tensor_parameter_117: class Program_weight_tensor_parameter_118: name = "parameter_118" + original_name = "linear_1487.b_0" shape = [1024] dtype = "float32" min_val = float("-0.918141") @@ -1309,6 +1428,7 @@ class Program_weight_tensor_parameter_118: class Program_weight_tensor_parameter_119: name = "parameter_119" + original_name = "linear_1487.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-0.833011") @@ -1320,6 +1440,7 @@ class Program_weight_tensor_parameter_119: class Program_weight_tensor_parameter_120: name = "parameter_120" + original_name = "linear_1486.b_0" shape = [4096] dtype = "float32" min_val = float("-0.293733") @@ -1331,6 +1452,7 @@ class Program_weight_tensor_parameter_120: class Program_weight_tensor_parameter_121: name = "parameter_121" + original_name = "linear_1486.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.199468") @@ -1342,6 +1464,7 @@ class Program_weight_tensor_parameter_121: class Program_weight_tensor_parameter_122: name = "parameter_122" + original_name = "linear_1485.b_0" shape = [1024] dtype = "float32" min_val = float("-0.126471") @@ -1353,6 +1476,7 @@ class Program_weight_tensor_parameter_122: class Program_weight_tensor_parameter_123: name = "parameter_123" + original_name = "linear_1485.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.218868") @@ -1364,6 +1488,7 @@ class Program_weight_tensor_parameter_123: class Program_weight_tensor_parameter_124: name = "parameter_124" + original_name = "linear_1484.b_0" shape = [1024] dtype = "float32" min_val = float("-0.130019") @@ -1375,6 +1500,7 @@ class Program_weight_tensor_parameter_124: class Program_weight_tensor_parameter_125: name = "parameter_125" + original_name = "linear_1484.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.160344") @@ -1386,6 +1512,7 @@ class Program_weight_tensor_parameter_125: class Program_weight_tensor_parameter_126: name = "parameter_126" + original_name = "linear_1483.b_0" shape = [1024] dtype = "float32" min_val = float("-1.82079") @@ -1397,6 +1524,7 @@ class Program_weight_tensor_parameter_126: class Program_weight_tensor_parameter_127: name = "parameter_127" + original_name = "linear_1483.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.293446") @@ -1408,6 +1536,7 @@ class Program_weight_tensor_parameter_127: class Program_weight_tensor_parameter_128: name = "parameter_128" + original_name = "linear_1482.b_0" shape = [1024] dtype = "float32" min_val = float("-0.554856") @@ -1419,6 +1548,7 @@ class Program_weight_tensor_parameter_128: class Program_weight_tensor_parameter_129: name = "parameter_129" + original_name = "linear_1482.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.212535") @@ -1430,6 +1560,7 @@ class Program_weight_tensor_parameter_129: class Program_weight_tensor_parameter_130: name = "parameter_130" + original_name = "layer_norm_510.b_0" shape = [1024] dtype = "float32" min_val = float("-0.896206") @@ -1441,6 +1572,7 @@ class Program_weight_tensor_parameter_130: class Program_weight_tensor_parameter_131: name = "parameter_131" + original_name = "layer_norm_510.w_0" shape = [1024] dtype = "float32" min_val = float("0.375429") @@ -1452,6 +1584,7 @@ class Program_weight_tensor_parameter_131: class Program_weight_tensor_parameter_132: name = "parameter_132" + original_name = "layer_norm_509.b_0" shape = [1024] dtype = "float32" min_val = float("-2.00872") @@ -1463,6 +1596,7 @@ class Program_weight_tensor_parameter_132: class Program_weight_tensor_parameter_133: name = "parameter_133" + original_name = "layer_norm_509.w_0" shape = [1024] dtype = "float32" min_val = float("0.649291") @@ -1474,6 +1608,7 @@ class Program_weight_tensor_parameter_133: class Program_weight_tensor_parameter_134: name = "parameter_134" + original_name = "linear_1481.b_0" shape = [1024] dtype = "float32" min_val = float("-0.841077") @@ -1485,6 +1620,7 @@ class Program_weight_tensor_parameter_134: class Program_weight_tensor_parameter_135: name = "parameter_135" + original_name = "linear_1481.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-2.43012") @@ -1496,6 +1632,7 @@ class Program_weight_tensor_parameter_135: class Program_weight_tensor_parameter_136: name = "parameter_136" + original_name = "linear_1480.b_0" shape = [4096] dtype = "float32" min_val = float("-0.299859") @@ -1507,6 +1644,7 @@ class Program_weight_tensor_parameter_136: class Program_weight_tensor_parameter_137: name = "parameter_137" + original_name = "linear_1480.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.237612") @@ -1518,6 +1656,7 @@ class Program_weight_tensor_parameter_137: class Program_weight_tensor_parameter_138: name = "parameter_138" + original_name = "linear_1479.b_0" shape = [1024] dtype = "float32" min_val = float("-0.131535") @@ -1529,6 +1668,7 @@ class Program_weight_tensor_parameter_138: class Program_weight_tensor_parameter_139: name = "parameter_139" + original_name = "linear_1479.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.514294") @@ -1540,6 +1680,7 @@ class Program_weight_tensor_parameter_139: class Program_weight_tensor_parameter_140: name = "parameter_140" + original_name = "linear_1478.b_0" shape = [1024] dtype = "float32" min_val = float("-0.277145") @@ -1551,6 +1692,7 @@ class Program_weight_tensor_parameter_140: class Program_weight_tensor_parameter_141: name = "parameter_141" + original_name = "linear_1478.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.18196") @@ -1562,6 +1704,7 @@ class Program_weight_tensor_parameter_141: class Program_weight_tensor_parameter_142: name = "parameter_142" + original_name = "linear_1477.b_0" shape = [1024] dtype = "float32" min_val = float("-1.9401") @@ -1573,6 +1716,7 @@ class Program_weight_tensor_parameter_142: class Program_weight_tensor_parameter_143: name = "parameter_143" + original_name = "linear_1477.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.281743") @@ -1584,6 +1728,7 @@ class Program_weight_tensor_parameter_143: class Program_weight_tensor_parameter_144: name = "parameter_144" + original_name = "linear_1476.b_0" shape = [1024] dtype = "float32" min_val = float("-0.477433") @@ -1595,6 +1740,7 @@ class Program_weight_tensor_parameter_144: class Program_weight_tensor_parameter_145: name = "parameter_145" + original_name = "linear_1476.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.293392") @@ -1606,6 +1752,7 @@ class Program_weight_tensor_parameter_145: class Program_weight_tensor_parameter_146: name = "parameter_146" + original_name = "layer_norm_508.b_0" shape = [1024] dtype = "float32" min_val = float("-2.14724") @@ -1617,6 +1764,7 @@ class Program_weight_tensor_parameter_146: class Program_weight_tensor_parameter_147: name = "parameter_147" + original_name = "layer_norm_508.w_0" shape = [1024] dtype = "float32" min_val = float("0.11918") @@ -1628,6 +1776,7 @@ class Program_weight_tensor_parameter_147: class Program_weight_tensor_parameter_148: name = "parameter_148" + original_name = "layer_norm_507.b_0" shape = [1024] dtype = "float32" min_val = float("-2.15367") @@ -1639,6 +1788,7 @@ class Program_weight_tensor_parameter_148: class Program_weight_tensor_parameter_149: name = "parameter_149" + original_name = "layer_norm_507.w_0" shape = [1024] dtype = "float32" min_val = float("0.673837") @@ -1650,6 +1800,7 @@ class Program_weight_tensor_parameter_149: class Program_weight_tensor_parameter_150: name = "parameter_150" + original_name = "linear_1475.b_0" shape = [1024] dtype = "float32" min_val = float("-0.549878") @@ -1661,6 +1812,7 @@ class Program_weight_tensor_parameter_150: class Program_weight_tensor_parameter_151: name = "parameter_151" + original_name = "linear_1475.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-1.42455") @@ -1672,6 +1824,7 @@ class Program_weight_tensor_parameter_151: class Program_weight_tensor_parameter_152: name = "parameter_152" + original_name = "linear_1474.b_0" shape = [4096] dtype = "float32" min_val = float("-0.246592") @@ -1683,6 +1836,7 @@ class Program_weight_tensor_parameter_152: class Program_weight_tensor_parameter_153: name = "parameter_153" + original_name = "linear_1474.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.272147") @@ -1694,6 +1848,7 @@ class Program_weight_tensor_parameter_153: class Program_weight_tensor_parameter_154: name = "parameter_154" + original_name = "linear_1473.b_0" shape = [1024] dtype = "float32" min_val = float("-0.134313") @@ -1705,6 +1860,7 @@ class Program_weight_tensor_parameter_154: class Program_weight_tensor_parameter_155: name = "parameter_155" + original_name = "linear_1473.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.226639") @@ -1716,6 +1872,7 @@ class Program_weight_tensor_parameter_155: class Program_weight_tensor_parameter_156: name = "parameter_156" + original_name = "linear_1472.b_0" shape = [1024] dtype = "float32" min_val = float("-0.143467") @@ -1727,6 +1884,7 @@ class Program_weight_tensor_parameter_156: class Program_weight_tensor_parameter_157: name = "parameter_157" + original_name = "linear_1472.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.168514") @@ -1738,6 +1896,7 @@ class Program_weight_tensor_parameter_157: class Program_weight_tensor_parameter_158: name = "parameter_158" + original_name = "linear_1471.b_0" shape = [1024] dtype = "float32" min_val = float("-1.39244") @@ -1749,6 +1908,7 @@ class Program_weight_tensor_parameter_158: class Program_weight_tensor_parameter_159: name = "parameter_159" + original_name = "linear_1471.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.344512") @@ -1760,6 +1920,7 @@ class Program_weight_tensor_parameter_159: class Program_weight_tensor_parameter_160: name = "parameter_160" + original_name = "linear_1470.b_0" shape = [1024] dtype = "float32" min_val = float("-0.597176") @@ -1771,6 +1932,7 @@ class Program_weight_tensor_parameter_160: class Program_weight_tensor_parameter_161: name = "parameter_161" + original_name = "linear_1470.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.344667") @@ -1782,6 +1944,7 @@ class Program_weight_tensor_parameter_161: class Program_weight_tensor_parameter_162: name = "parameter_162" + original_name = "layer_norm_506.b_0" shape = [1024] dtype = "float32" min_val = float("-1.34402") @@ -1793,6 +1956,7 @@ class Program_weight_tensor_parameter_162: class Program_weight_tensor_parameter_163: name = "parameter_163" + original_name = "layer_norm_506.w_0" shape = [1024] dtype = "float32" min_val = float("0.399463") @@ -1804,6 +1968,7 @@ class Program_weight_tensor_parameter_163: class Program_weight_tensor_parameter_164: name = "parameter_164" + original_name = "layer_norm_505.b_0" shape = [1024] dtype = "float32" min_val = float("-1.41746") @@ -1815,6 +1980,7 @@ class Program_weight_tensor_parameter_164: class Program_weight_tensor_parameter_165: name = "parameter_165" + original_name = "layer_norm_505.w_0" shape = [1024] dtype = "float32" min_val = float("0.652823") @@ -1826,6 +1992,7 @@ class Program_weight_tensor_parameter_165: class Program_weight_tensor_parameter_166: name = "parameter_166" + original_name = "linear_1469.b_0" shape = [1024] dtype = "float32" min_val = float("-0.79786") @@ -1837,6 +2004,7 @@ class Program_weight_tensor_parameter_166: class Program_weight_tensor_parameter_167: name = "parameter_167" + original_name = "linear_1469.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-0.625151") @@ -1848,6 +2016,7 @@ class Program_weight_tensor_parameter_167: class Program_weight_tensor_parameter_168: name = "parameter_168" + original_name = "linear_1468.b_0" shape = [4096] dtype = "float32" min_val = float("-0.300741") @@ -1859,6 +2028,7 @@ class Program_weight_tensor_parameter_168: class Program_weight_tensor_parameter_169: name = "parameter_169" + original_name = "linear_1468.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.285208") @@ -1870,6 +2040,7 @@ class Program_weight_tensor_parameter_169: class Program_weight_tensor_parameter_170: name = "parameter_170" + original_name = "linear_1467.b_0" shape = [1024] dtype = "float32" min_val = float("-0.146095") @@ -1881,6 +2052,7 @@ class Program_weight_tensor_parameter_170: class Program_weight_tensor_parameter_171: name = "parameter_171" + original_name = "linear_1467.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.280712") @@ -1892,6 +2064,7 @@ class Program_weight_tensor_parameter_171: class Program_weight_tensor_parameter_172: name = "parameter_172" + original_name = "linear_1466.b_0" shape = [1024] dtype = "float32" min_val = float("-0.198539") @@ -1903,6 +2076,7 @@ class Program_weight_tensor_parameter_172: class Program_weight_tensor_parameter_173: name = "parameter_173" + original_name = "linear_1466.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.186719") @@ -1914,6 +2088,7 @@ class Program_weight_tensor_parameter_173: class Program_weight_tensor_parameter_174: name = "parameter_174" + original_name = "linear_1465.b_0" shape = [1024] dtype = "float32" min_val = float("-1.07527") @@ -1925,6 +2100,7 @@ class Program_weight_tensor_parameter_174: class Program_weight_tensor_parameter_175: name = "parameter_175" + original_name = "linear_1465.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.301047") @@ -1936,6 +2112,7 @@ class Program_weight_tensor_parameter_175: class Program_weight_tensor_parameter_176: name = "parameter_176" + original_name = "linear_1464.b_0" shape = [1024] dtype = "float32" min_val = float("-0.453715") @@ -1947,6 +2124,7 @@ class Program_weight_tensor_parameter_176: class Program_weight_tensor_parameter_177: name = "parameter_177" + original_name = "linear_1464.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.294668") @@ -1958,6 +2136,7 @@ class Program_weight_tensor_parameter_177: class Program_weight_tensor_parameter_178: name = "parameter_178" + original_name = "layer_norm_504.b_0" shape = [1024] dtype = "float32" min_val = float("-0.409497") @@ -1969,6 +2148,7 @@ class Program_weight_tensor_parameter_178: class Program_weight_tensor_parameter_179: name = "parameter_179" + original_name = "layer_norm_504.w_0" shape = [1024] dtype = "float32" min_val = float("0.406718") @@ -1980,6 +2160,7 @@ class Program_weight_tensor_parameter_179: class Program_weight_tensor_parameter_180: name = "parameter_180" + original_name = "layer_norm_503.b_0" shape = [1024] dtype = "float32" min_val = float("-1.53891") @@ -1991,6 +2172,7 @@ class Program_weight_tensor_parameter_180: class Program_weight_tensor_parameter_181: name = "parameter_181" + original_name = "layer_norm_503.w_0" shape = [1024] dtype = "float32" min_val = float("0.659799") @@ -2002,6 +2184,7 @@ class Program_weight_tensor_parameter_181: class Program_weight_tensor_parameter_182: name = "parameter_182" + original_name = "linear_1463.b_0" shape = [1024] dtype = "float32" min_val = float("-0.665232") @@ -2013,6 +2196,7 @@ class Program_weight_tensor_parameter_182: class Program_weight_tensor_parameter_183: name = "parameter_183" + original_name = "linear_1463.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-1.34637") @@ -2024,6 +2208,7 @@ class Program_weight_tensor_parameter_183: class Program_weight_tensor_parameter_184: name = "parameter_184" + original_name = "linear_1462.b_0" shape = [4096] dtype = "float32" min_val = float("-0.251617") @@ -2035,6 +2220,7 @@ class Program_weight_tensor_parameter_184: class Program_weight_tensor_parameter_185: name = "parameter_185" + original_name = "linear_1462.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.40385") @@ -2046,6 +2232,7 @@ class Program_weight_tensor_parameter_185: class Program_weight_tensor_parameter_186: name = "parameter_186" + original_name = "linear_1461.b_0" shape = [1024] dtype = "float32" min_val = float("-0.179886") @@ -2057,6 +2244,7 @@ class Program_weight_tensor_parameter_186: class Program_weight_tensor_parameter_187: name = "parameter_187" + original_name = "linear_1461.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.25468") @@ -2068,6 +2256,7 @@ class Program_weight_tensor_parameter_187: class Program_weight_tensor_parameter_188: name = "parameter_188" + original_name = "linear_1460.b_0" shape = [1024] dtype = "float32" min_val = float("-0.18108") @@ -2079,6 +2268,7 @@ class Program_weight_tensor_parameter_188: class Program_weight_tensor_parameter_189: name = "parameter_189" + original_name = "linear_1460.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.169884") @@ -2090,6 +2280,7 @@ class Program_weight_tensor_parameter_189: class Program_weight_tensor_parameter_190: name = "parameter_190" + original_name = "linear_1459.b_0" shape = [1024] dtype = "float32" min_val = float("-0.797254") @@ -2101,6 +2292,7 @@ class Program_weight_tensor_parameter_190: class Program_weight_tensor_parameter_191: name = "parameter_191" + original_name = "linear_1459.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.266872") @@ -2112,6 +2304,7 @@ class Program_weight_tensor_parameter_191: class Program_weight_tensor_parameter_192: name = "parameter_192" + original_name = "linear_1458.b_0" shape = [1024] dtype = "float32" min_val = float("-0.399872") @@ -2123,6 +2316,7 @@ class Program_weight_tensor_parameter_192: class Program_weight_tensor_parameter_193: name = "parameter_193" + original_name = "linear_1458.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.309326") @@ -2134,6 +2328,7 @@ class Program_weight_tensor_parameter_193: class Program_weight_tensor_parameter_194: name = "parameter_194" + original_name = "layer_norm_502.b_0" shape = [1024] dtype = "float32" min_val = float("-0.529162") @@ -2145,6 +2340,7 @@ class Program_weight_tensor_parameter_194: class Program_weight_tensor_parameter_195: name = "parameter_195" + original_name = "layer_norm_502.w_0" shape = [1024] dtype = "float32" min_val = float("0.368106") @@ -2156,6 +2352,7 @@ class Program_weight_tensor_parameter_195: class Program_weight_tensor_parameter_196: name = "parameter_196" + original_name = "layer_norm_501.b_0" shape = [1024] dtype = "float32" min_val = float("-1.49305") @@ -2167,6 +2364,7 @@ class Program_weight_tensor_parameter_196: class Program_weight_tensor_parameter_197: name = "parameter_197" + original_name = "layer_norm_501.w_0" shape = [1024] dtype = "float32" min_val = float("0.674977") @@ -2178,6 +2376,7 @@ class Program_weight_tensor_parameter_197: class Program_weight_tensor_parameter_198: name = "parameter_198" + original_name = "linear_1457.b_0" shape = [1024] dtype = "float32" min_val = float("-0.642651") @@ -2189,6 +2388,7 @@ class Program_weight_tensor_parameter_198: class Program_weight_tensor_parameter_199: name = "parameter_199" + original_name = "linear_1457.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-1.36642") @@ -2200,6 +2400,7 @@ class Program_weight_tensor_parameter_199: class Program_weight_tensor_parameter_200: name = "parameter_200" + original_name = "linear_1456.b_0" shape = [4096] dtype = "float32" min_val = float("-0.250095") @@ -2211,6 +2412,7 @@ class Program_weight_tensor_parameter_200: class Program_weight_tensor_parameter_201: name = "parameter_201" + original_name = "linear_1456.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.606788") @@ -2222,6 +2424,7 @@ class Program_weight_tensor_parameter_201: class Program_weight_tensor_parameter_202: name = "parameter_202" + original_name = "linear_1455.b_0" shape = [1024] dtype = "float32" min_val = float("-0.256123") @@ -2233,6 +2436,7 @@ class Program_weight_tensor_parameter_202: class Program_weight_tensor_parameter_203: name = "parameter_203" + original_name = "linear_1455.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.346923") @@ -2244,6 +2448,7 @@ class Program_weight_tensor_parameter_203: class Program_weight_tensor_parameter_204: name = "parameter_204" + original_name = "linear_1454.b_0" shape = [1024] dtype = "float32" min_val = float("-0.18194") @@ -2255,6 +2460,7 @@ class Program_weight_tensor_parameter_204: class Program_weight_tensor_parameter_205: name = "parameter_205" + original_name = "linear_1454.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.163613") @@ -2266,6 +2472,7 @@ class Program_weight_tensor_parameter_205: class Program_weight_tensor_parameter_206: name = "parameter_206" + original_name = "linear_1453.b_0" shape = [1024] dtype = "float32" min_val = float("-1.14377") @@ -2277,6 +2484,7 @@ class Program_weight_tensor_parameter_206: class Program_weight_tensor_parameter_207: name = "parameter_207" + original_name = "linear_1453.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.314863") @@ -2288,6 +2496,7 @@ class Program_weight_tensor_parameter_207: class Program_weight_tensor_parameter_208: name = "parameter_208" + original_name = "linear_1452.b_0" shape = [1024] dtype = "float32" min_val = float("-0.4629") @@ -2299,6 +2508,7 @@ class Program_weight_tensor_parameter_208: class Program_weight_tensor_parameter_209: name = "parameter_209" + original_name = "linear_1452.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.340383") @@ -2310,6 +2520,7 @@ class Program_weight_tensor_parameter_209: class Program_weight_tensor_parameter_210: name = "parameter_210" + original_name = "layer_norm_500.b_0" shape = [1024] dtype = "float32" min_val = float("-0.624797") @@ -2321,6 +2532,7 @@ class Program_weight_tensor_parameter_210: class Program_weight_tensor_parameter_211: name = "parameter_211" + original_name = "layer_norm_500.w_0" shape = [1024] dtype = "float32" min_val = float("0.426274") @@ -2332,6 +2544,7 @@ class Program_weight_tensor_parameter_211: class Program_weight_tensor_parameter_212: name = "parameter_212" + original_name = "layer_norm_499.b_0" shape = [1024] dtype = "float32" min_val = float("-1.37062") @@ -2343,6 +2556,7 @@ class Program_weight_tensor_parameter_212: class Program_weight_tensor_parameter_213: name = "parameter_213" + original_name = "layer_norm_499.w_0" shape = [1024] dtype = "float32" min_val = float("0.685264") @@ -2354,6 +2568,7 @@ class Program_weight_tensor_parameter_213: class Program_weight_tensor_parameter_214: name = "parameter_214" + original_name = "linear_1451.b_0" shape = [1024] dtype = "float32" min_val = float("-0.67817") @@ -2365,6 +2580,7 @@ class Program_weight_tensor_parameter_214: class Program_weight_tensor_parameter_215: name = "parameter_215" + original_name = "linear_1451.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-1.37473") @@ -2376,6 +2592,7 @@ class Program_weight_tensor_parameter_215: class Program_weight_tensor_parameter_216: name = "parameter_216" + original_name = "linear_1450.b_0" shape = [4096] dtype = "float32" min_val = float("-0.256734") @@ -2387,6 +2604,7 @@ class Program_weight_tensor_parameter_216: class Program_weight_tensor_parameter_217: name = "parameter_217" + original_name = "linear_1450.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.502751") @@ -2398,6 +2616,7 @@ class Program_weight_tensor_parameter_217: class Program_weight_tensor_parameter_218: name = "parameter_218" + original_name = "linear_1449.b_0" shape = [1024] dtype = "float32" min_val = float("-0.234093") @@ -2409,6 +2628,7 @@ class Program_weight_tensor_parameter_218: class Program_weight_tensor_parameter_219: name = "parameter_219" + original_name = "linear_1449.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.286908") @@ -2420,6 +2640,7 @@ class Program_weight_tensor_parameter_219: class Program_weight_tensor_parameter_220: name = "parameter_220" + original_name = "linear_1448.b_0" shape = [1024] dtype = "float32" min_val = float("-0.164998") @@ -2431,6 +2652,7 @@ class Program_weight_tensor_parameter_220: class Program_weight_tensor_parameter_221: name = "parameter_221" + original_name = "linear_1448.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.155097") @@ -2442,6 +2664,7 @@ class Program_weight_tensor_parameter_221: class Program_weight_tensor_parameter_222: name = "parameter_222" + original_name = "linear_1447.b_0" shape = [1024] dtype = "float32" min_val = float("-0.973211") @@ -2453,6 +2676,7 @@ class Program_weight_tensor_parameter_222: class Program_weight_tensor_parameter_223: name = "parameter_223" + original_name = "linear_1447.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.286199") @@ -2464,6 +2688,7 @@ class Program_weight_tensor_parameter_223: class Program_weight_tensor_parameter_224: name = "parameter_224" + original_name = "linear_1446.b_0" shape = [1024] dtype = "float32" min_val = float("-0.384704") @@ -2475,6 +2700,7 @@ class Program_weight_tensor_parameter_224: class Program_weight_tensor_parameter_225: name = "parameter_225" + original_name = "linear_1446.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.257322") @@ -2486,6 +2712,7 @@ class Program_weight_tensor_parameter_225: class Program_weight_tensor_parameter_226: name = "parameter_226" + original_name = "layer_norm_498.b_0" shape = [1024] dtype = "float32" min_val = float("-0.416323") @@ -2497,6 +2724,7 @@ class Program_weight_tensor_parameter_226: class Program_weight_tensor_parameter_227: name = "parameter_227" + original_name = "layer_norm_498.w_0" shape = [1024] dtype = "float32" min_val = float("0.472291") @@ -2508,6 +2736,7 @@ class Program_weight_tensor_parameter_227: class Program_weight_tensor_parameter_228: name = "parameter_228" + original_name = "layer_norm_497.b_0" shape = [1024] dtype = "float32" min_val = float("-1.67723") @@ -2519,6 +2748,7 @@ class Program_weight_tensor_parameter_228: class Program_weight_tensor_parameter_229: name = "parameter_229" + original_name = "layer_norm_497.w_0" shape = [1024] dtype = "float32" min_val = float("0.736293") @@ -2530,6 +2760,7 @@ class Program_weight_tensor_parameter_229: class Program_weight_tensor_parameter_230: name = "parameter_230" + original_name = "linear_1445.b_0" shape = [1024] dtype = "float32" min_val = float("-0.321379") @@ -2541,6 +2772,7 @@ class Program_weight_tensor_parameter_230: class Program_weight_tensor_parameter_231: name = "parameter_231" + original_name = "linear_1445.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-0.909544") @@ -2552,6 +2784,7 @@ class Program_weight_tensor_parameter_231: class Program_weight_tensor_parameter_232: name = "parameter_232" + original_name = "linear_1444.b_0" shape = [4096] dtype = "float32" min_val = float("-0.199841") @@ -2563,6 +2796,7 @@ class Program_weight_tensor_parameter_232: class Program_weight_tensor_parameter_233: name = "parameter_233" + original_name = "linear_1444.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.504869") @@ -2574,6 +2808,7 @@ class Program_weight_tensor_parameter_233: class Program_weight_tensor_parameter_234: name = "parameter_234" + original_name = "linear_1443.b_0" shape = [1024] dtype = "float32" min_val = float("-0.16815") @@ -2585,6 +2820,7 @@ class Program_weight_tensor_parameter_234: class Program_weight_tensor_parameter_235: name = "parameter_235" + original_name = "linear_1443.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.251522") @@ -2596,6 +2832,7 @@ class Program_weight_tensor_parameter_235: class Program_weight_tensor_parameter_236: name = "parameter_236" + original_name = "linear_1442.b_0" shape = [1024] dtype = "float32" min_val = float("-0.225458") @@ -2607,6 +2844,7 @@ class Program_weight_tensor_parameter_236: class Program_weight_tensor_parameter_237: name = "parameter_237" + original_name = "linear_1442.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.230247") @@ -2618,6 +2856,7 @@ class Program_weight_tensor_parameter_237: class Program_weight_tensor_parameter_238: name = "parameter_238" + original_name = "linear_1441.b_0" shape = [1024] dtype = "float32" min_val = float("-0.886261") @@ -2629,6 +2868,7 @@ class Program_weight_tensor_parameter_238: class Program_weight_tensor_parameter_239: name = "parameter_239" + original_name = "linear_1441.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.293633") @@ -2640,6 +2880,7 @@ class Program_weight_tensor_parameter_239: class Program_weight_tensor_parameter_240: name = "parameter_240" + original_name = "linear_1440.b_0" shape = [1024] dtype = "float32" min_val = float("-0.407248") @@ -2651,6 +2892,7 @@ class Program_weight_tensor_parameter_240: class Program_weight_tensor_parameter_241: name = "parameter_241" + original_name = "linear_1440.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.289408") @@ -2662,6 +2904,7 @@ class Program_weight_tensor_parameter_241: class Program_weight_tensor_parameter_242: name = "parameter_242" + original_name = "layer_norm_496.b_0" shape = [1024] dtype = "float32" min_val = float("-0.626548") @@ -2673,6 +2916,7 @@ class Program_weight_tensor_parameter_242: class Program_weight_tensor_parameter_243: name = "parameter_243" + original_name = "layer_norm_496.w_0" shape = [1024] dtype = "float32" min_val = float("0.484988") @@ -2684,6 +2928,7 @@ class Program_weight_tensor_parameter_243: class Program_weight_tensor_parameter_244: name = "parameter_244" + original_name = "layer_norm_495.b_0" shape = [1024] dtype = "float32" min_val = float("-1.68536") @@ -2695,6 +2940,7 @@ class Program_weight_tensor_parameter_244: class Program_weight_tensor_parameter_245: name = "parameter_245" + original_name = "layer_norm_495.w_0" shape = [1024] dtype = "float32" min_val = float("0.632309") @@ -2706,6 +2952,7 @@ class Program_weight_tensor_parameter_245: class Program_weight_tensor_parameter_246: name = "parameter_246" + original_name = "linear_1439.b_0" shape = [1024] dtype = "float32" min_val = float("-0.174579") @@ -2717,6 +2964,7 @@ class Program_weight_tensor_parameter_246: class Program_weight_tensor_parameter_247: name = "parameter_247" + original_name = "linear_1439.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-1.64933") @@ -2728,6 +2976,7 @@ class Program_weight_tensor_parameter_247: class Program_weight_tensor_parameter_248: name = "parameter_248" + original_name = "linear_1438.b_0" shape = [4096] dtype = "float32" min_val = float("-0.194982") @@ -2739,6 +2988,7 @@ class Program_weight_tensor_parameter_248: class Program_weight_tensor_parameter_249: name = "parameter_249" + original_name = "linear_1438.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.310586") @@ -2750,6 +3000,7 @@ class Program_weight_tensor_parameter_249: class Program_weight_tensor_parameter_250: name = "parameter_250" + original_name = "linear_1437.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0774591") @@ -2761,6 +3012,7 @@ class Program_weight_tensor_parameter_250: class Program_weight_tensor_parameter_251: name = "parameter_251" + original_name = "linear_1437.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.518722") @@ -2772,6 +3024,7 @@ class Program_weight_tensor_parameter_251: class Program_weight_tensor_parameter_252: name = "parameter_252" + original_name = "linear_1436.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0741598") @@ -2783,6 +3036,7 @@ class Program_weight_tensor_parameter_252: class Program_weight_tensor_parameter_253: name = "parameter_253" + original_name = "linear_1436.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.146344") @@ -2794,6 +3048,7 @@ class Program_weight_tensor_parameter_253: class Program_weight_tensor_parameter_254: name = "parameter_254" + original_name = "linear_1435.b_0" shape = [1024] dtype = "float32" min_val = float("-0.889143") @@ -2805,6 +3060,7 @@ class Program_weight_tensor_parameter_254: class Program_weight_tensor_parameter_255: name = "parameter_255" + original_name = "linear_1435.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.299354") @@ -2816,6 +3072,7 @@ class Program_weight_tensor_parameter_255: class Program_weight_tensor_parameter_256: name = "parameter_256" + original_name = "linear_1434.b_0" shape = [1024] dtype = "float32" min_val = float("-0.53935") @@ -2827,6 +3084,7 @@ class Program_weight_tensor_parameter_256: class Program_weight_tensor_parameter_257: name = "parameter_257" + original_name = "linear_1434.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.286898") @@ -2838,6 +3096,7 @@ class Program_weight_tensor_parameter_257: class Program_weight_tensor_parameter_258: name = "parameter_258" + original_name = "layer_norm_494.b_0" shape = [1024] dtype = "float32" min_val = float("-1.14107") @@ -2849,6 +3108,7 @@ class Program_weight_tensor_parameter_258: class Program_weight_tensor_parameter_259: name = "parameter_259" + original_name = "layer_norm_494.w_0" shape = [1024] dtype = "float32" min_val = float("0.107997") @@ -2860,6 +3120,7 @@ class Program_weight_tensor_parameter_259: class Program_weight_tensor_parameter_260: name = "parameter_260" + original_name = "layer_norm_493.b_0" shape = [1024] dtype = "float32" min_val = float("-1.67549") @@ -2871,6 +3132,7 @@ class Program_weight_tensor_parameter_260: class Program_weight_tensor_parameter_261: name = "parameter_261" + original_name = "layer_norm_493.w_0" shape = [1024] dtype = "float32" min_val = float("0.811891") @@ -2882,6 +3144,7 @@ class Program_weight_tensor_parameter_261: class Program_weight_tensor_parameter_262: name = "parameter_262" + original_name = "linear_1433.b_0" shape = [1024] dtype = "float32" min_val = float("-0.194989") @@ -2893,6 +3156,7 @@ class Program_weight_tensor_parameter_262: class Program_weight_tensor_parameter_263: name = "parameter_263" + original_name = "linear_1433.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-1.27388") @@ -2904,6 +3168,7 @@ class Program_weight_tensor_parameter_263: class Program_weight_tensor_parameter_264: name = "parameter_264" + original_name = "linear_1432.b_0" shape = [4096] dtype = "float32" min_val = float("-0.217288") @@ -2915,6 +3180,7 @@ class Program_weight_tensor_parameter_264: class Program_weight_tensor_parameter_265: name = "parameter_265" + original_name = "linear_1432.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.438439") @@ -2926,6 +3192,7 @@ class Program_weight_tensor_parameter_265: class Program_weight_tensor_parameter_266: name = "parameter_266" + original_name = "linear_1431.b_0" shape = [1024] dtype = "float32" min_val = float("-0.122585") @@ -2937,6 +3204,7 @@ class Program_weight_tensor_parameter_266: class Program_weight_tensor_parameter_267: name = "parameter_267" + original_name = "linear_1431.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.233094") @@ -2948,6 +3216,7 @@ class Program_weight_tensor_parameter_267: class Program_weight_tensor_parameter_268: name = "parameter_268" + original_name = "linear_1430.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0420376") @@ -2959,6 +3228,7 @@ class Program_weight_tensor_parameter_268: class Program_weight_tensor_parameter_269: name = "parameter_269" + original_name = "linear_1430.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.156447") @@ -2970,6 +3240,7 @@ class Program_weight_tensor_parameter_269: class Program_weight_tensor_parameter_270: name = "parameter_270" + original_name = "linear_1429.b_0" shape = [1024] dtype = "float32" min_val = float("-0.492519") @@ -2981,6 +3252,7 @@ class Program_weight_tensor_parameter_270: class Program_weight_tensor_parameter_271: name = "parameter_271" + original_name = "linear_1429.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.339099") @@ -2992,6 +3264,7 @@ class Program_weight_tensor_parameter_271: class Program_weight_tensor_parameter_272: name = "parameter_272" + original_name = "linear_1428.b_0" shape = [1024] dtype = "float32" min_val = float("-0.357655") @@ -3003,6 +3276,7 @@ class Program_weight_tensor_parameter_272: class Program_weight_tensor_parameter_273: name = "parameter_273" + original_name = "linear_1428.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.33023") @@ -3014,6 +3288,7 @@ class Program_weight_tensor_parameter_273: class Program_weight_tensor_parameter_274: name = "parameter_274" + original_name = "layer_norm_492.b_0" shape = [1024] dtype = "float32" min_val = float("-1.28129") @@ -3025,6 +3300,7 @@ class Program_weight_tensor_parameter_274: class Program_weight_tensor_parameter_275: name = "parameter_275" + original_name = "layer_norm_492.w_0" shape = [1024] dtype = "float32" min_val = float("0.108529") @@ -3036,6 +3312,7 @@ class Program_weight_tensor_parameter_275: class Program_weight_tensor_parameter_276: name = "parameter_276" + original_name = "layer_norm_491.b_0" shape = [1024] dtype = "float32" min_val = float("-1.40004") @@ -3047,6 +3324,7 @@ class Program_weight_tensor_parameter_276: class Program_weight_tensor_parameter_277: name = "parameter_277" + original_name = "layer_norm_491.w_0" shape = [1024] dtype = "float32" min_val = float("0.747964") @@ -3058,6 +3336,7 @@ class Program_weight_tensor_parameter_277: class Program_weight_tensor_parameter_278: name = "parameter_278" + original_name = "linear_1427.b_0" shape = [1024] dtype = "float32" min_val = float("-0.196985") @@ -3069,6 +3348,7 @@ class Program_weight_tensor_parameter_278: class Program_weight_tensor_parameter_279: name = "parameter_279" + original_name = "linear_1427.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-0.964583") @@ -3080,6 +3360,7 @@ class Program_weight_tensor_parameter_279: class Program_weight_tensor_parameter_280: name = "parameter_280" + original_name = "linear_1426.b_0" shape = [4096] dtype = "float32" min_val = float("-0.173225") @@ -3091,6 +3372,7 @@ class Program_weight_tensor_parameter_280: class Program_weight_tensor_parameter_281: name = "parameter_281" + original_name = "linear_1426.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.199327") @@ -3102,6 +3384,7 @@ class Program_weight_tensor_parameter_281: class Program_weight_tensor_parameter_282: name = "parameter_282" + original_name = "linear_1425.b_0" shape = [1024] dtype = "float32" min_val = float("-0.140961") @@ -3113,6 +3396,7 @@ class Program_weight_tensor_parameter_282: class Program_weight_tensor_parameter_283: name = "parameter_283" + original_name = "linear_1425.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.214573") @@ -3124,6 +3408,7 @@ class Program_weight_tensor_parameter_283: class Program_weight_tensor_parameter_284: name = "parameter_284" + original_name = "linear_1424.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0441024") @@ -3135,6 +3420,7 @@ class Program_weight_tensor_parameter_284: class Program_weight_tensor_parameter_285: name = "parameter_285" + original_name = "linear_1424.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.142601") @@ -3146,6 +3432,7 @@ class Program_weight_tensor_parameter_285: class Program_weight_tensor_parameter_286: name = "parameter_286" + original_name = "linear_1423.b_0" shape = [1024] dtype = "float32" min_val = float("-0.736854") @@ -3157,6 +3444,7 @@ class Program_weight_tensor_parameter_286: class Program_weight_tensor_parameter_287: name = "parameter_287" + original_name = "linear_1423.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.516513") @@ -3168,6 +3456,7 @@ class Program_weight_tensor_parameter_287: class Program_weight_tensor_parameter_288: name = "parameter_288" + original_name = "linear_1422.b_0" shape = [1024] dtype = "float32" min_val = float("-0.396943") @@ -3179,6 +3468,7 @@ class Program_weight_tensor_parameter_288: class Program_weight_tensor_parameter_289: name = "parameter_289" + original_name = "linear_1422.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.310629") @@ -3190,6 +3480,7 @@ class Program_weight_tensor_parameter_289: class Program_weight_tensor_parameter_290: name = "parameter_290" + original_name = "layer_norm_490.b_0" shape = [1024] dtype = "float32" min_val = float("-1.19352") @@ -3201,6 +3492,7 @@ class Program_weight_tensor_parameter_290: class Program_weight_tensor_parameter_291: name = "parameter_291" + original_name = "layer_norm_490.w_0" shape = [1024] dtype = "float32" min_val = float("0.186839") @@ -3212,6 +3504,7 @@ class Program_weight_tensor_parameter_291: class Program_weight_tensor_parameter_292: name = "parameter_292" + original_name = "layer_norm_489.b_0" shape = [1024] dtype = "float32" min_val = float("-0.909079") @@ -3223,6 +3516,7 @@ class Program_weight_tensor_parameter_292: class Program_weight_tensor_parameter_293: name = "parameter_293" + original_name = "layer_norm_489.w_0" shape = [1024] dtype = "float32" min_val = float("0.735325") @@ -3234,6 +3528,7 @@ class Program_weight_tensor_parameter_293: class Program_weight_tensor_parameter_294: name = "parameter_294" + original_name = "linear_1421.b_0" shape = [1024] dtype = "float32" min_val = float("-0.230544") @@ -3245,6 +3540,7 @@ class Program_weight_tensor_parameter_294: class Program_weight_tensor_parameter_295: name = "parameter_295" + original_name = "linear_1421.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-0.743694") @@ -3256,6 +3552,7 @@ class Program_weight_tensor_parameter_295: class Program_weight_tensor_parameter_296: name = "parameter_296" + original_name = "linear_1420.b_0" shape = [4096] dtype = "float32" min_val = float("-0.148718") @@ -3267,6 +3564,7 @@ class Program_weight_tensor_parameter_296: class Program_weight_tensor_parameter_297: name = "parameter_297" + original_name = "linear_1420.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.235816") @@ -3278,6 +3576,7 @@ class Program_weight_tensor_parameter_297: class Program_weight_tensor_parameter_298: name = "parameter_298" + original_name = "linear_1419.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0701489") @@ -3289,6 +3588,7 @@ class Program_weight_tensor_parameter_298: class Program_weight_tensor_parameter_299: name = "parameter_299" + original_name = "linear_1419.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.202181") @@ -3300,6 +3600,7 @@ class Program_weight_tensor_parameter_299: class Program_weight_tensor_parameter_300: name = "parameter_300" + original_name = "linear_1418.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0397438") @@ -3311,6 +3612,7 @@ class Program_weight_tensor_parameter_300: class Program_weight_tensor_parameter_301: name = "parameter_301" + original_name = "linear_1418.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.161015") @@ -3322,6 +3624,7 @@ class Program_weight_tensor_parameter_301: class Program_weight_tensor_parameter_302: name = "parameter_302" + original_name = "linear_1417.b_0" shape = [1024] dtype = "float32" min_val = float("-0.615488") @@ -3333,6 +3636,7 @@ class Program_weight_tensor_parameter_302: class Program_weight_tensor_parameter_303: name = "parameter_303" + original_name = "linear_1417.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.309885") @@ -3344,6 +3648,7 @@ class Program_weight_tensor_parameter_303: class Program_weight_tensor_parameter_304: name = "parameter_304" + original_name = "linear_1416.b_0" shape = [1024] dtype = "float32" min_val = float("-0.36178") @@ -3355,6 +3660,7 @@ class Program_weight_tensor_parameter_304: class Program_weight_tensor_parameter_305: name = "parameter_305" + original_name = "linear_1416.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.311235") @@ -3366,6 +3672,7 @@ class Program_weight_tensor_parameter_305: class Program_weight_tensor_parameter_306: name = "parameter_306" + original_name = "layer_norm_488.b_0" shape = [1024] dtype = "float32" min_val = float("-0.479579") @@ -3377,6 +3684,7 @@ class Program_weight_tensor_parameter_306: class Program_weight_tensor_parameter_307: name = "parameter_307" + original_name = "layer_norm_488.w_0" shape = [1024] dtype = "float32" min_val = float("0.19717") @@ -3388,6 +3696,7 @@ class Program_weight_tensor_parameter_307: class Program_weight_tensor_parameter_308: name = "parameter_308" + original_name = "layer_norm_487.b_0" shape = [1024] dtype = "float32" min_val = float("-0.601447") @@ -3399,6 +3708,7 @@ class Program_weight_tensor_parameter_308: class Program_weight_tensor_parameter_309: name = "parameter_309" + original_name = "layer_norm_487.w_0" shape = [1024] dtype = "float32" min_val = float("0.71672") @@ -3410,6 +3720,7 @@ class Program_weight_tensor_parameter_309: class Program_weight_tensor_parameter_310: name = "parameter_310" + original_name = "linear_1415.b_0" shape = [1024] dtype = "float32" min_val = float("-0.226605") @@ -3421,6 +3732,7 @@ class Program_weight_tensor_parameter_310: class Program_weight_tensor_parameter_311: name = "parameter_311" + original_name = "linear_1415.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-0.604233") @@ -3432,6 +3744,7 @@ class Program_weight_tensor_parameter_311: class Program_weight_tensor_parameter_312: name = "parameter_312" + original_name = "linear_1414.b_0" shape = [4096] dtype = "float32" min_val = float("-0.160829") @@ -3443,6 +3756,7 @@ class Program_weight_tensor_parameter_312: class Program_weight_tensor_parameter_313: name = "parameter_313" + original_name = "linear_1414.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.256231") @@ -3454,6 +3768,7 @@ class Program_weight_tensor_parameter_313: class Program_weight_tensor_parameter_314: name = "parameter_314" + original_name = "linear_1413.b_0" shape = [1024] dtype = "float32" min_val = float("-0.108758") @@ -3465,6 +3780,7 @@ class Program_weight_tensor_parameter_314: class Program_weight_tensor_parameter_315: name = "parameter_315" + original_name = "linear_1413.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.173812") @@ -3476,6 +3792,7 @@ class Program_weight_tensor_parameter_315: class Program_weight_tensor_parameter_316: name = "parameter_316" + original_name = "linear_1412.b_0" shape = [1024] dtype = "float32" min_val = float("-0.143302") @@ -3487,6 +3804,7 @@ class Program_weight_tensor_parameter_316: class Program_weight_tensor_parameter_317: name = "parameter_317" + original_name = "linear_1412.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.150616") @@ -3498,6 +3816,7 @@ class Program_weight_tensor_parameter_317: class Program_weight_tensor_parameter_318: name = "parameter_318" + original_name = "linear_1411.b_0" shape = [1024] dtype = "float32" min_val = float("-0.646988") @@ -3509,6 +3828,7 @@ class Program_weight_tensor_parameter_318: class Program_weight_tensor_parameter_319: name = "parameter_319" + original_name = "linear_1411.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.317453") @@ -3520,6 +3840,7 @@ class Program_weight_tensor_parameter_319: class Program_weight_tensor_parameter_320: name = "parameter_320" + original_name = "linear_1410.b_0" shape = [1024] dtype = "float32" min_val = float("-0.330914") @@ -3531,6 +3852,7 @@ class Program_weight_tensor_parameter_320: class Program_weight_tensor_parameter_321: name = "parameter_321" + original_name = "linear_1410.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.291827") @@ -3542,6 +3864,7 @@ class Program_weight_tensor_parameter_321: class Program_weight_tensor_parameter_322: name = "parameter_322" + original_name = "layer_norm_486.b_0" shape = [1024] dtype = "float32" min_val = float("-0.202233") @@ -3553,6 +3876,7 @@ class Program_weight_tensor_parameter_322: class Program_weight_tensor_parameter_323: name = "parameter_323" + original_name = "layer_norm_486.w_0" shape = [1024] dtype = "float32" min_val = float("0.328862") @@ -3564,6 +3888,7 @@ class Program_weight_tensor_parameter_323: class Program_weight_tensor_parameter_324: name = "parameter_324" + original_name = "layer_norm_485.b_0" shape = [1024] dtype = "float32" min_val = float("-0.758595") @@ -3575,6 +3900,7 @@ class Program_weight_tensor_parameter_324: class Program_weight_tensor_parameter_325: name = "parameter_325" + original_name = "layer_norm_485.w_0" shape = [1024] dtype = "float32" min_val = float("0.761823") @@ -3586,6 +3912,7 @@ class Program_weight_tensor_parameter_325: class Program_weight_tensor_parameter_326: name = "parameter_326" + original_name = "linear_1409.b_0" shape = [1024] dtype = "float32" min_val = float("-0.114963") @@ -3597,6 +3924,7 @@ class Program_weight_tensor_parameter_326: class Program_weight_tensor_parameter_327: name = "parameter_327" + original_name = "linear_1409.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-0.630781") @@ -3608,6 +3936,7 @@ class Program_weight_tensor_parameter_327: class Program_weight_tensor_parameter_328: name = "parameter_328" + original_name = "linear_1408.b_0" shape = [4096] dtype = "float32" min_val = float("-0.172999") @@ -3619,6 +3948,7 @@ class Program_weight_tensor_parameter_328: class Program_weight_tensor_parameter_329: name = "parameter_329" + original_name = "linear_1408.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.225964") @@ -3630,6 +3960,7 @@ class Program_weight_tensor_parameter_329: class Program_weight_tensor_parameter_330: name = "parameter_330" + original_name = "linear_1407.b_0" shape = [1024] dtype = "float32" min_val = float("-0.234396") @@ -3641,6 +3972,7 @@ class Program_weight_tensor_parameter_330: class Program_weight_tensor_parameter_331: name = "parameter_331" + original_name = "linear_1407.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.190784") @@ -3652,6 +3984,7 @@ class Program_weight_tensor_parameter_331: class Program_weight_tensor_parameter_332: name = "parameter_332" + original_name = "linear_1406.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0869038") @@ -3663,6 +3996,7 @@ class Program_weight_tensor_parameter_332: class Program_weight_tensor_parameter_333: name = "parameter_333" + original_name = "linear_1406.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.163422") @@ -3674,6 +4008,7 @@ class Program_weight_tensor_parameter_333: class Program_weight_tensor_parameter_334: name = "parameter_334" + original_name = "linear_1405.b_0" shape = [1024] dtype = "float32" min_val = float("-0.333713") @@ -3685,6 +4020,7 @@ class Program_weight_tensor_parameter_334: class Program_weight_tensor_parameter_335: name = "parameter_335" + original_name = "linear_1405.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.231288") @@ -3696,6 +4032,7 @@ class Program_weight_tensor_parameter_335: class Program_weight_tensor_parameter_336: name = "parameter_336" + original_name = "linear_1404.b_0" shape = [1024] dtype = "float32" min_val = float("-0.293126") @@ -3707,6 +4044,7 @@ class Program_weight_tensor_parameter_336: class Program_weight_tensor_parameter_337: name = "parameter_337" + original_name = "linear_1404.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.255008") @@ -3718,6 +4056,7 @@ class Program_weight_tensor_parameter_337: class Program_weight_tensor_parameter_338: name = "parameter_338" + original_name = "layer_norm_484.b_0" shape = [1024] dtype = "float32" min_val = float("-0.27487") @@ -3729,6 +4068,7 @@ class Program_weight_tensor_parameter_338: class Program_weight_tensor_parameter_339: name = "parameter_339" + original_name = "layer_norm_484.w_0" shape = [1024] dtype = "float32" min_val = float("0.387424") @@ -3740,6 +4080,7 @@ class Program_weight_tensor_parameter_339: class Program_weight_tensor_parameter_340: name = "parameter_340" + original_name = "layer_norm_483.b_0" shape = [1024] dtype = "float32" min_val = float("-1.14492") @@ -3751,6 +4092,7 @@ class Program_weight_tensor_parameter_340: class Program_weight_tensor_parameter_341: name = "parameter_341" + original_name = "layer_norm_483.w_0" shape = [1024] dtype = "float32" min_val = float("0.827357") @@ -3762,6 +4104,7 @@ class Program_weight_tensor_parameter_341: class Program_weight_tensor_parameter_342: name = "parameter_342" + original_name = "linear_1403.b_0" shape = [1024] dtype = "float32" min_val = float("-0.327956") @@ -3773,6 +4116,7 @@ class Program_weight_tensor_parameter_342: class Program_weight_tensor_parameter_343: name = "parameter_343" + original_name = "linear_1403.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-0.596933") @@ -3784,6 +4128,7 @@ class Program_weight_tensor_parameter_343: class Program_weight_tensor_parameter_344: name = "parameter_344" + original_name = "linear_1402.b_0" shape = [4096] dtype = "float32" min_val = float("-0.141249") @@ -3795,6 +4140,7 @@ class Program_weight_tensor_parameter_344: class Program_weight_tensor_parameter_345: name = "parameter_345" + original_name = "linear_1402.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.229691") @@ -3806,6 +4152,7 @@ class Program_weight_tensor_parameter_345: class Program_weight_tensor_parameter_346: name = "parameter_346" + original_name = "linear_1401.b_0" shape = [1024] dtype = "float32" min_val = float("-0.140562") @@ -3817,6 +4164,7 @@ class Program_weight_tensor_parameter_346: class Program_weight_tensor_parameter_347: name = "parameter_347" + original_name = "linear_1401.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.268277") @@ -3828,6 +4176,7 @@ class Program_weight_tensor_parameter_347: class Program_weight_tensor_parameter_348: name = "parameter_348" + original_name = "linear_1400.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0919003") @@ -3839,6 +4188,7 @@ class Program_weight_tensor_parameter_348: class Program_weight_tensor_parameter_349: name = "parameter_349" + original_name = "linear_1400.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.182618") @@ -3850,6 +4200,7 @@ class Program_weight_tensor_parameter_349: class Program_weight_tensor_parameter_350: name = "parameter_350" + original_name = "linear_1399.b_0" shape = [1024] dtype = "float32" min_val = float("-0.371753") @@ -3861,6 +4212,7 @@ class Program_weight_tensor_parameter_350: class Program_weight_tensor_parameter_351: name = "parameter_351" + original_name = "linear_1399.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.334495") @@ -3872,6 +4224,7 @@ class Program_weight_tensor_parameter_351: class Program_weight_tensor_parameter_352: name = "parameter_352" + original_name = "linear_1398.b_0" shape = [1024] dtype = "float32" min_val = float("-0.309921") @@ -3883,6 +4236,7 @@ class Program_weight_tensor_parameter_352: class Program_weight_tensor_parameter_353: name = "parameter_353" + original_name = "linear_1398.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.26561") @@ -3894,6 +4248,7 @@ class Program_weight_tensor_parameter_353: class Program_weight_tensor_parameter_354: name = "parameter_354" + original_name = "layer_norm_482.b_0" shape = [1024] dtype = "float32" min_val = float("-0.247838") @@ -3905,6 +4260,7 @@ class Program_weight_tensor_parameter_354: class Program_weight_tensor_parameter_355: name = "parameter_355" + original_name = "layer_norm_482.w_0" shape = [1024] dtype = "float32" min_val = float("0.402601") @@ -3916,6 +4272,7 @@ class Program_weight_tensor_parameter_355: class Program_weight_tensor_parameter_356: name = "parameter_356" + original_name = "layer_norm_481.b_0" shape = [1024] dtype = "float32" min_val = float("-1.31776") @@ -3927,6 +4284,7 @@ class Program_weight_tensor_parameter_356: class Program_weight_tensor_parameter_357: name = "parameter_357" + original_name = "layer_norm_481.w_0" shape = [1024] dtype = "float32" min_val = float("0.738802") @@ -3938,6 +4296,7 @@ class Program_weight_tensor_parameter_357: class Program_weight_tensor_parameter_358: name = "parameter_358" + original_name = "linear_1397.b_0" shape = [1024] dtype = "float32" min_val = float("-0.471326") @@ -3949,6 +4308,7 @@ class Program_weight_tensor_parameter_358: class Program_weight_tensor_parameter_359: name = "parameter_359" + original_name = "linear_1397.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-0.48177") @@ -3960,6 +4320,7 @@ class Program_weight_tensor_parameter_359: class Program_weight_tensor_parameter_360: name = "parameter_360" + original_name = "linear_1396.b_0" shape = [4096] dtype = "float32" min_val = float("-0.133461") @@ -3971,6 +4332,7 @@ class Program_weight_tensor_parameter_360: class Program_weight_tensor_parameter_361: name = "parameter_361" + original_name = "linear_1396.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.2097") @@ -3982,6 +4344,7 @@ class Program_weight_tensor_parameter_361: class Program_weight_tensor_parameter_362: name = "parameter_362" + original_name = "linear_1395.b_0" shape = [1024] dtype = "float32" min_val = float("-0.162914") @@ -3993,6 +4356,7 @@ class Program_weight_tensor_parameter_362: class Program_weight_tensor_parameter_363: name = "parameter_363" + original_name = "linear_1395.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.792352") @@ -4004,6 +4368,7 @@ class Program_weight_tensor_parameter_363: class Program_weight_tensor_parameter_364: name = "parameter_364" + original_name = "linear_1394.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0514717") @@ -4015,6 +4380,7 @@ class Program_weight_tensor_parameter_364: class Program_weight_tensor_parameter_365: name = "parameter_365" + original_name = "linear_1394.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.142657") @@ -4026,6 +4392,7 @@ class Program_weight_tensor_parameter_365: class Program_weight_tensor_parameter_366: name = "parameter_366" + original_name = "linear_1393.b_0" shape = [1024] dtype = "float32" min_val = float("-0.327808") @@ -4037,6 +4404,7 @@ class Program_weight_tensor_parameter_366: class Program_weight_tensor_parameter_367: name = "parameter_367" + original_name = "linear_1393.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.443822") @@ -4048,6 +4416,7 @@ class Program_weight_tensor_parameter_367: class Program_weight_tensor_parameter_368: name = "parameter_368" + original_name = "linear_1392.b_0" shape = [1024] dtype = "float32" min_val = float("-0.374483") @@ -4059,6 +4428,7 @@ class Program_weight_tensor_parameter_368: class Program_weight_tensor_parameter_369: name = "parameter_369" + original_name = "linear_1392.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.327932") @@ -4070,6 +4440,7 @@ class Program_weight_tensor_parameter_369: class Program_weight_tensor_parameter_370: name = "parameter_370" + original_name = "layer_norm_480.b_0" shape = [1024] dtype = "float32" min_val = float("-0.378744") @@ -4081,6 +4452,7 @@ class Program_weight_tensor_parameter_370: class Program_weight_tensor_parameter_371: name = "parameter_371" + original_name = "layer_norm_480.w_0" shape = [1024] dtype = "float32" min_val = float("0.190282") @@ -4092,6 +4464,7 @@ class Program_weight_tensor_parameter_371: class Program_weight_tensor_parameter_372: name = "parameter_372" + original_name = "layer_norm_479.b_0" shape = [1024] dtype = "float32" min_val = float("-2.64098") @@ -4103,6 +4476,7 @@ class Program_weight_tensor_parameter_372: class Program_weight_tensor_parameter_373: name = "parameter_373" + original_name = "layer_norm_479.w_0" shape = [1024] dtype = "float32" min_val = float("0.809092") @@ -4114,6 +4488,7 @@ class Program_weight_tensor_parameter_373: class Program_weight_tensor_parameter_374: name = "parameter_374" + original_name = "linear_1391.b_0" shape = [1024] dtype = "float32" min_val = float("-0.436836") @@ -4125,6 +4500,7 @@ class Program_weight_tensor_parameter_374: class Program_weight_tensor_parameter_375: name = "parameter_375" + original_name = "linear_1391.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-0.728629") @@ -4136,6 +4512,7 @@ class Program_weight_tensor_parameter_375: class Program_weight_tensor_parameter_376: name = "parameter_376" + original_name = "linear_1390.b_0" shape = [4096] dtype = "float32" min_val = float("-0.223193") @@ -4147,6 +4524,7 @@ class Program_weight_tensor_parameter_376: class Program_weight_tensor_parameter_377: name = "parameter_377" + original_name = "linear_1390.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.425997") @@ -4158,6 +4536,7 @@ class Program_weight_tensor_parameter_377: class Program_weight_tensor_parameter_378: name = "parameter_378" + original_name = "linear_1389.b_0" shape = [1024] dtype = "float32" min_val = float("-0.535619") @@ -4169,6 +4548,7 @@ class Program_weight_tensor_parameter_378: class Program_weight_tensor_parameter_379: name = "parameter_379" + original_name = "linear_1389.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-1.06757") @@ -4180,6 +4560,7 @@ class Program_weight_tensor_parameter_379: class Program_weight_tensor_parameter_380: name = "parameter_380" + original_name = "linear_1388.b_0" shape = [1024] dtype = "float32" min_val = float("-0.343598") @@ -4191,6 +4572,7 @@ class Program_weight_tensor_parameter_380: class Program_weight_tensor_parameter_381: name = "parameter_381" + original_name = "linear_1388.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.136848") @@ -4202,6 +4584,7 @@ class Program_weight_tensor_parameter_381: class Program_weight_tensor_parameter_382: name = "parameter_382" + original_name = "linear_1387.b_0" shape = [1024] dtype = "float32" min_val = float("-0.45063") @@ -4213,6 +4596,7 @@ class Program_weight_tensor_parameter_382: class Program_weight_tensor_parameter_383: name = "parameter_383" + original_name = "linear_1387.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.444643") @@ -4224,6 +4608,7 @@ class Program_weight_tensor_parameter_383: class Program_weight_tensor_parameter_384: name = "parameter_384" + original_name = "linear_1386.b_0" shape = [1024] dtype = "float32" min_val = float("-0.959396") @@ -4235,6 +4620,7 @@ class Program_weight_tensor_parameter_384: class Program_weight_tensor_parameter_385: name = "parameter_385" + original_name = "linear_1386.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.438839") @@ -4246,6 +4632,7 @@ class Program_weight_tensor_parameter_385: class Program_weight_tensor_parameter_386: name = "parameter_386" + original_name = "layer_norm_478.b_0" shape = [1024] dtype = "float32" min_val = float("-0.574758") @@ -4257,6 +4644,7 @@ class Program_weight_tensor_parameter_386: class Program_weight_tensor_parameter_387: name = "parameter_387" + original_name = "layer_norm_478.w_0" shape = [1024] dtype = "float32" min_val = float("0.136694") @@ -4268,6 +4656,7 @@ class Program_weight_tensor_parameter_387: class Program_weight_tensor_parameter_388: name = "parameter_388" + original_name = "embedding_86.w_0" shape = [4, 1024] dtype = "float32" min_val = float("-0.174712") @@ -4279,6 +4668,7 @@ class Program_weight_tensor_parameter_388: class Program_weight_tensor_parameter_389: name = "parameter_389" + original_name = "embedding_85.w_0" shape = [512, 1024] dtype = "float32" min_val = float("-0.304575") @@ -4290,6 +4680,7 @@ class Program_weight_tensor_parameter_389: class Program_weight_tensor_parameter_390: name = "parameter_390" + original_name = "embedding_84.w_0" shape = [30522, 1024] dtype = "float32" min_val = float("-0.742746") diff --git a/paddle_samples/PaddleNLP/ernie-tiny/input_meta.py b/paddle_samples/PaddleNLP/ernie-tiny/input_meta.py index a32169d71..fcfb9adfd 100644 --- a/paddle_samples/PaddleNLP/ernie-tiny/input_meta.py +++ b/paddle_samples/PaddleNLP/ernie-tiny/input_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_data_0: name = "data_0" + original_name = "input_ids" shape = [1, 20] dtype = "int64" data = [ @@ -28,6 +29,7 @@ class Program_weight_tensor_data_0: class Program_weight_tensor_data_1: name = "data_1" + original_name = "token_type_ids" shape = [1, 20] dtype = "int64" data = [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0] diff --git a/paddle_samples/PaddleNLP/ernie-tiny/model.py b/paddle_samples/PaddleNLP/ernie-tiny/model.py index 8833237cb..629afad6b 100644 --- a/paddle_samples/PaddleNLP/ernie-tiny/model.py +++ b/paddle_samples/PaddleNLP/ernie-tiny/model.py @@ -121,10 +121,11 @@ def forward( # pd_op.embedding: (1x20x1024xf32) <- (1x20xi64, 600x1024xf32) embedding_1 = paddle._C_ops.embedding(subtract_0, parameter_53, -1, False) - del parameter_53 + del parameter_53, subtract_0 # pd_op.add: (1x20x1024xf32) <- (1x20x1024xf32, 1x20x1024xf32) add_0 = paddle._C_ops.add(embedding_0, embedding_1) + del embedding_0, embedding_1 # pd_op.embedding: (1x20x1024xf32) <- (1x20xi64, 2x1024xf32) embedding_2 = paddle._C_ops.embedding(data_1, parameter_52, -1, False) @@ -132,6 +133,7 @@ def forward( # pd_op.add: (1x20x1024xf32) <- (1x20x1024xf32, 1x20x1024xf32) add_1 = paddle._C_ops.add(add_0, embedding_2) + del add_0, embedding_2 # pd_op.layer_norm: (1x20x1024xf32, 1x20xf32, 1x20xf32) <- (1x20x1024xf32, 1024xf32, 1024xf32) layer_norm_0, layer_norm_1, layer_norm_2 = (lambda x, f: f(x))( @@ -140,44 +142,17 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_50, parameter_51 + del add_1, parameter_50, parameter_51 # pd_op.full: (1xf32) <- () full_4 = paddle._C_ops.full( [1], float("0.1"), paddle.float32, paddle.core.CPUPlace() ) - # pd_op.assign: (1xf32) <- (1xf32) - assign_0 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_1 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_2 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_3 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_4 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_5 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_6 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_7 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_8 = full_4 - # pd_op.dropout: (1x20x1024xf32, 1x20x1024xui8) <- (1x20x1024xf32, None, 1xf32) dropout_0, dropout_1 = (lambda x, f: f(x))( paddle._C_ops.dropout( - layer_norm_0, None, full_4, False, "upscale_in_train", 0, False + layer_norm_0, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -189,13 +164,14 @@ def forward( # pd_op.add: (1x20x1024xf32) <- (1x20x1024xf32, 1024xf32) add_2 = paddle._C_ops.add(matmul_0, parameter_48) - del parameter_48 + del matmul_0, parameter_48 # pd_op.full_int_array: (4xi64) <- () full_int_array_1 = [0, 0, 16, 64] # pd_op.reshape: (1x20x16x64xf32) <- (1x20x1024xf32, 4xi64) reshape_0 = paddle._C_ops.reshape(add_2, full_int_array_1) + del add_2 # pd_op.transpose: (1x16x20x64xf32) <- (1x20x16x64xf32) transpose_0 = paddle._C_ops.transpose(reshape_0, [0, 2, 1, 3]) @@ -207,7 +183,7 @@ def forward( # pd_op.add: (1x20x1024xf32) <- (1x20x1024xf32, 1024xf32) add_3 = paddle._C_ops.add(matmul_1, parameter_46) - del parameter_46 + del matmul_1, parameter_46 # pd_op.matmul: (1x20x1024xf32) <- (1x20x1024xf32, 1024x1024xf32) matmul_2 = paddle._C_ops.matmul(dropout_0, parameter_45, False, False) @@ -215,10 +191,11 @@ def forward( # pd_op.add: (1x20x1024xf32) <- (1x20x1024xf32, 1024xf32) add_4 = paddle._C_ops.add(matmul_2, parameter_44) - del parameter_44 + del matmul_2, parameter_44 # pd_op.reshape: (1x20x16x64xf32) <- (1x20x1024xf32, 4xi64) reshape_1 = paddle._C_ops.reshape(add_3, full_int_array_1) + del add_3 # pd_op.transpose: (1x16x20x64xf32) <- (1x20x16x64xf32) transpose_1 = paddle._C_ops.transpose(reshape_1, [0, 2, 1, 3]) @@ -226,6 +203,7 @@ def forward( # pd_op.reshape: (1x20x16x64xf32) <- (1x20x1024xf32, 4xi64) reshape_2 = paddle._C_ops.reshape(add_4, full_int_array_1) + del add_4 # pd_op.transpose: (1x16x20x64xf32) <- (1x20x16x64xf32) transpose_2 = paddle._C_ops.transpose(reshape_2, [0, 2, 1, 3]) @@ -236,21 +214,17 @@ def forward( [1], float("0.125"), paddle.float32, paddle.core.CPUPlace() ) - # pd_op.assign: (1xf32) <- (1xf32) - assign_9 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_10 = full_5 - # pd_op.scale: (1x16x20x64xf32) <- (1x16x20x64xf32, 1xf32) scale_1 = paddle._C_ops.scale(transpose_0, full_5, float("0"), True) del transpose_0 # pd_op.matmul: (1x16x20x20xf32) <- (1x16x20x64xf32, 1x16x20x64xf32) matmul_3 = paddle._C_ops.matmul(scale_1, transpose_1, False, True) + del scale_1, transpose_1 # pd_op.add: (1x16x20x20xf32) <- (1x16x20x20xf32, 1x1x1x20xf32) add_5 = paddle._C_ops.add(matmul_3, unsqueeze_0) + del matmul_3 # pd_op.softmax: (1x16x20x20xf32) <- (1x16x20x20xf32) softmax_0 = paddle._C_ops.softmax(add_5, -1) @@ -259,13 +233,15 @@ def forward( # pd_op.dropout: (1x16x20x20xf32, 1x16x20x20xui8) <- (1x16x20x20xf32, None, 1xf32) dropout_2, dropout_3 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_0, None, full_4, False, "upscale_in_train", 0, False + softmax_0, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_0 # pd_op.matmul: (1x16x20x64xf32) <- (1x16x20x20xf32, 1x16x20x64xf32) matmul_4 = paddle._C_ops.matmul(dropout_2, transpose_2, False, False) + del dropout_2, transpose_2 # pd_op.transpose: (1x20x16x64xf32) <- (1x16x20x64xf32) transpose_3 = paddle._C_ops.transpose(matmul_4, [0, 2, 1, 3]) @@ -276,19 +252,20 @@ def forward( # pd_op.reshape: (1x20x1024xf32) <- (1x20x16x64xf32, 3xi64) reshape_3 = paddle._C_ops.reshape(transpose_3, full_int_array_2) + del transpose_3 # pd_op.matmul: (1x20x1024xf32) <- (1x20x1024xf32, 1024x1024xf32) matmul_5 = paddle._C_ops.matmul(reshape_3, parameter_43, False, False) - del parameter_43 + del parameter_43, reshape_3 # pd_op.add: (1x20x1024xf32) <- (1x20x1024xf32, 1024xf32) add_6 = paddle._C_ops.add(matmul_5, parameter_42) - del parameter_42 + del matmul_5, parameter_42 # pd_op.dropout: (1x20x1024xf32, 1x20x1024xui8) <- (1x20x1024xf32, None, 1xf32) dropout_4, dropout_5 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_6, None, full_4, False, "upscale_in_train", 0, False + add_6, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -296,6 +273,7 @@ def forward( # pd_op.add: (1x20x1024xf32) <- (1x20x1024xf32, 1x20x1024xf32) add_7 = paddle._C_ops.add(dropout_0, dropout_4) + del dropout_0, dropout_4 # pd_op.layer_norm: (1x20x1024xf32, 1x20xf32, 1x20xf32) <- (1x20x1024xf32, 1024xf32, 1024xf32) layer_norm_3, layer_norm_4, layer_norm_5 = (lambda x, f: f(x))( @@ -304,7 +282,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_36, parameter_37 + del add_7, parameter_36, parameter_37 # pd_op.matmul: (1x20x4096xf32) <- (1x20x1024xf32, 1024x4096xf32) matmul_6 = paddle._C_ops.matmul(layer_norm_3, parameter_41, False, False) @@ -312,7 +290,7 @@ def forward( # pd_op.add: (1x20x4096xf32) <- (1x20x4096xf32, 4096xf32) add_8 = paddle._C_ops.add(matmul_6, parameter_40) - del parameter_40 + del matmul_6, parameter_40 # pd_op.relu: (1x20x4096xf32) <- (1x20x4096xf32) relu_0 = paddle._C_ops.relu(add_8) @@ -320,16 +298,16 @@ def forward( # pd_op.matmul: (1x20x1024xf32) <- (1x20x4096xf32, 4096x1024xf32) matmul_7 = paddle._C_ops.matmul(relu_0, parameter_39, False, False) - del parameter_39 + del parameter_39, relu_0 # pd_op.add: (1x20x1024xf32) <- (1x20x1024xf32, 1024xf32) add_9 = paddle._C_ops.add(matmul_7, parameter_38) - del parameter_38 + del matmul_7, parameter_38 # pd_op.dropout: (1x20x1024xf32, 1x20x1024xui8) <- (1x20x1024xf32, None, 1xf32) dropout_6, dropout_7 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_9, None, full_4, False, "upscale_in_train", 0, False + add_9, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -337,6 +315,7 @@ def forward( # pd_op.add: (1x20x1024xf32) <- (1x20x1024xf32, 1x20x1024xf32) add_10 = paddle._C_ops.add(layer_norm_3, dropout_6) + del dropout_6, layer_norm_3 # pd_op.layer_norm: (1x20x1024xf32, 1x20xf32, 1x20xf32) <- (1x20x1024xf32, 1024xf32, 1024xf32) layer_norm_6, layer_norm_7, layer_norm_8 = (lambda x, f: f(x))( @@ -345,7 +324,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_34, parameter_35 + del add_10, parameter_34, parameter_35 # pd_op.matmul: (1x20x1024xf32) <- (1x20x1024xf32, 1024x1024xf32) matmul_8 = paddle._C_ops.matmul(layer_norm_6, parameter_33, False, False) @@ -353,10 +332,11 @@ def forward( # pd_op.add: (1x20x1024xf32) <- (1x20x1024xf32, 1024xf32) add_11 = paddle._C_ops.add(matmul_8, parameter_32) - del parameter_32 + del matmul_8, parameter_32 # pd_op.reshape: (1x20x16x64xf32) <- (1x20x1024xf32, 4xi64) reshape_4 = paddle._C_ops.reshape(add_11, full_int_array_1) + del add_11 # pd_op.transpose: (1x16x20x64xf32) <- (1x20x16x64xf32) transpose_4 = paddle._C_ops.transpose(reshape_4, [0, 2, 1, 3]) @@ -368,7 +348,7 @@ def forward( # pd_op.add: (1x20x1024xf32) <- (1x20x1024xf32, 1024xf32) add_12 = paddle._C_ops.add(matmul_9, parameter_30) - del parameter_30 + del matmul_9, parameter_30 # pd_op.matmul: (1x20x1024xf32) <- (1x20x1024xf32, 1024x1024xf32) matmul_10 = paddle._C_ops.matmul(layer_norm_6, parameter_29, False, False) @@ -376,10 +356,11 @@ def forward( # pd_op.add: (1x20x1024xf32) <- (1x20x1024xf32, 1024xf32) add_13 = paddle._C_ops.add(matmul_10, parameter_28) - del parameter_28 + del matmul_10, parameter_28 # pd_op.reshape: (1x20x16x64xf32) <- (1x20x1024xf32, 4xi64) reshape_5 = paddle._C_ops.reshape(add_12, full_int_array_1) + del add_12 # pd_op.transpose: (1x16x20x64xf32) <- (1x20x16x64xf32) transpose_5 = paddle._C_ops.transpose(reshape_5, [0, 2, 1, 3]) @@ -387,6 +368,7 @@ def forward( # pd_op.reshape: (1x20x16x64xf32) <- (1x20x1024xf32, 4xi64) reshape_6 = paddle._C_ops.reshape(add_13, full_int_array_1) + del add_13 # pd_op.transpose: (1x16x20x64xf32) <- (1x20x16x64xf32) transpose_6 = paddle._C_ops.transpose(reshape_6, [0, 2, 1, 3]) @@ -398,9 +380,11 @@ def forward( # pd_op.matmul: (1x16x20x20xf32) <- (1x16x20x64xf32, 1x16x20x64xf32) matmul_11 = paddle._C_ops.matmul(scale_2, transpose_5, False, True) + del scale_2, transpose_5 # pd_op.add: (1x16x20x20xf32) <- (1x16x20x20xf32, 1x1x1x20xf32) add_14 = paddle._C_ops.add(matmul_11, unsqueeze_0) + del matmul_11 # pd_op.softmax: (1x16x20x20xf32) <- (1x16x20x20xf32) softmax_1 = paddle._C_ops.softmax(add_14, -1) @@ -409,13 +393,15 @@ def forward( # pd_op.dropout: (1x16x20x20xf32, 1x16x20x20xui8) <- (1x16x20x20xf32, None, 1xf32) dropout_8, dropout_9 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_1, None, full_4, False, "upscale_in_train", 0, False + softmax_1, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_1 # pd_op.matmul: (1x16x20x64xf32) <- (1x16x20x20xf32, 1x16x20x64xf32) matmul_12 = paddle._C_ops.matmul(dropout_8, transpose_6, False, False) + del dropout_8, transpose_6 # pd_op.transpose: (1x20x16x64xf32) <- (1x16x20x64xf32) transpose_7 = paddle._C_ops.transpose(matmul_12, [0, 2, 1, 3]) @@ -423,19 +409,20 @@ def forward( # pd_op.reshape: (1x20x1024xf32) <- (1x20x16x64xf32, 3xi64) reshape_7 = paddle._C_ops.reshape(transpose_7, full_int_array_2) + del transpose_7 # pd_op.matmul: (1x20x1024xf32) <- (1x20x1024xf32, 1024x1024xf32) matmul_13 = paddle._C_ops.matmul(reshape_7, parameter_27, False, False) - del parameter_27 + del parameter_27, reshape_7 # pd_op.add: (1x20x1024xf32) <- (1x20x1024xf32, 1024xf32) add_15 = paddle._C_ops.add(matmul_13, parameter_26) - del parameter_26 + del matmul_13, parameter_26 # pd_op.dropout: (1x20x1024xf32, 1x20x1024xui8) <- (1x20x1024xf32, None, 1xf32) dropout_10, dropout_11 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_15, None, full_4, False, "upscale_in_train", 0, False + add_15, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -443,6 +430,7 @@ def forward( # pd_op.add: (1x20x1024xf32) <- (1x20x1024xf32, 1x20x1024xf32) add_16 = paddle._C_ops.add(layer_norm_6, dropout_10) + del dropout_10, layer_norm_6 # pd_op.layer_norm: (1x20x1024xf32, 1x20xf32, 1x20xf32) <- (1x20x1024xf32, 1024xf32, 1024xf32) layer_norm_9, layer_norm_10, layer_norm_11 = (lambda x, f: f(x))( @@ -451,7 +439,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_20, parameter_21 + del add_16, parameter_20, parameter_21 # pd_op.matmul: (1x20x4096xf32) <- (1x20x1024xf32, 1024x4096xf32) matmul_14 = paddle._C_ops.matmul(layer_norm_9, parameter_25, False, False) @@ -459,7 +447,7 @@ def forward( # pd_op.add: (1x20x4096xf32) <- (1x20x4096xf32, 4096xf32) add_17 = paddle._C_ops.add(matmul_14, parameter_24) - del parameter_24 + del matmul_14, parameter_24 # pd_op.relu: (1x20x4096xf32) <- (1x20x4096xf32) relu_1 = paddle._C_ops.relu(add_17) @@ -467,16 +455,16 @@ def forward( # pd_op.matmul: (1x20x1024xf32) <- (1x20x4096xf32, 4096x1024xf32) matmul_15 = paddle._C_ops.matmul(relu_1, parameter_23, False, False) - del parameter_23 + del parameter_23, relu_1 # pd_op.add: (1x20x1024xf32) <- (1x20x1024xf32, 1024xf32) add_18 = paddle._C_ops.add(matmul_15, parameter_22) - del parameter_22 + del matmul_15, parameter_22 # pd_op.dropout: (1x20x1024xf32, 1x20x1024xui8) <- (1x20x1024xf32, None, 1xf32) dropout_12, dropout_13 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_18, None, full_4, False, "upscale_in_train", 0, False + add_18, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -484,6 +472,7 @@ def forward( # pd_op.add: (1x20x1024xf32) <- (1x20x1024xf32, 1x20x1024xf32) add_19 = paddle._C_ops.add(layer_norm_9, dropout_12) + del dropout_12, layer_norm_9 # pd_op.layer_norm: (1x20x1024xf32, 1x20xf32, 1x20xf32) <- (1x20x1024xf32, 1024xf32, 1024xf32) layer_norm_12, layer_norm_13, layer_norm_14 = (lambda x, f: f(x))( @@ -492,7 +481,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_18, parameter_19 + del add_19, parameter_18, parameter_19 # pd_op.matmul: (1x20x1024xf32) <- (1x20x1024xf32, 1024x1024xf32) matmul_16 = paddle._C_ops.matmul(layer_norm_12, parameter_17, False, False) @@ -500,10 +489,11 @@ def forward( # pd_op.add: (1x20x1024xf32) <- (1x20x1024xf32, 1024xf32) add_20 = paddle._C_ops.add(matmul_16, parameter_16) - del parameter_16 + del matmul_16, parameter_16 # pd_op.reshape: (1x20x16x64xf32) <- (1x20x1024xf32, 4xi64) reshape_8 = paddle._C_ops.reshape(add_20, full_int_array_1) + del add_20 # pd_op.transpose: (1x16x20x64xf32) <- (1x20x16x64xf32) transpose_8 = paddle._C_ops.transpose(reshape_8, [0, 2, 1, 3]) @@ -515,7 +505,7 @@ def forward( # pd_op.add: (1x20x1024xf32) <- (1x20x1024xf32, 1024xf32) add_21 = paddle._C_ops.add(matmul_17, parameter_14) - del parameter_14 + del matmul_17, parameter_14 # pd_op.matmul: (1x20x1024xf32) <- (1x20x1024xf32, 1024x1024xf32) matmul_18 = paddle._C_ops.matmul(layer_norm_12, parameter_13, False, False) @@ -523,10 +513,11 @@ def forward( # pd_op.add: (1x20x1024xf32) <- (1x20x1024xf32, 1024xf32) add_22 = paddle._C_ops.add(matmul_18, parameter_12) - del parameter_12 + del matmul_18, parameter_12 # pd_op.reshape: (1x20x16x64xf32) <- (1x20x1024xf32, 4xi64) reshape_9 = paddle._C_ops.reshape(add_21, full_int_array_1) + del add_21 # pd_op.transpose: (1x16x20x64xf32) <- (1x20x16x64xf32) transpose_9 = paddle._C_ops.transpose(reshape_9, [0, 2, 1, 3]) @@ -534,7 +525,7 @@ def forward( # pd_op.reshape: (1x20x16x64xf32) <- (1x20x1024xf32, 4xi64) reshape_10 = paddle._C_ops.reshape(add_22, full_int_array_1) - del full_int_array_1 + del add_22, full_int_array_1 # pd_op.transpose: (1x16x20x64xf32) <- (1x20x16x64xf32) transpose_10 = paddle._C_ops.transpose(reshape_10, [0, 2, 1, 3]) @@ -542,13 +533,15 @@ def forward( # pd_op.scale: (1x16x20x64xf32) <- (1x16x20x64xf32, 1xf32) scale_3 = paddle._C_ops.scale(transpose_8, full_5, float("0"), True) - del transpose_8 + del full_5, transpose_8 # pd_op.matmul: (1x16x20x20xf32) <- (1x16x20x64xf32, 1x16x20x64xf32) matmul_19 = paddle._C_ops.matmul(scale_3, transpose_9, False, True) + del scale_3, transpose_9 # pd_op.add: (1x16x20x20xf32) <- (1x16x20x20xf32, 1x1x1x20xf32) add_23 = paddle._C_ops.add(matmul_19, unsqueeze_0) + del matmul_19, unsqueeze_0 # pd_op.softmax: (1x16x20x20xf32) <- (1x16x20x20xf32) softmax_2 = paddle._C_ops.softmax(add_23, -1) @@ -557,13 +550,15 @@ def forward( # pd_op.dropout: (1x16x20x20xf32, 1x16x20x20xui8) <- (1x16x20x20xf32, None, 1xf32) dropout_14, dropout_15 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_2, None, full_4, False, "upscale_in_train", 0, False + softmax_2, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_2 # pd_op.matmul: (1x16x20x64xf32) <- (1x16x20x20xf32, 1x16x20x64xf32) matmul_20 = paddle._C_ops.matmul(dropout_14, transpose_10, False, False) + del dropout_14, transpose_10 # pd_op.transpose: (1x20x16x64xf32) <- (1x16x20x64xf32) transpose_11 = paddle._C_ops.transpose(matmul_20, [0, 2, 1, 3]) @@ -571,20 +566,20 @@ def forward( # pd_op.reshape: (1x20x1024xf32) <- (1x20x16x64xf32, 3xi64) reshape_11 = paddle._C_ops.reshape(transpose_11, full_int_array_2) - del full_int_array_2 + del full_int_array_2, transpose_11 # pd_op.matmul: (1x20x1024xf32) <- (1x20x1024xf32, 1024x1024xf32) matmul_21 = paddle._C_ops.matmul(reshape_11, parameter_11, False, False) - del parameter_11 + del parameter_11, reshape_11 # pd_op.add: (1x20x1024xf32) <- (1x20x1024xf32, 1024xf32) add_24 = paddle._C_ops.add(matmul_21, parameter_10) - del parameter_10 + del matmul_21, parameter_10 # pd_op.dropout: (1x20x1024xf32, 1x20x1024xui8) <- (1x20x1024xf32, None, 1xf32) dropout_16, dropout_17 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_24, None, full_4, False, "upscale_in_train", 0, False + add_24, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -592,6 +587,7 @@ def forward( # pd_op.add: (1x20x1024xf32) <- (1x20x1024xf32, 1x20x1024xf32) add_25 = paddle._C_ops.add(layer_norm_12, dropout_16) + del dropout_16, layer_norm_12 # pd_op.layer_norm: (1x20x1024xf32, 1x20xf32, 1x20xf32) <- (1x20x1024xf32, 1024xf32, 1024xf32) layer_norm_15, layer_norm_16, layer_norm_17 = (lambda x, f: f(x))( @@ -600,7 +596,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_4, parameter_5 + del add_25, parameter_4, parameter_5 # pd_op.matmul: (1x20x4096xf32) <- (1x20x1024xf32, 1024x4096xf32) matmul_22 = paddle._C_ops.matmul(layer_norm_15, parameter_9, False, False) @@ -608,7 +604,7 @@ def forward( # pd_op.add: (1x20x4096xf32) <- (1x20x4096xf32, 4096xf32) add_26 = paddle._C_ops.add(matmul_22, parameter_8) - del parameter_8 + del matmul_22, parameter_8 # pd_op.relu: (1x20x4096xf32) <- (1x20x4096xf32) relu_2 = paddle._C_ops.relu(add_26) @@ -616,23 +612,24 @@ def forward( # pd_op.matmul: (1x20x1024xf32) <- (1x20x4096xf32, 4096x1024xf32) matmul_23 = paddle._C_ops.matmul(relu_2, parameter_7, False, False) - del parameter_7 + del parameter_7, relu_2 # pd_op.add: (1x20x1024xf32) <- (1x20x1024xf32, 1024xf32) add_27 = paddle._C_ops.add(matmul_23, parameter_6) - del parameter_6 + del matmul_23, parameter_6 # pd_op.dropout: (1x20x1024xf32, 1x20x1024xui8) <- (1x20x1024xf32, None, 1xf32) dropout_18, dropout_19 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_27, None, full_4, False, "upscale_in_train", 0, False + add_27, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) - del add_27 + del add_27, full_4 # pd_op.add: (1x20x1024xf32) <- (1x20x1024xf32, 1x20x1024xf32) add_28 = paddle._C_ops.add(layer_norm_15, dropout_18) + del dropout_18, layer_norm_15 # pd_op.layer_norm: (1x20x1024xf32, 1x20xf32, 1x20xf32) <- (1x20x1024xf32, 1024xf32, 1024xf32) layer_norm_18, layer_norm_19, layer_norm_20 = (lambda x, f: f(x))( @@ -641,7 +638,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_2, parameter_3 + del add_28, parameter_2, parameter_3 # pd_op.full_int_array: (1xi64) <- () full_int_array_3 = [0] @@ -653,140 +650,18 @@ def forward( slice_0 = paddle._C_ops.slice( layer_norm_18, [1], full_int_array_3, full_int_array_4, [1], [1] ) + del full_int_array_3, full_int_array_4, layer_norm_18 # pd_op.matmul: (1x1024xf32) <- (1x1024xf32, 1024x1024xf32) matmul_24 = paddle._C_ops.matmul(slice_0, parameter_1, False, False) - del parameter_1 + del parameter_1, slice_0 # pd_op.add: (1x1024xf32) <- (1x1024xf32, 1024xf32) add_29 = paddle._C_ops.add(matmul_24, parameter_0) - del parameter_0 + del matmul_24, parameter_0 # pd_op.tanh: (1x1024xf32) <- (1x1024xf32) tanh_0 = paddle._C_ops.tanh(add_29) - del ( - add_0, - add_1, - add_10, - add_11, - add_12, - add_13, - add_16, - add_19, - add_2, - add_20, - add_21, - add_22, - add_25, - add_28, - add_29, - add_3, - add_4, - add_7, - assign_0, - assign_1, - assign_10, - assign_2, - assign_3, - assign_4, - assign_5, - assign_6, - assign_7, - assign_8, - assign_9, - dropout_0, - dropout_1, - dropout_10, - dropout_11, - dropout_12, - dropout_13, - dropout_14, - dropout_15, - dropout_16, - dropout_17, - dropout_18, - dropout_19, - dropout_2, - dropout_3, - dropout_4, - dropout_5, - dropout_6, - dropout_7, - dropout_8, - dropout_9, - embedding_0, - embedding_1, - embedding_2, - full_4, - full_5, - full_int_array_3, - full_int_array_4, - layer_norm_1, - layer_norm_10, - layer_norm_11, - layer_norm_12, - layer_norm_13, - layer_norm_14, - layer_norm_15, - layer_norm_16, - layer_norm_17, - layer_norm_18, - layer_norm_19, - layer_norm_2, - layer_norm_20, - layer_norm_3, - layer_norm_4, - layer_norm_5, - layer_norm_6, - layer_norm_7, - layer_norm_8, - layer_norm_9, - matmul_0, - matmul_1, - matmul_10, - matmul_11, - matmul_13, - matmul_14, - matmul_15, - matmul_16, - matmul_17, - matmul_18, - matmul_19, - matmul_2, - matmul_21, - matmul_22, - matmul_23, - matmul_24, - matmul_3, - matmul_5, - matmul_6, - matmul_7, - matmul_8, - matmul_9, - relu_0, - relu_1, - relu_2, - reshape_11, - reshape_3, - reshape_7, - scale_1, - scale_2, - scale_3, - slice_0, - softmax_0, - softmax_1, - softmax_2, - subtract_0, - transpose_1, - transpose_10, - transpose_11, - transpose_2, - transpose_3, - transpose_5, - transpose_6, - transpose_7, - transpose_9, - unsqueeze_0, - ) + del add_29 return tanh_0 diff --git a/paddle_samples/PaddleNLP/ernie-tiny/weight_meta.py b/paddle_samples/PaddleNLP/ernie-tiny/weight_meta.py index f67b0e9dc..52dbdb569 100644 --- a/paddle_samples/PaddleNLP/ernie-tiny/weight_meta.py +++ b/paddle_samples/PaddleNLP/ernie-tiny/weight_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_parameter_0: name = "parameter_0" + original_name = "linear_1549.b_0" shape = [1024] dtype = "float32" min_val = float("-0.111116") @@ -11,6 +12,7 @@ class Program_weight_tensor_parameter_0: class Program_weight_tensor_parameter_1: name = "parameter_1" + original_name = "linear_1549.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.344934") @@ -22,6 +24,7 @@ class Program_weight_tensor_parameter_1: class Program_weight_tensor_parameter_2: name = "parameter_2" + original_name = "layer_norm_533.b_0" shape = [1024] dtype = "float32" min_val = float("-1.38487") @@ -33,6 +36,7 @@ class Program_weight_tensor_parameter_2: class Program_weight_tensor_parameter_3: name = "parameter_3" + original_name = "layer_norm_533.w_0" shape = [1024] dtype = "float32" min_val = float("0.741124") @@ -44,6 +48,7 @@ class Program_weight_tensor_parameter_3: class Program_weight_tensor_parameter_4: name = "parameter_4" + original_name = "layer_norm_532.b_0" shape = [1024] dtype = "float32" min_val = float("-1.10527") @@ -55,6 +60,7 @@ class Program_weight_tensor_parameter_4: class Program_weight_tensor_parameter_5: name = "parameter_5" + original_name = "layer_norm_532.w_0" shape = [1024] dtype = "float32" min_val = float("0.108216") @@ -66,6 +72,7 @@ class Program_weight_tensor_parameter_5: class Program_weight_tensor_parameter_6: name = "parameter_6" + original_name = "linear_1548.b_0" shape = [1024] dtype = "float32" min_val = float("-0.55382") @@ -77,6 +84,7 @@ class Program_weight_tensor_parameter_6: class Program_weight_tensor_parameter_7: name = "parameter_7" + original_name = "linear_1548.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-1.65767") @@ -88,6 +96,7 @@ class Program_weight_tensor_parameter_7: class Program_weight_tensor_parameter_8: name = "parameter_8" + original_name = "linear_1547.b_0" shape = [4096] dtype = "float32" min_val = float("-2.09817") @@ -99,6 +108,7 @@ class Program_weight_tensor_parameter_8: class Program_weight_tensor_parameter_9: name = "parameter_9" + original_name = "linear_1547.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.739874") @@ -110,6 +120,7 @@ class Program_weight_tensor_parameter_9: class Program_weight_tensor_parameter_10: name = "parameter_10" + original_name = "linear_1546.b_0" shape = [1024] dtype = "float32" min_val = float("-0.28318") @@ -121,6 +132,7 @@ class Program_weight_tensor_parameter_10: class Program_weight_tensor_parameter_11: name = "parameter_11" + original_name = "linear_1546.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-1.01922") @@ -132,6 +144,7 @@ class Program_weight_tensor_parameter_11: class Program_weight_tensor_parameter_12: name = "parameter_12" + original_name = "linear_1545.b_0" shape = [1024] dtype = "float32" min_val = float("-0.486043") @@ -143,6 +156,7 @@ class Program_weight_tensor_parameter_12: class Program_weight_tensor_parameter_13: name = "parameter_13" + original_name = "linear_1545.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.587558") @@ -154,6 +168,7 @@ class Program_weight_tensor_parameter_13: class Program_weight_tensor_parameter_14: name = "parameter_14" + original_name = "linear_1544.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0843108") @@ -165,6 +180,7 @@ class Program_weight_tensor_parameter_14: class Program_weight_tensor_parameter_15: name = "parameter_15" + original_name = "linear_1544.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-1.33106") @@ -176,6 +192,7 @@ class Program_weight_tensor_parameter_15: class Program_weight_tensor_parameter_16: name = "parameter_16" + original_name = "linear_1543.b_0" shape = [1024] dtype = "float32" min_val = float("-1.31924") @@ -187,6 +204,7 @@ class Program_weight_tensor_parameter_16: class Program_weight_tensor_parameter_17: name = "parameter_17" + original_name = "linear_1543.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.688811") @@ -198,6 +216,7 @@ class Program_weight_tensor_parameter_17: class Program_weight_tensor_parameter_18: name = "parameter_18" + original_name = "layer_norm_531.b_0" shape = [1024] dtype = "float32" min_val = float("-0.832218") @@ -209,6 +228,7 @@ class Program_weight_tensor_parameter_18: class Program_weight_tensor_parameter_19: name = "parameter_19" + original_name = "layer_norm_531.w_0" shape = [1024] dtype = "float32" min_val = float("0.0774964") @@ -220,6 +240,7 @@ class Program_weight_tensor_parameter_19: class Program_weight_tensor_parameter_20: name = "parameter_20" + original_name = "layer_norm_530.b_0" shape = [1024] dtype = "float32" min_val = float("-5.55319") @@ -231,6 +252,7 @@ class Program_weight_tensor_parameter_20: class Program_weight_tensor_parameter_21: name = "parameter_21" + original_name = "layer_norm_530.w_0" shape = [1024] dtype = "float32" min_val = float("0.0478255") @@ -242,6 +264,7 @@ class Program_weight_tensor_parameter_21: class Program_weight_tensor_parameter_22: name = "parameter_22" + original_name = "linear_1542.b_0" shape = [1024] dtype = "float32" min_val = float("-1.06506") @@ -253,6 +276,7 @@ class Program_weight_tensor_parameter_22: class Program_weight_tensor_parameter_23: name = "parameter_23" + original_name = "linear_1542.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-9.23634") @@ -264,6 +288,7 @@ class Program_weight_tensor_parameter_23: class Program_weight_tensor_parameter_24: name = "parameter_24" + original_name = "linear_1541.b_0" shape = [4096] dtype = "float32" min_val = float("-0.713154") @@ -275,6 +300,7 @@ class Program_weight_tensor_parameter_24: class Program_weight_tensor_parameter_25: name = "parameter_25" + original_name = "linear_1541.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-1.42869") @@ -286,6 +312,7 @@ class Program_weight_tensor_parameter_25: class Program_weight_tensor_parameter_26: name = "parameter_26" + original_name = "linear_1540.b_0" shape = [1024] dtype = "float32" min_val = float("-0.824935") @@ -297,6 +324,7 @@ class Program_weight_tensor_parameter_26: class Program_weight_tensor_parameter_27: name = "parameter_27" + original_name = "linear_1540.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.819416") @@ -308,6 +336,7 @@ class Program_weight_tensor_parameter_27: class Program_weight_tensor_parameter_28: name = "parameter_28" + original_name = "linear_1539.b_0" shape = [1024] dtype = "float32" min_val = float("-0.758419") @@ -319,6 +348,7 @@ class Program_weight_tensor_parameter_28: class Program_weight_tensor_parameter_29: name = "parameter_29" + original_name = "linear_1539.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.423836") @@ -330,6 +360,7 @@ class Program_weight_tensor_parameter_29: class Program_weight_tensor_parameter_30: name = "parameter_30" + original_name = "linear_1538.b_0" shape = [1024] dtype = "float32" min_val = float("-0.105964") @@ -341,6 +372,7 @@ class Program_weight_tensor_parameter_30: class Program_weight_tensor_parameter_31: name = "parameter_31" + original_name = "linear_1538.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-1.19267") @@ -352,6 +384,7 @@ class Program_weight_tensor_parameter_31: class Program_weight_tensor_parameter_32: name = "parameter_32" + original_name = "linear_1537.b_0" shape = [1024] dtype = "float32" min_val = float("-1.40342") @@ -363,6 +396,7 @@ class Program_weight_tensor_parameter_32: class Program_weight_tensor_parameter_33: name = "parameter_33" + original_name = "linear_1537.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.552888") @@ -374,6 +408,7 @@ class Program_weight_tensor_parameter_33: class Program_weight_tensor_parameter_34: name = "parameter_34" + original_name = "layer_norm_529.b_0" shape = [1024] dtype = "float32" min_val = float("-1.38775") @@ -385,6 +420,7 @@ class Program_weight_tensor_parameter_34: class Program_weight_tensor_parameter_35: name = "parameter_35" + original_name = "layer_norm_529.w_0" shape = [1024] dtype = "float32" min_val = float("0.0383121") @@ -396,6 +432,7 @@ class Program_weight_tensor_parameter_35: class Program_weight_tensor_parameter_36: name = "parameter_36" + original_name = "layer_norm_528.b_0" shape = [1024] dtype = "float32" min_val = float("-6.06155") @@ -407,6 +444,7 @@ class Program_weight_tensor_parameter_36: class Program_weight_tensor_parameter_37: name = "parameter_37" + original_name = "layer_norm_528.w_0" shape = [1024] dtype = "float32" min_val = float("-0.00693505") @@ -418,6 +456,7 @@ class Program_weight_tensor_parameter_37: class Program_weight_tensor_parameter_38: name = "parameter_38" + original_name = "linear_1536.b_0" shape = [1024] dtype = "float32" min_val = float("-1.1372") @@ -429,6 +468,7 @@ class Program_weight_tensor_parameter_38: class Program_weight_tensor_parameter_39: name = "parameter_39" + original_name = "linear_1536.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-6.275") @@ -440,6 +480,7 @@ class Program_weight_tensor_parameter_39: class Program_weight_tensor_parameter_40: name = "parameter_40" + original_name = "linear_1535.b_0" shape = [4096] dtype = "float32" min_val = float("-0.953142") @@ -451,6 +492,7 @@ class Program_weight_tensor_parameter_40: class Program_weight_tensor_parameter_41: name = "parameter_41" + original_name = "linear_1535.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-1.80421") @@ -462,6 +504,7 @@ class Program_weight_tensor_parameter_41: class Program_weight_tensor_parameter_42: name = "parameter_42" + original_name = "linear_1534.b_0" shape = [1024] dtype = "float32" min_val = float("-0.452459") @@ -473,6 +516,7 @@ class Program_weight_tensor_parameter_42: class Program_weight_tensor_parameter_43: name = "parameter_43" + original_name = "linear_1534.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-1.29009") @@ -484,6 +528,7 @@ class Program_weight_tensor_parameter_43: class Program_weight_tensor_parameter_44: name = "parameter_44" + original_name = "linear_1533.b_0" shape = [1024] dtype = "float32" min_val = float("-0.451086") @@ -495,6 +540,7 @@ class Program_weight_tensor_parameter_44: class Program_weight_tensor_parameter_45: name = "parameter_45" + original_name = "linear_1533.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.510957") @@ -506,6 +552,7 @@ class Program_weight_tensor_parameter_45: class Program_weight_tensor_parameter_46: name = "parameter_46" + original_name = "linear_1532.b_0" shape = [1024] dtype = "float32" min_val = float("-0.124511") @@ -517,6 +564,7 @@ class Program_weight_tensor_parameter_46: class Program_weight_tensor_parameter_47: name = "parameter_47" + original_name = "linear_1532.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-1.06255") @@ -528,6 +576,7 @@ class Program_weight_tensor_parameter_47: class Program_weight_tensor_parameter_48: name = "parameter_48" + original_name = "linear_1531.b_0" shape = [1024] dtype = "float32" min_val = float("-2.19211") @@ -539,6 +588,7 @@ class Program_weight_tensor_parameter_48: class Program_weight_tensor_parameter_49: name = "parameter_49" + original_name = "linear_1531.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.974335") @@ -550,6 +600,7 @@ class Program_weight_tensor_parameter_49: class Program_weight_tensor_parameter_50: name = "parameter_50" + original_name = "layer_norm_527.b_0" shape = [1024] dtype = "float32" min_val = float("-0.378604") @@ -561,6 +612,7 @@ class Program_weight_tensor_parameter_50: class Program_weight_tensor_parameter_51: name = "parameter_51" + original_name = "layer_norm_527.w_0" shape = [1024] dtype = "float32" min_val = float("0.0842866") @@ -572,6 +624,7 @@ class Program_weight_tensor_parameter_51: class Program_weight_tensor_parameter_52: name = "parameter_52" + original_name = "embedding_89.w_0" shape = [2, 1024] dtype = "float32" min_val = float("-1.64942") @@ -583,6 +636,7 @@ class Program_weight_tensor_parameter_52: class Program_weight_tensor_parameter_53: name = "parameter_53" + original_name = "embedding_88.w_0" shape = [600, 1024] dtype = "float32" min_val = float("-0.907852") @@ -594,6 +648,7 @@ class Program_weight_tensor_parameter_53: class Program_weight_tensor_parameter_54: name = "parameter_54" + original_name = "embedding_87.w_0" shape = [50006, 1024] dtype = "float32" min_val = float("-1.34628") diff --git a/paddle_samples/PaddleNLP/ppminilm-6l-768h/input_meta.py b/paddle_samples/PaddleNLP/ppminilm-6l-768h/input_meta.py index 17034a664..f5ff6b58e 100644 --- a/paddle_samples/PaddleNLP/ppminilm-6l-768h/input_meta.py +++ b/paddle_samples/PaddleNLP/ppminilm-6l-768h/input_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_data_0: name = "data_0" + original_name = "input_ids" shape = [1, 27] dtype = "int64" data = [ @@ -35,6 +36,7 @@ class Program_weight_tensor_data_0: class Program_weight_tensor_data_1: name = "data_1" + original_name = "token_type_ids" shape = [1, 27] dtype = "int64" data = [ diff --git a/paddle_samples/PaddleNLP/ppminilm-6l-768h/model.py b/paddle_samples/PaddleNLP/ppminilm-6l-768h/model.py index ecae74a2a..311f371c9 100644 --- a/paddle_samples/PaddleNLP/ppminilm-6l-768h/model.py +++ b/paddle_samples/PaddleNLP/ppminilm-6l-768h/model.py @@ -172,7 +172,7 @@ def forward( # pd_op.embedding: (1x27x768xf32) <- (1x27xi64, 512x768xf32) embedding_1 = paddle._C_ops.embedding(subtract_0, parameter_101, -1, False) - del parameter_101 + del parameter_101, subtract_0 # pd_op.embedding: (1x27x768xf32) <- (1x27xi64, 2x768xf32) embedding_2 = paddle._C_ops.embedding(data_1, parameter_100, -1, False) @@ -180,9 +180,11 @@ def forward( # pd_op.add: (1x27x768xf32) <- (1x27x768xf32, 1x27x768xf32) add_0 = paddle._C_ops.add(embedding_0, embedding_1) + del embedding_0, embedding_1 # pd_op.add: (1x27x768xf32) <- (1x27x768xf32, 1x27x768xf32) add_1 = paddle._C_ops.add(add_0, embedding_2) + del add_0, embedding_2 # pd_op.layer_norm: (1x27x768xf32, 1x27xf32, 1x27xf32) <- (1x27x768xf32, 768xf32, 768xf32) layer_norm_0, layer_norm_1, layer_norm_2 = (lambda x, f: f(x))( @@ -191,71 +193,17 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_98, parameter_99 + del add_1, parameter_98, parameter_99 # pd_op.full: (1xf32) <- () full_4 = paddle._C_ops.full( [1], float("0.1"), paddle.float32, paddle.core.CPUPlace() ) - # pd_op.assign: (1xf32) <- (1xf32) - assign_0 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_1 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_2 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_3 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_4 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_5 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_6 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_7 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_8 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_9 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_10 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_11 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_12 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_13 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_14 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_15 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_16 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_17 = full_4 - # pd_op.dropout: (1x27x768xf32, 1x27x768xui8) <- (1x27x768xf32, None, 1xf32) dropout_0, dropout_1 = (lambda x, f: f(x))( paddle._C_ops.dropout( - layer_norm_0, None, full_4, False, "upscale_in_train", 0, False + layer_norm_0, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -267,13 +215,14 @@ def forward( # pd_op.add: (1x27x768xf32) <- (1x27x768xf32, 768xf32) add_2 = paddle._C_ops.add(matmul_0, parameter_96) - del parameter_96 + del matmul_0, parameter_96 # pd_op.full_int_array: (4xi64) <- () full_int_array_1 = [0, 0, 12, 64] # pd_op.reshape: (1x27x12x64xf32) <- (1x27x768xf32, 4xi64) reshape_0 = paddle._C_ops.reshape(add_2, full_int_array_1) + del add_2 # pd_op.transpose: (1x12x27x64xf32) <- (1x27x12x64xf32) transpose_0 = paddle._C_ops.transpose(reshape_0, [0, 2, 1, 3]) @@ -285,7 +234,7 @@ def forward( # pd_op.add: (1x27x768xf32) <- (1x27x768xf32, 768xf32) add_3 = paddle._C_ops.add(matmul_1, parameter_94) - del parameter_94 + del matmul_1, parameter_94 # pd_op.matmul: (1x27x768xf32) <- (1x27x768xf32, 768x768xf32) matmul_2 = paddle._C_ops.matmul(dropout_0, parameter_93, False, False) @@ -293,10 +242,11 @@ def forward( # pd_op.add: (1x27x768xf32) <- (1x27x768xf32, 768xf32) add_4 = paddle._C_ops.add(matmul_2, parameter_92) - del parameter_92 + del matmul_2, parameter_92 # pd_op.reshape: (1x27x12x64xf32) <- (1x27x768xf32, 4xi64) reshape_1 = paddle._C_ops.reshape(add_3, full_int_array_1) + del add_3 # pd_op.transpose: (1x12x27x64xf32) <- (1x27x12x64xf32) transpose_1 = paddle._C_ops.transpose(reshape_1, [0, 2, 1, 3]) @@ -304,6 +254,7 @@ def forward( # pd_op.reshape: (1x27x12x64xf32) <- (1x27x768xf32, 4xi64) reshape_2 = paddle._C_ops.reshape(add_4, full_int_array_1) + del add_4 # pd_op.transpose: (1x12x27x64xf32) <- (1x27x12x64xf32) transpose_2 = paddle._C_ops.transpose(reshape_2, [0, 2, 1, 3]) @@ -314,30 +265,17 @@ def forward( [1], float("0.125"), paddle.float32, paddle.core.CPUPlace() ) - # pd_op.assign: (1xf32) <- (1xf32) - assign_18 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_19 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_20 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_21 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_22 = full_5 - # pd_op.scale: (1x12x27x64xf32) <- (1x12x27x64xf32, 1xf32) scale_1 = paddle._C_ops.scale(transpose_0, full_5, float("0"), True) del transpose_0 # pd_op.matmul: (1x12x27x27xf32) <- (1x12x27x64xf32, 1x12x27x64xf32) matmul_3 = paddle._C_ops.matmul(scale_1, transpose_1, False, True) + del scale_1, transpose_1 # pd_op.add: (1x12x27x27xf32) <- (1x12x27x27xf32, 1x1x1x27xf32) add_5 = paddle._C_ops.add(matmul_3, unsqueeze_0) + del matmul_3 # pd_op.softmax: (1x12x27x27xf32) <- (1x12x27x27xf32) softmax_0 = paddle._C_ops.softmax(add_5, -1) @@ -346,13 +284,15 @@ def forward( # pd_op.dropout: (1x12x27x27xf32, 1x12x27x27xui8) <- (1x12x27x27xf32, None, 1xf32) dropout_2, dropout_3 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_0, None, full_4, False, "upscale_in_train", 0, False + softmax_0, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_0 # pd_op.matmul: (1x12x27x64xf32) <- (1x12x27x27xf32, 1x12x27x64xf32) matmul_4 = paddle._C_ops.matmul(dropout_2, transpose_2, False, False) + del dropout_2, transpose_2 # pd_op.transpose: (1x27x12x64xf32) <- (1x12x27x64xf32) transpose_3 = paddle._C_ops.transpose(matmul_4, [0, 2, 1, 3]) @@ -363,19 +303,20 @@ def forward( # pd_op.reshape: (1x27x768xf32) <- (1x27x12x64xf32, 3xi64) reshape_3 = paddle._C_ops.reshape(transpose_3, full_int_array_2) + del transpose_3 # pd_op.matmul: (1x27x768xf32) <- (1x27x768xf32, 768x768xf32) matmul_5 = paddle._C_ops.matmul(reshape_3, parameter_91, False, False) - del parameter_91 + del parameter_91, reshape_3 # pd_op.add: (1x27x768xf32) <- (1x27x768xf32, 768xf32) add_6 = paddle._C_ops.add(matmul_5, parameter_90) - del parameter_90 + del matmul_5, parameter_90 # pd_op.dropout: (1x27x768xf32, 1x27x768xui8) <- (1x27x768xf32, None, 1xf32) dropout_4, dropout_5 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_6, None, full_4, False, "upscale_in_train", 0, False + add_6, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -383,6 +324,7 @@ def forward( # pd_op.add: (1x27x768xf32) <- (1x27x768xf32, 1x27x768xf32) add_7 = paddle._C_ops.add(dropout_0, dropout_4) + del dropout_0, dropout_4 # pd_op.layer_norm: (1x27x768xf32, 1x27xf32, 1x27xf32) <- (1x27x768xf32, 768xf32, 768xf32) layer_norm_3, layer_norm_4, layer_norm_5 = (lambda x, f: f(x))( @@ -391,7 +333,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_84, parameter_85 + del add_7, parameter_84, parameter_85 # pd_op.matmul: (1x27x3072xf32) <- (1x27x768xf32, 768x3072xf32) matmul_6 = paddle._C_ops.matmul(layer_norm_3, parameter_89, False, False) @@ -399,7 +341,7 @@ def forward( # pd_op.add: (1x27x3072xf32) <- (1x27x3072xf32, 3072xf32) add_8 = paddle._C_ops.add(matmul_6, parameter_88) - del parameter_88 + del matmul_6, parameter_88 # pd_op.relu: (1x27x3072xf32) <- (1x27x3072xf32) relu_0 = paddle._C_ops.relu(add_8) @@ -407,16 +349,16 @@ def forward( # pd_op.matmul: (1x27x768xf32) <- (1x27x3072xf32, 3072x768xf32) matmul_7 = paddle._C_ops.matmul(relu_0, parameter_87, False, False) - del parameter_87 + del parameter_87, relu_0 # pd_op.add: (1x27x768xf32) <- (1x27x768xf32, 768xf32) add_9 = paddle._C_ops.add(matmul_7, parameter_86) - del parameter_86 + del matmul_7, parameter_86 # pd_op.dropout: (1x27x768xf32, 1x27x768xui8) <- (1x27x768xf32, None, 1xf32) dropout_6, dropout_7 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_9, None, full_4, False, "upscale_in_train", 0, False + add_9, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -424,6 +366,7 @@ def forward( # pd_op.add: (1x27x768xf32) <- (1x27x768xf32, 1x27x768xf32) add_10 = paddle._C_ops.add(layer_norm_3, dropout_6) + del dropout_6, layer_norm_3 # pd_op.layer_norm: (1x27x768xf32, 1x27xf32, 1x27xf32) <- (1x27x768xf32, 768xf32, 768xf32) layer_norm_6, layer_norm_7, layer_norm_8 = (lambda x, f: f(x))( @@ -432,7 +375,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_82, parameter_83 + del add_10, parameter_82, parameter_83 # pd_op.matmul: (1x27x768xf32) <- (1x27x768xf32, 768x768xf32) matmul_8 = paddle._C_ops.matmul(layer_norm_6, parameter_81, False, False) @@ -440,10 +383,11 @@ def forward( # pd_op.add: (1x27x768xf32) <- (1x27x768xf32, 768xf32) add_11 = paddle._C_ops.add(matmul_8, parameter_80) - del parameter_80 + del matmul_8, parameter_80 # pd_op.reshape: (1x27x12x64xf32) <- (1x27x768xf32, 4xi64) reshape_4 = paddle._C_ops.reshape(add_11, full_int_array_1) + del add_11 # pd_op.transpose: (1x12x27x64xf32) <- (1x27x12x64xf32) transpose_4 = paddle._C_ops.transpose(reshape_4, [0, 2, 1, 3]) @@ -455,7 +399,7 @@ def forward( # pd_op.add: (1x27x768xf32) <- (1x27x768xf32, 768xf32) add_12 = paddle._C_ops.add(matmul_9, parameter_78) - del parameter_78 + del matmul_9, parameter_78 # pd_op.matmul: (1x27x768xf32) <- (1x27x768xf32, 768x768xf32) matmul_10 = paddle._C_ops.matmul(layer_norm_6, parameter_77, False, False) @@ -463,10 +407,11 @@ def forward( # pd_op.add: (1x27x768xf32) <- (1x27x768xf32, 768xf32) add_13 = paddle._C_ops.add(matmul_10, parameter_76) - del parameter_76 + del matmul_10, parameter_76 # pd_op.reshape: (1x27x12x64xf32) <- (1x27x768xf32, 4xi64) reshape_5 = paddle._C_ops.reshape(add_12, full_int_array_1) + del add_12 # pd_op.transpose: (1x12x27x64xf32) <- (1x27x12x64xf32) transpose_5 = paddle._C_ops.transpose(reshape_5, [0, 2, 1, 3]) @@ -474,6 +419,7 @@ def forward( # pd_op.reshape: (1x27x12x64xf32) <- (1x27x768xf32, 4xi64) reshape_6 = paddle._C_ops.reshape(add_13, full_int_array_1) + del add_13 # pd_op.transpose: (1x12x27x64xf32) <- (1x27x12x64xf32) transpose_6 = paddle._C_ops.transpose(reshape_6, [0, 2, 1, 3]) @@ -485,9 +431,11 @@ def forward( # pd_op.matmul: (1x12x27x27xf32) <- (1x12x27x64xf32, 1x12x27x64xf32) matmul_11 = paddle._C_ops.matmul(scale_2, transpose_5, False, True) + del scale_2, transpose_5 # pd_op.add: (1x12x27x27xf32) <- (1x12x27x27xf32, 1x1x1x27xf32) add_14 = paddle._C_ops.add(matmul_11, unsqueeze_0) + del matmul_11 # pd_op.softmax: (1x12x27x27xf32) <- (1x12x27x27xf32) softmax_1 = paddle._C_ops.softmax(add_14, -1) @@ -496,13 +444,15 @@ def forward( # pd_op.dropout: (1x12x27x27xf32, 1x12x27x27xui8) <- (1x12x27x27xf32, None, 1xf32) dropout_8, dropout_9 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_1, None, full_4, False, "upscale_in_train", 0, False + softmax_1, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_1 # pd_op.matmul: (1x12x27x64xf32) <- (1x12x27x27xf32, 1x12x27x64xf32) matmul_12 = paddle._C_ops.matmul(dropout_8, transpose_6, False, False) + del dropout_8, transpose_6 # pd_op.transpose: (1x27x12x64xf32) <- (1x12x27x64xf32) transpose_7 = paddle._C_ops.transpose(matmul_12, [0, 2, 1, 3]) @@ -510,19 +460,20 @@ def forward( # pd_op.reshape: (1x27x768xf32) <- (1x27x12x64xf32, 3xi64) reshape_7 = paddle._C_ops.reshape(transpose_7, full_int_array_2) + del transpose_7 # pd_op.matmul: (1x27x768xf32) <- (1x27x768xf32, 768x768xf32) matmul_13 = paddle._C_ops.matmul(reshape_7, parameter_75, False, False) - del parameter_75 + del parameter_75, reshape_7 # pd_op.add: (1x27x768xf32) <- (1x27x768xf32, 768xf32) add_15 = paddle._C_ops.add(matmul_13, parameter_74) - del parameter_74 + del matmul_13, parameter_74 # pd_op.dropout: (1x27x768xf32, 1x27x768xui8) <- (1x27x768xf32, None, 1xf32) dropout_10, dropout_11 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_15, None, full_4, False, "upscale_in_train", 0, False + add_15, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -530,6 +481,7 @@ def forward( # pd_op.add: (1x27x768xf32) <- (1x27x768xf32, 1x27x768xf32) add_16 = paddle._C_ops.add(layer_norm_6, dropout_10) + del dropout_10, layer_norm_6 # pd_op.layer_norm: (1x27x768xf32, 1x27xf32, 1x27xf32) <- (1x27x768xf32, 768xf32, 768xf32) layer_norm_9, layer_norm_10, layer_norm_11 = (lambda x, f: f(x))( @@ -538,7 +490,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_68, parameter_69 + del add_16, parameter_68, parameter_69 # pd_op.matmul: (1x27x3072xf32) <- (1x27x768xf32, 768x3072xf32) matmul_14 = paddle._C_ops.matmul(layer_norm_9, parameter_73, False, False) @@ -546,7 +498,7 @@ def forward( # pd_op.add: (1x27x3072xf32) <- (1x27x3072xf32, 3072xf32) add_17 = paddle._C_ops.add(matmul_14, parameter_72) - del parameter_72 + del matmul_14, parameter_72 # pd_op.relu: (1x27x3072xf32) <- (1x27x3072xf32) relu_1 = paddle._C_ops.relu(add_17) @@ -554,16 +506,16 @@ def forward( # pd_op.matmul: (1x27x768xf32) <- (1x27x3072xf32, 3072x768xf32) matmul_15 = paddle._C_ops.matmul(relu_1, parameter_71, False, False) - del parameter_71 + del parameter_71, relu_1 # pd_op.add: (1x27x768xf32) <- (1x27x768xf32, 768xf32) add_18 = paddle._C_ops.add(matmul_15, parameter_70) - del parameter_70 + del matmul_15, parameter_70 # pd_op.dropout: (1x27x768xf32, 1x27x768xui8) <- (1x27x768xf32, None, 1xf32) dropout_12, dropout_13 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_18, None, full_4, False, "upscale_in_train", 0, False + add_18, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -571,6 +523,7 @@ def forward( # pd_op.add: (1x27x768xf32) <- (1x27x768xf32, 1x27x768xf32) add_19 = paddle._C_ops.add(layer_norm_9, dropout_12) + del dropout_12, layer_norm_9 # pd_op.layer_norm: (1x27x768xf32, 1x27xf32, 1x27xf32) <- (1x27x768xf32, 768xf32, 768xf32) layer_norm_12, layer_norm_13, layer_norm_14 = (lambda x, f: f(x))( @@ -579,7 +532,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_66, parameter_67 + del add_19, parameter_66, parameter_67 # pd_op.matmul: (1x27x768xf32) <- (1x27x768xf32, 768x768xf32) matmul_16 = paddle._C_ops.matmul(layer_norm_12, parameter_65, False, False) @@ -587,10 +540,11 @@ def forward( # pd_op.add: (1x27x768xf32) <- (1x27x768xf32, 768xf32) add_20 = paddle._C_ops.add(matmul_16, parameter_64) - del parameter_64 + del matmul_16, parameter_64 # pd_op.reshape: (1x27x12x64xf32) <- (1x27x768xf32, 4xi64) reshape_8 = paddle._C_ops.reshape(add_20, full_int_array_1) + del add_20 # pd_op.transpose: (1x12x27x64xf32) <- (1x27x12x64xf32) transpose_8 = paddle._C_ops.transpose(reshape_8, [0, 2, 1, 3]) @@ -602,7 +556,7 @@ def forward( # pd_op.add: (1x27x768xf32) <- (1x27x768xf32, 768xf32) add_21 = paddle._C_ops.add(matmul_17, parameter_62) - del parameter_62 + del matmul_17, parameter_62 # pd_op.matmul: (1x27x768xf32) <- (1x27x768xf32, 768x768xf32) matmul_18 = paddle._C_ops.matmul(layer_norm_12, parameter_61, False, False) @@ -610,10 +564,11 @@ def forward( # pd_op.add: (1x27x768xf32) <- (1x27x768xf32, 768xf32) add_22 = paddle._C_ops.add(matmul_18, parameter_60) - del parameter_60 + del matmul_18, parameter_60 # pd_op.reshape: (1x27x12x64xf32) <- (1x27x768xf32, 4xi64) reshape_9 = paddle._C_ops.reshape(add_21, full_int_array_1) + del add_21 # pd_op.transpose: (1x12x27x64xf32) <- (1x27x12x64xf32) transpose_9 = paddle._C_ops.transpose(reshape_9, [0, 2, 1, 3]) @@ -621,6 +576,7 @@ def forward( # pd_op.reshape: (1x27x12x64xf32) <- (1x27x768xf32, 4xi64) reshape_10 = paddle._C_ops.reshape(add_22, full_int_array_1) + del add_22 # pd_op.transpose: (1x12x27x64xf32) <- (1x27x12x64xf32) transpose_10 = paddle._C_ops.transpose(reshape_10, [0, 2, 1, 3]) @@ -632,9 +588,11 @@ def forward( # pd_op.matmul: (1x12x27x27xf32) <- (1x12x27x64xf32, 1x12x27x64xf32) matmul_19 = paddle._C_ops.matmul(scale_3, transpose_9, False, True) + del scale_3, transpose_9 # pd_op.add: (1x12x27x27xf32) <- (1x12x27x27xf32, 1x1x1x27xf32) add_23 = paddle._C_ops.add(matmul_19, unsqueeze_0) + del matmul_19 # pd_op.softmax: (1x12x27x27xf32) <- (1x12x27x27xf32) softmax_2 = paddle._C_ops.softmax(add_23, -1) @@ -643,13 +601,15 @@ def forward( # pd_op.dropout: (1x12x27x27xf32, 1x12x27x27xui8) <- (1x12x27x27xf32, None, 1xf32) dropout_14, dropout_15 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_2, None, full_4, False, "upscale_in_train", 0, False + softmax_2, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_2 # pd_op.matmul: (1x12x27x64xf32) <- (1x12x27x27xf32, 1x12x27x64xf32) matmul_20 = paddle._C_ops.matmul(dropout_14, transpose_10, False, False) + del dropout_14, transpose_10 # pd_op.transpose: (1x27x12x64xf32) <- (1x12x27x64xf32) transpose_11 = paddle._C_ops.transpose(matmul_20, [0, 2, 1, 3]) @@ -657,19 +617,20 @@ def forward( # pd_op.reshape: (1x27x768xf32) <- (1x27x12x64xf32, 3xi64) reshape_11 = paddle._C_ops.reshape(transpose_11, full_int_array_2) + del transpose_11 # pd_op.matmul: (1x27x768xf32) <- (1x27x768xf32, 768x768xf32) matmul_21 = paddle._C_ops.matmul(reshape_11, parameter_59, False, False) - del parameter_59 + del parameter_59, reshape_11 # pd_op.add: (1x27x768xf32) <- (1x27x768xf32, 768xf32) add_24 = paddle._C_ops.add(matmul_21, parameter_58) - del parameter_58 + del matmul_21, parameter_58 # pd_op.dropout: (1x27x768xf32, 1x27x768xui8) <- (1x27x768xf32, None, 1xf32) dropout_16, dropout_17 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_24, None, full_4, False, "upscale_in_train", 0, False + add_24, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -677,6 +638,7 @@ def forward( # pd_op.add: (1x27x768xf32) <- (1x27x768xf32, 1x27x768xf32) add_25 = paddle._C_ops.add(layer_norm_12, dropout_16) + del dropout_16, layer_norm_12 # pd_op.layer_norm: (1x27x768xf32, 1x27xf32, 1x27xf32) <- (1x27x768xf32, 768xf32, 768xf32) layer_norm_15, layer_norm_16, layer_norm_17 = (lambda x, f: f(x))( @@ -685,7 +647,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_52, parameter_53 + del add_25, parameter_52, parameter_53 # pd_op.matmul: (1x27x3072xf32) <- (1x27x768xf32, 768x3072xf32) matmul_22 = paddle._C_ops.matmul(layer_norm_15, parameter_57, False, False) @@ -693,7 +655,7 @@ def forward( # pd_op.add: (1x27x3072xf32) <- (1x27x3072xf32, 3072xf32) add_26 = paddle._C_ops.add(matmul_22, parameter_56) - del parameter_56 + del matmul_22, parameter_56 # pd_op.relu: (1x27x3072xf32) <- (1x27x3072xf32) relu_2 = paddle._C_ops.relu(add_26) @@ -701,16 +663,16 @@ def forward( # pd_op.matmul: (1x27x768xf32) <- (1x27x3072xf32, 3072x768xf32) matmul_23 = paddle._C_ops.matmul(relu_2, parameter_55, False, False) - del parameter_55 + del parameter_55, relu_2 # pd_op.add: (1x27x768xf32) <- (1x27x768xf32, 768xf32) add_27 = paddle._C_ops.add(matmul_23, parameter_54) - del parameter_54 + del matmul_23, parameter_54 # pd_op.dropout: (1x27x768xf32, 1x27x768xui8) <- (1x27x768xf32, None, 1xf32) dropout_18, dropout_19 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_27, None, full_4, False, "upscale_in_train", 0, False + add_27, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -718,6 +680,7 @@ def forward( # pd_op.add: (1x27x768xf32) <- (1x27x768xf32, 1x27x768xf32) add_28 = paddle._C_ops.add(layer_norm_15, dropout_18) + del dropout_18, layer_norm_15 # pd_op.layer_norm: (1x27x768xf32, 1x27xf32, 1x27xf32) <- (1x27x768xf32, 768xf32, 768xf32) layer_norm_18, layer_norm_19, layer_norm_20 = (lambda x, f: f(x))( @@ -726,7 +689,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_50, parameter_51 + del add_28, parameter_50, parameter_51 # pd_op.matmul: (1x27x768xf32) <- (1x27x768xf32, 768x768xf32) matmul_24 = paddle._C_ops.matmul(layer_norm_18, parameter_49, False, False) @@ -734,10 +697,11 @@ def forward( # pd_op.add: (1x27x768xf32) <- (1x27x768xf32, 768xf32) add_29 = paddle._C_ops.add(matmul_24, parameter_48) - del parameter_48 + del matmul_24, parameter_48 # pd_op.reshape: (1x27x12x64xf32) <- (1x27x768xf32, 4xi64) reshape_12 = paddle._C_ops.reshape(add_29, full_int_array_1) + del add_29 # pd_op.transpose: (1x12x27x64xf32) <- (1x27x12x64xf32) transpose_12 = paddle._C_ops.transpose(reshape_12, [0, 2, 1, 3]) @@ -749,7 +713,7 @@ def forward( # pd_op.add: (1x27x768xf32) <- (1x27x768xf32, 768xf32) add_30 = paddle._C_ops.add(matmul_25, parameter_46) - del parameter_46 + del matmul_25, parameter_46 # pd_op.matmul: (1x27x768xf32) <- (1x27x768xf32, 768x768xf32) matmul_26 = paddle._C_ops.matmul(layer_norm_18, parameter_45, False, False) @@ -757,10 +721,11 @@ def forward( # pd_op.add: (1x27x768xf32) <- (1x27x768xf32, 768xf32) add_31 = paddle._C_ops.add(matmul_26, parameter_44) - del parameter_44 + del matmul_26, parameter_44 # pd_op.reshape: (1x27x12x64xf32) <- (1x27x768xf32, 4xi64) reshape_13 = paddle._C_ops.reshape(add_30, full_int_array_1) + del add_30 # pd_op.transpose: (1x12x27x64xf32) <- (1x27x12x64xf32) transpose_13 = paddle._C_ops.transpose(reshape_13, [0, 2, 1, 3]) @@ -768,6 +733,7 @@ def forward( # pd_op.reshape: (1x27x12x64xf32) <- (1x27x768xf32, 4xi64) reshape_14 = paddle._C_ops.reshape(add_31, full_int_array_1) + del add_31 # pd_op.transpose: (1x12x27x64xf32) <- (1x27x12x64xf32) transpose_14 = paddle._C_ops.transpose(reshape_14, [0, 2, 1, 3]) @@ -779,9 +745,11 @@ def forward( # pd_op.matmul: (1x12x27x27xf32) <- (1x12x27x64xf32, 1x12x27x64xf32) matmul_27 = paddle._C_ops.matmul(scale_4, transpose_13, False, True) + del scale_4, transpose_13 # pd_op.add: (1x12x27x27xf32) <- (1x12x27x27xf32, 1x1x1x27xf32) add_32 = paddle._C_ops.add(matmul_27, unsqueeze_0) + del matmul_27 # pd_op.softmax: (1x12x27x27xf32) <- (1x12x27x27xf32) softmax_3 = paddle._C_ops.softmax(add_32, -1) @@ -790,13 +758,15 @@ def forward( # pd_op.dropout: (1x12x27x27xf32, 1x12x27x27xui8) <- (1x12x27x27xf32, None, 1xf32) dropout_20, dropout_21 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_3, None, full_4, False, "upscale_in_train", 0, False + softmax_3, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_3 # pd_op.matmul: (1x12x27x64xf32) <- (1x12x27x27xf32, 1x12x27x64xf32) matmul_28 = paddle._C_ops.matmul(dropout_20, transpose_14, False, False) + del dropout_20, transpose_14 # pd_op.transpose: (1x27x12x64xf32) <- (1x12x27x64xf32) transpose_15 = paddle._C_ops.transpose(matmul_28, [0, 2, 1, 3]) @@ -804,19 +774,20 @@ def forward( # pd_op.reshape: (1x27x768xf32) <- (1x27x12x64xf32, 3xi64) reshape_15 = paddle._C_ops.reshape(transpose_15, full_int_array_2) + del transpose_15 # pd_op.matmul: (1x27x768xf32) <- (1x27x768xf32, 768x768xf32) matmul_29 = paddle._C_ops.matmul(reshape_15, parameter_43, False, False) - del parameter_43 + del parameter_43, reshape_15 # pd_op.add: (1x27x768xf32) <- (1x27x768xf32, 768xf32) add_33 = paddle._C_ops.add(matmul_29, parameter_42) - del parameter_42 + del matmul_29, parameter_42 # pd_op.dropout: (1x27x768xf32, 1x27x768xui8) <- (1x27x768xf32, None, 1xf32) dropout_22, dropout_23 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_33, None, full_4, False, "upscale_in_train", 0, False + add_33, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -824,6 +795,7 @@ def forward( # pd_op.add: (1x27x768xf32) <- (1x27x768xf32, 1x27x768xf32) add_34 = paddle._C_ops.add(layer_norm_18, dropout_22) + del dropout_22, layer_norm_18 # pd_op.layer_norm: (1x27x768xf32, 1x27xf32, 1x27xf32) <- (1x27x768xf32, 768xf32, 768xf32) layer_norm_21, layer_norm_22, layer_norm_23 = (lambda x, f: f(x))( @@ -832,7 +804,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_36, parameter_37 + del add_34, parameter_36, parameter_37 # pd_op.matmul: (1x27x3072xf32) <- (1x27x768xf32, 768x3072xf32) matmul_30 = paddle._C_ops.matmul(layer_norm_21, parameter_41, False, False) @@ -840,7 +812,7 @@ def forward( # pd_op.add: (1x27x3072xf32) <- (1x27x3072xf32, 3072xf32) add_35 = paddle._C_ops.add(matmul_30, parameter_40) - del parameter_40 + del matmul_30, parameter_40 # pd_op.relu: (1x27x3072xf32) <- (1x27x3072xf32) relu_3 = paddle._C_ops.relu(add_35) @@ -848,16 +820,16 @@ def forward( # pd_op.matmul: (1x27x768xf32) <- (1x27x3072xf32, 3072x768xf32) matmul_31 = paddle._C_ops.matmul(relu_3, parameter_39, False, False) - del parameter_39 + del parameter_39, relu_3 # pd_op.add: (1x27x768xf32) <- (1x27x768xf32, 768xf32) add_36 = paddle._C_ops.add(matmul_31, parameter_38) - del parameter_38 + del matmul_31, parameter_38 # pd_op.dropout: (1x27x768xf32, 1x27x768xui8) <- (1x27x768xf32, None, 1xf32) dropout_24, dropout_25 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_36, None, full_4, False, "upscale_in_train", 0, False + add_36, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -865,6 +837,7 @@ def forward( # pd_op.add: (1x27x768xf32) <- (1x27x768xf32, 1x27x768xf32) add_37 = paddle._C_ops.add(layer_norm_21, dropout_24) + del dropout_24, layer_norm_21 # pd_op.layer_norm: (1x27x768xf32, 1x27xf32, 1x27xf32) <- (1x27x768xf32, 768xf32, 768xf32) layer_norm_24, layer_norm_25, layer_norm_26 = (lambda x, f: f(x))( @@ -873,7 +846,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_34, parameter_35 + del add_37, parameter_34, parameter_35 # pd_op.matmul: (1x27x768xf32) <- (1x27x768xf32, 768x768xf32) matmul_32 = paddle._C_ops.matmul(layer_norm_24, parameter_33, False, False) @@ -881,10 +854,11 @@ def forward( # pd_op.add: (1x27x768xf32) <- (1x27x768xf32, 768xf32) add_38 = paddle._C_ops.add(matmul_32, parameter_32) - del parameter_32 + del matmul_32, parameter_32 # pd_op.reshape: (1x27x12x64xf32) <- (1x27x768xf32, 4xi64) reshape_16 = paddle._C_ops.reshape(add_38, full_int_array_1) + del add_38 # pd_op.transpose: (1x12x27x64xf32) <- (1x27x12x64xf32) transpose_16 = paddle._C_ops.transpose(reshape_16, [0, 2, 1, 3]) @@ -896,7 +870,7 @@ def forward( # pd_op.add: (1x27x768xf32) <- (1x27x768xf32, 768xf32) add_39 = paddle._C_ops.add(matmul_33, parameter_30) - del parameter_30 + del matmul_33, parameter_30 # pd_op.matmul: (1x27x768xf32) <- (1x27x768xf32, 768x768xf32) matmul_34 = paddle._C_ops.matmul(layer_norm_24, parameter_29, False, False) @@ -904,10 +878,11 @@ def forward( # pd_op.add: (1x27x768xf32) <- (1x27x768xf32, 768xf32) add_40 = paddle._C_ops.add(matmul_34, parameter_28) - del parameter_28 + del matmul_34, parameter_28 # pd_op.reshape: (1x27x12x64xf32) <- (1x27x768xf32, 4xi64) reshape_17 = paddle._C_ops.reshape(add_39, full_int_array_1) + del add_39 # pd_op.transpose: (1x12x27x64xf32) <- (1x27x12x64xf32) transpose_17 = paddle._C_ops.transpose(reshape_17, [0, 2, 1, 3]) @@ -915,6 +890,7 @@ def forward( # pd_op.reshape: (1x27x12x64xf32) <- (1x27x768xf32, 4xi64) reshape_18 = paddle._C_ops.reshape(add_40, full_int_array_1) + del add_40 # pd_op.transpose: (1x12x27x64xf32) <- (1x27x12x64xf32) transpose_18 = paddle._C_ops.transpose(reshape_18, [0, 2, 1, 3]) @@ -926,9 +902,11 @@ def forward( # pd_op.matmul: (1x12x27x27xf32) <- (1x12x27x64xf32, 1x12x27x64xf32) matmul_35 = paddle._C_ops.matmul(scale_5, transpose_17, False, True) + del scale_5, transpose_17 # pd_op.add: (1x12x27x27xf32) <- (1x12x27x27xf32, 1x1x1x27xf32) add_41 = paddle._C_ops.add(matmul_35, unsqueeze_0) + del matmul_35 # pd_op.softmax: (1x12x27x27xf32) <- (1x12x27x27xf32) softmax_4 = paddle._C_ops.softmax(add_41, -1) @@ -937,13 +915,15 @@ def forward( # pd_op.dropout: (1x12x27x27xf32, 1x12x27x27xui8) <- (1x12x27x27xf32, None, 1xf32) dropout_26, dropout_27 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_4, None, full_4, False, "upscale_in_train", 0, False + softmax_4, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_4 # pd_op.matmul: (1x12x27x64xf32) <- (1x12x27x27xf32, 1x12x27x64xf32) matmul_36 = paddle._C_ops.matmul(dropout_26, transpose_18, False, False) + del dropout_26, transpose_18 # pd_op.transpose: (1x27x12x64xf32) <- (1x12x27x64xf32) transpose_19 = paddle._C_ops.transpose(matmul_36, [0, 2, 1, 3]) @@ -951,19 +931,20 @@ def forward( # pd_op.reshape: (1x27x768xf32) <- (1x27x12x64xf32, 3xi64) reshape_19 = paddle._C_ops.reshape(transpose_19, full_int_array_2) + del transpose_19 # pd_op.matmul: (1x27x768xf32) <- (1x27x768xf32, 768x768xf32) matmul_37 = paddle._C_ops.matmul(reshape_19, parameter_27, False, False) - del parameter_27 + del parameter_27, reshape_19 # pd_op.add: (1x27x768xf32) <- (1x27x768xf32, 768xf32) add_42 = paddle._C_ops.add(matmul_37, parameter_26) - del parameter_26 + del matmul_37, parameter_26 # pd_op.dropout: (1x27x768xf32, 1x27x768xui8) <- (1x27x768xf32, None, 1xf32) dropout_28, dropout_29 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_42, None, full_4, False, "upscale_in_train", 0, False + add_42, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -971,6 +952,7 @@ def forward( # pd_op.add: (1x27x768xf32) <- (1x27x768xf32, 1x27x768xf32) add_43 = paddle._C_ops.add(layer_norm_24, dropout_28) + del dropout_28, layer_norm_24 # pd_op.layer_norm: (1x27x768xf32, 1x27xf32, 1x27xf32) <- (1x27x768xf32, 768xf32, 768xf32) layer_norm_27, layer_norm_28, layer_norm_29 = (lambda x, f: f(x))( @@ -979,7 +961,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_20, parameter_21 + del add_43, parameter_20, parameter_21 # pd_op.matmul: (1x27x3072xf32) <- (1x27x768xf32, 768x3072xf32) matmul_38 = paddle._C_ops.matmul(layer_norm_27, parameter_25, False, False) @@ -987,7 +969,7 @@ def forward( # pd_op.add: (1x27x3072xf32) <- (1x27x3072xf32, 3072xf32) add_44 = paddle._C_ops.add(matmul_38, parameter_24) - del parameter_24 + del matmul_38, parameter_24 # pd_op.relu: (1x27x3072xf32) <- (1x27x3072xf32) relu_4 = paddle._C_ops.relu(add_44) @@ -995,16 +977,16 @@ def forward( # pd_op.matmul: (1x27x768xf32) <- (1x27x3072xf32, 3072x768xf32) matmul_39 = paddle._C_ops.matmul(relu_4, parameter_23, False, False) - del parameter_23 + del parameter_23, relu_4 # pd_op.add: (1x27x768xf32) <- (1x27x768xf32, 768xf32) add_45 = paddle._C_ops.add(matmul_39, parameter_22) - del parameter_22 + del matmul_39, parameter_22 # pd_op.dropout: (1x27x768xf32, 1x27x768xui8) <- (1x27x768xf32, None, 1xf32) dropout_30, dropout_31 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_45, None, full_4, False, "upscale_in_train", 0, False + add_45, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1012,6 +994,7 @@ def forward( # pd_op.add: (1x27x768xf32) <- (1x27x768xf32, 1x27x768xf32) add_46 = paddle._C_ops.add(layer_norm_27, dropout_30) + del dropout_30, layer_norm_27 # pd_op.layer_norm: (1x27x768xf32, 1x27xf32, 1x27xf32) <- (1x27x768xf32, 768xf32, 768xf32) layer_norm_30, layer_norm_31, layer_norm_32 = (lambda x, f: f(x))( @@ -1020,7 +1003,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_18, parameter_19 + del add_46, parameter_18, parameter_19 # pd_op.matmul: (1x27x768xf32) <- (1x27x768xf32, 768x768xf32) matmul_40 = paddle._C_ops.matmul(layer_norm_30, parameter_17, False, False) @@ -1028,10 +1011,11 @@ def forward( # pd_op.add: (1x27x768xf32) <- (1x27x768xf32, 768xf32) add_47 = paddle._C_ops.add(matmul_40, parameter_16) - del parameter_16 + del matmul_40, parameter_16 # pd_op.reshape: (1x27x12x64xf32) <- (1x27x768xf32, 4xi64) reshape_20 = paddle._C_ops.reshape(add_47, full_int_array_1) + del add_47 # pd_op.transpose: (1x12x27x64xf32) <- (1x27x12x64xf32) transpose_20 = paddle._C_ops.transpose(reshape_20, [0, 2, 1, 3]) @@ -1043,7 +1027,7 @@ def forward( # pd_op.add: (1x27x768xf32) <- (1x27x768xf32, 768xf32) add_48 = paddle._C_ops.add(matmul_41, parameter_14) - del parameter_14 + del matmul_41, parameter_14 # pd_op.matmul: (1x27x768xf32) <- (1x27x768xf32, 768x768xf32) matmul_42 = paddle._C_ops.matmul(layer_norm_30, parameter_13, False, False) @@ -1051,10 +1035,11 @@ def forward( # pd_op.add: (1x27x768xf32) <- (1x27x768xf32, 768xf32) add_49 = paddle._C_ops.add(matmul_42, parameter_12) - del parameter_12 + del matmul_42, parameter_12 # pd_op.reshape: (1x27x12x64xf32) <- (1x27x768xf32, 4xi64) reshape_21 = paddle._C_ops.reshape(add_48, full_int_array_1) + del add_48 # pd_op.transpose: (1x12x27x64xf32) <- (1x27x12x64xf32) transpose_21 = paddle._C_ops.transpose(reshape_21, [0, 2, 1, 3]) @@ -1062,7 +1047,7 @@ def forward( # pd_op.reshape: (1x27x12x64xf32) <- (1x27x768xf32, 4xi64) reshape_22 = paddle._C_ops.reshape(add_49, full_int_array_1) - del full_int_array_1 + del add_49, full_int_array_1 # pd_op.transpose: (1x12x27x64xf32) <- (1x27x12x64xf32) transpose_22 = paddle._C_ops.transpose(reshape_22, [0, 2, 1, 3]) @@ -1070,13 +1055,15 @@ def forward( # pd_op.scale: (1x12x27x64xf32) <- (1x12x27x64xf32, 1xf32) scale_6 = paddle._C_ops.scale(transpose_20, full_5, float("0"), True) - del transpose_20 + del full_5, transpose_20 # pd_op.matmul: (1x12x27x27xf32) <- (1x12x27x64xf32, 1x12x27x64xf32) matmul_43 = paddle._C_ops.matmul(scale_6, transpose_21, False, True) + del scale_6, transpose_21 # pd_op.add: (1x12x27x27xf32) <- (1x12x27x27xf32, 1x1x1x27xf32) add_50 = paddle._C_ops.add(matmul_43, unsqueeze_0) + del matmul_43, unsqueeze_0 # pd_op.softmax: (1x12x27x27xf32) <- (1x12x27x27xf32) softmax_5 = paddle._C_ops.softmax(add_50, -1) @@ -1085,13 +1072,15 @@ def forward( # pd_op.dropout: (1x12x27x27xf32, 1x12x27x27xui8) <- (1x12x27x27xf32, None, 1xf32) dropout_32, dropout_33 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_5, None, full_4, False, "upscale_in_train", 0, False + softmax_5, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_5 # pd_op.matmul: (1x12x27x64xf32) <- (1x12x27x27xf32, 1x12x27x64xf32) matmul_44 = paddle._C_ops.matmul(dropout_32, transpose_22, False, False) + del dropout_32, transpose_22 # pd_op.transpose: (1x27x12x64xf32) <- (1x12x27x64xf32) transpose_23 = paddle._C_ops.transpose(matmul_44, [0, 2, 1, 3]) @@ -1099,20 +1088,20 @@ def forward( # pd_op.reshape: (1x27x768xf32) <- (1x27x12x64xf32, 3xi64) reshape_23 = paddle._C_ops.reshape(transpose_23, full_int_array_2) - del full_int_array_2 + del full_int_array_2, transpose_23 # pd_op.matmul: (1x27x768xf32) <- (1x27x768xf32, 768x768xf32) matmul_45 = paddle._C_ops.matmul(reshape_23, parameter_11, False, False) - del parameter_11 + del parameter_11, reshape_23 # pd_op.add: (1x27x768xf32) <- (1x27x768xf32, 768xf32) add_51 = paddle._C_ops.add(matmul_45, parameter_10) - del parameter_10 + del matmul_45, parameter_10 # pd_op.dropout: (1x27x768xf32, 1x27x768xui8) <- (1x27x768xf32, None, 1xf32) dropout_34, dropout_35 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_51, None, full_4, False, "upscale_in_train", 0, False + add_51, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1120,6 +1109,7 @@ def forward( # pd_op.add: (1x27x768xf32) <- (1x27x768xf32, 1x27x768xf32) add_52 = paddle._C_ops.add(layer_norm_30, dropout_34) + del dropout_34, layer_norm_30 # pd_op.layer_norm: (1x27x768xf32, 1x27xf32, 1x27xf32) <- (1x27x768xf32, 768xf32, 768xf32) layer_norm_33, layer_norm_34, layer_norm_35 = (lambda x, f: f(x))( @@ -1128,7 +1118,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_4, parameter_5 + del add_52, parameter_4, parameter_5 # pd_op.matmul: (1x27x3072xf32) <- (1x27x768xf32, 768x3072xf32) matmul_46 = paddle._C_ops.matmul(layer_norm_33, parameter_9, False, False) @@ -1136,7 +1126,7 @@ def forward( # pd_op.add: (1x27x3072xf32) <- (1x27x3072xf32, 3072xf32) add_53 = paddle._C_ops.add(matmul_46, parameter_8) - del parameter_8 + del matmul_46, parameter_8 # pd_op.relu: (1x27x3072xf32) <- (1x27x3072xf32) relu_5 = paddle._C_ops.relu(add_53) @@ -1144,23 +1134,24 @@ def forward( # pd_op.matmul: (1x27x768xf32) <- (1x27x3072xf32, 3072x768xf32) matmul_47 = paddle._C_ops.matmul(relu_5, parameter_7, False, False) - del parameter_7 + del parameter_7, relu_5 # pd_op.add: (1x27x768xf32) <- (1x27x768xf32, 768xf32) add_54 = paddle._C_ops.add(matmul_47, parameter_6) - del parameter_6 + del matmul_47, parameter_6 # pd_op.dropout: (1x27x768xf32, 1x27x768xui8) <- (1x27x768xf32, None, 1xf32) dropout_36, dropout_37 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_54, None, full_4, False, "upscale_in_train", 0, False + add_54, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) - del add_54 + del add_54, full_4 # pd_op.add: (1x27x768xf32) <- (1x27x768xf32, 1x27x768xf32) add_55 = paddle._C_ops.add(layer_norm_33, dropout_36) + del dropout_36, layer_norm_33 # pd_op.layer_norm: (1x27x768xf32, 1x27xf32, 1x27xf32) <- (1x27x768xf32, 768xf32, 768xf32) layer_norm_36, layer_norm_37, layer_norm_38 = (lambda x, f: f(x))( @@ -1169,7 +1160,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_2, parameter_3 + del add_55, parameter_2, parameter_3 # pd_op.full_int_array: (1xi64) <- () full_int_array_3 = [0] @@ -1181,245 +1172,18 @@ def forward( slice_0 = paddle._C_ops.slice( layer_norm_36, [1], full_int_array_3, full_int_array_4, [1], [1] ) + del full_int_array_3, full_int_array_4, layer_norm_36 # pd_op.matmul: (1x768xf32) <- (1x768xf32, 768x768xf32) matmul_48 = paddle._C_ops.matmul(slice_0, parameter_1, False, False) - del parameter_1 + del parameter_1, slice_0 # pd_op.add: (1x768xf32) <- (1x768xf32, 768xf32) add_56 = paddle._C_ops.add(matmul_48, parameter_0) - del parameter_0 + del matmul_48, parameter_0 # pd_op.tanh: (1x768xf32) <- (1x768xf32) tanh_0 = paddle._C_ops.tanh(add_56) - del ( - add_0, - add_1, - add_10, - add_11, - add_12, - add_13, - add_16, - add_19, - add_2, - add_20, - add_21, - add_22, - add_25, - add_28, - add_29, - add_3, - add_30, - add_31, - add_34, - add_37, - add_38, - add_39, - add_4, - add_40, - add_43, - add_46, - add_47, - add_48, - add_49, - add_52, - add_55, - add_56, - add_7, - assign_0, - assign_1, - assign_10, - assign_11, - assign_12, - assign_13, - assign_14, - assign_15, - assign_16, - assign_17, - assign_18, - assign_19, - assign_2, - assign_20, - assign_21, - assign_22, - assign_3, - assign_4, - assign_5, - assign_6, - assign_7, - assign_8, - assign_9, - dropout_0, - dropout_1, - dropout_10, - dropout_11, - dropout_12, - dropout_13, - dropout_14, - dropout_15, - dropout_16, - dropout_17, - dropout_18, - dropout_19, - dropout_2, - dropout_20, - dropout_21, - dropout_22, - dropout_23, - dropout_24, - dropout_25, - dropout_26, - dropout_27, - dropout_28, - dropout_29, - dropout_3, - dropout_30, - dropout_31, - dropout_32, - dropout_33, - dropout_34, - dropout_35, - dropout_36, - dropout_37, - dropout_4, - dropout_5, - dropout_6, - dropout_7, - dropout_8, - dropout_9, - embedding_0, - embedding_1, - embedding_2, - full_4, - full_5, - full_int_array_3, - full_int_array_4, - layer_norm_1, - layer_norm_10, - layer_norm_11, - layer_norm_12, - layer_norm_13, - layer_norm_14, - layer_norm_15, - layer_norm_16, - layer_norm_17, - layer_norm_18, - layer_norm_19, - layer_norm_2, - layer_norm_20, - layer_norm_21, - layer_norm_22, - layer_norm_23, - layer_norm_24, - layer_norm_25, - layer_norm_26, - layer_norm_27, - layer_norm_28, - layer_norm_29, - layer_norm_3, - layer_norm_30, - layer_norm_31, - layer_norm_32, - layer_norm_33, - layer_norm_34, - layer_norm_35, - layer_norm_36, - layer_norm_37, - layer_norm_38, - layer_norm_4, - layer_norm_5, - layer_norm_6, - layer_norm_7, - layer_norm_8, - layer_norm_9, - matmul_0, - matmul_1, - matmul_10, - matmul_11, - matmul_13, - matmul_14, - matmul_15, - matmul_16, - matmul_17, - matmul_18, - matmul_19, - matmul_2, - matmul_21, - matmul_22, - matmul_23, - matmul_24, - matmul_25, - matmul_26, - matmul_27, - matmul_29, - matmul_3, - matmul_30, - matmul_31, - matmul_32, - matmul_33, - matmul_34, - matmul_35, - matmul_37, - matmul_38, - matmul_39, - matmul_40, - matmul_41, - matmul_42, - matmul_43, - matmul_45, - matmul_46, - matmul_47, - matmul_48, - matmul_5, - matmul_6, - matmul_7, - matmul_8, - matmul_9, - relu_0, - relu_1, - relu_2, - relu_3, - relu_4, - relu_5, - reshape_11, - reshape_15, - reshape_19, - reshape_23, - reshape_3, - reshape_7, - scale_1, - scale_2, - scale_3, - scale_4, - scale_5, - scale_6, - slice_0, - softmax_0, - softmax_1, - softmax_2, - softmax_3, - softmax_4, - softmax_5, - subtract_0, - transpose_1, - transpose_10, - transpose_11, - transpose_13, - transpose_14, - transpose_15, - transpose_17, - transpose_18, - transpose_19, - transpose_2, - transpose_21, - transpose_22, - transpose_23, - transpose_3, - transpose_5, - transpose_6, - transpose_7, - transpose_9, - unsqueeze_0, - ) + del add_56 return tanh_0 diff --git a/paddle_samples/PaddleNLP/ppminilm-6l-768h/weight_meta.py b/paddle_samples/PaddleNLP/ppminilm-6l-768h/weight_meta.py index 067f27be6..53ebe80f1 100644 --- a/paddle_samples/PaddleNLP/ppminilm-6l-768h/weight_meta.py +++ b/paddle_samples/PaddleNLP/ppminilm-6l-768h/weight_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_parameter_0: name = "parameter_0" + original_name = "linear_4146.b_0" shape = [768] dtype = "float32" data = None @@ -7,6 +8,7 @@ class Program_weight_tensor_parameter_0: class Program_weight_tensor_parameter_1: name = "parameter_1" + original_name = "linear_4146.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.0962126") @@ -18,6 +20,7 @@ class Program_weight_tensor_parameter_1: class Program_weight_tensor_parameter_2: name = "parameter_2" + original_name = "layer_norm_1430.b_0" shape = [768] dtype = "float32" data = None @@ -25,6 +28,7 @@ class Program_weight_tensor_parameter_2: class Program_weight_tensor_parameter_3: name = "parameter_3" + original_name = "layer_norm_1430.w_0" shape = [768] dtype = "float32" min_val = float("1.0") @@ -35,6 +39,7 @@ class Program_weight_tensor_parameter_3: class Program_weight_tensor_parameter_4: name = "parameter_4" + original_name = "layer_norm_1429.b_0" shape = [768] dtype = "float32" data = None @@ -42,6 +47,7 @@ class Program_weight_tensor_parameter_4: class Program_weight_tensor_parameter_5: name = "parameter_5" + original_name = "layer_norm_1429.w_0" shape = [768] dtype = "float32" min_val = float("1.0") @@ -52,6 +58,7 @@ class Program_weight_tensor_parameter_5: class Program_weight_tensor_parameter_6: name = "parameter_6" + original_name = "linear_4145.b_0" shape = [768] dtype = "float32" data = None @@ -59,6 +66,7 @@ class Program_weight_tensor_parameter_6: class Program_weight_tensor_parameter_7: name = "parameter_7" + original_name = "linear_4145.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-0.102023") @@ -70,6 +78,7 @@ class Program_weight_tensor_parameter_7: class Program_weight_tensor_parameter_8: name = "parameter_8" + original_name = "linear_4144.b_0" shape = [3072] dtype = "float32" data = None @@ -77,6 +86,7 @@ class Program_weight_tensor_parameter_8: class Program_weight_tensor_parameter_9: name = "parameter_9" + original_name = "linear_4144.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.104008") @@ -88,6 +98,7 @@ class Program_weight_tensor_parameter_9: class Program_weight_tensor_parameter_10: name = "parameter_10" + original_name = "linear_4143.b_0" shape = [768] dtype = "float32" data = None @@ -95,6 +106,7 @@ class Program_weight_tensor_parameter_10: class Program_weight_tensor_parameter_11: name = "parameter_11" + original_name = "linear_4143.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.0972951") @@ -106,6 +118,7 @@ class Program_weight_tensor_parameter_11: class Program_weight_tensor_parameter_12: name = "parameter_12" + original_name = "linear_4142.b_0" shape = [768] dtype = "float32" min_val = float("-0.272252") @@ -117,6 +130,7 @@ class Program_weight_tensor_parameter_12: class Program_weight_tensor_parameter_13: name = "parameter_13" + original_name = "linear_4142.w_0" shape = [768, 768] dtype = "float32" min_val = float("-1.13582") @@ -128,6 +142,7 @@ class Program_weight_tensor_parameter_13: class Program_weight_tensor_parameter_14: name = "parameter_14" + original_name = "linear_4141.b_0" shape = [768] dtype = "float32" min_val = float("-0.335672") @@ -139,6 +154,7 @@ class Program_weight_tensor_parameter_14: class Program_weight_tensor_parameter_15: name = "parameter_15" + original_name = "linear_4141.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.988077") @@ -150,6 +166,7 @@ class Program_weight_tensor_parameter_15: class Program_weight_tensor_parameter_16: name = "parameter_16" + original_name = "linear_4140.b_0" shape = [768] dtype = "float32" min_val = float("-1.38708") @@ -161,6 +178,7 @@ class Program_weight_tensor_parameter_16: class Program_weight_tensor_parameter_17: name = "parameter_17" + original_name = "linear_4140.w_0" shape = [768, 768] dtype = "float32" min_val = float("-1.07496") @@ -172,6 +190,7 @@ class Program_weight_tensor_parameter_17: class Program_weight_tensor_parameter_18: name = "parameter_18" + original_name = "layer_norm_1428.b_0" shape = [768] dtype = "float32" min_val = float("-0.368236") @@ -183,6 +202,7 @@ class Program_weight_tensor_parameter_18: class Program_weight_tensor_parameter_19: name = "parameter_19" + original_name = "layer_norm_1428.w_0" shape = [768] dtype = "float32" min_val = float("0.0446708") @@ -194,6 +214,7 @@ class Program_weight_tensor_parameter_19: class Program_weight_tensor_parameter_20: name = "parameter_20" + original_name = "layer_norm_1427.b_0" shape = [768] dtype = "float32" min_val = float("-3.97666") @@ -205,6 +226,7 @@ class Program_weight_tensor_parameter_20: class Program_weight_tensor_parameter_21: name = "parameter_21" + original_name = "layer_norm_1427.w_0" shape = [768] dtype = "float32" min_val = float("0.181201") @@ -216,6 +238,7 @@ class Program_weight_tensor_parameter_21: class Program_weight_tensor_parameter_22: name = "parameter_22" + original_name = "linear_4139.b_0" shape = [768] dtype = "float32" min_val = float("-0.454431") @@ -227,6 +250,7 @@ class Program_weight_tensor_parameter_22: class Program_weight_tensor_parameter_23: name = "parameter_23" + original_name = "linear_4139.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-4.14209") @@ -238,6 +262,7 @@ class Program_weight_tensor_parameter_23: class Program_weight_tensor_parameter_24: name = "parameter_24" + original_name = "linear_4138.b_0" shape = [3072] dtype = "float32" min_val = float("-1.1501") @@ -249,6 +274,7 @@ class Program_weight_tensor_parameter_24: class Program_weight_tensor_parameter_25: name = "parameter_25" + original_name = "linear_4138.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-2.87451") @@ -260,6 +286,7 @@ class Program_weight_tensor_parameter_25: class Program_weight_tensor_parameter_26: name = "parameter_26" + original_name = "linear_4137.b_0" shape = [768] dtype = "float32" min_val = float("-0.32939") @@ -271,6 +298,7 @@ class Program_weight_tensor_parameter_26: class Program_weight_tensor_parameter_27: name = "parameter_27" + original_name = "linear_4137.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.519336") @@ -282,6 +310,7 @@ class Program_weight_tensor_parameter_27: class Program_weight_tensor_parameter_28: name = "parameter_28" + original_name = "linear_4136.b_0" shape = [768] dtype = "float32" min_val = float("-0.121544") @@ -293,6 +322,7 @@ class Program_weight_tensor_parameter_28: class Program_weight_tensor_parameter_29: name = "parameter_29" + original_name = "linear_4136.w_0" shape = [768, 768] dtype = "float32" min_val = float("-1.1911") @@ -304,6 +334,7 @@ class Program_weight_tensor_parameter_29: class Program_weight_tensor_parameter_30: name = "parameter_30" + original_name = "linear_4135.b_0" shape = [768] dtype = "float32" min_val = float("-108.205") @@ -315,6 +346,7 @@ class Program_weight_tensor_parameter_30: class Program_weight_tensor_parameter_31: name = "parameter_31" + original_name = "linear_4135.w_0" shape = [768, 768] dtype = "float32" min_val = float("-2.37555") @@ -326,6 +358,7 @@ class Program_weight_tensor_parameter_31: class Program_weight_tensor_parameter_32: name = "parameter_32" + original_name = "linear_4134.b_0" shape = [768] dtype = "float32" min_val = float("-1.43141") @@ -337,6 +370,7 @@ class Program_weight_tensor_parameter_32: class Program_weight_tensor_parameter_33: name = "parameter_33" + original_name = "linear_4134.w_0" shape = [768, 768] dtype = "float32" min_val = float("-1.75537") @@ -348,6 +382,7 @@ class Program_weight_tensor_parameter_33: class Program_weight_tensor_parameter_34: name = "parameter_34" + original_name = "layer_norm_1426.b_0" shape = [768] dtype = "float32" min_val = float("-0.248581") @@ -359,6 +394,7 @@ class Program_weight_tensor_parameter_34: class Program_weight_tensor_parameter_35: name = "parameter_35" + original_name = "layer_norm_1426.w_0" shape = [768] dtype = "float32" min_val = float("0.0708327") @@ -370,6 +406,7 @@ class Program_weight_tensor_parameter_35: class Program_weight_tensor_parameter_36: name = "parameter_36" + original_name = "layer_norm_1425.b_0" shape = [768] dtype = "float32" min_val = float("-2.88754") @@ -381,6 +418,7 @@ class Program_weight_tensor_parameter_36: class Program_weight_tensor_parameter_37: name = "parameter_37" + original_name = "layer_norm_1425.w_0" shape = [768] dtype = "float32" min_val = float("0.213113") @@ -392,6 +430,7 @@ class Program_weight_tensor_parameter_37: class Program_weight_tensor_parameter_38: name = "parameter_38" + original_name = "linear_4133.b_0" shape = [768] dtype = "float32" min_val = float("-0.351474") @@ -403,6 +442,7 @@ class Program_weight_tensor_parameter_38: class Program_weight_tensor_parameter_39: name = "parameter_39" + original_name = "linear_4133.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-1.91562") @@ -414,6 +454,7 @@ class Program_weight_tensor_parameter_39: class Program_weight_tensor_parameter_40: name = "parameter_40" + original_name = "linear_4132.b_0" shape = [3072] dtype = "float32" min_val = float("-1.15322") @@ -425,6 +466,7 @@ class Program_weight_tensor_parameter_40: class Program_weight_tensor_parameter_41: name = "parameter_41" + original_name = "linear_4132.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.840755") @@ -436,6 +478,7 @@ class Program_weight_tensor_parameter_41: class Program_weight_tensor_parameter_42: name = "parameter_42" + original_name = "linear_4131.b_0" shape = [768] dtype = "float32" min_val = float("-0.182525") @@ -447,6 +490,7 @@ class Program_weight_tensor_parameter_42: class Program_weight_tensor_parameter_43: name = "parameter_43" + original_name = "linear_4131.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.768259") @@ -458,6 +502,7 @@ class Program_weight_tensor_parameter_43: class Program_weight_tensor_parameter_44: name = "parameter_44" + original_name = "linear_4130.b_0" shape = [768] dtype = "float32" min_val = float("-0.198252") @@ -469,6 +514,7 @@ class Program_weight_tensor_parameter_44: class Program_weight_tensor_parameter_45: name = "parameter_45" + original_name = "linear_4130.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.614578") @@ -480,6 +526,7 @@ class Program_weight_tensor_parameter_45: class Program_weight_tensor_parameter_46: name = "parameter_46" + original_name = "linear_4129.b_0" shape = [768] dtype = "float32" min_val = float("-65.625") @@ -491,6 +538,7 @@ class Program_weight_tensor_parameter_46: class Program_weight_tensor_parameter_47: name = "parameter_47" + original_name = "linear_4129.w_0" shape = [768, 768] dtype = "float32" min_val = float("-2.14395") @@ -502,6 +550,7 @@ class Program_weight_tensor_parameter_47: class Program_weight_tensor_parameter_48: name = "parameter_48" + original_name = "linear_4128.b_0" shape = [768] dtype = "float32" min_val = float("-1.42042") @@ -513,6 +562,7 @@ class Program_weight_tensor_parameter_48: class Program_weight_tensor_parameter_49: name = "parameter_49" + original_name = "linear_4128.w_0" shape = [768, 768] dtype = "float32" min_val = float("-1.64399") @@ -524,6 +574,7 @@ class Program_weight_tensor_parameter_49: class Program_weight_tensor_parameter_50: name = "parameter_50" + original_name = "layer_norm_1424.b_0" shape = [768] dtype = "float32" min_val = float("-0.368219") @@ -535,6 +586,7 @@ class Program_weight_tensor_parameter_50: class Program_weight_tensor_parameter_51: name = "parameter_51" + original_name = "layer_norm_1424.w_0" shape = [768] dtype = "float32" min_val = float("0.0278934") @@ -546,6 +598,7 @@ class Program_weight_tensor_parameter_51: class Program_weight_tensor_parameter_52: name = "parameter_52" + original_name = "layer_norm_1423.b_0" shape = [768] dtype = "float32" min_val = float("-4.49373") @@ -557,6 +610,7 @@ class Program_weight_tensor_parameter_52: class Program_weight_tensor_parameter_53: name = "parameter_53" + original_name = "layer_norm_1423.w_0" shape = [768] dtype = "float32" min_val = float("0.155214") @@ -568,6 +622,7 @@ class Program_weight_tensor_parameter_53: class Program_weight_tensor_parameter_54: name = "parameter_54" + original_name = "linear_4127.b_0" shape = [768] dtype = "float32" min_val = float("-0.848256") @@ -579,6 +634,7 @@ class Program_weight_tensor_parameter_54: class Program_weight_tensor_parameter_55: name = "parameter_55" + original_name = "linear_4127.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-48.6055") @@ -590,6 +646,7 @@ class Program_weight_tensor_parameter_55: class Program_weight_tensor_parameter_56: name = "parameter_56" + original_name = "linear_4126.b_0" shape = [3072] dtype = "float32" min_val = float("-0.719523") @@ -601,6 +658,7 @@ class Program_weight_tensor_parameter_56: class Program_weight_tensor_parameter_57: name = "parameter_57" + original_name = "linear_4126.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.922014") @@ -612,6 +670,7 @@ class Program_weight_tensor_parameter_57: class Program_weight_tensor_parameter_58: name = "parameter_58" + original_name = "linear_4125.b_0" shape = [768] dtype = "float32" min_val = float("-0.157812") @@ -623,6 +682,7 @@ class Program_weight_tensor_parameter_58: class Program_weight_tensor_parameter_59: name = "parameter_59" + original_name = "linear_4125.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.647719") @@ -634,6 +694,7 @@ class Program_weight_tensor_parameter_59: class Program_weight_tensor_parameter_60: name = "parameter_60" + original_name = "linear_4124.b_0" shape = [768] dtype = "float32" min_val = float("-0.145558") @@ -645,6 +706,7 @@ class Program_weight_tensor_parameter_60: class Program_weight_tensor_parameter_61: name = "parameter_61" + original_name = "linear_4124.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.534983") @@ -656,6 +718,7 @@ class Program_weight_tensor_parameter_61: class Program_weight_tensor_parameter_62: name = "parameter_62" + original_name = "linear_4123.b_0" shape = [768] dtype = "float32" min_val = float("-67.9141") @@ -667,6 +730,7 @@ class Program_weight_tensor_parameter_62: class Program_weight_tensor_parameter_63: name = "parameter_63" + original_name = "linear_4123.w_0" shape = [768, 768] dtype = "float32" min_val = float("-2.801") @@ -678,6 +742,7 @@ class Program_weight_tensor_parameter_63: class Program_weight_tensor_parameter_64: name = "parameter_64" + original_name = "linear_4122.b_0" shape = [768] dtype = "float32" min_val = float("-1.46384") @@ -689,6 +754,7 @@ class Program_weight_tensor_parameter_64: class Program_weight_tensor_parameter_65: name = "parameter_65" + original_name = "linear_4122.w_0" shape = [768, 768] dtype = "float32" min_val = float("-1.51038") @@ -700,6 +766,7 @@ class Program_weight_tensor_parameter_65: class Program_weight_tensor_parameter_66: name = "parameter_66" + original_name = "layer_norm_1422.b_0" shape = [768] dtype = "float32" min_val = float("-0.482284") @@ -711,6 +778,7 @@ class Program_weight_tensor_parameter_66: class Program_weight_tensor_parameter_67: name = "parameter_67" + original_name = "layer_norm_1422.w_0" shape = [768] dtype = "float32" min_val = float("0.0091019") @@ -722,6 +790,7 @@ class Program_weight_tensor_parameter_67: class Program_weight_tensor_parameter_68: name = "parameter_68" + original_name = "layer_norm_1421.b_0" shape = [768] dtype = "float32" min_val = float("-6.56632") @@ -733,6 +802,7 @@ class Program_weight_tensor_parameter_68: class Program_weight_tensor_parameter_69: name = "parameter_69" + original_name = "layer_norm_1421.w_0" shape = [768] dtype = "float32" min_val = float("0.135737") @@ -744,6 +814,7 @@ class Program_weight_tensor_parameter_69: class Program_weight_tensor_parameter_70: name = "parameter_70" + original_name = "linear_4121.b_0" shape = [768] dtype = "float32" min_val = float("-1.0718") @@ -755,6 +826,7 @@ class Program_weight_tensor_parameter_70: class Program_weight_tensor_parameter_71: name = "parameter_71" + original_name = "linear_4121.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-14.1658") @@ -766,6 +838,7 @@ class Program_weight_tensor_parameter_71: class Program_weight_tensor_parameter_72: name = "parameter_72" + original_name = "linear_4120.b_0" shape = [3072] dtype = "float32" min_val = float("-0.77803") @@ -777,6 +850,7 @@ class Program_weight_tensor_parameter_72: class Program_weight_tensor_parameter_73: name = "parameter_73" + original_name = "linear_4120.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-1.58726") @@ -788,6 +862,7 @@ class Program_weight_tensor_parameter_73: class Program_weight_tensor_parameter_74: name = "parameter_74" + original_name = "linear_4119.b_0" shape = [768] dtype = "float32" min_val = float("-0.187265") @@ -799,6 +874,7 @@ class Program_weight_tensor_parameter_74: class Program_weight_tensor_parameter_75: name = "parameter_75" + original_name = "linear_4119.w_0" shape = [768, 768] dtype = "float32" min_val = float("-1.48163") @@ -810,6 +886,7 @@ class Program_weight_tensor_parameter_75: class Program_weight_tensor_parameter_76: name = "parameter_76" + original_name = "linear_4118.b_0" shape = [768] dtype = "float32" min_val = float("-0.202134") @@ -821,6 +898,7 @@ class Program_weight_tensor_parameter_76: class Program_weight_tensor_parameter_77: name = "parameter_77" + original_name = "linear_4118.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.748471") @@ -832,6 +910,7 @@ class Program_weight_tensor_parameter_77: class Program_weight_tensor_parameter_78: name = "parameter_78" + original_name = "linear_4117.b_0" shape = [768] dtype = "float32" min_val = float("-58.014") @@ -843,6 +922,7 @@ class Program_weight_tensor_parameter_78: class Program_weight_tensor_parameter_79: name = "parameter_79" + original_name = "linear_4117.w_0" shape = [768, 768] dtype = "float32" min_val = float("-4.90016") @@ -854,6 +934,7 @@ class Program_weight_tensor_parameter_79: class Program_weight_tensor_parameter_80: name = "parameter_80" + original_name = "linear_4116.b_0" shape = [768] dtype = "float32" min_val = float("-1.84067") @@ -865,6 +946,7 @@ class Program_weight_tensor_parameter_80: class Program_weight_tensor_parameter_81: name = "parameter_81" + original_name = "linear_4116.w_0" shape = [768, 768] dtype = "float32" min_val = float("-1.77878") @@ -876,6 +958,7 @@ class Program_weight_tensor_parameter_81: class Program_weight_tensor_parameter_82: name = "parameter_82" + original_name = "layer_norm_1420.b_0" shape = [768] dtype = "float32" min_val = float("-0.952046") @@ -887,6 +970,7 @@ class Program_weight_tensor_parameter_82: class Program_weight_tensor_parameter_83: name = "parameter_83" + original_name = "layer_norm_1420.w_0" shape = [768] dtype = "float32" min_val = float("0.0031259") @@ -898,6 +982,7 @@ class Program_weight_tensor_parameter_83: class Program_weight_tensor_parameter_84: name = "parameter_84" + original_name = "layer_norm_1419.b_0" shape = [768] dtype = "float32" min_val = float("-6.65568") @@ -909,6 +994,7 @@ class Program_weight_tensor_parameter_84: class Program_weight_tensor_parameter_85: name = "parameter_85" + original_name = "layer_norm_1419.w_0" shape = [768] dtype = "float32" min_val = float("0.0682864") @@ -920,6 +1006,7 @@ class Program_weight_tensor_parameter_85: class Program_weight_tensor_parameter_86: name = "parameter_86" + original_name = "linear_4115.b_0" shape = [768] dtype = "float32" min_val = float("-1.69164") @@ -931,6 +1018,7 @@ class Program_weight_tensor_parameter_86: class Program_weight_tensor_parameter_87: name = "parameter_87" + original_name = "linear_4115.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-20.093") @@ -942,6 +1030,7 @@ class Program_weight_tensor_parameter_87: class Program_weight_tensor_parameter_88: name = "parameter_88" + original_name = "linear_4114.b_0" shape = [3072] dtype = "float32" min_val = float("-1.30696") @@ -953,6 +1042,7 @@ class Program_weight_tensor_parameter_88: class Program_weight_tensor_parameter_89: name = "parameter_89" + original_name = "linear_4114.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-3.65486") @@ -964,6 +1054,7 @@ class Program_weight_tensor_parameter_89: class Program_weight_tensor_parameter_90: name = "parameter_90" + original_name = "linear_4113.b_0" shape = [768] dtype = "float32" min_val = float("-0.465067") @@ -975,6 +1066,7 @@ class Program_weight_tensor_parameter_90: class Program_weight_tensor_parameter_91: name = "parameter_91" + original_name = "linear_4113.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.849994") @@ -986,6 +1078,7 @@ class Program_weight_tensor_parameter_91: class Program_weight_tensor_parameter_92: name = "parameter_92" + original_name = "linear_4112.b_0" shape = [768] dtype = "float32" min_val = float("-0.336198") @@ -997,6 +1090,7 @@ class Program_weight_tensor_parameter_92: class Program_weight_tensor_parameter_93: name = "parameter_93" + original_name = "linear_4112.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.316443") @@ -1008,6 +1102,7 @@ class Program_weight_tensor_parameter_93: class Program_weight_tensor_parameter_94: name = "parameter_94" + original_name = "linear_4111.b_0" shape = [768] dtype = "float32" min_val = float("-44.6574") @@ -1019,6 +1114,7 @@ class Program_weight_tensor_parameter_94: class Program_weight_tensor_parameter_95: name = "parameter_95" + original_name = "linear_4111.w_0" shape = [768, 768] dtype = "float32" min_val = float("-2.64678") @@ -1030,6 +1126,7 @@ class Program_weight_tensor_parameter_95: class Program_weight_tensor_parameter_96: name = "parameter_96" + original_name = "linear_4110.b_0" shape = [768] dtype = "float32" min_val = float("-3.11347") @@ -1041,6 +1138,7 @@ class Program_weight_tensor_parameter_96: class Program_weight_tensor_parameter_97: name = "parameter_97" + original_name = "linear_4110.w_0" shape = [768, 768] dtype = "float32" min_val = float("-1.44957") @@ -1052,6 +1150,7 @@ class Program_weight_tensor_parameter_97: class Program_weight_tensor_parameter_98: name = "parameter_98" + original_name = "layer_norm_1418.b_0" shape = [768] dtype = "float32" min_val = float("-0.301262") @@ -1063,6 +1162,7 @@ class Program_weight_tensor_parameter_98: class Program_weight_tensor_parameter_99: name = "parameter_99" + original_name = "layer_norm_1418.w_0" shape = [768] dtype = "float32" min_val = float("0.0474554") @@ -1074,6 +1174,7 @@ class Program_weight_tensor_parameter_99: class Program_weight_tensor_parameter_100: name = "parameter_100" + original_name = "embedding_255.w_0" shape = [2, 768] dtype = "float32" min_val = float("-3.76786") @@ -1085,6 +1186,7 @@ class Program_weight_tensor_parameter_100: class Program_weight_tensor_parameter_101: name = "parameter_101" + original_name = "embedding_254.w_0" shape = [512, 768] dtype = "float32" min_val = float("-2.88106") @@ -1096,6 +1198,7 @@ class Program_weight_tensor_parameter_101: class Program_weight_tensor_parameter_102: name = "parameter_102" + original_name = "embedding_253.w_0" shape = [21128, 768] dtype = "float32" min_val = float("-41.8125") diff --git a/paddle_samples/PaddleNLP/rocketqa-base-cross-encoder/input_meta.py b/paddle_samples/PaddleNLP/rocketqa-base-cross-encoder/input_meta.py index 4f25a05a9..055cd8978 100644 --- a/paddle_samples/PaddleNLP/rocketqa-base-cross-encoder/input_meta.py +++ b/paddle_samples/PaddleNLP/rocketqa-base-cross-encoder/input_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_data_0: name = "data_0" + original_name = "input_ids" shape = [1, 21] dtype = "int64" data = [ @@ -29,6 +30,7 @@ class Program_weight_tensor_data_0: class Program_weight_tensor_data_1: name = "data_1" + original_name = "token_type_ids" shape = [1, 21] dtype = "int64" data = [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0] diff --git a/paddle_samples/PaddleNLP/rocketqa-base-cross-encoder/model.py b/paddle_samples/PaddleNLP/rocketqa-base-cross-encoder/model.py index 304b0f9f0..60020c086 100644 --- a/paddle_samples/PaddleNLP/rocketqa-base-cross-encoder/model.py +++ b/paddle_samples/PaddleNLP/rocketqa-base-cross-encoder/model.py @@ -266,10 +266,11 @@ def forward( # pd_op.embedding: (1x21x768xf32) <- (1x21xi64, 2048x768xf32) embedding_1 = paddle._C_ops.embedding(subtract_0, parameter_198, -1, False) - del parameter_198 + del parameter_198, subtract_0 # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 1x21x768xf32) add_0 = paddle._C_ops.add(embedding_0, embedding_1) + del embedding_0, embedding_1 # pd_op.embedding: (1x21x768xf32) <- (1x21xi64, 4x768xf32) embedding_2 = paddle._C_ops.embedding(data_1, parameter_197, -1, False) @@ -277,6 +278,7 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 1x21x768xf32) add_1 = paddle._C_ops.add(add_0, embedding_2) + del add_0, embedding_2 # pd_op.full: (1xf32) <- () full_4 = paddle._C_ops.full( @@ -289,10 +291,11 @@ def forward( # pd_op.embedding: (1x21x768xf32) <- (1x21xi64, 3x768xf32) embedding_3 = paddle._C_ops.embedding(scale_1, parameter_196, -1, False) - del parameter_196 + del parameter_196, scale_1 # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 1x21x768xf32) add_2 = paddle._C_ops.add(add_1, embedding_3) + del add_1, embedding_3 # pd_op.layer_norm: (1x21x768xf32, 1x21xf32, 1x21xf32) <- (1x21x768xf32, 768xf32, 768xf32) layer_norm_0, layer_norm_1, layer_norm_2 = (lambda x, f: f(x))( @@ -301,125 +304,17 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_194, parameter_195 + del add_2, parameter_194, parameter_195 # pd_op.full: (1xf32) <- () full_5 = paddle._C_ops.full( [1], float("0.1"), paddle.float32, paddle.core.CPUPlace() ) - # pd_op.assign: (1xf32) <- (1xf32) - assign_0 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_1 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_2 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_3 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_4 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_5 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_6 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_7 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_8 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_9 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_10 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_11 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_12 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_13 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_14 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_15 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_16 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_17 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_18 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_19 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_20 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_21 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_22 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_23 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_24 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_25 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_26 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_27 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_28 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_29 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_30 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_31 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_32 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_33 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_34 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_35 = full_5 - # pd_op.dropout: (1x21x768xf32, 1x21x768xui8) <- (1x21x768xf32, None, 1xf32) dropout_0, dropout_1 = (lambda x, f: f(x))( paddle._C_ops.dropout( - layer_norm_0, None, full_5, False, "upscale_in_train", 0, False + layer_norm_0, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -431,13 +326,14 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_3 = paddle._C_ops.add(matmul_0, parameter_192) - del parameter_192 + del matmul_0, parameter_192 # pd_op.full_int_array: (4xi64) <- () full_int_array_1 = [0, 0, 12, 64] # pd_op.reshape: (1x21x12x64xf32) <- (1x21x768xf32, 4xi64) reshape_0 = paddle._C_ops.reshape(add_3, full_int_array_1) + del add_3 # pd_op.transpose: (1x12x21x64xf32) <- (1x21x12x64xf32) transpose_0 = paddle._C_ops.transpose(reshape_0, [0, 2, 1, 3]) @@ -449,7 +345,7 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_4 = paddle._C_ops.add(matmul_1, parameter_190) - del parameter_190 + del matmul_1, parameter_190 # pd_op.matmul: (1x21x768xf32) <- (1x21x768xf32, 768x768xf32) matmul_2 = paddle._C_ops.matmul(dropout_0, parameter_189, False, False) @@ -457,10 +353,11 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_5 = paddle._C_ops.add(matmul_2, parameter_188) - del parameter_188 + del matmul_2, parameter_188 # pd_op.reshape: (1x21x12x64xf32) <- (1x21x768xf32, 4xi64) reshape_1 = paddle._C_ops.reshape(add_4, full_int_array_1) + del add_4 # pd_op.transpose: (1x12x21x64xf32) <- (1x21x12x64xf32) transpose_1 = paddle._C_ops.transpose(reshape_1, [0, 2, 1, 3]) @@ -468,6 +365,7 @@ def forward( # pd_op.reshape: (1x21x12x64xf32) <- (1x21x768xf32, 4xi64) reshape_2 = paddle._C_ops.reshape(add_5, full_int_array_1) + del add_5 # pd_op.transpose: (1x12x21x64xf32) <- (1x21x12x64xf32) transpose_2 = paddle._C_ops.transpose(reshape_2, [0, 2, 1, 3]) @@ -478,48 +376,17 @@ def forward( [1], float("0.125"), paddle.float32, paddle.core.CPUPlace() ) - # pd_op.assign: (1xf32) <- (1xf32) - assign_36 = full_6 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_37 = full_6 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_38 = full_6 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_39 = full_6 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_40 = full_6 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_41 = full_6 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_42 = full_6 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_43 = full_6 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_44 = full_6 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_45 = full_6 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_46 = full_6 - # pd_op.scale: (1x12x21x64xf32) <- (1x12x21x64xf32, 1xf32) scale_2 = paddle._C_ops.scale(transpose_0, full_6, float("0"), True) del transpose_0 # pd_op.matmul: (1x12x21x21xf32) <- (1x12x21x64xf32, 1x12x21x64xf32) matmul_3 = paddle._C_ops.matmul(scale_2, transpose_1, False, True) + del scale_2, transpose_1 # pd_op.add: (1x12x21x21xf32) <- (1x12x21x21xf32, 1x1x1x21xf32) add_6 = paddle._C_ops.add(matmul_3, unsqueeze_0) + del matmul_3 # pd_op.softmax: (1x12x21x21xf32) <- (1x12x21x21xf32) softmax_0 = paddle._C_ops.softmax(add_6, -1) @@ -528,13 +395,15 @@ def forward( # pd_op.dropout: (1x12x21x21xf32, 1x12x21x21xui8) <- (1x12x21x21xf32, None, 1xf32) dropout_2, dropout_3 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_0, None, full_5, False, "upscale_in_train", 0, False + softmax_0, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_0 # pd_op.matmul: (1x12x21x64xf32) <- (1x12x21x21xf32, 1x12x21x64xf32) matmul_4 = paddle._C_ops.matmul(dropout_2, transpose_2, False, False) + del dropout_2, transpose_2 # pd_op.transpose: (1x21x12x64xf32) <- (1x12x21x64xf32) transpose_3 = paddle._C_ops.transpose(matmul_4, [0, 2, 1, 3]) @@ -545,19 +414,20 @@ def forward( # pd_op.reshape: (1x21x768xf32) <- (1x21x12x64xf32, 3xi64) reshape_3 = paddle._C_ops.reshape(transpose_3, full_int_array_2) + del transpose_3 # pd_op.matmul: (1x21x768xf32) <- (1x21x768xf32, 768x768xf32) matmul_5 = paddle._C_ops.matmul(reshape_3, parameter_187, False, False) - del parameter_187 + del parameter_187, reshape_3 # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_7 = paddle._C_ops.add(matmul_5, parameter_186) - del parameter_186 + del matmul_5, parameter_186 # pd_op.dropout: (1x21x768xf32, 1x21x768xui8) <- (1x21x768xf32, None, 1xf32) dropout_4, dropout_5 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_7, None, full_5, False, "upscale_in_train", 0, False + add_7, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -565,6 +435,7 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 1x21x768xf32) add_8 = paddle._C_ops.add(dropout_0, dropout_4) + del dropout_0, dropout_4 # pd_op.layer_norm: (1x21x768xf32, 1x21xf32, 1x21xf32) <- (1x21x768xf32, 768xf32, 768xf32) layer_norm_3, layer_norm_4, layer_norm_5 = (lambda x, f: f(x))( @@ -573,7 +444,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_180, parameter_181 + del add_8, parameter_180, parameter_181 # pd_op.matmul: (1x21x3072xf32) <- (1x21x768xf32, 768x3072xf32) matmul_6 = paddle._C_ops.matmul(layer_norm_3, parameter_185, False, False) @@ -581,23 +452,24 @@ def forward( # pd_op.add: (1x21x3072xf32) <- (1x21x3072xf32, 3072xf32) add_9 = paddle._C_ops.add(matmul_6, parameter_184) - del parameter_184 + del matmul_6, parameter_184 # pd_op.gelu: (1x21x3072xf32) <- (1x21x3072xf32) gelu_0 = paddle._C_ops.gelu(add_9, False) + del add_9 # pd_op.matmul: (1x21x768xf32) <- (1x21x3072xf32, 3072x768xf32) matmul_7 = paddle._C_ops.matmul(gelu_0, parameter_183, False, False) - del parameter_183 + del gelu_0, parameter_183 # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_10 = paddle._C_ops.add(matmul_7, parameter_182) - del parameter_182 + del matmul_7, parameter_182 # pd_op.dropout: (1x21x768xf32, 1x21x768xui8) <- (1x21x768xf32, None, 1xf32) dropout_6, dropout_7 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_10, None, full_5, False, "upscale_in_train", 0, False + add_10, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -605,6 +477,7 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 1x21x768xf32) add_11 = paddle._C_ops.add(layer_norm_3, dropout_6) + del dropout_6, layer_norm_3 # pd_op.layer_norm: (1x21x768xf32, 1x21xf32, 1x21xf32) <- (1x21x768xf32, 768xf32, 768xf32) layer_norm_6, layer_norm_7, layer_norm_8 = (lambda x, f: f(x))( @@ -613,7 +486,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_178, parameter_179 + del add_11, parameter_178, parameter_179 # pd_op.matmul: (1x21x768xf32) <- (1x21x768xf32, 768x768xf32) matmul_8 = paddle._C_ops.matmul(layer_norm_6, parameter_177, False, False) @@ -621,10 +494,11 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_12 = paddle._C_ops.add(matmul_8, parameter_176) - del parameter_176 + del matmul_8, parameter_176 # pd_op.reshape: (1x21x12x64xf32) <- (1x21x768xf32, 4xi64) reshape_4 = paddle._C_ops.reshape(add_12, full_int_array_1) + del add_12 # pd_op.transpose: (1x12x21x64xf32) <- (1x21x12x64xf32) transpose_4 = paddle._C_ops.transpose(reshape_4, [0, 2, 1, 3]) @@ -636,7 +510,7 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_13 = paddle._C_ops.add(matmul_9, parameter_174) - del parameter_174 + del matmul_9, parameter_174 # pd_op.matmul: (1x21x768xf32) <- (1x21x768xf32, 768x768xf32) matmul_10 = paddle._C_ops.matmul(layer_norm_6, parameter_173, False, False) @@ -644,10 +518,11 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_14 = paddle._C_ops.add(matmul_10, parameter_172) - del parameter_172 + del matmul_10, parameter_172 # pd_op.reshape: (1x21x12x64xf32) <- (1x21x768xf32, 4xi64) reshape_5 = paddle._C_ops.reshape(add_13, full_int_array_1) + del add_13 # pd_op.transpose: (1x12x21x64xf32) <- (1x21x12x64xf32) transpose_5 = paddle._C_ops.transpose(reshape_5, [0, 2, 1, 3]) @@ -655,6 +530,7 @@ def forward( # pd_op.reshape: (1x21x12x64xf32) <- (1x21x768xf32, 4xi64) reshape_6 = paddle._C_ops.reshape(add_14, full_int_array_1) + del add_14 # pd_op.transpose: (1x12x21x64xf32) <- (1x21x12x64xf32) transpose_6 = paddle._C_ops.transpose(reshape_6, [0, 2, 1, 3]) @@ -666,9 +542,11 @@ def forward( # pd_op.matmul: (1x12x21x21xf32) <- (1x12x21x64xf32, 1x12x21x64xf32) matmul_11 = paddle._C_ops.matmul(scale_3, transpose_5, False, True) + del scale_3, transpose_5 # pd_op.add: (1x12x21x21xf32) <- (1x12x21x21xf32, 1x1x1x21xf32) add_15 = paddle._C_ops.add(matmul_11, unsqueeze_0) + del matmul_11 # pd_op.softmax: (1x12x21x21xf32) <- (1x12x21x21xf32) softmax_1 = paddle._C_ops.softmax(add_15, -1) @@ -677,13 +555,15 @@ def forward( # pd_op.dropout: (1x12x21x21xf32, 1x12x21x21xui8) <- (1x12x21x21xf32, None, 1xf32) dropout_8, dropout_9 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_1, None, full_5, False, "upscale_in_train", 0, False + softmax_1, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_1 # pd_op.matmul: (1x12x21x64xf32) <- (1x12x21x21xf32, 1x12x21x64xf32) matmul_12 = paddle._C_ops.matmul(dropout_8, transpose_6, False, False) + del dropout_8, transpose_6 # pd_op.transpose: (1x21x12x64xf32) <- (1x12x21x64xf32) transpose_7 = paddle._C_ops.transpose(matmul_12, [0, 2, 1, 3]) @@ -691,19 +571,20 @@ def forward( # pd_op.reshape: (1x21x768xf32) <- (1x21x12x64xf32, 3xi64) reshape_7 = paddle._C_ops.reshape(transpose_7, full_int_array_2) + del transpose_7 # pd_op.matmul: (1x21x768xf32) <- (1x21x768xf32, 768x768xf32) matmul_13 = paddle._C_ops.matmul(reshape_7, parameter_171, False, False) - del parameter_171 + del parameter_171, reshape_7 # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_16 = paddle._C_ops.add(matmul_13, parameter_170) - del parameter_170 + del matmul_13, parameter_170 # pd_op.dropout: (1x21x768xf32, 1x21x768xui8) <- (1x21x768xf32, None, 1xf32) dropout_10, dropout_11 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_16, None, full_5, False, "upscale_in_train", 0, False + add_16, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -711,6 +592,7 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 1x21x768xf32) add_17 = paddle._C_ops.add(layer_norm_6, dropout_10) + del dropout_10, layer_norm_6 # pd_op.layer_norm: (1x21x768xf32, 1x21xf32, 1x21xf32) <- (1x21x768xf32, 768xf32, 768xf32) layer_norm_9, layer_norm_10, layer_norm_11 = (lambda x, f: f(x))( @@ -719,7 +601,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_164, parameter_165 + del add_17, parameter_164, parameter_165 # pd_op.matmul: (1x21x3072xf32) <- (1x21x768xf32, 768x3072xf32) matmul_14 = paddle._C_ops.matmul(layer_norm_9, parameter_169, False, False) @@ -727,23 +609,24 @@ def forward( # pd_op.add: (1x21x3072xf32) <- (1x21x3072xf32, 3072xf32) add_18 = paddle._C_ops.add(matmul_14, parameter_168) - del parameter_168 + del matmul_14, parameter_168 # pd_op.gelu: (1x21x3072xf32) <- (1x21x3072xf32) gelu_1 = paddle._C_ops.gelu(add_18, False) + del add_18 # pd_op.matmul: (1x21x768xf32) <- (1x21x3072xf32, 3072x768xf32) matmul_15 = paddle._C_ops.matmul(gelu_1, parameter_167, False, False) - del parameter_167 + del gelu_1, parameter_167 # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_19 = paddle._C_ops.add(matmul_15, parameter_166) - del parameter_166 + del matmul_15, parameter_166 # pd_op.dropout: (1x21x768xf32, 1x21x768xui8) <- (1x21x768xf32, None, 1xf32) dropout_12, dropout_13 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_19, None, full_5, False, "upscale_in_train", 0, False + add_19, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -751,6 +634,7 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 1x21x768xf32) add_20 = paddle._C_ops.add(layer_norm_9, dropout_12) + del dropout_12, layer_norm_9 # pd_op.layer_norm: (1x21x768xf32, 1x21xf32, 1x21xf32) <- (1x21x768xf32, 768xf32, 768xf32) layer_norm_12, layer_norm_13, layer_norm_14 = (lambda x, f: f(x))( @@ -759,7 +643,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_162, parameter_163 + del add_20, parameter_162, parameter_163 # pd_op.matmul: (1x21x768xf32) <- (1x21x768xf32, 768x768xf32) matmul_16 = paddle._C_ops.matmul(layer_norm_12, parameter_161, False, False) @@ -767,10 +651,11 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_21 = paddle._C_ops.add(matmul_16, parameter_160) - del parameter_160 + del matmul_16, parameter_160 # pd_op.reshape: (1x21x12x64xf32) <- (1x21x768xf32, 4xi64) reshape_8 = paddle._C_ops.reshape(add_21, full_int_array_1) + del add_21 # pd_op.transpose: (1x12x21x64xf32) <- (1x21x12x64xf32) transpose_8 = paddle._C_ops.transpose(reshape_8, [0, 2, 1, 3]) @@ -782,7 +667,7 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_22 = paddle._C_ops.add(matmul_17, parameter_158) - del parameter_158 + del matmul_17, parameter_158 # pd_op.matmul: (1x21x768xf32) <- (1x21x768xf32, 768x768xf32) matmul_18 = paddle._C_ops.matmul(layer_norm_12, parameter_157, False, False) @@ -790,10 +675,11 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_23 = paddle._C_ops.add(matmul_18, parameter_156) - del parameter_156 + del matmul_18, parameter_156 # pd_op.reshape: (1x21x12x64xf32) <- (1x21x768xf32, 4xi64) reshape_9 = paddle._C_ops.reshape(add_22, full_int_array_1) + del add_22 # pd_op.transpose: (1x12x21x64xf32) <- (1x21x12x64xf32) transpose_9 = paddle._C_ops.transpose(reshape_9, [0, 2, 1, 3]) @@ -801,6 +687,7 @@ def forward( # pd_op.reshape: (1x21x12x64xf32) <- (1x21x768xf32, 4xi64) reshape_10 = paddle._C_ops.reshape(add_23, full_int_array_1) + del add_23 # pd_op.transpose: (1x12x21x64xf32) <- (1x21x12x64xf32) transpose_10 = paddle._C_ops.transpose(reshape_10, [0, 2, 1, 3]) @@ -812,9 +699,11 @@ def forward( # pd_op.matmul: (1x12x21x21xf32) <- (1x12x21x64xf32, 1x12x21x64xf32) matmul_19 = paddle._C_ops.matmul(scale_4, transpose_9, False, True) + del scale_4, transpose_9 # pd_op.add: (1x12x21x21xf32) <- (1x12x21x21xf32, 1x1x1x21xf32) add_24 = paddle._C_ops.add(matmul_19, unsqueeze_0) + del matmul_19 # pd_op.softmax: (1x12x21x21xf32) <- (1x12x21x21xf32) softmax_2 = paddle._C_ops.softmax(add_24, -1) @@ -823,13 +712,15 @@ def forward( # pd_op.dropout: (1x12x21x21xf32, 1x12x21x21xui8) <- (1x12x21x21xf32, None, 1xf32) dropout_14, dropout_15 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_2, None, full_5, False, "upscale_in_train", 0, False + softmax_2, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_2 # pd_op.matmul: (1x12x21x64xf32) <- (1x12x21x21xf32, 1x12x21x64xf32) matmul_20 = paddle._C_ops.matmul(dropout_14, transpose_10, False, False) + del dropout_14, transpose_10 # pd_op.transpose: (1x21x12x64xf32) <- (1x12x21x64xf32) transpose_11 = paddle._C_ops.transpose(matmul_20, [0, 2, 1, 3]) @@ -837,19 +728,20 @@ def forward( # pd_op.reshape: (1x21x768xf32) <- (1x21x12x64xf32, 3xi64) reshape_11 = paddle._C_ops.reshape(transpose_11, full_int_array_2) + del transpose_11 # pd_op.matmul: (1x21x768xf32) <- (1x21x768xf32, 768x768xf32) matmul_21 = paddle._C_ops.matmul(reshape_11, parameter_155, False, False) - del parameter_155 + del parameter_155, reshape_11 # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_25 = paddle._C_ops.add(matmul_21, parameter_154) - del parameter_154 + del matmul_21, parameter_154 # pd_op.dropout: (1x21x768xf32, 1x21x768xui8) <- (1x21x768xf32, None, 1xf32) dropout_16, dropout_17 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_25, None, full_5, False, "upscale_in_train", 0, False + add_25, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -857,6 +749,7 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 1x21x768xf32) add_26 = paddle._C_ops.add(layer_norm_12, dropout_16) + del dropout_16, layer_norm_12 # pd_op.layer_norm: (1x21x768xf32, 1x21xf32, 1x21xf32) <- (1x21x768xf32, 768xf32, 768xf32) layer_norm_15, layer_norm_16, layer_norm_17 = (lambda x, f: f(x))( @@ -865,7 +758,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_148, parameter_149 + del add_26, parameter_148, parameter_149 # pd_op.matmul: (1x21x3072xf32) <- (1x21x768xf32, 768x3072xf32) matmul_22 = paddle._C_ops.matmul(layer_norm_15, parameter_153, False, False) @@ -873,23 +766,24 @@ def forward( # pd_op.add: (1x21x3072xf32) <- (1x21x3072xf32, 3072xf32) add_27 = paddle._C_ops.add(matmul_22, parameter_152) - del parameter_152 + del matmul_22, parameter_152 # pd_op.gelu: (1x21x3072xf32) <- (1x21x3072xf32) gelu_2 = paddle._C_ops.gelu(add_27, False) + del add_27 # pd_op.matmul: (1x21x768xf32) <- (1x21x3072xf32, 3072x768xf32) matmul_23 = paddle._C_ops.matmul(gelu_2, parameter_151, False, False) - del parameter_151 + del gelu_2, parameter_151 # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_28 = paddle._C_ops.add(matmul_23, parameter_150) - del parameter_150 + del matmul_23, parameter_150 # pd_op.dropout: (1x21x768xf32, 1x21x768xui8) <- (1x21x768xf32, None, 1xf32) dropout_18, dropout_19 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_28, None, full_5, False, "upscale_in_train", 0, False + add_28, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -897,6 +791,7 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 1x21x768xf32) add_29 = paddle._C_ops.add(layer_norm_15, dropout_18) + del dropout_18, layer_norm_15 # pd_op.layer_norm: (1x21x768xf32, 1x21xf32, 1x21xf32) <- (1x21x768xf32, 768xf32, 768xf32) layer_norm_18, layer_norm_19, layer_norm_20 = (lambda x, f: f(x))( @@ -905,7 +800,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_146, parameter_147 + del add_29, parameter_146, parameter_147 # pd_op.matmul: (1x21x768xf32) <- (1x21x768xf32, 768x768xf32) matmul_24 = paddle._C_ops.matmul(layer_norm_18, parameter_145, False, False) @@ -913,10 +808,11 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_30 = paddle._C_ops.add(matmul_24, parameter_144) - del parameter_144 + del matmul_24, parameter_144 # pd_op.reshape: (1x21x12x64xf32) <- (1x21x768xf32, 4xi64) reshape_12 = paddle._C_ops.reshape(add_30, full_int_array_1) + del add_30 # pd_op.transpose: (1x12x21x64xf32) <- (1x21x12x64xf32) transpose_12 = paddle._C_ops.transpose(reshape_12, [0, 2, 1, 3]) @@ -928,7 +824,7 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_31 = paddle._C_ops.add(matmul_25, parameter_142) - del parameter_142 + del matmul_25, parameter_142 # pd_op.matmul: (1x21x768xf32) <- (1x21x768xf32, 768x768xf32) matmul_26 = paddle._C_ops.matmul(layer_norm_18, parameter_141, False, False) @@ -936,10 +832,11 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_32 = paddle._C_ops.add(matmul_26, parameter_140) - del parameter_140 + del matmul_26, parameter_140 # pd_op.reshape: (1x21x12x64xf32) <- (1x21x768xf32, 4xi64) reshape_13 = paddle._C_ops.reshape(add_31, full_int_array_1) + del add_31 # pd_op.transpose: (1x12x21x64xf32) <- (1x21x12x64xf32) transpose_13 = paddle._C_ops.transpose(reshape_13, [0, 2, 1, 3]) @@ -947,6 +844,7 @@ def forward( # pd_op.reshape: (1x21x12x64xf32) <- (1x21x768xf32, 4xi64) reshape_14 = paddle._C_ops.reshape(add_32, full_int_array_1) + del add_32 # pd_op.transpose: (1x12x21x64xf32) <- (1x21x12x64xf32) transpose_14 = paddle._C_ops.transpose(reshape_14, [0, 2, 1, 3]) @@ -958,9 +856,11 @@ def forward( # pd_op.matmul: (1x12x21x21xf32) <- (1x12x21x64xf32, 1x12x21x64xf32) matmul_27 = paddle._C_ops.matmul(scale_5, transpose_13, False, True) + del scale_5, transpose_13 # pd_op.add: (1x12x21x21xf32) <- (1x12x21x21xf32, 1x1x1x21xf32) add_33 = paddle._C_ops.add(matmul_27, unsqueeze_0) + del matmul_27 # pd_op.softmax: (1x12x21x21xf32) <- (1x12x21x21xf32) softmax_3 = paddle._C_ops.softmax(add_33, -1) @@ -969,13 +869,15 @@ def forward( # pd_op.dropout: (1x12x21x21xf32, 1x12x21x21xui8) <- (1x12x21x21xf32, None, 1xf32) dropout_20, dropout_21 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_3, None, full_5, False, "upscale_in_train", 0, False + softmax_3, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_3 # pd_op.matmul: (1x12x21x64xf32) <- (1x12x21x21xf32, 1x12x21x64xf32) matmul_28 = paddle._C_ops.matmul(dropout_20, transpose_14, False, False) + del dropout_20, transpose_14 # pd_op.transpose: (1x21x12x64xf32) <- (1x12x21x64xf32) transpose_15 = paddle._C_ops.transpose(matmul_28, [0, 2, 1, 3]) @@ -983,19 +885,20 @@ def forward( # pd_op.reshape: (1x21x768xf32) <- (1x21x12x64xf32, 3xi64) reshape_15 = paddle._C_ops.reshape(transpose_15, full_int_array_2) + del transpose_15 # pd_op.matmul: (1x21x768xf32) <- (1x21x768xf32, 768x768xf32) matmul_29 = paddle._C_ops.matmul(reshape_15, parameter_139, False, False) - del parameter_139 + del parameter_139, reshape_15 # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_34 = paddle._C_ops.add(matmul_29, parameter_138) - del parameter_138 + del matmul_29, parameter_138 # pd_op.dropout: (1x21x768xf32, 1x21x768xui8) <- (1x21x768xf32, None, 1xf32) dropout_22, dropout_23 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_34, None, full_5, False, "upscale_in_train", 0, False + add_34, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1003,6 +906,7 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 1x21x768xf32) add_35 = paddle._C_ops.add(layer_norm_18, dropout_22) + del dropout_22, layer_norm_18 # pd_op.layer_norm: (1x21x768xf32, 1x21xf32, 1x21xf32) <- (1x21x768xf32, 768xf32, 768xf32) layer_norm_21, layer_norm_22, layer_norm_23 = (lambda x, f: f(x))( @@ -1011,7 +915,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_132, parameter_133 + del add_35, parameter_132, parameter_133 # pd_op.matmul: (1x21x3072xf32) <- (1x21x768xf32, 768x3072xf32) matmul_30 = paddle._C_ops.matmul(layer_norm_21, parameter_137, False, False) @@ -1019,23 +923,24 @@ def forward( # pd_op.add: (1x21x3072xf32) <- (1x21x3072xf32, 3072xf32) add_36 = paddle._C_ops.add(matmul_30, parameter_136) - del parameter_136 + del matmul_30, parameter_136 # pd_op.gelu: (1x21x3072xf32) <- (1x21x3072xf32) gelu_3 = paddle._C_ops.gelu(add_36, False) + del add_36 # pd_op.matmul: (1x21x768xf32) <- (1x21x3072xf32, 3072x768xf32) matmul_31 = paddle._C_ops.matmul(gelu_3, parameter_135, False, False) - del parameter_135 + del gelu_3, parameter_135 # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_37 = paddle._C_ops.add(matmul_31, parameter_134) - del parameter_134 + del matmul_31, parameter_134 # pd_op.dropout: (1x21x768xf32, 1x21x768xui8) <- (1x21x768xf32, None, 1xf32) dropout_24, dropout_25 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_37, None, full_5, False, "upscale_in_train", 0, False + add_37, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1043,6 +948,7 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 1x21x768xf32) add_38 = paddle._C_ops.add(layer_norm_21, dropout_24) + del dropout_24, layer_norm_21 # pd_op.layer_norm: (1x21x768xf32, 1x21xf32, 1x21xf32) <- (1x21x768xf32, 768xf32, 768xf32) layer_norm_24, layer_norm_25, layer_norm_26 = (lambda x, f: f(x))( @@ -1051,7 +957,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_130, parameter_131 + del add_38, parameter_130, parameter_131 # pd_op.matmul: (1x21x768xf32) <- (1x21x768xf32, 768x768xf32) matmul_32 = paddle._C_ops.matmul(layer_norm_24, parameter_129, False, False) @@ -1059,10 +965,11 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_39 = paddle._C_ops.add(matmul_32, parameter_128) - del parameter_128 + del matmul_32, parameter_128 # pd_op.reshape: (1x21x12x64xf32) <- (1x21x768xf32, 4xi64) reshape_16 = paddle._C_ops.reshape(add_39, full_int_array_1) + del add_39 # pd_op.transpose: (1x12x21x64xf32) <- (1x21x12x64xf32) transpose_16 = paddle._C_ops.transpose(reshape_16, [0, 2, 1, 3]) @@ -1074,7 +981,7 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_40 = paddle._C_ops.add(matmul_33, parameter_126) - del parameter_126 + del matmul_33, parameter_126 # pd_op.matmul: (1x21x768xf32) <- (1x21x768xf32, 768x768xf32) matmul_34 = paddle._C_ops.matmul(layer_norm_24, parameter_125, False, False) @@ -1082,10 +989,11 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_41 = paddle._C_ops.add(matmul_34, parameter_124) - del parameter_124 + del matmul_34, parameter_124 # pd_op.reshape: (1x21x12x64xf32) <- (1x21x768xf32, 4xi64) reshape_17 = paddle._C_ops.reshape(add_40, full_int_array_1) + del add_40 # pd_op.transpose: (1x12x21x64xf32) <- (1x21x12x64xf32) transpose_17 = paddle._C_ops.transpose(reshape_17, [0, 2, 1, 3]) @@ -1093,6 +1001,7 @@ def forward( # pd_op.reshape: (1x21x12x64xf32) <- (1x21x768xf32, 4xi64) reshape_18 = paddle._C_ops.reshape(add_41, full_int_array_1) + del add_41 # pd_op.transpose: (1x12x21x64xf32) <- (1x21x12x64xf32) transpose_18 = paddle._C_ops.transpose(reshape_18, [0, 2, 1, 3]) @@ -1104,9 +1013,11 @@ def forward( # pd_op.matmul: (1x12x21x21xf32) <- (1x12x21x64xf32, 1x12x21x64xf32) matmul_35 = paddle._C_ops.matmul(scale_6, transpose_17, False, True) + del scale_6, transpose_17 # pd_op.add: (1x12x21x21xf32) <- (1x12x21x21xf32, 1x1x1x21xf32) add_42 = paddle._C_ops.add(matmul_35, unsqueeze_0) + del matmul_35 # pd_op.softmax: (1x12x21x21xf32) <- (1x12x21x21xf32) softmax_4 = paddle._C_ops.softmax(add_42, -1) @@ -1115,13 +1026,15 @@ def forward( # pd_op.dropout: (1x12x21x21xf32, 1x12x21x21xui8) <- (1x12x21x21xf32, None, 1xf32) dropout_26, dropout_27 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_4, None, full_5, False, "upscale_in_train", 0, False + softmax_4, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_4 # pd_op.matmul: (1x12x21x64xf32) <- (1x12x21x21xf32, 1x12x21x64xf32) matmul_36 = paddle._C_ops.matmul(dropout_26, transpose_18, False, False) + del dropout_26, transpose_18 # pd_op.transpose: (1x21x12x64xf32) <- (1x12x21x64xf32) transpose_19 = paddle._C_ops.transpose(matmul_36, [0, 2, 1, 3]) @@ -1129,19 +1042,20 @@ def forward( # pd_op.reshape: (1x21x768xf32) <- (1x21x12x64xf32, 3xi64) reshape_19 = paddle._C_ops.reshape(transpose_19, full_int_array_2) + del transpose_19 # pd_op.matmul: (1x21x768xf32) <- (1x21x768xf32, 768x768xf32) matmul_37 = paddle._C_ops.matmul(reshape_19, parameter_123, False, False) - del parameter_123 + del parameter_123, reshape_19 # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_43 = paddle._C_ops.add(matmul_37, parameter_122) - del parameter_122 + del matmul_37, parameter_122 # pd_op.dropout: (1x21x768xf32, 1x21x768xui8) <- (1x21x768xf32, None, 1xf32) dropout_28, dropout_29 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_43, None, full_5, False, "upscale_in_train", 0, False + add_43, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1149,6 +1063,7 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 1x21x768xf32) add_44 = paddle._C_ops.add(layer_norm_24, dropout_28) + del dropout_28, layer_norm_24 # pd_op.layer_norm: (1x21x768xf32, 1x21xf32, 1x21xf32) <- (1x21x768xf32, 768xf32, 768xf32) layer_norm_27, layer_norm_28, layer_norm_29 = (lambda x, f: f(x))( @@ -1157,7 +1072,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_116, parameter_117 + del add_44, parameter_116, parameter_117 # pd_op.matmul: (1x21x3072xf32) <- (1x21x768xf32, 768x3072xf32) matmul_38 = paddle._C_ops.matmul(layer_norm_27, parameter_121, False, False) @@ -1165,23 +1080,24 @@ def forward( # pd_op.add: (1x21x3072xf32) <- (1x21x3072xf32, 3072xf32) add_45 = paddle._C_ops.add(matmul_38, parameter_120) - del parameter_120 + del matmul_38, parameter_120 # pd_op.gelu: (1x21x3072xf32) <- (1x21x3072xf32) gelu_4 = paddle._C_ops.gelu(add_45, False) + del add_45 # pd_op.matmul: (1x21x768xf32) <- (1x21x3072xf32, 3072x768xf32) matmul_39 = paddle._C_ops.matmul(gelu_4, parameter_119, False, False) - del parameter_119 + del gelu_4, parameter_119 # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_46 = paddle._C_ops.add(matmul_39, parameter_118) - del parameter_118 + del matmul_39, parameter_118 # pd_op.dropout: (1x21x768xf32, 1x21x768xui8) <- (1x21x768xf32, None, 1xf32) dropout_30, dropout_31 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_46, None, full_5, False, "upscale_in_train", 0, False + add_46, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1189,6 +1105,7 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 1x21x768xf32) add_47 = paddle._C_ops.add(layer_norm_27, dropout_30) + del dropout_30, layer_norm_27 # pd_op.layer_norm: (1x21x768xf32, 1x21xf32, 1x21xf32) <- (1x21x768xf32, 768xf32, 768xf32) layer_norm_30, layer_norm_31, layer_norm_32 = (lambda x, f: f(x))( @@ -1197,7 +1114,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_114, parameter_115 + del add_47, parameter_114, parameter_115 # pd_op.matmul: (1x21x768xf32) <- (1x21x768xf32, 768x768xf32) matmul_40 = paddle._C_ops.matmul(layer_norm_30, parameter_113, False, False) @@ -1205,10 +1122,11 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_48 = paddle._C_ops.add(matmul_40, parameter_112) - del parameter_112 + del matmul_40, parameter_112 # pd_op.reshape: (1x21x12x64xf32) <- (1x21x768xf32, 4xi64) reshape_20 = paddle._C_ops.reshape(add_48, full_int_array_1) + del add_48 # pd_op.transpose: (1x12x21x64xf32) <- (1x21x12x64xf32) transpose_20 = paddle._C_ops.transpose(reshape_20, [0, 2, 1, 3]) @@ -1220,7 +1138,7 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_49 = paddle._C_ops.add(matmul_41, parameter_110) - del parameter_110 + del matmul_41, parameter_110 # pd_op.matmul: (1x21x768xf32) <- (1x21x768xf32, 768x768xf32) matmul_42 = paddle._C_ops.matmul(layer_norm_30, parameter_109, False, False) @@ -1228,10 +1146,11 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_50 = paddle._C_ops.add(matmul_42, parameter_108) - del parameter_108 + del matmul_42, parameter_108 # pd_op.reshape: (1x21x12x64xf32) <- (1x21x768xf32, 4xi64) reshape_21 = paddle._C_ops.reshape(add_49, full_int_array_1) + del add_49 # pd_op.transpose: (1x12x21x64xf32) <- (1x21x12x64xf32) transpose_21 = paddle._C_ops.transpose(reshape_21, [0, 2, 1, 3]) @@ -1239,6 +1158,7 @@ def forward( # pd_op.reshape: (1x21x12x64xf32) <- (1x21x768xf32, 4xi64) reshape_22 = paddle._C_ops.reshape(add_50, full_int_array_1) + del add_50 # pd_op.transpose: (1x12x21x64xf32) <- (1x21x12x64xf32) transpose_22 = paddle._C_ops.transpose(reshape_22, [0, 2, 1, 3]) @@ -1250,9 +1170,11 @@ def forward( # pd_op.matmul: (1x12x21x21xf32) <- (1x12x21x64xf32, 1x12x21x64xf32) matmul_43 = paddle._C_ops.matmul(scale_7, transpose_21, False, True) + del scale_7, transpose_21 # pd_op.add: (1x12x21x21xf32) <- (1x12x21x21xf32, 1x1x1x21xf32) add_51 = paddle._C_ops.add(matmul_43, unsqueeze_0) + del matmul_43 # pd_op.softmax: (1x12x21x21xf32) <- (1x12x21x21xf32) softmax_5 = paddle._C_ops.softmax(add_51, -1) @@ -1261,13 +1183,15 @@ def forward( # pd_op.dropout: (1x12x21x21xf32, 1x12x21x21xui8) <- (1x12x21x21xf32, None, 1xf32) dropout_32, dropout_33 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_5, None, full_5, False, "upscale_in_train", 0, False + softmax_5, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_5 # pd_op.matmul: (1x12x21x64xf32) <- (1x12x21x21xf32, 1x12x21x64xf32) matmul_44 = paddle._C_ops.matmul(dropout_32, transpose_22, False, False) + del dropout_32, transpose_22 # pd_op.transpose: (1x21x12x64xf32) <- (1x12x21x64xf32) transpose_23 = paddle._C_ops.transpose(matmul_44, [0, 2, 1, 3]) @@ -1275,19 +1199,20 @@ def forward( # pd_op.reshape: (1x21x768xf32) <- (1x21x12x64xf32, 3xi64) reshape_23 = paddle._C_ops.reshape(transpose_23, full_int_array_2) + del transpose_23 # pd_op.matmul: (1x21x768xf32) <- (1x21x768xf32, 768x768xf32) matmul_45 = paddle._C_ops.matmul(reshape_23, parameter_107, False, False) - del parameter_107 + del parameter_107, reshape_23 # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_52 = paddle._C_ops.add(matmul_45, parameter_106) - del parameter_106 + del matmul_45, parameter_106 # pd_op.dropout: (1x21x768xf32, 1x21x768xui8) <- (1x21x768xf32, None, 1xf32) dropout_34, dropout_35 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_52, None, full_5, False, "upscale_in_train", 0, False + add_52, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1295,6 +1220,7 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 1x21x768xf32) add_53 = paddle._C_ops.add(layer_norm_30, dropout_34) + del dropout_34, layer_norm_30 # pd_op.layer_norm: (1x21x768xf32, 1x21xf32, 1x21xf32) <- (1x21x768xf32, 768xf32, 768xf32) layer_norm_33, layer_norm_34, layer_norm_35 = (lambda x, f: f(x))( @@ -1303,7 +1229,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_100, parameter_101 + del add_53, parameter_100, parameter_101 # pd_op.matmul: (1x21x3072xf32) <- (1x21x768xf32, 768x3072xf32) matmul_46 = paddle._C_ops.matmul(layer_norm_33, parameter_105, False, False) @@ -1311,23 +1237,24 @@ def forward( # pd_op.add: (1x21x3072xf32) <- (1x21x3072xf32, 3072xf32) add_54 = paddle._C_ops.add(matmul_46, parameter_104) - del parameter_104 + del matmul_46, parameter_104 # pd_op.gelu: (1x21x3072xf32) <- (1x21x3072xf32) gelu_5 = paddle._C_ops.gelu(add_54, False) + del add_54 # pd_op.matmul: (1x21x768xf32) <- (1x21x3072xf32, 3072x768xf32) matmul_47 = paddle._C_ops.matmul(gelu_5, parameter_103, False, False) - del parameter_103 + del gelu_5, parameter_103 # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_55 = paddle._C_ops.add(matmul_47, parameter_102) - del parameter_102 + del matmul_47, parameter_102 # pd_op.dropout: (1x21x768xf32, 1x21x768xui8) <- (1x21x768xf32, None, 1xf32) dropout_36, dropout_37 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_55, None, full_5, False, "upscale_in_train", 0, False + add_55, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1335,6 +1262,7 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 1x21x768xf32) add_56 = paddle._C_ops.add(layer_norm_33, dropout_36) + del dropout_36, layer_norm_33 # pd_op.layer_norm: (1x21x768xf32, 1x21xf32, 1x21xf32) <- (1x21x768xf32, 768xf32, 768xf32) layer_norm_36, layer_norm_37, layer_norm_38 = (lambda x, f: f(x))( @@ -1343,7 +1271,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_98, parameter_99 + del add_56, parameter_98, parameter_99 # pd_op.matmul: (1x21x768xf32) <- (1x21x768xf32, 768x768xf32) matmul_48 = paddle._C_ops.matmul(layer_norm_36, parameter_97, False, False) @@ -1351,10 +1279,11 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_57 = paddle._C_ops.add(matmul_48, parameter_96) - del parameter_96 + del matmul_48, parameter_96 # pd_op.reshape: (1x21x12x64xf32) <- (1x21x768xf32, 4xi64) reshape_24 = paddle._C_ops.reshape(add_57, full_int_array_1) + del add_57 # pd_op.transpose: (1x12x21x64xf32) <- (1x21x12x64xf32) transpose_24 = paddle._C_ops.transpose(reshape_24, [0, 2, 1, 3]) @@ -1366,7 +1295,7 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_58 = paddle._C_ops.add(matmul_49, parameter_94) - del parameter_94 + del matmul_49, parameter_94 # pd_op.matmul: (1x21x768xf32) <- (1x21x768xf32, 768x768xf32) matmul_50 = paddle._C_ops.matmul(layer_norm_36, parameter_93, False, False) @@ -1374,10 +1303,11 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_59 = paddle._C_ops.add(matmul_50, parameter_92) - del parameter_92 + del matmul_50, parameter_92 # pd_op.reshape: (1x21x12x64xf32) <- (1x21x768xf32, 4xi64) reshape_25 = paddle._C_ops.reshape(add_58, full_int_array_1) + del add_58 # pd_op.transpose: (1x12x21x64xf32) <- (1x21x12x64xf32) transpose_25 = paddle._C_ops.transpose(reshape_25, [0, 2, 1, 3]) @@ -1385,6 +1315,7 @@ def forward( # pd_op.reshape: (1x21x12x64xf32) <- (1x21x768xf32, 4xi64) reshape_26 = paddle._C_ops.reshape(add_59, full_int_array_1) + del add_59 # pd_op.transpose: (1x12x21x64xf32) <- (1x21x12x64xf32) transpose_26 = paddle._C_ops.transpose(reshape_26, [0, 2, 1, 3]) @@ -1396,9 +1327,11 @@ def forward( # pd_op.matmul: (1x12x21x21xf32) <- (1x12x21x64xf32, 1x12x21x64xf32) matmul_51 = paddle._C_ops.matmul(scale_8, transpose_25, False, True) + del scale_8, transpose_25 # pd_op.add: (1x12x21x21xf32) <- (1x12x21x21xf32, 1x1x1x21xf32) add_60 = paddle._C_ops.add(matmul_51, unsqueeze_0) + del matmul_51 # pd_op.softmax: (1x12x21x21xf32) <- (1x12x21x21xf32) softmax_6 = paddle._C_ops.softmax(add_60, -1) @@ -1407,13 +1340,15 @@ def forward( # pd_op.dropout: (1x12x21x21xf32, 1x12x21x21xui8) <- (1x12x21x21xf32, None, 1xf32) dropout_38, dropout_39 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_6, None, full_5, False, "upscale_in_train", 0, False + softmax_6, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_6 # pd_op.matmul: (1x12x21x64xf32) <- (1x12x21x21xf32, 1x12x21x64xf32) matmul_52 = paddle._C_ops.matmul(dropout_38, transpose_26, False, False) + del dropout_38, transpose_26 # pd_op.transpose: (1x21x12x64xf32) <- (1x12x21x64xf32) transpose_27 = paddle._C_ops.transpose(matmul_52, [0, 2, 1, 3]) @@ -1421,19 +1356,20 @@ def forward( # pd_op.reshape: (1x21x768xf32) <- (1x21x12x64xf32, 3xi64) reshape_27 = paddle._C_ops.reshape(transpose_27, full_int_array_2) + del transpose_27 # pd_op.matmul: (1x21x768xf32) <- (1x21x768xf32, 768x768xf32) matmul_53 = paddle._C_ops.matmul(reshape_27, parameter_91, False, False) - del parameter_91 + del parameter_91, reshape_27 # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_61 = paddle._C_ops.add(matmul_53, parameter_90) - del parameter_90 + del matmul_53, parameter_90 # pd_op.dropout: (1x21x768xf32, 1x21x768xui8) <- (1x21x768xf32, None, 1xf32) dropout_40, dropout_41 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_61, None, full_5, False, "upscale_in_train", 0, False + add_61, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1441,6 +1377,7 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 1x21x768xf32) add_62 = paddle._C_ops.add(layer_norm_36, dropout_40) + del dropout_40, layer_norm_36 # pd_op.layer_norm: (1x21x768xf32, 1x21xf32, 1x21xf32) <- (1x21x768xf32, 768xf32, 768xf32) layer_norm_39, layer_norm_40, layer_norm_41 = (lambda x, f: f(x))( @@ -1449,7 +1386,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_84, parameter_85 + del add_62, parameter_84, parameter_85 # pd_op.matmul: (1x21x3072xf32) <- (1x21x768xf32, 768x3072xf32) matmul_54 = paddle._C_ops.matmul(layer_norm_39, parameter_89, False, False) @@ -1457,23 +1394,24 @@ def forward( # pd_op.add: (1x21x3072xf32) <- (1x21x3072xf32, 3072xf32) add_63 = paddle._C_ops.add(matmul_54, parameter_88) - del parameter_88 + del matmul_54, parameter_88 # pd_op.gelu: (1x21x3072xf32) <- (1x21x3072xf32) gelu_6 = paddle._C_ops.gelu(add_63, False) + del add_63 # pd_op.matmul: (1x21x768xf32) <- (1x21x3072xf32, 3072x768xf32) matmul_55 = paddle._C_ops.matmul(gelu_6, parameter_87, False, False) - del parameter_87 + del gelu_6, parameter_87 # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_64 = paddle._C_ops.add(matmul_55, parameter_86) - del parameter_86 + del matmul_55, parameter_86 # pd_op.dropout: (1x21x768xf32, 1x21x768xui8) <- (1x21x768xf32, None, 1xf32) dropout_42, dropout_43 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_64, None, full_5, False, "upscale_in_train", 0, False + add_64, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1481,6 +1419,7 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 1x21x768xf32) add_65 = paddle._C_ops.add(layer_norm_39, dropout_42) + del dropout_42, layer_norm_39 # pd_op.layer_norm: (1x21x768xf32, 1x21xf32, 1x21xf32) <- (1x21x768xf32, 768xf32, 768xf32) layer_norm_42, layer_norm_43, layer_norm_44 = (lambda x, f: f(x))( @@ -1489,7 +1428,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_82, parameter_83 + del add_65, parameter_82, parameter_83 # pd_op.matmul: (1x21x768xf32) <- (1x21x768xf32, 768x768xf32) matmul_56 = paddle._C_ops.matmul(layer_norm_42, parameter_81, False, False) @@ -1497,10 +1436,11 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_66 = paddle._C_ops.add(matmul_56, parameter_80) - del parameter_80 + del matmul_56, parameter_80 # pd_op.reshape: (1x21x12x64xf32) <- (1x21x768xf32, 4xi64) reshape_28 = paddle._C_ops.reshape(add_66, full_int_array_1) + del add_66 # pd_op.transpose: (1x12x21x64xf32) <- (1x21x12x64xf32) transpose_28 = paddle._C_ops.transpose(reshape_28, [0, 2, 1, 3]) @@ -1512,7 +1452,7 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_67 = paddle._C_ops.add(matmul_57, parameter_78) - del parameter_78 + del matmul_57, parameter_78 # pd_op.matmul: (1x21x768xf32) <- (1x21x768xf32, 768x768xf32) matmul_58 = paddle._C_ops.matmul(layer_norm_42, parameter_77, False, False) @@ -1520,10 +1460,11 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_68 = paddle._C_ops.add(matmul_58, parameter_76) - del parameter_76 + del matmul_58, parameter_76 # pd_op.reshape: (1x21x12x64xf32) <- (1x21x768xf32, 4xi64) reshape_29 = paddle._C_ops.reshape(add_67, full_int_array_1) + del add_67 # pd_op.transpose: (1x12x21x64xf32) <- (1x21x12x64xf32) transpose_29 = paddle._C_ops.transpose(reshape_29, [0, 2, 1, 3]) @@ -1531,6 +1472,7 @@ def forward( # pd_op.reshape: (1x21x12x64xf32) <- (1x21x768xf32, 4xi64) reshape_30 = paddle._C_ops.reshape(add_68, full_int_array_1) + del add_68 # pd_op.transpose: (1x12x21x64xf32) <- (1x21x12x64xf32) transpose_30 = paddle._C_ops.transpose(reshape_30, [0, 2, 1, 3]) @@ -1542,9 +1484,11 @@ def forward( # pd_op.matmul: (1x12x21x21xf32) <- (1x12x21x64xf32, 1x12x21x64xf32) matmul_59 = paddle._C_ops.matmul(scale_9, transpose_29, False, True) + del scale_9, transpose_29 # pd_op.add: (1x12x21x21xf32) <- (1x12x21x21xf32, 1x1x1x21xf32) add_69 = paddle._C_ops.add(matmul_59, unsqueeze_0) + del matmul_59 # pd_op.softmax: (1x12x21x21xf32) <- (1x12x21x21xf32) softmax_7 = paddle._C_ops.softmax(add_69, -1) @@ -1553,13 +1497,15 @@ def forward( # pd_op.dropout: (1x12x21x21xf32, 1x12x21x21xui8) <- (1x12x21x21xf32, None, 1xf32) dropout_44, dropout_45 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_7, None, full_5, False, "upscale_in_train", 0, False + softmax_7, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_7 # pd_op.matmul: (1x12x21x64xf32) <- (1x12x21x21xf32, 1x12x21x64xf32) matmul_60 = paddle._C_ops.matmul(dropout_44, transpose_30, False, False) + del dropout_44, transpose_30 # pd_op.transpose: (1x21x12x64xf32) <- (1x12x21x64xf32) transpose_31 = paddle._C_ops.transpose(matmul_60, [0, 2, 1, 3]) @@ -1567,19 +1513,20 @@ def forward( # pd_op.reshape: (1x21x768xf32) <- (1x21x12x64xf32, 3xi64) reshape_31 = paddle._C_ops.reshape(transpose_31, full_int_array_2) + del transpose_31 # pd_op.matmul: (1x21x768xf32) <- (1x21x768xf32, 768x768xf32) matmul_61 = paddle._C_ops.matmul(reshape_31, parameter_75, False, False) - del parameter_75 + del parameter_75, reshape_31 # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_70 = paddle._C_ops.add(matmul_61, parameter_74) - del parameter_74 + del matmul_61, parameter_74 # pd_op.dropout: (1x21x768xf32, 1x21x768xui8) <- (1x21x768xf32, None, 1xf32) dropout_46, dropout_47 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_70, None, full_5, False, "upscale_in_train", 0, False + add_70, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1587,6 +1534,7 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 1x21x768xf32) add_71 = paddle._C_ops.add(layer_norm_42, dropout_46) + del dropout_46, layer_norm_42 # pd_op.layer_norm: (1x21x768xf32, 1x21xf32, 1x21xf32) <- (1x21x768xf32, 768xf32, 768xf32) layer_norm_45, layer_norm_46, layer_norm_47 = (lambda x, f: f(x))( @@ -1595,7 +1543,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_68, parameter_69 + del add_71, parameter_68, parameter_69 # pd_op.matmul: (1x21x3072xf32) <- (1x21x768xf32, 768x3072xf32) matmul_62 = paddle._C_ops.matmul(layer_norm_45, parameter_73, False, False) @@ -1603,23 +1551,24 @@ def forward( # pd_op.add: (1x21x3072xf32) <- (1x21x3072xf32, 3072xf32) add_72 = paddle._C_ops.add(matmul_62, parameter_72) - del parameter_72 + del matmul_62, parameter_72 # pd_op.gelu: (1x21x3072xf32) <- (1x21x3072xf32) gelu_7 = paddle._C_ops.gelu(add_72, False) + del add_72 # pd_op.matmul: (1x21x768xf32) <- (1x21x3072xf32, 3072x768xf32) matmul_63 = paddle._C_ops.matmul(gelu_7, parameter_71, False, False) - del parameter_71 + del gelu_7, parameter_71 # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_73 = paddle._C_ops.add(matmul_63, parameter_70) - del parameter_70 + del matmul_63, parameter_70 # pd_op.dropout: (1x21x768xf32, 1x21x768xui8) <- (1x21x768xf32, None, 1xf32) dropout_48, dropout_49 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_73, None, full_5, False, "upscale_in_train", 0, False + add_73, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1627,6 +1576,7 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 1x21x768xf32) add_74 = paddle._C_ops.add(layer_norm_45, dropout_48) + del dropout_48, layer_norm_45 # pd_op.layer_norm: (1x21x768xf32, 1x21xf32, 1x21xf32) <- (1x21x768xf32, 768xf32, 768xf32) layer_norm_48, layer_norm_49, layer_norm_50 = (lambda x, f: f(x))( @@ -1635,7 +1585,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_66, parameter_67 + del add_74, parameter_66, parameter_67 # pd_op.matmul: (1x21x768xf32) <- (1x21x768xf32, 768x768xf32) matmul_64 = paddle._C_ops.matmul(layer_norm_48, parameter_65, False, False) @@ -1643,10 +1593,11 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_75 = paddle._C_ops.add(matmul_64, parameter_64) - del parameter_64 + del matmul_64, parameter_64 # pd_op.reshape: (1x21x12x64xf32) <- (1x21x768xf32, 4xi64) reshape_32 = paddle._C_ops.reshape(add_75, full_int_array_1) + del add_75 # pd_op.transpose: (1x12x21x64xf32) <- (1x21x12x64xf32) transpose_32 = paddle._C_ops.transpose(reshape_32, [0, 2, 1, 3]) @@ -1658,7 +1609,7 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_76 = paddle._C_ops.add(matmul_65, parameter_62) - del parameter_62 + del matmul_65, parameter_62 # pd_op.matmul: (1x21x768xf32) <- (1x21x768xf32, 768x768xf32) matmul_66 = paddle._C_ops.matmul(layer_norm_48, parameter_61, False, False) @@ -1666,10 +1617,11 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_77 = paddle._C_ops.add(matmul_66, parameter_60) - del parameter_60 + del matmul_66, parameter_60 # pd_op.reshape: (1x21x12x64xf32) <- (1x21x768xf32, 4xi64) reshape_33 = paddle._C_ops.reshape(add_76, full_int_array_1) + del add_76 # pd_op.transpose: (1x12x21x64xf32) <- (1x21x12x64xf32) transpose_33 = paddle._C_ops.transpose(reshape_33, [0, 2, 1, 3]) @@ -1677,6 +1629,7 @@ def forward( # pd_op.reshape: (1x21x12x64xf32) <- (1x21x768xf32, 4xi64) reshape_34 = paddle._C_ops.reshape(add_77, full_int_array_1) + del add_77 # pd_op.transpose: (1x12x21x64xf32) <- (1x21x12x64xf32) transpose_34 = paddle._C_ops.transpose(reshape_34, [0, 2, 1, 3]) @@ -1688,9 +1641,11 @@ def forward( # pd_op.matmul: (1x12x21x21xf32) <- (1x12x21x64xf32, 1x12x21x64xf32) matmul_67 = paddle._C_ops.matmul(scale_10, transpose_33, False, True) + del scale_10, transpose_33 # pd_op.add: (1x12x21x21xf32) <- (1x12x21x21xf32, 1x1x1x21xf32) add_78 = paddle._C_ops.add(matmul_67, unsqueeze_0) + del matmul_67 # pd_op.softmax: (1x12x21x21xf32) <- (1x12x21x21xf32) softmax_8 = paddle._C_ops.softmax(add_78, -1) @@ -1699,13 +1654,15 @@ def forward( # pd_op.dropout: (1x12x21x21xf32, 1x12x21x21xui8) <- (1x12x21x21xf32, None, 1xf32) dropout_50, dropout_51 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_8, None, full_5, False, "upscale_in_train", 0, False + softmax_8, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_8 # pd_op.matmul: (1x12x21x64xf32) <- (1x12x21x21xf32, 1x12x21x64xf32) matmul_68 = paddle._C_ops.matmul(dropout_50, transpose_34, False, False) + del dropout_50, transpose_34 # pd_op.transpose: (1x21x12x64xf32) <- (1x12x21x64xf32) transpose_35 = paddle._C_ops.transpose(matmul_68, [0, 2, 1, 3]) @@ -1713,19 +1670,20 @@ def forward( # pd_op.reshape: (1x21x768xf32) <- (1x21x12x64xf32, 3xi64) reshape_35 = paddle._C_ops.reshape(transpose_35, full_int_array_2) + del transpose_35 # pd_op.matmul: (1x21x768xf32) <- (1x21x768xf32, 768x768xf32) matmul_69 = paddle._C_ops.matmul(reshape_35, parameter_59, False, False) - del parameter_59 + del parameter_59, reshape_35 # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_79 = paddle._C_ops.add(matmul_69, parameter_58) - del parameter_58 + del matmul_69, parameter_58 # pd_op.dropout: (1x21x768xf32, 1x21x768xui8) <- (1x21x768xf32, None, 1xf32) dropout_52, dropout_53 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_79, None, full_5, False, "upscale_in_train", 0, False + add_79, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1733,6 +1691,7 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 1x21x768xf32) add_80 = paddle._C_ops.add(layer_norm_48, dropout_52) + del dropout_52, layer_norm_48 # pd_op.layer_norm: (1x21x768xf32, 1x21xf32, 1x21xf32) <- (1x21x768xf32, 768xf32, 768xf32) layer_norm_51, layer_norm_52, layer_norm_53 = (lambda x, f: f(x))( @@ -1741,7 +1700,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_52, parameter_53 + del add_80, parameter_52, parameter_53 # pd_op.matmul: (1x21x3072xf32) <- (1x21x768xf32, 768x3072xf32) matmul_70 = paddle._C_ops.matmul(layer_norm_51, parameter_57, False, False) @@ -1749,23 +1708,24 @@ def forward( # pd_op.add: (1x21x3072xf32) <- (1x21x3072xf32, 3072xf32) add_81 = paddle._C_ops.add(matmul_70, parameter_56) - del parameter_56 + del matmul_70, parameter_56 # pd_op.gelu: (1x21x3072xf32) <- (1x21x3072xf32) gelu_8 = paddle._C_ops.gelu(add_81, False) + del add_81 # pd_op.matmul: (1x21x768xf32) <- (1x21x3072xf32, 3072x768xf32) matmul_71 = paddle._C_ops.matmul(gelu_8, parameter_55, False, False) - del parameter_55 + del gelu_8, parameter_55 # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_82 = paddle._C_ops.add(matmul_71, parameter_54) - del parameter_54 + del matmul_71, parameter_54 # pd_op.dropout: (1x21x768xf32, 1x21x768xui8) <- (1x21x768xf32, None, 1xf32) dropout_54, dropout_55 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_82, None, full_5, False, "upscale_in_train", 0, False + add_82, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1773,6 +1733,7 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 1x21x768xf32) add_83 = paddle._C_ops.add(layer_norm_51, dropout_54) + del dropout_54, layer_norm_51 # pd_op.layer_norm: (1x21x768xf32, 1x21xf32, 1x21xf32) <- (1x21x768xf32, 768xf32, 768xf32) layer_norm_54, layer_norm_55, layer_norm_56 = (lambda x, f: f(x))( @@ -1781,7 +1742,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_50, parameter_51 + del add_83, parameter_50, parameter_51 # pd_op.matmul: (1x21x768xf32) <- (1x21x768xf32, 768x768xf32) matmul_72 = paddle._C_ops.matmul(layer_norm_54, parameter_49, False, False) @@ -1789,10 +1750,11 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_84 = paddle._C_ops.add(matmul_72, parameter_48) - del parameter_48 + del matmul_72, parameter_48 # pd_op.reshape: (1x21x12x64xf32) <- (1x21x768xf32, 4xi64) reshape_36 = paddle._C_ops.reshape(add_84, full_int_array_1) + del add_84 # pd_op.transpose: (1x12x21x64xf32) <- (1x21x12x64xf32) transpose_36 = paddle._C_ops.transpose(reshape_36, [0, 2, 1, 3]) @@ -1804,7 +1766,7 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_85 = paddle._C_ops.add(matmul_73, parameter_46) - del parameter_46 + del matmul_73, parameter_46 # pd_op.matmul: (1x21x768xf32) <- (1x21x768xf32, 768x768xf32) matmul_74 = paddle._C_ops.matmul(layer_norm_54, parameter_45, False, False) @@ -1812,10 +1774,11 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_86 = paddle._C_ops.add(matmul_74, parameter_44) - del parameter_44 + del matmul_74, parameter_44 # pd_op.reshape: (1x21x12x64xf32) <- (1x21x768xf32, 4xi64) reshape_37 = paddle._C_ops.reshape(add_85, full_int_array_1) + del add_85 # pd_op.transpose: (1x12x21x64xf32) <- (1x21x12x64xf32) transpose_37 = paddle._C_ops.transpose(reshape_37, [0, 2, 1, 3]) @@ -1823,6 +1786,7 @@ def forward( # pd_op.reshape: (1x21x12x64xf32) <- (1x21x768xf32, 4xi64) reshape_38 = paddle._C_ops.reshape(add_86, full_int_array_1) + del add_86 # pd_op.transpose: (1x12x21x64xf32) <- (1x21x12x64xf32) transpose_38 = paddle._C_ops.transpose(reshape_38, [0, 2, 1, 3]) @@ -1834,9 +1798,11 @@ def forward( # pd_op.matmul: (1x12x21x21xf32) <- (1x12x21x64xf32, 1x12x21x64xf32) matmul_75 = paddle._C_ops.matmul(scale_11, transpose_37, False, True) + del scale_11, transpose_37 # pd_op.add: (1x12x21x21xf32) <- (1x12x21x21xf32, 1x1x1x21xf32) add_87 = paddle._C_ops.add(matmul_75, unsqueeze_0) + del matmul_75 # pd_op.softmax: (1x12x21x21xf32) <- (1x12x21x21xf32) softmax_9 = paddle._C_ops.softmax(add_87, -1) @@ -1845,13 +1811,15 @@ def forward( # pd_op.dropout: (1x12x21x21xf32, 1x12x21x21xui8) <- (1x12x21x21xf32, None, 1xf32) dropout_56, dropout_57 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_9, None, full_5, False, "upscale_in_train", 0, False + softmax_9, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_9 # pd_op.matmul: (1x12x21x64xf32) <- (1x12x21x21xf32, 1x12x21x64xf32) matmul_76 = paddle._C_ops.matmul(dropout_56, transpose_38, False, False) + del dropout_56, transpose_38 # pd_op.transpose: (1x21x12x64xf32) <- (1x12x21x64xf32) transpose_39 = paddle._C_ops.transpose(matmul_76, [0, 2, 1, 3]) @@ -1859,19 +1827,20 @@ def forward( # pd_op.reshape: (1x21x768xf32) <- (1x21x12x64xf32, 3xi64) reshape_39 = paddle._C_ops.reshape(transpose_39, full_int_array_2) + del transpose_39 # pd_op.matmul: (1x21x768xf32) <- (1x21x768xf32, 768x768xf32) matmul_77 = paddle._C_ops.matmul(reshape_39, parameter_43, False, False) - del parameter_43 + del parameter_43, reshape_39 # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_88 = paddle._C_ops.add(matmul_77, parameter_42) - del parameter_42 + del matmul_77, parameter_42 # pd_op.dropout: (1x21x768xf32, 1x21x768xui8) <- (1x21x768xf32, None, 1xf32) dropout_58, dropout_59 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_88, None, full_5, False, "upscale_in_train", 0, False + add_88, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1879,6 +1848,7 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 1x21x768xf32) add_89 = paddle._C_ops.add(layer_norm_54, dropout_58) + del dropout_58, layer_norm_54 # pd_op.layer_norm: (1x21x768xf32, 1x21xf32, 1x21xf32) <- (1x21x768xf32, 768xf32, 768xf32) layer_norm_57, layer_norm_58, layer_norm_59 = (lambda x, f: f(x))( @@ -1887,7 +1857,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_36, parameter_37 + del add_89, parameter_36, parameter_37 # pd_op.matmul: (1x21x3072xf32) <- (1x21x768xf32, 768x3072xf32) matmul_78 = paddle._C_ops.matmul(layer_norm_57, parameter_41, False, False) @@ -1895,23 +1865,24 @@ def forward( # pd_op.add: (1x21x3072xf32) <- (1x21x3072xf32, 3072xf32) add_90 = paddle._C_ops.add(matmul_78, parameter_40) - del parameter_40 + del matmul_78, parameter_40 # pd_op.gelu: (1x21x3072xf32) <- (1x21x3072xf32) gelu_9 = paddle._C_ops.gelu(add_90, False) + del add_90 # pd_op.matmul: (1x21x768xf32) <- (1x21x3072xf32, 3072x768xf32) matmul_79 = paddle._C_ops.matmul(gelu_9, parameter_39, False, False) - del parameter_39 + del gelu_9, parameter_39 # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_91 = paddle._C_ops.add(matmul_79, parameter_38) - del parameter_38 + del matmul_79, parameter_38 # pd_op.dropout: (1x21x768xf32, 1x21x768xui8) <- (1x21x768xf32, None, 1xf32) dropout_60, dropout_61 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_91, None, full_5, False, "upscale_in_train", 0, False + add_91, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1919,6 +1890,7 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 1x21x768xf32) add_92 = paddle._C_ops.add(layer_norm_57, dropout_60) + del dropout_60, layer_norm_57 # pd_op.layer_norm: (1x21x768xf32, 1x21xf32, 1x21xf32) <- (1x21x768xf32, 768xf32, 768xf32) layer_norm_60, layer_norm_61, layer_norm_62 = (lambda x, f: f(x))( @@ -1927,7 +1899,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_34, parameter_35 + del add_92, parameter_34, parameter_35 # pd_op.matmul: (1x21x768xf32) <- (1x21x768xf32, 768x768xf32) matmul_80 = paddle._C_ops.matmul(layer_norm_60, parameter_33, False, False) @@ -1935,10 +1907,11 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_93 = paddle._C_ops.add(matmul_80, parameter_32) - del parameter_32 + del matmul_80, parameter_32 # pd_op.reshape: (1x21x12x64xf32) <- (1x21x768xf32, 4xi64) reshape_40 = paddle._C_ops.reshape(add_93, full_int_array_1) + del add_93 # pd_op.transpose: (1x12x21x64xf32) <- (1x21x12x64xf32) transpose_40 = paddle._C_ops.transpose(reshape_40, [0, 2, 1, 3]) @@ -1950,7 +1923,7 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_94 = paddle._C_ops.add(matmul_81, parameter_30) - del parameter_30 + del matmul_81, parameter_30 # pd_op.matmul: (1x21x768xf32) <- (1x21x768xf32, 768x768xf32) matmul_82 = paddle._C_ops.matmul(layer_norm_60, parameter_29, False, False) @@ -1958,10 +1931,11 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_95 = paddle._C_ops.add(matmul_82, parameter_28) - del parameter_28 + del matmul_82, parameter_28 # pd_op.reshape: (1x21x12x64xf32) <- (1x21x768xf32, 4xi64) reshape_41 = paddle._C_ops.reshape(add_94, full_int_array_1) + del add_94 # pd_op.transpose: (1x12x21x64xf32) <- (1x21x12x64xf32) transpose_41 = paddle._C_ops.transpose(reshape_41, [0, 2, 1, 3]) @@ -1969,6 +1943,7 @@ def forward( # pd_op.reshape: (1x21x12x64xf32) <- (1x21x768xf32, 4xi64) reshape_42 = paddle._C_ops.reshape(add_95, full_int_array_1) + del add_95 # pd_op.transpose: (1x12x21x64xf32) <- (1x21x12x64xf32) transpose_42 = paddle._C_ops.transpose(reshape_42, [0, 2, 1, 3]) @@ -1980,9 +1955,11 @@ def forward( # pd_op.matmul: (1x12x21x21xf32) <- (1x12x21x64xf32, 1x12x21x64xf32) matmul_83 = paddle._C_ops.matmul(scale_12, transpose_41, False, True) + del scale_12, transpose_41 # pd_op.add: (1x12x21x21xf32) <- (1x12x21x21xf32, 1x1x1x21xf32) add_96 = paddle._C_ops.add(matmul_83, unsqueeze_0) + del matmul_83 # pd_op.softmax: (1x12x21x21xf32) <- (1x12x21x21xf32) softmax_10 = paddle._C_ops.softmax(add_96, -1) @@ -1991,13 +1968,15 @@ def forward( # pd_op.dropout: (1x12x21x21xf32, 1x12x21x21xui8) <- (1x12x21x21xf32, None, 1xf32) dropout_62, dropout_63 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_10, None, full_5, False, "upscale_in_train", 0, False + softmax_10, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_10 # pd_op.matmul: (1x12x21x64xf32) <- (1x12x21x21xf32, 1x12x21x64xf32) matmul_84 = paddle._C_ops.matmul(dropout_62, transpose_42, False, False) + del dropout_62, transpose_42 # pd_op.transpose: (1x21x12x64xf32) <- (1x12x21x64xf32) transpose_43 = paddle._C_ops.transpose(matmul_84, [0, 2, 1, 3]) @@ -2005,19 +1984,20 @@ def forward( # pd_op.reshape: (1x21x768xf32) <- (1x21x12x64xf32, 3xi64) reshape_43 = paddle._C_ops.reshape(transpose_43, full_int_array_2) + del transpose_43 # pd_op.matmul: (1x21x768xf32) <- (1x21x768xf32, 768x768xf32) matmul_85 = paddle._C_ops.matmul(reshape_43, parameter_27, False, False) - del parameter_27 + del parameter_27, reshape_43 # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_97 = paddle._C_ops.add(matmul_85, parameter_26) - del parameter_26 + del matmul_85, parameter_26 # pd_op.dropout: (1x21x768xf32, 1x21x768xui8) <- (1x21x768xf32, None, 1xf32) dropout_64, dropout_65 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_97, None, full_5, False, "upscale_in_train", 0, False + add_97, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -2025,6 +2005,7 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 1x21x768xf32) add_98 = paddle._C_ops.add(layer_norm_60, dropout_64) + del dropout_64, layer_norm_60 # pd_op.layer_norm: (1x21x768xf32, 1x21xf32, 1x21xf32) <- (1x21x768xf32, 768xf32, 768xf32) layer_norm_63, layer_norm_64, layer_norm_65 = (lambda x, f: f(x))( @@ -2033,7 +2014,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_20, parameter_21 + del add_98, parameter_20, parameter_21 # pd_op.matmul: (1x21x3072xf32) <- (1x21x768xf32, 768x3072xf32) matmul_86 = paddle._C_ops.matmul(layer_norm_63, parameter_25, False, False) @@ -2041,23 +2022,24 @@ def forward( # pd_op.add: (1x21x3072xf32) <- (1x21x3072xf32, 3072xf32) add_99 = paddle._C_ops.add(matmul_86, parameter_24) - del parameter_24 + del matmul_86, parameter_24 # pd_op.gelu: (1x21x3072xf32) <- (1x21x3072xf32) gelu_10 = paddle._C_ops.gelu(add_99, False) + del add_99 # pd_op.matmul: (1x21x768xf32) <- (1x21x3072xf32, 3072x768xf32) matmul_87 = paddle._C_ops.matmul(gelu_10, parameter_23, False, False) - del parameter_23 + del gelu_10, parameter_23 # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_100 = paddle._C_ops.add(matmul_87, parameter_22) - del parameter_22 + del matmul_87, parameter_22 # pd_op.dropout: (1x21x768xf32, 1x21x768xui8) <- (1x21x768xf32, None, 1xf32) dropout_66, dropout_67 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_100, None, full_5, False, "upscale_in_train", 0, False + add_100, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -2065,6 +2047,7 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 1x21x768xf32) add_101 = paddle._C_ops.add(layer_norm_63, dropout_66) + del dropout_66, layer_norm_63 # pd_op.layer_norm: (1x21x768xf32, 1x21xf32, 1x21xf32) <- (1x21x768xf32, 768xf32, 768xf32) layer_norm_66, layer_norm_67, layer_norm_68 = (lambda x, f: f(x))( @@ -2073,7 +2056,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_18, parameter_19 + del add_101, parameter_18, parameter_19 # pd_op.matmul: (1x21x768xf32) <- (1x21x768xf32, 768x768xf32) matmul_88 = paddle._C_ops.matmul(layer_norm_66, parameter_17, False, False) @@ -2081,10 +2064,11 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_102 = paddle._C_ops.add(matmul_88, parameter_16) - del parameter_16 + del matmul_88, parameter_16 # pd_op.reshape: (1x21x12x64xf32) <- (1x21x768xf32, 4xi64) reshape_44 = paddle._C_ops.reshape(add_102, full_int_array_1) + del add_102 # pd_op.transpose: (1x12x21x64xf32) <- (1x21x12x64xf32) transpose_44 = paddle._C_ops.transpose(reshape_44, [0, 2, 1, 3]) @@ -2096,7 +2080,7 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_103 = paddle._C_ops.add(matmul_89, parameter_14) - del parameter_14 + del matmul_89, parameter_14 # pd_op.matmul: (1x21x768xf32) <- (1x21x768xf32, 768x768xf32) matmul_90 = paddle._C_ops.matmul(layer_norm_66, parameter_13, False, False) @@ -2104,10 +2088,11 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_104 = paddle._C_ops.add(matmul_90, parameter_12) - del parameter_12 + del matmul_90, parameter_12 # pd_op.reshape: (1x21x12x64xf32) <- (1x21x768xf32, 4xi64) reshape_45 = paddle._C_ops.reshape(add_103, full_int_array_1) + del add_103 # pd_op.transpose: (1x12x21x64xf32) <- (1x21x12x64xf32) transpose_45 = paddle._C_ops.transpose(reshape_45, [0, 2, 1, 3]) @@ -2115,7 +2100,7 @@ def forward( # pd_op.reshape: (1x21x12x64xf32) <- (1x21x768xf32, 4xi64) reshape_46 = paddle._C_ops.reshape(add_104, full_int_array_1) - del full_int_array_1 + del add_104, full_int_array_1 # pd_op.transpose: (1x12x21x64xf32) <- (1x21x12x64xf32) transpose_46 = paddle._C_ops.transpose(reshape_46, [0, 2, 1, 3]) @@ -2123,13 +2108,15 @@ def forward( # pd_op.scale: (1x12x21x64xf32) <- (1x12x21x64xf32, 1xf32) scale_13 = paddle._C_ops.scale(transpose_44, full_6, float("0"), True) - del transpose_44 + del full_6, transpose_44 # pd_op.matmul: (1x12x21x21xf32) <- (1x12x21x64xf32, 1x12x21x64xf32) matmul_91 = paddle._C_ops.matmul(scale_13, transpose_45, False, True) + del scale_13, transpose_45 # pd_op.add: (1x12x21x21xf32) <- (1x12x21x21xf32, 1x1x1x21xf32) add_105 = paddle._C_ops.add(matmul_91, unsqueeze_0) + del matmul_91, unsqueeze_0 # pd_op.softmax: (1x12x21x21xf32) <- (1x12x21x21xf32) softmax_11 = paddle._C_ops.softmax(add_105, -1) @@ -2138,13 +2125,15 @@ def forward( # pd_op.dropout: (1x12x21x21xf32, 1x12x21x21xui8) <- (1x12x21x21xf32, None, 1xf32) dropout_68, dropout_69 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_11, None, full_5, False, "upscale_in_train", 0, False + softmax_11, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_11 # pd_op.matmul: (1x12x21x64xf32) <- (1x12x21x21xf32, 1x12x21x64xf32) matmul_92 = paddle._C_ops.matmul(dropout_68, transpose_46, False, False) + del dropout_68, transpose_46 # pd_op.transpose: (1x21x12x64xf32) <- (1x12x21x64xf32) transpose_47 = paddle._C_ops.transpose(matmul_92, [0, 2, 1, 3]) @@ -2152,20 +2141,20 @@ def forward( # pd_op.reshape: (1x21x768xf32) <- (1x21x12x64xf32, 3xi64) reshape_47 = paddle._C_ops.reshape(transpose_47, full_int_array_2) - del full_int_array_2 + del full_int_array_2, transpose_47 # pd_op.matmul: (1x21x768xf32) <- (1x21x768xf32, 768x768xf32) matmul_93 = paddle._C_ops.matmul(reshape_47, parameter_11, False, False) - del parameter_11 + del parameter_11, reshape_47 # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_106 = paddle._C_ops.add(matmul_93, parameter_10) - del parameter_10 + del matmul_93, parameter_10 # pd_op.dropout: (1x21x768xf32, 1x21x768xui8) <- (1x21x768xf32, None, 1xf32) dropout_70, dropout_71 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_106, None, full_5, False, "upscale_in_train", 0, False + add_106, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -2173,6 +2162,7 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 1x21x768xf32) add_107 = paddle._C_ops.add(layer_norm_66, dropout_70) + del dropout_70, layer_norm_66 # pd_op.layer_norm: (1x21x768xf32, 1x21xf32, 1x21xf32) <- (1x21x768xf32, 768xf32, 768xf32) layer_norm_69, layer_norm_70, layer_norm_71 = (lambda x, f: f(x))( @@ -2181,7 +2171,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_4, parameter_5 + del add_107, parameter_4, parameter_5 # pd_op.matmul: (1x21x3072xf32) <- (1x21x768xf32, 768x3072xf32) matmul_94 = paddle._C_ops.matmul(layer_norm_69, parameter_9, False, False) @@ -2189,30 +2179,32 @@ def forward( # pd_op.add: (1x21x3072xf32) <- (1x21x3072xf32, 3072xf32) add_108 = paddle._C_ops.add(matmul_94, parameter_8) - del parameter_8 + del matmul_94, parameter_8 # pd_op.gelu: (1x21x3072xf32) <- (1x21x3072xf32) gelu_11 = paddle._C_ops.gelu(add_108, False) + del add_108 # pd_op.matmul: (1x21x768xf32) <- (1x21x3072xf32, 3072x768xf32) matmul_95 = paddle._C_ops.matmul(gelu_11, parameter_7, False, False) - del parameter_7 + del gelu_11, parameter_7 # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_109 = paddle._C_ops.add(matmul_95, parameter_6) - del parameter_6 + del matmul_95, parameter_6 # pd_op.dropout: (1x21x768xf32, 1x21x768xui8) <- (1x21x768xf32, None, 1xf32) dropout_72, dropout_73 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_109, None, full_5, False, "upscale_in_train", 0, False + add_109, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) - del add_109 + del add_109, full_5 # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 1x21x768xf32) add_110 = paddle._C_ops.add(layer_norm_69, dropout_72) + del dropout_72, layer_norm_69 # pd_op.layer_norm: (1x21x768xf32, 1x21xf32, 1x21xf32) <- (1x21x768xf32, 768xf32, 768xf32) layer_norm_72, layer_norm_73, layer_norm_74 = (lambda x, f: f(x))( @@ -2221,7 +2213,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_2, parameter_3 + del add_110, parameter_2, parameter_3 # pd_op.full_int_array: (1xi64) <- () full_int_array_3 = [0] @@ -2233,470 +2225,18 @@ def forward( slice_0 = paddle._C_ops.slice( layer_norm_72, [1], full_int_array_3, full_int_array_4, [1], [1] ) + del full_int_array_3, full_int_array_4, layer_norm_72 # pd_op.matmul: (1x768xf32) <- (1x768xf32, 768x768xf32) matmul_96 = paddle._C_ops.matmul(slice_0, parameter_1, False, False) - del parameter_1 + del parameter_1, slice_0 # pd_op.add: (1x768xf32) <- (1x768xf32, 768xf32) add_111 = paddle._C_ops.add(matmul_96, parameter_0) - del parameter_0 + del matmul_96, parameter_0 # pd_op.tanh: (1x768xf32) <- (1x768xf32) tanh_0 = paddle._C_ops.tanh(add_111) - del ( - add_0, - add_1, - add_101, - add_102, - add_103, - add_104, - add_107, - add_108, - add_11, - add_110, - add_111, - add_12, - add_13, - add_14, - add_17, - add_18, - add_2, - add_20, - add_21, - add_22, - add_23, - add_26, - add_27, - add_29, - add_3, - add_30, - add_31, - add_32, - add_35, - add_36, - add_38, - add_39, - add_4, - add_40, - add_41, - add_44, - add_45, - add_47, - add_48, - add_49, - add_5, - add_50, - add_53, - add_54, - add_56, - add_57, - add_58, - add_59, - add_62, - add_63, - add_65, - add_66, - add_67, - add_68, - add_71, - add_72, - add_74, - add_75, - add_76, - add_77, - add_8, - add_80, - add_81, - add_83, - add_84, - add_85, - add_86, - add_89, - add_9, - add_90, - add_92, - add_93, - add_94, - add_95, - add_98, - add_99, - assign_0, - assign_1, - assign_10, - assign_11, - assign_12, - assign_13, - assign_14, - assign_15, - assign_16, - assign_17, - assign_18, - assign_19, - assign_2, - assign_20, - assign_21, - assign_22, - assign_23, - assign_24, - assign_25, - assign_26, - assign_27, - assign_28, - assign_29, - assign_3, - assign_30, - assign_31, - assign_32, - assign_33, - assign_34, - assign_35, - assign_36, - assign_37, - assign_38, - assign_39, - assign_4, - assign_40, - assign_41, - assign_42, - assign_43, - assign_44, - assign_45, - assign_46, - assign_5, - assign_6, - assign_7, - assign_8, - assign_9, - dropout_0, - dropout_1, - dropout_10, - dropout_11, - dropout_12, - dropout_13, - dropout_14, - dropout_15, - dropout_16, - dropout_17, - dropout_18, - dropout_19, - dropout_2, - dropout_20, - dropout_21, - dropout_22, - dropout_23, - dropout_24, - dropout_25, - dropout_26, - dropout_27, - dropout_28, - dropout_29, - dropout_3, - dropout_30, - dropout_31, - dropout_32, - dropout_33, - dropout_34, - dropout_35, - dropout_36, - dropout_37, - dropout_38, - dropout_39, - dropout_4, - dropout_40, - dropout_41, - dropout_42, - dropout_43, - dropout_44, - dropout_45, - dropout_46, - dropout_47, - dropout_48, - dropout_49, - dropout_5, - dropout_50, - dropout_51, - dropout_52, - dropout_53, - dropout_54, - dropout_55, - dropout_56, - dropout_57, - dropout_58, - dropout_59, - dropout_6, - dropout_60, - dropout_61, - dropout_62, - dropout_63, - dropout_64, - dropout_65, - dropout_66, - dropout_67, - dropout_68, - dropout_69, - dropout_7, - dropout_70, - dropout_71, - dropout_72, - dropout_73, - dropout_8, - dropout_9, - embedding_0, - embedding_1, - embedding_2, - embedding_3, - full_5, - full_6, - full_int_array_3, - full_int_array_4, - gelu_0, - gelu_1, - gelu_10, - gelu_11, - gelu_2, - gelu_3, - gelu_4, - gelu_5, - gelu_6, - gelu_7, - gelu_8, - gelu_9, - layer_norm_1, - layer_norm_10, - layer_norm_11, - layer_norm_12, - layer_norm_13, - layer_norm_14, - layer_norm_15, - layer_norm_16, - layer_norm_17, - layer_norm_18, - layer_norm_19, - layer_norm_2, - layer_norm_20, - layer_norm_21, - layer_norm_22, - layer_norm_23, - layer_norm_24, - layer_norm_25, - layer_norm_26, - layer_norm_27, - layer_norm_28, - layer_norm_29, - layer_norm_3, - layer_norm_30, - layer_norm_31, - layer_norm_32, - layer_norm_33, - layer_norm_34, - layer_norm_35, - layer_norm_36, - layer_norm_37, - layer_norm_38, - layer_norm_39, - layer_norm_4, - layer_norm_40, - layer_norm_41, - layer_norm_42, - layer_norm_43, - layer_norm_44, - layer_norm_45, - layer_norm_46, - layer_norm_47, - layer_norm_48, - layer_norm_49, - layer_norm_5, - layer_norm_50, - layer_norm_51, - layer_norm_52, - layer_norm_53, - layer_norm_54, - layer_norm_55, - layer_norm_56, - layer_norm_57, - layer_norm_58, - layer_norm_59, - layer_norm_6, - layer_norm_60, - layer_norm_61, - layer_norm_62, - layer_norm_63, - layer_norm_64, - layer_norm_65, - layer_norm_66, - layer_norm_67, - layer_norm_68, - layer_norm_69, - layer_norm_7, - layer_norm_70, - layer_norm_71, - layer_norm_72, - layer_norm_73, - layer_norm_74, - layer_norm_8, - layer_norm_9, - matmul_0, - matmul_1, - matmul_10, - matmul_11, - matmul_13, - matmul_14, - matmul_15, - matmul_16, - matmul_17, - matmul_18, - matmul_19, - matmul_2, - matmul_21, - matmul_22, - matmul_23, - matmul_24, - matmul_25, - matmul_26, - matmul_27, - matmul_29, - matmul_3, - matmul_30, - matmul_31, - matmul_32, - matmul_33, - matmul_34, - matmul_35, - matmul_37, - matmul_38, - matmul_39, - matmul_40, - matmul_41, - matmul_42, - matmul_43, - matmul_45, - matmul_46, - matmul_47, - matmul_48, - matmul_49, - matmul_5, - matmul_50, - matmul_51, - matmul_53, - matmul_54, - matmul_55, - matmul_56, - matmul_57, - matmul_58, - matmul_59, - matmul_6, - matmul_61, - matmul_62, - matmul_63, - matmul_64, - matmul_65, - matmul_66, - matmul_67, - matmul_69, - matmul_7, - matmul_70, - matmul_71, - matmul_72, - matmul_73, - matmul_74, - matmul_75, - matmul_77, - matmul_78, - matmul_79, - matmul_8, - matmul_80, - matmul_81, - matmul_82, - matmul_83, - matmul_85, - matmul_86, - matmul_87, - matmul_88, - matmul_89, - matmul_9, - matmul_90, - matmul_91, - matmul_93, - matmul_94, - matmul_95, - matmul_96, - reshape_11, - reshape_15, - reshape_19, - reshape_23, - reshape_27, - reshape_3, - reshape_31, - reshape_35, - reshape_39, - reshape_43, - reshape_47, - reshape_7, - scale_1, - scale_10, - scale_11, - scale_12, - scale_13, - scale_2, - scale_3, - scale_4, - scale_5, - scale_6, - scale_7, - scale_8, - scale_9, - slice_0, - softmax_0, - softmax_1, - softmax_10, - softmax_11, - softmax_2, - softmax_3, - softmax_4, - softmax_5, - softmax_6, - softmax_7, - softmax_8, - softmax_9, - subtract_0, - transpose_1, - transpose_10, - transpose_11, - transpose_13, - transpose_14, - transpose_15, - transpose_17, - transpose_18, - transpose_19, - transpose_2, - transpose_21, - transpose_22, - transpose_23, - transpose_25, - transpose_26, - transpose_27, - transpose_29, - transpose_3, - transpose_30, - transpose_31, - transpose_33, - transpose_34, - transpose_35, - transpose_37, - transpose_38, - transpose_39, - transpose_41, - transpose_42, - transpose_43, - transpose_45, - transpose_46, - transpose_47, - transpose_5, - transpose_6, - transpose_7, - transpose_9, - unsqueeze_0, - ) + del add_111 return tanh_0 diff --git a/paddle_samples/PaddleNLP/rocketqa-base-cross-encoder/weight_meta.py b/paddle_samples/PaddleNLP/rocketqa-base-cross-encoder/weight_meta.py index cde86f006..bde4d3ceb 100644 --- a/paddle_samples/PaddleNLP/rocketqa-base-cross-encoder/weight_meta.py +++ b/paddle_samples/PaddleNLP/rocketqa-base-cross-encoder/weight_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_parameter_0: name = "parameter_0" + original_name = "linear_1622.b_0" shape = [768] dtype = "float32" min_val = float("-0.498809") @@ -11,6 +12,7 @@ class Program_weight_tensor_parameter_0: class Program_weight_tensor_parameter_1: name = "parameter_1" + original_name = "linear_1622.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.326919") @@ -22,6 +24,7 @@ class Program_weight_tensor_parameter_1: class Program_weight_tensor_parameter_2: name = "parameter_2" + original_name = "layer_norm_558.b_0" shape = [768] dtype = "float32" min_val = float("-0.864449") @@ -33,6 +36,7 @@ class Program_weight_tensor_parameter_2: class Program_weight_tensor_parameter_3: name = "parameter_3" + original_name = "layer_norm_558.w_0" shape = [768] dtype = "float32" min_val = float("0.0875785") @@ -44,6 +48,7 @@ class Program_weight_tensor_parameter_3: class Program_weight_tensor_parameter_4: name = "parameter_4" + original_name = "layer_norm_557.b_0" shape = [768] dtype = "float32" min_val = float("-2.51454") @@ -55,6 +60,7 @@ class Program_weight_tensor_parameter_4: class Program_weight_tensor_parameter_5: name = "parameter_5" + original_name = "layer_norm_557.w_0" shape = [768] dtype = "float32" min_val = float("0.124387") @@ -66,6 +72,7 @@ class Program_weight_tensor_parameter_5: class Program_weight_tensor_parameter_6: name = "parameter_6" + original_name = "linear_1621.b_0" shape = [768] dtype = "float32" min_val = float("-1.63076") @@ -77,6 +84,7 @@ class Program_weight_tensor_parameter_6: class Program_weight_tensor_parameter_7: name = "parameter_7" + original_name = "linear_1621.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-1.39947") @@ -88,6 +96,7 @@ class Program_weight_tensor_parameter_7: class Program_weight_tensor_parameter_8: name = "parameter_8" + original_name = "linear_1620.b_0" shape = [3072] dtype = "float32" min_val = float("-2.39579") @@ -99,6 +108,7 @@ class Program_weight_tensor_parameter_8: class Program_weight_tensor_parameter_9: name = "parameter_9" + original_name = "linear_1620.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.607481") @@ -110,6 +120,7 @@ class Program_weight_tensor_parameter_9: class Program_weight_tensor_parameter_10: name = "parameter_10" + original_name = "linear_1619.b_0" shape = [768] dtype = "float32" min_val = float("-0.814027") @@ -121,6 +132,7 @@ class Program_weight_tensor_parameter_10: class Program_weight_tensor_parameter_11: name = "parameter_11" + original_name = "linear_1619.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.648169") @@ -132,6 +144,7 @@ class Program_weight_tensor_parameter_11: class Program_weight_tensor_parameter_12: name = "parameter_12" + original_name = "linear_1618.b_0" shape = [768] dtype = "float32" min_val = float("-0.811642") @@ -143,6 +156,7 @@ class Program_weight_tensor_parameter_12: class Program_weight_tensor_parameter_13: name = "parameter_13" + original_name = "linear_1618.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.309619") @@ -154,6 +168,7 @@ class Program_weight_tensor_parameter_13: class Program_weight_tensor_parameter_14: name = "parameter_14" + original_name = "linear_1617.b_0" shape = [768] dtype = "float32" min_val = float("-16.6109") @@ -165,6 +180,7 @@ class Program_weight_tensor_parameter_14: class Program_weight_tensor_parameter_15: name = "parameter_15" + original_name = "linear_1617.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.653577") @@ -176,6 +192,7 @@ class Program_weight_tensor_parameter_15: class Program_weight_tensor_parameter_16: name = "parameter_16" + original_name = "linear_1616.b_0" shape = [768] dtype = "float32" min_val = float("-3.43004") @@ -187,6 +204,7 @@ class Program_weight_tensor_parameter_16: class Program_weight_tensor_parameter_17: name = "parameter_17" + original_name = "linear_1616.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.697174") @@ -198,6 +216,7 @@ class Program_weight_tensor_parameter_17: class Program_weight_tensor_parameter_18: name = "parameter_18" + original_name = "layer_norm_556.b_0" shape = [768] dtype = "float32" min_val = float("-3.47037") @@ -209,6 +228,7 @@ class Program_weight_tensor_parameter_18: class Program_weight_tensor_parameter_19: name = "parameter_19" + original_name = "layer_norm_556.w_0" shape = [768] dtype = "float32" min_val = float("0.181489") @@ -220,6 +240,7 @@ class Program_weight_tensor_parameter_19: class Program_weight_tensor_parameter_20: name = "parameter_20" + original_name = "layer_norm_555.b_0" shape = [768] dtype = "float32" min_val = float("-5.16923") @@ -231,6 +252,7 @@ class Program_weight_tensor_parameter_20: class Program_weight_tensor_parameter_21: name = "parameter_21" + original_name = "layer_norm_555.w_0" shape = [768] dtype = "float32" min_val = float("0.504545") @@ -242,6 +264,7 @@ class Program_weight_tensor_parameter_21: class Program_weight_tensor_parameter_22: name = "parameter_22" + original_name = "linear_1615.b_0" shape = [768] dtype = "float32" min_val = float("-0.34357") @@ -253,6 +276,7 @@ class Program_weight_tensor_parameter_22: class Program_weight_tensor_parameter_23: name = "parameter_23" + original_name = "linear_1615.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-0.961195") @@ -264,6 +288,7 @@ class Program_weight_tensor_parameter_23: class Program_weight_tensor_parameter_24: name = "parameter_24" + original_name = "linear_1614.b_0" shape = [3072] dtype = "float32" min_val = float("-2.17209") @@ -275,6 +300,7 @@ class Program_weight_tensor_parameter_24: class Program_weight_tensor_parameter_25: name = "parameter_25" + original_name = "linear_1614.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.599701") @@ -286,6 +312,7 @@ class Program_weight_tensor_parameter_25: class Program_weight_tensor_parameter_26: name = "parameter_26" + original_name = "linear_1613.b_0" shape = [768] dtype = "float32" min_val = float("-0.245636") @@ -297,6 +324,7 @@ class Program_weight_tensor_parameter_26: class Program_weight_tensor_parameter_27: name = "parameter_27" + original_name = "linear_1613.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.891349") @@ -308,6 +336,7 @@ class Program_weight_tensor_parameter_27: class Program_weight_tensor_parameter_28: name = "parameter_28" + original_name = "linear_1612.b_0" shape = [768] dtype = "float32" min_val = float("-0.52809") @@ -319,6 +348,7 @@ class Program_weight_tensor_parameter_28: class Program_weight_tensor_parameter_29: name = "parameter_29" + original_name = "linear_1612.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.357214") @@ -330,6 +360,7 @@ class Program_weight_tensor_parameter_29: class Program_weight_tensor_parameter_30: name = "parameter_30" + original_name = "linear_1611.b_0" shape = [768] dtype = "float32" min_val = float("-18.3291") @@ -341,6 +372,7 @@ class Program_weight_tensor_parameter_30: class Program_weight_tensor_parameter_31: name = "parameter_31" + original_name = "linear_1611.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.467078") @@ -352,6 +384,7 @@ class Program_weight_tensor_parameter_31: class Program_weight_tensor_parameter_32: name = "parameter_32" + original_name = "linear_1610.b_0" shape = [768] dtype = "float32" min_val = float("-3.22234") @@ -363,6 +396,7 @@ class Program_weight_tensor_parameter_32: class Program_weight_tensor_parameter_33: name = "parameter_33" + original_name = "linear_1610.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.516459") @@ -374,6 +408,7 @@ class Program_weight_tensor_parameter_33: class Program_weight_tensor_parameter_34: name = "parameter_34" + original_name = "layer_norm_554.b_0" shape = [768] dtype = "float32" min_val = float("-0.530428") @@ -385,6 +420,7 @@ class Program_weight_tensor_parameter_34: class Program_weight_tensor_parameter_35: name = "parameter_35" + original_name = "layer_norm_554.w_0" shape = [768] dtype = "float32" min_val = float("0.2322") @@ -396,6 +432,7 @@ class Program_weight_tensor_parameter_35: class Program_weight_tensor_parameter_36: name = "parameter_36" + original_name = "layer_norm_553.b_0" shape = [768] dtype = "float32" min_val = float("-5.74524") @@ -407,6 +444,7 @@ class Program_weight_tensor_parameter_36: class Program_weight_tensor_parameter_37: name = "parameter_37" + original_name = "layer_norm_553.w_0" shape = [768] dtype = "float32" min_val = float("0.476349") @@ -418,6 +456,7 @@ class Program_weight_tensor_parameter_37: class Program_weight_tensor_parameter_38: name = "parameter_38" + original_name = "linear_1609.b_0" shape = [768] dtype = "float32" min_val = float("-1.21435") @@ -429,6 +468,7 @@ class Program_weight_tensor_parameter_38: class Program_weight_tensor_parameter_39: name = "parameter_39" + original_name = "linear_1609.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-1.2004") @@ -440,6 +480,7 @@ class Program_weight_tensor_parameter_39: class Program_weight_tensor_parameter_40: name = "parameter_40" + original_name = "linear_1608.b_0" shape = [3072] dtype = "float32" min_val = float("-1.41819") @@ -451,6 +492,7 @@ class Program_weight_tensor_parameter_40: class Program_weight_tensor_parameter_41: name = "parameter_41" + original_name = "linear_1608.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.601805") @@ -462,6 +504,7 @@ class Program_weight_tensor_parameter_41: class Program_weight_tensor_parameter_42: name = "parameter_42" + original_name = "linear_1607.b_0" shape = [768] dtype = "float32" min_val = float("-0.24089") @@ -473,6 +516,7 @@ class Program_weight_tensor_parameter_42: class Program_weight_tensor_parameter_43: name = "parameter_43" + original_name = "linear_1607.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.810296") @@ -484,6 +528,7 @@ class Program_weight_tensor_parameter_43: class Program_weight_tensor_parameter_44: name = "parameter_44" + original_name = "linear_1606.b_0" shape = [768] dtype = "float32" min_val = float("-0.80608") @@ -495,6 +540,7 @@ class Program_weight_tensor_parameter_44: class Program_weight_tensor_parameter_45: name = "parameter_45" + original_name = "linear_1606.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.459053") @@ -506,6 +552,7 @@ class Program_weight_tensor_parameter_45: class Program_weight_tensor_parameter_46: name = "parameter_46" + original_name = "linear_1605.b_0" shape = [768] dtype = "float32" min_val = float("-16.4834") @@ -517,6 +564,7 @@ class Program_weight_tensor_parameter_46: class Program_weight_tensor_parameter_47: name = "parameter_47" + original_name = "linear_1605.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.457381") @@ -528,6 +576,7 @@ class Program_weight_tensor_parameter_47: class Program_weight_tensor_parameter_48: name = "parameter_48" + original_name = "linear_1604.b_0" shape = [768] dtype = "float32" min_val = float("-3.06178") @@ -539,6 +588,7 @@ class Program_weight_tensor_parameter_48: class Program_weight_tensor_parameter_49: name = "parameter_49" + original_name = "linear_1604.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.415615") @@ -550,6 +600,7 @@ class Program_weight_tensor_parameter_49: class Program_weight_tensor_parameter_50: name = "parameter_50" + original_name = "layer_norm_552.b_0" shape = [768] dtype = "float32" min_val = float("-0.694078") @@ -561,6 +612,7 @@ class Program_weight_tensor_parameter_50: class Program_weight_tensor_parameter_51: name = "parameter_51" + original_name = "layer_norm_552.w_0" shape = [768] dtype = "float32" min_val = float("0.117814") @@ -572,6 +624,7 @@ class Program_weight_tensor_parameter_51: class Program_weight_tensor_parameter_52: name = "parameter_52" + original_name = "layer_norm_551.b_0" shape = [768] dtype = "float32" min_val = float("-8.1952") @@ -583,6 +636,7 @@ class Program_weight_tensor_parameter_52: class Program_weight_tensor_parameter_53: name = "parameter_53" + original_name = "layer_norm_551.w_0" shape = [768] dtype = "float32" min_val = float("0.477142") @@ -594,6 +648,7 @@ class Program_weight_tensor_parameter_53: class Program_weight_tensor_parameter_54: name = "parameter_54" + original_name = "linear_1603.b_0" shape = [768] dtype = "float32" min_val = float("-0.707219") @@ -605,6 +660,7 @@ class Program_weight_tensor_parameter_54: class Program_weight_tensor_parameter_55: name = "parameter_55" + original_name = "linear_1603.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-1.91938") @@ -616,6 +672,7 @@ class Program_weight_tensor_parameter_55: class Program_weight_tensor_parameter_56: name = "parameter_56" + original_name = "linear_1602.b_0" shape = [3072] dtype = "float32" min_val = float("-1.41629") @@ -627,6 +684,7 @@ class Program_weight_tensor_parameter_56: class Program_weight_tensor_parameter_57: name = "parameter_57" + original_name = "linear_1602.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.553806") @@ -638,6 +696,7 @@ class Program_weight_tensor_parameter_57: class Program_weight_tensor_parameter_58: name = "parameter_58" + original_name = "linear_1601.b_0" shape = [768] dtype = "float32" min_val = float("-0.158281") @@ -649,6 +708,7 @@ class Program_weight_tensor_parameter_58: class Program_weight_tensor_parameter_59: name = "parameter_59" + original_name = "linear_1601.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.368402") @@ -660,6 +720,7 @@ class Program_weight_tensor_parameter_59: class Program_weight_tensor_parameter_60: name = "parameter_60" + original_name = "linear_1600.b_0" shape = [768] dtype = "float32" min_val = float("-0.523903") @@ -671,6 +732,7 @@ class Program_weight_tensor_parameter_60: class Program_weight_tensor_parameter_61: name = "parameter_61" + original_name = "linear_1600.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.314065") @@ -682,6 +744,7 @@ class Program_weight_tensor_parameter_61: class Program_weight_tensor_parameter_62: name = "parameter_62" + original_name = "linear_1599.b_0" shape = [768] dtype = "float32" min_val = float("-13.719") @@ -693,6 +756,7 @@ class Program_weight_tensor_parameter_62: class Program_weight_tensor_parameter_63: name = "parameter_63" + original_name = "linear_1599.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.5027") @@ -704,6 +768,7 @@ class Program_weight_tensor_parameter_63: class Program_weight_tensor_parameter_64: name = "parameter_64" + original_name = "linear_1598.b_0" shape = [768] dtype = "float32" min_val = float("-3.2073") @@ -715,6 +780,7 @@ class Program_weight_tensor_parameter_64: class Program_weight_tensor_parameter_65: name = "parameter_65" + original_name = "linear_1598.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.395813") @@ -726,6 +792,7 @@ class Program_weight_tensor_parameter_65: class Program_weight_tensor_parameter_66: name = "parameter_66" + original_name = "layer_norm_550.b_0" shape = [768] dtype = "float32" min_val = float("-0.612544") @@ -737,6 +804,7 @@ class Program_weight_tensor_parameter_66: class Program_weight_tensor_parameter_67: name = "parameter_67" + original_name = "layer_norm_550.w_0" shape = [768] dtype = "float32" min_val = float("0.146122") @@ -748,6 +816,7 @@ class Program_weight_tensor_parameter_67: class Program_weight_tensor_parameter_68: name = "parameter_68" + original_name = "layer_norm_549.b_0" shape = [768] dtype = "float32" min_val = float("-7.49924") @@ -759,6 +828,7 @@ class Program_weight_tensor_parameter_68: class Program_weight_tensor_parameter_69: name = "parameter_69" + original_name = "layer_norm_549.w_0" shape = [768] dtype = "float32" min_val = float("0.330849") @@ -770,6 +840,7 @@ class Program_weight_tensor_parameter_69: class Program_weight_tensor_parameter_70: name = "parameter_70" + original_name = "linear_1597.b_0" shape = [768] dtype = "float32" min_val = float("-0.67516") @@ -781,6 +852,7 @@ class Program_weight_tensor_parameter_70: class Program_weight_tensor_parameter_71: name = "parameter_71" + original_name = "linear_1597.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-1.5706") @@ -792,6 +864,7 @@ class Program_weight_tensor_parameter_71: class Program_weight_tensor_parameter_72: name = "parameter_72" + original_name = "linear_1596.b_0" shape = [3072] dtype = "float32" min_val = float("-1.52043") @@ -803,6 +876,7 @@ class Program_weight_tensor_parameter_72: class Program_weight_tensor_parameter_73: name = "parameter_73" + original_name = "linear_1596.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.757403") @@ -814,6 +888,7 @@ class Program_weight_tensor_parameter_73: class Program_weight_tensor_parameter_74: name = "parameter_74" + original_name = "linear_1595.b_0" shape = [768] dtype = "float32" min_val = float("-0.249734") @@ -825,6 +900,7 @@ class Program_weight_tensor_parameter_74: class Program_weight_tensor_parameter_75: name = "parameter_75" + original_name = "linear_1595.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.382012") @@ -836,6 +912,7 @@ class Program_weight_tensor_parameter_75: class Program_weight_tensor_parameter_76: name = "parameter_76" + original_name = "linear_1594.b_0" shape = [768] dtype = "float32" min_val = float("-0.666095") @@ -847,6 +924,7 @@ class Program_weight_tensor_parameter_76: class Program_weight_tensor_parameter_77: name = "parameter_77" + original_name = "linear_1594.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.285693") @@ -858,6 +936,7 @@ class Program_weight_tensor_parameter_77: class Program_weight_tensor_parameter_78: name = "parameter_78" + original_name = "linear_1593.b_0" shape = [768] dtype = "float32" min_val = float("-8.67944") @@ -869,6 +948,7 @@ class Program_weight_tensor_parameter_78: class Program_weight_tensor_parameter_79: name = "parameter_79" + original_name = "linear_1593.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.551004") @@ -880,6 +960,7 @@ class Program_weight_tensor_parameter_79: class Program_weight_tensor_parameter_80: name = "parameter_80" + original_name = "linear_1592.b_0" shape = [768] dtype = "float32" min_val = float("-2.53412") @@ -891,6 +972,7 @@ class Program_weight_tensor_parameter_80: class Program_weight_tensor_parameter_81: name = "parameter_81" + original_name = "linear_1592.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.491208") @@ -902,6 +984,7 @@ class Program_weight_tensor_parameter_81: class Program_weight_tensor_parameter_82: name = "parameter_82" + original_name = "layer_norm_548.b_0" shape = [768] dtype = "float32" min_val = float("-0.892455") @@ -913,6 +996,7 @@ class Program_weight_tensor_parameter_82: class Program_weight_tensor_parameter_83: name = "parameter_83" + original_name = "layer_norm_548.w_0" shape = [768] dtype = "float32" min_val = float("0.11622") @@ -924,6 +1008,7 @@ class Program_weight_tensor_parameter_83: class Program_weight_tensor_parameter_84: name = "parameter_84" + original_name = "layer_norm_547.b_0" shape = [768] dtype = "float32" min_val = float("-8.49308") @@ -935,6 +1020,7 @@ class Program_weight_tensor_parameter_84: class Program_weight_tensor_parameter_85: name = "parameter_85" + original_name = "layer_norm_547.w_0" shape = [768] dtype = "float32" min_val = float("0.325524") @@ -946,6 +1032,7 @@ class Program_weight_tensor_parameter_85: class Program_weight_tensor_parameter_86: name = "parameter_86" + original_name = "linear_1591.b_0" shape = [768] dtype = "float32" min_val = float("-0.497673") @@ -957,6 +1044,7 @@ class Program_weight_tensor_parameter_86: class Program_weight_tensor_parameter_87: name = "parameter_87" + original_name = "linear_1591.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-1.64564") @@ -968,6 +1056,7 @@ class Program_weight_tensor_parameter_87: class Program_weight_tensor_parameter_88: name = "parameter_88" + original_name = "linear_1590.b_0" shape = [3072] dtype = "float32" min_val = float("-1.22099") @@ -979,6 +1068,7 @@ class Program_weight_tensor_parameter_88: class Program_weight_tensor_parameter_89: name = "parameter_89" + original_name = "linear_1590.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.601596") @@ -990,6 +1080,7 @@ class Program_weight_tensor_parameter_89: class Program_weight_tensor_parameter_90: name = "parameter_90" + original_name = "linear_1589.b_0" shape = [768] dtype = "float32" min_val = float("-0.364153") @@ -1001,6 +1092,7 @@ class Program_weight_tensor_parameter_90: class Program_weight_tensor_parameter_91: name = "parameter_91" + original_name = "linear_1589.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.481266") @@ -1012,6 +1104,7 @@ class Program_weight_tensor_parameter_91: class Program_weight_tensor_parameter_92: name = "parameter_92" + original_name = "linear_1588.b_0" shape = [768] dtype = "float32" min_val = float("-0.456344") @@ -1023,6 +1116,7 @@ class Program_weight_tensor_parameter_92: class Program_weight_tensor_parameter_93: name = "parameter_93" + original_name = "linear_1588.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.334676") @@ -1034,6 +1128,7 @@ class Program_weight_tensor_parameter_93: class Program_weight_tensor_parameter_94: name = "parameter_94" + original_name = "linear_1587.b_0" shape = [768] dtype = "float32" min_val = float("-5.33118") @@ -1045,6 +1140,7 @@ class Program_weight_tensor_parameter_94: class Program_weight_tensor_parameter_95: name = "parameter_95" + original_name = "linear_1587.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.561344") @@ -1056,6 +1152,7 @@ class Program_weight_tensor_parameter_95: class Program_weight_tensor_parameter_96: name = "parameter_96" + original_name = "linear_1586.b_0" shape = [768] dtype = "float32" min_val = float("-3.20684") @@ -1067,6 +1164,7 @@ class Program_weight_tensor_parameter_96: class Program_weight_tensor_parameter_97: name = "parameter_97" + original_name = "linear_1586.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.617535") @@ -1078,6 +1176,7 @@ class Program_weight_tensor_parameter_97: class Program_weight_tensor_parameter_98: name = "parameter_98" + original_name = "layer_norm_546.b_0" shape = [768] dtype = "float32" min_val = float("-1.16303") @@ -1089,6 +1188,7 @@ class Program_weight_tensor_parameter_98: class Program_weight_tensor_parameter_99: name = "parameter_99" + original_name = "layer_norm_546.w_0" shape = [768] dtype = "float32" min_val = float("0.106") @@ -1100,6 +1200,7 @@ class Program_weight_tensor_parameter_99: class Program_weight_tensor_parameter_100: name = "parameter_100" + original_name = "layer_norm_545.b_0" shape = [768] dtype = "float32" min_val = float("-9.2892") @@ -1111,6 +1212,7 @@ class Program_weight_tensor_parameter_100: class Program_weight_tensor_parameter_101: name = "parameter_101" + original_name = "layer_norm_545.w_0" shape = [768] dtype = "float32" min_val = float("0.0319994") @@ -1122,6 +1224,7 @@ class Program_weight_tensor_parameter_101: class Program_weight_tensor_parameter_102: name = "parameter_102" + original_name = "linear_1585.b_0" shape = [768] dtype = "float32" min_val = float("-0.409938") @@ -1133,6 +1236,7 @@ class Program_weight_tensor_parameter_102: class Program_weight_tensor_parameter_103: name = "parameter_103" + original_name = "linear_1585.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-1.45995") @@ -1144,6 +1248,7 @@ class Program_weight_tensor_parameter_103: class Program_weight_tensor_parameter_104: name = "parameter_104" + original_name = "linear_1584.b_0" shape = [3072] dtype = "float32" min_val = float("-1.21247") @@ -1155,6 +1260,7 @@ class Program_weight_tensor_parameter_104: class Program_weight_tensor_parameter_105: name = "parameter_105" + original_name = "linear_1584.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-1.13567") @@ -1166,6 +1272,7 @@ class Program_weight_tensor_parameter_105: class Program_weight_tensor_parameter_106: name = "parameter_106" + original_name = "linear_1583.b_0" shape = [768] dtype = "float32" min_val = float("-0.224812") @@ -1177,6 +1284,7 @@ class Program_weight_tensor_parameter_106: class Program_weight_tensor_parameter_107: name = "parameter_107" + original_name = "linear_1583.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.79954") @@ -1188,6 +1296,7 @@ class Program_weight_tensor_parameter_107: class Program_weight_tensor_parameter_108: name = "parameter_108" + original_name = "linear_1582.b_0" shape = [768] dtype = "float32" min_val = float("-0.504901") @@ -1199,6 +1308,7 @@ class Program_weight_tensor_parameter_108: class Program_weight_tensor_parameter_109: name = "parameter_109" + original_name = "linear_1582.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.576404") @@ -1210,6 +1320,7 @@ class Program_weight_tensor_parameter_109: class Program_weight_tensor_parameter_110: name = "parameter_110" + original_name = "linear_1581.b_0" shape = [768] dtype = "float32" min_val = float("-3.32848") @@ -1221,6 +1332,7 @@ class Program_weight_tensor_parameter_110: class Program_weight_tensor_parameter_111: name = "parameter_111" + original_name = "linear_1581.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.476796") @@ -1232,6 +1344,7 @@ class Program_weight_tensor_parameter_111: class Program_weight_tensor_parameter_112: name = "parameter_112" + original_name = "linear_1580.b_0" shape = [768] dtype = "float32" min_val = float("-2.60835") @@ -1243,6 +1356,7 @@ class Program_weight_tensor_parameter_112: class Program_weight_tensor_parameter_113: name = "parameter_113" + original_name = "linear_1580.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.497061") @@ -1254,6 +1368,7 @@ class Program_weight_tensor_parameter_113: class Program_weight_tensor_parameter_114: name = "parameter_114" + original_name = "layer_norm_544.b_0" shape = [768] dtype = "float32" min_val = float("-1.30966") @@ -1265,6 +1380,7 @@ class Program_weight_tensor_parameter_114: class Program_weight_tensor_parameter_115: name = "parameter_115" + original_name = "layer_norm_544.w_0" shape = [768] dtype = "float32" min_val = float("0.097049") @@ -1276,6 +1392,7 @@ class Program_weight_tensor_parameter_115: class Program_weight_tensor_parameter_116: name = "parameter_116" + original_name = "layer_norm_543.b_0" shape = [768] dtype = "float32" min_val = float("-8.08497") @@ -1287,6 +1404,7 @@ class Program_weight_tensor_parameter_116: class Program_weight_tensor_parameter_117: name = "parameter_117" + original_name = "layer_norm_543.w_0" shape = [768] dtype = "float32" min_val = float("0.446064") @@ -1298,6 +1416,7 @@ class Program_weight_tensor_parameter_117: class Program_weight_tensor_parameter_118: name = "parameter_118" + original_name = "linear_1579.b_0" shape = [768] dtype = "float32" min_val = float("-0.648548") @@ -1309,6 +1428,7 @@ class Program_weight_tensor_parameter_118: class Program_weight_tensor_parameter_119: name = "parameter_119" + original_name = "linear_1579.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-2.18128") @@ -1320,6 +1440,7 @@ class Program_weight_tensor_parameter_119: class Program_weight_tensor_parameter_120: name = "parameter_120" + original_name = "linear_1578.b_0" shape = [3072] dtype = "float32" min_val = float("-1.04575") @@ -1331,6 +1452,7 @@ class Program_weight_tensor_parameter_120: class Program_weight_tensor_parameter_121: name = "parameter_121" + original_name = "linear_1578.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.866082") @@ -1342,6 +1464,7 @@ class Program_weight_tensor_parameter_121: class Program_weight_tensor_parameter_122: name = "parameter_122" + original_name = "linear_1577.b_0" shape = [768] dtype = "float32" min_val = float("-0.40419") @@ -1353,6 +1476,7 @@ class Program_weight_tensor_parameter_122: class Program_weight_tensor_parameter_123: name = "parameter_123" + original_name = "linear_1577.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.324451") @@ -1364,6 +1488,7 @@ class Program_weight_tensor_parameter_123: class Program_weight_tensor_parameter_124: name = "parameter_124" + original_name = "linear_1576.b_0" shape = [768] dtype = "float32" min_val = float("-0.473353") @@ -1375,6 +1500,7 @@ class Program_weight_tensor_parameter_124: class Program_weight_tensor_parameter_125: name = "parameter_125" + original_name = "linear_1576.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.606651") @@ -1386,6 +1512,7 @@ class Program_weight_tensor_parameter_125: class Program_weight_tensor_parameter_126: name = "parameter_126" + original_name = "linear_1575.b_0" shape = [768] dtype = "float32" min_val = float("-2.9598") @@ -1397,6 +1524,7 @@ class Program_weight_tensor_parameter_126: class Program_weight_tensor_parameter_127: name = "parameter_127" + original_name = "linear_1575.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.476815") @@ -1408,6 +1536,7 @@ class Program_weight_tensor_parameter_127: class Program_weight_tensor_parameter_128: name = "parameter_128" + original_name = "linear_1574.b_0" shape = [768] dtype = "float32" min_val = float("-3.56495") @@ -1419,6 +1548,7 @@ class Program_weight_tensor_parameter_128: class Program_weight_tensor_parameter_129: name = "parameter_129" + original_name = "linear_1574.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.434912") @@ -1430,6 +1560,7 @@ class Program_weight_tensor_parameter_129: class Program_weight_tensor_parameter_130: name = "parameter_130" + original_name = "layer_norm_542.b_0" shape = [768] dtype = "float32" min_val = float("-1.0069") @@ -1441,6 +1572,7 @@ class Program_weight_tensor_parameter_130: class Program_weight_tensor_parameter_131: name = "parameter_131" + original_name = "layer_norm_542.w_0" shape = [768] dtype = "float32" min_val = float("0.234443") @@ -1452,6 +1584,7 @@ class Program_weight_tensor_parameter_131: class Program_weight_tensor_parameter_132: name = "parameter_132" + original_name = "layer_norm_541.b_0" shape = [768] dtype = "float32" min_val = float("-7.10172") @@ -1463,6 +1596,7 @@ class Program_weight_tensor_parameter_132: class Program_weight_tensor_parameter_133: name = "parameter_133" + original_name = "layer_norm_541.w_0" shape = [768] dtype = "float32" min_val = float("0.537385") @@ -1474,6 +1608,7 @@ class Program_weight_tensor_parameter_133: class Program_weight_tensor_parameter_134: name = "parameter_134" + original_name = "linear_1573.b_0" shape = [768] dtype = "float32" min_val = float("-0.539993") @@ -1485,6 +1620,7 @@ class Program_weight_tensor_parameter_134: class Program_weight_tensor_parameter_135: name = "parameter_135" + original_name = "linear_1573.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-4.44542") @@ -1496,6 +1632,7 @@ class Program_weight_tensor_parameter_135: class Program_weight_tensor_parameter_136: name = "parameter_136" + original_name = "linear_1572.b_0" shape = [3072] dtype = "float32" min_val = float("-1.30935") @@ -1507,6 +1644,7 @@ class Program_weight_tensor_parameter_136: class Program_weight_tensor_parameter_137: name = "parameter_137" + original_name = "linear_1572.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.640649") @@ -1518,6 +1656,7 @@ class Program_weight_tensor_parameter_137: class Program_weight_tensor_parameter_138: name = "parameter_138" + original_name = "linear_1571.b_0" shape = [768] dtype = "float32" min_val = float("-0.304537") @@ -1529,6 +1668,7 @@ class Program_weight_tensor_parameter_138: class Program_weight_tensor_parameter_139: name = "parameter_139" + original_name = "linear_1571.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.69361") @@ -1540,6 +1680,7 @@ class Program_weight_tensor_parameter_139: class Program_weight_tensor_parameter_140: name = "parameter_140" + original_name = "linear_1570.b_0" shape = [768] dtype = "float32" min_val = float("-0.480467") @@ -1551,6 +1692,7 @@ class Program_weight_tensor_parameter_140: class Program_weight_tensor_parameter_141: name = "parameter_141" + original_name = "linear_1570.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.479873") @@ -1562,6 +1704,7 @@ class Program_weight_tensor_parameter_141: class Program_weight_tensor_parameter_142: name = "parameter_142" + original_name = "linear_1569.b_0" shape = [768] dtype = "float32" min_val = float("-1.91661") @@ -1573,6 +1716,7 @@ class Program_weight_tensor_parameter_142: class Program_weight_tensor_parameter_143: name = "parameter_143" + original_name = "linear_1569.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.517174") @@ -1584,6 +1728,7 @@ class Program_weight_tensor_parameter_143: class Program_weight_tensor_parameter_144: name = "parameter_144" + original_name = "linear_1568.b_0" shape = [768] dtype = "float32" min_val = float("-2.94715") @@ -1595,6 +1740,7 @@ class Program_weight_tensor_parameter_144: class Program_weight_tensor_parameter_145: name = "parameter_145" + original_name = "linear_1568.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.58856") @@ -1606,6 +1752,7 @@ class Program_weight_tensor_parameter_145: class Program_weight_tensor_parameter_146: name = "parameter_146" + original_name = "layer_norm_540.b_0" shape = [768] dtype = "float32" min_val = float("-1.44132") @@ -1617,6 +1764,7 @@ class Program_weight_tensor_parameter_146: class Program_weight_tensor_parameter_147: name = "parameter_147" + original_name = "layer_norm_540.w_0" shape = [768] dtype = "float32" min_val = float("0.161303") @@ -1628,6 +1776,7 @@ class Program_weight_tensor_parameter_147: class Program_weight_tensor_parameter_148: name = "parameter_148" + original_name = "layer_norm_539.b_0" shape = [768] dtype = "float32" min_val = float("-6.1715") @@ -1639,6 +1788,7 @@ class Program_weight_tensor_parameter_148: class Program_weight_tensor_parameter_149: name = "parameter_149" + original_name = "layer_norm_539.w_0" shape = [768] dtype = "float32" min_val = float("0.670692") @@ -1650,6 +1800,7 @@ class Program_weight_tensor_parameter_149: class Program_weight_tensor_parameter_150: name = "parameter_150" + original_name = "linear_1567.b_0" shape = [768] dtype = "float32" min_val = float("-0.621517") @@ -1661,6 +1812,7 @@ class Program_weight_tensor_parameter_150: class Program_weight_tensor_parameter_151: name = "parameter_151" + original_name = "linear_1567.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-1.61474") @@ -1672,6 +1824,7 @@ class Program_weight_tensor_parameter_151: class Program_weight_tensor_parameter_152: name = "parameter_152" + original_name = "linear_1566.b_0" shape = [3072] dtype = "float32" min_val = float("-0.704806") @@ -1683,6 +1836,7 @@ class Program_weight_tensor_parameter_152: class Program_weight_tensor_parameter_153: name = "parameter_153" + original_name = "linear_1566.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.626374") @@ -1694,6 +1848,7 @@ class Program_weight_tensor_parameter_153: class Program_weight_tensor_parameter_154: name = "parameter_154" + original_name = "linear_1565.b_0" shape = [768] dtype = "float32" min_val = float("-0.263621") @@ -1705,6 +1860,7 @@ class Program_weight_tensor_parameter_154: class Program_weight_tensor_parameter_155: name = "parameter_155" + original_name = "linear_1565.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.601757") @@ -1716,6 +1872,7 @@ class Program_weight_tensor_parameter_155: class Program_weight_tensor_parameter_156: name = "parameter_156" + original_name = "linear_1564.b_0" shape = [768] dtype = "float32" min_val = float("-0.311646") @@ -1727,6 +1884,7 @@ class Program_weight_tensor_parameter_156: class Program_weight_tensor_parameter_157: name = "parameter_157" + original_name = "linear_1564.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.37793") @@ -1738,6 +1896,7 @@ class Program_weight_tensor_parameter_157: class Program_weight_tensor_parameter_158: name = "parameter_158" + original_name = "linear_1563.b_0" shape = [768] dtype = "float32" min_val = float("-2.67565") @@ -1749,6 +1908,7 @@ class Program_weight_tensor_parameter_158: class Program_weight_tensor_parameter_159: name = "parameter_159" + original_name = "linear_1563.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.484645") @@ -1760,6 +1920,7 @@ class Program_weight_tensor_parameter_159: class Program_weight_tensor_parameter_160: name = "parameter_160" + original_name = "linear_1562.b_0" shape = [768] dtype = "float32" min_val = float("-3.28419") @@ -1771,6 +1932,7 @@ class Program_weight_tensor_parameter_160: class Program_weight_tensor_parameter_161: name = "parameter_161" + original_name = "linear_1562.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.441279") @@ -1782,6 +1944,7 @@ class Program_weight_tensor_parameter_161: class Program_weight_tensor_parameter_162: name = "parameter_162" + original_name = "layer_norm_538.b_0" shape = [768] dtype = "float32" min_val = float("-0.926914") @@ -1793,6 +1956,7 @@ class Program_weight_tensor_parameter_162: class Program_weight_tensor_parameter_163: name = "parameter_163" + original_name = "layer_norm_538.w_0" shape = [768] dtype = "float32" min_val = float("0.136294") @@ -1804,6 +1968,7 @@ class Program_weight_tensor_parameter_163: class Program_weight_tensor_parameter_164: name = "parameter_164" + original_name = "layer_norm_537.b_0" shape = [768] dtype = "float32" min_val = float("-5.39258") @@ -1815,6 +1980,7 @@ class Program_weight_tensor_parameter_164: class Program_weight_tensor_parameter_165: name = "parameter_165" + original_name = "layer_norm_537.w_0" shape = [768] dtype = "float32" min_val = float("0.593965") @@ -1826,6 +1992,7 @@ class Program_weight_tensor_parameter_165: class Program_weight_tensor_parameter_166: name = "parameter_166" + original_name = "linear_1561.b_0" shape = [768] dtype = "float32" min_val = float("-0.738263") @@ -1837,6 +2004,7 @@ class Program_weight_tensor_parameter_166: class Program_weight_tensor_parameter_167: name = "parameter_167" + original_name = "linear_1561.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-3.80926") @@ -1848,6 +2016,7 @@ class Program_weight_tensor_parameter_167: class Program_weight_tensor_parameter_168: name = "parameter_168" + original_name = "linear_1560.b_0" shape = [3072] dtype = "float32" min_val = float("-0.989553") @@ -1859,6 +2028,7 @@ class Program_weight_tensor_parameter_168: class Program_weight_tensor_parameter_169: name = "parameter_169" + original_name = "linear_1560.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.586842") @@ -1870,6 +2040,7 @@ class Program_weight_tensor_parameter_169: class Program_weight_tensor_parameter_170: name = "parameter_170" + original_name = "linear_1559.b_0" shape = [768] dtype = "float32" min_val = float("-0.245289") @@ -1881,6 +2052,7 @@ class Program_weight_tensor_parameter_170: class Program_weight_tensor_parameter_171: name = "parameter_171" + original_name = "linear_1559.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.36488") @@ -1892,6 +2064,7 @@ class Program_weight_tensor_parameter_171: class Program_weight_tensor_parameter_172: name = "parameter_172" + original_name = "linear_1558.b_0" shape = [768] dtype = "float32" min_val = float("-0.369347") @@ -1903,6 +2076,7 @@ class Program_weight_tensor_parameter_172: class Program_weight_tensor_parameter_173: name = "parameter_173" + original_name = "linear_1558.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.423042") @@ -1914,6 +2088,7 @@ class Program_weight_tensor_parameter_173: class Program_weight_tensor_parameter_174: name = "parameter_174" + original_name = "linear_1557.b_0" shape = [768] dtype = "float32" min_val = float("-3.24227") @@ -1925,6 +2100,7 @@ class Program_weight_tensor_parameter_174: class Program_weight_tensor_parameter_175: name = "parameter_175" + original_name = "linear_1557.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.612718") @@ -1936,6 +2112,7 @@ class Program_weight_tensor_parameter_175: class Program_weight_tensor_parameter_176: name = "parameter_176" + original_name = "linear_1556.b_0" shape = [768] dtype = "float32" min_val = float("-2.76648") @@ -1947,6 +2124,7 @@ class Program_weight_tensor_parameter_176: class Program_weight_tensor_parameter_177: name = "parameter_177" + original_name = "linear_1556.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.398133") @@ -1958,6 +2136,7 @@ class Program_weight_tensor_parameter_177: class Program_weight_tensor_parameter_178: name = "parameter_178" + original_name = "layer_norm_536.b_0" shape = [768] dtype = "float32" min_val = float("-0.623046") @@ -1969,6 +2148,7 @@ class Program_weight_tensor_parameter_178: class Program_weight_tensor_parameter_179: name = "parameter_179" + original_name = "layer_norm_536.w_0" shape = [768] dtype = "float32" min_val = float("0.100446") @@ -1980,6 +2160,7 @@ class Program_weight_tensor_parameter_179: class Program_weight_tensor_parameter_180: name = "parameter_180" + original_name = "layer_norm_535.b_0" shape = [768] dtype = "float32" min_val = float("-7.9432") @@ -1991,6 +2172,7 @@ class Program_weight_tensor_parameter_180: class Program_weight_tensor_parameter_181: name = "parameter_181" + original_name = "layer_norm_535.w_0" shape = [768] dtype = "float32" min_val = float("0.348659") @@ -2002,6 +2184,7 @@ class Program_weight_tensor_parameter_181: class Program_weight_tensor_parameter_182: name = "parameter_182" + original_name = "linear_1555.b_0" shape = [768] dtype = "float32" min_val = float("-0.613216") @@ -2013,6 +2196,7 @@ class Program_weight_tensor_parameter_182: class Program_weight_tensor_parameter_183: name = "parameter_183" + original_name = "linear_1555.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-3.54399") @@ -2024,6 +2208,7 @@ class Program_weight_tensor_parameter_183: class Program_weight_tensor_parameter_184: name = "parameter_184" + original_name = "linear_1554.b_0" shape = [3072] dtype = "float32" min_val = float("-1.50515") @@ -2035,6 +2220,7 @@ class Program_weight_tensor_parameter_184: class Program_weight_tensor_parameter_185: name = "parameter_185" + original_name = "linear_1554.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.602425") @@ -2046,6 +2232,7 @@ class Program_weight_tensor_parameter_185: class Program_weight_tensor_parameter_186: name = "parameter_186" + original_name = "linear_1553.b_0" shape = [768] dtype = "float32" min_val = float("-0.595094") @@ -2057,6 +2244,7 @@ class Program_weight_tensor_parameter_186: class Program_weight_tensor_parameter_187: name = "parameter_187" + original_name = "linear_1553.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.384004") @@ -2068,6 +2256,7 @@ class Program_weight_tensor_parameter_187: class Program_weight_tensor_parameter_188: name = "parameter_188" + original_name = "linear_1552.b_0" shape = [768] dtype = "float32" min_val = float("-1.11316") @@ -2079,6 +2268,7 @@ class Program_weight_tensor_parameter_188: class Program_weight_tensor_parameter_189: name = "parameter_189" + original_name = "linear_1552.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.511461") @@ -2090,6 +2280,7 @@ class Program_weight_tensor_parameter_189: class Program_weight_tensor_parameter_190: name = "parameter_190" + original_name = "linear_1551.b_0" shape = [768] dtype = "float32" min_val = float("-1.6471") @@ -2101,6 +2292,7 @@ class Program_weight_tensor_parameter_190: class Program_weight_tensor_parameter_191: name = "parameter_191" + original_name = "linear_1551.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.450814") @@ -2112,6 +2304,7 @@ class Program_weight_tensor_parameter_191: class Program_weight_tensor_parameter_192: name = "parameter_192" + original_name = "linear_1550.b_0" shape = [768] dtype = "float32" min_val = float("-3.02949") @@ -2123,6 +2316,7 @@ class Program_weight_tensor_parameter_192: class Program_weight_tensor_parameter_193: name = "parameter_193" + original_name = "linear_1550.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.336782") @@ -2134,6 +2328,7 @@ class Program_weight_tensor_parameter_193: class Program_weight_tensor_parameter_194: name = "parameter_194" + original_name = "layer_norm_534.b_0" shape = [768] dtype = "float32" min_val = float("-0.556785") @@ -2145,6 +2340,7 @@ class Program_weight_tensor_parameter_194: class Program_weight_tensor_parameter_195: name = "parameter_195" + original_name = "layer_norm_534.w_0" shape = [768] dtype = "float32" min_val = float("0.135306") @@ -2156,6 +2352,7 @@ class Program_weight_tensor_parameter_195: class Program_weight_tensor_parameter_196: name = "parameter_196" + original_name = "embedding_93.w_0" shape = [3, 768] dtype = "float32" min_val = float("-0.194673") @@ -2167,6 +2364,7 @@ class Program_weight_tensor_parameter_196: class Program_weight_tensor_parameter_197: name = "parameter_197" + original_name = "embedding_92.w_0" shape = [4, 768] dtype = "float32" min_val = float("-0.339255") @@ -2178,6 +2376,7 @@ class Program_weight_tensor_parameter_197: class Program_weight_tensor_parameter_198: name = "parameter_198" + original_name = "embedding_91.w_0" shape = [2048, 768] dtype = "float32" min_val = float("-0.205464") @@ -2189,6 +2388,7 @@ class Program_weight_tensor_parameter_198: class Program_weight_tensor_parameter_199: name = "parameter_199" + original_name = "embedding_90.w_0" shape = [40000, 768] dtype = "float32" min_val = float("-0.866318") diff --git a/paddle_samples/PaddleNLP/rocketqa-medium-cross-encoder/input_meta.py b/paddle_samples/PaddleNLP/rocketqa-medium-cross-encoder/input_meta.py index 4f25a05a9..055cd8978 100644 --- a/paddle_samples/PaddleNLP/rocketqa-medium-cross-encoder/input_meta.py +++ b/paddle_samples/PaddleNLP/rocketqa-medium-cross-encoder/input_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_data_0: name = "data_0" + original_name = "input_ids" shape = [1, 21] dtype = "int64" data = [ @@ -29,6 +30,7 @@ class Program_weight_tensor_data_0: class Program_weight_tensor_data_1: name = "data_1" + original_name = "token_type_ids" shape = [1, 21] dtype = "int64" data = [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0] diff --git a/paddle_samples/PaddleNLP/rocketqa-medium-cross-encoder/model.py b/paddle_samples/PaddleNLP/rocketqa-medium-cross-encoder/model.py index 1f9f7b0d1..d2b31efdb 100644 --- a/paddle_samples/PaddleNLP/rocketqa-medium-cross-encoder/model.py +++ b/paddle_samples/PaddleNLP/rocketqa-medium-cross-encoder/model.py @@ -170,10 +170,11 @@ def forward( # pd_op.embedding: (1x21x768xf32) <- (1x21xi64, 2048x768xf32) embedding_1 = paddle._C_ops.embedding(subtract_0, parameter_102, -1, False) - del parameter_102 + del parameter_102, subtract_0 # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 1x21x768xf32) add_0 = paddle._C_ops.add(embedding_0, embedding_1) + del embedding_0, embedding_1 # pd_op.embedding: (1x21x768xf32) <- (1x21xi64, 4x768xf32) embedding_2 = paddle._C_ops.embedding(data_1, parameter_101, -1, False) @@ -181,6 +182,7 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 1x21x768xf32) add_1 = paddle._C_ops.add(add_0, embedding_2) + del add_0, embedding_2 # pd_op.full: (1xf32) <- () full_4 = paddle._C_ops.full( @@ -193,10 +195,11 @@ def forward( # pd_op.embedding: (1x21x768xf32) <- (1x21xi64, 16x768xf32) embedding_3 = paddle._C_ops.embedding(scale_1, parameter_100, -1, False) - del parameter_100 + del parameter_100, scale_1 # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 1x21x768xf32) add_2 = paddle._C_ops.add(add_1, embedding_3) + del add_1, embedding_3 # pd_op.layer_norm: (1x21x768xf32, 1x21xf32, 1x21xf32) <- (1x21x768xf32, 768xf32, 768xf32) layer_norm_0, layer_norm_1, layer_norm_2 = (lambda x, f: f(x))( @@ -205,71 +208,17 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_98, parameter_99 + del add_2, parameter_98, parameter_99 # pd_op.full: (1xf32) <- () full_5 = paddle._C_ops.full( [1], float("0.1"), paddle.float32, paddle.core.CPUPlace() ) - # pd_op.assign: (1xf32) <- (1xf32) - assign_0 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_1 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_2 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_3 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_4 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_5 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_6 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_7 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_8 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_9 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_10 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_11 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_12 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_13 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_14 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_15 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_16 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_17 = full_5 - # pd_op.dropout: (1x21x768xf32, 1x21x768xui8) <- (1x21x768xf32, None, 1xf32) dropout_0, dropout_1 = (lambda x, f: f(x))( paddle._C_ops.dropout( - layer_norm_0, None, full_5, False, "upscale_in_train", 0, False + layer_norm_0, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -281,13 +230,14 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_3 = paddle._C_ops.add(matmul_0, parameter_96) - del parameter_96 + del matmul_0, parameter_96 # pd_op.full_int_array: (4xi64) <- () full_int_array_1 = [0, 0, 12, 64] # pd_op.reshape: (1x21x12x64xf32) <- (1x21x768xf32, 4xi64) reshape_0 = paddle._C_ops.reshape(add_3, full_int_array_1) + del add_3 # pd_op.transpose: (1x12x21x64xf32) <- (1x21x12x64xf32) transpose_0 = paddle._C_ops.transpose(reshape_0, [0, 2, 1, 3]) @@ -299,7 +249,7 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_4 = paddle._C_ops.add(matmul_1, parameter_94) - del parameter_94 + del matmul_1, parameter_94 # pd_op.matmul: (1x21x768xf32) <- (1x21x768xf32, 768x768xf32) matmul_2 = paddle._C_ops.matmul(dropout_0, parameter_93, False, False) @@ -307,10 +257,11 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_5 = paddle._C_ops.add(matmul_2, parameter_92) - del parameter_92 + del matmul_2, parameter_92 # pd_op.reshape: (1x21x12x64xf32) <- (1x21x768xf32, 4xi64) reshape_1 = paddle._C_ops.reshape(add_4, full_int_array_1) + del add_4 # pd_op.transpose: (1x12x21x64xf32) <- (1x21x12x64xf32) transpose_1 = paddle._C_ops.transpose(reshape_1, [0, 2, 1, 3]) @@ -318,6 +269,7 @@ def forward( # pd_op.reshape: (1x21x12x64xf32) <- (1x21x768xf32, 4xi64) reshape_2 = paddle._C_ops.reshape(add_5, full_int_array_1) + del add_5 # pd_op.transpose: (1x12x21x64xf32) <- (1x21x12x64xf32) transpose_2 = paddle._C_ops.transpose(reshape_2, [0, 2, 1, 3]) @@ -328,30 +280,17 @@ def forward( [1], float("0.125"), paddle.float32, paddle.core.CPUPlace() ) - # pd_op.assign: (1xf32) <- (1xf32) - assign_18 = full_6 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_19 = full_6 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_20 = full_6 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_21 = full_6 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_22 = full_6 - # pd_op.scale: (1x12x21x64xf32) <- (1x12x21x64xf32, 1xf32) scale_2 = paddle._C_ops.scale(transpose_0, full_6, float("0"), True) del transpose_0 # pd_op.matmul: (1x12x21x21xf32) <- (1x12x21x64xf32, 1x12x21x64xf32) matmul_3 = paddle._C_ops.matmul(scale_2, transpose_1, False, True) + del scale_2, transpose_1 # pd_op.add: (1x12x21x21xf32) <- (1x12x21x21xf32, 1x1x1x21xf32) add_6 = paddle._C_ops.add(matmul_3, unsqueeze_0) + del matmul_3 # pd_op.softmax: (1x12x21x21xf32) <- (1x12x21x21xf32) softmax_0 = paddle._C_ops.softmax(add_6, -1) @@ -360,13 +299,15 @@ def forward( # pd_op.dropout: (1x12x21x21xf32, 1x12x21x21xui8) <- (1x12x21x21xf32, None, 1xf32) dropout_2, dropout_3 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_0, None, full_5, False, "upscale_in_train", 0, False + softmax_0, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_0 # pd_op.matmul: (1x12x21x64xf32) <- (1x12x21x21xf32, 1x12x21x64xf32) matmul_4 = paddle._C_ops.matmul(dropout_2, transpose_2, False, False) + del dropout_2, transpose_2 # pd_op.transpose: (1x21x12x64xf32) <- (1x12x21x64xf32) transpose_3 = paddle._C_ops.transpose(matmul_4, [0, 2, 1, 3]) @@ -377,19 +318,20 @@ def forward( # pd_op.reshape: (1x21x768xf32) <- (1x21x12x64xf32, 3xi64) reshape_3 = paddle._C_ops.reshape(transpose_3, full_int_array_2) + del transpose_3 # pd_op.matmul: (1x21x768xf32) <- (1x21x768xf32, 768x768xf32) matmul_5 = paddle._C_ops.matmul(reshape_3, parameter_91, False, False) - del parameter_91 + del parameter_91, reshape_3 # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_7 = paddle._C_ops.add(matmul_5, parameter_90) - del parameter_90 + del matmul_5, parameter_90 # pd_op.dropout: (1x21x768xf32, 1x21x768xui8) <- (1x21x768xf32, None, 1xf32) dropout_4, dropout_5 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_7, None, full_5, False, "upscale_in_train", 0, False + add_7, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -397,6 +339,7 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 1x21x768xf32) add_8 = paddle._C_ops.add(dropout_0, dropout_4) + del dropout_0, dropout_4 # pd_op.layer_norm: (1x21x768xf32, 1x21xf32, 1x21xf32) <- (1x21x768xf32, 768xf32, 768xf32) layer_norm_3, layer_norm_4, layer_norm_5 = (lambda x, f: f(x))( @@ -405,7 +348,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_84, parameter_85 + del add_8, parameter_84, parameter_85 # pd_op.matmul: (1x21x3072xf32) <- (1x21x768xf32, 768x3072xf32) matmul_6 = paddle._C_ops.matmul(layer_norm_3, parameter_89, False, False) @@ -413,23 +356,24 @@ def forward( # pd_op.add: (1x21x3072xf32) <- (1x21x3072xf32, 3072xf32) add_9 = paddle._C_ops.add(matmul_6, parameter_88) - del parameter_88 + del matmul_6, parameter_88 # pd_op.gelu: (1x21x3072xf32) <- (1x21x3072xf32) gelu_0 = paddle._C_ops.gelu(add_9, False) + del add_9 # pd_op.matmul: (1x21x768xf32) <- (1x21x3072xf32, 3072x768xf32) matmul_7 = paddle._C_ops.matmul(gelu_0, parameter_87, False, False) - del parameter_87 + del gelu_0, parameter_87 # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_10 = paddle._C_ops.add(matmul_7, parameter_86) - del parameter_86 + del matmul_7, parameter_86 # pd_op.dropout: (1x21x768xf32, 1x21x768xui8) <- (1x21x768xf32, None, 1xf32) dropout_6, dropout_7 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_10, None, full_5, False, "upscale_in_train", 0, False + add_10, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -437,6 +381,7 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 1x21x768xf32) add_11 = paddle._C_ops.add(layer_norm_3, dropout_6) + del dropout_6, layer_norm_3 # pd_op.layer_norm: (1x21x768xf32, 1x21xf32, 1x21xf32) <- (1x21x768xf32, 768xf32, 768xf32) layer_norm_6, layer_norm_7, layer_norm_8 = (lambda x, f: f(x))( @@ -445,7 +390,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_82, parameter_83 + del add_11, parameter_82, parameter_83 # pd_op.matmul: (1x21x768xf32) <- (1x21x768xf32, 768x768xf32) matmul_8 = paddle._C_ops.matmul(layer_norm_6, parameter_81, False, False) @@ -453,10 +398,11 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_12 = paddle._C_ops.add(matmul_8, parameter_80) - del parameter_80 + del matmul_8, parameter_80 # pd_op.reshape: (1x21x12x64xf32) <- (1x21x768xf32, 4xi64) reshape_4 = paddle._C_ops.reshape(add_12, full_int_array_1) + del add_12 # pd_op.transpose: (1x12x21x64xf32) <- (1x21x12x64xf32) transpose_4 = paddle._C_ops.transpose(reshape_4, [0, 2, 1, 3]) @@ -468,7 +414,7 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_13 = paddle._C_ops.add(matmul_9, parameter_78) - del parameter_78 + del matmul_9, parameter_78 # pd_op.matmul: (1x21x768xf32) <- (1x21x768xf32, 768x768xf32) matmul_10 = paddle._C_ops.matmul(layer_norm_6, parameter_77, False, False) @@ -476,10 +422,11 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_14 = paddle._C_ops.add(matmul_10, parameter_76) - del parameter_76 + del matmul_10, parameter_76 # pd_op.reshape: (1x21x12x64xf32) <- (1x21x768xf32, 4xi64) reshape_5 = paddle._C_ops.reshape(add_13, full_int_array_1) + del add_13 # pd_op.transpose: (1x12x21x64xf32) <- (1x21x12x64xf32) transpose_5 = paddle._C_ops.transpose(reshape_5, [0, 2, 1, 3]) @@ -487,6 +434,7 @@ def forward( # pd_op.reshape: (1x21x12x64xf32) <- (1x21x768xf32, 4xi64) reshape_6 = paddle._C_ops.reshape(add_14, full_int_array_1) + del add_14 # pd_op.transpose: (1x12x21x64xf32) <- (1x21x12x64xf32) transpose_6 = paddle._C_ops.transpose(reshape_6, [0, 2, 1, 3]) @@ -498,9 +446,11 @@ def forward( # pd_op.matmul: (1x12x21x21xf32) <- (1x12x21x64xf32, 1x12x21x64xf32) matmul_11 = paddle._C_ops.matmul(scale_3, transpose_5, False, True) + del scale_3, transpose_5 # pd_op.add: (1x12x21x21xf32) <- (1x12x21x21xf32, 1x1x1x21xf32) add_15 = paddle._C_ops.add(matmul_11, unsqueeze_0) + del matmul_11 # pd_op.softmax: (1x12x21x21xf32) <- (1x12x21x21xf32) softmax_1 = paddle._C_ops.softmax(add_15, -1) @@ -509,13 +459,15 @@ def forward( # pd_op.dropout: (1x12x21x21xf32, 1x12x21x21xui8) <- (1x12x21x21xf32, None, 1xf32) dropout_8, dropout_9 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_1, None, full_5, False, "upscale_in_train", 0, False + softmax_1, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_1 # pd_op.matmul: (1x12x21x64xf32) <- (1x12x21x21xf32, 1x12x21x64xf32) matmul_12 = paddle._C_ops.matmul(dropout_8, transpose_6, False, False) + del dropout_8, transpose_6 # pd_op.transpose: (1x21x12x64xf32) <- (1x12x21x64xf32) transpose_7 = paddle._C_ops.transpose(matmul_12, [0, 2, 1, 3]) @@ -523,19 +475,20 @@ def forward( # pd_op.reshape: (1x21x768xf32) <- (1x21x12x64xf32, 3xi64) reshape_7 = paddle._C_ops.reshape(transpose_7, full_int_array_2) + del transpose_7 # pd_op.matmul: (1x21x768xf32) <- (1x21x768xf32, 768x768xf32) matmul_13 = paddle._C_ops.matmul(reshape_7, parameter_75, False, False) - del parameter_75 + del parameter_75, reshape_7 # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_16 = paddle._C_ops.add(matmul_13, parameter_74) - del parameter_74 + del matmul_13, parameter_74 # pd_op.dropout: (1x21x768xf32, 1x21x768xui8) <- (1x21x768xf32, None, 1xf32) dropout_10, dropout_11 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_16, None, full_5, False, "upscale_in_train", 0, False + add_16, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -543,6 +496,7 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 1x21x768xf32) add_17 = paddle._C_ops.add(layer_norm_6, dropout_10) + del dropout_10, layer_norm_6 # pd_op.layer_norm: (1x21x768xf32, 1x21xf32, 1x21xf32) <- (1x21x768xf32, 768xf32, 768xf32) layer_norm_9, layer_norm_10, layer_norm_11 = (lambda x, f: f(x))( @@ -551,7 +505,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_68, parameter_69 + del add_17, parameter_68, parameter_69 # pd_op.matmul: (1x21x3072xf32) <- (1x21x768xf32, 768x3072xf32) matmul_14 = paddle._C_ops.matmul(layer_norm_9, parameter_73, False, False) @@ -559,23 +513,24 @@ def forward( # pd_op.add: (1x21x3072xf32) <- (1x21x3072xf32, 3072xf32) add_18 = paddle._C_ops.add(matmul_14, parameter_72) - del parameter_72 + del matmul_14, parameter_72 # pd_op.gelu: (1x21x3072xf32) <- (1x21x3072xf32) gelu_1 = paddle._C_ops.gelu(add_18, False) + del add_18 # pd_op.matmul: (1x21x768xf32) <- (1x21x3072xf32, 3072x768xf32) matmul_15 = paddle._C_ops.matmul(gelu_1, parameter_71, False, False) - del parameter_71 + del gelu_1, parameter_71 # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_19 = paddle._C_ops.add(matmul_15, parameter_70) - del parameter_70 + del matmul_15, parameter_70 # pd_op.dropout: (1x21x768xf32, 1x21x768xui8) <- (1x21x768xf32, None, 1xf32) dropout_12, dropout_13 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_19, None, full_5, False, "upscale_in_train", 0, False + add_19, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -583,6 +538,7 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 1x21x768xf32) add_20 = paddle._C_ops.add(layer_norm_9, dropout_12) + del dropout_12, layer_norm_9 # pd_op.layer_norm: (1x21x768xf32, 1x21xf32, 1x21xf32) <- (1x21x768xf32, 768xf32, 768xf32) layer_norm_12, layer_norm_13, layer_norm_14 = (lambda x, f: f(x))( @@ -591,7 +547,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_66, parameter_67 + del add_20, parameter_66, parameter_67 # pd_op.matmul: (1x21x768xf32) <- (1x21x768xf32, 768x768xf32) matmul_16 = paddle._C_ops.matmul(layer_norm_12, parameter_65, False, False) @@ -599,10 +555,11 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_21 = paddle._C_ops.add(matmul_16, parameter_64) - del parameter_64 + del matmul_16, parameter_64 # pd_op.reshape: (1x21x12x64xf32) <- (1x21x768xf32, 4xi64) reshape_8 = paddle._C_ops.reshape(add_21, full_int_array_1) + del add_21 # pd_op.transpose: (1x12x21x64xf32) <- (1x21x12x64xf32) transpose_8 = paddle._C_ops.transpose(reshape_8, [0, 2, 1, 3]) @@ -614,7 +571,7 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_22 = paddle._C_ops.add(matmul_17, parameter_62) - del parameter_62 + del matmul_17, parameter_62 # pd_op.matmul: (1x21x768xf32) <- (1x21x768xf32, 768x768xf32) matmul_18 = paddle._C_ops.matmul(layer_norm_12, parameter_61, False, False) @@ -622,10 +579,11 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_23 = paddle._C_ops.add(matmul_18, parameter_60) - del parameter_60 + del matmul_18, parameter_60 # pd_op.reshape: (1x21x12x64xf32) <- (1x21x768xf32, 4xi64) reshape_9 = paddle._C_ops.reshape(add_22, full_int_array_1) + del add_22 # pd_op.transpose: (1x12x21x64xf32) <- (1x21x12x64xf32) transpose_9 = paddle._C_ops.transpose(reshape_9, [0, 2, 1, 3]) @@ -633,6 +591,7 @@ def forward( # pd_op.reshape: (1x21x12x64xf32) <- (1x21x768xf32, 4xi64) reshape_10 = paddle._C_ops.reshape(add_23, full_int_array_1) + del add_23 # pd_op.transpose: (1x12x21x64xf32) <- (1x21x12x64xf32) transpose_10 = paddle._C_ops.transpose(reshape_10, [0, 2, 1, 3]) @@ -644,9 +603,11 @@ def forward( # pd_op.matmul: (1x12x21x21xf32) <- (1x12x21x64xf32, 1x12x21x64xf32) matmul_19 = paddle._C_ops.matmul(scale_4, transpose_9, False, True) + del scale_4, transpose_9 # pd_op.add: (1x12x21x21xf32) <- (1x12x21x21xf32, 1x1x1x21xf32) add_24 = paddle._C_ops.add(matmul_19, unsqueeze_0) + del matmul_19 # pd_op.softmax: (1x12x21x21xf32) <- (1x12x21x21xf32) softmax_2 = paddle._C_ops.softmax(add_24, -1) @@ -655,13 +616,15 @@ def forward( # pd_op.dropout: (1x12x21x21xf32, 1x12x21x21xui8) <- (1x12x21x21xf32, None, 1xf32) dropout_14, dropout_15 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_2, None, full_5, False, "upscale_in_train", 0, False + softmax_2, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_2 # pd_op.matmul: (1x12x21x64xf32) <- (1x12x21x21xf32, 1x12x21x64xf32) matmul_20 = paddle._C_ops.matmul(dropout_14, transpose_10, False, False) + del dropout_14, transpose_10 # pd_op.transpose: (1x21x12x64xf32) <- (1x12x21x64xf32) transpose_11 = paddle._C_ops.transpose(matmul_20, [0, 2, 1, 3]) @@ -669,19 +632,20 @@ def forward( # pd_op.reshape: (1x21x768xf32) <- (1x21x12x64xf32, 3xi64) reshape_11 = paddle._C_ops.reshape(transpose_11, full_int_array_2) + del transpose_11 # pd_op.matmul: (1x21x768xf32) <- (1x21x768xf32, 768x768xf32) matmul_21 = paddle._C_ops.matmul(reshape_11, parameter_59, False, False) - del parameter_59 + del parameter_59, reshape_11 # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_25 = paddle._C_ops.add(matmul_21, parameter_58) - del parameter_58 + del matmul_21, parameter_58 # pd_op.dropout: (1x21x768xf32, 1x21x768xui8) <- (1x21x768xf32, None, 1xf32) dropout_16, dropout_17 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_25, None, full_5, False, "upscale_in_train", 0, False + add_25, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -689,6 +653,7 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 1x21x768xf32) add_26 = paddle._C_ops.add(layer_norm_12, dropout_16) + del dropout_16, layer_norm_12 # pd_op.layer_norm: (1x21x768xf32, 1x21xf32, 1x21xf32) <- (1x21x768xf32, 768xf32, 768xf32) layer_norm_15, layer_norm_16, layer_norm_17 = (lambda x, f: f(x))( @@ -697,7 +662,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_52, parameter_53 + del add_26, parameter_52, parameter_53 # pd_op.matmul: (1x21x3072xf32) <- (1x21x768xf32, 768x3072xf32) matmul_22 = paddle._C_ops.matmul(layer_norm_15, parameter_57, False, False) @@ -705,23 +670,24 @@ def forward( # pd_op.add: (1x21x3072xf32) <- (1x21x3072xf32, 3072xf32) add_27 = paddle._C_ops.add(matmul_22, parameter_56) - del parameter_56 + del matmul_22, parameter_56 # pd_op.gelu: (1x21x3072xf32) <- (1x21x3072xf32) gelu_2 = paddle._C_ops.gelu(add_27, False) + del add_27 # pd_op.matmul: (1x21x768xf32) <- (1x21x3072xf32, 3072x768xf32) matmul_23 = paddle._C_ops.matmul(gelu_2, parameter_55, False, False) - del parameter_55 + del gelu_2, parameter_55 # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_28 = paddle._C_ops.add(matmul_23, parameter_54) - del parameter_54 + del matmul_23, parameter_54 # pd_op.dropout: (1x21x768xf32, 1x21x768xui8) <- (1x21x768xf32, None, 1xf32) dropout_18, dropout_19 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_28, None, full_5, False, "upscale_in_train", 0, False + add_28, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -729,6 +695,7 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 1x21x768xf32) add_29 = paddle._C_ops.add(layer_norm_15, dropout_18) + del dropout_18, layer_norm_15 # pd_op.layer_norm: (1x21x768xf32, 1x21xf32, 1x21xf32) <- (1x21x768xf32, 768xf32, 768xf32) layer_norm_18, layer_norm_19, layer_norm_20 = (lambda x, f: f(x))( @@ -737,7 +704,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_50, parameter_51 + del add_29, parameter_50, parameter_51 # pd_op.matmul: (1x21x768xf32) <- (1x21x768xf32, 768x768xf32) matmul_24 = paddle._C_ops.matmul(layer_norm_18, parameter_49, False, False) @@ -745,10 +712,11 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_30 = paddle._C_ops.add(matmul_24, parameter_48) - del parameter_48 + del matmul_24, parameter_48 # pd_op.reshape: (1x21x12x64xf32) <- (1x21x768xf32, 4xi64) reshape_12 = paddle._C_ops.reshape(add_30, full_int_array_1) + del add_30 # pd_op.transpose: (1x12x21x64xf32) <- (1x21x12x64xf32) transpose_12 = paddle._C_ops.transpose(reshape_12, [0, 2, 1, 3]) @@ -760,7 +728,7 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_31 = paddle._C_ops.add(matmul_25, parameter_46) - del parameter_46 + del matmul_25, parameter_46 # pd_op.matmul: (1x21x768xf32) <- (1x21x768xf32, 768x768xf32) matmul_26 = paddle._C_ops.matmul(layer_norm_18, parameter_45, False, False) @@ -768,10 +736,11 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_32 = paddle._C_ops.add(matmul_26, parameter_44) - del parameter_44 + del matmul_26, parameter_44 # pd_op.reshape: (1x21x12x64xf32) <- (1x21x768xf32, 4xi64) reshape_13 = paddle._C_ops.reshape(add_31, full_int_array_1) + del add_31 # pd_op.transpose: (1x12x21x64xf32) <- (1x21x12x64xf32) transpose_13 = paddle._C_ops.transpose(reshape_13, [0, 2, 1, 3]) @@ -779,6 +748,7 @@ def forward( # pd_op.reshape: (1x21x12x64xf32) <- (1x21x768xf32, 4xi64) reshape_14 = paddle._C_ops.reshape(add_32, full_int_array_1) + del add_32 # pd_op.transpose: (1x12x21x64xf32) <- (1x21x12x64xf32) transpose_14 = paddle._C_ops.transpose(reshape_14, [0, 2, 1, 3]) @@ -790,9 +760,11 @@ def forward( # pd_op.matmul: (1x12x21x21xf32) <- (1x12x21x64xf32, 1x12x21x64xf32) matmul_27 = paddle._C_ops.matmul(scale_5, transpose_13, False, True) + del scale_5, transpose_13 # pd_op.add: (1x12x21x21xf32) <- (1x12x21x21xf32, 1x1x1x21xf32) add_33 = paddle._C_ops.add(matmul_27, unsqueeze_0) + del matmul_27 # pd_op.softmax: (1x12x21x21xf32) <- (1x12x21x21xf32) softmax_3 = paddle._C_ops.softmax(add_33, -1) @@ -801,13 +773,15 @@ def forward( # pd_op.dropout: (1x12x21x21xf32, 1x12x21x21xui8) <- (1x12x21x21xf32, None, 1xf32) dropout_20, dropout_21 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_3, None, full_5, False, "upscale_in_train", 0, False + softmax_3, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_3 # pd_op.matmul: (1x12x21x64xf32) <- (1x12x21x21xf32, 1x12x21x64xf32) matmul_28 = paddle._C_ops.matmul(dropout_20, transpose_14, False, False) + del dropout_20, transpose_14 # pd_op.transpose: (1x21x12x64xf32) <- (1x12x21x64xf32) transpose_15 = paddle._C_ops.transpose(matmul_28, [0, 2, 1, 3]) @@ -815,19 +789,20 @@ def forward( # pd_op.reshape: (1x21x768xf32) <- (1x21x12x64xf32, 3xi64) reshape_15 = paddle._C_ops.reshape(transpose_15, full_int_array_2) + del transpose_15 # pd_op.matmul: (1x21x768xf32) <- (1x21x768xf32, 768x768xf32) matmul_29 = paddle._C_ops.matmul(reshape_15, parameter_43, False, False) - del parameter_43 + del parameter_43, reshape_15 # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_34 = paddle._C_ops.add(matmul_29, parameter_42) - del parameter_42 + del matmul_29, parameter_42 # pd_op.dropout: (1x21x768xf32, 1x21x768xui8) <- (1x21x768xf32, None, 1xf32) dropout_22, dropout_23 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_34, None, full_5, False, "upscale_in_train", 0, False + add_34, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -835,6 +810,7 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 1x21x768xf32) add_35 = paddle._C_ops.add(layer_norm_18, dropout_22) + del dropout_22, layer_norm_18 # pd_op.layer_norm: (1x21x768xf32, 1x21xf32, 1x21xf32) <- (1x21x768xf32, 768xf32, 768xf32) layer_norm_21, layer_norm_22, layer_norm_23 = (lambda x, f: f(x))( @@ -843,7 +819,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_36, parameter_37 + del add_35, parameter_36, parameter_37 # pd_op.matmul: (1x21x3072xf32) <- (1x21x768xf32, 768x3072xf32) matmul_30 = paddle._C_ops.matmul(layer_norm_21, parameter_41, False, False) @@ -851,23 +827,24 @@ def forward( # pd_op.add: (1x21x3072xf32) <- (1x21x3072xf32, 3072xf32) add_36 = paddle._C_ops.add(matmul_30, parameter_40) - del parameter_40 + del matmul_30, parameter_40 # pd_op.gelu: (1x21x3072xf32) <- (1x21x3072xf32) gelu_3 = paddle._C_ops.gelu(add_36, False) + del add_36 # pd_op.matmul: (1x21x768xf32) <- (1x21x3072xf32, 3072x768xf32) matmul_31 = paddle._C_ops.matmul(gelu_3, parameter_39, False, False) - del parameter_39 + del gelu_3, parameter_39 # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_37 = paddle._C_ops.add(matmul_31, parameter_38) - del parameter_38 + del matmul_31, parameter_38 # pd_op.dropout: (1x21x768xf32, 1x21x768xui8) <- (1x21x768xf32, None, 1xf32) dropout_24, dropout_25 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_37, None, full_5, False, "upscale_in_train", 0, False + add_37, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -875,6 +852,7 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 1x21x768xf32) add_38 = paddle._C_ops.add(layer_norm_21, dropout_24) + del dropout_24, layer_norm_21 # pd_op.layer_norm: (1x21x768xf32, 1x21xf32, 1x21xf32) <- (1x21x768xf32, 768xf32, 768xf32) layer_norm_24, layer_norm_25, layer_norm_26 = (lambda x, f: f(x))( @@ -883,7 +861,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_34, parameter_35 + del add_38, parameter_34, parameter_35 # pd_op.matmul: (1x21x768xf32) <- (1x21x768xf32, 768x768xf32) matmul_32 = paddle._C_ops.matmul(layer_norm_24, parameter_33, False, False) @@ -891,10 +869,11 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_39 = paddle._C_ops.add(matmul_32, parameter_32) - del parameter_32 + del matmul_32, parameter_32 # pd_op.reshape: (1x21x12x64xf32) <- (1x21x768xf32, 4xi64) reshape_16 = paddle._C_ops.reshape(add_39, full_int_array_1) + del add_39 # pd_op.transpose: (1x12x21x64xf32) <- (1x21x12x64xf32) transpose_16 = paddle._C_ops.transpose(reshape_16, [0, 2, 1, 3]) @@ -906,7 +885,7 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_40 = paddle._C_ops.add(matmul_33, parameter_30) - del parameter_30 + del matmul_33, parameter_30 # pd_op.matmul: (1x21x768xf32) <- (1x21x768xf32, 768x768xf32) matmul_34 = paddle._C_ops.matmul(layer_norm_24, parameter_29, False, False) @@ -914,10 +893,11 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_41 = paddle._C_ops.add(matmul_34, parameter_28) - del parameter_28 + del matmul_34, parameter_28 # pd_op.reshape: (1x21x12x64xf32) <- (1x21x768xf32, 4xi64) reshape_17 = paddle._C_ops.reshape(add_40, full_int_array_1) + del add_40 # pd_op.transpose: (1x12x21x64xf32) <- (1x21x12x64xf32) transpose_17 = paddle._C_ops.transpose(reshape_17, [0, 2, 1, 3]) @@ -925,6 +905,7 @@ def forward( # pd_op.reshape: (1x21x12x64xf32) <- (1x21x768xf32, 4xi64) reshape_18 = paddle._C_ops.reshape(add_41, full_int_array_1) + del add_41 # pd_op.transpose: (1x12x21x64xf32) <- (1x21x12x64xf32) transpose_18 = paddle._C_ops.transpose(reshape_18, [0, 2, 1, 3]) @@ -936,9 +917,11 @@ def forward( # pd_op.matmul: (1x12x21x21xf32) <- (1x12x21x64xf32, 1x12x21x64xf32) matmul_35 = paddle._C_ops.matmul(scale_6, transpose_17, False, True) + del scale_6, transpose_17 # pd_op.add: (1x12x21x21xf32) <- (1x12x21x21xf32, 1x1x1x21xf32) add_42 = paddle._C_ops.add(matmul_35, unsqueeze_0) + del matmul_35 # pd_op.softmax: (1x12x21x21xf32) <- (1x12x21x21xf32) softmax_4 = paddle._C_ops.softmax(add_42, -1) @@ -947,13 +930,15 @@ def forward( # pd_op.dropout: (1x12x21x21xf32, 1x12x21x21xui8) <- (1x12x21x21xf32, None, 1xf32) dropout_26, dropout_27 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_4, None, full_5, False, "upscale_in_train", 0, False + softmax_4, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_4 # pd_op.matmul: (1x12x21x64xf32) <- (1x12x21x21xf32, 1x12x21x64xf32) matmul_36 = paddle._C_ops.matmul(dropout_26, transpose_18, False, False) + del dropout_26, transpose_18 # pd_op.transpose: (1x21x12x64xf32) <- (1x12x21x64xf32) transpose_19 = paddle._C_ops.transpose(matmul_36, [0, 2, 1, 3]) @@ -961,19 +946,20 @@ def forward( # pd_op.reshape: (1x21x768xf32) <- (1x21x12x64xf32, 3xi64) reshape_19 = paddle._C_ops.reshape(transpose_19, full_int_array_2) + del transpose_19 # pd_op.matmul: (1x21x768xf32) <- (1x21x768xf32, 768x768xf32) matmul_37 = paddle._C_ops.matmul(reshape_19, parameter_27, False, False) - del parameter_27 + del parameter_27, reshape_19 # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_43 = paddle._C_ops.add(matmul_37, parameter_26) - del parameter_26 + del matmul_37, parameter_26 # pd_op.dropout: (1x21x768xf32, 1x21x768xui8) <- (1x21x768xf32, None, 1xf32) dropout_28, dropout_29 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_43, None, full_5, False, "upscale_in_train", 0, False + add_43, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -981,6 +967,7 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 1x21x768xf32) add_44 = paddle._C_ops.add(layer_norm_24, dropout_28) + del dropout_28, layer_norm_24 # pd_op.layer_norm: (1x21x768xf32, 1x21xf32, 1x21xf32) <- (1x21x768xf32, 768xf32, 768xf32) layer_norm_27, layer_norm_28, layer_norm_29 = (lambda x, f: f(x))( @@ -989,7 +976,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_20, parameter_21 + del add_44, parameter_20, parameter_21 # pd_op.matmul: (1x21x3072xf32) <- (1x21x768xf32, 768x3072xf32) matmul_38 = paddle._C_ops.matmul(layer_norm_27, parameter_25, False, False) @@ -997,23 +984,24 @@ def forward( # pd_op.add: (1x21x3072xf32) <- (1x21x3072xf32, 3072xf32) add_45 = paddle._C_ops.add(matmul_38, parameter_24) - del parameter_24 + del matmul_38, parameter_24 # pd_op.gelu: (1x21x3072xf32) <- (1x21x3072xf32) gelu_4 = paddle._C_ops.gelu(add_45, False) + del add_45 # pd_op.matmul: (1x21x768xf32) <- (1x21x3072xf32, 3072x768xf32) matmul_39 = paddle._C_ops.matmul(gelu_4, parameter_23, False, False) - del parameter_23 + del gelu_4, parameter_23 # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_46 = paddle._C_ops.add(matmul_39, parameter_22) - del parameter_22 + del matmul_39, parameter_22 # pd_op.dropout: (1x21x768xf32, 1x21x768xui8) <- (1x21x768xf32, None, 1xf32) dropout_30, dropout_31 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_46, None, full_5, False, "upscale_in_train", 0, False + add_46, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1021,6 +1009,7 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 1x21x768xf32) add_47 = paddle._C_ops.add(layer_norm_27, dropout_30) + del dropout_30, layer_norm_27 # pd_op.layer_norm: (1x21x768xf32, 1x21xf32, 1x21xf32) <- (1x21x768xf32, 768xf32, 768xf32) layer_norm_30, layer_norm_31, layer_norm_32 = (lambda x, f: f(x))( @@ -1029,7 +1018,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_18, parameter_19 + del add_47, parameter_18, parameter_19 # pd_op.matmul: (1x21x768xf32) <- (1x21x768xf32, 768x768xf32) matmul_40 = paddle._C_ops.matmul(layer_norm_30, parameter_17, False, False) @@ -1037,10 +1026,11 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_48 = paddle._C_ops.add(matmul_40, parameter_16) - del parameter_16 + del matmul_40, parameter_16 # pd_op.reshape: (1x21x12x64xf32) <- (1x21x768xf32, 4xi64) reshape_20 = paddle._C_ops.reshape(add_48, full_int_array_1) + del add_48 # pd_op.transpose: (1x12x21x64xf32) <- (1x21x12x64xf32) transpose_20 = paddle._C_ops.transpose(reshape_20, [0, 2, 1, 3]) @@ -1052,7 +1042,7 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_49 = paddle._C_ops.add(matmul_41, parameter_14) - del parameter_14 + del matmul_41, parameter_14 # pd_op.matmul: (1x21x768xf32) <- (1x21x768xf32, 768x768xf32) matmul_42 = paddle._C_ops.matmul(layer_norm_30, parameter_13, False, False) @@ -1060,10 +1050,11 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_50 = paddle._C_ops.add(matmul_42, parameter_12) - del parameter_12 + del matmul_42, parameter_12 # pd_op.reshape: (1x21x12x64xf32) <- (1x21x768xf32, 4xi64) reshape_21 = paddle._C_ops.reshape(add_49, full_int_array_1) + del add_49 # pd_op.transpose: (1x12x21x64xf32) <- (1x21x12x64xf32) transpose_21 = paddle._C_ops.transpose(reshape_21, [0, 2, 1, 3]) @@ -1071,7 +1062,7 @@ def forward( # pd_op.reshape: (1x21x12x64xf32) <- (1x21x768xf32, 4xi64) reshape_22 = paddle._C_ops.reshape(add_50, full_int_array_1) - del full_int_array_1 + del add_50, full_int_array_1 # pd_op.transpose: (1x12x21x64xf32) <- (1x21x12x64xf32) transpose_22 = paddle._C_ops.transpose(reshape_22, [0, 2, 1, 3]) @@ -1079,13 +1070,15 @@ def forward( # pd_op.scale: (1x12x21x64xf32) <- (1x12x21x64xf32, 1xf32) scale_7 = paddle._C_ops.scale(transpose_20, full_6, float("0"), True) - del transpose_20 + del full_6, transpose_20 # pd_op.matmul: (1x12x21x21xf32) <- (1x12x21x64xf32, 1x12x21x64xf32) matmul_43 = paddle._C_ops.matmul(scale_7, transpose_21, False, True) + del scale_7, transpose_21 # pd_op.add: (1x12x21x21xf32) <- (1x12x21x21xf32, 1x1x1x21xf32) add_51 = paddle._C_ops.add(matmul_43, unsqueeze_0) + del matmul_43, unsqueeze_0 # pd_op.softmax: (1x12x21x21xf32) <- (1x12x21x21xf32) softmax_5 = paddle._C_ops.softmax(add_51, -1) @@ -1094,13 +1087,15 @@ def forward( # pd_op.dropout: (1x12x21x21xf32, 1x12x21x21xui8) <- (1x12x21x21xf32, None, 1xf32) dropout_32, dropout_33 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_5, None, full_5, False, "upscale_in_train", 0, False + softmax_5, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_5 # pd_op.matmul: (1x12x21x64xf32) <- (1x12x21x21xf32, 1x12x21x64xf32) matmul_44 = paddle._C_ops.matmul(dropout_32, transpose_22, False, False) + del dropout_32, transpose_22 # pd_op.transpose: (1x21x12x64xf32) <- (1x12x21x64xf32) transpose_23 = paddle._C_ops.transpose(matmul_44, [0, 2, 1, 3]) @@ -1108,20 +1103,20 @@ def forward( # pd_op.reshape: (1x21x768xf32) <- (1x21x12x64xf32, 3xi64) reshape_23 = paddle._C_ops.reshape(transpose_23, full_int_array_2) - del full_int_array_2 + del full_int_array_2, transpose_23 # pd_op.matmul: (1x21x768xf32) <- (1x21x768xf32, 768x768xf32) matmul_45 = paddle._C_ops.matmul(reshape_23, parameter_11, False, False) - del parameter_11 + del parameter_11, reshape_23 # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_52 = paddle._C_ops.add(matmul_45, parameter_10) - del parameter_10 + del matmul_45, parameter_10 # pd_op.dropout: (1x21x768xf32, 1x21x768xui8) <- (1x21x768xf32, None, 1xf32) dropout_34, dropout_35 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_52, None, full_5, False, "upscale_in_train", 0, False + add_52, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1129,6 +1124,7 @@ def forward( # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 1x21x768xf32) add_53 = paddle._C_ops.add(layer_norm_30, dropout_34) + del dropout_34, layer_norm_30 # pd_op.layer_norm: (1x21x768xf32, 1x21xf32, 1x21xf32) <- (1x21x768xf32, 768xf32, 768xf32) layer_norm_33, layer_norm_34, layer_norm_35 = (lambda x, f: f(x))( @@ -1137,7 +1133,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_4, parameter_5 + del add_53, parameter_4, parameter_5 # pd_op.matmul: (1x21x3072xf32) <- (1x21x768xf32, 768x3072xf32) matmul_46 = paddle._C_ops.matmul(layer_norm_33, parameter_9, False, False) @@ -1145,30 +1141,32 @@ def forward( # pd_op.add: (1x21x3072xf32) <- (1x21x3072xf32, 3072xf32) add_54 = paddle._C_ops.add(matmul_46, parameter_8) - del parameter_8 + del matmul_46, parameter_8 # pd_op.gelu: (1x21x3072xf32) <- (1x21x3072xf32) gelu_5 = paddle._C_ops.gelu(add_54, False) + del add_54 # pd_op.matmul: (1x21x768xf32) <- (1x21x3072xf32, 3072x768xf32) matmul_47 = paddle._C_ops.matmul(gelu_5, parameter_7, False, False) - del parameter_7 + del gelu_5, parameter_7 # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 768xf32) add_55 = paddle._C_ops.add(matmul_47, parameter_6) - del parameter_6 + del matmul_47, parameter_6 # pd_op.dropout: (1x21x768xf32, 1x21x768xui8) <- (1x21x768xf32, None, 1xf32) dropout_36, dropout_37 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_55, None, full_5, False, "upscale_in_train", 0, False + add_55, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) - del add_55 + del add_55, full_5 # pd_op.add: (1x21x768xf32) <- (1x21x768xf32, 1x21x768xf32) add_56 = paddle._C_ops.add(layer_norm_33, dropout_36) + del dropout_36, layer_norm_33 # pd_op.layer_norm: (1x21x768xf32, 1x21xf32, 1x21xf32) <- (1x21x768xf32, 768xf32, 768xf32) layer_norm_36, layer_norm_37, layer_norm_38 = (lambda x, f: f(x))( @@ -1177,7 +1175,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_2, parameter_3 + del add_56, parameter_2, parameter_3 # pd_op.full_int_array: (1xi64) <- () full_int_array_3 = [0] @@ -1189,254 +1187,18 @@ def forward( slice_0 = paddle._C_ops.slice( layer_norm_36, [1], full_int_array_3, full_int_array_4, [1], [1] ) + del full_int_array_3, full_int_array_4, layer_norm_36 # pd_op.matmul: (1x768xf32) <- (1x768xf32, 768x768xf32) matmul_48 = paddle._C_ops.matmul(slice_0, parameter_1, False, False) - del parameter_1 + del parameter_1, slice_0 # pd_op.add: (1x768xf32) <- (1x768xf32, 768xf32) add_57 = paddle._C_ops.add(matmul_48, parameter_0) - del parameter_0 + del matmul_48, parameter_0 # pd_op.tanh: (1x768xf32) <- (1x768xf32) tanh_0 = paddle._C_ops.tanh(add_57) - del ( - add_0, - add_1, - add_11, - add_12, - add_13, - add_14, - add_17, - add_18, - add_2, - add_20, - add_21, - add_22, - add_23, - add_26, - add_27, - add_29, - add_3, - add_30, - add_31, - add_32, - add_35, - add_36, - add_38, - add_39, - add_4, - add_40, - add_41, - add_44, - add_45, - add_47, - add_48, - add_49, - add_5, - add_50, - add_53, - add_54, - add_56, - add_57, - add_8, - add_9, - assign_0, - assign_1, - assign_10, - assign_11, - assign_12, - assign_13, - assign_14, - assign_15, - assign_16, - assign_17, - assign_18, - assign_19, - assign_2, - assign_20, - assign_21, - assign_22, - assign_3, - assign_4, - assign_5, - assign_6, - assign_7, - assign_8, - assign_9, - dropout_0, - dropout_1, - dropout_10, - dropout_11, - dropout_12, - dropout_13, - dropout_14, - dropout_15, - dropout_16, - dropout_17, - dropout_18, - dropout_19, - dropout_2, - dropout_20, - dropout_21, - dropout_22, - dropout_23, - dropout_24, - dropout_25, - dropout_26, - dropout_27, - dropout_28, - dropout_29, - dropout_3, - dropout_30, - dropout_31, - dropout_32, - dropout_33, - dropout_34, - dropout_35, - dropout_36, - dropout_37, - dropout_4, - dropout_5, - dropout_6, - dropout_7, - dropout_8, - dropout_9, - embedding_0, - embedding_1, - embedding_2, - embedding_3, - full_5, - full_6, - full_int_array_3, - full_int_array_4, - gelu_0, - gelu_1, - gelu_2, - gelu_3, - gelu_4, - gelu_5, - layer_norm_1, - layer_norm_10, - layer_norm_11, - layer_norm_12, - layer_norm_13, - layer_norm_14, - layer_norm_15, - layer_norm_16, - layer_norm_17, - layer_norm_18, - layer_norm_19, - layer_norm_2, - layer_norm_20, - layer_norm_21, - layer_norm_22, - layer_norm_23, - layer_norm_24, - layer_norm_25, - layer_norm_26, - layer_norm_27, - layer_norm_28, - layer_norm_29, - layer_norm_3, - layer_norm_30, - layer_norm_31, - layer_norm_32, - layer_norm_33, - layer_norm_34, - layer_norm_35, - layer_norm_36, - layer_norm_37, - layer_norm_38, - layer_norm_4, - layer_norm_5, - layer_norm_6, - layer_norm_7, - layer_norm_8, - layer_norm_9, - matmul_0, - matmul_1, - matmul_10, - matmul_11, - matmul_13, - matmul_14, - matmul_15, - matmul_16, - matmul_17, - matmul_18, - matmul_19, - matmul_2, - matmul_21, - matmul_22, - matmul_23, - matmul_24, - matmul_25, - matmul_26, - matmul_27, - matmul_29, - matmul_3, - matmul_30, - matmul_31, - matmul_32, - matmul_33, - matmul_34, - matmul_35, - matmul_37, - matmul_38, - matmul_39, - matmul_40, - matmul_41, - matmul_42, - matmul_43, - matmul_45, - matmul_46, - matmul_47, - matmul_48, - matmul_5, - matmul_6, - matmul_7, - matmul_8, - matmul_9, - reshape_11, - reshape_15, - reshape_19, - reshape_23, - reshape_3, - reshape_7, - scale_1, - scale_2, - scale_3, - scale_4, - scale_5, - scale_6, - scale_7, - slice_0, - softmax_0, - softmax_1, - softmax_2, - softmax_3, - softmax_4, - softmax_5, - subtract_0, - transpose_1, - transpose_10, - transpose_11, - transpose_13, - transpose_14, - transpose_15, - transpose_17, - transpose_18, - transpose_19, - transpose_2, - transpose_21, - transpose_22, - transpose_23, - transpose_3, - transpose_5, - transpose_6, - transpose_7, - transpose_9, - unsqueeze_0, - ) + del add_57 return tanh_0 diff --git a/paddle_samples/PaddleNLP/rocketqa-medium-cross-encoder/weight_meta.py b/paddle_samples/PaddleNLP/rocketqa-medium-cross-encoder/weight_meta.py index 50c224e56..db2c4dbef 100644 --- a/paddle_samples/PaddleNLP/rocketqa-medium-cross-encoder/weight_meta.py +++ b/paddle_samples/PaddleNLP/rocketqa-medium-cross-encoder/weight_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_parameter_0: name = "parameter_0" + original_name = "linear_1659.b_0" shape = [768] dtype = "float32" min_val = float("-0.00423639") @@ -11,6 +12,7 @@ class Program_weight_tensor_parameter_0: class Program_weight_tensor_parameter_1: name = "parameter_1" + original_name = "linear_1659.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.0925458") @@ -22,6 +24,7 @@ class Program_weight_tensor_parameter_1: class Program_weight_tensor_parameter_2: name = "parameter_2" + original_name = "layer_norm_571.b_0" shape = [768] dtype = "float32" min_val = float("-0.615221") @@ -33,6 +36,7 @@ class Program_weight_tensor_parameter_2: class Program_weight_tensor_parameter_3: name = "parameter_3" + original_name = "layer_norm_571.w_0" shape = [768] dtype = "float32" min_val = float("0.444052") @@ -44,6 +48,7 @@ class Program_weight_tensor_parameter_3: class Program_weight_tensor_parameter_4: name = "parameter_4" + original_name = "layer_norm_570.b_0" shape = [768] dtype = "float32" min_val = float("-0.789388") @@ -55,6 +60,7 @@ class Program_weight_tensor_parameter_4: class Program_weight_tensor_parameter_5: name = "parameter_5" + original_name = "layer_norm_570.w_0" shape = [768] dtype = "float32" min_val = float("0.219723") @@ -66,6 +72,7 @@ class Program_weight_tensor_parameter_5: class Program_weight_tensor_parameter_6: name = "parameter_6" + original_name = "linear_1658.b_0" shape = [768] dtype = "float32" min_val = float("-0.21841") @@ -77,6 +84,7 @@ class Program_weight_tensor_parameter_6: class Program_weight_tensor_parameter_7: name = "parameter_7" + original_name = "linear_1658.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-1.78168") @@ -88,6 +96,7 @@ class Program_weight_tensor_parameter_7: class Program_weight_tensor_parameter_8: name = "parameter_8" + original_name = "linear_1657.b_0" shape = [3072] dtype = "float32" min_val = float("-0.383817") @@ -99,6 +108,7 @@ class Program_weight_tensor_parameter_8: class Program_weight_tensor_parameter_9: name = "parameter_9" + original_name = "linear_1657.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.424876") @@ -110,6 +120,7 @@ class Program_weight_tensor_parameter_9: class Program_weight_tensor_parameter_10: name = "parameter_10" + original_name = "linear_1656.b_0" shape = [768] dtype = "float32" min_val = float("-0.931707") @@ -121,6 +132,7 @@ class Program_weight_tensor_parameter_10: class Program_weight_tensor_parameter_11: name = "parameter_11" + original_name = "linear_1656.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.558119") @@ -132,6 +144,7 @@ class Program_weight_tensor_parameter_11: class Program_weight_tensor_parameter_12: name = "parameter_12" + original_name = "linear_1655.b_0" shape = [768] dtype = "float32" min_val = float("-0.156714") @@ -143,6 +156,7 @@ class Program_weight_tensor_parameter_12: class Program_weight_tensor_parameter_13: name = "parameter_13" + original_name = "linear_1655.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.271784") @@ -154,6 +168,7 @@ class Program_weight_tensor_parameter_13: class Program_weight_tensor_parameter_14: name = "parameter_14" + original_name = "linear_1654.b_0" shape = [768] dtype = "float32" min_val = float("-0.0212594") @@ -165,6 +180,7 @@ class Program_weight_tensor_parameter_14: class Program_weight_tensor_parameter_15: name = "parameter_15" + original_name = "linear_1654.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.373461") @@ -176,6 +192,7 @@ class Program_weight_tensor_parameter_15: class Program_weight_tensor_parameter_16: name = "parameter_16" + original_name = "linear_1653.b_0" shape = [768] dtype = "float32" min_val = float("-0.384231") @@ -187,6 +204,7 @@ class Program_weight_tensor_parameter_16: class Program_weight_tensor_parameter_17: name = "parameter_17" + original_name = "linear_1653.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.260382") @@ -198,6 +216,7 @@ class Program_weight_tensor_parameter_17: class Program_weight_tensor_parameter_18: name = "parameter_18" + original_name = "layer_norm_569.b_0" shape = [768] dtype = "float32" min_val = float("-0.71405") @@ -209,6 +228,7 @@ class Program_weight_tensor_parameter_18: class Program_weight_tensor_parameter_19: name = "parameter_19" + original_name = "layer_norm_569.w_0" shape = [768] dtype = "float32" min_val = float("0.652963") @@ -220,6 +240,7 @@ class Program_weight_tensor_parameter_19: class Program_weight_tensor_parameter_20: name = "parameter_20" + original_name = "layer_norm_568.b_0" shape = [768] dtype = "float32" min_val = float("-0.603316") @@ -231,6 +252,7 @@ class Program_weight_tensor_parameter_20: class Program_weight_tensor_parameter_21: name = "parameter_21" + original_name = "layer_norm_568.w_0" shape = [768] dtype = "float32" min_val = float("0.40396") @@ -242,6 +264,7 @@ class Program_weight_tensor_parameter_21: class Program_weight_tensor_parameter_22: name = "parameter_22" + original_name = "linear_1652.b_0" shape = [768] dtype = "float32" min_val = float("-0.233186") @@ -253,6 +276,7 @@ class Program_weight_tensor_parameter_22: class Program_weight_tensor_parameter_23: name = "parameter_23" + original_name = "linear_1652.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-0.575921") @@ -264,6 +288,7 @@ class Program_weight_tensor_parameter_23: class Program_weight_tensor_parameter_24: name = "parameter_24" + original_name = "linear_1651.b_0" shape = [3072] dtype = "float32" min_val = float("-0.434667") @@ -275,6 +300,7 @@ class Program_weight_tensor_parameter_24: class Program_weight_tensor_parameter_25: name = "parameter_25" + original_name = "linear_1651.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.458984") @@ -286,6 +312,7 @@ class Program_weight_tensor_parameter_25: class Program_weight_tensor_parameter_26: name = "parameter_26" + original_name = "linear_1650.b_0" shape = [768] dtype = "float32" min_val = float("-0.775907") @@ -297,6 +324,7 @@ class Program_weight_tensor_parameter_26: class Program_weight_tensor_parameter_27: name = "parameter_27" + original_name = "linear_1650.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.586457") @@ -308,6 +336,7 @@ class Program_weight_tensor_parameter_27: class Program_weight_tensor_parameter_28: name = "parameter_28" + original_name = "linear_1649.b_0" shape = [768] dtype = "float32" min_val = float("-0.12162") @@ -319,6 +348,7 @@ class Program_weight_tensor_parameter_28: class Program_weight_tensor_parameter_29: name = "parameter_29" + original_name = "linear_1649.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.354975") @@ -330,6 +360,7 @@ class Program_weight_tensor_parameter_29: class Program_weight_tensor_parameter_30: name = "parameter_30" + original_name = "linear_1648.b_0" shape = [768] dtype = "float32" min_val = float("-0.00324309") @@ -341,6 +372,7 @@ class Program_weight_tensor_parameter_30: class Program_weight_tensor_parameter_31: name = "parameter_31" + original_name = "linear_1648.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.319656") @@ -352,6 +384,7 @@ class Program_weight_tensor_parameter_31: class Program_weight_tensor_parameter_32: name = "parameter_32" + original_name = "linear_1647.b_0" shape = [768] dtype = "float32" min_val = float("-0.517274") @@ -363,6 +396,7 @@ class Program_weight_tensor_parameter_32: class Program_weight_tensor_parameter_33: name = "parameter_33" + original_name = "linear_1647.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.249151") @@ -374,6 +408,7 @@ class Program_weight_tensor_parameter_33: class Program_weight_tensor_parameter_34: name = "parameter_34" + original_name = "layer_norm_567.b_0" shape = [768] dtype = "float32" min_val = float("-0.505574") @@ -385,6 +420,7 @@ class Program_weight_tensor_parameter_34: class Program_weight_tensor_parameter_35: name = "parameter_35" + original_name = "layer_norm_567.w_0" shape = [768] dtype = "float32" min_val = float("0.694022") @@ -396,6 +432,7 @@ class Program_weight_tensor_parameter_35: class Program_weight_tensor_parameter_36: name = "parameter_36" + original_name = "layer_norm_566.b_0" shape = [768] dtype = "float32" min_val = float("-0.746436") @@ -407,6 +444,7 @@ class Program_weight_tensor_parameter_36: class Program_weight_tensor_parameter_37: name = "parameter_37" + original_name = "layer_norm_566.w_0" shape = [768] dtype = "float32" min_val = float("0.472442") @@ -418,6 +456,7 @@ class Program_weight_tensor_parameter_37: class Program_weight_tensor_parameter_38: name = "parameter_38" + original_name = "linear_1646.b_0" shape = [768] dtype = "float32" min_val = float("-0.210529") @@ -429,6 +468,7 @@ class Program_weight_tensor_parameter_38: class Program_weight_tensor_parameter_39: name = "parameter_39" + original_name = "linear_1646.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-1.39826") @@ -440,6 +480,7 @@ class Program_weight_tensor_parameter_39: class Program_weight_tensor_parameter_40: name = "parameter_40" + original_name = "linear_1645.b_0" shape = [3072] dtype = "float32" min_val = float("-0.47789") @@ -451,6 +492,7 @@ class Program_weight_tensor_parameter_40: class Program_weight_tensor_parameter_41: name = "parameter_41" + original_name = "linear_1645.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.452012") @@ -462,6 +504,7 @@ class Program_weight_tensor_parameter_41: class Program_weight_tensor_parameter_42: name = "parameter_42" + original_name = "linear_1644.b_0" shape = [768] dtype = "float32" min_val = float("-0.562414") @@ -473,6 +516,7 @@ class Program_weight_tensor_parameter_42: class Program_weight_tensor_parameter_43: name = "parameter_43" + original_name = "linear_1644.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.359355") @@ -484,6 +528,7 @@ class Program_weight_tensor_parameter_43: class Program_weight_tensor_parameter_44: name = "parameter_44" + original_name = "linear_1643.b_0" shape = [768] dtype = "float32" min_val = float("-0.157931") @@ -495,6 +540,7 @@ class Program_weight_tensor_parameter_44: class Program_weight_tensor_parameter_45: name = "parameter_45" + original_name = "linear_1643.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.580239") @@ -506,6 +552,7 @@ class Program_weight_tensor_parameter_45: class Program_weight_tensor_parameter_46: name = "parameter_46" + original_name = "linear_1642.b_0" shape = [768] dtype = "float32" min_val = float("-0.00289326") @@ -517,6 +564,7 @@ class Program_weight_tensor_parameter_46: class Program_weight_tensor_parameter_47: name = "parameter_47" + original_name = "linear_1642.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.498632") @@ -528,6 +576,7 @@ class Program_weight_tensor_parameter_47: class Program_weight_tensor_parameter_48: name = "parameter_48" + original_name = "linear_1641.b_0" shape = [768] dtype = "float32" min_val = float("-0.40305") @@ -539,6 +588,7 @@ class Program_weight_tensor_parameter_48: class Program_weight_tensor_parameter_49: name = "parameter_49" + original_name = "linear_1641.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.304816") @@ -550,6 +600,7 @@ class Program_weight_tensor_parameter_49: class Program_weight_tensor_parameter_50: name = "parameter_50" + original_name = "layer_norm_565.b_0" shape = [768] dtype = "float32" min_val = float("-0.625105") @@ -561,6 +612,7 @@ class Program_weight_tensor_parameter_50: class Program_weight_tensor_parameter_51: name = "parameter_51" + original_name = "layer_norm_565.w_0" shape = [768] dtype = "float32" min_val = float("0.802812") @@ -572,6 +624,7 @@ class Program_weight_tensor_parameter_51: class Program_weight_tensor_parameter_52: name = "parameter_52" + original_name = "layer_norm_564.b_0" shape = [768] dtype = "float32" min_val = float("-0.776603") @@ -583,6 +636,7 @@ class Program_weight_tensor_parameter_52: class Program_weight_tensor_parameter_53: name = "parameter_53" + original_name = "layer_norm_564.w_0" shape = [768] dtype = "float32" min_val = float("0.47484") @@ -594,6 +648,7 @@ class Program_weight_tensor_parameter_53: class Program_weight_tensor_parameter_54: name = "parameter_54" + original_name = "linear_1640.b_0" shape = [768] dtype = "float32" min_val = float("-0.211722") @@ -605,6 +660,7 @@ class Program_weight_tensor_parameter_54: class Program_weight_tensor_parameter_55: name = "parameter_55" + original_name = "linear_1640.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-1.78128") @@ -616,6 +672,7 @@ class Program_weight_tensor_parameter_55: class Program_weight_tensor_parameter_56: name = "parameter_56" + original_name = "linear_1639.b_0" shape = [3072] dtype = "float32" min_val = float("-0.541752") @@ -627,6 +684,7 @@ class Program_weight_tensor_parameter_56: class Program_weight_tensor_parameter_57: name = "parameter_57" + original_name = "linear_1639.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.552873") @@ -638,6 +696,7 @@ class Program_weight_tensor_parameter_57: class Program_weight_tensor_parameter_58: name = "parameter_58" + original_name = "linear_1638.b_0" shape = [768] dtype = "float32" min_val = float("-0.419921") @@ -649,6 +708,7 @@ class Program_weight_tensor_parameter_58: class Program_weight_tensor_parameter_59: name = "parameter_59" + original_name = "linear_1638.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.366951") @@ -660,6 +720,7 @@ class Program_weight_tensor_parameter_59: class Program_weight_tensor_parameter_60: name = "parameter_60" + original_name = "linear_1637.b_0" shape = [768] dtype = "float32" min_val = float("-0.15052") @@ -671,6 +732,7 @@ class Program_weight_tensor_parameter_60: class Program_weight_tensor_parameter_61: name = "parameter_61" + original_name = "linear_1637.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.324908") @@ -682,6 +744,7 @@ class Program_weight_tensor_parameter_61: class Program_weight_tensor_parameter_62: name = "parameter_62" + original_name = "linear_1636.b_0" shape = [768] dtype = "float32" min_val = float("-0.00116023") @@ -693,6 +756,7 @@ class Program_weight_tensor_parameter_62: class Program_weight_tensor_parameter_63: name = "parameter_63" + original_name = "linear_1636.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.587954") @@ -704,6 +768,7 @@ class Program_weight_tensor_parameter_63: class Program_weight_tensor_parameter_64: name = "parameter_64" + original_name = "linear_1635.b_0" shape = [768] dtype = "float32" min_val = float("-0.560503") @@ -715,6 +780,7 @@ class Program_weight_tensor_parameter_64: class Program_weight_tensor_parameter_65: name = "parameter_65" + original_name = "linear_1635.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.572374") @@ -726,6 +792,7 @@ class Program_weight_tensor_parameter_65: class Program_weight_tensor_parameter_66: name = "parameter_66" + original_name = "layer_norm_563.b_0" shape = [768] dtype = "float32" min_val = float("-0.814712") @@ -737,6 +804,7 @@ class Program_weight_tensor_parameter_66: class Program_weight_tensor_parameter_67: name = "parameter_67" + original_name = "layer_norm_563.w_0" shape = [768] dtype = "float32" min_val = float("0.820371") @@ -748,6 +816,7 @@ class Program_weight_tensor_parameter_67: class Program_weight_tensor_parameter_68: name = "parameter_68" + original_name = "layer_norm_562.b_0" shape = [768] dtype = "float32" min_val = float("-1.09345") @@ -759,6 +828,7 @@ class Program_weight_tensor_parameter_68: class Program_weight_tensor_parameter_69: name = "parameter_69" + original_name = "layer_norm_562.w_0" shape = [768] dtype = "float32" min_val = float("0.528376") @@ -770,6 +840,7 @@ class Program_weight_tensor_parameter_69: class Program_weight_tensor_parameter_70: name = "parameter_70" + original_name = "linear_1634.b_0" shape = [768] dtype = "float32" min_val = float("-0.246254") @@ -781,6 +852,7 @@ class Program_weight_tensor_parameter_70: class Program_weight_tensor_parameter_71: name = "parameter_71" + original_name = "linear_1634.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-2.90791") @@ -792,6 +864,7 @@ class Program_weight_tensor_parameter_71: class Program_weight_tensor_parameter_72: name = "parameter_72" + original_name = "linear_1633.b_0" shape = [3072] dtype = "float32" min_val = float("-0.417306") @@ -803,6 +876,7 @@ class Program_weight_tensor_parameter_72: class Program_weight_tensor_parameter_73: name = "parameter_73" + original_name = "linear_1633.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.421348") @@ -814,6 +888,7 @@ class Program_weight_tensor_parameter_73: class Program_weight_tensor_parameter_74: name = "parameter_74" + original_name = "linear_1632.b_0" shape = [768] dtype = "float32" min_val = float("-0.310579") @@ -825,6 +900,7 @@ class Program_weight_tensor_parameter_74: class Program_weight_tensor_parameter_75: name = "parameter_75" + original_name = "linear_1632.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.355029") @@ -836,6 +912,7 @@ class Program_weight_tensor_parameter_75: class Program_weight_tensor_parameter_76: name = "parameter_76" + original_name = "linear_1631.b_0" shape = [768] dtype = "float32" min_val = float("-0.104342") @@ -847,6 +924,7 @@ class Program_weight_tensor_parameter_76: class Program_weight_tensor_parameter_77: name = "parameter_77" + original_name = "linear_1631.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.235806") @@ -858,6 +936,7 @@ class Program_weight_tensor_parameter_77: class Program_weight_tensor_parameter_78: name = "parameter_78" + original_name = "linear_1630.b_0" shape = [768] dtype = "float32" min_val = float("-0.00224996") @@ -869,6 +948,7 @@ class Program_weight_tensor_parameter_78: class Program_weight_tensor_parameter_79: name = "parameter_79" + original_name = "linear_1630.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.551146") @@ -880,6 +960,7 @@ class Program_weight_tensor_parameter_79: class Program_weight_tensor_parameter_80: name = "parameter_80" + original_name = "linear_1629.b_0" shape = [768] dtype = "float32" min_val = float("-0.70474") @@ -891,6 +972,7 @@ class Program_weight_tensor_parameter_80: class Program_weight_tensor_parameter_81: name = "parameter_81" + original_name = "linear_1629.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.336716") @@ -902,6 +984,7 @@ class Program_weight_tensor_parameter_81: class Program_weight_tensor_parameter_82: name = "parameter_82" + original_name = "layer_norm_561.b_0" shape = [768] dtype = "float32" min_val = float("-1.19066") @@ -913,6 +996,7 @@ class Program_weight_tensor_parameter_82: class Program_weight_tensor_parameter_83: name = "parameter_83" + original_name = "layer_norm_561.w_0" shape = [768] dtype = "float32" min_val = float("0.753388") @@ -924,6 +1008,7 @@ class Program_weight_tensor_parameter_83: class Program_weight_tensor_parameter_84: name = "parameter_84" + original_name = "layer_norm_560.b_0" shape = [768] dtype = "float32" min_val = float("-2.66016") @@ -935,6 +1020,7 @@ class Program_weight_tensor_parameter_84: class Program_weight_tensor_parameter_85: name = "parameter_85" + original_name = "layer_norm_560.w_0" shape = [768] dtype = "float32" min_val = float("0.128009") @@ -946,6 +1032,7 @@ class Program_weight_tensor_parameter_85: class Program_weight_tensor_parameter_86: name = "parameter_86" + original_name = "linear_1628.b_0" shape = [768] dtype = "float32" min_val = float("-0.249888") @@ -957,6 +1044,7 @@ class Program_weight_tensor_parameter_86: class Program_weight_tensor_parameter_87: name = "parameter_87" + original_name = "linear_1628.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-6.3499") @@ -968,6 +1056,7 @@ class Program_weight_tensor_parameter_87: class Program_weight_tensor_parameter_88: name = "parameter_88" + original_name = "linear_1627.b_0" shape = [3072] dtype = "float32" min_val = float("-0.472484") @@ -979,6 +1068,7 @@ class Program_weight_tensor_parameter_88: class Program_weight_tensor_parameter_89: name = "parameter_89" + original_name = "linear_1627.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.955099") @@ -990,6 +1080,7 @@ class Program_weight_tensor_parameter_89: class Program_weight_tensor_parameter_90: name = "parameter_90" + original_name = "linear_1626.b_0" shape = [768] dtype = "float32" min_val = float("-0.24265") @@ -1001,6 +1092,7 @@ class Program_weight_tensor_parameter_90: class Program_weight_tensor_parameter_91: name = "parameter_91" + original_name = "linear_1626.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.715098") @@ -1012,6 +1104,7 @@ class Program_weight_tensor_parameter_91: class Program_weight_tensor_parameter_92: name = "parameter_92" + original_name = "linear_1625.b_0" shape = [768] dtype = "float32" min_val = float("-0.421609") @@ -1023,6 +1116,7 @@ class Program_weight_tensor_parameter_92: class Program_weight_tensor_parameter_93: name = "parameter_93" + original_name = "linear_1625.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.285378") @@ -1034,6 +1128,7 @@ class Program_weight_tensor_parameter_93: class Program_weight_tensor_parameter_94: name = "parameter_94" + original_name = "linear_1624.b_0" shape = [768] dtype = "float32" min_val = float("-0.000507141") @@ -1045,6 +1140,7 @@ class Program_weight_tensor_parameter_94: class Program_weight_tensor_parameter_95: name = "parameter_95" + original_name = "linear_1624.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.324055") @@ -1056,6 +1152,7 @@ class Program_weight_tensor_parameter_95: class Program_weight_tensor_parameter_96: name = "parameter_96" + original_name = "linear_1623.b_0" shape = [768] dtype = "float32" min_val = float("-1.03834") @@ -1067,6 +1164,7 @@ class Program_weight_tensor_parameter_96: class Program_weight_tensor_parameter_97: name = "parameter_97" + original_name = "linear_1623.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.421575") @@ -1078,6 +1176,7 @@ class Program_weight_tensor_parameter_97: class Program_weight_tensor_parameter_98: name = "parameter_98" + original_name = "layer_norm_559.b_0" shape = [768] dtype = "float32" min_val = float("-3.10681") @@ -1089,6 +1188,7 @@ class Program_weight_tensor_parameter_98: class Program_weight_tensor_parameter_99: name = "parameter_99" + original_name = "layer_norm_559.w_0" shape = [768] dtype = "float32" min_val = float("0.0784423") @@ -1100,6 +1200,7 @@ class Program_weight_tensor_parameter_99: class Program_weight_tensor_parameter_100: name = "parameter_100" + original_name = "embedding_97.w_0" shape = [16, 768] dtype = "float32" min_val = float("-0.0337877") @@ -1111,6 +1212,7 @@ class Program_weight_tensor_parameter_100: class Program_weight_tensor_parameter_101: name = "parameter_101" + original_name = "embedding_96.w_0" shape = [4, 768] dtype = "float32" min_val = float("-0.0737263") @@ -1122,6 +1224,7 @@ class Program_weight_tensor_parameter_101: class Program_weight_tensor_parameter_102: name = "parameter_102" + original_name = "embedding_95.w_0" shape = [2048, 768] dtype = "float32" min_val = float("-0.853686") @@ -1133,6 +1236,7 @@ class Program_weight_tensor_parameter_102: class Program_weight_tensor_parameter_103: name = "parameter_103" + original_name = "embedding_94.w_0" shape = [40000, 768] dtype = "float32" min_val = float("-1.13693") diff --git a/paddle_samples/PaddleNLP/rocketqa-mini-cross-encoder/input_meta.py b/paddle_samples/PaddleNLP/rocketqa-mini-cross-encoder/input_meta.py index 4f25a05a9..055cd8978 100644 --- a/paddle_samples/PaddleNLP/rocketqa-mini-cross-encoder/input_meta.py +++ b/paddle_samples/PaddleNLP/rocketqa-mini-cross-encoder/input_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_data_0: name = "data_0" + original_name = "input_ids" shape = [1, 21] dtype = "int64" data = [ @@ -29,6 +30,7 @@ class Program_weight_tensor_data_0: class Program_weight_tensor_data_1: name = "data_1" + original_name = "token_type_ids" shape = [1, 21] dtype = "int64" data = [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0] diff --git a/paddle_samples/PaddleNLP/rocketqa-mini-cross-encoder/model.py b/paddle_samples/PaddleNLP/rocketqa-mini-cross-encoder/model.py index d9f88a642..820b00da2 100644 --- a/paddle_samples/PaddleNLP/rocketqa-mini-cross-encoder/model.py +++ b/paddle_samples/PaddleNLP/rocketqa-mini-cross-encoder/model.py @@ -170,10 +170,11 @@ def forward( # pd_op.embedding: (1x21x384xf32) <- (1x21xi64, 2048x384xf32) embedding_1 = paddle._C_ops.embedding(subtract_0, parameter_102, -1, False) - del parameter_102 + del parameter_102, subtract_0 # pd_op.add: (1x21x384xf32) <- (1x21x384xf32, 1x21x384xf32) add_0 = paddle._C_ops.add(embedding_0, embedding_1) + del embedding_0, embedding_1 # pd_op.embedding: (1x21x384xf32) <- (1x21xi64, 4x384xf32) embedding_2 = paddle._C_ops.embedding(data_1, parameter_101, -1, False) @@ -181,6 +182,7 @@ def forward( # pd_op.add: (1x21x384xf32) <- (1x21x384xf32, 1x21x384xf32) add_1 = paddle._C_ops.add(add_0, embedding_2) + del add_0, embedding_2 # pd_op.full: (1xf32) <- () full_4 = paddle._C_ops.full( @@ -193,10 +195,11 @@ def forward( # pd_op.embedding: (1x21x384xf32) <- (1x21xi64, 16x384xf32) embedding_3 = paddle._C_ops.embedding(scale_1, parameter_100, -1, False) - del parameter_100 + del parameter_100, scale_1 # pd_op.add: (1x21x384xf32) <- (1x21x384xf32, 1x21x384xf32) add_2 = paddle._C_ops.add(add_1, embedding_3) + del add_1, embedding_3 # pd_op.layer_norm: (1x21x384xf32, 1x21xf32, 1x21xf32) <- (1x21x384xf32, 384xf32, 384xf32) layer_norm_0, layer_norm_1, layer_norm_2 = (lambda x, f: f(x))( @@ -205,71 +208,17 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_98, parameter_99 + del add_2, parameter_98, parameter_99 # pd_op.full: (1xf32) <- () full_5 = paddle._C_ops.full( [1], float("0.1"), paddle.float32, paddle.core.CPUPlace() ) - # pd_op.assign: (1xf32) <- (1xf32) - assign_0 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_1 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_2 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_3 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_4 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_5 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_6 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_7 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_8 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_9 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_10 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_11 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_12 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_13 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_14 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_15 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_16 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_17 = full_5 - # pd_op.dropout: (1x21x384xf32, 1x21x384xui8) <- (1x21x384xf32, None, 1xf32) dropout_0, dropout_1 = (lambda x, f: f(x))( paddle._C_ops.dropout( - layer_norm_0, None, full_5, False, "upscale_in_train", 0, False + layer_norm_0, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -281,13 +230,14 @@ def forward( # pd_op.add: (1x21x384xf32) <- (1x21x384xf32, 384xf32) add_3 = paddle._C_ops.add(matmul_0, parameter_96) - del parameter_96 + del matmul_0, parameter_96 # pd_op.full_int_array: (4xi64) <- () full_int_array_1 = [0, 0, 12, 32] # pd_op.reshape: (1x21x12x32xf32) <- (1x21x384xf32, 4xi64) reshape_0 = paddle._C_ops.reshape(add_3, full_int_array_1) + del add_3 # pd_op.transpose: (1x12x21x32xf32) <- (1x21x12x32xf32) transpose_0 = paddle._C_ops.transpose(reshape_0, [0, 2, 1, 3]) @@ -299,7 +249,7 @@ def forward( # pd_op.add: (1x21x384xf32) <- (1x21x384xf32, 384xf32) add_4 = paddle._C_ops.add(matmul_1, parameter_94) - del parameter_94 + del matmul_1, parameter_94 # pd_op.matmul: (1x21x384xf32) <- (1x21x384xf32, 384x384xf32) matmul_2 = paddle._C_ops.matmul(dropout_0, parameter_93, False, False) @@ -307,10 +257,11 @@ def forward( # pd_op.add: (1x21x384xf32) <- (1x21x384xf32, 384xf32) add_5 = paddle._C_ops.add(matmul_2, parameter_92) - del parameter_92 + del matmul_2, parameter_92 # pd_op.reshape: (1x21x12x32xf32) <- (1x21x384xf32, 4xi64) reshape_1 = paddle._C_ops.reshape(add_4, full_int_array_1) + del add_4 # pd_op.transpose: (1x12x21x32xf32) <- (1x21x12x32xf32) transpose_1 = paddle._C_ops.transpose(reshape_1, [0, 2, 1, 3]) @@ -318,6 +269,7 @@ def forward( # pd_op.reshape: (1x21x12x32xf32) <- (1x21x384xf32, 4xi64) reshape_2 = paddle._C_ops.reshape(add_5, full_int_array_1) + del add_5 # pd_op.transpose: (1x12x21x32xf32) <- (1x21x12x32xf32) transpose_2 = paddle._C_ops.transpose(reshape_2, [0, 2, 1, 3]) @@ -328,30 +280,17 @@ def forward( [1], float("0.176777"), paddle.float32, paddle.core.CPUPlace() ) - # pd_op.assign: (1xf32) <- (1xf32) - assign_18 = full_6 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_19 = full_6 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_20 = full_6 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_21 = full_6 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_22 = full_6 - # pd_op.scale: (1x12x21x32xf32) <- (1x12x21x32xf32, 1xf32) scale_2 = paddle._C_ops.scale(transpose_0, full_6, float("0"), True) del transpose_0 # pd_op.matmul: (1x12x21x21xf32) <- (1x12x21x32xf32, 1x12x21x32xf32) matmul_3 = paddle._C_ops.matmul(scale_2, transpose_1, False, True) + del scale_2, transpose_1 # pd_op.add: (1x12x21x21xf32) <- (1x12x21x21xf32, 1x1x1x21xf32) add_6 = paddle._C_ops.add(matmul_3, unsqueeze_0) + del matmul_3 # pd_op.softmax: (1x12x21x21xf32) <- (1x12x21x21xf32) softmax_0 = paddle._C_ops.softmax(add_6, -1) @@ -360,13 +299,15 @@ def forward( # pd_op.dropout: (1x12x21x21xf32, 1x12x21x21xui8) <- (1x12x21x21xf32, None, 1xf32) dropout_2, dropout_3 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_0, None, full_5, False, "upscale_in_train", 0, False + softmax_0, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_0 # pd_op.matmul: (1x12x21x32xf32) <- (1x12x21x21xf32, 1x12x21x32xf32) matmul_4 = paddle._C_ops.matmul(dropout_2, transpose_2, False, False) + del dropout_2, transpose_2 # pd_op.transpose: (1x21x12x32xf32) <- (1x12x21x32xf32) transpose_3 = paddle._C_ops.transpose(matmul_4, [0, 2, 1, 3]) @@ -377,19 +318,20 @@ def forward( # pd_op.reshape: (1x21x384xf32) <- (1x21x12x32xf32, 3xi64) reshape_3 = paddle._C_ops.reshape(transpose_3, full_int_array_2) + del transpose_3 # pd_op.matmul: (1x21x384xf32) <- (1x21x384xf32, 384x384xf32) matmul_5 = paddle._C_ops.matmul(reshape_3, parameter_91, False, False) - del parameter_91 + del parameter_91, reshape_3 # pd_op.add: (1x21x384xf32) <- (1x21x384xf32, 384xf32) add_7 = paddle._C_ops.add(matmul_5, parameter_90) - del parameter_90 + del matmul_5, parameter_90 # pd_op.dropout: (1x21x384xf32, 1x21x384xui8) <- (1x21x384xf32, None, 1xf32) dropout_4, dropout_5 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_7, None, full_5, False, "upscale_in_train", 0, False + add_7, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -397,6 +339,7 @@ def forward( # pd_op.add: (1x21x384xf32) <- (1x21x384xf32, 1x21x384xf32) add_8 = paddle._C_ops.add(dropout_0, dropout_4) + del dropout_0, dropout_4 # pd_op.layer_norm: (1x21x384xf32, 1x21xf32, 1x21xf32) <- (1x21x384xf32, 384xf32, 384xf32) layer_norm_3, layer_norm_4, layer_norm_5 = (lambda x, f: f(x))( @@ -405,7 +348,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_84, parameter_85 + del add_8, parameter_84, parameter_85 # pd_op.matmul: (1x21x1536xf32) <- (1x21x384xf32, 384x1536xf32) matmul_6 = paddle._C_ops.matmul(layer_norm_3, parameter_89, False, False) @@ -413,23 +356,24 @@ def forward( # pd_op.add: (1x21x1536xf32) <- (1x21x1536xf32, 1536xf32) add_9 = paddle._C_ops.add(matmul_6, parameter_88) - del parameter_88 + del matmul_6, parameter_88 # pd_op.gelu: (1x21x1536xf32) <- (1x21x1536xf32) gelu_0 = paddle._C_ops.gelu(add_9, False) + del add_9 # pd_op.matmul: (1x21x384xf32) <- (1x21x1536xf32, 1536x384xf32) matmul_7 = paddle._C_ops.matmul(gelu_0, parameter_87, False, False) - del parameter_87 + del gelu_0, parameter_87 # pd_op.add: (1x21x384xf32) <- (1x21x384xf32, 384xf32) add_10 = paddle._C_ops.add(matmul_7, parameter_86) - del parameter_86 + del matmul_7, parameter_86 # pd_op.dropout: (1x21x384xf32, 1x21x384xui8) <- (1x21x384xf32, None, 1xf32) dropout_6, dropout_7 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_10, None, full_5, False, "upscale_in_train", 0, False + add_10, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -437,6 +381,7 @@ def forward( # pd_op.add: (1x21x384xf32) <- (1x21x384xf32, 1x21x384xf32) add_11 = paddle._C_ops.add(layer_norm_3, dropout_6) + del dropout_6, layer_norm_3 # pd_op.layer_norm: (1x21x384xf32, 1x21xf32, 1x21xf32) <- (1x21x384xf32, 384xf32, 384xf32) layer_norm_6, layer_norm_7, layer_norm_8 = (lambda x, f: f(x))( @@ -445,7 +390,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_82, parameter_83 + del add_11, parameter_82, parameter_83 # pd_op.matmul: (1x21x384xf32) <- (1x21x384xf32, 384x384xf32) matmul_8 = paddle._C_ops.matmul(layer_norm_6, parameter_81, False, False) @@ -453,10 +398,11 @@ def forward( # pd_op.add: (1x21x384xf32) <- (1x21x384xf32, 384xf32) add_12 = paddle._C_ops.add(matmul_8, parameter_80) - del parameter_80 + del matmul_8, parameter_80 # pd_op.reshape: (1x21x12x32xf32) <- (1x21x384xf32, 4xi64) reshape_4 = paddle._C_ops.reshape(add_12, full_int_array_1) + del add_12 # pd_op.transpose: (1x12x21x32xf32) <- (1x21x12x32xf32) transpose_4 = paddle._C_ops.transpose(reshape_4, [0, 2, 1, 3]) @@ -468,7 +414,7 @@ def forward( # pd_op.add: (1x21x384xf32) <- (1x21x384xf32, 384xf32) add_13 = paddle._C_ops.add(matmul_9, parameter_78) - del parameter_78 + del matmul_9, parameter_78 # pd_op.matmul: (1x21x384xf32) <- (1x21x384xf32, 384x384xf32) matmul_10 = paddle._C_ops.matmul(layer_norm_6, parameter_77, False, False) @@ -476,10 +422,11 @@ def forward( # pd_op.add: (1x21x384xf32) <- (1x21x384xf32, 384xf32) add_14 = paddle._C_ops.add(matmul_10, parameter_76) - del parameter_76 + del matmul_10, parameter_76 # pd_op.reshape: (1x21x12x32xf32) <- (1x21x384xf32, 4xi64) reshape_5 = paddle._C_ops.reshape(add_13, full_int_array_1) + del add_13 # pd_op.transpose: (1x12x21x32xf32) <- (1x21x12x32xf32) transpose_5 = paddle._C_ops.transpose(reshape_5, [0, 2, 1, 3]) @@ -487,6 +434,7 @@ def forward( # pd_op.reshape: (1x21x12x32xf32) <- (1x21x384xf32, 4xi64) reshape_6 = paddle._C_ops.reshape(add_14, full_int_array_1) + del add_14 # pd_op.transpose: (1x12x21x32xf32) <- (1x21x12x32xf32) transpose_6 = paddle._C_ops.transpose(reshape_6, [0, 2, 1, 3]) @@ -498,9 +446,11 @@ def forward( # pd_op.matmul: (1x12x21x21xf32) <- (1x12x21x32xf32, 1x12x21x32xf32) matmul_11 = paddle._C_ops.matmul(scale_3, transpose_5, False, True) + del scale_3, transpose_5 # pd_op.add: (1x12x21x21xf32) <- (1x12x21x21xf32, 1x1x1x21xf32) add_15 = paddle._C_ops.add(matmul_11, unsqueeze_0) + del matmul_11 # pd_op.softmax: (1x12x21x21xf32) <- (1x12x21x21xf32) softmax_1 = paddle._C_ops.softmax(add_15, -1) @@ -509,13 +459,15 @@ def forward( # pd_op.dropout: (1x12x21x21xf32, 1x12x21x21xui8) <- (1x12x21x21xf32, None, 1xf32) dropout_8, dropout_9 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_1, None, full_5, False, "upscale_in_train", 0, False + softmax_1, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_1 # pd_op.matmul: (1x12x21x32xf32) <- (1x12x21x21xf32, 1x12x21x32xf32) matmul_12 = paddle._C_ops.matmul(dropout_8, transpose_6, False, False) + del dropout_8, transpose_6 # pd_op.transpose: (1x21x12x32xf32) <- (1x12x21x32xf32) transpose_7 = paddle._C_ops.transpose(matmul_12, [0, 2, 1, 3]) @@ -523,19 +475,20 @@ def forward( # pd_op.reshape: (1x21x384xf32) <- (1x21x12x32xf32, 3xi64) reshape_7 = paddle._C_ops.reshape(transpose_7, full_int_array_2) + del transpose_7 # pd_op.matmul: (1x21x384xf32) <- (1x21x384xf32, 384x384xf32) matmul_13 = paddle._C_ops.matmul(reshape_7, parameter_75, False, False) - del parameter_75 + del parameter_75, reshape_7 # pd_op.add: (1x21x384xf32) <- (1x21x384xf32, 384xf32) add_16 = paddle._C_ops.add(matmul_13, parameter_74) - del parameter_74 + del matmul_13, parameter_74 # pd_op.dropout: (1x21x384xf32, 1x21x384xui8) <- (1x21x384xf32, None, 1xf32) dropout_10, dropout_11 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_16, None, full_5, False, "upscale_in_train", 0, False + add_16, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -543,6 +496,7 @@ def forward( # pd_op.add: (1x21x384xf32) <- (1x21x384xf32, 1x21x384xf32) add_17 = paddle._C_ops.add(layer_norm_6, dropout_10) + del dropout_10, layer_norm_6 # pd_op.layer_norm: (1x21x384xf32, 1x21xf32, 1x21xf32) <- (1x21x384xf32, 384xf32, 384xf32) layer_norm_9, layer_norm_10, layer_norm_11 = (lambda x, f: f(x))( @@ -551,7 +505,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_68, parameter_69 + del add_17, parameter_68, parameter_69 # pd_op.matmul: (1x21x1536xf32) <- (1x21x384xf32, 384x1536xf32) matmul_14 = paddle._C_ops.matmul(layer_norm_9, parameter_73, False, False) @@ -559,23 +513,24 @@ def forward( # pd_op.add: (1x21x1536xf32) <- (1x21x1536xf32, 1536xf32) add_18 = paddle._C_ops.add(matmul_14, parameter_72) - del parameter_72 + del matmul_14, parameter_72 # pd_op.gelu: (1x21x1536xf32) <- (1x21x1536xf32) gelu_1 = paddle._C_ops.gelu(add_18, False) + del add_18 # pd_op.matmul: (1x21x384xf32) <- (1x21x1536xf32, 1536x384xf32) matmul_15 = paddle._C_ops.matmul(gelu_1, parameter_71, False, False) - del parameter_71 + del gelu_1, parameter_71 # pd_op.add: (1x21x384xf32) <- (1x21x384xf32, 384xf32) add_19 = paddle._C_ops.add(matmul_15, parameter_70) - del parameter_70 + del matmul_15, parameter_70 # pd_op.dropout: (1x21x384xf32, 1x21x384xui8) <- (1x21x384xf32, None, 1xf32) dropout_12, dropout_13 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_19, None, full_5, False, "upscale_in_train", 0, False + add_19, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -583,6 +538,7 @@ def forward( # pd_op.add: (1x21x384xf32) <- (1x21x384xf32, 1x21x384xf32) add_20 = paddle._C_ops.add(layer_norm_9, dropout_12) + del dropout_12, layer_norm_9 # pd_op.layer_norm: (1x21x384xf32, 1x21xf32, 1x21xf32) <- (1x21x384xf32, 384xf32, 384xf32) layer_norm_12, layer_norm_13, layer_norm_14 = (lambda x, f: f(x))( @@ -591,7 +547,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_66, parameter_67 + del add_20, parameter_66, parameter_67 # pd_op.matmul: (1x21x384xf32) <- (1x21x384xf32, 384x384xf32) matmul_16 = paddle._C_ops.matmul(layer_norm_12, parameter_65, False, False) @@ -599,10 +555,11 @@ def forward( # pd_op.add: (1x21x384xf32) <- (1x21x384xf32, 384xf32) add_21 = paddle._C_ops.add(matmul_16, parameter_64) - del parameter_64 + del matmul_16, parameter_64 # pd_op.reshape: (1x21x12x32xf32) <- (1x21x384xf32, 4xi64) reshape_8 = paddle._C_ops.reshape(add_21, full_int_array_1) + del add_21 # pd_op.transpose: (1x12x21x32xf32) <- (1x21x12x32xf32) transpose_8 = paddle._C_ops.transpose(reshape_8, [0, 2, 1, 3]) @@ -614,7 +571,7 @@ def forward( # pd_op.add: (1x21x384xf32) <- (1x21x384xf32, 384xf32) add_22 = paddle._C_ops.add(matmul_17, parameter_62) - del parameter_62 + del matmul_17, parameter_62 # pd_op.matmul: (1x21x384xf32) <- (1x21x384xf32, 384x384xf32) matmul_18 = paddle._C_ops.matmul(layer_norm_12, parameter_61, False, False) @@ -622,10 +579,11 @@ def forward( # pd_op.add: (1x21x384xf32) <- (1x21x384xf32, 384xf32) add_23 = paddle._C_ops.add(matmul_18, parameter_60) - del parameter_60 + del matmul_18, parameter_60 # pd_op.reshape: (1x21x12x32xf32) <- (1x21x384xf32, 4xi64) reshape_9 = paddle._C_ops.reshape(add_22, full_int_array_1) + del add_22 # pd_op.transpose: (1x12x21x32xf32) <- (1x21x12x32xf32) transpose_9 = paddle._C_ops.transpose(reshape_9, [0, 2, 1, 3]) @@ -633,6 +591,7 @@ def forward( # pd_op.reshape: (1x21x12x32xf32) <- (1x21x384xf32, 4xi64) reshape_10 = paddle._C_ops.reshape(add_23, full_int_array_1) + del add_23 # pd_op.transpose: (1x12x21x32xf32) <- (1x21x12x32xf32) transpose_10 = paddle._C_ops.transpose(reshape_10, [0, 2, 1, 3]) @@ -644,9 +603,11 @@ def forward( # pd_op.matmul: (1x12x21x21xf32) <- (1x12x21x32xf32, 1x12x21x32xf32) matmul_19 = paddle._C_ops.matmul(scale_4, transpose_9, False, True) + del scale_4, transpose_9 # pd_op.add: (1x12x21x21xf32) <- (1x12x21x21xf32, 1x1x1x21xf32) add_24 = paddle._C_ops.add(matmul_19, unsqueeze_0) + del matmul_19 # pd_op.softmax: (1x12x21x21xf32) <- (1x12x21x21xf32) softmax_2 = paddle._C_ops.softmax(add_24, -1) @@ -655,13 +616,15 @@ def forward( # pd_op.dropout: (1x12x21x21xf32, 1x12x21x21xui8) <- (1x12x21x21xf32, None, 1xf32) dropout_14, dropout_15 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_2, None, full_5, False, "upscale_in_train", 0, False + softmax_2, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_2 # pd_op.matmul: (1x12x21x32xf32) <- (1x12x21x21xf32, 1x12x21x32xf32) matmul_20 = paddle._C_ops.matmul(dropout_14, transpose_10, False, False) + del dropout_14, transpose_10 # pd_op.transpose: (1x21x12x32xf32) <- (1x12x21x32xf32) transpose_11 = paddle._C_ops.transpose(matmul_20, [0, 2, 1, 3]) @@ -669,19 +632,20 @@ def forward( # pd_op.reshape: (1x21x384xf32) <- (1x21x12x32xf32, 3xi64) reshape_11 = paddle._C_ops.reshape(transpose_11, full_int_array_2) + del transpose_11 # pd_op.matmul: (1x21x384xf32) <- (1x21x384xf32, 384x384xf32) matmul_21 = paddle._C_ops.matmul(reshape_11, parameter_59, False, False) - del parameter_59 + del parameter_59, reshape_11 # pd_op.add: (1x21x384xf32) <- (1x21x384xf32, 384xf32) add_25 = paddle._C_ops.add(matmul_21, parameter_58) - del parameter_58 + del matmul_21, parameter_58 # pd_op.dropout: (1x21x384xf32, 1x21x384xui8) <- (1x21x384xf32, None, 1xf32) dropout_16, dropout_17 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_25, None, full_5, False, "upscale_in_train", 0, False + add_25, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -689,6 +653,7 @@ def forward( # pd_op.add: (1x21x384xf32) <- (1x21x384xf32, 1x21x384xf32) add_26 = paddle._C_ops.add(layer_norm_12, dropout_16) + del dropout_16, layer_norm_12 # pd_op.layer_norm: (1x21x384xf32, 1x21xf32, 1x21xf32) <- (1x21x384xf32, 384xf32, 384xf32) layer_norm_15, layer_norm_16, layer_norm_17 = (lambda x, f: f(x))( @@ -697,7 +662,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_52, parameter_53 + del add_26, parameter_52, parameter_53 # pd_op.matmul: (1x21x1536xf32) <- (1x21x384xf32, 384x1536xf32) matmul_22 = paddle._C_ops.matmul(layer_norm_15, parameter_57, False, False) @@ -705,23 +670,24 @@ def forward( # pd_op.add: (1x21x1536xf32) <- (1x21x1536xf32, 1536xf32) add_27 = paddle._C_ops.add(matmul_22, parameter_56) - del parameter_56 + del matmul_22, parameter_56 # pd_op.gelu: (1x21x1536xf32) <- (1x21x1536xf32) gelu_2 = paddle._C_ops.gelu(add_27, False) + del add_27 # pd_op.matmul: (1x21x384xf32) <- (1x21x1536xf32, 1536x384xf32) matmul_23 = paddle._C_ops.matmul(gelu_2, parameter_55, False, False) - del parameter_55 + del gelu_2, parameter_55 # pd_op.add: (1x21x384xf32) <- (1x21x384xf32, 384xf32) add_28 = paddle._C_ops.add(matmul_23, parameter_54) - del parameter_54 + del matmul_23, parameter_54 # pd_op.dropout: (1x21x384xf32, 1x21x384xui8) <- (1x21x384xf32, None, 1xf32) dropout_18, dropout_19 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_28, None, full_5, False, "upscale_in_train", 0, False + add_28, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -729,6 +695,7 @@ def forward( # pd_op.add: (1x21x384xf32) <- (1x21x384xf32, 1x21x384xf32) add_29 = paddle._C_ops.add(layer_norm_15, dropout_18) + del dropout_18, layer_norm_15 # pd_op.layer_norm: (1x21x384xf32, 1x21xf32, 1x21xf32) <- (1x21x384xf32, 384xf32, 384xf32) layer_norm_18, layer_norm_19, layer_norm_20 = (lambda x, f: f(x))( @@ -737,7 +704,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_50, parameter_51 + del add_29, parameter_50, parameter_51 # pd_op.matmul: (1x21x384xf32) <- (1x21x384xf32, 384x384xf32) matmul_24 = paddle._C_ops.matmul(layer_norm_18, parameter_49, False, False) @@ -745,10 +712,11 @@ def forward( # pd_op.add: (1x21x384xf32) <- (1x21x384xf32, 384xf32) add_30 = paddle._C_ops.add(matmul_24, parameter_48) - del parameter_48 + del matmul_24, parameter_48 # pd_op.reshape: (1x21x12x32xf32) <- (1x21x384xf32, 4xi64) reshape_12 = paddle._C_ops.reshape(add_30, full_int_array_1) + del add_30 # pd_op.transpose: (1x12x21x32xf32) <- (1x21x12x32xf32) transpose_12 = paddle._C_ops.transpose(reshape_12, [0, 2, 1, 3]) @@ -760,7 +728,7 @@ def forward( # pd_op.add: (1x21x384xf32) <- (1x21x384xf32, 384xf32) add_31 = paddle._C_ops.add(matmul_25, parameter_46) - del parameter_46 + del matmul_25, parameter_46 # pd_op.matmul: (1x21x384xf32) <- (1x21x384xf32, 384x384xf32) matmul_26 = paddle._C_ops.matmul(layer_norm_18, parameter_45, False, False) @@ -768,10 +736,11 @@ def forward( # pd_op.add: (1x21x384xf32) <- (1x21x384xf32, 384xf32) add_32 = paddle._C_ops.add(matmul_26, parameter_44) - del parameter_44 + del matmul_26, parameter_44 # pd_op.reshape: (1x21x12x32xf32) <- (1x21x384xf32, 4xi64) reshape_13 = paddle._C_ops.reshape(add_31, full_int_array_1) + del add_31 # pd_op.transpose: (1x12x21x32xf32) <- (1x21x12x32xf32) transpose_13 = paddle._C_ops.transpose(reshape_13, [0, 2, 1, 3]) @@ -779,6 +748,7 @@ def forward( # pd_op.reshape: (1x21x12x32xf32) <- (1x21x384xf32, 4xi64) reshape_14 = paddle._C_ops.reshape(add_32, full_int_array_1) + del add_32 # pd_op.transpose: (1x12x21x32xf32) <- (1x21x12x32xf32) transpose_14 = paddle._C_ops.transpose(reshape_14, [0, 2, 1, 3]) @@ -790,9 +760,11 @@ def forward( # pd_op.matmul: (1x12x21x21xf32) <- (1x12x21x32xf32, 1x12x21x32xf32) matmul_27 = paddle._C_ops.matmul(scale_5, transpose_13, False, True) + del scale_5, transpose_13 # pd_op.add: (1x12x21x21xf32) <- (1x12x21x21xf32, 1x1x1x21xf32) add_33 = paddle._C_ops.add(matmul_27, unsqueeze_0) + del matmul_27 # pd_op.softmax: (1x12x21x21xf32) <- (1x12x21x21xf32) softmax_3 = paddle._C_ops.softmax(add_33, -1) @@ -801,13 +773,15 @@ def forward( # pd_op.dropout: (1x12x21x21xf32, 1x12x21x21xui8) <- (1x12x21x21xf32, None, 1xf32) dropout_20, dropout_21 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_3, None, full_5, False, "upscale_in_train", 0, False + softmax_3, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_3 # pd_op.matmul: (1x12x21x32xf32) <- (1x12x21x21xf32, 1x12x21x32xf32) matmul_28 = paddle._C_ops.matmul(dropout_20, transpose_14, False, False) + del dropout_20, transpose_14 # pd_op.transpose: (1x21x12x32xf32) <- (1x12x21x32xf32) transpose_15 = paddle._C_ops.transpose(matmul_28, [0, 2, 1, 3]) @@ -815,19 +789,20 @@ def forward( # pd_op.reshape: (1x21x384xf32) <- (1x21x12x32xf32, 3xi64) reshape_15 = paddle._C_ops.reshape(transpose_15, full_int_array_2) + del transpose_15 # pd_op.matmul: (1x21x384xf32) <- (1x21x384xf32, 384x384xf32) matmul_29 = paddle._C_ops.matmul(reshape_15, parameter_43, False, False) - del parameter_43 + del parameter_43, reshape_15 # pd_op.add: (1x21x384xf32) <- (1x21x384xf32, 384xf32) add_34 = paddle._C_ops.add(matmul_29, parameter_42) - del parameter_42 + del matmul_29, parameter_42 # pd_op.dropout: (1x21x384xf32, 1x21x384xui8) <- (1x21x384xf32, None, 1xf32) dropout_22, dropout_23 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_34, None, full_5, False, "upscale_in_train", 0, False + add_34, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -835,6 +810,7 @@ def forward( # pd_op.add: (1x21x384xf32) <- (1x21x384xf32, 1x21x384xf32) add_35 = paddle._C_ops.add(layer_norm_18, dropout_22) + del dropout_22, layer_norm_18 # pd_op.layer_norm: (1x21x384xf32, 1x21xf32, 1x21xf32) <- (1x21x384xf32, 384xf32, 384xf32) layer_norm_21, layer_norm_22, layer_norm_23 = (lambda x, f: f(x))( @@ -843,7 +819,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_36, parameter_37 + del add_35, parameter_36, parameter_37 # pd_op.matmul: (1x21x1536xf32) <- (1x21x384xf32, 384x1536xf32) matmul_30 = paddle._C_ops.matmul(layer_norm_21, parameter_41, False, False) @@ -851,23 +827,24 @@ def forward( # pd_op.add: (1x21x1536xf32) <- (1x21x1536xf32, 1536xf32) add_36 = paddle._C_ops.add(matmul_30, parameter_40) - del parameter_40 + del matmul_30, parameter_40 # pd_op.gelu: (1x21x1536xf32) <- (1x21x1536xf32) gelu_3 = paddle._C_ops.gelu(add_36, False) + del add_36 # pd_op.matmul: (1x21x384xf32) <- (1x21x1536xf32, 1536x384xf32) matmul_31 = paddle._C_ops.matmul(gelu_3, parameter_39, False, False) - del parameter_39 + del gelu_3, parameter_39 # pd_op.add: (1x21x384xf32) <- (1x21x384xf32, 384xf32) add_37 = paddle._C_ops.add(matmul_31, parameter_38) - del parameter_38 + del matmul_31, parameter_38 # pd_op.dropout: (1x21x384xf32, 1x21x384xui8) <- (1x21x384xf32, None, 1xf32) dropout_24, dropout_25 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_37, None, full_5, False, "upscale_in_train", 0, False + add_37, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -875,6 +852,7 @@ def forward( # pd_op.add: (1x21x384xf32) <- (1x21x384xf32, 1x21x384xf32) add_38 = paddle._C_ops.add(layer_norm_21, dropout_24) + del dropout_24, layer_norm_21 # pd_op.layer_norm: (1x21x384xf32, 1x21xf32, 1x21xf32) <- (1x21x384xf32, 384xf32, 384xf32) layer_norm_24, layer_norm_25, layer_norm_26 = (lambda x, f: f(x))( @@ -883,7 +861,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_34, parameter_35 + del add_38, parameter_34, parameter_35 # pd_op.matmul: (1x21x384xf32) <- (1x21x384xf32, 384x384xf32) matmul_32 = paddle._C_ops.matmul(layer_norm_24, parameter_33, False, False) @@ -891,10 +869,11 @@ def forward( # pd_op.add: (1x21x384xf32) <- (1x21x384xf32, 384xf32) add_39 = paddle._C_ops.add(matmul_32, parameter_32) - del parameter_32 + del matmul_32, parameter_32 # pd_op.reshape: (1x21x12x32xf32) <- (1x21x384xf32, 4xi64) reshape_16 = paddle._C_ops.reshape(add_39, full_int_array_1) + del add_39 # pd_op.transpose: (1x12x21x32xf32) <- (1x21x12x32xf32) transpose_16 = paddle._C_ops.transpose(reshape_16, [0, 2, 1, 3]) @@ -906,7 +885,7 @@ def forward( # pd_op.add: (1x21x384xf32) <- (1x21x384xf32, 384xf32) add_40 = paddle._C_ops.add(matmul_33, parameter_30) - del parameter_30 + del matmul_33, parameter_30 # pd_op.matmul: (1x21x384xf32) <- (1x21x384xf32, 384x384xf32) matmul_34 = paddle._C_ops.matmul(layer_norm_24, parameter_29, False, False) @@ -914,10 +893,11 @@ def forward( # pd_op.add: (1x21x384xf32) <- (1x21x384xf32, 384xf32) add_41 = paddle._C_ops.add(matmul_34, parameter_28) - del parameter_28 + del matmul_34, parameter_28 # pd_op.reshape: (1x21x12x32xf32) <- (1x21x384xf32, 4xi64) reshape_17 = paddle._C_ops.reshape(add_40, full_int_array_1) + del add_40 # pd_op.transpose: (1x12x21x32xf32) <- (1x21x12x32xf32) transpose_17 = paddle._C_ops.transpose(reshape_17, [0, 2, 1, 3]) @@ -925,6 +905,7 @@ def forward( # pd_op.reshape: (1x21x12x32xf32) <- (1x21x384xf32, 4xi64) reshape_18 = paddle._C_ops.reshape(add_41, full_int_array_1) + del add_41 # pd_op.transpose: (1x12x21x32xf32) <- (1x21x12x32xf32) transpose_18 = paddle._C_ops.transpose(reshape_18, [0, 2, 1, 3]) @@ -936,9 +917,11 @@ def forward( # pd_op.matmul: (1x12x21x21xf32) <- (1x12x21x32xf32, 1x12x21x32xf32) matmul_35 = paddle._C_ops.matmul(scale_6, transpose_17, False, True) + del scale_6, transpose_17 # pd_op.add: (1x12x21x21xf32) <- (1x12x21x21xf32, 1x1x1x21xf32) add_42 = paddle._C_ops.add(matmul_35, unsqueeze_0) + del matmul_35 # pd_op.softmax: (1x12x21x21xf32) <- (1x12x21x21xf32) softmax_4 = paddle._C_ops.softmax(add_42, -1) @@ -947,13 +930,15 @@ def forward( # pd_op.dropout: (1x12x21x21xf32, 1x12x21x21xui8) <- (1x12x21x21xf32, None, 1xf32) dropout_26, dropout_27 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_4, None, full_5, False, "upscale_in_train", 0, False + softmax_4, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_4 # pd_op.matmul: (1x12x21x32xf32) <- (1x12x21x21xf32, 1x12x21x32xf32) matmul_36 = paddle._C_ops.matmul(dropout_26, transpose_18, False, False) + del dropout_26, transpose_18 # pd_op.transpose: (1x21x12x32xf32) <- (1x12x21x32xf32) transpose_19 = paddle._C_ops.transpose(matmul_36, [0, 2, 1, 3]) @@ -961,19 +946,20 @@ def forward( # pd_op.reshape: (1x21x384xf32) <- (1x21x12x32xf32, 3xi64) reshape_19 = paddle._C_ops.reshape(transpose_19, full_int_array_2) + del transpose_19 # pd_op.matmul: (1x21x384xf32) <- (1x21x384xf32, 384x384xf32) matmul_37 = paddle._C_ops.matmul(reshape_19, parameter_27, False, False) - del parameter_27 + del parameter_27, reshape_19 # pd_op.add: (1x21x384xf32) <- (1x21x384xf32, 384xf32) add_43 = paddle._C_ops.add(matmul_37, parameter_26) - del parameter_26 + del matmul_37, parameter_26 # pd_op.dropout: (1x21x384xf32, 1x21x384xui8) <- (1x21x384xf32, None, 1xf32) dropout_28, dropout_29 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_43, None, full_5, False, "upscale_in_train", 0, False + add_43, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -981,6 +967,7 @@ def forward( # pd_op.add: (1x21x384xf32) <- (1x21x384xf32, 1x21x384xf32) add_44 = paddle._C_ops.add(layer_norm_24, dropout_28) + del dropout_28, layer_norm_24 # pd_op.layer_norm: (1x21x384xf32, 1x21xf32, 1x21xf32) <- (1x21x384xf32, 384xf32, 384xf32) layer_norm_27, layer_norm_28, layer_norm_29 = (lambda x, f: f(x))( @@ -989,7 +976,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_20, parameter_21 + del add_44, parameter_20, parameter_21 # pd_op.matmul: (1x21x1536xf32) <- (1x21x384xf32, 384x1536xf32) matmul_38 = paddle._C_ops.matmul(layer_norm_27, parameter_25, False, False) @@ -997,23 +984,24 @@ def forward( # pd_op.add: (1x21x1536xf32) <- (1x21x1536xf32, 1536xf32) add_45 = paddle._C_ops.add(matmul_38, parameter_24) - del parameter_24 + del matmul_38, parameter_24 # pd_op.gelu: (1x21x1536xf32) <- (1x21x1536xf32) gelu_4 = paddle._C_ops.gelu(add_45, False) + del add_45 # pd_op.matmul: (1x21x384xf32) <- (1x21x1536xf32, 1536x384xf32) matmul_39 = paddle._C_ops.matmul(gelu_4, parameter_23, False, False) - del parameter_23 + del gelu_4, parameter_23 # pd_op.add: (1x21x384xf32) <- (1x21x384xf32, 384xf32) add_46 = paddle._C_ops.add(matmul_39, parameter_22) - del parameter_22 + del matmul_39, parameter_22 # pd_op.dropout: (1x21x384xf32, 1x21x384xui8) <- (1x21x384xf32, None, 1xf32) dropout_30, dropout_31 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_46, None, full_5, False, "upscale_in_train", 0, False + add_46, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1021,6 +1009,7 @@ def forward( # pd_op.add: (1x21x384xf32) <- (1x21x384xf32, 1x21x384xf32) add_47 = paddle._C_ops.add(layer_norm_27, dropout_30) + del dropout_30, layer_norm_27 # pd_op.layer_norm: (1x21x384xf32, 1x21xf32, 1x21xf32) <- (1x21x384xf32, 384xf32, 384xf32) layer_norm_30, layer_norm_31, layer_norm_32 = (lambda x, f: f(x))( @@ -1029,7 +1018,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_18, parameter_19 + del add_47, parameter_18, parameter_19 # pd_op.matmul: (1x21x384xf32) <- (1x21x384xf32, 384x384xf32) matmul_40 = paddle._C_ops.matmul(layer_norm_30, parameter_17, False, False) @@ -1037,10 +1026,11 @@ def forward( # pd_op.add: (1x21x384xf32) <- (1x21x384xf32, 384xf32) add_48 = paddle._C_ops.add(matmul_40, parameter_16) - del parameter_16 + del matmul_40, parameter_16 # pd_op.reshape: (1x21x12x32xf32) <- (1x21x384xf32, 4xi64) reshape_20 = paddle._C_ops.reshape(add_48, full_int_array_1) + del add_48 # pd_op.transpose: (1x12x21x32xf32) <- (1x21x12x32xf32) transpose_20 = paddle._C_ops.transpose(reshape_20, [0, 2, 1, 3]) @@ -1052,7 +1042,7 @@ def forward( # pd_op.add: (1x21x384xf32) <- (1x21x384xf32, 384xf32) add_49 = paddle._C_ops.add(matmul_41, parameter_14) - del parameter_14 + del matmul_41, parameter_14 # pd_op.matmul: (1x21x384xf32) <- (1x21x384xf32, 384x384xf32) matmul_42 = paddle._C_ops.matmul(layer_norm_30, parameter_13, False, False) @@ -1060,10 +1050,11 @@ def forward( # pd_op.add: (1x21x384xf32) <- (1x21x384xf32, 384xf32) add_50 = paddle._C_ops.add(matmul_42, parameter_12) - del parameter_12 + del matmul_42, parameter_12 # pd_op.reshape: (1x21x12x32xf32) <- (1x21x384xf32, 4xi64) reshape_21 = paddle._C_ops.reshape(add_49, full_int_array_1) + del add_49 # pd_op.transpose: (1x12x21x32xf32) <- (1x21x12x32xf32) transpose_21 = paddle._C_ops.transpose(reshape_21, [0, 2, 1, 3]) @@ -1071,7 +1062,7 @@ def forward( # pd_op.reshape: (1x21x12x32xf32) <- (1x21x384xf32, 4xi64) reshape_22 = paddle._C_ops.reshape(add_50, full_int_array_1) - del full_int_array_1 + del add_50, full_int_array_1 # pd_op.transpose: (1x12x21x32xf32) <- (1x21x12x32xf32) transpose_22 = paddle._C_ops.transpose(reshape_22, [0, 2, 1, 3]) @@ -1079,13 +1070,15 @@ def forward( # pd_op.scale: (1x12x21x32xf32) <- (1x12x21x32xf32, 1xf32) scale_7 = paddle._C_ops.scale(transpose_20, full_6, float("0"), True) - del transpose_20 + del full_6, transpose_20 # pd_op.matmul: (1x12x21x21xf32) <- (1x12x21x32xf32, 1x12x21x32xf32) matmul_43 = paddle._C_ops.matmul(scale_7, transpose_21, False, True) + del scale_7, transpose_21 # pd_op.add: (1x12x21x21xf32) <- (1x12x21x21xf32, 1x1x1x21xf32) add_51 = paddle._C_ops.add(matmul_43, unsqueeze_0) + del matmul_43, unsqueeze_0 # pd_op.softmax: (1x12x21x21xf32) <- (1x12x21x21xf32) softmax_5 = paddle._C_ops.softmax(add_51, -1) @@ -1094,13 +1087,15 @@ def forward( # pd_op.dropout: (1x12x21x21xf32, 1x12x21x21xui8) <- (1x12x21x21xf32, None, 1xf32) dropout_32, dropout_33 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_5, None, full_5, False, "upscale_in_train", 0, False + softmax_5, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_5 # pd_op.matmul: (1x12x21x32xf32) <- (1x12x21x21xf32, 1x12x21x32xf32) matmul_44 = paddle._C_ops.matmul(dropout_32, transpose_22, False, False) + del dropout_32, transpose_22 # pd_op.transpose: (1x21x12x32xf32) <- (1x12x21x32xf32) transpose_23 = paddle._C_ops.transpose(matmul_44, [0, 2, 1, 3]) @@ -1108,20 +1103,20 @@ def forward( # pd_op.reshape: (1x21x384xf32) <- (1x21x12x32xf32, 3xi64) reshape_23 = paddle._C_ops.reshape(transpose_23, full_int_array_2) - del full_int_array_2 + del full_int_array_2, transpose_23 # pd_op.matmul: (1x21x384xf32) <- (1x21x384xf32, 384x384xf32) matmul_45 = paddle._C_ops.matmul(reshape_23, parameter_11, False, False) - del parameter_11 + del parameter_11, reshape_23 # pd_op.add: (1x21x384xf32) <- (1x21x384xf32, 384xf32) add_52 = paddle._C_ops.add(matmul_45, parameter_10) - del parameter_10 + del matmul_45, parameter_10 # pd_op.dropout: (1x21x384xf32, 1x21x384xui8) <- (1x21x384xf32, None, 1xf32) dropout_34, dropout_35 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_52, None, full_5, False, "upscale_in_train", 0, False + add_52, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1129,6 +1124,7 @@ def forward( # pd_op.add: (1x21x384xf32) <- (1x21x384xf32, 1x21x384xf32) add_53 = paddle._C_ops.add(layer_norm_30, dropout_34) + del dropout_34, layer_norm_30 # pd_op.layer_norm: (1x21x384xf32, 1x21xf32, 1x21xf32) <- (1x21x384xf32, 384xf32, 384xf32) layer_norm_33, layer_norm_34, layer_norm_35 = (lambda x, f: f(x))( @@ -1137,7 +1133,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_4, parameter_5 + del add_53, parameter_4, parameter_5 # pd_op.matmul: (1x21x1536xf32) <- (1x21x384xf32, 384x1536xf32) matmul_46 = paddle._C_ops.matmul(layer_norm_33, parameter_9, False, False) @@ -1145,30 +1141,32 @@ def forward( # pd_op.add: (1x21x1536xf32) <- (1x21x1536xf32, 1536xf32) add_54 = paddle._C_ops.add(matmul_46, parameter_8) - del parameter_8 + del matmul_46, parameter_8 # pd_op.gelu: (1x21x1536xf32) <- (1x21x1536xf32) gelu_5 = paddle._C_ops.gelu(add_54, False) + del add_54 # pd_op.matmul: (1x21x384xf32) <- (1x21x1536xf32, 1536x384xf32) matmul_47 = paddle._C_ops.matmul(gelu_5, parameter_7, False, False) - del parameter_7 + del gelu_5, parameter_7 # pd_op.add: (1x21x384xf32) <- (1x21x384xf32, 384xf32) add_55 = paddle._C_ops.add(matmul_47, parameter_6) - del parameter_6 + del matmul_47, parameter_6 # pd_op.dropout: (1x21x384xf32, 1x21x384xui8) <- (1x21x384xf32, None, 1xf32) dropout_36, dropout_37 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_55, None, full_5, False, "upscale_in_train", 0, False + add_55, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) - del add_55 + del add_55, full_5 # pd_op.add: (1x21x384xf32) <- (1x21x384xf32, 1x21x384xf32) add_56 = paddle._C_ops.add(layer_norm_33, dropout_36) + del dropout_36, layer_norm_33 # pd_op.layer_norm: (1x21x384xf32, 1x21xf32, 1x21xf32) <- (1x21x384xf32, 384xf32, 384xf32) layer_norm_36, layer_norm_37, layer_norm_38 = (lambda x, f: f(x))( @@ -1177,7 +1175,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_2, parameter_3 + del add_56, parameter_2, parameter_3 # pd_op.full_int_array: (1xi64) <- () full_int_array_3 = [0] @@ -1189,254 +1187,18 @@ def forward( slice_0 = paddle._C_ops.slice( layer_norm_36, [1], full_int_array_3, full_int_array_4, [1], [1] ) + del full_int_array_3, full_int_array_4, layer_norm_36 # pd_op.matmul: (1x384xf32) <- (1x384xf32, 384x384xf32) matmul_48 = paddle._C_ops.matmul(slice_0, parameter_1, False, False) - del parameter_1 + del parameter_1, slice_0 # pd_op.add: (1x384xf32) <- (1x384xf32, 384xf32) add_57 = paddle._C_ops.add(matmul_48, parameter_0) - del parameter_0 + del matmul_48, parameter_0 # pd_op.tanh: (1x384xf32) <- (1x384xf32) tanh_0 = paddle._C_ops.tanh(add_57) - del ( - add_0, - add_1, - add_11, - add_12, - add_13, - add_14, - add_17, - add_18, - add_2, - add_20, - add_21, - add_22, - add_23, - add_26, - add_27, - add_29, - add_3, - add_30, - add_31, - add_32, - add_35, - add_36, - add_38, - add_39, - add_4, - add_40, - add_41, - add_44, - add_45, - add_47, - add_48, - add_49, - add_5, - add_50, - add_53, - add_54, - add_56, - add_57, - add_8, - add_9, - assign_0, - assign_1, - assign_10, - assign_11, - assign_12, - assign_13, - assign_14, - assign_15, - assign_16, - assign_17, - assign_18, - assign_19, - assign_2, - assign_20, - assign_21, - assign_22, - assign_3, - assign_4, - assign_5, - assign_6, - assign_7, - assign_8, - assign_9, - dropout_0, - dropout_1, - dropout_10, - dropout_11, - dropout_12, - dropout_13, - dropout_14, - dropout_15, - dropout_16, - dropout_17, - dropout_18, - dropout_19, - dropout_2, - dropout_20, - dropout_21, - dropout_22, - dropout_23, - dropout_24, - dropout_25, - dropout_26, - dropout_27, - dropout_28, - dropout_29, - dropout_3, - dropout_30, - dropout_31, - dropout_32, - dropout_33, - dropout_34, - dropout_35, - dropout_36, - dropout_37, - dropout_4, - dropout_5, - dropout_6, - dropout_7, - dropout_8, - dropout_9, - embedding_0, - embedding_1, - embedding_2, - embedding_3, - full_5, - full_6, - full_int_array_3, - full_int_array_4, - gelu_0, - gelu_1, - gelu_2, - gelu_3, - gelu_4, - gelu_5, - layer_norm_1, - layer_norm_10, - layer_norm_11, - layer_norm_12, - layer_norm_13, - layer_norm_14, - layer_norm_15, - layer_norm_16, - layer_norm_17, - layer_norm_18, - layer_norm_19, - layer_norm_2, - layer_norm_20, - layer_norm_21, - layer_norm_22, - layer_norm_23, - layer_norm_24, - layer_norm_25, - layer_norm_26, - layer_norm_27, - layer_norm_28, - layer_norm_29, - layer_norm_3, - layer_norm_30, - layer_norm_31, - layer_norm_32, - layer_norm_33, - layer_norm_34, - layer_norm_35, - layer_norm_36, - layer_norm_37, - layer_norm_38, - layer_norm_4, - layer_norm_5, - layer_norm_6, - layer_norm_7, - layer_norm_8, - layer_norm_9, - matmul_0, - matmul_1, - matmul_10, - matmul_11, - matmul_13, - matmul_14, - matmul_15, - matmul_16, - matmul_17, - matmul_18, - matmul_19, - matmul_2, - matmul_21, - matmul_22, - matmul_23, - matmul_24, - matmul_25, - matmul_26, - matmul_27, - matmul_29, - matmul_3, - matmul_30, - matmul_31, - matmul_32, - matmul_33, - matmul_34, - matmul_35, - matmul_37, - matmul_38, - matmul_39, - matmul_40, - matmul_41, - matmul_42, - matmul_43, - matmul_45, - matmul_46, - matmul_47, - matmul_48, - matmul_5, - matmul_6, - matmul_7, - matmul_8, - matmul_9, - reshape_11, - reshape_15, - reshape_19, - reshape_23, - reshape_3, - reshape_7, - scale_1, - scale_2, - scale_3, - scale_4, - scale_5, - scale_6, - scale_7, - slice_0, - softmax_0, - softmax_1, - softmax_2, - softmax_3, - softmax_4, - softmax_5, - subtract_0, - transpose_1, - transpose_10, - transpose_11, - transpose_13, - transpose_14, - transpose_15, - transpose_17, - transpose_18, - transpose_19, - transpose_2, - transpose_21, - transpose_22, - transpose_23, - transpose_3, - transpose_5, - transpose_6, - transpose_7, - transpose_9, - unsqueeze_0, - ) + del add_57 return tanh_0 diff --git a/paddle_samples/PaddleNLP/rocketqa-mini-cross-encoder/weight_meta.py b/paddle_samples/PaddleNLP/rocketqa-mini-cross-encoder/weight_meta.py index 3b5464181..0402f3bd0 100644 --- a/paddle_samples/PaddleNLP/rocketqa-mini-cross-encoder/weight_meta.py +++ b/paddle_samples/PaddleNLP/rocketqa-mini-cross-encoder/weight_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_parameter_0: name = "parameter_0" + original_name = "linear_1721.b_0" shape = [384] dtype = "float32" min_val = float("-0.00256439") @@ -11,6 +12,7 @@ class Program_weight_tensor_parameter_0: class Program_weight_tensor_parameter_1: name = "parameter_1" + original_name = "linear_1721.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.081164") @@ -22,6 +24,7 @@ class Program_weight_tensor_parameter_1: class Program_weight_tensor_parameter_2: name = "parameter_2" + original_name = "layer_norm_593.b_0" shape = [384] dtype = "float32" min_val = float("-0.265784") @@ -33,6 +36,7 @@ class Program_weight_tensor_parameter_2: class Program_weight_tensor_parameter_3: name = "parameter_3" + original_name = "layer_norm_593.w_0" shape = [384] dtype = "float32" min_val = float("0.759393") @@ -44,6 +48,7 @@ class Program_weight_tensor_parameter_3: class Program_weight_tensor_parameter_4: name = "parameter_4" + original_name = "layer_norm_592.b_0" shape = [384] dtype = "float32" min_val = float("-1.178") @@ -55,6 +60,7 @@ class Program_weight_tensor_parameter_4: class Program_weight_tensor_parameter_5: name = "parameter_5" + original_name = "layer_norm_592.w_0" shape = [384] dtype = "float32" min_val = float("0.389675") @@ -66,6 +72,7 @@ class Program_weight_tensor_parameter_5: class Program_weight_tensor_parameter_6: name = "parameter_6" + original_name = "linear_1720.b_0" shape = [384] dtype = "float32" min_val = float("-0.199491") @@ -77,6 +84,7 @@ class Program_weight_tensor_parameter_6: class Program_weight_tensor_parameter_7: name = "parameter_7" + original_name = "linear_1720.w_0" shape = [1536, 384] dtype = "float32" min_val = float("-0.968364") @@ -88,6 +96,7 @@ class Program_weight_tensor_parameter_7: class Program_weight_tensor_parameter_8: name = "parameter_8" + original_name = "linear_1719.b_0" shape = [1536] dtype = "float32" min_val = float("-0.605898") @@ -99,6 +108,7 @@ class Program_weight_tensor_parameter_8: class Program_weight_tensor_parameter_9: name = "parameter_9" + original_name = "linear_1719.w_0" shape = [384, 1536] dtype = "float32" min_val = float("-0.60357") @@ -110,6 +120,7 @@ class Program_weight_tensor_parameter_9: class Program_weight_tensor_parameter_10: name = "parameter_10" + original_name = "linear_1718.b_0" shape = [384] dtype = "float32" min_val = float("-0.303614") @@ -121,6 +132,7 @@ class Program_weight_tensor_parameter_10: class Program_weight_tensor_parameter_11: name = "parameter_11" + original_name = "linear_1718.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.684151") @@ -132,6 +144,7 @@ class Program_weight_tensor_parameter_11: class Program_weight_tensor_parameter_12: name = "parameter_12" + original_name = "linear_1717.b_0" shape = [384] dtype = "float32" min_val = float("-0.130441") @@ -143,6 +156,7 @@ class Program_weight_tensor_parameter_12: class Program_weight_tensor_parameter_13: name = "parameter_13" + original_name = "linear_1717.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.299324") @@ -154,6 +168,7 @@ class Program_weight_tensor_parameter_13: class Program_weight_tensor_parameter_14: name = "parameter_14" + original_name = "linear_1716.b_0" shape = [384] dtype = "float32" min_val = float("-0.0260576") @@ -165,6 +180,7 @@ class Program_weight_tensor_parameter_14: class Program_weight_tensor_parameter_15: name = "parameter_15" + original_name = "linear_1716.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.334577") @@ -176,6 +192,7 @@ class Program_weight_tensor_parameter_15: class Program_weight_tensor_parameter_16: name = "parameter_16" + original_name = "linear_1715.b_0" shape = [384] dtype = "float32" min_val = float("-0.510457") @@ -187,6 +204,7 @@ class Program_weight_tensor_parameter_16: class Program_weight_tensor_parameter_17: name = "parameter_17" + original_name = "linear_1715.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.434075") @@ -198,6 +216,7 @@ class Program_weight_tensor_parameter_17: class Program_weight_tensor_parameter_18: name = "parameter_18" + original_name = "layer_norm_591.b_0" shape = [384] dtype = "float32" min_val = float("-0.497758") @@ -209,6 +228,7 @@ class Program_weight_tensor_parameter_18: class Program_weight_tensor_parameter_19: name = "parameter_19" + original_name = "layer_norm_591.w_0" shape = [384] dtype = "float32" min_val = float("0.633216") @@ -220,6 +240,7 @@ class Program_weight_tensor_parameter_19: class Program_weight_tensor_parameter_20: name = "parameter_20" + original_name = "layer_norm_590.b_0" shape = [384] dtype = "float32" min_val = float("-0.927368") @@ -231,6 +252,7 @@ class Program_weight_tensor_parameter_20: class Program_weight_tensor_parameter_21: name = "parameter_21" + original_name = "layer_norm_590.w_0" shape = [384] dtype = "float32" min_val = float("0.675674") @@ -242,6 +264,7 @@ class Program_weight_tensor_parameter_21: class Program_weight_tensor_parameter_22: name = "parameter_22" + original_name = "linear_1714.b_0" shape = [384] dtype = "float32" min_val = float("-0.185525") @@ -253,6 +276,7 @@ class Program_weight_tensor_parameter_22: class Program_weight_tensor_parameter_23: name = "parameter_23" + original_name = "linear_1714.w_0" shape = [1536, 384] dtype = "float32" min_val = float("-1.3537") @@ -264,6 +288,7 @@ class Program_weight_tensor_parameter_23: class Program_weight_tensor_parameter_24: name = "parameter_24" + original_name = "linear_1713.b_0" shape = [1536] dtype = "float32" min_val = float("-0.568222") @@ -275,6 +300,7 @@ class Program_weight_tensor_parameter_24: class Program_weight_tensor_parameter_25: name = "parameter_25" + original_name = "linear_1713.w_0" shape = [384, 1536] dtype = "float32" min_val = float("-0.507165") @@ -286,6 +312,7 @@ class Program_weight_tensor_parameter_25: class Program_weight_tensor_parameter_26: name = "parameter_26" + original_name = "linear_1712.b_0" shape = [384] dtype = "float32" min_val = float("-0.245455") @@ -297,6 +324,7 @@ class Program_weight_tensor_parameter_26: class Program_weight_tensor_parameter_27: name = "parameter_27" + original_name = "linear_1712.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.412449") @@ -308,6 +336,7 @@ class Program_weight_tensor_parameter_27: class Program_weight_tensor_parameter_28: name = "parameter_28" + original_name = "linear_1711.b_0" shape = [384] dtype = "float32" min_val = float("-0.161156") @@ -319,6 +348,7 @@ class Program_weight_tensor_parameter_28: class Program_weight_tensor_parameter_29: name = "parameter_29" + original_name = "linear_1711.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.342785") @@ -330,6 +360,7 @@ class Program_weight_tensor_parameter_29: class Program_weight_tensor_parameter_30: name = "parameter_30" + original_name = "linear_1710.b_0" shape = [384] dtype = "float32" min_val = float("-0.00574247") @@ -341,6 +372,7 @@ class Program_weight_tensor_parameter_30: class Program_weight_tensor_parameter_31: name = "parameter_31" + original_name = "linear_1710.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.361672") @@ -352,6 +384,7 @@ class Program_weight_tensor_parameter_31: class Program_weight_tensor_parameter_32: name = "parameter_32" + original_name = "linear_1709.b_0" shape = [384] dtype = "float32" min_val = float("-0.586434") @@ -363,6 +396,7 @@ class Program_weight_tensor_parameter_32: class Program_weight_tensor_parameter_33: name = "parameter_33" + original_name = "linear_1709.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.333574") @@ -374,6 +408,7 @@ class Program_weight_tensor_parameter_33: class Program_weight_tensor_parameter_34: name = "parameter_34" + original_name = "layer_norm_589.b_0" shape = [384] dtype = "float32" min_val = float("-0.718376") @@ -385,6 +420,7 @@ class Program_weight_tensor_parameter_34: class Program_weight_tensor_parameter_35: name = "parameter_35" + original_name = "layer_norm_589.w_0" shape = [384] dtype = "float32" min_val = float("0.825445") @@ -396,6 +432,7 @@ class Program_weight_tensor_parameter_35: class Program_weight_tensor_parameter_36: name = "parameter_36" + original_name = "layer_norm_588.b_0" shape = [384] dtype = "float32" min_val = float("-0.964334") @@ -407,6 +444,7 @@ class Program_weight_tensor_parameter_36: class Program_weight_tensor_parameter_37: name = "parameter_37" + original_name = "layer_norm_588.w_0" shape = [384] dtype = "float32" min_val = float("0.701612") @@ -418,6 +456,7 @@ class Program_weight_tensor_parameter_37: class Program_weight_tensor_parameter_38: name = "parameter_38" + original_name = "linear_1708.b_0" shape = [384] dtype = "float32" min_val = float("-0.154809") @@ -429,6 +468,7 @@ class Program_weight_tensor_parameter_38: class Program_weight_tensor_parameter_39: name = "parameter_39" + original_name = "linear_1708.w_0" shape = [1536, 384] dtype = "float32" min_val = float("-1.73948") @@ -440,6 +480,7 @@ class Program_weight_tensor_parameter_39: class Program_weight_tensor_parameter_40: name = "parameter_40" + original_name = "linear_1707.b_0" shape = [1536] dtype = "float32" min_val = float("-0.526366") @@ -451,6 +492,7 @@ class Program_weight_tensor_parameter_40: class Program_weight_tensor_parameter_41: name = "parameter_41" + original_name = "linear_1707.w_0" shape = [384, 1536] dtype = "float32" min_val = float("-0.477368") @@ -462,6 +504,7 @@ class Program_weight_tensor_parameter_41: class Program_weight_tensor_parameter_42: name = "parameter_42" + original_name = "linear_1706.b_0" shape = [384] dtype = "float32" min_val = float("-0.342679") @@ -473,6 +516,7 @@ class Program_weight_tensor_parameter_42: class Program_weight_tensor_parameter_43: name = "parameter_43" + original_name = "linear_1706.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.563855") @@ -484,6 +528,7 @@ class Program_weight_tensor_parameter_43: class Program_weight_tensor_parameter_44: name = "parameter_44" + original_name = "linear_1705.b_0" shape = [384] dtype = "float32" min_val = float("-0.133465") @@ -495,6 +540,7 @@ class Program_weight_tensor_parameter_44: class Program_weight_tensor_parameter_45: name = "parameter_45" + original_name = "linear_1705.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.282594") @@ -506,6 +552,7 @@ class Program_weight_tensor_parameter_45: class Program_weight_tensor_parameter_46: name = "parameter_46" + original_name = "linear_1704.b_0" shape = [384] dtype = "float32" min_val = float("-0.0032676") @@ -517,6 +564,7 @@ class Program_weight_tensor_parameter_46: class Program_weight_tensor_parameter_47: name = "parameter_47" + original_name = "linear_1704.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.305349") @@ -528,6 +576,7 @@ class Program_weight_tensor_parameter_47: class Program_weight_tensor_parameter_48: name = "parameter_48" + original_name = "linear_1703.b_0" shape = [384] dtype = "float32" min_val = float("-0.71351") @@ -539,6 +588,7 @@ class Program_weight_tensor_parameter_48: class Program_weight_tensor_parameter_49: name = "parameter_49" + original_name = "linear_1703.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.354357") @@ -550,6 +600,7 @@ class Program_weight_tensor_parameter_49: class Program_weight_tensor_parameter_50: name = "parameter_50" + original_name = "layer_norm_587.b_0" shape = [384] dtype = "float32" min_val = float("-0.800337") @@ -561,6 +612,7 @@ class Program_weight_tensor_parameter_50: class Program_weight_tensor_parameter_51: name = "parameter_51" + original_name = "layer_norm_587.w_0" shape = [384] dtype = "float32" min_val = float("0.917515") @@ -572,6 +624,7 @@ class Program_weight_tensor_parameter_51: class Program_weight_tensor_parameter_52: name = "parameter_52" + original_name = "layer_norm_586.b_0" shape = [384] dtype = "float32" min_val = float("-0.735475") @@ -583,6 +636,7 @@ class Program_weight_tensor_parameter_52: class Program_weight_tensor_parameter_53: name = "parameter_53" + original_name = "layer_norm_586.w_0" shape = [384] dtype = "float32" min_val = float("0.759203") @@ -594,6 +648,7 @@ class Program_weight_tensor_parameter_53: class Program_weight_tensor_parameter_54: name = "parameter_54" + original_name = "linear_1702.b_0" shape = [384] dtype = "float32" min_val = float("-0.263219") @@ -605,6 +660,7 @@ class Program_weight_tensor_parameter_54: class Program_weight_tensor_parameter_55: name = "parameter_55" + original_name = "linear_1702.w_0" shape = [1536, 384] dtype = "float32" min_val = float("-1.74112") @@ -616,6 +672,7 @@ class Program_weight_tensor_parameter_55: class Program_weight_tensor_parameter_56: name = "parameter_56" + original_name = "linear_1701.b_0" shape = [1536] dtype = "float32" min_val = float("-0.473519") @@ -627,6 +684,7 @@ class Program_weight_tensor_parameter_56: class Program_weight_tensor_parameter_57: name = "parameter_57" + original_name = "linear_1701.w_0" shape = [384, 1536] dtype = "float32" min_val = float("-0.643855") @@ -638,6 +696,7 @@ class Program_weight_tensor_parameter_57: class Program_weight_tensor_parameter_58: name = "parameter_58" + original_name = "linear_1700.b_0" shape = [384] dtype = "float32" min_val = float("-0.250901") @@ -649,6 +708,7 @@ class Program_weight_tensor_parameter_58: class Program_weight_tensor_parameter_59: name = "parameter_59" + original_name = "linear_1700.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.483821") @@ -660,6 +720,7 @@ class Program_weight_tensor_parameter_59: class Program_weight_tensor_parameter_60: name = "parameter_60" + original_name = "linear_1699.b_0" shape = [384] dtype = "float32" min_val = float("-0.121611") @@ -671,6 +732,7 @@ class Program_weight_tensor_parameter_60: class Program_weight_tensor_parameter_61: name = "parameter_61" + original_name = "linear_1699.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.295654") @@ -682,6 +744,7 @@ class Program_weight_tensor_parameter_61: class Program_weight_tensor_parameter_62: name = "parameter_62" + original_name = "linear_1698.b_0" shape = [384] dtype = "float32" min_val = float("-0.000948381") @@ -693,6 +756,7 @@ class Program_weight_tensor_parameter_62: class Program_weight_tensor_parameter_63: name = "parameter_63" + original_name = "linear_1698.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.714976") @@ -704,6 +768,7 @@ class Program_weight_tensor_parameter_63: class Program_weight_tensor_parameter_64: name = "parameter_64" + original_name = "linear_1697.b_0" shape = [384] dtype = "float32" min_val = float("-0.652352") @@ -715,6 +780,7 @@ class Program_weight_tensor_parameter_64: class Program_weight_tensor_parameter_65: name = "parameter_65" + original_name = "linear_1697.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.525362") @@ -726,6 +792,7 @@ class Program_weight_tensor_parameter_65: class Program_weight_tensor_parameter_66: name = "parameter_66" + original_name = "layer_norm_585.b_0" shape = [384] dtype = "float32" min_val = float("-0.547195") @@ -737,6 +804,7 @@ class Program_weight_tensor_parameter_66: class Program_weight_tensor_parameter_67: name = "parameter_67" + original_name = "layer_norm_585.w_0" shape = [384] dtype = "float32" min_val = float("0.992886") @@ -748,6 +816,7 @@ class Program_weight_tensor_parameter_67: class Program_weight_tensor_parameter_68: name = "parameter_68" + original_name = "layer_norm_584.b_0" shape = [384] dtype = "float32" min_val = float("-0.896357") @@ -759,6 +828,7 @@ class Program_weight_tensor_parameter_68: class Program_weight_tensor_parameter_69: name = "parameter_69" + original_name = "layer_norm_584.w_0" shape = [384] dtype = "float32" min_val = float("0.83989") @@ -770,6 +840,7 @@ class Program_weight_tensor_parameter_69: class Program_weight_tensor_parameter_70: name = "parameter_70" + original_name = "linear_1696.b_0" shape = [384] dtype = "float32" min_val = float("-0.203248") @@ -781,6 +852,7 @@ class Program_weight_tensor_parameter_70: class Program_weight_tensor_parameter_71: name = "parameter_71" + original_name = "linear_1696.w_0" shape = [1536, 384] dtype = "float32" min_val = float("-1.37074") @@ -792,6 +864,7 @@ class Program_weight_tensor_parameter_71: class Program_weight_tensor_parameter_72: name = "parameter_72" + original_name = "linear_1695.b_0" shape = [1536] dtype = "float32" min_val = float("-0.691778") @@ -803,6 +876,7 @@ class Program_weight_tensor_parameter_72: class Program_weight_tensor_parameter_73: name = "parameter_73" + original_name = "linear_1695.w_0" shape = [384, 1536] dtype = "float32" min_val = float("-0.55587") @@ -814,6 +888,7 @@ class Program_weight_tensor_parameter_73: class Program_weight_tensor_parameter_74: name = "parameter_74" + original_name = "linear_1694.b_0" shape = [384] dtype = "float32" min_val = float("-0.288079") @@ -825,6 +900,7 @@ class Program_weight_tensor_parameter_74: class Program_weight_tensor_parameter_75: name = "parameter_75" + original_name = "linear_1694.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.268353") @@ -836,6 +912,7 @@ class Program_weight_tensor_parameter_75: class Program_weight_tensor_parameter_76: name = "parameter_76" + original_name = "linear_1693.b_0" shape = [384] dtype = "float32" min_val = float("-0.116638") @@ -847,6 +924,7 @@ class Program_weight_tensor_parameter_76: class Program_weight_tensor_parameter_77: name = "parameter_77" + original_name = "linear_1693.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.225877") @@ -858,6 +936,7 @@ class Program_weight_tensor_parameter_77: class Program_weight_tensor_parameter_78: name = "parameter_78" + original_name = "linear_1692.b_0" shape = [384] dtype = "float32" min_val = float("-0.00167251") @@ -869,6 +948,7 @@ class Program_weight_tensor_parameter_78: class Program_weight_tensor_parameter_79: name = "parameter_79" + original_name = "linear_1692.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.432844") @@ -880,6 +960,7 @@ class Program_weight_tensor_parameter_79: class Program_weight_tensor_parameter_80: name = "parameter_80" + original_name = "linear_1691.b_0" shape = [384] dtype = "float32" min_val = float("-0.702922") @@ -891,6 +972,7 @@ class Program_weight_tensor_parameter_80: class Program_weight_tensor_parameter_81: name = "parameter_81" + original_name = "linear_1691.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.403141") @@ -902,6 +984,7 @@ class Program_weight_tensor_parameter_81: class Program_weight_tensor_parameter_82: name = "parameter_82" + original_name = "layer_norm_583.b_0" shape = [384] dtype = "float32" min_val = float("-1.21378") @@ -913,6 +996,7 @@ class Program_weight_tensor_parameter_82: class Program_weight_tensor_parameter_83: name = "parameter_83" + original_name = "layer_norm_583.w_0" shape = [384] dtype = "float32" min_val = float("0.99551") @@ -924,6 +1008,7 @@ class Program_weight_tensor_parameter_83: class Program_weight_tensor_parameter_84: name = "parameter_84" + original_name = "layer_norm_582.b_0" shape = [384] dtype = "float32" min_val = float("-3.99256") @@ -935,6 +1020,7 @@ class Program_weight_tensor_parameter_84: class Program_weight_tensor_parameter_85: name = "parameter_85" + original_name = "layer_norm_582.w_0" shape = [384] dtype = "float32" min_val = float("0.417885") @@ -946,6 +1032,7 @@ class Program_weight_tensor_parameter_85: class Program_weight_tensor_parameter_86: name = "parameter_86" + original_name = "linear_1690.b_0" shape = [384] dtype = "float32" min_val = float("-0.266796") @@ -957,6 +1044,7 @@ class Program_weight_tensor_parameter_86: class Program_weight_tensor_parameter_87: name = "parameter_87" + original_name = "linear_1690.w_0" shape = [1536, 384] dtype = "float32" min_val = float("-1.62381") @@ -968,6 +1056,7 @@ class Program_weight_tensor_parameter_87: class Program_weight_tensor_parameter_88: name = "parameter_88" + original_name = "linear_1689.b_0" shape = [1536] dtype = "float32" min_val = float("-0.540665") @@ -979,6 +1068,7 @@ class Program_weight_tensor_parameter_88: class Program_weight_tensor_parameter_89: name = "parameter_89" + original_name = "linear_1689.w_0" shape = [384, 1536] dtype = "float32" min_val = float("-0.381177") @@ -990,6 +1080,7 @@ class Program_weight_tensor_parameter_89: class Program_weight_tensor_parameter_90: name = "parameter_90" + original_name = "linear_1688.b_0" shape = [384] dtype = "float32" min_val = float("-0.34282") @@ -1001,6 +1092,7 @@ class Program_weight_tensor_parameter_90: class Program_weight_tensor_parameter_91: name = "parameter_91" + original_name = "linear_1688.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.304494") @@ -1012,6 +1104,7 @@ class Program_weight_tensor_parameter_91: class Program_weight_tensor_parameter_92: name = "parameter_92" + original_name = "linear_1687.b_0" shape = [384] dtype = "float32" min_val = float("-0.392436") @@ -1023,6 +1116,7 @@ class Program_weight_tensor_parameter_92: class Program_weight_tensor_parameter_93: name = "parameter_93" + original_name = "linear_1687.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.350792") @@ -1034,6 +1128,7 @@ class Program_weight_tensor_parameter_93: class Program_weight_tensor_parameter_94: name = "parameter_94" + original_name = "linear_1686.b_0" shape = [384] dtype = "float32" min_val = float("-0.000636372") @@ -1045,6 +1140,7 @@ class Program_weight_tensor_parameter_94: class Program_weight_tensor_parameter_95: name = "parameter_95" + original_name = "linear_1686.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.456937") @@ -1056,6 +1152,7 @@ class Program_weight_tensor_parameter_95: class Program_weight_tensor_parameter_96: name = "parameter_96" + original_name = "linear_1685.b_0" shape = [384] dtype = "float32" min_val = float("-1.24257") @@ -1067,6 +1164,7 @@ class Program_weight_tensor_parameter_96: class Program_weight_tensor_parameter_97: name = "parameter_97" + original_name = "linear_1685.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.621652") @@ -1078,6 +1176,7 @@ class Program_weight_tensor_parameter_97: class Program_weight_tensor_parameter_98: name = "parameter_98" + original_name = "layer_norm_581.b_0" shape = [384] dtype = "float32" min_val = float("-0.324497") @@ -1089,6 +1188,7 @@ class Program_weight_tensor_parameter_98: class Program_weight_tensor_parameter_99: name = "parameter_99" + original_name = "layer_norm_581.w_0" shape = [384] dtype = "float32" min_val = float("0.148805") @@ -1100,6 +1200,7 @@ class Program_weight_tensor_parameter_99: class Program_weight_tensor_parameter_100: name = "parameter_100" + original_name = "embedding_105.w_0" shape = [16, 384] dtype = "float32" min_val = float("-0.48256") @@ -1111,6 +1212,7 @@ class Program_weight_tensor_parameter_100: class Program_weight_tensor_parameter_101: name = "parameter_101" + original_name = "embedding_104.w_0" shape = [4, 384] dtype = "float32" min_val = float("-0.328651") @@ -1122,6 +1224,7 @@ class Program_weight_tensor_parameter_101: class Program_weight_tensor_parameter_102: name = "parameter_102" + original_name = "embedding_103.w_0" shape = [2048, 384] dtype = "float32" min_val = float("-0.345758") @@ -1133,10 +1236,11 @@ class Program_weight_tensor_parameter_102: class Program_weight_tensor_parameter_103: name = "parameter_103" + original_name = "embedding_102.w_0" shape = [40000, 384] dtype = "float32" min_val = float("-0.589332") max_val = float("0.572457") - mean = float("1.23116e-05") + mean = float("1.23115e-05") std = float("0.0360502") data = None diff --git a/paddle_samples/PaddleNLP/skep_ernie_1.0_large_ch/input_meta.py b/paddle_samples/PaddleNLP/skep_ernie_1.0_large_ch/input_meta.py index 97eb8a799..a4123e481 100644 --- a/paddle_samples/PaddleNLP/skep_ernie_1.0_large_ch/input_meta.py +++ b/paddle_samples/PaddleNLP/skep_ernie_1.0_large_ch/input_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_data_0: name = "data_0" + original_name = "input_ids" shape = [1, 11] dtype = "int64" data = [1, 811, 1257, 175, 29, 502, 130, 706, 3619, 12046, 2] @@ -7,6 +8,7 @@ class Program_weight_tensor_data_0: class Program_weight_tensor_data_1: name = "data_1" + original_name = "token_type_ids" shape = [1, 11] dtype = "int64" data = [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0] diff --git a/paddle_samples/PaddleNLP/skep_ernie_1.0_large_ch/model.py b/paddle_samples/PaddleNLP/skep_ernie_1.0_large_ch/model.py index 61b35494f..a2575ba0e 100644 --- a/paddle_samples/PaddleNLP/skep_ernie_1.0_large_ch/model.py +++ b/paddle_samples/PaddleNLP/skep_ernie_1.0_large_ch/model.py @@ -457,10 +457,11 @@ def forward( # pd_op.embedding: (1x11x1024xf32) <- (1x11xi64, 512x1024xf32) embedding_1 = paddle._C_ops.embedding(subtract_0, parameter_389, -1, False) - del parameter_389 + del parameter_389, subtract_0 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_0 = paddle._C_ops.add(embedding_0, embedding_1) + del embedding_0, embedding_1 # pd_op.embedding: (1x11x1024xf32) <- (1x11xi64, 4x1024xf32) embedding_2 = paddle._C_ops.embedding(data_1, parameter_388, -1, False) @@ -468,6 +469,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_1 = paddle._C_ops.add(add_0, embedding_2) + del add_0, embedding_2 # pd_op.layer_norm: (1x11x1024xf32, 1x11xf32, 1x11xf32) <- (1x11x1024xf32, 1024xf32, 1024xf32) layer_norm_0, layer_norm_1, layer_norm_2 = (lambda x, f: f(x))( @@ -476,233 +478,17 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_386, parameter_387 + del add_1, parameter_386, parameter_387 # pd_op.full: (1xf32) <- () full_4 = paddle._C_ops.full( [1], float("0.1"), paddle.float32, paddle.core.CPUPlace() ) - # pd_op.assign: (1xf32) <- (1xf32) - assign_0 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_1 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_2 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_3 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_4 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_5 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_6 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_7 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_8 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_9 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_10 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_11 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_12 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_13 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_14 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_15 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_16 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_17 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_18 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_19 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_20 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_21 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_22 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_23 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_24 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_25 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_26 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_27 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_28 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_29 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_30 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_31 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_32 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_33 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_34 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_35 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_36 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_37 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_38 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_39 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_40 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_41 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_42 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_43 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_44 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_45 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_46 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_47 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_48 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_49 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_50 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_51 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_52 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_53 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_54 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_55 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_56 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_57 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_58 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_59 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_60 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_61 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_62 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_63 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_64 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_65 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_66 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_67 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_68 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_69 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_70 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_71 = full_4 - # pd_op.dropout: (1x11x1024xf32, 1x11x1024xui8) <- (1x11x1024xf32, None, 1xf32) dropout_0, dropout_1 = (lambda x, f: f(x))( paddle._C_ops.dropout( - layer_norm_0, None, full_4, False, "upscale_in_train", 0, False + layer_norm_0, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -714,13 +500,14 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_2 = paddle._C_ops.add(matmul_0, parameter_384) - del parameter_384 + del matmul_0, parameter_384 # pd_op.full_int_array: (4xi64) <- () full_int_array_1 = [0, 0, 16, 64] # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_0 = paddle._C_ops.reshape(add_2, full_int_array_1) + del add_2 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_0 = paddle._C_ops.transpose(reshape_0, [0, 2, 1, 3]) @@ -732,7 +519,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_3 = paddle._C_ops.add(matmul_1, parameter_382) - del parameter_382 + del matmul_1, parameter_382 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_2 = paddle._C_ops.matmul(dropout_0, parameter_381, False, False) @@ -740,10 +527,11 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_4 = paddle._C_ops.add(matmul_2, parameter_380) - del parameter_380 + del matmul_2, parameter_380 # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_1 = paddle._C_ops.reshape(add_3, full_int_array_1) + del add_3 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_1 = paddle._C_ops.transpose(reshape_1, [0, 2, 1, 3]) @@ -751,6 +539,7 @@ def forward( # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_2 = paddle._C_ops.reshape(add_4, full_int_array_1) + del add_4 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_2 = paddle._C_ops.transpose(reshape_2, [0, 2, 1, 3]) @@ -761,84 +550,17 @@ def forward( [1], float("0.125"), paddle.float32, paddle.core.CPUPlace() ) - # pd_op.assign: (1xf32) <- (1xf32) - assign_72 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_73 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_74 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_75 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_76 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_77 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_78 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_79 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_80 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_81 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_82 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_83 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_84 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_85 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_86 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_87 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_88 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_89 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_90 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_91 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_92 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_93 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_94 = full_5 - # pd_op.scale: (1x16x11x64xf32) <- (1x16x11x64xf32, 1xf32) scale_1 = paddle._C_ops.scale(transpose_0, full_5, float("0"), True) del transpose_0 # pd_op.matmul: (1x16x11x11xf32) <- (1x16x11x64xf32, 1x16x11x64xf32) matmul_3 = paddle._C_ops.matmul(scale_1, transpose_1, False, True) + del scale_1, transpose_1 # pd_op.add: (1x16x11x11xf32) <- (1x16x11x11xf32, 1x1x1x11xf32) add_5 = paddle._C_ops.add(matmul_3, unsqueeze_0) + del matmul_3 # pd_op.softmax: (1x16x11x11xf32) <- (1x16x11x11xf32) softmax_0 = paddle._C_ops.softmax(add_5, -1) @@ -847,13 +569,15 @@ def forward( # pd_op.dropout: (1x16x11x11xf32, 1x16x11x11xui8) <- (1x16x11x11xf32, None, 1xf32) dropout_2, dropout_3 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_0, None, full_4, False, "upscale_in_train", 0, False + softmax_0, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_0 # pd_op.matmul: (1x16x11x64xf32) <- (1x16x11x11xf32, 1x16x11x64xf32) matmul_4 = paddle._C_ops.matmul(dropout_2, transpose_2, False, False) + del dropout_2, transpose_2 # pd_op.transpose: (1x11x16x64xf32) <- (1x16x11x64xf32) transpose_3 = paddle._C_ops.transpose(matmul_4, [0, 2, 1, 3]) @@ -864,19 +588,20 @@ def forward( # pd_op.reshape: (1x11x1024xf32) <- (1x11x16x64xf32, 3xi64) reshape_3 = paddle._C_ops.reshape(transpose_3, full_int_array_2) + del transpose_3 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_5 = paddle._C_ops.matmul(reshape_3, parameter_379, False, False) - del parameter_379 + del parameter_379, reshape_3 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_6 = paddle._C_ops.add(matmul_5, parameter_378) - del parameter_378 + del matmul_5, parameter_378 # pd_op.dropout: (1x11x1024xf32, 1x11x1024xui8) <- (1x11x1024xf32, None, 1xf32) dropout_4, dropout_5 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_6, None, full_4, False, "upscale_in_train", 0, False + add_6, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -884,6 +609,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_7 = paddle._C_ops.add(dropout_0, dropout_4) + del dropout_0, dropout_4 # pd_op.layer_norm: (1x11x1024xf32, 1x11xf32, 1x11xf32) <- (1x11x1024xf32, 1024xf32, 1024xf32) layer_norm_3, layer_norm_4, layer_norm_5 = (lambda x, f: f(x))( @@ -892,7 +618,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_372, parameter_373 + del add_7, parameter_372, parameter_373 # pd_op.matmul: (1x11x4096xf32) <- (1x11x1024xf32, 1024x4096xf32) matmul_6 = paddle._C_ops.matmul(layer_norm_3, parameter_377, False, False) @@ -900,7 +626,7 @@ def forward( # pd_op.add: (1x11x4096xf32) <- (1x11x4096xf32, 4096xf32) add_8 = paddle._C_ops.add(matmul_6, parameter_376) - del parameter_376 + del matmul_6, parameter_376 # pd_op.relu: (1x11x4096xf32) <- (1x11x4096xf32) relu_0 = paddle._C_ops.relu(add_8) @@ -908,16 +634,16 @@ def forward( # pd_op.matmul: (1x11x1024xf32) <- (1x11x4096xf32, 4096x1024xf32) matmul_7 = paddle._C_ops.matmul(relu_0, parameter_375, False, False) - del parameter_375 + del parameter_375, relu_0 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_9 = paddle._C_ops.add(matmul_7, parameter_374) - del parameter_374 + del matmul_7, parameter_374 # pd_op.dropout: (1x11x1024xf32, 1x11x1024xui8) <- (1x11x1024xf32, None, 1xf32) dropout_6, dropout_7 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_9, None, full_4, False, "upscale_in_train", 0, False + add_9, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -925,6 +651,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_10 = paddle._C_ops.add(layer_norm_3, dropout_6) + del dropout_6, layer_norm_3 # pd_op.layer_norm: (1x11x1024xf32, 1x11xf32, 1x11xf32) <- (1x11x1024xf32, 1024xf32, 1024xf32) layer_norm_6, layer_norm_7, layer_norm_8 = (lambda x, f: f(x))( @@ -933,7 +660,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_370, parameter_371 + del add_10, parameter_370, parameter_371 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_8 = paddle._C_ops.matmul(layer_norm_6, parameter_369, False, False) @@ -941,10 +668,11 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_11 = paddle._C_ops.add(matmul_8, parameter_368) - del parameter_368 + del matmul_8, parameter_368 # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_4 = paddle._C_ops.reshape(add_11, full_int_array_1) + del add_11 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_4 = paddle._C_ops.transpose(reshape_4, [0, 2, 1, 3]) @@ -956,7 +684,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_12 = paddle._C_ops.add(matmul_9, parameter_366) - del parameter_366 + del matmul_9, parameter_366 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_10 = paddle._C_ops.matmul(layer_norm_6, parameter_365, False, False) @@ -964,10 +692,11 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_13 = paddle._C_ops.add(matmul_10, parameter_364) - del parameter_364 + del matmul_10, parameter_364 # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_5 = paddle._C_ops.reshape(add_12, full_int_array_1) + del add_12 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_5 = paddle._C_ops.transpose(reshape_5, [0, 2, 1, 3]) @@ -975,6 +704,7 @@ def forward( # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_6 = paddle._C_ops.reshape(add_13, full_int_array_1) + del add_13 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_6 = paddle._C_ops.transpose(reshape_6, [0, 2, 1, 3]) @@ -986,9 +716,11 @@ def forward( # pd_op.matmul: (1x16x11x11xf32) <- (1x16x11x64xf32, 1x16x11x64xf32) matmul_11 = paddle._C_ops.matmul(scale_2, transpose_5, False, True) + del scale_2, transpose_5 # pd_op.add: (1x16x11x11xf32) <- (1x16x11x11xf32, 1x1x1x11xf32) add_14 = paddle._C_ops.add(matmul_11, unsqueeze_0) + del matmul_11 # pd_op.softmax: (1x16x11x11xf32) <- (1x16x11x11xf32) softmax_1 = paddle._C_ops.softmax(add_14, -1) @@ -997,13 +729,15 @@ def forward( # pd_op.dropout: (1x16x11x11xf32, 1x16x11x11xui8) <- (1x16x11x11xf32, None, 1xf32) dropout_8, dropout_9 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_1, None, full_4, False, "upscale_in_train", 0, False + softmax_1, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_1 # pd_op.matmul: (1x16x11x64xf32) <- (1x16x11x11xf32, 1x16x11x64xf32) matmul_12 = paddle._C_ops.matmul(dropout_8, transpose_6, False, False) + del dropout_8, transpose_6 # pd_op.transpose: (1x11x16x64xf32) <- (1x16x11x64xf32) transpose_7 = paddle._C_ops.transpose(matmul_12, [0, 2, 1, 3]) @@ -1011,19 +745,20 @@ def forward( # pd_op.reshape: (1x11x1024xf32) <- (1x11x16x64xf32, 3xi64) reshape_7 = paddle._C_ops.reshape(transpose_7, full_int_array_2) + del transpose_7 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_13 = paddle._C_ops.matmul(reshape_7, parameter_363, False, False) - del parameter_363 + del parameter_363, reshape_7 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_15 = paddle._C_ops.add(matmul_13, parameter_362) - del parameter_362 + del matmul_13, parameter_362 # pd_op.dropout: (1x11x1024xf32, 1x11x1024xui8) <- (1x11x1024xf32, None, 1xf32) dropout_10, dropout_11 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_15, None, full_4, False, "upscale_in_train", 0, False + add_15, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1031,6 +766,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_16 = paddle._C_ops.add(layer_norm_6, dropout_10) + del dropout_10, layer_norm_6 # pd_op.layer_norm: (1x11x1024xf32, 1x11xf32, 1x11xf32) <- (1x11x1024xf32, 1024xf32, 1024xf32) layer_norm_9, layer_norm_10, layer_norm_11 = (lambda x, f: f(x))( @@ -1039,7 +775,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_356, parameter_357 + del add_16, parameter_356, parameter_357 # pd_op.matmul: (1x11x4096xf32) <- (1x11x1024xf32, 1024x4096xf32) matmul_14 = paddle._C_ops.matmul(layer_norm_9, parameter_361, False, False) @@ -1047,7 +783,7 @@ def forward( # pd_op.add: (1x11x4096xf32) <- (1x11x4096xf32, 4096xf32) add_17 = paddle._C_ops.add(matmul_14, parameter_360) - del parameter_360 + del matmul_14, parameter_360 # pd_op.relu: (1x11x4096xf32) <- (1x11x4096xf32) relu_1 = paddle._C_ops.relu(add_17) @@ -1055,16 +791,16 @@ def forward( # pd_op.matmul: (1x11x1024xf32) <- (1x11x4096xf32, 4096x1024xf32) matmul_15 = paddle._C_ops.matmul(relu_1, parameter_359, False, False) - del parameter_359 + del parameter_359, relu_1 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_18 = paddle._C_ops.add(matmul_15, parameter_358) - del parameter_358 + del matmul_15, parameter_358 # pd_op.dropout: (1x11x1024xf32, 1x11x1024xui8) <- (1x11x1024xf32, None, 1xf32) dropout_12, dropout_13 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_18, None, full_4, False, "upscale_in_train", 0, False + add_18, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1072,6 +808,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_19 = paddle._C_ops.add(layer_norm_9, dropout_12) + del dropout_12, layer_norm_9 # pd_op.layer_norm: (1x11x1024xf32, 1x11xf32, 1x11xf32) <- (1x11x1024xf32, 1024xf32, 1024xf32) layer_norm_12, layer_norm_13, layer_norm_14 = (lambda x, f: f(x))( @@ -1080,7 +817,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_354, parameter_355 + del add_19, parameter_354, parameter_355 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_16 = paddle._C_ops.matmul(layer_norm_12, parameter_353, False, False) @@ -1088,10 +825,11 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_20 = paddle._C_ops.add(matmul_16, parameter_352) - del parameter_352 + del matmul_16, parameter_352 # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_8 = paddle._C_ops.reshape(add_20, full_int_array_1) + del add_20 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_8 = paddle._C_ops.transpose(reshape_8, [0, 2, 1, 3]) @@ -1103,7 +841,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_21 = paddle._C_ops.add(matmul_17, parameter_350) - del parameter_350 + del matmul_17, parameter_350 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_18 = paddle._C_ops.matmul(layer_norm_12, parameter_349, False, False) @@ -1111,10 +849,11 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_22 = paddle._C_ops.add(matmul_18, parameter_348) - del parameter_348 + del matmul_18, parameter_348 # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_9 = paddle._C_ops.reshape(add_21, full_int_array_1) + del add_21 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_9 = paddle._C_ops.transpose(reshape_9, [0, 2, 1, 3]) @@ -1122,6 +861,7 @@ def forward( # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_10 = paddle._C_ops.reshape(add_22, full_int_array_1) + del add_22 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_10 = paddle._C_ops.transpose(reshape_10, [0, 2, 1, 3]) @@ -1133,9 +873,11 @@ def forward( # pd_op.matmul: (1x16x11x11xf32) <- (1x16x11x64xf32, 1x16x11x64xf32) matmul_19 = paddle._C_ops.matmul(scale_3, transpose_9, False, True) + del scale_3, transpose_9 # pd_op.add: (1x16x11x11xf32) <- (1x16x11x11xf32, 1x1x1x11xf32) add_23 = paddle._C_ops.add(matmul_19, unsqueeze_0) + del matmul_19 # pd_op.softmax: (1x16x11x11xf32) <- (1x16x11x11xf32) softmax_2 = paddle._C_ops.softmax(add_23, -1) @@ -1144,13 +886,15 @@ def forward( # pd_op.dropout: (1x16x11x11xf32, 1x16x11x11xui8) <- (1x16x11x11xf32, None, 1xf32) dropout_14, dropout_15 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_2, None, full_4, False, "upscale_in_train", 0, False + softmax_2, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_2 # pd_op.matmul: (1x16x11x64xf32) <- (1x16x11x11xf32, 1x16x11x64xf32) matmul_20 = paddle._C_ops.matmul(dropout_14, transpose_10, False, False) + del dropout_14, transpose_10 # pd_op.transpose: (1x11x16x64xf32) <- (1x16x11x64xf32) transpose_11 = paddle._C_ops.transpose(matmul_20, [0, 2, 1, 3]) @@ -1158,19 +902,20 @@ def forward( # pd_op.reshape: (1x11x1024xf32) <- (1x11x16x64xf32, 3xi64) reshape_11 = paddle._C_ops.reshape(transpose_11, full_int_array_2) + del transpose_11 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_21 = paddle._C_ops.matmul(reshape_11, parameter_347, False, False) - del parameter_347 + del parameter_347, reshape_11 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_24 = paddle._C_ops.add(matmul_21, parameter_346) - del parameter_346 + del matmul_21, parameter_346 # pd_op.dropout: (1x11x1024xf32, 1x11x1024xui8) <- (1x11x1024xf32, None, 1xf32) dropout_16, dropout_17 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_24, None, full_4, False, "upscale_in_train", 0, False + add_24, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1178,6 +923,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_25 = paddle._C_ops.add(layer_norm_12, dropout_16) + del dropout_16, layer_norm_12 # pd_op.layer_norm: (1x11x1024xf32, 1x11xf32, 1x11xf32) <- (1x11x1024xf32, 1024xf32, 1024xf32) layer_norm_15, layer_norm_16, layer_norm_17 = (lambda x, f: f(x))( @@ -1186,7 +932,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_340, parameter_341 + del add_25, parameter_340, parameter_341 # pd_op.matmul: (1x11x4096xf32) <- (1x11x1024xf32, 1024x4096xf32) matmul_22 = paddle._C_ops.matmul(layer_norm_15, parameter_345, False, False) @@ -1194,7 +940,7 @@ def forward( # pd_op.add: (1x11x4096xf32) <- (1x11x4096xf32, 4096xf32) add_26 = paddle._C_ops.add(matmul_22, parameter_344) - del parameter_344 + del matmul_22, parameter_344 # pd_op.relu: (1x11x4096xf32) <- (1x11x4096xf32) relu_2 = paddle._C_ops.relu(add_26) @@ -1202,16 +948,16 @@ def forward( # pd_op.matmul: (1x11x1024xf32) <- (1x11x4096xf32, 4096x1024xf32) matmul_23 = paddle._C_ops.matmul(relu_2, parameter_343, False, False) - del parameter_343 + del parameter_343, relu_2 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_27 = paddle._C_ops.add(matmul_23, parameter_342) - del parameter_342 + del matmul_23, parameter_342 # pd_op.dropout: (1x11x1024xf32, 1x11x1024xui8) <- (1x11x1024xf32, None, 1xf32) dropout_18, dropout_19 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_27, None, full_4, False, "upscale_in_train", 0, False + add_27, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1219,6 +965,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_28 = paddle._C_ops.add(layer_norm_15, dropout_18) + del dropout_18, layer_norm_15 # pd_op.layer_norm: (1x11x1024xf32, 1x11xf32, 1x11xf32) <- (1x11x1024xf32, 1024xf32, 1024xf32) layer_norm_18, layer_norm_19, layer_norm_20 = (lambda x, f: f(x))( @@ -1227,7 +974,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_338, parameter_339 + del add_28, parameter_338, parameter_339 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_24 = paddle._C_ops.matmul(layer_norm_18, parameter_337, False, False) @@ -1235,10 +982,11 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_29 = paddle._C_ops.add(matmul_24, parameter_336) - del parameter_336 + del matmul_24, parameter_336 # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_12 = paddle._C_ops.reshape(add_29, full_int_array_1) + del add_29 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_12 = paddle._C_ops.transpose(reshape_12, [0, 2, 1, 3]) @@ -1250,7 +998,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_30 = paddle._C_ops.add(matmul_25, parameter_334) - del parameter_334 + del matmul_25, parameter_334 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_26 = paddle._C_ops.matmul(layer_norm_18, parameter_333, False, False) @@ -1258,10 +1006,11 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_31 = paddle._C_ops.add(matmul_26, parameter_332) - del parameter_332 + del matmul_26, parameter_332 # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_13 = paddle._C_ops.reshape(add_30, full_int_array_1) + del add_30 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_13 = paddle._C_ops.transpose(reshape_13, [0, 2, 1, 3]) @@ -1269,6 +1018,7 @@ def forward( # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_14 = paddle._C_ops.reshape(add_31, full_int_array_1) + del add_31 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_14 = paddle._C_ops.transpose(reshape_14, [0, 2, 1, 3]) @@ -1280,9 +1030,11 @@ def forward( # pd_op.matmul: (1x16x11x11xf32) <- (1x16x11x64xf32, 1x16x11x64xf32) matmul_27 = paddle._C_ops.matmul(scale_4, transpose_13, False, True) + del scale_4, transpose_13 # pd_op.add: (1x16x11x11xf32) <- (1x16x11x11xf32, 1x1x1x11xf32) add_32 = paddle._C_ops.add(matmul_27, unsqueeze_0) + del matmul_27 # pd_op.softmax: (1x16x11x11xf32) <- (1x16x11x11xf32) softmax_3 = paddle._C_ops.softmax(add_32, -1) @@ -1291,13 +1043,15 @@ def forward( # pd_op.dropout: (1x16x11x11xf32, 1x16x11x11xui8) <- (1x16x11x11xf32, None, 1xf32) dropout_20, dropout_21 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_3, None, full_4, False, "upscale_in_train", 0, False + softmax_3, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_3 # pd_op.matmul: (1x16x11x64xf32) <- (1x16x11x11xf32, 1x16x11x64xf32) matmul_28 = paddle._C_ops.matmul(dropout_20, transpose_14, False, False) + del dropout_20, transpose_14 # pd_op.transpose: (1x11x16x64xf32) <- (1x16x11x64xf32) transpose_15 = paddle._C_ops.transpose(matmul_28, [0, 2, 1, 3]) @@ -1305,19 +1059,20 @@ def forward( # pd_op.reshape: (1x11x1024xf32) <- (1x11x16x64xf32, 3xi64) reshape_15 = paddle._C_ops.reshape(transpose_15, full_int_array_2) + del transpose_15 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_29 = paddle._C_ops.matmul(reshape_15, parameter_331, False, False) - del parameter_331 + del parameter_331, reshape_15 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_33 = paddle._C_ops.add(matmul_29, parameter_330) - del parameter_330 + del matmul_29, parameter_330 # pd_op.dropout: (1x11x1024xf32, 1x11x1024xui8) <- (1x11x1024xf32, None, 1xf32) dropout_22, dropout_23 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_33, None, full_4, False, "upscale_in_train", 0, False + add_33, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1325,6 +1080,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_34 = paddle._C_ops.add(layer_norm_18, dropout_22) + del dropout_22, layer_norm_18 # pd_op.layer_norm: (1x11x1024xf32, 1x11xf32, 1x11xf32) <- (1x11x1024xf32, 1024xf32, 1024xf32) layer_norm_21, layer_norm_22, layer_norm_23 = (lambda x, f: f(x))( @@ -1333,7 +1089,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_324, parameter_325 + del add_34, parameter_324, parameter_325 # pd_op.matmul: (1x11x4096xf32) <- (1x11x1024xf32, 1024x4096xf32) matmul_30 = paddle._C_ops.matmul(layer_norm_21, parameter_329, False, False) @@ -1341,7 +1097,7 @@ def forward( # pd_op.add: (1x11x4096xf32) <- (1x11x4096xf32, 4096xf32) add_35 = paddle._C_ops.add(matmul_30, parameter_328) - del parameter_328 + del matmul_30, parameter_328 # pd_op.relu: (1x11x4096xf32) <- (1x11x4096xf32) relu_3 = paddle._C_ops.relu(add_35) @@ -1349,16 +1105,16 @@ def forward( # pd_op.matmul: (1x11x1024xf32) <- (1x11x4096xf32, 4096x1024xf32) matmul_31 = paddle._C_ops.matmul(relu_3, parameter_327, False, False) - del parameter_327 + del parameter_327, relu_3 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_36 = paddle._C_ops.add(matmul_31, parameter_326) - del parameter_326 + del matmul_31, parameter_326 # pd_op.dropout: (1x11x1024xf32, 1x11x1024xui8) <- (1x11x1024xf32, None, 1xf32) dropout_24, dropout_25 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_36, None, full_4, False, "upscale_in_train", 0, False + add_36, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1366,6 +1122,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_37 = paddle._C_ops.add(layer_norm_21, dropout_24) + del dropout_24, layer_norm_21 # pd_op.layer_norm: (1x11x1024xf32, 1x11xf32, 1x11xf32) <- (1x11x1024xf32, 1024xf32, 1024xf32) layer_norm_24, layer_norm_25, layer_norm_26 = (lambda x, f: f(x))( @@ -1374,7 +1131,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_322, parameter_323 + del add_37, parameter_322, parameter_323 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_32 = paddle._C_ops.matmul(layer_norm_24, parameter_321, False, False) @@ -1382,10 +1139,11 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_38 = paddle._C_ops.add(matmul_32, parameter_320) - del parameter_320 + del matmul_32, parameter_320 # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_16 = paddle._C_ops.reshape(add_38, full_int_array_1) + del add_38 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_16 = paddle._C_ops.transpose(reshape_16, [0, 2, 1, 3]) @@ -1397,7 +1155,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_39 = paddle._C_ops.add(matmul_33, parameter_318) - del parameter_318 + del matmul_33, parameter_318 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_34 = paddle._C_ops.matmul(layer_norm_24, parameter_317, False, False) @@ -1405,10 +1163,11 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_40 = paddle._C_ops.add(matmul_34, parameter_316) - del parameter_316 + del matmul_34, parameter_316 # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_17 = paddle._C_ops.reshape(add_39, full_int_array_1) + del add_39 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_17 = paddle._C_ops.transpose(reshape_17, [0, 2, 1, 3]) @@ -1416,6 +1175,7 @@ def forward( # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_18 = paddle._C_ops.reshape(add_40, full_int_array_1) + del add_40 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_18 = paddle._C_ops.transpose(reshape_18, [0, 2, 1, 3]) @@ -1427,9 +1187,11 @@ def forward( # pd_op.matmul: (1x16x11x11xf32) <- (1x16x11x64xf32, 1x16x11x64xf32) matmul_35 = paddle._C_ops.matmul(scale_5, transpose_17, False, True) + del scale_5, transpose_17 # pd_op.add: (1x16x11x11xf32) <- (1x16x11x11xf32, 1x1x1x11xf32) add_41 = paddle._C_ops.add(matmul_35, unsqueeze_0) + del matmul_35 # pd_op.softmax: (1x16x11x11xf32) <- (1x16x11x11xf32) softmax_4 = paddle._C_ops.softmax(add_41, -1) @@ -1438,13 +1200,15 @@ def forward( # pd_op.dropout: (1x16x11x11xf32, 1x16x11x11xui8) <- (1x16x11x11xf32, None, 1xf32) dropout_26, dropout_27 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_4, None, full_4, False, "upscale_in_train", 0, False + softmax_4, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_4 # pd_op.matmul: (1x16x11x64xf32) <- (1x16x11x11xf32, 1x16x11x64xf32) matmul_36 = paddle._C_ops.matmul(dropout_26, transpose_18, False, False) + del dropout_26, transpose_18 # pd_op.transpose: (1x11x16x64xf32) <- (1x16x11x64xf32) transpose_19 = paddle._C_ops.transpose(matmul_36, [0, 2, 1, 3]) @@ -1452,19 +1216,20 @@ def forward( # pd_op.reshape: (1x11x1024xf32) <- (1x11x16x64xf32, 3xi64) reshape_19 = paddle._C_ops.reshape(transpose_19, full_int_array_2) + del transpose_19 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_37 = paddle._C_ops.matmul(reshape_19, parameter_315, False, False) - del parameter_315 + del parameter_315, reshape_19 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_42 = paddle._C_ops.add(matmul_37, parameter_314) - del parameter_314 + del matmul_37, parameter_314 # pd_op.dropout: (1x11x1024xf32, 1x11x1024xui8) <- (1x11x1024xf32, None, 1xf32) dropout_28, dropout_29 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_42, None, full_4, False, "upscale_in_train", 0, False + add_42, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1472,6 +1237,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_43 = paddle._C_ops.add(layer_norm_24, dropout_28) + del dropout_28, layer_norm_24 # pd_op.layer_norm: (1x11x1024xf32, 1x11xf32, 1x11xf32) <- (1x11x1024xf32, 1024xf32, 1024xf32) layer_norm_27, layer_norm_28, layer_norm_29 = (lambda x, f: f(x))( @@ -1480,7 +1246,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_308, parameter_309 + del add_43, parameter_308, parameter_309 # pd_op.matmul: (1x11x4096xf32) <- (1x11x1024xf32, 1024x4096xf32) matmul_38 = paddle._C_ops.matmul(layer_norm_27, parameter_313, False, False) @@ -1488,7 +1254,7 @@ def forward( # pd_op.add: (1x11x4096xf32) <- (1x11x4096xf32, 4096xf32) add_44 = paddle._C_ops.add(matmul_38, parameter_312) - del parameter_312 + del matmul_38, parameter_312 # pd_op.relu: (1x11x4096xf32) <- (1x11x4096xf32) relu_4 = paddle._C_ops.relu(add_44) @@ -1496,16 +1262,16 @@ def forward( # pd_op.matmul: (1x11x1024xf32) <- (1x11x4096xf32, 4096x1024xf32) matmul_39 = paddle._C_ops.matmul(relu_4, parameter_311, False, False) - del parameter_311 + del parameter_311, relu_4 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_45 = paddle._C_ops.add(matmul_39, parameter_310) - del parameter_310 + del matmul_39, parameter_310 # pd_op.dropout: (1x11x1024xf32, 1x11x1024xui8) <- (1x11x1024xf32, None, 1xf32) dropout_30, dropout_31 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_45, None, full_4, False, "upscale_in_train", 0, False + add_45, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1513,6 +1279,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_46 = paddle._C_ops.add(layer_norm_27, dropout_30) + del dropout_30, layer_norm_27 # pd_op.layer_norm: (1x11x1024xf32, 1x11xf32, 1x11xf32) <- (1x11x1024xf32, 1024xf32, 1024xf32) layer_norm_30, layer_norm_31, layer_norm_32 = (lambda x, f: f(x))( @@ -1521,7 +1288,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_306, parameter_307 + del add_46, parameter_306, parameter_307 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_40 = paddle._C_ops.matmul(layer_norm_30, parameter_305, False, False) @@ -1529,10 +1296,11 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_47 = paddle._C_ops.add(matmul_40, parameter_304) - del parameter_304 + del matmul_40, parameter_304 # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_20 = paddle._C_ops.reshape(add_47, full_int_array_1) + del add_47 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_20 = paddle._C_ops.transpose(reshape_20, [0, 2, 1, 3]) @@ -1544,7 +1312,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_48 = paddle._C_ops.add(matmul_41, parameter_302) - del parameter_302 + del matmul_41, parameter_302 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_42 = paddle._C_ops.matmul(layer_norm_30, parameter_301, False, False) @@ -1552,10 +1320,11 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_49 = paddle._C_ops.add(matmul_42, parameter_300) - del parameter_300 + del matmul_42, parameter_300 # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_21 = paddle._C_ops.reshape(add_48, full_int_array_1) + del add_48 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_21 = paddle._C_ops.transpose(reshape_21, [0, 2, 1, 3]) @@ -1563,6 +1332,7 @@ def forward( # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_22 = paddle._C_ops.reshape(add_49, full_int_array_1) + del add_49 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_22 = paddle._C_ops.transpose(reshape_22, [0, 2, 1, 3]) @@ -1574,9 +1344,11 @@ def forward( # pd_op.matmul: (1x16x11x11xf32) <- (1x16x11x64xf32, 1x16x11x64xf32) matmul_43 = paddle._C_ops.matmul(scale_6, transpose_21, False, True) + del scale_6, transpose_21 # pd_op.add: (1x16x11x11xf32) <- (1x16x11x11xf32, 1x1x1x11xf32) add_50 = paddle._C_ops.add(matmul_43, unsqueeze_0) + del matmul_43 # pd_op.softmax: (1x16x11x11xf32) <- (1x16x11x11xf32) softmax_5 = paddle._C_ops.softmax(add_50, -1) @@ -1585,13 +1357,15 @@ def forward( # pd_op.dropout: (1x16x11x11xf32, 1x16x11x11xui8) <- (1x16x11x11xf32, None, 1xf32) dropout_32, dropout_33 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_5, None, full_4, False, "upscale_in_train", 0, False + softmax_5, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_5 # pd_op.matmul: (1x16x11x64xf32) <- (1x16x11x11xf32, 1x16x11x64xf32) matmul_44 = paddle._C_ops.matmul(dropout_32, transpose_22, False, False) + del dropout_32, transpose_22 # pd_op.transpose: (1x11x16x64xf32) <- (1x16x11x64xf32) transpose_23 = paddle._C_ops.transpose(matmul_44, [0, 2, 1, 3]) @@ -1599,19 +1373,20 @@ def forward( # pd_op.reshape: (1x11x1024xf32) <- (1x11x16x64xf32, 3xi64) reshape_23 = paddle._C_ops.reshape(transpose_23, full_int_array_2) + del transpose_23 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_45 = paddle._C_ops.matmul(reshape_23, parameter_299, False, False) - del parameter_299 + del parameter_299, reshape_23 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_51 = paddle._C_ops.add(matmul_45, parameter_298) - del parameter_298 + del matmul_45, parameter_298 # pd_op.dropout: (1x11x1024xf32, 1x11x1024xui8) <- (1x11x1024xf32, None, 1xf32) dropout_34, dropout_35 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_51, None, full_4, False, "upscale_in_train", 0, False + add_51, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1619,6 +1394,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_52 = paddle._C_ops.add(layer_norm_30, dropout_34) + del dropout_34, layer_norm_30 # pd_op.layer_norm: (1x11x1024xf32, 1x11xf32, 1x11xf32) <- (1x11x1024xf32, 1024xf32, 1024xf32) layer_norm_33, layer_norm_34, layer_norm_35 = (lambda x, f: f(x))( @@ -1627,7 +1403,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_292, parameter_293 + del add_52, parameter_292, parameter_293 # pd_op.matmul: (1x11x4096xf32) <- (1x11x1024xf32, 1024x4096xf32) matmul_46 = paddle._C_ops.matmul(layer_norm_33, parameter_297, False, False) @@ -1635,7 +1411,7 @@ def forward( # pd_op.add: (1x11x4096xf32) <- (1x11x4096xf32, 4096xf32) add_53 = paddle._C_ops.add(matmul_46, parameter_296) - del parameter_296 + del matmul_46, parameter_296 # pd_op.relu: (1x11x4096xf32) <- (1x11x4096xf32) relu_5 = paddle._C_ops.relu(add_53) @@ -1643,16 +1419,16 @@ def forward( # pd_op.matmul: (1x11x1024xf32) <- (1x11x4096xf32, 4096x1024xf32) matmul_47 = paddle._C_ops.matmul(relu_5, parameter_295, False, False) - del parameter_295 + del parameter_295, relu_5 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_54 = paddle._C_ops.add(matmul_47, parameter_294) - del parameter_294 + del matmul_47, parameter_294 # pd_op.dropout: (1x11x1024xf32, 1x11x1024xui8) <- (1x11x1024xf32, None, 1xf32) dropout_36, dropout_37 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_54, None, full_4, False, "upscale_in_train", 0, False + add_54, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1660,6 +1436,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_55 = paddle._C_ops.add(layer_norm_33, dropout_36) + del dropout_36, layer_norm_33 # pd_op.layer_norm: (1x11x1024xf32, 1x11xf32, 1x11xf32) <- (1x11x1024xf32, 1024xf32, 1024xf32) layer_norm_36, layer_norm_37, layer_norm_38 = (lambda x, f: f(x))( @@ -1668,7 +1445,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_290, parameter_291 + del add_55, parameter_290, parameter_291 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_48 = paddle._C_ops.matmul(layer_norm_36, parameter_289, False, False) @@ -1676,10 +1453,11 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_56 = paddle._C_ops.add(matmul_48, parameter_288) - del parameter_288 + del matmul_48, parameter_288 # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_24 = paddle._C_ops.reshape(add_56, full_int_array_1) + del add_56 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_24 = paddle._C_ops.transpose(reshape_24, [0, 2, 1, 3]) @@ -1691,7 +1469,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_57 = paddle._C_ops.add(matmul_49, parameter_286) - del parameter_286 + del matmul_49, parameter_286 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_50 = paddle._C_ops.matmul(layer_norm_36, parameter_285, False, False) @@ -1699,10 +1477,11 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_58 = paddle._C_ops.add(matmul_50, parameter_284) - del parameter_284 + del matmul_50, parameter_284 # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_25 = paddle._C_ops.reshape(add_57, full_int_array_1) + del add_57 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_25 = paddle._C_ops.transpose(reshape_25, [0, 2, 1, 3]) @@ -1710,6 +1489,7 @@ def forward( # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_26 = paddle._C_ops.reshape(add_58, full_int_array_1) + del add_58 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_26 = paddle._C_ops.transpose(reshape_26, [0, 2, 1, 3]) @@ -1721,9 +1501,11 @@ def forward( # pd_op.matmul: (1x16x11x11xf32) <- (1x16x11x64xf32, 1x16x11x64xf32) matmul_51 = paddle._C_ops.matmul(scale_7, transpose_25, False, True) + del scale_7, transpose_25 # pd_op.add: (1x16x11x11xf32) <- (1x16x11x11xf32, 1x1x1x11xf32) add_59 = paddle._C_ops.add(matmul_51, unsqueeze_0) + del matmul_51 # pd_op.softmax: (1x16x11x11xf32) <- (1x16x11x11xf32) softmax_6 = paddle._C_ops.softmax(add_59, -1) @@ -1732,13 +1514,15 @@ def forward( # pd_op.dropout: (1x16x11x11xf32, 1x16x11x11xui8) <- (1x16x11x11xf32, None, 1xf32) dropout_38, dropout_39 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_6, None, full_4, False, "upscale_in_train", 0, False + softmax_6, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_6 # pd_op.matmul: (1x16x11x64xf32) <- (1x16x11x11xf32, 1x16x11x64xf32) matmul_52 = paddle._C_ops.matmul(dropout_38, transpose_26, False, False) + del dropout_38, transpose_26 # pd_op.transpose: (1x11x16x64xf32) <- (1x16x11x64xf32) transpose_27 = paddle._C_ops.transpose(matmul_52, [0, 2, 1, 3]) @@ -1746,19 +1530,20 @@ def forward( # pd_op.reshape: (1x11x1024xf32) <- (1x11x16x64xf32, 3xi64) reshape_27 = paddle._C_ops.reshape(transpose_27, full_int_array_2) + del transpose_27 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_53 = paddle._C_ops.matmul(reshape_27, parameter_283, False, False) - del parameter_283 + del parameter_283, reshape_27 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_60 = paddle._C_ops.add(matmul_53, parameter_282) - del parameter_282 + del matmul_53, parameter_282 # pd_op.dropout: (1x11x1024xf32, 1x11x1024xui8) <- (1x11x1024xf32, None, 1xf32) dropout_40, dropout_41 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_60, None, full_4, False, "upscale_in_train", 0, False + add_60, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1766,6 +1551,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_61 = paddle._C_ops.add(layer_norm_36, dropout_40) + del dropout_40, layer_norm_36 # pd_op.layer_norm: (1x11x1024xf32, 1x11xf32, 1x11xf32) <- (1x11x1024xf32, 1024xf32, 1024xf32) layer_norm_39, layer_norm_40, layer_norm_41 = (lambda x, f: f(x))( @@ -1774,7 +1560,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_276, parameter_277 + del add_61, parameter_276, parameter_277 # pd_op.matmul: (1x11x4096xf32) <- (1x11x1024xf32, 1024x4096xf32) matmul_54 = paddle._C_ops.matmul(layer_norm_39, parameter_281, False, False) @@ -1782,7 +1568,7 @@ def forward( # pd_op.add: (1x11x4096xf32) <- (1x11x4096xf32, 4096xf32) add_62 = paddle._C_ops.add(matmul_54, parameter_280) - del parameter_280 + del matmul_54, parameter_280 # pd_op.relu: (1x11x4096xf32) <- (1x11x4096xf32) relu_6 = paddle._C_ops.relu(add_62) @@ -1790,16 +1576,16 @@ def forward( # pd_op.matmul: (1x11x1024xf32) <- (1x11x4096xf32, 4096x1024xf32) matmul_55 = paddle._C_ops.matmul(relu_6, parameter_279, False, False) - del parameter_279 + del parameter_279, relu_6 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_63 = paddle._C_ops.add(matmul_55, parameter_278) - del parameter_278 + del matmul_55, parameter_278 # pd_op.dropout: (1x11x1024xf32, 1x11x1024xui8) <- (1x11x1024xf32, None, 1xf32) dropout_42, dropout_43 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_63, None, full_4, False, "upscale_in_train", 0, False + add_63, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1807,6 +1593,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_64 = paddle._C_ops.add(layer_norm_39, dropout_42) + del dropout_42, layer_norm_39 # pd_op.layer_norm: (1x11x1024xf32, 1x11xf32, 1x11xf32) <- (1x11x1024xf32, 1024xf32, 1024xf32) layer_norm_42, layer_norm_43, layer_norm_44 = (lambda x, f: f(x))( @@ -1815,7 +1602,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_274, parameter_275 + del add_64, parameter_274, parameter_275 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_56 = paddle._C_ops.matmul(layer_norm_42, parameter_273, False, False) @@ -1823,10 +1610,11 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_65 = paddle._C_ops.add(matmul_56, parameter_272) - del parameter_272 + del matmul_56, parameter_272 # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_28 = paddle._C_ops.reshape(add_65, full_int_array_1) + del add_65 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_28 = paddle._C_ops.transpose(reshape_28, [0, 2, 1, 3]) @@ -1838,7 +1626,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_66 = paddle._C_ops.add(matmul_57, parameter_270) - del parameter_270 + del matmul_57, parameter_270 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_58 = paddle._C_ops.matmul(layer_norm_42, parameter_269, False, False) @@ -1846,10 +1634,11 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_67 = paddle._C_ops.add(matmul_58, parameter_268) - del parameter_268 + del matmul_58, parameter_268 # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_29 = paddle._C_ops.reshape(add_66, full_int_array_1) + del add_66 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_29 = paddle._C_ops.transpose(reshape_29, [0, 2, 1, 3]) @@ -1857,6 +1646,7 @@ def forward( # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_30 = paddle._C_ops.reshape(add_67, full_int_array_1) + del add_67 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_30 = paddle._C_ops.transpose(reshape_30, [0, 2, 1, 3]) @@ -1868,9 +1658,11 @@ def forward( # pd_op.matmul: (1x16x11x11xf32) <- (1x16x11x64xf32, 1x16x11x64xf32) matmul_59 = paddle._C_ops.matmul(scale_8, transpose_29, False, True) + del scale_8, transpose_29 # pd_op.add: (1x16x11x11xf32) <- (1x16x11x11xf32, 1x1x1x11xf32) add_68 = paddle._C_ops.add(matmul_59, unsqueeze_0) + del matmul_59 # pd_op.softmax: (1x16x11x11xf32) <- (1x16x11x11xf32) softmax_7 = paddle._C_ops.softmax(add_68, -1) @@ -1879,13 +1671,15 @@ def forward( # pd_op.dropout: (1x16x11x11xf32, 1x16x11x11xui8) <- (1x16x11x11xf32, None, 1xf32) dropout_44, dropout_45 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_7, None, full_4, False, "upscale_in_train", 0, False + softmax_7, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_7 # pd_op.matmul: (1x16x11x64xf32) <- (1x16x11x11xf32, 1x16x11x64xf32) matmul_60 = paddle._C_ops.matmul(dropout_44, transpose_30, False, False) + del dropout_44, transpose_30 # pd_op.transpose: (1x11x16x64xf32) <- (1x16x11x64xf32) transpose_31 = paddle._C_ops.transpose(matmul_60, [0, 2, 1, 3]) @@ -1893,19 +1687,20 @@ def forward( # pd_op.reshape: (1x11x1024xf32) <- (1x11x16x64xf32, 3xi64) reshape_31 = paddle._C_ops.reshape(transpose_31, full_int_array_2) + del transpose_31 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_61 = paddle._C_ops.matmul(reshape_31, parameter_267, False, False) - del parameter_267 + del parameter_267, reshape_31 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_69 = paddle._C_ops.add(matmul_61, parameter_266) - del parameter_266 + del matmul_61, parameter_266 # pd_op.dropout: (1x11x1024xf32, 1x11x1024xui8) <- (1x11x1024xf32, None, 1xf32) dropout_46, dropout_47 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_69, None, full_4, False, "upscale_in_train", 0, False + add_69, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1913,6 +1708,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_70 = paddle._C_ops.add(layer_norm_42, dropout_46) + del dropout_46, layer_norm_42 # pd_op.layer_norm: (1x11x1024xf32, 1x11xf32, 1x11xf32) <- (1x11x1024xf32, 1024xf32, 1024xf32) layer_norm_45, layer_norm_46, layer_norm_47 = (lambda x, f: f(x))( @@ -1921,7 +1717,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_260, parameter_261 + del add_70, parameter_260, parameter_261 # pd_op.matmul: (1x11x4096xf32) <- (1x11x1024xf32, 1024x4096xf32) matmul_62 = paddle._C_ops.matmul(layer_norm_45, parameter_265, False, False) @@ -1929,7 +1725,7 @@ def forward( # pd_op.add: (1x11x4096xf32) <- (1x11x4096xf32, 4096xf32) add_71 = paddle._C_ops.add(matmul_62, parameter_264) - del parameter_264 + del matmul_62, parameter_264 # pd_op.relu: (1x11x4096xf32) <- (1x11x4096xf32) relu_7 = paddle._C_ops.relu(add_71) @@ -1937,16 +1733,16 @@ def forward( # pd_op.matmul: (1x11x1024xf32) <- (1x11x4096xf32, 4096x1024xf32) matmul_63 = paddle._C_ops.matmul(relu_7, parameter_263, False, False) - del parameter_263 + del parameter_263, relu_7 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_72 = paddle._C_ops.add(matmul_63, parameter_262) - del parameter_262 + del matmul_63, parameter_262 # pd_op.dropout: (1x11x1024xf32, 1x11x1024xui8) <- (1x11x1024xf32, None, 1xf32) dropout_48, dropout_49 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_72, None, full_4, False, "upscale_in_train", 0, False + add_72, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1954,6 +1750,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_73 = paddle._C_ops.add(layer_norm_45, dropout_48) + del dropout_48, layer_norm_45 # pd_op.layer_norm: (1x11x1024xf32, 1x11xf32, 1x11xf32) <- (1x11x1024xf32, 1024xf32, 1024xf32) layer_norm_48, layer_norm_49, layer_norm_50 = (lambda x, f: f(x))( @@ -1962,7 +1759,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_258, parameter_259 + del add_73, parameter_258, parameter_259 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_64 = paddle._C_ops.matmul(layer_norm_48, parameter_257, False, False) @@ -1970,10 +1767,11 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_74 = paddle._C_ops.add(matmul_64, parameter_256) - del parameter_256 + del matmul_64, parameter_256 # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_32 = paddle._C_ops.reshape(add_74, full_int_array_1) + del add_74 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_32 = paddle._C_ops.transpose(reshape_32, [0, 2, 1, 3]) @@ -1985,7 +1783,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_75 = paddle._C_ops.add(matmul_65, parameter_254) - del parameter_254 + del matmul_65, parameter_254 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_66 = paddle._C_ops.matmul(layer_norm_48, parameter_253, False, False) @@ -1993,10 +1791,11 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_76 = paddle._C_ops.add(matmul_66, parameter_252) - del parameter_252 + del matmul_66, parameter_252 # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_33 = paddle._C_ops.reshape(add_75, full_int_array_1) + del add_75 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_33 = paddle._C_ops.transpose(reshape_33, [0, 2, 1, 3]) @@ -2004,6 +1803,7 @@ def forward( # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_34 = paddle._C_ops.reshape(add_76, full_int_array_1) + del add_76 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_34 = paddle._C_ops.transpose(reshape_34, [0, 2, 1, 3]) @@ -2015,9 +1815,11 @@ def forward( # pd_op.matmul: (1x16x11x11xf32) <- (1x16x11x64xf32, 1x16x11x64xf32) matmul_67 = paddle._C_ops.matmul(scale_9, transpose_33, False, True) + del scale_9, transpose_33 # pd_op.add: (1x16x11x11xf32) <- (1x16x11x11xf32, 1x1x1x11xf32) add_77 = paddle._C_ops.add(matmul_67, unsqueeze_0) + del matmul_67 # pd_op.softmax: (1x16x11x11xf32) <- (1x16x11x11xf32) softmax_8 = paddle._C_ops.softmax(add_77, -1) @@ -2026,13 +1828,15 @@ def forward( # pd_op.dropout: (1x16x11x11xf32, 1x16x11x11xui8) <- (1x16x11x11xf32, None, 1xf32) dropout_50, dropout_51 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_8, None, full_4, False, "upscale_in_train", 0, False + softmax_8, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_8 # pd_op.matmul: (1x16x11x64xf32) <- (1x16x11x11xf32, 1x16x11x64xf32) matmul_68 = paddle._C_ops.matmul(dropout_50, transpose_34, False, False) + del dropout_50, transpose_34 # pd_op.transpose: (1x11x16x64xf32) <- (1x16x11x64xf32) transpose_35 = paddle._C_ops.transpose(matmul_68, [0, 2, 1, 3]) @@ -2040,19 +1844,20 @@ def forward( # pd_op.reshape: (1x11x1024xf32) <- (1x11x16x64xf32, 3xi64) reshape_35 = paddle._C_ops.reshape(transpose_35, full_int_array_2) + del transpose_35 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_69 = paddle._C_ops.matmul(reshape_35, parameter_251, False, False) - del parameter_251 + del parameter_251, reshape_35 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_78 = paddle._C_ops.add(matmul_69, parameter_250) - del parameter_250 + del matmul_69, parameter_250 # pd_op.dropout: (1x11x1024xf32, 1x11x1024xui8) <- (1x11x1024xf32, None, 1xf32) dropout_52, dropout_53 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_78, None, full_4, False, "upscale_in_train", 0, False + add_78, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -2060,6 +1865,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_79 = paddle._C_ops.add(layer_norm_48, dropout_52) + del dropout_52, layer_norm_48 # pd_op.layer_norm: (1x11x1024xf32, 1x11xf32, 1x11xf32) <- (1x11x1024xf32, 1024xf32, 1024xf32) layer_norm_51, layer_norm_52, layer_norm_53 = (lambda x, f: f(x))( @@ -2068,7 +1874,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_244, parameter_245 + del add_79, parameter_244, parameter_245 # pd_op.matmul: (1x11x4096xf32) <- (1x11x1024xf32, 1024x4096xf32) matmul_70 = paddle._C_ops.matmul(layer_norm_51, parameter_249, False, False) @@ -2076,7 +1882,7 @@ def forward( # pd_op.add: (1x11x4096xf32) <- (1x11x4096xf32, 4096xf32) add_80 = paddle._C_ops.add(matmul_70, parameter_248) - del parameter_248 + del matmul_70, parameter_248 # pd_op.relu: (1x11x4096xf32) <- (1x11x4096xf32) relu_8 = paddle._C_ops.relu(add_80) @@ -2084,16 +1890,16 @@ def forward( # pd_op.matmul: (1x11x1024xf32) <- (1x11x4096xf32, 4096x1024xf32) matmul_71 = paddle._C_ops.matmul(relu_8, parameter_247, False, False) - del parameter_247 + del parameter_247, relu_8 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_81 = paddle._C_ops.add(matmul_71, parameter_246) - del parameter_246 + del matmul_71, parameter_246 # pd_op.dropout: (1x11x1024xf32, 1x11x1024xui8) <- (1x11x1024xf32, None, 1xf32) dropout_54, dropout_55 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_81, None, full_4, False, "upscale_in_train", 0, False + add_81, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -2101,6 +1907,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_82 = paddle._C_ops.add(layer_norm_51, dropout_54) + del dropout_54, layer_norm_51 # pd_op.layer_norm: (1x11x1024xf32, 1x11xf32, 1x11xf32) <- (1x11x1024xf32, 1024xf32, 1024xf32) layer_norm_54, layer_norm_55, layer_norm_56 = (lambda x, f: f(x))( @@ -2109,7 +1916,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_242, parameter_243 + del add_82, parameter_242, parameter_243 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_72 = paddle._C_ops.matmul(layer_norm_54, parameter_241, False, False) @@ -2117,10 +1924,11 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_83 = paddle._C_ops.add(matmul_72, parameter_240) - del parameter_240 + del matmul_72, parameter_240 # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_36 = paddle._C_ops.reshape(add_83, full_int_array_1) + del add_83 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_36 = paddle._C_ops.transpose(reshape_36, [0, 2, 1, 3]) @@ -2132,7 +1940,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_84 = paddle._C_ops.add(matmul_73, parameter_238) - del parameter_238 + del matmul_73, parameter_238 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_74 = paddle._C_ops.matmul(layer_norm_54, parameter_237, False, False) @@ -2140,10 +1948,11 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_85 = paddle._C_ops.add(matmul_74, parameter_236) - del parameter_236 + del matmul_74, parameter_236 # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_37 = paddle._C_ops.reshape(add_84, full_int_array_1) + del add_84 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_37 = paddle._C_ops.transpose(reshape_37, [0, 2, 1, 3]) @@ -2151,6 +1960,7 @@ def forward( # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_38 = paddle._C_ops.reshape(add_85, full_int_array_1) + del add_85 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_38 = paddle._C_ops.transpose(reshape_38, [0, 2, 1, 3]) @@ -2162,9 +1972,11 @@ def forward( # pd_op.matmul: (1x16x11x11xf32) <- (1x16x11x64xf32, 1x16x11x64xf32) matmul_75 = paddle._C_ops.matmul(scale_10, transpose_37, False, True) + del scale_10, transpose_37 # pd_op.add: (1x16x11x11xf32) <- (1x16x11x11xf32, 1x1x1x11xf32) add_86 = paddle._C_ops.add(matmul_75, unsqueeze_0) + del matmul_75 # pd_op.softmax: (1x16x11x11xf32) <- (1x16x11x11xf32) softmax_9 = paddle._C_ops.softmax(add_86, -1) @@ -2173,13 +1985,15 @@ def forward( # pd_op.dropout: (1x16x11x11xf32, 1x16x11x11xui8) <- (1x16x11x11xf32, None, 1xf32) dropout_56, dropout_57 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_9, None, full_4, False, "upscale_in_train", 0, False + softmax_9, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_9 # pd_op.matmul: (1x16x11x64xf32) <- (1x16x11x11xf32, 1x16x11x64xf32) matmul_76 = paddle._C_ops.matmul(dropout_56, transpose_38, False, False) + del dropout_56, transpose_38 # pd_op.transpose: (1x11x16x64xf32) <- (1x16x11x64xf32) transpose_39 = paddle._C_ops.transpose(matmul_76, [0, 2, 1, 3]) @@ -2187,19 +2001,20 @@ def forward( # pd_op.reshape: (1x11x1024xf32) <- (1x11x16x64xf32, 3xi64) reshape_39 = paddle._C_ops.reshape(transpose_39, full_int_array_2) + del transpose_39 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_77 = paddle._C_ops.matmul(reshape_39, parameter_235, False, False) - del parameter_235 + del parameter_235, reshape_39 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_87 = paddle._C_ops.add(matmul_77, parameter_234) - del parameter_234 + del matmul_77, parameter_234 # pd_op.dropout: (1x11x1024xf32, 1x11x1024xui8) <- (1x11x1024xf32, None, 1xf32) dropout_58, dropout_59 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_87, None, full_4, False, "upscale_in_train", 0, False + add_87, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -2207,6 +2022,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_88 = paddle._C_ops.add(layer_norm_54, dropout_58) + del dropout_58, layer_norm_54 # pd_op.layer_norm: (1x11x1024xf32, 1x11xf32, 1x11xf32) <- (1x11x1024xf32, 1024xf32, 1024xf32) layer_norm_57, layer_norm_58, layer_norm_59 = (lambda x, f: f(x))( @@ -2215,7 +2031,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_228, parameter_229 + del add_88, parameter_228, parameter_229 # pd_op.matmul: (1x11x4096xf32) <- (1x11x1024xf32, 1024x4096xf32) matmul_78 = paddle._C_ops.matmul(layer_norm_57, parameter_233, False, False) @@ -2223,7 +2039,7 @@ def forward( # pd_op.add: (1x11x4096xf32) <- (1x11x4096xf32, 4096xf32) add_89 = paddle._C_ops.add(matmul_78, parameter_232) - del parameter_232 + del matmul_78, parameter_232 # pd_op.relu: (1x11x4096xf32) <- (1x11x4096xf32) relu_9 = paddle._C_ops.relu(add_89) @@ -2231,16 +2047,16 @@ def forward( # pd_op.matmul: (1x11x1024xf32) <- (1x11x4096xf32, 4096x1024xf32) matmul_79 = paddle._C_ops.matmul(relu_9, parameter_231, False, False) - del parameter_231 + del parameter_231, relu_9 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_90 = paddle._C_ops.add(matmul_79, parameter_230) - del parameter_230 + del matmul_79, parameter_230 # pd_op.dropout: (1x11x1024xf32, 1x11x1024xui8) <- (1x11x1024xf32, None, 1xf32) dropout_60, dropout_61 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_90, None, full_4, False, "upscale_in_train", 0, False + add_90, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -2248,6 +2064,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_91 = paddle._C_ops.add(layer_norm_57, dropout_60) + del dropout_60, layer_norm_57 # pd_op.layer_norm: (1x11x1024xf32, 1x11xf32, 1x11xf32) <- (1x11x1024xf32, 1024xf32, 1024xf32) layer_norm_60, layer_norm_61, layer_norm_62 = (lambda x, f: f(x))( @@ -2256,7 +2073,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_226, parameter_227 + del add_91, parameter_226, parameter_227 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_80 = paddle._C_ops.matmul(layer_norm_60, parameter_225, False, False) @@ -2264,10 +2081,11 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_92 = paddle._C_ops.add(matmul_80, parameter_224) - del parameter_224 + del matmul_80, parameter_224 # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_40 = paddle._C_ops.reshape(add_92, full_int_array_1) + del add_92 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_40 = paddle._C_ops.transpose(reshape_40, [0, 2, 1, 3]) @@ -2279,7 +2097,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_93 = paddle._C_ops.add(matmul_81, parameter_222) - del parameter_222 + del matmul_81, parameter_222 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_82 = paddle._C_ops.matmul(layer_norm_60, parameter_221, False, False) @@ -2287,10 +2105,11 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_94 = paddle._C_ops.add(matmul_82, parameter_220) - del parameter_220 + del matmul_82, parameter_220 # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_41 = paddle._C_ops.reshape(add_93, full_int_array_1) + del add_93 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_41 = paddle._C_ops.transpose(reshape_41, [0, 2, 1, 3]) @@ -2298,6 +2117,7 @@ def forward( # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_42 = paddle._C_ops.reshape(add_94, full_int_array_1) + del add_94 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_42 = paddle._C_ops.transpose(reshape_42, [0, 2, 1, 3]) @@ -2309,9 +2129,11 @@ def forward( # pd_op.matmul: (1x16x11x11xf32) <- (1x16x11x64xf32, 1x16x11x64xf32) matmul_83 = paddle._C_ops.matmul(scale_11, transpose_41, False, True) + del scale_11, transpose_41 # pd_op.add: (1x16x11x11xf32) <- (1x16x11x11xf32, 1x1x1x11xf32) add_95 = paddle._C_ops.add(matmul_83, unsqueeze_0) + del matmul_83 # pd_op.softmax: (1x16x11x11xf32) <- (1x16x11x11xf32) softmax_10 = paddle._C_ops.softmax(add_95, -1) @@ -2320,13 +2142,15 @@ def forward( # pd_op.dropout: (1x16x11x11xf32, 1x16x11x11xui8) <- (1x16x11x11xf32, None, 1xf32) dropout_62, dropout_63 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_10, None, full_4, False, "upscale_in_train", 0, False + softmax_10, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_10 # pd_op.matmul: (1x16x11x64xf32) <- (1x16x11x11xf32, 1x16x11x64xf32) matmul_84 = paddle._C_ops.matmul(dropout_62, transpose_42, False, False) + del dropout_62, transpose_42 # pd_op.transpose: (1x11x16x64xf32) <- (1x16x11x64xf32) transpose_43 = paddle._C_ops.transpose(matmul_84, [0, 2, 1, 3]) @@ -2334,19 +2158,20 @@ def forward( # pd_op.reshape: (1x11x1024xf32) <- (1x11x16x64xf32, 3xi64) reshape_43 = paddle._C_ops.reshape(transpose_43, full_int_array_2) + del transpose_43 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_85 = paddle._C_ops.matmul(reshape_43, parameter_219, False, False) - del parameter_219 + del parameter_219, reshape_43 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_96 = paddle._C_ops.add(matmul_85, parameter_218) - del parameter_218 + del matmul_85, parameter_218 # pd_op.dropout: (1x11x1024xf32, 1x11x1024xui8) <- (1x11x1024xf32, None, 1xf32) dropout_64, dropout_65 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_96, None, full_4, False, "upscale_in_train", 0, False + add_96, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -2354,6 +2179,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_97 = paddle._C_ops.add(layer_norm_60, dropout_64) + del dropout_64, layer_norm_60 # pd_op.layer_norm: (1x11x1024xf32, 1x11xf32, 1x11xf32) <- (1x11x1024xf32, 1024xf32, 1024xf32) layer_norm_63, layer_norm_64, layer_norm_65 = (lambda x, f: f(x))( @@ -2362,7 +2188,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_212, parameter_213 + del add_97, parameter_212, parameter_213 # pd_op.matmul: (1x11x4096xf32) <- (1x11x1024xf32, 1024x4096xf32) matmul_86 = paddle._C_ops.matmul(layer_norm_63, parameter_217, False, False) @@ -2370,7 +2196,7 @@ def forward( # pd_op.add: (1x11x4096xf32) <- (1x11x4096xf32, 4096xf32) add_98 = paddle._C_ops.add(matmul_86, parameter_216) - del parameter_216 + del matmul_86, parameter_216 # pd_op.relu: (1x11x4096xf32) <- (1x11x4096xf32) relu_10 = paddle._C_ops.relu(add_98) @@ -2378,16 +2204,16 @@ def forward( # pd_op.matmul: (1x11x1024xf32) <- (1x11x4096xf32, 4096x1024xf32) matmul_87 = paddle._C_ops.matmul(relu_10, parameter_215, False, False) - del parameter_215 + del parameter_215, relu_10 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_99 = paddle._C_ops.add(matmul_87, parameter_214) - del parameter_214 + del matmul_87, parameter_214 # pd_op.dropout: (1x11x1024xf32, 1x11x1024xui8) <- (1x11x1024xf32, None, 1xf32) dropout_66, dropout_67 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_99, None, full_4, False, "upscale_in_train", 0, False + add_99, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -2395,6 +2221,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_100 = paddle._C_ops.add(layer_norm_63, dropout_66) + del dropout_66, layer_norm_63 # pd_op.layer_norm: (1x11x1024xf32, 1x11xf32, 1x11xf32) <- (1x11x1024xf32, 1024xf32, 1024xf32) layer_norm_66, layer_norm_67, layer_norm_68 = (lambda x, f: f(x))( @@ -2403,7 +2230,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_210, parameter_211 + del add_100, parameter_210, parameter_211 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_88 = paddle._C_ops.matmul(layer_norm_66, parameter_209, False, False) @@ -2411,10 +2238,11 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_101 = paddle._C_ops.add(matmul_88, parameter_208) - del parameter_208 + del matmul_88, parameter_208 # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_44 = paddle._C_ops.reshape(add_101, full_int_array_1) + del add_101 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_44 = paddle._C_ops.transpose(reshape_44, [0, 2, 1, 3]) @@ -2426,7 +2254,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_102 = paddle._C_ops.add(matmul_89, parameter_206) - del parameter_206 + del matmul_89, parameter_206 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_90 = paddle._C_ops.matmul(layer_norm_66, parameter_205, False, False) @@ -2434,10 +2262,11 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_103 = paddle._C_ops.add(matmul_90, parameter_204) - del parameter_204 + del matmul_90, parameter_204 # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_45 = paddle._C_ops.reshape(add_102, full_int_array_1) + del add_102 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_45 = paddle._C_ops.transpose(reshape_45, [0, 2, 1, 3]) @@ -2445,6 +2274,7 @@ def forward( # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_46 = paddle._C_ops.reshape(add_103, full_int_array_1) + del add_103 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_46 = paddle._C_ops.transpose(reshape_46, [0, 2, 1, 3]) @@ -2456,9 +2286,11 @@ def forward( # pd_op.matmul: (1x16x11x11xf32) <- (1x16x11x64xf32, 1x16x11x64xf32) matmul_91 = paddle._C_ops.matmul(scale_12, transpose_45, False, True) + del scale_12, transpose_45 # pd_op.add: (1x16x11x11xf32) <- (1x16x11x11xf32, 1x1x1x11xf32) add_104 = paddle._C_ops.add(matmul_91, unsqueeze_0) + del matmul_91 # pd_op.softmax: (1x16x11x11xf32) <- (1x16x11x11xf32) softmax_11 = paddle._C_ops.softmax(add_104, -1) @@ -2467,13 +2299,15 @@ def forward( # pd_op.dropout: (1x16x11x11xf32, 1x16x11x11xui8) <- (1x16x11x11xf32, None, 1xf32) dropout_68, dropout_69 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_11, None, full_4, False, "upscale_in_train", 0, False + softmax_11, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_11 # pd_op.matmul: (1x16x11x64xf32) <- (1x16x11x11xf32, 1x16x11x64xf32) matmul_92 = paddle._C_ops.matmul(dropout_68, transpose_46, False, False) + del dropout_68, transpose_46 # pd_op.transpose: (1x11x16x64xf32) <- (1x16x11x64xf32) transpose_47 = paddle._C_ops.transpose(matmul_92, [0, 2, 1, 3]) @@ -2481,19 +2315,20 @@ def forward( # pd_op.reshape: (1x11x1024xf32) <- (1x11x16x64xf32, 3xi64) reshape_47 = paddle._C_ops.reshape(transpose_47, full_int_array_2) + del transpose_47 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_93 = paddle._C_ops.matmul(reshape_47, parameter_203, False, False) - del parameter_203 + del parameter_203, reshape_47 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_105 = paddle._C_ops.add(matmul_93, parameter_202) - del parameter_202 + del matmul_93, parameter_202 # pd_op.dropout: (1x11x1024xf32, 1x11x1024xui8) <- (1x11x1024xf32, None, 1xf32) dropout_70, dropout_71 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_105, None, full_4, False, "upscale_in_train", 0, False + add_105, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -2501,6 +2336,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_106 = paddle._C_ops.add(layer_norm_66, dropout_70) + del dropout_70, layer_norm_66 # pd_op.layer_norm: (1x11x1024xf32, 1x11xf32, 1x11xf32) <- (1x11x1024xf32, 1024xf32, 1024xf32) layer_norm_69, layer_norm_70, layer_norm_71 = (lambda x, f: f(x))( @@ -2509,7 +2345,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_196, parameter_197 + del add_106, parameter_196, parameter_197 # pd_op.matmul: (1x11x4096xf32) <- (1x11x1024xf32, 1024x4096xf32) matmul_94 = paddle._C_ops.matmul(layer_norm_69, parameter_201, False, False) @@ -2517,7 +2353,7 @@ def forward( # pd_op.add: (1x11x4096xf32) <- (1x11x4096xf32, 4096xf32) add_107 = paddle._C_ops.add(matmul_94, parameter_200) - del parameter_200 + del matmul_94, parameter_200 # pd_op.relu: (1x11x4096xf32) <- (1x11x4096xf32) relu_11 = paddle._C_ops.relu(add_107) @@ -2525,16 +2361,16 @@ def forward( # pd_op.matmul: (1x11x1024xf32) <- (1x11x4096xf32, 4096x1024xf32) matmul_95 = paddle._C_ops.matmul(relu_11, parameter_199, False, False) - del parameter_199 + del parameter_199, relu_11 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_108 = paddle._C_ops.add(matmul_95, parameter_198) - del parameter_198 + del matmul_95, parameter_198 # pd_op.dropout: (1x11x1024xf32, 1x11x1024xui8) <- (1x11x1024xf32, None, 1xf32) dropout_72, dropout_73 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_108, None, full_4, False, "upscale_in_train", 0, False + add_108, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -2542,6 +2378,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_109 = paddle._C_ops.add(layer_norm_69, dropout_72) + del dropout_72, layer_norm_69 # pd_op.layer_norm: (1x11x1024xf32, 1x11xf32, 1x11xf32) <- (1x11x1024xf32, 1024xf32, 1024xf32) layer_norm_72, layer_norm_73, layer_norm_74 = (lambda x, f: f(x))( @@ -2550,7 +2387,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_194, parameter_195 + del add_109, parameter_194, parameter_195 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_96 = paddle._C_ops.matmul(layer_norm_72, parameter_193, False, False) @@ -2558,10 +2395,11 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_110 = paddle._C_ops.add(matmul_96, parameter_192) - del parameter_192 + del matmul_96, parameter_192 # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_48 = paddle._C_ops.reshape(add_110, full_int_array_1) + del add_110 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_48 = paddle._C_ops.transpose(reshape_48, [0, 2, 1, 3]) @@ -2573,7 +2411,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_111 = paddle._C_ops.add(matmul_97, parameter_190) - del parameter_190 + del matmul_97, parameter_190 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_98 = paddle._C_ops.matmul(layer_norm_72, parameter_189, False, False) @@ -2581,10 +2419,11 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_112 = paddle._C_ops.add(matmul_98, parameter_188) - del parameter_188 + del matmul_98, parameter_188 # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_49 = paddle._C_ops.reshape(add_111, full_int_array_1) + del add_111 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_49 = paddle._C_ops.transpose(reshape_49, [0, 2, 1, 3]) @@ -2592,6 +2431,7 @@ def forward( # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_50 = paddle._C_ops.reshape(add_112, full_int_array_1) + del add_112 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_50 = paddle._C_ops.transpose(reshape_50, [0, 2, 1, 3]) @@ -2603,9 +2443,11 @@ def forward( # pd_op.matmul: (1x16x11x11xf32) <- (1x16x11x64xf32, 1x16x11x64xf32) matmul_99 = paddle._C_ops.matmul(scale_13, transpose_49, False, True) + del scale_13, transpose_49 # pd_op.add: (1x16x11x11xf32) <- (1x16x11x11xf32, 1x1x1x11xf32) add_113 = paddle._C_ops.add(matmul_99, unsqueeze_0) + del matmul_99 # pd_op.softmax: (1x16x11x11xf32) <- (1x16x11x11xf32) softmax_12 = paddle._C_ops.softmax(add_113, -1) @@ -2614,13 +2456,15 @@ def forward( # pd_op.dropout: (1x16x11x11xf32, 1x16x11x11xui8) <- (1x16x11x11xf32, None, 1xf32) dropout_74, dropout_75 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_12, None, full_4, False, "upscale_in_train", 0, False + softmax_12, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_12 # pd_op.matmul: (1x16x11x64xf32) <- (1x16x11x11xf32, 1x16x11x64xf32) matmul_100 = paddle._C_ops.matmul(dropout_74, transpose_50, False, False) + del dropout_74, transpose_50 # pd_op.transpose: (1x11x16x64xf32) <- (1x16x11x64xf32) transpose_51 = paddle._C_ops.transpose(matmul_100, [0, 2, 1, 3]) @@ -2628,19 +2472,20 @@ def forward( # pd_op.reshape: (1x11x1024xf32) <- (1x11x16x64xf32, 3xi64) reshape_51 = paddle._C_ops.reshape(transpose_51, full_int_array_2) + del transpose_51 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_101 = paddle._C_ops.matmul(reshape_51, parameter_187, False, False) - del parameter_187 + del parameter_187, reshape_51 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_114 = paddle._C_ops.add(matmul_101, parameter_186) - del parameter_186 + del matmul_101, parameter_186 # pd_op.dropout: (1x11x1024xf32, 1x11x1024xui8) <- (1x11x1024xf32, None, 1xf32) dropout_76, dropout_77 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_114, None, full_4, False, "upscale_in_train", 0, False + add_114, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -2648,6 +2493,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_115 = paddle._C_ops.add(layer_norm_72, dropout_76) + del dropout_76, layer_norm_72 # pd_op.layer_norm: (1x11x1024xf32, 1x11xf32, 1x11xf32) <- (1x11x1024xf32, 1024xf32, 1024xf32) layer_norm_75, layer_norm_76, layer_norm_77 = (lambda x, f: f(x))( @@ -2656,7 +2502,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_180, parameter_181 + del add_115, parameter_180, parameter_181 # pd_op.matmul: (1x11x4096xf32) <- (1x11x1024xf32, 1024x4096xf32) matmul_102 = paddle._C_ops.matmul(layer_norm_75, parameter_185, False, False) @@ -2664,7 +2510,7 @@ def forward( # pd_op.add: (1x11x4096xf32) <- (1x11x4096xf32, 4096xf32) add_116 = paddle._C_ops.add(matmul_102, parameter_184) - del parameter_184 + del matmul_102, parameter_184 # pd_op.relu: (1x11x4096xf32) <- (1x11x4096xf32) relu_12 = paddle._C_ops.relu(add_116) @@ -2672,16 +2518,16 @@ def forward( # pd_op.matmul: (1x11x1024xf32) <- (1x11x4096xf32, 4096x1024xf32) matmul_103 = paddle._C_ops.matmul(relu_12, parameter_183, False, False) - del parameter_183 + del parameter_183, relu_12 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_117 = paddle._C_ops.add(matmul_103, parameter_182) - del parameter_182 + del matmul_103, parameter_182 # pd_op.dropout: (1x11x1024xf32, 1x11x1024xui8) <- (1x11x1024xf32, None, 1xf32) dropout_78, dropout_79 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_117, None, full_4, False, "upscale_in_train", 0, False + add_117, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -2689,6 +2535,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_118 = paddle._C_ops.add(layer_norm_75, dropout_78) + del dropout_78, layer_norm_75 # pd_op.layer_norm: (1x11x1024xf32, 1x11xf32, 1x11xf32) <- (1x11x1024xf32, 1024xf32, 1024xf32) layer_norm_78, layer_norm_79, layer_norm_80 = (lambda x, f: f(x))( @@ -2697,7 +2544,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_178, parameter_179 + del add_118, parameter_178, parameter_179 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_104 = paddle._C_ops.matmul(layer_norm_78, parameter_177, False, False) @@ -2705,10 +2552,11 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_119 = paddle._C_ops.add(matmul_104, parameter_176) - del parameter_176 + del matmul_104, parameter_176 # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_52 = paddle._C_ops.reshape(add_119, full_int_array_1) + del add_119 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_52 = paddle._C_ops.transpose(reshape_52, [0, 2, 1, 3]) @@ -2720,7 +2568,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_120 = paddle._C_ops.add(matmul_105, parameter_174) - del parameter_174 + del matmul_105, parameter_174 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_106 = paddle._C_ops.matmul(layer_norm_78, parameter_173, False, False) @@ -2728,10 +2576,11 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_121 = paddle._C_ops.add(matmul_106, parameter_172) - del parameter_172 + del matmul_106, parameter_172 # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_53 = paddle._C_ops.reshape(add_120, full_int_array_1) + del add_120 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_53 = paddle._C_ops.transpose(reshape_53, [0, 2, 1, 3]) @@ -2739,6 +2588,7 @@ def forward( # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_54 = paddle._C_ops.reshape(add_121, full_int_array_1) + del add_121 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_54 = paddle._C_ops.transpose(reshape_54, [0, 2, 1, 3]) @@ -2750,9 +2600,11 @@ def forward( # pd_op.matmul: (1x16x11x11xf32) <- (1x16x11x64xf32, 1x16x11x64xf32) matmul_107 = paddle._C_ops.matmul(scale_14, transpose_53, False, True) + del scale_14, transpose_53 # pd_op.add: (1x16x11x11xf32) <- (1x16x11x11xf32, 1x1x1x11xf32) add_122 = paddle._C_ops.add(matmul_107, unsqueeze_0) + del matmul_107 # pd_op.softmax: (1x16x11x11xf32) <- (1x16x11x11xf32) softmax_13 = paddle._C_ops.softmax(add_122, -1) @@ -2761,13 +2613,15 @@ def forward( # pd_op.dropout: (1x16x11x11xf32, 1x16x11x11xui8) <- (1x16x11x11xf32, None, 1xf32) dropout_80, dropout_81 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_13, None, full_4, False, "upscale_in_train", 0, False + softmax_13, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_13 # pd_op.matmul: (1x16x11x64xf32) <- (1x16x11x11xf32, 1x16x11x64xf32) matmul_108 = paddle._C_ops.matmul(dropout_80, transpose_54, False, False) + del dropout_80, transpose_54 # pd_op.transpose: (1x11x16x64xf32) <- (1x16x11x64xf32) transpose_55 = paddle._C_ops.transpose(matmul_108, [0, 2, 1, 3]) @@ -2775,19 +2629,20 @@ def forward( # pd_op.reshape: (1x11x1024xf32) <- (1x11x16x64xf32, 3xi64) reshape_55 = paddle._C_ops.reshape(transpose_55, full_int_array_2) + del transpose_55 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_109 = paddle._C_ops.matmul(reshape_55, parameter_171, False, False) - del parameter_171 + del parameter_171, reshape_55 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_123 = paddle._C_ops.add(matmul_109, parameter_170) - del parameter_170 + del matmul_109, parameter_170 # pd_op.dropout: (1x11x1024xf32, 1x11x1024xui8) <- (1x11x1024xf32, None, 1xf32) dropout_82, dropout_83 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_123, None, full_4, False, "upscale_in_train", 0, False + add_123, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -2795,6 +2650,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_124 = paddle._C_ops.add(layer_norm_78, dropout_82) + del dropout_82, layer_norm_78 # pd_op.layer_norm: (1x11x1024xf32, 1x11xf32, 1x11xf32) <- (1x11x1024xf32, 1024xf32, 1024xf32) layer_norm_81, layer_norm_82, layer_norm_83 = (lambda x, f: f(x))( @@ -2803,7 +2659,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_164, parameter_165 + del add_124, parameter_164, parameter_165 # pd_op.matmul: (1x11x4096xf32) <- (1x11x1024xf32, 1024x4096xf32) matmul_110 = paddle._C_ops.matmul(layer_norm_81, parameter_169, False, False) @@ -2811,7 +2667,7 @@ def forward( # pd_op.add: (1x11x4096xf32) <- (1x11x4096xf32, 4096xf32) add_125 = paddle._C_ops.add(matmul_110, parameter_168) - del parameter_168 + del matmul_110, parameter_168 # pd_op.relu: (1x11x4096xf32) <- (1x11x4096xf32) relu_13 = paddle._C_ops.relu(add_125) @@ -2819,16 +2675,16 @@ def forward( # pd_op.matmul: (1x11x1024xf32) <- (1x11x4096xf32, 4096x1024xf32) matmul_111 = paddle._C_ops.matmul(relu_13, parameter_167, False, False) - del parameter_167 + del parameter_167, relu_13 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_126 = paddle._C_ops.add(matmul_111, parameter_166) - del parameter_166 + del matmul_111, parameter_166 # pd_op.dropout: (1x11x1024xf32, 1x11x1024xui8) <- (1x11x1024xf32, None, 1xf32) dropout_84, dropout_85 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_126, None, full_4, False, "upscale_in_train", 0, False + add_126, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -2836,6 +2692,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_127 = paddle._C_ops.add(layer_norm_81, dropout_84) + del dropout_84, layer_norm_81 # pd_op.layer_norm: (1x11x1024xf32, 1x11xf32, 1x11xf32) <- (1x11x1024xf32, 1024xf32, 1024xf32) layer_norm_84, layer_norm_85, layer_norm_86 = (lambda x, f: f(x))( @@ -2844,7 +2701,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_162, parameter_163 + del add_127, parameter_162, parameter_163 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_112 = paddle._C_ops.matmul(layer_norm_84, parameter_161, False, False) @@ -2852,10 +2709,11 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_128 = paddle._C_ops.add(matmul_112, parameter_160) - del parameter_160 + del matmul_112, parameter_160 # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_56 = paddle._C_ops.reshape(add_128, full_int_array_1) + del add_128 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_56 = paddle._C_ops.transpose(reshape_56, [0, 2, 1, 3]) @@ -2867,7 +2725,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_129 = paddle._C_ops.add(matmul_113, parameter_158) - del parameter_158 + del matmul_113, parameter_158 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_114 = paddle._C_ops.matmul(layer_norm_84, parameter_157, False, False) @@ -2875,10 +2733,11 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_130 = paddle._C_ops.add(matmul_114, parameter_156) - del parameter_156 + del matmul_114, parameter_156 # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_57 = paddle._C_ops.reshape(add_129, full_int_array_1) + del add_129 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_57 = paddle._C_ops.transpose(reshape_57, [0, 2, 1, 3]) @@ -2886,6 +2745,7 @@ def forward( # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_58 = paddle._C_ops.reshape(add_130, full_int_array_1) + del add_130 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_58 = paddle._C_ops.transpose(reshape_58, [0, 2, 1, 3]) @@ -2897,9 +2757,11 @@ def forward( # pd_op.matmul: (1x16x11x11xf32) <- (1x16x11x64xf32, 1x16x11x64xf32) matmul_115 = paddle._C_ops.matmul(scale_15, transpose_57, False, True) + del scale_15, transpose_57 # pd_op.add: (1x16x11x11xf32) <- (1x16x11x11xf32, 1x1x1x11xf32) add_131 = paddle._C_ops.add(matmul_115, unsqueeze_0) + del matmul_115 # pd_op.softmax: (1x16x11x11xf32) <- (1x16x11x11xf32) softmax_14 = paddle._C_ops.softmax(add_131, -1) @@ -2908,13 +2770,15 @@ def forward( # pd_op.dropout: (1x16x11x11xf32, 1x16x11x11xui8) <- (1x16x11x11xf32, None, 1xf32) dropout_86, dropout_87 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_14, None, full_4, False, "upscale_in_train", 0, False + softmax_14, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_14 # pd_op.matmul: (1x16x11x64xf32) <- (1x16x11x11xf32, 1x16x11x64xf32) matmul_116 = paddle._C_ops.matmul(dropout_86, transpose_58, False, False) + del dropout_86, transpose_58 # pd_op.transpose: (1x11x16x64xf32) <- (1x16x11x64xf32) transpose_59 = paddle._C_ops.transpose(matmul_116, [0, 2, 1, 3]) @@ -2922,19 +2786,20 @@ def forward( # pd_op.reshape: (1x11x1024xf32) <- (1x11x16x64xf32, 3xi64) reshape_59 = paddle._C_ops.reshape(transpose_59, full_int_array_2) + del transpose_59 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_117 = paddle._C_ops.matmul(reshape_59, parameter_155, False, False) - del parameter_155 + del parameter_155, reshape_59 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_132 = paddle._C_ops.add(matmul_117, parameter_154) - del parameter_154 + del matmul_117, parameter_154 # pd_op.dropout: (1x11x1024xf32, 1x11x1024xui8) <- (1x11x1024xf32, None, 1xf32) dropout_88, dropout_89 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_132, None, full_4, False, "upscale_in_train", 0, False + add_132, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -2942,6 +2807,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_133 = paddle._C_ops.add(layer_norm_84, dropout_88) + del dropout_88, layer_norm_84 # pd_op.layer_norm: (1x11x1024xf32, 1x11xf32, 1x11xf32) <- (1x11x1024xf32, 1024xf32, 1024xf32) layer_norm_87, layer_norm_88, layer_norm_89 = (lambda x, f: f(x))( @@ -2950,7 +2816,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_148, parameter_149 + del add_133, parameter_148, parameter_149 # pd_op.matmul: (1x11x4096xf32) <- (1x11x1024xf32, 1024x4096xf32) matmul_118 = paddle._C_ops.matmul(layer_norm_87, parameter_153, False, False) @@ -2958,7 +2824,7 @@ def forward( # pd_op.add: (1x11x4096xf32) <- (1x11x4096xf32, 4096xf32) add_134 = paddle._C_ops.add(matmul_118, parameter_152) - del parameter_152 + del matmul_118, parameter_152 # pd_op.relu: (1x11x4096xf32) <- (1x11x4096xf32) relu_14 = paddle._C_ops.relu(add_134) @@ -2966,16 +2832,16 @@ def forward( # pd_op.matmul: (1x11x1024xf32) <- (1x11x4096xf32, 4096x1024xf32) matmul_119 = paddle._C_ops.matmul(relu_14, parameter_151, False, False) - del parameter_151 + del parameter_151, relu_14 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_135 = paddle._C_ops.add(matmul_119, parameter_150) - del parameter_150 + del matmul_119, parameter_150 # pd_op.dropout: (1x11x1024xf32, 1x11x1024xui8) <- (1x11x1024xf32, None, 1xf32) dropout_90, dropout_91 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_135, None, full_4, False, "upscale_in_train", 0, False + add_135, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -2983,6 +2849,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_136 = paddle._C_ops.add(layer_norm_87, dropout_90) + del dropout_90, layer_norm_87 # pd_op.layer_norm: (1x11x1024xf32, 1x11xf32, 1x11xf32) <- (1x11x1024xf32, 1024xf32, 1024xf32) layer_norm_90, layer_norm_91, layer_norm_92 = (lambda x, f: f(x))( @@ -2991,7 +2858,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_146, parameter_147 + del add_136, parameter_146, parameter_147 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_120 = paddle._C_ops.matmul(layer_norm_90, parameter_145, False, False) @@ -2999,10 +2866,11 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_137 = paddle._C_ops.add(matmul_120, parameter_144) - del parameter_144 + del matmul_120, parameter_144 # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_60 = paddle._C_ops.reshape(add_137, full_int_array_1) + del add_137 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_60 = paddle._C_ops.transpose(reshape_60, [0, 2, 1, 3]) @@ -3014,7 +2882,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_138 = paddle._C_ops.add(matmul_121, parameter_142) - del parameter_142 + del matmul_121, parameter_142 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_122 = paddle._C_ops.matmul(layer_norm_90, parameter_141, False, False) @@ -3022,10 +2890,11 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_139 = paddle._C_ops.add(matmul_122, parameter_140) - del parameter_140 + del matmul_122, parameter_140 # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_61 = paddle._C_ops.reshape(add_138, full_int_array_1) + del add_138 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_61 = paddle._C_ops.transpose(reshape_61, [0, 2, 1, 3]) @@ -3033,6 +2902,7 @@ def forward( # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_62 = paddle._C_ops.reshape(add_139, full_int_array_1) + del add_139 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_62 = paddle._C_ops.transpose(reshape_62, [0, 2, 1, 3]) @@ -3044,9 +2914,11 @@ def forward( # pd_op.matmul: (1x16x11x11xf32) <- (1x16x11x64xf32, 1x16x11x64xf32) matmul_123 = paddle._C_ops.matmul(scale_16, transpose_61, False, True) + del scale_16, transpose_61 # pd_op.add: (1x16x11x11xf32) <- (1x16x11x11xf32, 1x1x1x11xf32) add_140 = paddle._C_ops.add(matmul_123, unsqueeze_0) + del matmul_123 # pd_op.softmax: (1x16x11x11xf32) <- (1x16x11x11xf32) softmax_15 = paddle._C_ops.softmax(add_140, -1) @@ -3055,13 +2927,15 @@ def forward( # pd_op.dropout: (1x16x11x11xf32, 1x16x11x11xui8) <- (1x16x11x11xf32, None, 1xf32) dropout_92, dropout_93 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_15, None, full_4, False, "upscale_in_train", 0, False + softmax_15, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_15 # pd_op.matmul: (1x16x11x64xf32) <- (1x16x11x11xf32, 1x16x11x64xf32) matmul_124 = paddle._C_ops.matmul(dropout_92, transpose_62, False, False) + del dropout_92, transpose_62 # pd_op.transpose: (1x11x16x64xf32) <- (1x16x11x64xf32) transpose_63 = paddle._C_ops.transpose(matmul_124, [0, 2, 1, 3]) @@ -3069,19 +2943,20 @@ def forward( # pd_op.reshape: (1x11x1024xf32) <- (1x11x16x64xf32, 3xi64) reshape_63 = paddle._C_ops.reshape(transpose_63, full_int_array_2) + del transpose_63 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_125 = paddle._C_ops.matmul(reshape_63, parameter_139, False, False) - del parameter_139 + del parameter_139, reshape_63 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_141 = paddle._C_ops.add(matmul_125, parameter_138) - del parameter_138 + del matmul_125, parameter_138 # pd_op.dropout: (1x11x1024xf32, 1x11x1024xui8) <- (1x11x1024xf32, None, 1xf32) dropout_94, dropout_95 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_141, None, full_4, False, "upscale_in_train", 0, False + add_141, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -3089,6 +2964,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_142 = paddle._C_ops.add(layer_norm_90, dropout_94) + del dropout_94, layer_norm_90 # pd_op.layer_norm: (1x11x1024xf32, 1x11xf32, 1x11xf32) <- (1x11x1024xf32, 1024xf32, 1024xf32) layer_norm_93, layer_norm_94, layer_norm_95 = (lambda x, f: f(x))( @@ -3097,7 +2973,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_132, parameter_133 + del add_142, parameter_132, parameter_133 # pd_op.matmul: (1x11x4096xf32) <- (1x11x1024xf32, 1024x4096xf32) matmul_126 = paddle._C_ops.matmul(layer_norm_93, parameter_137, False, False) @@ -3105,7 +2981,7 @@ def forward( # pd_op.add: (1x11x4096xf32) <- (1x11x4096xf32, 4096xf32) add_143 = paddle._C_ops.add(matmul_126, parameter_136) - del parameter_136 + del matmul_126, parameter_136 # pd_op.relu: (1x11x4096xf32) <- (1x11x4096xf32) relu_15 = paddle._C_ops.relu(add_143) @@ -3113,16 +2989,16 @@ def forward( # pd_op.matmul: (1x11x1024xf32) <- (1x11x4096xf32, 4096x1024xf32) matmul_127 = paddle._C_ops.matmul(relu_15, parameter_135, False, False) - del parameter_135 + del parameter_135, relu_15 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_144 = paddle._C_ops.add(matmul_127, parameter_134) - del parameter_134 + del matmul_127, parameter_134 # pd_op.dropout: (1x11x1024xf32, 1x11x1024xui8) <- (1x11x1024xf32, None, 1xf32) dropout_96, dropout_97 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_144, None, full_4, False, "upscale_in_train", 0, False + add_144, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -3130,6 +3006,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_145 = paddle._C_ops.add(layer_norm_93, dropout_96) + del dropout_96, layer_norm_93 # pd_op.layer_norm: (1x11x1024xf32, 1x11xf32, 1x11xf32) <- (1x11x1024xf32, 1024xf32, 1024xf32) layer_norm_96, layer_norm_97, layer_norm_98 = (lambda x, f: f(x))( @@ -3138,7 +3015,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_130, parameter_131 + del add_145, parameter_130, parameter_131 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_128 = paddle._C_ops.matmul(layer_norm_96, parameter_129, False, False) @@ -3146,10 +3023,11 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_146 = paddle._C_ops.add(matmul_128, parameter_128) - del parameter_128 + del matmul_128, parameter_128 # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_64 = paddle._C_ops.reshape(add_146, full_int_array_1) + del add_146 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_64 = paddle._C_ops.transpose(reshape_64, [0, 2, 1, 3]) @@ -3161,7 +3039,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_147 = paddle._C_ops.add(matmul_129, parameter_126) - del parameter_126 + del matmul_129, parameter_126 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_130 = paddle._C_ops.matmul(layer_norm_96, parameter_125, False, False) @@ -3169,10 +3047,11 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_148 = paddle._C_ops.add(matmul_130, parameter_124) - del parameter_124 + del matmul_130, parameter_124 # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_65 = paddle._C_ops.reshape(add_147, full_int_array_1) + del add_147 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_65 = paddle._C_ops.transpose(reshape_65, [0, 2, 1, 3]) @@ -3180,6 +3059,7 @@ def forward( # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_66 = paddle._C_ops.reshape(add_148, full_int_array_1) + del add_148 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_66 = paddle._C_ops.transpose(reshape_66, [0, 2, 1, 3]) @@ -3191,9 +3071,11 @@ def forward( # pd_op.matmul: (1x16x11x11xf32) <- (1x16x11x64xf32, 1x16x11x64xf32) matmul_131 = paddle._C_ops.matmul(scale_17, transpose_65, False, True) + del scale_17, transpose_65 # pd_op.add: (1x16x11x11xf32) <- (1x16x11x11xf32, 1x1x1x11xf32) add_149 = paddle._C_ops.add(matmul_131, unsqueeze_0) + del matmul_131 # pd_op.softmax: (1x16x11x11xf32) <- (1x16x11x11xf32) softmax_16 = paddle._C_ops.softmax(add_149, -1) @@ -3202,13 +3084,15 @@ def forward( # pd_op.dropout: (1x16x11x11xf32, 1x16x11x11xui8) <- (1x16x11x11xf32, None, 1xf32) dropout_98, dropout_99 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_16, None, full_4, False, "upscale_in_train", 0, False + softmax_16, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_16 # pd_op.matmul: (1x16x11x64xf32) <- (1x16x11x11xf32, 1x16x11x64xf32) matmul_132 = paddle._C_ops.matmul(dropout_98, transpose_66, False, False) + del dropout_98, transpose_66 # pd_op.transpose: (1x11x16x64xf32) <- (1x16x11x64xf32) transpose_67 = paddle._C_ops.transpose(matmul_132, [0, 2, 1, 3]) @@ -3216,19 +3100,20 @@ def forward( # pd_op.reshape: (1x11x1024xf32) <- (1x11x16x64xf32, 3xi64) reshape_67 = paddle._C_ops.reshape(transpose_67, full_int_array_2) + del transpose_67 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_133 = paddle._C_ops.matmul(reshape_67, parameter_123, False, False) - del parameter_123 + del parameter_123, reshape_67 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_150 = paddle._C_ops.add(matmul_133, parameter_122) - del parameter_122 + del matmul_133, parameter_122 # pd_op.dropout: (1x11x1024xf32, 1x11x1024xui8) <- (1x11x1024xf32, None, 1xf32) dropout_100, dropout_101 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_150, None, full_4, False, "upscale_in_train", 0, False + add_150, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -3236,6 +3121,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_151 = paddle._C_ops.add(layer_norm_96, dropout_100) + del dropout_100, layer_norm_96 # pd_op.layer_norm: (1x11x1024xf32, 1x11xf32, 1x11xf32) <- (1x11x1024xf32, 1024xf32, 1024xf32) layer_norm_99, layer_norm_100, layer_norm_101 = (lambda x, f: f(x))( @@ -3244,7 +3130,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_116, parameter_117 + del add_151, parameter_116, parameter_117 # pd_op.matmul: (1x11x4096xf32) <- (1x11x1024xf32, 1024x4096xf32) matmul_134 = paddle._C_ops.matmul(layer_norm_99, parameter_121, False, False) @@ -3252,7 +3138,7 @@ def forward( # pd_op.add: (1x11x4096xf32) <- (1x11x4096xf32, 4096xf32) add_152 = paddle._C_ops.add(matmul_134, parameter_120) - del parameter_120 + del matmul_134, parameter_120 # pd_op.relu: (1x11x4096xf32) <- (1x11x4096xf32) relu_16 = paddle._C_ops.relu(add_152) @@ -3260,16 +3146,16 @@ def forward( # pd_op.matmul: (1x11x1024xf32) <- (1x11x4096xf32, 4096x1024xf32) matmul_135 = paddle._C_ops.matmul(relu_16, parameter_119, False, False) - del parameter_119 + del parameter_119, relu_16 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_153 = paddle._C_ops.add(matmul_135, parameter_118) - del parameter_118 + del matmul_135, parameter_118 # pd_op.dropout: (1x11x1024xf32, 1x11x1024xui8) <- (1x11x1024xf32, None, 1xf32) dropout_102, dropout_103 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_153, None, full_4, False, "upscale_in_train", 0, False + add_153, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -3277,6 +3163,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_154 = paddle._C_ops.add(layer_norm_99, dropout_102) + del dropout_102, layer_norm_99 # pd_op.layer_norm: (1x11x1024xf32, 1x11xf32, 1x11xf32) <- (1x11x1024xf32, 1024xf32, 1024xf32) layer_norm_102, layer_norm_103, layer_norm_104 = (lambda x, f: f(x))( @@ -3285,7 +3172,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_114, parameter_115 + del add_154, parameter_114, parameter_115 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_136 = paddle._C_ops.matmul(layer_norm_102, parameter_113, False, False) @@ -3293,10 +3180,11 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_155 = paddle._C_ops.add(matmul_136, parameter_112) - del parameter_112 + del matmul_136, parameter_112 # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_68 = paddle._C_ops.reshape(add_155, full_int_array_1) + del add_155 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_68 = paddle._C_ops.transpose(reshape_68, [0, 2, 1, 3]) @@ -3308,7 +3196,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_156 = paddle._C_ops.add(matmul_137, parameter_110) - del parameter_110 + del matmul_137, parameter_110 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_138 = paddle._C_ops.matmul(layer_norm_102, parameter_109, False, False) @@ -3316,10 +3204,11 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_157 = paddle._C_ops.add(matmul_138, parameter_108) - del parameter_108 + del matmul_138, parameter_108 # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_69 = paddle._C_ops.reshape(add_156, full_int_array_1) + del add_156 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_69 = paddle._C_ops.transpose(reshape_69, [0, 2, 1, 3]) @@ -3327,6 +3216,7 @@ def forward( # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_70 = paddle._C_ops.reshape(add_157, full_int_array_1) + del add_157 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_70 = paddle._C_ops.transpose(reshape_70, [0, 2, 1, 3]) @@ -3338,9 +3228,11 @@ def forward( # pd_op.matmul: (1x16x11x11xf32) <- (1x16x11x64xf32, 1x16x11x64xf32) matmul_139 = paddle._C_ops.matmul(scale_18, transpose_69, False, True) + del scale_18, transpose_69 # pd_op.add: (1x16x11x11xf32) <- (1x16x11x11xf32, 1x1x1x11xf32) add_158 = paddle._C_ops.add(matmul_139, unsqueeze_0) + del matmul_139 # pd_op.softmax: (1x16x11x11xf32) <- (1x16x11x11xf32) softmax_17 = paddle._C_ops.softmax(add_158, -1) @@ -3349,13 +3241,15 @@ def forward( # pd_op.dropout: (1x16x11x11xf32, 1x16x11x11xui8) <- (1x16x11x11xf32, None, 1xf32) dropout_104, dropout_105 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_17, None, full_4, False, "upscale_in_train", 0, False + softmax_17, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_17 # pd_op.matmul: (1x16x11x64xf32) <- (1x16x11x11xf32, 1x16x11x64xf32) matmul_140 = paddle._C_ops.matmul(dropout_104, transpose_70, False, False) + del dropout_104, transpose_70 # pd_op.transpose: (1x11x16x64xf32) <- (1x16x11x64xf32) transpose_71 = paddle._C_ops.transpose(matmul_140, [0, 2, 1, 3]) @@ -3363,19 +3257,20 @@ def forward( # pd_op.reshape: (1x11x1024xf32) <- (1x11x16x64xf32, 3xi64) reshape_71 = paddle._C_ops.reshape(transpose_71, full_int_array_2) + del transpose_71 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_141 = paddle._C_ops.matmul(reshape_71, parameter_107, False, False) - del parameter_107 + del parameter_107, reshape_71 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_159 = paddle._C_ops.add(matmul_141, parameter_106) - del parameter_106 + del matmul_141, parameter_106 # pd_op.dropout: (1x11x1024xf32, 1x11x1024xui8) <- (1x11x1024xf32, None, 1xf32) dropout_106, dropout_107 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_159, None, full_4, False, "upscale_in_train", 0, False + add_159, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -3383,6 +3278,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_160 = paddle._C_ops.add(layer_norm_102, dropout_106) + del dropout_106, layer_norm_102 # pd_op.layer_norm: (1x11x1024xf32, 1x11xf32, 1x11xf32) <- (1x11x1024xf32, 1024xf32, 1024xf32) layer_norm_105, layer_norm_106, layer_norm_107 = (lambda x, f: f(x))( @@ -3391,7 +3287,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_100, parameter_101 + del add_160, parameter_100, parameter_101 # pd_op.matmul: (1x11x4096xf32) <- (1x11x1024xf32, 1024x4096xf32) matmul_142 = paddle._C_ops.matmul(layer_norm_105, parameter_105, False, False) @@ -3399,7 +3295,7 @@ def forward( # pd_op.add: (1x11x4096xf32) <- (1x11x4096xf32, 4096xf32) add_161 = paddle._C_ops.add(matmul_142, parameter_104) - del parameter_104 + del matmul_142, parameter_104 # pd_op.relu: (1x11x4096xf32) <- (1x11x4096xf32) relu_17 = paddle._C_ops.relu(add_161) @@ -3407,16 +3303,16 @@ def forward( # pd_op.matmul: (1x11x1024xf32) <- (1x11x4096xf32, 4096x1024xf32) matmul_143 = paddle._C_ops.matmul(relu_17, parameter_103, False, False) - del parameter_103 + del parameter_103, relu_17 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_162 = paddle._C_ops.add(matmul_143, parameter_102) - del parameter_102 + del matmul_143, parameter_102 # pd_op.dropout: (1x11x1024xf32, 1x11x1024xui8) <- (1x11x1024xf32, None, 1xf32) dropout_108, dropout_109 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_162, None, full_4, False, "upscale_in_train", 0, False + add_162, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -3424,6 +3320,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_163 = paddle._C_ops.add(layer_norm_105, dropout_108) + del dropout_108, layer_norm_105 # pd_op.layer_norm: (1x11x1024xf32, 1x11xf32, 1x11xf32) <- (1x11x1024xf32, 1024xf32, 1024xf32) layer_norm_108, layer_norm_109, layer_norm_110 = (lambda x, f: f(x))( @@ -3432,7 +3329,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_98, parameter_99 + del add_163, parameter_98, parameter_99 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_144 = paddle._C_ops.matmul(layer_norm_108, parameter_97, False, False) @@ -3440,10 +3337,11 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_164 = paddle._C_ops.add(matmul_144, parameter_96) - del parameter_96 + del matmul_144, parameter_96 # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_72 = paddle._C_ops.reshape(add_164, full_int_array_1) + del add_164 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_72 = paddle._C_ops.transpose(reshape_72, [0, 2, 1, 3]) @@ -3455,7 +3353,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_165 = paddle._C_ops.add(matmul_145, parameter_94) - del parameter_94 + del matmul_145, parameter_94 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_146 = paddle._C_ops.matmul(layer_norm_108, parameter_93, False, False) @@ -3463,10 +3361,11 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_166 = paddle._C_ops.add(matmul_146, parameter_92) - del parameter_92 + del matmul_146, parameter_92 # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_73 = paddle._C_ops.reshape(add_165, full_int_array_1) + del add_165 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_73 = paddle._C_ops.transpose(reshape_73, [0, 2, 1, 3]) @@ -3474,6 +3373,7 @@ def forward( # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_74 = paddle._C_ops.reshape(add_166, full_int_array_1) + del add_166 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_74 = paddle._C_ops.transpose(reshape_74, [0, 2, 1, 3]) @@ -3485,9 +3385,11 @@ def forward( # pd_op.matmul: (1x16x11x11xf32) <- (1x16x11x64xf32, 1x16x11x64xf32) matmul_147 = paddle._C_ops.matmul(scale_19, transpose_73, False, True) + del scale_19, transpose_73 # pd_op.add: (1x16x11x11xf32) <- (1x16x11x11xf32, 1x1x1x11xf32) add_167 = paddle._C_ops.add(matmul_147, unsqueeze_0) + del matmul_147 # pd_op.softmax: (1x16x11x11xf32) <- (1x16x11x11xf32) softmax_18 = paddle._C_ops.softmax(add_167, -1) @@ -3496,13 +3398,15 @@ def forward( # pd_op.dropout: (1x16x11x11xf32, 1x16x11x11xui8) <- (1x16x11x11xf32, None, 1xf32) dropout_110, dropout_111 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_18, None, full_4, False, "upscale_in_train", 0, False + softmax_18, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_18 # pd_op.matmul: (1x16x11x64xf32) <- (1x16x11x11xf32, 1x16x11x64xf32) matmul_148 = paddle._C_ops.matmul(dropout_110, transpose_74, False, False) + del dropout_110, transpose_74 # pd_op.transpose: (1x11x16x64xf32) <- (1x16x11x64xf32) transpose_75 = paddle._C_ops.transpose(matmul_148, [0, 2, 1, 3]) @@ -3510,19 +3414,20 @@ def forward( # pd_op.reshape: (1x11x1024xf32) <- (1x11x16x64xf32, 3xi64) reshape_75 = paddle._C_ops.reshape(transpose_75, full_int_array_2) + del transpose_75 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_149 = paddle._C_ops.matmul(reshape_75, parameter_91, False, False) - del parameter_91 + del parameter_91, reshape_75 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_168 = paddle._C_ops.add(matmul_149, parameter_90) - del parameter_90 + del matmul_149, parameter_90 # pd_op.dropout: (1x11x1024xf32, 1x11x1024xui8) <- (1x11x1024xf32, None, 1xf32) dropout_112, dropout_113 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_168, None, full_4, False, "upscale_in_train", 0, False + add_168, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -3530,6 +3435,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_169 = paddle._C_ops.add(layer_norm_108, dropout_112) + del dropout_112, layer_norm_108 # pd_op.layer_norm: (1x11x1024xf32, 1x11xf32, 1x11xf32) <- (1x11x1024xf32, 1024xf32, 1024xf32) layer_norm_111, layer_norm_112, layer_norm_113 = (lambda x, f: f(x))( @@ -3538,7 +3444,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_84, parameter_85 + del add_169, parameter_84, parameter_85 # pd_op.matmul: (1x11x4096xf32) <- (1x11x1024xf32, 1024x4096xf32) matmul_150 = paddle._C_ops.matmul(layer_norm_111, parameter_89, False, False) @@ -3546,7 +3452,7 @@ def forward( # pd_op.add: (1x11x4096xf32) <- (1x11x4096xf32, 4096xf32) add_170 = paddle._C_ops.add(matmul_150, parameter_88) - del parameter_88 + del matmul_150, parameter_88 # pd_op.relu: (1x11x4096xf32) <- (1x11x4096xf32) relu_18 = paddle._C_ops.relu(add_170) @@ -3554,16 +3460,16 @@ def forward( # pd_op.matmul: (1x11x1024xf32) <- (1x11x4096xf32, 4096x1024xf32) matmul_151 = paddle._C_ops.matmul(relu_18, parameter_87, False, False) - del parameter_87 + del parameter_87, relu_18 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_171 = paddle._C_ops.add(matmul_151, parameter_86) - del parameter_86 + del matmul_151, parameter_86 # pd_op.dropout: (1x11x1024xf32, 1x11x1024xui8) <- (1x11x1024xf32, None, 1xf32) dropout_114, dropout_115 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_171, None, full_4, False, "upscale_in_train", 0, False + add_171, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -3571,6 +3477,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_172 = paddle._C_ops.add(layer_norm_111, dropout_114) + del dropout_114, layer_norm_111 # pd_op.layer_norm: (1x11x1024xf32, 1x11xf32, 1x11xf32) <- (1x11x1024xf32, 1024xf32, 1024xf32) layer_norm_114, layer_norm_115, layer_norm_116 = (lambda x, f: f(x))( @@ -3579,7 +3486,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_82, parameter_83 + del add_172, parameter_82, parameter_83 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_152 = paddle._C_ops.matmul(layer_norm_114, parameter_81, False, False) @@ -3587,10 +3494,11 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_173 = paddle._C_ops.add(matmul_152, parameter_80) - del parameter_80 + del matmul_152, parameter_80 # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_76 = paddle._C_ops.reshape(add_173, full_int_array_1) + del add_173 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_76 = paddle._C_ops.transpose(reshape_76, [0, 2, 1, 3]) @@ -3602,7 +3510,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_174 = paddle._C_ops.add(matmul_153, parameter_78) - del parameter_78 + del matmul_153, parameter_78 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_154 = paddle._C_ops.matmul(layer_norm_114, parameter_77, False, False) @@ -3610,10 +3518,11 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_175 = paddle._C_ops.add(matmul_154, parameter_76) - del parameter_76 + del matmul_154, parameter_76 # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_77 = paddle._C_ops.reshape(add_174, full_int_array_1) + del add_174 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_77 = paddle._C_ops.transpose(reshape_77, [0, 2, 1, 3]) @@ -3621,6 +3530,7 @@ def forward( # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_78 = paddle._C_ops.reshape(add_175, full_int_array_1) + del add_175 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_78 = paddle._C_ops.transpose(reshape_78, [0, 2, 1, 3]) @@ -3632,9 +3542,11 @@ def forward( # pd_op.matmul: (1x16x11x11xf32) <- (1x16x11x64xf32, 1x16x11x64xf32) matmul_155 = paddle._C_ops.matmul(scale_20, transpose_77, False, True) + del scale_20, transpose_77 # pd_op.add: (1x16x11x11xf32) <- (1x16x11x11xf32, 1x1x1x11xf32) add_176 = paddle._C_ops.add(matmul_155, unsqueeze_0) + del matmul_155 # pd_op.softmax: (1x16x11x11xf32) <- (1x16x11x11xf32) softmax_19 = paddle._C_ops.softmax(add_176, -1) @@ -3643,13 +3555,15 @@ def forward( # pd_op.dropout: (1x16x11x11xf32, 1x16x11x11xui8) <- (1x16x11x11xf32, None, 1xf32) dropout_116, dropout_117 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_19, None, full_4, False, "upscale_in_train", 0, False + softmax_19, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_19 # pd_op.matmul: (1x16x11x64xf32) <- (1x16x11x11xf32, 1x16x11x64xf32) matmul_156 = paddle._C_ops.matmul(dropout_116, transpose_78, False, False) + del dropout_116, transpose_78 # pd_op.transpose: (1x11x16x64xf32) <- (1x16x11x64xf32) transpose_79 = paddle._C_ops.transpose(matmul_156, [0, 2, 1, 3]) @@ -3657,19 +3571,20 @@ def forward( # pd_op.reshape: (1x11x1024xf32) <- (1x11x16x64xf32, 3xi64) reshape_79 = paddle._C_ops.reshape(transpose_79, full_int_array_2) + del transpose_79 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_157 = paddle._C_ops.matmul(reshape_79, parameter_75, False, False) - del parameter_75 + del parameter_75, reshape_79 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_177 = paddle._C_ops.add(matmul_157, parameter_74) - del parameter_74 + del matmul_157, parameter_74 # pd_op.dropout: (1x11x1024xf32, 1x11x1024xui8) <- (1x11x1024xf32, None, 1xf32) dropout_118, dropout_119 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_177, None, full_4, False, "upscale_in_train", 0, False + add_177, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -3677,6 +3592,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_178 = paddle._C_ops.add(layer_norm_114, dropout_118) + del dropout_118, layer_norm_114 # pd_op.layer_norm: (1x11x1024xf32, 1x11xf32, 1x11xf32) <- (1x11x1024xf32, 1024xf32, 1024xf32) layer_norm_117, layer_norm_118, layer_norm_119 = (lambda x, f: f(x))( @@ -3685,7 +3601,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_68, parameter_69 + del add_178, parameter_68, parameter_69 # pd_op.matmul: (1x11x4096xf32) <- (1x11x1024xf32, 1024x4096xf32) matmul_158 = paddle._C_ops.matmul(layer_norm_117, parameter_73, False, False) @@ -3693,7 +3609,7 @@ def forward( # pd_op.add: (1x11x4096xf32) <- (1x11x4096xf32, 4096xf32) add_179 = paddle._C_ops.add(matmul_158, parameter_72) - del parameter_72 + del matmul_158, parameter_72 # pd_op.relu: (1x11x4096xf32) <- (1x11x4096xf32) relu_19 = paddle._C_ops.relu(add_179) @@ -3701,16 +3617,16 @@ def forward( # pd_op.matmul: (1x11x1024xf32) <- (1x11x4096xf32, 4096x1024xf32) matmul_159 = paddle._C_ops.matmul(relu_19, parameter_71, False, False) - del parameter_71 + del parameter_71, relu_19 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_180 = paddle._C_ops.add(matmul_159, parameter_70) - del parameter_70 + del matmul_159, parameter_70 # pd_op.dropout: (1x11x1024xf32, 1x11x1024xui8) <- (1x11x1024xf32, None, 1xf32) dropout_120, dropout_121 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_180, None, full_4, False, "upscale_in_train", 0, False + add_180, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -3718,6 +3634,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_181 = paddle._C_ops.add(layer_norm_117, dropout_120) + del dropout_120, layer_norm_117 # pd_op.layer_norm: (1x11x1024xf32, 1x11xf32, 1x11xf32) <- (1x11x1024xf32, 1024xf32, 1024xf32) layer_norm_120, layer_norm_121, layer_norm_122 = (lambda x, f: f(x))( @@ -3726,7 +3643,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_66, parameter_67 + del add_181, parameter_66, parameter_67 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_160 = paddle._C_ops.matmul(layer_norm_120, parameter_65, False, False) @@ -3734,10 +3651,11 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_182 = paddle._C_ops.add(matmul_160, parameter_64) - del parameter_64 + del matmul_160, parameter_64 # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_80 = paddle._C_ops.reshape(add_182, full_int_array_1) + del add_182 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_80 = paddle._C_ops.transpose(reshape_80, [0, 2, 1, 3]) @@ -3749,7 +3667,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_183 = paddle._C_ops.add(matmul_161, parameter_62) - del parameter_62 + del matmul_161, parameter_62 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_162 = paddle._C_ops.matmul(layer_norm_120, parameter_61, False, False) @@ -3757,10 +3675,11 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_184 = paddle._C_ops.add(matmul_162, parameter_60) - del parameter_60 + del matmul_162, parameter_60 # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_81 = paddle._C_ops.reshape(add_183, full_int_array_1) + del add_183 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_81 = paddle._C_ops.transpose(reshape_81, [0, 2, 1, 3]) @@ -3768,6 +3687,7 @@ def forward( # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_82 = paddle._C_ops.reshape(add_184, full_int_array_1) + del add_184 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_82 = paddle._C_ops.transpose(reshape_82, [0, 2, 1, 3]) @@ -3779,9 +3699,11 @@ def forward( # pd_op.matmul: (1x16x11x11xf32) <- (1x16x11x64xf32, 1x16x11x64xf32) matmul_163 = paddle._C_ops.matmul(scale_21, transpose_81, False, True) + del scale_21, transpose_81 # pd_op.add: (1x16x11x11xf32) <- (1x16x11x11xf32, 1x1x1x11xf32) add_185 = paddle._C_ops.add(matmul_163, unsqueeze_0) + del matmul_163 # pd_op.softmax: (1x16x11x11xf32) <- (1x16x11x11xf32) softmax_20 = paddle._C_ops.softmax(add_185, -1) @@ -3790,13 +3712,15 @@ def forward( # pd_op.dropout: (1x16x11x11xf32, 1x16x11x11xui8) <- (1x16x11x11xf32, None, 1xf32) dropout_122, dropout_123 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_20, None, full_4, False, "upscale_in_train", 0, False + softmax_20, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_20 # pd_op.matmul: (1x16x11x64xf32) <- (1x16x11x11xf32, 1x16x11x64xf32) matmul_164 = paddle._C_ops.matmul(dropout_122, transpose_82, False, False) + del dropout_122, transpose_82 # pd_op.transpose: (1x11x16x64xf32) <- (1x16x11x64xf32) transpose_83 = paddle._C_ops.transpose(matmul_164, [0, 2, 1, 3]) @@ -3804,19 +3728,20 @@ def forward( # pd_op.reshape: (1x11x1024xf32) <- (1x11x16x64xf32, 3xi64) reshape_83 = paddle._C_ops.reshape(transpose_83, full_int_array_2) + del transpose_83 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_165 = paddle._C_ops.matmul(reshape_83, parameter_59, False, False) - del parameter_59 + del parameter_59, reshape_83 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_186 = paddle._C_ops.add(matmul_165, parameter_58) - del parameter_58 + del matmul_165, parameter_58 # pd_op.dropout: (1x11x1024xf32, 1x11x1024xui8) <- (1x11x1024xf32, None, 1xf32) dropout_124, dropout_125 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_186, None, full_4, False, "upscale_in_train", 0, False + add_186, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -3824,6 +3749,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_187 = paddle._C_ops.add(layer_norm_120, dropout_124) + del dropout_124, layer_norm_120 # pd_op.layer_norm: (1x11x1024xf32, 1x11xf32, 1x11xf32) <- (1x11x1024xf32, 1024xf32, 1024xf32) layer_norm_123, layer_norm_124, layer_norm_125 = (lambda x, f: f(x))( @@ -3832,7 +3758,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_52, parameter_53 + del add_187, parameter_52, parameter_53 # pd_op.matmul: (1x11x4096xf32) <- (1x11x1024xf32, 1024x4096xf32) matmul_166 = paddle._C_ops.matmul(layer_norm_123, parameter_57, False, False) @@ -3840,7 +3766,7 @@ def forward( # pd_op.add: (1x11x4096xf32) <- (1x11x4096xf32, 4096xf32) add_188 = paddle._C_ops.add(matmul_166, parameter_56) - del parameter_56 + del matmul_166, parameter_56 # pd_op.relu: (1x11x4096xf32) <- (1x11x4096xf32) relu_20 = paddle._C_ops.relu(add_188) @@ -3848,16 +3774,16 @@ def forward( # pd_op.matmul: (1x11x1024xf32) <- (1x11x4096xf32, 4096x1024xf32) matmul_167 = paddle._C_ops.matmul(relu_20, parameter_55, False, False) - del parameter_55 + del parameter_55, relu_20 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_189 = paddle._C_ops.add(matmul_167, parameter_54) - del parameter_54 + del matmul_167, parameter_54 # pd_op.dropout: (1x11x1024xf32, 1x11x1024xui8) <- (1x11x1024xf32, None, 1xf32) dropout_126, dropout_127 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_189, None, full_4, False, "upscale_in_train", 0, False + add_189, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -3865,6 +3791,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_190 = paddle._C_ops.add(layer_norm_123, dropout_126) + del dropout_126, layer_norm_123 # pd_op.layer_norm: (1x11x1024xf32, 1x11xf32, 1x11xf32) <- (1x11x1024xf32, 1024xf32, 1024xf32) layer_norm_126, layer_norm_127, layer_norm_128 = (lambda x, f: f(x))( @@ -3873,7 +3800,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_50, parameter_51 + del add_190, parameter_50, parameter_51 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_168 = paddle._C_ops.matmul(layer_norm_126, parameter_49, False, False) @@ -3881,10 +3808,11 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_191 = paddle._C_ops.add(matmul_168, parameter_48) - del parameter_48 + del matmul_168, parameter_48 # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_84 = paddle._C_ops.reshape(add_191, full_int_array_1) + del add_191 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_84 = paddle._C_ops.transpose(reshape_84, [0, 2, 1, 3]) @@ -3896,7 +3824,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_192 = paddle._C_ops.add(matmul_169, parameter_46) - del parameter_46 + del matmul_169, parameter_46 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_170 = paddle._C_ops.matmul(layer_norm_126, parameter_45, False, False) @@ -3904,10 +3832,11 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_193 = paddle._C_ops.add(matmul_170, parameter_44) - del parameter_44 + del matmul_170, parameter_44 # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_85 = paddle._C_ops.reshape(add_192, full_int_array_1) + del add_192 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_85 = paddle._C_ops.transpose(reshape_85, [0, 2, 1, 3]) @@ -3915,6 +3844,7 @@ def forward( # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_86 = paddle._C_ops.reshape(add_193, full_int_array_1) + del add_193 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_86 = paddle._C_ops.transpose(reshape_86, [0, 2, 1, 3]) @@ -3926,9 +3856,11 @@ def forward( # pd_op.matmul: (1x16x11x11xf32) <- (1x16x11x64xf32, 1x16x11x64xf32) matmul_171 = paddle._C_ops.matmul(scale_22, transpose_85, False, True) + del scale_22, transpose_85 # pd_op.add: (1x16x11x11xf32) <- (1x16x11x11xf32, 1x1x1x11xf32) add_194 = paddle._C_ops.add(matmul_171, unsqueeze_0) + del matmul_171 # pd_op.softmax: (1x16x11x11xf32) <- (1x16x11x11xf32) softmax_21 = paddle._C_ops.softmax(add_194, -1) @@ -3937,13 +3869,15 @@ def forward( # pd_op.dropout: (1x16x11x11xf32, 1x16x11x11xui8) <- (1x16x11x11xf32, None, 1xf32) dropout_128, dropout_129 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_21, None, full_4, False, "upscale_in_train", 0, False + softmax_21, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_21 # pd_op.matmul: (1x16x11x64xf32) <- (1x16x11x11xf32, 1x16x11x64xf32) matmul_172 = paddle._C_ops.matmul(dropout_128, transpose_86, False, False) + del dropout_128, transpose_86 # pd_op.transpose: (1x11x16x64xf32) <- (1x16x11x64xf32) transpose_87 = paddle._C_ops.transpose(matmul_172, [0, 2, 1, 3]) @@ -3951,19 +3885,20 @@ def forward( # pd_op.reshape: (1x11x1024xf32) <- (1x11x16x64xf32, 3xi64) reshape_87 = paddle._C_ops.reshape(transpose_87, full_int_array_2) + del transpose_87 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_173 = paddle._C_ops.matmul(reshape_87, parameter_43, False, False) - del parameter_43 + del parameter_43, reshape_87 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_195 = paddle._C_ops.add(matmul_173, parameter_42) - del parameter_42 + del matmul_173, parameter_42 # pd_op.dropout: (1x11x1024xf32, 1x11x1024xui8) <- (1x11x1024xf32, None, 1xf32) dropout_130, dropout_131 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_195, None, full_4, False, "upscale_in_train", 0, False + add_195, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -3971,6 +3906,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_196 = paddle._C_ops.add(layer_norm_126, dropout_130) + del dropout_130, layer_norm_126 # pd_op.layer_norm: (1x11x1024xf32, 1x11xf32, 1x11xf32) <- (1x11x1024xf32, 1024xf32, 1024xf32) layer_norm_129, layer_norm_130, layer_norm_131 = (lambda x, f: f(x))( @@ -3979,7 +3915,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_36, parameter_37 + del add_196, parameter_36, parameter_37 # pd_op.matmul: (1x11x4096xf32) <- (1x11x1024xf32, 1024x4096xf32) matmul_174 = paddle._C_ops.matmul(layer_norm_129, parameter_41, False, False) @@ -3987,7 +3923,7 @@ def forward( # pd_op.add: (1x11x4096xf32) <- (1x11x4096xf32, 4096xf32) add_197 = paddle._C_ops.add(matmul_174, parameter_40) - del parameter_40 + del matmul_174, parameter_40 # pd_op.relu: (1x11x4096xf32) <- (1x11x4096xf32) relu_21 = paddle._C_ops.relu(add_197) @@ -3995,16 +3931,16 @@ def forward( # pd_op.matmul: (1x11x1024xf32) <- (1x11x4096xf32, 4096x1024xf32) matmul_175 = paddle._C_ops.matmul(relu_21, parameter_39, False, False) - del parameter_39 + del parameter_39, relu_21 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_198 = paddle._C_ops.add(matmul_175, parameter_38) - del parameter_38 + del matmul_175, parameter_38 # pd_op.dropout: (1x11x1024xf32, 1x11x1024xui8) <- (1x11x1024xf32, None, 1xf32) dropout_132, dropout_133 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_198, None, full_4, False, "upscale_in_train", 0, False + add_198, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -4012,6 +3948,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_199 = paddle._C_ops.add(layer_norm_129, dropout_132) + del dropout_132, layer_norm_129 # pd_op.layer_norm: (1x11x1024xf32, 1x11xf32, 1x11xf32) <- (1x11x1024xf32, 1024xf32, 1024xf32) layer_norm_132, layer_norm_133, layer_norm_134 = (lambda x, f: f(x))( @@ -4020,7 +3957,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_34, parameter_35 + del add_199, parameter_34, parameter_35 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_176 = paddle._C_ops.matmul(layer_norm_132, parameter_33, False, False) @@ -4028,10 +3965,11 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_200 = paddle._C_ops.add(matmul_176, parameter_32) - del parameter_32 + del matmul_176, parameter_32 # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_88 = paddle._C_ops.reshape(add_200, full_int_array_1) + del add_200 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_88 = paddle._C_ops.transpose(reshape_88, [0, 2, 1, 3]) @@ -4043,7 +3981,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_201 = paddle._C_ops.add(matmul_177, parameter_30) - del parameter_30 + del matmul_177, parameter_30 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_178 = paddle._C_ops.matmul(layer_norm_132, parameter_29, False, False) @@ -4051,10 +3989,11 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_202 = paddle._C_ops.add(matmul_178, parameter_28) - del parameter_28 + del matmul_178, parameter_28 # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_89 = paddle._C_ops.reshape(add_201, full_int_array_1) + del add_201 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_89 = paddle._C_ops.transpose(reshape_89, [0, 2, 1, 3]) @@ -4062,6 +4001,7 @@ def forward( # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_90 = paddle._C_ops.reshape(add_202, full_int_array_1) + del add_202 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_90 = paddle._C_ops.transpose(reshape_90, [0, 2, 1, 3]) @@ -4073,9 +4013,11 @@ def forward( # pd_op.matmul: (1x16x11x11xf32) <- (1x16x11x64xf32, 1x16x11x64xf32) matmul_179 = paddle._C_ops.matmul(scale_23, transpose_89, False, True) + del scale_23, transpose_89 # pd_op.add: (1x16x11x11xf32) <- (1x16x11x11xf32, 1x1x1x11xf32) add_203 = paddle._C_ops.add(matmul_179, unsqueeze_0) + del matmul_179 # pd_op.softmax: (1x16x11x11xf32) <- (1x16x11x11xf32) softmax_22 = paddle._C_ops.softmax(add_203, -1) @@ -4084,13 +4026,15 @@ def forward( # pd_op.dropout: (1x16x11x11xf32, 1x16x11x11xui8) <- (1x16x11x11xf32, None, 1xf32) dropout_134, dropout_135 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_22, None, full_4, False, "upscale_in_train", 0, False + softmax_22, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_22 # pd_op.matmul: (1x16x11x64xf32) <- (1x16x11x11xf32, 1x16x11x64xf32) matmul_180 = paddle._C_ops.matmul(dropout_134, transpose_90, False, False) + del dropout_134, transpose_90 # pd_op.transpose: (1x11x16x64xf32) <- (1x16x11x64xf32) transpose_91 = paddle._C_ops.transpose(matmul_180, [0, 2, 1, 3]) @@ -4098,19 +4042,20 @@ def forward( # pd_op.reshape: (1x11x1024xf32) <- (1x11x16x64xf32, 3xi64) reshape_91 = paddle._C_ops.reshape(transpose_91, full_int_array_2) + del transpose_91 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_181 = paddle._C_ops.matmul(reshape_91, parameter_27, False, False) - del parameter_27 + del parameter_27, reshape_91 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_204 = paddle._C_ops.add(matmul_181, parameter_26) - del parameter_26 + del matmul_181, parameter_26 # pd_op.dropout: (1x11x1024xf32, 1x11x1024xui8) <- (1x11x1024xf32, None, 1xf32) dropout_136, dropout_137 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_204, None, full_4, False, "upscale_in_train", 0, False + add_204, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -4118,6 +4063,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_205 = paddle._C_ops.add(layer_norm_132, dropout_136) + del dropout_136, layer_norm_132 # pd_op.layer_norm: (1x11x1024xf32, 1x11xf32, 1x11xf32) <- (1x11x1024xf32, 1024xf32, 1024xf32) layer_norm_135, layer_norm_136, layer_norm_137 = (lambda x, f: f(x))( @@ -4126,7 +4072,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_20, parameter_21 + del add_205, parameter_20, parameter_21 # pd_op.matmul: (1x11x4096xf32) <- (1x11x1024xf32, 1024x4096xf32) matmul_182 = paddle._C_ops.matmul(layer_norm_135, parameter_25, False, False) @@ -4134,7 +4080,7 @@ def forward( # pd_op.add: (1x11x4096xf32) <- (1x11x4096xf32, 4096xf32) add_206 = paddle._C_ops.add(matmul_182, parameter_24) - del parameter_24 + del matmul_182, parameter_24 # pd_op.relu: (1x11x4096xf32) <- (1x11x4096xf32) relu_22 = paddle._C_ops.relu(add_206) @@ -4142,16 +4088,16 @@ def forward( # pd_op.matmul: (1x11x1024xf32) <- (1x11x4096xf32, 4096x1024xf32) matmul_183 = paddle._C_ops.matmul(relu_22, parameter_23, False, False) - del parameter_23 + del parameter_23, relu_22 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_207 = paddle._C_ops.add(matmul_183, parameter_22) - del parameter_22 + del matmul_183, parameter_22 # pd_op.dropout: (1x11x1024xf32, 1x11x1024xui8) <- (1x11x1024xf32, None, 1xf32) dropout_138, dropout_139 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_207, None, full_4, False, "upscale_in_train", 0, False + add_207, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -4159,6 +4105,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_208 = paddle._C_ops.add(layer_norm_135, dropout_138) + del dropout_138, layer_norm_135 # pd_op.layer_norm: (1x11x1024xf32, 1x11xf32, 1x11xf32) <- (1x11x1024xf32, 1024xf32, 1024xf32) layer_norm_138, layer_norm_139, layer_norm_140 = (lambda x, f: f(x))( @@ -4167,7 +4114,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_18, parameter_19 + del add_208, parameter_18, parameter_19 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_184 = paddle._C_ops.matmul(layer_norm_138, parameter_17, False, False) @@ -4175,10 +4122,11 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_209 = paddle._C_ops.add(matmul_184, parameter_16) - del parameter_16 + del matmul_184, parameter_16 # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_92 = paddle._C_ops.reshape(add_209, full_int_array_1) + del add_209 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_92 = paddle._C_ops.transpose(reshape_92, [0, 2, 1, 3]) @@ -4190,7 +4138,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_210 = paddle._C_ops.add(matmul_185, parameter_14) - del parameter_14 + del matmul_185, parameter_14 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_186 = paddle._C_ops.matmul(layer_norm_138, parameter_13, False, False) @@ -4198,10 +4146,11 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_211 = paddle._C_ops.add(matmul_186, parameter_12) - del parameter_12 + del matmul_186, parameter_12 # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_93 = paddle._C_ops.reshape(add_210, full_int_array_1) + del add_210 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_93 = paddle._C_ops.transpose(reshape_93, [0, 2, 1, 3]) @@ -4209,7 +4158,7 @@ def forward( # pd_op.reshape: (1x11x16x64xf32) <- (1x11x1024xf32, 4xi64) reshape_94 = paddle._C_ops.reshape(add_211, full_int_array_1) - del full_int_array_1 + del add_211, full_int_array_1 # pd_op.transpose: (1x16x11x64xf32) <- (1x11x16x64xf32) transpose_94 = paddle._C_ops.transpose(reshape_94, [0, 2, 1, 3]) @@ -4217,13 +4166,15 @@ def forward( # pd_op.scale: (1x16x11x64xf32) <- (1x16x11x64xf32, 1xf32) scale_24 = paddle._C_ops.scale(transpose_92, full_5, float("0"), True) - del transpose_92 + del full_5, transpose_92 # pd_op.matmul: (1x16x11x11xf32) <- (1x16x11x64xf32, 1x16x11x64xf32) matmul_187 = paddle._C_ops.matmul(scale_24, transpose_93, False, True) + del scale_24, transpose_93 # pd_op.add: (1x16x11x11xf32) <- (1x16x11x11xf32, 1x1x1x11xf32) add_212 = paddle._C_ops.add(matmul_187, unsqueeze_0) + del matmul_187, unsqueeze_0 # pd_op.softmax: (1x16x11x11xf32) <- (1x16x11x11xf32) softmax_23 = paddle._C_ops.softmax(add_212, -1) @@ -4232,13 +4183,15 @@ def forward( # pd_op.dropout: (1x16x11x11xf32, 1x16x11x11xui8) <- (1x16x11x11xf32, None, 1xf32) dropout_140, dropout_141 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_23, None, full_4, False, "upscale_in_train", 0, False + softmax_23, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_23 # pd_op.matmul: (1x16x11x64xf32) <- (1x16x11x11xf32, 1x16x11x64xf32) matmul_188 = paddle._C_ops.matmul(dropout_140, transpose_94, False, False) + del dropout_140, transpose_94 # pd_op.transpose: (1x11x16x64xf32) <- (1x16x11x64xf32) transpose_95 = paddle._C_ops.transpose(matmul_188, [0, 2, 1, 3]) @@ -4246,20 +4199,20 @@ def forward( # pd_op.reshape: (1x11x1024xf32) <- (1x11x16x64xf32, 3xi64) reshape_95 = paddle._C_ops.reshape(transpose_95, full_int_array_2) - del full_int_array_2 + del full_int_array_2, transpose_95 # pd_op.matmul: (1x11x1024xf32) <- (1x11x1024xf32, 1024x1024xf32) matmul_189 = paddle._C_ops.matmul(reshape_95, parameter_11, False, False) - del parameter_11 + del parameter_11, reshape_95 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_213 = paddle._C_ops.add(matmul_189, parameter_10) - del parameter_10 + del matmul_189, parameter_10 # pd_op.dropout: (1x11x1024xf32, 1x11x1024xui8) <- (1x11x1024xf32, None, 1xf32) dropout_142, dropout_143 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_213, None, full_4, False, "upscale_in_train", 0, False + add_213, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -4267,6 +4220,7 @@ def forward( # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_214 = paddle._C_ops.add(layer_norm_138, dropout_142) + del dropout_142, layer_norm_138 # pd_op.layer_norm: (1x11x1024xf32, 1x11xf32, 1x11xf32) <- (1x11x1024xf32, 1024xf32, 1024xf32) layer_norm_141, layer_norm_142, layer_norm_143 = (lambda x, f: f(x))( @@ -4275,7 +4229,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_4, parameter_5 + del add_214, parameter_4, parameter_5 # pd_op.matmul: (1x11x4096xf32) <- (1x11x1024xf32, 1024x4096xf32) matmul_190 = paddle._C_ops.matmul(layer_norm_141, parameter_9, False, False) @@ -4283,7 +4237,7 @@ def forward( # pd_op.add: (1x11x4096xf32) <- (1x11x4096xf32, 4096xf32) add_215 = paddle._C_ops.add(matmul_190, parameter_8) - del parameter_8 + del matmul_190, parameter_8 # pd_op.relu: (1x11x4096xf32) <- (1x11x4096xf32) relu_23 = paddle._C_ops.relu(add_215) @@ -4291,23 +4245,24 @@ def forward( # pd_op.matmul: (1x11x1024xf32) <- (1x11x4096xf32, 4096x1024xf32) matmul_191 = paddle._C_ops.matmul(relu_23, parameter_7, False, False) - del parameter_7 + del parameter_7, relu_23 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1024xf32) add_216 = paddle._C_ops.add(matmul_191, parameter_6) - del parameter_6 + del matmul_191, parameter_6 # pd_op.dropout: (1x11x1024xf32, 1x11x1024xui8) <- (1x11x1024xf32, None, 1xf32) dropout_144, dropout_145 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_216, None, full_4, False, "upscale_in_train", 0, False + add_216, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) - del add_216 + del add_216, full_4 # pd_op.add: (1x11x1024xf32) <- (1x11x1024xf32, 1x11x1024xf32) add_217 = paddle._C_ops.add(layer_norm_141, dropout_144) + del dropout_144, layer_norm_141 # pd_op.layer_norm: (1x11x1024xf32, 1x11xf32, 1x11xf32) <- (1x11x1024xf32, 1024xf32, 1024xf32) layer_norm_144, layer_norm_145, layer_norm_146 = (lambda x, f: f(x))( @@ -4316,7 +4271,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_2, parameter_3 + del add_217, parameter_2, parameter_3 # pd_op.full_int_array: (1xi64) <- () full_int_array_3 = [0] @@ -4328,875 +4283,18 @@ def forward( slice_0 = paddle._C_ops.slice( layer_norm_144, [1], full_int_array_3, full_int_array_4, [1], [1] ) + del full_int_array_3, full_int_array_4, layer_norm_144 # pd_op.matmul: (1x1024xf32) <- (1x1024xf32, 1024x1024xf32) matmul_192 = paddle._C_ops.matmul(slice_0, parameter_1, False, False) - del parameter_1 + del parameter_1, slice_0 # pd_op.add: (1x1024xf32) <- (1x1024xf32, 1024xf32) add_218 = paddle._C_ops.add(matmul_192, parameter_0) - del parameter_0 + del matmul_192, parameter_0 # pd_op.tanh: (1x1024xf32) <- (1x1024xf32) tanh_0 = paddle._C_ops.tanh(add_218) - del ( - add_0, - add_1, - add_10, - add_100, - add_101, - add_102, - add_103, - add_106, - add_109, - add_11, - add_110, - add_111, - add_112, - add_115, - add_118, - add_119, - add_12, - add_120, - add_121, - add_124, - add_127, - add_128, - add_129, - add_13, - add_130, - add_133, - add_136, - add_137, - add_138, - add_139, - add_142, - add_145, - add_146, - add_147, - add_148, - add_151, - add_154, - add_155, - add_156, - add_157, - add_16, - add_160, - add_163, - add_164, - add_165, - add_166, - add_169, - add_172, - add_173, - add_174, - add_175, - add_178, - add_181, - add_182, - add_183, - add_184, - add_187, - add_19, - add_190, - add_191, - add_192, - add_193, - add_196, - add_199, - add_2, - add_20, - add_200, - add_201, - add_202, - add_205, - add_208, - add_209, - add_21, - add_210, - add_211, - add_214, - add_217, - add_218, - add_22, - add_25, - add_28, - add_29, - add_3, - add_30, - add_31, - add_34, - add_37, - add_38, - add_39, - add_4, - add_40, - add_43, - add_46, - add_47, - add_48, - add_49, - add_52, - add_55, - add_56, - add_57, - add_58, - add_61, - add_64, - add_65, - add_66, - add_67, - add_7, - add_70, - add_73, - add_74, - add_75, - add_76, - add_79, - add_82, - add_83, - add_84, - add_85, - add_88, - add_91, - add_92, - add_93, - add_94, - add_97, - assign_0, - assign_1, - assign_10, - assign_11, - assign_12, - assign_13, - assign_14, - assign_15, - assign_16, - assign_17, - assign_18, - assign_19, - assign_2, - assign_20, - assign_21, - assign_22, - assign_23, - assign_24, - assign_25, - assign_26, - assign_27, - assign_28, - assign_29, - assign_3, - assign_30, - assign_31, - assign_32, - assign_33, - assign_34, - assign_35, - assign_36, - assign_37, - assign_38, - assign_39, - assign_4, - assign_40, - assign_41, - assign_42, - assign_43, - assign_44, - assign_45, - assign_46, - assign_47, - assign_48, - assign_49, - assign_5, - assign_50, - assign_51, - assign_52, - assign_53, - assign_54, - assign_55, - assign_56, - assign_57, - assign_58, - assign_59, - assign_6, - assign_60, - assign_61, - assign_62, - assign_63, - assign_64, - assign_65, - assign_66, - assign_67, - assign_68, - assign_69, - assign_7, - assign_70, - assign_71, - assign_72, - assign_73, - assign_74, - assign_75, - assign_76, - assign_77, - assign_78, - assign_79, - assign_8, - assign_80, - assign_81, - assign_82, - assign_83, - assign_84, - assign_85, - assign_86, - assign_87, - assign_88, - assign_89, - assign_9, - assign_90, - assign_91, - assign_92, - assign_93, - assign_94, - dropout_0, - dropout_1, - dropout_10, - dropout_100, - dropout_101, - dropout_102, - dropout_103, - dropout_104, - dropout_105, - dropout_106, - dropout_107, - dropout_108, - dropout_109, - dropout_11, - dropout_110, - dropout_111, - dropout_112, - dropout_113, - dropout_114, - dropout_115, - dropout_116, - dropout_117, - dropout_118, - dropout_119, - dropout_12, - dropout_120, - dropout_121, - dropout_122, - dropout_123, - dropout_124, - dropout_125, - dropout_126, - dropout_127, - dropout_128, - dropout_129, - dropout_13, - dropout_130, - dropout_131, - dropout_132, - dropout_133, - dropout_134, - dropout_135, - dropout_136, - dropout_137, - dropout_138, - dropout_139, - dropout_14, - dropout_140, - dropout_141, - dropout_142, - dropout_143, - dropout_144, - dropout_145, - dropout_15, - dropout_16, - dropout_17, - dropout_18, - dropout_19, - dropout_2, - dropout_20, - dropout_21, - dropout_22, - dropout_23, - dropout_24, - dropout_25, - dropout_26, - dropout_27, - dropout_28, - dropout_29, - dropout_3, - dropout_30, - dropout_31, - dropout_32, - dropout_33, - dropout_34, - dropout_35, - dropout_36, - dropout_37, - dropout_38, - dropout_39, - dropout_4, - dropout_40, - dropout_41, - dropout_42, - dropout_43, - dropout_44, - dropout_45, - dropout_46, - dropout_47, - dropout_48, - dropout_49, - dropout_5, - dropout_50, - dropout_51, - dropout_52, - dropout_53, - dropout_54, - dropout_55, - dropout_56, - dropout_57, - dropout_58, - dropout_59, - dropout_6, - dropout_60, - dropout_61, - dropout_62, - dropout_63, - dropout_64, - dropout_65, - dropout_66, - dropout_67, - dropout_68, - dropout_69, - dropout_7, - dropout_70, - dropout_71, - dropout_72, - dropout_73, - dropout_74, - dropout_75, - dropout_76, - dropout_77, - dropout_78, - dropout_79, - dropout_8, - dropout_80, - dropout_81, - dropout_82, - dropout_83, - dropout_84, - dropout_85, - dropout_86, - dropout_87, - dropout_88, - dropout_89, - dropout_9, - dropout_90, - dropout_91, - dropout_92, - dropout_93, - dropout_94, - dropout_95, - dropout_96, - dropout_97, - dropout_98, - dropout_99, - embedding_0, - embedding_1, - embedding_2, - full_4, - full_5, - full_int_array_3, - full_int_array_4, - layer_norm_1, - layer_norm_10, - layer_norm_100, - layer_norm_101, - layer_norm_102, - layer_norm_103, - layer_norm_104, - layer_norm_105, - layer_norm_106, - layer_norm_107, - layer_norm_108, - layer_norm_109, - layer_norm_11, - layer_norm_110, - layer_norm_111, - layer_norm_112, - layer_norm_113, - layer_norm_114, - layer_norm_115, - layer_norm_116, - layer_norm_117, - layer_norm_118, - layer_norm_119, - layer_norm_12, - layer_norm_120, - layer_norm_121, - layer_norm_122, - layer_norm_123, - layer_norm_124, - layer_norm_125, - layer_norm_126, - layer_norm_127, - layer_norm_128, - layer_norm_129, - layer_norm_13, - layer_norm_130, - layer_norm_131, - layer_norm_132, - layer_norm_133, - layer_norm_134, - layer_norm_135, - layer_norm_136, - layer_norm_137, - layer_norm_138, - layer_norm_139, - layer_norm_14, - layer_norm_140, - layer_norm_141, - layer_norm_142, - layer_norm_143, - layer_norm_144, - layer_norm_145, - layer_norm_146, - layer_norm_15, - layer_norm_16, - layer_norm_17, - layer_norm_18, - layer_norm_19, - layer_norm_2, - layer_norm_20, - layer_norm_21, - layer_norm_22, - layer_norm_23, - layer_norm_24, - layer_norm_25, - layer_norm_26, - layer_norm_27, - layer_norm_28, - layer_norm_29, - layer_norm_3, - layer_norm_30, - layer_norm_31, - layer_norm_32, - layer_norm_33, - layer_norm_34, - layer_norm_35, - layer_norm_36, - layer_norm_37, - layer_norm_38, - layer_norm_39, - layer_norm_4, - layer_norm_40, - layer_norm_41, - layer_norm_42, - layer_norm_43, - layer_norm_44, - layer_norm_45, - layer_norm_46, - layer_norm_47, - layer_norm_48, - layer_norm_49, - layer_norm_5, - layer_norm_50, - layer_norm_51, - layer_norm_52, - layer_norm_53, - layer_norm_54, - layer_norm_55, - layer_norm_56, - layer_norm_57, - layer_norm_58, - layer_norm_59, - layer_norm_6, - layer_norm_60, - layer_norm_61, - layer_norm_62, - layer_norm_63, - layer_norm_64, - layer_norm_65, - layer_norm_66, - layer_norm_67, - layer_norm_68, - layer_norm_69, - layer_norm_7, - layer_norm_70, - layer_norm_71, - layer_norm_72, - layer_norm_73, - layer_norm_74, - layer_norm_75, - layer_norm_76, - layer_norm_77, - layer_norm_78, - layer_norm_79, - layer_norm_8, - layer_norm_80, - layer_norm_81, - layer_norm_82, - layer_norm_83, - layer_norm_84, - layer_norm_85, - layer_norm_86, - layer_norm_87, - layer_norm_88, - layer_norm_89, - layer_norm_9, - layer_norm_90, - layer_norm_91, - layer_norm_92, - layer_norm_93, - layer_norm_94, - layer_norm_95, - layer_norm_96, - layer_norm_97, - layer_norm_98, - layer_norm_99, - matmul_0, - matmul_1, - matmul_10, - matmul_101, - matmul_102, - matmul_103, - matmul_104, - matmul_105, - matmul_106, - matmul_107, - matmul_109, - matmul_11, - matmul_110, - matmul_111, - matmul_112, - matmul_113, - matmul_114, - matmul_115, - matmul_117, - matmul_118, - matmul_119, - matmul_120, - matmul_121, - matmul_122, - matmul_123, - matmul_125, - matmul_126, - matmul_127, - matmul_128, - matmul_129, - matmul_13, - matmul_130, - matmul_131, - matmul_133, - matmul_134, - matmul_135, - matmul_136, - matmul_137, - matmul_138, - matmul_139, - matmul_14, - matmul_141, - matmul_142, - matmul_143, - matmul_144, - matmul_145, - matmul_146, - matmul_147, - matmul_149, - matmul_15, - matmul_150, - matmul_151, - matmul_152, - matmul_153, - matmul_154, - matmul_155, - matmul_157, - matmul_158, - matmul_159, - matmul_16, - matmul_160, - matmul_161, - matmul_162, - matmul_163, - matmul_165, - matmul_166, - matmul_167, - matmul_168, - matmul_169, - matmul_17, - matmul_170, - matmul_171, - matmul_173, - matmul_174, - matmul_175, - matmul_176, - matmul_177, - matmul_178, - matmul_179, - matmul_18, - matmul_181, - matmul_182, - matmul_183, - matmul_184, - matmul_185, - matmul_186, - matmul_187, - matmul_189, - matmul_19, - matmul_190, - matmul_191, - matmul_192, - matmul_2, - matmul_21, - matmul_22, - matmul_23, - matmul_24, - matmul_25, - matmul_26, - matmul_27, - matmul_29, - matmul_3, - matmul_30, - matmul_31, - matmul_32, - matmul_33, - matmul_34, - matmul_35, - matmul_37, - matmul_38, - matmul_39, - matmul_40, - matmul_41, - matmul_42, - matmul_43, - matmul_45, - matmul_46, - matmul_47, - matmul_48, - matmul_49, - matmul_5, - matmul_50, - matmul_51, - matmul_53, - matmul_54, - matmul_55, - matmul_56, - matmul_57, - matmul_58, - matmul_59, - matmul_6, - matmul_61, - matmul_62, - matmul_63, - matmul_64, - matmul_65, - matmul_66, - matmul_67, - matmul_69, - matmul_7, - matmul_70, - matmul_71, - matmul_72, - matmul_73, - matmul_74, - matmul_75, - matmul_77, - matmul_78, - matmul_79, - matmul_8, - matmul_80, - matmul_81, - matmul_82, - matmul_83, - matmul_85, - matmul_86, - matmul_87, - matmul_88, - matmul_89, - matmul_9, - matmul_90, - matmul_91, - matmul_93, - matmul_94, - matmul_95, - matmul_96, - matmul_97, - matmul_98, - matmul_99, - relu_0, - relu_1, - relu_10, - relu_11, - relu_12, - relu_13, - relu_14, - relu_15, - relu_16, - relu_17, - relu_18, - relu_19, - relu_2, - relu_20, - relu_21, - relu_22, - relu_23, - relu_3, - relu_4, - relu_5, - relu_6, - relu_7, - relu_8, - relu_9, - reshape_11, - reshape_15, - reshape_19, - reshape_23, - reshape_27, - reshape_3, - reshape_31, - reshape_35, - reshape_39, - reshape_43, - reshape_47, - reshape_51, - reshape_55, - reshape_59, - reshape_63, - reshape_67, - reshape_7, - reshape_71, - reshape_75, - reshape_79, - reshape_83, - reshape_87, - reshape_91, - reshape_95, - scale_1, - scale_10, - scale_11, - scale_12, - scale_13, - scale_14, - scale_15, - scale_16, - scale_17, - scale_18, - scale_19, - scale_2, - scale_20, - scale_21, - scale_22, - scale_23, - scale_24, - scale_3, - scale_4, - scale_5, - scale_6, - scale_7, - scale_8, - scale_9, - slice_0, - softmax_0, - softmax_1, - softmax_10, - softmax_11, - softmax_12, - softmax_13, - softmax_14, - softmax_15, - softmax_16, - softmax_17, - softmax_18, - softmax_19, - softmax_2, - softmax_20, - softmax_21, - softmax_22, - softmax_23, - softmax_3, - softmax_4, - softmax_5, - softmax_6, - softmax_7, - softmax_8, - softmax_9, - subtract_0, - transpose_1, - transpose_10, - transpose_11, - transpose_13, - transpose_14, - transpose_15, - transpose_17, - transpose_18, - transpose_19, - transpose_2, - transpose_21, - transpose_22, - transpose_23, - transpose_25, - transpose_26, - transpose_27, - transpose_29, - transpose_3, - transpose_30, - transpose_31, - transpose_33, - transpose_34, - transpose_35, - transpose_37, - transpose_38, - transpose_39, - transpose_41, - transpose_42, - transpose_43, - transpose_45, - transpose_46, - transpose_47, - transpose_49, - transpose_5, - transpose_50, - transpose_51, - transpose_53, - transpose_54, - transpose_55, - transpose_57, - transpose_58, - transpose_59, - transpose_6, - transpose_61, - transpose_62, - transpose_63, - transpose_65, - transpose_66, - transpose_67, - transpose_69, - transpose_7, - transpose_70, - transpose_71, - transpose_73, - transpose_74, - transpose_75, - transpose_77, - transpose_78, - transpose_79, - transpose_81, - transpose_82, - transpose_83, - transpose_85, - transpose_86, - transpose_87, - transpose_89, - transpose_9, - transpose_90, - transpose_91, - transpose_93, - transpose_94, - transpose_95, - unsqueeze_0, - ) + del add_218 return tanh_0 diff --git a/paddle_samples/PaddleNLP/skep_ernie_1.0_large_ch/weight_meta.py b/paddle_samples/PaddleNLP/skep_ernie_1.0_large_ch/weight_meta.py index b1f8c4a27..7fdc032a6 100644 --- a/paddle_samples/PaddleNLP/skep_ernie_1.0_large_ch/weight_meta.py +++ b/paddle_samples/PaddleNLP/skep_ernie_1.0_large_ch/weight_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_parameter_0: name = "parameter_0" + original_name = "linear_4436.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0671379") @@ -11,6 +12,7 @@ class Program_weight_tensor_parameter_0: class Program_weight_tensor_parameter_1: name = "parameter_1" + original_name = "linear_4436.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.194627") @@ -22,6 +24,7 @@ class Program_weight_tensor_parameter_1: class Program_weight_tensor_parameter_2: name = "parameter_2" + original_name = "layer_norm_1528.b_0" shape = [1024] dtype = "float32" min_val = float("-0.282979") @@ -33,6 +36,7 @@ class Program_weight_tensor_parameter_2: class Program_weight_tensor_parameter_3: name = "parameter_3" + original_name = "layer_norm_1528.w_0" shape = [1024] dtype = "float32" min_val = float("0.343393") @@ -44,6 +48,7 @@ class Program_weight_tensor_parameter_3: class Program_weight_tensor_parameter_4: name = "parameter_4" + original_name = "layer_norm_1527.b_0" shape = [1024] dtype = "float32" min_val = float("-0.344402") @@ -55,6 +60,7 @@ class Program_weight_tensor_parameter_4: class Program_weight_tensor_parameter_5: name = "parameter_5" + original_name = "layer_norm_1527.w_0" shape = [1024] dtype = "float32" min_val = float("0.898427") @@ -66,6 +72,7 @@ class Program_weight_tensor_parameter_5: class Program_weight_tensor_parameter_6: name = "parameter_6" + original_name = "linear_4435.b_0" shape = [1024] dtype = "float32" min_val = float("-0.107107") @@ -77,6 +84,7 @@ class Program_weight_tensor_parameter_6: class Program_weight_tensor_parameter_7: name = "parameter_7" + original_name = "linear_4435.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-0.294746") @@ -88,6 +96,7 @@ class Program_weight_tensor_parameter_7: class Program_weight_tensor_parameter_8: name = "parameter_8" + original_name = "linear_4434.b_0" shape = [4096] dtype = "float32" min_val = float("-0.273805") @@ -99,6 +108,7 @@ class Program_weight_tensor_parameter_8: class Program_weight_tensor_parameter_9: name = "parameter_9" + original_name = "linear_4434.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.289417") @@ -110,6 +120,7 @@ class Program_weight_tensor_parameter_9: class Program_weight_tensor_parameter_10: name = "parameter_10" + original_name = "linear_4433.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0979534") @@ -121,6 +132,7 @@ class Program_weight_tensor_parameter_10: class Program_weight_tensor_parameter_11: name = "parameter_11" + original_name = "linear_4433.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.166448") @@ -132,6 +144,7 @@ class Program_weight_tensor_parameter_11: class Program_weight_tensor_parameter_12: name = "parameter_12" + original_name = "linear_4432.b_0" shape = [1024] dtype = "float32" min_val = float("-0.121132") @@ -143,6 +156,7 @@ class Program_weight_tensor_parameter_12: class Program_weight_tensor_parameter_13: name = "parameter_13" + original_name = "linear_4432.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.171801") @@ -154,6 +168,7 @@ class Program_weight_tensor_parameter_13: class Program_weight_tensor_parameter_14: name = "parameter_14" + original_name = "linear_4431.b_0" shape = [1024] dtype = "float32" min_val = float("-15.6866") @@ -165,6 +180,7 @@ class Program_weight_tensor_parameter_14: class Program_weight_tensor_parameter_15: name = "parameter_15" + original_name = "linear_4431.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.241874") @@ -176,6 +192,7 @@ class Program_weight_tensor_parameter_15: class Program_weight_tensor_parameter_16: name = "parameter_16" + original_name = "linear_4430.b_0" shape = [1024] dtype = "float32" min_val = float("-0.321052") @@ -187,6 +204,7 @@ class Program_weight_tensor_parameter_16: class Program_weight_tensor_parameter_17: name = "parameter_17" + original_name = "linear_4430.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.416354") @@ -198,6 +216,7 @@ class Program_weight_tensor_parameter_17: class Program_weight_tensor_parameter_18: name = "parameter_18" + original_name = "layer_norm_1526.b_0" shape = [1024] dtype = "float32" min_val = float("-0.792417") @@ -209,6 +228,7 @@ class Program_weight_tensor_parameter_18: class Program_weight_tensor_parameter_19: name = "parameter_19" + original_name = "layer_norm_1526.w_0" shape = [1024] dtype = "float32" min_val = float("0.476952") @@ -220,6 +240,7 @@ class Program_weight_tensor_parameter_19: class Program_weight_tensor_parameter_20: name = "parameter_20" + original_name = "layer_norm_1525.b_0" shape = [1024] dtype = "float32" min_val = float("-0.674274") @@ -231,6 +252,7 @@ class Program_weight_tensor_parameter_20: class Program_weight_tensor_parameter_21: name = "parameter_21" + original_name = "layer_norm_1525.w_0" shape = [1024] dtype = "float32" min_val = float("0.91718") @@ -242,6 +264,7 @@ class Program_weight_tensor_parameter_21: class Program_weight_tensor_parameter_22: name = "parameter_22" + original_name = "linear_4429.b_0" shape = [1024] dtype = "float32" min_val = float("-0.234498") @@ -253,6 +276,7 @@ class Program_weight_tensor_parameter_22: class Program_weight_tensor_parameter_23: name = "parameter_23" + original_name = "linear_4429.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-0.322015") @@ -264,6 +288,7 @@ class Program_weight_tensor_parameter_23: class Program_weight_tensor_parameter_24: name = "parameter_24" + original_name = "linear_4428.b_0" shape = [4096] dtype = "float32" min_val = float("-0.172135") @@ -275,6 +300,7 @@ class Program_weight_tensor_parameter_24: class Program_weight_tensor_parameter_25: name = "parameter_25" + original_name = "linear_4428.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.221527") @@ -286,6 +312,7 @@ class Program_weight_tensor_parameter_25: class Program_weight_tensor_parameter_26: name = "parameter_26" + original_name = "linear_4427.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0955337") @@ -297,6 +324,7 @@ class Program_weight_tensor_parameter_26: class Program_weight_tensor_parameter_27: name = "parameter_27" + original_name = "linear_4427.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.173955") @@ -308,6 +336,7 @@ class Program_weight_tensor_parameter_27: class Program_weight_tensor_parameter_28: name = "parameter_28" + original_name = "linear_4426.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0507613") @@ -319,6 +348,7 @@ class Program_weight_tensor_parameter_28: class Program_weight_tensor_parameter_29: name = "parameter_29" + original_name = "linear_4426.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.173345") @@ -330,6 +360,7 @@ class Program_weight_tensor_parameter_29: class Program_weight_tensor_parameter_30: name = "parameter_30" + original_name = "linear_4425.b_0" shape = [1024] dtype = "float32" min_val = float("-19.3377") @@ -341,6 +372,7 @@ class Program_weight_tensor_parameter_30: class Program_weight_tensor_parameter_31: name = "parameter_31" + original_name = "linear_4425.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.243883") @@ -352,6 +384,7 @@ class Program_weight_tensor_parameter_31: class Program_weight_tensor_parameter_32: name = "parameter_32" + original_name = "linear_4424.b_0" shape = [1024] dtype = "float32" min_val = float("-0.422501") @@ -363,6 +396,7 @@ class Program_weight_tensor_parameter_32: class Program_weight_tensor_parameter_33: name = "parameter_33" + original_name = "linear_4424.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.375724") @@ -374,6 +408,7 @@ class Program_weight_tensor_parameter_33: class Program_weight_tensor_parameter_34: name = "parameter_34" + original_name = "layer_norm_1524.b_0" shape = [1024] dtype = "float32" min_val = float("-0.437628") @@ -385,6 +420,7 @@ class Program_weight_tensor_parameter_34: class Program_weight_tensor_parameter_35: name = "parameter_35" + original_name = "layer_norm_1524.w_0" shape = [1024] dtype = "float32" min_val = float("0.554111") @@ -396,6 +432,7 @@ class Program_weight_tensor_parameter_35: class Program_weight_tensor_parameter_36: name = "parameter_36" + original_name = "layer_norm_1523.b_0" shape = [1024] dtype = "float32" min_val = float("-1.13343") @@ -407,6 +444,7 @@ class Program_weight_tensor_parameter_36: class Program_weight_tensor_parameter_37: name = "parameter_37" + original_name = "layer_norm_1523.w_0" shape = [1024] dtype = "float32" min_val = float("0.908367") @@ -418,6 +456,7 @@ class Program_weight_tensor_parameter_37: class Program_weight_tensor_parameter_38: name = "parameter_38" + original_name = "linear_4423.b_0" shape = [1024] dtype = "float32" min_val = float("-0.257327") @@ -429,6 +468,7 @@ class Program_weight_tensor_parameter_38: class Program_weight_tensor_parameter_39: name = "parameter_39" + original_name = "linear_4423.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-0.487468") @@ -440,6 +480,7 @@ class Program_weight_tensor_parameter_39: class Program_weight_tensor_parameter_40: name = "parameter_40" + original_name = "linear_4422.b_0" shape = [4096] dtype = "float32" min_val = float("-0.143345") @@ -451,6 +492,7 @@ class Program_weight_tensor_parameter_40: class Program_weight_tensor_parameter_41: name = "parameter_41" + original_name = "linear_4422.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.436439") @@ -462,6 +504,7 @@ class Program_weight_tensor_parameter_41: class Program_weight_tensor_parameter_42: name = "parameter_42" + original_name = "linear_4421.b_0" shape = [1024] dtype = "float32" min_val = float("-0.141163") @@ -473,6 +516,7 @@ class Program_weight_tensor_parameter_42: class Program_weight_tensor_parameter_43: name = "parameter_43" + original_name = "linear_4421.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.27582") @@ -484,6 +528,7 @@ class Program_weight_tensor_parameter_43: class Program_weight_tensor_parameter_44: name = "parameter_44" + original_name = "linear_4420.b_0" shape = [1024] dtype = "float32" min_val = float("-0.134699") @@ -495,6 +540,7 @@ class Program_weight_tensor_parameter_44: class Program_weight_tensor_parameter_45: name = "parameter_45" + original_name = "linear_4420.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.157666") @@ -506,6 +552,7 @@ class Program_weight_tensor_parameter_45: class Program_weight_tensor_parameter_46: name = "parameter_46" + original_name = "linear_4419.b_0" shape = [1024] dtype = "float32" min_val = float("-13.0166") @@ -517,6 +564,7 @@ class Program_weight_tensor_parameter_46: class Program_weight_tensor_parameter_47: name = "parameter_47" + original_name = "linear_4419.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.219939") @@ -528,6 +576,7 @@ class Program_weight_tensor_parameter_47: class Program_weight_tensor_parameter_48: name = "parameter_48" + original_name = "linear_4418.b_0" shape = [1024] dtype = "float32" min_val = float("-0.337815") @@ -539,6 +588,7 @@ class Program_weight_tensor_parameter_48: class Program_weight_tensor_parameter_49: name = "parameter_49" + original_name = "linear_4418.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.321958") @@ -550,6 +600,7 @@ class Program_weight_tensor_parameter_49: class Program_weight_tensor_parameter_50: name = "parameter_50" + original_name = "layer_norm_1522.b_0" shape = [1024] dtype = "float32" min_val = float("-0.59306") @@ -561,6 +612,7 @@ class Program_weight_tensor_parameter_50: class Program_weight_tensor_parameter_51: name = "parameter_51" + original_name = "layer_norm_1522.w_0" shape = [1024] dtype = "float32" min_val = float("0.403322") @@ -572,6 +624,7 @@ class Program_weight_tensor_parameter_51: class Program_weight_tensor_parameter_52: name = "parameter_52" + original_name = "layer_norm_1521.b_0" shape = [1024] dtype = "float32" min_val = float("-1.32993") @@ -583,6 +636,7 @@ class Program_weight_tensor_parameter_52: class Program_weight_tensor_parameter_53: name = "parameter_53" + original_name = "layer_norm_1521.w_0" shape = [1024] dtype = "float32" min_val = float("0.883818") @@ -594,6 +648,7 @@ class Program_weight_tensor_parameter_53: class Program_weight_tensor_parameter_54: name = "parameter_54" + original_name = "linear_4417.b_0" shape = [1024] dtype = "float32" min_val = float("-0.395203") @@ -605,6 +660,7 @@ class Program_weight_tensor_parameter_54: class Program_weight_tensor_parameter_55: name = "parameter_55" + original_name = "linear_4417.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-0.465943") @@ -616,6 +672,7 @@ class Program_weight_tensor_parameter_55: class Program_weight_tensor_parameter_56: name = "parameter_56" + original_name = "linear_4416.b_0" shape = [4096] dtype = "float32" min_val = float("-0.140368") @@ -627,6 +684,7 @@ class Program_weight_tensor_parameter_56: class Program_weight_tensor_parameter_57: name = "parameter_57" + original_name = "linear_4416.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.744701") @@ -638,6 +696,7 @@ class Program_weight_tensor_parameter_57: class Program_weight_tensor_parameter_58: name = "parameter_58" + original_name = "linear_4415.b_0" shape = [1024] dtype = "float32" min_val = float("-0.143136") @@ -649,6 +708,7 @@ class Program_weight_tensor_parameter_58: class Program_weight_tensor_parameter_59: name = "parameter_59" + original_name = "linear_4415.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.210276") @@ -660,6 +720,7 @@ class Program_weight_tensor_parameter_59: class Program_weight_tensor_parameter_60: name = "parameter_60" + original_name = "linear_4414.b_0" shape = [1024] dtype = "float32" min_val = float("-0.096842") @@ -671,6 +732,7 @@ class Program_weight_tensor_parameter_60: class Program_weight_tensor_parameter_61: name = "parameter_61" + original_name = "linear_4414.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.167992") @@ -682,6 +744,7 @@ class Program_weight_tensor_parameter_61: class Program_weight_tensor_parameter_62: name = "parameter_62" + original_name = "linear_4413.b_0" shape = [1024] dtype = "float32" min_val = float("-12.7337") @@ -693,6 +756,7 @@ class Program_weight_tensor_parameter_62: class Program_weight_tensor_parameter_63: name = "parameter_63" + original_name = "linear_4413.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.668661") @@ -704,6 +768,7 @@ class Program_weight_tensor_parameter_63: class Program_weight_tensor_parameter_64: name = "parameter_64" + original_name = "linear_4412.b_0" shape = [1024] dtype = "float32" min_val = float("-0.277926") @@ -715,6 +780,7 @@ class Program_weight_tensor_parameter_64: class Program_weight_tensor_parameter_65: name = "parameter_65" + original_name = "linear_4412.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.233979") @@ -726,6 +792,7 @@ class Program_weight_tensor_parameter_65: class Program_weight_tensor_parameter_66: name = "parameter_66" + original_name = "layer_norm_1520.b_0" shape = [1024] dtype = "float32" min_val = float("-0.502607") @@ -737,6 +804,7 @@ class Program_weight_tensor_parameter_66: class Program_weight_tensor_parameter_67: name = "parameter_67" + original_name = "layer_norm_1520.w_0" shape = [1024] dtype = "float32" min_val = float("0.393191") @@ -748,6 +816,7 @@ class Program_weight_tensor_parameter_67: class Program_weight_tensor_parameter_68: name = "parameter_68" + original_name = "layer_norm_1519.b_0" shape = [1024] dtype = "float32" min_val = float("-1.58179") @@ -759,6 +828,7 @@ class Program_weight_tensor_parameter_68: class Program_weight_tensor_parameter_69: name = "parameter_69" + original_name = "layer_norm_1519.w_0" shape = [1024] dtype = "float32" min_val = float("0.880367") @@ -770,6 +840,7 @@ class Program_weight_tensor_parameter_69: class Program_weight_tensor_parameter_70: name = "parameter_70" + original_name = "linear_4411.b_0" shape = [1024] dtype = "float32" min_val = float("-0.714328") @@ -781,6 +852,7 @@ class Program_weight_tensor_parameter_70: class Program_weight_tensor_parameter_71: name = "parameter_71" + original_name = "linear_4411.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-0.788206") @@ -792,6 +864,7 @@ class Program_weight_tensor_parameter_71: class Program_weight_tensor_parameter_72: name = "parameter_72" + original_name = "linear_4410.b_0" shape = [4096] dtype = "float32" min_val = float("-0.186977") @@ -803,6 +876,7 @@ class Program_weight_tensor_parameter_72: class Program_weight_tensor_parameter_73: name = "parameter_73" + original_name = "linear_4410.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.525018") @@ -814,6 +888,7 @@ class Program_weight_tensor_parameter_73: class Program_weight_tensor_parameter_74: name = "parameter_74" + original_name = "linear_4409.b_0" shape = [1024] dtype = "float32" min_val = float("-0.10216") @@ -825,6 +900,7 @@ class Program_weight_tensor_parameter_74: class Program_weight_tensor_parameter_75: name = "parameter_75" + original_name = "linear_4409.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.286546") @@ -836,6 +912,7 @@ class Program_weight_tensor_parameter_75: class Program_weight_tensor_parameter_76: name = "parameter_76" + original_name = "linear_4408.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0612599") @@ -847,6 +924,7 @@ class Program_weight_tensor_parameter_76: class Program_weight_tensor_parameter_77: name = "parameter_77" + original_name = "linear_4408.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.151695") @@ -858,6 +936,7 @@ class Program_weight_tensor_parameter_77: class Program_weight_tensor_parameter_78: name = "parameter_78" + original_name = "linear_4407.b_0" shape = [1024] dtype = "float32" min_val = float("-14.6946") @@ -869,6 +948,7 @@ class Program_weight_tensor_parameter_78: class Program_weight_tensor_parameter_79: name = "parameter_79" + original_name = "linear_4407.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.652659") @@ -880,6 +960,7 @@ class Program_weight_tensor_parameter_79: class Program_weight_tensor_parameter_80: name = "parameter_80" + original_name = "linear_4406.b_0" shape = [1024] dtype = "float32" min_val = float("-0.303715") @@ -891,6 +972,7 @@ class Program_weight_tensor_parameter_80: class Program_weight_tensor_parameter_81: name = "parameter_81" + original_name = "linear_4406.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.224947") @@ -902,6 +984,7 @@ class Program_weight_tensor_parameter_81: class Program_weight_tensor_parameter_82: name = "parameter_82" + original_name = "layer_norm_1518.b_0" shape = [1024] dtype = "float32" min_val = float("-0.659071") @@ -913,6 +996,7 @@ class Program_weight_tensor_parameter_82: class Program_weight_tensor_parameter_83: name = "parameter_83" + original_name = "layer_norm_1518.w_0" shape = [1024] dtype = "float32" min_val = float("0.580017") @@ -924,6 +1008,7 @@ class Program_weight_tensor_parameter_83: class Program_weight_tensor_parameter_84: name = "parameter_84" + original_name = "layer_norm_1517.b_0" shape = [1024] dtype = "float32" min_val = float("-1.67413") @@ -935,6 +1020,7 @@ class Program_weight_tensor_parameter_84: class Program_weight_tensor_parameter_85: name = "parameter_85" + original_name = "layer_norm_1517.w_0" shape = [1024] dtype = "float32" min_val = float("0.876367") @@ -946,6 +1032,7 @@ class Program_weight_tensor_parameter_85: class Program_weight_tensor_parameter_86: name = "parameter_86" + original_name = "linear_4405.b_0" shape = [1024] dtype = "float32" min_val = float("-0.889271") @@ -957,6 +1044,7 @@ class Program_weight_tensor_parameter_86: class Program_weight_tensor_parameter_87: name = "parameter_87" + original_name = "linear_4405.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-0.982186") @@ -968,6 +1056,7 @@ class Program_weight_tensor_parameter_87: class Program_weight_tensor_parameter_88: name = "parameter_88" + original_name = "linear_4404.b_0" shape = [4096] dtype = "float32" min_val = float("-0.155853") @@ -979,6 +1068,7 @@ class Program_weight_tensor_parameter_88: class Program_weight_tensor_parameter_89: name = "parameter_89" + original_name = "linear_4404.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.377884") @@ -990,6 +1080,7 @@ class Program_weight_tensor_parameter_89: class Program_weight_tensor_parameter_90: name = "parameter_90" + original_name = "linear_4403.b_0" shape = [1024] dtype = "float32" min_val = float("-0.107526") @@ -1001,6 +1092,7 @@ class Program_weight_tensor_parameter_90: class Program_weight_tensor_parameter_91: name = "parameter_91" + original_name = "linear_4403.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.170694") @@ -1012,6 +1104,7 @@ class Program_weight_tensor_parameter_91: class Program_weight_tensor_parameter_92: name = "parameter_92" + original_name = "linear_4402.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0803002") @@ -1023,6 +1116,7 @@ class Program_weight_tensor_parameter_92: class Program_weight_tensor_parameter_93: name = "parameter_93" + original_name = "linear_4402.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.161767") @@ -1034,6 +1128,7 @@ class Program_weight_tensor_parameter_93: class Program_weight_tensor_parameter_94: name = "parameter_94" + original_name = "linear_4401.b_0" shape = [1024] dtype = "float32" min_val = float("-9.01977") @@ -1045,6 +1140,7 @@ class Program_weight_tensor_parameter_94: class Program_weight_tensor_parameter_95: name = "parameter_95" + original_name = "linear_4401.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.399516") @@ -1056,6 +1152,7 @@ class Program_weight_tensor_parameter_95: class Program_weight_tensor_parameter_96: name = "parameter_96" + original_name = "linear_4400.b_0" shape = [1024] dtype = "float32" min_val = float("-0.252241") @@ -1067,6 +1164,7 @@ class Program_weight_tensor_parameter_96: class Program_weight_tensor_parameter_97: name = "parameter_97" + original_name = "linear_4400.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.29496") @@ -1078,6 +1176,7 @@ class Program_weight_tensor_parameter_97: class Program_weight_tensor_parameter_98: name = "parameter_98" + original_name = "layer_norm_1516.b_0" shape = [1024] dtype = "float32" min_val = float("-0.743172") @@ -1089,6 +1188,7 @@ class Program_weight_tensor_parameter_98: class Program_weight_tensor_parameter_99: name = "parameter_99" + original_name = "layer_norm_1516.w_0" shape = [1024] dtype = "float32" min_val = float("0.649582") @@ -1100,6 +1200,7 @@ class Program_weight_tensor_parameter_99: class Program_weight_tensor_parameter_100: name = "parameter_100" + original_name = "layer_norm_1515.b_0" shape = [1024] dtype = "float32" min_val = float("-1.7495") @@ -1111,6 +1212,7 @@ class Program_weight_tensor_parameter_100: class Program_weight_tensor_parameter_101: name = "parameter_101" + original_name = "layer_norm_1515.w_0" shape = [1024] dtype = "float32" min_val = float("0.869435") @@ -1122,6 +1224,7 @@ class Program_weight_tensor_parameter_101: class Program_weight_tensor_parameter_102: name = "parameter_102" + original_name = "linear_4399.b_0" shape = [1024] dtype = "float32" min_val = float("-0.984901") @@ -1133,6 +1236,7 @@ class Program_weight_tensor_parameter_102: class Program_weight_tensor_parameter_103: name = "parameter_103" + original_name = "linear_4399.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-0.813271") @@ -1144,6 +1248,7 @@ class Program_weight_tensor_parameter_103: class Program_weight_tensor_parameter_104: name = "parameter_104" + original_name = "linear_4398.b_0" shape = [4096] dtype = "float32" min_val = float("-0.137839") @@ -1155,6 +1260,7 @@ class Program_weight_tensor_parameter_104: class Program_weight_tensor_parameter_105: name = "parameter_105" + original_name = "linear_4398.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.299116") @@ -1166,6 +1272,7 @@ class Program_weight_tensor_parameter_105: class Program_weight_tensor_parameter_106: name = "parameter_106" + original_name = "linear_4397.b_0" shape = [1024] dtype = "float32" min_val = float("-0.113628") @@ -1177,6 +1284,7 @@ class Program_weight_tensor_parameter_106: class Program_weight_tensor_parameter_107: name = "parameter_107" + original_name = "linear_4397.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.227156") @@ -1188,6 +1296,7 @@ class Program_weight_tensor_parameter_107: class Program_weight_tensor_parameter_108: name = "parameter_108" + original_name = "linear_4396.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0582153") @@ -1199,6 +1308,7 @@ class Program_weight_tensor_parameter_108: class Program_weight_tensor_parameter_109: name = "parameter_109" + original_name = "linear_4396.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.165559") @@ -1210,6 +1320,7 @@ class Program_weight_tensor_parameter_109: class Program_weight_tensor_parameter_110: name = "parameter_110" + original_name = "linear_4395.b_0" shape = [1024] dtype = "float32" min_val = float("-8.24145") @@ -1221,6 +1332,7 @@ class Program_weight_tensor_parameter_110: class Program_weight_tensor_parameter_111: name = "parameter_111" + original_name = "linear_4395.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.430112") @@ -1232,6 +1344,7 @@ class Program_weight_tensor_parameter_111: class Program_weight_tensor_parameter_112: name = "parameter_112" + original_name = "linear_4394.b_0" shape = [1024] dtype = "float32" min_val = float("-0.299611") @@ -1243,6 +1356,7 @@ class Program_weight_tensor_parameter_112: class Program_weight_tensor_parameter_113: name = "parameter_113" + original_name = "linear_4394.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.267537") @@ -1254,6 +1368,7 @@ class Program_weight_tensor_parameter_113: class Program_weight_tensor_parameter_114: name = "parameter_114" + original_name = "layer_norm_1514.b_0" shape = [1024] dtype = "float32" min_val = float("-0.73667") @@ -1265,6 +1380,7 @@ class Program_weight_tensor_parameter_114: class Program_weight_tensor_parameter_115: name = "parameter_115" + original_name = "layer_norm_1514.w_0" shape = [1024] dtype = "float32" min_val = float("0.649054") @@ -1276,6 +1392,7 @@ class Program_weight_tensor_parameter_115: class Program_weight_tensor_parameter_116: name = "parameter_116" + original_name = "layer_norm_1513.b_0" shape = [1024] dtype = "float32" min_val = float("-1.53654") @@ -1287,6 +1404,7 @@ class Program_weight_tensor_parameter_116: class Program_weight_tensor_parameter_117: name = "parameter_117" + original_name = "layer_norm_1513.w_0" shape = [1024] dtype = "float32" min_val = float("0.858606") @@ -1298,6 +1416,7 @@ class Program_weight_tensor_parameter_117: class Program_weight_tensor_parameter_118: name = "parameter_118" + original_name = "linear_4393.b_0" shape = [1024] dtype = "float32" min_val = float("-1.0386") @@ -1309,6 +1428,7 @@ class Program_weight_tensor_parameter_118: class Program_weight_tensor_parameter_119: name = "parameter_119" + original_name = "linear_4393.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-1.05719") @@ -1320,6 +1440,7 @@ class Program_weight_tensor_parameter_119: class Program_weight_tensor_parameter_120: name = "parameter_120" + original_name = "linear_4392.b_0" shape = [4096] dtype = "float32" min_val = float("-0.132885") @@ -1331,6 +1452,7 @@ class Program_weight_tensor_parameter_120: class Program_weight_tensor_parameter_121: name = "parameter_121" + original_name = "linear_4392.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.26813") @@ -1342,6 +1464,7 @@ class Program_weight_tensor_parameter_121: class Program_weight_tensor_parameter_122: name = "parameter_122" + original_name = "linear_4391.b_0" shape = [1024] dtype = "float32" min_val = float("-0.11324") @@ -1353,6 +1476,7 @@ class Program_weight_tensor_parameter_122: class Program_weight_tensor_parameter_123: name = "parameter_123" + original_name = "linear_4391.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.231549") @@ -1364,6 +1488,7 @@ class Program_weight_tensor_parameter_123: class Program_weight_tensor_parameter_124: name = "parameter_124" + original_name = "linear_4390.b_0" shape = [1024] dtype = "float32" min_val = float("-0.050542") @@ -1375,6 +1500,7 @@ class Program_weight_tensor_parameter_124: class Program_weight_tensor_parameter_125: name = "parameter_125" + original_name = "linear_4390.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.146931") @@ -1386,6 +1512,7 @@ class Program_weight_tensor_parameter_125: class Program_weight_tensor_parameter_126: name = "parameter_126" + original_name = "linear_4389.b_0" shape = [1024] dtype = "float32" min_val = float("-6.41581") @@ -1397,6 +1524,7 @@ class Program_weight_tensor_parameter_126: class Program_weight_tensor_parameter_127: name = "parameter_127" + original_name = "linear_4389.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.404083") @@ -1408,6 +1536,7 @@ class Program_weight_tensor_parameter_127: class Program_weight_tensor_parameter_128: name = "parameter_128" + original_name = "linear_4388.b_0" shape = [1024] dtype = "float32" min_val = float("-0.233182") @@ -1419,6 +1548,7 @@ class Program_weight_tensor_parameter_128: class Program_weight_tensor_parameter_129: name = "parameter_129" + original_name = "linear_4388.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.25552") @@ -1430,6 +1560,7 @@ class Program_weight_tensor_parameter_129: class Program_weight_tensor_parameter_130: name = "parameter_130" + original_name = "layer_norm_1512.b_0" shape = [1024] dtype = "float32" min_val = float("-0.702883") @@ -1441,6 +1572,7 @@ class Program_weight_tensor_parameter_130: class Program_weight_tensor_parameter_131: name = "parameter_131" + original_name = "layer_norm_1512.w_0" shape = [1024] dtype = "float32" min_val = float("0.659601") @@ -1452,6 +1584,7 @@ class Program_weight_tensor_parameter_131: class Program_weight_tensor_parameter_132: name = "parameter_132" + original_name = "layer_norm_1511.b_0" shape = [1024] dtype = "float32" min_val = float("-0.961858") @@ -1463,6 +1596,7 @@ class Program_weight_tensor_parameter_132: class Program_weight_tensor_parameter_133: name = "parameter_133" + original_name = "layer_norm_1511.w_0" shape = [1024] dtype = "float32" min_val = float("0.839306") @@ -1474,6 +1608,7 @@ class Program_weight_tensor_parameter_133: class Program_weight_tensor_parameter_134: name = "parameter_134" + original_name = "linear_4387.b_0" shape = [1024] dtype = "float32" min_val = float("-1.03147") @@ -1485,6 +1620,7 @@ class Program_weight_tensor_parameter_134: class Program_weight_tensor_parameter_135: name = "parameter_135" + original_name = "linear_4387.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-1.40921") @@ -1496,6 +1632,7 @@ class Program_weight_tensor_parameter_135: class Program_weight_tensor_parameter_136: name = "parameter_136" + original_name = "linear_4386.b_0" shape = [4096] dtype = "float32" min_val = float("-0.147209") @@ -1507,6 +1644,7 @@ class Program_weight_tensor_parameter_136: class Program_weight_tensor_parameter_137: name = "parameter_137" + original_name = "linear_4386.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.28167") @@ -1518,6 +1656,7 @@ class Program_weight_tensor_parameter_137: class Program_weight_tensor_parameter_138: name = "parameter_138" + original_name = "linear_4385.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0864816") @@ -1529,6 +1668,7 @@ class Program_weight_tensor_parameter_138: class Program_weight_tensor_parameter_139: name = "parameter_139" + original_name = "linear_4385.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.390454") @@ -1540,6 +1680,7 @@ class Program_weight_tensor_parameter_139: class Program_weight_tensor_parameter_140: name = "parameter_140" + original_name = "linear_4384.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0487462") @@ -1551,6 +1692,7 @@ class Program_weight_tensor_parameter_140: class Program_weight_tensor_parameter_141: name = "parameter_141" + original_name = "linear_4384.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.144965") @@ -1562,6 +1704,7 @@ class Program_weight_tensor_parameter_141: class Program_weight_tensor_parameter_142: name = "parameter_142" + original_name = "linear_4383.b_0" shape = [1024] dtype = "float32" min_val = float("-7.79869") @@ -1573,6 +1716,7 @@ class Program_weight_tensor_parameter_142: class Program_weight_tensor_parameter_143: name = "parameter_143" + original_name = "linear_4383.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.595891") @@ -1584,6 +1728,7 @@ class Program_weight_tensor_parameter_143: class Program_weight_tensor_parameter_144: name = "parameter_144" + original_name = "linear_4382.b_0" shape = [1024] dtype = "float32" min_val = float("-0.183592") @@ -1595,6 +1740,7 @@ class Program_weight_tensor_parameter_144: class Program_weight_tensor_parameter_145: name = "parameter_145" + original_name = "linear_4382.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.292882") @@ -1606,6 +1752,7 @@ class Program_weight_tensor_parameter_145: class Program_weight_tensor_parameter_146: name = "parameter_146" + original_name = "layer_norm_1510.b_0" shape = [1024] dtype = "float32" min_val = float("-0.385236") @@ -1617,6 +1764,7 @@ class Program_weight_tensor_parameter_146: class Program_weight_tensor_parameter_147: name = "parameter_147" + original_name = "layer_norm_1510.w_0" shape = [1024] dtype = "float32" min_val = float("0.703691") @@ -1628,6 +1776,7 @@ class Program_weight_tensor_parameter_147: class Program_weight_tensor_parameter_148: name = "parameter_148" + original_name = "layer_norm_1509.b_0" shape = [1024] dtype = "float32" min_val = float("-0.53284") @@ -1639,6 +1788,7 @@ class Program_weight_tensor_parameter_148: class Program_weight_tensor_parameter_149: name = "parameter_149" + original_name = "layer_norm_1509.w_0" shape = [1024] dtype = "float32" min_val = float("0.875238") @@ -1650,6 +1800,7 @@ class Program_weight_tensor_parameter_149: class Program_weight_tensor_parameter_150: name = "parameter_150" + original_name = "linear_4381.b_0" shape = [1024] dtype = "float32" min_val = float("-1.08931") @@ -1661,6 +1812,7 @@ class Program_weight_tensor_parameter_150: class Program_weight_tensor_parameter_151: name = "parameter_151" + original_name = "linear_4381.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-1.1203") @@ -1672,6 +1824,7 @@ class Program_weight_tensor_parameter_151: class Program_weight_tensor_parameter_152: name = "parameter_152" + original_name = "linear_4380.b_0" shape = [4096] dtype = "float32" min_val = float("-0.136849") @@ -1683,6 +1836,7 @@ class Program_weight_tensor_parameter_152: class Program_weight_tensor_parameter_153: name = "parameter_153" + original_name = "linear_4380.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.231231") @@ -1694,6 +1848,7 @@ class Program_weight_tensor_parameter_153: class Program_weight_tensor_parameter_154: name = "parameter_154" + original_name = "linear_4379.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0927901") @@ -1705,6 +1860,7 @@ class Program_weight_tensor_parameter_154: class Program_weight_tensor_parameter_155: name = "parameter_155" + original_name = "linear_4379.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.259299") @@ -1716,6 +1872,7 @@ class Program_weight_tensor_parameter_155: class Program_weight_tensor_parameter_156: name = "parameter_156" + original_name = "linear_4378.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0548148") @@ -1727,6 +1884,7 @@ class Program_weight_tensor_parameter_156: class Program_weight_tensor_parameter_157: name = "parameter_157" + original_name = "linear_4378.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.163605") @@ -1738,6 +1896,7 @@ class Program_weight_tensor_parameter_157: class Program_weight_tensor_parameter_158: name = "parameter_158" + original_name = "linear_4377.b_0" shape = [1024] dtype = "float32" min_val = float("-4.89877") @@ -1749,6 +1908,7 @@ class Program_weight_tensor_parameter_158: class Program_weight_tensor_parameter_159: name = "parameter_159" + original_name = "linear_4377.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.276939") @@ -1760,6 +1920,7 @@ class Program_weight_tensor_parameter_159: class Program_weight_tensor_parameter_160: name = "parameter_160" + original_name = "linear_4376.b_0" shape = [1024] dtype = "float32" min_val = float("-0.196304") @@ -1771,6 +1932,7 @@ class Program_weight_tensor_parameter_160: class Program_weight_tensor_parameter_161: name = "parameter_161" + original_name = "linear_4376.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.297006") @@ -1782,6 +1944,7 @@ class Program_weight_tensor_parameter_161: class Program_weight_tensor_parameter_162: name = "parameter_162" + original_name = "layer_norm_1508.b_0" shape = [1024] dtype = "float32" min_val = float("-0.131729") @@ -1793,6 +1956,7 @@ class Program_weight_tensor_parameter_162: class Program_weight_tensor_parameter_163: name = "parameter_163" + original_name = "layer_norm_1508.w_0" shape = [1024] dtype = "float32" min_val = float("0.538438") @@ -1804,6 +1968,7 @@ class Program_weight_tensor_parameter_163: class Program_weight_tensor_parameter_164: name = "parameter_164" + original_name = "layer_norm_1507.b_0" shape = [1024] dtype = "float32" min_val = float("-0.378343") @@ -1815,6 +1980,7 @@ class Program_weight_tensor_parameter_164: class Program_weight_tensor_parameter_165: name = "parameter_165" + original_name = "layer_norm_1507.w_0" shape = [1024] dtype = "float32" min_val = float("0.868218") @@ -1826,6 +1992,7 @@ class Program_weight_tensor_parameter_165: class Program_weight_tensor_parameter_166: name = "parameter_166" + original_name = "linear_4375.b_0" shape = [1024] dtype = "float32" min_val = float("-1.02504") @@ -1837,6 +2004,7 @@ class Program_weight_tensor_parameter_166: class Program_weight_tensor_parameter_167: name = "parameter_167" + original_name = "linear_4375.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-0.740766") @@ -1848,6 +2016,7 @@ class Program_weight_tensor_parameter_167: class Program_weight_tensor_parameter_168: name = "parameter_168" + original_name = "linear_4374.b_0" shape = [4096] dtype = "float32" min_val = float("-0.104707") @@ -1859,6 +2028,7 @@ class Program_weight_tensor_parameter_168: class Program_weight_tensor_parameter_169: name = "parameter_169" + original_name = "linear_4374.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.212528") @@ -1870,6 +2040,7 @@ class Program_weight_tensor_parameter_169: class Program_weight_tensor_parameter_170: name = "parameter_170" + original_name = "linear_4373.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0790037") @@ -1881,6 +2052,7 @@ class Program_weight_tensor_parameter_170: class Program_weight_tensor_parameter_171: name = "parameter_171" + original_name = "linear_4373.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.233005") @@ -1892,6 +2064,7 @@ class Program_weight_tensor_parameter_171: class Program_weight_tensor_parameter_172: name = "parameter_172" + original_name = "linear_4372.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0502381") @@ -1903,6 +2076,7 @@ class Program_weight_tensor_parameter_172: class Program_weight_tensor_parameter_173: name = "parameter_173" + original_name = "linear_4372.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.187032") @@ -1914,6 +2088,7 @@ class Program_weight_tensor_parameter_173: class Program_weight_tensor_parameter_174: name = "parameter_174" + original_name = "linear_4371.b_0" shape = [1024] dtype = "float32" min_val = float("-5.19654") @@ -1925,6 +2100,7 @@ class Program_weight_tensor_parameter_174: class Program_weight_tensor_parameter_175: name = "parameter_175" + original_name = "linear_4371.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.279211") @@ -1936,6 +2112,7 @@ class Program_weight_tensor_parameter_175: class Program_weight_tensor_parameter_176: name = "parameter_176" + original_name = "linear_4370.b_0" shape = [1024] dtype = "float32" min_val = float("-0.185898") @@ -1947,6 +2124,7 @@ class Program_weight_tensor_parameter_176: class Program_weight_tensor_parameter_177: name = "parameter_177" + original_name = "linear_4370.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.257178") @@ -1958,6 +2136,7 @@ class Program_weight_tensor_parameter_177: class Program_weight_tensor_parameter_178: name = "parameter_178" + original_name = "layer_norm_1506.b_0" shape = [1024] dtype = "float32" min_val = float("-0.200058") @@ -1969,6 +2148,7 @@ class Program_weight_tensor_parameter_178: class Program_weight_tensor_parameter_179: name = "parameter_179" + original_name = "layer_norm_1506.w_0" shape = [1024] dtype = "float32" min_val = float("0.395205") @@ -1980,6 +2160,7 @@ class Program_weight_tensor_parameter_179: class Program_weight_tensor_parameter_180: name = "parameter_180" + original_name = "layer_norm_1505.b_0" shape = [1024] dtype = "float32" min_val = float("-0.678026") @@ -1991,6 +2172,7 @@ class Program_weight_tensor_parameter_180: class Program_weight_tensor_parameter_181: name = "parameter_181" + original_name = "layer_norm_1505.w_0" shape = [1024] dtype = "float32" min_val = float("0.876021") @@ -2002,6 +2184,7 @@ class Program_weight_tensor_parameter_181: class Program_weight_tensor_parameter_182: name = "parameter_182" + original_name = "linear_4369.b_0" shape = [1024] dtype = "float32" min_val = float("-0.979984") @@ -2013,6 +2196,7 @@ class Program_weight_tensor_parameter_182: class Program_weight_tensor_parameter_183: name = "parameter_183" + original_name = "linear_4369.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-0.496973") @@ -2024,6 +2208,7 @@ class Program_weight_tensor_parameter_183: class Program_weight_tensor_parameter_184: name = "parameter_184" + original_name = "linear_4368.b_0" shape = [4096] dtype = "float32" min_val = float("-0.0970088") @@ -2035,6 +2220,7 @@ class Program_weight_tensor_parameter_184: class Program_weight_tensor_parameter_185: name = "parameter_185" + original_name = "linear_4368.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.267951") @@ -2046,6 +2232,7 @@ class Program_weight_tensor_parameter_185: class Program_weight_tensor_parameter_186: name = "parameter_186" + original_name = "linear_4367.b_0" shape = [1024] dtype = "float32" min_val = float("-0.145309") @@ -2057,6 +2244,7 @@ class Program_weight_tensor_parameter_186: class Program_weight_tensor_parameter_187: name = "parameter_187" + original_name = "linear_4367.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.197545") @@ -2068,6 +2256,7 @@ class Program_weight_tensor_parameter_187: class Program_weight_tensor_parameter_188: name = "parameter_188" + original_name = "linear_4366.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0467914") @@ -2079,6 +2268,7 @@ class Program_weight_tensor_parameter_188: class Program_weight_tensor_parameter_189: name = "parameter_189" + original_name = "linear_4366.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.163887") @@ -2090,6 +2280,7 @@ class Program_weight_tensor_parameter_189: class Program_weight_tensor_parameter_190: name = "parameter_190" + original_name = "linear_4365.b_0" shape = [1024] dtype = "float32" min_val = float("-4.16524") @@ -2101,6 +2292,7 @@ class Program_weight_tensor_parameter_190: class Program_weight_tensor_parameter_191: name = "parameter_191" + original_name = "linear_4365.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.221266") @@ -2112,6 +2304,7 @@ class Program_weight_tensor_parameter_191: class Program_weight_tensor_parameter_192: name = "parameter_192" + original_name = "linear_4364.b_0" shape = [1024] dtype = "float32" min_val = float("-0.206744") @@ -2123,6 +2316,7 @@ class Program_weight_tensor_parameter_192: class Program_weight_tensor_parameter_193: name = "parameter_193" + original_name = "linear_4364.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.188052") @@ -2134,6 +2328,7 @@ class Program_weight_tensor_parameter_193: class Program_weight_tensor_parameter_194: name = "parameter_194" + original_name = "layer_norm_1504.b_0" shape = [1024] dtype = "float32" min_val = float("-0.203856") @@ -2145,6 +2340,7 @@ class Program_weight_tensor_parameter_194: class Program_weight_tensor_parameter_195: name = "parameter_195" + original_name = "layer_norm_1504.w_0" shape = [1024] dtype = "float32" min_val = float("0.259144") @@ -2156,6 +2352,7 @@ class Program_weight_tensor_parameter_195: class Program_weight_tensor_parameter_196: name = "parameter_196" + original_name = "layer_norm_1503.b_0" shape = [1024] dtype = "float32" min_val = float("-0.987588") @@ -2167,6 +2364,7 @@ class Program_weight_tensor_parameter_196: class Program_weight_tensor_parameter_197: name = "parameter_197" + original_name = "layer_norm_1503.w_0" shape = [1024] dtype = "float32" min_val = float("0.886998") @@ -2178,6 +2376,7 @@ class Program_weight_tensor_parameter_197: class Program_weight_tensor_parameter_198: name = "parameter_198" + original_name = "linear_4363.b_0" shape = [1024] dtype = "float32" min_val = float("-0.693036") @@ -2189,6 +2388,7 @@ class Program_weight_tensor_parameter_198: class Program_weight_tensor_parameter_199: name = "parameter_199" + original_name = "linear_4363.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-0.643998") @@ -2200,6 +2400,7 @@ class Program_weight_tensor_parameter_199: class Program_weight_tensor_parameter_200: name = "parameter_200" + original_name = "linear_4362.b_0" shape = [4096] dtype = "float32" min_val = float("-0.0914531") @@ -2211,6 +2412,7 @@ class Program_weight_tensor_parameter_200: class Program_weight_tensor_parameter_201: name = "parameter_201" + original_name = "linear_4362.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.22384") @@ -2222,6 +2424,7 @@ class Program_weight_tensor_parameter_201: class Program_weight_tensor_parameter_202: name = "parameter_202" + original_name = "linear_4361.b_0" shape = [1024] dtype = "float32" min_val = float("-0.246896") @@ -2233,6 +2436,7 @@ class Program_weight_tensor_parameter_202: class Program_weight_tensor_parameter_203: name = "parameter_203" + original_name = "linear_4361.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.155957") @@ -2244,6 +2448,7 @@ class Program_weight_tensor_parameter_203: class Program_weight_tensor_parameter_204: name = "parameter_204" + original_name = "linear_4360.b_0" shape = [1024] dtype = "float32" min_val = float("-0.039385") @@ -2255,6 +2460,7 @@ class Program_weight_tensor_parameter_204: class Program_weight_tensor_parameter_205: name = "parameter_205" + original_name = "linear_4360.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.154352") @@ -2266,6 +2472,7 @@ class Program_weight_tensor_parameter_205: class Program_weight_tensor_parameter_206: name = "parameter_206" + original_name = "linear_4359.b_0" shape = [1024] dtype = "float32" min_val = float("-4.89885") @@ -2277,6 +2484,7 @@ class Program_weight_tensor_parameter_206: class Program_weight_tensor_parameter_207: name = "parameter_207" + original_name = "linear_4359.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.322668") @@ -2288,6 +2496,7 @@ class Program_weight_tensor_parameter_207: class Program_weight_tensor_parameter_208: name = "parameter_208" + original_name = "linear_4358.b_0" shape = [1024] dtype = "float32" min_val = float("-0.178589") @@ -2299,6 +2508,7 @@ class Program_weight_tensor_parameter_208: class Program_weight_tensor_parameter_209: name = "parameter_209" + original_name = "linear_4358.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.192735") @@ -2310,6 +2520,7 @@ class Program_weight_tensor_parameter_209: class Program_weight_tensor_parameter_210: name = "parameter_210" + original_name = "layer_norm_1502.b_0" shape = [1024] dtype = "float32" min_val = float("-0.234758") @@ -2321,6 +2532,7 @@ class Program_weight_tensor_parameter_210: class Program_weight_tensor_parameter_211: name = "parameter_211" + original_name = "layer_norm_1502.w_0" shape = [1024] dtype = "float32" min_val = float("0.301765") @@ -2332,6 +2544,7 @@ class Program_weight_tensor_parameter_211: class Program_weight_tensor_parameter_212: name = "parameter_212" + original_name = "layer_norm_1501.b_0" shape = [1024] dtype = "float32" min_val = float("-1.12608") @@ -2343,6 +2556,7 @@ class Program_weight_tensor_parameter_212: class Program_weight_tensor_parameter_213: name = "parameter_213" + original_name = "layer_norm_1501.w_0" shape = [1024] dtype = "float32" min_val = float("0.85224") @@ -2354,6 +2568,7 @@ class Program_weight_tensor_parameter_213: class Program_weight_tensor_parameter_214: name = "parameter_214" + original_name = "linear_4357.b_0" shape = [1024] dtype = "float32" min_val = float("-0.475892") @@ -2365,6 +2580,7 @@ class Program_weight_tensor_parameter_214: class Program_weight_tensor_parameter_215: name = "parameter_215" + original_name = "linear_4357.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-0.717258") @@ -2376,6 +2592,7 @@ class Program_weight_tensor_parameter_215: class Program_weight_tensor_parameter_216: name = "parameter_216" + original_name = "linear_4356.b_0" shape = [4096] dtype = "float32" min_val = float("-0.0982823") @@ -2387,6 +2604,7 @@ class Program_weight_tensor_parameter_216: class Program_weight_tensor_parameter_217: name = "parameter_217" + original_name = "linear_4356.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.308097") @@ -2398,6 +2616,7 @@ class Program_weight_tensor_parameter_217: class Program_weight_tensor_parameter_218: name = "parameter_218" + original_name = "linear_4355.b_0" shape = [1024] dtype = "float32" min_val = float("-0.239741") @@ -2409,6 +2628,7 @@ class Program_weight_tensor_parameter_218: class Program_weight_tensor_parameter_219: name = "parameter_219" + original_name = "linear_4355.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.223419") @@ -2420,6 +2640,7 @@ class Program_weight_tensor_parameter_219: class Program_weight_tensor_parameter_220: name = "parameter_220" + original_name = "linear_4354.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0305229") @@ -2431,6 +2652,7 @@ class Program_weight_tensor_parameter_220: class Program_weight_tensor_parameter_221: name = "parameter_221" + original_name = "linear_4354.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.184504") @@ -2442,6 +2664,7 @@ class Program_weight_tensor_parameter_221: class Program_weight_tensor_parameter_222: name = "parameter_222" + original_name = "linear_4353.b_0" shape = [1024] dtype = "float32" min_val = float("-4.08417") @@ -2453,6 +2676,7 @@ class Program_weight_tensor_parameter_222: class Program_weight_tensor_parameter_223: name = "parameter_223" + original_name = "linear_4353.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.474704") @@ -2464,6 +2688,7 @@ class Program_weight_tensor_parameter_223: class Program_weight_tensor_parameter_224: name = "parameter_224" + original_name = "linear_4352.b_0" shape = [1024] dtype = "float32" min_val = float("-0.215042") @@ -2475,6 +2700,7 @@ class Program_weight_tensor_parameter_224: class Program_weight_tensor_parameter_225: name = "parameter_225" + original_name = "linear_4352.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.229325") @@ -2486,6 +2712,7 @@ class Program_weight_tensor_parameter_225: class Program_weight_tensor_parameter_226: name = "parameter_226" + original_name = "layer_norm_1500.b_0" shape = [1024] dtype = "float32" min_val = float("-0.255976") @@ -2497,6 +2724,7 @@ class Program_weight_tensor_parameter_226: class Program_weight_tensor_parameter_227: name = "parameter_227" + original_name = "layer_norm_1500.w_0" shape = [1024] dtype = "float32" min_val = float("0.461913") @@ -2508,6 +2736,7 @@ class Program_weight_tensor_parameter_227: class Program_weight_tensor_parameter_228: name = "parameter_228" + original_name = "layer_norm_1499.b_0" shape = [1024] dtype = "float32" min_val = float("-1.0016") @@ -2519,6 +2748,7 @@ class Program_weight_tensor_parameter_228: class Program_weight_tensor_parameter_229: name = "parameter_229" + original_name = "layer_norm_1499.w_0" shape = [1024] dtype = "float32" min_val = float("0.891816") @@ -2530,6 +2760,7 @@ class Program_weight_tensor_parameter_229: class Program_weight_tensor_parameter_230: name = "parameter_230" + original_name = "linear_4351.b_0" shape = [1024] dtype = "float32" min_val = float("-0.334385") @@ -2541,6 +2772,7 @@ class Program_weight_tensor_parameter_230: class Program_weight_tensor_parameter_231: name = "parameter_231" + original_name = "linear_4351.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-1.32596") @@ -2552,6 +2784,7 @@ class Program_weight_tensor_parameter_231: class Program_weight_tensor_parameter_232: name = "parameter_232" + original_name = "linear_4350.b_0" shape = [4096] dtype = "float32" min_val = float("-0.102243") @@ -2563,6 +2796,7 @@ class Program_weight_tensor_parameter_232: class Program_weight_tensor_parameter_233: name = "parameter_233" + original_name = "linear_4350.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.306698") @@ -2574,6 +2808,7 @@ class Program_weight_tensor_parameter_233: class Program_weight_tensor_parameter_234: name = "parameter_234" + original_name = "linear_4349.b_0" shape = [1024] dtype = "float32" min_val = float("-0.247143") @@ -2585,6 +2820,7 @@ class Program_weight_tensor_parameter_234: class Program_weight_tensor_parameter_235: name = "parameter_235" + original_name = "linear_4349.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.242832") @@ -2596,6 +2832,7 @@ class Program_weight_tensor_parameter_235: class Program_weight_tensor_parameter_236: name = "parameter_236" + original_name = "linear_4348.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0395913") @@ -2607,6 +2844,7 @@ class Program_weight_tensor_parameter_236: class Program_weight_tensor_parameter_237: name = "parameter_237" + original_name = "linear_4348.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.192363") @@ -2618,6 +2856,7 @@ class Program_weight_tensor_parameter_237: class Program_weight_tensor_parameter_238: name = "parameter_238" + original_name = "linear_4347.b_0" shape = [1024] dtype = "float32" min_val = float("-3.39001") @@ -2629,6 +2868,7 @@ class Program_weight_tensor_parameter_238: class Program_weight_tensor_parameter_239: name = "parameter_239" + original_name = "linear_4347.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.358112") @@ -2640,6 +2880,7 @@ class Program_weight_tensor_parameter_239: class Program_weight_tensor_parameter_240: name = "parameter_240" + original_name = "linear_4346.b_0" shape = [1024] dtype = "float32" min_val = float("-0.232386") @@ -2651,6 +2892,7 @@ class Program_weight_tensor_parameter_240: class Program_weight_tensor_parameter_241: name = "parameter_241" + original_name = "linear_4346.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.249884") @@ -2662,6 +2904,7 @@ class Program_weight_tensor_parameter_241: class Program_weight_tensor_parameter_242: name = "parameter_242" + original_name = "layer_norm_1498.b_0" shape = [1024] dtype = "float32" min_val = float("-0.225377") @@ -2673,6 +2916,7 @@ class Program_weight_tensor_parameter_242: class Program_weight_tensor_parameter_243: name = "parameter_243" + original_name = "layer_norm_1498.w_0" shape = [1024] dtype = "float32" min_val = float("0.470048") @@ -2684,6 +2928,7 @@ class Program_weight_tensor_parameter_243: class Program_weight_tensor_parameter_244: name = "parameter_244" + original_name = "layer_norm_1497.b_0" shape = [1024] dtype = "float32" min_val = float("-1.0262") @@ -2695,6 +2940,7 @@ class Program_weight_tensor_parameter_244: class Program_weight_tensor_parameter_245: name = "parameter_245" + original_name = "layer_norm_1497.w_0" shape = [1024] dtype = "float32" min_val = float("0.871368") @@ -2706,6 +2952,7 @@ class Program_weight_tensor_parameter_245: class Program_weight_tensor_parameter_246: name = "parameter_246" + original_name = "linear_4345.b_0" shape = [1024] dtype = "float32" min_val = float("-0.308582") @@ -2717,6 +2964,7 @@ class Program_weight_tensor_parameter_246: class Program_weight_tensor_parameter_247: name = "parameter_247" + original_name = "linear_4345.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-1.75468") @@ -2728,6 +2976,7 @@ class Program_weight_tensor_parameter_247: class Program_weight_tensor_parameter_248: name = "parameter_248" + original_name = "linear_4344.b_0" shape = [4096] dtype = "float32" min_val = float("-0.0923743") @@ -2739,6 +2988,7 @@ class Program_weight_tensor_parameter_248: class Program_weight_tensor_parameter_249: name = "parameter_249" + original_name = "linear_4344.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.307683") @@ -2750,6 +3000,7 @@ class Program_weight_tensor_parameter_249: class Program_weight_tensor_parameter_250: name = "parameter_250" + original_name = "linear_4343.b_0" shape = [1024] dtype = "float32" min_val = float("-0.141345") @@ -2761,6 +3012,7 @@ class Program_weight_tensor_parameter_250: class Program_weight_tensor_parameter_251: name = "parameter_251" + original_name = "linear_4343.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.246243") @@ -2772,6 +3024,7 @@ class Program_weight_tensor_parameter_251: class Program_weight_tensor_parameter_252: name = "parameter_252" + original_name = "linear_4342.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0457099") @@ -2783,6 +3036,7 @@ class Program_weight_tensor_parameter_252: class Program_weight_tensor_parameter_253: name = "parameter_253" + original_name = "linear_4342.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.154555") @@ -2794,6 +3048,7 @@ class Program_weight_tensor_parameter_253: class Program_weight_tensor_parameter_254: name = "parameter_254" + original_name = "linear_4341.b_0" shape = [1024] dtype = "float32" min_val = float("-4.32052") @@ -2805,6 +3060,7 @@ class Program_weight_tensor_parameter_254: class Program_weight_tensor_parameter_255: name = "parameter_255" + original_name = "linear_4341.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.459937") @@ -2816,6 +3072,7 @@ class Program_weight_tensor_parameter_255: class Program_weight_tensor_parameter_256: name = "parameter_256" + original_name = "linear_4340.b_0" shape = [1024] dtype = "float32" min_val = float("-0.181096") @@ -2827,6 +3084,7 @@ class Program_weight_tensor_parameter_256: class Program_weight_tensor_parameter_257: name = "parameter_257" + original_name = "linear_4340.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.289184") @@ -2838,6 +3096,7 @@ class Program_weight_tensor_parameter_257: class Program_weight_tensor_parameter_258: name = "parameter_258" + original_name = "layer_norm_1496.b_0" shape = [1024] dtype = "float32" min_val = float("-0.26885") @@ -2849,6 +3108,7 @@ class Program_weight_tensor_parameter_258: class Program_weight_tensor_parameter_259: name = "parameter_259" + original_name = "layer_norm_1496.w_0" shape = [1024] dtype = "float32" min_val = float("0.498935") @@ -2860,6 +3120,7 @@ class Program_weight_tensor_parameter_259: class Program_weight_tensor_parameter_260: name = "parameter_260" + original_name = "layer_norm_1495.b_0" shape = [1024] dtype = "float32" min_val = float("-1.08955") @@ -2871,6 +3132,7 @@ class Program_weight_tensor_parameter_260: class Program_weight_tensor_parameter_261: name = "parameter_261" + original_name = "layer_norm_1495.w_0" shape = [1024] dtype = "float32" min_val = float("0.796824") @@ -2882,6 +3144,7 @@ class Program_weight_tensor_parameter_261: class Program_weight_tensor_parameter_262: name = "parameter_262" + original_name = "linear_4339.b_0" shape = [1024] dtype = "float32" min_val = float("-0.344467") @@ -2893,6 +3156,7 @@ class Program_weight_tensor_parameter_262: class Program_weight_tensor_parameter_263: name = "parameter_263" + original_name = "linear_4339.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-1.93241") @@ -2904,6 +3168,7 @@ class Program_weight_tensor_parameter_263: class Program_weight_tensor_parameter_264: name = "parameter_264" + original_name = "linear_4338.b_0" shape = [4096] dtype = "float32" min_val = float("-0.0913696") @@ -2915,6 +3180,7 @@ class Program_weight_tensor_parameter_264: class Program_weight_tensor_parameter_265: name = "parameter_265" + original_name = "linear_4338.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.241158") @@ -2926,6 +3192,7 @@ class Program_weight_tensor_parameter_265: class Program_weight_tensor_parameter_266: name = "parameter_266" + original_name = "linear_4337.b_0" shape = [1024] dtype = "float32" min_val = float("-0.144103") @@ -2937,6 +3204,7 @@ class Program_weight_tensor_parameter_266: class Program_weight_tensor_parameter_267: name = "parameter_267" + original_name = "linear_4337.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.2979") @@ -2948,6 +3216,7 @@ class Program_weight_tensor_parameter_267: class Program_weight_tensor_parameter_268: name = "parameter_268" + original_name = "linear_4336.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0454618") @@ -2959,6 +3228,7 @@ class Program_weight_tensor_parameter_268: class Program_weight_tensor_parameter_269: name = "parameter_269" + original_name = "linear_4336.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.273869") @@ -2970,6 +3240,7 @@ class Program_weight_tensor_parameter_269: class Program_weight_tensor_parameter_270: name = "parameter_270" + original_name = "linear_4335.b_0" shape = [1024] dtype = "float32" min_val = float("-4.07753") @@ -2981,6 +3252,7 @@ class Program_weight_tensor_parameter_270: class Program_weight_tensor_parameter_271: name = "parameter_271" + original_name = "linear_4335.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.314924") @@ -2992,6 +3264,7 @@ class Program_weight_tensor_parameter_271: class Program_weight_tensor_parameter_272: name = "parameter_272" + original_name = "linear_4334.b_0" shape = [1024] dtype = "float32" min_val = float("-0.162246") @@ -3003,6 +3276,7 @@ class Program_weight_tensor_parameter_272: class Program_weight_tensor_parameter_273: name = "parameter_273" + original_name = "linear_4334.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.302558") @@ -3014,6 +3288,7 @@ class Program_weight_tensor_parameter_273: class Program_weight_tensor_parameter_274: name = "parameter_274" + original_name = "layer_norm_1494.b_0" shape = [1024] dtype = "float32" min_val = float("-0.411918") @@ -3025,6 +3300,7 @@ class Program_weight_tensor_parameter_274: class Program_weight_tensor_parameter_275: name = "parameter_275" + original_name = "layer_norm_1494.w_0" shape = [1024] dtype = "float32" min_val = float("0.556571") @@ -3036,6 +3312,7 @@ class Program_weight_tensor_parameter_275: class Program_weight_tensor_parameter_276: name = "parameter_276" + original_name = "layer_norm_1493.b_0" shape = [1024] dtype = "float32" min_val = float("-1.14962") @@ -3047,6 +3324,7 @@ class Program_weight_tensor_parameter_276: class Program_weight_tensor_parameter_277: name = "parameter_277" + original_name = "layer_norm_1493.w_0" shape = [1024] dtype = "float32" min_val = float("0.798682") @@ -3058,6 +3336,7 @@ class Program_weight_tensor_parameter_277: class Program_weight_tensor_parameter_278: name = "parameter_278" + original_name = "linear_4333.b_0" shape = [1024] dtype = "float32" min_val = float("-0.362652") @@ -3069,6 +3348,7 @@ class Program_weight_tensor_parameter_278: class Program_weight_tensor_parameter_279: name = "parameter_279" + original_name = "linear_4333.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-1.91489") @@ -3080,6 +3360,7 @@ class Program_weight_tensor_parameter_279: class Program_weight_tensor_parameter_280: name = "parameter_280" + original_name = "linear_4332.b_0" shape = [4096] dtype = "float32" min_val = float("-0.106405") @@ -3091,6 +3372,7 @@ class Program_weight_tensor_parameter_280: class Program_weight_tensor_parameter_281: name = "parameter_281" + original_name = "linear_4332.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.229709") @@ -3102,6 +3384,7 @@ class Program_weight_tensor_parameter_281: class Program_weight_tensor_parameter_282: name = "parameter_282" + original_name = "linear_4331.b_0" shape = [1024] dtype = "float32" min_val = float("-0.140983") @@ -3113,6 +3396,7 @@ class Program_weight_tensor_parameter_282: class Program_weight_tensor_parameter_283: name = "parameter_283" + original_name = "linear_4331.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.400748") @@ -3124,6 +3408,7 @@ class Program_weight_tensor_parameter_283: class Program_weight_tensor_parameter_284: name = "parameter_284" + original_name = "linear_4330.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0454475") @@ -3135,6 +3420,7 @@ class Program_weight_tensor_parameter_284: class Program_weight_tensor_parameter_285: name = "parameter_285" + original_name = "linear_4330.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.152074") @@ -3146,6 +3432,7 @@ class Program_weight_tensor_parameter_285: class Program_weight_tensor_parameter_286: name = "parameter_286" + original_name = "linear_4329.b_0" shape = [1024] dtype = "float32" min_val = float("-4.23754") @@ -3157,6 +3444,7 @@ class Program_weight_tensor_parameter_286: class Program_weight_tensor_parameter_287: name = "parameter_287" + original_name = "linear_4329.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.297146") @@ -3168,6 +3456,7 @@ class Program_weight_tensor_parameter_287: class Program_weight_tensor_parameter_288: name = "parameter_288" + original_name = "linear_4328.b_0" shape = [1024] dtype = "float32" min_val = float("-0.17882") @@ -3179,6 +3468,7 @@ class Program_weight_tensor_parameter_288: class Program_weight_tensor_parameter_289: name = "parameter_289" + original_name = "linear_4328.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.281859") @@ -3190,6 +3480,7 @@ class Program_weight_tensor_parameter_289: class Program_weight_tensor_parameter_290: name = "parameter_290" + original_name = "layer_norm_1492.b_0" shape = [1024] dtype = "float32" min_val = float("-0.650817") @@ -3201,6 +3492,7 @@ class Program_weight_tensor_parameter_290: class Program_weight_tensor_parameter_291: name = "parameter_291" + original_name = "layer_norm_1492.w_0" shape = [1024] dtype = "float32" min_val = float("0.32314") @@ -3212,6 +3504,7 @@ class Program_weight_tensor_parameter_291: class Program_weight_tensor_parameter_292: name = "parameter_292" + original_name = "layer_norm_1491.b_0" shape = [1024] dtype = "float32" min_val = float("-1.15237") @@ -3223,6 +3516,7 @@ class Program_weight_tensor_parameter_292: class Program_weight_tensor_parameter_293: name = "parameter_293" + original_name = "layer_norm_1491.w_0" shape = [1024] dtype = "float32" min_val = float("0.83003") @@ -3234,6 +3528,7 @@ class Program_weight_tensor_parameter_293: class Program_weight_tensor_parameter_294: name = "parameter_294" + original_name = "linear_4327.b_0" shape = [1024] dtype = "float32" min_val = float("-0.359642") @@ -3245,6 +3540,7 @@ class Program_weight_tensor_parameter_294: class Program_weight_tensor_parameter_295: name = "parameter_295" + original_name = "linear_4327.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-3.27681") @@ -3256,6 +3552,7 @@ class Program_weight_tensor_parameter_295: class Program_weight_tensor_parameter_296: name = "parameter_296" + original_name = "linear_4326.b_0" shape = [4096] dtype = "float32" min_val = float("-0.0784912") @@ -3267,6 +3564,7 @@ class Program_weight_tensor_parameter_296: class Program_weight_tensor_parameter_297: name = "parameter_297" + original_name = "linear_4326.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.22014") @@ -3278,6 +3576,7 @@ class Program_weight_tensor_parameter_297: class Program_weight_tensor_parameter_298: name = "parameter_298" + original_name = "linear_4325.b_0" shape = [1024] dtype = "float32" min_val = float("-0.269479") @@ -3289,6 +3588,7 @@ class Program_weight_tensor_parameter_298: class Program_weight_tensor_parameter_299: name = "parameter_299" + original_name = "linear_4325.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.547823") @@ -3300,6 +3600,7 @@ class Program_weight_tensor_parameter_299: class Program_weight_tensor_parameter_300: name = "parameter_300" + original_name = "linear_4324.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0370498") @@ -3311,6 +3612,7 @@ class Program_weight_tensor_parameter_300: class Program_weight_tensor_parameter_301: name = "parameter_301" + original_name = "linear_4324.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.133564") @@ -3322,6 +3624,7 @@ class Program_weight_tensor_parameter_301: class Program_weight_tensor_parameter_302: name = "parameter_302" + original_name = "linear_4323.b_0" shape = [1024] dtype = "float32" min_val = float("-4.74029") @@ -3333,6 +3636,7 @@ class Program_weight_tensor_parameter_302: class Program_weight_tensor_parameter_303: name = "parameter_303" + original_name = "linear_4323.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.267074") @@ -3344,6 +3648,7 @@ class Program_weight_tensor_parameter_303: class Program_weight_tensor_parameter_304: name = "parameter_304" + original_name = "linear_4322.b_0" shape = [1024] dtype = "float32" min_val = float("-0.17035") @@ -3355,6 +3660,7 @@ class Program_weight_tensor_parameter_304: class Program_weight_tensor_parameter_305: name = "parameter_305" + original_name = "linear_4322.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.335498") @@ -3366,6 +3672,7 @@ class Program_weight_tensor_parameter_305: class Program_weight_tensor_parameter_306: name = "parameter_306" + original_name = "layer_norm_1490.b_0" shape = [1024] dtype = "float32" min_val = float("-0.525165") @@ -3377,6 +3684,7 @@ class Program_weight_tensor_parameter_306: class Program_weight_tensor_parameter_307: name = "parameter_307" + original_name = "layer_norm_1490.w_0" shape = [1024] dtype = "float32" min_val = float("0.298702") @@ -3388,6 +3696,7 @@ class Program_weight_tensor_parameter_307: class Program_weight_tensor_parameter_308: name = "parameter_308" + original_name = "layer_norm_1489.b_0" shape = [1024] dtype = "float32" min_val = float("-1.33929") @@ -3399,6 +3708,7 @@ class Program_weight_tensor_parameter_308: class Program_weight_tensor_parameter_309: name = "parameter_309" + original_name = "layer_norm_1489.w_0" shape = [1024] dtype = "float32" min_val = float("0.812179") @@ -3410,6 +3720,7 @@ class Program_weight_tensor_parameter_309: class Program_weight_tensor_parameter_310: name = "parameter_310" + original_name = "linear_4321.b_0" shape = [1024] dtype = "float32" min_val = float("-0.403276") @@ -3421,6 +3732,7 @@ class Program_weight_tensor_parameter_310: class Program_weight_tensor_parameter_311: name = "parameter_311" + original_name = "linear_4321.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-3.64969") @@ -3432,6 +3744,7 @@ class Program_weight_tensor_parameter_311: class Program_weight_tensor_parameter_312: name = "parameter_312" + original_name = "linear_4320.b_0" shape = [4096] dtype = "float32" min_val = float("-0.0931396") @@ -3443,6 +3756,7 @@ class Program_weight_tensor_parameter_312: class Program_weight_tensor_parameter_313: name = "parameter_313" + original_name = "linear_4320.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.247076") @@ -3454,6 +3768,7 @@ class Program_weight_tensor_parameter_313: class Program_weight_tensor_parameter_314: name = "parameter_314" + original_name = "linear_4319.b_0" shape = [1024] dtype = "float32" min_val = float("-0.169647") @@ -3465,6 +3780,7 @@ class Program_weight_tensor_parameter_314: class Program_weight_tensor_parameter_315: name = "parameter_315" + original_name = "linear_4319.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.519364") @@ -3476,6 +3792,7 @@ class Program_weight_tensor_parameter_315: class Program_weight_tensor_parameter_316: name = "parameter_316" + original_name = "linear_4318.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0578075") @@ -3487,6 +3804,7 @@ class Program_weight_tensor_parameter_316: class Program_weight_tensor_parameter_317: name = "parameter_317" + original_name = "linear_4318.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.137785") @@ -3498,6 +3816,7 @@ class Program_weight_tensor_parameter_317: class Program_weight_tensor_parameter_318: name = "parameter_318" + original_name = "linear_4317.b_0" shape = [1024] dtype = "float32" min_val = float("-4.79205") @@ -3509,6 +3828,7 @@ class Program_weight_tensor_parameter_318: class Program_weight_tensor_parameter_319: name = "parameter_319" + original_name = "linear_4317.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.285018") @@ -3520,6 +3840,7 @@ class Program_weight_tensor_parameter_319: class Program_weight_tensor_parameter_320: name = "parameter_320" + original_name = "linear_4316.b_0" shape = [1024] dtype = "float32" min_val = float("-0.224814") @@ -3531,6 +3852,7 @@ class Program_weight_tensor_parameter_320: class Program_weight_tensor_parameter_321: name = "parameter_321" + original_name = "linear_4316.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.230074") @@ -3542,6 +3864,7 @@ class Program_weight_tensor_parameter_321: class Program_weight_tensor_parameter_322: name = "parameter_322" + original_name = "layer_norm_1488.b_0" shape = [1024] dtype = "float32" min_val = float("-0.710258") @@ -3553,6 +3876,7 @@ class Program_weight_tensor_parameter_322: class Program_weight_tensor_parameter_323: name = "parameter_323" + original_name = "layer_norm_1488.w_0" shape = [1024] dtype = "float32" min_val = float("0.503958") @@ -3564,6 +3888,7 @@ class Program_weight_tensor_parameter_323: class Program_weight_tensor_parameter_324: name = "parameter_324" + original_name = "layer_norm_1487.b_0" shape = [1024] dtype = "float32" min_val = float("-1.35737") @@ -3575,6 +3900,7 @@ class Program_weight_tensor_parameter_324: class Program_weight_tensor_parameter_325: name = "parameter_325" + original_name = "layer_norm_1487.w_0" shape = [1024] dtype = "float32" min_val = float("0.84206") @@ -3586,6 +3912,7 @@ class Program_weight_tensor_parameter_325: class Program_weight_tensor_parameter_326: name = "parameter_326" + original_name = "linear_4315.b_0" shape = [1024] dtype = "float32" min_val = float("-0.562762") @@ -3597,6 +3924,7 @@ class Program_weight_tensor_parameter_326: class Program_weight_tensor_parameter_327: name = "parameter_327" + original_name = "linear_4315.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-2.91224") @@ -3608,6 +3936,7 @@ class Program_weight_tensor_parameter_327: class Program_weight_tensor_parameter_328: name = "parameter_328" + original_name = "linear_4314.b_0" shape = [4096] dtype = "float32" min_val = float("-0.0960693") @@ -3619,6 +3948,7 @@ class Program_weight_tensor_parameter_328: class Program_weight_tensor_parameter_329: name = "parameter_329" + original_name = "linear_4314.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.242379") @@ -3630,6 +3960,7 @@ class Program_weight_tensor_parameter_329: class Program_weight_tensor_parameter_330: name = "parameter_330" + original_name = "linear_4313.b_0" shape = [1024] dtype = "float32" min_val = float("-0.231323") @@ -3641,6 +3972,7 @@ class Program_weight_tensor_parameter_330: class Program_weight_tensor_parameter_331: name = "parameter_331" + original_name = "linear_4313.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.242201") @@ -3652,6 +3984,7 @@ class Program_weight_tensor_parameter_331: class Program_weight_tensor_parameter_332: name = "parameter_332" + original_name = "linear_4312.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0467409") @@ -3663,6 +3996,7 @@ class Program_weight_tensor_parameter_332: class Program_weight_tensor_parameter_333: name = "parameter_333" + original_name = "linear_4312.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.165015") @@ -3674,6 +4008,7 @@ class Program_weight_tensor_parameter_333: class Program_weight_tensor_parameter_334: name = "parameter_334" + original_name = "linear_4311.b_0" shape = [1024] dtype = "float32" min_val = float("-4.27284") @@ -3685,6 +4020,7 @@ class Program_weight_tensor_parameter_334: class Program_weight_tensor_parameter_335: name = "parameter_335" + original_name = "linear_4311.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.360685") @@ -3696,6 +4032,7 @@ class Program_weight_tensor_parameter_335: class Program_weight_tensor_parameter_336: name = "parameter_336" + original_name = "linear_4310.b_0" shape = [1024] dtype = "float32" min_val = float("-0.190094") @@ -3707,6 +4044,7 @@ class Program_weight_tensor_parameter_336: class Program_weight_tensor_parameter_337: name = "parameter_337" + original_name = "linear_4310.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.246565") @@ -3718,6 +4056,7 @@ class Program_weight_tensor_parameter_337: class Program_weight_tensor_parameter_338: name = "parameter_338" + original_name = "layer_norm_1486.b_0" shape = [1024] dtype = "float32" min_val = float("-0.867411") @@ -3729,6 +4068,7 @@ class Program_weight_tensor_parameter_338: class Program_weight_tensor_parameter_339: name = "parameter_339" + original_name = "layer_norm_1486.w_0" shape = [1024] dtype = "float32" min_val = float("0.58638") @@ -3740,6 +4080,7 @@ class Program_weight_tensor_parameter_339: class Program_weight_tensor_parameter_340: name = "parameter_340" + original_name = "layer_norm_1485.b_0" shape = [1024] dtype = "float32" min_val = float("-1.44422") @@ -3751,6 +4092,7 @@ class Program_weight_tensor_parameter_340: class Program_weight_tensor_parameter_341: name = "parameter_341" + original_name = "layer_norm_1485.w_0" shape = [1024] dtype = "float32" min_val = float("0.850759") @@ -3762,6 +4104,7 @@ class Program_weight_tensor_parameter_341: class Program_weight_tensor_parameter_342: name = "parameter_342" + original_name = "linear_4309.b_0" shape = [1024] dtype = "float32" min_val = float("-0.597662") @@ -3773,6 +4116,7 @@ class Program_weight_tensor_parameter_342: class Program_weight_tensor_parameter_343: name = "parameter_343" + original_name = "linear_4309.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-2.18835") @@ -3784,6 +4128,7 @@ class Program_weight_tensor_parameter_343: class Program_weight_tensor_parameter_344: name = "parameter_344" + original_name = "linear_4308.b_0" shape = [4096] dtype = "float32" min_val = float("-0.0910159") @@ -3795,6 +4140,7 @@ class Program_weight_tensor_parameter_344: class Program_weight_tensor_parameter_345: name = "parameter_345" + original_name = "linear_4308.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.222022") @@ -3806,6 +4152,7 @@ class Program_weight_tensor_parameter_345: class Program_weight_tensor_parameter_346: name = "parameter_346" + original_name = "linear_4307.b_0" shape = [1024] dtype = "float32" min_val = float("-0.19449") @@ -3817,6 +4164,7 @@ class Program_weight_tensor_parameter_346: class Program_weight_tensor_parameter_347: name = "parameter_347" + original_name = "linear_4307.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.272138") @@ -3828,6 +4176,7 @@ class Program_weight_tensor_parameter_347: class Program_weight_tensor_parameter_348: name = "parameter_348" + original_name = "linear_4306.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0607778") @@ -3839,6 +4188,7 @@ class Program_weight_tensor_parameter_348: class Program_weight_tensor_parameter_349: name = "parameter_349" + original_name = "linear_4306.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.130626") @@ -3850,6 +4200,7 @@ class Program_weight_tensor_parameter_349: class Program_weight_tensor_parameter_350: name = "parameter_350" + original_name = "linear_4305.b_0" shape = [1024] dtype = "float32" min_val = float("-3.96813") @@ -3861,6 +4212,7 @@ class Program_weight_tensor_parameter_350: class Program_weight_tensor_parameter_351: name = "parameter_351" + original_name = "linear_4305.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.35358") @@ -3872,6 +4224,7 @@ class Program_weight_tensor_parameter_351: class Program_weight_tensor_parameter_352: name = "parameter_352" + original_name = "linear_4304.b_0" shape = [1024] dtype = "float32" min_val = float("-0.184689") @@ -3883,6 +4236,7 @@ class Program_weight_tensor_parameter_352: class Program_weight_tensor_parameter_353: name = "parameter_353" + original_name = "linear_4304.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.246268") @@ -3894,6 +4248,7 @@ class Program_weight_tensor_parameter_353: class Program_weight_tensor_parameter_354: name = "parameter_354" + original_name = "layer_norm_1484.b_0" shape = [1024] dtype = "float32" min_val = float("-0.943756") @@ -3905,6 +4260,7 @@ class Program_weight_tensor_parameter_354: class Program_weight_tensor_parameter_355: name = "parameter_355" + original_name = "layer_norm_1484.w_0" shape = [1024] dtype = "float32" min_val = float("0.562619") @@ -3916,6 +4272,7 @@ class Program_weight_tensor_parameter_355: class Program_weight_tensor_parameter_356: name = "parameter_356" + original_name = "layer_norm_1483.b_0" shape = [1024] dtype = "float32" min_val = float("-1.41952") @@ -3927,6 +4284,7 @@ class Program_weight_tensor_parameter_356: class Program_weight_tensor_parameter_357: name = "parameter_357" + original_name = "layer_norm_1483.w_0" shape = [1024] dtype = "float32" min_val = float("0.840018") @@ -3938,6 +4296,7 @@ class Program_weight_tensor_parameter_357: class Program_weight_tensor_parameter_358: name = "parameter_358" + original_name = "linear_4303.b_0" shape = [1024] dtype = "float32" min_val = float("-0.577942") @@ -3949,6 +4308,7 @@ class Program_weight_tensor_parameter_358: class Program_weight_tensor_parameter_359: name = "parameter_359" + original_name = "linear_4303.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-1.71249") @@ -3960,6 +4320,7 @@ class Program_weight_tensor_parameter_359: class Program_weight_tensor_parameter_360: name = "parameter_360" + original_name = "linear_4302.b_0" shape = [4096] dtype = "float32" min_val = float("-0.0953369") @@ -3971,6 +4332,7 @@ class Program_weight_tensor_parameter_360: class Program_weight_tensor_parameter_361: name = "parameter_361" + original_name = "linear_4302.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.207642") @@ -3982,6 +4344,7 @@ class Program_weight_tensor_parameter_361: class Program_weight_tensor_parameter_362: name = "parameter_362" + original_name = "linear_4301.b_0" shape = [1024] dtype = "float32" min_val = float("-0.224879") @@ -3993,6 +4356,7 @@ class Program_weight_tensor_parameter_362: class Program_weight_tensor_parameter_363: name = "parameter_363" + original_name = "linear_4301.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.898207") @@ -4004,6 +4368,7 @@ class Program_weight_tensor_parameter_363: class Program_weight_tensor_parameter_364: name = "parameter_364" + original_name = "linear_4300.b_0" shape = [1024] dtype = "float32" min_val = float("-0.114147") @@ -4015,6 +4380,7 @@ class Program_weight_tensor_parameter_364: class Program_weight_tensor_parameter_365: name = "parameter_365" + original_name = "linear_4300.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.132032") @@ -4026,6 +4392,7 @@ class Program_weight_tensor_parameter_365: class Program_weight_tensor_parameter_366: name = "parameter_366" + original_name = "linear_4299.b_0" shape = [1024] dtype = "float32" min_val = float("-3.97696") @@ -4037,6 +4404,7 @@ class Program_weight_tensor_parameter_366: class Program_weight_tensor_parameter_367: name = "parameter_367" + original_name = "linear_4299.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.259797") @@ -4048,6 +4416,7 @@ class Program_weight_tensor_parameter_367: class Program_weight_tensor_parameter_368: name = "parameter_368" + original_name = "linear_4298.b_0" shape = [1024] dtype = "float32" min_val = float("-0.430703") @@ -4059,6 +4428,7 @@ class Program_weight_tensor_parameter_368: class Program_weight_tensor_parameter_369: name = "parameter_369" + original_name = "linear_4298.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.296863") @@ -4070,6 +4440,7 @@ class Program_weight_tensor_parameter_369: class Program_weight_tensor_parameter_370: name = "parameter_370" + original_name = "layer_norm_1482.b_0" shape = [1024] dtype = "float32" min_val = float("-0.909682") @@ -4081,6 +4452,7 @@ class Program_weight_tensor_parameter_370: class Program_weight_tensor_parameter_371: name = "parameter_371" + original_name = "layer_norm_1482.w_0" shape = [1024] dtype = "float32" min_val = float("0.332071") @@ -4092,6 +4464,7 @@ class Program_weight_tensor_parameter_371: class Program_weight_tensor_parameter_372: name = "parameter_372" + original_name = "layer_norm_1481.b_0" shape = [1024] dtype = "float32" min_val = float("-1.91985") @@ -4103,6 +4476,7 @@ class Program_weight_tensor_parameter_372: class Program_weight_tensor_parameter_373: name = "parameter_373" + original_name = "layer_norm_1481.w_0" shape = [1024] dtype = "float32" min_val = float("0.86443") @@ -4114,6 +4488,7 @@ class Program_weight_tensor_parameter_373: class Program_weight_tensor_parameter_374: name = "parameter_374" + original_name = "linear_4297.b_0" shape = [1024] dtype = "float32" min_val = float("-0.541737") @@ -4125,6 +4500,7 @@ class Program_weight_tensor_parameter_374: class Program_weight_tensor_parameter_375: name = "parameter_375" + original_name = "linear_4297.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-1.5876") @@ -4136,6 +4512,7 @@ class Program_weight_tensor_parameter_375: class Program_weight_tensor_parameter_376: name = "parameter_376" + original_name = "linear_4296.b_0" shape = [4096] dtype = "float32" min_val = float("-0.248739") @@ -4147,6 +4524,7 @@ class Program_weight_tensor_parameter_376: class Program_weight_tensor_parameter_377: name = "parameter_377" + original_name = "linear_4296.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.284186") @@ -4158,6 +4536,7 @@ class Program_weight_tensor_parameter_377: class Program_weight_tensor_parameter_378: name = "parameter_378" + original_name = "linear_4295.b_0" shape = [1024] dtype = "float32" min_val = float("-0.443254") @@ -4169,6 +4548,7 @@ class Program_weight_tensor_parameter_378: class Program_weight_tensor_parameter_379: name = "parameter_379" + original_name = "linear_4295.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.868548") @@ -4180,6 +4560,7 @@ class Program_weight_tensor_parameter_379: class Program_weight_tensor_parameter_380: name = "parameter_380" + original_name = "linear_4294.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0876111") @@ -4191,6 +4572,7 @@ class Program_weight_tensor_parameter_380: class Program_weight_tensor_parameter_381: name = "parameter_381" + original_name = "linear_4294.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.136108") @@ -4202,6 +4584,7 @@ class Program_weight_tensor_parameter_381: class Program_weight_tensor_parameter_382: name = "parameter_382" + original_name = "linear_4293.b_0" shape = [1024] dtype = "float32" min_val = float("-1.7628") @@ -4213,6 +4596,7 @@ class Program_weight_tensor_parameter_382: class Program_weight_tensor_parameter_383: name = "parameter_383" + original_name = "linear_4293.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.175664") @@ -4224,6 +4608,7 @@ class Program_weight_tensor_parameter_383: class Program_weight_tensor_parameter_384: name = "parameter_384" + original_name = "linear_4292.b_0" shape = [1024] dtype = "float32" min_val = float("-0.437152") @@ -4235,6 +4620,7 @@ class Program_weight_tensor_parameter_384: class Program_weight_tensor_parameter_385: name = "parameter_385" + original_name = "linear_4292.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.195578") @@ -4246,6 +4632,7 @@ class Program_weight_tensor_parameter_385: class Program_weight_tensor_parameter_386: name = "parameter_386" + original_name = "layer_norm_1480.b_0" shape = [1024] dtype = "float32" min_val = float("-0.676423") @@ -4257,6 +4644,7 @@ class Program_weight_tensor_parameter_386: class Program_weight_tensor_parameter_387: name = "parameter_387" + original_name = "layer_norm_1480.w_0" shape = [1024] dtype = "float32" min_val = float("0.500783") @@ -4268,6 +4656,7 @@ class Program_weight_tensor_parameter_387: class Program_weight_tensor_parameter_388: name = "parameter_388" + original_name = "embedding_261.w_0" shape = [4, 1024] dtype = "float32" min_val = float("-0.0571652") @@ -4279,6 +4668,7 @@ class Program_weight_tensor_parameter_388: class Program_weight_tensor_parameter_389: name = "parameter_389" + original_name = "embedding_260.w_0" shape = [512, 1024] dtype = "float32" min_val = float("-0.766039") @@ -4290,6 +4680,7 @@ class Program_weight_tensor_parameter_389: class Program_weight_tensor_parameter_390: name = "parameter_390" + original_name = "embedding_259.w_0" shape = [12800, 1024] dtype = "float32" min_val = float("-0.816515") diff --git a/paddle_samples/PaddleNLP/skep_ernie_2.0_large_en/input_meta.py b/paddle_samples/PaddleNLP/skep_ernie_2.0_large_en/input_meta.py index fd1f7db6f..060f275ed 100644 --- a/paddle_samples/PaddleNLP/skep_ernie_2.0_large_en/input_meta.py +++ b/paddle_samples/PaddleNLP/skep_ernie_2.0_large_en/input_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_data_0: name = "data_0" + original_name = "input_ids" shape = [1, 21] dtype = "int64" data = [ @@ -29,6 +30,7 @@ class Program_weight_tensor_data_0: class Program_weight_tensor_data_1: name = "data_1" + original_name = "token_type_ids" shape = [1, 21] dtype = "int64" data = [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0] diff --git a/paddle_samples/PaddleNLP/skep_ernie_2.0_large_en/model.py b/paddle_samples/PaddleNLP/skep_ernie_2.0_large_en/model.py index f52602daf..2fa6e8082 100644 --- a/paddle_samples/PaddleNLP/skep_ernie_2.0_large_en/model.py +++ b/paddle_samples/PaddleNLP/skep_ernie_2.0_large_en/model.py @@ -457,10 +457,11 @@ def forward( # pd_op.embedding: (1x21x1024xf32) <- (1x21xi64, 512x1024xf32) embedding_1 = paddle._C_ops.embedding(subtract_0, parameter_389, -1, False) - del parameter_389 + del parameter_389, subtract_0 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_0 = paddle._C_ops.add(embedding_0, embedding_1) + del embedding_0, embedding_1 # pd_op.embedding: (1x21x1024xf32) <- (1x21xi64, 4x1024xf32) embedding_2 = paddle._C_ops.embedding(data_1, parameter_388, -1, False) @@ -468,6 +469,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_1 = paddle._C_ops.add(add_0, embedding_2) + del add_0, embedding_2 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_0, layer_norm_1, layer_norm_2 = (lambda x, f: f(x))( @@ -476,233 +478,17 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_386, parameter_387 + del add_1, parameter_386, parameter_387 # pd_op.full: (1xf32) <- () full_4 = paddle._C_ops.full( [1], float("0.1"), paddle.float32, paddle.core.CPUPlace() ) - # pd_op.assign: (1xf32) <- (1xf32) - assign_0 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_1 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_2 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_3 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_4 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_5 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_6 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_7 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_8 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_9 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_10 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_11 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_12 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_13 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_14 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_15 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_16 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_17 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_18 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_19 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_20 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_21 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_22 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_23 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_24 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_25 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_26 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_27 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_28 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_29 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_30 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_31 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_32 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_33 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_34 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_35 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_36 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_37 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_38 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_39 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_40 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_41 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_42 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_43 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_44 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_45 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_46 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_47 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_48 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_49 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_50 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_51 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_52 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_53 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_54 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_55 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_56 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_57 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_58 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_59 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_60 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_61 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_62 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_63 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_64 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_65 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_66 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_67 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_68 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_69 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_70 = full_4 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_71 = full_4 - # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_0, dropout_1 = (lambda x, f: f(x))( paddle._C_ops.dropout( - layer_norm_0, None, full_4, False, "upscale_in_train", 0, False + layer_norm_0, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -714,13 +500,14 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_2 = paddle._C_ops.add(matmul_0, parameter_384) - del parameter_384 + del matmul_0, parameter_384 # pd_op.full_int_array: (4xi64) <- () full_int_array_1 = [0, 0, 16, 64] # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_0 = paddle._C_ops.reshape(add_2, full_int_array_1) + del add_2 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_0 = paddle._C_ops.transpose(reshape_0, [0, 2, 1, 3]) @@ -732,7 +519,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_3 = paddle._C_ops.add(matmul_1, parameter_382) - del parameter_382 + del matmul_1, parameter_382 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_2 = paddle._C_ops.matmul(dropout_0, parameter_381, False, False) @@ -740,10 +527,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_4 = paddle._C_ops.add(matmul_2, parameter_380) - del parameter_380 + del matmul_2, parameter_380 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_1 = paddle._C_ops.reshape(add_3, full_int_array_1) + del add_3 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_1 = paddle._C_ops.transpose(reshape_1, [0, 2, 1, 3]) @@ -751,6 +539,7 @@ def forward( # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_2 = paddle._C_ops.reshape(add_4, full_int_array_1) + del add_4 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_2 = paddle._C_ops.transpose(reshape_2, [0, 2, 1, 3]) @@ -761,84 +550,17 @@ def forward( [1], float("0.125"), paddle.float32, paddle.core.CPUPlace() ) - # pd_op.assign: (1xf32) <- (1xf32) - assign_72 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_73 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_74 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_75 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_76 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_77 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_78 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_79 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_80 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_81 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_82 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_83 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_84 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_85 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_86 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_87 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_88 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_89 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_90 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_91 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_92 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_93 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_94 = full_5 - # pd_op.scale: (1x16x21x64xf32) <- (1x16x21x64xf32, 1xf32) scale_1 = paddle._C_ops.scale(transpose_0, full_5, float("0"), True) del transpose_0 # pd_op.matmul: (1x16x21x21xf32) <- (1x16x21x64xf32, 1x16x21x64xf32) matmul_3 = paddle._C_ops.matmul(scale_1, transpose_1, False, True) + del scale_1, transpose_1 # pd_op.add: (1x16x21x21xf32) <- (1x16x21x21xf32, 1x1x1x21xf32) add_5 = paddle._C_ops.add(matmul_3, unsqueeze_0) + del matmul_3 # pd_op.softmax: (1x16x21x21xf32) <- (1x16x21x21xf32) softmax_0 = paddle._C_ops.softmax(add_5, -1) @@ -847,13 +569,15 @@ def forward( # pd_op.dropout: (1x16x21x21xf32, 1x16x21x21xui8) <- (1x16x21x21xf32, None, 1xf32) dropout_2, dropout_3 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_0, None, full_4, False, "upscale_in_train", 0, False + softmax_0, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_0 # pd_op.matmul: (1x16x21x64xf32) <- (1x16x21x21xf32, 1x16x21x64xf32) matmul_4 = paddle._C_ops.matmul(dropout_2, transpose_2, False, False) + del dropout_2, transpose_2 # pd_op.transpose: (1x21x16x64xf32) <- (1x16x21x64xf32) transpose_3 = paddle._C_ops.transpose(matmul_4, [0, 2, 1, 3]) @@ -864,19 +588,20 @@ def forward( # pd_op.reshape: (1x21x1024xf32) <- (1x21x16x64xf32, 3xi64) reshape_3 = paddle._C_ops.reshape(transpose_3, full_int_array_2) + del transpose_3 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_5 = paddle._C_ops.matmul(reshape_3, parameter_379, False, False) - del parameter_379 + del parameter_379, reshape_3 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_6 = paddle._C_ops.add(matmul_5, parameter_378) - del parameter_378 + del matmul_5, parameter_378 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_4, dropout_5 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_6, None, full_4, False, "upscale_in_train", 0, False + add_6, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -884,6 +609,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_7 = paddle._C_ops.add(dropout_0, dropout_4) + del dropout_0, dropout_4 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_3, layer_norm_4, layer_norm_5 = (lambda x, f: f(x))( @@ -892,7 +618,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_372, parameter_373 + del add_7, parameter_372, parameter_373 # pd_op.matmul: (1x21x4096xf32) <- (1x21x1024xf32, 1024x4096xf32) matmul_6 = paddle._C_ops.matmul(layer_norm_3, parameter_377, False, False) @@ -900,23 +626,24 @@ def forward( # pd_op.add: (1x21x4096xf32) <- (1x21x4096xf32, 4096xf32) add_8 = paddle._C_ops.add(matmul_6, parameter_376) - del parameter_376 + del matmul_6, parameter_376 # pd_op.gelu: (1x21x4096xf32) <- (1x21x4096xf32) gelu_0 = paddle._C_ops.gelu(add_8, False) + del add_8 # pd_op.matmul: (1x21x1024xf32) <- (1x21x4096xf32, 4096x1024xf32) matmul_7 = paddle._C_ops.matmul(gelu_0, parameter_375, False, False) - del parameter_375 + del gelu_0, parameter_375 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_9 = paddle._C_ops.add(matmul_7, parameter_374) - del parameter_374 + del matmul_7, parameter_374 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_6, dropout_7 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_9, None, full_4, False, "upscale_in_train", 0, False + add_9, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -924,6 +651,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_10 = paddle._C_ops.add(layer_norm_3, dropout_6) + del dropout_6, layer_norm_3 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_6, layer_norm_7, layer_norm_8 = (lambda x, f: f(x))( @@ -932,7 +660,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_370, parameter_371 + del add_10, parameter_370, parameter_371 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_8 = paddle._C_ops.matmul(layer_norm_6, parameter_369, False, False) @@ -940,10 +668,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_11 = paddle._C_ops.add(matmul_8, parameter_368) - del parameter_368 + del matmul_8, parameter_368 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_4 = paddle._C_ops.reshape(add_11, full_int_array_1) + del add_11 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_4 = paddle._C_ops.transpose(reshape_4, [0, 2, 1, 3]) @@ -955,7 +684,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_12 = paddle._C_ops.add(matmul_9, parameter_366) - del parameter_366 + del matmul_9, parameter_366 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_10 = paddle._C_ops.matmul(layer_norm_6, parameter_365, False, False) @@ -963,10 +692,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_13 = paddle._C_ops.add(matmul_10, parameter_364) - del parameter_364 + del matmul_10, parameter_364 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_5 = paddle._C_ops.reshape(add_12, full_int_array_1) + del add_12 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_5 = paddle._C_ops.transpose(reshape_5, [0, 2, 1, 3]) @@ -974,6 +704,7 @@ def forward( # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_6 = paddle._C_ops.reshape(add_13, full_int_array_1) + del add_13 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_6 = paddle._C_ops.transpose(reshape_6, [0, 2, 1, 3]) @@ -985,9 +716,11 @@ def forward( # pd_op.matmul: (1x16x21x21xf32) <- (1x16x21x64xf32, 1x16x21x64xf32) matmul_11 = paddle._C_ops.matmul(scale_2, transpose_5, False, True) + del scale_2, transpose_5 # pd_op.add: (1x16x21x21xf32) <- (1x16x21x21xf32, 1x1x1x21xf32) add_14 = paddle._C_ops.add(matmul_11, unsqueeze_0) + del matmul_11 # pd_op.softmax: (1x16x21x21xf32) <- (1x16x21x21xf32) softmax_1 = paddle._C_ops.softmax(add_14, -1) @@ -996,13 +729,15 @@ def forward( # pd_op.dropout: (1x16x21x21xf32, 1x16x21x21xui8) <- (1x16x21x21xf32, None, 1xf32) dropout_8, dropout_9 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_1, None, full_4, False, "upscale_in_train", 0, False + softmax_1, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_1 # pd_op.matmul: (1x16x21x64xf32) <- (1x16x21x21xf32, 1x16x21x64xf32) matmul_12 = paddle._C_ops.matmul(dropout_8, transpose_6, False, False) + del dropout_8, transpose_6 # pd_op.transpose: (1x21x16x64xf32) <- (1x16x21x64xf32) transpose_7 = paddle._C_ops.transpose(matmul_12, [0, 2, 1, 3]) @@ -1010,19 +745,20 @@ def forward( # pd_op.reshape: (1x21x1024xf32) <- (1x21x16x64xf32, 3xi64) reshape_7 = paddle._C_ops.reshape(transpose_7, full_int_array_2) + del transpose_7 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_13 = paddle._C_ops.matmul(reshape_7, parameter_363, False, False) - del parameter_363 + del parameter_363, reshape_7 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_15 = paddle._C_ops.add(matmul_13, parameter_362) - del parameter_362 + del matmul_13, parameter_362 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_10, dropout_11 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_15, None, full_4, False, "upscale_in_train", 0, False + add_15, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1030,6 +766,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_16 = paddle._C_ops.add(layer_norm_6, dropout_10) + del dropout_10, layer_norm_6 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_9, layer_norm_10, layer_norm_11 = (lambda x, f: f(x))( @@ -1038,7 +775,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_356, parameter_357 + del add_16, parameter_356, parameter_357 # pd_op.matmul: (1x21x4096xf32) <- (1x21x1024xf32, 1024x4096xf32) matmul_14 = paddle._C_ops.matmul(layer_norm_9, parameter_361, False, False) @@ -1046,23 +783,24 @@ def forward( # pd_op.add: (1x21x4096xf32) <- (1x21x4096xf32, 4096xf32) add_17 = paddle._C_ops.add(matmul_14, parameter_360) - del parameter_360 + del matmul_14, parameter_360 # pd_op.gelu: (1x21x4096xf32) <- (1x21x4096xf32) gelu_1 = paddle._C_ops.gelu(add_17, False) + del add_17 # pd_op.matmul: (1x21x1024xf32) <- (1x21x4096xf32, 4096x1024xf32) matmul_15 = paddle._C_ops.matmul(gelu_1, parameter_359, False, False) - del parameter_359 + del gelu_1, parameter_359 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_18 = paddle._C_ops.add(matmul_15, parameter_358) - del parameter_358 + del matmul_15, parameter_358 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_12, dropout_13 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_18, None, full_4, False, "upscale_in_train", 0, False + add_18, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1070,6 +808,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_19 = paddle._C_ops.add(layer_norm_9, dropout_12) + del dropout_12, layer_norm_9 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_12, layer_norm_13, layer_norm_14 = (lambda x, f: f(x))( @@ -1078,7 +817,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_354, parameter_355 + del add_19, parameter_354, parameter_355 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_16 = paddle._C_ops.matmul(layer_norm_12, parameter_353, False, False) @@ -1086,10 +825,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_20 = paddle._C_ops.add(matmul_16, parameter_352) - del parameter_352 + del matmul_16, parameter_352 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_8 = paddle._C_ops.reshape(add_20, full_int_array_1) + del add_20 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_8 = paddle._C_ops.transpose(reshape_8, [0, 2, 1, 3]) @@ -1101,7 +841,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_21 = paddle._C_ops.add(matmul_17, parameter_350) - del parameter_350 + del matmul_17, parameter_350 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_18 = paddle._C_ops.matmul(layer_norm_12, parameter_349, False, False) @@ -1109,10 +849,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_22 = paddle._C_ops.add(matmul_18, parameter_348) - del parameter_348 + del matmul_18, parameter_348 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_9 = paddle._C_ops.reshape(add_21, full_int_array_1) + del add_21 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_9 = paddle._C_ops.transpose(reshape_9, [0, 2, 1, 3]) @@ -1120,6 +861,7 @@ def forward( # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_10 = paddle._C_ops.reshape(add_22, full_int_array_1) + del add_22 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_10 = paddle._C_ops.transpose(reshape_10, [0, 2, 1, 3]) @@ -1131,9 +873,11 @@ def forward( # pd_op.matmul: (1x16x21x21xf32) <- (1x16x21x64xf32, 1x16x21x64xf32) matmul_19 = paddle._C_ops.matmul(scale_3, transpose_9, False, True) + del scale_3, transpose_9 # pd_op.add: (1x16x21x21xf32) <- (1x16x21x21xf32, 1x1x1x21xf32) add_23 = paddle._C_ops.add(matmul_19, unsqueeze_0) + del matmul_19 # pd_op.softmax: (1x16x21x21xf32) <- (1x16x21x21xf32) softmax_2 = paddle._C_ops.softmax(add_23, -1) @@ -1142,13 +886,15 @@ def forward( # pd_op.dropout: (1x16x21x21xf32, 1x16x21x21xui8) <- (1x16x21x21xf32, None, 1xf32) dropout_14, dropout_15 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_2, None, full_4, False, "upscale_in_train", 0, False + softmax_2, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_2 # pd_op.matmul: (1x16x21x64xf32) <- (1x16x21x21xf32, 1x16x21x64xf32) matmul_20 = paddle._C_ops.matmul(dropout_14, transpose_10, False, False) + del dropout_14, transpose_10 # pd_op.transpose: (1x21x16x64xf32) <- (1x16x21x64xf32) transpose_11 = paddle._C_ops.transpose(matmul_20, [0, 2, 1, 3]) @@ -1156,19 +902,20 @@ def forward( # pd_op.reshape: (1x21x1024xf32) <- (1x21x16x64xf32, 3xi64) reshape_11 = paddle._C_ops.reshape(transpose_11, full_int_array_2) + del transpose_11 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_21 = paddle._C_ops.matmul(reshape_11, parameter_347, False, False) - del parameter_347 + del parameter_347, reshape_11 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_24 = paddle._C_ops.add(matmul_21, parameter_346) - del parameter_346 + del matmul_21, parameter_346 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_16, dropout_17 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_24, None, full_4, False, "upscale_in_train", 0, False + add_24, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1176,6 +923,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_25 = paddle._C_ops.add(layer_norm_12, dropout_16) + del dropout_16, layer_norm_12 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_15, layer_norm_16, layer_norm_17 = (lambda x, f: f(x))( @@ -1184,7 +932,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_340, parameter_341 + del add_25, parameter_340, parameter_341 # pd_op.matmul: (1x21x4096xf32) <- (1x21x1024xf32, 1024x4096xf32) matmul_22 = paddle._C_ops.matmul(layer_norm_15, parameter_345, False, False) @@ -1192,23 +940,24 @@ def forward( # pd_op.add: (1x21x4096xf32) <- (1x21x4096xf32, 4096xf32) add_26 = paddle._C_ops.add(matmul_22, parameter_344) - del parameter_344 + del matmul_22, parameter_344 # pd_op.gelu: (1x21x4096xf32) <- (1x21x4096xf32) gelu_2 = paddle._C_ops.gelu(add_26, False) + del add_26 # pd_op.matmul: (1x21x1024xf32) <- (1x21x4096xf32, 4096x1024xf32) matmul_23 = paddle._C_ops.matmul(gelu_2, parameter_343, False, False) - del parameter_343 + del gelu_2, parameter_343 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_27 = paddle._C_ops.add(matmul_23, parameter_342) - del parameter_342 + del matmul_23, parameter_342 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_18, dropout_19 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_27, None, full_4, False, "upscale_in_train", 0, False + add_27, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1216,6 +965,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_28 = paddle._C_ops.add(layer_norm_15, dropout_18) + del dropout_18, layer_norm_15 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_18, layer_norm_19, layer_norm_20 = (lambda x, f: f(x))( @@ -1224,7 +974,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_338, parameter_339 + del add_28, parameter_338, parameter_339 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_24 = paddle._C_ops.matmul(layer_norm_18, parameter_337, False, False) @@ -1232,10 +982,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_29 = paddle._C_ops.add(matmul_24, parameter_336) - del parameter_336 + del matmul_24, parameter_336 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_12 = paddle._C_ops.reshape(add_29, full_int_array_1) + del add_29 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_12 = paddle._C_ops.transpose(reshape_12, [0, 2, 1, 3]) @@ -1247,7 +998,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_30 = paddle._C_ops.add(matmul_25, parameter_334) - del parameter_334 + del matmul_25, parameter_334 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_26 = paddle._C_ops.matmul(layer_norm_18, parameter_333, False, False) @@ -1255,10 +1006,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_31 = paddle._C_ops.add(matmul_26, parameter_332) - del parameter_332 + del matmul_26, parameter_332 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_13 = paddle._C_ops.reshape(add_30, full_int_array_1) + del add_30 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_13 = paddle._C_ops.transpose(reshape_13, [0, 2, 1, 3]) @@ -1266,6 +1018,7 @@ def forward( # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_14 = paddle._C_ops.reshape(add_31, full_int_array_1) + del add_31 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_14 = paddle._C_ops.transpose(reshape_14, [0, 2, 1, 3]) @@ -1277,9 +1030,11 @@ def forward( # pd_op.matmul: (1x16x21x21xf32) <- (1x16x21x64xf32, 1x16x21x64xf32) matmul_27 = paddle._C_ops.matmul(scale_4, transpose_13, False, True) + del scale_4, transpose_13 # pd_op.add: (1x16x21x21xf32) <- (1x16x21x21xf32, 1x1x1x21xf32) add_32 = paddle._C_ops.add(matmul_27, unsqueeze_0) + del matmul_27 # pd_op.softmax: (1x16x21x21xf32) <- (1x16x21x21xf32) softmax_3 = paddle._C_ops.softmax(add_32, -1) @@ -1288,13 +1043,15 @@ def forward( # pd_op.dropout: (1x16x21x21xf32, 1x16x21x21xui8) <- (1x16x21x21xf32, None, 1xf32) dropout_20, dropout_21 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_3, None, full_4, False, "upscale_in_train", 0, False + softmax_3, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_3 # pd_op.matmul: (1x16x21x64xf32) <- (1x16x21x21xf32, 1x16x21x64xf32) matmul_28 = paddle._C_ops.matmul(dropout_20, transpose_14, False, False) + del dropout_20, transpose_14 # pd_op.transpose: (1x21x16x64xf32) <- (1x16x21x64xf32) transpose_15 = paddle._C_ops.transpose(matmul_28, [0, 2, 1, 3]) @@ -1302,19 +1059,20 @@ def forward( # pd_op.reshape: (1x21x1024xf32) <- (1x21x16x64xf32, 3xi64) reshape_15 = paddle._C_ops.reshape(transpose_15, full_int_array_2) + del transpose_15 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_29 = paddle._C_ops.matmul(reshape_15, parameter_331, False, False) - del parameter_331 + del parameter_331, reshape_15 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_33 = paddle._C_ops.add(matmul_29, parameter_330) - del parameter_330 + del matmul_29, parameter_330 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_22, dropout_23 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_33, None, full_4, False, "upscale_in_train", 0, False + add_33, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1322,6 +1080,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_34 = paddle._C_ops.add(layer_norm_18, dropout_22) + del dropout_22, layer_norm_18 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_21, layer_norm_22, layer_norm_23 = (lambda x, f: f(x))( @@ -1330,7 +1089,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_324, parameter_325 + del add_34, parameter_324, parameter_325 # pd_op.matmul: (1x21x4096xf32) <- (1x21x1024xf32, 1024x4096xf32) matmul_30 = paddle._C_ops.matmul(layer_norm_21, parameter_329, False, False) @@ -1338,23 +1097,24 @@ def forward( # pd_op.add: (1x21x4096xf32) <- (1x21x4096xf32, 4096xf32) add_35 = paddle._C_ops.add(matmul_30, parameter_328) - del parameter_328 + del matmul_30, parameter_328 # pd_op.gelu: (1x21x4096xf32) <- (1x21x4096xf32) gelu_3 = paddle._C_ops.gelu(add_35, False) + del add_35 # pd_op.matmul: (1x21x1024xf32) <- (1x21x4096xf32, 4096x1024xf32) matmul_31 = paddle._C_ops.matmul(gelu_3, parameter_327, False, False) - del parameter_327 + del gelu_3, parameter_327 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_36 = paddle._C_ops.add(matmul_31, parameter_326) - del parameter_326 + del matmul_31, parameter_326 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_24, dropout_25 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_36, None, full_4, False, "upscale_in_train", 0, False + add_36, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1362,6 +1122,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_37 = paddle._C_ops.add(layer_norm_21, dropout_24) + del dropout_24, layer_norm_21 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_24, layer_norm_25, layer_norm_26 = (lambda x, f: f(x))( @@ -1370,7 +1131,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_322, parameter_323 + del add_37, parameter_322, parameter_323 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_32 = paddle._C_ops.matmul(layer_norm_24, parameter_321, False, False) @@ -1378,10 +1139,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_38 = paddle._C_ops.add(matmul_32, parameter_320) - del parameter_320 + del matmul_32, parameter_320 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_16 = paddle._C_ops.reshape(add_38, full_int_array_1) + del add_38 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_16 = paddle._C_ops.transpose(reshape_16, [0, 2, 1, 3]) @@ -1393,7 +1155,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_39 = paddle._C_ops.add(matmul_33, parameter_318) - del parameter_318 + del matmul_33, parameter_318 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_34 = paddle._C_ops.matmul(layer_norm_24, parameter_317, False, False) @@ -1401,10 +1163,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_40 = paddle._C_ops.add(matmul_34, parameter_316) - del parameter_316 + del matmul_34, parameter_316 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_17 = paddle._C_ops.reshape(add_39, full_int_array_1) + del add_39 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_17 = paddle._C_ops.transpose(reshape_17, [0, 2, 1, 3]) @@ -1412,6 +1175,7 @@ def forward( # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_18 = paddle._C_ops.reshape(add_40, full_int_array_1) + del add_40 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_18 = paddle._C_ops.transpose(reshape_18, [0, 2, 1, 3]) @@ -1423,9 +1187,11 @@ def forward( # pd_op.matmul: (1x16x21x21xf32) <- (1x16x21x64xf32, 1x16x21x64xf32) matmul_35 = paddle._C_ops.matmul(scale_5, transpose_17, False, True) + del scale_5, transpose_17 # pd_op.add: (1x16x21x21xf32) <- (1x16x21x21xf32, 1x1x1x21xf32) add_41 = paddle._C_ops.add(matmul_35, unsqueeze_0) + del matmul_35 # pd_op.softmax: (1x16x21x21xf32) <- (1x16x21x21xf32) softmax_4 = paddle._C_ops.softmax(add_41, -1) @@ -1434,13 +1200,15 @@ def forward( # pd_op.dropout: (1x16x21x21xf32, 1x16x21x21xui8) <- (1x16x21x21xf32, None, 1xf32) dropout_26, dropout_27 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_4, None, full_4, False, "upscale_in_train", 0, False + softmax_4, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_4 # pd_op.matmul: (1x16x21x64xf32) <- (1x16x21x21xf32, 1x16x21x64xf32) matmul_36 = paddle._C_ops.matmul(dropout_26, transpose_18, False, False) + del dropout_26, transpose_18 # pd_op.transpose: (1x21x16x64xf32) <- (1x16x21x64xf32) transpose_19 = paddle._C_ops.transpose(matmul_36, [0, 2, 1, 3]) @@ -1448,19 +1216,20 @@ def forward( # pd_op.reshape: (1x21x1024xf32) <- (1x21x16x64xf32, 3xi64) reshape_19 = paddle._C_ops.reshape(transpose_19, full_int_array_2) + del transpose_19 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_37 = paddle._C_ops.matmul(reshape_19, parameter_315, False, False) - del parameter_315 + del parameter_315, reshape_19 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_42 = paddle._C_ops.add(matmul_37, parameter_314) - del parameter_314 + del matmul_37, parameter_314 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_28, dropout_29 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_42, None, full_4, False, "upscale_in_train", 0, False + add_42, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1468,6 +1237,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_43 = paddle._C_ops.add(layer_norm_24, dropout_28) + del dropout_28, layer_norm_24 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_27, layer_norm_28, layer_norm_29 = (lambda x, f: f(x))( @@ -1476,7 +1246,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_308, parameter_309 + del add_43, parameter_308, parameter_309 # pd_op.matmul: (1x21x4096xf32) <- (1x21x1024xf32, 1024x4096xf32) matmul_38 = paddle._C_ops.matmul(layer_norm_27, parameter_313, False, False) @@ -1484,23 +1254,24 @@ def forward( # pd_op.add: (1x21x4096xf32) <- (1x21x4096xf32, 4096xf32) add_44 = paddle._C_ops.add(matmul_38, parameter_312) - del parameter_312 + del matmul_38, parameter_312 # pd_op.gelu: (1x21x4096xf32) <- (1x21x4096xf32) gelu_4 = paddle._C_ops.gelu(add_44, False) + del add_44 # pd_op.matmul: (1x21x1024xf32) <- (1x21x4096xf32, 4096x1024xf32) matmul_39 = paddle._C_ops.matmul(gelu_4, parameter_311, False, False) - del parameter_311 + del gelu_4, parameter_311 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_45 = paddle._C_ops.add(matmul_39, parameter_310) - del parameter_310 + del matmul_39, parameter_310 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_30, dropout_31 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_45, None, full_4, False, "upscale_in_train", 0, False + add_45, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1508,6 +1279,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_46 = paddle._C_ops.add(layer_norm_27, dropout_30) + del dropout_30, layer_norm_27 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_30, layer_norm_31, layer_norm_32 = (lambda x, f: f(x))( @@ -1516,7 +1288,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_306, parameter_307 + del add_46, parameter_306, parameter_307 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_40 = paddle._C_ops.matmul(layer_norm_30, parameter_305, False, False) @@ -1524,10 +1296,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_47 = paddle._C_ops.add(matmul_40, parameter_304) - del parameter_304 + del matmul_40, parameter_304 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_20 = paddle._C_ops.reshape(add_47, full_int_array_1) + del add_47 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_20 = paddle._C_ops.transpose(reshape_20, [0, 2, 1, 3]) @@ -1539,7 +1312,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_48 = paddle._C_ops.add(matmul_41, parameter_302) - del parameter_302 + del matmul_41, parameter_302 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_42 = paddle._C_ops.matmul(layer_norm_30, parameter_301, False, False) @@ -1547,10 +1320,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_49 = paddle._C_ops.add(matmul_42, parameter_300) - del parameter_300 + del matmul_42, parameter_300 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_21 = paddle._C_ops.reshape(add_48, full_int_array_1) + del add_48 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_21 = paddle._C_ops.transpose(reshape_21, [0, 2, 1, 3]) @@ -1558,6 +1332,7 @@ def forward( # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_22 = paddle._C_ops.reshape(add_49, full_int_array_1) + del add_49 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_22 = paddle._C_ops.transpose(reshape_22, [0, 2, 1, 3]) @@ -1569,9 +1344,11 @@ def forward( # pd_op.matmul: (1x16x21x21xf32) <- (1x16x21x64xf32, 1x16x21x64xf32) matmul_43 = paddle._C_ops.matmul(scale_6, transpose_21, False, True) + del scale_6, transpose_21 # pd_op.add: (1x16x21x21xf32) <- (1x16x21x21xf32, 1x1x1x21xf32) add_50 = paddle._C_ops.add(matmul_43, unsqueeze_0) + del matmul_43 # pd_op.softmax: (1x16x21x21xf32) <- (1x16x21x21xf32) softmax_5 = paddle._C_ops.softmax(add_50, -1) @@ -1580,13 +1357,15 @@ def forward( # pd_op.dropout: (1x16x21x21xf32, 1x16x21x21xui8) <- (1x16x21x21xf32, None, 1xf32) dropout_32, dropout_33 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_5, None, full_4, False, "upscale_in_train", 0, False + softmax_5, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_5 # pd_op.matmul: (1x16x21x64xf32) <- (1x16x21x21xf32, 1x16x21x64xf32) matmul_44 = paddle._C_ops.matmul(dropout_32, transpose_22, False, False) + del dropout_32, transpose_22 # pd_op.transpose: (1x21x16x64xf32) <- (1x16x21x64xf32) transpose_23 = paddle._C_ops.transpose(matmul_44, [0, 2, 1, 3]) @@ -1594,19 +1373,20 @@ def forward( # pd_op.reshape: (1x21x1024xf32) <- (1x21x16x64xf32, 3xi64) reshape_23 = paddle._C_ops.reshape(transpose_23, full_int_array_2) + del transpose_23 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_45 = paddle._C_ops.matmul(reshape_23, parameter_299, False, False) - del parameter_299 + del parameter_299, reshape_23 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_51 = paddle._C_ops.add(matmul_45, parameter_298) - del parameter_298 + del matmul_45, parameter_298 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_34, dropout_35 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_51, None, full_4, False, "upscale_in_train", 0, False + add_51, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1614,6 +1394,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_52 = paddle._C_ops.add(layer_norm_30, dropout_34) + del dropout_34, layer_norm_30 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_33, layer_norm_34, layer_norm_35 = (lambda x, f: f(x))( @@ -1622,7 +1403,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_292, parameter_293 + del add_52, parameter_292, parameter_293 # pd_op.matmul: (1x21x4096xf32) <- (1x21x1024xf32, 1024x4096xf32) matmul_46 = paddle._C_ops.matmul(layer_norm_33, parameter_297, False, False) @@ -1630,23 +1411,24 @@ def forward( # pd_op.add: (1x21x4096xf32) <- (1x21x4096xf32, 4096xf32) add_53 = paddle._C_ops.add(matmul_46, parameter_296) - del parameter_296 + del matmul_46, parameter_296 # pd_op.gelu: (1x21x4096xf32) <- (1x21x4096xf32) gelu_5 = paddle._C_ops.gelu(add_53, False) + del add_53 # pd_op.matmul: (1x21x1024xf32) <- (1x21x4096xf32, 4096x1024xf32) matmul_47 = paddle._C_ops.matmul(gelu_5, parameter_295, False, False) - del parameter_295 + del gelu_5, parameter_295 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_54 = paddle._C_ops.add(matmul_47, parameter_294) - del parameter_294 + del matmul_47, parameter_294 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_36, dropout_37 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_54, None, full_4, False, "upscale_in_train", 0, False + add_54, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1654,6 +1436,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_55 = paddle._C_ops.add(layer_norm_33, dropout_36) + del dropout_36, layer_norm_33 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_36, layer_norm_37, layer_norm_38 = (lambda x, f: f(x))( @@ -1662,7 +1445,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_290, parameter_291 + del add_55, parameter_290, parameter_291 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_48 = paddle._C_ops.matmul(layer_norm_36, parameter_289, False, False) @@ -1670,10 +1453,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_56 = paddle._C_ops.add(matmul_48, parameter_288) - del parameter_288 + del matmul_48, parameter_288 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_24 = paddle._C_ops.reshape(add_56, full_int_array_1) + del add_56 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_24 = paddle._C_ops.transpose(reshape_24, [0, 2, 1, 3]) @@ -1685,7 +1469,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_57 = paddle._C_ops.add(matmul_49, parameter_286) - del parameter_286 + del matmul_49, parameter_286 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_50 = paddle._C_ops.matmul(layer_norm_36, parameter_285, False, False) @@ -1693,10 +1477,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_58 = paddle._C_ops.add(matmul_50, parameter_284) - del parameter_284 + del matmul_50, parameter_284 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_25 = paddle._C_ops.reshape(add_57, full_int_array_1) + del add_57 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_25 = paddle._C_ops.transpose(reshape_25, [0, 2, 1, 3]) @@ -1704,6 +1489,7 @@ def forward( # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_26 = paddle._C_ops.reshape(add_58, full_int_array_1) + del add_58 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_26 = paddle._C_ops.transpose(reshape_26, [0, 2, 1, 3]) @@ -1715,9 +1501,11 @@ def forward( # pd_op.matmul: (1x16x21x21xf32) <- (1x16x21x64xf32, 1x16x21x64xf32) matmul_51 = paddle._C_ops.matmul(scale_7, transpose_25, False, True) + del scale_7, transpose_25 # pd_op.add: (1x16x21x21xf32) <- (1x16x21x21xf32, 1x1x1x21xf32) add_59 = paddle._C_ops.add(matmul_51, unsqueeze_0) + del matmul_51 # pd_op.softmax: (1x16x21x21xf32) <- (1x16x21x21xf32) softmax_6 = paddle._C_ops.softmax(add_59, -1) @@ -1726,13 +1514,15 @@ def forward( # pd_op.dropout: (1x16x21x21xf32, 1x16x21x21xui8) <- (1x16x21x21xf32, None, 1xf32) dropout_38, dropout_39 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_6, None, full_4, False, "upscale_in_train", 0, False + softmax_6, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_6 # pd_op.matmul: (1x16x21x64xf32) <- (1x16x21x21xf32, 1x16x21x64xf32) matmul_52 = paddle._C_ops.matmul(dropout_38, transpose_26, False, False) + del dropout_38, transpose_26 # pd_op.transpose: (1x21x16x64xf32) <- (1x16x21x64xf32) transpose_27 = paddle._C_ops.transpose(matmul_52, [0, 2, 1, 3]) @@ -1740,19 +1530,20 @@ def forward( # pd_op.reshape: (1x21x1024xf32) <- (1x21x16x64xf32, 3xi64) reshape_27 = paddle._C_ops.reshape(transpose_27, full_int_array_2) + del transpose_27 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_53 = paddle._C_ops.matmul(reshape_27, parameter_283, False, False) - del parameter_283 + del parameter_283, reshape_27 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_60 = paddle._C_ops.add(matmul_53, parameter_282) - del parameter_282 + del matmul_53, parameter_282 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_40, dropout_41 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_60, None, full_4, False, "upscale_in_train", 0, False + add_60, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1760,6 +1551,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_61 = paddle._C_ops.add(layer_norm_36, dropout_40) + del dropout_40, layer_norm_36 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_39, layer_norm_40, layer_norm_41 = (lambda x, f: f(x))( @@ -1768,7 +1560,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_276, parameter_277 + del add_61, parameter_276, parameter_277 # pd_op.matmul: (1x21x4096xf32) <- (1x21x1024xf32, 1024x4096xf32) matmul_54 = paddle._C_ops.matmul(layer_norm_39, parameter_281, False, False) @@ -1776,23 +1568,24 @@ def forward( # pd_op.add: (1x21x4096xf32) <- (1x21x4096xf32, 4096xf32) add_62 = paddle._C_ops.add(matmul_54, parameter_280) - del parameter_280 + del matmul_54, parameter_280 # pd_op.gelu: (1x21x4096xf32) <- (1x21x4096xf32) gelu_6 = paddle._C_ops.gelu(add_62, False) + del add_62 # pd_op.matmul: (1x21x1024xf32) <- (1x21x4096xf32, 4096x1024xf32) matmul_55 = paddle._C_ops.matmul(gelu_6, parameter_279, False, False) - del parameter_279 + del gelu_6, parameter_279 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_63 = paddle._C_ops.add(matmul_55, parameter_278) - del parameter_278 + del matmul_55, parameter_278 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_42, dropout_43 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_63, None, full_4, False, "upscale_in_train", 0, False + add_63, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1800,6 +1593,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_64 = paddle._C_ops.add(layer_norm_39, dropout_42) + del dropout_42, layer_norm_39 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_42, layer_norm_43, layer_norm_44 = (lambda x, f: f(x))( @@ -1808,7 +1602,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_274, parameter_275 + del add_64, parameter_274, parameter_275 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_56 = paddle._C_ops.matmul(layer_norm_42, parameter_273, False, False) @@ -1816,10 +1610,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_65 = paddle._C_ops.add(matmul_56, parameter_272) - del parameter_272 + del matmul_56, parameter_272 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_28 = paddle._C_ops.reshape(add_65, full_int_array_1) + del add_65 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_28 = paddle._C_ops.transpose(reshape_28, [0, 2, 1, 3]) @@ -1831,7 +1626,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_66 = paddle._C_ops.add(matmul_57, parameter_270) - del parameter_270 + del matmul_57, parameter_270 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_58 = paddle._C_ops.matmul(layer_norm_42, parameter_269, False, False) @@ -1839,10 +1634,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_67 = paddle._C_ops.add(matmul_58, parameter_268) - del parameter_268 + del matmul_58, parameter_268 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_29 = paddle._C_ops.reshape(add_66, full_int_array_1) + del add_66 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_29 = paddle._C_ops.transpose(reshape_29, [0, 2, 1, 3]) @@ -1850,6 +1646,7 @@ def forward( # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_30 = paddle._C_ops.reshape(add_67, full_int_array_1) + del add_67 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_30 = paddle._C_ops.transpose(reshape_30, [0, 2, 1, 3]) @@ -1861,9 +1658,11 @@ def forward( # pd_op.matmul: (1x16x21x21xf32) <- (1x16x21x64xf32, 1x16x21x64xf32) matmul_59 = paddle._C_ops.matmul(scale_8, transpose_29, False, True) + del scale_8, transpose_29 # pd_op.add: (1x16x21x21xf32) <- (1x16x21x21xf32, 1x1x1x21xf32) add_68 = paddle._C_ops.add(matmul_59, unsqueeze_0) + del matmul_59 # pd_op.softmax: (1x16x21x21xf32) <- (1x16x21x21xf32) softmax_7 = paddle._C_ops.softmax(add_68, -1) @@ -1872,13 +1671,15 @@ def forward( # pd_op.dropout: (1x16x21x21xf32, 1x16x21x21xui8) <- (1x16x21x21xf32, None, 1xf32) dropout_44, dropout_45 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_7, None, full_4, False, "upscale_in_train", 0, False + softmax_7, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_7 # pd_op.matmul: (1x16x21x64xf32) <- (1x16x21x21xf32, 1x16x21x64xf32) matmul_60 = paddle._C_ops.matmul(dropout_44, transpose_30, False, False) + del dropout_44, transpose_30 # pd_op.transpose: (1x21x16x64xf32) <- (1x16x21x64xf32) transpose_31 = paddle._C_ops.transpose(matmul_60, [0, 2, 1, 3]) @@ -1886,19 +1687,20 @@ def forward( # pd_op.reshape: (1x21x1024xf32) <- (1x21x16x64xf32, 3xi64) reshape_31 = paddle._C_ops.reshape(transpose_31, full_int_array_2) + del transpose_31 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_61 = paddle._C_ops.matmul(reshape_31, parameter_267, False, False) - del parameter_267 + del parameter_267, reshape_31 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_69 = paddle._C_ops.add(matmul_61, parameter_266) - del parameter_266 + del matmul_61, parameter_266 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_46, dropout_47 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_69, None, full_4, False, "upscale_in_train", 0, False + add_69, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1906,6 +1708,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_70 = paddle._C_ops.add(layer_norm_42, dropout_46) + del dropout_46, layer_norm_42 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_45, layer_norm_46, layer_norm_47 = (lambda x, f: f(x))( @@ -1914,7 +1717,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_260, parameter_261 + del add_70, parameter_260, parameter_261 # pd_op.matmul: (1x21x4096xf32) <- (1x21x1024xf32, 1024x4096xf32) matmul_62 = paddle._C_ops.matmul(layer_norm_45, parameter_265, False, False) @@ -1922,23 +1725,24 @@ def forward( # pd_op.add: (1x21x4096xf32) <- (1x21x4096xf32, 4096xf32) add_71 = paddle._C_ops.add(matmul_62, parameter_264) - del parameter_264 + del matmul_62, parameter_264 # pd_op.gelu: (1x21x4096xf32) <- (1x21x4096xf32) gelu_7 = paddle._C_ops.gelu(add_71, False) + del add_71 # pd_op.matmul: (1x21x1024xf32) <- (1x21x4096xf32, 4096x1024xf32) matmul_63 = paddle._C_ops.matmul(gelu_7, parameter_263, False, False) - del parameter_263 + del gelu_7, parameter_263 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_72 = paddle._C_ops.add(matmul_63, parameter_262) - del parameter_262 + del matmul_63, parameter_262 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_48, dropout_49 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_72, None, full_4, False, "upscale_in_train", 0, False + add_72, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1946,6 +1750,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_73 = paddle._C_ops.add(layer_norm_45, dropout_48) + del dropout_48, layer_norm_45 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_48, layer_norm_49, layer_norm_50 = (lambda x, f: f(x))( @@ -1954,7 +1759,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_258, parameter_259 + del add_73, parameter_258, parameter_259 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_64 = paddle._C_ops.matmul(layer_norm_48, parameter_257, False, False) @@ -1962,10 +1767,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_74 = paddle._C_ops.add(matmul_64, parameter_256) - del parameter_256 + del matmul_64, parameter_256 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_32 = paddle._C_ops.reshape(add_74, full_int_array_1) + del add_74 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_32 = paddle._C_ops.transpose(reshape_32, [0, 2, 1, 3]) @@ -1977,7 +1783,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_75 = paddle._C_ops.add(matmul_65, parameter_254) - del parameter_254 + del matmul_65, parameter_254 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_66 = paddle._C_ops.matmul(layer_norm_48, parameter_253, False, False) @@ -1985,10 +1791,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_76 = paddle._C_ops.add(matmul_66, parameter_252) - del parameter_252 + del matmul_66, parameter_252 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_33 = paddle._C_ops.reshape(add_75, full_int_array_1) + del add_75 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_33 = paddle._C_ops.transpose(reshape_33, [0, 2, 1, 3]) @@ -1996,6 +1803,7 @@ def forward( # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_34 = paddle._C_ops.reshape(add_76, full_int_array_1) + del add_76 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_34 = paddle._C_ops.transpose(reshape_34, [0, 2, 1, 3]) @@ -2007,9 +1815,11 @@ def forward( # pd_op.matmul: (1x16x21x21xf32) <- (1x16x21x64xf32, 1x16x21x64xf32) matmul_67 = paddle._C_ops.matmul(scale_9, transpose_33, False, True) + del scale_9, transpose_33 # pd_op.add: (1x16x21x21xf32) <- (1x16x21x21xf32, 1x1x1x21xf32) add_77 = paddle._C_ops.add(matmul_67, unsqueeze_0) + del matmul_67 # pd_op.softmax: (1x16x21x21xf32) <- (1x16x21x21xf32) softmax_8 = paddle._C_ops.softmax(add_77, -1) @@ -2018,13 +1828,15 @@ def forward( # pd_op.dropout: (1x16x21x21xf32, 1x16x21x21xui8) <- (1x16x21x21xf32, None, 1xf32) dropout_50, dropout_51 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_8, None, full_4, False, "upscale_in_train", 0, False + softmax_8, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_8 # pd_op.matmul: (1x16x21x64xf32) <- (1x16x21x21xf32, 1x16x21x64xf32) matmul_68 = paddle._C_ops.matmul(dropout_50, transpose_34, False, False) + del dropout_50, transpose_34 # pd_op.transpose: (1x21x16x64xf32) <- (1x16x21x64xf32) transpose_35 = paddle._C_ops.transpose(matmul_68, [0, 2, 1, 3]) @@ -2032,19 +1844,20 @@ def forward( # pd_op.reshape: (1x21x1024xf32) <- (1x21x16x64xf32, 3xi64) reshape_35 = paddle._C_ops.reshape(transpose_35, full_int_array_2) + del transpose_35 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_69 = paddle._C_ops.matmul(reshape_35, parameter_251, False, False) - del parameter_251 + del parameter_251, reshape_35 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_78 = paddle._C_ops.add(matmul_69, parameter_250) - del parameter_250 + del matmul_69, parameter_250 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_52, dropout_53 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_78, None, full_4, False, "upscale_in_train", 0, False + add_78, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -2052,6 +1865,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_79 = paddle._C_ops.add(layer_norm_48, dropout_52) + del dropout_52, layer_norm_48 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_51, layer_norm_52, layer_norm_53 = (lambda x, f: f(x))( @@ -2060,7 +1874,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_244, parameter_245 + del add_79, parameter_244, parameter_245 # pd_op.matmul: (1x21x4096xf32) <- (1x21x1024xf32, 1024x4096xf32) matmul_70 = paddle._C_ops.matmul(layer_norm_51, parameter_249, False, False) @@ -2068,23 +1882,24 @@ def forward( # pd_op.add: (1x21x4096xf32) <- (1x21x4096xf32, 4096xf32) add_80 = paddle._C_ops.add(matmul_70, parameter_248) - del parameter_248 + del matmul_70, parameter_248 # pd_op.gelu: (1x21x4096xf32) <- (1x21x4096xf32) gelu_8 = paddle._C_ops.gelu(add_80, False) + del add_80 # pd_op.matmul: (1x21x1024xf32) <- (1x21x4096xf32, 4096x1024xf32) matmul_71 = paddle._C_ops.matmul(gelu_8, parameter_247, False, False) - del parameter_247 + del gelu_8, parameter_247 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_81 = paddle._C_ops.add(matmul_71, parameter_246) - del parameter_246 + del matmul_71, parameter_246 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_54, dropout_55 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_81, None, full_4, False, "upscale_in_train", 0, False + add_81, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -2092,6 +1907,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_82 = paddle._C_ops.add(layer_norm_51, dropout_54) + del dropout_54, layer_norm_51 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_54, layer_norm_55, layer_norm_56 = (lambda x, f: f(x))( @@ -2100,7 +1916,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_242, parameter_243 + del add_82, parameter_242, parameter_243 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_72 = paddle._C_ops.matmul(layer_norm_54, parameter_241, False, False) @@ -2108,10 +1924,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_83 = paddle._C_ops.add(matmul_72, parameter_240) - del parameter_240 + del matmul_72, parameter_240 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_36 = paddle._C_ops.reshape(add_83, full_int_array_1) + del add_83 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_36 = paddle._C_ops.transpose(reshape_36, [0, 2, 1, 3]) @@ -2123,7 +1940,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_84 = paddle._C_ops.add(matmul_73, parameter_238) - del parameter_238 + del matmul_73, parameter_238 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_74 = paddle._C_ops.matmul(layer_norm_54, parameter_237, False, False) @@ -2131,10 +1948,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_85 = paddle._C_ops.add(matmul_74, parameter_236) - del parameter_236 + del matmul_74, parameter_236 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_37 = paddle._C_ops.reshape(add_84, full_int_array_1) + del add_84 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_37 = paddle._C_ops.transpose(reshape_37, [0, 2, 1, 3]) @@ -2142,6 +1960,7 @@ def forward( # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_38 = paddle._C_ops.reshape(add_85, full_int_array_1) + del add_85 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_38 = paddle._C_ops.transpose(reshape_38, [0, 2, 1, 3]) @@ -2153,9 +1972,11 @@ def forward( # pd_op.matmul: (1x16x21x21xf32) <- (1x16x21x64xf32, 1x16x21x64xf32) matmul_75 = paddle._C_ops.matmul(scale_10, transpose_37, False, True) + del scale_10, transpose_37 # pd_op.add: (1x16x21x21xf32) <- (1x16x21x21xf32, 1x1x1x21xf32) add_86 = paddle._C_ops.add(matmul_75, unsqueeze_0) + del matmul_75 # pd_op.softmax: (1x16x21x21xf32) <- (1x16x21x21xf32) softmax_9 = paddle._C_ops.softmax(add_86, -1) @@ -2164,13 +1985,15 @@ def forward( # pd_op.dropout: (1x16x21x21xf32, 1x16x21x21xui8) <- (1x16x21x21xf32, None, 1xf32) dropout_56, dropout_57 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_9, None, full_4, False, "upscale_in_train", 0, False + softmax_9, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_9 # pd_op.matmul: (1x16x21x64xf32) <- (1x16x21x21xf32, 1x16x21x64xf32) matmul_76 = paddle._C_ops.matmul(dropout_56, transpose_38, False, False) + del dropout_56, transpose_38 # pd_op.transpose: (1x21x16x64xf32) <- (1x16x21x64xf32) transpose_39 = paddle._C_ops.transpose(matmul_76, [0, 2, 1, 3]) @@ -2178,19 +2001,20 @@ def forward( # pd_op.reshape: (1x21x1024xf32) <- (1x21x16x64xf32, 3xi64) reshape_39 = paddle._C_ops.reshape(transpose_39, full_int_array_2) + del transpose_39 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_77 = paddle._C_ops.matmul(reshape_39, parameter_235, False, False) - del parameter_235 + del parameter_235, reshape_39 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_87 = paddle._C_ops.add(matmul_77, parameter_234) - del parameter_234 + del matmul_77, parameter_234 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_58, dropout_59 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_87, None, full_4, False, "upscale_in_train", 0, False + add_87, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -2198,6 +2022,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_88 = paddle._C_ops.add(layer_norm_54, dropout_58) + del dropout_58, layer_norm_54 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_57, layer_norm_58, layer_norm_59 = (lambda x, f: f(x))( @@ -2206,7 +2031,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_228, parameter_229 + del add_88, parameter_228, parameter_229 # pd_op.matmul: (1x21x4096xf32) <- (1x21x1024xf32, 1024x4096xf32) matmul_78 = paddle._C_ops.matmul(layer_norm_57, parameter_233, False, False) @@ -2214,23 +2039,24 @@ def forward( # pd_op.add: (1x21x4096xf32) <- (1x21x4096xf32, 4096xf32) add_89 = paddle._C_ops.add(matmul_78, parameter_232) - del parameter_232 + del matmul_78, parameter_232 # pd_op.gelu: (1x21x4096xf32) <- (1x21x4096xf32) gelu_9 = paddle._C_ops.gelu(add_89, False) + del add_89 # pd_op.matmul: (1x21x1024xf32) <- (1x21x4096xf32, 4096x1024xf32) matmul_79 = paddle._C_ops.matmul(gelu_9, parameter_231, False, False) - del parameter_231 + del gelu_9, parameter_231 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_90 = paddle._C_ops.add(matmul_79, parameter_230) - del parameter_230 + del matmul_79, parameter_230 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_60, dropout_61 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_90, None, full_4, False, "upscale_in_train", 0, False + add_90, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -2238,6 +2064,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_91 = paddle._C_ops.add(layer_norm_57, dropout_60) + del dropout_60, layer_norm_57 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_60, layer_norm_61, layer_norm_62 = (lambda x, f: f(x))( @@ -2246,7 +2073,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_226, parameter_227 + del add_91, parameter_226, parameter_227 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_80 = paddle._C_ops.matmul(layer_norm_60, parameter_225, False, False) @@ -2254,10 +2081,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_92 = paddle._C_ops.add(matmul_80, parameter_224) - del parameter_224 + del matmul_80, parameter_224 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_40 = paddle._C_ops.reshape(add_92, full_int_array_1) + del add_92 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_40 = paddle._C_ops.transpose(reshape_40, [0, 2, 1, 3]) @@ -2269,7 +2097,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_93 = paddle._C_ops.add(matmul_81, parameter_222) - del parameter_222 + del matmul_81, parameter_222 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_82 = paddle._C_ops.matmul(layer_norm_60, parameter_221, False, False) @@ -2277,10 +2105,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_94 = paddle._C_ops.add(matmul_82, parameter_220) - del parameter_220 + del matmul_82, parameter_220 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_41 = paddle._C_ops.reshape(add_93, full_int_array_1) + del add_93 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_41 = paddle._C_ops.transpose(reshape_41, [0, 2, 1, 3]) @@ -2288,6 +2117,7 @@ def forward( # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_42 = paddle._C_ops.reshape(add_94, full_int_array_1) + del add_94 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_42 = paddle._C_ops.transpose(reshape_42, [0, 2, 1, 3]) @@ -2299,9 +2129,11 @@ def forward( # pd_op.matmul: (1x16x21x21xf32) <- (1x16x21x64xf32, 1x16x21x64xf32) matmul_83 = paddle._C_ops.matmul(scale_11, transpose_41, False, True) + del scale_11, transpose_41 # pd_op.add: (1x16x21x21xf32) <- (1x16x21x21xf32, 1x1x1x21xf32) add_95 = paddle._C_ops.add(matmul_83, unsqueeze_0) + del matmul_83 # pd_op.softmax: (1x16x21x21xf32) <- (1x16x21x21xf32) softmax_10 = paddle._C_ops.softmax(add_95, -1) @@ -2310,13 +2142,15 @@ def forward( # pd_op.dropout: (1x16x21x21xf32, 1x16x21x21xui8) <- (1x16x21x21xf32, None, 1xf32) dropout_62, dropout_63 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_10, None, full_4, False, "upscale_in_train", 0, False + softmax_10, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_10 # pd_op.matmul: (1x16x21x64xf32) <- (1x16x21x21xf32, 1x16x21x64xf32) matmul_84 = paddle._C_ops.matmul(dropout_62, transpose_42, False, False) + del dropout_62, transpose_42 # pd_op.transpose: (1x21x16x64xf32) <- (1x16x21x64xf32) transpose_43 = paddle._C_ops.transpose(matmul_84, [0, 2, 1, 3]) @@ -2324,19 +2158,20 @@ def forward( # pd_op.reshape: (1x21x1024xf32) <- (1x21x16x64xf32, 3xi64) reshape_43 = paddle._C_ops.reshape(transpose_43, full_int_array_2) + del transpose_43 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_85 = paddle._C_ops.matmul(reshape_43, parameter_219, False, False) - del parameter_219 + del parameter_219, reshape_43 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_96 = paddle._C_ops.add(matmul_85, parameter_218) - del parameter_218 + del matmul_85, parameter_218 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_64, dropout_65 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_96, None, full_4, False, "upscale_in_train", 0, False + add_96, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -2344,6 +2179,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_97 = paddle._C_ops.add(layer_norm_60, dropout_64) + del dropout_64, layer_norm_60 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_63, layer_norm_64, layer_norm_65 = (lambda x, f: f(x))( @@ -2352,7 +2188,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_212, parameter_213 + del add_97, parameter_212, parameter_213 # pd_op.matmul: (1x21x4096xf32) <- (1x21x1024xf32, 1024x4096xf32) matmul_86 = paddle._C_ops.matmul(layer_norm_63, parameter_217, False, False) @@ -2360,23 +2196,24 @@ def forward( # pd_op.add: (1x21x4096xf32) <- (1x21x4096xf32, 4096xf32) add_98 = paddle._C_ops.add(matmul_86, parameter_216) - del parameter_216 + del matmul_86, parameter_216 # pd_op.gelu: (1x21x4096xf32) <- (1x21x4096xf32) gelu_10 = paddle._C_ops.gelu(add_98, False) + del add_98 # pd_op.matmul: (1x21x1024xf32) <- (1x21x4096xf32, 4096x1024xf32) matmul_87 = paddle._C_ops.matmul(gelu_10, parameter_215, False, False) - del parameter_215 + del gelu_10, parameter_215 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_99 = paddle._C_ops.add(matmul_87, parameter_214) - del parameter_214 + del matmul_87, parameter_214 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_66, dropout_67 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_99, None, full_4, False, "upscale_in_train", 0, False + add_99, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -2384,6 +2221,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_100 = paddle._C_ops.add(layer_norm_63, dropout_66) + del dropout_66, layer_norm_63 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_66, layer_norm_67, layer_norm_68 = (lambda x, f: f(x))( @@ -2392,7 +2230,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_210, parameter_211 + del add_100, parameter_210, parameter_211 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_88 = paddle._C_ops.matmul(layer_norm_66, parameter_209, False, False) @@ -2400,10 +2238,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_101 = paddle._C_ops.add(matmul_88, parameter_208) - del parameter_208 + del matmul_88, parameter_208 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_44 = paddle._C_ops.reshape(add_101, full_int_array_1) + del add_101 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_44 = paddle._C_ops.transpose(reshape_44, [0, 2, 1, 3]) @@ -2415,7 +2254,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_102 = paddle._C_ops.add(matmul_89, parameter_206) - del parameter_206 + del matmul_89, parameter_206 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_90 = paddle._C_ops.matmul(layer_norm_66, parameter_205, False, False) @@ -2423,10 +2262,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_103 = paddle._C_ops.add(matmul_90, parameter_204) - del parameter_204 + del matmul_90, parameter_204 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_45 = paddle._C_ops.reshape(add_102, full_int_array_1) + del add_102 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_45 = paddle._C_ops.transpose(reshape_45, [0, 2, 1, 3]) @@ -2434,6 +2274,7 @@ def forward( # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_46 = paddle._C_ops.reshape(add_103, full_int_array_1) + del add_103 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_46 = paddle._C_ops.transpose(reshape_46, [0, 2, 1, 3]) @@ -2445,9 +2286,11 @@ def forward( # pd_op.matmul: (1x16x21x21xf32) <- (1x16x21x64xf32, 1x16x21x64xf32) matmul_91 = paddle._C_ops.matmul(scale_12, transpose_45, False, True) + del scale_12, transpose_45 # pd_op.add: (1x16x21x21xf32) <- (1x16x21x21xf32, 1x1x1x21xf32) add_104 = paddle._C_ops.add(matmul_91, unsqueeze_0) + del matmul_91 # pd_op.softmax: (1x16x21x21xf32) <- (1x16x21x21xf32) softmax_11 = paddle._C_ops.softmax(add_104, -1) @@ -2456,13 +2299,15 @@ def forward( # pd_op.dropout: (1x16x21x21xf32, 1x16x21x21xui8) <- (1x16x21x21xf32, None, 1xf32) dropout_68, dropout_69 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_11, None, full_4, False, "upscale_in_train", 0, False + softmax_11, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_11 # pd_op.matmul: (1x16x21x64xf32) <- (1x16x21x21xf32, 1x16x21x64xf32) matmul_92 = paddle._C_ops.matmul(dropout_68, transpose_46, False, False) + del dropout_68, transpose_46 # pd_op.transpose: (1x21x16x64xf32) <- (1x16x21x64xf32) transpose_47 = paddle._C_ops.transpose(matmul_92, [0, 2, 1, 3]) @@ -2470,19 +2315,20 @@ def forward( # pd_op.reshape: (1x21x1024xf32) <- (1x21x16x64xf32, 3xi64) reshape_47 = paddle._C_ops.reshape(transpose_47, full_int_array_2) + del transpose_47 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_93 = paddle._C_ops.matmul(reshape_47, parameter_203, False, False) - del parameter_203 + del parameter_203, reshape_47 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_105 = paddle._C_ops.add(matmul_93, parameter_202) - del parameter_202 + del matmul_93, parameter_202 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_70, dropout_71 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_105, None, full_4, False, "upscale_in_train", 0, False + add_105, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -2490,6 +2336,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_106 = paddle._C_ops.add(layer_norm_66, dropout_70) + del dropout_70, layer_norm_66 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_69, layer_norm_70, layer_norm_71 = (lambda x, f: f(x))( @@ -2498,7 +2345,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_196, parameter_197 + del add_106, parameter_196, parameter_197 # pd_op.matmul: (1x21x4096xf32) <- (1x21x1024xf32, 1024x4096xf32) matmul_94 = paddle._C_ops.matmul(layer_norm_69, parameter_201, False, False) @@ -2506,23 +2353,24 @@ def forward( # pd_op.add: (1x21x4096xf32) <- (1x21x4096xf32, 4096xf32) add_107 = paddle._C_ops.add(matmul_94, parameter_200) - del parameter_200 + del matmul_94, parameter_200 # pd_op.gelu: (1x21x4096xf32) <- (1x21x4096xf32) gelu_11 = paddle._C_ops.gelu(add_107, False) + del add_107 # pd_op.matmul: (1x21x1024xf32) <- (1x21x4096xf32, 4096x1024xf32) matmul_95 = paddle._C_ops.matmul(gelu_11, parameter_199, False, False) - del parameter_199 + del gelu_11, parameter_199 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_108 = paddle._C_ops.add(matmul_95, parameter_198) - del parameter_198 + del matmul_95, parameter_198 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_72, dropout_73 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_108, None, full_4, False, "upscale_in_train", 0, False + add_108, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -2530,6 +2378,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_109 = paddle._C_ops.add(layer_norm_69, dropout_72) + del dropout_72, layer_norm_69 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_72, layer_norm_73, layer_norm_74 = (lambda x, f: f(x))( @@ -2538,7 +2387,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_194, parameter_195 + del add_109, parameter_194, parameter_195 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_96 = paddle._C_ops.matmul(layer_norm_72, parameter_193, False, False) @@ -2546,10 +2395,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_110 = paddle._C_ops.add(matmul_96, parameter_192) - del parameter_192 + del matmul_96, parameter_192 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_48 = paddle._C_ops.reshape(add_110, full_int_array_1) + del add_110 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_48 = paddle._C_ops.transpose(reshape_48, [0, 2, 1, 3]) @@ -2561,7 +2411,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_111 = paddle._C_ops.add(matmul_97, parameter_190) - del parameter_190 + del matmul_97, parameter_190 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_98 = paddle._C_ops.matmul(layer_norm_72, parameter_189, False, False) @@ -2569,10 +2419,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_112 = paddle._C_ops.add(matmul_98, parameter_188) - del parameter_188 + del matmul_98, parameter_188 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_49 = paddle._C_ops.reshape(add_111, full_int_array_1) + del add_111 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_49 = paddle._C_ops.transpose(reshape_49, [0, 2, 1, 3]) @@ -2580,6 +2431,7 @@ def forward( # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_50 = paddle._C_ops.reshape(add_112, full_int_array_1) + del add_112 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_50 = paddle._C_ops.transpose(reshape_50, [0, 2, 1, 3]) @@ -2591,9 +2443,11 @@ def forward( # pd_op.matmul: (1x16x21x21xf32) <- (1x16x21x64xf32, 1x16x21x64xf32) matmul_99 = paddle._C_ops.matmul(scale_13, transpose_49, False, True) + del scale_13, transpose_49 # pd_op.add: (1x16x21x21xf32) <- (1x16x21x21xf32, 1x1x1x21xf32) add_113 = paddle._C_ops.add(matmul_99, unsqueeze_0) + del matmul_99 # pd_op.softmax: (1x16x21x21xf32) <- (1x16x21x21xf32) softmax_12 = paddle._C_ops.softmax(add_113, -1) @@ -2602,13 +2456,15 @@ def forward( # pd_op.dropout: (1x16x21x21xf32, 1x16x21x21xui8) <- (1x16x21x21xf32, None, 1xf32) dropout_74, dropout_75 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_12, None, full_4, False, "upscale_in_train", 0, False + softmax_12, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_12 # pd_op.matmul: (1x16x21x64xf32) <- (1x16x21x21xf32, 1x16x21x64xf32) matmul_100 = paddle._C_ops.matmul(dropout_74, transpose_50, False, False) + del dropout_74, transpose_50 # pd_op.transpose: (1x21x16x64xf32) <- (1x16x21x64xf32) transpose_51 = paddle._C_ops.transpose(matmul_100, [0, 2, 1, 3]) @@ -2616,19 +2472,20 @@ def forward( # pd_op.reshape: (1x21x1024xf32) <- (1x21x16x64xf32, 3xi64) reshape_51 = paddle._C_ops.reshape(transpose_51, full_int_array_2) + del transpose_51 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_101 = paddle._C_ops.matmul(reshape_51, parameter_187, False, False) - del parameter_187 + del parameter_187, reshape_51 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_114 = paddle._C_ops.add(matmul_101, parameter_186) - del parameter_186 + del matmul_101, parameter_186 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_76, dropout_77 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_114, None, full_4, False, "upscale_in_train", 0, False + add_114, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -2636,6 +2493,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_115 = paddle._C_ops.add(layer_norm_72, dropout_76) + del dropout_76, layer_norm_72 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_75, layer_norm_76, layer_norm_77 = (lambda x, f: f(x))( @@ -2644,7 +2502,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_180, parameter_181 + del add_115, parameter_180, parameter_181 # pd_op.matmul: (1x21x4096xf32) <- (1x21x1024xf32, 1024x4096xf32) matmul_102 = paddle._C_ops.matmul(layer_norm_75, parameter_185, False, False) @@ -2652,23 +2510,24 @@ def forward( # pd_op.add: (1x21x4096xf32) <- (1x21x4096xf32, 4096xf32) add_116 = paddle._C_ops.add(matmul_102, parameter_184) - del parameter_184 + del matmul_102, parameter_184 # pd_op.gelu: (1x21x4096xf32) <- (1x21x4096xf32) gelu_12 = paddle._C_ops.gelu(add_116, False) + del add_116 # pd_op.matmul: (1x21x1024xf32) <- (1x21x4096xf32, 4096x1024xf32) matmul_103 = paddle._C_ops.matmul(gelu_12, parameter_183, False, False) - del parameter_183 + del gelu_12, parameter_183 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_117 = paddle._C_ops.add(matmul_103, parameter_182) - del parameter_182 + del matmul_103, parameter_182 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_78, dropout_79 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_117, None, full_4, False, "upscale_in_train", 0, False + add_117, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -2676,6 +2535,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_118 = paddle._C_ops.add(layer_norm_75, dropout_78) + del dropout_78, layer_norm_75 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_78, layer_norm_79, layer_norm_80 = (lambda x, f: f(x))( @@ -2684,7 +2544,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_178, parameter_179 + del add_118, parameter_178, parameter_179 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_104 = paddle._C_ops.matmul(layer_norm_78, parameter_177, False, False) @@ -2692,10 +2552,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_119 = paddle._C_ops.add(matmul_104, parameter_176) - del parameter_176 + del matmul_104, parameter_176 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_52 = paddle._C_ops.reshape(add_119, full_int_array_1) + del add_119 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_52 = paddle._C_ops.transpose(reshape_52, [0, 2, 1, 3]) @@ -2707,7 +2568,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_120 = paddle._C_ops.add(matmul_105, parameter_174) - del parameter_174 + del matmul_105, parameter_174 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_106 = paddle._C_ops.matmul(layer_norm_78, parameter_173, False, False) @@ -2715,10 +2576,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_121 = paddle._C_ops.add(matmul_106, parameter_172) - del parameter_172 + del matmul_106, parameter_172 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_53 = paddle._C_ops.reshape(add_120, full_int_array_1) + del add_120 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_53 = paddle._C_ops.transpose(reshape_53, [0, 2, 1, 3]) @@ -2726,6 +2588,7 @@ def forward( # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_54 = paddle._C_ops.reshape(add_121, full_int_array_1) + del add_121 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_54 = paddle._C_ops.transpose(reshape_54, [0, 2, 1, 3]) @@ -2737,9 +2600,11 @@ def forward( # pd_op.matmul: (1x16x21x21xf32) <- (1x16x21x64xf32, 1x16x21x64xf32) matmul_107 = paddle._C_ops.matmul(scale_14, transpose_53, False, True) + del scale_14, transpose_53 # pd_op.add: (1x16x21x21xf32) <- (1x16x21x21xf32, 1x1x1x21xf32) add_122 = paddle._C_ops.add(matmul_107, unsqueeze_0) + del matmul_107 # pd_op.softmax: (1x16x21x21xf32) <- (1x16x21x21xf32) softmax_13 = paddle._C_ops.softmax(add_122, -1) @@ -2748,13 +2613,15 @@ def forward( # pd_op.dropout: (1x16x21x21xf32, 1x16x21x21xui8) <- (1x16x21x21xf32, None, 1xf32) dropout_80, dropout_81 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_13, None, full_4, False, "upscale_in_train", 0, False + softmax_13, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_13 # pd_op.matmul: (1x16x21x64xf32) <- (1x16x21x21xf32, 1x16x21x64xf32) matmul_108 = paddle._C_ops.matmul(dropout_80, transpose_54, False, False) + del dropout_80, transpose_54 # pd_op.transpose: (1x21x16x64xf32) <- (1x16x21x64xf32) transpose_55 = paddle._C_ops.transpose(matmul_108, [0, 2, 1, 3]) @@ -2762,19 +2629,20 @@ def forward( # pd_op.reshape: (1x21x1024xf32) <- (1x21x16x64xf32, 3xi64) reshape_55 = paddle._C_ops.reshape(transpose_55, full_int_array_2) + del transpose_55 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_109 = paddle._C_ops.matmul(reshape_55, parameter_171, False, False) - del parameter_171 + del parameter_171, reshape_55 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_123 = paddle._C_ops.add(matmul_109, parameter_170) - del parameter_170 + del matmul_109, parameter_170 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_82, dropout_83 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_123, None, full_4, False, "upscale_in_train", 0, False + add_123, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -2782,6 +2650,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_124 = paddle._C_ops.add(layer_norm_78, dropout_82) + del dropout_82, layer_norm_78 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_81, layer_norm_82, layer_norm_83 = (lambda x, f: f(x))( @@ -2790,7 +2659,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_164, parameter_165 + del add_124, parameter_164, parameter_165 # pd_op.matmul: (1x21x4096xf32) <- (1x21x1024xf32, 1024x4096xf32) matmul_110 = paddle._C_ops.matmul(layer_norm_81, parameter_169, False, False) @@ -2798,23 +2667,24 @@ def forward( # pd_op.add: (1x21x4096xf32) <- (1x21x4096xf32, 4096xf32) add_125 = paddle._C_ops.add(matmul_110, parameter_168) - del parameter_168 + del matmul_110, parameter_168 # pd_op.gelu: (1x21x4096xf32) <- (1x21x4096xf32) gelu_13 = paddle._C_ops.gelu(add_125, False) + del add_125 # pd_op.matmul: (1x21x1024xf32) <- (1x21x4096xf32, 4096x1024xf32) matmul_111 = paddle._C_ops.matmul(gelu_13, parameter_167, False, False) - del parameter_167 + del gelu_13, parameter_167 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_126 = paddle._C_ops.add(matmul_111, parameter_166) - del parameter_166 + del matmul_111, parameter_166 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_84, dropout_85 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_126, None, full_4, False, "upscale_in_train", 0, False + add_126, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -2822,6 +2692,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_127 = paddle._C_ops.add(layer_norm_81, dropout_84) + del dropout_84, layer_norm_81 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_84, layer_norm_85, layer_norm_86 = (lambda x, f: f(x))( @@ -2830,7 +2701,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_162, parameter_163 + del add_127, parameter_162, parameter_163 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_112 = paddle._C_ops.matmul(layer_norm_84, parameter_161, False, False) @@ -2838,10 +2709,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_128 = paddle._C_ops.add(matmul_112, parameter_160) - del parameter_160 + del matmul_112, parameter_160 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_56 = paddle._C_ops.reshape(add_128, full_int_array_1) + del add_128 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_56 = paddle._C_ops.transpose(reshape_56, [0, 2, 1, 3]) @@ -2853,7 +2725,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_129 = paddle._C_ops.add(matmul_113, parameter_158) - del parameter_158 + del matmul_113, parameter_158 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_114 = paddle._C_ops.matmul(layer_norm_84, parameter_157, False, False) @@ -2861,10 +2733,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_130 = paddle._C_ops.add(matmul_114, parameter_156) - del parameter_156 + del matmul_114, parameter_156 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_57 = paddle._C_ops.reshape(add_129, full_int_array_1) + del add_129 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_57 = paddle._C_ops.transpose(reshape_57, [0, 2, 1, 3]) @@ -2872,6 +2745,7 @@ def forward( # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_58 = paddle._C_ops.reshape(add_130, full_int_array_1) + del add_130 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_58 = paddle._C_ops.transpose(reshape_58, [0, 2, 1, 3]) @@ -2883,9 +2757,11 @@ def forward( # pd_op.matmul: (1x16x21x21xf32) <- (1x16x21x64xf32, 1x16x21x64xf32) matmul_115 = paddle._C_ops.matmul(scale_15, transpose_57, False, True) + del scale_15, transpose_57 # pd_op.add: (1x16x21x21xf32) <- (1x16x21x21xf32, 1x1x1x21xf32) add_131 = paddle._C_ops.add(matmul_115, unsqueeze_0) + del matmul_115 # pd_op.softmax: (1x16x21x21xf32) <- (1x16x21x21xf32) softmax_14 = paddle._C_ops.softmax(add_131, -1) @@ -2894,13 +2770,15 @@ def forward( # pd_op.dropout: (1x16x21x21xf32, 1x16x21x21xui8) <- (1x16x21x21xf32, None, 1xf32) dropout_86, dropout_87 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_14, None, full_4, False, "upscale_in_train", 0, False + softmax_14, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_14 # pd_op.matmul: (1x16x21x64xf32) <- (1x16x21x21xf32, 1x16x21x64xf32) matmul_116 = paddle._C_ops.matmul(dropout_86, transpose_58, False, False) + del dropout_86, transpose_58 # pd_op.transpose: (1x21x16x64xf32) <- (1x16x21x64xf32) transpose_59 = paddle._C_ops.transpose(matmul_116, [0, 2, 1, 3]) @@ -2908,19 +2786,20 @@ def forward( # pd_op.reshape: (1x21x1024xf32) <- (1x21x16x64xf32, 3xi64) reshape_59 = paddle._C_ops.reshape(transpose_59, full_int_array_2) + del transpose_59 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_117 = paddle._C_ops.matmul(reshape_59, parameter_155, False, False) - del parameter_155 + del parameter_155, reshape_59 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_132 = paddle._C_ops.add(matmul_117, parameter_154) - del parameter_154 + del matmul_117, parameter_154 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_88, dropout_89 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_132, None, full_4, False, "upscale_in_train", 0, False + add_132, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -2928,6 +2807,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_133 = paddle._C_ops.add(layer_norm_84, dropout_88) + del dropout_88, layer_norm_84 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_87, layer_norm_88, layer_norm_89 = (lambda x, f: f(x))( @@ -2936,7 +2816,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_148, parameter_149 + del add_133, parameter_148, parameter_149 # pd_op.matmul: (1x21x4096xf32) <- (1x21x1024xf32, 1024x4096xf32) matmul_118 = paddle._C_ops.matmul(layer_norm_87, parameter_153, False, False) @@ -2944,23 +2824,24 @@ def forward( # pd_op.add: (1x21x4096xf32) <- (1x21x4096xf32, 4096xf32) add_134 = paddle._C_ops.add(matmul_118, parameter_152) - del parameter_152 + del matmul_118, parameter_152 # pd_op.gelu: (1x21x4096xf32) <- (1x21x4096xf32) gelu_14 = paddle._C_ops.gelu(add_134, False) + del add_134 # pd_op.matmul: (1x21x1024xf32) <- (1x21x4096xf32, 4096x1024xf32) matmul_119 = paddle._C_ops.matmul(gelu_14, parameter_151, False, False) - del parameter_151 + del gelu_14, parameter_151 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_135 = paddle._C_ops.add(matmul_119, parameter_150) - del parameter_150 + del matmul_119, parameter_150 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_90, dropout_91 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_135, None, full_4, False, "upscale_in_train", 0, False + add_135, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -2968,6 +2849,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_136 = paddle._C_ops.add(layer_norm_87, dropout_90) + del dropout_90, layer_norm_87 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_90, layer_norm_91, layer_norm_92 = (lambda x, f: f(x))( @@ -2976,7 +2858,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_146, parameter_147 + del add_136, parameter_146, parameter_147 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_120 = paddle._C_ops.matmul(layer_norm_90, parameter_145, False, False) @@ -2984,10 +2866,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_137 = paddle._C_ops.add(matmul_120, parameter_144) - del parameter_144 + del matmul_120, parameter_144 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_60 = paddle._C_ops.reshape(add_137, full_int_array_1) + del add_137 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_60 = paddle._C_ops.transpose(reshape_60, [0, 2, 1, 3]) @@ -2999,7 +2882,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_138 = paddle._C_ops.add(matmul_121, parameter_142) - del parameter_142 + del matmul_121, parameter_142 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_122 = paddle._C_ops.matmul(layer_norm_90, parameter_141, False, False) @@ -3007,10 +2890,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_139 = paddle._C_ops.add(matmul_122, parameter_140) - del parameter_140 + del matmul_122, parameter_140 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_61 = paddle._C_ops.reshape(add_138, full_int_array_1) + del add_138 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_61 = paddle._C_ops.transpose(reshape_61, [0, 2, 1, 3]) @@ -3018,6 +2902,7 @@ def forward( # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_62 = paddle._C_ops.reshape(add_139, full_int_array_1) + del add_139 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_62 = paddle._C_ops.transpose(reshape_62, [0, 2, 1, 3]) @@ -3029,9 +2914,11 @@ def forward( # pd_op.matmul: (1x16x21x21xf32) <- (1x16x21x64xf32, 1x16x21x64xf32) matmul_123 = paddle._C_ops.matmul(scale_16, transpose_61, False, True) + del scale_16, transpose_61 # pd_op.add: (1x16x21x21xf32) <- (1x16x21x21xf32, 1x1x1x21xf32) add_140 = paddle._C_ops.add(matmul_123, unsqueeze_0) + del matmul_123 # pd_op.softmax: (1x16x21x21xf32) <- (1x16x21x21xf32) softmax_15 = paddle._C_ops.softmax(add_140, -1) @@ -3040,13 +2927,15 @@ def forward( # pd_op.dropout: (1x16x21x21xf32, 1x16x21x21xui8) <- (1x16x21x21xf32, None, 1xf32) dropout_92, dropout_93 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_15, None, full_4, False, "upscale_in_train", 0, False + softmax_15, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_15 # pd_op.matmul: (1x16x21x64xf32) <- (1x16x21x21xf32, 1x16x21x64xf32) matmul_124 = paddle._C_ops.matmul(dropout_92, transpose_62, False, False) + del dropout_92, transpose_62 # pd_op.transpose: (1x21x16x64xf32) <- (1x16x21x64xf32) transpose_63 = paddle._C_ops.transpose(matmul_124, [0, 2, 1, 3]) @@ -3054,19 +2943,20 @@ def forward( # pd_op.reshape: (1x21x1024xf32) <- (1x21x16x64xf32, 3xi64) reshape_63 = paddle._C_ops.reshape(transpose_63, full_int_array_2) + del transpose_63 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_125 = paddle._C_ops.matmul(reshape_63, parameter_139, False, False) - del parameter_139 + del parameter_139, reshape_63 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_141 = paddle._C_ops.add(matmul_125, parameter_138) - del parameter_138 + del matmul_125, parameter_138 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_94, dropout_95 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_141, None, full_4, False, "upscale_in_train", 0, False + add_141, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -3074,6 +2964,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_142 = paddle._C_ops.add(layer_norm_90, dropout_94) + del dropout_94, layer_norm_90 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_93, layer_norm_94, layer_norm_95 = (lambda x, f: f(x))( @@ -3082,7 +2973,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_132, parameter_133 + del add_142, parameter_132, parameter_133 # pd_op.matmul: (1x21x4096xf32) <- (1x21x1024xf32, 1024x4096xf32) matmul_126 = paddle._C_ops.matmul(layer_norm_93, parameter_137, False, False) @@ -3090,23 +2981,24 @@ def forward( # pd_op.add: (1x21x4096xf32) <- (1x21x4096xf32, 4096xf32) add_143 = paddle._C_ops.add(matmul_126, parameter_136) - del parameter_136 + del matmul_126, parameter_136 # pd_op.gelu: (1x21x4096xf32) <- (1x21x4096xf32) gelu_15 = paddle._C_ops.gelu(add_143, False) + del add_143 # pd_op.matmul: (1x21x1024xf32) <- (1x21x4096xf32, 4096x1024xf32) matmul_127 = paddle._C_ops.matmul(gelu_15, parameter_135, False, False) - del parameter_135 + del gelu_15, parameter_135 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_144 = paddle._C_ops.add(matmul_127, parameter_134) - del parameter_134 + del matmul_127, parameter_134 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_96, dropout_97 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_144, None, full_4, False, "upscale_in_train", 0, False + add_144, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -3114,6 +3006,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_145 = paddle._C_ops.add(layer_norm_93, dropout_96) + del dropout_96, layer_norm_93 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_96, layer_norm_97, layer_norm_98 = (lambda x, f: f(x))( @@ -3122,7 +3015,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_130, parameter_131 + del add_145, parameter_130, parameter_131 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_128 = paddle._C_ops.matmul(layer_norm_96, parameter_129, False, False) @@ -3130,10 +3023,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_146 = paddle._C_ops.add(matmul_128, parameter_128) - del parameter_128 + del matmul_128, parameter_128 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_64 = paddle._C_ops.reshape(add_146, full_int_array_1) + del add_146 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_64 = paddle._C_ops.transpose(reshape_64, [0, 2, 1, 3]) @@ -3145,7 +3039,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_147 = paddle._C_ops.add(matmul_129, parameter_126) - del parameter_126 + del matmul_129, parameter_126 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_130 = paddle._C_ops.matmul(layer_norm_96, parameter_125, False, False) @@ -3153,10 +3047,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_148 = paddle._C_ops.add(matmul_130, parameter_124) - del parameter_124 + del matmul_130, parameter_124 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_65 = paddle._C_ops.reshape(add_147, full_int_array_1) + del add_147 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_65 = paddle._C_ops.transpose(reshape_65, [0, 2, 1, 3]) @@ -3164,6 +3059,7 @@ def forward( # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_66 = paddle._C_ops.reshape(add_148, full_int_array_1) + del add_148 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_66 = paddle._C_ops.transpose(reshape_66, [0, 2, 1, 3]) @@ -3175,9 +3071,11 @@ def forward( # pd_op.matmul: (1x16x21x21xf32) <- (1x16x21x64xf32, 1x16x21x64xf32) matmul_131 = paddle._C_ops.matmul(scale_17, transpose_65, False, True) + del scale_17, transpose_65 # pd_op.add: (1x16x21x21xf32) <- (1x16x21x21xf32, 1x1x1x21xf32) add_149 = paddle._C_ops.add(matmul_131, unsqueeze_0) + del matmul_131 # pd_op.softmax: (1x16x21x21xf32) <- (1x16x21x21xf32) softmax_16 = paddle._C_ops.softmax(add_149, -1) @@ -3186,13 +3084,15 @@ def forward( # pd_op.dropout: (1x16x21x21xf32, 1x16x21x21xui8) <- (1x16x21x21xf32, None, 1xf32) dropout_98, dropout_99 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_16, None, full_4, False, "upscale_in_train", 0, False + softmax_16, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_16 # pd_op.matmul: (1x16x21x64xf32) <- (1x16x21x21xf32, 1x16x21x64xf32) matmul_132 = paddle._C_ops.matmul(dropout_98, transpose_66, False, False) + del dropout_98, transpose_66 # pd_op.transpose: (1x21x16x64xf32) <- (1x16x21x64xf32) transpose_67 = paddle._C_ops.transpose(matmul_132, [0, 2, 1, 3]) @@ -3200,19 +3100,20 @@ def forward( # pd_op.reshape: (1x21x1024xf32) <- (1x21x16x64xf32, 3xi64) reshape_67 = paddle._C_ops.reshape(transpose_67, full_int_array_2) + del transpose_67 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_133 = paddle._C_ops.matmul(reshape_67, parameter_123, False, False) - del parameter_123 + del parameter_123, reshape_67 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_150 = paddle._C_ops.add(matmul_133, parameter_122) - del parameter_122 + del matmul_133, parameter_122 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_100, dropout_101 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_150, None, full_4, False, "upscale_in_train", 0, False + add_150, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -3220,6 +3121,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_151 = paddle._C_ops.add(layer_norm_96, dropout_100) + del dropout_100, layer_norm_96 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_99, layer_norm_100, layer_norm_101 = (lambda x, f: f(x))( @@ -3228,7 +3130,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_116, parameter_117 + del add_151, parameter_116, parameter_117 # pd_op.matmul: (1x21x4096xf32) <- (1x21x1024xf32, 1024x4096xf32) matmul_134 = paddle._C_ops.matmul(layer_norm_99, parameter_121, False, False) @@ -3236,23 +3138,24 @@ def forward( # pd_op.add: (1x21x4096xf32) <- (1x21x4096xf32, 4096xf32) add_152 = paddle._C_ops.add(matmul_134, parameter_120) - del parameter_120 + del matmul_134, parameter_120 # pd_op.gelu: (1x21x4096xf32) <- (1x21x4096xf32) gelu_16 = paddle._C_ops.gelu(add_152, False) + del add_152 # pd_op.matmul: (1x21x1024xf32) <- (1x21x4096xf32, 4096x1024xf32) matmul_135 = paddle._C_ops.matmul(gelu_16, parameter_119, False, False) - del parameter_119 + del gelu_16, parameter_119 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_153 = paddle._C_ops.add(matmul_135, parameter_118) - del parameter_118 + del matmul_135, parameter_118 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_102, dropout_103 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_153, None, full_4, False, "upscale_in_train", 0, False + add_153, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -3260,6 +3163,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_154 = paddle._C_ops.add(layer_norm_99, dropout_102) + del dropout_102, layer_norm_99 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_102, layer_norm_103, layer_norm_104 = (lambda x, f: f(x))( @@ -3268,7 +3172,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_114, parameter_115 + del add_154, parameter_114, parameter_115 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_136 = paddle._C_ops.matmul(layer_norm_102, parameter_113, False, False) @@ -3276,10 +3180,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_155 = paddle._C_ops.add(matmul_136, parameter_112) - del parameter_112 + del matmul_136, parameter_112 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_68 = paddle._C_ops.reshape(add_155, full_int_array_1) + del add_155 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_68 = paddle._C_ops.transpose(reshape_68, [0, 2, 1, 3]) @@ -3291,7 +3196,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_156 = paddle._C_ops.add(matmul_137, parameter_110) - del parameter_110 + del matmul_137, parameter_110 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_138 = paddle._C_ops.matmul(layer_norm_102, parameter_109, False, False) @@ -3299,10 +3204,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_157 = paddle._C_ops.add(matmul_138, parameter_108) - del parameter_108 + del matmul_138, parameter_108 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_69 = paddle._C_ops.reshape(add_156, full_int_array_1) + del add_156 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_69 = paddle._C_ops.transpose(reshape_69, [0, 2, 1, 3]) @@ -3310,6 +3216,7 @@ def forward( # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_70 = paddle._C_ops.reshape(add_157, full_int_array_1) + del add_157 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_70 = paddle._C_ops.transpose(reshape_70, [0, 2, 1, 3]) @@ -3321,9 +3228,11 @@ def forward( # pd_op.matmul: (1x16x21x21xf32) <- (1x16x21x64xf32, 1x16x21x64xf32) matmul_139 = paddle._C_ops.matmul(scale_18, transpose_69, False, True) + del scale_18, transpose_69 # pd_op.add: (1x16x21x21xf32) <- (1x16x21x21xf32, 1x1x1x21xf32) add_158 = paddle._C_ops.add(matmul_139, unsqueeze_0) + del matmul_139 # pd_op.softmax: (1x16x21x21xf32) <- (1x16x21x21xf32) softmax_17 = paddle._C_ops.softmax(add_158, -1) @@ -3332,13 +3241,15 @@ def forward( # pd_op.dropout: (1x16x21x21xf32, 1x16x21x21xui8) <- (1x16x21x21xf32, None, 1xf32) dropout_104, dropout_105 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_17, None, full_4, False, "upscale_in_train", 0, False + softmax_17, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_17 # pd_op.matmul: (1x16x21x64xf32) <- (1x16x21x21xf32, 1x16x21x64xf32) matmul_140 = paddle._C_ops.matmul(dropout_104, transpose_70, False, False) + del dropout_104, transpose_70 # pd_op.transpose: (1x21x16x64xf32) <- (1x16x21x64xf32) transpose_71 = paddle._C_ops.transpose(matmul_140, [0, 2, 1, 3]) @@ -3346,19 +3257,20 @@ def forward( # pd_op.reshape: (1x21x1024xf32) <- (1x21x16x64xf32, 3xi64) reshape_71 = paddle._C_ops.reshape(transpose_71, full_int_array_2) + del transpose_71 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_141 = paddle._C_ops.matmul(reshape_71, parameter_107, False, False) - del parameter_107 + del parameter_107, reshape_71 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_159 = paddle._C_ops.add(matmul_141, parameter_106) - del parameter_106 + del matmul_141, parameter_106 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_106, dropout_107 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_159, None, full_4, False, "upscale_in_train", 0, False + add_159, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -3366,6 +3278,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_160 = paddle._C_ops.add(layer_norm_102, dropout_106) + del dropout_106, layer_norm_102 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_105, layer_norm_106, layer_norm_107 = (lambda x, f: f(x))( @@ -3374,7 +3287,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_100, parameter_101 + del add_160, parameter_100, parameter_101 # pd_op.matmul: (1x21x4096xf32) <- (1x21x1024xf32, 1024x4096xf32) matmul_142 = paddle._C_ops.matmul(layer_norm_105, parameter_105, False, False) @@ -3382,23 +3295,24 @@ def forward( # pd_op.add: (1x21x4096xf32) <- (1x21x4096xf32, 4096xf32) add_161 = paddle._C_ops.add(matmul_142, parameter_104) - del parameter_104 + del matmul_142, parameter_104 # pd_op.gelu: (1x21x4096xf32) <- (1x21x4096xf32) gelu_17 = paddle._C_ops.gelu(add_161, False) + del add_161 # pd_op.matmul: (1x21x1024xf32) <- (1x21x4096xf32, 4096x1024xf32) matmul_143 = paddle._C_ops.matmul(gelu_17, parameter_103, False, False) - del parameter_103 + del gelu_17, parameter_103 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_162 = paddle._C_ops.add(matmul_143, parameter_102) - del parameter_102 + del matmul_143, parameter_102 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_108, dropout_109 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_162, None, full_4, False, "upscale_in_train", 0, False + add_162, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -3406,6 +3320,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_163 = paddle._C_ops.add(layer_norm_105, dropout_108) + del dropout_108, layer_norm_105 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_108, layer_norm_109, layer_norm_110 = (lambda x, f: f(x))( @@ -3414,7 +3329,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_98, parameter_99 + del add_163, parameter_98, parameter_99 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_144 = paddle._C_ops.matmul(layer_norm_108, parameter_97, False, False) @@ -3422,10 +3337,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_164 = paddle._C_ops.add(matmul_144, parameter_96) - del parameter_96 + del matmul_144, parameter_96 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_72 = paddle._C_ops.reshape(add_164, full_int_array_1) + del add_164 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_72 = paddle._C_ops.transpose(reshape_72, [0, 2, 1, 3]) @@ -3437,7 +3353,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_165 = paddle._C_ops.add(matmul_145, parameter_94) - del parameter_94 + del matmul_145, parameter_94 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_146 = paddle._C_ops.matmul(layer_norm_108, parameter_93, False, False) @@ -3445,10 +3361,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_166 = paddle._C_ops.add(matmul_146, parameter_92) - del parameter_92 + del matmul_146, parameter_92 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_73 = paddle._C_ops.reshape(add_165, full_int_array_1) + del add_165 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_73 = paddle._C_ops.transpose(reshape_73, [0, 2, 1, 3]) @@ -3456,6 +3373,7 @@ def forward( # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_74 = paddle._C_ops.reshape(add_166, full_int_array_1) + del add_166 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_74 = paddle._C_ops.transpose(reshape_74, [0, 2, 1, 3]) @@ -3467,9 +3385,11 @@ def forward( # pd_op.matmul: (1x16x21x21xf32) <- (1x16x21x64xf32, 1x16x21x64xf32) matmul_147 = paddle._C_ops.matmul(scale_19, transpose_73, False, True) + del scale_19, transpose_73 # pd_op.add: (1x16x21x21xf32) <- (1x16x21x21xf32, 1x1x1x21xf32) add_167 = paddle._C_ops.add(matmul_147, unsqueeze_0) + del matmul_147 # pd_op.softmax: (1x16x21x21xf32) <- (1x16x21x21xf32) softmax_18 = paddle._C_ops.softmax(add_167, -1) @@ -3478,13 +3398,15 @@ def forward( # pd_op.dropout: (1x16x21x21xf32, 1x16x21x21xui8) <- (1x16x21x21xf32, None, 1xf32) dropout_110, dropout_111 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_18, None, full_4, False, "upscale_in_train", 0, False + softmax_18, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_18 # pd_op.matmul: (1x16x21x64xf32) <- (1x16x21x21xf32, 1x16x21x64xf32) matmul_148 = paddle._C_ops.matmul(dropout_110, transpose_74, False, False) + del dropout_110, transpose_74 # pd_op.transpose: (1x21x16x64xf32) <- (1x16x21x64xf32) transpose_75 = paddle._C_ops.transpose(matmul_148, [0, 2, 1, 3]) @@ -3492,19 +3414,20 @@ def forward( # pd_op.reshape: (1x21x1024xf32) <- (1x21x16x64xf32, 3xi64) reshape_75 = paddle._C_ops.reshape(transpose_75, full_int_array_2) + del transpose_75 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_149 = paddle._C_ops.matmul(reshape_75, parameter_91, False, False) - del parameter_91 + del parameter_91, reshape_75 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_168 = paddle._C_ops.add(matmul_149, parameter_90) - del parameter_90 + del matmul_149, parameter_90 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_112, dropout_113 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_168, None, full_4, False, "upscale_in_train", 0, False + add_168, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -3512,6 +3435,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_169 = paddle._C_ops.add(layer_norm_108, dropout_112) + del dropout_112, layer_norm_108 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_111, layer_norm_112, layer_norm_113 = (lambda x, f: f(x))( @@ -3520,7 +3444,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_84, parameter_85 + del add_169, parameter_84, parameter_85 # pd_op.matmul: (1x21x4096xf32) <- (1x21x1024xf32, 1024x4096xf32) matmul_150 = paddle._C_ops.matmul(layer_norm_111, parameter_89, False, False) @@ -3528,23 +3452,24 @@ def forward( # pd_op.add: (1x21x4096xf32) <- (1x21x4096xf32, 4096xf32) add_170 = paddle._C_ops.add(matmul_150, parameter_88) - del parameter_88 + del matmul_150, parameter_88 # pd_op.gelu: (1x21x4096xf32) <- (1x21x4096xf32) gelu_18 = paddle._C_ops.gelu(add_170, False) + del add_170 # pd_op.matmul: (1x21x1024xf32) <- (1x21x4096xf32, 4096x1024xf32) matmul_151 = paddle._C_ops.matmul(gelu_18, parameter_87, False, False) - del parameter_87 + del gelu_18, parameter_87 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_171 = paddle._C_ops.add(matmul_151, parameter_86) - del parameter_86 + del matmul_151, parameter_86 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_114, dropout_115 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_171, None, full_4, False, "upscale_in_train", 0, False + add_171, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -3552,6 +3477,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_172 = paddle._C_ops.add(layer_norm_111, dropout_114) + del dropout_114, layer_norm_111 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_114, layer_norm_115, layer_norm_116 = (lambda x, f: f(x))( @@ -3560,7 +3486,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_82, parameter_83 + del add_172, parameter_82, parameter_83 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_152 = paddle._C_ops.matmul(layer_norm_114, parameter_81, False, False) @@ -3568,10 +3494,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_173 = paddle._C_ops.add(matmul_152, parameter_80) - del parameter_80 + del matmul_152, parameter_80 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_76 = paddle._C_ops.reshape(add_173, full_int_array_1) + del add_173 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_76 = paddle._C_ops.transpose(reshape_76, [0, 2, 1, 3]) @@ -3583,7 +3510,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_174 = paddle._C_ops.add(matmul_153, parameter_78) - del parameter_78 + del matmul_153, parameter_78 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_154 = paddle._C_ops.matmul(layer_norm_114, parameter_77, False, False) @@ -3591,10 +3518,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_175 = paddle._C_ops.add(matmul_154, parameter_76) - del parameter_76 + del matmul_154, parameter_76 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_77 = paddle._C_ops.reshape(add_174, full_int_array_1) + del add_174 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_77 = paddle._C_ops.transpose(reshape_77, [0, 2, 1, 3]) @@ -3602,6 +3530,7 @@ def forward( # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_78 = paddle._C_ops.reshape(add_175, full_int_array_1) + del add_175 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_78 = paddle._C_ops.transpose(reshape_78, [0, 2, 1, 3]) @@ -3613,9 +3542,11 @@ def forward( # pd_op.matmul: (1x16x21x21xf32) <- (1x16x21x64xf32, 1x16x21x64xf32) matmul_155 = paddle._C_ops.matmul(scale_20, transpose_77, False, True) + del scale_20, transpose_77 # pd_op.add: (1x16x21x21xf32) <- (1x16x21x21xf32, 1x1x1x21xf32) add_176 = paddle._C_ops.add(matmul_155, unsqueeze_0) + del matmul_155 # pd_op.softmax: (1x16x21x21xf32) <- (1x16x21x21xf32) softmax_19 = paddle._C_ops.softmax(add_176, -1) @@ -3624,13 +3555,15 @@ def forward( # pd_op.dropout: (1x16x21x21xf32, 1x16x21x21xui8) <- (1x16x21x21xf32, None, 1xf32) dropout_116, dropout_117 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_19, None, full_4, False, "upscale_in_train", 0, False + softmax_19, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_19 # pd_op.matmul: (1x16x21x64xf32) <- (1x16x21x21xf32, 1x16x21x64xf32) matmul_156 = paddle._C_ops.matmul(dropout_116, transpose_78, False, False) + del dropout_116, transpose_78 # pd_op.transpose: (1x21x16x64xf32) <- (1x16x21x64xf32) transpose_79 = paddle._C_ops.transpose(matmul_156, [0, 2, 1, 3]) @@ -3638,19 +3571,20 @@ def forward( # pd_op.reshape: (1x21x1024xf32) <- (1x21x16x64xf32, 3xi64) reshape_79 = paddle._C_ops.reshape(transpose_79, full_int_array_2) + del transpose_79 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_157 = paddle._C_ops.matmul(reshape_79, parameter_75, False, False) - del parameter_75 + del parameter_75, reshape_79 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_177 = paddle._C_ops.add(matmul_157, parameter_74) - del parameter_74 + del matmul_157, parameter_74 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_118, dropout_119 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_177, None, full_4, False, "upscale_in_train", 0, False + add_177, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -3658,6 +3592,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_178 = paddle._C_ops.add(layer_norm_114, dropout_118) + del dropout_118, layer_norm_114 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_117, layer_norm_118, layer_norm_119 = (lambda x, f: f(x))( @@ -3666,7 +3601,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_68, parameter_69 + del add_178, parameter_68, parameter_69 # pd_op.matmul: (1x21x4096xf32) <- (1x21x1024xf32, 1024x4096xf32) matmul_158 = paddle._C_ops.matmul(layer_norm_117, parameter_73, False, False) @@ -3674,23 +3609,24 @@ def forward( # pd_op.add: (1x21x4096xf32) <- (1x21x4096xf32, 4096xf32) add_179 = paddle._C_ops.add(matmul_158, parameter_72) - del parameter_72 + del matmul_158, parameter_72 # pd_op.gelu: (1x21x4096xf32) <- (1x21x4096xf32) gelu_19 = paddle._C_ops.gelu(add_179, False) + del add_179 # pd_op.matmul: (1x21x1024xf32) <- (1x21x4096xf32, 4096x1024xf32) matmul_159 = paddle._C_ops.matmul(gelu_19, parameter_71, False, False) - del parameter_71 + del gelu_19, parameter_71 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_180 = paddle._C_ops.add(matmul_159, parameter_70) - del parameter_70 + del matmul_159, parameter_70 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_120, dropout_121 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_180, None, full_4, False, "upscale_in_train", 0, False + add_180, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -3698,6 +3634,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_181 = paddle._C_ops.add(layer_norm_117, dropout_120) + del dropout_120, layer_norm_117 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_120, layer_norm_121, layer_norm_122 = (lambda x, f: f(x))( @@ -3706,7 +3643,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_66, parameter_67 + del add_181, parameter_66, parameter_67 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_160 = paddle._C_ops.matmul(layer_norm_120, parameter_65, False, False) @@ -3714,10 +3651,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_182 = paddle._C_ops.add(matmul_160, parameter_64) - del parameter_64 + del matmul_160, parameter_64 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_80 = paddle._C_ops.reshape(add_182, full_int_array_1) + del add_182 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_80 = paddle._C_ops.transpose(reshape_80, [0, 2, 1, 3]) @@ -3729,7 +3667,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_183 = paddle._C_ops.add(matmul_161, parameter_62) - del parameter_62 + del matmul_161, parameter_62 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_162 = paddle._C_ops.matmul(layer_norm_120, parameter_61, False, False) @@ -3737,10 +3675,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_184 = paddle._C_ops.add(matmul_162, parameter_60) - del parameter_60 + del matmul_162, parameter_60 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_81 = paddle._C_ops.reshape(add_183, full_int_array_1) + del add_183 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_81 = paddle._C_ops.transpose(reshape_81, [0, 2, 1, 3]) @@ -3748,6 +3687,7 @@ def forward( # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_82 = paddle._C_ops.reshape(add_184, full_int_array_1) + del add_184 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_82 = paddle._C_ops.transpose(reshape_82, [0, 2, 1, 3]) @@ -3759,9 +3699,11 @@ def forward( # pd_op.matmul: (1x16x21x21xf32) <- (1x16x21x64xf32, 1x16x21x64xf32) matmul_163 = paddle._C_ops.matmul(scale_21, transpose_81, False, True) + del scale_21, transpose_81 # pd_op.add: (1x16x21x21xf32) <- (1x16x21x21xf32, 1x1x1x21xf32) add_185 = paddle._C_ops.add(matmul_163, unsqueeze_0) + del matmul_163 # pd_op.softmax: (1x16x21x21xf32) <- (1x16x21x21xf32) softmax_20 = paddle._C_ops.softmax(add_185, -1) @@ -3770,13 +3712,15 @@ def forward( # pd_op.dropout: (1x16x21x21xf32, 1x16x21x21xui8) <- (1x16x21x21xf32, None, 1xf32) dropout_122, dropout_123 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_20, None, full_4, False, "upscale_in_train", 0, False + softmax_20, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_20 # pd_op.matmul: (1x16x21x64xf32) <- (1x16x21x21xf32, 1x16x21x64xf32) matmul_164 = paddle._C_ops.matmul(dropout_122, transpose_82, False, False) + del dropout_122, transpose_82 # pd_op.transpose: (1x21x16x64xf32) <- (1x16x21x64xf32) transpose_83 = paddle._C_ops.transpose(matmul_164, [0, 2, 1, 3]) @@ -3784,19 +3728,20 @@ def forward( # pd_op.reshape: (1x21x1024xf32) <- (1x21x16x64xf32, 3xi64) reshape_83 = paddle._C_ops.reshape(transpose_83, full_int_array_2) + del transpose_83 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_165 = paddle._C_ops.matmul(reshape_83, parameter_59, False, False) - del parameter_59 + del parameter_59, reshape_83 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_186 = paddle._C_ops.add(matmul_165, parameter_58) - del parameter_58 + del matmul_165, parameter_58 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_124, dropout_125 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_186, None, full_4, False, "upscale_in_train", 0, False + add_186, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -3804,6 +3749,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_187 = paddle._C_ops.add(layer_norm_120, dropout_124) + del dropout_124, layer_norm_120 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_123, layer_norm_124, layer_norm_125 = (lambda x, f: f(x))( @@ -3812,7 +3758,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_52, parameter_53 + del add_187, parameter_52, parameter_53 # pd_op.matmul: (1x21x4096xf32) <- (1x21x1024xf32, 1024x4096xf32) matmul_166 = paddle._C_ops.matmul(layer_norm_123, parameter_57, False, False) @@ -3820,23 +3766,24 @@ def forward( # pd_op.add: (1x21x4096xf32) <- (1x21x4096xf32, 4096xf32) add_188 = paddle._C_ops.add(matmul_166, parameter_56) - del parameter_56 + del matmul_166, parameter_56 # pd_op.gelu: (1x21x4096xf32) <- (1x21x4096xf32) gelu_20 = paddle._C_ops.gelu(add_188, False) + del add_188 # pd_op.matmul: (1x21x1024xf32) <- (1x21x4096xf32, 4096x1024xf32) matmul_167 = paddle._C_ops.matmul(gelu_20, parameter_55, False, False) - del parameter_55 + del gelu_20, parameter_55 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_189 = paddle._C_ops.add(matmul_167, parameter_54) - del parameter_54 + del matmul_167, parameter_54 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_126, dropout_127 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_189, None, full_4, False, "upscale_in_train", 0, False + add_189, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -3844,6 +3791,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_190 = paddle._C_ops.add(layer_norm_123, dropout_126) + del dropout_126, layer_norm_123 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_126, layer_norm_127, layer_norm_128 = (lambda x, f: f(x))( @@ -3852,7 +3800,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_50, parameter_51 + del add_190, parameter_50, parameter_51 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_168 = paddle._C_ops.matmul(layer_norm_126, parameter_49, False, False) @@ -3860,10 +3808,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_191 = paddle._C_ops.add(matmul_168, parameter_48) - del parameter_48 + del matmul_168, parameter_48 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_84 = paddle._C_ops.reshape(add_191, full_int_array_1) + del add_191 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_84 = paddle._C_ops.transpose(reshape_84, [0, 2, 1, 3]) @@ -3875,7 +3824,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_192 = paddle._C_ops.add(matmul_169, parameter_46) - del parameter_46 + del matmul_169, parameter_46 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_170 = paddle._C_ops.matmul(layer_norm_126, parameter_45, False, False) @@ -3883,10 +3832,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_193 = paddle._C_ops.add(matmul_170, parameter_44) - del parameter_44 + del matmul_170, parameter_44 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_85 = paddle._C_ops.reshape(add_192, full_int_array_1) + del add_192 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_85 = paddle._C_ops.transpose(reshape_85, [0, 2, 1, 3]) @@ -3894,6 +3844,7 @@ def forward( # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_86 = paddle._C_ops.reshape(add_193, full_int_array_1) + del add_193 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_86 = paddle._C_ops.transpose(reshape_86, [0, 2, 1, 3]) @@ -3905,9 +3856,11 @@ def forward( # pd_op.matmul: (1x16x21x21xf32) <- (1x16x21x64xf32, 1x16x21x64xf32) matmul_171 = paddle._C_ops.matmul(scale_22, transpose_85, False, True) + del scale_22, transpose_85 # pd_op.add: (1x16x21x21xf32) <- (1x16x21x21xf32, 1x1x1x21xf32) add_194 = paddle._C_ops.add(matmul_171, unsqueeze_0) + del matmul_171 # pd_op.softmax: (1x16x21x21xf32) <- (1x16x21x21xf32) softmax_21 = paddle._C_ops.softmax(add_194, -1) @@ -3916,13 +3869,15 @@ def forward( # pd_op.dropout: (1x16x21x21xf32, 1x16x21x21xui8) <- (1x16x21x21xf32, None, 1xf32) dropout_128, dropout_129 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_21, None, full_4, False, "upscale_in_train", 0, False + softmax_21, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_21 # pd_op.matmul: (1x16x21x64xf32) <- (1x16x21x21xf32, 1x16x21x64xf32) matmul_172 = paddle._C_ops.matmul(dropout_128, transpose_86, False, False) + del dropout_128, transpose_86 # pd_op.transpose: (1x21x16x64xf32) <- (1x16x21x64xf32) transpose_87 = paddle._C_ops.transpose(matmul_172, [0, 2, 1, 3]) @@ -3930,19 +3885,20 @@ def forward( # pd_op.reshape: (1x21x1024xf32) <- (1x21x16x64xf32, 3xi64) reshape_87 = paddle._C_ops.reshape(transpose_87, full_int_array_2) + del transpose_87 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_173 = paddle._C_ops.matmul(reshape_87, parameter_43, False, False) - del parameter_43 + del parameter_43, reshape_87 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_195 = paddle._C_ops.add(matmul_173, parameter_42) - del parameter_42 + del matmul_173, parameter_42 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_130, dropout_131 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_195, None, full_4, False, "upscale_in_train", 0, False + add_195, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -3950,6 +3906,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_196 = paddle._C_ops.add(layer_norm_126, dropout_130) + del dropout_130, layer_norm_126 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_129, layer_norm_130, layer_norm_131 = (lambda x, f: f(x))( @@ -3958,7 +3915,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_36, parameter_37 + del add_196, parameter_36, parameter_37 # pd_op.matmul: (1x21x4096xf32) <- (1x21x1024xf32, 1024x4096xf32) matmul_174 = paddle._C_ops.matmul(layer_norm_129, parameter_41, False, False) @@ -3966,23 +3923,24 @@ def forward( # pd_op.add: (1x21x4096xf32) <- (1x21x4096xf32, 4096xf32) add_197 = paddle._C_ops.add(matmul_174, parameter_40) - del parameter_40 + del matmul_174, parameter_40 # pd_op.gelu: (1x21x4096xf32) <- (1x21x4096xf32) gelu_21 = paddle._C_ops.gelu(add_197, False) + del add_197 # pd_op.matmul: (1x21x1024xf32) <- (1x21x4096xf32, 4096x1024xf32) matmul_175 = paddle._C_ops.matmul(gelu_21, parameter_39, False, False) - del parameter_39 + del gelu_21, parameter_39 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_198 = paddle._C_ops.add(matmul_175, parameter_38) - del parameter_38 + del matmul_175, parameter_38 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_132, dropout_133 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_198, None, full_4, False, "upscale_in_train", 0, False + add_198, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -3990,6 +3948,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_199 = paddle._C_ops.add(layer_norm_129, dropout_132) + del dropout_132, layer_norm_129 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_132, layer_norm_133, layer_norm_134 = (lambda x, f: f(x))( @@ -3998,7 +3957,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_34, parameter_35 + del add_199, parameter_34, parameter_35 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_176 = paddle._C_ops.matmul(layer_norm_132, parameter_33, False, False) @@ -4006,10 +3965,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_200 = paddle._C_ops.add(matmul_176, parameter_32) - del parameter_32 + del matmul_176, parameter_32 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_88 = paddle._C_ops.reshape(add_200, full_int_array_1) + del add_200 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_88 = paddle._C_ops.transpose(reshape_88, [0, 2, 1, 3]) @@ -4021,7 +3981,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_201 = paddle._C_ops.add(matmul_177, parameter_30) - del parameter_30 + del matmul_177, parameter_30 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_178 = paddle._C_ops.matmul(layer_norm_132, parameter_29, False, False) @@ -4029,10 +3989,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_202 = paddle._C_ops.add(matmul_178, parameter_28) - del parameter_28 + del matmul_178, parameter_28 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_89 = paddle._C_ops.reshape(add_201, full_int_array_1) + del add_201 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_89 = paddle._C_ops.transpose(reshape_89, [0, 2, 1, 3]) @@ -4040,6 +4001,7 @@ def forward( # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_90 = paddle._C_ops.reshape(add_202, full_int_array_1) + del add_202 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_90 = paddle._C_ops.transpose(reshape_90, [0, 2, 1, 3]) @@ -4051,9 +4013,11 @@ def forward( # pd_op.matmul: (1x16x21x21xf32) <- (1x16x21x64xf32, 1x16x21x64xf32) matmul_179 = paddle._C_ops.matmul(scale_23, transpose_89, False, True) + del scale_23, transpose_89 # pd_op.add: (1x16x21x21xf32) <- (1x16x21x21xf32, 1x1x1x21xf32) add_203 = paddle._C_ops.add(matmul_179, unsqueeze_0) + del matmul_179 # pd_op.softmax: (1x16x21x21xf32) <- (1x16x21x21xf32) softmax_22 = paddle._C_ops.softmax(add_203, -1) @@ -4062,13 +4026,15 @@ def forward( # pd_op.dropout: (1x16x21x21xf32, 1x16x21x21xui8) <- (1x16x21x21xf32, None, 1xf32) dropout_134, dropout_135 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_22, None, full_4, False, "upscale_in_train", 0, False + softmax_22, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_22 # pd_op.matmul: (1x16x21x64xf32) <- (1x16x21x21xf32, 1x16x21x64xf32) matmul_180 = paddle._C_ops.matmul(dropout_134, transpose_90, False, False) + del dropout_134, transpose_90 # pd_op.transpose: (1x21x16x64xf32) <- (1x16x21x64xf32) transpose_91 = paddle._C_ops.transpose(matmul_180, [0, 2, 1, 3]) @@ -4076,19 +4042,20 @@ def forward( # pd_op.reshape: (1x21x1024xf32) <- (1x21x16x64xf32, 3xi64) reshape_91 = paddle._C_ops.reshape(transpose_91, full_int_array_2) + del transpose_91 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_181 = paddle._C_ops.matmul(reshape_91, parameter_27, False, False) - del parameter_27 + del parameter_27, reshape_91 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_204 = paddle._C_ops.add(matmul_181, parameter_26) - del parameter_26 + del matmul_181, parameter_26 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_136, dropout_137 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_204, None, full_4, False, "upscale_in_train", 0, False + add_204, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -4096,6 +4063,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_205 = paddle._C_ops.add(layer_norm_132, dropout_136) + del dropout_136, layer_norm_132 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_135, layer_norm_136, layer_norm_137 = (lambda x, f: f(x))( @@ -4104,7 +4072,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_20, parameter_21 + del add_205, parameter_20, parameter_21 # pd_op.matmul: (1x21x4096xf32) <- (1x21x1024xf32, 1024x4096xf32) matmul_182 = paddle._C_ops.matmul(layer_norm_135, parameter_25, False, False) @@ -4112,23 +4080,24 @@ def forward( # pd_op.add: (1x21x4096xf32) <- (1x21x4096xf32, 4096xf32) add_206 = paddle._C_ops.add(matmul_182, parameter_24) - del parameter_24 + del matmul_182, parameter_24 # pd_op.gelu: (1x21x4096xf32) <- (1x21x4096xf32) gelu_22 = paddle._C_ops.gelu(add_206, False) + del add_206 # pd_op.matmul: (1x21x1024xf32) <- (1x21x4096xf32, 4096x1024xf32) matmul_183 = paddle._C_ops.matmul(gelu_22, parameter_23, False, False) - del parameter_23 + del gelu_22, parameter_23 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_207 = paddle._C_ops.add(matmul_183, parameter_22) - del parameter_22 + del matmul_183, parameter_22 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_138, dropout_139 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_207, None, full_4, False, "upscale_in_train", 0, False + add_207, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -4136,6 +4105,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_208 = paddle._C_ops.add(layer_norm_135, dropout_138) + del dropout_138, layer_norm_135 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_138, layer_norm_139, layer_norm_140 = (lambda x, f: f(x))( @@ -4144,7 +4114,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_18, parameter_19 + del add_208, parameter_18, parameter_19 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_184 = paddle._C_ops.matmul(layer_norm_138, parameter_17, False, False) @@ -4152,10 +4122,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_209 = paddle._C_ops.add(matmul_184, parameter_16) - del parameter_16 + del matmul_184, parameter_16 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_92 = paddle._C_ops.reshape(add_209, full_int_array_1) + del add_209 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_92 = paddle._C_ops.transpose(reshape_92, [0, 2, 1, 3]) @@ -4167,7 +4138,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_210 = paddle._C_ops.add(matmul_185, parameter_14) - del parameter_14 + del matmul_185, parameter_14 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_186 = paddle._C_ops.matmul(layer_norm_138, parameter_13, False, False) @@ -4175,10 +4146,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_211 = paddle._C_ops.add(matmul_186, parameter_12) - del parameter_12 + del matmul_186, parameter_12 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_93 = paddle._C_ops.reshape(add_210, full_int_array_1) + del add_210 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_93 = paddle._C_ops.transpose(reshape_93, [0, 2, 1, 3]) @@ -4186,7 +4158,7 @@ def forward( # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_94 = paddle._C_ops.reshape(add_211, full_int_array_1) - del full_int_array_1 + del add_211, full_int_array_1 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_94 = paddle._C_ops.transpose(reshape_94, [0, 2, 1, 3]) @@ -4194,13 +4166,15 @@ def forward( # pd_op.scale: (1x16x21x64xf32) <- (1x16x21x64xf32, 1xf32) scale_24 = paddle._C_ops.scale(transpose_92, full_5, float("0"), True) - del transpose_92 + del full_5, transpose_92 # pd_op.matmul: (1x16x21x21xf32) <- (1x16x21x64xf32, 1x16x21x64xf32) matmul_187 = paddle._C_ops.matmul(scale_24, transpose_93, False, True) + del scale_24, transpose_93 # pd_op.add: (1x16x21x21xf32) <- (1x16x21x21xf32, 1x1x1x21xf32) add_212 = paddle._C_ops.add(matmul_187, unsqueeze_0) + del matmul_187, unsqueeze_0 # pd_op.softmax: (1x16x21x21xf32) <- (1x16x21x21xf32) softmax_23 = paddle._C_ops.softmax(add_212, -1) @@ -4209,13 +4183,15 @@ def forward( # pd_op.dropout: (1x16x21x21xf32, 1x16x21x21xui8) <- (1x16x21x21xf32, None, 1xf32) dropout_140, dropout_141 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_23, None, full_4, False, "upscale_in_train", 0, False + softmax_23, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_23 # pd_op.matmul: (1x16x21x64xf32) <- (1x16x21x21xf32, 1x16x21x64xf32) matmul_188 = paddle._C_ops.matmul(dropout_140, transpose_94, False, False) + del dropout_140, transpose_94 # pd_op.transpose: (1x21x16x64xf32) <- (1x16x21x64xf32) transpose_95 = paddle._C_ops.transpose(matmul_188, [0, 2, 1, 3]) @@ -4223,20 +4199,20 @@ def forward( # pd_op.reshape: (1x21x1024xf32) <- (1x21x16x64xf32, 3xi64) reshape_95 = paddle._C_ops.reshape(transpose_95, full_int_array_2) - del full_int_array_2 + del full_int_array_2, transpose_95 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_189 = paddle._C_ops.matmul(reshape_95, parameter_11, False, False) - del parameter_11 + del parameter_11, reshape_95 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_213 = paddle._C_ops.add(matmul_189, parameter_10) - del parameter_10 + del matmul_189, parameter_10 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_142, dropout_143 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_213, None, full_4, False, "upscale_in_train", 0, False + add_213, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -4244,6 +4220,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_214 = paddle._C_ops.add(layer_norm_138, dropout_142) + del dropout_142, layer_norm_138 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_141, layer_norm_142, layer_norm_143 = (lambda x, f: f(x))( @@ -4252,7 +4229,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_4, parameter_5 + del add_214, parameter_4, parameter_5 # pd_op.matmul: (1x21x4096xf32) <- (1x21x1024xf32, 1024x4096xf32) matmul_190 = paddle._C_ops.matmul(layer_norm_141, parameter_9, False, False) @@ -4260,30 +4237,32 @@ def forward( # pd_op.add: (1x21x4096xf32) <- (1x21x4096xf32, 4096xf32) add_215 = paddle._C_ops.add(matmul_190, parameter_8) - del parameter_8 + del matmul_190, parameter_8 # pd_op.gelu: (1x21x4096xf32) <- (1x21x4096xf32) gelu_23 = paddle._C_ops.gelu(add_215, False) + del add_215 # pd_op.matmul: (1x21x1024xf32) <- (1x21x4096xf32, 4096x1024xf32) matmul_191 = paddle._C_ops.matmul(gelu_23, parameter_7, False, False) - del parameter_7 + del gelu_23, parameter_7 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_216 = paddle._C_ops.add(matmul_191, parameter_6) - del parameter_6 + del matmul_191, parameter_6 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_144, dropout_145 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_216, None, full_4, False, "upscale_in_train", 0, False + add_216, None, full_4, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) - del add_216 + del add_216, full_4 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_217 = paddle._C_ops.add(layer_norm_141, dropout_144) + del dropout_144, layer_norm_141 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_144, layer_norm_145, layer_norm_146 = (lambda x, f: f(x))( @@ -4292,7 +4271,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_2, parameter_3 + del add_217, parameter_2, parameter_3 # pd_op.full_int_array: (1xi64) <- () full_int_array_3 = [0] @@ -4304,899 +4283,18 @@ def forward( slice_0 = paddle._C_ops.slice( layer_norm_144, [1], full_int_array_3, full_int_array_4, [1], [1] ) + del full_int_array_3, full_int_array_4, layer_norm_144 # pd_op.matmul: (1x1024xf32) <- (1x1024xf32, 1024x1024xf32) matmul_192 = paddle._C_ops.matmul(slice_0, parameter_1, False, False) - del parameter_1 + del parameter_1, slice_0 # pd_op.add: (1x1024xf32) <- (1x1024xf32, 1024xf32) add_218 = paddle._C_ops.add(matmul_192, parameter_0) - del parameter_0 + del matmul_192, parameter_0 # pd_op.tanh: (1x1024xf32) <- (1x1024xf32) tanh_0 = paddle._C_ops.tanh(add_218) - del ( - add_0, - add_1, - add_10, - add_100, - add_101, - add_102, - add_103, - add_106, - add_107, - add_109, - add_11, - add_110, - add_111, - add_112, - add_115, - add_116, - add_118, - add_119, - add_12, - add_120, - add_121, - add_124, - add_125, - add_127, - add_128, - add_129, - add_13, - add_130, - add_133, - add_134, - add_136, - add_137, - add_138, - add_139, - add_142, - add_143, - add_145, - add_146, - add_147, - add_148, - add_151, - add_152, - add_154, - add_155, - add_156, - add_157, - add_16, - add_160, - add_161, - add_163, - add_164, - add_165, - add_166, - add_169, - add_17, - add_170, - add_172, - add_173, - add_174, - add_175, - add_178, - add_179, - add_181, - add_182, - add_183, - add_184, - add_187, - add_188, - add_19, - add_190, - add_191, - add_192, - add_193, - add_196, - add_197, - add_199, - add_2, - add_20, - add_200, - add_201, - add_202, - add_205, - add_206, - add_208, - add_209, - add_21, - add_210, - add_211, - add_214, - add_215, - add_217, - add_218, - add_22, - add_25, - add_26, - add_28, - add_29, - add_3, - add_30, - add_31, - add_34, - add_35, - add_37, - add_38, - add_39, - add_4, - add_40, - add_43, - add_44, - add_46, - add_47, - add_48, - add_49, - add_52, - add_53, - add_55, - add_56, - add_57, - add_58, - add_61, - add_62, - add_64, - add_65, - add_66, - add_67, - add_7, - add_70, - add_71, - add_73, - add_74, - add_75, - add_76, - add_79, - add_8, - add_80, - add_82, - add_83, - add_84, - add_85, - add_88, - add_89, - add_91, - add_92, - add_93, - add_94, - add_97, - add_98, - assign_0, - assign_1, - assign_10, - assign_11, - assign_12, - assign_13, - assign_14, - assign_15, - assign_16, - assign_17, - assign_18, - assign_19, - assign_2, - assign_20, - assign_21, - assign_22, - assign_23, - assign_24, - assign_25, - assign_26, - assign_27, - assign_28, - assign_29, - assign_3, - assign_30, - assign_31, - assign_32, - assign_33, - assign_34, - assign_35, - assign_36, - assign_37, - assign_38, - assign_39, - assign_4, - assign_40, - assign_41, - assign_42, - assign_43, - assign_44, - assign_45, - assign_46, - assign_47, - assign_48, - assign_49, - assign_5, - assign_50, - assign_51, - assign_52, - assign_53, - assign_54, - assign_55, - assign_56, - assign_57, - assign_58, - assign_59, - assign_6, - assign_60, - assign_61, - assign_62, - assign_63, - assign_64, - assign_65, - assign_66, - assign_67, - assign_68, - assign_69, - assign_7, - assign_70, - assign_71, - assign_72, - assign_73, - assign_74, - assign_75, - assign_76, - assign_77, - assign_78, - assign_79, - assign_8, - assign_80, - assign_81, - assign_82, - assign_83, - assign_84, - assign_85, - assign_86, - assign_87, - assign_88, - assign_89, - assign_9, - assign_90, - assign_91, - assign_92, - assign_93, - assign_94, - dropout_0, - dropout_1, - dropout_10, - dropout_100, - dropout_101, - dropout_102, - dropout_103, - dropout_104, - dropout_105, - dropout_106, - dropout_107, - dropout_108, - dropout_109, - dropout_11, - dropout_110, - dropout_111, - dropout_112, - dropout_113, - dropout_114, - dropout_115, - dropout_116, - dropout_117, - dropout_118, - dropout_119, - dropout_12, - dropout_120, - dropout_121, - dropout_122, - dropout_123, - dropout_124, - dropout_125, - dropout_126, - dropout_127, - dropout_128, - dropout_129, - dropout_13, - dropout_130, - dropout_131, - dropout_132, - dropout_133, - dropout_134, - dropout_135, - dropout_136, - dropout_137, - dropout_138, - dropout_139, - dropout_14, - dropout_140, - dropout_141, - dropout_142, - dropout_143, - dropout_144, - dropout_145, - dropout_15, - dropout_16, - dropout_17, - dropout_18, - dropout_19, - dropout_2, - dropout_20, - dropout_21, - dropout_22, - dropout_23, - dropout_24, - dropout_25, - dropout_26, - dropout_27, - dropout_28, - dropout_29, - dropout_3, - dropout_30, - dropout_31, - dropout_32, - dropout_33, - dropout_34, - dropout_35, - dropout_36, - dropout_37, - dropout_38, - dropout_39, - dropout_4, - dropout_40, - dropout_41, - dropout_42, - dropout_43, - dropout_44, - dropout_45, - dropout_46, - dropout_47, - dropout_48, - dropout_49, - dropout_5, - dropout_50, - dropout_51, - dropout_52, - dropout_53, - dropout_54, - dropout_55, - dropout_56, - dropout_57, - dropout_58, - dropout_59, - dropout_6, - dropout_60, - dropout_61, - dropout_62, - dropout_63, - dropout_64, - dropout_65, - dropout_66, - dropout_67, - dropout_68, - dropout_69, - dropout_7, - dropout_70, - dropout_71, - dropout_72, - dropout_73, - dropout_74, - dropout_75, - dropout_76, - dropout_77, - dropout_78, - dropout_79, - dropout_8, - dropout_80, - dropout_81, - dropout_82, - dropout_83, - dropout_84, - dropout_85, - dropout_86, - dropout_87, - dropout_88, - dropout_89, - dropout_9, - dropout_90, - dropout_91, - dropout_92, - dropout_93, - dropout_94, - dropout_95, - dropout_96, - dropout_97, - dropout_98, - dropout_99, - embedding_0, - embedding_1, - embedding_2, - full_4, - full_5, - full_int_array_3, - full_int_array_4, - gelu_0, - gelu_1, - gelu_10, - gelu_11, - gelu_12, - gelu_13, - gelu_14, - gelu_15, - gelu_16, - gelu_17, - gelu_18, - gelu_19, - gelu_2, - gelu_20, - gelu_21, - gelu_22, - gelu_23, - gelu_3, - gelu_4, - gelu_5, - gelu_6, - gelu_7, - gelu_8, - gelu_9, - layer_norm_1, - layer_norm_10, - layer_norm_100, - layer_norm_101, - layer_norm_102, - layer_norm_103, - layer_norm_104, - layer_norm_105, - layer_norm_106, - layer_norm_107, - layer_norm_108, - layer_norm_109, - layer_norm_11, - layer_norm_110, - layer_norm_111, - layer_norm_112, - layer_norm_113, - layer_norm_114, - layer_norm_115, - layer_norm_116, - layer_norm_117, - layer_norm_118, - layer_norm_119, - layer_norm_12, - layer_norm_120, - layer_norm_121, - layer_norm_122, - layer_norm_123, - layer_norm_124, - layer_norm_125, - layer_norm_126, - layer_norm_127, - layer_norm_128, - layer_norm_129, - layer_norm_13, - layer_norm_130, - layer_norm_131, - layer_norm_132, - layer_norm_133, - layer_norm_134, - layer_norm_135, - layer_norm_136, - layer_norm_137, - layer_norm_138, - layer_norm_139, - layer_norm_14, - layer_norm_140, - layer_norm_141, - layer_norm_142, - layer_norm_143, - layer_norm_144, - layer_norm_145, - layer_norm_146, - layer_norm_15, - layer_norm_16, - layer_norm_17, - layer_norm_18, - layer_norm_19, - layer_norm_2, - layer_norm_20, - layer_norm_21, - layer_norm_22, - layer_norm_23, - layer_norm_24, - layer_norm_25, - layer_norm_26, - layer_norm_27, - layer_norm_28, - layer_norm_29, - layer_norm_3, - layer_norm_30, - layer_norm_31, - layer_norm_32, - layer_norm_33, - layer_norm_34, - layer_norm_35, - layer_norm_36, - layer_norm_37, - layer_norm_38, - layer_norm_39, - layer_norm_4, - layer_norm_40, - layer_norm_41, - layer_norm_42, - layer_norm_43, - layer_norm_44, - layer_norm_45, - layer_norm_46, - layer_norm_47, - layer_norm_48, - layer_norm_49, - layer_norm_5, - layer_norm_50, - layer_norm_51, - layer_norm_52, - layer_norm_53, - layer_norm_54, - layer_norm_55, - layer_norm_56, - layer_norm_57, - layer_norm_58, - layer_norm_59, - layer_norm_6, - layer_norm_60, - layer_norm_61, - layer_norm_62, - layer_norm_63, - layer_norm_64, - layer_norm_65, - layer_norm_66, - layer_norm_67, - layer_norm_68, - layer_norm_69, - layer_norm_7, - layer_norm_70, - layer_norm_71, - layer_norm_72, - layer_norm_73, - layer_norm_74, - layer_norm_75, - layer_norm_76, - layer_norm_77, - layer_norm_78, - layer_norm_79, - layer_norm_8, - layer_norm_80, - layer_norm_81, - layer_norm_82, - layer_norm_83, - layer_norm_84, - layer_norm_85, - layer_norm_86, - layer_norm_87, - layer_norm_88, - layer_norm_89, - layer_norm_9, - layer_norm_90, - layer_norm_91, - layer_norm_92, - layer_norm_93, - layer_norm_94, - layer_norm_95, - layer_norm_96, - layer_norm_97, - layer_norm_98, - layer_norm_99, - matmul_0, - matmul_1, - matmul_10, - matmul_101, - matmul_102, - matmul_103, - matmul_104, - matmul_105, - matmul_106, - matmul_107, - matmul_109, - matmul_11, - matmul_110, - matmul_111, - matmul_112, - matmul_113, - matmul_114, - matmul_115, - matmul_117, - matmul_118, - matmul_119, - matmul_120, - matmul_121, - matmul_122, - matmul_123, - matmul_125, - matmul_126, - matmul_127, - matmul_128, - matmul_129, - matmul_13, - matmul_130, - matmul_131, - matmul_133, - matmul_134, - matmul_135, - matmul_136, - matmul_137, - matmul_138, - matmul_139, - matmul_14, - matmul_141, - matmul_142, - matmul_143, - matmul_144, - matmul_145, - matmul_146, - matmul_147, - matmul_149, - matmul_15, - matmul_150, - matmul_151, - matmul_152, - matmul_153, - matmul_154, - matmul_155, - matmul_157, - matmul_158, - matmul_159, - matmul_16, - matmul_160, - matmul_161, - matmul_162, - matmul_163, - matmul_165, - matmul_166, - matmul_167, - matmul_168, - matmul_169, - matmul_17, - matmul_170, - matmul_171, - matmul_173, - matmul_174, - matmul_175, - matmul_176, - matmul_177, - matmul_178, - matmul_179, - matmul_18, - matmul_181, - matmul_182, - matmul_183, - matmul_184, - matmul_185, - matmul_186, - matmul_187, - matmul_189, - matmul_19, - matmul_190, - matmul_191, - matmul_192, - matmul_2, - matmul_21, - matmul_22, - matmul_23, - matmul_24, - matmul_25, - matmul_26, - matmul_27, - matmul_29, - matmul_3, - matmul_30, - matmul_31, - matmul_32, - matmul_33, - matmul_34, - matmul_35, - matmul_37, - matmul_38, - matmul_39, - matmul_40, - matmul_41, - matmul_42, - matmul_43, - matmul_45, - matmul_46, - matmul_47, - matmul_48, - matmul_49, - matmul_5, - matmul_50, - matmul_51, - matmul_53, - matmul_54, - matmul_55, - matmul_56, - matmul_57, - matmul_58, - matmul_59, - matmul_6, - matmul_61, - matmul_62, - matmul_63, - matmul_64, - matmul_65, - matmul_66, - matmul_67, - matmul_69, - matmul_7, - matmul_70, - matmul_71, - matmul_72, - matmul_73, - matmul_74, - matmul_75, - matmul_77, - matmul_78, - matmul_79, - matmul_8, - matmul_80, - matmul_81, - matmul_82, - matmul_83, - matmul_85, - matmul_86, - matmul_87, - matmul_88, - matmul_89, - matmul_9, - matmul_90, - matmul_91, - matmul_93, - matmul_94, - matmul_95, - matmul_96, - matmul_97, - matmul_98, - matmul_99, - reshape_11, - reshape_15, - reshape_19, - reshape_23, - reshape_27, - reshape_3, - reshape_31, - reshape_35, - reshape_39, - reshape_43, - reshape_47, - reshape_51, - reshape_55, - reshape_59, - reshape_63, - reshape_67, - reshape_7, - reshape_71, - reshape_75, - reshape_79, - reshape_83, - reshape_87, - reshape_91, - reshape_95, - scale_1, - scale_10, - scale_11, - scale_12, - scale_13, - scale_14, - scale_15, - scale_16, - scale_17, - scale_18, - scale_19, - scale_2, - scale_20, - scale_21, - scale_22, - scale_23, - scale_24, - scale_3, - scale_4, - scale_5, - scale_6, - scale_7, - scale_8, - scale_9, - slice_0, - softmax_0, - softmax_1, - softmax_10, - softmax_11, - softmax_12, - softmax_13, - softmax_14, - softmax_15, - softmax_16, - softmax_17, - softmax_18, - softmax_19, - softmax_2, - softmax_20, - softmax_21, - softmax_22, - softmax_23, - softmax_3, - softmax_4, - softmax_5, - softmax_6, - softmax_7, - softmax_8, - softmax_9, - subtract_0, - transpose_1, - transpose_10, - transpose_11, - transpose_13, - transpose_14, - transpose_15, - transpose_17, - transpose_18, - transpose_19, - transpose_2, - transpose_21, - transpose_22, - transpose_23, - transpose_25, - transpose_26, - transpose_27, - transpose_29, - transpose_3, - transpose_30, - transpose_31, - transpose_33, - transpose_34, - transpose_35, - transpose_37, - transpose_38, - transpose_39, - transpose_41, - transpose_42, - transpose_43, - transpose_45, - transpose_46, - transpose_47, - transpose_49, - transpose_5, - transpose_50, - transpose_51, - transpose_53, - transpose_54, - transpose_55, - transpose_57, - transpose_58, - transpose_59, - transpose_6, - transpose_61, - transpose_62, - transpose_63, - transpose_65, - transpose_66, - transpose_67, - transpose_69, - transpose_7, - transpose_70, - transpose_71, - transpose_73, - transpose_74, - transpose_75, - transpose_77, - transpose_78, - transpose_79, - transpose_81, - transpose_82, - transpose_83, - transpose_85, - transpose_86, - transpose_87, - transpose_89, - transpose_9, - transpose_90, - transpose_91, - transpose_93, - transpose_94, - transpose_95, - unsqueeze_0, - ) + del add_218 return tanh_0 diff --git a/paddle_samples/PaddleNLP/skep_ernie_2.0_large_en/weight_meta.py b/paddle_samples/PaddleNLP/skep_ernie_2.0_large_en/weight_meta.py index 8343c5440..b99e31c8f 100644 --- a/paddle_samples/PaddleNLP/skep_ernie_2.0_large_en/weight_meta.py +++ b/paddle_samples/PaddleNLP/skep_ernie_2.0_large_en/weight_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_parameter_0: name = "parameter_0" + original_name = "linear_4291.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0964225") @@ -11,6 +12,7 @@ class Program_weight_tensor_parameter_0: class Program_weight_tensor_parameter_1: name = "parameter_1" + original_name = "linear_4291.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.206677") @@ -22,6 +24,7 @@ class Program_weight_tensor_parameter_1: class Program_weight_tensor_parameter_2: name = "parameter_2" + original_name = "layer_norm_1479.b_0" shape = [1024] dtype = "float32" min_val = float("-0.557199") @@ -33,6 +36,7 @@ class Program_weight_tensor_parameter_2: class Program_weight_tensor_parameter_3: name = "parameter_3" + original_name = "layer_norm_1479.w_0" shape = [1024] dtype = "float32" min_val = float("0.284239") @@ -44,6 +48,7 @@ class Program_weight_tensor_parameter_3: class Program_weight_tensor_parameter_4: name = "parameter_4" + original_name = "layer_norm_1478.b_0" shape = [1024] dtype = "float32" min_val = float("-0.227083") @@ -55,6 +60,7 @@ class Program_weight_tensor_parameter_4: class Program_weight_tensor_parameter_5: name = "parameter_5" + original_name = "layer_norm_1478.w_0" shape = [1024] dtype = "float32" min_val = float("0.756546") @@ -66,6 +72,7 @@ class Program_weight_tensor_parameter_5: class Program_weight_tensor_parameter_6: name = "parameter_6" + original_name = "linear_4290.b_0" shape = [1024] dtype = "float32" min_val = float("-0.190782") @@ -77,6 +84,7 @@ class Program_weight_tensor_parameter_6: class Program_weight_tensor_parameter_7: name = "parameter_7" + original_name = "linear_4290.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-0.848767") @@ -88,6 +96,7 @@ class Program_weight_tensor_parameter_7: class Program_weight_tensor_parameter_8: name = "parameter_8" + original_name = "linear_4289.b_0" shape = [4096] dtype = "float32" min_val = float("-0.372142") @@ -99,6 +108,7 @@ class Program_weight_tensor_parameter_8: class Program_weight_tensor_parameter_9: name = "parameter_9" + original_name = "linear_4289.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.205599") @@ -110,6 +120,7 @@ class Program_weight_tensor_parameter_9: class Program_weight_tensor_parameter_10: name = "parameter_10" + original_name = "linear_4288.b_0" shape = [1024] dtype = "float32" min_val = float("-0.155684") @@ -121,6 +132,7 @@ class Program_weight_tensor_parameter_10: class Program_weight_tensor_parameter_11: name = "parameter_11" + original_name = "linear_4288.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.181717") @@ -132,6 +144,7 @@ class Program_weight_tensor_parameter_11: class Program_weight_tensor_parameter_12: name = "parameter_12" + original_name = "linear_4287.b_0" shape = [1024] dtype = "float32" min_val = float("-0.110364") @@ -143,6 +156,7 @@ class Program_weight_tensor_parameter_12: class Program_weight_tensor_parameter_13: name = "parameter_13" + original_name = "linear_4287.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.206996") @@ -154,6 +168,7 @@ class Program_weight_tensor_parameter_13: class Program_weight_tensor_parameter_14: name = "parameter_14" + original_name = "linear_4286.b_0" shape = [1024] dtype = "float32" min_val = float("-5.7465") @@ -165,6 +180,7 @@ class Program_weight_tensor_parameter_14: class Program_weight_tensor_parameter_15: name = "parameter_15" + original_name = "linear_4286.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.525623") @@ -176,6 +192,7 @@ class Program_weight_tensor_parameter_15: class Program_weight_tensor_parameter_16: name = "parameter_16" + original_name = "linear_4285.b_0" shape = [1024] dtype = "float32" min_val = float("-0.593351") @@ -187,6 +204,7 @@ class Program_weight_tensor_parameter_16: class Program_weight_tensor_parameter_17: name = "parameter_17" + original_name = "linear_4285.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.340697") @@ -198,6 +216,7 @@ class Program_weight_tensor_parameter_17: class Program_weight_tensor_parameter_18: name = "parameter_18" + original_name = "layer_norm_1477.b_0" shape = [1024] dtype = "float32" min_val = float("-0.115539") @@ -209,6 +228,7 @@ class Program_weight_tensor_parameter_18: class Program_weight_tensor_parameter_19: name = "parameter_19" + original_name = "layer_norm_1477.w_0" shape = [1024] dtype = "float32" min_val = float("0.0514438") @@ -220,6 +240,7 @@ class Program_weight_tensor_parameter_19: class Program_weight_tensor_parameter_20: name = "parameter_20" + original_name = "layer_norm_1476.b_0" shape = [1024] dtype = "float32" min_val = float("-0.395873") @@ -231,6 +252,7 @@ class Program_weight_tensor_parameter_20: class Program_weight_tensor_parameter_21: name = "parameter_21" + original_name = "layer_norm_1476.w_0" shape = [1024] dtype = "float32" min_val = float("0.691651") @@ -242,6 +264,7 @@ class Program_weight_tensor_parameter_21: class Program_weight_tensor_parameter_22: name = "parameter_22" + original_name = "linear_4284.b_0" shape = [1024] dtype = "float32" min_val = float("-0.683503") @@ -253,6 +276,7 @@ class Program_weight_tensor_parameter_22: class Program_weight_tensor_parameter_23: name = "parameter_23" + original_name = "linear_4284.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-5.6009") @@ -264,6 +288,7 @@ class Program_weight_tensor_parameter_23: class Program_weight_tensor_parameter_24: name = "parameter_24" + original_name = "linear_4283.b_0" shape = [4096] dtype = "float32" min_val = float("-0.260091") @@ -275,6 +300,7 @@ class Program_weight_tensor_parameter_24: class Program_weight_tensor_parameter_25: name = "parameter_25" + original_name = "linear_4283.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.912912") @@ -286,6 +312,7 @@ class Program_weight_tensor_parameter_25: class Program_weight_tensor_parameter_26: name = "parameter_26" + original_name = "linear_4282.b_0" shape = [1024] dtype = "float32" min_val = float("-0.231532") @@ -297,6 +324,7 @@ class Program_weight_tensor_parameter_26: class Program_weight_tensor_parameter_27: name = "parameter_27" + original_name = "linear_4282.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.188875") @@ -308,6 +336,7 @@ class Program_weight_tensor_parameter_27: class Program_weight_tensor_parameter_28: name = "parameter_28" + original_name = "linear_4281.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0664895") @@ -319,6 +348,7 @@ class Program_weight_tensor_parameter_28: class Program_weight_tensor_parameter_29: name = "parameter_29" + original_name = "linear_4281.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.211943") @@ -330,6 +360,7 @@ class Program_weight_tensor_parameter_29: class Program_weight_tensor_parameter_30: name = "parameter_30" + original_name = "linear_4280.b_0" shape = [1024] dtype = "float32" min_val = float("-6.05831") @@ -341,6 +372,7 @@ class Program_weight_tensor_parameter_30: class Program_weight_tensor_parameter_31: name = "parameter_31" + original_name = "linear_4280.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.257823") @@ -352,6 +384,7 @@ class Program_weight_tensor_parameter_31: class Program_weight_tensor_parameter_32: name = "parameter_32" + original_name = "linear_4279.b_0" shape = [1024] dtype = "float32" min_val = float("-0.539692") @@ -363,6 +396,7 @@ class Program_weight_tensor_parameter_32: class Program_weight_tensor_parameter_33: name = "parameter_33" + original_name = "linear_4279.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.392863") @@ -374,6 +408,7 @@ class Program_weight_tensor_parameter_33: class Program_weight_tensor_parameter_34: name = "parameter_34" + original_name = "layer_norm_1475.b_0" shape = [1024] dtype = "float32" min_val = float("-0.392519") @@ -385,6 +420,7 @@ class Program_weight_tensor_parameter_34: class Program_weight_tensor_parameter_35: name = "parameter_35" + original_name = "layer_norm_1475.w_0" shape = [1024] dtype = "float32" min_val = float("0.066148") @@ -396,6 +432,7 @@ class Program_weight_tensor_parameter_35: class Program_weight_tensor_parameter_36: name = "parameter_36" + original_name = "layer_norm_1474.b_0" shape = [1024] dtype = "float32" min_val = float("-0.50683") @@ -407,6 +444,7 @@ class Program_weight_tensor_parameter_36: class Program_weight_tensor_parameter_37: name = "parameter_37" + original_name = "layer_norm_1474.w_0" shape = [1024] dtype = "float32" min_val = float("0.634109") @@ -418,6 +456,7 @@ class Program_weight_tensor_parameter_37: class Program_weight_tensor_parameter_38: name = "parameter_38" + original_name = "linear_4278.b_0" shape = [1024] dtype = "float32" min_val = float("-0.558364") @@ -429,6 +468,7 @@ class Program_weight_tensor_parameter_38: class Program_weight_tensor_parameter_39: name = "parameter_39" + original_name = "linear_4278.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-6.76184") @@ -440,6 +480,7 @@ class Program_weight_tensor_parameter_39: class Program_weight_tensor_parameter_40: name = "parameter_40" + original_name = "linear_4277.b_0" shape = [4096] dtype = "float32" min_val = float("-0.159331") @@ -451,6 +492,7 @@ class Program_weight_tensor_parameter_40: class Program_weight_tensor_parameter_41: name = "parameter_41" + original_name = "linear_4277.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-1.51378") @@ -462,6 +504,7 @@ class Program_weight_tensor_parameter_41: class Program_weight_tensor_parameter_42: name = "parameter_42" + original_name = "linear_4276.b_0" shape = [1024] dtype = "float32" min_val = float("-0.28841") @@ -473,6 +516,7 @@ class Program_weight_tensor_parameter_42: class Program_weight_tensor_parameter_43: name = "parameter_43" + original_name = "linear_4276.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.187442") @@ -484,6 +528,7 @@ class Program_weight_tensor_parameter_43: class Program_weight_tensor_parameter_44: name = "parameter_44" + original_name = "linear_4275.b_0" shape = [1024] dtype = "float32" min_val = float("-0.256224") @@ -495,6 +540,7 @@ class Program_weight_tensor_parameter_44: class Program_weight_tensor_parameter_45: name = "parameter_45" + original_name = "linear_4275.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.200321") @@ -506,6 +552,7 @@ class Program_weight_tensor_parameter_45: class Program_weight_tensor_parameter_46: name = "parameter_46" + original_name = "linear_4274.b_0" shape = [1024] dtype = "float32" min_val = float("-6.21819") @@ -517,6 +564,7 @@ class Program_weight_tensor_parameter_46: class Program_weight_tensor_parameter_47: name = "parameter_47" + original_name = "linear_4274.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.434202") @@ -528,6 +576,7 @@ class Program_weight_tensor_parameter_47: class Program_weight_tensor_parameter_48: name = "parameter_48" + original_name = "linear_4273.b_0" shape = [1024] dtype = "float32" min_val = float("-0.732496") @@ -539,6 +588,7 @@ class Program_weight_tensor_parameter_48: class Program_weight_tensor_parameter_49: name = "parameter_49" + original_name = "linear_4273.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.301956") @@ -550,6 +600,7 @@ class Program_weight_tensor_parameter_49: class Program_weight_tensor_parameter_50: name = "parameter_50" + original_name = "layer_norm_1473.b_0" shape = [1024] dtype = "float32" min_val = float("-0.183304") @@ -561,6 +612,7 @@ class Program_weight_tensor_parameter_50: class Program_weight_tensor_parameter_51: name = "parameter_51" + original_name = "layer_norm_1473.w_0" shape = [1024] dtype = "float32" min_val = float("0.0446676") @@ -572,6 +624,7 @@ class Program_weight_tensor_parameter_51: class Program_weight_tensor_parameter_52: name = "parameter_52" + original_name = "layer_norm_1472.b_0" shape = [1024] dtype = "float32" min_val = float("-1.18581") @@ -583,6 +636,7 @@ class Program_weight_tensor_parameter_52: class Program_weight_tensor_parameter_53: name = "parameter_53" + original_name = "layer_norm_1472.w_0" shape = [1024] dtype = "float32" min_val = float("0.584943") @@ -594,6 +648,7 @@ class Program_weight_tensor_parameter_53: class Program_weight_tensor_parameter_54: name = "parameter_54" + original_name = "linear_4272.b_0" shape = [1024] dtype = "float32" min_val = float("-0.706692") @@ -605,6 +660,7 @@ class Program_weight_tensor_parameter_54: class Program_weight_tensor_parameter_55: name = "parameter_55" + original_name = "linear_4272.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-5.03077") @@ -616,6 +672,7 @@ class Program_weight_tensor_parameter_55: class Program_weight_tensor_parameter_56: name = "parameter_56" + original_name = "linear_4271.b_0" shape = [4096] dtype = "float32" min_val = float("-0.217203") @@ -627,6 +684,7 @@ class Program_weight_tensor_parameter_56: class Program_weight_tensor_parameter_57: name = "parameter_57" + original_name = "linear_4271.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.446221") @@ -638,6 +696,7 @@ class Program_weight_tensor_parameter_57: class Program_weight_tensor_parameter_58: name = "parameter_58" + original_name = "linear_4270.b_0" shape = [1024] dtype = "float32" min_val = float("-0.334759") @@ -649,6 +708,7 @@ class Program_weight_tensor_parameter_58: class Program_weight_tensor_parameter_59: name = "parameter_59" + original_name = "linear_4270.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.204179") @@ -660,6 +720,7 @@ class Program_weight_tensor_parameter_59: class Program_weight_tensor_parameter_60: name = "parameter_60" + original_name = "linear_4269.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0923562") @@ -671,6 +732,7 @@ class Program_weight_tensor_parameter_60: class Program_weight_tensor_parameter_61: name = "parameter_61" + original_name = "linear_4269.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.210935") @@ -682,6 +744,7 @@ class Program_weight_tensor_parameter_61: class Program_weight_tensor_parameter_62: name = "parameter_62" + original_name = "linear_4268.b_0" shape = [1024] dtype = "float32" min_val = float("-6.72088") @@ -693,6 +756,7 @@ class Program_weight_tensor_parameter_62: class Program_weight_tensor_parameter_63: name = "parameter_63" + original_name = "linear_4268.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-1.95047") @@ -704,6 +768,7 @@ class Program_weight_tensor_parameter_63: class Program_weight_tensor_parameter_64: name = "parameter_64" + original_name = "linear_4267.b_0" shape = [1024] dtype = "float32" min_val = float("-0.746794") @@ -715,6 +780,7 @@ class Program_weight_tensor_parameter_64: class Program_weight_tensor_parameter_65: name = "parameter_65" + original_name = "linear_4267.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.373324") @@ -726,6 +792,7 @@ class Program_weight_tensor_parameter_65: class Program_weight_tensor_parameter_66: name = "parameter_66" + original_name = "layer_norm_1471.b_0" shape = [1024] dtype = "float32" min_val = float("-0.68651") @@ -737,6 +804,7 @@ class Program_weight_tensor_parameter_66: class Program_weight_tensor_parameter_67: name = "parameter_67" + original_name = "layer_norm_1471.w_0" shape = [1024] dtype = "float32" min_val = float("0.0112333") @@ -748,6 +816,7 @@ class Program_weight_tensor_parameter_67: class Program_weight_tensor_parameter_68: name = "parameter_68" + original_name = "layer_norm_1470.b_0" shape = [1024] dtype = "float32" min_val = float("-1.28879") @@ -759,6 +828,7 @@ class Program_weight_tensor_parameter_68: class Program_weight_tensor_parameter_69: name = "parameter_69" + original_name = "layer_norm_1470.w_0" shape = [1024] dtype = "float32" min_val = float("0.662829") @@ -770,6 +840,7 @@ class Program_weight_tensor_parameter_69: class Program_weight_tensor_parameter_70: name = "parameter_70" + original_name = "linear_4266.b_0" shape = [1024] dtype = "float32" min_val = float("-0.367769") @@ -781,6 +852,7 @@ class Program_weight_tensor_parameter_70: class Program_weight_tensor_parameter_71: name = "parameter_71" + original_name = "linear_4266.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-5.96343") @@ -792,6 +864,7 @@ class Program_weight_tensor_parameter_71: class Program_weight_tensor_parameter_72: name = "parameter_72" + original_name = "linear_4265.b_0" shape = [4096] dtype = "float32" min_val = float("-0.296464") @@ -803,6 +876,7 @@ class Program_weight_tensor_parameter_72: class Program_weight_tensor_parameter_73: name = "parameter_73" + original_name = "linear_4265.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.595273") @@ -814,6 +888,7 @@ class Program_weight_tensor_parameter_73: class Program_weight_tensor_parameter_74: name = "parameter_74" + original_name = "linear_4264.b_0" shape = [1024] dtype = "float32" min_val = float("-0.276061") @@ -825,6 +900,7 @@ class Program_weight_tensor_parameter_74: class Program_weight_tensor_parameter_75: name = "parameter_75" + original_name = "linear_4264.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.166253") @@ -836,6 +912,7 @@ class Program_weight_tensor_parameter_75: class Program_weight_tensor_parameter_76: name = "parameter_76" + original_name = "linear_4263.b_0" shape = [1024] dtype = "float32" min_val = float("-0.147808") @@ -847,6 +924,7 @@ class Program_weight_tensor_parameter_76: class Program_weight_tensor_parameter_77: name = "parameter_77" + original_name = "linear_4263.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.167294") @@ -858,6 +936,7 @@ class Program_weight_tensor_parameter_77: class Program_weight_tensor_parameter_78: name = "parameter_78" + original_name = "linear_4262.b_0" shape = [1024] dtype = "float32" min_val = float("-4.1333") @@ -869,6 +948,7 @@ class Program_weight_tensor_parameter_78: class Program_weight_tensor_parameter_79: name = "parameter_79" + original_name = "linear_4262.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.31173") @@ -880,6 +960,7 @@ class Program_weight_tensor_parameter_79: class Program_weight_tensor_parameter_80: name = "parameter_80" + original_name = "linear_4261.b_0" shape = [1024] dtype = "float32" min_val = float("-0.542118") @@ -891,6 +972,7 @@ class Program_weight_tensor_parameter_80: class Program_weight_tensor_parameter_81: name = "parameter_81" + original_name = "linear_4261.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.301893") @@ -902,6 +984,7 @@ class Program_weight_tensor_parameter_81: class Program_weight_tensor_parameter_82: name = "parameter_82" + original_name = "layer_norm_1469.b_0" shape = [1024] dtype = "float32" min_val = float("-0.37999") @@ -913,6 +996,7 @@ class Program_weight_tensor_parameter_82: class Program_weight_tensor_parameter_83: name = "parameter_83" + original_name = "layer_norm_1469.w_0" shape = [1024] dtype = "float32" min_val = float("0.205573") @@ -924,6 +1008,7 @@ class Program_weight_tensor_parameter_83: class Program_weight_tensor_parameter_84: name = "parameter_84" + original_name = "layer_norm_1468.b_0" shape = [1024] dtype = "float32" min_val = float("-1.78646") @@ -935,6 +1020,7 @@ class Program_weight_tensor_parameter_84: class Program_weight_tensor_parameter_85: name = "parameter_85" + original_name = "layer_norm_1468.w_0" shape = [1024] dtype = "float32" min_val = float("0.661085") @@ -946,6 +1032,7 @@ class Program_weight_tensor_parameter_85: class Program_weight_tensor_parameter_86: name = "parameter_86" + original_name = "linear_4260.b_0" shape = [1024] dtype = "float32" min_val = float("-0.239232") @@ -957,6 +1044,7 @@ class Program_weight_tensor_parameter_86: class Program_weight_tensor_parameter_87: name = "parameter_87" + original_name = "linear_4260.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-5.3042") @@ -968,6 +1056,7 @@ class Program_weight_tensor_parameter_87: class Program_weight_tensor_parameter_88: name = "parameter_88" + original_name = "linear_4259.b_0" shape = [4096] dtype = "float32" min_val = float("-0.294008") @@ -979,6 +1068,7 @@ class Program_weight_tensor_parameter_88: class Program_weight_tensor_parameter_89: name = "parameter_89" + original_name = "linear_4259.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-1.02776") @@ -990,6 +1080,7 @@ class Program_weight_tensor_parameter_89: class Program_weight_tensor_parameter_90: name = "parameter_90" + original_name = "linear_4258.b_0" shape = [1024] dtype = "float32" min_val = float("-0.165158") @@ -1001,6 +1092,7 @@ class Program_weight_tensor_parameter_90: class Program_weight_tensor_parameter_91: name = "parameter_91" + original_name = "linear_4258.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.153821") @@ -1012,6 +1104,7 @@ class Program_weight_tensor_parameter_91: class Program_weight_tensor_parameter_92: name = "parameter_92" + original_name = "linear_4257.b_0" shape = [1024] dtype = "float32" min_val = float("-0.17822") @@ -1023,6 +1116,7 @@ class Program_weight_tensor_parameter_92: class Program_weight_tensor_parameter_93: name = "parameter_93" + original_name = "linear_4257.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.196512") @@ -1034,6 +1128,7 @@ class Program_weight_tensor_parameter_93: class Program_weight_tensor_parameter_94: name = "parameter_94" + original_name = "linear_4256.b_0" shape = [1024] dtype = "float32" min_val = float("-5.89218") @@ -1045,6 +1140,7 @@ class Program_weight_tensor_parameter_94: class Program_weight_tensor_parameter_95: name = "parameter_95" + original_name = "linear_4256.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.287411") @@ -1056,6 +1152,7 @@ class Program_weight_tensor_parameter_95: class Program_weight_tensor_parameter_96: name = "parameter_96" + original_name = "linear_4255.b_0" shape = [1024] dtype = "float32" min_val = float("-0.481481") @@ -1067,6 +1164,7 @@ class Program_weight_tensor_parameter_96: class Program_weight_tensor_parameter_97: name = "parameter_97" + original_name = "linear_4255.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.354093") @@ -1078,6 +1176,7 @@ class Program_weight_tensor_parameter_97: class Program_weight_tensor_parameter_98: name = "parameter_98" + original_name = "layer_norm_1467.b_0" shape = [1024] dtype = "float32" min_val = float("-0.618261") @@ -1089,6 +1188,7 @@ class Program_weight_tensor_parameter_98: class Program_weight_tensor_parameter_99: name = "parameter_99" + original_name = "layer_norm_1467.w_0" shape = [1024] dtype = "float32" min_val = float("0.270417") @@ -1100,6 +1200,7 @@ class Program_weight_tensor_parameter_99: class Program_weight_tensor_parameter_100: name = "parameter_100" + original_name = "layer_norm_1466.b_0" shape = [1024] dtype = "float32" min_val = float("-1.66067") @@ -1111,6 +1212,7 @@ class Program_weight_tensor_parameter_100: class Program_weight_tensor_parameter_101: name = "parameter_101" + original_name = "layer_norm_1466.w_0" shape = [1024] dtype = "float32" min_val = float("0.678842") @@ -1122,6 +1224,7 @@ class Program_weight_tensor_parameter_101: class Program_weight_tensor_parameter_102: name = "parameter_102" + original_name = "linear_4254.b_0" shape = [1024] dtype = "float32" min_val = float("-0.663529") @@ -1133,6 +1236,7 @@ class Program_weight_tensor_parameter_102: class Program_weight_tensor_parameter_103: name = "parameter_103" + original_name = "linear_4254.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-4.82955") @@ -1144,6 +1248,7 @@ class Program_weight_tensor_parameter_103: class Program_weight_tensor_parameter_104: name = "parameter_104" + original_name = "linear_4253.b_0" shape = [4096] dtype = "float32" min_val = float("-0.285894") @@ -1155,6 +1260,7 @@ class Program_weight_tensor_parameter_104: class Program_weight_tensor_parameter_105: name = "parameter_105" + original_name = "linear_4253.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.539576") @@ -1166,6 +1272,7 @@ class Program_weight_tensor_parameter_105: class Program_weight_tensor_parameter_106: name = "parameter_106" + original_name = "linear_4252.b_0" shape = [1024] dtype = "float32" min_val = float("-0.123169") @@ -1177,6 +1284,7 @@ class Program_weight_tensor_parameter_106: class Program_weight_tensor_parameter_107: name = "parameter_107" + original_name = "linear_4252.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.177103") @@ -1188,6 +1296,7 @@ class Program_weight_tensor_parameter_107: class Program_weight_tensor_parameter_108: name = "parameter_108" + original_name = "linear_4251.b_0" shape = [1024] dtype = "float32" min_val = float("-0.226442") @@ -1199,6 +1308,7 @@ class Program_weight_tensor_parameter_108: class Program_weight_tensor_parameter_109: name = "parameter_109" + original_name = "linear_4251.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.174057") @@ -1210,6 +1320,7 @@ class Program_weight_tensor_parameter_109: class Program_weight_tensor_parameter_110: name = "parameter_110" + original_name = "linear_4250.b_0" shape = [1024] dtype = "float32" min_val = float("-1.7186") @@ -1221,6 +1332,7 @@ class Program_weight_tensor_parameter_110: class Program_weight_tensor_parameter_111: name = "parameter_111" + original_name = "linear_4250.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.280335") @@ -1232,6 +1344,7 @@ class Program_weight_tensor_parameter_111: class Program_weight_tensor_parameter_112: name = "parameter_112" + original_name = "linear_4249.b_0" shape = [1024] dtype = "float32" min_val = float("-0.458628") @@ -1243,6 +1356,7 @@ class Program_weight_tensor_parameter_112: class Program_weight_tensor_parameter_113: name = "parameter_113" + original_name = "linear_4249.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.244457") @@ -1254,6 +1368,7 @@ class Program_weight_tensor_parameter_113: class Program_weight_tensor_parameter_114: name = "parameter_114" + original_name = "layer_norm_1465.b_0" shape = [1024] dtype = "float32" min_val = float("-0.552241") @@ -1265,6 +1380,7 @@ class Program_weight_tensor_parameter_114: class Program_weight_tensor_parameter_115: name = "parameter_115" + original_name = "layer_norm_1465.w_0" shape = [1024] dtype = "float32" min_val = float("0.43101") @@ -1276,6 +1392,7 @@ class Program_weight_tensor_parameter_115: class Program_weight_tensor_parameter_116: name = "parameter_116" + original_name = "layer_norm_1464.b_0" shape = [1024] dtype = "float32" min_val = float("-1.66666") @@ -1287,6 +1404,7 @@ class Program_weight_tensor_parameter_116: class Program_weight_tensor_parameter_117: name = "parameter_117" + original_name = "layer_norm_1464.w_0" shape = [1024] dtype = "float32" min_val = float("0.651818") @@ -1298,6 +1416,7 @@ class Program_weight_tensor_parameter_117: class Program_weight_tensor_parameter_118: name = "parameter_118" + original_name = "linear_4248.b_0" shape = [1024] dtype = "float32" min_val = float("-0.917302") @@ -1309,6 +1428,7 @@ class Program_weight_tensor_parameter_118: class Program_weight_tensor_parameter_119: name = "parameter_119" + original_name = "linear_4248.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-0.831292") @@ -1320,6 +1440,7 @@ class Program_weight_tensor_parameter_119: class Program_weight_tensor_parameter_120: name = "parameter_120" + original_name = "linear_4247.b_0" shape = [4096] dtype = "float32" min_val = float("-0.292934") @@ -1331,6 +1452,7 @@ class Program_weight_tensor_parameter_120: class Program_weight_tensor_parameter_121: name = "parameter_121" + original_name = "linear_4247.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.197997") @@ -1342,6 +1464,7 @@ class Program_weight_tensor_parameter_121: class Program_weight_tensor_parameter_122: name = "parameter_122" + original_name = "linear_4246.b_0" shape = [1024] dtype = "float32" min_val = float("-0.1234") @@ -1353,6 +1476,7 @@ class Program_weight_tensor_parameter_122: class Program_weight_tensor_parameter_123: name = "parameter_123" + original_name = "linear_4246.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.219433") @@ -1364,6 +1488,7 @@ class Program_weight_tensor_parameter_123: class Program_weight_tensor_parameter_124: name = "parameter_124" + original_name = "linear_4245.b_0" shape = [1024] dtype = "float32" min_val = float("-0.120596") @@ -1375,6 +1500,7 @@ class Program_weight_tensor_parameter_124: class Program_weight_tensor_parameter_125: name = "parameter_125" + original_name = "linear_4245.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.160361") @@ -1386,6 +1512,7 @@ class Program_weight_tensor_parameter_125: class Program_weight_tensor_parameter_126: name = "parameter_126" + original_name = "linear_4244.b_0" shape = [1024] dtype = "float32" min_val = float("-2.03583") @@ -1397,6 +1524,7 @@ class Program_weight_tensor_parameter_126: class Program_weight_tensor_parameter_127: name = "parameter_127" + original_name = "linear_4244.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.29531") @@ -1408,6 +1536,7 @@ class Program_weight_tensor_parameter_127: class Program_weight_tensor_parameter_128: name = "parameter_128" + original_name = "linear_4243.b_0" shape = [1024] dtype = "float32" min_val = float("-0.554885") @@ -1419,6 +1548,7 @@ class Program_weight_tensor_parameter_128: class Program_weight_tensor_parameter_129: name = "parameter_129" + original_name = "linear_4243.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.2113") @@ -1430,6 +1560,7 @@ class Program_weight_tensor_parameter_129: class Program_weight_tensor_parameter_130: name = "parameter_130" + original_name = "layer_norm_1463.b_0" shape = [1024] dtype = "float32" min_val = float("-0.896358") @@ -1441,6 +1572,7 @@ class Program_weight_tensor_parameter_130: class Program_weight_tensor_parameter_131: name = "parameter_131" + original_name = "layer_norm_1463.w_0" shape = [1024] dtype = "float32" min_val = float("0.375239") @@ -1452,6 +1584,7 @@ class Program_weight_tensor_parameter_131: class Program_weight_tensor_parameter_132: name = "parameter_132" + original_name = "layer_norm_1462.b_0" shape = [1024] dtype = "float32" min_val = float("-2.00816") @@ -1463,6 +1596,7 @@ class Program_weight_tensor_parameter_132: class Program_weight_tensor_parameter_133: name = "parameter_133" + original_name = "layer_norm_1462.w_0" shape = [1024] dtype = "float32" min_val = float("0.650589") @@ -1474,6 +1608,7 @@ class Program_weight_tensor_parameter_133: class Program_weight_tensor_parameter_134: name = "parameter_134" + original_name = "linear_4242.b_0" shape = [1024] dtype = "float32" min_val = float("-0.840489") @@ -1485,6 +1620,7 @@ class Program_weight_tensor_parameter_134: class Program_weight_tensor_parameter_135: name = "parameter_135" + original_name = "linear_4242.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-2.4264") @@ -1496,6 +1632,7 @@ class Program_weight_tensor_parameter_135: class Program_weight_tensor_parameter_136: name = "parameter_136" + original_name = "linear_4241.b_0" shape = [4096] dtype = "float32" min_val = float("-0.298287") @@ -1507,6 +1644,7 @@ class Program_weight_tensor_parameter_136: class Program_weight_tensor_parameter_137: name = "parameter_137" + original_name = "linear_4241.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.235349") @@ -1518,6 +1656,7 @@ class Program_weight_tensor_parameter_137: class Program_weight_tensor_parameter_138: name = "parameter_138" + original_name = "linear_4240.b_0" shape = [1024] dtype = "float32" min_val = float("-0.133371") @@ -1529,6 +1668,7 @@ class Program_weight_tensor_parameter_138: class Program_weight_tensor_parameter_139: name = "parameter_139" + original_name = "linear_4240.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.513362") @@ -1540,6 +1680,7 @@ class Program_weight_tensor_parameter_139: class Program_weight_tensor_parameter_140: name = "parameter_140" + original_name = "linear_4239.b_0" shape = [1024] dtype = "float32" min_val = float("-0.278286") @@ -1551,6 +1692,7 @@ class Program_weight_tensor_parameter_140: class Program_weight_tensor_parameter_141: name = "parameter_141" + original_name = "linear_4239.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.182497") @@ -1562,6 +1704,7 @@ class Program_weight_tensor_parameter_141: class Program_weight_tensor_parameter_142: name = "parameter_142" + original_name = "linear_4238.b_0" shape = [1024] dtype = "float32" min_val = float("-2.02989") @@ -1573,6 +1716,7 @@ class Program_weight_tensor_parameter_142: class Program_weight_tensor_parameter_143: name = "parameter_143" + original_name = "linear_4238.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.281268") @@ -1584,6 +1728,7 @@ class Program_weight_tensor_parameter_143: class Program_weight_tensor_parameter_144: name = "parameter_144" + original_name = "linear_4237.b_0" shape = [1024] dtype = "float32" min_val = float("-0.478051") @@ -1595,6 +1740,7 @@ class Program_weight_tensor_parameter_144: class Program_weight_tensor_parameter_145: name = "parameter_145" + original_name = "linear_4237.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.294227") @@ -1606,6 +1752,7 @@ class Program_weight_tensor_parameter_145: class Program_weight_tensor_parameter_146: name = "parameter_146" + original_name = "layer_norm_1461.b_0" shape = [1024] dtype = "float32" min_val = float("-2.1473") @@ -1617,6 +1764,7 @@ class Program_weight_tensor_parameter_146: class Program_weight_tensor_parameter_147: name = "parameter_147" + original_name = "layer_norm_1461.w_0" shape = [1024] dtype = "float32" min_val = float("0.119336") @@ -1628,6 +1776,7 @@ class Program_weight_tensor_parameter_147: class Program_weight_tensor_parameter_148: name = "parameter_148" + original_name = "layer_norm_1460.b_0" shape = [1024] dtype = "float32" min_val = float("-2.15353") @@ -1639,6 +1788,7 @@ class Program_weight_tensor_parameter_148: class Program_weight_tensor_parameter_149: name = "parameter_149" + original_name = "layer_norm_1460.w_0" shape = [1024] dtype = "float32" min_val = float("0.677357") @@ -1650,6 +1800,7 @@ class Program_weight_tensor_parameter_149: class Program_weight_tensor_parameter_150: name = "parameter_150" + original_name = "linear_4236.b_0" shape = [1024] dtype = "float32" min_val = float("-0.550193") @@ -1661,6 +1812,7 @@ class Program_weight_tensor_parameter_150: class Program_weight_tensor_parameter_151: name = "parameter_151" + original_name = "linear_4236.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-1.42262") @@ -1672,6 +1824,7 @@ class Program_weight_tensor_parameter_151: class Program_weight_tensor_parameter_152: name = "parameter_152" + original_name = "linear_4235.b_0" shape = [4096] dtype = "float32" min_val = float("-0.245592") @@ -1683,6 +1836,7 @@ class Program_weight_tensor_parameter_152: class Program_weight_tensor_parameter_153: name = "parameter_153" + original_name = "linear_4235.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.272411") @@ -1694,6 +1848,7 @@ class Program_weight_tensor_parameter_153: class Program_weight_tensor_parameter_154: name = "parameter_154" + original_name = "linear_4234.b_0" shape = [1024] dtype = "float32" min_val = float("-0.13229") @@ -1705,6 +1860,7 @@ class Program_weight_tensor_parameter_154: class Program_weight_tensor_parameter_155: name = "parameter_155" + original_name = "linear_4234.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.22732") @@ -1716,6 +1872,7 @@ class Program_weight_tensor_parameter_155: class Program_weight_tensor_parameter_156: name = "parameter_156" + original_name = "linear_4233.b_0" shape = [1024] dtype = "float32" min_val = float("-0.140607") @@ -1727,6 +1884,7 @@ class Program_weight_tensor_parameter_156: class Program_weight_tensor_parameter_157: name = "parameter_157" + original_name = "linear_4233.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.167014") @@ -1738,6 +1896,7 @@ class Program_weight_tensor_parameter_157: class Program_weight_tensor_parameter_158: name = "parameter_158" + original_name = "linear_4232.b_0" shape = [1024] dtype = "float32" min_val = float("-1.14242") @@ -1749,6 +1908,7 @@ class Program_weight_tensor_parameter_158: class Program_weight_tensor_parameter_159: name = "parameter_159" + original_name = "linear_4232.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.345277") @@ -1760,6 +1920,7 @@ class Program_weight_tensor_parameter_159: class Program_weight_tensor_parameter_160: name = "parameter_160" + original_name = "linear_4231.b_0" shape = [1024] dtype = "float32" min_val = float("-0.602159") @@ -1771,6 +1932,7 @@ class Program_weight_tensor_parameter_160: class Program_weight_tensor_parameter_161: name = "parameter_161" + original_name = "linear_4231.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.345716") @@ -1782,6 +1944,7 @@ class Program_weight_tensor_parameter_161: class Program_weight_tensor_parameter_162: name = "parameter_162" + original_name = "layer_norm_1459.b_0" shape = [1024] dtype = "float32" min_val = float("-1.34428") @@ -1793,6 +1956,7 @@ class Program_weight_tensor_parameter_162: class Program_weight_tensor_parameter_163: name = "parameter_163" + original_name = "layer_norm_1459.w_0" shape = [1024] dtype = "float32" min_val = float("0.400195") @@ -1804,6 +1968,7 @@ class Program_weight_tensor_parameter_163: class Program_weight_tensor_parameter_164: name = "parameter_164" + original_name = "layer_norm_1458.b_0" shape = [1024] dtype = "float32" min_val = float("-1.41684") @@ -1815,6 +1980,7 @@ class Program_weight_tensor_parameter_164: class Program_weight_tensor_parameter_165: name = "parameter_165" + original_name = "layer_norm_1458.w_0" shape = [1024] dtype = "float32" min_val = float("0.652885") @@ -1826,6 +1992,7 @@ class Program_weight_tensor_parameter_165: class Program_weight_tensor_parameter_166: name = "parameter_166" + original_name = "linear_4230.b_0" shape = [1024] dtype = "float32" min_val = float("-0.799939") @@ -1837,6 +2004,7 @@ class Program_weight_tensor_parameter_166: class Program_weight_tensor_parameter_167: name = "parameter_167" + original_name = "linear_4230.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-0.625465") @@ -1848,6 +2016,7 @@ class Program_weight_tensor_parameter_167: class Program_weight_tensor_parameter_168: name = "parameter_168" + original_name = "linear_4229.b_0" shape = [4096] dtype = "float32" min_val = float("-0.300545") @@ -1859,6 +2028,7 @@ class Program_weight_tensor_parameter_168: class Program_weight_tensor_parameter_169: name = "parameter_169" + original_name = "linear_4229.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.286748") @@ -1870,6 +2040,7 @@ class Program_weight_tensor_parameter_169: class Program_weight_tensor_parameter_170: name = "parameter_170" + original_name = "linear_4228.b_0" shape = [1024] dtype = "float32" min_val = float("-0.143451") @@ -1881,6 +2052,7 @@ class Program_weight_tensor_parameter_170: class Program_weight_tensor_parameter_171: name = "parameter_171" + original_name = "linear_4228.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.280793") @@ -1892,6 +2064,7 @@ class Program_weight_tensor_parameter_171: class Program_weight_tensor_parameter_172: name = "parameter_172" + original_name = "linear_4227.b_0" shape = [1024] dtype = "float32" min_val = float("-0.195321") @@ -1903,6 +2076,7 @@ class Program_weight_tensor_parameter_172: class Program_weight_tensor_parameter_173: name = "parameter_173" + original_name = "linear_4227.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.18612") @@ -1914,6 +2088,7 @@ class Program_weight_tensor_parameter_173: class Program_weight_tensor_parameter_174: name = "parameter_174" + original_name = "linear_4226.b_0" shape = [1024] dtype = "float32" min_val = float("-1.25597") @@ -1925,6 +2100,7 @@ class Program_weight_tensor_parameter_174: class Program_weight_tensor_parameter_175: name = "parameter_175" + original_name = "linear_4226.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.300544") @@ -1936,6 +2112,7 @@ class Program_weight_tensor_parameter_175: class Program_weight_tensor_parameter_176: name = "parameter_176" + original_name = "linear_4225.b_0" shape = [1024] dtype = "float32" min_val = float("-0.454775") @@ -1947,6 +2124,7 @@ class Program_weight_tensor_parameter_176: class Program_weight_tensor_parameter_177: name = "parameter_177" + original_name = "linear_4225.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.293232") @@ -1958,6 +2136,7 @@ class Program_weight_tensor_parameter_177: class Program_weight_tensor_parameter_178: name = "parameter_178" + original_name = "layer_norm_1457.b_0" shape = [1024] dtype = "float32" min_val = float("-0.411469") @@ -1969,6 +2148,7 @@ class Program_weight_tensor_parameter_178: class Program_weight_tensor_parameter_179: name = "parameter_179" + original_name = "layer_norm_1457.w_0" shape = [1024] dtype = "float32" min_val = float("0.40705") @@ -1980,6 +2160,7 @@ class Program_weight_tensor_parameter_179: class Program_weight_tensor_parameter_180: name = "parameter_180" + original_name = "layer_norm_1456.b_0" shape = [1024] dtype = "float32" min_val = float("-1.53874") @@ -1991,6 +2172,7 @@ class Program_weight_tensor_parameter_180: class Program_weight_tensor_parameter_181: name = "parameter_181" + original_name = "layer_norm_1456.w_0" shape = [1024] dtype = "float32" min_val = float("0.660103") @@ -2002,6 +2184,7 @@ class Program_weight_tensor_parameter_181: class Program_weight_tensor_parameter_182: name = "parameter_182" + original_name = "linear_4224.b_0" shape = [1024] dtype = "float32" min_val = float("-0.667434") @@ -2013,6 +2196,7 @@ class Program_weight_tensor_parameter_182: class Program_weight_tensor_parameter_183: name = "parameter_183" + original_name = "linear_4224.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-1.34589") @@ -2024,6 +2208,7 @@ class Program_weight_tensor_parameter_183: class Program_weight_tensor_parameter_184: name = "parameter_184" + original_name = "linear_4223.b_0" shape = [4096] dtype = "float32" min_val = float("-0.252446") @@ -2035,6 +2220,7 @@ class Program_weight_tensor_parameter_184: class Program_weight_tensor_parameter_185: name = "parameter_185" + original_name = "linear_4223.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.405738") @@ -2046,6 +2232,7 @@ class Program_weight_tensor_parameter_185: class Program_weight_tensor_parameter_186: name = "parameter_186" + original_name = "linear_4222.b_0" shape = [1024] dtype = "float32" min_val = float("-0.182543") @@ -2057,6 +2244,7 @@ class Program_weight_tensor_parameter_186: class Program_weight_tensor_parameter_187: name = "parameter_187" + original_name = "linear_4222.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.253057") @@ -2068,6 +2256,7 @@ class Program_weight_tensor_parameter_187: class Program_weight_tensor_parameter_188: name = "parameter_188" + original_name = "linear_4221.b_0" shape = [1024] dtype = "float32" min_val = float("-0.17986") @@ -2079,6 +2268,7 @@ class Program_weight_tensor_parameter_188: class Program_weight_tensor_parameter_189: name = "parameter_189" + original_name = "linear_4221.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.169411") @@ -2090,6 +2280,7 @@ class Program_weight_tensor_parameter_189: class Program_weight_tensor_parameter_190: name = "parameter_190" + original_name = "linear_4220.b_0" shape = [1024] dtype = "float32" min_val = float("-0.848125") @@ -2101,6 +2292,7 @@ class Program_weight_tensor_parameter_190: class Program_weight_tensor_parameter_191: name = "parameter_191" + original_name = "linear_4220.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.266592") @@ -2112,6 +2304,7 @@ class Program_weight_tensor_parameter_191: class Program_weight_tensor_parameter_192: name = "parameter_192" + original_name = "linear_4219.b_0" shape = [1024] dtype = "float32" min_val = float("-0.401497") @@ -2123,6 +2316,7 @@ class Program_weight_tensor_parameter_192: class Program_weight_tensor_parameter_193: name = "parameter_193" + original_name = "linear_4219.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.309358") @@ -2134,6 +2328,7 @@ class Program_weight_tensor_parameter_193: class Program_weight_tensor_parameter_194: name = "parameter_194" + original_name = "layer_norm_1455.b_0" shape = [1024] dtype = "float32" min_val = float("-0.531196") @@ -2145,6 +2340,7 @@ class Program_weight_tensor_parameter_194: class Program_weight_tensor_parameter_195: name = "parameter_195" + original_name = "layer_norm_1455.w_0" shape = [1024] dtype = "float32" min_val = float("0.369097") @@ -2156,6 +2352,7 @@ class Program_weight_tensor_parameter_195: class Program_weight_tensor_parameter_196: name = "parameter_196" + original_name = "layer_norm_1454.b_0" shape = [1024] dtype = "float32" min_val = float("-1.49472") @@ -2167,6 +2364,7 @@ class Program_weight_tensor_parameter_196: class Program_weight_tensor_parameter_197: name = "parameter_197" + original_name = "layer_norm_1454.w_0" shape = [1024] dtype = "float32" min_val = float("0.674368") @@ -2178,6 +2376,7 @@ class Program_weight_tensor_parameter_197: class Program_weight_tensor_parameter_198: name = "parameter_198" + original_name = "linear_4218.b_0" shape = [1024] dtype = "float32" min_val = float("-0.645507") @@ -2189,6 +2388,7 @@ class Program_weight_tensor_parameter_198: class Program_weight_tensor_parameter_199: name = "parameter_199" + original_name = "linear_4218.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-1.36587") @@ -2200,6 +2400,7 @@ class Program_weight_tensor_parameter_199: class Program_weight_tensor_parameter_200: name = "parameter_200" + original_name = "linear_4217.b_0" shape = [4096] dtype = "float32" min_val = float("-0.248276") @@ -2211,6 +2412,7 @@ class Program_weight_tensor_parameter_200: class Program_weight_tensor_parameter_201: name = "parameter_201" + original_name = "linear_4217.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.604979") @@ -2222,6 +2424,7 @@ class Program_weight_tensor_parameter_201: class Program_weight_tensor_parameter_202: name = "parameter_202" + original_name = "linear_4216.b_0" shape = [1024] dtype = "float32" min_val = float("-0.254382") @@ -2233,6 +2436,7 @@ class Program_weight_tensor_parameter_202: class Program_weight_tensor_parameter_203: name = "parameter_203" + original_name = "linear_4216.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.345991") @@ -2244,6 +2448,7 @@ class Program_weight_tensor_parameter_203: class Program_weight_tensor_parameter_204: name = "parameter_204" + original_name = "linear_4215.b_0" shape = [1024] dtype = "float32" min_val = float("-0.18094") @@ -2255,6 +2460,7 @@ class Program_weight_tensor_parameter_204: class Program_weight_tensor_parameter_205: name = "parameter_205" + original_name = "linear_4215.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.163922") @@ -2266,6 +2472,7 @@ class Program_weight_tensor_parameter_205: class Program_weight_tensor_parameter_206: name = "parameter_206" + original_name = "linear_4214.b_0" shape = [1024] dtype = "float32" min_val = float("-0.898337") @@ -2277,6 +2484,7 @@ class Program_weight_tensor_parameter_206: class Program_weight_tensor_parameter_207: name = "parameter_207" + original_name = "linear_4214.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.315127") @@ -2288,6 +2496,7 @@ class Program_weight_tensor_parameter_207: class Program_weight_tensor_parameter_208: name = "parameter_208" + original_name = "linear_4213.b_0" shape = [1024] dtype = "float32" min_val = float("-0.464097") @@ -2299,6 +2508,7 @@ class Program_weight_tensor_parameter_208: class Program_weight_tensor_parameter_209: name = "parameter_209" + original_name = "linear_4213.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.339363") @@ -2310,6 +2520,7 @@ class Program_weight_tensor_parameter_209: class Program_weight_tensor_parameter_210: name = "parameter_210" + original_name = "layer_norm_1453.b_0" shape = [1024] dtype = "float32" min_val = float("-0.62676") @@ -2321,6 +2532,7 @@ class Program_weight_tensor_parameter_210: class Program_weight_tensor_parameter_211: name = "parameter_211" + original_name = "layer_norm_1453.w_0" shape = [1024] dtype = "float32" min_val = float("0.425947") @@ -2332,6 +2544,7 @@ class Program_weight_tensor_parameter_211: class Program_weight_tensor_parameter_212: name = "parameter_212" + original_name = "layer_norm_1452.b_0" shape = [1024] dtype = "float32" min_val = float("-1.37181") @@ -2343,6 +2556,7 @@ class Program_weight_tensor_parameter_212: class Program_weight_tensor_parameter_213: name = "parameter_213" + original_name = "layer_norm_1452.w_0" shape = [1024] dtype = "float32" min_val = float("0.68425") @@ -2354,6 +2568,7 @@ class Program_weight_tensor_parameter_213: class Program_weight_tensor_parameter_214: name = "parameter_214" + original_name = "linear_4212.b_0" shape = [1024] dtype = "float32" min_val = float("-0.67931") @@ -2365,6 +2580,7 @@ class Program_weight_tensor_parameter_214: class Program_weight_tensor_parameter_215: name = "parameter_215" + original_name = "linear_4212.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-1.37384") @@ -2376,6 +2592,7 @@ class Program_weight_tensor_parameter_215: class Program_weight_tensor_parameter_216: name = "parameter_216" + original_name = "linear_4211.b_0" shape = [4096] dtype = "float32" min_val = float("-0.254628") @@ -2387,6 +2604,7 @@ class Program_weight_tensor_parameter_216: class Program_weight_tensor_parameter_217: name = "parameter_217" + original_name = "linear_4211.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.500791") @@ -2398,6 +2616,7 @@ class Program_weight_tensor_parameter_217: class Program_weight_tensor_parameter_218: name = "parameter_218" + original_name = "linear_4210.b_0" shape = [1024] dtype = "float32" min_val = float("-0.233311") @@ -2409,6 +2628,7 @@ class Program_weight_tensor_parameter_218: class Program_weight_tensor_parameter_219: name = "parameter_219" + original_name = "linear_4210.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.288304") @@ -2420,6 +2640,7 @@ class Program_weight_tensor_parameter_219: class Program_weight_tensor_parameter_220: name = "parameter_220" + original_name = "linear_4209.b_0" shape = [1024] dtype = "float32" min_val = float("-0.163894") @@ -2431,6 +2652,7 @@ class Program_weight_tensor_parameter_220: class Program_weight_tensor_parameter_221: name = "parameter_221" + original_name = "linear_4209.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.153529") @@ -2442,6 +2664,7 @@ class Program_weight_tensor_parameter_221: class Program_weight_tensor_parameter_222: name = "parameter_222" + original_name = "linear_4208.b_0" shape = [1024] dtype = "float32" min_val = float("-0.996175") @@ -2453,6 +2676,7 @@ class Program_weight_tensor_parameter_222: class Program_weight_tensor_parameter_223: name = "parameter_223" + original_name = "linear_4208.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.284108") @@ -2464,6 +2688,7 @@ class Program_weight_tensor_parameter_223: class Program_weight_tensor_parameter_224: name = "parameter_224" + original_name = "linear_4207.b_0" shape = [1024] dtype = "float32" min_val = float("-0.385222") @@ -2475,6 +2700,7 @@ class Program_weight_tensor_parameter_224: class Program_weight_tensor_parameter_225: name = "parameter_225" + original_name = "linear_4207.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.257782") @@ -2486,6 +2712,7 @@ class Program_weight_tensor_parameter_225: class Program_weight_tensor_parameter_226: name = "parameter_226" + original_name = "layer_norm_1451.b_0" shape = [1024] dtype = "float32" min_val = float("-0.418317") @@ -2497,6 +2724,7 @@ class Program_weight_tensor_parameter_226: class Program_weight_tensor_parameter_227: name = "parameter_227" + original_name = "layer_norm_1451.w_0" shape = [1024] dtype = "float32" min_val = float("0.473695") @@ -2508,6 +2736,7 @@ class Program_weight_tensor_parameter_227: class Program_weight_tensor_parameter_228: name = "parameter_228" + original_name = "layer_norm_1450.b_0" shape = [1024] dtype = "float32" min_val = float("-1.67865") @@ -2519,6 +2748,7 @@ class Program_weight_tensor_parameter_228: class Program_weight_tensor_parameter_229: name = "parameter_229" + original_name = "layer_norm_1450.w_0" shape = [1024] dtype = "float32" min_val = float("0.737631") @@ -2530,6 +2760,7 @@ class Program_weight_tensor_parameter_229: class Program_weight_tensor_parameter_230: name = "parameter_230" + original_name = "linear_4206.b_0" shape = [1024] dtype = "float32" min_val = float("-0.323895") @@ -2541,6 +2772,7 @@ class Program_weight_tensor_parameter_230: class Program_weight_tensor_parameter_231: name = "parameter_231" + original_name = "linear_4206.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-0.909487") @@ -2552,6 +2784,7 @@ class Program_weight_tensor_parameter_231: class Program_weight_tensor_parameter_232: name = "parameter_232" + original_name = "linear_4205.b_0" shape = [4096] dtype = "float32" min_val = float("-0.199087") @@ -2563,6 +2796,7 @@ class Program_weight_tensor_parameter_232: class Program_weight_tensor_parameter_233: name = "parameter_233" + original_name = "linear_4205.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.503095") @@ -2574,6 +2808,7 @@ class Program_weight_tensor_parameter_233: class Program_weight_tensor_parameter_234: name = "parameter_234" + original_name = "linear_4204.b_0" shape = [1024] dtype = "float32" min_val = float("-0.168262") @@ -2585,6 +2820,7 @@ class Program_weight_tensor_parameter_234: class Program_weight_tensor_parameter_235: name = "parameter_235" + original_name = "linear_4204.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.253221") @@ -2596,6 +2832,7 @@ class Program_weight_tensor_parameter_235: class Program_weight_tensor_parameter_236: name = "parameter_236" + original_name = "linear_4203.b_0" shape = [1024] dtype = "float32" min_val = float("-0.22688") @@ -2607,6 +2844,7 @@ class Program_weight_tensor_parameter_236: class Program_weight_tensor_parameter_237: name = "parameter_237" + original_name = "linear_4203.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.229827") @@ -2618,6 +2856,7 @@ class Program_weight_tensor_parameter_237: class Program_weight_tensor_parameter_238: name = "parameter_238" + original_name = "linear_4202.b_0" shape = [1024] dtype = "float32" min_val = float("-1.06593") @@ -2629,6 +2868,7 @@ class Program_weight_tensor_parameter_238: class Program_weight_tensor_parameter_239: name = "parameter_239" + original_name = "linear_4202.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.293151") @@ -2640,6 +2880,7 @@ class Program_weight_tensor_parameter_239: class Program_weight_tensor_parameter_240: name = "parameter_240" + original_name = "linear_4201.b_0" shape = [1024] dtype = "float32" min_val = float("-0.411382") @@ -2651,6 +2892,7 @@ class Program_weight_tensor_parameter_240: class Program_weight_tensor_parameter_241: name = "parameter_241" + original_name = "linear_4201.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.289258") @@ -2662,6 +2904,7 @@ class Program_weight_tensor_parameter_241: class Program_weight_tensor_parameter_242: name = "parameter_242" + original_name = "layer_norm_1449.b_0" shape = [1024] dtype = "float32" min_val = float("-0.628532") @@ -2673,6 +2916,7 @@ class Program_weight_tensor_parameter_242: class Program_weight_tensor_parameter_243: name = "parameter_243" + original_name = "layer_norm_1449.w_0" shape = [1024] dtype = "float32" min_val = float("0.484172") @@ -2684,6 +2928,7 @@ class Program_weight_tensor_parameter_243: class Program_weight_tensor_parameter_244: name = "parameter_244" + original_name = "layer_norm_1448.b_0" shape = [1024] dtype = "float32" min_val = float("-1.6866") @@ -2695,6 +2940,7 @@ class Program_weight_tensor_parameter_244: class Program_weight_tensor_parameter_245: name = "parameter_245" + original_name = "layer_norm_1448.w_0" shape = [1024] dtype = "float32" min_val = float("0.631271") @@ -2706,6 +2952,7 @@ class Program_weight_tensor_parameter_245: class Program_weight_tensor_parameter_246: name = "parameter_246" + original_name = "linear_4200.b_0" shape = [1024] dtype = "float32" min_val = float("-0.175902") @@ -2717,6 +2964,7 @@ class Program_weight_tensor_parameter_246: class Program_weight_tensor_parameter_247: name = "parameter_247" + original_name = "linear_4200.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-1.65219") @@ -2728,6 +2976,7 @@ class Program_weight_tensor_parameter_247: class Program_weight_tensor_parameter_248: name = "parameter_248" + original_name = "linear_4199.b_0" shape = [4096] dtype = "float32" min_val = float("-0.192937") @@ -2739,6 +2988,7 @@ class Program_weight_tensor_parameter_248: class Program_weight_tensor_parameter_249: name = "parameter_249" + original_name = "linear_4199.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.305946") @@ -2750,6 +3000,7 @@ class Program_weight_tensor_parameter_249: class Program_weight_tensor_parameter_250: name = "parameter_250" + original_name = "linear_4198.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0780903") @@ -2761,6 +3012,7 @@ class Program_weight_tensor_parameter_250: class Program_weight_tensor_parameter_251: name = "parameter_251" + original_name = "linear_4198.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.517792") @@ -2772,6 +3024,7 @@ class Program_weight_tensor_parameter_251: class Program_weight_tensor_parameter_252: name = "parameter_252" + original_name = "linear_4197.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0640666") @@ -2783,6 +3036,7 @@ class Program_weight_tensor_parameter_252: class Program_weight_tensor_parameter_253: name = "parameter_253" + original_name = "linear_4197.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.14597") @@ -2794,6 +3048,7 @@ class Program_weight_tensor_parameter_253: class Program_weight_tensor_parameter_254: name = "parameter_254" + original_name = "linear_4196.b_0" shape = [1024] dtype = "float32" min_val = float("-0.869517") @@ -2805,6 +3060,7 @@ class Program_weight_tensor_parameter_254: class Program_weight_tensor_parameter_255: name = "parameter_255" + original_name = "linear_4196.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.299447") @@ -2816,6 +3072,7 @@ class Program_weight_tensor_parameter_255: class Program_weight_tensor_parameter_256: name = "parameter_256" + original_name = "linear_4195.b_0" shape = [1024] dtype = "float32" min_val = float("-0.537479") @@ -2827,6 +3084,7 @@ class Program_weight_tensor_parameter_256: class Program_weight_tensor_parameter_257: name = "parameter_257" + original_name = "linear_4195.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.287265") @@ -2838,6 +3096,7 @@ class Program_weight_tensor_parameter_257: class Program_weight_tensor_parameter_258: name = "parameter_258" + original_name = "layer_norm_1447.b_0" shape = [1024] dtype = "float32" min_val = float("-1.14141") @@ -2849,6 +3108,7 @@ class Program_weight_tensor_parameter_258: class Program_weight_tensor_parameter_259: name = "parameter_259" + original_name = "layer_norm_1447.w_0" shape = [1024] dtype = "float32" min_val = float("0.107967") @@ -2860,6 +3120,7 @@ class Program_weight_tensor_parameter_259: class Program_weight_tensor_parameter_260: name = "parameter_260" + original_name = "layer_norm_1446.b_0" shape = [1024] dtype = "float32" min_val = float("-1.6763") @@ -2871,6 +3132,7 @@ class Program_weight_tensor_parameter_260: class Program_weight_tensor_parameter_261: name = "parameter_261" + original_name = "layer_norm_1446.w_0" shape = [1024] dtype = "float32" min_val = float("0.811945") @@ -2882,6 +3144,7 @@ class Program_weight_tensor_parameter_261: class Program_weight_tensor_parameter_262: name = "parameter_262" + original_name = "linear_4194.b_0" shape = [1024] dtype = "float32" min_val = float("-0.197175") @@ -2893,6 +3156,7 @@ class Program_weight_tensor_parameter_262: class Program_weight_tensor_parameter_263: name = "parameter_263" + original_name = "linear_4194.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-1.27499") @@ -2904,6 +3168,7 @@ class Program_weight_tensor_parameter_263: class Program_weight_tensor_parameter_264: name = "parameter_264" + original_name = "linear_4193.b_0" shape = [4096] dtype = "float32" min_val = float("-0.218396") @@ -2915,6 +3180,7 @@ class Program_weight_tensor_parameter_264: class Program_weight_tensor_parameter_265: name = "parameter_265" + original_name = "linear_4193.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.438392") @@ -2926,6 +3192,7 @@ class Program_weight_tensor_parameter_265: class Program_weight_tensor_parameter_266: name = "parameter_266" + original_name = "linear_4192.b_0" shape = [1024] dtype = "float32" min_val = float("-0.122267") @@ -2937,6 +3204,7 @@ class Program_weight_tensor_parameter_266: class Program_weight_tensor_parameter_267: name = "parameter_267" + original_name = "linear_4192.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.234361") @@ -2948,6 +3216,7 @@ class Program_weight_tensor_parameter_267: class Program_weight_tensor_parameter_268: name = "parameter_268" + original_name = "linear_4191.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0412184") @@ -2959,6 +3228,7 @@ class Program_weight_tensor_parameter_268: class Program_weight_tensor_parameter_269: name = "parameter_269" + original_name = "linear_4191.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.156132") @@ -2970,6 +3240,7 @@ class Program_weight_tensor_parameter_269: class Program_weight_tensor_parameter_270: name = "parameter_270" + original_name = "linear_4190.b_0" shape = [1024] dtype = "float32" min_val = float("-0.50983") @@ -2981,6 +3252,7 @@ class Program_weight_tensor_parameter_270: class Program_weight_tensor_parameter_271: name = "parameter_271" + original_name = "linear_4190.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.338421") @@ -2992,6 +3264,7 @@ class Program_weight_tensor_parameter_271: class Program_weight_tensor_parameter_272: name = "parameter_272" + original_name = "linear_4189.b_0" shape = [1024] dtype = "float32" min_val = float("-0.354379") @@ -3003,6 +3276,7 @@ class Program_weight_tensor_parameter_272: class Program_weight_tensor_parameter_273: name = "parameter_273" + original_name = "linear_4189.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.330167") @@ -3014,6 +3288,7 @@ class Program_weight_tensor_parameter_273: class Program_weight_tensor_parameter_274: name = "parameter_274" + original_name = "layer_norm_1445.b_0" shape = [1024] dtype = "float32" min_val = float("-1.28205") @@ -3025,6 +3300,7 @@ class Program_weight_tensor_parameter_274: class Program_weight_tensor_parameter_275: name = "parameter_275" + original_name = "layer_norm_1445.w_0" shape = [1024] dtype = "float32" min_val = float("0.109015") @@ -3036,6 +3312,7 @@ class Program_weight_tensor_parameter_275: class Program_weight_tensor_parameter_276: name = "parameter_276" + original_name = "layer_norm_1444.b_0" shape = [1024] dtype = "float32" min_val = float("-1.40185") @@ -3047,6 +3324,7 @@ class Program_weight_tensor_parameter_276: class Program_weight_tensor_parameter_277: name = "parameter_277" + original_name = "layer_norm_1444.w_0" shape = [1024] dtype = "float32" min_val = float("0.746415") @@ -3058,6 +3336,7 @@ class Program_weight_tensor_parameter_277: class Program_weight_tensor_parameter_278: name = "parameter_278" + original_name = "linear_4188.b_0" shape = [1024] dtype = "float32" min_val = float("-0.196943") @@ -3069,6 +3348,7 @@ class Program_weight_tensor_parameter_278: class Program_weight_tensor_parameter_279: name = "parameter_279" + original_name = "linear_4188.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-0.966355") @@ -3080,6 +3360,7 @@ class Program_weight_tensor_parameter_279: class Program_weight_tensor_parameter_280: name = "parameter_280" + original_name = "linear_4187.b_0" shape = [4096] dtype = "float32" min_val = float("-0.168917") @@ -3091,6 +3372,7 @@ class Program_weight_tensor_parameter_280: class Program_weight_tensor_parameter_281: name = "parameter_281" + original_name = "linear_4187.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.200901") @@ -3102,6 +3384,7 @@ class Program_weight_tensor_parameter_281: class Program_weight_tensor_parameter_282: name = "parameter_282" + original_name = "linear_4186.b_0" shape = [1024] dtype = "float32" min_val = float("-0.140392") @@ -3113,6 +3396,7 @@ class Program_weight_tensor_parameter_282: class Program_weight_tensor_parameter_283: name = "parameter_283" + original_name = "linear_4186.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.214255") @@ -3124,6 +3408,7 @@ class Program_weight_tensor_parameter_283: class Program_weight_tensor_parameter_284: name = "parameter_284" + original_name = "linear_4185.b_0" shape = [1024] dtype = "float32" min_val = float("-0.042029") @@ -3135,6 +3420,7 @@ class Program_weight_tensor_parameter_284: class Program_weight_tensor_parameter_285: name = "parameter_285" + original_name = "linear_4185.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.140948") @@ -3146,6 +3432,7 @@ class Program_weight_tensor_parameter_285: class Program_weight_tensor_parameter_286: name = "parameter_286" + original_name = "linear_4184.b_0" shape = [1024] dtype = "float32" min_val = float("-0.664997") @@ -3157,6 +3444,7 @@ class Program_weight_tensor_parameter_286: class Program_weight_tensor_parameter_287: name = "parameter_287" + original_name = "linear_4184.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.515756") @@ -3168,6 +3456,7 @@ class Program_weight_tensor_parameter_287: class Program_weight_tensor_parameter_288: name = "parameter_288" + original_name = "linear_4183.b_0" shape = [1024] dtype = "float32" min_val = float("-0.396869") @@ -3179,6 +3468,7 @@ class Program_weight_tensor_parameter_288: class Program_weight_tensor_parameter_289: name = "parameter_289" + original_name = "linear_4183.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.311154") @@ -3190,6 +3480,7 @@ class Program_weight_tensor_parameter_289: class Program_weight_tensor_parameter_290: name = "parameter_290" + original_name = "layer_norm_1443.b_0" shape = [1024] dtype = "float32" min_val = float("-1.19393") @@ -3201,6 +3492,7 @@ class Program_weight_tensor_parameter_290: class Program_weight_tensor_parameter_291: name = "parameter_291" + original_name = "layer_norm_1443.w_0" shape = [1024] dtype = "float32" min_val = float("0.185133") @@ -3212,6 +3504,7 @@ class Program_weight_tensor_parameter_291: class Program_weight_tensor_parameter_292: name = "parameter_292" + original_name = "layer_norm_1442.b_0" shape = [1024] dtype = "float32" min_val = float("-0.910721") @@ -3223,6 +3516,7 @@ class Program_weight_tensor_parameter_292: class Program_weight_tensor_parameter_293: name = "parameter_293" + original_name = "layer_norm_1442.w_0" shape = [1024] dtype = "float32" min_val = float("0.736488") @@ -3234,6 +3528,7 @@ class Program_weight_tensor_parameter_293: class Program_weight_tensor_parameter_294: name = "parameter_294" + original_name = "linear_4182.b_0" shape = [1024] dtype = "float32" min_val = float("-0.233126") @@ -3245,6 +3540,7 @@ class Program_weight_tensor_parameter_294: class Program_weight_tensor_parameter_295: name = "parameter_295" + original_name = "linear_4182.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-0.744295") @@ -3256,6 +3552,7 @@ class Program_weight_tensor_parameter_295: class Program_weight_tensor_parameter_296: name = "parameter_296" + original_name = "linear_4181.b_0" shape = [4096] dtype = "float32" min_val = float("-0.148187") @@ -3267,6 +3564,7 @@ class Program_weight_tensor_parameter_296: class Program_weight_tensor_parameter_297: name = "parameter_297" + original_name = "linear_4181.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.235695") @@ -3278,6 +3576,7 @@ class Program_weight_tensor_parameter_297: class Program_weight_tensor_parameter_298: name = "parameter_298" + original_name = "linear_4180.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0699128") @@ -3289,6 +3588,7 @@ class Program_weight_tensor_parameter_298: class Program_weight_tensor_parameter_299: name = "parameter_299" + original_name = "linear_4180.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.20148") @@ -3300,6 +3600,7 @@ class Program_weight_tensor_parameter_299: class Program_weight_tensor_parameter_300: name = "parameter_300" + original_name = "linear_4179.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0400432") @@ -3311,6 +3612,7 @@ class Program_weight_tensor_parameter_300: class Program_weight_tensor_parameter_301: name = "parameter_301" + original_name = "linear_4179.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.159062") @@ -3322,6 +3624,7 @@ class Program_weight_tensor_parameter_301: class Program_weight_tensor_parameter_302: name = "parameter_302" + original_name = "linear_4178.b_0" shape = [1024] dtype = "float32" min_val = float("-0.559755") @@ -3333,6 +3636,7 @@ class Program_weight_tensor_parameter_302: class Program_weight_tensor_parameter_303: name = "parameter_303" + original_name = "linear_4178.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.311119") @@ -3344,6 +3648,7 @@ class Program_weight_tensor_parameter_303: class Program_weight_tensor_parameter_304: name = "parameter_304" + original_name = "linear_4177.b_0" shape = [1024] dtype = "float32" min_val = float("-0.363277") @@ -3355,6 +3660,7 @@ class Program_weight_tensor_parameter_304: class Program_weight_tensor_parameter_305: name = "parameter_305" + original_name = "linear_4177.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.311855") @@ -3366,6 +3672,7 @@ class Program_weight_tensor_parameter_305: class Program_weight_tensor_parameter_306: name = "parameter_306" + original_name = "layer_norm_1441.b_0" shape = [1024] dtype = "float32" min_val = float("-0.480509") @@ -3377,6 +3684,7 @@ class Program_weight_tensor_parameter_306: class Program_weight_tensor_parameter_307: name = "parameter_307" + original_name = "layer_norm_1441.w_0" shape = [1024] dtype = "float32" min_val = float("0.19545") @@ -3388,6 +3696,7 @@ class Program_weight_tensor_parameter_307: class Program_weight_tensor_parameter_308: name = "parameter_308" + original_name = "layer_norm_1440.b_0" shape = [1024] dtype = "float32" min_val = float("-0.60092") @@ -3399,6 +3708,7 @@ class Program_weight_tensor_parameter_308: class Program_weight_tensor_parameter_309: name = "parameter_309" + original_name = "layer_norm_1440.w_0" shape = [1024] dtype = "float32" min_val = float("0.716654") @@ -3410,6 +3720,7 @@ class Program_weight_tensor_parameter_309: class Program_weight_tensor_parameter_310: name = "parameter_310" + original_name = "linear_4176.b_0" shape = [1024] dtype = "float32" min_val = float("-0.227494") @@ -3421,6 +3732,7 @@ class Program_weight_tensor_parameter_310: class Program_weight_tensor_parameter_311: name = "parameter_311" + original_name = "linear_4176.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-0.602849") @@ -3432,6 +3744,7 @@ class Program_weight_tensor_parameter_311: class Program_weight_tensor_parameter_312: name = "parameter_312" + original_name = "linear_4175.b_0" shape = [4096] dtype = "float32" min_val = float("-0.16072") @@ -3443,6 +3756,7 @@ class Program_weight_tensor_parameter_312: class Program_weight_tensor_parameter_313: name = "parameter_313" + original_name = "linear_4175.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.256746") @@ -3454,6 +3768,7 @@ class Program_weight_tensor_parameter_313: class Program_weight_tensor_parameter_314: name = "parameter_314" + original_name = "linear_4174.b_0" shape = [1024] dtype = "float32" min_val = float("-0.109753") @@ -3465,6 +3780,7 @@ class Program_weight_tensor_parameter_314: class Program_weight_tensor_parameter_315: name = "parameter_315" + original_name = "linear_4174.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.173869") @@ -3476,6 +3792,7 @@ class Program_weight_tensor_parameter_315: class Program_weight_tensor_parameter_316: name = "parameter_316" + original_name = "linear_4173.b_0" shape = [1024] dtype = "float32" min_val = float("-0.141426") @@ -3487,6 +3804,7 @@ class Program_weight_tensor_parameter_316: class Program_weight_tensor_parameter_317: name = "parameter_317" + original_name = "linear_4173.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.151535") @@ -3498,6 +3816,7 @@ class Program_weight_tensor_parameter_317: class Program_weight_tensor_parameter_318: name = "parameter_318" + original_name = "linear_4172.b_0" shape = [1024] dtype = "float32" min_val = float("-0.493024") @@ -3509,6 +3828,7 @@ class Program_weight_tensor_parameter_318: class Program_weight_tensor_parameter_319: name = "parameter_319" + original_name = "linear_4172.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.318825") @@ -3520,6 +3840,7 @@ class Program_weight_tensor_parameter_319: class Program_weight_tensor_parameter_320: name = "parameter_320" + original_name = "linear_4171.b_0" shape = [1024] dtype = "float32" min_val = float("-0.331325") @@ -3531,6 +3852,7 @@ class Program_weight_tensor_parameter_320: class Program_weight_tensor_parameter_321: name = "parameter_321" + original_name = "linear_4171.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.290914") @@ -3542,6 +3864,7 @@ class Program_weight_tensor_parameter_321: class Program_weight_tensor_parameter_322: name = "parameter_322" + original_name = "layer_norm_1439.b_0" shape = [1024] dtype = "float32" min_val = float("-0.202867") @@ -3553,6 +3876,7 @@ class Program_weight_tensor_parameter_322: class Program_weight_tensor_parameter_323: name = "parameter_323" + original_name = "layer_norm_1439.w_0" shape = [1024] dtype = "float32" min_val = float("0.327822") @@ -3564,6 +3888,7 @@ class Program_weight_tensor_parameter_323: class Program_weight_tensor_parameter_324: name = "parameter_324" + original_name = "layer_norm_1438.b_0" shape = [1024] dtype = "float32" min_val = float("-0.758603") @@ -3575,6 +3900,7 @@ class Program_weight_tensor_parameter_324: class Program_weight_tensor_parameter_325: name = "parameter_325" + original_name = "layer_norm_1438.w_0" shape = [1024] dtype = "float32" min_val = float("0.761949") @@ -3586,6 +3912,7 @@ class Program_weight_tensor_parameter_325: class Program_weight_tensor_parameter_326: name = "parameter_326" + original_name = "linear_4170.b_0" shape = [1024] dtype = "float32" min_val = float("-0.116091") @@ -3597,6 +3924,7 @@ class Program_weight_tensor_parameter_326: class Program_weight_tensor_parameter_327: name = "parameter_327" + original_name = "linear_4170.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-0.632057") @@ -3608,6 +3936,7 @@ class Program_weight_tensor_parameter_327: class Program_weight_tensor_parameter_328: name = "parameter_328" + original_name = "linear_4169.b_0" shape = [4096] dtype = "float32" min_val = float("-0.172011") @@ -3619,6 +3948,7 @@ class Program_weight_tensor_parameter_328: class Program_weight_tensor_parameter_329: name = "parameter_329" + original_name = "linear_4169.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.224164") @@ -3630,6 +3960,7 @@ class Program_weight_tensor_parameter_329: class Program_weight_tensor_parameter_330: name = "parameter_330" + original_name = "linear_4168.b_0" shape = [1024] dtype = "float32" min_val = float("-0.236658") @@ -3641,6 +3972,7 @@ class Program_weight_tensor_parameter_330: class Program_weight_tensor_parameter_331: name = "parameter_331" + original_name = "linear_4168.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.19053") @@ -3652,6 +3984,7 @@ class Program_weight_tensor_parameter_331: class Program_weight_tensor_parameter_332: name = "parameter_332" + original_name = "linear_4167.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0867448") @@ -3663,6 +3996,7 @@ class Program_weight_tensor_parameter_332: class Program_weight_tensor_parameter_333: name = "parameter_333" + original_name = "linear_4167.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.164963") @@ -3674,6 +4008,7 @@ class Program_weight_tensor_parameter_333: class Program_weight_tensor_parameter_334: name = "parameter_334" + original_name = "linear_4166.b_0" shape = [1024] dtype = "float32" min_val = float("-0.445493") @@ -3685,6 +4020,7 @@ class Program_weight_tensor_parameter_334: class Program_weight_tensor_parameter_335: name = "parameter_335" + original_name = "linear_4166.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.230191") @@ -3696,6 +4032,7 @@ class Program_weight_tensor_parameter_335: class Program_weight_tensor_parameter_336: name = "parameter_336" + original_name = "linear_4165.b_0" shape = [1024] dtype = "float32" min_val = float("-0.292476") @@ -3707,6 +4044,7 @@ class Program_weight_tensor_parameter_336: class Program_weight_tensor_parameter_337: name = "parameter_337" + original_name = "linear_4165.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.255627") @@ -3718,6 +4056,7 @@ class Program_weight_tensor_parameter_337: class Program_weight_tensor_parameter_338: name = "parameter_338" + original_name = "layer_norm_1437.b_0" shape = [1024] dtype = "float32" min_val = float("-0.2759") @@ -3729,6 +4068,7 @@ class Program_weight_tensor_parameter_338: class Program_weight_tensor_parameter_339: name = "parameter_339" + original_name = "layer_norm_1437.w_0" shape = [1024] dtype = "float32" min_val = float("0.385794") @@ -3740,6 +4080,7 @@ class Program_weight_tensor_parameter_339: class Program_weight_tensor_parameter_340: name = "parameter_340" + original_name = "layer_norm_1436.b_0" shape = [1024] dtype = "float32" min_val = float("-1.14448") @@ -3751,6 +4092,7 @@ class Program_weight_tensor_parameter_340: class Program_weight_tensor_parameter_341: name = "parameter_341" + original_name = "layer_norm_1436.w_0" shape = [1024] dtype = "float32" min_val = float("0.826475") @@ -3762,6 +4104,7 @@ class Program_weight_tensor_parameter_341: class Program_weight_tensor_parameter_342: name = "parameter_342" + original_name = "linear_4164.b_0" shape = [1024] dtype = "float32" min_val = float("-0.329948") @@ -3773,6 +4116,7 @@ class Program_weight_tensor_parameter_342: class Program_weight_tensor_parameter_343: name = "parameter_343" + original_name = "linear_4164.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-0.597207") @@ -3784,6 +4128,7 @@ class Program_weight_tensor_parameter_343: class Program_weight_tensor_parameter_344: name = "parameter_344" + original_name = "linear_4163.b_0" shape = [4096] dtype = "float32" min_val = float("-0.13878") @@ -3795,6 +4140,7 @@ class Program_weight_tensor_parameter_344: class Program_weight_tensor_parameter_345: name = "parameter_345" + original_name = "linear_4163.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.22832") @@ -3806,6 +4152,7 @@ class Program_weight_tensor_parameter_345: class Program_weight_tensor_parameter_346: name = "parameter_346" + original_name = "linear_4162.b_0" shape = [1024] dtype = "float32" min_val = float("-0.138323") @@ -3817,6 +4164,7 @@ class Program_weight_tensor_parameter_346: class Program_weight_tensor_parameter_347: name = "parameter_347" + original_name = "linear_4162.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.268011") @@ -3828,6 +4176,7 @@ class Program_weight_tensor_parameter_347: class Program_weight_tensor_parameter_348: name = "parameter_348" + original_name = "linear_4161.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0914921") @@ -3839,6 +4188,7 @@ class Program_weight_tensor_parameter_348: class Program_weight_tensor_parameter_349: name = "parameter_349" + original_name = "linear_4161.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.1825") @@ -3850,6 +4200,7 @@ class Program_weight_tensor_parameter_349: class Program_weight_tensor_parameter_350: name = "parameter_350" + original_name = "linear_4160.b_0" shape = [1024] dtype = "float32" min_val = float("-0.398588") @@ -3861,6 +4212,7 @@ class Program_weight_tensor_parameter_350: class Program_weight_tensor_parameter_351: name = "parameter_351" + original_name = "linear_4160.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.336199") @@ -3872,6 +4224,7 @@ class Program_weight_tensor_parameter_351: class Program_weight_tensor_parameter_352: name = "parameter_352" + original_name = "linear_4159.b_0" shape = [1024] dtype = "float32" min_val = float("-0.309271") @@ -3883,6 +4236,7 @@ class Program_weight_tensor_parameter_352: class Program_weight_tensor_parameter_353: name = "parameter_353" + original_name = "linear_4159.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.264398") @@ -3894,6 +4248,7 @@ class Program_weight_tensor_parameter_353: class Program_weight_tensor_parameter_354: name = "parameter_354" + original_name = "layer_norm_1435.b_0" shape = [1024] dtype = "float32" min_val = float("-0.248926") @@ -3905,6 +4260,7 @@ class Program_weight_tensor_parameter_354: class Program_weight_tensor_parameter_355: name = "parameter_355" + original_name = "layer_norm_1435.w_0" shape = [1024] dtype = "float32" min_val = float("0.401081") @@ -3916,6 +4272,7 @@ class Program_weight_tensor_parameter_355: class Program_weight_tensor_parameter_356: name = "parameter_356" + original_name = "layer_norm_1434.b_0" shape = [1024] dtype = "float32" min_val = float("-1.3172") @@ -3927,6 +4284,7 @@ class Program_weight_tensor_parameter_356: class Program_weight_tensor_parameter_357: name = "parameter_357" + original_name = "layer_norm_1434.w_0" shape = [1024] dtype = "float32" min_val = float("0.739302") @@ -3938,6 +4296,7 @@ class Program_weight_tensor_parameter_357: class Program_weight_tensor_parameter_358: name = "parameter_358" + original_name = "linear_4158.b_0" shape = [1024] dtype = "float32" min_val = float("-0.473035") @@ -3949,6 +4308,7 @@ class Program_weight_tensor_parameter_358: class Program_weight_tensor_parameter_359: name = "parameter_359" + original_name = "linear_4158.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-0.481692") @@ -3960,6 +4320,7 @@ class Program_weight_tensor_parameter_359: class Program_weight_tensor_parameter_360: name = "parameter_360" + original_name = "linear_4157.b_0" shape = [4096] dtype = "float32" min_val = float("-0.133138") @@ -3971,6 +4332,7 @@ class Program_weight_tensor_parameter_360: class Program_weight_tensor_parameter_361: name = "parameter_361" + original_name = "linear_4157.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.209075") @@ -3982,6 +4344,7 @@ class Program_weight_tensor_parameter_361: class Program_weight_tensor_parameter_362: name = "parameter_362" + original_name = "linear_4156.b_0" shape = [1024] dtype = "float32" min_val = float("-0.160528") @@ -3993,6 +4356,7 @@ class Program_weight_tensor_parameter_362: class Program_weight_tensor_parameter_363: name = "parameter_363" + original_name = "linear_4156.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.79163") @@ -4004,6 +4368,7 @@ class Program_weight_tensor_parameter_363: class Program_weight_tensor_parameter_364: name = "parameter_364" + original_name = "linear_4155.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0517538") @@ -4015,6 +4380,7 @@ class Program_weight_tensor_parameter_364: class Program_weight_tensor_parameter_365: name = "parameter_365" + original_name = "linear_4155.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.141821") @@ -4026,6 +4392,7 @@ class Program_weight_tensor_parameter_365: class Program_weight_tensor_parameter_366: name = "parameter_366" + original_name = "linear_4154.b_0" shape = [1024] dtype = "float32" min_val = float("-0.413759") @@ -4037,6 +4404,7 @@ class Program_weight_tensor_parameter_366: class Program_weight_tensor_parameter_367: name = "parameter_367" + original_name = "linear_4154.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.442828") @@ -4048,6 +4416,7 @@ class Program_weight_tensor_parameter_367: class Program_weight_tensor_parameter_368: name = "parameter_368" + original_name = "linear_4153.b_0" shape = [1024] dtype = "float32" min_val = float("-0.37196") @@ -4059,6 +4428,7 @@ class Program_weight_tensor_parameter_368: class Program_weight_tensor_parameter_369: name = "parameter_369" + original_name = "linear_4153.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.326236") @@ -4070,6 +4440,7 @@ class Program_weight_tensor_parameter_369: class Program_weight_tensor_parameter_370: name = "parameter_370" + original_name = "layer_norm_1433.b_0" shape = [1024] dtype = "float32" min_val = float("-0.380914") @@ -4081,6 +4452,7 @@ class Program_weight_tensor_parameter_370: class Program_weight_tensor_parameter_371: name = "parameter_371" + original_name = "layer_norm_1433.w_0" shape = [1024] dtype = "float32" min_val = float("0.190392") @@ -4092,6 +4464,7 @@ class Program_weight_tensor_parameter_371: class Program_weight_tensor_parameter_372: name = "parameter_372" + original_name = "layer_norm_1432.b_0" shape = [1024] dtype = "float32" min_val = float("-2.6418") @@ -4103,6 +4476,7 @@ class Program_weight_tensor_parameter_372: class Program_weight_tensor_parameter_373: name = "parameter_373" + original_name = "layer_norm_1432.w_0" shape = [1024] dtype = "float32" min_val = float("0.808626") @@ -4114,6 +4488,7 @@ class Program_weight_tensor_parameter_373: class Program_weight_tensor_parameter_374: name = "parameter_374" + original_name = "linear_4152.b_0" shape = [1024] dtype = "float32" min_val = float("-0.43712") @@ -4125,6 +4500,7 @@ class Program_weight_tensor_parameter_374: class Program_weight_tensor_parameter_375: name = "parameter_375" + original_name = "linear_4152.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-0.730306") @@ -4136,6 +4512,7 @@ class Program_weight_tensor_parameter_375: class Program_weight_tensor_parameter_376: name = "parameter_376" + original_name = "linear_4151.b_0" shape = [4096] dtype = "float32" min_val = float("-0.224929") @@ -4147,6 +4524,7 @@ class Program_weight_tensor_parameter_376: class Program_weight_tensor_parameter_377: name = "parameter_377" + original_name = "linear_4151.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.42452") @@ -4158,6 +4536,7 @@ class Program_weight_tensor_parameter_377: class Program_weight_tensor_parameter_378: name = "parameter_378" + original_name = "linear_4150.b_0" shape = [1024] dtype = "float32" min_val = float("-0.536243") @@ -4169,6 +4548,7 @@ class Program_weight_tensor_parameter_378: class Program_weight_tensor_parameter_379: name = "parameter_379" + original_name = "linear_4150.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-1.06589") @@ -4180,6 +4560,7 @@ class Program_weight_tensor_parameter_379: class Program_weight_tensor_parameter_380: name = "parameter_380" + original_name = "linear_4149.b_0" shape = [1024] dtype = "float32" min_val = float("-0.342756") @@ -4191,6 +4572,7 @@ class Program_weight_tensor_parameter_380: class Program_weight_tensor_parameter_381: name = "parameter_381" + original_name = "linear_4149.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.136931") @@ -4202,6 +4584,7 @@ class Program_weight_tensor_parameter_381: class Program_weight_tensor_parameter_382: name = "parameter_382" + original_name = "linear_4148.b_0" shape = [1024] dtype = "float32" min_val = float("-0.469305") @@ -4213,6 +4596,7 @@ class Program_weight_tensor_parameter_382: class Program_weight_tensor_parameter_383: name = "parameter_383" + original_name = "linear_4148.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.445105") @@ -4224,6 +4608,7 @@ class Program_weight_tensor_parameter_383: class Program_weight_tensor_parameter_384: name = "parameter_384" + original_name = "linear_4147.b_0" shape = [1024] dtype = "float32" min_val = float("-0.95629") @@ -4235,6 +4620,7 @@ class Program_weight_tensor_parameter_384: class Program_weight_tensor_parameter_385: name = "parameter_385" + original_name = "linear_4147.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.440359") @@ -4246,6 +4632,7 @@ class Program_weight_tensor_parameter_385: class Program_weight_tensor_parameter_386: name = "parameter_386" + original_name = "layer_norm_1431.b_0" shape = [1024] dtype = "float32" min_val = float("-0.575655") @@ -4257,6 +4644,7 @@ class Program_weight_tensor_parameter_386: class Program_weight_tensor_parameter_387: name = "parameter_387" + original_name = "layer_norm_1431.w_0" shape = [1024] dtype = "float32" min_val = float("0.138355") @@ -4268,6 +4656,7 @@ class Program_weight_tensor_parameter_387: class Program_weight_tensor_parameter_388: name = "parameter_388" + original_name = "embedding_258.w_0" shape = [4, 1024] dtype = "float32" min_val = float("-0.174712") @@ -4279,6 +4668,7 @@ class Program_weight_tensor_parameter_388: class Program_weight_tensor_parameter_389: name = "parameter_389" + original_name = "embedding_257.w_0" shape = [512, 1024] dtype = "float32" min_val = float("-0.308933") @@ -4290,6 +4680,7 @@ class Program_weight_tensor_parameter_389: class Program_weight_tensor_parameter_390: name = "parameter_390" + original_name = "embedding_256.w_0" shape = [30522, 1024] dtype = "float32" min_val = float("-0.744221") diff --git a/paddle_samples/PaddleNLP/utc-large/input_meta.py b/paddle_samples/PaddleNLP/utc-large/input_meta.py index 2ad3baf05..4ec838a4a 100644 --- a/paddle_samples/PaddleNLP/utc-large/input_meta.py +++ b/paddle_samples/PaddleNLP/utc-large/input_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_data_0: name = "data_0" + original_name = "input_ids" shape = [1, 26] dtype = "int64" data = [ @@ -34,6 +35,7 @@ class Program_weight_tensor_data_0: class Program_weight_tensor_data_1: name = "data_1" + original_name = "token_type_ids" shape = [1, 26] dtype = "int64" data = [ diff --git a/paddle_samples/PaddleNLP/utc-large/model.py b/paddle_samples/PaddleNLP/utc-large/model.py index 59c876c76..033ec7a46 100644 --- a/paddle_samples/PaddleNLP/utc-large/model.py +++ b/paddle_samples/PaddleNLP/utc-large/model.py @@ -458,10 +458,11 @@ def forward( # pd_op.embedding: (1x26x1024xf32) <- (1x26xi64, 512x1024xf32) embedding_1 = paddle._C_ops.embedding(subtract_0, parameter_390, -1, False) - del parameter_390 + del parameter_390, subtract_0 # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1x26x1024xf32) add_0 = paddle._C_ops.add(embedding_0, embedding_1) + del embedding_0, embedding_1 # pd_op.embedding: (1x26x1024xf32) <- (1x26xi64, 2x1024xf32) embedding_2 = paddle._C_ops.embedding(data_1, parameter_389, -1, False) @@ -469,6 +470,7 @@ def forward( # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1x26x1024xf32) add_1 = paddle._C_ops.add(add_0, embedding_2) + del add_0, embedding_2 # pd_op.full: (1xf32) <- () full_4 = paddle._C_ops.full( @@ -481,10 +483,11 @@ def forward( # pd_op.embedding: (1x26x1024xf32) <- (1x26xi64, 3x1024xf32) embedding_3 = paddle._C_ops.embedding(scale_1, parameter_388, -1, False) - del parameter_388 + del parameter_388, scale_1 # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1x26x1024xf32) add_2 = paddle._C_ops.add(add_1, embedding_3) + del add_1, embedding_3 # pd_op.layer_norm: (1x26x1024xf32, 1x26xf32, 1x26xf32) <- (1x26x1024xf32, 1024xf32, 1024xf32) layer_norm_0, layer_norm_1, layer_norm_2 = (lambda x, f: f(x))( @@ -493,233 +496,17 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_386, parameter_387 + del add_2, parameter_386, parameter_387 # pd_op.full: (1xf32) <- () full_5 = paddle._C_ops.full( [1], float("0.1"), paddle.float32, paddle.core.CPUPlace() ) - # pd_op.assign: (1xf32) <- (1xf32) - assign_0 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_1 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_2 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_3 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_4 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_5 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_6 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_7 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_8 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_9 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_10 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_11 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_12 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_13 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_14 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_15 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_16 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_17 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_18 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_19 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_20 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_21 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_22 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_23 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_24 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_25 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_26 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_27 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_28 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_29 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_30 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_31 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_32 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_33 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_34 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_35 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_36 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_37 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_38 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_39 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_40 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_41 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_42 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_43 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_44 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_45 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_46 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_47 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_48 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_49 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_50 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_51 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_52 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_53 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_54 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_55 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_56 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_57 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_58 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_59 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_60 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_61 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_62 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_63 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_64 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_65 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_66 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_67 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_68 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_69 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_70 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_71 = full_5 - # pd_op.dropout: (1x26x1024xf32, 1x26x1024xui8) <- (1x26x1024xf32, None, 1xf32) dropout_0, dropout_1 = (lambda x, f: f(x))( paddle._C_ops.dropout( - layer_norm_0, None, full_5, False, "upscale_in_train", 0, False + layer_norm_0, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -731,13 +518,14 @@ def forward( # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1024xf32) add_3 = paddle._C_ops.add(matmul_0, parameter_384) - del parameter_384 + del matmul_0, parameter_384 # pd_op.full_int_array: (4xi64) <- () full_int_array_1 = [0, 0, 16, 64] # pd_op.reshape: (1x26x16x64xf32) <- (1x26x1024xf32, 4xi64) reshape_0 = paddle._C_ops.reshape(add_3, full_int_array_1) + del add_3 # pd_op.transpose: (1x16x26x64xf32) <- (1x26x16x64xf32) transpose_0 = paddle._C_ops.transpose(reshape_0, [0, 2, 1, 3]) @@ -749,7 +537,7 @@ def forward( # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1024xf32) add_4 = paddle._C_ops.add(matmul_1, parameter_382) - del parameter_382 + del matmul_1, parameter_382 # pd_op.matmul: (1x26x1024xf32) <- (1x26x1024xf32, 1024x1024xf32) matmul_2 = paddle._C_ops.matmul(dropout_0, parameter_381, False, False) @@ -757,10 +545,11 @@ def forward( # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1024xf32) add_5 = paddle._C_ops.add(matmul_2, parameter_380) - del parameter_380 + del matmul_2, parameter_380 # pd_op.reshape: (1x26x16x64xf32) <- (1x26x1024xf32, 4xi64) reshape_1 = paddle._C_ops.reshape(add_4, full_int_array_1) + del add_4 # pd_op.transpose: (1x16x26x64xf32) <- (1x26x16x64xf32) transpose_1 = paddle._C_ops.transpose(reshape_1, [0, 2, 1, 3]) @@ -768,6 +557,7 @@ def forward( # pd_op.reshape: (1x26x16x64xf32) <- (1x26x1024xf32, 4xi64) reshape_2 = paddle._C_ops.reshape(add_5, full_int_array_1) + del add_5 # pd_op.transpose: (1x16x26x64xf32) <- (1x26x16x64xf32) transpose_2 = paddle._C_ops.transpose(reshape_2, [0, 2, 1, 3]) @@ -778,84 +568,17 @@ def forward( [1], float("0.125"), paddle.float32, paddle.core.CPUPlace() ) - # pd_op.assign: (1xf32) <- (1xf32) - assign_72 = full_6 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_73 = full_6 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_74 = full_6 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_75 = full_6 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_76 = full_6 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_77 = full_6 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_78 = full_6 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_79 = full_6 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_80 = full_6 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_81 = full_6 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_82 = full_6 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_83 = full_6 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_84 = full_6 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_85 = full_6 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_86 = full_6 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_87 = full_6 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_88 = full_6 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_89 = full_6 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_90 = full_6 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_91 = full_6 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_92 = full_6 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_93 = full_6 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_94 = full_6 - # pd_op.scale: (1x16x26x64xf32) <- (1x16x26x64xf32, 1xf32) scale_2 = paddle._C_ops.scale(transpose_0, full_6, float("0"), True) del transpose_0 # pd_op.matmul: (1x16x26x26xf32) <- (1x16x26x64xf32, 1x16x26x64xf32) matmul_3 = paddle._C_ops.matmul(scale_2, transpose_1, False, True) + del scale_2, transpose_1 # pd_op.add: (1x16x26x26xf32) <- (1x16x26x26xf32, 1x1x1x26xf32) add_6 = paddle._C_ops.add(matmul_3, unsqueeze_0) + del matmul_3 # pd_op.softmax: (1x16x26x26xf32) <- (1x16x26x26xf32) softmax_0 = paddle._C_ops.softmax(add_6, -1) @@ -864,13 +587,15 @@ def forward( # pd_op.dropout: (1x16x26x26xf32, 1x16x26x26xui8) <- (1x16x26x26xf32, None, 1xf32) dropout_2, dropout_3 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_0, None, full_5, False, "upscale_in_train", 0, False + softmax_0, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_0 # pd_op.matmul: (1x16x26x64xf32) <- (1x16x26x26xf32, 1x16x26x64xf32) matmul_4 = paddle._C_ops.matmul(dropout_2, transpose_2, False, False) + del dropout_2, transpose_2 # pd_op.transpose: (1x26x16x64xf32) <- (1x16x26x64xf32) transpose_3 = paddle._C_ops.transpose(matmul_4, [0, 2, 1, 3]) @@ -881,19 +606,20 @@ def forward( # pd_op.reshape: (1x26x1024xf32) <- (1x26x16x64xf32, 3xi64) reshape_3 = paddle._C_ops.reshape(transpose_3, full_int_array_2) + del transpose_3 # pd_op.matmul: (1x26x1024xf32) <- (1x26x1024xf32, 1024x1024xf32) matmul_5 = paddle._C_ops.matmul(reshape_3, parameter_379, False, False) - del parameter_379 + del parameter_379, reshape_3 # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1024xf32) add_7 = paddle._C_ops.add(matmul_5, parameter_378) - del parameter_378 + del matmul_5, parameter_378 # pd_op.dropout: (1x26x1024xf32, 1x26x1024xui8) <- (1x26x1024xf32, None, 1xf32) dropout_4, dropout_5 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_7, None, full_5, False, "upscale_in_train", 0, False + add_7, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -901,6 +627,7 @@ def forward( # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1x26x1024xf32) add_8 = paddle._C_ops.add(dropout_0, dropout_4) + del dropout_0, dropout_4 # pd_op.layer_norm: (1x26x1024xf32, 1x26xf32, 1x26xf32) <- (1x26x1024xf32, 1024xf32, 1024xf32) layer_norm_3, layer_norm_4, layer_norm_5 = (lambda x, f: f(x))( @@ -909,7 +636,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_372, parameter_373 + del add_8, parameter_372, parameter_373 # pd_op.matmul: (1x26x3072xf32) <- (1x26x1024xf32, 1024x3072xf32) matmul_6 = paddle._C_ops.matmul(layer_norm_3, parameter_377, False, False) @@ -917,7 +644,7 @@ def forward( # pd_op.add: (1x26x3072xf32) <- (1x26x3072xf32, 3072xf32) add_9 = paddle._C_ops.add(matmul_6, parameter_376) - del parameter_376 + del matmul_6, parameter_376 # pd_op.relu: (1x26x3072xf32) <- (1x26x3072xf32) relu_0 = paddle._C_ops.relu(add_9) @@ -925,16 +652,16 @@ def forward( # pd_op.matmul: (1x26x1024xf32) <- (1x26x3072xf32, 3072x1024xf32) matmul_7 = paddle._C_ops.matmul(relu_0, parameter_375, False, False) - del parameter_375 + del parameter_375, relu_0 # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1024xf32) add_10 = paddle._C_ops.add(matmul_7, parameter_374) - del parameter_374 + del matmul_7, parameter_374 # pd_op.dropout: (1x26x1024xf32, 1x26x1024xui8) <- (1x26x1024xf32, None, 1xf32) dropout_6, dropout_7 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_10, None, full_5, False, "upscale_in_train", 0, False + add_10, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -942,6 +669,7 @@ def forward( # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1x26x1024xf32) add_11 = paddle._C_ops.add(layer_norm_3, dropout_6) + del dropout_6, layer_norm_3 # pd_op.layer_norm: (1x26x1024xf32, 1x26xf32, 1x26xf32) <- (1x26x1024xf32, 1024xf32, 1024xf32) layer_norm_6, layer_norm_7, layer_norm_8 = (lambda x, f: f(x))( @@ -950,7 +678,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_370, parameter_371 + del add_11, parameter_370, parameter_371 # pd_op.matmul: (1x26x1024xf32) <- (1x26x1024xf32, 1024x1024xf32) matmul_8 = paddle._C_ops.matmul(layer_norm_6, parameter_369, False, False) @@ -958,10 +686,11 @@ def forward( # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1024xf32) add_12 = paddle._C_ops.add(matmul_8, parameter_368) - del parameter_368 + del matmul_8, parameter_368 # pd_op.reshape: (1x26x16x64xf32) <- (1x26x1024xf32, 4xi64) reshape_4 = paddle._C_ops.reshape(add_12, full_int_array_1) + del add_12 # pd_op.transpose: (1x16x26x64xf32) <- (1x26x16x64xf32) transpose_4 = paddle._C_ops.transpose(reshape_4, [0, 2, 1, 3]) @@ -973,7 +702,7 @@ def forward( # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1024xf32) add_13 = paddle._C_ops.add(matmul_9, parameter_366) - del parameter_366 + del matmul_9, parameter_366 # pd_op.matmul: (1x26x1024xf32) <- (1x26x1024xf32, 1024x1024xf32) matmul_10 = paddle._C_ops.matmul(layer_norm_6, parameter_365, False, False) @@ -981,10 +710,11 @@ def forward( # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1024xf32) add_14 = paddle._C_ops.add(matmul_10, parameter_364) - del parameter_364 + del matmul_10, parameter_364 # pd_op.reshape: (1x26x16x64xf32) <- (1x26x1024xf32, 4xi64) reshape_5 = paddle._C_ops.reshape(add_13, full_int_array_1) + del add_13 # pd_op.transpose: (1x16x26x64xf32) <- (1x26x16x64xf32) transpose_5 = paddle._C_ops.transpose(reshape_5, [0, 2, 1, 3]) @@ -992,6 +722,7 @@ def forward( # pd_op.reshape: (1x26x16x64xf32) <- (1x26x1024xf32, 4xi64) reshape_6 = paddle._C_ops.reshape(add_14, full_int_array_1) + del add_14 # pd_op.transpose: (1x16x26x64xf32) <- (1x26x16x64xf32) transpose_6 = paddle._C_ops.transpose(reshape_6, [0, 2, 1, 3]) @@ -1003,9 +734,11 @@ def forward( # pd_op.matmul: (1x16x26x26xf32) <- (1x16x26x64xf32, 1x16x26x64xf32) matmul_11 = paddle._C_ops.matmul(scale_3, transpose_5, False, True) + del scale_3, transpose_5 # pd_op.add: (1x16x26x26xf32) <- (1x16x26x26xf32, 1x1x1x26xf32) add_15 = paddle._C_ops.add(matmul_11, unsqueeze_0) + del matmul_11 # pd_op.softmax: (1x16x26x26xf32) <- (1x16x26x26xf32) softmax_1 = paddle._C_ops.softmax(add_15, -1) @@ -1014,13 +747,15 @@ def forward( # pd_op.dropout: (1x16x26x26xf32, 1x16x26x26xui8) <- (1x16x26x26xf32, None, 1xf32) dropout_8, dropout_9 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_1, None, full_5, False, "upscale_in_train", 0, False + softmax_1, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_1 # pd_op.matmul: (1x16x26x64xf32) <- (1x16x26x26xf32, 1x16x26x64xf32) matmul_12 = paddle._C_ops.matmul(dropout_8, transpose_6, False, False) + del dropout_8, transpose_6 # pd_op.transpose: (1x26x16x64xf32) <- (1x16x26x64xf32) transpose_7 = paddle._C_ops.transpose(matmul_12, [0, 2, 1, 3]) @@ -1028,19 +763,20 @@ def forward( # pd_op.reshape: (1x26x1024xf32) <- (1x26x16x64xf32, 3xi64) reshape_7 = paddle._C_ops.reshape(transpose_7, full_int_array_2) + del transpose_7 # pd_op.matmul: (1x26x1024xf32) <- (1x26x1024xf32, 1024x1024xf32) matmul_13 = paddle._C_ops.matmul(reshape_7, parameter_363, False, False) - del parameter_363 + del parameter_363, reshape_7 # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1024xf32) add_16 = paddle._C_ops.add(matmul_13, parameter_362) - del parameter_362 + del matmul_13, parameter_362 # pd_op.dropout: (1x26x1024xf32, 1x26x1024xui8) <- (1x26x1024xf32, None, 1xf32) dropout_10, dropout_11 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_16, None, full_5, False, "upscale_in_train", 0, False + add_16, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1048,6 +784,7 @@ def forward( # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1x26x1024xf32) add_17 = paddle._C_ops.add(layer_norm_6, dropout_10) + del dropout_10, layer_norm_6 # pd_op.layer_norm: (1x26x1024xf32, 1x26xf32, 1x26xf32) <- (1x26x1024xf32, 1024xf32, 1024xf32) layer_norm_9, layer_norm_10, layer_norm_11 = (lambda x, f: f(x))( @@ -1056,7 +793,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_356, parameter_357 + del add_17, parameter_356, parameter_357 # pd_op.matmul: (1x26x3072xf32) <- (1x26x1024xf32, 1024x3072xf32) matmul_14 = paddle._C_ops.matmul(layer_norm_9, parameter_361, False, False) @@ -1064,7 +801,7 @@ def forward( # pd_op.add: (1x26x3072xf32) <- (1x26x3072xf32, 3072xf32) add_18 = paddle._C_ops.add(matmul_14, parameter_360) - del parameter_360 + del matmul_14, parameter_360 # pd_op.relu: (1x26x3072xf32) <- (1x26x3072xf32) relu_1 = paddle._C_ops.relu(add_18) @@ -1072,16 +809,16 @@ def forward( # pd_op.matmul: (1x26x1024xf32) <- (1x26x3072xf32, 3072x1024xf32) matmul_15 = paddle._C_ops.matmul(relu_1, parameter_359, False, False) - del parameter_359 + del parameter_359, relu_1 # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1024xf32) add_19 = paddle._C_ops.add(matmul_15, parameter_358) - del parameter_358 + del matmul_15, parameter_358 # pd_op.dropout: (1x26x1024xf32, 1x26x1024xui8) <- (1x26x1024xf32, None, 1xf32) dropout_12, dropout_13 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_19, None, full_5, False, "upscale_in_train", 0, False + add_19, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1089,6 +826,7 @@ def forward( # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1x26x1024xf32) add_20 = paddle._C_ops.add(layer_norm_9, dropout_12) + del dropout_12, layer_norm_9 # pd_op.layer_norm: (1x26x1024xf32, 1x26xf32, 1x26xf32) <- (1x26x1024xf32, 1024xf32, 1024xf32) layer_norm_12, layer_norm_13, layer_norm_14 = (lambda x, f: f(x))( @@ -1097,7 +835,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_354, parameter_355 + del add_20, parameter_354, parameter_355 # pd_op.matmul: (1x26x1024xf32) <- (1x26x1024xf32, 1024x1024xf32) matmul_16 = paddle._C_ops.matmul(layer_norm_12, parameter_353, False, False) @@ -1105,10 +843,11 @@ def forward( # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1024xf32) add_21 = paddle._C_ops.add(matmul_16, parameter_352) - del parameter_352 + del matmul_16, parameter_352 # pd_op.reshape: (1x26x16x64xf32) <- (1x26x1024xf32, 4xi64) reshape_8 = paddle._C_ops.reshape(add_21, full_int_array_1) + del add_21 # pd_op.transpose: (1x16x26x64xf32) <- (1x26x16x64xf32) transpose_8 = paddle._C_ops.transpose(reshape_8, [0, 2, 1, 3]) @@ -1120,7 +859,7 @@ def forward( # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1024xf32) add_22 = paddle._C_ops.add(matmul_17, parameter_350) - del parameter_350 + del matmul_17, parameter_350 # pd_op.matmul: (1x26x1024xf32) <- (1x26x1024xf32, 1024x1024xf32) matmul_18 = paddle._C_ops.matmul(layer_norm_12, parameter_349, False, False) @@ -1128,10 +867,11 @@ def forward( # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1024xf32) add_23 = paddle._C_ops.add(matmul_18, parameter_348) - del parameter_348 + del matmul_18, parameter_348 # pd_op.reshape: (1x26x16x64xf32) <- (1x26x1024xf32, 4xi64) reshape_9 = paddle._C_ops.reshape(add_22, full_int_array_1) + del add_22 # pd_op.transpose: (1x16x26x64xf32) <- (1x26x16x64xf32) transpose_9 = paddle._C_ops.transpose(reshape_9, [0, 2, 1, 3]) @@ -1139,6 +879,7 @@ def forward( # pd_op.reshape: (1x26x16x64xf32) <- (1x26x1024xf32, 4xi64) reshape_10 = paddle._C_ops.reshape(add_23, full_int_array_1) + del add_23 # pd_op.transpose: (1x16x26x64xf32) <- (1x26x16x64xf32) transpose_10 = paddle._C_ops.transpose(reshape_10, [0, 2, 1, 3]) @@ -1150,9 +891,11 @@ def forward( # pd_op.matmul: (1x16x26x26xf32) <- (1x16x26x64xf32, 1x16x26x64xf32) matmul_19 = paddle._C_ops.matmul(scale_4, transpose_9, False, True) + del scale_4, transpose_9 # pd_op.add: (1x16x26x26xf32) <- (1x16x26x26xf32, 1x1x1x26xf32) add_24 = paddle._C_ops.add(matmul_19, unsqueeze_0) + del matmul_19 # pd_op.softmax: (1x16x26x26xf32) <- (1x16x26x26xf32) softmax_2 = paddle._C_ops.softmax(add_24, -1) @@ -1161,13 +904,15 @@ def forward( # pd_op.dropout: (1x16x26x26xf32, 1x16x26x26xui8) <- (1x16x26x26xf32, None, 1xf32) dropout_14, dropout_15 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_2, None, full_5, False, "upscale_in_train", 0, False + softmax_2, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_2 # pd_op.matmul: (1x16x26x64xf32) <- (1x16x26x26xf32, 1x16x26x64xf32) matmul_20 = paddle._C_ops.matmul(dropout_14, transpose_10, False, False) + del dropout_14, transpose_10 # pd_op.transpose: (1x26x16x64xf32) <- (1x16x26x64xf32) transpose_11 = paddle._C_ops.transpose(matmul_20, [0, 2, 1, 3]) @@ -1175,19 +920,20 @@ def forward( # pd_op.reshape: (1x26x1024xf32) <- (1x26x16x64xf32, 3xi64) reshape_11 = paddle._C_ops.reshape(transpose_11, full_int_array_2) + del transpose_11 # pd_op.matmul: (1x26x1024xf32) <- (1x26x1024xf32, 1024x1024xf32) matmul_21 = paddle._C_ops.matmul(reshape_11, parameter_347, False, False) - del parameter_347 + del parameter_347, reshape_11 # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1024xf32) add_25 = paddle._C_ops.add(matmul_21, parameter_346) - del parameter_346 + del matmul_21, parameter_346 # pd_op.dropout: (1x26x1024xf32, 1x26x1024xui8) <- (1x26x1024xf32, None, 1xf32) dropout_16, dropout_17 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_25, None, full_5, False, "upscale_in_train", 0, False + add_25, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1195,6 +941,7 @@ def forward( # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1x26x1024xf32) add_26 = paddle._C_ops.add(layer_norm_12, dropout_16) + del dropout_16, layer_norm_12 # pd_op.layer_norm: (1x26x1024xf32, 1x26xf32, 1x26xf32) <- (1x26x1024xf32, 1024xf32, 1024xf32) layer_norm_15, layer_norm_16, layer_norm_17 = (lambda x, f: f(x))( @@ -1203,7 +950,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_340, parameter_341 + del add_26, parameter_340, parameter_341 # pd_op.matmul: (1x26x3072xf32) <- (1x26x1024xf32, 1024x3072xf32) matmul_22 = paddle._C_ops.matmul(layer_norm_15, parameter_345, False, False) @@ -1211,7 +958,7 @@ def forward( # pd_op.add: (1x26x3072xf32) <- (1x26x3072xf32, 3072xf32) add_27 = paddle._C_ops.add(matmul_22, parameter_344) - del parameter_344 + del matmul_22, parameter_344 # pd_op.relu: (1x26x3072xf32) <- (1x26x3072xf32) relu_2 = paddle._C_ops.relu(add_27) @@ -1219,16 +966,16 @@ def forward( # pd_op.matmul: (1x26x1024xf32) <- (1x26x3072xf32, 3072x1024xf32) matmul_23 = paddle._C_ops.matmul(relu_2, parameter_343, False, False) - del parameter_343 + del parameter_343, relu_2 # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1024xf32) add_28 = paddle._C_ops.add(matmul_23, parameter_342) - del parameter_342 + del matmul_23, parameter_342 # pd_op.dropout: (1x26x1024xf32, 1x26x1024xui8) <- (1x26x1024xf32, None, 1xf32) dropout_18, dropout_19 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_28, None, full_5, False, "upscale_in_train", 0, False + add_28, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1236,6 +983,7 @@ def forward( # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1x26x1024xf32) add_29 = paddle._C_ops.add(layer_norm_15, dropout_18) + del dropout_18, layer_norm_15 # pd_op.layer_norm: (1x26x1024xf32, 1x26xf32, 1x26xf32) <- (1x26x1024xf32, 1024xf32, 1024xf32) layer_norm_18, layer_norm_19, layer_norm_20 = (lambda x, f: f(x))( @@ -1244,7 +992,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_338, parameter_339 + del add_29, parameter_338, parameter_339 # pd_op.matmul: (1x26x1024xf32) <- (1x26x1024xf32, 1024x1024xf32) matmul_24 = paddle._C_ops.matmul(layer_norm_18, parameter_337, False, False) @@ -1252,10 +1000,11 @@ def forward( # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1024xf32) add_30 = paddle._C_ops.add(matmul_24, parameter_336) - del parameter_336 + del matmul_24, parameter_336 # pd_op.reshape: (1x26x16x64xf32) <- (1x26x1024xf32, 4xi64) reshape_12 = paddle._C_ops.reshape(add_30, full_int_array_1) + del add_30 # pd_op.transpose: (1x16x26x64xf32) <- (1x26x16x64xf32) transpose_12 = paddle._C_ops.transpose(reshape_12, [0, 2, 1, 3]) @@ -1267,7 +1016,7 @@ def forward( # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1024xf32) add_31 = paddle._C_ops.add(matmul_25, parameter_334) - del parameter_334 + del matmul_25, parameter_334 # pd_op.matmul: (1x26x1024xf32) <- (1x26x1024xf32, 1024x1024xf32) matmul_26 = paddle._C_ops.matmul(layer_norm_18, parameter_333, False, False) @@ -1275,10 +1024,11 @@ def forward( # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1024xf32) add_32 = paddle._C_ops.add(matmul_26, parameter_332) - del parameter_332 + del matmul_26, parameter_332 # pd_op.reshape: (1x26x16x64xf32) <- (1x26x1024xf32, 4xi64) reshape_13 = paddle._C_ops.reshape(add_31, full_int_array_1) + del add_31 # pd_op.transpose: (1x16x26x64xf32) <- (1x26x16x64xf32) transpose_13 = paddle._C_ops.transpose(reshape_13, [0, 2, 1, 3]) @@ -1286,6 +1036,7 @@ def forward( # pd_op.reshape: (1x26x16x64xf32) <- (1x26x1024xf32, 4xi64) reshape_14 = paddle._C_ops.reshape(add_32, full_int_array_1) + del add_32 # pd_op.transpose: (1x16x26x64xf32) <- (1x26x16x64xf32) transpose_14 = paddle._C_ops.transpose(reshape_14, [0, 2, 1, 3]) @@ -1297,9 +1048,11 @@ def forward( # pd_op.matmul: (1x16x26x26xf32) <- (1x16x26x64xf32, 1x16x26x64xf32) matmul_27 = paddle._C_ops.matmul(scale_5, transpose_13, False, True) + del scale_5, transpose_13 # pd_op.add: (1x16x26x26xf32) <- (1x16x26x26xf32, 1x1x1x26xf32) add_33 = paddle._C_ops.add(matmul_27, unsqueeze_0) + del matmul_27 # pd_op.softmax: (1x16x26x26xf32) <- (1x16x26x26xf32) softmax_3 = paddle._C_ops.softmax(add_33, -1) @@ -1308,13 +1061,15 @@ def forward( # pd_op.dropout: (1x16x26x26xf32, 1x16x26x26xui8) <- (1x16x26x26xf32, None, 1xf32) dropout_20, dropout_21 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_3, None, full_5, False, "upscale_in_train", 0, False + softmax_3, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_3 # pd_op.matmul: (1x16x26x64xf32) <- (1x16x26x26xf32, 1x16x26x64xf32) matmul_28 = paddle._C_ops.matmul(dropout_20, transpose_14, False, False) + del dropout_20, transpose_14 # pd_op.transpose: (1x26x16x64xf32) <- (1x16x26x64xf32) transpose_15 = paddle._C_ops.transpose(matmul_28, [0, 2, 1, 3]) @@ -1322,19 +1077,20 @@ def forward( # pd_op.reshape: (1x26x1024xf32) <- (1x26x16x64xf32, 3xi64) reshape_15 = paddle._C_ops.reshape(transpose_15, full_int_array_2) + del transpose_15 # pd_op.matmul: (1x26x1024xf32) <- (1x26x1024xf32, 1024x1024xf32) matmul_29 = paddle._C_ops.matmul(reshape_15, parameter_331, False, False) - del parameter_331 + del parameter_331, reshape_15 # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1024xf32) add_34 = paddle._C_ops.add(matmul_29, parameter_330) - del parameter_330 + del matmul_29, parameter_330 # pd_op.dropout: (1x26x1024xf32, 1x26x1024xui8) <- (1x26x1024xf32, None, 1xf32) dropout_22, dropout_23 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_34, None, full_5, False, "upscale_in_train", 0, False + add_34, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1342,6 +1098,7 @@ def forward( # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1x26x1024xf32) add_35 = paddle._C_ops.add(layer_norm_18, dropout_22) + del dropout_22, layer_norm_18 # pd_op.layer_norm: (1x26x1024xf32, 1x26xf32, 1x26xf32) <- (1x26x1024xf32, 1024xf32, 1024xf32) layer_norm_21, layer_norm_22, layer_norm_23 = (lambda x, f: f(x))( @@ -1350,7 +1107,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_324, parameter_325 + del add_35, parameter_324, parameter_325 # pd_op.matmul: (1x26x3072xf32) <- (1x26x1024xf32, 1024x3072xf32) matmul_30 = paddle._C_ops.matmul(layer_norm_21, parameter_329, False, False) @@ -1358,7 +1115,7 @@ def forward( # pd_op.add: (1x26x3072xf32) <- (1x26x3072xf32, 3072xf32) add_36 = paddle._C_ops.add(matmul_30, parameter_328) - del parameter_328 + del matmul_30, parameter_328 # pd_op.relu: (1x26x3072xf32) <- (1x26x3072xf32) relu_3 = paddle._C_ops.relu(add_36) @@ -1366,16 +1123,16 @@ def forward( # pd_op.matmul: (1x26x1024xf32) <- (1x26x3072xf32, 3072x1024xf32) matmul_31 = paddle._C_ops.matmul(relu_3, parameter_327, False, False) - del parameter_327 + del parameter_327, relu_3 # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1024xf32) add_37 = paddle._C_ops.add(matmul_31, parameter_326) - del parameter_326 + del matmul_31, parameter_326 # pd_op.dropout: (1x26x1024xf32, 1x26x1024xui8) <- (1x26x1024xf32, None, 1xf32) dropout_24, dropout_25 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_37, None, full_5, False, "upscale_in_train", 0, False + add_37, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1383,6 +1140,7 @@ def forward( # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1x26x1024xf32) add_38 = paddle._C_ops.add(layer_norm_21, dropout_24) + del dropout_24, layer_norm_21 # pd_op.layer_norm: (1x26x1024xf32, 1x26xf32, 1x26xf32) <- (1x26x1024xf32, 1024xf32, 1024xf32) layer_norm_24, layer_norm_25, layer_norm_26 = (lambda x, f: f(x))( @@ -1391,7 +1149,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_322, parameter_323 + del add_38, parameter_322, parameter_323 # pd_op.matmul: (1x26x1024xf32) <- (1x26x1024xf32, 1024x1024xf32) matmul_32 = paddle._C_ops.matmul(layer_norm_24, parameter_321, False, False) @@ -1399,10 +1157,11 @@ def forward( # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1024xf32) add_39 = paddle._C_ops.add(matmul_32, parameter_320) - del parameter_320 + del matmul_32, parameter_320 # pd_op.reshape: (1x26x16x64xf32) <- (1x26x1024xf32, 4xi64) reshape_16 = paddle._C_ops.reshape(add_39, full_int_array_1) + del add_39 # pd_op.transpose: (1x16x26x64xf32) <- (1x26x16x64xf32) transpose_16 = paddle._C_ops.transpose(reshape_16, [0, 2, 1, 3]) @@ -1414,7 +1173,7 @@ def forward( # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1024xf32) add_40 = paddle._C_ops.add(matmul_33, parameter_318) - del parameter_318 + del matmul_33, parameter_318 # pd_op.matmul: (1x26x1024xf32) <- (1x26x1024xf32, 1024x1024xf32) matmul_34 = paddle._C_ops.matmul(layer_norm_24, parameter_317, False, False) @@ -1422,10 +1181,11 @@ def forward( # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1024xf32) add_41 = paddle._C_ops.add(matmul_34, parameter_316) - del parameter_316 + del matmul_34, parameter_316 # pd_op.reshape: (1x26x16x64xf32) <- (1x26x1024xf32, 4xi64) reshape_17 = paddle._C_ops.reshape(add_40, full_int_array_1) + del add_40 # pd_op.transpose: (1x16x26x64xf32) <- (1x26x16x64xf32) transpose_17 = paddle._C_ops.transpose(reshape_17, [0, 2, 1, 3]) @@ -1433,6 +1193,7 @@ def forward( # pd_op.reshape: (1x26x16x64xf32) <- (1x26x1024xf32, 4xi64) reshape_18 = paddle._C_ops.reshape(add_41, full_int_array_1) + del add_41 # pd_op.transpose: (1x16x26x64xf32) <- (1x26x16x64xf32) transpose_18 = paddle._C_ops.transpose(reshape_18, [0, 2, 1, 3]) @@ -1444,9 +1205,11 @@ def forward( # pd_op.matmul: (1x16x26x26xf32) <- (1x16x26x64xf32, 1x16x26x64xf32) matmul_35 = paddle._C_ops.matmul(scale_6, transpose_17, False, True) + del scale_6, transpose_17 # pd_op.add: (1x16x26x26xf32) <- (1x16x26x26xf32, 1x1x1x26xf32) add_42 = paddle._C_ops.add(matmul_35, unsqueeze_0) + del matmul_35 # pd_op.softmax: (1x16x26x26xf32) <- (1x16x26x26xf32) softmax_4 = paddle._C_ops.softmax(add_42, -1) @@ -1455,13 +1218,15 @@ def forward( # pd_op.dropout: (1x16x26x26xf32, 1x16x26x26xui8) <- (1x16x26x26xf32, None, 1xf32) dropout_26, dropout_27 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_4, None, full_5, False, "upscale_in_train", 0, False + softmax_4, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_4 # pd_op.matmul: (1x16x26x64xf32) <- (1x16x26x26xf32, 1x16x26x64xf32) matmul_36 = paddle._C_ops.matmul(dropout_26, transpose_18, False, False) + del dropout_26, transpose_18 # pd_op.transpose: (1x26x16x64xf32) <- (1x16x26x64xf32) transpose_19 = paddle._C_ops.transpose(matmul_36, [0, 2, 1, 3]) @@ -1469,19 +1234,20 @@ def forward( # pd_op.reshape: (1x26x1024xf32) <- (1x26x16x64xf32, 3xi64) reshape_19 = paddle._C_ops.reshape(transpose_19, full_int_array_2) + del transpose_19 # pd_op.matmul: (1x26x1024xf32) <- (1x26x1024xf32, 1024x1024xf32) matmul_37 = paddle._C_ops.matmul(reshape_19, parameter_315, False, False) - del parameter_315 + del parameter_315, reshape_19 # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1024xf32) add_43 = paddle._C_ops.add(matmul_37, parameter_314) - del parameter_314 + del matmul_37, parameter_314 # pd_op.dropout: (1x26x1024xf32, 1x26x1024xui8) <- (1x26x1024xf32, None, 1xf32) dropout_28, dropout_29 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_43, None, full_5, False, "upscale_in_train", 0, False + add_43, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1489,6 +1255,7 @@ def forward( # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1x26x1024xf32) add_44 = paddle._C_ops.add(layer_norm_24, dropout_28) + del dropout_28, layer_norm_24 # pd_op.layer_norm: (1x26x1024xf32, 1x26xf32, 1x26xf32) <- (1x26x1024xf32, 1024xf32, 1024xf32) layer_norm_27, layer_norm_28, layer_norm_29 = (lambda x, f: f(x))( @@ -1497,7 +1264,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_308, parameter_309 + del add_44, parameter_308, parameter_309 # pd_op.matmul: (1x26x3072xf32) <- (1x26x1024xf32, 1024x3072xf32) matmul_38 = paddle._C_ops.matmul(layer_norm_27, parameter_313, False, False) @@ -1505,7 +1272,7 @@ def forward( # pd_op.add: (1x26x3072xf32) <- (1x26x3072xf32, 3072xf32) add_45 = paddle._C_ops.add(matmul_38, parameter_312) - del parameter_312 + del matmul_38, parameter_312 # pd_op.relu: (1x26x3072xf32) <- (1x26x3072xf32) relu_4 = paddle._C_ops.relu(add_45) @@ -1513,16 +1280,16 @@ def forward( # pd_op.matmul: (1x26x1024xf32) <- (1x26x3072xf32, 3072x1024xf32) matmul_39 = paddle._C_ops.matmul(relu_4, parameter_311, False, False) - del parameter_311 + del parameter_311, relu_4 # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1024xf32) add_46 = paddle._C_ops.add(matmul_39, parameter_310) - del parameter_310 + del matmul_39, parameter_310 # pd_op.dropout: (1x26x1024xf32, 1x26x1024xui8) <- (1x26x1024xf32, None, 1xf32) dropout_30, dropout_31 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_46, None, full_5, False, "upscale_in_train", 0, False + add_46, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1530,6 +1297,7 @@ def forward( # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1x26x1024xf32) add_47 = paddle._C_ops.add(layer_norm_27, dropout_30) + del dropout_30, layer_norm_27 # pd_op.layer_norm: (1x26x1024xf32, 1x26xf32, 1x26xf32) <- (1x26x1024xf32, 1024xf32, 1024xf32) layer_norm_30, layer_norm_31, layer_norm_32 = (lambda x, f: f(x))( @@ -1538,7 +1306,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_306, parameter_307 + del add_47, parameter_306, parameter_307 # pd_op.matmul: (1x26x1024xf32) <- (1x26x1024xf32, 1024x1024xf32) matmul_40 = paddle._C_ops.matmul(layer_norm_30, parameter_305, False, False) @@ -1546,10 +1314,11 @@ def forward( # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1024xf32) add_48 = paddle._C_ops.add(matmul_40, parameter_304) - del parameter_304 + del matmul_40, parameter_304 # pd_op.reshape: (1x26x16x64xf32) <- (1x26x1024xf32, 4xi64) reshape_20 = paddle._C_ops.reshape(add_48, full_int_array_1) + del add_48 # pd_op.transpose: (1x16x26x64xf32) <- (1x26x16x64xf32) transpose_20 = paddle._C_ops.transpose(reshape_20, [0, 2, 1, 3]) @@ -1561,7 +1330,7 @@ def forward( # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1024xf32) add_49 = paddle._C_ops.add(matmul_41, parameter_302) - del parameter_302 + del matmul_41, parameter_302 # pd_op.matmul: (1x26x1024xf32) <- (1x26x1024xf32, 1024x1024xf32) matmul_42 = paddle._C_ops.matmul(layer_norm_30, parameter_301, False, False) @@ -1569,10 +1338,11 @@ def forward( # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1024xf32) add_50 = paddle._C_ops.add(matmul_42, parameter_300) - del parameter_300 + del matmul_42, parameter_300 # pd_op.reshape: (1x26x16x64xf32) <- (1x26x1024xf32, 4xi64) reshape_21 = paddle._C_ops.reshape(add_49, full_int_array_1) + del add_49 # pd_op.transpose: (1x16x26x64xf32) <- (1x26x16x64xf32) transpose_21 = paddle._C_ops.transpose(reshape_21, [0, 2, 1, 3]) @@ -1580,6 +1350,7 @@ def forward( # pd_op.reshape: (1x26x16x64xf32) <- (1x26x1024xf32, 4xi64) reshape_22 = paddle._C_ops.reshape(add_50, full_int_array_1) + del add_50 # pd_op.transpose: (1x16x26x64xf32) <- (1x26x16x64xf32) transpose_22 = paddle._C_ops.transpose(reshape_22, [0, 2, 1, 3]) @@ -1591,9 +1362,11 @@ def forward( # pd_op.matmul: (1x16x26x26xf32) <- (1x16x26x64xf32, 1x16x26x64xf32) matmul_43 = paddle._C_ops.matmul(scale_7, transpose_21, False, True) + del scale_7, transpose_21 # pd_op.add: (1x16x26x26xf32) <- (1x16x26x26xf32, 1x1x1x26xf32) add_51 = paddle._C_ops.add(matmul_43, unsqueeze_0) + del matmul_43 # pd_op.softmax: (1x16x26x26xf32) <- (1x16x26x26xf32) softmax_5 = paddle._C_ops.softmax(add_51, -1) @@ -1602,13 +1375,15 @@ def forward( # pd_op.dropout: (1x16x26x26xf32, 1x16x26x26xui8) <- (1x16x26x26xf32, None, 1xf32) dropout_32, dropout_33 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_5, None, full_5, False, "upscale_in_train", 0, False + softmax_5, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_5 # pd_op.matmul: (1x16x26x64xf32) <- (1x16x26x26xf32, 1x16x26x64xf32) matmul_44 = paddle._C_ops.matmul(dropout_32, transpose_22, False, False) + del dropout_32, transpose_22 # pd_op.transpose: (1x26x16x64xf32) <- (1x16x26x64xf32) transpose_23 = paddle._C_ops.transpose(matmul_44, [0, 2, 1, 3]) @@ -1616,19 +1391,20 @@ def forward( # pd_op.reshape: (1x26x1024xf32) <- (1x26x16x64xf32, 3xi64) reshape_23 = paddle._C_ops.reshape(transpose_23, full_int_array_2) + del transpose_23 # pd_op.matmul: (1x26x1024xf32) <- (1x26x1024xf32, 1024x1024xf32) matmul_45 = paddle._C_ops.matmul(reshape_23, parameter_299, False, False) - del parameter_299 + del parameter_299, reshape_23 # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1024xf32) add_52 = paddle._C_ops.add(matmul_45, parameter_298) - del parameter_298 + del matmul_45, parameter_298 # pd_op.dropout: (1x26x1024xf32, 1x26x1024xui8) <- (1x26x1024xf32, None, 1xf32) dropout_34, dropout_35 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_52, None, full_5, False, "upscale_in_train", 0, False + add_52, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1636,6 +1412,7 @@ def forward( # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1x26x1024xf32) add_53 = paddle._C_ops.add(layer_norm_30, dropout_34) + del dropout_34, layer_norm_30 # pd_op.layer_norm: (1x26x1024xf32, 1x26xf32, 1x26xf32) <- (1x26x1024xf32, 1024xf32, 1024xf32) layer_norm_33, layer_norm_34, layer_norm_35 = (lambda x, f: f(x))( @@ -1644,7 +1421,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_292, parameter_293 + del add_53, parameter_292, parameter_293 # pd_op.matmul: (1x26x3072xf32) <- (1x26x1024xf32, 1024x3072xf32) matmul_46 = paddle._C_ops.matmul(layer_norm_33, parameter_297, False, False) @@ -1652,7 +1429,7 @@ def forward( # pd_op.add: (1x26x3072xf32) <- (1x26x3072xf32, 3072xf32) add_54 = paddle._C_ops.add(matmul_46, parameter_296) - del parameter_296 + del matmul_46, parameter_296 # pd_op.relu: (1x26x3072xf32) <- (1x26x3072xf32) relu_5 = paddle._C_ops.relu(add_54) @@ -1660,16 +1437,16 @@ def forward( # pd_op.matmul: (1x26x1024xf32) <- (1x26x3072xf32, 3072x1024xf32) matmul_47 = paddle._C_ops.matmul(relu_5, parameter_295, False, False) - del parameter_295 + del parameter_295, relu_5 # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1024xf32) add_55 = paddle._C_ops.add(matmul_47, parameter_294) - del parameter_294 + del matmul_47, parameter_294 # pd_op.dropout: (1x26x1024xf32, 1x26x1024xui8) <- (1x26x1024xf32, None, 1xf32) dropout_36, dropout_37 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_55, None, full_5, False, "upscale_in_train", 0, False + add_55, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1677,6 +1454,7 @@ def forward( # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1x26x1024xf32) add_56 = paddle._C_ops.add(layer_norm_33, dropout_36) + del dropout_36, layer_norm_33 # pd_op.layer_norm: (1x26x1024xf32, 1x26xf32, 1x26xf32) <- (1x26x1024xf32, 1024xf32, 1024xf32) layer_norm_36, layer_norm_37, layer_norm_38 = (lambda x, f: f(x))( @@ -1685,7 +1463,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_290, parameter_291 + del add_56, parameter_290, parameter_291 # pd_op.matmul: (1x26x1024xf32) <- (1x26x1024xf32, 1024x1024xf32) matmul_48 = paddle._C_ops.matmul(layer_norm_36, parameter_289, False, False) @@ -1693,10 +1471,11 @@ def forward( # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1024xf32) add_57 = paddle._C_ops.add(matmul_48, parameter_288) - del parameter_288 + del matmul_48, parameter_288 # pd_op.reshape: (1x26x16x64xf32) <- (1x26x1024xf32, 4xi64) reshape_24 = paddle._C_ops.reshape(add_57, full_int_array_1) + del add_57 # pd_op.transpose: (1x16x26x64xf32) <- (1x26x16x64xf32) transpose_24 = paddle._C_ops.transpose(reshape_24, [0, 2, 1, 3]) @@ -1708,7 +1487,7 @@ def forward( # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1024xf32) add_58 = paddle._C_ops.add(matmul_49, parameter_286) - del parameter_286 + del matmul_49, parameter_286 # pd_op.matmul: (1x26x1024xf32) <- (1x26x1024xf32, 1024x1024xf32) matmul_50 = paddle._C_ops.matmul(layer_norm_36, parameter_285, False, False) @@ -1716,10 +1495,11 @@ def forward( # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1024xf32) add_59 = paddle._C_ops.add(matmul_50, parameter_284) - del parameter_284 + del matmul_50, parameter_284 # pd_op.reshape: (1x26x16x64xf32) <- (1x26x1024xf32, 4xi64) reshape_25 = paddle._C_ops.reshape(add_58, full_int_array_1) + del add_58 # pd_op.transpose: (1x16x26x64xf32) <- (1x26x16x64xf32) transpose_25 = paddle._C_ops.transpose(reshape_25, [0, 2, 1, 3]) @@ -1727,6 +1507,7 @@ def forward( # pd_op.reshape: (1x26x16x64xf32) <- (1x26x1024xf32, 4xi64) reshape_26 = paddle._C_ops.reshape(add_59, full_int_array_1) + del add_59 # pd_op.transpose: (1x16x26x64xf32) <- (1x26x16x64xf32) transpose_26 = paddle._C_ops.transpose(reshape_26, [0, 2, 1, 3]) @@ -1738,9 +1519,11 @@ def forward( # pd_op.matmul: (1x16x26x26xf32) <- (1x16x26x64xf32, 1x16x26x64xf32) matmul_51 = paddle._C_ops.matmul(scale_8, transpose_25, False, True) + del scale_8, transpose_25 # pd_op.add: (1x16x26x26xf32) <- (1x16x26x26xf32, 1x1x1x26xf32) add_60 = paddle._C_ops.add(matmul_51, unsqueeze_0) + del matmul_51 # pd_op.softmax: (1x16x26x26xf32) <- (1x16x26x26xf32) softmax_6 = paddle._C_ops.softmax(add_60, -1) @@ -1749,13 +1532,15 @@ def forward( # pd_op.dropout: (1x16x26x26xf32, 1x16x26x26xui8) <- (1x16x26x26xf32, None, 1xf32) dropout_38, dropout_39 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_6, None, full_5, False, "upscale_in_train", 0, False + softmax_6, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_6 # pd_op.matmul: (1x16x26x64xf32) <- (1x16x26x26xf32, 1x16x26x64xf32) matmul_52 = paddle._C_ops.matmul(dropout_38, transpose_26, False, False) + del dropout_38, transpose_26 # pd_op.transpose: (1x26x16x64xf32) <- (1x16x26x64xf32) transpose_27 = paddle._C_ops.transpose(matmul_52, [0, 2, 1, 3]) @@ -1763,19 +1548,20 @@ def forward( # pd_op.reshape: (1x26x1024xf32) <- (1x26x16x64xf32, 3xi64) reshape_27 = paddle._C_ops.reshape(transpose_27, full_int_array_2) + del transpose_27 # pd_op.matmul: (1x26x1024xf32) <- (1x26x1024xf32, 1024x1024xf32) matmul_53 = paddle._C_ops.matmul(reshape_27, parameter_283, False, False) - del parameter_283 + del parameter_283, reshape_27 # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1024xf32) add_61 = paddle._C_ops.add(matmul_53, parameter_282) - del parameter_282 + del matmul_53, parameter_282 # pd_op.dropout: (1x26x1024xf32, 1x26x1024xui8) <- (1x26x1024xf32, None, 1xf32) dropout_40, dropout_41 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_61, None, full_5, False, "upscale_in_train", 0, False + add_61, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1783,6 +1569,7 @@ def forward( # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1x26x1024xf32) add_62 = paddle._C_ops.add(layer_norm_36, dropout_40) + del dropout_40, layer_norm_36 # pd_op.layer_norm: (1x26x1024xf32, 1x26xf32, 1x26xf32) <- (1x26x1024xf32, 1024xf32, 1024xf32) layer_norm_39, layer_norm_40, layer_norm_41 = (lambda x, f: f(x))( @@ -1791,7 +1578,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_276, parameter_277 + del add_62, parameter_276, parameter_277 # pd_op.matmul: (1x26x3072xf32) <- (1x26x1024xf32, 1024x3072xf32) matmul_54 = paddle._C_ops.matmul(layer_norm_39, parameter_281, False, False) @@ -1799,7 +1586,7 @@ def forward( # pd_op.add: (1x26x3072xf32) <- (1x26x3072xf32, 3072xf32) add_63 = paddle._C_ops.add(matmul_54, parameter_280) - del parameter_280 + del matmul_54, parameter_280 # pd_op.relu: (1x26x3072xf32) <- (1x26x3072xf32) relu_6 = paddle._C_ops.relu(add_63) @@ -1807,16 +1594,16 @@ def forward( # pd_op.matmul: (1x26x1024xf32) <- (1x26x3072xf32, 3072x1024xf32) matmul_55 = paddle._C_ops.matmul(relu_6, parameter_279, False, False) - del parameter_279 + del parameter_279, relu_6 # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1024xf32) add_64 = paddle._C_ops.add(matmul_55, parameter_278) - del parameter_278 + del matmul_55, parameter_278 # pd_op.dropout: (1x26x1024xf32, 1x26x1024xui8) <- (1x26x1024xf32, None, 1xf32) dropout_42, dropout_43 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_64, None, full_5, False, "upscale_in_train", 0, False + add_64, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1824,6 +1611,7 @@ def forward( # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1x26x1024xf32) add_65 = paddle._C_ops.add(layer_norm_39, dropout_42) + del dropout_42, layer_norm_39 # pd_op.layer_norm: (1x26x1024xf32, 1x26xf32, 1x26xf32) <- (1x26x1024xf32, 1024xf32, 1024xf32) layer_norm_42, layer_norm_43, layer_norm_44 = (lambda x, f: f(x))( @@ -1832,7 +1620,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_274, parameter_275 + del add_65, parameter_274, parameter_275 # pd_op.matmul: (1x26x1024xf32) <- (1x26x1024xf32, 1024x1024xf32) matmul_56 = paddle._C_ops.matmul(layer_norm_42, parameter_273, False, False) @@ -1840,10 +1628,11 @@ def forward( # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1024xf32) add_66 = paddle._C_ops.add(matmul_56, parameter_272) - del parameter_272 + del matmul_56, parameter_272 # pd_op.reshape: (1x26x16x64xf32) <- (1x26x1024xf32, 4xi64) reshape_28 = paddle._C_ops.reshape(add_66, full_int_array_1) + del add_66 # pd_op.transpose: (1x16x26x64xf32) <- (1x26x16x64xf32) transpose_28 = paddle._C_ops.transpose(reshape_28, [0, 2, 1, 3]) @@ -1855,7 +1644,7 @@ def forward( # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1024xf32) add_67 = paddle._C_ops.add(matmul_57, parameter_270) - del parameter_270 + del matmul_57, parameter_270 # pd_op.matmul: (1x26x1024xf32) <- (1x26x1024xf32, 1024x1024xf32) matmul_58 = paddle._C_ops.matmul(layer_norm_42, parameter_269, False, False) @@ -1863,10 +1652,11 @@ def forward( # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1024xf32) add_68 = paddle._C_ops.add(matmul_58, parameter_268) - del parameter_268 + del matmul_58, parameter_268 # pd_op.reshape: (1x26x16x64xf32) <- (1x26x1024xf32, 4xi64) reshape_29 = paddle._C_ops.reshape(add_67, full_int_array_1) + del add_67 # pd_op.transpose: (1x16x26x64xf32) <- (1x26x16x64xf32) transpose_29 = paddle._C_ops.transpose(reshape_29, [0, 2, 1, 3]) @@ -1874,6 +1664,7 @@ def forward( # pd_op.reshape: (1x26x16x64xf32) <- (1x26x1024xf32, 4xi64) reshape_30 = paddle._C_ops.reshape(add_68, full_int_array_1) + del add_68 # pd_op.transpose: (1x16x26x64xf32) <- (1x26x16x64xf32) transpose_30 = paddle._C_ops.transpose(reshape_30, [0, 2, 1, 3]) @@ -1885,9 +1676,11 @@ def forward( # pd_op.matmul: (1x16x26x26xf32) <- (1x16x26x64xf32, 1x16x26x64xf32) matmul_59 = paddle._C_ops.matmul(scale_9, transpose_29, False, True) + del scale_9, transpose_29 # pd_op.add: (1x16x26x26xf32) <- (1x16x26x26xf32, 1x1x1x26xf32) add_69 = paddle._C_ops.add(matmul_59, unsqueeze_0) + del matmul_59 # pd_op.softmax: (1x16x26x26xf32) <- (1x16x26x26xf32) softmax_7 = paddle._C_ops.softmax(add_69, -1) @@ -1896,13 +1689,15 @@ def forward( # pd_op.dropout: (1x16x26x26xf32, 1x16x26x26xui8) <- (1x16x26x26xf32, None, 1xf32) dropout_44, dropout_45 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_7, None, full_5, False, "upscale_in_train", 0, False + softmax_7, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_7 # pd_op.matmul: (1x16x26x64xf32) <- (1x16x26x26xf32, 1x16x26x64xf32) matmul_60 = paddle._C_ops.matmul(dropout_44, transpose_30, False, False) + del dropout_44, transpose_30 # pd_op.transpose: (1x26x16x64xf32) <- (1x16x26x64xf32) transpose_31 = paddle._C_ops.transpose(matmul_60, [0, 2, 1, 3]) @@ -1910,19 +1705,20 @@ def forward( # pd_op.reshape: (1x26x1024xf32) <- (1x26x16x64xf32, 3xi64) reshape_31 = paddle._C_ops.reshape(transpose_31, full_int_array_2) + del transpose_31 # pd_op.matmul: (1x26x1024xf32) <- (1x26x1024xf32, 1024x1024xf32) matmul_61 = paddle._C_ops.matmul(reshape_31, parameter_267, False, False) - del parameter_267 + del parameter_267, reshape_31 # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1024xf32) add_70 = paddle._C_ops.add(matmul_61, parameter_266) - del parameter_266 + del matmul_61, parameter_266 # pd_op.dropout: (1x26x1024xf32, 1x26x1024xui8) <- (1x26x1024xf32, None, 1xf32) dropout_46, dropout_47 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_70, None, full_5, False, "upscale_in_train", 0, False + add_70, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1930,6 +1726,7 @@ def forward( # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1x26x1024xf32) add_71 = paddle._C_ops.add(layer_norm_42, dropout_46) + del dropout_46, layer_norm_42 # pd_op.layer_norm: (1x26x1024xf32, 1x26xf32, 1x26xf32) <- (1x26x1024xf32, 1024xf32, 1024xf32) layer_norm_45, layer_norm_46, layer_norm_47 = (lambda x, f: f(x))( @@ -1938,7 +1735,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_260, parameter_261 + del add_71, parameter_260, parameter_261 # pd_op.matmul: (1x26x3072xf32) <- (1x26x1024xf32, 1024x3072xf32) matmul_62 = paddle._C_ops.matmul(layer_norm_45, parameter_265, False, False) @@ -1946,7 +1743,7 @@ def forward( # pd_op.add: (1x26x3072xf32) <- (1x26x3072xf32, 3072xf32) add_72 = paddle._C_ops.add(matmul_62, parameter_264) - del parameter_264 + del matmul_62, parameter_264 # pd_op.relu: (1x26x3072xf32) <- (1x26x3072xf32) relu_7 = paddle._C_ops.relu(add_72) @@ -1954,16 +1751,16 @@ def forward( # pd_op.matmul: (1x26x1024xf32) <- (1x26x3072xf32, 3072x1024xf32) matmul_63 = paddle._C_ops.matmul(relu_7, parameter_263, False, False) - del parameter_263 + del parameter_263, relu_7 # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1024xf32) add_73 = paddle._C_ops.add(matmul_63, parameter_262) - del parameter_262 + del matmul_63, parameter_262 # pd_op.dropout: (1x26x1024xf32, 1x26x1024xui8) <- (1x26x1024xf32, None, 1xf32) dropout_48, dropout_49 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_73, None, full_5, False, "upscale_in_train", 0, False + add_73, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1971,6 +1768,7 @@ def forward( # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1x26x1024xf32) add_74 = paddle._C_ops.add(layer_norm_45, dropout_48) + del dropout_48, layer_norm_45 # pd_op.layer_norm: (1x26x1024xf32, 1x26xf32, 1x26xf32) <- (1x26x1024xf32, 1024xf32, 1024xf32) layer_norm_48, layer_norm_49, layer_norm_50 = (lambda x, f: f(x))( @@ -1979,7 +1777,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_258, parameter_259 + del add_74, parameter_258, parameter_259 # pd_op.matmul: (1x26x1024xf32) <- (1x26x1024xf32, 1024x1024xf32) matmul_64 = paddle._C_ops.matmul(layer_norm_48, parameter_257, False, False) @@ -1987,10 +1785,11 @@ def forward( # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1024xf32) add_75 = paddle._C_ops.add(matmul_64, parameter_256) - del parameter_256 + del matmul_64, parameter_256 # pd_op.reshape: (1x26x16x64xf32) <- (1x26x1024xf32, 4xi64) reshape_32 = paddle._C_ops.reshape(add_75, full_int_array_1) + del add_75 # pd_op.transpose: (1x16x26x64xf32) <- (1x26x16x64xf32) transpose_32 = paddle._C_ops.transpose(reshape_32, [0, 2, 1, 3]) @@ -2002,7 +1801,7 @@ def forward( # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1024xf32) add_76 = paddle._C_ops.add(matmul_65, parameter_254) - del parameter_254 + del matmul_65, parameter_254 # pd_op.matmul: (1x26x1024xf32) <- (1x26x1024xf32, 1024x1024xf32) matmul_66 = paddle._C_ops.matmul(layer_norm_48, parameter_253, False, False) @@ -2010,10 +1809,11 @@ def forward( # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1024xf32) add_77 = paddle._C_ops.add(matmul_66, parameter_252) - del parameter_252 + del matmul_66, parameter_252 # pd_op.reshape: (1x26x16x64xf32) <- (1x26x1024xf32, 4xi64) reshape_33 = paddle._C_ops.reshape(add_76, full_int_array_1) + del add_76 # pd_op.transpose: (1x16x26x64xf32) <- (1x26x16x64xf32) transpose_33 = paddle._C_ops.transpose(reshape_33, [0, 2, 1, 3]) @@ -2021,6 +1821,7 @@ def forward( # pd_op.reshape: (1x26x16x64xf32) <- (1x26x1024xf32, 4xi64) reshape_34 = paddle._C_ops.reshape(add_77, full_int_array_1) + del add_77 # pd_op.transpose: (1x16x26x64xf32) <- (1x26x16x64xf32) transpose_34 = paddle._C_ops.transpose(reshape_34, [0, 2, 1, 3]) @@ -2032,9 +1833,11 @@ def forward( # pd_op.matmul: (1x16x26x26xf32) <- (1x16x26x64xf32, 1x16x26x64xf32) matmul_67 = paddle._C_ops.matmul(scale_10, transpose_33, False, True) + del scale_10, transpose_33 # pd_op.add: (1x16x26x26xf32) <- (1x16x26x26xf32, 1x1x1x26xf32) add_78 = paddle._C_ops.add(matmul_67, unsqueeze_0) + del matmul_67 # pd_op.softmax: (1x16x26x26xf32) <- (1x16x26x26xf32) softmax_8 = paddle._C_ops.softmax(add_78, -1) @@ -2043,13 +1846,15 @@ def forward( # pd_op.dropout: (1x16x26x26xf32, 1x16x26x26xui8) <- (1x16x26x26xf32, None, 1xf32) dropout_50, dropout_51 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_8, None, full_5, False, "upscale_in_train", 0, False + softmax_8, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_8 # pd_op.matmul: (1x16x26x64xf32) <- (1x16x26x26xf32, 1x16x26x64xf32) matmul_68 = paddle._C_ops.matmul(dropout_50, transpose_34, False, False) + del dropout_50, transpose_34 # pd_op.transpose: (1x26x16x64xf32) <- (1x16x26x64xf32) transpose_35 = paddle._C_ops.transpose(matmul_68, [0, 2, 1, 3]) @@ -2057,19 +1862,20 @@ def forward( # pd_op.reshape: (1x26x1024xf32) <- (1x26x16x64xf32, 3xi64) reshape_35 = paddle._C_ops.reshape(transpose_35, full_int_array_2) + del transpose_35 # pd_op.matmul: (1x26x1024xf32) <- (1x26x1024xf32, 1024x1024xf32) matmul_69 = paddle._C_ops.matmul(reshape_35, parameter_251, False, False) - del parameter_251 + del parameter_251, reshape_35 # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1024xf32) add_79 = paddle._C_ops.add(matmul_69, parameter_250) - del parameter_250 + del matmul_69, parameter_250 # pd_op.dropout: (1x26x1024xf32, 1x26x1024xui8) <- (1x26x1024xf32, None, 1xf32) dropout_52, dropout_53 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_79, None, full_5, False, "upscale_in_train", 0, False + add_79, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -2077,6 +1883,7 @@ def forward( # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1x26x1024xf32) add_80 = paddle._C_ops.add(layer_norm_48, dropout_52) + del dropout_52, layer_norm_48 # pd_op.layer_norm: (1x26x1024xf32, 1x26xf32, 1x26xf32) <- (1x26x1024xf32, 1024xf32, 1024xf32) layer_norm_51, layer_norm_52, layer_norm_53 = (lambda x, f: f(x))( @@ -2085,7 +1892,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_244, parameter_245 + del add_80, parameter_244, parameter_245 # pd_op.matmul: (1x26x3072xf32) <- (1x26x1024xf32, 1024x3072xf32) matmul_70 = paddle._C_ops.matmul(layer_norm_51, parameter_249, False, False) @@ -2093,7 +1900,7 @@ def forward( # pd_op.add: (1x26x3072xf32) <- (1x26x3072xf32, 3072xf32) add_81 = paddle._C_ops.add(matmul_70, parameter_248) - del parameter_248 + del matmul_70, parameter_248 # pd_op.relu: (1x26x3072xf32) <- (1x26x3072xf32) relu_8 = paddle._C_ops.relu(add_81) @@ -2101,16 +1908,16 @@ def forward( # pd_op.matmul: (1x26x1024xf32) <- (1x26x3072xf32, 3072x1024xf32) matmul_71 = paddle._C_ops.matmul(relu_8, parameter_247, False, False) - del parameter_247 + del parameter_247, relu_8 # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1024xf32) add_82 = paddle._C_ops.add(matmul_71, parameter_246) - del parameter_246 + del matmul_71, parameter_246 # pd_op.dropout: (1x26x1024xf32, 1x26x1024xui8) <- (1x26x1024xf32, None, 1xf32) dropout_54, dropout_55 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_82, None, full_5, False, "upscale_in_train", 0, False + add_82, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -2118,6 +1925,7 @@ def forward( # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1x26x1024xf32) add_83 = paddle._C_ops.add(layer_norm_51, dropout_54) + del dropout_54, layer_norm_51 # pd_op.layer_norm: (1x26x1024xf32, 1x26xf32, 1x26xf32) <- (1x26x1024xf32, 1024xf32, 1024xf32) layer_norm_54, layer_norm_55, layer_norm_56 = (lambda x, f: f(x))( @@ -2126,7 +1934,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_242, parameter_243 + del add_83, parameter_242, parameter_243 # pd_op.matmul: (1x26x1024xf32) <- (1x26x1024xf32, 1024x1024xf32) matmul_72 = paddle._C_ops.matmul(layer_norm_54, parameter_241, False, False) @@ -2134,10 +1942,11 @@ def forward( # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1024xf32) add_84 = paddle._C_ops.add(matmul_72, parameter_240) - del parameter_240 + del matmul_72, parameter_240 # pd_op.reshape: (1x26x16x64xf32) <- (1x26x1024xf32, 4xi64) reshape_36 = paddle._C_ops.reshape(add_84, full_int_array_1) + del add_84 # pd_op.transpose: (1x16x26x64xf32) <- (1x26x16x64xf32) transpose_36 = paddle._C_ops.transpose(reshape_36, [0, 2, 1, 3]) @@ -2149,7 +1958,7 @@ def forward( # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1024xf32) add_85 = paddle._C_ops.add(matmul_73, parameter_238) - del parameter_238 + del matmul_73, parameter_238 # pd_op.matmul: (1x26x1024xf32) <- (1x26x1024xf32, 1024x1024xf32) matmul_74 = paddle._C_ops.matmul(layer_norm_54, parameter_237, False, False) @@ -2157,10 +1966,11 @@ def forward( # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1024xf32) add_86 = paddle._C_ops.add(matmul_74, parameter_236) - del parameter_236 + del matmul_74, parameter_236 # pd_op.reshape: (1x26x16x64xf32) <- (1x26x1024xf32, 4xi64) reshape_37 = paddle._C_ops.reshape(add_85, full_int_array_1) + del add_85 # pd_op.transpose: (1x16x26x64xf32) <- (1x26x16x64xf32) transpose_37 = paddle._C_ops.transpose(reshape_37, [0, 2, 1, 3]) @@ -2168,6 +1978,7 @@ def forward( # pd_op.reshape: (1x26x16x64xf32) <- (1x26x1024xf32, 4xi64) reshape_38 = paddle._C_ops.reshape(add_86, full_int_array_1) + del add_86 # pd_op.transpose: (1x16x26x64xf32) <- (1x26x16x64xf32) transpose_38 = paddle._C_ops.transpose(reshape_38, [0, 2, 1, 3]) @@ -2179,9 +1990,11 @@ def forward( # pd_op.matmul: (1x16x26x26xf32) <- (1x16x26x64xf32, 1x16x26x64xf32) matmul_75 = paddle._C_ops.matmul(scale_11, transpose_37, False, True) + del scale_11, transpose_37 # pd_op.add: (1x16x26x26xf32) <- (1x16x26x26xf32, 1x1x1x26xf32) add_87 = paddle._C_ops.add(matmul_75, unsqueeze_0) + del matmul_75 # pd_op.softmax: (1x16x26x26xf32) <- (1x16x26x26xf32) softmax_9 = paddle._C_ops.softmax(add_87, -1) @@ -2190,13 +2003,15 @@ def forward( # pd_op.dropout: (1x16x26x26xf32, 1x16x26x26xui8) <- (1x16x26x26xf32, None, 1xf32) dropout_56, dropout_57 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_9, None, full_5, False, "upscale_in_train", 0, False + softmax_9, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_9 # pd_op.matmul: (1x16x26x64xf32) <- (1x16x26x26xf32, 1x16x26x64xf32) matmul_76 = paddle._C_ops.matmul(dropout_56, transpose_38, False, False) + del dropout_56, transpose_38 # pd_op.transpose: (1x26x16x64xf32) <- (1x16x26x64xf32) transpose_39 = paddle._C_ops.transpose(matmul_76, [0, 2, 1, 3]) @@ -2204,19 +2019,20 @@ def forward( # pd_op.reshape: (1x26x1024xf32) <- (1x26x16x64xf32, 3xi64) reshape_39 = paddle._C_ops.reshape(transpose_39, full_int_array_2) + del transpose_39 # pd_op.matmul: (1x26x1024xf32) <- (1x26x1024xf32, 1024x1024xf32) matmul_77 = paddle._C_ops.matmul(reshape_39, parameter_235, False, False) - del parameter_235 + del parameter_235, reshape_39 # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1024xf32) add_88 = paddle._C_ops.add(matmul_77, parameter_234) - del parameter_234 + del matmul_77, parameter_234 # pd_op.dropout: (1x26x1024xf32, 1x26x1024xui8) <- (1x26x1024xf32, None, 1xf32) dropout_58, dropout_59 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_88, None, full_5, False, "upscale_in_train", 0, False + add_88, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -2224,6 +2040,7 @@ def forward( # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1x26x1024xf32) add_89 = paddle._C_ops.add(layer_norm_54, dropout_58) + del dropout_58, layer_norm_54 # pd_op.layer_norm: (1x26x1024xf32, 1x26xf32, 1x26xf32) <- (1x26x1024xf32, 1024xf32, 1024xf32) layer_norm_57, layer_norm_58, layer_norm_59 = (lambda x, f: f(x))( @@ -2232,7 +2049,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_228, parameter_229 + del add_89, parameter_228, parameter_229 # pd_op.matmul: (1x26x3072xf32) <- (1x26x1024xf32, 1024x3072xf32) matmul_78 = paddle._C_ops.matmul(layer_norm_57, parameter_233, False, False) @@ -2240,7 +2057,7 @@ def forward( # pd_op.add: (1x26x3072xf32) <- (1x26x3072xf32, 3072xf32) add_90 = paddle._C_ops.add(matmul_78, parameter_232) - del parameter_232 + del matmul_78, parameter_232 # pd_op.relu: (1x26x3072xf32) <- (1x26x3072xf32) relu_9 = paddle._C_ops.relu(add_90) @@ -2248,16 +2065,16 @@ def forward( # pd_op.matmul: (1x26x1024xf32) <- (1x26x3072xf32, 3072x1024xf32) matmul_79 = paddle._C_ops.matmul(relu_9, parameter_231, False, False) - del parameter_231 + del parameter_231, relu_9 # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1024xf32) add_91 = paddle._C_ops.add(matmul_79, parameter_230) - del parameter_230 + del matmul_79, parameter_230 # pd_op.dropout: (1x26x1024xf32, 1x26x1024xui8) <- (1x26x1024xf32, None, 1xf32) dropout_60, dropout_61 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_91, None, full_5, False, "upscale_in_train", 0, False + add_91, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -2265,6 +2082,7 @@ def forward( # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1x26x1024xf32) add_92 = paddle._C_ops.add(layer_norm_57, dropout_60) + del dropout_60, layer_norm_57 # pd_op.layer_norm: (1x26x1024xf32, 1x26xf32, 1x26xf32) <- (1x26x1024xf32, 1024xf32, 1024xf32) layer_norm_60, layer_norm_61, layer_norm_62 = (lambda x, f: f(x))( @@ -2273,7 +2091,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_226, parameter_227 + del add_92, parameter_226, parameter_227 # pd_op.matmul: (1x26x1024xf32) <- (1x26x1024xf32, 1024x1024xf32) matmul_80 = paddle._C_ops.matmul(layer_norm_60, parameter_225, False, False) @@ -2281,10 +2099,11 @@ def forward( # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1024xf32) add_93 = paddle._C_ops.add(matmul_80, parameter_224) - del parameter_224 + del matmul_80, parameter_224 # pd_op.reshape: (1x26x16x64xf32) <- (1x26x1024xf32, 4xi64) reshape_40 = paddle._C_ops.reshape(add_93, full_int_array_1) + del add_93 # pd_op.transpose: (1x16x26x64xf32) <- (1x26x16x64xf32) transpose_40 = paddle._C_ops.transpose(reshape_40, [0, 2, 1, 3]) @@ -2296,7 +2115,7 @@ def forward( # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1024xf32) add_94 = paddle._C_ops.add(matmul_81, parameter_222) - del parameter_222 + del matmul_81, parameter_222 # pd_op.matmul: (1x26x1024xf32) <- (1x26x1024xf32, 1024x1024xf32) matmul_82 = paddle._C_ops.matmul(layer_norm_60, parameter_221, False, False) @@ -2304,10 +2123,11 @@ def forward( # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1024xf32) add_95 = paddle._C_ops.add(matmul_82, parameter_220) - del parameter_220 + del matmul_82, parameter_220 # pd_op.reshape: (1x26x16x64xf32) <- (1x26x1024xf32, 4xi64) reshape_41 = paddle._C_ops.reshape(add_94, full_int_array_1) + del add_94 # pd_op.transpose: (1x16x26x64xf32) <- (1x26x16x64xf32) transpose_41 = paddle._C_ops.transpose(reshape_41, [0, 2, 1, 3]) @@ -2315,6 +2135,7 @@ def forward( # pd_op.reshape: (1x26x16x64xf32) <- (1x26x1024xf32, 4xi64) reshape_42 = paddle._C_ops.reshape(add_95, full_int_array_1) + del add_95 # pd_op.transpose: (1x16x26x64xf32) <- (1x26x16x64xf32) transpose_42 = paddle._C_ops.transpose(reshape_42, [0, 2, 1, 3]) @@ -2326,9 +2147,11 @@ def forward( # pd_op.matmul: (1x16x26x26xf32) <- (1x16x26x64xf32, 1x16x26x64xf32) matmul_83 = paddle._C_ops.matmul(scale_12, transpose_41, False, True) + del scale_12, transpose_41 # pd_op.add: (1x16x26x26xf32) <- (1x16x26x26xf32, 1x1x1x26xf32) add_96 = paddle._C_ops.add(matmul_83, unsqueeze_0) + del matmul_83 # pd_op.softmax: (1x16x26x26xf32) <- (1x16x26x26xf32) softmax_10 = paddle._C_ops.softmax(add_96, -1) @@ -2337,13 +2160,15 @@ def forward( # pd_op.dropout: (1x16x26x26xf32, 1x16x26x26xui8) <- (1x16x26x26xf32, None, 1xf32) dropout_62, dropout_63 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_10, None, full_5, False, "upscale_in_train", 0, False + softmax_10, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_10 # pd_op.matmul: (1x16x26x64xf32) <- (1x16x26x26xf32, 1x16x26x64xf32) matmul_84 = paddle._C_ops.matmul(dropout_62, transpose_42, False, False) + del dropout_62, transpose_42 # pd_op.transpose: (1x26x16x64xf32) <- (1x16x26x64xf32) transpose_43 = paddle._C_ops.transpose(matmul_84, [0, 2, 1, 3]) @@ -2351,19 +2176,20 @@ def forward( # pd_op.reshape: (1x26x1024xf32) <- (1x26x16x64xf32, 3xi64) reshape_43 = paddle._C_ops.reshape(transpose_43, full_int_array_2) + del transpose_43 # pd_op.matmul: (1x26x1024xf32) <- (1x26x1024xf32, 1024x1024xf32) matmul_85 = paddle._C_ops.matmul(reshape_43, parameter_219, False, False) - del parameter_219 + del parameter_219, reshape_43 # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1024xf32) add_97 = paddle._C_ops.add(matmul_85, parameter_218) - del parameter_218 + del matmul_85, parameter_218 # pd_op.dropout: (1x26x1024xf32, 1x26x1024xui8) <- (1x26x1024xf32, None, 1xf32) dropout_64, dropout_65 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_97, None, full_5, False, "upscale_in_train", 0, False + add_97, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -2371,6 +2197,7 @@ def forward( # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1x26x1024xf32) add_98 = paddle._C_ops.add(layer_norm_60, dropout_64) + del dropout_64, layer_norm_60 # pd_op.layer_norm: (1x26x1024xf32, 1x26xf32, 1x26xf32) <- (1x26x1024xf32, 1024xf32, 1024xf32) layer_norm_63, layer_norm_64, layer_norm_65 = (lambda x, f: f(x))( @@ -2379,7 +2206,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_212, parameter_213 + del add_98, parameter_212, parameter_213 # pd_op.matmul: (1x26x3072xf32) <- (1x26x1024xf32, 1024x3072xf32) matmul_86 = paddle._C_ops.matmul(layer_norm_63, parameter_217, False, False) @@ -2387,7 +2214,7 @@ def forward( # pd_op.add: (1x26x3072xf32) <- (1x26x3072xf32, 3072xf32) add_99 = paddle._C_ops.add(matmul_86, parameter_216) - del parameter_216 + del matmul_86, parameter_216 # pd_op.relu: (1x26x3072xf32) <- (1x26x3072xf32) relu_10 = paddle._C_ops.relu(add_99) @@ -2395,16 +2222,16 @@ def forward( # pd_op.matmul: (1x26x1024xf32) <- (1x26x3072xf32, 3072x1024xf32) matmul_87 = paddle._C_ops.matmul(relu_10, parameter_215, False, False) - del parameter_215 + del parameter_215, relu_10 # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1024xf32) add_100 = paddle._C_ops.add(matmul_87, parameter_214) - del parameter_214 + del matmul_87, parameter_214 # pd_op.dropout: (1x26x1024xf32, 1x26x1024xui8) <- (1x26x1024xf32, None, 1xf32) dropout_66, dropout_67 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_100, None, full_5, False, "upscale_in_train", 0, False + add_100, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -2412,6 +2239,7 @@ def forward( # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1x26x1024xf32) add_101 = paddle._C_ops.add(layer_norm_63, dropout_66) + del dropout_66, layer_norm_63 # pd_op.layer_norm: (1x26x1024xf32, 1x26xf32, 1x26xf32) <- (1x26x1024xf32, 1024xf32, 1024xf32) layer_norm_66, layer_norm_67, layer_norm_68 = (lambda x, f: f(x))( @@ -2420,7 +2248,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_210, parameter_211 + del add_101, parameter_210, parameter_211 # pd_op.matmul: (1x26x1024xf32) <- (1x26x1024xf32, 1024x1024xf32) matmul_88 = paddle._C_ops.matmul(layer_norm_66, parameter_209, False, False) @@ -2428,10 +2256,11 @@ def forward( # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1024xf32) add_102 = paddle._C_ops.add(matmul_88, parameter_208) - del parameter_208 + del matmul_88, parameter_208 # pd_op.reshape: (1x26x16x64xf32) <- (1x26x1024xf32, 4xi64) reshape_44 = paddle._C_ops.reshape(add_102, full_int_array_1) + del add_102 # pd_op.transpose: (1x16x26x64xf32) <- (1x26x16x64xf32) transpose_44 = paddle._C_ops.transpose(reshape_44, [0, 2, 1, 3]) @@ -2443,7 +2272,7 @@ def forward( # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1024xf32) add_103 = paddle._C_ops.add(matmul_89, parameter_206) - del parameter_206 + del matmul_89, parameter_206 # pd_op.matmul: (1x26x1024xf32) <- (1x26x1024xf32, 1024x1024xf32) matmul_90 = paddle._C_ops.matmul(layer_norm_66, parameter_205, False, False) @@ -2451,10 +2280,11 @@ def forward( # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1024xf32) add_104 = paddle._C_ops.add(matmul_90, parameter_204) - del parameter_204 + del matmul_90, parameter_204 # pd_op.reshape: (1x26x16x64xf32) <- (1x26x1024xf32, 4xi64) reshape_45 = paddle._C_ops.reshape(add_103, full_int_array_1) + del add_103 # pd_op.transpose: (1x16x26x64xf32) <- (1x26x16x64xf32) transpose_45 = paddle._C_ops.transpose(reshape_45, [0, 2, 1, 3]) @@ -2462,6 +2292,7 @@ def forward( # pd_op.reshape: (1x26x16x64xf32) <- (1x26x1024xf32, 4xi64) reshape_46 = paddle._C_ops.reshape(add_104, full_int_array_1) + del add_104 # pd_op.transpose: (1x16x26x64xf32) <- (1x26x16x64xf32) transpose_46 = paddle._C_ops.transpose(reshape_46, [0, 2, 1, 3]) @@ -2473,9 +2304,11 @@ def forward( # pd_op.matmul: (1x16x26x26xf32) <- (1x16x26x64xf32, 1x16x26x64xf32) matmul_91 = paddle._C_ops.matmul(scale_13, transpose_45, False, True) + del scale_13, transpose_45 # pd_op.add: (1x16x26x26xf32) <- (1x16x26x26xf32, 1x1x1x26xf32) add_105 = paddle._C_ops.add(matmul_91, unsqueeze_0) + del matmul_91 # pd_op.softmax: (1x16x26x26xf32) <- (1x16x26x26xf32) softmax_11 = paddle._C_ops.softmax(add_105, -1) @@ -2484,13 +2317,15 @@ def forward( # pd_op.dropout: (1x16x26x26xf32, 1x16x26x26xui8) <- (1x16x26x26xf32, None, 1xf32) dropout_68, dropout_69 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_11, None, full_5, False, "upscale_in_train", 0, False + softmax_11, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_11 # pd_op.matmul: (1x16x26x64xf32) <- (1x16x26x26xf32, 1x16x26x64xf32) matmul_92 = paddle._C_ops.matmul(dropout_68, transpose_46, False, False) + del dropout_68, transpose_46 # pd_op.transpose: (1x26x16x64xf32) <- (1x16x26x64xf32) transpose_47 = paddle._C_ops.transpose(matmul_92, [0, 2, 1, 3]) @@ -2498,19 +2333,20 @@ def forward( # pd_op.reshape: (1x26x1024xf32) <- (1x26x16x64xf32, 3xi64) reshape_47 = paddle._C_ops.reshape(transpose_47, full_int_array_2) + del transpose_47 # pd_op.matmul: (1x26x1024xf32) <- (1x26x1024xf32, 1024x1024xf32) matmul_93 = paddle._C_ops.matmul(reshape_47, parameter_203, False, False) - del parameter_203 + del parameter_203, reshape_47 # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1024xf32) add_106 = paddle._C_ops.add(matmul_93, parameter_202) - del parameter_202 + del matmul_93, parameter_202 # pd_op.dropout: (1x26x1024xf32, 1x26x1024xui8) <- (1x26x1024xf32, None, 1xf32) dropout_70, dropout_71 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_106, None, full_5, False, "upscale_in_train", 0, False + add_106, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -2518,6 +2354,7 @@ def forward( # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1x26x1024xf32) add_107 = paddle._C_ops.add(layer_norm_66, dropout_70) + del dropout_70, layer_norm_66 # pd_op.layer_norm: (1x26x1024xf32, 1x26xf32, 1x26xf32) <- (1x26x1024xf32, 1024xf32, 1024xf32) layer_norm_69, layer_norm_70, layer_norm_71 = (lambda x, f: f(x))( @@ -2526,7 +2363,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_196, parameter_197 + del add_107, parameter_196, parameter_197 # pd_op.matmul: (1x26x3072xf32) <- (1x26x1024xf32, 1024x3072xf32) matmul_94 = paddle._C_ops.matmul(layer_norm_69, parameter_201, False, False) @@ -2534,7 +2371,7 @@ def forward( # pd_op.add: (1x26x3072xf32) <- (1x26x3072xf32, 3072xf32) add_108 = paddle._C_ops.add(matmul_94, parameter_200) - del parameter_200 + del matmul_94, parameter_200 # pd_op.relu: (1x26x3072xf32) <- (1x26x3072xf32) relu_11 = paddle._C_ops.relu(add_108) @@ -2542,16 +2379,16 @@ def forward( # pd_op.matmul: (1x26x1024xf32) <- (1x26x3072xf32, 3072x1024xf32) matmul_95 = paddle._C_ops.matmul(relu_11, parameter_199, False, False) - del parameter_199 + del parameter_199, relu_11 # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1024xf32) add_109 = paddle._C_ops.add(matmul_95, parameter_198) - del parameter_198 + del matmul_95, parameter_198 # pd_op.dropout: (1x26x1024xf32, 1x26x1024xui8) <- (1x26x1024xf32, None, 1xf32) dropout_72, dropout_73 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_109, None, full_5, False, "upscale_in_train", 0, False + add_109, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -2559,6 +2396,7 @@ def forward( # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1x26x1024xf32) add_110 = paddle._C_ops.add(layer_norm_69, dropout_72) + del dropout_72, layer_norm_69 # pd_op.layer_norm: (1x26x1024xf32, 1x26xf32, 1x26xf32) <- (1x26x1024xf32, 1024xf32, 1024xf32) layer_norm_72, layer_norm_73, layer_norm_74 = (lambda x, f: f(x))( @@ -2567,7 +2405,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_194, parameter_195 + del add_110, parameter_194, parameter_195 # pd_op.matmul: (1x26x1024xf32) <- (1x26x1024xf32, 1024x1024xf32) matmul_96 = paddle._C_ops.matmul(layer_norm_72, parameter_193, False, False) @@ -2575,10 +2413,11 @@ def forward( # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1024xf32) add_111 = paddle._C_ops.add(matmul_96, parameter_192) - del parameter_192 + del matmul_96, parameter_192 # pd_op.reshape: (1x26x16x64xf32) <- (1x26x1024xf32, 4xi64) reshape_48 = paddle._C_ops.reshape(add_111, full_int_array_1) + del add_111 # pd_op.transpose: (1x16x26x64xf32) <- (1x26x16x64xf32) transpose_48 = paddle._C_ops.transpose(reshape_48, [0, 2, 1, 3]) @@ -2590,7 +2429,7 @@ def forward( # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1024xf32) add_112 = paddle._C_ops.add(matmul_97, parameter_190) - del parameter_190 + del matmul_97, parameter_190 # pd_op.matmul: (1x26x1024xf32) <- (1x26x1024xf32, 1024x1024xf32) matmul_98 = paddle._C_ops.matmul(layer_norm_72, parameter_189, False, False) @@ -2598,10 +2437,11 @@ def forward( # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1024xf32) add_113 = paddle._C_ops.add(matmul_98, parameter_188) - del parameter_188 + del matmul_98, parameter_188 # pd_op.reshape: (1x26x16x64xf32) <- (1x26x1024xf32, 4xi64) reshape_49 = paddle._C_ops.reshape(add_112, full_int_array_1) + del add_112 # pd_op.transpose: (1x16x26x64xf32) <- (1x26x16x64xf32) transpose_49 = paddle._C_ops.transpose(reshape_49, [0, 2, 1, 3]) @@ -2609,6 +2449,7 @@ def forward( # pd_op.reshape: (1x26x16x64xf32) <- (1x26x1024xf32, 4xi64) reshape_50 = paddle._C_ops.reshape(add_113, full_int_array_1) + del add_113 # pd_op.transpose: (1x16x26x64xf32) <- (1x26x16x64xf32) transpose_50 = paddle._C_ops.transpose(reshape_50, [0, 2, 1, 3]) @@ -2620,9 +2461,11 @@ def forward( # pd_op.matmul: (1x16x26x26xf32) <- (1x16x26x64xf32, 1x16x26x64xf32) matmul_99 = paddle._C_ops.matmul(scale_14, transpose_49, False, True) + del scale_14, transpose_49 # pd_op.add: (1x16x26x26xf32) <- (1x16x26x26xf32, 1x1x1x26xf32) add_114 = paddle._C_ops.add(matmul_99, unsqueeze_0) + del matmul_99 # pd_op.softmax: (1x16x26x26xf32) <- (1x16x26x26xf32) softmax_12 = paddle._C_ops.softmax(add_114, -1) @@ -2631,13 +2474,15 @@ def forward( # pd_op.dropout: (1x16x26x26xf32, 1x16x26x26xui8) <- (1x16x26x26xf32, None, 1xf32) dropout_74, dropout_75 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_12, None, full_5, False, "upscale_in_train", 0, False + softmax_12, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_12 # pd_op.matmul: (1x16x26x64xf32) <- (1x16x26x26xf32, 1x16x26x64xf32) matmul_100 = paddle._C_ops.matmul(dropout_74, transpose_50, False, False) + del dropout_74, transpose_50 # pd_op.transpose: (1x26x16x64xf32) <- (1x16x26x64xf32) transpose_51 = paddle._C_ops.transpose(matmul_100, [0, 2, 1, 3]) @@ -2645,19 +2490,20 @@ def forward( # pd_op.reshape: (1x26x1024xf32) <- (1x26x16x64xf32, 3xi64) reshape_51 = paddle._C_ops.reshape(transpose_51, full_int_array_2) + del transpose_51 # pd_op.matmul: (1x26x1024xf32) <- (1x26x1024xf32, 1024x1024xf32) matmul_101 = paddle._C_ops.matmul(reshape_51, parameter_187, False, False) - del parameter_187 + del parameter_187, reshape_51 # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1024xf32) add_115 = paddle._C_ops.add(matmul_101, parameter_186) - del parameter_186 + del matmul_101, parameter_186 # pd_op.dropout: (1x26x1024xf32, 1x26x1024xui8) <- (1x26x1024xf32, None, 1xf32) dropout_76, dropout_77 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_115, None, full_5, False, "upscale_in_train", 0, False + add_115, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -2665,6 +2511,7 @@ def forward( # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1x26x1024xf32) add_116 = paddle._C_ops.add(layer_norm_72, dropout_76) + del dropout_76, layer_norm_72 # pd_op.layer_norm: (1x26x1024xf32, 1x26xf32, 1x26xf32) <- (1x26x1024xf32, 1024xf32, 1024xf32) layer_norm_75, layer_norm_76, layer_norm_77 = (lambda x, f: f(x))( @@ -2673,7 +2520,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_180, parameter_181 + del add_116, parameter_180, parameter_181 # pd_op.matmul: (1x26x3072xf32) <- (1x26x1024xf32, 1024x3072xf32) matmul_102 = paddle._C_ops.matmul(layer_norm_75, parameter_185, False, False) @@ -2681,7 +2528,7 @@ def forward( # pd_op.add: (1x26x3072xf32) <- (1x26x3072xf32, 3072xf32) add_117 = paddle._C_ops.add(matmul_102, parameter_184) - del parameter_184 + del matmul_102, parameter_184 # pd_op.relu: (1x26x3072xf32) <- (1x26x3072xf32) relu_12 = paddle._C_ops.relu(add_117) @@ -2689,16 +2536,16 @@ def forward( # pd_op.matmul: (1x26x1024xf32) <- (1x26x3072xf32, 3072x1024xf32) matmul_103 = paddle._C_ops.matmul(relu_12, parameter_183, False, False) - del parameter_183 + del parameter_183, relu_12 # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1024xf32) add_118 = paddle._C_ops.add(matmul_103, parameter_182) - del parameter_182 + del matmul_103, parameter_182 # pd_op.dropout: (1x26x1024xf32, 1x26x1024xui8) <- (1x26x1024xf32, None, 1xf32) dropout_78, dropout_79 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_118, None, full_5, False, "upscale_in_train", 0, False + add_118, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -2706,6 +2553,7 @@ def forward( # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1x26x1024xf32) add_119 = paddle._C_ops.add(layer_norm_75, dropout_78) + del dropout_78, layer_norm_75 # pd_op.layer_norm: (1x26x1024xf32, 1x26xf32, 1x26xf32) <- (1x26x1024xf32, 1024xf32, 1024xf32) layer_norm_78, layer_norm_79, layer_norm_80 = (lambda x, f: f(x))( @@ -2714,7 +2562,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_178, parameter_179 + del add_119, parameter_178, parameter_179 # pd_op.matmul: (1x26x1024xf32) <- (1x26x1024xf32, 1024x1024xf32) matmul_104 = paddle._C_ops.matmul(layer_norm_78, parameter_177, False, False) @@ -2722,10 +2570,11 @@ def forward( # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1024xf32) add_120 = paddle._C_ops.add(matmul_104, parameter_176) - del parameter_176 + del matmul_104, parameter_176 # pd_op.reshape: (1x26x16x64xf32) <- (1x26x1024xf32, 4xi64) reshape_52 = paddle._C_ops.reshape(add_120, full_int_array_1) + del add_120 # pd_op.transpose: (1x16x26x64xf32) <- (1x26x16x64xf32) transpose_52 = paddle._C_ops.transpose(reshape_52, [0, 2, 1, 3]) @@ -2737,7 +2586,7 @@ def forward( # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1024xf32) add_121 = paddle._C_ops.add(matmul_105, parameter_174) - del parameter_174 + del matmul_105, parameter_174 # pd_op.matmul: (1x26x1024xf32) <- (1x26x1024xf32, 1024x1024xf32) matmul_106 = paddle._C_ops.matmul(layer_norm_78, parameter_173, False, False) @@ -2745,10 +2594,11 @@ def forward( # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1024xf32) add_122 = paddle._C_ops.add(matmul_106, parameter_172) - del parameter_172 + del matmul_106, parameter_172 # pd_op.reshape: (1x26x16x64xf32) <- (1x26x1024xf32, 4xi64) reshape_53 = paddle._C_ops.reshape(add_121, full_int_array_1) + del add_121 # pd_op.transpose: (1x16x26x64xf32) <- (1x26x16x64xf32) transpose_53 = paddle._C_ops.transpose(reshape_53, [0, 2, 1, 3]) @@ -2756,6 +2606,7 @@ def forward( # pd_op.reshape: (1x26x16x64xf32) <- (1x26x1024xf32, 4xi64) reshape_54 = paddle._C_ops.reshape(add_122, full_int_array_1) + del add_122 # pd_op.transpose: (1x16x26x64xf32) <- (1x26x16x64xf32) transpose_54 = paddle._C_ops.transpose(reshape_54, [0, 2, 1, 3]) @@ -2767,9 +2618,11 @@ def forward( # pd_op.matmul: (1x16x26x26xf32) <- (1x16x26x64xf32, 1x16x26x64xf32) matmul_107 = paddle._C_ops.matmul(scale_15, transpose_53, False, True) + del scale_15, transpose_53 # pd_op.add: (1x16x26x26xf32) <- (1x16x26x26xf32, 1x1x1x26xf32) add_123 = paddle._C_ops.add(matmul_107, unsqueeze_0) + del matmul_107 # pd_op.softmax: (1x16x26x26xf32) <- (1x16x26x26xf32) softmax_13 = paddle._C_ops.softmax(add_123, -1) @@ -2778,13 +2631,15 @@ def forward( # pd_op.dropout: (1x16x26x26xf32, 1x16x26x26xui8) <- (1x16x26x26xf32, None, 1xf32) dropout_80, dropout_81 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_13, None, full_5, False, "upscale_in_train", 0, False + softmax_13, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_13 # pd_op.matmul: (1x16x26x64xf32) <- (1x16x26x26xf32, 1x16x26x64xf32) matmul_108 = paddle._C_ops.matmul(dropout_80, transpose_54, False, False) + del dropout_80, transpose_54 # pd_op.transpose: (1x26x16x64xf32) <- (1x16x26x64xf32) transpose_55 = paddle._C_ops.transpose(matmul_108, [0, 2, 1, 3]) @@ -2792,19 +2647,20 @@ def forward( # pd_op.reshape: (1x26x1024xf32) <- (1x26x16x64xf32, 3xi64) reshape_55 = paddle._C_ops.reshape(transpose_55, full_int_array_2) + del transpose_55 # pd_op.matmul: (1x26x1024xf32) <- (1x26x1024xf32, 1024x1024xf32) matmul_109 = paddle._C_ops.matmul(reshape_55, parameter_171, False, False) - del parameter_171 + del parameter_171, reshape_55 # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1024xf32) add_124 = paddle._C_ops.add(matmul_109, parameter_170) - del parameter_170 + del matmul_109, parameter_170 # pd_op.dropout: (1x26x1024xf32, 1x26x1024xui8) <- (1x26x1024xf32, None, 1xf32) dropout_82, dropout_83 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_124, None, full_5, False, "upscale_in_train", 0, False + add_124, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -2812,6 +2668,7 @@ def forward( # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1x26x1024xf32) add_125 = paddle._C_ops.add(layer_norm_78, dropout_82) + del dropout_82, layer_norm_78 # pd_op.layer_norm: (1x26x1024xf32, 1x26xf32, 1x26xf32) <- (1x26x1024xf32, 1024xf32, 1024xf32) layer_norm_81, layer_norm_82, layer_norm_83 = (lambda x, f: f(x))( @@ -2820,7 +2677,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_164, parameter_165 + del add_125, parameter_164, parameter_165 # pd_op.matmul: (1x26x3072xf32) <- (1x26x1024xf32, 1024x3072xf32) matmul_110 = paddle._C_ops.matmul(layer_norm_81, parameter_169, False, False) @@ -2828,7 +2685,7 @@ def forward( # pd_op.add: (1x26x3072xf32) <- (1x26x3072xf32, 3072xf32) add_126 = paddle._C_ops.add(matmul_110, parameter_168) - del parameter_168 + del matmul_110, parameter_168 # pd_op.relu: (1x26x3072xf32) <- (1x26x3072xf32) relu_13 = paddle._C_ops.relu(add_126) @@ -2836,16 +2693,16 @@ def forward( # pd_op.matmul: (1x26x1024xf32) <- (1x26x3072xf32, 3072x1024xf32) matmul_111 = paddle._C_ops.matmul(relu_13, parameter_167, False, False) - del parameter_167 + del parameter_167, relu_13 # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1024xf32) add_127 = paddle._C_ops.add(matmul_111, parameter_166) - del parameter_166 + del matmul_111, parameter_166 # pd_op.dropout: (1x26x1024xf32, 1x26x1024xui8) <- (1x26x1024xf32, None, 1xf32) dropout_84, dropout_85 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_127, None, full_5, False, "upscale_in_train", 0, False + add_127, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -2853,6 +2710,7 @@ def forward( # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1x26x1024xf32) add_128 = paddle._C_ops.add(layer_norm_81, dropout_84) + del dropout_84, layer_norm_81 # pd_op.layer_norm: (1x26x1024xf32, 1x26xf32, 1x26xf32) <- (1x26x1024xf32, 1024xf32, 1024xf32) layer_norm_84, layer_norm_85, layer_norm_86 = (lambda x, f: f(x))( @@ -2861,7 +2719,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_162, parameter_163 + del add_128, parameter_162, parameter_163 # pd_op.matmul: (1x26x1024xf32) <- (1x26x1024xf32, 1024x1024xf32) matmul_112 = paddle._C_ops.matmul(layer_norm_84, parameter_161, False, False) @@ -2869,10 +2727,11 @@ def forward( # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1024xf32) add_129 = paddle._C_ops.add(matmul_112, parameter_160) - del parameter_160 + del matmul_112, parameter_160 # pd_op.reshape: (1x26x16x64xf32) <- (1x26x1024xf32, 4xi64) reshape_56 = paddle._C_ops.reshape(add_129, full_int_array_1) + del add_129 # pd_op.transpose: (1x16x26x64xf32) <- (1x26x16x64xf32) transpose_56 = paddle._C_ops.transpose(reshape_56, [0, 2, 1, 3]) @@ -2884,7 +2743,7 @@ def forward( # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1024xf32) add_130 = paddle._C_ops.add(matmul_113, parameter_158) - del parameter_158 + del matmul_113, parameter_158 # pd_op.matmul: (1x26x1024xf32) <- (1x26x1024xf32, 1024x1024xf32) matmul_114 = paddle._C_ops.matmul(layer_norm_84, parameter_157, False, False) @@ -2892,10 +2751,11 @@ def forward( # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1024xf32) add_131 = paddle._C_ops.add(matmul_114, parameter_156) - del parameter_156 + del matmul_114, parameter_156 # pd_op.reshape: (1x26x16x64xf32) <- (1x26x1024xf32, 4xi64) reshape_57 = paddle._C_ops.reshape(add_130, full_int_array_1) + del add_130 # pd_op.transpose: (1x16x26x64xf32) <- (1x26x16x64xf32) transpose_57 = paddle._C_ops.transpose(reshape_57, [0, 2, 1, 3]) @@ -2903,6 +2763,7 @@ def forward( # pd_op.reshape: (1x26x16x64xf32) <- (1x26x1024xf32, 4xi64) reshape_58 = paddle._C_ops.reshape(add_131, full_int_array_1) + del add_131 # pd_op.transpose: (1x16x26x64xf32) <- (1x26x16x64xf32) transpose_58 = paddle._C_ops.transpose(reshape_58, [0, 2, 1, 3]) @@ -2914,9 +2775,11 @@ def forward( # pd_op.matmul: (1x16x26x26xf32) <- (1x16x26x64xf32, 1x16x26x64xf32) matmul_115 = paddle._C_ops.matmul(scale_16, transpose_57, False, True) + del scale_16, transpose_57 # pd_op.add: (1x16x26x26xf32) <- (1x16x26x26xf32, 1x1x1x26xf32) add_132 = paddle._C_ops.add(matmul_115, unsqueeze_0) + del matmul_115 # pd_op.softmax: (1x16x26x26xf32) <- (1x16x26x26xf32) softmax_14 = paddle._C_ops.softmax(add_132, -1) @@ -2925,13 +2788,15 @@ def forward( # pd_op.dropout: (1x16x26x26xf32, 1x16x26x26xui8) <- (1x16x26x26xf32, None, 1xf32) dropout_86, dropout_87 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_14, None, full_5, False, "upscale_in_train", 0, False + softmax_14, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_14 # pd_op.matmul: (1x16x26x64xf32) <- (1x16x26x26xf32, 1x16x26x64xf32) matmul_116 = paddle._C_ops.matmul(dropout_86, transpose_58, False, False) + del dropout_86, transpose_58 # pd_op.transpose: (1x26x16x64xf32) <- (1x16x26x64xf32) transpose_59 = paddle._C_ops.transpose(matmul_116, [0, 2, 1, 3]) @@ -2939,19 +2804,20 @@ def forward( # pd_op.reshape: (1x26x1024xf32) <- (1x26x16x64xf32, 3xi64) reshape_59 = paddle._C_ops.reshape(transpose_59, full_int_array_2) + del transpose_59 # pd_op.matmul: (1x26x1024xf32) <- (1x26x1024xf32, 1024x1024xf32) matmul_117 = paddle._C_ops.matmul(reshape_59, parameter_155, False, False) - del parameter_155 + del parameter_155, reshape_59 # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1024xf32) add_133 = paddle._C_ops.add(matmul_117, parameter_154) - del parameter_154 + del matmul_117, parameter_154 # pd_op.dropout: (1x26x1024xf32, 1x26x1024xui8) <- (1x26x1024xf32, None, 1xf32) dropout_88, dropout_89 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_133, None, full_5, False, "upscale_in_train", 0, False + add_133, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -2959,6 +2825,7 @@ def forward( # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1x26x1024xf32) add_134 = paddle._C_ops.add(layer_norm_84, dropout_88) + del dropout_88, layer_norm_84 # pd_op.layer_norm: (1x26x1024xf32, 1x26xf32, 1x26xf32) <- (1x26x1024xf32, 1024xf32, 1024xf32) layer_norm_87, layer_norm_88, layer_norm_89 = (lambda x, f: f(x))( @@ -2967,7 +2834,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_148, parameter_149 + del add_134, parameter_148, parameter_149 # pd_op.matmul: (1x26x3072xf32) <- (1x26x1024xf32, 1024x3072xf32) matmul_118 = paddle._C_ops.matmul(layer_norm_87, parameter_153, False, False) @@ -2975,7 +2842,7 @@ def forward( # pd_op.add: (1x26x3072xf32) <- (1x26x3072xf32, 3072xf32) add_135 = paddle._C_ops.add(matmul_118, parameter_152) - del parameter_152 + del matmul_118, parameter_152 # pd_op.relu: (1x26x3072xf32) <- (1x26x3072xf32) relu_14 = paddle._C_ops.relu(add_135) @@ -2983,16 +2850,16 @@ def forward( # pd_op.matmul: (1x26x1024xf32) <- (1x26x3072xf32, 3072x1024xf32) matmul_119 = paddle._C_ops.matmul(relu_14, parameter_151, False, False) - del parameter_151 + del parameter_151, relu_14 # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1024xf32) add_136 = paddle._C_ops.add(matmul_119, parameter_150) - del parameter_150 + del matmul_119, parameter_150 # pd_op.dropout: (1x26x1024xf32, 1x26x1024xui8) <- (1x26x1024xf32, None, 1xf32) dropout_90, dropout_91 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_136, None, full_5, False, "upscale_in_train", 0, False + add_136, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -3000,6 +2867,7 @@ def forward( # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1x26x1024xf32) add_137 = paddle._C_ops.add(layer_norm_87, dropout_90) + del dropout_90, layer_norm_87 # pd_op.layer_norm: (1x26x1024xf32, 1x26xf32, 1x26xf32) <- (1x26x1024xf32, 1024xf32, 1024xf32) layer_norm_90, layer_norm_91, layer_norm_92 = (lambda x, f: f(x))( @@ -3008,7 +2876,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_146, parameter_147 + del add_137, parameter_146, parameter_147 # pd_op.matmul: (1x26x1024xf32) <- (1x26x1024xf32, 1024x1024xf32) matmul_120 = paddle._C_ops.matmul(layer_norm_90, parameter_145, False, False) @@ -3016,10 +2884,11 @@ def forward( # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1024xf32) add_138 = paddle._C_ops.add(matmul_120, parameter_144) - del parameter_144 + del matmul_120, parameter_144 # pd_op.reshape: (1x26x16x64xf32) <- (1x26x1024xf32, 4xi64) reshape_60 = paddle._C_ops.reshape(add_138, full_int_array_1) + del add_138 # pd_op.transpose: (1x16x26x64xf32) <- (1x26x16x64xf32) transpose_60 = paddle._C_ops.transpose(reshape_60, [0, 2, 1, 3]) @@ -3031,7 +2900,7 @@ def forward( # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1024xf32) add_139 = paddle._C_ops.add(matmul_121, parameter_142) - del parameter_142 + del matmul_121, parameter_142 # pd_op.matmul: (1x26x1024xf32) <- (1x26x1024xf32, 1024x1024xf32) matmul_122 = paddle._C_ops.matmul(layer_norm_90, parameter_141, False, False) @@ -3039,10 +2908,11 @@ def forward( # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1024xf32) add_140 = paddle._C_ops.add(matmul_122, parameter_140) - del parameter_140 + del matmul_122, parameter_140 # pd_op.reshape: (1x26x16x64xf32) <- (1x26x1024xf32, 4xi64) reshape_61 = paddle._C_ops.reshape(add_139, full_int_array_1) + del add_139 # pd_op.transpose: (1x16x26x64xf32) <- (1x26x16x64xf32) transpose_61 = paddle._C_ops.transpose(reshape_61, [0, 2, 1, 3]) @@ -3050,6 +2920,7 @@ def forward( # pd_op.reshape: (1x26x16x64xf32) <- (1x26x1024xf32, 4xi64) reshape_62 = paddle._C_ops.reshape(add_140, full_int_array_1) + del add_140 # pd_op.transpose: (1x16x26x64xf32) <- (1x26x16x64xf32) transpose_62 = paddle._C_ops.transpose(reshape_62, [0, 2, 1, 3]) @@ -3061,9 +2932,11 @@ def forward( # pd_op.matmul: (1x16x26x26xf32) <- (1x16x26x64xf32, 1x16x26x64xf32) matmul_123 = paddle._C_ops.matmul(scale_17, transpose_61, False, True) + del scale_17, transpose_61 # pd_op.add: (1x16x26x26xf32) <- (1x16x26x26xf32, 1x1x1x26xf32) add_141 = paddle._C_ops.add(matmul_123, unsqueeze_0) + del matmul_123 # pd_op.softmax: (1x16x26x26xf32) <- (1x16x26x26xf32) softmax_15 = paddle._C_ops.softmax(add_141, -1) @@ -3072,13 +2945,15 @@ def forward( # pd_op.dropout: (1x16x26x26xf32, 1x16x26x26xui8) <- (1x16x26x26xf32, None, 1xf32) dropout_92, dropout_93 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_15, None, full_5, False, "upscale_in_train", 0, False + softmax_15, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_15 # pd_op.matmul: (1x16x26x64xf32) <- (1x16x26x26xf32, 1x16x26x64xf32) matmul_124 = paddle._C_ops.matmul(dropout_92, transpose_62, False, False) + del dropout_92, transpose_62 # pd_op.transpose: (1x26x16x64xf32) <- (1x16x26x64xf32) transpose_63 = paddle._C_ops.transpose(matmul_124, [0, 2, 1, 3]) @@ -3086,19 +2961,20 @@ def forward( # pd_op.reshape: (1x26x1024xf32) <- (1x26x16x64xf32, 3xi64) reshape_63 = paddle._C_ops.reshape(transpose_63, full_int_array_2) + del transpose_63 # pd_op.matmul: (1x26x1024xf32) <- (1x26x1024xf32, 1024x1024xf32) matmul_125 = paddle._C_ops.matmul(reshape_63, parameter_139, False, False) - del parameter_139 + del parameter_139, reshape_63 # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1024xf32) add_142 = paddle._C_ops.add(matmul_125, parameter_138) - del parameter_138 + del matmul_125, parameter_138 # pd_op.dropout: (1x26x1024xf32, 1x26x1024xui8) <- (1x26x1024xf32, None, 1xf32) dropout_94, dropout_95 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_142, None, full_5, False, "upscale_in_train", 0, False + add_142, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -3106,6 +2982,7 @@ def forward( # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1x26x1024xf32) add_143 = paddle._C_ops.add(layer_norm_90, dropout_94) + del dropout_94, layer_norm_90 # pd_op.layer_norm: (1x26x1024xf32, 1x26xf32, 1x26xf32) <- (1x26x1024xf32, 1024xf32, 1024xf32) layer_norm_93, layer_norm_94, layer_norm_95 = (lambda x, f: f(x))( @@ -3114,7 +2991,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_132, parameter_133 + del add_143, parameter_132, parameter_133 # pd_op.matmul: (1x26x3072xf32) <- (1x26x1024xf32, 1024x3072xf32) matmul_126 = paddle._C_ops.matmul(layer_norm_93, parameter_137, False, False) @@ -3122,7 +2999,7 @@ def forward( # pd_op.add: (1x26x3072xf32) <- (1x26x3072xf32, 3072xf32) add_144 = paddle._C_ops.add(matmul_126, parameter_136) - del parameter_136 + del matmul_126, parameter_136 # pd_op.relu: (1x26x3072xf32) <- (1x26x3072xf32) relu_15 = paddle._C_ops.relu(add_144) @@ -3130,16 +3007,16 @@ def forward( # pd_op.matmul: (1x26x1024xf32) <- (1x26x3072xf32, 3072x1024xf32) matmul_127 = paddle._C_ops.matmul(relu_15, parameter_135, False, False) - del parameter_135 + del parameter_135, relu_15 # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1024xf32) add_145 = paddle._C_ops.add(matmul_127, parameter_134) - del parameter_134 + del matmul_127, parameter_134 # pd_op.dropout: (1x26x1024xf32, 1x26x1024xui8) <- (1x26x1024xf32, None, 1xf32) dropout_96, dropout_97 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_145, None, full_5, False, "upscale_in_train", 0, False + add_145, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -3147,6 +3024,7 @@ def forward( # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1x26x1024xf32) add_146 = paddle._C_ops.add(layer_norm_93, dropout_96) + del dropout_96, layer_norm_93 # pd_op.layer_norm: (1x26x1024xf32, 1x26xf32, 1x26xf32) <- (1x26x1024xf32, 1024xf32, 1024xf32) layer_norm_96, layer_norm_97, layer_norm_98 = (lambda x, f: f(x))( @@ -3155,7 +3033,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_130, parameter_131 + del add_146, parameter_130, parameter_131 # pd_op.matmul: (1x26x1024xf32) <- (1x26x1024xf32, 1024x1024xf32) matmul_128 = paddle._C_ops.matmul(layer_norm_96, parameter_129, False, False) @@ -3163,10 +3041,11 @@ def forward( # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1024xf32) add_147 = paddle._C_ops.add(matmul_128, parameter_128) - del parameter_128 + del matmul_128, parameter_128 # pd_op.reshape: (1x26x16x64xf32) <- (1x26x1024xf32, 4xi64) reshape_64 = paddle._C_ops.reshape(add_147, full_int_array_1) + del add_147 # pd_op.transpose: (1x16x26x64xf32) <- (1x26x16x64xf32) transpose_64 = paddle._C_ops.transpose(reshape_64, [0, 2, 1, 3]) @@ -3178,7 +3057,7 @@ def forward( # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1024xf32) add_148 = paddle._C_ops.add(matmul_129, parameter_126) - del parameter_126 + del matmul_129, parameter_126 # pd_op.matmul: (1x26x1024xf32) <- (1x26x1024xf32, 1024x1024xf32) matmul_130 = paddle._C_ops.matmul(layer_norm_96, parameter_125, False, False) @@ -3186,10 +3065,11 @@ def forward( # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1024xf32) add_149 = paddle._C_ops.add(matmul_130, parameter_124) - del parameter_124 + del matmul_130, parameter_124 # pd_op.reshape: (1x26x16x64xf32) <- (1x26x1024xf32, 4xi64) reshape_65 = paddle._C_ops.reshape(add_148, full_int_array_1) + del add_148 # pd_op.transpose: (1x16x26x64xf32) <- (1x26x16x64xf32) transpose_65 = paddle._C_ops.transpose(reshape_65, [0, 2, 1, 3]) @@ -3197,6 +3077,7 @@ def forward( # pd_op.reshape: (1x26x16x64xf32) <- (1x26x1024xf32, 4xi64) reshape_66 = paddle._C_ops.reshape(add_149, full_int_array_1) + del add_149 # pd_op.transpose: (1x16x26x64xf32) <- (1x26x16x64xf32) transpose_66 = paddle._C_ops.transpose(reshape_66, [0, 2, 1, 3]) @@ -3208,9 +3089,11 @@ def forward( # pd_op.matmul: (1x16x26x26xf32) <- (1x16x26x64xf32, 1x16x26x64xf32) matmul_131 = paddle._C_ops.matmul(scale_18, transpose_65, False, True) + del scale_18, transpose_65 # pd_op.add: (1x16x26x26xf32) <- (1x16x26x26xf32, 1x1x1x26xf32) add_150 = paddle._C_ops.add(matmul_131, unsqueeze_0) + del matmul_131 # pd_op.softmax: (1x16x26x26xf32) <- (1x16x26x26xf32) softmax_16 = paddle._C_ops.softmax(add_150, -1) @@ -3219,13 +3102,15 @@ def forward( # pd_op.dropout: (1x16x26x26xf32, 1x16x26x26xui8) <- (1x16x26x26xf32, None, 1xf32) dropout_98, dropout_99 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_16, None, full_5, False, "upscale_in_train", 0, False + softmax_16, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_16 # pd_op.matmul: (1x16x26x64xf32) <- (1x16x26x26xf32, 1x16x26x64xf32) matmul_132 = paddle._C_ops.matmul(dropout_98, transpose_66, False, False) + del dropout_98, transpose_66 # pd_op.transpose: (1x26x16x64xf32) <- (1x16x26x64xf32) transpose_67 = paddle._C_ops.transpose(matmul_132, [0, 2, 1, 3]) @@ -3233,19 +3118,20 @@ def forward( # pd_op.reshape: (1x26x1024xf32) <- (1x26x16x64xf32, 3xi64) reshape_67 = paddle._C_ops.reshape(transpose_67, full_int_array_2) + del transpose_67 # pd_op.matmul: (1x26x1024xf32) <- (1x26x1024xf32, 1024x1024xf32) matmul_133 = paddle._C_ops.matmul(reshape_67, parameter_123, False, False) - del parameter_123 + del parameter_123, reshape_67 # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1024xf32) add_151 = paddle._C_ops.add(matmul_133, parameter_122) - del parameter_122 + del matmul_133, parameter_122 # pd_op.dropout: (1x26x1024xf32, 1x26x1024xui8) <- (1x26x1024xf32, None, 1xf32) dropout_100, dropout_101 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_151, None, full_5, False, "upscale_in_train", 0, False + add_151, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -3253,6 +3139,7 @@ def forward( # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1x26x1024xf32) add_152 = paddle._C_ops.add(layer_norm_96, dropout_100) + del dropout_100, layer_norm_96 # pd_op.layer_norm: (1x26x1024xf32, 1x26xf32, 1x26xf32) <- (1x26x1024xf32, 1024xf32, 1024xf32) layer_norm_99, layer_norm_100, layer_norm_101 = (lambda x, f: f(x))( @@ -3261,7 +3148,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_116, parameter_117 + del add_152, parameter_116, parameter_117 # pd_op.matmul: (1x26x3072xf32) <- (1x26x1024xf32, 1024x3072xf32) matmul_134 = paddle._C_ops.matmul(layer_norm_99, parameter_121, False, False) @@ -3269,7 +3156,7 @@ def forward( # pd_op.add: (1x26x3072xf32) <- (1x26x3072xf32, 3072xf32) add_153 = paddle._C_ops.add(matmul_134, parameter_120) - del parameter_120 + del matmul_134, parameter_120 # pd_op.relu: (1x26x3072xf32) <- (1x26x3072xf32) relu_16 = paddle._C_ops.relu(add_153) @@ -3277,16 +3164,16 @@ def forward( # pd_op.matmul: (1x26x1024xf32) <- (1x26x3072xf32, 3072x1024xf32) matmul_135 = paddle._C_ops.matmul(relu_16, parameter_119, False, False) - del parameter_119 + del parameter_119, relu_16 # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1024xf32) add_154 = paddle._C_ops.add(matmul_135, parameter_118) - del parameter_118 + del matmul_135, parameter_118 # pd_op.dropout: (1x26x1024xf32, 1x26x1024xui8) <- (1x26x1024xf32, None, 1xf32) dropout_102, dropout_103 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_154, None, full_5, False, "upscale_in_train", 0, False + add_154, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -3294,6 +3181,7 @@ def forward( # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1x26x1024xf32) add_155 = paddle._C_ops.add(layer_norm_99, dropout_102) + del dropout_102, layer_norm_99 # pd_op.layer_norm: (1x26x1024xf32, 1x26xf32, 1x26xf32) <- (1x26x1024xf32, 1024xf32, 1024xf32) layer_norm_102, layer_norm_103, layer_norm_104 = (lambda x, f: f(x))( @@ -3302,7 +3190,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_114, parameter_115 + del add_155, parameter_114, parameter_115 # pd_op.matmul: (1x26x1024xf32) <- (1x26x1024xf32, 1024x1024xf32) matmul_136 = paddle._C_ops.matmul(layer_norm_102, parameter_113, False, False) @@ -3310,10 +3198,11 @@ def forward( # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1024xf32) add_156 = paddle._C_ops.add(matmul_136, parameter_112) - del parameter_112 + del matmul_136, parameter_112 # pd_op.reshape: (1x26x16x64xf32) <- (1x26x1024xf32, 4xi64) reshape_68 = paddle._C_ops.reshape(add_156, full_int_array_1) + del add_156 # pd_op.transpose: (1x16x26x64xf32) <- (1x26x16x64xf32) transpose_68 = paddle._C_ops.transpose(reshape_68, [0, 2, 1, 3]) @@ -3325,7 +3214,7 @@ def forward( # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1024xf32) add_157 = paddle._C_ops.add(matmul_137, parameter_110) - del parameter_110 + del matmul_137, parameter_110 # pd_op.matmul: (1x26x1024xf32) <- (1x26x1024xf32, 1024x1024xf32) matmul_138 = paddle._C_ops.matmul(layer_norm_102, parameter_109, False, False) @@ -3333,10 +3222,11 @@ def forward( # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1024xf32) add_158 = paddle._C_ops.add(matmul_138, parameter_108) - del parameter_108 + del matmul_138, parameter_108 # pd_op.reshape: (1x26x16x64xf32) <- (1x26x1024xf32, 4xi64) reshape_69 = paddle._C_ops.reshape(add_157, full_int_array_1) + del add_157 # pd_op.transpose: (1x16x26x64xf32) <- (1x26x16x64xf32) transpose_69 = paddle._C_ops.transpose(reshape_69, [0, 2, 1, 3]) @@ -3344,6 +3234,7 @@ def forward( # pd_op.reshape: (1x26x16x64xf32) <- (1x26x1024xf32, 4xi64) reshape_70 = paddle._C_ops.reshape(add_158, full_int_array_1) + del add_158 # pd_op.transpose: (1x16x26x64xf32) <- (1x26x16x64xf32) transpose_70 = paddle._C_ops.transpose(reshape_70, [0, 2, 1, 3]) @@ -3355,9 +3246,11 @@ def forward( # pd_op.matmul: (1x16x26x26xf32) <- (1x16x26x64xf32, 1x16x26x64xf32) matmul_139 = paddle._C_ops.matmul(scale_19, transpose_69, False, True) + del scale_19, transpose_69 # pd_op.add: (1x16x26x26xf32) <- (1x16x26x26xf32, 1x1x1x26xf32) add_159 = paddle._C_ops.add(matmul_139, unsqueeze_0) + del matmul_139 # pd_op.softmax: (1x16x26x26xf32) <- (1x16x26x26xf32) softmax_17 = paddle._C_ops.softmax(add_159, -1) @@ -3366,13 +3259,15 @@ def forward( # pd_op.dropout: (1x16x26x26xf32, 1x16x26x26xui8) <- (1x16x26x26xf32, None, 1xf32) dropout_104, dropout_105 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_17, None, full_5, False, "upscale_in_train", 0, False + softmax_17, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_17 # pd_op.matmul: (1x16x26x64xf32) <- (1x16x26x26xf32, 1x16x26x64xf32) matmul_140 = paddle._C_ops.matmul(dropout_104, transpose_70, False, False) + del dropout_104, transpose_70 # pd_op.transpose: (1x26x16x64xf32) <- (1x16x26x64xf32) transpose_71 = paddle._C_ops.transpose(matmul_140, [0, 2, 1, 3]) @@ -3380,19 +3275,20 @@ def forward( # pd_op.reshape: (1x26x1024xf32) <- (1x26x16x64xf32, 3xi64) reshape_71 = paddle._C_ops.reshape(transpose_71, full_int_array_2) + del transpose_71 # pd_op.matmul: (1x26x1024xf32) <- (1x26x1024xf32, 1024x1024xf32) matmul_141 = paddle._C_ops.matmul(reshape_71, parameter_107, False, False) - del parameter_107 + del parameter_107, reshape_71 # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1024xf32) add_160 = paddle._C_ops.add(matmul_141, parameter_106) - del parameter_106 + del matmul_141, parameter_106 # pd_op.dropout: (1x26x1024xf32, 1x26x1024xui8) <- (1x26x1024xf32, None, 1xf32) dropout_106, dropout_107 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_160, None, full_5, False, "upscale_in_train", 0, False + add_160, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -3400,6 +3296,7 @@ def forward( # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1x26x1024xf32) add_161 = paddle._C_ops.add(layer_norm_102, dropout_106) + del dropout_106, layer_norm_102 # pd_op.layer_norm: (1x26x1024xf32, 1x26xf32, 1x26xf32) <- (1x26x1024xf32, 1024xf32, 1024xf32) layer_norm_105, layer_norm_106, layer_norm_107 = (lambda x, f: f(x))( @@ -3408,7 +3305,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_100, parameter_101 + del add_161, parameter_100, parameter_101 # pd_op.matmul: (1x26x3072xf32) <- (1x26x1024xf32, 1024x3072xf32) matmul_142 = paddle._C_ops.matmul(layer_norm_105, parameter_105, False, False) @@ -3416,7 +3313,7 @@ def forward( # pd_op.add: (1x26x3072xf32) <- (1x26x3072xf32, 3072xf32) add_162 = paddle._C_ops.add(matmul_142, parameter_104) - del parameter_104 + del matmul_142, parameter_104 # pd_op.relu: (1x26x3072xf32) <- (1x26x3072xf32) relu_17 = paddle._C_ops.relu(add_162) @@ -3424,16 +3321,16 @@ def forward( # pd_op.matmul: (1x26x1024xf32) <- (1x26x3072xf32, 3072x1024xf32) matmul_143 = paddle._C_ops.matmul(relu_17, parameter_103, False, False) - del parameter_103 + del parameter_103, relu_17 # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1024xf32) add_163 = paddle._C_ops.add(matmul_143, parameter_102) - del parameter_102 + del matmul_143, parameter_102 # pd_op.dropout: (1x26x1024xf32, 1x26x1024xui8) <- (1x26x1024xf32, None, 1xf32) dropout_108, dropout_109 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_163, None, full_5, False, "upscale_in_train", 0, False + add_163, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -3441,6 +3338,7 @@ def forward( # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1x26x1024xf32) add_164 = paddle._C_ops.add(layer_norm_105, dropout_108) + del dropout_108, layer_norm_105 # pd_op.layer_norm: (1x26x1024xf32, 1x26xf32, 1x26xf32) <- (1x26x1024xf32, 1024xf32, 1024xf32) layer_norm_108, layer_norm_109, layer_norm_110 = (lambda x, f: f(x))( @@ -3449,7 +3347,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_98, parameter_99 + del add_164, parameter_98, parameter_99 # pd_op.matmul: (1x26x1024xf32) <- (1x26x1024xf32, 1024x1024xf32) matmul_144 = paddle._C_ops.matmul(layer_norm_108, parameter_97, False, False) @@ -3457,10 +3355,11 @@ def forward( # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1024xf32) add_165 = paddle._C_ops.add(matmul_144, parameter_96) - del parameter_96 + del matmul_144, parameter_96 # pd_op.reshape: (1x26x16x64xf32) <- (1x26x1024xf32, 4xi64) reshape_72 = paddle._C_ops.reshape(add_165, full_int_array_1) + del add_165 # pd_op.transpose: (1x16x26x64xf32) <- (1x26x16x64xf32) transpose_72 = paddle._C_ops.transpose(reshape_72, [0, 2, 1, 3]) @@ -3472,7 +3371,7 @@ def forward( # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1024xf32) add_166 = paddle._C_ops.add(matmul_145, parameter_94) - del parameter_94 + del matmul_145, parameter_94 # pd_op.matmul: (1x26x1024xf32) <- (1x26x1024xf32, 1024x1024xf32) matmul_146 = paddle._C_ops.matmul(layer_norm_108, parameter_93, False, False) @@ -3480,10 +3379,11 @@ def forward( # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1024xf32) add_167 = paddle._C_ops.add(matmul_146, parameter_92) - del parameter_92 + del matmul_146, parameter_92 # pd_op.reshape: (1x26x16x64xf32) <- (1x26x1024xf32, 4xi64) reshape_73 = paddle._C_ops.reshape(add_166, full_int_array_1) + del add_166 # pd_op.transpose: (1x16x26x64xf32) <- (1x26x16x64xf32) transpose_73 = paddle._C_ops.transpose(reshape_73, [0, 2, 1, 3]) @@ -3491,6 +3391,7 @@ def forward( # pd_op.reshape: (1x26x16x64xf32) <- (1x26x1024xf32, 4xi64) reshape_74 = paddle._C_ops.reshape(add_167, full_int_array_1) + del add_167 # pd_op.transpose: (1x16x26x64xf32) <- (1x26x16x64xf32) transpose_74 = paddle._C_ops.transpose(reshape_74, [0, 2, 1, 3]) @@ -3502,9 +3403,11 @@ def forward( # pd_op.matmul: (1x16x26x26xf32) <- (1x16x26x64xf32, 1x16x26x64xf32) matmul_147 = paddle._C_ops.matmul(scale_20, transpose_73, False, True) + del scale_20, transpose_73 # pd_op.add: (1x16x26x26xf32) <- (1x16x26x26xf32, 1x1x1x26xf32) add_168 = paddle._C_ops.add(matmul_147, unsqueeze_0) + del matmul_147 # pd_op.softmax: (1x16x26x26xf32) <- (1x16x26x26xf32) softmax_18 = paddle._C_ops.softmax(add_168, -1) @@ -3513,13 +3416,15 @@ def forward( # pd_op.dropout: (1x16x26x26xf32, 1x16x26x26xui8) <- (1x16x26x26xf32, None, 1xf32) dropout_110, dropout_111 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_18, None, full_5, False, "upscale_in_train", 0, False + softmax_18, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_18 # pd_op.matmul: (1x16x26x64xf32) <- (1x16x26x26xf32, 1x16x26x64xf32) matmul_148 = paddle._C_ops.matmul(dropout_110, transpose_74, False, False) + del dropout_110, transpose_74 # pd_op.transpose: (1x26x16x64xf32) <- (1x16x26x64xf32) transpose_75 = paddle._C_ops.transpose(matmul_148, [0, 2, 1, 3]) @@ -3527,19 +3432,20 @@ def forward( # pd_op.reshape: (1x26x1024xf32) <- (1x26x16x64xf32, 3xi64) reshape_75 = paddle._C_ops.reshape(transpose_75, full_int_array_2) + del transpose_75 # pd_op.matmul: (1x26x1024xf32) <- (1x26x1024xf32, 1024x1024xf32) matmul_149 = paddle._C_ops.matmul(reshape_75, parameter_91, False, False) - del parameter_91 + del parameter_91, reshape_75 # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1024xf32) add_169 = paddle._C_ops.add(matmul_149, parameter_90) - del parameter_90 + del matmul_149, parameter_90 # pd_op.dropout: (1x26x1024xf32, 1x26x1024xui8) <- (1x26x1024xf32, None, 1xf32) dropout_112, dropout_113 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_169, None, full_5, False, "upscale_in_train", 0, False + add_169, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -3547,6 +3453,7 @@ def forward( # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1x26x1024xf32) add_170 = paddle._C_ops.add(layer_norm_108, dropout_112) + del dropout_112, layer_norm_108 # pd_op.layer_norm: (1x26x1024xf32, 1x26xf32, 1x26xf32) <- (1x26x1024xf32, 1024xf32, 1024xf32) layer_norm_111, layer_norm_112, layer_norm_113 = (lambda x, f: f(x))( @@ -3555,7 +3462,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_84, parameter_85 + del add_170, parameter_84, parameter_85 # pd_op.matmul: (1x26x3072xf32) <- (1x26x1024xf32, 1024x3072xf32) matmul_150 = paddle._C_ops.matmul(layer_norm_111, parameter_89, False, False) @@ -3563,7 +3470,7 @@ def forward( # pd_op.add: (1x26x3072xf32) <- (1x26x3072xf32, 3072xf32) add_171 = paddle._C_ops.add(matmul_150, parameter_88) - del parameter_88 + del matmul_150, parameter_88 # pd_op.relu: (1x26x3072xf32) <- (1x26x3072xf32) relu_18 = paddle._C_ops.relu(add_171) @@ -3571,16 +3478,16 @@ def forward( # pd_op.matmul: (1x26x1024xf32) <- (1x26x3072xf32, 3072x1024xf32) matmul_151 = paddle._C_ops.matmul(relu_18, parameter_87, False, False) - del parameter_87 + del parameter_87, relu_18 # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1024xf32) add_172 = paddle._C_ops.add(matmul_151, parameter_86) - del parameter_86 + del matmul_151, parameter_86 # pd_op.dropout: (1x26x1024xf32, 1x26x1024xui8) <- (1x26x1024xf32, None, 1xf32) dropout_114, dropout_115 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_172, None, full_5, False, "upscale_in_train", 0, False + add_172, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -3588,6 +3495,7 @@ def forward( # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1x26x1024xf32) add_173 = paddle._C_ops.add(layer_norm_111, dropout_114) + del dropout_114, layer_norm_111 # pd_op.layer_norm: (1x26x1024xf32, 1x26xf32, 1x26xf32) <- (1x26x1024xf32, 1024xf32, 1024xf32) layer_norm_114, layer_norm_115, layer_norm_116 = (lambda x, f: f(x))( @@ -3596,7 +3504,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_82, parameter_83 + del add_173, parameter_82, parameter_83 # pd_op.matmul: (1x26x1024xf32) <- (1x26x1024xf32, 1024x1024xf32) matmul_152 = paddle._C_ops.matmul(layer_norm_114, parameter_81, False, False) @@ -3604,10 +3512,11 @@ def forward( # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1024xf32) add_174 = paddle._C_ops.add(matmul_152, parameter_80) - del parameter_80 + del matmul_152, parameter_80 # pd_op.reshape: (1x26x16x64xf32) <- (1x26x1024xf32, 4xi64) reshape_76 = paddle._C_ops.reshape(add_174, full_int_array_1) + del add_174 # pd_op.transpose: (1x16x26x64xf32) <- (1x26x16x64xf32) transpose_76 = paddle._C_ops.transpose(reshape_76, [0, 2, 1, 3]) @@ -3619,7 +3528,7 @@ def forward( # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1024xf32) add_175 = paddle._C_ops.add(matmul_153, parameter_78) - del parameter_78 + del matmul_153, parameter_78 # pd_op.matmul: (1x26x1024xf32) <- (1x26x1024xf32, 1024x1024xf32) matmul_154 = paddle._C_ops.matmul(layer_norm_114, parameter_77, False, False) @@ -3627,10 +3536,11 @@ def forward( # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1024xf32) add_176 = paddle._C_ops.add(matmul_154, parameter_76) - del parameter_76 + del matmul_154, parameter_76 # pd_op.reshape: (1x26x16x64xf32) <- (1x26x1024xf32, 4xi64) reshape_77 = paddle._C_ops.reshape(add_175, full_int_array_1) + del add_175 # pd_op.transpose: (1x16x26x64xf32) <- (1x26x16x64xf32) transpose_77 = paddle._C_ops.transpose(reshape_77, [0, 2, 1, 3]) @@ -3638,6 +3548,7 @@ def forward( # pd_op.reshape: (1x26x16x64xf32) <- (1x26x1024xf32, 4xi64) reshape_78 = paddle._C_ops.reshape(add_176, full_int_array_1) + del add_176 # pd_op.transpose: (1x16x26x64xf32) <- (1x26x16x64xf32) transpose_78 = paddle._C_ops.transpose(reshape_78, [0, 2, 1, 3]) @@ -3649,9 +3560,11 @@ def forward( # pd_op.matmul: (1x16x26x26xf32) <- (1x16x26x64xf32, 1x16x26x64xf32) matmul_155 = paddle._C_ops.matmul(scale_21, transpose_77, False, True) + del scale_21, transpose_77 # pd_op.add: (1x16x26x26xf32) <- (1x16x26x26xf32, 1x1x1x26xf32) add_177 = paddle._C_ops.add(matmul_155, unsqueeze_0) + del matmul_155 # pd_op.softmax: (1x16x26x26xf32) <- (1x16x26x26xf32) softmax_19 = paddle._C_ops.softmax(add_177, -1) @@ -3660,13 +3573,15 @@ def forward( # pd_op.dropout: (1x16x26x26xf32, 1x16x26x26xui8) <- (1x16x26x26xf32, None, 1xf32) dropout_116, dropout_117 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_19, None, full_5, False, "upscale_in_train", 0, False + softmax_19, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_19 # pd_op.matmul: (1x16x26x64xf32) <- (1x16x26x26xf32, 1x16x26x64xf32) matmul_156 = paddle._C_ops.matmul(dropout_116, transpose_78, False, False) + del dropout_116, transpose_78 # pd_op.transpose: (1x26x16x64xf32) <- (1x16x26x64xf32) transpose_79 = paddle._C_ops.transpose(matmul_156, [0, 2, 1, 3]) @@ -3674,19 +3589,20 @@ def forward( # pd_op.reshape: (1x26x1024xf32) <- (1x26x16x64xf32, 3xi64) reshape_79 = paddle._C_ops.reshape(transpose_79, full_int_array_2) + del transpose_79 # pd_op.matmul: (1x26x1024xf32) <- (1x26x1024xf32, 1024x1024xf32) matmul_157 = paddle._C_ops.matmul(reshape_79, parameter_75, False, False) - del parameter_75 + del parameter_75, reshape_79 # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1024xf32) add_178 = paddle._C_ops.add(matmul_157, parameter_74) - del parameter_74 + del matmul_157, parameter_74 # pd_op.dropout: (1x26x1024xf32, 1x26x1024xui8) <- (1x26x1024xf32, None, 1xf32) dropout_118, dropout_119 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_178, None, full_5, False, "upscale_in_train", 0, False + add_178, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -3694,6 +3610,7 @@ def forward( # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1x26x1024xf32) add_179 = paddle._C_ops.add(layer_norm_114, dropout_118) + del dropout_118, layer_norm_114 # pd_op.layer_norm: (1x26x1024xf32, 1x26xf32, 1x26xf32) <- (1x26x1024xf32, 1024xf32, 1024xf32) layer_norm_117, layer_norm_118, layer_norm_119 = (lambda x, f: f(x))( @@ -3702,7 +3619,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_68, parameter_69 + del add_179, parameter_68, parameter_69 # pd_op.matmul: (1x26x3072xf32) <- (1x26x1024xf32, 1024x3072xf32) matmul_158 = paddle._C_ops.matmul(layer_norm_117, parameter_73, False, False) @@ -3710,7 +3627,7 @@ def forward( # pd_op.add: (1x26x3072xf32) <- (1x26x3072xf32, 3072xf32) add_180 = paddle._C_ops.add(matmul_158, parameter_72) - del parameter_72 + del matmul_158, parameter_72 # pd_op.relu: (1x26x3072xf32) <- (1x26x3072xf32) relu_19 = paddle._C_ops.relu(add_180) @@ -3718,16 +3635,16 @@ def forward( # pd_op.matmul: (1x26x1024xf32) <- (1x26x3072xf32, 3072x1024xf32) matmul_159 = paddle._C_ops.matmul(relu_19, parameter_71, False, False) - del parameter_71 + del parameter_71, relu_19 # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1024xf32) add_181 = paddle._C_ops.add(matmul_159, parameter_70) - del parameter_70 + del matmul_159, parameter_70 # pd_op.dropout: (1x26x1024xf32, 1x26x1024xui8) <- (1x26x1024xf32, None, 1xf32) dropout_120, dropout_121 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_181, None, full_5, False, "upscale_in_train", 0, False + add_181, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -3735,6 +3652,7 @@ def forward( # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1x26x1024xf32) add_182 = paddle._C_ops.add(layer_norm_117, dropout_120) + del dropout_120, layer_norm_117 # pd_op.layer_norm: (1x26x1024xf32, 1x26xf32, 1x26xf32) <- (1x26x1024xf32, 1024xf32, 1024xf32) layer_norm_120, layer_norm_121, layer_norm_122 = (lambda x, f: f(x))( @@ -3743,7 +3661,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_66, parameter_67 + del add_182, parameter_66, parameter_67 # pd_op.matmul: (1x26x1024xf32) <- (1x26x1024xf32, 1024x1024xf32) matmul_160 = paddle._C_ops.matmul(layer_norm_120, parameter_65, False, False) @@ -3751,10 +3669,11 @@ def forward( # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1024xf32) add_183 = paddle._C_ops.add(matmul_160, parameter_64) - del parameter_64 + del matmul_160, parameter_64 # pd_op.reshape: (1x26x16x64xf32) <- (1x26x1024xf32, 4xi64) reshape_80 = paddle._C_ops.reshape(add_183, full_int_array_1) + del add_183 # pd_op.transpose: (1x16x26x64xf32) <- (1x26x16x64xf32) transpose_80 = paddle._C_ops.transpose(reshape_80, [0, 2, 1, 3]) @@ -3766,7 +3685,7 @@ def forward( # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1024xf32) add_184 = paddle._C_ops.add(matmul_161, parameter_62) - del parameter_62 + del matmul_161, parameter_62 # pd_op.matmul: (1x26x1024xf32) <- (1x26x1024xf32, 1024x1024xf32) matmul_162 = paddle._C_ops.matmul(layer_norm_120, parameter_61, False, False) @@ -3774,10 +3693,11 @@ def forward( # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1024xf32) add_185 = paddle._C_ops.add(matmul_162, parameter_60) - del parameter_60 + del matmul_162, parameter_60 # pd_op.reshape: (1x26x16x64xf32) <- (1x26x1024xf32, 4xi64) reshape_81 = paddle._C_ops.reshape(add_184, full_int_array_1) + del add_184 # pd_op.transpose: (1x16x26x64xf32) <- (1x26x16x64xf32) transpose_81 = paddle._C_ops.transpose(reshape_81, [0, 2, 1, 3]) @@ -3785,6 +3705,7 @@ def forward( # pd_op.reshape: (1x26x16x64xf32) <- (1x26x1024xf32, 4xi64) reshape_82 = paddle._C_ops.reshape(add_185, full_int_array_1) + del add_185 # pd_op.transpose: (1x16x26x64xf32) <- (1x26x16x64xf32) transpose_82 = paddle._C_ops.transpose(reshape_82, [0, 2, 1, 3]) @@ -3796,9 +3717,11 @@ def forward( # pd_op.matmul: (1x16x26x26xf32) <- (1x16x26x64xf32, 1x16x26x64xf32) matmul_163 = paddle._C_ops.matmul(scale_22, transpose_81, False, True) + del scale_22, transpose_81 # pd_op.add: (1x16x26x26xf32) <- (1x16x26x26xf32, 1x1x1x26xf32) add_186 = paddle._C_ops.add(matmul_163, unsqueeze_0) + del matmul_163 # pd_op.softmax: (1x16x26x26xf32) <- (1x16x26x26xf32) softmax_20 = paddle._C_ops.softmax(add_186, -1) @@ -3807,13 +3730,15 @@ def forward( # pd_op.dropout: (1x16x26x26xf32, 1x16x26x26xui8) <- (1x16x26x26xf32, None, 1xf32) dropout_122, dropout_123 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_20, None, full_5, False, "upscale_in_train", 0, False + softmax_20, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_20 # pd_op.matmul: (1x16x26x64xf32) <- (1x16x26x26xf32, 1x16x26x64xf32) matmul_164 = paddle._C_ops.matmul(dropout_122, transpose_82, False, False) + del dropout_122, transpose_82 # pd_op.transpose: (1x26x16x64xf32) <- (1x16x26x64xf32) transpose_83 = paddle._C_ops.transpose(matmul_164, [0, 2, 1, 3]) @@ -3821,19 +3746,20 @@ def forward( # pd_op.reshape: (1x26x1024xf32) <- (1x26x16x64xf32, 3xi64) reshape_83 = paddle._C_ops.reshape(transpose_83, full_int_array_2) + del transpose_83 # pd_op.matmul: (1x26x1024xf32) <- (1x26x1024xf32, 1024x1024xf32) matmul_165 = paddle._C_ops.matmul(reshape_83, parameter_59, False, False) - del parameter_59 + del parameter_59, reshape_83 # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1024xf32) add_187 = paddle._C_ops.add(matmul_165, parameter_58) - del parameter_58 + del matmul_165, parameter_58 # pd_op.dropout: (1x26x1024xf32, 1x26x1024xui8) <- (1x26x1024xf32, None, 1xf32) dropout_124, dropout_125 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_187, None, full_5, False, "upscale_in_train", 0, False + add_187, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -3841,6 +3767,7 @@ def forward( # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1x26x1024xf32) add_188 = paddle._C_ops.add(layer_norm_120, dropout_124) + del dropout_124, layer_norm_120 # pd_op.layer_norm: (1x26x1024xf32, 1x26xf32, 1x26xf32) <- (1x26x1024xf32, 1024xf32, 1024xf32) layer_norm_123, layer_norm_124, layer_norm_125 = (lambda x, f: f(x))( @@ -3849,7 +3776,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_52, parameter_53 + del add_188, parameter_52, parameter_53 # pd_op.matmul: (1x26x3072xf32) <- (1x26x1024xf32, 1024x3072xf32) matmul_166 = paddle._C_ops.matmul(layer_norm_123, parameter_57, False, False) @@ -3857,7 +3784,7 @@ def forward( # pd_op.add: (1x26x3072xf32) <- (1x26x3072xf32, 3072xf32) add_189 = paddle._C_ops.add(matmul_166, parameter_56) - del parameter_56 + del matmul_166, parameter_56 # pd_op.relu: (1x26x3072xf32) <- (1x26x3072xf32) relu_20 = paddle._C_ops.relu(add_189) @@ -3865,16 +3792,16 @@ def forward( # pd_op.matmul: (1x26x1024xf32) <- (1x26x3072xf32, 3072x1024xf32) matmul_167 = paddle._C_ops.matmul(relu_20, parameter_55, False, False) - del parameter_55 + del parameter_55, relu_20 # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1024xf32) add_190 = paddle._C_ops.add(matmul_167, parameter_54) - del parameter_54 + del matmul_167, parameter_54 # pd_op.dropout: (1x26x1024xf32, 1x26x1024xui8) <- (1x26x1024xf32, None, 1xf32) dropout_126, dropout_127 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_190, None, full_5, False, "upscale_in_train", 0, False + add_190, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -3882,6 +3809,7 @@ def forward( # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1x26x1024xf32) add_191 = paddle._C_ops.add(layer_norm_123, dropout_126) + del dropout_126, layer_norm_123 # pd_op.layer_norm: (1x26x1024xf32, 1x26xf32, 1x26xf32) <- (1x26x1024xf32, 1024xf32, 1024xf32) layer_norm_126, layer_norm_127, layer_norm_128 = (lambda x, f: f(x))( @@ -3890,7 +3818,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_50, parameter_51 + del add_191, parameter_50, parameter_51 # pd_op.matmul: (1x26x1024xf32) <- (1x26x1024xf32, 1024x1024xf32) matmul_168 = paddle._C_ops.matmul(layer_norm_126, parameter_49, False, False) @@ -3898,10 +3826,11 @@ def forward( # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1024xf32) add_192 = paddle._C_ops.add(matmul_168, parameter_48) - del parameter_48 + del matmul_168, parameter_48 # pd_op.reshape: (1x26x16x64xf32) <- (1x26x1024xf32, 4xi64) reshape_84 = paddle._C_ops.reshape(add_192, full_int_array_1) + del add_192 # pd_op.transpose: (1x16x26x64xf32) <- (1x26x16x64xf32) transpose_84 = paddle._C_ops.transpose(reshape_84, [0, 2, 1, 3]) @@ -3913,7 +3842,7 @@ def forward( # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1024xf32) add_193 = paddle._C_ops.add(matmul_169, parameter_46) - del parameter_46 + del matmul_169, parameter_46 # pd_op.matmul: (1x26x1024xf32) <- (1x26x1024xf32, 1024x1024xf32) matmul_170 = paddle._C_ops.matmul(layer_norm_126, parameter_45, False, False) @@ -3921,10 +3850,11 @@ def forward( # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1024xf32) add_194 = paddle._C_ops.add(matmul_170, parameter_44) - del parameter_44 + del matmul_170, parameter_44 # pd_op.reshape: (1x26x16x64xf32) <- (1x26x1024xf32, 4xi64) reshape_85 = paddle._C_ops.reshape(add_193, full_int_array_1) + del add_193 # pd_op.transpose: (1x16x26x64xf32) <- (1x26x16x64xf32) transpose_85 = paddle._C_ops.transpose(reshape_85, [0, 2, 1, 3]) @@ -3932,6 +3862,7 @@ def forward( # pd_op.reshape: (1x26x16x64xf32) <- (1x26x1024xf32, 4xi64) reshape_86 = paddle._C_ops.reshape(add_194, full_int_array_1) + del add_194 # pd_op.transpose: (1x16x26x64xf32) <- (1x26x16x64xf32) transpose_86 = paddle._C_ops.transpose(reshape_86, [0, 2, 1, 3]) @@ -3943,9 +3874,11 @@ def forward( # pd_op.matmul: (1x16x26x26xf32) <- (1x16x26x64xf32, 1x16x26x64xf32) matmul_171 = paddle._C_ops.matmul(scale_23, transpose_85, False, True) + del scale_23, transpose_85 # pd_op.add: (1x16x26x26xf32) <- (1x16x26x26xf32, 1x1x1x26xf32) add_195 = paddle._C_ops.add(matmul_171, unsqueeze_0) + del matmul_171 # pd_op.softmax: (1x16x26x26xf32) <- (1x16x26x26xf32) softmax_21 = paddle._C_ops.softmax(add_195, -1) @@ -3954,13 +3887,15 @@ def forward( # pd_op.dropout: (1x16x26x26xf32, 1x16x26x26xui8) <- (1x16x26x26xf32, None, 1xf32) dropout_128, dropout_129 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_21, None, full_5, False, "upscale_in_train", 0, False + softmax_21, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_21 # pd_op.matmul: (1x16x26x64xf32) <- (1x16x26x26xf32, 1x16x26x64xf32) matmul_172 = paddle._C_ops.matmul(dropout_128, transpose_86, False, False) + del dropout_128, transpose_86 # pd_op.transpose: (1x26x16x64xf32) <- (1x16x26x64xf32) transpose_87 = paddle._C_ops.transpose(matmul_172, [0, 2, 1, 3]) @@ -3968,19 +3903,20 @@ def forward( # pd_op.reshape: (1x26x1024xf32) <- (1x26x16x64xf32, 3xi64) reshape_87 = paddle._C_ops.reshape(transpose_87, full_int_array_2) + del transpose_87 # pd_op.matmul: (1x26x1024xf32) <- (1x26x1024xf32, 1024x1024xf32) matmul_173 = paddle._C_ops.matmul(reshape_87, parameter_43, False, False) - del parameter_43 + del parameter_43, reshape_87 # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1024xf32) add_196 = paddle._C_ops.add(matmul_173, parameter_42) - del parameter_42 + del matmul_173, parameter_42 # pd_op.dropout: (1x26x1024xf32, 1x26x1024xui8) <- (1x26x1024xf32, None, 1xf32) dropout_130, dropout_131 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_196, None, full_5, False, "upscale_in_train", 0, False + add_196, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -3988,6 +3924,7 @@ def forward( # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1x26x1024xf32) add_197 = paddle._C_ops.add(layer_norm_126, dropout_130) + del dropout_130, layer_norm_126 # pd_op.layer_norm: (1x26x1024xf32, 1x26xf32, 1x26xf32) <- (1x26x1024xf32, 1024xf32, 1024xf32) layer_norm_129, layer_norm_130, layer_norm_131 = (lambda x, f: f(x))( @@ -3996,7 +3933,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_36, parameter_37 + del add_197, parameter_36, parameter_37 # pd_op.matmul: (1x26x3072xf32) <- (1x26x1024xf32, 1024x3072xf32) matmul_174 = paddle._C_ops.matmul(layer_norm_129, parameter_41, False, False) @@ -4004,7 +3941,7 @@ def forward( # pd_op.add: (1x26x3072xf32) <- (1x26x3072xf32, 3072xf32) add_198 = paddle._C_ops.add(matmul_174, parameter_40) - del parameter_40 + del matmul_174, parameter_40 # pd_op.relu: (1x26x3072xf32) <- (1x26x3072xf32) relu_21 = paddle._C_ops.relu(add_198) @@ -4012,16 +3949,16 @@ def forward( # pd_op.matmul: (1x26x1024xf32) <- (1x26x3072xf32, 3072x1024xf32) matmul_175 = paddle._C_ops.matmul(relu_21, parameter_39, False, False) - del parameter_39 + del parameter_39, relu_21 # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1024xf32) add_199 = paddle._C_ops.add(matmul_175, parameter_38) - del parameter_38 + del matmul_175, parameter_38 # pd_op.dropout: (1x26x1024xf32, 1x26x1024xui8) <- (1x26x1024xf32, None, 1xf32) dropout_132, dropout_133 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_199, None, full_5, False, "upscale_in_train", 0, False + add_199, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -4029,6 +3966,7 @@ def forward( # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1x26x1024xf32) add_200 = paddle._C_ops.add(layer_norm_129, dropout_132) + del dropout_132, layer_norm_129 # pd_op.layer_norm: (1x26x1024xf32, 1x26xf32, 1x26xf32) <- (1x26x1024xf32, 1024xf32, 1024xf32) layer_norm_132, layer_norm_133, layer_norm_134 = (lambda x, f: f(x))( @@ -4037,7 +3975,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_34, parameter_35 + del add_200, parameter_34, parameter_35 # pd_op.matmul: (1x26x1024xf32) <- (1x26x1024xf32, 1024x1024xf32) matmul_176 = paddle._C_ops.matmul(layer_norm_132, parameter_33, False, False) @@ -4045,10 +3983,11 @@ def forward( # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1024xf32) add_201 = paddle._C_ops.add(matmul_176, parameter_32) - del parameter_32 + del matmul_176, parameter_32 # pd_op.reshape: (1x26x16x64xf32) <- (1x26x1024xf32, 4xi64) reshape_88 = paddle._C_ops.reshape(add_201, full_int_array_1) + del add_201 # pd_op.transpose: (1x16x26x64xf32) <- (1x26x16x64xf32) transpose_88 = paddle._C_ops.transpose(reshape_88, [0, 2, 1, 3]) @@ -4060,7 +3999,7 @@ def forward( # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1024xf32) add_202 = paddle._C_ops.add(matmul_177, parameter_30) - del parameter_30 + del matmul_177, parameter_30 # pd_op.matmul: (1x26x1024xf32) <- (1x26x1024xf32, 1024x1024xf32) matmul_178 = paddle._C_ops.matmul(layer_norm_132, parameter_29, False, False) @@ -4068,10 +4007,11 @@ def forward( # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1024xf32) add_203 = paddle._C_ops.add(matmul_178, parameter_28) - del parameter_28 + del matmul_178, parameter_28 # pd_op.reshape: (1x26x16x64xf32) <- (1x26x1024xf32, 4xi64) reshape_89 = paddle._C_ops.reshape(add_202, full_int_array_1) + del add_202 # pd_op.transpose: (1x16x26x64xf32) <- (1x26x16x64xf32) transpose_89 = paddle._C_ops.transpose(reshape_89, [0, 2, 1, 3]) @@ -4079,6 +4019,7 @@ def forward( # pd_op.reshape: (1x26x16x64xf32) <- (1x26x1024xf32, 4xi64) reshape_90 = paddle._C_ops.reshape(add_203, full_int_array_1) + del add_203 # pd_op.transpose: (1x16x26x64xf32) <- (1x26x16x64xf32) transpose_90 = paddle._C_ops.transpose(reshape_90, [0, 2, 1, 3]) @@ -4090,9 +4031,11 @@ def forward( # pd_op.matmul: (1x16x26x26xf32) <- (1x16x26x64xf32, 1x16x26x64xf32) matmul_179 = paddle._C_ops.matmul(scale_24, transpose_89, False, True) + del scale_24, transpose_89 # pd_op.add: (1x16x26x26xf32) <- (1x16x26x26xf32, 1x1x1x26xf32) add_204 = paddle._C_ops.add(matmul_179, unsqueeze_0) + del matmul_179 # pd_op.softmax: (1x16x26x26xf32) <- (1x16x26x26xf32) softmax_22 = paddle._C_ops.softmax(add_204, -1) @@ -4101,13 +4044,15 @@ def forward( # pd_op.dropout: (1x16x26x26xf32, 1x16x26x26xui8) <- (1x16x26x26xf32, None, 1xf32) dropout_134, dropout_135 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_22, None, full_5, False, "upscale_in_train", 0, False + softmax_22, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_22 # pd_op.matmul: (1x16x26x64xf32) <- (1x16x26x26xf32, 1x16x26x64xf32) matmul_180 = paddle._C_ops.matmul(dropout_134, transpose_90, False, False) + del dropout_134, transpose_90 # pd_op.transpose: (1x26x16x64xf32) <- (1x16x26x64xf32) transpose_91 = paddle._C_ops.transpose(matmul_180, [0, 2, 1, 3]) @@ -4115,19 +4060,20 @@ def forward( # pd_op.reshape: (1x26x1024xf32) <- (1x26x16x64xf32, 3xi64) reshape_91 = paddle._C_ops.reshape(transpose_91, full_int_array_2) + del transpose_91 # pd_op.matmul: (1x26x1024xf32) <- (1x26x1024xf32, 1024x1024xf32) matmul_181 = paddle._C_ops.matmul(reshape_91, parameter_27, False, False) - del parameter_27 + del parameter_27, reshape_91 # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1024xf32) add_205 = paddle._C_ops.add(matmul_181, parameter_26) - del parameter_26 + del matmul_181, parameter_26 # pd_op.dropout: (1x26x1024xf32, 1x26x1024xui8) <- (1x26x1024xf32, None, 1xf32) dropout_136, dropout_137 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_205, None, full_5, False, "upscale_in_train", 0, False + add_205, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -4135,6 +4081,7 @@ def forward( # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1x26x1024xf32) add_206 = paddle._C_ops.add(layer_norm_132, dropout_136) + del dropout_136, layer_norm_132 # pd_op.layer_norm: (1x26x1024xf32, 1x26xf32, 1x26xf32) <- (1x26x1024xf32, 1024xf32, 1024xf32) layer_norm_135, layer_norm_136, layer_norm_137 = (lambda x, f: f(x))( @@ -4143,7 +4090,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_20, parameter_21 + del add_206, parameter_20, parameter_21 # pd_op.matmul: (1x26x3072xf32) <- (1x26x1024xf32, 1024x3072xf32) matmul_182 = paddle._C_ops.matmul(layer_norm_135, parameter_25, False, False) @@ -4151,7 +4098,7 @@ def forward( # pd_op.add: (1x26x3072xf32) <- (1x26x3072xf32, 3072xf32) add_207 = paddle._C_ops.add(matmul_182, parameter_24) - del parameter_24 + del matmul_182, parameter_24 # pd_op.relu: (1x26x3072xf32) <- (1x26x3072xf32) relu_22 = paddle._C_ops.relu(add_207) @@ -4159,16 +4106,16 @@ def forward( # pd_op.matmul: (1x26x1024xf32) <- (1x26x3072xf32, 3072x1024xf32) matmul_183 = paddle._C_ops.matmul(relu_22, parameter_23, False, False) - del parameter_23 + del parameter_23, relu_22 # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1024xf32) add_208 = paddle._C_ops.add(matmul_183, parameter_22) - del parameter_22 + del matmul_183, parameter_22 # pd_op.dropout: (1x26x1024xf32, 1x26x1024xui8) <- (1x26x1024xf32, None, 1xf32) dropout_138, dropout_139 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_208, None, full_5, False, "upscale_in_train", 0, False + add_208, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -4176,6 +4123,7 @@ def forward( # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1x26x1024xf32) add_209 = paddle._C_ops.add(layer_norm_135, dropout_138) + del dropout_138, layer_norm_135 # pd_op.layer_norm: (1x26x1024xf32, 1x26xf32, 1x26xf32) <- (1x26x1024xf32, 1024xf32, 1024xf32) layer_norm_138, layer_norm_139, layer_norm_140 = (lambda x, f: f(x))( @@ -4184,7 +4132,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_18, parameter_19 + del add_209, parameter_18, parameter_19 # pd_op.matmul: (1x26x1024xf32) <- (1x26x1024xf32, 1024x1024xf32) matmul_184 = paddle._C_ops.matmul(layer_norm_138, parameter_17, False, False) @@ -4192,10 +4140,11 @@ def forward( # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1024xf32) add_210 = paddle._C_ops.add(matmul_184, parameter_16) - del parameter_16 + del matmul_184, parameter_16 # pd_op.reshape: (1x26x16x64xf32) <- (1x26x1024xf32, 4xi64) reshape_92 = paddle._C_ops.reshape(add_210, full_int_array_1) + del add_210 # pd_op.transpose: (1x16x26x64xf32) <- (1x26x16x64xf32) transpose_92 = paddle._C_ops.transpose(reshape_92, [0, 2, 1, 3]) @@ -4207,7 +4156,7 @@ def forward( # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1024xf32) add_211 = paddle._C_ops.add(matmul_185, parameter_14) - del parameter_14 + del matmul_185, parameter_14 # pd_op.matmul: (1x26x1024xf32) <- (1x26x1024xf32, 1024x1024xf32) matmul_186 = paddle._C_ops.matmul(layer_norm_138, parameter_13, False, False) @@ -4215,10 +4164,11 @@ def forward( # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1024xf32) add_212 = paddle._C_ops.add(matmul_186, parameter_12) - del parameter_12 + del matmul_186, parameter_12 # pd_op.reshape: (1x26x16x64xf32) <- (1x26x1024xf32, 4xi64) reshape_93 = paddle._C_ops.reshape(add_211, full_int_array_1) + del add_211 # pd_op.transpose: (1x16x26x64xf32) <- (1x26x16x64xf32) transpose_93 = paddle._C_ops.transpose(reshape_93, [0, 2, 1, 3]) @@ -4226,7 +4176,7 @@ def forward( # pd_op.reshape: (1x26x16x64xf32) <- (1x26x1024xf32, 4xi64) reshape_94 = paddle._C_ops.reshape(add_212, full_int_array_1) - del full_int_array_1 + del add_212, full_int_array_1 # pd_op.transpose: (1x16x26x64xf32) <- (1x26x16x64xf32) transpose_94 = paddle._C_ops.transpose(reshape_94, [0, 2, 1, 3]) @@ -4234,13 +4184,15 @@ def forward( # pd_op.scale: (1x16x26x64xf32) <- (1x16x26x64xf32, 1xf32) scale_25 = paddle._C_ops.scale(transpose_92, full_6, float("0"), True) - del transpose_92 + del full_6, transpose_92 # pd_op.matmul: (1x16x26x26xf32) <- (1x16x26x64xf32, 1x16x26x64xf32) matmul_187 = paddle._C_ops.matmul(scale_25, transpose_93, False, True) + del scale_25, transpose_93 # pd_op.add: (1x16x26x26xf32) <- (1x16x26x26xf32, 1x1x1x26xf32) add_213 = paddle._C_ops.add(matmul_187, unsqueeze_0) + del matmul_187, unsqueeze_0 # pd_op.softmax: (1x16x26x26xf32) <- (1x16x26x26xf32) softmax_23 = paddle._C_ops.softmax(add_213, -1) @@ -4249,13 +4201,15 @@ def forward( # pd_op.dropout: (1x16x26x26xf32, 1x16x26x26xui8) <- (1x16x26x26xf32, None, 1xf32) dropout_140, dropout_141 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_23, None, full_5, False, "upscale_in_train", 0, False + softmax_23, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_23 # pd_op.matmul: (1x16x26x64xf32) <- (1x16x26x26xf32, 1x16x26x64xf32) matmul_188 = paddle._C_ops.matmul(dropout_140, transpose_94, False, False) + del dropout_140, transpose_94 # pd_op.transpose: (1x26x16x64xf32) <- (1x16x26x64xf32) transpose_95 = paddle._C_ops.transpose(matmul_188, [0, 2, 1, 3]) @@ -4263,20 +4217,20 @@ def forward( # pd_op.reshape: (1x26x1024xf32) <- (1x26x16x64xf32, 3xi64) reshape_95 = paddle._C_ops.reshape(transpose_95, full_int_array_2) - del full_int_array_2 + del full_int_array_2, transpose_95 # pd_op.matmul: (1x26x1024xf32) <- (1x26x1024xf32, 1024x1024xf32) matmul_189 = paddle._C_ops.matmul(reshape_95, parameter_11, False, False) - del parameter_11 + del parameter_11, reshape_95 # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1024xf32) add_214 = paddle._C_ops.add(matmul_189, parameter_10) - del parameter_10 + del matmul_189, parameter_10 # pd_op.dropout: (1x26x1024xf32, 1x26x1024xui8) <- (1x26x1024xf32, None, 1xf32) dropout_142, dropout_143 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_214, None, full_5, False, "upscale_in_train", 0, False + add_214, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -4284,6 +4238,7 @@ def forward( # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1x26x1024xf32) add_215 = paddle._C_ops.add(layer_norm_138, dropout_142) + del dropout_142, layer_norm_138 # pd_op.layer_norm: (1x26x1024xf32, 1x26xf32, 1x26xf32) <- (1x26x1024xf32, 1024xf32, 1024xf32) layer_norm_141, layer_norm_142, layer_norm_143 = (lambda x, f: f(x))( @@ -4292,7 +4247,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_4, parameter_5 + del add_215, parameter_4, parameter_5 # pd_op.matmul: (1x26x3072xf32) <- (1x26x1024xf32, 1024x3072xf32) matmul_190 = paddle._C_ops.matmul(layer_norm_141, parameter_9, False, False) @@ -4300,7 +4255,7 @@ def forward( # pd_op.add: (1x26x3072xf32) <- (1x26x3072xf32, 3072xf32) add_216 = paddle._C_ops.add(matmul_190, parameter_8) - del parameter_8 + del matmul_190, parameter_8 # pd_op.relu: (1x26x3072xf32) <- (1x26x3072xf32) relu_23 = paddle._C_ops.relu(add_216) @@ -4308,23 +4263,24 @@ def forward( # pd_op.matmul: (1x26x1024xf32) <- (1x26x3072xf32, 3072x1024xf32) matmul_191 = paddle._C_ops.matmul(relu_23, parameter_7, False, False) - del parameter_7 + del parameter_7, relu_23 # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1024xf32) add_217 = paddle._C_ops.add(matmul_191, parameter_6) - del parameter_6 + del matmul_191, parameter_6 # pd_op.dropout: (1x26x1024xf32, 1x26x1024xui8) <- (1x26x1024xf32, None, 1xf32) dropout_144, dropout_145 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_217, None, full_5, False, "upscale_in_train", 0, False + add_217, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) - del add_217 + del add_217, full_5 # pd_op.add: (1x26x1024xf32) <- (1x26x1024xf32, 1x26x1024xf32) add_218 = paddle._C_ops.add(layer_norm_141, dropout_144) + del dropout_144, layer_norm_141 # pd_op.layer_norm: (1x26x1024xf32, 1x26xf32, 1x26xf32) <- (1x26x1024xf32, 1024xf32, 1024xf32) layer_norm_144, layer_norm_145, layer_norm_146 = (lambda x, f: f(x))( @@ -4333,7 +4289,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_2, parameter_3 + del add_218, parameter_2, parameter_3 # pd_op.full_int_array: (1xi64) <- () full_int_array_3 = [0] @@ -4345,878 +4301,18 @@ def forward( slice_0 = paddle._C_ops.slice( layer_norm_144, [1], full_int_array_3, full_int_array_4, [1], [1] ) + del full_int_array_3, full_int_array_4, layer_norm_144 # pd_op.matmul: (1x1024xf32) <- (1x1024xf32, 1024x1024xf32) matmul_192 = paddle._C_ops.matmul(slice_0, parameter_1, False, False) - del parameter_1 + del parameter_1, slice_0 # pd_op.add: (1x1024xf32) <- (1x1024xf32, 1024xf32) add_219 = paddle._C_ops.add(matmul_192, parameter_0) - del parameter_0 + del matmul_192, parameter_0 # pd_op.tanh: (1x1024xf32) <- (1x1024xf32) tanh_0 = paddle._C_ops.tanh(add_219) - del ( - add_0, - add_1, - add_101, - add_102, - add_103, - add_104, - add_107, - add_11, - add_110, - add_111, - add_112, - add_113, - add_116, - add_119, - add_12, - add_120, - add_121, - add_122, - add_125, - add_128, - add_129, - add_13, - add_130, - add_131, - add_134, - add_137, - add_138, - add_139, - add_14, - add_140, - add_143, - add_146, - add_147, - add_148, - add_149, - add_152, - add_155, - add_156, - add_157, - add_158, - add_161, - add_164, - add_165, - add_166, - add_167, - add_17, - add_170, - add_173, - add_174, - add_175, - add_176, - add_179, - add_182, - add_183, - add_184, - add_185, - add_188, - add_191, - add_192, - add_193, - add_194, - add_197, - add_2, - add_20, - add_200, - add_201, - add_202, - add_203, - add_206, - add_209, - add_21, - add_210, - add_211, - add_212, - add_215, - add_218, - add_219, - add_22, - add_23, - add_26, - add_29, - add_3, - add_30, - add_31, - add_32, - add_35, - add_38, - add_39, - add_4, - add_40, - add_41, - add_44, - add_47, - add_48, - add_49, - add_5, - add_50, - add_53, - add_56, - add_57, - add_58, - add_59, - add_62, - add_65, - add_66, - add_67, - add_68, - add_71, - add_74, - add_75, - add_76, - add_77, - add_8, - add_80, - add_83, - add_84, - add_85, - add_86, - add_89, - add_92, - add_93, - add_94, - add_95, - add_98, - assign_0, - assign_1, - assign_10, - assign_11, - assign_12, - assign_13, - assign_14, - assign_15, - assign_16, - assign_17, - assign_18, - assign_19, - assign_2, - assign_20, - assign_21, - assign_22, - assign_23, - assign_24, - assign_25, - assign_26, - assign_27, - assign_28, - assign_29, - assign_3, - assign_30, - assign_31, - assign_32, - assign_33, - assign_34, - assign_35, - assign_36, - assign_37, - assign_38, - assign_39, - assign_4, - assign_40, - assign_41, - assign_42, - assign_43, - assign_44, - assign_45, - assign_46, - assign_47, - assign_48, - assign_49, - assign_5, - assign_50, - assign_51, - assign_52, - assign_53, - assign_54, - assign_55, - assign_56, - assign_57, - assign_58, - assign_59, - assign_6, - assign_60, - assign_61, - assign_62, - assign_63, - assign_64, - assign_65, - assign_66, - assign_67, - assign_68, - assign_69, - assign_7, - assign_70, - assign_71, - assign_72, - assign_73, - assign_74, - assign_75, - assign_76, - assign_77, - assign_78, - assign_79, - assign_8, - assign_80, - assign_81, - assign_82, - assign_83, - assign_84, - assign_85, - assign_86, - assign_87, - assign_88, - assign_89, - assign_9, - assign_90, - assign_91, - assign_92, - assign_93, - assign_94, - dropout_0, - dropout_1, - dropout_10, - dropout_100, - dropout_101, - dropout_102, - dropout_103, - dropout_104, - dropout_105, - dropout_106, - dropout_107, - dropout_108, - dropout_109, - dropout_11, - dropout_110, - dropout_111, - dropout_112, - dropout_113, - dropout_114, - dropout_115, - dropout_116, - dropout_117, - dropout_118, - dropout_119, - dropout_12, - dropout_120, - dropout_121, - dropout_122, - dropout_123, - dropout_124, - dropout_125, - dropout_126, - dropout_127, - dropout_128, - dropout_129, - dropout_13, - dropout_130, - dropout_131, - dropout_132, - dropout_133, - dropout_134, - dropout_135, - dropout_136, - dropout_137, - dropout_138, - dropout_139, - dropout_14, - dropout_140, - dropout_141, - dropout_142, - dropout_143, - dropout_144, - dropout_145, - dropout_15, - dropout_16, - dropout_17, - dropout_18, - dropout_19, - dropout_2, - dropout_20, - dropout_21, - dropout_22, - dropout_23, - dropout_24, - dropout_25, - dropout_26, - dropout_27, - dropout_28, - dropout_29, - dropout_3, - dropout_30, - dropout_31, - dropout_32, - dropout_33, - dropout_34, - dropout_35, - dropout_36, - dropout_37, - dropout_38, - dropout_39, - dropout_4, - dropout_40, - dropout_41, - dropout_42, - dropout_43, - dropout_44, - dropout_45, - dropout_46, - dropout_47, - dropout_48, - dropout_49, - dropout_5, - dropout_50, - dropout_51, - dropout_52, - dropout_53, - dropout_54, - dropout_55, - dropout_56, - dropout_57, - dropout_58, - dropout_59, - dropout_6, - dropout_60, - dropout_61, - dropout_62, - dropout_63, - dropout_64, - dropout_65, - dropout_66, - dropout_67, - dropout_68, - dropout_69, - dropout_7, - dropout_70, - dropout_71, - dropout_72, - dropout_73, - dropout_74, - dropout_75, - dropout_76, - dropout_77, - dropout_78, - dropout_79, - dropout_8, - dropout_80, - dropout_81, - dropout_82, - dropout_83, - dropout_84, - dropout_85, - dropout_86, - dropout_87, - dropout_88, - dropout_89, - dropout_9, - dropout_90, - dropout_91, - dropout_92, - dropout_93, - dropout_94, - dropout_95, - dropout_96, - dropout_97, - dropout_98, - dropout_99, - embedding_0, - embedding_1, - embedding_2, - embedding_3, - full_5, - full_6, - full_int_array_3, - full_int_array_4, - layer_norm_1, - layer_norm_10, - layer_norm_100, - layer_norm_101, - layer_norm_102, - layer_norm_103, - layer_norm_104, - layer_norm_105, - layer_norm_106, - layer_norm_107, - layer_norm_108, - layer_norm_109, - layer_norm_11, - layer_norm_110, - layer_norm_111, - layer_norm_112, - layer_norm_113, - layer_norm_114, - layer_norm_115, - layer_norm_116, - layer_norm_117, - layer_norm_118, - layer_norm_119, - layer_norm_12, - layer_norm_120, - layer_norm_121, - layer_norm_122, - layer_norm_123, - layer_norm_124, - layer_norm_125, - layer_norm_126, - layer_norm_127, - layer_norm_128, - layer_norm_129, - layer_norm_13, - layer_norm_130, - layer_norm_131, - layer_norm_132, - layer_norm_133, - layer_norm_134, - layer_norm_135, - layer_norm_136, - layer_norm_137, - layer_norm_138, - layer_norm_139, - layer_norm_14, - layer_norm_140, - layer_norm_141, - layer_norm_142, - layer_norm_143, - layer_norm_144, - layer_norm_145, - layer_norm_146, - layer_norm_15, - layer_norm_16, - layer_norm_17, - layer_norm_18, - layer_norm_19, - layer_norm_2, - layer_norm_20, - layer_norm_21, - layer_norm_22, - layer_norm_23, - layer_norm_24, - layer_norm_25, - layer_norm_26, - layer_norm_27, - layer_norm_28, - layer_norm_29, - layer_norm_3, - layer_norm_30, - layer_norm_31, - layer_norm_32, - layer_norm_33, - layer_norm_34, - layer_norm_35, - layer_norm_36, - layer_norm_37, - layer_norm_38, - layer_norm_39, - layer_norm_4, - layer_norm_40, - layer_norm_41, - layer_norm_42, - layer_norm_43, - layer_norm_44, - layer_norm_45, - layer_norm_46, - layer_norm_47, - layer_norm_48, - layer_norm_49, - layer_norm_5, - layer_norm_50, - layer_norm_51, - layer_norm_52, - layer_norm_53, - layer_norm_54, - layer_norm_55, - layer_norm_56, - layer_norm_57, - layer_norm_58, - layer_norm_59, - layer_norm_6, - layer_norm_60, - layer_norm_61, - layer_norm_62, - layer_norm_63, - layer_norm_64, - layer_norm_65, - layer_norm_66, - layer_norm_67, - layer_norm_68, - layer_norm_69, - layer_norm_7, - layer_norm_70, - layer_norm_71, - layer_norm_72, - layer_norm_73, - layer_norm_74, - layer_norm_75, - layer_norm_76, - layer_norm_77, - layer_norm_78, - layer_norm_79, - layer_norm_8, - layer_norm_80, - layer_norm_81, - layer_norm_82, - layer_norm_83, - layer_norm_84, - layer_norm_85, - layer_norm_86, - layer_norm_87, - layer_norm_88, - layer_norm_89, - layer_norm_9, - layer_norm_90, - layer_norm_91, - layer_norm_92, - layer_norm_93, - layer_norm_94, - layer_norm_95, - layer_norm_96, - layer_norm_97, - layer_norm_98, - layer_norm_99, - matmul_0, - matmul_1, - matmul_10, - matmul_101, - matmul_102, - matmul_103, - matmul_104, - matmul_105, - matmul_106, - matmul_107, - matmul_109, - matmul_11, - matmul_110, - matmul_111, - matmul_112, - matmul_113, - matmul_114, - matmul_115, - matmul_117, - matmul_118, - matmul_119, - matmul_120, - matmul_121, - matmul_122, - matmul_123, - matmul_125, - matmul_126, - matmul_127, - matmul_128, - matmul_129, - matmul_13, - matmul_130, - matmul_131, - matmul_133, - matmul_134, - matmul_135, - matmul_136, - matmul_137, - matmul_138, - matmul_139, - matmul_14, - matmul_141, - matmul_142, - matmul_143, - matmul_144, - matmul_145, - matmul_146, - matmul_147, - matmul_149, - matmul_15, - matmul_150, - matmul_151, - matmul_152, - matmul_153, - matmul_154, - matmul_155, - matmul_157, - matmul_158, - matmul_159, - matmul_16, - matmul_160, - matmul_161, - matmul_162, - matmul_163, - matmul_165, - matmul_166, - matmul_167, - matmul_168, - matmul_169, - matmul_17, - matmul_170, - matmul_171, - matmul_173, - matmul_174, - matmul_175, - matmul_176, - matmul_177, - matmul_178, - matmul_179, - matmul_18, - matmul_181, - matmul_182, - matmul_183, - matmul_184, - matmul_185, - matmul_186, - matmul_187, - matmul_189, - matmul_19, - matmul_190, - matmul_191, - matmul_192, - matmul_2, - matmul_21, - matmul_22, - matmul_23, - matmul_24, - matmul_25, - matmul_26, - matmul_27, - matmul_29, - matmul_3, - matmul_30, - matmul_31, - matmul_32, - matmul_33, - matmul_34, - matmul_35, - matmul_37, - matmul_38, - matmul_39, - matmul_40, - matmul_41, - matmul_42, - matmul_43, - matmul_45, - matmul_46, - matmul_47, - matmul_48, - matmul_49, - matmul_5, - matmul_50, - matmul_51, - matmul_53, - matmul_54, - matmul_55, - matmul_56, - matmul_57, - matmul_58, - matmul_59, - matmul_6, - matmul_61, - matmul_62, - matmul_63, - matmul_64, - matmul_65, - matmul_66, - matmul_67, - matmul_69, - matmul_7, - matmul_70, - matmul_71, - matmul_72, - matmul_73, - matmul_74, - matmul_75, - matmul_77, - matmul_78, - matmul_79, - matmul_8, - matmul_80, - matmul_81, - matmul_82, - matmul_83, - matmul_85, - matmul_86, - matmul_87, - matmul_88, - matmul_89, - matmul_9, - matmul_90, - matmul_91, - matmul_93, - matmul_94, - matmul_95, - matmul_96, - matmul_97, - matmul_98, - matmul_99, - relu_0, - relu_1, - relu_10, - relu_11, - relu_12, - relu_13, - relu_14, - relu_15, - relu_16, - relu_17, - relu_18, - relu_19, - relu_2, - relu_20, - relu_21, - relu_22, - relu_23, - relu_3, - relu_4, - relu_5, - relu_6, - relu_7, - relu_8, - relu_9, - reshape_11, - reshape_15, - reshape_19, - reshape_23, - reshape_27, - reshape_3, - reshape_31, - reshape_35, - reshape_39, - reshape_43, - reshape_47, - reshape_51, - reshape_55, - reshape_59, - reshape_63, - reshape_67, - reshape_7, - reshape_71, - reshape_75, - reshape_79, - reshape_83, - reshape_87, - reshape_91, - reshape_95, - scale_1, - scale_10, - scale_11, - scale_12, - scale_13, - scale_14, - scale_15, - scale_16, - scale_17, - scale_18, - scale_19, - scale_2, - scale_20, - scale_21, - scale_22, - scale_23, - scale_24, - scale_25, - scale_3, - scale_4, - scale_5, - scale_6, - scale_7, - scale_8, - scale_9, - slice_0, - softmax_0, - softmax_1, - softmax_10, - softmax_11, - softmax_12, - softmax_13, - softmax_14, - softmax_15, - softmax_16, - softmax_17, - softmax_18, - softmax_19, - softmax_2, - softmax_20, - softmax_21, - softmax_22, - softmax_23, - softmax_3, - softmax_4, - softmax_5, - softmax_6, - softmax_7, - softmax_8, - softmax_9, - subtract_0, - transpose_1, - transpose_10, - transpose_11, - transpose_13, - transpose_14, - transpose_15, - transpose_17, - transpose_18, - transpose_19, - transpose_2, - transpose_21, - transpose_22, - transpose_23, - transpose_25, - transpose_26, - transpose_27, - transpose_29, - transpose_3, - transpose_30, - transpose_31, - transpose_33, - transpose_34, - transpose_35, - transpose_37, - transpose_38, - transpose_39, - transpose_41, - transpose_42, - transpose_43, - transpose_45, - transpose_46, - transpose_47, - transpose_49, - transpose_5, - transpose_50, - transpose_51, - transpose_53, - transpose_54, - transpose_55, - transpose_57, - transpose_58, - transpose_59, - transpose_6, - transpose_61, - transpose_62, - transpose_63, - transpose_65, - transpose_66, - transpose_67, - transpose_69, - transpose_7, - transpose_70, - transpose_71, - transpose_73, - transpose_74, - transpose_75, - transpose_77, - transpose_78, - transpose_79, - transpose_81, - transpose_82, - transpose_83, - transpose_85, - transpose_86, - transpose_87, - transpose_89, - transpose_9, - transpose_90, - transpose_91, - transpose_93, - transpose_94, - transpose_95, - unsqueeze_0, - ) + del add_219 return tanh_0 diff --git a/paddle_samples/PaddleNLP/utc-large/weight_meta.py b/paddle_samples/PaddleNLP/utc-large/weight_meta.py index f9cadfb48..a315893c0 100644 --- a/paddle_samples/PaddleNLP/utc-large/weight_meta.py +++ b/paddle_samples/PaddleNLP/utc-large/weight_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_parameter_0: name = "parameter_0" + original_name = "linear_3409.b_0" shape = [1024] dtype = "float32" data = None @@ -7,6 +8,7 @@ class Program_weight_tensor_parameter_0: class Program_weight_tensor_parameter_1: name = "parameter_1" + original_name = "linear_3409.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.0399998") @@ -18,6 +20,7 @@ class Program_weight_tensor_parameter_1: class Program_weight_tensor_parameter_2: name = "parameter_2" + original_name = "layer_norm_1177.b_0" shape = [1024] dtype = "float32" min_val = float("-0.256506") @@ -29,6 +32,7 @@ class Program_weight_tensor_parameter_2: class Program_weight_tensor_parameter_3: name = "parameter_3" + original_name = "layer_norm_1177.w_0" shape = [1024] dtype = "float32" min_val = float("0.712868") @@ -40,6 +44,7 @@ class Program_weight_tensor_parameter_3: class Program_weight_tensor_parameter_4: name = "parameter_4" + original_name = "layer_norm_1176.b_0" shape = [1024] dtype = "float32" min_val = float("-0.531238") @@ -51,6 +56,7 @@ class Program_weight_tensor_parameter_4: class Program_weight_tensor_parameter_5: name = "parameter_5" + original_name = "layer_norm_1176.w_0" shape = [1024] dtype = "float32" min_val = float("0.447839") @@ -62,6 +68,7 @@ class Program_weight_tensor_parameter_5: class Program_weight_tensor_parameter_6: name = "parameter_6" + original_name = "linear_3408.b_0" shape = [1024] dtype = "float32" min_val = float("-0.149872") @@ -73,6 +80,7 @@ class Program_weight_tensor_parameter_6: class Program_weight_tensor_parameter_7: name = "parameter_7" + original_name = "linear_3408.w_0" shape = [3072, 1024] dtype = "float32" min_val = float("-0.489997") @@ -84,6 +92,7 @@ class Program_weight_tensor_parameter_7: class Program_weight_tensor_parameter_8: name = "parameter_8" + original_name = "linear_3407.b_0" shape = [3072] dtype = "float32" min_val = float("-0.480429") @@ -95,6 +104,7 @@ class Program_weight_tensor_parameter_8: class Program_weight_tensor_parameter_9: name = "parameter_9" + original_name = "linear_3407.w_0" shape = [1024, 3072] dtype = "float32" min_val = float("-0.508755") @@ -106,6 +116,7 @@ class Program_weight_tensor_parameter_9: class Program_weight_tensor_parameter_10: name = "parameter_10" + original_name = "linear_3406.b_0" shape = [1024] dtype = "float32" min_val = float("-0.203547") @@ -117,6 +128,7 @@ class Program_weight_tensor_parameter_10: class Program_weight_tensor_parameter_11: name = "parameter_11" + original_name = "linear_3406.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.177663") @@ -128,6 +140,7 @@ class Program_weight_tensor_parameter_11: class Program_weight_tensor_parameter_12: name = "parameter_12" + original_name = "linear_3405.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0927633") @@ -139,6 +152,7 @@ class Program_weight_tensor_parameter_12: class Program_weight_tensor_parameter_13: name = "parameter_13" + original_name = "linear_3405.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.187578") @@ -150,6 +164,7 @@ class Program_weight_tensor_parameter_13: class Program_weight_tensor_parameter_14: name = "parameter_14" + original_name = "linear_3404.b_0" shape = [1024] dtype = "float32" min_val = float("-101.392") @@ -161,6 +176,7 @@ class Program_weight_tensor_parameter_14: class Program_weight_tensor_parameter_15: name = "parameter_15" + original_name = "linear_3404.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.458621") @@ -172,6 +188,7 @@ class Program_weight_tensor_parameter_15: class Program_weight_tensor_parameter_16: name = "parameter_16" + original_name = "linear_3403.b_0" shape = [1024] dtype = "float32" min_val = float("-1.25096") @@ -183,6 +200,7 @@ class Program_weight_tensor_parameter_16: class Program_weight_tensor_parameter_17: name = "parameter_17" + original_name = "linear_3403.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-1.34382") @@ -194,6 +212,7 @@ class Program_weight_tensor_parameter_17: class Program_weight_tensor_parameter_18: name = "parameter_18" + original_name = "layer_norm_1175.b_0" shape = [1024] dtype = "float32" min_val = float("-0.92704") @@ -205,6 +224,7 @@ class Program_weight_tensor_parameter_18: class Program_weight_tensor_parameter_19: name = "parameter_19" + original_name = "layer_norm_1175.w_0" shape = [1024] dtype = "float32" min_val = float("0.18595") @@ -216,6 +236,7 @@ class Program_weight_tensor_parameter_19: class Program_weight_tensor_parameter_20: name = "parameter_20" + original_name = "layer_norm_1174.b_0" shape = [1024] dtype = "float32" min_val = float("-0.49776") @@ -227,6 +248,7 @@ class Program_weight_tensor_parameter_20: class Program_weight_tensor_parameter_21: name = "parameter_21" + original_name = "layer_norm_1174.w_0" shape = [1024] dtype = "float32" min_val = float("0.783923") @@ -238,6 +260,7 @@ class Program_weight_tensor_parameter_21: class Program_weight_tensor_parameter_22: name = "parameter_22" + original_name = "linear_3402.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0793206") @@ -249,6 +272,7 @@ class Program_weight_tensor_parameter_22: class Program_weight_tensor_parameter_23: name = "parameter_23" + original_name = "linear_3402.w_0" shape = [3072, 1024] dtype = "float32" min_val = float("-1.02188") @@ -260,6 +284,7 @@ class Program_weight_tensor_parameter_23: class Program_weight_tensor_parameter_24: name = "parameter_24" + original_name = "linear_3401.b_0" shape = [3072] dtype = "float32" min_val = float("-0.961136") @@ -271,6 +296,7 @@ class Program_weight_tensor_parameter_24: class Program_weight_tensor_parameter_25: name = "parameter_25" + original_name = "linear_3401.w_0" shape = [1024, 3072] dtype = "float32" min_val = float("-0.314021") @@ -282,6 +308,7 @@ class Program_weight_tensor_parameter_25: class Program_weight_tensor_parameter_26: name = "parameter_26" + original_name = "linear_3400.b_0" shape = [1024] dtype = "float32" min_val = float("-0.154093") @@ -293,6 +320,7 @@ class Program_weight_tensor_parameter_26: class Program_weight_tensor_parameter_27: name = "parameter_27" + original_name = "linear_3400.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.227369") @@ -304,6 +332,7 @@ class Program_weight_tensor_parameter_27: class Program_weight_tensor_parameter_28: name = "parameter_28" + original_name = "linear_3399.b_0" shape = [1024] dtype = "float32" min_val = float("-0.317023") @@ -315,6 +344,7 @@ class Program_weight_tensor_parameter_28: class Program_weight_tensor_parameter_29: name = "parameter_29" + original_name = "linear_3399.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.197659") @@ -326,6 +356,7 @@ class Program_weight_tensor_parameter_29: class Program_weight_tensor_parameter_30: name = "parameter_30" + original_name = "linear_3398.b_0" shape = [1024] dtype = "float32" min_val = float("-114.681") @@ -337,6 +368,7 @@ class Program_weight_tensor_parameter_30: class Program_weight_tensor_parameter_31: name = "parameter_31" + original_name = "linear_3398.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.594027") @@ -348,6 +380,7 @@ class Program_weight_tensor_parameter_31: class Program_weight_tensor_parameter_32: name = "parameter_32" + original_name = "linear_3397.b_0" shape = [1024] dtype = "float32" min_val = float("-0.820461") @@ -359,6 +392,7 @@ class Program_weight_tensor_parameter_32: class Program_weight_tensor_parameter_33: name = "parameter_33" + original_name = "linear_3397.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.355582") @@ -370,6 +404,7 @@ class Program_weight_tensor_parameter_33: class Program_weight_tensor_parameter_34: name = "parameter_34" + original_name = "layer_norm_1173.b_0" shape = [1024] dtype = "float32" min_val = float("-0.939664") @@ -381,6 +416,7 @@ class Program_weight_tensor_parameter_34: class Program_weight_tensor_parameter_35: name = "parameter_35" + original_name = "layer_norm_1173.w_0" shape = [1024] dtype = "float32" min_val = float("0.351417") @@ -392,6 +428,7 @@ class Program_weight_tensor_parameter_35: class Program_weight_tensor_parameter_36: name = "parameter_36" + original_name = "layer_norm_1172.b_0" shape = [1024] dtype = "float32" min_val = float("-0.688187") @@ -403,6 +440,7 @@ class Program_weight_tensor_parameter_36: class Program_weight_tensor_parameter_37: name = "parameter_37" + original_name = "layer_norm_1172.w_0" shape = [1024] dtype = "float32" min_val = float("0.775434") @@ -414,6 +452,7 @@ class Program_weight_tensor_parameter_37: class Program_weight_tensor_parameter_38: name = "parameter_38" + original_name = "linear_3396.b_0" shape = [1024] dtype = "float32" min_val = float("-0.334094") @@ -425,6 +464,7 @@ class Program_weight_tensor_parameter_38: class Program_weight_tensor_parameter_39: name = "parameter_39" + original_name = "linear_3396.w_0" shape = [3072, 1024] dtype = "float32" min_val = float("-0.463523") @@ -436,6 +476,7 @@ class Program_weight_tensor_parameter_39: class Program_weight_tensor_parameter_40: name = "parameter_40" + original_name = "linear_3395.b_0" shape = [3072] dtype = "float32" min_val = float("-0.762178") @@ -447,6 +488,7 @@ class Program_weight_tensor_parameter_40: class Program_weight_tensor_parameter_41: name = "parameter_41" + original_name = "linear_3395.w_0" shape = [1024, 3072] dtype = "float32" min_val = float("-1.17341") @@ -458,6 +500,7 @@ class Program_weight_tensor_parameter_41: class Program_weight_tensor_parameter_42: name = "parameter_42" + original_name = "linear_3394.b_0" shape = [1024] dtype = "float32" min_val = float("-0.357007") @@ -469,6 +512,7 @@ class Program_weight_tensor_parameter_42: class Program_weight_tensor_parameter_43: name = "parameter_43" + original_name = "linear_3394.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.186014") @@ -480,6 +524,7 @@ class Program_weight_tensor_parameter_43: class Program_weight_tensor_parameter_44: name = "parameter_44" + original_name = "linear_3393.b_0" shape = [1024] dtype = "float32" min_val = float("-0.615173") @@ -491,6 +536,7 @@ class Program_weight_tensor_parameter_44: class Program_weight_tensor_parameter_45: name = "parameter_45" + original_name = "linear_3393.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.220643") @@ -502,6 +548,7 @@ class Program_weight_tensor_parameter_45: class Program_weight_tensor_parameter_46: name = "parameter_46" + original_name = "linear_3392.b_0" shape = [1024] dtype = "float32" min_val = float("-116.814") @@ -513,6 +560,7 @@ class Program_weight_tensor_parameter_46: class Program_weight_tensor_parameter_47: name = "parameter_47" + original_name = "linear_3392.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.491821") @@ -524,6 +572,7 @@ class Program_weight_tensor_parameter_47: class Program_weight_tensor_parameter_48: name = "parameter_48" + original_name = "linear_3391.b_0" shape = [1024] dtype = "float32" min_val = float("-0.826495") @@ -535,6 +584,7 @@ class Program_weight_tensor_parameter_48: class Program_weight_tensor_parameter_49: name = "parameter_49" + original_name = "linear_3391.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.349877") @@ -546,6 +596,7 @@ class Program_weight_tensor_parameter_49: class Program_weight_tensor_parameter_50: name = "parameter_50" + original_name = "layer_norm_1171.b_0" shape = [1024] dtype = "float32" min_val = float("-0.6453") @@ -557,6 +608,7 @@ class Program_weight_tensor_parameter_50: class Program_weight_tensor_parameter_51: name = "parameter_51" + original_name = "layer_norm_1171.w_0" shape = [1024] dtype = "float32" min_val = float("0.32673") @@ -568,6 +620,7 @@ class Program_weight_tensor_parameter_51: class Program_weight_tensor_parameter_52: name = "parameter_52" + original_name = "layer_norm_1170.b_0" shape = [1024] dtype = "float32" min_val = float("-0.514739") @@ -579,6 +632,7 @@ class Program_weight_tensor_parameter_52: class Program_weight_tensor_parameter_53: name = "parameter_53" + original_name = "layer_norm_1170.w_0" shape = [1024] dtype = "float32" min_val = float("0.822808") @@ -590,6 +644,7 @@ class Program_weight_tensor_parameter_53: class Program_weight_tensor_parameter_54: name = "parameter_54" + original_name = "linear_3390.b_0" shape = [1024] dtype = "float32" min_val = float("-0.247413") @@ -601,6 +656,7 @@ class Program_weight_tensor_parameter_54: class Program_weight_tensor_parameter_55: name = "parameter_55" + original_name = "linear_3390.w_0" shape = [3072, 1024] dtype = "float32" min_val = float("-0.340775") @@ -612,6 +668,7 @@ class Program_weight_tensor_parameter_55: class Program_weight_tensor_parameter_56: name = "parameter_56" + original_name = "linear_3389.b_0" shape = [3072] dtype = "float32" min_val = float("-0.51657") @@ -623,6 +680,7 @@ class Program_weight_tensor_parameter_56: class Program_weight_tensor_parameter_57: name = "parameter_57" + original_name = "linear_3389.w_0" shape = [1024, 3072] dtype = "float32" min_val = float("-0.832333") @@ -634,6 +692,7 @@ class Program_weight_tensor_parameter_57: class Program_weight_tensor_parameter_58: name = "parameter_58" + original_name = "linear_3388.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0847723") @@ -645,6 +704,7 @@ class Program_weight_tensor_parameter_58: class Program_weight_tensor_parameter_59: name = "parameter_59" + original_name = "linear_3388.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.24998") @@ -656,6 +716,7 @@ class Program_weight_tensor_parameter_59: class Program_weight_tensor_parameter_60: name = "parameter_60" + original_name = "linear_3387.b_0" shape = [1024] dtype = "float32" min_val = float("-0.052044") @@ -667,6 +728,7 @@ class Program_weight_tensor_parameter_60: class Program_weight_tensor_parameter_61: name = "parameter_61" + original_name = "linear_3387.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.212692") @@ -678,6 +740,7 @@ class Program_weight_tensor_parameter_61: class Program_weight_tensor_parameter_62: name = "parameter_62" + original_name = "linear_3386.b_0" shape = [1024] dtype = "float32" min_val = float("-97.8228") @@ -689,6 +752,7 @@ class Program_weight_tensor_parameter_62: class Program_weight_tensor_parameter_63: name = "parameter_63" + original_name = "linear_3386.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.323862") @@ -700,6 +764,7 @@ class Program_weight_tensor_parameter_63: class Program_weight_tensor_parameter_64: name = "parameter_64" + original_name = "linear_3385.b_0" shape = [1024] dtype = "float32" min_val = float("-0.933782") @@ -711,6 +776,7 @@ class Program_weight_tensor_parameter_64: class Program_weight_tensor_parameter_65: name = "parameter_65" + original_name = "linear_3385.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.395542") @@ -722,6 +788,7 @@ class Program_weight_tensor_parameter_65: class Program_weight_tensor_parameter_66: name = "parameter_66" + original_name = "layer_norm_1169.b_0" shape = [1024] dtype = "float32" min_val = float("-0.605844") @@ -733,6 +800,7 @@ class Program_weight_tensor_parameter_66: class Program_weight_tensor_parameter_67: name = "parameter_67" + original_name = "layer_norm_1169.w_0" shape = [1024] dtype = "float32" min_val = float("0.317898") @@ -744,6 +812,7 @@ class Program_weight_tensor_parameter_67: class Program_weight_tensor_parameter_68: name = "parameter_68" + original_name = "layer_norm_1168.b_0" shape = [1024] dtype = "float32" min_val = float("-0.537132") @@ -755,6 +824,7 @@ class Program_weight_tensor_parameter_68: class Program_weight_tensor_parameter_69: name = "parameter_69" + original_name = "layer_norm_1168.w_0" shape = [1024] dtype = "float32" min_val = float("0.832432") @@ -766,6 +836,7 @@ class Program_weight_tensor_parameter_69: class Program_weight_tensor_parameter_70: name = "parameter_70" + original_name = "linear_3384.b_0" shape = [1024] dtype = "float32" min_val = float("-0.876172") @@ -777,6 +848,7 @@ class Program_weight_tensor_parameter_70: class Program_weight_tensor_parameter_71: name = "parameter_71" + original_name = "linear_3384.w_0" shape = [3072, 1024] dtype = "float32" min_val = float("-0.527782") @@ -788,6 +860,7 @@ class Program_weight_tensor_parameter_71: class Program_weight_tensor_parameter_72: name = "parameter_72" + original_name = "linear_3383.b_0" shape = [3072] dtype = "float32" min_val = float("-0.33473") @@ -799,6 +872,7 @@ class Program_weight_tensor_parameter_72: class Program_weight_tensor_parameter_73: name = "parameter_73" + original_name = "linear_3383.w_0" shape = [1024, 3072] dtype = "float32" min_val = float("-0.646091") @@ -810,6 +884,7 @@ class Program_weight_tensor_parameter_73: class Program_weight_tensor_parameter_74: name = "parameter_74" + original_name = "linear_3382.b_0" shape = [1024] dtype = "float32" min_val = float("-0.27059") @@ -821,6 +896,7 @@ class Program_weight_tensor_parameter_74: class Program_weight_tensor_parameter_75: name = "parameter_75" + original_name = "linear_3382.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.160108") @@ -832,6 +908,7 @@ class Program_weight_tensor_parameter_75: class Program_weight_tensor_parameter_76: name = "parameter_76" + original_name = "linear_3381.b_0" shape = [1024] dtype = "float32" min_val = float("-0.39612") @@ -843,6 +920,7 @@ class Program_weight_tensor_parameter_76: class Program_weight_tensor_parameter_77: name = "parameter_77" + original_name = "linear_3381.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.173578") @@ -854,6 +932,7 @@ class Program_weight_tensor_parameter_77: class Program_weight_tensor_parameter_78: name = "parameter_78" + original_name = "linear_3380.b_0" shape = [1024] dtype = "float32" min_val = float("-120.37") @@ -865,6 +944,7 @@ class Program_weight_tensor_parameter_78: class Program_weight_tensor_parameter_79: name = "parameter_79" + original_name = "linear_3380.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.372253") @@ -876,6 +956,7 @@ class Program_weight_tensor_parameter_79: class Program_weight_tensor_parameter_80: name = "parameter_80" + original_name = "linear_3379.b_0" shape = [1024] dtype = "float32" min_val = float("-1.06998") @@ -887,6 +968,7 @@ class Program_weight_tensor_parameter_80: class Program_weight_tensor_parameter_81: name = "parameter_81" + original_name = "linear_3379.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.332408") @@ -898,6 +980,7 @@ class Program_weight_tensor_parameter_81: class Program_weight_tensor_parameter_82: name = "parameter_82" + original_name = "layer_norm_1167.b_0" shape = [1024] dtype = "float32" min_val = float("-0.542002") @@ -909,6 +992,7 @@ class Program_weight_tensor_parameter_82: class Program_weight_tensor_parameter_83: name = "parameter_83" + original_name = "layer_norm_1167.w_0" shape = [1024] dtype = "float32" min_val = float("0.296532") @@ -920,6 +1004,7 @@ class Program_weight_tensor_parameter_83: class Program_weight_tensor_parameter_84: name = "parameter_84" + original_name = "layer_norm_1166.b_0" shape = [1024] dtype = "float32" min_val = float("-0.457177") @@ -931,6 +1016,7 @@ class Program_weight_tensor_parameter_84: class Program_weight_tensor_parameter_85: name = "parameter_85" + original_name = "layer_norm_1166.w_0" shape = [1024] dtype = "float32" min_val = float("0.840719") @@ -942,6 +1028,7 @@ class Program_weight_tensor_parameter_85: class Program_weight_tensor_parameter_86: name = "parameter_86" + original_name = "linear_3378.b_0" shape = [1024] dtype = "float32" min_val = float("-0.68468") @@ -953,6 +1040,7 @@ class Program_weight_tensor_parameter_86: class Program_weight_tensor_parameter_87: name = "parameter_87" + original_name = "linear_3378.w_0" shape = [3072, 1024] dtype = "float32" min_val = float("-0.324425") @@ -964,6 +1052,7 @@ class Program_weight_tensor_parameter_87: class Program_weight_tensor_parameter_88: name = "parameter_88" + original_name = "linear_3377.b_0" shape = [3072] dtype = "float32" min_val = float("-0.601107") @@ -975,6 +1064,7 @@ class Program_weight_tensor_parameter_88: class Program_weight_tensor_parameter_89: name = "parameter_89" + original_name = "linear_3377.w_0" shape = [1024, 3072] dtype = "float32" min_val = float("-0.889477") @@ -986,6 +1076,7 @@ class Program_weight_tensor_parameter_89: class Program_weight_tensor_parameter_90: name = "parameter_90" + original_name = "linear_3376.b_0" shape = [1024] dtype = "float32" min_val = float("-0.1916") @@ -997,6 +1088,7 @@ class Program_weight_tensor_parameter_90: class Program_weight_tensor_parameter_91: name = "parameter_91" + original_name = "linear_3376.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.184642") @@ -1008,6 +1100,7 @@ class Program_weight_tensor_parameter_91: class Program_weight_tensor_parameter_92: name = "parameter_92" + original_name = "linear_3375.b_0" shape = [1024] dtype = "float32" min_val = float("-0.116997") @@ -1019,6 +1112,7 @@ class Program_weight_tensor_parameter_92: class Program_weight_tensor_parameter_93: name = "parameter_93" + original_name = "linear_3375.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.181015") @@ -1030,6 +1124,7 @@ class Program_weight_tensor_parameter_93: class Program_weight_tensor_parameter_94: name = "parameter_94" + original_name = "linear_3374.b_0" shape = [1024] dtype = "float32" min_val = float("-131.023") @@ -1041,6 +1136,7 @@ class Program_weight_tensor_parameter_94: class Program_weight_tensor_parameter_95: name = "parameter_95" + original_name = "linear_3374.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.400135") @@ -1052,6 +1148,7 @@ class Program_weight_tensor_parameter_95: class Program_weight_tensor_parameter_96: name = "parameter_96" + original_name = "linear_3373.b_0" shape = [1024] dtype = "float32" min_val = float("-1.93699") @@ -1063,6 +1160,7 @@ class Program_weight_tensor_parameter_96: class Program_weight_tensor_parameter_97: name = "parameter_97" + original_name = "linear_3373.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-1.22943") @@ -1074,6 +1172,7 @@ class Program_weight_tensor_parameter_97: class Program_weight_tensor_parameter_98: name = "parameter_98" + original_name = "layer_norm_1165.b_0" shape = [1024] dtype = "float32" min_val = float("-0.278468") @@ -1085,6 +1184,7 @@ class Program_weight_tensor_parameter_98: class Program_weight_tensor_parameter_99: name = "parameter_99" + original_name = "layer_norm_1165.w_0" shape = [1024] dtype = "float32" min_val = float("0.285529") @@ -1096,6 +1196,7 @@ class Program_weight_tensor_parameter_99: class Program_weight_tensor_parameter_100: name = "parameter_100" + original_name = "layer_norm_1164.b_0" shape = [1024] dtype = "float32" min_val = float("-0.993221") @@ -1107,6 +1208,7 @@ class Program_weight_tensor_parameter_100: class Program_weight_tensor_parameter_101: name = "parameter_101" + original_name = "layer_norm_1164.w_0" shape = [1024] dtype = "float32" min_val = float("0.791543") @@ -1118,6 +1220,7 @@ class Program_weight_tensor_parameter_101: class Program_weight_tensor_parameter_102: name = "parameter_102" + original_name = "linear_3372.b_0" shape = [1024] dtype = "float32" min_val = float("-1.80392") @@ -1129,6 +1232,7 @@ class Program_weight_tensor_parameter_102: class Program_weight_tensor_parameter_103: name = "parameter_103" + original_name = "linear_3372.w_0" shape = [3072, 1024] dtype = "float32" min_val = float("-0.919148") @@ -1140,6 +1244,7 @@ class Program_weight_tensor_parameter_103: class Program_weight_tensor_parameter_104: name = "parameter_104" + original_name = "linear_3371.b_0" shape = [3072] dtype = "float32" min_val = float("-0.467576") @@ -1151,6 +1256,7 @@ class Program_weight_tensor_parameter_104: class Program_weight_tensor_parameter_105: name = "parameter_105" + original_name = "linear_3371.w_0" shape = [1024, 3072] dtype = "float32" min_val = float("-0.472305") @@ -1162,6 +1268,7 @@ class Program_weight_tensor_parameter_105: class Program_weight_tensor_parameter_106: name = "parameter_106" + original_name = "linear_3370.b_0" shape = [1024] dtype = "float32" min_val = float("-0.305707") @@ -1173,6 +1280,7 @@ class Program_weight_tensor_parameter_106: class Program_weight_tensor_parameter_107: name = "parameter_107" + original_name = "linear_3370.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.224886") @@ -1184,6 +1292,7 @@ class Program_weight_tensor_parameter_107: class Program_weight_tensor_parameter_108: name = "parameter_108" + original_name = "linear_3369.b_0" shape = [1024] dtype = "float32" min_val = float("-0.172979") @@ -1195,6 +1304,7 @@ class Program_weight_tensor_parameter_108: class Program_weight_tensor_parameter_109: name = "parameter_109" + original_name = "linear_3369.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.178275") @@ -1206,6 +1316,7 @@ class Program_weight_tensor_parameter_109: class Program_weight_tensor_parameter_110: name = "parameter_110" + original_name = "linear_3368.b_0" shape = [1024] dtype = "float32" min_val = float("-127.547") @@ -1217,6 +1328,7 @@ class Program_weight_tensor_parameter_110: class Program_weight_tensor_parameter_111: name = "parameter_111" + original_name = "linear_3368.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.376737") @@ -1228,6 +1340,7 @@ class Program_weight_tensor_parameter_111: class Program_weight_tensor_parameter_112: name = "parameter_112" + original_name = "linear_3367.b_0" shape = [1024] dtype = "float32" min_val = float("-0.929703") @@ -1239,6 +1352,7 @@ class Program_weight_tensor_parameter_112: class Program_weight_tensor_parameter_113: name = "parameter_113" + original_name = "linear_3367.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.336967") @@ -1250,6 +1364,7 @@ class Program_weight_tensor_parameter_113: class Program_weight_tensor_parameter_114: name = "parameter_114" + original_name = "layer_norm_1163.b_0" shape = [1024] dtype = "float32" min_val = float("-0.867253") @@ -1261,6 +1376,7 @@ class Program_weight_tensor_parameter_114: class Program_weight_tensor_parameter_115: name = "parameter_115" + original_name = "layer_norm_1163.w_0" shape = [1024] dtype = "float32" min_val = float("0.352619") @@ -1272,6 +1388,7 @@ class Program_weight_tensor_parameter_115: class Program_weight_tensor_parameter_116: name = "parameter_116" + original_name = "layer_norm_1162.b_0" shape = [1024] dtype = "float32" min_val = float("-1.25286") @@ -1283,6 +1400,7 @@ class Program_weight_tensor_parameter_116: class Program_weight_tensor_parameter_117: name = "parameter_117" + original_name = "layer_norm_1162.w_0" shape = [1024] dtype = "float32" min_val = float("0.733411") @@ -1294,6 +1412,7 @@ class Program_weight_tensor_parameter_117: class Program_weight_tensor_parameter_118: name = "parameter_118" + original_name = "linear_3366.b_0" shape = [1024] dtype = "float32" min_val = float("-2.69978") @@ -1305,6 +1424,7 @@ class Program_weight_tensor_parameter_118: class Program_weight_tensor_parameter_119: name = "parameter_119" + original_name = "linear_3366.w_0" shape = [3072, 1024] dtype = "float32" min_val = float("-0.432742") @@ -1316,6 +1436,7 @@ class Program_weight_tensor_parameter_119: class Program_weight_tensor_parameter_120: name = "parameter_120" + original_name = "linear_3365.b_0" shape = [3072] dtype = "float32" min_val = float("-0.521763") @@ -1327,6 +1448,7 @@ class Program_weight_tensor_parameter_120: class Program_weight_tensor_parameter_121: name = "parameter_121" + original_name = "linear_3365.w_0" shape = [1024, 3072] dtype = "float32" min_val = float("-0.43709") @@ -1338,6 +1460,7 @@ class Program_weight_tensor_parameter_121: class Program_weight_tensor_parameter_122: name = "parameter_122" + original_name = "linear_3364.b_0" shape = [1024] dtype = "float32" min_val = float("-0.199") @@ -1349,6 +1472,7 @@ class Program_weight_tensor_parameter_122: class Program_weight_tensor_parameter_123: name = "parameter_123" + original_name = "linear_3364.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.317888") @@ -1360,6 +1484,7 @@ class Program_weight_tensor_parameter_123: class Program_weight_tensor_parameter_124: name = "parameter_124" + original_name = "linear_3363.b_0" shape = [1024] dtype = "float32" min_val = float("-0.152513") @@ -1371,6 +1496,7 @@ class Program_weight_tensor_parameter_124: class Program_weight_tensor_parameter_125: name = "parameter_125" + original_name = "linear_3363.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.204415") @@ -1382,6 +1508,7 @@ class Program_weight_tensor_parameter_125: class Program_weight_tensor_parameter_126: name = "parameter_126" + original_name = "linear_3362.b_0" shape = [1024] dtype = "float32" min_val = float("-121.623") @@ -1393,6 +1520,7 @@ class Program_weight_tensor_parameter_126: class Program_weight_tensor_parameter_127: name = "parameter_127" + original_name = "linear_3362.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.391908") @@ -1404,6 +1532,7 @@ class Program_weight_tensor_parameter_127: class Program_weight_tensor_parameter_128: name = "parameter_128" + original_name = "linear_3361.b_0" shape = [1024] dtype = "float32" min_val = float("-0.902927") @@ -1415,6 +1544,7 @@ class Program_weight_tensor_parameter_128: class Program_weight_tensor_parameter_129: name = "parameter_129" + original_name = "linear_3361.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.284817") @@ -1426,6 +1556,7 @@ class Program_weight_tensor_parameter_129: class Program_weight_tensor_parameter_130: name = "parameter_130" + original_name = "layer_norm_1161.b_0" shape = [1024] dtype = "float32" min_val = float("-0.918519") @@ -1437,6 +1568,7 @@ class Program_weight_tensor_parameter_130: class Program_weight_tensor_parameter_131: name = "parameter_131" + original_name = "layer_norm_1161.w_0" shape = [1024] dtype = "float32" min_val = float("0.376512") @@ -1448,6 +1580,7 @@ class Program_weight_tensor_parameter_131: class Program_weight_tensor_parameter_132: name = "parameter_132" + original_name = "layer_norm_1160.b_0" shape = [1024] dtype = "float32" min_val = float("-1.21496") @@ -1459,6 +1592,7 @@ class Program_weight_tensor_parameter_132: class Program_weight_tensor_parameter_133: name = "parameter_133" + original_name = "layer_norm_1160.w_0" shape = [1024] dtype = "float32" min_val = float("0.496399") @@ -1470,6 +1604,7 @@ class Program_weight_tensor_parameter_133: class Program_weight_tensor_parameter_134: name = "parameter_134" + original_name = "linear_3360.b_0" shape = [1024] dtype = "float32" min_val = float("-2.22335") @@ -1481,6 +1616,7 @@ class Program_weight_tensor_parameter_134: class Program_weight_tensor_parameter_135: name = "parameter_135" + original_name = "linear_3360.w_0" shape = [3072, 1024] dtype = "float32" min_val = float("-0.816313") @@ -1492,6 +1628,7 @@ class Program_weight_tensor_parameter_135: class Program_weight_tensor_parameter_136: name = "parameter_136" + original_name = "linear_3359.b_0" shape = [3072] dtype = "float32" min_val = float("-0.522565") @@ -1503,6 +1640,7 @@ class Program_weight_tensor_parameter_136: class Program_weight_tensor_parameter_137: name = "parameter_137" + original_name = "linear_3359.w_0" shape = [1024, 3072] dtype = "float32" min_val = float("-0.510587") @@ -1514,6 +1652,7 @@ class Program_weight_tensor_parameter_137: class Program_weight_tensor_parameter_138: name = "parameter_138" + original_name = "linear_3358.b_0" shape = [1024] dtype = "float32" min_val = float("-0.382521") @@ -1525,6 +1664,7 @@ class Program_weight_tensor_parameter_138: class Program_weight_tensor_parameter_139: name = "parameter_139" + original_name = "linear_3358.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.256931") @@ -1536,6 +1676,7 @@ class Program_weight_tensor_parameter_139: class Program_weight_tensor_parameter_140: name = "parameter_140" + original_name = "linear_3357.b_0" shape = [1024] dtype = "float32" min_val = float("-0.210733") @@ -1547,6 +1688,7 @@ class Program_weight_tensor_parameter_140: class Program_weight_tensor_parameter_141: name = "parameter_141" + original_name = "linear_3357.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.208357") @@ -1558,6 +1700,7 @@ class Program_weight_tensor_parameter_141: class Program_weight_tensor_parameter_142: name = "parameter_142" + original_name = "linear_3356.b_0" shape = [1024] dtype = "float32" min_val = float("-123.908") @@ -1569,6 +1712,7 @@ class Program_weight_tensor_parameter_142: class Program_weight_tensor_parameter_143: name = "parameter_143" + original_name = "linear_3356.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.374401") @@ -1580,6 +1724,7 @@ class Program_weight_tensor_parameter_143: class Program_weight_tensor_parameter_144: name = "parameter_144" + original_name = "linear_3355.b_0" shape = [1024] dtype = "float32" min_val = float("-0.899303") @@ -1591,6 +1736,7 @@ class Program_weight_tensor_parameter_144: class Program_weight_tensor_parameter_145: name = "parameter_145" + original_name = "linear_3355.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.343765") @@ -1602,6 +1748,7 @@ class Program_weight_tensor_parameter_145: class Program_weight_tensor_parameter_146: name = "parameter_146" + original_name = "layer_norm_1159.b_0" shape = [1024] dtype = "float32" min_val = float("-0.439545") @@ -1613,6 +1760,7 @@ class Program_weight_tensor_parameter_146: class Program_weight_tensor_parameter_147: name = "parameter_147" + original_name = "layer_norm_1159.w_0" shape = [1024] dtype = "float32" min_val = float("0.0902309") @@ -1624,6 +1772,7 @@ class Program_weight_tensor_parameter_147: class Program_weight_tensor_parameter_148: name = "parameter_148" + original_name = "layer_norm_1158.b_0" shape = [1024] dtype = "float32" min_val = float("-1.5028") @@ -1635,6 +1784,7 @@ class Program_weight_tensor_parameter_148: class Program_weight_tensor_parameter_149: name = "parameter_149" + original_name = "layer_norm_1158.w_0" shape = [1024] dtype = "float32" min_val = float("0.625404") @@ -1646,6 +1796,7 @@ class Program_weight_tensor_parameter_149: class Program_weight_tensor_parameter_150: name = "parameter_150" + original_name = "linear_3354.b_0" shape = [1024] dtype = "float32" min_val = float("-1.6165") @@ -1657,6 +1808,7 @@ class Program_weight_tensor_parameter_150: class Program_weight_tensor_parameter_151: name = "parameter_151" + original_name = "linear_3354.w_0" shape = [3072, 1024] dtype = "float32" min_val = float("-1.02025") @@ -1668,6 +1820,7 @@ class Program_weight_tensor_parameter_151: class Program_weight_tensor_parameter_152: name = "parameter_152" + original_name = "linear_3353.b_0" shape = [3072] dtype = "float32" min_val = float("-0.540893") @@ -1679,6 +1832,7 @@ class Program_weight_tensor_parameter_152: class Program_weight_tensor_parameter_153: name = "parameter_153" + original_name = "linear_3353.w_0" shape = [1024, 3072] dtype = "float32" min_val = float("-0.746763") @@ -1690,6 +1844,7 @@ class Program_weight_tensor_parameter_153: class Program_weight_tensor_parameter_154: name = "parameter_154" + original_name = "linear_3352.b_0" shape = [1024] dtype = "float32" min_val = float("-0.387763") @@ -1701,6 +1856,7 @@ class Program_weight_tensor_parameter_154: class Program_weight_tensor_parameter_155: name = "parameter_155" + original_name = "linear_3352.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.206834") @@ -1712,6 +1868,7 @@ class Program_weight_tensor_parameter_155: class Program_weight_tensor_parameter_156: name = "parameter_156" + original_name = "linear_3351.b_0" shape = [1024] dtype = "float32" min_val = float("-0.27606") @@ -1723,6 +1880,7 @@ class Program_weight_tensor_parameter_156: class Program_weight_tensor_parameter_157: name = "parameter_157" + original_name = "linear_3351.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.180364") @@ -1734,6 +1892,7 @@ class Program_weight_tensor_parameter_157: class Program_weight_tensor_parameter_158: name = "parameter_158" + original_name = "linear_3350.b_0" shape = [1024] dtype = "float32" min_val = float("-97.7678") @@ -1745,6 +1904,7 @@ class Program_weight_tensor_parameter_158: class Program_weight_tensor_parameter_159: name = "parameter_159" + original_name = "linear_3350.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.360995") @@ -1756,6 +1916,7 @@ class Program_weight_tensor_parameter_159: class Program_weight_tensor_parameter_160: name = "parameter_160" + original_name = "linear_3349.b_0" shape = [1024] dtype = "float32" min_val = float("-0.97282") @@ -1767,6 +1928,7 @@ class Program_weight_tensor_parameter_160: class Program_weight_tensor_parameter_161: name = "parameter_161" + original_name = "linear_3349.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.329137") @@ -1778,6 +1940,7 @@ class Program_weight_tensor_parameter_161: class Program_weight_tensor_parameter_162: name = "parameter_162" + original_name = "layer_norm_1157.b_0" shape = [1024] dtype = "float32" min_val = float("-0.494129") @@ -1789,6 +1952,7 @@ class Program_weight_tensor_parameter_162: class Program_weight_tensor_parameter_163: name = "parameter_163" + original_name = "layer_norm_1157.w_0" shape = [1024] dtype = "float32" min_val = float("0.110627") @@ -1800,6 +1964,7 @@ class Program_weight_tensor_parameter_163: class Program_weight_tensor_parameter_164: name = "parameter_164" + original_name = "layer_norm_1156.b_0" shape = [1024] dtype = "float32" min_val = float("-1.42636") @@ -1811,6 +1976,7 @@ class Program_weight_tensor_parameter_164: class Program_weight_tensor_parameter_165: name = "parameter_165" + original_name = "layer_norm_1156.w_0" shape = [1024] dtype = "float32" min_val = float("0.711652") @@ -1822,6 +1988,7 @@ class Program_weight_tensor_parameter_165: class Program_weight_tensor_parameter_166: name = "parameter_166" + original_name = "linear_3348.b_0" shape = [1024] dtype = "float32" min_val = float("-2.58251") @@ -1833,6 +2000,7 @@ class Program_weight_tensor_parameter_166: class Program_weight_tensor_parameter_167: name = "parameter_167" + original_name = "linear_3348.w_0" shape = [3072, 1024] dtype = "float32" min_val = float("-0.998045") @@ -1844,6 +2012,7 @@ class Program_weight_tensor_parameter_167: class Program_weight_tensor_parameter_168: name = "parameter_168" + original_name = "linear_3347.b_0" shape = [3072] dtype = "float32" min_val = float("-0.45786") @@ -1855,6 +2024,7 @@ class Program_weight_tensor_parameter_168: class Program_weight_tensor_parameter_169: name = "parameter_169" + original_name = "linear_3347.w_0" shape = [1024, 3072] dtype = "float32" min_val = float("-0.482146") @@ -1866,6 +2036,7 @@ class Program_weight_tensor_parameter_169: class Program_weight_tensor_parameter_170: name = "parameter_170" + original_name = "linear_3346.b_0" shape = [1024] dtype = "float32" min_val = float("-0.318175") @@ -1877,6 +2048,7 @@ class Program_weight_tensor_parameter_170: class Program_weight_tensor_parameter_171: name = "parameter_171" + original_name = "linear_3346.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.301532") @@ -1888,6 +2060,7 @@ class Program_weight_tensor_parameter_171: class Program_weight_tensor_parameter_172: name = "parameter_172" + original_name = "linear_3345.b_0" shape = [1024] dtype = "float32" min_val = float("-0.244215") @@ -1899,6 +2072,7 @@ class Program_weight_tensor_parameter_172: class Program_weight_tensor_parameter_173: name = "parameter_173" + original_name = "linear_3345.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.191551") @@ -1910,6 +2084,7 @@ class Program_weight_tensor_parameter_173: class Program_weight_tensor_parameter_174: name = "parameter_174" + original_name = "linear_3344.b_0" shape = [1024] dtype = "float32" min_val = float("-88.9345") @@ -1921,6 +2096,7 @@ class Program_weight_tensor_parameter_174: class Program_weight_tensor_parameter_175: name = "parameter_175" + original_name = "linear_3344.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.386082") @@ -1932,6 +2108,7 @@ class Program_weight_tensor_parameter_175: class Program_weight_tensor_parameter_176: name = "parameter_176" + original_name = "linear_3343.b_0" shape = [1024] dtype = "float32" min_val = float("-0.893027") @@ -1943,6 +2120,7 @@ class Program_weight_tensor_parameter_176: class Program_weight_tensor_parameter_177: name = "parameter_177" + original_name = "linear_3343.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.32864") @@ -1954,6 +2132,7 @@ class Program_weight_tensor_parameter_177: class Program_weight_tensor_parameter_178: name = "parameter_178" + original_name = "layer_norm_1155.b_0" shape = [1024] dtype = "float32" min_val = float("-0.533644") @@ -1965,6 +2144,7 @@ class Program_weight_tensor_parameter_178: class Program_weight_tensor_parameter_179: name = "parameter_179" + original_name = "layer_norm_1155.w_0" shape = [1024] dtype = "float32" min_val = float("0.301428") @@ -1976,6 +2156,7 @@ class Program_weight_tensor_parameter_179: class Program_weight_tensor_parameter_180: name = "parameter_180" + original_name = "layer_norm_1154.b_0" shape = [1024] dtype = "float32" min_val = float("-0.930415") @@ -1987,6 +2168,7 @@ class Program_weight_tensor_parameter_180: class Program_weight_tensor_parameter_181: name = "parameter_181" + original_name = "layer_norm_1154.w_0" shape = [1024] dtype = "float32" min_val = float("0.754785") @@ -1998,6 +2180,7 @@ class Program_weight_tensor_parameter_181: class Program_weight_tensor_parameter_182: name = "parameter_182" + original_name = "linear_3342.b_0" shape = [1024] dtype = "float32" min_val = float("-2.96274") @@ -2009,6 +2192,7 @@ class Program_weight_tensor_parameter_182: class Program_weight_tensor_parameter_183: name = "parameter_183" + original_name = "linear_3342.w_0" shape = [3072, 1024] dtype = "float32" min_val = float("-0.811971") @@ -2020,6 +2204,7 @@ class Program_weight_tensor_parameter_183: class Program_weight_tensor_parameter_184: name = "parameter_184" + original_name = "linear_3341.b_0" shape = [3072] dtype = "float32" min_val = float("-0.295419") @@ -2031,6 +2216,7 @@ class Program_weight_tensor_parameter_184: class Program_weight_tensor_parameter_185: name = "parameter_185" + original_name = "linear_3341.w_0" shape = [1024, 3072] dtype = "float32" min_val = float("-0.423977") @@ -2042,6 +2228,7 @@ class Program_weight_tensor_parameter_185: class Program_weight_tensor_parameter_186: name = "parameter_186" + original_name = "linear_3340.b_0" shape = [1024] dtype = "float32" min_val = float("-0.165429") @@ -2053,6 +2240,7 @@ class Program_weight_tensor_parameter_186: class Program_weight_tensor_parameter_187: name = "parameter_187" + original_name = "linear_3340.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.181991") @@ -2064,6 +2252,7 @@ class Program_weight_tensor_parameter_187: class Program_weight_tensor_parameter_188: name = "parameter_188" + original_name = "linear_3339.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0812541") @@ -2075,6 +2264,7 @@ class Program_weight_tensor_parameter_188: class Program_weight_tensor_parameter_189: name = "parameter_189" + original_name = "linear_3339.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.185099") @@ -2086,6 +2276,7 @@ class Program_weight_tensor_parameter_189: class Program_weight_tensor_parameter_190: name = "parameter_190" + original_name = "linear_3338.b_0" shape = [1024] dtype = "float32" min_val = float("-71.6475") @@ -2097,6 +2288,7 @@ class Program_weight_tensor_parameter_190: class Program_weight_tensor_parameter_191: name = "parameter_191" + original_name = "linear_3338.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.360331") @@ -2108,6 +2300,7 @@ class Program_weight_tensor_parameter_191: class Program_weight_tensor_parameter_192: name = "parameter_192" + original_name = "linear_3337.b_0" shape = [1024] dtype = "float32" min_val = float("-0.835171") @@ -2119,6 +2312,7 @@ class Program_weight_tensor_parameter_192: class Program_weight_tensor_parameter_193: name = "parameter_193" + original_name = "linear_3337.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.294538") @@ -2130,6 +2324,7 @@ class Program_weight_tensor_parameter_193: class Program_weight_tensor_parameter_194: name = "parameter_194" + original_name = "layer_norm_1153.b_0" shape = [1024] dtype = "float32" min_val = float("-0.514991") @@ -2141,6 +2336,7 @@ class Program_weight_tensor_parameter_194: class Program_weight_tensor_parameter_195: name = "parameter_195" + original_name = "layer_norm_1153.w_0" shape = [1024] dtype = "float32" min_val = float("0.371559") @@ -2152,6 +2348,7 @@ class Program_weight_tensor_parameter_195: class Program_weight_tensor_parameter_196: name = "parameter_196" + original_name = "layer_norm_1152.b_0" shape = [1024] dtype = "float32" min_val = float("-0.959735") @@ -2163,6 +2360,7 @@ class Program_weight_tensor_parameter_196: class Program_weight_tensor_parameter_197: name = "parameter_197" + original_name = "layer_norm_1152.w_0" shape = [1024] dtype = "float32" min_val = float("0.755497") @@ -2174,6 +2372,7 @@ class Program_weight_tensor_parameter_197: class Program_weight_tensor_parameter_198: name = "parameter_198" + original_name = "linear_3336.b_0" shape = [1024] dtype = "float32" min_val = float("-3.36073") @@ -2185,6 +2384,7 @@ class Program_weight_tensor_parameter_198: class Program_weight_tensor_parameter_199: name = "parameter_199" + original_name = "linear_3336.w_0" shape = [3072, 1024] dtype = "float32" min_val = float("-0.558442") @@ -2196,6 +2396,7 @@ class Program_weight_tensor_parameter_199: class Program_weight_tensor_parameter_200: name = "parameter_200" + original_name = "linear_3335.b_0" shape = [3072] dtype = "float32" min_val = float("-0.302099") @@ -2207,6 +2408,7 @@ class Program_weight_tensor_parameter_200: class Program_weight_tensor_parameter_201: name = "parameter_201" + original_name = "linear_3335.w_0" shape = [1024, 3072] dtype = "float32" min_val = float("-0.371458") @@ -2218,6 +2420,7 @@ class Program_weight_tensor_parameter_201: class Program_weight_tensor_parameter_202: name = "parameter_202" + original_name = "linear_3334.b_0" shape = [1024] dtype = "float32" min_val = float("-0.34772") @@ -2229,6 +2432,7 @@ class Program_weight_tensor_parameter_202: class Program_weight_tensor_parameter_203: name = "parameter_203" + original_name = "linear_3334.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-1.40619") @@ -2240,6 +2444,7 @@ class Program_weight_tensor_parameter_203: class Program_weight_tensor_parameter_204: name = "parameter_204" + original_name = "linear_3333.b_0" shape = [1024] dtype = "float32" min_val = float("-0.226983") @@ -2251,6 +2456,7 @@ class Program_weight_tensor_parameter_204: class Program_weight_tensor_parameter_205: name = "parameter_205" + original_name = "linear_3333.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.162785") @@ -2262,6 +2468,7 @@ class Program_weight_tensor_parameter_205: class Program_weight_tensor_parameter_206: name = "parameter_206" + original_name = "linear_3332.b_0" shape = [1024] dtype = "float32" min_val = float("-50.3207") @@ -2273,6 +2480,7 @@ class Program_weight_tensor_parameter_206: class Program_weight_tensor_parameter_207: name = "parameter_207" + original_name = "linear_3332.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.401029") @@ -2284,6 +2492,7 @@ class Program_weight_tensor_parameter_207: class Program_weight_tensor_parameter_208: name = "parameter_208" + original_name = "linear_3331.b_0" shape = [1024] dtype = "float32" min_val = float("-0.791765") @@ -2295,6 +2504,7 @@ class Program_weight_tensor_parameter_208: class Program_weight_tensor_parameter_209: name = "parameter_209" + original_name = "linear_3331.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.31174") @@ -2306,6 +2516,7 @@ class Program_weight_tensor_parameter_209: class Program_weight_tensor_parameter_210: name = "parameter_210" + original_name = "layer_norm_1151.b_0" shape = [1024] dtype = "float32" min_val = float("-0.386486") @@ -2317,6 +2528,7 @@ class Program_weight_tensor_parameter_210: class Program_weight_tensor_parameter_211: name = "parameter_211" + original_name = "layer_norm_1151.w_0" shape = [1024] dtype = "float32" min_val = float("0.355836") @@ -2328,6 +2540,7 @@ class Program_weight_tensor_parameter_211: class Program_weight_tensor_parameter_212: name = "parameter_212" + original_name = "layer_norm_1150.b_0" shape = [1024] dtype = "float32" min_val = float("-1.15984") @@ -2339,6 +2552,7 @@ class Program_weight_tensor_parameter_212: class Program_weight_tensor_parameter_213: name = "parameter_213" + original_name = "layer_norm_1150.w_0" shape = [1024] dtype = "float32" min_val = float("0.773986") @@ -2350,6 +2564,7 @@ class Program_weight_tensor_parameter_213: class Program_weight_tensor_parameter_214: name = "parameter_214" + original_name = "linear_3330.b_0" shape = [1024] dtype = "float32" min_val = float("-3.11289") @@ -2361,6 +2576,7 @@ class Program_weight_tensor_parameter_214: class Program_weight_tensor_parameter_215: name = "parameter_215" + original_name = "linear_3330.w_0" shape = [3072, 1024] dtype = "float32" min_val = float("-0.934539") @@ -2372,6 +2588,7 @@ class Program_weight_tensor_parameter_215: class Program_weight_tensor_parameter_216: name = "parameter_216" + original_name = "linear_3329.b_0" shape = [3072] dtype = "float32" min_val = float("-0.386627") @@ -2383,6 +2600,7 @@ class Program_weight_tensor_parameter_216: class Program_weight_tensor_parameter_217: name = "parameter_217" + original_name = "linear_3329.w_0" shape = [1024, 3072] dtype = "float32" min_val = float("-0.483042") @@ -2394,6 +2612,7 @@ class Program_weight_tensor_parameter_217: class Program_weight_tensor_parameter_218: name = "parameter_218" + original_name = "linear_3328.b_0" shape = [1024] dtype = "float32" min_val = float("-0.577972") @@ -2405,6 +2624,7 @@ class Program_weight_tensor_parameter_218: class Program_weight_tensor_parameter_219: name = "parameter_219" + original_name = "linear_3328.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-2.08548") @@ -2416,6 +2636,7 @@ class Program_weight_tensor_parameter_219: class Program_weight_tensor_parameter_220: name = "parameter_220" + original_name = "linear_3327.b_0" shape = [1024] dtype = "float32" min_val = float("-0.176819") @@ -2427,6 +2648,7 @@ class Program_weight_tensor_parameter_220: class Program_weight_tensor_parameter_221: name = "parameter_221" + original_name = "linear_3327.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.411407") @@ -2438,6 +2660,7 @@ class Program_weight_tensor_parameter_221: class Program_weight_tensor_parameter_222: name = "parameter_222" + original_name = "linear_3326.b_0" shape = [1024] dtype = "float32" min_val = float("-30.3417") @@ -2449,6 +2672,7 @@ class Program_weight_tensor_parameter_222: class Program_weight_tensor_parameter_223: name = "parameter_223" + original_name = "linear_3326.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.368319") @@ -2460,6 +2684,7 @@ class Program_weight_tensor_parameter_223: class Program_weight_tensor_parameter_224: name = "parameter_224" + original_name = "linear_3325.b_0" shape = [1024] dtype = "float32" min_val = float("-0.859069") @@ -2471,6 +2696,7 @@ class Program_weight_tensor_parameter_224: class Program_weight_tensor_parameter_225: name = "parameter_225" + original_name = "linear_3325.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.397683") @@ -2482,6 +2708,7 @@ class Program_weight_tensor_parameter_225: class Program_weight_tensor_parameter_226: name = "parameter_226" + original_name = "layer_norm_1149.b_0" shape = [1024] dtype = "float32" min_val = float("-0.442596") @@ -2493,6 +2720,7 @@ class Program_weight_tensor_parameter_226: class Program_weight_tensor_parameter_227: name = "parameter_227" + original_name = "layer_norm_1149.w_0" shape = [1024] dtype = "float32" min_val = float("0.333406") @@ -2504,6 +2732,7 @@ class Program_weight_tensor_parameter_227: class Program_weight_tensor_parameter_228: name = "parameter_228" + original_name = "layer_norm_1148.b_0" shape = [1024] dtype = "float32" min_val = float("-1.07107") @@ -2515,6 +2744,7 @@ class Program_weight_tensor_parameter_228: class Program_weight_tensor_parameter_229: name = "parameter_229" + original_name = "layer_norm_1148.w_0" shape = [1024] dtype = "float32" min_val = float("0.757126") @@ -2526,6 +2756,7 @@ class Program_weight_tensor_parameter_229: class Program_weight_tensor_parameter_230: name = "parameter_230" + original_name = "linear_3324.b_0" shape = [1024] dtype = "float32" min_val = float("-2.96934") @@ -2537,6 +2768,7 @@ class Program_weight_tensor_parameter_230: class Program_weight_tensor_parameter_231: name = "parameter_231" + original_name = "linear_3324.w_0" shape = [3072, 1024] dtype = "float32" min_val = float("-0.585014") @@ -2548,6 +2780,7 @@ class Program_weight_tensor_parameter_231: class Program_weight_tensor_parameter_232: name = "parameter_232" + original_name = "linear_3323.b_0" shape = [3072] dtype = "float32" min_val = float("-0.405719") @@ -2559,6 +2792,7 @@ class Program_weight_tensor_parameter_232: class Program_weight_tensor_parameter_233: name = "parameter_233" + original_name = "linear_3323.w_0" shape = [1024, 3072] dtype = "float32" min_val = float("-0.443763") @@ -2570,6 +2804,7 @@ class Program_weight_tensor_parameter_233: class Program_weight_tensor_parameter_234: name = "parameter_234" + original_name = "linear_3322.b_0" shape = [1024] dtype = "float32" min_val = float("-0.384648") @@ -2581,6 +2816,7 @@ class Program_weight_tensor_parameter_234: class Program_weight_tensor_parameter_235: name = "parameter_235" + original_name = "linear_3322.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.46536") @@ -2592,6 +2828,7 @@ class Program_weight_tensor_parameter_235: class Program_weight_tensor_parameter_236: name = "parameter_236" + original_name = "linear_3321.b_0" shape = [1024] dtype = "float32" min_val = float("-0.34696") @@ -2603,6 +2840,7 @@ class Program_weight_tensor_parameter_236: class Program_weight_tensor_parameter_237: name = "parameter_237" + original_name = "linear_3321.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.203668") @@ -2614,6 +2852,7 @@ class Program_weight_tensor_parameter_237: class Program_weight_tensor_parameter_238: name = "parameter_238" + original_name = "linear_3320.b_0" shape = [1024] dtype = "float32" min_val = float("-40.3062") @@ -2625,6 +2864,7 @@ class Program_weight_tensor_parameter_238: class Program_weight_tensor_parameter_239: name = "parameter_239" + original_name = "linear_3320.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.433798") @@ -2636,6 +2876,7 @@ class Program_weight_tensor_parameter_239: class Program_weight_tensor_parameter_240: name = "parameter_240" + original_name = "linear_3319.b_0" shape = [1024] dtype = "float32" min_val = float("-0.71027") @@ -2647,6 +2888,7 @@ class Program_weight_tensor_parameter_240: class Program_weight_tensor_parameter_241: name = "parameter_241" + original_name = "linear_3319.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.381372") @@ -2658,6 +2900,7 @@ class Program_weight_tensor_parameter_241: class Program_weight_tensor_parameter_242: name = "parameter_242" + original_name = "layer_norm_1147.b_0" shape = [1024] dtype = "float32" min_val = float("-0.763102") @@ -2669,6 +2912,7 @@ class Program_weight_tensor_parameter_242: class Program_weight_tensor_parameter_243: name = "parameter_243" + original_name = "layer_norm_1147.w_0" shape = [1024] dtype = "float32" min_val = float("0.161468") @@ -2680,6 +2924,7 @@ class Program_weight_tensor_parameter_243: class Program_weight_tensor_parameter_244: name = "parameter_244" + original_name = "layer_norm_1146.b_0" shape = [1024] dtype = "float32" min_val = float("-3.19943") @@ -2691,6 +2936,7 @@ class Program_weight_tensor_parameter_244: class Program_weight_tensor_parameter_245: name = "parameter_245" + original_name = "layer_norm_1146.w_0" shape = [1024] dtype = "float32" min_val = float("0.784237") @@ -2702,6 +2948,7 @@ class Program_weight_tensor_parameter_245: class Program_weight_tensor_parameter_246: name = "parameter_246" + original_name = "linear_3318.b_0" shape = [1024] dtype = "float32" min_val = float("-2.1066") @@ -2713,6 +2960,7 @@ class Program_weight_tensor_parameter_246: class Program_weight_tensor_parameter_247: name = "parameter_247" + original_name = "linear_3318.w_0" shape = [3072, 1024] dtype = "float32" min_val = float("-1.10062") @@ -2724,6 +2972,7 @@ class Program_weight_tensor_parameter_247: class Program_weight_tensor_parameter_248: name = "parameter_248" + original_name = "linear_3317.b_0" shape = [3072] dtype = "float32" min_val = float("-0.286695") @@ -2735,6 +2984,7 @@ class Program_weight_tensor_parameter_248: class Program_weight_tensor_parameter_249: name = "parameter_249" + original_name = "linear_3317.w_0" shape = [1024, 3072] dtype = "float32" min_val = float("-0.349521") @@ -2746,6 +2996,7 @@ class Program_weight_tensor_parameter_249: class Program_weight_tensor_parameter_250: name = "parameter_250" + original_name = "linear_3316.b_0" shape = [1024] dtype = "float32" min_val = float("-0.137614") @@ -2757,6 +3008,7 @@ class Program_weight_tensor_parameter_250: class Program_weight_tensor_parameter_251: name = "parameter_251" + original_name = "linear_3316.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.182414") @@ -2768,6 +3020,7 @@ class Program_weight_tensor_parameter_251: class Program_weight_tensor_parameter_252: name = "parameter_252" + original_name = "linear_3315.b_0" shape = [1024] dtype = "float32" min_val = float("-0.136658") @@ -2779,6 +3032,7 @@ class Program_weight_tensor_parameter_252: class Program_weight_tensor_parameter_253: name = "parameter_253" + original_name = "linear_3315.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.311345") @@ -2790,6 +3044,7 @@ class Program_weight_tensor_parameter_253: class Program_weight_tensor_parameter_254: name = "parameter_254" + original_name = "linear_3314.b_0" shape = [1024] dtype = "float32" min_val = float("-27.8905") @@ -2801,6 +3056,7 @@ class Program_weight_tensor_parameter_254: class Program_weight_tensor_parameter_255: name = "parameter_255" + original_name = "linear_3314.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.425076") @@ -2812,6 +3068,7 @@ class Program_weight_tensor_parameter_255: class Program_weight_tensor_parameter_256: name = "parameter_256" + original_name = "linear_3313.b_0" shape = [1024] dtype = "float32" min_val = float("-0.918921") @@ -2823,6 +3080,7 @@ class Program_weight_tensor_parameter_256: class Program_weight_tensor_parameter_257: name = "parameter_257" + original_name = "linear_3313.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.35422") @@ -2834,6 +3092,7 @@ class Program_weight_tensor_parameter_257: class Program_weight_tensor_parameter_258: name = "parameter_258" + original_name = "layer_norm_1145.b_0" shape = [1024] dtype = "float32" min_val = float("-1.13853") @@ -2845,6 +3104,7 @@ class Program_weight_tensor_parameter_258: class Program_weight_tensor_parameter_259: name = "parameter_259" + original_name = "layer_norm_1145.w_0" shape = [1024] dtype = "float32" min_val = float("0.243789") @@ -2856,6 +3116,7 @@ class Program_weight_tensor_parameter_259: class Program_weight_tensor_parameter_260: name = "parameter_260" + original_name = "layer_norm_1144.b_0" shape = [1024] dtype = "float32" min_val = float("-2.54937") @@ -2867,6 +3128,7 @@ class Program_weight_tensor_parameter_260: class Program_weight_tensor_parameter_261: name = "parameter_261" + original_name = "layer_norm_1144.w_0" shape = [1024] dtype = "float32" min_val = float("0.796313") @@ -2878,6 +3140,7 @@ class Program_weight_tensor_parameter_261: class Program_weight_tensor_parameter_262: name = "parameter_262" + original_name = "linear_3312.b_0" shape = [1024] dtype = "float32" min_val = float("-1.15776") @@ -2889,6 +3152,7 @@ class Program_weight_tensor_parameter_262: class Program_weight_tensor_parameter_263: name = "parameter_263" + original_name = "linear_3312.w_0" shape = [3072, 1024] dtype = "float32" min_val = float("-1.56459") @@ -2900,6 +3164,7 @@ class Program_weight_tensor_parameter_263: class Program_weight_tensor_parameter_264: name = "parameter_264" + original_name = "linear_3311.b_0" shape = [3072] dtype = "float32" min_val = float("-0.243279") @@ -2911,6 +3176,7 @@ class Program_weight_tensor_parameter_264: class Program_weight_tensor_parameter_265: name = "parameter_265" + original_name = "linear_3311.w_0" shape = [1024, 3072] dtype = "float32" min_val = float("-0.506971") @@ -2922,6 +3188,7 @@ class Program_weight_tensor_parameter_265: class Program_weight_tensor_parameter_266: name = "parameter_266" + original_name = "linear_3310.b_0" shape = [1024] dtype = "float32" min_val = float("-0.353443") @@ -2933,6 +3200,7 @@ class Program_weight_tensor_parameter_266: class Program_weight_tensor_parameter_267: name = "parameter_267" + original_name = "linear_3310.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.380094") @@ -2944,6 +3212,7 @@ class Program_weight_tensor_parameter_267: class Program_weight_tensor_parameter_268: name = "parameter_268" + original_name = "linear_3309.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0898818") @@ -2955,6 +3224,7 @@ class Program_weight_tensor_parameter_268: class Program_weight_tensor_parameter_269: name = "parameter_269" + original_name = "linear_3309.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.186794") @@ -2966,6 +3236,7 @@ class Program_weight_tensor_parameter_269: class Program_weight_tensor_parameter_270: name = "parameter_270" + original_name = "linear_3308.b_0" shape = [1024] dtype = "float32" min_val = float("-29.5129") @@ -2977,6 +3248,7 @@ class Program_weight_tensor_parameter_270: class Program_weight_tensor_parameter_271: name = "parameter_271" + original_name = "linear_3308.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.347392") @@ -2988,6 +3260,7 @@ class Program_weight_tensor_parameter_271: class Program_weight_tensor_parameter_272: name = "parameter_272" + original_name = "linear_3307.b_0" shape = [1024] dtype = "float32" min_val = float("-0.906221") @@ -2999,6 +3272,7 @@ class Program_weight_tensor_parameter_272: class Program_weight_tensor_parameter_273: name = "parameter_273" + original_name = "linear_3307.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.319976") @@ -3010,6 +3284,7 @@ class Program_weight_tensor_parameter_273: class Program_weight_tensor_parameter_274: name = "parameter_274" + original_name = "layer_norm_1143.b_0" shape = [1024] dtype = "float32" min_val = float("-1.48846") @@ -3021,6 +3296,7 @@ class Program_weight_tensor_parameter_274: class Program_weight_tensor_parameter_275: name = "parameter_275" + original_name = "layer_norm_1143.w_0" shape = [1024] dtype = "float32" min_val = float("0.365271") @@ -3032,6 +3308,7 @@ class Program_weight_tensor_parameter_275: class Program_weight_tensor_parameter_276: name = "parameter_276" + original_name = "layer_norm_1142.b_0" shape = [1024] dtype = "float32" min_val = float("-2.58463") @@ -3043,6 +3320,7 @@ class Program_weight_tensor_parameter_276: class Program_weight_tensor_parameter_277: name = "parameter_277" + original_name = "layer_norm_1142.w_0" shape = [1024] dtype = "float32" min_val = float("0.756009") @@ -3054,6 +3332,7 @@ class Program_weight_tensor_parameter_277: class Program_weight_tensor_parameter_278: name = "parameter_278" + original_name = "linear_3306.b_0" shape = [1024] dtype = "float32" min_val = float("-1.19347") @@ -3065,6 +3344,7 @@ class Program_weight_tensor_parameter_278: class Program_weight_tensor_parameter_279: name = "parameter_279" + original_name = "linear_3306.w_0" shape = [3072, 1024] dtype = "float32" min_val = float("-3.40753") @@ -3076,6 +3356,7 @@ class Program_weight_tensor_parameter_279: class Program_weight_tensor_parameter_280: name = "parameter_280" + original_name = "linear_3305.b_0" shape = [3072] dtype = "float32" min_val = float("-0.222803") @@ -3087,6 +3368,7 @@ class Program_weight_tensor_parameter_280: class Program_weight_tensor_parameter_281: name = "parameter_281" + original_name = "linear_3305.w_0" shape = [1024, 3072] dtype = "float32" min_val = float("-0.27603") @@ -3098,6 +3380,7 @@ class Program_weight_tensor_parameter_281: class Program_weight_tensor_parameter_282: name = "parameter_282" + original_name = "linear_3304.b_0" shape = [1024] dtype = "float32" min_val = float("-0.2271") @@ -3109,6 +3392,7 @@ class Program_weight_tensor_parameter_282: class Program_weight_tensor_parameter_283: name = "parameter_283" + original_name = "linear_3304.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.324873") @@ -3120,6 +3404,7 @@ class Program_weight_tensor_parameter_283: class Program_weight_tensor_parameter_284: name = "parameter_284" + original_name = "linear_3303.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0890741") @@ -3131,6 +3416,7 @@ class Program_weight_tensor_parameter_284: class Program_weight_tensor_parameter_285: name = "parameter_285" + original_name = "linear_3303.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.191792") @@ -3142,6 +3428,7 @@ class Program_weight_tensor_parameter_285: class Program_weight_tensor_parameter_286: name = "parameter_286" + original_name = "linear_3302.b_0" shape = [1024] dtype = "float32" min_val = float("-36.6444") @@ -3153,6 +3440,7 @@ class Program_weight_tensor_parameter_286: class Program_weight_tensor_parameter_287: name = "parameter_287" + original_name = "linear_3302.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.347466") @@ -3164,6 +3452,7 @@ class Program_weight_tensor_parameter_287: class Program_weight_tensor_parameter_288: name = "parameter_288" + original_name = "linear_3301.b_0" shape = [1024] dtype = "float32" min_val = float("-0.79373") @@ -3175,6 +3464,7 @@ class Program_weight_tensor_parameter_288: class Program_weight_tensor_parameter_289: name = "parameter_289" + original_name = "linear_3301.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.388889") @@ -3186,6 +3476,7 @@ class Program_weight_tensor_parameter_289: class Program_weight_tensor_parameter_290: name = "parameter_290" + original_name = "layer_norm_1141.b_0" shape = [1024] dtype = "float32" min_val = float("-1.30698") @@ -3197,6 +3488,7 @@ class Program_weight_tensor_parameter_290: class Program_weight_tensor_parameter_291: name = "parameter_291" + original_name = "layer_norm_1141.w_0" shape = [1024] dtype = "float32" min_val = float("0.398626") @@ -3208,6 +3500,7 @@ class Program_weight_tensor_parameter_291: class Program_weight_tensor_parameter_292: name = "parameter_292" + original_name = "layer_norm_1140.b_0" shape = [1024] dtype = "float32" min_val = float("-2.13026") @@ -3219,6 +3512,7 @@ class Program_weight_tensor_parameter_292: class Program_weight_tensor_parameter_293: name = "parameter_293" + original_name = "layer_norm_1140.w_0" shape = [1024] dtype = "float32" min_val = float("0.625952") @@ -3230,6 +3524,7 @@ class Program_weight_tensor_parameter_293: class Program_weight_tensor_parameter_294: name = "parameter_294" + original_name = "linear_3300.b_0" shape = [1024] dtype = "float32" min_val = float("-1.37661") @@ -3241,6 +3536,7 @@ class Program_weight_tensor_parameter_294: class Program_weight_tensor_parameter_295: name = "parameter_295" + original_name = "linear_3300.w_0" shape = [3072, 1024] dtype = "float32" min_val = float("-6.29049") @@ -3252,6 +3548,7 @@ class Program_weight_tensor_parameter_295: class Program_weight_tensor_parameter_296: name = "parameter_296" + original_name = "linear_3299.b_0" shape = [3072] dtype = "float32" min_val = float("-0.234225") @@ -3263,6 +3560,7 @@ class Program_weight_tensor_parameter_296: class Program_weight_tensor_parameter_297: name = "parameter_297" + original_name = "linear_3299.w_0" shape = [1024, 3072] dtype = "float32" min_val = float("-0.31839") @@ -3274,6 +3572,7 @@ class Program_weight_tensor_parameter_297: class Program_weight_tensor_parameter_298: name = "parameter_298" + original_name = "linear_3298.b_0" shape = [1024] dtype = "float32" min_val = float("-0.232133") @@ -3285,6 +3584,7 @@ class Program_weight_tensor_parameter_298: class Program_weight_tensor_parameter_299: name = "parameter_299" + original_name = "linear_3298.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.332377") @@ -3296,6 +3596,7 @@ class Program_weight_tensor_parameter_299: class Program_weight_tensor_parameter_300: name = "parameter_300" + original_name = "linear_3297.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0793155") @@ -3307,6 +3608,7 @@ class Program_weight_tensor_parameter_300: class Program_weight_tensor_parameter_301: name = "parameter_301" + original_name = "linear_3297.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.199237") @@ -3318,6 +3620,7 @@ class Program_weight_tensor_parameter_301: class Program_weight_tensor_parameter_302: name = "parameter_302" + original_name = "linear_3296.b_0" shape = [1024] dtype = "float32" min_val = float("-32.3963") @@ -3329,6 +3632,7 @@ class Program_weight_tensor_parameter_302: class Program_weight_tensor_parameter_303: name = "parameter_303" + original_name = "linear_3296.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.328859") @@ -3340,6 +3644,7 @@ class Program_weight_tensor_parameter_303: class Program_weight_tensor_parameter_304: name = "parameter_304" + original_name = "linear_3295.b_0" shape = [1024] dtype = "float32" min_val = float("-0.811503") @@ -3351,6 +3656,7 @@ class Program_weight_tensor_parameter_304: class Program_weight_tensor_parameter_305: name = "parameter_305" + original_name = "linear_3295.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.34011") @@ -3362,6 +3668,7 @@ class Program_weight_tensor_parameter_305: class Program_weight_tensor_parameter_306: name = "parameter_306" + original_name = "layer_norm_1139.b_0" shape = [1024] dtype = "float32" min_val = float("-1.12317") @@ -3373,6 +3680,7 @@ class Program_weight_tensor_parameter_306: class Program_weight_tensor_parameter_307: name = "parameter_307" + original_name = "layer_norm_1139.w_0" shape = [1024] dtype = "float32" min_val = float("0.438732") @@ -3384,6 +3692,7 @@ class Program_weight_tensor_parameter_307: class Program_weight_tensor_parameter_308: name = "parameter_308" + original_name = "layer_norm_1138.b_0" shape = [1024] dtype = "float32" min_val = float("-1.58939") @@ -3395,6 +3704,7 @@ class Program_weight_tensor_parameter_308: class Program_weight_tensor_parameter_309: name = "parameter_309" + original_name = "layer_norm_1138.w_0" shape = [1024] dtype = "float32" min_val = float("0.692015") @@ -3406,6 +3716,7 @@ class Program_weight_tensor_parameter_309: class Program_weight_tensor_parameter_310: name = "parameter_310" + original_name = "linear_3294.b_0" shape = [1024] dtype = "float32" min_val = float("-1.54001") @@ -3417,6 +3728,7 @@ class Program_weight_tensor_parameter_310: class Program_weight_tensor_parameter_311: name = "parameter_311" + original_name = "linear_3294.w_0" shape = [3072, 1024] dtype = "float32" min_val = float("-7.07342") @@ -3428,6 +3740,7 @@ class Program_weight_tensor_parameter_311: class Program_weight_tensor_parameter_312: name = "parameter_312" + original_name = "linear_3293.b_0" shape = [3072] dtype = "float32" min_val = float("-0.235678") @@ -3439,6 +3752,7 @@ class Program_weight_tensor_parameter_312: class Program_weight_tensor_parameter_313: name = "parameter_313" + original_name = "linear_3293.w_0" shape = [1024, 3072] dtype = "float32" min_val = float("-0.543783") @@ -3450,6 +3764,7 @@ class Program_weight_tensor_parameter_313: class Program_weight_tensor_parameter_314: name = "parameter_314" + original_name = "linear_3292.b_0" shape = [1024] dtype = "float32" min_val = float("-0.340475") @@ -3461,6 +3776,7 @@ class Program_weight_tensor_parameter_314: class Program_weight_tensor_parameter_315: name = "parameter_315" + original_name = "linear_3292.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.429231") @@ -3472,6 +3788,7 @@ class Program_weight_tensor_parameter_315: class Program_weight_tensor_parameter_316: name = "parameter_316" + original_name = "linear_3291.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0870212") @@ -3483,6 +3800,7 @@ class Program_weight_tensor_parameter_316: class Program_weight_tensor_parameter_317: name = "parameter_317" + original_name = "linear_3291.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.225329") @@ -3494,6 +3812,7 @@ class Program_weight_tensor_parameter_317: class Program_weight_tensor_parameter_318: name = "parameter_318" + original_name = "linear_3290.b_0" shape = [1024] dtype = "float32" min_val = float("-23.2692") @@ -3505,6 +3824,7 @@ class Program_weight_tensor_parameter_318: class Program_weight_tensor_parameter_319: name = "parameter_319" + original_name = "linear_3290.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.330439") @@ -3516,6 +3836,7 @@ class Program_weight_tensor_parameter_319: class Program_weight_tensor_parameter_320: name = "parameter_320" + original_name = "linear_3289.b_0" shape = [1024] dtype = "float32" min_val = float("-0.937012") @@ -3527,6 +3848,7 @@ class Program_weight_tensor_parameter_320: class Program_weight_tensor_parameter_321: name = "parameter_321" + original_name = "linear_3289.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.384069") @@ -3538,6 +3860,7 @@ class Program_weight_tensor_parameter_321: class Program_weight_tensor_parameter_322: name = "parameter_322" + original_name = "layer_norm_1137.b_0" shape = [1024] dtype = "float32" min_val = float("-0.47526") @@ -3549,6 +3872,7 @@ class Program_weight_tensor_parameter_322: class Program_weight_tensor_parameter_323: name = "parameter_323" + original_name = "layer_norm_1137.w_0" shape = [1024] dtype = "float32" min_val = float("0.0914798") @@ -3560,6 +3884,7 @@ class Program_weight_tensor_parameter_323: class Program_weight_tensor_parameter_324: name = "parameter_324" + original_name = "layer_norm_1136.b_0" shape = [1024] dtype = "float32" min_val = float("-1.84961") @@ -3571,6 +3896,7 @@ class Program_weight_tensor_parameter_324: class Program_weight_tensor_parameter_325: name = "parameter_325" + original_name = "layer_norm_1136.w_0" shape = [1024] dtype = "float32" min_val = float("0.650388") @@ -3582,6 +3908,7 @@ class Program_weight_tensor_parameter_325: class Program_weight_tensor_parameter_326: name = "parameter_326" + original_name = "linear_3288.b_0" shape = [1024] dtype = "float32" min_val = float("-1.98624") @@ -3593,6 +3920,7 @@ class Program_weight_tensor_parameter_326: class Program_weight_tensor_parameter_327: name = "parameter_327" + original_name = "linear_3288.w_0" shape = [3072, 1024] dtype = "float32" min_val = float("-9.09443") @@ -3604,6 +3932,7 @@ class Program_weight_tensor_parameter_327: class Program_weight_tensor_parameter_328: name = "parameter_328" + original_name = "linear_3287.b_0" shape = [3072] dtype = "float32" min_val = float("-0.247981") @@ -3615,6 +3944,7 @@ class Program_weight_tensor_parameter_328: class Program_weight_tensor_parameter_329: name = "parameter_329" + original_name = "linear_3287.w_0" shape = [1024, 3072] dtype = "float32" min_val = float("-0.587541") @@ -3626,6 +3956,7 @@ class Program_weight_tensor_parameter_329: class Program_weight_tensor_parameter_330: name = "parameter_330" + original_name = "linear_3286.b_0" shape = [1024] dtype = "float32" min_val = float("-0.518518") @@ -3637,6 +3968,7 @@ class Program_weight_tensor_parameter_330: class Program_weight_tensor_parameter_331: name = "parameter_331" + original_name = "linear_3286.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.234722") @@ -3648,6 +3980,7 @@ class Program_weight_tensor_parameter_331: class Program_weight_tensor_parameter_332: name = "parameter_332" + original_name = "linear_3285.b_0" shape = [1024] dtype = "float32" min_val = float("-0.212552") @@ -3659,6 +3992,7 @@ class Program_weight_tensor_parameter_332: class Program_weight_tensor_parameter_333: name = "parameter_333" + original_name = "linear_3285.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.441802") @@ -3670,6 +4004,7 @@ class Program_weight_tensor_parameter_333: class Program_weight_tensor_parameter_334: name = "parameter_334" + original_name = "linear_3284.b_0" shape = [1024] dtype = "float32" min_val = float("-26.561") @@ -3681,6 +4016,7 @@ class Program_weight_tensor_parameter_334: class Program_weight_tensor_parameter_335: name = "parameter_335" + original_name = "linear_3284.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.723539") @@ -3692,6 +4028,7 @@ class Program_weight_tensor_parameter_335: class Program_weight_tensor_parameter_336: name = "parameter_336" + original_name = "linear_3283.b_0" shape = [1024] dtype = "float32" min_val = float("-0.756511") @@ -3703,6 +4040,7 @@ class Program_weight_tensor_parameter_336: class Program_weight_tensor_parameter_337: name = "parameter_337" + original_name = "linear_3283.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.544746") @@ -3714,6 +4052,7 @@ class Program_weight_tensor_parameter_337: class Program_weight_tensor_parameter_338: name = "parameter_338" + original_name = "layer_norm_1135.b_0" shape = [1024] dtype = "float32" min_val = float("-0.273104") @@ -3725,6 +4064,7 @@ class Program_weight_tensor_parameter_338: class Program_weight_tensor_parameter_339: name = "parameter_339" + original_name = "layer_norm_1135.w_0" shape = [1024] dtype = "float32" min_val = float("0.195772") @@ -3736,6 +4076,7 @@ class Program_weight_tensor_parameter_339: class Program_weight_tensor_parameter_340: name = "parameter_340" + original_name = "layer_norm_1134.b_0" shape = [1024] dtype = "float32" min_val = float("-2.37565") @@ -3747,6 +4088,7 @@ class Program_weight_tensor_parameter_340: class Program_weight_tensor_parameter_341: name = "parameter_341" + original_name = "layer_norm_1134.w_0" shape = [1024] dtype = "float32" min_val = float("0.391391") @@ -3758,6 +4100,7 @@ class Program_weight_tensor_parameter_341: class Program_weight_tensor_parameter_342: name = "parameter_342" + original_name = "linear_3282.b_0" shape = [1024] dtype = "float32" min_val = float("-2.53376") @@ -3769,6 +4112,7 @@ class Program_weight_tensor_parameter_342: class Program_weight_tensor_parameter_343: name = "parameter_343" + original_name = "linear_3282.w_0" shape = [3072, 1024] dtype = "float32" min_val = float("-5.22465") @@ -3780,6 +4124,7 @@ class Program_weight_tensor_parameter_343: class Program_weight_tensor_parameter_344: name = "parameter_344" + original_name = "linear_3281.b_0" shape = [3072] dtype = "float32" min_val = float("-0.239292") @@ -3791,6 +4136,7 @@ class Program_weight_tensor_parameter_344: class Program_weight_tensor_parameter_345: name = "parameter_345" + original_name = "linear_3281.w_0" shape = [1024, 3072] dtype = "float32" min_val = float("-0.574736") @@ -3802,6 +4148,7 @@ class Program_weight_tensor_parameter_345: class Program_weight_tensor_parameter_346: name = "parameter_346" + original_name = "linear_3280.b_0" shape = [1024] dtype = "float32" min_val = float("-0.469969") @@ -3813,6 +4160,7 @@ class Program_weight_tensor_parameter_346: class Program_weight_tensor_parameter_347: name = "parameter_347" + original_name = "linear_3280.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.441307") @@ -3824,6 +4172,7 @@ class Program_weight_tensor_parameter_347: class Program_weight_tensor_parameter_348: name = "parameter_348" + original_name = "linear_3279.b_0" shape = [1024] dtype = "float32" min_val = float("-0.36476") @@ -3835,6 +4184,7 @@ class Program_weight_tensor_parameter_348: class Program_weight_tensor_parameter_349: name = "parameter_349" + original_name = "linear_3279.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.34401") @@ -3846,6 +4196,7 @@ class Program_weight_tensor_parameter_349: class Program_weight_tensor_parameter_350: name = "parameter_350" + original_name = "linear_3278.b_0" shape = [1024] dtype = "float32" min_val = float("-14.3712") @@ -3857,6 +4208,7 @@ class Program_weight_tensor_parameter_350: class Program_weight_tensor_parameter_351: name = "parameter_351" + original_name = "linear_3278.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.441406") @@ -3868,6 +4220,7 @@ class Program_weight_tensor_parameter_351: class Program_weight_tensor_parameter_352: name = "parameter_352" + original_name = "linear_3277.b_0" shape = [1024] dtype = "float32" min_val = float("-0.732734") @@ -3879,6 +4232,7 @@ class Program_weight_tensor_parameter_352: class Program_weight_tensor_parameter_353: name = "parameter_353" + original_name = "linear_3277.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.305703") @@ -3890,6 +4244,7 @@ class Program_weight_tensor_parameter_353: class Program_weight_tensor_parameter_354: name = "parameter_354" + original_name = "layer_norm_1133.b_0" shape = [1024] dtype = "float32" min_val = float("-0.392564") @@ -3901,6 +4256,7 @@ class Program_weight_tensor_parameter_354: class Program_weight_tensor_parameter_355: name = "parameter_355" + original_name = "layer_norm_1133.w_0" shape = [1024] dtype = "float32" min_val = float("0.1122") @@ -3912,6 +4268,7 @@ class Program_weight_tensor_parameter_355: class Program_weight_tensor_parameter_356: name = "parameter_356" + original_name = "layer_norm_1132.b_0" shape = [1024] dtype = "float32" min_val = float("-3.07715") @@ -3923,6 +4280,7 @@ class Program_weight_tensor_parameter_356: class Program_weight_tensor_parameter_357: name = "parameter_357" + original_name = "layer_norm_1132.w_0" shape = [1024] dtype = "float32" min_val = float("0.727013") @@ -3934,6 +4292,7 @@ class Program_weight_tensor_parameter_357: class Program_weight_tensor_parameter_358: name = "parameter_358" + original_name = "linear_3276.b_0" shape = [1024] dtype = "float32" min_val = float("-2.24168") @@ -3945,6 +4304,7 @@ class Program_weight_tensor_parameter_358: class Program_weight_tensor_parameter_359: name = "parameter_359" + original_name = "linear_3276.w_0" shape = [3072, 1024] dtype = "float32" min_val = float("-1.75203") @@ -3956,6 +4316,7 @@ class Program_weight_tensor_parameter_359: class Program_weight_tensor_parameter_360: name = "parameter_360" + original_name = "linear_3275.b_0" shape = [3072] dtype = "float32" min_val = float("-0.307765") @@ -3967,6 +4328,7 @@ class Program_weight_tensor_parameter_360: class Program_weight_tensor_parameter_361: name = "parameter_361" + original_name = "linear_3275.w_0" shape = [1024, 3072] dtype = "float32" min_val = float("-0.316847") @@ -3978,6 +4340,7 @@ class Program_weight_tensor_parameter_361: class Program_weight_tensor_parameter_362: name = "parameter_362" + original_name = "linear_3274.b_0" shape = [1024] dtype = "float32" min_val = float("-0.359781") @@ -3989,6 +4352,7 @@ class Program_weight_tensor_parameter_362: class Program_weight_tensor_parameter_363: name = "parameter_363" + original_name = "linear_3274.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.353609") @@ -4000,6 +4364,7 @@ class Program_weight_tensor_parameter_363: class Program_weight_tensor_parameter_364: name = "parameter_364" + original_name = "linear_3273.b_0" shape = [1024] dtype = "float32" min_val = float("-0.8217") @@ -4011,6 +4376,7 @@ class Program_weight_tensor_parameter_364: class Program_weight_tensor_parameter_365: name = "parameter_365" + original_name = "linear_3273.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.191421") @@ -4022,6 +4388,7 @@ class Program_weight_tensor_parameter_365: class Program_weight_tensor_parameter_366: name = "parameter_366" + original_name = "linear_3272.b_0" shape = [1024] dtype = "float32" min_val = float("-7.61407") @@ -4033,6 +4400,7 @@ class Program_weight_tensor_parameter_366: class Program_weight_tensor_parameter_367: name = "parameter_367" + original_name = "linear_3272.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.388772") @@ -4044,6 +4412,7 @@ class Program_weight_tensor_parameter_367: class Program_weight_tensor_parameter_368: name = "parameter_368" + original_name = "linear_3271.b_0" shape = [1024] dtype = "float32" min_val = float("-1.3901") @@ -4055,6 +4424,7 @@ class Program_weight_tensor_parameter_368: class Program_weight_tensor_parameter_369: name = "parameter_369" + original_name = "linear_3271.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.286249") @@ -4066,6 +4436,7 @@ class Program_weight_tensor_parameter_369: class Program_weight_tensor_parameter_370: name = "parameter_370" + original_name = "layer_norm_1131.b_0" shape = [1024] dtype = "float32" min_val = float("-1.06177") @@ -4077,6 +4448,7 @@ class Program_weight_tensor_parameter_370: class Program_weight_tensor_parameter_371: name = "parameter_371" + original_name = "layer_norm_1131.w_0" shape = [1024] dtype = "float32" min_val = float("0.124651") @@ -4088,6 +4460,7 @@ class Program_weight_tensor_parameter_371: class Program_weight_tensor_parameter_372: name = "parameter_372" + original_name = "layer_norm_1130.b_0" shape = [1024] dtype = "float32" min_val = float("-3.64256") @@ -4099,6 +4472,7 @@ class Program_weight_tensor_parameter_372: class Program_weight_tensor_parameter_373: name = "parameter_373" + original_name = "layer_norm_1130.w_0" shape = [1024] dtype = "float32" min_val = float("0.75281") @@ -4110,6 +4484,7 @@ class Program_weight_tensor_parameter_373: class Program_weight_tensor_parameter_374: name = "parameter_374" + original_name = "linear_3270.b_0" shape = [1024] dtype = "float32" min_val = float("-1.76359") @@ -4121,6 +4496,7 @@ class Program_weight_tensor_parameter_374: class Program_weight_tensor_parameter_375: name = "parameter_375" + original_name = "linear_3270.w_0" shape = [3072, 1024] dtype = "float32" min_val = float("-0.984892") @@ -4132,6 +4508,7 @@ class Program_weight_tensor_parameter_375: class Program_weight_tensor_parameter_376: name = "parameter_376" + original_name = "linear_3269.b_0" shape = [3072] dtype = "float32" min_val = float("-0.394219") @@ -4143,6 +4520,7 @@ class Program_weight_tensor_parameter_376: class Program_weight_tensor_parameter_377: name = "parameter_377" + original_name = "linear_3269.w_0" shape = [1024, 3072] dtype = "float32" min_val = float("-0.329957") @@ -4154,6 +4532,7 @@ class Program_weight_tensor_parameter_377: class Program_weight_tensor_parameter_378: name = "parameter_378" + original_name = "linear_3268.b_0" shape = [1024] dtype = "float32" min_val = float("-0.470717") @@ -4165,6 +4544,7 @@ class Program_weight_tensor_parameter_378: class Program_weight_tensor_parameter_379: name = "parameter_379" + original_name = "linear_3268.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.517553") @@ -4176,6 +4556,7 @@ class Program_weight_tensor_parameter_379: class Program_weight_tensor_parameter_380: name = "parameter_380" + original_name = "linear_3267.b_0" shape = [1024] dtype = "float32" min_val = float("-0.843538") @@ -4187,6 +4568,7 @@ class Program_weight_tensor_parameter_380: class Program_weight_tensor_parameter_381: name = "parameter_381" + original_name = "linear_3267.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.418397") @@ -4198,6 +4580,7 @@ class Program_weight_tensor_parameter_381: class Program_weight_tensor_parameter_382: name = "parameter_382" + original_name = "linear_3266.b_0" shape = [1024] dtype = "float32" min_val = float("-4.2017") @@ -4209,6 +4592,7 @@ class Program_weight_tensor_parameter_382: class Program_weight_tensor_parameter_383: name = "parameter_383" + original_name = "linear_3266.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.27631") @@ -4220,6 +4604,7 @@ class Program_weight_tensor_parameter_383: class Program_weight_tensor_parameter_384: name = "parameter_384" + original_name = "linear_3265.b_0" shape = [1024] dtype = "float32" min_val = float("-1.67318") @@ -4231,6 +4616,7 @@ class Program_weight_tensor_parameter_384: class Program_weight_tensor_parameter_385: name = "parameter_385" + original_name = "linear_3265.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.487252") @@ -4242,6 +4628,7 @@ class Program_weight_tensor_parameter_385: class Program_weight_tensor_parameter_386: name = "parameter_386" + original_name = "layer_norm_1129.b_0" shape = [1024] dtype = "float32" min_val = float("-1.10377") @@ -4253,6 +4640,7 @@ class Program_weight_tensor_parameter_386: class Program_weight_tensor_parameter_387: name = "parameter_387" + original_name = "layer_norm_1129.w_0" shape = [1024] dtype = "float32" min_val = float("0.085907") @@ -4264,6 +4652,7 @@ class Program_weight_tensor_parameter_387: class Program_weight_tensor_parameter_388: name = "parameter_388" + original_name = "embedding_225.w_0" shape = [3, 1024] dtype = "float32" min_val = float("-0.0631576") @@ -4275,6 +4664,7 @@ class Program_weight_tensor_parameter_388: class Program_weight_tensor_parameter_389: name = "parameter_389" + original_name = "embedding_224.w_0" shape = [2, 1024] dtype = "float32" min_val = float("-0.736974") @@ -4286,6 +4676,7 @@ class Program_weight_tensor_parameter_389: class Program_weight_tensor_parameter_390: name = "parameter_390" + original_name = "embedding_223.w_0" shape = [512, 1024] dtype = "float32" min_val = float("-0.401593") @@ -4297,6 +4688,7 @@ class Program_weight_tensor_parameter_390: class Program_weight_tensor_parameter_391: name = "parameter_391" + original_name = "embedding_222.w_0" shape = [17965, 1024] dtype = "float32" min_val = float("-4.10967") diff --git a/paddle_samples/PaddleNLP/utc-xbase/input_meta.py b/paddle_samples/PaddleNLP/utc-xbase/input_meta.py index 4f25a05a9..055cd8978 100644 --- a/paddle_samples/PaddleNLP/utc-xbase/input_meta.py +++ b/paddle_samples/PaddleNLP/utc-xbase/input_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_data_0: name = "data_0" + original_name = "input_ids" shape = [1, 21] dtype = "int64" data = [ @@ -29,6 +30,7 @@ class Program_weight_tensor_data_0: class Program_weight_tensor_data_1: name = "data_1" + original_name = "token_type_ids" shape = [1, 21] dtype = "int64" data = [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0] diff --git a/paddle_samples/PaddleNLP/utc-xbase/model.py b/paddle_samples/PaddleNLP/utc-xbase/model.py index 805cdb2ef..806734f4d 100644 --- a/paddle_samples/PaddleNLP/utc-xbase/model.py +++ b/paddle_samples/PaddleNLP/utc-xbase/model.py @@ -394,10 +394,11 @@ def forward( # pd_op.embedding: (1x21x1024xf32) <- (1x21xi64, 2048x1024xf32) embedding_1 = paddle._C_ops.embedding(subtract_0, parameter_326, -1, False) - del parameter_326 + del parameter_326, subtract_0 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_0 = paddle._C_ops.add(embedding_0, embedding_1) + del embedding_0, embedding_1 # pd_op.embedding: (1x21x1024xf32) <- (1x21xi64, 4x1024xf32) embedding_2 = paddle._C_ops.embedding(data_1, parameter_325, -1, False) @@ -405,6 +406,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_1 = paddle._C_ops.add(add_0, embedding_2) + del add_0, embedding_2 # pd_op.full: (1xf32) <- () full_4 = paddle._C_ops.full( @@ -417,10 +419,11 @@ def forward( # pd_op.embedding: (1x21x1024xf32) <- (1x21xi64, 16x1024xf32) embedding_3 = paddle._C_ops.embedding(scale_1, parameter_324, -1, False) - del parameter_324 + del parameter_324, scale_1 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_2 = paddle._C_ops.add(add_1, embedding_3) + del add_1, embedding_3 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_0, layer_norm_1, layer_norm_2 = (lambda x, f: f(x))( @@ -429,197 +432,17 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_322, parameter_323 + del add_2, parameter_322, parameter_323 # pd_op.full: (1xf32) <- () full_5 = paddle._C_ops.full( [1], float("0.1"), paddle.float32, paddle.core.CPUPlace() ) - # pd_op.assign: (1xf32) <- (1xf32) - assign_0 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_1 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_2 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_3 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_4 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_5 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_6 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_7 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_8 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_9 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_10 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_11 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_12 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_13 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_14 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_15 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_16 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_17 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_18 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_19 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_20 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_21 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_22 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_23 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_24 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_25 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_26 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_27 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_28 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_29 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_30 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_31 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_32 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_33 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_34 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_35 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_36 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_37 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_38 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_39 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_40 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_41 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_42 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_43 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_44 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_45 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_46 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_47 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_48 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_49 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_50 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_51 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_52 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_53 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_54 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_55 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_56 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_57 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_58 = full_5 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_59 = full_5 - # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_0, dropout_1 = (lambda x, f: f(x))( paddle._C_ops.dropout( - layer_norm_0, None, full_5, False, "upscale_in_train", 0, False + layer_norm_0, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -631,13 +454,14 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_3 = paddle._C_ops.add(matmul_0, parameter_320) - del parameter_320 + del matmul_0, parameter_320 # pd_op.full_int_array: (4xi64) <- () full_int_array_1 = [0, 0, 16, 64] # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_0 = paddle._C_ops.reshape(add_3, full_int_array_1) + del add_3 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_0 = paddle._C_ops.transpose(reshape_0, [0, 2, 1, 3]) @@ -649,7 +473,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_4 = paddle._C_ops.add(matmul_1, parameter_318) - del parameter_318 + del matmul_1, parameter_318 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_2 = paddle._C_ops.matmul(dropout_0, parameter_317, False, False) @@ -657,10 +481,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_5 = paddle._C_ops.add(matmul_2, parameter_316) - del parameter_316 + del matmul_2, parameter_316 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_1 = paddle._C_ops.reshape(add_4, full_int_array_1) + del add_4 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_1 = paddle._C_ops.transpose(reshape_1, [0, 2, 1, 3]) @@ -668,6 +493,7 @@ def forward( # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_2 = paddle._C_ops.reshape(add_5, full_int_array_1) + del add_5 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_2 = paddle._C_ops.transpose(reshape_2, [0, 2, 1, 3]) @@ -678,72 +504,17 @@ def forward( [1], float("0.125"), paddle.float32, paddle.core.CPUPlace() ) - # pd_op.assign: (1xf32) <- (1xf32) - assign_60 = full_6 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_61 = full_6 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_62 = full_6 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_63 = full_6 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_64 = full_6 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_65 = full_6 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_66 = full_6 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_67 = full_6 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_68 = full_6 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_69 = full_6 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_70 = full_6 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_71 = full_6 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_72 = full_6 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_73 = full_6 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_74 = full_6 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_75 = full_6 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_76 = full_6 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_77 = full_6 - - # pd_op.assign: (1xf32) <- (1xf32) - assign_78 = full_6 - # pd_op.scale: (1x16x21x64xf32) <- (1x16x21x64xf32, 1xf32) scale_2 = paddle._C_ops.scale(transpose_0, full_6, float("0"), True) del transpose_0 # pd_op.matmul: (1x16x21x21xf32) <- (1x16x21x64xf32, 1x16x21x64xf32) matmul_3 = paddle._C_ops.matmul(scale_2, transpose_1, False, True) + del scale_2, transpose_1 # pd_op.add: (1x16x21x21xf32) <- (1x16x21x21xf32, 1x1x1x21xf32) add_6 = paddle._C_ops.add(matmul_3, unsqueeze_0) + del matmul_3 # pd_op.softmax: (1x16x21x21xf32) <- (1x16x21x21xf32) softmax_0 = paddle._C_ops.softmax(add_6, -1) @@ -752,13 +523,15 @@ def forward( # pd_op.dropout: (1x16x21x21xf32, 1x16x21x21xui8) <- (1x16x21x21xf32, None, 1xf32) dropout_2, dropout_3 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_0, None, full_5, False, "upscale_in_train", 0, False + softmax_0, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_0 # pd_op.matmul: (1x16x21x64xf32) <- (1x16x21x21xf32, 1x16x21x64xf32) matmul_4 = paddle._C_ops.matmul(dropout_2, transpose_2, False, False) + del dropout_2, transpose_2 # pd_op.transpose: (1x21x16x64xf32) <- (1x16x21x64xf32) transpose_3 = paddle._C_ops.transpose(matmul_4, [0, 2, 1, 3]) @@ -769,19 +542,20 @@ def forward( # pd_op.reshape: (1x21x1024xf32) <- (1x21x16x64xf32, 3xi64) reshape_3 = paddle._C_ops.reshape(transpose_3, full_int_array_2) + del transpose_3 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_5 = paddle._C_ops.matmul(reshape_3, parameter_315, False, False) - del parameter_315 + del parameter_315, reshape_3 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_7 = paddle._C_ops.add(matmul_5, parameter_314) - del parameter_314 + del matmul_5, parameter_314 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_4, dropout_5 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_7, None, full_5, False, "upscale_in_train", 0, False + add_7, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -789,6 +563,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_8 = paddle._C_ops.add(dropout_0, dropout_4) + del dropout_0, dropout_4 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_3, layer_norm_4, layer_norm_5 = (lambda x, f: f(x))( @@ -797,7 +572,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_308, parameter_309 + del add_8, parameter_308, parameter_309 # pd_op.matmul: (1x21x4096xf32) <- (1x21x1024xf32, 1024x4096xf32) matmul_6 = paddle._C_ops.matmul(layer_norm_3, parameter_313, False, False) @@ -805,23 +580,24 @@ def forward( # pd_op.add: (1x21x4096xf32) <- (1x21x4096xf32, 4096xf32) add_9 = paddle._C_ops.add(matmul_6, parameter_312) - del parameter_312 + del matmul_6, parameter_312 # pd_op.gelu: (1x21x4096xf32) <- (1x21x4096xf32) gelu_0 = paddle._C_ops.gelu(add_9, False) + del add_9 # pd_op.matmul: (1x21x1024xf32) <- (1x21x4096xf32, 4096x1024xf32) matmul_7 = paddle._C_ops.matmul(gelu_0, parameter_311, False, False) - del parameter_311 + del gelu_0, parameter_311 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_10 = paddle._C_ops.add(matmul_7, parameter_310) - del parameter_310 + del matmul_7, parameter_310 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_6, dropout_7 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_10, None, full_5, False, "upscale_in_train", 0, False + add_10, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -829,6 +605,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_11 = paddle._C_ops.add(layer_norm_3, dropout_6) + del dropout_6, layer_norm_3 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_6, layer_norm_7, layer_norm_8 = (lambda x, f: f(x))( @@ -837,7 +614,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_306, parameter_307 + del add_11, parameter_306, parameter_307 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_8 = paddle._C_ops.matmul(layer_norm_6, parameter_305, False, False) @@ -845,10 +622,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_12 = paddle._C_ops.add(matmul_8, parameter_304) - del parameter_304 + del matmul_8, parameter_304 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_4 = paddle._C_ops.reshape(add_12, full_int_array_1) + del add_12 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_4 = paddle._C_ops.transpose(reshape_4, [0, 2, 1, 3]) @@ -860,7 +638,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_13 = paddle._C_ops.add(matmul_9, parameter_302) - del parameter_302 + del matmul_9, parameter_302 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_10 = paddle._C_ops.matmul(layer_norm_6, parameter_301, False, False) @@ -868,10 +646,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_14 = paddle._C_ops.add(matmul_10, parameter_300) - del parameter_300 + del matmul_10, parameter_300 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_5 = paddle._C_ops.reshape(add_13, full_int_array_1) + del add_13 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_5 = paddle._C_ops.transpose(reshape_5, [0, 2, 1, 3]) @@ -879,6 +658,7 @@ def forward( # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_6 = paddle._C_ops.reshape(add_14, full_int_array_1) + del add_14 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_6 = paddle._C_ops.transpose(reshape_6, [0, 2, 1, 3]) @@ -890,9 +670,11 @@ def forward( # pd_op.matmul: (1x16x21x21xf32) <- (1x16x21x64xf32, 1x16x21x64xf32) matmul_11 = paddle._C_ops.matmul(scale_3, transpose_5, False, True) + del scale_3, transpose_5 # pd_op.add: (1x16x21x21xf32) <- (1x16x21x21xf32, 1x1x1x21xf32) add_15 = paddle._C_ops.add(matmul_11, unsqueeze_0) + del matmul_11 # pd_op.softmax: (1x16x21x21xf32) <- (1x16x21x21xf32) softmax_1 = paddle._C_ops.softmax(add_15, -1) @@ -901,13 +683,15 @@ def forward( # pd_op.dropout: (1x16x21x21xf32, 1x16x21x21xui8) <- (1x16x21x21xf32, None, 1xf32) dropout_8, dropout_9 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_1, None, full_5, False, "upscale_in_train", 0, False + softmax_1, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_1 # pd_op.matmul: (1x16x21x64xf32) <- (1x16x21x21xf32, 1x16x21x64xf32) matmul_12 = paddle._C_ops.matmul(dropout_8, transpose_6, False, False) + del dropout_8, transpose_6 # pd_op.transpose: (1x21x16x64xf32) <- (1x16x21x64xf32) transpose_7 = paddle._C_ops.transpose(matmul_12, [0, 2, 1, 3]) @@ -915,19 +699,20 @@ def forward( # pd_op.reshape: (1x21x1024xf32) <- (1x21x16x64xf32, 3xi64) reshape_7 = paddle._C_ops.reshape(transpose_7, full_int_array_2) + del transpose_7 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_13 = paddle._C_ops.matmul(reshape_7, parameter_299, False, False) - del parameter_299 + del parameter_299, reshape_7 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_16 = paddle._C_ops.add(matmul_13, parameter_298) - del parameter_298 + del matmul_13, parameter_298 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_10, dropout_11 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_16, None, full_5, False, "upscale_in_train", 0, False + add_16, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -935,6 +720,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_17 = paddle._C_ops.add(layer_norm_6, dropout_10) + del dropout_10, layer_norm_6 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_9, layer_norm_10, layer_norm_11 = (lambda x, f: f(x))( @@ -943,7 +729,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_292, parameter_293 + del add_17, parameter_292, parameter_293 # pd_op.matmul: (1x21x4096xf32) <- (1x21x1024xf32, 1024x4096xf32) matmul_14 = paddle._C_ops.matmul(layer_norm_9, parameter_297, False, False) @@ -951,23 +737,24 @@ def forward( # pd_op.add: (1x21x4096xf32) <- (1x21x4096xf32, 4096xf32) add_18 = paddle._C_ops.add(matmul_14, parameter_296) - del parameter_296 + del matmul_14, parameter_296 # pd_op.gelu: (1x21x4096xf32) <- (1x21x4096xf32) gelu_1 = paddle._C_ops.gelu(add_18, False) + del add_18 # pd_op.matmul: (1x21x1024xf32) <- (1x21x4096xf32, 4096x1024xf32) matmul_15 = paddle._C_ops.matmul(gelu_1, parameter_295, False, False) - del parameter_295 + del gelu_1, parameter_295 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_19 = paddle._C_ops.add(matmul_15, parameter_294) - del parameter_294 + del matmul_15, parameter_294 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_12, dropout_13 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_19, None, full_5, False, "upscale_in_train", 0, False + add_19, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -975,6 +762,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_20 = paddle._C_ops.add(layer_norm_9, dropout_12) + del dropout_12, layer_norm_9 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_12, layer_norm_13, layer_norm_14 = (lambda x, f: f(x))( @@ -983,7 +771,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_290, parameter_291 + del add_20, parameter_290, parameter_291 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_16 = paddle._C_ops.matmul(layer_norm_12, parameter_289, False, False) @@ -991,10 +779,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_21 = paddle._C_ops.add(matmul_16, parameter_288) - del parameter_288 + del matmul_16, parameter_288 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_8 = paddle._C_ops.reshape(add_21, full_int_array_1) + del add_21 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_8 = paddle._C_ops.transpose(reshape_8, [0, 2, 1, 3]) @@ -1006,7 +795,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_22 = paddle._C_ops.add(matmul_17, parameter_286) - del parameter_286 + del matmul_17, parameter_286 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_18 = paddle._C_ops.matmul(layer_norm_12, parameter_285, False, False) @@ -1014,10 +803,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_23 = paddle._C_ops.add(matmul_18, parameter_284) - del parameter_284 + del matmul_18, parameter_284 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_9 = paddle._C_ops.reshape(add_22, full_int_array_1) + del add_22 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_9 = paddle._C_ops.transpose(reshape_9, [0, 2, 1, 3]) @@ -1025,6 +815,7 @@ def forward( # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_10 = paddle._C_ops.reshape(add_23, full_int_array_1) + del add_23 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_10 = paddle._C_ops.transpose(reshape_10, [0, 2, 1, 3]) @@ -1036,9 +827,11 @@ def forward( # pd_op.matmul: (1x16x21x21xf32) <- (1x16x21x64xf32, 1x16x21x64xf32) matmul_19 = paddle._C_ops.matmul(scale_4, transpose_9, False, True) + del scale_4, transpose_9 # pd_op.add: (1x16x21x21xf32) <- (1x16x21x21xf32, 1x1x1x21xf32) add_24 = paddle._C_ops.add(matmul_19, unsqueeze_0) + del matmul_19 # pd_op.softmax: (1x16x21x21xf32) <- (1x16x21x21xf32) softmax_2 = paddle._C_ops.softmax(add_24, -1) @@ -1047,13 +840,15 @@ def forward( # pd_op.dropout: (1x16x21x21xf32, 1x16x21x21xui8) <- (1x16x21x21xf32, None, 1xf32) dropout_14, dropout_15 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_2, None, full_5, False, "upscale_in_train", 0, False + softmax_2, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_2 # pd_op.matmul: (1x16x21x64xf32) <- (1x16x21x21xf32, 1x16x21x64xf32) matmul_20 = paddle._C_ops.matmul(dropout_14, transpose_10, False, False) + del dropout_14, transpose_10 # pd_op.transpose: (1x21x16x64xf32) <- (1x16x21x64xf32) transpose_11 = paddle._C_ops.transpose(matmul_20, [0, 2, 1, 3]) @@ -1061,19 +856,20 @@ def forward( # pd_op.reshape: (1x21x1024xf32) <- (1x21x16x64xf32, 3xi64) reshape_11 = paddle._C_ops.reshape(transpose_11, full_int_array_2) + del transpose_11 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_21 = paddle._C_ops.matmul(reshape_11, parameter_283, False, False) - del parameter_283 + del parameter_283, reshape_11 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_25 = paddle._C_ops.add(matmul_21, parameter_282) - del parameter_282 + del matmul_21, parameter_282 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_16, dropout_17 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_25, None, full_5, False, "upscale_in_train", 0, False + add_25, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1081,6 +877,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_26 = paddle._C_ops.add(layer_norm_12, dropout_16) + del dropout_16, layer_norm_12 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_15, layer_norm_16, layer_norm_17 = (lambda x, f: f(x))( @@ -1089,7 +886,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_276, parameter_277 + del add_26, parameter_276, parameter_277 # pd_op.matmul: (1x21x4096xf32) <- (1x21x1024xf32, 1024x4096xf32) matmul_22 = paddle._C_ops.matmul(layer_norm_15, parameter_281, False, False) @@ -1097,23 +894,24 @@ def forward( # pd_op.add: (1x21x4096xf32) <- (1x21x4096xf32, 4096xf32) add_27 = paddle._C_ops.add(matmul_22, parameter_280) - del parameter_280 + del matmul_22, parameter_280 # pd_op.gelu: (1x21x4096xf32) <- (1x21x4096xf32) gelu_2 = paddle._C_ops.gelu(add_27, False) + del add_27 # pd_op.matmul: (1x21x1024xf32) <- (1x21x4096xf32, 4096x1024xf32) matmul_23 = paddle._C_ops.matmul(gelu_2, parameter_279, False, False) - del parameter_279 + del gelu_2, parameter_279 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_28 = paddle._C_ops.add(matmul_23, parameter_278) - del parameter_278 + del matmul_23, parameter_278 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_18, dropout_19 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_28, None, full_5, False, "upscale_in_train", 0, False + add_28, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1121,6 +919,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_29 = paddle._C_ops.add(layer_norm_15, dropout_18) + del dropout_18, layer_norm_15 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_18, layer_norm_19, layer_norm_20 = (lambda x, f: f(x))( @@ -1129,7 +928,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_274, parameter_275 + del add_29, parameter_274, parameter_275 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_24 = paddle._C_ops.matmul(layer_norm_18, parameter_273, False, False) @@ -1137,10 +936,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_30 = paddle._C_ops.add(matmul_24, parameter_272) - del parameter_272 + del matmul_24, parameter_272 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_12 = paddle._C_ops.reshape(add_30, full_int_array_1) + del add_30 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_12 = paddle._C_ops.transpose(reshape_12, [0, 2, 1, 3]) @@ -1152,7 +952,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_31 = paddle._C_ops.add(matmul_25, parameter_270) - del parameter_270 + del matmul_25, parameter_270 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_26 = paddle._C_ops.matmul(layer_norm_18, parameter_269, False, False) @@ -1160,10 +960,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_32 = paddle._C_ops.add(matmul_26, parameter_268) - del parameter_268 + del matmul_26, parameter_268 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_13 = paddle._C_ops.reshape(add_31, full_int_array_1) + del add_31 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_13 = paddle._C_ops.transpose(reshape_13, [0, 2, 1, 3]) @@ -1171,6 +972,7 @@ def forward( # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_14 = paddle._C_ops.reshape(add_32, full_int_array_1) + del add_32 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_14 = paddle._C_ops.transpose(reshape_14, [0, 2, 1, 3]) @@ -1182,9 +984,11 @@ def forward( # pd_op.matmul: (1x16x21x21xf32) <- (1x16x21x64xf32, 1x16x21x64xf32) matmul_27 = paddle._C_ops.matmul(scale_5, transpose_13, False, True) + del scale_5, transpose_13 # pd_op.add: (1x16x21x21xf32) <- (1x16x21x21xf32, 1x1x1x21xf32) add_33 = paddle._C_ops.add(matmul_27, unsqueeze_0) + del matmul_27 # pd_op.softmax: (1x16x21x21xf32) <- (1x16x21x21xf32) softmax_3 = paddle._C_ops.softmax(add_33, -1) @@ -1193,13 +997,15 @@ def forward( # pd_op.dropout: (1x16x21x21xf32, 1x16x21x21xui8) <- (1x16x21x21xf32, None, 1xf32) dropout_20, dropout_21 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_3, None, full_5, False, "upscale_in_train", 0, False + softmax_3, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_3 # pd_op.matmul: (1x16x21x64xf32) <- (1x16x21x21xf32, 1x16x21x64xf32) matmul_28 = paddle._C_ops.matmul(dropout_20, transpose_14, False, False) + del dropout_20, transpose_14 # pd_op.transpose: (1x21x16x64xf32) <- (1x16x21x64xf32) transpose_15 = paddle._C_ops.transpose(matmul_28, [0, 2, 1, 3]) @@ -1207,19 +1013,20 @@ def forward( # pd_op.reshape: (1x21x1024xf32) <- (1x21x16x64xf32, 3xi64) reshape_15 = paddle._C_ops.reshape(transpose_15, full_int_array_2) + del transpose_15 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_29 = paddle._C_ops.matmul(reshape_15, parameter_267, False, False) - del parameter_267 + del parameter_267, reshape_15 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_34 = paddle._C_ops.add(matmul_29, parameter_266) - del parameter_266 + del matmul_29, parameter_266 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_22, dropout_23 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_34, None, full_5, False, "upscale_in_train", 0, False + add_34, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1227,6 +1034,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_35 = paddle._C_ops.add(layer_norm_18, dropout_22) + del dropout_22, layer_norm_18 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_21, layer_norm_22, layer_norm_23 = (lambda x, f: f(x))( @@ -1235,7 +1043,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_260, parameter_261 + del add_35, parameter_260, parameter_261 # pd_op.matmul: (1x21x4096xf32) <- (1x21x1024xf32, 1024x4096xf32) matmul_30 = paddle._C_ops.matmul(layer_norm_21, parameter_265, False, False) @@ -1243,23 +1051,24 @@ def forward( # pd_op.add: (1x21x4096xf32) <- (1x21x4096xf32, 4096xf32) add_36 = paddle._C_ops.add(matmul_30, parameter_264) - del parameter_264 + del matmul_30, parameter_264 # pd_op.gelu: (1x21x4096xf32) <- (1x21x4096xf32) gelu_3 = paddle._C_ops.gelu(add_36, False) + del add_36 # pd_op.matmul: (1x21x1024xf32) <- (1x21x4096xf32, 4096x1024xf32) matmul_31 = paddle._C_ops.matmul(gelu_3, parameter_263, False, False) - del parameter_263 + del gelu_3, parameter_263 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_37 = paddle._C_ops.add(matmul_31, parameter_262) - del parameter_262 + del matmul_31, parameter_262 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_24, dropout_25 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_37, None, full_5, False, "upscale_in_train", 0, False + add_37, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1267,6 +1076,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_38 = paddle._C_ops.add(layer_norm_21, dropout_24) + del dropout_24, layer_norm_21 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_24, layer_norm_25, layer_norm_26 = (lambda x, f: f(x))( @@ -1275,7 +1085,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_258, parameter_259 + del add_38, parameter_258, parameter_259 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_32 = paddle._C_ops.matmul(layer_norm_24, parameter_257, False, False) @@ -1283,10 +1093,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_39 = paddle._C_ops.add(matmul_32, parameter_256) - del parameter_256 + del matmul_32, parameter_256 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_16 = paddle._C_ops.reshape(add_39, full_int_array_1) + del add_39 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_16 = paddle._C_ops.transpose(reshape_16, [0, 2, 1, 3]) @@ -1298,7 +1109,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_40 = paddle._C_ops.add(matmul_33, parameter_254) - del parameter_254 + del matmul_33, parameter_254 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_34 = paddle._C_ops.matmul(layer_norm_24, parameter_253, False, False) @@ -1306,10 +1117,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_41 = paddle._C_ops.add(matmul_34, parameter_252) - del parameter_252 + del matmul_34, parameter_252 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_17 = paddle._C_ops.reshape(add_40, full_int_array_1) + del add_40 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_17 = paddle._C_ops.transpose(reshape_17, [0, 2, 1, 3]) @@ -1317,6 +1129,7 @@ def forward( # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_18 = paddle._C_ops.reshape(add_41, full_int_array_1) + del add_41 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_18 = paddle._C_ops.transpose(reshape_18, [0, 2, 1, 3]) @@ -1328,9 +1141,11 @@ def forward( # pd_op.matmul: (1x16x21x21xf32) <- (1x16x21x64xf32, 1x16x21x64xf32) matmul_35 = paddle._C_ops.matmul(scale_6, transpose_17, False, True) + del scale_6, transpose_17 # pd_op.add: (1x16x21x21xf32) <- (1x16x21x21xf32, 1x1x1x21xf32) add_42 = paddle._C_ops.add(matmul_35, unsqueeze_0) + del matmul_35 # pd_op.softmax: (1x16x21x21xf32) <- (1x16x21x21xf32) softmax_4 = paddle._C_ops.softmax(add_42, -1) @@ -1339,13 +1154,15 @@ def forward( # pd_op.dropout: (1x16x21x21xf32, 1x16x21x21xui8) <- (1x16x21x21xf32, None, 1xf32) dropout_26, dropout_27 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_4, None, full_5, False, "upscale_in_train", 0, False + softmax_4, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_4 # pd_op.matmul: (1x16x21x64xf32) <- (1x16x21x21xf32, 1x16x21x64xf32) matmul_36 = paddle._C_ops.matmul(dropout_26, transpose_18, False, False) + del dropout_26, transpose_18 # pd_op.transpose: (1x21x16x64xf32) <- (1x16x21x64xf32) transpose_19 = paddle._C_ops.transpose(matmul_36, [0, 2, 1, 3]) @@ -1353,19 +1170,20 @@ def forward( # pd_op.reshape: (1x21x1024xf32) <- (1x21x16x64xf32, 3xi64) reshape_19 = paddle._C_ops.reshape(transpose_19, full_int_array_2) + del transpose_19 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_37 = paddle._C_ops.matmul(reshape_19, parameter_251, False, False) - del parameter_251 + del parameter_251, reshape_19 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_43 = paddle._C_ops.add(matmul_37, parameter_250) - del parameter_250 + del matmul_37, parameter_250 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_28, dropout_29 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_43, None, full_5, False, "upscale_in_train", 0, False + add_43, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1373,6 +1191,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_44 = paddle._C_ops.add(layer_norm_24, dropout_28) + del dropout_28, layer_norm_24 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_27, layer_norm_28, layer_norm_29 = (lambda x, f: f(x))( @@ -1381,7 +1200,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_244, parameter_245 + del add_44, parameter_244, parameter_245 # pd_op.matmul: (1x21x4096xf32) <- (1x21x1024xf32, 1024x4096xf32) matmul_38 = paddle._C_ops.matmul(layer_norm_27, parameter_249, False, False) @@ -1389,23 +1208,24 @@ def forward( # pd_op.add: (1x21x4096xf32) <- (1x21x4096xf32, 4096xf32) add_45 = paddle._C_ops.add(matmul_38, parameter_248) - del parameter_248 + del matmul_38, parameter_248 # pd_op.gelu: (1x21x4096xf32) <- (1x21x4096xf32) gelu_4 = paddle._C_ops.gelu(add_45, False) + del add_45 # pd_op.matmul: (1x21x1024xf32) <- (1x21x4096xf32, 4096x1024xf32) matmul_39 = paddle._C_ops.matmul(gelu_4, parameter_247, False, False) - del parameter_247 + del gelu_4, parameter_247 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_46 = paddle._C_ops.add(matmul_39, parameter_246) - del parameter_246 + del matmul_39, parameter_246 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_30, dropout_31 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_46, None, full_5, False, "upscale_in_train", 0, False + add_46, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1413,6 +1233,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_47 = paddle._C_ops.add(layer_norm_27, dropout_30) + del dropout_30, layer_norm_27 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_30, layer_norm_31, layer_norm_32 = (lambda x, f: f(x))( @@ -1421,7 +1242,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_242, parameter_243 + del add_47, parameter_242, parameter_243 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_40 = paddle._C_ops.matmul(layer_norm_30, parameter_241, False, False) @@ -1429,10 +1250,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_48 = paddle._C_ops.add(matmul_40, parameter_240) - del parameter_240 + del matmul_40, parameter_240 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_20 = paddle._C_ops.reshape(add_48, full_int_array_1) + del add_48 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_20 = paddle._C_ops.transpose(reshape_20, [0, 2, 1, 3]) @@ -1444,7 +1266,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_49 = paddle._C_ops.add(matmul_41, parameter_238) - del parameter_238 + del matmul_41, parameter_238 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_42 = paddle._C_ops.matmul(layer_norm_30, parameter_237, False, False) @@ -1452,10 +1274,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_50 = paddle._C_ops.add(matmul_42, parameter_236) - del parameter_236 + del matmul_42, parameter_236 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_21 = paddle._C_ops.reshape(add_49, full_int_array_1) + del add_49 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_21 = paddle._C_ops.transpose(reshape_21, [0, 2, 1, 3]) @@ -1463,6 +1286,7 @@ def forward( # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_22 = paddle._C_ops.reshape(add_50, full_int_array_1) + del add_50 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_22 = paddle._C_ops.transpose(reshape_22, [0, 2, 1, 3]) @@ -1474,9 +1298,11 @@ def forward( # pd_op.matmul: (1x16x21x21xf32) <- (1x16x21x64xf32, 1x16x21x64xf32) matmul_43 = paddle._C_ops.matmul(scale_7, transpose_21, False, True) + del scale_7, transpose_21 # pd_op.add: (1x16x21x21xf32) <- (1x16x21x21xf32, 1x1x1x21xf32) add_51 = paddle._C_ops.add(matmul_43, unsqueeze_0) + del matmul_43 # pd_op.softmax: (1x16x21x21xf32) <- (1x16x21x21xf32) softmax_5 = paddle._C_ops.softmax(add_51, -1) @@ -1485,13 +1311,15 @@ def forward( # pd_op.dropout: (1x16x21x21xf32, 1x16x21x21xui8) <- (1x16x21x21xf32, None, 1xf32) dropout_32, dropout_33 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_5, None, full_5, False, "upscale_in_train", 0, False + softmax_5, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_5 # pd_op.matmul: (1x16x21x64xf32) <- (1x16x21x21xf32, 1x16x21x64xf32) matmul_44 = paddle._C_ops.matmul(dropout_32, transpose_22, False, False) + del dropout_32, transpose_22 # pd_op.transpose: (1x21x16x64xf32) <- (1x16x21x64xf32) transpose_23 = paddle._C_ops.transpose(matmul_44, [0, 2, 1, 3]) @@ -1499,19 +1327,20 @@ def forward( # pd_op.reshape: (1x21x1024xf32) <- (1x21x16x64xf32, 3xi64) reshape_23 = paddle._C_ops.reshape(transpose_23, full_int_array_2) + del transpose_23 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_45 = paddle._C_ops.matmul(reshape_23, parameter_235, False, False) - del parameter_235 + del parameter_235, reshape_23 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_52 = paddle._C_ops.add(matmul_45, parameter_234) - del parameter_234 + del matmul_45, parameter_234 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_34, dropout_35 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_52, None, full_5, False, "upscale_in_train", 0, False + add_52, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1519,6 +1348,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_53 = paddle._C_ops.add(layer_norm_30, dropout_34) + del dropout_34, layer_norm_30 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_33, layer_norm_34, layer_norm_35 = (lambda x, f: f(x))( @@ -1527,7 +1357,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_228, parameter_229 + del add_53, parameter_228, parameter_229 # pd_op.matmul: (1x21x4096xf32) <- (1x21x1024xf32, 1024x4096xf32) matmul_46 = paddle._C_ops.matmul(layer_norm_33, parameter_233, False, False) @@ -1535,23 +1365,24 @@ def forward( # pd_op.add: (1x21x4096xf32) <- (1x21x4096xf32, 4096xf32) add_54 = paddle._C_ops.add(matmul_46, parameter_232) - del parameter_232 + del matmul_46, parameter_232 # pd_op.gelu: (1x21x4096xf32) <- (1x21x4096xf32) gelu_5 = paddle._C_ops.gelu(add_54, False) + del add_54 # pd_op.matmul: (1x21x1024xf32) <- (1x21x4096xf32, 4096x1024xf32) matmul_47 = paddle._C_ops.matmul(gelu_5, parameter_231, False, False) - del parameter_231 + del gelu_5, parameter_231 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_55 = paddle._C_ops.add(matmul_47, parameter_230) - del parameter_230 + del matmul_47, parameter_230 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_36, dropout_37 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_55, None, full_5, False, "upscale_in_train", 0, False + add_55, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1559,6 +1390,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_56 = paddle._C_ops.add(layer_norm_33, dropout_36) + del dropout_36, layer_norm_33 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_36, layer_norm_37, layer_norm_38 = (lambda x, f: f(x))( @@ -1567,7 +1399,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_226, parameter_227 + del add_56, parameter_226, parameter_227 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_48 = paddle._C_ops.matmul(layer_norm_36, parameter_225, False, False) @@ -1575,10 +1407,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_57 = paddle._C_ops.add(matmul_48, parameter_224) - del parameter_224 + del matmul_48, parameter_224 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_24 = paddle._C_ops.reshape(add_57, full_int_array_1) + del add_57 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_24 = paddle._C_ops.transpose(reshape_24, [0, 2, 1, 3]) @@ -1590,7 +1423,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_58 = paddle._C_ops.add(matmul_49, parameter_222) - del parameter_222 + del matmul_49, parameter_222 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_50 = paddle._C_ops.matmul(layer_norm_36, parameter_221, False, False) @@ -1598,10 +1431,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_59 = paddle._C_ops.add(matmul_50, parameter_220) - del parameter_220 + del matmul_50, parameter_220 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_25 = paddle._C_ops.reshape(add_58, full_int_array_1) + del add_58 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_25 = paddle._C_ops.transpose(reshape_25, [0, 2, 1, 3]) @@ -1609,6 +1443,7 @@ def forward( # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_26 = paddle._C_ops.reshape(add_59, full_int_array_1) + del add_59 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_26 = paddle._C_ops.transpose(reshape_26, [0, 2, 1, 3]) @@ -1620,9 +1455,11 @@ def forward( # pd_op.matmul: (1x16x21x21xf32) <- (1x16x21x64xf32, 1x16x21x64xf32) matmul_51 = paddle._C_ops.matmul(scale_8, transpose_25, False, True) + del scale_8, transpose_25 # pd_op.add: (1x16x21x21xf32) <- (1x16x21x21xf32, 1x1x1x21xf32) add_60 = paddle._C_ops.add(matmul_51, unsqueeze_0) + del matmul_51 # pd_op.softmax: (1x16x21x21xf32) <- (1x16x21x21xf32) softmax_6 = paddle._C_ops.softmax(add_60, -1) @@ -1631,13 +1468,15 @@ def forward( # pd_op.dropout: (1x16x21x21xf32, 1x16x21x21xui8) <- (1x16x21x21xf32, None, 1xf32) dropout_38, dropout_39 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_6, None, full_5, False, "upscale_in_train", 0, False + softmax_6, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_6 # pd_op.matmul: (1x16x21x64xf32) <- (1x16x21x21xf32, 1x16x21x64xf32) matmul_52 = paddle._C_ops.matmul(dropout_38, transpose_26, False, False) + del dropout_38, transpose_26 # pd_op.transpose: (1x21x16x64xf32) <- (1x16x21x64xf32) transpose_27 = paddle._C_ops.transpose(matmul_52, [0, 2, 1, 3]) @@ -1645,19 +1484,20 @@ def forward( # pd_op.reshape: (1x21x1024xf32) <- (1x21x16x64xf32, 3xi64) reshape_27 = paddle._C_ops.reshape(transpose_27, full_int_array_2) + del transpose_27 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_53 = paddle._C_ops.matmul(reshape_27, parameter_219, False, False) - del parameter_219 + del parameter_219, reshape_27 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_61 = paddle._C_ops.add(matmul_53, parameter_218) - del parameter_218 + del matmul_53, parameter_218 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_40, dropout_41 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_61, None, full_5, False, "upscale_in_train", 0, False + add_61, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1665,6 +1505,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_62 = paddle._C_ops.add(layer_norm_36, dropout_40) + del dropout_40, layer_norm_36 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_39, layer_norm_40, layer_norm_41 = (lambda x, f: f(x))( @@ -1673,7 +1514,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_212, parameter_213 + del add_62, parameter_212, parameter_213 # pd_op.matmul: (1x21x4096xf32) <- (1x21x1024xf32, 1024x4096xf32) matmul_54 = paddle._C_ops.matmul(layer_norm_39, parameter_217, False, False) @@ -1681,23 +1522,24 @@ def forward( # pd_op.add: (1x21x4096xf32) <- (1x21x4096xf32, 4096xf32) add_63 = paddle._C_ops.add(matmul_54, parameter_216) - del parameter_216 + del matmul_54, parameter_216 # pd_op.gelu: (1x21x4096xf32) <- (1x21x4096xf32) gelu_6 = paddle._C_ops.gelu(add_63, False) + del add_63 # pd_op.matmul: (1x21x1024xf32) <- (1x21x4096xf32, 4096x1024xf32) matmul_55 = paddle._C_ops.matmul(gelu_6, parameter_215, False, False) - del parameter_215 + del gelu_6, parameter_215 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_64 = paddle._C_ops.add(matmul_55, parameter_214) - del parameter_214 + del matmul_55, parameter_214 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_42, dropout_43 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_64, None, full_5, False, "upscale_in_train", 0, False + add_64, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1705,6 +1547,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_65 = paddle._C_ops.add(layer_norm_39, dropout_42) + del dropout_42, layer_norm_39 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_42, layer_norm_43, layer_norm_44 = (lambda x, f: f(x))( @@ -1713,7 +1556,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_210, parameter_211 + del add_65, parameter_210, parameter_211 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_56 = paddle._C_ops.matmul(layer_norm_42, parameter_209, False, False) @@ -1721,10 +1564,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_66 = paddle._C_ops.add(matmul_56, parameter_208) - del parameter_208 + del matmul_56, parameter_208 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_28 = paddle._C_ops.reshape(add_66, full_int_array_1) + del add_66 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_28 = paddle._C_ops.transpose(reshape_28, [0, 2, 1, 3]) @@ -1736,7 +1580,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_67 = paddle._C_ops.add(matmul_57, parameter_206) - del parameter_206 + del matmul_57, parameter_206 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_58 = paddle._C_ops.matmul(layer_norm_42, parameter_205, False, False) @@ -1744,10 +1588,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_68 = paddle._C_ops.add(matmul_58, parameter_204) - del parameter_204 + del matmul_58, parameter_204 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_29 = paddle._C_ops.reshape(add_67, full_int_array_1) + del add_67 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_29 = paddle._C_ops.transpose(reshape_29, [0, 2, 1, 3]) @@ -1755,6 +1600,7 @@ def forward( # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_30 = paddle._C_ops.reshape(add_68, full_int_array_1) + del add_68 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_30 = paddle._C_ops.transpose(reshape_30, [0, 2, 1, 3]) @@ -1766,9 +1612,11 @@ def forward( # pd_op.matmul: (1x16x21x21xf32) <- (1x16x21x64xf32, 1x16x21x64xf32) matmul_59 = paddle._C_ops.matmul(scale_9, transpose_29, False, True) + del scale_9, transpose_29 # pd_op.add: (1x16x21x21xf32) <- (1x16x21x21xf32, 1x1x1x21xf32) add_69 = paddle._C_ops.add(matmul_59, unsqueeze_0) + del matmul_59 # pd_op.softmax: (1x16x21x21xf32) <- (1x16x21x21xf32) softmax_7 = paddle._C_ops.softmax(add_69, -1) @@ -1777,13 +1625,15 @@ def forward( # pd_op.dropout: (1x16x21x21xf32, 1x16x21x21xui8) <- (1x16x21x21xf32, None, 1xf32) dropout_44, dropout_45 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_7, None, full_5, False, "upscale_in_train", 0, False + softmax_7, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_7 # pd_op.matmul: (1x16x21x64xf32) <- (1x16x21x21xf32, 1x16x21x64xf32) matmul_60 = paddle._C_ops.matmul(dropout_44, transpose_30, False, False) + del dropout_44, transpose_30 # pd_op.transpose: (1x21x16x64xf32) <- (1x16x21x64xf32) transpose_31 = paddle._C_ops.transpose(matmul_60, [0, 2, 1, 3]) @@ -1791,19 +1641,20 @@ def forward( # pd_op.reshape: (1x21x1024xf32) <- (1x21x16x64xf32, 3xi64) reshape_31 = paddle._C_ops.reshape(transpose_31, full_int_array_2) + del transpose_31 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_61 = paddle._C_ops.matmul(reshape_31, parameter_203, False, False) - del parameter_203 + del parameter_203, reshape_31 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_70 = paddle._C_ops.add(matmul_61, parameter_202) - del parameter_202 + del matmul_61, parameter_202 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_46, dropout_47 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_70, None, full_5, False, "upscale_in_train", 0, False + add_70, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1811,6 +1662,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_71 = paddle._C_ops.add(layer_norm_42, dropout_46) + del dropout_46, layer_norm_42 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_45, layer_norm_46, layer_norm_47 = (lambda x, f: f(x))( @@ -1819,7 +1671,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_196, parameter_197 + del add_71, parameter_196, parameter_197 # pd_op.matmul: (1x21x4096xf32) <- (1x21x1024xf32, 1024x4096xf32) matmul_62 = paddle._C_ops.matmul(layer_norm_45, parameter_201, False, False) @@ -1827,23 +1679,24 @@ def forward( # pd_op.add: (1x21x4096xf32) <- (1x21x4096xf32, 4096xf32) add_72 = paddle._C_ops.add(matmul_62, parameter_200) - del parameter_200 + del matmul_62, parameter_200 # pd_op.gelu: (1x21x4096xf32) <- (1x21x4096xf32) gelu_7 = paddle._C_ops.gelu(add_72, False) + del add_72 # pd_op.matmul: (1x21x1024xf32) <- (1x21x4096xf32, 4096x1024xf32) matmul_63 = paddle._C_ops.matmul(gelu_7, parameter_199, False, False) - del parameter_199 + del gelu_7, parameter_199 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_73 = paddle._C_ops.add(matmul_63, parameter_198) - del parameter_198 + del matmul_63, parameter_198 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_48, dropout_49 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_73, None, full_5, False, "upscale_in_train", 0, False + add_73, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1851,6 +1704,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_74 = paddle._C_ops.add(layer_norm_45, dropout_48) + del dropout_48, layer_norm_45 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_48, layer_norm_49, layer_norm_50 = (lambda x, f: f(x))( @@ -1859,7 +1713,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_194, parameter_195 + del add_74, parameter_194, parameter_195 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_64 = paddle._C_ops.matmul(layer_norm_48, parameter_193, False, False) @@ -1867,10 +1721,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_75 = paddle._C_ops.add(matmul_64, parameter_192) - del parameter_192 + del matmul_64, parameter_192 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_32 = paddle._C_ops.reshape(add_75, full_int_array_1) + del add_75 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_32 = paddle._C_ops.transpose(reshape_32, [0, 2, 1, 3]) @@ -1882,7 +1737,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_76 = paddle._C_ops.add(matmul_65, parameter_190) - del parameter_190 + del matmul_65, parameter_190 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_66 = paddle._C_ops.matmul(layer_norm_48, parameter_189, False, False) @@ -1890,10 +1745,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_77 = paddle._C_ops.add(matmul_66, parameter_188) - del parameter_188 + del matmul_66, parameter_188 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_33 = paddle._C_ops.reshape(add_76, full_int_array_1) + del add_76 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_33 = paddle._C_ops.transpose(reshape_33, [0, 2, 1, 3]) @@ -1901,6 +1757,7 @@ def forward( # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_34 = paddle._C_ops.reshape(add_77, full_int_array_1) + del add_77 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_34 = paddle._C_ops.transpose(reshape_34, [0, 2, 1, 3]) @@ -1912,9 +1769,11 @@ def forward( # pd_op.matmul: (1x16x21x21xf32) <- (1x16x21x64xf32, 1x16x21x64xf32) matmul_67 = paddle._C_ops.matmul(scale_10, transpose_33, False, True) + del scale_10, transpose_33 # pd_op.add: (1x16x21x21xf32) <- (1x16x21x21xf32, 1x1x1x21xf32) add_78 = paddle._C_ops.add(matmul_67, unsqueeze_0) + del matmul_67 # pd_op.softmax: (1x16x21x21xf32) <- (1x16x21x21xf32) softmax_8 = paddle._C_ops.softmax(add_78, -1) @@ -1923,13 +1782,15 @@ def forward( # pd_op.dropout: (1x16x21x21xf32, 1x16x21x21xui8) <- (1x16x21x21xf32, None, 1xf32) dropout_50, dropout_51 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_8, None, full_5, False, "upscale_in_train", 0, False + softmax_8, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_8 # pd_op.matmul: (1x16x21x64xf32) <- (1x16x21x21xf32, 1x16x21x64xf32) matmul_68 = paddle._C_ops.matmul(dropout_50, transpose_34, False, False) + del dropout_50, transpose_34 # pd_op.transpose: (1x21x16x64xf32) <- (1x16x21x64xf32) transpose_35 = paddle._C_ops.transpose(matmul_68, [0, 2, 1, 3]) @@ -1937,19 +1798,20 @@ def forward( # pd_op.reshape: (1x21x1024xf32) <- (1x21x16x64xf32, 3xi64) reshape_35 = paddle._C_ops.reshape(transpose_35, full_int_array_2) + del transpose_35 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_69 = paddle._C_ops.matmul(reshape_35, parameter_187, False, False) - del parameter_187 + del parameter_187, reshape_35 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_79 = paddle._C_ops.add(matmul_69, parameter_186) - del parameter_186 + del matmul_69, parameter_186 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_52, dropout_53 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_79, None, full_5, False, "upscale_in_train", 0, False + add_79, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1957,6 +1819,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_80 = paddle._C_ops.add(layer_norm_48, dropout_52) + del dropout_52, layer_norm_48 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_51, layer_norm_52, layer_norm_53 = (lambda x, f: f(x))( @@ -1965,7 +1828,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_180, parameter_181 + del add_80, parameter_180, parameter_181 # pd_op.matmul: (1x21x4096xf32) <- (1x21x1024xf32, 1024x4096xf32) matmul_70 = paddle._C_ops.matmul(layer_norm_51, parameter_185, False, False) @@ -1973,23 +1836,24 @@ def forward( # pd_op.add: (1x21x4096xf32) <- (1x21x4096xf32, 4096xf32) add_81 = paddle._C_ops.add(matmul_70, parameter_184) - del parameter_184 + del matmul_70, parameter_184 # pd_op.gelu: (1x21x4096xf32) <- (1x21x4096xf32) gelu_8 = paddle._C_ops.gelu(add_81, False) + del add_81 # pd_op.matmul: (1x21x1024xf32) <- (1x21x4096xf32, 4096x1024xf32) matmul_71 = paddle._C_ops.matmul(gelu_8, parameter_183, False, False) - del parameter_183 + del gelu_8, parameter_183 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_82 = paddle._C_ops.add(matmul_71, parameter_182) - del parameter_182 + del matmul_71, parameter_182 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_54, dropout_55 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_82, None, full_5, False, "upscale_in_train", 0, False + add_82, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -1997,6 +1861,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_83 = paddle._C_ops.add(layer_norm_51, dropout_54) + del dropout_54, layer_norm_51 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_54, layer_norm_55, layer_norm_56 = (lambda x, f: f(x))( @@ -2005,7 +1870,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_178, parameter_179 + del add_83, parameter_178, parameter_179 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_72 = paddle._C_ops.matmul(layer_norm_54, parameter_177, False, False) @@ -2013,10 +1878,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_84 = paddle._C_ops.add(matmul_72, parameter_176) - del parameter_176 + del matmul_72, parameter_176 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_36 = paddle._C_ops.reshape(add_84, full_int_array_1) + del add_84 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_36 = paddle._C_ops.transpose(reshape_36, [0, 2, 1, 3]) @@ -2028,7 +1894,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_85 = paddle._C_ops.add(matmul_73, parameter_174) - del parameter_174 + del matmul_73, parameter_174 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_74 = paddle._C_ops.matmul(layer_norm_54, parameter_173, False, False) @@ -2036,10 +1902,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_86 = paddle._C_ops.add(matmul_74, parameter_172) - del parameter_172 + del matmul_74, parameter_172 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_37 = paddle._C_ops.reshape(add_85, full_int_array_1) + del add_85 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_37 = paddle._C_ops.transpose(reshape_37, [0, 2, 1, 3]) @@ -2047,6 +1914,7 @@ def forward( # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_38 = paddle._C_ops.reshape(add_86, full_int_array_1) + del add_86 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_38 = paddle._C_ops.transpose(reshape_38, [0, 2, 1, 3]) @@ -2058,9 +1926,11 @@ def forward( # pd_op.matmul: (1x16x21x21xf32) <- (1x16x21x64xf32, 1x16x21x64xf32) matmul_75 = paddle._C_ops.matmul(scale_11, transpose_37, False, True) + del scale_11, transpose_37 # pd_op.add: (1x16x21x21xf32) <- (1x16x21x21xf32, 1x1x1x21xf32) add_87 = paddle._C_ops.add(matmul_75, unsqueeze_0) + del matmul_75 # pd_op.softmax: (1x16x21x21xf32) <- (1x16x21x21xf32) softmax_9 = paddle._C_ops.softmax(add_87, -1) @@ -2069,13 +1939,15 @@ def forward( # pd_op.dropout: (1x16x21x21xf32, 1x16x21x21xui8) <- (1x16x21x21xf32, None, 1xf32) dropout_56, dropout_57 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_9, None, full_5, False, "upscale_in_train", 0, False + softmax_9, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_9 # pd_op.matmul: (1x16x21x64xf32) <- (1x16x21x21xf32, 1x16x21x64xf32) matmul_76 = paddle._C_ops.matmul(dropout_56, transpose_38, False, False) + del dropout_56, transpose_38 # pd_op.transpose: (1x21x16x64xf32) <- (1x16x21x64xf32) transpose_39 = paddle._C_ops.transpose(matmul_76, [0, 2, 1, 3]) @@ -2083,19 +1955,20 @@ def forward( # pd_op.reshape: (1x21x1024xf32) <- (1x21x16x64xf32, 3xi64) reshape_39 = paddle._C_ops.reshape(transpose_39, full_int_array_2) + del transpose_39 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_77 = paddle._C_ops.matmul(reshape_39, parameter_171, False, False) - del parameter_171 + del parameter_171, reshape_39 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_88 = paddle._C_ops.add(matmul_77, parameter_170) - del parameter_170 + del matmul_77, parameter_170 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_58, dropout_59 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_88, None, full_5, False, "upscale_in_train", 0, False + add_88, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -2103,6 +1976,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_89 = paddle._C_ops.add(layer_norm_54, dropout_58) + del dropout_58, layer_norm_54 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_57, layer_norm_58, layer_norm_59 = (lambda x, f: f(x))( @@ -2111,7 +1985,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_164, parameter_165 + del add_89, parameter_164, parameter_165 # pd_op.matmul: (1x21x4096xf32) <- (1x21x1024xf32, 1024x4096xf32) matmul_78 = paddle._C_ops.matmul(layer_norm_57, parameter_169, False, False) @@ -2119,23 +1993,24 @@ def forward( # pd_op.add: (1x21x4096xf32) <- (1x21x4096xf32, 4096xf32) add_90 = paddle._C_ops.add(matmul_78, parameter_168) - del parameter_168 + del matmul_78, parameter_168 # pd_op.gelu: (1x21x4096xf32) <- (1x21x4096xf32) gelu_9 = paddle._C_ops.gelu(add_90, False) + del add_90 # pd_op.matmul: (1x21x1024xf32) <- (1x21x4096xf32, 4096x1024xf32) matmul_79 = paddle._C_ops.matmul(gelu_9, parameter_167, False, False) - del parameter_167 + del gelu_9, parameter_167 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_91 = paddle._C_ops.add(matmul_79, parameter_166) - del parameter_166 + del matmul_79, parameter_166 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_60, dropout_61 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_91, None, full_5, False, "upscale_in_train", 0, False + add_91, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -2143,6 +2018,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_92 = paddle._C_ops.add(layer_norm_57, dropout_60) + del dropout_60, layer_norm_57 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_60, layer_norm_61, layer_norm_62 = (lambda x, f: f(x))( @@ -2151,7 +2027,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_162, parameter_163 + del add_92, parameter_162, parameter_163 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_80 = paddle._C_ops.matmul(layer_norm_60, parameter_161, False, False) @@ -2159,10 +2035,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_93 = paddle._C_ops.add(matmul_80, parameter_160) - del parameter_160 + del matmul_80, parameter_160 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_40 = paddle._C_ops.reshape(add_93, full_int_array_1) + del add_93 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_40 = paddle._C_ops.transpose(reshape_40, [0, 2, 1, 3]) @@ -2174,7 +2051,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_94 = paddle._C_ops.add(matmul_81, parameter_158) - del parameter_158 + del matmul_81, parameter_158 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_82 = paddle._C_ops.matmul(layer_norm_60, parameter_157, False, False) @@ -2182,10 +2059,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_95 = paddle._C_ops.add(matmul_82, parameter_156) - del parameter_156 + del matmul_82, parameter_156 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_41 = paddle._C_ops.reshape(add_94, full_int_array_1) + del add_94 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_41 = paddle._C_ops.transpose(reshape_41, [0, 2, 1, 3]) @@ -2193,6 +2071,7 @@ def forward( # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_42 = paddle._C_ops.reshape(add_95, full_int_array_1) + del add_95 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_42 = paddle._C_ops.transpose(reshape_42, [0, 2, 1, 3]) @@ -2204,9 +2083,11 @@ def forward( # pd_op.matmul: (1x16x21x21xf32) <- (1x16x21x64xf32, 1x16x21x64xf32) matmul_83 = paddle._C_ops.matmul(scale_12, transpose_41, False, True) + del scale_12, transpose_41 # pd_op.add: (1x16x21x21xf32) <- (1x16x21x21xf32, 1x1x1x21xf32) add_96 = paddle._C_ops.add(matmul_83, unsqueeze_0) + del matmul_83 # pd_op.softmax: (1x16x21x21xf32) <- (1x16x21x21xf32) softmax_10 = paddle._C_ops.softmax(add_96, -1) @@ -2215,13 +2096,15 @@ def forward( # pd_op.dropout: (1x16x21x21xf32, 1x16x21x21xui8) <- (1x16x21x21xf32, None, 1xf32) dropout_62, dropout_63 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_10, None, full_5, False, "upscale_in_train", 0, False + softmax_10, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_10 # pd_op.matmul: (1x16x21x64xf32) <- (1x16x21x21xf32, 1x16x21x64xf32) matmul_84 = paddle._C_ops.matmul(dropout_62, transpose_42, False, False) + del dropout_62, transpose_42 # pd_op.transpose: (1x21x16x64xf32) <- (1x16x21x64xf32) transpose_43 = paddle._C_ops.transpose(matmul_84, [0, 2, 1, 3]) @@ -2229,19 +2112,20 @@ def forward( # pd_op.reshape: (1x21x1024xf32) <- (1x21x16x64xf32, 3xi64) reshape_43 = paddle._C_ops.reshape(transpose_43, full_int_array_2) + del transpose_43 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_85 = paddle._C_ops.matmul(reshape_43, parameter_155, False, False) - del parameter_155 + del parameter_155, reshape_43 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_97 = paddle._C_ops.add(matmul_85, parameter_154) - del parameter_154 + del matmul_85, parameter_154 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_64, dropout_65 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_97, None, full_5, False, "upscale_in_train", 0, False + add_97, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -2249,6 +2133,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_98 = paddle._C_ops.add(layer_norm_60, dropout_64) + del dropout_64, layer_norm_60 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_63, layer_norm_64, layer_norm_65 = (lambda x, f: f(x))( @@ -2257,7 +2142,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_148, parameter_149 + del add_98, parameter_148, parameter_149 # pd_op.matmul: (1x21x4096xf32) <- (1x21x1024xf32, 1024x4096xf32) matmul_86 = paddle._C_ops.matmul(layer_norm_63, parameter_153, False, False) @@ -2265,23 +2150,24 @@ def forward( # pd_op.add: (1x21x4096xf32) <- (1x21x4096xf32, 4096xf32) add_99 = paddle._C_ops.add(matmul_86, parameter_152) - del parameter_152 + del matmul_86, parameter_152 # pd_op.gelu: (1x21x4096xf32) <- (1x21x4096xf32) gelu_10 = paddle._C_ops.gelu(add_99, False) + del add_99 # pd_op.matmul: (1x21x1024xf32) <- (1x21x4096xf32, 4096x1024xf32) matmul_87 = paddle._C_ops.matmul(gelu_10, parameter_151, False, False) - del parameter_151 + del gelu_10, parameter_151 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_100 = paddle._C_ops.add(matmul_87, parameter_150) - del parameter_150 + del matmul_87, parameter_150 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_66, dropout_67 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_100, None, full_5, False, "upscale_in_train", 0, False + add_100, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -2289,6 +2175,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_101 = paddle._C_ops.add(layer_norm_63, dropout_66) + del dropout_66, layer_norm_63 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_66, layer_norm_67, layer_norm_68 = (lambda x, f: f(x))( @@ -2297,7 +2184,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_146, parameter_147 + del add_101, parameter_146, parameter_147 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_88 = paddle._C_ops.matmul(layer_norm_66, parameter_145, False, False) @@ -2305,10 +2192,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_102 = paddle._C_ops.add(matmul_88, parameter_144) - del parameter_144 + del matmul_88, parameter_144 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_44 = paddle._C_ops.reshape(add_102, full_int_array_1) + del add_102 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_44 = paddle._C_ops.transpose(reshape_44, [0, 2, 1, 3]) @@ -2320,7 +2208,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_103 = paddle._C_ops.add(matmul_89, parameter_142) - del parameter_142 + del matmul_89, parameter_142 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_90 = paddle._C_ops.matmul(layer_norm_66, parameter_141, False, False) @@ -2328,10 +2216,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_104 = paddle._C_ops.add(matmul_90, parameter_140) - del parameter_140 + del matmul_90, parameter_140 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_45 = paddle._C_ops.reshape(add_103, full_int_array_1) + del add_103 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_45 = paddle._C_ops.transpose(reshape_45, [0, 2, 1, 3]) @@ -2339,6 +2228,7 @@ def forward( # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_46 = paddle._C_ops.reshape(add_104, full_int_array_1) + del add_104 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_46 = paddle._C_ops.transpose(reshape_46, [0, 2, 1, 3]) @@ -2350,9 +2240,11 @@ def forward( # pd_op.matmul: (1x16x21x21xf32) <- (1x16x21x64xf32, 1x16x21x64xf32) matmul_91 = paddle._C_ops.matmul(scale_13, transpose_45, False, True) + del scale_13, transpose_45 # pd_op.add: (1x16x21x21xf32) <- (1x16x21x21xf32, 1x1x1x21xf32) add_105 = paddle._C_ops.add(matmul_91, unsqueeze_0) + del matmul_91 # pd_op.softmax: (1x16x21x21xf32) <- (1x16x21x21xf32) softmax_11 = paddle._C_ops.softmax(add_105, -1) @@ -2361,13 +2253,15 @@ def forward( # pd_op.dropout: (1x16x21x21xf32, 1x16x21x21xui8) <- (1x16x21x21xf32, None, 1xf32) dropout_68, dropout_69 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_11, None, full_5, False, "upscale_in_train", 0, False + softmax_11, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_11 # pd_op.matmul: (1x16x21x64xf32) <- (1x16x21x21xf32, 1x16x21x64xf32) matmul_92 = paddle._C_ops.matmul(dropout_68, transpose_46, False, False) + del dropout_68, transpose_46 # pd_op.transpose: (1x21x16x64xf32) <- (1x16x21x64xf32) transpose_47 = paddle._C_ops.transpose(matmul_92, [0, 2, 1, 3]) @@ -2375,19 +2269,20 @@ def forward( # pd_op.reshape: (1x21x1024xf32) <- (1x21x16x64xf32, 3xi64) reshape_47 = paddle._C_ops.reshape(transpose_47, full_int_array_2) + del transpose_47 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_93 = paddle._C_ops.matmul(reshape_47, parameter_139, False, False) - del parameter_139 + del parameter_139, reshape_47 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_106 = paddle._C_ops.add(matmul_93, parameter_138) - del parameter_138 + del matmul_93, parameter_138 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_70, dropout_71 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_106, None, full_5, False, "upscale_in_train", 0, False + add_106, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -2395,6 +2290,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_107 = paddle._C_ops.add(layer_norm_66, dropout_70) + del dropout_70, layer_norm_66 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_69, layer_norm_70, layer_norm_71 = (lambda x, f: f(x))( @@ -2403,7 +2299,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_132, parameter_133 + del add_107, parameter_132, parameter_133 # pd_op.matmul: (1x21x4096xf32) <- (1x21x1024xf32, 1024x4096xf32) matmul_94 = paddle._C_ops.matmul(layer_norm_69, parameter_137, False, False) @@ -2411,23 +2307,24 @@ def forward( # pd_op.add: (1x21x4096xf32) <- (1x21x4096xf32, 4096xf32) add_108 = paddle._C_ops.add(matmul_94, parameter_136) - del parameter_136 + del matmul_94, parameter_136 # pd_op.gelu: (1x21x4096xf32) <- (1x21x4096xf32) gelu_11 = paddle._C_ops.gelu(add_108, False) + del add_108 # pd_op.matmul: (1x21x1024xf32) <- (1x21x4096xf32, 4096x1024xf32) matmul_95 = paddle._C_ops.matmul(gelu_11, parameter_135, False, False) - del parameter_135 + del gelu_11, parameter_135 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_109 = paddle._C_ops.add(matmul_95, parameter_134) - del parameter_134 + del matmul_95, parameter_134 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_72, dropout_73 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_109, None, full_5, False, "upscale_in_train", 0, False + add_109, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -2435,6 +2332,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_110 = paddle._C_ops.add(layer_norm_69, dropout_72) + del dropout_72, layer_norm_69 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_72, layer_norm_73, layer_norm_74 = (lambda x, f: f(x))( @@ -2443,7 +2341,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_130, parameter_131 + del add_110, parameter_130, parameter_131 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_96 = paddle._C_ops.matmul(layer_norm_72, parameter_129, False, False) @@ -2451,10 +2349,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_111 = paddle._C_ops.add(matmul_96, parameter_128) - del parameter_128 + del matmul_96, parameter_128 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_48 = paddle._C_ops.reshape(add_111, full_int_array_1) + del add_111 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_48 = paddle._C_ops.transpose(reshape_48, [0, 2, 1, 3]) @@ -2466,7 +2365,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_112 = paddle._C_ops.add(matmul_97, parameter_126) - del parameter_126 + del matmul_97, parameter_126 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_98 = paddle._C_ops.matmul(layer_norm_72, parameter_125, False, False) @@ -2474,10 +2373,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_113 = paddle._C_ops.add(matmul_98, parameter_124) - del parameter_124 + del matmul_98, parameter_124 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_49 = paddle._C_ops.reshape(add_112, full_int_array_1) + del add_112 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_49 = paddle._C_ops.transpose(reshape_49, [0, 2, 1, 3]) @@ -2485,6 +2385,7 @@ def forward( # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_50 = paddle._C_ops.reshape(add_113, full_int_array_1) + del add_113 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_50 = paddle._C_ops.transpose(reshape_50, [0, 2, 1, 3]) @@ -2496,9 +2397,11 @@ def forward( # pd_op.matmul: (1x16x21x21xf32) <- (1x16x21x64xf32, 1x16x21x64xf32) matmul_99 = paddle._C_ops.matmul(scale_14, transpose_49, False, True) + del scale_14, transpose_49 # pd_op.add: (1x16x21x21xf32) <- (1x16x21x21xf32, 1x1x1x21xf32) add_114 = paddle._C_ops.add(matmul_99, unsqueeze_0) + del matmul_99 # pd_op.softmax: (1x16x21x21xf32) <- (1x16x21x21xf32) softmax_12 = paddle._C_ops.softmax(add_114, -1) @@ -2507,13 +2410,15 @@ def forward( # pd_op.dropout: (1x16x21x21xf32, 1x16x21x21xui8) <- (1x16x21x21xf32, None, 1xf32) dropout_74, dropout_75 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_12, None, full_5, False, "upscale_in_train", 0, False + softmax_12, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_12 # pd_op.matmul: (1x16x21x64xf32) <- (1x16x21x21xf32, 1x16x21x64xf32) matmul_100 = paddle._C_ops.matmul(dropout_74, transpose_50, False, False) + del dropout_74, transpose_50 # pd_op.transpose: (1x21x16x64xf32) <- (1x16x21x64xf32) transpose_51 = paddle._C_ops.transpose(matmul_100, [0, 2, 1, 3]) @@ -2521,19 +2426,20 @@ def forward( # pd_op.reshape: (1x21x1024xf32) <- (1x21x16x64xf32, 3xi64) reshape_51 = paddle._C_ops.reshape(transpose_51, full_int_array_2) + del transpose_51 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_101 = paddle._C_ops.matmul(reshape_51, parameter_123, False, False) - del parameter_123 + del parameter_123, reshape_51 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_115 = paddle._C_ops.add(matmul_101, parameter_122) - del parameter_122 + del matmul_101, parameter_122 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_76, dropout_77 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_115, None, full_5, False, "upscale_in_train", 0, False + add_115, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -2541,6 +2447,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_116 = paddle._C_ops.add(layer_norm_72, dropout_76) + del dropout_76, layer_norm_72 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_75, layer_norm_76, layer_norm_77 = (lambda x, f: f(x))( @@ -2549,7 +2456,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_116, parameter_117 + del add_116, parameter_116, parameter_117 # pd_op.matmul: (1x21x4096xf32) <- (1x21x1024xf32, 1024x4096xf32) matmul_102 = paddle._C_ops.matmul(layer_norm_75, parameter_121, False, False) @@ -2557,23 +2464,24 @@ def forward( # pd_op.add: (1x21x4096xf32) <- (1x21x4096xf32, 4096xf32) add_117 = paddle._C_ops.add(matmul_102, parameter_120) - del parameter_120 + del matmul_102, parameter_120 # pd_op.gelu: (1x21x4096xf32) <- (1x21x4096xf32) gelu_12 = paddle._C_ops.gelu(add_117, False) + del add_117 # pd_op.matmul: (1x21x1024xf32) <- (1x21x4096xf32, 4096x1024xf32) matmul_103 = paddle._C_ops.matmul(gelu_12, parameter_119, False, False) - del parameter_119 + del gelu_12, parameter_119 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_118 = paddle._C_ops.add(matmul_103, parameter_118) - del parameter_118 + del matmul_103, parameter_118 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_78, dropout_79 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_118, None, full_5, False, "upscale_in_train", 0, False + add_118, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -2581,6 +2489,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_119 = paddle._C_ops.add(layer_norm_75, dropout_78) + del dropout_78, layer_norm_75 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_78, layer_norm_79, layer_norm_80 = (lambda x, f: f(x))( @@ -2589,7 +2498,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_114, parameter_115 + del add_119, parameter_114, parameter_115 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_104 = paddle._C_ops.matmul(layer_norm_78, parameter_113, False, False) @@ -2597,10 +2506,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_120 = paddle._C_ops.add(matmul_104, parameter_112) - del parameter_112 + del matmul_104, parameter_112 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_52 = paddle._C_ops.reshape(add_120, full_int_array_1) + del add_120 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_52 = paddle._C_ops.transpose(reshape_52, [0, 2, 1, 3]) @@ -2612,7 +2522,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_121 = paddle._C_ops.add(matmul_105, parameter_110) - del parameter_110 + del matmul_105, parameter_110 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_106 = paddle._C_ops.matmul(layer_norm_78, parameter_109, False, False) @@ -2620,10 +2530,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_122 = paddle._C_ops.add(matmul_106, parameter_108) - del parameter_108 + del matmul_106, parameter_108 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_53 = paddle._C_ops.reshape(add_121, full_int_array_1) + del add_121 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_53 = paddle._C_ops.transpose(reshape_53, [0, 2, 1, 3]) @@ -2631,6 +2542,7 @@ def forward( # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_54 = paddle._C_ops.reshape(add_122, full_int_array_1) + del add_122 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_54 = paddle._C_ops.transpose(reshape_54, [0, 2, 1, 3]) @@ -2642,9 +2554,11 @@ def forward( # pd_op.matmul: (1x16x21x21xf32) <- (1x16x21x64xf32, 1x16x21x64xf32) matmul_107 = paddle._C_ops.matmul(scale_15, transpose_53, False, True) + del scale_15, transpose_53 # pd_op.add: (1x16x21x21xf32) <- (1x16x21x21xf32, 1x1x1x21xf32) add_123 = paddle._C_ops.add(matmul_107, unsqueeze_0) + del matmul_107 # pd_op.softmax: (1x16x21x21xf32) <- (1x16x21x21xf32) softmax_13 = paddle._C_ops.softmax(add_123, -1) @@ -2653,13 +2567,15 @@ def forward( # pd_op.dropout: (1x16x21x21xf32, 1x16x21x21xui8) <- (1x16x21x21xf32, None, 1xf32) dropout_80, dropout_81 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_13, None, full_5, False, "upscale_in_train", 0, False + softmax_13, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_13 # pd_op.matmul: (1x16x21x64xf32) <- (1x16x21x21xf32, 1x16x21x64xf32) matmul_108 = paddle._C_ops.matmul(dropout_80, transpose_54, False, False) + del dropout_80, transpose_54 # pd_op.transpose: (1x21x16x64xf32) <- (1x16x21x64xf32) transpose_55 = paddle._C_ops.transpose(matmul_108, [0, 2, 1, 3]) @@ -2667,19 +2583,20 @@ def forward( # pd_op.reshape: (1x21x1024xf32) <- (1x21x16x64xf32, 3xi64) reshape_55 = paddle._C_ops.reshape(transpose_55, full_int_array_2) + del transpose_55 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_109 = paddle._C_ops.matmul(reshape_55, parameter_107, False, False) - del parameter_107 + del parameter_107, reshape_55 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_124 = paddle._C_ops.add(matmul_109, parameter_106) - del parameter_106 + del matmul_109, parameter_106 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_82, dropout_83 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_124, None, full_5, False, "upscale_in_train", 0, False + add_124, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -2687,6 +2604,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_125 = paddle._C_ops.add(layer_norm_78, dropout_82) + del dropout_82, layer_norm_78 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_81, layer_norm_82, layer_norm_83 = (lambda x, f: f(x))( @@ -2695,7 +2613,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_100, parameter_101 + del add_125, parameter_100, parameter_101 # pd_op.matmul: (1x21x4096xf32) <- (1x21x1024xf32, 1024x4096xf32) matmul_110 = paddle._C_ops.matmul(layer_norm_81, parameter_105, False, False) @@ -2703,23 +2621,24 @@ def forward( # pd_op.add: (1x21x4096xf32) <- (1x21x4096xf32, 4096xf32) add_126 = paddle._C_ops.add(matmul_110, parameter_104) - del parameter_104 + del matmul_110, parameter_104 # pd_op.gelu: (1x21x4096xf32) <- (1x21x4096xf32) gelu_13 = paddle._C_ops.gelu(add_126, False) + del add_126 # pd_op.matmul: (1x21x1024xf32) <- (1x21x4096xf32, 4096x1024xf32) matmul_111 = paddle._C_ops.matmul(gelu_13, parameter_103, False, False) - del parameter_103 + del gelu_13, parameter_103 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_127 = paddle._C_ops.add(matmul_111, parameter_102) - del parameter_102 + del matmul_111, parameter_102 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_84, dropout_85 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_127, None, full_5, False, "upscale_in_train", 0, False + add_127, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -2727,6 +2646,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_128 = paddle._C_ops.add(layer_norm_81, dropout_84) + del dropout_84, layer_norm_81 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_84, layer_norm_85, layer_norm_86 = (lambda x, f: f(x))( @@ -2735,7 +2655,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_98, parameter_99 + del add_128, parameter_98, parameter_99 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_112 = paddle._C_ops.matmul(layer_norm_84, parameter_97, False, False) @@ -2743,10 +2663,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_129 = paddle._C_ops.add(matmul_112, parameter_96) - del parameter_96 + del matmul_112, parameter_96 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_56 = paddle._C_ops.reshape(add_129, full_int_array_1) + del add_129 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_56 = paddle._C_ops.transpose(reshape_56, [0, 2, 1, 3]) @@ -2758,7 +2679,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_130 = paddle._C_ops.add(matmul_113, parameter_94) - del parameter_94 + del matmul_113, parameter_94 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_114 = paddle._C_ops.matmul(layer_norm_84, parameter_93, False, False) @@ -2766,10 +2687,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_131 = paddle._C_ops.add(matmul_114, parameter_92) - del parameter_92 + del matmul_114, parameter_92 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_57 = paddle._C_ops.reshape(add_130, full_int_array_1) + del add_130 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_57 = paddle._C_ops.transpose(reshape_57, [0, 2, 1, 3]) @@ -2777,6 +2699,7 @@ def forward( # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_58 = paddle._C_ops.reshape(add_131, full_int_array_1) + del add_131 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_58 = paddle._C_ops.transpose(reshape_58, [0, 2, 1, 3]) @@ -2788,9 +2711,11 @@ def forward( # pd_op.matmul: (1x16x21x21xf32) <- (1x16x21x64xf32, 1x16x21x64xf32) matmul_115 = paddle._C_ops.matmul(scale_16, transpose_57, False, True) + del scale_16, transpose_57 # pd_op.add: (1x16x21x21xf32) <- (1x16x21x21xf32, 1x1x1x21xf32) add_132 = paddle._C_ops.add(matmul_115, unsqueeze_0) + del matmul_115 # pd_op.softmax: (1x16x21x21xf32) <- (1x16x21x21xf32) softmax_14 = paddle._C_ops.softmax(add_132, -1) @@ -2799,13 +2724,15 @@ def forward( # pd_op.dropout: (1x16x21x21xf32, 1x16x21x21xui8) <- (1x16x21x21xf32, None, 1xf32) dropout_86, dropout_87 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_14, None, full_5, False, "upscale_in_train", 0, False + softmax_14, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_14 # pd_op.matmul: (1x16x21x64xf32) <- (1x16x21x21xf32, 1x16x21x64xf32) matmul_116 = paddle._C_ops.matmul(dropout_86, transpose_58, False, False) + del dropout_86, transpose_58 # pd_op.transpose: (1x21x16x64xf32) <- (1x16x21x64xf32) transpose_59 = paddle._C_ops.transpose(matmul_116, [0, 2, 1, 3]) @@ -2813,19 +2740,20 @@ def forward( # pd_op.reshape: (1x21x1024xf32) <- (1x21x16x64xf32, 3xi64) reshape_59 = paddle._C_ops.reshape(transpose_59, full_int_array_2) + del transpose_59 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_117 = paddle._C_ops.matmul(reshape_59, parameter_91, False, False) - del parameter_91 + del parameter_91, reshape_59 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_133 = paddle._C_ops.add(matmul_117, parameter_90) - del parameter_90 + del matmul_117, parameter_90 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_88, dropout_89 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_133, None, full_5, False, "upscale_in_train", 0, False + add_133, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -2833,6 +2761,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_134 = paddle._C_ops.add(layer_norm_84, dropout_88) + del dropout_88, layer_norm_84 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_87, layer_norm_88, layer_norm_89 = (lambda x, f: f(x))( @@ -2841,7 +2770,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_84, parameter_85 + del add_134, parameter_84, parameter_85 # pd_op.matmul: (1x21x4096xf32) <- (1x21x1024xf32, 1024x4096xf32) matmul_118 = paddle._C_ops.matmul(layer_norm_87, parameter_89, False, False) @@ -2849,23 +2778,24 @@ def forward( # pd_op.add: (1x21x4096xf32) <- (1x21x4096xf32, 4096xf32) add_135 = paddle._C_ops.add(matmul_118, parameter_88) - del parameter_88 + del matmul_118, parameter_88 # pd_op.gelu: (1x21x4096xf32) <- (1x21x4096xf32) gelu_14 = paddle._C_ops.gelu(add_135, False) + del add_135 # pd_op.matmul: (1x21x1024xf32) <- (1x21x4096xf32, 4096x1024xf32) matmul_119 = paddle._C_ops.matmul(gelu_14, parameter_87, False, False) - del parameter_87 + del gelu_14, parameter_87 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_136 = paddle._C_ops.add(matmul_119, parameter_86) - del parameter_86 + del matmul_119, parameter_86 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_90, dropout_91 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_136, None, full_5, False, "upscale_in_train", 0, False + add_136, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -2873,6 +2803,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_137 = paddle._C_ops.add(layer_norm_87, dropout_90) + del dropout_90, layer_norm_87 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_90, layer_norm_91, layer_norm_92 = (lambda x, f: f(x))( @@ -2881,7 +2812,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_82, parameter_83 + del add_137, parameter_82, parameter_83 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_120 = paddle._C_ops.matmul(layer_norm_90, parameter_81, False, False) @@ -2889,10 +2820,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_138 = paddle._C_ops.add(matmul_120, parameter_80) - del parameter_80 + del matmul_120, parameter_80 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_60 = paddle._C_ops.reshape(add_138, full_int_array_1) + del add_138 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_60 = paddle._C_ops.transpose(reshape_60, [0, 2, 1, 3]) @@ -2904,7 +2836,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_139 = paddle._C_ops.add(matmul_121, parameter_78) - del parameter_78 + del matmul_121, parameter_78 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_122 = paddle._C_ops.matmul(layer_norm_90, parameter_77, False, False) @@ -2912,10 +2844,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_140 = paddle._C_ops.add(matmul_122, parameter_76) - del parameter_76 + del matmul_122, parameter_76 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_61 = paddle._C_ops.reshape(add_139, full_int_array_1) + del add_139 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_61 = paddle._C_ops.transpose(reshape_61, [0, 2, 1, 3]) @@ -2923,6 +2856,7 @@ def forward( # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_62 = paddle._C_ops.reshape(add_140, full_int_array_1) + del add_140 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_62 = paddle._C_ops.transpose(reshape_62, [0, 2, 1, 3]) @@ -2934,9 +2868,11 @@ def forward( # pd_op.matmul: (1x16x21x21xf32) <- (1x16x21x64xf32, 1x16x21x64xf32) matmul_123 = paddle._C_ops.matmul(scale_17, transpose_61, False, True) + del scale_17, transpose_61 # pd_op.add: (1x16x21x21xf32) <- (1x16x21x21xf32, 1x1x1x21xf32) add_141 = paddle._C_ops.add(matmul_123, unsqueeze_0) + del matmul_123 # pd_op.softmax: (1x16x21x21xf32) <- (1x16x21x21xf32) softmax_15 = paddle._C_ops.softmax(add_141, -1) @@ -2945,13 +2881,15 @@ def forward( # pd_op.dropout: (1x16x21x21xf32, 1x16x21x21xui8) <- (1x16x21x21xf32, None, 1xf32) dropout_92, dropout_93 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_15, None, full_5, False, "upscale_in_train", 0, False + softmax_15, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_15 # pd_op.matmul: (1x16x21x64xf32) <- (1x16x21x21xf32, 1x16x21x64xf32) matmul_124 = paddle._C_ops.matmul(dropout_92, transpose_62, False, False) + del dropout_92, transpose_62 # pd_op.transpose: (1x21x16x64xf32) <- (1x16x21x64xf32) transpose_63 = paddle._C_ops.transpose(matmul_124, [0, 2, 1, 3]) @@ -2959,19 +2897,20 @@ def forward( # pd_op.reshape: (1x21x1024xf32) <- (1x21x16x64xf32, 3xi64) reshape_63 = paddle._C_ops.reshape(transpose_63, full_int_array_2) + del transpose_63 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_125 = paddle._C_ops.matmul(reshape_63, parameter_75, False, False) - del parameter_75 + del parameter_75, reshape_63 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_142 = paddle._C_ops.add(matmul_125, parameter_74) - del parameter_74 + del matmul_125, parameter_74 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_94, dropout_95 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_142, None, full_5, False, "upscale_in_train", 0, False + add_142, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -2979,6 +2918,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_143 = paddle._C_ops.add(layer_norm_90, dropout_94) + del dropout_94, layer_norm_90 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_93, layer_norm_94, layer_norm_95 = (lambda x, f: f(x))( @@ -2987,7 +2927,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_68, parameter_69 + del add_143, parameter_68, parameter_69 # pd_op.matmul: (1x21x4096xf32) <- (1x21x1024xf32, 1024x4096xf32) matmul_126 = paddle._C_ops.matmul(layer_norm_93, parameter_73, False, False) @@ -2995,23 +2935,24 @@ def forward( # pd_op.add: (1x21x4096xf32) <- (1x21x4096xf32, 4096xf32) add_144 = paddle._C_ops.add(matmul_126, parameter_72) - del parameter_72 + del matmul_126, parameter_72 # pd_op.gelu: (1x21x4096xf32) <- (1x21x4096xf32) gelu_15 = paddle._C_ops.gelu(add_144, False) + del add_144 # pd_op.matmul: (1x21x1024xf32) <- (1x21x4096xf32, 4096x1024xf32) matmul_127 = paddle._C_ops.matmul(gelu_15, parameter_71, False, False) - del parameter_71 + del gelu_15, parameter_71 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_145 = paddle._C_ops.add(matmul_127, parameter_70) - del parameter_70 + del matmul_127, parameter_70 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_96, dropout_97 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_145, None, full_5, False, "upscale_in_train", 0, False + add_145, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -3019,6 +2960,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_146 = paddle._C_ops.add(layer_norm_93, dropout_96) + del dropout_96, layer_norm_93 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_96, layer_norm_97, layer_norm_98 = (lambda x, f: f(x))( @@ -3027,7 +2969,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_66, parameter_67 + del add_146, parameter_66, parameter_67 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_128 = paddle._C_ops.matmul(layer_norm_96, parameter_65, False, False) @@ -3035,10 +2977,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_147 = paddle._C_ops.add(matmul_128, parameter_64) - del parameter_64 + del matmul_128, parameter_64 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_64 = paddle._C_ops.reshape(add_147, full_int_array_1) + del add_147 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_64 = paddle._C_ops.transpose(reshape_64, [0, 2, 1, 3]) @@ -3050,7 +2993,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_148 = paddle._C_ops.add(matmul_129, parameter_62) - del parameter_62 + del matmul_129, parameter_62 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_130 = paddle._C_ops.matmul(layer_norm_96, parameter_61, False, False) @@ -3058,10 +3001,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_149 = paddle._C_ops.add(matmul_130, parameter_60) - del parameter_60 + del matmul_130, parameter_60 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_65 = paddle._C_ops.reshape(add_148, full_int_array_1) + del add_148 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_65 = paddle._C_ops.transpose(reshape_65, [0, 2, 1, 3]) @@ -3069,6 +3013,7 @@ def forward( # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_66 = paddle._C_ops.reshape(add_149, full_int_array_1) + del add_149 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_66 = paddle._C_ops.transpose(reshape_66, [0, 2, 1, 3]) @@ -3080,9 +3025,11 @@ def forward( # pd_op.matmul: (1x16x21x21xf32) <- (1x16x21x64xf32, 1x16x21x64xf32) matmul_131 = paddle._C_ops.matmul(scale_18, transpose_65, False, True) + del scale_18, transpose_65 # pd_op.add: (1x16x21x21xf32) <- (1x16x21x21xf32, 1x1x1x21xf32) add_150 = paddle._C_ops.add(matmul_131, unsqueeze_0) + del matmul_131 # pd_op.softmax: (1x16x21x21xf32) <- (1x16x21x21xf32) softmax_16 = paddle._C_ops.softmax(add_150, -1) @@ -3091,13 +3038,15 @@ def forward( # pd_op.dropout: (1x16x21x21xf32, 1x16x21x21xui8) <- (1x16x21x21xf32, None, 1xf32) dropout_98, dropout_99 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_16, None, full_5, False, "upscale_in_train", 0, False + softmax_16, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_16 # pd_op.matmul: (1x16x21x64xf32) <- (1x16x21x21xf32, 1x16x21x64xf32) matmul_132 = paddle._C_ops.matmul(dropout_98, transpose_66, False, False) + del dropout_98, transpose_66 # pd_op.transpose: (1x21x16x64xf32) <- (1x16x21x64xf32) transpose_67 = paddle._C_ops.transpose(matmul_132, [0, 2, 1, 3]) @@ -3105,19 +3054,20 @@ def forward( # pd_op.reshape: (1x21x1024xf32) <- (1x21x16x64xf32, 3xi64) reshape_67 = paddle._C_ops.reshape(transpose_67, full_int_array_2) + del transpose_67 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_133 = paddle._C_ops.matmul(reshape_67, parameter_59, False, False) - del parameter_59 + del parameter_59, reshape_67 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_151 = paddle._C_ops.add(matmul_133, parameter_58) - del parameter_58 + del matmul_133, parameter_58 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_100, dropout_101 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_151, None, full_5, False, "upscale_in_train", 0, False + add_151, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -3125,6 +3075,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_152 = paddle._C_ops.add(layer_norm_96, dropout_100) + del dropout_100, layer_norm_96 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_99, layer_norm_100, layer_norm_101 = (lambda x, f: f(x))( @@ -3133,7 +3084,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_52, parameter_53 + del add_152, parameter_52, parameter_53 # pd_op.matmul: (1x21x4096xf32) <- (1x21x1024xf32, 1024x4096xf32) matmul_134 = paddle._C_ops.matmul(layer_norm_99, parameter_57, False, False) @@ -3141,23 +3092,24 @@ def forward( # pd_op.add: (1x21x4096xf32) <- (1x21x4096xf32, 4096xf32) add_153 = paddle._C_ops.add(matmul_134, parameter_56) - del parameter_56 + del matmul_134, parameter_56 # pd_op.gelu: (1x21x4096xf32) <- (1x21x4096xf32) gelu_16 = paddle._C_ops.gelu(add_153, False) + del add_153 # pd_op.matmul: (1x21x1024xf32) <- (1x21x4096xf32, 4096x1024xf32) matmul_135 = paddle._C_ops.matmul(gelu_16, parameter_55, False, False) - del parameter_55 + del gelu_16, parameter_55 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_154 = paddle._C_ops.add(matmul_135, parameter_54) - del parameter_54 + del matmul_135, parameter_54 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_102, dropout_103 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_154, None, full_5, False, "upscale_in_train", 0, False + add_154, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -3165,6 +3117,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_155 = paddle._C_ops.add(layer_norm_99, dropout_102) + del dropout_102, layer_norm_99 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_102, layer_norm_103, layer_norm_104 = (lambda x, f: f(x))( @@ -3173,7 +3126,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_50, parameter_51 + del add_155, parameter_50, parameter_51 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_136 = paddle._C_ops.matmul(layer_norm_102, parameter_49, False, False) @@ -3181,10 +3134,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_156 = paddle._C_ops.add(matmul_136, parameter_48) - del parameter_48 + del matmul_136, parameter_48 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_68 = paddle._C_ops.reshape(add_156, full_int_array_1) + del add_156 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_68 = paddle._C_ops.transpose(reshape_68, [0, 2, 1, 3]) @@ -3196,7 +3150,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_157 = paddle._C_ops.add(matmul_137, parameter_46) - del parameter_46 + del matmul_137, parameter_46 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_138 = paddle._C_ops.matmul(layer_norm_102, parameter_45, False, False) @@ -3204,10 +3158,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_158 = paddle._C_ops.add(matmul_138, parameter_44) - del parameter_44 + del matmul_138, parameter_44 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_69 = paddle._C_ops.reshape(add_157, full_int_array_1) + del add_157 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_69 = paddle._C_ops.transpose(reshape_69, [0, 2, 1, 3]) @@ -3215,6 +3170,7 @@ def forward( # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_70 = paddle._C_ops.reshape(add_158, full_int_array_1) + del add_158 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_70 = paddle._C_ops.transpose(reshape_70, [0, 2, 1, 3]) @@ -3226,9 +3182,11 @@ def forward( # pd_op.matmul: (1x16x21x21xf32) <- (1x16x21x64xf32, 1x16x21x64xf32) matmul_139 = paddle._C_ops.matmul(scale_19, transpose_69, False, True) + del scale_19, transpose_69 # pd_op.add: (1x16x21x21xf32) <- (1x16x21x21xf32, 1x1x1x21xf32) add_159 = paddle._C_ops.add(matmul_139, unsqueeze_0) + del matmul_139 # pd_op.softmax: (1x16x21x21xf32) <- (1x16x21x21xf32) softmax_17 = paddle._C_ops.softmax(add_159, -1) @@ -3237,13 +3195,15 @@ def forward( # pd_op.dropout: (1x16x21x21xf32, 1x16x21x21xui8) <- (1x16x21x21xf32, None, 1xf32) dropout_104, dropout_105 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_17, None, full_5, False, "upscale_in_train", 0, False + softmax_17, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_17 # pd_op.matmul: (1x16x21x64xf32) <- (1x16x21x21xf32, 1x16x21x64xf32) matmul_140 = paddle._C_ops.matmul(dropout_104, transpose_70, False, False) + del dropout_104, transpose_70 # pd_op.transpose: (1x21x16x64xf32) <- (1x16x21x64xf32) transpose_71 = paddle._C_ops.transpose(matmul_140, [0, 2, 1, 3]) @@ -3251,19 +3211,20 @@ def forward( # pd_op.reshape: (1x21x1024xf32) <- (1x21x16x64xf32, 3xi64) reshape_71 = paddle._C_ops.reshape(transpose_71, full_int_array_2) + del transpose_71 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_141 = paddle._C_ops.matmul(reshape_71, parameter_43, False, False) - del parameter_43 + del parameter_43, reshape_71 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_160 = paddle._C_ops.add(matmul_141, parameter_42) - del parameter_42 + del matmul_141, parameter_42 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_106, dropout_107 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_160, None, full_5, False, "upscale_in_train", 0, False + add_160, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -3271,6 +3232,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_161 = paddle._C_ops.add(layer_norm_102, dropout_106) + del dropout_106, layer_norm_102 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_105, layer_norm_106, layer_norm_107 = (lambda x, f: f(x))( @@ -3279,7 +3241,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_36, parameter_37 + del add_161, parameter_36, parameter_37 # pd_op.matmul: (1x21x4096xf32) <- (1x21x1024xf32, 1024x4096xf32) matmul_142 = paddle._C_ops.matmul(layer_norm_105, parameter_41, False, False) @@ -3287,23 +3249,24 @@ def forward( # pd_op.add: (1x21x4096xf32) <- (1x21x4096xf32, 4096xf32) add_162 = paddle._C_ops.add(matmul_142, parameter_40) - del parameter_40 + del matmul_142, parameter_40 # pd_op.gelu: (1x21x4096xf32) <- (1x21x4096xf32) gelu_17 = paddle._C_ops.gelu(add_162, False) + del add_162 # pd_op.matmul: (1x21x1024xf32) <- (1x21x4096xf32, 4096x1024xf32) matmul_143 = paddle._C_ops.matmul(gelu_17, parameter_39, False, False) - del parameter_39 + del gelu_17, parameter_39 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_163 = paddle._C_ops.add(matmul_143, parameter_38) - del parameter_38 + del matmul_143, parameter_38 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_108, dropout_109 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_163, None, full_5, False, "upscale_in_train", 0, False + add_163, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -3311,6 +3274,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_164 = paddle._C_ops.add(layer_norm_105, dropout_108) + del dropout_108, layer_norm_105 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_108, layer_norm_109, layer_norm_110 = (lambda x, f: f(x))( @@ -3319,7 +3283,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_34, parameter_35 + del add_164, parameter_34, parameter_35 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_144 = paddle._C_ops.matmul(layer_norm_108, parameter_33, False, False) @@ -3327,10 +3291,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_165 = paddle._C_ops.add(matmul_144, parameter_32) - del parameter_32 + del matmul_144, parameter_32 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_72 = paddle._C_ops.reshape(add_165, full_int_array_1) + del add_165 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_72 = paddle._C_ops.transpose(reshape_72, [0, 2, 1, 3]) @@ -3342,7 +3307,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_166 = paddle._C_ops.add(matmul_145, parameter_30) - del parameter_30 + del matmul_145, parameter_30 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_146 = paddle._C_ops.matmul(layer_norm_108, parameter_29, False, False) @@ -3350,10 +3315,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_167 = paddle._C_ops.add(matmul_146, parameter_28) - del parameter_28 + del matmul_146, parameter_28 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_73 = paddle._C_ops.reshape(add_166, full_int_array_1) + del add_166 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_73 = paddle._C_ops.transpose(reshape_73, [0, 2, 1, 3]) @@ -3361,6 +3327,7 @@ def forward( # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_74 = paddle._C_ops.reshape(add_167, full_int_array_1) + del add_167 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_74 = paddle._C_ops.transpose(reshape_74, [0, 2, 1, 3]) @@ -3372,9 +3339,11 @@ def forward( # pd_op.matmul: (1x16x21x21xf32) <- (1x16x21x64xf32, 1x16x21x64xf32) matmul_147 = paddle._C_ops.matmul(scale_20, transpose_73, False, True) + del scale_20, transpose_73 # pd_op.add: (1x16x21x21xf32) <- (1x16x21x21xf32, 1x1x1x21xf32) add_168 = paddle._C_ops.add(matmul_147, unsqueeze_0) + del matmul_147 # pd_op.softmax: (1x16x21x21xf32) <- (1x16x21x21xf32) softmax_18 = paddle._C_ops.softmax(add_168, -1) @@ -3383,13 +3352,15 @@ def forward( # pd_op.dropout: (1x16x21x21xf32, 1x16x21x21xui8) <- (1x16x21x21xf32, None, 1xf32) dropout_110, dropout_111 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_18, None, full_5, False, "upscale_in_train", 0, False + softmax_18, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_18 # pd_op.matmul: (1x16x21x64xf32) <- (1x16x21x21xf32, 1x16x21x64xf32) matmul_148 = paddle._C_ops.matmul(dropout_110, transpose_74, False, False) + del dropout_110, transpose_74 # pd_op.transpose: (1x21x16x64xf32) <- (1x16x21x64xf32) transpose_75 = paddle._C_ops.transpose(matmul_148, [0, 2, 1, 3]) @@ -3397,19 +3368,20 @@ def forward( # pd_op.reshape: (1x21x1024xf32) <- (1x21x16x64xf32, 3xi64) reshape_75 = paddle._C_ops.reshape(transpose_75, full_int_array_2) + del transpose_75 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_149 = paddle._C_ops.matmul(reshape_75, parameter_27, False, False) - del parameter_27 + del parameter_27, reshape_75 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_169 = paddle._C_ops.add(matmul_149, parameter_26) - del parameter_26 + del matmul_149, parameter_26 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_112, dropout_113 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_169, None, full_5, False, "upscale_in_train", 0, False + add_169, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -3417,6 +3389,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_170 = paddle._C_ops.add(layer_norm_108, dropout_112) + del dropout_112, layer_norm_108 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_111, layer_norm_112, layer_norm_113 = (lambda x, f: f(x))( @@ -3425,7 +3398,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_20, parameter_21 + del add_170, parameter_20, parameter_21 # pd_op.matmul: (1x21x4096xf32) <- (1x21x1024xf32, 1024x4096xf32) matmul_150 = paddle._C_ops.matmul(layer_norm_111, parameter_25, False, False) @@ -3433,23 +3406,24 @@ def forward( # pd_op.add: (1x21x4096xf32) <- (1x21x4096xf32, 4096xf32) add_171 = paddle._C_ops.add(matmul_150, parameter_24) - del parameter_24 + del matmul_150, parameter_24 # pd_op.gelu: (1x21x4096xf32) <- (1x21x4096xf32) gelu_18 = paddle._C_ops.gelu(add_171, False) + del add_171 # pd_op.matmul: (1x21x1024xf32) <- (1x21x4096xf32, 4096x1024xf32) matmul_151 = paddle._C_ops.matmul(gelu_18, parameter_23, False, False) - del parameter_23 + del gelu_18, parameter_23 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_172 = paddle._C_ops.add(matmul_151, parameter_22) - del parameter_22 + del matmul_151, parameter_22 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_114, dropout_115 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_172, None, full_5, False, "upscale_in_train", 0, False + add_172, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -3457,6 +3431,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_173 = paddle._C_ops.add(layer_norm_111, dropout_114) + del dropout_114, layer_norm_111 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_114, layer_norm_115, layer_norm_116 = (lambda x, f: f(x))( @@ -3465,7 +3440,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_18, parameter_19 + del add_173, parameter_18, parameter_19 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_152 = paddle._C_ops.matmul(layer_norm_114, parameter_17, False, False) @@ -3473,10 +3448,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_174 = paddle._C_ops.add(matmul_152, parameter_16) - del parameter_16 + del matmul_152, parameter_16 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_76 = paddle._C_ops.reshape(add_174, full_int_array_1) + del add_174 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_76 = paddle._C_ops.transpose(reshape_76, [0, 2, 1, 3]) @@ -3488,7 +3464,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_175 = paddle._C_ops.add(matmul_153, parameter_14) - del parameter_14 + del matmul_153, parameter_14 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_154 = paddle._C_ops.matmul(layer_norm_114, parameter_13, False, False) @@ -3496,10 +3472,11 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_176 = paddle._C_ops.add(matmul_154, parameter_12) - del parameter_12 + del matmul_154, parameter_12 # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_77 = paddle._C_ops.reshape(add_175, full_int_array_1) + del add_175 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_77 = paddle._C_ops.transpose(reshape_77, [0, 2, 1, 3]) @@ -3507,7 +3484,7 @@ def forward( # pd_op.reshape: (1x21x16x64xf32) <- (1x21x1024xf32, 4xi64) reshape_78 = paddle._C_ops.reshape(add_176, full_int_array_1) - del full_int_array_1 + del add_176, full_int_array_1 # pd_op.transpose: (1x16x21x64xf32) <- (1x21x16x64xf32) transpose_78 = paddle._C_ops.transpose(reshape_78, [0, 2, 1, 3]) @@ -3515,13 +3492,15 @@ def forward( # pd_op.scale: (1x16x21x64xf32) <- (1x16x21x64xf32, 1xf32) scale_21 = paddle._C_ops.scale(transpose_76, full_6, float("0"), True) - del transpose_76 + del full_6, transpose_76 # pd_op.matmul: (1x16x21x21xf32) <- (1x16x21x64xf32, 1x16x21x64xf32) matmul_155 = paddle._C_ops.matmul(scale_21, transpose_77, False, True) + del scale_21, transpose_77 # pd_op.add: (1x16x21x21xf32) <- (1x16x21x21xf32, 1x1x1x21xf32) add_177 = paddle._C_ops.add(matmul_155, unsqueeze_0) + del matmul_155, unsqueeze_0 # pd_op.softmax: (1x16x21x21xf32) <- (1x16x21x21xf32) softmax_19 = paddle._C_ops.softmax(add_177, -1) @@ -3530,13 +3509,15 @@ def forward( # pd_op.dropout: (1x16x21x21xf32, 1x16x21x21xui8) <- (1x16x21x21xf32, None, 1xf32) dropout_116, dropout_117 = (lambda x, f: f(x))( paddle._C_ops.dropout( - softmax_19, None, full_5, False, "upscale_in_train", 0, False + softmax_19, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) + del softmax_19 # pd_op.matmul: (1x16x21x64xf32) <- (1x16x21x21xf32, 1x16x21x64xf32) matmul_156 = paddle._C_ops.matmul(dropout_116, transpose_78, False, False) + del dropout_116, transpose_78 # pd_op.transpose: (1x21x16x64xf32) <- (1x16x21x64xf32) transpose_79 = paddle._C_ops.transpose(matmul_156, [0, 2, 1, 3]) @@ -3544,20 +3525,20 @@ def forward( # pd_op.reshape: (1x21x1024xf32) <- (1x21x16x64xf32, 3xi64) reshape_79 = paddle._C_ops.reshape(transpose_79, full_int_array_2) - del full_int_array_2 + del full_int_array_2, transpose_79 # pd_op.matmul: (1x21x1024xf32) <- (1x21x1024xf32, 1024x1024xf32) matmul_157 = paddle._C_ops.matmul(reshape_79, parameter_11, False, False) - del parameter_11 + del parameter_11, reshape_79 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_178 = paddle._C_ops.add(matmul_157, parameter_10) - del parameter_10 + del matmul_157, parameter_10 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_118, dropout_119 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_178, None, full_5, False, "upscale_in_train", 0, False + add_178, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) @@ -3565,6 +3546,7 @@ def forward( # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_179 = paddle._C_ops.add(layer_norm_114, dropout_118) + del dropout_118, layer_norm_114 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_117, layer_norm_118, layer_norm_119 = (lambda x, f: f(x))( @@ -3573,7 +3555,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_4, parameter_5 + del add_179, parameter_4, parameter_5 # pd_op.matmul: (1x21x4096xf32) <- (1x21x1024xf32, 1024x4096xf32) matmul_158 = paddle._C_ops.matmul(layer_norm_117, parameter_9, False, False) @@ -3581,30 +3563,32 @@ def forward( # pd_op.add: (1x21x4096xf32) <- (1x21x4096xf32, 4096xf32) add_180 = paddle._C_ops.add(matmul_158, parameter_8) - del parameter_8 + del matmul_158, parameter_8 # pd_op.gelu: (1x21x4096xf32) <- (1x21x4096xf32) gelu_19 = paddle._C_ops.gelu(add_180, False) + del add_180 # pd_op.matmul: (1x21x1024xf32) <- (1x21x4096xf32, 4096x1024xf32) matmul_159 = paddle._C_ops.matmul(gelu_19, parameter_7, False, False) - del parameter_7 + del gelu_19, parameter_7 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1024xf32) add_181 = paddle._C_ops.add(matmul_159, parameter_6) - del parameter_6 + del matmul_159, parameter_6 # pd_op.dropout: (1x21x1024xf32, 1x21x1024xui8) <- (1x21x1024xf32, None, 1xf32) dropout_120, dropout_121 = (lambda x, f: f(x))( paddle._C_ops.dropout( - add_181, None, full_5, False, "upscale_in_train", 0, False + add_181, None, full_5, True, "upscale_in_train", 0, False ), lambda out: out if isinstance(out, (list, tuple)) else (out, None), ) - del add_181 + del add_181, full_5 # pd_op.add: (1x21x1024xf32) <- (1x21x1024xf32, 1x21x1024xf32) add_182 = paddle._C_ops.add(layer_norm_117, dropout_120) + del dropout_120, layer_norm_117 # pd_op.layer_norm: (1x21x1024xf32, 1x21xf32, 1x21xf32) <- (1x21x1024xf32, 1024xf32, 1024xf32) layer_norm_120, layer_norm_121, layer_norm_122 = (lambda x, f: f(x))( @@ -3613,7 +3597,7 @@ def forward( ), lambda out: out if isinstance(out, (list, tuple)) else (out, None, None), ) - del parameter_2, parameter_3 + del add_182, parameter_2, parameter_3 # pd_op.full_int_array: (1xi64) <- () full_int_array_3 = [0] @@ -3625,758 +3609,18 @@ def forward( slice_0 = paddle._C_ops.slice( layer_norm_120, [1], full_int_array_3, full_int_array_4, [1], [1] ) + del full_int_array_3, full_int_array_4, layer_norm_120 # pd_op.matmul: (1x1024xf32) <- (1x1024xf32, 1024x1024xf32) matmul_160 = paddle._C_ops.matmul(slice_0, parameter_1, False, False) - del parameter_1 + del parameter_1, slice_0 # pd_op.add: (1x1024xf32) <- (1x1024xf32, 1024xf32) add_183 = paddle._C_ops.add(matmul_160, parameter_0) - del parameter_0 + del matmul_160, parameter_0 # pd_op.tanh: (1x1024xf32) <- (1x1024xf32) tanh_0 = paddle._C_ops.tanh(add_183) - del ( - add_0, - add_1, - add_101, - add_102, - add_103, - add_104, - add_107, - add_108, - add_11, - add_110, - add_111, - add_112, - add_113, - add_116, - add_117, - add_119, - add_12, - add_120, - add_121, - add_122, - add_125, - add_126, - add_128, - add_129, - add_13, - add_130, - add_131, - add_134, - add_135, - add_137, - add_138, - add_139, - add_14, - add_140, - add_143, - add_144, - add_146, - add_147, - add_148, - add_149, - add_152, - add_153, - add_155, - add_156, - add_157, - add_158, - add_161, - add_162, - add_164, - add_165, - add_166, - add_167, - add_17, - add_170, - add_171, - add_173, - add_174, - add_175, - add_176, - add_179, - add_18, - add_180, - add_182, - add_183, - add_2, - add_20, - add_21, - add_22, - add_23, - add_26, - add_27, - add_29, - add_3, - add_30, - add_31, - add_32, - add_35, - add_36, - add_38, - add_39, - add_4, - add_40, - add_41, - add_44, - add_45, - add_47, - add_48, - add_49, - add_5, - add_50, - add_53, - add_54, - add_56, - add_57, - add_58, - add_59, - add_62, - add_63, - add_65, - add_66, - add_67, - add_68, - add_71, - add_72, - add_74, - add_75, - add_76, - add_77, - add_8, - add_80, - add_81, - add_83, - add_84, - add_85, - add_86, - add_89, - add_9, - add_90, - add_92, - add_93, - add_94, - add_95, - add_98, - add_99, - assign_0, - assign_1, - assign_10, - assign_11, - assign_12, - assign_13, - assign_14, - assign_15, - assign_16, - assign_17, - assign_18, - assign_19, - assign_2, - assign_20, - assign_21, - assign_22, - assign_23, - assign_24, - assign_25, - assign_26, - assign_27, - assign_28, - assign_29, - assign_3, - assign_30, - assign_31, - assign_32, - assign_33, - assign_34, - assign_35, - assign_36, - assign_37, - assign_38, - assign_39, - assign_4, - assign_40, - assign_41, - assign_42, - assign_43, - assign_44, - assign_45, - assign_46, - assign_47, - assign_48, - assign_49, - assign_5, - assign_50, - assign_51, - assign_52, - assign_53, - assign_54, - assign_55, - assign_56, - assign_57, - assign_58, - assign_59, - assign_6, - assign_60, - assign_61, - assign_62, - assign_63, - assign_64, - assign_65, - assign_66, - assign_67, - assign_68, - assign_69, - assign_7, - assign_70, - assign_71, - assign_72, - assign_73, - assign_74, - assign_75, - assign_76, - assign_77, - assign_78, - assign_8, - assign_9, - dropout_0, - dropout_1, - dropout_10, - dropout_100, - dropout_101, - dropout_102, - dropout_103, - dropout_104, - dropout_105, - dropout_106, - dropout_107, - dropout_108, - dropout_109, - dropout_11, - dropout_110, - dropout_111, - dropout_112, - dropout_113, - dropout_114, - dropout_115, - dropout_116, - dropout_117, - dropout_118, - dropout_119, - dropout_12, - dropout_120, - dropout_121, - dropout_13, - dropout_14, - dropout_15, - dropout_16, - dropout_17, - dropout_18, - dropout_19, - dropout_2, - dropout_20, - dropout_21, - dropout_22, - dropout_23, - dropout_24, - dropout_25, - dropout_26, - dropout_27, - dropout_28, - dropout_29, - dropout_3, - dropout_30, - dropout_31, - dropout_32, - dropout_33, - dropout_34, - dropout_35, - dropout_36, - dropout_37, - dropout_38, - dropout_39, - dropout_4, - dropout_40, - dropout_41, - dropout_42, - dropout_43, - dropout_44, - dropout_45, - dropout_46, - dropout_47, - dropout_48, - dropout_49, - dropout_5, - dropout_50, - dropout_51, - dropout_52, - dropout_53, - dropout_54, - dropout_55, - dropout_56, - dropout_57, - dropout_58, - dropout_59, - dropout_6, - dropout_60, - dropout_61, - dropout_62, - dropout_63, - dropout_64, - dropout_65, - dropout_66, - dropout_67, - dropout_68, - dropout_69, - dropout_7, - dropout_70, - dropout_71, - dropout_72, - dropout_73, - dropout_74, - dropout_75, - dropout_76, - dropout_77, - dropout_78, - dropout_79, - dropout_8, - dropout_80, - dropout_81, - dropout_82, - dropout_83, - dropout_84, - dropout_85, - dropout_86, - dropout_87, - dropout_88, - dropout_89, - dropout_9, - dropout_90, - dropout_91, - dropout_92, - dropout_93, - dropout_94, - dropout_95, - dropout_96, - dropout_97, - dropout_98, - dropout_99, - embedding_0, - embedding_1, - embedding_2, - embedding_3, - full_5, - full_6, - full_int_array_3, - full_int_array_4, - gelu_0, - gelu_1, - gelu_10, - gelu_11, - gelu_12, - gelu_13, - gelu_14, - gelu_15, - gelu_16, - gelu_17, - gelu_18, - gelu_19, - gelu_2, - gelu_3, - gelu_4, - gelu_5, - gelu_6, - gelu_7, - gelu_8, - gelu_9, - layer_norm_1, - layer_norm_10, - layer_norm_100, - layer_norm_101, - layer_norm_102, - layer_norm_103, - layer_norm_104, - layer_norm_105, - layer_norm_106, - layer_norm_107, - layer_norm_108, - layer_norm_109, - layer_norm_11, - layer_norm_110, - layer_norm_111, - layer_norm_112, - layer_norm_113, - layer_norm_114, - layer_norm_115, - layer_norm_116, - layer_norm_117, - layer_norm_118, - layer_norm_119, - layer_norm_12, - layer_norm_120, - layer_norm_121, - layer_norm_122, - layer_norm_13, - layer_norm_14, - layer_norm_15, - layer_norm_16, - layer_norm_17, - layer_norm_18, - layer_norm_19, - layer_norm_2, - layer_norm_20, - layer_norm_21, - layer_norm_22, - layer_norm_23, - layer_norm_24, - layer_norm_25, - layer_norm_26, - layer_norm_27, - layer_norm_28, - layer_norm_29, - layer_norm_3, - layer_norm_30, - layer_norm_31, - layer_norm_32, - layer_norm_33, - layer_norm_34, - layer_norm_35, - layer_norm_36, - layer_norm_37, - layer_norm_38, - layer_norm_39, - layer_norm_4, - layer_norm_40, - layer_norm_41, - layer_norm_42, - layer_norm_43, - layer_norm_44, - layer_norm_45, - layer_norm_46, - layer_norm_47, - layer_norm_48, - layer_norm_49, - layer_norm_5, - layer_norm_50, - layer_norm_51, - layer_norm_52, - layer_norm_53, - layer_norm_54, - layer_norm_55, - layer_norm_56, - layer_norm_57, - layer_norm_58, - layer_norm_59, - layer_norm_6, - layer_norm_60, - layer_norm_61, - layer_norm_62, - layer_norm_63, - layer_norm_64, - layer_norm_65, - layer_norm_66, - layer_norm_67, - layer_norm_68, - layer_norm_69, - layer_norm_7, - layer_norm_70, - layer_norm_71, - layer_norm_72, - layer_norm_73, - layer_norm_74, - layer_norm_75, - layer_norm_76, - layer_norm_77, - layer_norm_78, - layer_norm_79, - layer_norm_8, - layer_norm_80, - layer_norm_81, - layer_norm_82, - layer_norm_83, - layer_norm_84, - layer_norm_85, - layer_norm_86, - layer_norm_87, - layer_norm_88, - layer_norm_89, - layer_norm_9, - layer_norm_90, - layer_norm_91, - layer_norm_92, - layer_norm_93, - layer_norm_94, - layer_norm_95, - layer_norm_96, - layer_norm_97, - layer_norm_98, - layer_norm_99, - matmul_0, - matmul_1, - matmul_10, - matmul_101, - matmul_102, - matmul_103, - matmul_104, - matmul_105, - matmul_106, - matmul_107, - matmul_109, - matmul_11, - matmul_110, - matmul_111, - matmul_112, - matmul_113, - matmul_114, - matmul_115, - matmul_117, - matmul_118, - matmul_119, - matmul_120, - matmul_121, - matmul_122, - matmul_123, - matmul_125, - matmul_126, - matmul_127, - matmul_128, - matmul_129, - matmul_13, - matmul_130, - matmul_131, - matmul_133, - matmul_134, - matmul_135, - matmul_136, - matmul_137, - matmul_138, - matmul_139, - matmul_14, - matmul_141, - matmul_142, - matmul_143, - matmul_144, - matmul_145, - matmul_146, - matmul_147, - matmul_149, - matmul_15, - matmul_150, - matmul_151, - matmul_152, - matmul_153, - matmul_154, - matmul_155, - matmul_157, - matmul_158, - matmul_159, - matmul_16, - matmul_160, - matmul_17, - matmul_18, - matmul_19, - matmul_2, - matmul_21, - matmul_22, - matmul_23, - matmul_24, - matmul_25, - matmul_26, - matmul_27, - matmul_29, - matmul_3, - matmul_30, - matmul_31, - matmul_32, - matmul_33, - matmul_34, - matmul_35, - matmul_37, - matmul_38, - matmul_39, - matmul_40, - matmul_41, - matmul_42, - matmul_43, - matmul_45, - matmul_46, - matmul_47, - matmul_48, - matmul_49, - matmul_5, - matmul_50, - matmul_51, - matmul_53, - matmul_54, - matmul_55, - matmul_56, - matmul_57, - matmul_58, - matmul_59, - matmul_6, - matmul_61, - matmul_62, - matmul_63, - matmul_64, - matmul_65, - matmul_66, - matmul_67, - matmul_69, - matmul_7, - matmul_70, - matmul_71, - matmul_72, - matmul_73, - matmul_74, - matmul_75, - matmul_77, - matmul_78, - matmul_79, - matmul_8, - matmul_80, - matmul_81, - matmul_82, - matmul_83, - matmul_85, - matmul_86, - matmul_87, - matmul_88, - matmul_89, - matmul_9, - matmul_90, - matmul_91, - matmul_93, - matmul_94, - matmul_95, - matmul_96, - matmul_97, - matmul_98, - matmul_99, - reshape_11, - reshape_15, - reshape_19, - reshape_23, - reshape_27, - reshape_3, - reshape_31, - reshape_35, - reshape_39, - reshape_43, - reshape_47, - reshape_51, - reshape_55, - reshape_59, - reshape_63, - reshape_67, - reshape_7, - reshape_71, - reshape_75, - reshape_79, - scale_1, - scale_10, - scale_11, - scale_12, - scale_13, - scale_14, - scale_15, - scale_16, - scale_17, - scale_18, - scale_19, - scale_2, - scale_20, - scale_21, - scale_3, - scale_4, - scale_5, - scale_6, - scale_7, - scale_8, - scale_9, - slice_0, - softmax_0, - softmax_1, - softmax_10, - softmax_11, - softmax_12, - softmax_13, - softmax_14, - softmax_15, - softmax_16, - softmax_17, - softmax_18, - softmax_19, - softmax_2, - softmax_3, - softmax_4, - softmax_5, - softmax_6, - softmax_7, - softmax_8, - softmax_9, - subtract_0, - transpose_1, - transpose_10, - transpose_11, - transpose_13, - transpose_14, - transpose_15, - transpose_17, - transpose_18, - transpose_19, - transpose_2, - transpose_21, - transpose_22, - transpose_23, - transpose_25, - transpose_26, - transpose_27, - transpose_29, - transpose_3, - transpose_30, - transpose_31, - transpose_33, - transpose_34, - transpose_35, - transpose_37, - transpose_38, - transpose_39, - transpose_41, - transpose_42, - transpose_43, - transpose_45, - transpose_46, - transpose_47, - transpose_49, - transpose_5, - transpose_50, - transpose_51, - transpose_53, - transpose_54, - transpose_55, - transpose_57, - transpose_58, - transpose_59, - transpose_6, - transpose_61, - transpose_62, - transpose_63, - transpose_65, - transpose_66, - transpose_67, - transpose_69, - transpose_7, - transpose_70, - transpose_71, - transpose_73, - transpose_74, - transpose_75, - transpose_77, - transpose_78, - transpose_79, - transpose_9, - unsqueeze_0, - ) + del add_183 return tanh_0 diff --git a/paddle_samples/PaddleNLP/utc-xbase/weight_meta.py b/paddle_samples/PaddleNLP/utc-xbase/weight_meta.py index bdfb4d269..87e23de07 100644 --- a/paddle_samples/PaddleNLP/utc-xbase/weight_meta.py +++ b/paddle_samples/PaddleNLP/utc-xbase/weight_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_parameter_0: name = "parameter_0" + original_name = "linear_3673.b_0" shape = [1024] dtype = "float32" min_val = float("-0.109667") @@ -11,6 +12,7 @@ class Program_weight_tensor_parameter_0: class Program_weight_tensor_parameter_1: name = "parameter_1" + original_name = "linear_3673.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.228068") @@ -22,6 +24,7 @@ class Program_weight_tensor_parameter_1: class Program_weight_tensor_parameter_2: name = "parameter_2" + original_name = "layer_norm_1269.b_0" shape = [1024] dtype = "float32" min_val = float("-0.828803") @@ -33,6 +36,7 @@ class Program_weight_tensor_parameter_2: class Program_weight_tensor_parameter_3: name = "parameter_3" + original_name = "layer_norm_1269.w_0" shape = [1024] dtype = "float32" min_val = float("0.133532") @@ -44,6 +48,7 @@ class Program_weight_tensor_parameter_3: class Program_weight_tensor_parameter_4: name = "parameter_4" + original_name = "layer_norm_1268.b_0" shape = [1024] dtype = "float32" min_val = float("-1.44178") @@ -55,6 +60,7 @@ class Program_weight_tensor_parameter_4: class Program_weight_tensor_parameter_5: name = "parameter_5" + original_name = "layer_norm_1268.w_0" shape = [1024] dtype = "float32" min_val = float("0.756563") @@ -66,6 +72,7 @@ class Program_weight_tensor_parameter_5: class Program_weight_tensor_parameter_6: name = "parameter_6" + original_name = "linear_3672.b_0" shape = [1024] dtype = "float32" min_val = float("-0.224701") @@ -77,6 +84,7 @@ class Program_weight_tensor_parameter_6: class Program_weight_tensor_parameter_7: name = "parameter_7" + original_name = "linear_3672.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-0.960493") @@ -88,6 +96,7 @@ class Program_weight_tensor_parameter_7: class Program_weight_tensor_parameter_8: name = "parameter_8" + original_name = "linear_3671.b_0" shape = [4096] dtype = "float32" min_val = float("-0.287533") @@ -99,6 +108,7 @@ class Program_weight_tensor_parameter_8: class Program_weight_tensor_parameter_9: name = "parameter_9" + original_name = "linear_3671.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.505827") @@ -110,6 +120,7 @@ class Program_weight_tensor_parameter_9: class Program_weight_tensor_parameter_10: name = "parameter_10" + original_name = "linear_3670.b_0" shape = [1024] dtype = "float32" min_val = float("-0.671723") @@ -121,6 +132,7 @@ class Program_weight_tensor_parameter_10: class Program_weight_tensor_parameter_11: name = "parameter_11" + original_name = "linear_3670.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-1.02") @@ -132,6 +144,7 @@ class Program_weight_tensor_parameter_11: class Program_weight_tensor_parameter_12: name = "parameter_12" + original_name = "linear_3669.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0925895") @@ -143,6 +156,7 @@ class Program_weight_tensor_parameter_12: class Program_weight_tensor_parameter_13: name = "parameter_13" + original_name = "linear_3669.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.291031") @@ -154,6 +168,7 @@ class Program_weight_tensor_parameter_13: class Program_weight_tensor_parameter_14: name = "parameter_14" + original_name = "linear_3668.b_0" shape = [1024] dtype = "float32" min_val = float("-20.2271") @@ -165,6 +180,7 @@ class Program_weight_tensor_parameter_14: class Program_weight_tensor_parameter_15: name = "parameter_15" + original_name = "linear_3668.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.569125") @@ -176,6 +192,7 @@ class Program_weight_tensor_parameter_15: class Program_weight_tensor_parameter_16: name = "parameter_16" + original_name = "linear_3667.b_0" shape = [1024] dtype = "float32" min_val = float("-0.566006") @@ -187,6 +204,7 @@ class Program_weight_tensor_parameter_16: class Program_weight_tensor_parameter_17: name = "parameter_17" + original_name = "linear_3667.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.279315") @@ -198,6 +216,7 @@ class Program_weight_tensor_parameter_17: class Program_weight_tensor_parameter_18: name = "parameter_18" + original_name = "layer_norm_1267.b_0" shape = [1024] dtype = "float32" min_val = float("-1.16011") @@ -209,6 +228,7 @@ class Program_weight_tensor_parameter_18: class Program_weight_tensor_parameter_19: name = "parameter_19" + original_name = "layer_norm_1267.w_0" shape = [1024] dtype = "float32" min_val = float("0.303703") @@ -220,6 +240,7 @@ class Program_weight_tensor_parameter_19: class Program_weight_tensor_parameter_20: name = "parameter_20" + original_name = "layer_norm_1266.b_0" shape = [1024] dtype = "float32" min_val = float("-1.91225") @@ -231,6 +252,7 @@ class Program_weight_tensor_parameter_20: class Program_weight_tensor_parameter_21: name = "parameter_21" + original_name = "layer_norm_1266.w_0" shape = [1024] dtype = "float32" min_val = float("0.733577") @@ -242,6 +264,7 @@ class Program_weight_tensor_parameter_21: class Program_weight_tensor_parameter_22: name = "parameter_22" + original_name = "linear_3666.b_0" shape = [1024] dtype = "float32" min_val = float("-0.32819") @@ -253,6 +276,7 @@ class Program_weight_tensor_parameter_22: class Program_weight_tensor_parameter_23: name = "parameter_23" + original_name = "linear_3666.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-0.758702") @@ -264,6 +288,7 @@ class Program_weight_tensor_parameter_23: class Program_weight_tensor_parameter_24: name = "parameter_24" + original_name = "linear_3665.b_0" shape = [4096] dtype = "float32" min_val = float("-0.265841") @@ -275,6 +300,7 @@ class Program_weight_tensor_parameter_24: class Program_weight_tensor_parameter_25: name = "parameter_25" + original_name = "linear_3665.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.568146") @@ -286,6 +312,7 @@ class Program_weight_tensor_parameter_25: class Program_weight_tensor_parameter_26: name = "parameter_26" + original_name = "linear_3664.b_0" shape = [1024] dtype = "float32" min_val = float("-0.797868") @@ -297,6 +324,7 @@ class Program_weight_tensor_parameter_26: class Program_weight_tensor_parameter_27: name = "parameter_27" + original_name = "linear_3664.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.52824") @@ -308,6 +336,7 @@ class Program_weight_tensor_parameter_27: class Program_weight_tensor_parameter_28: name = "parameter_28" + original_name = "linear_3663.b_0" shape = [1024] dtype = "float32" min_val = float("-0.138759") @@ -319,6 +348,7 @@ class Program_weight_tensor_parameter_28: class Program_weight_tensor_parameter_29: name = "parameter_29" + original_name = "linear_3663.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.200107") @@ -330,6 +360,7 @@ class Program_weight_tensor_parameter_29: class Program_weight_tensor_parameter_30: name = "parameter_30" + original_name = "linear_3662.b_0" shape = [1024] dtype = "float32" min_val = float("-13.1723") @@ -341,6 +372,7 @@ class Program_weight_tensor_parameter_30: class Program_weight_tensor_parameter_31: name = "parameter_31" + original_name = "linear_3662.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.474989") @@ -352,6 +384,7 @@ class Program_weight_tensor_parameter_31: class Program_weight_tensor_parameter_32: name = "parameter_32" + original_name = "linear_3661.b_0" shape = [1024] dtype = "float32" min_val = float("-0.506805") @@ -363,6 +396,7 @@ class Program_weight_tensor_parameter_32: class Program_weight_tensor_parameter_33: name = "parameter_33" + original_name = "linear_3661.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.268329") @@ -374,6 +408,7 @@ class Program_weight_tensor_parameter_33: class Program_weight_tensor_parameter_34: name = "parameter_34" + original_name = "layer_norm_1265.b_0" shape = [1024] dtype = "float32" min_val = float("-1.09385") @@ -385,6 +420,7 @@ class Program_weight_tensor_parameter_34: class Program_weight_tensor_parameter_35: name = "parameter_35" + original_name = "layer_norm_1265.w_0" shape = [1024] dtype = "float32" min_val = float("0.406052") @@ -396,6 +432,7 @@ class Program_weight_tensor_parameter_35: class Program_weight_tensor_parameter_36: name = "parameter_36" + original_name = "layer_norm_1264.b_0" shape = [1024] dtype = "float32" min_val = float("-2.05657") @@ -407,6 +444,7 @@ class Program_weight_tensor_parameter_36: class Program_weight_tensor_parameter_37: name = "parameter_37" + original_name = "layer_norm_1264.w_0" shape = [1024] dtype = "float32" min_val = float("0.762874") @@ -418,6 +456,7 @@ class Program_weight_tensor_parameter_37: class Program_weight_tensor_parameter_38: name = "parameter_38" + original_name = "linear_3660.b_0" shape = [1024] dtype = "float32" min_val = float("-0.449178") @@ -429,6 +468,7 @@ class Program_weight_tensor_parameter_38: class Program_weight_tensor_parameter_39: name = "parameter_39" + original_name = "linear_3660.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-0.684172") @@ -440,6 +480,7 @@ class Program_weight_tensor_parameter_39: class Program_weight_tensor_parameter_40: name = "parameter_40" + original_name = "linear_3659.b_0" shape = [4096] dtype = "float32" min_val = float("-0.252285") @@ -451,6 +492,7 @@ class Program_weight_tensor_parameter_40: class Program_weight_tensor_parameter_41: name = "parameter_41" + original_name = "linear_3659.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.517035") @@ -462,6 +504,7 @@ class Program_weight_tensor_parameter_41: class Program_weight_tensor_parameter_42: name = "parameter_42" + original_name = "linear_3658.b_0" shape = [1024] dtype = "float32" min_val = float("-0.119642") @@ -473,6 +516,7 @@ class Program_weight_tensor_parameter_42: class Program_weight_tensor_parameter_43: name = "parameter_43" + original_name = "linear_3658.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.680798") @@ -484,6 +528,7 @@ class Program_weight_tensor_parameter_43: class Program_weight_tensor_parameter_44: name = "parameter_44" + original_name = "linear_3657.b_0" shape = [1024] dtype = "float32" min_val = float("-0.116182") @@ -495,6 +540,7 @@ class Program_weight_tensor_parameter_44: class Program_weight_tensor_parameter_45: name = "parameter_45" + original_name = "linear_3657.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.214823") @@ -506,6 +552,7 @@ class Program_weight_tensor_parameter_45: class Program_weight_tensor_parameter_46: name = "parameter_46" + original_name = "linear_3656.b_0" shape = [1024] dtype = "float32" min_val = float("-10.8041") @@ -517,6 +564,7 @@ class Program_weight_tensor_parameter_46: class Program_weight_tensor_parameter_47: name = "parameter_47" + original_name = "linear_3656.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.469957") @@ -528,6 +576,7 @@ class Program_weight_tensor_parameter_47: class Program_weight_tensor_parameter_48: name = "parameter_48" + original_name = "linear_3655.b_0" shape = [1024] dtype = "float32" min_val = float("-0.559159") @@ -539,6 +588,7 @@ class Program_weight_tensor_parameter_48: class Program_weight_tensor_parameter_49: name = "parameter_49" + original_name = "linear_3655.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.251893") @@ -550,6 +600,7 @@ class Program_weight_tensor_parameter_49: class Program_weight_tensor_parameter_50: name = "parameter_50" + original_name = "layer_norm_1263.b_0" shape = [1024] dtype = "float32" min_val = float("-0.732306") @@ -561,6 +612,7 @@ class Program_weight_tensor_parameter_50: class Program_weight_tensor_parameter_51: name = "parameter_51" + original_name = "layer_norm_1263.w_0" shape = [1024] dtype = "float32" min_val = float("0.406368") @@ -572,6 +624,7 @@ class Program_weight_tensor_parameter_51: class Program_weight_tensor_parameter_52: name = "parameter_52" + original_name = "layer_norm_1262.b_0" shape = [1024] dtype = "float32" min_val = float("-1.95255") @@ -583,6 +636,7 @@ class Program_weight_tensor_parameter_52: class Program_weight_tensor_parameter_53: name = "parameter_53" + original_name = "layer_norm_1262.w_0" shape = [1024] dtype = "float32" min_val = float("0.729948") @@ -594,6 +648,7 @@ class Program_weight_tensor_parameter_53: class Program_weight_tensor_parameter_54: name = "parameter_54" + original_name = "linear_3654.b_0" shape = [1024] dtype = "float32" min_val = float("-0.398544") @@ -605,6 +660,7 @@ class Program_weight_tensor_parameter_54: class Program_weight_tensor_parameter_55: name = "parameter_55" + original_name = "linear_3654.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-1.44077") @@ -616,6 +672,7 @@ class Program_weight_tensor_parameter_55: class Program_weight_tensor_parameter_56: name = "parameter_56" + original_name = "linear_3653.b_0" shape = [4096] dtype = "float32" min_val = float("-0.209874") @@ -627,6 +684,7 @@ class Program_weight_tensor_parameter_56: class Program_weight_tensor_parameter_57: name = "parameter_57" + original_name = "linear_3653.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.698412") @@ -638,6 +696,7 @@ class Program_weight_tensor_parameter_57: class Program_weight_tensor_parameter_58: name = "parameter_58" + original_name = "linear_3652.b_0" shape = [1024] dtype = "float32" min_val = float("-0.143755") @@ -649,6 +708,7 @@ class Program_weight_tensor_parameter_58: class Program_weight_tensor_parameter_59: name = "parameter_59" + original_name = "linear_3652.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.460259") @@ -660,6 +720,7 @@ class Program_weight_tensor_parameter_59: class Program_weight_tensor_parameter_60: name = "parameter_60" + original_name = "linear_3651.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0919018") @@ -671,6 +732,7 @@ class Program_weight_tensor_parameter_60: class Program_weight_tensor_parameter_61: name = "parameter_61" + original_name = "linear_3651.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.22218") @@ -682,6 +744,7 @@ class Program_weight_tensor_parameter_61: class Program_weight_tensor_parameter_62: name = "parameter_62" + original_name = "linear_3650.b_0" shape = [1024] dtype = "float32" min_val = float("-8.23645") @@ -693,6 +756,7 @@ class Program_weight_tensor_parameter_62: class Program_weight_tensor_parameter_63: name = "parameter_63" + original_name = "linear_3650.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.46862") @@ -704,6 +768,7 @@ class Program_weight_tensor_parameter_63: class Program_weight_tensor_parameter_64: name = "parameter_64" + original_name = "linear_3649.b_0" shape = [1024] dtype = "float32" min_val = float("-0.407792") @@ -715,6 +780,7 @@ class Program_weight_tensor_parameter_64: class Program_weight_tensor_parameter_65: name = "parameter_65" + original_name = "linear_3649.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.342689") @@ -726,6 +792,7 @@ class Program_weight_tensor_parameter_65: class Program_weight_tensor_parameter_66: name = "parameter_66" + original_name = "layer_norm_1261.b_0" shape = [1024] dtype = "float32" min_val = float("-0.651139") @@ -737,6 +804,7 @@ class Program_weight_tensor_parameter_66: class Program_weight_tensor_parameter_67: name = "parameter_67" + original_name = "layer_norm_1261.w_0" shape = [1024] dtype = "float32" min_val = float("0.4407") @@ -748,6 +816,7 @@ class Program_weight_tensor_parameter_67: class Program_weight_tensor_parameter_68: name = "parameter_68" + original_name = "layer_norm_1260.b_0" shape = [1024] dtype = "float32" min_val = float("-1.27358") @@ -759,6 +828,7 @@ class Program_weight_tensor_parameter_68: class Program_weight_tensor_parameter_69: name = "parameter_69" + original_name = "layer_norm_1260.w_0" shape = [1024] dtype = "float32" min_val = float("0.779986") @@ -770,6 +840,7 @@ class Program_weight_tensor_parameter_69: class Program_weight_tensor_parameter_70: name = "parameter_70" + original_name = "linear_3648.b_0" shape = [1024] dtype = "float32" min_val = float("-0.504222") @@ -781,6 +852,7 @@ class Program_weight_tensor_parameter_70: class Program_weight_tensor_parameter_71: name = "parameter_71" + original_name = "linear_3648.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-1.09472") @@ -792,6 +864,7 @@ class Program_weight_tensor_parameter_71: class Program_weight_tensor_parameter_72: name = "parameter_72" + original_name = "linear_3647.b_0" shape = [4096] dtype = "float32" min_val = float("-0.212922") @@ -803,6 +876,7 @@ class Program_weight_tensor_parameter_72: class Program_weight_tensor_parameter_73: name = "parameter_73" + original_name = "linear_3647.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.583319") @@ -814,6 +888,7 @@ class Program_weight_tensor_parameter_73: class Program_weight_tensor_parameter_74: name = "parameter_74" + original_name = "linear_3646.b_0" shape = [1024] dtype = "float32" min_val = float("-0.194831") @@ -825,6 +900,7 @@ class Program_weight_tensor_parameter_74: class Program_weight_tensor_parameter_75: name = "parameter_75" + original_name = "linear_3646.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.787399") @@ -836,6 +912,7 @@ class Program_weight_tensor_parameter_75: class Program_weight_tensor_parameter_76: name = "parameter_76" + original_name = "linear_3645.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0767059") @@ -847,6 +924,7 @@ class Program_weight_tensor_parameter_76: class Program_weight_tensor_parameter_77: name = "parameter_77" + original_name = "linear_3645.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.208358") @@ -858,6 +936,7 @@ class Program_weight_tensor_parameter_77: class Program_weight_tensor_parameter_78: name = "parameter_78" + original_name = "linear_3644.b_0" shape = [1024] dtype = "float32" min_val = float("-15.9508") @@ -869,6 +948,7 @@ class Program_weight_tensor_parameter_78: class Program_weight_tensor_parameter_79: name = "parameter_79" + original_name = "linear_3644.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.543048") @@ -880,6 +960,7 @@ class Program_weight_tensor_parameter_79: class Program_weight_tensor_parameter_80: name = "parameter_80" + original_name = "linear_3643.b_0" shape = [1024] dtype = "float32" min_val = float("-0.327272") @@ -891,6 +972,7 @@ class Program_weight_tensor_parameter_80: class Program_weight_tensor_parameter_81: name = "parameter_81" + original_name = "linear_3643.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.341118") @@ -902,6 +984,7 @@ class Program_weight_tensor_parameter_81: class Program_weight_tensor_parameter_82: name = "parameter_82" + original_name = "layer_norm_1259.b_0" shape = [1024] dtype = "float32" min_val = float("-0.349151") @@ -913,6 +996,7 @@ class Program_weight_tensor_parameter_82: class Program_weight_tensor_parameter_83: name = "parameter_83" + original_name = "layer_norm_1259.w_0" shape = [1024] dtype = "float32" min_val = float("0.467617") @@ -924,6 +1008,7 @@ class Program_weight_tensor_parameter_83: class Program_weight_tensor_parameter_84: name = "parameter_84" + original_name = "layer_norm_1258.b_0" shape = [1024] dtype = "float32" min_val = float("-0.877464") @@ -935,6 +1020,7 @@ class Program_weight_tensor_parameter_84: class Program_weight_tensor_parameter_85: name = "parameter_85" + original_name = "layer_norm_1258.w_0" shape = [1024] dtype = "float32" min_val = float("0.74926") @@ -946,6 +1032,7 @@ class Program_weight_tensor_parameter_85: class Program_weight_tensor_parameter_86: name = "parameter_86" + original_name = "linear_3642.b_0" shape = [1024] dtype = "float32" min_val = float("-0.318582") @@ -957,6 +1044,7 @@ class Program_weight_tensor_parameter_86: class Program_weight_tensor_parameter_87: name = "parameter_87" + original_name = "linear_3642.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-0.978036") @@ -968,6 +1056,7 @@ class Program_weight_tensor_parameter_87: class Program_weight_tensor_parameter_88: name = "parameter_88" + original_name = "linear_3641.b_0" shape = [4096] dtype = "float32" min_val = float("-0.246747") @@ -979,6 +1068,7 @@ class Program_weight_tensor_parameter_88: class Program_weight_tensor_parameter_89: name = "parameter_89" + original_name = "linear_3641.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.458404") @@ -990,6 +1080,7 @@ class Program_weight_tensor_parameter_89: class Program_weight_tensor_parameter_90: name = "parameter_90" + original_name = "linear_3640.b_0" shape = [1024] dtype = "float32" min_val = float("-0.225243") @@ -1001,6 +1092,7 @@ class Program_weight_tensor_parameter_90: class Program_weight_tensor_parameter_91: name = "parameter_91" + original_name = "linear_3640.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.595827") @@ -1012,6 +1104,7 @@ class Program_weight_tensor_parameter_91: class Program_weight_tensor_parameter_92: name = "parameter_92" + original_name = "linear_3639.b_0" shape = [1024] dtype = "float32" min_val = float("-0.10175") @@ -1023,6 +1116,7 @@ class Program_weight_tensor_parameter_92: class Program_weight_tensor_parameter_93: name = "parameter_93" + original_name = "linear_3639.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.203108") @@ -1034,6 +1128,7 @@ class Program_weight_tensor_parameter_93: class Program_weight_tensor_parameter_94: name = "parameter_94" + original_name = "linear_3638.b_0" shape = [1024] dtype = "float32" min_val = float("-6.20397") @@ -1045,6 +1140,7 @@ class Program_weight_tensor_parameter_94: class Program_weight_tensor_parameter_95: name = "parameter_95" + original_name = "linear_3638.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.390288") @@ -1056,6 +1152,7 @@ class Program_weight_tensor_parameter_95: class Program_weight_tensor_parameter_96: name = "parameter_96" + original_name = "linear_3637.b_0" shape = [1024] dtype = "float32" min_val = float("-0.425979") @@ -1067,6 +1164,7 @@ class Program_weight_tensor_parameter_96: class Program_weight_tensor_parameter_97: name = "parameter_97" + original_name = "linear_3637.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.348547") @@ -1078,6 +1176,7 @@ class Program_weight_tensor_parameter_97: class Program_weight_tensor_parameter_98: name = "parameter_98" + original_name = "layer_norm_1257.b_0" shape = [1024] dtype = "float32" min_val = float("-0.125942") @@ -1089,6 +1188,7 @@ class Program_weight_tensor_parameter_98: class Program_weight_tensor_parameter_99: name = "parameter_99" + original_name = "layer_norm_1257.w_0" shape = [1024] dtype = "float32" min_val = float("0.526406") @@ -1100,6 +1200,7 @@ class Program_weight_tensor_parameter_99: class Program_weight_tensor_parameter_100: name = "parameter_100" + original_name = "layer_norm_1256.b_0" shape = [1024] dtype = "float32" min_val = float("-0.641643") @@ -1111,6 +1212,7 @@ class Program_weight_tensor_parameter_100: class Program_weight_tensor_parameter_101: name = "parameter_101" + original_name = "layer_norm_1256.w_0" shape = [1024] dtype = "float32" min_val = float("0.76041") @@ -1122,6 +1224,7 @@ class Program_weight_tensor_parameter_101: class Program_weight_tensor_parameter_102: name = "parameter_102" + original_name = "linear_3636.b_0" shape = [1024] dtype = "float32" min_val = float("-0.398017") @@ -1133,6 +1236,7 @@ class Program_weight_tensor_parameter_102: class Program_weight_tensor_parameter_103: name = "parameter_103" + original_name = "linear_3636.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-1.02287") @@ -1144,6 +1248,7 @@ class Program_weight_tensor_parameter_103: class Program_weight_tensor_parameter_104: name = "parameter_104" + original_name = "linear_3635.b_0" shape = [4096] dtype = "float32" min_val = float("-0.25304") @@ -1155,6 +1260,7 @@ class Program_weight_tensor_parameter_104: class Program_weight_tensor_parameter_105: name = "parameter_105" + original_name = "linear_3635.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.540307") @@ -1166,6 +1272,7 @@ class Program_weight_tensor_parameter_105: class Program_weight_tensor_parameter_106: name = "parameter_106" + original_name = "linear_3634.b_0" shape = [1024] dtype = "float32" min_val = float("-0.143753") @@ -1177,6 +1284,7 @@ class Program_weight_tensor_parameter_106: class Program_weight_tensor_parameter_107: name = "parameter_107" + original_name = "linear_3634.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.425938") @@ -1188,6 +1296,7 @@ class Program_weight_tensor_parameter_107: class Program_weight_tensor_parameter_108: name = "parameter_108" + original_name = "linear_3633.b_0" shape = [1024] dtype = "float32" min_val = float("-0.184381") @@ -1199,6 +1308,7 @@ class Program_weight_tensor_parameter_108: class Program_weight_tensor_parameter_109: name = "parameter_109" + original_name = "linear_3633.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.218216") @@ -1210,6 +1320,7 @@ class Program_weight_tensor_parameter_109: class Program_weight_tensor_parameter_110: name = "parameter_110" + original_name = "linear_3632.b_0" shape = [1024] dtype = "float32" min_val = float("-6.96266") @@ -1221,6 +1332,7 @@ class Program_weight_tensor_parameter_110: class Program_weight_tensor_parameter_111: name = "parameter_111" + original_name = "linear_3632.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.34703") @@ -1232,6 +1344,7 @@ class Program_weight_tensor_parameter_111: class Program_weight_tensor_parameter_112: name = "parameter_112" + original_name = "linear_3631.b_0" shape = [1024] dtype = "float32" min_val = float("-0.405107") @@ -1243,6 +1356,7 @@ class Program_weight_tensor_parameter_112: class Program_weight_tensor_parameter_113: name = "parameter_113" + original_name = "linear_3631.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.292183") @@ -1254,6 +1368,7 @@ class Program_weight_tensor_parameter_113: class Program_weight_tensor_parameter_114: name = "parameter_114" + original_name = "layer_norm_1255.b_0" shape = [1024] dtype = "float32" min_val = float("-0.213844") @@ -1265,6 +1380,7 @@ class Program_weight_tensor_parameter_114: class Program_weight_tensor_parameter_115: name = "parameter_115" + original_name = "layer_norm_1255.w_0" shape = [1024] dtype = "float32" min_val = float("0.346154") @@ -1276,6 +1392,7 @@ class Program_weight_tensor_parameter_115: class Program_weight_tensor_parameter_116: name = "parameter_116" + original_name = "layer_norm_1254.b_0" shape = [1024] dtype = "float32" min_val = float("-0.991806") @@ -1287,6 +1404,7 @@ class Program_weight_tensor_parameter_116: class Program_weight_tensor_parameter_117: name = "parameter_117" + original_name = "layer_norm_1254.w_0" shape = [1024] dtype = "float32" min_val = float("0.769764") @@ -1298,6 +1416,7 @@ class Program_weight_tensor_parameter_117: class Program_weight_tensor_parameter_118: name = "parameter_118" + original_name = "linear_3630.b_0" shape = [1024] dtype = "float32" min_val = float("-0.449651") @@ -1309,6 +1428,7 @@ class Program_weight_tensor_parameter_118: class Program_weight_tensor_parameter_119: name = "parameter_119" + original_name = "linear_3630.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-0.601161") @@ -1320,6 +1440,7 @@ class Program_weight_tensor_parameter_119: class Program_weight_tensor_parameter_120: name = "parameter_120" + original_name = "linear_3629.b_0" shape = [4096] dtype = "float32" min_val = float("-0.318705") @@ -1331,6 +1452,7 @@ class Program_weight_tensor_parameter_120: class Program_weight_tensor_parameter_121: name = "parameter_121" + original_name = "linear_3629.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.361919") @@ -1342,6 +1464,7 @@ class Program_weight_tensor_parameter_121: class Program_weight_tensor_parameter_122: name = "parameter_122" + original_name = "linear_3628.b_0" shape = [1024] dtype = "float32" min_val = float("-0.368877") @@ -1353,6 +1476,7 @@ class Program_weight_tensor_parameter_122: class Program_weight_tensor_parameter_123: name = "parameter_123" + original_name = "linear_3628.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.290397") @@ -1364,6 +1488,7 @@ class Program_weight_tensor_parameter_123: class Program_weight_tensor_parameter_124: name = "parameter_124" + original_name = "linear_3627.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0771337") @@ -1375,6 +1500,7 @@ class Program_weight_tensor_parameter_124: class Program_weight_tensor_parameter_125: name = "parameter_125" + original_name = "linear_3627.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.21345") @@ -1386,6 +1512,7 @@ class Program_weight_tensor_parameter_125: class Program_weight_tensor_parameter_126: name = "parameter_126" + original_name = "linear_3626.b_0" shape = [1024] dtype = "float32" min_val = float("-4.83315") @@ -1397,6 +1524,7 @@ class Program_weight_tensor_parameter_126: class Program_weight_tensor_parameter_127: name = "parameter_127" + original_name = "linear_3626.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.316024") @@ -1408,6 +1536,7 @@ class Program_weight_tensor_parameter_127: class Program_weight_tensor_parameter_128: name = "parameter_128" + original_name = "linear_3625.b_0" shape = [1024] dtype = "float32" min_val = float("-0.490598") @@ -1419,6 +1548,7 @@ class Program_weight_tensor_parameter_128: class Program_weight_tensor_parameter_129: name = "parameter_129" + original_name = "linear_3625.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.304782") @@ -1430,6 +1560,7 @@ class Program_weight_tensor_parameter_129: class Program_weight_tensor_parameter_130: name = "parameter_130" + original_name = "layer_norm_1253.b_0" shape = [1024] dtype = "float32" min_val = float("-0.1898") @@ -1441,6 +1572,7 @@ class Program_weight_tensor_parameter_130: class Program_weight_tensor_parameter_131: name = "parameter_131" + original_name = "layer_norm_1253.w_0" shape = [1024] dtype = "float32" min_val = float("0.184788") @@ -1452,6 +1584,7 @@ class Program_weight_tensor_parameter_131: class Program_weight_tensor_parameter_132: name = "parameter_132" + original_name = "layer_norm_1252.b_0" shape = [1024] dtype = "float32" min_val = float("-1.62751") @@ -1463,6 +1596,7 @@ class Program_weight_tensor_parameter_132: class Program_weight_tensor_parameter_133: name = "parameter_133" + original_name = "layer_norm_1252.w_0" shape = [1024] dtype = "float32" min_val = float("0.711052") @@ -1474,6 +1608,7 @@ class Program_weight_tensor_parameter_133: class Program_weight_tensor_parameter_134: name = "parameter_134" + original_name = "linear_3624.b_0" shape = [1024] dtype = "float32" min_val = float("-0.239985") @@ -1485,6 +1620,7 @@ class Program_weight_tensor_parameter_134: class Program_weight_tensor_parameter_135: name = "parameter_135" + original_name = "linear_3624.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-0.770121") @@ -1496,6 +1632,7 @@ class Program_weight_tensor_parameter_135: class Program_weight_tensor_parameter_136: name = "parameter_136" + original_name = "linear_3623.b_0" shape = [4096] dtype = "float32" min_val = float("-0.234855") @@ -1507,6 +1644,7 @@ class Program_weight_tensor_parameter_136: class Program_weight_tensor_parameter_137: name = "parameter_137" + original_name = "linear_3623.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.408789") @@ -1518,6 +1656,7 @@ class Program_weight_tensor_parameter_137: class Program_weight_tensor_parameter_138: name = "parameter_138" + original_name = "linear_3622.b_0" shape = [1024] dtype = "float32" min_val = float("-0.71853") @@ -1529,6 +1668,7 @@ class Program_weight_tensor_parameter_138: class Program_weight_tensor_parameter_139: name = "parameter_139" + original_name = "linear_3622.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.382412") @@ -1540,6 +1680,7 @@ class Program_weight_tensor_parameter_139: class Program_weight_tensor_parameter_140: name = "parameter_140" + original_name = "linear_3621.b_0" shape = [1024] dtype = "float32" min_val = float("-0.110511") @@ -1551,6 +1692,7 @@ class Program_weight_tensor_parameter_140: class Program_weight_tensor_parameter_141: name = "parameter_141" + original_name = "linear_3621.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.314881") @@ -1562,6 +1704,7 @@ class Program_weight_tensor_parameter_141: class Program_weight_tensor_parameter_142: name = "parameter_142" + original_name = "linear_3620.b_0" shape = [1024] dtype = "float32" min_val = float("-6.00239") @@ -1573,6 +1716,7 @@ class Program_weight_tensor_parameter_142: class Program_weight_tensor_parameter_143: name = "parameter_143" + original_name = "linear_3620.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.308886") @@ -1584,6 +1728,7 @@ class Program_weight_tensor_parameter_143: class Program_weight_tensor_parameter_144: name = "parameter_144" + original_name = "linear_3619.b_0" shape = [1024] dtype = "float32" min_val = float("-0.518165") @@ -1595,6 +1740,7 @@ class Program_weight_tensor_parameter_144: class Program_weight_tensor_parameter_145: name = "parameter_145" + original_name = "linear_3619.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.320046") @@ -1606,6 +1752,7 @@ class Program_weight_tensor_parameter_145: class Program_weight_tensor_parameter_146: name = "parameter_146" + original_name = "layer_norm_1251.b_0" shape = [1024] dtype = "float32" min_val = float("-0.235557") @@ -1617,6 +1764,7 @@ class Program_weight_tensor_parameter_146: class Program_weight_tensor_parameter_147: name = "parameter_147" + original_name = "layer_norm_1251.w_0" shape = [1024] dtype = "float32" min_val = float("0.203569") @@ -1628,6 +1776,7 @@ class Program_weight_tensor_parameter_147: class Program_weight_tensor_parameter_148: name = "parameter_148" + original_name = "layer_norm_1250.b_0" shape = [1024] dtype = "float32" min_val = float("-2.07874") @@ -1639,6 +1788,7 @@ class Program_weight_tensor_parameter_148: class Program_weight_tensor_parameter_149: name = "parameter_149" + original_name = "layer_norm_1250.w_0" shape = [1024] dtype = "float32" min_val = float("0.768518") @@ -1650,6 +1800,7 @@ class Program_weight_tensor_parameter_149: class Program_weight_tensor_parameter_150: name = "parameter_150" + original_name = "linear_3618.b_0" shape = [1024] dtype = "float32" min_val = float("-0.302542") @@ -1661,6 +1812,7 @@ class Program_weight_tensor_parameter_150: class Program_weight_tensor_parameter_151: name = "parameter_151" + original_name = "linear_3618.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-1.33504") @@ -1672,6 +1824,7 @@ class Program_weight_tensor_parameter_151: class Program_weight_tensor_parameter_152: name = "parameter_152" + original_name = "linear_3617.b_0" shape = [4096] dtype = "float32" min_val = float("-0.193622") @@ -1683,6 +1836,7 @@ class Program_weight_tensor_parameter_152: class Program_weight_tensor_parameter_153: name = "parameter_153" + original_name = "linear_3617.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.522062") @@ -1694,6 +1848,7 @@ class Program_weight_tensor_parameter_153: class Program_weight_tensor_parameter_154: name = "parameter_154" + original_name = "linear_3616.b_0" shape = [1024] dtype = "float32" min_val = float("-0.836213") @@ -1705,6 +1860,7 @@ class Program_weight_tensor_parameter_154: class Program_weight_tensor_parameter_155: name = "parameter_155" + original_name = "linear_3616.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.377717") @@ -1716,6 +1872,7 @@ class Program_weight_tensor_parameter_155: class Program_weight_tensor_parameter_156: name = "parameter_156" + original_name = "linear_3615.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0661734") @@ -1727,6 +1884,7 @@ class Program_weight_tensor_parameter_156: class Program_weight_tensor_parameter_157: name = "parameter_157" + original_name = "linear_3615.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.19631") @@ -1738,6 +1896,7 @@ class Program_weight_tensor_parameter_157: class Program_weight_tensor_parameter_158: name = "parameter_158" + original_name = "linear_3614.b_0" shape = [1024] dtype = "float32" min_val = float("-4.46037") @@ -1749,6 +1908,7 @@ class Program_weight_tensor_parameter_158: class Program_weight_tensor_parameter_159: name = "parameter_159" + original_name = "linear_3614.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.540496") @@ -1760,6 +1920,7 @@ class Program_weight_tensor_parameter_159: class Program_weight_tensor_parameter_160: name = "parameter_160" + original_name = "linear_3613.b_0" shape = [1024] dtype = "float32" min_val = float("-0.432967") @@ -1771,6 +1932,7 @@ class Program_weight_tensor_parameter_160: class Program_weight_tensor_parameter_161: name = "parameter_161" + original_name = "linear_3613.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.30851") @@ -1782,6 +1944,7 @@ class Program_weight_tensor_parameter_161: class Program_weight_tensor_parameter_162: name = "parameter_162" + original_name = "layer_norm_1249.b_0" shape = [1024] dtype = "float32" min_val = float("-0.302107") @@ -1793,6 +1956,7 @@ class Program_weight_tensor_parameter_162: class Program_weight_tensor_parameter_163: name = "parameter_163" + original_name = "layer_norm_1249.w_0" shape = [1024] dtype = "float32" min_val = float("0.37308") @@ -1804,6 +1968,7 @@ class Program_weight_tensor_parameter_163: class Program_weight_tensor_parameter_164: name = "parameter_164" + original_name = "layer_norm_1248.b_0" shape = [1024] dtype = "float32" min_val = float("-1.93698") @@ -1815,6 +1980,7 @@ class Program_weight_tensor_parameter_164: class Program_weight_tensor_parameter_165: name = "parameter_165" + original_name = "layer_norm_1248.w_0" shape = [1024] dtype = "float32" min_val = float("0.819789") @@ -1826,6 +1992,7 @@ class Program_weight_tensor_parameter_165: class Program_weight_tensor_parameter_166: name = "parameter_166" + original_name = "linear_3612.b_0" shape = [1024] dtype = "float32" min_val = float("-0.492453") @@ -1837,6 +2004,7 @@ class Program_weight_tensor_parameter_166: class Program_weight_tensor_parameter_167: name = "parameter_167" + original_name = "linear_3612.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-1.84985") @@ -1848,6 +2016,7 @@ class Program_weight_tensor_parameter_167: class Program_weight_tensor_parameter_168: name = "parameter_168" + original_name = "linear_3611.b_0" shape = [4096] dtype = "float32" min_val = float("-0.167589") @@ -1859,6 +2028,7 @@ class Program_weight_tensor_parameter_168: class Program_weight_tensor_parameter_169: name = "parameter_169" + original_name = "linear_3611.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.459637") @@ -1870,6 +2040,7 @@ class Program_weight_tensor_parameter_169: class Program_weight_tensor_parameter_170: name = "parameter_170" + original_name = "linear_3610.b_0" shape = [1024] dtype = "float32" min_val = float("-0.691571") @@ -1881,6 +2052,7 @@ class Program_weight_tensor_parameter_170: class Program_weight_tensor_parameter_171: name = "parameter_171" + original_name = "linear_3610.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.355774") @@ -1892,6 +2064,7 @@ class Program_weight_tensor_parameter_171: class Program_weight_tensor_parameter_172: name = "parameter_172" + original_name = "linear_3609.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0808865") @@ -1903,6 +2076,7 @@ class Program_weight_tensor_parameter_172: class Program_weight_tensor_parameter_173: name = "parameter_173" + original_name = "linear_3609.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.248606") @@ -1914,6 +2088,7 @@ class Program_weight_tensor_parameter_173: class Program_weight_tensor_parameter_174: name = "parameter_174" + original_name = "linear_3608.b_0" shape = [1024] dtype = "float32" min_val = float("-4.02521") @@ -1925,6 +2100,7 @@ class Program_weight_tensor_parameter_174: class Program_weight_tensor_parameter_175: name = "parameter_175" + original_name = "linear_3608.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.350841") @@ -1936,6 +2112,7 @@ class Program_weight_tensor_parameter_175: class Program_weight_tensor_parameter_176: name = "parameter_176" + original_name = "linear_3607.b_0" shape = [1024] dtype = "float32" min_val = float("-0.552479") @@ -1947,6 +2124,7 @@ class Program_weight_tensor_parameter_176: class Program_weight_tensor_parameter_177: name = "parameter_177" + original_name = "linear_3607.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.292778") @@ -1958,6 +2136,7 @@ class Program_weight_tensor_parameter_177: class Program_weight_tensor_parameter_178: name = "parameter_178" + original_name = "layer_norm_1247.b_0" shape = [1024] dtype = "float32" min_val = float("-0.226098") @@ -1969,6 +2148,7 @@ class Program_weight_tensor_parameter_178: class Program_weight_tensor_parameter_179: name = "parameter_179" + original_name = "layer_norm_1247.w_0" shape = [1024] dtype = "float32" min_val = float("0.36922") @@ -1980,6 +2160,7 @@ class Program_weight_tensor_parameter_179: class Program_weight_tensor_parameter_180: name = "parameter_180" + original_name = "layer_norm_1246.b_0" shape = [1024] dtype = "float32" min_val = float("-1.88391") @@ -1991,6 +2172,7 @@ class Program_weight_tensor_parameter_180: class Program_weight_tensor_parameter_181: name = "parameter_181" + original_name = "layer_norm_1246.w_0" shape = [1024] dtype = "float32" min_val = float("0.841685") @@ -2002,6 +2184,7 @@ class Program_weight_tensor_parameter_181: class Program_weight_tensor_parameter_182: name = "parameter_182" + original_name = "linear_3606.b_0" shape = [1024] dtype = "float32" min_val = float("-0.640868") @@ -2013,6 +2196,7 @@ class Program_weight_tensor_parameter_182: class Program_weight_tensor_parameter_183: name = "parameter_183" + original_name = "linear_3606.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-2.22028") @@ -2024,6 +2208,7 @@ class Program_weight_tensor_parameter_183: class Program_weight_tensor_parameter_184: name = "parameter_184" + original_name = "linear_3605.b_0" shape = [4096] dtype = "float32" min_val = float("-0.207696") @@ -2035,6 +2220,7 @@ class Program_weight_tensor_parameter_184: class Program_weight_tensor_parameter_185: name = "parameter_185" + original_name = "linear_3605.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.40061") @@ -2046,6 +2232,7 @@ class Program_weight_tensor_parameter_185: class Program_weight_tensor_parameter_186: name = "parameter_186" + original_name = "linear_3604.b_0" shape = [1024] dtype = "float32" min_val = float("-0.423887") @@ -2057,6 +2244,7 @@ class Program_weight_tensor_parameter_186: class Program_weight_tensor_parameter_187: name = "parameter_187" + original_name = "linear_3604.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.294172") @@ -2068,6 +2256,7 @@ class Program_weight_tensor_parameter_187: class Program_weight_tensor_parameter_188: name = "parameter_188" + original_name = "linear_3603.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0868673") @@ -2079,6 +2268,7 @@ class Program_weight_tensor_parameter_188: class Program_weight_tensor_parameter_189: name = "parameter_189" + original_name = "linear_3603.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.222209") @@ -2090,6 +2280,7 @@ class Program_weight_tensor_parameter_189: class Program_weight_tensor_parameter_190: name = "parameter_190" + original_name = "linear_3602.b_0" shape = [1024] dtype = "float32" min_val = float("-5.25709") @@ -2101,6 +2292,7 @@ class Program_weight_tensor_parameter_190: class Program_weight_tensor_parameter_191: name = "parameter_191" + original_name = "linear_3602.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.734347") @@ -2112,6 +2304,7 @@ class Program_weight_tensor_parameter_191: class Program_weight_tensor_parameter_192: name = "parameter_192" + original_name = "linear_3601.b_0" shape = [1024] dtype = "float32" min_val = float("-0.39884") @@ -2123,6 +2316,7 @@ class Program_weight_tensor_parameter_192: class Program_weight_tensor_parameter_193: name = "parameter_193" + original_name = "linear_3601.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.39313") @@ -2134,6 +2328,7 @@ class Program_weight_tensor_parameter_193: class Program_weight_tensor_parameter_194: name = "parameter_194" + original_name = "layer_norm_1245.b_0" shape = [1024] dtype = "float32" min_val = float("-0.355357") @@ -2145,6 +2340,7 @@ class Program_weight_tensor_parameter_194: class Program_weight_tensor_parameter_195: name = "parameter_195" + original_name = "layer_norm_1245.w_0" shape = [1024] dtype = "float32" min_val = float("0.390035") @@ -2156,6 +2352,7 @@ class Program_weight_tensor_parameter_195: class Program_weight_tensor_parameter_196: name = "parameter_196" + original_name = "layer_norm_1244.b_0" shape = [1024] dtype = "float32" min_val = float("-2.03733") @@ -2167,6 +2364,7 @@ class Program_weight_tensor_parameter_196: class Program_weight_tensor_parameter_197: name = "parameter_197" + original_name = "layer_norm_1244.w_0" shape = [1024] dtype = "float32" min_val = float("0.817321") @@ -2178,6 +2376,7 @@ class Program_weight_tensor_parameter_197: class Program_weight_tensor_parameter_198: name = "parameter_198" + original_name = "linear_3600.b_0" shape = [1024] dtype = "float32" min_val = float("-0.57632") @@ -2189,6 +2388,7 @@ class Program_weight_tensor_parameter_198: class Program_weight_tensor_parameter_199: name = "parameter_199" + original_name = "linear_3600.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-2.1346") @@ -2200,6 +2400,7 @@ class Program_weight_tensor_parameter_199: class Program_weight_tensor_parameter_200: name = "parameter_200" + original_name = "linear_3599.b_0" shape = [4096] dtype = "float32" min_val = float("-0.152141") @@ -2211,6 +2412,7 @@ class Program_weight_tensor_parameter_200: class Program_weight_tensor_parameter_201: name = "parameter_201" + original_name = "linear_3599.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.341477") @@ -2222,6 +2424,7 @@ class Program_weight_tensor_parameter_201: class Program_weight_tensor_parameter_202: name = "parameter_202" + original_name = "linear_3598.b_0" shape = [1024] dtype = "float32" min_val = float("-0.599971") @@ -2233,6 +2436,7 @@ class Program_weight_tensor_parameter_202: class Program_weight_tensor_parameter_203: name = "parameter_203" + original_name = "linear_3598.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.34667") @@ -2244,6 +2448,7 @@ class Program_weight_tensor_parameter_203: class Program_weight_tensor_parameter_204: name = "parameter_204" + original_name = "linear_3597.b_0" shape = [1024] dtype = "float32" min_val = float("-0.106748") @@ -2255,6 +2460,7 @@ class Program_weight_tensor_parameter_204: class Program_weight_tensor_parameter_205: name = "parameter_205" + original_name = "linear_3597.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.332563") @@ -2266,6 +2472,7 @@ class Program_weight_tensor_parameter_205: class Program_weight_tensor_parameter_206: name = "parameter_206" + original_name = "linear_3596.b_0" shape = [1024] dtype = "float32" min_val = float("-4.9304") @@ -2277,6 +2484,7 @@ class Program_weight_tensor_parameter_206: class Program_weight_tensor_parameter_207: name = "parameter_207" + original_name = "linear_3596.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.46287") @@ -2288,6 +2496,7 @@ class Program_weight_tensor_parameter_207: class Program_weight_tensor_parameter_208: name = "parameter_208" + original_name = "linear_3595.b_0" shape = [1024] dtype = "float32" min_val = float("-0.353761") @@ -2299,6 +2508,7 @@ class Program_weight_tensor_parameter_208: class Program_weight_tensor_parameter_209: name = "parameter_209" + original_name = "linear_3595.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.332026") @@ -2310,6 +2520,7 @@ class Program_weight_tensor_parameter_209: class Program_weight_tensor_parameter_210: name = "parameter_210" + original_name = "layer_norm_1243.b_0" shape = [1024] dtype = "float32" min_val = float("-0.463938") @@ -2321,6 +2532,7 @@ class Program_weight_tensor_parameter_210: class Program_weight_tensor_parameter_211: name = "parameter_211" + original_name = "layer_norm_1243.w_0" shape = [1024] dtype = "float32" min_val = float("0.492569") @@ -2332,6 +2544,7 @@ class Program_weight_tensor_parameter_211: class Program_weight_tensor_parameter_212: name = "parameter_212" + original_name = "layer_norm_1242.b_0" shape = [1024] dtype = "float32" min_val = float("-1.97992") @@ -2343,6 +2556,7 @@ class Program_weight_tensor_parameter_212: class Program_weight_tensor_parameter_213: name = "parameter_213" + original_name = "layer_norm_1242.w_0" shape = [1024] dtype = "float32" min_val = float("0.844239") @@ -2354,6 +2568,7 @@ class Program_weight_tensor_parameter_213: class Program_weight_tensor_parameter_214: name = "parameter_214" + original_name = "linear_3594.b_0" shape = [1024] dtype = "float32" min_val = float("-0.398354") @@ -2365,6 +2580,7 @@ class Program_weight_tensor_parameter_214: class Program_weight_tensor_parameter_215: name = "parameter_215" + original_name = "linear_3594.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-2.4502") @@ -2376,6 +2592,7 @@ class Program_weight_tensor_parameter_215: class Program_weight_tensor_parameter_216: name = "parameter_216" + original_name = "linear_3593.b_0" shape = [4096] dtype = "float32" min_val = float("-0.184713") @@ -2387,6 +2604,7 @@ class Program_weight_tensor_parameter_216: class Program_weight_tensor_parameter_217: name = "parameter_217" + original_name = "linear_3593.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.331602") @@ -2398,6 +2616,7 @@ class Program_weight_tensor_parameter_217: class Program_weight_tensor_parameter_218: name = "parameter_218" + original_name = "linear_3592.b_0" shape = [1024] dtype = "float32" min_val = float("-0.404902") @@ -2409,6 +2628,7 @@ class Program_weight_tensor_parameter_218: class Program_weight_tensor_parameter_219: name = "parameter_219" + original_name = "linear_3592.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.349367") @@ -2420,6 +2640,7 @@ class Program_weight_tensor_parameter_219: class Program_weight_tensor_parameter_220: name = "parameter_220" + original_name = "linear_3591.b_0" shape = [1024] dtype = "float32" min_val = float("-0.10397") @@ -2431,6 +2652,7 @@ class Program_weight_tensor_parameter_220: class Program_weight_tensor_parameter_221: name = "parameter_221" + original_name = "linear_3591.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.207198") @@ -2442,6 +2664,7 @@ class Program_weight_tensor_parameter_221: class Program_weight_tensor_parameter_222: name = "parameter_222" + original_name = "linear_3590.b_0" shape = [1024] dtype = "float32" min_val = float("-5.03153") @@ -2453,6 +2676,7 @@ class Program_weight_tensor_parameter_222: class Program_weight_tensor_parameter_223: name = "parameter_223" + original_name = "linear_3590.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.528866") @@ -2464,6 +2688,7 @@ class Program_weight_tensor_parameter_223: class Program_weight_tensor_parameter_224: name = "parameter_224" + original_name = "linear_3589.b_0" shape = [1024] dtype = "float32" min_val = float("-0.458368") @@ -2475,6 +2700,7 @@ class Program_weight_tensor_parameter_224: class Program_weight_tensor_parameter_225: name = "parameter_225" + original_name = "linear_3589.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.369091") @@ -2486,6 +2712,7 @@ class Program_weight_tensor_parameter_225: class Program_weight_tensor_parameter_226: name = "parameter_226" + original_name = "layer_norm_1241.b_0" shape = [1024] dtype = "float32" min_val = float("-0.796062") @@ -2497,6 +2724,7 @@ class Program_weight_tensor_parameter_226: class Program_weight_tensor_parameter_227: name = "parameter_227" + original_name = "layer_norm_1241.w_0" shape = [1024] dtype = "float32" min_val = float("0.373671") @@ -2508,6 +2736,7 @@ class Program_weight_tensor_parameter_227: class Program_weight_tensor_parameter_228: name = "parameter_228" + original_name = "layer_norm_1240.b_0" shape = [1024] dtype = "float32" min_val = float("-1.9909") @@ -2519,6 +2748,7 @@ class Program_weight_tensor_parameter_228: class Program_weight_tensor_parameter_229: name = "parameter_229" + original_name = "layer_norm_1240.w_0" shape = [1024] dtype = "float32" min_val = float("0.834366") @@ -2530,6 +2760,7 @@ class Program_weight_tensor_parameter_229: class Program_weight_tensor_parameter_230: name = "parameter_230" + original_name = "linear_3588.b_0" shape = [1024] dtype = "float32" min_val = float("-0.416096") @@ -2541,6 +2772,7 @@ class Program_weight_tensor_parameter_230: class Program_weight_tensor_parameter_231: name = "parameter_231" + original_name = "linear_3588.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-3.83076") @@ -2552,6 +2784,7 @@ class Program_weight_tensor_parameter_231: class Program_weight_tensor_parameter_232: name = "parameter_232" + original_name = "linear_3587.b_0" shape = [4096] dtype = "float32" min_val = float("-0.139753") @@ -2563,6 +2796,7 @@ class Program_weight_tensor_parameter_232: class Program_weight_tensor_parameter_233: name = "parameter_233" + original_name = "linear_3587.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.652324") @@ -2574,6 +2808,7 @@ class Program_weight_tensor_parameter_233: class Program_weight_tensor_parameter_234: name = "parameter_234" + original_name = "linear_3586.b_0" shape = [1024] dtype = "float32" min_val = float("-0.688466") @@ -2585,6 +2820,7 @@ class Program_weight_tensor_parameter_234: class Program_weight_tensor_parameter_235: name = "parameter_235" + original_name = "linear_3586.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.474426") @@ -2596,6 +2832,7 @@ class Program_weight_tensor_parameter_235: class Program_weight_tensor_parameter_236: name = "parameter_236" + original_name = "linear_3585.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0657855") @@ -2607,6 +2844,7 @@ class Program_weight_tensor_parameter_236: class Program_weight_tensor_parameter_237: name = "parameter_237" + original_name = "linear_3585.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.23378") @@ -2618,6 +2856,7 @@ class Program_weight_tensor_parameter_237: class Program_weight_tensor_parameter_238: name = "parameter_238" + original_name = "linear_3584.b_0" shape = [1024] dtype = "float32" min_val = float("-5.60581") @@ -2629,6 +2868,7 @@ class Program_weight_tensor_parameter_238: class Program_weight_tensor_parameter_239: name = "parameter_239" + original_name = "linear_3584.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.376193") @@ -2640,6 +2880,7 @@ class Program_weight_tensor_parameter_239: class Program_weight_tensor_parameter_240: name = "parameter_240" + original_name = "linear_3583.b_0" shape = [1024] dtype = "float32" min_val = float("-0.452683") @@ -2651,6 +2892,7 @@ class Program_weight_tensor_parameter_240: class Program_weight_tensor_parameter_241: name = "parameter_241" + original_name = "linear_3583.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.421967") @@ -2662,6 +2904,7 @@ class Program_weight_tensor_parameter_241: class Program_weight_tensor_parameter_242: name = "parameter_242" + original_name = "layer_norm_1239.b_0" shape = [1024] dtype = "float32" min_val = float("-0.936374") @@ -2673,6 +2916,7 @@ class Program_weight_tensor_parameter_242: class Program_weight_tensor_parameter_243: name = "parameter_243" + original_name = "layer_norm_1239.w_0" shape = [1024] dtype = "float32" min_val = float("0.140298") @@ -2684,6 +2928,7 @@ class Program_weight_tensor_parameter_243: class Program_weight_tensor_parameter_244: name = "parameter_244" + original_name = "layer_norm_1238.b_0" shape = [1024] dtype = "float32" min_val = float("-2.0986") @@ -2695,6 +2940,7 @@ class Program_weight_tensor_parameter_244: class Program_weight_tensor_parameter_245: name = "parameter_245" + original_name = "layer_norm_1238.w_0" shape = [1024] dtype = "float32" min_val = float("0.785527") @@ -2706,6 +2952,7 @@ class Program_weight_tensor_parameter_245: class Program_weight_tensor_parameter_246: name = "parameter_246" + original_name = "linear_3582.b_0" shape = [1024] dtype = "float32" min_val = float("-0.410728") @@ -2717,6 +2964,7 @@ class Program_weight_tensor_parameter_246: class Program_weight_tensor_parameter_247: name = "parameter_247" + original_name = "linear_3582.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-4.49786") @@ -2728,6 +2976,7 @@ class Program_weight_tensor_parameter_247: class Program_weight_tensor_parameter_248: name = "parameter_248" + original_name = "linear_3581.b_0" shape = [4096] dtype = "float32" min_val = float("-0.133458") @@ -2739,6 +2988,7 @@ class Program_weight_tensor_parameter_248: class Program_weight_tensor_parameter_249: name = "parameter_249" + original_name = "linear_3581.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.535721") @@ -2750,6 +3000,7 @@ class Program_weight_tensor_parameter_249: class Program_weight_tensor_parameter_250: name = "parameter_250" + original_name = "linear_3580.b_0" shape = [1024] dtype = "float32" min_val = float("-0.454064") @@ -2761,6 +3012,7 @@ class Program_weight_tensor_parameter_250: class Program_weight_tensor_parameter_251: name = "parameter_251" + original_name = "linear_3580.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.528543") @@ -2772,6 +3024,7 @@ class Program_weight_tensor_parameter_251: class Program_weight_tensor_parameter_252: name = "parameter_252" + original_name = "linear_3579.b_0" shape = [1024] dtype = "float32" min_val = float("-0.12405") @@ -2783,6 +3036,7 @@ class Program_weight_tensor_parameter_252: class Program_weight_tensor_parameter_253: name = "parameter_253" + original_name = "linear_3579.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.189238") @@ -2794,6 +3048,7 @@ class Program_weight_tensor_parameter_253: class Program_weight_tensor_parameter_254: name = "parameter_254" + original_name = "linear_3578.b_0" shape = [1024] dtype = "float32" min_val = float("-5.89314") @@ -2805,6 +3060,7 @@ class Program_weight_tensor_parameter_254: class Program_weight_tensor_parameter_255: name = "parameter_255" + original_name = "linear_3578.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.347598") @@ -2816,6 +3072,7 @@ class Program_weight_tensor_parameter_255: class Program_weight_tensor_parameter_256: name = "parameter_256" + original_name = "linear_3577.b_0" shape = [1024] dtype = "float32" min_val = float("-0.456737") @@ -2827,6 +3084,7 @@ class Program_weight_tensor_parameter_256: class Program_weight_tensor_parameter_257: name = "parameter_257" + original_name = "linear_3577.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.310478") @@ -2838,6 +3096,7 @@ class Program_weight_tensor_parameter_257: class Program_weight_tensor_parameter_258: name = "parameter_258" + original_name = "layer_norm_1237.b_0" shape = [1024] dtype = "float32" min_val = float("-0.786046") @@ -2849,6 +3108,7 @@ class Program_weight_tensor_parameter_258: class Program_weight_tensor_parameter_259: name = "parameter_259" + original_name = "layer_norm_1237.w_0" shape = [1024] dtype = "float32" min_val = float("0.259398") @@ -2860,6 +3120,7 @@ class Program_weight_tensor_parameter_259: class Program_weight_tensor_parameter_260: name = "parameter_260" + original_name = "layer_norm_1236.b_0" shape = [1024] dtype = "float32" min_val = float("-1.90102") @@ -2871,6 +3132,7 @@ class Program_weight_tensor_parameter_260: class Program_weight_tensor_parameter_261: name = "parameter_261" + original_name = "layer_norm_1236.w_0" shape = [1024] dtype = "float32" min_val = float("0.76713") @@ -2882,6 +3144,7 @@ class Program_weight_tensor_parameter_261: class Program_weight_tensor_parameter_262: name = "parameter_262" + original_name = "linear_3576.b_0" shape = [1024] dtype = "float32" min_val = float("-0.390896") @@ -2893,6 +3156,7 @@ class Program_weight_tensor_parameter_262: class Program_weight_tensor_parameter_263: name = "parameter_263" + original_name = "linear_3576.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-3.5622") @@ -2904,6 +3168,7 @@ class Program_weight_tensor_parameter_263: class Program_weight_tensor_parameter_264: name = "parameter_264" + original_name = "linear_3575.b_0" shape = [4096] dtype = "float32" min_val = float("-0.150746") @@ -2915,6 +3180,7 @@ class Program_weight_tensor_parameter_264: class Program_weight_tensor_parameter_265: name = "parameter_265" + original_name = "linear_3575.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.620615") @@ -2926,6 +3192,7 @@ class Program_weight_tensor_parameter_265: class Program_weight_tensor_parameter_266: name = "parameter_266" + original_name = "linear_3574.b_0" shape = [1024] dtype = "float32" min_val = float("-0.527421") @@ -2937,6 +3204,7 @@ class Program_weight_tensor_parameter_266: class Program_weight_tensor_parameter_267: name = "parameter_267" + original_name = "linear_3574.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.506842") @@ -2948,6 +3216,7 @@ class Program_weight_tensor_parameter_267: class Program_weight_tensor_parameter_268: name = "parameter_268" + original_name = "linear_3573.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0944259") @@ -2959,6 +3228,7 @@ class Program_weight_tensor_parameter_268: class Program_weight_tensor_parameter_269: name = "parameter_269" + original_name = "linear_3573.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.254153") @@ -2970,6 +3240,7 @@ class Program_weight_tensor_parameter_269: class Program_weight_tensor_parameter_270: name = "parameter_270" + original_name = "linear_3572.b_0" shape = [1024] dtype = "float32" min_val = float("-4.6147") @@ -2981,6 +3252,7 @@ class Program_weight_tensor_parameter_270: class Program_weight_tensor_parameter_271: name = "parameter_271" + original_name = "linear_3572.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.439672") @@ -2992,6 +3264,7 @@ class Program_weight_tensor_parameter_271: class Program_weight_tensor_parameter_272: name = "parameter_272" + original_name = "linear_3571.b_0" shape = [1024] dtype = "float32" min_val = float("-0.354839") @@ -3003,6 +3276,7 @@ class Program_weight_tensor_parameter_272: class Program_weight_tensor_parameter_273: name = "parameter_273" + original_name = "linear_3571.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.296777") @@ -3014,6 +3288,7 @@ class Program_weight_tensor_parameter_273: class Program_weight_tensor_parameter_274: name = "parameter_274" + original_name = "layer_norm_1235.b_0" shape = [1024] dtype = "float32" min_val = float("-0.941908") @@ -3025,6 +3300,7 @@ class Program_weight_tensor_parameter_274: class Program_weight_tensor_parameter_275: name = "parameter_275" + original_name = "layer_norm_1235.w_0" shape = [1024] dtype = "float32" min_val = float("0.393158") @@ -3036,6 +3312,7 @@ class Program_weight_tensor_parameter_275: class Program_weight_tensor_parameter_276: name = "parameter_276" + original_name = "layer_norm_1234.b_0" shape = [1024] dtype = "float32" min_val = float("-1.78475") @@ -3047,6 +3324,7 @@ class Program_weight_tensor_parameter_276: class Program_weight_tensor_parameter_277: name = "parameter_277" + original_name = "layer_norm_1234.w_0" shape = [1024] dtype = "float32" min_val = float("0.815845") @@ -3058,6 +3336,7 @@ class Program_weight_tensor_parameter_277: class Program_weight_tensor_parameter_278: name = "parameter_278" + original_name = "linear_3570.b_0" shape = [1024] dtype = "float32" min_val = float("-0.447988") @@ -3069,6 +3348,7 @@ class Program_weight_tensor_parameter_278: class Program_weight_tensor_parameter_279: name = "parameter_279" + original_name = "linear_3570.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-1.67494") @@ -3080,6 +3360,7 @@ class Program_weight_tensor_parameter_279: class Program_weight_tensor_parameter_280: name = "parameter_280" + original_name = "linear_3569.b_0" shape = [4096] dtype = "float32" min_val = float("-0.130953") @@ -3091,6 +3372,7 @@ class Program_weight_tensor_parameter_280: class Program_weight_tensor_parameter_281: name = "parameter_281" + original_name = "linear_3569.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.337668") @@ -3102,6 +3384,7 @@ class Program_weight_tensor_parameter_281: class Program_weight_tensor_parameter_282: name = "parameter_282" + original_name = "linear_3568.b_0" shape = [1024] dtype = "float32" min_val = float("-0.305927") @@ -3113,6 +3396,7 @@ class Program_weight_tensor_parameter_282: class Program_weight_tensor_parameter_283: name = "parameter_283" + original_name = "linear_3568.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.28257") @@ -3124,6 +3408,7 @@ class Program_weight_tensor_parameter_283: class Program_weight_tensor_parameter_284: name = "parameter_284" + original_name = "linear_3567.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0991313") @@ -3135,6 +3420,7 @@ class Program_weight_tensor_parameter_284: class Program_weight_tensor_parameter_285: name = "parameter_285" + original_name = "linear_3567.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.189261") @@ -3146,6 +3432,7 @@ class Program_weight_tensor_parameter_285: class Program_weight_tensor_parameter_286: name = "parameter_286" + original_name = "linear_3566.b_0" shape = [1024] dtype = "float32" min_val = float("-4.72085") @@ -3157,6 +3444,7 @@ class Program_weight_tensor_parameter_286: class Program_weight_tensor_parameter_287: name = "parameter_287" + original_name = "linear_3566.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.417967") @@ -3168,6 +3456,7 @@ class Program_weight_tensor_parameter_287: class Program_weight_tensor_parameter_288: name = "parameter_288" + original_name = "linear_3565.b_0" shape = [1024] dtype = "float32" min_val = float("-0.440276") @@ -3179,6 +3468,7 @@ class Program_weight_tensor_parameter_288: class Program_weight_tensor_parameter_289: name = "parameter_289" + original_name = "linear_3565.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.311547") @@ -3190,6 +3480,7 @@ class Program_weight_tensor_parameter_289: class Program_weight_tensor_parameter_290: name = "parameter_290" + original_name = "layer_norm_1233.b_0" shape = [1024] dtype = "float32" min_val = float("-0.832812") @@ -3201,6 +3492,7 @@ class Program_weight_tensor_parameter_290: class Program_weight_tensor_parameter_291: name = "parameter_291" + original_name = "layer_norm_1233.w_0" shape = [1024] dtype = "float32" min_val = float("0.441643") @@ -3212,6 +3504,7 @@ class Program_weight_tensor_parameter_291: class Program_weight_tensor_parameter_292: name = "parameter_292" + original_name = "layer_norm_1232.b_0" shape = [1024] dtype = "float32" min_val = float("-1.57479") @@ -3223,6 +3516,7 @@ class Program_weight_tensor_parameter_292: class Program_weight_tensor_parameter_293: name = "parameter_293" + original_name = "layer_norm_1232.w_0" shape = [1024] dtype = "float32" min_val = float("0.821111") @@ -3234,6 +3528,7 @@ class Program_weight_tensor_parameter_293: class Program_weight_tensor_parameter_294: name = "parameter_294" + original_name = "linear_3564.b_0" shape = [1024] dtype = "float32" min_val = float("-0.360933") @@ -3245,6 +3540,7 @@ class Program_weight_tensor_parameter_294: class Program_weight_tensor_parameter_295: name = "parameter_295" + original_name = "linear_3564.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-1.08267") @@ -3256,6 +3552,7 @@ class Program_weight_tensor_parameter_295: class Program_weight_tensor_parameter_296: name = "parameter_296" + original_name = "linear_3563.b_0" shape = [4096] dtype = "float32" min_val = float("-0.125224") @@ -3267,6 +3564,7 @@ class Program_weight_tensor_parameter_296: class Program_weight_tensor_parameter_297: name = "parameter_297" + original_name = "linear_3563.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.320222") @@ -3278,6 +3576,7 @@ class Program_weight_tensor_parameter_297: class Program_weight_tensor_parameter_298: name = "parameter_298" + original_name = "linear_3562.b_0" shape = [1024] dtype = "float32" min_val = float("-0.316534") @@ -3289,6 +3588,7 @@ class Program_weight_tensor_parameter_298: class Program_weight_tensor_parameter_299: name = "parameter_299" + original_name = "linear_3562.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.894505") @@ -3300,6 +3600,7 @@ class Program_weight_tensor_parameter_299: class Program_weight_tensor_parameter_300: name = "parameter_300" + original_name = "linear_3561.b_0" shape = [1024] dtype = "float32" min_val = float("-0.158846") @@ -3311,6 +3612,7 @@ class Program_weight_tensor_parameter_300: class Program_weight_tensor_parameter_301: name = "parameter_301" + original_name = "linear_3561.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.208101") @@ -3322,6 +3624,7 @@ class Program_weight_tensor_parameter_301: class Program_weight_tensor_parameter_302: name = "parameter_302" + original_name = "linear_3560.b_0" shape = [1024] dtype = "float32" min_val = float("-5.15342") @@ -3333,6 +3636,7 @@ class Program_weight_tensor_parameter_302: class Program_weight_tensor_parameter_303: name = "parameter_303" + original_name = "linear_3560.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.394899") @@ -3344,6 +3648,7 @@ class Program_weight_tensor_parameter_303: class Program_weight_tensor_parameter_304: name = "parameter_304" + original_name = "linear_3559.b_0" shape = [1024] dtype = "float32" min_val = float("-0.772058") @@ -3355,6 +3660,7 @@ class Program_weight_tensor_parameter_304: class Program_weight_tensor_parameter_305: name = "parameter_305" + original_name = "linear_3559.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.379615") @@ -3366,6 +3672,7 @@ class Program_weight_tensor_parameter_305: class Program_weight_tensor_parameter_306: name = "parameter_306" + original_name = "layer_norm_1231.b_0" shape = [1024] dtype = "float32" min_val = float("-0.659021") @@ -3377,6 +3684,7 @@ class Program_weight_tensor_parameter_306: class Program_weight_tensor_parameter_307: name = "parameter_307" + original_name = "layer_norm_1231.w_0" shape = [1024] dtype = "float32" min_val = float("0.329969") @@ -3388,6 +3696,7 @@ class Program_weight_tensor_parameter_307: class Program_weight_tensor_parameter_308: name = "parameter_308" + original_name = "layer_norm_1230.b_0" shape = [1024] dtype = "float32" min_val = float("-1.89088") @@ -3399,6 +3708,7 @@ class Program_weight_tensor_parameter_308: class Program_weight_tensor_parameter_309: name = "parameter_309" + original_name = "layer_norm_1230.w_0" shape = [1024] dtype = "float32" min_val = float("0.861549") @@ -3410,6 +3720,7 @@ class Program_weight_tensor_parameter_309: class Program_weight_tensor_parameter_310: name = "parameter_310" + original_name = "linear_3558.b_0" shape = [1024] dtype = "float32" min_val = float("-0.45393") @@ -3421,6 +3732,7 @@ class Program_weight_tensor_parameter_310: class Program_weight_tensor_parameter_311: name = "parameter_311" + original_name = "linear_3558.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-1.01878") @@ -3432,6 +3744,7 @@ class Program_weight_tensor_parameter_311: class Program_weight_tensor_parameter_312: name = "parameter_312" + original_name = "linear_3557.b_0" shape = [4096] dtype = "float32" min_val = float("-0.427268") @@ -3443,6 +3756,7 @@ class Program_weight_tensor_parameter_312: class Program_weight_tensor_parameter_313: name = "parameter_313" + original_name = "linear_3557.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.503751") @@ -3454,6 +3768,7 @@ class Program_weight_tensor_parameter_313: class Program_weight_tensor_parameter_314: name = "parameter_314" + original_name = "linear_3556.b_0" shape = [1024] dtype = "float32" min_val = float("-0.301582") @@ -3465,6 +3780,7 @@ class Program_weight_tensor_parameter_314: class Program_weight_tensor_parameter_315: name = "parameter_315" + original_name = "linear_3556.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.619248") @@ -3476,6 +3792,7 @@ class Program_weight_tensor_parameter_315: class Program_weight_tensor_parameter_316: name = "parameter_316" + original_name = "linear_3555.b_0" shape = [1024] dtype = "float32" min_val = float("-0.113842") @@ -3487,6 +3804,7 @@ class Program_weight_tensor_parameter_316: class Program_weight_tensor_parameter_317: name = "parameter_317" + original_name = "linear_3555.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.180661") @@ -3498,6 +3816,7 @@ class Program_weight_tensor_parameter_317: class Program_weight_tensor_parameter_318: name = "parameter_318" + original_name = "linear_3554.b_0" shape = [1024] dtype = "float32" min_val = float("-2.2891") @@ -3509,6 +3828,7 @@ class Program_weight_tensor_parameter_318: class Program_weight_tensor_parameter_319: name = "parameter_319" + original_name = "linear_3554.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.30943") @@ -3520,6 +3840,7 @@ class Program_weight_tensor_parameter_319: class Program_weight_tensor_parameter_320: name = "parameter_320" + original_name = "linear_3553.b_0" shape = [1024] dtype = "float32" min_val = float("-0.621728") @@ -3531,6 +3852,7 @@ class Program_weight_tensor_parameter_320: class Program_weight_tensor_parameter_321: name = "parameter_321" + original_name = "linear_3553.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-0.245789") @@ -3542,6 +3864,7 @@ class Program_weight_tensor_parameter_321: class Program_weight_tensor_parameter_322: name = "parameter_322" + original_name = "layer_norm_1229.b_0" shape = [1024] dtype = "float32" min_val = float("-1.19149") @@ -3553,6 +3876,7 @@ class Program_weight_tensor_parameter_322: class Program_weight_tensor_parameter_323: name = "parameter_323" + original_name = "layer_norm_1229.w_0" shape = [1024] dtype = "float32" min_val = float("0.274043") @@ -3564,6 +3888,7 @@ class Program_weight_tensor_parameter_323: class Program_weight_tensor_parameter_324: name = "parameter_324" + original_name = "embedding_244.w_0" shape = [16, 1024] dtype = "float32" min_val = float("-0.0799812") @@ -3575,6 +3900,7 @@ class Program_weight_tensor_parameter_324: class Program_weight_tensor_parameter_325: name = "parameter_325" + original_name = "embedding_243.w_0" shape = [4, 1024] dtype = "float32" min_val = float("-0.23129") @@ -3586,6 +3912,7 @@ class Program_weight_tensor_parameter_325: class Program_weight_tensor_parameter_326: name = "parameter_326" + original_name = "embedding_242.w_0" shape = [2048, 1024] dtype = "float32" min_val = float("-0.782574") @@ -3597,6 +3924,7 @@ class Program_weight_tensor_parameter_326: class Program_weight_tensor_parameter_327: name = "parameter_327" + original_name = "embedding_241.w_0" shape = [39981, 1024] dtype = "float32" min_val = float("-1.04291") diff --git a/paddle_samples/PaddleX/Deeplabv3_Plus-R101/subgraph_1/input_meta.py b/paddle_samples/PaddleX/Deeplabv3_Plus-R101/subgraph_1/input_meta.py index 1c41d416b..b64e9a07b 100644 --- a/paddle_samples/PaddleX/Deeplabv3_Plus-R101/subgraph_1/input_meta.py +++ b/paddle_samples/PaddleX/Deeplabv3_Plus-R101/subgraph_1/input_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_data_0: name = "data_0" + original_name = "var_0" shape = [2, 3, 512, 1024] dtype = "float32" min_val = float("-1.0") diff --git a/paddle_samples/PaddleX/Deeplabv3_Plus-R101/subgraph_1/weight_meta.py b/paddle_samples/PaddleX/Deeplabv3_Plus-R101/subgraph_1/weight_meta.py index 025099617..bb794eb7e 100644 --- a/paddle_samples/PaddleX/Deeplabv3_Plus-R101/subgraph_1/weight_meta.py +++ b/paddle_samples/PaddleX/Deeplabv3_Plus-R101/subgraph_1/weight_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_parameter_0: name = "parameter_0" + original_name = "conv2d_120.b_0" shape = [2] dtype = "float32" min_val = float("0") @@ -9,6 +10,7 @@ class Program_weight_tensor_parameter_0: class Program_weight_tensor_parameter_1: name = "parameter_1" + original_name = "conv2d_120.w_0" shape = [2, 256, 1, 1] dtype = "float32" min_val = float("-0.278233") @@ -20,6 +22,7 @@ class Program_weight_tensor_parameter_1: class Program_weight_tensor_parameter_2: name = "parameter_2" + original_name = "batch_norm2d_119.b_0" shape = [256] dtype = "float32" data = None @@ -27,26 +30,31 @@ class Program_weight_tensor_parameter_2: class Program_weight_tensor_parameter_3: name = "parameter_3" + original_name = "batch_norm2d_119.w_0" shape = [256] dtype = "float32" min_val = float("1.0") max_val = float("1.0") mean = float("1.0") + std = float("0.0") data = None class Program_weight_tensor_parameter_4: name = "parameter_4" + original_name = "batch_norm2d_119.w_2" shape = [256] dtype = "float32" min_val = float("1.0") max_val = float("1.0") mean = float("1.0") + std = float("0.0") data = None class Program_weight_tensor_parameter_5: name = "parameter_5" + original_name = "batch_norm2d_119.w_1" shape = [256] dtype = "float32" data = None @@ -54,6 +62,7 @@ class Program_weight_tensor_parameter_5: class Program_weight_tensor_parameter_6: name = "parameter_6" + original_name = "conv2d_119.b_0" shape = [256] dtype = "float32" data = None @@ -61,6 +70,7 @@ class Program_weight_tensor_parameter_6: class Program_weight_tensor_parameter_7: name = "parameter_7" + original_name = "conv2d_119.w_0" shape = [256, 256, 1, 1] dtype = "float32" min_val = float("-0.350909") @@ -72,6 +82,7 @@ class Program_weight_tensor_parameter_7: class Program_weight_tensor_parameter_8: name = "parameter_8" + original_name = "batch_norm2d_118.b_0" shape = [256] dtype = "float32" data = None @@ -79,26 +90,31 @@ class Program_weight_tensor_parameter_8: class Program_weight_tensor_parameter_9: name = "parameter_9" + original_name = "batch_norm2d_118.w_0" shape = [256] dtype = "float32" min_val = float("1.0") max_val = float("1.0") mean = float("1.0") + std = float("0.0") data = None class Program_weight_tensor_parameter_10: name = "parameter_10" + original_name = "batch_norm2d_118.w_2" shape = [256] dtype = "float32" min_val = float("1.0") max_val = float("1.0") mean = float("1.0") + std = float("0.0") data = None class Program_weight_tensor_parameter_11: name = "parameter_11" + original_name = "batch_norm2d_118.w_1" shape = [256] dtype = "float32" data = None @@ -106,6 +122,7 @@ class Program_weight_tensor_parameter_11: class Program_weight_tensor_parameter_12: name = "parameter_12" + original_name = "conv2d_118.b_0" shape = [256] dtype = "float32" data = None @@ -113,6 +130,7 @@ class Program_weight_tensor_parameter_12: class Program_weight_tensor_parameter_13: name = "parameter_13" + original_name = "conv2d_118.w_0" shape = [256, 1, 3, 3] dtype = "float32" min_val = float("-0.115875") @@ -124,6 +142,7 @@ class Program_weight_tensor_parameter_13: class Program_weight_tensor_parameter_14: name = "parameter_14" + original_name = "batch_norm2d_117.b_0" shape = [256] dtype = "float32" data = None @@ -131,26 +150,31 @@ class Program_weight_tensor_parameter_14: class Program_weight_tensor_parameter_15: name = "parameter_15" + original_name = "batch_norm2d_117.w_0" shape = [256] dtype = "float32" min_val = float("1.0") max_val = float("1.0") mean = float("1.0") + std = float("0.0") data = None class Program_weight_tensor_parameter_16: name = "parameter_16" + original_name = "batch_norm2d_117.w_2" shape = [256] dtype = "float32" min_val = float("1.0") max_val = float("1.0") mean = float("1.0") + std = float("0.0") data = None class Program_weight_tensor_parameter_17: name = "parameter_17" + original_name = "batch_norm2d_117.w_1" shape = [256] dtype = "float32" data = None @@ -158,6 +182,7 @@ class Program_weight_tensor_parameter_17: class Program_weight_tensor_parameter_18: name = "parameter_18" + original_name = "conv2d_117.b_0" shape = [256] dtype = "float32" data = None @@ -165,6 +190,7 @@ class Program_weight_tensor_parameter_18: class Program_weight_tensor_parameter_19: name = "parameter_19" + original_name = "conv2d_117.w_0" shape = [256, 304, 1, 1] dtype = "float32" min_val = float("-0.357949") @@ -176,6 +202,7 @@ class Program_weight_tensor_parameter_19: class Program_weight_tensor_parameter_20: name = "parameter_20" + original_name = "batch_norm2d_116.b_0" shape = [304] dtype = "float32" data = None @@ -183,26 +210,31 @@ class Program_weight_tensor_parameter_20: class Program_weight_tensor_parameter_21: name = "parameter_21" + original_name = "batch_norm2d_116.w_0" shape = [304] dtype = "float32" min_val = float("1.0") max_val = float("1.0") mean = float("1.0") + std = float("0.0") data = None class Program_weight_tensor_parameter_22: name = "parameter_22" + original_name = "batch_norm2d_116.w_2" shape = [304] dtype = "float32" min_val = float("1.0") max_val = float("1.0") mean = float("1.0") + std = float("0.0") data = None class Program_weight_tensor_parameter_23: name = "parameter_23" + original_name = "batch_norm2d_116.w_1" shape = [304] dtype = "float32" data = None @@ -210,6 +242,7 @@ class Program_weight_tensor_parameter_23: class Program_weight_tensor_parameter_24: name = "parameter_24" + original_name = "conv2d_116.b_0" shape = [304] dtype = "float32" data = None @@ -217,6 +250,7 @@ class Program_weight_tensor_parameter_24: class Program_weight_tensor_parameter_25: name = "parameter_25" + original_name = "conv2d_116.w_0" shape = [304, 1, 3, 3] dtype = "float32" min_val = float("-0.0861833") @@ -228,6 +262,7 @@ class Program_weight_tensor_parameter_25: class Program_weight_tensor_parameter_26: name = "parameter_26" + original_name = "batch_norm2d_115.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -237,6 +272,7 @@ class Program_weight_tensor_parameter_26: class Program_weight_tensor_parameter_27: name = "parameter_27" + original_name = "batch_norm2d_115.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -246,6 +282,7 @@ class Program_weight_tensor_parameter_27: class Program_weight_tensor_parameter_28: name = "parameter_28" + original_name = "batch_norm2d_115.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -255,6 +292,7 @@ class Program_weight_tensor_parameter_28: class Program_weight_tensor_parameter_29: name = "parameter_29" + original_name = "batch_norm2d_115.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -264,6 +302,7 @@ class Program_weight_tensor_parameter_29: class Program_weight_tensor_parameter_30: name = "parameter_30" + original_name = "conv2d_115.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -273,6 +312,7 @@ class Program_weight_tensor_parameter_30: class Program_weight_tensor_parameter_31: name = "parameter_31" + original_name = "conv2d_115.w_0" shape = [48, 256, 1, 1] dtype = "float32" min_val = float("-0.329901") @@ -284,6 +324,7 @@ class Program_weight_tensor_parameter_31: class Program_weight_tensor_parameter_32: name = "parameter_32" + original_name = "batch_norm2d_114.b_0" shape = [256] dtype = "float32" data = None @@ -291,26 +332,31 @@ class Program_weight_tensor_parameter_32: class Program_weight_tensor_parameter_33: name = "parameter_33" + original_name = "batch_norm2d_114.w_0" shape = [256] dtype = "float32" min_val = float("1.0") max_val = float("1.0") mean = float("1.0") + std = float("0.0") data = None class Program_weight_tensor_parameter_34: name = "parameter_34" + original_name = "batch_norm2d_114.w_2" shape = [256] dtype = "float32" min_val = float("1.0") max_val = float("1.0") mean = float("1.0") + std = float("0.0") data = None class Program_weight_tensor_parameter_35: name = "parameter_35" + original_name = "batch_norm2d_114.w_1" shape = [256] dtype = "float32" data = None @@ -318,6 +364,7 @@ class Program_weight_tensor_parameter_35: class Program_weight_tensor_parameter_36: name = "parameter_36" + original_name = "conv2d_114.b_0" shape = [256] dtype = "float32" data = None @@ -325,6 +372,7 @@ class Program_weight_tensor_parameter_36: class Program_weight_tensor_parameter_37: name = "parameter_37" + original_name = "conv2d_114.w_0" shape = [256, 1280, 1, 1] dtype = "float32" min_val = float("-0.176027") @@ -336,6 +384,7 @@ class Program_weight_tensor_parameter_37: class Program_weight_tensor_parameter_38: name = "parameter_38" + original_name = "batch_norm2d_113.b_0" shape = [256] dtype = "float32" data = None @@ -343,26 +392,31 @@ class Program_weight_tensor_parameter_38: class Program_weight_tensor_parameter_39: name = "parameter_39" + original_name = "batch_norm2d_113.w_0" shape = [256] dtype = "float32" min_val = float("1.0") max_val = float("1.0") mean = float("1.0") + std = float("0.0") data = None class Program_weight_tensor_parameter_40: name = "parameter_40" + original_name = "batch_norm2d_113.w_2" shape = [256] dtype = "float32" min_val = float("1.0") max_val = float("1.0") mean = float("1.0") + std = float("0.0") data = None class Program_weight_tensor_parameter_41: name = "parameter_41" + original_name = "batch_norm2d_113.w_1" shape = [256] dtype = "float32" data = None @@ -370,6 +424,7 @@ class Program_weight_tensor_parameter_41: class Program_weight_tensor_parameter_42: name = "parameter_42" + original_name = "conv2d_113.w_0" shape = [256, 2048, 1, 1] dtype = "float32" min_val = float("-0.139592") @@ -381,6 +436,7 @@ class Program_weight_tensor_parameter_42: class Program_weight_tensor_parameter_43: name = "parameter_43" + original_name = "batch_norm2d_112.b_0" shape = [256] dtype = "float32" data = None @@ -388,26 +444,31 @@ class Program_weight_tensor_parameter_43: class Program_weight_tensor_parameter_44: name = "parameter_44" + original_name = "batch_norm2d_112.w_0" shape = [256] dtype = "float32" min_val = float("1.0") max_val = float("1.0") mean = float("1.0") + std = float("0.0") data = None class Program_weight_tensor_parameter_45: name = "parameter_45" + original_name = "batch_norm2d_112.w_2" shape = [256] dtype = "float32" min_val = float("1.0") max_val = float("1.0") mean = float("1.0") + std = float("0.0") data = None class Program_weight_tensor_parameter_46: name = "parameter_46" + original_name = "batch_norm2d_112.w_1" shape = [256] dtype = "float32" data = None @@ -415,6 +476,7 @@ class Program_weight_tensor_parameter_46: class Program_weight_tensor_parameter_47: name = "parameter_47" + original_name = "conv2d_112.b_0" shape = [256] dtype = "float32" data = None @@ -422,6 +484,7 @@ class Program_weight_tensor_parameter_47: class Program_weight_tensor_parameter_48: name = "parameter_48" + original_name = "conv2d_112.w_0" shape = [256, 2048, 1, 1] dtype = "float32" min_val = float("-0.13993") @@ -433,6 +496,7 @@ class Program_weight_tensor_parameter_48: class Program_weight_tensor_parameter_49: name = "parameter_49" + original_name = "batch_norm2d_111.b_0" shape = [2048] dtype = "float32" data = None @@ -440,26 +504,31 @@ class Program_weight_tensor_parameter_49: class Program_weight_tensor_parameter_50: name = "parameter_50" + original_name = "batch_norm2d_111.w_0" shape = [2048] dtype = "float32" min_val = float("1.0") max_val = float("1.0") mean = float("1.0") + std = float("0.0") data = None class Program_weight_tensor_parameter_51: name = "parameter_51" + original_name = "batch_norm2d_111.w_2" shape = [2048] dtype = "float32" min_val = float("1.0") max_val = float("1.0") mean = float("1.0") + std = float("0.0") data = None class Program_weight_tensor_parameter_52: name = "parameter_52" + original_name = "batch_norm2d_111.w_1" shape = [2048] dtype = "float32" data = None @@ -467,6 +536,7 @@ class Program_weight_tensor_parameter_52: class Program_weight_tensor_parameter_53: name = "parameter_53" + original_name = "conv2d_111.b_0" shape = [2048] dtype = "float32" data = None @@ -474,6 +544,7 @@ class Program_weight_tensor_parameter_53: class Program_weight_tensor_parameter_54: name = "parameter_54" + original_name = "conv2d_111.w_0" shape = [2048, 1, 3, 3] dtype = "float32" min_val = float("-0.0410822") @@ -485,6 +556,7 @@ class Program_weight_tensor_parameter_54: class Program_weight_tensor_parameter_55: name = "parameter_55" + original_name = "batch_norm2d_110.b_0" shape = [256] dtype = "float32" data = None @@ -492,26 +564,31 @@ class Program_weight_tensor_parameter_55: class Program_weight_tensor_parameter_56: name = "parameter_56" + original_name = "batch_norm2d_110.w_0" shape = [256] dtype = "float32" min_val = float("1.0") max_val = float("1.0") mean = float("1.0") + std = float("0.0") data = None class Program_weight_tensor_parameter_57: name = "parameter_57" + original_name = "batch_norm2d_110.w_2" shape = [256] dtype = "float32" min_val = float("1.0") max_val = float("1.0") mean = float("1.0") + std = float("0.0") data = None class Program_weight_tensor_parameter_58: name = "parameter_58" + original_name = "batch_norm2d_110.w_1" shape = [256] dtype = "float32" data = None @@ -519,6 +596,7 @@ class Program_weight_tensor_parameter_58: class Program_weight_tensor_parameter_59: name = "parameter_59" + original_name = "conv2d_110.b_0" shape = [256] dtype = "float32" data = None @@ -526,6 +604,7 @@ class Program_weight_tensor_parameter_59: class Program_weight_tensor_parameter_60: name = "parameter_60" + original_name = "conv2d_110.w_0" shape = [256, 2048, 1, 1] dtype = "float32" min_val = float("-0.166783") @@ -537,6 +616,7 @@ class Program_weight_tensor_parameter_60: class Program_weight_tensor_parameter_61: name = "parameter_61" + original_name = "batch_norm2d_109.b_0" shape = [2048] dtype = "float32" data = None @@ -544,26 +624,31 @@ class Program_weight_tensor_parameter_61: class Program_weight_tensor_parameter_62: name = "parameter_62" + original_name = "batch_norm2d_109.w_0" shape = [2048] dtype = "float32" min_val = float("1.0") max_val = float("1.0") mean = float("1.0") + std = float("0.0") data = None class Program_weight_tensor_parameter_63: name = "parameter_63" + original_name = "batch_norm2d_109.w_2" shape = [2048] dtype = "float32" min_val = float("1.0") max_val = float("1.0") mean = float("1.0") + std = float("0.0") data = None class Program_weight_tensor_parameter_64: name = "parameter_64" + original_name = "batch_norm2d_109.w_1" shape = [2048] dtype = "float32" data = None @@ -571,6 +656,7 @@ class Program_weight_tensor_parameter_64: class Program_weight_tensor_parameter_65: name = "parameter_65" + original_name = "conv2d_109.b_0" shape = [2048] dtype = "float32" data = None @@ -578,6 +664,7 @@ class Program_weight_tensor_parameter_65: class Program_weight_tensor_parameter_66: name = "parameter_66" + original_name = "conv2d_109.w_0" shape = [2048, 1, 3, 3] dtype = "float32" min_val = float("-0.0432663") @@ -589,6 +676,7 @@ class Program_weight_tensor_parameter_66: class Program_weight_tensor_parameter_67: name = "parameter_67" + original_name = "batch_norm2d_108.b_0" shape = [256] dtype = "float32" data = None @@ -596,26 +684,31 @@ class Program_weight_tensor_parameter_67: class Program_weight_tensor_parameter_68: name = "parameter_68" + original_name = "batch_norm2d_108.w_0" shape = [256] dtype = "float32" min_val = float("1.0") max_val = float("1.0") mean = float("1.0") + std = float("0.0") data = None class Program_weight_tensor_parameter_69: name = "parameter_69" + original_name = "batch_norm2d_108.w_2" shape = [256] dtype = "float32" min_val = float("1.0") max_val = float("1.0") mean = float("1.0") + std = float("0.0") data = None class Program_weight_tensor_parameter_70: name = "parameter_70" + original_name = "batch_norm2d_108.w_1" shape = [256] dtype = "float32" data = None @@ -623,6 +716,7 @@ class Program_weight_tensor_parameter_70: class Program_weight_tensor_parameter_71: name = "parameter_71" + original_name = "conv2d_108.b_0" shape = [256] dtype = "float32" data = None @@ -630,6 +724,7 @@ class Program_weight_tensor_parameter_71: class Program_weight_tensor_parameter_72: name = "parameter_72" + original_name = "conv2d_108.w_0" shape = [256, 2048, 1, 1] dtype = "float32" min_val = float("-0.148217") @@ -641,6 +736,7 @@ class Program_weight_tensor_parameter_72: class Program_weight_tensor_parameter_73: name = "parameter_73" + original_name = "batch_norm2d_107.b_0" shape = [2048] dtype = "float32" data = None @@ -648,26 +744,31 @@ class Program_weight_tensor_parameter_73: class Program_weight_tensor_parameter_74: name = "parameter_74" + original_name = "batch_norm2d_107.w_0" shape = [2048] dtype = "float32" min_val = float("1.0") max_val = float("1.0") mean = float("1.0") + std = float("0.0") data = None class Program_weight_tensor_parameter_75: name = "parameter_75" + original_name = "batch_norm2d_107.w_2" shape = [2048] dtype = "float32" min_val = float("1.0") max_val = float("1.0") mean = float("1.0") + std = float("0.0") data = None class Program_weight_tensor_parameter_76: name = "parameter_76" + original_name = "batch_norm2d_107.w_1" shape = [2048] dtype = "float32" data = None @@ -675,6 +776,7 @@ class Program_weight_tensor_parameter_76: class Program_weight_tensor_parameter_77: name = "parameter_77" + original_name = "conv2d_107.b_0" shape = [2048] dtype = "float32" data = None @@ -682,6 +784,7 @@ class Program_weight_tensor_parameter_77: class Program_weight_tensor_parameter_78: name = "parameter_78" + original_name = "conv2d_107.w_0" shape = [2048, 1, 3, 3] dtype = "float32" min_val = float("-0.0422734") @@ -693,6 +796,7 @@ class Program_weight_tensor_parameter_78: class Program_weight_tensor_parameter_79: name = "parameter_79" + original_name = "batch_norm2d_106.b_0" shape = [256] dtype = "float32" data = None @@ -700,26 +804,31 @@ class Program_weight_tensor_parameter_79: class Program_weight_tensor_parameter_80: name = "parameter_80" + original_name = "batch_norm2d_106.w_0" shape = [256] dtype = "float32" min_val = float("1.0") max_val = float("1.0") mean = float("1.0") + std = float("0.0") data = None class Program_weight_tensor_parameter_81: name = "parameter_81" + original_name = "batch_norm2d_106.w_2" shape = [256] dtype = "float32" min_val = float("1.0") max_val = float("1.0") mean = float("1.0") + std = float("0.0") data = None class Program_weight_tensor_parameter_82: name = "parameter_82" + original_name = "batch_norm2d_106.w_1" shape = [256] dtype = "float32" data = None @@ -727,6 +836,7 @@ class Program_weight_tensor_parameter_82: class Program_weight_tensor_parameter_83: name = "parameter_83" + original_name = "conv2d_106.b_0" shape = [256] dtype = "float32" data = None @@ -734,6 +844,7 @@ class Program_weight_tensor_parameter_83: class Program_weight_tensor_parameter_84: name = "parameter_84" + original_name = "conv2d_106.w_0" shape = [256, 2048, 1, 1] dtype = "float32" min_val = float("-0.144774") @@ -745,6 +856,7 @@ class Program_weight_tensor_parameter_84: class Program_weight_tensor_parameter_85: name = "parameter_85" + original_name = "batch_norm2d_105.b_0" shape = [2048] dtype = "float32" min_val = float("0.0827274") @@ -756,6 +868,7 @@ class Program_weight_tensor_parameter_85: class Program_weight_tensor_parameter_86: name = "parameter_86" + original_name = "batch_norm2d_105.w_0" shape = [2048] dtype = "float32" min_val = float("0.398192") @@ -767,6 +880,7 @@ class Program_weight_tensor_parameter_86: class Program_weight_tensor_parameter_87: name = "parameter_87" + original_name = "batch_norm2d_105.w_2" shape = [2048] dtype = "float32" min_val = float("0.000114954") @@ -778,6 +892,7 @@ class Program_weight_tensor_parameter_87: class Program_weight_tensor_parameter_88: name = "parameter_88" + original_name = "batch_norm2d_105.w_1" shape = [2048] dtype = "float32" min_val = float("-0.0425386") @@ -789,6 +904,7 @@ class Program_weight_tensor_parameter_88: class Program_weight_tensor_parameter_89: name = "parameter_89" + original_name = "conv2d_105.w_0" shape = [2048, 512, 1, 1] dtype = "float32" min_val = float("-0.133717") @@ -800,6 +916,7 @@ class Program_weight_tensor_parameter_89: class Program_weight_tensor_parameter_90: name = "parameter_90" + original_name = "batch_norm2d_104.b_0" shape = [512] dtype = "float32" min_val = float("-0.216342") @@ -811,6 +928,7 @@ class Program_weight_tensor_parameter_90: class Program_weight_tensor_parameter_91: name = "parameter_91" + original_name = "batch_norm2d_104.w_0" shape = [512] dtype = "float32" min_val = float("0.137888") @@ -822,6 +940,7 @@ class Program_weight_tensor_parameter_91: class Program_weight_tensor_parameter_92: name = "parameter_92" + original_name = "batch_norm2d_104.w_2" shape = [512] dtype = "float32" min_val = float("0.000717996") @@ -833,6 +952,7 @@ class Program_weight_tensor_parameter_92: class Program_weight_tensor_parameter_93: name = "parameter_93" + original_name = "batch_norm2d_104.w_1" shape = [512] dtype = "float32" min_val = float("-0.0644819") @@ -844,6 +964,7 @@ class Program_weight_tensor_parameter_93: class Program_weight_tensor_parameter_94: name = "parameter_94" + original_name = "conv2d_104.w_0" shape = [512, 512, 3, 3] dtype = "float32" min_val = float("-0.17946") @@ -855,6 +976,7 @@ class Program_weight_tensor_parameter_94: class Program_weight_tensor_parameter_95: name = "parameter_95" + original_name = "batch_norm2d_103.b_0" shape = [512] dtype = "float32" min_val = float("-0.273174") @@ -866,6 +988,7 @@ class Program_weight_tensor_parameter_95: class Program_weight_tensor_parameter_96: name = "parameter_96" + original_name = "batch_norm2d_103.w_0" shape = [512] dtype = "float32" min_val = float("0.0716476") @@ -877,6 +1000,7 @@ class Program_weight_tensor_parameter_96: class Program_weight_tensor_parameter_97: name = "parameter_97" + original_name = "batch_norm2d_103.w_2" shape = [512] dtype = "float32" min_val = float("0.000327466") @@ -888,6 +1012,7 @@ class Program_weight_tensor_parameter_97: class Program_weight_tensor_parameter_98: name = "parameter_98" + original_name = "batch_norm2d_103.w_1" shape = [512] dtype = "float32" min_val = float("-0.0710303") @@ -899,6 +1024,7 @@ class Program_weight_tensor_parameter_98: class Program_weight_tensor_parameter_99: name = "parameter_99" + original_name = "conv2d_103.w_0" shape = [512, 2048, 1, 1] dtype = "float32" min_val = float("-0.526026") @@ -910,6 +1036,7 @@ class Program_weight_tensor_parameter_99: class Program_weight_tensor_parameter_100: name = "parameter_100" + original_name = "batch_norm2d_102.b_0" shape = [2048] dtype = "float32" min_val = float("-0.291208") @@ -921,6 +1048,7 @@ class Program_weight_tensor_parameter_100: class Program_weight_tensor_parameter_101: name = "parameter_101" + original_name = "batch_norm2d_102.w_0" shape = [2048] dtype = "float32" min_val = float("-0.0102351") @@ -932,6 +1060,7 @@ class Program_weight_tensor_parameter_101: class Program_weight_tensor_parameter_102: name = "parameter_102" + original_name = "batch_norm2d_102.w_2" shape = [2048] dtype = "float32" min_val = float("7.00605e-05") @@ -943,6 +1072,7 @@ class Program_weight_tensor_parameter_102: class Program_weight_tensor_parameter_103: name = "parameter_103" + original_name = "batch_norm2d_102.w_1" shape = [2048] dtype = "float32" min_val = float("-0.0579006") @@ -954,6 +1084,7 @@ class Program_weight_tensor_parameter_103: class Program_weight_tensor_parameter_104: name = "parameter_104" + original_name = "conv2d_102.w_0" shape = [2048, 512, 1, 1] dtype = "float32" min_val = float("-0.293566") @@ -965,6 +1096,7 @@ class Program_weight_tensor_parameter_104: class Program_weight_tensor_parameter_105: name = "parameter_105" + original_name = "batch_norm2d_101.b_0" shape = [512] dtype = "float32" min_val = float("-0.263418") @@ -976,6 +1108,7 @@ class Program_weight_tensor_parameter_105: class Program_weight_tensor_parameter_106: name = "parameter_106" + original_name = "batch_norm2d_101.w_0" shape = [512] dtype = "float32" min_val = float("0.0750012") @@ -987,6 +1120,7 @@ class Program_weight_tensor_parameter_106: class Program_weight_tensor_parameter_107: name = "parameter_107" + original_name = "batch_norm2d_101.w_2" shape = [512] dtype = "float32" min_val = float("0.00142611") @@ -998,6 +1132,7 @@ class Program_weight_tensor_parameter_107: class Program_weight_tensor_parameter_108: name = "parameter_108" + original_name = "batch_norm2d_101.w_1" shape = [512] dtype = "float32" min_val = float("-0.264804") @@ -1009,6 +1144,7 @@ class Program_weight_tensor_parameter_108: class Program_weight_tensor_parameter_109: name = "parameter_109" + original_name = "conv2d_101.w_0" shape = [512, 512, 3, 3] dtype = "float32" min_val = float("-0.359546") @@ -1020,6 +1156,7 @@ class Program_weight_tensor_parameter_109: class Program_weight_tensor_parameter_110: name = "parameter_110" + original_name = "batch_norm2d_100.b_0" shape = [512] dtype = "float32" min_val = float("-0.197871") @@ -1031,6 +1168,7 @@ class Program_weight_tensor_parameter_110: class Program_weight_tensor_parameter_111: name = "parameter_111" + original_name = "batch_norm2d_100.w_0" shape = [512] dtype = "float32" min_val = float("0.0754074") @@ -1042,6 +1180,7 @@ class Program_weight_tensor_parameter_111: class Program_weight_tensor_parameter_112: name = "parameter_112" + original_name = "batch_norm2d_100.w_2" shape = [512] dtype = "float32" min_val = float("0.000242041") @@ -1053,6 +1192,7 @@ class Program_weight_tensor_parameter_112: class Program_weight_tensor_parameter_113: name = "parameter_113" + original_name = "batch_norm2d_100.w_1" shape = [512] dtype = "float32" min_val = float("-0.0453947") @@ -1064,6 +1204,7 @@ class Program_weight_tensor_parameter_113: class Program_weight_tensor_parameter_114: name = "parameter_114" + original_name = "conv2d_100.w_0" shape = [512, 2048, 1, 1] dtype = "float32" min_val = float("-0.301798") @@ -1075,6 +1216,7 @@ class Program_weight_tensor_parameter_114: class Program_weight_tensor_parameter_115: name = "parameter_115" + original_name = "batch_norm2d_99.b_0" shape = [2048] dtype = "float32" min_val = float("-0.148764") @@ -1086,6 +1228,7 @@ class Program_weight_tensor_parameter_115: class Program_weight_tensor_parameter_116: name = "parameter_116" + original_name = "batch_norm2d_99.w_0" shape = [2048] dtype = "float32" min_val = float("0.0435349") @@ -1097,6 +1240,7 @@ class Program_weight_tensor_parameter_116: class Program_weight_tensor_parameter_117: name = "parameter_117" + original_name = "batch_norm2d_99.w_2" shape = [2048] dtype = "float32" min_val = float("0.000769876") @@ -1108,6 +1252,7 @@ class Program_weight_tensor_parameter_117: class Program_weight_tensor_parameter_118: name = "parameter_118" + original_name = "batch_norm2d_99.w_1" shape = [2048] dtype = "float32" min_val = float("-0.352413") @@ -1119,6 +1264,7 @@ class Program_weight_tensor_parameter_118: class Program_weight_tensor_parameter_119: name = "parameter_119" + original_name = "conv2d_99.w_0" shape = [2048, 1024, 1, 1] dtype = "float32" min_val = float("-0.746467") @@ -1130,6 +1276,7 @@ class Program_weight_tensor_parameter_119: class Program_weight_tensor_parameter_120: name = "parameter_120" + original_name = "batch_norm2d_98.b_0" shape = [2048] dtype = "float32" min_val = float("-0.148764") @@ -1141,6 +1288,7 @@ class Program_weight_tensor_parameter_120: class Program_weight_tensor_parameter_121: name = "parameter_121" + original_name = "batch_norm2d_98.w_0" shape = [2048] dtype = "float32" min_val = float("-0.101896") @@ -1152,6 +1300,7 @@ class Program_weight_tensor_parameter_121: class Program_weight_tensor_parameter_122: name = "parameter_122" + original_name = "batch_norm2d_98.w_2" shape = [2048] dtype = "float32" min_val = float("4.32131e-05") @@ -1163,6 +1312,7 @@ class Program_weight_tensor_parameter_122: class Program_weight_tensor_parameter_123: name = "parameter_123" + original_name = "batch_norm2d_98.w_1" shape = [2048] dtype = "float32" min_val = float("-0.0303601") @@ -1174,6 +1324,7 @@ class Program_weight_tensor_parameter_123: class Program_weight_tensor_parameter_124: name = "parameter_124" + original_name = "conv2d_98.w_0" shape = [2048, 512, 1, 1] dtype = "float32" min_val = float("-0.100081") @@ -1185,6 +1336,7 @@ class Program_weight_tensor_parameter_124: class Program_weight_tensor_parameter_125: name = "parameter_125" + original_name = "batch_norm2d_97.b_0" shape = [512] dtype = "float32" min_val = float("-0.131196") @@ -1196,6 +1348,7 @@ class Program_weight_tensor_parameter_125: class Program_weight_tensor_parameter_126: name = "parameter_126" + original_name = "batch_norm2d_97.w_0" shape = [512] dtype = "float32" min_val = float("0.0656177") @@ -1207,6 +1360,7 @@ class Program_weight_tensor_parameter_126: class Program_weight_tensor_parameter_127: name = "parameter_127" + original_name = "batch_norm2d_97.w_2" shape = [512] dtype = "float32" min_val = float("0.000436737") @@ -1218,6 +1372,7 @@ class Program_weight_tensor_parameter_127: class Program_weight_tensor_parameter_128: name = "parameter_128" + original_name = "batch_norm2d_97.w_1" shape = [512] dtype = "float32" min_val = float("-0.333681") @@ -1229,6 +1384,7 @@ class Program_weight_tensor_parameter_128: class Program_weight_tensor_parameter_129: name = "parameter_129" + original_name = "conv2d_97.w_0" shape = [512, 512, 3, 3] dtype = "float32" min_val = float("-0.354123") @@ -1240,6 +1396,7 @@ class Program_weight_tensor_parameter_129: class Program_weight_tensor_parameter_130: name = "parameter_130" + original_name = "batch_norm2d_96.b_0" shape = [512] dtype = "float32" min_val = float("-0.195116") @@ -1251,6 +1408,7 @@ class Program_weight_tensor_parameter_130: class Program_weight_tensor_parameter_131: name = "parameter_131" + original_name = "batch_norm2d_96.w_0" shape = [512] dtype = "float32" min_val = float("0.0903273") @@ -1262,6 +1420,7 @@ class Program_weight_tensor_parameter_131: class Program_weight_tensor_parameter_132: name = "parameter_132" + original_name = "batch_norm2d_96.w_2" shape = [512] dtype = "float32" min_val = float("0.0020334") @@ -1273,6 +1432,7 @@ class Program_weight_tensor_parameter_132: class Program_weight_tensor_parameter_133: name = "parameter_133" + original_name = "batch_norm2d_96.w_1" shape = [512] dtype = "float32" min_val = float("-0.18973") @@ -1284,6 +1444,7 @@ class Program_weight_tensor_parameter_133: class Program_weight_tensor_parameter_134: name = "parameter_134" + original_name = "conv2d_96.w_0" shape = [512, 1024, 1, 1] dtype = "float32" min_val = float("-0.100513") @@ -1295,6 +1456,7 @@ class Program_weight_tensor_parameter_134: class Program_weight_tensor_parameter_135: name = "parameter_135" + original_name = "batch_norm2d_95.b_0" shape = [1024] dtype = "float32" min_val = float("-0.206403") @@ -1306,6 +1468,7 @@ class Program_weight_tensor_parameter_135: class Program_weight_tensor_parameter_136: name = "parameter_136" + original_name = "batch_norm2d_95.w_0" shape = [1024] dtype = "float32" min_val = float("-0.180513") @@ -1317,6 +1480,7 @@ class Program_weight_tensor_parameter_136: class Program_weight_tensor_parameter_137: name = "parameter_137" + original_name = "batch_norm2d_95.w_2" shape = [1024] dtype = "float32" min_val = float("1.12905e-05") @@ -1328,6 +1492,7 @@ class Program_weight_tensor_parameter_137: class Program_weight_tensor_parameter_138: name = "parameter_138" + original_name = "batch_norm2d_95.w_1" shape = [1024] dtype = "float32" min_val = float("-0.027597") @@ -1339,6 +1504,7 @@ class Program_weight_tensor_parameter_138: class Program_weight_tensor_parameter_139: name = "parameter_139" + original_name = "conv2d_95.w_0" shape = [1024, 256, 1, 1] dtype = "float32" min_val = float("-0.18543") @@ -1350,6 +1516,7 @@ class Program_weight_tensor_parameter_139: class Program_weight_tensor_parameter_140: name = "parameter_140" + original_name = "batch_norm2d_94.b_0" shape = [256] dtype = "float32" min_val = float("-0.265797") @@ -1361,6 +1528,7 @@ class Program_weight_tensor_parameter_140: class Program_weight_tensor_parameter_141: name = "parameter_141" + original_name = "batch_norm2d_94.w_0" shape = [256] dtype = "float32" min_val = float("0.071971") @@ -1372,6 +1540,7 @@ class Program_weight_tensor_parameter_141: class Program_weight_tensor_parameter_142: name = "parameter_142" + original_name = "batch_norm2d_94.w_2" shape = [256] dtype = "float32" min_val = float("0.000303703") @@ -1383,6 +1552,7 @@ class Program_weight_tensor_parameter_142: class Program_weight_tensor_parameter_143: name = "parameter_143" + original_name = "batch_norm2d_94.w_1" shape = [256] dtype = "float32" min_val = float("-0.0438911") @@ -1394,6 +1564,7 @@ class Program_weight_tensor_parameter_143: class Program_weight_tensor_parameter_144: name = "parameter_144" + original_name = "conv2d_94.w_0" shape = [256, 256, 3, 3] dtype = "float32" min_val = float("-0.146711") @@ -1405,6 +1576,7 @@ class Program_weight_tensor_parameter_144: class Program_weight_tensor_parameter_145: name = "parameter_145" + original_name = "batch_norm2d_93.b_0" shape = [256] dtype = "float32" min_val = float("-0.294242") @@ -1416,6 +1588,7 @@ class Program_weight_tensor_parameter_145: class Program_weight_tensor_parameter_146: name = "parameter_146" + original_name = "batch_norm2d_93.w_0" shape = [256] dtype = "float32" min_val = float("7.24191e-42") @@ -1427,6 +1600,7 @@ class Program_weight_tensor_parameter_146: class Program_weight_tensor_parameter_147: name = "parameter_147" + original_name = "batch_norm2d_93.w_2" shape = [256] dtype = "float32" min_val = float("5.60519e-45") @@ -1438,6 +1612,7 @@ class Program_weight_tensor_parameter_147: class Program_weight_tensor_parameter_148: name = "parameter_148" + original_name = "batch_norm2d_93.w_1" shape = [256] dtype = "float32" min_val = float("-0.241326") @@ -1449,6 +1624,7 @@ class Program_weight_tensor_parameter_148: class Program_weight_tensor_parameter_149: name = "parameter_149" + original_name = "conv2d_93.w_0" shape = [256, 1024, 1, 1] dtype = "float32" min_val = float("-0.285422") @@ -1460,6 +1636,7 @@ class Program_weight_tensor_parameter_149: class Program_weight_tensor_parameter_150: name = "parameter_150" + original_name = "batch_norm2d_92.b_0" shape = [1024] dtype = "float32" min_val = float("-0.138375") @@ -1471,6 +1648,7 @@ class Program_weight_tensor_parameter_150: class Program_weight_tensor_parameter_151: name = "parameter_151" + original_name = "batch_norm2d_92.w_0" shape = [1024] dtype = "float32" min_val = float("-0.0772375") @@ -1482,6 +1660,7 @@ class Program_weight_tensor_parameter_151: class Program_weight_tensor_parameter_152: name = "parameter_152" + original_name = "batch_norm2d_92.w_2" shape = [1024] dtype = "float32" min_val = float("9.79997e-06") @@ -1493,6 +1672,7 @@ class Program_weight_tensor_parameter_152: class Program_weight_tensor_parameter_153: name = "parameter_153" + original_name = "batch_norm2d_92.w_1" shape = [1024] dtype = "float32" min_val = float("-0.0297335") @@ -1504,6 +1684,7 @@ class Program_weight_tensor_parameter_153: class Program_weight_tensor_parameter_154: name = "parameter_154" + original_name = "conv2d_92.w_0" shape = [1024, 256, 1, 1] dtype = "float32" min_val = float("-0.227452") @@ -1515,6 +1696,7 @@ class Program_weight_tensor_parameter_154: class Program_weight_tensor_parameter_155: name = "parameter_155" + original_name = "batch_norm2d_91.b_0" shape = [256] dtype = "float32" min_val = float("-0.305461") @@ -1526,6 +1708,7 @@ class Program_weight_tensor_parameter_155: class Program_weight_tensor_parameter_156: name = "parameter_156" + original_name = "batch_norm2d_91.w_0" shape = [256] dtype = "float32" min_val = float("0.0733176") @@ -1537,6 +1720,7 @@ class Program_weight_tensor_parameter_156: class Program_weight_tensor_parameter_157: name = "parameter_157" + original_name = "batch_norm2d_91.w_2" shape = [256] dtype = "float32" min_val = float("0.000409322") @@ -1548,6 +1732,7 @@ class Program_weight_tensor_parameter_157: class Program_weight_tensor_parameter_158: name = "parameter_158" + original_name = "batch_norm2d_91.w_1" shape = [256] dtype = "float32" min_val = float("-0.0383431") @@ -1559,6 +1744,7 @@ class Program_weight_tensor_parameter_158: class Program_weight_tensor_parameter_159: name = "parameter_159" + original_name = "conv2d_91.w_0" shape = [256, 256, 3, 3] dtype = "float32" min_val = float("-0.130468") @@ -1570,6 +1756,7 @@ class Program_weight_tensor_parameter_159: class Program_weight_tensor_parameter_160: name = "parameter_160" + original_name = "batch_norm2d_90.b_0" shape = [256] dtype = "float32" min_val = float("-0.286373") @@ -1581,6 +1768,7 @@ class Program_weight_tensor_parameter_160: class Program_weight_tensor_parameter_161: name = "parameter_161" + original_name = "batch_norm2d_90.w_0" shape = [256] dtype = "float32" min_val = float("0.0675077") @@ -1592,6 +1780,7 @@ class Program_weight_tensor_parameter_161: class Program_weight_tensor_parameter_162: name = "parameter_162" + original_name = "batch_norm2d_90.w_2" shape = [256] dtype = "float32" min_val = float("0.00199446") @@ -1603,6 +1792,7 @@ class Program_weight_tensor_parameter_162: class Program_weight_tensor_parameter_163: name = "parameter_163" + original_name = "batch_norm2d_90.w_1" shape = [256] dtype = "float32" min_val = float("-0.152589") @@ -1614,6 +1804,7 @@ class Program_weight_tensor_parameter_163: class Program_weight_tensor_parameter_164: name = "parameter_164" + original_name = "conv2d_90.w_0" shape = [256, 1024, 1, 1] dtype = "float32" min_val = float("-0.131884") @@ -1625,6 +1816,7 @@ class Program_weight_tensor_parameter_164: class Program_weight_tensor_parameter_165: name = "parameter_165" + original_name = "batch_norm2d_89.b_0" shape = [1024] dtype = "float32" min_val = float("-0.192306") @@ -1636,6 +1828,7 @@ class Program_weight_tensor_parameter_165: class Program_weight_tensor_parameter_166: name = "parameter_166" + original_name = "batch_norm2d_89.w_0" shape = [1024] dtype = "float32" min_val = float("-0.101866") @@ -1647,6 +1840,7 @@ class Program_weight_tensor_parameter_166: class Program_weight_tensor_parameter_167: name = "parameter_167" + original_name = "batch_norm2d_89.w_2" shape = [1024] dtype = "float32" min_val = float("2.95398e-07") @@ -1658,6 +1852,7 @@ class Program_weight_tensor_parameter_167: class Program_weight_tensor_parameter_168: name = "parameter_168" + original_name = "batch_norm2d_89.w_1" shape = [1024] dtype = "float32" min_val = float("-0.0293132") @@ -1669,6 +1864,7 @@ class Program_weight_tensor_parameter_168: class Program_weight_tensor_parameter_169: name = "parameter_169" + original_name = "conv2d_89.w_0" shape = [1024, 256, 1, 1] dtype = "float32" min_val = float("-0.18324") @@ -1680,6 +1876,7 @@ class Program_weight_tensor_parameter_169: class Program_weight_tensor_parameter_170: name = "parameter_170" + original_name = "batch_norm2d_88.b_0" shape = [256] dtype = "float32" min_val = float("-0.151589") @@ -1691,6 +1888,7 @@ class Program_weight_tensor_parameter_170: class Program_weight_tensor_parameter_171: name = "parameter_171" + original_name = "batch_norm2d_88.w_0" shape = [256] dtype = "float32" min_val = float("0.0669419") @@ -1702,6 +1900,7 @@ class Program_weight_tensor_parameter_171: class Program_weight_tensor_parameter_172: name = "parameter_172" + original_name = "batch_norm2d_88.w_2" shape = [256] dtype = "float32" min_val = float("0.00035768") @@ -1713,6 +1912,7 @@ class Program_weight_tensor_parameter_172: class Program_weight_tensor_parameter_173: name = "parameter_173" + original_name = "batch_norm2d_88.w_1" shape = [256] dtype = "float32" min_val = float("-0.070715") @@ -1724,6 +1924,7 @@ class Program_weight_tensor_parameter_173: class Program_weight_tensor_parameter_174: name = "parameter_174" + original_name = "conv2d_88.w_0" shape = [256, 256, 3, 3] dtype = "float32" min_val = float("-0.127768") @@ -1735,6 +1936,7 @@ class Program_weight_tensor_parameter_174: class Program_weight_tensor_parameter_175: name = "parameter_175" + original_name = "batch_norm2d_87.b_0" shape = [256] dtype = "float32" min_val = float("-0.308452") @@ -1746,6 +1948,7 @@ class Program_weight_tensor_parameter_175: class Program_weight_tensor_parameter_176: name = "parameter_176" + original_name = "batch_norm2d_87.w_0" shape = [256] dtype = "float32" min_val = float("0.0696605") @@ -1757,6 +1960,7 @@ class Program_weight_tensor_parameter_176: class Program_weight_tensor_parameter_177: name = "parameter_177" + original_name = "batch_norm2d_87.w_2" shape = [256] dtype = "float32" min_val = float("0.00196375") @@ -1768,6 +1972,7 @@ class Program_weight_tensor_parameter_177: class Program_weight_tensor_parameter_178: name = "parameter_178" + original_name = "batch_norm2d_87.w_1" shape = [256] dtype = "float32" min_val = float("-0.183315") @@ -1779,6 +1984,7 @@ class Program_weight_tensor_parameter_178: class Program_weight_tensor_parameter_179: name = "parameter_179" + original_name = "conv2d_87.w_0" shape = [256, 1024, 1, 1] dtype = "float32" min_val = float("-0.327476") @@ -1790,6 +1996,7 @@ class Program_weight_tensor_parameter_179: class Program_weight_tensor_parameter_180: name = "parameter_180" + original_name = "batch_norm2d_86.b_0" shape = [1024] dtype = "float32" min_val = float("-0.157325") @@ -1801,6 +2008,7 @@ class Program_weight_tensor_parameter_180: class Program_weight_tensor_parameter_181: name = "parameter_181" + original_name = "batch_norm2d_86.w_0" shape = [1024] dtype = "float32" min_val = float("-0.155552") @@ -1812,6 +2020,7 @@ class Program_weight_tensor_parameter_181: class Program_weight_tensor_parameter_182: name = "parameter_182" + original_name = "batch_norm2d_86.w_2" shape = [1024] dtype = "float32" min_val = float("4.43334e-07") @@ -1823,6 +2032,7 @@ class Program_weight_tensor_parameter_182: class Program_weight_tensor_parameter_183: name = "parameter_183" + original_name = "batch_norm2d_86.w_1" shape = [1024] dtype = "float32" min_val = float("-0.0202518") @@ -1834,6 +2044,7 @@ class Program_weight_tensor_parameter_183: class Program_weight_tensor_parameter_184: name = "parameter_184" + original_name = "conv2d_86.w_0" shape = [1024, 256, 1, 1] dtype = "float32" min_val = float("-0.165613") @@ -1845,6 +2056,7 @@ class Program_weight_tensor_parameter_184: class Program_weight_tensor_parameter_185: name = "parameter_185" + original_name = "batch_norm2d_85.b_0" shape = [256] dtype = "float32" min_val = float("-0.311108") @@ -1856,6 +2068,7 @@ class Program_weight_tensor_parameter_185: class Program_weight_tensor_parameter_186: name = "parameter_186" + original_name = "batch_norm2d_85.w_0" shape = [256] dtype = "float32" min_val = float("0.0572317") @@ -1867,6 +2080,7 @@ class Program_weight_tensor_parameter_186: class Program_weight_tensor_parameter_187: name = "parameter_187" + original_name = "batch_norm2d_85.w_2" shape = [256] dtype = "float32" min_val = float("0.000339152") @@ -1878,6 +2092,7 @@ class Program_weight_tensor_parameter_187: class Program_weight_tensor_parameter_188: name = "parameter_188" + original_name = "batch_norm2d_85.w_1" shape = [256] dtype = "float32" min_val = float("-0.0636495") @@ -1889,6 +2104,7 @@ class Program_weight_tensor_parameter_188: class Program_weight_tensor_parameter_189: name = "parameter_189" + original_name = "conv2d_85.w_0" shape = [256, 256, 3, 3] dtype = "float32" min_val = float("-0.132645") @@ -1900,6 +2116,7 @@ class Program_weight_tensor_parameter_189: class Program_weight_tensor_parameter_190: name = "parameter_190" + original_name = "batch_norm2d_84.b_0" shape = [256] dtype = "float32" min_val = float("-0.183586") @@ -1911,6 +2128,7 @@ class Program_weight_tensor_parameter_190: class Program_weight_tensor_parameter_191: name = "parameter_191" + original_name = "batch_norm2d_84.w_0" shape = [256] dtype = "float32" min_val = float("0.0546606") @@ -1922,6 +2140,7 @@ class Program_weight_tensor_parameter_191: class Program_weight_tensor_parameter_192: name = "parameter_192" + original_name = "batch_norm2d_84.w_2" shape = [256] dtype = "float32" min_val = float("0.00170534") @@ -1933,6 +2152,7 @@ class Program_weight_tensor_parameter_192: class Program_weight_tensor_parameter_193: name = "parameter_193" + original_name = "batch_norm2d_84.w_1" shape = [256] dtype = "float32" min_val = float("-0.172921") @@ -1944,6 +2164,7 @@ class Program_weight_tensor_parameter_193: class Program_weight_tensor_parameter_194: name = "parameter_194" + original_name = "conv2d_84.w_0" shape = [256, 1024, 1, 1] dtype = "float32" min_val = float("-0.155053") @@ -1955,6 +2176,7 @@ class Program_weight_tensor_parameter_194: class Program_weight_tensor_parameter_195: name = "parameter_195" + original_name = "batch_norm2d_83.b_0" shape = [1024] dtype = "float32" min_val = float("-0.2296") @@ -1966,6 +2188,7 @@ class Program_weight_tensor_parameter_195: class Program_weight_tensor_parameter_196: name = "parameter_196" + original_name = "batch_norm2d_83.w_0" shape = [1024] dtype = "float32" min_val = float("-0.074123") @@ -1977,6 +2200,7 @@ class Program_weight_tensor_parameter_196: class Program_weight_tensor_parameter_197: name = "parameter_197" + original_name = "batch_norm2d_83.w_2" shape = [1024] dtype = "float32" min_val = float("1.59533e-06") @@ -1988,6 +2212,7 @@ class Program_weight_tensor_parameter_197: class Program_weight_tensor_parameter_198: name = "parameter_198" + original_name = "batch_norm2d_83.w_1" shape = [1024] dtype = "float32" min_val = float("-0.11021") @@ -1999,6 +2224,7 @@ class Program_weight_tensor_parameter_198: class Program_weight_tensor_parameter_199: name = "parameter_199" + original_name = "conv2d_83.w_0" shape = [1024, 256, 1, 1] dtype = "float32" min_val = float("-0.479697") @@ -2010,6 +2236,7 @@ class Program_weight_tensor_parameter_199: class Program_weight_tensor_parameter_200: name = "parameter_200" + original_name = "batch_norm2d_82.b_0" shape = [256] dtype = "float32" min_val = float("-0.136562") @@ -2021,6 +2248,7 @@ class Program_weight_tensor_parameter_200: class Program_weight_tensor_parameter_201: name = "parameter_201" + original_name = "batch_norm2d_82.w_0" shape = [256] dtype = "float32" min_val = float("0.0475024") @@ -2032,6 +2260,7 @@ class Program_weight_tensor_parameter_201: class Program_weight_tensor_parameter_202: name = "parameter_202" + original_name = "batch_norm2d_82.w_2" shape = [256] dtype = "float32" min_val = float("0.000246573") @@ -2043,6 +2272,7 @@ class Program_weight_tensor_parameter_202: class Program_weight_tensor_parameter_203: name = "parameter_203" + original_name = "batch_norm2d_82.w_1" shape = [256] dtype = "float32" min_val = float("-0.0781337") @@ -2054,6 +2284,7 @@ class Program_weight_tensor_parameter_203: class Program_weight_tensor_parameter_204: name = "parameter_204" + original_name = "conv2d_82.w_0" shape = [256, 256, 3, 3] dtype = "float32" min_val = float("-0.245573") @@ -2065,6 +2296,7 @@ class Program_weight_tensor_parameter_204: class Program_weight_tensor_parameter_205: name = "parameter_205" + original_name = "batch_norm2d_81.b_0" shape = [256] dtype = "float32" min_val = float("-0.162884") @@ -2076,6 +2308,7 @@ class Program_weight_tensor_parameter_205: class Program_weight_tensor_parameter_206: name = "parameter_206" + original_name = "batch_norm2d_81.w_0" shape = [256] dtype = "float32" min_val = float("0.0544083") @@ -2087,6 +2320,7 @@ class Program_weight_tensor_parameter_206: class Program_weight_tensor_parameter_207: name = "parameter_207" + original_name = "batch_norm2d_81.w_2" shape = [256] dtype = "float32" min_val = float("0.00184023") @@ -2098,6 +2332,7 @@ class Program_weight_tensor_parameter_207: class Program_weight_tensor_parameter_208: name = "parameter_208" + original_name = "batch_norm2d_81.w_1" shape = [256] dtype = "float32" min_val = float("-0.134317") @@ -2109,6 +2344,7 @@ class Program_weight_tensor_parameter_208: class Program_weight_tensor_parameter_209: name = "parameter_209" + original_name = "conv2d_81.w_0" shape = [256, 1024, 1, 1] dtype = "float32" min_val = float("-0.219317") @@ -2120,6 +2356,7 @@ class Program_weight_tensor_parameter_209: class Program_weight_tensor_parameter_210: name = "parameter_210" + original_name = "batch_norm2d_80.b_0" shape = [1024] dtype = "float32" min_val = float("-0.118821") @@ -2131,6 +2368,7 @@ class Program_weight_tensor_parameter_210: class Program_weight_tensor_parameter_211: name = "parameter_211" + original_name = "batch_norm2d_80.w_0" shape = [1024] dtype = "float32" min_val = float("-0.0644223") @@ -2142,6 +2380,7 @@ class Program_weight_tensor_parameter_211: class Program_weight_tensor_parameter_212: name = "parameter_212" + original_name = "batch_norm2d_80.w_2" shape = [1024] dtype = "float32" min_val = float("1.10508e-27") @@ -2153,6 +2392,7 @@ class Program_weight_tensor_parameter_212: class Program_weight_tensor_parameter_213: name = "parameter_213" + original_name = "batch_norm2d_80.w_1" shape = [1024] dtype = "float32" min_val = float("-0.017754") @@ -2164,6 +2404,7 @@ class Program_weight_tensor_parameter_213: class Program_weight_tensor_parameter_214: name = "parameter_214" + original_name = "conv2d_80.w_0" shape = [1024, 256, 1, 1] dtype = "float32" min_val = float("-0.19161") @@ -2175,6 +2416,7 @@ class Program_weight_tensor_parameter_214: class Program_weight_tensor_parameter_215: name = "parameter_215" + original_name = "batch_norm2d_79.b_0" shape = [256] dtype = "float32" min_val = float("-0.128163") @@ -2186,6 +2428,7 @@ class Program_weight_tensor_parameter_215: class Program_weight_tensor_parameter_216: name = "parameter_216" + original_name = "batch_norm2d_79.w_0" shape = [256] dtype = "float32" min_val = float("0.0506961") @@ -2197,6 +2440,7 @@ class Program_weight_tensor_parameter_216: class Program_weight_tensor_parameter_217: name = "parameter_217" + original_name = "batch_norm2d_79.w_2" shape = [256] dtype = "float32" min_val = float("0.000352798") @@ -2208,6 +2452,7 @@ class Program_weight_tensor_parameter_217: class Program_weight_tensor_parameter_218: name = "parameter_218" + original_name = "batch_norm2d_79.w_1" shape = [256] dtype = "float32" min_val = float("-0.0403924") @@ -2219,6 +2464,7 @@ class Program_weight_tensor_parameter_218: class Program_weight_tensor_parameter_219: name = "parameter_219" + original_name = "conv2d_79.w_0" shape = [256, 256, 3, 3] dtype = "float32" min_val = float("-0.13259") @@ -2230,6 +2476,7 @@ class Program_weight_tensor_parameter_219: class Program_weight_tensor_parameter_220: name = "parameter_220" + original_name = "batch_norm2d_78.b_0" shape = [256] dtype = "float32" min_val = float("-0.161042") @@ -2241,6 +2488,7 @@ class Program_weight_tensor_parameter_220: class Program_weight_tensor_parameter_221: name = "parameter_221" + original_name = "batch_norm2d_78.w_0" shape = [256] dtype = "float32" min_val = float("0.0516501") @@ -2252,6 +2500,7 @@ class Program_weight_tensor_parameter_221: class Program_weight_tensor_parameter_222: name = "parameter_222" + original_name = "batch_norm2d_78.w_2" shape = [256] dtype = "float32" min_val = float("0.00200027") @@ -2263,6 +2512,7 @@ class Program_weight_tensor_parameter_222: class Program_weight_tensor_parameter_223: name = "parameter_223" + original_name = "batch_norm2d_78.w_1" shape = [256] dtype = "float32" min_val = float("-0.112165") @@ -2274,6 +2524,7 @@ class Program_weight_tensor_parameter_223: class Program_weight_tensor_parameter_224: name = "parameter_224" + original_name = "conv2d_78.w_0" shape = [256, 1024, 1, 1] dtype = "float32" min_val = float("-0.209198") @@ -2285,6 +2536,7 @@ class Program_weight_tensor_parameter_224: class Program_weight_tensor_parameter_225: name = "parameter_225" + original_name = "batch_norm2d_77.b_0" shape = [1024] dtype = "float32" min_val = float("-0.144365") @@ -2296,6 +2548,7 @@ class Program_weight_tensor_parameter_225: class Program_weight_tensor_parameter_226: name = "parameter_226" + original_name = "batch_norm2d_77.w_0" shape = [1024] dtype = "float32" min_val = float("-0.0877915") @@ -2307,6 +2560,7 @@ class Program_weight_tensor_parameter_226: class Program_weight_tensor_parameter_227: name = "parameter_227" + original_name = "batch_norm2d_77.w_2" shape = [1024] dtype = "float32" min_val = float("5.36924e-07") @@ -2318,6 +2572,7 @@ class Program_weight_tensor_parameter_227: class Program_weight_tensor_parameter_228: name = "parameter_228" + original_name = "batch_norm2d_77.w_1" shape = [1024] dtype = "float32" min_val = float("-0.0667076") @@ -2329,6 +2584,7 @@ class Program_weight_tensor_parameter_228: class Program_weight_tensor_parameter_229: name = "parameter_229" + original_name = "conv2d_77.w_0" shape = [1024, 256, 1, 1] dtype = "float32" min_val = float("-0.373221") @@ -2340,6 +2596,7 @@ class Program_weight_tensor_parameter_229: class Program_weight_tensor_parameter_230: name = "parameter_230" + original_name = "batch_norm2d_76.b_0" shape = [256] dtype = "float32" min_val = float("-0.148653") @@ -2351,6 +2608,7 @@ class Program_weight_tensor_parameter_230: class Program_weight_tensor_parameter_231: name = "parameter_231" + original_name = "batch_norm2d_76.w_0" shape = [256] dtype = "float32" min_val = float("0.0469178") @@ -2362,6 +2620,7 @@ class Program_weight_tensor_parameter_231: class Program_weight_tensor_parameter_232: name = "parameter_232" + original_name = "batch_norm2d_76.w_2" shape = [256] dtype = "float32" min_val = float("0.000240204") @@ -2373,6 +2632,7 @@ class Program_weight_tensor_parameter_232: class Program_weight_tensor_parameter_233: name = "parameter_233" + original_name = "batch_norm2d_76.w_1" shape = [256] dtype = "float32" min_val = float("-0.0457173") @@ -2384,6 +2644,7 @@ class Program_weight_tensor_parameter_233: class Program_weight_tensor_parameter_234: name = "parameter_234" + original_name = "conv2d_76.w_0" shape = [256, 256, 3, 3] dtype = "float32" min_val = float("-0.268807") @@ -2395,6 +2656,7 @@ class Program_weight_tensor_parameter_234: class Program_weight_tensor_parameter_235: name = "parameter_235" + original_name = "batch_norm2d_75.b_0" shape = [256] dtype = "float32" min_val = float("-0.196785") @@ -2406,6 +2668,7 @@ class Program_weight_tensor_parameter_235: class Program_weight_tensor_parameter_236: name = "parameter_236" + original_name = "batch_norm2d_75.w_0" shape = [256] dtype = "float32" min_val = float("0.0504373") @@ -2417,6 +2680,7 @@ class Program_weight_tensor_parameter_236: class Program_weight_tensor_parameter_237: name = "parameter_237" + original_name = "batch_norm2d_75.w_2" shape = [256] dtype = "float32" min_val = float("0.00154363") @@ -2428,6 +2692,7 @@ class Program_weight_tensor_parameter_237: class Program_weight_tensor_parameter_238: name = "parameter_238" + original_name = "batch_norm2d_75.w_1" shape = [256] dtype = "float32" min_val = float("-0.103539") @@ -2439,6 +2704,7 @@ class Program_weight_tensor_parameter_238: class Program_weight_tensor_parameter_239: name = "parameter_239" + original_name = "conv2d_75.w_0" shape = [256, 1024, 1, 1] dtype = "float32" min_val = float("-0.173457") @@ -2450,6 +2716,7 @@ class Program_weight_tensor_parameter_239: class Program_weight_tensor_parameter_240: name = "parameter_240" + original_name = "batch_norm2d_74.b_0" shape = [1024] dtype = "float32" min_val = float("-0.154772") @@ -2461,6 +2728,7 @@ class Program_weight_tensor_parameter_240: class Program_weight_tensor_parameter_241: name = "parameter_241" + original_name = "batch_norm2d_74.w_0" shape = [1024] dtype = "float32" min_val = float("-0.0862494") @@ -2472,6 +2740,7 @@ class Program_weight_tensor_parameter_241: class Program_weight_tensor_parameter_242: name = "parameter_242" + original_name = "batch_norm2d_74.w_2" shape = [1024] dtype = "float32" min_val = float("1.21888e-10") @@ -2483,6 +2752,7 @@ class Program_weight_tensor_parameter_242: class Program_weight_tensor_parameter_243: name = "parameter_243" + original_name = "batch_norm2d_74.w_1" shape = [1024] dtype = "float32" min_val = float("-0.0330924") @@ -2494,6 +2764,7 @@ class Program_weight_tensor_parameter_243: class Program_weight_tensor_parameter_244: name = "parameter_244" + original_name = "conv2d_74.w_0" shape = [1024, 256, 1, 1] dtype = "float32" min_val = float("-0.18676") @@ -2505,6 +2776,7 @@ class Program_weight_tensor_parameter_244: class Program_weight_tensor_parameter_245: name = "parameter_245" + original_name = "batch_norm2d_73.b_0" shape = [256] dtype = "float32" min_val = float("-0.292217") @@ -2516,6 +2788,7 @@ class Program_weight_tensor_parameter_245: class Program_weight_tensor_parameter_246: name = "parameter_246" + original_name = "batch_norm2d_73.w_0" shape = [256] dtype = "float32" min_val = float("0.0585411") @@ -2527,6 +2800,7 @@ class Program_weight_tensor_parameter_246: class Program_weight_tensor_parameter_247: name = "parameter_247" + original_name = "batch_norm2d_73.w_2" shape = [256] dtype = "float32" min_val = float("0.00036765") @@ -2538,6 +2812,7 @@ class Program_weight_tensor_parameter_247: class Program_weight_tensor_parameter_248: name = "parameter_248" + original_name = "batch_norm2d_73.w_1" shape = [256] dtype = "float32" min_val = float("-0.0637849") @@ -2549,6 +2824,7 @@ class Program_weight_tensor_parameter_248: class Program_weight_tensor_parameter_249: name = "parameter_249" + original_name = "conv2d_73.w_0" shape = [256, 256, 3, 3] dtype = "float32" min_val = float("-0.122804") @@ -2560,6 +2836,7 @@ class Program_weight_tensor_parameter_249: class Program_weight_tensor_parameter_250: name = "parameter_250" + original_name = "batch_norm2d_72.b_0" shape = [256] dtype = "float32" min_val = float("-0.1479") @@ -2571,6 +2848,7 @@ class Program_weight_tensor_parameter_250: class Program_weight_tensor_parameter_251: name = "parameter_251" + original_name = "batch_norm2d_72.w_0" shape = [256] dtype = "float32" min_val = float("6.39272e-42") @@ -2582,6 +2860,7 @@ class Program_weight_tensor_parameter_251: class Program_weight_tensor_parameter_252: name = "parameter_252" + original_name = "batch_norm2d_72.w_2" shape = [256] dtype = "float32" min_val = float("5.60519e-45") @@ -2593,6 +2872,7 @@ class Program_weight_tensor_parameter_252: class Program_weight_tensor_parameter_253: name = "parameter_253" + original_name = "batch_norm2d_72.w_1" shape = [256] dtype = "float32" min_val = float("-0.119982") @@ -2604,6 +2884,7 @@ class Program_weight_tensor_parameter_253: class Program_weight_tensor_parameter_254: name = "parameter_254" + original_name = "conv2d_72.w_0" shape = [256, 1024, 1, 1] dtype = "float32" min_val = float("-0.17648") @@ -2615,6 +2896,7 @@ class Program_weight_tensor_parameter_254: class Program_weight_tensor_parameter_255: name = "parameter_255" + original_name = "batch_norm2d_71.b_0" shape = [1024] dtype = "float32" min_val = float("-0.181404") @@ -2626,6 +2908,7 @@ class Program_weight_tensor_parameter_255: class Program_weight_tensor_parameter_256: name = "parameter_256" + original_name = "batch_norm2d_71.w_0" shape = [1024] dtype = "float32" min_val = float("-0.067036") @@ -2637,6 +2920,7 @@ class Program_weight_tensor_parameter_256: class Program_weight_tensor_parameter_257: name = "parameter_257" + original_name = "batch_norm2d_71.w_2" shape = [1024] dtype = "float32" min_val = float("1.25122e-06") @@ -2648,6 +2932,7 @@ class Program_weight_tensor_parameter_257: class Program_weight_tensor_parameter_258: name = "parameter_258" + original_name = "batch_norm2d_71.w_1" shape = [1024] dtype = "float32" min_val = float("-0.0225367") @@ -2659,6 +2944,7 @@ class Program_weight_tensor_parameter_258: class Program_weight_tensor_parameter_259: name = "parameter_259" + original_name = "conv2d_71.w_0" shape = [1024, 256, 1, 1] dtype = "float32" min_val = float("-0.203571") @@ -2670,6 +2956,7 @@ class Program_weight_tensor_parameter_259: class Program_weight_tensor_parameter_260: name = "parameter_260" + original_name = "batch_norm2d_70.b_0" shape = [256] dtype = "float32" min_val = float("-0.142678") @@ -2681,6 +2968,7 @@ class Program_weight_tensor_parameter_260: class Program_weight_tensor_parameter_261: name = "parameter_261" + original_name = "batch_norm2d_70.w_0" shape = [256] dtype = "float32" min_val = float("0.0525285") @@ -2692,6 +2980,7 @@ class Program_weight_tensor_parameter_261: class Program_weight_tensor_parameter_262: name = "parameter_262" + original_name = "batch_norm2d_70.w_2" shape = [256] dtype = "float32" min_val = float("0.000315005") @@ -2703,6 +2992,7 @@ class Program_weight_tensor_parameter_262: class Program_weight_tensor_parameter_263: name = "parameter_263" + original_name = "batch_norm2d_70.w_1" shape = [256] dtype = "float32" min_val = float("-0.0438057") @@ -2714,6 +3004,7 @@ class Program_weight_tensor_parameter_263: class Program_weight_tensor_parameter_264: name = "parameter_264" + original_name = "conv2d_70.w_0" shape = [256, 256, 3, 3] dtype = "float32" min_val = float("-0.157069") @@ -2725,6 +3016,7 @@ class Program_weight_tensor_parameter_264: class Program_weight_tensor_parameter_265: name = "parameter_265" + original_name = "batch_norm2d_69.b_0" shape = [256] dtype = "float32" min_val = float("-0.263298") @@ -2736,6 +3028,7 @@ class Program_weight_tensor_parameter_265: class Program_weight_tensor_parameter_266: name = "parameter_266" + original_name = "batch_norm2d_69.w_0" shape = [256] dtype = "float32" min_val = float("-4.54021e-42") @@ -2747,6 +3040,7 @@ class Program_weight_tensor_parameter_266: class Program_weight_tensor_parameter_267: name = "parameter_267" + original_name = "batch_norm2d_69.w_2" shape = [256] dtype = "float32" min_val = float("5.60519e-45") @@ -2758,6 +3052,7 @@ class Program_weight_tensor_parameter_267: class Program_weight_tensor_parameter_268: name = "parameter_268" + original_name = "batch_norm2d_69.w_1" shape = [256] dtype = "float32" min_val = float("-0.101473") @@ -2769,6 +3064,7 @@ class Program_weight_tensor_parameter_268: class Program_weight_tensor_parameter_269: name = "parameter_269" + original_name = "conv2d_69.w_0" shape = [256, 1024, 1, 1] dtype = "float32" min_val = float("-0.184965") @@ -2780,6 +3076,7 @@ class Program_weight_tensor_parameter_269: class Program_weight_tensor_parameter_270: name = "parameter_270" + original_name = "batch_norm2d_68.b_0" shape = [1024] dtype = "float32" min_val = float("-0.158868") @@ -2791,6 +3088,7 @@ class Program_weight_tensor_parameter_270: class Program_weight_tensor_parameter_271: name = "parameter_271" + original_name = "batch_norm2d_68.w_0" shape = [1024] dtype = "float32" min_val = float("-0.0788502") @@ -2802,6 +3100,7 @@ class Program_weight_tensor_parameter_271: class Program_weight_tensor_parameter_272: name = "parameter_272" + original_name = "batch_norm2d_68.w_2" shape = [1024] dtype = "float32" min_val = float("9.42686e-07") @@ -2813,6 +3112,7 @@ class Program_weight_tensor_parameter_272: class Program_weight_tensor_parameter_273: name = "parameter_273" + original_name = "batch_norm2d_68.w_1" shape = [1024] dtype = "float32" min_val = float("-0.0225805") @@ -2824,6 +3124,7 @@ class Program_weight_tensor_parameter_273: class Program_weight_tensor_parameter_274: name = "parameter_274" + original_name = "conv2d_68.w_0" shape = [1024, 256, 1, 1] dtype = "float32" min_val = float("-0.193111") @@ -2835,6 +3136,7 @@ class Program_weight_tensor_parameter_274: class Program_weight_tensor_parameter_275: name = "parameter_275" + original_name = "batch_norm2d_67.b_0" shape = [256] dtype = "float32" min_val = float("-0.126881") @@ -2846,6 +3148,7 @@ class Program_weight_tensor_parameter_275: class Program_weight_tensor_parameter_276: name = "parameter_276" + original_name = "batch_norm2d_67.w_0" shape = [256] dtype = "float32" min_val = float("0.0538402") @@ -2857,6 +3160,7 @@ class Program_weight_tensor_parameter_276: class Program_weight_tensor_parameter_277: name = "parameter_277" + original_name = "batch_norm2d_67.w_2" shape = [256] dtype = "float32" min_val = float("0.00048507") @@ -2868,6 +3172,7 @@ class Program_weight_tensor_parameter_277: class Program_weight_tensor_parameter_278: name = "parameter_278" + original_name = "batch_norm2d_67.w_1" shape = [256] dtype = "float32" min_val = float("-0.0562529") @@ -2879,6 +3184,7 @@ class Program_weight_tensor_parameter_278: class Program_weight_tensor_parameter_279: name = "parameter_279" + original_name = "conv2d_67.w_0" shape = [256, 256, 3, 3] dtype = "float32" min_val = float("-0.111533") @@ -2890,6 +3196,7 @@ class Program_weight_tensor_parameter_279: class Program_weight_tensor_parameter_280: name = "parameter_280" + original_name = "batch_norm2d_66.b_0" shape = [256] dtype = "float32" min_val = float("-0.177625") @@ -2901,6 +3208,7 @@ class Program_weight_tensor_parameter_280: class Program_weight_tensor_parameter_281: name = "parameter_281" + original_name = "batch_norm2d_66.w_0" shape = [256] dtype = "float32" min_val = float("0.0510496") @@ -2912,6 +3220,7 @@ class Program_weight_tensor_parameter_281: class Program_weight_tensor_parameter_282: name = "parameter_282" + original_name = "batch_norm2d_66.w_2" shape = [256] dtype = "float32" min_val = float("0.00170635") @@ -2923,6 +3232,7 @@ class Program_weight_tensor_parameter_282: class Program_weight_tensor_parameter_283: name = "parameter_283" + original_name = "batch_norm2d_66.w_1" shape = [256] dtype = "float32" min_val = float("-0.147881") @@ -2934,6 +3244,7 @@ class Program_weight_tensor_parameter_283: class Program_weight_tensor_parameter_284: name = "parameter_284" + original_name = "conv2d_66.w_0" shape = [256, 1024, 1, 1] dtype = "float32" min_val = float("-0.190949") @@ -2945,6 +3256,7 @@ class Program_weight_tensor_parameter_284: class Program_weight_tensor_parameter_285: name = "parameter_285" + original_name = "batch_norm2d_65.b_0" shape = [1024] dtype = "float32" min_val = float("-0.169765") @@ -2956,6 +3268,7 @@ class Program_weight_tensor_parameter_285: class Program_weight_tensor_parameter_286: name = "parameter_286" + original_name = "batch_norm2d_65.w_0" shape = [1024] dtype = "float32" min_val = float("-0.112945") @@ -2967,6 +3280,7 @@ class Program_weight_tensor_parameter_286: class Program_weight_tensor_parameter_287: name = "parameter_287" + original_name = "batch_norm2d_65.w_2" shape = [1024] dtype = "float32" min_val = float("5.60519e-45") @@ -2978,6 +3292,7 @@ class Program_weight_tensor_parameter_287: class Program_weight_tensor_parameter_288: name = "parameter_288" + original_name = "batch_norm2d_65.w_1" shape = [1024] dtype = "float32" min_val = float("-0.0260425") @@ -2989,6 +3304,7 @@ class Program_weight_tensor_parameter_288: class Program_weight_tensor_parameter_289: name = "parameter_289" + original_name = "conv2d_65.w_0" shape = [1024, 256, 1, 1] dtype = "float32" min_val = float("-0.151513") @@ -3000,6 +3316,7 @@ class Program_weight_tensor_parameter_289: class Program_weight_tensor_parameter_290: name = "parameter_290" + original_name = "batch_norm2d_64.b_0" shape = [256] dtype = "float32" min_val = float("-0.119003") @@ -3011,6 +3328,7 @@ class Program_weight_tensor_parameter_290: class Program_weight_tensor_parameter_291: name = "parameter_291" + original_name = "batch_norm2d_64.w_0" shape = [256] dtype = "float32" min_val = float("0.0584529") @@ -3022,6 +3340,7 @@ class Program_weight_tensor_parameter_291: class Program_weight_tensor_parameter_292: name = "parameter_292" + original_name = "batch_norm2d_64.w_2" shape = [256] dtype = "float32" min_val = float("0.000528709") @@ -3033,6 +3352,7 @@ class Program_weight_tensor_parameter_292: class Program_weight_tensor_parameter_293: name = "parameter_293" + original_name = "batch_norm2d_64.w_1" shape = [256] dtype = "float32" min_val = float("-0.0602747") @@ -3044,6 +3364,7 @@ class Program_weight_tensor_parameter_293: class Program_weight_tensor_parameter_294: name = "parameter_294" + original_name = "conv2d_64.w_0" shape = [256, 256, 3, 3] dtype = "float32" min_val = float("-0.0802216") @@ -3055,6 +3376,7 @@ class Program_weight_tensor_parameter_294: class Program_weight_tensor_parameter_295: name = "parameter_295" + original_name = "batch_norm2d_63.b_0" shape = [256] dtype = "float32" min_val = float("-0.13678") @@ -3066,6 +3388,7 @@ class Program_weight_tensor_parameter_295: class Program_weight_tensor_parameter_296: name = "parameter_296" + original_name = "batch_norm2d_63.w_0" shape = [256] dtype = "float32" min_val = float("0.0602401") @@ -3077,6 +3400,7 @@ class Program_weight_tensor_parameter_296: class Program_weight_tensor_parameter_297: name = "parameter_297" + original_name = "batch_norm2d_63.w_2" shape = [256] dtype = "float32" min_val = float("0.00184796") @@ -3088,6 +3412,7 @@ class Program_weight_tensor_parameter_297: class Program_weight_tensor_parameter_298: name = "parameter_298" + original_name = "batch_norm2d_63.w_1" shape = [256] dtype = "float32" min_val = float("-0.112543") @@ -3099,6 +3424,7 @@ class Program_weight_tensor_parameter_298: class Program_weight_tensor_parameter_299: name = "parameter_299" + original_name = "conv2d_63.w_0" shape = [256, 1024, 1, 1] dtype = "float32" min_val = float("-0.159615") @@ -3110,6 +3436,7 @@ class Program_weight_tensor_parameter_299: class Program_weight_tensor_parameter_300: name = "parameter_300" + original_name = "batch_norm2d_62.b_0" shape = [1024] dtype = "float32" min_val = float("-0.119579") @@ -3121,6 +3448,7 @@ class Program_weight_tensor_parameter_300: class Program_weight_tensor_parameter_301: name = "parameter_301" + original_name = "batch_norm2d_62.w_0" shape = [1024] dtype = "float32" min_val = float("-0.111251") @@ -3132,6 +3460,7 @@ class Program_weight_tensor_parameter_301: class Program_weight_tensor_parameter_302: name = "parameter_302" + original_name = "batch_norm2d_62.w_2" shape = [1024] dtype = "float32" min_val = float("5.60519e-45") @@ -3143,6 +3472,7 @@ class Program_weight_tensor_parameter_302: class Program_weight_tensor_parameter_303: name = "parameter_303" + original_name = "batch_norm2d_62.w_1" shape = [1024] dtype = "float32" min_val = float("-0.0229368") @@ -3154,6 +3484,7 @@ class Program_weight_tensor_parameter_303: class Program_weight_tensor_parameter_304: name = "parameter_304" + original_name = "conv2d_62.w_0" shape = [1024, 256, 1, 1] dtype = "float32" min_val = float("-0.135953") @@ -3165,6 +3496,7 @@ class Program_weight_tensor_parameter_304: class Program_weight_tensor_parameter_305: name = "parameter_305" + original_name = "batch_norm2d_61.b_0" shape = [256] dtype = "float32" min_val = float("-0.142617") @@ -3176,6 +3508,7 @@ class Program_weight_tensor_parameter_305: class Program_weight_tensor_parameter_306: name = "parameter_306" + original_name = "batch_norm2d_61.w_0" shape = [256] dtype = "float32" min_val = float("0.0604692") @@ -3187,6 +3520,7 @@ class Program_weight_tensor_parameter_306: class Program_weight_tensor_parameter_307: name = "parameter_307" + original_name = "batch_norm2d_61.w_2" shape = [256] dtype = "float32" min_val = float("0.000498543") @@ -3198,6 +3532,7 @@ class Program_weight_tensor_parameter_307: class Program_weight_tensor_parameter_308: name = "parameter_308" + original_name = "batch_norm2d_61.w_1" shape = [256] dtype = "float32" min_val = float("-0.158943") @@ -3209,6 +3544,7 @@ class Program_weight_tensor_parameter_308: class Program_weight_tensor_parameter_309: name = "parameter_309" + original_name = "conv2d_61.w_0" shape = [256, 256, 3, 3] dtype = "float32" min_val = float("-0.101181") @@ -3220,6 +3556,7 @@ class Program_weight_tensor_parameter_309: class Program_weight_tensor_parameter_310: name = "parameter_310" + original_name = "batch_norm2d_60.b_0" shape = [256] dtype = "float32" min_val = float("-0.147466") @@ -3231,6 +3568,7 @@ class Program_weight_tensor_parameter_310: class Program_weight_tensor_parameter_311: name = "parameter_311" + original_name = "batch_norm2d_60.w_0" shape = [256] dtype = "float32" min_val = float("0.0493016") @@ -3242,6 +3580,7 @@ class Program_weight_tensor_parameter_311: class Program_weight_tensor_parameter_312: name = "parameter_312" + original_name = "batch_norm2d_60.w_2" shape = [256] dtype = "float32" min_val = float("0.0015957") @@ -3253,6 +3592,7 @@ class Program_weight_tensor_parameter_312: class Program_weight_tensor_parameter_313: name = "parameter_313" + original_name = "batch_norm2d_60.w_1" shape = [256] dtype = "float32" min_val = float("-0.176171") @@ -3264,6 +3604,7 @@ class Program_weight_tensor_parameter_313: class Program_weight_tensor_parameter_314: name = "parameter_314" + original_name = "conv2d_60.w_0" shape = [256, 1024, 1, 1] dtype = "float32" min_val = float("-0.269775") @@ -3275,6 +3616,7 @@ class Program_weight_tensor_parameter_314: class Program_weight_tensor_parameter_315: name = "parameter_315" + original_name = "batch_norm2d_59.b_0" shape = [1024] dtype = "float32" min_val = float("-0.157031") @@ -3286,6 +3628,7 @@ class Program_weight_tensor_parameter_315: class Program_weight_tensor_parameter_316: name = "parameter_316" + original_name = "batch_norm2d_59.w_0" shape = [1024] dtype = "float32" min_val = float("-0.0900119") @@ -3297,6 +3640,7 @@ class Program_weight_tensor_parameter_316: class Program_weight_tensor_parameter_317: name = "parameter_317" + original_name = "batch_norm2d_59.w_2" shape = [1024] dtype = "float32" min_val = float("5.60519e-45") @@ -3308,6 +3652,7 @@ class Program_weight_tensor_parameter_317: class Program_weight_tensor_parameter_318: name = "parameter_318" + original_name = "batch_norm2d_59.w_1" shape = [1024] dtype = "float32" min_val = float("-0.0139065") @@ -3319,6 +3664,7 @@ class Program_weight_tensor_parameter_318: class Program_weight_tensor_parameter_319: name = "parameter_319" + original_name = "conv2d_59.w_0" shape = [1024, 256, 1, 1] dtype = "float32" min_val = float("-0.201005") @@ -3330,6 +3676,7 @@ class Program_weight_tensor_parameter_319: class Program_weight_tensor_parameter_320: name = "parameter_320" + original_name = "batch_norm2d_58.b_0" shape = [256] dtype = "float32" min_val = float("-0.122277") @@ -3341,6 +3688,7 @@ class Program_weight_tensor_parameter_320: class Program_weight_tensor_parameter_321: name = "parameter_321" + original_name = "batch_norm2d_58.w_0" shape = [256] dtype = "float32" min_val = float("0.0539781") @@ -3352,6 +3700,7 @@ class Program_weight_tensor_parameter_321: class Program_weight_tensor_parameter_322: name = "parameter_322" + original_name = "batch_norm2d_58.w_2" shape = [256] dtype = "float32" min_val = float("0.000414495") @@ -3363,6 +3712,7 @@ class Program_weight_tensor_parameter_322: class Program_weight_tensor_parameter_323: name = "parameter_323" + original_name = "batch_norm2d_58.w_1" shape = [256] dtype = "float32" min_val = float("-0.0649243") @@ -3374,6 +3724,7 @@ class Program_weight_tensor_parameter_323: class Program_weight_tensor_parameter_324: name = "parameter_324" + original_name = "conv2d_58.w_0" shape = [256, 256, 3, 3] dtype = "float32" min_val = float("-0.126597") @@ -3385,6 +3736,7 @@ class Program_weight_tensor_parameter_324: class Program_weight_tensor_parameter_325: name = "parameter_325" + original_name = "batch_norm2d_57.b_0" shape = [256] dtype = "float32" min_val = float("-0.121751") @@ -3396,6 +3748,7 @@ class Program_weight_tensor_parameter_325: class Program_weight_tensor_parameter_326: name = "parameter_326" + original_name = "batch_norm2d_57.w_0" shape = [256] dtype = "float32" min_val = float("0.0471067") @@ -3407,6 +3760,7 @@ class Program_weight_tensor_parameter_326: class Program_weight_tensor_parameter_327: name = "parameter_327" + original_name = "batch_norm2d_57.w_2" shape = [256] dtype = "float32" min_val = float("0.00170937") @@ -3418,6 +3772,7 @@ class Program_weight_tensor_parameter_327: class Program_weight_tensor_parameter_328: name = "parameter_328" + original_name = "batch_norm2d_57.w_1" shape = [256] dtype = "float32" min_val = float("-0.126432") @@ -3429,6 +3784,7 @@ class Program_weight_tensor_parameter_328: class Program_weight_tensor_parameter_329: name = "parameter_329" + original_name = "conv2d_57.w_0" shape = [256, 1024, 1, 1] dtype = "float32" min_val = float("-0.147816") @@ -3440,6 +3796,7 @@ class Program_weight_tensor_parameter_329: class Program_weight_tensor_parameter_330: name = "parameter_330" + original_name = "batch_norm2d_56.b_0" shape = [1024] dtype = "float32" min_val = float("-0.105613") @@ -3451,6 +3808,7 @@ class Program_weight_tensor_parameter_330: class Program_weight_tensor_parameter_331: name = "parameter_331" + original_name = "batch_norm2d_56.w_0" shape = [1024] dtype = "float32" min_val = float("-0.118786") @@ -3462,6 +3820,7 @@ class Program_weight_tensor_parameter_331: class Program_weight_tensor_parameter_332: name = "parameter_332" + original_name = "batch_norm2d_56.w_2" shape = [1024] dtype = "float32" min_val = float("5.60519e-45") @@ -3473,6 +3832,7 @@ class Program_weight_tensor_parameter_332: class Program_weight_tensor_parameter_333: name = "parameter_333" + original_name = "batch_norm2d_56.w_1" shape = [1024] dtype = "float32" min_val = float("-0.0232002") @@ -3484,6 +3844,7 @@ class Program_weight_tensor_parameter_333: class Program_weight_tensor_parameter_334: name = "parameter_334" + original_name = "conv2d_56.w_0" shape = [1024, 256, 1, 1] dtype = "float32" min_val = float("-0.25673") @@ -3495,6 +3856,7 @@ class Program_weight_tensor_parameter_334: class Program_weight_tensor_parameter_335: name = "parameter_335" + original_name = "batch_norm2d_55.b_0" shape = [256] dtype = "float32" min_val = float("-0.361962") @@ -3506,6 +3868,7 @@ class Program_weight_tensor_parameter_335: class Program_weight_tensor_parameter_336: name = "parameter_336" + original_name = "batch_norm2d_55.w_0" shape = [256] dtype = "float32" min_val = float("0.0547539") @@ -3517,6 +3880,7 @@ class Program_weight_tensor_parameter_336: class Program_weight_tensor_parameter_337: name = "parameter_337" + original_name = "batch_norm2d_55.w_2" shape = [256] dtype = "float32" min_val = float("0.000475605") @@ -3528,6 +3892,7 @@ class Program_weight_tensor_parameter_337: class Program_weight_tensor_parameter_338: name = "parameter_338" + original_name = "batch_norm2d_55.w_1" shape = [256] dtype = "float32" min_val = float("-0.13866") @@ -3539,6 +3904,7 @@ class Program_weight_tensor_parameter_338: class Program_weight_tensor_parameter_339: name = "parameter_339" + original_name = "conv2d_55.w_0" shape = [256, 256, 3, 3] dtype = "float32" min_val = float("-0.156502") @@ -3550,6 +3916,7 @@ class Program_weight_tensor_parameter_339: class Program_weight_tensor_parameter_340: name = "parameter_340" + original_name = "batch_norm2d_54.b_0" shape = [256] dtype = "float32" min_val = float("-0.333298") @@ -3561,6 +3928,7 @@ class Program_weight_tensor_parameter_340: class Program_weight_tensor_parameter_341: name = "parameter_341" + original_name = "batch_norm2d_54.w_0" shape = [256] dtype = "float32" min_val = float("0.0471103") @@ -3572,6 +3940,7 @@ class Program_weight_tensor_parameter_341: class Program_weight_tensor_parameter_342: name = "parameter_342" + original_name = "batch_norm2d_54.w_2" shape = [256] dtype = "float32" min_val = float("0.00119363") @@ -3583,6 +3952,7 @@ class Program_weight_tensor_parameter_342: class Program_weight_tensor_parameter_343: name = "parameter_343" + original_name = "batch_norm2d_54.w_1" shape = [256] dtype = "float32" min_val = float("-0.131256") @@ -3594,6 +3964,7 @@ class Program_weight_tensor_parameter_343: class Program_weight_tensor_parameter_344: name = "parameter_344" + original_name = "conv2d_54.w_0" shape = [256, 1024, 1, 1] dtype = "float32" min_val = float("-0.169202") @@ -3605,6 +3976,7 @@ class Program_weight_tensor_parameter_344: class Program_weight_tensor_parameter_345: name = "parameter_345" + original_name = "batch_norm2d_53.b_0" shape = [1024] dtype = "float32" min_val = float("-0.14296") @@ -3616,6 +3988,7 @@ class Program_weight_tensor_parameter_345: class Program_weight_tensor_parameter_346: name = "parameter_346" + original_name = "batch_norm2d_53.w_0" shape = [1024] dtype = "float32" min_val = float("-0.0977966") @@ -3627,6 +4000,7 @@ class Program_weight_tensor_parameter_346: class Program_weight_tensor_parameter_347: name = "parameter_347" + original_name = "batch_norm2d_53.w_2" shape = [1024] dtype = "float32" min_val = float("5.60519e-45") @@ -3638,6 +4012,7 @@ class Program_weight_tensor_parameter_347: class Program_weight_tensor_parameter_348: name = "parameter_348" + original_name = "batch_norm2d_53.w_1" shape = [1024] dtype = "float32" min_val = float("-0.0163465") @@ -3649,6 +4024,7 @@ class Program_weight_tensor_parameter_348: class Program_weight_tensor_parameter_349: name = "parameter_349" + original_name = "conv2d_53.w_0" shape = [1024, 256, 1, 1] dtype = "float32" min_val = float("-0.190243") @@ -3660,6 +4036,7 @@ class Program_weight_tensor_parameter_349: class Program_weight_tensor_parameter_350: name = "parameter_350" + original_name = "batch_norm2d_52.b_0" shape = [256] dtype = "float32" min_val = float("-0.150391") @@ -3671,6 +4048,7 @@ class Program_weight_tensor_parameter_350: class Program_weight_tensor_parameter_351: name = "parameter_351" + original_name = "batch_norm2d_52.w_0" shape = [256] dtype = "float32" min_val = float("0.0479986") @@ -3682,6 +4060,7 @@ class Program_weight_tensor_parameter_351: class Program_weight_tensor_parameter_352: name = "parameter_352" + original_name = "batch_norm2d_52.w_2" shape = [256] dtype = "float32" min_val = float("0.000481484") @@ -3693,6 +4072,7 @@ class Program_weight_tensor_parameter_352: class Program_weight_tensor_parameter_353: name = "parameter_353" + original_name = "batch_norm2d_52.w_1" shape = [256] dtype = "float32" min_val = float("-0.0997624") @@ -3704,6 +4084,7 @@ class Program_weight_tensor_parameter_353: class Program_weight_tensor_parameter_354: name = "parameter_354" + original_name = "conv2d_52.w_0" shape = [256, 256, 3, 3] dtype = "float32" min_val = float("-0.126122") @@ -3715,6 +4096,7 @@ class Program_weight_tensor_parameter_354: class Program_weight_tensor_parameter_355: name = "parameter_355" + original_name = "batch_norm2d_51.b_0" shape = [256] dtype = "float32" min_val = float("-0.1715") @@ -3726,6 +4108,7 @@ class Program_weight_tensor_parameter_355: class Program_weight_tensor_parameter_356: name = "parameter_356" + original_name = "batch_norm2d_51.w_0" shape = [256] dtype = "float32" min_val = float("0.0470394") @@ -3737,6 +4120,7 @@ class Program_weight_tensor_parameter_356: class Program_weight_tensor_parameter_357: name = "parameter_357" + original_name = "batch_norm2d_51.w_2" shape = [256] dtype = "float32" min_val = float("0.00156337") @@ -3748,6 +4132,7 @@ class Program_weight_tensor_parameter_357: class Program_weight_tensor_parameter_358: name = "parameter_358" + original_name = "batch_norm2d_51.w_1" shape = [256] dtype = "float32" min_val = float("-0.0805164") @@ -3759,6 +4144,7 @@ class Program_weight_tensor_parameter_358: class Program_weight_tensor_parameter_359: name = "parameter_359" + original_name = "conv2d_51.w_0" shape = [256, 1024, 1, 1] dtype = "float32" min_val = float("-0.135523") @@ -3770,6 +4156,7 @@ class Program_weight_tensor_parameter_359: class Program_weight_tensor_parameter_360: name = "parameter_360" + original_name = "batch_norm2d_50.b_0" shape = [1024] dtype = "float32" min_val = float("-0.110059") @@ -3781,6 +4168,7 @@ class Program_weight_tensor_parameter_360: class Program_weight_tensor_parameter_361: name = "parameter_361" + original_name = "batch_norm2d_50.w_0" shape = [1024] dtype = "float32" min_val = float("-0.0931681") @@ -3792,6 +4180,7 @@ class Program_weight_tensor_parameter_361: class Program_weight_tensor_parameter_362: name = "parameter_362" + original_name = "batch_norm2d_50.w_2" shape = [1024] dtype = "float32" min_val = float("5.60519e-45") @@ -3803,6 +4192,7 @@ class Program_weight_tensor_parameter_362: class Program_weight_tensor_parameter_363: name = "parameter_363" + original_name = "batch_norm2d_50.w_1" shape = [1024] dtype = "float32" min_val = float("-0.018444") @@ -3814,6 +4204,7 @@ class Program_weight_tensor_parameter_363: class Program_weight_tensor_parameter_364: name = "parameter_364" + original_name = "conv2d_50.w_0" shape = [1024, 256, 1, 1] dtype = "float32" min_val = float("-0.139429") @@ -3825,6 +4216,7 @@ class Program_weight_tensor_parameter_364: class Program_weight_tensor_parameter_365: name = "parameter_365" + original_name = "batch_norm2d_49.b_0" shape = [256] dtype = "float32" min_val = float("-0.157348") @@ -3836,6 +4228,7 @@ class Program_weight_tensor_parameter_365: class Program_weight_tensor_parameter_366: name = "parameter_366" + original_name = "batch_norm2d_49.w_0" shape = [256] dtype = "float32" min_val = float("0.0530454") @@ -3847,6 +4240,7 @@ class Program_weight_tensor_parameter_366: class Program_weight_tensor_parameter_367: name = "parameter_367" + original_name = "batch_norm2d_49.w_2" shape = [256] dtype = "float32" min_val = float("0.000397306") @@ -3858,6 +4252,7 @@ class Program_weight_tensor_parameter_367: class Program_weight_tensor_parameter_368: name = "parameter_368" + original_name = "batch_norm2d_49.w_1" shape = [256] dtype = "float32" min_val = float("-0.107012") @@ -3869,6 +4264,7 @@ class Program_weight_tensor_parameter_368: class Program_weight_tensor_parameter_369: name = "parameter_369" + original_name = "conv2d_49.w_0" shape = [256, 256, 3, 3] dtype = "float32" min_val = float("-0.117755") @@ -3880,6 +4276,7 @@ class Program_weight_tensor_parameter_369: class Program_weight_tensor_parameter_370: name = "parameter_370" + original_name = "batch_norm2d_48.b_0" shape = [256] dtype = "float32" min_val = float("-0.181549") @@ -3891,6 +4288,7 @@ class Program_weight_tensor_parameter_370: class Program_weight_tensor_parameter_371: name = "parameter_371" + original_name = "batch_norm2d_48.w_0" shape = [256] dtype = "float32" min_val = float("0.0575181") @@ -3902,6 +4300,7 @@ class Program_weight_tensor_parameter_371: class Program_weight_tensor_parameter_372: name = "parameter_372" + original_name = "batch_norm2d_48.w_2" shape = [256] dtype = "float32" min_val = float("0.00120071") @@ -3913,6 +4312,7 @@ class Program_weight_tensor_parameter_372: class Program_weight_tensor_parameter_373: name = "parameter_373" + original_name = "batch_norm2d_48.w_1" shape = [256] dtype = "float32" min_val = float("-0.127524") @@ -3924,6 +4324,7 @@ class Program_weight_tensor_parameter_373: class Program_weight_tensor_parameter_374: name = "parameter_374" + original_name = "conv2d_48.w_0" shape = [256, 1024, 1, 1] dtype = "float32" min_val = float("-0.198103") @@ -3935,6 +4336,7 @@ class Program_weight_tensor_parameter_374: class Program_weight_tensor_parameter_375: name = "parameter_375" + original_name = "batch_norm2d_47.b_0" shape = [1024] dtype = "float32" min_val = float("-0.183645") @@ -3946,6 +4348,7 @@ class Program_weight_tensor_parameter_375: class Program_weight_tensor_parameter_376: name = "parameter_376" + original_name = "batch_norm2d_47.w_0" shape = [1024] dtype = "float32" min_val = float("-0.0497403") @@ -3957,6 +4360,7 @@ class Program_weight_tensor_parameter_376: class Program_weight_tensor_parameter_377: name = "parameter_377" + original_name = "batch_norm2d_47.w_2" shape = [1024] dtype = "float32" min_val = float("5.60519e-45") @@ -3968,6 +4372,7 @@ class Program_weight_tensor_parameter_377: class Program_weight_tensor_parameter_378: name = "parameter_378" + original_name = "batch_norm2d_47.w_1" shape = [1024] dtype = "float32" min_val = float("-0.028365") @@ -3979,6 +4384,7 @@ class Program_weight_tensor_parameter_378: class Program_weight_tensor_parameter_379: name = "parameter_379" + original_name = "conv2d_47.w_0" shape = [1024, 256, 1, 1] dtype = "float32" min_val = float("-0.220192") @@ -3990,6 +4396,7 @@ class Program_weight_tensor_parameter_379: class Program_weight_tensor_parameter_380: name = "parameter_380" + original_name = "batch_norm2d_46.b_0" shape = [256] dtype = "float32" min_val = float("-0.261715") @@ -4001,6 +4408,7 @@ class Program_weight_tensor_parameter_380: class Program_weight_tensor_parameter_381: name = "parameter_381" + original_name = "batch_norm2d_46.w_0" shape = [256] dtype = "float32" min_val = float("0.0650689") @@ -4012,6 +4420,7 @@ class Program_weight_tensor_parameter_381: class Program_weight_tensor_parameter_382: name = "parameter_382" + original_name = "batch_norm2d_46.w_2" shape = [256] dtype = "float32" min_val = float("0.000682362") @@ -4023,6 +4432,7 @@ class Program_weight_tensor_parameter_382: class Program_weight_tensor_parameter_383: name = "parameter_383" + original_name = "batch_norm2d_46.w_1" shape = [256] dtype = "float32" min_val = float("-0.0922706") @@ -4034,6 +4444,7 @@ class Program_weight_tensor_parameter_383: class Program_weight_tensor_parameter_384: name = "parameter_384" + original_name = "conv2d_46.w_0" shape = [256, 256, 3, 3] dtype = "float32" min_val = float("-0.165137") @@ -4045,6 +4456,7 @@ class Program_weight_tensor_parameter_384: class Program_weight_tensor_parameter_385: name = "parameter_385" + original_name = "batch_norm2d_45.b_0" shape = [256] dtype = "float32" min_val = float("-0.175628") @@ -4056,6 +4468,7 @@ class Program_weight_tensor_parameter_385: class Program_weight_tensor_parameter_386: name = "parameter_386" + original_name = "batch_norm2d_45.w_0" shape = [256] dtype = "float32" min_val = float("0.066233") @@ -4067,6 +4480,7 @@ class Program_weight_tensor_parameter_386: class Program_weight_tensor_parameter_387: name = "parameter_387" + original_name = "batch_norm2d_45.w_2" shape = [256] dtype = "float32" min_val = float("0.00199584") @@ -4078,6 +4492,7 @@ class Program_weight_tensor_parameter_387: class Program_weight_tensor_parameter_388: name = "parameter_388" + original_name = "batch_norm2d_45.w_1" shape = [256] dtype = "float32" min_val = float("-0.113064") @@ -4089,6 +4504,7 @@ class Program_weight_tensor_parameter_388: class Program_weight_tensor_parameter_389: name = "parameter_389" + original_name = "conv2d_45.w_0" shape = [256, 1024, 1, 1] dtype = "float32" min_val = float("-0.196158") @@ -4100,6 +4516,7 @@ class Program_weight_tensor_parameter_389: class Program_weight_tensor_parameter_390: name = "parameter_390" + original_name = "batch_norm2d_44.b_0" shape = [1024] dtype = "float32" min_val = float("-0.150438") @@ -4111,6 +4528,7 @@ class Program_weight_tensor_parameter_390: class Program_weight_tensor_parameter_391: name = "parameter_391" + original_name = "batch_norm2d_44.w_0" shape = [1024] dtype = "float32" min_val = float("-0.0526699") @@ -4122,6 +4540,7 @@ class Program_weight_tensor_parameter_391: class Program_weight_tensor_parameter_392: name = "parameter_392" + original_name = "batch_norm2d_44.w_2" shape = [1024] dtype = "float32" min_val = float("5.60519e-45") @@ -4133,6 +4552,7 @@ class Program_weight_tensor_parameter_392: class Program_weight_tensor_parameter_393: name = "parameter_393" + original_name = "batch_norm2d_44.w_1" shape = [1024] dtype = "float32" min_val = float("-0.0221359") @@ -4144,6 +4564,7 @@ class Program_weight_tensor_parameter_393: class Program_weight_tensor_parameter_394: name = "parameter_394" + original_name = "conv2d_44.w_0" shape = [1024, 256, 1, 1] dtype = "float32" min_val = float("-0.149936") @@ -4155,6 +4576,7 @@ class Program_weight_tensor_parameter_394: class Program_weight_tensor_parameter_395: name = "parameter_395" + original_name = "batch_norm2d_43.b_0" shape = [256] dtype = "float32" min_val = float("-0.153784") @@ -4166,6 +4588,7 @@ class Program_weight_tensor_parameter_395: class Program_weight_tensor_parameter_396: name = "parameter_396" + original_name = "batch_norm2d_43.w_0" shape = [256] dtype = "float32" min_val = float("0.0553597") @@ -4177,6 +4600,7 @@ class Program_weight_tensor_parameter_396: class Program_weight_tensor_parameter_397: name = "parameter_397" + original_name = "batch_norm2d_43.w_2" shape = [256] dtype = "float32" min_val = float("0.000499374") @@ -4188,6 +4612,7 @@ class Program_weight_tensor_parameter_397: class Program_weight_tensor_parameter_398: name = "parameter_398" + original_name = "batch_norm2d_43.w_1" shape = [256] dtype = "float32" min_val = float("-0.0774878") @@ -4199,6 +4624,7 @@ class Program_weight_tensor_parameter_398: class Program_weight_tensor_parameter_399: name = "parameter_399" + original_name = "conv2d_43.w_0" shape = [256, 256, 3, 3] dtype = "float32" min_val = float("-0.115989") @@ -4210,6 +4636,7 @@ class Program_weight_tensor_parameter_399: class Program_weight_tensor_parameter_400: name = "parameter_400" + original_name = "batch_norm2d_42.b_0" shape = [256] dtype = "float32" min_val = float("-0.143976") @@ -4221,6 +4648,7 @@ class Program_weight_tensor_parameter_400: class Program_weight_tensor_parameter_401: name = "parameter_401" + original_name = "batch_norm2d_42.w_0" shape = [256] dtype = "float32" min_val = float("0.0597508") @@ -4232,6 +4660,7 @@ class Program_weight_tensor_parameter_401: class Program_weight_tensor_parameter_402: name = "parameter_402" + original_name = "batch_norm2d_42.w_2" shape = [256] dtype = "float32" min_val = float("0.00153534") @@ -4243,6 +4672,7 @@ class Program_weight_tensor_parameter_402: class Program_weight_tensor_parameter_403: name = "parameter_403" + original_name = "batch_norm2d_42.w_1" shape = [256] dtype = "float32" min_val = float("-0.165527") @@ -4254,6 +4684,7 @@ class Program_weight_tensor_parameter_403: class Program_weight_tensor_parameter_404: name = "parameter_404" + original_name = "conv2d_42.w_0" shape = [256, 1024, 1, 1] dtype = "float32" min_val = float("-0.125145") @@ -4265,6 +4696,7 @@ class Program_weight_tensor_parameter_404: class Program_weight_tensor_parameter_405: name = "parameter_405" + original_name = "batch_norm2d_41.b_0" shape = [1024] dtype = "float32" min_val = float("-0.206426") @@ -4276,6 +4708,7 @@ class Program_weight_tensor_parameter_405: class Program_weight_tensor_parameter_406: name = "parameter_406" + original_name = "batch_norm2d_41.w_0" shape = [1024] dtype = "float32" min_val = float("-0.0464987") @@ -4287,6 +4720,7 @@ class Program_weight_tensor_parameter_406: class Program_weight_tensor_parameter_407: name = "parameter_407" + original_name = "batch_norm2d_41.w_2" shape = [1024] dtype = "float32" min_val = float("5.60519e-45") @@ -4298,6 +4732,7 @@ class Program_weight_tensor_parameter_407: class Program_weight_tensor_parameter_408: name = "parameter_408" + original_name = "batch_norm2d_41.w_1" shape = [1024] dtype = "float32" min_val = float("-0.0218367") @@ -4309,6 +4744,7 @@ class Program_weight_tensor_parameter_408: class Program_weight_tensor_parameter_409: name = "parameter_409" + original_name = "conv2d_41.w_0" shape = [1024, 256, 1, 1] dtype = "float32" min_val = float("-0.124601") @@ -4320,6 +4756,7 @@ class Program_weight_tensor_parameter_409: class Program_weight_tensor_parameter_410: name = "parameter_410" + original_name = "batch_norm2d_40.b_0" shape = [256] dtype = "float32" min_val = float("-0.168556") @@ -4331,6 +4768,7 @@ class Program_weight_tensor_parameter_410: class Program_weight_tensor_parameter_411: name = "parameter_411" + original_name = "batch_norm2d_40.w_0" shape = [256] dtype = "float32" min_val = float("0.0540817") @@ -4342,6 +4780,7 @@ class Program_weight_tensor_parameter_411: class Program_weight_tensor_parameter_412: name = "parameter_412" + original_name = "batch_norm2d_40.w_2" shape = [256] dtype = "float32" min_val = float("0.000486622") @@ -4353,6 +4792,7 @@ class Program_weight_tensor_parameter_412: class Program_weight_tensor_parameter_413: name = "parameter_413" + original_name = "batch_norm2d_40.w_1" shape = [256] dtype = "float32" min_val = float("-0.0689682") @@ -4364,6 +4804,7 @@ class Program_weight_tensor_parameter_413: class Program_weight_tensor_parameter_414: name = "parameter_414" + original_name = "conv2d_40.w_0" shape = [256, 256, 3, 3] dtype = "float32" min_val = float("-0.149872") @@ -4375,6 +4816,7 @@ class Program_weight_tensor_parameter_414: class Program_weight_tensor_parameter_415: name = "parameter_415" + original_name = "batch_norm2d_39.b_0" shape = [256] dtype = "float32" min_val = float("-0.125849") @@ -4386,6 +4828,7 @@ class Program_weight_tensor_parameter_415: class Program_weight_tensor_parameter_416: name = "parameter_416" + original_name = "batch_norm2d_39.w_0" shape = [256] dtype = "float32" min_val = float("0.0581656") @@ -4397,6 +4840,7 @@ class Program_weight_tensor_parameter_416: class Program_weight_tensor_parameter_417: name = "parameter_417" + original_name = "batch_norm2d_39.w_2" shape = [256] dtype = "float32" min_val = float("0.00159382") @@ -4408,6 +4852,7 @@ class Program_weight_tensor_parameter_417: class Program_weight_tensor_parameter_418: name = "parameter_418" + original_name = "batch_norm2d_39.w_1" shape = [256] dtype = "float32" min_val = float("-0.16078") @@ -4419,6 +4864,7 @@ class Program_weight_tensor_parameter_418: class Program_weight_tensor_parameter_419: name = "parameter_419" + original_name = "conv2d_39.w_0" shape = [256, 1024, 1, 1] dtype = "float32" min_val = float("-0.206113") @@ -4430,6 +4876,7 @@ class Program_weight_tensor_parameter_419: class Program_weight_tensor_parameter_420: name = "parameter_420" + original_name = "batch_norm2d_38.b_0" shape = [1024] dtype = "float32" min_val = float("-0.137891") @@ -4441,6 +4888,7 @@ class Program_weight_tensor_parameter_420: class Program_weight_tensor_parameter_421: name = "parameter_421" + original_name = "batch_norm2d_38.w_0" shape = [1024] dtype = "float32" min_val = float("-0.129481") @@ -4452,6 +4900,7 @@ class Program_weight_tensor_parameter_421: class Program_weight_tensor_parameter_422: name = "parameter_422" + original_name = "batch_norm2d_38.w_2" shape = [1024] dtype = "float32" min_val = float("5.60519e-45") @@ -4463,6 +4912,7 @@ class Program_weight_tensor_parameter_422: class Program_weight_tensor_parameter_423: name = "parameter_423" + original_name = "batch_norm2d_38.w_1" shape = [1024] dtype = "float32" min_val = float("-0.0303839") @@ -4474,6 +4924,7 @@ class Program_weight_tensor_parameter_423: class Program_weight_tensor_parameter_424: name = "parameter_424" + original_name = "conv2d_38.w_0" shape = [1024, 256, 1, 1] dtype = "float32" min_val = float("-0.133517") @@ -4485,6 +4936,7 @@ class Program_weight_tensor_parameter_424: class Program_weight_tensor_parameter_425: name = "parameter_425" + original_name = "batch_norm2d_37.b_0" shape = [256] dtype = "float32" min_val = float("-0.395405") @@ -4496,6 +4948,7 @@ class Program_weight_tensor_parameter_425: class Program_weight_tensor_parameter_426: name = "parameter_426" + original_name = "batch_norm2d_37.w_0" shape = [256] dtype = "float32" min_val = float("0.0522077") @@ -4507,6 +4960,7 @@ class Program_weight_tensor_parameter_426: class Program_weight_tensor_parameter_427: name = "parameter_427" + original_name = "batch_norm2d_37.w_2" shape = [256] dtype = "float32" min_val = float("0.00061321") @@ -4518,6 +4972,7 @@ class Program_weight_tensor_parameter_427: class Program_weight_tensor_parameter_428: name = "parameter_428" + original_name = "batch_norm2d_37.w_1" shape = [256] dtype = "float32" min_val = float("-0.177767") @@ -4529,6 +4984,7 @@ class Program_weight_tensor_parameter_428: class Program_weight_tensor_parameter_429: name = "parameter_429" + original_name = "conv2d_37.w_0" shape = [256, 256, 3, 3] dtype = "float32" min_val = float("-0.0892948") @@ -4540,6 +4996,7 @@ class Program_weight_tensor_parameter_429: class Program_weight_tensor_parameter_430: name = "parameter_430" + original_name = "batch_norm2d_36.b_0" shape = [256] dtype = "float32" min_val = float("-0.128915") @@ -4551,6 +5008,7 @@ class Program_weight_tensor_parameter_430: class Program_weight_tensor_parameter_431: name = "parameter_431" + original_name = "batch_norm2d_36.w_0" shape = [256] dtype = "float32" min_val = float("0.0464295") @@ -4562,6 +5020,7 @@ class Program_weight_tensor_parameter_431: class Program_weight_tensor_parameter_432: name = "parameter_432" + original_name = "batch_norm2d_36.w_2" shape = [256] dtype = "float32" min_val = float("0.00125646") @@ -4573,6 +5032,7 @@ class Program_weight_tensor_parameter_432: class Program_weight_tensor_parameter_433: name = "parameter_433" + original_name = "batch_norm2d_36.w_1" shape = [256] dtype = "float32" min_val = float("-0.170545") @@ -4584,6 +5044,7 @@ class Program_weight_tensor_parameter_433: class Program_weight_tensor_parameter_434: name = "parameter_434" + original_name = "conv2d_36.w_0" shape = [256, 1024, 1, 1] dtype = "float32" min_val = float("-0.187673") @@ -4595,6 +5056,7 @@ class Program_weight_tensor_parameter_434: class Program_weight_tensor_parameter_435: name = "parameter_435" + original_name = "batch_norm2d_35.b_0" shape = [1024] dtype = "float32" min_val = float("-0.17385") @@ -4606,6 +5068,7 @@ class Program_weight_tensor_parameter_435: class Program_weight_tensor_parameter_436: name = "parameter_436" + original_name = "batch_norm2d_35.w_0" shape = [1024] dtype = "float32" min_val = float("-0.0858251") @@ -4617,6 +5080,7 @@ class Program_weight_tensor_parameter_436: class Program_weight_tensor_parameter_437: name = "parameter_437" + original_name = "batch_norm2d_35.w_2" shape = [1024] dtype = "float32" min_val = float("5.60519e-45") @@ -4628,6 +5092,7 @@ class Program_weight_tensor_parameter_437: class Program_weight_tensor_parameter_438: name = "parameter_438" + original_name = "batch_norm2d_35.w_1" shape = [1024] dtype = "float32" min_val = float("-0.0571547") @@ -4639,6 +5104,7 @@ class Program_weight_tensor_parameter_438: class Program_weight_tensor_parameter_439: name = "parameter_439" + original_name = "conv2d_35.w_0" shape = [1024, 256, 1, 1] dtype = "float32" min_val = float("-0.336971") @@ -4650,6 +5116,7 @@ class Program_weight_tensor_parameter_439: class Program_weight_tensor_parameter_440: name = "parameter_440" + original_name = "batch_norm2d_34.b_0" shape = [256] dtype = "float32" min_val = float("-0.188655") @@ -4661,6 +5128,7 @@ class Program_weight_tensor_parameter_440: class Program_weight_tensor_parameter_441: name = "parameter_441" + original_name = "batch_norm2d_34.w_0" shape = [256] dtype = "float32" min_val = float("0.0683378") @@ -4672,6 +5140,7 @@ class Program_weight_tensor_parameter_441: class Program_weight_tensor_parameter_442: name = "parameter_442" + original_name = "batch_norm2d_34.w_2" shape = [256] dtype = "float32" min_val = float("0.000900066") @@ -4683,6 +5152,7 @@ class Program_weight_tensor_parameter_442: class Program_weight_tensor_parameter_443: name = "parameter_443" + original_name = "batch_norm2d_34.w_1" shape = [256] dtype = "float32" min_val = float("-0.145351") @@ -4694,6 +5164,7 @@ class Program_weight_tensor_parameter_443: class Program_weight_tensor_parameter_444: name = "parameter_444" + original_name = "conv2d_34.w_0" shape = [256, 256, 3, 3] dtype = "float32" min_val = float("-0.132562") @@ -4705,6 +5176,7 @@ class Program_weight_tensor_parameter_444: class Program_weight_tensor_parameter_445: name = "parameter_445" + original_name = "batch_norm2d_33.b_0" shape = [256] dtype = "float32" min_val = float("-0.144211") @@ -4716,6 +5188,7 @@ class Program_weight_tensor_parameter_445: class Program_weight_tensor_parameter_446: name = "parameter_446" + original_name = "batch_norm2d_33.w_0" shape = [256] dtype = "float32" min_val = float("0.0696308") @@ -4727,6 +5200,7 @@ class Program_weight_tensor_parameter_446: class Program_weight_tensor_parameter_447: name = "parameter_447" + original_name = "batch_norm2d_33.w_2" shape = [256] dtype = "float32" min_val = float("0.00197131") @@ -4738,6 +5212,7 @@ class Program_weight_tensor_parameter_447: class Program_weight_tensor_parameter_448: name = "parameter_448" + original_name = "batch_norm2d_33.w_1" shape = [256] dtype = "float32" min_val = float("-0.131424") @@ -4749,6 +5224,7 @@ class Program_weight_tensor_parameter_448: class Program_weight_tensor_parameter_449: name = "parameter_449" + original_name = "conv2d_33.w_0" shape = [256, 1024, 1, 1] dtype = "float32" min_val = float("-0.172959") @@ -4760,6 +5236,7 @@ class Program_weight_tensor_parameter_449: class Program_weight_tensor_parameter_450: name = "parameter_450" + original_name = "batch_norm2d_32.b_0" shape = [1024] dtype = "float32" min_val = float("-0.167666") @@ -4771,6 +5248,7 @@ class Program_weight_tensor_parameter_450: class Program_weight_tensor_parameter_451: name = "parameter_451" + original_name = "batch_norm2d_32.w_0" shape = [1024] dtype = "float32" min_val = float("-0.112383") @@ -4782,6 +5260,7 @@ class Program_weight_tensor_parameter_451: class Program_weight_tensor_parameter_452: name = "parameter_452" + original_name = "batch_norm2d_32.w_2" shape = [1024] dtype = "float32" min_val = float("5.60519e-45") @@ -4793,6 +5272,7 @@ class Program_weight_tensor_parameter_452: class Program_weight_tensor_parameter_453: name = "parameter_453" + original_name = "batch_norm2d_32.w_1" shape = [1024] dtype = "float32" min_val = float("-0.0534777") @@ -4804,6 +5284,7 @@ class Program_weight_tensor_parameter_453: class Program_weight_tensor_parameter_454: name = "parameter_454" + original_name = "conv2d_32.w_0" shape = [1024, 256, 1, 1] dtype = "float32" min_val = float("-0.282625") @@ -4815,6 +5296,7 @@ class Program_weight_tensor_parameter_454: class Program_weight_tensor_parameter_455: name = "parameter_455" + original_name = "batch_norm2d_31.b_0" shape = [256] dtype = "float32" min_val = float("-0.391536") @@ -4826,6 +5308,7 @@ class Program_weight_tensor_parameter_455: class Program_weight_tensor_parameter_456: name = "parameter_456" + original_name = "batch_norm2d_31.w_0" shape = [256] dtype = "float32" min_val = float("0.0474037") @@ -4837,6 +5320,7 @@ class Program_weight_tensor_parameter_456: class Program_weight_tensor_parameter_457: name = "parameter_457" + original_name = "batch_norm2d_31.w_2" shape = [256] dtype = "float32" min_val = float("0.000825777") @@ -4848,6 +5332,7 @@ class Program_weight_tensor_parameter_457: class Program_weight_tensor_parameter_458: name = "parameter_458" + original_name = "batch_norm2d_31.w_1" shape = [256] dtype = "float32" min_val = float("-0.284886") @@ -4859,6 +5344,7 @@ class Program_weight_tensor_parameter_458: class Program_weight_tensor_parameter_459: name = "parameter_459" + original_name = "conv2d_31.w_0" shape = [256, 256, 3, 3] dtype = "float32" min_val = float("-0.174341") @@ -4870,6 +5356,7 @@ class Program_weight_tensor_parameter_459: class Program_weight_tensor_parameter_460: name = "parameter_460" + original_name = "batch_norm2d_30.b_0" shape = [256] dtype = "float32" min_val = float("-0.254122") @@ -4881,6 +5368,7 @@ class Program_weight_tensor_parameter_460: class Program_weight_tensor_parameter_461: name = "parameter_461" + original_name = "batch_norm2d_30.w_0" shape = [256] dtype = "float32" min_val = float("0.0522015") @@ -4892,6 +5380,7 @@ class Program_weight_tensor_parameter_461: class Program_weight_tensor_parameter_462: name = "parameter_462" + original_name = "batch_norm2d_30.w_2" shape = [256] dtype = "float32" min_val = float("0.00139353") @@ -4903,6 +5392,7 @@ class Program_weight_tensor_parameter_462: class Program_weight_tensor_parameter_463: name = "parameter_463" + original_name = "batch_norm2d_30.w_1" shape = [256] dtype = "float32" min_val = float("-0.325423") @@ -4914,6 +5404,7 @@ class Program_weight_tensor_parameter_463: class Program_weight_tensor_parameter_464: name = "parameter_464" + original_name = "conv2d_30.w_0" shape = [256, 1024, 1, 1] dtype = "float32" min_val = float("-0.143796") @@ -4925,6 +5416,7 @@ class Program_weight_tensor_parameter_464: class Program_weight_tensor_parameter_465: name = "parameter_465" + original_name = "batch_norm2d_29.b_0" shape = [1024] dtype = "float32" min_val = float("-0.070916") @@ -4936,6 +5428,7 @@ class Program_weight_tensor_parameter_465: class Program_weight_tensor_parameter_466: name = "parameter_466" + original_name = "batch_norm2d_29.w_0" shape = [1024] dtype = "float32" min_val = float("-0.00377393") @@ -4947,6 +5440,7 @@ class Program_weight_tensor_parameter_466: class Program_weight_tensor_parameter_467: name = "parameter_467" + original_name = "batch_norm2d_29.w_2" shape = [1024] dtype = "float32" min_val = float("5.60519e-45") @@ -4958,6 +5452,7 @@ class Program_weight_tensor_parameter_467: class Program_weight_tensor_parameter_468: name = "parameter_468" + original_name = "batch_norm2d_29.w_1" shape = [1024] dtype = "float32" min_val = float("-0.131816") @@ -4969,6 +5464,7 @@ class Program_weight_tensor_parameter_468: class Program_weight_tensor_parameter_469: name = "parameter_469" + original_name = "conv2d_29.w_0" shape = [1024, 512, 1, 1] dtype = "float32" min_val = float("-0.357426") @@ -4980,6 +5476,7 @@ class Program_weight_tensor_parameter_469: class Program_weight_tensor_parameter_470: name = "parameter_470" + original_name = "batch_norm2d_28.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0709161") @@ -4991,6 +5488,7 @@ class Program_weight_tensor_parameter_470: class Program_weight_tensor_parameter_471: name = "parameter_471" + original_name = "batch_norm2d_28.w_0" shape = [1024] dtype = "float32" min_val = float("-0.0815912") @@ -5002,6 +5500,7 @@ class Program_weight_tensor_parameter_471: class Program_weight_tensor_parameter_472: name = "parameter_472" + original_name = "batch_norm2d_28.w_2" shape = [1024] dtype = "float32" min_val = float("5.60519e-45") @@ -5013,6 +5512,7 @@ class Program_weight_tensor_parameter_472: class Program_weight_tensor_parameter_473: name = "parameter_473" + original_name = "batch_norm2d_28.w_1" shape = [1024] dtype = "float32" min_val = float("-0.05334") @@ -5024,6 +5524,7 @@ class Program_weight_tensor_parameter_473: class Program_weight_tensor_parameter_474: name = "parameter_474" + original_name = "conv2d_28.w_0" shape = [1024, 256, 1, 1] dtype = "float32" min_val = float("-0.130086") @@ -5035,6 +5536,7 @@ class Program_weight_tensor_parameter_474: class Program_weight_tensor_parameter_475: name = "parameter_475" + original_name = "batch_norm2d_27.b_0" shape = [256] dtype = "float32" min_val = float("-0.132216") @@ -5046,6 +5548,7 @@ class Program_weight_tensor_parameter_475: class Program_weight_tensor_parameter_476: name = "parameter_476" + original_name = "batch_norm2d_27.w_0" shape = [256] dtype = "float32" min_val = float("0.0621491") @@ -5057,6 +5560,7 @@ class Program_weight_tensor_parameter_476: class Program_weight_tensor_parameter_477: name = "parameter_477" + original_name = "batch_norm2d_27.w_2" shape = [256] dtype = "float32" min_val = float("0.000498966") @@ -5068,6 +5572,7 @@ class Program_weight_tensor_parameter_477: class Program_weight_tensor_parameter_478: name = "parameter_478" + original_name = "batch_norm2d_27.w_1" shape = [256] dtype = "float32" min_val = float("-0.0500929") @@ -5079,6 +5584,7 @@ class Program_weight_tensor_parameter_478: class Program_weight_tensor_parameter_479: name = "parameter_479" + original_name = "conv2d_27.w_0" shape = [256, 256, 3, 3] dtype = "float32" min_val = float("-0.0972044") @@ -5090,6 +5596,7 @@ class Program_weight_tensor_parameter_479: class Program_weight_tensor_parameter_480: name = "parameter_480" + original_name = "batch_norm2d_26.b_0" shape = [256] dtype = "float32" min_val = float("-0.200827") @@ -5101,6 +5608,7 @@ class Program_weight_tensor_parameter_480: class Program_weight_tensor_parameter_481: name = "parameter_481" + original_name = "batch_norm2d_26.w_0" shape = [256] dtype = "float32" min_val = float("0.0806692") @@ -5112,6 +5620,7 @@ class Program_weight_tensor_parameter_481: class Program_weight_tensor_parameter_482: name = "parameter_482" + original_name = "batch_norm2d_26.w_2" shape = [256] dtype = "float32" min_val = float("0.00175098") @@ -5123,6 +5632,7 @@ class Program_weight_tensor_parameter_482: class Program_weight_tensor_parameter_483: name = "parameter_483" + original_name = "batch_norm2d_26.w_1" shape = [256] dtype = "float32" min_val = float("-0.154164") @@ -5134,6 +5644,7 @@ class Program_weight_tensor_parameter_483: class Program_weight_tensor_parameter_484: name = "parameter_484" + original_name = "conv2d_26.w_0" shape = [256, 512, 1, 1] dtype = "float32" min_val = float("-0.126123") @@ -5145,6 +5656,7 @@ class Program_weight_tensor_parameter_484: class Program_weight_tensor_parameter_485: name = "parameter_485" + original_name = "batch_norm2d_25.b_0" shape = [512] dtype = "float32" min_val = float("-0.213783") @@ -5156,6 +5668,7 @@ class Program_weight_tensor_parameter_485: class Program_weight_tensor_parameter_486: name = "parameter_486" + original_name = "batch_norm2d_25.w_0" shape = [512] dtype = "float32" min_val = float("-0.132737") @@ -5167,6 +5680,7 @@ class Program_weight_tensor_parameter_486: class Program_weight_tensor_parameter_487: name = "parameter_487" + original_name = "batch_norm2d_25.w_2" shape = [512] dtype = "float32" min_val = float("1.17707e-05") @@ -5178,6 +5692,7 @@ class Program_weight_tensor_parameter_487: class Program_weight_tensor_parameter_488: name = "parameter_488" + original_name = "batch_norm2d_25.w_1" shape = [512] dtype = "float32" min_val = float("-0.0737471") @@ -5189,6 +5704,7 @@ class Program_weight_tensor_parameter_488: class Program_weight_tensor_parameter_489: name = "parameter_489" + original_name = "conv2d_25.w_0" shape = [512, 128, 1, 1] dtype = "float32" min_val = float("-0.19458") @@ -5200,6 +5716,7 @@ class Program_weight_tensor_parameter_489: class Program_weight_tensor_parameter_490: name = "parameter_490" + original_name = "batch_norm2d_24.b_0" shape = [128] dtype = "float32" min_val = float("-0.188612") @@ -5211,6 +5728,7 @@ class Program_weight_tensor_parameter_490: class Program_weight_tensor_parameter_491: name = "parameter_491" + original_name = "batch_norm2d_24.w_0" shape = [128] dtype = "float32" min_val = float("0.0808353") @@ -5222,6 +5740,7 @@ class Program_weight_tensor_parameter_491: class Program_weight_tensor_parameter_492: name = "parameter_492" + original_name = "batch_norm2d_24.w_2" shape = [128] dtype = "float32" min_val = float("0.00107675") @@ -5233,6 +5752,7 @@ class Program_weight_tensor_parameter_492: class Program_weight_tensor_parameter_493: name = "parameter_493" + original_name = "batch_norm2d_24.w_1" shape = [128] dtype = "float32" min_val = float("-0.0859865") @@ -5244,6 +5764,7 @@ class Program_weight_tensor_parameter_493: class Program_weight_tensor_parameter_494: name = "parameter_494" + original_name = "conv2d_24.w_0" shape = [128, 128, 3, 3] dtype = "float32" min_val = float("-0.136146") @@ -5255,6 +5776,7 @@ class Program_weight_tensor_parameter_494: class Program_weight_tensor_parameter_495: name = "parameter_495" + original_name = "batch_norm2d_23.b_0" shape = [128] dtype = "float32" min_val = float("-0.133446") @@ -5266,6 +5788,7 @@ class Program_weight_tensor_parameter_495: class Program_weight_tensor_parameter_496: name = "parameter_496" + original_name = "batch_norm2d_23.w_0" shape = [128] dtype = "float32" min_val = float("0.0832541") @@ -5277,6 +5800,7 @@ class Program_weight_tensor_parameter_496: class Program_weight_tensor_parameter_497: name = "parameter_497" + original_name = "batch_norm2d_23.w_2" shape = [128] dtype = "float32" min_val = float("0.00151614") @@ -5288,6 +5812,7 @@ class Program_weight_tensor_parameter_497: class Program_weight_tensor_parameter_498: name = "parameter_498" + original_name = "batch_norm2d_23.w_1" shape = [128] dtype = "float32" min_val = float("-0.0779984") @@ -5299,6 +5824,7 @@ class Program_weight_tensor_parameter_498: class Program_weight_tensor_parameter_499: name = "parameter_499" + original_name = "conv2d_23.w_0" shape = [128, 512, 1, 1] dtype = "float32" min_val = float("-0.116246") @@ -5310,6 +5836,7 @@ class Program_weight_tensor_parameter_499: class Program_weight_tensor_parameter_500: name = "parameter_500" + original_name = "batch_norm2d_22.b_0" shape = [512] dtype = "float32" min_val = float("-0.206455") @@ -5321,6 +5848,7 @@ class Program_weight_tensor_parameter_500: class Program_weight_tensor_parameter_501: name = "parameter_501" + original_name = "batch_norm2d_22.w_0" shape = [512] dtype = "float32" min_val = float("-0.0548628") @@ -5332,6 +5860,7 @@ class Program_weight_tensor_parameter_501: class Program_weight_tensor_parameter_502: name = "parameter_502" + original_name = "batch_norm2d_22.w_2" shape = [512] dtype = "float32" min_val = float("5.60519e-45") @@ -5343,6 +5872,7 @@ class Program_weight_tensor_parameter_502: class Program_weight_tensor_parameter_503: name = "parameter_503" + original_name = "batch_norm2d_22.w_1" shape = [512] dtype = "float32" min_val = float("-0.0518352") @@ -5354,6 +5884,7 @@ class Program_weight_tensor_parameter_503: class Program_weight_tensor_parameter_504: name = "parameter_504" + original_name = "conv2d_22.w_0" shape = [512, 128, 1, 1] dtype = "float32" min_val = float("-0.153612") @@ -5365,6 +5896,7 @@ class Program_weight_tensor_parameter_504: class Program_weight_tensor_parameter_505: name = "parameter_505" + original_name = "batch_norm2d_21.b_0" shape = [128] dtype = "float32" min_val = float("-0.18239") @@ -5376,6 +5908,7 @@ class Program_weight_tensor_parameter_505: class Program_weight_tensor_parameter_506: name = "parameter_506" + original_name = "batch_norm2d_21.w_0" shape = [128] dtype = "float32" min_val = float("0.0744946") @@ -5387,6 +5920,7 @@ class Program_weight_tensor_parameter_506: class Program_weight_tensor_parameter_507: name = "parameter_507" + original_name = "batch_norm2d_21.w_2" shape = [128] dtype = "float32" min_val = float("0.00122704") @@ -5398,6 +5932,7 @@ class Program_weight_tensor_parameter_507: class Program_weight_tensor_parameter_508: name = "parameter_508" + original_name = "batch_norm2d_21.w_1" shape = [128] dtype = "float32" min_val = float("-0.09244") @@ -5409,6 +5944,7 @@ class Program_weight_tensor_parameter_508: class Program_weight_tensor_parameter_509: name = "parameter_509" + original_name = "conv2d_21.w_0" shape = [128, 128, 3, 3] dtype = "float32" min_val = float("-0.175254") @@ -5420,6 +5956,7 @@ class Program_weight_tensor_parameter_509: class Program_weight_tensor_parameter_510: name = "parameter_510" + original_name = "batch_norm2d_20.b_0" shape = [128] dtype = "float32" min_val = float("-0.264595") @@ -5431,6 +5968,7 @@ class Program_weight_tensor_parameter_510: class Program_weight_tensor_parameter_511: name = "parameter_511" + original_name = "batch_norm2d_20.w_0" shape = [128] dtype = "float32" min_val = float("0.0727919") @@ -5442,6 +5980,7 @@ class Program_weight_tensor_parameter_511: class Program_weight_tensor_parameter_512: name = "parameter_512" + original_name = "batch_norm2d_20.w_2" shape = [128] dtype = "float32" min_val = float("0.00204811") @@ -5453,6 +5992,7 @@ class Program_weight_tensor_parameter_512: class Program_weight_tensor_parameter_513: name = "parameter_513" + original_name = "batch_norm2d_20.w_1" shape = [128] dtype = "float32" min_val = float("-0.292813") @@ -5464,6 +6004,7 @@ class Program_weight_tensor_parameter_513: class Program_weight_tensor_parameter_514: name = "parameter_514" + original_name = "conv2d_20.w_0" shape = [128, 512, 1, 1] dtype = "float32" min_val = float("-0.158225") @@ -5475,6 +6016,7 @@ class Program_weight_tensor_parameter_514: class Program_weight_tensor_parameter_515: name = "parameter_515" + original_name = "batch_norm2d_19.b_0" shape = [512] dtype = "float32" min_val = float("-0.272751") @@ -5486,6 +6028,7 @@ class Program_weight_tensor_parameter_515: class Program_weight_tensor_parameter_516: name = "parameter_516" + original_name = "batch_norm2d_19.w_0" shape = [512] dtype = "float32" min_val = float("-0.0773685") @@ -5497,6 +6040,7 @@ class Program_weight_tensor_parameter_516: class Program_weight_tensor_parameter_517: name = "parameter_517" + original_name = "batch_norm2d_19.w_2" shape = [512] dtype = "float32" min_val = float("5.60519e-45") @@ -5508,6 +6052,7 @@ class Program_weight_tensor_parameter_517: class Program_weight_tensor_parameter_518: name = "parameter_518" + original_name = "batch_norm2d_19.w_1" shape = [512] dtype = "float32" min_val = float("-0.0518335") @@ -5519,6 +6064,7 @@ class Program_weight_tensor_parameter_518: class Program_weight_tensor_parameter_519: name = "parameter_519" + original_name = "conv2d_19.w_0" shape = [512, 128, 1, 1] dtype = "float32" min_val = float("-0.212086") @@ -5530,6 +6076,7 @@ class Program_weight_tensor_parameter_519: class Program_weight_tensor_parameter_520: name = "parameter_520" + original_name = "batch_norm2d_18.b_0" shape = [128] dtype = "float32" min_val = float("-0.207375") @@ -5541,6 +6088,7 @@ class Program_weight_tensor_parameter_520: class Program_weight_tensor_parameter_521: name = "parameter_521" + original_name = "batch_norm2d_18.w_0" shape = [128] dtype = "float32" min_val = float("0.0714045") @@ -5552,6 +6100,7 @@ class Program_weight_tensor_parameter_521: class Program_weight_tensor_parameter_522: name = "parameter_522" + original_name = "batch_norm2d_18.w_2" shape = [128] dtype = "float32" min_val = float("0.00167628") @@ -5563,6 +6112,7 @@ class Program_weight_tensor_parameter_522: class Program_weight_tensor_parameter_523: name = "parameter_523" + original_name = "batch_norm2d_18.w_1" shape = [128] dtype = "float32" min_val = float("-0.388887") @@ -5574,6 +6124,7 @@ class Program_weight_tensor_parameter_523: class Program_weight_tensor_parameter_524: name = "parameter_524" + original_name = "conv2d_18.w_0" shape = [128, 128, 3, 3] dtype = "float32" min_val = float("-0.239234") @@ -5585,6 +6136,7 @@ class Program_weight_tensor_parameter_524: class Program_weight_tensor_parameter_525: name = "parameter_525" + original_name = "batch_norm2d_17.b_0" shape = [128] dtype = "float32" min_val = float("-0.150355") @@ -5596,6 +6148,7 @@ class Program_weight_tensor_parameter_525: class Program_weight_tensor_parameter_526: name = "parameter_526" + original_name = "batch_norm2d_17.w_0" shape = [128] dtype = "float32" min_val = float("0.0835381") @@ -5607,6 +6160,7 @@ class Program_weight_tensor_parameter_526: class Program_weight_tensor_parameter_527: name = "parameter_527" + original_name = "batch_norm2d_17.w_2" shape = [128] dtype = "float32" min_val = float("0.00224778") @@ -5618,6 +6172,7 @@ class Program_weight_tensor_parameter_527: class Program_weight_tensor_parameter_528: name = "parameter_528" + original_name = "batch_norm2d_17.w_1" shape = [128] dtype = "float32" min_val = float("-0.267656") @@ -5629,6 +6184,7 @@ class Program_weight_tensor_parameter_528: class Program_weight_tensor_parameter_529: name = "parameter_529" + original_name = "conv2d_17.w_0" shape = [128, 512, 1, 1] dtype = "float32" min_val = float("-0.192112") @@ -5640,6 +6196,7 @@ class Program_weight_tensor_parameter_529: class Program_weight_tensor_parameter_530: name = "parameter_530" + original_name = "batch_norm2d_16.b_0" shape = [512] dtype = "float32" min_val = float("-0.174138") @@ -5651,6 +6208,7 @@ class Program_weight_tensor_parameter_530: class Program_weight_tensor_parameter_531: name = "parameter_531" + original_name = "batch_norm2d_16.w_0" shape = [512] dtype = "float32" min_val = float("-0.0299296") @@ -5662,6 +6220,7 @@ class Program_weight_tensor_parameter_531: class Program_weight_tensor_parameter_532: name = "parameter_532" + original_name = "batch_norm2d_16.w_2" shape = [512] dtype = "float32" min_val = float("5.60519e-45") @@ -5673,6 +6232,7 @@ class Program_weight_tensor_parameter_532: class Program_weight_tensor_parameter_533: name = "parameter_533" + original_name = "batch_norm2d_16.w_1" shape = [512] dtype = "float32" min_val = float("-0.162818") @@ -5684,6 +6244,7 @@ class Program_weight_tensor_parameter_533: class Program_weight_tensor_parameter_534: name = "parameter_534" + original_name = "conv2d_16.w_0" shape = [512, 256, 1, 1] dtype = "float32" min_val = float("-0.52586") @@ -5695,6 +6256,7 @@ class Program_weight_tensor_parameter_534: class Program_weight_tensor_parameter_535: name = "parameter_535" + original_name = "batch_norm2d_15.b_0" shape = [512] dtype = "float32" min_val = float("-0.174138") @@ -5706,6 +6268,7 @@ class Program_weight_tensor_parameter_535: class Program_weight_tensor_parameter_536: name = "parameter_536" + original_name = "batch_norm2d_15.w_0" shape = [512] dtype = "float32" min_val = float("-0.0561179") @@ -5717,6 +6280,7 @@ class Program_weight_tensor_parameter_536: class Program_weight_tensor_parameter_537: name = "parameter_537" + original_name = "batch_norm2d_15.w_2" shape = [512] dtype = "float32" min_val = float("5.60519e-45") @@ -5728,6 +6292,7 @@ class Program_weight_tensor_parameter_537: class Program_weight_tensor_parameter_538: name = "parameter_538" + original_name = "batch_norm2d_15.w_1" shape = [512] dtype = "float32" min_val = float("-0.0692909") @@ -5739,6 +6304,7 @@ class Program_weight_tensor_parameter_538: class Program_weight_tensor_parameter_539: name = "parameter_539" + original_name = "conv2d_15.w_0" shape = [512, 128, 1, 1] dtype = "float32" min_val = float("-0.207854") @@ -5750,6 +6316,7 @@ class Program_weight_tensor_parameter_539: class Program_weight_tensor_parameter_540: name = "parameter_540" + original_name = "batch_norm2d_14.b_0" shape = [128] dtype = "float32" min_val = float("-0.125217") @@ -5761,6 +6328,7 @@ class Program_weight_tensor_parameter_540: class Program_weight_tensor_parameter_541: name = "parameter_541" + original_name = "batch_norm2d_14.w_0" shape = [128] dtype = "float32" min_val = float("0.0760892") @@ -5772,6 +6340,7 @@ class Program_weight_tensor_parameter_541: class Program_weight_tensor_parameter_542: name = "parameter_542" + original_name = "batch_norm2d_14.w_2" shape = [128] dtype = "float32" min_val = float("0.000951838") @@ -5783,6 +6352,7 @@ class Program_weight_tensor_parameter_542: class Program_weight_tensor_parameter_543: name = "parameter_543" + original_name = "batch_norm2d_14.w_1" shape = [128] dtype = "float32" min_val = float("-0.126163") @@ -5794,6 +6364,7 @@ class Program_weight_tensor_parameter_543: class Program_weight_tensor_parameter_544: name = "parameter_544" + original_name = "conv2d_14.w_0" shape = [128, 128, 3, 3] dtype = "float32" min_val = float("-0.096496") @@ -5805,6 +6376,7 @@ class Program_weight_tensor_parameter_544: class Program_weight_tensor_parameter_545: name = "parameter_545" + original_name = "batch_norm2d_13.b_0" shape = [128] dtype = "float32" min_val = float("-0.175984") @@ -5816,6 +6388,7 @@ class Program_weight_tensor_parameter_545: class Program_weight_tensor_parameter_546: name = "parameter_546" + original_name = "batch_norm2d_13.w_0" shape = [128] dtype = "float32" min_val = float("0.0741341") @@ -5827,6 +6400,7 @@ class Program_weight_tensor_parameter_546: class Program_weight_tensor_parameter_547: name = "parameter_547" + original_name = "batch_norm2d_13.w_2" shape = [128] dtype = "float32" min_val = float("0.00194663") @@ -5838,6 +6412,7 @@ class Program_weight_tensor_parameter_547: class Program_weight_tensor_parameter_548: name = "parameter_548" + original_name = "batch_norm2d_13.w_1" shape = [128] dtype = "float32" min_val = float("-0.185112") @@ -5849,6 +6424,7 @@ class Program_weight_tensor_parameter_548: class Program_weight_tensor_parameter_549: name = "parameter_549" + original_name = "conv2d_13.w_0" shape = [128, 256, 1, 1] dtype = "float32" min_val = float("-0.16558") @@ -5860,6 +6436,7 @@ class Program_weight_tensor_parameter_549: class Program_weight_tensor_parameter_550: name = "parameter_550" + original_name = "batch_norm2d_12.b_0" shape = [256] dtype = "float32" min_val = float("-0.242305") @@ -5871,6 +6448,7 @@ class Program_weight_tensor_parameter_550: class Program_weight_tensor_parameter_551: name = "parameter_551" + original_name = "batch_norm2d_12.w_0" shape = [256] dtype = "float32" min_val = float("-0.0764245") @@ -5882,6 +6460,7 @@ class Program_weight_tensor_parameter_551: class Program_weight_tensor_parameter_552: name = "parameter_552" + original_name = "batch_norm2d_12.w_2" shape = [256] dtype = "float32" min_val = float("1.47993e-05") @@ -5893,6 +6472,7 @@ class Program_weight_tensor_parameter_552: class Program_weight_tensor_parameter_553: name = "parameter_553" + original_name = "batch_norm2d_12.w_1" shape = [256] dtype = "float32" min_val = float("-0.070885") @@ -5904,6 +6484,7 @@ class Program_weight_tensor_parameter_553: class Program_weight_tensor_parameter_554: name = "parameter_554" + original_name = "conv2d_12.w_0" shape = [256, 64, 1, 1] dtype = "float32" min_val = float("-0.169234") @@ -5915,6 +6496,7 @@ class Program_weight_tensor_parameter_554: class Program_weight_tensor_parameter_555: name = "parameter_555" + original_name = "batch_norm2d_11.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -5924,6 +6506,7 @@ class Program_weight_tensor_parameter_555: class Program_weight_tensor_parameter_556: name = "parameter_556" + original_name = "batch_norm2d_11.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -5933,6 +6516,7 @@ class Program_weight_tensor_parameter_556: class Program_weight_tensor_parameter_557: name = "parameter_557" + original_name = "batch_norm2d_11.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -5942,6 +6526,7 @@ class Program_weight_tensor_parameter_557: class Program_weight_tensor_parameter_558: name = "parameter_558" + original_name = "batch_norm2d_11.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -5951,6 +6536,7 @@ class Program_weight_tensor_parameter_558: class Program_weight_tensor_parameter_559: name = "parameter_559" + original_name = "conv2d_11.w_0" shape = [64, 64, 3, 3] dtype = "float32" min_val = float("-0.197603") @@ -5962,6 +6548,7 @@ class Program_weight_tensor_parameter_559: class Program_weight_tensor_parameter_560: name = "parameter_560" + original_name = "batch_norm2d_10.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -5971,6 +6558,7 @@ class Program_weight_tensor_parameter_560: class Program_weight_tensor_parameter_561: name = "parameter_561" + original_name = "batch_norm2d_10.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -5980,6 +6568,7 @@ class Program_weight_tensor_parameter_561: class Program_weight_tensor_parameter_562: name = "parameter_562" + original_name = "batch_norm2d_10.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -5989,6 +6578,7 @@ class Program_weight_tensor_parameter_562: class Program_weight_tensor_parameter_563: name = "parameter_563" + original_name = "batch_norm2d_10.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -5998,6 +6588,7 @@ class Program_weight_tensor_parameter_563: class Program_weight_tensor_parameter_564: name = "parameter_564" + original_name = "conv2d_10.w_0" shape = [64, 256, 1, 1] dtype = "float32" min_val = float("-0.186245") @@ -6009,6 +6600,7 @@ class Program_weight_tensor_parameter_564: class Program_weight_tensor_parameter_565: name = "parameter_565" + original_name = "batch_norm2d_9.b_0" shape = [256] dtype = "float32" min_val = float("-0.20351") @@ -6020,6 +6612,7 @@ class Program_weight_tensor_parameter_565: class Program_weight_tensor_parameter_566: name = "parameter_566" + original_name = "batch_norm2d_9.w_0" shape = [256] dtype = "float32" min_val = float("-0.0916299") @@ -6031,6 +6624,7 @@ class Program_weight_tensor_parameter_566: class Program_weight_tensor_parameter_567: name = "parameter_567" + original_name = "batch_norm2d_9.w_2" shape = [256] dtype = "float32" min_val = float("5.60519e-45") @@ -6042,6 +6636,7 @@ class Program_weight_tensor_parameter_567: class Program_weight_tensor_parameter_568: name = "parameter_568" + original_name = "batch_norm2d_9.w_1" shape = [256] dtype = "float32" min_val = float("-0.154898") @@ -6053,6 +6648,7 @@ class Program_weight_tensor_parameter_568: class Program_weight_tensor_parameter_569: name = "parameter_569" + original_name = "conv2d_9.w_0" shape = [256, 64, 1, 1] dtype = "float32" min_val = float("-0.228203") @@ -6064,6 +6660,7 @@ class Program_weight_tensor_parameter_569: class Program_weight_tensor_parameter_570: name = "parameter_570" + original_name = "batch_norm2d_8.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -6073,6 +6670,7 @@ class Program_weight_tensor_parameter_570: class Program_weight_tensor_parameter_571: name = "parameter_571" + original_name = "batch_norm2d_8.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -6082,6 +6680,7 @@ class Program_weight_tensor_parameter_571: class Program_weight_tensor_parameter_572: name = "parameter_572" + original_name = "batch_norm2d_8.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -6091,6 +6690,7 @@ class Program_weight_tensor_parameter_572: class Program_weight_tensor_parameter_573: name = "parameter_573" + original_name = "batch_norm2d_8.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -6100,6 +6700,7 @@ class Program_weight_tensor_parameter_573: class Program_weight_tensor_parameter_574: name = "parameter_574" + original_name = "conv2d_8.w_0" shape = [64, 64, 3, 3] dtype = "float32" min_val = float("-0.215985") @@ -6111,6 +6712,7 @@ class Program_weight_tensor_parameter_574: class Program_weight_tensor_parameter_575: name = "parameter_575" + original_name = "batch_norm2d_7.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -6120,6 +6722,7 @@ class Program_weight_tensor_parameter_575: class Program_weight_tensor_parameter_576: name = "parameter_576" + original_name = "batch_norm2d_7.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -6129,6 +6732,7 @@ class Program_weight_tensor_parameter_576: class Program_weight_tensor_parameter_577: name = "parameter_577" + original_name = "batch_norm2d_7.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -6138,6 +6742,7 @@ class Program_weight_tensor_parameter_577: class Program_weight_tensor_parameter_578: name = "parameter_578" + original_name = "batch_norm2d_7.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -6147,6 +6752,7 @@ class Program_weight_tensor_parameter_578: class Program_weight_tensor_parameter_579: name = "parameter_579" + original_name = "conv2d_7.w_0" shape = [64, 256, 1, 1] dtype = "float32" min_val = float("-0.217827") @@ -6158,6 +6764,7 @@ class Program_weight_tensor_parameter_579: class Program_weight_tensor_parameter_580: name = "parameter_580" + original_name = "batch_norm2d_6.b_0" shape = [256] dtype = "float32" min_val = float("-0.126803") @@ -6169,6 +6776,7 @@ class Program_weight_tensor_parameter_580: class Program_weight_tensor_parameter_581: name = "parameter_581" + original_name = "batch_norm2d_6.w_0" shape = [256] dtype = "float32" min_val = float("-0.0899233") @@ -6180,6 +6788,7 @@ class Program_weight_tensor_parameter_581: class Program_weight_tensor_parameter_582: name = "parameter_582" + original_name = "batch_norm2d_6.w_2" shape = [256] dtype = "float32" min_val = float("5.60519e-45") @@ -6191,6 +6800,7 @@ class Program_weight_tensor_parameter_582: class Program_weight_tensor_parameter_583: name = "parameter_583" + original_name = "batch_norm2d_6.w_1" shape = [256] dtype = "float32" min_val = float("-0.527159") @@ -6202,6 +6812,7 @@ class Program_weight_tensor_parameter_583: class Program_weight_tensor_parameter_584: name = "parameter_584" + original_name = "conv2d_6.w_0" shape = [256, 64, 1, 1] dtype = "float32" min_val = float("-0.736433") @@ -6213,6 +6824,7 @@ class Program_weight_tensor_parameter_584: class Program_weight_tensor_parameter_585: name = "parameter_585" + original_name = "batch_norm2d_5.b_0" shape = [256] dtype = "float32" min_val = float("-0.126803") @@ -6224,6 +6836,7 @@ class Program_weight_tensor_parameter_585: class Program_weight_tensor_parameter_586: name = "parameter_586" + original_name = "batch_norm2d_5.w_0" shape = [256] dtype = "float32" min_val = float("-0.0793541") @@ -6235,6 +6848,7 @@ class Program_weight_tensor_parameter_586: class Program_weight_tensor_parameter_587: name = "parameter_587" + original_name = "batch_norm2d_5.w_2" shape = [256] dtype = "float32" min_val = float("5.60519e-45") @@ -6246,6 +6860,7 @@ class Program_weight_tensor_parameter_587: class Program_weight_tensor_parameter_588: name = "parameter_588" + original_name = "batch_norm2d_5.w_1" shape = [256] dtype = "float32" min_val = float("-0.096113") @@ -6257,6 +6872,7 @@ class Program_weight_tensor_parameter_588: class Program_weight_tensor_parameter_589: name = "parameter_589" + original_name = "conv2d_5.w_0" shape = [256, 64, 1, 1] dtype = "float32" min_val = float("-0.22305") @@ -6268,6 +6884,7 @@ class Program_weight_tensor_parameter_589: class Program_weight_tensor_parameter_590: name = "parameter_590" + original_name = "batch_norm2d_4.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -6277,6 +6894,7 @@ class Program_weight_tensor_parameter_590: class Program_weight_tensor_parameter_591: name = "parameter_591" + original_name = "batch_norm2d_4.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -6286,6 +6904,7 @@ class Program_weight_tensor_parameter_591: class Program_weight_tensor_parameter_592: name = "parameter_592" + original_name = "batch_norm2d_4.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -6295,6 +6914,7 @@ class Program_weight_tensor_parameter_592: class Program_weight_tensor_parameter_593: name = "parameter_593" + original_name = "batch_norm2d_4.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -6304,6 +6924,7 @@ class Program_weight_tensor_parameter_593: class Program_weight_tensor_parameter_594: name = "parameter_594" + original_name = "conv2d_4.w_0" shape = [64, 64, 3, 3] dtype = "float32" min_val = float("-0.376788") @@ -6315,6 +6936,7 @@ class Program_weight_tensor_parameter_594: class Program_weight_tensor_parameter_595: name = "parameter_595" + original_name = "batch_norm2d_3.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -6324,6 +6946,7 @@ class Program_weight_tensor_parameter_595: class Program_weight_tensor_parameter_596: name = "parameter_596" + original_name = "batch_norm2d_3.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -6333,6 +6956,7 @@ class Program_weight_tensor_parameter_596: class Program_weight_tensor_parameter_597: name = "parameter_597" + original_name = "batch_norm2d_3.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -6342,6 +6966,7 @@ class Program_weight_tensor_parameter_597: class Program_weight_tensor_parameter_598: name = "parameter_598" + original_name = "batch_norm2d_3.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -6351,6 +6976,7 @@ class Program_weight_tensor_parameter_598: class Program_weight_tensor_parameter_599: name = "parameter_599" + original_name = "conv2d_3.w_0" shape = [64, 64, 1, 1] dtype = "float32" min_val = float("-0.451502") @@ -6362,6 +6988,7 @@ class Program_weight_tensor_parameter_599: class Program_weight_tensor_parameter_600: name = "parameter_600" + original_name = "batch_norm2d_2.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -6371,6 +6998,7 @@ class Program_weight_tensor_parameter_600: class Program_weight_tensor_parameter_601: name = "parameter_601" + original_name = "batch_norm2d_2.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -6380,6 +7008,7 @@ class Program_weight_tensor_parameter_601: class Program_weight_tensor_parameter_602: name = "parameter_602" + original_name = "batch_norm2d_2.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -6389,6 +7018,7 @@ class Program_weight_tensor_parameter_602: class Program_weight_tensor_parameter_603: name = "parameter_603" + original_name = "batch_norm2d_2.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -6398,6 +7028,7 @@ class Program_weight_tensor_parameter_603: class Program_weight_tensor_parameter_604: name = "parameter_604" + original_name = "conv2d_2.w_0" shape = [64, 32, 3, 3] dtype = "float32" min_val = float("-0.457575") @@ -6409,6 +7040,7 @@ class Program_weight_tensor_parameter_604: class Program_weight_tensor_parameter_605: name = "parameter_605" + original_name = "batch_norm2d_1.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -6418,6 +7050,7 @@ class Program_weight_tensor_parameter_605: class Program_weight_tensor_parameter_606: name = "parameter_606" + original_name = "batch_norm2d_1.w_0" shape = [32] dtype = "float32" min_val = float("0") @@ -6427,6 +7060,7 @@ class Program_weight_tensor_parameter_606: class Program_weight_tensor_parameter_607: name = "parameter_607" + original_name = "batch_norm2d_1.w_2" shape = [32] dtype = "float32" min_val = float("0") @@ -6436,6 +7070,7 @@ class Program_weight_tensor_parameter_607: class Program_weight_tensor_parameter_608: name = "parameter_608" + original_name = "batch_norm2d_1.w_1" shape = [32] dtype = "float32" min_val = float("0") @@ -6445,6 +7080,7 @@ class Program_weight_tensor_parameter_608: class Program_weight_tensor_parameter_609: name = "parameter_609" + original_name = "conv2d_1.w_0" shape = [32, 32, 3, 3] dtype = "float32" min_val = float("-0.734028") @@ -6456,6 +7092,7 @@ class Program_weight_tensor_parameter_609: class Program_weight_tensor_parameter_610: name = "parameter_610" + original_name = "batch_norm2d_0.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -6465,6 +7102,7 @@ class Program_weight_tensor_parameter_610: class Program_weight_tensor_parameter_611: name = "parameter_611" + original_name = "batch_norm2d_0.w_0" shape = [32] dtype = "float32" min_val = float("0") @@ -6474,6 +7112,7 @@ class Program_weight_tensor_parameter_611: class Program_weight_tensor_parameter_612: name = "parameter_612" + original_name = "batch_norm2d_0.w_2" shape = [32] dtype = "float32" min_val = float("0") @@ -6483,6 +7122,7 @@ class Program_weight_tensor_parameter_612: class Program_weight_tensor_parameter_613: name = "parameter_613" + original_name = "batch_norm2d_0.w_1" shape = [32] dtype = "float32" min_val = float("0") @@ -6492,6 +7132,7 @@ class Program_weight_tensor_parameter_613: class Program_weight_tensor_parameter_614: name = "parameter_614" + original_name = "conv2d_0.w_0" shape = [32, 3, 3, 3] dtype = "float32" min_val = float("-0.756744") diff --git a/paddle_samples/PaddleX/Deeplabv3_Plus-R50/subgraph_1/input_meta.py b/paddle_samples/PaddleX/Deeplabv3_Plus-R50/subgraph_1/input_meta.py index 60db74418..ff4d55767 100644 --- a/paddle_samples/PaddleX/Deeplabv3_Plus-R50/subgraph_1/input_meta.py +++ b/paddle_samples/PaddleX/Deeplabv3_Plus-R50/subgraph_1/input_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_data_0: name = "data_0" + original_name = "var_0" shape = [2, 3, 512, 1024] dtype = "float32" min_val = float("-1.0") diff --git a/paddle_samples/PaddleX/Deeplabv3_Plus-R50/subgraph_1/weight_meta.py b/paddle_samples/PaddleX/Deeplabv3_Plus-R50/subgraph_1/weight_meta.py index eb43c6c87..d75a94032 100644 --- a/paddle_samples/PaddleX/Deeplabv3_Plus-R50/subgraph_1/weight_meta.py +++ b/paddle_samples/PaddleX/Deeplabv3_Plus-R50/subgraph_1/weight_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_parameter_0: name = "parameter_0" + original_name = "conv2d_69.b_0" shape = [2] dtype = "float32" min_val = float("0") @@ -9,6 +10,7 @@ class Program_weight_tensor_parameter_0: class Program_weight_tensor_parameter_1: name = "parameter_1" + original_name = "conv2d_69.w_0" shape = [2, 256, 1, 1] dtype = "float32" min_val = float("-0.30596") @@ -20,6 +22,7 @@ class Program_weight_tensor_parameter_1: class Program_weight_tensor_parameter_2: name = "parameter_2" + original_name = "batch_norm2d_68.b_0" shape = [256] dtype = "float32" data = None @@ -27,6 +30,7 @@ class Program_weight_tensor_parameter_2: class Program_weight_tensor_parameter_3: name = "parameter_3" + original_name = "batch_norm2d_68.w_0" shape = [256] dtype = "float32" min_val = float("1.0") @@ -37,6 +41,7 @@ class Program_weight_tensor_parameter_3: class Program_weight_tensor_parameter_4: name = "parameter_4" + original_name = "batch_norm2d_68.w_2" shape = [256] dtype = "float32" min_val = float("1.0") @@ -47,6 +52,7 @@ class Program_weight_tensor_parameter_4: class Program_weight_tensor_parameter_5: name = "parameter_5" + original_name = "batch_norm2d_68.w_1" shape = [256] dtype = "float32" data = None @@ -54,6 +60,7 @@ class Program_weight_tensor_parameter_5: class Program_weight_tensor_parameter_6: name = "parameter_6" + original_name = "conv2d_68.b_0" shape = [256] dtype = "float32" data = None @@ -61,6 +68,7 @@ class Program_weight_tensor_parameter_6: class Program_weight_tensor_parameter_7: name = "parameter_7" + original_name = "conv2d_68.w_0" shape = [256, 256, 1, 1] dtype = "float32" min_val = float("-0.380838") @@ -72,6 +80,7 @@ class Program_weight_tensor_parameter_7: class Program_weight_tensor_parameter_8: name = "parameter_8" + original_name = "batch_norm2d_67.b_0" shape = [256] dtype = "float32" data = None @@ -79,26 +88,31 @@ class Program_weight_tensor_parameter_8: class Program_weight_tensor_parameter_9: name = "parameter_9" + original_name = "batch_norm2d_67.w_0" shape = [256] dtype = "float32" min_val = float("1.0") max_val = float("1.0") mean = float("1.0") + std = float("0.0") data = None class Program_weight_tensor_parameter_10: name = "parameter_10" + original_name = "batch_norm2d_67.w_2" shape = [256] dtype = "float32" min_val = float("1.0") max_val = float("1.0") mean = float("1.0") + std = float("0.0") data = None class Program_weight_tensor_parameter_11: name = "parameter_11" + original_name = "batch_norm2d_67.w_1" shape = [256] dtype = "float32" data = None @@ -106,6 +120,7 @@ class Program_weight_tensor_parameter_11: class Program_weight_tensor_parameter_12: name = "parameter_12" + original_name = "conv2d_67.b_0" shape = [256] dtype = "float32" data = None @@ -113,6 +128,7 @@ class Program_weight_tensor_parameter_12: class Program_weight_tensor_parameter_13: name = "parameter_13" + original_name = "conv2d_67.w_0" shape = [256, 1, 3, 3] dtype = "float32" min_val = float("-0.100088") @@ -124,6 +140,7 @@ class Program_weight_tensor_parameter_13: class Program_weight_tensor_parameter_14: name = "parameter_14" + original_name = "batch_norm2d_66.b_0" shape = [256] dtype = "float32" data = None @@ -131,26 +148,31 @@ class Program_weight_tensor_parameter_14: class Program_weight_tensor_parameter_15: name = "parameter_15" + original_name = "batch_norm2d_66.w_0" shape = [256] dtype = "float32" min_val = float("1.0") max_val = float("1.0") mean = float("1.0") + std = float("0.0") data = None class Program_weight_tensor_parameter_16: name = "parameter_16" + original_name = "batch_norm2d_66.w_2" shape = [256] dtype = "float32" min_val = float("1.0") max_val = float("1.0") mean = float("1.0") + std = float("0.0") data = None class Program_weight_tensor_parameter_17: name = "parameter_17" + original_name = "batch_norm2d_66.w_1" shape = [256] dtype = "float32" data = None @@ -158,6 +180,7 @@ class Program_weight_tensor_parameter_17: class Program_weight_tensor_parameter_18: name = "parameter_18" + original_name = "conv2d_66.b_0" shape = [256] dtype = "float32" data = None @@ -165,6 +188,7 @@ class Program_weight_tensor_parameter_18: class Program_weight_tensor_parameter_19: name = "parameter_19" + original_name = "conv2d_66.w_0" shape = [256, 304, 1, 1] dtype = "float32" min_val = float("-0.363711") @@ -176,6 +200,7 @@ class Program_weight_tensor_parameter_19: class Program_weight_tensor_parameter_20: name = "parameter_20" + original_name = "batch_norm2d_65.b_0" shape = [304] dtype = "float32" data = None @@ -183,26 +208,31 @@ class Program_weight_tensor_parameter_20: class Program_weight_tensor_parameter_21: name = "parameter_21" + original_name = "batch_norm2d_65.w_0" shape = [304] dtype = "float32" min_val = float("1.0") max_val = float("1.0") mean = float("1.0") + std = float("0.0") data = None class Program_weight_tensor_parameter_22: name = "parameter_22" + original_name = "batch_norm2d_65.w_2" shape = [304] dtype = "float32" min_val = float("1.0") max_val = float("1.0") mean = float("1.0") + std = float("0.0") data = None class Program_weight_tensor_parameter_23: name = "parameter_23" + original_name = "batch_norm2d_65.w_1" shape = [304] dtype = "float32" data = None @@ -210,6 +240,7 @@ class Program_weight_tensor_parameter_23: class Program_weight_tensor_parameter_24: name = "parameter_24" + original_name = "conv2d_65.b_0" shape = [304] dtype = "float32" data = None @@ -217,6 +248,7 @@ class Program_weight_tensor_parameter_24: class Program_weight_tensor_parameter_25: name = "parameter_25" + original_name = "conv2d_65.w_0" shape = [304, 1, 3, 3] dtype = "float32" min_val = float("-0.0929631") @@ -228,6 +260,7 @@ class Program_weight_tensor_parameter_25: class Program_weight_tensor_parameter_26: name = "parameter_26" + original_name = "batch_norm2d_64.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -237,6 +270,7 @@ class Program_weight_tensor_parameter_26: class Program_weight_tensor_parameter_27: name = "parameter_27" + original_name = "batch_norm2d_64.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -246,6 +280,7 @@ class Program_weight_tensor_parameter_27: class Program_weight_tensor_parameter_28: name = "parameter_28" + original_name = "batch_norm2d_64.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -255,6 +290,7 @@ class Program_weight_tensor_parameter_28: class Program_weight_tensor_parameter_29: name = "parameter_29" + original_name = "batch_norm2d_64.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -264,6 +300,7 @@ class Program_weight_tensor_parameter_29: class Program_weight_tensor_parameter_30: name = "parameter_30" + original_name = "conv2d_64.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -273,6 +310,7 @@ class Program_weight_tensor_parameter_30: class Program_weight_tensor_parameter_31: name = "parameter_31" + original_name = "conv2d_64.w_0" shape = [48, 256, 1, 1] dtype = "float32" min_val = float("-0.314136") @@ -284,6 +322,7 @@ class Program_weight_tensor_parameter_31: class Program_weight_tensor_parameter_32: name = "parameter_32" + original_name = "batch_norm2d_63.b_0" shape = [256] dtype = "float32" data = None @@ -291,26 +330,31 @@ class Program_weight_tensor_parameter_32: class Program_weight_tensor_parameter_33: name = "parameter_33" + original_name = "batch_norm2d_63.w_0" shape = [256] dtype = "float32" min_val = float("1.0") max_val = float("1.0") mean = float("1.0") + std = float("0.0") data = None class Program_weight_tensor_parameter_34: name = "parameter_34" + original_name = "batch_norm2d_63.w_2" shape = [256] dtype = "float32" min_val = float("1.0") max_val = float("1.0") mean = float("1.0") + std = float("0.0") data = None class Program_weight_tensor_parameter_35: name = "parameter_35" + original_name = "batch_norm2d_63.w_1" shape = [256] dtype = "float32" data = None @@ -318,6 +362,7 @@ class Program_weight_tensor_parameter_35: class Program_weight_tensor_parameter_36: name = "parameter_36" + original_name = "conv2d_63.b_0" shape = [256] dtype = "float32" data = None @@ -325,6 +370,7 @@ class Program_weight_tensor_parameter_36: class Program_weight_tensor_parameter_37: name = "parameter_37" + original_name = "conv2d_63.w_0" shape = [256, 1280, 1, 1] dtype = "float32" min_val = float("-0.183748") @@ -336,6 +382,7 @@ class Program_weight_tensor_parameter_37: class Program_weight_tensor_parameter_38: name = "parameter_38" + original_name = "batch_norm2d_62.b_0" shape = [256] dtype = "float32" data = None @@ -343,26 +390,31 @@ class Program_weight_tensor_parameter_38: class Program_weight_tensor_parameter_39: name = "parameter_39" + original_name = "batch_norm2d_62.w_0" shape = [256] dtype = "float32" min_val = float("1.0") max_val = float("1.0") mean = float("1.0") + std = float("0.0") data = None class Program_weight_tensor_parameter_40: name = "parameter_40" + original_name = "batch_norm2d_62.w_2" shape = [256] dtype = "float32" min_val = float("1.0") max_val = float("1.0") mean = float("1.0") + std = float("0.0") data = None class Program_weight_tensor_parameter_41: name = "parameter_41" + original_name = "batch_norm2d_62.w_1" shape = [256] dtype = "float32" data = None @@ -370,6 +422,7 @@ class Program_weight_tensor_parameter_41: class Program_weight_tensor_parameter_42: name = "parameter_42" + original_name = "conv2d_62.w_0" shape = [256, 2048, 1, 1] dtype = "float32" min_val = float("-0.149363") @@ -381,6 +434,7 @@ class Program_weight_tensor_parameter_42: class Program_weight_tensor_parameter_43: name = "parameter_43" + original_name = "batch_norm2d_61.b_0" shape = [256] dtype = "float32" data = None @@ -388,26 +442,31 @@ class Program_weight_tensor_parameter_43: class Program_weight_tensor_parameter_44: name = "parameter_44" + original_name = "batch_norm2d_61.w_0" shape = [256] dtype = "float32" min_val = float("1.0") max_val = float("1.0") mean = float("1.0") + std = float("0.0") data = None class Program_weight_tensor_parameter_45: name = "parameter_45" + original_name = "batch_norm2d_61.w_2" shape = [256] dtype = "float32" min_val = float("1.0") max_val = float("1.0") mean = float("1.0") + std = float("0.0") data = None class Program_weight_tensor_parameter_46: name = "parameter_46" + original_name = "batch_norm2d_61.w_1" shape = [256] dtype = "float32" data = None @@ -415,6 +474,7 @@ class Program_weight_tensor_parameter_46: class Program_weight_tensor_parameter_47: name = "parameter_47" + original_name = "conv2d_61.b_0" shape = [256] dtype = "float32" data = None @@ -422,6 +482,7 @@ class Program_weight_tensor_parameter_47: class Program_weight_tensor_parameter_48: name = "parameter_48" + original_name = "conv2d_61.w_0" shape = [256, 2048, 1, 1] dtype = "float32" min_val = float("-0.153666") @@ -433,6 +494,7 @@ class Program_weight_tensor_parameter_48: class Program_weight_tensor_parameter_49: name = "parameter_49" + original_name = "batch_norm2d_60.b_0" shape = [2048] dtype = "float32" data = None @@ -440,26 +502,31 @@ class Program_weight_tensor_parameter_49: class Program_weight_tensor_parameter_50: name = "parameter_50" + original_name = "batch_norm2d_60.w_0" shape = [2048] dtype = "float32" min_val = float("1.0") max_val = float("1.0") mean = float("1.0") + std = float("0.0") data = None class Program_weight_tensor_parameter_51: name = "parameter_51" + original_name = "batch_norm2d_60.w_2" shape = [2048] dtype = "float32" min_val = float("1.0") max_val = float("1.0") mean = float("1.0") + std = float("0.0") data = None class Program_weight_tensor_parameter_52: name = "parameter_52" + original_name = "batch_norm2d_60.w_1" shape = [2048] dtype = "float32" data = None @@ -467,6 +534,7 @@ class Program_weight_tensor_parameter_52: class Program_weight_tensor_parameter_53: name = "parameter_53" + original_name = "conv2d_60.b_0" shape = [2048] dtype = "float32" data = None @@ -474,6 +542,7 @@ class Program_weight_tensor_parameter_53: class Program_weight_tensor_parameter_54: name = "parameter_54" + original_name = "conv2d_60.w_0" shape = [2048, 1, 3, 3] dtype = "float32" min_val = float("-0.0419179") @@ -485,6 +554,7 @@ class Program_weight_tensor_parameter_54: class Program_weight_tensor_parameter_55: name = "parameter_55" + original_name = "batch_norm2d_59.b_0" shape = [256] dtype = "float32" data = None @@ -492,26 +562,31 @@ class Program_weight_tensor_parameter_55: class Program_weight_tensor_parameter_56: name = "parameter_56" + original_name = "batch_norm2d_59.w_0" shape = [256] dtype = "float32" min_val = float("1.0") max_val = float("1.0") mean = float("1.0") + std = float("0.0") data = None class Program_weight_tensor_parameter_57: name = "parameter_57" + original_name = "batch_norm2d_59.w_2" shape = [256] dtype = "float32" min_val = float("1.0") max_val = float("1.0") mean = float("1.0") + std = float("0.0") data = None class Program_weight_tensor_parameter_58: name = "parameter_58" + original_name = "batch_norm2d_59.w_1" shape = [256] dtype = "float32" data = None @@ -519,6 +594,7 @@ class Program_weight_tensor_parameter_58: class Program_weight_tensor_parameter_59: name = "parameter_59" + original_name = "conv2d_59.b_0" shape = [256] dtype = "float32" data = None @@ -526,6 +602,7 @@ class Program_weight_tensor_parameter_59: class Program_weight_tensor_parameter_60: name = "parameter_60" + original_name = "conv2d_59.w_0" shape = [256, 2048, 1, 1] dtype = "float32" min_val = float("-0.142097") @@ -537,6 +614,7 @@ class Program_weight_tensor_parameter_60: class Program_weight_tensor_parameter_61: name = "parameter_61" + original_name = "batch_norm2d_58.b_0" shape = [2048] dtype = "float32" data = None @@ -544,26 +622,31 @@ class Program_weight_tensor_parameter_61: class Program_weight_tensor_parameter_62: name = "parameter_62" + original_name = "batch_norm2d_58.w_0" shape = [2048] dtype = "float32" min_val = float("1.0") max_val = float("1.0") mean = float("1.0") + std = float("0.0") data = None class Program_weight_tensor_parameter_63: name = "parameter_63" + original_name = "batch_norm2d_58.w_2" shape = [2048] dtype = "float32" min_val = float("1.0") max_val = float("1.0") mean = float("1.0") + std = float("0.0") data = None class Program_weight_tensor_parameter_64: name = "parameter_64" + original_name = "batch_norm2d_58.w_1" shape = [2048] dtype = "float32" data = None @@ -571,6 +654,7 @@ class Program_weight_tensor_parameter_64: class Program_weight_tensor_parameter_65: name = "parameter_65" + original_name = "conv2d_58.b_0" shape = [2048] dtype = "float32" data = None @@ -578,6 +662,7 @@ class Program_weight_tensor_parameter_65: class Program_weight_tensor_parameter_66: name = "parameter_66" + original_name = "conv2d_58.w_0" shape = [2048, 1, 3, 3] dtype = "float32" min_val = float("-0.0427812") @@ -589,6 +674,7 @@ class Program_weight_tensor_parameter_66: class Program_weight_tensor_parameter_67: name = "parameter_67" + original_name = "batch_norm2d_57.b_0" shape = [256] dtype = "float32" data = None @@ -596,26 +682,31 @@ class Program_weight_tensor_parameter_67: class Program_weight_tensor_parameter_68: name = "parameter_68" + original_name = "batch_norm2d_57.w_0" shape = [256] dtype = "float32" min_val = float("1.0") max_val = float("1.0") mean = float("1.0") + std = float("0.0") data = None class Program_weight_tensor_parameter_69: name = "parameter_69" + original_name = "batch_norm2d_57.w_2" shape = [256] dtype = "float32" min_val = float("1.0") max_val = float("1.0") mean = float("1.0") + std = float("0.0") data = None class Program_weight_tensor_parameter_70: name = "parameter_70" + original_name = "batch_norm2d_57.w_1" shape = [256] dtype = "float32" data = None @@ -623,6 +714,7 @@ class Program_weight_tensor_parameter_70: class Program_weight_tensor_parameter_71: name = "parameter_71" + original_name = "conv2d_57.b_0" shape = [256] dtype = "float32" data = None @@ -630,6 +722,7 @@ class Program_weight_tensor_parameter_71: class Program_weight_tensor_parameter_72: name = "parameter_72" + original_name = "conv2d_57.w_0" shape = [256, 2048, 1, 1] dtype = "float32" min_val = float("-0.145608") @@ -641,6 +734,7 @@ class Program_weight_tensor_parameter_72: class Program_weight_tensor_parameter_73: name = "parameter_73" + original_name = "batch_norm2d_56.b_0" shape = [2048] dtype = "float32" data = None @@ -648,26 +742,31 @@ class Program_weight_tensor_parameter_73: class Program_weight_tensor_parameter_74: name = "parameter_74" + original_name = "batch_norm2d_56.w_0" shape = [2048] dtype = "float32" min_val = float("1.0") max_val = float("1.0") mean = float("1.0") + std = float("0.0") data = None class Program_weight_tensor_parameter_75: name = "parameter_75" + original_name = "batch_norm2d_56.w_2" shape = [2048] dtype = "float32" min_val = float("1.0") max_val = float("1.0") mean = float("1.0") + std = float("0.0") data = None class Program_weight_tensor_parameter_76: name = "parameter_76" + original_name = "batch_norm2d_56.w_1" shape = [2048] dtype = "float32" data = None @@ -675,6 +774,7 @@ class Program_weight_tensor_parameter_76: class Program_weight_tensor_parameter_77: name = "parameter_77" + original_name = "conv2d_56.b_0" shape = [2048] dtype = "float32" data = None @@ -682,6 +782,7 @@ class Program_weight_tensor_parameter_77: class Program_weight_tensor_parameter_78: name = "parameter_78" + original_name = "conv2d_56.w_0" shape = [2048, 1, 3, 3] dtype = "float32" min_val = float("-0.0419395") @@ -693,6 +794,7 @@ class Program_weight_tensor_parameter_78: class Program_weight_tensor_parameter_79: name = "parameter_79" + original_name = "batch_norm2d_55.b_0" shape = [256] dtype = "float32" data = None @@ -700,26 +802,31 @@ class Program_weight_tensor_parameter_79: class Program_weight_tensor_parameter_80: name = "parameter_80" + original_name = "batch_norm2d_55.w_0" shape = [256] dtype = "float32" min_val = float("1.0") max_val = float("1.0") mean = float("1.0") + std = float("0.0") data = None class Program_weight_tensor_parameter_81: name = "parameter_81" + original_name = "batch_norm2d_55.w_2" shape = [256] dtype = "float32" min_val = float("1.0") max_val = float("1.0") mean = float("1.0") + std = float("0.0") data = None class Program_weight_tensor_parameter_82: name = "parameter_82" + original_name = "batch_norm2d_55.w_1" shape = [256] dtype = "float32" data = None @@ -727,6 +834,7 @@ class Program_weight_tensor_parameter_82: class Program_weight_tensor_parameter_83: name = "parameter_83" + original_name = "conv2d_55.b_0" shape = [256] dtype = "float32" data = None @@ -734,6 +842,7 @@ class Program_weight_tensor_parameter_83: class Program_weight_tensor_parameter_84: name = "parameter_84" + original_name = "conv2d_55.w_0" shape = [256, 2048, 1, 1] dtype = "float32" min_val = float("-0.140792") @@ -745,6 +854,7 @@ class Program_weight_tensor_parameter_84: class Program_weight_tensor_parameter_85: name = "parameter_85" + original_name = "batch_norm2d_54.b_0" shape = [2048] dtype = "float32" min_val = float("-0.282166") @@ -756,6 +866,7 @@ class Program_weight_tensor_parameter_85: class Program_weight_tensor_parameter_86: name = "parameter_86" + original_name = "batch_norm2d_54.w_0" shape = [2048] dtype = "float32" min_val = float("0.714089") @@ -767,6 +878,7 @@ class Program_weight_tensor_parameter_86: class Program_weight_tensor_parameter_87: name = "parameter_87" + original_name = "batch_norm2d_54.w_2" shape = [2048] dtype = "float32" min_val = float("0.000386176") @@ -778,6 +890,7 @@ class Program_weight_tensor_parameter_87: class Program_weight_tensor_parameter_88: name = "parameter_88" + original_name = "batch_norm2d_54.w_1" shape = [2048] dtype = "float32" min_val = float("-0.016995") @@ -789,6 +902,7 @@ class Program_weight_tensor_parameter_88: class Program_weight_tensor_parameter_89: name = "parameter_89" + original_name = "conv2d_54.w_0" shape = [2048, 512, 1, 1] dtype = "float32" min_val = float("-0.234447") @@ -800,6 +914,7 @@ class Program_weight_tensor_parameter_89: class Program_weight_tensor_parameter_90: name = "parameter_90" + original_name = "batch_norm2d_53.b_0" shape = [512] dtype = "float32" min_val = float("-0.397288") @@ -811,6 +926,7 @@ class Program_weight_tensor_parameter_90: class Program_weight_tensor_parameter_91: name = "parameter_91" + original_name = "batch_norm2d_53.w_0" shape = [512] dtype = "float32" min_val = float("0.109729") @@ -822,6 +938,7 @@ class Program_weight_tensor_parameter_91: class Program_weight_tensor_parameter_92: name = "parameter_92" + original_name = "batch_norm2d_53.w_2" shape = [512] dtype = "float32" min_val = float("0.00483171") @@ -833,6 +950,7 @@ class Program_weight_tensor_parameter_92: class Program_weight_tensor_parameter_93: name = "parameter_93" + original_name = "batch_norm2d_53.w_1" shape = [512] dtype = "float32" min_val = float("-0.411076") @@ -844,6 +962,7 @@ class Program_weight_tensor_parameter_93: class Program_weight_tensor_parameter_94: name = "parameter_94" + original_name = "conv2d_53.w_0" shape = [512, 512, 3, 3] dtype = "float32" min_val = float("-0.455835") @@ -855,6 +974,7 @@ class Program_weight_tensor_parameter_94: class Program_weight_tensor_parameter_95: name = "parameter_95" + original_name = "batch_norm2d_52.b_0" shape = [512] dtype = "float32" min_val = float("-0.402652") @@ -866,6 +986,7 @@ class Program_weight_tensor_parameter_95: class Program_weight_tensor_parameter_96: name = "parameter_96" + original_name = "batch_norm2d_52.w_0" shape = [512] dtype = "float32" min_val = float("0.0818051") @@ -877,6 +998,7 @@ class Program_weight_tensor_parameter_96: class Program_weight_tensor_parameter_97: name = "parameter_97" + original_name = "batch_norm2d_52.w_2" shape = [512] dtype = "float32" min_val = float("0.00297223") @@ -888,6 +1010,7 @@ class Program_weight_tensor_parameter_97: class Program_weight_tensor_parameter_98: name = "parameter_98" + original_name = "batch_norm2d_52.w_1" shape = [512] dtype = "float32" min_val = float("-0.0819767") @@ -899,6 +1022,7 @@ class Program_weight_tensor_parameter_98: class Program_weight_tensor_parameter_99: name = "parameter_99" + original_name = "conv2d_52.w_0" shape = [512, 2048, 1, 1] dtype = "float32" min_val = float("-0.791667") @@ -910,6 +1034,7 @@ class Program_weight_tensor_parameter_99: class Program_weight_tensor_parameter_100: name = "parameter_100" + original_name = "batch_norm2d_51.b_0" shape = [2048] dtype = "float32" min_val = float("-0.445082") @@ -921,6 +1046,7 @@ class Program_weight_tensor_parameter_100: class Program_weight_tensor_parameter_101: name = "parameter_101" + original_name = "batch_norm2d_51.w_0" shape = [2048] dtype = "float32" min_val = float("-0.0510686") @@ -932,6 +1058,7 @@ class Program_weight_tensor_parameter_101: class Program_weight_tensor_parameter_102: name = "parameter_102" + original_name = "batch_norm2d_51.w_2" shape = [2048] dtype = "float32" min_val = float("0.000911003") @@ -943,6 +1070,7 @@ class Program_weight_tensor_parameter_102: class Program_weight_tensor_parameter_103: name = "parameter_103" + original_name = "batch_norm2d_51.w_1" shape = [2048] dtype = "float32" min_val = float("-0.858086") @@ -954,6 +1082,7 @@ class Program_weight_tensor_parameter_103: class Program_weight_tensor_parameter_104: name = "parameter_104" + original_name = "conv2d_51.w_0" shape = [2048, 512, 1, 1] dtype = "float32" min_val = float("-1.21905") @@ -965,6 +1094,7 @@ class Program_weight_tensor_parameter_104: class Program_weight_tensor_parameter_105: name = "parameter_105" + original_name = "batch_norm2d_50.b_0" shape = [512] dtype = "float32" min_val = float("-0.334352") @@ -976,6 +1106,7 @@ class Program_weight_tensor_parameter_105: class Program_weight_tensor_parameter_106: name = "parameter_106" + original_name = "batch_norm2d_50.w_0" shape = [512] dtype = "float32" min_val = float("0.134261") @@ -987,6 +1118,7 @@ class Program_weight_tensor_parameter_106: class Program_weight_tensor_parameter_107: name = "parameter_107" + original_name = "batch_norm2d_50.w_2" shape = [512] dtype = "float32" min_val = float("0.0166705") @@ -998,6 +1130,7 @@ class Program_weight_tensor_parameter_107: class Program_weight_tensor_parameter_108: name = "parameter_108" + original_name = "batch_norm2d_50.w_1" shape = [512] dtype = "float32" min_val = float("-0.884374") @@ -1009,6 +1142,7 @@ class Program_weight_tensor_parameter_108: class Program_weight_tensor_parameter_109: name = "parameter_109" + original_name = "conv2d_50.w_0" shape = [512, 512, 3, 3] dtype = "float32" min_val = float("-0.537283") @@ -1020,6 +1154,7 @@ class Program_weight_tensor_parameter_109: class Program_weight_tensor_parameter_110: name = "parameter_110" + original_name = "batch_norm2d_49.b_0" shape = [512] dtype = "float32" min_val = float("-0.266159") @@ -1031,6 +1166,7 @@ class Program_weight_tensor_parameter_110: class Program_weight_tensor_parameter_111: name = "parameter_111" + original_name = "batch_norm2d_49.w_0" shape = [512] dtype = "float32" min_val = float("0.114282") @@ -1042,6 +1178,7 @@ class Program_weight_tensor_parameter_111: class Program_weight_tensor_parameter_112: name = "parameter_112" + original_name = "batch_norm2d_49.w_2" shape = [512] dtype = "float32" min_val = float("0.00260214") @@ -1053,6 +1190,7 @@ class Program_weight_tensor_parameter_112: class Program_weight_tensor_parameter_113: name = "parameter_113" + original_name = "batch_norm2d_49.w_1" shape = [512] dtype = "float32" min_val = float("-0.176514") @@ -1064,6 +1202,7 @@ class Program_weight_tensor_parameter_113: class Program_weight_tensor_parameter_114: name = "parameter_114" + original_name = "conv2d_49.w_0" shape = [512, 2048, 1, 1] dtype = "float32" min_val = float("-0.489221") @@ -1075,6 +1214,7 @@ class Program_weight_tensor_parameter_114: class Program_weight_tensor_parameter_115: name = "parameter_115" + original_name = "batch_norm2d_48.b_0" shape = [2048] dtype = "float32" min_val = float("-0.284104") @@ -1086,6 +1226,7 @@ class Program_weight_tensor_parameter_115: class Program_weight_tensor_parameter_116: name = "parameter_116" + original_name = "batch_norm2d_48.w_0" shape = [2048] dtype = "float32" min_val = float("0.0373721") @@ -1097,6 +1238,7 @@ class Program_weight_tensor_parameter_116: class Program_weight_tensor_parameter_117: name = "parameter_117" + original_name = "batch_norm2d_48.w_2" shape = [2048] dtype = "float32" min_val = float("0.00248082") @@ -1108,6 +1250,7 @@ class Program_weight_tensor_parameter_117: class Program_weight_tensor_parameter_118: name = "parameter_118" + original_name = "batch_norm2d_48.w_1" shape = [2048] dtype = "float32" min_val = float("-0.556372") @@ -1119,6 +1262,7 @@ class Program_weight_tensor_parameter_118: class Program_weight_tensor_parameter_119: name = "parameter_119" + original_name = "conv2d_48.w_0" shape = [2048, 1024, 1, 1] dtype = "float32" min_val = float("-0.540116") @@ -1130,6 +1274,7 @@ class Program_weight_tensor_parameter_119: class Program_weight_tensor_parameter_120: name = "parameter_120" + original_name = "batch_norm2d_47.b_0" shape = [2048] dtype = "float32" min_val = float("-0.284104") @@ -1141,6 +1286,7 @@ class Program_weight_tensor_parameter_120: class Program_weight_tensor_parameter_121: name = "parameter_121" + original_name = "batch_norm2d_47.w_0" shape = [2048] dtype = "float32" min_val = float("0.0248532") @@ -1152,6 +1298,7 @@ class Program_weight_tensor_parameter_121: class Program_weight_tensor_parameter_122: name = "parameter_122" + original_name = "batch_norm2d_47.w_2" shape = [2048] dtype = "float32" min_val = float("0.00188508") @@ -1163,6 +1310,7 @@ class Program_weight_tensor_parameter_122: class Program_weight_tensor_parameter_123: name = "parameter_123" + original_name = "batch_norm2d_47.w_1" shape = [2048] dtype = "float32" min_val = float("-0.145631") @@ -1174,6 +1322,7 @@ class Program_weight_tensor_parameter_123: class Program_weight_tensor_parameter_124: name = "parameter_124" + original_name = "conv2d_47.w_0" shape = [2048, 512, 1, 1] dtype = "float32" min_val = float("-0.547884") @@ -1185,6 +1334,7 @@ class Program_weight_tensor_parameter_124: class Program_weight_tensor_parameter_125: name = "parameter_125" + original_name = "batch_norm2d_46.b_0" shape = [512] dtype = "float32" min_val = float("-0.267678") @@ -1196,6 +1346,7 @@ class Program_weight_tensor_parameter_125: class Program_weight_tensor_parameter_126: name = "parameter_126" + original_name = "batch_norm2d_46.w_0" shape = [512] dtype = "float32" min_val = float("0.115875") @@ -1207,6 +1358,7 @@ class Program_weight_tensor_parameter_126: class Program_weight_tensor_parameter_127: name = "parameter_127" + original_name = "batch_norm2d_46.w_2" shape = [512] dtype = "float32" min_val = float("0.00683442") @@ -1218,6 +1370,7 @@ class Program_weight_tensor_parameter_127: class Program_weight_tensor_parameter_128: name = "parameter_128" + original_name = "batch_norm2d_46.w_1" shape = [512] dtype = "float32" min_val = float("-0.356659") @@ -1229,6 +1382,7 @@ class Program_weight_tensor_parameter_128: class Program_weight_tensor_parameter_129: name = "parameter_129" + original_name = "conv2d_46.w_0" shape = [512, 512, 3, 3] dtype = "float32" min_val = float("-0.290132") @@ -1240,6 +1394,7 @@ class Program_weight_tensor_parameter_129: class Program_weight_tensor_parameter_130: name = "parameter_130" + original_name = "batch_norm2d_45.b_0" shape = [512] dtype = "float32" min_val = float("-0.455306") @@ -1251,6 +1406,7 @@ class Program_weight_tensor_parameter_130: class Program_weight_tensor_parameter_131: name = "parameter_131" + original_name = "batch_norm2d_45.w_0" shape = [512] dtype = "float32" min_val = float("0.0956012") @@ -1262,6 +1418,7 @@ class Program_weight_tensor_parameter_131: class Program_weight_tensor_parameter_132: name = "parameter_132" + original_name = "batch_norm2d_45.w_2" shape = [512] dtype = "float32" min_val = float("0.00995626") @@ -1273,6 +1430,7 @@ class Program_weight_tensor_parameter_132: class Program_weight_tensor_parameter_133: name = "parameter_133" + original_name = "batch_norm2d_45.w_1" shape = [512] dtype = "float32" min_val = float("-0.272144") @@ -1284,6 +1442,7 @@ class Program_weight_tensor_parameter_133: class Program_weight_tensor_parameter_134: name = "parameter_134" + original_name = "conv2d_45.w_0" shape = [512, 1024, 1, 1] dtype = "float32" min_val = float("-0.254182") @@ -1295,6 +1454,7 @@ class Program_weight_tensor_parameter_134: class Program_weight_tensor_parameter_135: name = "parameter_135" + original_name = "batch_norm2d_44.b_0" shape = [1024] dtype = "float32" min_val = float("-0.494356") @@ -1306,6 +1466,7 @@ class Program_weight_tensor_parameter_135: class Program_weight_tensor_parameter_136: name = "parameter_136" + original_name = "batch_norm2d_44.w_0" shape = [1024] dtype = "float32" min_val = float("-0.114335") @@ -1317,6 +1478,7 @@ class Program_weight_tensor_parameter_136: class Program_weight_tensor_parameter_137: name = "parameter_137" + original_name = "batch_norm2d_44.w_2" shape = [1024] dtype = "float32" min_val = float("0.000758928") @@ -1328,6 +1490,7 @@ class Program_weight_tensor_parameter_137: class Program_weight_tensor_parameter_138: name = "parameter_138" + original_name = "batch_norm2d_44.w_1" shape = [1024] dtype = "float32" min_val = float("-0.196626") @@ -1339,6 +1502,7 @@ class Program_weight_tensor_parameter_138: class Program_weight_tensor_parameter_139: name = "parameter_139" + original_name = "conv2d_44.w_0" shape = [1024, 256, 1, 1] dtype = "float32" min_val = float("-0.516216") @@ -1350,6 +1514,7 @@ class Program_weight_tensor_parameter_139: class Program_weight_tensor_parameter_140: name = "parameter_140" + original_name = "batch_norm2d_43.b_0" shape = [256] dtype = "float32" min_val = float("-0.628594") @@ -1361,6 +1526,7 @@ class Program_weight_tensor_parameter_140: class Program_weight_tensor_parameter_141: name = "parameter_141" + original_name = "batch_norm2d_43.w_0" shape = [256] dtype = "float32" min_val = float("0.129324") @@ -1372,6 +1538,7 @@ class Program_weight_tensor_parameter_141: class Program_weight_tensor_parameter_142: name = "parameter_142" + original_name = "batch_norm2d_43.w_2" shape = [256] dtype = "float32" min_val = float("0.0113363") @@ -1383,6 +1550,7 @@ class Program_weight_tensor_parameter_142: class Program_weight_tensor_parameter_143: name = "parameter_143" + original_name = "batch_norm2d_43.w_1" shape = [256] dtype = "float32" min_val = float("-0.469623") @@ -1394,6 +1562,7 @@ class Program_weight_tensor_parameter_143: class Program_weight_tensor_parameter_144: name = "parameter_144" + original_name = "conv2d_43.w_0" shape = [256, 256, 3, 3] dtype = "float32" min_val = float("-0.439431") @@ -1405,6 +1574,7 @@ class Program_weight_tensor_parameter_144: class Program_weight_tensor_parameter_145: name = "parameter_145" + original_name = "batch_norm2d_42.b_0" shape = [256] dtype = "float32" min_val = float("-0.465171") @@ -1416,6 +1586,7 @@ class Program_weight_tensor_parameter_145: class Program_weight_tensor_parameter_146: name = "parameter_146" + original_name = "batch_norm2d_42.w_0" shape = [256] dtype = "float32" min_val = float("0.119903") @@ -1427,6 +1598,7 @@ class Program_weight_tensor_parameter_146: class Program_weight_tensor_parameter_147: name = "parameter_147" + original_name = "batch_norm2d_42.w_2" shape = [256] dtype = "float32" min_val = float("0.0136785") @@ -1438,6 +1610,7 @@ class Program_weight_tensor_parameter_147: class Program_weight_tensor_parameter_148: name = "parameter_148" + original_name = "batch_norm2d_42.w_1" shape = [256] dtype = "float32" min_val = float("-0.267117") @@ -1449,6 +1622,7 @@ class Program_weight_tensor_parameter_148: class Program_weight_tensor_parameter_149: name = "parameter_149" + original_name = "conv2d_42.w_0" shape = [256, 1024, 1, 1] dtype = "float32" min_val = float("-0.386117") @@ -1460,6 +1634,7 @@ class Program_weight_tensor_parameter_149: class Program_weight_tensor_parameter_150: name = "parameter_150" + original_name = "batch_norm2d_41.b_0" shape = [1024] dtype = "float32" min_val = float("-0.505123") @@ -1471,6 +1646,7 @@ class Program_weight_tensor_parameter_150: class Program_weight_tensor_parameter_151: name = "parameter_151" + original_name = "batch_norm2d_41.w_0" shape = [1024] dtype = "float32" min_val = float("-0.071253") @@ -1482,6 +1658,7 @@ class Program_weight_tensor_parameter_151: class Program_weight_tensor_parameter_152: name = "parameter_152" + original_name = "batch_norm2d_41.w_2" shape = [1024] dtype = "float32" min_val = float("0.000556931") @@ -1493,6 +1670,7 @@ class Program_weight_tensor_parameter_152: class Program_weight_tensor_parameter_153: name = "parameter_153" + original_name = "batch_norm2d_41.w_1" shape = [1024] dtype = "float32" min_val = float("-0.194132") @@ -1504,6 +1682,7 @@ class Program_weight_tensor_parameter_153: class Program_weight_tensor_parameter_154: name = "parameter_154" + original_name = "conv2d_41.w_0" shape = [1024, 256, 1, 1] dtype = "float32" min_val = float("-0.680876") @@ -1515,6 +1694,7 @@ class Program_weight_tensor_parameter_154: class Program_weight_tensor_parameter_155: name = "parameter_155" + original_name = "batch_norm2d_40.b_0" shape = [256] dtype = "float32" min_val = float("-0.69082") @@ -1526,6 +1706,7 @@ class Program_weight_tensor_parameter_155: class Program_weight_tensor_parameter_156: name = "parameter_156" + original_name = "batch_norm2d_40.w_0" shape = [256] dtype = "float32" min_val = float("0.126233") @@ -1537,6 +1718,7 @@ class Program_weight_tensor_parameter_156: class Program_weight_tensor_parameter_157: name = "parameter_157" + original_name = "batch_norm2d_40.w_2" shape = [256] dtype = "float32" min_val = float("0.0117291") @@ -1548,6 +1730,7 @@ class Program_weight_tensor_parameter_157: class Program_weight_tensor_parameter_158: name = "parameter_158" + original_name = "batch_norm2d_40.w_1" shape = [256] dtype = "float32" min_val = float("-0.434858") @@ -1559,6 +1742,7 @@ class Program_weight_tensor_parameter_158: class Program_weight_tensor_parameter_159: name = "parameter_159" + original_name = "conv2d_40.w_0" shape = [256, 256, 3, 3] dtype = "float32" min_val = float("-0.308659") @@ -1570,6 +1754,7 @@ class Program_weight_tensor_parameter_159: class Program_weight_tensor_parameter_160: name = "parameter_160" + original_name = "batch_norm2d_39.b_0" shape = [256] dtype = "float32" min_val = float("-0.490268") @@ -1581,6 +1766,7 @@ class Program_weight_tensor_parameter_160: class Program_weight_tensor_parameter_161: name = "parameter_161" + original_name = "batch_norm2d_39.w_0" shape = [256] dtype = "float32" min_val = float("0.123382") @@ -1592,6 +1778,7 @@ class Program_weight_tensor_parameter_161: class Program_weight_tensor_parameter_162: name = "parameter_162" + original_name = "batch_norm2d_39.w_2" shape = [256] dtype = "float32" min_val = float("0.0127077") @@ -1603,6 +1790,7 @@ class Program_weight_tensor_parameter_162: class Program_weight_tensor_parameter_163: name = "parameter_163" + original_name = "batch_norm2d_39.w_1" shape = [256] dtype = "float32" min_val = float("-0.231346") @@ -1614,6 +1802,7 @@ class Program_weight_tensor_parameter_163: class Program_weight_tensor_parameter_164: name = "parameter_164" + original_name = "conv2d_39.w_0" shape = [256, 1024, 1, 1] dtype = "float32" min_val = float("-0.460862") @@ -1625,6 +1814,7 @@ class Program_weight_tensor_parameter_164: class Program_weight_tensor_parameter_165: name = "parameter_165" + original_name = "batch_norm2d_38.b_0" shape = [1024] dtype = "float32" min_val = float("-0.351691") @@ -1636,6 +1826,7 @@ class Program_weight_tensor_parameter_165: class Program_weight_tensor_parameter_166: name = "parameter_166" + original_name = "batch_norm2d_38.w_0" shape = [1024] dtype = "float32" min_val = float("-0.0798891") @@ -1647,6 +1838,7 @@ class Program_weight_tensor_parameter_166: class Program_weight_tensor_parameter_167: name = "parameter_167" + original_name = "batch_norm2d_38.w_2" shape = [1024] dtype = "float32" min_val = float("0.00029933") @@ -1658,6 +1850,7 @@ class Program_weight_tensor_parameter_167: class Program_weight_tensor_parameter_168: name = "parameter_168" + original_name = "batch_norm2d_38.w_1" shape = [1024] dtype = "float32" min_val = float("-0.135183") @@ -1669,6 +1862,7 @@ class Program_weight_tensor_parameter_168: class Program_weight_tensor_parameter_169: name = "parameter_169" + original_name = "conv2d_38.w_0" shape = [1024, 256, 1, 1] dtype = "float32" min_val = float("-0.412226") @@ -1680,6 +1874,7 @@ class Program_weight_tensor_parameter_169: class Program_weight_tensor_parameter_170: name = "parameter_170" + original_name = "batch_norm2d_37.b_0" shape = [256] dtype = "float32" min_val = float("-0.327533") @@ -1691,6 +1886,7 @@ class Program_weight_tensor_parameter_170: class Program_weight_tensor_parameter_171: name = "parameter_171" + original_name = "batch_norm2d_37.w_0" shape = [256] dtype = "float32" min_val = float("0.124773") @@ -1702,6 +1898,7 @@ class Program_weight_tensor_parameter_171: class Program_weight_tensor_parameter_172: name = "parameter_172" + original_name = "batch_norm2d_37.w_2" shape = [256] dtype = "float32" min_val = float("0.0132998") @@ -1713,6 +1910,7 @@ class Program_weight_tensor_parameter_172: class Program_weight_tensor_parameter_173: name = "parameter_173" + original_name = "batch_norm2d_37.w_1" shape = [256] dtype = "float32" min_val = float("-0.291819") @@ -1724,6 +1922,7 @@ class Program_weight_tensor_parameter_173: class Program_weight_tensor_parameter_174: name = "parameter_174" + original_name = "conv2d_37.w_0" shape = [256, 256, 3, 3] dtype = "float32" min_val = float("-0.389441") @@ -1735,6 +1934,7 @@ class Program_weight_tensor_parameter_174: class Program_weight_tensor_parameter_175: name = "parameter_175" + original_name = "batch_norm2d_36.b_0" shape = [256] dtype = "float32" min_val = float("-0.510875") @@ -1746,6 +1946,7 @@ class Program_weight_tensor_parameter_175: class Program_weight_tensor_parameter_176: name = "parameter_176" + original_name = "batch_norm2d_36.w_0" shape = [256] dtype = "float32" min_val = float("0.117607") @@ -1757,6 +1958,7 @@ class Program_weight_tensor_parameter_176: class Program_weight_tensor_parameter_177: name = "parameter_177" + original_name = "batch_norm2d_36.w_2" shape = [256] dtype = "float32" min_val = float("0.0131457") @@ -1768,6 +1970,7 @@ class Program_weight_tensor_parameter_177: class Program_weight_tensor_parameter_178: name = "parameter_178" + original_name = "batch_norm2d_36.w_1" shape = [256] dtype = "float32" min_val = float("-0.320565") @@ -1779,6 +1982,7 @@ class Program_weight_tensor_parameter_178: class Program_weight_tensor_parameter_179: name = "parameter_179" + original_name = "conv2d_36.w_0" shape = [256, 1024, 1, 1] dtype = "float32" min_val = float("-0.415199") @@ -1790,6 +1994,7 @@ class Program_weight_tensor_parameter_179: class Program_weight_tensor_parameter_180: name = "parameter_180" + original_name = "batch_norm2d_35.b_0" shape = [1024] dtype = "float32" min_val = float("-0.420672") @@ -1801,6 +2006,7 @@ class Program_weight_tensor_parameter_180: class Program_weight_tensor_parameter_181: name = "parameter_181" + original_name = "batch_norm2d_35.w_0" shape = [1024] dtype = "float32" min_val = float("-0.0631926") @@ -1812,6 +2018,7 @@ class Program_weight_tensor_parameter_181: class Program_weight_tensor_parameter_182: name = "parameter_182" + original_name = "batch_norm2d_35.w_2" shape = [1024] dtype = "float32" min_val = float("0.000697383") @@ -1823,6 +2030,7 @@ class Program_weight_tensor_parameter_182: class Program_weight_tensor_parameter_183: name = "parameter_183" + original_name = "batch_norm2d_35.w_1" shape = [1024] dtype = "float32" min_val = float("-0.161883") @@ -1834,6 +2042,7 @@ class Program_weight_tensor_parameter_183: class Program_weight_tensor_parameter_184: name = "parameter_184" + original_name = "conv2d_35.w_0" shape = [1024, 256, 1, 1] dtype = "float32" min_val = float("-0.411167") @@ -1845,6 +2054,7 @@ class Program_weight_tensor_parameter_184: class Program_weight_tensor_parameter_185: name = "parameter_185" + original_name = "batch_norm2d_34.b_0" shape = [256] dtype = "float32" min_val = float("-0.280491") @@ -1856,6 +2066,7 @@ class Program_weight_tensor_parameter_185: class Program_weight_tensor_parameter_186: name = "parameter_186" + original_name = "batch_norm2d_34.w_0" shape = [256] dtype = "float32" min_val = float("0.128527") @@ -1867,6 +2078,7 @@ class Program_weight_tensor_parameter_186: class Program_weight_tensor_parameter_187: name = "parameter_187" + original_name = "batch_norm2d_34.w_2" shape = [256] dtype = "float32" min_val = float("0.0181118") @@ -1878,6 +2090,7 @@ class Program_weight_tensor_parameter_187: class Program_weight_tensor_parameter_188: name = "parameter_188" + original_name = "batch_norm2d_34.w_1" shape = [256] dtype = "float32" min_val = float("-0.608816") @@ -1889,6 +2102,7 @@ class Program_weight_tensor_parameter_188: class Program_weight_tensor_parameter_189: name = "parameter_189" + original_name = "conv2d_34.w_0" shape = [256, 256, 3, 3] dtype = "float32" min_val = float("-0.29439") @@ -1900,6 +2114,7 @@ class Program_weight_tensor_parameter_189: class Program_weight_tensor_parameter_190: name = "parameter_190" + original_name = "batch_norm2d_33.b_0" shape = [256] dtype = "float32" min_val = float("-0.411786") @@ -1911,6 +2126,7 @@ class Program_weight_tensor_parameter_190: class Program_weight_tensor_parameter_191: name = "parameter_191" + original_name = "batch_norm2d_33.w_0" shape = [256] dtype = "float32" min_val = float("0.122476") @@ -1922,6 +2138,7 @@ class Program_weight_tensor_parameter_191: class Program_weight_tensor_parameter_192: name = "parameter_192" + original_name = "batch_norm2d_33.w_2" shape = [256] dtype = "float32" min_val = float("0.0136736") @@ -1933,6 +2150,7 @@ class Program_weight_tensor_parameter_192: class Program_weight_tensor_parameter_193: name = "parameter_193" + original_name = "batch_norm2d_33.w_1" shape = [256] dtype = "float32" min_val = float("-0.350375") @@ -1944,6 +2162,7 @@ class Program_weight_tensor_parameter_193: class Program_weight_tensor_parameter_194: name = "parameter_194" + original_name = "conv2d_33.w_0" shape = [256, 1024, 1, 1] dtype = "float32" min_val = float("-0.562702") @@ -1955,6 +2174,7 @@ class Program_weight_tensor_parameter_194: class Program_weight_tensor_parameter_195: name = "parameter_195" + original_name = "batch_norm2d_32.b_0" shape = [1024] dtype = "float32" min_val = float("-0.448255") @@ -1966,6 +2186,7 @@ class Program_weight_tensor_parameter_195: class Program_weight_tensor_parameter_196: name = "parameter_196" + original_name = "batch_norm2d_32.w_0" shape = [1024] dtype = "float32" min_val = float("-1.00724e-23") @@ -1977,6 +2198,7 @@ class Program_weight_tensor_parameter_196: class Program_weight_tensor_parameter_197: name = "parameter_197" + original_name = "batch_norm2d_32.w_2" shape = [1024] dtype = "float32" min_val = float("5.60519e-45") @@ -1988,6 +2210,7 @@ class Program_weight_tensor_parameter_197: class Program_weight_tensor_parameter_198: name = "parameter_198" + original_name = "batch_norm2d_32.w_1" shape = [1024] dtype = "float32" min_val = float("-0.238083") @@ -1999,6 +2222,7 @@ class Program_weight_tensor_parameter_198: class Program_weight_tensor_parameter_199: name = "parameter_199" + original_name = "conv2d_32.w_0" shape = [1024, 256, 1, 1] dtype = "float32" min_val = float("-0.723318") @@ -2010,6 +2234,7 @@ class Program_weight_tensor_parameter_199: class Program_weight_tensor_parameter_200: name = "parameter_200" + original_name = "batch_norm2d_31.b_0" shape = [256] dtype = "float32" min_val = float("-0.946089") @@ -2021,6 +2246,7 @@ class Program_weight_tensor_parameter_200: class Program_weight_tensor_parameter_201: name = "parameter_201" + original_name = "batch_norm2d_31.w_0" shape = [256] dtype = "float32" min_val = float("0.140842") @@ -2032,6 +2258,7 @@ class Program_weight_tensor_parameter_201: class Program_weight_tensor_parameter_202: name = "parameter_202" + original_name = "batch_norm2d_31.w_2" shape = [256] dtype = "float32" min_val = float("0.0257037") @@ -2043,6 +2270,7 @@ class Program_weight_tensor_parameter_202: class Program_weight_tensor_parameter_203: name = "parameter_203" + original_name = "batch_norm2d_31.w_1" shape = [256] dtype = "float32" min_val = float("-1.34447") @@ -2054,6 +2282,7 @@ class Program_weight_tensor_parameter_203: class Program_weight_tensor_parameter_204: name = "parameter_204" + original_name = "conv2d_31.w_0" shape = [256, 256, 3, 3] dtype = "float32" min_val = float("-0.594571") @@ -2065,6 +2294,7 @@ class Program_weight_tensor_parameter_204: class Program_weight_tensor_parameter_205: name = "parameter_205" + original_name = "batch_norm2d_30.b_0" shape = [256] dtype = "float32" min_val = float("-0.350109") @@ -2076,6 +2306,7 @@ class Program_weight_tensor_parameter_205: class Program_weight_tensor_parameter_206: name = "parameter_206" + original_name = "batch_norm2d_30.w_0" shape = [256] dtype = "float32" min_val = float("0.132203") @@ -2087,6 +2318,7 @@ class Program_weight_tensor_parameter_206: class Program_weight_tensor_parameter_207: name = "parameter_207" + original_name = "batch_norm2d_30.w_2" shape = [256] dtype = "float32" min_val = float("0.0143599") @@ -2098,6 +2330,7 @@ class Program_weight_tensor_parameter_207: class Program_weight_tensor_parameter_208: name = "parameter_208" + original_name = "batch_norm2d_30.w_1" shape = [256] dtype = "float32" min_val = float("-1.17498") @@ -2109,6 +2342,7 @@ class Program_weight_tensor_parameter_208: class Program_weight_tensor_parameter_209: name = "parameter_209" + original_name = "conv2d_30.w_0" shape = [256, 1024, 1, 1] dtype = "float32" min_val = float("-0.460684") @@ -2120,6 +2354,7 @@ class Program_weight_tensor_parameter_209: class Program_weight_tensor_parameter_210: name = "parameter_210" + original_name = "batch_norm2d_29.b_0" shape = [1024] dtype = "float32" min_val = float("-0.297989") @@ -2131,6 +2366,7 @@ class Program_weight_tensor_parameter_210: class Program_weight_tensor_parameter_211: name = "parameter_211" + original_name = "batch_norm2d_29.w_0" shape = [1024] dtype = "float32" min_val = float("-2.10863e-13") @@ -2142,6 +2378,7 @@ class Program_weight_tensor_parameter_211: class Program_weight_tensor_parameter_212: name = "parameter_212" + original_name = "batch_norm2d_29.w_2" shape = [1024] dtype = "float32" min_val = float("5.60519e-45") @@ -2153,6 +2390,7 @@ class Program_weight_tensor_parameter_212: class Program_weight_tensor_parameter_213: name = "parameter_213" + original_name = "batch_norm2d_29.w_1" shape = [1024] dtype = "float32" min_val = float("-0.488374") @@ -2164,6 +2402,7 @@ class Program_weight_tensor_parameter_213: class Program_weight_tensor_parameter_214: name = "parameter_214" + original_name = "conv2d_29.w_0" shape = [1024, 512, 1, 1] dtype = "float32" min_val = float("-0.714351") @@ -2175,6 +2414,7 @@ class Program_weight_tensor_parameter_214: class Program_weight_tensor_parameter_215: name = "parameter_215" + original_name = "batch_norm2d_28.b_0" shape = [1024] dtype = "float32" min_val = float("-0.297989") @@ -2186,6 +2426,7 @@ class Program_weight_tensor_parameter_215: class Program_weight_tensor_parameter_216: name = "parameter_216" + original_name = "batch_norm2d_28.w_0" shape = [1024] dtype = "float32" min_val = float("-0.0884224") @@ -2197,6 +2438,7 @@ class Program_weight_tensor_parameter_216: class Program_weight_tensor_parameter_217: name = "parameter_217" + original_name = "batch_norm2d_28.w_2" shape = [1024] dtype = "float32" min_val = float("5.60519e-45") @@ -2208,6 +2450,7 @@ class Program_weight_tensor_parameter_217: class Program_weight_tensor_parameter_218: name = "parameter_218" + original_name = "batch_norm2d_28.w_1" shape = [1024] dtype = "float32" min_val = float("-0.2498") @@ -2219,6 +2462,7 @@ class Program_weight_tensor_parameter_218: class Program_weight_tensor_parameter_219: name = "parameter_219" + original_name = "conv2d_28.w_0" shape = [1024, 256, 1, 1] dtype = "float32" min_val = float("-0.315775") @@ -2230,6 +2474,7 @@ class Program_weight_tensor_parameter_219: class Program_weight_tensor_parameter_220: name = "parameter_220" + original_name = "batch_norm2d_27.b_0" shape = [256] dtype = "float32" min_val = float("-0.403554") @@ -2241,6 +2486,7 @@ class Program_weight_tensor_parameter_220: class Program_weight_tensor_parameter_221: name = "parameter_221" + original_name = "batch_norm2d_27.w_0" shape = [256] dtype = "float32" min_val = float("0.119874") @@ -2252,6 +2498,7 @@ class Program_weight_tensor_parameter_221: class Program_weight_tensor_parameter_222: name = "parameter_222" + original_name = "batch_norm2d_27.w_2" shape = [256] dtype = "float32" min_val = float("0.00748281") @@ -2263,6 +2510,7 @@ class Program_weight_tensor_parameter_222: class Program_weight_tensor_parameter_223: name = "parameter_223" + original_name = "batch_norm2d_27.w_1" shape = [256] dtype = "float32" min_val = float("-0.210233") @@ -2274,6 +2522,7 @@ class Program_weight_tensor_parameter_223: class Program_weight_tensor_parameter_224: name = "parameter_224" + original_name = "conv2d_27.w_0" shape = [256, 256, 3, 3] dtype = "float32" min_val = float("-0.373469") @@ -2285,6 +2534,7 @@ class Program_weight_tensor_parameter_224: class Program_weight_tensor_parameter_225: name = "parameter_225" + original_name = "batch_norm2d_26.b_0" shape = [256] dtype = "float32" min_val = float("-0.472221") @@ -2296,6 +2546,7 @@ class Program_weight_tensor_parameter_225: class Program_weight_tensor_parameter_226: name = "parameter_226" + original_name = "batch_norm2d_26.w_0" shape = [256] dtype = "float32" min_val = float("0.128796") @@ -2307,6 +2558,7 @@ class Program_weight_tensor_parameter_226: class Program_weight_tensor_parameter_227: name = "parameter_227" + original_name = "batch_norm2d_26.w_2" shape = [256] dtype = "float32" min_val = float("0.0166948") @@ -2318,6 +2570,7 @@ class Program_weight_tensor_parameter_227: class Program_weight_tensor_parameter_228: name = "parameter_228" + original_name = "batch_norm2d_26.w_1" shape = [256] dtype = "float32" min_val = float("-0.336503") @@ -2329,6 +2582,7 @@ class Program_weight_tensor_parameter_228: class Program_weight_tensor_parameter_229: name = "parameter_229" + original_name = "conv2d_26.w_0" shape = [256, 512, 1, 1] dtype = "float32" min_val = float("-0.286994") @@ -2340,6 +2594,7 @@ class Program_weight_tensor_parameter_229: class Program_weight_tensor_parameter_230: name = "parameter_230" + original_name = "batch_norm2d_25.b_0" shape = [512] dtype = "float32" min_val = float("-0.351865") @@ -2351,6 +2606,7 @@ class Program_weight_tensor_parameter_230: class Program_weight_tensor_parameter_231: name = "parameter_231" + original_name = "batch_norm2d_25.w_0" shape = [512] dtype = "float32" min_val = float("-0.150051") @@ -2362,6 +2618,7 @@ class Program_weight_tensor_parameter_231: class Program_weight_tensor_parameter_232: name = "parameter_232" + original_name = "batch_norm2d_25.w_2" shape = [512] dtype = "float32" min_val = float("0.000305778") @@ -2373,6 +2630,7 @@ class Program_weight_tensor_parameter_232: class Program_weight_tensor_parameter_233: name = "parameter_233" + original_name = "batch_norm2d_25.w_1" shape = [512] dtype = "float32" min_val = float("-0.182206") @@ -2384,6 +2642,7 @@ class Program_weight_tensor_parameter_233: class Program_weight_tensor_parameter_234: name = "parameter_234" + original_name = "conv2d_25.w_0" shape = [512, 128, 1, 1] dtype = "float32" min_val = float("-0.437671") @@ -2395,6 +2654,7 @@ class Program_weight_tensor_parameter_234: class Program_weight_tensor_parameter_235: name = "parameter_235" + original_name = "batch_norm2d_24.b_0" shape = [128] dtype = "float32" min_val = float("-0.288327") @@ -2406,6 +2666,7 @@ class Program_weight_tensor_parameter_235: class Program_weight_tensor_parameter_236: name = "parameter_236" + original_name = "batch_norm2d_24.w_0" shape = [128] dtype = "float32" min_val = float("0.13864") @@ -2417,6 +2678,7 @@ class Program_weight_tensor_parameter_236: class Program_weight_tensor_parameter_237: name = "parameter_237" + original_name = "batch_norm2d_24.w_2" shape = [128] dtype = "float32" min_val = float("0.0139091") @@ -2428,6 +2690,7 @@ class Program_weight_tensor_parameter_237: class Program_weight_tensor_parameter_238: name = "parameter_238" + original_name = "batch_norm2d_24.w_1" shape = [128] dtype = "float32" min_val = float("-0.372374") @@ -2439,6 +2702,7 @@ class Program_weight_tensor_parameter_238: class Program_weight_tensor_parameter_239: name = "parameter_239" + original_name = "conv2d_24.w_0" shape = [128, 128, 3, 3] dtype = "float32" min_val = float("-0.304806") @@ -2450,6 +2714,7 @@ class Program_weight_tensor_parameter_239: class Program_weight_tensor_parameter_240: name = "parameter_240" + original_name = "batch_norm2d_23.b_0" shape = [128] dtype = "float32" min_val = float("-0.394085") @@ -2461,6 +2726,7 @@ class Program_weight_tensor_parameter_240: class Program_weight_tensor_parameter_241: name = "parameter_241" + original_name = "batch_norm2d_23.w_0" shape = [128] dtype = "float32" min_val = float("0.114849") @@ -2472,6 +2738,7 @@ class Program_weight_tensor_parameter_241: class Program_weight_tensor_parameter_242: name = "parameter_242" + original_name = "batch_norm2d_23.w_2" shape = [128] dtype = "float32" min_val = float("0.0206183") @@ -2483,6 +2750,7 @@ class Program_weight_tensor_parameter_242: class Program_weight_tensor_parameter_243: name = "parameter_243" + original_name = "batch_norm2d_23.w_1" shape = [128] dtype = "float32" min_val = float("-0.508035") @@ -2494,6 +2762,7 @@ class Program_weight_tensor_parameter_243: class Program_weight_tensor_parameter_244: name = "parameter_244" + original_name = "conv2d_23.w_0" shape = [128, 512, 1, 1] dtype = "float32" min_val = float("-0.338475") @@ -2505,6 +2774,7 @@ class Program_weight_tensor_parameter_244: class Program_weight_tensor_parameter_245: name = "parameter_245" + original_name = "batch_norm2d_22.b_0" shape = [512] dtype = "float32" min_val = float("-0.278261") @@ -2516,6 +2786,7 @@ class Program_weight_tensor_parameter_245: class Program_weight_tensor_parameter_246: name = "parameter_246" + original_name = "batch_norm2d_22.w_0" shape = [512] dtype = "float32" min_val = float("-0.0943411") @@ -2527,6 +2798,7 @@ class Program_weight_tensor_parameter_246: class Program_weight_tensor_parameter_247: name = "parameter_247" + original_name = "batch_norm2d_22.w_2" shape = [512] dtype = "float32" min_val = float("1.28703e-35") @@ -2538,6 +2810,7 @@ class Program_weight_tensor_parameter_247: class Program_weight_tensor_parameter_248: name = "parameter_248" + original_name = "batch_norm2d_22.w_1" shape = [512] dtype = "float32" min_val = float("-0.222267") @@ -2549,6 +2822,7 @@ class Program_weight_tensor_parameter_248: class Program_weight_tensor_parameter_249: name = "parameter_249" + original_name = "conv2d_22.w_0" shape = [512, 128, 1, 1] dtype = "float32" min_val = float("-0.394291") @@ -2560,6 +2834,7 @@ class Program_weight_tensor_parameter_249: class Program_weight_tensor_parameter_250: name = "parameter_250" + original_name = "batch_norm2d_21.b_0" shape = [128] dtype = "float32" min_val = float("-0.574827") @@ -2571,6 +2846,7 @@ class Program_weight_tensor_parameter_250: class Program_weight_tensor_parameter_251: name = "parameter_251" + original_name = "batch_norm2d_21.w_0" shape = [128] dtype = "float32" min_val = float("0.145856") @@ -2582,6 +2858,7 @@ class Program_weight_tensor_parameter_251: class Program_weight_tensor_parameter_252: name = "parameter_252" + original_name = "batch_norm2d_21.w_2" shape = [128] dtype = "float32" min_val = float("0.0140519") @@ -2593,6 +2870,7 @@ class Program_weight_tensor_parameter_252: class Program_weight_tensor_parameter_253: name = "parameter_253" + original_name = "batch_norm2d_21.w_1" shape = [128] dtype = "float32" min_val = float("-0.642487") @@ -2604,6 +2882,7 @@ class Program_weight_tensor_parameter_253: class Program_weight_tensor_parameter_254: name = "parameter_254" + original_name = "conv2d_21.w_0" shape = [128, 128, 3, 3] dtype = "float32" min_val = float("-0.330961") @@ -2615,6 +2894,7 @@ class Program_weight_tensor_parameter_254: class Program_weight_tensor_parameter_255: name = "parameter_255" + original_name = "batch_norm2d_20.b_0" shape = [128] dtype = "float32" min_val = float("-0.394333") @@ -2626,6 +2906,7 @@ class Program_weight_tensor_parameter_255: class Program_weight_tensor_parameter_256: name = "parameter_256" + original_name = "batch_norm2d_20.w_0" shape = [128] dtype = "float32" min_val = float("0.122299") @@ -2637,6 +2918,7 @@ class Program_weight_tensor_parameter_256: class Program_weight_tensor_parameter_257: name = "parameter_257" + original_name = "batch_norm2d_20.w_2" shape = [128] dtype = "float32" min_val = float("0.0114481") @@ -2648,6 +2930,7 @@ class Program_weight_tensor_parameter_257: class Program_weight_tensor_parameter_258: name = "parameter_258" + original_name = "batch_norm2d_20.w_1" shape = [128] dtype = "float32" min_val = float("-0.417545") @@ -2659,6 +2942,7 @@ class Program_weight_tensor_parameter_258: class Program_weight_tensor_parameter_259: name = "parameter_259" + original_name = "conv2d_20.w_0" shape = [128, 512, 1, 1] dtype = "float32" min_val = float("-0.346156") @@ -2670,6 +2954,7 @@ class Program_weight_tensor_parameter_259: class Program_weight_tensor_parameter_260: name = "parameter_260" + original_name = "batch_norm2d_19.b_0" shape = [512] dtype = "float32" min_val = float("-0.455431") @@ -2681,6 +2966,7 @@ class Program_weight_tensor_parameter_260: class Program_weight_tensor_parameter_261: name = "parameter_261" + original_name = "batch_norm2d_19.w_0" shape = [512] dtype = "float32" min_val = float("-0.0672166") @@ -2692,6 +2978,7 @@ class Program_weight_tensor_parameter_261: class Program_weight_tensor_parameter_262: name = "parameter_262" + original_name = "batch_norm2d_19.w_2" shape = [512] dtype = "float32" min_val = float("3.4359e-36") @@ -2703,6 +2990,7 @@ class Program_weight_tensor_parameter_262: class Program_weight_tensor_parameter_263: name = "parameter_263" + original_name = "batch_norm2d_19.w_1" shape = [512] dtype = "float32" min_val = float("-0.205145") @@ -2714,6 +3002,7 @@ class Program_weight_tensor_parameter_263: class Program_weight_tensor_parameter_264: name = "parameter_264" + original_name = "conv2d_19.w_0" shape = [512, 128, 1, 1] dtype = "float32" min_val = float("-0.411486") @@ -2725,6 +3014,7 @@ class Program_weight_tensor_parameter_264: class Program_weight_tensor_parameter_265: name = "parameter_265" + original_name = "batch_norm2d_18.b_0" shape = [128] dtype = "float32" min_val = float("-0.243617") @@ -2736,6 +3026,7 @@ class Program_weight_tensor_parameter_265: class Program_weight_tensor_parameter_266: name = "parameter_266" + original_name = "batch_norm2d_18.w_0" shape = [128] dtype = "float32" min_val = float("0.142974") @@ -2747,6 +3038,7 @@ class Program_weight_tensor_parameter_266: class Program_weight_tensor_parameter_267: name = "parameter_267" + original_name = "batch_norm2d_18.w_2" shape = [128] dtype = "float32" min_val = float("0.0210301") @@ -2758,6 +3050,7 @@ class Program_weight_tensor_parameter_267: class Program_weight_tensor_parameter_268: name = "parameter_268" + original_name = "batch_norm2d_18.w_1" shape = [128] dtype = "float32" min_val = float("-0.958791") @@ -2769,6 +3062,7 @@ class Program_weight_tensor_parameter_268: class Program_weight_tensor_parameter_269: name = "parameter_269" + original_name = "conv2d_18.w_0" shape = [128, 128, 3, 3] dtype = "float32" min_val = float("-0.356737") @@ -2780,6 +3074,7 @@ class Program_weight_tensor_parameter_269: class Program_weight_tensor_parameter_270: name = "parameter_270" + original_name = "batch_norm2d_17.b_0" shape = [128] dtype = "float32" min_val = float("-0.26234") @@ -2791,6 +3086,7 @@ class Program_weight_tensor_parameter_270: class Program_weight_tensor_parameter_271: name = "parameter_271" + original_name = "batch_norm2d_17.w_0" shape = [128] dtype = "float32" min_val = float("0.135421") @@ -2802,6 +3098,7 @@ class Program_weight_tensor_parameter_271: class Program_weight_tensor_parameter_272: name = "parameter_272" + original_name = "batch_norm2d_17.w_2" shape = [128] dtype = "float32" min_val = float("0.0169231") @@ -2813,6 +3110,7 @@ class Program_weight_tensor_parameter_272: class Program_weight_tensor_parameter_273: name = "parameter_273" + original_name = "batch_norm2d_17.w_1" shape = [128] dtype = "float32" min_val = float("-1.0924") @@ -2824,6 +3122,7 @@ class Program_weight_tensor_parameter_273: class Program_weight_tensor_parameter_274: name = "parameter_274" + original_name = "conv2d_17.w_0" shape = [128, 512, 1, 1] dtype = "float32" min_val = float("-0.562909") @@ -2835,6 +3134,7 @@ class Program_weight_tensor_parameter_274: class Program_weight_tensor_parameter_275: name = "parameter_275" + original_name = "batch_norm2d_16.b_0" shape = [512] dtype = "float32" min_val = float("-0.370036") @@ -2846,6 +3146,7 @@ class Program_weight_tensor_parameter_275: class Program_weight_tensor_parameter_276: name = "parameter_276" + original_name = "batch_norm2d_16.w_0" shape = [512] dtype = "float32" min_val = float("-7.30756e-19") @@ -2857,6 +3158,7 @@ class Program_weight_tensor_parameter_276: class Program_weight_tensor_parameter_277: name = "parameter_277" + original_name = "batch_norm2d_16.w_2" shape = [512] dtype = "float32" min_val = float("5.60519e-45") @@ -2868,6 +3170,7 @@ class Program_weight_tensor_parameter_277: class Program_weight_tensor_parameter_278: name = "parameter_278" + original_name = "batch_norm2d_16.w_1" shape = [512] dtype = "float32" min_val = float("-0.667008") @@ -2879,6 +3182,7 @@ class Program_weight_tensor_parameter_278: class Program_weight_tensor_parameter_279: name = "parameter_279" + original_name = "conv2d_16.w_0" shape = [512, 256, 1, 1] dtype = "float32" min_val = float("-0.723694") @@ -2890,6 +3194,7 @@ class Program_weight_tensor_parameter_279: class Program_weight_tensor_parameter_280: name = "parameter_280" + original_name = "batch_norm2d_15.b_0" shape = [512] dtype = "float32" min_val = float("-0.370036") @@ -2901,6 +3206,7 @@ class Program_weight_tensor_parameter_280: class Program_weight_tensor_parameter_281: name = "parameter_281" + original_name = "batch_norm2d_15.w_0" shape = [512] dtype = "float32" min_val = float("-0.0332384") @@ -2912,6 +3218,7 @@ class Program_weight_tensor_parameter_281: class Program_weight_tensor_parameter_282: name = "parameter_282" + original_name = "batch_norm2d_15.w_2" shape = [512] dtype = "float32" min_val = float("5.60519e-45") @@ -2923,6 +3230,7 @@ class Program_weight_tensor_parameter_282: class Program_weight_tensor_parameter_283: name = "parameter_283" + original_name = "batch_norm2d_15.w_1" shape = [512] dtype = "float32" min_val = float("-0.216692") @@ -2934,6 +3242,7 @@ class Program_weight_tensor_parameter_283: class Program_weight_tensor_parameter_284: name = "parameter_284" + original_name = "conv2d_15.w_0" shape = [512, 128, 1, 1] dtype = "float32" min_val = float("-0.329659") @@ -2945,6 +3254,7 @@ class Program_weight_tensor_parameter_284: class Program_weight_tensor_parameter_285: name = "parameter_285" + original_name = "batch_norm2d_14.b_0" shape = [128] dtype = "float32" min_val = float("-0.382095") @@ -2956,6 +3266,7 @@ class Program_weight_tensor_parameter_285: class Program_weight_tensor_parameter_286: name = "parameter_286" + original_name = "batch_norm2d_14.w_0" shape = [128] dtype = "float32" min_val = float("0.128756") @@ -2967,6 +3278,7 @@ class Program_weight_tensor_parameter_286: class Program_weight_tensor_parameter_287: name = "parameter_287" + original_name = "batch_norm2d_14.w_2" shape = [128] dtype = "float32" min_val = float("0.00985592") @@ -2978,6 +3290,7 @@ class Program_weight_tensor_parameter_287: class Program_weight_tensor_parameter_288: name = "parameter_288" + original_name = "batch_norm2d_14.w_1" shape = [128] dtype = "float32" min_val = float("-0.362586") @@ -2989,6 +3302,7 @@ class Program_weight_tensor_parameter_288: class Program_weight_tensor_parameter_289: name = "parameter_289" + original_name = "conv2d_14.w_0" shape = [128, 128, 3, 3] dtype = "float32" min_val = float("-0.231993") @@ -3000,6 +3314,7 @@ class Program_weight_tensor_parameter_289: class Program_weight_tensor_parameter_290: name = "parameter_290" + original_name = "batch_norm2d_13.b_0" shape = [128] dtype = "float32" min_val = float("-0.47523") @@ -3011,6 +3326,7 @@ class Program_weight_tensor_parameter_290: class Program_weight_tensor_parameter_291: name = "parameter_291" + original_name = "batch_norm2d_13.w_0" shape = [128] dtype = "float32" min_val = float("0.164114") @@ -3022,6 +3338,7 @@ class Program_weight_tensor_parameter_291: class Program_weight_tensor_parameter_292: name = "parameter_292" + original_name = "batch_norm2d_13.w_2" shape = [128] dtype = "float32" min_val = float("0.0232963") @@ -3033,6 +3350,7 @@ class Program_weight_tensor_parameter_292: class Program_weight_tensor_parameter_293: name = "parameter_293" + original_name = "batch_norm2d_13.w_1" shape = [128] dtype = "float32" min_val = float("-0.553745") @@ -3044,6 +3362,7 @@ class Program_weight_tensor_parameter_293: class Program_weight_tensor_parameter_294: name = "parameter_294" + original_name = "conv2d_13.w_0" shape = [128, 256, 1, 1] dtype = "float32" min_val = float("-0.434561") @@ -3055,6 +3374,7 @@ class Program_weight_tensor_parameter_294: class Program_weight_tensor_parameter_295: name = "parameter_295" + original_name = "batch_norm2d_12.b_0" shape = [256] dtype = "float32" min_val = float("-0.304404") @@ -3066,6 +3386,7 @@ class Program_weight_tensor_parameter_295: class Program_weight_tensor_parameter_296: name = "parameter_296" + original_name = "batch_norm2d_12.w_0" shape = [256] dtype = "float32" min_val = float("-0.118238") @@ -3077,6 +3398,7 @@ class Program_weight_tensor_parameter_296: class Program_weight_tensor_parameter_297: name = "parameter_297" + original_name = "batch_norm2d_12.w_2" shape = [256] dtype = "float32" min_val = float("0.000216192") @@ -3088,6 +3410,7 @@ class Program_weight_tensor_parameter_297: class Program_weight_tensor_parameter_298: name = "parameter_298" + original_name = "batch_norm2d_12.w_1" shape = [256] dtype = "float32" min_val = float("-0.258") @@ -3099,6 +3422,7 @@ class Program_weight_tensor_parameter_298: class Program_weight_tensor_parameter_299: name = "parameter_299" + original_name = "conv2d_12.w_0" shape = [256, 64, 1, 1] dtype = "float32" min_val = float("-0.322187") @@ -3110,6 +3434,7 @@ class Program_weight_tensor_parameter_299: class Program_weight_tensor_parameter_300: name = "parameter_300" + original_name = "batch_norm2d_11.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -3119,6 +3444,7 @@ class Program_weight_tensor_parameter_300: class Program_weight_tensor_parameter_301: name = "parameter_301" + original_name = "batch_norm2d_11.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -3128,6 +3454,7 @@ class Program_weight_tensor_parameter_301: class Program_weight_tensor_parameter_302: name = "parameter_302" + original_name = "batch_norm2d_11.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -3137,6 +3464,7 @@ class Program_weight_tensor_parameter_302: class Program_weight_tensor_parameter_303: name = "parameter_303" + original_name = "batch_norm2d_11.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -3146,6 +3474,7 @@ class Program_weight_tensor_parameter_303: class Program_weight_tensor_parameter_304: name = "parameter_304" + original_name = "conv2d_11.w_0" shape = [64, 64, 3, 3] dtype = "float32" min_val = float("-0.393733") @@ -3157,6 +3486,7 @@ class Program_weight_tensor_parameter_304: class Program_weight_tensor_parameter_305: name = "parameter_305" + original_name = "batch_norm2d_10.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -3166,6 +3496,7 @@ class Program_weight_tensor_parameter_305: class Program_weight_tensor_parameter_306: name = "parameter_306" + original_name = "batch_norm2d_10.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -3175,6 +3506,7 @@ class Program_weight_tensor_parameter_306: class Program_weight_tensor_parameter_307: name = "parameter_307" + original_name = "batch_norm2d_10.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -3184,6 +3516,7 @@ class Program_weight_tensor_parameter_307: class Program_weight_tensor_parameter_308: name = "parameter_308" + original_name = "batch_norm2d_10.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -3193,6 +3526,7 @@ class Program_weight_tensor_parameter_308: class Program_weight_tensor_parameter_309: name = "parameter_309" + original_name = "conv2d_10.w_0" shape = [64, 256, 1, 1] dtype = "float32" min_val = float("-0.364472") @@ -3204,6 +3538,7 @@ class Program_weight_tensor_parameter_309: class Program_weight_tensor_parameter_310: name = "parameter_310" + original_name = "batch_norm2d_9.b_0" shape = [256] dtype = "float32" min_val = float("-0.462393") @@ -3215,6 +3550,7 @@ class Program_weight_tensor_parameter_310: class Program_weight_tensor_parameter_311: name = "parameter_311" + original_name = "batch_norm2d_9.w_0" shape = [256] dtype = "float32" min_val = float("-0.0375522") @@ -3226,6 +3562,7 @@ class Program_weight_tensor_parameter_311: class Program_weight_tensor_parameter_312: name = "parameter_312" + original_name = "batch_norm2d_9.w_2" shape = [256] dtype = "float32" min_val = float("0.000317126") @@ -3237,6 +3574,7 @@ class Program_weight_tensor_parameter_312: class Program_weight_tensor_parameter_313: name = "parameter_313" + original_name = "batch_norm2d_9.w_1" shape = [256] dtype = "float32" min_val = float("-0.270805") @@ -3248,6 +3586,7 @@ class Program_weight_tensor_parameter_313: class Program_weight_tensor_parameter_314: name = "parameter_314" + original_name = "conv2d_9.w_0" shape = [256, 64, 1, 1] dtype = "float32" min_val = float("-0.425601") @@ -3259,6 +3598,7 @@ class Program_weight_tensor_parameter_314: class Program_weight_tensor_parameter_315: name = "parameter_315" + original_name = "batch_norm2d_8.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -3268,6 +3608,7 @@ class Program_weight_tensor_parameter_315: class Program_weight_tensor_parameter_316: name = "parameter_316" + original_name = "batch_norm2d_8.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -3277,6 +3618,7 @@ class Program_weight_tensor_parameter_316: class Program_weight_tensor_parameter_317: name = "parameter_317" + original_name = "batch_norm2d_8.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -3286,6 +3628,7 @@ class Program_weight_tensor_parameter_317: class Program_weight_tensor_parameter_318: name = "parameter_318" + original_name = "batch_norm2d_8.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -3295,6 +3638,7 @@ class Program_weight_tensor_parameter_318: class Program_weight_tensor_parameter_319: name = "parameter_319" + original_name = "conv2d_8.w_0" shape = [64, 64, 3, 3] dtype = "float32" min_val = float("-0.365322") @@ -3306,6 +3650,7 @@ class Program_weight_tensor_parameter_319: class Program_weight_tensor_parameter_320: name = "parameter_320" + original_name = "batch_norm2d_7.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -3315,6 +3660,7 @@ class Program_weight_tensor_parameter_320: class Program_weight_tensor_parameter_321: name = "parameter_321" + original_name = "batch_norm2d_7.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -3324,6 +3670,7 @@ class Program_weight_tensor_parameter_321: class Program_weight_tensor_parameter_322: name = "parameter_322" + original_name = "batch_norm2d_7.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -3333,6 +3680,7 @@ class Program_weight_tensor_parameter_322: class Program_weight_tensor_parameter_323: name = "parameter_323" + original_name = "batch_norm2d_7.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -3342,6 +3690,7 @@ class Program_weight_tensor_parameter_323: class Program_weight_tensor_parameter_324: name = "parameter_324" + original_name = "conv2d_7.w_0" shape = [64, 256, 1, 1] dtype = "float32" min_val = float("-0.369789") @@ -3353,6 +3702,7 @@ class Program_weight_tensor_parameter_324: class Program_weight_tensor_parameter_325: name = "parameter_325" + original_name = "batch_norm2d_6.b_0" shape = [256] dtype = "float32" min_val = float("-0.254622") @@ -3364,6 +3714,7 @@ class Program_weight_tensor_parameter_325: class Program_weight_tensor_parameter_326: name = "parameter_326" + original_name = "batch_norm2d_6.w_0" shape = [256] dtype = "float32" min_val = float("-0.0321339") @@ -3375,6 +3726,7 @@ class Program_weight_tensor_parameter_326: class Program_weight_tensor_parameter_327: name = "parameter_327" + original_name = "batch_norm2d_6.w_2" shape = [256] dtype = "float32" min_val = float("5.60519e-45") @@ -3386,6 +3738,7 @@ class Program_weight_tensor_parameter_327: class Program_weight_tensor_parameter_328: name = "parameter_328" + original_name = "batch_norm2d_6.w_1" shape = [256] dtype = "float32" min_val = float("-1.35275") @@ -3397,6 +3750,7 @@ class Program_weight_tensor_parameter_328: class Program_weight_tensor_parameter_329: name = "parameter_329" + original_name = "conv2d_6.w_0" shape = [256, 64, 1, 1] dtype = "float32" min_val = float("-0.983408") @@ -3408,6 +3762,7 @@ class Program_weight_tensor_parameter_329: class Program_weight_tensor_parameter_330: name = "parameter_330" + original_name = "batch_norm2d_5.b_0" shape = [256] dtype = "float32" min_val = float("-0.254622") @@ -3419,6 +3774,7 @@ class Program_weight_tensor_parameter_330: class Program_weight_tensor_parameter_331: name = "parameter_331" + original_name = "batch_norm2d_5.w_0" shape = [256] dtype = "float32" min_val = float("-0.110997") @@ -3430,6 +3786,7 @@ class Program_weight_tensor_parameter_331: class Program_weight_tensor_parameter_332: name = "parameter_332" + original_name = "batch_norm2d_5.w_2" shape = [256] dtype = "float32" min_val = float("5.60519e-45") @@ -3441,6 +3798,7 @@ class Program_weight_tensor_parameter_332: class Program_weight_tensor_parameter_333: name = "parameter_333" + original_name = "batch_norm2d_5.w_1" shape = [256] dtype = "float32" min_val = float("-0.426076") @@ -3452,6 +3810,7 @@ class Program_weight_tensor_parameter_333: class Program_weight_tensor_parameter_334: name = "parameter_334" + original_name = "conv2d_5.w_0" shape = [256, 64, 1, 1] dtype = "float32" min_val = float("-0.493531") @@ -3463,6 +3822,7 @@ class Program_weight_tensor_parameter_334: class Program_weight_tensor_parameter_335: name = "parameter_335" + original_name = "batch_norm2d_4.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -3472,6 +3832,7 @@ class Program_weight_tensor_parameter_335: class Program_weight_tensor_parameter_336: name = "parameter_336" + original_name = "batch_norm2d_4.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -3481,6 +3842,7 @@ class Program_weight_tensor_parameter_336: class Program_weight_tensor_parameter_337: name = "parameter_337" + original_name = "batch_norm2d_4.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -3490,6 +3852,7 @@ class Program_weight_tensor_parameter_337: class Program_weight_tensor_parameter_338: name = "parameter_338" + original_name = "batch_norm2d_4.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -3499,6 +3862,7 @@ class Program_weight_tensor_parameter_338: class Program_weight_tensor_parameter_339: name = "parameter_339" + original_name = "conv2d_4.w_0" shape = [64, 64, 3, 3] dtype = "float32" min_val = float("-0.594026") @@ -3510,6 +3874,7 @@ class Program_weight_tensor_parameter_339: class Program_weight_tensor_parameter_340: name = "parameter_340" + original_name = "batch_norm2d_3.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -3519,6 +3884,7 @@ class Program_weight_tensor_parameter_340: class Program_weight_tensor_parameter_341: name = "parameter_341" + original_name = "batch_norm2d_3.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -3528,6 +3894,7 @@ class Program_weight_tensor_parameter_341: class Program_weight_tensor_parameter_342: name = "parameter_342" + original_name = "batch_norm2d_3.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -3537,6 +3904,7 @@ class Program_weight_tensor_parameter_342: class Program_weight_tensor_parameter_343: name = "parameter_343" + original_name = "batch_norm2d_3.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -3546,6 +3914,7 @@ class Program_weight_tensor_parameter_343: class Program_weight_tensor_parameter_344: name = "parameter_344" + original_name = "conv2d_3.w_0" shape = [64, 64, 1, 1] dtype = "float32" min_val = float("-0.911477") @@ -3557,6 +3926,7 @@ class Program_weight_tensor_parameter_344: class Program_weight_tensor_parameter_345: name = "parameter_345" + original_name = "batch_norm2d_2.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -3566,6 +3936,7 @@ class Program_weight_tensor_parameter_345: class Program_weight_tensor_parameter_346: name = "parameter_346" + original_name = "batch_norm2d_2.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -3575,6 +3946,7 @@ class Program_weight_tensor_parameter_346: class Program_weight_tensor_parameter_347: name = "parameter_347" + original_name = "batch_norm2d_2.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -3584,6 +3956,7 @@ class Program_weight_tensor_parameter_347: class Program_weight_tensor_parameter_348: name = "parameter_348" + original_name = "batch_norm2d_2.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -3593,6 +3966,7 @@ class Program_weight_tensor_parameter_348: class Program_weight_tensor_parameter_349: name = "parameter_349" + original_name = "conv2d_2.w_0" shape = [64, 32, 3, 3] dtype = "float32" min_val = float("-0.825257") @@ -3604,6 +3978,7 @@ class Program_weight_tensor_parameter_349: class Program_weight_tensor_parameter_350: name = "parameter_350" + original_name = "batch_norm2d_1.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -3613,6 +3988,7 @@ class Program_weight_tensor_parameter_350: class Program_weight_tensor_parameter_351: name = "parameter_351" + original_name = "batch_norm2d_1.w_0" shape = [32] dtype = "float32" min_val = float("0") @@ -3622,6 +3998,7 @@ class Program_weight_tensor_parameter_351: class Program_weight_tensor_parameter_352: name = "parameter_352" + original_name = "batch_norm2d_1.w_2" shape = [32] dtype = "float32" min_val = float("0") @@ -3631,6 +4008,7 @@ class Program_weight_tensor_parameter_352: class Program_weight_tensor_parameter_353: name = "parameter_353" + original_name = "batch_norm2d_1.w_1" shape = [32] dtype = "float32" min_val = float("0") @@ -3640,6 +4018,7 @@ class Program_weight_tensor_parameter_353: class Program_weight_tensor_parameter_354: name = "parameter_354" + original_name = "conv2d_1.w_0" shape = [32, 32, 3, 3] dtype = "float32" min_val = float("-1.31725") @@ -3651,6 +4030,7 @@ class Program_weight_tensor_parameter_354: class Program_weight_tensor_parameter_355: name = "parameter_355" + original_name = "batch_norm2d_0.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -3660,6 +4040,7 @@ class Program_weight_tensor_parameter_355: class Program_weight_tensor_parameter_356: name = "parameter_356" + original_name = "batch_norm2d_0.w_0" shape = [32] dtype = "float32" min_val = float("0") @@ -3669,6 +4050,7 @@ class Program_weight_tensor_parameter_356: class Program_weight_tensor_parameter_357: name = "parameter_357" + original_name = "batch_norm2d_0.w_2" shape = [32] dtype = "float32" min_val = float("0") @@ -3678,6 +4060,7 @@ class Program_weight_tensor_parameter_357: class Program_weight_tensor_parameter_358: name = "parameter_358" + original_name = "batch_norm2d_0.w_1" shape = [32] dtype = "float32" min_val = float("0") @@ -3687,6 +4070,7 @@ class Program_weight_tensor_parameter_358: class Program_weight_tensor_parameter_359: name = "parameter_359" + original_name = "conv2d_0.w_0" shape = [32, 3, 3, 3] dtype = "float32" min_val = float("-1.4988") From 746cdf15ffd15ec93422a2fa55dd2800680e0bcf Mon Sep 17 00:00:00 2001 From: Liu Yiqun Date: Tue, 20 Jan 2026 17:14:24 +0800 Subject: [PATCH 05/10] Add original_name for serveral PaddleCV samples. --- .../subgraph_0/input_meta.py | 5 + .../subgraph_0/weight_meta.py | 171 + .../subgraph_2/input_meta.py | 4 + .../subgraph_2/weight_meta.py | 171 + .../Deeplabv3-R101/subgraph_1/input_meta.py | 1 + .../Deeplabv3-R101/subgraph_1/weight_meta.py | 579 ++++ .../Deeplabv3-R50/subgraph_0/input_meta.py | 1 + .../Deeplabv3-R50/subgraph_0/weight_meta.py | 312 ++ .../FasterNet-L/subgraph_2/input_meta.py | 1 + .../FasterNet-L/subgraph_2/weight_meta.py | 219 ++ .../FasterNet-T0/subgraph_1/input_meta.py | 1 + .../FasterNet-T0/subgraph_1/weight_meta.py | 114 + .../FasterNet-T1/subgraph_1/input_meta.py | 1 + .../FasterNet-T1/subgraph_1/weight_meta.py | 114 + .../FasterNet-T1/subgraph_2/input_meta.py | 1 + .../FasterNet-T1/subgraph_2/weight_meta.py | 114 + .../subgraph_14/input_meta.py | 5 + .../subgraph_14/weight_meta.py | 12 + .../subgraph_17/input_meta.py | 5 + .../subgraph_17/weight_meta.py | 12 + .../subgraph_9/input_meta.py | 5 + .../subgraph_9/weight_meta.py | 24 + .../Mask-RT-DETR-H/subgraph_5/input_meta.py | 19 + .../Mask-RT-DETR-H/subgraph_5/weight_meta.py | 159 + .../Mask-RT-DETR-L/subgraph_11/input_meta.py | 23 + .../Mask-RT-DETR-L/subgraph_11/weight_meta.py | 159 + .../Mask-RT-DETR-L/subgraph_12/input_meta.py | 21 + .../Mask-RT-DETR-L/subgraph_12/weight_meta.py | 159 + .../Mask-RT-DETR-L/subgraph_18/input_meta.py | 23 + .../Mask-RT-DETR-L/subgraph_18/weight_meta.py | 159 + .../Mask-RT-DETR-M/subgraph_3/input_meta.py | 19 + .../Mask-RT-DETR-M/subgraph_3/weight_meta.py | 159 + .../Mask-RT-DETR-S/subgraph_10/input_meta.py | 19 + .../Mask-RT-DETR-S/subgraph_10/weight_meta.py | 159 + .../Mask-RT-DETR-S/subgraph_9/input_meta.py | 21 + .../Mask-RT-DETR-S/subgraph_9/weight_meta.py | 159 + .../MaskFormer_small/subgraph_2/input_meta.py | 50 + .../subgraph_2/weight_meta.py | 500 +++ .../MaskFormer_tiny/subgraph_0/input_meta.py | 26 + .../MaskFormer_tiny/subgraph_0/weight_meta.py | 356 +++ .../MobileFaceNet/subgraph_0/input_meta.py | 3 + .../MobileFaceNet/subgraph_0/weight_meta.py | 283 ++ .../MobileFaceNet/subgraph_2/input_meta.py | 3 + .../MobileFaceNet/subgraph_2/weight_meta.py | 283 ++ .../PaddleX/MobileNetV2_x1_0/input_meta.py | 1 + .../PaddleX/MobileNetV2_x1_0/weight_meta.py | 267 ++ .../PaddleX/MobileNetV2_x2_0/input_meta.py | 1 + .../PaddleX/MobileNetV2_x2_0/weight_meta.py | 267 ++ .../subgraph_1/input_meta.py | 1 + .../subgraph_1/weight_meta.py | 270 ++ .../subgraph_0/input_meta.py | 1 + .../subgraph_0/weight_meta.py | 270 ++ .../subgraph_0/input_meta.py | 1 + .../subgraph_0/weight_meta.py | 270 ++ .../subgraph_1/input_meta.py | 1 + .../subgraph_1/weight_meta.py | 270 ++ .../subgraph_1/input_meta.py | 1 + .../subgraph_1/weight_meta.py | 270 ++ .../subgraph_0/input_meta.py | 1 + .../subgraph_0/weight_meta.py | 214 ++ .../subgraph_0/input_meta.py | 1 + .../subgraph_0/weight_meta.py | 214 ++ .../subgraph_0/input_meta.py | 1 + .../subgraph_0/weight_meta.py | 214 ++ .../subgraph_1/input_meta.py | 1 + .../subgraph_1/weight_meta.py | 214 ++ .../subgraph_0/input_meta.py | 1 + .../subgraph_0/weight_meta.py | 214 ++ .../subgraph_0/input_meta.py | 1 + .../subgraph_0/weight_meta.py | 512 +++ .../subgraph_2/input_meta.py | 1 + .../subgraph_2/weight_meta.py | 387 +++ .../subgraph_1/input_meta.py | 36 + .../subgraph_1/weight_meta.py | 631 ++++ .../subgraph_2/input_meta.py | 31 + .../subgraph_2/weight_meta.py | 506 +++ .../PP-HGNetV2-B0_ML/subgraph_1/input_meta.py | 64 + .../subgraph_1/weight_meta.py | 233 ++ .../PP-HGNetV2-B0_ML/subgraph_2/input_meta.py | 64 + .../subgraph_2/weight_meta.py | 233 ++ .../PP-HGNetV2-B4_ML/subgraph_0/input_meta.py | 2 + .../subgraph_0/weight_meta.py | 423 +++ .../PP-HGNetV2-B6_ML/subgraph_1/input_meta.py | 2 + .../subgraph_1/weight_meta.py | 893 ++++++ .../PP-LCNet_x1_0_ML/subgraph_0/input_meta.py | 2 + .../subgraph_0/weight_meta.py | 166 + .../PP-LiteSeg-B/subgraph_5/input_meta.py | 6 + .../PP-LiteSeg-B/subgraph_5/weight_meta.py | 84 + .../PP-OCRv4_mobile_seal_det/input_meta.py | 105 + .../PP-OCRv4_mobile_seal_det/weight_meta.py | 793 +++++ .../subgraph_0/input_meta.py | 1 + .../subgraph_0/weight_meta.py | 367 +++ .../subgraph_1/input_meta.py | 1 + .../subgraph_1/weight_meta.py | 354 +++ .../subgraph_0/input_meta.py | 1 + .../subgraph_0/weight_meta.py | 220 ++ .../subgraph_1/input_meta.py | 4 + .../subgraph_1/weight_meta.py | 83 + .../subgraph_4/input_meta.py | 4 + .../subgraph_4/weight_meta.py | 83 + .../PP-OCRv4_server_seal_det/input_meta.py | 1 + .../PP-OCRv4_server_seal_det/weight_meta.py | 367 +++ .../PP-ShiTuV2_det/subgraph_1/input_meta.py | 4 + .../PP-ShiTuV2_det/subgraph_1/weight_meta.py | 2790 ++++++++++++----- .../PP-ShiTuV2_det/subgraph_5/input_meta.py | 1 + .../PP-ShiTuV2_det/subgraph_5/weight_meta.py | 486 +++ .../PP-ShiTuV2_rec/subgraph_1/input_meta.py | 1 + .../PP-ShiTuV2_rec/subgraph_1/weight_meta.py | 271 ++ .../PP-YOLOE-R-L/subgraph_3/input_meta.py | 34 + .../PP-YOLOE-R-L/subgraph_3/weight_meta.py | 698 +++++ .../PP-YOLOE_plus-L/subgraph_4/input_meta.py | 19 + .../PP-YOLOE_plus-L/subgraph_4/weight_meta.py | 698 +++++ .../PP-YOLOE_plus-M/subgraph_1/input_meta.py | 1 - .../PP-YOLOE_plus-S/subgraph_2/input_meta.py | 7 + .../PP-YOLOE_plus-S/subgraph_2/weight_meta.py | 368 +++ .../PP-YOLOE_plus-X/subgraph_1/input_meta.py | 25 + .../PP-YOLOE_plus-X/subgraph_1/weight_meta.py | 863 +++++ .../subgraph_2/input_meta.py | 19 + .../subgraph_2/weight_meta.py | 373 +++ .../subgraph_6/input_meta.py | 27 + .../subgraph_6/weight_meta.py | 738 +++++ .../subgraph_0/input_meta.py | 11 + .../subgraph_0/weight_meta.py | 215 ++ .../subgraph_3/input_meta.py | 1 - .../subgraph_0/input_meta.py | 11 + .../subgraph_0/weight_meta.py | 365 +++ .../subgraph_6/input_meta.py | 27 + .../subgraph_6/weight_meta.py | 738 +++++ .../subgraph_0/input_meta.py | 2 + .../subgraph_0/weight_meta.py | 522 +++ .../PicoDet-S/subgraph_12/input_meta.py | 2 + .../PicoDet-S/subgraph_12/weight_meta.py | 442 +++ .../PicoDet-S/subgraph_13/input_meta.py | 2 + .../PicoDet-S/subgraph_13/weight_meta.py | 442 +++ .../PicoDet-XS/subgraph_4/input_meta.py | 2 + .../PicoDet-XS/subgraph_4/weight_meta.py | 442 +++ .../PicoDet-XS/subgraph_6/input_meta.py | 2 + .../PicoDet-XS/subgraph_6/weight_meta.py | 442 +++ paddle_samples/PaddleX/ResNet34/input_meta.py | 1 + .../PaddleX/ResNet34/weight_meta.py | 187 ++ .../PaddleX/SOLOv2/subgraph_0/input_meta.py | 1 + .../PaddleX/SOLOv2/subgraph_0/weight_meta.py | 333 ++ .../PaddleX/SOLOv2/subgraph_2/input_meta.py | 1 + .../PaddleX/SOLOv2/subgraph_2/weight_meta.py | 333 ++ .../SegFormer-B1/subgraph_0/input_meta.py | 1 + .../SegFormer-B1/subgraph_0/weight_meta.py | 191 ++ .../SegFormer-B2/subgraph_1/input_meta.py | 1 + .../SegFormer-B2/subgraph_1/weight_meta.py | 347 ++ .../SegFormer-B3/subgraph_0/input_meta.py | 1 + .../SegFormer-B3/subgraph_0/weight_meta.py | 587 ++++ .../SegFormer-B4/subgraph_0/input_meta.py | 1 + .../SegFormer-B4/subgraph_0/weight_meta.py | 847 +++++ .../SegFormer-B5/subgraph_1/input_meta.py | 1 + .../SegFormer-B5/subgraph_1/weight_meta.py | 1067 +++++++ .../subgraph_0/input_meta.py | 49 + .../subgraph_0/weight_meta.py | 305 ++ .../subgraph_0/input_meta.py | 49 + .../subgraph_0/weight_meta.py | 305 ++ .../subgraph_1/input_meta.py | 49 + .../subgraph_1/weight_meta.py | 305 ++ .../subgraph_1/input_meta.py | 49 + .../subgraph_1/weight_meta.py | 305 ++ .../subgraph_0/input_meta.py | 49 + .../subgraph_0/weight_meta.py | 305 ++ .../subgraph_2/input_meta.py | 49 + .../subgraph_2/weight_meta.py | 305 ++ .../subgraph_0/input_meta.py | 49 + .../subgraph_0/weight_meta.py | 305 ++ .../subgraph_2/input_meta.py | 49 + .../subgraph_2/weight_meta.py | 305 ++ .../subgraph_1/input_meta.py | 25 + .../subgraph_1/weight_meta.py | 161 + .../subgraph_2/input_meta.py | 25 + .../subgraph_2/weight_meta.py | 161 + .../ch_SVTRv2_rec/subgraph_10/input_meta.py | 1 + .../ch_SVTRv2_rec/subgraph_10/weight_meta.py | 53 + .../ch_SVTRv2_rec/subgraph_11/input_meta.py | 4 + .../ch_SVTRv2_rec/subgraph_11/weight_meta.py | 72 + .../ch_SVTRv2_rec/subgraph_13/input_meta.py | 1 + .../ch_SVTRv2_rec/subgraph_13/weight_meta.py | 53 + .../ch_SVTRv2_rec/subgraph_3/input_meta.py | 1 + .../ch_SVTRv2_rec/subgraph_3/weight_meta.py | 220 ++ .../ch_SVTRv2_rec/subgraph_7/input_meta.py | 7 + .../ch_SVTRv2_rec/subgraph_7/weight_meta.py | 72 + 184 files changed, 32170 insertions(+), 770 deletions(-) diff --git a/paddle_samples/PaddleX/CLIP_vit_base_patch16_448_ML/subgraph_0/input_meta.py b/paddle_samples/PaddleX/CLIP_vit_base_patch16_448_ML/subgraph_0/input_meta.py index a1e08e7e9..cdb8d4192 100644 --- a/paddle_samples/PaddleX/CLIP_vit_base_patch16_448_ML/subgraph_0/input_meta.py +++ b/paddle_samples/PaddleX/CLIP_vit_base_patch16_448_ML/subgraph_0/input_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_data_0: name = "data_0" + original_name = "param_4" shape = [1, 1, 768] dtype = "float32" min_val = float("-1.57892") @@ -11,6 +12,7 @@ class Program_weight_tensor_data_0: class Program_weight_tensor_data_1: name = "data_1" + original_name = "param_5" shape = [1, 197, 768] dtype = "float32" min_val = float("-0.226016") @@ -22,6 +24,7 @@ class Program_weight_tensor_data_1: class Program_weight_tensor_data_2: name = "data_2" + original_name = "param_6" shape = [33, 768] dtype = "float32" min_val = float("-0.0865287") @@ -33,6 +36,7 @@ class Program_weight_tensor_data_2: class Program_weight_tensor_data_3: name = "data_3" + original_name = "args_3" shape = [] dtype = "int64" data = [7] @@ -40,6 +44,7 @@ class Program_weight_tensor_data_3: class Program_weight_tensor_data_4: name = "data_4" + original_name = "var_336" shape = [7, 3, 448, 448] dtype = "float32" min_val = float("-2.1179") diff --git a/paddle_samples/PaddleX/CLIP_vit_base_patch16_448_ML/subgraph_0/weight_meta.py b/paddle_samples/PaddleX/CLIP_vit_base_patch16_448_ML/subgraph_0/weight_meta.py index 289aa9e9d..4c1bc2c60 100644 --- a/paddle_samples/PaddleX/CLIP_vit_base_patch16_448_ML/subgraph_0/weight_meta.py +++ b/paddle_samples/PaddleX/CLIP_vit_base_patch16_448_ML/subgraph_0/weight_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_parameter_0: name = "parameter_0" + original_name = "conv2d_2.b_0" shape = [33] dtype = "float32" min_val = float("0") @@ -9,6 +10,7 @@ class Program_weight_tensor_parameter_0: class Program_weight_tensor_parameter_1: name = "parameter_1" + original_name = "conv2d_2.w_0" shape = [33, 768, 1, 1] dtype = "float32" min_val = float("-0.205661") @@ -20,6 +22,7 @@ class Program_weight_tensor_parameter_1: class Program_weight_tensor_parameter_2: name = "parameter_2" + original_name = "layer_norm_28.b_0" shape = [768] dtype = "float32" min_val = float("-0.00549064") @@ -31,6 +34,7 @@ class Program_weight_tensor_parameter_2: class Program_weight_tensor_parameter_3: name = "parameter_3" + original_name = "layer_norm_28.w_0" shape = [768] dtype = "float32" min_val = float("0.959046") @@ -42,6 +46,7 @@ class Program_weight_tensor_parameter_3: class Program_weight_tensor_parameter_4: name = "parameter_4" + original_name = "linear_58.b_0" shape = [768] dtype = "float32" min_val = float("-0.0224886") @@ -53,6 +58,7 @@ class Program_weight_tensor_parameter_4: class Program_weight_tensor_parameter_5: name = "parameter_5" + original_name = "linear_58.w_0" shape = [2048, 768] dtype = "float32" min_val = float("-0.0712805") @@ -64,6 +70,7 @@ class Program_weight_tensor_parameter_5: class Program_weight_tensor_parameter_6: name = "parameter_6" + original_name = "linear_57.b_0" shape = [2048] dtype = "float32" min_val = float("-0.00846867") @@ -75,6 +82,7 @@ class Program_weight_tensor_parameter_6: class Program_weight_tensor_parameter_7: name = "parameter_7" + original_name = "linear_57.w_0" shape = [768, 2048] dtype = "float32" min_val = float("-0.0563459") @@ -86,6 +94,7 @@ class Program_weight_tensor_parameter_7: class Program_weight_tensor_parameter_8: name = "parameter_8" + original_name = "layer_norm_27.b_0" shape = [768] dtype = "float32" min_val = float("-0.0479148") @@ -97,6 +106,7 @@ class Program_weight_tensor_parameter_8: class Program_weight_tensor_parameter_9: name = "parameter_9" + original_name = "layer_norm_27.w_0" shape = [768] dtype = "float32" min_val = float("0.989513") @@ -108,6 +118,7 @@ class Program_weight_tensor_parameter_9: class Program_weight_tensor_parameter_10: name = "parameter_10" + original_name = "linear_56.b_0" shape = [768] dtype = "float32" min_val = float("-0.0208882") @@ -119,6 +130,7 @@ class Program_weight_tensor_parameter_10: class Program_weight_tensor_parameter_11: name = "parameter_11" + original_name = "linear_56.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.120595") @@ -130,6 +142,7 @@ class Program_weight_tensor_parameter_11: class Program_weight_tensor_parameter_12: name = "parameter_12" + original_name = "linear_55.b_0" shape = [768] dtype = "float32" min_val = float("-0.00575747") @@ -141,6 +154,7 @@ class Program_weight_tensor_parameter_12: class Program_weight_tensor_parameter_13: name = "parameter_13" + original_name = "linear_55.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.0745472") @@ -152,6 +166,7 @@ class Program_weight_tensor_parameter_13: class Program_weight_tensor_parameter_14: name = "parameter_14" + original_name = "linear_54.b_0" shape = [768] dtype = "float32" data = None @@ -159,6 +174,7 @@ class Program_weight_tensor_parameter_14: class Program_weight_tensor_parameter_15: name = "parameter_15" + original_name = "linear_54.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.0624998") @@ -170,6 +186,7 @@ class Program_weight_tensor_parameter_15: class Program_weight_tensor_parameter_16: name = "parameter_16" + original_name = "linear_53.b_0" shape = [768] dtype = "float32" data = None @@ -177,6 +194,7 @@ class Program_weight_tensor_parameter_16: class Program_weight_tensor_parameter_17: name = "parameter_17" + original_name = "linear_53.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.0625") @@ -188,6 +206,7 @@ class Program_weight_tensor_parameter_17: class Program_weight_tensor_parameter_18: name = "parameter_18" + original_name = "layer_norm_26.b_0" shape = [768] dtype = "float32" min_val = float("-0.0329438") @@ -199,6 +218,7 @@ class Program_weight_tensor_parameter_18: class Program_weight_tensor_parameter_19: name = "parameter_19" + original_name = "layer_norm_26.w_0" shape = [768] dtype = "float32" min_val = float("0.995161") @@ -210,6 +230,7 @@ class Program_weight_tensor_parameter_19: class Program_weight_tensor_parameter_20: name = "parameter_20" + original_name = "conv2d_1.b_0" shape = [768] dtype = "float32" min_val = float("-0.008174") @@ -221,6 +242,7 @@ class Program_weight_tensor_parameter_20: class Program_weight_tensor_parameter_21: name = "parameter_21" + original_name = "conv2d_1.w_0" shape = [768, 768, 1, 1] dtype = "float32" min_val = float("-0.255873") @@ -232,6 +254,7 @@ class Program_weight_tensor_parameter_21: class Program_weight_tensor_parameter_22: name = "parameter_22" + original_name = "layer_norm_25.b_0" shape = [768] dtype = "float32" min_val = float("-1.66673") @@ -243,6 +266,7 @@ class Program_weight_tensor_parameter_22: class Program_weight_tensor_parameter_23: name = "parameter_23" + original_name = "layer_norm_25.w_0" shape = [768] dtype = "float32" min_val = float("-0.19603") @@ -254,6 +278,7 @@ class Program_weight_tensor_parameter_23: class Program_weight_tensor_parameter_24: name = "parameter_24" + original_name = "linear_47.b_0" shape = [768] dtype = "float32" min_val = float("-2.91586") @@ -265,6 +290,7 @@ class Program_weight_tensor_parameter_24: class Program_weight_tensor_parameter_25: name = "parameter_25" + original_name = "linear_47.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-0.115215") @@ -276,6 +302,7 @@ class Program_weight_tensor_parameter_25: class Program_weight_tensor_parameter_26: name = "parameter_26" + original_name = "linear_46.b_0" shape = [3072] dtype = "float32" min_val = float("-1.36362") @@ -287,6 +314,7 @@ class Program_weight_tensor_parameter_26: class Program_weight_tensor_parameter_27: name = "parameter_27" + original_name = "linear_46.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.161895") @@ -298,6 +326,7 @@ class Program_weight_tensor_parameter_27: class Program_weight_tensor_parameter_28: name = "parameter_28" + original_name = "layer_norm_24.b_0" shape = [768] dtype = "float32" min_val = float("-2.30851") @@ -309,6 +338,7 @@ class Program_weight_tensor_parameter_28: class Program_weight_tensor_parameter_29: name = "parameter_29" + original_name = "layer_norm_24.w_0" shape = [768] dtype = "float32" min_val = float("0.889446") @@ -320,6 +350,7 @@ class Program_weight_tensor_parameter_29: class Program_weight_tensor_parameter_30: name = "parameter_30" + original_name = "linear_45.b_0" shape = [768] dtype = "float32" min_val = float("-7.19411") @@ -331,6 +362,7 @@ class Program_weight_tensor_parameter_30: class Program_weight_tensor_parameter_31: name = "parameter_31" + original_name = "linear_45.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.160982") @@ -342,6 +374,7 @@ class Program_weight_tensor_parameter_31: class Program_weight_tensor_parameter_32: name = "parameter_32" + original_name = "linear_44.b_0" shape = [2304] dtype = "float32" min_val = float("-2.2147") @@ -353,6 +386,7 @@ class Program_weight_tensor_parameter_32: class Program_weight_tensor_parameter_33: name = "parameter_33" + original_name = "linear_44.w_0" shape = [768, 2304] dtype = "float32" min_val = float("-0.15186") @@ -364,6 +398,7 @@ class Program_weight_tensor_parameter_33: class Program_weight_tensor_parameter_34: name = "parameter_34" + original_name = "layer_norm_23.b_0" shape = [768] dtype = "float32" min_val = float("-2.09522") @@ -375,6 +410,7 @@ class Program_weight_tensor_parameter_34: class Program_weight_tensor_parameter_35: name = "parameter_35" + original_name = "layer_norm_23.w_0" shape = [768] dtype = "float32" min_val = float("0.346598") @@ -386,6 +422,7 @@ class Program_weight_tensor_parameter_35: class Program_weight_tensor_parameter_36: name = "parameter_36" + original_name = "linear_43.b_0" shape = [768] dtype = "float32" min_val = float("-0.351064") @@ -397,6 +434,7 @@ class Program_weight_tensor_parameter_36: class Program_weight_tensor_parameter_37: name = "parameter_37" + original_name = "linear_43.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-0.308212") @@ -408,6 +446,7 @@ class Program_weight_tensor_parameter_37: class Program_weight_tensor_parameter_38: name = "parameter_38" + original_name = "linear_42.b_0" shape = [3072] dtype = "float32" min_val = float("-1.66805") @@ -419,6 +458,7 @@ class Program_weight_tensor_parameter_38: class Program_weight_tensor_parameter_39: name = "parameter_39" + original_name = "linear_42.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.0922761") @@ -430,6 +470,7 @@ class Program_weight_tensor_parameter_39: class Program_weight_tensor_parameter_40: name = "parameter_40" + original_name = "layer_norm_22.b_0" shape = [768] dtype = "float32" min_val = float("-3.30835") @@ -441,6 +482,7 @@ class Program_weight_tensor_parameter_40: class Program_weight_tensor_parameter_41: name = "parameter_41" + original_name = "layer_norm_22.w_0" shape = [768] dtype = "float32" min_val = float("0.0765754") @@ -452,6 +494,7 @@ class Program_weight_tensor_parameter_41: class Program_weight_tensor_parameter_42: name = "parameter_42" + original_name = "linear_41.b_0" shape = [768] dtype = "float32" min_val = float("-0.354107") @@ -463,6 +506,7 @@ class Program_weight_tensor_parameter_42: class Program_weight_tensor_parameter_43: name = "parameter_43" + original_name = "linear_41.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.22582") @@ -474,6 +518,7 @@ class Program_weight_tensor_parameter_43: class Program_weight_tensor_parameter_44: name = "parameter_44" + original_name = "linear_40.b_0" shape = [2304] dtype = "float32" min_val = float("-3.1714") @@ -485,6 +530,7 @@ class Program_weight_tensor_parameter_44: class Program_weight_tensor_parameter_45: name = "parameter_45" + original_name = "linear_40.w_0" shape = [768, 2304] dtype = "float32" min_val = float("-0.170246") @@ -496,6 +542,7 @@ class Program_weight_tensor_parameter_45: class Program_weight_tensor_parameter_46: name = "parameter_46" + original_name = "layer_norm_21.b_0" shape = [768] dtype = "float32" min_val = float("-1.92449") @@ -507,6 +554,7 @@ class Program_weight_tensor_parameter_46: class Program_weight_tensor_parameter_47: name = "parameter_47" + original_name = "layer_norm_21.w_0" shape = [768] dtype = "float32" min_val = float("0.190346") @@ -518,6 +566,7 @@ class Program_weight_tensor_parameter_47: class Program_weight_tensor_parameter_48: name = "parameter_48" + original_name = "linear_39.b_0" shape = [768] dtype = "float32" min_val = float("-0.78645") @@ -529,6 +578,7 @@ class Program_weight_tensor_parameter_48: class Program_weight_tensor_parameter_49: name = "parameter_49" + original_name = "linear_39.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-0.232162") @@ -540,6 +590,7 @@ class Program_weight_tensor_parameter_49: class Program_weight_tensor_parameter_50: name = "parameter_50" + original_name = "linear_38.b_0" shape = [3072] dtype = "float32" min_val = float("-1.15743") @@ -551,6 +602,7 @@ class Program_weight_tensor_parameter_50: class Program_weight_tensor_parameter_51: name = "parameter_51" + original_name = "linear_38.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.0871569") @@ -562,6 +614,7 @@ class Program_weight_tensor_parameter_51: class Program_weight_tensor_parameter_52: name = "parameter_52" + original_name = "layer_norm_20.b_0" shape = [768] dtype = "float32" min_val = float("-3.74045") @@ -573,6 +626,7 @@ class Program_weight_tensor_parameter_52: class Program_weight_tensor_parameter_53: name = "parameter_53" + original_name = "layer_norm_20.w_0" shape = [768] dtype = "float32" min_val = float("0.214725") @@ -584,6 +638,7 @@ class Program_weight_tensor_parameter_53: class Program_weight_tensor_parameter_54: name = "parameter_54" + original_name = "linear_37.b_0" shape = [768] dtype = "float32" min_val = float("-0.706976") @@ -595,6 +650,7 @@ class Program_weight_tensor_parameter_54: class Program_weight_tensor_parameter_55: name = "parameter_55" + original_name = "linear_37.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.188515") @@ -606,6 +662,7 @@ class Program_weight_tensor_parameter_55: class Program_weight_tensor_parameter_56: name = "parameter_56" + original_name = "linear_36.b_0" shape = [2304] dtype = "float32" min_val = float("-2.92975") @@ -617,6 +674,7 @@ class Program_weight_tensor_parameter_56: class Program_weight_tensor_parameter_57: name = "parameter_57" + original_name = "linear_36.w_0" shape = [768, 2304] dtype = "float32" min_val = float("-0.137954") @@ -628,6 +686,7 @@ class Program_weight_tensor_parameter_57: class Program_weight_tensor_parameter_58: name = "parameter_58" + original_name = "layer_norm_19.b_0" shape = [768] dtype = "float32" min_val = float("-2.08994") @@ -639,6 +698,7 @@ class Program_weight_tensor_parameter_58: class Program_weight_tensor_parameter_59: name = "parameter_59" + original_name = "layer_norm_19.w_0" shape = [768] dtype = "float32" min_val = float("0.296997") @@ -650,6 +710,7 @@ class Program_weight_tensor_parameter_59: class Program_weight_tensor_parameter_60: name = "parameter_60" + original_name = "linear_35.b_0" shape = [768] dtype = "float32" min_val = float("-0.832353") @@ -661,6 +722,7 @@ class Program_weight_tensor_parameter_60: class Program_weight_tensor_parameter_61: name = "parameter_61" + original_name = "linear_35.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-0.311415") @@ -672,6 +734,7 @@ class Program_weight_tensor_parameter_61: class Program_weight_tensor_parameter_62: name = "parameter_62" + original_name = "linear_34.b_0" shape = [3072] dtype = "float32" min_val = float("-0.95784") @@ -683,6 +746,7 @@ class Program_weight_tensor_parameter_62: class Program_weight_tensor_parameter_63: name = "parameter_63" + original_name = "linear_34.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.128392") @@ -694,6 +758,7 @@ class Program_weight_tensor_parameter_63: class Program_weight_tensor_parameter_64: name = "parameter_64" + original_name = "layer_norm_18.b_0" shape = [768] dtype = "float32" min_val = float("-2.78726") @@ -705,6 +770,7 @@ class Program_weight_tensor_parameter_64: class Program_weight_tensor_parameter_65: name = "parameter_65" + original_name = "layer_norm_18.w_0" shape = [768] dtype = "float32" min_val = float("0.293647") @@ -716,6 +782,7 @@ class Program_weight_tensor_parameter_65: class Program_weight_tensor_parameter_66: name = "parameter_66" + original_name = "linear_33.b_0" shape = [768] dtype = "float32" min_val = float("-0.536813") @@ -727,6 +794,7 @@ class Program_weight_tensor_parameter_66: class Program_weight_tensor_parameter_67: name = "parameter_67" + original_name = "linear_33.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.131648") @@ -738,6 +806,7 @@ class Program_weight_tensor_parameter_67: class Program_weight_tensor_parameter_68: name = "parameter_68" + original_name = "linear_32.b_0" shape = [2304] dtype = "float32" min_val = float("-2.66188") @@ -749,6 +818,7 @@ class Program_weight_tensor_parameter_68: class Program_weight_tensor_parameter_69: name = "parameter_69" + original_name = "linear_32.w_0" shape = [768, 2304] dtype = "float32" min_val = float("-0.143223") @@ -760,6 +830,7 @@ class Program_weight_tensor_parameter_69: class Program_weight_tensor_parameter_70: name = "parameter_70" + original_name = "layer_norm_17.b_0" shape = [768] dtype = "float32" min_val = float("-2.57068") @@ -771,6 +842,7 @@ class Program_weight_tensor_parameter_70: class Program_weight_tensor_parameter_71: name = "parameter_71" + original_name = "layer_norm_17.w_0" shape = [768] dtype = "float32" min_val = float("0.368158") @@ -782,6 +854,7 @@ class Program_weight_tensor_parameter_71: class Program_weight_tensor_parameter_72: name = "parameter_72" + original_name = "linear_31.b_0" shape = [768] dtype = "float32" min_val = float("-1.51263") @@ -793,6 +866,7 @@ class Program_weight_tensor_parameter_72: class Program_weight_tensor_parameter_73: name = "parameter_73" + original_name = "linear_31.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-0.193308") @@ -804,6 +878,7 @@ class Program_weight_tensor_parameter_73: class Program_weight_tensor_parameter_74: name = "parameter_74" + original_name = "linear_30.b_0" shape = [3072] dtype = "float32" min_val = float("-1.03227") @@ -815,6 +890,7 @@ class Program_weight_tensor_parameter_74: class Program_weight_tensor_parameter_75: name = "parameter_75" + original_name = "linear_30.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.316096") @@ -826,6 +902,7 @@ class Program_weight_tensor_parameter_75: class Program_weight_tensor_parameter_76: name = "parameter_76" + original_name = "layer_norm_16.b_0" shape = [768] dtype = "float32" min_val = float("-7.35983") @@ -837,6 +914,7 @@ class Program_weight_tensor_parameter_76: class Program_weight_tensor_parameter_77: name = "parameter_77" + original_name = "layer_norm_16.w_0" shape = [768] dtype = "float32" min_val = float("0.622149") @@ -848,6 +926,7 @@ class Program_weight_tensor_parameter_77: class Program_weight_tensor_parameter_78: name = "parameter_78" + original_name = "linear_29.b_0" shape = [768] dtype = "float32" min_val = float("-1.97119") @@ -859,6 +938,7 @@ class Program_weight_tensor_parameter_78: class Program_weight_tensor_parameter_79: name = "parameter_79" + original_name = "linear_29.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.0920979") @@ -870,6 +950,7 @@ class Program_weight_tensor_parameter_79: class Program_weight_tensor_parameter_80: name = "parameter_80" + original_name = "linear_28.b_0" shape = [2304] dtype = "float32" min_val = float("-2.52528") @@ -881,6 +962,7 @@ class Program_weight_tensor_parameter_80: class Program_weight_tensor_parameter_81: name = "parameter_81" + original_name = "linear_28.w_0" shape = [768, 2304] dtype = "float32" min_val = float("-0.161749") @@ -892,6 +974,7 @@ class Program_weight_tensor_parameter_81: class Program_weight_tensor_parameter_82: name = "parameter_82" + original_name = "layer_norm_15.b_0" shape = [768] dtype = "float32" min_val = float("-2.47657") @@ -903,6 +986,7 @@ class Program_weight_tensor_parameter_82: class Program_weight_tensor_parameter_83: name = "parameter_83" + original_name = "layer_norm_15.w_0" shape = [768] dtype = "float32" min_val = float("0.283797") @@ -914,6 +998,7 @@ class Program_weight_tensor_parameter_83: class Program_weight_tensor_parameter_84: name = "parameter_84" + original_name = "linear_27.b_0" shape = [768] dtype = "float32" min_val = float("-0.655382") @@ -925,6 +1010,7 @@ class Program_weight_tensor_parameter_84: class Program_weight_tensor_parameter_85: name = "parameter_85" + original_name = "linear_27.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-0.198635") @@ -936,6 +1022,7 @@ class Program_weight_tensor_parameter_85: class Program_weight_tensor_parameter_86: name = "parameter_86" + original_name = "linear_26.b_0" shape = [3072] dtype = "float32" min_val = float("-1.27721") @@ -947,6 +1034,7 @@ class Program_weight_tensor_parameter_86: class Program_weight_tensor_parameter_87: name = "parameter_87" + original_name = "linear_26.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.16075") @@ -958,6 +1046,7 @@ class Program_weight_tensor_parameter_87: class Program_weight_tensor_parameter_88: name = "parameter_88" + original_name = "layer_norm_14.b_0" shape = [768] dtype = "float32" min_val = float("-7.89944") @@ -969,6 +1058,7 @@ class Program_weight_tensor_parameter_88: class Program_weight_tensor_parameter_89: name = "parameter_89" + original_name = "layer_norm_14.w_0" shape = [768] dtype = "float32" min_val = float("0.616027") @@ -980,6 +1070,7 @@ class Program_weight_tensor_parameter_89: class Program_weight_tensor_parameter_90: name = "parameter_90" + original_name = "linear_25.b_0" shape = [768] dtype = "float32" min_val = float("-1.12289") @@ -991,6 +1082,7 @@ class Program_weight_tensor_parameter_90: class Program_weight_tensor_parameter_91: name = "parameter_91" + original_name = "linear_25.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.11091") @@ -1002,6 +1094,7 @@ class Program_weight_tensor_parameter_91: class Program_weight_tensor_parameter_92: name = "parameter_92" + original_name = "linear_24.b_0" shape = [2304] dtype = "float32" min_val = float("-2.52948") @@ -1013,6 +1106,7 @@ class Program_weight_tensor_parameter_92: class Program_weight_tensor_parameter_93: name = "parameter_93" + original_name = "linear_24.w_0" shape = [768, 2304] dtype = "float32" min_val = float("-0.185187") @@ -1024,6 +1118,7 @@ class Program_weight_tensor_parameter_93: class Program_weight_tensor_parameter_94: name = "parameter_94" + original_name = "layer_norm_13.b_0" shape = [768] dtype = "float32" min_val = float("-2.43769") @@ -1035,6 +1130,7 @@ class Program_weight_tensor_parameter_94: class Program_weight_tensor_parameter_95: name = "parameter_95" + original_name = "layer_norm_13.w_0" shape = [768] dtype = "float32" min_val = float("0.422635") @@ -1046,6 +1142,7 @@ class Program_weight_tensor_parameter_95: class Program_weight_tensor_parameter_96: name = "parameter_96" + original_name = "linear_23.b_0" shape = [768] dtype = "float32" min_val = float("-0.243277") @@ -1057,6 +1154,7 @@ class Program_weight_tensor_parameter_96: class Program_weight_tensor_parameter_97: name = "parameter_97" + original_name = "linear_23.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-0.250877") @@ -1068,6 +1166,7 @@ class Program_weight_tensor_parameter_97: class Program_weight_tensor_parameter_98: name = "parameter_98" + original_name = "linear_22.b_0" shape = [3072] dtype = "float32" min_val = float("-1.02128") @@ -1079,6 +1178,7 @@ class Program_weight_tensor_parameter_98: class Program_weight_tensor_parameter_99: name = "parameter_99" + original_name = "linear_22.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.179363") @@ -1090,6 +1190,7 @@ class Program_weight_tensor_parameter_99: class Program_weight_tensor_parameter_100: name = "parameter_100" + original_name = "layer_norm_12.b_0" shape = [768] dtype = "float32" min_val = float("-5.01805") @@ -1101,6 +1202,7 @@ class Program_weight_tensor_parameter_100: class Program_weight_tensor_parameter_101: name = "parameter_101" + original_name = "layer_norm_12.w_0" shape = [768] dtype = "float32" min_val = float("0.758959") @@ -1112,6 +1214,7 @@ class Program_weight_tensor_parameter_101: class Program_weight_tensor_parameter_102: name = "parameter_102" + original_name = "linear_21.b_0" shape = [768] dtype = "float32" min_val = float("-0.995215") @@ -1123,6 +1226,7 @@ class Program_weight_tensor_parameter_102: class Program_weight_tensor_parameter_103: name = "parameter_103" + original_name = "linear_21.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.101734") @@ -1134,6 +1238,7 @@ class Program_weight_tensor_parameter_103: class Program_weight_tensor_parameter_104: name = "parameter_104" + original_name = "linear_20.b_0" shape = [2304] dtype = "float32" min_val = float("-2.92805") @@ -1145,6 +1250,7 @@ class Program_weight_tensor_parameter_104: class Program_weight_tensor_parameter_105: name = "parameter_105" + original_name = "linear_20.w_0" shape = [768, 2304] dtype = "float32" min_val = float("-0.165573") @@ -1156,6 +1262,7 @@ class Program_weight_tensor_parameter_105: class Program_weight_tensor_parameter_106: name = "parameter_106" + original_name = "layer_norm_11.b_0" shape = [768] dtype = "float32" min_val = float("-2.45883") @@ -1167,6 +1274,7 @@ class Program_weight_tensor_parameter_106: class Program_weight_tensor_parameter_107: name = "parameter_107" + original_name = "layer_norm_11.w_0" shape = [768] dtype = "float32" min_val = float("0.473003") @@ -1178,6 +1286,7 @@ class Program_weight_tensor_parameter_107: class Program_weight_tensor_parameter_108: name = "parameter_108" + original_name = "linear_19.b_0" shape = [768] dtype = "float32" min_val = float("-0.251239") @@ -1189,6 +1298,7 @@ class Program_weight_tensor_parameter_108: class Program_weight_tensor_parameter_109: name = "parameter_109" + original_name = "linear_19.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-0.32802") @@ -1200,6 +1310,7 @@ class Program_weight_tensor_parameter_109: class Program_weight_tensor_parameter_110: name = "parameter_110" + original_name = "linear_18.b_0" shape = [3072] dtype = "float32" min_val = float("-0.972244") @@ -1211,6 +1322,7 @@ class Program_weight_tensor_parameter_110: class Program_weight_tensor_parameter_111: name = "parameter_111" + original_name = "linear_18.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.173923") @@ -1222,6 +1334,7 @@ class Program_weight_tensor_parameter_111: class Program_weight_tensor_parameter_112: name = "parameter_112" + original_name = "layer_norm_10.b_0" shape = [768] dtype = "float32" min_val = float("-2.49497") @@ -1233,6 +1346,7 @@ class Program_weight_tensor_parameter_112: class Program_weight_tensor_parameter_113: name = "parameter_113" + original_name = "layer_norm_10.w_0" shape = [768] dtype = "float32" min_val = float("0.537359") @@ -1244,6 +1358,7 @@ class Program_weight_tensor_parameter_113: class Program_weight_tensor_parameter_114: name = "parameter_114" + original_name = "linear_17.b_0" shape = [768] dtype = "float32" min_val = float("-0.62237") @@ -1255,6 +1370,7 @@ class Program_weight_tensor_parameter_114: class Program_weight_tensor_parameter_115: name = "parameter_115" + original_name = "linear_17.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.069168") @@ -1266,6 +1382,7 @@ class Program_weight_tensor_parameter_115: class Program_weight_tensor_parameter_116: name = "parameter_116" + original_name = "linear_16.b_0" shape = [2304] dtype = "float32" min_val = float("-2.65286") @@ -1277,6 +1394,7 @@ class Program_weight_tensor_parameter_116: class Program_weight_tensor_parameter_117: name = "parameter_117" + original_name = "linear_16.w_0" shape = [768, 2304] dtype = "float32" min_val = float("-0.150103") @@ -1288,6 +1406,7 @@ class Program_weight_tensor_parameter_117: class Program_weight_tensor_parameter_118: name = "parameter_118" + original_name = "layer_norm_9.b_0" shape = [768] dtype = "float32" min_val = float("-2.26176") @@ -1299,6 +1418,7 @@ class Program_weight_tensor_parameter_118: class Program_weight_tensor_parameter_119: name = "parameter_119" + original_name = "layer_norm_9.w_0" shape = [768] dtype = "float32" min_val = float("0.522107") @@ -1310,6 +1430,7 @@ class Program_weight_tensor_parameter_119: class Program_weight_tensor_parameter_120: name = "parameter_120" + original_name = "linear_15.b_0" shape = [768] dtype = "float32" min_val = float("-0.335366") @@ -1321,6 +1442,7 @@ class Program_weight_tensor_parameter_120: class Program_weight_tensor_parameter_121: name = "parameter_121" + original_name = "linear_15.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-0.314974") @@ -1332,6 +1454,7 @@ class Program_weight_tensor_parameter_121: class Program_weight_tensor_parameter_122: name = "parameter_122" + original_name = "linear_14.b_0" shape = [3072] dtype = "float32" min_val = float("-1.17018") @@ -1343,6 +1466,7 @@ class Program_weight_tensor_parameter_122: class Program_weight_tensor_parameter_123: name = "parameter_123" + original_name = "linear_14.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.18718") @@ -1354,6 +1478,7 @@ class Program_weight_tensor_parameter_123: class Program_weight_tensor_parameter_124: name = "parameter_124" + original_name = "layer_norm_8.b_0" shape = [768] dtype = "float32" min_val = float("-1.99446") @@ -1365,6 +1490,7 @@ class Program_weight_tensor_parameter_124: class Program_weight_tensor_parameter_125: name = "parameter_125" + original_name = "layer_norm_8.w_0" shape = [768] dtype = "float32" min_val = float("0.362634") @@ -1376,6 +1502,7 @@ class Program_weight_tensor_parameter_125: class Program_weight_tensor_parameter_126: name = "parameter_126" + original_name = "linear_13.b_0" shape = [768] dtype = "float32" min_val = float("-0.767585") @@ -1387,6 +1514,7 @@ class Program_weight_tensor_parameter_126: class Program_weight_tensor_parameter_127: name = "parameter_127" + original_name = "linear_13.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.0845959") @@ -1398,6 +1526,7 @@ class Program_weight_tensor_parameter_127: class Program_weight_tensor_parameter_128: name = "parameter_128" + original_name = "linear_12.b_0" shape = [2304] dtype = "float32" min_val = float("-2.75091") @@ -1409,6 +1538,7 @@ class Program_weight_tensor_parameter_128: class Program_weight_tensor_parameter_129: name = "parameter_129" + original_name = "linear_12.w_0" shape = [768, 2304] dtype = "float32" min_val = float("-0.165258") @@ -1420,6 +1550,7 @@ class Program_weight_tensor_parameter_129: class Program_weight_tensor_parameter_130: name = "parameter_130" + original_name = "layer_norm_7.b_0" shape = [768] dtype = "float32" min_val = float("-1.89018") @@ -1431,6 +1562,7 @@ class Program_weight_tensor_parameter_130: class Program_weight_tensor_parameter_131: name = "parameter_131" + original_name = "layer_norm_7.w_0" shape = [768] dtype = "float32" min_val = float("0.362591") @@ -1442,6 +1574,7 @@ class Program_weight_tensor_parameter_131: class Program_weight_tensor_parameter_132: name = "parameter_132" + original_name = "linear_11.b_0" shape = [768] dtype = "float32" min_val = float("-0.575409") @@ -1453,6 +1586,7 @@ class Program_weight_tensor_parameter_132: class Program_weight_tensor_parameter_133: name = "parameter_133" + original_name = "linear_11.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-0.349476") @@ -1464,6 +1598,7 @@ class Program_weight_tensor_parameter_133: class Program_weight_tensor_parameter_134: name = "parameter_134" + original_name = "linear_10.b_0" shape = [3072] dtype = "float32" min_val = float("-1.22001") @@ -1475,6 +1610,7 @@ class Program_weight_tensor_parameter_134: class Program_weight_tensor_parameter_135: name = "parameter_135" + original_name = "linear_10.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.220208") @@ -1486,6 +1622,7 @@ class Program_weight_tensor_parameter_135: class Program_weight_tensor_parameter_136: name = "parameter_136" + original_name = "layer_norm_6.b_0" shape = [768] dtype = "float32" min_val = float("-2.2078") @@ -1497,6 +1634,7 @@ class Program_weight_tensor_parameter_136: class Program_weight_tensor_parameter_137: name = "parameter_137" + original_name = "layer_norm_6.w_0" shape = [768] dtype = "float32" min_val = float("0.31903") @@ -1508,6 +1646,7 @@ class Program_weight_tensor_parameter_137: class Program_weight_tensor_parameter_138: name = "parameter_138" + original_name = "linear_9.b_0" shape = [768] dtype = "float32" min_val = float("-1.03172") @@ -1519,6 +1658,7 @@ class Program_weight_tensor_parameter_138: class Program_weight_tensor_parameter_139: name = "parameter_139" + original_name = "linear_9.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.104568") @@ -1530,6 +1670,7 @@ class Program_weight_tensor_parameter_139: class Program_weight_tensor_parameter_140: name = "parameter_140" + original_name = "linear_8.b_0" shape = [2304] dtype = "float32" min_val = float("-2.36341") @@ -1541,6 +1682,7 @@ class Program_weight_tensor_parameter_140: class Program_weight_tensor_parameter_141: name = "parameter_141" + original_name = "linear_8.w_0" shape = [768, 2304] dtype = "float32" min_val = float("-0.166713") @@ -1552,6 +1694,7 @@ class Program_weight_tensor_parameter_141: class Program_weight_tensor_parameter_142: name = "parameter_142" + original_name = "layer_norm_5.b_0" shape = [768] dtype = "float32" min_val = float("-1.93409") @@ -1563,6 +1706,7 @@ class Program_weight_tensor_parameter_142: class Program_weight_tensor_parameter_143: name = "parameter_143" + original_name = "layer_norm_5.w_0" shape = [768] dtype = "float32" min_val = float("0.148613") @@ -1574,6 +1718,7 @@ class Program_weight_tensor_parameter_143: class Program_weight_tensor_parameter_144: name = "parameter_144" + original_name = "linear_7.b_0" shape = [768] dtype = "float32" min_val = float("-0.969605") @@ -1585,6 +1730,7 @@ class Program_weight_tensor_parameter_144: class Program_weight_tensor_parameter_145: name = "parameter_145" + original_name = "linear_7.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-0.325278") @@ -1596,6 +1742,7 @@ class Program_weight_tensor_parameter_145: class Program_weight_tensor_parameter_146: name = "parameter_146" + original_name = "linear_6.b_0" shape = [3072] dtype = "float32" min_val = float("-1.34357") @@ -1607,6 +1754,7 @@ class Program_weight_tensor_parameter_146: class Program_weight_tensor_parameter_147: name = "parameter_147" + original_name = "linear_6.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.199966") @@ -1618,6 +1766,7 @@ class Program_weight_tensor_parameter_147: class Program_weight_tensor_parameter_148: name = "parameter_148" + original_name = "layer_norm_4.b_0" shape = [768] dtype = "float32" min_val = float("-1.53153") @@ -1629,6 +1778,7 @@ class Program_weight_tensor_parameter_148: class Program_weight_tensor_parameter_149: name = "parameter_149" + original_name = "layer_norm_4.w_0" shape = [768] dtype = "float32" min_val = float("0.0575891") @@ -1640,6 +1790,7 @@ class Program_weight_tensor_parameter_149: class Program_weight_tensor_parameter_150: name = "parameter_150" + original_name = "linear_5.b_0" shape = [768] dtype = "float32" min_val = float("-0.995867") @@ -1651,6 +1802,7 @@ class Program_weight_tensor_parameter_150: class Program_weight_tensor_parameter_151: name = "parameter_151" + original_name = "linear_5.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.219077") @@ -1662,6 +1814,7 @@ class Program_weight_tensor_parameter_151: class Program_weight_tensor_parameter_152: name = "parameter_152" + original_name = "linear_4.b_0" shape = [2304] dtype = "float32" min_val = float("-2.13837") @@ -1673,6 +1826,7 @@ class Program_weight_tensor_parameter_152: class Program_weight_tensor_parameter_153: name = "parameter_153" + original_name = "linear_4.w_0" shape = [768, 2304] dtype = "float32" min_val = float("-0.207629") @@ -1684,6 +1838,7 @@ class Program_weight_tensor_parameter_153: class Program_weight_tensor_parameter_154: name = "parameter_154" + original_name = "layer_norm_3.b_0" shape = [768] dtype = "float32" min_val = float("-1.35332") @@ -1695,6 +1850,7 @@ class Program_weight_tensor_parameter_154: class Program_weight_tensor_parameter_155: name = "parameter_155" + original_name = "layer_norm_3.w_0" shape = [768] dtype = "float32" min_val = float("-0.000824287") @@ -1706,6 +1862,7 @@ class Program_weight_tensor_parameter_155: class Program_weight_tensor_parameter_156: name = "parameter_156" + original_name = "linear_3.b_0" shape = [768] dtype = "float32" min_val = float("-0.375516") @@ -1717,6 +1874,7 @@ class Program_weight_tensor_parameter_156: class Program_weight_tensor_parameter_157: name = "parameter_157" + original_name = "linear_3.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-0.224006") @@ -1728,6 +1886,7 @@ class Program_weight_tensor_parameter_157: class Program_weight_tensor_parameter_158: name = "parameter_158" + original_name = "linear_2.b_0" shape = [3072] dtype = "float32" min_val = float("-1.48737") @@ -1739,6 +1898,7 @@ class Program_weight_tensor_parameter_158: class Program_weight_tensor_parameter_159: name = "parameter_159" + original_name = "linear_2.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.277739") @@ -1750,6 +1910,7 @@ class Program_weight_tensor_parameter_159: class Program_weight_tensor_parameter_160: name = "parameter_160" + original_name = "layer_norm_2.b_0" shape = [768] dtype = "float32" min_val = float("-1.23187") @@ -1761,6 +1922,7 @@ class Program_weight_tensor_parameter_160: class Program_weight_tensor_parameter_161: name = "parameter_161" + original_name = "layer_norm_2.w_0" shape = [768] dtype = "float32" min_val = float("0.227053") @@ -1772,6 +1934,7 @@ class Program_weight_tensor_parameter_161: class Program_weight_tensor_parameter_162: name = "parameter_162" + original_name = "linear_1.b_0" shape = [768] dtype = "float32" min_val = float("-0.495867") @@ -1783,6 +1946,7 @@ class Program_weight_tensor_parameter_162: class Program_weight_tensor_parameter_163: name = "parameter_163" + original_name = "linear_1.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.171608") @@ -1794,6 +1958,7 @@ class Program_weight_tensor_parameter_163: class Program_weight_tensor_parameter_164: name = "parameter_164" + original_name = "linear_0.b_0" shape = [2304] dtype = "float32" min_val = float("-2.2833") @@ -1805,6 +1970,7 @@ class Program_weight_tensor_parameter_164: class Program_weight_tensor_parameter_165: name = "parameter_165" + original_name = "linear_0.w_0" shape = [768, 2304] dtype = "float32" min_val = float("-0.191934") @@ -1816,6 +1982,7 @@ class Program_weight_tensor_parameter_165: class Program_weight_tensor_parameter_166: name = "parameter_166" + original_name = "layer_norm_1.b_0" shape = [768] dtype = "float32" min_val = float("-1.12782") @@ -1827,6 +1994,7 @@ class Program_weight_tensor_parameter_166: class Program_weight_tensor_parameter_167: name = "parameter_167" + original_name = "layer_norm_1.w_0" shape = [768] dtype = "float32" min_val = float("-0.0108757") @@ -1838,6 +2006,7 @@ class Program_weight_tensor_parameter_167: class Program_weight_tensor_parameter_168: name = "parameter_168" + original_name = "layer_norm_0.b_0" shape = [768] dtype = "float32" min_val = float("-0.318354") @@ -1849,6 +2018,7 @@ class Program_weight_tensor_parameter_168: class Program_weight_tensor_parameter_169: name = "parameter_169" + original_name = "layer_norm_0.w_0" shape = [768] dtype = "float32" min_val = float("-0.00685189") @@ -1860,6 +2030,7 @@ class Program_weight_tensor_parameter_169: class Program_weight_tensor_parameter_170: name = "parameter_170" + original_name = "conv2d_0.w_0" shape = [768, 3, 16, 16] dtype = "float32" min_val = float("-0.146742") diff --git a/paddle_samples/PaddleX/CLIP_vit_base_patch16_448_ML/subgraph_2/input_meta.py b/paddle_samples/PaddleX/CLIP_vit_base_patch16_448_ML/subgraph_2/input_meta.py index ba83016d7..364bdf3e5 100644 --- a/paddle_samples/PaddleX/CLIP_vit_base_patch16_448_ML/subgraph_2/input_meta.py +++ b/paddle_samples/PaddleX/CLIP_vit_base_patch16_448_ML/subgraph_2/input_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_data_0: name = "data_0" + original_name = "param_0" shape = [1, 1, 768] dtype = "float32" min_val = float("-1.56196") @@ -11,6 +12,7 @@ class Program_weight_tensor_data_0: class Program_weight_tensor_data_1: name = "data_1" + original_name = "param_1" shape = [1, 197, 768] dtype = "float32" min_val = float("-0.220928") @@ -22,6 +24,7 @@ class Program_weight_tensor_data_1: class Program_weight_tensor_data_2: name = "data_2" + original_name = "param_2" shape = [33, 768] dtype = "float32" min_val = float("-0.0865287") @@ -33,6 +36,7 @@ class Program_weight_tensor_data_2: class Program_weight_tensor_data_3: name = "data_3" + original_name = "var_0" shape = [8, 3, 448, 448] dtype = "float32" min_val = float("-2.1179") diff --git a/paddle_samples/PaddleX/CLIP_vit_base_patch16_448_ML/subgraph_2/weight_meta.py b/paddle_samples/PaddleX/CLIP_vit_base_patch16_448_ML/subgraph_2/weight_meta.py index cde56d704..ce41d6f28 100644 --- a/paddle_samples/PaddleX/CLIP_vit_base_patch16_448_ML/subgraph_2/weight_meta.py +++ b/paddle_samples/PaddleX/CLIP_vit_base_patch16_448_ML/subgraph_2/weight_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_parameter_0: name = "parameter_0" + original_name = "conv2d_2.b_0" shape = [33] dtype = "float32" min_val = float("0") @@ -9,6 +10,7 @@ class Program_weight_tensor_parameter_0: class Program_weight_tensor_parameter_1: name = "parameter_1" + original_name = "conv2d_2.w_0" shape = [33, 768, 1, 1] dtype = "float32" min_val = float("-0.209469") @@ -20,6 +22,7 @@ class Program_weight_tensor_parameter_1: class Program_weight_tensor_parameter_2: name = "parameter_2" + original_name = "layer_norm_28.b_0" shape = [768] dtype = "float32" min_val = float("-0.00173656") @@ -31,6 +34,7 @@ class Program_weight_tensor_parameter_2: class Program_weight_tensor_parameter_3: name = "parameter_3" + original_name = "layer_norm_28.w_0" shape = [768] dtype = "float32" min_val = float("0.978248") @@ -42,6 +46,7 @@ class Program_weight_tensor_parameter_3: class Program_weight_tensor_parameter_4: name = "parameter_4" + original_name = "linear_58.b_0" shape = [768] dtype = "float32" min_val = float("-0.00929943") @@ -53,6 +58,7 @@ class Program_weight_tensor_parameter_4: class Program_weight_tensor_parameter_5: name = "parameter_5" + original_name = "linear_58.w_0" shape = [2048, 768] dtype = "float32" min_val = float("-0.0559707") @@ -64,6 +70,7 @@ class Program_weight_tensor_parameter_5: class Program_weight_tensor_parameter_6: name = "parameter_6" + original_name = "linear_57.b_0" shape = [2048] dtype = "float32" min_val = float("-0.00583159") @@ -75,6 +82,7 @@ class Program_weight_tensor_parameter_6: class Program_weight_tensor_parameter_7: name = "parameter_7" + original_name = "linear_57.w_0" shape = [768, 2048] dtype = "float32" min_val = float("-0.0523044") @@ -86,6 +94,7 @@ class Program_weight_tensor_parameter_7: class Program_weight_tensor_parameter_8: name = "parameter_8" + original_name = "layer_norm_27.b_0" shape = [768] dtype = "float32" min_val = float("-0.022251") @@ -97,6 +106,7 @@ class Program_weight_tensor_parameter_8: class Program_weight_tensor_parameter_9: name = "parameter_9" + original_name = "layer_norm_27.w_0" shape = [768] dtype = "float32" min_val = float("0.994222") @@ -108,6 +118,7 @@ class Program_weight_tensor_parameter_9: class Program_weight_tensor_parameter_10: name = "parameter_10" + original_name = "linear_56.b_0" shape = [768] dtype = "float32" min_val = float("-0.00741432") @@ -119,6 +130,7 @@ class Program_weight_tensor_parameter_10: class Program_weight_tensor_parameter_11: name = "parameter_11" + original_name = "linear_56.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.0798364") @@ -130,6 +142,7 @@ class Program_weight_tensor_parameter_11: class Program_weight_tensor_parameter_12: name = "parameter_12" + original_name = "linear_55.b_0" shape = [768] dtype = "float32" min_val = float("-0.00266821") @@ -141,6 +154,7 @@ class Program_weight_tensor_parameter_12: class Program_weight_tensor_parameter_13: name = "parameter_13" + original_name = "linear_55.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.0703408") @@ -152,6 +166,7 @@ class Program_weight_tensor_parameter_13: class Program_weight_tensor_parameter_14: name = "parameter_14" + original_name = "linear_54.b_0" shape = [768] dtype = "float32" data = None @@ -159,6 +174,7 @@ class Program_weight_tensor_parameter_14: class Program_weight_tensor_parameter_15: name = "parameter_15" + original_name = "linear_54.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.0624998") @@ -170,6 +186,7 @@ class Program_weight_tensor_parameter_15: class Program_weight_tensor_parameter_16: name = "parameter_16" + original_name = "linear_53.b_0" shape = [768] dtype = "float32" data = None @@ -177,6 +194,7 @@ class Program_weight_tensor_parameter_16: class Program_weight_tensor_parameter_17: name = "parameter_17" + original_name = "linear_53.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.0625") @@ -188,6 +206,7 @@ class Program_weight_tensor_parameter_17: class Program_weight_tensor_parameter_18: name = "parameter_18" + original_name = "layer_norm_26.b_0" shape = [768] dtype = "float32" min_val = float("-0.0170882") @@ -199,6 +218,7 @@ class Program_weight_tensor_parameter_18: class Program_weight_tensor_parameter_19: name = "parameter_19" + original_name = "layer_norm_26.w_0" shape = [768] dtype = "float32" min_val = float("0.99706") @@ -210,6 +230,7 @@ class Program_weight_tensor_parameter_19: class Program_weight_tensor_parameter_20: name = "parameter_20" + original_name = "conv2d_1.b_0" shape = [768] dtype = "float32" min_val = float("-0.00374967") @@ -221,6 +242,7 @@ class Program_weight_tensor_parameter_20: class Program_weight_tensor_parameter_21: name = "parameter_21" + original_name = "conv2d_1.w_0" shape = [768, 768, 1, 1] dtype = "float32" min_val = float("-0.251184") @@ -232,6 +254,7 @@ class Program_weight_tensor_parameter_21: class Program_weight_tensor_parameter_22: name = "parameter_22" + original_name = "layer_norm_25.b_0" shape = [768] dtype = "float32" min_val = float("-1.66268") @@ -243,6 +266,7 @@ class Program_weight_tensor_parameter_22: class Program_weight_tensor_parameter_23: name = "parameter_23" + original_name = "layer_norm_25.w_0" shape = [768] dtype = "float32" min_val = float("-0.196553") @@ -254,6 +278,7 @@ class Program_weight_tensor_parameter_23: class Program_weight_tensor_parameter_24: name = "parameter_24" + original_name = "linear_47.b_0" shape = [768] dtype = "float32" min_val = float("-2.91333") @@ -265,6 +290,7 @@ class Program_weight_tensor_parameter_24: class Program_weight_tensor_parameter_25: name = "parameter_25" + original_name = "linear_47.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-0.11469") @@ -276,6 +302,7 @@ class Program_weight_tensor_parameter_25: class Program_weight_tensor_parameter_26: name = "parameter_26" + original_name = "linear_46.b_0" shape = [3072] dtype = "float32" min_val = float("-1.36305") @@ -287,6 +314,7 @@ class Program_weight_tensor_parameter_26: class Program_weight_tensor_parameter_27: name = "parameter_27" + original_name = "linear_46.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.160164") @@ -298,6 +326,7 @@ class Program_weight_tensor_parameter_27: class Program_weight_tensor_parameter_28: name = "parameter_28" + original_name = "layer_norm_24.b_0" shape = [768] dtype = "float32" min_val = float("-2.31009") @@ -309,6 +338,7 @@ class Program_weight_tensor_parameter_28: class Program_weight_tensor_parameter_29: name = "parameter_29" + original_name = "layer_norm_24.w_0" shape = [768] dtype = "float32" min_val = float("0.888643") @@ -320,6 +350,7 @@ class Program_weight_tensor_parameter_29: class Program_weight_tensor_parameter_30: name = "parameter_30" + original_name = "linear_45.b_0" shape = [768] dtype = "float32" min_val = float("-7.19312") @@ -331,6 +362,7 @@ class Program_weight_tensor_parameter_30: class Program_weight_tensor_parameter_31: name = "parameter_31" + original_name = "linear_45.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.15897") @@ -342,6 +374,7 @@ class Program_weight_tensor_parameter_31: class Program_weight_tensor_parameter_32: name = "parameter_32" + original_name = "linear_44.b_0" shape = [2304] dtype = "float32" min_val = float("-2.21317") @@ -353,6 +386,7 @@ class Program_weight_tensor_parameter_32: class Program_weight_tensor_parameter_33: name = "parameter_33" + original_name = "linear_44.w_0" shape = [768, 2304] dtype = "float32" min_val = float("-0.149614") @@ -364,6 +398,7 @@ class Program_weight_tensor_parameter_33: class Program_weight_tensor_parameter_34: name = "parameter_34" + original_name = "layer_norm_23.b_0" shape = [768] dtype = "float32" min_val = float("-2.09741") @@ -375,6 +410,7 @@ class Program_weight_tensor_parameter_34: class Program_weight_tensor_parameter_35: name = "parameter_35" + original_name = "layer_norm_23.w_0" shape = [768] dtype = "float32" min_val = float("0.345147") @@ -386,6 +422,7 @@ class Program_weight_tensor_parameter_35: class Program_weight_tensor_parameter_36: name = "parameter_36" + original_name = "linear_43.b_0" shape = [768] dtype = "float32" min_val = float("-0.349931") @@ -397,6 +434,7 @@ class Program_weight_tensor_parameter_36: class Program_weight_tensor_parameter_37: name = "parameter_37" + original_name = "linear_43.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-0.301543") @@ -408,6 +446,7 @@ class Program_weight_tensor_parameter_37: class Program_weight_tensor_parameter_38: name = "parameter_38" + original_name = "linear_42.b_0" shape = [3072] dtype = "float32" min_val = float("-1.66871") @@ -419,6 +458,7 @@ class Program_weight_tensor_parameter_38: class Program_weight_tensor_parameter_39: name = "parameter_39" + original_name = "linear_42.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.0919095") @@ -430,6 +470,7 @@ class Program_weight_tensor_parameter_39: class Program_weight_tensor_parameter_40: name = "parameter_40" + original_name = "layer_norm_22.b_0" shape = [768] dtype = "float32" min_val = float("-3.30111") @@ -441,6 +482,7 @@ class Program_weight_tensor_parameter_40: class Program_weight_tensor_parameter_41: name = "parameter_41" + original_name = "layer_norm_22.w_0" shape = [768] dtype = "float32" min_val = float("0.0782615") @@ -452,6 +494,7 @@ class Program_weight_tensor_parameter_41: class Program_weight_tensor_parameter_42: name = "parameter_42" + original_name = "linear_41.b_0" shape = [768] dtype = "float32" min_val = float("-0.352172") @@ -463,6 +506,7 @@ class Program_weight_tensor_parameter_42: class Program_weight_tensor_parameter_43: name = "parameter_43" + original_name = "linear_41.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.224819") @@ -474,6 +518,7 @@ class Program_weight_tensor_parameter_43: class Program_weight_tensor_parameter_44: name = "parameter_44" + original_name = "linear_40.b_0" shape = [2304] dtype = "float32" min_val = float("-3.16968") @@ -485,6 +530,7 @@ class Program_weight_tensor_parameter_44: class Program_weight_tensor_parameter_45: name = "parameter_45" + original_name = "linear_40.w_0" shape = [768, 2304] dtype = "float32" min_val = float("-0.166357") @@ -496,6 +542,7 @@ class Program_weight_tensor_parameter_45: class Program_weight_tensor_parameter_46: name = "parameter_46" + original_name = "layer_norm_21.b_0" shape = [768] dtype = "float32" min_val = float("-1.92357") @@ -507,6 +554,7 @@ class Program_weight_tensor_parameter_46: class Program_weight_tensor_parameter_47: name = "parameter_47" + original_name = "layer_norm_21.w_0" shape = [768] dtype = "float32" min_val = float("0.190894") @@ -518,6 +566,7 @@ class Program_weight_tensor_parameter_47: class Program_weight_tensor_parameter_48: name = "parameter_48" + original_name = "linear_39.b_0" shape = [768] dtype = "float32" min_val = float("-0.783883") @@ -529,6 +578,7 @@ class Program_weight_tensor_parameter_48: class Program_weight_tensor_parameter_49: name = "parameter_49" + original_name = "linear_39.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-0.233022") @@ -540,6 +590,7 @@ class Program_weight_tensor_parameter_49: class Program_weight_tensor_parameter_50: name = "parameter_50" + original_name = "linear_38.b_0" shape = [3072] dtype = "float32" min_val = float("-1.15674") @@ -551,6 +602,7 @@ class Program_weight_tensor_parameter_50: class Program_weight_tensor_parameter_51: name = "parameter_51" + original_name = "linear_38.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.0867691") @@ -562,6 +614,7 @@ class Program_weight_tensor_parameter_51: class Program_weight_tensor_parameter_52: name = "parameter_52" + original_name = "layer_norm_20.b_0" shape = [768] dtype = "float32" min_val = float("-3.73717") @@ -573,6 +626,7 @@ class Program_weight_tensor_parameter_52: class Program_weight_tensor_parameter_53: name = "parameter_53" + original_name = "layer_norm_20.w_0" shape = [768] dtype = "float32" min_val = float("0.218019") @@ -584,6 +638,7 @@ class Program_weight_tensor_parameter_53: class Program_weight_tensor_parameter_54: name = "parameter_54" + original_name = "linear_37.b_0" shape = [768] dtype = "float32" min_val = float("-0.706593") @@ -595,6 +650,7 @@ class Program_weight_tensor_parameter_54: class Program_weight_tensor_parameter_55: name = "parameter_55" + original_name = "linear_37.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.19023") @@ -606,6 +662,7 @@ class Program_weight_tensor_parameter_55: class Program_weight_tensor_parameter_56: name = "parameter_56" + original_name = "linear_36.b_0" shape = [2304] dtype = "float32" min_val = float("-2.92966") @@ -617,6 +674,7 @@ class Program_weight_tensor_parameter_56: class Program_weight_tensor_parameter_57: name = "parameter_57" + original_name = "linear_36.w_0" shape = [768, 2304] dtype = "float32" min_val = float("-0.140196") @@ -628,6 +686,7 @@ class Program_weight_tensor_parameter_57: class Program_weight_tensor_parameter_58: name = "parameter_58" + original_name = "layer_norm_19.b_0" shape = [768] dtype = "float32" min_val = float("-2.0869") @@ -639,6 +698,7 @@ class Program_weight_tensor_parameter_58: class Program_weight_tensor_parameter_59: name = "parameter_59" + original_name = "layer_norm_19.w_0" shape = [768] dtype = "float32" min_val = float("0.299495") @@ -650,6 +710,7 @@ class Program_weight_tensor_parameter_59: class Program_weight_tensor_parameter_60: name = "parameter_60" + original_name = "linear_35.b_0" shape = [768] dtype = "float32" min_val = float("-0.831269") @@ -661,6 +722,7 @@ class Program_weight_tensor_parameter_60: class Program_weight_tensor_parameter_61: name = "parameter_61" + original_name = "linear_35.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-0.313957") @@ -672,6 +734,7 @@ class Program_weight_tensor_parameter_61: class Program_weight_tensor_parameter_62: name = "parameter_62" + original_name = "linear_34.b_0" shape = [3072] dtype = "float32" min_val = float("-0.957513") @@ -683,6 +746,7 @@ class Program_weight_tensor_parameter_62: class Program_weight_tensor_parameter_63: name = "parameter_63" + original_name = "linear_34.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.131615") @@ -694,6 +758,7 @@ class Program_weight_tensor_parameter_63: class Program_weight_tensor_parameter_64: name = "parameter_64" + original_name = "layer_norm_18.b_0" shape = [768] dtype = "float32" min_val = float("-2.78421") @@ -705,6 +770,7 @@ class Program_weight_tensor_parameter_64: class Program_weight_tensor_parameter_65: name = "parameter_65" + original_name = "layer_norm_18.w_0" shape = [768] dtype = "float32" min_val = float("0.290691") @@ -716,6 +782,7 @@ class Program_weight_tensor_parameter_65: class Program_weight_tensor_parameter_66: name = "parameter_66" + original_name = "linear_33.b_0" shape = [768] dtype = "float32" min_val = float("-0.541811") @@ -727,6 +794,7 @@ class Program_weight_tensor_parameter_66: class Program_weight_tensor_parameter_67: name = "parameter_67" + original_name = "linear_33.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.134193") @@ -738,6 +806,7 @@ class Program_weight_tensor_parameter_67: class Program_weight_tensor_parameter_68: name = "parameter_68" + original_name = "linear_32.b_0" shape = [2304] dtype = "float32" min_val = float("-2.66556") @@ -749,6 +818,7 @@ class Program_weight_tensor_parameter_68: class Program_weight_tensor_parameter_69: name = "parameter_69" + original_name = "linear_32.w_0" shape = [768, 2304] dtype = "float32" min_val = float("-0.142037") @@ -760,6 +830,7 @@ class Program_weight_tensor_parameter_69: class Program_weight_tensor_parameter_70: name = "parameter_70" + original_name = "layer_norm_17.b_0" shape = [768] dtype = "float32" min_val = float("-2.57032") @@ -771,6 +842,7 @@ class Program_weight_tensor_parameter_70: class Program_weight_tensor_parameter_71: name = "parameter_71" + original_name = "layer_norm_17.w_0" shape = [768] dtype = "float32" min_val = float("0.368544") @@ -782,6 +854,7 @@ class Program_weight_tensor_parameter_71: class Program_weight_tensor_parameter_72: name = "parameter_72" + original_name = "linear_31.b_0" shape = [768] dtype = "float32" min_val = float("-1.5116") @@ -793,6 +866,7 @@ class Program_weight_tensor_parameter_72: class Program_weight_tensor_parameter_73: name = "parameter_73" + original_name = "linear_31.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-0.192765") @@ -804,6 +878,7 @@ class Program_weight_tensor_parameter_73: class Program_weight_tensor_parameter_74: name = "parameter_74" + original_name = "linear_30.b_0" shape = [3072] dtype = "float32" min_val = float("-1.03028") @@ -815,6 +890,7 @@ class Program_weight_tensor_parameter_74: class Program_weight_tensor_parameter_75: name = "parameter_75" + original_name = "linear_30.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.320952") @@ -826,6 +902,7 @@ class Program_weight_tensor_parameter_75: class Program_weight_tensor_parameter_76: name = "parameter_76" + original_name = "layer_norm_16.b_0" shape = [768] dtype = "float32" min_val = float("-7.3634") @@ -837,6 +914,7 @@ class Program_weight_tensor_parameter_76: class Program_weight_tensor_parameter_77: name = "parameter_77" + original_name = "layer_norm_16.w_0" shape = [768] dtype = "float32" min_val = float("0.61816") @@ -848,6 +926,7 @@ class Program_weight_tensor_parameter_77: class Program_weight_tensor_parameter_78: name = "parameter_78" + original_name = "linear_29.b_0" shape = [768] dtype = "float32" min_val = float("-1.97478") @@ -859,6 +938,7 @@ class Program_weight_tensor_parameter_78: class Program_weight_tensor_parameter_79: name = "parameter_79" + original_name = "linear_29.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.0912049") @@ -870,6 +950,7 @@ class Program_weight_tensor_parameter_79: class Program_weight_tensor_parameter_80: name = "parameter_80" + original_name = "linear_28.b_0" shape = [2304] dtype = "float32" min_val = float("-2.52247") @@ -881,6 +962,7 @@ class Program_weight_tensor_parameter_80: class Program_weight_tensor_parameter_81: name = "parameter_81" + original_name = "linear_28.w_0" shape = [768, 2304] dtype = "float32" min_val = float("-0.159709") @@ -892,6 +974,7 @@ class Program_weight_tensor_parameter_81: class Program_weight_tensor_parameter_82: name = "parameter_82" + original_name = "layer_norm_15.b_0" shape = [768] dtype = "float32" min_val = float("-2.47639") @@ -903,6 +986,7 @@ class Program_weight_tensor_parameter_82: class Program_weight_tensor_parameter_83: name = "parameter_83" + original_name = "layer_norm_15.w_0" shape = [768] dtype = "float32" min_val = float("0.28326") @@ -914,6 +998,7 @@ class Program_weight_tensor_parameter_83: class Program_weight_tensor_parameter_84: name = "parameter_84" + original_name = "linear_27.b_0" shape = [768] dtype = "float32" min_val = float("-0.656186") @@ -925,6 +1010,7 @@ class Program_weight_tensor_parameter_84: class Program_weight_tensor_parameter_85: name = "parameter_85" + original_name = "linear_27.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-0.198722") @@ -936,6 +1022,7 @@ class Program_weight_tensor_parameter_85: class Program_weight_tensor_parameter_86: name = "parameter_86" + original_name = "linear_26.b_0" shape = [3072] dtype = "float32" min_val = float("-1.28156") @@ -947,6 +1034,7 @@ class Program_weight_tensor_parameter_86: class Program_weight_tensor_parameter_87: name = "parameter_87" + original_name = "linear_26.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.165945") @@ -958,6 +1046,7 @@ class Program_weight_tensor_parameter_87: class Program_weight_tensor_parameter_88: name = "parameter_88" + original_name = "layer_norm_14.b_0" shape = [768] dtype = "float32" min_val = float("-7.90225") @@ -969,6 +1058,7 @@ class Program_weight_tensor_parameter_88: class Program_weight_tensor_parameter_89: name = "parameter_89" + original_name = "layer_norm_14.w_0" shape = [768] dtype = "float32" min_val = float("0.615221") @@ -980,6 +1070,7 @@ class Program_weight_tensor_parameter_89: class Program_weight_tensor_parameter_90: name = "parameter_90" + original_name = "linear_25.b_0" shape = [768] dtype = "float32" min_val = float("-1.12654") @@ -991,6 +1082,7 @@ class Program_weight_tensor_parameter_90: class Program_weight_tensor_parameter_91: name = "parameter_91" + original_name = "linear_25.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.111449") @@ -1002,6 +1094,7 @@ class Program_weight_tensor_parameter_91: class Program_weight_tensor_parameter_92: name = "parameter_92" + original_name = "linear_24.b_0" shape = [2304] dtype = "float32" min_val = float("-2.52852") @@ -1013,6 +1106,7 @@ class Program_weight_tensor_parameter_92: class Program_weight_tensor_parameter_93: name = "parameter_93" + original_name = "linear_24.w_0" shape = [768, 2304] dtype = "float32" min_val = float("-0.179875") @@ -1024,6 +1118,7 @@ class Program_weight_tensor_parameter_93: class Program_weight_tensor_parameter_94: name = "parameter_94" + original_name = "layer_norm_13.b_0" shape = [768] dtype = "float32" min_val = float("-2.43728") @@ -1035,6 +1130,7 @@ class Program_weight_tensor_parameter_94: class Program_weight_tensor_parameter_95: name = "parameter_95" + original_name = "layer_norm_13.w_0" shape = [768] dtype = "float32" min_val = float("0.422955") @@ -1046,6 +1142,7 @@ class Program_weight_tensor_parameter_95: class Program_weight_tensor_parameter_96: name = "parameter_96" + original_name = "linear_23.b_0" shape = [768] dtype = "float32" min_val = float("-0.242907") @@ -1057,6 +1154,7 @@ class Program_weight_tensor_parameter_96: class Program_weight_tensor_parameter_97: name = "parameter_97" + original_name = "linear_23.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-0.252255") @@ -1068,6 +1166,7 @@ class Program_weight_tensor_parameter_97: class Program_weight_tensor_parameter_98: name = "parameter_98" + original_name = "linear_22.b_0" shape = [3072] dtype = "float32" min_val = float("-1.02261") @@ -1079,6 +1178,7 @@ class Program_weight_tensor_parameter_98: class Program_weight_tensor_parameter_99: name = "parameter_99" + original_name = "linear_22.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.178997") @@ -1090,6 +1190,7 @@ class Program_weight_tensor_parameter_99: class Program_weight_tensor_parameter_100: name = "parameter_100" + original_name = "layer_norm_12.b_0" shape = [768] dtype = "float32" min_val = float("-5.01811") @@ -1101,6 +1202,7 @@ class Program_weight_tensor_parameter_100: class Program_weight_tensor_parameter_101: name = "parameter_101" + original_name = "layer_norm_12.w_0" shape = [768] dtype = "float32" min_val = float("0.756998") @@ -1112,6 +1214,7 @@ class Program_weight_tensor_parameter_101: class Program_weight_tensor_parameter_102: name = "parameter_102" + original_name = "linear_21.b_0" shape = [768] dtype = "float32" min_val = float("-0.997953") @@ -1123,6 +1226,7 @@ class Program_weight_tensor_parameter_102: class Program_weight_tensor_parameter_103: name = "parameter_103" + original_name = "linear_21.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.0987804") @@ -1134,6 +1238,7 @@ class Program_weight_tensor_parameter_103: class Program_weight_tensor_parameter_104: name = "parameter_104" + original_name = "linear_20.b_0" shape = [2304] dtype = "float32" min_val = float("-2.92641") @@ -1145,6 +1250,7 @@ class Program_weight_tensor_parameter_104: class Program_weight_tensor_parameter_105: name = "parameter_105" + original_name = "linear_20.w_0" shape = [768, 2304] dtype = "float32" min_val = float("-0.15969") @@ -1156,6 +1262,7 @@ class Program_weight_tensor_parameter_105: class Program_weight_tensor_parameter_106: name = "parameter_106" + original_name = "layer_norm_11.b_0" shape = [768] dtype = "float32" min_val = float("-2.4625") @@ -1167,6 +1274,7 @@ class Program_weight_tensor_parameter_106: class Program_weight_tensor_parameter_107: name = "parameter_107" + original_name = "layer_norm_11.w_0" shape = [768] dtype = "float32" min_val = float("0.470648") @@ -1178,6 +1286,7 @@ class Program_weight_tensor_parameter_107: class Program_weight_tensor_parameter_108: name = "parameter_108" + original_name = "linear_19.b_0" shape = [768] dtype = "float32" min_val = float("-0.250379") @@ -1189,6 +1298,7 @@ class Program_weight_tensor_parameter_108: class Program_weight_tensor_parameter_109: name = "parameter_109" + original_name = "linear_19.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-0.315509") @@ -1200,6 +1310,7 @@ class Program_weight_tensor_parameter_109: class Program_weight_tensor_parameter_110: name = "parameter_110" + original_name = "linear_18.b_0" shape = [3072] dtype = "float32" min_val = float("-0.973659") @@ -1211,6 +1322,7 @@ class Program_weight_tensor_parameter_110: class Program_weight_tensor_parameter_111: name = "parameter_111" + original_name = "linear_18.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.171095") @@ -1222,6 +1334,7 @@ class Program_weight_tensor_parameter_111: class Program_weight_tensor_parameter_112: name = "parameter_112" + original_name = "layer_norm_10.b_0" shape = [768] dtype = "float32" min_val = float("-2.48863") @@ -1233,6 +1346,7 @@ class Program_weight_tensor_parameter_112: class Program_weight_tensor_parameter_113: name = "parameter_113" + original_name = "layer_norm_10.w_0" shape = [768] dtype = "float32" min_val = float("0.534553") @@ -1244,6 +1358,7 @@ class Program_weight_tensor_parameter_113: class Program_weight_tensor_parameter_114: name = "parameter_114" + original_name = "linear_17.b_0" shape = [768] dtype = "float32" min_val = float("-0.622629") @@ -1255,6 +1370,7 @@ class Program_weight_tensor_parameter_114: class Program_weight_tensor_parameter_115: name = "parameter_115" + original_name = "linear_17.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.0700505") @@ -1266,6 +1382,7 @@ class Program_weight_tensor_parameter_115: class Program_weight_tensor_parameter_116: name = "parameter_116" + original_name = "linear_16.b_0" shape = [2304] dtype = "float32" min_val = float("-2.65289") @@ -1277,6 +1394,7 @@ class Program_weight_tensor_parameter_116: class Program_weight_tensor_parameter_117: name = "parameter_117" + original_name = "linear_16.w_0" shape = [768, 2304] dtype = "float32" min_val = float("-0.148306") @@ -1288,6 +1406,7 @@ class Program_weight_tensor_parameter_117: class Program_weight_tensor_parameter_118: name = "parameter_118" + original_name = "layer_norm_9.b_0" shape = [768] dtype = "float32" min_val = float("-2.26006") @@ -1299,6 +1418,7 @@ class Program_weight_tensor_parameter_118: class Program_weight_tensor_parameter_119: name = "parameter_119" + original_name = "layer_norm_9.w_0" shape = [768] dtype = "float32" min_val = float("0.518355") @@ -1310,6 +1430,7 @@ class Program_weight_tensor_parameter_119: class Program_weight_tensor_parameter_120: name = "parameter_120" + original_name = "linear_15.b_0" shape = [768] dtype = "float32" min_val = float("-0.335812") @@ -1321,6 +1442,7 @@ class Program_weight_tensor_parameter_120: class Program_weight_tensor_parameter_121: name = "parameter_121" + original_name = "linear_15.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-0.309881") @@ -1332,6 +1454,7 @@ class Program_weight_tensor_parameter_121: class Program_weight_tensor_parameter_122: name = "parameter_122" + original_name = "linear_14.b_0" shape = [3072] dtype = "float32" min_val = float("-1.17332") @@ -1343,6 +1466,7 @@ class Program_weight_tensor_parameter_122: class Program_weight_tensor_parameter_123: name = "parameter_123" + original_name = "linear_14.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.188677") @@ -1354,6 +1478,7 @@ class Program_weight_tensor_parameter_123: class Program_weight_tensor_parameter_124: name = "parameter_124" + original_name = "layer_norm_8.b_0" shape = [768] dtype = "float32" min_val = float("-1.98716") @@ -1365,6 +1490,7 @@ class Program_weight_tensor_parameter_124: class Program_weight_tensor_parameter_125: name = "parameter_125" + original_name = "layer_norm_8.w_0" shape = [768] dtype = "float32" min_val = float("0.361211") @@ -1376,6 +1502,7 @@ class Program_weight_tensor_parameter_125: class Program_weight_tensor_parameter_126: name = "parameter_126" + original_name = "linear_13.b_0" shape = [768] dtype = "float32" min_val = float("-0.765817") @@ -1387,6 +1514,7 @@ class Program_weight_tensor_parameter_126: class Program_weight_tensor_parameter_127: name = "parameter_127" + original_name = "linear_13.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.085365") @@ -1398,6 +1526,7 @@ class Program_weight_tensor_parameter_127: class Program_weight_tensor_parameter_128: name = "parameter_128" + original_name = "linear_12.b_0" shape = [2304] dtype = "float32" min_val = float("-2.75515") @@ -1409,6 +1538,7 @@ class Program_weight_tensor_parameter_128: class Program_weight_tensor_parameter_129: name = "parameter_129" + original_name = "linear_12.w_0" shape = [768, 2304] dtype = "float32" min_val = float("-0.162762") @@ -1420,6 +1550,7 @@ class Program_weight_tensor_parameter_129: class Program_weight_tensor_parameter_130: name = "parameter_130" + original_name = "layer_norm_7.b_0" shape = [768] dtype = "float32" min_val = float("-1.89674") @@ -1431,6 +1562,7 @@ class Program_weight_tensor_parameter_130: class Program_weight_tensor_parameter_131: name = "parameter_131" + original_name = "layer_norm_7.w_0" shape = [768] dtype = "float32" min_val = float("0.364325") @@ -1442,6 +1574,7 @@ class Program_weight_tensor_parameter_131: class Program_weight_tensor_parameter_132: name = "parameter_132" + original_name = "linear_11.b_0" shape = [768] dtype = "float32" min_val = float("-0.575908") @@ -1453,6 +1586,7 @@ class Program_weight_tensor_parameter_132: class Program_weight_tensor_parameter_133: name = "parameter_133" + original_name = "linear_11.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-0.334646") @@ -1464,6 +1598,7 @@ class Program_weight_tensor_parameter_133: class Program_weight_tensor_parameter_134: name = "parameter_134" + original_name = "linear_10.b_0" shape = [3072] dtype = "float32" min_val = float("-1.21946") @@ -1475,6 +1610,7 @@ class Program_weight_tensor_parameter_134: class Program_weight_tensor_parameter_135: name = "parameter_135" + original_name = "linear_10.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.21773") @@ -1486,6 +1622,7 @@ class Program_weight_tensor_parameter_135: class Program_weight_tensor_parameter_136: name = "parameter_136" + original_name = "layer_norm_6.b_0" shape = [768] dtype = "float32" min_val = float("-2.20074") @@ -1497,6 +1634,7 @@ class Program_weight_tensor_parameter_136: class Program_weight_tensor_parameter_137: name = "parameter_137" + original_name = "layer_norm_6.w_0" shape = [768] dtype = "float32" min_val = float("0.320372") @@ -1508,6 +1646,7 @@ class Program_weight_tensor_parameter_137: class Program_weight_tensor_parameter_138: name = "parameter_138" + original_name = "linear_9.b_0" shape = [768] dtype = "float32" min_val = float("-1.02768") @@ -1519,6 +1658,7 @@ class Program_weight_tensor_parameter_138: class Program_weight_tensor_parameter_139: name = "parameter_139" + original_name = "linear_9.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.105803") @@ -1530,6 +1670,7 @@ class Program_weight_tensor_parameter_139: class Program_weight_tensor_parameter_140: name = "parameter_140" + original_name = "linear_8.b_0" shape = [2304] dtype = "float32" min_val = float("-2.36277") @@ -1541,6 +1682,7 @@ class Program_weight_tensor_parameter_140: class Program_weight_tensor_parameter_141: name = "parameter_141" + original_name = "linear_8.w_0" shape = [768, 2304] dtype = "float32" min_val = float("-0.164516") @@ -1552,6 +1694,7 @@ class Program_weight_tensor_parameter_141: class Program_weight_tensor_parameter_142: name = "parameter_142" + original_name = "layer_norm_5.b_0" shape = [768] dtype = "float32" min_val = float("-1.94278") @@ -1563,6 +1706,7 @@ class Program_weight_tensor_parameter_142: class Program_weight_tensor_parameter_143: name = "parameter_143" + original_name = "layer_norm_5.w_0" shape = [768] dtype = "float32" min_val = float("0.147256") @@ -1574,6 +1718,7 @@ class Program_weight_tensor_parameter_143: class Program_weight_tensor_parameter_144: name = "parameter_144" + original_name = "linear_7.b_0" shape = [768] dtype = "float32" min_val = float("-0.967616") @@ -1585,6 +1730,7 @@ class Program_weight_tensor_parameter_144: class Program_weight_tensor_parameter_145: name = "parameter_145" + original_name = "linear_7.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-0.323132") @@ -1596,6 +1742,7 @@ class Program_weight_tensor_parameter_145: class Program_weight_tensor_parameter_146: name = "parameter_146" + original_name = "linear_6.b_0" shape = [3072] dtype = "float32" min_val = float("-1.34354") @@ -1607,6 +1754,7 @@ class Program_weight_tensor_parameter_146: class Program_weight_tensor_parameter_147: name = "parameter_147" + original_name = "linear_6.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.198593") @@ -1618,6 +1766,7 @@ class Program_weight_tensor_parameter_147: class Program_weight_tensor_parameter_148: name = "parameter_148" + original_name = "layer_norm_4.b_0" shape = [768] dtype = "float32" min_val = float("-1.52628") @@ -1629,6 +1778,7 @@ class Program_weight_tensor_parameter_148: class Program_weight_tensor_parameter_149: name = "parameter_149" + original_name = "layer_norm_4.w_0" shape = [768] dtype = "float32" min_val = float("0.0635571") @@ -1640,6 +1790,7 @@ class Program_weight_tensor_parameter_149: class Program_weight_tensor_parameter_150: name = "parameter_150" + original_name = "linear_5.b_0" shape = [768] dtype = "float32" min_val = float("-0.993845") @@ -1651,6 +1802,7 @@ class Program_weight_tensor_parameter_150: class Program_weight_tensor_parameter_151: name = "parameter_151" + original_name = "linear_5.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.211366") @@ -1662,6 +1814,7 @@ class Program_weight_tensor_parameter_151: class Program_weight_tensor_parameter_152: name = "parameter_152" + original_name = "linear_4.b_0" shape = [2304] dtype = "float32" min_val = float("-2.13255") @@ -1673,6 +1826,7 @@ class Program_weight_tensor_parameter_152: class Program_weight_tensor_parameter_153: name = "parameter_153" + original_name = "linear_4.w_0" shape = [768, 2304] dtype = "float32" min_val = float("-0.212923") @@ -1684,6 +1838,7 @@ class Program_weight_tensor_parameter_153: class Program_weight_tensor_parameter_154: name = "parameter_154" + original_name = "layer_norm_3.b_0" shape = [768] dtype = "float32" min_val = float("-1.36312") @@ -1695,6 +1850,7 @@ class Program_weight_tensor_parameter_154: class Program_weight_tensor_parameter_155: name = "parameter_155" + original_name = "layer_norm_3.w_0" shape = [768] dtype = "float32" min_val = float("-6.07658e-05") @@ -1706,6 +1862,7 @@ class Program_weight_tensor_parameter_155: class Program_weight_tensor_parameter_156: name = "parameter_156" + original_name = "linear_3.b_0" shape = [768] dtype = "float32" min_val = float("-0.376214") @@ -1717,6 +1874,7 @@ class Program_weight_tensor_parameter_156: class Program_weight_tensor_parameter_157: name = "parameter_157" + original_name = "linear_3.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-0.221614") @@ -1728,6 +1886,7 @@ class Program_weight_tensor_parameter_157: class Program_weight_tensor_parameter_158: name = "parameter_158" + original_name = "linear_2.b_0" shape = [3072] dtype = "float32" min_val = float("-1.48757") @@ -1739,6 +1898,7 @@ class Program_weight_tensor_parameter_158: class Program_weight_tensor_parameter_159: name = "parameter_159" + original_name = "linear_2.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.280627") @@ -1750,6 +1910,7 @@ class Program_weight_tensor_parameter_159: class Program_weight_tensor_parameter_160: name = "parameter_160" + original_name = "layer_norm_2.b_0" shape = [768] dtype = "float32" min_val = float("-1.22237") @@ -1761,6 +1922,7 @@ class Program_weight_tensor_parameter_160: class Program_weight_tensor_parameter_161: name = "parameter_161" + original_name = "layer_norm_2.w_0" shape = [768] dtype = "float32" min_val = float("0.217935") @@ -1772,6 +1934,7 @@ class Program_weight_tensor_parameter_161: class Program_weight_tensor_parameter_162: name = "parameter_162" + original_name = "linear_1.b_0" shape = [768] dtype = "float32" min_val = float("-0.487797") @@ -1783,6 +1946,7 @@ class Program_weight_tensor_parameter_162: class Program_weight_tensor_parameter_163: name = "parameter_163" + original_name = "linear_1.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.185322") @@ -1794,6 +1958,7 @@ class Program_weight_tensor_parameter_163: class Program_weight_tensor_parameter_164: name = "parameter_164" + original_name = "linear_0.b_0" shape = [2304] dtype = "float32" min_val = float("-2.27756") @@ -1805,6 +1970,7 @@ class Program_weight_tensor_parameter_164: class Program_weight_tensor_parameter_165: name = "parameter_165" + original_name = "linear_0.w_0" shape = [768, 2304] dtype = "float32" min_val = float("-0.18819") @@ -1816,6 +1982,7 @@ class Program_weight_tensor_parameter_165: class Program_weight_tensor_parameter_166: name = "parameter_166" + original_name = "layer_norm_1.b_0" shape = [768] dtype = "float32" min_val = float("-1.14852") @@ -1827,6 +1994,7 @@ class Program_weight_tensor_parameter_166: class Program_weight_tensor_parameter_167: name = "parameter_167" + original_name = "layer_norm_1.w_0" shape = [768] dtype = "float32" min_val = float("-0.00204292") @@ -1838,6 +2006,7 @@ class Program_weight_tensor_parameter_167: class Program_weight_tensor_parameter_168: name = "parameter_168" + original_name = "layer_norm_0.b_0" shape = [768] dtype = "float32" min_val = float("-0.304392") @@ -1849,6 +2018,7 @@ class Program_weight_tensor_parameter_168: class Program_weight_tensor_parameter_169: name = "parameter_169" + original_name = "layer_norm_0.w_0" shape = [768] dtype = "float32" min_val = float("-0.00126423") @@ -1860,6 +2030,7 @@ class Program_weight_tensor_parameter_169: class Program_weight_tensor_parameter_170: name = "parameter_170" + original_name = "conv2d_0.w_0" shape = [768, 3, 16, 16] dtype = "float32" min_val = float("-0.14734") diff --git a/paddle_samples/PaddleX/Deeplabv3-R101/subgraph_1/input_meta.py b/paddle_samples/PaddleX/Deeplabv3-R101/subgraph_1/input_meta.py index 0da4d0f1e..d5d8d554c 100644 --- a/paddle_samples/PaddleX/Deeplabv3-R101/subgraph_1/input_meta.py +++ b/paddle_samples/PaddleX/Deeplabv3-R101/subgraph_1/input_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_data_0: name = "data_0" + original_name = "var_0" shape = [2, 3, 512, 1024] dtype = "float32" min_val = float("-1.0") diff --git a/paddle_samples/PaddleX/Deeplabv3-R101/subgraph_1/weight_meta.py b/paddle_samples/PaddleX/Deeplabv3-R101/subgraph_1/weight_meta.py index dc1d70af9..ba685b938 100644 --- a/paddle_samples/PaddleX/Deeplabv3-R101/subgraph_1/weight_meta.py +++ b/paddle_samples/PaddleX/Deeplabv3-R101/subgraph_1/weight_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_parameter_0: name = "parameter_0" + original_name = "conv2d_112.b_0" shape = [2] dtype = "float32" min_val = float("0") @@ -9,6 +10,7 @@ class Program_weight_tensor_parameter_0: class Program_weight_tensor_parameter_1: name = "parameter_1" + original_name = "conv2d_112.w_0" shape = [2, 256, 1, 1] dtype = "float32" min_val = float("-0.237801") @@ -20,6 +22,7 @@ class Program_weight_tensor_parameter_1: class Program_weight_tensor_parameter_2: name = "parameter_2" + original_name = "batch_norm2d_111.b_0" shape = [256] dtype = "float32" data = None @@ -27,26 +30,31 @@ class Program_weight_tensor_parameter_2: class Program_weight_tensor_parameter_3: name = "parameter_3" + original_name = "batch_norm2d_111.w_0" shape = [256] dtype = "float32" min_val = float("1.0") max_val = float("1.0") mean = float("1.0") + std = float("0.0") data = None class Program_weight_tensor_parameter_4: name = "parameter_4" + original_name = "batch_norm2d_111.w_2" shape = [256] dtype = "float32" min_val = float("1.0") max_val = float("1.0") mean = float("1.0") + std = float("0.0") data = None class Program_weight_tensor_parameter_5: name = "parameter_5" + original_name = "batch_norm2d_111.w_1" shape = [256] dtype = "float32" data = None @@ -54,6 +62,7 @@ class Program_weight_tensor_parameter_5: class Program_weight_tensor_parameter_6: name = "parameter_6" + original_name = "conv2d_111.b_0" shape = [256] dtype = "float32" data = None @@ -61,6 +70,7 @@ class Program_weight_tensor_parameter_6: class Program_weight_tensor_parameter_7: name = "parameter_7" + original_name = "conv2d_111.w_0" shape = [256, 1280, 1, 1] dtype = "float32" min_val = float("-0.179873") @@ -72,6 +82,7 @@ class Program_weight_tensor_parameter_7: class Program_weight_tensor_parameter_8: name = "parameter_8" + original_name = "batch_norm2d_110.b_0" shape = [256] dtype = "float32" data = None @@ -79,26 +90,31 @@ class Program_weight_tensor_parameter_8: class Program_weight_tensor_parameter_9: name = "parameter_9" + original_name = "batch_norm2d_110.w_0" shape = [256] dtype = "float32" min_val = float("1.0") max_val = float("1.0") mean = float("1.0") + std = float("0.0") data = None class Program_weight_tensor_parameter_10: name = "parameter_10" + original_name = "batch_norm2d_110.w_2" shape = [256] dtype = "float32" min_val = float("1.0") max_val = float("1.0") mean = float("1.0") + std = float("0.0") data = None class Program_weight_tensor_parameter_11: name = "parameter_11" + original_name = "batch_norm2d_110.w_1" shape = [256] dtype = "float32" data = None @@ -106,6 +122,7 @@ class Program_weight_tensor_parameter_11: class Program_weight_tensor_parameter_12: name = "parameter_12" + original_name = "conv2d_110.w_0" shape = [256, 2048, 1, 1] dtype = "float32" min_val = float("-0.15373") @@ -117,6 +134,7 @@ class Program_weight_tensor_parameter_12: class Program_weight_tensor_parameter_13: name = "parameter_13" + original_name = "batch_norm2d_109.b_0" shape = [256] dtype = "float32" data = None @@ -124,26 +142,31 @@ class Program_weight_tensor_parameter_13: class Program_weight_tensor_parameter_14: name = "parameter_14" + original_name = "batch_norm2d_109.w_0" shape = [256] dtype = "float32" min_val = float("1.0") max_val = float("1.0") mean = float("1.0") + std = float("0.0") data = None class Program_weight_tensor_parameter_15: name = "parameter_15" + original_name = "batch_norm2d_109.w_2" shape = [256] dtype = "float32" min_val = float("1.0") max_val = float("1.0") mean = float("1.0") + std = float("0.0") data = None class Program_weight_tensor_parameter_16: name = "parameter_16" + original_name = "batch_norm2d_109.w_1" shape = [256] dtype = "float32" data = None @@ -151,6 +174,7 @@ class Program_weight_tensor_parameter_16: class Program_weight_tensor_parameter_17: name = "parameter_17" + original_name = "conv2d_109.b_0" shape = [256] dtype = "float32" data = None @@ -158,6 +182,7 @@ class Program_weight_tensor_parameter_17: class Program_weight_tensor_parameter_18: name = "parameter_18" + original_name = "conv2d_109.w_0" shape = [256, 2048, 3, 3] dtype = "float32" min_val = float("-0.0521362") @@ -169,6 +194,7 @@ class Program_weight_tensor_parameter_18: class Program_weight_tensor_parameter_19: name = "parameter_19" + original_name = "batch_norm2d_108.b_0" shape = [256] dtype = "float32" data = None @@ -176,26 +202,31 @@ class Program_weight_tensor_parameter_19: class Program_weight_tensor_parameter_20: name = "parameter_20" + original_name = "batch_norm2d_108.w_0" shape = [256] dtype = "float32" min_val = float("1.0") max_val = float("1.0") mean = float("1.0") + std = float("0.0") data = None class Program_weight_tensor_parameter_21: name = "parameter_21" + original_name = "batch_norm2d_108.w_2" shape = [256] dtype = "float32" min_val = float("1.0") max_val = float("1.0") mean = float("1.0") + std = float("0.0") data = None class Program_weight_tensor_parameter_22: name = "parameter_22" + original_name = "batch_norm2d_108.w_1" shape = [256] dtype = "float32" data = None @@ -203,6 +234,7 @@ class Program_weight_tensor_parameter_22: class Program_weight_tensor_parameter_23: name = "parameter_23" + original_name = "conv2d_108.b_0" shape = [256] dtype = "float32" data = None @@ -210,6 +242,7 @@ class Program_weight_tensor_parameter_23: class Program_weight_tensor_parameter_24: name = "parameter_24" + original_name = "conv2d_108.w_0" shape = [256, 2048, 3, 3] dtype = "float32" min_val = float("-0.0529906") @@ -221,6 +254,7 @@ class Program_weight_tensor_parameter_24: class Program_weight_tensor_parameter_25: name = "parameter_25" + original_name = "batch_norm2d_107.b_0" shape = [256] dtype = "float32" data = None @@ -228,26 +262,31 @@ class Program_weight_tensor_parameter_25: class Program_weight_tensor_parameter_26: name = "parameter_26" + original_name = "batch_norm2d_107.w_0" shape = [256] dtype = "float32" min_val = float("1.0") max_val = float("1.0") mean = float("1.0") + std = float("0.0") data = None class Program_weight_tensor_parameter_27: name = "parameter_27" + original_name = "batch_norm2d_107.w_2" shape = [256] dtype = "float32" min_val = float("1.0") max_val = float("1.0") mean = float("1.0") + std = float("0.0") data = None class Program_weight_tensor_parameter_28: name = "parameter_28" + original_name = "batch_norm2d_107.w_1" shape = [256] dtype = "float32" data = None @@ -255,6 +294,7 @@ class Program_weight_tensor_parameter_28: class Program_weight_tensor_parameter_29: name = "parameter_29" + original_name = "conv2d_107.b_0" shape = [256] dtype = "float32" data = None @@ -262,6 +302,7 @@ class Program_weight_tensor_parameter_29: class Program_weight_tensor_parameter_30: name = "parameter_30" + original_name = "conv2d_107.w_0" shape = [256, 2048, 3, 3] dtype = "float32" min_val = float("-0.0540393") @@ -273,6 +314,7 @@ class Program_weight_tensor_parameter_30: class Program_weight_tensor_parameter_31: name = "parameter_31" + original_name = "batch_norm2d_106.b_0" shape = [256] dtype = "float32" data = None @@ -280,26 +322,31 @@ class Program_weight_tensor_parameter_31: class Program_weight_tensor_parameter_32: name = "parameter_32" + original_name = "batch_norm2d_106.w_0" shape = [256] dtype = "float32" min_val = float("1.0") max_val = float("1.0") mean = float("1.0") + std = float("0.0") data = None class Program_weight_tensor_parameter_33: name = "parameter_33" + original_name = "batch_norm2d_106.w_2" shape = [256] dtype = "float32" min_val = float("1.0") max_val = float("1.0") mean = float("1.0") + std = float("0.0") data = None class Program_weight_tensor_parameter_34: name = "parameter_34" + original_name = "batch_norm2d_106.w_1" shape = [256] dtype = "float32" data = None @@ -307,6 +354,7 @@ class Program_weight_tensor_parameter_34: class Program_weight_tensor_parameter_35: name = "parameter_35" + original_name = "conv2d_106.b_0" shape = [256] dtype = "float32" data = None @@ -314,6 +362,7 @@ class Program_weight_tensor_parameter_35: class Program_weight_tensor_parameter_36: name = "parameter_36" + original_name = "conv2d_106.w_0" shape = [256, 2048, 1, 1] dtype = "float32" min_val = float("-0.148355") @@ -325,6 +374,7 @@ class Program_weight_tensor_parameter_36: class Program_weight_tensor_parameter_37: name = "parameter_37" + original_name = "batch_norm2d_105.b_0" shape = [2048] dtype = "float32" min_val = float("0.0827274") @@ -336,6 +386,7 @@ class Program_weight_tensor_parameter_37: class Program_weight_tensor_parameter_38: name = "parameter_38" + original_name = "batch_norm2d_105.w_0" shape = [2048] dtype = "float32" min_val = float("0.398192") @@ -347,6 +398,7 @@ class Program_weight_tensor_parameter_38: class Program_weight_tensor_parameter_39: name = "parameter_39" + original_name = "batch_norm2d_105.w_2" shape = [2048] dtype = "float32" min_val = float("0.000114954") @@ -358,6 +410,7 @@ class Program_weight_tensor_parameter_39: class Program_weight_tensor_parameter_40: name = "parameter_40" + original_name = "batch_norm2d_105.w_1" shape = [2048] dtype = "float32" min_val = float("-0.0425386") @@ -369,6 +422,7 @@ class Program_weight_tensor_parameter_40: class Program_weight_tensor_parameter_41: name = "parameter_41" + original_name = "conv2d_105.w_0" shape = [2048, 512, 1, 1] dtype = "float32" min_val = float("-0.133717") @@ -380,6 +434,7 @@ class Program_weight_tensor_parameter_41: class Program_weight_tensor_parameter_42: name = "parameter_42" + original_name = "batch_norm2d_104.b_0" shape = [512] dtype = "float32" min_val = float("-0.216342") @@ -391,6 +446,7 @@ class Program_weight_tensor_parameter_42: class Program_weight_tensor_parameter_43: name = "parameter_43" + original_name = "batch_norm2d_104.w_0" shape = [512] dtype = "float32" min_val = float("0.137888") @@ -402,6 +458,7 @@ class Program_weight_tensor_parameter_43: class Program_weight_tensor_parameter_44: name = "parameter_44" + original_name = "batch_norm2d_104.w_2" shape = [512] dtype = "float32" min_val = float("0.000717996") @@ -413,6 +470,7 @@ class Program_weight_tensor_parameter_44: class Program_weight_tensor_parameter_45: name = "parameter_45" + original_name = "batch_norm2d_104.w_1" shape = [512] dtype = "float32" min_val = float("-0.0644819") @@ -424,6 +482,7 @@ class Program_weight_tensor_parameter_45: class Program_weight_tensor_parameter_46: name = "parameter_46" + original_name = "conv2d_104.w_0" shape = [512, 512, 3, 3] dtype = "float32" min_val = float("-0.17946") @@ -435,6 +494,7 @@ class Program_weight_tensor_parameter_46: class Program_weight_tensor_parameter_47: name = "parameter_47" + original_name = "batch_norm2d_103.b_0" shape = [512] dtype = "float32" min_val = float("-0.273174") @@ -446,6 +506,7 @@ class Program_weight_tensor_parameter_47: class Program_weight_tensor_parameter_48: name = "parameter_48" + original_name = "batch_norm2d_103.w_0" shape = [512] dtype = "float32" min_val = float("0.0716476") @@ -457,6 +518,7 @@ class Program_weight_tensor_parameter_48: class Program_weight_tensor_parameter_49: name = "parameter_49" + original_name = "batch_norm2d_103.w_2" shape = [512] dtype = "float32" min_val = float("0.000327466") @@ -468,6 +530,7 @@ class Program_weight_tensor_parameter_49: class Program_weight_tensor_parameter_50: name = "parameter_50" + original_name = "batch_norm2d_103.w_1" shape = [512] dtype = "float32" min_val = float("-0.0710303") @@ -479,6 +542,7 @@ class Program_weight_tensor_parameter_50: class Program_weight_tensor_parameter_51: name = "parameter_51" + original_name = "conv2d_103.w_0" shape = [512, 2048, 1, 1] dtype = "float32" min_val = float("-0.526026") @@ -490,6 +554,7 @@ class Program_weight_tensor_parameter_51: class Program_weight_tensor_parameter_52: name = "parameter_52" + original_name = "batch_norm2d_102.b_0" shape = [2048] dtype = "float32" min_val = float("-0.291208") @@ -501,6 +566,7 @@ class Program_weight_tensor_parameter_52: class Program_weight_tensor_parameter_53: name = "parameter_53" + original_name = "batch_norm2d_102.w_0" shape = [2048] dtype = "float32" min_val = float("-0.0102351") @@ -512,6 +578,7 @@ class Program_weight_tensor_parameter_53: class Program_weight_tensor_parameter_54: name = "parameter_54" + original_name = "batch_norm2d_102.w_2" shape = [2048] dtype = "float32" min_val = float("7.00605e-05") @@ -523,6 +590,7 @@ class Program_weight_tensor_parameter_54: class Program_weight_tensor_parameter_55: name = "parameter_55" + original_name = "batch_norm2d_102.w_1" shape = [2048] dtype = "float32" min_val = float("-0.0579006") @@ -534,6 +602,7 @@ class Program_weight_tensor_parameter_55: class Program_weight_tensor_parameter_56: name = "parameter_56" + original_name = "conv2d_102.w_0" shape = [2048, 512, 1, 1] dtype = "float32" min_val = float("-0.293566") @@ -545,6 +614,7 @@ class Program_weight_tensor_parameter_56: class Program_weight_tensor_parameter_57: name = "parameter_57" + original_name = "batch_norm2d_101.b_0" shape = [512] dtype = "float32" min_val = float("-0.263418") @@ -556,6 +626,7 @@ class Program_weight_tensor_parameter_57: class Program_weight_tensor_parameter_58: name = "parameter_58" + original_name = "batch_norm2d_101.w_0" shape = [512] dtype = "float32" min_val = float("0.0750012") @@ -567,6 +638,7 @@ class Program_weight_tensor_parameter_58: class Program_weight_tensor_parameter_59: name = "parameter_59" + original_name = "batch_norm2d_101.w_2" shape = [512] dtype = "float32" min_val = float("0.00142611") @@ -578,6 +650,7 @@ class Program_weight_tensor_parameter_59: class Program_weight_tensor_parameter_60: name = "parameter_60" + original_name = "batch_norm2d_101.w_1" shape = [512] dtype = "float32" min_val = float("-0.264804") @@ -589,6 +662,7 @@ class Program_weight_tensor_parameter_60: class Program_weight_tensor_parameter_61: name = "parameter_61" + original_name = "conv2d_101.w_0" shape = [512, 512, 3, 3] dtype = "float32" min_val = float("-0.359546") @@ -600,6 +674,7 @@ class Program_weight_tensor_parameter_61: class Program_weight_tensor_parameter_62: name = "parameter_62" + original_name = "batch_norm2d_100.b_0" shape = [512] dtype = "float32" min_val = float("-0.197871") @@ -611,6 +686,7 @@ class Program_weight_tensor_parameter_62: class Program_weight_tensor_parameter_63: name = "parameter_63" + original_name = "batch_norm2d_100.w_0" shape = [512] dtype = "float32" min_val = float("0.0754074") @@ -622,6 +698,7 @@ class Program_weight_tensor_parameter_63: class Program_weight_tensor_parameter_64: name = "parameter_64" + original_name = "batch_norm2d_100.w_2" shape = [512] dtype = "float32" min_val = float("0.000242041") @@ -633,6 +710,7 @@ class Program_weight_tensor_parameter_64: class Program_weight_tensor_parameter_65: name = "parameter_65" + original_name = "batch_norm2d_100.w_1" shape = [512] dtype = "float32" min_val = float("-0.0453947") @@ -644,6 +722,7 @@ class Program_weight_tensor_parameter_65: class Program_weight_tensor_parameter_66: name = "parameter_66" + original_name = "conv2d_100.w_0" shape = [512, 2048, 1, 1] dtype = "float32" min_val = float("-0.301798") @@ -655,6 +734,7 @@ class Program_weight_tensor_parameter_66: class Program_weight_tensor_parameter_67: name = "parameter_67" + original_name = "batch_norm2d_99.b_0" shape = [2048] dtype = "float32" min_val = float("-0.148764") @@ -666,6 +746,7 @@ class Program_weight_tensor_parameter_67: class Program_weight_tensor_parameter_68: name = "parameter_68" + original_name = "batch_norm2d_99.w_0" shape = [2048] dtype = "float32" min_val = float("0.0435349") @@ -677,6 +758,7 @@ class Program_weight_tensor_parameter_68: class Program_weight_tensor_parameter_69: name = "parameter_69" + original_name = "batch_norm2d_99.w_2" shape = [2048] dtype = "float32" min_val = float("0.000769876") @@ -688,6 +770,7 @@ class Program_weight_tensor_parameter_69: class Program_weight_tensor_parameter_70: name = "parameter_70" + original_name = "batch_norm2d_99.w_1" shape = [2048] dtype = "float32" min_val = float("-0.352413") @@ -699,6 +782,7 @@ class Program_weight_tensor_parameter_70: class Program_weight_tensor_parameter_71: name = "parameter_71" + original_name = "conv2d_99.w_0" shape = [2048, 1024, 1, 1] dtype = "float32" min_val = float("-0.746467") @@ -710,6 +794,7 @@ class Program_weight_tensor_parameter_71: class Program_weight_tensor_parameter_72: name = "parameter_72" + original_name = "batch_norm2d_98.b_0" shape = [2048] dtype = "float32" min_val = float("-0.148764") @@ -721,6 +806,7 @@ class Program_weight_tensor_parameter_72: class Program_weight_tensor_parameter_73: name = "parameter_73" + original_name = "batch_norm2d_98.w_0" shape = [2048] dtype = "float32" min_val = float("-0.101896") @@ -732,6 +818,7 @@ class Program_weight_tensor_parameter_73: class Program_weight_tensor_parameter_74: name = "parameter_74" + original_name = "batch_norm2d_98.w_2" shape = [2048] dtype = "float32" min_val = float("4.32131e-05") @@ -743,6 +830,7 @@ class Program_weight_tensor_parameter_74: class Program_weight_tensor_parameter_75: name = "parameter_75" + original_name = "batch_norm2d_98.w_1" shape = [2048] dtype = "float32" min_val = float("-0.0303601") @@ -754,6 +842,7 @@ class Program_weight_tensor_parameter_75: class Program_weight_tensor_parameter_76: name = "parameter_76" + original_name = "conv2d_98.w_0" shape = [2048, 512, 1, 1] dtype = "float32" min_val = float("-0.100081") @@ -765,6 +854,7 @@ class Program_weight_tensor_parameter_76: class Program_weight_tensor_parameter_77: name = "parameter_77" + original_name = "batch_norm2d_97.b_0" shape = [512] dtype = "float32" min_val = float("-0.131196") @@ -776,6 +866,7 @@ class Program_weight_tensor_parameter_77: class Program_weight_tensor_parameter_78: name = "parameter_78" + original_name = "batch_norm2d_97.w_0" shape = [512] dtype = "float32" min_val = float("0.0656177") @@ -787,6 +878,7 @@ class Program_weight_tensor_parameter_78: class Program_weight_tensor_parameter_79: name = "parameter_79" + original_name = "batch_norm2d_97.w_2" shape = [512] dtype = "float32" min_val = float("0.000436737") @@ -798,6 +890,7 @@ class Program_weight_tensor_parameter_79: class Program_weight_tensor_parameter_80: name = "parameter_80" + original_name = "batch_norm2d_97.w_1" shape = [512] dtype = "float32" min_val = float("-0.333681") @@ -809,6 +902,7 @@ class Program_weight_tensor_parameter_80: class Program_weight_tensor_parameter_81: name = "parameter_81" + original_name = "conv2d_97.w_0" shape = [512, 512, 3, 3] dtype = "float32" min_val = float("-0.354123") @@ -820,6 +914,7 @@ class Program_weight_tensor_parameter_81: class Program_weight_tensor_parameter_82: name = "parameter_82" + original_name = "batch_norm2d_96.b_0" shape = [512] dtype = "float32" min_val = float("-0.195116") @@ -831,6 +926,7 @@ class Program_weight_tensor_parameter_82: class Program_weight_tensor_parameter_83: name = "parameter_83" + original_name = "batch_norm2d_96.w_0" shape = [512] dtype = "float32" min_val = float("0.0903273") @@ -842,6 +938,7 @@ class Program_weight_tensor_parameter_83: class Program_weight_tensor_parameter_84: name = "parameter_84" + original_name = "batch_norm2d_96.w_2" shape = [512] dtype = "float32" min_val = float("0.0020334") @@ -853,6 +950,7 @@ class Program_weight_tensor_parameter_84: class Program_weight_tensor_parameter_85: name = "parameter_85" + original_name = "batch_norm2d_96.w_1" shape = [512] dtype = "float32" min_val = float("-0.18973") @@ -864,6 +962,7 @@ class Program_weight_tensor_parameter_85: class Program_weight_tensor_parameter_86: name = "parameter_86" + original_name = "conv2d_96.w_0" shape = [512, 1024, 1, 1] dtype = "float32" min_val = float("-0.100513") @@ -875,6 +974,7 @@ class Program_weight_tensor_parameter_86: class Program_weight_tensor_parameter_87: name = "parameter_87" + original_name = "batch_norm2d_95.b_0" shape = [1024] dtype = "float32" min_val = float("-0.206403") @@ -886,6 +986,7 @@ class Program_weight_tensor_parameter_87: class Program_weight_tensor_parameter_88: name = "parameter_88" + original_name = "batch_norm2d_95.w_0" shape = [1024] dtype = "float32" min_val = float("-0.180513") @@ -897,6 +998,7 @@ class Program_weight_tensor_parameter_88: class Program_weight_tensor_parameter_89: name = "parameter_89" + original_name = "batch_norm2d_95.w_2" shape = [1024] dtype = "float32" min_val = float("1.12905e-05") @@ -908,6 +1010,7 @@ class Program_weight_tensor_parameter_89: class Program_weight_tensor_parameter_90: name = "parameter_90" + original_name = "batch_norm2d_95.w_1" shape = [1024] dtype = "float32" min_val = float("-0.027597") @@ -919,6 +1022,7 @@ class Program_weight_tensor_parameter_90: class Program_weight_tensor_parameter_91: name = "parameter_91" + original_name = "conv2d_95.w_0" shape = [1024, 256, 1, 1] dtype = "float32" min_val = float("-0.18543") @@ -930,6 +1034,7 @@ class Program_weight_tensor_parameter_91: class Program_weight_tensor_parameter_92: name = "parameter_92" + original_name = "batch_norm2d_94.b_0" shape = [256] dtype = "float32" min_val = float("-0.265797") @@ -941,6 +1046,7 @@ class Program_weight_tensor_parameter_92: class Program_weight_tensor_parameter_93: name = "parameter_93" + original_name = "batch_norm2d_94.w_0" shape = [256] dtype = "float32" min_val = float("0.071971") @@ -952,6 +1058,7 @@ class Program_weight_tensor_parameter_93: class Program_weight_tensor_parameter_94: name = "parameter_94" + original_name = "batch_norm2d_94.w_2" shape = [256] dtype = "float32" min_val = float("0.000303703") @@ -963,6 +1070,7 @@ class Program_weight_tensor_parameter_94: class Program_weight_tensor_parameter_95: name = "parameter_95" + original_name = "batch_norm2d_94.w_1" shape = [256] dtype = "float32" min_val = float("-0.0438911") @@ -974,6 +1082,7 @@ class Program_weight_tensor_parameter_95: class Program_weight_tensor_parameter_96: name = "parameter_96" + original_name = "conv2d_94.w_0" shape = [256, 256, 3, 3] dtype = "float32" min_val = float("-0.146711") @@ -985,6 +1094,7 @@ class Program_weight_tensor_parameter_96: class Program_weight_tensor_parameter_97: name = "parameter_97" + original_name = "batch_norm2d_93.b_0" shape = [256] dtype = "float32" min_val = float("-0.294242") @@ -996,6 +1106,7 @@ class Program_weight_tensor_parameter_97: class Program_weight_tensor_parameter_98: name = "parameter_98" + original_name = "batch_norm2d_93.w_0" shape = [256] dtype = "float32" min_val = float("7.24191e-42") @@ -1007,6 +1118,7 @@ class Program_weight_tensor_parameter_98: class Program_weight_tensor_parameter_99: name = "parameter_99" + original_name = "batch_norm2d_93.w_2" shape = [256] dtype = "float32" min_val = float("5.60519e-45") @@ -1018,6 +1130,7 @@ class Program_weight_tensor_parameter_99: class Program_weight_tensor_parameter_100: name = "parameter_100" + original_name = "batch_norm2d_93.w_1" shape = [256] dtype = "float32" min_val = float("-0.241326") @@ -1029,6 +1142,7 @@ class Program_weight_tensor_parameter_100: class Program_weight_tensor_parameter_101: name = "parameter_101" + original_name = "conv2d_93.w_0" shape = [256, 1024, 1, 1] dtype = "float32" min_val = float("-0.285422") @@ -1040,6 +1154,7 @@ class Program_weight_tensor_parameter_101: class Program_weight_tensor_parameter_102: name = "parameter_102" + original_name = "batch_norm2d_92.b_0" shape = [1024] dtype = "float32" min_val = float("-0.138375") @@ -1051,6 +1166,7 @@ class Program_weight_tensor_parameter_102: class Program_weight_tensor_parameter_103: name = "parameter_103" + original_name = "batch_norm2d_92.w_0" shape = [1024] dtype = "float32" min_val = float("-0.0772375") @@ -1062,6 +1178,7 @@ class Program_weight_tensor_parameter_103: class Program_weight_tensor_parameter_104: name = "parameter_104" + original_name = "batch_norm2d_92.w_2" shape = [1024] dtype = "float32" min_val = float("9.79997e-06") @@ -1073,6 +1190,7 @@ class Program_weight_tensor_parameter_104: class Program_weight_tensor_parameter_105: name = "parameter_105" + original_name = "batch_norm2d_92.w_1" shape = [1024] dtype = "float32" min_val = float("-0.0297335") @@ -1084,6 +1202,7 @@ class Program_weight_tensor_parameter_105: class Program_weight_tensor_parameter_106: name = "parameter_106" + original_name = "conv2d_92.w_0" shape = [1024, 256, 1, 1] dtype = "float32" min_val = float("-0.227452") @@ -1095,6 +1214,7 @@ class Program_weight_tensor_parameter_106: class Program_weight_tensor_parameter_107: name = "parameter_107" + original_name = "batch_norm2d_91.b_0" shape = [256] dtype = "float32" min_val = float("-0.305461") @@ -1106,6 +1226,7 @@ class Program_weight_tensor_parameter_107: class Program_weight_tensor_parameter_108: name = "parameter_108" + original_name = "batch_norm2d_91.w_0" shape = [256] dtype = "float32" min_val = float("0.0733176") @@ -1117,6 +1238,7 @@ class Program_weight_tensor_parameter_108: class Program_weight_tensor_parameter_109: name = "parameter_109" + original_name = "batch_norm2d_91.w_2" shape = [256] dtype = "float32" min_val = float("0.000409322") @@ -1128,6 +1250,7 @@ class Program_weight_tensor_parameter_109: class Program_weight_tensor_parameter_110: name = "parameter_110" + original_name = "batch_norm2d_91.w_1" shape = [256] dtype = "float32" min_val = float("-0.0383431") @@ -1139,6 +1262,7 @@ class Program_weight_tensor_parameter_110: class Program_weight_tensor_parameter_111: name = "parameter_111" + original_name = "conv2d_91.w_0" shape = [256, 256, 3, 3] dtype = "float32" min_val = float("-0.130468") @@ -1150,6 +1274,7 @@ class Program_weight_tensor_parameter_111: class Program_weight_tensor_parameter_112: name = "parameter_112" + original_name = "batch_norm2d_90.b_0" shape = [256] dtype = "float32" min_val = float("-0.286373") @@ -1161,6 +1286,7 @@ class Program_weight_tensor_parameter_112: class Program_weight_tensor_parameter_113: name = "parameter_113" + original_name = "batch_norm2d_90.w_0" shape = [256] dtype = "float32" min_val = float("0.0675077") @@ -1172,6 +1298,7 @@ class Program_weight_tensor_parameter_113: class Program_weight_tensor_parameter_114: name = "parameter_114" + original_name = "batch_norm2d_90.w_2" shape = [256] dtype = "float32" min_val = float("0.00199446") @@ -1183,6 +1310,7 @@ class Program_weight_tensor_parameter_114: class Program_weight_tensor_parameter_115: name = "parameter_115" + original_name = "batch_norm2d_90.w_1" shape = [256] dtype = "float32" min_val = float("-0.152589") @@ -1194,6 +1322,7 @@ class Program_weight_tensor_parameter_115: class Program_weight_tensor_parameter_116: name = "parameter_116" + original_name = "conv2d_90.w_0" shape = [256, 1024, 1, 1] dtype = "float32" min_val = float("-0.131884") @@ -1205,6 +1334,7 @@ class Program_weight_tensor_parameter_116: class Program_weight_tensor_parameter_117: name = "parameter_117" + original_name = "batch_norm2d_89.b_0" shape = [1024] dtype = "float32" min_val = float("-0.192306") @@ -1216,6 +1346,7 @@ class Program_weight_tensor_parameter_117: class Program_weight_tensor_parameter_118: name = "parameter_118" + original_name = "batch_norm2d_89.w_0" shape = [1024] dtype = "float32" min_val = float("-0.101866") @@ -1227,6 +1358,7 @@ class Program_weight_tensor_parameter_118: class Program_weight_tensor_parameter_119: name = "parameter_119" + original_name = "batch_norm2d_89.w_2" shape = [1024] dtype = "float32" min_val = float("2.95398e-07") @@ -1238,6 +1370,7 @@ class Program_weight_tensor_parameter_119: class Program_weight_tensor_parameter_120: name = "parameter_120" + original_name = "batch_norm2d_89.w_1" shape = [1024] dtype = "float32" min_val = float("-0.0293132") @@ -1249,6 +1382,7 @@ class Program_weight_tensor_parameter_120: class Program_weight_tensor_parameter_121: name = "parameter_121" + original_name = "conv2d_89.w_0" shape = [1024, 256, 1, 1] dtype = "float32" min_val = float("-0.18324") @@ -1260,6 +1394,7 @@ class Program_weight_tensor_parameter_121: class Program_weight_tensor_parameter_122: name = "parameter_122" + original_name = "batch_norm2d_88.b_0" shape = [256] dtype = "float32" min_val = float("-0.151589") @@ -1271,6 +1406,7 @@ class Program_weight_tensor_parameter_122: class Program_weight_tensor_parameter_123: name = "parameter_123" + original_name = "batch_norm2d_88.w_0" shape = [256] dtype = "float32" min_val = float("0.0669419") @@ -1282,6 +1418,7 @@ class Program_weight_tensor_parameter_123: class Program_weight_tensor_parameter_124: name = "parameter_124" + original_name = "batch_norm2d_88.w_2" shape = [256] dtype = "float32" min_val = float("0.00035768") @@ -1293,6 +1430,7 @@ class Program_weight_tensor_parameter_124: class Program_weight_tensor_parameter_125: name = "parameter_125" + original_name = "batch_norm2d_88.w_1" shape = [256] dtype = "float32" min_val = float("-0.070715") @@ -1304,6 +1442,7 @@ class Program_weight_tensor_parameter_125: class Program_weight_tensor_parameter_126: name = "parameter_126" + original_name = "conv2d_88.w_0" shape = [256, 256, 3, 3] dtype = "float32" min_val = float("-0.127768") @@ -1315,6 +1454,7 @@ class Program_weight_tensor_parameter_126: class Program_weight_tensor_parameter_127: name = "parameter_127" + original_name = "batch_norm2d_87.b_0" shape = [256] dtype = "float32" min_val = float("-0.308452") @@ -1326,6 +1466,7 @@ class Program_weight_tensor_parameter_127: class Program_weight_tensor_parameter_128: name = "parameter_128" + original_name = "batch_norm2d_87.w_0" shape = [256] dtype = "float32" min_val = float("0.0696605") @@ -1337,6 +1478,7 @@ class Program_weight_tensor_parameter_128: class Program_weight_tensor_parameter_129: name = "parameter_129" + original_name = "batch_norm2d_87.w_2" shape = [256] dtype = "float32" min_val = float("0.00196375") @@ -1348,6 +1490,7 @@ class Program_weight_tensor_parameter_129: class Program_weight_tensor_parameter_130: name = "parameter_130" + original_name = "batch_norm2d_87.w_1" shape = [256] dtype = "float32" min_val = float("-0.183315") @@ -1359,6 +1502,7 @@ class Program_weight_tensor_parameter_130: class Program_weight_tensor_parameter_131: name = "parameter_131" + original_name = "conv2d_87.w_0" shape = [256, 1024, 1, 1] dtype = "float32" min_val = float("-0.327476") @@ -1370,6 +1514,7 @@ class Program_weight_tensor_parameter_131: class Program_weight_tensor_parameter_132: name = "parameter_132" + original_name = "batch_norm2d_86.b_0" shape = [1024] dtype = "float32" min_val = float("-0.157325") @@ -1381,6 +1526,7 @@ class Program_weight_tensor_parameter_132: class Program_weight_tensor_parameter_133: name = "parameter_133" + original_name = "batch_norm2d_86.w_0" shape = [1024] dtype = "float32" min_val = float("-0.155552") @@ -1392,6 +1538,7 @@ class Program_weight_tensor_parameter_133: class Program_weight_tensor_parameter_134: name = "parameter_134" + original_name = "batch_norm2d_86.w_2" shape = [1024] dtype = "float32" min_val = float("4.43334e-07") @@ -1403,6 +1550,7 @@ class Program_weight_tensor_parameter_134: class Program_weight_tensor_parameter_135: name = "parameter_135" + original_name = "batch_norm2d_86.w_1" shape = [1024] dtype = "float32" min_val = float("-0.0202518") @@ -1414,6 +1562,7 @@ class Program_weight_tensor_parameter_135: class Program_weight_tensor_parameter_136: name = "parameter_136" + original_name = "conv2d_86.w_0" shape = [1024, 256, 1, 1] dtype = "float32" min_val = float("-0.165613") @@ -1425,6 +1574,7 @@ class Program_weight_tensor_parameter_136: class Program_weight_tensor_parameter_137: name = "parameter_137" + original_name = "batch_norm2d_85.b_0" shape = [256] dtype = "float32" min_val = float("-0.311108") @@ -1436,6 +1586,7 @@ class Program_weight_tensor_parameter_137: class Program_weight_tensor_parameter_138: name = "parameter_138" + original_name = "batch_norm2d_85.w_0" shape = [256] dtype = "float32" min_val = float("0.0572317") @@ -1447,6 +1598,7 @@ class Program_weight_tensor_parameter_138: class Program_weight_tensor_parameter_139: name = "parameter_139" + original_name = "batch_norm2d_85.w_2" shape = [256] dtype = "float32" min_val = float("0.000339152") @@ -1458,6 +1610,7 @@ class Program_weight_tensor_parameter_139: class Program_weight_tensor_parameter_140: name = "parameter_140" + original_name = "batch_norm2d_85.w_1" shape = [256] dtype = "float32" min_val = float("-0.0636495") @@ -1469,6 +1622,7 @@ class Program_weight_tensor_parameter_140: class Program_weight_tensor_parameter_141: name = "parameter_141" + original_name = "conv2d_85.w_0" shape = [256, 256, 3, 3] dtype = "float32" min_val = float("-0.132645") @@ -1480,6 +1634,7 @@ class Program_weight_tensor_parameter_141: class Program_weight_tensor_parameter_142: name = "parameter_142" + original_name = "batch_norm2d_84.b_0" shape = [256] dtype = "float32" min_val = float("-0.183586") @@ -1491,6 +1646,7 @@ class Program_weight_tensor_parameter_142: class Program_weight_tensor_parameter_143: name = "parameter_143" + original_name = "batch_norm2d_84.w_0" shape = [256] dtype = "float32" min_val = float("0.0546606") @@ -1502,6 +1658,7 @@ class Program_weight_tensor_parameter_143: class Program_weight_tensor_parameter_144: name = "parameter_144" + original_name = "batch_norm2d_84.w_2" shape = [256] dtype = "float32" min_val = float("0.00170534") @@ -1513,6 +1670,7 @@ class Program_weight_tensor_parameter_144: class Program_weight_tensor_parameter_145: name = "parameter_145" + original_name = "batch_norm2d_84.w_1" shape = [256] dtype = "float32" min_val = float("-0.172921") @@ -1524,6 +1682,7 @@ class Program_weight_tensor_parameter_145: class Program_weight_tensor_parameter_146: name = "parameter_146" + original_name = "conv2d_84.w_0" shape = [256, 1024, 1, 1] dtype = "float32" min_val = float("-0.155053") @@ -1535,6 +1694,7 @@ class Program_weight_tensor_parameter_146: class Program_weight_tensor_parameter_147: name = "parameter_147" + original_name = "batch_norm2d_83.b_0" shape = [1024] dtype = "float32" min_val = float("-0.2296") @@ -1546,6 +1706,7 @@ class Program_weight_tensor_parameter_147: class Program_weight_tensor_parameter_148: name = "parameter_148" + original_name = "batch_norm2d_83.w_0" shape = [1024] dtype = "float32" min_val = float("-0.074123") @@ -1557,6 +1718,7 @@ class Program_weight_tensor_parameter_148: class Program_weight_tensor_parameter_149: name = "parameter_149" + original_name = "batch_norm2d_83.w_2" shape = [1024] dtype = "float32" min_val = float("1.59533e-06") @@ -1568,6 +1730,7 @@ class Program_weight_tensor_parameter_149: class Program_weight_tensor_parameter_150: name = "parameter_150" + original_name = "batch_norm2d_83.w_1" shape = [1024] dtype = "float32" min_val = float("-0.11021") @@ -1579,6 +1742,7 @@ class Program_weight_tensor_parameter_150: class Program_weight_tensor_parameter_151: name = "parameter_151" + original_name = "conv2d_83.w_0" shape = [1024, 256, 1, 1] dtype = "float32" min_val = float("-0.479697") @@ -1590,6 +1754,7 @@ class Program_weight_tensor_parameter_151: class Program_weight_tensor_parameter_152: name = "parameter_152" + original_name = "batch_norm2d_82.b_0" shape = [256] dtype = "float32" min_val = float("-0.136562") @@ -1601,6 +1766,7 @@ class Program_weight_tensor_parameter_152: class Program_weight_tensor_parameter_153: name = "parameter_153" + original_name = "batch_norm2d_82.w_0" shape = [256] dtype = "float32" min_val = float("0.0475024") @@ -1612,6 +1778,7 @@ class Program_weight_tensor_parameter_153: class Program_weight_tensor_parameter_154: name = "parameter_154" + original_name = "batch_norm2d_82.w_2" shape = [256] dtype = "float32" min_val = float("0.000246573") @@ -1623,6 +1790,7 @@ class Program_weight_tensor_parameter_154: class Program_weight_tensor_parameter_155: name = "parameter_155" + original_name = "batch_norm2d_82.w_1" shape = [256] dtype = "float32" min_val = float("-0.0781337") @@ -1634,6 +1802,7 @@ class Program_weight_tensor_parameter_155: class Program_weight_tensor_parameter_156: name = "parameter_156" + original_name = "conv2d_82.w_0" shape = [256, 256, 3, 3] dtype = "float32" min_val = float("-0.245573") @@ -1645,6 +1814,7 @@ class Program_weight_tensor_parameter_156: class Program_weight_tensor_parameter_157: name = "parameter_157" + original_name = "batch_norm2d_81.b_0" shape = [256] dtype = "float32" min_val = float("-0.162884") @@ -1656,6 +1826,7 @@ class Program_weight_tensor_parameter_157: class Program_weight_tensor_parameter_158: name = "parameter_158" + original_name = "batch_norm2d_81.w_0" shape = [256] dtype = "float32" min_val = float("0.0544083") @@ -1667,6 +1838,7 @@ class Program_weight_tensor_parameter_158: class Program_weight_tensor_parameter_159: name = "parameter_159" + original_name = "batch_norm2d_81.w_2" shape = [256] dtype = "float32" min_val = float("0.00184023") @@ -1678,6 +1850,7 @@ class Program_weight_tensor_parameter_159: class Program_weight_tensor_parameter_160: name = "parameter_160" + original_name = "batch_norm2d_81.w_1" shape = [256] dtype = "float32" min_val = float("-0.134317") @@ -1689,6 +1862,7 @@ class Program_weight_tensor_parameter_160: class Program_weight_tensor_parameter_161: name = "parameter_161" + original_name = "conv2d_81.w_0" shape = [256, 1024, 1, 1] dtype = "float32" min_val = float("-0.219317") @@ -1700,6 +1874,7 @@ class Program_weight_tensor_parameter_161: class Program_weight_tensor_parameter_162: name = "parameter_162" + original_name = "batch_norm2d_80.b_0" shape = [1024] dtype = "float32" min_val = float("-0.118821") @@ -1711,6 +1886,7 @@ class Program_weight_tensor_parameter_162: class Program_weight_tensor_parameter_163: name = "parameter_163" + original_name = "batch_norm2d_80.w_0" shape = [1024] dtype = "float32" min_val = float("-0.0644223") @@ -1722,6 +1898,7 @@ class Program_weight_tensor_parameter_163: class Program_weight_tensor_parameter_164: name = "parameter_164" + original_name = "batch_norm2d_80.w_2" shape = [1024] dtype = "float32" min_val = float("1.10508e-27") @@ -1733,6 +1910,7 @@ class Program_weight_tensor_parameter_164: class Program_weight_tensor_parameter_165: name = "parameter_165" + original_name = "batch_norm2d_80.w_1" shape = [1024] dtype = "float32" min_val = float("-0.017754") @@ -1744,6 +1922,7 @@ class Program_weight_tensor_parameter_165: class Program_weight_tensor_parameter_166: name = "parameter_166" + original_name = "conv2d_80.w_0" shape = [1024, 256, 1, 1] dtype = "float32" min_val = float("-0.19161") @@ -1755,6 +1934,7 @@ class Program_weight_tensor_parameter_166: class Program_weight_tensor_parameter_167: name = "parameter_167" + original_name = "batch_norm2d_79.b_0" shape = [256] dtype = "float32" min_val = float("-0.128163") @@ -1766,6 +1946,7 @@ class Program_weight_tensor_parameter_167: class Program_weight_tensor_parameter_168: name = "parameter_168" + original_name = "batch_norm2d_79.w_0" shape = [256] dtype = "float32" min_val = float("0.0506961") @@ -1777,6 +1958,7 @@ class Program_weight_tensor_parameter_168: class Program_weight_tensor_parameter_169: name = "parameter_169" + original_name = "batch_norm2d_79.w_2" shape = [256] dtype = "float32" min_val = float("0.000352798") @@ -1788,6 +1970,7 @@ class Program_weight_tensor_parameter_169: class Program_weight_tensor_parameter_170: name = "parameter_170" + original_name = "batch_norm2d_79.w_1" shape = [256] dtype = "float32" min_val = float("-0.0403924") @@ -1799,6 +1982,7 @@ class Program_weight_tensor_parameter_170: class Program_weight_tensor_parameter_171: name = "parameter_171" + original_name = "conv2d_79.w_0" shape = [256, 256, 3, 3] dtype = "float32" min_val = float("-0.13259") @@ -1810,6 +1994,7 @@ class Program_weight_tensor_parameter_171: class Program_weight_tensor_parameter_172: name = "parameter_172" + original_name = "batch_norm2d_78.b_0" shape = [256] dtype = "float32" min_val = float("-0.161042") @@ -1821,6 +2006,7 @@ class Program_weight_tensor_parameter_172: class Program_weight_tensor_parameter_173: name = "parameter_173" + original_name = "batch_norm2d_78.w_0" shape = [256] dtype = "float32" min_val = float("0.0516501") @@ -1832,6 +2018,7 @@ class Program_weight_tensor_parameter_173: class Program_weight_tensor_parameter_174: name = "parameter_174" + original_name = "batch_norm2d_78.w_2" shape = [256] dtype = "float32" min_val = float("0.00200027") @@ -1843,6 +2030,7 @@ class Program_weight_tensor_parameter_174: class Program_weight_tensor_parameter_175: name = "parameter_175" + original_name = "batch_norm2d_78.w_1" shape = [256] dtype = "float32" min_val = float("-0.112165") @@ -1854,6 +2042,7 @@ class Program_weight_tensor_parameter_175: class Program_weight_tensor_parameter_176: name = "parameter_176" + original_name = "conv2d_78.w_0" shape = [256, 1024, 1, 1] dtype = "float32" min_val = float("-0.209198") @@ -1865,6 +2054,7 @@ class Program_weight_tensor_parameter_176: class Program_weight_tensor_parameter_177: name = "parameter_177" + original_name = "batch_norm2d_77.b_0" shape = [1024] dtype = "float32" min_val = float("-0.144365") @@ -1876,6 +2066,7 @@ class Program_weight_tensor_parameter_177: class Program_weight_tensor_parameter_178: name = "parameter_178" + original_name = "batch_norm2d_77.w_0" shape = [1024] dtype = "float32" min_val = float("-0.0877915") @@ -1887,6 +2078,7 @@ class Program_weight_tensor_parameter_178: class Program_weight_tensor_parameter_179: name = "parameter_179" + original_name = "batch_norm2d_77.w_2" shape = [1024] dtype = "float32" min_val = float("5.36924e-07") @@ -1898,6 +2090,7 @@ class Program_weight_tensor_parameter_179: class Program_weight_tensor_parameter_180: name = "parameter_180" + original_name = "batch_norm2d_77.w_1" shape = [1024] dtype = "float32" min_val = float("-0.0667076") @@ -1909,6 +2102,7 @@ class Program_weight_tensor_parameter_180: class Program_weight_tensor_parameter_181: name = "parameter_181" + original_name = "conv2d_77.w_0" shape = [1024, 256, 1, 1] dtype = "float32" min_val = float("-0.373221") @@ -1920,6 +2114,7 @@ class Program_weight_tensor_parameter_181: class Program_weight_tensor_parameter_182: name = "parameter_182" + original_name = "batch_norm2d_76.b_0" shape = [256] dtype = "float32" min_val = float("-0.148653") @@ -1931,6 +2126,7 @@ class Program_weight_tensor_parameter_182: class Program_weight_tensor_parameter_183: name = "parameter_183" + original_name = "batch_norm2d_76.w_0" shape = [256] dtype = "float32" min_val = float("0.0469178") @@ -1942,6 +2138,7 @@ class Program_weight_tensor_parameter_183: class Program_weight_tensor_parameter_184: name = "parameter_184" + original_name = "batch_norm2d_76.w_2" shape = [256] dtype = "float32" min_val = float("0.000240204") @@ -1953,6 +2150,7 @@ class Program_weight_tensor_parameter_184: class Program_weight_tensor_parameter_185: name = "parameter_185" + original_name = "batch_norm2d_76.w_1" shape = [256] dtype = "float32" min_val = float("-0.0457173") @@ -1964,6 +2162,7 @@ class Program_weight_tensor_parameter_185: class Program_weight_tensor_parameter_186: name = "parameter_186" + original_name = "conv2d_76.w_0" shape = [256, 256, 3, 3] dtype = "float32" min_val = float("-0.268807") @@ -1975,6 +2174,7 @@ class Program_weight_tensor_parameter_186: class Program_weight_tensor_parameter_187: name = "parameter_187" + original_name = "batch_norm2d_75.b_0" shape = [256] dtype = "float32" min_val = float("-0.196785") @@ -1986,6 +2186,7 @@ class Program_weight_tensor_parameter_187: class Program_weight_tensor_parameter_188: name = "parameter_188" + original_name = "batch_norm2d_75.w_0" shape = [256] dtype = "float32" min_val = float("0.0504373") @@ -1997,6 +2198,7 @@ class Program_weight_tensor_parameter_188: class Program_weight_tensor_parameter_189: name = "parameter_189" + original_name = "batch_norm2d_75.w_2" shape = [256] dtype = "float32" min_val = float("0.00154363") @@ -2008,6 +2210,7 @@ class Program_weight_tensor_parameter_189: class Program_weight_tensor_parameter_190: name = "parameter_190" + original_name = "batch_norm2d_75.w_1" shape = [256] dtype = "float32" min_val = float("-0.103539") @@ -2019,6 +2222,7 @@ class Program_weight_tensor_parameter_190: class Program_weight_tensor_parameter_191: name = "parameter_191" + original_name = "conv2d_75.w_0" shape = [256, 1024, 1, 1] dtype = "float32" min_val = float("-0.173457") @@ -2030,6 +2234,7 @@ class Program_weight_tensor_parameter_191: class Program_weight_tensor_parameter_192: name = "parameter_192" + original_name = "batch_norm2d_74.b_0" shape = [1024] dtype = "float32" min_val = float("-0.154772") @@ -2041,6 +2246,7 @@ class Program_weight_tensor_parameter_192: class Program_weight_tensor_parameter_193: name = "parameter_193" + original_name = "batch_norm2d_74.w_0" shape = [1024] dtype = "float32" min_val = float("-0.0862494") @@ -2052,6 +2258,7 @@ class Program_weight_tensor_parameter_193: class Program_weight_tensor_parameter_194: name = "parameter_194" + original_name = "batch_norm2d_74.w_2" shape = [1024] dtype = "float32" min_val = float("1.21888e-10") @@ -2063,6 +2270,7 @@ class Program_weight_tensor_parameter_194: class Program_weight_tensor_parameter_195: name = "parameter_195" + original_name = "batch_norm2d_74.w_1" shape = [1024] dtype = "float32" min_val = float("-0.0330924") @@ -2074,6 +2282,7 @@ class Program_weight_tensor_parameter_195: class Program_weight_tensor_parameter_196: name = "parameter_196" + original_name = "conv2d_74.w_0" shape = [1024, 256, 1, 1] dtype = "float32" min_val = float("-0.18676") @@ -2085,6 +2294,7 @@ class Program_weight_tensor_parameter_196: class Program_weight_tensor_parameter_197: name = "parameter_197" + original_name = "batch_norm2d_73.b_0" shape = [256] dtype = "float32" min_val = float("-0.292217") @@ -2096,6 +2306,7 @@ class Program_weight_tensor_parameter_197: class Program_weight_tensor_parameter_198: name = "parameter_198" + original_name = "batch_norm2d_73.w_0" shape = [256] dtype = "float32" min_val = float("0.0585411") @@ -2107,6 +2318,7 @@ class Program_weight_tensor_parameter_198: class Program_weight_tensor_parameter_199: name = "parameter_199" + original_name = "batch_norm2d_73.w_2" shape = [256] dtype = "float32" min_val = float("0.00036765") @@ -2118,6 +2330,7 @@ class Program_weight_tensor_parameter_199: class Program_weight_tensor_parameter_200: name = "parameter_200" + original_name = "batch_norm2d_73.w_1" shape = [256] dtype = "float32" min_val = float("-0.0637849") @@ -2129,6 +2342,7 @@ class Program_weight_tensor_parameter_200: class Program_weight_tensor_parameter_201: name = "parameter_201" + original_name = "conv2d_73.w_0" shape = [256, 256, 3, 3] dtype = "float32" min_val = float("-0.122804") @@ -2140,6 +2354,7 @@ class Program_weight_tensor_parameter_201: class Program_weight_tensor_parameter_202: name = "parameter_202" + original_name = "batch_norm2d_72.b_0" shape = [256] dtype = "float32" min_val = float("-0.1479") @@ -2151,6 +2366,7 @@ class Program_weight_tensor_parameter_202: class Program_weight_tensor_parameter_203: name = "parameter_203" + original_name = "batch_norm2d_72.w_0" shape = [256] dtype = "float32" min_val = float("6.39272e-42") @@ -2162,6 +2378,7 @@ class Program_weight_tensor_parameter_203: class Program_weight_tensor_parameter_204: name = "parameter_204" + original_name = "batch_norm2d_72.w_2" shape = [256] dtype = "float32" min_val = float("5.60519e-45") @@ -2173,6 +2390,7 @@ class Program_weight_tensor_parameter_204: class Program_weight_tensor_parameter_205: name = "parameter_205" + original_name = "batch_norm2d_72.w_1" shape = [256] dtype = "float32" min_val = float("-0.119982") @@ -2184,6 +2402,7 @@ class Program_weight_tensor_parameter_205: class Program_weight_tensor_parameter_206: name = "parameter_206" + original_name = "conv2d_72.w_0" shape = [256, 1024, 1, 1] dtype = "float32" min_val = float("-0.17648") @@ -2195,6 +2414,7 @@ class Program_weight_tensor_parameter_206: class Program_weight_tensor_parameter_207: name = "parameter_207" + original_name = "batch_norm2d_71.b_0" shape = [1024] dtype = "float32" min_val = float("-0.181404") @@ -2206,6 +2426,7 @@ class Program_weight_tensor_parameter_207: class Program_weight_tensor_parameter_208: name = "parameter_208" + original_name = "batch_norm2d_71.w_0" shape = [1024] dtype = "float32" min_val = float("-0.067036") @@ -2217,6 +2438,7 @@ class Program_weight_tensor_parameter_208: class Program_weight_tensor_parameter_209: name = "parameter_209" + original_name = "batch_norm2d_71.w_2" shape = [1024] dtype = "float32" min_val = float("1.25122e-06") @@ -2228,6 +2450,7 @@ class Program_weight_tensor_parameter_209: class Program_weight_tensor_parameter_210: name = "parameter_210" + original_name = "batch_norm2d_71.w_1" shape = [1024] dtype = "float32" min_val = float("-0.0225367") @@ -2239,6 +2462,7 @@ class Program_weight_tensor_parameter_210: class Program_weight_tensor_parameter_211: name = "parameter_211" + original_name = "conv2d_71.w_0" shape = [1024, 256, 1, 1] dtype = "float32" min_val = float("-0.203571") @@ -2250,6 +2474,7 @@ class Program_weight_tensor_parameter_211: class Program_weight_tensor_parameter_212: name = "parameter_212" + original_name = "batch_norm2d_70.b_0" shape = [256] dtype = "float32" min_val = float("-0.142678") @@ -2261,6 +2486,7 @@ class Program_weight_tensor_parameter_212: class Program_weight_tensor_parameter_213: name = "parameter_213" + original_name = "batch_norm2d_70.w_0" shape = [256] dtype = "float32" min_val = float("0.0525285") @@ -2272,6 +2498,7 @@ class Program_weight_tensor_parameter_213: class Program_weight_tensor_parameter_214: name = "parameter_214" + original_name = "batch_norm2d_70.w_2" shape = [256] dtype = "float32" min_val = float("0.000315005") @@ -2283,6 +2510,7 @@ class Program_weight_tensor_parameter_214: class Program_weight_tensor_parameter_215: name = "parameter_215" + original_name = "batch_norm2d_70.w_1" shape = [256] dtype = "float32" min_val = float("-0.0438057") @@ -2294,6 +2522,7 @@ class Program_weight_tensor_parameter_215: class Program_weight_tensor_parameter_216: name = "parameter_216" + original_name = "conv2d_70.w_0" shape = [256, 256, 3, 3] dtype = "float32" min_val = float("-0.157069") @@ -2305,6 +2534,7 @@ class Program_weight_tensor_parameter_216: class Program_weight_tensor_parameter_217: name = "parameter_217" + original_name = "batch_norm2d_69.b_0" shape = [256] dtype = "float32" min_val = float("-0.263298") @@ -2316,6 +2546,7 @@ class Program_weight_tensor_parameter_217: class Program_weight_tensor_parameter_218: name = "parameter_218" + original_name = "batch_norm2d_69.w_0" shape = [256] dtype = "float32" min_val = float("-4.54021e-42") @@ -2327,6 +2558,7 @@ class Program_weight_tensor_parameter_218: class Program_weight_tensor_parameter_219: name = "parameter_219" + original_name = "batch_norm2d_69.w_2" shape = [256] dtype = "float32" min_val = float("5.60519e-45") @@ -2338,6 +2570,7 @@ class Program_weight_tensor_parameter_219: class Program_weight_tensor_parameter_220: name = "parameter_220" + original_name = "batch_norm2d_69.w_1" shape = [256] dtype = "float32" min_val = float("-0.101473") @@ -2349,6 +2582,7 @@ class Program_weight_tensor_parameter_220: class Program_weight_tensor_parameter_221: name = "parameter_221" + original_name = "conv2d_69.w_0" shape = [256, 1024, 1, 1] dtype = "float32" min_val = float("-0.184965") @@ -2360,6 +2594,7 @@ class Program_weight_tensor_parameter_221: class Program_weight_tensor_parameter_222: name = "parameter_222" + original_name = "batch_norm2d_68.b_0" shape = [1024] dtype = "float32" min_val = float("-0.158868") @@ -2371,6 +2606,7 @@ class Program_weight_tensor_parameter_222: class Program_weight_tensor_parameter_223: name = "parameter_223" + original_name = "batch_norm2d_68.w_0" shape = [1024] dtype = "float32" min_val = float("-0.0788502") @@ -2382,6 +2618,7 @@ class Program_weight_tensor_parameter_223: class Program_weight_tensor_parameter_224: name = "parameter_224" + original_name = "batch_norm2d_68.w_2" shape = [1024] dtype = "float32" min_val = float("9.42686e-07") @@ -2393,6 +2630,7 @@ class Program_weight_tensor_parameter_224: class Program_weight_tensor_parameter_225: name = "parameter_225" + original_name = "batch_norm2d_68.w_1" shape = [1024] dtype = "float32" min_val = float("-0.0225805") @@ -2404,6 +2642,7 @@ class Program_weight_tensor_parameter_225: class Program_weight_tensor_parameter_226: name = "parameter_226" + original_name = "conv2d_68.w_0" shape = [1024, 256, 1, 1] dtype = "float32" min_val = float("-0.193111") @@ -2415,6 +2654,7 @@ class Program_weight_tensor_parameter_226: class Program_weight_tensor_parameter_227: name = "parameter_227" + original_name = "batch_norm2d_67.b_0" shape = [256] dtype = "float32" min_val = float("-0.126881") @@ -2426,6 +2666,7 @@ class Program_weight_tensor_parameter_227: class Program_weight_tensor_parameter_228: name = "parameter_228" + original_name = "batch_norm2d_67.w_0" shape = [256] dtype = "float32" min_val = float("0.0538402") @@ -2437,6 +2678,7 @@ class Program_weight_tensor_parameter_228: class Program_weight_tensor_parameter_229: name = "parameter_229" + original_name = "batch_norm2d_67.w_2" shape = [256] dtype = "float32" min_val = float("0.00048507") @@ -2448,6 +2690,7 @@ class Program_weight_tensor_parameter_229: class Program_weight_tensor_parameter_230: name = "parameter_230" + original_name = "batch_norm2d_67.w_1" shape = [256] dtype = "float32" min_val = float("-0.0562529") @@ -2459,6 +2702,7 @@ class Program_weight_tensor_parameter_230: class Program_weight_tensor_parameter_231: name = "parameter_231" + original_name = "conv2d_67.w_0" shape = [256, 256, 3, 3] dtype = "float32" min_val = float("-0.111533") @@ -2470,6 +2714,7 @@ class Program_weight_tensor_parameter_231: class Program_weight_tensor_parameter_232: name = "parameter_232" + original_name = "batch_norm2d_66.b_0" shape = [256] dtype = "float32" min_val = float("-0.177625") @@ -2481,6 +2726,7 @@ class Program_weight_tensor_parameter_232: class Program_weight_tensor_parameter_233: name = "parameter_233" + original_name = "batch_norm2d_66.w_0" shape = [256] dtype = "float32" min_val = float("0.0510496") @@ -2492,6 +2738,7 @@ class Program_weight_tensor_parameter_233: class Program_weight_tensor_parameter_234: name = "parameter_234" + original_name = "batch_norm2d_66.w_2" shape = [256] dtype = "float32" min_val = float("0.00170635") @@ -2503,6 +2750,7 @@ class Program_weight_tensor_parameter_234: class Program_weight_tensor_parameter_235: name = "parameter_235" + original_name = "batch_norm2d_66.w_1" shape = [256] dtype = "float32" min_val = float("-0.147881") @@ -2514,6 +2762,7 @@ class Program_weight_tensor_parameter_235: class Program_weight_tensor_parameter_236: name = "parameter_236" + original_name = "conv2d_66.w_0" shape = [256, 1024, 1, 1] dtype = "float32" min_val = float("-0.190949") @@ -2525,6 +2774,7 @@ class Program_weight_tensor_parameter_236: class Program_weight_tensor_parameter_237: name = "parameter_237" + original_name = "batch_norm2d_65.b_0" shape = [1024] dtype = "float32" min_val = float("-0.169765") @@ -2536,6 +2786,7 @@ class Program_weight_tensor_parameter_237: class Program_weight_tensor_parameter_238: name = "parameter_238" + original_name = "batch_norm2d_65.w_0" shape = [1024] dtype = "float32" min_val = float("-0.112945") @@ -2547,6 +2798,7 @@ class Program_weight_tensor_parameter_238: class Program_weight_tensor_parameter_239: name = "parameter_239" + original_name = "batch_norm2d_65.w_2" shape = [1024] dtype = "float32" min_val = float("5.60519e-45") @@ -2558,6 +2810,7 @@ class Program_weight_tensor_parameter_239: class Program_weight_tensor_parameter_240: name = "parameter_240" + original_name = "batch_norm2d_65.w_1" shape = [1024] dtype = "float32" min_val = float("-0.0260425") @@ -2569,6 +2822,7 @@ class Program_weight_tensor_parameter_240: class Program_weight_tensor_parameter_241: name = "parameter_241" + original_name = "conv2d_65.w_0" shape = [1024, 256, 1, 1] dtype = "float32" min_val = float("-0.151513") @@ -2580,6 +2834,7 @@ class Program_weight_tensor_parameter_241: class Program_weight_tensor_parameter_242: name = "parameter_242" + original_name = "batch_norm2d_64.b_0" shape = [256] dtype = "float32" min_val = float("-0.119003") @@ -2591,6 +2846,7 @@ class Program_weight_tensor_parameter_242: class Program_weight_tensor_parameter_243: name = "parameter_243" + original_name = "batch_norm2d_64.w_0" shape = [256] dtype = "float32" min_val = float("0.0584529") @@ -2602,6 +2858,7 @@ class Program_weight_tensor_parameter_243: class Program_weight_tensor_parameter_244: name = "parameter_244" + original_name = "batch_norm2d_64.w_2" shape = [256] dtype = "float32" min_val = float("0.000528709") @@ -2613,6 +2870,7 @@ class Program_weight_tensor_parameter_244: class Program_weight_tensor_parameter_245: name = "parameter_245" + original_name = "batch_norm2d_64.w_1" shape = [256] dtype = "float32" min_val = float("-0.0602747") @@ -2624,6 +2882,7 @@ class Program_weight_tensor_parameter_245: class Program_weight_tensor_parameter_246: name = "parameter_246" + original_name = "conv2d_64.w_0" shape = [256, 256, 3, 3] dtype = "float32" min_val = float("-0.0802216") @@ -2635,6 +2894,7 @@ class Program_weight_tensor_parameter_246: class Program_weight_tensor_parameter_247: name = "parameter_247" + original_name = "batch_norm2d_63.b_0" shape = [256] dtype = "float32" min_val = float("-0.13678") @@ -2646,6 +2906,7 @@ class Program_weight_tensor_parameter_247: class Program_weight_tensor_parameter_248: name = "parameter_248" + original_name = "batch_norm2d_63.w_0" shape = [256] dtype = "float32" min_val = float("0.0602401") @@ -2657,6 +2918,7 @@ class Program_weight_tensor_parameter_248: class Program_weight_tensor_parameter_249: name = "parameter_249" + original_name = "batch_norm2d_63.w_2" shape = [256] dtype = "float32" min_val = float("0.00184796") @@ -2668,6 +2930,7 @@ class Program_weight_tensor_parameter_249: class Program_weight_tensor_parameter_250: name = "parameter_250" + original_name = "batch_norm2d_63.w_1" shape = [256] dtype = "float32" min_val = float("-0.112543") @@ -2679,6 +2942,7 @@ class Program_weight_tensor_parameter_250: class Program_weight_tensor_parameter_251: name = "parameter_251" + original_name = "conv2d_63.w_0" shape = [256, 1024, 1, 1] dtype = "float32" min_val = float("-0.159615") @@ -2690,6 +2954,7 @@ class Program_weight_tensor_parameter_251: class Program_weight_tensor_parameter_252: name = "parameter_252" + original_name = "batch_norm2d_62.b_0" shape = [1024] dtype = "float32" min_val = float("-0.119579") @@ -2701,6 +2966,7 @@ class Program_weight_tensor_parameter_252: class Program_weight_tensor_parameter_253: name = "parameter_253" + original_name = "batch_norm2d_62.w_0" shape = [1024] dtype = "float32" min_val = float("-0.111251") @@ -2712,6 +2978,7 @@ class Program_weight_tensor_parameter_253: class Program_weight_tensor_parameter_254: name = "parameter_254" + original_name = "batch_norm2d_62.w_2" shape = [1024] dtype = "float32" min_val = float("5.60519e-45") @@ -2723,6 +2990,7 @@ class Program_weight_tensor_parameter_254: class Program_weight_tensor_parameter_255: name = "parameter_255" + original_name = "batch_norm2d_62.w_1" shape = [1024] dtype = "float32" min_val = float("-0.0229368") @@ -2734,6 +3002,7 @@ class Program_weight_tensor_parameter_255: class Program_weight_tensor_parameter_256: name = "parameter_256" + original_name = "conv2d_62.w_0" shape = [1024, 256, 1, 1] dtype = "float32" min_val = float("-0.135953") @@ -2745,6 +3014,7 @@ class Program_weight_tensor_parameter_256: class Program_weight_tensor_parameter_257: name = "parameter_257" + original_name = "batch_norm2d_61.b_0" shape = [256] dtype = "float32" min_val = float("-0.142617") @@ -2756,6 +3026,7 @@ class Program_weight_tensor_parameter_257: class Program_weight_tensor_parameter_258: name = "parameter_258" + original_name = "batch_norm2d_61.w_0" shape = [256] dtype = "float32" min_val = float("0.0604692") @@ -2767,6 +3038,7 @@ class Program_weight_tensor_parameter_258: class Program_weight_tensor_parameter_259: name = "parameter_259" + original_name = "batch_norm2d_61.w_2" shape = [256] dtype = "float32" min_val = float("0.000498543") @@ -2778,6 +3050,7 @@ class Program_weight_tensor_parameter_259: class Program_weight_tensor_parameter_260: name = "parameter_260" + original_name = "batch_norm2d_61.w_1" shape = [256] dtype = "float32" min_val = float("-0.158943") @@ -2789,6 +3062,7 @@ class Program_weight_tensor_parameter_260: class Program_weight_tensor_parameter_261: name = "parameter_261" + original_name = "conv2d_61.w_0" shape = [256, 256, 3, 3] dtype = "float32" min_val = float("-0.101181") @@ -2800,6 +3074,7 @@ class Program_weight_tensor_parameter_261: class Program_weight_tensor_parameter_262: name = "parameter_262" + original_name = "batch_norm2d_60.b_0" shape = [256] dtype = "float32" min_val = float("-0.147466") @@ -2811,6 +3086,7 @@ class Program_weight_tensor_parameter_262: class Program_weight_tensor_parameter_263: name = "parameter_263" + original_name = "batch_norm2d_60.w_0" shape = [256] dtype = "float32" min_val = float("0.0493016") @@ -2822,6 +3098,7 @@ class Program_weight_tensor_parameter_263: class Program_weight_tensor_parameter_264: name = "parameter_264" + original_name = "batch_norm2d_60.w_2" shape = [256] dtype = "float32" min_val = float("0.0015957") @@ -2833,6 +3110,7 @@ class Program_weight_tensor_parameter_264: class Program_weight_tensor_parameter_265: name = "parameter_265" + original_name = "batch_norm2d_60.w_1" shape = [256] dtype = "float32" min_val = float("-0.176171") @@ -2844,6 +3122,7 @@ class Program_weight_tensor_parameter_265: class Program_weight_tensor_parameter_266: name = "parameter_266" + original_name = "conv2d_60.w_0" shape = [256, 1024, 1, 1] dtype = "float32" min_val = float("-0.269775") @@ -2855,6 +3134,7 @@ class Program_weight_tensor_parameter_266: class Program_weight_tensor_parameter_267: name = "parameter_267" + original_name = "batch_norm2d_59.b_0" shape = [1024] dtype = "float32" min_val = float("-0.157031") @@ -2866,6 +3146,7 @@ class Program_weight_tensor_parameter_267: class Program_weight_tensor_parameter_268: name = "parameter_268" + original_name = "batch_norm2d_59.w_0" shape = [1024] dtype = "float32" min_val = float("-0.0900119") @@ -2877,6 +3158,7 @@ class Program_weight_tensor_parameter_268: class Program_weight_tensor_parameter_269: name = "parameter_269" + original_name = "batch_norm2d_59.w_2" shape = [1024] dtype = "float32" min_val = float("5.60519e-45") @@ -2888,6 +3170,7 @@ class Program_weight_tensor_parameter_269: class Program_weight_tensor_parameter_270: name = "parameter_270" + original_name = "batch_norm2d_59.w_1" shape = [1024] dtype = "float32" min_val = float("-0.0139065") @@ -2899,6 +3182,7 @@ class Program_weight_tensor_parameter_270: class Program_weight_tensor_parameter_271: name = "parameter_271" + original_name = "conv2d_59.w_0" shape = [1024, 256, 1, 1] dtype = "float32" min_val = float("-0.201005") @@ -2910,6 +3194,7 @@ class Program_weight_tensor_parameter_271: class Program_weight_tensor_parameter_272: name = "parameter_272" + original_name = "batch_norm2d_58.b_0" shape = [256] dtype = "float32" min_val = float("-0.122277") @@ -2921,6 +3206,7 @@ class Program_weight_tensor_parameter_272: class Program_weight_tensor_parameter_273: name = "parameter_273" + original_name = "batch_norm2d_58.w_0" shape = [256] dtype = "float32" min_val = float("0.0539781") @@ -2932,6 +3218,7 @@ class Program_weight_tensor_parameter_273: class Program_weight_tensor_parameter_274: name = "parameter_274" + original_name = "batch_norm2d_58.w_2" shape = [256] dtype = "float32" min_val = float("0.000414495") @@ -2943,6 +3230,7 @@ class Program_weight_tensor_parameter_274: class Program_weight_tensor_parameter_275: name = "parameter_275" + original_name = "batch_norm2d_58.w_1" shape = [256] dtype = "float32" min_val = float("-0.0649243") @@ -2954,6 +3242,7 @@ class Program_weight_tensor_parameter_275: class Program_weight_tensor_parameter_276: name = "parameter_276" + original_name = "conv2d_58.w_0" shape = [256, 256, 3, 3] dtype = "float32" min_val = float("-0.126597") @@ -2965,6 +3254,7 @@ class Program_weight_tensor_parameter_276: class Program_weight_tensor_parameter_277: name = "parameter_277" + original_name = "batch_norm2d_57.b_0" shape = [256] dtype = "float32" min_val = float("-0.121751") @@ -2976,6 +3266,7 @@ class Program_weight_tensor_parameter_277: class Program_weight_tensor_parameter_278: name = "parameter_278" + original_name = "batch_norm2d_57.w_0" shape = [256] dtype = "float32" min_val = float("0.0471067") @@ -2987,6 +3278,7 @@ class Program_weight_tensor_parameter_278: class Program_weight_tensor_parameter_279: name = "parameter_279" + original_name = "batch_norm2d_57.w_2" shape = [256] dtype = "float32" min_val = float("0.00170937") @@ -2998,6 +3290,7 @@ class Program_weight_tensor_parameter_279: class Program_weight_tensor_parameter_280: name = "parameter_280" + original_name = "batch_norm2d_57.w_1" shape = [256] dtype = "float32" min_val = float("-0.126432") @@ -3009,6 +3302,7 @@ class Program_weight_tensor_parameter_280: class Program_weight_tensor_parameter_281: name = "parameter_281" + original_name = "conv2d_57.w_0" shape = [256, 1024, 1, 1] dtype = "float32" min_val = float("-0.147816") @@ -3020,6 +3314,7 @@ class Program_weight_tensor_parameter_281: class Program_weight_tensor_parameter_282: name = "parameter_282" + original_name = "batch_norm2d_56.b_0" shape = [1024] dtype = "float32" min_val = float("-0.105613") @@ -3031,6 +3326,7 @@ class Program_weight_tensor_parameter_282: class Program_weight_tensor_parameter_283: name = "parameter_283" + original_name = "batch_norm2d_56.w_0" shape = [1024] dtype = "float32" min_val = float("-0.118786") @@ -3042,6 +3338,7 @@ class Program_weight_tensor_parameter_283: class Program_weight_tensor_parameter_284: name = "parameter_284" + original_name = "batch_norm2d_56.w_2" shape = [1024] dtype = "float32" min_val = float("5.60519e-45") @@ -3053,6 +3350,7 @@ class Program_weight_tensor_parameter_284: class Program_weight_tensor_parameter_285: name = "parameter_285" + original_name = "batch_norm2d_56.w_1" shape = [1024] dtype = "float32" min_val = float("-0.0232002") @@ -3064,6 +3362,7 @@ class Program_weight_tensor_parameter_285: class Program_weight_tensor_parameter_286: name = "parameter_286" + original_name = "conv2d_56.w_0" shape = [1024, 256, 1, 1] dtype = "float32" min_val = float("-0.25673") @@ -3075,6 +3374,7 @@ class Program_weight_tensor_parameter_286: class Program_weight_tensor_parameter_287: name = "parameter_287" + original_name = "batch_norm2d_55.b_0" shape = [256] dtype = "float32" min_val = float("-0.361962") @@ -3086,6 +3386,7 @@ class Program_weight_tensor_parameter_287: class Program_weight_tensor_parameter_288: name = "parameter_288" + original_name = "batch_norm2d_55.w_0" shape = [256] dtype = "float32" min_val = float("0.0547539") @@ -3097,6 +3398,7 @@ class Program_weight_tensor_parameter_288: class Program_weight_tensor_parameter_289: name = "parameter_289" + original_name = "batch_norm2d_55.w_2" shape = [256] dtype = "float32" min_val = float("0.000475605") @@ -3108,6 +3410,7 @@ class Program_weight_tensor_parameter_289: class Program_weight_tensor_parameter_290: name = "parameter_290" + original_name = "batch_norm2d_55.w_1" shape = [256] dtype = "float32" min_val = float("-0.13866") @@ -3119,6 +3422,7 @@ class Program_weight_tensor_parameter_290: class Program_weight_tensor_parameter_291: name = "parameter_291" + original_name = "conv2d_55.w_0" shape = [256, 256, 3, 3] dtype = "float32" min_val = float("-0.156502") @@ -3130,6 +3434,7 @@ class Program_weight_tensor_parameter_291: class Program_weight_tensor_parameter_292: name = "parameter_292" + original_name = "batch_norm2d_54.b_0" shape = [256] dtype = "float32" min_val = float("-0.333298") @@ -3141,6 +3446,7 @@ class Program_weight_tensor_parameter_292: class Program_weight_tensor_parameter_293: name = "parameter_293" + original_name = "batch_norm2d_54.w_0" shape = [256] dtype = "float32" min_val = float("0.0471103") @@ -3152,6 +3458,7 @@ class Program_weight_tensor_parameter_293: class Program_weight_tensor_parameter_294: name = "parameter_294" + original_name = "batch_norm2d_54.w_2" shape = [256] dtype = "float32" min_val = float("0.00119363") @@ -3163,6 +3470,7 @@ class Program_weight_tensor_parameter_294: class Program_weight_tensor_parameter_295: name = "parameter_295" + original_name = "batch_norm2d_54.w_1" shape = [256] dtype = "float32" min_val = float("-0.131256") @@ -3174,6 +3482,7 @@ class Program_weight_tensor_parameter_295: class Program_weight_tensor_parameter_296: name = "parameter_296" + original_name = "conv2d_54.w_0" shape = [256, 1024, 1, 1] dtype = "float32" min_val = float("-0.169202") @@ -3185,6 +3494,7 @@ class Program_weight_tensor_parameter_296: class Program_weight_tensor_parameter_297: name = "parameter_297" + original_name = "batch_norm2d_53.b_0" shape = [1024] dtype = "float32" min_val = float("-0.14296") @@ -3196,6 +3506,7 @@ class Program_weight_tensor_parameter_297: class Program_weight_tensor_parameter_298: name = "parameter_298" + original_name = "batch_norm2d_53.w_0" shape = [1024] dtype = "float32" min_val = float("-0.0977966") @@ -3207,6 +3518,7 @@ class Program_weight_tensor_parameter_298: class Program_weight_tensor_parameter_299: name = "parameter_299" + original_name = "batch_norm2d_53.w_2" shape = [1024] dtype = "float32" min_val = float("5.60519e-45") @@ -3218,6 +3530,7 @@ class Program_weight_tensor_parameter_299: class Program_weight_tensor_parameter_300: name = "parameter_300" + original_name = "batch_norm2d_53.w_1" shape = [1024] dtype = "float32" min_val = float("-0.0163465") @@ -3229,6 +3542,7 @@ class Program_weight_tensor_parameter_300: class Program_weight_tensor_parameter_301: name = "parameter_301" + original_name = "conv2d_53.w_0" shape = [1024, 256, 1, 1] dtype = "float32" min_val = float("-0.190243") @@ -3240,6 +3554,7 @@ class Program_weight_tensor_parameter_301: class Program_weight_tensor_parameter_302: name = "parameter_302" + original_name = "batch_norm2d_52.b_0" shape = [256] dtype = "float32" min_val = float("-0.150391") @@ -3251,6 +3566,7 @@ class Program_weight_tensor_parameter_302: class Program_weight_tensor_parameter_303: name = "parameter_303" + original_name = "batch_norm2d_52.w_0" shape = [256] dtype = "float32" min_val = float("0.0479986") @@ -3262,6 +3578,7 @@ class Program_weight_tensor_parameter_303: class Program_weight_tensor_parameter_304: name = "parameter_304" + original_name = "batch_norm2d_52.w_2" shape = [256] dtype = "float32" min_val = float("0.000481484") @@ -3273,6 +3590,7 @@ class Program_weight_tensor_parameter_304: class Program_weight_tensor_parameter_305: name = "parameter_305" + original_name = "batch_norm2d_52.w_1" shape = [256] dtype = "float32" min_val = float("-0.0997624") @@ -3284,6 +3602,7 @@ class Program_weight_tensor_parameter_305: class Program_weight_tensor_parameter_306: name = "parameter_306" + original_name = "conv2d_52.w_0" shape = [256, 256, 3, 3] dtype = "float32" min_val = float("-0.126122") @@ -3295,6 +3614,7 @@ class Program_weight_tensor_parameter_306: class Program_weight_tensor_parameter_307: name = "parameter_307" + original_name = "batch_norm2d_51.b_0" shape = [256] dtype = "float32" min_val = float("-0.1715") @@ -3306,6 +3626,7 @@ class Program_weight_tensor_parameter_307: class Program_weight_tensor_parameter_308: name = "parameter_308" + original_name = "batch_norm2d_51.w_0" shape = [256] dtype = "float32" min_val = float("0.0470394") @@ -3317,6 +3638,7 @@ class Program_weight_tensor_parameter_308: class Program_weight_tensor_parameter_309: name = "parameter_309" + original_name = "batch_norm2d_51.w_2" shape = [256] dtype = "float32" min_val = float("0.00156337") @@ -3328,6 +3650,7 @@ class Program_weight_tensor_parameter_309: class Program_weight_tensor_parameter_310: name = "parameter_310" + original_name = "batch_norm2d_51.w_1" shape = [256] dtype = "float32" min_val = float("-0.0805164") @@ -3339,6 +3662,7 @@ class Program_weight_tensor_parameter_310: class Program_weight_tensor_parameter_311: name = "parameter_311" + original_name = "conv2d_51.w_0" shape = [256, 1024, 1, 1] dtype = "float32" min_val = float("-0.135523") @@ -3350,6 +3674,7 @@ class Program_weight_tensor_parameter_311: class Program_weight_tensor_parameter_312: name = "parameter_312" + original_name = "batch_norm2d_50.b_0" shape = [1024] dtype = "float32" min_val = float("-0.110059") @@ -3361,6 +3686,7 @@ class Program_weight_tensor_parameter_312: class Program_weight_tensor_parameter_313: name = "parameter_313" + original_name = "batch_norm2d_50.w_0" shape = [1024] dtype = "float32" min_val = float("-0.0931681") @@ -3372,6 +3698,7 @@ class Program_weight_tensor_parameter_313: class Program_weight_tensor_parameter_314: name = "parameter_314" + original_name = "batch_norm2d_50.w_2" shape = [1024] dtype = "float32" min_val = float("5.60519e-45") @@ -3383,6 +3710,7 @@ class Program_weight_tensor_parameter_314: class Program_weight_tensor_parameter_315: name = "parameter_315" + original_name = "batch_norm2d_50.w_1" shape = [1024] dtype = "float32" min_val = float("-0.018444") @@ -3394,6 +3722,7 @@ class Program_weight_tensor_parameter_315: class Program_weight_tensor_parameter_316: name = "parameter_316" + original_name = "conv2d_50.w_0" shape = [1024, 256, 1, 1] dtype = "float32" min_val = float("-0.139429") @@ -3405,6 +3734,7 @@ class Program_weight_tensor_parameter_316: class Program_weight_tensor_parameter_317: name = "parameter_317" + original_name = "batch_norm2d_49.b_0" shape = [256] dtype = "float32" min_val = float("-0.157348") @@ -3416,6 +3746,7 @@ class Program_weight_tensor_parameter_317: class Program_weight_tensor_parameter_318: name = "parameter_318" + original_name = "batch_norm2d_49.w_0" shape = [256] dtype = "float32" min_val = float("0.0530454") @@ -3427,6 +3758,7 @@ class Program_weight_tensor_parameter_318: class Program_weight_tensor_parameter_319: name = "parameter_319" + original_name = "batch_norm2d_49.w_2" shape = [256] dtype = "float32" min_val = float("0.000397306") @@ -3438,6 +3770,7 @@ class Program_weight_tensor_parameter_319: class Program_weight_tensor_parameter_320: name = "parameter_320" + original_name = "batch_norm2d_49.w_1" shape = [256] dtype = "float32" min_val = float("-0.107012") @@ -3449,6 +3782,7 @@ class Program_weight_tensor_parameter_320: class Program_weight_tensor_parameter_321: name = "parameter_321" + original_name = "conv2d_49.w_0" shape = [256, 256, 3, 3] dtype = "float32" min_val = float("-0.117755") @@ -3460,6 +3794,7 @@ class Program_weight_tensor_parameter_321: class Program_weight_tensor_parameter_322: name = "parameter_322" + original_name = "batch_norm2d_48.b_0" shape = [256] dtype = "float32" min_val = float("-0.181549") @@ -3471,6 +3806,7 @@ class Program_weight_tensor_parameter_322: class Program_weight_tensor_parameter_323: name = "parameter_323" + original_name = "batch_norm2d_48.w_0" shape = [256] dtype = "float32" min_val = float("0.0575181") @@ -3482,6 +3818,7 @@ class Program_weight_tensor_parameter_323: class Program_weight_tensor_parameter_324: name = "parameter_324" + original_name = "batch_norm2d_48.w_2" shape = [256] dtype = "float32" min_val = float("0.00120071") @@ -3493,6 +3830,7 @@ class Program_weight_tensor_parameter_324: class Program_weight_tensor_parameter_325: name = "parameter_325" + original_name = "batch_norm2d_48.w_1" shape = [256] dtype = "float32" min_val = float("-0.127524") @@ -3504,6 +3842,7 @@ class Program_weight_tensor_parameter_325: class Program_weight_tensor_parameter_326: name = "parameter_326" + original_name = "conv2d_48.w_0" shape = [256, 1024, 1, 1] dtype = "float32" min_val = float("-0.198103") @@ -3515,6 +3854,7 @@ class Program_weight_tensor_parameter_326: class Program_weight_tensor_parameter_327: name = "parameter_327" + original_name = "batch_norm2d_47.b_0" shape = [1024] dtype = "float32" min_val = float("-0.183645") @@ -3526,6 +3866,7 @@ class Program_weight_tensor_parameter_327: class Program_weight_tensor_parameter_328: name = "parameter_328" + original_name = "batch_norm2d_47.w_0" shape = [1024] dtype = "float32" min_val = float("-0.0497403") @@ -3537,6 +3878,7 @@ class Program_weight_tensor_parameter_328: class Program_weight_tensor_parameter_329: name = "parameter_329" + original_name = "batch_norm2d_47.w_2" shape = [1024] dtype = "float32" min_val = float("5.60519e-45") @@ -3548,6 +3890,7 @@ class Program_weight_tensor_parameter_329: class Program_weight_tensor_parameter_330: name = "parameter_330" + original_name = "batch_norm2d_47.w_1" shape = [1024] dtype = "float32" min_val = float("-0.028365") @@ -3559,6 +3902,7 @@ class Program_weight_tensor_parameter_330: class Program_weight_tensor_parameter_331: name = "parameter_331" + original_name = "conv2d_47.w_0" shape = [1024, 256, 1, 1] dtype = "float32" min_val = float("-0.220192") @@ -3570,6 +3914,7 @@ class Program_weight_tensor_parameter_331: class Program_weight_tensor_parameter_332: name = "parameter_332" + original_name = "batch_norm2d_46.b_0" shape = [256] dtype = "float32" min_val = float("-0.261715") @@ -3581,6 +3926,7 @@ class Program_weight_tensor_parameter_332: class Program_weight_tensor_parameter_333: name = "parameter_333" + original_name = "batch_norm2d_46.w_0" shape = [256] dtype = "float32" min_val = float("0.0650689") @@ -3592,6 +3938,7 @@ class Program_weight_tensor_parameter_333: class Program_weight_tensor_parameter_334: name = "parameter_334" + original_name = "batch_norm2d_46.w_2" shape = [256] dtype = "float32" min_val = float("0.000682362") @@ -3603,6 +3950,7 @@ class Program_weight_tensor_parameter_334: class Program_weight_tensor_parameter_335: name = "parameter_335" + original_name = "batch_norm2d_46.w_1" shape = [256] dtype = "float32" min_val = float("-0.0922706") @@ -3614,6 +3962,7 @@ class Program_weight_tensor_parameter_335: class Program_weight_tensor_parameter_336: name = "parameter_336" + original_name = "conv2d_46.w_0" shape = [256, 256, 3, 3] dtype = "float32" min_val = float("-0.165137") @@ -3625,6 +3974,7 @@ class Program_weight_tensor_parameter_336: class Program_weight_tensor_parameter_337: name = "parameter_337" + original_name = "batch_norm2d_45.b_0" shape = [256] dtype = "float32" min_val = float("-0.175628") @@ -3636,6 +3986,7 @@ class Program_weight_tensor_parameter_337: class Program_weight_tensor_parameter_338: name = "parameter_338" + original_name = "batch_norm2d_45.w_0" shape = [256] dtype = "float32" min_val = float("0.066233") @@ -3647,6 +3998,7 @@ class Program_weight_tensor_parameter_338: class Program_weight_tensor_parameter_339: name = "parameter_339" + original_name = "batch_norm2d_45.w_2" shape = [256] dtype = "float32" min_val = float("0.00199584") @@ -3658,6 +4010,7 @@ class Program_weight_tensor_parameter_339: class Program_weight_tensor_parameter_340: name = "parameter_340" + original_name = "batch_norm2d_45.w_1" shape = [256] dtype = "float32" min_val = float("-0.113064") @@ -3669,6 +4022,7 @@ class Program_weight_tensor_parameter_340: class Program_weight_tensor_parameter_341: name = "parameter_341" + original_name = "conv2d_45.w_0" shape = [256, 1024, 1, 1] dtype = "float32" min_val = float("-0.196158") @@ -3680,6 +4034,7 @@ class Program_weight_tensor_parameter_341: class Program_weight_tensor_parameter_342: name = "parameter_342" + original_name = "batch_norm2d_44.b_0" shape = [1024] dtype = "float32" min_val = float("-0.150438") @@ -3691,6 +4046,7 @@ class Program_weight_tensor_parameter_342: class Program_weight_tensor_parameter_343: name = "parameter_343" + original_name = "batch_norm2d_44.w_0" shape = [1024] dtype = "float32" min_val = float("-0.0526699") @@ -3702,6 +4058,7 @@ class Program_weight_tensor_parameter_343: class Program_weight_tensor_parameter_344: name = "parameter_344" + original_name = "batch_norm2d_44.w_2" shape = [1024] dtype = "float32" min_val = float("5.60519e-45") @@ -3713,6 +4070,7 @@ class Program_weight_tensor_parameter_344: class Program_weight_tensor_parameter_345: name = "parameter_345" + original_name = "batch_norm2d_44.w_1" shape = [1024] dtype = "float32" min_val = float("-0.0221359") @@ -3724,6 +4082,7 @@ class Program_weight_tensor_parameter_345: class Program_weight_tensor_parameter_346: name = "parameter_346" + original_name = "conv2d_44.w_0" shape = [1024, 256, 1, 1] dtype = "float32" min_val = float("-0.149936") @@ -3735,6 +4094,7 @@ class Program_weight_tensor_parameter_346: class Program_weight_tensor_parameter_347: name = "parameter_347" + original_name = "batch_norm2d_43.b_0" shape = [256] dtype = "float32" min_val = float("-0.153784") @@ -3746,6 +4106,7 @@ class Program_weight_tensor_parameter_347: class Program_weight_tensor_parameter_348: name = "parameter_348" + original_name = "batch_norm2d_43.w_0" shape = [256] dtype = "float32" min_val = float("0.0553597") @@ -3757,6 +4118,7 @@ class Program_weight_tensor_parameter_348: class Program_weight_tensor_parameter_349: name = "parameter_349" + original_name = "batch_norm2d_43.w_2" shape = [256] dtype = "float32" min_val = float("0.000499374") @@ -3768,6 +4130,7 @@ class Program_weight_tensor_parameter_349: class Program_weight_tensor_parameter_350: name = "parameter_350" + original_name = "batch_norm2d_43.w_1" shape = [256] dtype = "float32" min_val = float("-0.0774878") @@ -3779,6 +4142,7 @@ class Program_weight_tensor_parameter_350: class Program_weight_tensor_parameter_351: name = "parameter_351" + original_name = "conv2d_43.w_0" shape = [256, 256, 3, 3] dtype = "float32" min_val = float("-0.115989") @@ -3790,6 +4154,7 @@ class Program_weight_tensor_parameter_351: class Program_weight_tensor_parameter_352: name = "parameter_352" + original_name = "batch_norm2d_42.b_0" shape = [256] dtype = "float32" min_val = float("-0.143976") @@ -3801,6 +4166,7 @@ class Program_weight_tensor_parameter_352: class Program_weight_tensor_parameter_353: name = "parameter_353" + original_name = "batch_norm2d_42.w_0" shape = [256] dtype = "float32" min_val = float("0.0597508") @@ -3812,6 +4178,7 @@ class Program_weight_tensor_parameter_353: class Program_weight_tensor_parameter_354: name = "parameter_354" + original_name = "batch_norm2d_42.w_2" shape = [256] dtype = "float32" min_val = float("0.00153534") @@ -3823,6 +4190,7 @@ class Program_weight_tensor_parameter_354: class Program_weight_tensor_parameter_355: name = "parameter_355" + original_name = "batch_norm2d_42.w_1" shape = [256] dtype = "float32" min_val = float("-0.165527") @@ -3834,6 +4202,7 @@ class Program_weight_tensor_parameter_355: class Program_weight_tensor_parameter_356: name = "parameter_356" + original_name = "conv2d_42.w_0" shape = [256, 1024, 1, 1] dtype = "float32" min_val = float("-0.125145") @@ -3845,6 +4214,7 @@ class Program_weight_tensor_parameter_356: class Program_weight_tensor_parameter_357: name = "parameter_357" + original_name = "batch_norm2d_41.b_0" shape = [1024] dtype = "float32" min_val = float("-0.206426") @@ -3856,6 +4226,7 @@ class Program_weight_tensor_parameter_357: class Program_weight_tensor_parameter_358: name = "parameter_358" + original_name = "batch_norm2d_41.w_0" shape = [1024] dtype = "float32" min_val = float("-0.0464987") @@ -3867,6 +4238,7 @@ class Program_weight_tensor_parameter_358: class Program_weight_tensor_parameter_359: name = "parameter_359" + original_name = "batch_norm2d_41.w_2" shape = [1024] dtype = "float32" min_val = float("5.60519e-45") @@ -3878,6 +4250,7 @@ class Program_weight_tensor_parameter_359: class Program_weight_tensor_parameter_360: name = "parameter_360" + original_name = "batch_norm2d_41.w_1" shape = [1024] dtype = "float32" min_val = float("-0.0218367") @@ -3889,6 +4262,7 @@ class Program_weight_tensor_parameter_360: class Program_weight_tensor_parameter_361: name = "parameter_361" + original_name = "conv2d_41.w_0" shape = [1024, 256, 1, 1] dtype = "float32" min_val = float("-0.124601") @@ -3900,6 +4274,7 @@ class Program_weight_tensor_parameter_361: class Program_weight_tensor_parameter_362: name = "parameter_362" + original_name = "batch_norm2d_40.b_0" shape = [256] dtype = "float32" min_val = float("-0.168556") @@ -3911,6 +4286,7 @@ class Program_weight_tensor_parameter_362: class Program_weight_tensor_parameter_363: name = "parameter_363" + original_name = "batch_norm2d_40.w_0" shape = [256] dtype = "float32" min_val = float("0.0540817") @@ -3922,6 +4298,7 @@ class Program_weight_tensor_parameter_363: class Program_weight_tensor_parameter_364: name = "parameter_364" + original_name = "batch_norm2d_40.w_2" shape = [256] dtype = "float32" min_val = float("0.000486622") @@ -3933,6 +4310,7 @@ class Program_weight_tensor_parameter_364: class Program_weight_tensor_parameter_365: name = "parameter_365" + original_name = "batch_norm2d_40.w_1" shape = [256] dtype = "float32" min_val = float("-0.0689682") @@ -3944,6 +4322,7 @@ class Program_weight_tensor_parameter_365: class Program_weight_tensor_parameter_366: name = "parameter_366" + original_name = "conv2d_40.w_0" shape = [256, 256, 3, 3] dtype = "float32" min_val = float("-0.149872") @@ -3955,6 +4334,7 @@ class Program_weight_tensor_parameter_366: class Program_weight_tensor_parameter_367: name = "parameter_367" + original_name = "batch_norm2d_39.b_0" shape = [256] dtype = "float32" min_val = float("-0.125849") @@ -3966,6 +4346,7 @@ class Program_weight_tensor_parameter_367: class Program_weight_tensor_parameter_368: name = "parameter_368" + original_name = "batch_norm2d_39.w_0" shape = [256] dtype = "float32" min_val = float("0.0581656") @@ -3977,6 +4358,7 @@ class Program_weight_tensor_parameter_368: class Program_weight_tensor_parameter_369: name = "parameter_369" + original_name = "batch_norm2d_39.w_2" shape = [256] dtype = "float32" min_val = float("0.00159382") @@ -3988,6 +4370,7 @@ class Program_weight_tensor_parameter_369: class Program_weight_tensor_parameter_370: name = "parameter_370" + original_name = "batch_norm2d_39.w_1" shape = [256] dtype = "float32" min_val = float("-0.16078") @@ -3999,6 +4382,7 @@ class Program_weight_tensor_parameter_370: class Program_weight_tensor_parameter_371: name = "parameter_371" + original_name = "conv2d_39.w_0" shape = [256, 1024, 1, 1] dtype = "float32" min_val = float("-0.206113") @@ -4010,6 +4394,7 @@ class Program_weight_tensor_parameter_371: class Program_weight_tensor_parameter_372: name = "parameter_372" + original_name = "batch_norm2d_38.b_0" shape = [1024] dtype = "float32" min_val = float("-0.137891") @@ -4021,6 +4406,7 @@ class Program_weight_tensor_parameter_372: class Program_weight_tensor_parameter_373: name = "parameter_373" + original_name = "batch_norm2d_38.w_0" shape = [1024] dtype = "float32" min_val = float("-0.129481") @@ -4032,6 +4418,7 @@ class Program_weight_tensor_parameter_373: class Program_weight_tensor_parameter_374: name = "parameter_374" + original_name = "batch_norm2d_38.w_2" shape = [1024] dtype = "float32" min_val = float("5.60519e-45") @@ -4043,6 +4430,7 @@ class Program_weight_tensor_parameter_374: class Program_weight_tensor_parameter_375: name = "parameter_375" + original_name = "batch_norm2d_38.w_1" shape = [1024] dtype = "float32" min_val = float("-0.0303839") @@ -4054,6 +4442,7 @@ class Program_weight_tensor_parameter_375: class Program_weight_tensor_parameter_376: name = "parameter_376" + original_name = "conv2d_38.w_0" shape = [1024, 256, 1, 1] dtype = "float32" min_val = float("-0.133517") @@ -4065,6 +4454,7 @@ class Program_weight_tensor_parameter_376: class Program_weight_tensor_parameter_377: name = "parameter_377" + original_name = "batch_norm2d_37.b_0" shape = [256] dtype = "float32" min_val = float("-0.395405") @@ -4076,6 +4466,7 @@ class Program_weight_tensor_parameter_377: class Program_weight_tensor_parameter_378: name = "parameter_378" + original_name = "batch_norm2d_37.w_0" shape = [256] dtype = "float32" min_val = float("0.0522077") @@ -4087,6 +4478,7 @@ class Program_weight_tensor_parameter_378: class Program_weight_tensor_parameter_379: name = "parameter_379" + original_name = "batch_norm2d_37.w_2" shape = [256] dtype = "float32" min_val = float("0.00061321") @@ -4098,6 +4490,7 @@ class Program_weight_tensor_parameter_379: class Program_weight_tensor_parameter_380: name = "parameter_380" + original_name = "batch_norm2d_37.w_1" shape = [256] dtype = "float32" min_val = float("-0.177767") @@ -4109,6 +4502,7 @@ class Program_weight_tensor_parameter_380: class Program_weight_tensor_parameter_381: name = "parameter_381" + original_name = "conv2d_37.w_0" shape = [256, 256, 3, 3] dtype = "float32" min_val = float("-0.0892948") @@ -4120,6 +4514,7 @@ class Program_weight_tensor_parameter_381: class Program_weight_tensor_parameter_382: name = "parameter_382" + original_name = "batch_norm2d_36.b_0" shape = [256] dtype = "float32" min_val = float("-0.128915") @@ -4131,6 +4526,7 @@ class Program_weight_tensor_parameter_382: class Program_weight_tensor_parameter_383: name = "parameter_383" + original_name = "batch_norm2d_36.w_0" shape = [256] dtype = "float32" min_val = float("0.0464295") @@ -4142,6 +4538,7 @@ class Program_weight_tensor_parameter_383: class Program_weight_tensor_parameter_384: name = "parameter_384" + original_name = "batch_norm2d_36.w_2" shape = [256] dtype = "float32" min_val = float("0.00125646") @@ -4153,6 +4550,7 @@ class Program_weight_tensor_parameter_384: class Program_weight_tensor_parameter_385: name = "parameter_385" + original_name = "batch_norm2d_36.w_1" shape = [256] dtype = "float32" min_val = float("-0.170545") @@ -4164,6 +4562,7 @@ class Program_weight_tensor_parameter_385: class Program_weight_tensor_parameter_386: name = "parameter_386" + original_name = "conv2d_36.w_0" shape = [256, 1024, 1, 1] dtype = "float32" min_val = float("-0.187673") @@ -4175,6 +4574,7 @@ class Program_weight_tensor_parameter_386: class Program_weight_tensor_parameter_387: name = "parameter_387" + original_name = "batch_norm2d_35.b_0" shape = [1024] dtype = "float32" min_val = float("-0.17385") @@ -4186,6 +4586,7 @@ class Program_weight_tensor_parameter_387: class Program_weight_tensor_parameter_388: name = "parameter_388" + original_name = "batch_norm2d_35.w_0" shape = [1024] dtype = "float32" min_val = float("-0.0858251") @@ -4197,6 +4598,7 @@ class Program_weight_tensor_parameter_388: class Program_weight_tensor_parameter_389: name = "parameter_389" + original_name = "batch_norm2d_35.w_2" shape = [1024] dtype = "float32" min_val = float("5.60519e-45") @@ -4208,6 +4610,7 @@ class Program_weight_tensor_parameter_389: class Program_weight_tensor_parameter_390: name = "parameter_390" + original_name = "batch_norm2d_35.w_1" shape = [1024] dtype = "float32" min_val = float("-0.0571547") @@ -4219,6 +4622,7 @@ class Program_weight_tensor_parameter_390: class Program_weight_tensor_parameter_391: name = "parameter_391" + original_name = "conv2d_35.w_0" shape = [1024, 256, 1, 1] dtype = "float32" min_val = float("-0.336971") @@ -4230,6 +4634,7 @@ class Program_weight_tensor_parameter_391: class Program_weight_tensor_parameter_392: name = "parameter_392" + original_name = "batch_norm2d_34.b_0" shape = [256] dtype = "float32" min_val = float("-0.188655") @@ -4241,6 +4646,7 @@ class Program_weight_tensor_parameter_392: class Program_weight_tensor_parameter_393: name = "parameter_393" + original_name = "batch_norm2d_34.w_0" shape = [256] dtype = "float32" min_val = float("0.0683378") @@ -4252,6 +4658,7 @@ class Program_weight_tensor_parameter_393: class Program_weight_tensor_parameter_394: name = "parameter_394" + original_name = "batch_norm2d_34.w_2" shape = [256] dtype = "float32" min_val = float("0.000900066") @@ -4263,6 +4670,7 @@ class Program_weight_tensor_parameter_394: class Program_weight_tensor_parameter_395: name = "parameter_395" + original_name = "batch_norm2d_34.w_1" shape = [256] dtype = "float32" min_val = float("-0.145351") @@ -4274,6 +4682,7 @@ class Program_weight_tensor_parameter_395: class Program_weight_tensor_parameter_396: name = "parameter_396" + original_name = "conv2d_34.w_0" shape = [256, 256, 3, 3] dtype = "float32" min_val = float("-0.132562") @@ -4285,6 +4694,7 @@ class Program_weight_tensor_parameter_396: class Program_weight_tensor_parameter_397: name = "parameter_397" + original_name = "batch_norm2d_33.b_0" shape = [256] dtype = "float32" min_val = float("-0.144211") @@ -4296,6 +4706,7 @@ class Program_weight_tensor_parameter_397: class Program_weight_tensor_parameter_398: name = "parameter_398" + original_name = "batch_norm2d_33.w_0" shape = [256] dtype = "float32" min_val = float("0.0696308") @@ -4307,6 +4718,7 @@ class Program_weight_tensor_parameter_398: class Program_weight_tensor_parameter_399: name = "parameter_399" + original_name = "batch_norm2d_33.w_2" shape = [256] dtype = "float32" min_val = float("0.00197131") @@ -4318,6 +4730,7 @@ class Program_weight_tensor_parameter_399: class Program_weight_tensor_parameter_400: name = "parameter_400" + original_name = "batch_norm2d_33.w_1" shape = [256] dtype = "float32" min_val = float("-0.131424") @@ -4329,6 +4742,7 @@ class Program_weight_tensor_parameter_400: class Program_weight_tensor_parameter_401: name = "parameter_401" + original_name = "conv2d_33.w_0" shape = [256, 1024, 1, 1] dtype = "float32" min_val = float("-0.172959") @@ -4340,6 +4754,7 @@ class Program_weight_tensor_parameter_401: class Program_weight_tensor_parameter_402: name = "parameter_402" + original_name = "batch_norm2d_32.b_0" shape = [1024] dtype = "float32" min_val = float("-0.167666") @@ -4351,6 +4766,7 @@ class Program_weight_tensor_parameter_402: class Program_weight_tensor_parameter_403: name = "parameter_403" + original_name = "batch_norm2d_32.w_0" shape = [1024] dtype = "float32" min_val = float("-0.112383") @@ -4362,6 +4778,7 @@ class Program_weight_tensor_parameter_403: class Program_weight_tensor_parameter_404: name = "parameter_404" + original_name = "batch_norm2d_32.w_2" shape = [1024] dtype = "float32" min_val = float("5.60519e-45") @@ -4373,6 +4790,7 @@ class Program_weight_tensor_parameter_404: class Program_weight_tensor_parameter_405: name = "parameter_405" + original_name = "batch_norm2d_32.w_1" shape = [1024] dtype = "float32" min_val = float("-0.0534777") @@ -4384,6 +4802,7 @@ class Program_weight_tensor_parameter_405: class Program_weight_tensor_parameter_406: name = "parameter_406" + original_name = "conv2d_32.w_0" shape = [1024, 256, 1, 1] dtype = "float32" min_val = float("-0.282625") @@ -4395,6 +4814,7 @@ class Program_weight_tensor_parameter_406: class Program_weight_tensor_parameter_407: name = "parameter_407" + original_name = "batch_norm2d_31.b_0" shape = [256] dtype = "float32" min_val = float("-0.391536") @@ -4406,6 +4826,7 @@ class Program_weight_tensor_parameter_407: class Program_weight_tensor_parameter_408: name = "parameter_408" + original_name = "batch_norm2d_31.w_0" shape = [256] dtype = "float32" min_val = float("0.0474037") @@ -4417,6 +4838,7 @@ class Program_weight_tensor_parameter_408: class Program_weight_tensor_parameter_409: name = "parameter_409" + original_name = "batch_norm2d_31.w_2" shape = [256] dtype = "float32" min_val = float("0.000825777") @@ -4428,6 +4850,7 @@ class Program_weight_tensor_parameter_409: class Program_weight_tensor_parameter_410: name = "parameter_410" + original_name = "batch_norm2d_31.w_1" shape = [256] dtype = "float32" min_val = float("-0.284886") @@ -4439,6 +4862,7 @@ class Program_weight_tensor_parameter_410: class Program_weight_tensor_parameter_411: name = "parameter_411" + original_name = "conv2d_31.w_0" shape = [256, 256, 3, 3] dtype = "float32" min_val = float("-0.174341") @@ -4450,6 +4874,7 @@ class Program_weight_tensor_parameter_411: class Program_weight_tensor_parameter_412: name = "parameter_412" + original_name = "batch_norm2d_30.b_0" shape = [256] dtype = "float32" min_val = float("-0.254122") @@ -4461,6 +4886,7 @@ class Program_weight_tensor_parameter_412: class Program_weight_tensor_parameter_413: name = "parameter_413" + original_name = "batch_norm2d_30.w_0" shape = [256] dtype = "float32" min_val = float("0.0522015") @@ -4472,6 +4898,7 @@ class Program_weight_tensor_parameter_413: class Program_weight_tensor_parameter_414: name = "parameter_414" + original_name = "batch_norm2d_30.w_2" shape = [256] dtype = "float32" min_val = float("0.00139353") @@ -4483,6 +4910,7 @@ class Program_weight_tensor_parameter_414: class Program_weight_tensor_parameter_415: name = "parameter_415" + original_name = "batch_norm2d_30.w_1" shape = [256] dtype = "float32" min_val = float("-0.325423") @@ -4494,6 +4922,7 @@ class Program_weight_tensor_parameter_415: class Program_weight_tensor_parameter_416: name = "parameter_416" + original_name = "conv2d_30.w_0" shape = [256, 1024, 1, 1] dtype = "float32" min_val = float("-0.143796") @@ -4505,6 +4934,7 @@ class Program_weight_tensor_parameter_416: class Program_weight_tensor_parameter_417: name = "parameter_417" + original_name = "batch_norm2d_29.b_0" shape = [1024] dtype = "float32" min_val = float("-0.070916") @@ -4516,6 +4946,7 @@ class Program_weight_tensor_parameter_417: class Program_weight_tensor_parameter_418: name = "parameter_418" + original_name = "batch_norm2d_29.w_0" shape = [1024] dtype = "float32" min_val = float("-0.00377393") @@ -4527,6 +4958,7 @@ class Program_weight_tensor_parameter_418: class Program_weight_tensor_parameter_419: name = "parameter_419" + original_name = "batch_norm2d_29.w_2" shape = [1024] dtype = "float32" min_val = float("5.60519e-45") @@ -4538,6 +4970,7 @@ class Program_weight_tensor_parameter_419: class Program_weight_tensor_parameter_420: name = "parameter_420" + original_name = "batch_norm2d_29.w_1" shape = [1024] dtype = "float32" min_val = float("-0.131816") @@ -4549,6 +4982,7 @@ class Program_weight_tensor_parameter_420: class Program_weight_tensor_parameter_421: name = "parameter_421" + original_name = "conv2d_29.w_0" shape = [1024, 512, 1, 1] dtype = "float32" min_val = float("-0.357426") @@ -4560,6 +4994,7 @@ class Program_weight_tensor_parameter_421: class Program_weight_tensor_parameter_422: name = "parameter_422" + original_name = "batch_norm2d_28.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0709161") @@ -4571,6 +5006,7 @@ class Program_weight_tensor_parameter_422: class Program_weight_tensor_parameter_423: name = "parameter_423" + original_name = "batch_norm2d_28.w_0" shape = [1024] dtype = "float32" min_val = float("-0.0815912") @@ -4582,6 +5018,7 @@ class Program_weight_tensor_parameter_423: class Program_weight_tensor_parameter_424: name = "parameter_424" + original_name = "batch_norm2d_28.w_2" shape = [1024] dtype = "float32" min_val = float("5.60519e-45") @@ -4593,6 +5030,7 @@ class Program_weight_tensor_parameter_424: class Program_weight_tensor_parameter_425: name = "parameter_425" + original_name = "batch_norm2d_28.w_1" shape = [1024] dtype = "float32" min_val = float("-0.05334") @@ -4604,6 +5042,7 @@ class Program_weight_tensor_parameter_425: class Program_weight_tensor_parameter_426: name = "parameter_426" + original_name = "conv2d_28.w_0" shape = [1024, 256, 1, 1] dtype = "float32" min_val = float("-0.130086") @@ -4615,6 +5054,7 @@ class Program_weight_tensor_parameter_426: class Program_weight_tensor_parameter_427: name = "parameter_427" + original_name = "batch_norm2d_27.b_0" shape = [256] dtype = "float32" min_val = float("-0.132216") @@ -4626,6 +5066,7 @@ class Program_weight_tensor_parameter_427: class Program_weight_tensor_parameter_428: name = "parameter_428" + original_name = "batch_norm2d_27.w_0" shape = [256] dtype = "float32" min_val = float("0.0621491") @@ -4637,6 +5078,7 @@ class Program_weight_tensor_parameter_428: class Program_weight_tensor_parameter_429: name = "parameter_429" + original_name = "batch_norm2d_27.w_2" shape = [256] dtype = "float32" min_val = float("0.000498966") @@ -4648,6 +5090,7 @@ class Program_weight_tensor_parameter_429: class Program_weight_tensor_parameter_430: name = "parameter_430" + original_name = "batch_norm2d_27.w_1" shape = [256] dtype = "float32" min_val = float("-0.0500929") @@ -4659,6 +5102,7 @@ class Program_weight_tensor_parameter_430: class Program_weight_tensor_parameter_431: name = "parameter_431" + original_name = "conv2d_27.w_0" shape = [256, 256, 3, 3] dtype = "float32" min_val = float("-0.0972044") @@ -4670,6 +5114,7 @@ class Program_weight_tensor_parameter_431: class Program_weight_tensor_parameter_432: name = "parameter_432" + original_name = "batch_norm2d_26.b_0" shape = [256] dtype = "float32" min_val = float("-0.200827") @@ -4681,6 +5126,7 @@ class Program_weight_tensor_parameter_432: class Program_weight_tensor_parameter_433: name = "parameter_433" + original_name = "batch_norm2d_26.w_0" shape = [256] dtype = "float32" min_val = float("0.0806692") @@ -4692,6 +5138,7 @@ class Program_weight_tensor_parameter_433: class Program_weight_tensor_parameter_434: name = "parameter_434" + original_name = "batch_norm2d_26.w_2" shape = [256] dtype = "float32" min_val = float("0.00175098") @@ -4703,6 +5150,7 @@ class Program_weight_tensor_parameter_434: class Program_weight_tensor_parameter_435: name = "parameter_435" + original_name = "batch_norm2d_26.w_1" shape = [256] dtype = "float32" min_val = float("-0.154164") @@ -4714,6 +5162,7 @@ class Program_weight_tensor_parameter_435: class Program_weight_tensor_parameter_436: name = "parameter_436" + original_name = "conv2d_26.w_0" shape = [256, 512, 1, 1] dtype = "float32" min_val = float("-0.126123") @@ -4725,6 +5174,7 @@ class Program_weight_tensor_parameter_436: class Program_weight_tensor_parameter_437: name = "parameter_437" + original_name = "batch_norm2d_25.b_0" shape = [512] dtype = "float32" min_val = float("-0.213783") @@ -4736,6 +5186,7 @@ class Program_weight_tensor_parameter_437: class Program_weight_tensor_parameter_438: name = "parameter_438" + original_name = "batch_norm2d_25.w_0" shape = [512] dtype = "float32" min_val = float("-0.132737") @@ -4747,6 +5198,7 @@ class Program_weight_tensor_parameter_438: class Program_weight_tensor_parameter_439: name = "parameter_439" + original_name = "batch_norm2d_25.w_2" shape = [512] dtype = "float32" min_val = float("1.17707e-05") @@ -4758,6 +5210,7 @@ class Program_weight_tensor_parameter_439: class Program_weight_tensor_parameter_440: name = "parameter_440" + original_name = "batch_norm2d_25.w_1" shape = [512] dtype = "float32" min_val = float("-0.0737471") @@ -4769,6 +5222,7 @@ class Program_weight_tensor_parameter_440: class Program_weight_tensor_parameter_441: name = "parameter_441" + original_name = "conv2d_25.w_0" shape = [512, 128, 1, 1] dtype = "float32" min_val = float("-0.19458") @@ -4780,6 +5234,7 @@ class Program_weight_tensor_parameter_441: class Program_weight_tensor_parameter_442: name = "parameter_442" + original_name = "batch_norm2d_24.b_0" shape = [128] dtype = "float32" min_val = float("-0.188612") @@ -4791,6 +5246,7 @@ class Program_weight_tensor_parameter_442: class Program_weight_tensor_parameter_443: name = "parameter_443" + original_name = "batch_norm2d_24.w_0" shape = [128] dtype = "float32" min_val = float("0.0808353") @@ -4802,6 +5258,7 @@ class Program_weight_tensor_parameter_443: class Program_weight_tensor_parameter_444: name = "parameter_444" + original_name = "batch_norm2d_24.w_2" shape = [128] dtype = "float32" min_val = float("0.00107675") @@ -4813,6 +5270,7 @@ class Program_weight_tensor_parameter_444: class Program_weight_tensor_parameter_445: name = "parameter_445" + original_name = "batch_norm2d_24.w_1" shape = [128] dtype = "float32" min_val = float("-0.0859865") @@ -4824,6 +5282,7 @@ class Program_weight_tensor_parameter_445: class Program_weight_tensor_parameter_446: name = "parameter_446" + original_name = "conv2d_24.w_0" shape = [128, 128, 3, 3] dtype = "float32" min_val = float("-0.136146") @@ -4835,6 +5294,7 @@ class Program_weight_tensor_parameter_446: class Program_weight_tensor_parameter_447: name = "parameter_447" + original_name = "batch_norm2d_23.b_0" shape = [128] dtype = "float32" min_val = float("-0.133446") @@ -4846,6 +5306,7 @@ class Program_weight_tensor_parameter_447: class Program_weight_tensor_parameter_448: name = "parameter_448" + original_name = "batch_norm2d_23.w_0" shape = [128] dtype = "float32" min_val = float("0.0832541") @@ -4857,6 +5318,7 @@ class Program_weight_tensor_parameter_448: class Program_weight_tensor_parameter_449: name = "parameter_449" + original_name = "batch_norm2d_23.w_2" shape = [128] dtype = "float32" min_val = float("0.00151614") @@ -4868,6 +5330,7 @@ class Program_weight_tensor_parameter_449: class Program_weight_tensor_parameter_450: name = "parameter_450" + original_name = "batch_norm2d_23.w_1" shape = [128] dtype = "float32" min_val = float("-0.0779984") @@ -4879,6 +5342,7 @@ class Program_weight_tensor_parameter_450: class Program_weight_tensor_parameter_451: name = "parameter_451" + original_name = "conv2d_23.w_0" shape = [128, 512, 1, 1] dtype = "float32" min_val = float("-0.116246") @@ -4890,6 +5354,7 @@ class Program_weight_tensor_parameter_451: class Program_weight_tensor_parameter_452: name = "parameter_452" + original_name = "batch_norm2d_22.b_0" shape = [512] dtype = "float32" min_val = float("-0.206455") @@ -4901,6 +5366,7 @@ class Program_weight_tensor_parameter_452: class Program_weight_tensor_parameter_453: name = "parameter_453" + original_name = "batch_norm2d_22.w_0" shape = [512] dtype = "float32" min_val = float("-0.0548628") @@ -4912,6 +5378,7 @@ class Program_weight_tensor_parameter_453: class Program_weight_tensor_parameter_454: name = "parameter_454" + original_name = "batch_norm2d_22.w_2" shape = [512] dtype = "float32" min_val = float("5.60519e-45") @@ -4923,6 +5390,7 @@ class Program_weight_tensor_parameter_454: class Program_weight_tensor_parameter_455: name = "parameter_455" + original_name = "batch_norm2d_22.w_1" shape = [512] dtype = "float32" min_val = float("-0.0518352") @@ -4934,6 +5402,7 @@ class Program_weight_tensor_parameter_455: class Program_weight_tensor_parameter_456: name = "parameter_456" + original_name = "conv2d_22.w_0" shape = [512, 128, 1, 1] dtype = "float32" min_val = float("-0.153612") @@ -4945,6 +5414,7 @@ class Program_weight_tensor_parameter_456: class Program_weight_tensor_parameter_457: name = "parameter_457" + original_name = "batch_norm2d_21.b_0" shape = [128] dtype = "float32" min_val = float("-0.18239") @@ -4956,6 +5426,7 @@ class Program_weight_tensor_parameter_457: class Program_weight_tensor_parameter_458: name = "parameter_458" + original_name = "batch_norm2d_21.w_0" shape = [128] dtype = "float32" min_val = float("0.0744946") @@ -4967,6 +5438,7 @@ class Program_weight_tensor_parameter_458: class Program_weight_tensor_parameter_459: name = "parameter_459" + original_name = "batch_norm2d_21.w_2" shape = [128] dtype = "float32" min_val = float("0.00122704") @@ -4978,6 +5450,7 @@ class Program_weight_tensor_parameter_459: class Program_weight_tensor_parameter_460: name = "parameter_460" + original_name = "batch_norm2d_21.w_1" shape = [128] dtype = "float32" min_val = float("-0.09244") @@ -4989,6 +5462,7 @@ class Program_weight_tensor_parameter_460: class Program_weight_tensor_parameter_461: name = "parameter_461" + original_name = "conv2d_21.w_0" shape = [128, 128, 3, 3] dtype = "float32" min_val = float("-0.175254") @@ -5000,6 +5474,7 @@ class Program_weight_tensor_parameter_461: class Program_weight_tensor_parameter_462: name = "parameter_462" + original_name = "batch_norm2d_20.b_0" shape = [128] dtype = "float32" min_val = float("-0.264595") @@ -5011,6 +5486,7 @@ class Program_weight_tensor_parameter_462: class Program_weight_tensor_parameter_463: name = "parameter_463" + original_name = "batch_norm2d_20.w_0" shape = [128] dtype = "float32" min_val = float("0.0727919") @@ -5022,6 +5498,7 @@ class Program_weight_tensor_parameter_463: class Program_weight_tensor_parameter_464: name = "parameter_464" + original_name = "batch_norm2d_20.w_2" shape = [128] dtype = "float32" min_val = float("0.00204811") @@ -5033,6 +5510,7 @@ class Program_weight_tensor_parameter_464: class Program_weight_tensor_parameter_465: name = "parameter_465" + original_name = "batch_norm2d_20.w_1" shape = [128] dtype = "float32" min_val = float("-0.292813") @@ -5044,6 +5522,7 @@ class Program_weight_tensor_parameter_465: class Program_weight_tensor_parameter_466: name = "parameter_466" + original_name = "conv2d_20.w_0" shape = [128, 512, 1, 1] dtype = "float32" min_val = float("-0.158225") @@ -5055,6 +5534,7 @@ class Program_weight_tensor_parameter_466: class Program_weight_tensor_parameter_467: name = "parameter_467" + original_name = "batch_norm2d_19.b_0" shape = [512] dtype = "float32" min_val = float("-0.272751") @@ -5066,6 +5546,7 @@ class Program_weight_tensor_parameter_467: class Program_weight_tensor_parameter_468: name = "parameter_468" + original_name = "batch_norm2d_19.w_0" shape = [512] dtype = "float32" min_val = float("-0.0773685") @@ -5077,6 +5558,7 @@ class Program_weight_tensor_parameter_468: class Program_weight_tensor_parameter_469: name = "parameter_469" + original_name = "batch_norm2d_19.w_2" shape = [512] dtype = "float32" min_val = float("5.60519e-45") @@ -5088,6 +5570,7 @@ class Program_weight_tensor_parameter_469: class Program_weight_tensor_parameter_470: name = "parameter_470" + original_name = "batch_norm2d_19.w_1" shape = [512] dtype = "float32" min_val = float("-0.0518335") @@ -5099,6 +5582,7 @@ class Program_weight_tensor_parameter_470: class Program_weight_tensor_parameter_471: name = "parameter_471" + original_name = "conv2d_19.w_0" shape = [512, 128, 1, 1] dtype = "float32" min_val = float("-0.212086") @@ -5110,6 +5594,7 @@ class Program_weight_tensor_parameter_471: class Program_weight_tensor_parameter_472: name = "parameter_472" + original_name = "batch_norm2d_18.b_0" shape = [128] dtype = "float32" min_val = float("-0.207375") @@ -5121,6 +5606,7 @@ class Program_weight_tensor_parameter_472: class Program_weight_tensor_parameter_473: name = "parameter_473" + original_name = "batch_norm2d_18.w_0" shape = [128] dtype = "float32" min_val = float("0.0714045") @@ -5132,6 +5618,7 @@ class Program_weight_tensor_parameter_473: class Program_weight_tensor_parameter_474: name = "parameter_474" + original_name = "batch_norm2d_18.w_2" shape = [128] dtype = "float32" min_val = float("0.00167628") @@ -5143,6 +5630,7 @@ class Program_weight_tensor_parameter_474: class Program_weight_tensor_parameter_475: name = "parameter_475" + original_name = "batch_norm2d_18.w_1" shape = [128] dtype = "float32" min_val = float("-0.388887") @@ -5154,6 +5642,7 @@ class Program_weight_tensor_parameter_475: class Program_weight_tensor_parameter_476: name = "parameter_476" + original_name = "conv2d_18.w_0" shape = [128, 128, 3, 3] dtype = "float32" min_val = float("-0.239234") @@ -5165,6 +5654,7 @@ class Program_weight_tensor_parameter_476: class Program_weight_tensor_parameter_477: name = "parameter_477" + original_name = "batch_norm2d_17.b_0" shape = [128] dtype = "float32" min_val = float("-0.150355") @@ -5176,6 +5666,7 @@ class Program_weight_tensor_parameter_477: class Program_weight_tensor_parameter_478: name = "parameter_478" + original_name = "batch_norm2d_17.w_0" shape = [128] dtype = "float32" min_val = float("0.0835381") @@ -5187,6 +5678,7 @@ class Program_weight_tensor_parameter_478: class Program_weight_tensor_parameter_479: name = "parameter_479" + original_name = "batch_norm2d_17.w_2" shape = [128] dtype = "float32" min_val = float("0.00224778") @@ -5198,6 +5690,7 @@ class Program_weight_tensor_parameter_479: class Program_weight_tensor_parameter_480: name = "parameter_480" + original_name = "batch_norm2d_17.w_1" shape = [128] dtype = "float32" min_val = float("-0.267656") @@ -5209,6 +5702,7 @@ class Program_weight_tensor_parameter_480: class Program_weight_tensor_parameter_481: name = "parameter_481" + original_name = "conv2d_17.w_0" shape = [128, 512, 1, 1] dtype = "float32" min_val = float("-0.192112") @@ -5220,6 +5714,7 @@ class Program_weight_tensor_parameter_481: class Program_weight_tensor_parameter_482: name = "parameter_482" + original_name = "batch_norm2d_16.b_0" shape = [512] dtype = "float32" min_val = float("-0.174138") @@ -5231,6 +5726,7 @@ class Program_weight_tensor_parameter_482: class Program_weight_tensor_parameter_483: name = "parameter_483" + original_name = "batch_norm2d_16.w_0" shape = [512] dtype = "float32" min_val = float("-0.0299296") @@ -5242,6 +5738,7 @@ class Program_weight_tensor_parameter_483: class Program_weight_tensor_parameter_484: name = "parameter_484" + original_name = "batch_norm2d_16.w_2" shape = [512] dtype = "float32" min_val = float("5.60519e-45") @@ -5253,6 +5750,7 @@ class Program_weight_tensor_parameter_484: class Program_weight_tensor_parameter_485: name = "parameter_485" + original_name = "batch_norm2d_16.w_1" shape = [512] dtype = "float32" min_val = float("-0.162818") @@ -5264,6 +5762,7 @@ class Program_weight_tensor_parameter_485: class Program_weight_tensor_parameter_486: name = "parameter_486" + original_name = "conv2d_16.w_0" shape = [512, 256, 1, 1] dtype = "float32" min_val = float("-0.52586") @@ -5275,6 +5774,7 @@ class Program_weight_tensor_parameter_486: class Program_weight_tensor_parameter_487: name = "parameter_487" + original_name = "batch_norm2d_15.b_0" shape = [512] dtype = "float32" min_val = float("-0.174138") @@ -5286,6 +5786,7 @@ class Program_weight_tensor_parameter_487: class Program_weight_tensor_parameter_488: name = "parameter_488" + original_name = "batch_norm2d_15.w_0" shape = [512] dtype = "float32" min_val = float("-0.0561179") @@ -5297,6 +5798,7 @@ class Program_weight_tensor_parameter_488: class Program_weight_tensor_parameter_489: name = "parameter_489" + original_name = "batch_norm2d_15.w_2" shape = [512] dtype = "float32" min_val = float("5.60519e-45") @@ -5308,6 +5810,7 @@ class Program_weight_tensor_parameter_489: class Program_weight_tensor_parameter_490: name = "parameter_490" + original_name = "batch_norm2d_15.w_1" shape = [512] dtype = "float32" min_val = float("-0.0692909") @@ -5319,6 +5822,7 @@ class Program_weight_tensor_parameter_490: class Program_weight_tensor_parameter_491: name = "parameter_491" + original_name = "conv2d_15.w_0" shape = [512, 128, 1, 1] dtype = "float32" min_val = float("-0.207854") @@ -5330,6 +5834,7 @@ class Program_weight_tensor_parameter_491: class Program_weight_tensor_parameter_492: name = "parameter_492" + original_name = "batch_norm2d_14.b_0" shape = [128] dtype = "float32" min_val = float("-0.125217") @@ -5341,6 +5846,7 @@ class Program_weight_tensor_parameter_492: class Program_weight_tensor_parameter_493: name = "parameter_493" + original_name = "batch_norm2d_14.w_0" shape = [128] dtype = "float32" min_val = float("0.0760892") @@ -5352,6 +5858,7 @@ class Program_weight_tensor_parameter_493: class Program_weight_tensor_parameter_494: name = "parameter_494" + original_name = "batch_norm2d_14.w_2" shape = [128] dtype = "float32" min_val = float("0.000951838") @@ -5363,6 +5870,7 @@ class Program_weight_tensor_parameter_494: class Program_weight_tensor_parameter_495: name = "parameter_495" + original_name = "batch_norm2d_14.w_1" shape = [128] dtype = "float32" min_val = float("-0.126163") @@ -5374,6 +5882,7 @@ class Program_weight_tensor_parameter_495: class Program_weight_tensor_parameter_496: name = "parameter_496" + original_name = "conv2d_14.w_0" shape = [128, 128, 3, 3] dtype = "float32" min_val = float("-0.096496") @@ -5385,6 +5894,7 @@ class Program_weight_tensor_parameter_496: class Program_weight_tensor_parameter_497: name = "parameter_497" + original_name = "batch_norm2d_13.b_0" shape = [128] dtype = "float32" min_val = float("-0.175984") @@ -5396,6 +5906,7 @@ class Program_weight_tensor_parameter_497: class Program_weight_tensor_parameter_498: name = "parameter_498" + original_name = "batch_norm2d_13.w_0" shape = [128] dtype = "float32" min_val = float("0.0741341") @@ -5407,6 +5918,7 @@ class Program_weight_tensor_parameter_498: class Program_weight_tensor_parameter_499: name = "parameter_499" + original_name = "batch_norm2d_13.w_2" shape = [128] dtype = "float32" min_val = float("0.00194663") @@ -5418,6 +5930,7 @@ class Program_weight_tensor_parameter_499: class Program_weight_tensor_parameter_500: name = "parameter_500" + original_name = "batch_norm2d_13.w_1" shape = [128] dtype = "float32" min_val = float("-0.185112") @@ -5429,6 +5942,7 @@ class Program_weight_tensor_parameter_500: class Program_weight_tensor_parameter_501: name = "parameter_501" + original_name = "conv2d_13.w_0" shape = [128, 256, 1, 1] dtype = "float32" min_val = float("-0.16558") @@ -5440,6 +5954,7 @@ class Program_weight_tensor_parameter_501: class Program_weight_tensor_parameter_502: name = "parameter_502" + original_name = "batch_norm2d_12.b_0" shape = [256] dtype = "float32" min_val = float("-0.242305") @@ -5451,6 +5966,7 @@ class Program_weight_tensor_parameter_502: class Program_weight_tensor_parameter_503: name = "parameter_503" + original_name = "batch_norm2d_12.w_0" shape = [256] dtype = "float32" min_val = float("-0.0764245") @@ -5462,6 +5978,7 @@ class Program_weight_tensor_parameter_503: class Program_weight_tensor_parameter_504: name = "parameter_504" + original_name = "batch_norm2d_12.w_2" shape = [256] dtype = "float32" min_val = float("1.47993e-05") @@ -5473,6 +5990,7 @@ class Program_weight_tensor_parameter_504: class Program_weight_tensor_parameter_505: name = "parameter_505" + original_name = "batch_norm2d_12.w_1" shape = [256] dtype = "float32" min_val = float("-0.070885") @@ -5484,6 +6002,7 @@ class Program_weight_tensor_parameter_505: class Program_weight_tensor_parameter_506: name = "parameter_506" + original_name = "conv2d_12.w_0" shape = [256, 64, 1, 1] dtype = "float32" min_val = float("-0.169234") @@ -5495,6 +6014,7 @@ class Program_weight_tensor_parameter_506: class Program_weight_tensor_parameter_507: name = "parameter_507" + original_name = "batch_norm2d_11.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -5504,6 +6024,7 @@ class Program_weight_tensor_parameter_507: class Program_weight_tensor_parameter_508: name = "parameter_508" + original_name = "batch_norm2d_11.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -5513,6 +6034,7 @@ class Program_weight_tensor_parameter_508: class Program_weight_tensor_parameter_509: name = "parameter_509" + original_name = "batch_norm2d_11.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -5522,6 +6044,7 @@ class Program_weight_tensor_parameter_509: class Program_weight_tensor_parameter_510: name = "parameter_510" + original_name = "batch_norm2d_11.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -5531,6 +6054,7 @@ class Program_weight_tensor_parameter_510: class Program_weight_tensor_parameter_511: name = "parameter_511" + original_name = "conv2d_11.w_0" shape = [64, 64, 3, 3] dtype = "float32" min_val = float("-0.197603") @@ -5542,6 +6066,7 @@ class Program_weight_tensor_parameter_511: class Program_weight_tensor_parameter_512: name = "parameter_512" + original_name = "batch_norm2d_10.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -5551,6 +6076,7 @@ class Program_weight_tensor_parameter_512: class Program_weight_tensor_parameter_513: name = "parameter_513" + original_name = "batch_norm2d_10.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -5560,6 +6086,7 @@ class Program_weight_tensor_parameter_513: class Program_weight_tensor_parameter_514: name = "parameter_514" + original_name = "batch_norm2d_10.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -5569,6 +6096,7 @@ class Program_weight_tensor_parameter_514: class Program_weight_tensor_parameter_515: name = "parameter_515" + original_name = "batch_norm2d_10.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -5578,6 +6106,7 @@ class Program_weight_tensor_parameter_515: class Program_weight_tensor_parameter_516: name = "parameter_516" + original_name = "conv2d_10.w_0" shape = [64, 256, 1, 1] dtype = "float32" min_val = float("-0.186245") @@ -5589,6 +6118,7 @@ class Program_weight_tensor_parameter_516: class Program_weight_tensor_parameter_517: name = "parameter_517" + original_name = "batch_norm2d_9.b_0" shape = [256] dtype = "float32" min_val = float("-0.20351") @@ -5600,6 +6130,7 @@ class Program_weight_tensor_parameter_517: class Program_weight_tensor_parameter_518: name = "parameter_518" + original_name = "batch_norm2d_9.w_0" shape = [256] dtype = "float32" min_val = float("-0.0916299") @@ -5611,6 +6142,7 @@ class Program_weight_tensor_parameter_518: class Program_weight_tensor_parameter_519: name = "parameter_519" + original_name = "batch_norm2d_9.w_2" shape = [256] dtype = "float32" min_val = float("5.60519e-45") @@ -5622,6 +6154,7 @@ class Program_weight_tensor_parameter_519: class Program_weight_tensor_parameter_520: name = "parameter_520" + original_name = "batch_norm2d_9.w_1" shape = [256] dtype = "float32" min_val = float("-0.154898") @@ -5633,6 +6166,7 @@ class Program_weight_tensor_parameter_520: class Program_weight_tensor_parameter_521: name = "parameter_521" + original_name = "conv2d_9.w_0" shape = [256, 64, 1, 1] dtype = "float32" min_val = float("-0.228203") @@ -5644,6 +6178,7 @@ class Program_weight_tensor_parameter_521: class Program_weight_tensor_parameter_522: name = "parameter_522" + original_name = "batch_norm2d_8.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -5653,6 +6188,7 @@ class Program_weight_tensor_parameter_522: class Program_weight_tensor_parameter_523: name = "parameter_523" + original_name = "batch_norm2d_8.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -5662,6 +6198,7 @@ class Program_weight_tensor_parameter_523: class Program_weight_tensor_parameter_524: name = "parameter_524" + original_name = "batch_norm2d_8.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -5671,6 +6208,7 @@ class Program_weight_tensor_parameter_524: class Program_weight_tensor_parameter_525: name = "parameter_525" + original_name = "batch_norm2d_8.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -5680,6 +6218,7 @@ class Program_weight_tensor_parameter_525: class Program_weight_tensor_parameter_526: name = "parameter_526" + original_name = "conv2d_8.w_0" shape = [64, 64, 3, 3] dtype = "float32" min_val = float("-0.215985") @@ -5691,6 +6230,7 @@ class Program_weight_tensor_parameter_526: class Program_weight_tensor_parameter_527: name = "parameter_527" + original_name = "batch_norm2d_7.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -5700,6 +6240,7 @@ class Program_weight_tensor_parameter_527: class Program_weight_tensor_parameter_528: name = "parameter_528" + original_name = "batch_norm2d_7.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -5709,6 +6250,7 @@ class Program_weight_tensor_parameter_528: class Program_weight_tensor_parameter_529: name = "parameter_529" + original_name = "batch_norm2d_7.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -5718,6 +6260,7 @@ class Program_weight_tensor_parameter_529: class Program_weight_tensor_parameter_530: name = "parameter_530" + original_name = "batch_norm2d_7.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -5727,6 +6270,7 @@ class Program_weight_tensor_parameter_530: class Program_weight_tensor_parameter_531: name = "parameter_531" + original_name = "conv2d_7.w_0" shape = [64, 256, 1, 1] dtype = "float32" min_val = float("-0.217827") @@ -5738,6 +6282,7 @@ class Program_weight_tensor_parameter_531: class Program_weight_tensor_parameter_532: name = "parameter_532" + original_name = "batch_norm2d_6.b_0" shape = [256] dtype = "float32" min_val = float("-0.126803") @@ -5749,6 +6294,7 @@ class Program_weight_tensor_parameter_532: class Program_weight_tensor_parameter_533: name = "parameter_533" + original_name = "batch_norm2d_6.w_0" shape = [256] dtype = "float32" min_val = float("-0.0899233") @@ -5760,6 +6306,7 @@ class Program_weight_tensor_parameter_533: class Program_weight_tensor_parameter_534: name = "parameter_534" + original_name = "batch_norm2d_6.w_2" shape = [256] dtype = "float32" min_val = float("5.60519e-45") @@ -5771,6 +6318,7 @@ class Program_weight_tensor_parameter_534: class Program_weight_tensor_parameter_535: name = "parameter_535" + original_name = "batch_norm2d_6.w_1" shape = [256] dtype = "float32" min_val = float("-0.527159") @@ -5782,6 +6330,7 @@ class Program_weight_tensor_parameter_535: class Program_weight_tensor_parameter_536: name = "parameter_536" + original_name = "conv2d_6.w_0" shape = [256, 64, 1, 1] dtype = "float32" min_val = float("-0.736433") @@ -5793,6 +6342,7 @@ class Program_weight_tensor_parameter_536: class Program_weight_tensor_parameter_537: name = "parameter_537" + original_name = "batch_norm2d_5.b_0" shape = [256] dtype = "float32" min_val = float("-0.126803") @@ -5804,6 +6354,7 @@ class Program_weight_tensor_parameter_537: class Program_weight_tensor_parameter_538: name = "parameter_538" + original_name = "batch_norm2d_5.w_0" shape = [256] dtype = "float32" min_val = float("-0.0793541") @@ -5815,6 +6366,7 @@ class Program_weight_tensor_parameter_538: class Program_weight_tensor_parameter_539: name = "parameter_539" + original_name = "batch_norm2d_5.w_2" shape = [256] dtype = "float32" min_val = float("5.60519e-45") @@ -5826,6 +6378,7 @@ class Program_weight_tensor_parameter_539: class Program_weight_tensor_parameter_540: name = "parameter_540" + original_name = "batch_norm2d_5.w_1" shape = [256] dtype = "float32" min_val = float("-0.096113") @@ -5837,6 +6390,7 @@ class Program_weight_tensor_parameter_540: class Program_weight_tensor_parameter_541: name = "parameter_541" + original_name = "conv2d_5.w_0" shape = [256, 64, 1, 1] dtype = "float32" min_val = float("-0.22305") @@ -5848,6 +6402,7 @@ class Program_weight_tensor_parameter_541: class Program_weight_tensor_parameter_542: name = "parameter_542" + original_name = "batch_norm2d_4.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -5857,6 +6412,7 @@ class Program_weight_tensor_parameter_542: class Program_weight_tensor_parameter_543: name = "parameter_543" + original_name = "batch_norm2d_4.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -5866,6 +6422,7 @@ class Program_weight_tensor_parameter_543: class Program_weight_tensor_parameter_544: name = "parameter_544" + original_name = "batch_norm2d_4.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -5875,6 +6432,7 @@ class Program_weight_tensor_parameter_544: class Program_weight_tensor_parameter_545: name = "parameter_545" + original_name = "batch_norm2d_4.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -5884,6 +6442,7 @@ class Program_weight_tensor_parameter_545: class Program_weight_tensor_parameter_546: name = "parameter_546" + original_name = "conv2d_4.w_0" shape = [64, 64, 3, 3] dtype = "float32" min_val = float("-0.376788") @@ -5895,6 +6454,7 @@ class Program_weight_tensor_parameter_546: class Program_weight_tensor_parameter_547: name = "parameter_547" + original_name = "batch_norm2d_3.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -5904,6 +6464,7 @@ class Program_weight_tensor_parameter_547: class Program_weight_tensor_parameter_548: name = "parameter_548" + original_name = "batch_norm2d_3.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -5913,6 +6474,7 @@ class Program_weight_tensor_parameter_548: class Program_weight_tensor_parameter_549: name = "parameter_549" + original_name = "batch_norm2d_3.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -5922,6 +6484,7 @@ class Program_weight_tensor_parameter_549: class Program_weight_tensor_parameter_550: name = "parameter_550" + original_name = "batch_norm2d_3.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -5931,6 +6494,7 @@ class Program_weight_tensor_parameter_550: class Program_weight_tensor_parameter_551: name = "parameter_551" + original_name = "conv2d_3.w_0" shape = [64, 64, 1, 1] dtype = "float32" min_val = float("-0.451502") @@ -5942,6 +6506,7 @@ class Program_weight_tensor_parameter_551: class Program_weight_tensor_parameter_552: name = "parameter_552" + original_name = "batch_norm2d_2.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -5951,6 +6516,7 @@ class Program_weight_tensor_parameter_552: class Program_weight_tensor_parameter_553: name = "parameter_553" + original_name = "batch_norm2d_2.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -5960,6 +6526,7 @@ class Program_weight_tensor_parameter_553: class Program_weight_tensor_parameter_554: name = "parameter_554" + original_name = "batch_norm2d_2.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -5969,6 +6536,7 @@ class Program_weight_tensor_parameter_554: class Program_weight_tensor_parameter_555: name = "parameter_555" + original_name = "batch_norm2d_2.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -5978,6 +6546,7 @@ class Program_weight_tensor_parameter_555: class Program_weight_tensor_parameter_556: name = "parameter_556" + original_name = "conv2d_2.w_0" shape = [64, 32, 3, 3] dtype = "float32" min_val = float("-0.457575") @@ -5989,6 +6558,7 @@ class Program_weight_tensor_parameter_556: class Program_weight_tensor_parameter_557: name = "parameter_557" + original_name = "batch_norm2d_1.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -5998,6 +6568,7 @@ class Program_weight_tensor_parameter_557: class Program_weight_tensor_parameter_558: name = "parameter_558" + original_name = "batch_norm2d_1.w_0" shape = [32] dtype = "float32" min_val = float("0") @@ -6007,6 +6578,7 @@ class Program_weight_tensor_parameter_558: class Program_weight_tensor_parameter_559: name = "parameter_559" + original_name = "batch_norm2d_1.w_2" shape = [32] dtype = "float32" min_val = float("0") @@ -6016,6 +6588,7 @@ class Program_weight_tensor_parameter_559: class Program_weight_tensor_parameter_560: name = "parameter_560" + original_name = "batch_norm2d_1.w_1" shape = [32] dtype = "float32" min_val = float("0") @@ -6025,6 +6598,7 @@ class Program_weight_tensor_parameter_560: class Program_weight_tensor_parameter_561: name = "parameter_561" + original_name = "conv2d_1.w_0" shape = [32, 32, 3, 3] dtype = "float32" min_val = float("-0.734028") @@ -6036,6 +6610,7 @@ class Program_weight_tensor_parameter_561: class Program_weight_tensor_parameter_562: name = "parameter_562" + original_name = "batch_norm2d_0.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -6045,6 +6620,7 @@ class Program_weight_tensor_parameter_562: class Program_weight_tensor_parameter_563: name = "parameter_563" + original_name = "batch_norm2d_0.w_0" shape = [32] dtype = "float32" min_val = float("0") @@ -6054,6 +6630,7 @@ class Program_weight_tensor_parameter_563: class Program_weight_tensor_parameter_564: name = "parameter_564" + original_name = "batch_norm2d_0.w_2" shape = [32] dtype = "float32" min_val = float("0") @@ -6063,6 +6640,7 @@ class Program_weight_tensor_parameter_564: class Program_weight_tensor_parameter_565: name = "parameter_565" + original_name = "batch_norm2d_0.w_1" shape = [32] dtype = "float32" min_val = float("0") @@ -6072,6 +6650,7 @@ class Program_weight_tensor_parameter_565: class Program_weight_tensor_parameter_566: name = "parameter_566" + original_name = "conv2d_0.w_0" shape = [32, 3, 3, 3] dtype = "float32" min_val = float("-0.756744") diff --git a/paddle_samples/PaddleX/Deeplabv3-R50/subgraph_0/input_meta.py b/paddle_samples/PaddleX/Deeplabv3-R50/subgraph_0/input_meta.py index a312c2c39..1283ab1c8 100644 --- a/paddle_samples/PaddleX/Deeplabv3-R50/subgraph_0/input_meta.py +++ b/paddle_samples/PaddleX/Deeplabv3-R50/subgraph_0/input_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_data_0: name = "data_0" + original_name = "var_0" shape = [2, 3, 512, 1024] dtype = "float32" min_val = float("-1.0") diff --git a/paddle_samples/PaddleX/Deeplabv3-R50/subgraph_0/weight_meta.py b/paddle_samples/PaddleX/Deeplabv3-R50/subgraph_0/weight_meta.py index 1fd55fb2b..d900b3cd0 100644 --- a/paddle_samples/PaddleX/Deeplabv3-R50/subgraph_0/weight_meta.py +++ b/paddle_samples/PaddleX/Deeplabv3-R50/subgraph_0/weight_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_parameter_0: name = "parameter_0" + original_name = "conv2d_61.b_0" shape = [2] dtype = "float32" min_val = float("0") @@ -9,6 +10,7 @@ class Program_weight_tensor_parameter_0: class Program_weight_tensor_parameter_1: name = "parameter_1" + original_name = "conv2d_61.w_0" shape = [2, 256, 1, 1] dtype = "float32" min_val = float("-0.22829") @@ -20,6 +22,7 @@ class Program_weight_tensor_parameter_1: class Program_weight_tensor_parameter_2: name = "parameter_2" + original_name = "batch_norm2d_60.b_0" shape = [256] dtype = "float32" data = None @@ -27,6 +30,7 @@ class Program_weight_tensor_parameter_2: class Program_weight_tensor_parameter_3: name = "parameter_3" + original_name = "batch_norm2d_60.w_0" shape = [256] dtype = "float32" min_val = float("1.0") @@ -37,6 +41,7 @@ class Program_weight_tensor_parameter_3: class Program_weight_tensor_parameter_4: name = "parameter_4" + original_name = "batch_norm2d_60.w_2" shape = [256] dtype = "float32" min_val = float("1.0") @@ -47,6 +52,7 @@ class Program_weight_tensor_parameter_4: class Program_weight_tensor_parameter_5: name = "parameter_5" + original_name = "batch_norm2d_60.w_1" shape = [256] dtype = "float32" data = None @@ -54,6 +60,7 @@ class Program_weight_tensor_parameter_5: class Program_weight_tensor_parameter_6: name = "parameter_6" + original_name = "conv2d_60.b_0" shape = [256] dtype = "float32" data = None @@ -61,6 +68,7 @@ class Program_weight_tensor_parameter_6: class Program_weight_tensor_parameter_7: name = "parameter_7" + original_name = "conv2d_60.w_0" shape = [256, 1280, 1, 1] dtype = "float32" min_val = float("-0.180365") @@ -72,6 +80,7 @@ class Program_weight_tensor_parameter_7: class Program_weight_tensor_parameter_8: name = "parameter_8" + original_name = "batch_norm2d_59.b_0" shape = [256] dtype = "float32" data = None @@ -79,6 +88,7 @@ class Program_weight_tensor_parameter_8: class Program_weight_tensor_parameter_9: name = "parameter_9" + original_name = "batch_norm2d_59.w_0" shape = [256] dtype = "float32" min_val = float("1.0") @@ -89,6 +99,7 @@ class Program_weight_tensor_parameter_9: class Program_weight_tensor_parameter_10: name = "parameter_10" + original_name = "batch_norm2d_59.w_2" shape = [256] dtype = "float32" min_val = float("1.0") @@ -99,6 +110,7 @@ class Program_weight_tensor_parameter_10: class Program_weight_tensor_parameter_11: name = "parameter_11" + original_name = "batch_norm2d_59.w_1" shape = [256] dtype = "float32" data = None @@ -106,6 +118,7 @@ class Program_weight_tensor_parameter_11: class Program_weight_tensor_parameter_12: name = "parameter_12" + original_name = "conv2d_59.w_0" shape = [256, 2048, 1, 1] dtype = "float32" min_val = float("-0.14635") @@ -117,6 +130,7 @@ class Program_weight_tensor_parameter_12: class Program_weight_tensor_parameter_13: name = "parameter_13" + original_name = "batch_norm2d_58.b_0" shape = [256] dtype = "float32" data = None @@ -124,6 +138,7 @@ class Program_weight_tensor_parameter_13: class Program_weight_tensor_parameter_14: name = "parameter_14" + original_name = "batch_norm2d_58.w_0" shape = [256] dtype = "float32" min_val = float("1.0") @@ -134,6 +149,7 @@ class Program_weight_tensor_parameter_14: class Program_weight_tensor_parameter_15: name = "parameter_15" + original_name = "batch_norm2d_58.w_2" shape = [256] dtype = "float32" min_val = float("1.0") @@ -144,6 +160,7 @@ class Program_weight_tensor_parameter_15: class Program_weight_tensor_parameter_16: name = "parameter_16" + original_name = "batch_norm2d_58.w_1" shape = [256] dtype = "float32" data = None @@ -151,6 +168,7 @@ class Program_weight_tensor_parameter_16: class Program_weight_tensor_parameter_17: name = "parameter_17" + original_name = "conv2d_58.b_0" shape = [256] dtype = "float32" data = None @@ -158,6 +176,7 @@ class Program_weight_tensor_parameter_17: class Program_weight_tensor_parameter_18: name = "parameter_18" + original_name = "conv2d_58.w_0" shape = [256, 2048, 3, 3] dtype = "float32" min_val = float("-0.055991") @@ -169,6 +188,7 @@ class Program_weight_tensor_parameter_18: class Program_weight_tensor_parameter_19: name = "parameter_19" + original_name = "batch_norm2d_57.b_0" shape = [256] dtype = "float32" data = None @@ -176,6 +196,7 @@ class Program_weight_tensor_parameter_19: class Program_weight_tensor_parameter_20: name = "parameter_20" + original_name = "batch_norm2d_57.w_0" shape = [256] dtype = "float32" min_val = float("1.0") @@ -186,6 +207,7 @@ class Program_weight_tensor_parameter_20: class Program_weight_tensor_parameter_21: name = "parameter_21" + original_name = "batch_norm2d_57.w_2" shape = [256] dtype = "float32" min_val = float("1.0") @@ -196,6 +218,7 @@ class Program_weight_tensor_parameter_21: class Program_weight_tensor_parameter_22: name = "parameter_22" + original_name = "batch_norm2d_57.w_1" shape = [256] dtype = "float32" data = None @@ -203,6 +226,7 @@ class Program_weight_tensor_parameter_22: class Program_weight_tensor_parameter_23: name = "parameter_23" + original_name = "conv2d_57.b_0" shape = [256] dtype = "float32" data = None @@ -210,6 +234,7 @@ class Program_weight_tensor_parameter_23: class Program_weight_tensor_parameter_24: name = "parameter_24" + original_name = "conv2d_57.w_0" shape = [256, 2048, 3, 3] dtype = "float32" min_val = float("-0.0518542") @@ -221,6 +246,7 @@ class Program_weight_tensor_parameter_24: class Program_weight_tensor_parameter_25: name = "parameter_25" + original_name = "batch_norm2d_56.b_0" shape = [256] dtype = "float32" data = None @@ -228,6 +254,7 @@ class Program_weight_tensor_parameter_25: class Program_weight_tensor_parameter_26: name = "parameter_26" + original_name = "batch_norm2d_56.w_0" shape = [256] dtype = "float32" min_val = float("1.0") @@ -238,6 +265,7 @@ class Program_weight_tensor_parameter_26: class Program_weight_tensor_parameter_27: name = "parameter_27" + original_name = "batch_norm2d_56.w_2" shape = [256] dtype = "float32" min_val = float("1.0") @@ -248,6 +276,7 @@ class Program_weight_tensor_parameter_27: class Program_weight_tensor_parameter_28: name = "parameter_28" + original_name = "batch_norm2d_56.w_1" shape = [256] dtype = "float32" data = None @@ -255,6 +284,7 @@ class Program_weight_tensor_parameter_28: class Program_weight_tensor_parameter_29: name = "parameter_29" + original_name = "conv2d_56.b_0" shape = [256] dtype = "float32" data = None @@ -262,6 +292,7 @@ class Program_weight_tensor_parameter_29: class Program_weight_tensor_parameter_30: name = "parameter_30" + original_name = "conv2d_56.w_0" shape = [256, 2048, 3, 3] dtype = "float32" min_val = float("-0.0543418") @@ -273,6 +304,7 @@ class Program_weight_tensor_parameter_30: class Program_weight_tensor_parameter_31: name = "parameter_31" + original_name = "batch_norm2d_55.b_0" shape = [256] dtype = "float32" data = None @@ -280,6 +312,7 @@ class Program_weight_tensor_parameter_31: class Program_weight_tensor_parameter_32: name = "parameter_32" + original_name = "batch_norm2d_55.w_0" shape = [256] dtype = "float32" min_val = float("1.0") @@ -290,6 +323,7 @@ class Program_weight_tensor_parameter_32: class Program_weight_tensor_parameter_33: name = "parameter_33" + original_name = "batch_norm2d_55.w_2" shape = [256] dtype = "float32" min_val = float("1.0") @@ -300,6 +334,7 @@ class Program_weight_tensor_parameter_33: class Program_weight_tensor_parameter_34: name = "parameter_34" + original_name = "batch_norm2d_55.w_1" shape = [256] dtype = "float32" data = None @@ -307,6 +342,7 @@ class Program_weight_tensor_parameter_34: class Program_weight_tensor_parameter_35: name = "parameter_35" + original_name = "conv2d_55.b_0" shape = [256] dtype = "float32" data = None @@ -314,6 +350,7 @@ class Program_weight_tensor_parameter_35: class Program_weight_tensor_parameter_36: name = "parameter_36" + original_name = "conv2d_55.w_0" shape = [256, 2048, 1, 1] dtype = "float32" min_val = float("-0.150029") @@ -325,6 +362,7 @@ class Program_weight_tensor_parameter_36: class Program_weight_tensor_parameter_37: name = "parameter_37" + original_name = "batch_norm2d_54.b_0" shape = [2048] dtype = "float32" min_val = float("-0.282166") @@ -336,6 +374,7 @@ class Program_weight_tensor_parameter_37: class Program_weight_tensor_parameter_38: name = "parameter_38" + original_name = "batch_norm2d_54.w_0" shape = [2048] dtype = "float32" min_val = float("0.714089") @@ -347,6 +386,7 @@ class Program_weight_tensor_parameter_38: class Program_weight_tensor_parameter_39: name = "parameter_39" + original_name = "batch_norm2d_54.w_2" shape = [2048] dtype = "float32" min_val = float("0.000386176") @@ -358,6 +398,7 @@ class Program_weight_tensor_parameter_39: class Program_weight_tensor_parameter_40: name = "parameter_40" + original_name = "batch_norm2d_54.w_1" shape = [2048] dtype = "float32" min_val = float("-0.016995") @@ -369,6 +410,7 @@ class Program_weight_tensor_parameter_40: class Program_weight_tensor_parameter_41: name = "parameter_41" + original_name = "conv2d_54.w_0" shape = [2048, 512, 1, 1] dtype = "float32" min_val = float("-0.234447") @@ -380,6 +422,7 @@ class Program_weight_tensor_parameter_41: class Program_weight_tensor_parameter_42: name = "parameter_42" + original_name = "batch_norm2d_53.b_0" shape = [512] dtype = "float32" min_val = float("-0.397288") @@ -391,6 +434,7 @@ class Program_weight_tensor_parameter_42: class Program_weight_tensor_parameter_43: name = "parameter_43" + original_name = "batch_norm2d_53.w_0" shape = [512] dtype = "float32" min_val = float("0.109729") @@ -402,6 +446,7 @@ class Program_weight_tensor_parameter_43: class Program_weight_tensor_parameter_44: name = "parameter_44" + original_name = "batch_norm2d_53.w_2" shape = [512] dtype = "float32" min_val = float("0.00483171") @@ -413,6 +458,7 @@ class Program_weight_tensor_parameter_44: class Program_weight_tensor_parameter_45: name = "parameter_45" + original_name = "batch_norm2d_53.w_1" shape = [512] dtype = "float32" min_val = float("-0.411076") @@ -424,6 +470,7 @@ class Program_weight_tensor_parameter_45: class Program_weight_tensor_parameter_46: name = "parameter_46" + original_name = "conv2d_53.w_0" shape = [512, 512, 3, 3] dtype = "float32" min_val = float("-0.455835") @@ -435,6 +482,7 @@ class Program_weight_tensor_parameter_46: class Program_weight_tensor_parameter_47: name = "parameter_47" + original_name = "batch_norm2d_52.b_0" shape = [512] dtype = "float32" min_val = float("-0.402652") @@ -446,6 +494,7 @@ class Program_weight_tensor_parameter_47: class Program_weight_tensor_parameter_48: name = "parameter_48" + original_name = "batch_norm2d_52.w_0" shape = [512] dtype = "float32" min_val = float("0.0818051") @@ -457,6 +506,7 @@ class Program_weight_tensor_parameter_48: class Program_weight_tensor_parameter_49: name = "parameter_49" + original_name = "batch_norm2d_52.w_2" shape = [512] dtype = "float32" min_val = float("0.00297223") @@ -468,6 +518,7 @@ class Program_weight_tensor_parameter_49: class Program_weight_tensor_parameter_50: name = "parameter_50" + original_name = "batch_norm2d_52.w_1" shape = [512] dtype = "float32" min_val = float("-0.0819767") @@ -479,6 +530,7 @@ class Program_weight_tensor_parameter_50: class Program_weight_tensor_parameter_51: name = "parameter_51" + original_name = "conv2d_52.w_0" shape = [512, 2048, 1, 1] dtype = "float32" min_val = float("-0.791667") @@ -490,6 +542,7 @@ class Program_weight_tensor_parameter_51: class Program_weight_tensor_parameter_52: name = "parameter_52" + original_name = "batch_norm2d_51.b_0" shape = [2048] dtype = "float32" min_val = float("-0.445082") @@ -501,6 +554,7 @@ class Program_weight_tensor_parameter_52: class Program_weight_tensor_parameter_53: name = "parameter_53" + original_name = "batch_norm2d_51.w_0" shape = [2048] dtype = "float32" min_val = float("-0.0510686") @@ -512,6 +566,7 @@ class Program_weight_tensor_parameter_53: class Program_weight_tensor_parameter_54: name = "parameter_54" + original_name = "batch_norm2d_51.w_2" shape = [2048] dtype = "float32" min_val = float("0.000911003") @@ -523,6 +578,7 @@ class Program_weight_tensor_parameter_54: class Program_weight_tensor_parameter_55: name = "parameter_55" + original_name = "batch_norm2d_51.w_1" shape = [2048] dtype = "float32" min_val = float("-0.858086") @@ -534,6 +590,7 @@ class Program_weight_tensor_parameter_55: class Program_weight_tensor_parameter_56: name = "parameter_56" + original_name = "conv2d_51.w_0" shape = [2048, 512, 1, 1] dtype = "float32" min_val = float("-1.21905") @@ -545,6 +602,7 @@ class Program_weight_tensor_parameter_56: class Program_weight_tensor_parameter_57: name = "parameter_57" + original_name = "batch_norm2d_50.b_0" shape = [512] dtype = "float32" min_val = float("-0.334352") @@ -556,6 +614,7 @@ class Program_weight_tensor_parameter_57: class Program_weight_tensor_parameter_58: name = "parameter_58" + original_name = "batch_norm2d_50.w_0" shape = [512] dtype = "float32" min_val = float("0.134261") @@ -567,6 +626,7 @@ class Program_weight_tensor_parameter_58: class Program_weight_tensor_parameter_59: name = "parameter_59" + original_name = "batch_norm2d_50.w_2" shape = [512] dtype = "float32" min_val = float("0.0166705") @@ -578,6 +638,7 @@ class Program_weight_tensor_parameter_59: class Program_weight_tensor_parameter_60: name = "parameter_60" + original_name = "batch_norm2d_50.w_1" shape = [512] dtype = "float32" min_val = float("-0.884374") @@ -589,6 +650,7 @@ class Program_weight_tensor_parameter_60: class Program_weight_tensor_parameter_61: name = "parameter_61" + original_name = "conv2d_50.w_0" shape = [512, 512, 3, 3] dtype = "float32" min_val = float("-0.537283") @@ -600,6 +662,7 @@ class Program_weight_tensor_parameter_61: class Program_weight_tensor_parameter_62: name = "parameter_62" + original_name = "batch_norm2d_49.b_0" shape = [512] dtype = "float32" min_val = float("-0.266159") @@ -611,6 +674,7 @@ class Program_weight_tensor_parameter_62: class Program_weight_tensor_parameter_63: name = "parameter_63" + original_name = "batch_norm2d_49.w_0" shape = [512] dtype = "float32" min_val = float("0.114282") @@ -622,6 +686,7 @@ class Program_weight_tensor_parameter_63: class Program_weight_tensor_parameter_64: name = "parameter_64" + original_name = "batch_norm2d_49.w_2" shape = [512] dtype = "float32" min_val = float("0.00260214") @@ -633,6 +698,7 @@ class Program_weight_tensor_parameter_64: class Program_weight_tensor_parameter_65: name = "parameter_65" + original_name = "batch_norm2d_49.w_1" shape = [512] dtype = "float32" min_val = float("-0.176514") @@ -644,6 +710,7 @@ class Program_weight_tensor_parameter_65: class Program_weight_tensor_parameter_66: name = "parameter_66" + original_name = "conv2d_49.w_0" shape = [512, 2048, 1, 1] dtype = "float32" min_val = float("-0.489221") @@ -655,6 +722,7 @@ class Program_weight_tensor_parameter_66: class Program_weight_tensor_parameter_67: name = "parameter_67" + original_name = "batch_norm2d_48.b_0" shape = [2048] dtype = "float32" min_val = float("-0.284104") @@ -666,6 +734,7 @@ class Program_weight_tensor_parameter_67: class Program_weight_tensor_parameter_68: name = "parameter_68" + original_name = "batch_norm2d_48.w_0" shape = [2048] dtype = "float32" min_val = float("0.0373721") @@ -677,6 +746,7 @@ class Program_weight_tensor_parameter_68: class Program_weight_tensor_parameter_69: name = "parameter_69" + original_name = "batch_norm2d_48.w_2" shape = [2048] dtype = "float32" min_val = float("0.00248082") @@ -688,6 +758,7 @@ class Program_weight_tensor_parameter_69: class Program_weight_tensor_parameter_70: name = "parameter_70" + original_name = "batch_norm2d_48.w_1" shape = [2048] dtype = "float32" min_val = float("-0.556372") @@ -699,6 +770,7 @@ class Program_weight_tensor_parameter_70: class Program_weight_tensor_parameter_71: name = "parameter_71" + original_name = "conv2d_48.w_0" shape = [2048, 1024, 1, 1] dtype = "float32" min_val = float("-0.540116") @@ -710,6 +782,7 @@ class Program_weight_tensor_parameter_71: class Program_weight_tensor_parameter_72: name = "parameter_72" + original_name = "batch_norm2d_47.b_0" shape = [2048] dtype = "float32" min_val = float("-0.284104") @@ -721,6 +794,7 @@ class Program_weight_tensor_parameter_72: class Program_weight_tensor_parameter_73: name = "parameter_73" + original_name = "batch_norm2d_47.w_0" shape = [2048] dtype = "float32" min_val = float("0.0248532") @@ -732,6 +806,7 @@ class Program_weight_tensor_parameter_73: class Program_weight_tensor_parameter_74: name = "parameter_74" + original_name = "batch_norm2d_47.w_2" shape = [2048] dtype = "float32" min_val = float("0.00188508") @@ -743,6 +818,7 @@ class Program_weight_tensor_parameter_74: class Program_weight_tensor_parameter_75: name = "parameter_75" + original_name = "batch_norm2d_47.w_1" shape = [2048] dtype = "float32" min_val = float("-0.145631") @@ -754,6 +830,7 @@ class Program_weight_tensor_parameter_75: class Program_weight_tensor_parameter_76: name = "parameter_76" + original_name = "conv2d_47.w_0" shape = [2048, 512, 1, 1] dtype = "float32" min_val = float("-0.547884") @@ -765,6 +842,7 @@ class Program_weight_tensor_parameter_76: class Program_weight_tensor_parameter_77: name = "parameter_77" + original_name = "batch_norm2d_46.b_0" shape = [512] dtype = "float32" min_val = float("-0.267678") @@ -776,6 +854,7 @@ class Program_weight_tensor_parameter_77: class Program_weight_tensor_parameter_78: name = "parameter_78" + original_name = "batch_norm2d_46.w_0" shape = [512] dtype = "float32" min_val = float("0.115875") @@ -787,6 +866,7 @@ class Program_weight_tensor_parameter_78: class Program_weight_tensor_parameter_79: name = "parameter_79" + original_name = "batch_norm2d_46.w_2" shape = [512] dtype = "float32" min_val = float("0.00683442") @@ -798,6 +878,7 @@ class Program_weight_tensor_parameter_79: class Program_weight_tensor_parameter_80: name = "parameter_80" + original_name = "batch_norm2d_46.w_1" shape = [512] dtype = "float32" min_val = float("-0.356659") @@ -809,6 +890,7 @@ class Program_weight_tensor_parameter_80: class Program_weight_tensor_parameter_81: name = "parameter_81" + original_name = "conv2d_46.w_0" shape = [512, 512, 3, 3] dtype = "float32" min_val = float("-0.290132") @@ -820,6 +902,7 @@ class Program_weight_tensor_parameter_81: class Program_weight_tensor_parameter_82: name = "parameter_82" + original_name = "batch_norm2d_45.b_0" shape = [512] dtype = "float32" min_val = float("-0.455306") @@ -831,6 +914,7 @@ class Program_weight_tensor_parameter_82: class Program_weight_tensor_parameter_83: name = "parameter_83" + original_name = "batch_norm2d_45.w_0" shape = [512] dtype = "float32" min_val = float("0.0956012") @@ -842,6 +926,7 @@ class Program_weight_tensor_parameter_83: class Program_weight_tensor_parameter_84: name = "parameter_84" + original_name = "batch_norm2d_45.w_2" shape = [512] dtype = "float32" min_val = float("0.00995626") @@ -853,6 +938,7 @@ class Program_weight_tensor_parameter_84: class Program_weight_tensor_parameter_85: name = "parameter_85" + original_name = "batch_norm2d_45.w_1" shape = [512] dtype = "float32" min_val = float("-0.272144") @@ -864,6 +950,7 @@ class Program_weight_tensor_parameter_85: class Program_weight_tensor_parameter_86: name = "parameter_86" + original_name = "conv2d_45.w_0" shape = [512, 1024, 1, 1] dtype = "float32" min_val = float("-0.254182") @@ -875,6 +962,7 @@ class Program_weight_tensor_parameter_86: class Program_weight_tensor_parameter_87: name = "parameter_87" + original_name = "batch_norm2d_44.b_0" shape = [1024] dtype = "float32" min_val = float("-0.494356") @@ -886,6 +974,7 @@ class Program_weight_tensor_parameter_87: class Program_weight_tensor_parameter_88: name = "parameter_88" + original_name = "batch_norm2d_44.w_0" shape = [1024] dtype = "float32" min_val = float("-0.114335") @@ -897,6 +986,7 @@ class Program_weight_tensor_parameter_88: class Program_weight_tensor_parameter_89: name = "parameter_89" + original_name = "batch_norm2d_44.w_2" shape = [1024] dtype = "float32" min_val = float("0.000758928") @@ -908,6 +998,7 @@ class Program_weight_tensor_parameter_89: class Program_weight_tensor_parameter_90: name = "parameter_90" + original_name = "batch_norm2d_44.w_1" shape = [1024] dtype = "float32" min_val = float("-0.196626") @@ -919,6 +1010,7 @@ class Program_weight_tensor_parameter_90: class Program_weight_tensor_parameter_91: name = "parameter_91" + original_name = "conv2d_44.w_0" shape = [1024, 256, 1, 1] dtype = "float32" min_val = float("-0.516216") @@ -930,6 +1022,7 @@ class Program_weight_tensor_parameter_91: class Program_weight_tensor_parameter_92: name = "parameter_92" + original_name = "batch_norm2d_43.b_0" shape = [256] dtype = "float32" min_val = float("-0.628594") @@ -941,6 +1034,7 @@ class Program_weight_tensor_parameter_92: class Program_weight_tensor_parameter_93: name = "parameter_93" + original_name = "batch_norm2d_43.w_0" shape = [256] dtype = "float32" min_val = float("0.129324") @@ -952,6 +1046,7 @@ class Program_weight_tensor_parameter_93: class Program_weight_tensor_parameter_94: name = "parameter_94" + original_name = "batch_norm2d_43.w_2" shape = [256] dtype = "float32" min_val = float("0.0113363") @@ -963,6 +1058,7 @@ class Program_weight_tensor_parameter_94: class Program_weight_tensor_parameter_95: name = "parameter_95" + original_name = "batch_norm2d_43.w_1" shape = [256] dtype = "float32" min_val = float("-0.469623") @@ -974,6 +1070,7 @@ class Program_weight_tensor_parameter_95: class Program_weight_tensor_parameter_96: name = "parameter_96" + original_name = "conv2d_43.w_0" shape = [256, 256, 3, 3] dtype = "float32" min_val = float("-0.439431") @@ -985,6 +1082,7 @@ class Program_weight_tensor_parameter_96: class Program_weight_tensor_parameter_97: name = "parameter_97" + original_name = "batch_norm2d_42.b_0" shape = [256] dtype = "float32" min_val = float("-0.465171") @@ -996,6 +1094,7 @@ class Program_weight_tensor_parameter_97: class Program_weight_tensor_parameter_98: name = "parameter_98" + original_name = "batch_norm2d_42.w_0" shape = [256] dtype = "float32" min_val = float("0.119903") @@ -1007,6 +1106,7 @@ class Program_weight_tensor_parameter_98: class Program_weight_tensor_parameter_99: name = "parameter_99" + original_name = "batch_norm2d_42.w_2" shape = [256] dtype = "float32" min_val = float("0.0136785") @@ -1018,6 +1118,7 @@ class Program_weight_tensor_parameter_99: class Program_weight_tensor_parameter_100: name = "parameter_100" + original_name = "batch_norm2d_42.w_1" shape = [256] dtype = "float32" min_val = float("-0.267117") @@ -1029,6 +1130,7 @@ class Program_weight_tensor_parameter_100: class Program_weight_tensor_parameter_101: name = "parameter_101" + original_name = "conv2d_42.w_0" shape = [256, 1024, 1, 1] dtype = "float32" min_val = float("-0.386117") @@ -1040,6 +1142,7 @@ class Program_weight_tensor_parameter_101: class Program_weight_tensor_parameter_102: name = "parameter_102" + original_name = "batch_norm2d_41.b_0" shape = [1024] dtype = "float32" min_val = float("-0.505123") @@ -1051,6 +1154,7 @@ class Program_weight_tensor_parameter_102: class Program_weight_tensor_parameter_103: name = "parameter_103" + original_name = "batch_norm2d_41.w_0" shape = [1024] dtype = "float32" min_val = float("-0.071253") @@ -1062,6 +1166,7 @@ class Program_weight_tensor_parameter_103: class Program_weight_tensor_parameter_104: name = "parameter_104" + original_name = "batch_norm2d_41.w_2" shape = [1024] dtype = "float32" min_val = float("0.000556931") @@ -1073,6 +1178,7 @@ class Program_weight_tensor_parameter_104: class Program_weight_tensor_parameter_105: name = "parameter_105" + original_name = "batch_norm2d_41.w_1" shape = [1024] dtype = "float32" min_val = float("-0.194132") @@ -1084,6 +1190,7 @@ class Program_weight_tensor_parameter_105: class Program_weight_tensor_parameter_106: name = "parameter_106" + original_name = "conv2d_41.w_0" shape = [1024, 256, 1, 1] dtype = "float32" min_val = float("-0.680876") @@ -1095,6 +1202,7 @@ class Program_weight_tensor_parameter_106: class Program_weight_tensor_parameter_107: name = "parameter_107" + original_name = "batch_norm2d_40.b_0" shape = [256] dtype = "float32" min_val = float("-0.69082") @@ -1106,6 +1214,7 @@ class Program_weight_tensor_parameter_107: class Program_weight_tensor_parameter_108: name = "parameter_108" + original_name = "batch_norm2d_40.w_0" shape = [256] dtype = "float32" min_val = float("0.126233") @@ -1117,6 +1226,7 @@ class Program_weight_tensor_parameter_108: class Program_weight_tensor_parameter_109: name = "parameter_109" + original_name = "batch_norm2d_40.w_2" shape = [256] dtype = "float32" min_val = float("0.0117291") @@ -1128,6 +1238,7 @@ class Program_weight_tensor_parameter_109: class Program_weight_tensor_parameter_110: name = "parameter_110" + original_name = "batch_norm2d_40.w_1" shape = [256] dtype = "float32" min_val = float("-0.434858") @@ -1139,6 +1250,7 @@ class Program_weight_tensor_parameter_110: class Program_weight_tensor_parameter_111: name = "parameter_111" + original_name = "conv2d_40.w_0" shape = [256, 256, 3, 3] dtype = "float32" min_val = float("-0.308659") @@ -1150,6 +1262,7 @@ class Program_weight_tensor_parameter_111: class Program_weight_tensor_parameter_112: name = "parameter_112" + original_name = "batch_norm2d_39.b_0" shape = [256] dtype = "float32" min_val = float("-0.490268") @@ -1161,6 +1274,7 @@ class Program_weight_tensor_parameter_112: class Program_weight_tensor_parameter_113: name = "parameter_113" + original_name = "batch_norm2d_39.w_0" shape = [256] dtype = "float32" min_val = float("0.123382") @@ -1172,6 +1286,7 @@ class Program_weight_tensor_parameter_113: class Program_weight_tensor_parameter_114: name = "parameter_114" + original_name = "batch_norm2d_39.w_2" shape = [256] dtype = "float32" min_val = float("0.0127077") @@ -1183,6 +1298,7 @@ class Program_weight_tensor_parameter_114: class Program_weight_tensor_parameter_115: name = "parameter_115" + original_name = "batch_norm2d_39.w_1" shape = [256] dtype = "float32" min_val = float("-0.231346") @@ -1194,6 +1310,7 @@ class Program_weight_tensor_parameter_115: class Program_weight_tensor_parameter_116: name = "parameter_116" + original_name = "conv2d_39.w_0" shape = [256, 1024, 1, 1] dtype = "float32" min_val = float("-0.460862") @@ -1205,6 +1322,7 @@ class Program_weight_tensor_parameter_116: class Program_weight_tensor_parameter_117: name = "parameter_117" + original_name = "batch_norm2d_38.b_0" shape = [1024] dtype = "float32" min_val = float("-0.351691") @@ -1216,6 +1334,7 @@ class Program_weight_tensor_parameter_117: class Program_weight_tensor_parameter_118: name = "parameter_118" + original_name = "batch_norm2d_38.w_0" shape = [1024] dtype = "float32" min_val = float("-0.0798891") @@ -1227,6 +1346,7 @@ class Program_weight_tensor_parameter_118: class Program_weight_tensor_parameter_119: name = "parameter_119" + original_name = "batch_norm2d_38.w_2" shape = [1024] dtype = "float32" min_val = float("0.00029933") @@ -1238,6 +1358,7 @@ class Program_weight_tensor_parameter_119: class Program_weight_tensor_parameter_120: name = "parameter_120" + original_name = "batch_norm2d_38.w_1" shape = [1024] dtype = "float32" min_val = float("-0.135183") @@ -1249,6 +1370,7 @@ class Program_weight_tensor_parameter_120: class Program_weight_tensor_parameter_121: name = "parameter_121" + original_name = "conv2d_38.w_0" shape = [1024, 256, 1, 1] dtype = "float32" min_val = float("-0.412226") @@ -1260,6 +1382,7 @@ class Program_weight_tensor_parameter_121: class Program_weight_tensor_parameter_122: name = "parameter_122" + original_name = "batch_norm2d_37.b_0" shape = [256] dtype = "float32" min_val = float("-0.327533") @@ -1271,6 +1394,7 @@ class Program_weight_tensor_parameter_122: class Program_weight_tensor_parameter_123: name = "parameter_123" + original_name = "batch_norm2d_37.w_0" shape = [256] dtype = "float32" min_val = float("0.124773") @@ -1282,6 +1406,7 @@ class Program_weight_tensor_parameter_123: class Program_weight_tensor_parameter_124: name = "parameter_124" + original_name = "batch_norm2d_37.w_2" shape = [256] dtype = "float32" min_val = float("0.0132998") @@ -1293,6 +1418,7 @@ class Program_weight_tensor_parameter_124: class Program_weight_tensor_parameter_125: name = "parameter_125" + original_name = "batch_norm2d_37.w_1" shape = [256] dtype = "float32" min_val = float("-0.291819") @@ -1304,6 +1430,7 @@ class Program_weight_tensor_parameter_125: class Program_weight_tensor_parameter_126: name = "parameter_126" + original_name = "conv2d_37.w_0" shape = [256, 256, 3, 3] dtype = "float32" min_val = float("-0.389441") @@ -1315,6 +1442,7 @@ class Program_weight_tensor_parameter_126: class Program_weight_tensor_parameter_127: name = "parameter_127" + original_name = "batch_norm2d_36.b_0" shape = [256] dtype = "float32" min_val = float("-0.510875") @@ -1326,6 +1454,7 @@ class Program_weight_tensor_parameter_127: class Program_weight_tensor_parameter_128: name = "parameter_128" + original_name = "batch_norm2d_36.w_0" shape = [256] dtype = "float32" min_val = float("0.117607") @@ -1337,6 +1466,7 @@ class Program_weight_tensor_parameter_128: class Program_weight_tensor_parameter_129: name = "parameter_129" + original_name = "batch_norm2d_36.w_2" shape = [256] dtype = "float32" min_val = float("0.0131457") @@ -1348,6 +1478,7 @@ class Program_weight_tensor_parameter_129: class Program_weight_tensor_parameter_130: name = "parameter_130" + original_name = "batch_norm2d_36.w_1" shape = [256] dtype = "float32" min_val = float("-0.320565") @@ -1359,6 +1490,7 @@ class Program_weight_tensor_parameter_130: class Program_weight_tensor_parameter_131: name = "parameter_131" + original_name = "conv2d_36.w_0" shape = [256, 1024, 1, 1] dtype = "float32" min_val = float("-0.415199") @@ -1370,6 +1502,7 @@ class Program_weight_tensor_parameter_131: class Program_weight_tensor_parameter_132: name = "parameter_132" + original_name = "batch_norm2d_35.b_0" shape = [1024] dtype = "float32" min_val = float("-0.420672") @@ -1381,6 +1514,7 @@ class Program_weight_tensor_parameter_132: class Program_weight_tensor_parameter_133: name = "parameter_133" + original_name = "batch_norm2d_35.w_0" shape = [1024] dtype = "float32" min_val = float("-0.0631926") @@ -1392,6 +1526,7 @@ class Program_weight_tensor_parameter_133: class Program_weight_tensor_parameter_134: name = "parameter_134" + original_name = "batch_norm2d_35.w_2" shape = [1024] dtype = "float32" min_val = float("0.000697383") @@ -1403,6 +1538,7 @@ class Program_weight_tensor_parameter_134: class Program_weight_tensor_parameter_135: name = "parameter_135" + original_name = "batch_norm2d_35.w_1" shape = [1024] dtype = "float32" min_val = float("-0.161883") @@ -1414,6 +1550,7 @@ class Program_weight_tensor_parameter_135: class Program_weight_tensor_parameter_136: name = "parameter_136" + original_name = "conv2d_35.w_0" shape = [1024, 256, 1, 1] dtype = "float32" min_val = float("-0.411167") @@ -1425,6 +1562,7 @@ class Program_weight_tensor_parameter_136: class Program_weight_tensor_parameter_137: name = "parameter_137" + original_name = "batch_norm2d_34.b_0" shape = [256] dtype = "float32" min_val = float("-0.280491") @@ -1436,6 +1574,7 @@ class Program_weight_tensor_parameter_137: class Program_weight_tensor_parameter_138: name = "parameter_138" + original_name = "batch_norm2d_34.w_0" shape = [256] dtype = "float32" min_val = float("0.128527") @@ -1447,6 +1586,7 @@ class Program_weight_tensor_parameter_138: class Program_weight_tensor_parameter_139: name = "parameter_139" + original_name = "batch_norm2d_34.w_2" shape = [256] dtype = "float32" min_val = float("0.0181118") @@ -1458,6 +1598,7 @@ class Program_weight_tensor_parameter_139: class Program_weight_tensor_parameter_140: name = "parameter_140" + original_name = "batch_norm2d_34.w_1" shape = [256] dtype = "float32" min_val = float("-0.608816") @@ -1469,6 +1610,7 @@ class Program_weight_tensor_parameter_140: class Program_weight_tensor_parameter_141: name = "parameter_141" + original_name = "conv2d_34.w_0" shape = [256, 256, 3, 3] dtype = "float32" min_val = float("-0.29439") @@ -1480,6 +1622,7 @@ class Program_weight_tensor_parameter_141: class Program_weight_tensor_parameter_142: name = "parameter_142" + original_name = "batch_norm2d_33.b_0" shape = [256] dtype = "float32" min_val = float("-0.411786") @@ -1491,6 +1634,7 @@ class Program_weight_tensor_parameter_142: class Program_weight_tensor_parameter_143: name = "parameter_143" + original_name = "batch_norm2d_33.w_0" shape = [256] dtype = "float32" min_val = float("0.122476") @@ -1502,6 +1646,7 @@ class Program_weight_tensor_parameter_143: class Program_weight_tensor_parameter_144: name = "parameter_144" + original_name = "batch_norm2d_33.w_2" shape = [256] dtype = "float32" min_val = float("0.0136736") @@ -1513,6 +1658,7 @@ class Program_weight_tensor_parameter_144: class Program_weight_tensor_parameter_145: name = "parameter_145" + original_name = "batch_norm2d_33.w_1" shape = [256] dtype = "float32" min_val = float("-0.350375") @@ -1524,6 +1670,7 @@ class Program_weight_tensor_parameter_145: class Program_weight_tensor_parameter_146: name = "parameter_146" + original_name = "conv2d_33.w_0" shape = [256, 1024, 1, 1] dtype = "float32" min_val = float("-0.562702") @@ -1535,6 +1682,7 @@ class Program_weight_tensor_parameter_146: class Program_weight_tensor_parameter_147: name = "parameter_147" + original_name = "batch_norm2d_32.b_0" shape = [1024] dtype = "float32" min_val = float("-0.448255") @@ -1546,6 +1694,7 @@ class Program_weight_tensor_parameter_147: class Program_weight_tensor_parameter_148: name = "parameter_148" + original_name = "batch_norm2d_32.w_0" shape = [1024] dtype = "float32" min_val = float("-1.00724e-23") @@ -1557,6 +1706,7 @@ class Program_weight_tensor_parameter_148: class Program_weight_tensor_parameter_149: name = "parameter_149" + original_name = "batch_norm2d_32.w_2" shape = [1024] dtype = "float32" min_val = float("5.60519e-45") @@ -1568,6 +1718,7 @@ class Program_weight_tensor_parameter_149: class Program_weight_tensor_parameter_150: name = "parameter_150" + original_name = "batch_norm2d_32.w_1" shape = [1024] dtype = "float32" min_val = float("-0.238083") @@ -1579,6 +1730,7 @@ class Program_weight_tensor_parameter_150: class Program_weight_tensor_parameter_151: name = "parameter_151" + original_name = "conv2d_32.w_0" shape = [1024, 256, 1, 1] dtype = "float32" min_val = float("-0.723318") @@ -1590,6 +1742,7 @@ class Program_weight_tensor_parameter_151: class Program_weight_tensor_parameter_152: name = "parameter_152" + original_name = "batch_norm2d_31.b_0" shape = [256] dtype = "float32" min_val = float("-0.946089") @@ -1601,6 +1754,7 @@ class Program_weight_tensor_parameter_152: class Program_weight_tensor_parameter_153: name = "parameter_153" + original_name = "batch_norm2d_31.w_0" shape = [256] dtype = "float32" min_val = float("0.140842") @@ -1612,6 +1766,7 @@ class Program_weight_tensor_parameter_153: class Program_weight_tensor_parameter_154: name = "parameter_154" + original_name = "batch_norm2d_31.w_2" shape = [256] dtype = "float32" min_val = float("0.0257037") @@ -1623,6 +1778,7 @@ class Program_weight_tensor_parameter_154: class Program_weight_tensor_parameter_155: name = "parameter_155" + original_name = "batch_norm2d_31.w_1" shape = [256] dtype = "float32" min_val = float("-1.34447") @@ -1634,6 +1790,7 @@ class Program_weight_tensor_parameter_155: class Program_weight_tensor_parameter_156: name = "parameter_156" + original_name = "conv2d_31.w_0" shape = [256, 256, 3, 3] dtype = "float32" min_val = float("-0.594571") @@ -1645,6 +1802,7 @@ class Program_weight_tensor_parameter_156: class Program_weight_tensor_parameter_157: name = "parameter_157" + original_name = "batch_norm2d_30.b_0" shape = [256] dtype = "float32" min_val = float("-0.350109") @@ -1656,6 +1814,7 @@ class Program_weight_tensor_parameter_157: class Program_weight_tensor_parameter_158: name = "parameter_158" + original_name = "batch_norm2d_30.w_0" shape = [256] dtype = "float32" min_val = float("0.132203") @@ -1667,6 +1826,7 @@ class Program_weight_tensor_parameter_158: class Program_weight_tensor_parameter_159: name = "parameter_159" + original_name = "batch_norm2d_30.w_2" shape = [256] dtype = "float32" min_val = float("0.0143599") @@ -1678,6 +1838,7 @@ class Program_weight_tensor_parameter_159: class Program_weight_tensor_parameter_160: name = "parameter_160" + original_name = "batch_norm2d_30.w_1" shape = [256] dtype = "float32" min_val = float("-1.17498") @@ -1689,6 +1850,7 @@ class Program_weight_tensor_parameter_160: class Program_weight_tensor_parameter_161: name = "parameter_161" + original_name = "conv2d_30.w_0" shape = [256, 1024, 1, 1] dtype = "float32" min_val = float("-0.460684") @@ -1700,6 +1862,7 @@ class Program_weight_tensor_parameter_161: class Program_weight_tensor_parameter_162: name = "parameter_162" + original_name = "batch_norm2d_29.b_0" shape = [1024] dtype = "float32" min_val = float("-0.297989") @@ -1711,6 +1874,7 @@ class Program_weight_tensor_parameter_162: class Program_weight_tensor_parameter_163: name = "parameter_163" + original_name = "batch_norm2d_29.w_0" shape = [1024] dtype = "float32" min_val = float("-2.10863e-13") @@ -1722,6 +1886,7 @@ class Program_weight_tensor_parameter_163: class Program_weight_tensor_parameter_164: name = "parameter_164" + original_name = "batch_norm2d_29.w_2" shape = [1024] dtype = "float32" min_val = float("5.60519e-45") @@ -1733,6 +1898,7 @@ class Program_weight_tensor_parameter_164: class Program_weight_tensor_parameter_165: name = "parameter_165" + original_name = "batch_norm2d_29.w_1" shape = [1024] dtype = "float32" min_val = float("-0.488374") @@ -1744,6 +1910,7 @@ class Program_weight_tensor_parameter_165: class Program_weight_tensor_parameter_166: name = "parameter_166" + original_name = "conv2d_29.w_0" shape = [1024, 512, 1, 1] dtype = "float32" min_val = float("-0.714351") @@ -1755,6 +1922,7 @@ class Program_weight_tensor_parameter_166: class Program_weight_tensor_parameter_167: name = "parameter_167" + original_name = "batch_norm2d_28.b_0" shape = [1024] dtype = "float32" min_val = float("-0.297989") @@ -1766,6 +1934,7 @@ class Program_weight_tensor_parameter_167: class Program_weight_tensor_parameter_168: name = "parameter_168" + original_name = "batch_norm2d_28.w_0" shape = [1024] dtype = "float32" min_val = float("-0.0884224") @@ -1777,6 +1946,7 @@ class Program_weight_tensor_parameter_168: class Program_weight_tensor_parameter_169: name = "parameter_169" + original_name = "batch_norm2d_28.w_2" shape = [1024] dtype = "float32" min_val = float("5.60519e-45") @@ -1788,6 +1958,7 @@ class Program_weight_tensor_parameter_169: class Program_weight_tensor_parameter_170: name = "parameter_170" + original_name = "batch_norm2d_28.w_1" shape = [1024] dtype = "float32" min_val = float("-0.2498") @@ -1799,6 +1970,7 @@ class Program_weight_tensor_parameter_170: class Program_weight_tensor_parameter_171: name = "parameter_171" + original_name = "conv2d_28.w_0" shape = [1024, 256, 1, 1] dtype = "float32" min_val = float("-0.315775") @@ -1810,6 +1982,7 @@ class Program_weight_tensor_parameter_171: class Program_weight_tensor_parameter_172: name = "parameter_172" + original_name = "batch_norm2d_27.b_0" shape = [256] dtype = "float32" min_val = float("-0.403554") @@ -1821,6 +1994,7 @@ class Program_weight_tensor_parameter_172: class Program_weight_tensor_parameter_173: name = "parameter_173" + original_name = "batch_norm2d_27.w_0" shape = [256] dtype = "float32" min_val = float("0.119874") @@ -1832,6 +2006,7 @@ class Program_weight_tensor_parameter_173: class Program_weight_tensor_parameter_174: name = "parameter_174" + original_name = "batch_norm2d_27.w_2" shape = [256] dtype = "float32" min_val = float("0.00748281") @@ -1843,6 +2018,7 @@ class Program_weight_tensor_parameter_174: class Program_weight_tensor_parameter_175: name = "parameter_175" + original_name = "batch_norm2d_27.w_1" shape = [256] dtype = "float32" min_val = float("-0.210233") @@ -1854,6 +2030,7 @@ class Program_weight_tensor_parameter_175: class Program_weight_tensor_parameter_176: name = "parameter_176" + original_name = "conv2d_27.w_0" shape = [256, 256, 3, 3] dtype = "float32" min_val = float("-0.373469") @@ -1865,6 +2042,7 @@ class Program_weight_tensor_parameter_176: class Program_weight_tensor_parameter_177: name = "parameter_177" + original_name = "batch_norm2d_26.b_0" shape = [256] dtype = "float32" min_val = float("-0.472221") @@ -1876,6 +2054,7 @@ class Program_weight_tensor_parameter_177: class Program_weight_tensor_parameter_178: name = "parameter_178" + original_name = "batch_norm2d_26.w_0" shape = [256] dtype = "float32" min_val = float("0.128796") @@ -1887,6 +2066,7 @@ class Program_weight_tensor_parameter_178: class Program_weight_tensor_parameter_179: name = "parameter_179" + original_name = "batch_norm2d_26.w_2" shape = [256] dtype = "float32" min_val = float("0.0166948") @@ -1898,6 +2078,7 @@ class Program_weight_tensor_parameter_179: class Program_weight_tensor_parameter_180: name = "parameter_180" + original_name = "batch_norm2d_26.w_1" shape = [256] dtype = "float32" min_val = float("-0.336503") @@ -1909,6 +2090,7 @@ class Program_weight_tensor_parameter_180: class Program_weight_tensor_parameter_181: name = "parameter_181" + original_name = "conv2d_26.w_0" shape = [256, 512, 1, 1] dtype = "float32" min_val = float("-0.286994") @@ -1920,6 +2102,7 @@ class Program_weight_tensor_parameter_181: class Program_weight_tensor_parameter_182: name = "parameter_182" + original_name = "batch_norm2d_25.b_0" shape = [512] dtype = "float32" min_val = float("-0.351865") @@ -1931,6 +2114,7 @@ class Program_weight_tensor_parameter_182: class Program_weight_tensor_parameter_183: name = "parameter_183" + original_name = "batch_norm2d_25.w_0" shape = [512] dtype = "float32" min_val = float("-0.150051") @@ -1942,6 +2126,7 @@ class Program_weight_tensor_parameter_183: class Program_weight_tensor_parameter_184: name = "parameter_184" + original_name = "batch_norm2d_25.w_2" shape = [512] dtype = "float32" min_val = float("0.000305778") @@ -1953,6 +2138,7 @@ class Program_weight_tensor_parameter_184: class Program_weight_tensor_parameter_185: name = "parameter_185" + original_name = "batch_norm2d_25.w_1" shape = [512] dtype = "float32" min_val = float("-0.182206") @@ -1964,6 +2150,7 @@ class Program_weight_tensor_parameter_185: class Program_weight_tensor_parameter_186: name = "parameter_186" + original_name = "conv2d_25.w_0" shape = [512, 128, 1, 1] dtype = "float32" min_val = float("-0.437671") @@ -1975,6 +2162,7 @@ class Program_weight_tensor_parameter_186: class Program_weight_tensor_parameter_187: name = "parameter_187" + original_name = "batch_norm2d_24.b_0" shape = [128] dtype = "float32" min_val = float("-0.288327") @@ -1986,6 +2174,7 @@ class Program_weight_tensor_parameter_187: class Program_weight_tensor_parameter_188: name = "parameter_188" + original_name = "batch_norm2d_24.w_0" shape = [128] dtype = "float32" min_val = float("0.13864") @@ -1997,6 +2186,7 @@ class Program_weight_tensor_parameter_188: class Program_weight_tensor_parameter_189: name = "parameter_189" + original_name = "batch_norm2d_24.w_2" shape = [128] dtype = "float32" min_val = float("0.0139091") @@ -2008,6 +2198,7 @@ class Program_weight_tensor_parameter_189: class Program_weight_tensor_parameter_190: name = "parameter_190" + original_name = "batch_norm2d_24.w_1" shape = [128] dtype = "float32" min_val = float("-0.372374") @@ -2019,6 +2210,7 @@ class Program_weight_tensor_parameter_190: class Program_weight_tensor_parameter_191: name = "parameter_191" + original_name = "conv2d_24.w_0" shape = [128, 128, 3, 3] dtype = "float32" min_val = float("-0.304806") @@ -2030,6 +2222,7 @@ class Program_weight_tensor_parameter_191: class Program_weight_tensor_parameter_192: name = "parameter_192" + original_name = "batch_norm2d_23.b_0" shape = [128] dtype = "float32" min_val = float("-0.394085") @@ -2041,6 +2234,7 @@ class Program_weight_tensor_parameter_192: class Program_weight_tensor_parameter_193: name = "parameter_193" + original_name = "batch_norm2d_23.w_0" shape = [128] dtype = "float32" min_val = float("0.114849") @@ -2052,6 +2246,7 @@ class Program_weight_tensor_parameter_193: class Program_weight_tensor_parameter_194: name = "parameter_194" + original_name = "batch_norm2d_23.w_2" shape = [128] dtype = "float32" min_val = float("0.0206183") @@ -2063,6 +2258,7 @@ class Program_weight_tensor_parameter_194: class Program_weight_tensor_parameter_195: name = "parameter_195" + original_name = "batch_norm2d_23.w_1" shape = [128] dtype = "float32" min_val = float("-0.508035") @@ -2074,6 +2270,7 @@ class Program_weight_tensor_parameter_195: class Program_weight_tensor_parameter_196: name = "parameter_196" + original_name = "conv2d_23.w_0" shape = [128, 512, 1, 1] dtype = "float32" min_val = float("-0.338475") @@ -2085,6 +2282,7 @@ class Program_weight_tensor_parameter_196: class Program_weight_tensor_parameter_197: name = "parameter_197" + original_name = "batch_norm2d_22.b_0" shape = [512] dtype = "float32" min_val = float("-0.278261") @@ -2096,6 +2294,7 @@ class Program_weight_tensor_parameter_197: class Program_weight_tensor_parameter_198: name = "parameter_198" + original_name = "batch_norm2d_22.w_0" shape = [512] dtype = "float32" min_val = float("-0.0943411") @@ -2107,6 +2306,7 @@ class Program_weight_tensor_parameter_198: class Program_weight_tensor_parameter_199: name = "parameter_199" + original_name = "batch_norm2d_22.w_2" shape = [512] dtype = "float32" min_val = float("1.28703e-35") @@ -2118,6 +2318,7 @@ class Program_weight_tensor_parameter_199: class Program_weight_tensor_parameter_200: name = "parameter_200" + original_name = "batch_norm2d_22.w_1" shape = [512] dtype = "float32" min_val = float("-0.222267") @@ -2129,6 +2330,7 @@ class Program_weight_tensor_parameter_200: class Program_weight_tensor_parameter_201: name = "parameter_201" + original_name = "conv2d_22.w_0" shape = [512, 128, 1, 1] dtype = "float32" min_val = float("-0.394291") @@ -2140,6 +2342,7 @@ class Program_weight_tensor_parameter_201: class Program_weight_tensor_parameter_202: name = "parameter_202" + original_name = "batch_norm2d_21.b_0" shape = [128] dtype = "float32" min_val = float("-0.574827") @@ -2151,6 +2354,7 @@ class Program_weight_tensor_parameter_202: class Program_weight_tensor_parameter_203: name = "parameter_203" + original_name = "batch_norm2d_21.w_0" shape = [128] dtype = "float32" min_val = float("0.145856") @@ -2162,6 +2366,7 @@ class Program_weight_tensor_parameter_203: class Program_weight_tensor_parameter_204: name = "parameter_204" + original_name = "batch_norm2d_21.w_2" shape = [128] dtype = "float32" min_val = float("0.0140519") @@ -2173,6 +2378,7 @@ class Program_weight_tensor_parameter_204: class Program_weight_tensor_parameter_205: name = "parameter_205" + original_name = "batch_norm2d_21.w_1" shape = [128] dtype = "float32" min_val = float("-0.642487") @@ -2184,6 +2390,7 @@ class Program_weight_tensor_parameter_205: class Program_weight_tensor_parameter_206: name = "parameter_206" + original_name = "conv2d_21.w_0" shape = [128, 128, 3, 3] dtype = "float32" min_val = float("-0.330961") @@ -2195,6 +2402,7 @@ class Program_weight_tensor_parameter_206: class Program_weight_tensor_parameter_207: name = "parameter_207" + original_name = "batch_norm2d_20.b_0" shape = [128] dtype = "float32" min_val = float("-0.394333") @@ -2206,6 +2414,7 @@ class Program_weight_tensor_parameter_207: class Program_weight_tensor_parameter_208: name = "parameter_208" + original_name = "batch_norm2d_20.w_0" shape = [128] dtype = "float32" min_val = float("0.122299") @@ -2217,6 +2426,7 @@ class Program_weight_tensor_parameter_208: class Program_weight_tensor_parameter_209: name = "parameter_209" + original_name = "batch_norm2d_20.w_2" shape = [128] dtype = "float32" min_val = float("0.0114481") @@ -2228,6 +2438,7 @@ class Program_weight_tensor_parameter_209: class Program_weight_tensor_parameter_210: name = "parameter_210" + original_name = "batch_norm2d_20.w_1" shape = [128] dtype = "float32" min_val = float("-0.417545") @@ -2239,6 +2450,7 @@ class Program_weight_tensor_parameter_210: class Program_weight_tensor_parameter_211: name = "parameter_211" + original_name = "conv2d_20.w_0" shape = [128, 512, 1, 1] dtype = "float32" min_val = float("-0.346156") @@ -2250,6 +2462,7 @@ class Program_weight_tensor_parameter_211: class Program_weight_tensor_parameter_212: name = "parameter_212" + original_name = "batch_norm2d_19.b_0" shape = [512] dtype = "float32" min_val = float("-0.455431") @@ -2261,6 +2474,7 @@ class Program_weight_tensor_parameter_212: class Program_weight_tensor_parameter_213: name = "parameter_213" + original_name = "batch_norm2d_19.w_0" shape = [512] dtype = "float32" min_val = float("-0.0672166") @@ -2272,6 +2486,7 @@ class Program_weight_tensor_parameter_213: class Program_weight_tensor_parameter_214: name = "parameter_214" + original_name = "batch_norm2d_19.w_2" shape = [512] dtype = "float32" min_val = float("3.4359e-36") @@ -2283,6 +2498,7 @@ class Program_weight_tensor_parameter_214: class Program_weight_tensor_parameter_215: name = "parameter_215" + original_name = "batch_norm2d_19.w_1" shape = [512] dtype = "float32" min_val = float("-0.205145") @@ -2294,6 +2510,7 @@ class Program_weight_tensor_parameter_215: class Program_weight_tensor_parameter_216: name = "parameter_216" + original_name = "conv2d_19.w_0" shape = [512, 128, 1, 1] dtype = "float32" min_val = float("-0.411486") @@ -2305,6 +2522,7 @@ class Program_weight_tensor_parameter_216: class Program_weight_tensor_parameter_217: name = "parameter_217" + original_name = "batch_norm2d_18.b_0" shape = [128] dtype = "float32" min_val = float("-0.243617") @@ -2316,6 +2534,7 @@ class Program_weight_tensor_parameter_217: class Program_weight_tensor_parameter_218: name = "parameter_218" + original_name = "batch_norm2d_18.w_0" shape = [128] dtype = "float32" min_val = float("0.142974") @@ -2327,6 +2546,7 @@ class Program_weight_tensor_parameter_218: class Program_weight_tensor_parameter_219: name = "parameter_219" + original_name = "batch_norm2d_18.w_2" shape = [128] dtype = "float32" min_val = float("0.0210301") @@ -2338,6 +2558,7 @@ class Program_weight_tensor_parameter_219: class Program_weight_tensor_parameter_220: name = "parameter_220" + original_name = "batch_norm2d_18.w_1" shape = [128] dtype = "float32" min_val = float("-0.958791") @@ -2349,6 +2570,7 @@ class Program_weight_tensor_parameter_220: class Program_weight_tensor_parameter_221: name = "parameter_221" + original_name = "conv2d_18.w_0" shape = [128, 128, 3, 3] dtype = "float32" min_val = float("-0.356737") @@ -2360,6 +2582,7 @@ class Program_weight_tensor_parameter_221: class Program_weight_tensor_parameter_222: name = "parameter_222" + original_name = "batch_norm2d_17.b_0" shape = [128] dtype = "float32" min_val = float("-0.26234") @@ -2371,6 +2594,7 @@ class Program_weight_tensor_parameter_222: class Program_weight_tensor_parameter_223: name = "parameter_223" + original_name = "batch_norm2d_17.w_0" shape = [128] dtype = "float32" min_val = float("0.135421") @@ -2382,6 +2606,7 @@ class Program_weight_tensor_parameter_223: class Program_weight_tensor_parameter_224: name = "parameter_224" + original_name = "batch_norm2d_17.w_2" shape = [128] dtype = "float32" min_val = float("0.0169231") @@ -2393,6 +2618,7 @@ class Program_weight_tensor_parameter_224: class Program_weight_tensor_parameter_225: name = "parameter_225" + original_name = "batch_norm2d_17.w_1" shape = [128] dtype = "float32" min_val = float("-1.0924") @@ -2404,6 +2630,7 @@ class Program_weight_tensor_parameter_225: class Program_weight_tensor_parameter_226: name = "parameter_226" + original_name = "conv2d_17.w_0" shape = [128, 512, 1, 1] dtype = "float32" min_val = float("-0.562909") @@ -2415,6 +2642,7 @@ class Program_weight_tensor_parameter_226: class Program_weight_tensor_parameter_227: name = "parameter_227" + original_name = "batch_norm2d_16.b_0" shape = [512] dtype = "float32" min_val = float("-0.370036") @@ -2426,6 +2654,7 @@ class Program_weight_tensor_parameter_227: class Program_weight_tensor_parameter_228: name = "parameter_228" + original_name = "batch_norm2d_16.w_0" shape = [512] dtype = "float32" min_val = float("-7.30756e-19") @@ -2437,6 +2666,7 @@ class Program_weight_tensor_parameter_228: class Program_weight_tensor_parameter_229: name = "parameter_229" + original_name = "batch_norm2d_16.w_2" shape = [512] dtype = "float32" min_val = float("5.60519e-45") @@ -2448,6 +2678,7 @@ class Program_weight_tensor_parameter_229: class Program_weight_tensor_parameter_230: name = "parameter_230" + original_name = "batch_norm2d_16.w_1" shape = [512] dtype = "float32" min_val = float("-0.667008") @@ -2459,6 +2690,7 @@ class Program_weight_tensor_parameter_230: class Program_weight_tensor_parameter_231: name = "parameter_231" + original_name = "conv2d_16.w_0" shape = [512, 256, 1, 1] dtype = "float32" min_val = float("-0.723694") @@ -2470,6 +2702,7 @@ class Program_weight_tensor_parameter_231: class Program_weight_tensor_parameter_232: name = "parameter_232" + original_name = "batch_norm2d_15.b_0" shape = [512] dtype = "float32" min_val = float("-0.370036") @@ -2481,6 +2714,7 @@ class Program_weight_tensor_parameter_232: class Program_weight_tensor_parameter_233: name = "parameter_233" + original_name = "batch_norm2d_15.w_0" shape = [512] dtype = "float32" min_val = float("-0.0332384") @@ -2492,6 +2726,7 @@ class Program_weight_tensor_parameter_233: class Program_weight_tensor_parameter_234: name = "parameter_234" + original_name = "batch_norm2d_15.w_2" shape = [512] dtype = "float32" min_val = float("5.60519e-45") @@ -2503,6 +2738,7 @@ class Program_weight_tensor_parameter_234: class Program_weight_tensor_parameter_235: name = "parameter_235" + original_name = "batch_norm2d_15.w_1" shape = [512] dtype = "float32" min_val = float("-0.216692") @@ -2514,6 +2750,7 @@ class Program_weight_tensor_parameter_235: class Program_weight_tensor_parameter_236: name = "parameter_236" + original_name = "conv2d_15.w_0" shape = [512, 128, 1, 1] dtype = "float32" min_val = float("-0.329659") @@ -2525,6 +2762,7 @@ class Program_weight_tensor_parameter_236: class Program_weight_tensor_parameter_237: name = "parameter_237" + original_name = "batch_norm2d_14.b_0" shape = [128] dtype = "float32" min_val = float("-0.382095") @@ -2536,6 +2774,7 @@ class Program_weight_tensor_parameter_237: class Program_weight_tensor_parameter_238: name = "parameter_238" + original_name = "batch_norm2d_14.w_0" shape = [128] dtype = "float32" min_val = float("0.128756") @@ -2547,6 +2786,7 @@ class Program_weight_tensor_parameter_238: class Program_weight_tensor_parameter_239: name = "parameter_239" + original_name = "batch_norm2d_14.w_2" shape = [128] dtype = "float32" min_val = float("0.00985592") @@ -2558,6 +2798,7 @@ class Program_weight_tensor_parameter_239: class Program_weight_tensor_parameter_240: name = "parameter_240" + original_name = "batch_norm2d_14.w_1" shape = [128] dtype = "float32" min_val = float("-0.362586") @@ -2569,6 +2810,7 @@ class Program_weight_tensor_parameter_240: class Program_weight_tensor_parameter_241: name = "parameter_241" + original_name = "conv2d_14.w_0" shape = [128, 128, 3, 3] dtype = "float32" min_val = float("-0.231993") @@ -2580,6 +2822,7 @@ class Program_weight_tensor_parameter_241: class Program_weight_tensor_parameter_242: name = "parameter_242" + original_name = "batch_norm2d_13.b_0" shape = [128] dtype = "float32" min_val = float("-0.47523") @@ -2591,6 +2834,7 @@ class Program_weight_tensor_parameter_242: class Program_weight_tensor_parameter_243: name = "parameter_243" + original_name = "batch_norm2d_13.w_0" shape = [128] dtype = "float32" min_val = float("0.164114") @@ -2602,6 +2846,7 @@ class Program_weight_tensor_parameter_243: class Program_weight_tensor_parameter_244: name = "parameter_244" + original_name = "batch_norm2d_13.w_2" shape = [128] dtype = "float32" min_val = float("0.0232963") @@ -2613,6 +2858,7 @@ class Program_weight_tensor_parameter_244: class Program_weight_tensor_parameter_245: name = "parameter_245" + original_name = "batch_norm2d_13.w_1" shape = [128] dtype = "float32" min_val = float("-0.553745") @@ -2624,6 +2870,7 @@ class Program_weight_tensor_parameter_245: class Program_weight_tensor_parameter_246: name = "parameter_246" + original_name = "conv2d_13.w_0" shape = [128, 256, 1, 1] dtype = "float32" min_val = float("-0.434561") @@ -2635,6 +2882,7 @@ class Program_weight_tensor_parameter_246: class Program_weight_tensor_parameter_247: name = "parameter_247" + original_name = "batch_norm2d_12.b_0" shape = [256] dtype = "float32" min_val = float("-0.304404") @@ -2646,6 +2894,7 @@ class Program_weight_tensor_parameter_247: class Program_weight_tensor_parameter_248: name = "parameter_248" + original_name = "batch_norm2d_12.w_0" shape = [256] dtype = "float32" min_val = float("-0.118238") @@ -2657,6 +2906,7 @@ class Program_weight_tensor_parameter_248: class Program_weight_tensor_parameter_249: name = "parameter_249" + original_name = "batch_norm2d_12.w_2" shape = [256] dtype = "float32" min_val = float("0.000216192") @@ -2668,6 +2918,7 @@ class Program_weight_tensor_parameter_249: class Program_weight_tensor_parameter_250: name = "parameter_250" + original_name = "batch_norm2d_12.w_1" shape = [256] dtype = "float32" min_val = float("-0.258") @@ -2679,6 +2930,7 @@ class Program_weight_tensor_parameter_250: class Program_weight_tensor_parameter_251: name = "parameter_251" + original_name = "conv2d_12.w_0" shape = [256, 64, 1, 1] dtype = "float32" min_val = float("-0.322187") @@ -2690,6 +2942,7 @@ class Program_weight_tensor_parameter_251: class Program_weight_tensor_parameter_252: name = "parameter_252" + original_name = "batch_norm2d_11.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -2699,6 +2952,7 @@ class Program_weight_tensor_parameter_252: class Program_weight_tensor_parameter_253: name = "parameter_253" + original_name = "batch_norm2d_11.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -2708,6 +2962,7 @@ class Program_weight_tensor_parameter_253: class Program_weight_tensor_parameter_254: name = "parameter_254" + original_name = "batch_norm2d_11.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -2717,6 +2972,7 @@ class Program_weight_tensor_parameter_254: class Program_weight_tensor_parameter_255: name = "parameter_255" + original_name = "batch_norm2d_11.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -2726,6 +2982,7 @@ class Program_weight_tensor_parameter_255: class Program_weight_tensor_parameter_256: name = "parameter_256" + original_name = "conv2d_11.w_0" shape = [64, 64, 3, 3] dtype = "float32" min_val = float("-0.393733") @@ -2737,6 +2994,7 @@ class Program_weight_tensor_parameter_256: class Program_weight_tensor_parameter_257: name = "parameter_257" + original_name = "batch_norm2d_10.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -2746,6 +3004,7 @@ class Program_weight_tensor_parameter_257: class Program_weight_tensor_parameter_258: name = "parameter_258" + original_name = "batch_norm2d_10.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -2755,6 +3014,7 @@ class Program_weight_tensor_parameter_258: class Program_weight_tensor_parameter_259: name = "parameter_259" + original_name = "batch_norm2d_10.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -2764,6 +3024,7 @@ class Program_weight_tensor_parameter_259: class Program_weight_tensor_parameter_260: name = "parameter_260" + original_name = "batch_norm2d_10.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -2773,6 +3034,7 @@ class Program_weight_tensor_parameter_260: class Program_weight_tensor_parameter_261: name = "parameter_261" + original_name = "conv2d_10.w_0" shape = [64, 256, 1, 1] dtype = "float32" min_val = float("-0.364472") @@ -2784,6 +3046,7 @@ class Program_weight_tensor_parameter_261: class Program_weight_tensor_parameter_262: name = "parameter_262" + original_name = "batch_norm2d_9.b_0" shape = [256] dtype = "float32" min_val = float("-0.462393") @@ -2795,6 +3058,7 @@ class Program_weight_tensor_parameter_262: class Program_weight_tensor_parameter_263: name = "parameter_263" + original_name = "batch_norm2d_9.w_0" shape = [256] dtype = "float32" min_val = float("-0.0375522") @@ -2806,6 +3070,7 @@ class Program_weight_tensor_parameter_263: class Program_weight_tensor_parameter_264: name = "parameter_264" + original_name = "batch_norm2d_9.w_2" shape = [256] dtype = "float32" min_val = float("0.000317126") @@ -2817,6 +3082,7 @@ class Program_weight_tensor_parameter_264: class Program_weight_tensor_parameter_265: name = "parameter_265" + original_name = "batch_norm2d_9.w_1" shape = [256] dtype = "float32" min_val = float("-0.270805") @@ -2828,6 +3094,7 @@ class Program_weight_tensor_parameter_265: class Program_weight_tensor_parameter_266: name = "parameter_266" + original_name = "conv2d_9.w_0" shape = [256, 64, 1, 1] dtype = "float32" min_val = float("-0.425601") @@ -2839,6 +3106,7 @@ class Program_weight_tensor_parameter_266: class Program_weight_tensor_parameter_267: name = "parameter_267" + original_name = "batch_norm2d_8.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -2848,6 +3116,7 @@ class Program_weight_tensor_parameter_267: class Program_weight_tensor_parameter_268: name = "parameter_268" + original_name = "batch_norm2d_8.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -2857,6 +3126,7 @@ class Program_weight_tensor_parameter_268: class Program_weight_tensor_parameter_269: name = "parameter_269" + original_name = "batch_norm2d_8.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -2866,6 +3136,7 @@ class Program_weight_tensor_parameter_269: class Program_weight_tensor_parameter_270: name = "parameter_270" + original_name = "batch_norm2d_8.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -2875,6 +3146,7 @@ class Program_weight_tensor_parameter_270: class Program_weight_tensor_parameter_271: name = "parameter_271" + original_name = "conv2d_8.w_0" shape = [64, 64, 3, 3] dtype = "float32" min_val = float("-0.365322") @@ -2886,6 +3158,7 @@ class Program_weight_tensor_parameter_271: class Program_weight_tensor_parameter_272: name = "parameter_272" + original_name = "batch_norm2d_7.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -2895,6 +3168,7 @@ class Program_weight_tensor_parameter_272: class Program_weight_tensor_parameter_273: name = "parameter_273" + original_name = "batch_norm2d_7.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -2904,6 +3178,7 @@ class Program_weight_tensor_parameter_273: class Program_weight_tensor_parameter_274: name = "parameter_274" + original_name = "batch_norm2d_7.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -2913,6 +3188,7 @@ class Program_weight_tensor_parameter_274: class Program_weight_tensor_parameter_275: name = "parameter_275" + original_name = "batch_norm2d_7.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -2922,6 +3198,7 @@ class Program_weight_tensor_parameter_275: class Program_weight_tensor_parameter_276: name = "parameter_276" + original_name = "conv2d_7.w_0" shape = [64, 256, 1, 1] dtype = "float32" min_val = float("-0.369789") @@ -2933,6 +3210,7 @@ class Program_weight_tensor_parameter_276: class Program_weight_tensor_parameter_277: name = "parameter_277" + original_name = "batch_norm2d_6.b_0" shape = [256] dtype = "float32" min_val = float("-0.254622") @@ -2944,6 +3222,7 @@ class Program_weight_tensor_parameter_277: class Program_weight_tensor_parameter_278: name = "parameter_278" + original_name = "batch_norm2d_6.w_0" shape = [256] dtype = "float32" min_val = float("-0.0321339") @@ -2955,6 +3234,7 @@ class Program_weight_tensor_parameter_278: class Program_weight_tensor_parameter_279: name = "parameter_279" + original_name = "batch_norm2d_6.w_2" shape = [256] dtype = "float32" min_val = float("5.60519e-45") @@ -2966,6 +3246,7 @@ class Program_weight_tensor_parameter_279: class Program_weight_tensor_parameter_280: name = "parameter_280" + original_name = "batch_norm2d_6.w_1" shape = [256] dtype = "float32" min_val = float("-1.35275") @@ -2977,6 +3258,7 @@ class Program_weight_tensor_parameter_280: class Program_weight_tensor_parameter_281: name = "parameter_281" + original_name = "conv2d_6.w_0" shape = [256, 64, 1, 1] dtype = "float32" min_val = float("-0.983408") @@ -2988,6 +3270,7 @@ class Program_weight_tensor_parameter_281: class Program_weight_tensor_parameter_282: name = "parameter_282" + original_name = "batch_norm2d_5.b_0" shape = [256] dtype = "float32" min_val = float("-0.254622") @@ -2999,6 +3282,7 @@ class Program_weight_tensor_parameter_282: class Program_weight_tensor_parameter_283: name = "parameter_283" + original_name = "batch_norm2d_5.w_0" shape = [256] dtype = "float32" min_val = float("-0.110997") @@ -3010,6 +3294,7 @@ class Program_weight_tensor_parameter_283: class Program_weight_tensor_parameter_284: name = "parameter_284" + original_name = "batch_norm2d_5.w_2" shape = [256] dtype = "float32" min_val = float("5.60519e-45") @@ -3021,6 +3306,7 @@ class Program_weight_tensor_parameter_284: class Program_weight_tensor_parameter_285: name = "parameter_285" + original_name = "batch_norm2d_5.w_1" shape = [256] dtype = "float32" min_val = float("-0.426076") @@ -3032,6 +3318,7 @@ class Program_weight_tensor_parameter_285: class Program_weight_tensor_parameter_286: name = "parameter_286" + original_name = "conv2d_5.w_0" shape = [256, 64, 1, 1] dtype = "float32" min_val = float("-0.493531") @@ -3043,6 +3330,7 @@ class Program_weight_tensor_parameter_286: class Program_weight_tensor_parameter_287: name = "parameter_287" + original_name = "batch_norm2d_4.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -3052,6 +3340,7 @@ class Program_weight_tensor_parameter_287: class Program_weight_tensor_parameter_288: name = "parameter_288" + original_name = "batch_norm2d_4.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -3061,6 +3350,7 @@ class Program_weight_tensor_parameter_288: class Program_weight_tensor_parameter_289: name = "parameter_289" + original_name = "batch_norm2d_4.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -3070,6 +3360,7 @@ class Program_weight_tensor_parameter_289: class Program_weight_tensor_parameter_290: name = "parameter_290" + original_name = "batch_norm2d_4.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -3079,6 +3370,7 @@ class Program_weight_tensor_parameter_290: class Program_weight_tensor_parameter_291: name = "parameter_291" + original_name = "conv2d_4.w_0" shape = [64, 64, 3, 3] dtype = "float32" min_val = float("-0.594026") @@ -3090,6 +3382,7 @@ class Program_weight_tensor_parameter_291: class Program_weight_tensor_parameter_292: name = "parameter_292" + original_name = "batch_norm2d_3.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -3099,6 +3392,7 @@ class Program_weight_tensor_parameter_292: class Program_weight_tensor_parameter_293: name = "parameter_293" + original_name = "batch_norm2d_3.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -3108,6 +3402,7 @@ class Program_weight_tensor_parameter_293: class Program_weight_tensor_parameter_294: name = "parameter_294" + original_name = "batch_norm2d_3.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -3117,6 +3412,7 @@ class Program_weight_tensor_parameter_294: class Program_weight_tensor_parameter_295: name = "parameter_295" + original_name = "batch_norm2d_3.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -3126,6 +3422,7 @@ class Program_weight_tensor_parameter_295: class Program_weight_tensor_parameter_296: name = "parameter_296" + original_name = "conv2d_3.w_0" shape = [64, 64, 1, 1] dtype = "float32" min_val = float("-0.911477") @@ -3137,6 +3434,7 @@ class Program_weight_tensor_parameter_296: class Program_weight_tensor_parameter_297: name = "parameter_297" + original_name = "batch_norm2d_2.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -3146,6 +3444,7 @@ class Program_weight_tensor_parameter_297: class Program_weight_tensor_parameter_298: name = "parameter_298" + original_name = "batch_norm2d_2.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -3155,6 +3454,7 @@ class Program_weight_tensor_parameter_298: class Program_weight_tensor_parameter_299: name = "parameter_299" + original_name = "batch_norm2d_2.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -3164,6 +3464,7 @@ class Program_weight_tensor_parameter_299: class Program_weight_tensor_parameter_300: name = "parameter_300" + original_name = "batch_norm2d_2.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -3173,6 +3474,7 @@ class Program_weight_tensor_parameter_300: class Program_weight_tensor_parameter_301: name = "parameter_301" + original_name = "conv2d_2.w_0" shape = [64, 32, 3, 3] dtype = "float32" min_val = float("-0.825257") @@ -3184,6 +3486,7 @@ class Program_weight_tensor_parameter_301: class Program_weight_tensor_parameter_302: name = "parameter_302" + original_name = "batch_norm2d_1.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -3193,6 +3496,7 @@ class Program_weight_tensor_parameter_302: class Program_weight_tensor_parameter_303: name = "parameter_303" + original_name = "batch_norm2d_1.w_0" shape = [32] dtype = "float32" min_val = float("0") @@ -3202,6 +3506,7 @@ class Program_weight_tensor_parameter_303: class Program_weight_tensor_parameter_304: name = "parameter_304" + original_name = "batch_norm2d_1.w_2" shape = [32] dtype = "float32" min_val = float("0") @@ -3211,6 +3516,7 @@ class Program_weight_tensor_parameter_304: class Program_weight_tensor_parameter_305: name = "parameter_305" + original_name = "batch_norm2d_1.w_1" shape = [32] dtype = "float32" min_val = float("0") @@ -3220,6 +3526,7 @@ class Program_weight_tensor_parameter_305: class Program_weight_tensor_parameter_306: name = "parameter_306" + original_name = "conv2d_1.w_0" shape = [32, 32, 3, 3] dtype = "float32" min_val = float("-1.31725") @@ -3231,6 +3538,7 @@ class Program_weight_tensor_parameter_306: class Program_weight_tensor_parameter_307: name = "parameter_307" + original_name = "batch_norm2d_0.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -3240,6 +3548,7 @@ class Program_weight_tensor_parameter_307: class Program_weight_tensor_parameter_308: name = "parameter_308" + original_name = "batch_norm2d_0.w_0" shape = [32] dtype = "float32" min_val = float("0") @@ -3249,6 +3558,7 @@ class Program_weight_tensor_parameter_308: class Program_weight_tensor_parameter_309: name = "parameter_309" + original_name = "batch_norm2d_0.w_2" shape = [32] dtype = "float32" min_val = float("0") @@ -3258,6 +3568,7 @@ class Program_weight_tensor_parameter_309: class Program_weight_tensor_parameter_310: name = "parameter_310" + original_name = "batch_norm2d_0.w_1" shape = [32] dtype = "float32" min_val = float("0") @@ -3267,6 +3578,7 @@ class Program_weight_tensor_parameter_310: class Program_weight_tensor_parameter_311: name = "parameter_311" + original_name = "conv2d_0.w_0" shape = [32, 3, 3, 3] dtype = "float32" min_val = float("-1.4988") diff --git a/paddle_samples/PaddleX/FasterNet-L/subgraph_2/input_meta.py b/paddle_samples/PaddleX/FasterNet-L/subgraph_2/input_meta.py index 050f8d358..28cdfe82b 100644 --- a/paddle_samples/PaddleX/FasterNet-L/subgraph_2/input_meta.py +++ b/paddle_samples/PaddleX/FasterNet-L/subgraph_2/input_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_data_0: name = "data_0" + original_name = "var_456" shape = [60, 3, 224, 224] dtype = "float32" min_val = float("-2.1179") diff --git a/paddle_samples/PaddleX/FasterNet-L/subgraph_2/weight_meta.py b/paddle_samples/PaddleX/FasterNet-L/subgraph_2/weight_meta.py index bfde1de1d..d6bf91265 100644 --- a/paddle_samples/PaddleX/FasterNet-L/subgraph_2/weight_meta.py +++ b/paddle_samples/PaddleX/FasterNet-L/subgraph_2/weight_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_parameter_0: name = "parameter_0" + original_name = "linear_0.b_0" shape = [102] dtype = "float32" min_val = float("-0.00338778") @@ -11,6 +12,7 @@ class Program_weight_tensor_parameter_0: class Program_weight_tensor_parameter_1: name = "parameter_1" + original_name = "linear_0.w_0" shape = [1280, 102] dtype = "float32" min_val = float("-0.0851378") @@ -22,6 +24,7 @@ class Program_weight_tensor_parameter_1: class Program_weight_tensor_parameter_2: name = "parameter_2" + original_name = "conv2d_88.w_0" shape = [1280, 1536, 1, 1] dtype = "float32" min_val = float("-0.995539") @@ -33,6 +36,7 @@ class Program_weight_tensor_parameter_2: class Program_weight_tensor_parameter_3: name = "parameter_3" + original_name = "conv2d_86.w_0" shape = [1536, 3072, 1, 1] dtype = "float32" min_val = float("-0.792863") @@ -44,6 +48,7 @@ class Program_weight_tensor_parameter_3: class Program_weight_tensor_parameter_4: name = "parameter_4" + original_name = "batch_norm2d_31.b_0" shape = [3072] dtype = "float32" min_val = float("-0.29347") @@ -55,6 +60,7 @@ class Program_weight_tensor_parameter_4: class Program_weight_tensor_parameter_5: name = "parameter_5" + original_name = "batch_norm2d_31.w_0" shape = [3072] dtype = "float32" min_val = float("-0.00274576") @@ -66,6 +72,7 @@ class Program_weight_tensor_parameter_5: class Program_weight_tensor_parameter_6: name = "parameter_6" + original_name = "batch_norm2d_31.w_2" shape = [3072] dtype = "float32" min_val = float("2.01721e-08") @@ -77,6 +84,7 @@ class Program_weight_tensor_parameter_6: class Program_weight_tensor_parameter_7: name = "parameter_7" + original_name = "batch_norm2d_31.w_1" shape = [3072] dtype = "float32" min_val = float("-2.18936") @@ -88,6 +96,7 @@ class Program_weight_tensor_parameter_7: class Program_weight_tensor_parameter_8: name = "parameter_8" + original_name = "conv2d_85.w_0" shape = [3072, 1536, 1, 1] dtype = "float32" min_val = float("-0.535563") @@ -99,6 +108,7 @@ class Program_weight_tensor_parameter_8: class Program_weight_tensor_parameter_9: name = "parameter_9" + original_name = "conv2d_87.w_0" shape = [384, 384, 3, 3] dtype = "float32" min_val = float("-0.511347") @@ -110,6 +120,7 @@ class Program_weight_tensor_parameter_9: class Program_weight_tensor_parameter_10: name = "parameter_10" + original_name = "conv2d_83.w_0" shape = [1536, 3072, 1, 1] dtype = "float32" min_val = float("-0.776099") @@ -121,6 +132,7 @@ class Program_weight_tensor_parameter_10: class Program_weight_tensor_parameter_11: name = "parameter_11" + original_name = "batch_norm2d_30.b_0" shape = [3072] dtype = "float32" min_val = float("-0.270749") @@ -132,6 +144,7 @@ class Program_weight_tensor_parameter_11: class Program_weight_tensor_parameter_12: name = "parameter_12" + original_name = "batch_norm2d_30.w_0" shape = [3072] dtype = "float32" min_val = float("-0.00332062") @@ -143,6 +156,7 @@ class Program_weight_tensor_parameter_12: class Program_weight_tensor_parameter_13: name = "parameter_13" + original_name = "batch_norm2d_30.w_2" shape = [3072] dtype = "float32" min_val = float("3.43233e-08") @@ -154,6 +168,7 @@ class Program_weight_tensor_parameter_13: class Program_weight_tensor_parameter_14: name = "parameter_14" + original_name = "batch_norm2d_30.w_1" shape = [3072] dtype = "float32" min_val = float("-1.91314") @@ -165,6 +180,7 @@ class Program_weight_tensor_parameter_14: class Program_weight_tensor_parameter_15: name = "parameter_15" + original_name = "conv2d_82.w_0" shape = [3072, 1536, 1, 1] dtype = "float32" min_val = float("-0.45466") @@ -176,6 +192,7 @@ class Program_weight_tensor_parameter_15: class Program_weight_tensor_parameter_16: name = "parameter_16" + original_name = "conv2d_84.w_0" shape = [384, 384, 3, 3] dtype = "float32" min_val = float("-0.389557") @@ -187,6 +204,7 @@ class Program_weight_tensor_parameter_16: class Program_weight_tensor_parameter_17: name = "parameter_17" + original_name = "conv2d_80.w_0" shape = [1536, 3072, 1, 1] dtype = "float32" min_val = float("-0.641912") @@ -198,6 +216,7 @@ class Program_weight_tensor_parameter_17: class Program_weight_tensor_parameter_18: name = "parameter_18" + original_name = "batch_norm2d_29.b_0" shape = [3072] dtype = "float32" min_val = float("-0.26071") @@ -209,6 +228,7 @@ class Program_weight_tensor_parameter_18: class Program_weight_tensor_parameter_19: name = "parameter_19" + original_name = "batch_norm2d_29.w_0" shape = [3072] dtype = "float32" min_val = float("-0.00202214") @@ -220,6 +240,7 @@ class Program_weight_tensor_parameter_19: class Program_weight_tensor_parameter_20: name = "parameter_20" + original_name = "batch_norm2d_29.w_2" shape = [3072] dtype = "float32" min_val = float("1.93677e-08") @@ -231,6 +252,7 @@ class Program_weight_tensor_parameter_20: class Program_weight_tensor_parameter_21: name = "parameter_21" + original_name = "batch_norm2d_29.w_1" shape = [3072] dtype = "float32" min_val = float("-2.48758") @@ -242,6 +264,7 @@ class Program_weight_tensor_parameter_21: class Program_weight_tensor_parameter_22: name = "parameter_22" + original_name = "conv2d_79.w_0" shape = [3072, 1536, 1, 1] dtype = "float32" min_val = float("-0.552919") @@ -253,6 +276,7 @@ class Program_weight_tensor_parameter_22: class Program_weight_tensor_parameter_23: name = "parameter_23" + original_name = "conv2d_81.w_0" shape = [384, 384, 3, 3] dtype = "float32" min_val = float("-0.343608") @@ -264,6 +288,7 @@ class Program_weight_tensor_parameter_23: class Program_weight_tensor_parameter_24: name = "parameter_24" + original_name = "batch_norm2d_28.b_0" shape = [1536] dtype = "float32" min_val = float("-0.355309") @@ -275,6 +300,7 @@ class Program_weight_tensor_parameter_24: class Program_weight_tensor_parameter_25: name = "parameter_25" + original_name = "batch_norm2d_28.w_0" shape = [1536] dtype = "float32" min_val = float("-0.0251692") @@ -286,6 +312,7 @@ class Program_weight_tensor_parameter_25: class Program_weight_tensor_parameter_26: name = "parameter_26" + original_name = "batch_norm2d_28.w_2" shape = [1536] dtype = "float32" min_val = float("0.127456") @@ -297,6 +324,7 @@ class Program_weight_tensor_parameter_26: class Program_weight_tensor_parameter_27: name = "parameter_27" + original_name = "batch_norm2d_28.w_1" shape = [1536] dtype = "float32" min_val = float("-1.58491") @@ -308,6 +336,7 @@ class Program_weight_tensor_parameter_27: class Program_weight_tensor_parameter_28: name = "parameter_28" + original_name = "conv2d_78.w_0" shape = [1536, 768, 2, 2] dtype = "float32" min_val = float("-0.712223") @@ -319,6 +348,7 @@ class Program_weight_tensor_parameter_28: class Program_weight_tensor_parameter_29: name = "parameter_29" + original_name = "conv2d_76.w_0" shape = [768, 1536, 1, 1] dtype = "float32" min_val = float("-0.396799") @@ -330,6 +360,7 @@ class Program_weight_tensor_parameter_29: class Program_weight_tensor_parameter_30: name = "parameter_30" + original_name = "batch_norm2d_27.b_0" shape = [1536] dtype = "float32" min_val = float("-0.217757") @@ -341,6 +372,7 @@ class Program_weight_tensor_parameter_30: class Program_weight_tensor_parameter_31: name = "parameter_31" + original_name = "batch_norm2d_27.w_0" shape = [1536] dtype = "float32" min_val = float("-7.04814e-05") @@ -352,6 +384,7 @@ class Program_weight_tensor_parameter_31: class Program_weight_tensor_parameter_32: name = "parameter_32" + original_name = "batch_norm2d_27.w_2" shape = [1536] dtype = "float32" min_val = float("1.40995e-08") @@ -363,6 +396,7 @@ class Program_weight_tensor_parameter_32: class Program_weight_tensor_parameter_33: name = "parameter_33" + original_name = "batch_norm2d_27.w_1" shape = [1536] dtype = "float32" min_val = float("-1.58361") @@ -374,6 +408,7 @@ class Program_weight_tensor_parameter_33: class Program_weight_tensor_parameter_34: name = "parameter_34" + original_name = "conv2d_75.w_0" shape = [1536, 768, 1, 1] dtype = "float32" min_val = float("-0.495763") @@ -385,6 +420,7 @@ class Program_weight_tensor_parameter_34: class Program_weight_tensor_parameter_35: name = "parameter_35" + original_name = "conv2d_77.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.413055") @@ -396,6 +432,7 @@ class Program_weight_tensor_parameter_35: class Program_weight_tensor_parameter_36: name = "parameter_36" + original_name = "conv2d_73.w_0" shape = [768, 1536, 1, 1] dtype = "float32" min_val = float("-0.43895") @@ -407,6 +444,7 @@ class Program_weight_tensor_parameter_36: class Program_weight_tensor_parameter_37: name = "parameter_37" + original_name = "batch_norm2d_26.b_0" shape = [1536] dtype = "float32" min_val = float("-0.318451") @@ -418,6 +456,7 @@ class Program_weight_tensor_parameter_37: class Program_weight_tensor_parameter_38: name = "parameter_38" + original_name = "batch_norm2d_26.w_0" shape = [1536] dtype = "float32" min_val = float("-8.25041e-05") @@ -429,6 +468,7 @@ class Program_weight_tensor_parameter_38: class Program_weight_tensor_parameter_39: name = "parameter_39" + original_name = "batch_norm2d_26.w_2" shape = [1536] dtype = "float32" min_val = float("4.28966e-09") @@ -440,6 +480,7 @@ class Program_weight_tensor_parameter_39: class Program_weight_tensor_parameter_40: name = "parameter_40" + original_name = "batch_norm2d_26.w_1" shape = [1536] dtype = "float32" min_val = float("-1.42516") @@ -451,6 +492,7 @@ class Program_weight_tensor_parameter_40: class Program_weight_tensor_parameter_41: name = "parameter_41" + original_name = "conv2d_72.w_0" shape = [1536, 768, 1, 1] dtype = "float32" min_val = float("-0.441145") @@ -462,6 +504,7 @@ class Program_weight_tensor_parameter_41: class Program_weight_tensor_parameter_42: name = "parameter_42" + original_name = "conv2d_74.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.30324") @@ -473,6 +516,7 @@ class Program_weight_tensor_parameter_42: class Program_weight_tensor_parameter_43: name = "parameter_43" + original_name = "conv2d_70.w_0" shape = [768, 1536, 1, 1] dtype = "float32" min_val = float("-0.42877") @@ -484,6 +528,7 @@ class Program_weight_tensor_parameter_43: class Program_weight_tensor_parameter_44: name = "parameter_44" + original_name = "batch_norm2d_25.b_0" shape = [1536] dtype = "float32" min_val = float("-0.364668") @@ -495,6 +540,7 @@ class Program_weight_tensor_parameter_44: class Program_weight_tensor_parameter_45: name = "parameter_45" + original_name = "batch_norm2d_25.w_0" shape = [1536] dtype = "float32" min_val = float("-7.73636e-05") @@ -506,6 +552,7 @@ class Program_weight_tensor_parameter_45: class Program_weight_tensor_parameter_46: name = "parameter_46" + original_name = "batch_norm2d_25.w_2" shape = [1536] dtype = "float32" min_val = float("2.2796e-08") @@ -517,6 +564,7 @@ class Program_weight_tensor_parameter_46: class Program_weight_tensor_parameter_47: name = "parameter_47" + original_name = "batch_norm2d_25.w_1" shape = [1536] dtype = "float32" min_val = float("-1.35992") @@ -528,6 +576,7 @@ class Program_weight_tensor_parameter_47: class Program_weight_tensor_parameter_48: name = "parameter_48" + original_name = "conv2d_69.w_0" shape = [1536, 768, 1, 1] dtype = "float32" min_val = float("-0.383911") @@ -539,6 +588,7 @@ class Program_weight_tensor_parameter_48: class Program_weight_tensor_parameter_49: name = "parameter_49" + original_name = "conv2d_71.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.322324") @@ -550,6 +600,7 @@ class Program_weight_tensor_parameter_49: class Program_weight_tensor_parameter_50: name = "parameter_50" + original_name = "conv2d_67.w_0" shape = [768, 1536, 1, 1] dtype = "float32" min_val = float("-0.524073") @@ -561,6 +612,7 @@ class Program_weight_tensor_parameter_50: class Program_weight_tensor_parameter_51: name = "parameter_51" + original_name = "batch_norm2d_24.b_0" shape = [1536] dtype = "float32" min_val = float("-0.329001") @@ -572,6 +624,7 @@ class Program_weight_tensor_parameter_51: class Program_weight_tensor_parameter_52: name = "parameter_52" + original_name = "batch_norm2d_24.w_0" shape = [1536] dtype = "float32" min_val = float("-6.76644e-05") @@ -583,6 +636,7 @@ class Program_weight_tensor_parameter_52: class Program_weight_tensor_parameter_53: name = "parameter_53" + original_name = "batch_norm2d_24.w_2" shape = [1536] dtype = "float32" min_val = float("5.29791e-09") @@ -594,6 +648,7 @@ class Program_weight_tensor_parameter_53: class Program_weight_tensor_parameter_54: name = "parameter_54" + original_name = "batch_norm2d_24.w_1" shape = [1536] dtype = "float32" min_val = float("-2.40514") @@ -605,6 +660,7 @@ class Program_weight_tensor_parameter_54: class Program_weight_tensor_parameter_55: name = "parameter_55" + original_name = "conv2d_66.w_0" shape = [1536, 768, 1, 1] dtype = "float32" min_val = float("-0.481457") @@ -616,6 +672,7 @@ class Program_weight_tensor_parameter_55: class Program_weight_tensor_parameter_56: name = "parameter_56" + original_name = "conv2d_68.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.355065") @@ -627,6 +684,7 @@ class Program_weight_tensor_parameter_56: class Program_weight_tensor_parameter_57: name = "parameter_57" + original_name = "conv2d_64.w_0" shape = [768, 1536, 1, 1] dtype = "float32" min_val = float("-0.528305") @@ -638,6 +696,7 @@ class Program_weight_tensor_parameter_57: class Program_weight_tensor_parameter_58: name = "parameter_58" + original_name = "batch_norm2d_23.b_0" shape = [1536] dtype = "float32" min_val = float("-0.295005") @@ -649,6 +708,7 @@ class Program_weight_tensor_parameter_58: class Program_weight_tensor_parameter_59: name = "parameter_59" + original_name = "batch_norm2d_23.w_0" shape = [1536] dtype = "float32" min_val = float("-2.19265e-05") @@ -660,6 +720,7 @@ class Program_weight_tensor_parameter_59: class Program_weight_tensor_parameter_60: name = "parameter_60" + original_name = "batch_norm2d_23.w_2" shape = [1536] dtype = "float32" min_val = float("1.82631e-08") @@ -671,6 +732,7 @@ class Program_weight_tensor_parameter_60: class Program_weight_tensor_parameter_61: name = "parameter_61" + original_name = "batch_norm2d_23.w_1" shape = [1536] dtype = "float32" min_val = float("-1.51048") @@ -682,6 +744,7 @@ class Program_weight_tensor_parameter_61: class Program_weight_tensor_parameter_62: name = "parameter_62" + original_name = "conv2d_63.w_0" shape = [1536, 768, 1, 1] dtype = "float32" min_val = float("-0.45972") @@ -693,6 +756,7 @@ class Program_weight_tensor_parameter_62: class Program_weight_tensor_parameter_63: name = "parameter_63" + original_name = "conv2d_65.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.365823") @@ -704,6 +768,7 @@ class Program_weight_tensor_parameter_63: class Program_weight_tensor_parameter_64: name = "parameter_64" + original_name = "conv2d_61.w_0" shape = [768, 1536, 1, 1] dtype = "float32" min_val = float("-0.371079") @@ -715,6 +780,7 @@ class Program_weight_tensor_parameter_64: class Program_weight_tensor_parameter_65: name = "parameter_65" + original_name = "batch_norm2d_22.b_0" shape = [1536] dtype = "float32" min_val = float("-0.288538") @@ -726,6 +792,7 @@ class Program_weight_tensor_parameter_65: class Program_weight_tensor_parameter_66: name = "parameter_66" + original_name = "batch_norm2d_22.w_0" shape = [1536] dtype = "float32" min_val = float("-4.37362e-06") @@ -737,6 +804,7 @@ class Program_weight_tensor_parameter_66: class Program_weight_tensor_parameter_67: name = "parameter_67" + original_name = "batch_norm2d_22.w_2" shape = [1536] dtype = "float32" min_val = float("1.60948e-09") @@ -748,6 +816,7 @@ class Program_weight_tensor_parameter_67: class Program_weight_tensor_parameter_68: name = "parameter_68" + original_name = "batch_norm2d_22.w_1" shape = [1536] dtype = "float32" min_val = float("-1.3801") @@ -759,6 +828,7 @@ class Program_weight_tensor_parameter_68: class Program_weight_tensor_parameter_69: name = "parameter_69" + original_name = "conv2d_60.w_0" shape = [1536, 768, 1, 1] dtype = "float32" min_val = float("-0.469812") @@ -770,6 +840,7 @@ class Program_weight_tensor_parameter_69: class Program_weight_tensor_parameter_70: name = "parameter_70" + original_name = "conv2d_62.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.264255") @@ -781,6 +852,7 @@ class Program_weight_tensor_parameter_70: class Program_weight_tensor_parameter_71: name = "parameter_71" + original_name = "conv2d_58.w_0" shape = [768, 1536, 1, 1] dtype = "float32" min_val = float("-0.362958") @@ -792,6 +864,7 @@ class Program_weight_tensor_parameter_71: class Program_weight_tensor_parameter_72: name = "parameter_72" + original_name = "batch_norm2d_21.b_0" shape = [1536] dtype = "float32" min_val = float("-0.198076") @@ -803,6 +876,7 @@ class Program_weight_tensor_parameter_72: class Program_weight_tensor_parameter_73: name = "parameter_73" + original_name = "batch_norm2d_21.w_0" shape = [1536] dtype = "float32" min_val = float("0.0307269") @@ -814,6 +888,7 @@ class Program_weight_tensor_parameter_73: class Program_weight_tensor_parameter_74: name = "parameter_74" + original_name = "batch_norm2d_21.w_2" shape = [1536] dtype = "float32" min_val = float("0.0443646") @@ -825,6 +900,7 @@ class Program_weight_tensor_parameter_74: class Program_weight_tensor_parameter_75: name = "parameter_75" + original_name = "batch_norm2d_21.w_1" shape = [1536] dtype = "float32" min_val = float("-1.48634") @@ -836,6 +912,7 @@ class Program_weight_tensor_parameter_75: class Program_weight_tensor_parameter_76: name = "parameter_76" + original_name = "conv2d_57.w_0" shape = [1536, 768, 1, 1] dtype = "float32" min_val = float("-0.524887") @@ -847,6 +924,7 @@ class Program_weight_tensor_parameter_76: class Program_weight_tensor_parameter_77: name = "parameter_77" + original_name = "conv2d_59.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.260939") @@ -858,6 +936,7 @@ class Program_weight_tensor_parameter_77: class Program_weight_tensor_parameter_78: name = "parameter_78" + original_name = "conv2d_55.w_0" shape = [768, 1536, 1, 1] dtype = "float32" min_val = float("-0.419473") @@ -869,6 +948,7 @@ class Program_weight_tensor_parameter_78: class Program_weight_tensor_parameter_79: name = "parameter_79" + original_name = "batch_norm2d_20.b_0" shape = [1536] dtype = "float32" min_val = float("-0.185466") @@ -880,6 +960,7 @@ class Program_weight_tensor_parameter_79: class Program_weight_tensor_parameter_80: name = "parameter_80" + original_name = "batch_norm2d_20.w_0" shape = [1536] dtype = "float32" min_val = float("-1.10052e-05") @@ -891,6 +972,7 @@ class Program_weight_tensor_parameter_80: class Program_weight_tensor_parameter_81: name = "parameter_81" + original_name = "batch_norm2d_20.w_2" shape = [1536] dtype = "float32" min_val = float("3.89352e-09") @@ -902,6 +984,7 @@ class Program_weight_tensor_parameter_81: class Program_weight_tensor_parameter_82: name = "parameter_82" + original_name = "batch_norm2d_20.w_1" shape = [1536] dtype = "float32" min_val = float("-1.33201") @@ -913,6 +996,7 @@ class Program_weight_tensor_parameter_82: class Program_weight_tensor_parameter_83: name = "parameter_83" + original_name = "conv2d_54.w_0" shape = [1536, 768, 1, 1] dtype = "float32" min_val = float("-0.588254") @@ -924,6 +1008,7 @@ class Program_weight_tensor_parameter_83: class Program_weight_tensor_parameter_84: name = "parameter_84" + original_name = "conv2d_56.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.274569") @@ -935,6 +1020,7 @@ class Program_weight_tensor_parameter_84: class Program_weight_tensor_parameter_85: name = "parameter_85" + original_name = "conv2d_52.w_0" shape = [768, 1536, 1, 1] dtype = "float32" min_val = float("-0.495482") @@ -946,6 +1032,7 @@ class Program_weight_tensor_parameter_85: class Program_weight_tensor_parameter_86: name = "parameter_86" + original_name = "batch_norm2d_19.b_0" shape = [1536] dtype = "float32" min_val = float("-0.178221") @@ -957,6 +1044,7 @@ class Program_weight_tensor_parameter_86: class Program_weight_tensor_parameter_87: name = "parameter_87" + original_name = "batch_norm2d_19.w_0" shape = [1536] dtype = "float32" min_val = float("6.13292e-06") @@ -968,6 +1056,7 @@ class Program_weight_tensor_parameter_87: class Program_weight_tensor_parameter_88: name = "parameter_88" + original_name = "batch_norm2d_19.w_2" shape = [1536] dtype = "float32" min_val = float("5.81444e-09") @@ -979,6 +1068,7 @@ class Program_weight_tensor_parameter_88: class Program_weight_tensor_parameter_89: name = "parameter_89" + original_name = "batch_norm2d_19.w_1" shape = [1536] dtype = "float32" min_val = float("-1.29155") @@ -990,6 +1080,7 @@ class Program_weight_tensor_parameter_89: class Program_weight_tensor_parameter_90: name = "parameter_90" + original_name = "conv2d_51.w_0" shape = [1536, 768, 1, 1] dtype = "float32" min_val = float("-0.472933") @@ -1001,6 +1092,7 @@ class Program_weight_tensor_parameter_90: class Program_weight_tensor_parameter_91: name = "parameter_91" + original_name = "conv2d_53.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.310057") @@ -1012,6 +1104,7 @@ class Program_weight_tensor_parameter_91: class Program_weight_tensor_parameter_92: name = "parameter_92" + original_name = "conv2d_49.w_0" shape = [768, 1536, 1, 1] dtype = "float32" min_val = float("-0.421435") @@ -1023,6 +1116,7 @@ class Program_weight_tensor_parameter_92: class Program_weight_tensor_parameter_93: name = "parameter_93" + original_name = "batch_norm2d_18.b_0" shape = [1536] dtype = "float32" min_val = float("-0.196902") @@ -1034,6 +1128,7 @@ class Program_weight_tensor_parameter_93: class Program_weight_tensor_parameter_94: name = "parameter_94" + original_name = "batch_norm2d_18.w_0" shape = [1536] dtype = "float32" min_val = float("-3.53839e-05") @@ -1045,6 +1140,7 @@ class Program_weight_tensor_parameter_94: class Program_weight_tensor_parameter_95: name = "parameter_95" + original_name = "batch_norm2d_18.w_2" shape = [1536] dtype = "float32" min_val = float("3.86892e-09") @@ -1056,6 +1152,7 @@ class Program_weight_tensor_parameter_95: class Program_weight_tensor_parameter_96: name = "parameter_96" + original_name = "batch_norm2d_18.w_1" shape = [1536] dtype = "float32" min_val = float("-1.19611") @@ -1067,6 +1164,7 @@ class Program_weight_tensor_parameter_96: class Program_weight_tensor_parameter_97: name = "parameter_97" + original_name = "conv2d_48.w_0" shape = [1536, 768, 1, 1] dtype = "float32" min_val = float("-0.41966") @@ -1078,6 +1176,7 @@ class Program_weight_tensor_parameter_97: class Program_weight_tensor_parameter_98: name = "parameter_98" + original_name = "conv2d_50.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.32187") @@ -1089,6 +1188,7 @@ class Program_weight_tensor_parameter_98: class Program_weight_tensor_parameter_99: name = "parameter_99" + original_name = "conv2d_46.w_0" shape = [768, 1536, 1, 1] dtype = "float32" min_val = float("-0.556169") @@ -1100,6 +1200,7 @@ class Program_weight_tensor_parameter_99: class Program_weight_tensor_parameter_100: name = "parameter_100" + original_name = "batch_norm2d_17.b_0" shape = [1536] dtype = "float32" min_val = float("-0.175585") @@ -1111,6 +1212,7 @@ class Program_weight_tensor_parameter_100: class Program_weight_tensor_parameter_101: name = "parameter_101" + original_name = "batch_norm2d_17.w_0" shape = [1536] dtype = "float32" min_val = float("-0.0001845") @@ -1122,6 +1224,7 @@ class Program_weight_tensor_parameter_101: class Program_weight_tensor_parameter_102: name = "parameter_102" + original_name = "batch_norm2d_17.w_2" shape = [1536] dtype = "float32" min_val = float("1.99519e-08") @@ -1133,6 +1236,7 @@ class Program_weight_tensor_parameter_102: class Program_weight_tensor_parameter_103: name = "parameter_103" + original_name = "batch_norm2d_17.w_1" shape = [1536] dtype = "float32" min_val = float("-1.10836") @@ -1144,6 +1248,7 @@ class Program_weight_tensor_parameter_103: class Program_weight_tensor_parameter_104: name = "parameter_104" + original_name = "conv2d_45.w_0" shape = [1536, 768, 1, 1] dtype = "float32" min_val = float("-0.58053") @@ -1155,6 +1260,7 @@ class Program_weight_tensor_parameter_104: class Program_weight_tensor_parameter_105: name = "parameter_105" + original_name = "conv2d_47.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.415979") @@ -1166,6 +1272,7 @@ class Program_weight_tensor_parameter_105: class Program_weight_tensor_parameter_106: name = "parameter_106" + original_name = "conv2d_43.w_0" shape = [768, 1536, 1, 1] dtype = "float32" min_val = float("-0.569745") @@ -1177,6 +1284,7 @@ class Program_weight_tensor_parameter_106: class Program_weight_tensor_parameter_107: name = "parameter_107" + original_name = "batch_norm2d_16.b_0" shape = [1536] dtype = "float32" min_val = float("-0.214829") @@ -1188,6 +1296,7 @@ class Program_weight_tensor_parameter_107: class Program_weight_tensor_parameter_108: name = "parameter_108" + original_name = "batch_norm2d_16.w_0" shape = [1536] dtype = "float32" min_val = float("-0.000169927") @@ -1199,6 +1308,7 @@ class Program_weight_tensor_parameter_108: class Program_weight_tensor_parameter_109: name = "parameter_109" + original_name = "batch_norm2d_16.w_2" shape = [1536] dtype = "float32" min_val = float("1.92936e-08") @@ -1210,6 +1320,7 @@ class Program_weight_tensor_parameter_109: class Program_weight_tensor_parameter_110: name = "parameter_110" + original_name = "batch_norm2d_16.w_1" shape = [1536] dtype = "float32" min_val = float("-1.122") @@ -1221,6 +1332,7 @@ class Program_weight_tensor_parameter_110: class Program_weight_tensor_parameter_111: name = "parameter_111" + original_name = "conv2d_42.w_0" shape = [1536, 768, 1, 1] dtype = "float32" min_val = float("-0.464481") @@ -1232,6 +1344,7 @@ class Program_weight_tensor_parameter_111: class Program_weight_tensor_parameter_112: name = "parameter_112" + original_name = "conv2d_44.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.272569") @@ -1243,6 +1356,7 @@ class Program_weight_tensor_parameter_112: class Program_weight_tensor_parameter_113: name = "parameter_113" + original_name = "conv2d_40.w_0" shape = [768, 1536, 1, 1] dtype = "float32" min_val = float("-0.469476") @@ -1254,6 +1368,7 @@ class Program_weight_tensor_parameter_113: class Program_weight_tensor_parameter_114: name = "parameter_114" + original_name = "batch_norm2d_15.b_0" shape = [1536] dtype = "float32" min_val = float("-0.173285") @@ -1265,6 +1380,7 @@ class Program_weight_tensor_parameter_114: class Program_weight_tensor_parameter_115: name = "parameter_115" + original_name = "batch_norm2d_15.w_0" shape = [1536] dtype = "float32" min_val = float("-0.00171177") @@ -1276,6 +1392,7 @@ class Program_weight_tensor_parameter_115: class Program_weight_tensor_parameter_116: name = "parameter_116" + original_name = "batch_norm2d_15.w_2" shape = [1536] dtype = "float32" min_val = float("1.55073e-08") @@ -1287,6 +1404,7 @@ class Program_weight_tensor_parameter_116: class Program_weight_tensor_parameter_117: name = "parameter_117" + original_name = "batch_norm2d_15.w_1" shape = [1536] dtype = "float32" min_val = float("-1.08779") @@ -1298,6 +1416,7 @@ class Program_weight_tensor_parameter_117: class Program_weight_tensor_parameter_118: name = "parameter_118" + original_name = "conv2d_39.w_0" shape = [1536, 768, 1, 1] dtype = "float32" min_val = float("-0.621944") @@ -1309,6 +1428,7 @@ class Program_weight_tensor_parameter_118: class Program_weight_tensor_parameter_119: name = "parameter_119" + original_name = "conv2d_41.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.357478") @@ -1320,6 +1440,7 @@ class Program_weight_tensor_parameter_119: class Program_weight_tensor_parameter_120: name = "parameter_120" + original_name = "conv2d_37.w_0" shape = [768, 1536, 1, 1] dtype = "float32" min_val = float("-0.547576") @@ -1331,6 +1452,7 @@ class Program_weight_tensor_parameter_120: class Program_weight_tensor_parameter_121: name = "parameter_121" + original_name = "batch_norm2d_14.b_0" shape = [1536] dtype = "float32" min_val = float("-0.160204") @@ -1342,6 +1464,7 @@ class Program_weight_tensor_parameter_121: class Program_weight_tensor_parameter_122: name = "parameter_122" + original_name = "batch_norm2d_14.w_0" shape = [1536] dtype = "float32" min_val = float("-0.000760759") @@ -1353,6 +1476,7 @@ class Program_weight_tensor_parameter_122: class Program_weight_tensor_parameter_123: name = "parameter_123" + original_name = "batch_norm2d_14.w_2" shape = [1536] dtype = "float32" min_val = float("3.42846e-08") @@ -1364,6 +1488,7 @@ class Program_weight_tensor_parameter_123: class Program_weight_tensor_parameter_124: name = "parameter_124" + original_name = "batch_norm2d_14.w_1" shape = [1536] dtype = "float32" min_val = float("-0.859252") @@ -1375,6 +1500,7 @@ class Program_weight_tensor_parameter_124: class Program_weight_tensor_parameter_125: name = "parameter_125" + original_name = "conv2d_36.w_0" shape = [1536, 768, 1, 1] dtype = "float32" min_val = float("-0.582799") @@ -1386,6 +1512,7 @@ class Program_weight_tensor_parameter_125: class Program_weight_tensor_parameter_126: name = "parameter_126" + original_name = "conv2d_38.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.312486") @@ -1397,6 +1524,7 @@ class Program_weight_tensor_parameter_126: class Program_weight_tensor_parameter_127: name = "parameter_127" + original_name = "conv2d_34.w_0" shape = [768, 1536, 1, 1] dtype = "float32" min_val = float("-0.421018") @@ -1408,6 +1536,7 @@ class Program_weight_tensor_parameter_127: class Program_weight_tensor_parameter_128: name = "parameter_128" + original_name = "batch_norm2d_13.b_0" shape = [1536] dtype = "float32" min_val = float("-0.156978") @@ -1419,6 +1548,7 @@ class Program_weight_tensor_parameter_128: class Program_weight_tensor_parameter_129: name = "parameter_129" + original_name = "batch_norm2d_13.w_0" shape = [1536] dtype = "float32" min_val = float("-0.0008913") @@ -1430,6 +1560,7 @@ class Program_weight_tensor_parameter_129: class Program_weight_tensor_parameter_130: name = "parameter_130" + original_name = "batch_norm2d_13.w_2" shape = [1536] dtype = "float32" min_val = float("6.35773e-09") @@ -1441,6 +1572,7 @@ class Program_weight_tensor_parameter_130: class Program_weight_tensor_parameter_131: name = "parameter_131" + original_name = "batch_norm2d_13.w_1" shape = [1536] dtype = "float32" min_val = float("-1.1519") @@ -1452,6 +1584,7 @@ class Program_weight_tensor_parameter_131: class Program_weight_tensor_parameter_132: name = "parameter_132" + original_name = "conv2d_33.w_0" shape = [1536, 768, 1, 1] dtype = "float32" min_val = float("-0.630783") @@ -1463,6 +1596,7 @@ class Program_weight_tensor_parameter_132: class Program_weight_tensor_parameter_133: name = "parameter_133" + original_name = "conv2d_35.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.319123") @@ -1474,6 +1608,7 @@ class Program_weight_tensor_parameter_133: class Program_weight_tensor_parameter_134: name = "parameter_134" + original_name = "conv2d_31.w_0" shape = [768, 1536, 1, 1] dtype = "float32" min_val = float("-0.542877") @@ -1485,6 +1620,7 @@ class Program_weight_tensor_parameter_134: class Program_weight_tensor_parameter_135: name = "parameter_135" + original_name = "batch_norm2d_12.b_0" shape = [1536] dtype = "float32" min_val = float("-0.16228") @@ -1496,6 +1632,7 @@ class Program_weight_tensor_parameter_135: class Program_weight_tensor_parameter_136: name = "parameter_136" + original_name = "batch_norm2d_12.w_0" shape = [1536] dtype = "float32" min_val = float("-0.000514413") @@ -1507,6 +1644,7 @@ class Program_weight_tensor_parameter_136: class Program_weight_tensor_parameter_137: name = "parameter_137" + original_name = "batch_norm2d_12.w_2" shape = [1536] dtype = "float32" min_val = float("1.2314e-08") @@ -1518,6 +1656,7 @@ class Program_weight_tensor_parameter_137: class Program_weight_tensor_parameter_138: name = "parameter_138" + original_name = "batch_norm2d_12.w_1" shape = [1536] dtype = "float32" min_val = float("-0.777999") @@ -1529,6 +1668,7 @@ class Program_weight_tensor_parameter_138: class Program_weight_tensor_parameter_139: name = "parameter_139" + original_name = "conv2d_30.w_0" shape = [1536, 768, 1, 1] dtype = "float32" min_val = float("-0.459032") @@ -1540,6 +1680,7 @@ class Program_weight_tensor_parameter_139: class Program_weight_tensor_parameter_140: name = "parameter_140" + original_name = "conv2d_32.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.407741") @@ -1551,6 +1692,7 @@ class Program_weight_tensor_parameter_140: class Program_weight_tensor_parameter_141: name = "parameter_141" + original_name = "conv2d_28.w_0" shape = [768, 1536, 1, 1] dtype = "float32" min_val = float("-0.497684") @@ -1562,6 +1704,7 @@ class Program_weight_tensor_parameter_141: class Program_weight_tensor_parameter_142: name = "parameter_142" + original_name = "batch_norm2d_11.b_0" shape = [1536] dtype = "float32" min_val = float("-0.150792") @@ -1573,6 +1716,7 @@ class Program_weight_tensor_parameter_142: class Program_weight_tensor_parameter_143: name = "parameter_143" + original_name = "batch_norm2d_11.w_0" shape = [1536] dtype = "float32" min_val = float("-0.00222853") @@ -1584,6 +1728,7 @@ class Program_weight_tensor_parameter_143: class Program_weight_tensor_parameter_144: name = "parameter_144" + original_name = "batch_norm2d_11.w_2" shape = [1536] dtype = "float32" min_val = float("5.21087e-09") @@ -1595,6 +1740,7 @@ class Program_weight_tensor_parameter_144: class Program_weight_tensor_parameter_145: name = "parameter_145" + original_name = "batch_norm2d_11.w_1" shape = [1536] dtype = "float32" min_val = float("-0.952451") @@ -1606,6 +1752,7 @@ class Program_weight_tensor_parameter_145: class Program_weight_tensor_parameter_146: name = "parameter_146" + original_name = "conv2d_27.w_0" shape = [1536, 768, 1, 1] dtype = "float32" min_val = float("-0.474375") @@ -1617,6 +1764,7 @@ class Program_weight_tensor_parameter_146: class Program_weight_tensor_parameter_147: name = "parameter_147" + original_name = "conv2d_29.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.323066") @@ -1628,6 +1776,7 @@ class Program_weight_tensor_parameter_147: class Program_weight_tensor_parameter_148: name = "parameter_148" + original_name = "conv2d_25.w_0" shape = [768, 1536, 1, 1] dtype = "float32" min_val = float("-0.462102") @@ -1639,6 +1788,7 @@ class Program_weight_tensor_parameter_148: class Program_weight_tensor_parameter_149: name = "parameter_149" + original_name = "batch_norm2d_10.b_0" shape = [1536] dtype = "float32" min_val = float("-0.35695") @@ -1650,6 +1800,7 @@ class Program_weight_tensor_parameter_149: class Program_weight_tensor_parameter_150: name = "parameter_150" + original_name = "batch_norm2d_10.w_0" shape = [1536] dtype = "float32" min_val = float("-0.00125723") @@ -1661,6 +1812,7 @@ class Program_weight_tensor_parameter_150: class Program_weight_tensor_parameter_151: name = "parameter_151" + original_name = "batch_norm2d_10.w_2" shape = [1536] dtype = "float32" min_val = float("7.51662e-09") @@ -1672,6 +1824,7 @@ class Program_weight_tensor_parameter_151: class Program_weight_tensor_parameter_152: name = "parameter_152" + original_name = "batch_norm2d_10.w_1" shape = [1536] dtype = "float32" min_val = float("-0.717346") @@ -1683,6 +1836,7 @@ class Program_weight_tensor_parameter_152: class Program_weight_tensor_parameter_153: name = "parameter_153" + original_name = "conv2d_24.w_0" shape = [1536, 768, 1, 1] dtype = "float32" min_val = float("-0.549117") @@ -1694,6 +1848,7 @@ class Program_weight_tensor_parameter_153: class Program_weight_tensor_parameter_154: name = "parameter_154" + original_name = "conv2d_26.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.360171") @@ -1705,6 +1860,7 @@ class Program_weight_tensor_parameter_154: class Program_weight_tensor_parameter_155: name = "parameter_155" + original_name = "batch_norm2d_9.b_0" shape = [768] dtype = "float32" min_val = float("-0.446058") @@ -1716,6 +1872,7 @@ class Program_weight_tensor_parameter_155: class Program_weight_tensor_parameter_156: name = "parameter_156" + original_name = "batch_norm2d_9.w_0" shape = [768] dtype = "float32" min_val = float("0.00064216") @@ -1727,6 +1884,7 @@ class Program_weight_tensor_parameter_156: class Program_weight_tensor_parameter_157: name = "parameter_157" + original_name = "batch_norm2d_9.w_2" shape = [768] dtype = "float32" min_val = float("0.074055") @@ -1738,6 +1896,7 @@ class Program_weight_tensor_parameter_157: class Program_weight_tensor_parameter_158: name = "parameter_158" + original_name = "batch_norm2d_9.w_1" shape = [768] dtype = "float32" min_val = float("-1.12888") @@ -1749,6 +1908,7 @@ class Program_weight_tensor_parameter_158: class Program_weight_tensor_parameter_159: name = "parameter_159" + original_name = "conv2d_23.w_0" shape = [768, 384, 2, 2] dtype = "float32" min_val = float("-0.64127") @@ -1760,6 +1920,7 @@ class Program_weight_tensor_parameter_159: class Program_weight_tensor_parameter_160: name = "parameter_160" + original_name = "conv2d_21.w_0" shape = [384, 768, 1, 1] dtype = "float32" min_val = float("-0.455305") @@ -1771,6 +1932,7 @@ class Program_weight_tensor_parameter_160: class Program_weight_tensor_parameter_161: name = "parameter_161" + original_name = "batch_norm2d_8.b_0" shape = [768] dtype = "float32" min_val = float("-0.195275") @@ -1782,6 +1944,7 @@ class Program_weight_tensor_parameter_161: class Program_weight_tensor_parameter_162: name = "parameter_162" + original_name = "batch_norm2d_8.w_0" shape = [768] dtype = "float32" min_val = float("0.0495166") @@ -1793,6 +1956,7 @@ class Program_weight_tensor_parameter_162: class Program_weight_tensor_parameter_163: name = "parameter_163" + original_name = "batch_norm2d_8.w_2" shape = [768] dtype = "float32" min_val = float("0.0111406") @@ -1804,6 +1968,7 @@ class Program_weight_tensor_parameter_163: class Program_weight_tensor_parameter_164: name = "parameter_164" + original_name = "batch_norm2d_8.w_1" shape = [768] dtype = "float32" min_val = float("-0.629735") @@ -1815,6 +1980,7 @@ class Program_weight_tensor_parameter_164: class Program_weight_tensor_parameter_165: name = "parameter_165" + original_name = "conv2d_20.w_0" shape = [768, 384, 1, 1] dtype = "float32" min_val = float("-0.471485") @@ -1826,6 +1992,7 @@ class Program_weight_tensor_parameter_165: class Program_weight_tensor_parameter_166: name = "parameter_166" + original_name = "conv2d_22.w_0" shape = [96, 96, 3, 3] dtype = "float32" min_val = float("-0.332823") @@ -1837,6 +2004,7 @@ class Program_weight_tensor_parameter_166: class Program_weight_tensor_parameter_167: name = "parameter_167" + original_name = "conv2d_18.w_0" shape = [384, 768, 1, 1] dtype = "float32" min_val = float("-0.460281") @@ -1848,6 +2016,7 @@ class Program_weight_tensor_parameter_167: class Program_weight_tensor_parameter_168: name = "parameter_168" + original_name = "batch_norm2d_7.b_0" shape = [768] dtype = "float32" min_val = float("-0.257379") @@ -1859,6 +2028,7 @@ class Program_weight_tensor_parameter_168: class Program_weight_tensor_parameter_169: name = "parameter_169" + original_name = "batch_norm2d_7.w_0" shape = [768] dtype = "float32" min_val = float("0.0331909") @@ -1870,6 +2040,7 @@ class Program_weight_tensor_parameter_169: class Program_weight_tensor_parameter_170: name = "parameter_170" + original_name = "batch_norm2d_7.w_2" shape = [768] dtype = "float32" min_val = float("0.00650611") @@ -1881,6 +2052,7 @@ class Program_weight_tensor_parameter_170: class Program_weight_tensor_parameter_171: name = "parameter_171" + original_name = "batch_norm2d_7.w_1" shape = [768] dtype = "float32" min_val = float("-0.442582") @@ -1892,6 +2064,7 @@ class Program_weight_tensor_parameter_171: class Program_weight_tensor_parameter_172: name = "parameter_172" + original_name = "conv2d_17.w_0" shape = [768, 384, 1, 1] dtype = "float32" min_val = float("-0.46642") @@ -1903,6 +2076,7 @@ class Program_weight_tensor_parameter_172: class Program_weight_tensor_parameter_173: name = "parameter_173" + original_name = "conv2d_19.w_0" shape = [96, 96, 3, 3] dtype = "float32" min_val = float("-0.335425") @@ -1914,6 +2088,7 @@ class Program_weight_tensor_parameter_173: class Program_weight_tensor_parameter_174: name = "parameter_174" + original_name = "conv2d_15.w_0" shape = [384, 768, 1, 1] dtype = "float32" min_val = float("-0.39159") @@ -1925,6 +2100,7 @@ class Program_weight_tensor_parameter_174: class Program_weight_tensor_parameter_175: name = "parameter_175" + original_name = "batch_norm2d_6.b_0" shape = [768] dtype = "float32" min_val = float("-0.227285") @@ -1936,6 +2112,7 @@ class Program_weight_tensor_parameter_175: class Program_weight_tensor_parameter_176: name = "parameter_176" + original_name = "batch_norm2d_6.w_0" shape = [768] dtype = "float32" min_val = float("0.0091738") @@ -1947,6 +2124,7 @@ class Program_weight_tensor_parameter_176: class Program_weight_tensor_parameter_177: name = "parameter_177" + original_name = "batch_norm2d_6.w_2" shape = [768] dtype = "float32" min_val = float("0.00489244") @@ -1958,6 +2136,7 @@ class Program_weight_tensor_parameter_177: class Program_weight_tensor_parameter_178: name = "parameter_178" + original_name = "batch_norm2d_6.w_1" shape = [768] dtype = "float32" min_val = float("-0.519454") @@ -1969,6 +2148,7 @@ class Program_weight_tensor_parameter_178: class Program_weight_tensor_parameter_179: name = "parameter_179" + original_name = "conv2d_14.w_0" shape = [768, 384, 1, 1] dtype = "float32" min_val = float("-0.351532") @@ -1980,6 +2160,7 @@ class Program_weight_tensor_parameter_179: class Program_weight_tensor_parameter_180: name = "parameter_180" + original_name = "conv2d_16.w_0" shape = [96, 96, 3, 3] dtype = "float32" min_val = float("-0.359459") @@ -1991,6 +2172,7 @@ class Program_weight_tensor_parameter_180: class Program_weight_tensor_parameter_181: name = "parameter_181" + original_name = "conv2d_12.w_0" shape = [384, 768, 1, 1] dtype = "float32" min_val = float("-0.763782") @@ -2002,6 +2184,7 @@ class Program_weight_tensor_parameter_181: class Program_weight_tensor_parameter_182: name = "parameter_182" + original_name = "batch_norm2d_5.b_0" shape = [768] dtype = "float32" min_val = float("-0.244577") @@ -2013,6 +2196,7 @@ class Program_weight_tensor_parameter_182: class Program_weight_tensor_parameter_183: name = "parameter_183" + original_name = "batch_norm2d_5.w_0" shape = [768] dtype = "float32" min_val = float("-0.00133899") @@ -2024,6 +2208,7 @@ class Program_weight_tensor_parameter_183: class Program_weight_tensor_parameter_184: name = "parameter_184" + original_name = "batch_norm2d_5.w_2" shape = [768] dtype = "float32" min_val = float("1.15577e-08") @@ -2035,6 +2220,7 @@ class Program_weight_tensor_parameter_184: class Program_weight_tensor_parameter_185: name = "parameter_185" + original_name = "batch_norm2d_5.w_1" shape = [768] dtype = "float32" min_val = float("-0.441419") @@ -2046,6 +2232,7 @@ class Program_weight_tensor_parameter_185: class Program_weight_tensor_parameter_186: name = "parameter_186" + original_name = "conv2d_11.w_0" shape = [768, 384, 1, 1] dtype = "float32" min_val = float("-0.532942") @@ -2057,6 +2244,7 @@ class Program_weight_tensor_parameter_186: class Program_weight_tensor_parameter_187: name = "parameter_187" + original_name = "conv2d_13.w_0" shape = [96, 96, 3, 3] dtype = "float32" min_val = float("-0.364193") @@ -2068,6 +2256,7 @@ class Program_weight_tensor_parameter_187: class Program_weight_tensor_parameter_188: name = "parameter_188" + original_name = "batch_norm2d_4.b_0" shape = [384] dtype = "float32" min_val = float("-0.137062") @@ -2079,6 +2268,7 @@ class Program_weight_tensor_parameter_188: class Program_weight_tensor_parameter_189: name = "parameter_189" + original_name = "batch_norm2d_4.w_0" shape = [384] dtype = "float32" min_val = float("0.0122036") @@ -2090,6 +2280,7 @@ class Program_weight_tensor_parameter_189: class Program_weight_tensor_parameter_190: name = "parameter_190" + original_name = "batch_norm2d_4.w_2" shape = [384] dtype = "float32" min_val = float("0.0213914") @@ -2101,6 +2292,7 @@ class Program_weight_tensor_parameter_190: class Program_weight_tensor_parameter_191: name = "parameter_191" + original_name = "batch_norm2d_4.w_1" shape = [384] dtype = "float32" min_val = float("-0.557845") @@ -2112,6 +2304,7 @@ class Program_weight_tensor_parameter_191: class Program_weight_tensor_parameter_192: name = "parameter_192" + original_name = "conv2d_10.w_0" shape = [384, 192, 2, 2] dtype = "float32" min_val = float("-0.493315") @@ -2123,6 +2316,7 @@ class Program_weight_tensor_parameter_192: class Program_weight_tensor_parameter_193: name = "parameter_193" + original_name = "conv2d_8.w_0" shape = [192, 384, 1, 1] dtype = "float32" min_val = float("-0.383636") @@ -2134,6 +2328,7 @@ class Program_weight_tensor_parameter_193: class Program_weight_tensor_parameter_194: name = "parameter_194" + original_name = "batch_norm2d_3.b_0" shape = [384] dtype = "float32" min_val = float("-0.171066") @@ -2145,6 +2340,7 @@ class Program_weight_tensor_parameter_194: class Program_weight_tensor_parameter_195: name = "parameter_195" + original_name = "batch_norm2d_3.w_0" shape = [384] dtype = "float32" min_val = float("0.0143999") @@ -2156,6 +2352,7 @@ class Program_weight_tensor_parameter_195: class Program_weight_tensor_parameter_196: name = "parameter_196" + original_name = "batch_norm2d_3.w_2" shape = [384] dtype = "float32" min_val = float("0.00157374") @@ -2167,6 +2364,7 @@ class Program_weight_tensor_parameter_196: class Program_weight_tensor_parameter_197: name = "parameter_197" + original_name = "batch_norm2d_3.w_1" shape = [384] dtype = "float32" min_val = float("-0.212639") @@ -2178,6 +2376,7 @@ class Program_weight_tensor_parameter_197: class Program_weight_tensor_parameter_198: name = "parameter_198" + original_name = "conv2d_7.w_0" shape = [384, 192, 1, 1] dtype = "float32" min_val = float("-0.45981") @@ -2189,6 +2388,7 @@ class Program_weight_tensor_parameter_198: class Program_weight_tensor_parameter_199: name = "parameter_199" + original_name = "conv2d_9.w_0" shape = [48, 48, 3, 3] dtype = "float32" min_val = float("-0.339738") @@ -2200,6 +2400,7 @@ class Program_weight_tensor_parameter_199: class Program_weight_tensor_parameter_200: name = "parameter_200" + original_name = "conv2d_5.w_0" shape = [192, 384, 1, 1] dtype = "float32" min_val = float("-0.554372") @@ -2211,6 +2412,7 @@ class Program_weight_tensor_parameter_200: class Program_weight_tensor_parameter_201: name = "parameter_201" + original_name = "batch_norm2d_2.b_0" shape = [384] dtype = "float32" min_val = float("-0.126871") @@ -2222,6 +2424,7 @@ class Program_weight_tensor_parameter_201: class Program_weight_tensor_parameter_202: name = "parameter_202" + original_name = "batch_norm2d_2.w_0" shape = [384] dtype = "float32" min_val = float("0.0498806") @@ -2233,6 +2436,7 @@ class Program_weight_tensor_parameter_202: class Program_weight_tensor_parameter_203: name = "parameter_203" + original_name = "batch_norm2d_2.w_2" shape = [384] dtype = "float32" min_val = float("0.00112645") @@ -2244,6 +2448,7 @@ class Program_weight_tensor_parameter_203: class Program_weight_tensor_parameter_204: name = "parameter_204" + original_name = "batch_norm2d_2.w_1" shape = [384] dtype = "float32" min_val = float("-0.173821") @@ -2255,6 +2460,7 @@ class Program_weight_tensor_parameter_204: class Program_weight_tensor_parameter_205: name = "parameter_205" + original_name = "conv2d_4.w_0" shape = [384, 192, 1, 1] dtype = "float32" min_val = float("-0.481591") @@ -2266,6 +2472,7 @@ class Program_weight_tensor_parameter_205: class Program_weight_tensor_parameter_206: name = "parameter_206" + original_name = "conv2d_6.w_0" shape = [48, 48, 3, 3] dtype = "float32" min_val = float("-0.386428") @@ -2277,6 +2484,7 @@ class Program_weight_tensor_parameter_206: class Program_weight_tensor_parameter_207: name = "parameter_207" + original_name = "conv2d_2.w_0" shape = [192, 384, 1, 1] dtype = "float32" min_val = float("-0.407018") @@ -2288,6 +2496,7 @@ class Program_weight_tensor_parameter_207: class Program_weight_tensor_parameter_208: name = "parameter_208" + original_name = "batch_norm2d_1.b_0" shape = [384] dtype = "float32" min_val = float("-0.1664") @@ -2299,6 +2508,7 @@ class Program_weight_tensor_parameter_208: class Program_weight_tensor_parameter_209: name = "parameter_209" + original_name = "batch_norm2d_1.w_0" shape = [384] dtype = "float32" min_val = float("-0.000137082") @@ -2310,6 +2520,7 @@ class Program_weight_tensor_parameter_209: class Program_weight_tensor_parameter_210: name = "parameter_210" + original_name = "batch_norm2d_1.w_2" shape = [384] dtype = "float32" min_val = float("6.59057e-12") @@ -2321,6 +2532,7 @@ class Program_weight_tensor_parameter_210: class Program_weight_tensor_parameter_211: name = "parameter_211" + original_name = "batch_norm2d_1.w_1" shape = [384] dtype = "float32" min_val = float("-0.246698") @@ -2332,6 +2544,7 @@ class Program_weight_tensor_parameter_211: class Program_weight_tensor_parameter_212: name = "parameter_212" + original_name = "conv2d_1.w_0" shape = [384, 192, 1, 1] dtype = "float32" min_val = float("-2.05381") @@ -2343,6 +2556,7 @@ class Program_weight_tensor_parameter_212: class Program_weight_tensor_parameter_213: name = "parameter_213" + original_name = "conv2d_3.w_0" shape = [48, 48, 3, 3] dtype = "float32" min_val = float("-0.656575") @@ -2354,6 +2568,7 @@ class Program_weight_tensor_parameter_213: class Program_weight_tensor_parameter_214: name = "parameter_214" + original_name = "batch_norm2d_0.b_0" shape = [192] dtype = "float32" min_val = float("-0.180182") @@ -2365,6 +2580,7 @@ class Program_weight_tensor_parameter_214: class Program_weight_tensor_parameter_215: name = "parameter_215" + original_name = "batch_norm2d_0.w_0" shape = [192] dtype = "float32" min_val = float("-0.00710445") @@ -2376,6 +2592,7 @@ class Program_weight_tensor_parameter_215: class Program_weight_tensor_parameter_216: name = "parameter_216" + original_name = "batch_norm2d_0.w_2" shape = [192] dtype = "float32" min_val = float("0.00206879") @@ -2387,6 +2604,7 @@ class Program_weight_tensor_parameter_216: class Program_weight_tensor_parameter_217: name = "parameter_217" + original_name = "batch_norm2d_0.w_1" shape = [192] dtype = "float32" min_val = float("-0.170337") @@ -2398,6 +2616,7 @@ class Program_weight_tensor_parameter_217: class Program_weight_tensor_parameter_218: name = "parameter_218" + original_name = "conv2d_0.w_0" shape = [192, 3, 4, 4] dtype = "float32" min_val = float("-0.295166") diff --git a/paddle_samples/PaddleX/FasterNet-T0/subgraph_1/input_meta.py b/paddle_samples/PaddleX/FasterNet-T0/subgraph_1/input_meta.py index d50f92c45..a3f24694e 100644 --- a/paddle_samples/PaddleX/FasterNet-T0/subgraph_1/input_meta.py +++ b/paddle_samples/PaddleX/FasterNet-T0/subgraph_1/input_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_data_0: name = "data_0" + original_name = "var_0" shape = [128, 3, 224, 224] dtype = "float32" min_val = float("-2.1179") diff --git a/paddle_samples/PaddleX/FasterNet-T0/subgraph_1/weight_meta.py b/paddle_samples/PaddleX/FasterNet-T0/subgraph_1/weight_meta.py index 542a838e8..b31684231 100644 --- a/paddle_samples/PaddleX/FasterNet-T0/subgraph_1/weight_meta.py +++ b/paddle_samples/PaddleX/FasterNet-T0/subgraph_1/weight_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_parameter_0: name = "parameter_0" + original_name = "linear_0.b_0" shape = [102] dtype = "float32" data = None @@ -7,6 +8,7 @@ class Program_weight_tensor_parameter_0: class Program_weight_tensor_parameter_1: name = "parameter_1" + original_name = "linear_0.w_0" shape = [1280, 102] dtype = "float32" min_val = float("-0.0861183") @@ -18,6 +20,7 @@ class Program_weight_tensor_parameter_1: class Program_weight_tensor_parameter_2: name = "parameter_2" + original_name = "conv2d_43.w_0" shape = [1280, 320, 1, 1] dtype = "float32" min_val = float("-1.30145") @@ -29,6 +32,7 @@ class Program_weight_tensor_parameter_2: class Program_weight_tensor_parameter_3: name = "parameter_3" + original_name = "conv2d_41.w_0" shape = [320, 640, 1, 1] dtype = "float32" min_val = float("-2.34502") @@ -40,6 +44,7 @@ class Program_weight_tensor_parameter_3: class Program_weight_tensor_parameter_4: name = "parameter_4" + original_name = "batch_norm2d_16.b_0" shape = [640] dtype = "float32" min_val = float("-3.22377") @@ -51,6 +56,7 @@ class Program_weight_tensor_parameter_4: class Program_weight_tensor_parameter_5: name = "parameter_5" + original_name = "batch_norm2d_16.w_0" shape = [640] dtype = "float32" min_val = float("0.195941") @@ -62,6 +68,7 @@ class Program_weight_tensor_parameter_5: class Program_weight_tensor_parameter_6: name = "parameter_6" + original_name = "batch_norm2d_16.w_2" shape = [640] dtype = "float32" min_val = float("43.1267") @@ -73,6 +80,7 @@ class Program_weight_tensor_parameter_6: class Program_weight_tensor_parameter_7: name = "parameter_7" + original_name = "batch_norm2d_16.w_1" shape = [640] dtype = "float32" min_val = float("-3.58898") @@ -84,6 +92,7 @@ class Program_weight_tensor_parameter_7: class Program_weight_tensor_parameter_8: name = "parameter_8" + original_name = "conv2d_40.w_0" shape = [640, 320, 1, 1] dtype = "float32" min_val = float("-2.58741") @@ -95,6 +104,7 @@ class Program_weight_tensor_parameter_8: class Program_weight_tensor_parameter_9: name = "parameter_9" + original_name = "conv2d_42.w_0" shape = [80, 80, 3, 3] dtype = "float32" min_val = float("-1.05712") @@ -106,6 +116,7 @@ class Program_weight_tensor_parameter_9: class Program_weight_tensor_parameter_10: name = "parameter_10" + original_name = "conv2d_38.w_0" shape = [320, 640, 1, 1] dtype = "float32" min_val = float("-1.72649") @@ -117,6 +128,7 @@ class Program_weight_tensor_parameter_10: class Program_weight_tensor_parameter_11: name = "parameter_11" + original_name = "batch_norm2d_15.b_0" shape = [640] dtype = "float32" min_val = float("-2.40291") @@ -128,6 +140,7 @@ class Program_weight_tensor_parameter_11: class Program_weight_tensor_parameter_12: name = "parameter_12" + original_name = "batch_norm2d_15.w_0" shape = [640] dtype = "float32" min_val = float("0.169422") @@ -139,6 +152,7 @@ class Program_weight_tensor_parameter_12: class Program_weight_tensor_parameter_13: name = "parameter_13" + original_name = "batch_norm2d_15.w_2" shape = [640] dtype = "float32" min_val = float("12.9165") @@ -150,6 +164,7 @@ class Program_weight_tensor_parameter_13: class Program_weight_tensor_parameter_14: name = "parameter_14" + original_name = "batch_norm2d_15.w_1" shape = [640] dtype = "float32" min_val = float("-10.6163") @@ -161,6 +176,7 @@ class Program_weight_tensor_parameter_14: class Program_weight_tensor_parameter_15: name = "parameter_15" + original_name = "conv2d_37.w_0" shape = [640, 320, 1, 1] dtype = "float32" min_val = float("-2.16494") @@ -172,6 +188,7 @@ class Program_weight_tensor_parameter_15: class Program_weight_tensor_parameter_16: name = "parameter_16" + original_name = "conv2d_39.w_0" shape = [80, 80, 3, 3] dtype = "float32" min_val = float("-2.15269") @@ -183,6 +200,7 @@ class Program_weight_tensor_parameter_16: class Program_weight_tensor_parameter_17: name = "parameter_17" + original_name = "batch_norm2d_14.b_0" shape = [320] dtype = "float32" min_val = float("-0.38275") @@ -194,6 +212,7 @@ class Program_weight_tensor_parameter_17: class Program_weight_tensor_parameter_18: name = "parameter_18" + original_name = "batch_norm2d_14.w_0" shape = [320] dtype = "float32" min_val = float("0.0445683") @@ -205,6 +224,7 @@ class Program_weight_tensor_parameter_18: class Program_weight_tensor_parameter_19: name = "parameter_19" + original_name = "batch_norm2d_14.w_2" shape = [320] dtype = "float32" min_val = float("1656.74") @@ -216,6 +236,7 @@ class Program_weight_tensor_parameter_19: class Program_weight_tensor_parameter_20: name = "parameter_20" + original_name = "batch_norm2d_14.w_1" shape = [320] dtype = "float32" min_val = float("-244.296") @@ -227,6 +248,7 @@ class Program_weight_tensor_parameter_20: class Program_weight_tensor_parameter_21: name = "parameter_21" + original_name = "conv2d_36.w_0" shape = [320, 160, 2, 2] dtype = "float32" min_val = float("-2.3106") @@ -238,6 +260,7 @@ class Program_weight_tensor_parameter_21: class Program_weight_tensor_parameter_22: name = "parameter_22" + original_name = "conv2d_34.w_0" shape = [160, 320, 1, 1] dtype = "float32" min_val = float("-2.34633") @@ -249,6 +272,7 @@ class Program_weight_tensor_parameter_22: class Program_weight_tensor_parameter_23: name = "parameter_23" + original_name = "batch_norm2d_13.b_0" shape = [320] dtype = "float32" min_val = float("-1.75704") @@ -260,6 +284,7 @@ class Program_weight_tensor_parameter_23: class Program_weight_tensor_parameter_24: name = "parameter_24" + original_name = "batch_norm2d_13.w_0" shape = [320] dtype = "float32" min_val = float("0.50175") @@ -271,6 +296,7 @@ class Program_weight_tensor_parameter_24: class Program_weight_tensor_parameter_25: name = "parameter_25" + original_name = "batch_norm2d_13.w_2" shape = [320] dtype = "float32" min_val = float("277.204") @@ -282,6 +308,7 @@ class Program_weight_tensor_parameter_25: class Program_weight_tensor_parameter_26: name = "parameter_26" + original_name = "batch_norm2d_13.w_1" shape = [320] dtype = "float32" min_val = float("-93.3597") @@ -293,6 +320,7 @@ class Program_weight_tensor_parameter_26: class Program_weight_tensor_parameter_27: name = "parameter_27" + original_name = "conv2d_33.w_0" shape = [320, 160, 1, 1] dtype = "float32" min_val = float("-2.27852") @@ -304,6 +332,7 @@ class Program_weight_tensor_parameter_27: class Program_weight_tensor_parameter_28: name = "parameter_28" + original_name = "conv2d_35.w_0" shape = [40, 40, 3, 3] dtype = "float32" min_val = float("-1.07355") @@ -315,6 +344,7 @@ class Program_weight_tensor_parameter_28: class Program_weight_tensor_parameter_29: name = "parameter_29" + original_name = "conv2d_31.w_0" shape = [160, 320, 1, 1] dtype = "float32" min_val = float("-2.38353") @@ -326,6 +356,7 @@ class Program_weight_tensor_parameter_29: class Program_weight_tensor_parameter_30: name = "parameter_30" + original_name = "batch_norm2d_12.b_0" shape = [320] dtype = "float32" min_val = float("-1.73404") @@ -337,6 +368,7 @@ class Program_weight_tensor_parameter_30: class Program_weight_tensor_parameter_31: name = "parameter_31" + original_name = "batch_norm2d_12.w_0" shape = [320] dtype = "float32" min_val = float("0.367442") @@ -348,6 +380,7 @@ class Program_weight_tensor_parameter_31: class Program_weight_tensor_parameter_32: name = "parameter_32" + original_name = "batch_norm2d_12.w_2" shape = [320] dtype = "float32" min_val = float("157.784") @@ -359,6 +392,7 @@ class Program_weight_tensor_parameter_32: class Program_weight_tensor_parameter_33: name = "parameter_33" + original_name = "batch_norm2d_12.w_1" shape = [320] dtype = "float32" min_val = float("-117.727") @@ -370,6 +404,7 @@ class Program_weight_tensor_parameter_33: class Program_weight_tensor_parameter_34: name = "parameter_34" + original_name = "conv2d_30.w_0" shape = [320, 160, 1, 1] dtype = "float32" min_val = float("-2.19397") @@ -381,6 +416,7 @@ class Program_weight_tensor_parameter_34: class Program_weight_tensor_parameter_35: name = "parameter_35" + original_name = "conv2d_32.w_0" shape = [40, 40, 3, 3] dtype = "float32" min_val = float("-1.13801") @@ -392,6 +428,7 @@ class Program_weight_tensor_parameter_35: class Program_weight_tensor_parameter_36: name = "parameter_36" + original_name = "conv2d_28.w_0" shape = [160, 320, 1, 1] dtype = "float32" min_val = float("-2.37843") @@ -403,6 +440,7 @@ class Program_weight_tensor_parameter_36: class Program_weight_tensor_parameter_37: name = "parameter_37" + original_name = "batch_norm2d_11.b_0" shape = [320] dtype = "float32" min_val = float("-1.75063") @@ -414,6 +452,7 @@ class Program_weight_tensor_parameter_37: class Program_weight_tensor_parameter_38: name = "parameter_38" + original_name = "batch_norm2d_11.w_0" shape = [320] dtype = "float32" min_val = float("0.300191") @@ -425,6 +464,7 @@ class Program_weight_tensor_parameter_38: class Program_weight_tensor_parameter_39: name = "parameter_39" + original_name = "batch_norm2d_11.w_2" shape = [320] dtype = "float32" min_val = float("115.908") @@ -436,6 +476,7 @@ class Program_weight_tensor_parameter_39: class Program_weight_tensor_parameter_40: name = "parameter_40" + original_name = "batch_norm2d_11.w_1" shape = [320] dtype = "float32" min_val = float("-93.808") @@ -447,6 +488,7 @@ class Program_weight_tensor_parameter_40: class Program_weight_tensor_parameter_41: name = "parameter_41" + original_name = "conv2d_27.w_0" shape = [320, 160, 1, 1] dtype = "float32" min_val = float("-2.25214") @@ -458,6 +500,7 @@ class Program_weight_tensor_parameter_41: class Program_weight_tensor_parameter_42: name = "parameter_42" + original_name = "conv2d_29.w_0" shape = [40, 40, 3, 3] dtype = "float32" min_val = float("-1.22762") @@ -469,6 +512,7 @@ class Program_weight_tensor_parameter_42: class Program_weight_tensor_parameter_43: name = "parameter_43" + original_name = "conv2d_25.w_0" shape = [160, 320, 1, 1] dtype = "float32" min_val = float("-2.01307") @@ -480,6 +524,7 @@ class Program_weight_tensor_parameter_43: class Program_weight_tensor_parameter_44: name = "parameter_44" + original_name = "batch_norm2d_10.b_0" shape = [320] dtype = "float32" min_val = float("-1.5061") @@ -491,6 +536,7 @@ class Program_weight_tensor_parameter_44: class Program_weight_tensor_parameter_45: name = "parameter_45" + original_name = "batch_norm2d_10.w_0" shape = [320] dtype = "float32" min_val = float("0.233548") @@ -502,6 +548,7 @@ class Program_weight_tensor_parameter_45: class Program_weight_tensor_parameter_46: name = "parameter_46" + original_name = "batch_norm2d_10.w_2" shape = [320] dtype = "float32" min_val = float("78.1731") @@ -513,6 +560,7 @@ class Program_weight_tensor_parameter_46: class Program_weight_tensor_parameter_47: name = "parameter_47" + original_name = "batch_norm2d_10.w_1" shape = [320] dtype = "float32" min_val = float("-60.4873") @@ -524,6 +572,7 @@ class Program_weight_tensor_parameter_47: class Program_weight_tensor_parameter_48: name = "parameter_48" + original_name = "conv2d_24.w_0" shape = [320, 160, 1, 1] dtype = "float32" min_val = float("-2.04087") @@ -535,6 +584,7 @@ class Program_weight_tensor_parameter_48: class Program_weight_tensor_parameter_49: name = "parameter_49" + original_name = "conv2d_26.w_0" shape = [40, 40, 3, 3] dtype = "float32" min_val = float("-1.66886") @@ -546,6 +596,7 @@ class Program_weight_tensor_parameter_49: class Program_weight_tensor_parameter_50: name = "parameter_50" + original_name = "conv2d_22.w_0" shape = [160, 320, 1, 1] dtype = "float32" min_val = float("-2.08546") @@ -557,6 +608,7 @@ class Program_weight_tensor_parameter_50: class Program_weight_tensor_parameter_51: name = "parameter_51" + original_name = "batch_norm2d_9.b_0" shape = [320] dtype = "float32" min_val = float("-2.03017") @@ -568,6 +620,7 @@ class Program_weight_tensor_parameter_51: class Program_weight_tensor_parameter_52: name = "parameter_52" + original_name = "batch_norm2d_9.w_0" shape = [320] dtype = "float32" min_val = float("0.200585") @@ -579,6 +632,7 @@ class Program_weight_tensor_parameter_52: class Program_weight_tensor_parameter_53: name = "parameter_53" + original_name = "batch_norm2d_9.w_2" shape = [320] dtype = "float32" min_val = float("30.0456") @@ -590,6 +644,7 @@ class Program_weight_tensor_parameter_53: class Program_weight_tensor_parameter_54: name = "parameter_54" + original_name = "batch_norm2d_9.w_1" shape = [320] dtype = "float32" min_val = float("-122.219") @@ -601,6 +656,7 @@ class Program_weight_tensor_parameter_54: class Program_weight_tensor_parameter_55: name = "parameter_55" + original_name = "conv2d_21.w_0" shape = [320, 160, 1, 1] dtype = "float32" min_val = float("-2.20477") @@ -612,6 +668,7 @@ class Program_weight_tensor_parameter_55: class Program_weight_tensor_parameter_56: name = "parameter_56" + original_name = "conv2d_23.w_0" shape = [40, 40, 3, 3] dtype = "float32" min_val = float("-1.54952") @@ -623,6 +680,7 @@ class Program_weight_tensor_parameter_56: class Program_weight_tensor_parameter_57: name = "parameter_57" + original_name = "conv2d_19.w_0" shape = [160, 320, 1, 1] dtype = "float32" min_val = float("-1.92693") @@ -634,6 +692,7 @@ class Program_weight_tensor_parameter_57: class Program_weight_tensor_parameter_58: name = "parameter_58" + original_name = "batch_norm2d_8.b_0" shape = [320] dtype = "float32" min_val = float("-1.5439") @@ -645,6 +704,7 @@ class Program_weight_tensor_parameter_58: class Program_weight_tensor_parameter_59: name = "parameter_59" + original_name = "batch_norm2d_8.w_0" shape = [320] dtype = "float32" min_val = float("0.15538") @@ -656,6 +716,7 @@ class Program_weight_tensor_parameter_59: class Program_weight_tensor_parameter_60: name = "parameter_60" + original_name = "batch_norm2d_8.w_2" shape = [320] dtype = "float32" min_val = float("11.1617") @@ -667,6 +728,7 @@ class Program_weight_tensor_parameter_60: class Program_weight_tensor_parameter_61: name = "parameter_61" + original_name = "batch_norm2d_8.w_1" shape = [320] dtype = "float32" min_val = float("-75.1346") @@ -678,6 +740,7 @@ class Program_weight_tensor_parameter_61: class Program_weight_tensor_parameter_62: name = "parameter_62" + original_name = "conv2d_18.w_0" shape = [320, 160, 1, 1] dtype = "float32" min_val = float("-2.6015") @@ -689,6 +752,7 @@ class Program_weight_tensor_parameter_62: class Program_weight_tensor_parameter_63: name = "parameter_63" + original_name = "conv2d_20.w_0" shape = [40, 40, 3, 3] dtype = "float32" min_val = float("-1.72255") @@ -700,6 +764,7 @@ class Program_weight_tensor_parameter_63: class Program_weight_tensor_parameter_64: name = "parameter_64" + original_name = "conv2d_16.w_0" shape = [160, 320, 1, 1] dtype = "float32" min_val = float("-2.14367") @@ -711,6 +776,7 @@ class Program_weight_tensor_parameter_64: class Program_weight_tensor_parameter_65: name = "parameter_65" + original_name = "batch_norm2d_7.b_0" shape = [320] dtype = "float32" min_val = float("-1.66179") @@ -722,6 +788,7 @@ class Program_weight_tensor_parameter_65: class Program_weight_tensor_parameter_66: name = "parameter_66" + original_name = "batch_norm2d_7.w_0" shape = [320] dtype = "float32" min_val = float("0.138334") @@ -733,6 +800,7 @@ class Program_weight_tensor_parameter_66: class Program_weight_tensor_parameter_67: name = "parameter_67" + original_name = "batch_norm2d_7.w_2" shape = [320] dtype = "float32" min_val = float("10.4824") @@ -744,6 +812,7 @@ class Program_weight_tensor_parameter_67: class Program_weight_tensor_parameter_68: name = "parameter_68" + original_name = "batch_norm2d_7.w_1" shape = [320] dtype = "float32" min_val = float("-39.3742") @@ -755,6 +824,7 @@ class Program_weight_tensor_parameter_68: class Program_weight_tensor_parameter_69: name = "parameter_69" + original_name = "conv2d_15.w_0" shape = [320, 160, 1, 1] dtype = "float32" min_val = float("-1.98501") @@ -766,6 +836,7 @@ class Program_weight_tensor_parameter_69: class Program_weight_tensor_parameter_70: name = "parameter_70" + original_name = "conv2d_17.w_0" shape = [40, 40, 3, 3] dtype = "float32" min_val = float("-1.66839") @@ -777,6 +848,7 @@ class Program_weight_tensor_parameter_70: class Program_weight_tensor_parameter_71: name = "parameter_71" + original_name = "conv2d_13.w_0" shape = [160, 320, 1, 1] dtype = "float32" min_val = float("-2.2387") @@ -788,6 +860,7 @@ class Program_weight_tensor_parameter_71: class Program_weight_tensor_parameter_72: name = "parameter_72" + original_name = "batch_norm2d_6.b_0" shape = [320] dtype = "float32" min_val = float("-1.37278") @@ -799,6 +872,7 @@ class Program_weight_tensor_parameter_72: class Program_weight_tensor_parameter_73: name = "parameter_73" + original_name = "batch_norm2d_6.w_0" shape = [320] dtype = "float32" min_val = float("0.0986435") @@ -810,6 +884,7 @@ class Program_weight_tensor_parameter_73: class Program_weight_tensor_parameter_74: name = "parameter_74" + original_name = "batch_norm2d_6.w_2" shape = [320] dtype = "float32" min_val = float("6.24234") @@ -821,6 +896,7 @@ class Program_weight_tensor_parameter_74: class Program_weight_tensor_parameter_75: name = "parameter_75" + original_name = "batch_norm2d_6.w_1" shape = [320] dtype = "float32" min_val = float("-22.6372") @@ -832,6 +908,7 @@ class Program_weight_tensor_parameter_75: class Program_weight_tensor_parameter_76: name = "parameter_76" + original_name = "conv2d_12.w_0" shape = [320, 160, 1, 1] dtype = "float32" min_val = float("-2.60726") @@ -843,6 +920,7 @@ class Program_weight_tensor_parameter_76: class Program_weight_tensor_parameter_77: name = "parameter_77" + original_name = "conv2d_14.w_0" shape = [40, 40, 3, 3] dtype = "float32" min_val = float("-1.81922") @@ -854,6 +932,7 @@ class Program_weight_tensor_parameter_77: class Program_weight_tensor_parameter_78: name = "parameter_78" + original_name = "batch_norm2d_5.b_0" shape = [160] dtype = "float32" min_val = float("-1.43576") @@ -865,6 +944,7 @@ class Program_weight_tensor_parameter_78: class Program_weight_tensor_parameter_79: name = "parameter_79" + original_name = "batch_norm2d_5.w_0" shape = [160] dtype = "float32" min_val = float("0.109511") @@ -876,6 +956,7 @@ class Program_weight_tensor_parameter_79: class Program_weight_tensor_parameter_80: name = "parameter_80" + original_name = "batch_norm2d_5.w_2" shape = [160] dtype = "float32" min_val = float("388.977") @@ -887,6 +968,7 @@ class Program_weight_tensor_parameter_80: class Program_weight_tensor_parameter_81: name = "parameter_81" + original_name = "batch_norm2d_5.w_1" shape = [160] dtype = "float32" min_val = float("-44.6769") @@ -898,6 +980,7 @@ class Program_weight_tensor_parameter_81: class Program_weight_tensor_parameter_82: name = "parameter_82" + original_name = "conv2d_11.w_0" shape = [160, 80, 2, 2] dtype = "float32" min_val = float("-1.76845") @@ -909,6 +992,7 @@ class Program_weight_tensor_parameter_82: class Program_weight_tensor_parameter_83: name = "parameter_83" + original_name = "conv2d_9.w_0" shape = [80, 160, 1, 1] dtype = "float32" min_val = float("-2.17258") @@ -920,6 +1004,7 @@ class Program_weight_tensor_parameter_83: class Program_weight_tensor_parameter_84: name = "parameter_84" + original_name = "batch_norm2d_4.b_0" shape = [160] dtype = "float32" min_val = float("-1.71082") @@ -931,6 +1016,7 @@ class Program_weight_tensor_parameter_84: class Program_weight_tensor_parameter_85: name = "parameter_85" + original_name = "batch_norm2d_4.w_0" shape = [160] dtype = "float32" min_val = float("0.523331") @@ -942,6 +1028,7 @@ class Program_weight_tensor_parameter_85: class Program_weight_tensor_parameter_86: name = "parameter_86" + original_name = "batch_norm2d_4.w_2" shape = [160] dtype = "float32" min_val = float("16.534") @@ -953,6 +1040,7 @@ class Program_weight_tensor_parameter_86: class Program_weight_tensor_parameter_87: name = "parameter_87" + original_name = "batch_norm2d_4.w_1" shape = [160] dtype = "float32" min_val = float("-8.30639") @@ -964,6 +1052,7 @@ class Program_weight_tensor_parameter_87: class Program_weight_tensor_parameter_88: name = "parameter_88" + original_name = "conv2d_8.w_0" shape = [160, 80, 1, 1] dtype = "float32" min_val = float("-2.19078") @@ -975,6 +1064,7 @@ class Program_weight_tensor_parameter_88: class Program_weight_tensor_parameter_89: name = "parameter_89" + original_name = "conv2d_10.w_0" shape = [20, 20, 3, 3] dtype = "float32" min_val = float("-1.435") @@ -986,6 +1076,7 @@ class Program_weight_tensor_parameter_89: class Program_weight_tensor_parameter_90: name = "parameter_90" + original_name = "conv2d_6.w_0" shape = [80, 160, 1, 1] dtype = "float32" min_val = float("-2.40332") @@ -997,6 +1088,7 @@ class Program_weight_tensor_parameter_90: class Program_weight_tensor_parameter_91: name = "parameter_91" + original_name = "batch_norm2d_3.b_0" shape = [160] dtype = "float32" min_val = float("-1.72903") @@ -1008,6 +1100,7 @@ class Program_weight_tensor_parameter_91: class Program_weight_tensor_parameter_92: name = "parameter_92" + original_name = "batch_norm2d_3.w_0" shape = [160] dtype = "float32" min_val = float("0.302703") @@ -1019,6 +1112,7 @@ class Program_weight_tensor_parameter_92: class Program_weight_tensor_parameter_93: name = "parameter_93" + original_name = "batch_norm2d_3.w_2" shape = [160] dtype = "float32" min_val = float("7.23398") @@ -1030,6 +1124,7 @@ class Program_weight_tensor_parameter_93: class Program_weight_tensor_parameter_94: name = "parameter_94" + original_name = "batch_norm2d_3.w_1" shape = [160] dtype = "float32" min_val = float("-13.6405") @@ -1041,6 +1136,7 @@ class Program_weight_tensor_parameter_94: class Program_weight_tensor_parameter_95: name = "parameter_95" + original_name = "conv2d_5.w_0" shape = [160, 80, 1, 1] dtype = "float32" min_val = float("-2.09213") @@ -1052,6 +1148,7 @@ class Program_weight_tensor_parameter_95: class Program_weight_tensor_parameter_96: name = "parameter_96" + original_name = "conv2d_7.w_0" shape = [20, 20, 3, 3] dtype = "float32" min_val = float("-1.38734") @@ -1063,6 +1160,7 @@ class Program_weight_tensor_parameter_96: class Program_weight_tensor_parameter_97: name = "parameter_97" + original_name = "batch_norm2d_2.b_0" shape = [80] dtype = "float32" min_val = float("-0.848234") @@ -1074,6 +1172,7 @@ class Program_weight_tensor_parameter_97: class Program_weight_tensor_parameter_98: name = "parameter_98" + original_name = "batch_norm2d_2.w_0" shape = [80] dtype = "float32" min_val = float("0.285761") @@ -1085,6 +1184,7 @@ class Program_weight_tensor_parameter_98: class Program_weight_tensor_parameter_99: name = "parameter_99" + original_name = "batch_norm2d_2.w_2" shape = [80] dtype = "float32" min_val = float("330.977") @@ -1096,6 +1196,7 @@ class Program_weight_tensor_parameter_99: class Program_weight_tensor_parameter_100: name = "parameter_100" + original_name = "batch_norm2d_2.w_1" shape = [80] dtype = "float32" min_val = float("-64.6527") @@ -1107,6 +1208,7 @@ class Program_weight_tensor_parameter_100: class Program_weight_tensor_parameter_101: name = "parameter_101" + original_name = "conv2d_4.w_0" shape = [80, 40, 2, 2] dtype = "float32" min_val = float("-1.93591") @@ -1118,6 +1220,7 @@ class Program_weight_tensor_parameter_101: class Program_weight_tensor_parameter_102: name = "parameter_102" + original_name = "conv2d_2.w_0" shape = [40, 80, 1, 1] dtype = "float32" min_val = float("-3.33116") @@ -1129,6 +1232,7 @@ class Program_weight_tensor_parameter_102: class Program_weight_tensor_parameter_103: name = "parameter_103" + original_name = "batch_norm2d_1.b_0" shape = [80] dtype = "float32" min_val = float("-1.69928") @@ -1140,6 +1244,7 @@ class Program_weight_tensor_parameter_103: class Program_weight_tensor_parameter_104: name = "parameter_104" + original_name = "batch_norm2d_1.w_0" shape = [80] dtype = "float32" min_val = float("0.670717") @@ -1151,6 +1256,7 @@ class Program_weight_tensor_parameter_104: class Program_weight_tensor_parameter_105: name = "parameter_105" + original_name = "batch_norm2d_1.w_2" shape = [80] dtype = "float32" min_val = float("0.980977") @@ -1162,6 +1268,7 @@ class Program_weight_tensor_parameter_105: class Program_weight_tensor_parameter_106: name = "parameter_106" + original_name = "batch_norm2d_1.w_1" shape = [80] dtype = "float32" min_val = float("-8.49422") @@ -1173,6 +1280,7 @@ class Program_weight_tensor_parameter_106: class Program_weight_tensor_parameter_107: name = "parameter_107" + original_name = "conv2d_1.w_0" shape = [80, 40, 1, 1] dtype = "float32" min_val = float("-4.19382") @@ -1184,6 +1292,7 @@ class Program_weight_tensor_parameter_107: class Program_weight_tensor_parameter_108: name = "parameter_108" + original_name = "conv2d_3.w_0" shape = [10, 10, 3, 3] dtype = "float32" min_val = float("-1.70078") @@ -1195,6 +1304,7 @@ class Program_weight_tensor_parameter_108: class Program_weight_tensor_parameter_109: name = "parameter_109" + original_name = "batch_norm2d_0.b_0" shape = [40] dtype = "float32" min_val = float("0") @@ -1204,6 +1314,7 @@ class Program_weight_tensor_parameter_109: class Program_weight_tensor_parameter_110: name = "parameter_110" + original_name = "batch_norm2d_0.w_0" shape = [40] dtype = "float32" min_val = float("0") @@ -1213,6 +1324,7 @@ class Program_weight_tensor_parameter_110: class Program_weight_tensor_parameter_111: name = "parameter_111" + original_name = "batch_norm2d_0.w_2" shape = [40] dtype = "float32" min_val = float("0") @@ -1222,6 +1334,7 @@ class Program_weight_tensor_parameter_111: class Program_weight_tensor_parameter_112: name = "parameter_112" + original_name = "batch_norm2d_0.w_1" shape = [40] dtype = "float32" min_val = float("0") @@ -1231,6 +1344,7 @@ class Program_weight_tensor_parameter_112: class Program_weight_tensor_parameter_113: name = "parameter_113" + original_name = "conv2d_0.w_0" shape = [40, 3, 4, 4] dtype = "float32" min_val = float("-1.5863") diff --git a/paddle_samples/PaddleX/FasterNet-T1/subgraph_1/input_meta.py b/paddle_samples/PaddleX/FasterNet-T1/subgraph_1/input_meta.py index 9768d334d..08971911c 100644 --- a/paddle_samples/PaddleX/FasterNet-T1/subgraph_1/input_meta.py +++ b/paddle_samples/PaddleX/FasterNet-T1/subgraph_1/input_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_data_0: name = "data_0" + original_name = "var_216" shape = [124, 3, 224, 224] dtype = "float32" min_val = float("-2.1179") diff --git a/paddle_samples/PaddleX/FasterNet-T1/subgraph_1/weight_meta.py b/paddle_samples/PaddleX/FasterNet-T1/subgraph_1/weight_meta.py index 0990421b0..614f43f56 100644 --- a/paddle_samples/PaddleX/FasterNet-T1/subgraph_1/weight_meta.py +++ b/paddle_samples/PaddleX/FasterNet-T1/subgraph_1/weight_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_parameter_0: name = "parameter_0" + original_name = "linear_0.b_0" shape = [102] dtype = "float32" min_val = float("-0.00235572") @@ -11,6 +12,7 @@ class Program_weight_tensor_parameter_0: class Program_weight_tensor_parameter_1: name = "parameter_1" + original_name = "linear_0.w_0" shape = [1280, 102] dtype = "float32" min_val = float("-0.0894211") @@ -22,6 +24,7 @@ class Program_weight_tensor_parameter_1: class Program_weight_tensor_parameter_2: name = "parameter_2" + original_name = "conv2d_43.w_0" shape = [1280, 512, 1, 1] dtype = "float32" min_val = float("-1.56204") @@ -33,6 +36,7 @@ class Program_weight_tensor_parameter_2: class Program_weight_tensor_parameter_3: name = "parameter_3" + original_name = "conv2d_41.w_0" shape = [512, 1024, 1, 1] dtype = "float32" min_val = float("-1.2719") @@ -44,6 +48,7 @@ class Program_weight_tensor_parameter_3: class Program_weight_tensor_parameter_4: name = "parameter_4" + original_name = "batch_norm2d_16.b_0" shape = [1024] dtype = "float32" min_val = float("-2.86973") @@ -55,6 +60,7 @@ class Program_weight_tensor_parameter_4: class Program_weight_tensor_parameter_5: name = "parameter_5" + original_name = "batch_norm2d_16.w_0" shape = [1024] dtype = "float32" min_val = float("0.0220399") @@ -66,6 +72,7 @@ class Program_weight_tensor_parameter_5: class Program_weight_tensor_parameter_6: name = "parameter_6" + original_name = "batch_norm2d_16.w_2" shape = [1024] dtype = "float32" min_val = float("31.4645") @@ -77,6 +84,7 @@ class Program_weight_tensor_parameter_6: class Program_weight_tensor_parameter_7: name = "parameter_7" + original_name = "batch_norm2d_16.w_1" shape = [1024] dtype = "float32" min_val = float("-9.36448") @@ -88,6 +96,7 @@ class Program_weight_tensor_parameter_7: class Program_weight_tensor_parameter_8: name = "parameter_8" + original_name = "conv2d_40.w_0" shape = [1024, 512, 1, 1] dtype = "float32" min_val = float("-1.89282") @@ -99,6 +108,7 @@ class Program_weight_tensor_parameter_8: class Program_weight_tensor_parameter_9: name = "parameter_9" + original_name = "conv2d_42.w_0" shape = [128, 128, 3, 3] dtype = "float32" min_val = float("-1.41039") @@ -110,6 +120,7 @@ class Program_weight_tensor_parameter_9: class Program_weight_tensor_parameter_10: name = "parameter_10" + original_name = "conv2d_38.w_0" shape = [512, 1024, 1, 1] dtype = "float32" min_val = float("-1.1371") @@ -121,6 +132,7 @@ class Program_weight_tensor_parameter_10: class Program_weight_tensor_parameter_11: name = "parameter_11" + original_name = "batch_norm2d_15.b_0" shape = [1024] dtype = "float32" min_val = float("-2.33717") @@ -132,6 +144,7 @@ class Program_weight_tensor_parameter_11: class Program_weight_tensor_parameter_12: name = "parameter_12" + original_name = "batch_norm2d_15.w_0" shape = [1024] dtype = "float32" min_val = float("-0.00530456") @@ -143,6 +156,7 @@ class Program_weight_tensor_parameter_12: class Program_weight_tensor_parameter_13: name = "parameter_13" + original_name = "batch_norm2d_15.w_2" shape = [1024] dtype = "float32" min_val = float("23.7499") @@ -154,6 +168,7 @@ class Program_weight_tensor_parameter_13: class Program_weight_tensor_parameter_14: name = "parameter_14" + original_name = "batch_norm2d_15.w_1" shape = [1024] dtype = "float32" min_val = float("-13.1036") @@ -165,6 +180,7 @@ class Program_weight_tensor_parameter_14: class Program_weight_tensor_parameter_15: name = "parameter_15" + original_name = "conv2d_37.w_0" shape = [1024, 512, 1, 1] dtype = "float32" min_val = float("-1.72583") @@ -176,6 +192,7 @@ class Program_weight_tensor_parameter_15: class Program_weight_tensor_parameter_16: name = "parameter_16" + original_name = "conv2d_39.w_0" shape = [128, 128, 3, 3] dtype = "float32" min_val = float("-1.71472") @@ -187,6 +204,7 @@ class Program_weight_tensor_parameter_16: class Program_weight_tensor_parameter_17: name = "parameter_17" + original_name = "batch_norm2d_14.b_0" shape = [512] dtype = "float32" min_val = float("-0.446601") @@ -198,6 +216,7 @@ class Program_weight_tensor_parameter_17: class Program_weight_tensor_parameter_18: name = "parameter_18" + original_name = "batch_norm2d_14.w_0" shape = [512] dtype = "float32" min_val = float("0.0716077") @@ -209,6 +228,7 @@ class Program_weight_tensor_parameter_18: class Program_weight_tensor_parameter_19: name = "parameter_19" + original_name = "batch_norm2d_14.w_2" shape = [512] dtype = "float32" min_val = float("763.939") @@ -220,6 +240,7 @@ class Program_weight_tensor_parameter_19: class Program_weight_tensor_parameter_20: name = "parameter_20" + original_name = "batch_norm2d_14.w_1" shape = [512] dtype = "float32" min_val = float("-58.5679") @@ -231,6 +252,7 @@ class Program_weight_tensor_parameter_20: class Program_weight_tensor_parameter_21: name = "parameter_21" + original_name = "conv2d_36.w_0" shape = [512, 256, 2, 2] dtype = "float32" min_val = float("-1.71319") @@ -242,6 +264,7 @@ class Program_weight_tensor_parameter_21: class Program_weight_tensor_parameter_22: name = "parameter_22" + original_name = "conv2d_34.w_0" shape = [256, 512, 1, 1] dtype = "float32" min_val = float("-2.26721") @@ -253,6 +276,7 @@ class Program_weight_tensor_parameter_22: class Program_weight_tensor_parameter_23: name = "parameter_23" + original_name = "batch_norm2d_13.b_0" shape = [512] dtype = "float32" min_val = float("-1.77091") @@ -264,6 +288,7 @@ class Program_weight_tensor_parameter_23: class Program_weight_tensor_parameter_24: name = "parameter_24" + original_name = "batch_norm2d_13.w_0" shape = [512] dtype = "float32" min_val = float("0.325668") @@ -275,6 +300,7 @@ class Program_weight_tensor_parameter_24: class Program_weight_tensor_parameter_25: name = "parameter_25" + original_name = "batch_norm2d_13.w_2" shape = [512] dtype = "float32" min_val = float("140.024") @@ -286,6 +312,7 @@ class Program_weight_tensor_parameter_25: class Program_weight_tensor_parameter_26: name = "parameter_26" + original_name = "batch_norm2d_13.w_1" shape = [512] dtype = "float32" min_val = float("-41.7132") @@ -297,6 +324,7 @@ class Program_weight_tensor_parameter_26: class Program_weight_tensor_parameter_27: name = "parameter_27" + original_name = "conv2d_33.w_0" shape = [512, 256, 1, 1] dtype = "float32" min_val = float("-1.64179") @@ -308,6 +336,7 @@ class Program_weight_tensor_parameter_27: class Program_weight_tensor_parameter_28: name = "parameter_28" + original_name = "conv2d_35.w_0" shape = [64, 64, 3, 3] dtype = "float32" min_val = float("-0.920591") @@ -319,6 +348,7 @@ class Program_weight_tensor_parameter_28: class Program_weight_tensor_parameter_29: name = "parameter_29" + original_name = "conv2d_31.w_0" shape = [256, 512, 1, 1] dtype = "float32" min_val = float("-3.10395") @@ -330,6 +360,7 @@ class Program_weight_tensor_parameter_29: class Program_weight_tensor_parameter_30: name = "parameter_30" + original_name = "batch_norm2d_12.b_0" shape = [512] dtype = "float32" min_val = float("-2.04568") @@ -341,6 +372,7 @@ class Program_weight_tensor_parameter_30: class Program_weight_tensor_parameter_31: name = "parameter_31" + original_name = "batch_norm2d_12.w_0" shape = [512] dtype = "float32" min_val = float("0.279229") @@ -352,6 +384,7 @@ class Program_weight_tensor_parameter_31: class Program_weight_tensor_parameter_32: name = "parameter_32" + original_name = "batch_norm2d_12.w_2" shape = [512] dtype = "float32" min_val = float("83.7205") @@ -363,6 +396,7 @@ class Program_weight_tensor_parameter_32: class Program_weight_tensor_parameter_33: name = "parameter_33" + original_name = "batch_norm2d_12.w_1" shape = [512] dtype = "float32" min_val = float("-48.885") @@ -374,6 +408,7 @@ class Program_weight_tensor_parameter_33: class Program_weight_tensor_parameter_34: name = "parameter_34" + original_name = "conv2d_30.w_0" shape = [512, 256, 1, 1] dtype = "float32" min_val = float("-1.69473") @@ -385,6 +420,7 @@ class Program_weight_tensor_parameter_34: class Program_weight_tensor_parameter_35: name = "parameter_35" + original_name = "conv2d_32.w_0" shape = [64, 64, 3, 3] dtype = "float32" min_val = float("-0.819536") @@ -396,6 +432,7 @@ class Program_weight_tensor_parameter_35: class Program_weight_tensor_parameter_36: name = "parameter_36" + original_name = "conv2d_28.w_0" shape = [256, 512, 1, 1] dtype = "float32" min_val = float("-1.95301") @@ -407,6 +444,7 @@ class Program_weight_tensor_parameter_36: class Program_weight_tensor_parameter_37: name = "parameter_37" + original_name = "batch_norm2d_11.b_0" shape = [512] dtype = "float32" min_val = float("-1.75779") @@ -418,6 +456,7 @@ class Program_weight_tensor_parameter_37: class Program_weight_tensor_parameter_38: name = "parameter_38" + original_name = "batch_norm2d_11.w_0" shape = [512] dtype = "float32" min_val = float("0.16471") @@ -429,6 +468,7 @@ class Program_weight_tensor_parameter_38: class Program_weight_tensor_parameter_39: name = "parameter_39" + original_name = "batch_norm2d_11.w_2" shape = [512] dtype = "float32" min_val = float("85.7683") @@ -440,6 +480,7 @@ class Program_weight_tensor_parameter_39: class Program_weight_tensor_parameter_40: name = "parameter_40" + original_name = "batch_norm2d_11.w_1" shape = [512] dtype = "float32" min_val = float("-42.9588") @@ -451,6 +492,7 @@ class Program_weight_tensor_parameter_40: class Program_weight_tensor_parameter_41: name = "parameter_41" + original_name = "conv2d_27.w_0" shape = [512, 256, 1, 1] dtype = "float32" min_val = float("-1.70913") @@ -462,6 +504,7 @@ class Program_weight_tensor_parameter_41: class Program_weight_tensor_parameter_42: name = "parameter_42" + original_name = "conv2d_29.w_0" shape = [64, 64, 3, 3] dtype = "float32" min_val = float("-1.02977") @@ -473,6 +516,7 @@ class Program_weight_tensor_parameter_42: class Program_weight_tensor_parameter_43: name = "parameter_43" + original_name = "conv2d_25.w_0" shape = [256, 512, 1, 1] dtype = "float32" min_val = float("-2.22364") @@ -484,6 +528,7 @@ class Program_weight_tensor_parameter_43: class Program_weight_tensor_parameter_44: name = "parameter_44" + original_name = "batch_norm2d_10.b_0" shape = [512] dtype = "float32" min_val = float("-1.51903") @@ -495,6 +540,7 @@ class Program_weight_tensor_parameter_44: class Program_weight_tensor_parameter_45: name = "parameter_45" + original_name = "batch_norm2d_10.w_0" shape = [512] dtype = "float32" min_val = float("0.140214") @@ -506,6 +552,7 @@ class Program_weight_tensor_parameter_45: class Program_weight_tensor_parameter_46: name = "parameter_46" + original_name = "batch_norm2d_10.w_2" shape = [512] dtype = "float32" min_val = float("48.9873") @@ -517,6 +564,7 @@ class Program_weight_tensor_parameter_46: class Program_weight_tensor_parameter_47: name = "parameter_47" + original_name = "batch_norm2d_10.w_1" shape = [512] dtype = "float32" min_val = float("-41.9956") @@ -528,6 +576,7 @@ class Program_weight_tensor_parameter_47: class Program_weight_tensor_parameter_48: name = "parameter_48" + original_name = "conv2d_24.w_0" shape = [512, 256, 1, 1] dtype = "float32" min_val = float("-1.7411") @@ -539,6 +588,7 @@ class Program_weight_tensor_parameter_48: class Program_weight_tensor_parameter_49: name = "parameter_49" + original_name = "conv2d_26.w_0" shape = [64, 64, 3, 3] dtype = "float32" min_val = float("-1.16125") @@ -550,6 +600,7 @@ class Program_weight_tensor_parameter_49: class Program_weight_tensor_parameter_50: name = "parameter_50" + original_name = "conv2d_22.w_0" shape = [256, 512, 1, 1] dtype = "float32" min_val = float("-2.09195") @@ -561,6 +612,7 @@ class Program_weight_tensor_parameter_50: class Program_weight_tensor_parameter_51: name = "parameter_51" + original_name = "batch_norm2d_9.b_0" shape = [512] dtype = "float32" min_val = float("-1.87036") @@ -572,6 +624,7 @@ class Program_weight_tensor_parameter_51: class Program_weight_tensor_parameter_52: name = "parameter_52" + original_name = "batch_norm2d_9.w_0" shape = [512] dtype = "float32" min_val = float("0.146153") @@ -583,6 +636,7 @@ class Program_weight_tensor_parameter_52: class Program_weight_tensor_parameter_53: name = "parameter_53" + original_name = "batch_norm2d_9.w_2" shape = [512] dtype = "float32" min_val = float("18.3063") @@ -594,6 +648,7 @@ class Program_weight_tensor_parameter_53: class Program_weight_tensor_parameter_54: name = "parameter_54" + original_name = "batch_norm2d_9.w_1" shape = [512] dtype = "float32" min_val = float("-38.1848") @@ -605,6 +660,7 @@ class Program_weight_tensor_parameter_54: class Program_weight_tensor_parameter_55: name = "parameter_55" + original_name = "conv2d_21.w_0" shape = [512, 256, 1, 1] dtype = "float32" min_val = float("-1.58366") @@ -616,6 +672,7 @@ class Program_weight_tensor_parameter_55: class Program_weight_tensor_parameter_56: name = "parameter_56" + original_name = "conv2d_23.w_0" shape = [64, 64, 3, 3] dtype = "float32" min_val = float("-1.1544") @@ -627,6 +684,7 @@ class Program_weight_tensor_parameter_56: class Program_weight_tensor_parameter_57: name = "parameter_57" + original_name = "conv2d_19.w_0" shape = [256, 512, 1, 1] dtype = "float32" min_val = float("-2.17671") @@ -638,6 +696,7 @@ class Program_weight_tensor_parameter_57: class Program_weight_tensor_parameter_58: name = "parameter_58" + original_name = "batch_norm2d_8.b_0" shape = [512] dtype = "float32" min_val = float("-1.81219") @@ -649,6 +708,7 @@ class Program_weight_tensor_parameter_58: class Program_weight_tensor_parameter_59: name = "parameter_59" + original_name = "batch_norm2d_8.w_0" shape = [512] dtype = "float32" min_val = float("0.135106") @@ -660,6 +720,7 @@ class Program_weight_tensor_parameter_59: class Program_weight_tensor_parameter_60: name = "parameter_60" + original_name = "batch_norm2d_8.w_2" shape = [512] dtype = "float32" min_val = float("12.0597") @@ -671,6 +732,7 @@ class Program_weight_tensor_parameter_60: class Program_weight_tensor_parameter_61: name = "parameter_61" + original_name = "batch_norm2d_8.w_1" shape = [512] dtype = "float32" min_val = float("-31.5479") @@ -682,6 +744,7 @@ class Program_weight_tensor_parameter_61: class Program_weight_tensor_parameter_62: name = "parameter_62" + original_name = "conv2d_18.w_0" shape = [512, 256, 1, 1] dtype = "float32" min_val = float("-1.77038") @@ -693,6 +756,7 @@ class Program_weight_tensor_parameter_62: class Program_weight_tensor_parameter_63: name = "parameter_63" + original_name = "conv2d_20.w_0" shape = [64, 64, 3, 3] dtype = "float32" min_val = float("-1.1654") @@ -704,6 +768,7 @@ class Program_weight_tensor_parameter_63: class Program_weight_tensor_parameter_64: name = "parameter_64" + original_name = "conv2d_16.w_0" shape = [256, 512, 1, 1] dtype = "float32" min_val = float("-2.63217") @@ -715,6 +780,7 @@ class Program_weight_tensor_parameter_64: class Program_weight_tensor_parameter_65: name = "parameter_65" + original_name = "batch_norm2d_7.b_0" shape = [512] dtype = "float32" min_val = float("-1.58512") @@ -726,6 +792,7 @@ class Program_weight_tensor_parameter_65: class Program_weight_tensor_parameter_66: name = "parameter_66" + original_name = "batch_norm2d_7.w_0" shape = [512] dtype = "float32" min_val = float("0.100669") @@ -737,6 +804,7 @@ class Program_weight_tensor_parameter_66: class Program_weight_tensor_parameter_67: name = "parameter_67" + original_name = "batch_norm2d_7.w_2" shape = [512] dtype = "float32" min_val = float("6.08017") @@ -748,6 +816,7 @@ class Program_weight_tensor_parameter_67: class Program_weight_tensor_parameter_68: name = "parameter_68" + original_name = "batch_norm2d_7.w_1" shape = [512] dtype = "float32" min_val = float("-26.5748") @@ -759,6 +828,7 @@ class Program_weight_tensor_parameter_68: class Program_weight_tensor_parameter_69: name = "parameter_69" + original_name = "conv2d_15.w_0" shape = [512, 256, 1, 1] dtype = "float32" min_val = float("-1.63299") @@ -770,6 +840,7 @@ class Program_weight_tensor_parameter_69: class Program_weight_tensor_parameter_70: name = "parameter_70" + original_name = "conv2d_17.w_0" shape = [64, 64, 3, 3] dtype = "float32" min_val = float("-1.36254") @@ -781,6 +852,7 @@ class Program_weight_tensor_parameter_70: class Program_weight_tensor_parameter_71: name = "parameter_71" + original_name = "conv2d_13.w_0" shape = [256, 512, 1, 1] dtype = "float32" min_val = float("-2.57655") @@ -792,6 +864,7 @@ class Program_weight_tensor_parameter_71: class Program_weight_tensor_parameter_72: name = "parameter_72" + original_name = "batch_norm2d_6.b_0" shape = [512] dtype = "float32" min_val = float("-1.21342") @@ -803,6 +876,7 @@ class Program_weight_tensor_parameter_72: class Program_weight_tensor_parameter_73: name = "parameter_73" + original_name = "batch_norm2d_6.w_0" shape = [512] dtype = "float32" min_val = float("0.0985472") @@ -814,6 +888,7 @@ class Program_weight_tensor_parameter_73: class Program_weight_tensor_parameter_74: name = "parameter_74" + original_name = "batch_norm2d_6.w_2" shape = [512] dtype = "float32" min_val = float("6.34811") @@ -825,6 +900,7 @@ class Program_weight_tensor_parameter_74: class Program_weight_tensor_parameter_75: name = "parameter_75" + original_name = "batch_norm2d_6.w_1" shape = [512] dtype = "float32" min_val = float("-42.7247") @@ -836,6 +912,7 @@ class Program_weight_tensor_parameter_75: class Program_weight_tensor_parameter_76: name = "parameter_76" + original_name = "conv2d_12.w_0" shape = [512, 256, 1, 1] dtype = "float32" min_val = float("-2.46312") @@ -847,6 +924,7 @@ class Program_weight_tensor_parameter_76: class Program_weight_tensor_parameter_77: name = "parameter_77" + original_name = "conv2d_14.w_0" shape = [64, 64, 3, 3] dtype = "float32" min_val = float("-1.46055") @@ -858,6 +936,7 @@ class Program_weight_tensor_parameter_77: class Program_weight_tensor_parameter_78: name = "parameter_78" + original_name = "batch_norm2d_5.b_0" shape = [256] dtype = "float32" min_val = float("-0.739149") @@ -869,6 +948,7 @@ class Program_weight_tensor_parameter_78: class Program_weight_tensor_parameter_79: name = "parameter_79" + original_name = "batch_norm2d_5.w_0" shape = [256] dtype = "float32" min_val = float("0.0959448") @@ -880,6 +960,7 @@ class Program_weight_tensor_parameter_79: class Program_weight_tensor_parameter_80: name = "parameter_80" + original_name = "batch_norm2d_5.w_2" shape = [256] dtype = "float32" min_val = float("167.913") @@ -891,6 +972,7 @@ class Program_weight_tensor_parameter_80: class Program_weight_tensor_parameter_81: name = "parameter_81" + original_name = "batch_norm2d_5.w_1" shape = [256] dtype = "float32" min_val = float("-23.1004") @@ -902,6 +984,7 @@ class Program_weight_tensor_parameter_81: class Program_weight_tensor_parameter_82: name = "parameter_82" + original_name = "conv2d_11.w_0" shape = [256, 128, 2, 2] dtype = "float32" min_val = float("-1.63006") @@ -913,6 +996,7 @@ class Program_weight_tensor_parameter_82: class Program_weight_tensor_parameter_83: name = "parameter_83" + original_name = "conv2d_9.w_0" shape = [128, 256, 1, 1] dtype = "float32" min_val = float("-1.85891") @@ -924,6 +1008,7 @@ class Program_weight_tensor_parameter_83: class Program_weight_tensor_parameter_84: name = "parameter_84" + original_name = "batch_norm2d_4.b_0" shape = [256] dtype = "float32" min_val = float("-1.40524") @@ -935,6 +1020,7 @@ class Program_weight_tensor_parameter_84: class Program_weight_tensor_parameter_85: name = "parameter_85" + original_name = "batch_norm2d_4.w_0" shape = [256] dtype = "float32" min_val = float("0.365038") @@ -946,6 +1032,7 @@ class Program_weight_tensor_parameter_85: class Program_weight_tensor_parameter_86: name = "parameter_86" + original_name = "batch_norm2d_4.w_2" shape = [256] dtype = "float32" min_val = float("8.50652") @@ -957,6 +1044,7 @@ class Program_weight_tensor_parameter_86: class Program_weight_tensor_parameter_87: name = "parameter_87" + original_name = "batch_norm2d_4.w_1" shape = [256] dtype = "float32" min_val = float("-4.521") @@ -968,6 +1056,7 @@ class Program_weight_tensor_parameter_87: class Program_weight_tensor_parameter_88: name = "parameter_88" + original_name = "conv2d_8.w_0" shape = [256, 128, 1, 1] dtype = "float32" min_val = float("-1.55295") @@ -979,6 +1068,7 @@ class Program_weight_tensor_parameter_88: class Program_weight_tensor_parameter_89: name = "parameter_89" + original_name = "conv2d_10.w_0" shape = [32, 32, 3, 3] dtype = "float32" min_val = float("-1.26079") @@ -990,6 +1080,7 @@ class Program_weight_tensor_parameter_89: class Program_weight_tensor_parameter_90: name = "parameter_90" + original_name = "conv2d_6.w_0" shape = [128, 256, 1, 1] dtype = "float32" min_val = float("-1.76604") @@ -1001,6 +1092,7 @@ class Program_weight_tensor_parameter_90: class Program_weight_tensor_parameter_91: name = "parameter_91" + original_name = "batch_norm2d_3.b_0" shape = [256] dtype = "float32" min_val = float("-1.65319") @@ -1012,6 +1104,7 @@ class Program_weight_tensor_parameter_91: class Program_weight_tensor_parameter_92: name = "parameter_92" + original_name = "batch_norm2d_3.w_0" shape = [256] dtype = "float32" min_val = float("0.261292") @@ -1023,6 +1116,7 @@ class Program_weight_tensor_parameter_92: class Program_weight_tensor_parameter_93: name = "parameter_93" + original_name = "batch_norm2d_3.w_2" shape = [256] dtype = "float32" min_val = float("6.40282") @@ -1034,6 +1128,7 @@ class Program_weight_tensor_parameter_93: class Program_weight_tensor_parameter_94: name = "parameter_94" + original_name = "batch_norm2d_3.w_1" shape = [256] dtype = "float32" min_val = float("-9.70504") @@ -1045,6 +1140,7 @@ class Program_weight_tensor_parameter_94: class Program_weight_tensor_parameter_95: name = "parameter_95" + original_name = "conv2d_5.w_0" shape = [256, 128, 1, 1] dtype = "float32" min_val = float("-1.66078") @@ -1056,6 +1152,7 @@ class Program_weight_tensor_parameter_95: class Program_weight_tensor_parameter_96: name = "parameter_96" + original_name = "conv2d_7.w_0" shape = [32, 32, 3, 3] dtype = "float32" min_val = float("-1.73857") @@ -1067,6 +1164,7 @@ class Program_weight_tensor_parameter_96: class Program_weight_tensor_parameter_97: name = "parameter_97" + original_name = "batch_norm2d_2.b_0" shape = [128] dtype = "float32" min_val = float("-0.74219") @@ -1078,6 +1176,7 @@ class Program_weight_tensor_parameter_97: class Program_weight_tensor_parameter_98: name = "parameter_98" + original_name = "batch_norm2d_2.w_0" shape = [128] dtype = "float32" min_val = float("0.175491") @@ -1089,6 +1188,7 @@ class Program_weight_tensor_parameter_98: class Program_weight_tensor_parameter_99: name = "parameter_99" + original_name = "batch_norm2d_2.w_2" shape = [128] dtype = "float32" min_val = float("208.091") @@ -1100,6 +1200,7 @@ class Program_weight_tensor_parameter_99: class Program_weight_tensor_parameter_100: name = "parameter_100" + original_name = "batch_norm2d_2.w_1" shape = [128] dtype = "float32" min_val = float("-46.5844") @@ -1111,6 +1212,7 @@ class Program_weight_tensor_parameter_100: class Program_weight_tensor_parameter_101: name = "parameter_101" + original_name = "conv2d_4.w_0" shape = [128, 64, 2, 2] dtype = "float32" min_val = float("-1.51232") @@ -1122,6 +1224,7 @@ class Program_weight_tensor_parameter_101: class Program_weight_tensor_parameter_102: name = "parameter_102" + original_name = "conv2d_2.w_0" shape = [64, 128, 1, 1] dtype = "float32" min_val = float("-2.46293") @@ -1133,6 +1236,7 @@ class Program_weight_tensor_parameter_102: class Program_weight_tensor_parameter_103: name = "parameter_103" + original_name = "batch_norm2d_1.b_0" shape = [128] dtype = "float32" min_val = float("-1.47638") @@ -1144,6 +1248,7 @@ class Program_weight_tensor_parameter_103: class Program_weight_tensor_parameter_104: name = "parameter_104" + original_name = "batch_norm2d_1.w_0" shape = [128] dtype = "float32" min_val = float("0.468111") @@ -1155,6 +1260,7 @@ class Program_weight_tensor_parameter_104: class Program_weight_tensor_parameter_105: name = "parameter_105" + original_name = "batch_norm2d_1.w_2" shape = [128] dtype = "float32" min_val = float("0.186558") @@ -1166,6 +1272,7 @@ class Program_weight_tensor_parameter_105: class Program_weight_tensor_parameter_106: name = "parameter_106" + original_name = "batch_norm2d_1.w_1" shape = [128] dtype = "float32" min_val = float("-2.37611") @@ -1177,6 +1284,7 @@ class Program_weight_tensor_parameter_106: class Program_weight_tensor_parameter_107: name = "parameter_107" + original_name = "conv2d_1.w_0" shape = [128, 64, 1, 1] dtype = "float32" min_val = float("-2.2026") @@ -1188,6 +1296,7 @@ class Program_weight_tensor_parameter_107: class Program_weight_tensor_parameter_108: name = "parameter_108" + original_name = "conv2d_3.w_0" shape = [16, 16, 3, 3] dtype = "float32" min_val = float("-1.5412") @@ -1199,6 +1308,7 @@ class Program_weight_tensor_parameter_108: class Program_weight_tensor_parameter_109: name = "parameter_109" + original_name = "batch_norm2d_0.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -1208,6 +1318,7 @@ class Program_weight_tensor_parameter_109: class Program_weight_tensor_parameter_110: name = "parameter_110" + original_name = "batch_norm2d_0.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -1217,6 +1328,7 @@ class Program_weight_tensor_parameter_110: class Program_weight_tensor_parameter_111: name = "parameter_111" + original_name = "batch_norm2d_0.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -1226,6 +1338,7 @@ class Program_weight_tensor_parameter_111: class Program_weight_tensor_parameter_112: name = "parameter_112" + original_name = "batch_norm2d_0.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -1235,6 +1348,7 @@ class Program_weight_tensor_parameter_112: class Program_weight_tensor_parameter_113: name = "parameter_113" + original_name = "conv2d_0.w_0" shape = [64, 3, 4, 4] dtype = "float32" min_val = float("-1.0045") diff --git a/paddle_samples/PaddleX/FasterNet-T1/subgraph_2/input_meta.py b/paddle_samples/PaddleX/FasterNet-T1/subgraph_2/input_meta.py index 302f27845..34ccf9d70 100644 --- a/paddle_samples/PaddleX/FasterNet-T1/subgraph_2/input_meta.py +++ b/paddle_samples/PaddleX/FasterNet-T1/subgraph_2/input_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_data_0: name = "data_0" + original_name = "var_0" shape = [128, 3, 224, 224] dtype = "float32" min_val = float("-2.1179") diff --git a/paddle_samples/PaddleX/FasterNet-T1/subgraph_2/weight_meta.py b/paddle_samples/PaddleX/FasterNet-T1/subgraph_2/weight_meta.py index 51b100e10..c6961a945 100644 --- a/paddle_samples/PaddleX/FasterNet-T1/subgraph_2/weight_meta.py +++ b/paddle_samples/PaddleX/FasterNet-T1/subgraph_2/weight_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_parameter_0: name = "parameter_0" + original_name = "linear_0.b_0" shape = [102] dtype = "float32" data = None @@ -7,6 +8,7 @@ class Program_weight_tensor_parameter_0: class Program_weight_tensor_parameter_1: name = "parameter_1" + original_name = "linear_0.w_0" shape = [1280, 102] dtype = "float32" min_val = float("-0.0861183") @@ -18,6 +20,7 @@ class Program_weight_tensor_parameter_1: class Program_weight_tensor_parameter_2: name = "parameter_2" + original_name = "conv2d_43.w_0" shape = [1280, 512, 1, 1] dtype = "float32" min_val = float("-1.5621") @@ -29,6 +32,7 @@ class Program_weight_tensor_parameter_2: class Program_weight_tensor_parameter_3: name = "parameter_3" + original_name = "conv2d_41.w_0" shape = [512, 1024, 1, 1] dtype = "float32" min_val = float("-1.26988") @@ -40,6 +44,7 @@ class Program_weight_tensor_parameter_3: class Program_weight_tensor_parameter_4: name = "parameter_4" + original_name = "batch_norm2d_16.b_0" shape = [1024] dtype = "float32" min_val = float("-2.87022") @@ -51,6 +56,7 @@ class Program_weight_tensor_parameter_4: class Program_weight_tensor_parameter_5: name = "parameter_5" + original_name = "batch_norm2d_16.w_0" shape = [1024] dtype = "float32" min_val = float("0.0225535") @@ -62,6 +68,7 @@ class Program_weight_tensor_parameter_5: class Program_weight_tensor_parameter_6: name = "parameter_6" + original_name = "batch_norm2d_16.w_2" shape = [1024] dtype = "float32" min_val = float("31.9152") @@ -73,6 +80,7 @@ class Program_weight_tensor_parameter_6: class Program_weight_tensor_parameter_7: name = "parameter_7" + original_name = "batch_norm2d_16.w_1" shape = [1024] dtype = "float32" min_val = float("-9.85412") @@ -84,6 +92,7 @@ class Program_weight_tensor_parameter_7: class Program_weight_tensor_parameter_8: name = "parameter_8" + original_name = "conv2d_40.w_0" shape = [1024, 512, 1, 1] dtype = "float32" min_val = float("-1.89407") @@ -95,6 +104,7 @@ class Program_weight_tensor_parameter_8: class Program_weight_tensor_parameter_9: name = "parameter_9" + original_name = "conv2d_42.w_0" shape = [128, 128, 3, 3] dtype = "float32" min_val = float("-1.4125") @@ -106,6 +116,7 @@ class Program_weight_tensor_parameter_9: class Program_weight_tensor_parameter_10: name = "parameter_10" + original_name = "conv2d_38.w_0" shape = [512, 1024, 1, 1] dtype = "float32" min_val = float("-1.13787") @@ -117,6 +128,7 @@ class Program_weight_tensor_parameter_10: class Program_weight_tensor_parameter_11: name = "parameter_11" + original_name = "batch_norm2d_15.b_0" shape = [1024] dtype = "float32" min_val = float("-2.33591") @@ -128,6 +140,7 @@ class Program_weight_tensor_parameter_11: class Program_weight_tensor_parameter_12: name = "parameter_12" + original_name = "batch_norm2d_15.w_0" shape = [1024] dtype = "float32" min_val = float("-0.00504681") @@ -139,6 +152,7 @@ class Program_weight_tensor_parameter_12: class Program_weight_tensor_parameter_13: name = "parameter_13" + original_name = "batch_norm2d_15.w_2" shape = [1024] dtype = "float32" min_val = float("23.0733") @@ -150,6 +164,7 @@ class Program_weight_tensor_parameter_13: class Program_weight_tensor_parameter_14: name = "parameter_14" + original_name = "batch_norm2d_15.w_1" shape = [1024] dtype = "float32" min_val = float("-13.15") @@ -161,6 +176,7 @@ class Program_weight_tensor_parameter_14: class Program_weight_tensor_parameter_15: name = "parameter_15" + original_name = "conv2d_37.w_0" shape = [1024, 512, 1, 1] dtype = "float32" min_val = float("-1.72533") @@ -172,6 +188,7 @@ class Program_weight_tensor_parameter_15: class Program_weight_tensor_parameter_16: name = "parameter_16" + original_name = "conv2d_39.w_0" shape = [128, 128, 3, 3] dtype = "float32" min_val = float("-1.71417") @@ -183,6 +200,7 @@ class Program_weight_tensor_parameter_16: class Program_weight_tensor_parameter_17: name = "parameter_17" + original_name = "batch_norm2d_14.b_0" shape = [512] dtype = "float32" min_val = float("-0.445128") @@ -194,6 +212,7 @@ class Program_weight_tensor_parameter_17: class Program_weight_tensor_parameter_18: name = "parameter_18" + original_name = "batch_norm2d_14.w_0" shape = [512] dtype = "float32" min_val = float("0.0733845") @@ -205,6 +224,7 @@ class Program_weight_tensor_parameter_18: class Program_weight_tensor_parameter_19: name = "parameter_19" + original_name = "batch_norm2d_14.w_2" shape = [512] dtype = "float32" min_val = float("824.163") @@ -216,6 +236,7 @@ class Program_weight_tensor_parameter_19: class Program_weight_tensor_parameter_20: name = "parameter_20" + original_name = "batch_norm2d_14.w_1" shape = [512] dtype = "float32" min_val = float("-58.8661") @@ -227,6 +248,7 @@ class Program_weight_tensor_parameter_20: class Program_weight_tensor_parameter_21: name = "parameter_21" + original_name = "conv2d_36.w_0" shape = [512, 256, 2, 2] dtype = "float32" min_val = float("-1.71114") @@ -238,6 +260,7 @@ class Program_weight_tensor_parameter_21: class Program_weight_tensor_parameter_22: name = "parameter_22" + original_name = "conv2d_34.w_0" shape = [256, 512, 1, 1] dtype = "float32" min_val = float("-2.27018") @@ -249,6 +272,7 @@ class Program_weight_tensor_parameter_22: class Program_weight_tensor_parameter_23: name = "parameter_23" + original_name = "batch_norm2d_13.b_0" shape = [512] dtype = "float32" min_val = float("-1.77216") @@ -260,6 +284,7 @@ class Program_weight_tensor_parameter_23: class Program_weight_tensor_parameter_24: name = "parameter_24" + original_name = "batch_norm2d_13.w_0" shape = [512] dtype = "float32" min_val = float("0.328603") @@ -271,6 +296,7 @@ class Program_weight_tensor_parameter_24: class Program_weight_tensor_parameter_25: name = "parameter_25" + original_name = "batch_norm2d_13.w_2" shape = [512] dtype = "float32" min_val = float("114.987") @@ -282,6 +308,7 @@ class Program_weight_tensor_parameter_25: class Program_weight_tensor_parameter_26: name = "parameter_26" + original_name = "batch_norm2d_13.w_1" shape = [512] dtype = "float32" min_val = float("-42.5944") @@ -293,6 +320,7 @@ class Program_weight_tensor_parameter_26: class Program_weight_tensor_parameter_27: name = "parameter_27" + original_name = "conv2d_33.w_0" shape = [512, 256, 1, 1] dtype = "float32" min_val = float("-1.64082") @@ -304,6 +332,7 @@ class Program_weight_tensor_parameter_27: class Program_weight_tensor_parameter_28: name = "parameter_28" + original_name = "conv2d_35.w_0" shape = [64, 64, 3, 3] dtype = "float32" min_val = float("-0.920203") @@ -315,6 +344,7 @@ class Program_weight_tensor_parameter_28: class Program_weight_tensor_parameter_29: name = "parameter_29" + original_name = "conv2d_31.w_0" shape = [256, 512, 1, 1] dtype = "float32" min_val = float("-3.10505") @@ -326,6 +356,7 @@ class Program_weight_tensor_parameter_29: class Program_weight_tensor_parameter_30: name = "parameter_30" + original_name = "batch_norm2d_12.b_0" shape = [512] dtype = "float32" min_val = float("-2.04782") @@ -337,6 +368,7 @@ class Program_weight_tensor_parameter_30: class Program_weight_tensor_parameter_31: name = "parameter_31" + original_name = "batch_norm2d_12.w_0" shape = [512] dtype = "float32" min_val = float("0.279325") @@ -348,6 +380,7 @@ class Program_weight_tensor_parameter_31: class Program_weight_tensor_parameter_32: name = "parameter_32" + original_name = "batch_norm2d_12.w_2" shape = [512] dtype = "float32" min_val = float("91.8719") @@ -359,6 +392,7 @@ class Program_weight_tensor_parameter_32: class Program_weight_tensor_parameter_33: name = "parameter_33" + original_name = "batch_norm2d_12.w_1" shape = [512] dtype = "float32" min_val = float("-48.0355") @@ -370,6 +404,7 @@ class Program_weight_tensor_parameter_33: class Program_weight_tensor_parameter_34: name = "parameter_34" + original_name = "conv2d_30.w_0" shape = [512, 256, 1, 1] dtype = "float32" min_val = float("-1.69221") @@ -381,6 +416,7 @@ class Program_weight_tensor_parameter_34: class Program_weight_tensor_parameter_35: name = "parameter_35" + original_name = "conv2d_32.w_0" shape = [64, 64, 3, 3] dtype = "float32" min_val = float("-0.818272") @@ -392,6 +428,7 @@ class Program_weight_tensor_parameter_35: class Program_weight_tensor_parameter_36: name = "parameter_36" + original_name = "conv2d_28.w_0" shape = [256, 512, 1, 1] dtype = "float32" min_val = float("-1.95144") @@ -403,6 +440,7 @@ class Program_weight_tensor_parameter_36: class Program_weight_tensor_parameter_37: name = "parameter_37" + original_name = "batch_norm2d_11.b_0" shape = [512] dtype = "float32" min_val = float("-1.7583") @@ -414,6 +452,7 @@ class Program_weight_tensor_parameter_37: class Program_weight_tensor_parameter_38: name = "parameter_38" + original_name = "batch_norm2d_11.w_0" shape = [512] dtype = "float32" min_val = float("0.163608") @@ -425,6 +464,7 @@ class Program_weight_tensor_parameter_38: class Program_weight_tensor_parameter_39: name = "parameter_39" + original_name = "batch_norm2d_11.w_2" shape = [512] dtype = "float32" min_val = float("82.7929") @@ -436,6 +476,7 @@ class Program_weight_tensor_parameter_39: class Program_weight_tensor_parameter_40: name = "parameter_40" + original_name = "batch_norm2d_11.w_1" shape = [512] dtype = "float32" min_val = float("-41.8256") @@ -447,6 +488,7 @@ class Program_weight_tensor_parameter_40: class Program_weight_tensor_parameter_41: name = "parameter_41" + original_name = "conv2d_27.w_0" shape = [512, 256, 1, 1] dtype = "float32" min_val = float("-1.70647") @@ -458,6 +500,7 @@ class Program_weight_tensor_parameter_41: class Program_weight_tensor_parameter_42: name = "parameter_42" + original_name = "conv2d_29.w_0" shape = [64, 64, 3, 3] dtype = "float32" min_val = float("-1.02948") @@ -469,6 +512,7 @@ class Program_weight_tensor_parameter_42: class Program_weight_tensor_parameter_43: name = "parameter_43" + original_name = "conv2d_25.w_0" shape = [256, 512, 1, 1] dtype = "float32" min_val = float("-2.22119") @@ -480,6 +524,7 @@ class Program_weight_tensor_parameter_43: class Program_weight_tensor_parameter_44: name = "parameter_44" + original_name = "batch_norm2d_10.b_0" shape = [512] dtype = "float32" min_val = float("-1.51948") @@ -491,6 +536,7 @@ class Program_weight_tensor_parameter_44: class Program_weight_tensor_parameter_45: name = "parameter_45" + original_name = "batch_norm2d_10.w_0" shape = [512] dtype = "float32" min_val = float("0.141094") @@ -502,6 +548,7 @@ class Program_weight_tensor_parameter_45: class Program_weight_tensor_parameter_46: name = "parameter_46" + original_name = "batch_norm2d_10.w_2" shape = [512] dtype = "float32" min_val = float("52.3193") @@ -513,6 +560,7 @@ class Program_weight_tensor_parameter_46: class Program_weight_tensor_parameter_47: name = "parameter_47" + original_name = "batch_norm2d_10.w_1" shape = [512] dtype = "float32" min_val = float("-40.4954") @@ -524,6 +572,7 @@ class Program_weight_tensor_parameter_47: class Program_weight_tensor_parameter_48: name = "parameter_48" + original_name = "conv2d_24.w_0" shape = [512, 256, 1, 1] dtype = "float32" min_val = float("-1.74061") @@ -535,6 +584,7 @@ class Program_weight_tensor_parameter_48: class Program_weight_tensor_parameter_49: name = "parameter_49" + original_name = "conv2d_26.w_0" shape = [64, 64, 3, 3] dtype = "float32" min_val = float("-1.15947") @@ -546,6 +596,7 @@ class Program_weight_tensor_parameter_49: class Program_weight_tensor_parameter_50: name = "parameter_50" + original_name = "conv2d_22.w_0" shape = [256, 512, 1, 1] dtype = "float32" min_val = float("-2.09294") @@ -557,6 +608,7 @@ class Program_weight_tensor_parameter_50: class Program_weight_tensor_parameter_51: name = "parameter_51" + original_name = "batch_norm2d_9.b_0" shape = [512] dtype = "float32" min_val = float("-1.86842") @@ -568,6 +620,7 @@ class Program_weight_tensor_parameter_51: class Program_weight_tensor_parameter_52: name = "parameter_52" + original_name = "batch_norm2d_9.w_0" shape = [512] dtype = "float32" min_val = float("0.145913") @@ -579,6 +632,7 @@ class Program_weight_tensor_parameter_52: class Program_weight_tensor_parameter_53: name = "parameter_53" + original_name = "batch_norm2d_9.w_2" shape = [512] dtype = "float32" min_val = float("16.8886") @@ -590,6 +644,7 @@ class Program_weight_tensor_parameter_53: class Program_weight_tensor_parameter_54: name = "parameter_54" + original_name = "batch_norm2d_9.w_1" shape = [512] dtype = "float32" min_val = float("-37.4029") @@ -601,6 +656,7 @@ class Program_weight_tensor_parameter_54: class Program_weight_tensor_parameter_55: name = "parameter_55" + original_name = "conv2d_21.w_0" shape = [512, 256, 1, 1] dtype = "float32" min_val = float("-1.58189") @@ -612,6 +668,7 @@ class Program_weight_tensor_parameter_55: class Program_weight_tensor_parameter_56: name = "parameter_56" + original_name = "conv2d_23.w_0" shape = [64, 64, 3, 3] dtype = "float32" min_val = float("-1.15548") @@ -623,6 +680,7 @@ class Program_weight_tensor_parameter_56: class Program_weight_tensor_parameter_57: name = "parameter_57" + original_name = "conv2d_19.w_0" shape = [256, 512, 1, 1] dtype = "float32" min_val = float("-2.17508") @@ -634,6 +692,7 @@ class Program_weight_tensor_parameter_57: class Program_weight_tensor_parameter_58: name = "parameter_58" + original_name = "batch_norm2d_8.b_0" shape = [512] dtype = "float32" min_val = float("-1.81278") @@ -645,6 +704,7 @@ class Program_weight_tensor_parameter_58: class Program_weight_tensor_parameter_59: name = "parameter_59" + original_name = "batch_norm2d_8.w_0" shape = [512] dtype = "float32" min_val = float("0.134971") @@ -656,6 +716,7 @@ class Program_weight_tensor_parameter_59: class Program_weight_tensor_parameter_60: name = "parameter_60" + original_name = "batch_norm2d_8.w_2" shape = [512] dtype = "float32" min_val = float("12.4826") @@ -667,6 +728,7 @@ class Program_weight_tensor_parameter_60: class Program_weight_tensor_parameter_61: name = "parameter_61" + original_name = "batch_norm2d_8.w_1" shape = [512] dtype = "float32" min_val = float("-29.8724") @@ -678,6 +740,7 @@ class Program_weight_tensor_parameter_61: class Program_weight_tensor_parameter_62: name = "parameter_62" + original_name = "conv2d_18.w_0" shape = [512, 256, 1, 1] dtype = "float32" min_val = float("-1.76996") @@ -689,6 +752,7 @@ class Program_weight_tensor_parameter_62: class Program_weight_tensor_parameter_63: name = "parameter_63" + original_name = "conv2d_20.w_0" shape = [64, 64, 3, 3] dtype = "float32" min_val = float("-1.16487") @@ -700,6 +764,7 @@ class Program_weight_tensor_parameter_63: class Program_weight_tensor_parameter_64: name = "parameter_64" + original_name = "conv2d_16.w_0" shape = [256, 512, 1, 1] dtype = "float32" min_val = float("-2.63149") @@ -711,6 +776,7 @@ class Program_weight_tensor_parameter_64: class Program_weight_tensor_parameter_65: name = "parameter_65" + original_name = "batch_norm2d_7.b_0" shape = [512] dtype = "float32" min_val = float("-1.58625") @@ -722,6 +788,7 @@ class Program_weight_tensor_parameter_65: class Program_weight_tensor_parameter_66: name = "parameter_66" + original_name = "batch_norm2d_7.w_0" shape = [512] dtype = "float32" min_val = float("0.10206") @@ -733,6 +800,7 @@ class Program_weight_tensor_parameter_66: class Program_weight_tensor_parameter_67: name = "parameter_67" + original_name = "batch_norm2d_7.w_2" shape = [512] dtype = "float32" min_val = float("5.62581") @@ -744,6 +812,7 @@ class Program_weight_tensor_parameter_67: class Program_weight_tensor_parameter_68: name = "parameter_68" + original_name = "batch_norm2d_7.w_1" shape = [512] dtype = "float32" min_val = float("-26.6392") @@ -755,6 +824,7 @@ class Program_weight_tensor_parameter_68: class Program_weight_tensor_parameter_69: name = "parameter_69" + original_name = "conv2d_15.w_0" shape = [512, 256, 1, 1] dtype = "float32" min_val = float("-1.63046") @@ -766,6 +836,7 @@ class Program_weight_tensor_parameter_69: class Program_weight_tensor_parameter_70: name = "parameter_70" + original_name = "conv2d_17.w_0" shape = [64, 64, 3, 3] dtype = "float32" min_val = float("-1.36512") @@ -777,6 +848,7 @@ class Program_weight_tensor_parameter_70: class Program_weight_tensor_parameter_71: name = "parameter_71" + original_name = "conv2d_13.w_0" shape = [256, 512, 1, 1] dtype = "float32" min_val = float("-2.57951") @@ -788,6 +860,7 @@ class Program_weight_tensor_parameter_71: class Program_weight_tensor_parameter_72: name = "parameter_72" + original_name = "batch_norm2d_6.b_0" shape = [512] dtype = "float32" min_val = float("-1.2133") @@ -799,6 +872,7 @@ class Program_weight_tensor_parameter_72: class Program_weight_tensor_parameter_73: name = "parameter_73" + original_name = "batch_norm2d_6.w_0" shape = [512] dtype = "float32" min_val = float("0.0991139") @@ -810,6 +884,7 @@ class Program_weight_tensor_parameter_73: class Program_weight_tensor_parameter_74: name = "parameter_74" + original_name = "batch_norm2d_6.w_2" shape = [512] dtype = "float32" min_val = float("6.78242") @@ -821,6 +896,7 @@ class Program_weight_tensor_parameter_74: class Program_weight_tensor_parameter_75: name = "parameter_75" + original_name = "batch_norm2d_6.w_1" shape = [512] dtype = "float32" min_val = float("-42.7284") @@ -832,6 +908,7 @@ class Program_weight_tensor_parameter_75: class Program_weight_tensor_parameter_76: name = "parameter_76" + original_name = "conv2d_12.w_0" shape = [512, 256, 1, 1] dtype = "float32" min_val = float("-2.45992") @@ -843,6 +920,7 @@ class Program_weight_tensor_parameter_76: class Program_weight_tensor_parameter_77: name = "parameter_77" + original_name = "conv2d_14.w_0" shape = [64, 64, 3, 3] dtype = "float32" min_val = float("-1.46099") @@ -854,6 +932,7 @@ class Program_weight_tensor_parameter_77: class Program_weight_tensor_parameter_78: name = "parameter_78" + original_name = "batch_norm2d_5.b_0" shape = [256] dtype = "float32" min_val = float("-0.739135") @@ -865,6 +944,7 @@ class Program_weight_tensor_parameter_78: class Program_weight_tensor_parameter_79: name = "parameter_79" + original_name = "batch_norm2d_5.w_0" shape = [256] dtype = "float32" min_val = float("0.0963561") @@ -876,6 +956,7 @@ class Program_weight_tensor_parameter_79: class Program_weight_tensor_parameter_80: name = "parameter_80" + original_name = "batch_norm2d_5.w_2" shape = [256] dtype = "float32" min_val = float("165.786") @@ -887,6 +968,7 @@ class Program_weight_tensor_parameter_80: class Program_weight_tensor_parameter_81: name = "parameter_81" + original_name = "batch_norm2d_5.w_1" shape = [256] dtype = "float32" min_val = float("-22.7878") @@ -898,6 +980,7 @@ class Program_weight_tensor_parameter_81: class Program_weight_tensor_parameter_82: name = "parameter_82" + original_name = "conv2d_11.w_0" shape = [256, 128, 2, 2] dtype = "float32" min_val = float("-1.62867") @@ -909,6 +992,7 @@ class Program_weight_tensor_parameter_82: class Program_weight_tensor_parameter_83: name = "parameter_83" + original_name = "conv2d_9.w_0" shape = [128, 256, 1, 1] dtype = "float32" min_val = float("-1.85731") @@ -920,6 +1004,7 @@ class Program_weight_tensor_parameter_83: class Program_weight_tensor_parameter_84: name = "parameter_84" + original_name = "batch_norm2d_4.b_0" shape = [256] dtype = "float32" min_val = float("-1.40623") @@ -931,6 +1016,7 @@ class Program_weight_tensor_parameter_84: class Program_weight_tensor_parameter_85: name = "parameter_85" + original_name = "batch_norm2d_4.w_0" shape = [256] dtype = "float32" min_val = float("0.36565") @@ -942,6 +1028,7 @@ class Program_weight_tensor_parameter_85: class Program_weight_tensor_parameter_86: name = "parameter_86" + original_name = "batch_norm2d_4.w_2" shape = [256] dtype = "float32" min_val = float("8.64435") @@ -953,6 +1040,7 @@ class Program_weight_tensor_parameter_86: class Program_weight_tensor_parameter_87: name = "parameter_87" + original_name = "batch_norm2d_4.w_1" shape = [256] dtype = "float32" min_val = float("-4.28921") @@ -964,6 +1052,7 @@ class Program_weight_tensor_parameter_87: class Program_weight_tensor_parameter_88: name = "parameter_88" + original_name = "conv2d_8.w_0" shape = [256, 128, 1, 1] dtype = "float32" min_val = float("-1.55226") @@ -975,6 +1064,7 @@ class Program_weight_tensor_parameter_88: class Program_weight_tensor_parameter_89: name = "parameter_89" + original_name = "conv2d_10.w_0" shape = [32, 32, 3, 3] dtype = "float32" min_val = float("-1.26034") @@ -986,6 +1076,7 @@ class Program_weight_tensor_parameter_89: class Program_weight_tensor_parameter_90: name = "parameter_90" + original_name = "conv2d_6.w_0" shape = [128, 256, 1, 1] dtype = "float32" min_val = float("-1.76629") @@ -997,6 +1088,7 @@ class Program_weight_tensor_parameter_90: class Program_weight_tensor_parameter_91: name = "parameter_91" + original_name = "batch_norm2d_3.b_0" shape = [256] dtype = "float32" min_val = float("-1.65293") @@ -1008,6 +1100,7 @@ class Program_weight_tensor_parameter_91: class Program_weight_tensor_parameter_92: name = "parameter_92" + original_name = "batch_norm2d_3.w_0" shape = [256] dtype = "float32" min_val = float("0.259995") @@ -1019,6 +1112,7 @@ class Program_weight_tensor_parameter_92: class Program_weight_tensor_parameter_93: name = "parameter_93" + original_name = "batch_norm2d_3.w_2" shape = [256] dtype = "float32" min_val = float("6.59544") @@ -1030,6 +1124,7 @@ class Program_weight_tensor_parameter_93: class Program_weight_tensor_parameter_94: name = "parameter_94" + original_name = "batch_norm2d_3.w_1" shape = [256] dtype = "float32" min_val = float("-9.70948") @@ -1041,6 +1136,7 @@ class Program_weight_tensor_parameter_94: class Program_weight_tensor_parameter_95: name = "parameter_95" + original_name = "conv2d_5.w_0" shape = [256, 128, 1, 1] dtype = "float32" min_val = float("-1.66044") @@ -1052,6 +1148,7 @@ class Program_weight_tensor_parameter_95: class Program_weight_tensor_parameter_96: name = "parameter_96" + original_name = "conv2d_7.w_0" shape = [32, 32, 3, 3] dtype = "float32" min_val = float("-1.73986") @@ -1063,6 +1160,7 @@ class Program_weight_tensor_parameter_96: class Program_weight_tensor_parameter_97: name = "parameter_97" + original_name = "batch_norm2d_2.b_0" shape = [128] dtype = "float32" min_val = float("-0.742217") @@ -1074,6 +1172,7 @@ class Program_weight_tensor_parameter_97: class Program_weight_tensor_parameter_98: name = "parameter_98" + original_name = "batch_norm2d_2.w_0" shape = [128] dtype = "float32" min_val = float("0.175038") @@ -1085,6 +1184,7 @@ class Program_weight_tensor_parameter_98: class Program_weight_tensor_parameter_99: name = "parameter_99" + original_name = "batch_norm2d_2.w_2" shape = [128] dtype = "float32" min_val = float("194.727") @@ -1096,6 +1196,7 @@ class Program_weight_tensor_parameter_99: class Program_weight_tensor_parameter_100: name = "parameter_100" + original_name = "batch_norm2d_2.w_1" shape = [128] dtype = "float32" min_val = float("-47.9886") @@ -1107,6 +1208,7 @@ class Program_weight_tensor_parameter_100: class Program_weight_tensor_parameter_101: name = "parameter_101" + original_name = "conv2d_4.w_0" shape = [128, 64, 2, 2] dtype = "float32" min_val = float("-1.51342") @@ -1118,6 +1220,7 @@ class Program_weight_tensor_parameter_101: class Program_weight_tensor_parameter_102: name = "parameter_102" + original_name = "conv2d_2.w_0" shape = [64, 128, 1, 1] dtype = "float32" min_val = float("-2.46571") @@ -1129,6 +1232,7 @@ class Program_weight_tensor_parameter_102: class Program_weight_tensor_parameter_103: name = "parameter_103" + original_name = "batch_norm2d_1.b_0" shape = [128] dtype = "float32" min_val = float("-1.47553") @@ -1140,6 +1244,7 @@ class Program_weight_tensor_parameter_103: class Program_weight_tensor_parameter_104: name = "parameter_104" + original_name = "batch_norm2d_1.w_0" shape = [128] dtype = "float32" min_val = float("0.469342") @@ -1151,6 +1256,7 @@ class Program_weight_tensor_parameter_104: class Program_weight_tensor_parameter_105: name = "parameter_105" + original_name = "batch_norm2d_1.w_2" shape = [128] dtype = "float32" min_val = float("0.207494") @@ -1162,6 +1268,7 @@ class Program_weight_tensor_parameter_105: class Program_weight_tensor_parameter_106: name = "parameter_106" + original_name = "batch_norm2d_1.w_1" shape = [128] dtype = "float32" min_val = float("-2.3779") @@ -1173,6 +1280,7 @@ class Program_weight_tensor_parameter_106: class Program_weight_tensor_parameter_107: name = "parameter_107" + original_name = "conv2d_1.w_0" shape = [128, 64, 1, 1] dtype = "float32" min_val = float("-2.20433") @@ -1184,6 +1292,7 @@ class Program_weight_tensor_parameter_107: class Program_weight_tensor_parameter_108: name = "parameter_108" + original_name = "conv2d_3.w_0" shape = [16, 16, 3, 3] dtype = "float32" min_val = float("-1.54425") @@ -1195,6 +1304,7 @@ class Program_weight_tensor_parameter_108: class Program_weight_tensor_parameter_109: name = "parameter_109" + original_name = "batch_norm2d_0.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -1204,6 +1314,7 @@ class Program_weight_tensor_parameter_109: class Program_weight_tensor_parameter_110: name = "parameter_110" + original_name = "batch_norm2d_0.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -1213,6 +1324,7 @@ class Program_weight_tensor_parameter_110: class Program_weight_tensor_parameter_111: name = "parameter_111" + original_name = "batch_norm2d_0.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -1222,6 +1334,7 @@ class Program_weight_tensor_parameter_111: class Program_weight_tensor_parameter_112: name = "parameter_112" + original_name = "batch_norm2d_0.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -1231,6 +1344,7 @@ class Program_weight_tensor_parameter_112: class Program_weight_tensor_parameter_113: name = "parameter_113" + original_name = "conv2d_0.w_0" shape = [64, 3, 4, 4] dtype = "float32" min_val = float("-1.00411") diff --git a/paddle_samples/PaddleX/FasterRCNN-Swin-Tiny-FPN/subgraph_14/input_meta.py b/paddle_samples/PaddleX/FasterRCNN-Swin-Tiny-FPN/subgraph_14/input_meta.py index 389a455f6..1d951be9b 100644 --- a/paddle_samples/PaddleX/FasterRCNN-Swin-Tiny-FPN/subgraph_14/input_meta.py +++ b/paddle_samples/PaddleX/FasterRCNN-Swin-Tiny-FPN/subgraph_14/input_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_data_0: name = "data_0" + original_name = "param_70" shape = [169, 12] dtype = "float32" min_val = float("-9.66443") @@ -11,6 +12,7 @@ class Program_weight_tensor_data_0: class Program_weight_tensor_data_1: name = "data_1" + original_name = "args_1" shape = [] dtype = "int64" data = [3] @@ -18,6 +20,7 @@ class Program_weight_tensor_data_1: class Program_weight_tensor_data_2: name = "data_2" + original_name = "var_8192" shape = [2, 1920, 384] dtype = "float32" min_val = float("-138.555") @@ -29,6 +32,7 @@ class Program_weight_tensor_data_2: class Program_weight_tensor_data_3: name = "data_3" + original_name = "var_8193" shape = [42, 49, 49] dtype = "float32" min_val = float("-100.0") @@ -39,6 +43,7 @@ class Program_weight_tensor_data_3: class Program_weight_tensor_data_4: name = "data_4" + original_name = "var_8211" shape = [49, 49] dtype = "int64" min_val = 0 diff --git a/paddle_samples/PaddleX/FasterRCNN-Swin-Tiny-FPN/subgraph_14/weight_meta.py b/paddle_samples/PaddleX/FasterRCNN-Swin-Tiny-FPN/subgraph_14/weight_meta.py index 7e99d557d..ed3959ed5 100644 --- a/paddle_samples/PaddleX/FasterRCNN-Swin-Tiny-FPN/subgraph_14/weight_meta.py +++ b/paddle_samples/PaddleX/FasterRCNN-Swin-Tiny-FPN/subgraph_14/weight_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_parameter_0: name = "parameter_0" + original_name = "linear_41.b_0" shape = [384] dtype = "float32" min_val = float("-0.573867") @@ -11,6 +12,7 @@ class Program_weight_tensor_parameter_0: class Program_weight_tensor_parameter_1: name = "parameter_1" + original_name = "linear_41.w_0" shape = [1536, 384] dtype = "float32" min_val = float("-1.25327") @@ -22,6 +24,7 @@ class Program_weight_tensor_parameter_1: class Program_weight_tensor_parameter_2: name = "parameter_2" + original_name = "linear_40.b_0" shape = [1536] dtype = "float32" min_val = float("-2.81387") @@ -33,6 +36,7 @@ class Program_weight_tensor_parameter_2: class Program_weight_tensor_parameter_3: name = "parameter_3" + original_name = "linear_40.w_0" shape = [384, 1536] dtype = "float32" min_val = float("-0.56189") @@ -44,6 +48,7 @@ class Program_weight_tensor_parameter_3: class Program_weight_tensor_parameter_4: name = "parameter_4" + original_name = "layer_norm_22.b_0" shape = [384] dtype = "float32" min_val = float("-1.59366") @@ -55,6 +60,7 @@ class Program_weight_tensor_parameter_4: class Program_weight_tensor_parameter_5: name = "parameter_5" + original_name = "layer_norm_22.w_0" shape = [384] dtype = "float32" min_val = float("0.229577") @@ -66,6 +72,7 @@ class Program_weight_tensor_parameter_5: class Program_weight_tensor_parameter_6: name = "parameter_6" + original_name = "linear_39.b_0" shape = [384] dtype = "float32" min_val = float("-0.650481") @@ -77,6 +84,7 @@ class Program_weight_tensor_parameter_6: class Program_weight_tensor_parameter_7: name = "parameter_7" + original_name = "linear_39.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.418155") @@ -88,6 +96,7 @@ class Program_weight_tensor_parameter_7: class Program_weight_tensor_parameter_8: name = "parameter_8" + original_name = "linear_38.b_0" shape = [1152] dtype = "float32" min_val = float("-1.44132") @@ -99,6 +108,7 @@ class Program_weight_tensor_parameter_8: class Program_weight_tensor_parameter_9: name = "parameter_9" + original_name = "linear_38.w_0" shape = [384, 1152] dtype = "float32" min_val = float("-0.342818") @@ -110,6 +120,7 @@ class Program_weight_tensor_parameter_9: class Program_weight_tensor_parameter_10: name = "parameter_10" + original_name = "layer_norm_21.b_0" shape = [384] dtype = "float32" min_val = float("-1.39738") @@ -121,6 +132,7 @@ class Program_weight_tensor_parameter_10: class Program_weight_tensor_parameter_11: name = "parameter_11" + original_name = "layer_norm_21.w_0" shape = [384] dtype = "float32" min_val = float("0.222467") diff --git a/paddle_samples/PaddleX/FasterRCNN-Swin-Tiny-FPN/subgraph_17/input_meta.py b/paddle_samples/PaddleX/FasterRCNN-Swin-Tiny-FPN/subgraph_17/input_meta.py index 89f210e17..c030cc159 100644 --- a/paddle_samples/PaddleX/FasterRCNN-Swin-Tiny-FPN/subgraph_17/input_meta.py +++ b/paddle_samples/PaddleX/FasterRCNN-Swin-Tiny-FPN/subgraph_17/input_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_data_0: name = "data_0" + original_name = "param_62" shape = [169, 6] dtype = "float32" min_val = float("-6.72354") @@ -11,6 +12,7 @@ class Program_weight_tensor_data_0: class Program_weight_tensor_data_1: name = "data_1" + original_name = "args_1" shape = [] dtype = "int64" data = [3] @@ -18,6 +20,7 @@ class Program_weight_tensor_data_1: class Program_weight_tensor_data_2: name = "data_2" + original_name = "var_7690" shape = [2, 7680, 192] dtype = "float32" min_val = float("-3.05039") @@ -29,6 +32,7 @@ class Program_weight_tensor_data_2: class Program_weight_tensor_data_3: name = "data_3" + original_name = "var_7691" shape = [168, 49, 49] dtype = "float32" min_val = float("-100.0") @@ -39,6 +43,7 @@ class Program_weight_tensor_data_3: class Program_weight_tensor_data_4: name = "data_4" + original_name = "var_7709" shape = [49, 49] dtype = "int64" min_val = 0 diff --git a/paddle_samples/PaddleX/FasterRCNN-Swin-Tiny-FPN/subgraph_17/weight_meta.py b/paddle_samples/PaddleX/FasterRCNN-Swin-Tiny-FPN/subgraph_17/weight_meta.py index a216fb5a4..67c6a6eed 100644 --- a/paddle_samples/PaddleX/FasterRCNN-Swin-Tiny-FPN/subgraph_17/weight_meta.py +++ b/paddle_samples/PaddleX/FasterRCNN-Swin-Tiny-FPN/subgraph_17/weight_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_parameter_0: name = "parameter_0" + original_name = "linear_16.b_0" shape = [192] dtype = "float32" min_val = float("-0.531807") @@ -11,6 +12,7 @@ class Program_weight_tensor_parameter_0: class Program_weight_tensor_parameter_1: name = "parameter_1" + original_name = "linear_16.w_0" shape = [768, 192] dtype = "float32" min_val = float("-0.33168") @@ -22,6 +24,7 @@ class Program_weight_tensor_parameter_1: class Program_weight_tensor_parameter_2: name = "parameter_2" + original_name = "linear_15.b_0" shape = [768] dtype = "float32" min_val = float("-1.22843") @@ -33,6 +36,7 @@ class Program_weight_tensor_parameter_2: class Program_weight_tensor_parameter_3: name = "parameter_3" + original_name = "linear_15.w_0" shape = [192, 768] dtype = "float32" min_val = float("-0.33779") @@ -44,6 +48,7 @@ class Program_weight_tensor_parameter_3: class Program_weight_tensor_parameter_4: name = "parameter_4" + original_name = "layer_norm_9.b_0" shape = [192] dtype = "float32" min_val = float("-1.13741") @@ -55,6 +60,7 @@ class Program_weight_tensor_parameter_4: class Program_weight_tensor_parameter_5: name = "parameter_5" + original_name = "layer_norm_9.w_0" shape = [192] dtype = "float32" min_val = float("0.68955") @@ -66,6 +72,7 @@ class Program_weight_tensor_parameter_5: class Program_weight_tensor_parameter_6: name = "parameter_6" + original_name = "linear_14.b_0" shape = [192] dtype = "float32" min_val = float("-0.400882") @@ -77,6 +84,7 @@ class Program_weight_tensor_parameter_6: class Program_weight_tensor_parameter_7: name = "parameter_7" + original_name = "linear_14.w_0" shape = [192, 192] dtype = "float32" min_val = float("-0.204501") @@ -88,6 +96,7 @@ class Program_weight_tensor_parameter_7: class Program_weight_tensor_parameter_8: name = "parameter_8" + original_name = "linear_13.b_0" shape = [576] dtype = "float32" min_val = float("-1.41263") @@ -99,6 +108,7 @@ class Program_weight_tensor_parameter_8: class Program_weight_tensor_parameter_9: name = "parameter_9" + original_name = "linear_13.w_0" shape = [192, 576] dtype = "float32" min_val = float("-0.409737") @@ -110,6 +120,7 @@ class Program_weight_tensor_parameter_9: class Program_weight_tensor_parameter_10: name = "parameter_10" + original_name = "layer_norm_8.b_0" shape = [192] dtype = "float32" min_val = float("-1.36881") @@ -121,6 +132,7 @@ class Program_weight_tensor_parameter_10: class Program_weight_tensor_parameter_11: name = "parameter_11" + original_name = "layer_norm_8.w_0" shape = [192] dtype = "float32" min_val = float("0.300192") diff --git a/paddle_samples/PaddleX/FasterRCNN-Swin-Tiny-FPN/subgraph_9/input_meta.py b/paddle_samples/PaddleX/FasterRCNN-Swin-Tiny-FPN/subgraph_9/input_meta.py index 35e4ae5f9..9cb89c7c0 100644 --- a/paddle_samples/PaddleX/FasterRCNN-Swin-Tiny-FPN/subgraph_9/input_meta.py +++ b/paddle_samples/PaddleX/FasterRCNN-Swin-Tiny-FPN/subgraph_9/input_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_data_0: name = "data_0" + original_name = "param_56" shape = [169, 24] dtype = "float32" min_val = float("-11.2016") @@ -11,6 +12,7 @@ class Program_weight_tensor_data_0: class Program_weight_tensor_data_1: name = "data_1" + original_name = "param_57" shape = [169, 24] dtype = "float32" min_val = float("-11.466") @@ -22,6 +24,7 @@ class Program_weight_tensor_data_1: class Program_weight_tensor_data_2: name = "data_2" + original_name = "var_7030" shape = [2, 494, 768] dtype = "float32" min_val = float("-65.6259") @@ -33,6 +36,7 @@ class Program_weight_tensor_data_2: class Program_weight_tensor_data_3: name = "data_3" + original_name = "var_7069" shape = [49, 49] dtype = "int64" min_val = 0 @@ -42,6 +46,7 @@ class Program_weight_tensor_data_3: class Program_weight_tensor_data_4: name = "data_4" + original_name = "var_7126" shape = [49, 49] dtype = "int64" min_val = 0 diff --git a/paddle_samples/PaddleX/FasterRCNN-Swin-Tiny-FPN/subgraph_9/weight_meta.py b/paddle_samples/PaddleX/FasterRCNN-Swin-Tiny-FPN/subgraph_9/weight_meta.py index d48dd07e2..99b135e13 100644 --- a/paddle_samples/PaddleX/FasterRCNN-Swin-Tiny-FPN/subgraph_9/weight_meta.py +++ b/paddle_samples/PaddleX/FasterRCNN-Swin-Tiny-FPN/subgraph_9/weight_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_parameter_0: name = "parameter_0" + original_name = "linear_50.b_0" shape = [768] dtype = "float32" min_val = float("-5.81438") @@ -11,6 +12,7 @@ class Program_weight_tensor_parameter_0: class Program_weight_tensor_parameter_1: name = "parameter_1" + original_name = "linear_50.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-1.44605") @@ -22,6 +24,7 @@ class Program_weight_tensor_parameter_1: class Program_weight_tensor_parameter_2: name = "parameter_2" + original_name = "linear_49.b_0" shape = [3072] dtype = "float32" min_val = float("-1.20901") @@ -33,6 +36,7 @@ class Program_weight_tensor_parameter_2: class Program_weight_tensor_parameter_3: name = "parameter_3" + original_name = "linear_49.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.338827") @@ -44,6 +48,7 @@ class Program_weight_tensor_parameter_3: class Program_weight_tensor_parameter_4: name = "parameter_4" + original_name = "layer_norm_27.b_0" shape = [768] dtype = "float32" min_val = float("-1.52521") @@ -55,6 +60,7 @@ class Program_weight_tensor_parameter_4: class Program_weight_tensor_parameter_5: name = "parameter_5" + original_name = "layer_norm_27.w_0" shape = [768] dtype = "float32" min_val = float("0.00321708") @@ -66,6 +72,7 @@ class Program_weight_tensor_parameter_5: class Program_weight_tensor_parameter_6: name = "parameter_6" + original_name = "linear_48.b_0" shape = [768] dtype = "float32" min_val = float("-8.84715") @@ -77,6 +84,7 @@ class Program_weight_tensor_parameter_6: class Program_weight_tensor_parameter_7: name = "parameter_7" + original_name = "linear_48.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.795614") @@ -88,6 +96,7 @@ class Program_weight_tensor_parameter_7: class Program_weight_tensor_parameter_8: name = "parameter_8" + original_name = "linear_47.b_0" shape = [2304] dtype = "float32" min_val = float("-2.15807") @@ -99,6 +108,7 @@ class Program_weight_tensor_parameter_8: class Program_weight_tensor_parameter_9: name = "parameter_9" + original_name = "linear_47.w_0" shape = [768, 2304] dtype = "float32" min_val = float("-0.382858") @@ -110,6 +120,7 @@ class Program_weight_tensor_parameter_9: class Program_weight_tensor_parameter_10: name = "parameter_10" + original_name = "layer_norm_26.b_0" shape = [768] dtype = "float32" min_val = float("-2.22172") @@ -121,6 +132,7 @@ class Program_weight_tensor_parameter_10: class Program_weight_tensor_parameter_11: name = "parameter_11" + original_name = "layer_norm_26.w_0" shape = [768] dtype = "float32" min_val = float("0.00866992") @@ -132,6 +144,7 @@ class Program_weight_tensor_parameter_11: class Program_weight_tensor_parameter_12: name = "parameter_12" + original_name = "linear_46.b_0" shape = [768] dtype = "float32" min_val = float("-1.74389") @@ -143,6 +156,7 @@ class Program_weight_tensor_parameter_12: class Program_weight_tensor_parameter_13: name = "parameter_13" + original_name = "linear_46.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-1.36222") @@ -154,6 +168,7 @@ class Program_weight_tensor_parameter_13: class Program_weight_tensor_parameter_14: name = "parameter_14" + original_name = "linear_45.b_0" shape = [3072] dtype = "float32" min_val = float("-0.818577") @@ -165,6 +180,7 @@ class Program_weight_tensor_parameter_14: class Program_weight_tensor_parameter_15: name = "parameter_15" + original_name = "linear_45.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.373612") @@ -176,6 +192,7 @@ class Program_weight_tensor_parameter_15: class Program_weight_tensor_parameter_16: name = "parameter_16" + original_name = "layer_norm_25.b_0" shape = [768] dtype = "float32" min_val = float("-2.63537") @@ -187,6 +204,7 @@ class Program_weight_tensor_parameter_16: class Program_weight_tensor_parameter_17: name = "parameter_17" + original_name = "layer_norm_25.w_0" shape = [768] dtype = "float32" min_val = float("-0.184584") @@ -198,6 +216,7 @@ class Program_weight_tensor_parameter_17: class Program_weight_tensor_parameter_18: name = "parameter_18" + original_name = "linear_44.b_0" shape = [768] dtype = "float32" min_val = float("-3.71363") @@ -209,6 +228,7 @@ class Program_weight_tensor_parameter_18: class Program_weight_tensor_parameter_19: name = "parameter_19" + original_name = "linear_44.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.80812") @@ -220,6 +240,7 @@ class Program_weight_tensor_parameter_19: class Program_weight_tensor_parameter_20: name = "parameter_20" + original_name = "linear_43.b_0" shape = [2304] dtype = "float32" min_val = float("-2.11019") @@ -231,6 +252,7 @@ class Program_weight_tensor_parameter_20: class Program_weight_tensor_parameter_21: name = "parameter_21" + original_name = "linear_43.w_0" shape = [768, 2304] dtype = "float32" min_val = float("-0.276151") @@ -242,6 +264,7 @@ class Program_weight_tensor_parameter_21: class Program_weight_tensor_parameter_22: name = "parameter_22" + original_name = "layer_norm_24.b_0" shape = [768] dtype = "float32" min_val = float("-1.36686") @@ -253,6 +276,7 @@ class Program_weight_tensor_parameter_22: class Program_weight_tensor_parameter_23: name = "parameter_23" + original_name = "layer_norm_24.w_0" shape = [768] dtype = "float32" min_val = float("-0.00404748") diff --git a/paddle_samples/PaddleX/Mask-RT-DETR-H/subgraph_5/input_meta.py b/paddle_samples/PaddleX/Mask-RT-DETR-H/subgraph_5/input_meta.py index 953031ec6..0780ed721 100644 --- a/paddle_samples/PaddleX/Mask-RT-DETR-H/subgraph_5/input_meta.py +++ b/paddle_samples/PaddleX/Mask-RT-DETR-H/subgraph_5/input_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_data_0: name = "data_0" + original_name = "param_20" shape = [2, 256] dtype = "float32" min_val = float("-3.40036") @@ -11,6 +12,7 @@ class Program_weight_tensor_data_0: class Program_weight_tensor_data_1: name = "data_1" + original_name = "param_21" shape = [256, 768] dtype = "float32" min_val = float("-0.629611") @@ -22,6 +24,7 @@ class Program_weight_tensor_data_1: class Program_weight_tensor_data_2: name = "data_2" + original_name = "param_22" shape = [768] dtype = "float32" min_val = float("-0.745722") @@ -33,6 +36,7 @@ class Program_weight_tensor_data_2: class Program_weight_tensor_data_3: name = "data_3" + original_name = "param_23" shape = [256, 768] dtype = "float32" min_val = float("-1.12133") @@ -44,6 +48,7 @@ class Program_weight_tensor_data_3: class Program_weight_tensor_data_4: name = "data_4" + original_name = "param_24" shape = [768] dtype = "float32" min_val = float("-0.300444") @@ -55,6 +60,7 @@ class Program_weight_tensor_data_4: class Program_weight_tensor_data_5: name = "data_5" + original_name = "param_25" shape = [256, 768] dtype = "float32" min_val = float("-1.52103") @@ -66,6 +72,7 @@ class Program_weight_tensor_data_5: class Program_weight_tensor_data_6: name = "data_6" + original_name = "param_26" shape = [768] dtype = "float32" min_val = float("-0.355645") @@ -77,6 +84,7 @@ class Program_weight_tensor_data_6: class Program_weight_tensor_data_7: name = "data_7" + original_name = "param_27" shape = [256, 768] dtype = "float32" min_val = float("-0.935949") @@ -88,6 +96,7 @@ class Program_weight_tensor_data_7: class Program_weight_tensor_data_8: name = "data_8" + original_name = "param_28" shape = [768] dtype = "float32" min_val = float("-0.318404") @@ -99,6 +108,7 @@ class Program_weight_tensor_data_8: class Program_weight_tensor_data_9: name = "data_9" + original_name = "param_29" shape = [256, 768] dtype = "float32" min_val = float("-0.702972") @@ -110,6 +120,7 @@ class Program_weight_tensor_data_9: class Program_weight_tensor_data_10: name = "data_10" + original_name = "param_30" shape = [768] dtype = "float32" min_val = float("-0.198399") @@ -121,6 +132,7 @@ class Program_weight_tensor_data_10: class Program_weight_tensor_data_11: name = "data_11" + original_name = "param_31" shape = [256, 768] dtype = "float32" min_val = float("-0.51271") @@ -132,6 +144,7 @@ class Program_weight_tensor_data_11: class Program_weight_tensor_data_12: name = "data_12" + original_name = "param_32" shape = [768] dtype = "float32" min_val = float("-0.149108") @@ -143,6 +156,7 @@ class Program_weight_tensor_data_12: class Program_weight_tensor_data_13: name = "data_13" + original_name = "var_2723" shape = [1, 512, 84, 84] dtype = "float32" min_val = float("-0.556929") @@ -154,6 +168,7 @@ class Program_weight_tensor_data_13: class Program_weight_tensor_data_14: name = "data_14" + original_name = "var_2724" shape = [1, 512, 42, 42] dtype = "float32" min_val = float("-0.556929") @@ -165,6 +180,7 @@ class Program_weight_tensor_data_14: class Program_weight_tensor_data_15: name = "data_15" + original_name = "var_2725" shape = [1, 512, 21, 21] dtype = "float32" min_val = float("-0.556929") @@ -176,6 +192,7 @@ class Program_weight_tensor_data_15: class Program_weight_tensor_data_16: name = "data_16" + original_name = "var_2726" shape = [1, 128, 168, 168] dtype = "float32" min_val = float("-21.5963") @@ -187,6 +204,7 @@ class Program_weight_tensor_data_16: class Program_weight_tensor_data_17: name = "data_17" + original_name = "var_2729" shape = [1, 1] dtype = "int32" data = [0] @@ -194,6 +212,7 @@ class Program_weight_tensor_data_17: class Program_weight_tensor_data_18: name = "data_18" + original_name = "var_2730" shape = [1, 4] dtype = "float32" data = [0.376923, 0.5, 0.728846, 1.0] diff --git a/paddle_samples/PaddleX/Mask-RT-DETR-H/subgraph_5/weight_meta.py b/paddle_samples/PaddleX/Mask-RT-DETR-H/subgraph_5/weight_meta.py index 0552318e6..e214b91e7 100644 --- a/paddle_samples/PaddleX/Mask-RT-DETR-H/subgraph_5/weight_meta.py +++ b/paddle_samples/PaddleX/Mask-RT-DETR-H/subgraph_5/weight_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_parameter_0: name = "parameter_0" + original_name = "layer_norm_4.b_0_deepcopy_155" shape = [256] dtype = "float32" min_val = float("-1.08828") @@ -11,6 +12,7 @@ class Program_weight_tensor_parameter_0: class Program_weight_tensor_parameter_1: name = "parameter_1" + original_name = "layer_norm_4.w_0_deepcopy_154" shape = [256] dtype = "float32" min_val = float("0.15948") @@ -22,6 +24,7 @@ class Program_weight_tensor_parameter_1: class Program_weight_tensor_parameter_2: name = "parameter_2" + original_name = "linear_9.b_0_deepcopy_153" shape = [256] dtype = "float32" min_val = float("-0.638735") @@ -33,6 +36,7 @@ class Program_weight_tensor_parameter_2: class Program_weight_tensor_parameter_3: name = "parameter_3" + original_name = "linear_9.w_0_deepcopy_152" shape = [1024, 256] dtype = "float32" min_val = float("-0.756613") @@ -44,6 +48,7 @@ class Program_weight_tensor_parameter_3: class Program_weight_tensor_parameter_4: name = "parameter_4" + original_name = "linear_8.b_0_deepcopy_151" shape = [1024] dtype = "float32" min_val = float("-0.259211") @@ -55,6 +60,7 @@ class Program_weight_tensor_parameter_4: class Program_weight_tensor_parameter_5: name = "parameter_5" + original_name = "linear_8.w_0_deepcopy_150" shape = [256, 1024] dtype = "float32" min_val = float("-0.548407") @@ -66,6 +72,7 @@ class Program_weight_tensor_parameter_5: class Program_weight_tensor_parameter_6: name = "parameter_6" + original_name = "layer_norm_3.b_0_deepcopy_149" shape = [256] dtype = "float32" min_val = float("-0.264206") @@ -77,6 +84,7 @@ class Program_weight_tensor_parameter_6: class Program_weight_tensor_parameter_7: name = "parameter_7" + original_name = "layer_norm_3.w_0_deepcopy_148" shape = [256] dtype = "float32" min_val = float("0.819435") @@ -88,6 +96,7 @@ class Program_weight_tensor_parameter_7: class Program_weight_tensor_parameter_8: name = "parameter_8" + original_name = "linear_7.b_0_deepcopy_147" shape = [256] dtype = "float32" min_val = float("-0.228026") @@ -99,6 +108,7 @@ class Program_weight_tensor_parameter_8: class Program_weight_tensor_parameter_9: name = "parameter_9" + original_name = "linear_7.w_0_deepcopy_146" shape = [256, 256] dtype = "float32" min_val = float("-0.512949") @@ -110,6 +120,7 @@ class Program_weight_tensor_parameter_9: class Program_weight_tensor_parameter_10: name = "parameter_10" + original_name = "linear_5.b_0_deepcopy_143" shape = [96] dtype = "float32" min_val = float("-0.106224") @@ -121,6 +132,7 @@ class Program_weight_tensor_parameter_10: class Program_weight_tensor_parameter_11: name = "parameter_11" + original_name = "linear_5.w_0_deepcopy_142" shape = [256, 96] dtype = "float32" min_val = float("-0.380522") @@ -132,6 +144,7 @@ class Program_weight_tensor_parameter_11: class Program_weight_tensor_parameter_12: name = "parameter_12" + original_name = "linear_4.b_0_deepcopy_141" shape = [192] dtype = "float32" min_val = float("-4.65477") @@ -143,6 +156,7 @@ class Program_weight_tensor_parameter_12: class Program_weight_tensor_parameter_13: name = "parameter_13" + original_name = "linear_4.w_0_deepcopy_140" shape = [256, 192] dtype = "float32" min_val = float("-0.686338") @@ -154,6 +168,7 @@ class Program_weight_tensor_parameter_13: class Program_weight_tensor_parameter_14: name = "parameter_14" + original_name = "linear_6.b_0_deepcopy_145" shape = [256] dtype = "float32" min_val = float("-0.125582") @@ -165,6 +180,7 @@ class Program_weight_tensor_parameter_14: class Program_weight_tensor_parameter_15: name = "parameter_15" + original_name = "linear_6.w_0_deepcopy_144" shape = [256, 256] dtype = "float32" min_val = float("-0.492568") @@ -176,6 +192,7 @@ class Program_weight_tensor_parameter_15: class Program_weight_tensor_parameter_16: name = "parameter_16" + original_name = "layer_norm_2.b_0_deepcopy_139" shape = [256] dtype = "float32" min_val = float("-0.396249") @@ -187,6 +204,7 @@ class Program_weight_tensor_parameter_16: class Program_weight_tensor_parameter_17: name = "parameter_17" + original_name = "layer_norm_2.w_0_deepcopy_138" shape = [256] dtype = "float32" min_val = float("0.702463") @@ -198,6 +216,7 @@ class Program_weight_tensor_parameter_17: class Program_weight_tensor_parameter_18: name = "parameter_18" + original_name = "linear_3.b_0_deepcopy_137" shape = [256] dtype = "float32" min_val = float("-0.403858") @@ -209,6 +228,7 @@ class Program_weight_tensor_parameter_18: class Program_weight_tensor_parameter_19: name = "parameter_19" + original_name = "linear_3.w_0_deepcopy_136" shape = [256, 256] dtype = "float32" min_val = float("-0.422391") @@ -220,6 +240,7 @@ class Program_weight_tensor_parameter_19: class Program_weight_tensor_parameter_20: name = "parameter_20" + original_name = "layer_norm_4.b_0_deepcopy_133" shape = [256] dtype = "float32" min_val = float("-1.13462") @@ -231,6 +252,7 @@ class Program_weight_tensor_parameter_20: class Program_weight_tensor_parameter_21: name = "parameter_21" + original_name = "layer_norm_4.w_0_deepcopy_132" shape = [256] dtype = "float32" min_val = float("0.22348") @@ -242,6 +264,7 @@ class Program_weight_tensor_parameter_21: class Program_weight_tensor_parameter_22: name = "parameter_22" + original_name = "linear_9.b_0_deepcopy_131" shape = [256] dtype = "float32" min_val = float("-0.458782") @@ -253,6 +276,7 @@ class Program_weight_tensor_parameter_22: class Program_weight_tensor_parameter_23: name = "parameter_23" + original_name = "linear_9.w_0_deepcopy_130" shape = [1024, 256] dtype = "float32" min_val = float("-1.23508") @@ -264,6 +288,7 @@ class Program_weight_tensor_parameter_23: class Program_weight_tensor_parameter_24: name = "parameter_24" + original_name = "linear_8.b_0_deepcopy_129" shape = [1024] dtype = "float32" min_val = float("-0.29742") @@ -275,6 +300,7 @@ class Program_weight_tensor_parameter_24: class Program_weight_tensor_parameter_25: name = "parameter_25" + original_name = "linear_8.w_0_deepcopy_128" shape = [256, 1024] dtype = "float32" min_val = float("-0.531935") @@ -286,6 +312,7 @@ class Program_weight_tensor_parameter_25: class Program_weight_tensor_parameter_26: name = "parameter_26" + original_name = "layer_norm_3.b_0_deepcopy_127" shape = [256] dtype = "float32" min_val = float("-0.291642") @@ -297,6 +324,7 @@ class Program_weight_tensor_parameter_26: class Program_weight_tensor_parameter_27: name = "parameter_27" + original_name = "layer_norm_3.w_0_deepcopy_126" shape = [256] dtype = "float32" min_val = float("0.790209") @@ -308,6 +336,7 @@ class Program_weight_tensor_parameter_27: class Program_weight_tensor_parameter_28: name = "parameter_28" + original_name = "linear_7.b_0_deepcopy_125" shape = [256] dtype = "float32" min_val = float("-0.261482") @@ -319,6 +348,7 @@ class Program_weight_tensor_parameter_28: class Program_weight_tensor_parameter_29: name = "parameter_29" + original_name = "linear_7.w_0_deepcopy_124" shape = [256, 256] dtype = "float32" min_val = float("-0.566123") @@ -330,6 +360,7 @@ class Program_weight_tensor_parameter_29: class Program_weight_tensor_parameter_30: name = "parameter_30" + original_name = "linear_5.b_0_deepcopy_121" shape = [96] dtype = "float32" min_val = float("-0.37826") @@ -341,6 +372,7 @@ class Program_weight_tensor_parameter_30: class Program_weight_tensor_parameter_31: name = "parameter_31" + original_name = "linear_5.w_0_deepcopy_120" shape = [256, 96] dtype = "float32" min_val = float("-0.78109") @@ -352,6 +384,7 @@ class Program_weight_tensor_parameter_31: class Program_weight_tensor_parameter_32: name = "parameter_32" + original_name = "linear_4.b_0_deepcopy_119" shape = [192] dtype = "float32" min_val = float("-4.57834") @@ -363,6 +396,7 @@ class Program_weight_tensor_parameter_32: class Program_weight_tensor_parameter_33: name = "parameter_33" + original_name = "linear_4.w_0_deepcopy_118" shape = [256, 192] dtype = "float32" min_val = float("-0.640833") @@ -374,6 +408,7 @@ class Program_weight_tensor_parameter_33: class Program_weight_tensor_parameter_34: name = "parameter_34" + original_name = "linear_6.b_0_deepcopy_123" shape = [256] dtype = "float32" min_val = float("-0.209067") @@ -385,6 +420,7 @@ class Program_weight_tensor_parameter_34: class Program_weight_tensor_parameter_35: name = "parameter_35" + original_name = "linear_6.w_0_deepcopy_122" shape = [256, 256] dtype = "float32" min_val = float("-0.436521") @@ -396,6 +432,7 @@ class Program_weight_tensor_parameter_35: class Program_weight_tensor_parameter_36: name = "parameter_36" + original_name = "layer_norm_2.b_0_deepcopy_117" shape = [256] dtype = "float32" min_val = float("-0.602921") @@ -407,6 +444,7 @@ class Program_weight_tensor_parameter_36: class Program_weight_tensor_parameter_37: name = "parameter_37" + original_name = "layer_norm_2.w_0_deepcopy_116" shape = [256] dtype = "float32" min_val = float("0.134909") @@ -418,6 +456,7 @@ class Program_weight_tensor_parameter_37: class Program_weight_tensor_parameter_38: name = "parameter_38" + original_name = "linear_3.b_0_deepcopy_115" shape = [256] dtype = "float32" min_val = float("-1.67188") @@ -429,6 +468,7 @@ class Program_weight_tensor_parameter_38: class Program_weight_tensor_parameter_39: name = "parameter_39" + original_name = "linear_3.w_0_deepcopy_114" shape = [256, 256] dtype = "float32" min_val = float("-1.42317") @@ -440,6 +480,7 @@ class Program_weight_tensor_parameter_39: class Program_weight_tensor_parameter_40: name = "parameter_40" + original_name = "layer_norm_4.b_0_deepcopy_111" shape = [256] dtype = "float32" min_val = float("-0.585382") @@ -451,6 +492,7 @@ class Program_weight_tensor_parameter_40: class Program_weight_tensor_parameter_41: name = "parameter_41" + original_name = "layer_norm_4.w_0_deepcopy_110" shape = [256] dtype = "float32" min_val = float("0.501918") @@ -462,6 +504,7 @@ class Program_weight_tensor_parameter_41: class Program_weight_tensor_parameter_42: name = "parameter_42" + original_name = "linear_9.b_0_deepcopy_109" shape = [256] dtype = "float32" min_val = float("-0.766699") @@ -473,6 +516,7 @@ class Program_weight_tensor_parameter_42: class Program_weight_tensor_parameter_43: name = "parameter_43" + original_name = "linear_9.w_0_deepcopy_108" shape = [1024, 256] dtype = "float32" min_val = float("-1.7147") @@ -484,6 +528,7 @@ class Program_weight_tensor_parameter_43: class Program_weight_tensor_parameter_44: name = "parameter_44" + original_name = "linear_8.b_0_deepcopy_107" shape = [1024] dtype = "float32" min_val = float("-0.623843") @@ -495,6 +540,7 @@ class Program_weight_tensor_parameter_44: class Program_weight_tensor_parameter_45: name = "parameter_45" + original_name = "linear_8.w_0_deepcopy_106" shape = [256, 1024] dtype = "float32" min_val = float("-0.832171") @@ -506,6 +552,7 @@ class Program_weight_tensor_parameter_45: class Program_weight_tensor_parameter_46: name = "parameter_46" + original_name = "layer_norm_3.b_0_deepcopy_105" shape = [256] dtype = "float32" min_val = float("-0.450064") @@ -517,6 +564,7 @@ class Program_weight_tensor_parameter_46: class Program_weight_tensor_parameter_47: name = "parameter_47" + original_name = "layer_norm_3.w_0_deepcopy_104" shape = [256] dtype = "float32" min_val = float("0.838193") @@ -528,6 +576,7 @@ class Program_weight_tensor_parameter_47: class Program_weight_tensor_parameter_48: name = "parameter_48" + original_name = "linear_7.b_0_deepcopy_103" shape = [256] dtype = "float32" min_val = float("-0.291914") @@ -539,6 +588,7 @@ class Program_weight_tensor_parameter_48: class Program_weight_tensor_parameter_49: name = "parameter_49" + original_name = "linear_7.w_0_deepcopy_102" shape = [256, 256] dtype = "float32" min_val = float("-0.587292") @@ -550,6 +600,7 @@ class Program_weight_tensor_parameter_49: class Program_weight_tensor_parameter_50: name = "parameter_50" + original_name = "linear_5.b_0_deepcopy_99" shape = [96] dtype = "float32" min_val = float("-0.29105") @@ -561,6 +612,7 @@ class Program_weight_tensor_parameter_50: class Program_weight_tensor_parameter_51: name = "parameter_51" + original_name = "linear_5.w_0_deepcopy_98" shape = [256, 96] dtype = "float32" min_val = float("-1.08274") @@ -572,6 +624,7 @@ class Program_weight_tensor_parameter_51: class Program_weight_tensor_parameter_52: name = "parameter_52" + original_name = "linear_4.b_0_deepcopy_97" shape = [192] dtype = "float32" min_val = float("-4.52721") @@ -583,6 +636,7 @@ class Program_weight_tensor_parameter_52: class Program_weight_tensor_parameter_53: name = "parameter_53" + original_name = "linear_4.w_0_deepcopy_96" shape = [256, 192] dtype = "float32" min_val = float("-0.981726") @@ -594,6 +648,7 @@ class Program_weight_tensor_parameter_53: class Program_weight_tensor_parameter_54: name = "parameter_54" + original_name = "linear_6.b_0_deepcopy_101" shape = [256] dtype = "float32" min_val = float("-0.170452") @@ -605,6 +660,7 @@ class Program_weight_tensor_parameter_54: class Program_weight_tensor_parameter_55: name = "parameter_55" + original_name = "linear_6.w_0_deepcopy_100" shape = [256, 256] dtype = "float32" min_val = float("-0.484173") @@ -616,6 +672,7 @@ class Program_weight_tensor_parameter_55: class Program_weight_tensor_parameter_56: name = "parameter_56" + original_name = "layer_norm_2.b_0_deepcopy_95" shape = [256] dtype = "float32" min_val = float("-0.828853") @@ -627,6 +684,7 @@ class Program_weight_tensor_parameter_56: class Program_weight_tensor_parameter_57: name = "parameter_57" + original_name = "layer_norm_2.w_0_deepcopy_94" shape = [256] dtype = "float32" min_val = float("0.264938") @@ -638,6 +696,7 @@ class Program_weight_tensor_parameter_57: class Program_weight_tensor_parameter_58: name = "parameter_58" + original_name = "linear_3.b_0_deepcopy_93" shape = [256] dtype = "float32" min_val = float("-0.854488") @@ -649,6 +708,7 @@ class Program_weight_tensor_parameter_58: class Program_weight_tensor_parameter_59: name = "parameter_59" + original_name = "linear_3.w_0_deepcopy_92" shape = [256, 256] dtype = "float32" min_val = float("-0.752087") @@ -660,6 +720,7 @@ class Program_weight_tensor_parameter_59: class Program_weight_tensor_parameter_60: name = "parameter_60" + original_name = "layer_norm_4.b_0_deepcopy_89" shape = [256] dtype = "float32" min_val = float("-1.34459") @@ -671,6 +732,7 @@ class Program_weight_tensor_parameter_60: class Program_weight_tensor_parameter_61: name = "parameter_61" + original_name = "layer_norm_4.w_0_deepcopy_88" shape = [256] dtype = "float32" min_val = float("0.440873") @@ -682,6 +744,7 @@ class Program_weight_tensor_parameter_61: class Program_weight_tensor_parameter_62: name = "parameter_62" + original_name = "linear_9.b_0_deepcopy_87" shape = [256] dtype = "float32" min_val = float("-1.07526") @@ -693,6 +756,7 @@ class Program_weight_tensor_parameter_62: class Program_weight_tensor_parameter_63: name = "parameter_63" + original_name = "linear_9.w_0_deepcopy_86" shape = [1024, 256] dtype = "float32" min_val = float("-0.937008") @@ -704,6 +768,7 @@ class Program_weight_tensor_parameter_63: class Program_weight_tensor_parameter_64: name = "parameter_64" + original_name = "linear_8.b_0_deepcopy_85" shape = [1024] dtype = "float32" min_val = float("-0.520745") @@ -715,6 +780,7 @@ class Program_weight_tensor_parameter_64: class Program_weight_tensor_parameter_65: name = "parameter_65" + original_name = "linear_8.w_0_deepcopy_84" shape = [256, 1024] dtype = "float32" min_val = float("-0.953411") @@ -726,6 +792,7 @@ class Program_weight_tensor_parameter_65: class Program_weight_tensor_parameter_66: name = "parameter_66" + original_name = "layer_norm_3.b_0_deepcopy_83" shape = [256] dtype = "float32" min_val = float("-0.741091") @@ -737,6 +804,7 @@ class Program_weight_tensor_parameter_66: class Program_weight_tensor_parameter_67: name = "parameter_67" + original_name = "layer_norm_3.w_0_deepcopy_82" shape = [256] dtype = "float32" min_val = float("0.816865") @@ -748,6 +816,7 @@ class Program_weight_tensor_parameter_67: class Program_weight_tensor_parameter_68: name = "parameter_68" + original_name = "linear_7.b_0_deepcopy_81" shape = [256] dtype = "float32" min_val = float("-0.570317") @@ -759,6 +828,7 @@ class Program_weight_tensor_parameter_68: class Program_weight_tensor_parameter_69: name = "parameter_69" + original_name = "linear_7.w_0_deepcopy_80" shape = [256, 256] dtype = "float32" min_val = float("-0.475444") @@ -770,6 +840,7 @@ class Program_weight_tensor_parameter_69: class Program_weight_tensor_parameter_70: name = "parameter_70" + original_name = "linear_5.b_0_deepcopy_77" shape = [96] dtype = "float32" min_val = float("-0.426585") @@ -781,6 +852,7 @@ class Program_weight_tensor_parameter_70: class Program_weight_tensor_parameter_71: name = "parameter_71" + original_name = "linear_5.w_0_deepcopy_76" shape = [256, 96] dtype = "float32" min_val = float("-1.46776") @@ -792,6 +864,7 @@ class Program_weight_tensor_parameter_71: class Program_weight_tensor_parameter_72: name = "parameter_72" + original_name = "linear_4.b_0_deepcopy_75" shape = [192] dtype = "float32" min_val = float("-4.28183") @@ -803,6 +876,7 @@ class Program_weight_tensor_parameter_72: class Program_weight_tensor_parameter_73: name = "parameter_73" + original_name = "linear_4.w_0_deepcopy_74" shape = [256, 192] dtype = "float32" min_val = float("-0.851503") @@ -814,6 +888,7 @@ class Program_weight_tensor_parameter_73: class Program_weight_tensor_parameter_74: name = "parameter_74" + original_name = "linear_6.b_0_deepcopy_79" shape = [256] dtype = "float32" min_val = float("-0.209855") @@ -825,6 +900,7 @@ class Program_weight_tensor_parameter_74: class Program_weight_tensor_parameter_75: name = "parameter_75" + original_name = "linear_6.w_0_deepcopy_78" shape = [256, 256] dtype = "float32" min_val = float("-0.530781") @@ -836,6 +912,7 @@ class Program_weight_tensor_parameter_75: class Program_weight_tensor_parameter_76: name = "parameter_76" + original_name = "layer_norm_2.b_0_deepcopy_73" shape = [256] dtype = "float32" min_val = float("-0.849895") @@ -847,6 +924,7 @@ class Program_weight_tensor_parameter_76: class Program_weight_tensor_parameter_77: name = "parameter_77" + original_name = "layer_norm_2.w_0_deepcopy_72" shape = [256] dtype = "float32" min_val = float("0.286865") @@ -858,6 +936,7 @@ class Program_weight_tensor_parameter_77: class Program_weight_tensor_parameter_78: name = "parameter_78" + original_name = "linear_3.b_0_deepcopy_71" shape = [256] dtype = "float32" min_val = float("-0.512192") @@ -869,6 +948,7 @@ class Program_weight_tensor_parameter_78: class Program_weight_tensor_parameter_79: name = "parameter_79" + original_name = "linear_3.w_0_deepcopy_70" shape = [256, 256] dtype = "float32" min_val = float("-0.802719") @@ -880,6 +960,7 @@ class Program_weight_tensor_parameter_79: class Program_weight_tensor_parameter_80: name = "parameter_80" + original_name = "layer_norm_4.b_0_deepcopy_67" shape = [256] dtype = "float32" min_val = float("-1.31924") @@ -891,6 +972,7 @@ class Program_weight_tensor_parameter_80: class Program_weight_tensor_parameter_81: name = "parameter_81" + original_name = "layer_norm_4.w_0_deepcopy_66" shape = [256] dtype = "float32" min_val = float("0.505696") @@ -902,6 +984,7 @@ class Program_weight_tensor_parameter_81: class Program_weight_tensor_parameter_82: name = "parameter_82" + original_name = "linear_9.b_0_deepcopy_65" shape = [256] dtype = "float32" min_val = float("-0.301069") @@ -913,6 +996,7 @@ class Program_weight_tensor_parameter_82: class Program_weight_tensor_parameter_83: name = "parameter_83" + original_name = "linear_9.w_0_deepcopy_64" shape = [1024, 256] dtype = "float32" min_val = float("-0.863167") @@ -924,6 +1008,7 @@ class Program_weight_tensor_parameter_83: class Program_weight_tensor_parameter_84: name = "parameter_84" + original_name = "linear_8.b_0_deepcopy_63" shape = [1024] dtype = "float32" min_val = float("-0.637175") @@ -935,6 +1020,7 @@ class Program_weight_tensor_parameter_84: class Program_weight_tensor_parameter_85: name = "parameter_85" + original_name = "linear_8.w_0_deepcopy_62" shape = [256, 1024] dtype = "float32" min_val = float("-0.989559") @@ -946,6 +1032,7 @@ class Program_weight_tensor_parameter_85: class Program_weight_tensor_parameter_86: name = "parameter_86" + original_name = "layer_norm_3.b_0_deepcopy_61" shape = [256] dtype = "float32" min_val = float("-0.32114") @@ -957,6 +1044,7 @@ class Program_weight_tensor_parameter_86: class Program_weight_tensor_parameter_87: name = "parameter_87" + original_name = "layer_norm_3.w_0_deepcopy_60" shape = [256] dtype = "float32" min_val = float("0.844879") @@ -968,6 +1056,7 @@ class Program_weight_tensor_parameter_87: class Program_weight_tensor_parameter_88: name = "parameter_88" + original_name = "linear_7.b_0_deepcopy_59" shape = [256] dtype = "float32" min_val = float("-0.358051") @@ -979,6 +1068,7 @@ class Program_weight_tensor_parameter_88: class Program_weight_tensor_parameter_89: name = "parameter_89" + original_name = "linear_7.w_0_deepcopy_58" shape = [256, 256] dtype = "float32" min_val = float("-0.531293") @@ -990,6 +1080,7 @@ class Program_weight_tensor_parameter_89: class Program_weight_tensor_parameter_90: name = "parameter_90" + original_name = "linear_5.b_0_deepcopy_55" shape = [96] dtype = "float32" min_val = float("-0.222233") @@ -1001,6 +1092,7 @@ class Program_weight_tensor_parameter_90: class Program_weight_tensor_parameter_91: name = "parameter_91" + original_name = "linear_5.w_0_deepcopy_54" shape = [256, 96] dtype = "float32" min_val = float("-1.30554") @@ -1012,6 +1104,7 @@ class Program_weight_tensor_parameter_91: class Program_weight_tensor_parameter_92: name = "parameter_92" + original_name = "linear_4.b_0_deepcopy_53" shape = [192] dtype = "float32" min_val = float("-4.68891") @@ -1023,6 +1116,7 @@ class Program_weight_tensor_parameter_92: class Program_weight_tensor_parameter_93: name = "parameter_93" + original_name = "linear_4.w_0_deepcopy_52" shape = [256, 192] dtype = "float32" min_val = float("-1.82708") @@ -1034,6 +1128,7 @@ class Program_weight_tensor_parameter_93: class Program_weight_tensor_parameter_94: name = "parameter_94" + original_name = "linear_6.b_0_deepcopy_57" shape = [256] dtype = "float32" min_val = float("-0.227255") @@ -1045,6 +1140,7 @@ class Program_weight_tensor_parameter_94: class Program_weight_tensor_parameter_95: name = "parameter_95" + original_name = "linear_6.w_0_deepcopy_56" shape = [256, 256] dtype = "float32" min_val = float("-0.521616") @@ -1056,6 +1152,7 @@ class Program_weight_tensor_parameter_95: class Program_weight_tensor_parameter_96: name = "parameter_96" + original_name = "layer_norm_2.b_0_deepcopy_51" shape = [256] dtype = "float32" min_val = float("-0.624182") @@ -1067,6 +1164,7 @@ class Program_weight_tensor_parameter_96: class Program_weight_tensor_parameter_97: name = "parameter_97" + original_name = "layer_norm_2.w_0_deepcopy_50" shape = [256] dtype = "float32" min_val = float("0.381649") @@ -1078,6 +1176,7 @@ class Program_weight_tensor_parameter_97: class Program_weight_tensor_parameter_98: name = "parameter_98" + original_name = "linear_3.b_0_deepcopy_49" shape = [256] dtype = "float32" min_val = float("-0.396139") @@ -1089,6 +1188,7 @@ class Program_weight_tensor_parameter_98: class Program_weight_tensor_parameter_99: name = "parameter_99" + original_name = "linear_3.w_0_deepcopy_48" shape = [256, 256] dtype = "float32" min_val = float("-1.15889") @@ -1100,6 +1200,7 @@ class Program_weight_tensor_parameter_99: class Program_weight_tensor_parameter_100: name = "parameter_100" + original_name = "layer_norm_4.b_0_deepcopy_45" shape = [256] dtype = "float32" min_val = float("-1.19916") @@ -1111,6 +1212,7 @@ class Program_weight_tensor_parameter_100: class Program_weight_tensor_parameter_101: name = "parameter_101" + original_name = "layer_norm_4.w_0_deepcopy_44" shape = [256] dtype = "float32" min_val = float("0.387824") @@ -1122,6 +1224,7 @@ class Program_weight_tensor_parameter_101: class Program_weight_tensor_parameter_102: name = "parameter_102" + original_name = "linear_9.b_0_deepcopy_43" shape = [256] dtype = "float32" min_val = float("-0.564282") @@ -1133,6 +1236,7 @@ class Program_weight_tensor_parameter_102: class Program_weight_tensor_parameter_103: name = "parameter_103" + original_name = "linear_9.w_0_deepcopy_42" shape = [1024, 256] dtype = "float32" min_val = float("-0.869138") @@ -1144,6 +1248,7 @@ class Program_weight_tensor_parameter_103: class Program_weight_tensor_parameter_104: name = "parameter_104" + original_name = "linear_8.b_0_deepcopy_41" shape = [1024] dtype = "float32" min_val = float("-0.479375") @@ -1155,6 +1260,7 @@ class Program_weight_tensor_parameter_104: class Program_weight_tensor_parameter_105: name = "parameter_105" + original_name = "linear_8.w_0_deepcopy_40" shape = [256, 1024] dtype = "float32" min_val = float("-0.662676") @@ -1166,6 +1272,7 @@ class Program_weight_tensor_parameter_105: class Program_weight_tensor_parameter_106: name = "parameter_106" + original_name = "layer_norm_3.b_0_deepcopy_39" shape = [256] dtype = "float32" min_val = float("-0.337199") @@ -1177,6 +1284,7 @@ class Program_weight_tensor_parameter_106: class Program_weight_tensor_parameter_107: name = "parameter_107" + original_name = "layer_norm_3.w_0_deepcopy_38" shape = [256] dtype = "float32" min_val = float("0.871378") @@ -1188,6 +1296,7 @@ class Program_weight_tensor_parameter_107: class Program_weight_tensor_parameter_108: name = "parameter_108" + original_name = "linear_7.b_0_deepcopy_37" shape = [256] dtype = "float32" min_val = float("-0.257891") @@ -1199,6 +1308,7 @@ class Program_weight_tensor_parameter_108: class Program_weight_tensor_parameter_109: name = "parameter_109" + original_name = "linear_7.w_0_deepcopy_36" shape = [256, 256] dtype = "float32" min_val = float("-0.532757") @@ -1210,6 +1320,7 @@ class Program_weight_tensor_parameter_109: class Program_weight_tensor_parameter_110: name = "parameter_110" + original_name = "linear_5.b_0_deepcopy_33" shape = [96] dtype = "float32" min_val = float("-0.440132") @@ -1221,6 +1332,7 @@ class Program_weight_tensor_parameter_110: class Program_weight_tensor_parameter_111: name = "parameter_111" + original_name = "linear_5.w_0_deepcopy_32" shape = [256, 96] dtype = "float32" min_val = float("-0.557306") @@ -1232,6 +1344,7 @@ class Program_weight_tensor_parameter_111: class Program_weight_tensor_parameter_112: name = "parameter_112" + original_name = "linear_4.b_0_deepcopy_31" shape = [192] dtype = "float32" min_val = float("-6.06192") @@ -1243,6 +1356,7 @@ class Program_weight_tensor_parameter_112: class Program_weight_tensor_parameter_113: name = "parameter_113" + original_name = "linear_4.w_0_deepcopy_30" shape = [256, 192] dtype = "float32" min_val = float("-4.54142") @@ -1254,6 +1368,7 @@ class Program_weight_tensor_parameter_113: class Program_weight_tensor_parameter_114: name = "parameter_114" + original_name = "linear_6.b_0_deepcopy_35" shape = [256] dtype = "float32" min_val = float("-0.248139") @@ -1265,6 +1380,7 @@ class Program_weight_tensor_parameter_114: class Program_weight_tensor_parameter_115: name = "parameter_115" + original_name = "linear_6.w_0_deepcopy_34" shape = [256, 256] dtype = "float32" min_val = float("-0.485781") @@ -1276,6 +1392,7 @@ class Program_weight_tensor_parameter_115: class Program_weight_tensor_parameter_116: name = "parameter_116" + original_name = "layer_norm_2.b_0_deepcopy_29" shape = [256] dtype = "float32" min_val = float("-0.346039") @@ -1287,6 +1404,7 @@ class Program_weight_tensor_parameter_116: class Program_weight_tensor_parameter_117: name = "parameter_117" + original_name = "layer_norm_2.w_0_deepcopy_28" shape = [256] dtype = "float32" min_val = float("0.396042") @@ -1298,6 +1416,7 @@ class Program_weight_tensor_parameter_117: class Program_weight_tensor_parameter_118: name = "parameter_118" + original_name = "linear_3.b_0_deepcopy_27" shape = [256] dtype = "float32" min_val = float("-0.788951") @@ -1309,6 +1428,7 @@ class Program_weight_tensor_parameter_118: class Program_weight_tensor_parameter_119: name = "parameter_119" + original_name = "linear_3.w_0_deepcopy_26" shape = [256, 256] dtype = "float32" min_val = float("-1.63946") @@ -1320,6 +1440,7 @@ class Program_weight_tensor_parameter_119: class Program_weight_tensor_parameter_120: name = "parameter_120" + original_name = "linear_11.b_0" shape = [256] dtype = "float32" min_val = float("-0.418575") @@ -1331,6 +1452,7 @@ class Program_weight_tensor_parameter_120: class Program_weight_tensor_parameter_121: name = "parameter_121" + original_name = "linear_11.w_0" shape = [512, 256] dtype = "float32" min_val = float("-3.98602") @@ -1342,6 +1464,7 @@ class Program_weight_tensor_parameter_121: class Program_weight_tensor_parameter_122: name = "parameter_122" + original_name = "linear_10.b_0" shape = [512] dtype = "float32" min_val = float("-0.499851") @@ -1353,6 +1476,7 @@ class Program_weight_tensor_parameter_122: class Program_weight_tensor_parameter_123: name = "parameter_123" + original_name = "linear_10.w_0" shape = [4, 512] dtype = "float32" min_val = float("-12.5849") @@ -1364,6 +1488,7 @@ class Program_weight_tensor_parameter_123: class Program_weight_tensor_parameter_124: name = "parameter_124" + original_name = "linear_14.b_0" shape = [128] dtype = "float32" min_val = float("-0.141721") @@ -1375,6 +1500,7 @@ class Program_weight_tensor_parameter_124: class Program_weight_tensor_parameter_125: name = "parameter_125" + original_name = "linear_14.w_0" shape = [256, 128] dtype = "float32" min_val = float("-0.435355") @@ -1386,6 +1512,7 @@ class Program_weight_tensor_parameter_125: class Program_weight_tensor_parameter_126: name = "parameter_126" + original_name = "linear_13.b_0" shape = [256] dtype = "float32" min_val = float("-0.135588") @@ -1397,6 +1524,7 @@ class Program_weight_tensor_parameter_126: class Program_weight_tensor_parameter_127: name = "parameter_127" + original_name = "linear_13.w_0" shape = [256, 256] dtype = "float32" min_val = float("-0.930858") @@ -1408,6 +1536,7 @@ class Program_weight_tensor_parameter_127: class Program_weight_tensor_parameter_128: name = "parameter_128" + original_name = "linear_12.b_0" shape = [256] dtype = "float32" min_val = float("-0.191066") @@ -1419,6 +1548,7 @@ class Program_weight_tensor_parameter_128: class Program_weight_tensor_parameter_129: name = "parameter_129" + original_name = "linear_12.w_0" shape = [256, 256] dtype = "float32" min_val = float("-0.315793") @@ -1430,6 +1560,7 @@ class Program_weight_tensor_parameter_129: class Program_weight_tensor_parameter_130: name = "parameter_130" + original_name = "layer_norm_6.b_0" shape = [256] dtype = "float32" min_val = float("-0.250476") @@ -1441,6 +1572,7 @@ class Program_weight_tensor_parameter_130: class Program_weight_tensor_parameter_131: name = "parameter_131" + original_name = "layer_norm_6.w_0" shape = [256] dtype = "float32" min_val = float("0.794544") @@ -1452,6 +1584,7 @@ class Program_weight_tensor_parameter_131: class Program_weight_tensor_parameter_132: name = "parameter_132" + original_name = "linear_19.b_0" shape = [4] dtype = "float32" min_val = float("0") @@ -1461,6 +1594,7 @@ class Program_weight_tensor_parameter_132: class Program_weight_tensor_parameter_133: name = "parameter_133" + original_name = "linear_19.w_0" shape = [256, 4] dtype = "float32" min_val = float("-0.509244") @@ -1472,6 +1606,7 @@ class Program_weight_tensor_parameter_133: class Program_weight_tensor_parameter_134: name = "parameter_134" + original_name = "linear_18.b_0" shape = [256] dtype = "float32" min_val = float("-0.202204") @@ -1483,6 +1618,7 @@ class Program_weight_tensor_parameter_134: class Program_weight_tensor_parameter_135: name = "parameter_135" + original_name = "linear_18.w_0" shape = [256, 256] dtype = "float32" min_val = float("-1.90312") @@ -1494,6 +1630,7 @@ class Program_weight_tensor_parameter_135: class Program_weight_tensor_parameter_136: name = "parameter_136" + original_name = "linear_17.b_0" shape = [256] dtype = "float32" min_val = float("-0.193429") @@ -1505,6 +1642,7 @@ class Program_weight_tensor_parameter_136: class Program_weight_tensor_parameter_137: name = "parameter_137" + original_name = "linear_17.w_0" shape = [256, 256] dtype = "float32" min_val = float("-0.34687") @@ -1516,6 +1654,7 @@ class Program_weight_tensor_parameter_137: class Program_weight_tensor_parameter_138: name = "parameter_138" + original_name = "linear_16.b_0" shape = [2] dtype = "float32" min_val = float("0") @@ -1525,6 +1664,7 @@ class Program_weight_tensor_parameter_138: class Program_weight_tensor_parameter_139: name = "parameter_139" + original_name = "linear_16.w_0" shape = [256, 2] dtype = "float32" min_val = float("-0.0624725") @@ -1536,6 +1676,7 @@ class Program_weight_tensor_parameter_139: class Program_weight_tensor_parameter_140: name = "parameter_140" + original_name = "layer_norm_5.b_0" shape = [256] dtype = "float32" min_val = float("-0.76069") @@ -1547,6 +1688,7 @@ class Program_weight_tensor_parameter_140: class Program_weight_tensor_parameter_141: name = "parameter_141" + original_name = "layer_norm_5.w_0" shape = [256] dtype = "float32" min_val = float("-0.168207") @@ -1558,6 +1700,7 @@ class Program_weight_tensor_parameter_141: class Program_weight_tensor_parameter_142: name = "parameter_142" + original_name = "linear_15.b_0" shape = [256] dtype = "float32" min_val = float("-0.69633") @@ -1569,6 +1712,7 @@ class Program_weight_tensor_parameter_142: class Program_weight_tensor_parameter_143: name = "parameter_143" + original_name = "linear_15.w_0" shape = [256, 256] dtype = "float32" min_val = float("-1.17933") @@ -1580,6 +1724,7 @@ class Program_weight_tensor_parameter_143: class Program_weight_tensor_parameter_144: name = "parameter_144" + original_name = "batch_norm2d_396.b_0" shape = [256] dtype = "float32" min_val = float("-0.164808") @@ -1591,6 +1736,7 @@ class Program_weight_tensor_parameter_144: class Program_weight_tensor_parameter_145: name = "parameter_145" + original_name = "batch_norm2d_396.w_0" shape = [256] dtype = "float32" min_val = float("0.714905") @@ -1602,6 +1748,7 @@ class Program_weight_tensor_parameter_145: class Program_weight_tensor_parameter_146: name = "parameter_146" + original_name = "batch_norm2d_396.w_2" shape = [256] dtype = "float32" min_val = float("287.428") @@ -1613,6 +1760,7 @@ class Program_weight_tensor_parameter_146: class Program_weight_tensor_parameter_147: name = "parameter_147" + original_name = "batch_norm2d_396.w_1" shape = [256] dtype = "float32" min_val = float("-3.2488") @@ -1624,6 +1772,7 @@ class Program_weight_tensor_parameter_147: class Program_weight_tensor_parameter_148: name = "parameter_148" + original_name = "conv2d_223.w_0" shape = [256, 512, 1, 1] dtype = "float32" min_val = float("-0.571843") @@ -1635,6 +1784,7 @@ class Program_weight_tensor_parameter_148: class Program_weight_tensor_parameter_149: name = "parameter_149" + original_name = "batch_norm2d_395.b_0" shape = [256] dtype = "float32" min_val = float("-0.276862") @@ -1646,6 +1796,7 @@ class Program_weight_tensor_parameter_149: class Program_weight_tensor_parameter_150: name = "parameter_150" + original_name = "batch_norm2d_395.w_0" shape = [256] dtype = "float32" min_val = float("0.628871") @@ -1657,6 +1808,7 @@ class Program_weight_tensor_parameter_150: class Program_weight_tensor_parameter_151: name = "parameter_151" + original_name = "batch_norm2d_395.w_2" shape = [256] dtype = "float32" min_val = float("373.6") @@ -1668,6 +1820,7 @@ class Program_weight_tensor_parameter_151: class Program_weight_tensor_parameter_152: name = "parameter_152" + original_name = "batch_norm2d_395.w_1" shape = [256] dtype = "float32" min_val = float("-2.98223") @@ -1679,6 +1832,7 @@ class Program_weight_tensor_parameter_152: class Program_weight_tensor_parameter_153: name = "parameter_153" + original_name = "conv2d_222.w_0" shape = [256, 512, 1, 1] dtype = "float32" min_val = float("-0.67272") @@ -1690,6 +1844,7 @@ class Program_weight_tensor_parameter_153: class Program_weight_tensor_parameter_154: name = "parameter_154" + original_name = "batch_norm2d_394.b_0" shape = [256] dtype = "float32" min_val = float("-0.16536") @@ -1701,6 +1856,7 @@ class Program_weight_tensor_parameter_154: class Program_weight_tensor_parameter_155: name = "parameter_155" + original_name = "batch_norm2d_394.w_0" shape = [256] dtype = "float32" min_val = float("0.580973") @@ -1712,6 +1868,7 @@ class Program_weight_tensor_parameter_155: class Program_weight_tensor_parameter_156: name = "parameter_156" + original_name = "batch_norm2d_394.w_2" shape = [256] dtype = "float32" min_val = float("341.57") @@ -1723,6 +1880,7 @@ class Program_weight_tensor_parameter_156: class Program_weight_tensor_parameter_157: name = "parameter_157" + original_name = "batch_norm2d_394.w_1" shape = [256] dtype = "float32" min_val = float("-3.08787") @@ -1734,6 +1892,7 @@ class Program_weight_tensor_parameter_157: class Program_weight_tensor_parameter_158: name = "parameter_158" + original_name = "conv2d_221.w_0" shape = [256, 512, 1, 1] dtype = "float32" min_val = float("-0.572443") diff --git a/paddle_samples/PaddleX/Mask-RT-DETR-L/subgraph_11/input_meta.py b/paddle_samples/PaddleX/Mask-RT-DETR-L/subgraph_11/input_meta.py index 3abe37196..274b1883a 100644 --- a/paddle_samples/PaddleX/Mask-RT-DETR-L/subgraph_11/input_meta.py +++ b/paddle_samples/PaddleX/Mask-RT-DETR-L/subgraph_11/input_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_data_0: name = "data_0" + original_name = "param_25" shape = [2, 256] dtype = "float32" min_val = float("-3.27767") @@ -11,6 +12,7 @@ class Program_weight_tensor_data_0: class Program_weight_tensor_data_1: name = "data_1" + original_name = "param_26" shape = [256, 768] dtype = "float32" min_val = float("-0.388813") @@ -22,6 +24,7 @@ class Program_weight_tensor_data_1: class Program_weight_tensor_data_2: name = "data_2" + original_name = "param_27" shape = [768] dtype = "float32" min_val = float("-0.226413") @@ -33,6 +36,7 @@ class Program_weight_tensor_data_2: class Program_weight_tensor_data_3: name = "data_3" + original_name = "param_28" shape = [256, 768] dtype = "float32" min_val = float("-0.431728") @@ -44,6 +48,7 @@ class Program_weight_tensor_data_3: class Program_weight_tensor_data_4: name = "data_4" + original_name = "param_29" shape = [768] dtype = "float32" min_val = float("-0.155516") @@ -55,6 +60,7 @@ class Program_weight_tensor_data_4: class Program_weight_tensor_data_5: name = "data_5" + original_name = "param_30" shape = [256, 768] dtype = "float32" min_val = float("-0.358802") @@ -66,6 +72,7 @@ class Program_weight_tensor_data_5: class Program_weight_tensor_data_6: name = "data_6" + original_name = "param_31" shape = [768] dtype = "float32" min_val = float("-0.119078") @@ -77,6 +84,7 @@ class Program_weight_tensor_data_6: class Program_weight_tensor_data_7: name = "data_7" + original_name = "param_32" shape = [256, 768] dtype = "float32" min_val = float("-0.372305") @@ -88,6 +96,7 @@ class Program_weight_tensor_data_7: class Program_weight_tensor_data_8: name = "data_8" + original_name = "param_33" shape = [768] dtype = "float32" min_val = float("-0.0974898") @@ -99,6 +108,7 @@ class Program_weight_tensor_data_8: class Program_weight_tensor_data_9: name = "data_9" + original_name = "param_34" shape = [256, 768] dtype = "float32" min_val = float("-0.347611") @@ -110,6 +120,7 @@ class Program_weight_tensor_data_9: class Program_weight_tensor_data_10: name = "data_10" + original_name = "param_35" shape = [768] dtype = "float32" min_val = float("-0.109375") @@ -121,6 +132,7 @@ class Program_weight_tensor_data_10: class Program_weight_tensor_data_11: name = "data_11" + original_name = "param_36" shape = [256, 768] dtype = "float32" min_val = float("-0.295065") @@ -132,6 +144,7 @@ class Program_weight_tensor_data_11: class Program_weight_tensor_data_12: name = "data_12" + original_name = "param_37" shape = [768] dtype = "float32" min_val = float("-0.0793127") @@ -143,6 +156,7 @@ class Program_weight_tensor_data_12: class Program_weight_tensor_data_13: name = "data_13" + original_name = "args_13" shape = [] dtype = "int64" data = [200] @@ -150,6 +164,7 @@ class Program_weight_tensor_data_13: class Program_weight_tensor_data_14: name = "data_14" + original_name = "args_14" shape = [] dtype = "int64" data = [200] @@ -157,6 +172,7 @@ class Program_weight_tensor_data_14: class Program_weight_tensor_data_15: name = "data_15" + original_name = "var_8886" shape = [2, 256, 100, 100] dtype = "float32" min_val = float("-0.556929") @@ -168,6 +184,7 @@ class Program_weight_tensor_data_15: class Program_weight_tensor_data_16: name = "data_16" + original_name = "var_8887" shape = [2, 256, 50, 50] dtype = "float32" min_val = float("-0.556927") @@ -179,6 +196,7 @@ class Program_weight_tensor_data_16: class Program_weight_tensor_data_17: name = "data_17" + original_name = "var_8888" shape = [2, 256, 25, 25] dtype = "float32" min_val = float("-0.556924") @@ -190,6 +208,7 @@ class Program_weight_tensor_data_17: class Program_weight_tensor_data_18: name = "data_18" + original_name = "var_8889" shape = [2, 32, 200, 200] dtype = "float32" min_val = float("-16.272") @@ -201,6 +220,7 @@ class Program_weight_tensor_data_18: class Program_weight_tensor_data_19: name = "data_19" + original_name = "var_8893" shape = [1, 1] dtype = "int32" data = [0] @@ -208,6 +228,7 @@ class Program_weight_tensor_data_19: class Program_weight_tensor_data_20: name = "data_20" + original_name = "var_8894" shape = [1, 1] dtype = "int32" data = [1] @@ -215,6 +236,7 @@ class Program_weight_tensor_data_20: class Program_weight_tensor_data_21: name = "data_21" + original_name = "var_8895" shape = [1, 4] dtype = "float32" data = [0.5, 0.651724, 1.0, 0.696552] @@ -222,6 +244,7 @@ class Program_weight_tensor_data_21: class Program_weight_tensor_data_22: name = "data_22" + original_name = "var_8896" shape = [1, 4] dtype = "float32" data = [0.606696, 0.503521, 0.559821, 0.992958] diff --git a/paddle_samples/PaddleX/Mask-RT-DETR-L/subgraph_11/weight_meta.py b/paddle_samples/PaddleX/Mask-RT-DETR-L/subgraph_11/weight_meta.py index ba73367f4..fcfc82685 100644 --- a/paddle_samples/PaddleX/Mask-RT-DETR-L/subgraph_11/weight_meta.py +++ b/paddle_samples/PaddleX/Mask-RT-DETR-L/subgraph_11/weight_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_parameter_0: name = "parameter_0" + original_name = "layer_norm_4.b_0_deepcopy_143" shape = [256] dtype = "float32" min_val = float("-0.235318") @@ -11,6 +12,7 @@ class Program_weight_tensor_parameter_0: class Program_weight_tensor_parameter_1: name = "parameter_1" + original_name = "layer_norm_4.w_0_deepcopy_142" shape = [256] dtype = "float32" min_val = float("0.669133") @@ -22,6 +24,7 @@ class Program_weight_tensor_parameter_1: class Program_weight_tensor_parameter_2: name = "parameter_2" + original_name = "linear_9.b_0_deepcopy_141" shape = [256] dtype = "float32" min_val = float("-0.229282") @@ -33,6 +36,7 @@ class Program_weight_tensor_parameter_2: class Program_weight_tensor_parameter_3: name = "parameter_3" + original_name = "linear_9.w_0_deepcopy_140" shape = [1024, 256] dtype = "float32" min_val = float("-0.411993") @@ -44,6 +48,7 @@ class Program_weight_tensor_parameter_3: class Program_weight_tensor_parameter_4: name = "parameter_4" + original_name = "linear_8.b_0_deepcopy_139" shape = [1024] dtype = "float32" min_val = float("-0.160422") @@ -55,6 +60,7 @@ class Program_weight_tensor_parameter_4: class Program_weight_tensor_parameter_5: name = "parameter_5" + original_name = "linear_8.w_0_deepcopy_138" shape = [256, 1024] dtype = "float32" min_val = float("-0.314147") @@ -66,6 +72,7 @@ class Program_weight_tensor_parameter_5: class Program_weight_tensor_parameter_6: name = "parameter_6" + original_name = "layer_norm_3.b_0_deepcopy_137" shape = [256] dtype = "float32" min_val = float("-0.165694") @@ -77,6 +84,7 @@ class Program_weight_tensor_parameter_6: class Program_weight_tensor_parameter_7: name = "parameter_7" + original_name = "layer_norm_3.w_0_deepcopy_136" shape = [256] dtype = "float32" min_val = float("0.819601") @@ -88,6 +96,7 @@ class Program_weight_tensor_parameter_7: class Program_weight_tensor_parameter_8: name = "parameter_8" + original_name = "linear_7.b_0_deepcopy_135" shape = [256] dtype = "float32" min_val = float("-0.167097") @@ -99,6 +108,7 @@ class Program_weight_tensor_parameter_8: class Program_weight_tensor_parameter_9: name = "parameter_9" + original_name = "linear_7.w_0_deepcopy_134" shape = [256, 256] dtype = "float32" min_val = float("-0.274856") @@ -110,6 +120,7 @@ class Program_weight_tensor_parameter_9: class Program_weight_tensor_parameter_10: name = "parameter_10" + original_name = "linear_5.b_0_deepcopy_131" shape = [96] dtype = "float32" min_val = float("-0.0530413") @@ -121,6 +132,7 @@ class Program_weight_tensor_parameter_10: class Program_weight_tensor_parameter_11: name = "parameter_11" + original_name = "linear_5.w_0_deepcopy_130" shape = [256, 96] dtype = "float32" min_val = float("-0.351493") @@ -132,6 +144,7 @@ class Program_weight_tensor_parameter_11: class Program_weight_tensor_parameter_12: name = "parameter_12" + original_name = "linear_4.b_0_deepcopy_129" shape = [192] dtype = "float32" min_val = float("-4.13705") @@ -143,6 +156,7 @@ class Program_weight_tensor_parameter_12: class Program_weight_tensor_parameter_13: name = "parameter_13" + original_name = "linear_4.w_0_deepcopy_128" shape = [256, 192] dtype = "float32" min_val = float("-0.329086") @@ -154,6 +168,7 @@ class Program_weight_tensor_parameter_13: class Program_weight_tensor_parameter_14: name = "parameter_14" + original_name = "linear_6.b_0_deepcopy_133" shape = [256] dtype = "float32" min_val = float("-0.0672903") @@ -165,6 +180,7 @@ class Program_weight_tensor_parameter_14: class Program_weight_tensor_parameter_15: name = "parameter_15" + original_name = "linear_6.w_0_deepcopy_132" shape = [256, 256] dtype = "float32" min_val = float("-0.291328") @@ -176,6 +192,7 @@ class Program_weight_tensor_parameter_15: class Program_weight_tensor_parameter_16: name = "parameter_16" + original_name = "layer_norm_2.b_0_deepcopy_127" shape = [256] dtype = "float32" min_val = float("-0.113191") @@ -187,6 +204,7 @@ class Program_weight_tensor_parameter_16: class Program_weight_tensor_parameter_17: name = "parameter_17" + original_name = "layer_norm_2.w_0_deepcopy_126" shape = [256] dtype = "float32" min_val = float("0.91342") @@ -198,6 +216,7 @@ class Program_weight_tensor_parameter_17: class Program_weight_tensor_parameter_18: name = "parameter_18" + original_name = "linear_3.b_0_deepcopy_125" shape = [256] dtype = "float32" min_val = float("-0.402995") @@ -209,6 +228,7 @@ class Program_weight_tensor_parameter_18: class Program_weight_tensor_parameter_19: name = "parameter_19" + original_name = "linear_3.w_0_deepcopy_124" shape = [256, 256] dtype = "float32" min_val = float("-0.385778") @@ -220,6 +240,7 @@ class Program_weight_tensor_parameter_19: class Program_weight_tensor_parameter_20: name = "parameter_20" + original_name = "layer_norm_4.b_0_deepcopy_121" shape = [256] dtype = "float32" min_val = float("-0.242023") @@ -231,6 +252,7 @@ class Program_weight_tensor_parameter_20: class Program_weight_tensor_parameter_21: name = "parameter_21" + original_name = "layer_norm_4.w_0_deepcopy_120" shape = [256] dtype = "float32" min_val = float("0.651354") @@ -242,6 +264,7 @@ class Program_weight_tensor_parameter_21: class Program_weight_tensor_parameter_22: name = "parameter_22" + original_name = "linear_9.b_0_deepcopy_119" shape = [256] dtype = "float32" min_val = float("-0.242045") @@ -253,6 +276,7 @@ class Program_weight_tensor_parameter_22: class Program_weight_tensor_parameter_23: name = "parameter_23" + original_name = "linear_9.w_0_deepcopy_118" shape = [1024, 256] dtype = "float32" min_val = float("-0.319603") @@ -264,6 +288,7 @@ class Program_weight_tensor_parameter_23: class Program_weight_tensor_parameter_24: name = "parameter_24" + original_name = "linear_8.b_0_deepcopy_117" shape = [1024] dtype = "float32" min_val = float("-0.177691") @@ -275,6 +300,7 @@ class Program_weight_tensor_parameter_24: class Program_weight_tensor_parameter_25: name = "parameter_25" + original_name = "linear_8.w_0_deepcopy_116" shape = [256, 1024] dtype = "float32" min_val = float("-0.295095") @@ -286,6 +312,7 @@ class Program_weight_tensor_parameter_25: class Program_weight_tensor_parameter_26: name = "parameter_26" + original_name = "layer_norm_3.b_0_deepcopy_115" shape = [256] dtype = "float32" min_val = float("-0.0836084") @@ -297,6 +324,7 @@ class Program_weight_tensor_parameter_26: class Program_weight_tensor_parameter_27: name = "parameter_27" + original_name = "layer_norm_3.w_0_deepcopy_114" shape = [256] dtype = "float32" min_val = float("0.841125") @@ -308,6 +336,7 @@ class Program_weight_tensor_parameter_27: class Program_weight_tensor_parameter_28: name = "parameter_28" + original_name = "linear_7.b_0_deepcopy_113" shape = [256] dtype = "float32" min_val = float("-0.100835") @@ -319,6 +348,7 @@ class Program_weight_tensor_parameter_28: class Program_weight_tensor_parameter_29: name = "parameter_29" + original_name = "linear_7.w_0_deepcopy_112" shape = [256, 256] dtype = "float32" min_val = float("-0.258674") @@ -330,6 +360,7 @@ class Program_weight_tensor_parameter_29: class Program_weight_tensor_parameter_30: name = "parameter_30" + original_name = "linear_5.b_0_deepcopy_109" shape = [96] dtype = "float32" min_val = float("-0.113839") @@ -341,6 +372,7 @@ class Program_weight_tensor_parameter_30: class Program_weight_tensor_parameter_31: name = "parameter_31" + original_name = "linear_5.w_0_deepcopy_108" shape = [256, 96] dtype = "float32" min_val = float("-0.310426") @@ -352,6 +384,7 @@ class Program_weight_tensor_parameter_31: class Program_weight_tensor_parameter_32: name = "parameter_32" + original_name = "linear_4.b_0_deepcopy_107" shape = [192] dtype = "float32" min_val = float("-4.16001") @@ -363,6 +396,7 @@ class Program_weight_tensor_parameter_32: class Program_weight_tensor_parameter_33: name = "parameter_33" + original_name = "linear_4.w_0_deepcopy_106" shape = [256, 192] dtype = "float32" min_val = float("-0.506003") @@ -374,6 +408,7 @@ class Program_weight_tensor_parameter_33: class Program_weight_tensor_parameter_34: name = "parameter_34" + original_name = "linear_6.b_0_deepcopy_111" shape = [256] dtype = "float32" min_val = float("-0.0981877") @@ -385,6 +420,7 @@ class Program_weight_tensor_parameter_34: class Program_weight_tensor_parameter_35: name = "parameter_35" + original_name = "linear_6.w_0_deepcopy_110" shape = [256, 256] dtype = "float32" min_val = float("-0.305591") @@ -396,6 +432,7 @@ class Program_weight_tensor_parameter_35: class Program_weight_tensor_parameter_36: name = "parameter_36" + original_name = "layer_norm_2.b_0_deepcopy_105" shape = [256] dtype = "float32" min_val = float("-0.113256") @@ -407,6 +444,7 @@ class Program_weight_tensor_parameter_36: class Program_weight_tensor_parameter_37: name = "parameter_37" + original_name = "layer_norm_2.w_0_deepcopy_104" shape = [256] dtype = "float32" min_val = float("0.809153") @@ -418,6 +456,7 @@ class Program_weight_tensor_parameter_37: class Program_weight_tensor_parameter_38: name = "parameter_38" + original_name = "linear_3.b_0_deepcopy_103" shape = [256] dtype = "float32" min_val = float("-0.148367") @@ -429,6 +468,7 @@ class Program_weight_tensor_parameter_38: class Program_weight_tensor_parameter_39: name = "parameter_39" + original_name = "linear_3.w_0_deepcopy_102" shape = [256, 256] dtype = "float32" min_val = float("-0.275028") @@ -440,6 +480,7 @@ class Program_weight_tensor_parameter_39: class Program_weight_tensor_parameter_40: name = "parameter_40" + original_name = "layer_norm_4.b_0_deepcopy_99" shape = [256] dtype = "float32" min_val = float("-0.152569") @@ -451,6 +492,7 @@ class Program_weight_tensor_parameter_40: class Program_weight_tensor_parameter_41: name = "parameter_41" + original_name = "layer_norm_4.w_0_deepcopy_98" shape = [256] dtype = "float32" min_val = float("0.71104") @@ -462,6 +504,7 @@ class Program_weight_tensor_parameter_41: class Program_weight_tensor_parameter_42: name = "parameter_42" + original_name = "linear_9.b_0_deepcopy_97" shape = [256] dtype = "float32" min_val = float("-0.145011") @@ -473,6 +516,7 @@ class Program_weight_tensor_parameter_42: class Program_weight_tensor_parameter_43: name = "parameter_43" + original_name = "linear_9.w_0_deepcopy_96" shape = [1024, 256] dtype = "float32" min_val = float("-0.340016") @@ -484,6 +528,7 @@ class Program_weight_tensor_parameter_43: class Program_weight_tensor_parameter_44: name = "parameter_44" + original_name = "linear_8.b_0_deepcopy_95" shape = [1024] dtype = "float32" min_val = float("-0.194362") @@ -495,6 +540,7 @@ class Program_weight_tensor_parameter_44: class Program_weight_tensor_parameter_45: name = "parameter_45" + original_name = "linear_8.w_0_deepcopy_94" shape = [256, 1024] dtype = "float32" min_val = float("-0.32099") @@ -506,6 +552,7 @@ class Program_weight_tensor_parameter_45: class Program_weight_tensor_parameter_46: name = "parameter_46" + original_name = "layer_norm_3.b_0_deepcopy_93" shape = [256] dtype = "float32" min_val = float("-0.105715") @@ -517,6 +564,7 @@ class Program_weight_tensor_parameter_46: class Program_weight_tensor_parameter_47: name = "parameter_47" + original_name = "layer_norm_3.w_0_deepcopy_92" shape = [256] dtype = "float32" min_val = float("0.875368") @@ -528,6 +576,7 @@ class Program_weight_tensor_parameter_47: class Program_weight_tensor_parameter_48: name = "parameter_48" + original_name = "linear_7.b_0_deepcopy_91" shape = [256] dtype = "float32" min_val = float("-0.125952") @@ -539,6 +588,7 @@ class Program_weight_tensor_parameter_48: class Program_weight_tensor_parameter_49: name = "parameter_49" + original_name = "linear_7.w_0_deepcopy_90" shape = [256, 256] dtype = "float32" min_val = float("-0.280394") @@ -550,6 +600,7 @@ class Program_weight_tensor_parameter_49: class Program_weight_tensor_parameter_50: name = "parameter_50" + original_name = "linear_5.b_0_deepcopy_87" shape = [96] dtype = "float32" min_val = float("-0.102302") @@ -561,6 +612,7 @@ class Program_weight_tensor_parameter_50: class Program_weight_tensor_parameter_51: name = "parameter_51" + original_name = "linear_5.w_0_deepcopy_86" shape = [256, 96] dtype = "float32" min_val = float("-0.343369") @@ -572,6 +624,7 @@ class Program_weight_tensor_parameter_51: class Program_weight_tensor_parameter_52: name = "parameter_52" + original_name = "linear_4.b_0_deepcopy_85" shape = [192] dtype = "float32" min_val = float("-4.23366") @@ -583,6 +636,7 @@ class Program_weight_tensor_parameter_52: class Program_weight_tensor_parameter_53: name = "parameter_53" + original_name = "linear_4.w_0_deepcopy_84" shape = [256, 192] dtype = "float32" min_val = float("-0.43715") @@ -594,6 +648,7 @@ class Program_weight_tensor_parameter_53: class Program_weight_tensor_parameter_54: name = "parameter_54" + original_name = "linear_6.b_0_deepcopy_89" shape = [256] dtype = "float32" min_val = float("-0.0959492") @@ -605,6 +660,7 @@ class Program_weight_tensor_parameter_54: class Program_weight_tensor_parameter_55: name = "parameter_55" + original_name = "linear_6.w_0_deepcopy_88" shape = [256, 256] dtype = "float32" min_val = float("-0.312213") @@ -616,6 +672,7 @@ class Program_weight_tensor_parameter_55: class Program_weight_tensor_parameter_56: name = "parameter_56" + original_name = "layer_norm_2.b_0_deepcopy_83" shape = [256] dtype = "float32" min_val = float("-0.099881") @@ -627,6 +684,7 @@ class Program_weight_tensor_parameter_56: class Program_weight_tensor_parameter_57: name = "parameter_57" + original_name = "layer_norm_2.w_0_deepcopy_82" shape = [256] dtype = "float32" min_val = float("0.805496") @@ -638,6 +696,7 @@ class Program_weight_tensor_parameter_57: class Program_weight_tensor_parameter_58: name = "parameter_58" + original_name = "linear_3.b_0_deepcopy_81" shape = [256] dtype = "float32" min_val = float("-0.1321") @@ -649,6 +708,7 @@ class Program_weight_tensor_parameter_58: class Program_weight_tensor_parameter_59: name = "parameter_59" + original_name = "linear_3.w_0_deepcopy_80" shape = [256, 256] dtype = "float32" min_val = float("-0.356366") @@ -660,6 +720,7 @@ class Program_weight_tensor_parameter_59: class Program_weight_tensor_parameter_60: name = "parameter_60" + original_name = "layer_norm_4.b_0_deepcopy_77" shape = [256] dtype = "float32" min_val = float("-0.091823") @@ -671,6 +732,7 @@ class Program_weight_tensor_parameter_60: class Program_weight_tensor_parameter_61: name = "parameter_61" + original_name = "layer_norm_4.w_0_deepcopy_76" shape = [256] dtype = "float32" min_val = float("0.763869") @@ -682,6 +744,7 @@ class Program_weight_tensor_parameter_61: class Program_weight_tensor_parameter_62: name = "parameter_62" + original_name = "linear_9.b_0_deepcopy_75" shape = [256] dtype = "float32" min_val = float("-0.11669") @@ -693,6 +756,7 @@ class Program_weight_tensor_parameter_62: class Program_weight_tensor_parameter_63: name = "parameter_63" + original_name = "linear_9.w_0_deepcopy_74" shape = [1024, 256] dtype = "float32" min_val = float("-0.424501") @@ -704,6 +768,7 @@ class Program_weight_tensor_parameter_63: class Program_weight_tensor_parameter_64: name = "parameter_64" + original_name = "linear_8.b_0_deepcopy_73" shape = [1024] dtype = "float32" min_val = float("-0.265827") @@ -715,6 +780,7 @@ class Program_weight_tensor_parameter_64: class Program_weight_tensor_parameter_65: name = "parameter_65" + original_name = "linear_8.w_0_deepcopy_72" shape = [256, 1024] dtype = "float32" min_val = float("-0.349265") @@ -726,6 +792,7 @@ class Program_weight_tensor_parameter_65: class Program_weight_tensor_parameter_66: name = "parameter_66" + original_name = "layer_norm_3.b_0_deepcopy_71" shape = [256] dtype = "float32" min_val = float("-0.107568") @@ -737,6 +804,7 @@ class Program_weight_tensor_parameter_66: class Program_weight_tensor_parameter_67: name = "parameter_67" + original_name = "layer_norm_3.w_0_deepcopy_70" shape = [256] dtype = "float32" min_val = float("0.87042") @@ -748,6 +816,7 @@ class Program_weight_tensor_parameter_67: class Program_weight_tensor_parameter_68: name = "parameter_68" + original_name = "linear_7.b_0_deepcopy_69" shape = [256] dtype = "float32" min_val = float("-0.0853663") @@ -759,6 +828,7 @@ class Program_weight_tensor_parameter_68: class Program_weight_tensor_parameter_69: name = "parameter_69" + original_name = "linear_7.w_0_deepcopy_68" shape = [256, 256] dtype = "float32" min_val = float("-0.29553") @@ -770,6 +840,7 @@ class Program_weight_tensor_parameter_69: class Program_weight_tensor_parameter_70: name = "parameter_70" + original_name = "linear_5.b_0_deepcopy_65" shape = [96] dtype = "float32" min_val = float("-0.151328") @@ -781,6 +852,7 @@ class Program_weight_tensor_parameter_70: class Program_weight_tensor_parameter_71: name = "parameter_71" + original_name = "linear_5.w_0_deepcopy_64" shape = [256, 96] dtype = "float32" min_val = float("-0.502133") @@ -792,6 +864,7 @@ class Program_weight_tensor_parameter_71: class Program_weight_tensor_parameter_72: name = "parameter_72" + original_name = "linear_4.b_0_deepcopy_63" shape = [192] dtype = "float32" min_val = float("-4.03718") @@ -803,6 +876,7 @@ class Program_weight_tensor_parameter_72: class Program_weight_tensor_parameter_73: name = "parameter_73" + original_name = "linear_4.w_0_deepcopy_62" shape = [256, 192] dtype = "float32" min_val = float("-0.531345") @@ -814,6 +888,7 @@ class Program_weight_tensor_parameter_73: class Program_weight_tensor_parameter_74: name = "parameter_74" + original_name = "linear_6.b_0_deepcopy_67" shape = [256] dtype = "float32" min_val = float("-0.116264") @@ -825,6 +900,7 @@ class Program_weight_tensor_parameter_74: class Program_weight_tensor_parameter_75: name = "parameter_75" + original_name = "linear_6.w_0_deepcopy_66" shape = [256, 256] dtype = "float32" min_val = float("-0.311207") @@ -836,6 +912,7 @@ class Program_weight_tensor_parameter_75: class Program_weight_tensor_parameter_76: name = "parameter_76" + original_name = "layer_norm_2.b_0_deepcopy_61" shape = [256] dtype = "float32" min_val = float("-0.115737") @@ -847,6 +924,7 @@ class Program_weight_tensor_parameter_76: class Program_weight_tensor_parameter_77: name = "parameter_77" + original_name = "layer_norm_2.w_0_deepcopy_60" shape = [256] dtype = "float32" min_val = float("0.738236") @@ -858,6 +936,7 @@ class Program_weight_tensor_parameter_77: class Program_weight_tensor_parameter_78: name = "parameter_78" + original_name = "linear_3.b_0_deepcopy_59" shape = [256] dtype = "float32" min_val = float("-0.197148") @@ -869,6 +948,7 @@ class Program_weight_tensor_parameter_78: class Program_weight_tensor_parameter_79: name = "parameter_79" + original_name = "linear_3.w_0_deepcopy_58" shape = [256, 256] dtype = "float32" min_val = float("-0.298777") @@ -880,6 +960,7 @@ class Program_weight_tensor_parameter_79: class Program_weight_tensor_parameter_80: name = "parameter_80" + original_name = "layer_norm_4.b_0_deepcopy_55" shape = [256] dtype = "float32" min_val = float("-0.102009") @@ -891,6 +972,7 @@ class Program_weight_tensor_parameter_80: class Program_weight_tensor_parameter_81: name = "parameter_81" + original_name = "layer_norm_4.w_0_deepcopy_54" shape = [256] dtype = "float32" min_val = float("0.773911") @@ -902,6 +984,7 @@ class Program_weight_tensor_parameter_81: class Program_weight_tensor_parameter_82: name = "parameter_82" + original_name = "linear_9.b_0_deepcopy_53" shape = [256] dtype = "float32" min_val = float("-0.120821") @@ -913,6 +996,7 @@ class Program_weight_tensor_parameter_82: class Program_weight_tensor_parameter_83: name = "parameter_83" + original_name = "linear_9.w_0_deepcopy_52" shape = [1024, 256] dtype = "float32" min_val = float("-0.517246") @@ -924,6 +1008,7 @@ class Program_weight_tensor_parameter_83: class Program_weight_tensor_parameter_84: name = "parameter_84" + original_name = "linear_8.b_0_deepcopy_51" shape = [1024] dtype = "float32" min_val = float("-0.287466") @@ -935,6 +1020,7 @@ class Program_weight_tensor_parameter_84: class Program_weight_tensor_parameter_85: name = "parameter_85" + original_name = "linear_8.w_0_deepcopy_50" shape = [256, 1024] dtype = "float32" min_val = float("-0.467369") @@ -946,6 +1032,7 @@ class Program_weight_tensor_parameter_85: class Program_weight_tensor_parameter_86: name = "parameter_86" + original_name = "layer_norm_3.b_0_deepcopy_49" shape = [256] dtype = "float32" min_val = float("-0.117943") @@ -957,6 +1044,7 @@ class Program_weight_tensor_parameter_86: class Program_weight_tensor_parameter_87: name = "parameter_87" + original_name = "layer_norm_3.w_0_deepcopy_48" shape = [256] dtype = "float32" min_val = float("0.878238") @@ -968,6 +1056,7 @@ class Program_weight_tensor_parameter_87: class Program_weight_tensor_parameter_88: name = "parameter_88" + original_name = "linear_7.b_0_deepcopy_47" shape = [256] dtype = "float32" min_val = float("-0.106038") @@ -979,6 +1068,7 @@ class Program_weight_tensor_parameter_88: class Program_weight_tensor_parameter_89: name = "parameter_89" + original_name = "linear_7.w_0_deepcopy_46" shape = [256, 256] dtype = "float32" min_val = float("-0.303226") @@ -990,6 +1080,7 @@ class Program_weight_tensor_parameter_89: class Program_weight_tensor_parameter_90: name = "parameter_90" + original_name = "linear_5.b_0_deepcopy_43" shape = [96] dtype = "float32" min_val = float("-0.172097") @@ -1001,6 +1092,7 @@ class Program_weight_tensor_parameter_90: class Program_weight_tensor_parameter_91: name = "parameter_91" + original_name = "linear_5.w_0_deepcopy_42" shape = [256, 96] dtype = "float32" min_val = float("-0.606411") @@ -1012,6 +1104,7 @@ class Program_weight_tensor_parameter_91: class Program_weight_tensor_parameter_92: name = "parameter_92" + original_name = "linear_4.b_0_deepcopy_41" shape = [192] dtype = "float32" min_val = float("-4.20814") @@ -1023,6 +1116,7 @@ class Program_weight_tensor_parameter_92: class Program_weight_tensor_parameter_93: name = "parameter_93" + original_name = "linear_4.w_0_deepcopy_40" shape = [256, 192] dtype = "float32" min_val = float("-0.620158") @@ -1034,6 +1128,7 @@ class Program_weight_tensor_parameter_93: class Program_weight_tensor_parameter_94: name = "parameter_94" + original_name = "linear_6.b_0_deepcopy_45" shape = [256] dtype = "float32" min_val = float("-0.140434") @@ -1045,6 +1140,7 @@ class Program_weight_tensor_parameter_94: class Program_weight_tensor_parameter_95: name = "parameter_95" + original_name = "linear_6.w_0_deepcopy_44" shape = [256, 256] dtype = "float32" min_val = float("-0.31598") @@ -1056,6 +1152,7 @@ class Program_weight_tensor_parameter_95: class Program_weight_tensor_parameter_96: name = "parameter_96" + original_name = "layer_norm_2.b_0_deepcopy_39" shape = [256] dtype = "float32" min_val = float("-0.160636") @@ -1067,6 +1164,7 @@ class Program_weight_tensor_parameter_96: class Program_weight_tensor_parameter_97: name = "parameter_97" + original_name = "layer_norm_2.w_0_deepcopy_38" shape = [256] dtype = "float32" min_val = float("0.687008") @@ -1078,6 +1176,7 @@ class Program_weight_tensor_parameter_97: class Program_weight_tensor_parameter_98: name = "parameter_98" + original_name = "linear_3.b_0_deepcopy_37" shape = [256] dtype = "float32" min_val = float("-0.182585") @@ -1089,6 +1188,7 @@ class Program_weight_tensor_parameter_98: class Program_weight_tensor_parameter_99: name = "parameter_99" + original_name = "linear_3.w_0_deepcopy_36" shape = [256, 256] dtype = "float32" min_val = float("-0.357272") @@ -1100,6 +1200,7 @@ class Program_weight_tensor_parameter_99: class Program_weight_tensor_parameter_100: name = "parameter_100" + original_name = "layer_norm_4.b_0_deepcopy_33" shape = [256] dtype = "float32" min_val = float("-0.15818") @@ -1111,6 +1212,7 @@ class Program_weight_tensor_parameter_100: class Program_weight_tensor_parameter_101: name = "parameter_101" + original_name = "layer_norm_4.w_0_deepcopy_32" shape = [256] dtype = "float32" min_val = float("0.787488") @@ -1122,6 +1224,7 @@ class Program_weight_tensor_parameter_101: class Program_weight_tensor_parameter_102: name = "parameter_102" + original_name = "linear_9.b_0_deepcopy_31" shape = [256] dtype = "float32" min_val = float("-0.135321") @@ -1133,6 +1236,7 @@ class Program_weight_tensor_parameter_102: class Program_weight_tensor_parameter_103: name = "parameter_103" + original_name = "linear_9.w_0_deepcopy_30" shape = [1024, 256] dtype = "float32" min_val = float("-0.375069") @@ -1144,6 +1248,7 @@ class Program_weight_tensor_parameter_103: class Program_weight_tensor_parameter_104: name = "parameter_104" + original_name = "linear_8.b_0_deepcopy_29" shape = [1024] dtype = "float32" min_val = float("-0.277239") @@ -1155,6 +1260,7 @@ class Program_weight_tensor_parameter_104: class Program_weight_tensor_parameter_105: name = "parameter_105" + original_name = "linear_8.w_0_deepcopy_28" shape = [256, 1024] dtype = "float32" min_val = float("-0.351256") @@ -1166,6 +1272,7 @@ class Program_weight_tensor_parameter_105: class Program_weight_tensor_parameter_106: name = "parameter_106" + original_name = "layer_norm_3.b_0_deepcopy_27" shape = [256] dtype = "float32" min_val = float("-0.146369") @@ -1177,6 +1284,7 @@ class Program_weight_tensor_parameter_106: class Program_weight_tensor_parameter_107: name = "parameter_107" + original_name = "layer_norm_3.w_0_deepcopy_26" shape = [256] dtype = "float32" min_val = float("0.876909") @@ -1188,6 +1296,7 @@ class Program_weight_tensor_parameter_107: class Program_weight_tensor_parameter_108: name = "parameter_108" + original_name = "linear_7.b_0_deepcopy_25" shape = [256] dtype = "float32" min_val = float("-0.114928") @@ -1199,6 +1308,7 @@ class Program_weight_tensor_parameter_108: class Program_weight_tensor_parameter_109: name = "parameter_109" + original_name = "linear_7.w_0_deepcopy_24" shape = [256, 256] dtype = "float32" min_val = float("-0.297003") @@ -1210,6 +1320,7 @@ class Program_weight_tensor_parameter_109: class Program_weight_tensor_parameter_110: name = "parameter_110" + original_name = "linear_5.b_0_deepcopy_21" shape = [96] dtype = "float32" min_val = float("-0.20168") @@ -1221,6 +1332,7 @@ class Program_weight_tensor_parameter_110: class Program_weight_tensor_parameter_111: name = "parameter_111" + original_name = "linear_5.w_0_deepcopy_20" shape = [256, 96] dtype = "float32" min_val = float("-0.547941") @@ -1232,6 +1344,7 @@ class Program_weight_tensor_parameter_111: class Program_weight_tensor_parameter_112: name = "parameter_112" + original_name = "linear_4.b_0_deepcopy_19" shape = [192] dtype = "float32" min_val = float("-4.25868") @@ -1243,6 +1356,7 @@ class Program_weight_tensor_parameter_112: class Program_weight_tensor_parameter_113: name = "parameter_113" + original_name = "linear_4.w_0_deepcopy_18" shape = [256, 192] dtype = "float32" min_val = float("-0.678738") @@ -1254,6 +1368,7 @@ class Program_weight_tensor_parameter_113: class Program_weight_tensor_parameter_114: name = "parameter_114" + original_name = "linear_6.b_0_deepcopy_23" shape = [256] dtype = "float32" min_val = float("-0.259033") @@ -1265,6 +1380,7 @@ class Program_weight_tensor_parameter_114: class Program_weight_tensor_parameter_115: name = "parameter_115" + original_name = "linear_6.w_0_deepcopy_22" shape = [256, 256] dtype = "float32" min_val = float("-0.330366") @@ -1276,6 +1392,7 @@ class Program_weight_tensor_parameter_115: class Program_weight_tensor_parameter_116: name = "parameter_116" + original_name = "layer_norm_2.b_0_deepcopy_17" shape = [256] dtype = "float32" min_val = float("-0.211202") @@ -1287,6 +1404,7 @@ class Program_weight_tensor_parameter_116: class Program_weight_tensor_parameter_117: name = "parameter_117" + original_name = "layer_norm_2.w_0_deepcopy_16" shape = [256] dtype = "float32" min_val = float("0.687804") @@ -1298,6 +1416,7 @@ class Program_weight_tensor_parameter_117: class Program_weight_tensor_parameter_118: name = "parameter_118" + original_name = "linear_3.b_0_deepcopy_15" shape = [256] dtype = "float32" min_val = float("-0.270326") @@ -1309,6 +1428,7 @@ class Program_weight_tensor_parameter_118: class Program_weight_tensor_parameter_119: name = "parameter_119" + original_name = "linear_3.w_0_deepcopy_14" shape = [256, 256] dtype = "float32" min_val = float("-0.319159") @@ -1320,6 +1440,7 @@ class Program_weight_tensor_parameter_119: class Program_weight_tensor_parameter_120: name = "parameter_120" + original_name = "linear_11.b_0" shape = [256] dtype = "float32" min_val = float("-0.17368") @@ -1331,6 +1452,7 @@ class Program_weight_tensor_parameter_120: class Program_weight_tensor_parameter_121: name = "parameter_121" + original_name = "linear_11.w_0" shape = [512, 256] dtype = "float32" min_val = float("-1.72547") @@ -1342,6 +1464,7 @@ class Program_weight_tensor_parameter_121: class Program_weight_tensor_parameter_122: name = "parameter_122" + original_name = "linear_10.b_0" shape = [512] dtype = "float32" min_val = float("-0.498659") @@ -1353,6 +1476,7 @@ class Program_weight_tensor_parameter_122: class Program_weight_tensor_parameter_123: name = "parameter_123" + original_name = "linear_10.w_0" shape = [4, 512] dtype = "float32" min_val = float("-3.55189") @@ -1364,6 +1488,7 @@ class Program_weight_tensor_parameter_123: class Program_weight_tensor_parameter_124: name = "parameter_124" + original_name = "linear_14.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -1373,6 +1498,7 @@ class Program_weight_tensor_parameter_124: class Program_weight_tensor_parameter_125: name = "parameter_125" + original_name = "linear_14.w_0" shape = [256, 32] dtype = "float32" min_val = float("-0.739126") @@ -1384,6 +1510,7 @@ class Program_weight_tensor_parameter_125: class Program_weight_tensor_parameter_126: name = "parameter_126" + original_name = "linear_13.b_0" shape = [256] dtype = "float32" min_val = float("-0.144131") @@ -1395,6 +1522,7 @@ class Program_weight_tensor_parameter_126: class Program_weight_tensor_parameter_127: name = "parameter_127" + original_name = "linear_13.w_0" shape = [256, 256] dtype = "float32" min_val = float("-1.29129") @@ -1406,6 +1534,7 @@ class Program_weight_tensor_parameter_127: class Program_weight_tensor_parameter_128: name = "parameter_128" + original_name = "linear_12.b_0" shape = [256] dtype = "float32" min_val = float("-0.462414") @@ -1417,6 +1546,7 @@ class Program_weight_tensor_parameter_128: class Program_weight_tensor_parameter_129: name = "parameter_129" + original_name = "linear_12.w_0" shape = [256, 256] dtype = "float32" min_val = float("-0.377232") @@ -1428,6 +1558,7 @@ class Program_weight_tensor_parameter_129: class Program_weight_tensor_parameter_130: name = "parameter_130" + original_name = "layer_norm_6.b_0" shape = [256] dtype = "float32" min_val = float("-0.756978") @@ -1439,6 +1570,7 @@ class Program_weight_tensor_parameter_130: class Program_weight_tensor_parameter_131: name = "parameter_131" + original_name = "layer_norm_6.w_0" shape = [256] dtype = "float32" min_val = float("0.769186") @@ -1450,6 +1582,7 @@ class Program_weight_tensor_parameter_131: class Program_weight_tensor_parameter_132: name = "parameter_132" + original_name = "linear_19.b_0" shape = [4] dtype = "float32" min_val = float("0") @@ -1459,6 +1592,7 @@ class Program_weight_tensor_parameter_132: class Program_weight_tensor_parameter_133: name = "parameter_133" + original_name = "linear_19.w_0" shape = [256, 4] dtype = "float32" min_val = float("-0.571372") @@ -1470,6 +1604,7 @@ class Program_weight_tensor_parameter_133: class Program_weight_tensor_parameter_134: name = "parameter_134" + original_name = "linear_18.b_0" shape = [256] dtype = "float32" min_val = float("-0.257747") @@ -1481,6 +1616,7 @@ class Program_weight_tensor_parameter_134: class Program_weight_tensor_parameter_135: name = "parameter_135" + original_name = "linear_18.w_0" shape = [256, 256] dtype = "float32" min_val = float("-2.25347") @@ -1492,6 +1628,7 @@ class Program_weight_tensor_parameter_135: class Program_weight_tensor_parameter_136: name = "parameter_136" + original_name = "linear_17.b_0" shape = [256] dtype = "float32" min_val = float("-0.332821") @@ -1503,6 +1640,7 @@ class Program_weight_tensor_parameter_136: class Program_weight_tensor_parameter_137: name = "parameter_137" + original_name = "linear_17.w_0" shape = [256, 256] dtype = "float32" min_val = float("-0.355112") @@ -1514,6 +1652,7 @@ class Program_weight_tensor_parameter_137: class Program_weight_tensor_parameter_138: name = "parameter_138" + original_name = "linear_16.b_0" shape = [2] dtype = "float32" min_val = float("0") @@ -1523,6 +1662,7 @@ class Program_weight_tensor_parameter_138: class Program_weight_tensor_parameter_139: name = "parameter_139" + original_name = "linear_16.w_0" shape = [256, 2] dtype = "float32" min_val = float("-0.062324") @@ -1534,6 +1674,7 @@ class Program_weight_tensor_parameter_139: class Program_weight_tensor_parameter_140: name = "parameter_140" + original_name = "layer_norm_5.b_0" shape = [256] dtype = "float32" min_val = float("-0.63679") @@ -1545,6 +1686,7 @@ class Program_weight_tensor_parameter_140: class Program_weight_tensor_parameter_141: name = "parameter_141" + original_name = "layer_norm_5.w_0" shape = [256] dtype = "float32" min_val = float("0.726323") @@ -1556,6 +1698,7 @@ class Program_weight_tensor_parameter_141: class Program_weight_tensor_parameter_142: name = "parameter_142" + original_name = "linear_15.b_0" shape = [256] dtype = "float32" min_val = float("-0.560806") @@ -1567,6 +1710,7 @@ class Program_weight_tensor_parameter_142: class Program_weight_tensor_parameter_143: name = "parameter_143" + original_name = "linear_15.w_0" shape = [256, 256] dtype = "float32" min_val = float("-0.417298") @@ -1578,6 +1722,7 @@ class Program_weight_tensor_parameter_143: class Program_weight_tensor_parameter_144: name = "parameter_144" + original_name = "batch_norm2d_208.b_0" shape = [256] dtype = "float32" min_val = float("-0.452751") @@ -1589,6 +1734,7 @@ class Program_weight_tensor_parameter_144: class Program_weight_tensor_parameter_145: name = "parameter_145" + original_name = "batch_norm2d_208.w_0" shape = [256] dtype = "float32" min_val = float("0.760864") @@ -1600,6 +1746,7 @@ class Program_weight_tensor_parameter_145: class Program_weight_tensor_parameter_146: name = "parameter_146" + original_name = "batch_norm2d_208.w_2" shape = [256] dtype = "float32" min_val = float("6.99231") @@ -1611,6 +1758,7 @@ class Program_weight_tensor_parameter_146: class Program_weight_tensor_parameter_147: name = "parameter_147" + original_name = "batch_norm2d_208.w_1" shape = [256] dtype = "float32" min_val = float("-2.09351") @@ -1622,6 +1770,7 @@ class Program_weight_tensor_parameter_147: class Program_weight_tensor_parameter_148: name = "parameter_148" + original_name = "conv2d_129.w_0" shape = [256, 256, 1, 1] dtype = "float32" min_val = float("-0.342562") @@ -1633,6 +1782,7 @@ class Program_weight_tensor_parameter_148: class Program_weight_tensor_parameter_149: name = "parameter_149" + original_name = "batch_norm2d_207.b_0" shape = [256] dtype = "float32" min_val = float("-0.31344") @@ -1644,6 +1794,7 @@ class Program_weight_tensor_parameter_149: class Program_weight_tensor_parameter_150: name = "parameter_150" + original_name = "batch_norm2d_207.w_0" shape = [256] dtype = "float32" min_val = float("0.812157") @@ -1655,6 +1806,7 @@ class Program_weight_tensor_parameter_150: class Program_weight_tensor_parameter_151: name = "parameter_151" + original_name = "batch_norm2d_207.w_2" shape = [256] dtype = "float32" min_val = float("11.6006") @@ -1666,6 +1818,7 @@ class Program_weight_tensor_parameter_151: class Program_weight_tensor_parameter_152: name = "parameter_152" + original_name = "batch_norm2d_207.w_1" shape = [256] dtype = "float32" min_val = float("-2.03196") @@ -1677,6 +1830,7 @@ class Program_weight_tensor_parameter_152: class Program_weight_tensor_parameter_153: name = "parameter_153" + original_name = "conv2d_128.w_0" shape = [256, 256, 1, 1] dtype = "float32" min_val = float("-0.361381") @@ -1688,6 +1842,7 @@ class Program_weight_tensor_parameter_153: class Program_weight_tensor_parameter_154: name = "parameter_154" + original_name = "batch_norm2d_206.b_0" shape = [256] dtype = "float32" min_val = float("-0.252679") @@ -1699,6 +1854,7 @@ class Program_weight_tensor_parameter_154: class Program_weight_tensor_parameter_155: name = "parameter_155" + original_name = "batch_norm2d_206.w_0" shape = [256] dtype = "float32" min_val = float("0.725283") @@ -1710,6 +1866,7 @@ class Program_weight_tensor_parameter_155: class Program_weight_tensor_parameter_156: name = "parameter_156" + original_name = "batch_norm2d_206.w_2" shape = [256] dtype = "float32" min_val = float("13.0832") @@ -1721,6 +1878,7 @@ class Program_weight_tensor_parameter_156: class Program_weight_tensor_parameter_157: name = "parameter_157" + original_name = "batch_norm2d_206.w_1" shape = [256] dtype = "float32" min_val = float("-1.76855") @@ -1732,6 +1890,7 @@ class Program_weight_tensor_parameter_157: class Program_weight_tensor_parameter_158: name = "parameter_158" + original_name = "conv2d_127.w_0" shape = [256, 256, 1, 1] dtype = "float32" min_val = float("-0.357759") diff --git a/paddle_samples/PaddleX/Mask-RT-DETR-L/subgraph_12/input_meta.py b/paddle_samples/PaddleX/Mask-RT-DETR-L/subgraph_12/input_meta.py index 27812e84c..c16b9e305 100644 --- a/paddle_samples/PaddleX/Mask-RT-DETR-L/subgraph_12/input_meta.py +++ b/paddle_samples/PaddleX/Mask-RT-DETR-L/subgraph_12/input_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_data_0: name = "data_0" + original_name = "param_10" shape = [2, 256] dtype = "float32" min_val = float("-3.27767") @@ -11,6 +12,7 @@ class Program_weight_tensor_data_0: class Program_weight_tensor_data_1: name = "data_1" + original_name = "param_11" shape = [256, 768] dtype = "float32" min_val = float("-0.388813") @@ -22,6 +24,7 @@ class Program_weight_tensor_data_1: class Program_weight_tensor_data_2: name = "data_2" + original_name = "param_12" shape = [768] dtype = "float32" min_val = float("-0.226413") @@ -33,6 +36,7 @@ class Program_weight_tensor_data_2: class Program_weight_tensor_data_3: name = "data_3" + original_name = "param_13" shape = [256, 768] dtype = "float32" min_val = float("-0.431728") @@ -44,6 +48,7 @@ class Program_weight_tensor_data_3: class Program_weight_tensor_data_4: name = "data_4" + original_name = "param_14" shape = [768] dtype = "float32" min_val = float("-0.155516") @@ -55,6 +60,7 @@ class Program_weight_tensor_data_4: class Program_weight_tensor_data_5: name = "data_5" + original_name = "param_15" shape = [256, 768] dtype = "float32" min_val = float("-0.358802") @@ -66,6 +72,7 @@ class Program_weight_tensor_data_5: class Program_weight_tensor_data_6: name = "data_6" + original_name = "param_16" shape = [768] dtype = "float32" min_val = float("-0.119078") @@ -77,6 +84,7 @@ class Program_weight_tensor_data_6: class Program_weight_tensor_data_7: name = "data_7" + original_name = "param_17" shape = [256, 768] dtype = "float32" min_val = float("-0.372305") @@ -88,6 +96,7 @@ class Program_weight_tensor_data_7: class Program_weight_tensor_data_8: name = "data_8" + original_name = "param_18" shape = [768] dtype = "float32" min_val = float("-0.0974899") @@ -99,6 +108,7 @@ class Program_weight_tensor_data_8: class Program_weight_tensor_data_9: name = "data_9" + original_name = "param_19" shape = [256, 768] dtype = "float32" min_val = float("-0.347611") @@ -110,6 +120,7 @@ class Program_weight_tensor_data_9: class Program_weight_tensor_data_10: name = "data_10" + original_name = "param_20" shape = [768] dtype = "float32" min_val = float("-0.109375") @@ -121,6 +132,7 @@ class Program_weight_tensor_data_10: class Program_weight_tensor_data_11: name = "data_11" + original_name = "param_21" shape = [256, 768] dtype = "float32" min_val = float("-0.295065") @@ -132,6 +144,7 @@ class Program_weight_tensor_data_11: class Program_weight_tensor_data_12: name = "data_12" + original_name = "param_22" shape = [768] dtype = "float32" min_val = float("-0.0793127") @@ -143,6 +156,7 @@ class Program_weight_tensor_data_12: class Program_weight_tensor_data_13: name = "data_13" + original_name = "var_1766" shape = [2, 256, 96, 96] dtype = "float32" min_val = float("-0.556929") @@ -154,6 +168,7 @@ class Program_weight_tensor_data_13: class Program_weight_tensor_data_14: name = "data_14" + original_name = "var_1767" shape = [2, 256, 48, 48] dtype = "float32" min_val = float("-0.556928") @@ -165,6 +180,7 @@ class Program_weight_tensor_data_14: class Program_weight_tensor_data_15: name = "data_15" + original_name = "var_1768" shape = [2, 256, 24, 24] dtype = "float32" min_val = float("-0.556925") @@ -176,6 +192,7 @@ class Program_weight_tensor_data_15: class Program_weight_tensor_data_16: name = "data_16" + original_name = "var_1769" shape = [2, 32, 192, 192] dtype = "float32" min_val = float("-19.8716") @@ -187,6 +204,7 @@ class Program_weight_tensor_data_16: class Program_weight_tensor_data_17: name = "data_17" + original_name = "var_1773" shape = [1, 1] dtype = "int32" data = [0] @@ -194,6 +212,7 @@ class Program_weight_tensor_data_17: class Program_weight_tensor_data_18: name = "data_18" + original_name = "var_1774" shape = [1, 1] dtype = "int32" data = [0] @@ -201,6 +220,7 @@ class Program_weight_tensor_data_18: class Program_weight_tensor_data_19: name = "data_19" + original_name = "var_1775" shape = [1, 4] dtype = "float32" data = [0.466667, 0.5, 0.856497, 1.0] @@ -208,6 +228,7 @@ class Program_weight_tensor_data_19: class Program_weight_tensor_data_20: name = "data_20" + original_name = "var_1776" shape = [1, 4] dtype = "float32" data = [0.472296, 0.413278, 0.800401, 0.826555] diff --git a/paddle_samples/PaddleX/Mask-RT-DETR-L/subgraph_12/weight_meta.py b/paddle_samples/PaddleX/Mask-RT-DETR-L/subgraph_12/weight_meta.py index 9e0d390f0..2a0276ced 100644 --- a/paddle_samples/PaddleX/Mask-RT-DETR-L/subgraph_12/weight_meta.py +++ b/paddle_samples/PaddleX/Mask-RT-DETR-L/subgraph_12/weight_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_parameter_0: name = "parameter_0" + original_name = "layer_norm_4.b_0_deepcopy_143" shape = [256] dtype = "float32" min_val = float("-0.235319") @@ -11,6 +12,7 @@ class Program_weight_tensor_parameter_0: class Program_weight_tensor_parameter_1: name = "parameter_1" + original_name = "layer_norm_4.w_0_deepcopy_142" shape = [256] dtype = "float32" min_val = float("0.669133") @@ -22,6 +24,7 @@ class Program_weight_tensor_parameter_1: class Program_weight_tensor_parameter_2: name = "parameter_2" + original_name = "linear_9.b_0_deepcopy_141" shape = [256] dtype = "float32" min_val = float("-0.229282") @@ -33,6 +36,7 @@ class Program_weight_tensor_parameter_2: class Program_weight_tensor_parameter_3: name = "parameter_3" + original_name = "linear_9.w_0_deepcopy_140" shape = [1024, 256] dtype = "float32" min_val = float("-0.411993") @@ -44,6 +48,7 @@ class Program_weight_tensor_parameter_3: class Program_weight_tensor_parameter_4: name = "parameter_4" + original_name = "linear_8.b_0_deepcopy_139" shape = [1024] dtype = "float32" min_val = float("-0.160422") @@ -55,6 +60,7 @@ class Program_weight_tensor_parameter_4: class Program_weight_tensor_parameter_5: name = "parameter_5" + original_name = "linear_8.w_0_deepcopy_138" shape = [256, 1024] dtype = "float32" min_val = float("-0.314147") @@ -66,6 +72,7 @@ class Program_weight_tensor_parameter_5: class Program_weight_tensor_parameter_6: name = "parameter_6" + original_name = "layer_norm_3.b_0_deepcopy_137" shape = [256] dtype = "float32" min_val = float("-0.165694") @@ -77,6 +84,7 @@ class Program_weight_tensor_parameter_6: class Program_weight_tensor_parameter_7: name = "parameter_7" + original_name = "layer_norm_3.w_0_deepcopy_136" shape = [256] dtype = "float32" min_val = float("0.819601") @@ -88,6 +96,7 @@ class Program_weight_tensor_parameter_7: class Program_weight_tensor_parameter_8: name = "parameter_8" + original_name = "linear_7.b_0_deepcopy_135" shape = [256] dtype = "float32" min_val = float("-0.167098") @@ -99,6 +108,7 @@ class Program_weight_tensor_parameter_8: class Program_weight_tensor_parameter_9: name = "parameter_9" + original_name = "linear_7.w_0_deepcopy_134" shape = [256, 256] dtype = "float32" min_val = float("-0.274856") @@ -110,6 +120,7 @@ class Program_weight_tensor_parameter_9: class Program_weight_tensor_parameter_10: name = "parameter_10" + original_name = "linear_5.b_0_deepcopy_131" shape = [96] dtype = "float32" min_val = float("-0.0530414") @@ -121,6 +132,7 @@ class Program_weight_tensor_parameter_10: class Program_weight_tensor_parameter_11: name = "parameter_11" + original_name = "linear_5.w_0_deepcopy_130" shape = [256, 96] dtype = "float32" min_val = float("-0.351493") @@ -132,6 +144,7 @@ class Program_weight_tensor_parameter_11: class Program_weight_tensor_parameter_12: name = "parameter_12" + original_name = "linear_4.b_0_deepcopy_129" shape = [192] dtype = "float32" min_val = float("-4.13705") @@ -143,6 +156,7 @@ class Program_weight_tensor_parameter_12: class Program_weight_tensor_parameter_13: name = "parameter_13" + original_name = "linear_4.w_0_deepcopy_128" shape = [256, 192] dtype = "float32" min_val = float("-0.329086") @@ -154,6 +168,7 @@ class Program_weight_tensor_parameter_13: class Program_weight_tensor_parameter_14: name = "parameter_14" + original_name = "linear_6.b_0_deepcopy_133" shape = [256] dtype = "float32" min_val = float("-0.0672904") @@ -165,6 +180,7 @@ class Program_weight_tensor_parameter_14: class Program_weight_tensor_parameter_15: name = "parameter_15" + original_name = "linear_6.w_0_deepcopy_132" shape = [256, 256] dtype = "float32" min_val = float("-0.291328") @@ -176,6 +192,7 @@ class Program_weight_tensor_parameter_15: class Program_weight_tensor_parameter_16: name = "parameter_16" + original_name = "layer_norm_2.b_0_deepcopy_127" shape = [256] dtype = "float32" min_val = float("-0.11319") @@ -187,6 +204,7 @@ class Program_weight_tensor_parameter_16: class Program_weight_tensor_parameter_17: name = "parameter_17" + original_name = "layer_norm_2.w_0_deepcopy_126" shape = [256] dtype = "float32" min_val = float("0.91342") @@ -198,6 +216,7 @@ class Program_weight_tensor_parameter_17: class Program_weight_tensor_parameter_18: name = "parameter_18" + original_name = "linear_3.b_0_deepcopy_125" shape = [256] dtype = "float32" min_val = float("-0.402996") @@ -209,6 +228,7 @@ class Program_weight_tensor_parameter_18: class Program_weight_tensor_parameter_19: name = "parameter_19" + original_name = "linear_3.w_0_deepcopy_124" shape = [256, 256] dtype = "float32" min_val = float("-0.385778") @@ -220,6 +240,7 @@ class Program_weight_tensor_parameter_19: class Program_weight_tensor_parameter_20: name = "parameter_20" + original_name = "layer_norm_4.b_0_deepcopy_121" shape = [256] dtype = "float32" min_val = float("-0.242023") @@ -231,6 +252,7 @@ class Program_weight_tensor_parameter_20: class Program_weight_tensor_parameter_21: name = "parameter_21" + original_name = "layer_norm_4.w_0_deepcopy_120" shape = [256] dtype = "float32" min_val = float("0.651354") @@ -242,6 +264,7 @@ class Program_weight_tensor_parameter_21: class Program_weight_tensor_parameter_22: name = "parameter_22" + original_name = "linear_9.b_0_deepcopy_119" shape = [256] dtype = "float32" min_val = float("-0.242045") @@ -253,6 +276,7 @@ class Program_weight_tensor_parameter_22: class Program_weight_tensor_parameter_23: name = "parameter_23" + original_name = "linear_9.w_0_deepcopy_118" shape = [1024, 256] dtype = "float32" min_val = float("-0.319603") @@ -264,6 +288,7 @@ class Program_weight_tensor_parameter_23: class Program_weight_tensor_parameter_24: name = "parameter_24" + original_name = "linear_8.b_0_deepcopy_117" shape = [1024] dtype = "float32" min_val = float("-0.177691") @@ -275,6 +300,7 @@ class Program_weight_tensor_parameter_24: class Program_weight_tensor_parameter_25: name = "parameter_25" + original_name = "linear_8.w_0_deepcopy_116" shape = [256, 1024] dtype = "float32" min_val = float("-0.295095") @@ -286,6 +312,7 @@ class Program_weight_tensor_parameter_25: class Program_weight_tensor_parameter_26: name = "parameter_26" + original_name = "layer_norm_3.b_0_deepcopy_115" shape = [256] dtype = "float32" min_val = float("-0.0836085") @@ -297,6 +324,7 @@ class Program_weight_tensor_parameter_26: class Program_weight_tensor_parameter_27: name = "parameter_27" + original_name = "layer_norm_3.w_0_deepcopy_114" shape = [256] dtype = "float32" min_val = float("0.841125") @@ -308,6 +336,7 @@ class Program_weight_tensor_parameter_27: class Program_weight_tensor_parameter_28: name = "parameter_28" + original_name = "linear_7.b_0_deepcopy_113" shape = [256] dtype = "float32" min_val = float("-0.100835") @@ -319,6 +348,7 @@ class Program_weight_tensor_parameter_28: class Program_weight_tensor_parameter_29: name = "parameter_29" + original_name = "linear_7.w_0_deepcopy_112" shape = [256, 256] dtype = "float32" min_val = float("-0.258674") @@ -330,6 +360,7 @@ class Program_weight_tensor_parameter_29: class Program_weight_tensor_parameter_30: name = "parameter_30" + original_name = "linear_5.b_0_deepcopy_109" shape = [96] dtype = "float32" min_val = float("-0.113839") @@ -341,6 +372,7 @@ class Program_weight_tensor_parameter_30: class Program_weight_tensor_parameter_31: name = "parameter_31" + original_name = "linear_5.w_0_deepcopy_108" shape = [256, 96] dtype = "float32" min_val = float("-0.310425") @@ -352,6 +384,7 @@ class Program_weight_tensor_parameter_31: class Program_weight_tensor_parameter_32: name = "parameter_32" + original_name = "linear_4.b_0_deepcopy_107" shape = [192] dtype = "float32" min_val = float("-4.16001") @@ -363,6 +396,7 @@ class Program_weight_tensor_parameter_32: class Program_weight_tensor_parameter_33: name = "parameter_33" + original_name = "linear_4.w_0_deepcopy_106" shape = [256, 192] dtype = "float32" min_val = float("-0.506003") @@ -374,6 +408,7 @@ class Program_weight_tensor_parameter_33: class Program_weight_tensor_parameter_34: name = "parameter_34" + original_name = "linear_6.b_0_deepcopy_111" shape = [256] dtype = "float32" min_val = float("-0.0981878") @@ -385,6 +420,7 @@ class Program_weight_tensor_parameter_34: class Program_weight_tensor_parameter_35: name = "parameter_35" + original_name = "linear_6.w_0_deepcopy_110" shape = [256, 256] dtype = "float32" min_val = float("-0.305591") @@ -396,6 +432,7 @@ class Program_weight_tensor_parameter_35: class Program_weight_tensor_parameter_36: name = "parameter_36" + original_name = "layer_norm_2.b_0_deepcopy_105" shape = [256] dtype = "float32" min_val = float("-0.113256") @@ -407,6 +444,7 @@ class Program_weight_tensor_parameter_36: class Program_weight_tensor_parameter_37: name = "parameter_37" + original_name = "layer_norm_2.w_0_deepcopy_104" shape = [256] dtype = "float32" min_val = float("0.809154") @@ -418,6 +456,7 @@ class Program_weight_tensor_parameter_37: class Program_weight_tensor_parameter_38: name = "parameter_38" + original_name = "linear_3.b_0_deepcopy_103" shape = [256] dtype = "float32" min_val = float("-0.148367") @@ -429,6 +468,7 @@ class Program_weight_tensor_parameter_38: class Program_weight_tensor_parameter_39: name = "parameter_39" + original_name = "linear_3.w_0_deepcopy_102" shape = [256, 256] dtype = "float32" min_val = float("-0.275028") @@ -440,6 +480,7 @@ class Program_weight_tensor_parameter_39: class Program_weight_tensor_parameter_40: name = "parameter_40" + original_name = "layer_norm_4.b_0_deepcopy_99" shape = [256] dtype = "float32" min_val = float("-0.152569") @@ -451,6 +492,7 @@ class Program_weight_tensor_parameter_40: class Program_weight_tensor_parameter_41: name = "parameter_41" + original_name = "layer_norm_4.w_0_deepcopy_98" shape = [256] dtype = "float32" min_val = float("0.71104") @@ -462,6 +504,7 @@ class Program_weight_tensor_parameter_41: class Program_weight_tensor_parameter_42: name = "parameter_42" + original_name = "linear_9.b_0_deepcopy_97" shape = [256] dtype = "float32" min_val = float("-0.145011") @@ -473,6 +516,7 @@ class Program_weight_tensor_parameter_42: class Program_weight_tensor_parameter_43: name = "parameter_43" + original_name = "linear_9.w_0_deepcopy_96" shape = [1024, 256] dtype = "float32" min_val = float("-0.340016") @@ -484,6 +528,7 @@ class Program_weight_tensor_parameter_43: class Program_weight_tensor_parameter_44: name = "parameter_44" + original_name = "linear_8.b_0_deepcopy_95" shape = [1024] dtype = "float32" min_val = float("-0.194362") @@ -495,6 +540,7 @@ class Program_weight_tensor_parameter_44: class Program_weight_tensor_parameter_45: name = "parameter_45" + original_name = "linear_8.w_0_deepcopy_94" shape = [256, 1024] dtype = "float32" min_val = float("-0.32099") @@ -506,6 +552,7 @@ class Program_weight_tensor_parameter_45: class Program_weight_tensor_parameter_46: name = "parameter_46" + original_name = "layer_norm_3.b_0_deepcopy_93" shape = [256] dtype = "float32" min_val = float("-0.105715") @@ -517,6 +564,7 @@ class Program_weight_tensor_parameter_46: class Program_weight_tensor_parameter_47: name = "parameter_47" + original_name = "layer_norm_3.w_0_deepcopy_92" shape = [256] dtype = "float32" min_val = float("0.875368") @@ -528,6 +576,7 @@ class Program_weight_tensor_parameter_47: class Program_weight_tensor_parameter_48: name = "parameter_48" + original_name = "linear_7.b_0_deepcopy_91" shape = [256] dtype = "float32" min_val = float("-0.125952") @@ -539,6 +588,7 @@ class Program_weight_tensor_parameter_48: class Program_weight_tensor_parameter_49: name = "parameter_49" + original_name = "linear_7.w_0_deepcopy_90" shape = [256, 256] dtype = "float32" min_val = float("-0.280394") @@ -550,6 +600,7 @@ class Program_weight_tensor_parameter_49: class Program_weight_tensor_parameter_50: name = "parameter_50" + original_name = "linear_5.b_0_deepcopy_87" shape = [96] dtype = "float32" min_val = float("-0.102302") @@ -561,6 +612,7 @@ class Program_weight_tensor_parameter_50: class Program_weight_tensor_parameter_51: name = "parameter_51" + original_name = "linear_5.w_0_deepcopy_86" shape = [256, 96] dtype = "float32" min_val = float("-0.343369") @@ -572,6 +624,7 @@ class Program_weight_tensor_parameter_51: class Program_weight_tensor_parameter_52: name = "parameter_52" + original_name = "linear_4.b_0_deepcopy_85" shape = [192] dtype = "float32" min_val = float("-4.23366") @@ -583,6 +636,7 @@ class Program_weight_tensor_parameter_52: class Program_weight_tensor_parameter_53: name = "parameter_53" + original_name = "linear_4.w_0_deepcopy_84" shape = [256, 192] dtype = "float32" min_val = float("-0.43715") @@ -594,6 +648,7 @@ class Program_weight_tensor_parameter_53: class Program_weight_tensor_parameter_54: name = "parameter_54" + original_name = "linear_6.b_0_deepcopy_89" shape = [256] dtype = "float32" min_val = float("-0.0959492") @@ -605,6 +660,7 @@ class Program_weight_tensor_parameter_54: class Program_weight_tensor_parameter_55: name = "parameter_55" + original_name = "linear_6.w_0_deepcopy_88" shape = [256, 256] dtype = "float32" min_val = float("-0.312213") @@ -616,6 +672,7 @@ class Program_weight_tensor_parameter_55: class Program_weight_tensor_parameter_56: name = "parameter_56" + original_name = "layer_norm_2.b_0_deepcopy_83" shape = [256] dtype = "float32" min_val = float("-0.0998809") @@ -627,6 +684,7 @@ class Program_weight_tensor_parameter_56: class Program_weight_tensor_parameter_57: name = "parameter_57" + original_name = "layer_norm_2.w_0_deepcopy_82" shape = [256] dtype = "float32" min_val = float("0.805495") @@ -638,6 +696,7 @@ class Program_weight_tensor_parameter_57: class Program_weight_tensor_parameter_58: name = "parameter_58" + original_name = "linear_3.b_0_deepcopy_81" shape = [256] dtype = "float32" min_val = float("-0.1321") @@ -649,6 +708,7 @@ class Program_weight_tensor_parameter_58: class Program_weight_tensor_parameter_59: name = "parameter_59" + original_name = "linear_3.w_0_deepcopy_80" shape = [256, 256] dtype = "float32" min_val = float("-0.356365") @@ -660,6 +720,7 @@ class Program_weight_tensor_parameter_59: class Program_weight_tensor_parameter_60: name = "parameter_60" + original_name = "layer_norm_4.b_0_deepcopy_77" shape = [256] dtype = "float32" min_val = float("-0.0918231") @@ -671,6 +732,7 @@ class Program_weight_tensor_parameter_60: class Program_weight_tensor_parameter_61: name = "parameter_61" + original_name = "layer_norm_4.w_0_deepcopy_76" shape = [256] dtype = "float32" min_val = float("0.763869") @@ -682,6 +744,7 @@ class Program_weight_tensor_parameter_61: class Program_weight_tensor_parameter_62: name = "parameter_62" + original_name = "linear_9.b_0_deepcopy_75" shape = [256] dtype = "float32" min_val = float("-0.11669") @@ -693,6 +756,7 @@ class Program_weight_tensor_parameter_62: class Program_weight_tensor_parameter_63: name = "parameter_63" + original_name = "linear_9.w_0_deepcopy_74" shape = [1024, 256] dtype = "float32" min_val = float("-0.4245") @@ -704,6 +768,7 @@ class Program_weight_tensor_parameter_63: class Program_weight_tensor_parameter_64: name = "parameter_64" + original_name = "linear_8.b_0_deepcopy_73" shape = [1024] dtype = "float32" min_val = float("-0.265827") @@ -715,6 +780,7 @@ class Program_weight_tensor_parameter_64: class Program_weight_tensor_parameter_65: name = "parameter_65" + original_name = "linear_8.w_0_deepcopy_72" shape = [256, 1024] dtype = "float32" min_val = float("-0.349265") @@ -726,6 +792,7 @@ class Program_weight_tensor_parameter_65: class Program_weight_tensor_parameter_66: name = "parameter_66" + original_name = "layer_norm_3.b_0_deepcopy_71" shape = [256] dtype = "float32" min_val = float("-0.107568") @@ -737,6 +804,7 @@ class Program_weight_tensor_parameter_66: class Program_weight_tensor_parameter_67: name = "parameter_67" + original_name = "layer_norm_3.w_0_deepcopy_70" shape = [256] dtype = "float32" min_val = float("0.87042") @@ -748,6 +816,7 @@ class Program_weight_tensor_parameter_67: class Program_weight_tensor_parameter_68: name = "parameter_68" + original_name = "linear_7.b_0_deepcopy_69" shape = [256] dtype = "float32" min_val = float("-0.0853662") @@ -759,6 +828,7 @@ class Program_weight_tensor_parameter_68: class Program_weight_tensor_parameter_69: name = "parameter_69" + original_name = "linear_7.w_0_deepcopy_68" shape = [256, 256] dtype = "float32" min_val = float("-0.29553") @@ -770,6 +840,7 @@ class Program_weight_tensor_parameter_69: class Program_weight_tensor_parameter_70: name = "parameter_70" + original_name = "linear_5.b_0_deepcopy_65" shape = [96] dtype = "float32" min_val = float("-0.151328") @@ -781,6 +852,7 @@ class Program_weight_tensor_parameter_70: class Program_weight_tensor_parameter_71: name = "parameter_71" + original_name = "linear_5.w_0_deepcopy_64" shape = [256, 96] dtype = "float32" min_val = float("-0.502133") @@ -792,6 +864,7 @@ class Program_weight_tensor_parameter_71: class Program_weight_tensor_parameter_72: name = "parameter_72" + original_name = "linear_4.b_0_deepcopy_63" shape = [192] dtype = "float32" min_val = float("-4.03718") @@ -803,6 +876,7 @@ class Program_weight_tensor_parameter_72: class Program_weight_tensor_parameter_73: name = "parameter_73" + original_name = "linear_4.w_0_deepcopy_62" shape = [256, 192] dtype = "float32" min_val = float("-0.531345") @@ -814,6 +888,7 @@ class Program_weight_tensor_parameter_73: class Program_weight_tensor_parameter_74: name = "parameter_74" + original_name = "linear_6.b_0_deepcopy_67" shape = [256] dtype = "float32" min_val = float("-0.116263") @@ -825,6 +900,7 @@ class Program_weight_tensor_parameter_74: class Program_weight_tensor_parameter_75: name = "parameter_75" + original_name = "linear_6.w_0_deepcopy_66" shape = [256, 256] dtype = "float32" min_val = float("-0.311206") @@ -836,6 +912,7 @@ class Program_weight_tensor_parameter_75: class Program_weight_tensor_parameter_76: name = "parameter_76" + original_name = "layer_norm_2.b_0_deepcopy_61" shape = [256] dtype = "float32" min_val = float("-0.115736") @@ -847,6 +924,7 @@ class Program_weight_tensor_parameter_76: class Program_weight_tensor_parameter_77: name = "parameter_77" + original_name = "layer_norm_2.w_0_deepcopy_60" shape = [256] dtype = "float32" min_val = float("0.738237") @@ -858,6 +936,7 @@ class Program_weight_tensor_parameter_77: class Program_weight_tensor_parameter_78: name = "parameter_78" + original_name = "linear_3.b_0_deepcopy_59" shape = [256] dtype = "float32" min_val = float("-0.197148") @@ -869,6 +948,7 @@ class Program_weight_tensor_parameter_78: class Program_weight_tensor_parameter_79: name = "parameter_79" + original_name = "linear_3.w_0_deepcopy_58" shape = [256, 256] dtype = "float32" min_val = float("-0.298777") @@ -880,6 +960,7 @@ class Program_weight_tensor_parameter_79: class Program_weight_tensor_parameter_80: name = "parameter_80" + original_name = "layer_norm_4.b_0_deepcopy_55" shape = [256] dtype = "float32" min_val = float("-0.102009") @@ -891,6 +972,7 @@ class Program_weight_tensor_parameter_80: class Program_weight_tensor_parameter_81: name = "parameter_81" + original_name = "layer_norm_4.w_0_deepcopy_54" shape = [256] dtype = "float32" min_val = float("0.773911") @@ -902,6 +984,7 @@ class Program_weight_tensor_parameter_81: class Program_weight_tensor_parameter_82: name = "parameter_82" + original_name = "linear_9.b_0_deepcopy_53" shape = [256] dtype = "float32" min_val = float("-0.120821") @@ -913,6 +996,7 @@ class Program_weight_tensor_parameter_82: class Program_weight_tensor_parameter_83: name = "parameter_83" + original_name = "linear_9.w_0_deepcopy_52" shape = [1024, 256] dtype = "float32" min_val = float("-0.517246") @@ -924,6 +1008,7 @@ class Program_weight_tensor_parameter_83: class Program_weight_tensor_parameter_84: name = "parameter_84" + original_name = "linear_8.b_0_deepcopy_51" shape = [1024] dtype = "float32" min_val = float("-0.287466") @@ -935,6 +1020,7 @@ class Program_weight_tensor_parameter_84: class Program_weight_tensor_parameter_85: name = "parameter_85" + original_name = "linear_8.w_0_deepcopy_50" shape = [256, 1024] dtype = "float32" min_val = float("-0.467369") @@ -946,6 +1032,7 @@ class Program_weight_tensor_parameter_85: class Program_weight_tensor_parameter_86: name = "parameter_86" + original_name = "layer_norm_3.b_0_deepcopy_49" shape = [256] dtype = "float32" min_val = float("-0.117943") @@ -957,6 +1044,7 @@ class Program_weight_tensor_parameter_86: class Program_weight_tensor_parameter_87: name = "parameter_87" + original_name = "layer_norm_3.w_0_deepcopy_48" shape = [256] dtype = "float32" min_val = float("0.878238") @@ -968,6 +1056,7 @@ class Program_weight_tensor_parameter_87: class Program_weight_tensor_parameter_88: name = "parameter_88" + original_name = "linear_7.b_0_deepcopy_47" shape = [256] dtype = "float32" min_val = float("-0.106038") @@ -979,6 +1068,7 @@ class Program_weight_tensor_parameter_88: class Program_weight_tensor_parameter_89: name = "parameter_89" + original_name = "linear_7.w_0_deepcopy_46" shape = [256, 256] dtype = "float32" min_val = float("-0.303226") @@ -990,6 +1080,7 @@ class Program_weight_tensor_parameter_89: class Program_weight_tensor_parameter_90: name = "parameter_90" + original_name = "linear_5.b_0_deepcopy_43" shape = [96] dtype = "float32" min_val = float("-0.172098") @@ -1001,6 +1092,7 @@ class Program_weight_tensor_parameter_90: class Program_weight_tensor_parameter_91: name = "parameter_91" + original_name = "linear_5.w_0_deepcopy_42" shape = [256, 96] dtype = "float32" min_val = float("-0.606411") @@ -1012,6 +1104,7 @@ class Program_weight_tensor_parameter_91: class Program_weight_tensor_parameter_92: name = "parameter_92" + original_name = "linear_4.b_0_deepcopy_41" shape = [192] dtype = "float32" min_val = float("-4.20814") @@ -1023,6 +1116,7 @@ class Program_weight_tensor_parameter_92: class Program_weight_tensor_parameter_93: name = "parameter_93" + original_name = "linear_4.w_0_deepcopy_40" shape = [256, 192] dtype = "float32" min_val = float("-0.620158") @@ -1034,6 +1128,7 @@ class Program_weight_tensor_parameter_93: class Program_weight_tensor_parameter_94: name = "parameter_94" + original_name = "linear_6.b_0_deepcopy_45" shape = [256] dtype = "float32" min_val = float("-0.140434") @@ -1045,6 +1140,7 @@ class Program_weight_tensor_parameter_94: class Program_weight_tensor_parameter_95: name = "parameter_95" + original_name = "linear_6.w_0_deepcopy_44" shape = [256, 256] dtype = "float32" min_val = float("-0.31598") @@ -1056,6 +1152,7 @@ class Program_weight_tensor_parameter_95: class Program_weight_tensor_parameter_96: name = "parameter_96" + original_name = "layer_norm_2.b_0_deepcopy_39" shape = [256] dtype = "float32" min_val = float("-0.160636") @@ -1067,6 +1164,7 @@ class Program_weight_tensor_parameter_96: class Program_weight_tensor_parameter_97: name = "parameter_97" + original_name = "layer_norm_2.w_0_deepcopy_38" shape = [256] dtype = "float32" min_val = float("0.687008") @@ -1078,6 +1176,7 @@ class Program_weight_tensor_parameter_97: class Program_weight_tensor_parameter_98: name = "parameter_98" + original_name = "linear_3.b_0_deepcopy_37" shape = [256] dtype = "float32" min_val = float("-0.182585") @@ -1089,6 +1188,7 @@ class Program_weight_tensor_parameter_98: class Program_weight_tensor_parameter_99: name = "parameter_99" + original_name = "linear_3.w_0_deepcopy_36" shape = [256, 256] dtype = "float32" min_val = float("-0.357272") @@ -1100,6 +1200,7 @@ class Program_weight_tensor_parameter_99: class Program_weight_tensor_parameter_100: name = "parameter_100" + original_name = "layer_norm_4.b_0_deepcopy_33" shape = [256] dtype = "float32" min_val = float("-0.15818") @@ -1111,6 +1212,7 @@ class Program_weight_tensor_parameter_100: class Program_weight_tensor_parameter_101: name = "parameter_101" + original_name = "layer_norm_4.w_0_deepcopy_32" shape = [256] dtype = "float32" min_val = float("0.787488") @@ -1122,6 +1224,7 @@ class Program_weight_tensor_parameter_101: class Program_weight_tensor_parameter_102: name = "parameter_102" + original_name = "linear_9.b_0_deepcopy_31" shape = [256] dtype = "float32" min_val = float("-0.135321") @@ -1133,6 +1236,7 @@ class Program_weight_tensor_parameter_102: class Program_weight_tensor_parameter_103: name = "parameter_103" + original_name = "linear_9.w_0_deepcopy_30" shape = [1024, 256] dtype = "float32" min_val = float("-0.375068") @@ -1144,6 +1248,7 @@ class Program_weight_tensor_parameter_103: class Program_weight_tensor_parameter_104: name = "parameter_104" + original_name = "linear_8.b_0_deepcopy_29" shape = [1024] dtype = "float32" min_val = float("-0.277239") @@ -1155,6 +1260,7 @@ class Program_weight_tensor_parameter_104: class Program_weight_tensor_parameter_105: name = "parameter_105" + original_name = "linear_8.w_0_deepcopy_28" shape = [256, 1024] dtype = "float32" min_val = float("-0.351256") @@ -1166,6 +1272,7 @@ class Program_weight_tensor_parameter_105: class Program_weight_tensor_parameter_106: name = "parameter_106" + original_name = "layer_norm_3.b_0_deepcopy_27" shape = [256] dtype = "float32" min_val = float("-0.146369") @@ -1177,6 +1284,7 @@ class Program_weight_tensor_parameter_106: class Program_weight_tensor_parameter_107: name = "parameter_107" + original_name = "layer_norm_3.w_0_deepcopy_26" shape = [256] dtype = "float32" min_val = float("0.876909") @@ -1188,6 +1296,7 @@ class Program_weight_tensor_parameter_107: class Program_weight_tensor_parameter_108: name = "parameter_108" + original_name = "linear_7.b_0_deepcopy_25" shape = [256] dtype = "float32" min_val = float("-0.114928") @@ -1199,6 +1308,7 @@ class Program_weight_tensor_parameter_108: class Program_weight_tensor_parameter_109: name = "parameter_109" + original_name = "linear_7.w_0_deepcopy_24" shape = [256, 256] dtype = "float32" min_val = float("-0.297003") @@ -1210,6 +1320,7 @@ class Program_weight_tensor_parameter_109: class Program_weight_tensor_parameter_110: name = "parameter_110" + original_name = "linear_5.b_0_deepcopy_21" shape = [96] dtype = "float32" min_val = float("-0.201679") @@ -1221,6 +1332,7 @@ class Program_weight_tensor_parameter_110: class Program_weight_tensor_parameter_111: name = "parameter_111" + original_name = "linear_5.w_0_deepcopy_20" shape = [256, 96] dtype = "float32" min_val = float("-0.547941") @@ -1232,6 +1344,7 @@ class Program_weight_tensor_parameter_111: class Program_weight_tensor_parameter_112: name = "parameter_112" + original_name = "linear_4.b_0_deepcopy_19" shape = [192] dtype = "float32" min_val = float("-4.25868") @@ -1243,6 +1356,7 @@ class Program_weight_tensor_parameter_112: class Program_weight_tensor_parameter_113: name = "parameter_113" + original_name = "linear_4.w_0_deepcopy_18" shape = [256, 192] dtype = "float32" min_val = float("-0.678738") @@ -1254,6 +1368,7 @@ class Program_weight_tensor_parameter_113: class Program_weight_tensor_parameter_114: name = "parameter_114" + original_name = "linear_6.b_0_deepcopy_23" shape = [256] dtype = "float32" min_val = float("-0.259034") @@ -1265,6 +1380,7 @@ class Program_weight_tensor_parameter_114: class Program_weight_tensor_parameter_115: name = "parameter_115" + original_name = "linear_6.w_0_deepcopy_22" shape = [256, 256] dtype = "float32" min_val = float("-0.330366") @@ -1276,6 +1392,7 @@ class Program_weight_tensor_parameter_115: class Program_weight_tensor_parameter_116: name = "parameter_116" + original_name = "layer_norm_2.b_0_deepcopy_17" shape = [256] dtype = "float32" min_val = float("-0.211202") @@ -1287,6 +1404,7 @@ class Program_weight_tensor_parameter_116: class Program_weight_tensor_parameter_117: name = "parameter_117" + original_name = "layer_norm_2.w_0_deepcopy_16" shape = [256] dtype = "float32" min_val = float("0.687804") @@ -1298,6 +1416,7 @@ class Program_weight_tensor_parameter_117: class Program_weight_tensor_parameter_118: name = "parameter_118" + original_name = "linear_3.b_0_deepcopy_15" shape = [256] dtype = "float32" min_val = float("-0.270326") @@ -1309,6 +1428,7 @@ class Program_weight_tensor_parameter_118: class Program_weight_tensor_parameter_119: name = "parameter_119" + original_name = "linear_3.w_0_deepcopy_14" shape = [256, 256] dtype = "float32" min_val = float("-0.319159") @@ -1320,6 +1440,7 @@ class Program_weight_tensor_parameter_119: class Program_weight_tensor_parameter_120: name = "parameter_120" + original_name = "linear_11.b_0" shape = [256] dtype = "float32" min_val = float("-0.17368") @@ -1331,6 +1452,7 @@ class Program_weight_tensor_parameter_120: class Program_weight_tensor_parameter_121: name = "parameter_121" + original_name = "linear_11.w_0" shape = [512, 256] dtype = "float32" min_val = float("-1.72547") @@ -1342,6 +1464,7 @@ class Program_weight_tensor_parameter_121: class Program_weight_tensor_parameter_122: name = "parameter_122" + original_name = "linear_10.b_0" shape = [512] dtype = "float32" min_val = float("-0.498659") @@ -1353,6 +1476,7 @@ class Program_weight_tensor_parameter_122: class Program_weight_tensor_parameter_123: name = "parameter_123" + original_name = "linear_10.w_0" shape = [4, 512] dtype = "float32" min_val = float("-3.55189") @@ -1364,6 +1488,7 @@ class Program_weight_tensor_parameter_123: class Program_weight_tensor_parameter_124: name = "parameter_124" + original_name = "linear_14.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -1373,6 +1498,7 @@ class Program_weight_tensor_parameter_124: class Program_weight_tensor_parameter_125: name = "parameter_125" + original_name = "linear_14.w_0" shape = [256, 32] dtype = "float32" min_val = float("-0.739126") @@ -1384,6 +1510,7 @@ class Program_weight_tensor_parameter_125: class Program_weight_tensor_parameter_126: name = "parameter_126" + original_name = "linear_13.b_0" shape = [256] dtype = "float32" min_val = float("-0.144131") @@ -1395,6 +1522,7 @@ class Program_weight_tensor_parameter_126: class Program_weight_tensor_parameter_127: name = "parameter_127" + original_name = "linear_13.w_0" shape = [256, 256] dtype = "float32" min_val = float("-1.29129") @@ -1406,6 +1534,7 @@ class Program_weight_tensor_parameter_127: class Program_weight_tensor_parameter_128: name = "parameter_128" + original_name = "linear_12.b_0" shape = [256] dtype = "float32" min_val = float("-0.462414") @@ -1417,6 +1546,7 @@ class Program_weight_tensor_parameter_128: class Program_weight_tensor_parameter_129: name = "parameter_129" + original_name = "linear_12.w_0" shape = [256, 256] dtype = "float32" min_val = float("-0.377233") @@ -1428,6 +1558,7 @@ class Program_weight_tensor_parameter_129: class Program_weight_tensor_parameter_130: name = "parameter_130" + original_name = "layer_norm_6.b_0" shape = [256] dtype = "float32" min_val = float("-0.756979") @@ -1439,6 +1570,7 @@ class Program_weight_tensor_parameter_130: class Program_weight_tensor_parameter_131: name = "parameter_131" + original_name = "layer_norm_6.w_0" shape = [256] dtype = "float32" min_val = float("0.769186") @@ -1450,6 +1582,7 @@ class Program_weight_tensor_parameter_131: class Program_weight_tensor_parameter_132: name = "parameter_132" + original_name = "linear_19.b_0" shape = [4] dtype = "float32" min_val = float("0") @@ -1459,6 +1592,7 @@ class Program_weight_tensor_parameter_132: class Program_weight_tensor_parameter_133: name = "parameter_133" + original_name = "linear_19.w_0" shape = [256, 4] dtype = "float32" min_val = float("-0.571372") @@ -1470,6 +1604,7 @@ class Program_weight_tensor_parameter_133: class Program_weight_tensor_parameter_134: name = "parameter_134" + original_name = "linear_18.b_0" shape = [256] dtype = "float32" min_val = float("-0.257747") @@ -1481,6 +1616,7 @@ class Program_weight_tensor_parameter_134: class Program_weight_tensor_parameter_135: name = "parameter_135" + original_name = "linear_18.w_0" shape = [256, 256] dtype = "float32" min_val = float("-2.25347") @@ -1492,6 +1628,7 @@ class Program_weight_tensor_parameter_135: class Program_weight_tensor_parameter_136: name = "parameter_136" + original_name = "linear_17.b_0" shape = [256] dtype = "float32" min_val = float("-0.332821") @@ -1503,6 +1640,7 @@ class Program_weight_tensor_parameter_136: class Program_weight_tensor_parameter_137: name = "parameter_137" + original_name = "linear_17.w_0" shape = [256, 256] dtype = "float32" min_val = float("-0.355112") @@ -1514,6 +1652,7 @@ class Program_weight_tensor_parameter_137: class Program_weight_tensor_parameter_138: name = "parameter_138" + original_name = "linear_16.b_0" shape = [2] dtype = "float32" min_val = float("0") @@ -1523,6 +1662,7 @@ class Program_weight_tensor_parameter_138: class Program_weight_tensor_parameter_139: name = "parameter_139" + original_name = "linear_16.w_0" shape = [256, 2] dtype = "float32" min_val = float("-0.0623241") @@ -1534,6 +1674,7 @@ class Program_weight_tensor_parameter_139: class Program_weight_tensor_parameter_140: name = "parameter_140" + original_name = "layer_norm_5.b_0" shape = [256] dtype = "float32" min_val = float("-0.63679") @@ -1545,6 +1686,7 @@ class Program_weight_tensor_parameter_140: class Program_weight_tensor_parameter_141: name = "parameter_141" + original_name = "layer_norm_5.w_0" shape = [256] dtype = "float32" min_val = float("0.726323") @@ -1556,6 +1698,7 @@ class Program_weight_tensor_parameter_141: class Program_weight_tensor_parameter_142: name = "parameter_142" + original_name = "linear_15.b_0" shape = [256] dtype = "float32" min_val = float("-0.560806") @@ -1567,6 +1710,7 @@ class Program_weight_tensor_parameter_142: class Program_weight_tensor_parameter_143: name = "parameter_143" + original_name = "linear_15.w_0" shape = [256, 256] dtype = "float32" min_val = float("-0.417298") @@ -1578,6 +1722,7 @@ class Program_weight_tensor_parameter_143: class Program_weight_tensor_parameter_144: name = "parameter_144" + original_name = "batch_norm2d_208.b_0" shape = [256] dtype = "float32" min_val = float("-0.452751") @@ -1589,6 +1734,7 @@ class Program_weight_tensor_parameter_144: class Program_weight_tensor_parameter_145: name = "parameter_145" + original_name = "batch_norm2d_208.w_0" shape = [256] dtype = "float32" min_val = float("0.760864") @@ -1600,6 +1746,7 @@ class Program_weight_tensor_parameter_145: class Program_weight_tensor_parameter_146: name = "parameter_146" + original_name = "batch_norm2d_208.w_2" shape = [256] dtype = "float32" min_val = float("7.20578") @@ -1611,6 +1758,7 @@ class Program_weight_tensor_parameter_146: class Program_weight_tensor_parameter_147: name = "parameter_147" + original_name = "batch_norm2d_208.w_1" shape = [256] dtype = "float32" min_val = float("-2.09419") @@ -1622,6 +1770,7 @@ class Program_weight_tensor_parameter_147: class Program_weight_tensor_parameter_148: name = "parameter_148" + original_name = "conv2d_129.w_0" shape = [256, 256, 1, 1] dtype = "float32" min_val = float("-0.342562") @@ -1633,6 +1782,7 @@ class Program_weight_tensor_parameter_148: class Program_weight_tensor_parameter_149: name = "parameter_149" + original_name = "batch_norm2d_207.b_0" shape = [256] dtype = "float32" min_val = float("-0.31344") @@ -1644,6 +1794,7 @@ class Program_weight_tensor_parameter_149: class Program_weight_tensor_parameter_150: name = "parameter_150" + original_name = "batch_norm2d_207.w_0" shape = [256] dtype = "float32" min_val = float("0.812157") @@ -1655,6 +1806,7 @@ class Program_weight_tensor_parameter_150: class Program_weight_tensor_parameter_151: name = "parameter_151" + original_name = "batch_norm2d_207.w_2" shape = [256] dtype = "float32" min_val = float("11.7728") @@ -1666,6 +1818,7 @@ class Program_weight_tensor_parameter_151: class Program_weight_tensor_parameter_152: name = "parameter_152" + original_name = "batch_norm2d_207.w_1" shape = [256] dtype = "float32" min_val = float("-2.01368") @@ -1677,6 +1830,7 @@ class Program_weight_tensor_parameter_152: class Program_weight_tensor_parameter_153: name = "parameter_153" + original_name = "conv2d_128.w_0" shape = [256, 256, 1, 1] dtype = "float32" min_val = float("-0.361381") @@ -1688,6 +1842,7 @@ class Program_weight_tensor_parameter_153: class Program_weight_tensor_parameter_154: name = "parameter_154" + original_name = "batch_norm2d_206.b_0" shape = [256] dtype = "float32" min_val = float("-0.252679") @@ -1699,6 +1854,7 @@ class Program_weight_tensor_parameter_154: class Program_weight_tensor_parameter_155: name = "parameter_155" + original_name = "batch_norm2d_206.w_0" shape = [256] dtype = "float32" min_val = float("0.725283") @@ -1710,6 +1866,7 @@ class Program_weight_tensor_parameter_155: class Program_weight_tensor_parameter_156: name = "parameter_156" + original_name = "batch_norm2d_206.w_2" shape = [256] dtype = "float32" min_val = float("13.4483") @@ -1721,6 +1878,7 @@ class Program_weight_tensor_parameter_156: class Program_weight_tensor_parameter_157: name = "parameter_157" + original_name = "batch_norm2d_206.w_1" shape = [256] dtype = "float32" min_val = float("-1.7452") @@ -1732,6 +1890,7 @@ class Program_weight_tensor_parameter_157: class Program_weight_tensor_parameter_158: name = "parameter_158" + original_name = "conv2d_127.w_0" shape = [256, 256, 1, 1] dtype = "float32" min_val = float("-0.357759") diff --git a/paddle_samples/PaddleX/Mask-RT-DETR-L/subgraph_18/input_meta.py b/paddle_samples/PaddleX/Mask-RT-DETR-L/subgraph_18/input_meta.py index f845b9d34..a925498ce 100644 --- a/paddle_samples/PaddleX/Mask-RT-DETR-L/subgraph_18/input_meta.py +++ b/paddle_samples/PaddleX/Mask-RT-DETR-L/subgraph_18/input_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_data_0: name = "data_0" + original_name = "param_60" shape = [2, 256] dtype = "float32" min_val = float("-3.27765") @@ -11,6 +12,7 @@ class Program_weight_tensor_data_0: class Program_weight_tensor_data_1: name = "data_1" + original_name = "param_61" shape = [256, 768] dtype = "float32" min_val = float("-0.388823") @@ -22,6 +24,7 @@ class Program_weight_tensor_data_1: class Program_weight_tensor_data_2: name = "data_2" + original_name = "param_62" shape = [768] dtype = "float32" min_val = float("-0.226414") @@ -33,6 +36,7 @@ class Program_weight_tensor_data_2: class Program_weight_tensor_data_3: name = "data_3" + original_name = "param_63" shape = [256, 768] dtype = "float32" min_val = float("-0.431741") @@ -44,6 +48,7 @@ class Program_weight_tensor_data_3: class Program_weight_tensor_data_4: name = "data_4" + original_name = "param_64" shape = [768] dtype = "float32" min_val = float("-0.155515") @@ -55,6 +60,7 @@ class Program_weight_tensor_data_4: class Program_weight_tensor_data_5: name = "data_5" + original_name = "param_65" shape = [256, 768] dtype = "float32" min_val = float("-0.358807") @@ -66,6 +72,7 @@ class Program_weight_tensor_data_5: class Program_weight_tensor_data_6: name = "data_6" + original_name = "param_66" shape = [768] dtype = "float32" min_val = float("-0.119088") @@ -77,6 +84,7 @@ class Program_weight_tensor_data_6: class Program_weight_tensor_data_7: name = "data_7" + original_name = "param_67" shape = [256, 768] dtype = "float32" min_val = float("-0.3723") @@ -88,6 +96,7 @@ class Program_weight_tensor_data_7: class Program_weight_tensor_data_8: name = "data_8" + original_name = "param_68" shape = [768] dtype = "float32" min_val = float("-0.0974734") @@ -99,6 +108,7 @@ class Program_weight_tensor_data_8: class Program_weight_tensor_data_9: name = "data_9" + original_name = "param_69" shape = [256, 768] dtype = "float32" min_val = float("-0.347624") @@ -110,6 +120,7 @@ class Program_weight_tensor_data_9: class Program_weight_tensor_data_10: name = "data_10" + original_name = "param_70" shape = [768] dtype = "float32" min_val = float("-0.109379") @@ -121,6 +132,7 @@ class Program_weight_tensor_data_10: class Program_weight_tensor_data_11: name = "data_11" + original_name = "param_71" shape = [256, 768] dtype = "float32" min_val = float("-0.295072") @@ -132,6 +144,7 @@ class Program_weight_tensor_data_11: class Program_weight_tensor_data_12: name = "data_12" + original_name = "param_72" shape = [768] dtype = "float32" min_val = float("-0.0793131") @@ -143,6 +156,7 @@ class Program_weight_tensor_data_12: class Program_weight_tensor_data_13: name = "data_13" + original_name = "args_13" shape = [] dtype = "int64" data = [144] @@ -150,6 +164,7 @@ class Program_weight_tensor_data_13: class Program_weight_tensor_data_14: name = "data_14" + original_name = "args_14" shape = [] dtype = "int64" data = [144] @@ -157,6 +172,7 @@ class Program_weight_tensor_data_14: class Program_weight_tensor_data_15: name = "data_15" + original_name = "var_17738" shape = [2, 256, 72, 72] dtype = "float32" min_val = float("-0.556929") @@ -168,6 +184,7 @@ class Program_weight_tensor_data_15: class Program_weight_tensor_data_16: name = "data_16" + original_name = "var_17739" shape = [2, 256, 36, 36] dtype = "float32" min_val = float("-0.556927") @@ -179,6 +196,7 @@ class Program_weight_tensor_data_16: class Program_weight_tensor_data_17: name = "data_17" + original_name = "var_17740" shape = [2, 256, 18, 18] dtype = "float32" min_val = float("-0.556913") @@ -190,6 +208,7 @@ class Program_weight_tensor_data_17: class Program_weight_tensor_data_18: name = "data_18" + original_name = "var_17741" shape = [2, 32, 144, 144] dtype = "float32" min_val = float("-24.9318") @@ -201,6 +220,7 @@ class Program_weight_tensor_data_18: class Program_weight_tensor_data_19: name = "data_19" + original_name = "var_17745" shape = [1, 1] dtype = "int32" data = [1] @@ -208,6 +228,7 @@ class Program_weight_tensor_data_19: class Program_weight_tensor_data_20: name = "data_20" + original_name = "var_17746" shape = [1, 1] dtype = "int32" data = [1] @@ -215,6 +236,7 @@ class Program_weight_tensor_data_20: class Program_weight_tensor_data_21: name = "data_21" + original_name = "var_17747" shape = [1, 4] dtype = "float32" data = [0.561914, 0.52875, 0.651032, 0.94] @@ -222,6 +244,7 @@ class Program_weight_tensor_data_21: class Program_weight_tensor_data_22: name = "data_22" + original_name = "var_17748" shape = [1, 4] dtype = "float32" data = [0.456522, 0.501887, 0.626598, 0.993711] diff --git a/paddle_samples/PaddleX/Mask-RT-DETR-L/subgraph_18/weight_meta.py b/paddle_samples/PaddleX/Mask-RT-DETR-L/subgraph_18/weight_meta.py index 3a7a4e5b8..b187aa94a 100644 --- a/paddle_samples/PaddleX/Mask-RT-DETR-L/subgraph_18/weight_meta.py +++ b/paddle_samples/PaddleX/Mask-RT-DETR-L/subgraph_18/weight_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_parameter_0: name = "parameter_0" + original_name = "layer_norm_4.b_0_deepcopy_143" shape = [256] dtype = "float32" min_val = float("-0.235304") @@ -11,6 +12,7 @@ class Program_weight_tensor_parameter_0: class Program_weight_tensor_parameter_1: name = "parameter_1" + original_name = "layer_norm_4.w_0_deepcopy_142" shape = [256] dtype = "float32" min_val = float("0.669122") @@ -22,6 +24,7 @@ class Program_weight_tensor_parameter_1: class Program_weight_tensor_parameter_2: name = "parameter_2" + original_name = "linear_9.b_0_deepcopy_141" shape = [256] dtype = "float32" min_val = float("-0.229266") @@ -33,6 +36,7 @@ class Program_weight_tensor_parameter_2: class Program_weight_tensor_parameter_3: name = "parameter_3" + original_name = "linear_9.w_0_deepcopy_140" shape = [1024, 256] dtype = "float32" min_val = float("-0.411993") @@ -44,6 +48,7 @@ class Program_weight_tensor_parameter_3: class Program_weight_tensor_parameter_4: name = "parameter_4" + original_name = "linear_8.b_0_deepcopy_139" shape = [1024] dtype = "float32" min_val = float("-0.160422") @@ -55,6 +60,7 @@ class Program_weight_tensor_parameter_4: class Program_weight_tensor_parameter_5: name = "parameter_5" + original_name = "linear_8.w_0_deepcopy_138" shape = [256, 1024] dtype = "float32" min_val = float("-0.314147") @@ -66,6 +72,7 @@ class Program_weight_tensor_parameter_5: class Program_weight_tensor_parameter_6: name = "parameter_6" + original_name = "layer_norm_3.b_0_deepcopy_137" shape = [256] dtype = "float32" min_val = float("-0.165677") @@ -77,6 +84,7 @@ class Program_weight_tensor_parameter_6: class Program_weight_tensor_parameter_7: name = "parameter_7" + original_name = "layer_norm_3.w_0_deepcopy_136" shape = [256] dtype = "float32" min_val = float("0.819597") @@ -88,6 +96,7 @@ class Program_weight_tensor_parameter_7: class Program_weight_tensor_parameter_8: name = "parameter_8" + original_name = "linear_7.b_0_deepcopy_135" shape = [256] dtype = "float32" min_val = float("-0.167081") @@ -99,6 +108,7 @@ class Program_weight_tensor_parameter_8: class Program_weight_tensor_parameter_9: name = "parameter_9" + original_name = "linear_7.w_0_deepcopy_134" shape = [256, 256] dtype = "float32" min_val = float("-0.274836") @@ -110,6 +120,7 @@ class Program_weight_tensor_parameter_9: class Program_weight_tensor_parameter_10: name = "parameter_10" + original_name = "linear_5.b_0_deepcopy_131" shape = [96] dtype = "float32" min_val = float("-0.0530445") @@ -121,6 +132,7 @@ class Program_weight_tensor_parameter_10: class Program_weight_tensor_parameter_11: name = "parameter_11" + original_name = "linear_5.w_0_deepcopy_130" shape = [256, 96] dtype = "float32" min_val = float("-0.351484") @@ -132,6 +144,7 @@ class Program_weight_tensor_parameter_11: class Program_weight_tensor_parameter_12: name = "parameter_12" + original_name = "linear_4.b_0_deepcopy_129" shape = [192] dtype = "float32" min_val = float("-4.13705") @@ -143,6 +156,7 @@ class Program_weight_tensor_parameter_12: class Program_weight_tensor_parameter_13: name = "parameter_13" + original_name = "linear_4.w_0_deepcopy_128" shape = [256, 192] dtype = "float32" min_val = float("-0.329092") @@ -154,6 +168,7 @@ class Program_weight_tensor_parameter_13: class Program_weight_tensor_parameter_14: name = "parameter_14" + original_name = "linear_6.b_0_deepcopy_133" shape = [256] dtype = "float32" min_val = float("-0.0672969") @@ -165,6 +180,7 @@ class Program_weight_tensor_parameter_14: class Program_weight_tensor_parameter_15: name = "parameter_15" + original_name = "linear_6.w_0_deepcopy_132" shape = [256, 256] dtype = "float32" min_val = float("-0.291325") @@ -176,6 +192,7 @@ class Program_weight_tensor_parameter_15: class Program_weight_tensor_parameter_16: name = "parameter_16" + original_name = "layer_norm_2.b_0_deepcopy_127" shape = [256] dtype = "float32" min_val = float("-0.113193") @@ -187,6 +204,7 @@ class Program_weight_tensor_parameter_16: class Program_weight_tensor_parameter_17: name = "parameter_17" + original_name = "layer_norm_2.w_0_deepcopy_126" shape = [256] dtype = "float32" min_val = float("0.913411") @@ -198,6 +216,7 @@ class Program_weight_tensor_parameter_17: class Program_weight_tensor_parameter_18: name = "parameter_18" + original_name = "linear_3.b_0_deepcopy_125" shape = [256] dtype = "float32" min_val = float("-0.402983") @@ -209,6 +228,7 @@ class Program_weight_tensor_parameter_18: class Program_weight_tensor_parameter_19: name = "parameter_19" + original_name = "linear_3.w_0_deepcopy_124" shape = [256, 256] dtype = "float32" min_val = float("-0.385772") @@ -220,6 +240,7 @@ class Program_weight_tensor_parameter_19: class Program_weight_tensor_parameter_20: name = "parameter_20" + original_name = "layer_norm_4.b_0_deepcopy_121" shape = [256] dtype = "float32" min_val = float("-0.242033") @@ -231,6 +252,7 @@ class Program_weight_tensor_parameter_20: class Program_weight_tensor_parameter_21: name = "parameter_21" + original_name = "layer_norm_4.w_0_deepcopy_120" shape = [256] dtype = "float32" min_val = float("0.651342") @@ -242,6 +264,7 @@ class Program_weight_tensor_parameter_21: class Program_weight_tensor_parameter_22: name = "parameter_22" + original_name = "linear_9.b_0_deepcopy_119" shape = [256] dtype = "float32" min_val = float("-0.242028") @@ -253,6 +276,7 @@ class Program_weight_tensor_parameter_22: class Program_weight_tensor_parameter_23: name = "parameter_23" + original_name = "linear_9.w_0_deepcopy_118" shape = [1024, 256] dtype = "float32" min_val = float("-0.319618") @@ -264,6 +288,7 @@ class Program_weight_tensor_parameter_23: class Program_weight_tensor_parameter_24: name = "parameter_24" + original_name = "linear_8.b_0_deepcopy_117" shape = [1024] dtype = "float32" min_val = float("-0.177691") @@ -275,6 +300,7 @@ class Program_weight_tensor_parameter_24: class Program_weight_tensor_parameter_25: name = "parameter_25" + original_name = "linear_8.w_0_deepcopy_116" shape = [256, 1024] dtype = "float32" min_val = float("-0.295095") @@ -286,6 +312,7 @@ class Program_weight_tensor_parameter_25: class Program_weight_tensor_parameter_26: name = "parameter_26" + original_name = "layer_norm_3.b_0_deepcopy_115" shape = [256] dtype = "float32" min_val = float("-0.0835911") @@ -297,6 +324,7 @@ class Program_weight_tensor_parameter_26: class Program_weight_tensor_parameter_27: name = "parameter_27" + original_name = "layer_norm_3.w_0_deepcopy_114" shape = [256] dtype = "float32" min_val = float("0.84112") @@ -308,6 +336,7 @@ class Program_weight_tensor_parameter_27: class Program_weight_tensor_parameter_28: name = "parameter_28" + original_name = "linear_7.b_0_deepcopy_113" shape = [256] dtype = "float32" min_val = float("-0.100836") @@ -319,6 +348,7 @@ class Program_weight_tensor_parameter_28: class Program_weight_tensor_parameter_29: name = "parameter_29" + original_name = "linear_7.w_0_deepcopy_112" shape = [256, 256] dtype = "float32" min_val = float("-0.258658") @@ -330,6 +360,7 @@ class Program_weight_tensor_parameter_29: class Program_weight_tensor_parameter_30: name = "parameter_30" + original_name = "linear_5.b_0_deepcopy_109" shape = [96] dtype = "float32" min_val = float("-0.113829") @@ -341,6 +372,7 @@ class Program_weight_tensor_parameter_30: class Program_weight_tensor_parameter_31: name = "parameter_31" + original_name = "linear_5.w_0_deepcopy_108" shape = [256, 96] dtype = "float32" min_val = float("-0.310433") @@ -352,6 +384,7 @@ class Program_weight_tensor_parameter_31: class Program_weight_tensor_parameter_32: name = "parameter_32" + original_name = "linear_4.b_0_deepcopy_107" shape = [192] dtype = "float32" min_val = float("-4.16001") @@ -363,6 +396,7 @@ class Program_weight_tensor_parameter_32: class Program_weight_tensor_parameter_33: name = "parameter_33" + original_name = "linear_4.w_0_deepcopy_106" shape = [256, 192] dtype = "float32" min_val = float("-0.505993") @@ -374,6 +408,7 @@ class Program_weight_tensor_parameter_33: class Program_weight_tensor_parameter_34: name = "parameter_34" + original_name = "linear_6.b_0_deepcopy_111" shape = [256] dtype = "float32" min_val = float("-0.0981688") @@ -385,6 +420,7 @@ class Program_weight_tensor_parameter_34: class Program_weight_tensor_parameter_35: name = "parameter_35" + original_name = "linear_6.w_0_deepcopy_110" shape = [256, 256] dtype = "float32" min_val = float("-0.305597") @@ -396,6 +432,7 @@ class Program_weight_tensor_parameter_35: class Program_weight_tensor_parameter_36: name = "parameter_36" + original_name = "layer_norm_2.b_0_deepcopy_105" shape = [256] dtype = "float32" min_val = float("-0.113243") @@ -407,6 +444,7 @@ class Program_weight_tensor_parameter_36: class Program_weight_tensor_parameter_37: name = "parameter_37" + original_name = "layer_norm_2.w_0_deepcopy_104" shape = [256] dtype = "float32" min_val = float("0.809139") @@ -418,6 +456,7 @@ class Program_weight_tensor_parameter_37: class Program_weight_tensor_parameter_38: name = "parameter_38" + original_name = "linear_3.b_0_deepcopy_103" shape = [256] dtype = "float32" min_val = float("-0.148352") @@ -429,6 +468,7 @@ class Program_weight_tensor_parameter_38: class Program_weight_tensor_parameter_39: name = "parameter_39" + original_name = "linear_3.w_0_deepcopy_102" shape = [256, 256] dtype = "float32" min_val = float("-0.275025") @@ -440,6 +480,7 @@ class Program_weight_tensor_parameter_39: class Program_weight_tensor_parameter_40: name = "parameter_40" + original_name = "layer_norm_4.b_0_deepcopy_99" shape = [256] dtype = "float32" min_val = float("-0.152573") @@ -451,6 +492,7 @@ class Program_weight_tensor_parameter_40: class Program_weight_tensor_parameter_41: name = "parameter_41" + original_name = "layer_norm_4.w_0_deepcopy_98" shape = [256] dtype = "float32" min_val = float("0.711054") @@ -462,6 +504,7 @@ class Program_weight_tensor_parameter_41: class Program_weight_tensor_parameter_42: name = "parameter_42" + original_name = "linear_9.b_0_deepcopy_97" shape = [256] dtype = "float32" min_val = float("-0.144996") @@ -473,6 +516,7 @@ class Program_weight_tensor_parameter_42: class Program_weight_tensor_parameter_43: name = "parameter_43" + original_name = "linear_9.w_0_deepcopy_96" shape = [1024, 256] dtype = "float32" min_val = float("-0.340004") @@ -484,6 +528,7 @@ class Program_weight_tensor_parameter_43: class Program_weight_tensor_parameter_44: name = "parameter_44" + original_name = "linear_8.b_0_deepcopy_95" shape = [1024] dtype = "float32" min_val = float("-0.194359") @@ -495,6 +540,7 @@ class Program_weight_tensor_parameter_44: class Program_weight_tensor_parameter_45: name = "parameter_45" + original_name = "linear_8.w_0_deepcopy_94" shape = [256, 1024] dtype = "float32" min_val = float("-0.320983") @@ -506,6 +552,7 @@ class Program_weight_tensor_parameter_45: class Program_weight_tensor_parameter_46: name = "parameter_46" + original_name = "layer_norm_3.b_0_deepcopy_93" shape = [256] dtype = "float32" min_val = float("-0.105718") @@ -517,6 +564,7 @@ class Program_weight_tensor_parameter_46: class Program_weight_tensor_parameter_47: name = "parameter_47" + original_name = "layer_norm_3.w_0_deepcopy_92" shape = [256] dtype = "float32" min_val = float("0.875372") @@ -528,6 +576,7 @@ class Program_weight_tensor_parameter_47: class Program_weight_tensor_parameter_48: name = "parameter_48" + original_name = "linear_7.b_0_deepcopy_91" shape = [256] dtype = "float32" min_val = float("-0.125954") @@ -539,6 +588,7 @@ class Program_weight_tensor_parameter_48: class Program_weight_tensor_parameter_49: name = "parameter_49" + original_name = "linear_7.w_0_deepcopy_90" shape = [256, 256] dtype = "float32" min_val = float("-0.280382") @@ -550,6 +600,7 @@ class Program_weight_tensor_parameter_49: class Program_weight_tensor_parameter_50: name = "parameter_50" + original_name = "linear_5.b_0_deepcopy_87" shape = [96] dtype = "float32" min_val = float("-0.10229") @@ -561,6 +612,7 @@ class Program_weight_tensor_parameter_50: class Program_weight_tensor_parameter_51: name = "parameter_51" + original_name = "linear_5.w_0_deepcopy_86" shape = [256, 96] dtype = "float32" min_val = float("-0.343367") @@ -572,6 +624,7 @@ class Program_weight_tensor_parameter_51: class Program_weight_tensor_parameter_52: name = "parameter_52" + original_name = "linear_4.b_0_deepcopy_85" shape = [192] dtype = "float32" min_val = float("-4.23368") @@ -583,6 +636,7 @@ class Program_weight_tensor_parameter_52: class Program_weight_tensor_parameter_53: name = "parameter_53" + original_name = "linear_4.w_0_deepcopy_84" shape = [256, 192] dtype = "float32" min_val = float("-0.437147") @@ -594,6 +648,7 @@ class Program_weight_tensor_parameter_53: class Program_weight_tensor_parameter_54: name = "parameter_54" + original_name = "linear_6.b_0_deepcopy_89" shape = [256] dtype = "float32" min_val = float("-0.0959403") @@ -605,6 +660,7 @@ class Program_weight_tensor_parameter_54: class Program_weight_tensor_parameter_55: name = "parameter_55" + original_name = "linear_6.w_0_deepcopy_88" shape = [256, 256] dtype = "float32" min_val = float("-0.312222") @@ -616,6 +672,7 @@ class Program_weight_tensor_parameter_55: class Program_weight_tensor_parameter_56: name = "parameter_56" + original_name = "layer_norm_2.b_0_deepcopy_83" shape = [256] dtype = "float32" min_val = float("-0.0998866") @@ -627,6 +684,7 @@ class Program_weight_tensor_parameter_56: class Program_weight_tensor_parameter_57: name = "parameter_57" + original_name = "layer_norm_2.w_0_deepcopy_82" shape = [256] dtype = "float32" min_val = float("0.805504") @@ -638,6 +696,7 @@ class Program_weight_tensor_parameter_57: class Program_weight_tensor_parameter_58: name = "parameter_58" + original_name = "linear_3.b_0_deepcopy_81" shape = [256] dtype = "float32" min_val = float("-0.132093") @@ -649,6 +708,7 @@ class Program_weight_tensor_parameter_58: class Program_weight_tensor_parameter_59: name = "parameter_59" + original_name = "linear_3.w_0_deepcopy_80" shape = [256, 256] dtype = "float32" min_val = float("-0.356367") @@ -660,6 +720,7 @@ class Program_weight_tensor_parameter_59: class Program_weight_tensor_parameter_60: name = "parameter_60" + original_name = "layer_norm_4.b_0_deepcopy_77" shape = [256] dtype = "float32" min_val = float("-0.0918197") @@ -671,6 +732,7 @@ class Program_weight_tensor_parameter_60: class Program_weight_tensor_parameter_61: name = "parameter_61" + original_name = "layer_norm_4.w_0_deepcopy_76" shape = [256] dtype = "float32" min_val = float("0.763868") @@ -682,6 +744,7 @@ class Program_weight_tensor_parameter_61: class Program_weight_tensor_parameter_62: name = "parameter_62" + original_name = "linear_9.b_0_deepcopy_75" shape = [256] dtype = "float32" min_val = float("-0.116682") @@ -693,6 +756,7 @@ class Program_weight_tensor_parameter_62: class Program_weight_tensor_parameter_63: name = "parameter_63" + original_name = "linear_9.w_0_deepcopy_74" shape = [1024, 256] dtype = "float32" min_val = float("-0.424504") @@ -704,6 +768,7 @@ class Program_weight_tensor_parameter_63: class Program_weight_tensor_parameter_64: name = "parameter_64" + original_name = "linear_8.b_0_deepcopy_73" shape = [1024] dtype = "float32" min_val = float("-0.265825") @@ -715,6 +780,7 @@ class Program_weight_tensor_parameter_64: class Program_weight_tensor_parameter_65: name = "parameter_65" + original_name = "linear_8.w_0_deepcopy_72" shape = [256, 1024] dtype = "float32" min_val = float("-0.34927") @@ -726,6 +792,7 @@ class Program_weight_tensor_parameter_65: class Program_weight_tensor_parameter_66: name = "parameter_66" + original_name = "layer_norm_3.b_0_deepcopy_71" shape = [256] dtype = "float32" min_val = float("-0.107562") @@ -737,6 +804,7 @@ class Program_weight_tensor_parameter_66: class Program_weight_tensor_parameter_67: name = "parameter_67" + original_name = "layer_norm_3.w_0_deepcopy_70" shape = [256] dtype = "float32" min_val = float("0.870413") @@ -748,6 +816,7 @@ class Program_weight_tensor_parameter_67: class Program_weight_tensor_parameter_68: name = "parameter_68" + original_name = "linear_7.b_0_deepcopy_69" shape = [256] dtype = "float32" min_val = float("-0.0853606") @@ -759,6 +828,7 @@ class Program_weight_tensor_parameter_68: class Program_weight_tensor_parameter_69: name = "parameter_69" + original_name = "linear_7.w_0_deepcopy_68" shape = [256, 256] dtype = "float32" min_val = float("-0.295519") @@ -770,6 +840,7 @@ class Program_weight_tensor_parameter_69: class Program_weight_tensor_parameter_70: name = "parameter_70" + original_name = "linear_5.b_0_deepcopy_65" shape = [96] dtype = "float32" min_val = float("-0.15134") @@ -781,6 +852,7 @@ class Program_weight_tensor_parameter_70: class Program_weight_tensor_parameter_71: name = "parameter_71" + original_name = "linear_5.w_0_deepcopy_64" shape = [256, 96] dtype = "float32" min_val = float("-0.502123") @@ -792,6 +864,7 @@ class Program_weight_tensor_parameter_71: class Program_weight_tensor_parameter_72: name = "parameter_72" + original_name = "linear_4.b_0_deepcopy_63" shape = [192] dtype = "float32" min_val = float("-4.03717") @@ -803,6 +876,7 @@ class Program_weight_tensor_parameter_72: class Program_weight_tensor_parameter_73: name = "parameter_73" + original_name = "linear_4.w_0_deepcopy_62" shape = [256, 192] dtype = "float32" min_val = float("-0.531354") @@ -814,6 +888,7 @@ class Program_weight_tensor_parameter_73: class Program_weight_tensor_parameter_74: name = "parameter_74" + original_name = "linear_6.b_0_deepcopy_67" shape = [256] dtype = "float32" min_val = float("-0.116269") @@ -825,6 +900,7 @@ class Program_weight_tensor_parameter_74: class Program_weight_tensor_parameter_75: name = "parameter_75" + original_name = "linear_6.w_0_deepcopy_66" shape = [256, 256] dtype = "float32" min_val = float("-0.311209") @@ -836,6 +912,7 @@ class Program_weight_tensor_parameter_75: class Program_weight_tensor_parameter_76: name = "parameter_76" + original_name = "layer_norm_2.b_0_deepcopy_61" shape = [256] dtype = "float32" min_val = float("-0.115747") @@ -847,6 +924,7 @@ class Program_weight_tensor_parameter_76: class Program_weight_tensor_parameter_77: name = "parameter_77" + original_name = "layer_norm_2.w_0_deepcopy_60" shape = [256] dtype = "float32" min_val = float("0.738231") @@ -858,6 +936,7 @@ class Program_weight_tensor_parameter_77: class Program_weight_tensor_parameter_78: name = "parameter_78" + original_name = "linear_3.b_0_deepcopy_59" shape = [256] dtype = "float32" min_val = float("-0.197139") @@ -869,6 +948,7 @@ class Program_weight_tensor_parameter_78: class Program_weight_tensor_parameter_79: name = "parameter_79" + original_name = "linear_3.w_0_deepcopy_58" shape = [256, 256] dtype = "float32" min_val = float("-0.298774") @@ -880,6 +960,7 @@ class Program_weight_tensor_parameter_79: class Program_weight_tensor_parameter_80: name = "parameter_80" + original_name = "layer_norm_4.b_0_deepcopy_55" shape = [256] dtype = "float32" min_val = float("-0.102022") @@ -891,6 +972,7 @@ class Program_weight_tensor_parameter_80: class Program_weight_tensor_parameter_81: name = "parameter_81" + original_name = "layer_norm_4.w_0_deepcopy_54" shape = [256] dtype = "float32" min_val = float("0.773895") @@ -902,6 +984,7 @@ class Program_weight_tensor_parameter_81: class Program_weight_tensor_parameter_82: name = "parameter_82" + original_name = "linear_9.b_0_deepcopy_53" shape = [256] dtype = "float32" min_val = float("-0.120819") @@ -913,6 +996,7 @@ class Program_weight_tensor_parameter_82: class Program_weight_tensor_parameter_83: name = "parameter_83" + original_name = "linear_9.w_0_deepcopy_52" shape = [1024, 256] dtype = "float32" min_val = float("-0.517248") @@ -924,6 +1008,7 @@ class Program_weight_tensor_parameter_83: class Program_weight_tensor_parameter_84: name = "parameter_84" + original_name = "linear_8.b_0_deepcopy_51" shape = [1024] dtype = "float32" min_val = float("-0.287466") @@ -935,6 +1020,7 @@ class Program_weight_tensor_parameter_84: class Program_weight_tensor_parameter_85: name = "parameter_85" + original_name = "linear_8.w_0_deepcopy_50" shape = [256, 1024] dtype = "float32" min_val = float("-0.467368") @@ -946,6 +1032,7 @@ class Program_weight_tensor_parameter_85: class Program_weight_tensor_parameter_86: name = "parameter_86" + original_name = "layer_norm_3.b_0_deepcopy_49" shape = [256] dtype = "float32" min_val = float("-0.117937") @@ -957,6 +1044,7 @@ class Program_weight_tensor_parameter_86: class Program_weight_tensor_parameter_87: name = "parameter_87" + original_name = "layer_norm_3.w_0_deepcopy_48" shape = [256] dtype = "float32" min_val = float("0.878245") @@ -968,6 +1056,7 @@ class Program_weight_tensor_parameter_87: class Program_weight_tensor_parameter_88: name = "parameter_88" + original_name = "linear_7.b_0_deepcopy_47" shape = [256] dtype = "float32" min_val = float("-0.106038") @@ -979,6 +1068,7 @@ class Program_weight_tensor_parameter_88: class Program_weight_tensor_parameter_89: name = "parameter_89" + original_name = "linear_7.w_0_deepcopy_46" shape = [256, 256] dtype = "float32" min_val = float("-0.303225") @@ -990,6 +1080,7 @@ class Program_weight_tensor_parameter_89: class Program_weight_tensor_parameter_90: name = "parameter_90" + original_name = "linear_5.b_0_deepcopy_43" shape = [96] dtype = "float32" min_val = float("-0.172088") @@ -1001,6 +1092,7 @@ class Program_weight_tensor_parameter_90: class Program_weight_tensor_parameter_91: name = "parameter_91" + original_name = "linear_5.w_0_deepcopy_42" shape = [256, 96] dtype = "float32" min_val = float("-0.606402") @@ -1012,6 +1104,7 @@ class Program_weight_tensor_parameter_91: class Program_weight_tensor_parameter_92: name = "parameter_92" + original_name = "linear_4.b_0_deepcopy_41" shape = [192] dtype = "float32" min_val = float("-4.20814") @@ -1023,6 +1116,7 @@ class Program_weight_tensor_parameter_92: class Program_weight_tensor_parameter_93: name = "parameter_93" + original_name = "linear_4.w_0_deepcopy_40" shape = [256, 192] dtype = "float32" min_val = float("-0.620158") @@ -1034,6 +1128,7 @@ class Program_weight_tensor_parameter_93: class Program_weight_tensor_parameter_94: name = "parameter_94" + original_name = "linear_6.b_0_deepcopy_45" shape = [256] dtype = "float32" min_val = float("-0.140435") @@ -1045,6 +1140,7 @@ class Program_weight_tensor_parameter_94: class Program_weight_tensor_parameter_95: name = "parameter_95" + original_name = "linear_6.w_0_deepcopy_44" shape = [256, 256] dtype = "float32" min_val = float("-0.315991") @@ -1056,6 +1152,7 @@ class Program_weight_tensor_parameter_95: class Program_weight_tensor_parameter_96: name = "parameter_96" + original_name = "layer_norm_2.b_0_deepcopy_39" shape = [256] dtype = "float32" min_val = float("-0.160634") @@ -1067,6 +1164,7 @@ class Program_weight_tensor_parameter_96: class Program_weight_tensor_parameter_97: name = "parameter_97" + original_name = "layer_norm_2.w_0_deepcopy_38" shape = [256] dtype = "float32" min_val = float("0.68701") @@ -1078,6 +1176,7 @@ class Program_weight_tensor_parameter_97: class Program_weight_tensor_parameter_98: name = "parameter_98" + original_name = "linear_3.b_0_deepcopy_37" shape = [256] dtype = "float32" min_val = float("-0.182583") @@ -1089,6 +1188,7 @@ class Program_weight_tensor_parameter_98: class Program_weight_tensor_parameter_99: name = "parameter_99" + original_name = "linear_3.w_0_deepcopy_36" shape = [256, 256] dtype = "float32" min_val = float("-0.357265") @@ -1100,6 +1200,7 @@ class Program_weight_tensor_parameter_99: class Program_weight_tensor_parameter_100: name = "parameter_100" + original_name = "layer_norm_4.b_0_deepcopy_33" shape = [256] dtype = "float32" min_val = float("-0.158175") @@ -1111,6 +1212,7 @@ class Program_weight_tensor_parameter_100: class Program_weight_tensor_parameter_101: name = "parameter_101" + original_name = "layer_norm_4.w_0_deepcopy_32" shape = [256] dtype = "float32" min_val = float("0.787475") @@ -1122,6 +1224,7 @@ class Program_weight_tensor_parameter_101: class Program_weight_tensor_parameter_102: name = "parameter_102" + original_name = "linear_9.b_0_deepcopy_31" shape = [256] dtype = "float32" min_val = float("-0.135313") @@ -1133,6 +1236,7 @@ class Program_weight_tensor_parameter_102: class Program_weight_tensor_parameter_103: name = "parameter_103" + original_name = "linear_9.w_0_deepcopy_30" shape = [1024, 256] dtype = "float32" min_val = float("-0.375063") @@ -1144,6 +1248,7 @@ class Program_weight_tensor_parameter_103: class Program_weight_tensor_parameter_104: name = "parameter_104" + original_name = "linear_8.b_0_deepcopy_29" shape = [1024] dtype = "float32" min_val = float("-0.277239") @@ -1155,6 +1260,7 @@ class Program_weight_tensor_parameter_104: class Program_weight_tensor_parameter_105: name = "parameter_105" + original_name = "linear_8.w_0_deepcopy_28" shape = [256, 1024] dtype = "float32" min_val = float("-0.351262") @@ -1166,6 +1272,7 @@ class Program_weight_tensor_parameter_105: class Program_weight_tensor_parameter_106: name = "parameter_106" + original_name = "layer_norm_3.b_0_deepcopy_27" shape = [256] dtype = "float32" min_val = float("-0.146354") @@ -1177,6 +1284,7 @@ class Program_weight_tensor_parameter_106: class Program_weight_tensor_parameter_107: name = "parameter_107" + original_name = "layer_norm_3.w_0_deepcopy_26" shape = [256] dtype = "float32" min_val = float("0.87691") @@ -1188,6 +1296,7 @@ class Program_weight_tensor_parameter_107: class Program_weight_tensor_parameter_108: name = "parameter_108" + original_name = "linear_7.b_0_deepcopy_25" shape = [256] dtype = "float32" min_val = float("-0.114909") @@ -1199,6 +1308,7 @@ class Program_weight_tensor_parameter_108: class Program_weight_tensor_parameter_109: name = "parameter_109" + original_name = "linear_7.w_0_deepcopy_24" shape = [256, 256] dtype = "float32" min_val = float("-0.297005") @@ -1210,6 +1320,7 @@ class Program_weight_tensor_parameter_109: class Program_weight_tensor_parameter_110: name = "parameter_110" + original_name = "linear_5.b_0_deepcopy_21" shape = [96] dtype = "float32" min_val = float("-0.201674") @@ -1221,6 +1332,7 @@ class Program_weight_tensor_parameter_110: class Program_weight_tensor_parameter_111: name = "parameter_111" + original_name = "linear_5.w_0_deepcopy_20" shape = [256, 96] dtype = "float32" min_val = float("-0.547945") @@ -1232,6 +1344,7 @@ class Program_weight_tensor_parameter_111: class Program_weight_tensor_parameter_112: name = "parameter_112" + original_name = "linear_4.b_0_deepcopy_19" shape = [192] dtype = "float32" min_val = float("-4.25867") @@ -1243,6 +1356,7 @@ class Program_weight_tensor_parameter_112: class Program_weight_tensor_parameter_113: name = "parameter_113" + original_name = "linear_4.w_0_deepcopy_18" shape = [256, 192] dtype = "float32" min_val = float("-0.678728") @@ -1254,6 +1368,7 @@ class Program_weight_tensor_parameter_113: class Program_weight_tensor_parameter_114: name = "parameter_114" + original_name = "linear_6.b_0_deepcopy_23" shape = [256] dtype = "float32" min_val = float("-0.259039") @@ -1265,6 +1380,7 @@ class Program_weight_tensor_parameter_114: class Program_weight_tensor_parameter_115: name = "parameter_115" + original_name = "linear_6.w_0_deepcopy_22" shape = [256, 256] dtype = "float32" min_val = float("-0.330377") @@ -1276,6 +1392,7 @@ class Program_weight_tensor_parameter_115: class Program_weight_tensor_parameter_116: name = "parameter_116" + original_name = "layer_norm_2.b_0_deepcopy_17" shape = [256] dtype = "float32" min_val = float("-0.211205") @@ -1287,6 +1404,7 @@ class Program_weight_tensor_parameter_116: class Program_weight_tensor_parameter_117: name = "parameter_117" + original_name = "layer_norm_2.w_0_deepcopy_16" shape = [256] dtype = "float32" min_val = float("0.687817") @@ -1298,6 +1416,7 @@ class Program_weight_tensor_parameter_117: class Program_weight_tensor_parameter_118: name = "parameter_118" + original_name = "linear_3.b_0_deepcopy_15" shape = [256] dtype = "float32" min_val = float("-0.270329") @@ -1309,6 +1428,7 @@ class Program_weight_tensor_parameter_118: class Program_weight_tensor_parameter_119: name = "parameter_119" + original_name = "linear_3.w_0_deepcopy_14" shape = [256, 256] dtype = "float32" min_val = float("-0.319173") @@ -1320,6 +1440,7 @@ class Program_weight_tensor_parameter_119: class Program_weight_tensor_parameter_120: name = "parameter_120" + original_name = "linear_11.b_0" shape = [256] dtype = "float32" min_val = float("-0.173673") @@ -1331,6 +1452,7 @@ class Program_weight_tensor_parameter_120: class Program_weight_tensor_parameter_121: name = "parameter_121" + original_name = "linear_11.w_0" shape = [512, 256] dtype = "float32" min_val = float("-1.72547") @@ -1342,6 +1464,7 @@ class Program_weight_tensor_parameter_121: class Program_weight_tensor_parameter_122: name = "parameter_122" + original_name = "linear_10.b_0" shape = [512] dtype = "float32" min_val = float("-0.498659") @@ -1353,6 +1476,7 @@ class Program_weight_tensor_parameter_122: class Program_weight_tensor_parameter_123: name = "parameter_123" + original_name = "linear_10.w_0" shape = [4, 512] dtype = "float32" min_val = float("-3.55188") @@ -1364,6 +1488,7 @@ class Program_weight_tensor_parameter_123: class Program_weight_tensor_parameter_124: name = "parameter_124" + original_name = "linear_14.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -1373,6 +1498,7 @@ class Program_weight_tensor_parameter_124: class Program_weight_tensor_parameter_125: name = "parameter_125" + original_name = "linear_14.w_0" shape = [256, 32] dtype = "float32" min_val = float("-0.739139") @@ -1384,6 +1510,7 @@ class Program_weight_tensor_parameter_125: class Program_weight_tensor_parameter_126: name = "parameter_126" + original_name = "linear_13.b_0" shape = [256] dtype = "float32" min_val = float("-0.14415") @@ -1395,6 +1522,7 @@ class Program_weight_tensor_parameter_126: class Program_weight_tensor_parameter_127: name = "parameter_127" + original_name = "linear_13.w_0" shape = [256, 256] dtype = "float32" min_val = float("-1.2913") @@ -1406,6 +1534,7 @@ class Program_weight_tensor_parameter_127: class Program_weight_tensor_parameter_128: name = "parameter_128" + original_name = "linear_12.b_0" shape = [256] dtype = "float32" min_val = float("-0.46243") @@ -1417,6 +1546,7 @@ class Program_weight_tensor_parameter_128: class Program_weight_tensor_parameter_129: name = "parameter_129" + original_name = "linear_12.w_0" shape = [256, 256] dtype = "float32" min_val = float("-0.377217") @@ -1428,6 +1558,7 @@ class Program_weight_tensor_parameter_129: class Program_weight_tensor_parameter_130: name = "parameter_130" + original_name = "layer_norm_6.b_0" shape = [256] dtype = "float32" min_val = float("-0.756964") @@ -1439,6 +1570,7 @@ class Program_weight_tensor_parameter_130: class Program_weight_tensor_parameter_131: name = "parameter_131" + original_name = "layer_norm_6.w_0" shape = [256] dtype = "float32" min_val = float("0.769167") @@ -1450,6 +1582,7 @@ class Program_weight_tensor_parameter_131: class Program_weight_tensor_parameter_132: name = "parameter_132" + original_name = "linear_19.b_0" shape = [4] dtype = "float32" min_val = float("0") @@ -1459,6 +1592,7 @@ class Program_weight_tensor_parameter_132: class Program_weight_tensor_parameter_133: name = "parameter_133" + original_name = "linear_19.w_0" shape = [256, 4] dtype = "float32" min_val = float("-0.571369") @@ -1470,6 +1604,7 @@ class Program_weight_tensor_parameter_133: class Program_weight_tensor_parameter_134: name = "parameter_134" + original_name = "linear_18.b_0" shape = [256] dtype = "float32" min_val = float("-0.257734") @@ -1481,6 +1616,7 @@ class Program_weight_tensor_parameter_134: class Program_weight_tensor_parameter_135: name = "parameter_135" + original_name = "linear_18.w_0" shape = [256, 256] dtype = "float32" min_val = float("-2.25347") @@ -1492,6 +1628,7 @@ class Program_weight_tensor_parameter_135: class Program_weight_tensor_parameter_136: name = "parameter_136" + original_name = "linear_17.b_0" shape = [256] dtype = "float32" min_val = float("-0.332837") @@ -1503,6 +1640,7 @@ class Program_weight_tensor_parameter_136: class Program_weight_tensor_parameter_137: name = "parameter_137" + original_name = "linear_17.w_0" shape = [256, 256] dtype = "float32" min_val = float("-0.3551") @@ -1514,6 +1652,7 @@ class Program_weight_tensor_parameter_137: class Program_weight_tensor_parameter_138: name = "parameter_138" + original_name = "linear_16.b_0" shape = [2] dtype = "float32" min_val = float("0") @@ -1523,6 +1662,7 @@ class Program_weight_tensor_parameter_138: class Program_weight_tensor_parameter_139: name = "parameter_139" + original_name = "linear_16.w_0" shape = [256, 2] dtype = "float32" min_val = float("-0.0623197") @@ -1534,6 +1674,7 @@ class Program_weight_tensor_parameter_139: class Program_weight_tensor_parameter_140: name = "parameter_140" + original_name = "layer_norm_5.b_0" shape = [256] dtype = "float32" min_val = float("-0.636776") @@ -1545,6 +1686,7 @@ class Program_weight_tensor_parameter_140: class Program_weight_tensor_parameter_141: name = "parameter_141" + original_name = "layer_norm_5.w_0" shape = [256] dtype = "float32" min_val = float("0.726317") @@ -1556,6 +1698,7 @@ class Program_weight_tensor_parameter_141: class Program_weight_tensor_parameter_142: name = "parameter_142" + original_name = "linear_15.b_0" shape = [256] dtype = "float32" min_val = float("-0.560802") @@ -1567,6 +1710,7 @@ class Program_weight_tensor_parameter_142: class Program_weight_tensor_parameter_143: name = "parameter_143" + original_name = "linear_15.w_0" shape = [256, 256] dtype = "float32" min_val = float("-0.417286") @@ -1578,6 +1722,7 @@ class Program_weight_tensor_parameter_143: class Program_weight_tensor_parameter_144: name = "parameter_144" + original_name = "batch_norm2d_208.b_0" shape = [256] dtype = "float32" min_val = float("-0.452767") @@ -1589,6 +1734,7 @@ class Program_weight_tensor_parameter_144: class Program_weight_tensor_parameter_145: name = "parameter_145" + original_name = "batch_norm2d_208.w_0" shape = [256] dtype = "float32" min_val = float("0.760863") @@ -1600,6 +1746,7 @@ class Program_weight_tensor_parameter_145: class Program_weight_tensor_parameter_146: name = "parameter_146" + original_name = "batch_norm2d_208.w_2" shape = [256] dtype = "float32" min_val = float("6.38446") @@ -1611,6 +1758,7 @@ class Program_weight_tensor_parameter_146: class Program_weight_tensor_parameter_147: name = "parameter_147" + original_name = "batch_norm2d_208.w_1" shape = [256] dtype = "float32" min_val = float("-2.19169") @@ -1622,6 +1770,7 @@ class Program_weight_tensor_parameter_147: class Program_weight_tensor_parameter_148: name = "parameter_148" + original_name = "conv2d_129.w_0" shape = [256, 256, 1, 1] dtype = "float32" min_val = float("-0.34255") @@ -1633,6 +1782,7 @@ class Program_weight_tensor_parameter_148: class Program_weight_tensor_parameter_149: name = "parameter_149" + original_name = "batch_norm2d_207.b_0" shape = [256] dtype = "float32" min_val = float("-0.313442") @@ -1644,6 +1794,7 @@ class Program_weight_tensor_parameter_149: class Program_weight_tensor_parameter_150: name = "parameter_150" + original_name = "batch_norm2d_207.w_0" shape = [256] dtype = "float32" min_val = float("0.812166") @@ -1655,6 +1806,7 @@ class Program_weight_tensor_parameter_150: class Program_weight_tensor_parameter_151: name = "parameter_151" + original_name = "batch_norm2d_207.w_2" shape = [256] dtype = "float32" min_val = float("10.2251") @@ -1666,6 +1818,7 @@ class Program_weight_tensor_parameter_151: class Program_weight_tensor_parameter_152: name = "parameter_152" + original_name = "batch_norm2d_207.w_1" shape = [256] dtype = "float32" min_val = float("-2.15563") @@ -1677,6 +1830,7 @@ class Program_weight_tensor_parameter_152: class Program_weight_tensor_parameter_153: name = "parameter_153" + original_name = "conv2d_128.w_0" shape = [256, 256, 1, 1] dtype = "float32" min_val = float("-0.361386") @@ -1688,6 +1842,7 @@ class Program_weight_tensor_parameter_153: class Program_weight_tensor_parameter_154: name = "parameter_154" + original_name = "batch_norm2d_206.b_0" shape = [256] dtype = "float32" min_val = float("-0.252684") @@ -1699,6 +1854,7 @@ class Program_weight_tensor_parameter_154: class Program_weight_tensor_parameter_155: name = "parameter_155" + original_name = "batch_norm2d_206.w_0" shape = [256] dtype = "float32" min_val = float("0.725279") @@ -1710,6 +1866,7 @@ class Program_weight_tensor_parameter_155: class Program_weight_tensor_parameter_156: name = "parameter_156" + original_name = "batch_norm2d_206.w_2" shape = [256] dtype = "float32" min_val = float("10.8762") @@ -1721,6 +1878,7 @@ class Program_weight_tensor_parameter_156: class Program_weight_tensor_parameter_157: name = "parameter_157" + original_name = "batch_norm2d_206.w_1" shape = [256] dtype = "float32" min_val = float("-2.01493") @@ -1732,6 +1890,7 @@ class Program_weight_tensor_parameter_157: class Program_weight_tensor_parameter_158: name = "parameter_158" + original_name = "conv2d_127.w_0" shape = [256, 256, 1, 1] dtype = "float32" min_val = float("-0.357751") diff --git a/paddle_samples/PaddleX/Mask-RT-DETR-M/subgraph_3/input_meta.py b/paddle_samples/PaddleX/Mask-RT-DETR-M/subgraph_3/input_meta.py index e5bf5b355..c67b243ec 100644 --- a/paddle_samples/PaddleX/Mask-RT-DETR-M/subgraph_3/input_meta.py +++ b/paddle_samples/PaddleX/Mask-RT-DETR-M/subgraph_3/input_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_data_0: name = "data_0" + original_name = "param_10" shape = [2, 256] dtype = "float32" min_val = float("-2.69103") @@ -11,6 +12,7 @@ class Program_weight_tensor_data_0: class Program_weight_tensor_data_1: name = "data_1" + original_name = "param_11" shape = [256, 768] dtype = "float32" min_val = float("-0.493981") @@ -22,6 +24,7 @@ class Program_weight_tensor_data_1: class Program_weight_tensor_data_2: name = "data_2" + original_name = "param_12" shape = [768] dtype = "float32" min_val = float("-0.349259") @@ -33,6 +36,7 @@ class Program_weight_tensor_data_2: class Program_weight_tensor_data_3: name = "data_3" + original_name = "param_13" shape = [256, 768] dtype = "float32" min_val = float("-0.535905") @@ -44,6 +48,7 @@ class Program_weight_tensor_data_3: class Program_weight_tensor_data_4: name = "data_4" + original_name = "param_14" shape = [768] dtype = "float32" min_val = float("-0.259251") @@ -55,6 +60,7 @@ class Program_weight_tensor_data_4: class Program_weight_tensor_data_5: name = "data_5" + original_name = "param_15" shape = [256, 768] dtype = "float32" min_val = float("-0.496456") @@ -66,6 +72,7 @@ class Program_weight_tensor_data_5: class Program_weight_tensor_data_6: name = "data_6" + original_name = "param_16" shape = [768] dtype = "float32" min_val = float("-0.23482") @@ -77,6 +84,7 @@ class Program_weight_tensor_data_6: class Program_weight_tensor_data_7: name = "data_7" + original_name = "param_17" shape = [256, 768] dtype = "float32" min_val = float("-0.516936") @@ -88,6 +96,7 @@ class Program_weight_tensor_data_7: class Program_weight_tensor_data_8: name = "data_8" + original_name = "param_18" shape = [768] dtype = "float32" min_val = float("-0.160416") @@ -99,6 +108,7 @@ class Program_weight_tensor_data_8: class Program_weight_tensor_data_9: name = "data_9" + original_name = "param_19" shape = [256, 768] dtype = "float32" min_val = float("-0.498378") @@ -110,6 +120,7 @@ class Program_weight_tensor_data_9: class Program_weight_tensor_data_10: name = "data_10" + original_name = "param_20" shape = [768] dtype = "float32" min_val = float("-0.182169") @@ -121,6 +132,7 @@ class Program_weight_tensor_data_10: class Program_weight_tensor_data_11: name = "data_11" + original_name = "param_21" shape = [256, 768] dtype = "float32" min_val = float("-0.441833") @@ -132,6 +144,7 @@ class Program_weight_tensor_data_11: class Program_weight_tensor_data_12: name = "data_12" + original_name = "param_22" shape = [768] dtype = "float32" min_val = float("-0.108312") @@ -143,6 +156,7 @@ class Program_weight_tensor_data_12: class Program_weight_tensor_data_13: name = "data_13" + original_name = "var_1658" shape = [1, 256, 68, 68] dtype = "float32" min_val = float("-0.278465") @@ -154,6 +168,7 @@ class Program_weight_tensor_data_13: class Program_weight_tensor_data_14: name = "data_14" + original_name = "var_1659" shape = [1, 256, 34, 34] dtype = "float32" min_val = float("-0.278465") @@ -165,6 +180,7 @@ class Program_weight_tensor_data_14: class Program_weight_tensor_data_15: name = "data_15" + original_name = "var_1660" shape = [1, 256, 17, 17] dtype = "float32" min_val = float("-0.278465") @@ -176,6 +192,7 @@ class Program_weight_tensor_data_15: class Program_weight_tensor_data_16: name = "data_16" + original_name = "var_1661" shape = [1, 128, 136, 136] dtype = "float32" min_val = float("-35.2707") @@ -187,6 +204,7 @@ class Program_weight_tensor_data_16: class Program_weight_tensor_data_17: name = "data_17" + original_name = "var_1664" shape = [1, 1] dtype = "int32" data = [0] @@ -194,6 +212,7 @@ class Program_weight_tensor_data_17: class Program_weight_tensor_data_18: name = "data_18" + original_name = "var_1665" shape = [1, 4] dtype = "float32" data = [0.494674, 0.791309, 0.196934, 0.299688] diff --git a/paddle_samples/PaddleX/Mask-RT-DETR-M/subgraph_3/weight_meta.py b/paddle_samples/PaddleX/Mask-RT-DETR-M/subgraph_3/weight_meta.py index fa506a78b..e3f1a7cfe 100644 --- a/paddle_samples/PaddleX/Mask-RT-DETR-M/subgraph_3/weight_meta.py +++ b/paddle_samples/PaddleX/Mask-RT-DETR-M/subgraph_3/weight_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_parameter_0: name = "parameter_0" + original_name = "layer_norm_4.b_0_deepcopy_143" shape = [256] dtype = "float32" min_val = float("-0.257848") @@ -11,6 +12,7 @@ class Program_weight_tensor_parameter_0: class Program_weight_tensor_parameter_1: name = "parameter_1" + original_name = "layer_norm_4.w_0_deepcopy_142" shape = [256] dtype = "float32" min_val = float("0.570463") @@ -22,6 +24,7 @@ class Program_weight_tensor_parameter_1: class Program_weight_tensor_parameter_2: name = "parameter_2" + original_name = "linear_9.b_0_deepcopy_141" shape = [256] dtype = "float32" min_val = float("-0.277375") @@ -33,6 +36,7 @@ class Program_weight_tensor_parameter_2: class Program_weight_tensor_parameter_3: name = "parameter_3" + original_name = "linear_9.w_0_deepcopy_140" shape = [1024, 256] dtype = "float32" min_val = float("-0.537982") @@ -44,6 +48,7 @@ class Program_weight_tensor_parameter_3: class Program_weight_tensor_parameter_4: name = "parameter_4" + original_name = "linear_8.b_0_deepcopy_139" shape = [1024] dtype = "float32" min_val = float("-0.195968") @@ -55,6 +60,7 @@ class Program_weight_tensor_parameter_4: class Program_weight_tensor_parameter_5: name = "parameter_5" + original_name = "linear_8.w_0_deepcopy_138" shape = [256, 1024] dtype = "float32" min_val = float("-0.418274") @@ -66,6 +72,7 @@ class Program_weight_tensor_parameter_5: class Program_weight_tensor_parameter_6: name = "parameter_6" + original_name = "layer_norm_3.b_0_deepcopy_137" shape = [256] dtype = "float32" min_val = float("-0.220991") @@ -77,6 +84,7 @@ class Program_weight_tensor_parameter_6: class Program_weight_tensor_parameter_7: name = "parameter_7" + original_name = "layer_norm_3.w_0_deepcopy_136" shape = [256] dtype = "float32" min_val = float("0.707561") @@ -88,6 +96,7 @@ class Program_weight_tensor_parameter_7: class Program_weight_tensor_parameter_8: name = "parameter_8" + original_name = "linear_7.b_0_deepcopy_135" shape = [256] dtype = "float32" min_val = float("-0.211832") @@ -99,6 +108,7 @@ class Program_weight_tensor_parameter_8: class Program_weight_tensor_parameter_9: name = "parameter_9" + original_name = "linear_7.w_0_deepcopy_134" shape = [256, 256] dtype = "float32" min_val = float("-0.529502") @@ -110,6 +120,7 @@ class Program_weight_tensor_parameter_9: class Program_weight_tensor_parameter_10: name = "parameter_10" + original_name = "linear_5.b_0_deepcopy_131" shape = [96] dtype = "float32" min_val = float("-0.110283") @@ -121,6 +132,7 @@ class Program_weight_tensor_parameter_10: class Program_weight_tensor_parameter_11: name = "parameter_11" + original_name = "linear_5.w_0_deepcopy_130" shape = [256, 96] dtype = "float32" min_val = float("-0.368816") @@ -132,6 +144,7 @@ class Program_weight_tensor_parameter_11: class Program_weight_tensor_parameter_12: name = "parameter_12" + original_name = "linear_4.b_0_deepcopy_129" shape = [192] dtype = "float32" min_val = float("-4.17248") @@ -143,6 +156,7 @@ class Program_weight_tensor_parameter_12: class Program_weight_tensor_parameter_13: name = "parameter_13" + original_name = "linear_4.w_0_deepcopy_128" shape = [256, 192] dtype = "float32" min_val = float("-0.594662") @@ -154,6 +168,7 @@ class Program_weight_tensor_parameter_13: class Program_weight_tensor_parameter_14: name = "parameter_14" + original_name = "linear_6.b_0_deepcopy_133" shape = [256] dtype = "float32" min_val = float("-0.136066") @@ -165,6 +180,7 @@ class Program_weight_tensor_parameter_14: class Program_weight_tensor_parameter_15: name = "parameter_15" + original_name = "linear_6.w_0_deepcopy_132" shape = [256, 256] dtype = "float32" min_val = float("-0.320062") @@ -176,6 +192,7 @@ class Program_weight_tensor_parameter_15: class Program_weight_tensor_parameter_16: name = "parameter_16" + original_name = "layer_norm_2.b_0_deepcopy_127" shape = [256] dtype = "float32" min_val = float("-0.139328") @@ -187,6 +204,7 @@ class Program_weight_tensor_parameter_16: class Program_weight_tensor_parameter_17: name = "parameter_17" + original_name = "layer_norm_2.w_0_deepcopy_126" shape = [256] dtype = "float32" min_val = float("0.822527") @@ -198,6 +216,7 @@ class Program_weight_tensor_parameter_17: class Program_weight_tensor_parameter_18: name = "parameter_18" + original_name = "linear_3.b_0_deepcopy_125" shape = [256] dtype = "float32" min_val = float("-0.407685") @@ -209,6 +228,7 @@ class Program_weight_tensor_parameter_18: class Program_weight_tensor_parameter_19: name = "parameter_19" + original_name = "linear_3.w_0_deepcopy_124" shape = [256, 256] dtype = "float32" min_val = float("-0.401918") @@ -220,6 +240,7 @@ class Program_weight_tensor_parameter_19: class Program_weight_tensor_parameter_20: name = "parameter_20" + original_name = "layer_norm_4.b_0_deepcopy_121" shape = [256] dtype = "float32" min_val = float("-0.165821") @@ -231,6 +252,7 @@ class Program_weight_tensor_parameter_20: class Program_weight_tensor_parameter_21: name = "parameter_21" + original_name = "layer_norm_4.w_0_deepcopy_120" shape = [256] dtype = "float32" min_val = float("0.51545") @@ -242,6 +264,7 @@ class Program_weight_tensor_parameter_21: class Program_weight_tensor_parameter_22: name = "parameter_22" + original_name = "linear_9.b_0_deepcopy_119" shape = [256] dtype = "float32" min_val = float("-0.196345") @@ -253,6 +276,7 @@ class Program_weight_tensor_parameter_22: class Program_weight_tensor_parameter_23: name = "parameter_23" + original_name = "linear_9.w_0_deepcopy_118" shape = [1024, 256] dtype = "float32" min_val = float("-0.463641") @@ -264,6 +288,7 @@ class Program_weight_tensor_parameter_23: class Program_weight_tensor_parameter_24: name = "parameter_24" + original_name = "linear_8.b_0_deepcopy_117" shape = [1024] dtype = "float32" min_val = float("-0.209801") @@ -275,6 +300,7 @@ class Program_weight_tensor_parameter_24: class Program_weight_tensor_parameter_25: name = "parameter_25" + original_name = "linear_8.w_0_deepcopy_116" shape = [256, 1024] dtype = "float32" min_val = float("-0.429552") @@ -286,6 +312,7 @@ class Program_weight_tensor_parameter_25: class Program_weight_tensor_parameter_26: name = "parameter_26" + original_name = "layer_norm_3.b_0_deepcopy_115" shape = [256] dtype = "float32" min_val = float("-0.165935") @@ -297,6 +324,7 @@ class Program_weight_tensor_parameter_26: class Program_weight_tensor_parameter_27: name = "parameter_27" + original_name = "layer_norm_3.w_0_deepcopy_114" shape = [256] dtype = "float32" min_val = float("0.760126") @@ -308,6 +336,7 @@ class Program_weight_tensor_parameter_27: class Program_weight_tensor_parameter_28: name = "parameter_28" + original_name = "linear_7.b_0_deepcopy_113" shape = [256] dtype = "float32" min_val = float("-0.175664") @@ -319,6 +348,7 @@ class Program_weight_tensor_parameter_28: class Program_weight_tensor_parameter_29: name = "parameter_29" + original_name = "linear_7.w_0_deepcopy_112" shape = [256, 256] dtype = "float32" min_val = float("-0.581637") @@ -330,6 +360,7 @@ class Program_weight_tensor_parameter_29: class Program_weight_tensor_parameter_30: name = "parameter_30" + original_name = "linear_5.b_0_deepcopy_109" shape = [96] dtype = "float32" min_val = float("-0.172914") @@ -341,6 +372,7 @@ class Program_weight_tensor_parameter_30: class Program_weight_tensor_parameter_31: name = "parameter_31" + original_name = "linear_5.w_0_deepcopy_108" shape = [256, 96] dtype = "float32" min_val = float("-0.364284") @@ -352,6 +384,7 @@ class Program_weight_tensor_parameter_31: class Program_weight_tensor_parameter_32: name = "parameter_32" + original_name = "linear_4.b_0_deepcopy_107" shape = [192] dtype = "float32" min_val = float("-4.29171") @@ -363,6 +396,7 @@ class Program_weight_tensor_parameter_32: class Program_weight_tensor_parameter_33: name = "parameter_33" + original_name = "linear_4.w_0_deepcopy_106" shape = [256, 192] dtype = "float32" min_val = float("-0.74258") @@ -374,6 +408,7 @@ class Program_weight_tensor_parameter_33: class Program_weight_tensor_parameter_34: name = "parameter_34" + original_name = "linear_6.b_0_deepcopy_111" shape = [256] dtype = "float32" min_val = float("-0.197216") @@ -385,6 +420,7 @@ class Program_weight_tensor_parameter_34: class Program_weight_tensor_parameter_35: name = "parameter_35" + original_name = "linear_6.w_0_deepcopy_110" shape = [256, 256] dtype = "float32" min_val = float("-0.391063") @@ -396,6 +432,7 @@ class Program_weight_tensor_parameter_35: class Program_weight_tensor_parameter_36: name = "parameter_36" + original_name = "layer_norm_2.b_0_deepcopy_105" shape = [256] dtype = "float32" min_val = float("-0.196236") @@ -407,6 +444,7 @@ class Program_weight_tensor_parameter_36: class Program_weight_tensor_parameter_37: name = "parameter_37" + original_name = "layer_norm_2.w_0_deepcopy_104" shape = [256] dtype = "float32" min_val = float("0.576399") @@ -418,6 +456,7 @@ class Program_weight_tensor_parameter_37: class Program_weight_tensor_parameter_38: name = "parameter_38" + original_name = "linear_3.b_0_deepcopy_103" shape = [256] dtype = "float32" min_val = float("-0.229334") @@ -429,6 +468,7 @@ class Program_weight_tensor_parameter_38: class Program_weight_tensor_parameter_39: name = "parameter_39" + original_name = "linear_3.w_0_deepcopy_102" shape = [256, 256] dtype = "float32" min_val = float("-0.468118") @@ -440,6 +480,7 @@ class Program_weight_tensor_parameter_39: class Program_weight_tensor_parameter_40: name = "parameter_40" + original_name = "layer_norm_4.b_0_deepcopy_99" shape = [256] dtype = "float32" min_val = float("-0.125057") @@ -451,6 +492,7 @@ class Program_weight_tensor_parameter_40: class Program_weight_tensor_parameter_41: name = "parameter_41" + original_name = "layer_norm_4.w_0_deepcopy_98" shape = [256] dtype = "float32" min_val = float("0.582964") @@ -462,6 +504,7 @@ class Program_weight_tensor_parameter_41: class Program_weight_tensor_parameter_42: name = "parameter_42" + original_name = "linear_9.b_0_deepcopy_97" shape = [256] dtype = "float32" min_val = float("-0.175962") @@ -473,6 +516,7 @@ class Program_weight_tensor_parameter_42: class Program_weight_tensor_parameter_43: name = "parameter_43" + original_name = "linear_9.w_0_deepcopy_96" shape = [1024, 256] dtype = "float32" min_val = float("-0.485647") @@ -484,6 +528,7 @@ class Program_weight_tensor_parameter_43: class Program_weight_tensor_parameter_44: name = "parameter_44" + original_name = "linear_8.b_0_deepcopy_95" shape = [1024] dtype = "float32" min_val = float("-0.218792") @@ -495,6 +540,7 @@ class Program_weight_tensor_parameter_44: class Program_weight_tensor_parameter_45: name = "parameter_45" + original_name = "linear_8.w_0_deepcopy_94" shape = [256, 1024] dtype = "float32" min_val = float("-0.428192") @@ -506,6 +552,7 @@ class Program_weight_tensor_parameter_45: class Program_weight_tensor_parameter_46: name = "parameter_46" + original_name = "layer_norm_3.b_0_deepcopy_93" shape = [256] dtype = "float32" min_val = float("-0.177081") @@ -517,6 +564,7 @@ class Program_weight_tensor_parameter_46: class Program_weight_tensor_parameter_47: name = "parameter_47" + original_name = "layer_norm_3.w_0_deepcopy_92" shape = [256] dtype = "float32" min_val = float("0.826158") @@ -528,6 +576,7 @@ class Program_weight_tensor_parameter_47: class Program_weight_tensor_parameter_48: name = "parameter_48" + original_name = "linear_7.b_0_deepcopy_91" shape = [256] dtype = "float32" min_val = float("-0.154622") @@ -539,6 +588,7 @@ class Program_weight_tensor_parameter_48: class Program_weight_tensor_parameter_49: name = "parameter_49" + original_name = "linear_7.w_0_deepcopy_90" shape = [256, 256] dtype = "float32" min_val = float("-0.472119") @@ -550,6 +600,7 @@ class Program_weight_tensor_parameter_49: class Program_weight_tensor_parameter_50: name = "parameter_50" + original_name = "linear_5.b_0_deepcopy_87" shape = [96] dtype = "float32" min_val = float("-0.2411") @@ -561,6 +612,7 @@ class Program_weight_tensor_parameter_50: class Program_weight_tensor_parameter_51: name = "parameter_51" + original_name = "linear_5.w_0_deepcopy_86" shape = [256, 96] dtype = "float32" min_val = float("-0.350853") @@ -572,6 +624,7 @@ class Program_weight_tensor_parameter_51: class Program_weight_tensor_parameter_52: name = "parameter_52" + original_name = "linear_4.b_0_deepcopy_85" shape = [192] dtype = "float32" min_val = float("-4.32846") @@ -583,6 +636,7 @@ class Program_weight_tensor_parameter_52: class Program_weight_tensor_parameter_53: name = "parameter_53" + original_name = "linear_4.w_0_deepcopy_84" shape = [256, 192] dtype = "float32" min_val = float("-0.653274") @@ -594,6 +648,7 @@ class Program_weight_tensor_parameter_53: class Program_weight_tensor_parameter_54: name = "parameter_54" + original_name = "linear_6.b_0_deepcopy_89" shape = [256] dtype = "float32" min_val = float("-0.222378") @@ -605,6 +660,7 @@ class Program_weight_tensor_parameter_54: class Program_weight_tensor_parameter_55: name = "parameter_55" + original_name = "linear_6.w_0_deepcopy_88" shape = [256, 256] dtype = "float32" min_val = float("-0.3707") @@ -616,6 +672,7 @@ class Program_weight_tensor_parameter_55: class Program_weight_tensor_parameter_56: name = "parameter_56" + original_name = "layer_norm_2.b_0_deepcopy_83" shape = [256] dtype = "float32" min_val = float("-0.13598") @@ -627,6 +684,7 @@ class Program_weight_tensor_parameter_56: class Program_weight_tensor_parameter_57: name = "parameter_57" + original_name = "layer_norm_2.w_0_deepcopy_82" shape = [256] dtype = "float32" min_val = float("0.577729") @@ -638,6 +696,7 @@ class Program_weight_tensor_parameter_57: class Program_weight_tensor_parameter_58: name = "parameter_58" + original_name = "linear_3.b_0_deepcopy_81" shape = [256] dtype = "float32" min_val = float("-0.197881") @@ -649,6 +708,7 @@ class Program_weight_tensor_parameter_58: class Program_weight_tensor_parameter_59: name = "parameter_59" + original_name = "linear_3.w_0_deepcopy_80" shape = [256, 256] dtype = "float32" min_val = float("-0.49689") @@ -660,6 +720,7 @@ class Program_weight_tensor_parameter_59: class Program_weight_tensor_parameter_60: name = "parameter_60" + original_name = "layer_norm_4.b_0_deepcopy_77" shape = [256] dtype = "float32" min_val = float("-0.104953") @@ -671,6 +732,7 @@ class Program_weight_tensor_parameter_60: class Program_weight_tensor_parameter_61: name = "parameter_61" + original_name = "layer_norm_4.w_0_deepcopy_76" shape = [256] dtype = "float32" min_val = float("0.615191") @@ -682,6 +744,7 @@ class Program_weight_tensor_parameter_61: class Program_weight_tensor_parameter_62: name = "parameter_62" + original_name = "linear_9.b_0_deepcopy_75" shape = [256] dtype = "float32" min_val = float("-0.209475") @@ -693,6 +756,7 @@ class Program_weight_tensor_parameter_62: class Program_weight_tensor_parameter_63: name = "parameter_63" + original_name = "linear_9.w_0_deepcopy_74" shape = [1024, 256] dtype = "float32" min_val = float("-0.526995") @@ -704,6 +768,7 @@ class Program_weight_tensor_parameter_63: class Program_weight_tensor_parameter_64: name = "parameter_64" + original_name = "linear_8.b_0_deepcopy_73" shape = [1024] dtype = "float32" min_val = float("-0.229651") @@ -715,6 +780,7 @@ class Program_weight_tensor_parameter_64: class Program_weight_tensor_parameter_65: name = "parameter_65" + original_name = "linear_8.w_0_deepcopy_72" shape = [256, 1024] dtype = "float32" min_val = float("-0.450684") @@ -726,6 +792,7 @@ class Program_weight_tensor_parameter_65: class Program_weight_tensor_parameter_66: name = "parameter_66" + original_name = "layer_norm_3.b_0_deepcopy_71" shape = [256] dtype = "float32" min_val = float("-0.15251") @@ -737,6 +804,7 @@ class Program_weight_tensor_parameter_66: class Program_weight_tensor_parameter_67: name = "parameter_67" + original_name = "layer_norm_3.w_0_deepcopy_70" shape = [256] dtype = "float32" min_val = float("0.83256") @@ -748,6 +816,7 @@ class Program_weight_tensor_parameter_67: class Program_weight_tensor_parameter_68: name = "parameter_68" + original_name = "linear_7.b_0_deepcopy_69" shape = [256] dtype = "float32" min_val = float("-0.148285") @@ -759,6 +828,7 @@ class Program_weight_tensor_parameter_68: class Program_weight_tensor_parameter_69: name = "parameter_69" + original_name = "linear_7.w_0_deepcopy_68" shape = [256, 256] dtype = "float32" min_val = float("-0.507227") @@ -770,6 +840,7 @@ class Program_weight_tensor_parameter_69: class Program_weight_tensor_parameter_70: name = "parameter_70" + original_name = "linear_5.b_0_deepcopy_65" shape = [96] dtype = "float32" min_val = float("-0.298007") @@ -781,6 +852,7 @@ class Program_weight_tensor_parameter_70: class Program_weight_tensor_parameter_71: name = "parameter_71" + original_name = "linear_5.w_0_deepcopy_64" shape = [256, 96] dtype = "float32" min_val = float("-0.594586") @@ -792,6 +864,7 @@ class Program_weight_tensor_parameter_71: class Program_weight_tensor_parameter_72: name = "parameter_72" + original_name = "linear_4.b_0_deepcopy_63" shape = [192] dtype = "float32" min_val = float("-4.24696") @@ -803,6 +876,7 @@ class Program_weight_tensor_parameter_72: class Program_weight_tensor_parameter_73: name = "parameter_73" + original_name = "linear_4.w_0_deepcopy_62" shape = [256, 192] dtype = "float32" min_val = float("-0.587607") @@ -814,6 +888,7 @@ class Program_weight_tensor_parameter_73: class Program_weight_tensor_parameter_74: name = "parameter_74" + original_name = "linear_6.b_0_deepcopy_67" shape = [256] dtype = "float32" min_val = float("-0.173497") @@ -825,6 +900,7 @@ class Program_weight_tensor_parameter_74: class Program_weight_tensor_parameter_75: name = "parameter_75" + original_name = "linear_6.w_0_deepcopy_66" shape = [256, 256] dtype = "float32" min_val = float("-0.440777") @@ -836,6 +912,7 @@ class Program_weight_tensor_parameter_75: class Program_weight_tensor_parameter_76: name = "parameter_76" + original_name = "layer_norm_2.b_0_deepcopy_61" shape = [256] dtype = "float32" min_val = float("-0.124013") @@ -847,6 +924,7 @@ class Program_weight_tensor_parameter_76: class Program_weight_tensor_parameter_77: name = "parameter_77" + original_name = "layer_norm_2.w_0_deepcopy_60" shape = [256] dtype = "float32" min_val = float("0.555376") @@ -858,6 +936,7 @@ class Program_weight_tensor_parameter_77: class Program_weight_tensor_parameter_78: name = "parameter_78" + original_name = "linear_3.b_0_deepcopy_59" shape = [256] dtype = "float32" min_val = float("-0.28193") @@ -869,6 +948,7 @@ class Program_weight_tensor_parameter_78: class Program_weight_tensor_parameter_79: name = "parameter_79" + original_name = "linear_3.w_0_deepcopy_58" shape = [256, 256] dtype = "float32" min_val = float("-0.664597") @@ -880,6 +960,7 @@ class Program_weight_tensor_parameter_79: class Program_weight_tensor_parameter_80: name = "parameter_80" + original_name = "layer_norm_4.b_0_deepcopy_55" shape = [256] dtype = "float32" min_val = float("-0.131243") @@ -891,6 +972,7 @@ class Program_weight_tensor_parameter_80: class Program_weight_tensor_parameter_81: name = "parameter_81" + original_name = "layer_norm_4.w_0_deepcopy_54" shape = [256] dtype = "float32" min_val = float("0.620092") @@ -902,6 +984,7 @@ class Program_weight_tensor_parameter_81: class Program_weight_tensor_parameter_82: name = "parameter_82" + original_name = "linear_9.b_0_deepcopy_53" shape = [256] dtype = "float32" min_val = float("-0.186161") @@ -913,6 +996,7 @@ class Program_weight_tensor_parameter_82: class Program_weight_tensor_parameter_83: name = "parameter_83" + original_name = "linear_9.w_0_deepcopy_52" shape = [1024, 256] dtype = "float32" min_val = float("-0.650339") @@ -924,6 +1008,7 @@ class Program_weight_tensor_parameter_83: class Program_weight_tensor_parameter_84: name = "parameter_84" + original_name = "linear_8.b_0_deepcopy_51" shape = [1024] dtype = "float32" min_val = float("-0.294994") @@ -935,6 +1020,7 @@ class Program_weight_tensor_parameter_84: class Program_weight_tensor_parameter_85: name = "parameter_85" + original_name = "linear_8.w_0_deepcopy_50" shape = [256, 1024] dtype = "float32" min_val = float("-0.552137") @@ -946,6 +1032,7 @@ class Program_weight_tensor_parameter_85: class Program_weight_tensor_parameter_86: name = "parameter_86" + original_name = "layer_norm_3.b_0_deepcopy_49" shape = [256] dtype = "float32" min_val = float("-0.164228") @@ -957,6 +1044,7 @@ class Program_weight_tensor_parameter_86: class Program_weight_tensor_parameter_87: name = "parameter_87" + original_name = "layer_norm_3.w_0_deepcopy_48" shape = [256] dtype = "float32" min_val = float("0.821907") @@ -968,6 +1056,7 @@ class Program_weight_tensor_parameter_87: class Program_weight_tensor_parameter_88: name = "parameter_88" + original_name = "linear_7.b_0_deepcopy_47" shape = [256] dtype = "float32" min_val = float("-0.134866") @@ -979,6 +1068,7 @@ class Program_weight_tensor_parameter_88: class Program_weight_tensor_parameter_89: name = "parameter_89" + original_name = "linear_7.w_0_deepcopy_46" shape = [256, 256] dtype = "float32" min_val = float("-0.427895") @@ -990,6 +1080,7 @@ class Program_weight_tensor_parameter_89: class Program_weight_tensor_parameter_90: name = "parameter_90" + original_name = "linear_5.b_0_deepcopy_43" shape = [96] dtype = "float32" min_val = float("-0.194148") @@ -1001,6 +1092,7 @@ class Program_weight_tensor_parameter_90: class Program_weight_tensor_parameter_91: name = "parameter_91" + original_name = "linear_5.w_0_deepcopy_42" shape = [256, 96] dtype = "float32" min_val = float("-0.669615") @@ -1012,6 +1104,7 @@ class Program_weight_tensor_parameter_91: class Program_weight_tensor_parameter_92: name = "parameter_92" + original_name = "linear_4.b_0_deepcopy_41" shape = [192] dtype = "float32" min_val = float("-4.30897") @@ -1023,6 +1116,7 @@ class Program_weight_tensor_parameter_92: class Program_weight_tensor_parameter_93: name = "parameter_93" + original_name = "linear_4.w_0_deepcopy_40" shape = [256, 192] dtype = "float32" min_val = float("-0.803796") @@ -1034,6 +1128,7 @@ class Program_weight_tensor_parameter_93: class Program_weight_tensor_parameter_94: name = "parameter_94" + original_name = "linear_6.b_0_deepcopy_45" shape = [256] dtype = "float32" min_val = float("-0.271884") @@ -1045,6 +1140,7 @@ class Program_weight_tensor_parameter_94: class Program_weight_tensor_parameter_95: name = "parameter_95" + original_name = "linear_6.w_0_deepcopy_44" shape = [256, 256] dtype = "float32" min_val = float("-0.425265") @@ -1056,6 +1152,7 @@ class Program_weight_tensor_parameter_95: class Program_weight_tensor_parameter_96: name = "parameter_96" + original_name = "layer_norm_2.b_0_deepcopy_39" shape = [256] dtype = "float32" min_val = float("-0.162712") @@ -1067,6 +1164,7 @@ class Program_weight_tensor_parameter_96: class Program_weight_tensor_parameter_97: name = "parameter_97" + original_name = "layer_norm_2.w_0_deepcopy_38" shape = [256] dtype = "float32" min_val = float("0.381827") @@ -1078,6 +1176,7 @@ class Program_weight_tensor_parameter_97: class Program_weight_tensor_parameter_98: name = "parameter_98" + original_name = "linear_3.b_0_deepcopy_37" shape = [256] dtype = "float32" min_val = float("-0.574018") @@ -1089,6 +1188,7 @@ class Program_weight_tensor_parameter_98: class Program_weight_tensor_parameter_99: name = "parameter_99" + original_name = "linear_3.w_0_deepcopy_36" shape = [256, 256] dtype = "float32" min_val = float("-0.7134") @@ -1100,6 +1200,7 @@ class Program_weight_tensor_parameter_99: class Program_weight_tensor_parameter_100: name = "parameter_100" + original_name = "layer_norm_4.b_0_deepcopy_33" shape = [256] dtype = "float32" min_val = float("-0.201337") @@ -1111,6 +1212,7 @@ class Program_weight_tensor_parameter_100: class Program_weight_tensor_parameter_101: name = "parameter_101" + original_name = "layer_norm_4.w_0_deepcopy_32" shape = [256] dtype = "float32" min_val = float("0.615198") @@ -1122,6 +1224,7 @@ class Program_weight_tensor_parameter_101: class Program_weight_tensor_parameter_102: name = "parameter_102" + original_name = "linear_9.b_0_deepcopy_31" shape = [256] dtype = "float32" min_val = float("-0.531165") @@ -1133,6 +1236,7 @@ class Program_weight_tensor_parameter_102: class Program_weight_tensor_parameter_103: name = "parameter_103" + original_name = "linear_9.w_0_deepcopy_30" shape = [1024, 256] dtype = "float32" min_val = float("-0.720231") @@ -1144,6 +1248,7 @@ class Program_weight_tensor_parameter_103: class Program_weight_tensor_parameter_104: name = "parameter_104" + original_name = "linear_8.b_0_deepcopy_29" shape = [1024] dtype = "float32" min_val = float("-0.312767") @@ -1155,6 +1260,7 @@ class Program_weight_tensor_parameter_104: class Program_weight_tensor_parameter_105: name = "parameter_105" + original_name = "linear_8.w_0_deepcopy_28" shape = [256, 1024] dtype = "float32" min_val = float("-0.518351") @@ -1166,6 +1272,7 @@ class Program_weight_tensor_parameter_105: class Program_weight_tensor_parameter_106: name = "parameter_106" + original_name = "layer_norm_3.b_0_deepcopy_27" shape = [256] dtype = "float32" min_val = float("-0.201598") @@ -1177,6 +1284,7 @@ class Program_weight_tensor_parameter_106: class Program_weight_tensor_parameter_107: name = "parameter_107" + original_name = "layer_norm_3.w_0_deepcopy_26" shape = [256] dtype = "float32" min_val = float("0.755676") @@ -1188,6 +1296,7 @@ class Program_weight_tensor_parameter_107: class Program_weight_tensor_parameter_108: name = "parameter_108" + original_name = "linear_7.b_0_deepcopy_25" shape = [256] dtype = "float32" min_val = float("-0.131418") @@ -1199,6 +1308,7 @@ class Program_weight_tensor_parameter_108: class Program_weight_tensor_parameter_109: name = "parameter_109" + original_name = "linear_7.w_0_deepcopy_24" shape = [256, 256] dtype = "float32" min_val = float("-0.514779") @@ -1210,6 +1320,7 @@ class Program_weight_tensor_parameter_109: class Program_weight_tensor_parameter_110: name = "parameter_110" + original_name = "linear_5.b_0_deepcopy_21" shape = [96] dtype = "float32" min_val = float("-0.14749") @@ -1221,6 +1332,7 @@ class Program_weight_tensor_parameter_110: class Program_weight_tensor_parameter_111: name = "parameter_111" + original_name = "linear_5.w_0_deepcopy_20" shape = [256, 96] dtype = "float32" min_val = float("-0.329684") @@ -1232,6 +1344,7 @@ class Program_weight_tensor_parameter_111: class Program_weight_tensor_parameter_112: name = "parameter_112" + original_name = "linear_4.b_0_deepcopy_19" shape = [192] dtype = "float32" min_val = float("-4.05128") @@ -1243,6 +1356,7 @@ class Program_weight_tensor_parameter_112: class Program_weight_tensor_parameter_113: name = "parameter_113" + original_name = "linear_4.w_0_deepcopy_18" shape = [256, 192] dtype = "float32" min_val = float("-0.583759") @@ -1254,6 +1368,7 @@ class Program_weight_tensor_parameter_113: class Program_weight_tensor_parameter_114: name = "parameter_114" + original_name = "linear_6.b_0_deepcopy_23" shape = [256] dtype = "float32" min_val = float("-0.279755") @@ -1265,6 +1380,7 @@ class Program_weight_tensor_parameter_114: class Program_weight_tensor_parameter_115: name = "parameter_115" + original_name = "linear_6.w_0_deepcopy_22" shape = [256, 256] dtype = "float32" min_val = float("-0.380746") @@ -1276,6 +1392,7 @@ class Program_weight_tensor_parameter_115: class Program_weight_tensor_parameter_116: name = "parameter_116" + original_name = "layer_norm_2.b_0_deepcopy_17" shape = [256] dtype = "float32" min_val = float("-0.323486") @@ -1287,6 +1404,7 @@ class Program_weight_tensor_parameter_116: class Program_weight_tensor_parameter_117: name = "parameter_117" + original_name = "layer_norm_2.w_0_deepcopy_16" shape = [256] dtype = "float32" min_val = float("0.500781") @@ -1298,6 +1416,7 @@ class Program_weight_tensor_parameter_117: class Program_weight_tensor_parameter_118: name = "parameter_118" + original_name = "linear_3.b_0_deepcopy_15" shape = [256] dtype = "float32" min_val = float("-1.26318") @@ -1309,6 +1428,7 @@ class Program_weight_tensor_parameter_118: class Program_weight_tensor_parameter_119: name = "parameter_119" + original_name = "linear_3.w_0_deepcopy_14" shape = [256, 256] dtype = "float32" min_val = float("-0.949722") @@ -1320,6 +1440,7 @@ class Program_weight_tensor_parameter_119: class Program_weight_tensor_parameter_120: name = "parameter_120" + original_name = "linear_11.b_0" shape = [256] dtype = "float32" min_val = float("-0.53675") @@ -1331,6 +1452,7 @@ class Program_weight_tensor_parameter_120: class Program_weight_tensor_parameter_121: name = "parameter_121" + original_name = "linear_11.w_0" shape = [512, 256] dtype = "float32" min_val = float("-2.64187") @@ -1342,6 +1464,7 @@ class Program_weight_tensor_parameter_121: class Program_weight_tensor_parameter_122: name = "parameter_122" + original_name = "linear_10.b_0" shape = [512] dtype = "float32" min_val = float("-0.496871") @@ -1353,6 +1476,7 @@ class Program_weight_tensor_parameter_122: class Program_weight_tensor_parameter_123: name = "parameter_123" + original_name = "linear_10.w_0" shape = [4, 512] dtype = "float32" min_val = float("-5.17564") @@ -1364,6 +1488,7 @@ class Program_weight_tensor_parameter_123: class Program_weight_tensor_parameter_124: name = "parameter_124" + original_name = "linear_14.b_0" shape = [128] dtype = "float32" min_val = float("-0.235086") @@ -1375,6 +1500,7 @@ class Program_weight_tensor_parameter_124: class Program_weight_tensor_parameter_125: name = "parameter_125" + original_name = "linear_14.w_0" shape = [256, 128] dtype = "float32" min_val = float("-0.77282") @@ -1386,6 +1512,7 @@ class Program_weight_tensor_parameter_125: class Program_weight_tensor_parameter_126: name = "parameter_126" + original_name = "linear_13.b_0" shape = [256] dtype = "float32" min_val = float("-0.467767") @@ -1397,6 +1524,7 @@ class Program_weight_tensor_parameter_126: class Program_weight_tensor_parameter_127: name = "parameter_127" + original_name = "linear_13.w_0" shape = [256, 256] dtype = "float32" min_val = float("-1.43517") @@ -1408,6 +1536,7 @@ class Program_weight_tensor_parameter_127: class Program_weight_tensor_parameter_128: name = "parameter_128" + original_name = "linear_12.b_0" shape = [256] dtype = "float32" min_val = float("-0.514653") @@ -1419,6 +1548,7 @@ class Program_weight_tensor_parameter_128: class Program_weight_tensor_parameter_129: name = "parameter_129" + original_name = "linear_12.w_0" shape = [256, 256] dtype = "float32" min_val = float("-0.556973") @@ -1430,6 +1560,7 @@ class Program_weight_tensor_parameter_129: class Program_weight_tensor_parameter_130: name = "parameter_130" + original_name = "layer_norm_6.b_0" shape = [256] dtype = "float32" min_val = float("-1.14669") @@ -1441,6 +1572,7 @@ class Program_weight_tensor_parameter_130: class Program_weight_tensor_parameter_131: name = "parameter_131" + original_name = "layer_norm_6.w_0" shape = [256] dtype = "float32" min_val = float("0.885796") @@ -1452,6 +1584,7 @@ class Program_weight_tensor_parameter_131: class Program_weight_tensor_parameter_132: name = "parameter_132" + original_name = "linear_19.b_0" shape = [4] dtype = "float32" min_val = float("0") @@ -1461,6 +1594,7 @@ class Program_weight_tensor_parameter_132: class Program_weight_tensor_parameter_133: name = "parameter_133" + original_name = "linear_19.w_0" shape = [256, 4] dtype = "float32" min_val = float("-0.468809") @@ -1472,6 +1606,7 @@ class Program_weight_tensor_parameter_133: class Program_weight_tensor_parameter_134: name = "parameter_134" + original_name = "linear_18.b_0" shape = [256] dtype = "float32" min_val = float("-0.326378") @@ -1483,6 +1618,7 @@ class Program_weight_tensor_parameter_134: class Program_weight_tensor_parameter_135: name = "parameter_135" + original_name = "linear_18.w_0" shape = [256, 256] dtype = "float32" min_val = float("-4.06851") @@ -1494,6 +1630,7 @@ class Program_weight_tensor_parameter_135: class Program_weight_tensor_parameter_136: name = "parameter_136" + original_name = "linear_17.b_0" shape = [256] dtype = "float32" min_val = float("-0.451628") @@ -1505,6 +1642,7 @@ class Program_weight_tensor_parameter_136: class Program_weight_tensor_parameter_137: name = "parameter_137" + original_name = "linear_17.w_0" shape = [256, 256] dtype = "float32" min_val = float("-0.484331") @@ -1516,6 +1654,7 @@ class Program_weight_tensor_parameter_137: class Program_weight_tensor_parameter_138: name = "parameter_138" + original_name = "linear_16.b_0" shape = [2] dtype = "float32" min_val = float("0") @@ -1525,6 +1664,7 @@ class Program_weight_tensor_parameter_138: class Program_weight_tensor_parameter_139: name = "parameter_139" + original_name = "linear_16.w_0" shape = [256, 2] dtype = "float32" min_val = float("-0.0619545") @@ -1536,6 +1676,7 @@ class Program_weight_tensor_parameter_139: class Program_weight_tensor_parameter_140: name = "parameter_140" + original_name = "layer_norm_5.b_0" shape = [256] dtype = "float32" min_val = float("-1.11617") @@ -1547,6 +1688,7 @@ class Program_weight_tensor_parameter_140: class Program_weight_tensor_parameter_141: name = "parameter_141" + original_name = "layer_norm_5.w_0" shape = [256] dtype = "float32" min_val = float("0.526477") @@ -1558,6 +1700,7 @@ class Program_weight_tensor_parameter_141: class Program_weight_tensor_parameter_142: name = "parameter_142" + original_name = "linear_15.b_0" shape = [256] dtype = "float32" min_val = float("-0.944365") @@ -1569,6 +1712,7 @@ class Program_weight_tensor_parameter_142: class Program_weight_tensor_parameter_143: name = "parameter_143" + original_name = "linear_15.w_0" shape = [256, 256] dtype = "float32" min_val = float("-0.875333") @@ -1580,6 +1724,7 @@ class Program_weight_tensor_parameter_143: class Program_weight_tensor_parameter_144: name = "parameter_144" + original_name = "batch_norm2d_172.b_0" shape = [256] dtype = "float32" min_val = float("-0.256558") @@ -1591,6 +1736,7 @@ class Program_weight_tensor_parameter_144: class Program_weight_tensor_parameter_145: name = "parameter_145" + original_name = "batch_norm2d_172.w_0" shape = [256] dtype = "float32" min_val = float("0.708988") @@ -1602,6 +1748,7 @@ class Program_weight_tensor_parameter_145: class Program_weight_tensor_parameter_146: name = "parameter_146" + original_name = "batch_norm2d_172.w_2" shape = [256] dtype = "float32" min_val = float("3.76937") @@ -1613,6 +1760,7 @@ class Program_weight_tensor_parameter_146: class Program_weight_tensor_parameter_147: name = "parameter_147" + original_name = "batch_norm2d_172.w_1" shape = [256] dtype = "float32" min_val = float("-1.40143") @@ -1624,6 +1772,7 @@ class Program_weight_tensor_parameter_147: class Program_weight_tensor_parameter_148: name = "parameter_148" + original_name = "conv2d_113.w_0" shape = [256, 256, 1, 1] dtype = "float32" min_val = float("-0.438309") @@ -1635,6 +1784,7 @@ class Program_weight_tensor_parameter_148: class Program_weight_tensor_parameter_149: name = "parameter_149" + original_name = "batch_norm2d_171.b_0" shape = [256] dtype = "float32" min_val = float("-0.19543") @@ -1646,6 +1796,7 @@ class Program_weight_tensor_parameter_149: class Program_weight_tensor_parameter_150: name = "parameter_150" + original_name = "batch_norm2d_171.w_0" shape = [256] dtype = "float32" min_val = float("0.639519") @@ -1657,6 +1808,7 @@ class Program_weight_tensor_parameter_150: class Program_weight_tensor_parameter_151: name = "parameter_151" + original_name = "batch_norm2d_171.w_2" shape = [256] dtype = "float32" min_val = float("33.9632") @@ -1668,6 +1820,7 @@ class Program_weight_tensor_parameter_151: class Program_weight_tensor_parameter_152: name = "parameter_152" + original_name = "batch_norm2d_171.w_1" shape = [256] dtype = "float32" min_val = float("-1.43804") @@ -1679,6 +1832,7 @@ class Program_weight_tensor_parameter_152: class Program_weight_tensor_parameter_153: name = "parameter_153" + original_name = "conv2d_112.w_0" shape = [256, 256, 1, 1] dtype = "float32" min_val = float("-0.558902") @@ -1690,6 +1844,7 @@ class Program_weight_tensor_parameter_153: class Program_weight_tensor_parameter_154: name = "parameter_154" + original_name = "batch_norm2d_170.b_0" shape = [256] dtype = "float32" min_val = float("-0.210248") @@ -1701,6 +1856,7 @@ class Program_weight_tensor_parameter_154: class Program_weight_tensor_parameter_155: name = "parameter_155" + original_name = "batch_norm2d_170.w_0" shape = [256] dtype = "float32" min_val = float("0.584569") @@ -1712,6 +1868,7 @@ class Program_weight_tensor_parameter_155: class Program_weight_tensor_parameter_156: name = "parameter_156" + original_name = "batch_norm2d_170.w_2" shape = [256] dtype = "float32" min_val = float("17.1254") @@ -1723,6 +1880,7 @@ class Program_weight_tensor_parameter_156: class Program_weight_tensor_parameter_157: name = "parameter_157" + original_name = "batch_norm2d_170.w_1" shape = [256] dtype = "float32" min_val = float("-0.661493") @@ -1734,6 +1892,7 @@ class Program_weight_tensor_parameter_157: class Program_weight_tensor_parameter_158: name = "parameter_158" + original_name = "conv2d_111.w_0" shape = [256, 256, 1, 1] dtype = "float32" min_val = float("-0.570548") diff --git a/paddle_samples/PaddleX/Mask-RT-DETR-S/subgraph_10/input_meta.py b/paddle_samples/PaddleX/Mask-RT-DETR-S/subgraph_10/input_meta.py index 987c9a5f6..0d4a08448 100644 --- a/paddle_samples/PaddleX/Mask-RT-DETR-S/subgraph_10/input_meta.py +++ b/paddle_samples/PaddleX/Mask-RT-DETR-S/subgraph_10/input_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_data_0: name = "data_0" + original_name = "param_10" shape = [2, 256] dtype = "float32" min_val = float("-2.77622") @@ -11,6 +12,7 @@ class Program_weight_tensor_data_0: class Program_weight_tensor_data_1: name = "data_1" + original_name = "param_11" shape = [256, 768] dtype = "float32" min_val = float("-0.443256") @@ -22,6 +24,7 @@ class Program_weight_tensor_data_1: class Program_weight_tensor_data_2: name = "data_2" + original_name = "param_12" shape = [768] dtype = "float32" min_val = float("-0.256517") @@ -33,6 +36,7 @@ class Program_weight_tensor_data_2: class Program_weight_tensor_data_3: name = "data_3" + original_name = "param_13" shape = [256, 768] dtype = "float32" min_val = float("-0.428645") @@ -44,6 +48,7 @@ class Program_weight_tensor_data_3: class Program_weight_tensor_data_4: name = "data_4" + original_name = "param_14" shape = [768] dtype = "float32" min_val = float("-0.203884") @@ -55,6 +60,7 @@ class Program_weight_tensor_data_4: class Program_weight_tensor_data_5: name = "data_5" + original_name = "param_15" shape = [256, 768] dtype = "float32" min_val = float("-0.45735") @@ -66,6 +72,7 @@ class Program_weight_tensor_data_5: class Program_weight_tensor_data_6: name = "data_6" + original_name = "param_16" shape = [768] dtype = "float32" min_val = float("-0.173772") @@ -77,6 +84,7 @@ class Program_weight_tensor_data_6: class Program_weight_tensor_data_7: name = "data_7" + original_name = "param_17" shape = [256, 768] dtype = "float32" min_val = float("-0.440415") @@ -88,6 +96,7 @@ class Program_weight_tensor_data_7: class Program_weight_tensor_data_8: name = "data_8" + original_name = "param_18" shape = [768] dtype = "float32" min_val = float("-0.132569") @@ -99,6 +108,7 @@ class Program_weight_tensor_data_8: class Program_weight_tensor_data_9: name = "data_9" + original_name = "param_19" shape = [256, 768] dtype = "float32" min_val = float("-0.43989") @@ -110,6 +120,7 @@ class Program_weight_tensor_data_9: class Program_weight_tensor_data_10: name = "data_10" + original_name = "param_20" shape = [768] dtype = "float32" min_val = float("-0.0943557") @@ -121,6 +132,7 @@ class Program_weight_tensor_data_10: class Program_weight_tensor_data_11: name = "data_11" + original_name = "param_21" shape = [256, 768] dtype = "float32" min_val = float("-0.409089") @@ -132,6 +144,7 @@ class Program_weight_tensor_data_11: class Program_weight_tensor_data_12: name = "data_12" + original_name = "param_22" shape = [768] dtype = "float32" min_val = float("-0.0834688") @@ -143,6 +156,7 @@ class Program_weight_tensor_data_12: class Program_weight_tensor_data_13: name = "data_13" + original_name = "var_1511" shape = [1, 256, 80, 80] dtype = "float32" min_val = float("-0.278465") @@ -154,6 +168,7 @@ class Program_weight_tensor_data_13: class Program_weight_tensor_data_14: name = "data_14" + original_name = "var_1512" shape = [1, 256, 40, 40] dtype = "float32" min_val = float("-0.278465") @@ -165,6 +180,7 @@ class Program_weight_tensor_data_14: class Program_weight_tensor_data_15: name = "data_15" + original_name = "var_1513" shape = [1, 256, 20, 20] dtype = "float32" min_val = float("-0.278465") @@ -176,6 +192,7 @@ class Program_weight_tensor_data_15: class Program_weight_tensor_data_16: name = "data_16" + original_name = "var_1514" shape = [1, 128, 160, 160] dtype = "float32" min_val = float("-68.1084") @@ -187,6 +204,7 @@ class Program_weight_tensor_data_16: class Program_weight_tensor_data_17: name = "data_17" + original_name = "var_1517" shape = [1, 1] dtype = "int32" data = [0] @@ -194,6 +212,7 @@ class Program_weight_tensor_data_17: class Program_weight_tensor_data_18: name = "data_18" + original_name = "var_1518" shape = [1, 4] dtype = "float32" data = [0.814055, 0.448795, 0.37189, 0.66179] diff --git a/paddle_samples/PaddleX/Mask-RT-DETR-S/subgraph_10/weight_meta.py b/paddle_samples/PaddleX/Mask-RT-DETR-S/subgraph_10/weight_meta.py index afc2fff48..b6b7ff8ca 100644 --- a/paddle_samples/PaddleX/Mask-RT-DETR-S/subgraph_10/weight_meta.py +++ b/paddle_samples/PaddleX/Mask-RT-DETR-S/subgraph_10/weight_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_parameter_0: name = "parameter_0" + original_name = "layer_norm_4.b_0_deepcopy_143" shape = [256] dtype = "float32" min_val = float("-0.257064") @@ -11,6 +12,7 @@ class Program_weight_tensor_parameter_0: class Program_weight_tensor_parameter_1: name = "parameter_1" + original_name = "layer_norm_4.w_0_deepcopy_142" shape = [256] dtype = "float32" min_val = float("0.681351") @@ -22,6 +24,7 @@ class Program_weight_tensor_parameter_1: class Program_weight_tensor_parameter_2: name = "parameter_2" + original_name = "linear_9.b_0_deepcopy_141" shape = [256] dtype = "float32" min_val = float("-0.225936") @@ -33,6 +36,7 @@ class Program_weight_tensor_parameter_2: class Program_weight_tensor_parameter_3: name = "parameter_3" + original_name = "linear_9.w_0_deepcopy_140" shape = [1024, 256] dtype = "float32" min_val = float("-0.479777") @@ -44,6 +48,7 @@ class Program_weight_tensor_parameter_3: class Program_weight_tensor_parameter_4: name = "parameter_4" + original_name = "linear_8.b_0_deepcopy_139" shape = [1024] dtype = "float32" min_val = float("-0.175465") @@ -55,6 +60,7 @@ class Program_weight_tensor_parameter_4: class Program_weight_tensor_parameter_5: name = "parameter_5" + original_name = "linear_8.w_0_deepcopy_138" shape = [256, 1024] dtype = "float32" min_val = float("-0.394227") @@ -66,6 +72,7 @@ class Program_weight_tensor_parameter_5: class Program_weight_tensor_parameter_6: name = "parameter_6" + original_name = "layer_norm_3.b_0_deepcopy_137" shape = [256] dtype = "float32" min_val = float("-0.148679") @@ -77,6 +84,7 @@ class Program_weight_tensor_parameter_6: class Program_weight_tensor_parameter_7: name = "parameter_7" + original_name = "layer_norm_3.w_0_deepcopy_136" shape = [256] dtype = "float32" min_val = float("0.811588") @@ -88,6 +96,7 @@ class Program_weight_tensor_parameter_7: class Program_weight_tensor_parameter_8: name = "parameter_8" + original_name = "linear_7.b_0_deepcopy_135" shape = [256] dtype = "float32" min_val = float("-0.16884") @@ -99,6 +108,7 @@ class Program_weight_tensor_parameter_8: class Program_weight_tensor_parameter_9: name = "parameter_9" + original_name = "linear_7.w_0_deepcopy_134" shape = [256, 256] dtype = "float32" min_val = float("-0.296648") @@ -110,6 +120,7 @@ class Program_weight_tensor_parameter_9: class Program_weight_tensor_parameter_10: name = "parameter_10" + original_name = "linear_5.b_0_deepcopy_131" shape = [96] dtype = "float32" min_val = float("-0.10222") @@ -121,6 +132,7 @@ class Program_weight_tensor_parameter_10: class Program_weight_tensor_parameter_11: name = "parameter_11" + original_name = "linear_5.w_0_deepcopy_130" shape = [256, 96] dtype = "float32" min_val = float("-0.434355") @@ -132,6 +144,7 @@ class Program_weight_tensor_parameter_11: class Program_weight_tensor_parameter_12: name = "parameter_12" + original_name = "linear_4.b_0_deepcopy_129" shape = [192] dtype = "float32" min_val = float("-4.26608") @@ -143,6 +156,7 @@ class Program_weight_tensor_parameter_12: class Program_weight_tensor_parameter_13: name = "parameter_13" + original_name = "linear_4.w_0_deepcopy_128" shape = [256, 192] dtype = "float32" min_val = float("-0.534499") @@ -154,6 +168,7 @@ class Program_weight_tensor_parameter_13: class Program_weight_tensor_parameter_14: name = "parameter_14" + original_name = "linear_6.b_0_deepcopy_133" shape = [256] dtype = "float32" min_val = float("-0.0804551") @@ -165,6 +180,7 @@ class Program_weight_tensor_parameter_14: class Program_weight_tensor_parameter_15: name = "parameter_15" + original_name = "linear_6.w_0_deepcopy_132" shape = [256, 256] dtype = "float32" min_val = float("-0.274243") @@ -176,6 +192,7 @@ class Program_weight_tensor_parameter_15: class Program_weight_tensor_parameter_16: name = "parameter_16" + original_name = "layer_norm_2.b_0_deepcopy_127" shape = [256] dtype = "float32" min_val = float("-0.166604") @@ -187,6 +204,7 @@ class Program_weight_tensor_parameter_16: class Program_weight_tensor_parameter_17: name = "parameter_17" + original_name = "layer_norm_2.w_0_deepcopy_126" shape = [256] dtype = "float32" min_val = float("0.967437") @@ -198,6 +216,7 @@ class Program_weight_tensor_parameter_17: class Program_weight_tensor_parameter_18: name = "parameter_18" + original_name = "linear_3.b_0_deepcopy_125" shape = [256] dtype = "float32" min_val = float("-0.599141") @@ -209,6 +228,7 @@ class Program_weight_tensor_parameter_18: class Program_weight_tensor_parameter_19: name = "parameter_19" + original_name = "linear_3.w_0_deepcopy_124" shape = [256, 256] dtype = "float32" min_val = float("-0.50621") @@ -220,6 +240,7 @@ class Program_weight_tensor_parameter_19: class Program_weight_tensor_parameter_20: name = "parameter_20" + original_name = "layer_norm_4.b_0_deepcopy_121" shape = [256] dtype = "float32" min_val = float("-0.166835") @@ -231,6 +252,7 @@ class Program_weight_tensor_parameter_20: class Program_weight_tensor_parameter_21: name = "parameter_21" + original_name = "layer_norm_4.w_0_deepcopy_120" shape = [256] dtype = "float32" min_val = float("0.680908") @@ -242,6 +264,7 @@ class Program_weight_tensor_parameter_21: class Program_weight_tensor_parameter_22: name = "parameter_22" + original_name = "linear_9.b_0_deepcopy_119" shape = [256] dtype = "float32" min_val = float("-0.17171") @@ -253,6 +276,7 @@ class Program_weight_tensor_parameter_22: class Program_weight_tensor_parameter_23: name = "parameter_23" + original_name = "linear_9.w_0_deepcopy_118" shape = [1024, 256] dtype = "float32" min_val = float("-0.386224") @@ -264,6 +288,7 @@ class Program_weight_tensor_parameter_23: class Program_weight_tensor_parameter_24: name = "parameter_24" + original_name = "linear_8.b_0_deepcopy_117" shape = [1024] dtype = "float32" min_val = float("-0.176651") @@ -275,6 +300,7 @@ class Program_weight_tensor_parameter_24: class Program_weight_tensor_parameter_25: name = "parameter_25" + original_name = "linear_8.w_0_deepcopy_116" shape = [256, 1024] dtype = "float32" min_val = float("-0.372818") @@ -286,6 +312,7 @@ class Program_weight_tensor_parameter_25: class Program_weight_tensor_parameter_26: name = "parameter_26" + original_name = "layer_norm_3.b_0_deepcopy_115" shape = [256] dtype = "float32" min_val = float("-0.125251") @@ -297,6 +324,7 @@ class Program_weight_tensor_parameter_26: class Program_weight_tensor_parameter_27: name = "parameter_27" + original_name = "layer_norm_3.w_0_deepcopy_114" shape = [256] dtype = "float32" min_val = float("0.85913") @@ -308,6 +336,7 @@ class Program_weight_tensor_parameter_27: class Program_weight_tensor_parameter_28: name = "parameter_28" + original_name = "linear_7.b_0_deepcopy_113" shape = [256] dtype = "float32" min_val = float("-0.139561") @@ -319,6 +348,7 @@ class Program_weight_tensor_parameter_28: class Program_weight_tensor_parameter_29: name = "parameter_29" + original_name = "linear_7.w_0_deepcopy_112" shape = [256, 256] dtype = "float32" min_val = float("-0.315092") @@ -330,6 +360,7 @@ class Program_weight_tensor_parameter_29: class Program_weight_tensor_parameter_30: name = "parameter_30" + original_name = "linear_5.b_0_deepcopy_109" shape = [96] dtype = "float32" min_val = float("-0.0653593") @@ -341,6 +372,7 @@ class Program_weight_tensor_parameter_30: class Program_weight_tensor_parameter_31: name = "parameter_31" + original_name = "linear_5.w_0_deepcopy_108" shape = [256, 96] dtype = "float32" min_val = float("-0.347212") @@ -352,6 +384,7 @@ class Program_weight_tensor_parameter_31: class Program_weight_tensor_parameter_32: name = "parameter_32" + original_name = "linear_4.b_0_deepcopy_107" shape = [192] dtype = "float32" min_val = float("-4.22646") @@ -363,6 +396,7 @@ class Program_weight_tensor_parameter_32: class Program_weight_tensor_parameter_33: name = "parameter_33" + original_name = "linear_4.w_0_deepcopy_106" shape = [256, 192] dtype = "float32" min_val = float("-0.48362") @@ -374,6 +408,7 @@ class Program_weight_tensor_parameter_33: class Program_weight_tensor_parameter_34: name = "parameter_34" + original_name = "linear_6.b_0_deepcopy_111" shape = [256] dtype = "float32" min_val = float("-0.17612") @@ -385,6 +420,7 @@ class Program_weight_tensor_parameter_34: class Program_weight_tensor_parameter_35: name = "parameter_35" + original_name = "linear_6.w_0_deepcopy_110" shape = [256, 256] dtype = "float32" min_val = float("-0.312752") @@ -396,6 +432,7 @@ class Program_weight_tensor_parameter_35: class Program_weight_tensor_parameter_36: name = "parameter_36" + original_name = "layer_norm_2.b_0_deepcopy_105" shape = [256] dtype = "float32" min_val = float("-0.173944") @@ -407,6 +444,7 @@ class Program_weight_tensor_parameter_36: class Program_weight_tensor_parameter_37: name = "parameter_37" + original_name = "layer_norm_2.w_0_deepcopy_104" shape = [256] dtype = "float32" min_val = float("0.83639") @@ -418,6 +456,7 @@ class Program_weight_tensor_parameter_37: class Program_weight_tensor_parameter_38: name = "parameter_38" + original_name = "linear_3.b_0_deepcopy_103" shape = [256] dtype = "float32" min_val = float("-0.232223") @@ -429,6 +468,7 @@ class Program_weight_tensor_parameter_38: class Program_weight_tensor_parameter_39: name = "parameter_39" + original_name = "linear_3.w_0_deepcopy_102" shape = [256, 256] dtype = "float32" min_val = float("-0.323369") @@ -440,6 +480,7 @@ class Program_weight_tensor_parameter_39: class Program_weight_tensor_parameter_40: name = "parameter_40" + original_name = "layer_norm_4.b_0_deepcopy_99" shape = [256] dtype = "float32" min_val = float("-0.11938") @@ -451,6 +492,7 @@ class Program_weight_tensor_parameter_40: class Program_weight_tensor_parameter_41: name = "parameter_41" + original_name = "layer_norm_4.w_0_deepcopy_98" shape = [256] dtype = "float32" min_val = float("0.703938") @@ -462,6 +504,7 @@ class Program_weight_tensor_parameter_41: class Program_weight_tensor_parameter_42: name = "parameter_42" + original_name = "linear_9.b_0_deepcopy_97" shape = [256] dtype = "float32" min_val = float("-0.112619") @@ -473,6 +516,7 @@ class Program_weight_tensor_parameter_42: class Program_weight_tensor_parameter_43: name = "parameter_43" + original_name = "linear_9.w_0_deepcopy_96" shape = [1024, 256] dtype = "float32" min_val = float("-0.429034") @@ -484,6 +528,7 @@ class Program_weight_tensor_parameter_43: class Program_weight_tensor_parameter_44: name = "parameter_44" + original_name = "linear_8.b_0_deepcopy_95" shape = [1024] dtype = "float32" min_val = float("-0.20779") @@ -495,6 +540,7 @@ class Program_weight_tensor_parameter_44: class Program_weight_tensor_parameter_45: name = "parameter_45" + original_name = "linear_8.w_0_deepcopy_94" shape = [256, 1024] dtype = "float32" min_val = float("-0.392607") @@ -506,6 +552,7 @@ class Program_weight_tensor_parameter_45: class Program_weight_tensor_parameter_46: name = "parameter_46" + original_name = "layer_norm_3.b_0_deepcopy_93" shape = [256] dtype = "float32" min_val = float("-0.141018") @@ -517,6 +564,7 @@ class Program_weight_tensor_parameter_46: class Program_weight_tensor_parameter_47: name = "parameter_47" + original_name = "layer_norm_3.w_0_deepcopy_92" shape = [256] dtype = "float32" min_val = float("0.904757") @@ -528,6 +576,7 @@ class Program_weight_tensor_parameter_47: class Program_weight_tensor_parameter_48: name = "parameter_48" + original_name = "linear_7.b_0_deepcopy_91" shape = [256] dtype = "float32" min_val = float("-0.134183") @@ -539,6 +588,7 @@ class Program_weight_tensor_parameter_48: class Program_weight_tensor_parameter_49: name = "parameter_49" + original_name = "linear_7.w_0_deepcopy_90" shape = [256, 256] dtype = "float32" min_val = float("-0.331426") @@ -550,6 +600,7 @@ class Program_weight_tensor_parameter_49: class Program_weight_tensor_parameter_50: name = "parameter_50" + original_name = "linear_5.b_0_deepcopy_87" shape = [96] dtype = "float32" min_val = float("-0.210064") @@ -561,6 +612,7 @@ class Program_weight_tensor_parameter_50: class Program_weight_tensor_parameter_51: name = "parameter_51" + original_name = "linear_5.w_0_deepcopy_86" shape = [256, 96] dtype = "float32" min_val = float("-0.393653") @@ -572,6 +624,7 @@ class Program_weight_tensor_parameter_51: class Program_weight_tensor_parameter_52: name = "parameter_52" + original_name = "linear_4.b_0_deepcopy_85" shape = [192] dtype = "float32" min_val = float("-4.25234") @@ -583,6 +636,7 @@ class Program_weight_tensor_parameter_52: class Program_weight_tensor_parameter_53: name = "parameter_53" + original_name = "linear_4.w_0_deepcopy_84" shape = [256, 192] dtype = "float32" min_val = float("-0.508175") @@ -594,6 +648,7 @@ class Program_weight_tensor_parameter_53: class Program_weight_tensor_parameter_54: name = "parameter_54" + original_name = "linear_6.b_0_deepcopy_89" shape = [256] dtype = "float32" min_val = float("-0.137544") @@ -605,6 +660,7 @@ class Program_weight_tensor_parameter_54: class Program_weight_tensor_parameter_55: name = "parameter_55" + original_name = "linear_6.w_0_deepcopy_88" shape = [256, 256] dtype = "float32" min_val = float("-0.352594") @@ -616,6 +672,7 @@ class Program_weight_tensor_parameter_55: class Program_weight_tensor_parameter_56: name = "parameter_56" + original_name = "layer_norm_2.b_0_deepcopy_83" shape = [256] dtype = "float32" min_val = float("-0.326569") @@ -627,6 +684,7 @@ class Program_weight_tensor_parameter_56: class Program_weight_tensor_parameter_57: name = "parameter_57" + original_name = "layer_norm_2.w_0_deepcopy_82" shape = [256] dtype = "float32" min_val = float("0.733128") @@ -638,6 +696,7 @@ class Program_weight_tensor_parameter_57: class Program_weight_tensor_parameter_58: name = "parameter_58" + original_name = "linear_3.b_0_deepcopy_81" shape = [256] dtype = "float32" min_val = float("-0.18088") @@ -649,6 +708,7 @@ class Program_weight_tensor_parameter_58: class Program_weight_tensor_parameter_59: name = "parameter_59" + original_name = "linear_3.w_0_deepcopy_80" shape = [256, 256] dtype = "float32" min_val = float("-0.264534") @@ -660,6 +720,7 @@ class Program_weight_tensor_parameter_59: class Program_weight_tensor_parameter_60: name = "parameter_60" + original_name = "layer_norm_4.b_0_deepcopy_77" shape = [256] dtype = "float32" min_val = float("-0.146108") @@ -671,6 +732,7 @@ class Program_weight_tensor_parameter_60: class Program_weight_tensor_parameter_61: name = "parameter_61" + original_name = "layer_norm_4.w_0_deepcopy_76" shape = [256] dtype = "float32" min_val = float("0.750443") @@ -682,6 +744,7 @@ class Program_weight_tensor_parameter_61: class Program_weight_tensor_parameter_62: name = "parameter_62" + original_name = "linear_9.b_0_deepcopy_75" shape = [256] dtype = "float32" min_val = float("-0.155642") @@ -693,6 +756,7 @@ class Program_weight_tensor_parameter_62: class Program_weight_tensor_parameter_63: name = "parameter_63" + original_name = "linear_9.w_0_deepcopy_74" shape = [1024, 256] dtype = "float32" min_val = float("-0.420891") @@ -704,6 +768,7 @@ class Program_weight_tensor_parameter_63: class Program_weight_tensor_parameter_64: name = "parameter_64" + original_name = "linear_8.b_0_deepcopy_73" shape = [1024] dtype = "float32" min_val = float("-0.199189") @@ -715,6 +780,7 @@ class Program_weight_tensor_parameter_64: class Program_weight_tensor_parameter_65: name = "parameter_65" + original_name = "linear_8.w_0_deepcopy_72" shape = [256, 1024] dtype = "float32" min_val = float("-0.378273") @@ -726,6 +792,7 @@ class Program_weight_tensor_parameter_65: class Program_weight_tensor_parameter_66: name = "parameter_66" + original_name = "layer_norm_3.b_0_deepcopy_71" shape = [256] dtype = "float32" min_val = float("-0.148874") @@ -737,6 +804,7 @@ class Program_weight_tensor_parameter_66: class Program_weight_tensor_parameter_67: name = "parameter_67" + original_name = "layer_norm_3.w_0_deepcopy_70" shape = [256] dtype = "float32" min_val = float("0.88501") @@ -748,6 +816,7 @@ class Program_weight_tensor_parameter_67: class Program_weight_tensor_parameter_68: name = "parameter_68" + original_name = "linear_7.b_0_deepcopy_69" shape = [256] dtype = "float32" min_val = float("-0.129583") @@ -759,6 +828,7 @@ class Program_weight_tensor_parameter_68: class Program_weight_tensor_parameter_69: name = "parameter_69" + original_name = "linear_7.w_0_deepcopy_68" shape = [256, 256] dtype = "float32" min_val = float("-0.375377") @@ -770,6 +840,7 @@ class Program_weight_tensor_parameter_69: class Program_weight_tensor_parameter_70: name = "parameter_70" + original_name = "linear_5.b_0_deepcopy_65" shape = [96] dtype = "float32" min_val = float("-0.176126") @@ -781,6 +852,7 @@ class Program_weight_tensor_parameter_70: class Program_weight_tensor_parameter_71: name = "parameter_71" + original_name = "linear_5.w_0_deepcopy_64" shape = [256, 96] dtype = "float32" min_val = float("-0.510471") @@ -792,6 +864,7 @@ class Program_weight_tensor_parameter_71: class Program_weight_tensor_parameter_72: name = "parameter_72" + original_name = "linear_4.b_0_deepcopy_63" shape = [192] dtype = "float32" min_val = float("-4.40858") @@ -803,6 +876,7 @@ class Program_weight_tensor_parameter_72: class Program_weight_tensor_parameter_73: name = "parameter_73" + original_name = "linear_4.w_0_deepcopy_62" shape = [256, 192] dtype = "float32" min_val = float("-0.538114") @@ -814,6 +888,7 @@ class Program_weight_tensor_parameter_73: class Program_weight_tensor_parameter_74: name = "parameter_74" + original_name = "linear_6.b_0_deepcopy_67" shape = [256] dtype = "float32" min_val = float("-0.164598") @@ -825,6 +900,7 @@ class Program_weight_tensor_parameter_74: class Program_weight_tensor_parameter_75: name = "parameter_75" + original_name = "linear_6.w_0_deepcopy_66" shape = [256, 256] dtype = "float32" min_val = float("-0.341213") @@ -836,6 +912,7 @@ class Program_weight_tensor_parameter_75: class Program_weight_tensor_parameter_76: name = "parameter_76" + original_name = "layer_norm_2.b_0_deepcopy_61" shape = [256] dtype = "float32" min_val = float("-0.222443") @@ -847,6 +924,7 @@ class Program_weight_tensor_parameter_76: class Program_weight_tensor_parameter_77: name = "parameter_77" + original_name = "layer_norm_2.w_0_deepcopy_60" shape = [256] dtype = "float32" min_val = float("0.456413") @@ -858,6 +936,7 @@ class Program_weight_tensor_parameter_77: class Program_weight_tensor_parameter_78: name = "parameter_78" + original_name = "linear_3.b_0_deepcopy_59" shape = [256] dtype = "float32" min_val = float("-0.191954") @@ -869,6 +948,7 @@ class Program_weight_tensor_parameter_78: class Program_weight_tensor_parameter_79: name = "parameter_79" + original_name = "linear_3.w_0_deepcopy_58" shape = [256, 256] dtype = "float32" min_val = float("-0.307968") @@ -880,6 +960,7 @@ class Program_weight_tensor_parameter_79: class Program_weight_tensor_parameter_80: name = "parameter_80" + original_name = "layer_norm_4.b_0_deepcopy_55" shape = [256] dtype = "float32" min_val = float("-0.363536") @@ -891,6 +972,7 @@ class Program_weight_tensor_parameter_80: class Program_weight_tensor_parameter_81: name = "parameter_81" + original_name = "layer_norm_4.w_0_deepcopy_54" shape = [256] dtype = "float32" min_val = float("0.788861") @@ -902,6 +984,7 @@ class Program_weight_tensor_parameter_81: class Program_weight_tensor_parameter_82: name = "parameter_82" + original_name = "linear_9.b_0_deepcopy_53" shape = [256] dtype = "float32" min_val = float("-0.283662") @@ -913,6 +996,7 @@ class Program_weight_tensor_parameter_82: class Program_weight_tensor_parameter_83: name = "parameter_83" + original_name = "linear_9.w_0_deepcopy_52" shape = [1024, 256] dtype = "float32" min_val = float("-0.564152") @@ -924,6 +1008,7 @@ class Program_weight_tensor_parameter_83: class Program_weight_tensor_parameter_84: name = "parameter_84" + original_name = "linear_8.b_0_deepcopy_51" shape = [1024] dtype = "float32" min_val = float("-0.37974") @@ -935,6 +1020,7 @@ class Program_weight_tensor_parameter_84: class Program_weight_tensor_parameter_85: name = "parameter_85" + original_name = "linear_8.w_0_deepcopy_50" shape = [256, 1024] dtype = "float32" min_val = float("-0.423302") @@ -946,6 +1032,7 @@ class Program_weight_tensor_parameter_85: class Program_weight_tensor_parameter_86: name = "parameter_86" + original_name = "layer_norm_3.b_0_deepcopy_49" shape = [256] dtype = "float32" min_val = float("-0.136404") @@ -957,6 +1044,7 @@ class Program_weight_tensor_parameter_86: class Program_weight_tensor_parameter_87: name = "parameter_87" + original_name = "layer_norm_3.w_0_deepcopy_48" shape = [256] dtype = "float32" min_val = float("0.859837") @@ -968,6 +1056,7 @@ class Program_weight_tensor_parameter_87: class Program_weight_tensor_parameter_88: name = "parameter_88" + original_name = "linear_7.b_0_deepcopy_47" shape = [256] dtype = "float32" min_val = float("-0.120963") @@ -979,6 +1068,7 @@ class Program_weight_tensor_parameter_88: class Program_weight_tensor_parameter_89: name = "parameter_89" + original_name = "linear_7.w_0_deepcopy_46" shape = [256, 256] dtype = "float32" min_val = float("-0.354921") @@ -990,6 +1080,7 @@ class Program_weight_tensor_parameter_89: class Program_weight_tensor_parameter_90: name = "parameter_90" + original_name = "linear_5.b_0_deepcopy_43" shape = [96] dtype = "float32" min_val = float("-0.112888") @@ -1001,6 +1092,7 @@ class Program_weight_tensor_parameter_90: class Program_weight_tensor_parameter_91: name = "parameter_91" + original_name = "linear_5.w_0_deepcopy_42" shape = [256, 96] dtype = "float32" min_val = float("-0.60773") @@ -1012,6 +1104,7 @@ class Program_weight_tensor_parameter_91: class Program_weight_tensor_parameter_92: name = "parameter_92" + original_name = "linear_4.b_0_deepcopy_41" shape = [192] dtype = "float32" min_val = float("-4.22324") @@ -1023,6 +1116,7 @@ class Program_weight_tensor_parameter_92: class Program_weight_tensor_parameter_93: name = "parameter_93" + original_name = "linear_4.w_0_deepcopy_40" shape = [256, 192] dtype = "float32" min_val = float("-0.621387") @@ -1034,6 +1128,7 @@ class Program_weight_tensor_parameter_93: class Program_weight_tensor_parameter_94: name = "parameter_94" + original_name = "linear_6.b_0_deepcopy_45" shape = [256] dtype = "float32" min_val = float("-0.208168") @@ -1045,6 +1140,7 @@ class Program_weight_tensor_parameter_94: class Program_weight_tensor_parameter_95: name = "parameter_95" + original_name = "linear_6.w_0_deepcopy_44" shape = [256, 256] dtype = "float32" min_val = float("-0.327069") @@ -1056,6 +1152,7 @@ class Program_weight_tensor_parameter_95: class Program_weight_tensor_parameter_96: name = "parameter_96" + original_name = "layer_norm_2.b_0_deepcopy_39" shape = [256] dtype = "float32" min_val = float("-0.298101") @@ -1067,6 +1164,7 @@ class Program_weight_tensor_parameter_96: class Program_weight_tensor_parameter_97: name = "parameter_97" + original_name = "layer_norm_2.w_0_deepcopy_38" shape = [256] dtype = "float32" min_val = float("0.633658") @@ -1078,6 +1176,7 @@ class Program_weight_tensor_parameter_97: class Program_weight_tensor_parameter_98: name = "parameter_98" + original_name = "linear_3.b_0_deepcopy_37" shape = [256] dtype = "float32" min_val = float("-0.269304") @@ -1089,6 +1188,7 @@ class Program_weight_tensor_parameter_98: class Program_weight_tensor_parameter_99: name = "parameter_99" + original_name = "linear_3.w_0_deepcopy_36" shape = [256, 256] dtype = "float32" min_val = float("-0.425369") @@ -1100,6 +1200,7 @@ class Program_weight_tensor_parameter_99: class Program_weight_tensor_parameter_100: name = "parameter_100" + original_name = "layer_norm_4.b_0_deepcopy_33" shape = [256] dtype = "float32" min_val = float("-0.193777") @@ -1111,6 +1212,7 @@ class Program_weight_tensor_parameter_100: class Program_weight_tensor_parameter_101: name = "parameter_101" + original_name = "layer_norm_4.w_0_deepcopy_32" shape = [256] dtype = "float32" min_val = float("0.715436") @@ -1122,6 +1224,7 @@ class Program_weight_tensor_parameter_101: class Program_weight_tensor_parameter_102: name = "parameter_102" + original_name = "linear_9.b_0_deepcopy_31" shape = [256] dtype = "float32" min_val = float("-0.332257") @@ -1133,6 +1236,7 @@ class Program_weight_tensor_parameter_102: class Program_weight_tensor_parameter_103: name = "parameter_103" + original_name = "linear_9.w_0_deepcopy_30" shape = [1024, 256] dtype = "float32" min_val = float("-0.645388") @@ -1144,6 +1248,7 @@ class Program_weight_tensor_parameter_103: class Program_weight_tensor_parameter_104: name = "parameter_104" + original_name = "linear_8.b_0_deepcopy_29" shape = [1024] dtype = "float32" min_val = float("-0.297516") @@ -1155,6 +1260,7 @@ class Program_weight_tensor_parameter_104: class Program_weight_tensor_parameter_105: name = "parameter_105" + original_name = "linear_8.w_0_deepcopy_28" shape = [256, 1024] dtype = "float32" min_val = float("-0.439839") @@ -1166,6 +1272,7 @@ class Program_weight_tensor_parameter_105: class Program_weight_tensor_parameter_106: name = "parameter_106" + original_name = "layer_norm_3.b_0_deepcopy_27" shape = [256] dtype = "float32" min_val = float("-0.221757") @@ -1177,6 +1284,7 @@ class Program_weight_tensor_parameter_106: class Program_weight_tensor_parameter_107: name = "parameter_107" + original_name = "layer_norm_3.w_0_deepcopy_26" shape = [256] dtype = "float32" min_val = float("0.870477") @@ -1188,6 +1296,7 @@ class Program_weight_tensor_parameter_107: class Program_weight_tensor_parameter_108: name = "parameter_108" + original_name = "linear_7.b_0_deepcopy_25" shape = [256] dtype = "float32" min_val = float("-0.102291") @@ -1199,6 +1308,7 @@ class Program_weight_tensor_parameter_108: class Program_weight_tensor_parameter_109: name = "parameter_109" + original_name = "linear_7.w_0_deepcopy_24" shape = [256, 256] dtype = "float32" min_val = float("-0.333316") @@ -1210,6 +1320,7 @@ class Program_weight_tensor_parameter_109: class Program_weight_tensor_parameter_110: name = "parameter_110" + original_name = "linear_5.b_0_deepcopy_21" shape = [96] dtype = "float32" min_val = float("-0.240902") @@ -1221,6 +1332,7 @@ class Program_weight_tensor_parameter_110: class Program_weight_tensor_parameter_111: name = "parameter_111" + original_name = "linear_5.w_0_deepcopy_20" shape = [256, 96] dtype = "float32" min_val = float("-0.477203") @@ -1232,6 +1344,7 @@ class Program_weight_tensor_parameter_111: class Program_weight_tensor_parameter_112: name = "parameter_112" + original_name = "linear_4.b_0_deepcopy_19" shape = [192] dtype = "float32" min_val = float("-4.19973") @@ -1243,6 +1356,7 @@ class Program_weight_tensor_parameter_112: class Program_weight_tensor_parameter_113: name = "parameter_113" + original_name = "linear_4.w_0_deepcopy_18" shape = [256, 192] dtype = "float32" min_val = float("-0.626627") @@ -1254,6 +1368,7 @@ class Program_weight_tensor_parameter_113: class Program_weight_tensor_parameter_114: name = "parameter_114" + original_name = "linear_6.b_0_deepcopy_23" shape = [256] dtype = "float32" min_val = float("-0.284116") @@ -1265,6 +1380,7 @@ class Program_weight_tensor_parameter_114: class Program_weight_tensor_parameter_115: name = "parameter_115" + original_name = "linear_6.w_0_deepcopy_22" shape = [256, 256] dtype = "float32" min_val = float("-0.319663") @@ -1276,6 +1392,7 @@ class Program_weight_tensor_parameter_115: class Program_weight_tensor_parameter_116: name = "parameter_116" + original_name = "layer_norm_2.b_0_deepcopy_17" shape = [256] dtype = "float32" min_val = float("-0.231617") @@ -1287,6 +1404,7 @@ class Program_weight_tensor_parameter_116: class Program_weight_tensor_parameter_117: name = "parameter_117" + original_name = "layer_norm_2.w_0_deepcopy_16" shape = [256] dtype = "float32" min_val = float("0.641496") @@ -1298,6 +1416,7 @@ class Program_weight_tensor_parameter_117: class Program_weight_tensor_parameter_118: name = "parameter_118" + original_name = "linear_3.b_0_deepcopy_15" shape = [256] dtype = "float32" min_val = float("-0.345909") @@ -1309,6 +1428,7 @@ class Program_weight_tensor_parameter_118: class Program_weight_tensor_parameter_119: name = "parameter_119" + original_name = "linear_3.w_0_deepcopy_14" shape = [256, 256] dtype = "float32" min_val = float("-0.397334") @@ -1320,6 +1440,7 @@ class Program_weight_tensor_parameter_119: class Program_weight_tensor_parameter_120: name = "parameter_120" + original_name = "linear_11.b_0" shape = [256] dtype = "float32" min_val = float("-0.175103") @@ -1331,6 +1452,7 @@ class Program_weight_tensor_parameter_120: class Program_weight_tensor_parameter_121: name = "parameter_121" + original_name = "linear_11.w_0" shape = [512, 256] dtype = "float32" min_val = float("-2.82063") @@ -1342,6 +1464,7 @@ class Program_weight_tensor_parameter_121: class Program_weight_tensor_parameter_122: name = "parameter_122" + original_name = "linear_10.b_0" shape = [512] dtype = "float32" min_val = float("-0.497467") @@ -1353,6 +1476,7 @@ class Program_weight_tensor_parameter_122: class Program_weight_tensor_parameter_123: name = "parameter_123" + original_name = "linear_10.w_0" shape = [4, 512] dtype = "float32" min_val = float("-4.9862") @@ -1364,6 +1488,7 @@ class Program_weight_tensor_parameter_123: class Program_weight_tensor_parameter_124: name = "parameter_124" + original_name = "linear_14.b_0" shape = [128] dtype = "float32" min_val = float("-0.121007") @@ -1375,6 +1500,7 @@ class Program_weight_tensor_parameter_124: class Program_weight_tensor_parameter_125: name = "parameter_125" + original_name = "linear_14.w_0" shape = [256, 128] dtype = "float32" min_val = float("-1.10795") @@ -1386,6 +1512,7 @@ class Program_weight_tensor_parameter_125: class Program_weight_tensor_parameter_126: name = "parameter_126" + original_name = "linear_13.b_0" shape = [256] dtype = "float32" min_val = float("-0.440023") @@ -1397,6 +1524,7 @@ class Program_weight_tensor_parameter_126: class Program_weight_tensor_parameter_127: name = "parameter_127" + original_name = "linear_13.w_0" shape = [256, 256] dtype = "float32" min_val = float("-1.7908") @@ -1408,6 +1536,7 @@ class Program_weight_tensor_parameter_127: class Program_weight_tensor_parameter_128: name = "parameter_128" + original_name = "linear_12.b_0" shape = [256] dtype = "float32" min_val = float("-0.591299") @@ -1419,6 +1548,7 @@ class Program_weight_tensor_parameter_128: class Program_weight_tensor_parameter_129: name = "parameter_129" + original_name = "linear_12.w_0" shape = [256, 256] dtype = "float32" min_val = float("-0.442718") @@ -1430,6 +1560,7 @@ class Program_weight_tensor_parameter_129: class Program_weight_tensor_parameter_130: name = "parameter_130" + original_name = "layer_norm_6.b_0" shape = [256] dtype = "float32" min_val = float("-1.03308") @@ -1441,6 +1572,7 @@ class Program_weight_tensor_parameter_130: class Program_weight_tensor_parameter_131: name = "parameter_131" + original_name = "layer_norm_6.w_0" shape = [256] dtype = "float32" min_val = float("0.467465") @@ -1452,6 +1584,7 @@ class Program_weight_tensor_parameter_131: class Program_weight_tensor_parameter_132: name = "parameter_132" + original_name = "linear_19.b_0" shape = [4] dtype = "float32" min_val = float("0") @@ -1461,6 +1594,7 @@ class Program_weight_tensor_parameter_132: class Program_weight_tensor_parameter_133: name = "parameter_133" + original_name = "linear_19.w_0" shape = [256, 4] dtype = "float32" min_val = float("-0.627751") @@ -1472,6 +1606,7 @@ class Program_weight_tensor_parameter_133: class Program_weight_tensor_parameter_134: name = "parameter_134" + original_name = "linear_18.b_0" shape = [256] dtype = "float32" min_val = float("-0.325889") @@ -1483,6 +1618,7 @@ class Program_weight_tensor_parameter_134: class Program_weight_tensor_parameter_135: name = "parameter_135" + original_name = "linear_18.w_0" shape = [256, 256] dtype = "float32" min_val = float("-4.22837") @@ -1494,6 +1630,7 @@ class Program_weight_tensor_parameter_135: class Program_weight_tensor_parameter_136: name = "parameter_136" + original_name = "linear_17.b_0" shape = [256] dtype = "float32" min_val = float("-0.425869") @@ -1505,6 +1642,7 @@ class Program_weight_tensor_parameter_136: class Program_weight_tensor_parameter_137: name = "parameter_137" + original_name = "linear_17.w_0" shape = [256, 256] dtype = "float32" min_val = float("-0.378004") @@ -1516,6 +1654,7 @@ class Program_weight_tensor_parameter_137: class Program_weight_tensor_parameter_138: name = "parameter_138" + original_name = "linear_16.b_0" shape = [2] dtype = "float32" min_val = float("0") @@ -1525,6 +1664,7 @@ class Program_weight_tensor_parameter_138: class Program_weight_tensor_parameter_139: name = "parameter_139" + original_name = "linear_16.w_0" shape = [256, 2] dtype = "float32" min_val = float("-0.0623489") @@ -1536,6 +1676,7 @@ class Program_weight_tensor_parameter_139: class Program_weight_tensor_parameter_140: name = "parameter_140" + original_name = "layer_norm_5.b_0" shape = [256] dtype = "float32" min_val = float("-0.95956") @@ -1547,6 +1688,7 @@ class Program_weight_tensor_parameter_140: class Program_weight_tensor_parameter_141: name = "parameter_141" + original_name = "layer_norm_5.w_0" shape = [256] dtype = "float32" min_val = float("0.729018") @@ -1558,6 +1700,7 @@ class Program_weight_tensor_parameter_141: class Program_weight_tensor_parameter_142: name = "parameter_142" + original_name = "linear_15.b_0" shape = [256] dtype = "float32" min_val = float("-0.842445") @@ -1569,6 +1712,7 @@ class Program_weight_tensor_parameter_142: class Program_weight_tensor_parameter_143: name = "parameter_143" + original_name = "linear_15.w_0" shape = [256, 256] dtype = "float32" min_val = float("-0.54123") @@ -1580,6 +1724,7 @@ class Program_weight_tensor_parameter_143: class Program_weight_tensor_parameter_144: name = "parameter_144" + original_name = "batch_norm2d_136.b_0" shape = [256] dtype = "float32" min_val = float("-0.313374") @@ -1591,6 +1736,7 @@ class Program_weight_tensor_parameter_144: class Program_weight_tensor_parameter_145: name = "parameter_145" + original_name = "batch_norm2d_136.w_0" shape = [256] dtype = "float32" min_val = float("0.784517") @@ -1602,6 +1748,7 @@ class Program_weight_tensor_parameter_145: class Program_weight_tensor_parameter_146: name = "parameter_146" + original_name = "batch_norm2d_136.w_2" shape = [256] dtype = "float32" min_val = float("2.22403") @@ -1613,6 +1760,7 @@ class Program_weight_tensor_parameter_146: class Program_weight_tensor_parameter_147: name = "parameter_147" + original_name = "batch_norm2d_136.w_1" shape = [256] dtype = "float32" min_val = float("-0.867835") @@ -1624,6 +1772,7 @@ class Program_weight_tensor_parameter_147: class Program_weight_tensor_parameter_148: name = "parameter_148" + original_name = "conv2d_95.w_0" shape = [256, 256, 1, 1] dtype = "float32" min_val = float("-0.344502") @@ -1635,6 +1784,7 @@ class Program_weight_tensor_parameter_148: class Program_weight_tensor_parameter_149: name = "parameter_149" + original_name = "batch_norm2d_135.b_0" shape = [256] dtype = "float32" min_val = float("-0.243726") @@ -1646,6 +1796,7 @@ class Program_weight_tensor_parameter_149: class Program_weight_tensor_parameter_150: name = "parameter_150" + original_name = "batch_norm2d_135.w_0" shape = [256] dtype = "float32" min_val = float("0.711516") @@ -1657,6 +1808,7 @@ class Program_weight_tensor_parameter_150: class Program_weight_tensor_parameter_151: name = "parameter_151" + original_name = "batch_norm2d_135.w_2" shape = [256] dtype = "float32" min_val = float("6.30256") @@ -1668,6 +1820,7 @@ class Program_weight_tensor_parameter_151: class Program_weight_tensor_parameter_152: name = "parameter_152" + original_name = "batch_norm2d_135.w_1" shape = [256] dtype = "float32" min_val = float("-0.705253") @@ -1679,6 +1832,7 @@ class Program_weight_tensor_parameter_152: class Program_weight_tensor_parameter_153: name = "parameter_153" + original_name = "conv2d_94.w_0" shape = [256, 256, 1, 1] dtype = "float32" min_val = float("-0.458921") @@ -1690,6 +1844,7 @@ class Program_weight_tensor_parameter_153: class Program_weight_tensor_parameter_154: name = "parameter_154" + original_name = "batch_norm2d_134.b_0" shape = [256] dtype = "float32" min_val = float("-0.415519") @@ -1701,6 +1856,7 @@ class Program_weight_tensor_parameter_154: class Program_weight_tensor_parameter_155: name = "parameter_155" + original_name = "batch_norm2d_134.w_0" shape = [256] dtype = "float32" min_val = float("0.696872") @@ -1712,6 +1868,7 @@ class Program_weight_tensor_parameter_155: class Program_weight_tensor_parameter_156: name = "parameter_156" + original_name = "batch_norm2d_134.w_2" shape = [256] dtype = "float32" min_val = float("3.95818") @@ -1723,6 +1880,7 @@ class Program_weight_tensor_parameter_156: class Program_weight_tensor_parameter_157: name = "parameter_157" + original_name = "batch_norm2d_134.w_1" shape = [256] dtype = "float32" min_val = float("-0.721968") @@ -1734,6 +1892,7 @@ class Program_weight_tensor_parameter_157: class Program_weight_tensor_parameter_158: name = "parameter_158" + original_name = "conv2d_93.w_0" shape = [256, 256, 1, 1] dtype = "float32" min_val = float("-0.420632") diff --git a/paddle_samples/PaddleX/Mask-RT-DETR-S/subgraph_9/input_meta.py b/paddle_samples/PaddleX/Mask-RT-DETR-S/subgraph_9/input_meta.py index 6cfbc9766..9142409de 100644 --- a/paddle_samples/PaddleX/Mask-RT-DETR-S/subgraph_9/input_meta.py +++ b/paddle_samples/PaddleX/Mask-RT-DETR-S/subgraph_9/input_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_data_0: name = "data_0" + original_name = "param_60" shape = [2, 256] dtype = "float32" min_val = float("-2.77619") @@ -11,6 +12,7 @@ class Program_weight_tensor_data_0: class Program_weight_tensor_data_1: name = "data_1" + original_name = "param_61" shape = [256, 768] dtype = "float32" min_val = float("-0.443238") @@ -22,6 +24,7 @@ class Program_weight_tensor_data_1: class Program_weight_tensor_data_2: name = "data_2" + original_name = "param_62" shape = [768] dtype = "float32" min_val = float("-0.256503") @@ -33,6 +36,7 @@ class Program_weight_tensor_data_2: class Program_weight_tensor_data_3: name = "data_3" + original_name = "param_63" shape = [256, 768] dtype = "float32" min_val = float("-0.42865") @@ -44,6 +48,7 @@ class Program_weight_tensor_data_3: class Program_weight_tensor_data_4: name = "data_4" + original_name = "param_64" shape = [768] dtype = "float32" min_val = float("-0.203861") @@ -55,6 +60,7 @@ class Program_weight_tensor_data_4: class Program_weight_tensor_data_5: name = "data_5" + original_name = "param_65" shape = [256, 768] dtype = "float32" min_val = float("-0.457353") @@ -66,6 +72,7 @@ class Program_weight_tensor_data_5: class Program_weight_tensor_data_6: name = "data_6" + original_name = "param_66" shape = [768] dtype = "float32" min_val = float("-0.173776") @@ -77,6 +84,7 @@ class Program_weight_tensor_data_6: class Program_weight_tensor_data_7: name = "data_7" + original_name = "param_67" shape = [256, 768] dtype = "float32" min_val = float("-0.440406") @@ -88,6 +96,7 @@ class Program_weight_tensor_data_7: class Program_weight_tensor_data_8: name = "data_8" + original_name = "param_68" shape = [768] dtype = "float32" min_val = float("-0.132557") @@ -99,6 +108,7 @@ class Program_weight_tensor_data_8: class Program_weight_tensor_data_9: name = "data_9" + original_name = "param_69" shape = [256, 768] dtype = "float32" min_val = float("-0.439901") @@ -110,6 +120,7 @@ class Program_weight_tensor_data_9: class Program_weight_tensor_data_10: name = "data_10" + original_name = "param_70" shape = [768] dtype = "float32" min_val = float("-0.0943525") @@ -121,6 +132,7 @@ class Program_weight_tensor_data_10: class Program_weight_tensor_data_11: name = "data_11" + original_name = "param_71" shape = [256, 768] dtype = "float32" min_val = float("-0.409097") @@ -132,6 +144,7 @@ class Program_weight_tensor_data_11: class Program_weight_tensor_data_12: name = "data_12" + original_name = "param_72" shape = [768] dtype = "float32" min_val = float("-0.0834445") @@ -143,6 +156,7 @@ class Program_weight_tensor_data_12: class Program_weight_tensor_data_13: name = "data_13" + original_name = "args_13" shape = [] dtype = "int64" data = [160] @@ -150,6 +164,7 @@ class Program_weight_tensor_data_13: class Program_weight_tensor_data_14: name = "data_14" + original_name = "args_14" shape = [] dtype = "int64" data = [160] @@ -157,6 +172,7 @@ class Program_weight_tensor_data_14: class Program_weight_tensor_data_15: name = "data_15" + original_name = "var_16348" shape = [1, 256, 80, 80] dtype = "float32" min_val = float("-0.278465") @@ -168,6 +184,7 @@ class Program_weight_tensor_data_15: class Program_weight_tensor_data_16: name = "data_16" + original_name = "var_16349" shape = [1, 256, 40, 40] dtype = "float32" min_val = float("-0.278465") @@ -179,6 +196,7 @@ class Program_weight_tensor_data_16: class Program_weight_tensor_data_17: name = "data_17" + original_name = "var_16350" shape = [1, 256, 20, 20] dtype = "float32" min_val = float("-0.278465") @@ -190,6 +208,7 @@ class Program_weight_tensor_data_17: class Program_weight_tensor_data_18: name = "data_18" + original_name = "var_16351" shape = [1, 128, 160, 160] dtype = "float32" min_val = float("-31.9008") @@ -201,6 +220,7 @@ class Program_weight_tensor_data_18: class Program_weight_tensor_data_19: name = "data_19" + original_name = "var_16354" shape = [1, 1] dtype = "int32" data = [1] @@ -208,6 +228,7 @@ class Program_weight_tensor_data_19: class Program_weight_tensor_data_20: name = "data_20" + original_name = "var_16355" shape = [1, 4] dtype = "float32" data = [0.438086, 0.52875, 0.651032, 0.94] diff --git a/paddle_samples/PaddleX/Mask-RT-DETR-S/subgraph_9/weight_meta.py b/paddle_samples/PaddleX/Mask-RT-DETR-S/subgraph_9/weight_meta.py index 6876725f2..c2922ea2b 100644 --- a/paddle_samples/PaddleX/Mask-RT-DETR-S/subgraph_9/weight_meta.py +++ b/paddle_samples/PaddleX/Mask-RT-DETR-S/subgraph_9/weight_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_parameter_0: name = "parameter_0" + original_name = "layer_norm_4.b_0_deepcopy_143" shape = [256] dtype = "float32" min_val = float("-0.257077") @@ -11,6 +12,7 @@ class Program_weight_tensor_parameter_0: class Program_weight_tensor_parameter_1: name = "parameter_1" + original_name = "layer_norm_4.w_0_deepcopy_142" shape = [256] dtype = "float32" min_val = float("0.681366") @@ -22,6 +24,7 @@ class Program_weight_tensor_parameter_1: class Program_weight_tensor_parameter_2: name = "parameter_2" + original_name = "linear_9.b_0_deepcopy_141" shape = [256] dtype = "float32" min_val = float("-0.225948") @@ -33,6 +36,7 @@ class Program_weight_tensor_parameter_2: class Program_weight_tensor_parameter_3: name = "parameter_3" + original_name = "linear_9.w_0_deepcopy_140" shape = [1024, 256] dtype = "float32" min_val = float("-0.479769") @@ -44,6 +48,7 @@ class Program_weight_tensor_parameter_3: class Program_weight_tensor_parameter_4: name = "parameter_4" + original_name = "linear_8.b_0_deepcopy_139" shape = [1024] dtype = "float32" min_val = float("-0.175478") @@ -55,6 +60,7 @@ class Program_weight_tensor_parameter_4: class Program_weight_tensor_parameter_5: name = "parameter_5" + original_name = "linear_8.w_0_deepcopy_138" shape = [256, 1024] dtype = "float32" min_val = float("-0.394211") @@ -66,6 +72,7 @@ class Program_weight_tensor_parameter_5: class Program_weight_tensor_parameter_6: name = "parameter_6" + original_name = "layer_norm_3.b_0_deepcopy_137" shape = [256] dtype = "float32" min_val = float("-0.148705") @@ -77,6 +84,7 @@ class Program_weight_tensor_parameter_6: class Program_weight_tensor_parameter_7: name = "parameter_7" + original_name = "layer_norm_3.w_0_deepcopy_136" shape = [256] dtype = "float32" min_val = float("0.811608") @@ -88,6 +96,7 @@ class Program_weight_tensor_parameter_7: class Program_weight_tensor_parameter_8: name = "parameter_8" + original_name = "linear_7.b_0_deepcopy_135" shape = [256] dtype = "float32" min_val = float("-0.1688") @@ -99,6 +108,7 @@ class Program_weight_tensor_parameter_8: class Program_weight_tensor_parameter_9: name = "parameter_9" + original_name = "linear_7.w_0_deepcopy_134" shape = [256, 256] dtype = "float32" min_val = float("-0.296662") @@ -110,6 +120,7 @@ class Program_weight_tensor_parameter_9: class Program_weight_tensor_parameter_10: name = "parameter_10" + original_name = "linear_5.b_0_deepcopy_131" shape = [96] dtype = "float32" min_val = float("-0.102205") @@ -121,6 +132,7 @@ class Program_weight_tensor_parameter_10: class Program_weight_tensor_parameter_11: name = "parameter_11" + original_name = "linear_5.w_0_deepcopy_130" shape = [256, 96] dtype = "float32" min_val = float("-0.434342") @@ -132,6 +144,7 @@ class Program_weight_tensor_parameter_11: class Program_weight_tensor_parameter_12: name = "parameter_12" + original_name = "linear_4.b_0_deepcopy_129" shape = [192] dtype = "float32" min_val = float("-4.2661") @@ -143,6 +156,7 @@ class Program_weight_tensor_parameter_12: class Program_weight_tensor_parameter_13: name = "parameter_13" + original_name = "linear_4.w_0_deepcopy_128" shape = [256, 192] dtype = "float32" min_val = float("-0.534478") @@ -154,6 +168,7 @@ class Program_weight_tensor_parameter_13: class Program_weight_tensor_parameter_14: name = "parameter_14" + original_name = "linear_6.b_0_deepcopy_133" shape = [256] dtype = "float32" min_val = float("-0.0804182") @@ -165,6 +180,7 @@ class Program_weight_tensor_parameter_14: class Program_weight_tensor_parameter_15: name = "parameter_15" + original_name = "linear_6.w_0_deepcopy_132" shape = [256, 256] dtype = "float32" min_val = float("-0.274237") @@ -176,6 +192,7 @@ class Program_weight_tensor_parameter_15: class Program_weight_tensor_parameter_16: name = "parameter_16" + original_name = "layer_norm_2.b_0_deepcopy_127" shape = [256] dtype = "float32" min_val = float("-0.166629") @@ -187,6 +204,7 @@ class Program_weight_tensor_parameter_16: class Program_weight_tensor_parameter_17: name = "parameter_17" + original_name = "layer_norm_2.w_0_deepcopy_126" shape = [256] dtype = "float32" min_val = float("0.96747") @@ -198,6 +216,7 @@ class Program_weight_tensor_parameter_17: class Program_weight_tensor_parameter_18: name = "parameter_18" + original_name = "linear_3.b_0_deepcopy_125" shape = [256] dtype = "float32" min_val = float("-0.5991") @@ -209,6 +228,7 @@ class Program_weight_tensor_parameter_18: class Program_weight_tensor_parameter_19: name = "parameter_19" + original_name = "linear_3.w_0_deepcopy_124" shape = [256, 256] dtype = "float32" min_val = float("-0.506166") @@ -220,6 +240,7 @@ class Program_weight_tensor_parameter_19: class Program_weight_tensor_parameter_20: name = "parameter_20" + original_name = "layer_norm_4.b_0_deepcopy_121" shape = [256] dtype = "float32" min_val = float("-0.166846") @@ -231,6 +252,7 @@ class Program_weight_tensor_parameter_20: class Program_weight_tensor_parameter_21: name = "parameter_21" + original_name = "layer_norm_4.w_0_deepcopy_120" shape = [256] dtype = "float32" min_val = float("0.680942") @@ -242,6 +264,7 @@ class Program_weight_tensor_parameter_21: class Program_weight_tensor_parameter_22: name = "parameter_22" + original_name = "linear_9.b_0_deepcopy_119" shape = [256] dtype = "float32" min_val = float("-0.171694") @@ -253,6 +276,7 @@ class Program_weight_tensor_parameter_22: class Program_weight_tensor_parameter_23: name = "parameter_23" + original_name = "linear_9.w_0_deepcopy_118" shape = [1024, 256] dtype = "float32" min_val = float("-0.386215") @@ -264,6 +288,7 @@ class Program_weight_tensor_parameter_23: class Program_weight_tensor_parameter_24: name = "parameter_24" + original_name = "linear_8.b_0_deepcopy_117" shape = [1024] dtype = "float32" min_val = float("-0.17664") @@ -275,6 +300,7 @@ class Program_weight_tensor_parameter_24: class Program_weight_tensor_parameter_25: name = "parameter_25" + original_name = "linear_8.w_0_deepcopy_116" shape = [256, 1024] dtype = "float32" min_val = float("-0.372825") @@ -286,6 +312,7 @@ class Program_weight_tensor_parameter_25: class Program_weight_tensor_parameter_26: name = "parameter_26" + original_name = "layer_norm_3.b_0_deepcopy_115" shape = [256] dtype = "float32" min_val = float("-0.125284") @@ -297,6 +324,7 @@ class Program_weight_tensor_parameter_26: class Program_weight_tensor_parameter_27: name = "parameter_27" + original_name = "layer_norm_3.w_0_deepcopy_114" shape = [256] dtype = "float32" min_val = float("0.859133") @@ -308,6 +336,7 @@ class Program_weight_tensor_parameter_27: class Program_weight_tensor_parameter_28: name = "parameter_28" + original_name = "linear_7.b_0_deepcopy_113" shape = [256] dtype = "float32" min_val = float("-0.139522") @@ -319,6 +348,7 @@ class Program_weight_tensor_parameter_28: class Program_weight_tensor_parameter_29: name = "parameter_29" + original_name = "linear_7.w_0_deepcopy_112" shape = [256, 256] dtype = "float32" min_val = float("-0.315115") @@ -330,6 +360,7 @@ class Program_weight_tensor_parameter_29: class Program_weight_tensor_parameter_30: name = "parameter_30" + original_name = "linear_5.b_0_deepcopy_109" shape = [96] dtype = "float32" min_val = float("-0.065347") @@ -341,6 +372,7 @@ class Program_weight_tensor_parameter_30: class Program_weight_tensor_parameter_31: name = "parameter_31" + original_name = "linear_5.w_0_deepcopy_108" shape = [256, 96] dtype = "float32" min_val = float("-0.347224") @@ -352,6 +384,7 @@ class Program_weight_tensor_parameter_31: class Program_weight_tensor_parameter_32: name = "parameter_32" + original_name = "linear_4.b_0_deepcopy_107" shape = [192] dtype = "float32" min_val = float("-4.22649") @@ -363,6 +396,7 @@ class Program_weight_tensor_parameter_32: class Program_weight_tensor_parameter_33: name = "parameter_33" + original_name = "linear_4.w_0_deepcopy_106" shape = [256, 192] dtype = "float32" min_val = float("-0.483596") @@ -374,6 +408,7 @@ class Program_weight_tensor_parameter_33: class Program_weight_tensor_parameter_34: name = "parameter_34" + original_name = "linear_6.b_0_deepcopy_111" shape = [256] dtype = "float32" min_val = float("-0.176126") @@ -385,6 +420,7 @@ class Program_weight_tensor_parameter_34: class Program_weight_tensor_parameter_35: name = "parameter_35" + original_name = "linear_6.w_0_deepcopy_110" shape = [256, 256] dtype = "float32" min_val = float("-0.312742") @@ -396,6 +432,7 @@ class Program_weight_tensor_parameter_35: class Program_weight_tensor_parameter_36: name = "parameter_36" + original_name = "layer_norm_2.b_0_deepcopy_105" shape = [256] dtype = "float32" min_val = float("-0.173981") @@ -407,6 +444,7 @@ class Program_weight_tensor_parameter_36: class Program_weight_tensor_parameter_37: name = "parameter_37" + original_name = "layer_norm_2.w_0_deepcopy_104" shape = [256] dtype = "float32" min_val = float("0.836381") @@ -418,6 +456,7 @@ class Program_weight_tensor_parameter_37: class Program_weight_tensor_parameter_38: name = "parameter_38" + original_name = "linear_3.b_0_deepcopy_103" shape = [256] dtype = "float32" min_val = float("-0.232186") @@ -429,6 +468,7 @@ class Program_weight_tensor_parameter_38: class Program_weight_tensor_parameter_39: name = "parameter_39" + original_name = "linear_3.w_0_deepcopy_102" shape = [256, 256] dtype = "float32" min_val = float("-0.323358") @@ -440,6 +480,7 @@ class Program_weight_tensor_parameter_39: class Program_weight_tensor_parameter_40: name = "parameter_40" + original_name = "layer_norm_4.b_0_deepcopy_99" shape = [256] dtype = "float32" min_val = float("-0.119385") @@ -451,6 +492,7 @@ class Program_weight_tensor_parameter_40: class Program_weight_tensor_parameter_41: name = "parameter_41" + original_name = "layer_norm_4.w_0_deepcopy_98" shape = [256] dtype = "float32" min_val = float("0.703915") @@ -462,6 +504,7 @@ class Program_weight_tensor_parameter_41: class Program_weight_tensor_parameter_42: name = "parameter_42" + original_name = "linear_9.b_0_deepcopy_97" shape = [256] dtype = "float32" min_val = float("-0.112626") @@ -473,6 +516,7 @@ class Program_weight_tensor_parameter_42: class Program_weight_tensor_parameter_43: name = "parameter_43" + original_name = "linear_9.w_0_deepcopy_96" shape = [1024, 256] dtype = "float32" min_val = float("-0.429047") @@ -484,6 +528,7 @@ class Program_weight_tensor_parameter_43: class Program_weight_tensor_parameter_44: name = "parameter_44" + original_name = "linear_8.b_0_deepcopy_95" shape = [1024] dtype = "float32" min_val = float("-0.207784") @@ -495,6 +540,7 @@ class Program_weight_tensor_parameter_44: class Program_weight_tensor_parameter_45: name = "parameter_45" + original_name = "linear_8.w_0_deepcopy_94" shape = [256, 1024] dtype = "float32" min_val = float("-0.392619") @@ -506,6 +552,7 @@ class Program_weight_tensor_parameter_45: class Program_weight_tensor_parameter_46: name = "parameter_46" + original_name = "layer_norm_3.b_0_deepcopy_93" shape = [256] dtype = "float32" min_val = float("-0.140992") @@ -517,6 +564,7 @@ class Program_weight_tensor_parameter_46: class Program_weight_tensor_parameter_47: name = "parameter_47" + original_name = "layer_norm_3.w_0_deepcopy_92" shape = [256] dtype = "float32" min_val = float("0.904747") @@ -528,6 +576,7 @@ class Program_weight_tensor_parameter_47: class Program_weight_tensor_parameter_48: name = "parameter_48" + original_name = "linear_7.b_0_deepcopy_91" shape = [256] dtype = "float32" min_val = float("-0.134226") @@ -539,6 +588,7 @@ class Program_weight_tensor_parameter_48: class Program_weight_tensor_parameter_49: name = "parameter_49" + original_name = "linear_7.w_0_deepcopy_90" shape = [256, 256] dtype = "float32" min_val = float("-0.331404") @@ -550,6 +600,7 @@ class Program_weight_tensor_parameter_49: class Program_weight_tensor_parameter_50: name = "parameter_50" + original_name = "linear_5.b_0_deepcopy_87" shape = [96] dtype = "float32" min_val = float("-0.210059") @@ -561,6 +612,7 @@ class Program_weight_tensor_parameter_50: class Program_weight_tensor_parameter_51: name = "parameter_51" + original_name = "linear_5.w_0_deepcopy_86" shape = [256, 96] dtype = "float32" min_val = float("-0.393647") @@ -572,6 +624,7 @@ class Program_weight_tensor_parameter_51: class Program_weight_tensor_parameter_52: name = "parameter_52" + original_name = "linear_4.b_0_deepcopy_85" shape = [192] dtype = "float32" min_val = float("-4.25235") @@ -583,6 +636,7 @@ class Program_weight_tensor_parameter_52: class Program_weight_tensor_parameter_53: name = "parameter_53" + original_name = "linear_4.w_0_deepcopy_84" shape = [256, 192] dtype = "float32" min_val = float("-0.50817") @@ -594,6 +648,7 @@ class Program_weight_tensor_parameter_53: class Program_weight_tensor_parameter_54: name = "parameter_54" + original_name = "linear_6.b_0_deepcopy_89" shape = [256] dtype = "float32" min_val = float("-0.137511") @@ -605,6 +660,7 @@ class Program_weight_tensor_parameter_54: class Program_weight_tensor_parameter_55: name = "parameter_55" + original_name = "linear_6.w_0_deepcopy_88" shape = [256, 256] dtype = "float32" min_val = float("-0.35258") @@ -616,6 +672,7 @@ class Program_weight_tensor_parameter_55: class Program_weight_tensor_parameter_56: name = "parameter_56" + original_name = "layer_norm_2.b_0_deepcopy_83" shape = [256] dtype = "float32" min_val = float("-0.326547") @@ -627,6 +684,7 @@ class Program_weight_tensor_parameter_56: class Program_weight_tensor_parameter_57: name = "parameter_57" + original_name = "layer_norm_2.w_0_deepcopy_82" shape = [256] dtype = "float32" min_val = float("0.733154") @@ -638,6 +696,7 @@ class Program_weight_tensor_parameter_57: class Program_weight_tensor_parameter_58: name = "parameter_58" + original_name = "linear_3.b_0_deepcopy_81" shape = [256] dtype = "float32" min_val = float("-0.180855") @@ -649,6 +708,7 @@ class Program_weight_tensor_parameter_58: class Program_weight_tensor_parameter_59: name = "parameter_59" + original_name = "linear_3.w_0_deepcopy_80" shape = [256, 256] dtype = "float32" min_val = float("-0.264527") @@ -660,6 +720,7 @@ class Program_weight_tensor_parameter_59: class Program_weight_tensor_parameter_60: name = "parameter_60" + original_name = "layer_norm_4.b_0_deepcopy_77" shape = [256] dtype = "float32" min_val = float("-0.146117") @@ -671,6 +732,7 @@ class Program_weight_tensor_parameter_60: class Program_weight_tensor_parameter_61: name = "parameter_61" + original_name = "layer_norm_4.w_0_deepcopy_76" shape = [256] dtype = "float32" min_val = float("0.750467") @@ -682,6 +744,7 @@ class Program_weight_tensor_parameter_61: class Program_weight_tensor_parameter_62: name = "parameter_62" + original_name = "linear_9.b_0_deepcopy_75" shape = [256] dtype = "float32" min_val = float("-0.155623") @@ -693,6 +756,7 @@ class Program_weight_tensor_parameter_62: class Program_weight_tensor_parameter_63: name = "parameter_63" + original_name = "linear_9.w_0_deepcopy_74" shape = [1024, 256] dtype = "float32" min_val = float("-0.420884") @@ -704,6 +768,7 @@ class Program_weight_tensor_parameter_63: class Program_weight_tensor_parameter_64: name = "parameter_64" + original_name = "linear_8.b_0_deepcopy_73" shape = [1024] dtype = "float32" min_val = float("-0.199214") @@ -715,6 +780,7 @@ class Program_weight_tensor_parameter_64: class Program_weight_tensor_parameter_65: name = "parameter_65" + original_name = "linear_8.w_0_deepcopy_72" shape = [256, 1024] dtype = "float32" min_val = float("-0.378287") @@ -726,6 +792,7 @@ class Program_weight_tensor_parameter_65: class Program_weight_tensor_parameter_66: name = "parameter_66" + original_name = "layer_norm_3.b_0_deepcopy_71" shape = [256] dtype = "float32" min_val = float("-0.148889") @@ -737,6 +804,7 @@ class Program_weight_tensor_parameter_66: class Program_weight_tensor_parameter_67: name = "parameter_67" + original_name = "layer_norm_3.w_0_deepcopy_70" shape = [256] dtype = "float32" min_val = float("0.884987") @@ -748,6 +816,7 @@ class Program_weight_tensor_parameter_67: class Program_weight_tensor_parameter_68: name = "parameter_68" + original_name = "linear_7.b_0_deepcopy_69" shape = [256] dtype = "float32" min_val = float("-0.129623") @@ -759,6 +828,7 @@ class Program_weight_tensor_parameter_68: class Program_weight_tensor_parameter_69: name = "parameter_69" + original_name = "linear_7.w_0_deepcopy_68" shape = [256, 256] dtype = "float32" min_val = float("-0.375354") @@ -770,6 +840,7 @@ class Program_weight_tensor_parameter_69: class Program_weight_tensor_parameter_70: name = "parameter_70" + original_name = "linear_5.b_0_deepcopy_65" shape = [96] dtype = "float32" min_val = float("-0.176131") @@ -781,6 +852,7 @@ class Program_weight_tensor_parameter_70: class Program_weight_tensor_parameter_71: name = "parameter_71" + original_name = "linear_5.w_0_deepcopy_64" shape = [256, 96] dtype = "float32" min_val = float("-0.510437") @@ -792,6 +864,7 @@ class Program_weight_tensor_parameter_71: class Program_weight_tensor_parameter_72: name = "parameter_72" + original_name = "linear_4.b_0_deepcopy_63" shape = [192] dtype = "float32" min_val = float("-4.40857") @@ -803,6 +876,7 @@ class Program_weight_tensor_parameter_72: class Program_weight_tensor_parameter_73: name = "parameter_73" + original_name = "linear_4.w_0_deepcopy_62" shape = [256, 192] dtype = "float32" min_val = float("-0.538118") @@ -814,6 +888,7 @@ class Program_weight_tensor_parameter_73: class Program_weight_tensor_parameter_74: name = "parameter_74" + original_name = "linear_6.b_0_deepcopy_67" shape = [256] dtype = "float32" min_val = float("-0.164591") @@ -825,6 +900,7 @@ class Program_weight_tensor_parameter_74: class Program_weight_tensor_parameter_75: name = "parameter_75" + original_name = "linear_6.w_0_deepcopy_66" shape = [256, 256] dtype = "float32" min_val = float("-0.341205") @@ -836,6 +912,7 @@ class Program_weight_tensor_parameter_75: class Program_weight_tensor_parameter_76: name = "parameter_76" + original_name = "layer_norm_2.b_0_deepcopy_61" shape = [256] dtype = "float32" min_val = float("-0.222436") @@ -847,6 +924,7 @@ class Program_weight_tensor_parameter_76: class Program_weight_tensor_parameter_77: name = "parameter_77" + original_name = "layer_norm_2.w_0_deepcopy_60" shape = [256] dtype = "float32" min_val = float("0.456384") @@ -858,6 +936,7 @@ class Program_weight_tensor_parameter_77: class Program_weight_tensor_parameter_78: name = "parameter_78" + original_name = "linear_3.b_0_deepcopy_59" shape = [256] dtype = "float32" min_val = float("-0.191942") @@ -869,6 +948,7 @@ class Program_weight_tensor_parameter_78: class Program_weight_tensor_parameter_79: name = "parameter_79" + original_name = "linear_3.w_0_deepcopy_58" shape = [256, 256] dtype = "float32" min_val = float("-0.307978") @@ -880,6 +960,7 @@ class Program_weight_tensor_parameter_79: class Program_weight_tensor_parameter_80: name = "parameter_80" + original_name = "layer_norm_4.b_0_deepcopy_55" shape = [256] dtype = "float32" min_val = float("-0.363568") @@ -891,6 +972,7 @@ class Program_weight_tensor_parameter_80: class Program_weight_tensor_parameter_81: name = "parameter_81" + original_name = "layer_norm_4.w_0_deepcopy_54" shape = [256] dtype = "float32" min_val = float("0.78889") @@ -902,6 +984,7 @@ class Program_weight_tensor_parameter_81: class Program_weight_tensor_parameter_82: name = "parameter_82" + original_name = "linear_9.b_0_deepcopy_53" shape = [256] dtype = "float32" min_val = float("-0.283697") @@ -913,6 +996,7 @@ class Program_weight_tensor_parameter_82: class Program_weight_tensor_parameter_83: name = "parameter_83" + original_name = "linear_9.w_0_deepcopy_52" shape = [1024, 256] dtype = "float32" min_val = float("-0.564167") @@ -924,6 +1008,7 @@ class Program_weight_tensor_parameter_83: class Program_weight_tensor_parameter_84: name = "parameter_84" + original_name = "linear_8.b_0_deepcopy_51" shape = [1024] dtype = "float32" min_val = float("-0.37974") @@ -935,6 +1020,7 @@ class Program_weight_tensor_parameter_84: class Program_weight_tensor_parameter_85: name = "parameter_85" + original_name = "linear_8.w_0_deepcopy_50" shape = [256, 1024] dtype = "float32" min_val = float("-0.423296") @@ -946,6 +1032,7 @@ class Program_weight_tensor_parameter_85: class Program_weight_tensor_parameter_86: name = "parameter_86" + original_name = "layer_norm_3.b_0_deepcopy_49" shape = [256] dtype = "float32" min_val = float("-0.136425") @@ -957,6 +1044,7 @@ class Program_weight_tensor_parameter_86: class Program_weight_tensor_parameter_87: name = "parameter_87" + original_name = "layer_norm_3.w_0_deepcopy_48" shape = [256] dtype = "float32" min_val = float("0.859815") @@ -968,6 +1056,7 @@ class Program_weight_tensor_parameter_87: class Program_weight_tensor_parameter_88: name = "parameter_88" + original_name = "linear_7.b_0_deepcopy_47" shape = [256] dtype = "float32" min_val = float("-0.120992") @@ -979,6 +1068,7 @@ class Program_weight_tensor_parameter_88: class Program_weight_tensor_parameter_89: name = "parameter_89" + original_name = "linear_7.w_0_deepcopy_46" shape = [256, 256] dtype = "float32" min_val = float("-0.354922") @@ -990,6 +1080,7 @@ class Program_weight_tensor_parameter_89: class Program_weight_tensor_parameter_90: name = "parameter_90" + original_name = "linear_5.b_0_deepcopy_43" shape = [96] dtype = "float32" min_val = float("-0.112906") @@ -1001,6 +1092,7 @@ class Program_weight_tensor_parameter_90: class Program_weight_tensor_parameter_91: name = "parameter_91" + original_name = "linear_5.w_0_deepcopy_42" shape = [256, 96] dtype = "float32" min_val = float("-0.607715") @@ -1012,6 +1104,7 @@ class Program_weight_tensor_parameter_91: class Program_weight_tensor_parameter_92: name = "parameter_92" + original_name = "linear_4.b_0_deepcopy_41" shape = [192] dtype = "float32" min_val = float("-4.22325") @@ -1023,6 +1116,7 @@ class Program_weight_tensor_parameter_92: class Program_weight_tensor_parameter_93: name = "parameter_93" + original_name = "linear_4.w_0_deepcopy_40" shape = [256, 192] dtype = "float32" min_val = float("-0.621372") @@ -1034,6 +1128,7 @@ class Program_weight_tensor_parameter_93: class Program_weight_tensor_parameter_94: name = "parameter_94" + original_name = "linear_6.b_0_deepcopy_45" shape = [256] dtype = "float32" min_val = float("-0.208167") @@ -1045,6 +1140,7 @@ class Program_weight_tensor_parameter_94: class Program_weight_tensor_parameter_95: name = "parameter_95" + original_name = "linear_6.w_0_deepcopy_44" shape = [256, 256] dtype = "float32" min_val = float("-0.327044") @@ -1056,6 +1152,7 @@ class Program_weight_tensor_parameter_95: class Program_weight_tensor_parameter_96: name = "parameter_96" + original_name = "layer_norm_2.b_0_deepcopy_39" shape = [256] dtype = "float32" min_val = float("-0.298091") @@ -1067,6 +1164,7 @@ class Program_weight_tensor_parameter_96: class Program_weight_tensor_parameter_97: name = "parameter_97" + original_name = "layer_norm_2.w_0_deepcopy_38" shape = [256] dtype = "float32" min_val = float("0.633639") @@ -1078,6 +1176,7 @@ class Program_weight_tensor_parameter_97: class Program_weight_tensor_parameter_98: name = "parameter_98" + original_name = "linear_3.b_0_deepcopy_37" shape = [256] dtype = "float32" min_val = float("-0.269271") @@ -1089,6 +1188,7 @@ class Program_weight_tensor_parameter_98: class Program_weight_tensor_parameter_99: name = "parameter_99" + original_name = "linear_3.w_0_deepcopy_36" shape = [256, 256] dtype = "float32" min_val = float("-0.425386") @@ -1100,6 +1200,7 @@ class Program_weight_tensor_parameter_99: class Program_weight_tensor_parameter_100: name = "parameter_100" + original_name = "layer_norm_4.b_0_deepcopy_33" shape = [256] dtype = "float32" min_val = float("-0.193797") @@ -1111,6 +1212,7 @@ class Program_weight_tensor_parameter_100: class Program_weight_tensor_parameter_101: name = "parameter_101" + original_name = "layer_norm_4.w_0_deepcopy_32" shape = [256] dtype = "float32" min_val = float("0.715448") @@ -1122,6 +1224,7 @@ class Program_weight_tensor_parameter_101: class Program_weight_tensor_parameter_102: name = "parameter_102" + original_name = "linear_9.b_0_deepcopy_31" shape = [256] dtype = "float32" min_val = float("-0.332269") @@ -1133,6 +1236,7 @@ class Program_weight_tensor_parameter_102: class Program_weight_tensor_parameter_103: name = "parameter_103" + original_name = "linear_9.w_0_deepcopy_30" shape = [1024, 256] dtype = "float32" min_val = float("-0.645393") @@ -1144,6 +1248,7 @@ class Program_weight_tensor_parameter_103: class Program_weight_tensor_parameter_104: name = "parameter_104" + original_name = "linear_8.b_0_deepcopy_29" shape = [1024] dtype = "float32" min_val = float("-0.297501") @@ -1155,6 +1260,7 @@ class Program_weight_tensor_parameter_104: class Program_weight_tensor_parameter_105: name = "parameter_105" + original_name = "linear_8.w_0_deepcopy_28" shape = [256, 1024] dtype = "float32" min_val = float("-0.439843") @@ -1166,6 +1272,7 @@ class Program_weight_tensor_parameter_105: class Program_weight_tensor_parameter_106: name = "parameter_106" + original_name = "layer_norm_3.b_0_deepcopy_27" shape = [256] dtype = "float32" min_val = float("-0.22177") @@ -1177,6 +1284,7 @@ class Program_weight_tensor_parameter_106: class Program_weight_tensor_parameter_107: name = "parameter_107" + original_name = "layer_norm_3.w_0_deepcopy_26" shape = [256] dtype = "float32" min_val = float("0.870443") @@ -1188,6 +1296,7 @@ class Program_weight_tensor_parameter_107: class Program_weight_tensor_parameter_108: name = "parameter_108" + original_name = "linear_7.b_0_deepcopy_25" shape = [256] dtype = "float32" min_val = float("-0.102303") @@ -1199,6 +1308,7 @@ class Program_weight_tensor_parameter_108: class Program_weight_tensor_parameter_109: name = "parameter_109" + original_name = "linear_7.w_0_deepcopy_24" shape = [256, 256] dtype = "float32" min_val = float("-0.333307") @@ -1210,6 +1320,7 @@ class Program_weight_tensor_parameter_109: class Program_weight_tensor_parameter_110: name = "parameter_110" + original_name = "linear_5.b_0_deepcopy_21" shape = [96] dtype = "float32" min_val = float("-0.240894") @@ -1221,6 +1332,7 @@ class Program_weight_tensor_parameter_110: class Program_weight_tensor_parameter_111: name = "parameter_111" + original_name = "linear_5.w_0_deepcopy_20" shape = [256, 96] dtype = "float32" min_val = float("-0.477188") @@ -1232,6 +1344,7 @@ class Program_weight_tensor_parameter_111: class Program_weight_tensor_parameter_112: name = "parameter_112" + original_name = "linear_4.b_0_deepcopy_19" shape = [192] dtype = "float32" min_val = float("-4.19973") @@ -1243,6 +1356,7 @@ class Program_weight_tensor_parameter_112: class Program_weight_tensor_parameter_113: name = "parameter_113" + original_name = "linear_4.w_0_deepcopy_18" shape = [256, 192] dtype = "float32" min_val = float("-0.626619") @@ -1254,6 +1368,7 @@ class Program_weight_tensor_parameter_113: class Program_weight_tensor_parameter_114: name = "parameter_114" + original_name = "linear_6.b_0_deepcopy_23" shape = [256] dtype = "float32" min_val = float("-0.284102") @@ -1265,6 +1380,7 @@ class Program_weight_tensor_parameter_114: class Program_weight_tensor_parameter_115: name = "parameter_115" + original_name = "linear_6.w_0_deepcopy_22" shape = [256, 256] dtype = "float32" min_val = float("-0.319682") @@ -1276,6 +1392,7 @@ class Program_weight_tensor_parameter_115: class Program_weight_tensor_parameter_116: name = "parameter_116" + original_name = "layer_norm_2.b_0_deepcopy_17" shape = [256] dtype = "float32" min_val = float("-0.231613") @@ -1287,6 +1404,7 @@ class Program_weight_tensor_parameter_116: class Program_weight_tensor_parameter_117: name = "parameter_117" + original_name = "layer_norm_2.w_0_deepcopy_16" shape = [256] dtype = "float32" min_val = float("0.641499") @@ -1298,6 +1416,7 @@ class Program_weight_tensor_parameter_117: class Program_weight_tensor_parameter_118: name = "parameter_118" + original_name = "linear_3.b_0_deepcopy_15" shape = [256] dtype = "float32" min_val = float("-0.345921") @@ -1309,6 +1428,7 @@ class Program_weight_tensor_parameter_118: class Program_weight_tensor_parameter_119: name = "parameter_119" + original_name = "linear_3.w_0_deepcopy_14" shape = [256, 256] dtype = "float32" min_val = float("-0.397349") @@ -1320,6 +1440,7 @@ class Program_weight_tensor_parameter_119: class Program_weight_tensor_parameter_120: name = "parameter_120" + original_name = "linear_11.b_0" shape = [256] dtype = "float32" min_val = float("-0.1751") @@ -1331,6 +1452,7 @@ class Program_weight_tensor_parameter_120: class Program_weight_tensor_parameter_121: name = "parameter_121" + original_name = "linear_11.w_0" shape = [512, 256] dtype = "float32" min_val = float("-2.82062") @@ -1342,6 +1464,7 @@ class Program_weight_tensor_parameter_121: class Program_weight_tensor_parameter_122: name = "parameter_122" + original_name = "linear_10.b_0" shape = [512] dtype = "float32" min_val = float("-0.497467") @@ -1353,6 +1476,7 @@ class Program_weight_tensor_parameter_122: class Program_weight_tensor_parameter_123: name = "parameter_123" + original_name = "linear_10.w_0" shape = [4, 512] dtype = "float32" min_val = float("-4.98618") @@ -1364,6 +1488,7 @@ class Program_weight_tensor_parameter_123: class Program_weight_tensor_parameter_124: name = "parameter_124" + original_name = "linear_14.b_0" shape = [128] dtype = "float32" min_val = float("-0.121052") @@ -1375,6 +1500,7 @@ class Program_weight_tensor_parameter_124: class Program_weight_tensor_parameter_125: name = "parameter_125" + original_name = "linear_14.w_0" shape = [256, 128] dtype = "float32" min_val = float("-1.10794") @@ -1386,6 +1512,7 @@ class Program_weight_tensor_parameter_125: class Program_weight_tensor_parameter_126: name = "parameter_126" + original_name = "linear_13.b_0" shape = [256] dtype = "float32" min_val = float("-0.439998") @@ -1397,6 +1524,7 @@ class Program_weight_tensor_parameter_126: class Program_weight_tensor_parameter_127: name = "parameter_127" + original_name = "linear_13.w_0" shape = [256, 256] dtype = "float32" min_val = float("-1.79083") @@ -1408,6 +1536,7 @@ class Program_weight_tensor_parameter_127: class Program_weight_tensor_parameter_128: name = "parameter_128" + original_name = "linear_12.b_0" shape = [256] dtype = "float32" min_val = float("-0.591331") @@ -1419,6 +1548,7 @@ class Program_weight_tensor_parameter_128: class Program_weight_tensor_parameter_129: name = "parameter_129" + original_name = "linear_12.w_0" shape = [256, 256] dtype = "float32" min_val = float("-0.442693") @@ -1430,6 +1560,7 @@ class Program_weight_tensor_parameter_129: class Program_weight_tensor_parameter_130: name = "parameter_130" + original_name = "layer_norm_6.b_0" shape = [256] dtype = "float32" min_val = float("-1.03312") @@ -1441,6 +1572,7 @@ class Program_weight_tensor_parameter_130: class Program_weight_tensor_parameter_131: name = "parameter_131" + original_name = "layer_norm_6.w_0" shape = [256] dtype = "float32" min_val = float("0.467435") @@ -1452,6 +1584,7 @@ class Program_weight_tensor_parameter_131: class Program_weight_tensor_parameter_132: name = "parameter_132" + original_name = "linear_19.b_0" shape = [4] dtype = "float32" min_val = float("0") @@ -1461,6 +1594,7 @@ class Program_weight_tensor_parameter_132: class Program_weight_tensor_parameter_133: name = "parameter_133" + original_name = "linear_19.w_0" shape = [256, 4] dtype = "float32" min_val = float("-0.627713") @@ -1472,6 +1606,7 @@ class Program_weight_tensor_parameter_133: class Program_weight_tensor_parameter_134: name = "parameter_134" + original_name = "linear_18.b_0" shape = [256] dtype = "float32" min_val = float("-0.325887") @@ -1483,6 +1618,7 @@ class Program_weight_tensor_parameter_134: class Program_weight_tensor_parameter_135: name = "parameter_135" + original_name = "linear_18.w_0" shape = [256, 256] dtype = "float32" min_val = float("-4.22838") @@ -1494,6 +1630,7 @@ class Program_weight_tensor_parameter_135: class Program_weight_tensor_parameter_136: name = "parameter_136" + original_name = "linear_17.b_0" shape = [256] dtype = "float32" min_val = float("-0.425893") @@ -1505,6 +1642,7 @@ class Program_weight_tensor_parameter_136: class Program_weight_tensor_parameter_137: name = "parameter_137" + original_name = "linear_17.w_0" shape = [256, 256] dtype = "float32" min_val = float("-0.377984") @@ -1516,6 +1654,7 @@ class Program_weight_tensor_parameter_137: class Program_weight_tensor_parameter_138: name = "parameter_138" + original_name = "linear_16.b_0" shape = [2] dtype = "float32" min_val = float("0") @@ -1525,6 +1664,7 @@ class Program_weight_tensor_parameter_138: class Program_weight_tensor_parameter_139: name = "parameter_139" + original_name = "linear_16.w_0" shape = [256, 2] dtype = "float32" min_val = float("-0.0623768") @@ -1536,6 +1676,7 @@ class Program_weight_tensor_parameter_139: class Program_weight_tensor_parameter_140: name = "parameter_140" + original_name = "layer_norm_5.b_0" shape = [256] dtype = "float32" min_val = float("-0.959546") @@ -1547,6 +1688,7 @@ class Program_weight_tensor_parameter_140: class Program_weight_tensor_parameter_141: name = "parameter_141" + original_name = "layer_norm_5.w_0" shape = [256] dtype = "float32" min_val = float("0.729011") @@ -1558,6 +1700,7 @@ class Program_weight_tensor_parameter_141: class Program_weight_tensor_parameter_142: name = "parameter_142" + original_name = "linear_15.b_0" shape = [256] dtype = "float32" min_val = float("-0.842428") @@ -1569,6 +1712,7 @@ class Program_weight_tensor_parameter_142: class Program_weight_tensor_parameter_143: name = "parameter_143" + original_name = "linear_15.w_0" shape = [256, 256] dtype = "float32" min_val = float("-0.541244") @@ -1580,6 +1724,7 @@ class Program_weight_tensor_parameter_143: class Program_weight_tensor_parameter_144: name = "parameter_144" + original_name = "batch_norm2d_136.b_0" shape = [256] dtype = "float32" min_val = float("-0.313387") @@ -1591,6 +1736,7 @@ class Program_weight_tensor_parameter_144: class Program_weight_tensor_parameter_145: name = "parameter_145" + original_name = "batch_norm2d_136.w_0" shape = [256] dtype = "float32" min_val = float("0.784497") @@ -1602,6 +1748,7 @@ class Program_weight_tensor_parameter_145: class Program_weight_tensor_parameter_146: name = "parameter_146" + original_name = "batch_norm2d_136.w_2" shape = [256] dtype = "float32" min_val = float("1.77816") @@ -1613,6 +1760,7 @@ class Program_weight_tensor_parameter_146: class Program_weight_tensor_parameter_147: name = "parameter_147" + original_name = "batch_norm2d_136.w_1" shape = [256] dtype = "float32" min_val = float("-1.07592") @@ -1624,6 +1772,7 @@ class Program_weight_tensor_parameter_147: class Program_weight_tensor_parameter_148: name = "parameter_148" + original_name = "conv2d_95.w_0" shape = [256, 256, 1, 1] dtype = "float32" min_val = float("-0.344515") @@ -1635,6 +1784,7 @@ class Program_weight_tensor_parameter_148: class Program_weight_tensor_parameter_149: name = "parameter_149" + original_name = "batch_norm2d_135.b_0" shape = [256] dtype = "float32" min_val = float("-0.243712") @@ -1646,6 +1796,7 @@ class Program_weight_tensor_parameter_149: class Program_weight_tensor_parameter_150: name = "parameter_150" + original_name = "batch_norm2d_135.w_0" shape = [256] dtype = "float32" min_val = float("0.711525") @@ -1657,6 +1808,7 @@ class Program_weight_tensor_parameter_150: class Program_weight_tensor_parameter_151: name = "parameter_151" + original_name = "batch_norm2d_135.w_2" shape = [256] dtype = "float32" min_val = float("5.42446") @@ -1668,6 +1820,7 @@ class Program_weight_tensor_parameter_151: class Program_weight_tensor_parameter_152: name = "parameter_152" + original_name = "batch_norm2d_135.w_1" shape = [256] dtype = "float32" min_val = float("-0.664401") @@ -1679,6 +1832,7 @@ class Program_weight_tensor_parameter_152: class Program_weight_tensor_parameter_153: name = "parameter_153" + original_name = "conv2d_94.w_0" shape = [256, 256, 1, 1] dtype = "float32" min_val = float("-0.458922") @@ -1690,6 +1844,7 @@ class Program_weight_tensor_parameter_153: class Program_weight_tensor_parameter_154: name = "parameter_154" + original_name = "batch_norm2d_134.b_0" shape = [256] dtype = "float32" min_val = float("-0.415495") @@ -1701,6 +1856,7 @@ class Program_weight_tensor_parameter_154: class Program_weight_tensor_parameter_155: name = "parameter_155" + original_name = "batch_norm2d_134.w_0" shape = [256] dtype = "float32" min_val = float("0.696872") @@ -1712,6 +1868,7 @@ class Program_weight_tensor_parameter_155: class Program_weight_tensor_parameter_156: name = "parameter_156" + original_name = "batch_norm2d_134.w_2" shape = [256] dtype = "float32" min_val = float("2.75026") @@ -1723,6 +1880,7 @@ class Program_weight_tensor_parameter_156: class Program_weight_tensor_parameter_157: name = "parameter_157" + original_name = "batch_norm2d_134.w_1" shape = [256] dtype = "float32" min_val = float("-0.951364") @@ -1734,6 +1892,7 @@ class Program_weight_tensor_parameter_157: class Program_weight_tensor_parameter_158: name = "parameter_158" + original_name = "conv2d_93.w_0" shape = [256, 256, 1, 1] dtype = "float32" min_val = float("-0.42062") diff --git a/paddle_samples/PaddleX/MaskFormer_small/subgraph_2/input_meta.py b/paddle_samples/PaddleX/MaskFormer_small/subgraph_2/input_meta.py index 3cc6da05a..30359876f 100644 --- a/paddle_samples/PaddleX/MaskFormer_small/subgraph_2/input_meta.py +++ b/paddle_samples/PaddleX/MaskFormer_small/subgraph_2/input_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_data_0: name = "data_0" + original_name = "param_0" shape = [169, 3] dtype = "float32" min_val = float("-13.6244") @@ -11,6 +12,7 @@ class Program_weight_tensor_data_0: class Program_weight_tensor_data_1: name = "data_1" + original_name = "param_1" shape = [169, 3] dtype = "float32" min_val = float("-10.1725") @@ -22,6 +24,7 @@ class Program_weight_tensor_data_1: class Program_weight_tensor_data_2: name = "data_2" + original_name = "param_10" shape = [169, 12] dtype = "float32" min_val = float("-13.5494") @@ -33,6 +36,7 @@ class Program_weight_tensor_data_2: class Program_weight_tensor_data_3: name = "data_3" + original_name = "param_11" shape = [169, 12] dtype = "float32" min_val = float("-13.2241") @@ -44,6 +48,7 @@ class Program_weight_tensor_data_3: class Program_weight_tensor_data_4: name = "data_4" + original_name = "param_12" shape = [169, 12] dtype = "float32" min_val = float("-12.7894") @@ -55,6 +60,7 @@ class Program_weight_tensor_data_4: class Program_weight_tensor_data_5: name = "data_5" + original_name = "param_13" shape = [169, 12] dtype = "float32" min_val = float("-16.0148") @@ -66,6 +72,7 @@ class Program_weight_tensor_data_5: class Program_weight_tensor_data_6: name = "data_6" + original_name = "param_14" shape = [169, 12] dtype = "float32" min_val = float("-11.2492") @@ -77,6 +84,7 @@ class Program_weight_tensor_data_6: class Program_weight_tensor_data_7: name = "data_7" + original_name = "param_15" shape = [169, 12] dtype = "float32" min_val = float("-11.7707") @@ -88,6 +96,7 @@ class Program_weight_tensor_data_7: class Program_weight_tensor_data_8: name = "data_8" + original_name = "param_16" shape = [169, 12] dtype = "float32" min_val = float("-21.5965") @@ -99,6 +108,7 @@ class Program_weight_tensor_data_8: class Program_weight_tensor_data_9: name = "data_9" + original_name = "param_17" shape = [169, 12] dtype = "float32" min_val = float("-6.82914") @@ -110,6 +120,7 @@ class Program_weight_tensor_data_9: class Program_weight_tensor_data_10: name = "data_10" + original_name = "param_18" shape = [169, 12] dtype = "float32" min_val = float("-10.565") @@ -121,6 +132,7 @@ class Program_weight_tensor_data_10: class Program_weight_tensor_data_11: name = "data_11" + original_name = "param_19" shape = [169, 12] dtype = "float32" min_val = float("-5.64113") @@ -132,6 +144,7 @@ class Program_weight_tensor_data_11: class Program_weight_tensor_data_12: name = "data_12" + original_name = "param_2" shape = [169, 6] dtype = "float32" min_val = float("-10.4008") @@ -143,6 +156,7 @@ class Program_weight_tensor_data_12: class Program_weight_tensor_data_13: name = "data_13" + original_name = "param_20" shape = [169, 12] dtype = "float32" min_val = float("-15.6832") @@ -154,6 +168,7 @@ class Program_weight_tensor_data_13: class Program_weight_tensor_data_14: name = "data_14" + original_name = "param_21" shape = [169, 12] dtype = "float32" min_val = float("-21.99") @@ -165,6 +180,7 @@ class Program_weight_tensor_data_14: class Program_weight_tensor_data_15: name = "data_15" + original_name = "param_22" shape = [169, 24] dtype = "float32" min_val = float("-22.8913") @@ -176,6 +192,7 @@ class Program_weight_tensor_data_15: class Program_weight_tensor_data_16: name = "data_16" + original_name = "param_23" shape = [169, 24] dtype = "float32" min_val = float("-26.7541") @@ -187,6 +204,7 @@ class Program_weight_tensor_data_16: class Program_weight_tensor_data_17: name = "data_17" + original_name = "param_24" shape = [100, 256] dtype = "float32" min_val = float("-4.05563") @@ -198,6 +216,7 @@ class Program_weight_tensor_data_17: class Program_weight_tensor_data_18: name = "data_18" + original_name = "param_3" shape = [169, 6] dtype = "float32" min_val = float("-11.2985") @@ -209,6 +228,7 @@ class Program_weight_tensor_data_18: class Program_weight_tensor_data_19: name = "data_19" + original_name = "param_4" shape = [169, 12] dtype = "float32" min_val = float("-13.0402") @@ -220,6 +240,7 @@ class Program_weight_tensor_data_19: class Program_weight_tensor_data_20: name = "data_20" + original_name = "param_5" shape = [169, 12] dtype = "float32" min_val = float("-15.2582") @@ -231,6 +252,7 @@ class Program_weight_tensor_data_20: class Program_weight_tensor_data_21: name = "data_21" + original_name = "param_6" shape = [169, 12] dtype = "float32" min_val = float("-11.9454") @@ -242,6 +264,7 @@ class Program_weight_tensor_data_21: class Program_weight_tensor_data_22: name = "data_22" + original_name = "param_7" shape = [169, 12] dtype = "float32" min_val = float("-13.3033") @@ -253,6 +276,7 @@ class Program_weight_tensor_data_22: class Program_weight_tensor_data_23: name = "data_23" + original_name = "param_8" shape = [169, 12] dtype = "float32" min_val = float("-11.7905") @@ -264,6 +288,7 @@ class Program_weight_tensor_data_23: class Program_weight_tensor_data_24: name = "data_24" + original_name = "param_9" shape = [169, 12] dtype = "float32" min_val = float("-12.1116") @@ -275,6 +300,7 @@ class Program_weight_tensor_data_24: class Program_weight_tensor_data_25: name = "data_25" + original_name = "var_0" shape = [2, 3, 512, 512] dtype = "float32" min_val = float("-2.1179") @@ -286,6 +312,7 @@ class Program_weight_tensor_data_25: class Program_weight_tensor_data_26: name = "data_26" + original_name = "var_1054" shape = [49, 49] dtype = "int64" min_val = 0 @@ -295,6 +322,7 @@ class Program_weight_tensor_data_26: class Program_weight_tensor_data_27: name = "data_27" + original_name = "var_1120" shape = [49, 49] dtype = "int64" min_val = 0 @@ -304,6 +332,7 @@ class Program_weight_tensor_data_27: class Program_weight_tensor_data_28: name = "data_28" + original_name = "var_1181" shape = [49, 49] dtype = "int64" min_val = 0 @@ -313,6 +342,7 @@ class Program_weight_tensor_data_28: class Program_weight_tensor_data_29: name = "data_29" + original_name = "var_1247" shape = [49, 49] dtype = "int64" min_val = 0 @@ -322,6 +352,7 @@ class Program_weight_tensor_data_29: class Program_weight_tensor_data_30: name = "data_30" + original_name = "var_1308" shape = [49, 49] dtype = "int64" min_val = 0 @@ -331,6 +362,7 @@ class Program_weight_tensor_data_30: class Program_weight_tensor_data_31: name = "data_31" + original_name = "var_1374" shape = [49, 49] dtype = "int64" min_val = 0 @@ -340,6 +372,7 @@ class Program_weight_tensor_data_31: class Program_weight_tensor_data_32: name = "data_32" + original_name = "var_1435" shape = [49, 49] dtype = "int64" min_val = 0 @@ -349,6 +382,7 @@ class Program_weight_tensor_data_32: class Program_weight_tensor_data_33: name = "data_33" + original_name = "var_1535" shape = [49, 49] dtype = "int64" min_val = 0 @@ -358,6 +392,7 @@ class Program_weight_tensor_data_33: class Program_weight_tensor_data_34: name = "data_34" + original_name = "var_1596" shape = [49, 49] dtype = "int64" min_val = 0 @@ -367,6 +402,7 @@ class Program_weight_tensor_data_34: class Program_weight_tensor_data_35: name = "data_35" + original_name = "var_197" shape = [49, 49] dtype = "int64" min_val = 0 @@ -376,6 +412,7 @@ class Program_weight_tensor_data_35: class Program_weight_tensor_data_36: name = "data_36" + original_name = "var_258" shape = [49, 49] dtype = "int64" min_val = 0 @@ -385,6 +422,7 @@ class Program_weight_tensor_data_36: class Program_weight_tensor_data_37: name = "data_37" + original_name = "var_358" shape = [49, 49] dtype = "int64" min_val = 0 @@ -394,6 +432,7 @@ class Program_weight_tensor_data_37: class Program_weight_tensor_data_38: name = "data_38" + original_name = "var_419" shape = [49, 49] dtype = "int64" min_val = 0 @@ -403,6 +442,7 @@ class Program_weight_tensor_data_38: class Program_weight_tensor_data_39: name = "data_39" + original_name = "var_485" shape = [49, 49] dtype = "int64" min_val = 0 @@ -412,6 +452,7 @@ class Program_weight_tensor_data_39: class Program_weight_tensor_data_40: name = "data_40" + original_name = "var_50" shape = [49, 49] dtype = "int64" min_val = 0 @@ -421,6 +462,7 @@ class Program_weight_tensor_data_40: class Program_weight_tensor_data_41: name = "data_41" + original_name = "var_546" shape = [49, 49] dtype = "int64" min_val = 0 @@ -430,6 +472,7 @@ class Program_weight_tensor_data_41: class Program_weight_tensor_data_42: name = "data_42" + original_name = "var_612" shape = [49, 49] dtype = "int64" min_val = 0 @@ -439,6 +482,7 @@ class Program_weight_tensor_data_42: class Program_weight_tensor_data_43: name = "data_43" + original_name = "var_673" shape = [49, 49] dtype = "int64" min_val = 0 @@ -448,6 +492,7 @@ class Program_weight_tensor_data_43: class Program_weight_tensor_data_44: name = "data_44" + original_name = "var_739" shape = [49, 49] dtype = "int64" min_val = 0 @@ -457,6 +502,7 @@ class Program_weight_tensor_data_44: class Program_weight_tensor_data_45: name = "data_45" + original_name = "var_800" shape = [49, 49] dtype = "int64" min_val = 0 @@ -466,6 +512,7 @@ class Program_weight_tensor_data_45: class Program_weight_tensor_data_46: name = "data_46" + original_name = "var_866" shape = [49, 49] dtype = "int64" min_val = 0 @@ -475,6 +522,7 @@ class Program_weight_tensor_data_46: class Program_weight_tensor_data_47: name = "data_47" + original_name = "var_927" shape = [49, 49] dtype = "int64" min_val = 0 @@ -484,6 +532,7 @@ class Program_weight_tensor_data_47: class Program_weight_tensor_data_48: name = "data_48" + original_name = "var_97" shape = [49, 49] dtype = "int64" min_val = 0 @@ -493,6 +542,7 @@ class Program_weight_tensor_data_48: class Program_weight_tensor_data_49: name = "data_49" + original_name = "var_993" shape = [49, 49] dtype = "int64" min_val = 0 diff --git a/paddle_samples/PaddleX/MaskFormer_small/subgraph_2/weight_meta.py b/paddle_samples/PaddleX/MaskFormer_small/subgraph_2/weight_meta.py index 059a9b0bd..5404b3203 100644 --- a/paddle_samples/PaddleX/MaskFormer_small/subgraph_2/weight_meta.py +++ b/paddle_samples/PaddleX/MaskFormer_small/subgraph_2/weight_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_parameter_0: name = "parameter_0" + original_name = "linear_118.b_0" shape = [256] dtype = "float32" min_val = float("-0.0619398") @@ -11,6 +12,7 @@ class Program_weight_tensor_parameter_0: class Program_weight_tensor_parameter_1: name = "parameter_1" + original_name = "linear_118.w_0" shape = [256, 256] dtype = "float32" min_val = float("-0.0624999") @@ -22,6 +24,7 @@ class Program_weight_tensor_parameter_1: class Program_weight_tensor_parameter_2: name = "parameter_2" + original_name = "linear_117.b_0" shape = [256] dtype = "float32" min_val = float("-0.061877") @@ -33,6 +36,7 @@ class Program_weight_tensor_parameter_2: class Program_weight_tensor_parameter_3: name = "parameter_3" + original_name = "linear_117.w_0" shape = [256, 256] dtype = "float32" min_val = float("-0.062497") @@ -44,6 +48,7 @@ class Program_weight_tensor_parameter_3: class Program_weight_tensor_parameter_4: name = "parameter_4" + original_name = "linear_116.b_0" shape = [256] dtype = "float32" min_val = float("-0.0621395") @@ -55,6 +60,7 @@ class Program_weight_tensor_parameter_4: class Program_weight_tensor_parameter_5: name = "parameter_5" + original_name = "linear_116.w_0" shape = [256, 256] dtype = "float32" min_val = float("-0.0624948") @@ -66,6 +72,7 @@ class Program_weight_tensor_parameter_5: class Program_weight_tensor_parameter_6: name = "parameter_6" + original_name = "linear_115.b_0" shape = [3] dtype = "float32" min_val = float("0") @@ -75,6 +82,7 @@ class Program_weight_tensor_parameter_6: class Program_weight_tensor_parameter_7: name = "parameter_7" + original_name = "linear_115.w_0" shape = [256, 3] dtype = "float32" min_val = float("-0.0624409") @@ -86,6 +94,7 @@ class Program_weight_tensor_parameter_7: class Program_weight_tensor_parameter_8: name = "parameter_8" + original_name = "layer_norm_60.b_0_deepcopy_155" shape = [256] dtype = "float32" data = None @@ -93,6 +102,7 @@ class Program_weight_tensor_parameter_8: class Program_weight_tensor_parameter_9: name = "parameter_9" + original_name = "layer_norm_60.w_0_deepcopy_154" shape = [256] dtype = "float32" min_val = float("1.0") @@ -103,6 +113,7 @@ class Program_weight_tensor_parameter_9: class Program_weight_tensor_parameter_10: name = "parameter_10" + original_name = "linear_114.b_0_deepcopy_149" shape = [256] dtype = "float32" data = None @@ -110,6 +121,7 @@ class Program_weight_tensor_parameter_10: class Program_weight_tensor_parameter_11: name = "parameter_11" + original_name = "linear_114.w_0_deepcopy_148" shape = [2048, 256] dtype = "float32" min_val = float("-0.051031") @@ -121,6 +133,7 @@ class Program_weight_tensor_parameter_11: class Program_weight_tensor_parameter_12: name = "parameter_12" + original_name = "linear_113.b_0_deepcopy_147" shape = [2048] dtype = "float32" data = None @@ -128,6 +141,7 @@ class Program_weight_tensor_parameter_12: class Program_weight_tensor_parameter_13: name = "parameter_13" + original_name = "linear_113.w_0_deepcopy_146" shape = [256, 2048] dtype = "float32" min_val = float("-0.051031") @@ -139,6 +153,7 @@ class Program_weight_tensor_parameter_13: class Program_weight_tensor_parameter_14: name = "parameter_14" + original_name = "layer_norm_59.b_0_deepcopy_153" shape = [256] dtype = "float32" data = None @@ -146,6 +161,7 @@ class Program_weight_tensor_parameter_14: class Program_weight_tensor_parameter_15: name = "parameter_15" + original_name = "layer_norm_59.w_0_deepcopy_152" shape = [256] dtype = "float32" min_val = float("1.0") @@ -156,6 +172,7 @@ class Program_weight_tensor_parameter_15: class Program_weight_tensor_parameter_16: name = "parameter_16" + original_name = "linear_112.b_0_deepcopy_145" shape = [256] dtype = "float32" data = None @@ -163,6 +180,7 @@ class Program_weight_tensor_parameter_16: class Program_weight_tensor_parameter_17: name = "parameter_17" + original_name = "linear_112.w_0_deepcopy_144" shape = [256, 256] dtype = "float32" min_val = float("-0.10825") @@ -174,6 +192,7 @@ class Program_weight_tensor_parameter_17: class Program_weight_tensor_parameter_18: name = "parameter_18" + original_name = "linear_111.b_0_deepcopy_143" shape = [256] dtype = "float32" data = None @@ -181,6 +200,7 @@ class Program_weight_tensor_parameter_18: class Program_weight_tensor_parameter_19: name = "parameter_19" + original_name = "linear_111.w_0_deepcopy_142" shape = [256, 256] dtype = "float32" min_val = float("-0.108247") @@ -192,6 +212,7 @@ class Program_weight_tensor_parameter_19: class Program_weight_tensor_parameter_20: name = "parameter_20" + original_name = "linear_110.b_0_deepcopy_141" shape = [256] dtype = "float32" data = None @@ -199,6 +220,7 @@ class Program_weight_tensor_parameter_20: class Program_weight_tensor_parameter_21: name = "parameter_21" + original_name = "linear_110.w_0_deepcopy_140" shape = [256, 256] dtype = "float32" min_val = float("-0.108252") @@ -210,6 +232,7 @@ class Program_weight_tensor_parameter_21: class Program_weight_tensor_parameter_22: name = "parameter_22" + original_name = "linear_109.b_0_deepcopy_139" shape = [256] dtype = "float32" data = None @@ -217,6 +240,7 @@ class Program_weight_tensor_parameter_22: class Program_weight_tensor_parameter_23: name = "parameter_23" + original_name = "linear_109.w_0_deepcopy_138" shape = [256, 256] dtype = "float32" min_val = float("-0.108252") @@ -228,6 +252,7 @@ class Program_weight_tensor_parameter_23: class Program_weight_tensor_parameter_24: name = "parameter_24" + original_name = "layer_norm_58.b_0_deepcopy_151" shape = [256] dtype = "float32" data = None @@ -235,6 +260,7 @@ class Program_weight_tensor_parameter_24: class Program_weight_tensor_parameter_25: name = "parameter_25" + original_name = "layer_norm_58.w_0_deepcopy_150" shape = [256] dtype = "float32" min_val = float("1.0") @@ -245,6 +271,7 @@ class Program_weight_tensor_parameter_25: class Program_weight_tensor_parameter_26: name = "parameter_26" + original_name = "linear_108.b_0_deepcopy_137" shape = [256] dtype = "float32" data = None @@ -252,6 +279,7 @@ class Program_weight_tensor_parameter_26: class Program_weight_tensor_parameter_27: name = "parameter_27" + original_name = "linear_108.w_0_deepcopy_136" shape = [256, 256] dtype = "float32" min_val = float("-0.108252") @@ -263,6 +291,7 @@ class Program_weight_tensor_parameter_27: class Program_weight_tensor_parameter_28: name = "parameter_28" + original_name = "linear_107.b_0_deepcopy_135" shape = [256] dtype = "float32" data = None @@ -270,6 +299,7 @@ class Program_weight_tensor_parameter_28: class Program_weight_tensor_parameter_29: name = "parameter_29" + original_name = "linear_107.w_0_deepcopy_134" shape = [256, 256] dtype = "float32" min_val = float("-0.108251") @@ -281,6 +311,7 @@ class Program_weight_tensor_parameter_29: class Program_weight_tensor_parameter_30: name = "parameter_30" + original_name = "linear_106.b_0_deepcopy_133" shape = [256] dtype = "float32" data = None @@ -288,6 +319,7 @@ class Program_weight_tensor_parameter_30: class Program_weight_tensor_parameter_31: name = "parameter_31" + original_name = "linear_106.w_0_deepcopy_132" shape = [256, 256] dtype = "float32" min_val = float("-0.108252") @@ -299,6 +331,7 @@ class Program_weight_tensor_parameter_31: class Program_weight_tensor_parameter_32: name = "parameter_32" + original_name = "linear_105.b_0_deepcopy_131" shape = [256] dtype = "float32" data = None @@ -306,6 +339,7 @@ class Program_weight_tensor_parameter_32: class Program_weight_tensor_parameter_33: name = "parameter_33" + original_name = "linear_105.w_0_deepcopy_130" shape = [256, 256] dtype = "float32" min_val = float("-0.108251") @@ -317,6 +351,7 @@ class Program_weight_tensor_parameter_33: class Program_weight_tensor_parameter_34: name = "parameter_34" + original_name = "layer_norm_60.b_0_deepcopy_129" shape = [256] dtype = "float32" data = None @@ -324,6 +359,7 @@ class Program_weight_tensor_parameter_34: class Program_weight_tensor_parameter_35: name = "parameter_35" + original_name = "layer_norm_60.w_0_deepcopy_128" shape = [256] dtype = "float32" min_val = float("1.0") @@ -334,6 +370,7 @@ class Program_weight_tensor_parameter_35: class Program_weight_tensor_parameter_36: name = "parameter_36" + original_name = "linear_114.b_0_deepcopy_123" shape = [256] dtype = "float32" data = None @@ -341,6 +378,7 @@ class Program_weight_tensor_parameter_36: class Program_weight_tensor_parameter_37: name = "parameter_37" + original_name = "linear_114.w_0_deepcopy_122" shape = [2048, 256] dtype = "float32" min_val = float("-0.0510309") @@ -352,6 +390,7 @@ class Program_weight_tensor_parameter_37: class Program_weight_tensor_parameter_38: name = "parameter_38" + original_name = "linear_113.b_0_deepcopy_121" shape = [2048] dtype = "float32" data = None @@ -359,6 +398,7 @@ class Program_weight_tensor_parameter_38: class Program_weight_tensor_parameter_39: name = "parameter_39" + original_name = "linear_113.w_0_deepcopy_120" shape = [256, 2048] dtype = "float32" min_val = float("-0.0510308") @@ -370,6 +410,7 @@ class Program_weight_tensor_parameter_39: class Program_weight_tensor_parameter_40: name = "parameter_40" + original_name = "layer_norm_59.b_0_deepcopy_127" shape = [256] dtype = "float32" data = None @@ -377,6 +418,7 @@ class Program_weight_tensor_parameter_40: class Program_weight_tensor_parameter_41: name = "parameter_41" + original_name = "layer_norm_59.w_0_deepcopy_126" shape = [256] dtype = "float32" min_val = float("1.0") @@ -387,6 +429,7 @@ class Program_weight_tensor_parameter_41: class Program_weight_tensor_parameter_42: name = "parameter_42" + original_name = "linear_112.b_0_deepcopy_119" shape = [256] dtype = "float32" data = None @@ -394,6 +437,7 @@ class Program_weight_tensor_parameter_42: class Program_weight_tensor_parameter_43: name = "parameter_43" + original_name = "linear_112.w_0_deepcopy_118" shape = [256, 256] dtype = "float32" min_val = float("-0.10825") @@ -405,6 +449,7 @@ class Program_weight_tensor_parameter_43: class Program_weight_tensor_parameter_44: name = "parameter_44" + original_name = "linear_111.b_0_deepcopy_117" shape = [256] dtype = "float32" data = None @@ -412,6 +457,7 @@ class Program_weight_tensor_parameter_44: class Program_weight_tensor_parameter_45: name = "parameter_45" + original_name = "linear_111.w_0_deepcopy_116" shape = [256, 256] dtype = "float32" min_val = float("-0.108247") @@ -423,6 +469,7 @@ class Program_weight_tensor_parameter_45: class Program_weight_tensor_parameter_46: name = "parameter_46" + original_name = "linear_110.b_0_deepcopy_115" shape = [256] dtype = "float32" data = None @@ -430,6 +477,7 @@ class Program_weight_tensor_parameter_46: class Program_weight_tensor_parameter_47: name = "parameter_47" + original_name = "linear_110.w_0_deepcopy_114" shape = [256, 256] dtype = "float32" min_val = float("-0.108252") @@ -441,6 +489,7 @@ class Program_weight_tensor_parameter_47: class Program_weight_tensor_parameter_48: name = "parameter_48" + original_name = "linear_109.b_0_deepcopy_113" shape = [256] dtype = "float32" data = None @@ -448,6 +497,7 @@ class Program_weight_tensor_parameter_48: class Program_weight_tensor_parameter_49: name = "parameter_49" + original_name = "linear_109.w_0_deepcopy_112" shape = [256, 256] dtype = "float32" min_val = float("-0.108252") @@ -459,6 +509,7 @@ class Program_weight_tensor_parameter_49: class Program_weight_tensor_parameter_50: name = "parameter_50" + original_name = "layer_norm_58.b_0_deepcopy_125" shape = [256] dtype = "float32" data = None @@ -466,6 +517,7 @@ class Program_weight_tensor_parameter_50: class Program_weight_tensor_parameter_51: name = "parameter_51" + original_name = "layer_norm_58.w_0_deepcopy_124" shape = [256] dtype = "float32" min_val = float("1.0") @@ -476,6 +528,7 @@ class Program_weight_tensor_parameter_51: class Program_weight_tensor_parameter_52: name = "parameter_52" + original_name = "linear_108.b_0_deepcopy_111" shape = [256] dtype = "float32" data = None @@ -483,6 +536,7 @@ class Program_weight_tensor_parameter_52: class Program_weight_tensor_parameter_53: name = "parameter_53" + original_name = "linear_108.w_0_deepcopy_110" shape = [256, 256] dtype = "float32" min_val = float("-0.108252") @@ -494,6 +548,7 @@ class Program_weight_tensor_parameter_53: class Program_weight_tensor_parameter_54: name = "parameter_54" + original_name = "linear_107.b_0_deepcopy_109" shape = [256] dtype = "float32" data = None @@ -501,6 +556,7 @@ class Program_weight_tensor_parameter_54: class Program_weight_tensor_parameter_55: name = "parameter_55" + original_name = "linear_107.w_0_deepcopy_108" shape = [256, 256] dtype = "float32" min_val = float("-0.108251") @@ -512,6 +568,7 @@ class Program_weight_tensor_parameter_55: class Program_weight_tensor_parameter_56: name = "parameter_56" + original_name = "linear_106.b_0_deepcopy_107" shape = [256] dtype = "float32" data = None @@ -519,6 +576,7 @@ class Program_weight_tensor_parameter_56: class Program_weight_tensor_parameter_57: name = "parameter_57" + original_name = "linear_106.w_0_deepcopy_106" shape = [256, 256] dtype = "float32" min_val = float("-0.108252") @@ -530,6 +588,7 @@ class Program_weight_tensor_parameter_57: class Program_weight_tensor_parameter_58: name = "parameter_58" + original_name = "linear_105.b_0_deepcopy_105" shape = [256] dtype = "float32" data = None @@ -537,6 +596,7 @@ class Program_weight_tensor_parameter_58: class Program_weight_tensor_parameter_59: name = "parameter_59" + original_name = "linear_105.w_0_deepcopy_104" shape = [256, 256] dtype = "float32" min_val = float("-0.108251") @@ -548,6 +608,7 @@ class Program_weight_tensor_parameter_59: class Program_weight_tensor_parameter_60: name = "parameter_60" + original_name = "layer_norm_60.b_0_deepcopy_103" shape = [256] dtype = "float32" data = None @@ -555,6 +616,7 @@ class Program_weight_tensor_parameter_60: class Program_weight_tensor_parameter_61: name = "parameter_61" + original_name = "layer_norm_60.w_0_deepcopy_102" shape = [256] dtype = "float32" min_val = float("1.0") @@ -565,6 +627,7 @@ class Program_weight_tensor_parameter_61: class Program_weight_tensor_parameter_62: name = "parameter_62" + original_name = "linear_114.b_0_deepcopy_97" shape = [256] dtype = "float32" data = None @@ -572,6 +635,7 @@ class Program_weight_tensor_parameter_62: class Program_weight_tensor_parameter_63: name = "parameter_63" + original_name = "linear_114.w_0_deepcopy_96" shape = [2048, 256] dtype = "float32" min_val = float("-0.0510309") @@ -583,6 +647,7 @@ class Program_weight_tensor_parameter_63: class Program_weight_tensor_parameter_64: name = "parameter_64" + original_name = "linear_113.b_0_deepcopy_95" shape = [2048] dtype = "float32" data = None @@ -590,6 +655,7 @@ class Program_weight_tensor_parameter_64: class Program_weight_tensor_parameter_65: name = "parameter_65" + original_name = "linear_113.w_0_deepcopy_94" shape = [256, 2048] dtype = "float32" min_val = float("-0.0510309") @@ -601,6 +667,7 @@ class Program_weight_tensor_parameter_65: class Program_weight_tensor_parameter_66: name = "parameter_66" + original_name = "layer_norm_59.b_0_deepcopy_101" shape = [256] dtype = "float32" data = None @@ -608,6 +675,7 @@ class Program_weight_tensor_parameter_66: class Program_weight_tensor_parameter_67: name = "parameter_67" + original_name = "layer_norm_59.w_0_deepcopy_100" shape = [256] dtype = "float32" min_val = float("1.0") @@ -618,6 +686,7 @@ class Program_weight_tensor_parameter_67: class Program_weight_tensor_parameter_68: name = "parameter_68" + original_name = "linear_112.b_0_deepcopy_93" shape = [256] dtype = "float32" data = None @@ -625,6 +694,7 @@ class Program_weight_tensor_parameter_68: class Program_weight_tensor_parameter_69: name = "parameter_69" + original_name = "linear_112.w_0_deepcopy_92" shape = [256, 256] dtype = "float32" min_val = float("-0.10825") @@ -636,6 +706,7 @@ class Program_weight_tensor_parameter_69: class Program_weight_tensor_parameter_70: name = "parameter_70" + original_name = "linear_111.b_0_deepcopy_91" shape = [256] dtype = "float32" data = None @@ -643,6 +714,7 @@ class Program_weight_tensor_parameter_70: class Program_weight_tensor_parameter_71: name = "parameter_71" + original_name = "linear_111.w_0_deepcopy_90" shape = [256, 256] dtype = "float32" min_val = float("-0.108247") @@ -654,6 +726,7 @@ class Program_weight_tensor_parameter_71: class Program_weight_tensor_parameter_72: name = "parameter_72" + original_name = "linear_110.b_0_deepcopy_89" shape = [256] dtype = "float32" data = None @@ -661,6 +734,7 @@ class Program_weight_tensor_parameter_72: class Program_weight_tensor_parameter_73: name = "parameter_73" + original_name = "linear_110.w_0_deepcopy_88" shape = [256, 256] dtype = "float32" min_val = float("-0.108252") @@ -672,6 +746,7 @@ class Program_weight_tensor_parameter_73: class Program_weight_tensor_parameter_74: name = "parameter_74" + original_name = "linear_109.b_0_deepcopy_87" shape = [256] dtype = "float32" data = None @@ -679,6 +754,7 @@ class Program_weight_tensor_parameter_74: class Program_weight_tensor_parameter_75: name = "parameter_75" + original_name = "linear_109.w_0_deepcopy_86" shape = [256, 256] dtype = "float32" min_val = float("-0.108252") @@ -690,6 +766,7 @@ class Program_weight_tensor_parameter_75: class Program_weight_tensor_parameter_76: name = "parameter_76" + original_name = "layer_norm_58.b_0_deepcopy_99" shape = [256] dtype = "float32" data = None @@ -697,6 +774,7 @@ class Program_weight_tensor_parameter_76: class Program_weight_tensor_parameter_77: name = "parameter_77" + original_name = "layer_norm_58.w_0_deepcopy_98" shape = [256] dtype = "float32" min_val = float("1.0") @@ -707,6 +785,7 @@ class Program_weight_tensor_parameter_77: class Program_weight_tensor_parameter_78: name = "parameter_78" + original_name = "linear_108.b_0_deepcopy_85" shape = [256] dtype = "float32" data = None @@ -714,6 +793,7 @@ class Program_weight_tensor_parameter_78: class Program_weight_tensor_parameter_79: name = "parameter_79" + original_name = "linear_108.w_0_deepcopy_84" shape = [256, 256] dtype = "float32" min_val = float("-0.108252") @@ -725,6 +805,7 @@ class Program_weight_tensor_parameter_79: class Program_weight_tensor_parameter_80: name = "parameter_80" + original_name = "linear_107.b_0_deepcopy_83" shape = [256] dtype = "float32" data = None @@ -732,6 +813,7 @@ class Program_weight_tensor_parameter_80: class Program_weight_tensor_parameter_81: name = "parameter_81" + original_name = "linear_107.w_0_deepcopy_82" shape = [256, 256] dtype = "float32" min_val = float("-0.108251") @@ -743,6 +825,7 @@ class Program_weight_tensor_parameter_81: class Program_weight_tensor_parameter_82: name = "parameter_82" + original_name = "linear_106.b_0_deepcopy_81" shape = [256] dtype = "float32" data = None @@ -750,6 +833,7 @@ class Program_weight_tensor_parameter_82: class Program_weight_tensor_parameter_83: name = "parameter_83" + original_name = "linear_106.w_0_deepcopy_80" shape = [256, 256] dtype = "float32" min_val = float("-0.108252") @@ -761,6 +845,7 @@ class Program_weight_tensor_parameter_83: class Program_weight_tensor_parameter_84: name = "parameter_84" + original_name = "linear_105.b_0_deepcopy_79" shape = [256] dtype = "float32" data = None @@ -768,6 +853,7 @@ class Program_weight_tensor_parameter_84: class Program_weight_tensor_parameter_85: name = "parameter_85" + original_name = "linear_105.w_0_deepcopy_78" shape = [256, 256] dtype = "float32" min_val = float("-0.108251") @@ -779,6 +865,7 @@ class Program_weight_tensor_parameter_85: class Program_weight_tensor_parameter_86: name = "parameter_86" + original_name = "layer_norm_60.b_0_deepcopy_77" shape = [256] dtype = "float32" data = None @@ -786,6 +873,7 @@ class Program_weight_tensor_parameter_86: class Program_weight_tensor_parameter_87: name = "parameter_87" + original_name = "layer_norm_60.w_0_deepcopy_76" shape = [256] dtype = "float32" min_val = float("1.0") @@ -796,6 +884,7 @@ class Program_weight_tensor_parameter_87: class Program_weight_tensor_parameter_88: name = "parameter_88" + original_name = "linear_114.b_0_deepcopy_71" shape = [256] dtype = "float32" data = None @@ -803,6 +892,7 @@ class Program_weight_tensor_parameter_88: class Program_weight_tensor_parameter_89: name = "parameter_89" + original_name = "linear_114.w_0_deepcopy_70" shape = [2048, 256] dtype = "float32" min_val = float("-0.051031") @@ -814,6 +904,7 @@ class Program_weight_tensor_parameter_89: class Program_weight_tensor_parameter_90: name = "parameter_90" + original_name = "linear_113.b_0_deepcopy_69" shape = [2048] dtype = "float32" data = None @@ -821,6 +912,7 @@ class Program_weight_tensor_parameter_90: class Program_weight_tensor_parameter_91: name = "parameter_91" + original_name = "linear_113.w_0_deepcopy_68" shape = [256, 2048] dtype = "float32" min_val = float("-0.051031") @@ -832,6 +924,7 @@ class Program_weight_tensor_parameter_91: class Program_weight_tensor_parameter_92: name = "parameter_92" + original_name = "layer_norm_59.b_0_deepcopy_75" shape = [256] dtype = "float32" data = None @@ -839,6 +932,7 @@ class Program_weight_tensor_parameter_92: class Program_weight_tensor_parameter_93: name = "parameter_93" + original_name = "layer_norm_59.w_0_deepcopy_74" shape = [256] dtype = "float32" min_val = float("1.0") @@ -849,6 +943,7 @@ class Program_weight_tensor_parameter_93: class Program_weight_tensor_parameter_94: name = "parameter_94" + original_name = "linear_112.b_0_deepcopy_67" shape = [256] dtype = "float32" data = None @@ -856,6 +951,7 @@ class Program_weight_tensor_parameter_94: class Program_weight_tensor_parameter_95: name = "parameter_95" + original_name = "linear_112.w_0_deepcopy_66" shape = [256, 256] dtype = "float32" min_val = float("-0.10825") @@ -867,6 +963,7 @@ class Program_weight_tensor_parameter_95: class Program_weight_tensor_parameter_96: name = "parameter_96" + original_name = "linear_111.b_0_deepcopy_65" shape = [256] dtype = "float32" data = None @@ -874,6 +971,7 @@ class Program_weight_tensor_parameter_96: class Program_weight_tensor_parameter_97: name = "parameter_97" + original_name = "linear_111.w_0_deepcopy_64" shape = [256, 256] dtype = "float32" min_val = float("-0.108247") @@ -885,6 +983,7 @@ class Program_weight_tensor_parameter_97: class Program_weight_tensor_parameter_98: name = "parameter_98" + original_name = "linear_110.b_0_deepcopy_63" shape = [256] dtype = "float32" data = None @@ -892,6 +991,7 @@ class Program_weight_tensor_parameter_98: class Program_weight_tensor_parameter_99: name = "parameter_99" + original_name = "linear_110.w_0_deepcopy_62" shape = [256, 256] dtype = "float32" min_val = float("-0.108252") @@ -903,6 +1003,7 @@ class Program_weight_tensor_parameter_99: class Program_weight_tensor_parameter_100: name = "parameter_100" + original_name = "linear_109.b_0_deepcopy_61" shape = [256] dtype = "float32" data = None @@ -910,6 +1011,7 @@ class Program_weight_tensor_parameter_100: class Program_weight_tensor_parameter_101: name = "parameter_101" + original_name = "linear_109.w_0_deepcopy_60" shape = [256, 256] dtype = "float32" min_val = float("-0.108252") @@ -921,6 +1023,7 @@ class Program_weight_tensor_parameter_101: class Program_weight_tensor_parameter_102: name = "parameter_102" + original_name = "layer_norm_58.b_0_deepcopy_73" shape = [256] dtype = "float32" data = None @@ -928,6 +1031,7 @@ class Program_weight_tensor_parameter_102: class Program_weight_tensor_parameter_103: name = "parameter_103" + original_name = "layer_norm_58.w_0_deepcopy_72" shape = [256] dtype = "float32" min_val = float("1.0") @@ -938,6 +1042,7 @@ class Program_weight_tensor_parameter_103: class Program_weight_tensor_parameter_104: name = "parameter_104" + original_name = "linear_108.b_0_deepcopy_59" shape = [256] dtype = "float32" data = None @@ -945,6 +1050,7 @@ class Program_weight_tensor_parameter_104: class Program_weight_tensor_parameter_105: name = "parameter_105" + original_name = "linear_108.w_0_deepcopy_58" shape = [256, 256] dtype = "float32" min_val = float("-0.108252") @@ -956,6 +1062,7 @@ class Program_weight_tensor_parameter_105: class Program_weight_tensor_parameter_106: name = "parameter_106" + original_name = "linear_107.b_0_deepcopy_57" shape = [256] dtype = "float32" data = None @@ -963,6 +1070,7 @@ class Program_weight_tensor_parameter_106: class Program_weight_tensor_parameter_107: name = "parameter_107" + original_name = "linear_107.w_0_deepcopy_56" shape = [256, 256] dtype = "float32" min_val = float("-0.108251") @@ -974,6 +1082,7 @@ class Program_weight_tensor_parameter_107: class Program_weight_tensor_parameter_108: name = "parameter_108" + original_name = "linear_106.b_0_deepcopy_55" shape = [256] dtype = "float32" data = None @@ -981,6 +1090,7 @@ class Program_weight_tensor_parameter_108: class Program_weight_tensor_parameter_109: name = "parameter_109" + original_name = "linear_106.w_0_deepcopy_54" shape = [256, 256] dtype = "float32" min_val = float("-0.108252") @@ -992,6 +1102,7 @@ class Program_weight_tensor_parameter_109: class Program_weight_tensor_parameter_110: name = "parameter_110" + original_name = "linear_105.b_0_deepcopy_53" shape = [256] dtype = "float32" data = None @@ -999,6 +1110,7 @@ class Program_weight_tensor_parameter_110: class Program_weight_tensor_parameter_111: name = "parameter_111" + original_name = "linear_105.w_0_deepcopy_52" shape = [256, 256] dtype = "float32" min_val = float("-0.108251") @@ -1010,6 +1122,7 @@ class Program_weight_tensor_parameter_111: class Program_weight_tensor_parameter_112: name = "parameter_112" + original_name = "layer_norm_60.b_0_deepcopy_51" shape = [256] dtype = "float32" data = None @@ -1017,6 +1130,7 @@ class Program_weight_tensor_parameter_112: class Program_weight_tensor_parameter_113: name = "parameter_113" + original_name = "layer_norm_60.w_0_deepcopy_50" shape = [256] dtype = "float32" min_val = float("1.0") @@ -1027,6 +1141,7 @@ class Program_weight_tensor_parameter_113: class Program_weight_tensor_parameter_114: name = "parameter_114" + original_name = "linear_114.b_0_deepcopy_45" shape = [256] dtype = "float32" data = None @@ -1034,6 +1149,7 @@ class Program_weight_tensor_parameter_114: class Program_weight_tensor_parameter_115: name = "parameter_115" + original_name = "linear_114.w_0_deepcopy_44" shape = [2048, 256] dtype = "float32" min_val = float("-0.0510307") @@ -1045,6 +1161,7 @@ class Program_weight_tensor_parameter_115: class Program_weight_tensor_parameter_116: name = "parameter_116" + original_name = "linear_113.b_0_deepcopy_43" shape = [2048] dtype = "float32" data = None @@ -1052,6 +1169,7 @@ class Program_weight_tensor_parameter_116: class Program_weight_tensor_parameter_117: name = "parameter_117" + original_name = "linear_113.w_0_deepcopy_42" shape = [256, 2048] dtype = "float32" min_val = float("-0.051031") @@ -1063,6 +1181,7 @@ class Program_weight_tensor_parameter_117: class Program_weight_tensor_parameter_118: name = "parameter_118" + original_name = "layer_norm_59.b_0_deepcopy_49" shape = [256] dtype = "float32" data = None @@ -1070,6 +1189,7 @@ class Program_weight_tensor_parameter_118: class Program_weight_tensor_parameter_119: name = "parameter_119" + original_name = "layer_norm_59.w_0_deepcopy_48" shape = [256] dtype = "float32" min_val = float("1.0") @@ -1080,6 +1200,7 @@ class Program_weight_tensor_parameter_119: class Program_weight_tensor_parameter_120: name = "parameter_120" + original_name = "linear_112.b_0_deepcopy_41" shape = [256] dtype = "float32" data = None @@ -1087,6 +1208,7 @@ class Program_weight_tensor_parameter_120: class Program_weight_tensor_parameter_121: name = "parameter_121" + original_name = "linear_112.w_0_deepcopy_40" shape = [256, 256] dtype = "float32" min_val = float("-0.10825") @@ -1098,6 +1220,7 @@ class Program_weight_tensor_parameter_121: class Program_weight_tensor_parameter_122: name = "parameter_122" + original_name = "linear_111.b_0_deepcopy_39" shape = [256] dtype = "float32" data = None @@ -1105,6 +1228,7 @@ class Program_weight_tensor_parameter_122: class Program_weight_tensor_parameter_123: name = "parameter_123" + original_name = "linear_111.w_0_deepcopy_38" shape = [256, 256] dtype = "float32" min_val = float("-0.108247") @@ -1116,6 +1240,7 @@ class Program_weight_tensor_parameter_123: class Program_weight_tensor_parameter_124: name = "parameter_124" + original_name = "linear_110.b_0_deepcopy_37" shape = [256] dtype = "float32" data = None @@ -1123,6 +1248,7 @@ class Program_weight_tensor_parameter_124: class Program_weight_tensor_parameter_125: name = "parameter_125" + original_name = "linear_110.w_0_deepcopy_36" shape = [256, 256] dtype = "float32" min_val = float("-0.108252") @@ -1134,6 +1260,7 @@ class Program_weight_tensor_parameter_125: class Program_weight_tensor_parameter_126: name = "parameter_126" + original_name = "linear_109.b_0_deepcopy_35" shape = [256] dtype = "float32" data = None @@ -1141,6 +1268,7 @@ class Program_weight_tensor_parameter_126: class Program_weight_tensor_parameter_127: name = "parameter_127" + original_name = "linear_109.w_0_deepcopy_34" shape = [256, 256] dtype = "float32" min_val = float("-0.108252") @@ -1152,6 +1280,7 @@ class Program_weight_tensor_parameter_127: class Program_weight_tensor_parameter_128: name = "parameter_128" + original_name = "layer_norm_58.b_0_deepcopy_47" shape = [256] dtype = "float32" data = None @@ -1159,6 +1288,7 @@ class Program_weight_tensor_parameter_128: class Program_weight_tensor_parameter_129: name = "parameter_129" + original_name = "layer_norm_58.w_0_deepcopy_46" shape = [256] dtype = "float32" min_val = float("1.0") @@ -1169,6 +1299,7 @@ class Program_weight_tensor_parameter_129: class Program_weight_tensor_parameter_130: name = "parameter_130" + original_name = "linear_108.b_0_deepcopy_33" shape = [256] dtype = "float32" data = None @@ -1176,6 +1307,7 @@ class Program_weight_tensor_parameter_130: class Program_weight_tensor_parameter_131: name = "parameter_131" + original_name = "linear_108.w_0_deepcopy_32" shape = [256, 256] dtype = "float32" min_val = float("-0.108252") @@ -1187,6 +1319,7 @@ class Program_weight_tensor_parameter_131: class Program_weight_tensor_parameter_132: name = "parameter_132" + original_name = "linear_107.b_0_deepcopy_31" shape = [256] dtype = "float32" data = None @@ -1194,6 +1327,7 @@ class Program_weight_tensor_parameter_132: class Program_weight_tensor_parameter_133: name = "parameter_133" + original_name = "linear_107.w_0_deepcopy_30" shape = [256, 256] dtype = "float32" min_val = float("-0.108251") @@ -1205,6 +1339,7 @@ class Program_weight_tensor_parameter_133: class Program_weight_tensor_parameter_134: name = "parameter_134" + original_name = "linear_106.b_0_deepcopy_29" shape = [256] dtype = "float32" data = None @@ -1212,6 +1347,7 @@ class Program_weight_tensor_parameter_134: class Program_weight_tensor_parameter_135: name = "parameter_135" + original_name = "linear_106.w_0_deepcopy_28" shape = [256, 256] dtype = "float32" min_val = float("-0.108252") @@ -1223,6 +1359,7 @@ class Program_weight_tensor_parameter_135: class Program_weight_tensor_parameter_136: name = "parameter_136" + original_name = "linear_105.b_0_deepcopy_27" shape = [256] dtype = "float32" data = None @@ -1230,6 +1367,7 @@ class Program_weight_tensor_parameter_136: class Program_weight_tensor_parameter_137: name = "parameter_137" + original_name = "linear_105.w_0_deepcopy_26" shape = [256, 256] dtype = "float32" min_val = float("-0.108251") @@ -1241,6 +1379,7 @@ class Program_weight_tensor_parameter_137: class Program_weight_tensor_parameter_138: name = "parameter_138" + original_name = "layer_norm_61.b_0" shape = [256] dtype = "float32" data = None @@ -1248,6 +1387,7 @@ class Program_weight_tensor_parameter_138: class Program_weight_tensor_parameter_139: name = "parameter_139" + original_name = "layer_norm_61.w_0" shape = [256] dtype = "float32" min_val = float("1.0") @@ -1258,6 +1398,7 @@ class Program_weight_tensor_parameter_139: class Program_weight_tensor_parameter_140: name = "parameter_140" + original_name = "layer_norm_60.b_0_deepcopy_25" shape = [256] dtype = "float32" data = None @@ -1265,6 +1406,7 @@ class Program_weight_tensor_parameter_140: class Program_weight_tensor_parameter_141: name = "parameter_141" + original_name = "layer_norm_60.w_0_deepcopy_24" shape = [256] dtype = "float32" min_val = float("1.0") @@ -1275,6 +1417,7 @@ class Program_weight_tensor_parameter_141: class Program_weight_tensor_parameter_142: name = "parameter_142" + original_name = "linear_114.b_0_deepcopy_19" shape = [256] dtype = "float32" data = None @@ -1282,6 +1425,7 @@ class Program_weight_tensor_parameter_142: class Program_weight_tensor_parameter_143: name = "parameter_143" + original_name = "linear_114.w_0_deepcopy_18" shape = [2048, 256] dtype = "float32" min_val = float("-0.051031") @@ -1293,6 +1437,7 @@ class Program_weight_tensor_parameter_143: class Program_weight_tensor_parameter_144: name = "parameter_144" + original_name = "linear_113.b_0_deepcopy_17" shape = [2048] dtype = "float32" data = None @@ -1300,6 +1445,7 @@ class Program_weight_tensor_parameter_144: class Program_weight_tensor_parameter_145: name = "parameter_145" + original_name = "linear_113.w_0_deepcopy_16" shape = [256, 2048] dtype = "float32" min_val = float("-0.0510309") @@ -1311,6 +1457,7 @@ class Program_weight_tensor_parameter_145: class Program_weight_tensor_parameter_146: name = "parameter_146" + original_name = "layer_norm_59.b_0_deepcopy_23" shape = [256] dtype = "float32" data = None @@ -1318,6 +1465,7 @@ class Program_weight_tensor_parameter_146: class Program_weight_tensor_parameter_147: name = "parameter_147" + original_name = "layer_norm_59.w_0_deepcopy_22" shape = [256] dtype = "float32" min_val = float("1.0") @@ -1328,6 +1476,7 @@ class Program_weight_tensor_parameter_147: class Program_weight_tensor_parameter_148: name = "parameter_148" + original_name = "linear_112.b_0_deepcopy_15" shape = [256] dtype = "float32" data = None @@ -1335,6 +1484,7 @@ class Program_weight_tensor_parameter_148: class Program_weight_tensor_parameter_149: name = "parameter_149" + original_name = "linear_112.w_0_deepcopy_14" shape = [256, 256] dtype = "float32" min_val = float("-0.10825") @@ -1346,6 +1496,7 @@ class Program_weight_tensor_parameter_149: class Program_weight_tensor_parameter_150: name = "parameter_150" + original_name = "linear_111.b_0_deepcopy_13" shape = [256] dtype = "float32" data = None @@ -1353,6 +1504,7 @@ class Program_weight_tensor_parameter_150: class Program_weight_tensor_parameter_151: name = "parameter_151" + original_name = "linear_111.w_0_deepcopy_12" shape = [256, 256] dtype = "float32" min_val = float("-0.108247") @@ -1364,6 +1516,7 @@ class Program_weight_tensor_parameter_151: class Program_weight_tensor_parameter_152: name = "parameter_152" + original_name = "linear_110.b_0_deepcopy_11" shape = [256] dtype = "float32" data = None @@ -1371,6 +1524,7 @@ class Program_weight_tensor_parameter_152: class Program_weight_tensor_parameter_153: name = "parameter_153" + original_name = "linear_110.w_0_deepcopy_10" shape = [256, 256] dtype = "float32" min_val = float("-0.108252") @@ -1382,6 +1536,7 @@ class Program_weight_tensor_parameter_153: class Program_weight_tensor_parameter_154: name = "parameter_154" + original_name = "linear_109.b_0_deepcopy_9" shape = [256] dtype = "float32" data = None @@ -1389,6 +1544,7 @@ class Program_weight_tensor_parameter_154: class Program_weight_tensor_parameter_155: name = "parameter_155" + original_name = "linear_109.w_0_deepcopy_8" shape = [256, 256] dtype = "float32" min_val = float("-0.108252") @@ -1400,6 +1556,7 @@ class Program_weight_tensor_parameter_155: class Program_weight_tensor_parameter_156: name = "parameter_156" + original_name = "layer_norm_58.b_0_deepcopy_21" shape = [256] dtype = "float32" data = None @@ -1407,6 +1564,7 @@ class Program_weight_tensor_parameter_156: class Program_weight_tensor_parameter_157: name = "parameter_157" + original_name = "layer_norm_58.w_0_deepcopy_20" shape = [256] dtype = "float32" min_val = float("1.0") @@ -1417,6 +1575,7 @@ class Program_weight_tensor_parameter_157: class Program_weight_tensor_parameter_158: name = "parameter_158" + original_name = "linear_108.b_0_deepcopy_7" shape = [256] dtype = "float32" data = None @@ -1424,6 +1583,7 @@ class Program_weight_tensor_parameter_158: class Program_weight_tensor_parameter_159: name = "parameter_159" + original_name = "linear_108.w_0_deepcopy_6" shape = [256, 256] dtype = "float32" min_val = float("-0.108252") @@ -1435,6 +1595,7 @@ class Program_weight_tensor_parameter_159: class Program_weight_tensor_parameter_160: name = "parameter_160" + original_name = "linear_107.b_0_deepcopy_5" shape = [256] dtype = "float32" data = None @@ -1442,6 +1603,7 @@ class Program_weight_tensor_parameter_160: class Program_weight_tensor_parameter_161: name = "parameter_161" + original_name = "linear_107.w_0_deepcopy_4" shape = [256, 256] dtype = "float32" min_val = float("-0.108251") @@ -1453,6 +1615,7 @@ class Program_weight_tensor_parameter_161: class Program_weight_tensor_parameter_162: name = "parameter_162" + original_name = "linear_106.b_0_deepcopy_3" shape = [256] dtype = "float32" data = None @@ -1460,6 +1623,7 @@ class Program_weight_tensor_parameter_162: class Program_weight_tensor_parameter_163: name = "parameter_163" + original_name = "linear_106.w_0_deepcopy_2" shape = [256, 256] dtype = "float32" min_val = float("-0.108252") @@ -1471,6 +1635,7 @@ class Program_weight_tensor_parameter_163: class Program_weight_tensor_parameter_164: name = "parameter_164" + original_name = "linear_105.b_0_deepcopy_1" shape = [256] dtype = "float32" data = None @@ -1478,6 +1643,7 @@ class Program_weight_tensor_parameter_164: class Program_weight_tensor_parameter_165: name = "parameter_165" + original_name = "linear_105.w_0_deepcopy_0" shape = [256, 256] dtype = "float32" min_val = float("-0.108251") @@ -1489,6 +1655,7 @@ class Program_weight_tensor_parameter_165: class Program_weight_tensor_parameter_166: name = "parameter_166" + original_name = "conv2d_9.b_0" shape = [256] dtype = "float32" data = None @@ -1496,6 +1663,7 @@ class Program_weight_tensor_parameter_166: class Program_weight_tensor_parameter_167: name = "parameter_167" + original_name = "conv2d_9.w_0" shape = [256, 768, 1, 1] dtype = "float32" min_val = float("-0.0624983") @@ -1507,6 +1675,7 @@ class Program_weight_tensor_parameter_167: class Program_weight_tensor_parameter_168: name = "parameter_168" + original_name = "conv2d_8.b_0" shape = [256] dtype = "float32" data = None @@ -1514,6 +1683,7 @@ class Program_weight_tensor_parameter_168: class Program_weight_tensor_parameter_169: name = "parameter_169" + original_name = "conv2d_8.w_0" shape = [256, 256, 3, 3] dtype = "float32" min_val = float("-0.0360843") @@ -1525,6 +1695,7 @@ class Program_weight_tensor_parameter_169: class Program_weight_tensor_parameter_170: name = "parameter_170" + original_name = "group_norm_1.b_0" shape = [256] dtype = "float32" data = None @@ -1532,6 +1703,7 @@ class Program_weight_tensor_parameter_170: class Program_weight_tensor_parameter_171: name = "parameter_171" + original_name = "group_norm_1.w_0" shape = [256] dtype = "float32" min_val = float("-0.107791") @@ -1543,6 +1715,7 @@ class Program_weight_tensor_parameter_171: class Program_weight_tensor_parameter_172: name = "parameter_172" + original_name = "conv2d_2.w_0" shape = [256, 256, 3, 3] dtype = "float32" min_val = float("-0.0360841") @@ -1554,6 +1727,7 @@ class Program_weight_tensor_parameter_172: class Program_weight_tensor_parameter_173: name = "parameter_173" + original_name = "group_norm_0.b_0" shape = [256] dtype = "float32" data = None @@ -1561,6 +1735,7 @@ class Program_weight_tensor_parameter_173: class Program_weight_tensor_parameter_174: name = "parameter_174" + original_name = "group_norm_0.w_0" shape = [256] dtype = "float32" min_val = float("-0.108122") @@ -1572,6 +1747,7 @@ class Program_weight_tensor_parameter_174: class Program_weight_tensor_parameter_175: name = "parameter_175" + original_name = "conv2d_1.w_0" shape = [256, 96, 1, 1] dtype = "float32" min_val = float("-0.176776") @@ -1583,6 +1759,7 @@ class Program_weight_tensor_parameter_175: class Program_weight_tensor_parameter_176: name = "parameter_176" + original_name = "group_norm_3.b_0" shape = [256] dtype = "float32" data = None @@ -1590,6 +1767,7 @@ class Program_weight_tensor_parameter_176: class Program_weight_tensor_parameter_177: name = "parameter_177" + original_name = "group_norm_3.w_0" shape = [256] dtype = "float32" min_val = float("-0.107366") @@ -1601,6 +1779,7 @@ class Program_weight_tensor_parameter_177: class Program_weight_tensor_parameter_178: name = "parameter_178" + original_name = "conv2d_4.w_0" shape = [256, 256, 3, 3] dtype = "float32" min_val = float("-0.0360841") @@ -1612,6 +1791,7 @@ class Program_weight_tensor_parameter_178: class Program_weight_tensor_parameter_179: name = "parameter_179" + original_name = "group_norm_2.b_0" shape = [256] dtype = "float32" data = None @@ -1619,6 +1799,7 @@ class Program_weight_tensor_parameter_179: class Program_weight_tensor_parameter_180: name = "parameter_180" + original_name = "group_norm_2.w_0" shape = [256] dtype = "float32" min_val = float("-0.107236") @@ -1630,6 +1811,7 @@ class Program_weight_tensor_parameter_180: class Program_weight_tensor_parameter_181: name = "parameter_181" + original_name = "conv2d_3.w_0" shape = [256, 192, 1, 1] dtype = "float32" min_val = float("-0.124994") @@ -1641,6 +1823,7 @@ class Program_weight_tensor_parameter_181: class Program_weight_tensor_parameter_182: name = "parameter_182" + original_name = "group_norm_5.b_0" shape = [256] dtype = "float32" data = None @@ -1648,6 +1831,7 @@ class Program_weight_tensor_parameter_182: class Program_weight_tensor_parameter_183: name = "parameter_183" + original_name = "group_norm_5.w_0" shape = [256] dtype = "float32" min_val = float("-0.107132") @@ -1659,6 +1843,7 @@ class Program_weight_tensor_parameter_183: class Program_weight_tensor_parameter_184: name = "parameter_184" + original_name = "conv2d_6.w_0" shape = [256, 256, 3, 3] dtype = "float32" min_val = float("-0.0360844") @@ -1670,6 +1855,7 @@ class Program_weight_tensor_parameter_184: class Program_weight_tensor_parameter_185: name = "parameter_185" + original_name = "group_norm_4.b_0" shape = [256] dtype = "float32" data = None @@ -1677,6 +1863,7 @@ class Program_weight_tensor_parameter_185: class Program_weight_tensor_parameter_186: name = "parameter_186" + original_name = "group_norm_4.w_0" shape = [256] dtype = "float32" min_val = float("-0.107274") @@ -1688,6 +1875,7 @@ class Program_weight_tensor_parameter_186: class Program_weight_tensor_parameter_187: name = "parameter_187" + original_name = "conv2d_5.w_0" shape = [256, 384, 1, 1] dtype = "float32" min_val = float("-0.0883844") @@ -1699,6 +1887,7 @@ class Program_weight_tensor_parameter_187: class Program_weight_tensor_parameter_188: name = "parameter_188" + original_name = "group_norm_6.b_0" shape = [256] dtype = "float32" data = None @@ -1706,6 +1895,7 @@ class Program_weight_tensor_parameter_188: class Program_weight_tensor_parameter_189: name = "parameter_189" + original_name = "group_norm_6.w_0" shape = [256] dtype = "float32" min_val = float("-0.105307") @@ -1717,6 +1907,7 @@ class Program_weight_tensor_parameter_189: class Program_weight_tensor_parameter_190: name = "parameter_190" + original_name = "conv2d_7.w_0" shape = [256, 768, 3, 3] dtype = "float32" min_val = float("-0.0208333") @@ -1728,6 +1919,7 @@ class Program_weight_tensor_parameter_190: class Program_weight_tensor_parameter_191: name = "parameter_191" + original_name = "layer_norm_55.b_0" shape = [768] dtype = "float32" data = None @@ -1735,6 +1927,7 @@ class Program_weight_tensor_parameter_191: class Program_weight_tensor_parameter_192: name = "parameter_192" + original_name = "layer_norm_55.w_0" shape = [768] dtype = "float32" min_val = float("1.0") @@ -1745,6 +1938,7 @@ class Program_weight_tensor_parameter_192: class Program_weight_tensor_parameter_193: name = "parameter_193" + original_name = "linear_98.b_0" shape = [768] dtype = "float32" min_val = float("-7.8427") @@ -1756,6 +1950,7 @@ class Program_weight_tensor_parameter_193: class Program_weight_tensor_parameter_194: name = "parameter_194" + original_name = "linear_98.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-2.13627") @@ -1767,6 +1962,7 @@ class Program_weight_tensor_parameter_194: class Program_weight_tensor_parameter_195: name = "parameter_195" + original_name = "linear_97.b_0" shape = [3072] dtype = "float32" min_val = float("-7.65925") @@ -1778,6 +1974,7 @@ class Program_weight_tensor_parameter_195: class Program_weight_tensor_parameter_196: name = "parameter_196" + original_name = "linear_97.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-1.40798") @@ -1789,6 +1986,7 @@ class Program_weight_tensor_parameter_196: class Program_weight_tensor_parameter_197: name = "parameter_197" + original_name = "layer_norm_51.b_0" shape = [768] dtype = "float32" min_val = float("-7.05455") @@ -1800,6 +1998,7 @@ class Program_weight_tensor_parameter_197: class Program_weight_tensor_parameter_198: name = "parameter_198" + original_name = "layer_norm_51.w_0" shape = [768] dtype = "float32" min_val = float("-0.291477") @@ -1811,6 +2010,7 @@ class Program_weight_tensor_parameter_198: class Program_weight_tensor_parameter_199: name = "parameter_199" + original_name = "linear_96.b_0" shape = [768] dtype = "float32" min_val = float("-11.8044") @@ -1822,6 +2022,7 @@ class Program_weight_tensor_parameter_199: class Program_weight_tensor_parameter_200: name = "parameter_200" + original_name = "linear_96.w_0" shape = [768, 768] dtype = "float32" min_val = float("-1.72301") @@ -1833,6 +2034,7 @@ class Program_weight_tensor_parameter_200: class Program_weight_tensor_parameter_201: name = "parameter_201" + original_name = "linear_95.b_0" shape = [2304] dtype = "float32" min_val = float("-7.31156") @@ -1844,6 +2046,7 @@ class Program_weight_tensor_parameter_201: class Program_weight_tensor_parameter_202: name = "parameter_202" + original_name = "linear_95.w_0" shape = [768, 2304] dtype = "float32" min_val = float("-0.509874") @@ -1855,6 +2058,7 @@ class Program_weight_tensor_parameter_202: class Program_weight_tensor_parameter_203: name = "parameter_203" + original_name = "layer_norm_50.b_0" shape = [768] dtype = "float32" min_val = float("-7.10294") @@ -1866,6 +2070,7 @@ class Program_weight_tensor_parameter_203: class Program_weight_tensor_parameter_204: name = "parameter_204" + original_name = "layer_norm_50.w_0" shape = [768] dtype = "float32" min_val = float("0.00304417") @@ -1877,6 +2082,7 @@ class Program_weight_tensor_parameter_204: class Program_weight_tensor_parameter_205: name = "parameter_205" + original_name = "linear_94.b_0" shape = [768] dtype = "float32" min_val = float("-5.52452") @@ -1888,6 +2094,7 @@ class Program_weight_tensor_parameter_205: class Program_weight_tensor_parameter_206: name = "parameter_206" + original_name = "linear_94.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-2.01228") @@ -1899,6 +2106,7 @@ class Program_weight_tensor_parameter_206: class Program_weight_tensor_parameter_207: name = "parameter_207" + original_name = "linear_93.b_0" shape = [3072] dtype = "float32" min_val = float("-6.01539") @@ -1910,6 +2118,7 @@ class Program_weight_tensor_parameter_207: class Program_weight_tensor_parameter_208: name = "parameter_208" + original_name = "linear_93.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.43767") @@ -1921,6 +2130,7 @@ class Program_weight_tensor_parameter_208: class Program_weight_tensor_parameter_209: name = "parameter_209" + original_name = "layer_norm_49.b_0" shape = [768] dtype = "float32" min_val = float("-18.4878") @@ -1932,6 +2142,7 @@ class Program_weight_tensor_parameter_209: class Program_weight_tensor_parameter_210: name = "parameter_210" + original_name = "layer_norm_49.w_0" shape = [768] dtype = "float32" min_val = float("-0.0755746") @@ -1943,6 +2154,7 @@ class Program_weight_tensor_parameter_210: class Program_weight_tensor_parameter_211: name = "parameter_211" + original_name = "linear_92.b_0" shape = [768] dtype = "float32" min_val = float("-5.11022") @@ -1954,6 +2166,7 @@ class Program_weight_tensor_parameter_211: class Program_weight_tensor_parameter_212: name = "parameter_212" + original_name = "linear_92.w_0" shape = [768, 768] dtype = "float32" min_val = float("-1.51202") @@ -1965,6 +2178,7 @@ class Program_weight_tensor_parameter_212: class Program_weight_tensor_parameter_213: name = "parameter_213" + original_name = "linear_91.b_0" shape = [2304] dtype = "float32" min_val = float("-6.81097") @@ -1976,6 +2190,7 @@ class Program_weight_tensor_parameter_213: class Program_weight_tensor_parameter_214: name = "parameter_214" + original_name = "linear_91.w_0" shape = [768, 2304] dtype = "float32" min_val = float("-0.494403") @@ -1987,6 +2202,7 @@ class Program_weight_tensor_parameter_214: class Program_weight_tensor_parameter_215: name = "parameter_215" + original_name = "layer_norm_48.b_0" shape = [768] dtype = "float32" min_val = float("-3.91771") @@ -1998,6 +2214,7 @@ class Program_weight_tensor_parameter_215: class Program_weight_tensor_parameter_216: name = "parameter_216" + original_name = "layer_norm_48.w_0" shape = [768] dtype = "float32" min_val = float("-0.00323922") @@ -2009,6 +2226,7 @@ class Program_weight_tensor_parameter_216: class Program_weight_tensor_parameter_217: name = "parameter_217" + original_name = "layer_norm_54.b_0" shape = [384] dtype = "float32" data = None @@ -2016,6 +2234,7 @@ class Program_weight_tensor_parameter_217: class Program_weight_tensor_parameter_218: name = "parameter_218" + original_name = "layer_norm_54.w_0" shape = [384] dtype = "float32" min_val = float("1.0") @@ -2027,6 +2246,7 @@ class Program_weight_tensor_parameter_218: class Program_weight_tensor_parameter_219: name = "parameter_219" + original_name = "linear_90.w_0" shape = [1536, 768] dtype = "float32" min_val = float("-1.17259") @@ -2038,6 +2258,7 @@ class Program_weight_tensor_parameter_219: class Program_weight_tensor_parameter_220: name = "parameter_220" + original_name = "layer_norm_47.b_0" shape = [1536] dtype = "float32" min_val = float("-1.18188") @@ -2049,6 +2270,7 @@ class Program_weight_tensor_parameter_220: class Program_weight_tensor_parameter_221: name = "parameter_221" + original_name = "layer_norm_47.w_0" shape = [1536] dtype = "float32" min_val = float("0.143443") @@ -2060,6 +2282,7 @@ class Program_weight_tensor_parameter_221: class Program_weight_tensor_parameter_222: name = "parameter_222" + original_name = "linear_89.b_0" shape = [384] dtype = "float32" min_val = float("-7.84877") @@ -2071,6 +2294,7 @@ class Program_weight_tensor_parameter_222: class Program_weight_tensor_parameter_223: name = "parameter_223" + original_name = "linear_89.w_0" shape = [1536, 384] dtype = "float32" min_val = float("-2.35838") @@ -2082,6 +2306,7 @@ class Program_weight_tensor_parameter_223: class Program_weight_tensor_parameter_224: name = "parameter_224" + original_name = "linear_88.b_0" shape = [1536] dtype = "float32" min_val = float("-5.393") @@ -2093,6 +2318,7 @@ class Program_weight_tensor_parameter_224: class Program_weight_tensor_parameter_225: name = "parameter_225" + original_name = "linear_88.w_0" shape = [384, 1536] dtype = "float32" min_val = float("-0.87558") @@ -2104,6 +2330,7 @@ class Program_weight_tensor_parameter_225: class Program_weight_tensor_parameter_226: name = "parameter_226" + original_name = "layer_norm_46.b_0" shape = [384] dtype = "float32" min_val = float("-8.80398") @@ -2115,6 +2342,7 @@ class Program_weight_tensor_parameter_226: class Program_weight_tensor_parameter_227: name = "parameter_227" + original_name = "layer_norm_46.w_0" shape = [384] dtype = "float32" min_val = float("0.919731") @@ -2126,6 +2354,7 @@ class Program_weight_tensor_parameter_227: class Program_weight_tensor_parameter_228: name = "parameter_228" + original_name = "linear_87.b_0" shape = [384] dtype = "float32" min_val = float("-3.30452") @@ -2137,6 +2366,7 @@ class Program_weight_tensor_parameter_228: class Program_weight_tensor_parameter_229: name = "parameter_229" + original_name = "linear_87.w_0" shape = [384, 384] dtype = "float32" min_val = float("-1.35306") @@ -2148,6 +2378,7 @@ class Program_weight_tensor_parameter_229: class Program_weight_tensor_parameter_230: name = "parameter_230" + original_name = "linear_86.b_0" shape = [1152] dtype = "float32" min_val = float("-7.37382") @@ -2159,6 +2390,7 @@ class Program_weight_tensor_parameter_230: class Program_weight_tensor_parameter_231: name = "parameter_231" + original_name = "linear_86.w_0" shape = [384, 1152] dtype = "float32" min_val = float("-0.642222") @@ -2170,6 +2402,7 @@ class Program_weight_tensor_parameter_231: class Program_weight_tensor_parameter_232: name = "parameter_232" + original_name = "layer_norm_45.b_0" shape = [384] dtype = "float32" min_val = float("-1.79459") @@ -2181,6 +2414,7 @@ class Program_weight_tensor_parameter_232: class Program_weight_tensor_parameter_233: name = "parameter_233" + original_name = "layer_norm_45.w_0" shape = [384] dtype = "float32" min_val = float("0.147375") @@ -2192,6 +2426,7 @@ class Program_weight_tensor_parameter_233: class Program_weight_tensor_parameter_234: name = "parameter_234" + original_name = "linear_85.b_0" shape = [384] dtype = "float32" min_val = float("-1.63457") @@ -2203,6 +2438,7 @@ class Program_weight_tensor_parameter_234: class Program_weight_tensor_parameter_235: name = "parameter_235" + original_name = "linear_85.w_0" shape = [1536, 384] dtype = "float32" min_val = float("-2.04924") @@ -2214,6 +2450,7 @@ class Program_weight_tensor_parameter_235: class Program_weight_tensor_parameter_236: name = "parameter_236" + original_name = "linear_84.b_0" shape = [1536] dtype = "float32" min_val = float("-5.10001") @@ -2225,6 +2462,7 @@ class Program_weight_tensor_parameter_236: class Program_weight_tensor_parameter_237: name = "parameter_237" + original_name = "linear_84.w_0" shape = [384, 1536] dtype = "float32" min_val = float("-1.15585") @@ -2236,6 +2474,7 @@ class Program_weight_tensor_parameter_237: class Program_weight_tensor_parameter_238: name = "parameter_238" + original_name = "layer_norm_44.b_0" shape = [384] dtype = "float32" min_val = float("-8.2005") @@ -2247,6 +2486,7 @@ class Program_weight_tensor_parameter_238: class Program_weight_tensor_parameter_239: name = "parameter_239" + original_name = "layer_norm_44.w_0" shape = [384] dtype = "float32" min_val = float("0.799803") @@ -2258,6 +2498,7 @@ class Program_weight_tensor_parameter_239: class Program_weight_tensor_parameter_240: name = "parameter_240" + original_name = "linear_83.b_0" shape = [384] dtype = "float32" min_val = float("-2.67407") @@ -2269,6 +2510,7 @@ class Program_weight_tensor_parameter_240: class Program_weight_tensor_parameter_241: name = "parameter_241" + original_name = "linear_83.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.650043") @@ -2280,6 +2522,7 @@ class Program_weight_tensor_parameter_241: class Program_weight_tensor_parameter_242: name = "parameter_242" + original_name = "linear_82.b_0" shape = [1152] dtype = "float32" min_val = float("-6.66009") @@ -2291,6 +2534,7 @@ class Program_weight_tensor_parameter_242: class Program_weight_tensor_parameter_243: name = "parameter_243" + original_name = "linear_82.w_0" shape = [384, 1152] dtype = "float32" min_val = float("-0.461448") @@ -2302,6 +2546,7 @@ class Program_weight_tensor_parameter_243: class Program_weight_tensor_parameter_244: name = "parameter_244" + original_name = "layer_norm_43.b_0" shape = [384] dtype = "float32" min_val = float("-1.52346") @@ -2313,6 +2558,7 @@ class Program_weight_tensor_parameter_244: class Program_weight_tensor_parameter_245: name = "parameter_245" + original_name = "layer_norm_43.w_0" shape = [384] dtype = "float32" min_val = float("0.157076") @@ -2324,6 +2570,7 @@ class Program_weight_tensor_parameter_245: class Program_weight_tensor_parameter_246: name = "parameter_246" + original_name = "linear_81.b_0" shape = [384] dtype = "float32" min_val = float("-1.52422") @@ -2335,6 +2582,7 @@ class Program_weight_tensor_parameter_246: class Program_weight_tensor_parameter_247: name = "parameter_247" + original_name = "linear_81.w_0" shape = [1536, 384] dtype = "float32" min_val = float("-1.66223") @@ -2346,6 +2594,7 @@ class Program_weight_tensor_parameter_247: class Program_weight_tensor_parameter_248: name = "parameter_248" + original_name = "linear_80.b_0" shape = [1536] dtype = "float32" min_val = float("-5.18373") @@ -2357,6 +2606,7 @@ class Program_weight_tensor_parameter_248: class Program_weight_tensor_parameter_249: name = "parameter_249" + original_name = "linear_80.w_0" shape = [384, 1536] dtype = "float32" min_val = float("-1.20777") @@ -2368,6 +2618,7 @@ class Program_weight_tensor_parameter_249: class Program_weight_tensor_parameter_250: name = "parameter_250" + original_name = "layer_norm_42.b_0" shape = [384] dtype = "float32" min_val = float("-6.61675") @@ -2379,6 +2630,7 @@ class Program_weight_tensor_parameter_250: class Program_weight_tensor_parameter_251: name = "parameter_251" + original_name = "layer_norm_42.w_0" shape = [384] dtype = "float32" min_val = float("0.670386") @@ -2390,6 +2642,7 @@ class Program_weight_tensor_parameter_251: class Program_weight_tensor_parameter_252: name = "parameter_252" + original_name = "linear_79.b_0" shape = [384] dtype = "float32" min_val = float("-1.34129") @@ -2401,6 +2654,7 @@ class Program_weight_tensor_parameter_252: class Program_weight_tensor_parameter_253: name = "parameter_253" + original_name = "linear_79.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.423975") @@ -2412,6 +2666,7 @@ class Program_weight_tensor_parameter_253: class Program_weight_tensor_parameter_254: name = "parameter_254" + original_name = "linear_78.b_0" shape = [1152] dtype = "float32" min_val = float("-7.1346") @@ -2423,6 +2678,7 @@ class Program_weight_tensor_parameter_254: class Program_weight_tensor_parameter_255: name = "parameter_255" + original_name = "linear_78.w_0" shape = [384, 1152] dtype = "float32" min_val = float("-0.421479") @@ -2434,6 +2690,7 @@ class Program_weight_tensor_parameter_255: class Program_weight_tensor_parameter_256: name = "parameter_256" + original_name = "layer_norm_41.b_0" shape = [384] dtype = "float32" min_val = float("-0.59226") @@ -2445,6 +2702,7 @@ class Program_weight_tensor_parameter_256: class Program_weight_tensor_parameter_257: name = "parameter_257" + original_name = "layer_norm_41.w_0" shape = [384] dtype = "float32" min_val = float("0.0360341") @@ -2456,6 +2714,7 @@ class Program_weight_tensor_parameter_257: class Program_weight_tensor_parameter_258: name = "parameter_258" + original_name = "linear_77.b_0" shape = [384] dtype = "float32" min_val = float("-1.54795") @@ -2467,6 +2726,7 @@ class Program_weight_tensor_parameter_258: class Program_weight_tensor_parameter_259: name = "parameter_259" + original_name = "linear_77.w_0" shape = [1536, 384] dtype = "float32" min_val = float("-0.755414") @@ -2478,6 +2738,7 @@ class Program_weight_tensor_parameter_259: class Program_weight_tensor_parameter_260: name = "parameter_260" + original_name = "linear_76.b_0" shape = [1536] dtype = "float32" min_val = float("-5.08707") @@ -2489,6 +2750,7 @@ class Program_weight_tensor_parameter_260: class Program_weight_tensor_parameter_261: name = "parameter_261" + original_name = "linear_76.w_0" shape = [384, 1536] dtype = "float32" min_val = float("-0.5499") @@ -2500,6 +2762,7 @@ class Program_weight_tensor_parameter_261: class Program_weight_tensor_parameter_262: name = "parameter_262" + original_name = "layer_norm_40.b_0" shape = [384] dtype = "float32" min_val = float("-5.38069") @@ -2511,6 +2774,7 @@ class Program_weight_tensor_parameter_262: class Program_weight_tensor_parameter_263: name = "parameter_263" + original_name = "layer_norm_40.w_0" shape = [384] dtype = "float32" min_val = float("0.598278") @@ -2522,6 +2786,7 @@ class Program_weight_tensor_parameter_263: class Program_weight_tensor_parameter_264: name = "parameter_264" + original_name = "linear_75.b_0" shape = [384] dtype = "float32" min_val = float("-1.48892") @@ -2533,6 +2798,7 @@ class Program_weight_tensor_parameter_264: class Program_weight_tensor_parameter_265: name = "parameter_265" + original_name = "linear_75.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.726442") @@ -2544,6 +2810,7 @@ class Program_weight_tensor_parameter_265: class Program_weight_tensor_parameter_266: name = "parameter_266" + original_name = "linear_74.b_0" shape = [1152] dtype = "float32" min_val = float("-6.64044") @@ -2555,6 +2822,7 @@ class Program_weight_tensor_parameter_266: class Program_weight_tensor_parameter_267: name = "parameter_267" + original_name = "linear_74.w_0" shape = [384, 1152] dtype = "float32" min_val = float("-0.630656") @@ -2566,6 +2834,7 @@ class Program_weight_tensor_parameter_267: class Program_weight_tensor_parameter_268: name = "parameter_268" + original_name = "layer_norm_39.b_0" shape = [384] dtype = "float32" min_val = float("-1.17042") @@ -2577,6 +2846,7 @@ class Program_weight_tensor_parameter_268: class Program_weight_tensor_parameter_269: name = "parameter_269" + original_name = "layer_norm_39.w_0" shape = [384] dtype = "float32" min_val = float("0.149416") @@ -2588,6 +2858,7 @@ class Program_weight_tensor_parameter_269: class Program_weight_tensor_parameter_270: name = "parameter_270" + original_name = "linear_73.b_0" shape = [384] dtype = "float32" min_val = float("-1.30153") @@ -2599,6 +2870,7 @@ class Program_weight_tensor_parameter_270: class Program_weight_tensor_parameter_271: name = "parameter_271" + original_name = "linear_73.w_0" shape = [1536, 384] dtype = "float32" min_val = float("-0.844984") @@ -2610,6 +2882,7 @@ class Program_weight_tensor_parameter_271: class Program_weight_tensor_parameter_272: name = "parameter_272" + original_name = "linear_72.b_0" shape = [1536] dtype = "float32" min_val = float("-4.83716") @@ -2621,6 +2894,7 @@ class Program_weight_tensor_parameter_272: class Program_weight_tensor_parameter_273: name = "parameter_273" + original_name = "linear_72.w_0" shape = [384, 1536] dtype = "float32" min_val = float("-0.762709") @@ -2632,6 +2906,7 @@ class Program_weight_tensor_parameter_273: class Program_weight_tensor_parameter_274: name = "parameter_274" + original_name = "layer_norm_38.b_0" shape = [384] dtype = "float32" min_val = float("-2.3715") @@ -2643,6 +2918,7 @@ class Program_weight_tensor_parameter_274: class Program_weight_tensor_parameter_275: name = "parameter_275" + original_name = "layer_norm_38.w_0" shape = [384] dtype = "float32" min_val = float("0.511289") @@ -2654,6 +2930,7 @@ class Program_weight_tensor_parameter_275: class Program_weight_tensor_parameter_276: name = "parameter_276" + original_name = "linear_71.b_0" shape = [384] dtype = "float32" min_val = float("-1.07573") @@ -2665,6 +2942,7 @@ class Program_weight_tensor_parameter_276: class Program_weight_tensor_parameter_277: name = "parameter_277" + original_name = "linear_71.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.806431") @@ -2676,6 +2954,7 @@ class Program_weight_tensor_parameter_277: class Program_weight_tensor_parameter_278: name = "parameter_278" + original_name = "linear_70.b_0" shape = [1152] dtype = "float32" min_val = float("-7.88966") @@ -2687,6 +2966,7 @@ class Program_weight_tensor_parameter_278: class Program_weight_tensor_parameter_279: name = "parameter_279" + original_name = "linear_70.w_0" shape = [384, 1152] dtype = "float32" min_val = float("-0.575804") @@ -2698,6 +2978,7 @@ class Program_weight_tensor_parameter_279: class Program_weight_tensor_parameter_280: name = "parameter_280" + original_name = "layer_norm_37.b_0" shape = [384] dtype = "float32" min_val = float("-2.409") @@ -2709,6 +2990,7 @@ class Program_weight_tensor_parameter_280: class Program_weight_tensor_parameter_281: name = "parameter_281" + original_name = "layer_norm_37.w_0" shape = [384] dtype = "float32" min_val = float("0.2557") @@ -2720,6 +3002,7 @@ class Program_weight_tensor_parameter_281: class Program_weight_tensor_parameter_282: name = "parameter_282" + original_name = "linear_69.b_0" shape = [384] dtype = "float32" min_val = float("-1.72326") @@ -2731,6 +3014,7 @@ class Program_weight_tensor_parameter_282: class Program_weight_tensor_parameter_283: name = "parameter_283" + original_name = "linear_69.w_0" shape = [1536, 384] dtype = "float32" min_val = float("-0.954852") @@ -2742,6 +3026,7 @@ class Program_weight_tensor_parameter_283: class Program_weight_tensor_parameter_284: name = "parameter_284" + original_name = "linear_68.b_0" shape = [1536] dtype = "float32" min_val = float("-4.77488") @@ -2753,6 +3038,7 @@ class Program_weight_tensor_parameter_284: class Program_weight_tensor_parameter_285: name = "parameter_285" + original_name = "linear_68.w_0" shape = [384, 1536] dtype = "float32" min_val = float("-0.673223") @@ -2764,6 +3050,7 @@ class Program_weight_tensor_parameter_285: class Program_weight_tensor_parameter_286: name = "parameter_286" + original_name = "layer_norm_36.b_0" shape = [384] dtype = "float32" min_val = float("-2.68383") @@ -2775,6 +3062,7 @@ class Program_weight_tensor_parameter_286: class Program_weight_tensor_parameter_287: name = "parameter_287" + original_name = "layer_norm_36.w_0" shape = [384] dtype = "float32" min_val = float("0.393495") @@ -2786,6 +3074,7 @@ class Program_weight_tensor_parameter_287: class Program_weight_tensor_parameter_288: name = "parameter_288" + original_name = "linear_67.b_0" shape = [384] dtype = "float32" min_val = float("-1.73033") @@ -2797,6 +3086,7 @@ class Program_weight_tensor_parameter_288: class Program_weight_tensor_parameter_289: name = "parameter_289" + original_name = "linear_67.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.937806") @@ -2808,6 +3098,7 @@ class Program_weight_tensor_parameter_289: class Program_weight_tensor_parameter_290: name = "parameter_290" + original_name = "linear_66.b_0" shape = [1152] dtype = "float32" min_val = float("-6.46573") @@ -2819,6 +3110,7 @@ class Program_weight_tensor_parameter_290: class Program_weight_tensor_parameter_291: name = "parameter_291" + original_name = "linear_66.w_0" shape = [384, 1152] dtype = "float32" min_val = float("-0.447358") @@ -2830,6 +3122,7 @@ class Program_weight_tensor_parameter_291: class Program_weight_tensor_parameter_292: name = "parameter_292" + original_name = "layer_norm_35.b_0" shape = [384] dtype = "float32" min_val = float("-4.08543") @@ -2841,6 +3134,7 @@ class Program_weight_tensor_parameter_292: class Program_weight_tensor_parameter_293: name = "parameter_293" + original_name = "layer_norm_35.w_0" shape = [384] dtype = "float32" min_val = float("0.376255") @@ -2852,6 +3146,7 @@ class Program_weight_tensor_parameter_293: class Program_weight_tensor_parameter_294: name = "parameter_294" + original_name = "linear_65.b_0" shape = [384] dtype = "float32" min_val = float("-2.54107") @@ -2863,6 +3158,7 @@ class Program_weight_tensor_parameter_294: class Program_weight_tensor_parameter_295: name = "parameter_295" + original_name = "linear_65.w_0" shape = [1536, 384] dtype = "float32" min_val = float("-1.06038") @@ -2874,6 +3170,7 @@ class Program_weight_tensor_parameter_295: class Program_weight_tensor_parameter_296: name = "parameter_296" + original_name = "linear_64.b_0" shape = [1536] dtype = "float32" min_val = float("-7.28658") @@ -2885,6 +3182,7 @@ class Program_weight_tensor_parameter_296: class Program_weight_tensor_parameter_297: name = "parameter_297" + original_name = "linear_64.w_0" shape = [384, 1536] dtype = "float32" min_val = float("-0.67299") @@ -2896,6 +3194,7 @@ class Program_weight_tensor_parameter_297: class Program_weight_tensor_parameter_298: name = "parameter_298" + original_name = "layer_norm_34.b_0" shape = [384] dtype = "float32" min_val = float("-3.21753") @@ -2907,6 +3206,7 @@ class Program_weight_tensor_parameter_298: class Program_weight_tensor_parameter_299: name = "parameter_299" + original_name = "layer_norm_34.w_0" shape = [384] dtype = "float32" min_val = float("0.62258") @@ -2918,6 +3218,7 @@ class Program_weight_tensor_parameter_299: class Program_weight_tensor_parameter_300: name = "parameter_300" + original_name = "linear_63.b_0" shape = [384] dtype = "float32" min_val = float("-1.14977") @@ -2929,6 +3230,7 @@ class Program_weight_tensor_parameter_300: class Program_weight_tensor_parameter_301: name = "parameter_301" + original_name = "linear_63.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.580466") @@ -2940,6 +3242,7 @@ class Program_weight_tensor_parameter_301: class Program_weight_tensor_parameter_302: name = "parameter_302" + original_name = "linear_62.b_0" shape = [1152] dtype = "float32" min_val = float("-6.71017") @@ -2951,6 +3254,7 @@ class Program_weight_tensor_parameter_302: class Program_weight_tensor_parameter_303: name = "parameter_303" + original_name = "linear_62.w_0" shape = [384, 1152] dtype = "float32" min_val = float("-0.454211") @@ -2962,6 +3266,7 @@ class Program_weight_tensor_parameter_303: class Program_weight_tensor_parameter_304: name = "parameter_304" + original_name = "layer_norm_33.b_0" shape = [384] dtype = "float32" min_val = float("-3.97081") @@ -2973,6 +3278,7 @@ class Program_weight_tensor_parameter_304: class Program_weight_tensor_parameter_305: name = "parameter_305" + original_name = "layer_norm_33.w_0" shape = [384] dtype = "float32" min_val = float("0.333199") @@ -2984,6 +3290,7 @@ class Program_weight_tensor_parameter_305: class Program_weight_tensor_parameter_306: name = "parameter_306" + original_name = "linear_61.b_0" shape = [384] dtype = "float32" min_val = float("-2.57988") @@ -2995,6 +3302,7 @@ class Program_weight_tensor_parameter_306: class Program_weight_tensor_parameter_307: name = "parameter_307" + original_name = "linear_61.w_0" shape = [1536, 384] dtype = "float32" min_val = float("-1.12995") @@ -3006,6 +3314,7 @@ class Program_weight_tensor_parameter_307: class Program_weight_tensor_parameter_308: name = "parameter_308" + original_name = "linear_60.b_0" shape = [1536] dtype = "float32" min_val = float("-8.36899") @@ -3017,6 +3326,7 @@ class Program_weight_tensor_parameter_308: class Program_weight_tensor_parameter_309: name = "parameter_309" + original_name = "linear_60.w_0" shape = [384, 1536] dtype = "float32" min_val = float("-0.728203") @@ -3028,6 +3338,7 @@ class Program_weight_tensor_parameter_309: class Program_weight_tensor_parameter_310: name = "parameter_310" + original_name = "layer_norm_32.b_0" shape = [384] dtype = "float32" min_val = float("-2.96812") @@ -3039,6 +3350,7 @@ class Program_weight_tensor_parameter_310: class Program_weight_tensor_parameter_311: name = "parameter_311" + original_name = "layer_norm_32.w_0" shape = [384] dtype = "float32" min_val = float("0.60936") @@ -3050,6 +3362,7 @@ class Program_weight_tensor_parameter_311: class Program_weight_tensor_parameter_312: name = "parameter_312" + original_name = "linear_59.b_0" shape = [384] dtype = "float32" min_val = float("-1.60122") @@ -3061,6 +3374,7 @@ class Program_weight_tensor_parameter_312: class Program_weight_tensor_parameter_313: name = "parameter_313" + original_name = "linear_59.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.473329") @@ -3072,6 +3386,7 @@ class Program_weight_tensor_parameter_313: class Program_weight_tensor_parameter_314: name = "parameter_314" + original_name = "linear_58.b_0" shape = [1152] dtype = "float32" min_val = float("-6.15011") @@ -3083,6 +3398,7 @@ class Program_weight_tensor_parameter_314: class Program_weight_tensor_parameter_315: name = "parameter_315" + original_name = "linear_58.w_0" shape = [384, 1152] dtype = "float32" min_val = float("-0.467695") @@ -3094,6 +3410,7 @@ class Program_weight_tensor_parameter_315: class Program_weight_tensor_parameter_316: name = "parameter_316" + original_name = "layer_norm_31.b_0" shape = [384] dtype = "float32" min_val = float("-4.31834") @@ -3105,6 +3422,7 @@ class Program_weight_tensor_parameter_316: class Program_weight_tensor_parameter_317: name = "parameter_317" + original_name = "layer_norm_31.w_0" shape = [384] dtype = "float32" min_val = float("0.258324") @@ -3116,6 +3434,7 @@ class Program_weight_tensor_parameter_317: class Program_weight_tensor_parameter_318: name = "parameter_318" + original_name = "linear_57.b_0" shape = [384] dtype = "float32" min_val = float("-1.21956") @@ -3127,6 +3446,7 @@ class Program_weight_tensor_parameter_318: class Program_weight_tensor_parameter_319: name = "parameter_319" + original_name = "linear_57.w_0" shape = [1536, 384] dtype = "float32" min_val = float("-1.189") @@ -3138,6 +3458,7 @@ class Program_weight_tensor_parameter_319: class Program_weight_tensor_parameter_320: name = "parameter_320" + original_name = "linear_56.b_0" shape = [1536] dtype = "float32" min_val = float("-6.92092") @@ -3149,6 +3470,7 @@ class Program_weight_tensor_parameter_320: class Program_weight_tensor_parameter_321: name = "parameter_321" + original_name = "linear_56.w_0" shape = [384, 1536] dtype = "float32" min_val = float("-0.722332") @@ -3160,6 +3482,7 @@ class Program_weight_tensor_parameter_321: class Program_weight_tensor_parameter_322: name = "parameter_322" + original_name = "layer_norm_30.b_0" shape = [384] dtype = "float32" min_val = float("-3.04891") @@ -3171,6 +3494,7 @@ class Program_weight_tensor_parameter_322: class Program_weight_tensor_parameter_323: name = "parameter_323" + original_name = "layer_norm_30.w_0" shape = [384] dtype = "float32" min_val = float("0.595418") @@ -3182,6 +3506,7 @@ class Program_weight_tensor_parameter_323: class Program_weight_tensor_parameter_324: name = "parameter_324" + original_name = "linear_55.b_0" shape = [384] dtype = "float32" min_val = float("-1.00124") @@ -3193,6 +3518,7 @@ class Program_weight_tensor_parameter_324: class Program_weight_tensor_parameter_325: name = "parameter_325" + original_name = "linear_55.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.333799") @@ -3204,6 +3530,7 @@ class Program_weight_tensor_parameter_325: class Program_weight_tensor_parameter_326: name = "parameter_326" + original_name = "linear_54.b_0" shape = [1152] dtype = "float32" min_val = float("-4.96308") @@ -3215,6 +3542,7 @@ class Program_weight_tensor_parameter_326: class Program_weight_tensor_parameter_327: name = "parameter_327" + original_name = "linear_54.w_0" shape = [384, 1152] dtype = "float32" min_val = float("-0.537365") @@ -3226,6 +3554,7 @@ class Program_weight_tensor_parameter_327: class Program_weight_tensor_parameter_328: name = "parameter_328" + original_name = "layer_norm_29.b_0" shape = [384] dtype = "float32" min_val = float("-4.5908") @@ -3237,6 +3566,7 @@ class Program_weight_tensor_parameter_328: class Program_weight_tensor_parameter_329: name = "parameter_329" + original_name = "layer_norm_29.w_0" shape = [384] dtype = "float32" min_val = float("0.243951") @@ -3248,6 +3578,7 @@ class Program_weight_tensor_parameter_329: class Program_weight_tensor_parameter_330: name = "parameter_330" + original_name = "linear_53.b_0" shape = [384] dtype = "float32" min_val = float("-1.404") @@ -3259,6 +3590,7 @@ class Program_weight_tensor_parameter_330: class Program_weight_tensor_parameter_331: name = "parameter_331" + original_name = "linear_53.w_0" shape = [1536, 384] dtype = "float32" min_val = float("-0.64012") @@ -3270,6 +3602,7 @@ class Program_weight_tensor_parameter_331: class Program_weight_tensor_parameter_332: name = "parameter_332" + original_name = "linear_52.b_0" shape = [1536] dtype = "float32" min_val = float("-9.64632") @@ -3281,6 +3614,7 @@ class Program_weight_tensor_parameter_332: class Program_weight_tensor_parameter_333: name = "parameter_333" + original_name = "linear_52.w_0" shape = [384, 1536] dtype = "float32" min_val = float("-0.628208") @@ -3292,6 +3626,7 @@ class Program_weight_tensor_parameter_333: class Program_weight_tensor_parameter_334: name = "parameter_334" + original_name = "layer_norm_28.b_0" shape = [384] dtype = "float32" min_val = float("-2.89532") @@ -3303,6 +3638,7 @@ class Program_weight_tensor_parameter_334: class Program_weight_tensor_parameter_335: name = "parameter_335" + original_name = "layer_norm_28.w_0" shape = [384] dtype = "float32" min_val = float("0.526601") @@ -3314,6 +3650,7 @@ class Program_weight_tensor_parameter_335: class Program_weight_tensor_parameter_336: name = "parameter_336" + original_name = "linear_51.b_0" shape = [384] dtype = "float32" min_val = float("-1.51566") @@ -3325,6 +3662,7 @@ class Program_weight_tensor_parameter_336: class Program_weight_tensor_parameter_337: name = "parameter_337" + original_name = "linear_51.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.729258") @@ -3336,6 +3674,7 @@ class Program_weight_tensor_parameter_337: class Program_weight_tensor_parameter_338: name = "parameter_338" + original_name = "linear_50.b_0" shape = [1152] dtype = "float32" min_val = float("-5.14222") @@ -3347,6 +3686,7 @@ class Program_weight_tensor_parameter_338: class Program_weight_tensor_parameter_339: name = "parameter_339" + original_name = "linear_50.w_0" shape = [384, 1152] dtype = "float32" min_val = float("-0.572713") @@ -3358,6 +3698,7 @@ class Program_weight_tensor_parameter_339: class Program_weight_tensor_parameter_340: name = "parameter_340" + original_name = "layer_norm_27.b_0" shape = [384] dtype = "float32" min_val = float("-4.08867") @@ -3369,6 +3710,7 @@ class Program_weight_tensor_parameter_340: class Program_weight_tensor_parameter_341: name = "parameter_341" + original_name = "layer_norm_27.w_0" shape = [384] dtype = "float32" min_val = float("0.239561") @@ -3380,6 +3722,7 @@ class Program_weight_tensor_parameter_341: class Program_weight_tensor_parameter_342: name = "parameter_342" + original_name = "linear_49.b_0" shape = [384] dtype = "float32" min_val = float("-1.1093") @@ -3391,6 +3734,7 @@ class Program_weight_tensor_parameter_342: class Program_weight_tensor_parameter_343: name = "parameter_343" + original_name = "linear_49.w_0" shape = [1536, 384] dtype = "float32" min_val = float("-0.718471") @@ -3402,6 +3746,7 @@ class Program_weight_tensor_parameter_343: class Program_weight_tensor_parameter_344: name = "parameter_344" + original_name = "linear_48.b_0" shape = [1536] dtype = "float32" min_val = float("-5.67662") @@ -3413,6 +3758,7 @@ class Program_weight_tensor_parameter_344: class Program_weight_tensor_parameter_345: name = "parameter_345" + original_name = "linear_48.w_0" shape = [384, 1536] dtype = "float32" min_val = float("-0.687415") @@ -3424,6 +3770,7 @@ class Program_weight_tensor_parameter_345: class Program_weight_tensor_parameter_346: name = "parameter_346" + original_name = "layer_norm_26.b_0" shape = [384] dtype = "float32" min_val = float("-3.70984") @@ -3435,6 +3782,7 @@ class Program_weight_tensor_parameter_346: class Program_weight_tensor_parameter_347: name = "parameter_347" + original_name = "layer_norm_26.w_0" shape = [384] dtype = "float32" min_val = float("0.533325") @@ -3446,6 +3794,7 @@ class Program_weight_tensor_parameter_347: class Program_weight_tensor_parameter_348: name = "parameter_348" + original_name = "linear_47.b_0" shape = [384] dtype = "float32" min_val = float("-1.38298") @@ -3457,6 +3806,7 @@ class Program_weight_tensor_parameter_348: class Program_weight_tensor_parameter_349: name = "parameter_349" + original_name = "linear_47.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.821947") @@ -3468,6 +3818,7 @@ class Program_weight_tensor_parameter_349: class Program_weight_tensor_parameter_350: name = "parameter_350" + original_name = "linear_46.b_0" shape = [1152] dtype = "float32" min_val = float("-4.94495") @@ -3479,6 +3830,7 @@ class Program_weight_tensor_parameter_350: class Program_weight_tensor_parameter_351: name = "parameter_351" + original_name = "linear_46.w_0" shape = [384, 1152] dtype = "float32" min_val = float("-0.606119") @@ -3490,6 +3842,7 @@ class Program_weight_tensor_parameter_351: class Program_weight_tensor_parameter_352: name = "parameter_352" + original_name = "layer_norm_25.b_0" shape = [384] dtype = "float32" min_val = float("-3.64079") @@ -3501,6 +3854,7 @@ class Program_weight_tensor_parameter_352: class Program_weight_tensor_parameter_353: name = "parameter_353" + original_name = "layer_norm_25.w_0" shape = [384] dtype = "float32" min_val = float("0.25278") @@ -3512,6 +3866,7 @@ class Program_weight_tensor_parameter_353: class Program_weight_tensor_parameter_354: name = "parameter_354" + original_name = "linear_45.b_0" shape = [384] dtype = "float32" min_val = float("-1.26844") @@ -3523,6 +3878,7 @@ class Program_weight_tensor_parameter_354: class Program_weight_tensor_parameter_355: name = "parameter_355" + original_name = "linear_45.w_0" shape = [1536, 384] dtype = "float32" min_val = float("-0.705548") @@ -3534,6 +3890,7 @@ class Program_weight_tensor_parameter_355: class Program_weight_tensor_parameter_356: name = "parameter_356" + original_name = "linear_44.b_0" shape = [1536] dtype = "float32" min_val = float("-5.12857") @@ -3545,6 +3902,7 @@ class Program_weight_tensor_parameter_356: class Program_weight_tensor_parameter_357: name = "parameter_357" + original_name = "linear_44.w_0" shape = [384, 1536] dtype = "float32" min_val = float("-0.5558") @@ -3556,6 +3914,7 @@ class Program_weight_tensor_parameter_357: class Program_weight_tensor_parameter_358: name = "parameter_358" + original_name = "layer_norm_24.b_0" shape = [384] dtype = "float32" min_val = float("-3.75593") @@ -3567,6 +3926,7 @@ class Program_weight_tensor_parameter_358: class Program_weight_tensor_parameter_359: name = "parameter_359" + original_name = "layer_norm_24.w_0" shape = [384] dtype = "float32" min_val = float("0.50125") @@ -3578,6 +3938,7 @@ class Program_weight_tensor_parameter_359: class Program_weight_tensor_parameter_360: name = "parameter_360" + original_name = "linear_43.b_0" shape = [384] dtype = "float32" min_val = float("-1.46327") @@ -3589,6 +3950,7 @@ class Program_weight_tensor_parameter_360: class Program_weight_tensor_parameter_361: name = "parameter_361" + original_name = "linear_43.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.716069") @@ -3600,6 +3962,7 @@ class Program_weight_tensor_parameter_361: class Program_weight_tensor_parameter_362: name = "parameter_362" + original_name = "linear_42.b_0" shape = [1152] dtype = "float32" min_val = float("-3.71036") @@ -3611,6 +3974,7 @@ class Program_weight_tensor_parameter_362: class Program_weight_tensor_parameter_363: name = "parameter_363" + original_name = "linear_42.w_0" shape = [384, 1152] dtype = "float32" min_val = float("-0.456241") @@ -3622,6 +3986,7 @@ class Program_weight_tensor_parameter_363: class Program_weight_tensor_parameter_364: name = "parameter_364" + original_name = "layer_norm_23.b_0" shape = [384] dtype = "float32" min_val = float("-3.47193") @@ -3633,6 +3998,7 @@ class Program_weight_tensor_parameter_364: class Program_weight_tensor_parameter_365: name = "parameter_365" + original_name = "layer_norm_23.w_0" shape = [384] dtype = "float32" min_val = float("0.210565") @@ -3644,6 +4010,7 @@ class Program_weight_tensor_parameter_365: class Program_weight_tensor_parameter_366: name = "parameter_366" + original_name = "linear_41.b_0" shape = [384] dtype = "float32" min_val = float("-1.15477") @@ -3655,6 +4022,7 @@ class Program_weight_tensor_parameter_366: class Program_weight_tensor_parameter_367: name = "parameter_367" + original_name = "linear_41.w_0" shape = [1536, 384] dtype = "float32" min_val = float("-0.639952") @@ -3666,6 +4034,7 @@ class Program_weight_tensor_parameter_367: class Program_weight_tensor_parameter_368: name = "parameter_368" + original_name = "linear_40.b_0" shape = [1536] dtype = "float32" min_val = float("-5.43379") @@ -3677,6 +4046,7 @@ class Program_weight_tensor_parameter_368: class Program_weight_tensor_parameter_369: name = "parameter_369" + original_name = "linear_40.w_0" shape = [384, 1536] dtype = "float32" min_val = float("-0.674227") @@ -3688,6 +4058,7 @@ class Program_weight_tensor_parameter_369: class Program_weight_tensor_parameter_370: name = "parameter_370" + original_name = "layer_norm_22.b_0" shape = [384] dtype = "float32" min_val = float("-4.02772") @@ -3699,6 +4070,7 @@ class Program_weight_tensor_parameter_370: class Program_weight_tensor_parameter_371: name = "parameter_371" + original_name = "layer_norm_22.w_0" shape = [384] dtype = "float32" min_val = float("0.355152") @@ -3710,6 +4082,7 @@ class Program_weight_tensor_parameter_371: class Program_weight_tensor_parameter_372: name = "parameter_372" + original_name = "linear_39.b_0" shape = [384] dtype = "float32" min_val = float("-1.36797") @@ -3721,6 +4094,7 @@ class Program_weight_tensor_parameter_372: class Program_weight_tensor_parameter_373: name = "parameter_373" + original_name = "linear_39.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.601018") @@ -3732,6 +4106,7 @@ class Program_weight_tensor_parameter_373: class Program_weight_tensor_parameter_374: name = "parameter_374" + original_name = "linear_38.b_0" shape = [1152] dtype = "float32" min_val = float("-4.19132") @@ -3743,6 +4118,7 @@ class Program_weight_tensor_parameter_374: class Program_weight_tensor_parameter_375: name = "parameter_375" + original_name = "linear_38.w_0" shape = [384, 1152] dtype = "float32" min_val = float("-0.461236") @@ -3754,6 +4130,7 @@ class Program_weight_tensor_parameter_375: class Program_weight_tensor_parameter_376: name = "parameter_376" + original_name = "layer_norm_21.b_0" shape = [384] dtype = "float32" min_val = float("-3.1804") @@ -3765,6 +4142,7 @@ class Program_weight_tensor_parameter_376: class Program_weight_tensor_parameter_377: name = "parameter_377" + original_name = "layer_norm_21.w_0" shape = [384] dtype = "float32" min_val = float("0.204104") @@ -3776,6 +4154,7 @@ class Program_weight_tensor_parameter_377: class Program_weight_tensor_parameter_378: name = "parameter_378" + original_name = "linear_37.b_0" shape = [384] dtype = "float32" min_val = float("-1.2134") @@ -3787,6 +4166,7 @@ class Program_weight_tensor_parameter_378: class Program_weight_tensor_parameter_379: name = "parameter_379" + original_name = "linear_37.w_0" shape = [1536, 384] dtype = "float32" min_val = float("-0.57784") @@ -3798,6 +4178,7 @@ class Program_weight_tensor_parameter_379: class Program_weight_tensor_parameter_380: name = "parameter_380" + original_name = "linear_36.b_0" shape = [1536] dtype = "float32" min_val = float("-5.32214") @@ -3809,6 +4190,7 @@ class Program_weight_tensor_parameter_380: class Program_weight_tensor_parameter_381: name = "parameter_381" + original_name = "linear_36.w_0" shape = [384, 1536] dtype = "float32" min_val = float("-0.59506") @@ -3820,6 +4202,7 @@ class Program_weight_tensor_parameter_381: class Program_weight_tensor_parameter_382: name = "parameter_382" + original_name = "layer_norm_20.b_0" shape = [384] dtype = "float32" min_val = float("-4.21981") @@ -3831,6 +4214,7 @@ class Program_weight_tensor_parameter_382: class Program_weight_tensor_parameter_383: name = "parameter_383" + original_name = "layer_norm_20.w_0" shape = [384] dtype = "float32" min_val = float("0.386956") @@ -3842,6 +4226,7 @@ class Program_weight_tensor_parameter_383: class Program_weight_tensor_parameter_384: name = "parameter_384" + original_name = "linear_35.b_0" shape = [384] dtype = "float32" min_val = float("-1.6503") @@ -3853,6 +4238,7 @@ class Program_weight_tensor_parameter_384: class Program_weight_tensor_parameter_385: name = "parameter_385" + original_name = "linear_35.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.665253") @@ -3864,6 +4250,7 @@ class Program_weight_tensor_parameter_385: class Program_weight_tensor_parameter_386: name = "parameter_386" + original_name = "linear_34.b_0" shape = [1152] dtype = "float32" min_val = float("-5.18755") @@ -3875,6 +4262,7 @@ class Program_weight_tensor_parameter_386: class Program_weight_tensor_parameter_387: name = "parameter_387" + original_name = "linear_34.w_0" shape = [384, 1152] dtype = "float32" min_val = float("-0.380591") @@ -3886,6 +4274,7 @@ class Program_weight_tensor_parameter_387: class Program_weight_tensor_parameter_388: name = "parameter_388" + original_name = "layer_norm_19.b_0" shape = [384] dtype = "float32" min_val = float("-3.50029") @@ -3897,6 +4286,7 @@ class Program_weight_tensor_parameter_388: class Program_weight_tensor_parameter_389: name = "parameter_389" + original_name = "layer_norm_19.w_0" shape = [384] dtype = "float32" min_val = float("0.12845") @@ -3908,6 +4298,7 @@ class Program_weight_tensor_parameter_389: class Program_weight_tensor_parameter_390: name = "parameter_390" + original_name = "linear_33.b_0" shape = [384] dtype = "float32" min_val = float("-1.36973") @@ -3919,6 +4310,7 @@ class Program_weight_tensor_parameter_390: class Program_weight_tensor_parameter_391: name = "parameter_391" + original_name = "linear_33.w_0" shape = [1536, 384] dtype = "float32" min_val = float("-0.733098") @@ -3930,6 +4322,7 @@ class Program_weight_tensor_parameter_391: class Program_weight_tensor_parameter_392: name = "parameter_392" + original_name = "linear_32.b_0" shape = [1536] dtype = "float32" min_val = float("-5.67193") @@ -3941,6 +4334,7 @@ class Program_weight_tensor_parameter_392: class Program_weight_tensor_parameter_393: name = "parameter_393" + original_name = "linear_32.w_0" shape = [384, 1536] dtype = "float32" min_val = float("-0.571878") @@ -3952,6 +4346,7 @@ class Program_weight_tensor_parameter_393: class Program_weight_tensor_parameter_394: name = "parameter_394" + original_name = "layer_norm_18.b_0" shape = [384] dtype = "float32" min_val = float("-4.46615") @@ -3963,6 +4358,7 @@ class Program_weight_tensor_parameter_394: class Program_weight_tensor_parameter_395: name = "parameter_395" + original_name = "layer_norm_18.w_0" shape = [384] dtype = "float32" min_val = float("-0.00102235") @@ -3974,6 +4370,7 @@ class Program_weight_tensor_parameter_395: class Program_weight_tensor_parameter_396: name = "parameter_396" + original_name = "linear_31.b_0" shape = [384] dtype = "float32" min_val = float("-0.896298") @@ -3985,6 +4382,7 @@ class Program_weight_tensor_parameter_396: class Program_weight_tensor_parameter_397: name = "parameter_397" + original_name = "linear_31.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.442041") @@ -3996,6 +4394,7 @@ class Program_weight_tensor_parameter_397: class Program_weight_tensor_parameter_398: name = "parameter_398" + original_name = "linear_30.b_0" shape = [1152] dtype = "float32" min_val = float("-3.21664") @@ -4007,6 +4406,7 @@ class Program_weight_tensor_parameter_398: class Program_weight_tensor_parameter_399: name = "parameter_399" + original_name = "linear_30.w_0" shape = [384, 1152] dtype = "float32" min_val = float("-0.539962") @@ -4018,6 +4418,7 @@ class Program_weight_tensor_parameter_399: class Program_weight_tensor_parameter_400: name = "parameter_400" + original_name = "layer_norm_17.b_0" shape = [384] dtype = "float32" min_val = float("-2.98689") @@ -4029,6 +4430,7 @@ class Program_weight_tensor_parameter_400: class Program_weight_tensor_parameter_401: name = "parameter_401" + original_name = "layer_norm_17.w_0" shape = [384] dtype = "float32" min_val = float("-0.00691647") @@ -4040,6 +4442,7 @@ class Program_weight_tensor_parameter_401: class Program_weight_tensor_parameter_402: name = "parameter_402" + original_name = "linear_29.b_0" shape = [384] dtype = "float32" min_val = float("-1.55968") @@ -4051,6 +4454,7 @@ class Program_weight_tensor_parameter_402: class Program_weight_tensor_parameter_403: name = "parameter_403" + original_name = "linear_29.w_0" shape = [1536, 384] dtype = "float32" min_val = float("-0.660884") @@ -4062,6 +4466,7 @@ class Program_weight_tensor_parameter_403: class Program_weight_tensor_parameter_404: name = "parameter_404" + original_name = "linear_28.b_0" shape = [1536] dtype = "float32" min_val = float("-5.42016") @@ -4073,6 +4478,7 @@ class Program_weight_tensor_parameter_404: class Program_weight_tensor_parameter_405: name = "parameter_405" + original_name = "linear_28.w_0" shape = [384, 1536] dtype = "float32" min_val = float("-0.543028") @@ -4084,6 +4490,7 @@ class Program_weight_tensor_parameter_405: class Program_weight_tensor_parameter_406: name = "parameter_406" + original_name = "layer_norm_16.b_0" shape = [384] dtype = "float32" min_val = float("-4.10652") @@ -4095,6 +4502,7 @@ class Program_weight_tensor_parameter_406: class Program_weight_tensor_parameter_407: name = "parameter_407" + original_name = "layer_norm_16.w_0" shape = [384] dtype = "float32" min_val = float("0.00332775") @@ -4106,6 +4514,7 @@ class Program_weight_tensor_parameter_407: class Program_weight_tensor_parameter_408: name = "parameter_408" + original_name = "linear_27.b_0" shape = [384] dtype = "float32" min_val = float("-0.925735") @@ -4117,6 +4526,7 @@ class Program_weight_tensor_parameter_408: class Program_weight_tensor_parameter_409: name = "parameter_409" + original_name = "linear_27.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.502551") @@ -4128,6 +4538,7 @@ class Program_weight_tensor_parameter_409: class Program_weight_tensor_parameter_410: name = "parameter_410" + original_name = "linear_26.b_0" shape = [1152] dtype = "float32" min_val = float("-3.70529") @@ -4139,6 +4550,7 @@ class Program_weight_tensor_parameter_410: class Program_weight_tensor_parameter_411: name = "parameter_411" + original_name = "linear_26.w_0" shape = [384, 1152] dtype = "float32" min_val = float("-0.459269") @@ -4150,6 +4562,7 @@ class Program_weight_tensor_parameter_411: class Program_weight_tensor_parameter_412: name = "parameter_412" + original_name = "layer_norm_15.b_0" shape = [384] dtype = "float32" min_val = float("-3.17884") @@ -4161,6 +4574,7 @@ class Program_weight_tensor_parameter_412: class Program_weight_tensor_parameter_413: name = "parameter_413" + original_name = "layer_norm_15.w_0" shape = [384] dtype = "float32" min_val = float("-0.00690475") @@ -4172,6 +4586,7 @@ class Program_weight_tensor_parameter_413: class Program_weight_tensor_parameter_414: name = "parameter_414" + original_name = "linear_25.b_0" shape = [384] dtype = "float32" min_val = float("-1.56459") @@ -4183,6 +4598,7 @@ class Program_weight_tensor_parameter_414: class Program_weight_tensor_parameter_415: name = "parameter_415" + original_name = "linear_25.w_0" shape = [1536, 384] dtype = "float32" min_val = float("-0.534191") @@ -4194,6 +4610,7 @@ class Program_weight_tensor_parameter_415: class Program_weight_tensor_parameter_416: name = "parameter_416" + original_name = "linear_24.b_0" shape = [1536] dtype = "float32" min_val = float("-5.471") @@ -4205,6 +4622,7 @@ class Program_weight_tensor_parameter_416: class Program_weight_tensor_parameter_417: name = "parameter_417" + original_name = "linear_24.w_0" shape = [384, 1536] dtype = "float32" min_val = float("-0.705037") @@ -4216,6 +4634,7 @@ class Program_weight_tensor_parameter_417: class Program_weight_tensor_parameter_418: name = "parameter_418" + original_name = "layer_norm_14.b_0" shape = [384] dtype = "float32" min_val = float("-2.77399") @@ -4227,6 +4646,7 @@ class Program_weight_tensor_parameter_418: class Program_weight_tensor_parameter_419: name = "parameter_419" + original_name = "layer_norm_14.w_0" shape = [384] dtype = "float32" min_val = float("-0.00153011") @@ -4238,6 +4658,7 @@ class Program_weight_tensor_parameter_419: class Program_weight_tensor_parameter_420: name = "parameter_420" + original_name = "linear_23.b_0" shape = [384] dtype = "float32" min_val = float("-0.882424") @@ -4249,6 +4670,7 @@ class Program_weight_tensor_parameter_420: class Program_weight_tensor_parameter_421: name = "parameter_421" + original_name = "linear_23.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.841758") @@ -4260,6 +4682,7 @@ class Program_weight_tensor_parameter_421: class Program_weight_tensor_parameter_422: name = "parameter_422" + original_name = "linear_22.b_0" shape = [1152] dtype = "float32" min_val = float("-3.42565") @@ -4271,6 +4694,7 @@ class Program_weight_tensor_parameter_422: class Program_weight_tensor_parameter_423: name = "parameter_423" + original_name = "linear_22.w_0" shape = [384, 1152] dtype = "float32" min_val = float("-0.464748") @@ -4282,6 +4706,7 @@ class Program_weight_tensor_parameter_423: class Program_weight_tensor_parameter_424: name = "parameter_424" + original_name = "layer_norm_13.b_0" shape = [384] dtype = "float32" min_val = float("-1.18072") @@ -4293,6 +4718,7 @@ class Program_weight_tensor_parameter_424: class Program_weight_tensor_parameter_425: name = "parameter_425" + original_name = "layer_norm_13.w_0" shape = [384] dtype = "float32" min_val = float("-0.011215") @@ -4304,6 +4730,7 @@ class Program_weight_tensor_parameter_425: class Program_weight_tensor_parameter_426: name = "parameter_426" + original_name = "linear_21.b_0" shape = [384] dtype = "float32" min_val = float("-2.05865") @@ -4315,6 +4742,7 @@ class Program_weight_tensor_parameter_426: class Program_weight_tensor_parameter_427: name = "parameter_427" + original_name = "linear_21.w_0" shape = [1536, 384] dtype = "float32" min_val = float("-0.782334") @@ -4326,6 +4754,7 @@ class Program_weight_tensor_parameter_427: class Program_weight_tensor_parameter_428: name = "parameter_428" + original_name = "linear_20.b_0" shape = [1536] dtype = "float32" min_val = float("-5.41117") @@ -4337,6 +4766,7 @@ class Program_weight_tensor_parameter_428: class Program_weight_tensor_parameter_429: name = "parameter_429" + original_name = "linear_20.w_0" shape = [384, 1536] dtype = "float32" min_val = float("-0.523359") @@ -4348,6 +4778,7 @@ class Program_weight_tensor_parameter_429: class Program_weight_tensor_parameter_430: name = "parameter_430" + original_name = "layer_norm_12.b_0" shape = [384] dtype = "float32" min_val = float("-2.33563") @@ -4359,6 +4790,7 @@ class Program_weight_tensor_parameter_430: class Program_weight_tensor_parameter_431: name = "parameter_431" + original_name = "layer_norm_12.w_0" shape = [384] dtype = "float32" min_val = float("-0.00924296") @@ -4370,6 +4802,7 @@ class Program_weight_tensor_parameter_431: class Program_weight_tensor_parameter_432: name = "parameter_432" + original_name = "linear_19.b_0" shape = [384] dtype = "float32" min_val = float("-1.76653") @@ -4381,6 +4814,7 @@ class Program_weight_tensor_parameter_432: class Program_weight_tensor_parameter_433: name = "parameter_433" + original_name = "linear_19.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.578849") @@ -4392,6 +4826,7 @@ class Program_weight_tensor_parameter_433: class Program_weight_tensor_parameter_434: name = "parameter_434" + original_name = "linear_18.b_0" shape = [1152] dtype = "float32" min_val = float("-4.34906") @@ -4403,6 +4838,7 @@ class Program_weight_tensor_parameter_434: class Program_weight_tensor_parameter_435: name = "parameter_435" + original_name = "linear_18.w_0" shape = [384, 1152] dtype = "float32" min_val = float("-0.434441") @@ -4414,6 +4850,7 @@ class Program_weight_tensor_parameter_435: class Program_weight_tensor_parameter_436: name = "parameter_436" + original_name = "layer_norm_11.b_0" shape = [384] dtype = "float32" min_val = float("-0.160743") @@ -4425,6 +4862,7 @@ class Program_weight_tensor_parameter_436: class Program_weight_tensor_parameter_437: name = "parameter_437" + original_name = "layer_norm_11.w_0" shape = [384] dtype = "float32" min_val = float("-0.00772838") @@ -4436,6 +4874,7 @@ class Program_weight_tensor_parameter_437: class Program_weight_tensor_parameter_438: name = "parameter_438" + original_name = "layer_norm_53.b_0" shape = [192] dtype = "float32" data = None @@ -4443,6 +4882,7 @@ class Program_weight_tensor_parameter_438: class Program_weight_tensor_parameter_439: name = "parameter_439" + original_name = "layer_norm_53.w_0" shape = [192] dtype = "float32" min_val = float("1.0") @@ -4453,6 +4893,7 @@ class Program_weight_tensor_parameter_439: class Program_weight_tensor_parameter_440: name = "parameter_440" + original_name = "linear_17.w_0" shape = [768, 384] dtype = "float32" min_val = float("-0.357094") @@ -4464,6 +4905,7 @@ class Program_weight_tensor_parameter_440: class Program_weight_tensor_parameter_441: name = "parameter_441" + original_name = "layer_norm_10.b_0" shape = [768] dtype = "float32" min_val = float("-1.35845") @@ -4475,6 +4917,7 @@ class Program_weight_tensor_parameter_441: class Program_weight_tensor_parameter_442: name = "parameter_442" + original_name = "layer_norm_10.w_0" shape = [768] dtype = "float32" min_val = float("0.250203") @@ -4486,6 +4929,7 @@ class Program_weight_tensor_parameter_442: class Program_weight_tensor_parameter_443: name = "parameter_443" + original_name = "linear_16.b_0" shape = [192] dtype = "float32" min_val = float("-5.65409") @@ -4497,6 +4941,7 @@ class Program_weight_tensor_parameter_443: class Program_weight_tensor_parameter_444: name = "parameter_444" + original_name = "linear_16.w_0" shape = [768, 192] dtype = "float32" min_val = float("-1.01845") @@ -4508,6 +4953,7 @@ class Program_weight_tensor_parameter_444: class Program_weight_tensor_parameter_445: name = "parameter_445" + original_name = "linear_15.b_0" shape = [768] dtype = "float32" min_val = float("-3.30376") @@ -4519,6 +4965,7 @@ class Program_weight_tensor_parameter_445: class Program_weight_tensor_parameter_446: name = "parameter_446" + original_name = "linear_15.w_0" shape = [192, 768] dtype = "float32" min_val = float("-0.486469") @@ -4530,6 +4977,7 @@ class Program_weight_tensor_parameter_446: class Program_weight_tensor_parameter_447: name = "parameter_447" + original_name = "layer_norm_9.b_0" shape = [192] dtype = "float32" min_val = float("-1.19854") @@ -4541,6 +4989,7 @@ class Program_weight_tensor_parameter_447: class Program_weight_tensor_parameter_448: name = "parameter_448" + original_name = "layer_norm_9.w_0" shape = [192] dtype = "float32" min_val = float("0.387802") @@ -4552,6 +5001,7 @@ class Program_weight_tensor_parameter_448: class Program_weight_tensor_parameter_449: name = "parameter_449" + original_name = "linear_14.b_0" shape = [192] dtype = "float32" min_val = float("-4.43349") @@ -4563,6 +5013,7 @@ class Program_weight_tensor_parameter_449: class Program_weight_tensor_parameter_450: name = "parameter_450" + original_name = "linear_14.w_0" shape = [192, 192] dtype = "float32" min_val = float("-0.277136") @@ -4574,6 +5025,7 @@ class Program_weight_tensor_parameter_450: class Program_weight_tensor_parameter_451: name = "parameter_451" + original_name = "linear_13.b_0" shape = [576] dtype = "float32" min_val = float("-3.70476") @@ -4585,6 +5037,7 @@ class Program_weight_tensor_parameter_451: class Program_weight_tensor_parameter_452: name = "parameter_452" + original_name = "linear_13.w_0" shape = [192, 576] dtype = "float32" min_val = float("-0.614933") @@ -4596,6 +5049,7 @@ class Program_weight_tensor_parameter_452: class Program_weight_tensor_parameter_453: name = "parameter_453" + original_name = "layer_norm_8.b_0" shape = [192] dtype = "float32" min_val = float("-0.708") @@ -4607,6 +5061,7 @@ class Program_weight_tensor_parameter_453: class Program_weight_tensor_parameter_454: name = "parameter_454" + original_name = "layer_norm_8.w_0" shape = [192] dtype = "float32" min_val = float("0.486079") @@ -4618,6 +5073,7 @@ class Program_weight_tensor_parameter_454: class Program_weight_tensor_parameter_455: name = "parameter_455" + original_name = "linear_12.b_0" shape = [192] dtype = "float32" min_val = float("-6.42398") @@ -4629,6 +5085,7 @@ class Program_weight_tensor_parameter_455: class Program_weight_tensor_parameter_456: name = "parameter_456" + original_name = "linear_12.w_0" shape = [768, 192] dtype = "float32" min_val = float("-0.562966") @@ -4640,6 +5097,7 @@ class Program_weight_tensor_parameter_456: class Program_weight_tensor_parameter_457: name = "parameter_457" + original_name = "linear_11.b_0" shape = [768] dtype = "float32" min_val = float("-2.88993") @@ -4651,6 +5109,7 @@ class Program_weight_tensor_parameter_457: class Program_weight_tensor_parameter_458: name = "parameter_458" + original_name = "linear_11.w_0" shape = [192, 768] dtype = "float32" min_val = float("-0.413941") @@ -4662,6 +5121,7 @@ class Program_weight_tensor_parameter_458: class Program_weight_tensor_parameter_459: name = "parameter_459" + original_name = "layer_norm_7.b_0" shape = [192] dtype = "float32" min_val = float("-0.556142") @@ -4673,6 +5133,7 @@ class Program_weight_tensor_parameter_459: class Program_weight_tensor_parameter_460: name = "parameter_460" + original_name = "layer_norm_7.w_0" shape = [192] dtype = "float32" min_val = float("-0.00685324") @@ -4684,6 +5145,7 @@ class Program_weight_tensor_parameter_460: class Program_weight_tensor_parameter_461: name = "parameter_461" + original_name = "linear_10.b_0" shape = [192] dtype = "float32" min_val = float("-5.87772") @@ -4695,6 +5157,7 @@ class Program_weight_tensor_parameter_461: class Program_weight_tensor_parameter_462: name = "parameter_462" + original_name = "linear_10.w_0" shape = [192, 192] dtype = "float32" min_val = float("-0.384763") @@ -4706,6 +5169,7 @@ class Program_weight_tensor_parameter_462: class Program_weight_tensor_parameter_463: name = "parameter_463" + original_name = "linear_9.b_0" shape = [576] dtype = "float32" min_val = float("-2.53186") @@ -4717,6 +5181,7 @@ class Program_weight_tensor_parameter_463: class Program_weight_tensor_parameter_464: name = "parameter_464" + original_name = "linear_9.w_0" shape = [192, 576] dtype = "float32" min_val = float("-0.540943") @@ -4728,6 +5193,7 @@ class Program_weight_tensor_parameter_464: class Program_weight_tensor_parameter_465: name = "parameter_465" + original_name = "layer_norm_6.b_0" shape = [192] dtype = "float32" min_val = float("-0.329145") @@ -4739,6 +5205,7 @@ class Program_weight_tensor_parameter_465: class Program_weight_tensor_parameter_466: name = "parameter_466" + original_name = "layer_norm_6.w_0" shape = [192] dtype = "float32" min_val = float("-0.00330349") @@ -4750,6 +5217,7 @@ class Program_weight_tensor_parameter_466: class Program_weight_tensor_parameter_467: name = "parameter_467" + original_name = "layer_norm_52.b_0" shape = [96] dtype = "float32" data = None @@ -4757,6 +5225,7 @@ class Program_weight_tensor_parameter_467: class Program_weight_tensor_parameter_468: name = "parameter_468" + original_name = "layer_norm_52.w_0" shape = [96] dtype = "float32" min_val = float("1.0") @@ -4767,6 +5236,7 @@ class Program_weight_tensor_parameter_468: class Program_weight_tensor_parameter_469: name = "parameter_469" + original_name = "linear_8.w_0" shape = [384, 192] dtype = "float32" min_val = float("-0.412246") @@ -4778,6 +5248,7 @@ class Program_weight_tensor_parameter_469: class Program_weight_tensor_parameter_470: name = "parameter_470" + original_name = "layer_norm_5.b_0" shape = [384] dtype = "float32" min_val = float("-1.76269") @@ -4789,6 +5260,7 @@ class Program_weight_tensor_parameter_470: class Program_weight_tensor_parameter_471: name = "parameter_471" + original_name = "layer_norm_5.w_0" shape = [384] dtype = "float32" min_val = float("0.259424") @@ -4800,6 +5272,7 @@ class Program_weight_tensor_parameter_471: class Program_weight_tensor_parameter_472: name = "parameter_472" + original_name = "linear_7.b_0" shape = [96] dtype = "float32" min_val = float("-1.4916") @@ -4811,6 +5284,7 @@ class Program_weight_tensor_parameter_472: class Program_weight_tensor_parameter_473: name = "parameter_473" + original_name = "linear_7.w_0" shape = [384, 96] dtype = "float32" min_val = float("-0.654892") @@ -4822,6 +5296,7 @@ class Program_weight_tensor_parameter_473: class Program_weight_tensor_parameter_474: name = "parameter_474" + original_name = "linear_6.b_0" shape = [384] dtype = "float32" min_val = float("-3.89129") @@ -4833,6 +5308,7 @@ class Program_weight_tensor_parameter_474: class Program_weight_tensor_parameter_475: name = "parameter_475" + original_name = "linear_6.w_0" shape = [96, 384] dtype = "float32" min_val = float("-0.396774") @@ -4844,6 +5320,7 @@ class Program_weight_tensor_parameter_475: class Program_weight_tensor_parameter_476: name = "parameter_476" + original_name = "layer_norm_4.b_0" shape = [96] dtype = "float32" min_val = float("-2.51497") @@ -4855,6 +5332,7 @@ class Program_weight_tensor_parameter_476: class Program_weight_tensor_parameter_477: name = "parameter_477" + original_name = "layer_norm_4.w_0" shape = [96] dtype = "float32" min_val = float("-0.722583") @@ -4866,6 +5344,7 @@ class Program_weight_tensor_parameter_477: class Program_weight_tensor_parameter_478: name = "parameter_478" + original_name = "linear_5.b_0" shape = [96] dtype = "float32" min_val = float("-1.2732") @@ -4877,6 +5356,7 @@ class Program_weight_tensor_parameter_478: class Program_weight_tensor_parameter_479: name = "parameter_479" + original_name = "linear_5.w_0" shape = [96, 96] dtype = "float32" min_val = float("-0.312308") @@ -4888,6 +5368,7 @@ class Program_weight_tensor_parameter_479: class Program_weight_tensor_parameter_480: name = "parameter_480" + original_name = "linear_4.b_0" shape = [288] dtype = "float32" min_val = float("-4.67732") @@ -4899,6 +5380,7 @@ class Program_weight_tensor_parameter_480: class Program_weight_tensor_parameter_481: name = "parameter_481" + original_name = "linear_4.w_0" shape = [96, 288] dtype = "float32" min_val = float("-0.483863") @@ -4910,6 +5392,7 @@ class Program_weight_tensor_parameter_481: class Program_weight_tensor_parameter_482: name = "parameter_482" + original_name = "layer_norm_3.b_0" shape = [96] dtype = "float32" min_val = float("-1.93978") @@ -4921,6 +5404,7 @@ class Program_weight_tensor_parameter_482: class Program_weight_tensor_parameter_483: name = "parameter_483" + original_name = "layer_norm_3.w_0" shape = [96] dtype = "float32" min_val = float("0.458121") @@ -4932,6 +5416,7 @@ class Program_weight_tensor_parameter_483: class Program_weight_tensor_parameter_484: name = "parameter_484" + original_name = "linear_3.b_0" shape = [96] dtype = "float32" min_val = float("-2.85777") @@ -4943,6 +5428,7 @@ class Program_weight_tensor_parameter_484: class Program_weight_tensor_parameter_485: name = "parameter_485" + original_name = "linear_3.w_0" shape = [384, 96] dtype = "float32" min_val = float("-0.663805") @@ -4954,6 +5440,7 @@ class Program_weight_tensor_parameter_485: class Program_weight_tensor_parameter_486: name = "parameter_486" + original_name = "linear_2.b_0" shape = [384] dtype = "float32" min_val = float("-2.99421") @@ -4965,6 +5452,7 @@ class Program_weight_tensor_parameter_486: class Program_weight_tensor_parameter_487: name = "parameter_487" + original_name = "linear_2.w_0" shape = [96, 384] dtype = "float32" min_val = float("-0.49021") @@ -4976,6 +5464,7 @@ class Program_weight_tensor_parameter_487: class Program_weight_tensor_parameter_488: name = "parameter_488" + original_name = "layer_norm_2.b_0" shape = [96] dtype = "float32" min_val = float("-2.18691") @@ -4987,6 +5476,7 @@ class Program_weight_tensor_parameter_488: class Program_weight_tensor_parameter_489: name = "parameter_489" + original_name = "layer_norm_2.w_0" shape = [96] dtype = "float32" min_val = float("-0.00102264") @@ -4998,6 +5488,7 @@ class Program_weight_tensor_parameter_489: class Program_weight_tensor_parameter_490: name = "parameter_490" + original_name = "linear_1.b_0" shape = [96] dtype = "float32" min_val = float("-1.9372") @@ -5009,6 +5500,7 @@ class Program_weight_tensor_parameter_490: class Program_weight_tensor_parameter_491: name = "parameter_491" + original_name = "linear_1.w_0" shape = [96, 96] dtype = "float32" min_val = float("-0.447579") @@ -5020,6 +5512,7 @@ class Program_weight_tensor_parameter_491: class Program_weight_tensor_parameter_492: name = "parameter_492" + original_name = "linear_0.b_0" shape = [288] dtype = "float32" min_val = float("-4.66158") @@ -5031,6 +5524,7 @@ class Program_weight_tensor_parameter_492: class Program_weight_tensor_parameter_493: name = "parameter_493" + original_name = "linear_0.w_0" shape = [96, 288] dtype = "float32" min_val = float("-0.356196") @@ -5042,6 +5536,7 @@ class Program_weight_tensor_parameter_493: class Program_weight_tensor_parameter_494: name = "parameter_494" + original_name = "layer_norm_1.b_0" shape = [96] dtype = "float32" min_val = float("-1.75712") @@ -5053,6 +5548,7 @@ class Program_weight_tensor_parameter_494: class Program_weight_tensor_parameter_495: name = "parameter_495" + original_name = "layer_norm_1.w_0" shape = [96] dtype = "float32" min_val = float("-0.00132859") @@ -5064,6 +5560,7 @@ class Program_weight_tensor_parameter_495: class Program_weight_tensor_parameter_496: name = "parameter_496" + original_name = "layer_norm_0.b_0" shape = [96] dtype = "float32" min_val = float("-3.28992") @@ -5075,6 +5572,7 @@ class Program_weight_tensor_parameter_496: class Program_weight_tensor_parameter_497: name = "parameter_497" + original_name = "layer_norm_0.w_0" shape = [96] dtype = "float32" min_val = float("-0.0888211") @@ -5086,6 +5584,7 @@ class Program_weight_tensor_parameter_497: class Program_weight_tensor_parameter_498: name = "parameter_498" + original_name = "conv2d_0.b_0" shape = [96] dtype = "float32" min_val = float("-1.44354") @@ -5097,6 +5596,7 @@ class Program_weight_tensor_parameter_498: class Program_weight_tensor_parameter_499: name = "parameter_499" + original_name = "conv2d_0.w_0" shape = [96, 3, 4, 4] dtype = "float32" min_val = float("-0.283527") diff --git a/paddle_samples/PaddleX/MaskFormer_tiny/subgraph_0/input_meta.py b/paddle_samples/PaddleX/MaskFormer_tiny/subgraph_0/input_meta.py index 10eb67e14..a4a1a3063 100644 --- a/paddle_samples/PaddleX/MaskFormer_tiny/subgraph_0/input_meta.py +++ b/paddle_samples/PaddleX/MaskFormer_tiny/subgraph_0/input_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_data_0: name = "data_0" + original_name = "param_0" shape = [169, 3] dtype = "float32" min_val = float("-13.6244") @@ -11,6 +12,7 @@ class Program_weight_tensor_data_0: class Program_weight_tensor_data_1: name = "data_1" + original_name = "param_1" shape = [169, 3] dtype = "float32" min_val = float("-10.1725") @@ -22,6 +24,7 @@ class Program_weight_tensor_data_1: class Program_weight_tensor_data_2: name = "data_2" + original_name = "param_10" shape = [169, 24] dtype = "float32" min_val = float("-22.8913") @@ -33,6 +36,7 @@ class Program_weight_tensor_data_2: class Program_weight_tensor_data_3: name = "data_3" + original_name = "param_11" shape = [169, 24] dtype = "float32" min_val = float("-26.7541") @@ -44,6 +48,7 @@ class Program_weight_tensor_data_3: class Program_weight_tensor_data_4: name = "data_4" + original_name = "param_12" shape = [100, 256] dtype = "float32" min_val = float("-3.74592") @@ -55,6 +60,7 @@ class Program_weight_tensor_data_4: class Program_weight_tensor_data_5: name = "data_5" + original_name = "param_2" shape = [169, 6] dtype = "float32" min_val = float("-10.4008") @@ -66,6 +72,7 @@ class Program_weight_tensor_data_5: class Program_weight_tensor_data_6: name = "data_6" + original_name = "param_3" shape = [169, 6] dtype = "float32" min_val = float("-11.2985") @@ -77,6 +84,7 @@ class Program_weight_tensor_data_6: class Program_weight_tensor_data_7: name = "data_7" + original_name = "param_4" shape = [169, 12] dtype = "float32" min_val = float("-13.0402") @@ -88,6 +96,7 @@ class Program_weight_tensor_data_7: class Program_weight_tensor_data_8: name = "data_8" + original_name = "param_5" shape = [169, 12] dtype = "float32" min_val = float("-15.2582") @@ -99,6 +108,7 @@ class Program_weight_tensor_data_8: class Program_weight_tensor_data_9: name = "data_9" + original_name = "param_6" shape = [169, 12] dtype = "float32" min_val = float("-11.9454") @@ -110,6 +120,7 @@ class Program_weight_tensor_data_9: class Program_weight_tensor_data_10: name = "data_10" + original_name = "param_7" shape = [169, 12] dtype = "float32" min_val = float("-13.3033") @@ -121,6 +132,7 @@ class Program_weight_tensor_data_10: class Program_weight_tensor_data_11: name = "data_11" + original_name = "param_8" shape = [169, 12] dtype = "float32" min_val = float("-11.7905") @@ -132,6 +144,7 @@ class Program_weight_tensor_data_11: class Program_weight_tensor_data_12: name = "data_12" + original_name = "param_9" shape = [169, 12] dtype = "float32" min_val = float("-12.1116") @@ -143,6 +156,7 @@ class Program_weight_tensor_data_12: class Program_weight_tensor_data_13: name = "data_13" + original_name = "var_0" shape = [2, 3, 512, 512] dtype = "float32" min_val = float("-2.1179") @@ -154,6 +168,7 @@ class Program_weight_tensor_data_13: class Program_weight_tensor_data_14: name = "data_14" + original_name = "var_197" shape = [49, 49] dtype = "int64" min_val = 0 @@ -163,6 +178,7 @@ class Program_weight_tensor_data_14: class Program_weight_tensor_data_15: name = "data_15" + original_name = "var_258" shape = [49, 49] dtype = "int64" min_val = 0 @@ -172,6 +188,7 @@ class Program_weight_tensor_data_15: class Program_weight_tensor_data_16: name = "data_16" + original_name = "var_358" shape = [49, 49] dtype = "int64" min_val = 0 @@ -181,6 +198,7 @@ class Program_weight_tensor_data_16: class Program_weight_tensor_data_17: name = "data_17" + original_name = "var_419" shape = [49, 49] dtype = "int64" min_val = 0 @@ -190,6 +208,7 @@ class Program_weight_tensor_data_17: class Program_weight_tensor_data_18: name = "data_18" + original_name = "var_485" shape = [49, 49] dtype = "int64" min_val = 0 @@ -199,6 +218,7 @@ class Program_weight_tensor_data_18: class Program_weight_tensor_data_19: name = "data_19" + original_name = "var_50" shape = [49, 49] dtype = "int64" min_val = 0 @@ -208,6 +228,7 @@ class Program_weight_tensor_data_19: class Program_weight_tensor_data_20: name = "data_20" + original_name = "var_546" shape = [49, 49] dtype = "int64" min_val = 0 @@ -217,6 +238,7 @@ class Program_weight_tensor_data_20: class Program_weight_tensor_data_21: name = "data_21" + original_name = "var_612" shape = [49, 49] dtype = "int64" min_val = 0 @@ -226,6 +248,7 @@ class Program_weight_tensor_data_21: class Program_weight_tensor_data_22: name = "data_22" + original_name = "var_673" shape = [49, 49] dtype = "int64" min_val = 0 @@ -235,6 +258,7 @@ class Program_weight_tensor_data_22: class Program_weight_tensor_data_23: name = "data_23" + original_name = "var_773" shape = [49, 49] dtype = "int64" min_val = 0 @@ -244,6 +268,7 @@ class Program_weight_tensor_data_23: class Program_weight_tensor_data_24: name = "data_24" + original_name = "var_834" shape = [49, 49] dtype = "int64" min_val = 0 @@ -253,6 +278,7 @@ class Program_weight_tensor_data_24: class Program_weight_tensor_data_25: name = "data_25" + original_name = "var_97" shape = [49, 49] dtype = "int64" min_val = 0 diff --git a/paddle_samples/PaddleX/MaskFormer_tiny/subgraph_0/weight_meta.py b/paddle_samples/PaddleX/MaskFormer_tiny/subgraph_0/weight_meta.py index 578ba45c4..e0fe4818f 100644 --- a/paddle_samples/PaddleX/MaskFormer_tiny/subgraph_0/weight_meta.py +++ b/paddle_samples/PaddleX/MaskFormer_tiny/subgraph_0/weight_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_parameter_0: name = "parameter_0" + original_name = "linear_70.b_0" shape = [256] dtype = "float32" min_val = float("-0.0619741") @@ -11,6 +12,7 @@ class Program_weight_tensor_parameter_0: class Program_weight_tensor_parameter_1: name = "parameter_1" + original_name = "linear_70.w_0" shape = [256, 256] dtype = "float32" min_val = float("-0.0624993") @@ -22,6 +24,7 @@ class Program_weight_tensor_parameter_1: class Program_weight_tensor_parameter_2: name = "parameter_2" + original_name = "linear_69.b_0" shape = [256] dtype = "float32" min_val = float("-0.061831") @@ -33,6 +36,7 @@ class Program_weight_tensor_parameter_2: class Program_weight_tensor_parameter_3: name = "parameter_3" + original_name = "linear_69.w_0" shape = [256, 256] dtype = "float32" min_val = float("-0.0624973") @@ -44,6 +48,7 @@ class Program_weight_tensor_parameter_3: class Program_weight_tensor_parameter_4: name = "parameter_4" + original_name = "linear_68.b_0" shape = [256] dtype = "float32" min_val = float("-0.0610588") @@ -55,6 +60,7 @@ class Program_weight_tensor_parameter_4: class Program_weight_tensor_parameter_5: name = "parameter_5" + original_name = "linear_68.w_0" shape = [256, 256] dtype = "float32" min_val = float("-0.0625") @@ -66,6 +72,7 @@ class Program_weight_tensor_parameter_5: class Program_weight_tensor_parameter_6: name = "parameter_6" + original_name = "linear_67.b_0" shape = [3] dtype = "float32" min_val = float("0") @@ -75,6 +82,7 @@ class Program_weight_tensor_parameter_6: class Program_weight_tensor_parameter_7: name = "parameter_7" + original_name = "linear_67.w_0" shape = [256, 3] dtype = "float32" min_val = float("-0.0624929") @@ -86,6 +94,7 @@ class Program_weight_tensor_parameter_7: class Program_weight_tensor_parameter_8: name = "parameter_8" + original_name = "layer_norm_36.b_0_deepcopy_155" shape = [256] dtype = "float32" data = None @@ -93,6 +102,7 @@ class Program_weight_tensor_parameter_8: class Program_weight_tensor_parameter_9: name = "parameter_9" + original_name = "layer_norm_36.w_0_deepcopy_154" shape = [256] dtype = "float32" min_val = float("1.0") @@ -103,6 +113,7 @@ class Program_weight_tensor_parameter_9: class Program_weight_tensor_parameter_10: name = "parameter_10" + original_name = "linear_66.b_0_deepcopy_149" shape = [256] dtype = "float32" data = None @@ -110,6 +121,7 @@ class Program_weight_tensor_parameter_10: class Program_weight_tensor_parameter_11: name = "parameter_11" + original_name = "linear_66.w_0_deepcopy_148" shape = [2048, 256] dtype = "float32" min_val = float("-0.0510303") @@ -121,6 +133,7 @@ class Program_weight_tensor_parameter_11: class Program_weight_tensor_parameter_12: name = "parameter_12" + original_name = "linear_65.b_0_deepcopy_147" shape = [2048] dtype = "float32" data = None @@ -128,6 +141,7 @@ class Program_weight_tensor_parameter_12: class Program_weight_tensor_parameter_13: name = "parameter_13" + original_name = "linear_65.w_0_deepcopy_146" shape = [256, 2048] dtype = "float32" min_val = float("-0.0510308") @@ -139,6 +153,7 @@ class Program_weight_tensor_parameter_13: class Program_weight_tensor_parameter_14: name = "parameter_14" + original_name = "layer_norm_35.b_0_deepcopy_153" shape = [256] dtype = "float32" data = None @@ -146,6 +161,7 @@ class Program_weight_tensor_parameter_14: class Program_weight_tensor_parameter_15: name = "parameter_15" + original_name = "layer_norm_35.w_0_deepcopy_152" shape = [256] dtype = "float32" min_val = float("1.0") @@ -156,6 +172,7 @@ class Program_weight_tensor_parameter_15: class Program_weight_tensor_parameter_16: name = "parameter_16" + original_name = "linear_64.b_0_deepcopy_145" shape = [256] dtype = "float32" data = None @@ -163,6 +180,7 @@ class Program_weight_tensor_parameter_16: class Program_weight_tensor_parameter_17: name = "parameter_17" + original_name = "linear_64.w_0_deepcopy_144" shape = [256, 256] dtype = "float32" min_val = float("-0.10825") @@ -174,6 +192,7 @@ class Program_weight_tensor_parameter_17: class Program_weight_tensor_parameter_18: name = "parameter_18" + original_name = "linear_63.b_0_deepcopy_143" shape = [256] dtype = "float32" data = None @@ -181,6 +200,7 @@ class Program_weight_tensor_parameter_18: class Program_weight_tensor_parameter_19: name = "parameter_19" + original_name = "linear_63.w_0_deepcopy_142" shape = [256, 256] dtype = "float32" min_val = float("-0.108253") @@ -192,6 +212,7 @@ class Program_weight_tensor_parameter_19: class Program_weight_tensor_parameter_20: name = "parameter_20" + original_name = "linear_62.b_0_deepcopy_141" shape = [256] dtype = "float32" data = None @@ -199,6 +220,7 @@ class Program_weight_tensor_parameter_20: class Program_weight_tensor_parameter_21: name = "parameter_21" + original_name = "linear_62.w_0_deepcopy_140" shape = [256, 256] dtype = "float32" min_val = float("-0.108251") @@ -210,6 +232,7 @@ class Program_weight_tensor_parameter_21: class Program_weight_tensor_parameter_22: name = "parameter_22" + original_name = "linear_61.b_0_deepcopy_139" shape = [256] dtype = "float32" data = None @@ -217,6 +240,7 @@ class Program_weight_tensor_parameter_22: class Program_weight_tensor_parameter_23: name = "parameter_23" + original_name = "linear_61.w_0_deepcopy_138" shape = [256, 256] dtype = "float32" min_val = float("-0.108253") @@ -228,6 +252,7 @@ class Program_weight_tensor_parameter_23: class Program_weight_tensor_parameter_24: name = "parameter_24" + original_name = "layer_norm_34.b_0_deepcopy_151" shape = [256] dtype = "float32" data = None @@ -235,6 +260,7 @@ class Program_weight_tensor_parameter_24: class Program_weight_tensor_parameter_25: name = "parameter_25" + original_name = "layer_norm_34.w_0_deepcopy_150" shape = [256] dtype = "float32" min_val = float("1.0") @@ -245,6 +271,7 @@ class Program_weight_tensor_parameter_25: class Program_weight_tensor_parameter_26: name = "parameter_26" + original_name = "linear_60.b_0_deepcopy_137" shape = [256] dtype = "float32" data = None @@ -252,6 +279,7 @@ class Program_weight_tensor_parameter_26: class Program_weight_tensor_parameter_27: name = "parameter_27" + original_name = "linear_60.w_0_deepcopy_136" shape = [256, 256] dtype = "float32" min_val = float("-0.108243") @@ -263,6 +291,7 @@ class Program_weight_tensor_parameter_27: class Program_weight_tensor_parameter_28: name = "parameter_28" + original_name = "linear_59.b_0_deepcopy_135" shape = [256] dtype = "float32" data = None @@ -270,6 +299,7 @@ class Program_weight_tensor_parameter_28: class Program_weight_tensor_parameter_29: name = "parameter_29" + original_name = "linear_59.w_0_deepcopy_134" shape = [256, 256] dtype = "float32" min_val = float("-0.108252") @@ -281,6 +311,7 @@ class Program_weight_tensor_parameter_29: class Program_weight_tensor_parameter_30: name = "parameter_30" + original_name = "linear_58.b_0_deepcopy_133" shape = [256] dtype = "float32" data = None @@ -288,6 +319,7 @@ class Program_weight_tensor_parameter_30: class Program_weight_tensor_parameter_31: name = "parameter_31" + original_name = "linear_58.w_0_deepcopy_132" shape = [256, 256] dtype = "float32" min_val = float("-0.108252") @@ -299,6 +331,7 @@ class Program_weight_tensor_parameter_31: class Program_weight_tensor_parameter_32: name = "parameter_32" + original_name = "linear_57.b_0_deepcopy_131" shape = [256] dtype = "float32" data = None @@ -306,6 +339,7 @@ class Program_weight_tensor_parameter_32: class Program_weight_tensor_parameter_33: name = "parameter_33" + original_name = "linear_57.w_0_deepcopy_130" shape = [256, 256] dtype = "float32" min_val = float("-0.10825") @@ -317,6 +351,7 @@ class Program_weight_tensor_parameter_33: class Program_weight_tensor_parameter_34: name = "parameter_34" + original_name = "layer_norm_36.b_0_deepcopy_129" shape = [256] dtype = "float32" data = None @@ -324,6 +359,7 @@ class Program_weight_tensor_parameter_34: class Program_weight_tensor_parameter_35: name = "parameter_35" + original_name = "layer_norm_36.w_0_deepcopy_128" shape = [256] dtype = "float32" min_val = float("1.0") @@ -334,6 +370,7 @@ class Program_weight_tensor_parameter_35: class Program_weight_tensor_parameter_36: name = "parameter_36" + original_name = "linear_66.b_0_deepcopy_123" shape = [256] dtype = "float32" data = None @@ -341,6 +378,7 @@ class Program_weight_tensor_parameter_36: class Program_weight_tensor_parameter_37: name = "parameter_37" + original_name = "linear_66.w_0_deepcopy_122" shape = [2048, 256] dtype = "float32" min_val = float("-0.0510309") @@ -352,6 +390,7 @@ class Program_weight_tensor_parameter_37: class Program_weight_tensor_parameter_38: name = "parameter_38" + original_name = "linear_65.b_0_deepcopy_121" shape = [2048] dtype = "float32" data = None @@ -359,6 +398,7 @@ class Program_weight_tensor_parameter_38: class Program_weight_tensor_parameter_39: name = "parameter_39" + original_name = "linear_65.w_0_deepcopy_120" shape = [256, 2048] dtype = "float32" min_val = float("-0.051031") @@ -370,6 +410,7 @@ class Program_weight_tensor_parameter_39: class Program_weight_tensor_parameter_40: name = "parameter_40" + original_name = "layer_norm_35.b_0_deepcopy_127" shape = [256] dtype = "float32" data = None @@ -377,6 +418,7 @@ class Program_weight_tensor_parameter_40: class Program_weight_tensor_parameter_41: name = "parameter_41" + original_name = "layer_norm_35.w_0_deepcopy_126" shape = [256] dtype = "float32" min_val = float("1.0") @@ -387,6 +429,7 @@ class Program_weight_tensor_parameter_41: class Program_weight_tensor_parameter_42: name = "parameter_42" + original_name = "linear_64.b_0_deepcopy_119" shape = [256] dtype = "float32" data = None @@ -394,6 +437,7 @@ class Program_weight_tensor_parameter_42: class Program_weight_tensor_parameter_43: name = "parameter_43" + original_name = "linear_64.w_0_deepcopy_118" shape = [256, 256] dtype = "float32" min_val = float("-0.10825") @@ -405,6 +449,7 @@ class Program_weight_tensor_parameter_43: class Program_weight_tensor_parameter_44: name = "parameter_44" + original_name = "linear_63.b_0_deepcopy_117" shape = [256] dtype = "float32" data = None @@ -412,6 +457,7 @@ class Program_weight_tensor_parameter_44: class Program_weight_tensor_parameter_45: name = "parameter_45" + original_name = "linear_63.w_0_deepcopy_116" shape = [256, 256] dtype = "float32" min_val = float("-0.108253") @@ -423,6 +469,7 @@ class Program_weight_tensor_parameter_45: class Program_weight_tensor_parameter_46: name = "parameter_46" + original_name = "linear_62.b_0_deepcopy_115" shape = [256] dtype = "float32" data = None @@ -430,6 +477,7 @@ class Program_weight_tensor_parameter_46: class Program_weight_tensor_parameter_47: name = "parameter_47" + original_name = "linear_62.w_0_deepcopy_114" shape = [256, 256] dtype = "float32" min_val = float("-0.108251") @@ -441,6 +489,7 @@ class Program_weight_tensor_parameter_47: class Program_weight_tensor_parameter_48: name = "parameter_48" + original_name = "linear_61.b_0_deepcopy_113" shape = [256] dtype = "float32" data = None @@ -448,6 +497,7 @@ class Program_weight_tensor_parameter_48: class Program_weight_tensor_parameter_49: name = "parameter_49" + original_name = "linear_61.w_0_deepcopy_112" shape = [256, 256] dtype = "float32" min_val = float("-0.108253") @@ -459,6 +509,7 @@ class Program_weight_tensor_parameter_49: class Program_weight_tensor_parameter_50: name = "parameter_50" + original_name = "layer_norm_34.b_0_deepcopy_125" shape = [256] dtype = "float32" data = None @@ -466,6 +517,7 @@ class Program_weight_tensor_parameter_50: class Program_weight_tensor_parameter_51: name = "parameter_51" + original_name = "layer_norm_34.w_0_deepcopy_124" shape = [256] dtype = "float32" min_val = float("1.0") @@ -476,6 +528,7 @@ class Program_weight_tensor_parameter_51: class Program_weight_tensor_parameter_52: name = "parameter_52" + original_name = "linear_60.b_0_deepcopy_111" shape = [256] dtype = "float32" data = None @@ -483,6 +536,7 @@ class Program_weight_tensor_parameter_52: class Program_weight_tensor_parameter_53: name = "parameter_53" + original_name = "linear_60.w_0_deepcopy_110" shape = [256, 256] dtype = "float32" min_val = float("-0.108243") @@ -494,6 +548,7 @@ class Program_weight_tensor_parameter_53: class Program_weight_tensor_parameter_54: name = "parameter_54" + original_name = "linear_59.b_0_deepcopy_109" shape = [256] dtype = "float32" data = None @@ -501,6 +556,7 @@ class Program_weight_tensor_parameter_54: class Program_weight_tensor_parameter_55: name = "parameter_55" + original_name = "linear_59.w_0_deepcopy_108" shape = [256, 256] dtype = "float32" min_val = float("-0.108252") @@ -512,6 +568,7 @@ class Program_weight_tensor_parameter_55: class Program_weight_tensor_parameter_56: name = "parameter_56" + original_name = "linear_58.b_0_deepcopy_107" shape = [256] dtype = "float32" data = None @@ -519,6 +576,7 @@ class Program_weight_tensor_parameter_56: class Program_weight_tensor_parameter_57: name = "parameter_57" + original_name = "linear_58.w_0_deepcopy_106" shape = [256, 256] dtype = "float32" min_val = float("-0.108252") @@ -530,6 +588,7 @@ class Program_weight_tensor_parameter_57: class Program_weight_tensor_parameter_58: name = "parameter_58" + original_name = "linear_57.b_0_deepcopy_105" shape = [256] dtype = "float32" data = None @@ -537,6 +596,7 @@ class Program_weight_tensor_parameter_58: class Program_weight_tensor_parameter_59: name = "parameter_59" + original_name = "linear_57.w_0_deepcopy_104" shape = [256, 256] dtype = "float32" min_val = float("-0.10825") @@ -548,6 +608,7 @@ class Program_weight_tensor_parameter_59: class Program_weight_tensor_parameter_60: name = "parameter_60" + original_name = "layer_norm_36.b_0_deepcopy_103" shape = [256] dtype = "float32" data = None @@ -555,6 +616,7 @@ class Program_weight_tensor_parameter_60: class Program_weight_tensor_parameter_61: name = "parameter_61" + original_name = "layer_norm_36.w_0_deepcopy_102" shape = [256] dtype = "float32" min_val = float("1.0") @@ -565,6 +627,7 @@ class Program_weight_tensor_parameter_61: class Program_weight_tensor_parameter_62: name = "parameter_62" + original_name = "linear_66.b_0_deepcopy_97" shape = [256] dtype = "float32" data = None @@ -572,6 +635,7 @@ class Program_weight_tensor_parameter_62: class Program_weight_tensor_parameter_63: name = "parameter_63" + original_name = "linear_66.w_0_deepcopy_96" shape = [2048, 256] dtype = "float32" min_val = float("-0.0510307") @@ -583,6 +647,7 @@ class Program_weight_tensor_parameter_63: class Program_weight_tensor_parameter_64: name = "parameter_64" + original_name = "linear_65.b_0_deepcopy_95" shape = [2048] dtype = "float32" data = None @@ -590,6 +655,7 @@ class Program_weight_tensor_parameter_64: class Program_weight_tensor_parameter_65: name = "parameter_65" + original_name = "linear_65.w_0_deepcopy_94" shape = [256, 2048] dtype = "float32" min_val = float("-0.0510306") @@ -601,6 +667,7 @@ class Program_weight_tensor_parameter_65: class Program_weight_tensor_parameter_66: name = "parameter_66" + original_name = "layer_norm_35.b_0_deepcopy_101" shape = [256] dtype = "float32" data = None @@ -608,6 +675,7 @@ class Program_weight_tensor_parameter_66: class Program_weight_tensor_parameter_67: name = "parameter_67" + original_name = "layer_norm_35.w_0_deepcopy_100" shape = [256] dtype = "float32" min_val = float("1.0") @@ -618,6 +686,7 @@ class Program_weight_tensor_parameter_67: class Program_weight_tensor_parameter_68: name = "parameter_68" + original_name = "linear_64.b_0_deepcopy_93" shape = [256] dtype = "float32" data = None @@ -625,6 +694,7 @@ class Program_weight_tensor_parameter_68: class Program_weight_tensor_parameter_69: name = "parameter_69" + original_name = "linear_64.w_0_deepcopy_92" shape = [256, 256] dtype = "float32" min_val = float("-0.10825") @@ -636,6 +706,7 @@ class Program_weight_tensor_parameter_69: class Program_weight_tensor_parameter_70: name = "parameter_70" + original_name = "linear_63.b_0_deepcopy_91" shape = [256] dtype = "float32" data = None @@ -643,6 +714,7 @@ class Program_weight_tensor_parameter_70: class Program_weight_tensor_parameter_71: name = "parameter_71" + original_name = "linear_63.w_0_deepcopy_90" shape = [256, 256] dtype = "float32" min_val = float("-0.108253") @@ -654,6 +726,7 @@ class Program_weight_tensor_parameter_71: class Program_weight_tensor_parameter_72: name = "parameter_72" + original_name = "linear_62.b_0_deepcopy_89" shape = [256] dtype = "float32" data = None @@ -661,6 +734,7 @@ class Program_weight_tensor_parameter_72: class Program_weight_tensor_parameter_73: name = "parameter_73" + original_name = "linear_62.w_0_deepcopy_88" shape = [256, 256] dtype = "float32" min_val = float("-0.108251") @@ -672,6 +746,7 @@ class Program_weight_tensor_parameter_73: class Program_weight_tensor_parameter_74: name = "parameter_74" + original_name = "linear_61.b_0_deepcopy_87" shape = [256] dtype = "float32" data = None @@ -679,6 +754,7 @@ class Program_weight_tensor_parameter_74: class Program_weight_tensor_parameter_75: name = "parameter_75" + original_name = "linear_61.w_0_deepcopy_86" shape = [256, 256] dtype = "float32" min_val = float("-0.108253") @@ -690,6 +766,7 @@ class Program_weight_tensor_parameter_75: class Program_weight_tensor_parameter_76: name = "parameter_76" + original_name = "layer_norm_34.b_0_deepcopy_99" shape = [256] dtype = "float32" data = None @@ -697,6 +774,7 @@ class Program_weight_tensor_parameter_76: class Program_weight_tensor_parameter_77: name = "parameter_77" + original_name = "layer_norm_34.w_0_deepcopy_98" shape = [256] dtype = "float32" min_val = float("1.0") @@ -707,6 +785,7 @@ class Program_weight_tensor_parameter_77: class Program_weight_tensor_parameter_78: name = "parameter_78" + original_name = "linear_60.b_0_deepcopy_85" shape = [256] dtype = "float32" data = None @@ -714,6 +793,7 @@ class Program_weight_tensor_parameter_78: class Program_weight_tensor_parameter_79: name = "parameter_79" + original_name = "linear_60.w_0_deepcopy_84" shape = [256, 256] dtype = "float32" min_val = float("-0.108243") @@ -725,6 +805,7 @@ class Program_weight_tensor_parameter_79: class Program_weight_tensor_parameter_80: name = "parameter_80" + original_name = "linear_59.b_0_deepcopy_83" shape = [256] dtype = "float32" data = None @@ -732,6 +813,7 @@ class Program_weight_tensor_parameter_80: class Program_weight_tensor_parameter_81: name = "parameter_81" + original_name = "linear_59.w_0_deepcopy_82" shape = [256, 256] dtype = "float32" min_val = float("-0.108252") @@ -743,6 +825,7 @@ class Program_weight_tensor_parameter_81: class Program_weight_tensor_parameter_82: name = "parameter_82" + original_name = "linear_58.b_0_deepcopy_81" shape = [256] dtype = "float32" data = None @@ -750,6 +833,7 @@ class Program_weight_tensor_parameter_82: class Program_weight_tensor_parameter_83: name = "parameter_83" + original_name = "linear_58.w_0_deepcopy_80" shape = [256, 256] dtype = "float32" min_val = float("-0.108252") @@ -761,6 +845,7 @@ class Program_weight_tensor_parameter_83: class Program_weight_tensor_parameter_84: name = "parameter_84" + original_name = "linear_57.b_0_deepcopy_79" shape = [256] dtype = "float32" data = None @@ -768,6 +853,7 @@ class Program_weight_tensor_parameter_84: class Program_weight_tensor_parameter_85: name = "parameter_85" + original_name = "linear_57.w_0_deepcopy_78" shape = [256, 256] dtype = "float32" min_val = float("-0.10825") @@ -779,6 +865,7 @@ class Program_weight_tensor_parameter_85: class Program_weight_tensor_parameter_86: name = "parameter_86" + original_name = "layer_norm_36.b_0_deepcopy_77" shape = [256] dtype = "float32" data = None @@ -786,6 +873,7 @@ class Program_weight_tensor_parameter_86: class Program_weight_tensor_parameter_87: name = "parameter_87" + original_name = "layer_norm_36.w_0_deepcopy_76" shape = [256] dtype = "float32" min_val = float("1.0") @@ -796,6 +884,7 @@ class Program_weight_tensor_parameter_87: class Program_weight_tensor_parameter_88: name = "parameter_88" + original_name = "linear_66.b_0_deepcopy_71" shape = [256] dtype = "float32" data = None @@ -803,6 +892,7 @@ class Program_weight_tensor_parameter_88: class Program_weight_tensor_parameter_89: name = "parameter_89" + original_name = "linear_66.w_0_deepcopy_70" shape = [2048, 256] dtype = "float32" min_val = float("-0.051031") @@ -814,6 +904,7 @@ class Program_weight_tensor_parameter_89: class Program_weight_tensor_parameter_90: name = "parameter_90" + original_name = "linear_65.b_0_deepcopy_69" shape = [2048] dtype = "float32" data = None @@ -821,6 +912,7 @@ class Program_weight_tensor_parameter_90: class Program_weight_tensor_parameter_91: name = "parameter_91" + original_name = "linear_65.w_0_deepcopy_68" shape = [256, 2048] dtype = "float32" min_val = float("-0.0510308") @@ -832,6 +924,7 @@ class Program_weight_tensor_parameter_91: class Program_weight_tensor_parameter_92: name = "parameter_92" + original_name = "layer_norm_35.b_0_deepcopy_75" shape = [256] dtype = "float32" data = None @@ -839,6 +932,7 @@ class Program_weight_tensor_parameter_92: class Program_weight_tensor_parameter_93: name = "parameter_93" + original_name = "layer_norm_35.w_0_deepcopy_74" shape = [256] dtype = "float32" min_val = float("1.0") @@ -849,6 +943,7 @@ class Program_weight_tensor_parameter_93: class Program_weight_tensor_parameter_94: name = "parameter_94" + original_name = "linear_64.b_0_deepcopy_67" shape = [256] dtype = "float32" data = None @@ -856,6 +951,7 @@ class Program_weight_tensor_parameter_94: class Program_weight_tensor_parameter_95: name = "parameter_95" + original_name = "linear_64.w_0_deepcopy_66" shape = [256, 256] dtype = "float32" min_val = float("-0.10825") @@ -867,6 +963,7 @@ class Program_weight_tensor_parameter_95: class Program_weight_tensor_parameter_96: name = "parameter_96" + original_name = "linear_63.b_0_deepcopy_65" shape = [256] dtype = "float32" data = None @@ -874,6 +971,7 @@ class Program_weight_tensor_parameter_96: class Program_weight_tensor_parameter_97: name = "parameter_97" + original_name = "linear_63.w_0_deepcopy_64" shape = [256, 256] dtype = "float32" min_val = float("-0.108253") @@ -885,6 +983,7 @@ class Program_weight_tensor_parameter_97: class Program_weight_tensor_parameter_98: name = "parameter_98" + original_name = "linear_62.b_0_deepcopy_63" shape = [256] dtype = "float32" data = None @@ -892,6 +991,7 @@ class Program_weight_tensor_parameter_98: class Program_weight_tensor_parameter_99: name = "parameter_99" + original_name = "linear_62.w_0_deepcopy_62" shape = [256, 256] dtype = "float32" min_val = float("-0.108251") @@ -903,6 +1003,7 @@ class Program_weight_tensor_parameter_99: class Program_weight_tensor_parameter_100: name = "parameter_100" + original_name = "linear_61.b_0_deepcopy_61" shape = [256] dtype = "float32" data = None @@ -910,6 +1011,7 @@ class Program_weight_tensor_parameter_100: class Program_weight_tensor_parameter_101: name = "parameter_101" + original_name = "linear_61.w_0_deepcopy_60" shape = [256, 256] dtype = "float32" min_val = float("-0.108253") @@ -921,6 +1023,7 @@ class Program_weight_tensor_parameter_101: class Program_weight_tensor_parameter_102: name = "parameter_102" + original_name = "layer_norm_34.b_0_deepcopy_73" shape = [256] dtype = "float32" data = None @@ -928,6 +1031,7 @@ class Program_weight_tensor_parameter_102: class Program_weight_tensor_parameter_103: name = "parameter_103" + original_name = "layer_norm_34.w_0_deepcopy_72" shape = [256] dtype = "float32" min_val = float("1.0") @@ -938,6 +1042,7 @@ class Program_weight_tensor_parameter_103: class Program_weight_tensor_parameter_104: name = "parameter_104" + original_name = "linear_60.b_0_deepcopy_59" shape = [256] dtype = "float32" data = None @@ -945,6 +1050,7 @@ class Program_weight_tensor_parameter_104: class Program_weight_tensor_parameter_105: name = "parameter_105" + original_name = "linear_60.w_0_deepcopy_58" shape = [256, 256] dtype = "float32" min_val = float("-0.108243") @@ -956,6 +1062,7 @@ class Program_weight_tensor_parameter_105: class Program_weight_tensor_parameter_106: name = "parameter_106" + original_name = "linear_59.b_0_deepcopy_57" shape = [256] dtype = "float32" data = None @@ -963,6 +1070,7 @@ class Program_weight_tensor_parameter_106: class Program_weight_tensor_parameter_107: name = "parameter_107" + original_name = "linear_59.w_0_deepcopy_56" shape = [256, 256] dtype = "float32" min_val = float("-0.108252") @@ -974,6 +1082,7 @@ class Program_weight_tensor_parameter_107: class Program_weight_tensor_parameter_108: name = "parameter_108" + original_name = "linear_58.b_0_deepcopy_55" shape = [256] dtype = "float32" data = None @@ -981,6 +1090,7 @@ class Program_weight_tensor_parameter_108: class Program_weight_tensor_parameter_109: name = "parameter_109" + original_name = "linear_58.w_0_deepcopy_54" shape = [256, 256] dtype = "float32" min_val = float("-0.108252") @@ -992,6 +1102,7 @@ class Program_weight_tensor_parameter_109: class Program_weight_tensor_parameter_110: name = "parameter_110" + original_name = "linear_57.b_0_deepcopy_53" shape = [256] dtype = "float32" data = None @@ -999,6 +1110,7 @@ class Program_weight_tensor_parameter_110: class Program_weight_tensor_parameter_111: name = "parameter_111" + original_name = "linear_57.w_0_deepcopy_52" shape = [256, 256] dtype = "float32" min_val = float("-0.10825") @@ -1010,6 +1122,7 @@ class Program_weight_tensor_parameter_111: class Program_weight_tensor_parameter_112: name = "parameter_112" + original_name = "layer_norm_36.b_0_deepcopy_51" shape = [256] dtype = "float32" data = None @@ -1017,6 +1130,7 @@ class Program_weight_tensor_parameter_112: class Program_weight_tensor_parameter_113: name = "parameter_113" + original_name = "layer_norm_36.w_0_deepcopy_50" shape = [256] dtype = "float32" min_val = float("1.0") @@ -1027,6 +1141,7 @@ class Program_weight_tensor_parameter_113: class Program_weight_tensor_parameter_114: name = "parameter_114" + original_name = "linear_66.b_0_deepcopy_45" shape = [256] dtype = "float32" data = None @@ -1034,6 +1149,7 @@ class Program_weight_tensor_parameter_114: class Program_weight_tensor_parameter_115: name = "parameter_115" + original_name = "linear_66.w_0_deepcopy_44" shape = [2048, 256] dtype = "float32" min_val = float("-0.051031") @@ -1045,6 +1161,7 @@ class Program_weight_tensor_parameter_115: class Program_weight_tensor_parameter_116: name = "parameter_116" + original_name = "linear_65.b_0_deepcopy_43" shape = [2048] dtype = "float32" data = None @@ -1052,6 +1169,7 @@ class Program_weight_tensor_parameter_116: class Program_weight_tensor_parameter_117: name = "parameter_117" + original_name = "linear_65.w_0_deepcopy_42" shape = [256, 2048] dtype = "float32" min_val = float("-0.0510309") @@ -1063,6 +1181,7 @@ class Program_weight_tensor_parameter_117: class Program_weight_tensor_parameter_118: name = "parameter_118" + original_name = "layer_norm_35.b_0_deepcopy_49" shape = [256] dtype = "float32" data = None @@ -1070,6 +1189,7 @@ class Program_weight_tensor_parameter_118: class Program_weight_tensor_parameter_119: name = "parameter_119" + original_name = "layer_norm_35.w_0_deepcopy_48" shape = [256] dtype = "float32" min_val = float("1.0") @@ -1080,6 +1200,7 @@ class Program_weight_tensor_parameter_119: class Program_weight_tensor_parameter_120: name = "parameter_120" + original_name = "linear_64.b_0_deepcopy_41" shape = [256] dtype = "float32" data = None @@ -1087,6 +1208,7 @@ class Program_weight_tensor_parameter_120: class Program_weight_tensor_parameter_121: name = "parameter_121" + original_name = "linear_64.w_0_deepcopy_40" shape = [256, 256] dtype = "float32" min_val = float("-0.10825") @@ -1098,6 +1220,7 @@ class Program_weight_tensor_parameter_121: class Program_weight_tensor_parameter_122: name = "parameter_122" + original_name = "linear_63.b_0_deepcopy_39" shape = [256] dtype = "float32" data = None @@ -1105,6 +1228,7 @@ class Program_weight_tensor_parameter_122: class Program_weight_tensor_parameter_123: name = "parameter_123" + original_name = "linear_63.w_0_deepcopy_38" shape = [256, 256] dtype = "float32" min_val = float("-0.108253") @@ -1116,6 +1240,7 @@ class Program_weight_tensor_parameter_123: class Program_weight_tensor_parameter_124: name = "parameter_124" + original_name = "linear_62.b_0_deepcopy_37" shape = [256] dtype = "float32" data = None @@ -1123,6 +1248,7 @@ class Program_weight_tensor_parameter_124: class Program_weight_tensor_parameter_125: name = "parameter_125" + original_name = "linear_62.w_0_deepcopy_36" shape = [256, 256] dtype = "float32" min_val = float("-0.108251") @@ -1134,6 +1260,7 @@ class Program_weight_tensor_parameter_125: class Program_weight_tensor_parameter_126: name = "parameter_126" + original_name = "linear_61.b_0_deepcopy_35" shape = [256] dtype = "float32" data = None @@ -1141,6 +1268,7 @@ class Program_weight_tensor_parameter_126: class Program_weight_tensor_parameter_127: name = "parameter_127" + original_name = "linear_61.w_0_deepcopy_34" shape = [256, 256] dtype = "float32" min_val = float("-0.108253") @@ -1152,6 +1280,7 @@ class Program_weight_tensor_parameter_127: class Program_weight_tensor_parameter_128: name = "parameter_128" + original_name = "layer_norm_34.b_0_deepcopy_47" shape = [256] dtype = "float32" data = None @@ -1159,6 +1288,7 @@ class Program_weight_tensor_parameter_128: class Program_weight_tensor_parameter_129: name = "parameter_129" + original_name = "layer_norm_34.w_0_deepcopy_46" shape = [256] dtype = "float32" min_val = float("1.0") @@ -1169,6 +1299,7 @@ class Program_weight_tensor_parameter_129: class Program_weight_tensor_parameter_130: name = "parameter_130" + original_name = "linear_60.b_0_deepcopy_33" shape = [256] dtype = "float32" data = None @@ -1176,6 +1307,7 @@ class Program_weight_tensor_parameter_130: class Program_weight_tensor_parameter_131: name = "parameter_131" + original_name = "linear_60.w_0_deepcopy_32" shape = [256, 256] dtype = "float32" min_val = float("-0.108243") @@ -1187,6 +1319,7 @@ class Program_weight_tensor_parameter_131: class Program_weight_tensor_parameter_132: name = "parameter_132" + original_name = "linear_59.b_0_deepcopy_31" shape = [256] dtype = "float32" data = None @@ -1194,6 +1327,7 @@ class Program_weight_tensor_parameter_132: class Program_weight_tensor_parameter_133: name = "parameter_133" + original_name = "linear_59.w_0_deepcopy_30" shape = [256, 256] dtype = "float32" min_val = float("-0.108252") @@ -1205,6 +1339,7 @@ class Program_weight_tensor_parameter_133: class Program_weight_tensor_parameter_134: name = "parameter_134" + original_name = "linear_58.b_0_deepcopy_29" shape = [256] dtype = "float32" data = None @@ -1212,6 +1347,7 @@ class Program_weight_tensor_parameter_134: class Program_weight_tensor_parameter_135: name = "parameter_135" + original_name = "linear_58.w_0_deepcopy_28" shape = [256, 256] dtype = "float32" min_val = float("-0.108252") @@ -1223,6 +1359,7 @@ class Program_weight_tensor_parameter_135: class Program_weight_tensor_parameter_136: name = "parameter_136" + original_name = "linear_57.b_0_deepcopy_27" shape = [256] dtype = "float32" data = None @@ -1230,6 +1367,7 @@ class Program_weight_tensor_parameter_136: class Program_weight_tensor_parameter_137: name = "parameter_137" + original_name = "linear_57.w_0_deepcopy_26" shape = [256, 256] dtype = "float32" min_val = float("-0.10825") @@ -1241,6 +1379,7 @@ class Program_weight_tensor_parameter_137: class Program_weight_tensor_parameter_138: name = "parameter_138" + original_name = "layer_norm_37.b_0" shape = [256] dtype = "float32" data = None @@ -1248,6 +1387,7 @@ class Program_weight_tensor_parameter_138: class Program_weight_tensor_parameter_139: name = "parameter_139" + original_name = "layer_norm_37.w_0" shape = [256] dtype = "float32" min_val = float("1.0") @@ -1258,6 +1398,7 @@ class Program_weight_tensor_parameter_139: class Program_weight_tensor_parameter_140: name = "parameter_140" + original_name = "layer_norm_36.b_0_deepcopy_25" shape = [256] dtype = "float32" data = None @@ -1265,6 +1406,7 @@ class Program_weight_tensor_parameter_140: class Program_weight_tensor_parameter_141: name = "parameter_141" + original_name = "layer_norm_36.w_0_deepcopy_24" shape = [256] dtype = "float32" min_val = float("1.0") @@ -1275,6 +1417,7 @@ class Program_weight_tensor_parameter_141: class Program_weight_tensor_parameter_142: name = "parameter_142" + original_name = "linear_66.b_0_deepcopy_19" shape = [256] dtype = "float32" data = None @@ -1282,6 +1425,7 @@ class Program_weight_tensor_parameter_142: class Program_weight_tensor_parameter_143: name = "parameter_143" + original_name = "linear_66.w_0_deepcopy_18" shape = [2048, 256] dtype = "float32" min_val = float("-0.0510307") @@ -1293,6 +1437,7 @@ class Program_weight_tensor_parameter_143: class Program_weight_tensor_parameter_144: name = "parameter_144" + original_name = "linear_65.b_0_deepcopy_17" shape = [2048] dtype = "float32" data = None @@ -1300,6 +1445,7 @@ class Program_weight_tensor_parameter_144: class Program_weight_tensor_parameter_145: name = "parameter_145" + original_name = "linear_65.w_0_deepcopy_16" shape = [256, 2048] dtype = "float32" min_val = float("-0.0510309") @@ -1311,6 +1457,7 @@ class Program_weight_tensor_parameter_145: class Program_weight_tensor_parameter_146: name = "parameter_146" + original_name = "layer_norm_35.b_0_deepcopy_23" shape = [256] dtype = "float32" data = None @@ -1318,6 +1465,7 @@ class Program_weight_tensor_parameter_146: class Program_weight_tensor_parameter_147: name = "parameter_147" + original_name = "layer_norm_35.w_0_deepcopy_22" shape = [256] dtype = "float32" min_val = float("1.0") @@ -1328,6 +1476,7 @@ class Program_weight_tensor_parameter_147: class Program_weight_tensor_parameter_148: name = "parameter_148" + original_name = "linear_64.b_0_deepcopy_15" shape = [256] dtype = "float32" data = None @@ -1335,6 +1484,7 @@ class Program_weight_tensor_parameter_148: class Program_weight_tensor_parameter_149: name = "parameter_149" + original_name = "linear_64.w_0_deepcopy_14" shape = [256, 256] dtype = "float32" min_val = float("-0.10825") @@ -1346,6 +1496,7 @@ class Program_weight_tensor_parameter_149: class Program_weight_tensor_parameter_150: name = "parameter_150" + original_name = "linear_63.b_0_deepcopy_13" shape = [256] dtype = "float32" data = None @@ -1353,6 +1504,7 @@ class Program_weight_tensor_parameter_150: class Program_weight_tensor_parameter_151: name = "parameter_151" + original_name = "linear_63.w_0_deepcopy_12" shape = [256, 256] dtype = "float32" min_val = float("-0.108253") @@ -1364,6 +1516,7 @@ class Program_weight_tensor_parameter_151: class Program_weight_tensor_parameter_152: name = "parameter_152" + original_name = "linear_62.b_0_deepcopy_11" shape = [256] dtype = "float32" data = None @@ -1371,6 +1524,7 @@ class Program_weight_tensor_parameter_152: class Program_weight_tensor_parameter_153: name = "parameter_153" + original_name = "linear_62.w_0_deepcopy_10" shape = [256, 256] dtype = "float32" min_val = float("-0.108251") @@ -1382,6 +1536,7 @@ class Program_weight_tensor_parameter_153: class Program_weight_tensor_parameter_154: name = "parameter_154" + original_name = "linear_61.b_0_deepcopy_9" shape = [256] dtype = "float32" data = None @@ -1389,6 +1544,7 @@ class Program_weight_tensor_parameter_154: class Program_weight_tensor_parameter_155: name = "parameter_155" + original_name = "linear_61.w_0_deepcopy_8" shape = [256, 256] dtype = "float32" min_val = float("-0.108253") @@ -1400,6 +1556,7 @@ class Program_weight_tensor_parameter_155: class Program_weight_tensor_parameter_156: name = "parameter_156" + original_name = "layer_norm_34.b_0_deepcopy_21" shape = [256] dtype = "float32" data = None @@ -1407,6 +1564,7 @@ class Program_weight_tensor_parameter_156: class Program_weight_tensor_parameter_157: name = "parameter_157" + original_name = "layer_norm_34.w_0_deepcopy_20" shape = [256] dtype = "float32" min_val = float("1.0") @@ -1417,6 +1575,7 @@ class Program_weight_tensor_parameter_157: class Program_weight_tensor_parameter_158: name = "parameter_158" + original_name = "linear_60.b_0_deepcopy_7" shape = [256] dtype = "float32" data = None @@ -1424,6 +1583,7 @@ class Program_weight_tensor_parameter_158: class Program_weight_tensor_parameter_159: name = "parameter_159" + original_name = "linear_60.w_0_deepcopy_6" shape = [256, 256] dtype = "float32" min_val = float("-0.108243") @@ -1435,6 +1595,7 @@ class Program_weight_tensor_parameter_159: class Program_weight_tensor_parameter_160: name = "parameter_160" + original_name = "linear_59.b_0_deepcopy_5" shape = [256] dtype = "float32" data = None @@ -1442,6 +1603,7 @@ class Program_weight_tensor_parameter_160: class Program_weight_tensor_parameter_161: name = "parameter_161" + original_name = "linear_59.w_0_deepcopy_4" shape = [256, 256] dtype = "float32" min_val = float("-0.108252") @@ -1453,6 +1615,7 @@ class Program_weight_tensor_parameter_161: class Program_weight_tensor_parameter_162: name = "parameter_162" + original_name = "linear_58.b_0_deepcopy_3" shape = [256] dtype = "float32" data = None @@ -1460,6 +1623,7 @@ class Program_weight_tensor_parameter_162: class Program_weight_tensor_parameter_163: name = "parameter_163" + original_name = "linear_58.w_0_deepcopy_2" shape = [256, 256] dtype = "float32" min_val = float("-0.108252") @@ -1471,6 +1635,7 @@ class Program_weight_tensor_parameter_163: class Program_weight_tensor_parameter_164: name = "parameter_164" + original_name = "linear_57.b_0_deepcopy_1" shape = [256] dtype = "float32" data = None @@ -1478,6 +1643,7 @@ class Program_weight_tensor_parameter_164: class Program_weight_tensor_parameter_165: name = "parameter_165" + original_name = "linear_57.w_0_deepcopy_0" shape = [256, 256] dtype = "float32" min_val = float("-0.10825") @@ -1489,6 +1655,7 @@ class Program_weight_tensor_parameter_165: class Program_weight_tensor_parameter_166: name = "parameter_166" + original_name = "conv2d_9.b_0" shape = [256] dtype = "float32" data = None @@ -1496,6 +1663,7 @@ class Program_weight_tensor_parameter_166: class Program_weight_tensor_parameter_167: name = "parameter_167" + original_name = "conv2d_9.w_0" shape = [256, 768, 1, 1] dtype = "float32" min_val = float("-0.0624997") @@ -1507,6 +1675,7 @@ class Program_weight_tensor_parameter_167: class Program_weight_tensor_parameter_168: name = "parameter_168" + original_name = "conv2d_8.b_0" shape = [256] dtype = "float32" data = None @@ -1514,6 +1683,7 @@ class Program_weight_tensor_parameter_168: class Program_weight_tensor_parameter_169: name = "parameter_169" + original_name = "conv2d_8.w_0" shape = [256, 256, 3, 3] dtype = "float32" min_val = float("-0.0360844") @@ -1525,6 +1695,7 @@ class Program_weight_tensor_parameter_169: class Program_weight_tensor_parameter_170: name = "parameter_170" + original_name = "group_norm_1.b_0" shape = [256] dtype = "float32" data = None @@ -1532,6 +1703,7 @@ class Program_weight_tensor_parameter_170: class Program_weight_tensor_parameter_171: name = "parameter_171" + original_name = "group_norm_1.w_0" shape = [256] dtype = "float32" min_val = float("-0.106959") @@ -1543,6 +1715,7 @@ class Program_weight_tensor_parameter_171: class Program_weight_tensor_parameter_172: name = "parameter_172" + original_name = "conv2d_2.w_0" shape = [256, 256, 3, 3] dtype = "float32" min_val = float("-0.0360838") @@ -1554,6 +1727,7 @@ class Program_weight_tensor_parameter_172: class Program_weight_tensor_parameter_173: name = "parameter_173" + original_name = "group_norm_0.b_0" shape = [256] dtype = "float32" data = None @@ -1561,6 +1735,7 @@ class Program_weight_tensor_parameter_173: class Program_weight_tensor_parameter_174: name = "parameter_174" + original_name = "group_norm_0.w_0" shape = [256] dtype = "float32" min_val = float("-0.10821") @@ -1572,6 +1747,7 @@ class Program_weight_tensor_parameter_174: class Program_weight_tensor_parameter_175: name = "parameter_175" + original_name = "conv2d_1.w_0" shape = [256, 96, 1, 1] dtype = "float32" min_val = float("-0.176771") @@ -1583,6 +1759,7 @@ class Program_weight_tensor_parameter_175: class Program_weight_tensor_parameter_176: name = "parameter_176" + original_name = "group_norm_3.b_0" shape = [256] dtype = "float32" data = None @@ -1590,6 +1767,7 @@ class Program_weight_tensor_parameter_176: class Program_weight_tensor_parameter_177: name = "parameter_177" + original_name = "group_norm_3.w_0" shape = [256] dtype = "float32" min_val = float("-0.107216") @@ -1601,6 +1779,7 @@ class Program_weight_tensor_parameter_177: class Program_weight_tensor_parameter_178: name = "parameter_178" + original_name = "conv2d_4.w_0" shape = [256, 256, 3, 3] dtype = "float32" min_val = float("-0.0360843") @@ -1612,6 +1791,7 @@ class Program_weight_tensor_parameter_178: class Program_weight_tensor_parameter_179: name = "parameter_179" + original_name = "group_norm_2.b_0" shape = [256] dtype = "float32" data = None @@ -1619,6 +1799,7 @@ class Program_weight_tensor_parameter_179: class Program_weight_tensor_parameter_180: name = "parameter_180" + original_name = "group_norm_2.w_0" shape = [256] dtype = "float32" min_val = float("-0.107711") @@ -1630,6 +1811,7 @@ class Program_weight_tensor_parameter_180: class Program_weight_tensor_parameter_181: name = "parameter_181" + original_name = "conv2d_3.w_0" shape = [256, 192, 1, 1] dtype = "float32" min_val = float("-0.124999") @@ -1641,6 +1823,7 @@ class Program_weight_tensor_parameter_181: class Program_weight_tensor_parameter_182: name = "parameter_182" + original_name = "group_norm_5.b_0" shape = [256] dtype = "float32" data = None @@ -1648,6 +1831,7 @@ class Program_weight_tensor_parameter_182: class Program_weight_tensor_parameter_183: name = "parameter_183" + original_name = "group_norm_5.w_0" shape = [256] dtype = "float32" min_val = float("-0.106325") @@ -1659,6 +1843,7 @@ class Program_weight_tensor_parameter_183: class Program_weight_tensor_parameter_184: name = "parameter_184" + original_name = "conv2d_6.w_0" shape = [256, 256, 3, 3] dtype = "float32" min_val = float("-0.0360843") @@ -1670,6 +1855,7 @@ class Program_weight_tensor_parameter_184: class Program_weight_tensor_parameter_185: name = "parameter_185" + original_name = "group_norm_4.b_0" shape = [256] dtype = "float32" data = None @@ -1677,6 +1863,7 @@ class Program_weight_tensor_parameter_185: class Program_weight_tensor_parameter_186: name = "parameter_186" + original_name = "group_norm_4.w_0" shape = [256] dtype = "float32" min_val = float("-0.107962") @@ -1688,6 +1875,7 @@ class Program_weight_tensor_parameter_186: class Program_weight_tensor_parameter_187: name = "parameter_187" + original_name = "conv2d_5.w_0" shape = [256, 384, 1, 1] dtype = "float32" min_val = float("-0.0883853") @@ -1699,6 +1887,7 @@ class Program_weight_tensor_parameter_187: class Program_weight_tensor_parameter_188: name = "parameter_188" + original_name = "group_norm_6.b_0" shape = [256] dtype = "float32" data = None @@ -1706,6 +1895,7 @@ class Program_weight_tensor_parameter_188: class Program_weight_tensor_parameter_189: name = "parameter_189" + original_name = "group_norm_6.w_0" shape = [256] dtype = "float32" min_val = float("-0.107326") @@ -1717,6 +1907,7 @@ class Program_weight_tensor_parameter_189: class Program_weight_tensor_parameter_190: name = "parameter_190" + original_name = "conv2d_7.w_0" shape = [256, 768, 3, 3] dtype = "float32" min_val = float("-0.0208333") @@ -1728,6 +1919,7 @@ class Program_weight_tensor_parameter_190: class Program_weight_tensor_parameter_191: name = "parameter_191" + original_name = "layer_norm_31.b_0" shape = [768] dtype = "float32" data = None @@ -1735,6 +1927,7 @@ class Program_weight_tensor_parameter_191: class Program_weight_tensor_parameter_192: name = "parameter_192" + original_name = "layer_norm_31.w_0" shape = [768] dtype = "float32" min_val = float("1.0") @@ -1745,6 +1938,7 @@ class Program_weight_tensor_parameter_192: class Program_weight_tensor_parameter_193: name = "parameter_193" + original_name = "linear_50.b_0" shape = [768] dtype = "float32" min_val = float("-7.8427") @@ -1756,6 +1950,7 @@ class Program_weight_tensor_parameter_193: class Program_weight_tensor_parameter_194: name = "parameter_194" + original_name = "linear_50.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-2.13627") @@ -1767,6 +1962,7 @@ class Program_weight_tensor_parameter_194: class Program_weight_tensor_parameter_195: name = "parameter_195" + original_name = "linear_49.b_0" shape = [3072] dtype = "float32" min_val = float("-7.65925") @@ -1778,6 +1974,7 @@ class Program_weight_tensor_parameter_195: class Program_weight_tensor_parameter_196: name = "parameter_196" + original_name = "linear_49.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-1.40798") @@ -1789,6 +1986,7 @@ class Program_weight_tensor_parameter_196: class Program_weight_tensor_parameter_197: name = "parameter_197" + original_name = "layer_norm_27.b_0" shape = [768] dtype = "float32" min_val = float("-7.05455") @@ -1800,6 +1998,7 @@ class Program_weight_tensor_parameter_197: class Program_weight_tensor_parameter_198: name = "parameter_198" + original_name = "layer_norm_27.w_0" shape = [768] dtype = "float32" min_val = float("-0.291477") @@ -1811,6 +2010,7 @@ class Program_weight_tensor_parameter_198: class Program_weight_tensor_parameter_199: name = "parameter_199" + original_name = "linear_48.b_0" shape = [768] dtype = "float32" min_val = float("-11.8044") @@ -1822,6 +2022,7 @@ class Program_weight_tensor_parameter_199: class Program_weight_tensor_parameter_200: name = "parameter_200" + original_name = "linear_48.w_0" shape = [768, 768] dtype = "float32" min_val = float("-1.72301") @@ -1833,6 +2034,7 @@ class Program_weight_tensor_parameter_200: class Program_weight_tensor_parameter_201: name = "parameter_201" + original_name = "linear_47.b_0" shape = [2304] dtype = "float32" min_val = float("-7.31156") @@ -1844,6 +2046,7 @@ class Program_weight_tensor_parameter_201: class Program_weight_tensor_parameter_202: name = "parameter_202" + original_name = "linear_47.w_0" shape = [768, 2304] dtype = "float32" min_val = float("-0.509874") @@ -1855,6 +2058,7 @@ class Program_weight_tensor_parameter_202: class Program_weight_tensor_parameter_203: name = "parameter_203" + original_name = "layer_norm_26.b_0" shape = [768] dtype = "float32" min_val = float("-7.10294") @@ -1866,6 +2070,7 @@ class Program_weight_tensor_parameter_203: class Program_weight_tensor_parameter_204: name = "parameter_204" + original_name = "layer_norm_26.w_0" shape = [768] dtype = "float32" min_val = float("0.00304417") @@ -1877,6 +2082,7 @@ class Program_weight_tensor_parameter_204: class Program_weight_tensor_parameter_205: name = "parameter_205" + original_name = "linear_46.b_0" shape = [768] dtype = "float32" min_val = float("-5.52452") @@ -1888,6 +2094,7 @@ class Program_weight_tensor_parameter_205: class Program_weight_tensor_parameter_206: name = "parameter_206" + original_name = "linear_46.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-2.01228") @@ -1899,6 +2106,7 @@ class Program_weight_tensor_parameter_206: class Program_weight_tensor_parameter_207: name = "parameter_207" + original_name = "linear_45.b_0" shape = [3072] dtype = "float32" min_val = float("-6.01539") @@ -1910,6 +2118,7 @@ class Program_weight_tensor_parameter_207: class Program_weight_tensor_parameter_208: name = "parameter_208" + original_name = "linear_45.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.43767") @@ -1921,6 +2130,7 @@ class Program_weight_tensor_parameter_208: class Program_weight_tensor_parameter_209: name = "parameter_209" + original_name = "layer_norm_25.b_0" shape = [768] dtype = "float32" min_val = float("-18.4878") @@ -1932,6 +2142,7 @@ class Program_weight_tensor_parameter_209: class Program_weight_tensor_parameter_210: name = "parameter_210" + original_name = "layer_norm_25.w_0" shape = [768] dtype = "float32" min_val = float("-0.0755746") @@ -1943,6 +2154,7 @@ class Program_weight_tensor_parameter_210: class Program_weight_tensor_parameter_211: name = "parameter_211" + original_name = "linear_44.b_0" shape = [768] dtype = "float32" min_val = float("-5.11022") @@ -1954,6 +2166,7 @@ class Program_weight_tensor_parameter_211: class Program_weight_tensor_parameter_212: name = "parameter_212" + original_name = "linear_44.w_0" shape = [768, 768] dtype = "float32" min_val = float("-1.51202") @@ -1965,6 +2178,7 @@ class Program_weight_tensor_parameter_212: class Program_weight_tensor_parameter_213: name = "parameter_213" + original_name = "linear_43.b_0" shape = [2304] dtype = "float32" min_val = float("-6.81097") @@ -1976,6 +2190,7 @@ class Program_weight_tensor_parameter_213: class Program_weight_tensor_parameter_214: name = "parameter_214" + original_name = "linear_43.w_0" shape = [768, 2304] dtype = "float32" min_val = float("-0.494403") @@ -1987,6 +2202,7 @@ class Program_weight_tensor_parameter_214: class Program_weight_tensor_parameter_215: name = "parameter_215" + original_name = "layer_norm_24.b_0" shape = [768] dtype = "float32" min_val = float("-3.91771") @@ -1998,6 +2214,7 @@ class Program_weight_tensor_parameter_215: class Program_weight_tensor_parameter_216: name = "parameter_216" + original_name = "layer_norm_24.w_0" shape = [768] dtype = "float32" min_val = float("-0.00323922") @@ -2009,6 +2226,7 @@ class Program_weight_tensor_parameter_216: class Program_weight_tensor_parameter_217: name = "parameter_217" + original_name = "layer_norm_30.b_0" shape = [384] dtype = "float32" data = None @@ -2016,6 +2234,7 @@ class Program_weight_tensor_parameter_217: class Program_weight_tensor_parameter_218: name = "parameter_218" + original_name = "layer_norm_30.w_0" shape = [384] dtype = "float32" min_val = float("1.0") @@ -2027,6 +2246,7 @@ class Program_weight_tensor_parameter_218: class Program_weight_tensor_parameter_219: name = "parameter_219" + original_name = "linear_42.w_0" shape = [1536, 768] dtype = "float32" min_val = float("-1.17259") @@ -2038,6 +2258,7 @@ class Program_weight_tensor_parameter_219: class Program_weight_tensor_parameter_220: name = "parameter_220" + original_name = "layer_norm_23.b_0" shape = [1536] dtype = "float32" min_val = float("-1.18188") @@ -2049,6 +2270,7 @@ class Program_weight_tensor_parameter_220: class Program_weight_tensor_parameter_221: name = "parameter_221" + original_name = "layer_norm_23.w_0" shape = [1536] dtype = "float32" min_val = float("0.143443") @@ -2060,6 +2282,7 @@ class Program_weight_tensor_parameter_221: class Program_weight_tensor_parameter_222: name = "parameter_222" + original_name = "linear_41.b_0" shape = [384] dtype = "float32" min_val = float("-1.15477") @@ -2071,6 +2294,7 @@ class Program_weight_tensor_parameter_222: class Program_weight_tensor_parameter_223: name = "parameter_223" + original_name = "linear_41.w_0" shape = [1536, 384] dtype = "float32" min_val = float("-0.639952") @@ -2082,6 +2306,7 @@ class Program_weight_tensor_parameter_223: class Program_weight_tensor_parameter_224: name = "parameter_224" + original_name = "linear_40.b_0" shape = [1536] dtype = "float32" min_val = float("-5.43379") @@ -2093,6 +2318,7 @@ class Program_weight_tensor_parameter_224: class Program_weight_tensor_parameter_225: name = "parameter_225" + original_name = "linear_40.w_0" shape = [384, 1536] dtype = "float32" min_val = float("-0.674227") @@ -2104,6 +2330,7 @@ class Program_weight_tensor_parameter_225: class Program_weight_tensor_parameter_226: name = "parameter_226" + original_name = "layer_norm_22.b_0" shape = [384] dtype = "float32" min_val = float("-4.02772") @@ -2115,6 +2342,7 @@ class Program_weight_tensor_parameter_226: class Program_weight_tensor_parameter_227: name = "parameter_227" + original_name = "layer_norm_22.w_0" shape = [384] dtype = "float32" min_val = float("0.355152") @@ -2126,6 +2354,7 @@ class Program_weight_tensor_parameter_227: class Program_weight_tensor_parameter_228: name = "parameter_228" + original_name = "linear_39.b_0" shape = [384] dtype = "float32" min_val = float("-1.36797") @@ -2137,6 +2366,7 @@ class Program_weight_tensor_parameter_228: class Program_weight_tensor_parameter_229: name = "parameter_229" + original_name = "linear_39.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.601018") @@ -2148,6 +2378,7 @@ class Program_weight_tensor_parameter_229: class Program_weight_tensor_parameter_230: name = "parameter_230" + original_name = "linear_38.b_0" shape = [1152] dtype = "float32" min_val = float("-4.19132") @@ -2159,6 +2390,7 @@ class Program_weight_tensor_parameter_230: class Program_weight_tensor_parameter_231: name = "parameter_231" + original_name = "linear_38.w_0" shape = [384, 1152] dtype = "float32" min_val = float("-0.461236") @@ -2170,6 +2402,7 @@ class Program_weight_tensor_parameter_231: class Program_weight_tensor_parameter_232: name = "parameter_232" + original_name = "layer_norm_21.b_0" shape = [384] dtype = "float32" min_val = float("-3.1804") @@ -2181,6 +2414,7 @@ class Program_weight_tensor_parameter_232: class Program_weight_tensor_parameter_233: name = "parameter_233" + original_name = "layer_norm_21.w_0" shape = [384] dtype = "float32" min_val = float("0.204104") @@ -2192,6 +2426,7 @@ class Program_weight_tensor_parameter_233: class Program_weight_tensor_parameter_234: name = "parameter_234" + original_name = "linear_37.b_0" shape = [384] dtype = "float32" min_val = float("-1.2134") @@ -2203,6 +2438,7 @@ class Program_weight_tensor_parameter_234: class Program_weight_tensor_parameter_235: name = "parameter_235" + original_name = "linear_37.w_0" shape = [1536, 384] dtype = "float32" min_val = float("-0.57784") @@ -2214,6 +2450,7 @@ class Program_weight_tensor_parameter_235: class Program_weight_tensor_parameter_236: name = "parameter_236" + original_name = "linear_36.b_0" shape = [1536] dtype = "float32" min_val = float("-5.32214") @@ -2225,6 +2462,7 @@ class Program_weight_tensor_parameter_236: class Program_weight_tensor_parameter_237: name = "parameter_237" + original_name = "linear_36.w_0" shape = [384, 1536] dtype = "float32" min_val = float("-0.59506") @@ -2236,6 +2474,7 @@ class Program_weight_tensor_parameter_237: class Program_weight_tensor_parameter_238: name = "parameter_238" + original_name = "layer_norm_20.b_0" shape = [384] dtype = "float32" min_val = float("-4.21981") @@ -2247,6 +2486,7 @@ class Program_weight_tensor_parameter_238: class Program_weight_tensor_parameter_239: name = "parameter_239" + original_name = "layer_norm_20.w_0" shape = [384] dtype = "float32" min_val = float("0.386956") @@ -2258,6 +2498,7 @@ class Program_weight_tensor_parameter_239: class Program_weight_tensor_parameter_240: name = "parameter_240" + original_name = "linear_35.b_0" shape = [384] dtype = "float32" min_val = float("-1.6503") @@ -2269,6 +2510,7 @@ class Program_weight_tensor_parameter_240: class Program_weight_tensor_parameter_241: name = "parameter_241" + original_name = "linear_35.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.665253") @@ -2280,6 +2522,7 @@ class Program_weight_tensor_parameter_241: class Program_weight_tensor_parameter_242: name = "parameter_242" + original_name = "linear_34.b_0" shape = [1152] dtype = "float32" min_val = float("-5.18755") @@ -2291,6 +2534,7 @@ class Program_weight_tensor_parameter_242: class Program_weight_tensor_parameter_243: name = "parameter_243" + original_name = "linear_34.w_0" shape = [384, 1152] dtype = "float32" min_val = float("-0.380591") @@ -2302,6 +2546,7 @@ class Program_weight_tensor_parameter_243: class Program_weight_tensor_parameter_244: name = "parameter_244" + original_name = "layer_norm_19.b_0" shape = [384] dtype = "float32" min_val = float("-3.50029") @@ -2313,6 +2558,7 @@ class Program_weight_tensor_parameter_244: class Program_weight_tensor_parameter_245: name = "parameter_245" + original_name = "layer_norm_19.w_0" shape = [384] dtype = "float32" min_val = float("0.12845") @@ -2324,6 +2570,7 @@ class Program_weight_tensor_parameter_245: class Program_weight_tensor_parameter_246: name = "parameter_246" + original_name = "linear_33.b_0" shape = [384] dtype = "float32" min_val = float("-1.36973") @@ -2335,6 +2582,7 @@ class Program_weight_tensor_parameter_246: class Program_weight_tensor_parameter_247: name = "parameter_247" + original_name = "linear_33.w_0" shape = [1536, 384] dtype = "float32" min_val = float("-0.733098") @@ -2346,6 +2594,7 @@ class Program_weight_tensor_parameter_247: class Program_weight_tensor_parameter_248: name = "parameter_248" + original_name = "linear_32.b_0" shape = [1536] dtype = "float32" min_val = float("-5.67193") @@ -2357,6 +2606,7 @@ class Program_weight_tensor_parameter_248: class Program_weight_tensor_parameter_249: name = "parameter_249" + original_name = "linear_32.w_0" shape = [384, 1536] dtype = "float32" min_val = float("-0.571878") @@ -2368,6 +2618,7 @@ class Program_weight_tensor_parameter_249: class Program_weight_tensor_parameter_250: name = "parameter_250" + original_name = "layer_norm_18.b_0" shape = [384] dtype = "float32" min_val = float("-4.46615") @@ -2379,6 +2630,7 @@ class Program_weight_tensor_parameter_250: class Program_weight_tensor_parameter_251: name = "parameter_251" + original_name = "layer_norm_18.w_0" shape = [384] dtype = "float32" min_val = float("-0.00102235") @@ -2390,6 +2642,7 @@ class Program_weight_tensor_parameter_251: class Program_weight_tensor_parameter_252: name = "parameter_252" + original_name = "linear_31.b_0" shape = [384] dtype = "float32" min_val = float("-0.896298") @@ -2401,6 +2654,7 @@ class Program_weight_tensor_parameter_252: class Program_weight_tensor_parameter_253: name = "parameter_253" + original_name = "linear_31.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.442041") @@ -2412,6 +2666,7 @@ class Program_weight_tensor_parameter_253: class Program_weight_tensor_parameter_254: name = "parameter_254" + original_name = "linear_30.b_0" shape = [1152] dtype = "float32" min_val = float("-3.21664") @@ -2423,6 +2678,7 @@ class Program_weight_tensor_parameter_254: class Program_weight_tensor_parameter_255: name = "parameter_255" + original_name = "linear_30.w_0" shape = [384, 1152] dtype = "float32" min_val = float("-0.539962") @@ -2434,6 +2690,7 @@ class Program_weight_tensor_parameter_255: class Program_weight_tensor_parameter_256: name = "parameter_256" + original_name = "layer_norm_17.b_0" shape = [384] dtype = "float32" min_val = float("-2.98689") @@ -2445,6 +2702,7 @@ class Program_weight_tensor_parameter_256: class Program_weight_tensor_parameter_257: name = "parameter_257" + original_name = "layer_norm_17.w_0" shape = [384] dtype = "float32" min_val = float("-0.00691647") @@ -2456,6 +2714,7 @@ class Program_weight_tensor_parameter_257: class Program_weight_tensor_parameter_258: name = "parameter_258" + original_name = "linear_29.b_0" shape = [384] dtype = "float32" min_val = float("-1.55968") @@ -2467,6 +2726,7 @@ class Program_weight_tensor_parameter_258: class Program_weight_tensor_parameter_259: name = "parameter_259" + original_name = "linear_29.w_0" shape = [1536, 384] dtype = "float32" min_val = float("-0.660884") @@ -2478,6 +2738,7 @@ class Program_weight_tensor_parameter_259: class Program_weight_tensor_parameter_260: name = "parameter_260" + original_name = "linear_28.b_0" shape = [1536] dtype = "float32" min_val = float("-5.42016") @@ -2489,6 +2750,7 @@ class Program_weight_tensor_parameter_260: class Program_weight_tensor_parameter_261: name = "parameter_261" + original_name = "linear_28.w_0" shape = [384, 1536] dtype = "float32" min_val = float("-0.543028") @@ -2500,6 +2762,7 @@ class Program_weight_tensor_parameter_261: class Program_weight_tensor_parameter_262: name = "parameter_262" + original_name = "layer_norm_16.b_0" shape = [384] dtype = "float32" min_val = float("-4.10652") @@ -2511,6 +2774,7 @@ class Program_weight_tensor_parameter_262: class Program_weight_tensor_parameter_263: name = "parameter_263" + original_name = "layer_norm_16.w_0" shape = [384] dtype = "float32" min_val = float("0.00332775") @@ -2522,6 +2786,7 @@ class Program_weight_tensor_parameter_263: class Program_weight_tensor_parameter_264: name = "parameter_264" + original_name = "linear_27.b_0" shape = [384] dtype = "float32" min_val = float("-0.925735") @@ -2533,6 +2798,7 @@ class Program_weight_tensor_parameter_264: class Program_weight_tensor_parameter_265: name = "parameter_265" + original_name = "linear_27.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.502551") @@ -2544,6 +2810,7 @@ class Program_weight_tensor_parameter_265: class Program_weight_tensor_parameter_266: name = "parameter_266" + original_name = "linear_26.b_0" shape = [1152] dtype = "float32" min_val = float("-3.70529") @@ -2555,6 +2822,7 @@ class Program_weight_tensor_parameter_266: class Program_weight_tensor_parameter_267: name = "parameter_267" + original_name = "linear_26.w_0" shape = [384, 1152] dtype = "float32" min_val = float("-0.459269") @@ -2566,6 +2834,7 @@ class Program_weight_tensor_parameter_267: class Program_weight_tensor_parameter_268: name = "parameter_268" + original_name = "layer_norm_15.b_0" shape = [384] dtype = "float32" min_val = float("-3.17884") @@ -2577,6 +2846,7 @@ class Program_weight_tensor_parameter_268: class Program_weight_tensor_parameter_269: name = "parameter_269" + original_name = "layer_norm_15.w_0" shape = [384] dtype = "float32" min_val = float("-0.00690475") @@ -2588,6 +2858,7 @@ class Program_weight_tensor_parameter_269: class Program_weight_tensor_parameter_270: name = "parameter_270" + original_name = "linear_25.b_0" shape = [384] dtype = "float32" min_val = float("-1.56459") @@ -2599,6 +2870,7 @@ class Program_weight_tensor_parameter_270: class Program_weight_tensor_parameter_271: name = "parameter_271" + original_name = "linear_25.w_0" shape = [1536, 384] dtype = "float32" min_val = float("-0.534191") @@ -2610,6 +2882,7 @@ class Program_weight_tensor_parameter_271: class Program_weight_tensor_parameter_272: name = "parameter_272" + original_name = "linear_24.b_0" shape = [1536] dtype = "float32" min_val = float("-5.471") @@ -2621,6 +2894,7 @@ class Program_weight_tensor_parameter_272: class Program_weight_tensor_parameter_273: name = "parameter_273" + original_name = "linear_24.w_0" shape = [384, 1536] dtype = "float32" min_val = float("-0.705037") @@ -2632,6 +2906,7 @@ class Program_weight_tensor_parameter_273: class Program_weight_tensor_parameter_274: name = "parameter_274" + original_name = "layer_norm_14.b_0" shape = [384] dtype = "float32" min_val = float("-2.77399") @@ -2643,6 +2918,7 @@ class Program_weight_tensor_parameter_274: class Program_weight_tensor_parameter_275: name = "parameter_275" + original_name = "layer_norm_14.w_0" shape = [384] dtype = "float32" min_val = float("-0.00153011") @@ -2654,6 +2930,7 @@ class Program_weight_tensor_parameter_275: class Program_weight_tensor_parameter_276: name = "parameter_276" + original_name = "linear_23.b_0" shape = [384] dtype = "float32" min_val = float("-0.882424") @@ -2665,6 +2942,7 @@ class Program_weight_tensor_parameter_276: class Program_weight_tensor_parameter_277: name = "parameter_277" + original_name = "linear_23.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.841758") @@ -2676,6 +2954,7 @@ class Program_weight_tensor_parameter_277: class Program_weight_tensor_parameter_278: name = "parameter_278" + original_name = "linear_22.b_0" shape = [1152] dtype = "float32" min_val = float("-3.42565") @@ -2687,6 +2966,7 @@ class Program_weight_tensor_parameter_278: class Program_weight_tensor_parameter_279: name = "parameter_279" + original_name = "linear_22.w_0" shape = [384, 1152] dtype = "float32" min_val = float("-0.464748") @@ -2698,6 +2978,7 @@ class Program_weight_tensor_parameter_279: class Program_weight_tensor_parameter_280: name = "parameter_280" + original_name = "layer_norm_13.b_0" shape = [384] dtype = "float32" min_val = float("-1.18072") @@ -2709,6 +2990,7 @@ class Program_weight_tensor_parameter_280: class Program_weight_tensor_parameter_281: name = "parameter_281" + original_name = "layer_norm_13.w_0" shape = [384] dtype = "float32" min_val = float("-0.011215") @@ -2720,6 +3002,7 @@ class Program_weight_tensor_parameter_281: class Program_weight_tensor_parameter_282: name = "parameter_282" + original_name = "linear_21.b_0" shape = [384] dtype = "float32" min_val = float("-2.05865") @@ -2731,6 +3014,7 @@ class Program_weight_tensor_parameter_282: class Program_weight_tensor_parameter_283: name = "parameter_283" + original_name = "linear_21.w_0" shape = [1536, 384] dtype = "float32" min_val = float("-0.782334") @@ -2742,6 +3026,7 @@ class Program_weight_tensor_parameter_283: class Program_weight_tensor_parameter_284: name = "parameter_284" + original_name = "linear_20.b_0" shape = [1536] dtype = "float32" min_val = float("-5.41117") @@ -2753,6 +3038,7 @@ class Program_weight_tensor_parameter_284: class Program_weight_tensor_parameter_285: name = "parameter_285" + original_name = "linear_20.w_0" shape = [384, 1536] dtype = "float32" min_val = float("-0.523359") @@ -2764,6 +3050,7 @@ class Program_weight_tensor_parameter_285: class Program_weight_tensor_parameter_286: name = "parameter_286" + original_name = "layer_norm_12.b_0" shape = [384] dtype = "float32" min_val = float("-2.33563") @@ -2775,6 +3062,7 @@ class Program_weight_tensor_parameter_286: class Program_weight_tensor_parameter_287: name = "parameter_287" + original_name = "layer_norm_12.w_0" shape = [384] dtype = "float32" min_val = float("-0.00924296") @@ -2786,6 +3074,7 @@ class Program_weight_tensor_parameter_287: class Program_weight_tensor_parameter_288: name = "parameter_288" + original_name = "linear_19.b_0" shape = [384] dtype = "float32" min_val = float("-1.76653") @@ -2797,6 +3086,7 @@ class Program_weight_tensor_parameter_288: class Program_weight_tensor_parameter_289: name = "parameter_289" + original_name = "linear_19.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.578849") @@ -2808,6 +3098,7 @@ class Program_weight_tensor_parameter_289: class Program_weight_tensor_parameter_290: name = "parameter_290" + original_name = "linear_18.b_0" shape = [1152] dtype = "float32" min_val = float("-4.34906") @@ -2819,6 +3110,7 @@ class Program_weight_tensor_parameter_290: class Program_weight_tensor_parameter_291: name = "parameter_291" + original_name = "linear_18.w_0" shape = [384, 1152] dtype = "float32" min_val = float("-0.434441") @@ -2830,6 +3122,7 @@ class Program_weight_tensor_parameter_291: class Program_weight_tensor_parameter_292: name = "parameter_292" + original_name = "layer_norm_11.b_0" shape = [384] dtype = "float32" min_val = float("-0.160743") @@ -2841,6 +3134,7 @@ class Program_weight_tensor_parameter_292: class Program_weight_tensor_parameter_293: name = "parameter_293" + original_name = "layer_norm_11.w_0" shape = [384] dtype = "float32" min_val = float("-0.00772838") @@ -2852,6 +3146,7 @@ class Program_weight_tensor_parameter_293: class Program_weight_tensor_parameter_294: name = "parameter_294" + original_name = "layer_norm_29.b_0" shape = [192] dtype = "float32" data = None @@ -2859,6 +3154,7 @@ class Program_weight_tensor_parameter_294: class Program_weight_tensor_parameter_295: name = "parameter_295" + original_name = "layer_norm_29.w_0" shape = [192] dtype = "float32" min_val = float("1.0") @@ -2869,6 +3165,7 @@ class Program_weight_tensor_parameter_295: class Program_weight_tensor_parameter_296: name = "parameter_296" + original_name = "linear_17.w_0" shape = [768, 384] dtype = "float32" min_val = float("-0.357094") @@ -2880,6 +3177,7 @@ class Program_weight_tensor_parameter_296: class Program_weight_tensor_parameter_297: name = "parameter_297" + original_name = "layer_norm_10.b_0" shape = [768] dtype = "float32" min_val = float("-1.35845") @@ -2891,6 +3189,7 @@ class Program_weight_tensor_parameter_297: class Program_weight_tensor_parameter_298: name = "parameter_298" + original_name = "layer_norm_10.w_0" shape = [768] dtype = "float32" min_val = float("0.250203") @@ -2902,6 +3201,7 @@ class Program_weight_tensor_parameter_298: class Program_weight_tensor_parameter_299: name = "parameter_299" + original_name = "linear_16.b_0" shape = [192] dtype = "float32" min_val = float("-5.65409") @@ -2913,6 +3213,7 @@ class Program_weight_tensor_parameter_299: class Program_weight_tensor_parameter_300: name = "parameter_300" + original_name = "linear_16.w_0" shape = [768, 192] dtype = "float32" min_val = float("-1.01845") @@ -2924,6 +3225,7 @@ class Program_weight_tensor_parameter_300: class Program_weight_tensor_parameter_301: name = "parameter_301" + original_name = "linear_15.b_0" shape = [768] dtype = "float32" min_val = float("-3.30376") @@ -2935,6 +3237,7 @@ class Program_weight_tensor_parameter_301: class Program_weight_tensor_parameter_302: name = "parameter_302" + original_name = "linear_15.w_0" shape = [192, 768] dtype = "float32" min_val = float("-0.486469") @@ -2946,6 +3249,7 @@ class Program_weight_tensor_parameter_302: class Program_weight_tensor_parameter_303: name = "parameter_303" + original_name = "layer_norm_9.b_0" shape = [192] dtype = "float32" min_val = float("-1.19854") @@ -2957,6 +3261,7 @@ class Program_weight_tensor_parameter_303: class Program_weight_tensor_parameter_304: name = "parameter_304" + original_name = "layer_norm_9.w_0" shape = [192] dtype = "float32" min_val = float("0.387802") @@ -2968,6 +3273,7 @@ class Program_weight_tensor_parameter_304: class Program_weight_tensor_parameter_305: name = "parameter_305" + original_name = "linear_14.b_0" shape = [192] dtype = "float32" min_val = float("-4.43349") @@ -2979,6 +3285,7 @@ class Program_weight_tensor_parameter_305: class Program_weight_tensor_parameter_306: name = "parameter_306" + original_name = "linear_14.w_0" shape = [192, 192] dtype = "float32" min_val = float("-0.277136") @@ -2990,6 +3297,7 @@ class Program_weight_tensor_parameter_306: class Program_weight_tensor_parameter_307: name = "parameter_307" + original_name = "linear_13.b_0" shape = [576] dtype = "float32" min_val = float("-3.70476") @@ -3001,6 +3309,7 @@ class Program_weight_tensor_parameter_307: class Program_weight_tensor_parameter_308: name = "parameter_308" + original_name = "linear_13.w_0" shape = [192, 576] dtype = "float32" min_val = float("-0.614933") @@ -3012,6 +3321,7 @@ class Program_weight_tensor_parameter_308: class Program_weight_tensor_parameter_309: name = "parameter_309" + original_name = "layer_norm_8.b_0" shape = [192] dtype = "float32" min_val = float("-0.708") @@ -3023,6 +3333,7 @@ class Program_weight_tensor_parameter_309: class Program_weight_tensor_parameter_310: name = "parameter_310" + original_name = "layer_norm_8.w_0" shape = [192] dtype = "float32" min_val = float("0.486079") @@ -3034,6 +3345,7 @@ class Program_weight_tensor_parameter_310: class Program_weight_tensor_parameter_311: name = "parameter_311" + original_name = "linear_12.b_0" shape = [192] dtype = "float32" min_val = float("-6.42398") @@ -3045,6 +3357,7 @@ class Program_weight_tensor_parameter_311: class Program_weight_tensor_parameter_312: name = "parameter_312" + original_name = "linear_12.w_0" shape = [768, 192] dtype = "float32" min_val = float("-0.562966") @@ -3056,6 +3369,7 @@ class Program_weight_tensor_parameter_312: class Program_weight_tensor_parameter_313: name = "parameter_313" + original_name = "linear_11.b_0" shape = [768] dtype = "float32" min_val = float("-2.88993") @@ -3067,6 +3381,7 @@ class Program_weight_tensor_parameter_313: class Program_weight_tensor_parameter_314: name = "parameter_314" + original_name = "linear_11.w_0" shape = [192, 768] dtype = "float32" min_val = float("-0.413941") @@ -3078,6 +3393,7 @@ class Program_weight_tensor_parameter_314: class Program_weight_tensor_parameter_315: name = "parameter_315" + original_name = "layer_norm_7.b_0" shape = [192] dtype = "float32" min_val = float("-0.556142") @@ -3089,6 +3405,7 @@ class Program_weight_tensor_parameter_315: class Program_weight_tensor_parameter_316: name = "parameter_316" + original_name = "layer_norm_7.w_0" shape = [192] dtype = "float32" min_val = float("-0.00685324") @@ -3100,6 +3417,7 @@ class Program_weight_tensor_parameter_316: class Program_weight_tensor_parameter_317: name = "parameter_317" + original_name = "linear_10.b_0" shape = [192] dtype = "float32" min_val = float("-5.87772") @@ -3111,6 +3429,7 @@ class Program_weight_tensor_parameter_317: class Program_weight_tensor_parameter_318: name = "parameter_318" + original_name = "linear_10.w_0" shape = [192, 192] dtype = "float32" min_val = float("-0.384763") @@ -3122,6 +3441,7 @@ class Program_weight_tensor_parameter_318: class Program_weight_tensor_parameter_319: name = "parameter_319" + original_name = "linear_9.b_0" shape = [576] dtype = "float32" min_val = float("-2.53186") @@ -3133,6 +3453,7 @@ class Program_weight_tensor_parameter_319: class Program_weight_tensor_parameter_320: name = "parameter_320" + original_name = "linear_9.w_0" shape = [192, 576] dtype = "float32" min_val = float("-0.540943") @@ -3144,6 +3465,7 @@ class Program_weight_tensor_parameter_320: class Program_weight_tensor_parameter_321: name = "parameter_321" + original_name = "layer_norm_6.b_0" shape = [192] dtype = "float32" min_val = float("-0.329145") @@ -3155,6 +3477,7 @@ class Program_weight_tensor_parameter_321: class Program_weight_tensor_parameter_322: name = "parameter_322" + original_name = "layer_norm_6.w_0" shape = [192] dtype = "float32" min_val = float("-0.00330349") @@ -3166,6 +3489,7 @@ class Program_weight_tensor_parameter_322: class Program_weight_tensor_parameter_323: name = "parameter_323" + original_name = "layer_norm_28.b_0" shape = [96] dtype = "float32" data = None @@ -3173,6 +3497,7 @@ class Program_weight_tensor_parameter_323: class Program_weight_tensor_parameter_324: name = "parameter_324" + original_name = "layer_norm_28.w_0" shape = [96] dtype = "float32" min_val = float("1.0") @@ -3183,6 +3508,7 @@ class Program_weight_tensor_parameter_324: class Program_weight_tensor_parameter_325: name = "parameter_325" + original_name = "linear_8.w_0" shape = [384, 192] dtype = "float32" min_val = float("-0.412246") @@ -3194,6 +3520,7 @@ class Program_weight_tensor_parameter_325: class Program_weight_tensor_parameter_326: name = "parameter_326" + original_name = "layer_norm_5.b_0" shape = [384] dtype = "float32" min_val = float("-1.76269") @@ -3205,6 +3532,7 @@ class Program_weight_tensor_parameter_326: class Program_weight_tensor_parameter_327: name = "parameter_327" + original_name = "layer_norm_5.w_0" shape = [384] dtype = "float32" min_val = float("0.259424") @@ -3216,6 +3544,7 @@ class Program_weight_tensor_parameter_327: class Program_weight_tensor_parameter_328: name = "parameter_328" + original_name = "linear_7.b_0" shape = [96] dtype = "float32" min_val = float("-1.4916") @@ -3227,6 +3556,7 @@ class Program_weight_tensor_parameter_328: class Program_weight_tensor_parameter_329: name = "parameter_329" + original_name = "linear_7.w_0" shape = [384, 96] dtype = "float32" min_val = float("-0.654892") @@ -3238,6 +3568,7 @@ class Program_weight_tensor_parameter_329: class Program_weight_tensor_parameter_330: name = "parameter_330" + original_name = "linear_6.b_0" shape = [384] dtype = "float32" min_val = float("-3.89129") @@ -3249,6 +3580,7 @@ class Program_weight_tensor_parameter_330: class Program_weight_tensor_parameter_331: name = "parameter_331" + original_name = "linear_6.w_0" shape = [96, 384] dtype = "float32" min_val = float("-0.396774") @@ -3260,6 +3592,7 @@ class Program_weight_tensor_parameter_331: class Program_weight_tensor_parameter_332: name = "parameter_332" + original_name = "layer_norm_4.b_0" shape = [96] dtype = "float32" min_val = float("-2.51497") @@ -3271,6 +3604,7 @@ class Program_weight_tensor_parameter_332: class Program_weight_tensor_parameter_333: name = "parameter_333" + original_name = "layer_norm_4.w_0" shape = [96] dtype = "float32" min_val = float("-0.722583") @@ -3282,6 +3616,7 @@ class Program_weight_tensor_parameter_333: class Program_weight_tensor_parameter_334: name = "parameter_334" + original_name = "linear_5.b_0" shape = [96] dtype = "float32" min_val = float("-1.2732") @@ -3293,6 +3628,7 @@ class Program_weight_tensor_parameter_334: class Program_weight_tensor_parameter_335: name = "parameter_335" + original_name = "linear_5.w_0" shape = [96, 96] dtype = "float32" min_val = float("-0.312308") @@ -3304,6 +3640,7 @@ class Program_weight_tensor_parameter_335: class Program_weight_tensor_parameter_336: name = "parameter_336" + original_name = "linear_4.b_0" shape = [288] dtype = "float32" min_val = float("-4.67732") @@ -3315,6 +3652,7 @@ class Program_weight_tensor_parameter_336: class Program_weight_tensor_parameter_337: name = "parameter_337" + original_name = "linear_4.w_0" shape = [96, 288] dtype = "float32" min_val = float("-0.483863") @@ -3326,6 +3664,7 @@ class Program_weight_tensor_parameter_337: class Program_weight_tensor_parameter_338: name = "parameter_338" + original_name = "layer_norm_3.b_0" shape = [96] dtype = "float32" min_val = float("-1.93978") @@ -3337,6 +3676,7 @@ class Program_weight_tensor_parameter_338: class Program_weight_tensor_parameter_339: name = "parameter_339" + original_name = "layer_norm_3.w_0" shape = [96] dtype = "float32" min_val = float("0.458121") @@ -3348,6 +3688,7 @@ class Program_weight_tensor_parameter_339: class Program_weight_tensor_parameter_340: name = "parameter_340" + original_name = "linear_3.b_0" shape = [96] dtype = "float32" min_val = float("-2.85777") @@ -3359,6 +3700,7 @@ class Program_weight_tensor_parameter_340: class Program_weight_tensor_parameter_341: name = "parameter_341" + original_name = "linear_3.w_0" shape = [384, 96] dtype = "float32" min_val = float("-0.663805") @@ -3370,6 +3712,7 @@ class Program_weight_tensor_parameter_341: class Program_weight_tensor_parameter_342: name = "parameter_342" + original_name = "linear_2.b_0" shape = [384] dtype = "float32" min_val = float("-2.99421") @@ -3381,6 +3724,7 @@ class Program_weight_tensor_parameter_342: class Program_weight_tensor_parameter_343: name = "parameter_343" + original_name = "linear_2.w_0" shape = [96, 384] dtype = "float32" min_val = float("-0.49021") @@ -3392,6 +3736,7 @@ class Program_weight_tensor_parameter_343: class Program_weight_tensor_parameter_344: name = "parameter_344" + original_name = "layer_norm_2.b_0" shape = [96] dtype = "float32" min_val = float("-2.18691") @@ -3403,6 +3748,7 @@ class Program_weight_tensor_parameter_344: class Program_weight_tensor_parameter_345: name = "parameter_345" + original_name = "layer_norm_2.w_0" shape = [96] dtype = "float32" min_val = float("-0.00102264") @@ -3414,6 +3760,7 @@ class Program_weight_tensor_parameter_345: class Program_weight_tensor_parameter_346: name = "parameter_346" + original_name = "linear_1.b_0" shape = [96] dtype = "float32" min_val = float("-1.9372") @@ -3425,6 +3772,7 @@ class Program_weight_tensor_parameter_346: class Program_weight_tensor_parameter_347: name = "parameter_347" + original_name = "linear_1.w_0" shape = [96, 96] dtype = "float32" min_val = float("-0.447579") @@ -3436,6 +3784,7 @@ class Program_weight_tensor_parameter_347: class Program_weight_tensor_parameter_348: name = "parameter_348" + original_name = "linear_0.b_0" shape = [288] dtype = "float32" min_val = float("-4.66158") @@ -3447,6 +3796,7 @@ class Program_weight_tensor_parameter_348: class Program_weight_tensor_parameter_349: name = "parameter_349" + original_name = "linear_0.w_0" shape = [96, 288] dtype = "float32" min_val = float("-0.356196") @@ -3458,6 +3808,7 @@ class Program_weight_tensor_parameter_349: class Program_weight_tensor_parameter_350: name = "parameter_350" + original_name = "layer_norm_1.b_0" shape = [96] dtype = "float32" min_val = float("-1.75712") @@ -3469,6 +3820,7 @@ class Program_weight_tensor_parameter_350: class Program_weight_tensor_parameter_351: name = "parameter_351" + original_name = "layer_norm_1.w_0" shape = [96] dtype = "float32" min_val = float("-0.00132859") @@ -3480,6 +3832,7 @@ class Program_weight_tensor_parameter_351: class Program_weight_tensor_parameter_352: name = "parameter_352" + original_name = "layer_norm_0.b_0" shape = [96] dtype = "float32" min_val = float("-3.28992") @@ -3491,6 +3844,7 @@ class Program_weight_tensor_parameter_352: class Program_weight_tensor_parameter_353: name = "parameter_353" + original_name = "layer_norm_0.w_0" shape = [96] dtype = "float32" min_val = float("-0.0888211") @@ -3502,6 +3856,7 @@ class Program_weight_tensor_parameter_353: class Program_weight_tensor_parameter_354: name = "parameter_354" + original_name = "conv2d_0.b_0" shape = [96] dtype = "float32" min_val = float("-1.44354") @@ -3513,6 +3868,7 @@ class Program_weight_tensor_parameter_354: class Program_weight_tensor_parameter_355: name = "parameter_355" + original_name = "conv2d_0.w_0" shape = [96, 3, 4, 4] dtype = "float32" min_val = float("-0.283527") diff --git a/paddle_samples/PaddleX/MobileFaceNet/subgraph_0/input_meta.py b/paddle_samples/PaddleX/MobileFaceNet/subgraph_0/input_meta.py index 5565e68ed..ec58aac6d 100644 --- a/paddle_samples/PaddleX/MobileFaceNet/subgraph_0/input_meta.py +++ b/paddle_samples/PaddleX/MobileFaceNet/subgraph_0/input_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_data_0: name = "data_0" + original_name = "param_0" shape = [128, 93431] dtype = "float32" min_val = float("-0.410486") @@ -11,6 +12,7 @@ class Program_weight_tensor_data_0: class Program_weight_tensor_data_1: name = "data_1" + original_name = "var_0" shape = [128, 3, 112, 112] dtype = "float32" min_val = float("-1.0") @@ -22,6 +24,7 @@ class Program_weight_tensor_data_1: class Program_weight_tensor_data_2: name = "data_2" + original_name = "var_1" shape = [128, 1] dtype = "int64" min_val = 1 diff --git a/paddle_samples/PaddleX/MobileFaceNet/subgraph_0/weight_meta.py b/paddle_samples/PaddleX/MobileFaceNet/subgraph_0/weight_meta.py index 4eb999ff2..c7330d191 100644 --- a/paddle_samples/PaddleX/MobileFaceNet/subgraph_0/weight_meta.py +++ b/paddle_samples/PaddleX/MobileFaceNet/subgraph_0/weight_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_parameter_0: name = "parameter_0" + original_name = "create_parameter_149.w_0" shape = [128] dtype = "float32" min_val = float("-0.0479561") @@ -11,6 +12,7 @@ class Program_weight_tensor_parameter_0: class Program_weight_tensor_parameter_1: name = "parameter_1" + original_name = "create_parameter_148.w_0" shape = [128] dtype = "float32" min_val = float("1.36107") @@ -22,6 +24,7 @@ class Program_weight_tensor_parameter_1: class Program_weight_tensor_parameter_2: name = "parameter_2" + original_name = "batch_norm2d_49.w_2" shape = [128] dtype = "float32" min_val = float("6.37803") @@ -33,6 +36,7 @@ class Program_weight_tensor_parameter_2: class Program_weight_tensor_parameter_3: name = "parameter_3" + original_name = "batch_norm2d_49.w_1" shape = [128] dtype = "float32" min_val = float("-5.74265") @@ -44,6 +48,7 @@ class Program_weight_tensor_parameter_3: class Program_weight_tensor_parameter_4: name = "parameter_4" + original_name = "create_parameter_147.w_0" shape = [128, 512, 1, 1] dtype = "float32" min_val = float("-0.41632") @@ -55,6 +60,7 @@ class Program_weight_tensor_parameter_4: class Program_weight_tensor_parameter_5: name = "parameter_5" + original_name = "create_parameter_146.w_0" shape = [512] dtype = "float32" min_val = float("-2.27627") @@ -66,6 +72,7 @@ class Program_weight_tensor_parameter_5: class Program_weight_tensor_parameter_6: name = "parameter_6" + original_name = "create_parameter_145.w_0" shape = [512] dtype = "float32" min_val = float("-1.88216") @@ -77,6 +84,7 @@ class Program_weight_tensor_parameter_6: class Program_weight_tensor_parameter_7: name = "parameter_7" + original_name = "batch_norm2d_48.w_2" shape = [512] dtype = "float32" min_val = float("5.60519e-45") @@ -88,6 +96,7 @@ class Program_weight_tensor_parameter_7: class Program_weight_tensor_parameter_8: name = "parameter_8" + original_name = "batch_norm2d_48.w_1" shape = [512] dtype = "float32" min_val = float("-13.5753") @@ -99,6 +108,7 @@ class Program_weight_tensor_parameter_8: class Program_weight_tensor_parameter_9: name = "parameter_9" + original_name = "create_parameter_144.w_0" shape = [512, 1, 7, 7] dtype = "float32" min_val = float("-0.488965") @@ -110,6 +120,7 @@ class Program_weight_tensor_parameter_9: class Program_weight_tensor_parameter_10: name = "parameter_10" + original_name = "p_re_lu_32.w_0" shape = [512] dtype = "float32" min_val = float("-3.27426") @@ -121,6 +132,7 @@ class Program_weight_tensor_parameter_10: class Program_weight_tensor_parameter_11: name = "parameter_11" + original_name = "create_parameter_143.w_0" shape = [512] dtype = "float32" min_val = float("-2.25426") @@ -132,6 +144,7 @@ class Program_weight_tensor_parameter_11: class Program_weight_tensor_parameter_12: name = "parameter_12" + original_name = "create_parameter_142.w_0" shape = [512] dtype = "float32" min_val = float("-3.65712") @@ -143,6 +156,7 @@ class Program_weight_tensor_parameter_12: class Program_weight_tensor_parameter_13: name = "parameter_13" + original_name = "batch_norm2d_47.w_2" shape = [512] dtype = "float32" min_val = float("6.38599e-08") @@ -154,6 +168,7 @@ class Program_weight_tensor_parameter_13: class Program_weight_tensor_parameter_14: name = "parameter_14" + original_name = "batch_norm2d_47.w_1" shape = [512] dtype = "float32" min_val = float("-7.17799") @@ -165,6 +180,7 @@ class Program_weight_tensor_parameter_14: class Program_weight_tensor_parameter_15: name = "parameter_15" + original_name = "create_parameter_141.w_0" shape = [512, 128, 1, 1] dtype = "float32" min_val = float("-0.474424") @@ -176,6 +192,7 @@ class Program_weight_tensor_parameter_15: class Program_weight_tensor_parameter_16: name = "parameter_16" + original_name = "create_parameter_140.w_0" shape = [128] dtype = "float32" min_val = float("-2.47906") @@ -187,6 +204,7 @@ class Program_weight_tensor_parameter_16: class Program_weight_tensor_parameter_17: name = "parameter_17" + original_name = "create_parameter_139.w_0" shape = [128] dtype = "float32" min_val = float("0.462517") @@ -198,6 +216,7 @@ class Program_weight_tensor_parameter_17: class Program_weight_tensor_parameter_18: name = "parameter_18" + original_name = "batch_norm2d_46.w_2" shape = [128] dtype = "float32" min_val = float("0.48464") @@ -209,6 +228,7 @@ class Program_weight_tensor_parameter_18: class Program_weight_tensor_parameter_19: name = "parameter_19" + original_name = "batch_norm2d_46.w_1" shape = [128] dtype = "float32" min_val = float("-2.69658") @@ -220,6 +240,7 @@ class Program_weight_tensor_parameter_19: class Program_weight_tensor_parameter_20: name = "parameter_20" + original_name = "create_parameter_138.w_0" shape = [128, 256, 1, 1] dtype = "float32" min_val = float("-0.371552") @@ -231,6 +252,7 @@ class Program_weight_tensor_parameter_20: class Program_weight_tensor_parameter_21: name = "parameter_21" + original_name = "p_re_lu_31.w_0" shape = [256] dtype = "float32" min_val = float("-2.5338") @@ -242,6 +264,7 @@ class Program_weight_tensor_parameter_21: class Program_weight_tensor_parameter_22: name = "parameter_22" + original_name = "create_parameter_137.w_0" shape = [256] dtype = "float32" min_val = float("-5.91928") @@ -253,6 +276,7 @@ class Program_weight_tensor_parameter_22: class Program_weight_tensor_parameter_23: name = "parameter_23" + original_name = "create_parameter_136.w_0" shape = [256] dtype = "float32" min_val = float("0.279144") @@ -264,6 +288,7 @@ class Program_weight_tensor_parameter_23: class Program_weight_tensor_parameter_24: name = "parameter_24" + original_name = "batch_norm2d_45.w_2" shape = [256] dtype = "float32" min_val = float("0.000522155") @@ -275,6 +300,7 @@ class Program_weight_tensor_parameter_24: class Program_weight_tensor_parameter_25: name = "parameter_25" + original_name = "batch_norm2d_45.w_1" shape = [256] dtype = "float32" min_val = float("-1.61751") @@ -286,6 +312,7 @@ class Program_weight_tensor_parameter_25: class Program_weight_tensor_parameter_26: name = "parameter_26" + original_name = "create_parameter_135.w_0" shape = [256, 1, 3, 3] dtype = "float32" min_val = float("-0.737647") @@ -297,6 +324,7 @@ class Program_weight_tensor_parameter_26: class Program_weight_tensor_parameter_27: name = "parameter_27" + original_name = "p_re_lu_30.w_0" shape = [256] dtype = "float32" min_val = float("-25.4405") @@ -308,6 +336,7 @@ class Program_weight_tensor_parameter_27: class Program_weight_tensor_parameter_28: name = "parameter_28" + original_name = "create_parameter_134.w_0" shape = [256] dtype = "float32" min_val = float("-2.91458") @@ -319,6 +348,7 @@ class Program_weight_tensor_parameter_28: class Program_weight_tensor_parameter_29: name = "parameter_29" + original_name = "create_parameter_133.w_0" shape = [256] dtype = "float32" min_val = float("-0.39488") @@ -330,6 +360,7 @@ class Program_weight_tensor_parameter_29: class Program_weight_tensor_parameter_30: name = "parameter_30" + original_name = "batch_norm2d_44.w_2" shape = [256] dtype = "float32" min_val = float("1.80376") @@ -341,6 +372,7 @@ class Program_weight_tensor_parameter_30: class Program_weight_tensor_parameter_31: name = "parameter_31" + original_name = "batch_norm2d_44.w_1" shape = [256] dtype = "float32" min_val = float("-3.01711") @@ -352,6 +384,7 @@ class Program_weight_tensor_parameter_31: class Program_weight_tensor_parameter_32: name = "parameter_32" + original_name = "create_parameter_132.w_0" shape = [256, 128, 1, 1] dtype = "float32" min_val = float("-0.412343") @@ -363,6 +396,7 @@ class Program_weight_tensor_parameter_32: class Program_weight_tensor_parameter_33: name = "parameter_33" + original_name = "create_parameter_131.w_0" shape = [128] dtype = "float32" min_val = float("-2.02579") @@ -374,6 +408,7 @@ class Program_weight_tensor_parameter_33: class Program_weight_tensor_parameter_34: name = "parameter_34" + original_name = "create_parameter_130.w_0" shape = [128] dtype = "float32" min_val = float("0.548783") @@ -385,6 +420,7 @@ class Program_weight_tensor_parameter_34: class Program_weight_tensor_parameter_35: name = "parameter_35" + original_name = "batch_norm2d_43.w_2" shape = [128] dtype = "float32" min_val = float("0.5086") @@ -396,6 +432,7 @@ class Program_weight_tensor_parameter_35: class Program_weight_tensor_parameter_36: name = "parameter_36" + original_name = "batch_norm2d_43.w_1" shape = [128] dtype = "float32" min_val = float("-2.29034") @@ -407,6 +444,7 @@ class Program_weight_tensor_parameter_36: class Program_weight_tensor_parameter_37: name = "parameter_37" + original_name = "create_parameter_129.w_0" shape = [128, 256, 1, 1] dtype = "float32" min_val = float("-0.34896") @@ -418,6 +456,7 @@ class Program_weight_tensor_parameter_37: class Program_weight_tensor_parameter_38: name = "parameter_38" + original_name = "p_re_lu_29.w_0" shape = [256] dtype = "float32" min_val = float("-2.96755") @@ -429,6 +468,7 @@ class Program_weight_tensor_parameter_38: class Program_weight_tensor_parameter_39: name = "parameter_39" + original_name = "create_parameter_128.w_0" shape = [256] dtype = "float32" min_val = float("-7.93746") @@ -440,6 +480,7 @@ class Program_weight_tensor_parameter_39: class Program_weight_tensor_parameter_40: name = "parameter_40" + original_name = "create_parameter_127.w_0" shape = [256] dtype = "float32" min_val = float("-0.0284306") @@ -451,6 +492,7 @@ class Program_weight_tensor_parameter_40: class Program_weight_tensor_parameter_41: name = "parameter_41" + original_name = "batch_norm2d_42.w_2" shape = [256] dtype = "float32" min_val = float("0.000773831") @@ -462,6 +504,7 @@ class Program_weight_tensor_parameter_41: class Program_weight_tensor_parameter_42: name = "parameter_42" + original_name = "batch_norm2d_42.w_1" shape = [256] dtype = "float32" min_val = float("-0.712629") @@ -473,6 +516,7 @@ class Program_weight_tensor_parameter_42: class Program_weight_tensor_parameter_43: name = "parameter_43" + original_name = "create_parameter_126.w_0" shape = [256, 1, 3, 3] dtype = "float32" min_val = float("-0.403541") @@ -484,6 +528,7 @@ class Program_weight_tensor_parameter_43: class Program_weight_tensor_parameter_44: name = "parameter_44" + original_name = "p_re_lu_28.w_0" shape = [256] dtype = "float32" min_val = float("-9.21684") @@ -495,6 +540,7 @@ class Program_weight_tensor_parameter_44: class Program_weight_tensor_parameter_45: name = "parameter_45" + original_name = "create_parameter_125.w_0" shape = [256] dtype = "float32" min_val = float("-1.40624") @@ -506,6 +552,7 @@ class Program_weight_tensor_parameter_45: class Program_weight_tensor_parameter_46: name = "parameter_46" + original_name = "create_parameter_124.w_0" shape = [256] dtype = "float32" min_val = float("-0.727924") @@ -517,6 +564,7 @@ class Program_weight_tensor_parameter_46: class Program_weight_tensor_parameter_47: name = "parameter_47" + original_name = "batch_norm2d_41.w_2" shape = [256] dtype = "float32" min_val = float("0.457189") @@ -528,6 +576,7 @@ class Program_weight_tensor_parameter_47: class Program_weight_tensor_parameter_48: name = "parameter_48" + original_name = "batch_norm2d_41.w_1" shape = [256] dtype = "float32" min_val = float("-1.72527") @@ -539,6 +588,7 @@ class Program_weight_tensor_parameter_48: class Program_weight_tensor_parameter_49: name = "parameter_49" + original_name = "create_parameter_123.w_0" shape = [256, 128, 1, 1] dtype = "float32" min_val = float("-0.322308") @@ -550,6 +600,7 @@ class Program_weight_tensor_parameter_49: class Program_weight_tensor_parameter_50: name = "parameter_50" + original_name = "create_parameter_122.w_0" shape = [128] dtype = "float32" min_val = float("-1.64941") @@ -561,6 +612,7 @@ class Program_weight_tensor_parameter_50: class Program_weight_tensor_parameter_51: name = "parameter_51" + original_name = "create_parameter_121.w_0" shape = [128] dtype = "float32" min_val = float("0.930199") @@ -572,6 +624,7 @@ class Program_weight_tensor_parameter_51: class Program_weight_tensor_parameter_52: name = "parameter_52" + original_name = "batch_norm2d_40.w_2" shape = [128] dtype = "float32" min_val = float("2.21887") @@ -583,6 +636,7 @@ class Program_weight_tensor_parameter_52: class Program_weight_tensor_parameter_53: name = "parameter_53" + original_name = "batch_norm2d_40.w_1" shape = [128] dtype = "float32" min_val = float("-5.62287") @@ -594,6 +648,7 @@ class Program_weight_tensor_parameter_53: class Program_weight_tensor_parameter_54: name = "parameter_54" + original_name = "create_parameter_120.w_0" shape = [128, 512, 1, 1] dtype = "float32" min_val = float("-0.332522") @@ -605,6 +660,7 @@ class Program_weight_tensor_parameter_54: class Program_weight_tensor_parameter_55: name = "parameter_55" + original_name = "p_re_lu_27.w_0" shape = [512] dtype = "float32" min_val = float("-1.7556") @@ -616,6 +672,7 @@ class Program_weight_tensor_parameter_55: class Program_weight_tensor_parameter_56: name = "parameter_56" + original_name = "create_parameter_119.w_0" shape = [512] dtype = "float32" min_val = float("-9.48434") @@ -627,6 +684,7 @@ class Program_weight_tensor_parameter_56: class Program_weight_tensor_parameter_57: name = "parameter_57" + original_name = "create_parameter_118.w_0" shape = [512] dtype = "float32" min_val = float("0.489082") @@ -638,6 +696,7 @@ class Program_weight_tensor_parameter_57: class Program_weight_tensor_parameter_58: name = "parameter_58" + original_name = "batch_norm2d_39.w_2" shape = [512] dtype = "float32" min_val = float("0.000275093") @@ -649,6 +708,7 @@ class Program_weight_tensor_parameter_58: class Program_weight_tensor_parameter_59: name = "parameter_59" + original_name = "batch_norm2d_39.w_1" shape = [512] dtype = "float32" min_val = float("-1.02003") @@ -660,6 +720,7 @@ class Program_weight_tensor_parameter_59: class Program_weight_tensor_parameter_60: name = "parameter_60" + original_name = "create_parameter_117.w_0" shape = [512, 1, 3, 3] dtype = "float32" min_val = float("-0.333892") @@ -671,6 +732,7 @@ class Program_weight_tensor_parameter_60: class Program_weight_tensor_parameter_61: name = "parameter_61" + original_name = "p_re_lu_26.w_0" shape = [512] dtype = "float32" min_val = float("-7.22925") @@ -682,6 +744,7 @@ class Program_weight_tensor_parameter_61: class Program_weight_tensor_parameter_62: name = "parameter_62" + original_name = "create_parameter_116.w_0" shape = [512] dtype = "float32" min_val = float("-1.57395") @@ -693,6 +756,7 @@ class Program_weight_tensor_parameter_62: class Program_weight_tensor_parameter_63: name = "parameter_63" + original_name = "create_parameter_115.w_0" shape = [512] dtype = "float32" min_val = float("-0.253051") @@ -704,6 +768,7 @@ class Program_weight_tensor_parameter_63: class Program_weight_tensor_parameter_64: name = "parameter_64" + original_name = "batch_norm2d_38.w_2" shape = [512] dtype = "float32" min_val = float("6.67056") @@ -715,6 +780,7 @@ class Program_weight_tensor_parameter_64: class Program_weight_tensor_parameter_65: name = "parameter_65" + original_name = "batch_norm2d_38.w_1" shape = [512] dtype = "float32" min_val = float("-14.1001") @@ -726,6 +792,7 @@ class Program_weight_tensor_parameter_65: class Program_weight_tensor_parameter_66: name = "parameter_66" + original_name = "create_parameter_114.w_0" shape = [512, 128, 1, 1] dtype = "float32" min_val = float("-0.395576") @@ -737,6 +804,7 @@ class Program_weight_tensor_parameter_66: class Program_weight_tensor_parameter_67: name = "parameter_67" + original_name = "create_parameter_113.w_0" shape = [128] dtype = "float32" min_val = float("-2.37154") @@ -748,6 +816,7 @@ class Program_weight_tensor_parameter_67: class Program_weight_tensor_parameter_68: name = "parameter_68" + original_name = "create_parameter_112.w_0" shape = [128] dtype = "float32" min_val = float("1.10453") @@ -759,6 +828,7 @@ class Program_weight_tensor_parameter_68: class Program_weight_tensor_parameter_69: name = "parameter_69" + original_name = "batch_norm2d_37.w_2" shape = [128] dtype = "float32" min_val = float("0.790137") @@ -770,6 +840,7 @@ class Program_weight_tensor_parameter_69: class Program_weight_tensor_parameter_70: name = "parameter_70" + original_name = "batch_norm2d_37.w_1" shape = [128] dtype = "float32" min_val = float("-2.78433") @@ -781,6 +852,7 @@ class Program_weight_tensor_parameter_70: class Program_weight_tensor_parameter_71: name = "parameter_71" + original_name = "create_parameter_111.w_0" shape = [128, 256, 1, 1] dtype = "float32" min_val = float("-0.324941") @@ -792,6 +864,7 @@ class Program_weight_tensor_parameter_71: class Program_weight_tensor_parameter_72: name = "parameter_72" + original_name = "p_re_lu_25.w_0" shape = [256] dtype = "float32" min_val = float("-3.24905") @@ -803,6 +876,7 @@ class Program_weight_tensor_parameter_72: class Program_weight_tensor_parameter_73: name = "parameter_73" + original_name = "create_parameter_110.w_0" shape = [256] dtype = "float32" min_val = float("-4.17003") @@ -814,6 +888,7 @@ class Program_weight_tensor_parameter_73: class Program_weight_tensor_parameter_74: name = "parameter_74" + original_name = "create_parameter_109.w_0" shape = [256] dtype = "float32" min_val = float("0.289089") @@ -825,6 +900,7 @@ class Program_weight_tensor_parameter_74: class Program_weight_tensor_parameter_75: name = "parameter_75" + original_name = "batch_norm2d_36.w_2" shape = [256] dtype = "float32" min_val = float("0.00048774") @@ -836,6 +912,7 @@ class Program_weight_tensor_parameter_75: class Program_weight_tensor_parameter_76: name = "parameter_76" + original_name = "batch_norm2d_36.w_1" shape = [256] dtype = "float32" min_val = float("-1.58586") @@ -847,6 +924,7 @@ class Program_weight_tensor_parameter_76: class Program_weight_tensor_parameter_77: name = "parameter_77" + original_name = "create_parameter_108.w_0" shape = [256, 1, 3, 3] dtype = "float32" min_val = float("-0.707676") @@ -858,6 +936,7 @@ class Program_weight_tensor_parameter_77: class Program_weight_tensor_parameter_78: name = "parameter_78" + original_name = "p_re_lu_24.w_0" shape = [256] dtype = "float32" min_val = float("-11.3139") @@ -869,6 +948,7 @@ class Program_weight_tensor_parameter_78: class Program_weight_tensor_parameter_79: name = "parameter_79" + original_name = "create_parameter_107.w_0" shape = [256] dtype = "float32" min_val = float("-1.82379") @@ -880,6 +960,7 @@ class Program_weight_tensor_parameter_79: class Program_weight_tensor_parameter_80: name = "parameter_80" + original_name = "create_parameter_106.w_0" shape = [256] dtype = "float32" min_val = float("0.359466") @@ -891,6 +972,7 @@ class Program_weight_tensor_parameter_80: class Program_weight_tensor_parameter_81: name = "parameter_81" + original_name = "batch_norm2d_35.w_2" shape = [256] dtype = "float32" min_val = float("5.59852") @@ -902,6 +984,7 @@ class Program_weight_tensor_parameter_81: class Program_weight_tensor_parameter_82: name = "parameter_82" + original_name = "batch_norm2d_35.w_1" shape = [256] dtype = "float32" min_val = float("-11.2864") @@ -913,6 +996,7 @@ class Program_weight_tensor_parameter_82: class Program_weight_tensor_parameter_83: name = "parameter_83" + original_name = "create_parameter_105.w_0" shape = [256, 128, 1, 1] dtype = "float32" min_val = float("-0.361286") @@ -924,6 +1008,7 @@ class Program_weight_tensor_parameter_83: class Program_weight_tensor_parameter_84: name = "parameter_84" + original_name = "create_parameter_104.w_0" shape = [128] dtype = "float32" min_val = float("-2.38519") @@ -935,6 +1020,7 @@ class Program_weight_tensor_parameter_84: class Program_weight_tensor_parameter_85: name = "parameter_85" + original_name = "create_parameter_103.w_0" shape = [128] dtype = "float32" min_val = float("0.897397") @@ -946,6 +1032,7 @@ class Program_weight_tensor_parameter_85: class Program_weight_tensor_parameter_86: name = "parameter_86" + original_name = "batch_norm2d_34.w_2" shape = [128] dtype = "float32" min_val = float("0.77063") @@ -957,6 +1044,7 @@ class Program_weight_tensor_parameter_86: class Program_weight_tensor_parameter_87: name = "parameter_87" + original_name = "batch_norm2d_34.w_1" shape = [128] dtype = "float32" min_val = float("-3.01662") @@ -968,6 +1056,7 @@ class Program_weight_tensor_parameter_87: class Program_weight_tensor_parameter_88: name = "parameter_88" + original_name = "create_parameter_102.w_0" shape = [128, 256, 1, 1] dtype = "float32" min_val = float("-0.330042") @@ -979,6 +1068,7 @@ class Program_weight_tensor_parameter_88: class Program_weight_tensor_parameter_89: name = "parameter_89" + original_name = "p_re_lu_23.w_0" shape = [256] dtype = "float32" min_val = float("-2.36374") @@ -990,6 +1080,7 @@ class Program_weight_tensor_parameter_89: class Program_weight_tensor_parameter_90: name = "parameter_90" + original_name = "create_parameter_101.w_0" shape = [256] dtype = "float32" min_val = float("-5.71402") @@ -1001,6 +1092,7 @@ class Program_weight_tensor_parameter_90: class Program_weight_tensor_parameter_91: name = "parameter_91" + original_name = "create_parameter_100.w_0" shape = [256] dtype = "float32" min_val = float("0.349548") @@ -1012,6 +1104,7 @@ class Program_weight_tensor_parameter_91: class Program_weight_tensor_parameter_92: name = "parameter_92" + original_name = "batch_norm2d_33.w_2" shape = [256] dtype = "float32" min_val = float("0.000875338") @@ -1023,6 +1116,7 @@ class Program_weight_tensor_parameter_92: class Program_weight_tensor_parameter_93: name = "parameter_93" + original_name = "batch_norm2d_33.w_1" shape = [256] dtype = "float32" min_val = float("-3.13102") @@ -1034,6 +1128,7 @@ class Program_weight_tensor_parameter_93: class Program_weight_tensor_parameter_94: name = "parameter_94" + original_name = "create_parameter_99.w_0" shape = [256, 1, 3, 3] dtype = "float32" min_val = float("-0.64319") @@ -1045,6 +1140,7 @@ class Program_weight_tensor_parameter_94: class Program_weight_tensor_parameter_95: name = "parameter_95" + original_name = "p_re_lu_22.w_0" shape = [256] dtype = "float32" min_val = float("-7.46313") @@ -1056,6 +1152,7 @@ class Program_weight_tensor_parameter_95: class Program_weight_tensor_parameter_96: name = "parameter_96" + original_name = "create_parameter_98.w_0" shape = [256] dtype = "float32" min_val = float("-1.5107") @@ -1067,6 +1164,7 @@ class Program_weight_tensor_parameter_96: class Program_weight_tensor_parameter_97: name = "parameter_97" + original_name = "create_parameter_97.w_0" shape = [256] dtype = "float32" min_val = float("0.461346") @@ -1078,6 +1176,7 @@ class Program_weight_tensor_parameter_97: class Program_weight_tensor_parameter_98: name = "parameter_98" + original_name = "batch_norm2d_32.w_2" shape = [256] dtype = "float32" min_val = float("4.03328") @@ -1089,6 +1188,7 @@ class Program_weight_tensor_parameter_98: class Program_weight_tensor_parameter_99: name = "parameter_99" + original_name = "batch_norm2d_32.w_1" shape = [256] dtype = "float32" min_val = float("-9.55226") @@ -1100,6 +1200,7 @@ class Program_weight_tensor_parameter_99: class Program_weight_tensor_parameter_100: name = "parameter_100" + original_name = "create_parameter_96.w_0" shape = [256, 128, 1, 1] dtype = "float32" min_val = float("-0.331624") @@ -1111,6 +1212,7 @@ class Program_weight_tensor_parameter_100: class Program_weight_tensor_parameter_101: name = "parameter_101" + original_name = "create_parameter_95.w_0" shape = [128] dtype = "float32" min_val = float("-2.44459") @@ -1122,6 +1224,7 @@ class Program_weight_tensor_parameter_101: class Program_weight_tensor_parameter_102: name = "parameter_102" + original_name = "create_parameter_94.w_0" shape = [128] dtype = "float32" min_val = float("0.717076") @@ -1133,6 +1236,7 @@ class Program_weight_tensor_parameter_102: class Program_weight_tensor_parameter_103: name = "parameter_103" + original_name = "batch_norm2d_31.w_2" shape = [128] dtype = "float32" min_val = float("0.807336") @@ -1144,6 +1248,7 @@ class Program_weight_tensor_parameter_103: class Program_weight_tensor_parameter_104: name = "parameter_104" + original_name = "batch_norm2d_31.w_1" shape = [128] dtype = "float32" min_val = float("-3.13433") @@ -1155,6 +1260,7 @@ class Program_weight_tensor_parameter_104: class Program_weight_tensor_parameter_105: name = "parameter_105" + original_name = "create_parameter_93.w_0" shape = [128, 256, 1, 1] dtype = "float32" min_val = float("-0.383748") @@ -1166,6 +1272,7 @@ class Program_weight_tensor_parameter_105: class Program_weight_tensor_parameter_106: name = "parameter_106" + original_name = "p_re_lu_21.w_0" shape = [256] dtype = "float32" min_val = float("-2.57138") @@ -1177,6 +1284,7 @@ class Program_weight_tensor_parameter_106: class Program_weight_tensor_parameter_107: name = "parameter_107" + original_name = "create_parameter_92.w_0" shape = [256] dtype = "float32" min_val = float("-3.1317") @@ -1188,6 +1296,7 @@ class Program_weight_tensor_parameter_107: class Program_weight_tensor_parameter_108: name = "parameter_108" + original_name = "create_parameter_91.w_0" shape = [256] dtype = "float32" min_val = float("0.329055") @@ -1199,6 +1308,7 @@ class Program_weight_tensor_parameter_108: class Program_weight_tensor_parameter_109: name = "parameter_109" + original_name = "batch_norm2d_30.w_2" shape = [256] dtype = "float32" min_val = float("0.000892741") @@ -1210,6 +1320,7 @@ class Program_weight_tensor_parameter_109: class Program_weight_tensor_parameter_110: name = "parameter_110" + original_name = "batch_norm2d_30.w_1" shape = [256] dtype = "float32" min_val = float("-2.10856") @@ -1221,6 +1332,7 @@ class Program_weight_tensor_parameter_110: class Program_weight_tensor_parameter_111: name = "parameter_111" + original_name = "create_parameter_90.w_0" shape = [256, 1, 3, 3] dtype = "float32" min_val = float("-0.661179") @@ -1232,6 +1344,7 @@ class Program_weight_tensor_parameter_111: class Program_weight_tensor_parameter_112: name = "parameter_112" + original_name = "p_re_lu_20.w_0" shape = [256] dtype = "float32" min_val = float("-9.84087") @@ -1243,6 +1356,7 @@ class Program_weight_tensor_parameter_112: class Program_weight_tensor_parameter_113: name = "parameter_113" + original_name = "create_parameter_89.w_0" shape = [256] dtype = "float32" min_val = float("-1.5471") @@ -1254,6 +1368,7 @@ class Program_weight_tensor_parameter_113: class Program_weight_tensor_parameter_114: name = "parameter_114" + original_name = "create_parameter_88.w_0" shape = [256] dtype = "float32" min_val = float("-0.57587") @@ -1265,6 +1380,7 @@ class Program_weight_tensor_parameter_114: class Program_weight_tensor_parameter_115: name = "parameter_115" + original_name = "batch_norm2d_29.w_2" shape = [256] dtype = "float32" min_val = float("2.74716") @@ -1276,6 +1392,7 @@ class Program_weight_tensor_parameter_115: class Program_weight_tensor_parameter_116: name = "parameter_116" + original_name = "batch_norm2d_29.w_1" shape = [256] dtype = "float32" min_val = float("-10.4925") @@ -1287,6 +1404,7 @@ class Program_weight_tensor_parameter_116: class Program_weight_tensor_parameter_117: name = "parameter_117" + original_name = "create_parameter_87.w_0" shape = [256, 128, 1, 1] dtype = "float32" min_val = float("-0.361988") @@ -1298,6 +1416,7 @@ class Program_weight_tensor_parameter_117: class Program_weight_tensor_parameter_118: name = "parameter_118" + original_name = "create_parameter_86.w_0" shape = [128] dtype = "float32" min_val = float("-2.68412") @@ -1309,6 +1428,7 @@ class Program_weight_tensor_parameter_118: class Program_weight_tensor_parameter_119: name = "parameter_119" + original_name = "create_parameter_85.w_0" shape = [128] dtype = "float32" min_val = float("0.587941") @@ -1320,6 +1440,7 @@ class Program_weight_tensor_parameter_119: class Program_weight_tensor_parameter_120: name = "parameter_120" + original_name = "batch_norm2d_28.w_2" shape = [128] dtype = "float32" min_val = float("0.614758") @@ -1331,6 +1452,7 @@ class Program_weight_tensor_parameter_120: class Program_weight_tensor_parameter_121: name = "parameter_121" + original_name = "batch_norm2d_28.w_1" shape = [128] dtype = "float32" min_val = float("-2.80184") @@ -1342,6 +1464,7 @@ class Program_weight_tensor_parameter_121: class Program_weight_tensor_parameter_122: name = "parameter_122" + original_name = "create_parameter_84.w_0" shape = [128, 256, 1, 1] dtype = "float32" min_val = float("-0.417326") @@ -1353,6 +1476,7 @@ class Program_weight_tensor_parameter_122: class Program_weight_tensor_parameter_123: name = "parameter_123" + original_name = "p_re_lu_19.w_0" shape = [256] dtype = "float32" min_val = float("-1.60386") @@ -1364,6 +1488,7 @@ class Program_weight_tensor_parameter_123: class Program_weight_tensor_parameter_124: name = "parameter_124" + original_name = "create_parameter_83.w_0" shape = [256] dtype = "float32" min_val = float("-4.05393") @@ -1375,6 +1500,7 @@ class Program_weight_tensor_parameter_124: class Program_weight_tensor_parameter_125: name = "parameter_125" + original_name = "create_parameter_82.w_0" shape = [256] dtype = "float32" min_val = float("-1.54244") @@ -1386,6 +1512,7 @@ class Program_weight_tensor_parameter_125: class Program_weight_tensor_parameter_126: name = "parameter_126" + original_name = "batch_norm2d_27.w_2" shape = [256] dtype = "float32" min_val = float("2.42211e-08") @@ -1397,6 +1524,7 @@ class Program_weight_tensor_parameter_126: class Program_weight_tensor_parameter_127: name = "parameter_127" + original_name = "batch_norm2d_27.w_1" shape = [256] dtype = "float32" min_val = float("-1.8864") @@ -1408,6 +1536,7 @@ class Program_weight_tensor_parameter_127: class Program_weight_tensor_parameter_128: name = "parameter_128" + original_name = "create_parameter_81.w_0" shape = [256, 1, 3, 3] dtype = "float32" min_val = float("-0.60388") @@ -1419,6 +1548,7 @@ class Program_weight_tensor_parameter_128: class Program_weight_tensor_parameter_129: name = "parameter_129" + original_name = "p_re_lu_18.w_0" shape = [256] dtype = "float32" min_val = float("-11.9168") @@ -1430,6 +1560,7 @@ class Program_weight_tensor_parameter_129: class Program_weight_tensor_parameter_130: name = "parameter_130" + original_name = "create_parameter_80.w_0" shape = [256] dtype = "float32" min_val = float("-1.84185") @@ -1441,6 +1572,7 @@ class Program_weight_tensor_parameter_130: class Program_weight_tensor_parameter_131: name = "parameter_131" + original_name = "create_parameter_79.w_0" shape = [256] dtype = "float32" min_val = float("-0.0172566") @@ -1452,6 +1584,7 @@ class Program_weight_tensor_parameter_131: class Program_weight_tensor_parameter_132: name = "parameter_132" + original_name = "batch_norm2d_26.w_2" shape = [256] dtype = "float32" min_val = float("1.72376") @@ -1463,6 +1596,7 @@ class Program_weight_tensor_parameter_132: class Program_weight_tensor_parameter_133: name = "parameter_133" + original_name = "batch_norm2d_26.w_1" shape = [256] dtype = "float32" min_val = float("-6.25629") @@ -1474,6 +1608,7 @@ class Program_weight_tensor_parameter_133: class Program_weight_tensor_parameter_134: name = "parameter_134" + original_name = "create_parameter_78.w_0" shape = [256, 128, 1, 1] dtype = "float32" min_val = float("-0.316146") @@ -1485,6 +1620,7 @@ class Program_weight_tensor_parameter_134: class Program_weight_tensor_parameter_135: name = "parameter_135" + original_name = "create_parameter_77.w_0" shape = [128] dtype = "float32" min_val = float("-2.42536") @@ -1496,6 +1632,7 @@ class Program_weight_tensor_parameter_135: class Program_weight_tensor_parameter_136: name = "parameter_136" + original_name = "create_parameter_76.w_0" shape = [128] dtype = "float32" min_val = float("0.530225") @@ -1507,6 +1644,7 @@ class Program_weight_tensor_parameter_136: class Program_weight_tensor_parameter_137: name = "parameter_137" + original_name = "batch_norm2d_25.w_2" shape = [128] dtype = "float32" min_val = float("0.987776") @@ -1518,6 +1656,7 @@ class Program_weight_tensor_parameter_137: class Program_weight_tensor_parameter_138: name = "parameter_138" + original_name = "batch_norm2d_25.w_1" shape = [128] dtype = "float32" min_val = float("-3.22718") @@ -1529,6 +1668,7 @@ class Program_weight_tensor_parameter_138: class Program_weight_tensor_parameter_139: name = "parameter_139" + original_name = "create_parameter_75.w_0" shape = [128, 256, 1, 1] dtype = "float32" min_val = float("-0.360028") @@ -1540,6 +1680,7 @@ class Program_weight_tensor_parameter_139: class Program_weight_tensor_parameter_140: name = "parameter_140" + original_name = "p_re_lu_17.w_0" shape = [256] dtype = "float32" min_val = float("-1.76538") @@ -1551,6 +1692,7 @@ class Program_weight_tensor_parameter_140: class Program_weight_tensor_parameter_141: name = "parameter_141" + original_name = "create_parameter_74.w_0" shape = [256] dtype = "float32" min_val = float("-2.85406") @@ -1562,6 +1704,7 @@ class Program_weight_tensor_parameter_141: class Program_weight_tensor_parameter_142: name = "parameter_142" + original_name = "create_parameter_73.w_0" shape = [256] dtype = "float32" min_val = float("0.386561") @@ -1573,6 +1716,7 @@ class Program_weight_tensor_parameter_142: class Program_weight_tensor_parameter_143: name = "parameter_143" + original_name = "batch_norm2d_24.w_2" shape = [256] dtype = "float32" min_val = float("0.000447341") @@ -1584,6 +1728,7 @@ class Program_weight_tensor_parameter_143: class Program_weight_tensor_parameter_144: name = "parameter_144" + original_name = "batch_norm2d_24.w_1" shape = [256] dtype = "float32" min_val = float("-1.04021") @@ -1595,6 +1740,7 @@ class Program_weight_tensor_parameter_144: class Program_weight_tensor_parameter_145: name = "parameter_145" + original_name = "create_parameter_72.w_0" shape = [256, 1, 3, 3] dtype = "float32" min_val = float("-0.40418") @@ -1606,6 +1752,7 @@ class Program_weight_tensor_parameter_145: class Program_weight_tensor_parameter_146: name = "parameter_146" + original_name = "p_re_lu_16.w_0" shape = [256] dtype = "float32" min_val = float("-5.60664") @@ -1617,6 +1764,7 @@ class Program_weight_tensor_parameter_146: class Program_weight_tensor_parameter_147: name = "parameter_147" + original_name = "create_parameter_71.w_0" shape = [256] dtype = "float32" min_val = float("-1.73164") @@ -1628,6 +1776,7 @@ class Program_weight_tensor_parameter_147: class Program_weight_tensor_parameter_148: name = "parameter_148" + original_name = "create_parameter_70.w_0" shape = [256] dtype = "float32" min_val = float("0.515911") @@ -1639,6 +1788,7 @@ class Program_weight_tensor_parameter_148: class Program_weight_tensor_parameter_149: name = "parameter_149" + original_name = "batch_norm2d_23.w_2" shape = [256] dtype = "float32" min_val = float("1.32477") @@ -1650,6 +1800,7 @@ class Program_weight_tensor_parameter_149: class Program_weight_tensor_parameter_150: name = "parameter_150" + original_name = "batch_norm2d_23.w_1" shape = [256] dtype = "float32" min_val = float("-4.1593") @@ -1661,6 +1812,7 @@ class Program_weight_tensor_parameter_150: class Program_weight_tensor_parameter_151: name = "parameter_151" + original_name = "create_parameter_69.w_0" shape = [256, 128, 1, 1] dtype = "float32" min_val = float("-0.352573") @@ -1672,6 +1824,7 @@ class Program_weight_tensor_parameter_151: class Program_weight_tensor_parameter_152: name = "parameter_152" + original_name = "create_parameter_68.w_0" shape = [128] dtype = "float32" min_val = float("-1.96956") @@ -1683,6 +1836,7 @@ class Program_weight_tensor_parameter_152: class Program_weight_tensor_parameter_153: name = "parameter_153" + original_name = "create_parameter_67.w_0" shape = [128] dtype = "float32" min_val = float("0.530148") @@ -1694,6 +1848,7 @@ class Program_weight_tensor_parameter_153: class Program_weight_tensor_parameter_154: name = "parameter_154" + original_name = "batch_norm2d_22.w_2" shape = [128] dtype = "float32" min_val = float("1.01401") @@ -1705,6 +1860,7 @@ class Program_weight_tensor_parameter_154: class Program_weight_tensor_parameter_155: name = "parameter_155" + original_name = "batch_norm2d_22.w_1" shape = [128] dtype = "float32" min_val = float("-4.36325") @@ -1716,6 +1872,7 @@ class Program_weight_tensor_parameter_155: class Program_weight_tensor_parameter_156: name = "parameter_156" + original_name = "create_parameter_66.w_0" shape = [128, 256, 1, 1] dtype = "float32" min_val = float("-0.334618") @@ -1727,6 +1884,7 @@ class Program_weight_tensor_parameter_156: class Program_weight_tensor_parameter_157: name = "parameter_157" + original_name = "p_re_lu_15.w_0" shape = [256] dtype = "float32" min_val = float("-2.27078") @@ -1738,6 +1896,7 @@ class Program_weight_tensor_parameter_157: class Program_weight_tensor_parameter_158: name = "parameter_158" + original_name = "create_parameter_65.w_0" shape = [256] dtype = "float32" min_val = float("-2.31758") @@ -1749,6 +1908,7 @@ class Program_weight_tensor_parameter_158: class Program_weight_tensor_parameter_159: name = "parameter_159" + original_name = "create_parameter_64.w_0" shape = [256] dtype = "float32" min_val = float("0.575027") @@ -1760,6 +1920,7 @@ class Program_weight_tensor_parameter_159: class Program_weight_tensor_parameter_160: name = "parameter_160" + original_name = "batch_norm2d_21.w_2" shape = [256] dtype = "float32" min_val = float("0.00639319") @@ -1771,6 +1932,7 @@ class Program_weight_tensor_parameter_160: class Program_weight_tensor_parameter_161: name = "parameter_161" + original_name = "batch_norm2d_21.w_1" shape = [256] dtype = "float32" min_val = float("-1.14366") @@ -1782,6 +1944,7 @@ class Program_weight_tensor_parameter_161: class Program_weight_tensor_parameter_162: name = "parameter_162" + original_name = "create_parameter_63.w_0" shape = [256, 1, 3, 3] dtype = "float32" min_val = float("-0.395838") @@ -1793,6 +1956,7 @@ class Program_weight_tensor_parameter_162: class Program_weight_tensor_parameter_163: name = "parameter_163" + original_name = "p_re_lu_14.w_0" shape = [256] dtype = "float32" min_val = float("-1.36936") @@ -1804,6 +1968,7 @@ class Program_weight_tensor_parameter_163: class Program_weight_tensor_parameter_164: name = "parameter_164" + original_name = "create_parameter_62.w_0" shape = [256] dtype = "float32" min_val = float("-1.4657") @@ -1815,6 +1980,7 @@ class Program_weight_tensor_parameter_164: class Program_weight_tensor_parameter_165: name = "parameter_165" + original_name = "create_parameter_61.w_0" shape = [256] dtype = "float32" min_val = float("0.227443") @@ -1826,6 +1992,7 @@ class Program_weight_tensor_parameter_165: class Program_weight_tensor_parameter_166: name = "parameter_166" + original_name = "batch_norm2d_20.w_2" shape = [256] dtype = "float32" min_val = float("0.307633") @@ -1837,6 +2004,7 @@ class Program_weight_tensor_parameter_166: class Program_weight_tensor_parameter_167: name = "parameter_167" + original_name = "batch_norm2d_20.w_1" shape = [256] dtype = "float32" min_val = float("-1.91138") @@ -1848,6 +2016,7 @@ class Program_weight_tensor_parameter_167: class Program_weight_tensor_parameter_168: name = "parameter_168" + original_name = "create_parameter_60.w_0" shape = [256, 128, 1, 1] dtype = "float32" min_val = float("-0.428357") @@ -1859,6 +2028,7 @@ class Program_weight_tensor_parameter_168: class Program_weight_tensor_parameter_169: name = "parameter_169" + original_name = "create_parameter_59.w_0" shape = [128] dtype = "float32" min_val = float("-2.14417") @@ -1870,6 +2040,7 @@ class Program_weight_tensor_parameter_169: class Program_weight_tensor_parameter_170: name = "parameter_170" + original_name = "create_parameter_58.w_0" shape = [128] dtype = "float32" min_val = float("0.607841") @@ -1881,6 +2052,7 @@ class Program_weight_tensor_parameter_170: class Program_weight_tensor_parameter_171: name = "parameter_171" + original_name = "batch_norm2d_19.w_2" shape = [128] dtype = "float32" min_val = float("1.12113") @@ -1892,6 +2064,7 @@ class Program_weight_tensor_parameter_171: class Program_weight_tensor_parameter_172: name = "parameter_172" + original_name = "batch_norm2d_19.w_1" shape = [128] dtype = "float32" min_val = float("-4.92335") @@ -1903,6 +2076,7 @@ class Program_weight_tensor_parameter_172: class Program_weight_tensor_parameter_173: name = "parameter_173" + original_name = "create_parameter_57.w_0" shape = [128, 256, 1, 1] dtype = "float32" min_val = float("-0.544527") @@ -1914,6 +2088,7 @@ class Program_weight_tensor_parameter_173: class Program_weight_tensor_parameter_174: name = "parameter_174" + original_name = "p_re_lu_13.w_0" shape = [256] dtype = "float32" min_val = float("-1.4521") @@ -1925,6 +2100,7 @@ class Program_weight_tensor_parameter_174: class Program_weight_tensor_parameter_175: name = "parameter_175" + original_name = "create_parameter_56.w_0" shape = [256] dtype = "float32" min_val = float("-2.48052") @@ -1936,6 +2112,7 @@ class Program_weight_tensor_parameter_175: class Program_weight_tensor_parameter_176: name = "parameter_176" + original_name = "create_parameter_55.w_0" shape = [256] dtype = "float32" min_val = float("0.417075") @@ -1947,6 +2124,7 @@ class Program_weight_tensor_parameter_176: class Program_weight_tensor_parameter_177: name = "parameter_177" + original_name = "batch_norm2d_18.w_2" shape = [256] dtype = "float32" min_val = float("0.000215633") @@ -1958,6 +2136,7 @@ class Program_weight_tensor_parameter_177: class Program_weight_tensor_parameter_178: name = "parameter_178" + original_name = "batch_norm2d_18.w_1" shape = [256] dtype = "float32" min_val = float("-12.2918") @@ -1969,6 +2148,7 @@ class Program_weight_tensor_parameter_178: class Program_weight_tensor_parameter_179: name = "parameter_179" + original_name = "create_parameter_54.w_0" shape = [256, 1, 3, 3] dtype = "float32" min_val = float("-0.241329") @@ -1980,6 +2160,7 @@ class Program_weight_tensor_parameter_179: class Program_weight_tensor_parameter_180: name = "parameter_180" + original_name = "p_re_lu_12.w_0" shape = [256] dtype = "float32" min_val = float("-1.35817") @@ -1991,6 +2172,7 @@ class Program_weight_tensor_parameter_180: class Program_weight_tensor_parameter_181: name = "parameter_181" + original_name = "create_parameter_53.w_0" shape = [256] dtype = "float32" min_val = float("-1.40557") @@ -2002,6 +2184,7 @@ class Program_weight_tensor_parameter_181: class Program_weight_tensor_parameter_182: name = "parameter_182" + original_name = "create_parameter_52.w_0" shape = [256] dtype = "float32" min_val = float("0.322521") @@ -2013,6 +2196,7 @@ class Program_weight_tensor_parameter_182: class Program_weight_tensor_parameter_183: name = "parameter_183" + original_name = "batch_norm2d_17.w_2" shape = [256] dtype = "float32" min_val = float("1.41019") @@ -2024,6 +2208,7 @@ class Program_weight_tensor_parameter_183: class Program_weight_tensor_parameter_184: name = "parameter_184" + original_name = "batch_norm2d_17.w_1" shape = [256] dtype = "float32" min_val = float("-6.88009") @@ -2035,6 +2220,7 @@ class Program_weight_tensor_parameter_184: class Program_weight_tensor_parameter_185: name = "parameter_185" + original_name = "create_parameter_51.w_0" shape = [256, 64, 1, 1] dtype = "float32" min_val = float("-0.575668") @@ -2046,6 +2232,7 @@ class Program_weight_tensor_parameter_185: class Program_weight_tensor_parameter_186: name = "parameter_186" + original_name = "create_parameter_50.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -2055,6 +2242,7 @@ class Program_weight_tensor_parameter_186: class Program_weight_tensor_parameter_187: name = "parameter_187" + original_name = "create_parameter_49.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -2064,6 +2252,7 @@ class Program_weight_tensor_parameter_187: class Program_weight_tensor_parameter_188: name = "parameter_188" + original_name = "batch_norm2d_16.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -2073,6 +2262,7 @@ class Program_weight_tensor_parameter_188: class Program_weight_tensor_parameter_189: name = "parameter_189" + original_name = "batch_norm2d_16.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -2082,6 +2272,7 @@ class Program_weight_tensor_parameter_189: class Program_weight_tensor_parameter_190: name = "parameter_190" + original_name = "create_parameter_48.w_0" shape = [64, 128, 1, 1] dtype = "float32" min_val = float("-0.347509") @@ -2093,6 +2284,7 @@ class Program_weight_tensor_parameter_190: class Program_weight_tensor_parameter_191: name = "parameter_191" + original_name = "p_re_lu_11.w_0" shape = [128] dtype = "float32" min_val = float("-2.62566") @@ -2104,6 +2296,7 @@ class Program_weight_tensor_parameter_191: class Program_weight_tensor_parameter_192: name = "parameter_192" + original_name = "create_parameter_47.w_0" shape = [128] dtype = "float32" min_val = float("-5.45737") @@ -2115,6 +2308,7 @@ class Program_weight_tensor_parameter_192: class Program_weight_tensor_parameter_193: name = "parameter_193" + original_name = "create_parameter_46.w_0" shape = [128] dtype = "float32" min_val = float("0.473821") @@ -2126,6 +2320,7 @@ class Program_weight_tensor_parameter_193: class Program_weight_tensor_parameter_194: name = "parameter_194" + original_name = "batch_norm2d_15.w_2" shape = [128] dtype = "float32" min_val = float("0.000554323") @@ -2137,6 +2332,7 @@ class Program_weight_tensor_parameter_194: class Program_weight_tensor_parameter_195: name = "parameter_195" + original_name = "batch_norm2d_15.w_1" shape = [128] dtype = "float32" min_val = float("-1.54943") @@ -2148,6 +2344,7 @@ class Program_weight_tensor_parameter_195: class Program_weight_tensor_parameter_196: name = "parameter_196" + original_name = "create_parameter_45.w_0" shape = [128, 1, 3, 3] dtype = "float32" min_val = float("-0.871849") @@ -2159,6 +2356,7 @@ class Program_weight_tensor_parameter_196: class Program_weight_tensor_parameter_197: name = "parameter_197" + original_name = "p_re_lu_10.w_0" shape = [128] dtype = "float32" min_val = float("-1.27318") @@ -2170,6 +2368,7 @@ class Program_weight_tensor_parameter_197: class Program_weight_tensor_parameter_198: name = "parameter_198" + original_name = "create_parameter_44.w_0" shape = [128] dtype = "float32" min_val = float("-1.4089") @@ -2181,6 +2380,7 @@ class Program_weight_tensor_parameter_198: class Program_weight_tensor_parameter_199: name = "parameter_199" + original_name = "create_parameter_43.w_0" shape = [128] dtype = "float32" min_val = float("0.126469") @@ -2192,6 +2392,7 @@ class Program_weight_tensor_parameter_199: class Program_weight_tensor_parameter_200: name = "parameter_200" + original_name = "batch_norm2d_14.w_2" shape = [128] dtype = "float32" min_val = float("1.42744") @@ -2203,6 +2404,7 @@ class Program_weight_tensor_parameter_200: class Program_weight_tensor_parameter_201: name = "parameter_201" + original_name = "batch_norm2d_14.w_1" shape = [128] dtype = "float32" min_val = float("-4.00374") @@ -2214,6 +2416,7 @@ class Program_weight_tensor_parameter_201: class Program_weight_tensor_parameter_202: name = "parameter_202" + original_name = "create_parameter_42.w_0" shape = [128, 64, 1, 1] dtype = "float32" min_val = float("-0.319115") @@ -2225,6 +2428,7 @@ class Program_weight_tensor_parameter_202: class Program_weight_tensor_parameter_203: name = "parameter_203" + original_name = "create_parameter_41.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -2234,6 +2438,7 @@ class Program_weight_tensor_parameter_203: class Program_weight_tensor_parameter_204: name = "parameter_204" + original_name = "create_parameter_40.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -2243,6 +2448,7 @@ class Program_weight_tensor_parameter_204: class Program_weight_tensor_parameter_205: name = "parameter_205" + original_name = "batch_norm2d_13.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -2252,6 +2458,7 @@ class Program_weight_tensor_parameter_205: class Program_weight_tensor_parameter_206: name = "parameter_206" + original_name = "batch_norm2d_13.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -2261,6 +2468,7 @@ class Program_weight_tensor_parameter_206: class Program_weight_tensor_parameter_207: name = "parameter_207" + original_name = "create_parameter_39.w_0" shape = [64, 128, 1, 1] dtype = "float32" min_val = float("-0.480608") @@ -2272,6 +2480,7 @@ class Program_weight_tensor_parameter_207: class Program_weight_tensor_parameter_208: name = "parameter_208" + original_name = "p_re_lu_9.w_0" shape = [128] dtype = "float32" min_val = float("-1.95619") @@ -2283,6 +2492,7 @@ class Program_weight_tensor_parameter_208: class Program_weight_tensor_parameter_209: name = "parameter_209" + original_name = "create_parameter_38.w_0" shape = [128] dtype = "float32" min_val = float("-4.68014") @@ -2294,6 +2504,7 @@ class Program_weight_tensor_parameter_209: class Program_weight_tensor_parameter_210: name = "parameter_210" + original_name = "create_parameter_37.w_0" shape = [128] dtype = "float32" min_val = float("0.480384") @@ -2305,6 +2516,7 @@ class Program_weight_tensor_parameter_210: class Program_weight_tensor_parameter_211: name = "parameter_211" + original_name = "batch_norm2d_12.w_2" shape = [128] dtype = "float32" min_val = float("0.00041237") @@ -2316,6 +2528,7 @@ class Program_weight_tensor_parameter_211: class Program_weight_tensor_parameter_212: name = "parameter_212" + original_name = "batch_norm2d_12.w_1" shape = [128] dtype = "float32" min_val = float("-1.18143") @@ -2327,6 +2540,7 @@ class Program_weight_tensor_parameter_212: class Program_weight_tensor_parameter_213: name = "parameter_213" + original_name = "create_parameter_36.w_0" shape = [128, 1, 3, 3] dtype = "float32" min_val = float("-0.449287") @@ -2338,6 +2552,7 @@ class Program_weight_tensor_parameter_213: class Program_weight_tensor_parameter_214: name = "parameter_214" + original_name = "p_re_lu_8.w_0" shape = [128] dtype = "float32" min_val = float("-1.20466") @@ -2349,6 +2564,7 @@ class Program_weight_tensor_parameter_214: class Program_weight_tensor_parameter_215: name = "parameter_215" + original_name = "create_parameter_35.w_0" shape = [128] dtype = "float32" min_val = float("-1.45572") @@ -2360,6 +2576,7 @@ class Program_weight_tensor_parameter_215: class Program_weight_tensor_parameter_216: name = "parameter_216" + original_name = "create_parameter_34.w_0" shape = [128] dtype = "float32" min_val = float("0.379786") @@ -2371,6 +2588,7 @@ class Program_weight_tensor_parameter_216: class Program_weight_tensor_parameter_217: name = "parameter_217" + original_name = "batch_norm2d_11.w_2" shape = [128] dtype = "float32" min_val = float("0.830185") @@ -2382,6 +2600,7 @@ class Program_weight_tensor_parameter_217: class Program_weight_tensor_parameter_218: name = "parameter_218" + original_name = "batch_norm2d_11.w_1" shape = [128] dtype = "float32" min_val = float("-4.05866") @@ -2393,6 +2612,7 @@ class Program_weight_tensor_parameter_218: class Program_weight_tensor_parameter_219: name = "parameter_219" + original_name = "create_parameter_33.w_0" shape = [128, 64, 1, 1] dtype = "float32" min_val = float("-0.314175") @@ -2404,6 +2624,7 @@ class Program_weight_tensor_parameter_219: class Program_weight_tensor_parameter_220: name = "parameter_220" + original_name = "create_parameter_32.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -2413,6 +2634,7 @@ class Program_weight_tensor_parameter_220: class Program_weight_tensor_parameter_221: name = "parameter_221" + original_name = "create_parameter_31.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -2422,6 +2644,7 @@ class Program_weight_tensor_parameter_221: class Program_weight_tensor_parameter_222: name = "parameter_222" + original_name = "batch_norm2d_10.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -2431,6 +2654,7 @@ class Program_weight_tensor_parameter_222: class Program_weight_tensor_parameter_223: name = "parameter_223" + original_name = "batch_norm2d_10.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -2440,6 +2664,7 @@ class Program_weight_tensor_parameter_223: class Program_weight_tensor_parameter_224: name = "parameter_224" + original_name = "create_parameter_30.w_0" shape = [64, 128, 1, 1] dtype = "float32" min_val = float("-0.324898") @@ -2451,6 +2676,7 @@ class Program_weight_tensor_parameter_224: class Program_weight_tensor_parameter_225: name = "parameter_225" + original_name = "p_re_lu_7.w_0" shape = [128] dtype = "float32" min_val = float("-2.08631") @@ -2462,6 +2688,7 @@ class Program_weight_tensor_parameter_225: class Program_weight_tensor_parameter_226: name = "parameter_226" + original_name = "create_parameter_29.w_0" shape = [128] dtype = "float32" min_val = float("-2.41764") @@ -2473,6 +2700,7 @@ class Program_weight_tensor_parameter_226: class Program_weight_tensor_parameter_227: name = "parameter_227" + original_name = "create_parameter_28.w_0" shape = [128] dtype = "float32" min_val = float("0.512532") @@ -2484,6 +2712,7 @@ class Program_weight_tensor_parameter_227: class Program_weight_tensor_parameter_228: name = "parameter_228" + original_name = "batch_norm2d_9.w_2" shape = [128] dtype = "float32" min_val = float("0.00436705") @@ -2495,6 +2724,7 @@ class Program_weight_tensor_parameter_228: class Program_weight_tensor_parameter_229: name = "parameter_229" + original_name = "batch_norm2d_9.w_1" shape = [128] dtype = "float32" min_val = float("-1.21529") @@ -2506,6 +2736,7 @@ class Program_weight_tensor_parameter_229: class Program_weight_tensor_parameter_230: name = "parameter_230" + original_name = "create_parameter_27.w_0" shape = [128, 1, 3, 3] dtype = "float32" min_val = float("-0.549605") @@ -2517,6 +2748,7 @@ class Program_weight_tensor_parameter_230: class Program_weight_tensor_parameter_231: name = "parameter_231" + original_name = "p_re_lu_6.w_0" shape = [128] dtype = "float32" min_val = float("-1.16685") @@ -2528,6 +2760,7 @@ class Program_weight_tensor_parameter_231: class Program_weight_tensor_parameter_232: name = "parameter_232" + original_name = "create_parameter_26.w_0" shape = [128] dtype = "float32" min_val = float("-1.57629") @@ -2539,6 +2772,7 @@ class Program_weight_tensor_parameter_232: class Program_weight_tensor_parameter_233: name = "parameter_233" + original_name = "create_parameter_25.w_0" shape = [128] dtype = "float32" min_val = float("0.587851") @@ -2550,6 +2784,7 @@ class Program_weight_tensor_parameter_233: class Program_weight_tensor_parameter_234: name = "parameter_234" + original_name = "batch_norm2d_8.w_2" shape = [128] dtype = "float32" min_val = float("0.505666") @@ -2561,6 +2796,7 @@ class Program_weight_tensor_parameter_234: class Program_weight_tensor_parameter_235: name = "parameter_235" + original_name = "batch_norm2d_8.w_1" shape = [128] dtype = "float32" min_val = float("-2.38386") @@ -2572,6 +2808,7 @@ class Program_weight_tensor_parameter_235: class Program_weight_tensor_parameter_236: name = "parameter_236" + original_name = "create_parameter_24.w_0" shape = [128, 64, 1, 1] dtype = "float32" min_val = float("-0.306139") @@ -2583,6 +2820,7 @@ class Program_weight_tensor_parameter_236: class Program_weight_tensor_parameter_237: name = "parameter_237" + original_name = "create_parameter_23.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -2592,6 +2830,7 @@ class Program_weight_tensor_parameter_237: class Program_weight_tensor_parameter_238: name = "parameter_238" + original_name = "create_parameter_22.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -2601,6 +2840,7 @@ class Program_weight_tensor_parameter_238: class Program_weight_tensor_parameter_239: name = "parameter_239" + original_name = "batch_norm2d_7.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -2610,6 +2850,7 @@ class Program_weight_tensor_parameter_239: class Program_weight_tensor_parameter_240: name = "parameter_240" + original_name = "batch_norm2d_7.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -2619,6 +2860,7 @@ class Program_weight_tensor_parameter_240: class Program_weight_tensor_parameter_241: name = "parameter_241" + original_name = "create_parameter_21.w_0" shape = [64, 128, 1, 1] dtype = "float32" min_val = float("-0.332293") @@ -2630,6 +2872,7 @@ class Program_weight_tensor_parameter_241: class Program_weight_tensor_parameter_242: name = "parameter_242" + original_name = "p_re_lu_5.w_0" shape = [128] dtype = "float32" min_val = float("-1.78892") @@ -2641,6 +2884,7 @@ class Program_weight_tensor_parameter_242: class Program_weight_tensor_parameter_243: name = "parameter_243" + original_name = "create_parameter_20.w_0" shape = [128] dtype = "float32" min_val = float("-3.25661") @@ -2652,6 +2896,7 @@ class Program_weight_tensor_parameter_243: class Program_weight_tensor_parameter_244: name = "parameter_244" + original_name = "create_parameter_19.w_0" shape = [128] dtype = "float32" min_val = float("0.432232") @@ -2663,6 +2908,7 @@ class Program_weight_tensor_parameter_244: class Program_weight_tensor_parameter_245: name = "parameter_245" + original_name = "batch_norm2d_6.w_2" shape = [128] dtype = "float32" min_val = float("0.00280592") @@ -2674,6 +2920,7 @@ class Program_weight_tensor_parameter_245: class Program_weight_tensor_parameter_246: name = "parameter_246" + original_name = "batch_norm2d_6.w_1" shape = [128] dtype = "float32" min_val = float("-2.33954") @@ -2685,6 +2932,7 @@ class Program_weight_tensor_parameter_246: class Program_weight_tensor_parameter_247: name = "parameter_247" + original_name = "create_parameter_18.w_0" shape = [128, 1, 3, 3] dtype = "float32" min_val = float("-0.451795") @@ -2696,6 +2944,7 @@ class Program_weight_tensor_parameter_247: class Program_weight_tensor_parameter_248: name = "parameter_248" + original_name = "p_re_lu_4.w_0" shape = [128] dtype = "float32" min_val = float("-1.1523") @@ -2707,6 +2956,7 @@ class Program_weight_tensor_parameter_248: class Program_weight_tensor_parameter_249: name = "parameter_249" + original_name = "create_parameter_17.w_0" shape = [128] dtype = "float32" min_val = float("-1.19202") @@ -2718,6 +2968,7 @@ class Program_weight_tensor_parameter_249: class Program_weight_tensor_parameter_250: name = "parameter_250" + original_name = "create_parameter_16.w_0" shape = [128] dtype = "float32" min_val = float("0.250075") @@ -2729,6 +2980,7 @@ class Program_weight_tensor_parameter_250: class Program_weight_tensor_parameter_251: name = "parameter_251" + original_name = "batch_norm2d_5.w_2" shape = [128] dtype = "float32" min_val = float("0.146916") @@ -2740,6 +2992,7 @@ class Program_weight_tensor_parameter_251: class Program_weight_tensor_parameter_252: name = "parameter_252" + original_name = "batch_norm2d_5.w_1" shape = [128] dtype = "float32" min_val = float("-1.61376") @@ -2751,6 +3004,7 @@ class Program_weight_tensor_parameter_252: class Program_weight_tensor_parameter_253: name = "parameter_253" + original_name = "create_parameter_15.w_0" shape = [128, 64, 1, 1] dtype = "float32" min_val = float("-0.473151") @@ -2762,6 +3016,7 @@ class Program_weight_tensor_parameter_253: class Program_weight_tensor_parameter_254: name = "parameter_254" + original_name = "create_parameter_14.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -2771,6 +3026,7 @@ class Program_weight_tensor_parameter_254: class Program_weight_tensor_parameter_255: name = "parameter_255" + original_name = "create_parameter_13.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -2780,6 +3036,7 @@ class Program_weight_tensor_parameter_255: class Program_weight_tensor_parameter_256: name = "parameter_256" + original_name = "batch_norm2d_4.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -2789,6 +3046,7 @@ class Program_weight_tensor_parameter_256: class Program_weight_tensor_parameter_257: name = "parameter_257" + original_name = "batch_norm2d_4.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -2798,6 +3056,7 @@ class Program_weight_tensor_parameter_257: class Program_weight_tensor_parameter_258: name = "parameter_258" + original_name = "create_parameter_12.w_0" shape = [64, 128, 1, 1] dtype = "float32" min_val = float("-0.364858") @@ -2809,6 +3068,7 @@ class Program_weight_tensor_parameter_258: class Program_weight_tensor_parameter_259: name = "parameter_259" + original_name = "p_re_lu_3.w_0" shape = [128] dtype = "float32" min_val = float("-2.26817") @@ -2820,6 +3080,7 @@ class Program_weight_tensor_parameter_259: class Program_weight_tensor_parameter_260: name = "parameter_260" + original_name = "create_parameter_11.w_0" shape = [128] dtype = "float32" min_val = float("-4.92465") @@ -2831,6 +3092,7 @@ class Program_weight_tensor_parameter_260: class Program_weight_tensor_parameter_261: name = "parameter_261" + original_name = "create_parameter_10.w_0" shape = [128] dtype = "float32" min_val = float("0.334952") @@ -2842,6 +3104,7 @@ class Program_weight_tensor_parameter_261: class Program_weight_tensor_parameter_262: name = "parameter_262" + original_name = "batch_norm2d_3.w_2" shape = [128] dtype = "float32" min_val = float("0.00141397") @@ -2853,6 +3116,7 @@ class Program_weight_tensor_parameter_262: class Program_weight_tensor_parameter_263: name = "parameter_263" + original_name = "batch_norm2d_3.w_1" shape = [128] dtype = "float32" min_val = float("-3.33303") @@ -2864,6 +3128,7 @@ class Program_weight_tensor_parameter_263: class Program_weight_tensor_parameter_264: name = "parameter_264" + original_name = "create_parameter_9.w_0" shape = [128, 1, 3, 3] dtype = "float32" min_val = float("-0.222675") @@ -2875,6 +3140,7 @@ class Program_weight_tensor_parameter_264: class Program_weight_tensor_parameter_265: name = "parameter_265" + original_name = "p_re_lu_2.w_0" shape = [128] dtype = "float32" min_val = float("-1.95234") @@ -2886,6 +3152,7 @@ class Program_weight_tensor_parameter_265: class Program_weight_tensor_parameter_266: name = "parameter_266" + original_name = "create_parameter_8.w_0" shape = [128] dtype = "float32" min_val = float("-1.37448") @@ -2897,6 +3164,7 @@ class Program_weight_tensor_parameter_266: class Program_weight_tensor_parameter_267: name = "parameter_267" + original_name = "create_parameter_7.w_0" shape = [128] dtype = "float32" min_val = float("0.136873") @@ -2908,6 +3176,7 @@ class Program_weight_tensor_parameter_267: class Program_weight_tensor_parameter_268: name = "parameter_268" + original_name = "batch_norm2d_2.w_2" shape = [128] dtype = "float32" min_val = float("0.0763601") @@ -2919,6 +3188,7 @@ class Program_weight_tensor_parameter_268: class Program_weight_tensor_parameter_269: name = "parameter_269" + original_name = "batch_norm2d_2.w_1" shape = [128] dtype = "float32" min_val = float("-2.36766") @@ -2930,6 +3200,7 @@ class Program_weight_tensor_parameter_269: class Program_weight_tensor_parameter_270: name = "parameter_270" + original_name = "create_parameter_6.w_0" shape = [128, 64, 1, 1] dtype = "float32" min_val = float("-0.673241") @@ -2941,6 +3212,7 @@ class Program_weight_tensor_parameter_270: class Program_weight_tensor_parameter_271: name = "parameter_271" + original_name = "p_re_lu_1.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -2950,6 +3222,7 @@ class Program_weight_tensor_parameter_271: class Program_weight_tensor_parameter_272: name = "parameter_272" + original_name = "create_parameter_5.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -2959,6 +3232,7 @@ class Program_weight_tensor_parameter_272: class Program_weight_tensor_parameter_273: name = "parameter_273" + original_name = "create_parameter_4.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -2968,6 +3242,7 @@ class Program_weight_tensor_parameter_273: class Program_weight_tensor_parameter_274: name = "parameter_274" + original_name = "batch_norm2d_1.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -2977,6 +3252,7 @@ class Program_weight_tensor_parameter_274: class Program_weight_tensor_parameter_275: name = "parameter_275" + original_name = "batch_norm2d_1.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -2986,6 +3262,7 @@ class Program_weight_tensor_parameter_275: class Program_weight_tensor_parameter_276: name = "parameter_276" + original_name = "create_parameter_3.w_0" shape = [64, 1, 3, 3] dtype = "float32" min_val = float("-0.446798") @@ -2997,6 +3274,7 @@ class Program_weight_tensor_parameter_276: class Program_weight_tensor_parameter_277: name = "parameter_277" + original_name = "p_re_lu_0.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -3006,6 +3284,7 @@ class Program_weight_tensor_parameter_277: class Program_weight_tensor_parameter_278: name = "parameter_278" + original_name = "create_parameter_2.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -3015,6 +3294,7 @@ class Program_weight_tensor_parameter_278: class Program_weight_tensor_parameter_279: name = "parameter_279" + original_name = "create_parameter_1.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -3024,6 +3304,7 @@ class Program_weight_tensor_parameter_279: class Program_weight_tensor_parameter_280: name = "parameter_280" + original_name = "batch_norm2d_0.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -3033,6 +3314,7 @@ class Program_weight_tensor_parameter_280: class Program_weight_tensor_parameter_281: name = "parameter_281" + original_name = "batch_norm2d_0.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -3042,6 +3324,7 @@ class Program_weight_tensor_parameter_281: class Program_weight_tensor_parameter_282: name = "parameter_282" + original_name = "create_parameter_0.w_0" shape = [64, 3, 3, 3] dtype = "float32" min_val = float("-1.27744") diff --git a/paddle_samples/PaddleX/MobileFaceNet/subgraph_2/input_meta.py b/paddle_samples/PaddleX/MobileFaceNet/subgraph_2/input_meta.py index 8a5b93099..d963f3835 100644 --- a/paddle_samples/PaddleX/MobileFaceNet/subgraph_2/input_meta.py +++ b/paddle_samples/PaddleX/MobileFaceNet/subgraph_2/input_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_data_0: name = "data_0" + original_name = "param_1" shape = [128, 93431] dtype = "float32" min_val = float("-0.409481") @@ -11,6 +12,7 @@ class Program_weight_tensor_data_0: class Program_weight_tensor_data_1: name = "data_1" + original_name = "var_178" shape = [104, 3, 112, 112] dtype = "float32" min_val = float("-1.0") @@ -22,6 +24,7 @@ class Program_weight_tensor_data_1: class Program_weight_tensor_data_2: name = "data_2" + original_name = "var_179" shape = [104, 1] dtype = "int64" min_val = 9 diff --git a/paddle_samples/PaddleX/MobileFaceNet/subgraph_2/weight_meta.py b/paddle_samples/PaddleX/MobileFaceNet/subgraph_2/weight_meta.py index ba0cb6eb1..6a7376849 100644 --- a/paddle_samples/PaddleX/MobileFaceNet/subgraph_2/weight_meta.py +++ b/paddle_samples/PaddleX/MobileFaceNet/subgraph_2/weight_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_parameter_0: name = "parameter_0" + original_name = "create_parameter_149.w_0" shape = [128] dtype = "float32" min_val = float("-0.0476865") @@ -11,6 +12,7 @@ class Program_weight_tensor_parameter_0: class Program_weight_tensor_parameter_1: name = "parameter_1" + original_name = "create_parameter_148.w_0" shape = [128] dtype = "float32" min_val = float("1.36121") @@ -22,6 +24,7 @@ class Program_weight_tensor_parameter_1: class Program_weight_tensor_parameter_2: name = "parameter_2" + original_name = "batch_norm2d_49.w_2" shape = [128] dtype = "float32" min_val = float("6.52615") @@ -33,6 +36,7 @@ class Program_weight_tensor_parameter_2: class Program_weight_tensor_parameter_3: name = "parameter_3" + original_name = "batch_norm2d_49.w_1" shape = [128] dtype = "float32" min_val = float("-5.75616") @@ -44,6 +48,7 @@ class Program_weight_tensor_parameter_3: class Program_weight_tensor_parameter_4: name = "parameter_4" + original_name = "create_parameter_147.w_0" shape = [128, 512, 1, 1] dtype = "float32" min_val = float("-0.419765") @@ -55,6 +60,7 @@ class Program_weight_tensor_parameter_4: class Program_weight_tensor_parameter_5: name = "parameter_5" + original_name = "create_parameter_146.w_0" shape = [512] dtype = "float32" min_val = float("-2.27707") @@ -66,6 +72,7 @@ class Program_weight_tensor_parameter_5: class Program_weight_tensor_parameter_6: name = "parameter_6" + original_name = "create_parameter_145.w_0" shape = [512] dtype = "float32" min_val = float("-1.8795") @@ -77,6 +84,7 @@ class Program_weight_tensor_parameter_6: class Program_weight_tensor_parameter_7: name = "parameter_7" + original_name = "batch_norm2d_48.w_2" shape = [512] dtype = "float32" min_val = float("5.60519e-45") @@ -88,6 +96,7 @@ class Program_weight_tensor_parameter_7: class Program_weight_tensor_parameter_8: name = "parameter_8" + original_name = "batch_norm2d_48.w_1" shape = [512] dtype = "float32" min_val = float("-13.2592") @@ -99,6 +108,7 @@ class Program_weight_tensor_parameter_8: class Program_weight_tensor_parameter_9: name = "parameter_9" + original_name = "create_parameter_144.w_0" shape = [512, 1, 7, 7] dtype = "float32" min_val = float("-0.490107") @@ -110,6 +120,7 @@ class Program_weight_tensor_parameter_9: class Program_weight_tensor_parameter_10: name = "parameter_10" + original_name = "p_re_lu_32.w_0" shape = [512] dtype = "float32" min_val = float("-3.2782") @@ -121,6 +132,7 @@ class Program_weight_tensor_parameter_10: class Program_weight_tensor_parameter_11: name = "parameter_11" + original_name = "create_parameter_143.w_0" shape = [512] dtype = "float32" min_val = float("-2.25383") @@ -132,6 +144,7 @@ class Program_weight_tensor_parameter_11: class Program_weight_tensor_parameter_12: name = "parameter_12" + original_name = "create_parameter_142.w_0" shape = [512] dtype = "float32" min_val = float("-3.65308") @@ -143,6 +156,7 @@ class Program_weight_tensor_parameter_12: class Program_weight_tensor_parameter_13: name = "parameter_13" + original_name = "batch_norm2d_47.w_2" shape = [512] dtype = "float32" min_val = float("6.22182e-08") @@ -154,6 +168,7 @@ class Program_weight_tensor_parameter_13: class Program_weight_tensor_parameter_14: name = "parameter_14" + original_name = "batch_norm2d_47.w_1" shape = [512] dtype = "float32" min_val = float("-7.17742") @@ -165,6 +180,7 @@ class Program_weight_tensor_parameter_14: class Program_weight_tensor_parameter_15: name = "parameter_15" + original_name = "create_parameter_141.w_0" shape = [512, 128, 1, 1] dtype = "float32" min_val = float("-0.478835") @@ -176,6 +192,7 @@ class Program_weight_tensor_parameter_15: class Program_weight_tensor_parameter_16: name = "parameter_16" + original_name = "create_parameter_140.w_0" shape = [128] dtype = "float32" min_val = float("-2.47816") @@ -187,6 +204,7 @@ class Program_weight_tensor_parameter_16: class Program_weight_tensor_parameter_17: name = "parameter_17" + original_name = "create_parameter_139.w_0" shape = [128] dtype = "float32" min_val = float("0.463178") @@ -198,6 +216,7 @@ class Program_weight_tensor_parameter_17: class Program_weight_tensor_parameter_18: name = "parameter_18" + original_name = "batch_norm2d_46.w_2" shape = [128] dtype = "float32" min_val = float("0.494628") @@ -209,6 +228,7 @@ class Program_weight_tensor_parameter_18: class Program_weight_tensor_parameter_19: name = "parameter_19" + original_name = "batch_norm2d_46.w_1" shape = [128] dtype = "float32" min_val = float("-2.70147") @@ -220,6 +240,7 @@ class Program_weight_tensor_parameter_19: class Program_weight_tensor_parameter_20: name = "parameter_20" + original_name = "create_parameter_138.w_0" shape = [128, 256, 1, 1] dtype = "float32" min_val = float("-0.368353") @@ -231,6 +252,7 @@ class Program_weight_tensor_parameter_20: class Program_weight_tensor_parameter_21: name = "parameter_21" + original_name = "p_re_lu_31.w_0" shape = [256] dtype = "float32" min_val = float("-2.53663") @@ -242,6 +264,7 @@ class Program_weight_tensor_parameter_21: class Program_weight_tensor_parameter_22: name = "parameter_22" + original_name = "create_parameter_137.w_0" shape = [256] dtype = "float32" min_val = float("-5.92081") @@ -253,6 +276,7 @@ class Program_weight_tensor_parameter_22: class Program_weight_tensor_parameter_23: name = "parameter_23" + original_name = "create_parameter_136.w_0" shape = [256] dtype = "float32" min_val = float("0.281625") @@ -264,6 +288,7 @@ class Program_weight_tensor_parameter_23: class Program_weight_tensor_parameter_24: name = "parameter_24" + original_name = "batch_norm2d_45.w_2" shape = [256] dtype = "float32" min_val = float("0.000528962") @@ -275,6 +300,7 @@ class Program_weight_tensor_parameter_24: class Program_weight_tensor_parameter_25: name = "parameter_25" + original_name = "batch_norm2d_45.w_1" shape = [256] dtype = "float32" min_val = float("-1.61635") @@ -286,6 +312,7 @@ class Program_weight_tensor_parameter_25: class Program_weight_tensor_parameter_26: name = "parameter_26" + original_name = "create_parameter_135.w_0" shape = [256, 1, 3, 3] dtype = "float32" min_val = float("-0.736995") @@ -297,6 +324,7 @@ class Program_weight_tensor_parameter_26: class Program_weight_tensor_parameter_27: name = "parameter_27" + original_name = "p_re_lu_30.w_0" shape = [256] dtype = "float32" min_val = float("-25.4404") @@ -308,6 +336,7 @@ class Program_weight_tensor_parameter_27: class Program_weight_tensor_parameter_28: name = "parameter_28" + original_name = "create_parameter_134.w_0" shape = [256] dtype = "float32" min_val = float("-2.90987") @@ -319,6 +348,7 @@ class Program_weight_tensor_parameter_28: class Program_weight_tensor_parameter_29: name = "parameter_29" + original_name = "create_parameter_133.w_0" shape = [256] dtype = "float32" min_val = float("-0.398413") @@ -330,6 +360,7 @@ class Program_weight_tensor_parameter_29: class Program_weight_tensor_parameter_30: name = "parameter_30" + original_name = "batch_norm2d_44.w_2" shape = [256] dtype = "float32" min_val = float("1.79849") @@ -341,6 +372,7 @@ class Program_weight_tensor_parameter_30: class Program_weight_tensor_parameter_31: name = "parameter_31" + original_name = "batch_norm2d_44.w_1" shape = [256] dtype = "float32" min_val = float("-3.02133") @@ -352,6 +384,7 @@ class Program_weight_tensor_parameter_31: class Program_weight_tensor_parameter_32: name = "parameter_32" + original_name = "create_parameter_132.w_0" shape = [256, 128, 1, 1] dtype = "float32" min_val = float("-0.415184") @@ -363,6 +396,7 @@ class Program_weight_tensor_parameter_32: class Program_weight_tensor_parameter_33: name = "parameter_33" + original_name = "create_parameter_131.w_0" shape = [128] dtype = "float32" min_val = float("-2.0248") @@ -374,6 +408,7 @@ class Program_weight_tensor_parameter_33: class Program_weight_tensor_parameter_34: name = "parameter_34" + original_name = "create_parameter_130.w_0" shape = [128] dtype = "float32" min_val = float("0.55114") @@ -385,6 +420,7 @@ class Program_weight_tensor_parameter_34: class Program_weight_tensor_parameter_35: name = "parameter_35" + original_name = "batch_norm2d_43.w_2" shape = [128] dtype = "float32" min_val = float("0.517643") @@ -396,6 +432,7 @@ class Program_weight_tensor_parameter_35: class Program_weight_tensor_parameter_36: name = "parameter_36" + original_name = "batch_norm2d_43.w_1" shape = [128] dtype = "float32" min_val = float("-2.28421") @@ -407,6 +444,7 @@ class Program_weight_tensor_parameter_36: class Program_weight_tensor_parameter_37: name = "parameter_37" + original_name = "create_parameter_129.w_0" shape = [128, 256, 1, 1] dtype = "float32" min_val = float("-0.349282") @@ -418,6 +456,7 @@ class Program_weight_tensor_parameter_37: class Program_weight_tensor_parameter_38: name = "parameter_38" + original_name = "p_re_lu_29.w_0" shape = [256] dtype = "float32" min_val = float("-2.96541") @@ -429,6 +468,7 @@ class Program_weight_tensor_parameter_38: class Program_weight_tensor_parameter_39: name = "parameter_39" + original_name = "create_parameter_128.w_0" shape = [256] dtype = "float32" min_val = float("-7.93896") @@ -440,6 +480,7 @@ class Program_weight_tensor_parameter_39: class Program_weight_tensor_parameter_40: name = "parameter_40" + original_name = "create_parameter_127.w_0" shape = [256] dtype = "float32" min_val = float("-0.0238939") @@ -451,6 +492,7 @@ class Program_weight_tensor_parameter_40: class Program_weight_tensor_parameter_41: name = "parameter_41" + original_name = "batch_norm2d_42.w_2" shape = [256] dtype = "float32" min_val = float("0.000774138") @@ -462,6 +504,7 @@ class Program_weight_tensor_parameter_41: class Program_weight_tensor_parameter_42: name = "parameter_42" + original_name = "batch_norm2d_42.w_1" shape = [256] dtype = "float32" min_val = float("-0.721419") @@ -473,6 +516,7 @@ class Program_weight_tensor_parameter_42: class Program_weight_tensor_parameter_43: name = "parameter_43" + original_name = "create_parameter_126.w_0" shape = [256, 1, 3, 3] dtype = "float32" min_val = float("-0.405248") @@ -484,6 +528,7 @@ class Program_weight_tensor_parameter_43: class Program_weight_tensor_parameter_44: name = "parameter_44" + original_name = "p_re_lu_28.w_0" shape = [256] dtype = "float32" min_val = float("-9.21716") @@ -495,6 +540,7 @@ class Program_weight_tensor_parameter_44: class Program_weight_tensor_parameter_45: name = "parameter_45" + original_name = "create_parameter_125.w_0" shape = [256] dtype = "float32" min_val = float("-1.40485") @@ -506,6 +552,7 @@ class Program_weight_tensor_parameter_45: class Program_weight_tensor_parameter_46: name = "parameter_46" + original_name = "create_parameter_124.w_0" shape = [256] dtype = "float32" min_val = float("-0.726111") @@ -517,6 +564,7 @@ class Program_weight_tensor_parameter_46: class Program_weight_tensor_parameter_47: name = "parameter_47" + original_name = "batch_norm2d_41.w_2" shape = [256] dtype = "float32" min_val = float("0.47672") @@ -528,6 +576,7 @@ class Program_weight_tensor_parameter_47: class Program_weight_tensor_parameter_48: name = "parameter_48" + original_name = "batch_norm2d_41.w_1" shape = [256] dtype = "float32" min_val = float("-1.7287") @@ -539,6 +588,7 @@ class Program_weight_tensor_parameter_48: class Program_weight_tensor_parameter_49: name = "parameter_49" + original_name = "create_parameter_123.w_0" shape = [256, 128, 1, 1] dtype = "float32" min_val = float("-0.323891") @@ -550,6 +600,7 @@ class Program_weight_tensor_parameter_49: class Program_weight_tensor_parameter_50: name = "parameter_50" + original_name = "create_parameter_122.w_0" shape = [128] dtype = "float32" min_val = float("-1.64813") @@ -561,6 +612,7 @@ class Program_weight_tensor_parameter_50: class Program_weight_tensor_parameter_51: name = "parameter_51" + original_name = "create_parameter_121.w_0" shape = [128] dtype = "float32" min_val = float("0.932106") @@ -572,6 +624,7 @@ class Program_weight_tensor_parameter_51: class Program_weight_tensor_parameter_52: name = "parameter_52" + original_name = "batch_norm2d_40.w_2" shape = [128] dtype = "float32" min_val = float("2.27565") @@ -583,6 +636,7 @@ class Program_weight_tensor_parameter_52: class Program_weight_tensor_parameter_53: name = "parameter_53" + original_name = "batch_norm2d_40.w_1" shape = [128] dtype = "float32" min_val = float("-5.60257") @@ -594,6 +648,7 @@ class Program_weight_tensor_parameter_53: class Program_weight_tensor_parameter_54: name = "parameter_54" + original_name = "create_parameter_120.w_0" shape = [128, 512, 1, 1] dtype = "float32" min_val = float("-0.328238") @@ -605,6 +660,7 @@ class Program_weight_tensor_parameter_54: class Program_weight_tensor_parameter_55: name = "parameter_55" + original_name = "p_re_lu_27.w_0" shape = [512] dtype = "float32" min_val = float("-1.75454") @@ -616,6 +672,7 @@ class Program_weight_tensor_parameter_55: class Program_weight_tensor_parameter_56: name = "parameter_56" + original_name = "create_parameter_119.w_0" shape = [512] dtype = "float32" min_val = float("-9.48303") @@ -627,6 +684,7 @@ class Program_weight_tensor_parameter_56: class Program_weight_tensor_parameter_57: name = "parameter_57" + original_name = "create_parameter_118.w_0" shape = [512] dtype = "float32" min_val = float("0.491687") @@ -638,6 +696,7 @@ class Program_weight_tensor_parameter_57: class Program_weight_tensor_parameter_58: name = "parameter_58" + original_name = "batch_norm2d_39.w_2" shape = [512] dtype = "float32" min_val = float("0.000257255") @@ -649,6 +708,7 @@ class Program_weight_tensor_parameter_58: class Program_weight_tensor_parameter_59: name = "parameter_59" + original_name = "batch_norm2d_39.w_1" shape = [512] dtype = "float32" min_val = float("-1.0238") @@ -660,6 +720,7 @@ class Program_weight_tensor_parameter_59: class Program_weight_tensor_parameter_60: name = "parameter_60" + original_name = "create_parameter_117.w_0" shape = [512, 1, 3, 3] dtype = "float32" min_val = float("-0.334637") @@ -671,6 +732,7 @@ class Program_weight_tensor_parameter_60: class Program_weight_tensor_parameter_61: name = "parameter_61" + original_name = "p_re_lu_26.w_0" shape = [512] dtype = "float32" min_val = float("-7.23302") @@ -682,6 +744,7 @@ class Program_weight_tensor_parameter_61: class Program_weight_tensor_parameter_62: name = "parameter_62" + original_name = "create_parameter_116.w_0" shape = [512] dtype = "float32" min_val = float("-1.57745") @@ -693,6 +756,7 @@ class Program_weight_tensor_parameter_62: class Program_weight_tensor_parameter_63: name = "parameter_63" + original_name = "create_parameter_115.w_0" shape = [512] dtype = "float32" min_val = float("-0.249546") @@ -704,6 +768,7 @@ class Program_weight_tensor_parameter_63: class Program_weight_tensor_parameter_64: name = "parameter_64" + original_name = "batch_norm2d_38.w_2" shape = [512] dtype = "float32" min_val = float("6.64366") @@ -715,6 +780,7 @@ class Program_weight_tensor_parameter_64: class Program_weight_tensor_parameter_65: name = "parameter_65" + original_name = "batch_norm2d_38.w_1" shape = [512] dtype = "float32" min_val = float("-14.1029") @@ -726,6 +792,7 @@ class Program_weight_tensor_parameter_65: class Program_weight_tensor_parameter_66: name = "parameter_66" + original_name = "create_parameter_114.w_0" shape = [512, 128, 1, 1] dtype = "float32" min_val = float("-0.393752") @@ -737,6 +804,7 @@ class Program_weight_tensor_parameter_66: class Program_weight_tensor_parameter_67: name = "parameter_67" + original_name = "create_parameter_113.w_0" shape = [128] dtype = "float32" min_val = float("-2.37038") @@ -748,6 +816,7 @@ class Program_weight_tensor_parameter_67: class Program_weight_tensor_parameter_68: name = "parameter_68" + original_name = "create_parameter_112.w_0" shape = [128] dtype = "float32" min_val = float("1.10728") @@ -759,6 +828,7 @@ class Program_weight_tensor_parameter_68: class Program_weight_tensor_parameter_69: name = "parameter_69" + original_name = "batch_norm2d_37.w_2" shape = [128] dtype = "float32" min_val = float("0.796922") @@ -770,6 +840,7 @@ class Program_weight_tensor_parameter_69: class Program_weight_tensor_parameter_70: name = "parameter_70" + original_name = "batch_norm2d_37.w_1" shape = [128] dtype = "float32" min_val = float("-2.78868") @@ -781,6 +852,7 @@ class Program_weight_tensor_parameter_70: class Program_weight_tensor_parameter_71: name = "parameter_71" + original_name = "create_parameter_111.w_0" shape = [128, 256, 1, 1] dtype = "float32" min_val = float("-0.321147") @@ -792,6 +864,7 @@ class Program_weight_tensor_parameter_71: class Program_weight_tensor_parameter_72: name = "parameter_72" + original_name = "p_re_lu_25.w_0" shape = [256] dtype = "float32" min_val = float("-3.2468") @@ -803,6 +876,7 @@ class Program_weight_tensor_parameter_72: class Program_weight_tensor_parameter_73: name = "parameter_73" + original_name = "create_parameter_110.w_0" shape = [256] dtype = "float32" min_val = float("-4.17068") @@ -814,6 +888,7 @@ class Program_weight_tensor_parameter_73: class Program_weight_tensor_parameter_74: name = "parameter_74" + original_name = "create_parameter_109.w_0" shape = [256] dtype = "float32" min_val = float("0.288061") @@ -825,6 +900,7 @@ class Program_weight_tensor_parameter_74: class Program_weight_tensor_parameter_75: name = "parameter_75" + original_name = "batch_norm2d_36.w_2" shape = [256] dtype = "float32" min_val = float("0.000484347") @@ -836,6 +912,7 @@ class Program_weight_tensor_parameter_75: class Program_weight_tensor_parameter_76: name = "parameter_76" + original_name = "batch_norm2d_36.w_1" shape = [256] dtype = "float32" min_val = float("-1.58316") @@ -847,6 +924,7 @@ class Program_weight_tensor_parameter_76: class Program_weight_tensor_parameter_77: name = "parameter_77" + original_name = "create_parameter_108.w_0" shape = [256, 1, 3, 3] dtype = "float32" min_val = float("-0.709561") @@ -858,6 +936,7 @@ class Program_weight_tensor_parameter_77: class Program_weight_tensor_parameter_78: name = "parameter_78" + original_name = "p_re_lu_24.w_0" shape = [256] dtype = "float32" min_val = float("-11.3124") @@ -869,6 +948,7 @@ class Program_weight_tensor_parameter_78: class Program_weight_tensor_parameter_79: name = "parameter_79" + original_name = "create_parameter_107.w_0" shape = [256] dtype = "float32" min_val = float("-1.8262") @@ -880,6 +960,7 @@ class Program_weight_tensor_parameter_79: class Program_weight_tensor_parameter_80: name = "parameter_80" + original_name = "create_parameter_106.w_0" shape = [256] dtype = "float32" min_val = float("0.357815") @@ -891,6 +972,7 @@ class Program_weight_tensor_parameter_80: class Program_weight_tensor_parameter_81: name = "parameter_81" + original_name = "batch_norm2d_35.w_2" shape = [256] dtype = "float32" min_val = float("5.51238") @@ -902,6 +984,7 @@ class Program_weight_tensor_parameter_81: class Program_weight_tensor_parameter_82: name = "parameter_82" + original_name = "batch_norm2d_35.w_1" shape = [256] dtype = "float32" min_val = float("-11.273") @@ -913,6 +996,7 @@ class Program_weight_tensor_parameter_82: class Program_weight_tensor_parameter_83: name = "parameter_83" + original_name = "create_parameter_105.w_0" shape = [256, 128, 1, 1] dtype = "float32" min_val = float("-0.364678") @@ -924,6 +1008,7 @@ class Program_weight_tensor_parameter_83: class Program_weight_tensor_parameter_84: name = "parameter_84" + original_name = "create_parameter_104.w_0" shape = [128] dtype = "float32" min_val = float("-2.38461") @@ -935,6 +1020,7 @@ class Program_weight_tensor_parameter_84: class Program_weight_tensor_parameter_85: name = "parameter_85" + original_name = "create_parameter_103.w_0" shape = [128] dtype = "float32" min_val = float("0.89689") @@ -946,6 +1032,7 @@ class Program_weight_tensor_parameter_85: class Program_weight_tensor_parameter_86: name = "parameter_86" + original_name = "batch_norm2d_34.w_2" shape = [128] dtype = "float32" min_val = float("0.774718") @@ -957,6 +1044,7 @@ class Program_weight_tensor_parameter_86: class Program_weight_tensor_parameter_87: name = "parameter_87" + original_name = "batch_norm2d_34.w_1" shape = [128] dtype = "float32" min_val = float("-3.03624") @@ -968,6 +1056,7 @@ class Program_weight_tensor_parameter_87: class Program_weight_tensor_parameter_88: name = "parameter_88" + original_name = "create_parameter_102.w_0" shape = [128, 256, 1, 1] dtype = "float32" min_val = float("-0.330712") @@ -979,6 +1068,7 @@ class Program_weight_tensor_parameter_88: class Program_weight_tensor_parameter_89: name = "parameter_89" + original_name = "p_re_lu_23.w_0" shape = [256] dtype = "float32" min_val = float("-2.3646") @@ -990,6 +1080,7 @@ class Program_weight_tensor_parameter_89: class Program_weight_tensor_parameter_90: name = "parameter_90" + original_name = "create_parameter_101.w_0" shape = [256] dtype = "float32" min_val = float("-5.71461") @@ -1001,6 +1092,7 @@ class Program_weight_tensor_parameter_90: class Program_weight_tensor_parameter_91: name = "parameter_91" + original_name = "create_parameter_100.w_0" shape = [256] dtype = "float32" min_val = float("0.349337") @@ -1012,6 +1104,7 @@ class Program_weight_tensor_parameter_91: class Program_weight_tensor_parameter_92: name = "parameter_92" + original_name = "batch_norm2d_33.w_2" shape = [256] dtype = "float32" min_val = float("0.000879841") @@ -1023,6 +1116,7 @@ class Program_weight_tensor_parameter_92: class Program_weight_tensor_parameter_93: name = "parameter_93" + original_name = "batch_norm2d_33.w_1" shape = [256] dtype = "float32" min_val = float("-3.11596") @@ -1034,6 +1128,7 @@ class Program_weight_tensor_parameter_93: class Program_weight_tensor_parameter_94: name = "parameter_94" + original_name = "create_parameter_99.w_0" shape = [256, 1, 3, 3] dtype = "float32" min_val = float("-0.643663") @@ -1045,6 +1140,7 @@ class Program_weight_tensor_parameter_94: class Program_weight_tensor_parameter_95: name = "parameter_95" + original_name = "p_re_lu_22.w_0" shape = [256] dtype = "float32" min_val = float("-7.46299") @@ -1056,6 +1152,7 @@ class Program_weight_tensor_parameter_95: class Program_weight_tensor_parameter_96: name = "parameter_96" + original_name = "create_parameter_98.w_0" shape = [256] dtype = "float32" min_val = float("-1.51356") @@ -1067,6 +1164,7 @@ class Program_weight_tensor_parameter_96: class Program_weight_tensor_parameter_97: name = "parameter_97" + original_name = "create_parameter_97.w_0" shape = [256] dtype = "float32" min_val = float("0.458434") @@ -1078,6 +1176,7 @@ class Program_weight_tensor_parameter_97: class Program_weight_tensor_parameter_98: name = "parameter_98" + original_name = "batch_norm2d_32.w_2" shape = [256] dtype = "float32" min_val = float("4.09567") @@ -1089,6 +1188,7 @@ class Program_weight_tensor_parameter_98: class Program_weight_tensor_parameter_99: name = "parameter_99" + original_name = "batch_norm2d_32.w_1" shape = [256] dtype = "float32" min_val = float("-9.56507") @@ -1100,6 +1200,7 @@ class Program_weight_tensor_parameter_99: class Program_weight_tensor_parameter_100: name = "parameter_100" + original_name = "create_parameter_96.w_0" shape = [256, 128, 1, 1] dtype = "float32" min_val = float("-0.334524") @@ -1111,6 +1212,7 @@ class Program_weight_tensor_parameter_100: class Program_weight_tensor_parameter_101: name = "parameter_101" + original_name = "create_parameter_95.w_0" shape = [128] dtype = "float32" min_val = float("-2.44412") @@ -1122,6 +1224,7 @@ class Program_weight_tensor_parameter_101: class Program_weight_tensor_parameter_102: name = "parameter_102" + original_name = "create_parameter_94.w_0" shape = [128] dtype = "float32" min_val = float("0.713177") @@ -1133,6 +1236,7 @@ class Program_weight_tensor_parameter_102: class Program_weight_tensor_parameter_103: name = "parameter_103" + original_name = "batch_norm2d_31.w_2" shape = [128] dtype = "float32" min_val = float("0.80313") @@ -1144,6 +1248,7 @@ class Program_weight_tensor_parameter_103: class Program_weight_tensor_parameter_104: name = "parameter_104" + original_name = "batch_norm2d_31.w_1" shape = [128] dtype = "float32" min_val = float("-3.12884") @@ -1155,6 +1260,7 @@ class Program_weight_tensor_parameter_104: class Program_weight_tensor_parameter_105: name = "parameter_105" + original_name = "create_parameter_93.w_0" shape = [128, 256, 1, 1] dtype = "float32" min_val = float("-0.385883") @@ -1166,6 +1272,7 @@ class Program_weight_tensor_parameter_105: class Program_weight_tensor_parameter_106: name = "parameter_106" + original_name = "p_re_lu_21.w_0" shape = [256] dtype = "float32" min_val = float("-2.56897") @@ -1177,6 +1284,7 @@ class Program_weight_tensor_parameter_106: class Program_weight_tensor_parameter_107: name = "parameter_107" + original_name = "create_parameter_92.w_0" shape = [256] dtype = "float32" min_val = float("-3.13198") @@ -1188,6 +1296,7 @@ class Program_weight_tensor_parameter_107: class Program_weight_tensor_parameter_108: name = "parameter_108" + original_name = "create_parameter_91.w_0" shape = [256] dtype = "float32" min_val = float("0.333235") @@ -1199,6 +1308,7 @@ class Program_weight_tensor_parameter_108: class Program_weight_tensor_parameter_109: name = "parameter_109" + original_name = "batch_norm2d_30.w_2" shape = [256] dtype = "float32" min_val = float("0.000848788") @@ -1210,6 +1320,7 @@ class Program_weight_tensor_parameter_109: class Program_weight_tensor_parameter_110: name = "parameter_110" + original_name = "batch_norm2d_30.w_1" shape = [256] dtype = "float32" min_val = float("-2.10281") @@ -1221,6 +1332,7 @@ class Program_weight_tensor_parameter_110: class Program_weight_tensor_parameter_111: name = "parameter_111" + original_name = "create_parameter_90.w_0" shape = [256, 1, 3, 3] dtype = "float32" min_val = float("-0.662732") @@ -1232,6 +1344,7 @@ class Program_weight_tensor_parameter_111: class Program_weight_tensor_parameter_112: name = "parameter_112" + original_name = "p_re_lu_20.w_0" shape = [256] dtype = "float32" min_val = float("-9.84596") @@ -1243,6 +1356,7 @@ class Program_weight_tensor_parameter_112: class Program_weight_tensor_parameter_113: name = "parameter_113" + original_name = "create_parameter_89.w_0" shape = [256] dtype = "float32" min_val = float("-1.5506") @@ -1254,6 +1368,7 @@ class Program_weight_tensor_parameter_113: class Program_weight_tensor_parameter_114: name = "parameter_114" + original_name = "create_parameter_88.w_0" shape = [256] dtype = "float32" min_val = float("-0.57792") @@ -1265,6 +1380,7 @@ class Program_weight_tensor_parameter_114: class Program_weight_tensor_parameter_115: name = "parameter_115" + original_name = "batch_norm2d_29.w_2" shape = [256] dtype = "float32" min_val = float("2.77883") @@ -1276,6 +1392,7 @@ class Program_weight_tensor_parameter_115: class Program_weight_tensor_parameter_116: name = "parameter_116" + original_name = "batch_norm2d_29.w_1" shape = [256] dtype = "float32" min_val = float("-10.5") @@ -1287,6 +1404,7 @@ class Program_weight_tensor_parameter_116: class Program_weight_tensor_parameter_117: name = "parameter_117" + original_name = "create_parameter_87.w_0" shape = [256, 128, 1, 1] dtype = "float32" min_val = float("-0.359085") @@ -1298,6 +1416,7 @@ class Program_weight_tensor_parameter_117: class Program_weight_tensor_parameter_118: name = "parameter_118" + original_name = "create_parameter_86.w_0" shape = [128] dtype = "float32" min_val = float("-2.68514") @@ -1309,6 +1428,7 @@ class Program_weight_tensor_parameter_118: class Program_weight_tensor_parameter_119: name = "parameter_119" + original_name = "create_parameter_85.w_0" shape = [128] dtype = "float32" min_val = float("0.590275") @@ -1320,6 +1440,7 @@ class Program_weight_tensor_parameter_119: class Program_weight_tensor_parameter_120: name = "parameter_120" + original_name = "batch_norm2d_28.w_2" shape = [128] dtype = "float32" min_val = float("0.617482") @@ -1331,6 +1452,7 @@ class Program_weight_tensor_parameter_120: class Program_weight_tensor_parameter_121: name = "parameter_121" + original_name = "batch_norm2d_28.w_1" shape = [128] dtype = "float32" min_val = float("-2.78932") @@ -1342,6 +1464,7 @@ class Program_weight_tensor_parameter_121: class Program_weight_tensor_parameter_122: name = "parameter_122" + original_name = "create_parameter_84.w_0" shape = [128, 256, 1, 1] dtype = "float32" min_val = float("-0.41904") @@ -1353,6 +1476,7 @@ class Program_weight_tensor_parameter_122: class Program_weight_tensor_parameter_123: name = "parameter_123" + original_name = "p_re_lu_19.w_0" shape = [256] dtype = "float32" min_val = float("-1.60295") @@ -1364,6 +1488,7 @@ class Program_weight_tensor_parameter_123: class Program_weight_tensor_parameter_124: name = "parameter_124" + original_name = "create_parameter_83.w_0" shape = [256] dtype = "float32" min_val = float("-4.05436") @@ -1375,6 +1500,7 @@ class Program_weight_tensor_parameter_124: class Program_weight_tensor_parameter_125: name = "parameter_125" + original_name = "create_parameter_82.w_0" shape = [256] dtype = "float32" min_val = float("-1.53997") @@ -1386,6 +1512,7 @@ class Program_weight_tensor_parameter_125: class Program_weight_tensor_parameter_126: name = "parameter_126" + original_name = "batch_norm2d_27.w_2" shape = [256] dtype = "float32" min_val = float("2.26014e-08") @@ -1397,6 +1524,7 @@ class Program_weight_tensor_parameter_126: class Program_weight_tensor_parameter_127: name = "parameter_127" + original_name = "batch_norm2d_27.w_1" shape = [256] dtype = "float32" min_val = float("-1.88275") @@ -1408,6 +1536,7 @@ class Program_weight_tensor_parameter_127: class Program_weight_tensor_parameter_128: name = "parameter_128" + original_name = "create_parameter_81.w_0" shape = [256, 1, 3, 3] dtype = "float32" min_val = float("-0.601298") @@ -1419,6 +1548,7 @@ class Program_weight_tensor_parameter_128: class Program_weight_tensor_parameter_129: name = "parameter_129" + original_name = "p_re_lu_18.w_0" shape = [256] dtype = "float32" min_val = float("-11.9158") @@ -1430,6 +1560,7 @@ class Program_weight_tensor_parameter_129: class Program_weight_tensor_parameter_130: name = "parameter_130" + original_name = "create_parameter_80.w_0" shape = [256] dtype = "float32" min_val = float("-1.84547") @@ -1441,6 +1572,7 @@ class Program_weight_tensor_parameter_130: class Program_weight_tensor_parameter_131: name = "parameter_131" + original_name = "create_parameter_79.w_0" shape = [256] dtype = "float32" min_val = float("-0.0147656") @@ -1452,6 +1584,7 @@ class Program_weight_tensor_parameter_131: class Program_weight_tensor_parameter_132: name = "parameter_132" + original_name = "batch_norm2d_26.w_2" shape = [256] dtype = "float32" min_val = float("1.73138") @@ -1463,6 +1596,7 @@ class Program_weight_tensor_parameter_132: class Program_weight_tensor_parameter_133: name = "parameter_133" + original_name = "batch_norm2d_26.w_1" shape = [256] dtype = "float32" min_val = float("-6.2661") @@ -1474,6 +1608,7 @@ class Program_weight_tensor_parameter_133: class Program_weight_tensor_parameter_134: name = "parameter_134" + original_name = "create_parameter_78.w_0" shape = [256, 128, 1, 1] dtype = "float32" min_val = float("-0.316691") @@ -1485,6 +1620,7 @@ class Program_weight_tensor_parameter_134: class Program_weight_tensor_parameter_135: name = "parameter_135" + original_name = "create_parameter_77.w_0" shape = [128] dtype = "float32" min_val = float("-2.42461") @@ -1496,6 +1632,7 @@ class Program_weight_tensor_parameter_135: class Program_weight_tensor_parameter_136: name = "parameter_136" + original_name = "create_parameter_76.w_0" shape = [128] dtype = "float32" min_val = float("0.530248") @@ -1507,6 +1644,7 @@ class Program_weight_tensor_parameter_136: class Program_weight_tensor_parameter_137: name = "parameter_137" + original_name = "batch_norm2d_25.w_2" shape = [128] dtype = "float32" min_val = float("0.982216") @@ -1518,6 +1656,7 @@ class Program_weight_tensor_parameter_137: class Program_weight_tensor_parameter_138: name = "parameter_138" + original_name = "batch_norm2d_25.w_1" shape = [128] dtype = "float32" min_val = float("-3.23035") @@ -1529,6 +1668,7 @@ class Program_weight_tensor_parameter_138: class Program_weight_tensor_parameter_139: name = "parameter_139" + original_name = "create_parameter_75.w_0" shape = [128, 256, 1, 1] dtype = "float32" min_val = float("-0.359199") @@ -1540,6 +1680,7 @@ class Program_weight_tensor_parameter_139: class Program_weight_tensor_parameter_140: name = "parameter_140" + original_name = "p_re_lu_17.w_0" shape = [256] dtype = "float32" min_val = float("-1.76432") @@ -1551,6 +1692,7 @@ class Program_weight_tensor_parameter_140: class Program_weight_tensor_parameter_141: name = "parameter_141" + original_name = "create_parameter_74.w_0" shape = [256] dtype = "float32" min_val = float("-2.85251") @@ -1562,6 +1704,7 @@ class Program_weight_tensor_parameter_141: class Program_weight_tensor_parameter_142: name = "parameter_142" + original_name = "create_parameter_73.w_0" shape = [256] dtype = "float32" min_val = float("0.38289") @@ -1573,6 +1716,7 @@ class Program_weight_tensor_parameter_142: class Program_weight_tensor_parameter_143: name = "parameter_143" + original_name = "batch_norm2d_24.w_2" shape = [256] dtype = "float32" min_val = float("0.000421309") @@ -1584,6 +1728,7 @@ class Program_weight_tensor_parameter_143: class Program_weight_tensor_parameter_144: name = "parameter_144" + original_name = "batch_norm2d_24.w_1" shape = [256] dtype = "float32" min_val = float("-1.04254") @@ -1595,6 +1740,7 @@ class Program_weight_tensor_parameter_144: class Program_weight_tensor_parameter_145: name = "parameter_145" + original_name = "create_parameter_72.w_0" shape = [256, 1, 3, 3] dtype = "float32" min_val = float("-0.402921") @@ -1606,6 +1752,7 @@ class Program_weight_tensor_parameter_145: class Program_weight_tensor_parameter_146: name = "parameter_146" + original_name = "p_re_lu_16.w_0" shape = [256] dtype = "float32" min_val = float("-5.60544") @@ -1617,6 +1764,7 @@ class Program_weight_tensor_parameter_146: class Program_weight_tensor_parameter_147: name = "parameter_147" + original_name = "create_parameter_71.w_0" shape = [256] dtype = "float32" min_val = float("-1.73573") @@ -1628,6 +1776,7 @@ class Program_weight_tensor_parameter_147: class Program_weight_tensor_parameter_148: name = "parameter_148" + original_name = "create_parameter_70.w_0" shape = [256] dtype = "float32" min_val = float("0.511197") @@ -1639,6 +1788,7 @@ class Program_weight_tensor_parameter_148: class Program_weight_tensor_parameter_149: name = "parameter_149" + original_name = "batch_norm2d_23.w_2" shape = [256] dtype = "float32" min_val = float("1.29315") @@ -1650,6 +1800,7 @@ class Program_weight_tensor_parameter_149: class Program_weight_tensor_parameter_150: name = "parameter_150" + original_name = "batch_norm2d_23.w_1" shape = [256] dtype = "float32" min_val = float("-4.15903") @@ -1661,6 +1812,7 @@ class Program_weight_tensor_parameter_150: class Program_weight_tensor_parameter_151: name = "parameter_151" + original_name = "create_parameter_69.w_0" shape = [256, 128, 1, 1] dtype = "float32" min_val = float("-0.352596") @@ -1672,6 +1824,7 @@ class Program_weight_tensor_parameter_151: class Program_weight_tensor_parameter_152: name = "parameter_152" + original_name = "create_parameter_68.w_0" shape = [128] dtype = "float32" min_val = float("-1.96833") @@ -1683,6 +1836,7 @@ class Program_weight_tensor_parameter_152: class Program_weight_tensor_parameter_153: name = "parameter_153" + original_name = "create_parameter_67.w_0" shape = [128] dtype = "float32" min_val = float("0.525904") @@ -1694,6 +1848,7 @@ class Program_weight_tensor_parameter_153: class Program_weight_tensor_parameter_154: name = "parameter_154" + original_name = "batch_norm2d_22.w_2" shape = [128] dtype = "float32" min_val = float("1.02621") @@ -1705,6 +1860,7 @@ class Program_weight_tensor_parameter_154: class Program_weight_tensor_parameter_155: name = "parameter_155" + original_name = "batch_norm2d_22.w_1" shape = [128] dtype = "float32" min_val = float("-4.36954") @@ -1716,6 +1872,7 @@ class Program_weight_tensor_parameter_155: class Program_weight_tensor_parameter_156: name = "parameter_156" + original_name = "create_parameter_66.w_0" shape = [128, 256, 1, 1] dtype = "float32" min_val = float("-0.331118") @@ -1727,6 +1884,7 @@ class Program_weight_tensor_parameter_156: class Program_weight_tensor_parameter_157: name = "parameter_157" + original_name = "p_re_lu_15.w_0" shape = [256] dtype = "float32" min_val = float("-2.26721") @@ -1738,6 +1896,7 @@ class Program_weight_tensor_parameter_157: class Program_weight_tensor_parameter_158: name = "parameter_158" + original_name = "create_parameter_65.w_0" shape = [256] dtype = "float32" min_val = float("-2.31612") @@ -1749,6 +1908,7 @@ class Program_weight_tensor_parameter_158: class Program_weight_tensor_parameter_159: name = "parameter_159" + original_name = "create_parameter_64.w_0" shape = [256] dtype = "float32" min_val = float("0.571292") @@ -1760,6 +1920,7 @@ class Program_weight_tensor_parameter_159: class Program_weight_tensor_parameter_160: name = "parameter_160" + original_name = "batch_norm2d_21.w_2" shape = [256] dtype = "float32" min_val = float("0.00626217") @@ -1771,6 +1932,7 @@ class Program_weight_tensor_parameter_160: class Program_weight_tensor_parameter_161: name = "parameter_161" + original_name = "batch_norm2d_21.w_1" shape = [256] dtype = "float32" min_val = float("-1.14919") @@ -1782,6 +1944,7 @@ class Program_weight_tensor_parameter_161: class Program_weight_tensor_parameter_162: name = "parameter_162" + original_name = "create_parameter_63.w_0" shape = [256, 1, 3, 3] dtype = "float32" min_val = float("-0.399955") @@ -1793,6 +1956,7 @@ class Program_weight_tensor_parameter_162: class Program_weight_tensor_parameter_163: name = "parameter_163" + original_name = "p_re_lu_14.w_0" shape = [256] dtype = "float32" min_val = float("-1.37009") @@ -1804,6 +1968,7 @@ class Program_weight_tensor_parameter_163: class Program_weight_tensor_parameter_164: name = "parameter_164" + original_name = "create_parameter_62.w_0" shape = [256] dtype = "float32" min_val = float("-1.47056") @@ -1815,6 +1980,7 @@ class Program_weight_tensor_parameter_164: class Program_weight_tensor_parameter_165: name = "parameter_165" + original_name = "create_parameter_61.w_0" shape = [256] dtype = "float32" min_val = float("0.227376") @@ -1826,6 +1992,7 @@ class Program_weight_tensor_parameter_165: class Program_weight_tensor_parameter_166: name = "parameter_166" + original_name = "batch_norm2d_20.w_2" shape = [256] dtype = "float32" min_val = float("0.313107") @@ -1837,6 +2004,7 @@ class Program_weight_tensor_parameter_166: class Program_weight_tensor_parameter_167: name = "parameter_167" + original_name = "batch_norm2d_20.w_1" shape = [256] dtype = "float32" min_val = float("-1.9076") @@ -1848,6 +2016,7 @@ class Program_weight_tensor_parameter_167: class Program_weight_tensor_parameter_168: name = "parameter_168" + original_name = "create_parameter_60.w_0" shape = [256, 128, 1, 1] dtype = "float32" min_val = float("-0.427353") @@ -1859,6 +2028,7 @@ class Program_weight_tensor_parameter_168: class Program_weight_tensor_parameter_169: name = "parameter_169" + original_name = "create_parameter_59.w_0" shape = [128] dtype = "float32" min_val = float("-2.14289") @@ -1870,6 +2040,7 @@ class Program_weight_tensor_parameter_169: class Program_weight_tensor_parameter_170: name = "parameter_170" + original_name = "create_parameter_58.w_0" shape = [128] dtype = "float32" min_val = float("0.604782") @@ -1881,6 +2052,7 @@ class Program_weight_tensor_parameter_170: class Program_weight_tensor_parameter_171: name = "parameter_171" + original_name = "batch_norm2d_19.w_2" shape = [128] dtype = "float32" min_val = float("1.13442") @@ -1892,6 +2064,7 @@ class Program_weight_tensor_parameter_171: class Program_weight_tensor_parameter_172: name = "parameter_172" + original_name = "batch_norm2d_19.w_1" shape = [128] dtype = "float32" min_val = float("-4.93474") @@ -1903,6 +2076,7 @@ class Program_weight_tensor_parameter_172: class Program_weight_tensor_parameter_173: name = "parameter_173" + original_name = "create_parameter_57.w_0" shape = [128, 256, 1, 1] dtype = "float32" min_val = float("-0.548485") @@ -1914,6 +2088,7 @@ class Program_weight_tensor_parameter_173: class Program_weight_tensor_parameter_174: name = "parameter_174" + original_name = "p_re_lu_13.w_0" shape = [256] dtype = "float32" min_val = float("-1.45452") @@ -1925,6 +2100,7 @@ class Program_weight_tensor_parameter_174: class Program_weight_tensor_parameter_175: name = "parameter_175" + original_name = "create_parameter_56.w_0" shape = [256] dtype = "float32" min_val = float("-2.47756") @@ -1936,6 +2112,7 @@ class Program_weight_tensor_parameter_175: class Program_weight_tensor_parameter_176: name = "parameter_176" + original_name = "create_parameter_55.w_0" shape = [256] dtype = "float32" min_val = float("0.415456") @@ -1947,6 +2124,7 @@ class Program_weight_tensor_parameter_176: class Program_weight_tensor_parameter_177: name = "parameter_177" + original_name = "batch_norm2d_18.w_2" shape = [256] dtype = "float32" min_val = float("0.000210179") @@ -1958,6 +2136,7 @@ class Program_weight_tensor_parameter_177: class Program_weight_tensor_parameter_178: name = "parameter_178" + original_name = "batch_norm2d_18.w_1" shape = [256] dtype = "float32" min_val = float("-12.3337") @@ -1969,6 +2148,7 @@ class Program_weight_tensor_parameter_178: class Program_weight_tensor_parameter_179: name = "parameter_179" + original_name = "create_parameter_54.w_0" shape = [256, 1, 3, 3] dtype = "float32" min_val = float("-0.239819") @@ -1980,6 +2160,7 @@ class Program_weight_tensor_parameter_179: class Program_weight_tensor_parameter_180: name = "parameter_180" + original_name = "p_re_lu_12.w_0" shape = [256] dtype = "float32" min_val = float("-1.35518") @@ -1991,6 +2172,7 @@ class Program_weight_tensor_parameter_180: class Program_weight_tensor_parameter_181: name = "parameter_181" + original_name = "create_parameter_53.w_0" shape = [256] dtype = "float32" min_val = float("-1.40819") @@ -2002,6 +2184,7 @@ class Program_weight_tensor_parameter_181: class Program_weight_tensor_parameter_182: name = "parameter_182" + original_name = "create_parameter_52.w_0" shape = [256] dtype = "float32" min_val = float("0.325755") @@ -2013,6 +2196,7 @@ class Program_weight_tensor_parameter_182: class Program_weight_tensor_parameter_183: name = "parameter_183" + original_name = "batch_norm2d_17.w_2" shape = [256] dtype = "float32" min_val = float("1.35819") @@ -2024,6 +2208,7 @@ class Program_weight_tensor_parameter_183: class Program_weight_tensor_parameter_184: name = "parameter_184" + original_name = "batch_norm2d_17.w_1" shape = [256] dtype = "float32" min_val = float("-6.88019") @@ -2035,6 +2220,7 @@ class Program_weight_tensor_parameter_184: class Program_weight_tensor_parameter_185: name = "parameter_185" + original_name = "create_parameter_51.w_0" shape = [256, 64, 1, 1] dtype = "float32" min_val = float("-0.578467") @@ -2046,6 +2232,7 @@ class Program_weight_tensor_parameter_185: class Program_weight_tensor_parameter_186: name = "parameter_186" + original_name = "create_parameter_50.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -2055,6 +2242,7 @@ class Program_weight_tensor_parameter_186: class Program_weight_tensor_parameter_187: name = "parameter_187" + original_name = "create_parameter_49.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -2064,6 +2252,7 @@ class Program_weight_tensor_parameter_187: class Program_weight_tensor_parameter_188: name = "parameter_188" + original_name = "batch_norm2d_16.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -2073,6 +2262,7 @@ class Program_weight_tensor_parameter_188: class Program_weight_tensor_parameter_189: name = "parameter_189" + original_name = "batch_norm2d_16.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -2082,6 +2272,7 @@ class Program_weight_tensor_parameter_189: class Program_weight_tensor_parameter_190: name = "parameter_190" + original_name = "create_parameter_48.w_0" shape = [64, 128, 1, 1] dtype = "float32" min_val = float("-0.344653") @@ -2093,6 +2284,7 @@ class Program_weight_tensor_parameter_190: class Program_weight_tensor_parameter_191: name = "parameter_191" + original_name = "p_re_lu_11.w_0" shape = [128] dtype = "float32" min_val = float("-2.62574") @@ -2104,6 +2296,7 @@ class Program_weight_tensor_parameter_191: class Program_weight_tensor_parameter_192: name = "parameter_192" + original_name = "create_parameter_47.w_0" shape = [128] dtype = "float32" min_val = float("-5.46044") @@ -2115,6 +2308,7 @@ class Program_weight_tensor_parameter_192: class Program_weight_tensor_parameter_193: name = "parameter_193" + original_name = "create_parameter_46.w_0" shape = [128] dtype = "float32" min_val = float("0.477466") @@ -2126,6 +2320,7 @@ class Program_weight_tensor_parameter_193: class Program_weight_tensor_parameter_194: name = "parameter_194" + original_name = "batch_norm2d_15.w_2" shape = [128] dtype = "float32" min_val = float("0.000549945") @@ -2137,6 +2332,7 @@ class Program_weight_tensor_parameter_194: class Program_weight_tensor_parameter_195: name = "parameter_195" + original_name = "batch_norm2d_15.w_1" shape = [128] dtype = "float32" min_val = float("-1.55093") @@ -2148,6 +2344,7 @@ class Program_weight_tensor_parameter_195: class Program_weight_tensor_parameter_196: name = "parameter_196" + original_name = "create_parameter_45.w_0" shape = [128, 1, 3, 3] dtype = "float32" min_val = float("-0.870726") @@ -2159,6 +2356,7 @@ class Program_weight_tensor_parameter_196: class Program_weight_tensor_parameter_197: name = "parameter_197" + original_name = "p_re_lu_10.w_0" shape = [128] dtype = "float32" min_val = float("-1.27593") @@ -2170,6 +2368,7 @@ class Program_weight_tensor_parameter_197: class Program_weight_tensor_parameter_198: name = "parameter_198" + original_name = "create_parameter_44.w_0" shape = [128] dtype = "float32" min_val = float("-1.40877") @@ -2181,6 +2380,7 @@ class Program_weight_tensor_parameter_198: class Program_weight_tensor_parameter_199: name = "parameter_199" + original_name = "create_parameter_43.w_0" shape = [128] dtype = "float32" min_val = float("0.121505") @@ -2192,6 +2392,7 @@ class Program_weight_tensor_parameter_199: class Program_weight_tensor_parameter_200: name = "parameter_200" + original_name = "batch_norm2d_14.w_2" shape = [128] dtype = "float32" min_val = float("1.44566") @@ -2203,6 +2404,7 @@ class Program_weight_tensor_parameter_200: class Program_weight_tensor_parameter_201: name = "parameter_201" + original_name = "batch_norm2d_14.w_1" shape = [128] dtype = "float32" min_val = float("-4.00828") @@ -2214,6 +2416,7 @@ class Program_weight_tensor_parameter_201: class Program_weight_tensor_parameter_202: name = "parameter_202" + original_name = "create_parameter_42.w_0" shape = [128, 64, 1, 1] dtype = "float32" min_val = float("-0.319622") @@ -2225,6 +2428,7 @@ class Program_weight_tensor_parameter_202: class Program_weight_tensor_parameter_203: name = "parameter_203" + original_name = "create_parameter_41.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -2234,6 +2438,7 @@ class Program_weight_tensor_parameter_203: class Program_weight_tensor_parameter_204: name = "parameter_204" + original_name = "create_parameter_40.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -2243,6 +2448,7 @@ class Program_weight_tensor_parameter_204: class Program_weight_tensor_parameter_205: name = "parameter_205" + original_name = "batch_norm2d_13.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -2252,6 +2458,7 @@ class Program_weight_tensor_parameter_205: class Program_weight_tensor_parameter_206: name = "parameter_206" + original_name = "batch_norm2d_13.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -2261,6 +2468,7 @@ class Program_weight_tensor_parameter_206: class Program_weight_tensor_parameter_207: name = "parameter_207" + original_name = "create_parameter_39.w_0" shape = [64, 128, 1, 1] dtype = "float32" min_val = float("-0.477866") @@ -2272,6 +2480,7 @@ class Program_weight_tensor_parameter_207: class Program_weight_tensor_parameter_208: name = "parameter_208" + original_name = "p_re_lu_9.w_0" shape = [128] dtype = "float32" min_val = float("-1.95747") @@ -2283,6 +2492,7 @@ class Program_weight_tensor_parameter_208: class Program_weight_tensor_parameter_209: name = "parameter_209" + original_name = "create_parameter_38.w_0" shape = [128] dtype = "float32" min_val = float("-4.67779") @@ -2294,6 +2504,7 @@ class Program_weight_tensor_parameter_209: class Program_weight_tensor_parameter_210: name = "parameter_210" + original_name = "create_parameter_37.w_0" shape = [128] dtype = "float32" min_val = float("0.478135") @@ -2305,6 +2516,7 @@ class Program_weight_tensor_parameter_210: class Program_weight_tensor_parameter_211: name = "parameter_211" + original_name = "batch_norm2d_12.w_2" shape = [128] dtype = "float32" min_val = float("0.000431738") @@ -2316,6 +2528,7 @@ class Program_weight_tensor_parameter_211: class Program_weight_tensor_parameter_212: name = "parameter_212" + original_name = "batch_norm2d_12.w_1" shape = [128] dtype = "float32" min_val = float("-1.1806") @@ -2327,6 +2540,7 @@ class Program_weight_tensor_parameter_212: class Program_weight_tensor_parameter_213: name = "parameter_213" + original_name = "create_parameter_36.w_0" shape = [128, 1, 3, 3] dtype = "float32" min_val = float("-0.448925") @@ -2338,6 +2552,7 @@ class Program_weight_tensor_parameter_213: class Program_weight_tensor_parameter_214: name = "parameter_214" + original_name = "p_re_lu_8.w_0" shape = [128] dtype = "float32" min_val = float("-1.20297") @@ -2349,6 +2564,7 @@ class Program_weight_tensor_parameter_214: class Program_weight_tensor_parameter_215: name = "parameter_215" + original_name = "create_parameter_35.w_0" shape = [128] dtype = "float32" min_val = float("-1.45412") @@ -2360,6 +2576,7 @@ class Program_weight_tensor_parameter_215: class Program_weight_tensor_parameter_216: name = "parameter_216" + original_name = "create_parameter_34.w_0" shape = [128] dtype = "float32" min_val = float("0.378773") @@ -2371,6 +2588,7 @@ class Program_weight_tensor_parameter_216: class Program_weight_tensor_parameter_217: name = "parameter_217" + original_name = "batch_norm2d_11.w_2" shape = [128] dtype = "float32" min_val = float("0.834681") @@ -2382,6 +2600,7 @@ class Program_weight_tensor_parameter_217: class Program_weight_tensor_parameter_218: name = "parameter_218" + original_name = "batch_norm2d_11.w_1" shape = [128] dtype = "float32" min_val = float("-4.04353") @@ -2393,6 +2612,7 @@ class Program_weight_tensor_parameter_218: class Program_weight_tensor_parameter_219: name = "parameter_219" + original_name = "create_parameter_33.w_0" shape = [128, 64, 1, 1] dtype = "float32" min_val = float("-0.312178") @@ -2404,6 +2624,7 @@ class Program_weight_tensor_parameter_219: class Program_weight_tensor_parameter_220: name = "parameter_220" + original_name = "create_parameter_32.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -2413,6 +2634,7 @@ class Program_weight_tensor_parameter_220: class Program_weight_tensor_parameter_221: name = "parameter_221" + original_name = "create_parameter_31.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -2422,6 +2644,7 @@ class Program_weight_tensor_parameter_221: class Program_weight_tensor_parameter_222: name = "parameter_222" + original_name = "batch_norm2d_10.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -2431,6 +2654,7 @@ class Program_weight_tensor_parameter_222: class Program_weight_tensor_parameter_223: name = "parameter_223" + original_name = "batch_norm2d_10.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -2440,6 +2664,7 @@ class Program_weight_tensor_parameter_223: class Program_weight_tensor_parameter_224: name = "parameter_224" + original_name = "create_parameter_30.w_0" shape = [64, 128, 1, 1] dtype = "float32" min_val = float("-0.324848") @@ -2451,6 +2676,7 @@ class Program_weight_tensor_parameter_224: class Program_weight_tensor_parameter_225: name = "parameter_225" + original_name = "p_re_lu_7.w_0" shape = [128] dtype = "float32" min_val = float("-2.09098") @@ -2462,6 +2688,7 @@ class Program_weight_tensor_parameter_225: class Program_weight_tensor_parameter_226: name = "parameter_226" + original_name = "create_parameter_29.w_0" shape = [128] dtype = "float32" min_val = float("-2.41672") @@ -2473,6 +2700,7 @@ class Program_weight_tensor_parameter_226: class Program_weight_tensor_parameter_227: name = "parameter_227" + original_name = "create_parameter_28.w_0" shape = [128] dtype = "float32" min_val = float("0.510125") @@ -2484,6 +2712,7 @@ class Program_weight_tensor_parameter_227: class Program_weight_tensor_parameter_228: name = "parameter_228" + original_name = "batch_norm2d_9.w_2" shape = [128] dtype = "float32" min_val = float("0.00434379") @@ -2495,6 +2724,7 @@ class Program_weight_tensor_parameter_228: class Program_weight_tensor_parameter_229: name = "parameter_229" + original_name = "batch_norm2d_9.w_1" shape = [128] dtype = "float32" min_val = float("-1.21694") @@ -2506,6 +2736,7 @@ class Program_weight_tensor_parameter_229: class Program_weight_tensor_parameter_230: name = "parameter_230" + original_name = "create_parameter_27.w_0" shape = [128, 1, 3, 3] dtype = "float32" min_val = float("-0.55184") @@ -2517,6 +2748,7 @@ class Program_weight_tensor_parameter_230: class Program_weight_tensor_parameter_231: name = "parameter_231" + original_name = "p_re_lu_6.w_0" shape = [128] dtype = "float32" min_val = float("-1.16996") @@ -2528,6 +2760,7 @@ class Program_weight_tensor_parameter_231: class Program_weight_tensor_parameter_232: name = "parameter_232" + original_name = "create_parameter_26.w_0" shape = [128] dtype = "float32" min_val = float("-1.57551") @@ -2539,6 +2772,7 @@ class Program_weight_tensor_parameter_232: class Program_weight_tensor_parameter_233: name = "parameter_233" + original_name = "create_parameter_25.w_0" shape = [128] dtype = "float32" min_val = float("0.586447") @@ -2550,6 +2784,7 @@ class Program_weight_tensor_parameter_233: class Program_weight_tensor_parameter_234: name = "parameter_234" + original_name = "batch_norm2d_8.w_2" shape = [128] dtype = "float32" min_val = float("0.51473") @@ -2561,6 +2796,7 @@ class Program_weight_tensor_parameter_234: class Program_weight_tensor_parameter_235: name = "parameter_235" + original_name = "batch_norm2d_8.w_1" shape = [128] dtype = "float32" min_val = float("-2.3871") @@ -2572,6 +2808,7 @@ class Program_weight_tensor_parameter_235: class Program_weight_tensor_parameter_236: name = "parameter_236" + original_name = "create_parameter_24.w_0" shape = [128, 64, 1, 1] dtype = "float32" min_val = float("-0.308041") @@ -2583,6 +2820,7 @@ class Program_weight_tensor_parameter_236: class Program_weight_tensor_parameter_237: name = "parameter_237" + original_name = "create_parameter_23.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -2592,6 +2830,7 @@ class Program_weight_tensor_parameter_237: class Program_weight_tensor_parameter_238: name = "parameter_238" + original_name = "create_parameter_22.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -2601,6 +2840,7 @@ class Program_weight_tensor_parameter_238: class Program_weight_tensor_parameter_239: name = "parameter_239" + original_name = "batch_norm2d_7.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -2610,6 +2850,7 @@ class Program_weight_tensor_parameter_239: class Program_weight_tensor_parameter_240: name = "parameter_240" + original_name = "batch_norm2d_7.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -2619,6 +2860,7 @@ class Program_weight_tensor_parameter_240: class Program_weight_tensor_parameter_241: name = "parameter_241" + original_name = "create_parameter_21.w_0" shape = [64, 128, 1, 1] dtype = "float32" min_val = float("-0.328409") @@ -2630,6 +2872,7 @@ class Program_weight_tensor_parameter_241: class Program_weight_tensor_parameter_242: name = "parameter_242" + original_name = "p_re_lu_5.w_0" shape = [128] dtype = "float32" min_val = float("-1.79154") @@ -2641,6 +2884,7 @@ class Program_weight_tensor_parameter_242: class Program_weight_tensor_parameter_243: name = "parameter_243" + original_name = "create_parameter_20.w_0" shape = [128] dtype = "float32" min_val = float("-3.25818") @@ -2652,6 +2896,7 @@ class Program_weight_tensor_parameter_243: class Program_weight_tensor_parameter_244: name = "parameter_244" + original_name = "create_parameter_19.w_0" shape = [128] dtype = "float32" min_val = float("0.432455") @@ -2663,6 +2908,7 @@ class Program_weight_tensor_parameter_244: class Program_weight_tensor_parameter_245: name = "parameter_245" + original_name = "batch_norm2d_6.w_2" shape = [128] dtype = "float32" min_val = float("0.00289582") @@ -2674,6 +2920,7 @@ class Program_weight_tensor_parameter_245: class Program_weight_tensor_parameter_246: name = "parameter_246" + original_name = "batch_norm2d_6.w_1" shape = [128] dtype = "float32" min_val = float("-2.33721") @@ -2685,6 +2932,7 @@ class Program_weight_tensor_parameter_246: class Program_weight_tensor_parameter_247: name = "parameter_247" + original_name = "create_parameter_18.w_0" shape = [128, 1, 3, 3] dtype = "float32" min_val = float("-0.449424") @@ -2696,6 +2944,7 @@ class Program_weight_tensor_parameter_247: class Program_weight_tensor_parameter_248: name = "parameter_248" + original_name = "p_re_lu_4.w_0" shape = [128] dtype = "float32" min_val = float("-1.15459") @@ -2707,6 +2956,7 @@ class Program_weight_tensor_parameter_248: class Program_weight_tensor_parameter_249: name = "parameter_249" + original_name = "create_parameter_17.w_0" shape = [128] dtype = "float32" min_val = float("-1.19098") @@ -2718,6 +2968,7 @@ class Program_weight_tensor_parameter_249: class Program_weight_tensor_parameter_250: name = "parameter_250" + original_name = "create_parameter_16.w_0" shape = [128] dtype = "float32" min_val = float("0.253252") @@ -2729,6 +2980,7 @@ class Program_weight_tensor_parameter_250: class Program_weight_tensor_parameter_251: name = "parameter_251" + original_name = "batch_norm2d_5.w_2" shape = [128] dtype = "float32" min_val = float("0.146828") @@ -2740,6 +2992,7 @@ class Program_weight_tensor_parameter_251: class Program_weight_tensor_parameter_252: name = "parameter_252" + original_name = "batch_norm2d_5.w_1" shape = [128] dtype = "float32" min_val = float("-1.61425") @@ -2751,6 +3004,7 @@ class Program_weight_tensor_parameter_252: class Program_weight_tensor_parameter_253: name = "parameter_253" + original_name = "create_parameter_15.w_0" shape = [128, 64, 1, 1] dtype = "float32" min_val = float("-0.47187") @@ -2762,6 +3016,7 @@ class Program_weight_tensor_parameter_253: class Program_weight_tensor_parameter_254: name = "parameter_254" + original_name = "create_parameter_14.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -2771,6 +3026,7 @@ class Program_weight_tensor_parameter_254: class Program_weight_tensor_parameter_255: name = "parameter_255" + original_name = "create_parameter_13.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -2780,6 +3036,7 @@ class Program_weight_tensor_parameter_255: class Program_weight_tensor_parameter_256: name = "parameter_256" + original_name = "batch_norm2d_4.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -2789,6 +3046,7 @@ class Program_weight_tensor_parameter_256: class Program_weight_tensor_parameter_257: name = "parameter_257" + original_name = "batch_norm2d_4.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -2798,6 +3056,7 @@ class Program_weight_tensor_parameter_257: class Program_weight_tensor_parameter_258: name = "parameter_258" + original_name = "create_parameter_12.w_0" shape = [64, 128, 1, 1] dtype = "float32" min_val = float("-0.369357") @@ -2809,6 +3068,7 @@ class Program_weight_tensor_parameter_258: class Program_weight_tensor_parameter_259: name = "parameter_259" + original_name = "p_re_lu_3.w_0" shape = [128] dtype = "float32" min_val = float("-2.27215") @@ -2820,6 +3080,7 @@ class Program_weight_tensor_parameter_259: class Program_weight_tensor_parameter_260: name = "parameter_260" + original_name = "create_parameter_11.w_0" shape = [128] dtype = "float32" min_val = float("-4.92387") @@ -2831,6 +3092,7 @@ class Program_weight_tensor_parameter_260: class Program_weight_tensor_parameter_261: name = "parameter_261" + original_name = "create_parameter_10.w_0" shape = [128] dtype = "float32" min_val = float("0.338414") @@ -2842,6 +3104,7 @@ class Program_weight_tensor_parameter_261: class Program_weight_tensor_parameter_262: name = "parameter_262" + original_name = "batch_norm2d_3.w_2" shape = [128] dtype = "float32" min_val = float("0.00140203") @@ -2853,6 +3116,7 @@ class Program_weight_tensor_parameter_262: class Program_weight_tensor_parameter_263: name = "parameter_263" + original_name = "batch_norm2d_3.w_1" shape = [128] dtype = "float32" min_val = float("-3.38099") @@ -2864,6 +3128,7 @@ class Program_weight_tensor_parameter_263: class Program_weight_tensor_parameter_264: name = "parameter_264" + original_name = "create_parameter_9.w_0" shape = [128, 1, 3, 3] dtype = "float32" min_val = float("-0.222082") @@ -2875,6 +3140,7 @@ class Program_weight_tensor_parameter_264: class Program_weight_tensor_parameter_265: name = "parameter_265" + original_name = "p_re_lu_2.w_0" shape = [128] dtype = "float32" min_val = float("-1.95447") @@ -2886,6 +3152,7 @@ class Program_weight_tensor_parameter_265: class Program_weight_tensor_parameter_266: name = "parameter_266" + original_name = "create_parameter_8.w_0" shape = [128] dtype = "float32" min_val = float("-1.37572") @@ -2897,6 +3164,7 @@ class Program_weight_tensor_parameter_266: class Program_weight_tensor_parameter_267: name = "parameter_267" + original_name = "create_parameter_7.w_0" shape = [128] dtype = "float32" min_val = float("0.136703") @@ -2908,6 +3176,7 @@ class Program_weight_tensor_parameter_267: class Program_weight_tensor_parameter_268: name = "parameter_268" + original_name = "batch_norm2d_2.w_2" shape = [128] dtype = "float32" min_val = float("0.0838861") @@ -2919,6 +3188,7 @@ class Program_weight_tensor_parameter_268: class Program_weight_tensor_parameter_269: name = "parameter_269" + original_name = "batch_norm2d_2.w_1" shape = [128] dtype = "float32" min_val = float("-2.36651") @@ -2930,6 +3200,7 @@ class Program_weight_tensor_parameter_269: class Program_weight_tensor_parameter_270: name = "parameter_270" + original_name = "create_parameter_6.w_0" shape = [128, 64, 1, 1] dtype = "float32" min_val = float("-0.670918") @@ -2941,6 +3212,7 @@ class Program_weight_tensor_parameter_270: class Program_weight_tensor_parameter_271: name = "parameter_271" + original_name = "p_re_lu_1.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -2950,6 +3222,7 @@ class Program_weight_tensor_parameter_271: class Program_weight_tensor_parameter_272: name = "parameter_272" + original_name = "create_parameter_5.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -2959,6 +3232,7 @@ class Program_weight_tensor_parameter_272: class Program_weight_tensor_parameter_273: name = "parameter_273" + original_name = "create_parameter_4.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -2968,6 +3242,7 @@ class Program_weight_tensor_parameter_273: class Program_weight_tensor_parameter_274: name = "parameter_274" + original_name = "batch_norm2d_1.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -2977,6 +3252,7 @@ class Program_weight_tensor_parameter_274: class Program_weight_tensor_parameter_275: name = "parameter_275" + original_name = "batch_norm2d_1.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -2986,6 +3262,7 @@ class Program_weight_tensor_parameter_275: class Program_weight_tensor_parameter_276: name = "parameter_276" + original_name = "create_parameter_3.w_0" shape = [64, 1, 3, 3] dtype = "float32" min_val = float("-0.443397") @@ -2997,6 +3274,7 @@ class Program_weight_tensor_parameter_276: class Program_weight_tensor_parameter_277: name = "parameter_277" + original_name = "p_re_lu_0.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -3006,6 +3284,7 @@ class Program_weight_tensor_parameter_277: class Program_weight_tensor_parameter_278: name = "parameter_278" + original_name = "create_parameter_2.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -3015,6 +3294,7 @@ class Program_weight_tensor_parameter_278: class Program_weight_tensor_parameter_279: name = "parameter_279" + original_name = "create_parameter_1.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -3024,6 +3304,7 @@ class Program_weight_tensor_parameter_279: class Program_weight_tensor_parameter_280: name = "parameter_280" + original_name = "batch_norm2d_0.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -3033,6 +3314,7 @@ class Program_weight_tensor_parameter_280: class Program_weight_tensor_parameter_281: name = "parameter_281" + original_name = "batch_norm2d_0.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -3042,6 +3324,7 @@ class Program_weight_tensor_parameter_281: class Program_weight_tensor_parameter_282: name = "parameter_282" + original_name = "create_parameter_0.w_0" shape = [64, 3, 3, 3] dtype = "float32" min_val = float("-1.28059") diff --git a/paddle_samples/PaddleX/MobileNetV2_x1_0/input_meta.py b/paddle_samples/PaddleX/MobileNetV2_x1_0/input_meta.py index 1c7927023..bcf17dbfb 100644 --- a/paddle_samples/PaddleX/MobileNetV2_x1_0/input_meta.py +++ b/paddle_samples/PaddleX/MobileNetV2_x1_0/input_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_data_0: name = "data_0" + original_name = "var_0" shape = [256, 3, 224, 224] dtype = "float32" min_val = float("-2.1179") diff --git a/paddle_samples/PaddleX/MobileNetV2_x1_0/weight_meta.py b/paddle_samples/PaddleX/MobileNetV2_x1_0/weight_meta.py index c7e17d1ec..ce7037b58 100644 --- a/paddle_samples/PaddleX/MobileNetV2_x1_0/weight_meta.py +++ b/paddle_samples/PaddleX/MobileNetV2_x1_0/weight_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_parameter_0: name = "parameter_0" + original_name = "fc10_offset" shape = [102] dtype = "float32" data = None @@ -7,6 +8,7 @@ class Program_weight_tensor_parameter_0: class Program_weight_tensor_parameter_1: name = "parameter_1" + original_name = "fc10_weights" shape = [1280, 102] dtype = "float32" min_val = float("-0.0658901") @@ -18,6 +20,7 @@ class Program_weight_tensor_parameter_1: class Program_weight_tensor_parameter_2: name = "parameter_2" + original_name = "conv9_bn_offset" shape = [1280] dtype = "float32" min_val = float("-0.515879") @@ -29,6 +32,7 @@ class Program_weight_tensor_parameter_2: class Program_weight_tensor_parameter_3: name = "parameter_3" + original_name = "conv9_bn_scale" shape = [1280] dtype = "float32" min_val = float("1.51137") @@ -40,6 +44,7 @@ class Program_weight_tensor_parameter_3: class Program_weight_tensor_parameter_4: name = "parameter_4" + original_name = "conv9_bn_variance" shape = [1280] dtype = "float32" min_val = float("0.0453368") @@ -51,6 +56,7 @@ class Program_weight_tensor_parameter_4: class Program_weight_tensor_parameter_5: name = "parameter_5" + original_name = "conv9_bn_mean" shape = [1280] dtype = "float32" min_val = float("-0.000202692") @@ -62,6 +68,7 @@ class Program_weight_tensor_parameter_5: class Program_weight_tensor_parameter_6: name = "parameter_6" + original_name = "conv9_weights" shape = [1280, 320, 1, 1] dtype = "float32" min_val = float("-0.286049") @@ -73,6 +80,7 @@ class Program_weight_tensor_parameter_6: class Program_weight_tensor_parameter_7: name = "parameter_7" + original_name = "conv8_1_linear_bn_offset" shape = [320] dtype = "float32" min_val = float("5.06265e-06") @@ -84,6 +92,7 @@ class Program_weight_tensor_parameter_7: class Program_weight_tensor_parameter_8: name = "parameter_8" + original_name = "conv8_1_linear_bn_scale" shape = [320] dtype = "float32" min_val = float("0.284913") @@ -95,6 +104,7 @@ class Program_weight_tensor_parameter_8: class Program_weight_tensor_parameter_9: name = "parameter_9" + original_name = "conv8_1_linear_bn_variance" shape = [320] dtype = "float32" min_val = float("0.0206656") @@ -106,6 +116,7 @@ class Program_weight_tensor_parameter_9: class Program_weight_tensor_parameter_10: name = "parameter_10" + original_name = "conv8_1_linear_bn_mean" shape = [320] dtype = "float32" min_val = float("-1.01281") @@ -117,6 +128,7 @@ class Program_weight_tensor_parameter_10: class Program_weight_tensor_parameter_11: name = "parameter_11" + original_name = "conv8_1_linear_weights" shape = [320, 960, 1, 1] dtype = "float32" min_val = float("-0.324488") @@ -128,6 +140,7 @@ class Program_weight_tensor_parameter_11: class Program_weight_tensor_parameter_12: name = "parameter_12" + original_name = "conv8_1_dwise_bn_offset" shape = [960] dtype = "float32" min_val = float("-0.531185") @@ -139,6 +152,7 @@ class Program_weight_tensor_parameter_12: class Program_weight_tensor_parameter_13: name = "parameter_13" + original_name = "conv8_1_dwise_bn_scale" shape = [960] dtype = "float32" min_val = float("5.2452e-06") @@ -150,6 +164,7 @@ class Program_weight_tensor_parameter_13: class Program_weight_tensor_parameter_14: name = "parameter_14" + original_name = "conv8_1_dwise_bn_variance" shape = [960] dtype = "float32" min_val = float("5.60519e-45") @@ -161,6 +176,7 @@ class Program_weight_tensor_parameter_14: class Program_weight_tensor_parameter_15: name = "parameter_15" + original_name = "conv8_1_dwise_bn_mean" shape = [960] dtype = "float32" min_val = float("-0.17594") @@ -172,6 +188,7 @@ class Program_weight_tensor_parameter_15: class Program_weight_tensor_parameter_16: name = "parameter_16" + original_name = "conv8_1_dwise_weights" shape = [960, 1, 3, 3] dtype = "float32" min_val = float("-0.187452") @@ -183,6 +200,7 @@ class Program_weight_tensor_parameter_16: class Program_weight_tensor_parameter_17: name = "parameter_17" + original_name = "conv8_1_expand_bn_offset" shape = [960] dtype = "float32" min_val = float("-0.569152") @@ -194,6 +212,7 @@ class Program_weight_tensor_parameter_17: class Program_weight_tensor_parameter_18: name = "parameter_18" + original_name = "conv8_1_expand_bn_scale" shape = [960] dtype = "float32" min_val = float("4.45914e-07") @@ -205,6 +224,7 @@ class Program_weight_tensor_parameter_18: class Program_weight_tensor_parameter_19: name = "parameter_19" + original_name = "conv8_1_expand_bn_variance" shape = [960] dtype = "float32" min_val = float("1.54069e-10") @@ -216,6 +236,7 @@ class Program_weight_tensor_parameter_19: class Program_weight_tensor_parameter_20: name = "parameter_20" + original_name = "conv8_1_expand_bn_mean" shape = [960] dtype = "float32" min_val = float("-0.00633716") @@ -227,6 +248,7 @@ class Program_weight_tensor_parameter_20: class Program_weight_tensor_parameter_21: name = "parameter_21" + original_name = "conv8_1_expand_weights" shape = [960, 160, 1, 1] dtype = "float32" min_val = float("-0.585291") @@ -238,6 +260,7 @@ class Program_weight_tensor_parameter_21: class Program_weight_tensor_parameter_22: name = "parameter_22" + original_name = "conv7_3_linear_bn_offset" shape = [160] dtype = "float32" min_val = float("0.000626738") @@ -249,6 +272,7 @@ class Program_weight_tensor_parameter_22: class Program_weight_tensor_parameter_23: name = "parameter_23" + original_name = "conv7_3_linear_bn_scale" shape = [160] dtype = "float32" min_val = float("0.208359") @@ -260,6 +284,7 @@ class Program_weight_tensor_parameter_23: class Program_weight_tensor_parameter_24: name = "parameter_24" + original_name = "conv7_3_linear_bn_variance" shape = [160] dtype = "float32" min_val = float("0.0173921") @@ -271,6 +296,7 @@ class Program_weight_tensor_parameter_24: class Program_weight_tensor_parameter_25: name = "parameter_25" + original_name = "conv7_3_linear_bn_mean" shape = [160] dtype = "float32" min_val = float("-0.567566") @@ -282,6 +308,7 @@ class Program_weight_tensor_parameter_25: class Program_weight_tensor_parameter_26: name = "parameter_26" + original_name = "conv7_3_linear_weights" shape = [160, 960, 1, 1] dtype = "float32" min_val = float("-0.230647") @@ -293,6 +320,7 @@ class Program_weight_tensor_parameter_26: class Program_weight_tensor_parameter_27: name = "parameter_27" + original_name = "conv7_3_dwise_bn_offset" shape = [960] dtype = "float32" min_val = float("-0.555728") @@ -304,6 +332,7 @@ class Program_weight_tensor_parameter_27: class Program_weight_tensor_parameter_28: name = "parameter_28" + original_name = "conv7_3_dwise_bn_scale" shape = [960] dtype = "float32" min_val = float("3.53796e-06") @@ -315,6 +344,7 @@ class Program_weight_tensor_parameter_28: class Program_weight_tensor_parameter_29: name = "parameter_29" + original_name = "conv7_3_dwise_bn_variance" shape = [960] dtype = "float32" min_val = float("5.60519e-45") @@ -326,6 +356,7 @@ class Program_weight_tensor_parameter_29: class Program_weight_tensor_parameter_30: name = "parameter_30" + original_name = "conv7_3_dwise_bn_mean" shape = [960] dtype = "float32" min_val = float("-0.207315") @@ -337,6 +368,7 @@ class Program_weight_tensor_parameter_30: class Program_weight_tensor_parameter_31: name = "parameter_31" + original_name = "conv7_3_dwise_weights" shape = [960, 1, 3, 3] dtype = "float32" min_val = float("-0.599263") @@ -348,6 +380,7 @@ class Program_weight_tensor_parameter_31: class Program_weight_tensor_parameter_32: name = "parameter_32" + original_name = "conv7_3_expand_bn_offset" shape = [960] dtype = "float32" min_val = float("-0.26419") @@ -359,6 +392,7 @@ class Program_weight_tensor_parameter_32: class Program_weight_tensor_parameter_33: name = "parameter_33" + original_name = "conv7_3_expand_bn_scale" shape = [960] dtype = "float32" min_val = float("-4.81338e-06") @@ -370,6 +404,7 @@ class Program_weight_tensor_parameter_33: class Program_weight_tensor_parameter_34: name = "parameter_34" + original_name = "conv7_3_expand_bn_variance" shape = [960] dtype = "float32" min_val = float("6.49831e-11") @@ -381,6 +416,7 @@ class Program_weight_tensor_parameter_34: class Program_weight_tensor_parameter_35: name = "parameter_35" + original_name = "conv7_3_expand_bn_mean" shape = [960] dtype = "float32" min_val = float("-0.00296418") @@ -392,6 +428,7 @@ class Program_weight_tensor_parameter_35: class Program_weight_tensor_parameter_36: name = "parameter_36" + original_name = "conv7_3_expand_weights" shape = [960, 160, 1, 1] dtype = "float32" min_val = float("-0.308279") @@ -403,6 +440,7 @@ class Program_weight_tensor_parameter_36: class Program_weight_tensor_parameter_37: name = "parameter_37" + original_name = "conv7_2_linear_bn_offset" shape = [160] dtype = "float32" min_val = float("0.000649794") @@ -414,6 +452,7 @@ class Program_weight_tensor_parameter_37: class Program_weight_tensor_parameter_38: name = "parameter_38" + original_name = "conv7_2_linear_bn_scale" shape = [160] dtype = "float32" min_val = float("0.1822") @@ -425,6 +464,7 @@ class Program_weight_tensor_parameter_38: class Program_weight_tensor_parameter_39: name = "parameter_39" + original_name = "conv7_2_linear_bn_variance" shape = [160] dtype = "float32" min_val = float("0.0181743") @@ -436,6 +476,7 @@ class Program_weight_tensor_parameter_39: class Program_weight_tensor_parameter_40: name = "parameter_40" + original_name = "conv7_2_linear_bn_mean" shape = [160] dtype = "float32" min_val = float("-0.698936") @@ -447,6 +488,7 @@ class Program_weight_tensor_parameter_40: class Program_weight_tensor_parameter_41: name = "parameter_41" + original_name = "conv7_2_linear_weights" shape = [160, 960, 1, 1] dtype = "float32" min_val = float("-0.269475") @@ -458,6 +500,7 @@ class Program_weight_tensor_parameter_41: class Program_weight_tensor_parameter_42: name = "parameter_42" + original_name = "conv7_2_dwise_bn_offset" shape = [960] dtype = "float32" min_val = float("-0.677281") @@ -469,6 +512,7 @@ class Program_weight_tensor_parameter_42: class Program_weight_tensor_parameter_43: name = "parameter_43" + original_name = "conv7_2_dwise_bn_scale" shape = [960] dtype = "float32" min_val = float("4.49504e-06") @@ -480,6 +524,7 @@ class Program_weight_tensor_parameter_43: class Program_weight_tensor_parameter_44: name = "parameter_44" + original_name = "conv7_2_dwise_bn_variance" shape = [960] dtype = "float32" min_val = float("5.60519e-45") @@ -491,6 +536,7 @@ class Program_weight_tensor_parameter_44: class Program_weight_tensor_parameter_45: name = "parameter_45" + original_name = "conv7_2_dwise_bn_mean" shape = [960] dtype = "float32" min_val = float("-0.234563") @@ -502,6 +548,7 @@ class Program_weight_tensor_parameter_45: class Program_weight_tensor_parameter_46: name = "parameter_46" + original_name = "conv7_2_dwise_weights" shape = [960, 1, 3, 3] dtype = "float32" min_val = float("-0.559277") @@ -513,6 +560,7 @@ class Program_weight_tensor_parameter_46: class Program_weight_tensor_parameter_47: name = "parameter_47" + original_name = "conv7_2_expand_bn_offset" shape = [960] dtype = "float32" min_val = float("-0.19169") @@ -524,6 +572,7 @@ class Program_weight_tensor_parameter_47: class Program_weight_tensor_parameter_48: name = "parameter_48" + original_name = "conv7_2_expand_bn_scale" shape = [960] dtype = "float32" min_val = float("-0.0766565") @@ -535,6 +584,7 @@ class Program_weight_tensor_parameter_48: class Program_weight_tensor_parameter_49: name = "parameter_49" + original_name = "conv7_2_expand_bn_variance" shape = [960] dtype = "float32" min_val = float("6.26843e-11") @@ -546,6 +596,7 @@ class Program_weight_tensor_parameter_49: class Program_weight_tensor_parameter_50: name = "parameter_50" + original_name = "conv7_2_expand_bn_mean" shape = [960] dtype = "float32" min_val = float("-0.00114222") @@ -557,6 +608,7 @@ class Program_weight_tensor_parameter_50: class Program_weight_tensor_parameter_51: name = "parameter_51" + original_name = "conv7_2_expand_weights" shape = [960, 160, 1, 1] dtype = "float32" min_val = float("-0.265441") @@ -568,6 +620,7 @@ class Program_weight_tensor_parameter_51: class Program_weight_tensor_parameter_52: name = "parameter_52" + original_name = "conv7_1_linear_bn_offset" shape = [160] dtype = "float32" min_val = float("0.000546865") @@ -579,6 +632,7 @@ class Program_weight_tensor_parameter_52: class Program_weight_tensor_parameter_53: name = "parameter_53" + original_name = "conv7_1_linear_bn_scale" shape = [160] dtype = "float32" min_val = float("0.305364") @@ -590,6 +644,7 @@ class Program_weight_tensor_parameter_53: class Program_weight_tensor_parameter_54: name = "parameter_54" + original_name = "conv7_1_linear_bn_variance" shape = [160] dtype = "float32" min_val = float("0.0662108") @@ -601,6 +656,7 @@ class Program_weight_tensor_parameter_54: class Program_weight_tensor_parameter_55: name = "parameter_55" + original_name = "conv7_1_linear_bn_mean" shape = [160] dtype = "float32" min_val = float("-1.4787") @@ -612,6 +668,7 @@ class Program_weight_tensor_parameter_55: class Program_weight_tensor_parameter_56: name = "parameter_56" + original_name = "conv7_1_linear_weights" shape = [160, 576, 1, 1] dtype = "float32" min_val = float("-0.342484") @@ -623,6 +680,7 @@ class Program_weight_tensor_parameter_56: class Program_weight_tensor_parameter_57: name = "parameter_57" + original_name = "conv7_1_dwise_bn_offset" shape = [576] dtype = "float32" min_val = float("-0.265612") @@ -634,6 +692,7 @@ class Program_weight_tensor_parameter_57: class Program_weight_tensor_parameter_58: name = "parameter_58" + original_name = "conv7_1_dwise_bn_scale" shape = [576] dtype = "float32" min_val = float("5.79456e-06") @@ -645,6 +704,7 @@ class Program_weight_tensor_parameter_58: class Program_weight_tensor_parameter_59: name = "parameter_59" + original_name = "conv7_1_dwise_bn_variance" shape = [576] dtype = "float32" min_val = float("5.60519e-45") @@ -656,6 +716,7 @@ class Program_weight_tensor_parameter_59: class Program_weight_tensor_parameter_60: name = "parameter_60" + original_name = "conv7_1_dwise_bn_mean" shape = [576] dtype = "float32" min_val = float("-0.167296") @@ -667,6 +728,7 @@ class Program_weight_tensor_parameter_60: class Program_weight_tensor_parameter_61: name = "parameter_61" + original_name = "conv7_1_dwise_weights" shape = [576, 1, 3, 3] dtype = "float32" min_val = float("-0.286527") @@ -678,6 +740,7 @@ class Program_weight_tensor_parameter_61: class Program_weight_tensor_parameter_62: name = "parameter_62" + original_name = "conv7_1_expand_bn_offset" shape = [576] dtype = "float32" min_val = float("-0.456777") @@ -689,6 +752,7 @@ class Program_weight_tensor_parameter_62: class Program_weight_tensor_parameter_63: name = "parameter_63" + original_name = "conv7_1_expand_bn_scale" shape = [576] dtype = "float32" min_val = float("-4.80622e-05") @@ -700,6 +764,7 @@ class Program_weight_tensor_parameter_63: class Program_weight_tensor_parameter_64: name = "parameter_64" + original_name = "conv7_1_expand_bn_variance" shape = [576] dtype = "float32" min_val = float("1.26477e-10") @@ -711,6 +776,7 @@ class Program_weight_tensor_parameter_64: class Program_weight_tensor_parameter_65: name = "parameter_65" + original_name = "conv7_1_expand_bn_mean" shape = [576] dtype = "float32" min_val = float("-0.0579126") @@ -722,6 +788,7 @@ class Program_weight_tensor_parameter_65: class Program_weight_tensor_parameter_66: name = "parameter_66" + original_name = "conv7_1_expand_weights" shape = [576, 96, 1, 1] dtype = "float32" min_val = float("-0.476618") @@ -733,6 +800,7 @@ class Program_weight_tensor_parameter_66: class Program_weight_tensor_parameter_67: name = "parameter_67" + original_name = "conv6_3_linear_bn_offset" shape = [96] dtype = "float32" min_val = float("-0.0298874") @@ -744,6 +812,7 @@ class Program_weight_tensor_parameter_67: class Program_weight_tensor_parameter_68: name = "parameter_68" + original_name = "conv6_3_linear_bn_scale" shape = [96] dtype = "float32" min_val = float("0.101089") @@ -755,6 +824,7 @@ class Program_weight_tensor_parameter_68: class Program_weight_tensor_parameter_69: name = "parameter_69" + original_name = "conv6_3_linear_bn_variance" shape = [96] dtype = "float32" min_val = float("0.00999286") @@ -766,6 +836,7 @@ class Program_weight_tensor_parameter_69: class Program_weight_tensor_parameter_70: name = "parameter_70" + original_name = "conv6_3_linear_bn_mean" shape = [96] dtype = "float32" min_val = float("-0.398622") @@ -777,6 +848,7 @@ class Program_weight_tensor_parameter_70: class Program_weight_tensor_parameter_71: name = "parameter_71" + original_name = "conv6_3_linear_weights" shape = [96, 576, 1, 1] dtype = "float32" min_val = float("-0.306892") @@ -788,6 +860,7 @@ class Program_weight_tensor_parameter_71: class Program_weight_tensor_parameter_72: name = "parameter_72" + original_name = "conv6_3_dwise_bn_offset" shape = [576] dtype = "float32" min_val = float("-0.778854") @@ -799,6 +872,7 @@ class Program_weight_tensor_parameter_72: class Program_weight_tensor_parameter_73: name = "parameter_73" + original_name = "conv6_3_dwise_bn_scale" shape = [576] dtype = "float32" min_val = float("-0.0893874") @@ -810,6 +884,7 @@ class Program_weight_tensor_parameter_73: class Program_weight_tensor_parameter_74: name = "parameter_74" + original_name = "conv6_3_dwise_bn_variance" shape = [576] dtype = "float32" min_val = float("5.60519e-45") @@ -821,6 +896,7 @@ class Program_weight_tensor_parameter_74: class Program_weight_tensor_parameter_75: name = "parameter_75" + original_name = "conv6_3_dwise_bn_mean" shape = [576] dtype = "float32" min_val = float("-0.195586") @@ -832,6 +908,7 @@ class Program_weight_tensor_parameter_75: class Program_weight_tensor_parameter_76: name = "parameter_76" + original_name = "conv6_3_dwise_weights" shape = [576, 1, 3, 3] dtype = "float32" min_val = float("-0.50904") @@ -843,6 +920,7 @@ class Program_weight_tensor_parameter_76: class Program_weight_tensor_parameter_77: name = "parameter_77" + original_name = "conv6_3_expand_bn_offset" shape = [576] dtype = "float32" min_val = float("-0.385181") @@ -854,6 +932,7 @@ class Program_weight_tensor_parameter_77: class Program_weight_tensor_parameter_78: name = "parameter_78" + original_name = "conv6_3_expand_bn_scale" shape = [576] dtype = "float32" min_val = float("-4.80665e-06") @@ -865,6 +944,7 @@ class Program_weight_tensor_parameter_78: class Program_weight_tensor_parameter_79: name = "parameter_79" + original_name = "conv6_3_expand_bn_variance" shape = [576] dtype = "float32" min_val = float("6.34614e-11") @@ -876,6 +956,7 @@ class Program_weight_tensor_parameter_79: class Program_weight_tensor_parameter_80: name = "parameter_80" + original_name = "conv6_3_expand_bn_mean" shape = [576] dtype = "float32" min_val = float("-0.0423281") @@ -887,6 +968,7 @@ class Program_weight_tensor_parameter_80: class Program_weight_tensor_parameter_81: name = "parameter_81" + original_name = "conv6_3_expand_weights" shape = [576, 96, 1, 1] dtype = "float32" min_val = float("-0.320122") @@ -898,6 +980,7 @@ class Program_weight_tensor_parameter_81: class Program_weight_tensor_parameter_82: name = "parameter_82" + original_name = "conv6_2_linear_bn_offset" shape = [96] dtype = "float32" min_val = float("-0.0415129") @@ -909,6 +992,7 @@ class Program_weight_tensor_parameter_82: class Program_weight_tensor_parameter_83: name = "parameter_83" + original_name = "conv6_2_linear_bn_scale" shape = [96] dtype = "float32" min_val = float("0.0905051") @@ -920,6 +1004,7 @@ class Program_weight_tensor_parameter_83: class Program_weight_tensor_parameter_84: name = "parameter_84" + original_name = "conv6_2_linear_bn_variance" shape = [96] dtype = "float32" min_val = float("0.0102661") @@ -931,6 +1016,7 @@ class Program_weight_tensor_parameter_84: class Program_weight_tensor_parameter_85: name = "parameter_85" + original_name = "conv6_2_linear_bn_mean" shape = [96] dtype = "float32" min_val = float("-0.555293") @@ -942,6 +1028,7 @@ class Program_weight_tensor_parameter_85: class Program_weight_tensor_parameter_86: name = "parameter_86" + original_name = "conv6_2_linear_weights" shape = [96, 576, 1, 1] dtype = "float32" min_val = float("-0.280609") @@ -953,6 +1040,7 @@ class Program_weight_tensor_parameter_86: class Program_weight_tensor_parameter_87: name = "parameter_87" + original_name = "conv6_2_dwise_bn_offset" shape = [576] dtype = "float32" min_val = float("-0.494951") @@ -964,6 +1052,7 @@ class Program_weight_tensor_parameter_87: class Program_weight_tensor_parameter_88: name = "parameter_88" + original_name = "conv6_2_dwise_bn_scale" shape = [576] dtype = "float32" min_val = float("-0.00262835") @@ -975,6 +1064,7 @@ class Program_weight_tensor_parameter_88: class Program_weight_tensor_parameter_89: name = "parameter_89" + original_name = "conv6_2_dwise_bn_variance" shape = [576] dtype = "float32" min_val = float("5.60519e-45") @@ -986,6 +1076,7 @@ class Program_weight_tensor_parameter_89: class Program_weight_tensor_parameter_90: name = "parameter_90" + original_name = "conv6_2_dwise_bn_mean" shape = [576] dtype = "float32" min_val = float("-0.183864") @@ -997,6 +1088,7 @@ class Program_weight_tensor_parameter_90: class Program_weight_tensor_parameter_91: name = "parameter_91" + original_name = "conv6_2_dwise_weights" shape = [576, 1, 3, 3] dtype = "float32" min_val = float("-0.650009") @@ -1008,6 +1100,7 @@ class Program_weight_tensor_parameter_91: class Program_weight_tensor_parameter_92: name = "parameter_92" + original_name = "conv6_2_expand_bn_offset" shape = [576] dtype = "float32" min_val = float("-0.317215") @@ -1019,6 +1112,7 @@ class Program_weight_tensor_parameter_92: class Program_weight_tensor_parameter_93: name = "parameter_93" + original_name = "conv6_2_expand_bn_scale" shape = [576] dtype = "float32" min_val = float("-5.05088e-05") @@ -1030,6 +1124,7 @@ class Program_weight_tensor_parameter_93: class Program_weight_tensor_parameter_94: name = "parameter_94" + original_name = "conv6_2_expand_bn_variance" shape = [576] dtype = "float32" min_val = float("3.73005e-11") @@ -1041,6 +1136,7 @@ class Program_weight_tensor_parameter_94: class Program_weight_tensor_parameter_95: name = "parameter_95" + original_name = "conv6_2_expand_bn_mean" shape = [576] dtype = "float32" min_val = float("-0.0294991") @@ -1052,6 +1148,7 @@ class Program_weight_tensor_parameter_95: class Program_weight_tensor_parameter_96: name = "parameter_96" + original_name = "conv6_2_expand_weights" shape = [576, 96, 1, 1] dtype = "float32" min_val = float("-0.415294") @@ -1063,6 +1160,7 @@ class Program_weight_tensor_parameter_96: class Program_weight_tensor_parameter_97: name = "parameter_97" + original_name = "conv6_1_linear_bn_offset" shape = [96] dtype = "float32" min_val = float("-0.0634994") @@ -1074,6 +1172,7 @@ class Program_weight_tensor_parameter_97: class Program_weight_tensor_parameter_98: name = "parameter_98" + original_name = "conv6_1_linear_bn_scale" shape = [96] dtype = "float32" min_val = float("0.319786") @@ -1085,6 +1184,7 @@ class Program_weight_tensor_parameter_98: class Program_weight_tensor_parameter_99: name = "parameter_99" + original_name = "conv6_1_linear_bn_variance" shape = [96] dtype = "float32" min_val = float("0.0567135") @@ -1096,6 +1196,7 @@ class Program_weight_tensor_parameter_99: class Program_weight_tensor_parameter_100: name = "parameter_100" + original_name = "conv6_1_linear_bn_mean" shape = [96] dtype = "float32" min_val = float("-0.698804") @@ -1107,6 +1208,7 @@ class Program_weight_tensor_parameter_100: class Program_weight_tensor_parameter_101: name = "parameter_101" + original_name = "conv6_1_linear_weights" shape = [96, 384, 1, 1] dtype = "float32" min_val = float("-0.561591") @@ -1118,6 +1220,7 @@ class Program_weight_tensor_parameter_101: class Program_weight_tensor_parameter_102: name = "parameter_102" + original_name = "conv6_1_dwise_bn_offset" shape = [384] dtype = "float32" min_val = float("-0.364855") @@ -1129,6 +1232,7 @@ class Program_weight_tensor_parameter_102: class Program_weight_tensor_parameter_103: name = "parameter_103" + original_name = "conv6_1_dwise_bn_scale" shape = [384] dtype = "float32" min_val = float("5.9427e-05") @@ -1140,6 +1244,7 @@ class Program_weight_tensor_parameter_103: class Program_weight_tensor_parameter_104: name = "parameter_104" + original_name = "conv6_1_dwise_bn_variance" shape = [384] dtype = "float32" min_val = float("5.60519e-45") @@ -1151,6 +1256,7 @@ class Program_weight_tensor_parameter_104: class Program_weight_tensor_parameter_105: name = "parameter_105" + original_name = "conv6_1_dwise_bn_mean" shape = [384] dtype = "float32" min_val = float("-0.366835") @@ -1162,6 +1268,7 @@ class Program_weight_tensor_parameter_105: class Program_weight_tensor_parameter_106: name = "parameter_106" + original_name = "conv6_1_dwise_weights" shape = [384, 1, 3, 3] dtype = "float32" min_val = float("-0.602977") @@ -1173,6 +1280,7 @@ class Program_weight_tensor_parameter_106: class Program_weight_tensor_parameter_107: name = "parameter_107" + original_name = "conv6_1_expand_bn_offset" shape = [384] dtype = "float32" min_val = float("-0.225637") @@ -1184,6 +1292,7 @@ class Program_weight_tensor_parameter_107: class Program_weight_tensor_parameter_108: name = "parameter_108" + original_name = "conv6_1_expand_bn_scale" shape = [384] dtype = "float32" min_val = float("8.4513e-07") @@ -1195,6 +1304,7 @@ class Program_weight_tensor_parameter_108: class Program_weight_tensor_parameter_109: name = "parameter_109" + original_name = "conv6_1_expand_bn_variance" shape = [384] dtype = "float32" min_val = float("3.80421e-08") @@ -1206,6 +1316,7 @@ class Program_weight_tensor_parameter_109: class Program_weight_tensor_parameter_110: name = "parameter_110" + original_name = "conv6_1_expand_bn_mean" shape = [384] dtype = "float32" min_val = float("-0.0753244") @@ -1217,6 +1328,7 @@ class Program_weight_tensor_parameter_110: class Program_weight_tensor_parameter_111: name = "parameter_111" + original_name = "conv6_1_expand_weights" shape = [384, 64, 1, 1] dtype = "float32" min_val = float("-0.565511") @@ -1228,6 +1340,7 @@ class Program_weight_tensor_parameter_111: class Program_weight_tensor_parameter_112: name = "parameter_112" + original_name = "conv5_4_linear_bn_offset" shape = [64] dtype = "float32" min_val = float("0") @@ -1237,6 +1350,7 @@ class Program_weight_tensor_parameter_112: class Program_weight_tensor_parameter_113: name = "parameter_113" + original_name = "conv5_4_linear_bn_scale" shape = [64] dtype = "float32" min_val = float("0") @@ -1246,6 +1360,7 @@ class Program_weight_tensor_parameter_113: class Program_weight_tensor_parameter_114: name = "parameter_114" + original_name = "conv5_4_linear_bn_variance" shape = [64] dtype = "float32" min_val = float("0") @@ -1255,6 +1370,7 @@ class Program_weight_tensor_parameter_114: class Program_weight_tensor_parameter_115: name = "parameter_115" + original_name = "conv5_4_linear_bn_mean" shape = [64] dtype = "float32" min_val = float("0") @@ -1264,6 +1380,7 @@ class Program_weight_tensor_parameter_115: class Program_weight_tensor_parameter_116: name = "parameter_116" + original_name = "conv5_4_linear_weights" shape = [64, 384, 1, 1] dtype = "float32" min_val = float("-0.245534") @@ -1275,6 +1392,7 @@ class Program_weight_tensor_parameter_116: class Program_weight_tensor_parameter_117: name = "parameter_117" + original_name = "conv5_4_dwise_bn_offset" shape = [384] dtype = "float32" min_val = float("-0.46956") @@ -1286,6 +1404,7 @@ class Program_weight_tensor_parameter_117: class Program_weight_tensor_parameter_118: name = "parameter_118" + original_name = "conv5_4_dwise_bn_scale" shape = [384] dtype = "float32" min_val = float("-0.0351645") @@ -1297,6 +1416,7 @@ class Program_weight_tensor_parameter_118: class Program_weight_tensor_parameter_119: name = "parameter_119" + original_name = "conv5_4_dwise_bn_variance" shape = [384] dtype = "float32" min_val = float("5.60519e-45") @@ -1308,6 +1428,7 @@ class Program_weight_tensor_parameter_119: class Program_weight_tensor_parameter_120: name = "parameter_120" + original_name = "conv5_4_dwise_bn_mean" shape = [384] dtype = "float32" min_val = float("-0.152416") @@ -1319,6 +1440,7 @@ class Program_weight_tensor_parameter_120: class Program_weight_tensor_parameter_121: name = "parameter_121" + original_name = "conv5_4_dwise_weights" shape = [384, 1, 3, 3] dtype = "float32" min_val = float("-0.484497") @@ -1330,6 +1452,7 @@ class Program_weight_tensor_parameter_121: class Program_weight_tensor_parameter_122: name = "parameter_122" + original_name = "conv5_4_expand_bn_offset" shape = [384] dtype = "float32" min_val = float("-0.282772") @@ -1341,6 +1464,7 @@ class Program_weight_tensor_parameter_122: class Program_weight_tensor_parameter_123: name = "parameter_123" + original_name = "conv5_4_expand_bn_scale" shape = [384] dtype = "float32" min_val = float("-2.42021e-05") @@ -1352,6 +1476,7 @@ class Program_weight_tensor_parameter_123: class Program_weight_tensor_parameter_124: name = "parameter_124" + original_name = "conv5_4_expand_bn_variance" shape = [384] dtype = "float32" min_val = float("6.76194e-11") @@ -1363,6 +1488,7 @@ class Program_weight_tensor_parameter_124: class Program_weight_tensor_parameter_125: name = "parameter_125" + original_name = "conv5_4_expand_bn_mean" shape = [384] dtype = "float32" min_val = float("-0.04728") @@ -1374,6 +1500,7 @@ class Program_weight_tensor_parameter_125: class Program_weight_tensor_parameter_126: name = "parameter_126" + original_name = "conv5_4_expand_weights" shape = [384, 64, 1, 1] dtype = "float32" min_val = float("-0.335295") @@ -1385,6 +1512,7 @@ class Program_weight_tensor_parameter_126: class Program_weight_tensor_parameter_127: name = "parameter_127" + original_name = "conv5_3_linear_bn_offset" shape = [64] dtype = "float32" min_val = float("0") @@ -1394,6 +1522,7 @@ class Program_weight_tensor_parameter_127: class Program_weight_tensor_parameter_128: name = "parameter_128" + original_name = "conv5_3_linear_bn_scale" shape = [64] dtype = "float32" min_val = float("0") @@ -1403,6 +1532,7 @@ class Program_weight_tensor_parameter_128: class Program_weight_tensor_parameter_129: name = "parameter_129" + original_name = "conv5_3_linear_bn_variance" shape = [64] dtype = "float32" min_val = float("0") @@ -1412,6 +1542,7 @@ class Program_weight_tensor_parameter_129: class Program_weight_tensor_parameter_130: name = "parameter_130" + original_name = "conv5_3_linear_bn_mean" shape = [64] dtype = "float32" min_val = float("0") @@ -1421,6 +1552,7 @@ class Program_weight_tensor_parameter_130: class Program_weight_tensor_parameter_131: name = "parameter_131" + original_name = "conv5_3_linear_weights" shape = [64, 384, 1, 1] dtype = "float32" min_val = float("-0.38466") @@ -1432,6 +1564,7 @@ class Program_weight_tensor_parameter_131: class Program_weight_tensor_parameter_132: name = "parameter_132" + original_name = "conv5_3_dwise_bn_offset" shape = [384] dtype = "float32" min_val = float("-0.729309") @@ -1443,6 +1576,7 @@ class Program_weight_tensor_parameter_132: class Program_weight_tensor_parameter_133: name = "parameter_133" + original_name = "conv5_3_dwise_bn_scale" shape = [384] dtype = "float32" min_val = float("-0.000125833") @@ -1454,6 +1588,7 @@ class Program_weight_tensor_parameter_133: class Program_weight_tensor_parameter_134: name = "parameter_134" + original_name = "conv5_3_dwise_bn_variance" shape = [384] dtype = "float32" min_val = float("5.60519e-45") @@ -1465,6 +1600,7 @@ class Program_weight_tensor_parameter_134: class Program_weight_tensor_parameter_135: name = "parameter_135" + original_name = "conv5_3_dwise_bn_mean" shape = [384] dtype = "float32" min_val = float("-0.182929") @@ -1476,6 +1612,7 @@ class Program_weight_tensor_parameter_135: class Program_weight_tensor_parameter_136: name = "parameter_136" + original_name = "conv5_3_dwise_weights" shape = [384, 1, 3, 3] dtype = "float32" min_val = float("-0.470247") @@ -1487,6 +1624,7 @@ class Program_weight_tensor_parameter_136: class Program_weight_tensor_parameter_137: name = "parameter_137" + original_name = "conv5_3_expand_bn_offset" shape = [384] dtype = "float32" min_val = float("-0.268102") @@ -1498,6 +1636,7 @@ class Program_weight_tensor_parameter_137: class Program_weight_tensor_parameter_138: name = "parameter_138" + original_name = "conv5_3_expand_bn_scale" shape = [384] dtype = "float32" min_val = float("-5.22238e-07") @@ -1509,6 +1648,7 @@ class Program_weight_tensor_parameter_138: class Program_weight_tensor_parameter_139: name = "parameter_139" + original_name = "conv5_3_expand_bn_variance" shape = [384] dtype = "float32" min_val = float("1.05776e-10") @@ -1520,6 +1660,7 @@ class Program_weight_tensor_parameter_139: class Program_weight_tensor_parameter_140: name = "parameter_140" + original_name = "conv5_3_expand_bn_mean" shape = [384] dtype = "float32" min_val = float("-0.0424079") @@ -1531,6 +1672,7 @@ class Program_weight_tensor_parameter_140: class Program_weight_tensor_parameter_141: name = "parameter_141" + original_name = "conv5_3_expand_weights" shape = [384, 64, 1, 1] dtype = "float32" min_val = float("-0.403788") @@ -1542,6 +1684,7 @@ class Program_weight_tensor_parameter_141: class Program_weight_tensor_parameter_142: name = "parameter_142" + original_name = "conv5_2_linear_bn_offset" shape = [64] dtype = "float32" min_val = float("0") @@ -1551,6 +1694,7 @@ class Program_weight_tensor_parameter_142: class Program_weight_tensor_parameter_143: name = "parameter_143" + original_name = "conv5_2_linear_bn_scale" shape = [64] dtype = "float32" min_val = float("0") @@ -1560,6 +1704,7 @@ class Program_weight_tensor_parameter_143: class Program_weight_tensor_parameter_144: name = "parameter_144" + original_name = "conv5_2_linear_bn_variance" shape = [64] dtype = "float32" min_val = float("0") @@ -1569,6 +1714,7 @@ class Program_weight_tensor_parameter_144: class Program_weight_tensor_parameter_145: name = "parameter_145" + original_name = "conv5_2_linear_bn_mean" shape = [64] dtype = "float32" min_val = float("0") @@ -1578,6 +1724,7 @@ class Program_weight_tensor_parameter_145: class Program_weight_tensor_parameter_146: name = "parameter_146" + original_name = "conv5_2_linear_weights" shape = [64, 384, 1, 1] dtype = "float32" min_val = float("-0.340783") @@ -1589,6 +1736,7 @@ class Program_weight_tensor_parameter_146: class Program_weight_tensor_parameter_147: name = "parameter_147" + original_name = "conv5_2_dwise_bn_offset" shape = [384] dtype = "float32" min_val = float("-0.455545") @@ -1600,6 +1748,7 @@ class Program_weight_tensor_parameter_147: class Program_weight_tensor_parameter_148: name = "parameter_148" + original_name = "conv5_2_dwise_bn_scale" shape = [384] dtype = "float32" min_val = float("-0.000546031") @@ -1611,6 +1760,7 @@ class Program_weight_tensor_parameter_148: class Program_weight_tensor_parameter_149: name = "parameter_149" + original_name = "conv5_2_dwise_bn_variance" shape = [384] dtype = "float32" min_val = float("5.60519e-45") @@ -1622,6 +1772,7 @@ class Program_weight_tensor_parameter_149: class Program_weight_tensor_parameter_150: name = "parameter_150" + original_name = "conv5_2_dwise_bn_mean" shape = [384] dtype = "float32" min_val = float("-0.158135") @@ -1633,6 +1784,7 @@ class Program_weight_tensor_parameter_150: class Program_weight_tensor_parameter_151: name = "parameter_151" + original_name = "conv5_2_dwise_weights" shape = [384, 1, 3, 3] dtype = "float32" min_val = float("-0.661338") @@ -1644,6 +1796,7 @@ class Program_weight_tensor_parameter_151: class Program_weight_tensor_parameter_152: name = "parameter_152" + original_name = "conv5_2_expand_bn_offset" shape = [384] dtype = "float32" min_val = float("-0.254279") @@ -1655,6 +1808,7 @@ class Program_weight_tensor_parameter_152: class Program_weight_tensor_parameter_153: name = "parameter_153" + original_name = "conv5_2_expand_bn_scale" shape = [384] dtype = "float32" min_val = float("-8.93419e-06") @@ -1666,6 +1820,7 @@ class Program_weight_tensor_parameter_153: class Program_weight_tensor_parameter_154: name = "parameter_154" + original_name = "conv5_2_expand_bn_variance" shape = [384] dtype = "float32" min_val = float("8.73098e-11") @@ -1677,6 +1832,7 @@ class Program_weight_tensor_parameter_154: class Program_weight_tensor_parameter_155: name = "parameter_155" + original_name = "conv5_2_expand_bn_mean" shape = [384] dtype = "float32" min_val = float("-0.014719") @@ -1688,6 +1844,7 @@ class Program_weight_tensor_parameter_155: class Program_weight_tensor_parameter_156: name = "parameter_156" + original_name = "conv5_2_expand_weights" shape = [384, 64, 1, 1] dtype = "float32" min_val = float("-0.306903") @@ -1699,6 +1856,7 @@ class Program_weight_tensor_parameter_156: class Program_weight_tensor_parameter_157: name = "parameter_157" + original_name = "conv5_1_linear_bn_offset" shape = [64] dtype = "float32" min_val = float("0") @@ -1708,6 +1866,7 @@ class Program_weight_tensor_parameter_157: class Program_weight_tensor_parameter_158: name = "parameter_158" + original_name = "conv5_1_linear_bn_scale" shape = [64] dtype = "float32" min_val = float("0") @@ -1717,6 +1876,7 @@ class Program_weight_tensor_parameter_158: class Program_weight_tensor_parameter_159: name = "parameter_159" + original_name = "conv5_1_linear_bn_variance" shape = [64] dtype = "float32" min_val = float("0") @@ -1726,6 +1886,7 @@ class Program_weight_tensor_parameter_159: class Program_weight_tensor_parameter_160: name = "parameter_160" + original_name = "conv5_1_linear_bn_mean" shape = [64] dtype = "float32" min_val = float("0") @@ -1735,6 +1896,7 @@ class Program_weight_tensor_parameter_160: class Program_weight_tensor_parameter_161: name = "parameter_161" + original_name = "conv5_1_linear_weights" shape = [64, 192, 1, 1] dtype = "float32" min_val = float("-0.498261") @@ -1746,6 +1908,7 @@ class Program_weight_tensor_parameter_161: class Program_weight_tensor_parameter_162: name = "parameter_162" + original_name = "conv5_1_dwise_bn_offset" shape = [192] dtype = "float32" min_val = float("-0.0977629") @@ -1757,6 +1920,7 @@ class Program_weight_tensor_parameter_162: class Program_weight_tensor_parameter_163: name = "parameter_163" + original_name = "conv5_1_dwise_bn_scale" shape = [192] dtype = "float32" min_val = float("0.167084") @@ -1768,6 +1932,7 @@ class Program_weight_tensor_parameter_163: class Program_weight_tensor_parameter_164: name = "parameter_164" + original_name = "conv5_1_dwise_bn_variance" shape = [192] dtype = "float32" min_val = float("0.00114") @@ -1779,6 +1944,7 @@ class Program_weight_tensor_parameter_164: class Program_weight_tensor_parameter_165: name = "parameter_165" + original_name = "conv5_1_dwise_bn_mean" shape = [192] dtype = "float32" min_val = float("-0.411906") @@ -1790,6 +1956,7 @@ class Program_weight_tensor_parameter_165: class Program_weight_tensor_parameter_166: name = "parameter_166" + original_name = "conv5_1_dwise_weights" shape = [192, 1, 3, 3] dtype = "float32" min_val = float("-0.342455") @@ -1801,6 +1968,7 @@ class Program_weight_tensor_parameter_166: class Program_weight_tensor_parameter_167: name = "parameter_167" + original_name = "conv5_1_expand_bn_offset" shape = [192] dtype = "float32" min_val = float("-0.319807") @@ -1812,6 +1980,7 @@ class Program_weight_tensor_parameter_167: class Program_weight_tensor_parameter_168: name = "parameter_168" + original_name = "conv5_1_expand_bn_scale" shape = [192] dtype = "float32" min_val = float("0.0696795") @@ -1823,6 +1992,7 @@ class Program_weight_tensor_parameter_168: class Program_weight_tensor_parameter_169: name = "parameter_169" + original_name = "conv5_1_expand_bn_variance" shape = [192] dtype = "float32" min_val = float("0.213599") @@ -1834,6 +2004,7 @@ class Program_weight_tensor_parameter_169: class Program_weight_tensor_parameter_170: name = "parameter_170" + original_name = "conv5_1_expand_bn_mean" shape = [192] dtype = "float32" min_val = float("-0.0279844") @@ -1845,6 +2016,7 @@ class Program_weight_tensor_parameter_170: class Program_weight_tensor_parameter_171: name = "parameter_171" + original_name = "conv5_1_expand_weights" shape = [192, 32, 1, 1] dtype = "float32" min_val = float("-0.710155") @@ -1856,6 +2028,7 @@ class Program_weight_tensor_parameter_171: class Program_weight_tensor_parameter_172: name = "parameter_172" + original_name = "conv4_3_linear_bn_offset" shape = [32] dtype = "float32" min_val = float("0") @@ -1865,6 +2038,7 @@ class Program_weight_tensor_parameter_172: class Program_weight_tensor_parameter_173: name = "parameter_173" + original_name = "conv4_3_linear_bn_scale" shape = [32] dtype = "float32" min_val = float("0") @@ -1874,6 +2048,7 @@ class Program_weight_tensor_parameter_173: class Program_weight_tensor_parameter_174: name = "parameter_174" + original_name = "conv4_3_linear_bn_variance" shape = [32] dtype = "float32" min_val = float("0") @@ -1883,6 +2058,7 @@ class Program_weight_tensor_parameter_174: class Program_weight_tensor_parameter_175: name = "parameter_175" + original_name = "conv4_3_linear_bn_mean" shape = [32] dtype = "float32" min_val = float("0") @@ -1892,6 +2068,7 @@ class Program_weight_tensor_parameter_175: class Program_weight_tensor_parameter_176: name = "parameter_176" + original_name = "conv4_3_linear_weights" shape = [32, 192, 1, 1] dtype = "float32" min_val = float("-0.398713") @@ -1903,6 +2080,7 @@ class Program_weight_tensor_parameter_176: class Program_weight_tensor_parameter_177: name = "parameter_177" + original_name = "conv4_3_dwise_bn_offset" shape = [192] dtype = "float32" min_val = float("-0.488235") @@ -1914,6 +2092,7 @@ class Program_weight_tensor_parameter_177: class Program_weight_tensor_parameter_178: name = "parameter_178" + original_name = "conv4_3_dwise_bn_scale" shape = [192] dtype = "float32" min_val = float("1.12696e-05") @@ -1925,6 +2104,7 @@ class Program_weight_tensor_parameter_178: class Program_weight_tensor_parameter_179: name = "parameter_179" + original_name = "conv4_3_dwise_bn_variance" shape = [192] dtype = "float32" min_val = float("5.60519e-45") @@ -1936,6 +2116,7 @@ class Program_weight_tensor_parameter_179: class Program_weight_tensor_parameter_180: name = "parameter_180" + original_name = "conv4_3_dwise_bn_mean" shape = [192] dtype = "float32" min_val = float("-0.303607") @@ -1947,6 +2128,7 @@ class Program_weight_tensor_parameter_180: class Program_weight_tensor_parameter_181: name = "parameter_181" + original_name = "conv4_3_dwise_weights" shape = [192, 1, 3, 3] dtype = "float32" min_val = float("-0.625628") @@ -1958,6 +2140,7 @@ class Program_weight_tensor_parameter_181: class Program_weight_tensor_parameter_182: name = "parameter_182" + original_name = "conv4_3_expand_bn_offset" shape = [192] dtype = "float32" min_val = float("-0.254761") @@ -1969,6 +2152,7 @@ class Program_weight_tensor_parameter_182: class Program_weight_tensor_parameter_183: name = "parameter_183" + original_name = "conv4_3_expand_bn_scale" shape = [192] dtype = "float32" min_val = float("2.31088e-06") @@ -1980,6 +2164,7 @@ class Program_weight_tensor_parameter_183: class Program_weight_tensor_parameter_184: name = "parameter_184" + original_name = "conv4_3_expand_bn_variance" shape = [192] dtype = "float32" min_val = float("5.80786e-10") @@ -1991,6 +2176,7 @@ class Program_weight_tensor_parameter_184: class Program_weight_tensor_parameter_185: name = "parameter_185" + original_name = "conv4_3_expand_bn_mean" shape = [192] dtype = "float32" min_val = float("-0.0364331") @@ -2002,6 +2188,7 @@ class Program_weight_tensor_parameter_185: class Program_weight_tensor_parameter_186: name = "parameter_186" + original_name = "conv4_3_expand_weights" shape = [192, 32, 1, 1] dtype = "float32" min_val = float("-0.413455") @@ -2013,6 +2200,7 @@ class Program_weight_tensor_parameter_186: class Program_weight_tensor_parameter_187: name = "parameter_187" + original_name = "conv4_2_linear_bn_offset" shape = [32] dtype = "float32" min_val = float("0") @@ -2022,6 +2210,7 @@ class Program_weight_tensor_parameter_187: class Program_weight_tensor_parameter_188: name = "parameter_188" + original_name = "conv4_2_linear_bn_scale" shape = [32] dtype = "float32" min_val = float("0") @@ -2031,6 +2220,7 @@ class Program_weight_tensor_parameter_188: class Program_weight_tensor_parameter_189: name = "parameter_189" + original_name = "conv4_2_linear_bn_variance" shape = [32] dtype = "float32" min_val = float("0") @@ -2040,6 +2230,7 @@ class Program_weight_tensor_parameter_189: class Program_weight_tensor_parameter_190: name = "parameter_190" + original_name = "conv4_2_linear_bn_mean" shape = [32] dtype = "float32" min_val = float("0") @@ -2049,6 +2240,7 @@ class Program_weight_tensor_parameter_190: class Program_weight_tensor_parameter_191: name = "parameter_191" + original_name = "conv4_2_linear_weights" shape = [32, 192, 1, 1] dtype = "float32" min_val = float("-0.39256") @@ -2060,6 +2252,7 @@ class Program_weight_tensor_parameter_191: class Program_weight_tensor_parameter_192: name = "parameter_192" + original_name = "conv4_2_dwise_bn_offset" shape = [192] dtype = "float32" min_val = float("-0.462771") @@ -2071,6 +2264,7 @@ class Program_weight_tensor_parameter_192: class Program_weight_tensor_parameter_193: name = "parameter_193" + original_name = "conv4_2_dwise_bn_scale" shape = [192] dtype = "float32" min_val = float("6.46046e-06") @@ -2082,6 +2276,7 @@ class Program_weight_tensor_parameter_193: class Program_weight_tensor_parameter_194: name = "parameter_194" + original_name = "conv4_2_dwise_bn_variance" shape = [192] dtype = "float32" min_val = float("5.60519e-45") @@ -2093,6 +2288,7 @@ class Program_weight_tensor_parameter_194: class Program_weight_tensor_parameter_195: name = "parameter_195" + original_name = "conv4_2_dwise_bn_mean" shape = [192] dtype = "float32" min_val = float("-0.190409") @@ -2104,6 +2300,7 @@ class Program_weight_tensor_parameter_195: class Program_weight_tensor_parameter_196: name = "parameter_196" + original_name = "conv4_2_dwise_weights" shape = [192, 1, 3, 3] dtype = "float32" min_val = float("-0.681306") @@ -2115,6 +2312,7 @@ class Program_weight_tensor_parameter_196: class Program_weight_tensor_parameter_197: name = "parameter_197" + original_name = "conv4_2_expand_bn_offset" shape = [192] dtype = "float32" min_val = float("-0.3752") @@ -2126,6 +2324,7 @@ class Program_weight_tensor_parameter_197: class Program_weight_tensor_parameter_198: name = "parameter_198" + original_name = "conv4_2_expand_bn_scale" shape = [192] dtype = "float32" min_val = float("1.68656e-06") @@ -2137,6 +2336,7 @@ class Program_weight_tensor_parameter_198: class Program_weight_tensor_parameter_199: name = "parameter_199" + original_name = "conv4_2_expand_bn_variance" shape = [192] dtype = "float32" min_val = float("1.54172e-10") @@ -2148,6 +2348,7 @@ class Program_weight_tensor_parameter_199: class Program_weight_tensor_parameter_200: name = "parameter_200" + original_name = "conv4_2_expand_bn_mean" shape = [192] dtype = "float32" min_val = float("-0.0294191") @@ -2159,6 +2360,7 @@ class Program_weight_tensor_parameter_200: class Program_weight_tensor_parameter_201: name = "parameter_201" + original_name = "conv4_2_expand_weights" shape = [192, 32, 1, 1] dtype = "float32" min_val = float("-0.48809") @@ -2170,6 +2372,7 @@ class Program_weight_tensor_parameter_201: class Program_weight_tensor_parameter_202: name = "parameter_202" + original_name = "conv4_1_linear_bn_offset" shape = [32] dtype = "float32" min_val = float("0") @@ -2179,6 +2382,7 @@ class Program_weight_tensor_parameter_202: class Program_weight_tensor_parameter_203: name = "parameter_203" + original_name = "conv4_1_linear_bn_scale" shape = [32] dtype = "float32" min_val = float("0") @@ -2188,6 +2392,7 @@ class Program_weight_tensor_parameter_203: class Program_weight_tensor_parameter_204: name = "parameter_204" + original_name = "conv4_1_linear_bn_variance" shape = [32] dtype = "float32" min_val = float("0") @@ -2197,6 +2402,7 @@ class Program_weight_tensor_parameter_204: class Program_weight_tensor_parameter_205: name = "parameter_205" + original_name = "conv4_1_linear_bn_mean" shape = [32] dtype = "float32" min_val = float("0") @@ -2206,6 +2412,7 @@ class Program_weight_tensor_parameter_205: class Program_weight_tensor_parameter_206: name = "parameter_206" + original_name = "conv4_1_linear_weights" shape = [32, 144, 1, 1] dtype = "float32" min_val = float("-0.912872") @@ -2217,6 +2424,7 @@ class Program_weight_tensor_parameter_206: class Program_weight_tensor_parameter_207: name = "parameter_207" + original_name = "conv4_1_dwise_bn_offset" shape = [144] dtype = "float32" min_val = float("-0.261459") @@ -2228,6 +2436,7 @@ class Program_weight_tensor_parameter_207: class Program_weight_tensor_parameter_208: name = "parameter_208" + original_name = "conv4_1_dwise_bn_scale" shape = [144] dtype = "float32" min_val = float("0.154565") @@ -2239,6 +2448,7 @@ class Program_weight_tensor_parameter_208: class Program_weight_tensor_parameter_209: name = "parameter_209" + original_name = "conv4_1_dwise_bn_variance" shape = [144] dtype = "float32" min_val = float("0.000328699") @@ -2250,6 +2460,7 @@ class Program_weight_tensor_parameter_209: class Program_weight_tensor_parameter_210: name = "parameter_210" + original_name = "conv4_1_dwise_bn_mean" shape = [144] dtype = "float32" min_val = float("-0.481826") @@ -2261,6 +2472,7 @@ class Program_weight_tensor_parameter_210: class Program_weight_tensor_parameter_211: name = "parameter_211" + original_name = "conv4_1_dwise_weights" shape = [144, 1, 3, 3] dtype = "float32" min_val = float("-0.347734") @@ -2272,6 +2484,7 @@ class Program_weight_tensor_parameter_211: class Program_weight_tensor_parameter_212: name = "parameter_212" + original_name = "conv4_1_expand_bn_offset" shape = [144] dtype = "float32" min_val = float("-0.430761") @@ -2283,6 +2496,7 @@ class Program_weight_tensor_parameter_212: class Program_weight_tensor_parameter_213: name = "parameter_213" + original_name = "conv4_1_expand_bn_scale" shape = [144] dtype = "float32" min_val = float("0.0705883") @@ -2294,6 +2508,7 @@ class Program_weight_tensor_parameter_213: class Program_weight_tensor_parameter_214: name = "parameter_214" + original_name = "conv4_1_expand_bn_variance" shape = [144] dtype = "float32" min_val = float("0.191773") @@ -2305,6 +2520,7 @@ class Program_weight_tensor_parameter_214: class Program_weight_tensor_parameter_215: name = "parameter_215" + original_name = "conv4_1_expand_bn_mean" shape = [144] dtype = "float32" min_val = float("-0.0875977") @@ -2316,6 +2532,7 @@ class Program_weight_tensor_parameter_215: class Program_weight_tensor_parameter_216: name = "parameter_216" + original_name = "conv4_1_expand_weights" shape = [144, 24, 1, 1] dtype = "float32" min_val = float("-1.17786") @@ -2327,6 +2544,7 @@ class Program_weight_tensor_parameter_216: class Program_weight_tensor_parameter_217: name = "parameter_217" + original_name = "conv3_2_linear_bn_offset" shape = [24] dtype = "float32" min_val = float("0") @@ -2336,6 +2554,7 @@ class Program_weight_tensor_parameter_217: class Program_weight_tensor_parameter_218: name = "parameter_218" + original_name = "conv3_2_linear_bn_scale" shape = [24] dtype = "float32" min_val = float("0") @@ -2345,6 +2564,7 @@ class Program_weight_tensor_parameter_218: class Program_weight_tensor_parameter_219: name = "parameter_219" + original_name = "conv3_2_linear_bn_variance" shape = [24] dtype = "float32" min_val = float("0") @@ -2354,6 +2574,7 @@ class Program_weight_tensor_parameter_219: class Program_weight_tensor_parameter_220: name = "parameter_220" + original_name = "conv3_2_linear_bn_mean" shape = [24] dtype = "float32" min_val = float("0") @@ -2363,6 +2584,7 @@ class Program_weight_tensor_parameter_220: class Program_weight_tensor_parameter_221: name = "parameter_221" + original_name = "conv3_2_linear_weights" shape = [24, 144, 1, 1] dtype = "float32" min_val = float("-0.651639") @@ -2374,6 +2596,7 @@ class Program_weight_tensor_parameter_221: class Program_weight_tensor_parameter_222: name = "parameter_222" + original_name = "conv3_2_dwise_bn_offset" shape = [144] dtype = "float32" min_val = float("-0.41175") @@ -2385,6 +2608,7 @@ class Program_weight_tensor_parameter_222: class Program_weight_tensor_parameter_223: name = "parameter_223" + original_name = "conv3_2_dwise_bn_scale" shape = [144] dtype = "float32" min_val = float("1.55025e-05") @@ -2396,6 +2620,7 @@ class Program_weight_tensor_parameter_223: class Program_weight_tensor_parameter_224: name = "parameter_224" + original_name = "conv3_2_dwise_bn_variance" shape = [144] dtype = "float32" min_val = float("5.60519e-45") @@ -2407,6 +2632,7 @@ class Program_weight_tensor_parameter_224: class Program_weight_tensor_parameter_225: name = "parameter_225" + original_name = "conv3_2_dwise_bn_mean" shape = [144] dtype = "float32" min_val = float("-0.811784") @@ -2418,6 +2644,7 @@ class Program_weight_tensor_parameter_225: class Program_weight_tensor_parameter_226: name = "parameter_226" + original_name = "conv3_2_dwise_weights" shape = [144, 1, 3, 3] dtype = "float32" min_val = float("-0.93789") @@ -2429,6 +2656,7 @@ class Program_weight_tensor_parameter_226: class Program_weight_tensor_parameter_227: name = "parameter_227" + original_name = "conv3_2_expand_bn_offset" shape = [144] dtype = "float32" min_val = float("-0.38103") @@ -2440,6 +2668,7 @@ class Program_weight_tensor_parameter_227: class Program_weight_tensor_parameter_228: name = "parameter_228" + original_name = "conv3_2_expand_bn_scale" shape = [144] dtype = "float32" min_val = float("5.12993e-06") @@ -2451,6 +2680,7 @@ class Program_weight_tensor_parameter_228: class Program_weight_tensor_parameter_229: name = "parameter_229" + original_name = "conv3_2_expand_bn_variance" shape = [144] dtype = "float32" min_val = float("9.8049e-10") @@ -2462,6 +2692,7 @@ class Program_weight_tensor_parameter_229: class Program_weight_tensor_parameter_230: name = "parameter_230" + original_name = "conv3_2_expand_bn_mean" shape = [144] dtype = "float32" min_val = float("-0.057612") @@ -2473,6 +2704,7 @@ class Program_weight_tensor_parameter_230: class Program_weight_tensor_parameter_231: name = "parameter_231" + original_name = "conv3_2_expand_weights" shape = [144, 24, 1, 1] dtype = "float32" min_val = float("-0.66654") @@ -2484,6 +2716,7 @@ class Program_weight_tensor_parameter_231: class Program_weight_tensor_parameter_232: name = "parameter_232" + original_name = "conv3_1_linear_bn_offset" shape = [24] dtype = "float32" min_val = float("0") @@ -2493,6 +2726,7 @@ class Program_weight_tensor_parameter_232: class Program_weight_tensor_parameter_233: name = "parameter_233" + original_name = "conv3_1_linear_bn_scale" shape = [24] dtype = "float32" min_val = float("0") @@ -2502,6 +2736,7 @@ class Program_weight_tensor_parameter_233: class Program_weight_tensor_parameter_234: name = "parameter_234" + original_name = "conv3_1_linear_bn_variance" shape = [24] dtype = "float32" min_val = float("0") @@ -2511,6 +2746,7 @@ class Program_weight_tensor_parameter_234: class Program_weight_tensor_parameter_235: name = "parameter_235" + original_name = "conv3_1_linear_bn_mean" shape = [24] dtype = "float32" min_val = float("0") @@ -2520,6 +2756,7 @@ class Program_weight_tensor_parameter_235: class Program_weight_tensor_parameter_236: name = "parameter_236" + original_name = "conv3_1_linear_weights" shape = [24, 96, 1, 1] dtype = "float32" min_val = float("-1.33829") @@ -2531,6 +2768,7 @@ class Program_weight_tensor_parameter_236: class Program_weight_tensor_parameter_237: name = "parameter_237" + original_name = "conv3_1_dwise_bn_offset" shape = [96] dtype = "float32" min_val = float("-0.360715") @@ -2542,6 +2780,7 @@ class Program_weight_tensor_parameter_237: class Program_weight_tensor_parameter_238: name = "parameter_238" + original_name = "conv3_1_dwise_bn_scale" shape = [96] dtype = "float32" min_val = float("0.119495") @@ -2553,6 +2792,7 @@ class Program_weight_tensor_parameter_238: class Program_weight_tensor_parameter_239: name = "parameter_239" + original_name = "conv3_1_dwise_bn_variance" shape = [96] dtype = "float32" min_val = float("0.001568") @@ -2564,6 +2804,7 @@ class Program_weight_tensor_parameter_239: class Program_weight_tensor_parameter_240: name = "parameter_240" + original_name = "conv3_1_dwise_bn_mean" shape = [96] dtype = "float32" min_val = float("-0.984839") @@ -2575,6 +2816,7 @@ class Program_weight_tensor_parameter_240: class Program_weight_tensor_parameter_241: name = "parameter_241" + original_name = "conv3_1_dwise_weights" shape = [96, 1, 3, 3] dtype = "float32" min_val = float("-0.465258") @@ -2586,6 +2828,7 @@ class Program_weight_tensor_parameter_241: class Program_weight_tensor_parameter_242: name = "parameter_242" + original_name = "conv3_1_expand_bn_offset" shape = [96] dtype = "float32" min_val = float("-0.466104") @@ -2597,6 +2840,7 @@ class Program_weight_tensor_parameter_242: class Program_weight_tensor_parameter_243: name = "parameter_243" + original_name = "conv3_1_expand_bn_scale" shape = [96] dtype = "float32" min_val = float("0.0854064") @@ -2608,6 +2852,7 @@ class Program_weight_tensor_parameter_243: class Program_weight_tensor_parameter_244: name = "parameter_244" + original_name = "conv3_1_expand_bn_variance" shape = [96] dtype = "float32" min_val = float("0.0642253") @@ -2619,6 +2864,7 @@ class Program_weight_tensor_parameter_244: class Program_weight_tensor_parameter_245: name = "parameter_245" + original_name = "conv3_1_expand_bn_mean" shape = [96] dtype = "float32" min_val = float("-8.58092e-06") @@ -2630,6 +2876,7 @@ class Program_weight_tensor_parameter_245: class Program_weight_tensor_parameter_246: name = "parameter_246" + original_name = "conv3_1_expand_weights" shape = [96, 16, 1, 1] dtype = "float32" min_val = float("-0.98358") @@ -2641,6 +2888,7 @@ class Program_weight_tensor_parameter_246: class Program_weight_tensor_parameter_247: name = "parameter_247" + original_name = "conv2_1_linear_bn_offset" shape = [16] dtype = "float32" min_val = float("0") @@ -2650,6 +2898,7 @@ class Program_weight_tensor_parameter_247: class Program_weight_tensor_parameter_248: name = "parameter_248" + original_name = "conv2_1_linear_bn_scale" shape = [16] dtype = "float32" min_val = float("0") @@ -2659,6 +2908,7 @@ class Program_weight_tensor_parameter_248: class Program_weight_tensor_parameter_249: name = "parameter_249" + original_name = "conv2_1_linear_bn_variance" shape = [16] dtype = "float32" min_val = float("0") @@ -2668,6 +2918,7 @@ class Program_weight_tensor_parameter_249: class Program_weight_tensor_parameter_250: name = "parameter_250" + original_name = "conv2_1_linear_bn_mean" shape = [16] dtype = "float32" min_val = float("0") @@ -2677,6 +2928,7 @@ class Program_weight_tensor_parameter_250: class Program_weight_tensor_parameter_251: name = "parameter_251" + original_name = "conv2_1_linear_weights" shape = [16, 32, 1, 1] dtype = "float32" min_val = float("-1.13687") @@ -2688,6 +2940,7 @@ class Program_weight_tensor_parameter_251: class Program_weight_tensor_parameter_252: name = "parameter_252" + original_name = "conv2_1_dwise_bn_offset" shape = [32] dtype = "float32" min_val = float("0") @@ -2697,6 +2950,7 @@ class Program_weight_tensor_parameter_252: class Program_weight_tensor_parameter_253: name = "parameter_253" + original_name = "conv2_1_dwise_bn_scale" shape = [32] dtype = "float32" min_val = float("0") @@ -2706,6 +2960,7 @@ class Program_weight_tensor_parameter_253: class Program_weight_tensor_parameter_254: name = "parameter_254" + original_name = "conv2_1_dwise_bn_variance" shape = [32] dtype = "float32" min_val = float("0") @@ -2715,6 +2970,7 @@ class Program_weight_tensor_parameter_254: class Program_weight_tensor_parameter_255: name = "parameter_255" + original_name = "conv2_1_dwise_bn_mean" shape = [32] dtype = "float32" min_val = float("0") @@ -2724,6 +2980,7 @@ class Program_weight_tensor_parameter_255: class Program_weight_tensor_parameter_256: name = "parameter_256" + original_name = "conv2_1_dwise_weights" shape = [32, 1, 3, 3] dtype = "float32" min_val = float("-2.39341") @@ -2735,6 +2992,7 @@ class Program_weight_tensor_parameter_256: class Program_weight_tensor_parameter_257: name = "parameter_257" + original_name = "conv2_1_expand_bn_offset" shape = [32] dtype = "float32" min_val = float("0") @@ -2744,6 +3002,7 @@ class Program_weight_tensor_parameter_257: class Program_weight_tensor_parameter_258: name = "parameter_258" + original_name = "conv2_1_expand_bn_scale" shape = [32] dtype = "float32" min_val = float("0") @@ -2753,6 +3012,7 @@ class Program_weight_tensor_parameter_258: class Program_weight_tensor_parameter_259: name = "parameter_259" + original_name = "conv2_1_expand_bn_variance" shape = [32] dtype = "float32" min_val = float("0") @@ -2762,6 +3022,7 @@ class Program_weight_tensor_parameter_259: class Program_weight_tensor_parameter_260: name = "parameter_260" + original_name = "conv2_1_expand_bn_mean" shape = [32] dtype = "float32" min_val = float("0") @@ -2771,6 +3032,7 @@ class Program_weight_tensor_parameter_260: class Program_weight_tensor_parameter_261: name = "parameter_261" + original_name = "conv2_1_expand_weights" shape = [32, 32, 1, 1] dtype = "float32" min_val = float("-1.06852") @@ -2782,6 +3044,7 @@ class Program_weight_tensor_parameter_261: class Program_weight_tensor_parameter_262: name = "parameter_262" + original_name = "conv1_1_bn_offset" shape = [32] dtype = "float32" min_val = float("0") @@ -2791,6 +3054,7 @@ class Program_weight_tensor_parameter_262: class Program_weight_tensor_parameter_263: name = "parameter_263" + original_name = "conv1_1_bn_scale" shape = [32] dtype = "float32" min_val = float("0") @@ -2800,6 +3064,7 @@ class Program_weight_tensor_parameter_263: class Program_weight_tensor_parameter_264: name = "parameter_264" + original_name = "conv1_1_bn_variance" shape = [32] dtype = "float32" min_val = float("0") @@ -2809,6 +3074,7 @@ class Program_weight_tensor_parameter_264: class Program_weight_tensor_parameter_265: name = "parameter_265" + original_name = "conv1_1_bn_mean" shape = [32] dtype = "float32" min_val = float("0") @@ -2818,6 +3084,7 @@ class Program_weight_tensor_parameter_265: class Program_weight_tensor_parameter_266: name = "parameter_266" + original_name = "conv1_1_weights" shape = [32, 3, 3, 3] dtype = "float32" min_val = float("-0.967471") diff --git a/paddle_samples/PaddleX/MobileNetV2_x2_0/input_meta.py b/paddle_samples/PaddleX/MobileNetV2_x2_0/input_meta.py index 795f9bcfe..6d7dac1c9 100644 --- a/paddle_samples/PaddleX/MobileNetV2_x2_0/input_meta.py +++ b/paddle_samples/PaddleX/MobileNetV2_x2_0/input_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_data_0: name = "data_0" + original_name = "var_0" shape = [128, 3, 224, 224] dtype = "float32" min_val = float("-2.1179") diff --git a/paddle_samples/PaddleX/MobileNetV2_x2_0/weight_meta.py b/paddle_samples/PaddleX/MobileNetV2_x2_0/weight_meta.py index 06b4ab8c7..be6c17cb5 100644 --- a/paddle_samples/PaddleX/MobileNetV2_x2_0/weight_meta.py +++ b/paddle_samples/PaddleX/MobileNetV2_x2_0/weight_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_parameter_0: name = "parameter_0" + original_name = "fc10_offset" shape = [102] dtype = "float32" data = None @@ -7,6 +8,7 @@ class Program_weight_tensor_parameter_0: class Program_weight_tensor_parameter_1: name = "parameter_1" + original_name = "fc10_weights" shape = [2560, 102] dtype = "float32" min_val = float("-0.0474755") @@ -18,6 +20,7 @@ class Program_weight_tensor_parameter_1: class Program_weight_tensor_parameter_2: name = "parameter_2" + original_name = "conv9_bn_offset" shape = [2560] dtype = "float32" min_val = float("-0.409093") @@ -29,6 +32,7 @@ class Program_weight_tensor_parameter_2: class Program_weight_tensor_parameter_3: name = "parameter_3" + original_name = "conv9_bn_scale" shape = [2560] dtype = "float32" min_val = float("1.07225") @@ -40,6 +44,7 @@ class Program_weight_tensor_parameter_3: class Program_weight_tensor_parameter_4: name = "parameter_4" + original_name = "conv9_bn_variance" shape = [2560] dtype = "float32" min_val = float("0.0212158") @@ -51,6 +56,7 @@ class Program_weight_tensor_parameter_4: class Program_weight_tensor_parameter_5: name = "parameter_5" + original_name = "conv9_bn_mean" shape = [2560] dtype = "float32" min_val = float("-5.88004e-07") @@ -62,6 +68,7 @@ class Program_weight_tensor_parameter_5: class Program_weight_tensor_parameter_6: name = "parameter_6" + original_name = "conv9_weights" shape = [2560, 640, 1, 1] dtype = "float32" min_val = float("-0.138165") @@ -73,6 +80,7 @@ class Program_weight_tensor_parameter_6: class Program_weight_tensor_parameter_7: name = "parameter_7" + original_name = "conv8_1_linear_bn_offset" shape = [640] dtype = "float32" min_val = float("-7.93787e-07") @@ -84,6 +92,7 @@ class Program_weight_tensor_parameter_7: class Program_weight_tensor_parameter_8: name = "parameter_8" + original_name = "conv8_1_linear_bn_scale" shape = [640] dtype = "float32" min_val = float("0.154623") @@ -95,6 +104,7 @@ class Program_weight_tensor_parameter_8: class Program_weight_tensor_parameter_9: name = "parameter_9" + original_name = "conv8_1_linear_bn_variance" shape = [640] dtype = "float32" min_val = float("0.00456705") @@ -106,6 +116,7 @@ class Program_weight_tensor_parameter_9: class Program_weight_tensor_parameter_10: name = "parameter_10" + original_name = "conv8_1_linear_bn_mean" shape = [640] dtype = "float32" min_val = float("-0.870601") @@ -117,6 +128,7 @@ class Program_weight_tensor_parameter_10: class Program_weight_tensor_parameter_11: name = "parameter_11" + original_name = "conv8_1_linear_weights" shape = [640, 1920, 1, 1] dtype = "float32" min_val = float("-0.243989") @@ -128,6 +140,7 @@ class Program_weight_tensor_parameter_11: class Program_weight_tensor_parameter_12: name = "parameter_12" + original_name = "conv8_1_dwise_bn_offset" shape = [1920] dtype = "float32" min_val = float("-0.259583") @@ -139,6 +152,7 @@ class Program_weight_tensor_parameter_12: class Program_weight_tensor_parameter_13: name = "parameter_13" + original_name = "conv8_1_dwise_bn_scale" shape = [1920] dtype = "float32" min_val = float("2.04938e-11") @@ -150,6 +164,7 @@ class Program_weight_tensor_parameter_13: class Program_weight_tensor_parameter_14: name = "parameter_14" + original_name = "conv8_1_dwise_bn_variance" shape = [1920] dtype = "float32" min_val = float("5.60519e-45") @@ -161,6 +176,7 @@ class Program_weight_tensor_parameter_14: class Program_weight_tensor_parameter_15: name = "parameter_15" + original_name = "conv8_1_dwise_bn_mean" shape = [1920] dtype = "float32" min_val = float("-0.355269") @@ -172,6 +188,7 @@ class Program_weight_tensor_parameter_15: class Program_weight_tensor_parameter_16: name = "parameter_16" + original_name = "conv8_1_dwise_weights" shape = [1920, 1, 3, 3] dtype = "float32" min_val = float("-0.212842") @@ -183,6 +200,7 @@ class Program_weight_tensor_parameter_16: class Program_weight_tensor_parameter_17: name = "parameter_17" + original_name = "conv8_1_expand_bn_offset" shape = [1920] dtype = "float32" min_val = float("-0.22522") @@ -194,6 +212,7 @@ class Program_weight_tensor_parameter_17: class Program_weight_tensor_parameter_18: name = "parameter_18" + original_name = "conv8_1_expand_bn_scale" shape = [1920] dtype = "float32" min_val = float("-3.70037e-06") @@ -205,6 +224,7 @@ class Program_weight_tensor_parameter_18: class Program_weight_tensor_parameter_19: name = "parameter_19" + original_name = "conv8_1_expand_bn_variance" shape = [1920] dtype = "float32" min_val = float("8.99498e-22") @@ -216,6 +236,7 @@ class Program_weight_tensor_parameter_19: class Program_weight_tensor_parameter_20: name = "parameter_20" + original_name = "conv8_1_expand_bn_mean" shape = [1920] dtype = "float32" min_val = float("-1.43524e-06") @@ -227,6 +248,7 @@ class Program_weight_tensor_parameter_20: class Program_weight_tensor_parameter_21: name = "parameter_21" + original_name = "conv8_1_expand_weights" shape = [1920, 320, 1, 1] dtype = "float32" min_val = float("-0.68243") @@ -238,6 +260,7 @@ class Program_weight_tensor_parameter_21: class Program_weight_tensor_parameter_22: name = "parameter_22" + original_name = "conv7_3_linear_bn_offset" shape = [320] dtype = "float32" min_val = float("-2.44299e-07") @@ -249,6 +272,7 @@ class Program_weight_tensor_parameter_22: class Program_weight_tensor_parameter_23: name = "parameter_23" + original_name = "conv7_3_linear_bn_scale" shape = [320] dtype = "float32" min_val = float("0.0629004") @@ -260,6 +284,7 @@ class Program_weight_tensor_parameter_23: class Program_weight_tensor_parameter_24: name = "parameter_24" + original_name = "conv7_3_linear_bn_variance" shape = [320] dtype = "float32" min_val = float("0.00431427") @@ -271,6 +296,7 @@ class Program_weight_tensor_parameter_24: class Program_weight_tensor_parameter_25: name = "parameter_25" + original_name = "conv7_3_linear_bn_mean" shape = [320] dtype = "float32" min_val = float("-0.397212") @@ -282,6 +308,7 @@ class Program_weight_tensor_parameter_25: class Program_weight_tensor_parameter_26: name = "parameter_26" + original_name = "conv7_3_linear_weights" shape = [320, 1920, 1, 1] dtype = "float32" min_val = float("-0.486695") @@ -293,6 +320,7 @@ class Program_weight_tensor_parameter_26: class Program_weight_tensor_parameter_27: name = "parameter_27" + original_name = "conv7_3_dwise_bn_offset" shape = [1920] dtype = "float32" min_val = float("-0.623428") @@ -304,6 +332,7 @@ class Program_weight_tensor_parameter_27: class Program_weight_tensor_parameter_28: name = "parameter_28" + original_name = "conv7_3_dwise_bn_scale" shape = [1920] dtype = "float32" min_val = float("1.28867e-11") @@ -315,6 +344,7 @@ class Program_weight_tensor_parameter_28: class Program_weight_tensor_parameter_29: name = "parameter_29" + original_name = "conv7_3_dwise_bn_variance" shape = [1920] dtype = "float32" min_val = float("5.60519e-45") @@ -326,6 +356,7 @@ class Program_weight_tensor_parameter_29: class Program_weight_tensor_parameter_30: name = "parameter_30" + original_name = "conv7_3_dwise_bn_mean" shape = [1920] dtype = "float32" min_val = float("-0.315971") @@ -337,6 +368,7 @@ class Program_weight_tensor_parameter_30: class Program_weight_tensor_parameter_31: name = "parameter_31" + original_name = "conv7_3_dwise_weights" shape = [1920, 1, 3, 3] dtype = "float32" min_val = float("-0.452421") @@ -348,6 +380,7 @@ class Program_weight_tensor_parameter_31: class Program_weight_tensor_parameter_32: name = "parameter_32" + original_name = "conv7_3_expand_bn_offset" shape = [1920] dtype = "float32" min_val = float("-0.187336") @@ -359,6 +392,7 @@ class Program_weight_tensor_parameter_32: class Program_weight_tensor_parameter_33: name = "parameter_33" + original_name = "conv7_3_expand_bn_scale" shape = [1920] dtype = "float32" min_val = float("-0.109549") @@ -370,6 +404,7 @@ class Program_weight_tensor_parameter_33: class Program_weight_tensor_parameter_34: name = "parameter_34" + original_name = "conv7_3_expand_bn_variance" shape = [1920] dtype = "float32" min_val = float("5.50956e-22") @@ -381,6 +416,7 @@ class Program_weight_tensor_parameter_34: class Program_weight_tensor_parameter_35: name = "parameter_35" + original_name = "conv7_3_expand_bn_mean" shape = [1920] dtype = "float32" min_val = float("-8.26422e-07") @@ -392,6 +428,7 @@ class Program_weight_tensor_parameter_35: class Program_weight_tensor_parameter_36: name = "parameter_36" + original_name = "conv7_3_expand_weights" shape = [1920, 320, 1, 1] dtype = "float32" min_val = float("-0.310963") @@ -403,6 +440,7 @@ class Program_weight_tensor_parameter_36: class Program_weight_tensor_parameter_37: name = "parameter_37" + original_name = "conv7_2_linear_bn_offset" shape = [320] dtype = "float32" min_val = float("-3.42187e-07") @@ -414,6 +452,7 @@ class Program_weight_tensor_parameter_37: class Program_weight_tensor_parameter_38: name = "parameter_38" + original_name = "conv7_2_linear_bn_scale" shape = [320] dtype = "float32" min_val = float("0.0754754") @@ -425,6 +464,7 @@ class Program_weight_tensor_parameter_38: class Program_weight_tensor_parameter_39: name = "parameter_39" + original_name = "conv7_2_linear_bn_variance" shape = [320] dtype = "float32" min_val = float("0.00621525") @@ -436,6 +476,7 @@ class Program_weight_tensor_parameter_39: class Program_weight_tensor_parameter_40: name = "parameter_40" + original_name = "conv7_2_linear_bn_mean" shape = [320] dtype = "float32" min_val = float("-0.450697") @@ -447,6 +488,7 @@ class Program_weight_tensor_parameter_40: class Program_weight_tensor_parameter_41: name = "parameter_41" + original_name = "conv7_2_linear_weights" shape = [320, 1920, 1, 1] dtype = "float32" min_val = float("-0.282633") @@ -458,6 +500,7 @@ class Program_weight_tensor_parameter_41: class Program_weight_tensor_parameter_42: name = "parameter_42" + original_name = "conv7_2_dwise_bn_offset" shape = [1920] dtype = "float32" min_val = float("-0.470106") @@ -469,6 +512,7 @@ class Program_weight_tensor_parameter_42: class Program_weight_tensor_parameter_43: name = "parameter_43" + original_name = "conv7_2_dwise_bn_scale" shape = [1920] dtype = "float32" min_val = float("1.36903e-11") @@ -480,6 +524,7 @@ class Program_weight_tensor_parameter_43: class Program_weight_tensor_parameter_44: name = "parameter_44" + original_name = "conv7_2_dwise_bn_variance" shape = [1920] dtype = "float32" min_val = float("5.60519e-45") @@ -491,6 +536,7 @@ class Program_weight_tensor_parameter_44: class Program_weight_tensor_parameter_45: name = "parameter_45" + original_name = "conv7_2_dwise_bn_mean" shape = [1920] dtype = "float32" min_val = float("-0.243979") @@ -502,6 +548,7 @@ class Program_weight_tensor_parameter_45: class Program_weight_tensor_parameter_46: name = "parameter_46" + original_name = "conv7_2_dwise_weights" shape = [1920, 1, 3, 3] dtype = "float32" min_val = float("-0.571675") @@ -513,6 +560,7 @@ class Program_weight_tensor_parameter_46: class Program_weight_tensor_parameter_47: name = "parameter_47" + original_name = "conv7_2_expand_bn_offset" shape = [1920] dtype = "float32" min_val = float("-0.158206") @@ -524,6 +572,7 @@ class Program_weight_tensor_parameter_47: class Program_weight_tensor_parameter_48: name = "parameter_48" + original_name = "conv7_2_expand_bn_scale" shape = [1920] dtype = "float32" min_val = float("-0.112173") @@ -535,6 +584,7 @@ class Program_weight_tensor_parameter_48: class Program_weight_tensor_parameter_49: name = "parameter_49" + original_name = "conv7_2_expand_bn_variance" shape = [1920] dtype = "float32" min_val = float("3.69036e-22") @@ -546,6 +596,7 @@ class Program_weight_tensor_parameter_49: class Program_weight_tensor_parameter_50: name = "parameter_50" + original_name = "conv7_2_expand_bn_mean" shape = [1920] dtype = "float32" min_val = float("-4.72125e-07") @@ -557,6 +608,7 @@ class Program_weight_tensor_parameter_50: class Program_weight_tensor_parameter_51: name = "parameter_51" + original_name = "conv7_2_expand_weights" shape = [1920, 320, 1, 1] dtype = "float32" min_val = float("-0.295077") @@ -568,6 +620,7 @@ class Program_weight_tensor_parameter_51: class Program_weight_tensor_parameter_52: name = "parameter_52" + original_name = "conv7_1_linear_bn_offset" shape = [320] dtype = "float32" min_val = float("-3.76072e-07") @@ -579,6 +632,7 @@ class Program_weight_tensor_parameter_52: class Program_weight_tensor_parameter_53: name = "parameter_53" + original_name = "conv7_1_linear_bn_scale" shape = [320] dtype = "float32" min_val = float("0.136361") @@ -590,6 +644,7 @@ class Program_weight_tensor_parameter_53: class Program_weight_tensor_parameter_54: name = "parameter_54" + original_name = "conv7_1_linear_bn_variance" shape = [320] dtype = "float32" min_val = float("0.00888174") @@ -601,6 +656,7 @@ class Program_weight_tensor_parameter_54: class Program_weight_tensor_parameter_55: name = "parameter_55" + original_name = "conv7_1_linear_bn_mean" shape = [320] dtype = "float32" min_val = float("-1.52688") @@ -612,6 +668,7 @@ class Program_weight_tensor_parameter_55: class Program_weight_tensor_parameter_56: name = "parameter_56" + original_name = "conv7_1_linear_weights" shape = [320, 1152, 1, 1] dtype = "float32" min_val = float("-0.531469") @@ -623,6 +680,7 @@ class Program_weight_tensor_parameter_56: class Program_weight_tensor_parameter_57: name = "parameter_57" + original_name = "conv7_1_dwise_bn_offset" shape = [1152] dtype = "float32" min_val = float("-0.190585") @@ -634,6 +692,7 @@ class Program_weight_tensor_parameter_57: class Program_weight_tensor_parameter_58: name = "parameter_58" + original_name = "conv7_1_dwise_bn_scale" shape = [1152] dtype = "float32" min_val = float("1.89888e-11") @@ -645,6 +704,7 @@ class Program_weight_tensor_parameter_58: class Program_weight_tensor_parameter_59: name = "parameter_59" + original_name = "conv7_1_dwise_bn_variance" shape = [1152] dtype = "float32" min_val = float("5.60519e-45") @@ -656,6 +716,7 @@ class Program_weight_tensor_parameter_59: class Program_weight_tensor_parameter_60: name = "parameter_60" + original_name = "conv7_1_dwise_bn_mean" shape = [1152] dtype = "float32" min_val = float("-0.198018") @@ -667,6 +728,7 @@ class Program_weight_tensor_parameter_60: class Program_weight_tensor_parameter_61: name = "parameter_61" + original_name = "conv7_1_dwise_weights" shape = [1152, 1, 3, 3] dtype = "float32" min_val = float("-0.256558") @@ -678,6 +740,7 @@ class Program_weight_tensor_parameter_61: class Program_weight_tensor_parameter_62: name = "parameter_62" + original_name = "conv7_1_expand_bn_offset" shape = [1152] dtype = "float32" min_val = float("-0.433717") @@ -689,6 +752,7 @@ class Program_weight_tensor_parameter_62: class Program_weight_tensor_parameter_63: name = "parameter_63" + original_name = "conv7_1_expand_bn_scale" shape = [1152] dtype = "float32" min_val = float("-0.0686943") @@ -700,6 +764,7 @@ class Program_weight_tensor_parameter_63: class Program_weight_tensor_parameter_64: name = "parameter_64" + original_name = "conv7_1_expand_bn_variance" shape = [1152] dtype = "float32" min_val = float("9.05528e-22") @@ -711,6 +776,7 @@ class Program_weight_tensor_parameter_64: class Program_weight_tensor_parameter_65: name = "parameter_65" + original_name = "conv7_1_expand_bn_mean" shape = [1152] dtype = "float32" min_val = float("-9.56226e-07") @@ -722,6 +788,7 @@ class Program_weight_tensor_parameter_65: class Program_weight_tensor_parameter_66: name = "parameter_66" + original_name = "conv7_1_expand_weights" shape = [1152, 192, 1, 1] dtype = "float32" min_val = float("-0.467675") @@ -733,6 +800,7 @@ class Program_weight_tensor_parameter_66: class Program_weight_tensor_parameter_67: name = "parameter_67" + original_name = "conv6_3_linear_bn_offset" shape = [192] dtype = "float32" min_val = float("-2.1977e-07") @@ -744,6 +812,7 @@ class Program_weight_tensor_parameter_67: class Program_weight_tensor_parameter_68: name = "parameter_68" + original_name = "conv6_3_linear_bn_scale" shape = [192] dtype = "float32" min_val = float("-0.0503631") @@ -755,6 +824,7 @@ class Program_weight_tensor_parameter_68: class Program_weight_tensor_parameter_69: name = "parameter_69" + original_name = "conv6_3_linear_bn_variance" shape = [192] dtype = "float32" min_val = float("0.00114648") @@ -766,6 +836,7 @@ class Program_weight_tensor_parameter_69: class Program_weight_tensor_parameter_70: name = "parameter_70" + original_name = "conv6_3_linear_bn_mean" shape = [192] dtype = "float32" min_val = float("-0.340468") @@ -777,6 +848,7 @@ class Program_weight_tensor_parameter_70: class Program_weight_tensor_parameter_71: name = "parameter_71" + original_name = "conv6_3_linear_weights" shape = [192, 1152, 1, 1] dtype = "float32" min_val = float("-0.564185") @@ -788,6 +860,7 @@ class Program_weight_tensor_parameter_71: class Program_weight_tensor_parameter_72: name = "parameter_72" + original_name = "conv6_3_dwise_bn_offset" shape = [1152] dtype = "float32" min_val = float("-0.412132") @@ -799,6 +872,7 @@ class Program_weight_tensor_parameter_72: class Program_weight_tensor_parameter_73: name = "parameter_73" + original_name = "conv6_3_dwise_bn_scale" shape = [1152] dtype = "float32" min_val = float("1.82927e-11") @@ -810,6 +884,7 @@ class Program_weight_tensor_parameter_73: class Program_weight_tensor_parameter_74: name = "parameter_74" + original_name = "conv6_3_dwise_bn_variance" shape = [1152] dtype = "float32" min_val = float("5.60519e-45") @@ -821,6 +896,7 @@ class Program_weight_tensor_parameter_74: class Program_weight_tensor_parameter_75: name = "parameter_75" + original_name = "conv6_3_dwise_bn_mean" shape = [1152] dtype = "float32" min_val = float("-0.162084") @@ -832,6 +908,7 @@ class Program_weight_tensor_parameter_75: class Program_weight_tensor_parameter_76: name = "parameter_76" + original_name = "conv6_3_dwise_weights" shape = [1152, 1, 3, 3] dtype = "float32" min_val = float("-0.576102") @@ -843,6 +920,7 @@ class Program_weight_tensor_parameter_76: class Program_weight_tensor_parameter_77: name = "parameter_77" + original_name = "conv6_3_expand_bn_offset" shape = [1152] dtype = "float32" min_val = float("-0.665768") @@ -854,6 +932,7 @@ class Program_weight_tensor_parameter_77: class Program_weight_tensor_parameter_78: name = "parameter_78" + original_name = "conv6_3_expand_bn_scale" shape = [1152] dtype = "float32" min_val = float("-0.0947745") @@ -865,6 +944,7 @@ class Program_weight_tensor_parameter_78: class Program_weight_tensor_parameter_79: name = "parameter_79" + original_name = "conv6_3_expand_bn_variance" shape = [1152] dtype = "float32" min_val = float("5.50622e-22") @@ -876,6 +956,7 @@ class Program_weight_tensor_parameter_79: class Program_weight_tensor_parameter_80: name = "parameter_80" + original_name = "conv6_3_expand_bn_mean" shape = [1152] dtype = "float32" min_val = float("-4.94479e-07") @@ -887,6 +968,7 @@ class Program_weight_tensor_parameter_80: class Program_weight_tensor_parameter_81: name = "parameter_81" + original_name = "conv6_3_expand_weights" shape = [1152, 192, 1, 1] dtype = "float32" min_val = float("-0.357049") @@ -898,6 +980,7 @@ class Program_weight_tensor_parameter_81: class Program_weight_tensor_parameter_82: name = "parameter_82" + original_name = "conv6_2_linear_bn_offset" shape = [192] dtype = "float32" min_val = float("-2.11979e-07") @@ -909,6 +992,7 @@ class Program_weight_tensor_parameter_82: class Program_weight_tensor_parameter_83: name = "parameter_83" + original_name = "conv6_2_linear_bn_scale" shape = [192] dtype = "float32" min_val = float("-0.037586") @@ -920,6 +1004,7 @@ class Program_weight_tensor_parameter_83: class Program_weight_tensor_parameter_84: name = "parameter_84" + original_name = "conv6_2_linear_bn_variance" shape = [192] dtype = "float32" min_val = float("0.00133759") @@ -931,6 +1016,7 @@ class Program_weight_tensor_parameter_84: class Program_weight_tensor_parameter_85: name = "parameter_85" + original_name = "conv6_2_linear_bn_mean" shape = [192] dtype = "float32" min_val = float("-0.351809") @@ -942,6 +1028,7 @@ class Program_weight_tensor_parameter_85: class Program_weight_tensor_parameter_86: name = "parameter_86" + original_name = "conv6_2_linear_weights" shape = [192, 1152, 1, 1] dtype = "float32" min_val = float("-0.401722") @@ -953,6 +1040,7 @@ class Program_weight_tensor_parameter_86: class Program_weight_tensor_parameter_87: name = "parameter_87" + original_name = "conv6_2_dwise_bn_offset" shape = [1152] dtype = "float32" min_val = float("-0.75416") @@ -964,6 +1052,7 @@ class Program_weight_tensor_parameter_87: class Program_weight_tensor_parameter_88: name = "parameter_88" + original_name = "conv6_2_dwise_bn_scale" shape = [1152] dtype = "float32" min_val = float("1.71918e-11") @@ -975,6 +1064,7 @@ class Program_weight_tensor_parameter_88: class Program_weight_tensor_parameter_89: name = "parameter_89" + original_name = "conv6_2_dwise_bn_variance" shape = [1152] dtype = "float32" min_val = float("5.60519e-45") @@ -986,6 +1076,7 @@ class Program_weight_tensor_parameter_89: class Program_weight_tensor_parameter_90: name = "parameter_90" + original_name = "conv6_2_dwise_bn_mean" shape = [1152] dtype = "float32" min_val = float("-0.273853") @@ -997,6 +1088,7 @@ class Program_weight_tensor_parameter_90: class Program_weight_tensor_parameter_91: name = "parameter_91" + original_name = "conv6_2_dwise_weights" shape = [1152, 1, 3, 3] dtype = "float32" min_val = float("-0.581963") @@ -1008,6 +1100,7 @@ class Program_weight_tensor_parameter_91: class Program_weight_tensor_parameter_92: name = "parameter_92" + original_name = "conv6_2_expand_bn_offset" shape = [1152] dtype = "float32" min_val = float("-0.378841") @@ -1019,6 +1112,7 @@ class Program_weight_tensor_parameter_92: class Program_weight_tensor_parameter_93: name = "parameter_93" + original_name = "conv6_2_expand_bn_scale" shape = [1152] dtype = "float32" min_val = float("-0.112428") @@ -1030,6 +1124,7 @@ class Program_weight_tensor_parameter_93: class Program_weight_tensor_parameter_94: name = "parameter_94" + original_name = "conv6_2_expand_bn_variance" shape = [1152] dtype = "float32" min_val = float("6.25592e-22") @@ -1041,6 +1136,7 @@ class Program_weight_tensor_parameter_94: class Program_weight_tensor_parameter_95: name = "parameter_95" + original_name = "conv6_2_expand_bn_mean" shape = [1152] dtype = "float32" min_val = float("-2.47039e-07") @@ -1052,6 +1148,7 @@ class Program_weight_tensor_parameter_95: class Program_weight_tensor_parameter_96: name = "parameter_96" + original_name = "conv6_2_expand_weights" shape = [1152, 192, 1, 1] dtype = "float32" min_val = float("-0.311649") @@ -1063,6 +1160,7 @@ class Program_weight_tensor_parameter_96: class Program_weight_tensor_parameter_97: name = "parameter_97" + original_name = "conv6_1_linear_bn_offset" shape = [192] dtype = "float32" min_val = float("-2.85897e-07") @@ -1074,6 +1172,7 @@ class Program_weight_tensor_parameter_97: class Program_weight_tensor_parameter_98: name = "parameter_98" + original_name = "conv6_1_linear_bn_scale" shape = [192] dtype = "float32" min_val = float("0.211009") @@ -1085,6 +1184,7 @@ class Program_weight_tensor_parameter_98: class Program_weight_tensor_parameter_99: name = "parameter_99" + original_name = "conv6_1_linear_bn_variance" shape = [192] dtype = "float32" min_val = float("0.034918") @@ -1096,6 +1196,7 @@ class Program_weight_tensor_parameter_99: class Program_weight_tensor_parameter_100: name = "parameter_100" + original_name = "conv6_1_linear_bn_mean" shape = [192] dtype = "float32" min_val = float("-0.879679") @@ -1107,6 +1208,7 @@ class Program_weight_tensor_parameter_100: class Program_weight_tensor_parameter_101: name = "parameter_101" + original_name = "conv6_1_linear_weights" shape = [192, 768, 1, 1] dtype = "float32" min_val = float("-0.501629") @@ -1118,6 +1220,7 @@ class Program_weight_tensor_parameter_101: class Program_weight_tensor_parameter_102: name = "parameter_102" + original_name = "conv6_1_dwise_bn_offset" shape = [768] dtype = "float32" min_val = float("-0.505404") @@ -1129,6 +1232,7 @@ class Program_weight_tensor_parameter_102: class Program_weight_tensor_parameter_103: name = "parameter_103" + original_name = "conv6_1_dwise_bn_scale" shape = [768] dtype = "float32" min_val = float("3.01887e-11") @@ -1140,6 +1244,7 @@ class Program_weight_tensor_parameter_103: class Program_weight_tensor_parameter_104: name = "parameter_104" + original_name = "conv6_1_dwise_bn_variance" shape = [768] dtype = "float32" min_val = float("5.60519e-45") @@ -1151,6 +1256,7 @@ class Program_weight_tensor_parameter_104: class Program_weight_tensor_parameter_105: name = "parameter_105" + original_name = "conv6_1_dwise_bn_mean" shape = [768] dtype = "float32" min_val = float("-0.266286") @@ -1162,6 +1268,7 @@ class Program_weight_tensor_parameter_105: class Program_weight_tensor_parameter_106: name = "parameter_106" + original_name = "conv6_1_dwise_weights" shape = [768, 1, 3, 3] dtype = "float32" min_val = float("-0.567354") @@ -1173,6 +1280,7 @@ class Program_weight_tensor_parameter_106: class Program_weight_tensor_parameter_107: name = "parameter_107" + original_name = "conv6_1_expand_bn_offset" shape = [768] dtype = "float32" min_val = float("-0.449363") @@ -1184,6 +1292,7 @@ class Program_weight_tensor_parameter_107: class Program_weight_tensor_parameter_108: name = "parameter_108" + original_name = "conv6_1_expand_bn_scale" shape = [768] dtype = "float32" min_val = float("-2.96645e-06") @@ -1195,6 +1304,7 @@ class Program_weight_tensor_parameter_108: class Program_weight_tensor_parameter_109: name = "parameter_109" + original_name = "conv6_1_expand_bn_variance" shape = [768] dtype = "float32" min_val = float("2.38644e-21") @@ -1206,6 +1316,7 @@ class Program_weight_tensor_parameter_109: class Program_weight_tensor_parameter_110: name = "parameter_110" + original_name = "conv6_1_expand_bn_mean" shape = [768] dtype = "float32" min_val = float("-1.14609e-06") @@ -1217,6 +1328,7 @@ class Program_weight_tensor_parameter_110: class Program_weight_tensor_parameter_111: name = "parameter_111" + original_name = "conv6_1_expand_weights" shape = [768, 128, 1, 1] dtype = "float32" min_val = float("-0.686659") @@ -1228,6 +1340,7 @@ class Program_weight_tensor_parameter_111: class Program_weight_tensor_parameter_112: name = "parameter_112" + original_name = "conv5_4_linear_bn_offset" shape = [128] dtype = "float32" min_val = float("-3.0331e-07") @@ -1239,6 +1352,7 @@ class Program_weight_tensor_parameter_112: class Program_weight_tensor_parameter_113: name = "parameter_113" + original_name = "conv5_4_linear_bn_scale" shape = [128] dtype = "float32" min_val = float("0.046647") @@ -1250,6 +1364,7 @@ class Program_weight_tensor_parameter_113: class Program_weight_tensor_parameter_114: name = "parameter_114" + original_name = "conv5_4_linear_bn_variance" shape = [128] dtype = "float32" min_val = float("0.00301309") @@ -1261,6 +1376,7 @@ class Program_weight_tensor_parameter_114: class Program_weight_tensor_parameter_115: name = "parameter_115" + original_name = "conv5_4_linear_bn_mean" shape = [128] dtype = "float32" min_val = float("-0.481459") @@ -1272,6 +1388,7 @@ class Program_weight_tensor_parameter_115: class Program_weight_tensor_parameter_116: name = "parameter_116" + original_name = "conv5_4_linear_weights" shape = [128, 768, 1, 1] dtype = "float32" min_val = float("-0.343883") @@ -1283,6 +1400,7 @@ class Program_weight_tensor_parameter_116: class Program_weight_tensor_parameter_117: name = "parameter_117" + original_name = "conv5_4_dwise_bn_offset" shape = [768] dtype = "float32" min_val = float("-0.425673") @@ -1294,6 +1412,7 @@ class Program_weight_tensor_parameter_117: class Program_weight_tensor_parameter_118: name = "parameter_118" + original_name = "conv5_4_dwise_bn_scale" shape = [768] dtype = "float32" min_val = float("1.67014e-11") @@ -1305,6 +1424,7 @@ class Program_weight_tensor_parameter_118: class Program_weight_tensor_parameter_119: name = "parameter_119" + original_name = "conv5_4_dwise_bn_variance" shape = [768] dtype = "float32" min_val = float("5.60519e-45") @@ -1316,6 +1436,7 @@ class Program_weight_tensor_parameter_119: class Program_weight_tensor_parameter_120: name = "parameter_120" + original_name = "conv5_4_dwise_bn_mean" shape = [768] dtype = "float32" min_val = float("-0.295067") @@ -1327,6 +1448,7 @@ class Program_weight_tensor_parameter_120: class Program_weight_tensor_parameter_121: name = "parameter_121" + original_name = "conv5_4_dwise_weights" shape = [768, 1, 3, 3] dtype = "float32" min_val = float("-0.46512") @@ -1338,6 +1460,7 @@ class Program_weight_tensor_parameter_121: class Program_weight_tensor_parameter_122: name = "parameter_122" + original_name = "conv5_4_expand_bn_offset" shape = [768] dtype = "float32" min_val = float("-0.182083") @@ -1349,6 +1472,7 @@ class Program_weight_tensor_parameter_122: class Program_weight_tensor_parameter_123: name = "parameter_123" + original_name = "conv5_4_expand_bn_scale" shape = [768] dtype = "float32" min_val = float("-0.104548") @@ -1360,6 +1484,7 @@ class Program_weight_tensor_parameter_123: class Program_weight_tensor_parameter_124: name = "parameter_124" + original_name = "conv5_4_expand_bn_variance" shape = [768] dtype = "float32" min_val = float("1.07626e-21") @@ -1371,6 +1496,7 @@ class Program_weight_tensor_parameter_124: class Program_weight_tensor_parameter_125: name = "parameter_125" + original_name = "conv5_4_expand_bn_mean" shape = [768] dtype = "float32" min_val = float("-6.13003e-07") @@ -1382,6 +1508,7 @@ class Program_weight_tensor_parameter_125: class Program_weight_tensor_parameter_126: name = "parameter_126" + original_name = "conv5_4_expand_weights" shape = [768, 128, 1, 1] dtype = "float32" min_val = float("-0.320932") @@ -1393,6 +1520,7 @@ class Program_weight_tensor_parameter_126: class Program_weight_tensor_parameter_127: name = "parameter_127" + original_name = "conv5_3_linear_bn_offset" shape = [128] dtype = "float32" min_val = float("-2.86251e-07") @@ -1404,6 +1532,7 @@ class Program_weight_tensor_parameter_127: class Program_weight_tensor_parameter_128: name = "parameter_128" + original_name = "conv5_3_linear_bn_scale" shape = [128] dtype = "float32" min_val = float("0.0482937") @@ -1415,6 +1544,7 @@ class Program_weight_tensor_parameter_128: class Program_weight_tensor_parameter_129: name = "parameter_129" + original_name = "conv5_3_linear_bn_variance" shape = [128] dtype = "float32" min_val = float("0.00416822") @@ -1426,6 +1556,7 @@ class Program_weight_tensor_parameter_129: class Program_weight_tensor_parameter_130: name = "parameter_130" + original_name = "conv5_3_linear_bn_mean" shape = [128] dtype = "float32" min_val = float("-0.40317") @@ -1437,6 +1568,7 @@ class Program_weight_tensor_parameter_130: class Program_weight_tensor_parameter_131: name = "parameter_131" + original_name = "conv5_3_linear_weights" shape = [128, 768, 1, 1] dtype = "float32" min_val = float("-0.574202") @@ -1448,6 +1580,7 @@ class Program_weight_tensor_parameter_131: class Program_weight_tensor_parameter_132: name = "parameter_132" + original_name = "conv5_3_dwise_bn_offset" shape = [768] dtype = "float32" min_val = float("-0.508623") @@ -1459,6 +1592,7 @@ class Program_weight_tensor_parameter_132: class Program_weight_tensor_parameter_133: name = "parameter_133" + original_name = "conv5_3_dwise_bn_scale" shape = [768] dtype = "float32" min_val = float("1.39312e-11") @@ -1470,6 +1604,7 @@ class Program_weight_tensor_parameter_133: class Program_weight_tensor_parameter_134: name = "parameter_134" + original_name = "conv5_3_dwise_bn_variance" shape = [768] dtype = "float32" min_val = float("5.60519e-45") @@ -1481,6 +1616,7 @@ class Program_weight_tensor_parameter_134: class Program_weight_tensor_parameter_135: name = "parameter_135" + original_name = "conv5_3_dwise_bn_mean" shape = [768] dtype = "float32" min_val = float("-0.195873") @@ -1492,6 +1628,7 @@ class Program_weight_tensor_parameter_135: class Program_weight_tensor_parameter_136: name = "parameter_136" + original_name = "conv5_3_dwise_weights" shape = [768, 1, 3, 3] dtype = "float32" min_val = float("-0.569465") @@ -1503,6 +1640,7 @@ class Program_weight_tensor_parameter_136: class Program_weight_tensor_parameter_137: name = "parameter_137" + original_name = "conv5_3_expand_bn_offset" shape = [768] dtype = "float32" min_val = float("-0.147825") @@ -1514,6 +1652,7 @@ class Program_weight_tensor_parameter_137: class Program_weight_tensor_parameter_138: name = "parameter_138" + original_name = "conv5_3_expand_bn_scale" shape = [768] dtype = "float32" min_val = float("-0.0969862") @@ -1525,6 +1664,7 @@ class Program_weight_tensor_parameter_138: class Program_weight_tensor_parameter_139: name = "parameter_139" + original_name = "conv5_3_expand_bn_variance" shape = [768] dtype = "float32" min_val = float("1.34535e-21") @@ -1536,6 +1676,7 @@ class Program_weight_tensor_parameter_139: class Program_weight_tensor_parameter_140: name = "parameter_140" + original_name = "conv5_3_expand_bn_mean" shape = [768] dtype = "float32" min_val = float("-4.26923e-07") @@ -1547,6 +1688,7 @@ class Program_weight_tensor_parameter_140: class Program_weight_tensor_parameter_141: name = "parameter_141" + original_name = "conv5_3_expand_weights" shape = [768, 128, 1, 1] dtype = "float32" min_val = float("-0.601237") @@ -1558,6 +1700,7 @@ class Program_weight_tensor_parameter_141: class Program_weight_tensor_parameter_142: name = "parameter_142" + original_name = "conv5_2_linear_bn_offset" shape = [128] dtype = "float32" min_val = float("-2.40423e-07") @@ -1569,6 +1712,7 @@ class Program_weight_tensor_parameter_142: class Program_weight_tensor_parameter_143: name = "parameter_143" + original_name = "conv5_2_linear_bn_scale" shape = [128] dtype = "float32" min_val = float("0.0902598") @@ -1580,6 +1724,7 @@ class Program_weight_tensor_parameter_143: class Program_weight_tensor_parameter_144: name = "parameter_144" + original_name = "conv5_2_linear_bn_variance" shape = [128] dtype = "float32" min_val = float("0.0074343") @@ -1591,6 +1736,7 @@ class Program_weight_tensor_parameter_144: class Program_weight_tensor_parameter_145: name = "parameter_145" + original_name = "conv5_2_linear_bn_mean" shape = [128] dtype = "float32" min_val = float("-0.364802") @@ -1602,6 +1748,7 @@ class Program_weight_tensor_parameter_145: class Program_weight_tensor_parameter_146: name = "parameter_146" + original_name = "conv5_2_linear_weights" shape = [128, 768, 1, 1] dtype = "float32" min_val = float("-0.328559") @@ -1613,6 +1760,7 @@ class Program_weight_tensor_parameter_146: class Program_weight_tensor_parameter_147: name = "parameter_147" + original_name = "conv5_2_dwise_bn_offset" shape = [768] dtype = "float32" min_val = float("-0.440413") @@ -1624,6 +1772,7 @@ class Program_weight_tensor_parameter_147: class Program_weight_tensor_parameter_148: name = "parameter_148" + original_name = "conv5_2_dwise_bn_scale" shape = [768] dtype = "float32" min_val = float("2.08697e-11") @@ -1635,6 +1784,7 @@ class Program_weight_tensor_parameter_148: class Program_weight_tensor_parameter_149: name = "parameter_149" + original_name = "conv5_2_dwise_bn_variance" shape = [768] dtype = "float32" min_val = float("5.60519e-45") @@ -1646,6 +1796,7 @@ class Program_weight_tensor_parameter_149: class Program_weight_tensor_parameter_150: name = "parameter_150" + original_name = "conv5_2_dwise_bn_mean" shape = [768] dtype = "float32" min_val = float("-0.152065") @@ -1657,6 +1808,7 @@ class Program_weight_tensor_parameter_150: class Program_weight_tensor_parameter_151: name = "parameter_151" + original_name = "conv5_2_dwise_weights" shape = [768, 1, 3, 3] dtype = "float32" min_val = float("-0.512442") @@ -1668,6 +1820,7 @@ class Program_weight_tensor_parameter_151: class Program_weight_tensor_parameter_152: name = "parameter_152" + original_name = "conv5_2_expand_bn_offset" shape = [768] dtype = "float32" min_val = float("-0.175324") @@ -1679,6 +1832,7 @@ class Program_weight_tensor_parameter_152: class Program_weight_tensor_parameter_153: name = "parameter_153" + original_name = "conv5_2_expand_bn_scale" shape = [768] dtype = "float32" min_val = float("-0.130853") @@ -1690,6 +1844,7 @@ class Program_weight_tensor_parameter_153: class Program_weight_tensor_parameter_154: name = "parameter_154" + original_name = "conv5_2_expand_bn_variance" shape = [768] dtype = "float32" min_val = float("1.00438e-21") @@ -1701,6 +1856,7 @@ class Program_weight_tensor_parameter_154: class Program_weight_tensor_parameter_155: name = "parameter_155" + original_name = "conv5_2_expand_bn_mean" shape = [768] dtype = "float32" min_val = float("-1.9898e-07") @@ -1712,6 +1868,7 @@ class Program_weight_tensor_parameter_155: class Program_weight_tensor_parameter_156: name = "parameter_156" + original_name = "conv5_2_expand_weights" shape = [768, 128, 1, 1] dtype = "float32" min_val = float("-0.318679") @@ -1723,6 +1880,7 @@ class Program_weight_tensor_parameter_156: class Program_weight_tensor_parameter_157: name = "parameter_157" + original_name = "conv5_1_linear_bn_offset" shape = [128] dtype = "float32" min_val = float("-2.9443e-07") @@ -1734,6 +1892,7 @@ class Program_weight_tensor_parameter_157: class Program_weight_tensor_parameter_158: name = "parameter_158" + original_name = "conv5_1_linear_bn_scale" shape = [128] dtype = "float32" min_val = float("0.0828847") @@ -1745,6 +1904,7 @@ class Program_weight_tensor_parameter_158: class Program_weight_tensor_parameter_159: name = "parameter_159" + original_name = "conv5_1_linear_bn_variance" shape = [128] dtype = "float32" min_val = float("0.0143345") @@ -1756,6 +1916,7 @@ class Program_weight_tensor_parameter_159: class Program_weight_tensor_parameter_160: name = "parameter_160" + original_name = "conv5_1_linear_bn_mean" shape = [128] dtype = "float32" min_val = float("-1.33261") @@ -1767,6 +1928,7 @@ class Program_weight_tensor_parameter_160: class Program_weight_tensor_parameter_161: name = "parameter_161" + original_name = "conv5_1_linear_weights" shape = [128, 384, 1, 1] dtype = "float32" min_val = float("-0.522209") @@ -1778,6 +1940,7 @@ class Program_weight_tensor_parameter_161: class Program_weight_tensor_parameter_162: name = "parameter_162" + original_name = "conv5_1_dwise_bn_offset" shape = [384] dtype = "float32" min_val = float("-0.123964") @@ -1789,6 +1952,7 @@ class Program_weight_tensor_parameter_162: class Program_weight_tensor_parameter_163: name = "parameter_163" + original_name = "conv5_1_dwise_bn_scale" shape = [384] dtype = "float32" min_val = float("3.18048e-11") @@ -1800,6 +1964,7 @@ class Program_weight_tensor_parameter_163: class Program_weight_tensor_parameter_164: name = "parameter_164" + original_name = "conv5_1_dwise_bn_variance" shape = [384] dtype = "float32" min_val = float("5.60519e-45") @@ -1811,6 +1976,7 @@ class Program_weight_tensor_parameter_164: class Program_weight_tensor_parameter_165: name = "parameter_165" + original_name = "conv5_1_dwise_bn_mean" shape = [384] dtype = "float32" min_val = float("-0.27978") @@ -1822,6 +1988,7 @@ class Program_weight_tensor_parameter_165: class Program_weight_tensor_parameter_166: name = "parameter_166" + original_name = "conv5_1_dwise_weights" shape = [384, 1, 3, 3] dtype = "float32" min_val = float("-0.23857") @@ -1833,6 +2000,7 @@ class Program_weight_tensor_parameter_166: class Program_weight_tensor_parameter_167: name = "parameter_167" + original_name = "conv5_1_expand_bn_offset" shape = [384] dtype = "float32" min_val = float("-0.351776") @@ -1844,6 +2012,7 @@ class Program_weight_tensor_parameter_167: class Program_weight_tensor_parameter_168: name = "parameter_168" + original_name = "conv5_1_expand_bn_scale" shape = [384] dtype = "float32" min_val = float("-0.216771") @@ -1855,6 +2024,7 @@ class Program_weight_tensor_parameter_168: class Program_weight_tensor_parameter_169: name = "parameter_169" + original_name = "conv5_1_expand_bn_variance" shape = [384] dtype = "float32" min_val = float("4.51937e-21") @@ -1866,6 +2036,7 @@ class Program_weight_tensor_parameter_169: class Program_weight_tensor_parameter_170: name = "parameter_170" + original_name = "conv5_1_expand_bn_mean" shape = [384] dtype = "float32" min_val = float("-7.93089e-07") @@ -1877,6 +2048,7 @@ class Program_weight_tensor_parameter_170: class Program_weight_tensor_parameter_171: name = "parameter_171" + original_name = "conv5_1_expand_weights" shape = [384, 64, 1, 1] dtype = "float32" min_val = float("-0.685614") @@ -1888,6 +2060,7 @@ class Program_weight_tensor_parameter_171: class Program_weight_tensor_parameter_172: name = "parameter_172" + original_name = "conv4_3_linear_bn_offset" shape = [64] dtype = "float32" min_val = float("0") @@ -1897,6 +2070,7 @@ class Program_weight_tensor_parameter_172: class Program_weight_tensor_parameter_173: name = "parameter_173" + original_name = "conv4_3_linear_bn_scale" shape = [64] dtype = "float32" min_val = float("0") @@ -1906,6 +2080,7 @@ class Program_weight_tensor_parameter_173: class Program_weight_tensor_parameter_174: name = "parameter_174" + original_name = "conv4_3_linear_bn_variance" shape = [64] dtype = "float32" min_val = float("0") @@ -1915,6 +2090,7 @@ class Program_weight_tensor_parameter_174: class Program_weight_tensor_parameter_175: name = "parameter_175" + original_name = "conv4_3_linear_bn_mean" shape = [64] dtype = "float32" min_val = float("0") @@ -1924,6 +2100,7 @@ class Program_weight_tensor_parameter_175: class Program_weight_tensor_parameter_176: name = "parameter_176" + original_name = "conv4_3_linear_weights" shape = [64, 384, 1, 1] dtype = "float32" min_val = float("-0.36142") @@ -1935,6 +2112,7 @@ class Program_weight_tensor_parameter_176: class Program_weight_tensor_parameter_177: name = "parameter_177" + original_name = "conv4_3_dwise_bn_offset" shape = [384] dtype = "float32" min_val = float("-0.496257") @@ -1946,6 +2124,7 @@ class Program_weight_tensor_parameter_177: class Program_weight_tensor_parameter_178: name = "parameter_178" + original_name = "conv4_3_dwise_bn_scale" shape = [384] dtype = "float32" min_val = float("2.77833e-11") @@ -1957,6 +2136,7 @@ class Program_weight_tensor_parameter_178: class Program_weight_tensor_parameter_179: name = "parameter_179" + original_name = "conv4_3_dwise_bn_variance" shape = [384] dtype = "float32" min_val = float("5.60519e-45") @@ -1968,6 +2148,7 @@ class Program_weight_tensor_parameter_179: class Program_weight_tensor_parameter_180: name = "parameter_180" + original_name = "conv4_3_dwise_bn_mean" shape = [384] dtype = "float32" min_val = float("-0.189671") @@ -1979,6 +2160,7 @@ class Program_weight_tensor_parameter_180: class Program_weight_tensor_parameter_181: name = "parameter_181" + original_name = "conv4_3_dwise_weights" shape = [384, 1, 3, 3] dtype = "float32" min_val = float("-0.568739") @@ -1990,6 +2172,7 @@ class Program_weight_tensor_parameter_181: class Program_weight_tensor_parameter_182: name = "parameter_182" + original_name = "conv4_3_expand_bn_offset" shape = [384] dtype = "float32" min_val = float("-0.198826") @@ -2001,6 +2184,7 @@ class Program_weight_tensor_parameter_182: class Program_weight_tensor_parameter_183: name = "parameter_183" + original_name = "conv4_3_expand_bn_scale" shape = [384] dtype = "float32" min_val = float("-0.132979") @@ -2012,6 +2196,7 @@ class Program_weight_tensor_parameter_183: class Program_weight_tensor_parameter_184: name = "parameter_184" + original_name = "conv4_3_expand_bn_variance" shape = [384] dtype = "float32" min_val = float("2.85439e-21") @@ -2023,6 +2208,7 @@ class Program_weight_tensor_parameter_184: class Program_weight_tensor_parameter_185: name = "parameter_185" + original_name = "conv4_3_expand_bn_mean" shape = [384] dtype = "float32" min_val = float("-3.09328e-07") @@ -2034,6 +2220,7 @@ class Program_weight_tensor_parameter_185: class Program_weight_tensor_parameter_186: name = "parameter_186" + original_name = "conv4_3_expand_weights" shape = [384, 64, 1, 1] dtype = "float32" min_val = float("-0.44282") @@ -2045,6 +2232,7 @@ class Program_weight_tensor_parameter_186: class Program_weight_tensor_parameter_187: name = "parameter_187" + original_name = "conv4_2_linear_bn_offset" shape = [64] dtype = "float32" min_val = float("0") @@ -2054,6 +2242,7 @@ class Program_weight_tensor_parameter_187: class Program_weight_tensor_parameter_188: name = "parameter_188" + original_name = "conv4_2_linear_bn_scale" shape = [64] dtype = "float32" min_val = float("0") @@ -2063,6 +2252,7 @@ class Program_weight_tensor_parameter_188: class Program_weight_tensor_parameter_189: name = "parameter_189" + original_name = "conv4_2_linear_bn_variance" shape = [64] dtype = "float32" min_val = float("0") @@ -2072,6 +2262,7 @@ class Program_weight_tensor_parameter_189: class Program_weight_tensor_parameter_190: name = "parameter_190" + original_name = "conv4_2_linear_bn_mean" shape = [64] dtype = "float32" min_val = float("0") @@ -2081,6 +2272,7 @@ class Program_weight_tensor_parameter_190: class Program_weight_tensor_parameter_191: name = "parameter_191" + original_name = "conv4_2_linear_weights" shape = [64, 384, 1, 1] dtype = "float32" min_val = float("-0.368421") @@ -2092,6 +2284,7 @@ class Program_weight_tensor_parameter_191: class Program_weight_tensor_parameter_192: name = "parameter_192" + original_name = "conv4_2_dwise_bn_offset" shape = [384] dtype = "float32" min_val = float("-0.379269") @@ -2103,6 +2296,7 @@ class Program_weight_tensor_parameter_192: class Program_weight_tensor_parameter_193: name = "parameter_193" + original_name = "conv4_2_dwise_bn_scale" shape = [384] dtype = "float32" min_val = float("2.37346e-11") @@ -2114,6 +2308,7 @@ class Program_weight_tensor_parameter_193: class Program_weight_tensor_parameter_194: name = "parameter_194" + original_name = "conv4_2_dwise_bn_variance" shape = [384] dtype = "float32" min_val = float("5.60519e-45") @@ -2125,6 +2320,7 @@ class Program_weight_tensor_parameter_194: class Program_weight_tensor_parameter_195: name = "parameter_195" + original_name = "conv4_2_dwise_bn_mean" shape = [384] dtype = "float32" min_val = float("-0.17321") @@ -2136,6 +2332,7 @@ class Program_weight_tensor_parameter_195: class Program_weight_tensor_parameter_196: name = "parameter_196" + original_name = "conv4_2_dwise_weights" shape = [384, 1, 3, 3] dtype = "float32" min_val = float("-0.535525") @@ -2147,6 +2344,7 @@ class Program_weight_tensor_parameter_196: class Program_weight_tensor_parameter_197: name = "parameter_197" + original_name = "conv4_2_expand_bn_offset" shape = [384] dtype = "float32" min_val = float("-0.21279") @@ -2158,6 +2356,7 @@ class Program_weight_tensor_parameter_197: class Program_weight_tensor_parameter_198: name = "parameter_198" + original_name = "conv4_2_expand_bn_scale" shape = [384] dtype = "float32" min_val = float("-0.13955") @@ -2169,6 +2368,7 @@ class Program_weight_tensor_parameter_198: class Program_weight_tensor_parameter_199: name = "parameter_199" + original_name = "conv4_2_expand_bn_variance" shape = [384] dtype = "float32" min_val = float("2.04175e-21") @@ -2180,6 +2380,7 @@ class Program_weight_tensor_parameter_199: class Program_weight_tensor_parameter_200: name = "parameter_200" + original_name = "conv4_2_expand_bn_mean" shape = [384] dtype = "float32" min_val = float("-1.92038e-07") @@ -2191,6 +2392,7 @@ class Program_weight_tensor_parameter_200: class Program_weight_tensor_parameter_201: name = "parameter_201" + original_name = "conv4_2_expand_weights" shape = [384, 64, 1, 1] dtype = "float32" min_val = float("-0.390754") @@ -2202,6 +2404,7 @@ class Program_weight_tensor_parameter_201: class Program_weight_tensor_parameter_202: name = "parameter_202" + original_name = "conv4_1_linear_bn_offset" shape = [64] dtype = "float32" min_val = float("0") @@ -2211,6 +2414,7 @@ class Program_weight_tensor_parameter_202: class Program_weight_tensor_parameter_203: name = "parameter_203" + original_name = "conv4_1_linear_bn_scale" shape = [64] dtype = "float32" min_val = float("0") @@ -2220,6 +2424,7 @@ class Program_weight_tensor_parameter_203: class Program_weight_tensor_parameter_204: name = "parameter_204" + original_name = "conv4_1_linear_bn_variance" shape = [64] dtype = "float32" min_val = float("0") @@ -2229,6 +2434,7 @@ class Program_weight_tensor_parameter_204: class Program_weight_tensor_parameter_205: name = "parameter_205" + original_name = "conv4_1_linear_bn_mean" shape = [64] dtype = "float32" min_val = float("0") @@ -2238,6 +2444,7 @@ class Program_weight_tensor_parameter_205: class Program_weight_tensor_parameter_206: name = "parameter_206" + original_name = "conv4_1_linear_weights" shape = [64, 288, 1, 1] dtype = "float32" min_val = float("-0.758605") @@ -2249,6 +2456,7 @@ class Program_weight_tensor_parameter_206: class Program_weight_tensor_parameter_207: name = "parameter_207" + original_name = "conv4_1_dwise_bn_offset" shape = [288] dtype = "float32" min_val = float("-0.292074") @@ -2260,6 +2468,7 @@ class Program_weight_tensor_parameter_207: class Program_weight_tensor_parameter_208: name = "parameter_208" + original_name = "conv4_1_dwise_bn_scale" shape = [288] dtype = "float32" min_val = float("2.3603e-11") @@ -2271,6 +2480,7 @@ class Program_weight_tensor_parameter_208: class Program_weight_tensor_parameter_209: name = "parameter_209" + original_name = "conv4_1_dwise_bn_variance" shape = [288] dtype = "float32" min_val = float("5.60519e-45") @@ -2282,6 +2492,7 @@ class Program_weight_tensor_parameter_209: class Program_weight_tensor_parameter_210: name = "parameter_210" + original_name = "conv4_1_dwise_bn_mean" shape = [288] dtype = "float32" min_val = float("-0.388107") @@ -2293,6 +2504,7 @@ class Program_weight_tensor_parameter_210: class Program_weight_tensor_parameter_211: name = "parameter_211" + original_name = "conv4_1_dwise_weights" shape = [288, 1, 3, 3] dtype = "float32" min_val = float("-0.291649") @@ -2304,6 +2516,7 @@ class Program_weight_tensor_parameter_211: class Program_weight_tensor_parameter_212: name = "parameter_212" + original_name = "conv4_1_expand_bn_offset" shape = [288] dtype = "float32" min_val = float("-0.460139") @@ -2315,6 +2528,7 @@ class Program_weight_tensor_parameter_212: class Program_weight_tensor_parameter_213: name = "parameter_213" + original_name = "conv4_1_expand_bn_scale" shape = [288] dtype = "float32" min_val = float("-0.137911") @@ -2326,6 +2540,7 @@ class Program_weight_tensor_parameter_213: class Program_weight_tensor_parameter_214: name = "parameter_214" + original_name = "conv4_1_expand_bn_variance" shape = [288] dtype = "float32" min_val = float("1.36533e-21") @@ -2337,6 +2552,7 @@ class Program_weight_tensor_parameter_214: class Program_weight_tensor_parameter_215: name = "parameter_215" + original_name = "conv4_1_expand_bn_mean" shape = [288] dtype = "float32" min_val = float("-1.54279e-06") @@ -2348,6 +2564,7 @@ class Program_weight_tensor_parameter_215: class Program_weight_tensor_parameter_216: name = "parameter_216" + original_name = "conv4_1_expand_weights" shape = [288, 48, 1, 1] dtype = "float32" min_val = float("-0.929308") @@ -2359,6 +2576,7 @@ class Program_weight_tensor_parameter_216: class Program_weight_tensor_parameter_217: name = "parameter_217" + original_name = "conv3_2_linear_bn_offset" shape = [48] dtype = "float32" min_val = float("0") @@ -2368,6 +2586,7 @@ class Program_weight_tensor_parameter_217: class Program_weight_tensor_parameter_218: name = "parameter_218" + original_name = "conv3_2_linear_bn_scale" shape = [48] dtype = "float32" min_val = float("0") @@ -2377,6 +2596,7 @@ class Program_weight_tensor_parameter_218: class Program_weight_tensor_parameter_219: name = "parameter_219" + original_name = "conv3_2_linear_bn_variance" shape = [48] dtype = "float32" min_val = float("0") @@ -2386,6 +2606,7 @@ class Program_weight_tensor_parameter_219: class Program_weight_tensor_parameter_220: name = "parameter_220" + original_name = "conv3_2_linear_bn_mean" shape = [48] dtype = "float32" min_val = float("0") @@ -2395,6 +2616,7 @@ class Program_weight_tensor_parameter_220: class Program_weight_tensor_parameter_221: name = "parameter_221" + original_name = "conv3_2_linear_weights" shape = [48, 288, 1, 1] dtype = "float32" min_val = float("-0.614002") @@ -2406,6 +2628,7 @@ class Program_weight_tensor_parameter_221: class Program_weight_tensor_parameter_222: name = "parameter_222" + original_name = "conv3_2_dwise_bn_offset" shape = [288] dtype = "float32" min_val = float("-0.295047") @@ -2417,6 +2640,7 @@ class Program_weight_tensor_parameter_222: class Program_weight_tensor_parameter_223: name = "parameter_223" + original_name = "conv3_2_dwise_bn_scale" shape = [288] dtype = "float32" min_val = float("-0.200539") @@ -2428,6 +2652,7 @@ class Program_weight_tensor_parameter_223: class Program_weight_tensor_parameter_224: name = "parameter_224" + original_name = "conv3_2_dwise_bn_variance" shape = [288] dtype = "float32" min_val = float("5.60519e-45") @@ -2439,6 +2664,7 @@ class Program_weight_tensor_parameter_224: class Program_weight_tensor_parameter_225: name = "parameter_225" + original_name = "conv3_2_dwise_bn_mean" shape = [288] dtype = "float32" min_val = float("-0.623729") @@ -2450,6 +2676,7 @@ class Program_weight_tensor_parameter_225: class Program_weight_tensor_parameter_226: name = "parameter_226" + original_name = "conv3_2_dwise_weights" shape = [288, 1, 3, 3] dtype = "float32" min_val = float("-0.663279") @@ -2461,6 +2688,7 @@ class Program_weight_tensor_parameter_226: class Program_weight_tensor_parameter_227: name = "parameter_227" + original_name = "conv3_2_expand_bn_offset" shape = [288] dtype = "float32" min_val = float("-0.212978") @@ -2472,6 +2700,7 @@ class Program_weight_tensor_parameter_227: class Program_weight_tensor_parameter_228: name = "parameter_228" + original_name = "conv3_2_expand_bn_scale" shape = [288] dtype = "float32" min_val = float("-0.208963") @@ -2483,6 +2712,7 @@ class Program_weight_tensor_parameter_228: class Program_weight_tensor_parameter_229: name = "parameter_229" + original_name = "conv3_2_expand_bn_variance" shape = [288] dtype = "float32" min_val = float("1.9233e-21") @@ -2494,6 +2724,7 @@ class Program_weight_tensor_parameter_229: class Program_weight_tensor_parameter_230: name = "parameter_230" + original_name = "conv3_2_expand_bn_mean" shape = [288] dtype = "float32" min_val = float("-6.65191e-07") @@ -2505,6 +2736,7 @@ class Program_weight_tensor_parameter_230: class Program_weight_tensor_parameter_231: name = "parameter_231" + original_name = "conv3_2_expand_weights" shape = [288, 48, 1, 1] dtype = "float32" min_val = float("-0.519184") @@ -2516,6 +2748,7 @@ class Program_weight_tensor_parameter_231: class Program_weight_tensor_parameter_232: name = "parameter_232" + original_name = "conv3_1_linear_bn_offset" shape = [48] dtype = "float32" min_val = float("0") @@ -2525,6 +2758,7 @@ class Program_weight_tensor_parameter_232: class Program_weight_tensor_parameter_233: name = "parameter_233" + original_name = "conv3_1_linear_bn_scale" shape = [48] dtype = "float32" min_val = float("0") @@ -2534,6 +2768,7 @@ class Program_weight_tensor_parameter_233: class Program_weight_tensor_parameter_234: name = "parameter_234" + original_name = "conv3_1_linear_bn_variance" shape = [48] dtype = "float32" min_val = float("0") @@ -2543,6 +2778,7 @@ class Program_weight_tensor_parameter_234: class Program_weight_tensor_parameter_235: name = "parameter_235" + original_name = "conv3_1_linear_bn_mean" shape = [48] dtype = "float32" min_val = float("0") @@ -2552,6 +2788,7 @@ class Program_weight_tensor_parameter_235: class Program_weight_tensor_parameter_236: name = "parameter_236" + original_name = "conv3_1_linear_weights" shape = [48, 192, 1, 1] dtype = "float32" min_val = float("-0.943636") @@ -2563,6 +2800,7 @@ class Program_weight_tensor_parameter_236: class Program_weight_tensor_parameter_237: name = "parameter_237" + original_name = "conv3_1_dwise_bn_offset" shape = [192] dtype = "float32" min_val = float("-0.299798") @@ -2574,6 +2812,7 @@ class Program_weight_tensor_parameter_237: class Program_weight_tensor_parameter_238: name = "parameter_238" + original_name = "conv3_1_dwise_bn_scale" shape = [192] dtype = "float32" min_val = float("1.83327e-10") @@ -2585,6 +2824,7 @@ class Program_weight_tensor_parameter_238: class Program_weight_tensor_parameter_239: name = "parameter_239" + original_name = "conv3_1_dwise_bn_variance" shape = [192] dtype = "float32" min_val = float("5.60519e-45") @@ -2596,6 +2836,7 @@ class Program_weight_tensor_parameter_239: class Program_weight_tensor_parameter_240: name = "parameter_240" + original_name = "conv3_1_dwise_bn_mean" shape = [192] dtype = "float32" min_val = float("-0.406683") @@ -2607,6 +2848,7 @@ class Program_weight_tensor_parameter_240: class Program_weight_tensor_parameter_241: name = "parameter_241" + original_name = "conv3_1_dwise_weights" shape = [192, 1, 3, 3] dtype = "float32" min_val = float("-0.388552") @@ -2618,6 +2860,7 @@ class Program_weight_tensor_parameter_241: class Program_weight_tensor_parameter_242: name = "parameter_242" + original_name = "conv3_1_expand_bn_offset" shape = [192] dtype = "float32" min_val = float("-0.466511") @@ -2629,6 +2872,7 @@ class Program_weight_tensor_parameter_242: class Program_weight_tensor_parameter_243: name = "parameter_243" + original_name = "conv3_1_expand_bn_scale" shape = [192] dtype = "float32" min_val = float("-0.123873") @@ -2640,6 +2884,7 @@ class Program_weight_tensor_parameter_243: class Program_weight_tensor_parameter_244: name = "parameter_244" + original_name = "conv3_1_expand_bn_variance" shape = [192] dtype = "float32" min_val = float("4.52098e-20") @@ -2651,6 +2896,7 @@ class Program_weight_tensor_parameter_244: class Program_weight_tensor_parameter_245: name = "parameter_245" + original_name = "conv3_1_expand_bn_mean" shape = [192] dtype = "float32" min_val = float("-2.53685e-06") @@ -2662,6 +2908,7 @@ class Program_weight_tensor_parameter_245: class Program_weight_tensor_parameter_246: name = "parameter_246" + original_name = "conv3_1_expand_weights" shape = [192, 32, 1, 1] dtype = "float32" min_val = float("-1.19344") @@ -2673,6 +2920,7 @@ class Program_weight_tensor_parameter_246: class Program_weight_tensor_parameter_247: name = "parameter_247" + original_name = "conv2_1_linear_bn_offset" shape = [32] dtype = "float32" min_val = float("0") @@ -2682,6 +2930,7 @@ class Program_weight_tensor_parameter_247: class Program_weight_tensor_parameter_248: name = "parameter_248" + original_name = "conv2_1_linear_bn_scale" shape = [32] dtype = "float32" min_val = float("0") @@ -2691,6 +2940,7 @@ class Program_weight_tensor_parameter_248: class Program_weight_tensor_parameter_249: name = "parameter_249" + original_name = "conv2_1_linear_bn_variance" shape = [32] dtype = "float32" min_val = float("0") @@ -2700,6 +2950,7 @@ class Program_weight_tensor_parameter_249: class Program_weight_tensor_parameter_250: name = "parameter_250" + original_name = "conv2_1_linear_bn_mean" shape = [32] dtype = "float32" min_val = float("0") @@ -2709,6 +2960,7 @@ class Program_weight_tensor_parameter_250: class Program_weight_tensor_parameter_251: name = "parameter_251" + original_name = "conv2_1_linear_weights" shape = [32, 64, 1, 1] dtype = "float32" min_val = float("-1.13692") @@ -2720,6 +2972,7 @@ class Program_weight_tensor_parameter_251: class Program_weight_tensor_parameter_252: name = "parameter_252" + original_name = "conv2_1_dwise_bn_offset" shape = [64] dtype = "float32" min_val = float("0") @@ -2729,6 +2982,7 @@ class Program_weight_tensor_parameter_252: class Program_weight_tensor_parameter_253: name = "parameter_253" + original_name = "conv2_1_dwise_bn_scale" shape = [64] dtype = "float32" min_val = float("0") @@ -2738,6 +2992,7 @@ class Program_weight_tensor_parameter_253: class Program_weight_tensor_parameter_254: name = "parameter_254" + original_name = "conv2_1_dwise_bn_variance" shape = [64] dtype = "float32" min_val = float("0") @@ -2747,6 +3002,7 @@ class Program_weight_tensor_parameter_254: class Program_weight_tensor_parameter_255: name = "parameter_255" + original_name = "conv2_1_dwise_bn_mean" shape = [64] dtype = "float32" min_val = float("0") @@ -2756,6 +3012,7 @@ class Program_weight_tensor_parameter_255: class Program_weight_tensor_parameter_256: name = "parameter_256" + original_name = "conv2_1_dwise_weights" shape = [64, 1, 3, 3] dtype = "float32" min_val = float("-1.74189") @@ -2767,6 +3024,7 @@ class Program_weight_tensor_parameter_256: class Program_weight_tensor_parameter_257: name = "parameter_257" + original_name = "conv2_1_expand_bn_offset" shape = [64] dtype = "float32" min_val = float("0") @@ -2776,6 +3034,7 @@ class Program_weight_tensor_parameter_257: class Program_weight_tensor_parameter_258: name = "parameter_258" + original_name = "conv2_1_expand_bn_scale" shape = [64] dtype = "float32" min_val = float("0") @@ -2785,6 +3044,7 @@ class Program_weight_tensor_parameter_258: class Program_weight_tensor_parameter_259: name = "parameter_259" + original_name = "conv2_1_expand_bn_variance" shape = [64] dtype = "float32" min_val = float("0") @@ -2794,6 +3054,7 @@ class Program_weight_tensor_parameter_259: class Program_weight_tensor_parameter_260: name = "parameter_260" + original_name = "conv2_1_expand_bn_mean" shape = [64] dtype = "float32" min_val = float("0") @@ -2803,6 +3064,7 @@ class Program_weight_tensor_parameter_260: class Program_weight_tensor_parameter_261: name = "parameter_261" + original_name = "conv2_1_expand_weights" shape = [64, 64, 1, 1] dtype = "float32" min_val = float("-0.895546") @@ -2814,6 +3076,7 @@ class Program_weight_tensor_parameter_261: class Program_weight_tensor_parameter_262: name = "parameter_262" + original_name = "conv1_1_bn_offset" shape = [64] dtype = "float32" min_val = float("0") @@ -2823,6 +3086,7 @@ class Program_weight_tensor_parameter_262: class Program_weight_tensor_parameter_263: name = "parameter_263" + original_name = "conv1_1_bn_scale" shape = [64] dtype = "float32" min_val = float("0") @@ -2832,6 +3096,7 @@ class Program_weight_tensor_parameter_263: class Program_weight_tensor_parameter_264: name = "parameter_264" + original_name = "conv1_1_bn_variance" shape = [64] dtype = "float32" min_val = float("0") @@ -2841,6 +3106,7 @@ class Program_weight_tensor_parameter_264: class Program_weight_tensor_parameter_265: name = "parameter_265" + original_name = "conv1_1_bn_mean" shape = [64] dtype = "float32" min_val = float("0") @@ -2850,6 +3116,7 @@ class Program_weight_tensor_parameter_265: class Program_weight_tensor_parameter_266: name = "parameter_266" + original_name = "conv1_1_weights" shape = [64, 3, 3, 3] dtype = "float32" min_val = float("-0.980946") diff --git a/paddle_samples/PaddleX/MobileNetV3_large_x0_35/subgraph_1/input_meta.py b/paddle_samples/PaddleX/MobileNetV3_large_x0_35/subgraph_1/input_meta.py index 795f9bcfe..6d7dac1c9 100644 --- a/paddle_samples/PaddleX/MobileNetV3_large_x0_35/subgraph_1/input_meta.py +++ b/paddle_samples/PaddleX/MobileNetV3_large_x0_35/subgraph_1/input_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_data_0: name = "data_0" + original_name = "var_0" shape = [128, 3, 224, 224] dtype = "float32" min_val = float("-2.1179") diff --git a/paddle_samples/PaddleX/MobileNetV3_large_x0_35/subgraph_1/weight_meta.py b/paddle_samples/PaddleX/MobileNetV3_large_x0_35/subgraph_1/weight_meta.py index 9976629eb..7a108a420 100644 --- a/paddle_samples/PaddleX/MobileNetV3_large_x0_35/subgraph_1/weight_meta.py +++ b/paddle_samples/PaddleX/MobileNetV3_large_x0_35/subgraph_1/weight_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_parameter_0: name = "parameter_0" + original_name = "linear_0.b_0" shape = [102] dtype = "float32" data = None @@ -7,6 +8,7 @@ class Program_weight_tensor_parameter_0: class Program_weight_tensor_parameter_1: name = "parameter_1" + original_name = "linear_0.w_0" shape = [1280, 102] dtype = "float32" min_val = float("-0.0658888") @@ -18,6 +20,7 @@ class Program_weight_tensor_parameter_1: class Program_weight_tensor_parameter_2: name = "parameter_2" + original_name = "conv2d_63.w_0" shape = [1280, 336, 1, 1] dtype = "float32" min_val = float("-0.797415") @@ -29,6 +32,7 @@ class Program_weight_tensor_parameter_2: class Program_weight_tensor_parameter_3: name = "parameter_3" + original_name = "batch_norm_46.b_0" shape = [336] dtype = "float32" min_val = float("-12.7312") @@ -40,6 +44,7 @@ class Program_weight_tensor_parameter_3: class Program_weight_tensor_parameter_4: name = "parameter_4" + original_name = "batch_norm_46.w_0" shape = [336] dtype = "float32" min_val = float("2.70503") @@ -51,6 +56,7 @@ class Program_weight_tensor_parameter_4: class Program_weight_tensor_parameter_5: name = "parameter_5" + original_name = "batch_norm_46.w_2" shape = [336] dtype = "float32" min_val = float("50.0128") @@ -62,6 +68,7 @@ class Program_weight_tensor_parameter_5: class Program_weight_tensor_parameter_6: name = "parameter_6" + original_name = "batch_norm_46.w_1" shape = [336] dtype = "float32" min_val = float("-1.26245e-05") @@ -73,6 +80,7 @@ class Program_weight_tensor_parameter_6: class Program_weight_tensor_parameter_7: name = "parameter_7" + original_name = "conv2d_62.w_0" shape = [336, 56, 1, 1] dtype = "float32" min_val = float("-1.32997") @@ -84,6 +92,7 @@ class Program_weight_tensor_parameter_7: class Program_weight_tensor_parameter_8: name = "parameter_8" + original_name = "batch_norm_45.b_0" shape = [56] dtype = "float32" min_val = float("0") @@ -93,6 +102,7 @@ class Program_weight_tensor_parameter_8: class Program_weight_tensor_parameter_9: name = "parameter_9" + original_name = "batch_norm_45.w_0" shape = [56] dtype = "float32" min_val = float("0") @@ -102,6 +112,7 @@ class Program_weight_tensor_parameter_9: class Program_weight_tensor_parameter_10: name = "parameter_10" + original_name = "batch_norm_45.w_2" shape = [56] dtype = "float32" min_val = float("0") @@ -111,6 +122,7 @@ class Program_weight_tensor_parameter_10: class Program_weight_tensor_parameter_11: name = "parameter_11" + original_name = "batch_norm_45.w_1" shape = [56] dtype = "float32" min_val = float("0") @@ -120,6 +132,7 @@ class Program_weight_tensor_parameter_11: class Program_weight_tensor_parameter_12: name = "parameter_12" + original_name = "conv2d_61.w_0" shape = [56, 336, 1, 1] dtype = "float32" min_val = float("-0.711538") @@ -131,6 +144,7 @@ class Program_weight_tensor_parameter_12: class Program_weight_tensor_parameter_13: name = "parameter_13" + original_name = "conv2d_60.b_0" shape = [336] dtype = "float32" min_val = float("-0.688471") @@ -142,6 +156,7 @@ class Program_weight_tensor_parameter_13: class Program_weight_tensor_parameter_14: name = "parameter_14" + original_name = "conv2d_60.w_0" shape = [336, 84, 1, 1] dtype = "float32" min_val = float("-0.558542") @@ -153,6 +168,7 @@ class Program_weight_tensor_parameter_14: class Program_weight_tensor_parameter_15: name = "parameter_15" + original_name = "conv2d_59.b_0" shape = [84] dtype = "float32" min_val = float("-0.136014") @@ -164,6 +180,7 @@ class Program_weight_tensor_parameter_15: class Program_weight_tensor_parameter_16: name = "parameter_16" + original_name = "conv2d_59.w_0" shape = [84, 336, 1, 1] dtype = "float32" min_val = float("-0.603048") @@ -175,6 +192,7 @@ class Program_weight_tensor_parameter_16: class Program_weight_tensor_parameter_17: name = "parameter_17" + original_name = "batch_norm_44.b_0" shape = [336] dtype = "float32" min_val = float("-3.51271") @@ -186,6 +204,7 @@ class Program_weight_tensor_parameter_17: class Program_weight_tensor_parameter_18: name = "parameter_18" + original_name = "batch_norm_44.w_0" shape = [336] dtype = "float32" min_val = float("0.107752") @@ -197,6 +216,7 @@ class Program_weight_tensor_parameter_18: class Program_weight_tensor_parameter_19: name = "parameter_19" + original_name = "batch_norm_44.w_2" shape = [336] dtype = "float32" min_val = float("6.586e-05") @@ -208,6 +228,7 @@ class Program_weight_tensor_parameter_19: class Program_weight_tensor_parameter_20: name = "parameter_20" + original_name = "batch_norm_44.w_1" shape = [336] dtype = "float32" min_val = float("-3.7609") @@ -219,6 +240,7 @@ class Program_weight_tensor_parameter_20: class Program_weight_tensor_parameter_21: name = "parameter_21" + original_name = "conv2d_58.w_0" shape = [336, 1, 5, 5] dtype = "float32" min_val = float("-0.978072") @@ -230,6 +252,7 @@ class Program_weight_tensor_parameter_21: class Program_weight_tensor_parameter_22: name = "parameter_22" + original_name = "batch_norm_43.b_0" shape = [336] dtype = "float32" min_val = float("-3.53109") @@ -241,6 +264,7 @@ class Program_weight_tensor_parameter_22: class Program_weight_tensor_parameter_23: name = "parameter_23" + original_name = "batch_norm_43.w_0" shape = [336] dtype = "float32" min_val = float("-1.18346") @@ -252,6 +276,7 @@ class Program_weight_tensor_parameter_23: class Program_weight_tensor_parameter_24: name = "parameter_24" + original_name = "batch_norm_43.w_2" shape = [336] dtype = "float32" min_val = float("6.38172") @@ -263,6 +288,7 @@ class Program_weight_tensor_parameter_24: class Program_weight_tensor_parameter_25: name = "parameter_25" + original_name = "batch_norm_43.w_1" shape = [336] dtype = "float32" min_val = float("-9.75298e-06") @@ -274,6 +300,7 @@ class Program_weight_tensor_parameter_25: class Program_weight_tensor_parameter_26: name = "parameter_26" + original_name = "conv2d_57.w_0" shape = [336, 56, 1, 1] dtype = "float32" min_val = float("-0.568171") @@ -285,6 +312,7 @@ class Program_weight_tensor_parameter_26: class Program_weight_tensor_parameter_27: name = "parameter_27" + original_name = "batch_norm_42.b_0" shape = [56] dtype = "float32" min_val = float("0") @@ -294,6 +322,7 @@ class Program_weight_tensor_parameter_27: class Program_weight_tensor_parameter_28: name = "parameter_28" + original_name = "batch_norm_42.w_0" shape = [56] dtype = "float32" min_val = float("0") @@ -303,6 +332,7 @@ class Program_weight_tensor_parameter_28: class Program_weight_tensor_parameter_29: name = "parameter_29" + original_name = "batch_norm_42.w_2" shape = [56] dtype = "float32" min_val = float("0") @@ -312,6 +342,7 @@ class Program_weight_tensor_parameter_29: class Program_weight_tensor_parameter_30: name = "parameter_30" + original_name = "batch_norm_42.w_1" shape = [56] dtype = "float32" min_val = float("0") @@ -321,6 +352,7 @@ class Program_weight_tensor_parameter_30: class Program_weight_tensor_parameter_31: name = "parameter_31" + original_name = "conv2d_56.w_0" shape = [56, 336, 1, 1] dtype = "float32" min_val = float("-0.633387") @@ -332,6 +364,7 @@ class Program_weight_tensor_parameter_31: class Program_weight_tensor_parameter_32: name = "parameter_32" + original_name = "conv2d_55.b_0" shape = [336] dtype = "float32" min_val = float("-0.323771") @@ -343,6 +376,7 @@ class Program_weight_tensor_parameter_32: class Program_weight_tensor_parameter_33: name = "parameter_33" + original_name = "conv2d_55.w_0" shape = [336, 84, 1, 1] dtype = "float32" min_val = float("-0.637123") @@ -354,6 +388,7 @@ class Program_weight_tensor_parameter_33: class Program_weight_tensor_parameter_34: name = "parameter_34" + original_name = "conv2d_54.b_0" shape = [84] dtype = "float32" min_val = float("-0.119348") @@ -365,6 +400,7 @@ class Program_weight_tensor_parameter_34: class Program_weight_tensor_parameter_35: name = "parameter_35" + original_name = "conv2d_54.w_0" shape = [84, 336, 1, 1] dtype = "float32" min_val = float("-0.893996") @@ -376,6 +412,7 @@ class Program_weight_tensor_parameter_35: class Program_weight_tensor_parameter_36: name = "parameter_36" + original_name = "batch_norm_41.b_0" shape = [336] dtype = "float32" min_val = float("-4.50558") @@ -387,6 +424,7 @@ class Program_weight_tensor_parameter_36: class Program_weight_tensor_parameter_37: name = "parameter_37" + original_name = "batch_norm_41.w_0" shape = [336] dtype = "float32" min_val = float("0.356013") @@ -398,6 +436,7 @@ class Program_weight_tensor_parameter_37: class Program_weight_tensor_parameter_38: name = "parameter_38" + original_name = "batch_norm_41.w_2" shape = [336] dtype = "float32" min_val = float("0.000150831") @@ -409,6 +448,7 @@ class Program_weight_tensor_parameter_38: class Program_weight_tensor_parameter_39: name = "parameter_39" + original_name = "batch_norm_41.w_1" shape = [336] dtype = "float32" min_val = float("-4.93938") @@ -420,6 +460,7 @@ class Program_weight_tensor_parameter_39: class Program_weight_tensor_parameter_40: name = "parameter_40" + original_name = "conv2d_53.w_0" shape = [336, 1, 5, 5] dtype = "float32" min_val = float("-1.0208") @@ -431,6 +472,7 @@ class Program_weight_tensor_parameter_40: class Program_weight_tensor_parameter_41: name = "parameter_41" + original_name = "batch_norm_40.b_0" shape = [336] dtype = "float32" min_val = float("-3.50603") @@ -442,6 +484,7 @@ class Program_weight_tensor_parameter_41: class Program_weight_tensor_parameter_42: name = "parameter_42" + original_name = "batch_norm_40.w_0" shape = [336] dtype = "float32" min_val = float("-0.322644") @@ -453,6 +496,7 @@ class Program_weight_tensor_parameter_42: class Program_weight_tensor_parameter_43: name = "parameter_43" + original_name = "batch_norm_40.w_2" shape = [336] dtype = "float32" min_val = float("9.09732") @@ -464,6 +508,7 @@ class Program_weight_tensor_parameter_43: class Program_weight_tensor_parameter_44: name = "parameter_44" + original_name = "batch_norm_40.w_1" shape = [336] dtype = "float32" min_val = float("-3.70001e-06") @@ -475,6 +520,7 @@ class Program_weight_tensor_parameter_44: class Program_weight_tensor_parameter_45: name = "parameter_45" + original_name = "conv2d_52.w_0" shape = [336, 56, 1, 1] dtype = "float32" min_val = float("-0.546904") @@ -486,6 +532,7 @@ class Program_weight_tensor_parameter_45: class Program_weight_tensor_parameter_46: name = "parameter_46" + original_name = "batch_norm_39.b_0" shape = [56] dtype = "float32" min_val = float("0") @@ -495,6 +542,7 @@ class Program_weight_tensor_parameter_46: class Program_weight_tensor_parameter_47: name = "parameter_47" + original_name = "batch_norm_39.w_0" shape = [56] dtype = "float32" min_val = float("0") @@ -504,6 +552,7 @@ class Program_weight_tensor_parameter_47: class Program_weight_tensor_parameter_48: name = "parameter_48" + original_name = "batch_norm_39.w_2" shape = [56] dtype = "float32" min_val = float("0") @@ -513,6 +562,7 @@ class Program_weight_tensor_parameter_48: class Program_weight_tensor_parameter_49: name = "parameter_49" + original_name = "batch_norm_39.w_1" shape = [56] dtype = "float32" min_val = float("0") @@ -522,6 +572,7 @@ class Program_weight_tensor_parameter_49: class Program_weight_tensor_parameter_50: name = "parameter_50" + original_name = "conv2d_51.w_0" shape = [56, 232, 1, 1] dtype = "float32" min_val = float("-1.06909") @@ -533,6 +584,7 @@ class Program_weight_tensor_parameter_50: class Program_weight_tensor_parameter_51: name = "parameter_51" + original_name = "conv2d_50.b_0" shape = [232] dtype = "float32" min_val = float("-0.419397") @@ -544,6 +596,7 @@ class Program_weight_tensor_parameter_51: class Program_weight_tensor_parameter_52: name = "parameter_52" + original_name = "conv2d_50.w_0" shape = [232, 58, 1, 1] dtype = "float32" min_val = float("-0.937691") @@ -555,6 +608,7 @@ class Program_weight_tensor_parameter_52: class Program_weight_tensor_parameter_53: name = "parameter_53" + original_name = "conv2d_49.b_0" shape = [58] dtype = "float32" min_val = float("0") @@ -564,6 +618,7 @@ class Program_weight_tensor_parameter_53: class Program_weight_tensor_parameter_54: name = "parameter_54" + original_name = "conv2d_49.w_0" shape = [58, 232, 1, 1] dtype = "float32" min_val = float("-0.658062") @@ -575,6 +630,7 @@ class Program_weight_tensor_parameter_54: class Program_weight_tensor_parameter_55: name = "parameter_55" + original_name = "batch_norm_38.b_0" shape = [232] dtype = "float32" min_val = float("-2.31321") @@ -586,6 +642,7 @@ class Program_weight_tensor_parameter_55: class Program_weight_tensor_parameter_56: name = "parameter_56" + original_name = "batch_norm_38.w_0" shape = [232] dtype = "float32" min_val = float("0.274919") @@ -597,6 +654,7 @@ class Program_weight_tensor_parameter_56: class Program_weight_tensor_parameter_57: name = "parameter_57" + original_name = "batch_norm_38.w_2" shape = [232] dtype = "float32" min_val = float("0.000264627") @@ -608,6 +666,7 @@ class Program_weight_tensor_parameter_57: class Program_weight_tensor_parameter_58: name = "parameter_58" + original_name = "batch_norm_38.w_1" shape = [232] dtype = "float32" min_val = float("-11.524") @@ -619,6 +678,7 @@ class Program_weight_tensor_parameter_58: class Program_weight_tensor_parameter_59: name = "parameter_59" + original_name = "conv2d_48.w_0" shape = [232, 1, 5, 5] dtype = "float32" min_val = float("-0.404835") @@ -630,6 +690,7 @@ class Program_weight_tensor_parameter_59: class Program_weight_tensor_parameter_60: name = "parameter_60" + original_name = "batch_norm_37.b_0" shape = [232] dtype = "float32" min_val = float("-3.85738") @@ -641,6 +702,7 @@ class Program_weight_tensor_parameter_60: class Program_weight_tensor_parameter_61: name = "parameter_61" + original_name = "batch_norm_37.w_0" shape = [232] dtype = "float32" min_val = float("-1.34618") @@ -652,6 +714,7 @@ class Program_weight_tensor_parameter_61: class Program_weight_tensor_parameter_62: name = "parameter_62" + original_name = "batch_norm_37.w_2" shape = [232] dtype = "float32" min_val = float("22.0105") @@ -663,6 +726,7 @@ class Program_weight_tensor_parameter_62: class Program_weight_tensor_parameter_63: name = "parameter_63" + original_name = "batch_norm_37.w_1" shape = [232] dtype = "float32" min_val = float("-1.14001e-05") @@ -674,6 +738,7 @@ class Program_weight_tensor_parameter_63: class Program_weight_tensor_parameter_64: name = "parameter_64" + original_name = "conv2d_47.w_0" shape = [232, 40, 1, 1] dtype = "float32" min_val = float("-0.740918") @@ -685,6 +750,7 @@ class Program_weight_tensor_parameter_64: class Program_weight_tensor_parameter_65: name = "parameter_65" + original_name = "batch_norm_36.b_0" shape = [40] dtype = "float32" min_val = float("0") @@ -694,6 +760,7 @@ class Program_weight_tensor_parameter_65: class Program_weight_tensor_parameter_66: name = "parameter_66" + original_name = "batch_norm_36.w_0" shape = [40] dtype = "float32" min_val = float("0") @@ -703,6 +770,7 @@ class Program_weight_tensor_parameter_66: class Program_weight_tensor_parameter_67: name = "parameter_67" + original_name = "batch_norm_36.w_2" shape = [40] dtype = "float32" min_val = float("0") @@ -712,6 +780,7 @@ class Program_weight_tensor_parameter_67: class Program_weight_tensor_parameter_68: name = "parameter_68" + original_name = "batch_norm_36.w_1" shape = [40] dtype = "float32" min_val = float("0") @@ -721,6 +790,7 @@ class Program_weight_tensor_parameter_68: class Program_weight_tensor_parameter_69: name = "parameter_69" + original_name = "conv2d_46.w_0" shape = [40, 232, 1, 1] dtype = "float32" min_val = float("-0.645461") @@ -732,6 +802,7 @@ class Program_weight_tensor_parameter_69: class Program_weight_tensor_parameter_70: name = "parameter_70" + original_name = "conv2d_45.b_0" shape = [232] dtype = "float32" min_val = float("-0.372776") @@ -743,6 +814,7 @@ class Program_weight_tensor_parameter_70: class Program_weight_tensor_parameter_71: name = "parameter_71" + original_name = "conv2d_45.w_0" shape = [232, 58, 1, 1] dtype = "float32" min_val = float("-0.732399") @@ -754,6 +826,7 @@ class Program_weight_tensor_parameter_71: class Program_weight_tensor_parameter_72: name = "parameter_72" + original_name = "conv2d_44.b_0" shape = [58] dtype = "float32" min_val = float("0") @@ -763,6 +836,7 @@ class Program_weight_tensor_parameter_72: class Program_weight_tensor_parameter_73: name = "parameter_73" + original_name = "conv2d_44.w_0" shape = [58, 232, 1, 1] dtype = "float32" min_val = float("-0.96416") @@ -774,6 +848,7 @@ class Program_weight_tensor_parameter_73: class Program_weight_tensor_parameter_74: name = "parameter_74" + original_name = "batch_norm_35.b_0" shape = [232] dtype = "float32" min_val = float("-4.81927") @@ -785,6 +860,7 @@ class Program_weight_tensor_parameter_74: class Program_weight_tensor_parameter_75: name = "parameter_75" + original_name = "batch_norm_35.w_0" shape = [232] dtype = "float32" min_val = float("0.197465") @@ -796,6 +872,7 @@ class Program_weight_tensor_parameter_75: class Program_weight_tensor_parameter_76: name = "parameter_76" + original_name = "batch_norm_35.w_2" shape = [232] dtype = "float32" min_val = float("0.000123798") @@ -807,6 +884,7 @@ class Program_weight_tensor_parameter_76: class Program_weight_tensor_parameter_77: name = "parameter_77" + original_name = "batch_norm_35.w_1" shape = [232] dtype = "float32" min_val = float("-3.59429") @@ -818,6 +896,7 @@ class Program_weight_tensor_parameter_77: class Program_weight_tensor_parameter_78: name = "parameter_78" + original_name = "conv2d_43.w_0" shape = [232, 1, 3, 3] dtype = "float32" min_val = float("-0.895537") @@ -829,6 +908,7 @@ class Program_weight_tensor_parameter_78: class Program_weight_tensor_parameter_79: name = "parameter_79" + original_name = "batch_norm_34.b_0" shape = [232] dtype = "float32" min_val = float("-3.13185") @@ -840,6 +920,7 @@ class Program_weight_tensor_parameter_79: class Program_weight_tensor_parameter_80: name = "parameter_80" + original_name = "batch_norm_34.w_0" shape = [232] dtype = "float32" min_val = float("-0.640948") @@ -851,6 +932,7 @@ class Program_weight_tensor_parameter_80: class Program_weight_tensor_parameter_81: name = "parameter_81" + original_name = "batch_norm_34.w_2" shape = [232] dtype = "float32" min_val = float("7.60394") @@ -862,6 +944,7 @@ class Program_weight_tensor_parameter_81: class Program_weight_tensor_parameter_82: name = "parameter_82" + original_name = "batch_norm_34.w_1" shape = [232] dtype = "float32" min_val = float("-4.73143e-06") @@ -873,6 +956,7 @@ class Program_weight_tensor_parameter_82: class Program_weight_tensor_parameter_83: name = "parameter_83" + original_name = "conv2d_42.w_0" shape = [232, 40, 1, 1] dtype = "float32" min_val = float("-0.723486") @@ -884,6 +968,7 @@ class Program_weight_tensor_parameter_83: class Program_weight_tensor_parameter_84: name = "parameter_84" + original_name = "batch_norm_33.b_0" shape = [40] dtype = "float32" min_val = float("0") @@ -893,6 +978,7 @@ class Program_weight_tensor_parameter_84: class Program_weight_tensor_parameter_85: name = "parameter_85" + original_name = "batch_norm_33.w_0" shape = [40] dtype = "float32" min_val = float("0") @@ -902,6 +988,7 @@ class Program_weight_tensor_parameter_85: class Program_weight_tensor_parameter_86: name = "parameter_86" + original_name = "batch_norm_33.w_2" shape = [40] dtype = "float32" min_val = float("0") @@ -911,6 +998,7 @@ class Program_weight_tensor_parameter_86: class Program_weight_tensor_parameter_87: name = "parameter_87" + original_name = "batch_norm_33.w_1" shape = [40] dtype = "float32" min_val = float("0") @@ -920,6 +1008,7 @@ class Program_weight_tensor_parameter_87: class Program_weight_tensor_parameter_88: name = "parameter_88" + original_name = "conv2d_41.w_0" shape = [40, 168, 1, 1] dtype = "float32" min_val = float("-0.804418") @@ -931,6 +1020,7 @@ class Program_weight_tensor_parameter_88: class Program_weight_tensor_parameter_89: name = "parameter_89" + original_name = "conv2d_40.b_0" shape = [168] dtype = "float32" min_val = float("-0.503427") @@ -942,6 +1032,7 @@ class Program_weight_tensor_parameter_89: class Program_weight_tensor_parameter_90: name = "parameter_90" + original_name = "conv2d_40.w_0" shape = [168, 42, 1, 1] dtype = "float32" min_val = float("-1.12128") @@ -953,6 +1044,7 @@ class Program_weight_tensor_parameter_90: class Program_weight_tensor_parameter_91: name = "parameter_91" + original_name = "conv2d_39.b_0" shape = [42] dtype = "float32" min_val = float("0") @@ -962,6 +1054,7 @@ class Program_weight_tensor_parameter_91: class Program_weight_tensor_parameter_92: name = "parameter_92" + original_name = "conv2d_39.w_0" shape = [42, 168, 1, 1] dtype = "float32" min_val = float("-1.10887") @@ -973,6 +1066,7 @@ class Program_weight_tensor_parameter_92: class Program_weight_tensor_parameter_93: name = "parameter_93" + original_name = "batch_norm_32.b_0" shape = [168] dtype = "float32" min_val = float("-2.02416") @@ -984,6 +1078,7 @@ class Program_weight_tensor_parameter_93: class Program_weight_tensor_parameter_94: name = "parameter_94" + original_name = "batch_norm_32.w_0" shape = [168] dtype = "float32" min_val = float("-0.373414") @@ -995,6 +1090,7 @@ class Program_weight_tensor_parameter_94: class Program_weight_tensor_parameter_95: name = "parameter_95" + original_name = "batch_norm_32.w_2" shape = [168] dtype = "float32" min_val = float("0.000238601") @@ -1006,6 +1102,7 @@ class Program_weight_tensor_parameter_95: class Program_weight_tensor_parameter_96: name = "parameter_96" + original_name = "batch_norm_32.w_1" shape = [168] dtype = "float32" min_val = float("-12.9869") @@ -1017,6 +1114,7 @@ class Program_weight_tensor_parameter_96: class Program_weight_tensor_parameter_97: name = "parameter_97" + original_name = "conv2d_38.w_0" shape = [168, 1, 3, 3] dtype = "float32" min_val = float("-1.47211") @@ -1028,6 +1126,7 @@ class Program_weight_tensor_parameter_97: class Program_weight_tensor_parameter_98: name = "parameter_98" + original_name = "batch_norm_31.b_0" shape = [168] dtype = "float32" min_val = float("-2.73562") @@ -1039,6 +1138,7 @@ class Program_weight_tensor_parameter_98: class Program_weight_tensor_parameter_99: name = "parameter_99" + original_name = "batch_norm_31.w_0" shape = [168] dtype = "float32" min_val = float("-7.56043") @@ -1050,6 +1150,7 @@ class Program_weight_tensor_parameter_99: class Program_weight_tensor_parameter_100: name = "parameter_100" + original_name = "batch_norm_31.w_2" shape = [168] dtype = "float32" min_val = float("33.8432") @@ -1061,6 +1162,7 @@ class Program_weight_tensor_parameter_100: class Program_weight_tensor_parameter_101: name = "parameter_101" + original_name = "batch_norm_31.w_1" shape = [168] dtype = "float32" min_val = float("-2.54346e-05") @@ -1072,6 +1174,7 @@ class Program_weight_tensor_parameter_101: class Program_weight_tensor_parameter_102: name = "parameter_102" + original_name = "conv2d_37.w_0" shape = [168, 32, 1, 1] dtype = "float32" min_val = float("-1.10432") @@ -1083,6 +1186,7 @@ class Program_weight_tensor_parameter_102: class Program_weight_tensor_parameter_103: name = "parameter_103" + original_name = "batch_norm_30.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -1092,6 +1196,7 @@ class Program_weight_tensor_parameter_103: class Program_weight_tensor_parameter_104: name = "parameter_104" + original_name = "batch_norm_30.w_0" shape = [32] dtype = "float32" min_val = float("0") @@ -1101,6 +1206,7 @@ class Program_weight_tensor_parameter_104: class Program_weight_tensor_parameter_105: name = "parameter_105" + original_name = "batch_norm_30.w_2" shape = [32] dtype = "float32" min_val = float("0") @@ -1110,6 +1216,7 @@ class Program_weight_tensor_parameter_105: class Program_weight_tensor_parameter_106: name = "parameter_106" + original_name = "batch_norm_30.w_1" shape = [32] dtype = "float32" min_val = float("0") @@ -1119,6 +1226,7 @@ class Program_weight_tensor_parameter_106: class Program_weight_tensor_parameter_107: name = "parameter_107" + original_name = "conv2d_36.w_0" shape = [32, 64, 1, 1] dtype = "float32" min_val = float("-1.36445") @@ -1130,6 +1238,7 @@ class Program_weight_tensor_parameter_107: class Program_weight_tensor_parameter_108: name = "parameter_108" + original_name = "batch_norm_29.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -1139,6 +1248,7 @@ class Program_weight_tensor_parameter_108: class Program_weight_tensor_parameter_109: name = "parameter_109" + original_name = "batch_norm_29.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -1148,6 +1258,7 @@ class Program_weight_tensor_parameter_109: class Program_weight_tensor_parameter_110: name = "parameter_110" + original_name = "batch_norm_29.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -1157,6 +1268,7 @@ class Program_weight_tensor_parameter_110: class Program_weight_tensor_parameter_111: name = "parameter_111" + original_name = "batch_norm_29.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -1166,6 +1278,7 @@ class Program_weight_tensor_parameter_111: class Program_weight_tensor_parameter_112: name = "parameter_112" + original_name = "conv2d_35.w_0" shape = [64, 1, 3, 3] dtype = "float32" min_val = float("-1.03936") @@ -1177,6 +1290,7 @@ class Program_weight_tensor_parameter_112: class Program_weight_tensor_parameter_113: name = "parameter_113" + original_name = "batch_norm_28.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -1186,6 +1300,7 @@ class Program_weight_tensor_parameter_113: class Program_weight_tensor_parameter_114: name = "parameter_114" + original_name = "batch_norm_28.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -1195,6 +1310,7 @@ class Program_weight_tensor_parameter_114: class Program_weight_tensor_parameter_115: name = "parameter_115" + original_name = "batch_norm_28.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -1204,6 +1320,7 @@ class Program_weight_tensor_parameter_115: class Program_weight_tensor_parameter_116: name = "parameter_116" + original_name = "batch_norm_28.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -1213,6 +1330,7 @@ class Program_weight_tensor_parameter_116: class Program_weight_tensor_parameter_117: name = "parameter_117" + original_name = "conv2d_34.w_0" shape = [64, 32, 1, 1] dtype = "float32" min_val = float("-1.30232") @@ -1224,6 +1342,7 @@ class Program_weight_tensor_parameter_117: class Program_weight_tensor_parameter_118: name = "parameter_118" + original_name = "batch_norm_27.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -1233,6 +1352,7 @@ class Program_weight_tensor_parameter_118: class Program_weight_tensor_parameter_119: name = "parameter_119" + original_name = "batch_norm_27.w_0" shape = [32] dtype = "float32" min_val = float("0") @@ -1242,6 +1362,7 @@ class Program_weight_tensor_parameter_119: class Program_weight_tensor_parameter_120: name = "parameter_120" + original_name = "batch_norm_27.w_2" shape = [32] dtype = "float32" min_val = float("0") @@ -1251,6 +1372,7 @@ class Program_weight_tensor_parameter_120: class Program_weight_tensor_parameter_121: name = "parameter_121" + original_name = "batch_norm_27.w_1" shape = [32] dtype = "float32" min_val = float("0") @@ -1260,6 +1382,7 @@ class Program_weight_tensor_parameter_121: class Program_weight_tensor_parameter_122: name = "parameter_122" + original_name = "conv2d_33.w_0" shape = [32, 64, 1, 1] dtype = "float32" min_val = float("-0.619016") @@ -1271,6 +1394,7 @@ class Program_weight_tensor_parameter_122: class Program_weight_tensor_parameter_123: name = "parameter_123" + original_name = "batch_norm_26.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -1280,6 +1404,7 @@ class Program_weight_tensor_parameter_123: class Program_weight_tensor_parameter_124: name = "parameter_124" + original_name = "batch_norm_26.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -1289,6 +1414,7 @@ class Program_weight_tensor_parameter_124: class Program_weight_tensor_parameter_125: name = "parameter_125" + original_name = "batch_norm_26.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -1298,6 +1424,7 @@ class Program_weight_tensor_parameter_125: class Program_weight_tensor_parameter_126: name = "parameter_126" + original_name = "batch_norm_26.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -1307,6 +1434,7 @@ class Program_weight_tensor_parameter_126: class Program_weight_tensor_parameter_127: name = "parameter_127" + original_name = "conv2d_32.w_0" shape = [64, 1, 3, 3] dtype = "float32" min_val = float("-1.00592") @@ -1318,6 +1446,7 @@ class Program_weight_tensor_parameter_127: class Program_weight_tensor_parameter_128: name = "parameter_128" + original_name = "batch_norm_25.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -1327,6 +1456,7 @@ class Program_weight_tensor_parameter_128: class Program_weight_tensor_parameter_129: name = "parameter_129" + original_name = "batch_norm_25.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -1336,6 +1466,7 @@ class Program_weight_tensor_parameter_129: class Program_weight_tensor_parameter_130: name = "parameter_130" + original_name = "batch_norm_25.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -1345,6 +1476,7 @@ class Program_weight_tensor_parameter_130: class Program_weight_tensor_parameter_131: name = "parameter_131" + original_name = "batch_norm_25.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -1354,6 +1486,7 @@ class Program_weight_tensor_parameter_131: class Program_weight_tensor_parameter_132: name = "parameter_132" + original_name = "conv2d_31.w_0" shape = [64, 32, 1, 1] dtype = "float32" min_val = float("-1.04582") @@ -1365,6 +1498,7 @@ class Program_weight_tensor_parameter_132: class Program_weight_tensor_parameter_133: name = "parameter_133" + original_name = "batch_norm_24.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -1374,6 +1508,7 @@ class Program_weight_tensor_parameter_133: class Program_weight_tensor_parameter_134: name = "parameter_134" + original_name = "batch_norm_24.w_0" shape = [32] dtype = "float32" min_val = float("0") @@ -1383,6 +1518,7 @@ class Program_weight_tensor_parameter_134: class Program_weight_tensor_parameter_135: name = "parameter_135" + original_name = "batch_norm_24.w_2" shape = [32] dtype = "float32" min_val = float("0") @@ -1392,6 +1528,7 @@ class Program_weight_tensor_parameter_135: class Program_weight_tensor_parameter_136: name = "parameter_136" + original_name = "batch_norm_24.w_1" shape = [32] dtype = "float32" min_val = float("0") @@ -1401,6 +1538,7 @@ class Program_weight_tensor_parameter_136: class Program_weight_tensor_parameter_137: name = "parameter_137" + original_name = "conv2d_30.w_0" shape = [32, 72, 1, 1] dtype = "float32" min_val = float("-0.88015") @@ -1412,6 +1550,7 @@ class Program_weight_tensor_parameter_137: class Program_weight_tensor_parameter_138: name = "parameter_138" + original_name = "batch_norm_23.b_0" shape = [72] dtype = "float32" min_val = float("-1.89056") @@ -1423,6 +1562,7 @@ class Program_weight_tensor_parameter_138: class Program_weight_tensor_parameter_139: name = "parameter_139" + original_name = "batch_norm_23.w_0" shape = [72] dtype = "float32" min_val = float("0.698852") @@ -1434,6 +1574,7 @@ class Program_weight_tensor_parameter_139: class Program_weight_tensor_parameter_140: name = "parameter_140" + original_name = "batch_norm_23.w_2" shape = [72] dtype = "float32" min_val = float("8.07485e-05") @@ -1445,6 +1586,7 @@ class Program_weight_tensor_parameter_140: class Program_weight_tensor_parameter_141: name = "parameter_141" + original_name = "batch_norm_23.w_1" shape = [72] dtype = "float32" min_val = float("-11.8427") @@ -1456,6 +1598,7 @@ class Program_weight_tensor_parameter_141: class Program_weight_tensor_parameter_142: name = "parameter_142" + original_name = "conv2d_29.w_0" shape = [72, 1, 3, 3] dtype = "float32" min_val = float("-0.937211") @@ -1467,6 +1610,7 @@ class Program_weight_tensor_parameter_142: class Program_weight_tensor_parameter_143: name = "parameter_143" + original_name = "batch_norm_22.b_0" shape = [72] dtype = "float32" min_val = float("-1.92262") @@ -1478,6 +1622,7 @@ class Program_weight_tensor_parameter_143: class Program_weight_tensor_parameter_144: name = "parameter_144" + original_name = "batch_norm_22.w_0" shape = [72] dtype = "float32" min_val = float("-0.18199") @@ -1489,6 +1634,7 @@ class Program_weight_tensor_parameter_144: class Program_weight_tensor_parameter_145: name = "parameter_145" + original_name = "batch_norm_22.w_2" shape = [72] dtype = "float32" min_val = float("13.1559") @@ -1500,6 +1646,7 @@ class Program_weight_tensor_parameter_145: class Program_weight_tensor_parameter_146: name = "parameter_146" + original_name = "batch_norm_22.w_1" shape = [72] dtype = "float32" min_val = float("-6.47258e-06") @@ -1511,6 +1658,7 @@ class Program_weight_tensor_parameter_146: class Program_weight_tensor_parameter_147: name = "parameter_147" + original_name = "conv2d_28.w_0" shape = [72, 32, 1, 1] dtype = "float32" min_val = float("-0.642541") @@ -1522,6 +1670,7 @@ class Program_weight_tensor_parameter_147: class Program_weight_tensor_parameter_148: name = "parameter_148" + original_name = "batch_norm_21.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -1531,6 +1680,7 @@ class Program_weight_tensor_parameter_148: class Program_weight_tensor_parameter_149: name = "parameter_149" + original_name = "batch_norm_21.w_0" shape = [32] dtype = "float32" min_val = float("0") @@ -1540,6 +1690,7 @@ class Program_weight_tensor_parameter_149: class Program_weight_tensor_parameter_150: name = "parameter_150" + original_name = "batch_norm_21.w_2" shape = [32] dtype = "float32" min_val = float("0") @@ -1549,6 +1700,7 @@ class Program_weight_tensor_parameter_150: class Program_weight_tensor_parameter_151: name = "parameter_151" + original_name = "batch_norm_21.w_1" shape = [32] dtype = "float32" min_val = float("0") @@ -1558,6 +1710,7 @@ class Program_weight_tensor_parameter_151: class Program_weight_tensor_parameter_152: name = "parameter_152" + original_name = "conv2d_27.w_0" shape = [32, 88, 1, 1] dtype = "float32" min_val = float("-1.18028") @@ -1569,6 +1722,7 @@ class Program_weight_tensor_parameter_152: class Program_weight_tensor_parameter_153: name = "parameter_153" + original_name = "batch_norm_20.b_0" shape = [88] dtype = "float32" min_val = float("-2.89929") @@ -1580,6 +1734,7 @@ class Program_weight_tensor_parameter_153: class Program_weight_tensor_parameter_154: name = "parameter_154" + original_name = "batch_norm_20.w_0" shape = [88] dtype = "float32" min_val = float("0.452273") @@ -1591,6 +1746,7 @@ class Program_weight_tensor_parameter_154: class Program_weight_tensor_parameter_155: name = "parameter_155" + original_name = "batch_norm_20.w_2" shape = [88] dtype = "float32" min_val = float("0.00126151") @@ -1602,6 +1758,7 @@ class Program_weight_tensor_parameter_155: class Program_weight_tensor_parameter_156: name = "parameter_156" + original_name = "batch_norm_20.w_1" shape = [88] dtype = "float32" min_val = float("-7.17926") @@ -1613,6 +1770,7 @@ class Program_weight_tensor_parameter_156: class Program_weight_tensor_parameter_157: name = "parameter_157" + original_name = "conv2d_26.w_0" shape = [88, 1, 3, 3] dtype = "float32" min_val = float("-0.630557") @@ -1624,6 +1782,7 @@ class Program_weight_tensor_parameter_157: class Program_weight_tensor_parameter_158: name = "parameter_158" + original_name = "batch_norm_19.b_0" shape = [88] dtype = "float32" min_val = float("-2.79304") @@ -1635,6 +1794,7 @@ class Program_weight_tensor_parameter_158: class Program_weight_tensor_parameter_159: name = "parameter_159" + original_name = "batch_norm_19.w_0" shape = [88] dtype = "float32" min_val = float("0.117592") @@ -1646,6 +1806,7 @@ class Program_weight_tensor_parameter_159: class Program_weight_tensor_parameter_160: name = "parameter_160" + original_name = "batch_norm_19.w_2" shape = [88] dtype = "float32" min_val = float("46.9038") @@ -1657,6 +1818,7 @@ class Program_weight_tensor_parameter_160: class Program_weight_tensor_parameter_161: name = "parameter_161" + original_name = "batch_norm_19.w_1" shape = [88] dtype = "float32" min_val = float("-3.28095e-05") @@ -1668,6 +1830,7 @@ class Program_weight_tensor_parameter_161: class Program_weight_tensor_parameter_162: name = "parameter_162" + original_name = "conv2d_25.w_0" shape = [88, 16, 1, 1] dtype = "float32" min_val = float("-1.73101") @@ -1679,6 +1842,7 @@ class Program_weight_tensor_parameter_162: class Program_weight_tensor_parameter_163: name = "parameter_163" + original_name = "batch_norm_18.b_0" shape = [16] dtype = "float32" min_val = float("0") @@ -1688,6 +1852,7 @@ class Program_weight_tensor_parameter_163: class Program_weight_tensor_parameter_164: name = "parameter_164" + original_name = "batch_norm_18.w_0" shape = [16] dtype = "float32" min_val = float("0") @@ -1697,6 +1862,7 @@ class Program_weight_tensor_parameter_164: class Program_weight_tensor_parameter_165: name = "parameter_165" + original_name = "batch_norm_18.w_2" shape = [16] dtype = "float32" min_val = float("0") @@ -1706,6 +1872,7 @@ class Program_weight_tensor_parameter_165: class Program_weight_tensor_parameter_166: name = "parameter_166" + original_name = "batch_norm_18.w_1" shape = [16] dtype = "float32" min_val = float("0") @@ -1715,6 +1882,7 @@ class Program_weight_tensor_parameter_166: class Program_weight_tensor_parameter_167: name = "parameter_167" + original_name = "conv2d_24.w_0" shape = [16, 40, 1, 1] dtype = "float32" min_val = float("-1.08442") @@ -1726,6 +1894,7 @@ class Program_weight_tensor_parameter_167: class Program_weight_tensor_parameter_168: name = "parameter_168" + original_name = "conv2d_23.b_0" shape = [40] dtype = "float32" min_val = float("0") @@ -1735,6 +1904,7 @@ class Program_weight_tensor_parameter_168: class Program_weight_tensor_parameter_169: name = "parameter_169" + original_name = "conv2d_23.w_0" shape = [40, 10, 1, 1] dtype = "float32" min_val = float("-0.786781") @@ -1746,6 +1916,7 @@ class Program_weight_tensor_parameter_169: class Program_weight_tensor_parameter_170: name = "parameter_170" + original_name = "conv2d_22.b_0" shape = [10] dtype = "float32" min_val = float("0") @@ -1755,6 +1926,7 @@ class Program_weight_tensor_parameter_170: class Program_weight_tensor_parameter_171: name = "parameter_171" + original_name = "conv2d_22.w_0" shape = [10, 40, 1, 1] dtype = "float32" min_val = float("-0.733925") @@ -1766,6 +1938,7 @@ class Program_weight_tensor_parameter_171: class Program_weight_tensor_parameter_172: name = "parameter_172" + original_name = "batch_norm_17.b_0" shape = [40] dtype = "float32" min_val = float("0") @@ -1775,6 +1948,7 @@ class Program_weight_tensor_parameter_172: class Program_weight_tensor_parameter_173: name = "parameter_173" + original_name = "batch_norm_17.w_0" shape = [40] dtype = "float32" min_val = float("0") @@ -1784,6 +1958,7 @@ class Program_weight_tensor_parameter_173: class Program_weight_tensor_parameter_174: name = "parameter_174" + original_name = "batch_norm_17.w_2" shape = [40] dtype = "float32" min_val = float("0") @@ -1793,6 +1968,7 @@ class Program_weight_tensor_parameter_174: class Program_weight_tensor_parameter_175: name = "parameter_175" + original_name = "batch_norm_17.w_1" shape = [40] dtype = "float32" min_val = float("0") @@ -1802,6 +1978,7 @@ class Program_weight_tensor_parameter_175: class Program_weight_tensor_parameter_176: name = "parameter_176" + original_name = "conv2d_21.w_0" shape = [40, 1, 5, 5] dtype = "float32" min_val = float("-1.35371") @@ -1813,6 +1990,7 @@ class Program_weight_tensor_parameter_176: class Program_weight_tensor_parameter_177: name = "parameter_177" + original_name = "batch_norm_16.b_0" shape = [40] dtype = "float32" min_val = float("0") @@ -1822,6 +2000,7 @@ class Program_weight_tensor_parameter_177: class Program_weight_tensor_parameter_178: name = "parameter_178" + original_name = "batch_norm_16.w_0" shape = [40] dtype = "float32" min_val = float("0") @@ -1831,6 +2010,7 @@ class Program_weight_tensor_parameter_178: class Program_weight_tensor_parameter_179: name = "parameter_179" + original_name = "batch_norm_16.w_2" shape = [40] dtype = "float32" min_val = float("0") @@ -1840,6 +2020,7 @@ class Program_weight_tensor_parameter_179: class Program_weight_tensor_parameter_180: name = "parameter_180" + original_name = "batch_norm_16.w_1" shape = [40] dtype = "float32" min_val = float("0") @@ -1849,6 +2030,7 @@ class Program_weight_tensor_parameter_180: class Program_weight_tensor_parameter_181: name = "parameter_181" + original_name = "conv2d_20.w_0" shape = [40, 16, 1, 1] dtype = "float32" min_val = float("-0.832364") @@ -1860,6 +2042,7 @@ class Program_weight_tensor_parameter_181: class Program_weight_tensor_parameter_182: name = "parameter_182" + original_name = "batch_norm_15.b_0" shape = [16] dtype = "float32" min_val = float("0") @@ -1869,6 +2052,7 @@ class Program_weight_tensor_parameter_182: class Program_weight_tensor_parameter_183: name = "parameter_183" + original_name = "batch_norm_15.w_0" shape = [16] dtype = "float32" min_val = float("0") @@ -1878,6 +2062,7 @@ class Program_weight_tensor_parameter_183: class Program_weight_tensor_parameter_184: name = "parameter_184" + original_name = "batch_norm_15.w_2" shape = [16] dtype = "float32" min_val = float("0") @@ -1887,6 +2072,7 @@ class Program_weight_tensor_parameter_184: class Program_weight_tensor_parameter_185: name = "parameter_185" + original_name = "batch_norm_15.w_1" shape = [16] dtype = "float32" min_val = float("0") @@ -1896,6 +2082,7 @@ class Program_weight_tensor_parameter_185: class Program_weight_tensor_parameter_186: name = "parameter_186" + original_name = "conv2d_19.w_0" shape = [16, 40, 1, 1] dtype = "float32" min_val = float("-0.90344") @@ -1907,6 +2094,7 @@ class Program_weight_tensor_parameter_186: class Program_weight_tensor_parameter_187: name = "parameter_187" + original_name = "conv2d_18.b_0" shape = [40] dtype = "float32" min_val = float("0") @@ -1916,6 +2104,7 @@ class Program_weight_tensor_parameter_187: class Program_weight_tensor_parameter_188: name = "parameter_188" + original_name = "conv2d_18.w_0" shape = [40, 10, 1, 1] dtype = "float32" min_val = float("-0.798656") @@ -1927,6 +2116,7 @@ class Program_weight_tensor_parameter_188: class Program_weight_tensor_parameter_189: name = "parameter_189" + original_name = "conv2d_17.b_0" shape = [10] dtype = "float32" min_val = float("0") @@ -1936,6 +2126,7 @@ class Program_weight_tensor_parameter_189: class Program_weight_tensor_parameter_190: name = "parameter_190" + original_name = "conv2d_17.w_0" shape = [10, 40, 1, 1] dtype = "float32" min_val = float("-0.621303") @@ -1947,6 +2138,7 @@ class Program_weight_tensor_parameter_190: class Program_weight_tensor_parameter_191: name = "parameter_191" + original_name = "batch_norm_14.b_0" shape = [40] dtype = "float32" min_val = float("0") @@ -1956,6 +2148,7 @@ class Program_weight_tensor_parameter_191: class Program_weight_tensor_parameter_192: name = "parameter_192" + original_name = "batch_norm_14.w_0" shape = [40] dtype = "float32" min_val = float("0") @@ -1965,6 +2158,7 @@ class Program_weight_tensor_parameter_192: class Program_weight_tensor_parameter_193: name = "parameter_193" + original_name = "batch_norm_14.w_2" shape = [40] dtype = "float32" min_val = float("0") @@ -1974,6 +2168,7 @@ class Program_weight_tensor_parameter_193: class Program_weight_tensor_parameter_194: name = "parameter_194" + original_name = "batch_norm_14.w_1" shape = [40] dtype = "float32" min_val = float("0") @@ -1983,6 +2178,7 @@ class Program_weight_tensor_parameter_194: class Program_weight_tensor_parameter_195: name = "parameter_195" + original_name = "conv2d_16.w_0" shape = [40, 1, 5, 5] dtype = "float32" min_val = float("-1.39292") @@ -1994,6 +2190,7 @@ class Program_weight_tensor_parameter_195: class Program_weight_tensor_parameter_196: name = "parameter_196" + original_name = "batch_norm_13.b_0" shape = [40] dtype = "float32" min_val = float("0") @@ -2003,6 +2200,7 @@ class Program_weight_tensor_parameter_196: class Program_weight_tensor_parameter_197: name = "parameter_197" + original_name = "batch_norm_13.w_0" shape = [40] dtype = "float32" min_val = float("0") @@ -2012,6 +2210,7 @@ class Program_weight_tensor_parameter_197: class Program_weight_tensor_parameter_198: name = "parameter_198" + original_name = "batch_norm_13.w_2" shape = [40] dtype = "float32" min_val = float("0") @@ -2021,6 +2220,7 @@ class Program_weight_tensor_parameter_198: class Program_weight_tensor_parameter_199: name = "parameter_199" + original_name = "batch_norm_13.w_1" shape = [40] dtype = "float32" min_val = float("0") @@ -2030,6 +2230,7 @@ class Program_weight_tensor_parameter_199: class Program_weight_tensor_parameter_200: name = "parameter_200" + original_name = "conv2d_15.w_0" shape = [40, 16, 1, 1] dtype = "float32" min_val = float("-1.04639") @@ -2041,6 +2242,7 @@ class Program_weight_tensor_parameter_200: class Program_weight_tensor_parameter_201: name = "parameter_201" + original_name = "batch_norm_12.b_0" shape = [16] dtype = "float32" min_val = float("0") @@ -2050,6 +2252,7 @@ class Program_weight_tensor_parameter_201: class Program_weight_tensor_parameter_202: name = "parameter_202" + original_name = "batch_norm_12.w_0" shape = [16] dtype = "float32" min_val = float("0") @@ -2059,6 +2262,7 @@ class Program_weight_tensor_parameter_202: class Program_weight_tensor_parameter_203: name = "parameter_203" + original_name = "batch_norm_12.w_2" shape = [16] dtype = "float32" min_val = float("0") @@ -2068,6 +2272,7 @@ class Program_weight_tensor_parameter_203: class Program_weight_tensor_parameter_204: name = "parameter_204" + original_name = "batch_norm_12.w_1" shape = [16] dtype = "float32" min_val = float("0") @@ -2077,6 +2282,7 @@ class Program_weight_tensor_parameter_204: class Program_weight_tensor_parameter_205: name = "parameter_205" + original_name = "conv2d_14.w_0" shape = [16, 24, 1, 1] dtype = "float32" min_val = float("-1.67548") @@ -2088,6 +2294,7 @@ class Program_weight_tensor_parameter_205: class Program_weight_tensor_parameter_206: name = "parameter_206" + original_name = "conv2d_13.b_0" shape = [24] dtype = "float32" min_val = float("0") @@ -2097,6 +2304,7 @@ class Program_weight_tensor_parameter_206: class Program_weight_tensor_parameter_207: name = "parameter_207" + original_name = "conv2d_13.w_0" shape = [24, 6, 1, 1] dtype = "float32" min_val = float("-1.56296e-13") @@ -2108,6 +2316,7 @@ class Program_weight_tensor_parameter_207: class Program_weight_tensor_parameter_208: name = "parameter_208" + original_name = "conv2d_12.b_0" shape = [6] dtype = "float32" min_val = float("0") @@ -2117,6 +2326,7 @@ class Program_weight_tensor_parameter_208: class Program_weight_tensor_parameter_209: name = "parameter_209" + original_name = "conv2d_12.w_0" shape = [6, 24, 1, 1] dtype = "float32" min_val = float("-1.89298e-13") @@ -2128,6 +2338,7 @@ class Program_weight_tensor_parameter_209: class Program_weight_tensor_parameter_210: name = "parameter_210" + original_name = "batch_norm_11.b_0" shape = [24] dtype = "float32" min_val = float("0") @@ -2137,6 +2348,7 @@ class Program_weight_tensor_parameter_210: class Program_weight_tensor_parameter_211: name = "parameter_211" + original_name = "batch_norm_11.w_0" shape = [24] dtype = "float32" min_val = float("0") @@ -2146,6 +2358,7 @@ class Program_weight_tensor_parameter_211: class Program_weight_tensor_parameter_212: name = "parameter_212" + original_name = "batch_norm_11.w_2" shape = [24] dtype = "float32" min_val = float("0") @@ -2155,6 +2368,7 @@ class Program_weight_tensor_parameter_212: class Program_weight_tensor_parameter_213: name = "parameter_213" + original_name = "batch_norm_11.w_1" shape = [24] dtype = "float32" min_val = float("0") @@ -2164,6 +2378,7 @@ class Program_weight_tensor_parameter_213: class Program_weight_tensor_parameter_214: name = "parameter_214" + original_name = "conv2d_11.w_0" shape = [24, 1, 5, 5] dtype = "float32" min_val = float("-1.1626") @@ -2175,6 +2390,7 @@ class Program_weight_tensor_parameter_214: class Program_weight_tensor_parameter_215: name = "parameter_215" + original_name = "batch_norm_10.b_0" shape = [24] dtype = "float32" min_val = float("0") @@ -2184,6 +2400,7 @@ class Program_weight_tensor_parameter_215: class Program_weight_tensor_parameter_216: name = "parameter_216" + original_name = "batch_norm_10.w_0" shape = [24] dtype = "float32" min_val = float("0") @@ -2193,6 +2410,7 @@ class Program_weight_tensor_parameter_216: class Program_weight_tensor_parameter_217: name = "parameter_217" + original_name = "batch_norm_10.w_2" shape = [24] dtype = "float32" min_val = float("0") @@ -2202,6 +2420,7 @@ class Program_weight_tensor_parameter_217: class Program_weight_tensor_parameter_218: name = "parameter_218" + original_name = "batch_norm_10.w_1" shape = [24] dtype = "float32" min_val = float("0") @@ -2211,6 +2430,7 @@ class Program_weight_tensor_parameter_218: class Program_weight_tensor_parameter_219: name = "parameter_219" + original_name = "conv2d_10.w_0" shape = [24, 8, 1, 1] dtype = "float32" min_val = float("-2.17072") @@ -2222,6 +2442,7 @@ class Program_weight_tensor_parameter_219: class Program_weight_tensor_parameter_220: name = "parameter_220" + original_name = "batch_norm_9.b_0" shape = [8] dtype = "float32" min_val = float("0") @@ -2231,6 +2452,7 @@ class Program_weight_tensor_parameter_220: class Program_weight_tensor_parameter_221: name = "parameter_221" + original_name = "batch_norm_9.w_0" shape = [8] dtype = "float32" min_val = float("0") @@ -2240,6 +2462,7 @@ class Program_weight_tensor_parameter_221: class Program_weight_tensor_parameter_222: name = "parameter_222" + original_name = "batch_norm_9.w_2" shape = [8] dtype = "float32" min_val = float("0") @@ -2249,6 +2472,7 @@ class Program_weight_tensor_parameter_222: class Program_weight_tensor_parameter_223: name = "parameter_223" + original_name = "batch_norm_9.w_1" shape = [8] dtype = "float32" min_val = float("0") @@ -2258,6 +2482,7 @@ class Program_weight_tensor_parameter_223: class Program_weight_tensor_parameter_224: name = "parameter_224" + original_name = "conv2d_9.w_0" shape = [8, 24, 1, 1] dtype = "float32" min_val = float("-1.33004") @@ -2269,6 +2494,7 @@ class Program_weight_tensor_parameter_224: class Program_weight_tensor_parameter_225: name = "parameter_225" + original_name = "batch_norm_8.b_0" shape = [24] dtype = "float32" min_val = float("0") @@ -2278,6 +2504,7 @@ class Program_weight_tensor_parameter_225: class Program_weight_tensor_parameter_226: name = "parameter_226" + original_name = "batch_norm_8.w_0" shape = [24] dtype = "float32" min_val = float("0") @@ -2287,6 +2514,7 @@ class Program_weight_tensor_parameter_226: class Program_weight_tensor_parameter_227: name = "parameter_227" + original_name = "batch_norm_8.w_2" shape = [24] dtype = "float32" min_val = float("0") @@ -2296,6 +2524,7 @@ class Program_weight_tensor_parameter_227: class Program_weight_tensor_parameter_228: name = "parameter_228" + original_name = "batch_norm_8.w_1" shape = [24] dtype = "float32" min_val = float("0") @@ -2305,6 +2534,7 @@ class Program_weight_tensor_parameter_228: class Program_weight_tensor_parameter_229: name = "parameter_229" + original_name = "conv2d_8.w_0" shape = [24, 1, 3, 3] dtype = "float32" min_val = float("-1.76237") @@ -2316,6 +2546,7 @@ class Program_weight_tensor_parameter_229: class Program_weight_tensor_parameter_230: name = "parameter_230" + original_name = "batch_norm_7.b_0" shape = [24] dtype = "float32" min_val = float("0") @@ -2325,6 +2556,7 @@ class Program_weight_tensor_parameter_230: class Program_weight_tensor_parameter_231: name = "parameter_231" + original_name = "batch_norm_7.w_0" shape = [24] dtype = "float32" min_val = float("0") @@ -2334,6 +2566,7 @@ class Program_weight_tensor_parameter_231: class Program_weight_tensor_parameter_232: name = "parameter_232" + original_name = "batch_norm_7.w_2" shape = [24] dtype = "float32" min_val = float("0") @@ -2343,6 +2576,7 @@ class Program_weight_tensor_parameter_232: class Program_weight_tensor_parameter_233: name = "parameter_233" + original_name = "batch_norm_7.w_1" shape = [24] dtype = "float32" min_val = float("0") @@ -2352,6 +2586,7 @@ class Program_weight_tensor_parameter_233: class Program_weight_tensor_parameter_234: name = "parameter_234" + original_name = "conv2d_7.w_0" shape = [24, 8, 1, 1] dtype = "float32" min_val = float("-1.46358") @@ -2363,6 +2598,7 @@ class Program_weight_tensor_parameter_234: class Program_weight_tensor_parameter_235: name = "parameter_235" + original_name = "batch_norm_6.b_0" shape = [8] dtype = "float32" min_val = float("0") @@ -2372,6 +2608,7 @@ class Program_weight_tensor_parameter_235: class Program_weight_tensor_parameter_236: name = "parameter_236" + original_name = "batch_norm_6.w_0" shape = [8] dtype = "float32" min_val = float("0") @@ -2381,6 +2618,7 @@ class Program_weight_tensor_parameter_236: class Program_weight_tensor_parameter_237: name = "parameter_237" + original_name = "batch_norm_6.w_2" shape = [8] dtype = "float32" min_val = float("0") @@ -2390,6 +2628,7 @@ class Program_weight_tensor_parameter_237: class Program_weight_tensor_parameter_238: name = "parameter_238" + original_name = "batch_norm_6.w_1" shape = [8] dtype = "float32" min_val = float("0") @@ -2399,6 +2638,7 @@ class Program_weight_tensor_parameter_238: class Program_weight_tensor_parameter_239: name = "parameter_239" + original_name = "conv2d_6.w_0" shape = [8, 24, 1, 1] dtype = "float32" min_val = float("-2.01604") @@ -2410,6 +2650,7 @@ class Program_weight_tensor_parameter_239: class Program_weight_tensor_parameter_240: name = "parameter_240" + original_name = "batch_norm_5.b_0" shape = [24] dtype = "float32" min_val = float("0") @@ -2419,6 +2660,7 @@ class Program_weight_tensor_parameter_240: class Program_weight_tensor_parameter_241: name = "parameter_241" + original_name = "batch_norm_5.w_0" shape = [24] dtype = "float32" min_val = float("0") @@ -2428,6 +2670,7 @@ class Program_weight_tensor_parameter_241: class Program_weight_tensor_parameter_242: name = "parameter_242" + original_name = "batch_norm_5.w_2" shape = [24] dtype = "float32" min_val = float("0") @@ -2437,6 +2680,7 @@ class Program_weight_tensor_parameter_242: class Program_weight_tensor_parameter_243: name = "parameter_243" + original_name = "batch_norm_5.w_1" shape = [24] dtype = "float32" min_val = float("0") @@ -2446,6 +2690,7 @@ class Program_weight_tensor_parameter_243: class Program_weight_tensor_parameter_244: name = "parameter_244" + original_name = "conv2d_5.w_0" shape = [24, 1, 3, 3] dtype = "float32" min_val = float("-1.01716") @@ -2457,6 +2702,7 @@ class Program_weight_tensor_parameter_244: class Program_weight_tensor_parameter_245: name = "parameter_245" + original_name = "batch_norm_4.b_0" shape = [24] dtype = "float32" min_val = float("0") @@ -2466,6 +2712,7 @@ class Program_weight_tensor_parameter_245: class Program_weight_tensor_parameter_246: name = "parameter_246" + original_name = "batch_norm_4.w_0" shape = [24] dtype = "float32" min_val = float("0") @@ -2475,6 +2722,7 @@ class Program_weight_tensor_parameter_246: class Program_weight_tensor_parameter_247: name = "parameter_247" + original_name = "batch_norm_4.w_2" shape = [24] dtype = "float32" min_val = float("0") @@ -2484,6 +2732,7 @@ class Program_weight_tensor_parameter_247: class Program_weight_tensor_parameter_248: name = "parameter_248" + original_name = "batch_norm_4.w_1" shape = [24] dtype = "float32" min_val = float("0") @@ -2493,6 +2742,7 @@ class Program_weight_tensor_parameter_248: class Program_weight_tensor_parameter_249: name = "parameter_249" + original_name = "conv2d_4.w_0" shape = [24, 8, 1, 1] dtype = "float32" min_val = float("-2.11845") @@ -2504,6 +2754,7 @@ class Program_weight_tensor_parameter_249: class Program_weight_tensor_parameter_250: name = "parameter_250" + original_name = "batch_norm_3.b_0" shape = [8] dtype = "float32" min_val = float("0") @@ -2513,6 +2764,7 @@ class Program_weight_tensor_parameter_250: class Program_weight_tensor_parameter_251: name = "parameter_251" + original_name = "batch_norm_3.w_0" shape = [8] dtype = "float32" min_val = float("0") @@ -2522,6 +2774,7 @@ class Program_weight_tensor_parameter_251: class Program_weight_tensor_parameter_252: name = "parameter_252" + original_name = "batch_norm_3.w_2" shape = [8] dtype = "float32" min_val = float("0") @@ -2531,6 +2784,7 @@ class Program_weight_tensor_parameter_252: class Program_weight_tensor_parameter_253: name = "parameter_253" + original_name = "batch_norm_3.w_1" shape = [8] dtype = "float32" min_val = float("0") @@ -2540,6 +2794,7 @@ class Program_weight_tensor_parameter_253: class Program_weight_tensor_parameter_254: name = "parameter_254" + original_name = "conv2d_3.w_0" shape = [8, 8, 1, 1] dtype = "float32" min_val = float("0") @@ -2549,6 +2804,7 @@ class Program_weight_tensor_parameter_254: class Program_weight_tensor_parameter_255: name = "parameter_255" + original_name = "batch_norm_2.b_0" shape = [8] dtype = "float32" min_val = float("0") @@ -2558,6 +2814,7 @@ class Program_weight_tensor_parameter_255: class Program_weight_tensor_parameter_256: name = "parameter_256" + original_name = "batch_norm_2.w_0" shape = [8] dtype = "float32" min_val = float("0") @@ -2567,6 +2824,7 @@ class Program_weight_tensor_parameter_256: class Program_weight_tensor_parameter_257: name = "parameter_257" + original_name = "batch_norm_2.w_2" shape = [8] dtype = "float32" min_val = float("0") @@ -2576,6 +2834,7 @@ class Program_weight_tensor_parameter_257: class Program_weight_tensor_parameter_258: name = "parameter_258" + original_name = "batch_norm_2.w_1" shape = [8] dtype = "float32" min_val = float("0") @@ -2585,6 +2844,7 @@ class Program_weight_tensor_parameter_258: class Program_weight_tensor_parameter_259: name = "parameter_259" + original_name = "conv2d_2.w_0" shape = [8, 1, 3, 3] dtype = "float32" min_val = float("-1.61055") @@ -2596,6 +2856,7 @@ class Program_weight_tensor_parameter_259: class Program_weight_tensor_parameter_260: name = "parameter_260" + original_name = "batch_norm_1.b_0" shape = [8] dtype = "float32" min_val = float("0") @@ -2605,6 +2866,7 @@ class Program_weight_tensor_parameter_260: class Program_weight_tensor_parameter_261: name = "parameter_261" + original_name = "batch_norm_1.w_0" shape = [8] dtype = "float32" min_val = float("0") @@ -2614,6 +2876,7 @@ class Program_weight_tensor_parameter_261: class Program_weight_tensor_parameter_262: name = "parameter_262" + original_name = "batch_norm_1.w_2" shape = [8] dtype = "float32" min_val = float("0") @@ -2623,6 +2886,7 @@ class Program_weight_tensor_parameter_262: class Program_weight_tensor_parameter_263: name = "parameter_263" + original_name = "batch_norm_1.w_1" shape = [8] dtype = "float32" min_val = float("0") @@ -2632,6 +2896,7 @@ class Program_weight_tensor_parameter_263: class Program_weight_tensor_parameter_264: name = "parameter_264" + original_name = "conv2d_1.w_0" shape = [8, 8, 1, 1] dtype = "float32" min_val = float("0") @@ -2641,6 +2906,7 @@ class Program_weight_tensor_parameter_264: class Program_weight_tensor_parameter_265: name = "parameter_265" + original_name = "batch_norm_0.b_0" shape = [8] dtype = "float32" min_val = float("0") @@ -2650,6 +2916,7 @@ class Program_weight_tensor_parameter_265: class Program_weight_tensor_parameter_266: name = "parameter_266" + original_name = "batch_norm_0.w_0" shape = [8] dtype = "float32" min_val = float("0") @@ -2659,6 +2926,7 @@ class Program_weight_tensor_parameter_266: class Program_weight_tensor_parameter_267: name = "parameter_267" + original_name = "batch_norm_0.w_2" shape = [8] dtype = "float32" min_val = float("0") @@ -2668,6 +2936,7 @@ class Program_weight_tensor_parameter_267: class Program_weight_tensor_parameter_268: name = "parameter_268" + original_name = "batch_norm_0.w_1" shape = [8] dtype = "float32" min_val = float("0") @@ -2677,6 +2946,7 @@ class Program_weight_tensor_parameter_268: class Program_weight_tensor_parameter_269: name = "parameter_269" + original_name = "conv2d_0.w_0" shape = [8, 3, 3, 3] dtype = "float32" min_val = float("-2.99168") diff --git a/paddle_samples/PaddleX/MobileNetV3_large_x0_5/subgraph_0/input_meta.py b/paddle_samples/PaddleX/MobileNetV3_large_x0_5/subgraph_0/input_meta.py index 795f9bcfe..6d7dac1c9 100644 --- a/paddle_samples/PaddleX/MobileNetV3_large_x0_5/subgraph_0/input_meta.py +++ b/paddle_samples/PaddleX/MobileNetV3_large_x0_5/subgraph_0/input_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_data_0: name = "data_0" + original_name = "var_0" shape = [128, 3, 224, 224] dtype = "float32" min_val = float("-2.1179") diff --git a/paddle_samples/PaddleX/MobileNetV3_large_x0_5/subgraph_0/weight_meta.py b/paddle_samples/PaddleX/MobileNetV3_large_x0_5/subgraph_0/weight_meta.py index 30114f1dc..ccacc1a97 100644 --- a/paddle_samples/PaddleX/MobileNetV3_large_x0_5/subgraph_0/weight_meta.py +++ b/paddle_samples/PaddleX/MobileNetV3_large_x0_5/subgraph_0/weight_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_parameter_0: name = "parameter_0" + original_name = "linear_0.b_0" shape = [102] dtype = "float32" data = None @@ -7,6 +8,7 @@ class Program_weight_tensor_parameter_0: class Program_weight_tensor_parameter_1: name = "parameter_1" + original_name = "linear_0.w_0" shape = [1280, 102] dtype = "float32" min_val = float("-0.0658888") @@ -18,6 +20,7 @@ class Program_weight_tensor_parameter_1: class Program_weight_tensor_parameter_2: name = "parameter_2" + original_name = "conv2d_63.w_0" shape = [1280, 480, 1, 1] dtype = "float32" min_val = float("-0.580823") @@ -29,6 +32,7 @@ class Program_weight_tensor_parameter_2: class Program_weight_tensor_parameter_3: name = "parameter_3" + original_name = "batch_norm_46.b_0" shape = [480] dtype = "float32" min_val = float("-10.3034") @@ -40,6 +44,7 @@ class Program_weight_tensor_parameter_3: class Program_weight_tensor_parameter_4: name = "parameter_4" + original_name = "batch_norm_46.w_0" shape = [480] dtype = "float32" min_val = float("2.52624") @@ -51,6 +56,7 @@ class Program_weight_tensor_parameter_4: class Program_weight_tensor_parameter_5: name = "parameter_5" + original_name = "batch_norm_46.w_2" shape = [480] dtype = "float32" min_val = float("34.0685") @@ -62,6 +68,7 @@ class Program_weight_tensor_parameter_5: class Program_weight_tensor_parameter_6: name = "parameter_6" + original_name = "batch_norm_46.w_1" shape = [480] dtype = "float32" min_val = float("-1.62218e-05") @@ -73,6 +80,7 @@ class Program_weight_tensor_parameter_6: class Program_weight_tensor_parameter_7: name = "parameter_7" + original_name = "conv2d_62.w_0" shape = [480, 80, 1, 1] dtype = "float32" min_val = float("-1.58227") @@ -84,6 +92,7 @@ class Program_weight_tensor_parameter_7: class Program_weight_tensor_parameter_8: name = "parameter_8" + original_name = "batch_norm_45.b_0" shape = [80] dtype = "float32" min_val = float("-2.48069e-06") @@ -95,6 +104,7 @@ class Program_weight_tensor_parameter_8: class Program_weight_tensor_parameter_9: name = "parameter_9" + original_name = "batch_norm_45.w_0" shape = [80] dtype = "float32" min_val = float("1.66187") @@ -106,6 +116,7 @@ class Program_weight_tensor_parameter_9: class Program_weight_tensor_parameter_10: name = "parameter_10" + original_name = "batch_norm_45.w_2" shape = [80] dtype = "float32" min_val = float("0.0531797") @@ -117,6 +128,7 @@ class Program_weight_tensor_parameter_10: class Program_weight_tensor_parameter_11: name = "parameter_11" + original_name = "batch_norm_45.w_1" shape = [80] dtype = "float32" min_val = float("-0.359942") @@ -128,6 +140,7 @@ class Program_weight_tensor_parameter_11: class Program_weight_tensor_parameter_12: name = "parameter_12" + original_name = "conv2d_61.w_0" shape = [80, 480, 1, 1] dtype = "float32" min_val = float("-0.627034") @@ -139,6 +152,7 @@ class Program_weight_tensor_parameter_12: class Program_weight_tensor_parameter_13: name = "parameter_13" + original_name = "conv2d_60.b_0" shape = [480] dtype = "float32" min_val = float("-0.912623") @@ -150,6 +164,7 @@ class Program_weight_tensor_parameter_13: class Program_weight_tensor_parameter_14: name = "parameter_14" + original_name = "conv2d_60.w_0" shape = [480, 120, 1, 1] dtype = "float32" min_val = float("-0.727886") @@ -161,6 +176,7 @@ class Program_weight_tensor_parameter_14: class Program_weight_tensor_parameter_15: name = "parameter_15" + original_name = "conv2d_59.b_0" shape = [120] dtype = "float32" min_val = float("-0.115518") @@ -172,6 +188,7 @@ class Program_weight_tensor_parameter_15: class Program_weight_tensor_parameter_16: name = "parameter_16" + original_name = "conv2d_59.w_0" shape = [120, 480, 1, 1] dtype = "float32" min_val = float("-0.644708") @@ -183,6 +200,7 @@ class Program_weight_tensor_parameter_16: class Program_weight_tensor_parameter_17: name = "parameter_17" + original_name = "batch_norm_44.b_0" shape = [480] dtype = "float32" min_val = float("-5.9007") @@ -194,6 +212,7 @@ class Program_weight_tensor_parameter_17: class Program_weight_tensor_parameter_18: name = "parameter_18" + original_name = "batch_norm_44.w_0" shape = [480] dtype = "float32" min_val = float("0.162836") @@ -205,6 +224,7 @@ class Program_weight_tensor_parameter_18: class Program_weight_tensor_parameter_19: name = "parameter_19" + original_name = "batch_norm_44.w_2" shape = [480] dtype = "float32" min_val = float("9.24052e-05") @@ -216,6 +236,7 @@ class Program_weight_tensor_parameter_19: class Program_weight_tensor_parameter_20: name = "parameter_20" + original_name = "batch_norm_44.w_1" shape = [480] dtype = "float32" min_val = float("-2.99003") @@ -227,6 +248,7 @@ class Program_weight_tensor_parameter_20: class Program_weight_tensor_parameter_21: name = "parameter_21" + original_name = "conv2d_58.w_0" shape = [480, 1, 5, 5] dtype = "float32" min_val = float("-0.702944") @@ -238,6 +260,7 @@ class Program_weight_tensor_parameter_21: class Program_weight_tensor_parameter_22: name = "parameter_22" + original_name = "batch_norm_43.b_0" shape = [480] dtype = "float32" min_val = float("-3.57336") @@ -249,6 +272,7 @@ class Program_weight_tensor_parameter_22: class Program_weight_tensor_parameter_23: name = "parameter_23" + original_name = "batch_norm_43.w_0" shape = [480] dtype = "float32" min_val = float("-1.36216") @@ -260,6 +284,7 @@ class Program_weight_tensor_parameter_23: class Program_weight_tensor_parameter_24: name = "parameter_24" + original_name = "batch_norm_43.w_2" shape = [480] dtype = "float32" min_val = float("6.53337") @@ -271,6 +296,7 @@ class Program_weight_tensor_parameter_24: class Program_weight_tensor_parameter_25: name = "parameter_25" + original_name = "batch_norm_43.w_1" shape = [480] dtype = "float32" min_val = float("-8.06679e-06") @@ -282,6 +308,7 @@ class Program_weight_tensor_parameter_25: class Program_weight_tensor_parameter_26: name = "parameter_26" + original_name = "conv2d_57.w_0" shape = [480, 80, 1, 1] dtype = "float32" min_val = float("-0.574413") @@ -293,6 +320,7 @@ class Program_weight_tensor_parameter_26: class Program_weight_tensor_parameter_27: name = "parameter_27" + original_name = "batch_norm_42.b_0" shape = [80] dtype = "float32" min_val = float("-2.50957e-06") @@ -304,6 +332,7 @@ class Program_weight_tensor_parameter_27: class Program_weight_tensor_parameter_28: name = "parameter_28" + original_name = "batch_norm_42.w_0" shape = [80] dtype = "float32" min_val = float("0.864096") @@ -315,6 +344,7 @@ class Program_weight_tensor_parameter_28: class Program_weight_tensor_parameter_29: name = "parameter_29" + original_name = "batch_norm_42.w_2" shape = [80] dtype = "float32" min_val = float("0.0540053") @@ -326,6 +356,7 @@ class Program_weight_tensor_parameter_29: class Program_weight_tensor_parameter_30: name = "parameter_30" + original_name = "batch_norm_42.w_1" shape = [80] dtype = "float32" min_val = float("-0.518445") @@ -337,6 +368,7 @@ class Program_weight_tensor_parameter_30: class Program_weight_tensor_parameter_31: name = "parameter_31" + original_name = "conv2d_56.w_0" shape = [80, 480, 1, 1] dtype = "float32" min_val = float("-0.547161") @@ -348,6 +380,7 @@ class Program_weight_tensor_parameter_31: class Program_weight_tensor_parameter_32: name = "parameter_32" + original_name = "conv2d_55.b_0" shape = [480] dtype = "float32" min_val = float("-0.350915") @@ -359,6 +392,7 @@ class Program_weight_tensor_parameter_32: class Program_weight_tensor_parameter_33: name = "parameter_33" + original_name = "conv2d_55.w_0" shape = [480, 120, 1, 1] dtype = "float32" min_val = float("-0.43498") @@ -370,6 +404,7 @@ class Program_weight_tensor_parameter_33: class Program_weight_tensor_parameter_34: name = "parameter_34" + original_name = "conv2d_54.b_0" shape = [120] dtype = "float32" min_val = float("-0.0975236") @@ -381,6 +416,7 @@ class Program_weight_tensor_parameter_34: class Program_weight_tensor_parameter_35: name = "parameter_35" + original_name = "conv2d_54.w_0" shape = [120, 480, 1, 1] dtype = "float32" min_val = float("-0.810099") @@ -392,6 +428,7 @@ class Program_weight_tensor_parameter_35: class Program_weight_tensor_parameter_36: name = "parameter_36" + original_name = "batch_norm_41.b_0" shape = [480] dtype = "float32" min_val = float("-3.37143") @@ -403,6 +440,7 @@ class Program_weight_tensor_parameter_36: class Program_weight_tensor_parameter_37: name = "parameter_37" + original_name = "batch_norm_41.w_0" shape = [480] dtype = "float32" min_val = float("0.269123") @@ -414,6 +452,7 @@ class Program_weight_tensor_parameter_37: class Program_weight_tensor_parameter_38: name = "parameter_38" + original_name = "batch_norm_41.w_2" shape = [480] dtype = "float32" min_val = float("0.000105688") @@ -425,6 +464,7 @@ class Program_weight_tensor_parameter_38: class Program_weight_tensor_parameter_39: name = "parameter_39" + original_name = "batch_norm_41.w_1" shape = [480] dtype = "float32" min_val = float("-6.99758") @@ -436,6 +476,7 @@ class Program_weight_tensor_parameter_39: class Program_weight_tensor_parameter_40: name = "parameter_40" + original_name = "conv2d_53.w_0" shape = [480, 1, 5, 5] dtype = "float32" min_val = float("-0.824259") @@ -447,6 +488,7 @@ class Program_weight_tensor_parameter_40: class Program_weight_tensor_parameter_41: name = "parameter_41" + original_name = "batch_norm_40.b_0" shape = [480] dtype = "float32" min_val = float("-3.88159") @@ -458,6 +500,7 @@ class Program_weight_tensor_parameter_41: class Program_weight_tensor_parameter_42: name = "parameter_42" + original_name = "batch_norm_40.w_0" shape = [480] dtype = "float32" min_val = float("-0.212371") @@ -469,6 +512,7 @@ class Program_weight_tensor_parameter_42: class Program_weight_tensor_parameter_43: name = "parameter_43" + original_name = "batch_norm_40.w_2" shape = [480] dtype = "float32" min_val = float("6.39243") @@ -480,6 +524,7 @@ class Program_weight_tensor_parameter_43: class Program_weight_tensor_parameter_44: name = "parameter_44" + original_name = "batch_norm_40.w_1" shape = [480] dtype = "float32" min_val = float("-3.91635e-06") @@ -491,6 +536,7 @@ class Program_weight_tensor_parameter_44: class Program_weight_tensor_parameter_45: name = "parameter_45" + original_name = "conv2d_52.w_0" shape = [480, 80, 1, 1] dtype = "float32" min_val = float("-0.764259") @@ -502,6 +548,7 @@ class Program_weight_tensor_parameter_45: class Program_weight_tensor_parameter_46: name = "parameter_46" + original_name = "batch_norm_39.b_0" shape = [80] dtype = "float32" min_val = float("-3.77978e-06") @@ -513,6 +560,7 @@ class Program_weight_tensor_parameter_46: class Program_weight_tensor_parameter_47: name = "parameter_47" + original_name = "batch_norm_39.w_0" shape = [80] dtype = "float32" min_val = float("2.79365") @@ -524,6 +572,7 @@ class Program_weight_tensor_parameter_47: class Program_weight_tensor_parameter_48: name = "parameter_48" + original_name = "batch_norm_39.w_2" shape = [80] dtype = "float32" min_val = float("0.0867921") @@ -535,6 +584,7 @@ class Program_weight_tensor_parameter_48: class Program_weight_tensor_parameter_49: name = "parameter_49" + original_name = "batch_norm_39.w_1" shape = [80] dtype = "float32" min_val = float("-0.655323") @@ -546,6 +596,7 @@ class Program_weight_tensor_parameter_49: class Program_weight_tensor_parameter_50: name = "parameter_50" + original_name = "conv2d_51.w_0" shape = [80, 336, 1, 1] dtype = "float32" min_val = float("-0.929654") @@ -557,6 +608,7 @@ class Program_weight_tensor_parameter_50: class Program_weight_tensor_parameter_51: name = "parameter_51" + original_name = "conv2d_50.b_0" shape = [336] dtype = "float32" min_val = float("-0.364391") @@ -568,6 +620,7 @@ class Program_weight_tensor_parameter_51: class Program_weight_tensor_parameter_52: name = "parameter_52" + original_name = "conv2d_50.w_0" shape = [336, 84, 1, 1] dtype = "float32" min_val = float("-0.715124") @@ -579,6 +632,7 @@ class Program_weight_tensor_parameter_52: class Program_weight_tensor_parameter_53: name = "parameter_53" + original_name = "conv2d_49.b_0" shape = [84] dtype = "float32" min_val = float("-0.0966891") @@ -590,6 +644,7 @@ class Program_weight_tensor_parameter_53: class Program_weight_tensor_parameter_54: name = "parameter_54" + original_name = "conv2d_49.w_0" shape = [84, 336, 1, 1] dtype = "float32" min_val = float("-0.483499") @@ -601,6 +656,7 @@ class Program_weight_tensor_parameter_54: class Program_weight_tensor_parameter_55: name = "parameter_55" + original_name = "batch_norm_38.b_0" shape = [336] dtype = "float32" min_val = float("-2.31619") @@ -612,6 +668,7 @@ class Program_weight_tensor_parameter_55: class Program_weight_tensor_parameter_56: name = "parameter_56" + original_name = "batch_norm_38.w_0" shape = [336] dtype = "float32" min_val = float("-0.804971") @@ -623,6 +680,7 @@ class Program_weight_tensor_parameter_56: class Program_weight_tensor_parameter_57: name = "parameter_57" + original_name = "batch_norm_38.w_2" shape = [336] dtype = "float32" min_val = float("0.000137531") @@ -634,6 +692,7 @@ class Program_weight_tensor_parameter_57: class Program_weight_tensor_parameter_58: name = "parameter_58" + original_name = "batch_norm_38.w_1" shape = [336] dtype = "float32" min_val = float("-12.3743") @@ -645,6 +704,7 @@ class Program_weight_tensor_parameter_58: class Program_weight_tensor_parameter_59: name = "parameter_59" + original_name = "conv2d_48.w_0" shape = [336, 1, 5, 5] dtype = "float32" min_val = float("-0.43205") @@ -656,6 +716,7 @@ class Program_weight_tensor_parameter_59: class Program_weight_tensor_parameter_60: name = "parameter_60" + original_name = "batch_norm_37.b_0" shape = [336] dtype = "float32" min_val = float("-3.62699") @@ -667,6 +728,7 @@ class Program_weight_tensor_parameter_60: class Program_weight_tensor_parameter_61: name = "parameter_61" + original_name = "batch_norm_37.w_0" shape = [336] dtype = "float32" min_val = float("-1.17063") @@ -678,6 +740,7 @@ class Program_weight_tensor_parameter_61: class Program_weight_tensor_parameter_62: name = "parameter_62" + original_name = "batch_norm_37.w_2" shape = [336] dtype = "float32" min_val = float("11.8318") @@ -689,6 +752,7 @@ class Program_weight_tensor_parameter_62: class Program_weight_tensor_parameter_63: name = "parameter_63" + original_name = "batch_norm_37.w_1" shape = [336] dtype = "float32" min_val = float("-8.36682e-06") @@ -700,6 +764,7 @@ class Program_weight_tensor_parameter_63: class Program_weight_tensor_parameter_64: name = "parameter_64" + original_name = "conv2d_47.w_0" shape = [336, 56, 1, 1] dtype = "float32" min_val = float("-0.616462") @@ -711,6 +776,7 @@ class Program_weight_tensor_parameter_64: class Program_weight_tensor_parameter_65: name = "parameter_65" + original_name = "batch_norm_36.b_0" shape = [56] dtype = "float32" min_val = float("0") @@ -720,6 +786,7 @@ class Program_weight_tensor_parameter_65: class Program_weight_tensor_parameter_66: name = "parameter_66" + original_name = "batch_norm_36.w_0" shape = [56] dtype = "float32" min_val = float("0") @@ -729,6 +796,7 @@ class Program_weight_tensor_parameter_66: class Program_weight_tensor_parameter_67: name = "parameter_67" + original_name = "batch_norm_36.w_2" shape = [56] dtype = "float32" min_val = float("0") @@ -738,6 +806,7 @@ class Program_weight_tensor_parameter_67: class Program_weight_tensor_parameter_68: name = "parameter_68" + original_name = "batch_norm_36.w_1" shape = [56] dtype = "float32" min_val = float("0") @@ -747,6 +816,7 @@ class Program_weight_tensor_parameter_68: class Program_weight_tensor_parameter_69: name = "parameter_69" + original_name = "conv2d_46.w_0" shape = [56, 336, 1, 1] dtype = "float32" min_val = float("-0.688862") @@ -758,6 +828,7 @@ class Program_weight_tensor_parameter_69: class Program_weight_tensor_parameter_70: name = "parameter_70" + original_name = "conv2d_45.b_0" shape = [336] dtype = "float32" min_val = float("-0.302611") @@ -769,6 +840,7 @@ class Program_weight_tensor_parameter_70: class Program_weight_tensor_parameter_71: name = "parameter_71" + original_name = "conv2d_45.w_0" shape = [336, 84, 1, 1] dtype = "float32" min_val = float("-0.640586") @@ -780,6 +852,7 @@ class Program_weight_tensor_parameter_71: class Program_weight_tensor_parameter_72: name = "parameter_72" + original_name = "conv2d_44.b_0" shape = [84] dtype = "float32" min_val = float("-0.0635902") @@ -791,6 +864,7 @@ class Program_weight_tensor_parameter_72: class Program_weight_tensor_parameter_73: name = "parameter_73" + original_name = "conv2d_44.w_0" shape = [84, 336, 1, 1] dtype = "float32" min_val = float("-0.885985") @@ -802,6 +876,7 @@ class Program_weight_tensor_parameter_73: class Program_weight_tensor_parameter_74: name = "parameter_74" + original_name = "batch_norm_35.b_0" shape = [336] dtype = "float32" min_val = float("-5.6082") @@ -813,6 +888,7 @@ class Program_weight_tensor_parameter_74: class Program_weight_tensor_parameter_75: name = "parameter_75" + original_name = "batch_norm_35.w_0" shape = [336] dtype = "float32" min_val = float("0.193434") @@ -824,6 +900,7 @@ class Program_weight_tensor_parameter_75: class Program_weight_tensor_parameter_76: name = "parameter_76" + original_name = "batch_norm_35.w_2" shape = [336] dtype = "float32" min_val = float("6.25827e-05") @@ -835,6 +912,7 @@ class Program_weight_tensor_parameter_76: class Program_weight_tensor_parameter_77: name = "parameter_77" + original_name = "batch_norm_35.w_1" shape = [336] dtype = "float32" min_val = float("-2.40829") @@ -846,6 +924,7 @@ class Program_weight_tensor_parameter_77: class Program_weight_tensor_parameter_78: name = "parameter_78" + original_name = "conv2d_43.w_0" shape = [336, 1, 3, 3] dtype = "float32" min_val = float("-0.777343") @@ -857,6 +936,7 @@ class Program_weight_tensor_parameter_78: class Program_weight_tensor_parameter_79: name = "parameter_79" + original_name = "batch_norm_34.b_0" shape = [336] dtype = "float32" min_val = float("-3.83973") @@ -868,6 +948,7 @@ class Program_weight_tensor_parameter_79: class Program_weight_tensor_parameter_80: name = "parameter_80" + original_name = "batch_norm_34.w_0" shape = [336] dtype = "float32" min_val = float("-0.541454") @@ -879,6 +960,7 @@ class Program_weight_tensor_parameter_80: class Program_weight_tensor_parameter_81: name = "parameter_81" + original_name = "batch_norm_34.w_2" shape = [336] dtype = "float32" min_val = float("5.28538") @@ -890,6 +972,7 @@ class Program_weight_tensor_parameter_81: class Program_weight_tensor_parameter_82: name = "parameter_82" + original_name = "batch_norm_34.w_1" shape = [336] dtype = "float32" min_val = float("-4.5976e-06") @@ -901,6 +984,7 @@ class Program_weight_tensor_parameter_82: class Program_weight_tensor_parameter_83: name = "parameter_83" + original_name = "conv2d_42.w_0" shape = [336, 56, 1, 1] dtype = "float32" min_val = float("-0.615976") @@ -912,6 +996,7 @@ class Program_weight_tensor_parameter_83: class Program_weight_tensor_parameter_84: name = "parameter_84" + original_name = "batch_norm_33.b_0" shape = [56] dtype = "float32" min_val = float("0") @@ -921,6 +1006,7 @@ class Program_weight_tensor_parameter_84: class Program_weight_tensor_parameter_85: name = "parameter_85" + original_name = "batch_norm_33.w_0" shape = [56] dtype = "float32" min_val = float("0") @@ -930,6 +1016,7 @@ class Program_weight_tensor_parameter_85: class Program_weight_tensor_parameter_86: name = "parameter_86" + original_name = "batch_norm_33.w_2" shape = [56] dtype = "float32" min_val = float("0") @@ -939,6 +1026,7 @@ class Program_weight_tensor_parameter_86: class Program_weight_tensor_parameter_87: name = "parameter_87" + original_name = "batch_norm_33.w_1" shape = [56] dtype = "float32" min_val = float("0") @@ -948,6 +1036,7 @@ class Program_weight_tensor_parameter_87: class Program_weight_tensor_parameter_88: name = "parameter_88" + original_name = "conv2d_41.w_0" shape = [56, 240, 1, 1] dtype = "float32" min_val = float("-0.687844") @@ -959,6 +1048,7 @@ class Program_weight_tensor_parameter_88: class Program_weight_tensor_parameter_89: name = "parameter_89" + original_name = "conv2d_40.b_0" shape = [240] dtype = "float32" min_val = float("-0.3876") @@ -970,6 +1060,7 @@ class Program_weight_tensor_parameter_89: class Program_weight_tensor_parameter_90: name = "parameter_90" + original_name = "conv2d_40.w_0" shape = [240, 60, 1, 1] dtype = "float32" min_val = float("-0.73428") @@ -981,6 +1072,7 @@ class Program_weight_tensor_parameter_90: class Program_weight_tensor_parameter_91: name = "parameter_91" + original_name = "conv2d_39.b_0" shape = [60] dtype = "float32" min_val = float("0") @@ -990,6 +1082,7 @@ class Program_weight_tensor_parameter_91: class Program_weight_tensor_parameter_92: name = "parameter_92" + original_name = "conv2d_39.w_0" shape = [60, 240, 1, 1] dtype = "float32" min_val = float("-0.960755") @@ -1001,6 +1094,7 @@ class Program_weight_tensor_parameter_92: class Program_weight_tensor_parameter_93: name = "parameter_93" + original_name = "batch_norm_32.b_0" shape = [240] dtype = "float32" min_val = float("-1.73446") @@ -1012,6 +1106,7 @@ class Program_weight_tensor_parameter_93: class Program_weight_tensor_parameter_94: name = "parameter_94" + original_name = "batch_norm_32.w_0" shape = [240] dtype = "float32" min_val = float("0.259654") @@ -1023,6 +1118,7 @@ class Program_weight_tensor_parameter_94: class Program_weight_tensor_parameter_95: name = "parameter_95" + original_name = "batch_norm_32.w_2" shape = [240] dtype = "float32" min_val = float("0.000167677") @@ -1034,6 +1130,7 @@ class Program_weight_tensor_parameter_95: class Program_weight_tensor_parameter_96: name = "parameter_96" + original_name = "batch_norm_32.w_1" shape = [240] dtype = "float32" min_val = float("-16.3074") @@ -1045,6 +1142,7 @@ class Program_weight_tensor_parameter_96: class Program_weight_tensor_parameter_97: name = "parameter_97" + original_name = "conv2d_38.w_0" shape = [240, 1, 3, 3] dtype = "float32" min_val = float("-1.12687") @@ -1056,6 +1154,7 @@ class Program_weight_tensor_parameter_97: class Program_weight_tensor_parameter_98: name = "parameter_98" + original_name = "batch_norm_31.b_0" shape = [240] dtype = "float32" min_val = float("-2.28532") @@ -1067,6 +1166,7 @@ class Program_weight_tensor_parameter_98: class Program_weight_tensor_parameter_99: name = "parameter_99" + original_name = "batch_norm_31.w_0" shape = [240] dtype = "float32" min_val = float("-0.352207") @@ -1078,6 +1178,7 @@ class Program_weight_tensor_parameter_99: class Program_weight_tensor_parameter_100: name = "parameter_100" + original_name = "batch_norm_31.w_2" shape = [240] dtype = "float32" min_val = float("9.49135") @@ -1089,6 +1190,7 @@ class Program_weight_tensor_parameter_100: class Program_weight_tensor_parameter_101: name = "parameter_101" + original_name = "batch_norm_31.w_1" shape = [240] dtype = "float32" min_val = float("-1.36796e-05") @@ -1100,6 +1202,7 @@ class Program_weight_tensor_parameter_101: class Program_weight_tensor_parameter_102: name = "parameter_102" + original_name = "conv2d_37.w_0" shape = [240, 40, 1, 1] dtype = "float32" min_val = float("-1.04598") @@ -1111,6 +1214,7 @@ class Program_weight_tensor_parameter_102: class Program_weight_tensor_parameter_103: name = "parameter_103" + original_name = "batch_norm_30.b_0" shape = [40] dtype = "float32" min_val = float("0") @@ -1120,6 +1224,7 @@ class Program_weight_tensor_parameter_103: class Program_weight_tensor_parameter_104: name = "parameter_104" + original_name = "batch_norm_30.w_0" shape = [40] dtype = "float32" min_val = float("0") @@ -1129,6 +1234,7 @@ class Program_weight_tensor_parameter_104: class Program_weight_tensor_parameter_105: name = "parameter_105" + original_name = "batch_norm_30.w_2" shape = [40] dtype = "float32" min_val = float("0") @@ -1138,6 +1244,7 @@ class Program_weight_tensor_parameter_105: class Program_weight_tensor_parameter_106: name = "parameter_106" + original_name = "batch_norm_30.w_1" shape = [40] dtype = "float32" min_val = float("0") @@ -1147,6 +1254,7 @@ class Program_weight_tensor_parameter_106: class Program_weight_tensor_parameter_107: name = "parameter_107" + original_name = "conv2d_36.w_0" shape = [40, 96, 1, 1] dtype = "float32" min_val = float("-0.9455") @@ -1158,6 +1266,7 @@ class Program_weight_tensor_parameter_107: class Program_weight_tensor_parameter_108: name = "parameter_108" + original_name = "batch_norm_29.b_0" shape = [96] dtype = "float32" min_val = float("-2.90736") @@ -1169,6 +1278,7 @@ class Program_weight_tensor_parameter_108: class Program_weight_tensor_parameter_109: name = "parameter_109" + original_name = "batch_norm_29.w_0" shape = [96] dtype = "float32" min_val = float("0.356411") @@ -1180,6 +1290,7 @@ class Program_weight_tensor_parameter_109: class Program_weight_tensor_parameter_110: name = "parameter_110" + original_name = "batch_norm_29.w_2" shape = [96] dtype = "float32" min_val = float("8.61013e-05") @@ -1191,6 +1302,7 @@ class Program_weight_tensor_parameter_110: class Program_weight_tensor_parameter_111: name = "parameter_111" + original_name = "batch_norm_29.w_1" shape = [96] dtype = "float32" min_val = float("-2.05985") @@ -1202,6 +1314,7 @@ class Program_weight_tensor_parameter_111: class Program_weight_tensor_parameter_112: name = "parameter_112" + original_name = "conv2d_35.w_0" shape = [96, 1, 3, 3] dtype = "float32" min_val = float("-0.766634") @@ -1213,6 +1326,7 @@ class Program_weight_tensor_parameter_112: class Program_weight_tensor_parameter_113: name = "parameter_113" + original_name = "batch_norm_28.b_0" shape = [96] dtype = "float32" min_val = float("-2.94365") @@ -1224,6 +1338,7 @@ class Program_weight_tensor_parameter_113: class Program_weight_tensor_parameter_114: name = "parameter_114" + original_name = "batch_norm_28.w_0" shape = [96] dtype = "float32" min_val = float("-0.250529") @@ -1235,6 +1350,7 @@ class Program_weight_tensor_parameter_114: class Program_weight_tensor_parameter_115: name = "parameter_115" + original_name = "batch_norm_28.w_2" shape = [96] dtype = "float32" min_val = float("12.2176") @@ -1246,6 +1362,7 @@ class Program_weight_tensor_parameter_115: class Program_weight_tensor_parameter_116: name = "parameter_116" + original_name = "batch_norm_28.w_1" shape = [96] dtype = "float32" min_val = float("-8.00924e-06") @@ -1257,6 +1374,7 @@ class Program_weight_tensor_parameter_116: class Program_weight_tensor_parameter_117: name = "parameter_117" + original_name = "conv2d_34.w_0" shape = [96, 40, 1, 1] dtype = "float32" min_val = float("-1.47171") @@ -1268,6 +1386,7 @@ class Program_weight_tensor_parameter_117: class Program_weight_tensor_parameter_118: name = "parameter_118" + original_name = "batch_norm_27.b_0" shape = [40] dtype = "float32" min_val = float("0") @@ -1277,6 +1396,7 @@ class Program_weight_tensor_parameter_118: class Program_weight_tensor_parameter_119: name = "parameter_119" + original_name = "batch_norm_27.w_0" shape = [40] dtype = "float32" min_val = float("0") @@ -1286,6 +1406,7 @@ class Program_weight_tensor_parameter_119: class Program_weight_tensor_parameter_120: name = "parameter_120" + original_name = "batch_norm_27.w_2" shape = [40] dtype = "float32" min_val = float("0") @@ -1295,6 +1416,7 @@ class Program_weight_tensor_parameter_120: class Program_weight_tensor_parameter_121: name = "parameter_121" + original_name = "batch_norm_27.w_1" shape = [40] dtype = "float32" min_val = float("0") @@ -1304,6 +1426,7 @@ class Program_weight_tensor_parameter_121: class Program_weight_tensor_parameter_122: name = "parameter_122" + original_name = "conv2d_33.w_0" shape = [40, 96, 1, 1] dtype = "float32" min_val = float("-0.891732") @@ -1315,6 +1438,7 @@ class Program_weight_tensor_parameter_122: class Program_weight_tensor_parameter_123: name = "parameter_123" + original_name = "batch_norm_26.b_0" shape = [96] dtype = "float32" min_val = float("-3.19176") @@ -1326,6 +1450,7 @@ class Program_weight_tensor_parameter_123: class Program_weight_tensor_parameter_124: name = "parameter_124" + original_name = "batch_norm_26.w_0" shape = [96] dtype = "float32" min_val = float("0.623799") @@ -1337,6 +1462,7 @@ class Program_weight_tensor_parameter_124: class Program_weight_tensor_parameter_125: name = "parameter_125" + original_name = "batch_norm_26.w_2" shape = [96] dtype = "float32" min_val = float("0.000110248") @@ -1348,6 +1474,7 @@ class Program_weight_tensor_parameter_125: class Program_weight_tensor_parameter_126: name = "parameter_126" + original_name = "batch_norm_26.w_1" shape = [96] dtype = "float32" min_val = float("-2.90018") @@ -1359,6 +1486,7 @@ class Program_weight_tensor_parameter_126: class Program_weight_tensor_parameter_127: name = "parameter_127" + original_name = "conv2d_32.w_0" shape = [96, 1, 3, 3] dtype = "float32" min_val = float("-0.757558") @@ -1370,6 +1498,7 @@ class Program_weight_tensor_parameter_127: class Program_weight_tensor_parameter_128: name = "parameter_128" + original_name = "batch_norm_25.b_0" shape = [96] dtype = "float32" min_val = float("-2.55868") @@ -1381,6 +1510,7 @@ class Program_weight_tensor_parameter_128: class Program_weight_tensor_parameter_129: name = "parameter_129" + original_name = "batch_norm_25.w_0" shape = [96] dtype = "float32" min_val = float("-0.496536") @@ -1392,6 +1522,7 @@ class Program_weight_tensor_parameter_129: class Program_weight_tensor_parameter_130: name = "parameter_130" + original_name = "batch_norm_25.w_2" shape = [96] dtype = "float32" min_val = float("12.4555") @@ -1403,6 +1534,7 @@ class Program_weight_tensor_parameter_130: class Program_weight_tensor_parameter_131: name = "parameter_131" + original_name = "batch_norm_25.w_1" shape = [96] dtype = "float32" min_val = float("-8.73606e-06") @@ -1414,6 +1546,7 @@ class Program_weight_tensor_parameter_131: class Program_weight_tensor_parameter_132: name = "parameter_132" + original_name = "conv2d_31.w_0" shape = [96, 40, 1, 1] dtype = "float32" min_val = float("-0.838911") @@ -1425,6 +1558,7 @@ class Program_weight_tensor_parameter_132: class Program_weight_tensor_parameter_133: name = "parameter_133" + original_name = "batch_norm_24.b_0" shape = [40] dtype = "float32" min_val = float("0") @@ -1434,6 +1568,7 @@ class Program_weight_tensor_parameter_133: class Program_weight_tensor_parameter_134: name = "parameter_134" + original_name = "batch_norm_24.w_0" shape = [40] dtype = "float32" min_val = float("0") @@ -1443,6 +1578,7 @@ class Program_weight_tensor_parameter_134: class Program_weight_tensor_parameter_135: name = "parameter_135" + original_name = "batch_norm_24.w_2" shape = [40] dtype = "float32" min_val = float("0") @@ -1452,6 +1588,7 @@ class Program_weight_tensor_parameter_135: class Program_weight_tensor_parameter_136: name = "parameter_136" + original_name = "batch_norm_24.w_1" shape = [40] dtype = "float32" min_val = float("0") @@ -1461,6 +1598,7 @@ class Program_weight_tensor_parameter_136: class Program_weight_tensor_parameter_137: name = "parameter_137" + original_name = "conv2d_30.w_0" shape = [40, 104, 1, 1] dtype = "float32" min_val = float("-0.68535") @@ -1472,6 +1610,7 @@ class Program_weight_tensor_parameter_137: class Program_weight_tensor_parameter_138: name = "parameter_138" + original_name = "batch_norm_23.b_0" shape = [104] dtype = "float32" min_val = float("-2.24451") @@ -1483,6 +1622,7 @@ class Program_weight_tensor_parameter_138: class Program_weight_tensor_parameter_139: name = "parameter_139" + original_name = "batch_norm_23.w_0" shape = [104] dtype = "float32" min_val = float("0.187077") @@ -1494,6 +1634,7 @@ class Program_weight_tensor_parameter_139: class Program_weight_tensor_parameter_140: name = "parameter_140" + original_name = "batch_norm_23.w_2" shape = [104] dtype = "float32" min_val = float("0.000140825") @@ -1505,6 +1646,7 @@ class Program_weight_tensor_parameter_140: class Program_weight_tensor_parameter_141: name = "parameter_141" + original_name = "batch_norm_23.w_1" shape = [104] dtype = "float32" min_val = float("-8.09244") @@ -1516,6 +1658,7 @@ class Program_weight_tensor_parameter_141: class Program_weight_tensor_parameter_142: name = "parameter_142" + original_name = "conv2d_29.w_0" shape = [104, 1, 3, 3] dtype = "float32" min_val = float("-0.821249") @@ -1527,6 +1670,7 @@ class Program_weight_tensor_parameter_142: class Program_weight_tensor_parameter_143: name = "parameter_143" + original_name = "batch_norm_22.b_0" shape = [104] dtype = "float32" min_val = float("-1.87689") @@ -1538,6 +1682,7 @@ class Program_weight_tensor_parameter_143: class Program_weight_tensor_parameter_144: name = "parameter_144" + original_name = "batch_norm_22.w_0" shape = [104] dtype = "float32" min_val = float("-0.0355361") @@ -1549,6 +1694,7 @@ class Program_weight_tensor_parameter_144: class Program_weight_tensor_parameter_145: name = "parameter_145" + original_name = "batch_norm_22.w_2" shape = [104] dtype = "float32" min_val = float("5.54799") @@ -1560,6 +1706,7 @@ class Program_weight_tensor_parameter_145: class Program_weight_tensor_parameter_146: name = "parameter_146" + original_name = "batch_norm_22.w_1" shape = [104] dtype = "float32" min_val = float("-2.99764e-06") @@ -1571,6 +1718,7 @@ class Program_weight_tensor_parameter_146: class Program_weight_tensor_parameter_147: name = "parameter_147" + original_name = "conv2d_28.w_0" shape = [104, 40, 1, 1] dtype = "float32" min_val = float("-0.62063") @@ -1582,6 +1730,7 @@ class Program_weight_tensor_parameter_147: class Program_weight_tensor_parameter_148: name = "parameter_148" + original_name = "batch_norm_21.b_0" shape = [40] dtype = "float32" min_val = float("0") @@ -1591,6 +1740,7 @@ class Program_weight_tensor_parameter_148: class Program_weight_tensor_parameter_149: name = "parameter_149" + original_name = "batch_norm_21.w_0" shape = [40] dtype = "float32" min_val = float("0") @@ -1600,6 +1750,7 @@ class Program_weight_tensor_parameter_149: class Program_weight_tensor_parameter_150: name = "parameter_150" + original_name = "batch_norm_21.w_2" shape = [40] dtype = "float32" min_val = float("0") @@ -1609,6 +1760,7 @@ class Program_weight_tensor_parameter_150: class Program_weight_tensor_parameter_151: name = "parameter_151" + original_name = "batch_norm_21.w_1" shape = [40] dtype = "float32" min_val = float("0") @@ -1618,6 +1770,7 @@ class Program_weight_tensor_parameter_151: class Program_weight_tensor_parameter_152: name = "parameter_152" + original_name = "conv2d_27.w_0" shape = [40, 120, 1, 1] dtype = "float32" min_val = float("-1.43018") @@ -1629,6 +1782,7 @@ class Program_weight_tensor_parameter_152: class Program_weight_tensor_parameter_153: name = "parameter_153" + original_name = "batch_norm_20.b_0" shape = [120] dtype = "float32" min_val = float("-1.42656") @@ -1640,6 +1794,7 @@ class Program_weight_tensor_parameter_153: class Program_weight_tensor_parameter_154: name = "parameter_154" + original_name = "batch_norm_20.w_0" shape = [120] dtype = "float32" min_val = float("0.409872") @@ -1651,6 +1806,7 @@ class Program_weight_tensor_parameter_154: class Program_weight_tensor_parameter_155: name = "parameter_155" + original_name = "batch_norm_20.w_2" shape = [120] dtype = "float32" min_val = float("5.63917e-05") @@ -1662,6 +1818,7 @@ class Program_weight_tensor_parameter_155: class Program_weight_tensor_parameter_156: name = "parameter_156" + original_name = "batch_norm_20.w_1" shape = [120] dtype = "float32" min_val = float("-5.51925") @@ -1673,6 +1830,7 @@ class Program_weight_tensor_parameter_156: class Program_weight_tensor_parameter_157: name = "parameter_157" + original_name = "conv2d_26.w_0" shape = [120, 1, 3, 3] dtype = "float32" min_val = float("-0.494566") @@ -1684,6 +1842,7 @@ class Program_weight_tensor_parameter_157: class Program_weight_tensor_parameter_158: name = "parameter_158" + original_name = "batch_norm_19.b_0" shape = [120] dtype = "float32" min_val = float("-2.03548") @@ -1695,6 +1854,7 @@ class Program_weight_tensor_parameter_158: class Program_weight_tensor_parameter_159: name = "parameter_159" + original_name = "batch_norm_19.w_0" shape = [120] dtype = "float32" min_val = float("-0.0419506") @@ -1706,6 +1866,7 @@ class Program_weight_tensor_parameter_159: class Program_weight_tensor_parameter_160: name = "parameter_160" + original_name = "batch_norm_19.w_2" shape = [120] dtype = "float32" min_val = float("23.369") @@ -1717,6 +1878,7 @@ class Program_weight_tensor_parameter_160: class Program_weight_tensor_parameter_161: name = "parameter_161" + original_name = "batch_norm_19.w_1" shape = [120] dtype = "float32" min_val = float("-3.59612e-05") @@ -1728,6 +1890,7 @@ class Program_weight_tensor_parameter_161: class Program_weight_tensor_parameter_162: name = "parameter_162" + original_name = "conv2d_25.w_0" shape = [120, 24, 1, 1] dtype = "float32" min_val = float("-1.36196") @@ -1739,6 +1902,7 @@ class Program_weight_tensor_parameter_162: class Program_weight_tensor_parameter_163: name = "parameter_163" + original_name = "batch_norm_18.b_0" shape = [24] dtype = "float32" min_val = float("0") @@ -1748,6 +1912,7 @@ class Program_weight_tensor_parameter_163: class Program_weight_tensor_parameter_164: name = "parameter_164" + original_name = "batch_norm_18.w_0" shape = [24] dtype = "float32" min_val = float("0") @@ -1757,6 +1922,7 @@ class Program_weight_tensor_parameter_164: class Program_weight_tensor_parameter_165: name = "parameter_165" + original_name = "batch_norm_18.w_2" shape = [24] dtype = "float32" min_val = float("0") @@ -1766,6 +1932,7 @@ class Program_weight_tensor_parameter_165: class Program_weight_tensor_parameter_166: name = "parameter_166" + original_name = "batch_norm_18.w_1" shape = [24] dtype = "float32" min_val = float("0") @@ -1775,6 +1942,7 @@ class Program_weight_tensor_parameter_166: class Program_weight_tensor_parameter_167: name = "parameter_167" + original_name = "conv2d_24.w_0" shape = [24, 64, 1, 1] dtype = "float32" min_val = float("-0.799287") @@ -1786,6 +1954,7 @@ class Program_weight_tensor_parameter_167: class Program_weight_tensor_parameter_168: name = "parameter_168" + original_name = "conv2d_23.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -1795,6 +1964,7 @@ class Program_weight_tensor_parameter_168: class Program_weight_tensor_parameter_169: name = "parameter_169" + original_name = "conv2d_23.w_0" shape = [64, 16, 1, 1] dtype = "float32" min_val = float("-0.832081") @@ -1806,6 +1976,7 @@ class Program_weight_tensor_parameter_169: class Program_weight_tensor_parameter_170: name = "parameter_170" + original_name = "conv2d_22.b_0" shape = [16] dtype = "float32" min_val = float("0") @@ -1815,6 +1986,7 @@ class Program_weight_tensor_parameter_170: class Program_weight_tensor_parameter_171: name = "parameter_171" + original_name = "conv2d_22.w_0" shape = [16, 64, 1, 1] dtype = "float32" min_val = float("-0.957694") @@ -1826,6 +1998,7 @@ class Program_weight_tensor_parameter_171: class Program_weight_tensor_parameter_172: name = "parameter_172" + original_name = "batch_norm_17.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -1835,6 +2008,7 @@ class Program_weight_tensor_parameter_172: class Program_weight_tensor_parameter_173: name = "parameter_173" + original_name = "batch_norm_17.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -1844,6 +2018,7 @@ class Program_weight_tensor_parameter_173: class Program_weight_tensor_parameter_174: name = "parameter_174" + original_name = "batch_norm_17.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -1853,6 +2028,7 @@ class Program_weight_tensor_parameter_174: class Program_weight_tensor_parameter_175: name = "parameter_175" + original_name = "batch_norm_17.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -1862,6 +2038,7 @@ class Program_weight_tensor_parameter_175: class Program_weight_tensor_parameter_176: name = "parameter_176" + original_name = "conv2d_21.w_0" shape = [64, 1, 5, 5] dtype = "float32" min_val = float("-1.07452") @@ -1873,6 +2050,7 @@ class Program_weight_tensor_parameter_176: class Program_weight_tensor_parameter_177: name = "parameter_177" + original_name = "batch_norm_16.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -1882,6 +2060,7 @@ class Program_weight_tensor_parameter_177: class Program_weight_tensor_parameter_178: name = "parameter_178" + original_name = "batch_norm_16.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -1891,6 +2070,7 @@ class Program_weight_tensor_parameter_178: class Program_weight_tensor_parameter_179: name = "parameter_179" + original_name = "batch_norm_16.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -1900,6 +2080,7 @@ class Program_weight_tensor_parameter_179: class Program_weight_tensor_parameter_180: name = "parameter_180" + original_name = "batch_norm_16.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -1909,6 +2090,7 @@ class Program_weight_tensor_parameter_180: class Program_weight_tensor_parameter_181: name = "parameter_181" + original_name = "conv2d_20.w_0" shape = [64, 24, 1, 1] dtype = "float32" min_val = float("-1.06359") @@ -1920,6 +2102,7 @@ class Program_weight_tensor_parameter_181: class Program_weight_tensor_parameter_182: name = "parameter_182" + original_name = "batch_norm_15.b_0" shape = [24] dtype = "float32" min_val = float("0") @@ -1929,6 +2112,7 @@ class Program_weight_tensor_parameter_182: class Program_weight_tensor_parameter_183: name = "parameter_183" + original_name = "batch_norm_15.w_0" shape = [24] dtype = "float32" min_val = float("0") @@ -1938,6 +2122,7 @@ class Program_weight_tensor_parameter_183: class Program_weight_tensor_parameter_184: name = "parameter_184" + original_name = "batch_norm_15.w_2" shape = [24] dtype = "float32" min_val = float("0") @@ -1947,6 +2132,7 @@ class Program_weight_tensor_parameter_184: class Program_weight_tensor_parameter_185: name = "parameter_185" + original_name = "batch_norm_15.w_1" shape = [24] dtype = "float32" min_val = float("0") @@ -1956,6 +2142,7 @@ class Program_weight_tensor_parameter_185: class Program_weight_tensor_parameter_186: name = "parameter_186" + original_name = "conv2d_19.w_0" shape = [24, 64, 1, 1] dtype = "float32" min_val = float("-0.882603") @@ -1967,6 +2154,7 @@ class Program_weight_tensor_parameter_186: class Program_weight_tensor_parameter_187: name = "parameter_187" + original_name = "conv2d_18.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -1976,6 +2164,7 @@ class Program_weight_tensor_parameter_187: class Program_weight_tensor_parameter_188: name = "parameter_188" + original_name = "conv2d_18.w_0" shape = [64, 16, 1, 1] dtype = "float32" min_val = float("-0.703183") @@ -1987,6 +2176,7 @@ class Program_weight_tensor_parameter_188: class Program_weight_tensor_parameter_189: name = "parameter_189" + original_name = "conv2d_17.b_0" shape = [16] dtype = "float32" min_val = float("0") @@ -1996,6 +2186,7 @@ class Program_weight_tensor_parameter_189: class Program_weight_tensor_parameter_190: name = "parameter_190" + original_name = "conv2d_17.w_0" shape = [16, 64, 1, 1] dtype = "float32" min_val = float("-1.5768") @@ -2007,6 +2198,7 @@ class Program_weight_tensor_parameter_190: class Program_weight_tensor_parameter_191: name = "parameter_191" + original_name = "batch_norm_14.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -2016,6 +2208,7 @@ class Program_weight_tensor_parameter_191: class Program_weight_tensor_parameter_192: name = "parameter_192" + original_name = "batch_norm_14.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -2025,6 +2218,7 @@ class Program_weight_tensor_parameter_192: class Program_weight_tensor_parameter_193: name = "parameter_193" + original_name = "batch_norm_14.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -2034,6 +2228,7 @@ class Program_weight_tensor_parameter_193: class Program_weight_tensor_parameter_194: name = "parameter_194" + original_name = "batch_norm_14.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -2043,6 +2238,7 @@ class Program_weight_tensor_parameter_194: class Program_weight_tensor_parameter_195: name = "parameter_195" + original_name = "conv2d_16.w_0" shape = [64, 1, 5, 5] dtype = "float32" min_val = float("-1.23397") @@ -2054,6 +2250,7 @@ class Program_weight_tensor_parameter_195: class Program_weight_tensor_parameter_196: name = "parameter_196" + original_name = "batch_norm_13.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -2063,6 +2260,7 @@ class Program_weight_tensor_parameter_196: class Program_weight_tensor_parameter_197: name = "parameter_197" + original_name = "batch_norm_13.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -2072,6 +2270,7 @@ class Program_weight_tensor_parameter_197: class Program_weight_tensor_parameter_198: name = "parameter_198" + original_name = "batch_norm_13.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -2081,6 +2280,7 @@ class Program_weight_tensor_parameter_198: class Program_weight_tensor_parameter_199: name = "parameter_199" + original_name = "batch_norm_13.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -2090,6 +2290,7 @@ class Program_weight_tensor_parameter_199: class Program_weight_tensor_parameter_200: name = "parameter_200" + original_name = "conv2d_15.w_0" shape = [64, 24, 1, 1] dtype = "float32" min_val = float("-1.0741") @@ -2101,6 +2302,7 @@ class Program_weight_tensor_parameter_200: class Program_weight_tensor_parameter_201: name = "parameter_201" + original_name = "batch_norm_12.b_0" shape = [24] dtype = "float32" min_val = float("0") @@ -2110,6 +2312,7 @@ class Program_weight_tensor_parameter_201: class Program_weight_tensor_parameter_202: name = "parameter_202" + original_name = "batch_norm_12.w_0" shape = [24] dtype = "float32" min_val = float("0") @@ -2119,6 +2322,7 @@ class Program_weight_tensor_parameter_202: class Program_weight_tensor_parameter_203: name = "parameter_203" + original_name = "batch_norm_12.w_2" shape = [24] dtype = "float32" min_val = float("0") @@ -2128,6 +2332,7 @@ class Program_weight_tensor_parameter_203: class Program_weight_tensor_parameter_204: name = "parameter_204" + original_name = "batch_norm_12.w_1" shape = [24] dtype = "float32" min_val = float("0") @@ -2137,6 +2342,7 @@ class Program_weight_tensor_parameter_204: class Program_weight_tensor_parameter_205: name = "parameter_205" + original_name = "conv2d_14.w_0" shape = [24, 40, 1, 1] dtype = "float32" min_val = float("-1.47384") @@ -2148,6 +2354,7 @@ class Program_weight_tensor_parameter_205: class Program_weight_tensor_parameter_206: name = "parameter_206" + original_name = "conv2d_13.b_0" shape = [40] dtype = "float32" min_val = float("0") @@ -2157,6 +2364,7 @@ class Program_weight_tensor_parameter_206: class Program_weight_tensor_parameter_207: name = "parameter_207" + original_name = "conv2d_13.w_0" shape = [40, 10, 1, 1] dtype = "float32" min_val = float("-2.86316e-13") @@ -2168,6 +2376,7 @@ class Program_weight_tensor_parameter_207: class Program_weight_tensor_parameter_208: name = "parameter_208" + original_name = "conv2d_12.b_0" shape = [10] dtype = "float32" min_val = float("0") @@ -2177,6 +2386,7 @@ class Program_weight_tensor_parameter_208: class Program_weight_tensor_parameter_209: name = "parameter_209" + original_name = "conv2d_12.w_0" shape = [10, 40, 1, 1] dtype = "float32" min_val = float("-5.77377e-13") @@ -2188,6 +2398,7 @@ class Program_weight_tensor_parameter_209: class Program_weight_tensor_parameter_210: name = "parameter_210" + original_name = "batch_norm_11.b_0" shape = [40] dtype = "float32" min_val = float("0") @@ -2197,6 +2408,7 @@ class Program_weight_tensor_parameter_210: class Program_weight_tensor_parameter_211: name = "parameter_211" + original_name = "batch_norm_11.w_0" shape = [40] dtype = "float32" min_val = float("0") @@ -2206,6 +2418,7 @@ class Program_weight_tensor_parameter_211: class Program_weight_tensor_parameter_212: name = "parameter_212" + original_name = "batch_norm_11.w_2" shape = [40] dtype = "float32" min_val = float("0") @@ -2215,6 +2428,7 @@ class Program_weight_tensor_parameter_212: class Program_weight_tensor_parameter_213: name = "parameter_213" + original_name = "batch_norm_11.w_1" shape = [40] dtype = "float32" min_val = float("0") @@ -2224,6 +2438,7 @@ class Program_weight_tensor_parameter_213: class Program_weight_tensor_parameter_214: name = "parameter_214" + original_name = "conv2d_11.w_0" shape = [40, 1, 5, 5] dtype = "float32" min_val = float("-0.968468") @@ -2235,6 +2450,7 @@ class Program_weight_tensor_parameter_214: class Program_weight_tensor_parameter_215: name = "parameter_215" + original_name = "batch_norm_10.b_0" shape = [40] dtype = "float32" min_val = float("0") @@ -2244,6 +2460,7 @@ class Program_weight_tensor_parameter_215: class Program_weight_tensor_parameter_216: name = "parameter_216" + original_name = "batch_norm_10.w_0" shape = [40] dtype = "float32" min_val = float("0") @@ -2253,6 +2470,7 @@ class Program_weight_tensor_parameter_216: class Program_weight_tensor_parameter_217: name = "parameter_217" + original_name = "batch_norm_10.w_2" shape = [40] dtype = "float32" min_val = float("0") @@ -2262,6 +2480,7 @@ class Program_weight_tensor_parameter_217: class Program_weight_tensor_parameter_218: name = "parameter_218" + original_name = "batch_norm_10.w_1" shape = [40] dtype = "float32" min_val = float("0") @@ -2271,6 +2490,7 @@ class Program_weight_tensor_parameter_218: class Program_weight_tensor_parameter_219: name = "parameter_219" + original_name = "conv2d_10.w_0" shape = [40, 16, 1, 1] dtype = "float32" min_val = float("-1.73538") @@ -2282,6 +2502,7 @@ class Program_weight_tensor_parameter_219: class Program_weight_tensor_parameter_220: name = "parameter_220" + original_name = "batch_norm_9.b_0" shape = [16] dtype = "float32" min_val = float("0") @@ -2291,6 +2512,7 @@ class Program_weight_tensor_parameter_220: class Program_weight_tensor_parameter_221: name = "parameter_221" + original_name = "batch_norm_9.w_0" shape = [16] dtype = "float32" min_val = float("0") @@ -2300,6 +2522,7 @@ class Program_weight_tensor_parameter_221: class Program_weight_tensor_parameter_222: name = "parameter_222" + original_name = "batch_norm_9.w_2" shape = [16] dtype = "float32" min_val = float("0") @@ -2309,6 +2532,7 @@ class Program_weight_tensor_parameter_222: class Program_weight_tensor_parameter_223: name = "parameter_223" + original_name = "batch_norm_9.w_1" shape = [16] dtype = "float32" min_val = float("0") @@ -2318,6 +2542,7 @@ class Program_weight_tensor_parameter_223: class Program_weight_tensor_parameter_224: name = "parameter_224" + original_name = "conv2d_9.w_0" shape = [16, 40, 1, 1] dtype = "float32" min_val = float("-1.08973") @@ -2329,6 +2554,7 @@ class Program_weight_tensor_parameter_224: class Program_weight_tensor_parameter_225: name = "parameter_225" + original_name = "batch_norm_8.b_0" shape = [40] dtype = "float32" min_val = float("0") @@ -2338,6 +2564,7 @@ class Program_weight_tensor_parameter_225: class Program_weight_tensor_parameter_226: name = "parameter_226" + original_name = "batch_norm_8.w_0" shape = [40] dtype = "float32" min_val = float("0") @@ -2347,6 +2574,7 @@ class Program_weight_tensor_parameter_226: class Program_weight_tensor_parameter_227: name = "parameter_227" + original_name = "batch_norm_8.w_2" shape = [40] dtype = "float32" min_val = float("0") @@ -2356,6 +2584,7 @@ class Program_weight_tensor_parameter_227: class Program_weight_tensor_parameter_228: name = "parameter_228" + original_name = "batch_norm_8.w_1" shape = [40] dtype = "float32" min_val = float("0") @@ -2365,6 +2594,7 @@ class Program_weight_tensor_parameter_228: class Program_weight_tensor_parameter_229: name = "parameter_229" + original_name = "conv2d_8.w_0" shape = [40, 1, 3, 3] dtype = "float32" min_val = float("-1.42618") @@ -2376,6 +2606,7 @@ class Program_weight_tensor_parameter_229: class Program_weight_tensor_parameter_230: name = "parameter_230" + original_name = "batch_norm_7.b_0" shape = [40] dtype = "float32" min_val = float("0") @@ -2385,6 +2616,7 @@ class Program_weight_tensor_parameter_230: class Program_weight_tensor_parameter_231: name = "parameter_231" + original_name = "batch_norm_7.w_0" shape = [40] dtype = "float32" min_val = float("0") @@ -2394,6 +2626,7 @@ class Program_weight_tensor_parameter_231: class Program_weight_tensor_parameter_232: name = "parameter_232" + original_name = "batch_norm_7.w_2" shape = [40] dtype = "float32" min_val = float("0") @@ -2403,6 +2636,7 @@ class Program_weight_tensor_parameter_232: class Program_weight_tensor_parameter_233: name = "parameter_233" + original_name = "batch_norm_7.w_1" shape = [40] dtype = "float32" min_val = float("0") @@ -2412,6 +2646,7 @@ class Program_weight_tensor_parameter_233: class Program_weight_tensor_parameter_234: name = "parameter_234" + original_name = "conv2d_7.w_0" shape = [40, 16, 1, 1] dtype = "float32" min_val = float("-0.995556") @@ -2423,6 +2658,7 @@ class Program_weight_tensor_parameter_234: class Program_weight_tensor_parameter_235: name = "parameter_235" + original_name = "batch_norm_6.b_0" shape = [16] dtype = "float32" min_val = float("0") @@ -2432,6 +2668,7 @@ class Program_weight_tensor_parameter_235: class Program_weight_tensor_parameter_236: name = "parameter_236" + original_name = "batch_norm_6.w_0" shape = [16] dtype = "float32" min_val = float("0") @@ -2441,6 +2678,7 @@ class Program_weight_tensor_parameter_236: class Program_weight_tensor_parameter_237: name = "parameter_237" + original_name = "batch_norm_6.w_2" shape = [16] dtype = "float32" min_val = float("0") @@ -2450,6 +2688,7 @@ class Program_weight_tensor_parameter_237: class Program_weight_tensor_parameter_238: name = "parameter_238" + original_name = "batch_norm_6.w_1" shape = [16] dtype = "float32" min_val = float("0") @@ -2459,6 +2698,7 @@ class Program_weight_tensor_parameter_238: class Program_weight_tensor_parameter_239: name = "parameter_239" + original_name = "conv2d_6.w_0" shape = [16, 32, 1, 1] dtype = "float32" min_val = float("-1.69633") @@ -2470,6 +2710,7 @@ class Program_weight_tensor_parameter_239: class Program_weight_tensor_parameter_240: name = "parameter_240" + original_name = "batch_norm_5.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -2479,6 +2720,7 @@ class Program_weight_tensor_parameter_240: class Program_weight_tensor_parameter_241: name = "parameter_241" + original_name = "batch_norm_5.w_0" shape = [32] dtype = "float32" min_val = float("0") @@ -2488,6 +2730,7 @@ class Program_weight_tensor_parameter_241: class Program_weight_tensor_parameter_242: name = "parameter_242" + original_name = "batch_norm_5.w_2" shape = [32] dtype = "float32" min_val = float("0") @@ -2497,6 +2740,7 @@ class Program_weight_tensor_parameter_242: class Program_weight_tensor_parameter_243: name = "parameter_243" + original_name = "batch_norm_5.w_1" shape = [32] dtype = "float32" min_val = float("0") @@ -2506,6 +2750,7 @@ class Program_weight_tensor_parameter_243: class Program_weight_tensor_parameter_244: name = "parameter_244" + original_name = "conv2d_5.w_0" shape = [32, 1, 3, 3] dtype = "float32" min_val = float("-0.971538") @@ -2517,6 +2762,7 @@ class Program_weight_tensor_parameter_244: class Program_weight_tensor_parameter_245: name = "parameter_245" + original_name = "batch_norm_4.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -2526,6 +2772,7 @@ class Program_weight_tensor_parameter_245: class Program_weight_tensor_parameter_246: name = "parameter_246" + original_name = "batch_norm_4.w_0" shape = [32] dtype = "float32" min_val = float("0") @@ -2535,6 +2782,7 @@ class Program_weight_tensor_parameter_246: class Program_weight_tensor_parameter_247: name = "parameter_247" + original_name = "batch_norm_4.w_2" shape = [32] dtype = "float32" min_val = float("0") @@ -2544,6 +2792,7 @@ class Program_weight_tensor_parameter_247: class Program_weight_tensor_parameter_248: name = "parameter_248" + original_name = "batch_norm_4.w_1" shape = [32] dtype = "float32" min_val = float("0") @@ -2553,6 +2802,7 @@ class Program_weight_tensor_parameter_248: class Program_weight_tensor_parameter_249: name = "parameter_249" + original_name = "conv2d_4.w_0" shape = [32, 8, 1, 1] dtype = "float32" min_val = float("-2.05033") @@ -2564,6 +2814,7 @@ class Program_weight_tensor_parameter_249: class Program_weight_tensor_parameter_250: name = "parameter_250" + original_name = "batch_norm_3.b_0" shape = [8] dtype = "float32" min_val = float("0") @@ -2573,6 +2824,7 @@ class Program_weight_tensor_parameter_250: class Program_weight_tensor_parameter_251: name = "parameter_251" + original_name = "batch_norm_3.w_0" shape = [8] dtype = "float32" min_val = float("0") @@ -2582,6 +2834,7 @@ class Program_weight_tensor_parameter_251: class Program_weight_tensor_parameter_252: name = "parameter_252" + original_name = "batch_norm_3.w_2" shape = [8] dtype = "float32" min_val = float("0") @@ -2591,6 +2844,7 @@ class Program_weight_tensor_parameter_252: class Program_weight_tensor_parameter_253: name = "parameter_253" + original_name = "batch_norm_3.w_1" shape = [8] dtype = "float32" min_val = float("0") @@ -2600,6 +2854,7 @@ class Program_weight_tensor_parameter_253: class Program_weight_tensor_parameter_254: name = "parameter_254" + original_name = "conv2d_3.w_0" shape = [8, 8, 1, 1] dtype = "float32" min_val = float("0") @@ -2609,6 +2864,7 @@ class Program_weight_tensor_parameter_254: class Program_weight_tensor_parameter_255: name = "parameter_255" + original_name = "batch_norm_2.b_0" shape = [8] dtype = "float32" min_val = float("0") @@ -2618,6 +2874,7 @@ class Program_weight_tensor_parameter_255: class Program_weight_tensor_parameter_256: name = "parameter_256" + original_name = "batch_norm_2.w_0" shape = [8] dtype = "float32" min_val = float("0") @@ -2627,6 +2884,7 @@ class Program_weight_tensor_parameter_256: class Program_weight_tensor_parameter_257: name = "parameter_257" + original_name = "batch_norm_2.w_2" shape = [8] dtype = "float32" min_val = float("0") @@ -2636,6 +2894,7 @@ class Program_weight_tensor_parameter_257: class Program_weight_tensor_parameter_258: name = "parameter_258" + original_name = "batch_norm_2.w_1" shape = [8] dtype = "float32" min_val = float("0") @@ -2645,6 +2904,7 @@ class Program_weight_tensor_parameter_258: class Program_weight_tensor_parameter_259: name = "parameter_259" + original_name = "conv2d_2.w_0" shape = [8, 1, 3, 3] dtype = "float32" min_val = float("-1.87906") @@ -2656,6 +2916,7 @@ class Program_weight_tensor_parameter_259: class Program_weight_tensor_parameter_260: name = "parameter_260" + original_name = "batch_norm_1.b_0" shape = [8] dtype = "float32" min_val = float("0") @@ -2665,6 +2926,7 @@ class Program_weight_tensor_parameter_260: class Program_weight_tensor_parameter_261: name = "parameter_261" + original_name = "batch_norm_1.w_0" shape = [8] dtype = "float32" min_val = float("0") @@ -2674,6 +2936,7 @@ class Program_weight_tensor_parameter_261: class Program_weight_tensor_parameter_262: name = "parameter_262" + original_name = "batch_norm_1.w_2" shape = [8] dtype = "float32" min_val = float("0") @@ -2683,6 +2946,7 @@ class Program_weight_tensor_parameter_262: class Program_weight_tensor_parameter_263: name = "parameter_263" + original_name = "batch_norm_1.w_1" shape = [8] dtype = "float32" min_val = float("0") @@ -2692,6 +2956,7 @@ class Program_weight_tensor_parameter_263: class Program_weight_tensor_parameter_264: name = "parameter_264" + original_name = "conv2d_1.w_0" shape = [8, 8, 1, 1] dtype = "float32" min_val = float("0") @@ -2701,6 +2966,7 @@ class Program_weight_tensor_parameter_264: class Program_weight_tensor_parameter_265: name = "parameter_265" + original_name = "batch_norm_0.b_0" shape = [8] dtype = "float32" min_val = float("0") @@ -2710,6 +2976,7 @@ class Program_weight_tensor_parameter_265: class Program_weight_tensor_parameter_266: name = "parameter_266" + original_name = "batch_norm_0.w_0" shape = [8] dtype = "float32" min_val = float("0") @@ -2719,6 +2986,7 @@ class Program_weight_tensor_parameter_266: class Program_weight_tensor_parameter_267: name = "parameter_267" + original_name = "batch_norm_0.w_2" shape = [8] dtype = "float32" min_val = float("0") @@ -2728,6 +2996,7 @@ class Program_weight_tensor_parameter_267: class Program_weight_tensor_parameter_268: name = "parameter_268" + original_name = "batch_norm_0.w_1" shape = [8] dtype = "float32" min_val = float("0") @@ -2737,6 +3006,7 @@ class Program_weight_tensor_parameter_268: class Program_weight_tensor_parameter_269: name = "parameter_269" + original_name = "conv2d_0.w_0" shape = [8, 3, 3, 3] dtype = "float32" min_val = float("-3.40429") diff --git a/paddle_samples/PaddleX/MobileNetV3_large_x0_75/subgraph_0/input_meta.py b/paddle_samples/PaddleX/MobileNetV3_large_x0_75/subgraph_0/input_meta.py index 795f9bcfe..6d7dac1c9 100644 --- a/paddle_samples/PaddleX/MobileNetV3_large_x0_75/subgraph_0/input_meta.py +++ b/paddle_samples/PaddleX/MobileNetV3_large_x0_75/subgraph_0/input_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_data_0: name = "data_0" + original_name = "var_0" shape = [128, 3, 224, 224] dtype = "float32" min_val = float("-2.1179") diff --git a/paddle_samples/PaddleX/MobileNetV3_large_x0_75/subgraph_0/weight_meta.py b/paddle_samples/PaddleX/MobileNetV3_large_x0_75/subgraph_0/weight_meta.py index b78f07b41..2fc3f641b 100644 --- a/paddle_samples/PaddleX/MobileNetV3_large_x0_75/subgraph_0/weight_meta.py +++ b/paddle_samples/PaddleX/MobileNetV3_large_x0_75/subgraph_0/weight_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_parameter_0: name = "parameter_0" + original_name = "linear_0.b_0" shape = [102] dtype = "float32" data = None @@ -7,6 +8,7 @@ class Program_weight_tensor_parameter_0: class Program_weight_tensor_parameter_1: name = "parameter_1" + original_name = "linear_0.w_0" shape = [1280, 102] dtype = "float32" min_val = float("-0.0658889") @@ -18,6 +20,7 @@ class Program_weight_tensor_parameter_1: class Program_weight_tensor_parameter_2: name = "parameter_2" + original_name = "conv2d_63.w_0" shape = [1280, 720, 1, 1] dtype = "float32" min_val = float("-0.46837") @@ -29,6 +32,7 @@ class Program_weight_tensor_parameter_2: class Program_weight_tensor_parameter_3: name = "parameter_3" + original_name = "batch_norm_46.b_0" shape = [720] dtype = "float32" min_val = float("-8.57225") @@ -40,6 +44,7 @@ class Program_weight_tensor_parameter_3: class Program_weight_tensor_parameter_4: name = "parameter_4" + original_name = "batch_norm_46.w_0" shape = [720] dtype = "float32" min_val = float("1.65564") @@ -51,6 +56,7 @@ class Program_weight_tensor_parameter_4: class Program_weight_tensor_parameter_5: name = "parameter_5" + original_name = "batch_norm_46.w_2" shape = [720] dtype = "float32" min_val = float("18.722") @@ -62,6 +68,7 @@ class Program_weight_tensor_parameter_5: class Program_weight_tensor_parameter_6: name = "parameter_6" + original_name = "batch_norm_46.w_1" shape = [720] dtype = "float32" min_val = float("-1.0648e-05") @@ -73,6 +80,7 @@ class Program_weight_tensor_parameter_6: class Program_weight_tensor_parameter_7: name = "parameter_7" + original_name = "conv2d_62.w_0" shape = [720, 120, 1, 1] dtype = "float32" min_val = float("-0.695536") @@ -84,6 +92,7 @@ class Program_weight_tensor_parameter_7: class Program_weight_tensor_parameter_8: name = "parameter_8" + original_name = "batch_norm_45.b_0" shape = [120] dtype = "float32" min_val = float("-2.35481e-06") @@ -95,6 +104,7 @@ class Program_weight_tensor_parameter_8: class Program_weight_tensor_parameter_9: name = "parameter_9" + original_name = "batch_norm_45.w_0" shape = [120] dtype = "float32" min_val = float("1.81269") @@ -106,6 +116,7 @@ class Program_weight_tensor_parameter_9: class Program_weight_tensor_parameter_10: name = "parameter_10" + original_name = "batch_norm_45.w_2" shape = [120] dtype = "float32" min_val = float("0.0430573") @@ -117,6 +128,7 @@ class Program_weight_tensor_parameter_10: class Program_weight_tensor_parameter_11: name = "parameter_11" + original_name = "batch_norm_45.w_1" shape = [120] dtype = "float32" min_val = float("-0.339458") @@ -128,6 +140,7 @@ class Program_weight_tensor_parameter_11: class Program_weight_tensor_parameter_12: name = "parameter_12" + original_name = "conv2d_61.w_0" shape = [120, 720, 1, 1] dtype = "float32" min_val = float("-0.294698") @@ -139,6 +152,7 @@ class Program_weight_tensor_parameter_12: class Program_weight_tensor_parameter_13: name = "parameter_13" + original_name = "conv2d_60.b_0" shape = [720] dtype = "float32" min_val = float("-0.863123") @@ -150,6 +164,7 @@ class Program_weight_tensor_parameter_13: class Program_weight_tensor_parameter_14: name = "parameter_14" + original_name = "conv2d_60.w_0" shape = [720, 180, 1, 1] dtype = "float32" min_val = float("-0.634716") @@ -161,6 +176,7 @@ class Program_weight_tensor_parameter_14: class Program_weight_tensor_parameter_15: name = "parameter_15" + original_name = "conv2d_59.b_0" shape = [180] dtype = "float32" min_val = float("-0.084701") @@ -172,6 +188,7 @@ class Program_weight_tensor_parameter_15: class Program_weight_tensor_parameter_16: name = "parameter_16" + original_name = "conv2d_59.w_0" shape = [180, 720, 1, 1] dtype = "float32" min_val = float("-0.47523") @@ -183,6 +200,7 @@ class Program_weight_tensor_parameter_16: class Program_weight_tensor_parameter_17: name = "parameter_17" + original_name = "batch_norm_44.b_0" shape = [720] dtype = "float32" min_val = float("-5.46876") @@ -194,6 +212,7 @@ class Program_weight_tensor_parameter_17: class Program_weight_tensor_parameter_18: name = "parameter_18" + original_name = "batch_norm_44.w_0" shape = [720] dtype = "float32" min_val = float("0.195235") @@ -205,6 +224,7 @@ class Program_weight_tensor_parameter_18: class Program_weight_tensor_parameter_19: name = "parameter_19" + original_name = "batch_norm_44.w_2" shape = [720] dtype = "float32" min_val = float("0.000110481") @@ -216,6 +236,7 @@ class Program_weight_tensor_parameter_19: class Program_weight_tensor_parameter_20: name = "parameter_20" + original_name = "batch_norm_44.w_1" shape = [720] dtype = "float32" min_val = float("-3.7083") @@ -227,6 +248,7 @@ class Program_weight_tensor_parameter_20: class Program_weight_tensor_parameter_21: name = "parameter_21" + original_name = "conv2d_58.w_0" shape = [720, 1, 5, 5] dtype = "float32" min_val = float("-0.641292") @@ -238,6 +260,7 @@ class Program_weight_tensor_parameter_21: class Program_weight_tensor_parameter_22: name = "parameter_22" + original_name = "batch_norm_43.b_0" shape = [720] dtype = "float32" min_val = float("-4.01612") @@ -249,6 +272,7 @@ class Program_weight_tensor_parameter_22: class Program_weight_tensor_parameter_23: name = "parameter_23" + original_name = "batch_norm_43.w_0" shape = [720] dtype = "float32" min_val = float("-0.940435") @@ -260,6 +284,7 @@ class Program_weight_tensor_parameter_23: class Program_weight_tensor_parameter_24: name = "parameter_24" + original_name = "batch_norm_43.w_2" shape = [720] dtype = "float32" min_val = float("5.02264") @@ -271,6 +296,7 @@ class Program_weight_tensor_parameter_24: class Program_weight_tensor_parameter_25: name = "parameter_25" + original_name = "batch_norm_43.w_1" shape = [720] dtype = "float32" min_val = float("-8.49782e-06") @@ -282,6 +308,7 @@ class Program_weight_tensor_parameter_25: class Program_weight_tensor_parameter_26: name = "parameter_26" + original_name = "conv2d_57.w_0" shape = [720, 120, 1, 1] dtype = "float32" min_val = float("-0.396064") @@ -293,6 +320,7 @@ class Program_weight_tensor_parameter_26: class Program_weight_tensor_parameter_27: name = "parameter_27" + original_name = "batch_norm_42.b_0" shape = [120] dtype = "float32" min_val = float("-2.50983e-06") @@ -304,6 +332,7 @@ class Program_weight_tensor_parameter_27: class Program_weight_tensor_parameter_28: name = "parameter_28" + original_name = "batch_norm_42.w_0" shape = [120] dtype = "float32" min_val = float("0.975457") @@ -315,6 +344,7 @@ class Program_weight_tensor_parameter_28: class Program_weight_tensor_parameter_29: name = "parameter_29" + original_name = "batch_norm_42.w_2" shape = [120] dtype = "float32" min_val = float("0.0517678") @@ -326,6 +356,7 @@ class Program_weight_tensor_parameter_29: class Program_weight_tensor_parameter_30: name = "parameter_30" + original_name = "batch_norm_42.w_1" shape = [120] dtype = "float32" min_val = float("-0.296751") @@ -337,6 +368,7 @@ class Program_weight_tensor_parameter_30: class Program_weight_tensor_parameter_31: name = "parameter_31" + original_name = "conv2d_56.w_0" shape = [120, 720, 1, 1] dtype = "float32" min_val = float("-0.330074") @@ -348,6 +380,7 @@ class Program_weight_tensor_parameter_31: class Program_weight_tensor_parameter_32: name = "parameter_32" + original_name = "conv2d_55.b_0" shape = [720] dtype = "float32" min_val = float("-0.233065") @@ -359,6 +392,7 @@ class Program_weight_tensor_parameter_32: class Program_weight_tensor_parameter_33: name = "parameter_33" + original_name = "conv2d_55.w_0" shape = [720, 180, 1, 1] dtype = "float32" min_val = float("-0.473329") @@ -370,6 +404,7 @@ class Program_weight_tensor_parameter_33: class Program_weight_tensor_parameter_34: name = "parameter_34" + original_name = "conv2d_54.b_0" shape = [180] dtype = "float32" min_val = float("-0.052022") @@ -381,6 +416,7 @@ class Program_weight_tensor_parameter_34: class Program_weight_tensor_parameter_35: name = "parameter_35" + original_name = "conv2d_54.w_0" shape = [180, 720, 1, 1] dtype = "float32" min_val = float("-0.410206") @@ -392,6 +428,7 @@ class Program_weight_tensor_parameter_35: class Program_weight_tensor_parameter_36: name = "parameter_36" + original_name = "batch_norm_41.b_0" shape = [720] dtype = "float32" min_val = float("-3.50398") @@ -403,6 +440,7 @@ class Program_weight_tensor_parameter_36: class Program_weight_tensor_parameter_37: name = "parameter_37" + original_name = "batch_norm_41.w_0" shape = [720] dtype = "float32" min_val = float("0.206224") @@ -414,6 +452,7 @@ class Program_weight_tensor_parameter_37: class Program_weight_tensor_parameter_38: name = "parameter_38" + original_name = "batch_norm_41.w_2" shape = [720] dtype = "float32" min_val = float("0.000122239") @@ -425,6 +464,7 @@ class Program_weight_tensor_parameter_38: class Program_weight_tensor_parameter_39: name = "parameter_39" + original_name = "batch_norm_41.w_1" shape = [720] dtype = "float32" min_val = float("-1.91099") @@ -436,6 +476,7 @@ class Program_weight_tensor_parameter_39: class Program_weight_tensor_parameter_40: name = "parameter_40" + original_name = "conv2d_53.w_0" shape = [720, 1, 5, 5] dtype = "float32" min_val = float("-0.607011") @@ -447,6 +488,7 @@ class Program_weight_tensor_parameter_40: class Program_weight_tensor_parameter_41: name = "parameter_41" + original_name = "batch_norm_40.b_0" shape = [720] dtype = "float32" min_val = float("-3.19753") @@ -458,6 +500,7 @@ class Program_weight_tensor_parameter_41: class Program_weight_tensor_parameter_42: name = "parameter_42" + original_name = "batch_norm_40.w_0" shape = [720] dtype = "float32" min_val = float("-0.0661638") @@ -469,6 +512,7 @@ class Program_weight_tensor_parameter_42: class Program_weight_tensor_parameter_43: name = "parameter_43" + original_name = "batch_norm_40.w_2" shape = [720] dtype = "float32" min_val = float("3.94119") @@ -480,6 +524,7 @@ class Program_weight_tensor_parameter_43: class Program_weight_tensor_parameter_44: name = "parameter_44" + original_name = "batch_norm_40.w_1" shape = [720] dtype = "float32" min_val = float("-3.78814e-06") @@ -491,6 +536,7 @@ class Program_weight_tensor_parameter_44: class Program_weight_tensor_parameter_45: name = "parameter_45" + original_name = "conv2d_52.w_0" shape = [720, 120, 1, 1] dtype = "float32" min_val = float("-0.347929") @@ -502,6 +548,7 @@ class Program_weight_tensor_parameter_45: class Program_weight_tensor_parameter_46: name = "parameter_46" + original_name = "batch_norm_39.b_0" shape = [120] dtype = "float32" min_val = float("-2.34769e-06") @@ -513,6 +560,7 @@ class Program_weight_tensor_parameter_46: class Program_weight_tensor_parameter_47: name = "parameter_47" + original_name = "batch_norm_39.w_0" shape = [120] dtype = "float32" min_val = float("2.53557") @@ -524,6 +572,7 @@ class Program_weight_tensor_parameter_47: class Program_weight_tensor_parameter_48: name = "parameter_48" + original_name = "batch_norm_39.w_2" shape = [120] dtype = "float32" min_val = float("0.149443") @@ -535,6 +584,7 @@ class Program_weight_tensor_parameter_48: class Program_weight_tensor_parameter_49: name = "parameter_49" + original_name = "batch_norm_39.w_1" shape = [120] dtype = "float32" min_val = float("-0.556891") @@ -546,6 +596,7 @@ class Program_weight_tensor_parameter_49: class Program_weight_tensor_parameter_50: name = "parameter_50" + original_name = "conv2d_51.w_0" shape = [120, 504, 1, 1] dtype = "float32" min_val = float("-0.635785") @@ -557,6 +608,7 @@ class Program_weight_tensor_parameter_50: class Program_weight_tensor_parameter_51: name = "parameter_51" + original_name = "conv2d_50.b_0" shape = [504] dtype = "float32" min_val = float("-0.294356") @@ -568,6 +620,7 @@ class Program_weight_tensor_parameter_51: class Program_weight_tensor_parameter_52: name = "parameter_52" + original_name = "conv2d_50.w_0" shape = [504, 126, 1, 1] dtype = "float32" min_val = float("-0.690762") @@ -579,6 +632,7 @@ class Program_weight_tensor_parameter_52: class Program_weight_tensor_parameter_53: name = "parameter_53" + original_name = "conv2d_49.b_0" shape = [126] dtype = "float32" min_val = float("-0.0453208") @@ -590,6 +644,7 @@ class Program_weight_tensor_parameter_53: class Program_weight_tensor_parameter_54: name = "parameter_54" + original_name = "conv2d_49.w_0" shape = [126, 504, 1, 1] dtype = "float32" min_val = float("-0.523911") @@ -601,6 +656,7 @@ class Program_weight_tensor_parameter_54: class Program_weight_tensor_parameter_55: name = "parameter_55" + original_name = "batch_norm_38.b_0" shape = [504] dtype = "float32" min_val = float("-2.10228") @@ -612,6 +668,7 @@ class Program_weight_tensor_parameter_55: class Program_weight_tensor_parameter_56: name = "parameter_56" + original_name = "batch_norm_38.w_0" shape = [504] dtype = "float32" min_val = float("0.287653") @@ -623,6 +680,7 @@ class Program_weight_tensor_parameter_56: class Program_weight_tensor_parameter_57: name = "parameter_57" + original_name = "batch_norm_38.w_2" shape = [504] dtype = "float32" min_val = float("0.000113446") @@ -634,6 +692,7 @@ class Program_weight_tensor_parameter_57: class Program_weight_tensor_parameter_58: name = "parameter_58" + original_name = "batch_norm_38.w_1" shape = [504] dtype = "float32" min_val = float("-5.22728") @@ -645,6 +704,7 @@ class Program_weight_tensor_parameter_58: class Program_weight_tensor_parameter_59: name = "parameter_59" + original_name = "conv2d_48.w_0" shape = [504, 1, 5, 5] dtype = "float32" min_val = float("-0.398182") @@ -656,6 +716,7 @@ class Program_weight_tensor_parameter_59: class Program_weight_tensor_parameter_60: name = "parameter_60" + original_name = "batch_norm_37.b_0" shape = [504] dtype = "float32" min_val = float("-3.59078") @@ -667,6 +728,7 @@ class Program_weight_tensor_parameter_60: class Program_weight_tensor_parameter_61: name = "parameter_61" + original_name = "batch_norm_37.w_0" shape = [504] dtype = "float32" min_val = float("-1.20024") @@ -678,6 +740,7 @@ class Program_weight_tensor_parameter_61: class Program_weight_tensor_parameter_62: name = "parameter_62" + original_name = "batch_norm_37.w_2" shape = [504] dtype = "float32" min_val = float("7.20351") @@ -689,6 +752,7 @@ class Program_weight_tensor_parameter_62: class Program_weight_tensor_parameter_63: name = "parameter_63" + original_name = "batch_norm_37.w_1" shape = [504] dtype = "float32" min_val = float("-4.87624e-06") @@ -700,6 +764,7 @@ class Program_weight_tensor_parameter_63: class Program_weight_tensor_parameter_64: name = "parameter_64" + original_name = "conv2d_47.w_0" shape = [504, 88, 1, 1] dtype = "float32" min_val = float("-0.548745") @@ -711,6 +776,7 @@ class Program_weight_tensor_parameter_64: class Program_weight_tensor_parameter_65: name = "parameter_65" + original_name = "batch_norm_36.b_0" shape = [88] dtype = "float32" min_val = float("-1.27936e-06") @@ -722,6 +788,7 @@ class Program_weight_tensor_parameter_65: class Program_weight_tensor_parameter_66: name = "parameter_66" + original_name = "batch_norm_36.w_0" shape = [88] dtype = "float32" min_val = float("1.03318") @@ -733,6 +800,7 @@ class Program_weight_tensor_parameter_66: class Program_weight_tensor_parameter_67: name = "parameter_67" + original_name = "batch_norm_36.w_2" shape = [88] dtype = "float32" min_val = float("0.0287246") @@ -744,6 +812,7 @@ class Program_weight_tensor_parameter_67: class Program_weight_tensor_parameter_68: name = "parameter_68" + original_name = "batch_norm_36.w_1" shape = [88] dtype = "float32" min_val = float("-0.303803") @@ -755,6 +824,7 @@ class Program_weight_tensor_parameter_68: class Program_weight_tensor_parameter_69: name = "parameter_69" + original_name = "conv2d_46.w_0" shape = [88, 504, 1, 1] dtype = "float32" min_val = float("-0.463461") @@ -766,6 +836,7 @@ class Program_weight_tensor_parameter_69: class Program_weight_tensor_parameter_70: name = "parameter_70" + original_name = "conv2d_45.b_0" shape = [504] dtype = "float32" min_val = float("-0.218766") @@ -777,6 +848,7 @@ class Program_weight_tensor_parameter_70: class Program_weight_tensor_parameter_71: name = "parameter_71" + original_name = "conv2d_45.w_0" shape = [504, 126, 1, 1] dtype = "float32" min_val = float("-0.643379") @@ -788,6 +860,7 @@ class Program_weight_tensor_parameter_71: class Program_weight_tensor_parameter_72: name = "parameter_72" + original_name = "conv2d_44.b_0" shape = [126] dtype = "float32" min_val = float("-0.0197441") @@ -799,6 +872,7 @@ class Program_weight_tensor_parameter_72: class Program_weight_tensor_parameter_73: name = "parameter_73" + original_name = "conv2d_44.w_0" shape = [126, 504, 1, 1] dtype = "float32" min_val = float("-0.797233") @@ -810,6 +884,7 @@ class Program_weight_tensor_parameter_73: class Program_weight_tensor_parameter_74: name = "parameter_74" + original_name = "batch_norm_35.b_0" shape = [504] dtype = "float32" min_val = float("-4.19902") @@ -821,6 +896,7 @@ class Program_weight_tensor_parameter_74: class Program_weight_tensor_parameter_75: name = "parameter_75" + original_name = "batch_norm_35.w_0" shape = [504] dtype = "float32" min_val = float("0.20254") @@ -832,6 +908,7 @@ class Program_weight_tensor_parameter_75: class Program_weight_tensor_parameter_76: name = "parameter_76" + original_name = "batch_norm_35.w_2" shape = [504] dtype = "float32" min_val = float("0.000115978") @@ -843,6 +920,7 @@ class Program_weight_tensor_parameter_76: class Program_weight_tensor_parameter_77: name = "parameter_77" + original_name = "batch_norm_35.w_1" shape = [504] dtype = "float32" min_val = float("-3.03334") @@ -854,6 +932,7 @@ class Program_weight_tensor_parameter_77: class Program_weight_tensor_parameter_78: name = "parameter_78" + original_name = "conv2d_43.w_0" shape = [504, 1, 3, 3] dtype = "float32" min_val = float("-0.636408") @@ -865,6 +944,7 @@ class Program_weight_tensor_parameter_78: class Program_weight_tensor_parameter_79: name = "parameter_79" + original_name = "batch_norm_34.b_0" shape = [504] dtype = "float32" min_val = float("-2.84548") @@ -876,6 +956,7 @@ class Program_weight_tensor_parameter_79: class Program_weight_tensor_parameter_80: name = "parameter_80" + original_name = "batch_norm_34.w_0" shape = [504] dtype = "float32" min_val = float("-0.720555") @@ -887,6 +968,7 @@ class Program_weight_tensor_parameter_80: class Program_weight_tensor_parameter_81: name = "parameter_81" + original_name = "batch_norm_34.w_2" shape = [504] dtype = "float32" min_val = float("3.94906") @@ -898,6 +980,7 @@ class Program_weight_tensor_parameter_81: class Program_weight_tensor_parameter_82: name = "parameter_82" + original_name = "batch_norm_34.w_1" shape = [504] dtype = "float32" min_val = float("-2.17694e-06") @@ -909,6 +992,7 @@ class Program_weight_tensor_parameter_82: class Program_weight_tensor_parameter_83: name = "parameter_83" + original_name = "conv2d_42.w_0" shape = [504, 88, 1, 1] dtype = "float32" min_val = float("-0.395402") @@ -920,6 +1004,7 @@ class Program_weight_tensor_parameter_83: class Program_weight_tensor_parameter_84: name = "parameter_84" + original_name = "batch_norm_33.b_0" shape = [88] dtype = "float32" min_val = float("-2.58721e-06") @@ -931,6 +1016,7 @@ class Program_weight_tensor_parameter_84: class Program_weight_tensor_parameter_85: name = "parameter_85" + original_name = "batch_norm_33.w_0" shape = [88] dtype = "float32" min_val = float("2.50744") @@ -942,6 +1028,7 @@ class Program_weight_tensor_parameter_85: class Program_weight_tensor_parameter_86: name = "parameter_86" + original_name = "batch_norm_33.w_2" shape = [88] dtype = "float32" min_val = float("0.0985202") @@ -953,6 +1040,7 @@ class Program_weight_tensor_parameter_86: class Program_weight_tensor_parameter_87: name = "parameter_87" + original_name = "batch_norm_33.w_1" shape = [88] dtype = "float32" min_val = float("-0.597276") @@ -964,6 +1052,7 @@ class Program_weight_tensor_parameter_87: class Program_weight_tensor_parameter_88: name = "parameter_88" + original_name = "conv2d_41.w_0" shape = [88, 360, 1, 1] dtype = "float32" min_val = float("-0.742602") @@ -975,6 +1064,7 @@ class Program_weight_tensor_parameter_88: class Program_weight_tensor_parameter_89: name = "parameter_89" + original_name = "conv2d_40.b_0" shape = [360] dtype = "float32" min_val = float("-0.331864") @@ -986,6 +1076,7 @@ class Program_weight_tensor_parameter_89: class Program_weight_tensor_parameter_90: name = "parameter_90" + original_name = "conv2d_40.w_0" shape = [360, 90, 1, 1] dtype = "float32" min_val = float("-0.826345") @@ -997,6 +1088,7 @@ class Program_weight_tensor_parameter_90: class Program_weight_tensor_parameter_91: name = "parameter_91" + original_name = "conv2d_39.b_0" shape = [90] dtype = "float32" min_val = float("-0.0761577") @@ -1008,6 +1100,7 @@ class Program_weight_tensor_parameter_91: class Program_weight_tensor_parameter_92: name = "parameter_92" + original_name = "conv2d_39.w_0" shape = [90, 360, 1, 1] dtype = "float32" min_val = float("-1.03586") @@ -1019,6 +1112,7 @@ class Program_weight_tensor_parameter_92: class Program_weight_tensor_parameter_93: name = "parameter_93" + original_name = "batch_norm_32.b_0" shape = [360] dtype = "float32" min_val = float("-1.99234") @@ -1030,6 +1124,7 @@ class Program_weight_tensor_parameter_93: class Program_weight_tensor_parameter_94: name = "parameter_94" + original_name = "batch_norm_32.w_0" shape = [360] dtype = "float32" min_val = float("0.309052") @@ -1041,6 +1136,7 @@ class Program_weight_tensor_parameter_94: class Program_weight_tensor_parameter_95: name = "parameter_95" + original_name = "batch_norm_32.w_2" shape = [360] dtype = "float32" min_val = float("0.00012227") @@ -1052,6 +1148,7 @@ class Program_weight_tensor_parameter_95: class Program_weight_tensor_parameter_96: name = "parameter_96" + original_name = "batch_norm_32.w_1" shape = [360] dtype = "float32" min_val = float("-3.97308") @@ -1063,6 +1160,7 @@ class Program_weight_tensor_parameter_96: class Program_weight_tensor_parameter_97: name = "parameter_97" + original_name = "conv2d_38.w_0" shape = [360, 1, 3, 3] dtype = "float32" min_val = float("-1.05193") @@ -1074,6 +1172,7 @@ class Program_weight_tensor_parameter_97: class Program_weight_tensor_parameter_98: name = "parameter_98" + original_name = "batch_norm_31.b_0" shape = [360] dtype = "float32" min_val = float("-2.33837") @@ -1085,6 +1184,7 @@ class Program_weight_tensor_parameter_98: class Program_weight_tensor_parameter_99: name = "parameter_99" + original_name = "batch_norm_31.w_0" shape = [360] dtype = "float32" min_val = float("-1.84226") @@ -1096,6 +1196,7 @@ class Program_weight_tensor_parameter_99: class Program_weight_tensor_parameter_100: name = "parameter_100" + original_name = "batch_norm_31.w_2" shape = [360] dtype = "float32" min_val = float("11.8328") @@ -1107,6 +1208,7 @@ class Program_weight_tensor_parameter_100: class Program_weight_tensor_parameter_101: name = "parameter_101" + original_name = "batch_norm_31.w_1" shape = [360] dtype = "float32" min_val = float("-7.75129e-06") @@ -1118,6 +1220,7 @@ class Program_weight_tensor_parameter_101: class Program_weight_tensor_parameter_102: name = "parameter_102" + original_name = "conv2d_37.w_0" shape = [360, 64, 1, 1] dtype = "float32" min_val = float("-0.710854") @@ -1129,6 +1232,7 @@ class Program_weight_tensor_parameter_102: class Program_weight_tensor_parameter_103: name = "parameter_103" + original_name = "batch_norm_30.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -1138,6 +1242,7 @@ class Program_weight_tensor_parameter_103: class Program_weight_tensor_parameter_104: name = "parameter_104" + original_name = "batch_norm_30.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -1147,6 +1252,7 @@ class Program_weight_tensor_parameter_104: class Program_weight_tensor_parameter_105: name = "parameter_105" + original_name = "batch_norm_30.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -1156,6 +1262,7 @@ class Program_weight_tensor_parameter_105: class Program_weight_tensor_parameter_106: name = "parameter_106" + original_name = "batch_norm_30.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -1165,6 +1272,7 @@ class Program_weight_tensor_parameter_106: class Program_weight_tensor_parameter_107: name = "parameter_107" + original_name = "conv2d_36.w_0" shape = [64, 136, 1, 1] dtype = "float32" min_val = float("-0.585797") @@ -1176,6 +1284,7 @@ class Program_weight_tensor_parameter_107: class Program_weight_tensor_parameter_108: name = "parameter_108" + original_name = "batch_norm_29.b_0" shape = [136] dtype = "float32" min_val = float("-2.68023") @@ -1187,6 +1296,7 @@ class Program_weight_tensor_parameter_108: class Program_weight_tensor_parameter_109: name = "parameter_109" + original_name = "batch_norm_29.w_0" shape = [136] dtype = "float32" min_val = float("0.30344") @@ -1198,6 +1308,7 @@ class Program_weight_tensor_parameter_109: class Program_weight_tensor_parameter_110: name = "parameter_110" + original_name = "batch_norm_29.w_2" shape = [136] dtype = "float32" min_val = float("0.000108379") @@ -1209,6 +1320,7 @@ class Program_weight_tensor_parameter_110: class Program_weight_tensor_parameter_111: name = "parameter_111" + original_name = "batch_norm_29.w_1" shape = [136] dtype = "float32" min_val = float("-2.2265") @@ -1220,6 +1332,7 @@ class Program_weight_tensor_parameter_111: class Program_weight_tensor_parameter_112: name = "parameter_112" + original_name = "conv2d_35.w_0" shape = [136, 1, 3, 3] dtype = "float32" min_val = float("-0.561665") @@ -1231,6 +1344,7 @@ class Program_weight_tensor_parameter_112: class Program_weight_tensor_parameter_113: name = "parameter_113" + original_name = "batch_norm_28.b_0" shape = [136] dtype = "float32" min_val = float("-2.53786") @@ -1242,6 +1356,7 @@ class Program_weight_tensor_parameter_113: class Program_weight_tensor_parameter_114: name = "parameter_114" + original_name = "batch_norm_28.w_0" shape = [136] dtype = "float32" min_val = float("-0.899339") @@ -1253,6 +1368,7 @@ class Program_weight_tensor_parameter_114: class Program_weight_tensor_parameter_115: name = "parameter_115" + original_name = "batch_norm_28.w_2" shape = [136] dtype = "float32" min_val = float("6.72388") @@ -1264,6 +1380,7 @@ class Program_weight_tensor_parameter_115: class Program_weight_tensor_parameter_116: name = "parameter_116" + original_name = "batch_norm_28.w_1" shape = [136] dtype = "float32" min_val = float("-4.49021e-06") @@ -1275,6 +1392,7 @@ class Program_weight_tensor_parameter_116: class Program_weight_tensor_parameter_117: name = "parameter_117" + original_name = "conv2d_34.w_0" shape = [136, 64, 1, 1] dtype = "float32" min_val = float("-0.859069") @@ -1286,6 +1404,7 @@ class Program_weight_tensor_parameter_117: class Program_weight_tensor_parameter_118: name = "parameter_118" + original_name = "batch_norm_27.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -1295,6 +1414,7 @@ class Program_weight_tensor_parameter_118: class Program_weight_tensor_parameter_119: name = "parameter_119" + original_name = "batch_norm_27.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -1304,6 +1424,7 @@ class Program_weight_tensor_parameter_119: class Program_weight_tensor_parameter_120: name = "parameter_120" + original_name = "batch_norm_27.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -1313,6 +1434,7 @@ class Program_weight_tensor_parameter_120: class Program_weight_tensor_parameter_121: name = "parameter_121" + original_name = "batch_norm_27.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -1322,6 +1444,7 @@ class Program_weight_tensor_parameter_121: class Program_weight_tensor_parameter_122: name = "parameter_122" + original_name = "conv2d_33.w_0" shape = [64, 136, 1, 1] dtype = "float32" min_val = float("-0.628194") @@ -1333,6 +1456,7 @@ class Program_weight_tensor_parameter_122: class Program_weight_tensor_parameter_123: name = "parameter_123" + original_name = "batch_norm_26.b_0" shape = [136] dtype = "float32" min_val = float("-2.18304") @@ -1344,6 +1468,7 @@ class Program_weight_tensor_parameter_123: class Program_weight_tensor_parameter_124: name = "parameter_124" + original_name = "batch_norm_26.w_0" shape = [136] dtype = "float32" min_val = float("0.149736") @@ -1355,6 +1480,7 @@ class Program_weight_tensor_parameter_124: class Program_weight_tensor_parameter_125: name = "parameter_125" + original_name = "batch_norm_26.w_2" shape = [136] dtype = "float32" min_val = float("6.71884e-05") @@ -1366,6 +1492,7 @@ class Program_weight_tensor_parameter_125: class Program_weight_tensor_parameter_126: name = "parameter_126" + original_name = "batch_norm_26.w_1" shape = [136] dtype = "float32" min_val = float("-2.7798") @@ -1377,6 +1504,7 @@ class Program_weight_tensor_parameter_126: class Program_weight_tensor_parameter_127: name = "parameter_127" + original_name = "conv2d_32.w_0" shape = [136, 1, 3, 3] dtype = "float32" min_val = float("-0.578846") @@ -1388,6 +1516,7 @@ class Program_weight_tensor_parameter_127: class Program_weight_tensor_parameter_128: name = "parameter_128" + original_name = "batch_norm_25.b_0" shape = [136] dtype = "float32" min_val = float("-1.95157") @@ -1399,6 +1528,7 @@ class Program_weight_tensor_parameter_128: class Program_weight_tensor_parameter_129: name = "parameter_129" + original_name = "batch_norm_25.w_0" shape = [136] dtype = "float32" min_val = float("-0.527812") @@ -1410,6 +1540,7 @@ class Program_weight_tensor_parameter_129: class Program_weight_tensor_parameter_130: name = "parameter_130" + original_name = "batch_norm_25.w_2" shape = [136] dtype = "float32" min_val = float("8.10774") @@ -1421,6 +1552,7 @@ class Program_weight_tensor_parameter_130: class Program_weight_tensor_parameter_131: name = "parameter_131" + original_name = "batch_norm_25.w_1" shape = [136] dtype = "float32" min_val = float("-3.98717e-06") @@ -1432,6 +1564,7 @@ class Program_weight_tensor_parameter_131: class Program_weight_tensor_parameter_132: name = "parameter_132" + original_name = "conv2d_31.w_0" shape = [136, 64, 1, 1] dtype = "float32" min_val = float("-0.861406") @@ -1443,6 +1576,7 @@ class Program_weight_tensor_parameter_132: class Program_weight_tensor_parameter_133: name = "parameter_133" + original_name = "batch_norm_24.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -1452,6 +1586,7 @@ class Program_weight_tensor_parameter_133: class Program_weight_tensor_parameter_134: name = "parameter_134" + original_name = "batch_norm_24.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -1461,6 +1596,7 @@ class Program_weight_tensor_parameter_134: class Program_weight_tensor_parameter_135: name = "parameter_135" + original_name = "batch_norm_24.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -1470,6 +1606,7 @@ class Program_weight_tensor_parameter_135: class Program_weight_tensor_parameter_136: name = "parameter_136" + original_name = "batch_norm_24.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -1479,6 +1616,7 @@ class Program_weight_tensor_parameter_136: class Program_weight_tensor_parameter_137: name = "parameter_137" + original_name = "conv2d_30.w_0" shape = [64, 152, 1, 1] dtype = "float32" min_val = float("-0.734319") @@ -1490,6 +1628,7 @@ class Program_weight_tensor_parameter_137: class Program_weight_tensor_parameter_138: name = "parameter_138" + original_name = "batch_norm_23.b_0" shape = [152] dtype = "float32" min_val = float("-2.87089") @@ -1501,6 +1640,7 @@ class Program_weight_tensor_parameter_138: class Program_weight_tensor_parameter_139: name = "parameter_139" + original_name = "batch_norm_23.w_0" shape = [152] dtype = "float32" min_val = float("0.0941322") @@ -1512,6 +1652,7 @@ class Program_weight_tensor_parameter_139: class Program_weight_tensor_parameter_140: name = "parameter_140" + original_name = "batch_norm_23.w_2" shape = [152] dtype = "float32" min_val = float("8.26851e-05") @@ -1523,6 +1664,7 @@ class Program_weight_tensor_parameter_140: class Program_weight_tensor_parameter_141: name = "parameter_141" + original_name = "batch_norm_23.w_1" shape = [152] dtype = "float32" min_val = float("-9.86311") @@ -1534,6 +1676,7 @@ class Program_weight_tensor_parameter_141: class Program_weight_tensor_parameter_142: name = "parameter_142" + original_name = "conv2d_29.w_0" shape = [152, 1, 3, 3] dtype = "float32" min_val = float("-0.636663") @@ -1545,6 +1688,7 @@ class Program_weight_tensor_parameter_142: class Program_weight_tensor_parameter_143: name = "parameter_143" + original_name = "batch_norm_22.b_0" shape = [152] dtype = "float32" min_val = float("-2.60576") @@ -1556,6 +1700,7 @@ class Program_weight_tensor_parameter_143: class Program_weight_tensor_parameter_144: name = "parameter_144" + original_name = "batch_norm_22.w_0" shape = [152] dtype = "float32" min_val = float("-0.230744") @@ -1567,6 +1712,7 @@ class Program_weight_tensor_parameter_144: class Program_weight_tensor_parameter_145: name = "parameter_145" + original_name = "batch_norm_22.w_2" shape = [152] dtype = "float32" min_val = float("6.37535") @@ -1578,6 +1724,7 @@ class Program_weight_tensor_parameter_145: class Program_weight_tensor_parameter_146: name = "parameter_146" + original_name = "batch_norm_22.w_1" shape = [152] dtype = "float32" min_val = float("-2.24098e-06") @@ -1589,6 +1736,7 @@ class Program_weight_tensor_parameter_146: class Program_weight_tensor_parameter_147: name = "parameter_147" + original_name = "conv2d_28.w_0" shape = [152, 64, 1, 1] dtype = "float32" min_val = float("-0.557448") @@ -1600,6 +1748,7 @@ class Program_weight_tensor_parameter_147: class Program_weight_tensor_parameter_148: name = "parameter_148" + original_name = "batch_norm_21.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -1609,6 +1758,7 @@ class Program_weight_tensor_parameter_148: class Program_weight_tensor_parameter_149: name = "parameter_149" + original_name = "batch_norm_21.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -1618,6 +1768,7 @@ class Program_weight_tensor_parameter_149: class Program_weight_tensor_parameter_150: name = "parameter_150" + original_name = "batch_norm_21.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -1627,6 +1778,7 @@ class Program_weight_tensor_parameter_150: class Program_weight_tensor_parameter_151: name = "parameter_151" + original_name = "batch_norm_21.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -1636,6 +1788,7 @@ class Program_weight_tensor_parameter_151: class Program_weight_tensor_parameter_152: name = "parameter_152" + original_name = "conv2d_27.w_0" shape = [64, 184, 1, 1] dtype = "float32" min_val = float("-1.0319") @@ -1647,6 +1800,7 @@ class Program_weight_tensor_parameter_152: class Program_weight_tensor_parameter_153: name = "parameter_153" + original_name = "batch_norm_20.b_0" shape = [184] dtype = "float32" min_val = float("-1.2342") @@ -1658,6 +1812,7 @@ class Program_weight_tensor_parameter_153: class Program_weight_tensor_parameter_154: name = "parameter_154" + original_name = "batch_norm_20.w_0" shape = [184] dtype = "float32" min_val = float("0.370526") @@ -1669,6 +1824,7 @@ class Program_weight_tensor_parameter_154: class Program_weight_tensor_parameter_155: name = "parameter_155" + original_name = "batch_norm_20.w_2" shape = [184] dtype = "float32" min_val = float("3.94438e-05") @@ -1680,6 +1836,7 @@ class Program_weight_tensor_parameter_155: class Program_weight_tensor_parameter_156: name = "parameter_156" + original_name = "batch_norm_20.w_1" shape = [184] dtype = "float32" min_val = float("-3.06516") @@ -1691,6 +1848,7 @@ class Program_weight_tensor_parameter_156: class Program_weight_tensor_parameter_157: name = "parameter_157" + original_name = "conv2d_26.w_0" shape = [184, 1, 3, 3] dtype = "float32" min_val = float("-0.447293") @@ -1702,6 +1860,7 @@ class Program_weight_tensor_parameter_157: class Program_weight_tensor_parameter_158: name = "parameter_158" + original_name = "batch_norm_19.b_0" shape = [184] dtype = "float32" min_val = float("-2.53947") @@ -1713,6 +1872,7 @@ class Program_weight_tensor_parameter_158: class Program_weight_tensor_parameter_159: name = "parameter_159" + original_name = "batch_norm_19.w_0" shape = [184] dtype = "float32" min_val = float("0.0382549") @@ -1724,6 +1884,7 @@ class Program_weight_tensor_parameter_159: class Program_weight_tensor_parameter_160: name = "parameter_160" + original_name = "batch_norm_19.w_2" shape = [184] dtype = "float32" min_val = float("12.9524") @@ -1735,6 +1896,7 @@ class Program_weight_tensor_parameter_160: class Program_weight_tensor_parameter_161: name = "parameter_161" + original_name = "batch_norm_19.w_1" shape = [184] dtype = "float32" min_val = float("-2.05619e-05") @@ -1746,6 +1908,7 @@ class Program_weight_tensor_parameter_161: class Program_weight_tensor_parameter_162: name = "parameter_162" + original_name = "conv2d_25.w_0" shape = [184, 32, 1, 1] dtype = "float32" min_val = float("-0.931903") @@ -1757,6 +1920,7 @@ class Program_weight_tensor_parameter_162: class Program_weight_tensor_parameter_163: name = "parameter_163" + original_name = "batch_norm_18.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -1766,6 +1930,7 @@ class Program_weight_tensor_parameter_163: class Program_weight_tensor_parameter_164: name = "parameter_164" + original_name = "batch_norm_18.w_0" shape = [32] dtype = "float32" min_val = float("0") @@ -1775,6 +1940,7 @@ class Program_weight_tensor_parameter_164: class Program_weight_tensor_parameter_165: name = "parameter_165" + original_name = "batch_norm_18.w_2" shape = [32] dtype = "float32" min_val = float("0") @@ -1784,6 +1950,7 @@ class Program_weight_tensor_parameter_165: class Program_weight_tensor_parameter_166: name = "parameter_166" + original_name = "batch_norm_18.w_1" shape = [32] dtype = "float32" min_val = float("0") @@ -1793,6 +1960,7 @@ class Program_weight_tensor_parameter_166: class Program_weight_tensor_parameter_167: name = "parameter_167" + original_name = "conv2d_24.w_0" shape = [32, 88, 1, 1] dtype = "float32" min_val = float("-0.66288") @@ -1804,6 +1972,7 @@ class Program_weight_tensor_parameter_167: class Program_weight_tensor_parameter_168: name = "parameter_168" + original_name = "conv2d_23.b_0" shape = [88] dtype = "float32" min_val = float("-0.168609") @@ -1815,6 +1984,7 @@ class Program_weight_tensor_parameter_168: class Program_weight_tensor_parameter_169: name = "parameter_169" + original_name = "conv2d_23.w_0" shape = [88, 22, 1, 1] dtype = "float32" min_val = float("-0.67169") @@ -1826,6 +1996,7 @@ class Program_weight_tensor_parameter_169: class Program_weight_tensor_parameter_170: name = "parameter_170" + original_name = "conv2d_22.b_0" shape = [22] dtype = "float32" min_val = float("0") @@ -1835,6 +2006,7 @@ class Program_weight_tensor_parameter_170: class Program_weight_tensor_parameter_171: name = "parameter_171" + original_name = "conv2d_22.w_0" shape = [22, 88, 1, 1] dtype = "float32" min_val = float("-0.790067") @@ -1846,6 +2018,7 @@ class Program_weight_tensor_parameter_171: class Program_weight_tensor_parameter_172: name = "parameter_172" + original_name = "batch_norm_17.b_0" shape = [88] dtype = "float32" min_val = float("-5.27957") @@ -1857,6 +2030,7 @@ class Program_weight_tensor_parameter_172: class Program_weight_tensor_parameter_173: name = "parameter_173" + original_name = "batch_norm_17.w_0" shape = [88] dtype = "float32" min_val = float("0.271866") @@ -1868,6 +2042,7 @@ class Program_weight_tensor_parameter_173: class Program_weight_tensor_parameter_174: name = "parameter_174" + original_name = "batch_norm_17.w_2" shape = [88] dtype = "float32" min_val = float("0.0230687") @@ -1879,6 +2054,7 @@ class Program_weight_tensor_parameter_174: class Program_weight_tensor_parameter_175: name = "parameter_175" + original_name = "batch_norm_17.w_1" shape = [88] dtype = "float32" min_val = float("-2.68518") @@ -1890,6 +2066,7 @@ class Program_weight_tensor_parameter_175: class Program_weight_tensor_parameter_176: name = "parameter_176" + original_name = "conv2d_21.w_0" shape = [88, 1, 5, 5] dtype = "float32" min_val = float("-1.00694") @@ -1901,6 +2078,7 @@ class Program_weight_tensor_parameter_176: class Program_weight_tensor_parameter_177: name = "parameter_177" + original_name = "batch_norm_16.b_0" shape = [88] dtype = "float32" min_val = float("-2.06485") @@ -1912,6 +2090,7 @@ class Program_weight_tensor_parameter_177: class Program_weight_tensor_parameter_178: name = "parameter_178" + original_name = "batch_norm_16.w_0" shape = [88] dtype = "float32" min_val = float("0.596613") @@ -1923,6 +2102,7 @@ class Program_weight_tensor_parameter_178: class Program_weight_tensor_parameter_179: name = "parameter_179" + original_name = "batch_norm_16.w_2" shape = [88] dtype = "float32" min_val = float("15.1148") @@ -1934,6 +2114,7 @@ class Program_weight_tensor_parameter_179: class Program_weight_tensor_parameter_180: name = "parameter_180" + original_name = "batch_norm_16.w_1" shape = [88] dtype = "float32" min_val = float("-9.77207e-06") @@ -1945,6 +2126,7 @@ class Program_weight_tensor_parameter_180: class Program_weight_tensor_parameter_181: name = "parameter_181" + original_name = "conv2d_20.w_0" shape = [88, 32, 1, 1] dtype = "float32" min_val = float("-0.9059") @@ -1956,6 +2138,7 @@ class Program_weight_tensor_parameter_181: class Program_weight_tensor_parameter_182: name = "parameter_182" + original_name = "batch_norm_15.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -1965,6 +2148,7 @@ class Program_weight_tensor_parameter_182: class Program_weight_tensor_parameter_183: name = "parameter_183" + original_name = "batch_norm_15.w_0" shape = [32] dtype = "float32" min_val = float("0") @@ -1974,6 +2158,7 @@ class Program_weight_tensor_parameter_183: class Program_weight_tensor_parameter_184: name = "parameter_184" + original_name = "batch_norm_15.w_2" shape = [32] dtype = "float32" min_val = float("0") @@ -1983,6 +2168,7 @@ class Program_weight_tensor_parameter_184: class Program_weight_tensor_parameter_185: name = "parameter_185" + original_name = "batch_norm_15.w_1" shape = [32] dtype = "float32" min_val = float("0") @@ -1992,6 +2178,7 @@ class Program_weight_tensor_parameter_185: class Program_weight_tensor_parameter_186: name = "parameter_186" + original_name = "conv2d_19.w_0" shape = [32, 88, 1, 1] dtype = "float32" min_val = float("-0.660313") @@ -2003,6 +2190,7 @@ class Program_weight_tensor_parameter_186: class Program_weight_tensor_parameter_187: name = "parameter_187" + original_name = "conv2d_18.b_0" shape = [88] dtype = "float32" min_val = float("-0.240632") @@ -2014,6 +2202,7 @@ class Program_weight_tensor_parameter_187: class Program_weight_tensor_parameter_188: name = "parameter_188" + original_name = "conv2d_18.w_0" shape = [88, 22, 1, 1] dtype = "float32" min_val = float("-0.658799") @@ -2025,6 +2214,7 @@ class Program_weight_tensor_parameter_188: class Program_weight_tensor_parameter_189: name = "parameter_189" + original_name = "conv2d_17.b_0" shape = [22] dtype = "float32" min_val = float("0") @@ -2034,6 +2224,7 @@ class Program_weight_tensor_parameter_189: class Program_weight_tensor_parameter_190: name = "parameter_190" + original_name = "conv2d_17.w_0" shape = [22, 88, 1, 1] dtype = "float32" min_val = float("-1.38519") @@ -2045,6 +2236,7 @@ class Program_weight_tensor_parameter_190: class Program_weight_tensor_parameter_191: name = "parameter_191" + original_name = "batch_norm_14.b_0" shape = [88] dtype = "float32" min_val = float("-2.27873") @@ -2056,6 +2248,7 @@ class Program_weight_tensor_parameter_191: class Program_weight_tensor_parameter_192: name = "parameter_192" + original_name = "batch_norm_14.w_0" shape = [88] dtype = "float32" min_val = float("0.419434") @@ -2067,6 +2260,7 @@ class Program_weight_tensor_parameter_192: class Program_weight_tensor_parameter_193: name = "parameter_193" + original_name = "batch_norm_14.w_2" shape = [88] dtype = "float32" min_val = float("0.0111586") @@ -2078,6 +2272,7 @@ class Program_weight_tensor_parameter_193: class Program_weight_tensor_parameter_194: name = "parameter_194" + original_name = "batch_norm_14.w_1" shape = [88] dtype = "float32" min_val = float("-16.6302") @@ -2089,6 +2284,7 @@ class Program_weight_tensor_parameter_194: class Program_weight_tensor_parameter_195: name = "parameter_195" + original_name = "conv2d_16.w_0" shape = [88, 1, 5, 5] dtype = "float32" min_val = float("-1.01116") @@ -2100,6 +2296,7 @@ class Program_weight_tensor_parameter_195: class Program_weight_tensor_parameter_196: name = "parameter_196" + original_name = "batch_norm_13.b_0" shape = [88] dtype = "float32" min_val = float("-2.68261") @@ -2111,6 +2308,7 @@ class Program_weight_tensor_parameter_196: class Program_weight_tensor_parameter_197: name = "parameter_197" + original_name = "batch_norm_13.w_0" shape = [88] dtype = "float32" min_val = float("0.391703") @@ -2122,6 +2320,7 @@ class Program_weight_tensor_parameter_197: class Program_weight_tensor_parameter_198: name = "parameter_198" + original_name = "batch_norm_13.w_2" shape = [88] dtype = "float32" min_val = float("8.33198") @@ -2133,6 +2332,7 @@ class Program_weight_tensor_parameter_198: class Program_weight_tensor_parameter_199: name = "parameter_199" + original_name = "batch_norm_13.w_1" shape = [88] dtype = "float32" min_val = float("-6.09474e-06") @@ -2144,6 +2344,7 @@ class Program_weight_tensor_parameter_199: class Program_weight_tensor_parameter_200: name = "parameter_200" + original_name = "conv2d_15.w_0" shape = [88, 32, 1, 1] dtype = "float32" min_val = float("-0.670744") @@ -2155,6 +2356,7 @@ class Program_weight_tensor_parameter_200: class Program_weight_tensor_parameter_201: name = "parameter_201" + original_name = "batch_norm_12.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -2164,6 +2366,7 @@ class Program_weight_tensor_parameter_201: class Program_weight_tensor_parameter_202: name = "parameter_202" + original_name = "batch_norm_12.w_0" shape = [32] dtype = "float32" min_val = float("0") @@ -2173,6 +2376,7 @@ class Program_weight_tensor_parameter_202: class Program_weight_tensor_parameter_203: name = "parameter_203" + original_name = "batch_norm_12.w_2" shape = [32] dtype = "float32" min_val = float("0") @@ -2182,6 +2386,7 @@ class Program_weight_tensor_parameter_203: class Program_weight_tensor_parameter_204: name = "parameter_204" + original_name = "batch_norm_12.w_1" shape = [32] dtype = "float32" min_val = float("0") @@ -2191,6 +2396,7 @@ class Program_weight_tensor_parameter_204: class Program_weight_tensor_parameter_205: name = "parameter_205" + original_name = "conv2d_14.w_0" shape = [32, 56, 1, 1] dtype = "float32" min_val = float("-1.61644") @@ -2202,6 +2408,7 @@ class Program_weight_tensor_parameter_205: class Program_weight_tensor_parameter_206: name = "parameter_206" + original_name = "conv2d_13.b_0" shape = [56] dtype = "float32" min_val = float("0") @@ -2211,6 +2418,7 @@ class Program_weight_tensor_parameter_206: class Program_weight_tensor_parameter_207: name = "parameter_207" + original_name = "conv2d_13.w_0" shape = [56, 14, 1, 1] dtype = "float32" min_val = float("-4.50801e-13") @@ -2222,6 +2430,7 @@ class Program_weight_tensor_parameter_207: class Program_weight_tensor_parameter_208: name = "parameter_208" + original_name = "conv2d_12.b_0" shape = [14] dtype = "float32" min_val = float("0") @@ -2231,6 +2440,7 @@ class Program_weight_tensor_parameter_208: class Program_weight_tensor_parameter_209: name = "parameter_209" + original_name = "conv2d_12.w_0" shape = [14, 56, 1, 1] dtype = "float32" min_val = float("-9.4712e-13") @@ -2242,6 +2452,7 @@ class Program_weight_tensor_parameter_209: class Program_weight_tensor_parameter_210: name = "parameter_210" + original_name = "batch_norm_11.b_0" shape = [56] dtype = "float32" min_val = float("0") @@ -2251,6 +2462,7 @@ class Program_weight_tensor_parameter_210: class Program_weight_tensor_parameter_211: name = "parameter_211" + original_name = "batch_norm_11.w_0" shape = [56] dtype = "float32" min_val = float("0") @@ -2260,6 +2472,7 @@ class Program_weight_tensor_parameter_211: class Program_weight_tensor_parameter_212: name = "parameter_212" + original_name = "batch_norm_11.w_2" shape = [56] dtype = "float32" min_val = float("0") @@ -2269,6 +2482,7 @@ class Program_weight_tensor_parameter_212: class Program_weight_tensor_parameter_213: name = "parameter_213" + original_name = "batch_norm_11.w_1" shape = [56] dtype = "float32" min_val = float("0") @@ -2278,6 +2492,7 @@ class Program_weight_tensor_parameter_213: class Program_weight_tensor_parameter_214: name = "parameter_214" + original_name = "conv2d_11.w_0" shape = [56, 1, 5, 5] dtype = "float32" min_val = float("-0.842694") @@ -2289,6 +2504,7 @@ class Program_weight_tensor_parameter_214: class Program_weight_tensor_parameter_215: name = "parameter_215" + original_name = "batch_norm_10.b_0" shape = [56] dtype = "float32" min_val = float("0") @@ -2298,6 +2514,7 @@ class Program_weight_tensor_parameter_215: class Program_weight_tensor_parameter_216: name = "parameter_216" + original_name = "batch_norm_10.w_0" shape = [56] dtype = "float32" min_val = float("0") @@ -2307,6 +2524,7 @@ class Program_weight_tensor_parameter_216: class Program_weight_tensor_parameter_217: name = "parameter_217" + original_name = "batch_norm_10.w_2" shape = [56] dtype = "float32" min_val = float("0") @@ -2316,6 +2534,7 @@ class Program_weight_tensor_parameter_217: class Program_weight_tensor_parameter_218: name = "parameter_218" + original_name = "batch_norm_10.w_1" shape = [56] dtype = "float32" min_val = float("0") @@ -2325,6 +2544,7 @@ class Program_weight_tensor_parameter_218: class Program_weight_tensor_parameter_219: name = "parameter_219" + original_name = "conv2d_10.w_0" shape = [56, 24, 1, 1] dtype = "float32" min_val = float("-2.15983") @@ -2336,6 +2556,7 @@ class Program_weight_tensor_parameter_219: class Program_weight_tensor_parameter_220: name = "parameter_220" + original_name = "batch_norm_9.b_0" shape = [24] dtype = "float32" min_val = float("0") @@ -2345,6 +2566,7 @@ class Program_weight_tensor_parameter_220: class Program_weight_tensor_parameter_221: name = "parameter_221" + original_name = "batch_norm_9.w_0" shape = [24] dtype = "float32" min_val = float("0") @@ -2354,6 +2576,7 @@ class Program_weight_tensor_parameter_221: class Program_weight_tensor_parameter_222: name = "parameter_222" + original_name = "batch_norm_9.w_2" shape = [24] dtype = "float32" min_val = float("0") @@ -2363,6 +2586,7 @@ class Program_weight_tensor_parameter_222: class Program_weight_tensor_parameter_223: name = "parameter_223" + original_name = "batch_norm_9.w_1" shape = [24] dtype = "float32" min_val = float("0") @@ -2372,6 +2596,7 @@ class Program_weight_tensor_parameter_223: class Program_weight_tensor_parameter_224: name = "parameter_224" + original_name = "conv2d_9.w_0" shape = [24, 56, 1, 1] dtype = "float32" min_val = float("-0.83597") @@ -2383,6 +2608,7 @@ class Program_weight_tensor_parameter_224: class Program_weight_tensor_parameter_225: name = "parameter_225" + original_name = "batch_norm_8.b_0" shape = [56] dtype = "float32" min_val = float("0") @@ -2392,6 +2618,7 @@ class Program_weight_tensor_parameter_225: class Program_weight_tensor_parameter_226: name = "parameter_226" + original_name = "batch_norm_8.w_0" shape = [56] dtype = "float32" min_val = float("0") @@ -2401,6 +2628,7 @@ class Program_weight_tensor_parameter_226: class Program_weight_tensor_parameter_227: name = "parameter_227" + original_name = "batch_norm_8.w_2" shape = [56] dtype = "float32" min_val = float("0") @@ -2410,6 +2638,7 @@ class Program_weight_tensor_parameter_227: class Program_weight_tensor_parameter_228: name = "parameter_228" + original_name = "batch_norm_8.w_1" shape = [56] dtype = "float32" min_val = float("0") @@ -2419,6 +2648,7 @@ class Program_weight_tensor_parameter_228: class Program_weight_tensor_parameter_229: name = "parameter_229" + original_name = "conv2d_8.w_0" shape = [56, 1, 3, 3] dtype = "float32" min_val = float("-1.6145") @@ -2430,6 +2660,7 @@ class Program_weight_tensor_parameter_229: class Program_weight_tensor_parameter_230: name = "parameter_230" + original_name = "batch_norm_7.b_0" shape = [56] dtype = "float32" min_val = float("0") @@ -2439,6 +2670,7 @@ class Program_weight_tensor_parameter_230: class Program_weight_tensor_parameter_231: name = "parameter_231" + original_name = "batch_norm_7.w_0" shape = [56] dtype = "float32" min_val = float("0") @@ -2448,6 +2680,7 @@ class Program_weight_tensor_parameter_231: class Program_weight_tensor_parameter_232: name = "parameter_232" + original_name = "batch_norm_7.w_2" shape = [56] dtype = "float32" min_val = float("0") @@ -2457,6 +2690,7 @@ class Program_weight_tensor_parameter_232: class Program_weight_tensor_parameter_233: name = "parameter_233" + original_name = "batch_norm_7.w_1" shape = [56] dtype = "float32" min_val = float("0") @@ -2466,6 +2700,7 @@ class Program_weight_tensor_parameter_233: class Program_weight_tensor_parameter_234: name = "parameter_234" + original_name = "conv2d_7.w_0" shape = [56, 24, 1, 1] dtype = "float32" min_val = float("-1.29776") @@ -2477,6 +2712,7 @@ class Program_weight_tensor_parameter_234: class Program_weight_tensor_parameter_235: name = "parameter_235" + original_name = "batch_norm_6.b_0" shape = [24] dtype = "float32" min_val = float("0") @@ -2486,6 +2722,7 @@ class Program_weight_tensor_parameter_235: class Program_weight_tensor_parameter_236: name = "parameter_236" + original_name = "batch_norm_6.w_0" shape = [24] dtype = "float32" min_val = float("0") @@ -2495,6 +2732,7 @@ class Program_weight_tensor_parameter_236: class Program_weight_tensor_parameter_237: name = "parameter_237" + original_name = "batch_norm_6.w_2" shape = [24] dtype = "float32" min_val = float("0") @@ -2504,6 +2742,7 @@ class Program_weight_tensor_parameter_237: class Program_weight_tensor_parameter_238: name = "parameter_238" + original_name = "batch_norm_6.w_1" shape = [24] dtype = "float32" min_val = float("0") @@ -2513,6 +2752,7 @@ class Program_weight_tensor_parameter_238: class Program_weight_tensor_parameter_239: name = "parameter_239" + original_name = "conv2d_6.w_0" shape = [24, 48, 1, 1] dtype = "float32" min_val = float("-1.71038") @@ -2524,6 +2764,7 @@ class Program_weight_tensor_parameter_239: class Program_weight_tensor_parameter_240: name = "parameter_240" + original_name = "batch_norm_5.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -2533,6 +2774,7 @@ class Program_weight_tensor_parameter_240: class Program_weight_tensor_parameter_241: name = "parameter_241" + original_name = "batch_norm_5.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -2542,6 +2784,7 @@ class Program_weight_tensor_parameter_241: class Program_weight_tensor_parameter_242: name = "parameter_242" + original_name = "batch_norm_5.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -2551,6 +2794,7 @@ class Program_weight_tensor_parameter_242: class Program_weight_tensor_parameter_243: name = "parameter_243" + original_name = "batch_norm_5.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -2560,6 +2804,7 @@ class Program_weight_tensor_parameter_243: class Program_weight_tensor_parameter_244: name = "parameter_244" + original_name = "conv2d_5.w_0" shape = [48, 1, 3, 3] dtype = "float32" min_val = float("-0.652995") @@ -2571,6 +2816,7 @@ class Program_weight_tensor_parameter_244: class Program_weight_tensor_parameter_245: name = "parameter_245" + original_name = "batch_norm_4.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -2580,6 +2826,7 @@ class Program_weight_tensor_parameter_245: class Program_weight_tensor_parameter_246: name = "parameter_246" + original_name = "batch_norm_4.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -2589,6 +2836,7 @@ class Program_weight_tensor_parameter_246: class Program_weight_tensor_parameter_247: name = "parameter_247" + original_name = "batch_norm_4.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -2598,6 +2846,7 @@ class Program_weight_tensor_parameter_247: class Program_weight_tensor_parameter_248: name = "parameter_248" + original_name = "batch_norm_4.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -2607,6 +2856,7 @@ class Program_weight_tensor_parameter_248: class Program_weight_tensor_parameter_249: name = "parameter_249" + original_name = "conv2d_4.w_0" shape = [48, 16, 1, 1] dtype = "float32" min_val = float("-1.51828") @@ -2618,6 +2868,7 @@ class Program_weight_tensor_parameter_249: class Program_weight_tensor_parameter_250: name = "parameter_250" + original_name = "batch_norm_3.b_0" shape = [16] dtype = "float32" min_val = float("0") @@ -2627,6 +2878,7 @@ class Program_weight_tensor_parameter_250: class Program_weight_tensor_parameter_251: name = "parameter_251" + original_name = "batch_norm_3.w_0" shape = [16] dtype = "float32" min_val = float("0") @@ -2636,6 +2888,7 @@ class Program_weight_tensor_parameter_251: class Program_weight_tensor_parameter_252: name = "parameter_252" + original_name = "batch_norm_3.w_2" shape = [16] dtype = "float32" min_val = float("0") @@ -2645,6 +2898,7 @@ class Program_weight_tensor_parameter_252: class Program_weight_tensor_parameter_253: name = "parameter_253" + original_name = "batch_norm_3.w_1" shape = [16] dtype = "float32" min_val = float("0") @@ -2654,6 +2908,7 @@ class Program_weight_tensor_parameter_253: class Program_weight_tensor_parameter_254: name = "parameter_254" + original_name = "conv2d_3.w_0" shape = [16, 16, 1, 1] dtype = "float32" min_val = float("-1.28614") @@ -2665,6 +2920,7 @@ class Program_weight_tensor_parameter_254: class Program_weight_tensor_parameter_255: name = "parameter_255" + original_name = "batch_norm_2.b_0" shape = [16] dtype = "float32" min_val = float("0") @@ -2674,6 +2930,7 @@ class Program_weight_tensor_parameter_255: class Program_weight_tensor_parameter_256: name = "parameter_256" + original_name = "batch_norm_2.w_0" shape = [16] dtype = "float32" min_val = float("0") @@ -2683,6 +2940,7 @@ class Program_weight_tensor_parameter_256: class Program_weight_tensor_parameter_257: name = "parameter_257" + original_name = "batch_norm_2.w_2" shape = [16] dtype = "float32" min_val = float("0") @@ -2692,6 +2950,7 @@ class Program_weight_tensor_parameter_257: class Program_weight_tensor_parameter_258: name = "parameter_258" + original_name = "batch_norm_2.w_1" shape = [16] dtype = "float32" min_val = float("0") @@ -2701,6 +2960,7 @@ class Program_weight_tensor_parameter_258: class Program_weight_tensor_parameter_259: name = "parameter_259" + original_name = "conv2d_2.w_0" shape = [16, 1, 3, 3] dtype = "float32" min_val = float("-2.98312") @@ -2712,6 +2972,7 @@ class Program_weight_tensor_parameter_259: class Program_weight_tensor_parameter_260: name = "parameter_260" + original_name = "batch_norm_1.b_0" shape = [16] dtype = "float32" min_val = float("0") @@ -2721,6 +2982,7 @@ class Program_weight_tensor_parameter_260: class Program_weight_tensor_parameter_261: name = "parameter_261" + original_name = "batch_norm_1.w_0" shape = [16] dtype = "float32" min_val = float("0") @@ -2730,6 +2992,7 @@ class Program_weight_tensor_parameter_261: class Program_weight_tensor_parameter_262: name = "parameter_262" + original_name = "batch_norm_1.w_2" shape = [16] dtype = "float32" min_val = float("0") @@ -2739,6 +3002,7 @@ class Program_weight_tensor_parameter_262: class Program_weight_tensor_parameter_263: name = "parameter_263" + original_name = "batch_norm_1.w_1" shape = [16] dtype = "float32" min_val = float("0") @@ -2748,6 +3012,7 @@ class Program_weight_tensor_parameter_263: class Program_weight_tensor_parameter_264: name = "parameter_264" + original_name = "conv2d_1.w_0" shape = [16, 16, 1, 1] dtype = "float32" min_val = float("-1.52015") @@ -2759,6 +3024,7 @@ class Program_weight_tensor_parameter_264: class Program_weight_tensor_parameter_265: name = "parameter_265" + original_name = "batch_norm_0.b_0" shape = [16] dtype = "float32" min_val = float("0") @@ -2768,6 +3034,7 @@ class Program_weight_tensor_parameter_265: class Program_weight_tensor_parameter_266: name = "parameter_266" + original_name = "batch_norm_0.w_0" shape = [16] dtype = "float32" min_val = float("0") @@ -2777,6 +3044,7 @@ class Program_weight_tensor_parameter_266: class Program_weight_tensor_parameter_267: name = "parameter_267" + original_name = "batch_norm_0.w_2" shape = [16] dtype = "float32" min_val = float("0") @@ -2786,6 +3054,7 @@ class Program_weight_tensor_parameter_267: class Program_weight_tensor_parameter_268: name = "parameter_268" + original_name = "batch_norm_0.w_1" shape = [16] dtype = "float32" min_val = float("0") @@ -2795,6 +3064,7 @@ class Program_weight_tensor_parameter_268: class Program_weight_tensor_parameter_269: name = "parameter_269" + original_name = "conv2d_0.w_0" shape = [16, 3, 3, 3] dtype = "float32" min_val = float("-2.18792") diff --git a/paddle_samples/PaddleX/MobileNetV3_large_x1_0/subgraph_1/input_meta.py b/paddle_samples/PaddleX/MobileNetV3_large_x1_0/subgraph_1/input_meta.py index 474a83446..300b3d10a 100644 --- a/paddle_samples/PaddleX/MobileNetV3_large_x1_0/subgraph_1/input_meta.py +++ b/paddle_samples/PaddleX/MobileNetV3_large_x1_0/subgraph_1/input_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_data_0: name = "data_0" + original_name = "var_191" shape = [124, 3, 224, 224] dtype = "float32" min_val = float("-2.1179") diff --git a/paddle_samples/PaddleX/MobileNetV3_large_x1_0/subgraph_1/weight_meta.py b/paddle_samples/PaddleX/MobileNetV3_large_x1_0/subgraph_1/weight_meta.py index 9c806c36a..0dece087e 100644 --- a/paddle_samples/PaddleX/MobileNetV3_large_x1_0/subgraph_1/weight_meta.py +++ b/paddle_samples/PaddleX/MobileNetV3_large_x1_0/subgraph_1/weight_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_parameter_0: name = "parameter_0" + original_name = "linear_0.b_0" shape = [102] dtype = "float32" min_val = float("-0.00244098") @@ -11,6 +12,7 @@ class Program_weight_tensor_parameter_0: class Program_weight_tensor_parameter_1: name = "parameter_1" + original_name = "linear_0.w_0" shape = [1280, 102] dtype = "float32" min_val = float("-0.0676653") @@ -22,6 +24,7 @@ class Program_weight_tensor_parameter_1: class Program_weight_tensor_parameter_2: name = "parameter_2" + original_name = "conv2d_63.w_0" shape = [1280, 960, 1, 1] dtype = "float32" min_val = float("-0.60852") @@ -33,6 +36,7 @@ class Program_weight_tensor_parameter_2: class Program_weight_tensor_parameter_3: name = "parameter_3" + original_name = "batch_norm_46.b_0" shape = [960] dtype = "float32" min_val = float("-7.93193") @@ -44,6 +48,7 @@ class Program_weight_tensor_parameter_3: class Program_weight_tensor_parameter_4: name = "parameter_4" + original_name = "batch_norm_46.w_0" shape = [960] dtype = "float32" min_val = float("1.04844") @@ -55,6 +60,7 @@ class Program_weight_tensor_parameter_4: class Program_weight_tensor_parameter_5: name = "parameter_5" + original_name = "batch_norm_46.w_2" shape = [960] dtype = "float32" min_val = float("13.1302") @@ -66,6 +72,7 @@ class Program_weight_tensor_parameter_5: class Program_weight_tensor_parameter_6: name = "parameter_6" + original_name = "batch_norm_46.w_1" shape = [960] dtype = "float32" min_val = float("-1.49423e-05") @@ -77,6 +84,7 @@ class Program_weight_tensor_parameter_6: class Program_weight_tensor_parameter_7: name = "parameter_7" + original_name = "conv2d_62.w_0" shape = [960, 160, 1, 1] dtype = "float32" min_val = float("-1.14473") @@ -88,6 +96,7 @@ class Program_weight_tensor_parameter_7: class Program_weight_tensor_parameter_8: name = "parameter_8" + original_name = "batch_norm_45.b_0" shape = [160] dtype = "float32" min_val = float("-1.57317e-06") @@ -99,6 +108,7 @@ class Program_weight_tensor_parameter_8: class Program_weight_tensor_parameter_9: name = "parameter_9" + original_name = "batch_norm_45.w_0" shape = [160] dtype = "float32" min_val = float("1.63583") @@ -110,6 +120,7 @@ class Program_weight_tensor_parameter_9: class Program_weight_tensor_parameter_10: name = "parameter_10" + original_name = "batch_norm_45.w_2" shape = [160] dtype = "float32" min_val = float("0.084753") @@ -121,6 +132,7 @@ class Program_weight_tensor_parameter_10: class Program_weight_tensor_parameter_11: name = "parameter_11" + original_name = "batch_norm_45.w_1" shape = [160] dtype = "float32" min_val = float("-0.64012") @@ -132,6 +144,7 @@ class Program_weight_tensor_parameter_11: class Program_weight_tensor_parameter_12: name = "parameter_12" + original_name = "conv2d_61.w_0" shape = [160, 960, 1, 1] dtype = "float32" min_val = float("-0.354054") @@ -143,6 +156,7 @@ class Program_weight_tensor_parameter_12: class Program_weight_tensor_parameter_13: name = "parameter_13" + original_name = "conv2d_60.b_0" shape = [960] dtype = "float32" min_val = float("-0.917951") @@ -154,6 +168,7 @@ class Program_weight_tensor_parameter_13: class Program_weight_tensor_parameter_14: name = "parameter_14" + original_name = "conv2d_60.w_0" shape = [960, 240, 1, 1] dtype = "float32" min_val = float("-0.715262") @@ -165,6 +180,7 @@ class Program_weight_tensor_parameter_14: class Program_weight_tensor_parameter_15: name = "parameter_15" + original_name = "conv2d_59.b_0" shape = [240] dtype = "float32" min_val = float("-0.0595683") @@ -176,6 +192,7 @@ class Program_weight_tensor_parameter_15: class Program_weight_tensor_parameter_16: name = "parameter_16" + original_name = "conv2d_59.w_0" shape = [240, 960, 1, 1] dtype = "float32" min_val = float("-0.352981") @@ -187,6 +204,7 @@ class Program_weight_tensor_parameter_16: class Program_weight_tensor_parameter_17: name = "parameter_17" + original_name = "batch_norm_44.b_0" shape = [960] dtype = "float32" min_val = float("-4.11874") @@ -198,6 +216,7 @@ class Program_weight_tensor_parameter_17: class Program_weight_tensor_parameter_18: name = "parameter_18" + original_name = "batch_norm_44.w_0" shape = [960] dtype = "float32" min_val = float("0.15711") @@ -209,6 +228,7 @@ class Program_weight_tensor_parameter_18: class Program_weight_tensor_parameter_19: name = "parameter_19" + original_name = "batch_norm_44.w_2" shape = [960] dtype = "float32" min_val = float("4.46047e-05") @@ -220,6 +240,7 @@ class Program_weight_tensor_parameter_19: class Program_weight_tensor_parameter_20: name = "parameter_20" + original_name = "batch_norm_44.w_1" shape = [960] dtype = "float32" min_val = float("-1.45741") @@ -231,6 +252,7 @@ class Program_weight_tensor_parameter_20: class Program_weight_tensor_parameter_21: name = "parameter_21" + original_name = "conv2d_58.w_0" shape = [960, 1, 5, 5] dtype = "float32" min_val = float("-0.670905") @@ -242,6 +264,7 @@ class Program_weight_tensor_parameter_21: class Program_weight_tensor_parameter_22: name = "parameter_22" + original_name = "batch_norm_43.b_0" shape = [960] dtype = "float32" min_val = float("-4.48236") @@ -253,6 +276,7 @@ class Program_weight_tensor_parameter_22: class Program_weight_tensor_parameter_23: name = "parameter_23" + original_name = "batch_norm_43.w_0" shape = [960] dtype = "float32" min_val = float("-0.617643") @@ -264,6 +288,7 @@ class Program_weight_tensor_parameter_23: class Program_weight_tensor_parameter_24: name = "parameter_24" + original_name = "batch_norm_43.w_2" shape = [960] dtype = "float32" min_val = float("2.25015") @@ -275,6 +300,7 @@ class Program_weight_tensor_parameter_24: class Program_weight_tensor_parameter_25: name = "parameter_25" + original_name = "batch_norm_43.w_1" shape = [960] dtype = "float32" min_val = float("-6.67037e-06") @@ -286,6 +312,7 @@ class Program_weight_tensor_parameter_25: class Program_weight_tensor_parameter_26: name = "parameter_26" + original_name = "conv2d_57.w_0" shape = [960, 160, 1, 1] dtype = "float32" min_val = float("-0.579094") @@ -297,6 +324,7 @@ class Program_weight_tensor_parameter_26: class Program_weight_tensor_parameter_27: name = "parameter_27" + original_name = "batch_norm_42.b_0" shape = [160] dtype = "float32" min_val = float("-2.25862e-06") @@ -308,6 +336,7 @@ class Program_weight_tensor_parameter_27: class Program_weight_tensor_parameter_28: name = "parameter_28" + original_name = "batch_norm_42.w_0" shape = [160] dtype = "float32" min_val = float("0.919471") @@ -319,6 +348,7 @@ class Program_weight_tensor_parameter_28: class Program_weight_tensor_parameter_29: name = "parameter_29" + original_name = "batch_norm_42.w_2" shape = [160] dtype = "float32" min_val = float("0.0760391") @@ -330,6 +360,7 @@ class Program_weight_tensor_parameter_29: class Program_weight_tensor_parameter_30: name = "parameter_30" + original_name = "batch_norm_42.w_1" shape = [160] dtype = "float32" min_val = float("-0.515434") @@ -341,6 +372,7 @@ class Program_weight_tensor_parameter_30: class Program_weight_tensor_parameter_31: name = "parameter_31" + original_name = "conv2d_56.w_0" shape = [160, 960, 1, 1] dtype = "float32" min_val = float("-0.38516") @@ -352,6 +384,7 @@ class Program_weight_tensor_parameter_31: class Program_weight_tensor_parameter_32: name = "parameter_32" + original_name = "conv2d_55.b_0" shape = [960] dtype = "float32" min_val = float("-0.199724") @@ -363,6 +396,7 @@ class Program_weight_tensor_parameter_32: class Program_weight_tensor_parameter_33: name = "parameter_33" + original_name = "conv2d_55.w_0" shape = [960, 240, 1, 1] dtype = "float32" min_val = float("-0.360401") @@ -374,6 +408,7 @@ class Program_weight_tensor_parameter_33: class Program_weight_tensor_parameter_34: name = "parameter_34" + original_name = "conv2d_54.b_0" shape = [240] dtype = "float32" min_val = float("-0.047157") @@ -385,6 +420,7 @@ class Program_weight_tensor_parameter_34: class Program_weight_tensor_parameter_35: name = "parameter_35" + original_name = "conv2d_54.w_0" shape = [240, 960, 1, 1] dtype = "float32" min_val = float("-0.362833") @@ -396,6 +432,7 @@ class Program_weight_tensor_parameter_35: class Program_weight_tensor_parameter_36: name = "parameter_36" + original_name = "batch_norm_41.b_0" shape = [960] dtype = "float32" min_val = float("-4.85964") @@ -407,6 +444,7 @@ class Program_weight_tensor_parameter_36: class Program_weight_tensor_parameter_37: name = "parameter_37" + original_name = "batch_norm_41.w_0" shape = [960] dtype = "float32" min_val = float("0.386584") @@ -418,6 +456,7 @@ class Program_weight_tensor_parameter_37: class Program_weight_tensor_parameter_38: name = "parameter_38" + original_name = "batch_norm_41.w_2" shape = [960] dtype = "float32" min_val = float("0.000115022") @@ -429,6 +468,7 @@ class Program_weight_tensor_parameter_38: class Program_weight_tensor_parameter_39: name = "parameter_39" + original_name = "batch_norm_41.w_1" shape = [960] dtype = "float32" min_val = float("-1.62183") @@ -440,6 +480,7 @@ class Program_weight_tensor_parameter_39: class Program_weight_tensor_parameter_40: name = "parameter_40" + original_name = "conv2d_53.w_0" shape = [960, 1, 5, 5] dtype = "float32" min_val = float("-0.572992") @@ -451,6 +492,7 @@ class Program_weight_tensor_parameter_40: class Program_weight_tensor_parameter_41: name = "parameter_41" + original_name = "batch_norm_40.b_0" shape = [960] dtype = "float32" min_val = float("-3.64815") @@ -462,6 +504,7 @@ class Program_weight_tensor_parameter_41: class Program_weight_tensor_parameter_42: name = "parameter_42" + original_name = "batch_norm_40.w_0" shape = [960] dtype = "float32" min_val = float("-0.706069") @@ -473,6 +516,7 @@ class Program_weight_tensor_parameter_42: class Program_weight_tensor_parameter_43: name = "parameter_43" + original_name = "batch_norm_40.w_2" shape = [960] dtype = "float32" min_val = float("2.38642") @@ -484,6 +528,7 @@ class Program_weight_tensor_parameter_43: class Program_weight_tensor_parameter_44: name = "parameter_44" + original_name = "batch_norm_40.w_1" shape = [960] dtype = "float32" min_val = float("-3.58688e-06") @@ -495,6 +540,7 @@ class Program_weight_tensor_parameter_44: class Program_weight_tensor_parameter_45: name = "parameter_45" + original_name = "conv2d_52.w_0" shape = [960, 160, 1, 1] dtype = "float32" min_val = float("-0.541989") @@ -506,6 +552,7 @@ class Program_weight_tensor_parameter_45: class Program_weight_tensor_parameter_46: name = "parameter_46" + original_name = "batch_norm_39.b_0" shape = [160] dtype = "float32" min_val = float("-2.49198e-06") @@ -517,6 +564,7 @@ class Program_weight_tensor_parameter_46: class Program_weight_tensor_parameter_47: name = "parameter_47" + original_name = "batch_norm_39.w_0" shape = [160] dtype = "float32" min_val = float("2.07887") @@ -528,6 +576,7 @@ class Program_weight_tensor_parameter_47: class Program_weight_tensor_parameter_48: name = "parameter_48" + original_name = "batch_norm_39.w_2" shape = [160] dtype = "float32" min_val = float("0.10537") @@ -539,6 +588,7 @@ class Program_weight_tensor_parameter_48: class Program_weight_tensor_parameter_49: name = "parameter_49" + original_name = "batch_norm_39.w_1" shape = [160] dtype = "float32" min_val = float("-0.737338") @@ -550,6 +600,7 @@ class Program_weight_tensor_parameter_49: class Program_weight_tensor_parameter_50: name = "parameter_50" + original_name = "conv2d_51.w_0" shape = [160, 672, 1, 1] dtype = "float32" min_val = float("-0.717712") @@ -561,6 +612,7 @@ class Program_weight_tensor_parameter_50: class Program_weight_tensor_parameter_51: name = "parameter_51" + original_name = "conv2d_50.b_0" shape = [672] dtype = "float32" min_val = float("-0.348848") @@ -572,6 +624,7 @@ class Program_weight_tensor_parameter_51: class Program_weight_tensor_parameter_52: name = "parameter_52" + original_name = "conv2d_50.w_0" shape = [672, 168, 1, 1] dtype = "float32" min_val = float("-0.653859") @@ -583,6 +636,7 @@ class Program_weight_tensor_parameter_52: class Program_weight_tensor_parameter_53: name = "parameter_53" + original_name = "conv2d_49.b_0" shape = [168] dtype = "float32" min_val = float("-0.0389658") @@ -594,6 +648,7 @@ class Program_weight_tensor_parameter_53: class Program_weight_tensor_parameter_54: name = "parameter_54" + original_name = "conv2d_49.w_0" shape = [168, 672, 1, 1] dtype = "float32" min_val = float("-0.543314") @@ -605,6 +660,7 @@ class Program_weight_tensor_parameter_54: class Program_weight_tensor_parameter_55: name = "parameter_55" + original_name = "batch_norm_38.b_0" shape = [672] dtype = "float32" min_val = float("-2.82412") @@ -616,6 +672,7 @@ class Program_weight_tensor_parameter_55: class Program_weight_tensor_parameter_56: name = "parameter_56" + original_name = "batch_norm_38.w_0" shape = [672] dtype = "float32" min_val = float("0.190345") @@ -627,6 +684,7 @@ class Program_weight_tensor_parameter_56: class Program_weight_tensor_parameter_57: name = "parameter_57" + original_name = "batch_norm_38.w_2" shape = [672] dtype = "float32" min_val = float("0.000100282") @@ -638,6 +696,7 @@ class Program_weight_tensor_parameter_57: class Program_weight_tensor_parameter_58: name = "parameter_58" + original_name = "batch_norm_38.w_1" shape = [672] dtype = "float32" min_val = float("-4.53833") @@ -649,6 +708,7 @@ class Program_weight_tensor_parameter_58: class Program_weight_tensor_parameter_59: name = "parameter_59" + original_name = "conv2d_48.w_0" shape = [672, 1, 5, 5] dtype = "float32" min_val = float("-0.506136") @@ -660,6 +720,7 @@ class Program_weight_tensor_parameter_59: class Program_weight_tensor_parameter_60: name = "parameter_60" + original_name = "batch_norm_37.b_0" shape = [672] dtype = "float32" min_val = float("-3.80572") @@ -671,6 +732,7 @@ class Program_weight_tensor_parameter_60: class Program_weight_tensor_parameter_61: name = "parameter_61" + original_name = "batch_norm_37.w_0" shape = [672] dtype = "float32" min_val = float("-0.997597") @@ -682,6 +744,7 @@ class Program_weight_tensor_parameter_61: class Program_weight_tensor_parameter_62: name = "parameter_62" + original_name = "batch_norm_37.w_2" shape = [672] dtype = "float32" min_val = float("5.94858") @@ -693,6 +756,7 @@ class Program_weight_tensor_parameter_62: class Program_weight_tensor_parameter_63: name = "parameter_63" + original_name = "batch_norm_37.w_1" shape = [672] dtype = "float32" min_val = float("-5.95937e-06") @@ -704,6 +768,7 @@ class Program_weight_tensor_parameter_63: class Program_weight_tensor_parameter_64: name = "parameter_64" + original_name = "conv2d_47.w_0" shape = [672, 112, 1, 1] dtype = "float32" min_val = float("-0.786461") @@ -715,6 +780,7 @@ class Program_weight_tensor_parameter_64: class Program_weight_tensor_parameter_65: name = "parameter_65" + original_name = "batch_norm_36.b_0" shape = [112] dtype = "float32" min_val = float("-3.23877e-06") @@ -726,6 +792,7 @@ class Program_weight_tensor_parameter_65: class Program_weight_tensor_parameter_66: name = "parameter_66" + original_name = "batch_norm_36.w_0" shape = [112] dtype = "float32" min_val = float("0.872453") @@ -737,6 +804,7 @@ class Program_weight_tensor_parameter_66: class Program_weight_tensor_parameter_67: name = "parameter_67" + original_name = "batch_norm_36.w_2" shape = [112] dtype = "float32" min_val = float("0.0271354") @@ -748,6 +816,7 @@ class Program_weight_tensor_parameter_67: class Program_weight_tensor_parameter_68: name = "parameter_68" + original_name = "batch_norm_36.w_1" shape = [112] dtype = "float32" min_val = float("-0.310959") @@ -759,6 +828,7 @@ class Program_weight_tensor_parameter_68: class Program_weight_tensor_parameter_69: name = "parameter_69" + original_name = "conv2d_46.w_0" shape = [112, 672, 1, 1] dtype = "float32" min_val = float("-0.582762") @@ -770,6 +840,7 @@ class Program_weight_tensor_parameter_69: class Program_weight_tensor_parameter_70: name = "parameter_70" + original_name = "conv2d_45.b_0" shape = [672] dtype = "float32" min_val = float("-0.19265") @@ -781,6 +852,7 @@ class Program_weight_tensor_parameter_70: class Program_weight_tensor_parameter_71: name = "parameter_71" + original_name = "conv2d_45.w_0" shape = [672, 168, 1, 1] dtype = "float32" min_val = float("-0.457763") @@ -792,6 +864,7 @@ class Program_weight_tensor_parameter_71: class Program_weight_tensor_parameter_72: name = "parameter_72" + original_name = "conv2d_44.b_0" shape = [168] dtype = "float32" min_val = float("-0.0425561") @@ -803,6 +876,7 @@ class Program_weight_tensor_parameter_72: class Program_weight_tensor_parameter_73: name = "parameter_73" + original_name = "conv2d_44.w_0" shape = [168, 672, 1, 1] dtype = "float32" min_val = float("-0.687204") @@ -814,6 +888,7 @@ class Program_weight_tensor_parameter_73: class Program_weight_tensor_parameter_74: name = "parameter_74" + original_name = "batch_norm_35.b_0" shape = [672] dtype = "float32" min_val = float("-3.0495") @@ -825,6 +900,7 @@ class Program_weight_tensor_parameter_74: class Program_weight_tensor_parameter_75: name = "parameter_75" + original_name = "batch_norm_35.w_0" shape = [672] dtype = "float32" min_val = float("0.181173") @@ -836,6 +912,7 @@ class Program_weight_tensor_parameter_75: class Program_weight_tensor_parameter_76: name = "parameter_76" + original_name = "batch_norm_35.w_2" shape = [672] dtype = "float32" min_val = float("7.72249e-05") @@ -847,6 +924,7 @@ class Program_weight_tensor_parameter_76: class Program_weight_tensor_parameter_77: name = "parameter_77" + original_name = "batch_norm_35.w_1" shape = [672] dtype = "float32" min_val = float("-2.64815") @@ -858,6 +936,7 @@ class Program_weight_tensor_parameter_77: class Program_weight_tensor_parameter_78: name = "parameter_78" + original_name = "conv2d_43.w_0" shape = [672, 1, 3, 3] dtype = "float32" min_val = float("-0.718112") @@ -869,6 +948,7 @@ class Program_weight_tensor_parameter_78: class Program_weight_tensor_parameter_79: name = "parameter_79" + original_name = "batch_norm_34.b_0" shape = [672] dtype = "float32" min_val = float("-3.53873") @@ -880,6 +960,7 @@ class Program_weight_tensor_parameter_79: class Program_weight_tensor_parameter_80: name = "parameter_80" + original_name = "batch_norm_34.w_0" shape = [672] dtype = "float32" min_val = float("-0.291747") @@ -891,6 +972,7 @@ class Program_weight_tensor_parameter_80: class Program_weight_tensor_parameter_81: name = "parameter_81" + original_name = "batch_norm_34.w_2" shape = [672] dtype = "float32" min_val = float("3.44234") @@ -902,6 +984,7 @@ class Program_weight_tensor_parameter_81: class Program_weight_tensor_parameter_82: name = "parameter_82" + original_name = "batch_norm_34.w_1" shape = [672] dtype = "float32" min_val = float("-3.78328e-06") @@ -913,6 +996,7 @@ class Program_weight_tensor_parameter_82: class Program_weight_tensor_parameter_83: name = "parameter_83" + original_name = "conv2d_42.w_0" shape = [672, 112, 1, 1] dtype = "float32" min_val = float("-0.565329") @@ -924,6 +1008,7 @@ class Program_weight_tensor_parameter_83: class Program_weight_tensor_parameter_84: name = "parameter_84" + original_name = "batch_norm_33.b_0" shape = [112] dtype = "float32" min_val = float("-2.58045e-06") @@ -935,6 +1020,7 @@ class Program_weight_tensor_parameter_84: class Program_weight_tensor_parameter_85: name = "parameter_85" + original_name = "batch_norm_33.w_0" shape = [112] dtype = "float32" min_val = float("2.1733") @@ -946,6 +1032,7 @@ class Program_weight_tensor_parameter_85: class Program_weight_tensor_parameter_86: name = "parameter_86" + original_name = "batch_norm_33.w_2" shape = [112] dtype = "float32" min_val = float("0.0596829") @@ -957,6 +1044,7 @@ class Program_weight_tensor_parameter_86: class Program_weight_tensor_parameter_87: name = "parameter_87" + original_name = "batch_norm_33.w_1" shape = [112] dtype = "float32" min_val = float("-0.526685") @@ -968,6 +1056,7 @@ class Program_weight_tensor_parameter_87: class Program_weight_tensor_parameter_88: name = "parameter_88" + original_name = "conv2d_41.w_0" shape = [112, 480, 1, 1] dtype = "float32" min_val = float("-0.816297") @@ -979,6 +1068,7 @@ class Program_weight_tensor_parameter_88: class Program_weight_tensor_parameter_89: name = "parameter_89" + original_name = "conv2d_40.b_0" shape = [480] dtype = "float32" min_val = float("-0.312372") @@ -990,6 +1080,7 @@ class Program_weight_tensor_parameter_89: class Program_weight_tensor_parameter_90: name = "parameter_90" + original_name = "conv2d_40.w_0" shape = [480, 120, 1, 1] dtype = "float32" min_val = float("-1.7023") @@ -1001,6 +1092,7 @@ class Program_weight_tensor_parameter_90: class Program_weight_tensor_parameter_91: name = "parameter_91" + original_name = "conv2d_39.b_0" shape = [120] dtype = "float32" min_val = float("-0.238304") @@ -1012,6 +1104,7 @@ class Program_weight_tensor_parameter_91: class Program_weight_tensor_parameter_92: name = "parameter_92" + original_name = "conv2d_39.w_0" shape = [120, 480, 1, 1] dtype = "float32" min_val = float("-0.968733") @@ -1023,6 +1116,7 @@ class Program_weight_tensor_parameter_92: class Program_weight_tensor_parameter_93: name = "parameter_93" + original_name = "batch_norm_32.b_0" shape = [480] dtype = "float32" min_val = float("-4.89681") @@ -1034,6 +1128,7 @@ class Program_weight_tensor_parameter_93: class Program_weight_tensor_parameter_94: name = "parameter_94" + original_name = "batch_norm_32.w_0" shape = [480] dtype = "float32" min_val = float("0.23559") @@ -1045,6 +1140,7 @@ class Program_weight_tensor_parameter_94: class Program_weight_tensor_parameter_95: name = "parameter_95" + original_name = "batch_norm_32.w_2" shape = [480] dtype = "float32" min_val = float("7.10746e-05") @@ -1056,6 +1152,7 @@ class Program_weight_tensor_parameter_95: class Program_weight_tensor_parameter_96: name = "parameter_96" + original_name = "batch_norm_32.w_1" shape = [480] dtype = "float32" min_val = float("-7.05523") @@ -1067,6 +1164,7 @@ class Program_weight_tensor_parameter_96: class Program_weight_tensor_parameter_97: name = "parameter_97" + original_name = "conv2d_38.w_0" shape = [480, 1, 3, 3] dtype = "float32" min_val = float("-0.911618") @@ -1078,6 +1176,7 @@ class Program_weight_tensor_parameter_97: class Program_weight_tensor_parameter_98: name = "parameter_98" + original_name = "batch_norm_31.b_0" shape = [480] dtype = "float32" min_val = float("-4.51436") @@ -1089,6 +1188,7 @@ class Program_weight_tensor_parameter_98: class Program_weight_tensor_parameter_99: name = "parameter_99" + original_name = "batch_norm_31.w_0" shape = [480] dtype = "float32" min_val = float("-0.991543") @@ -1100,6 +1200,7 @@ class Program_weight_tensor_parameter_99: class Program_weight_tensor_parameter_100: name = "parameter_100" + original_name = "batch_norm_31.w_2" shape = [480] dtype = "float32" min_val = float("9.33819") @@ -1111,6 +1212,7 @@ class Program_weight_tensor_parameter_100: class Program_weight_tensor_parameter_101: name = "parameter_101" + original_name = "batch_norm_31.w_1" shape = [480] dtype = "float32" min_val = float("-1.90985e-05") @@ -1122,6 +1224,7 @@ class Program_weight_tensor_parameter_101: class Program_weight_tensor_parameter_102: name = "parameter_102" + original_name = "conv2d_37.w_0" shape = [480, 80, 1, 1] dtype = "float32" min_val = float("-0.788288") @@ -1133,6 +1236,7 @@ class Program_weight_tensor_parameter_102: class Program_weight_tensor_parameter_103: name = "parameter_103" + original_name = "batch_norm_30.b_0" shape = [80] dtype = "float32" min_val = float("-2.56299e-06") @@ -1144,6 +1248,7 @@ class Program_weight_tensor_parameter_103: class Program_weight_tensor_parameter_104: name = "parameter_104" + original_name = "batch_norm_30.w_0" shape = [80] dtype = "float32" min_val = float("0.633228") @@ -1155,6 +1260,7 @@ class Program_weight_tensor_parameter_104: class Program_weight_tensor_parameter_105: name = "parameter_105" + original_name = "batch_norm_30.w_2" shape = [80] dtype = "float32" min_val = float("0.139372") @@ -1166,6 +1272,7 @@ class Program_weight_tensor_parameter_105: class Program_weight_tensor_parameter_106: name = "parameter_106" + original_name = "batch_norm_30.w_1" shape = [80] dtype = "float32" min_val = float("-3.17364") @@ -1177,6 +1284,7 @@ class Program_weight_tensor_parameter_106: class Program_weight_tensor_parameter_107: name = "parameter_107" + original_name = "conv2d_36.w_0" shape = [80, 184, 1, 1] dtype = "float32" min_val = float("-0.791077") @@ -1188,6 +1296,7 @@ class Program_weight_tensor_parameter_107: class Program_weight_tensor_parameter_108: name = "parameter_108" + original_name = "batch_norm_29.b_0" shape = [184] dtype = "float32" min_val = float("-3.08713") @@ -1199,6 +1308,7 @@ class Program_weight_tensor_parameter_108: class Program_weight_tensor_parameter_109: name = "parameter_109" + original_name = "batch_norm_29.w_0" shape = [184] dtype = "float32" min_val = float("0.33694") @@ -1210,6 +1320,7 @@ class Program_weight_tensor_parameter_109: class Program_weight_tensor_parameter_110: name = "parameter_110" + original_name = "batch_norm_29.w_2" shape = [184] dtype = "float32" min_val = float("0.000114573") @@ -1221,6 +1332,7 @@ class Program_weight_tensor_parameter_110: class Program_weight_tensor_parameter_111: name = "parameter_111" + original_name = "batch_norm_29.w_1" shape = [184] dtype = "float32" min_val = float("-1.57137") @@ -1232,6 +1344,7 @@ class Program_weight_tensor_parameter_111: class Program_weight_tensor_parameter_112: name = "parameter_112" + original_name = "conv2d_35.w_0" shape = [184, 1, 3, 3] dtype = "float32" min_val = float("-0.735884") @@ -1243,6 +1356,7 @@ class Program_weight_tensor_parameter_112: class Program_weight_tensor_parameter_113: name = "parameter_113" + original_name = "batch_norm_28.b_0" shape = [184] dtype = "float32" min_val = float("-3.52441") @@ -1254,6 +1368,7 @@ class Program_weight_tensor_parameter_113: class Program_weight_tensor_parameter_114: name = "parameter_114" + original_name = "batch_norm_28.w_0" shape = [184] dtype = "float32" min_val = float("0.0565259") @@ -1265,6 +1380,7 @@ class Program_weight_tensor_parameter_114: class Program_weight_tensor_parameter_115: name = "parameter_115" + original_name = "batch_norm_28.w_2" shape = [184] dtype = "float32" min_val = float("8.2054") @@ -1276,6 +1392,7 @@ class Program_weight_tensor_parameter_115: class Program_weight_tensor_parameter_116: name = "parameter_116" + original_name = "batch_norm_28.w_1" shape = [184] dtype = "float32" min_val = float("-8.05556e-06") @@ -1287,6 +1404,7 @@ class Program_weight_tensor_parameter_116: class Program_weight_tensor_parameter_117: name = "parameter_117" + original_name = "conv2d_34.w_0" shape = [184, 80, 1, 1] dtype = "float32" min_val = float("-0.929282") @@ -1298,6 +1416,7 @@ class Program_weight_tensor_parameter_117: class Program_weight_tensor_parameter_118: name = "parameter_118" + original_name = "batch_norm_27.b_0" shape = [80] dtype = "float32" min_val = float("-2.99623e-06") @@ -1309,6 +1428,7 @@ class Program_weight_tensor_parameter_118: class Program_weight_tensor_parameter_119: name = "parameter_119" + original_name = "batch_norm_27.w_0" shape = [80] dtype = "float32" min_val = float("0.512375") @@ -1320,6 +1440,7 @@ class Program_weight_tensor_parameter_119: class Program_weight_tensor_parameter_120: name = "parameter_120" + original_name = "batch_norm_27.w_2" shape = [80] dtype = "float32" min_val = float("0.0200513") @@ -1331,6 +1452,7 @@ class Program_weight_tensor_parameter_120: class Program_weight_tensor_parameter_121: name = "parameter_121" + original_name = "batch_norm_27.w_1" shape = [80] dtype = "float32" min_val = float("-2.00709") @@ -1342,6 +1464,7 @@ class Program_weight_tensor_parameter_121: class Program_weight_tensor_parameter_122: name = "parameter_122" + original_name = "conv2d_33.w_0" shape = [80, 184, 1, 1] dtype = "float32" min_val = float("-0.500645") @@ -1353,6 +1476,7 @@ class Program_weight_tensor_parameter_122: class Program_weight_tensor_parameter_123: name = "parameter_123" + original_name = "batch_norm_26.b_0" shape = [184] dtype = "float32" min_val = float("-2.25544") @@ -1364,6 +1488,7 @@ class Program_weight_tensor_parameter_123: class Program_weight_tensor_parameter_124: name = "parameter_124" + original_name = "batch_norm_26.w_0" shape = [184] dtype = "float32" min_val = float("0.140281") @@ -1375,6 +1500,7 @@ class Program_weight_tensor_parameter_124: class Program_weight_tensor_parameter_125: name = "parameter_125" + original_name = "batch_norm_26.w_2" shape = [184] dtype = "float32" min_val = float("7.83479e-05") @@ -1386,6 +1512,7 @@ class Program_weight_tensor_parameter_125: class Program_weight_tensor_parameter_126: name = "parameter_126" + original_name = "batch_norm_26.w_1" shape = [184] dtype = "float32" min_val = float("-2.09114") @@ -1397,6 +1524,7 @@ class Program_weight_tensor_parameter_126: class Program_weight_tensor_parameter_127: name = "parameter_127" + original_name = "conv2d_32.w_0" shape = [184, 1, 3, 3] dtype = "float32" min_val = float("-0.664416") @@ -1408,6 +1536,7 @@ class Program_weight_tensor_parameter_127: class Program_weight_tensor_parameter_128: name = "parameter_128" + original_name = "batch_norm_25.b_0" shape = [184] dtype = "float32" min_val = float("-3.49358") @@ -1419,6 +1548,7 @@ class Program_weight_tensor_parameter_128: class Program_weight_tensor_parameter_129: name = "parameter_129" + original_name = "batch_norm_25.w_0" shape = [184] dtype = "float32" min_val = float("-1.32507") @@ -1430,6 +1560,7 @@ class Program_weight_tensor_parameter_129: class Program_weight_tensor_parameter_130: name = "parameter_130" + original_name = "batch_norm_25.w_2" shape = [184] dtype = "float32" min_val = float("6.36965") @@ -1441,6 +1572,7 @@ class Program_weight_tensor_parameter_130: class Program_weight_tensor_parameter_131: name = "parameter_131" + original_name = "batch_norm_25.w_1" shape = [184] dtype = "float32" min_val = float("-5.38293e-06") @@ -1452,6 +1584,7 @@ class Program_weight_tensor_parameter_131: class Program_weight_tensor_parameter_132: name = "parameter_132" + original_name = "conv2d_31.w_0" shape = [184, 80, 1, 1] dtype = "float32" min_val = float("-0.723578") @@ -1463,6 +1596,7 @@ class Program_weight_tensor_parameter_132: class Program_weight_tensor_parameter_133: name = "parameter_133" + original_name = "batch_norm_24.b_0" shape = [80] dtype = "float32" min_val = float("-2.92183e-06") @@ -1474,6 +1608,7 @@ class Program_weight_tensor_parameter_133: class Program_weight_tensor_parameter_134: name = "parameter_134" + original_name = "batch_norm_24.w_0" shape = [80] dtype = "float32" min_val = float("0.615486") @@ -1485,6 +1620,7 @@ class Program_weight_tensor_parameter_134: class Program_weight_tensor_parameter_135: name = "parameter_135" + original_name = "batch_norm_24.w_2" shape = [80] dtype = "float32" min_val = float("0.00649837") @@ -1496,6 +1632,7 @@ class Program_weight_tensor_parameter_135: class Program_weight_tensor_parameter_136: name = "parameter_136" + original_name = "batch_norm_24.w_1" shape = [80] dtype = "float32" min_val = float("-2.72076") @@ -1507,6 +1644,7 @@ class Program_weight_tensor_parameter_136: class Program_weight_tensor_parameter_137: name = "parameter_137" + original_name = "conv2d_30.w_0" shape = [80, 200, 1, 1] dtype = "float32" min_val = float("-0.716648") @@ -1518,6 +1656,7 @@ class Program_weight_tensor_parameter_137: class Program_weight_tensor_parameter_138: name = "parameter_138" + original_name = "batch_norm_23.b_0" shape = [200] dtype = "float32" min_val = float("-3.52827") @@ -1529,6 +1668,7 @@ class Program_weight_tensor_parameter_138: class Program_weight_tensor_parameter_139: name = "parameter_139" + original_name = "batch_norm_23.w_0" shape = [200] dtype = "float32" min_val = float("0.134102") @@ -1540,6 +1680,7 @@ class Program_weight_tensor_parameter_139: class Program_weight_tensor_parameter_140: name = "parameter_140" + original_name = "batch_norm_23.w_2" shape = [200] dtype = "float32" min_val = float("0.000116902") @@ -1551,6 +1692,7 @@ class Program_weight_tensor_parameter_140: class Program_weight_tensor_parameter_141: name = "parameter_141" + original_name = "batch_norm_23.w_1" shape = [200] dtype = "float32" min_val = float("-5.10496") @@ -1562,6 +1704,7 @@ class Program_weight_tensor_parameter_141: class Program_weight_tensor_parameter_142: name = "parameter_142" + original_name = "conv2d_29.w_0" shape = [200, 1, 3, 3] dtype = "float32" min_val = float("-0.677084") @@ -1573,6 +1716,7 @@ class Program_weight_tensor_parameter_142: class Program_weight_tensor_parameter_143: name = "parameter_143" + original_name = "batch_norm_22.b_0" shape = [200] dtype = "float32" min_val = float("-5.26121") @@ -1584,6 +1728,7 @@ class Program_weight_tensor_parameter_143: class Program_weight_tensor_parameter_144: name = "parameter_144" + original_name = "batch_norm_22.w_0" shape = [200] dtype = "float32" min_val = float("-0.230809") @@ -1595,6 +1740,7 @@ class Program_weight_tensor_parameter_144: class Program_weight_tensor_parameter_145: name = "parameter_145" + original_name = "batch_norm_22.w_2" shape = [200] dtype = "float32" min_val = float("4.5139") @@ -1606,6 +1752,7 @@ class Program_weight_tensor_parameter_145: class Program_weight_tensor_parameter_146: name = "parameter_146" + original_name = "batch_norm_22.w_1" shape = [200] dtype = "float32" min_val = float("-4.6129e-06") @@ -1617,6 +1764,7 @@ class Program_weight_tensor_parameter_146: class Program_weight_tensor_parameter_147: name = "parameter_147" + original_name = "conv2d_28.w_0" shape = [200, 80, 1, 1] dtype = "float32" min_val = float("-0.57245") @@ -1628,6 +1776,7 @@ class Program_weight_tensor_parameter_147: class Program_weight_tensor_parameter_148: name = "parameter_148" + original_name = "batch_norm_21.b_0" shape = [80] dtype = "float32" min_val = float("-3.27838e-06") @@ -1639,6 +1788,7 @@ class Program_weight_tensor_parameter_148: class Program_weight_tensor_parameter_149: name = "parameter_149" + original_name = "batch_norm_21.w_0" shape = [80] dtype = "float32" min_val = float("0.903267") @@ -1650,6 +1800,7 @@ class Program_weight_tensor_parameter_149: class Program_weight_tensor_parameter_150: name = "parameter_150" + original_name = "batch_norm_21.w_2" shape = [80] dtype = "float32" min_val = float("0.795533") @@ -1661,6 +1812,7 @@ class Program_weight_tensor_parameter_150: class Program_weight_tensor_parameter_151: name = "parameter_151" + original_name = "batch_norm_21.w_1" shape = [80] dtype = "float32" min_val = float("-10.1932") @@ -1672,6 +1824,7 @@ class Program_weight_tensor_parameter_151: class Program_weight_tensor_parameter_152: name = "parameter_152" + original_name = "conv2d_27.w_0" shape = [80, 240, 1, 1] dtype = "float32" min_val = float("-0.820887") @@ -1683,6 +1836,7 @@ class Program_weight_tensor_parameter_152: class Program_weight_tensor_parameter_153: name = "parameter_153" + original_name = "batch_norm_20.b_0" shape = [240] dtype = "float32" min_val = float("-1.96145") @@ -1694,6 +1848,7 @@ class Program_weight_tensor_parameter_153: class Program_weight_tensor_parameter_154: name = "parameter_154" + original_name = "batch_norm_20.w_0" shape = [240] dtype = "float32" min_val = float("0.455213") @@ -1705,6 +1860,7 @@ class Program_weight_tensor_parameter_154: class Program_weight_tensor_parameter_155: name = "parameter_155" + original_name = "batch_norm_20.w_2" shape = [240] dtype = "float32" min_val = float("0.000213177") @@ -1716,6 +1872,7 @@ class Program_weight_tensor_parameter_155: class Program_weight_tensor_parameter_156: name = "parameter_156" + original_name = "batch_norm_20.w_1" shape = [240] dtype = "float32" min_val = float("-2.91317") @@ -1727,6 +1884,7 @@ class Program_weight_tensor_parameter_156: class Program_weight_tensor_parameter_157: name = "parameter_157" + original_name = "conv2d_26.w_0" shape = [240, 1, 3, 3] dtype = "float32" min_val = float("-0.454744") @@ -1738,6 +1896,7 @@ class Program_weight_tensor_parameter_157: class Program_weight_tensor_parameter_158: name = "parameter_158" + original_name = "batch_norm_19.b_0" shape = [240] dtype = "float32" min_val = float("-3.7168") @@ -1749,6 +1908,7 @@ class Program_weight_tensor_parameter_158: class Program_weight_tensor_parameter_159: name = "parameter_159" + original_name = "batch_norm_19.w_0" shape = [240] dtype = "float32" min_val = float("-0.599628") @@ -1760,6 +1920,7 @@ class Program_weight_tensor_parameter_159: class Program_weight_tensor_parameter_160: name = "parameter_160" + original_name = "batch_norm_19.w_2" shape = [240] dtype = "float32" min_val = float("12.6637") @@ -1771,6 +1932,7 @@ class Program_weight_tensor_parameter_160: class Program_weight_tensor_parameter_161: name = "parameter_161" + original_name = "batch_norm_19.w_1" shape = [240] dtype = "float32" min_val = float("-1.81288e-05") @@ -1782,6 +1944,7 @@ class Program_weight_tensor_parameter_161: class Program_weight_tensor_parameter_162: name = "parameter_162" + original_name = "conv2d_25.w_0" shape = [240, 40, 1, 1] dtype = "float32" min_val = float("-1.04656") @@ -1793,6 +1956,7 @@ class Program_weight_tensor_parameter_162: class Program_weight_tensor_parameter_163: name = "parameter_163" + original_name = "batch_norm_18.b_0" shape = [40] dtype = "float32" min_val = float("0") @@ -1802,6 +1966,7 @@ class Program_weight_tensor_parameter_163: class Program_weight_tensor_parameter_164: name = "parameter_164" + original_name = "batch_norm_18.w_0" shape = [40] dtype = "float32" min_val = float("0") @@ -1811,6 +1976,7 @@ class Program_weight_tensor_parameter_164: class Program_weight_tensor_parameter_165: name = "parameter_165" + original_name = "batch_norm_18.w_2" shape = [40] dtype = "float32" min_val = float("0") @@ -1820,6 +1986,7 @@ class Program_weight_tensor_parameter_165: class Program_weight_tensor_parameter_166: name = "parameter_166" + original_name = "batch_norm_18.w_1" shape = [40] dtype = "float32" min_val = float("0") @@ -1829,6 +1996,7 @@ class Program_weight_tensor_parameter_166: class Program_weight_tensor_parameter_167: name = "parameter_167" + original_name = "conv2d_24.w_0" shape = [40, 120, 1, 1] dtype = "float32" min_val = float("-0.638971") @@ -1840,6 +2008,7 @@ class Program_weight_tensor_parameter_167: class Program_weight_tensor_parameter_168: name = "parameter_168" + original_name = "conv2d_23.b_0" shape = [120] dtype = "float32" min_val = float("-0.208798") @@ -1851,6 +2020,7 @@ class Program_weight_tensor_parameter_168: class Program_weight_tensor_parameter_169: name = "parameter_169" + original_name = "conv2d_23.w_0" shape = [120, 30, 1, 1] dtype = "float32" min_val = float("-0.917489") @@ -1862,6 +2032,7 @@ class Program_weight_tensor_parameter_169: class Program_weight_tensor_parameter_170: name = "parameter_170" + original_name = "conv2d_22.b_0" shape = [30] dtype = "float32" min_val = float("0") @@ -1871,6 +2042,7 @@ class Program_weight_tensor_parameter_170: class Program_weight_tensor_parameter_171: name = "parameter_171" + original_name = "conv2d_22.w_0" shape = [30, 120, 1, 1] dtype = "float32" min_val = float("-1.22326") @@ -1882,6 +2054,7 @@ class Program_weight_tensor_parameter_171: class Program_weight_tensor_parameter_172: name = "parameter_172" + original_name = "batch_norm_17.b_0" shape = [120] dtype = "float32" min_val = float("-7.18435") @@ -1893,6 +2066,7 @@ class Program_weight_tensor_parameter_172: class Program_weight_tensor_parameter_173: name = "parameter_173" + original_name = "batch_norm_17.w_0" shape = [120] dtype = "float32" min_val = float("0.250863") @@ -1904,6 +2078,7 @@ class Program_weight_tensor_parameter_173: class Program_weight_tensor_parameter_174: name = "parameter_174" + original_name = "batch_norm_17.w_2" shape = [120] dtype = "float32" min_val = float("0.016154") @@ -1915,6 +2090,7 @@ class Program_weight_tensor_parameter_174: class Program_weight_tensor_parameter_175: name = "parameter_175" + original_name = "batch_norm_17.w_1" shape = [120] dtype = "float32" min_val = float("-6.65649") @@ -1926,6 +2102,7 @@ class Program_weight_tensor_parameter_175: class Program_weight_tensor_parameter_176: name = "parameter_176" + original_name = "conv2d_21.w_0" shape = [120, 1, 5, 5] dtype = "float32" min_val = float("-0.904495") @@ -1937,6 +2114,7 @@ class Program_weight_tensor_parameter_176: class Program_weight_tensor_parameter_177: name = "parameter_177" + original_name = "batch_norm_16.b_0" shape = [120] dtype = "float32" min_val = float("-4.01876") @@ -1948,6 +2126,7 @@ class Program_weight_tensor_parameter_177: class Program_weight_tensor_parameter_178: name = "parameter_178" + original_name = "batch_norm_16.w_0" shape = [120] dtype = "float32" min_val = float("0.464227") @@ -1959,6 +2138,7 @@ class Program_weight_tensor_parameter_178: class Program_weight_tensor_parameter_179: name = "parameter_179" + original_name = "batch_norm_16.w_2" shape = [120] dtype = "float32" min_val = float("9.68358") @@ -1970,6 +2150,7 @@ class Program_weight_tensor_parameter_179: class Program_weight_tensor_parameter_180: name = "parameter_180" + original_name = "batch_norm_16.w_1" shape = [120] dtype = "float32" min_val = float("-9.60327e-06") @@ -1981,6 +2162,7 @@ class Program_weight_tensor_parameter_180: class Program_weight_tensor_parameter_181: name = "parameter_181" + original_name = "conv2d_20.w_0" shape = [120, 40, 1, 1] dtype = "float32" min_val = float("-0.697068") @@ -1992,6 +2174,7 @@ class Program_weight_tensor_parameter_181: class Program_weight_tensor_parameter_182: name = "parameter_182" + original_name = "batch_norm_15.b_0" shape = [40] dtype = "float32" min_val = float("0") @@ -2001,6 +2184,7 @@ class Program_weight_tensor_parameter_182: class Program_weight_tensor_parameter_183: name = "parameter_183" + original_name = "batch_norm_15.w_0" shape = [40] dtype = "float32" min_val = float("0") @@ -2010,6 +2194,7 @@ class Program_weight_tensor_parameter_183: class Program_weight_tensor_parameter_184: name = "parameter_184" + original_name = "batch_norm_15.w_2" shape = [40] dtype = "float32" min_val = float("0") @@ -2019,6 +2204,7 @@ class Program_weight_tensor_parameter_184: class Program_weight_tensor_parameter_185: name = "parameter_185" + original_name = "batch_norm_15.w_1" shape = [40] dtype = "float32" min_val = float("0") @@ -2028,6 +2214,7 @@ class Program_weight_tensor_parameter_185: class Program_weight_tensor_parameter_186: name = "parameter_186" + original_name = "conv2d_19.w_0" shape = [40, 120, 1, 1] dtype = "float32" min_val = float("-0.837616") @@ -2039,6 +2226,7 @@ class Program_weight_tensor_parameter_186: class Program_weight_tensor_parameter_187: name = "parameter_187" + original_name = "conv2d_18.b_0" shape = [120] dtype = "float32" min_val = float("-0.194223") @@ -2050,6 +2238,7 @@ class Program_weight_tensor_parameter_187: class Program_weight_tensor_parameter_188: name = "parameter_188" + original_name = "conv2d_18.w_0" shape = [120, 30, 1, 1] dtype = "float32" min_val = float("-0.848496") @@ -2061,6 +2250,7 @@ class Program_weight_tensor_parameter_188: class Program_weight_tensor_parameter_189: name = "parameter_189" + original_name = "conv2d_17.b_0" shape = [30] dtype = "float32" min_val = float("0") @@ -2070,6 +2260,7 @@ class Program_weight_tensor_parameter_189: class Program_weight_tensor_parameter_190: name = "parameter_190" + original_name = "conv2d_17.w_0" shape = [30, 120, 1, 1] dtype = "float32" min_val = float("-1.86266") @@ -2081,6 +2272,7 @@ class Program_weight_tensor_parameter_190: class Program_weight_tensor_parameter_191: name = "parameter_191" + original_name = "batch_norm_14.b_0" shape = [120] dtype = "float32" min_val = float("-6.59654") @@ -2092,6 +2284,7 @@ class Program_weight_tensor_parameter_191: class Program_weight_tensor_parameter_192: name = "parameter_192" + original_name = "batch_norm_14.w_0" shape = [120] dtype = "float32" min_val = float("0.262511") @@ -2103,6 +2296,7 @@ class Program_weight_tensor_parameter_192: class Program_weight_tensor_parameter_193: name = "parameter_193" + original_name = "batch_norm_14.w_2" shape = [120] dtype = "float32" min_val = float("0.00258724") @@ -2114,6 +2308,7 @@ class Program_weight_tensor_parameter_193: class Program_weight_tensor_parameter_194: name = "parameter_194" + original_name = "batch_norm_14.w_1" shape = [120] dtype = "float32" min_val = float("-9.7533") @@ -2125,6 +2320,7 @@ class Program_weight_tensor_parameter_194: class Program_weight_tensor_parameter_195: name = "parameter_195" + original_name = "conv2d_16.w_0" shape = [120, 1, 5, 5] dtype = "float32" min_val = float("-0.962315") @@ -2136,6 +2332,7 @@ class Program_weight_tensor_parameter_195: class Program_weight_tensor_parameter_196: name = "parameter_196" + original_name = "batch_norm_13.b_0" shape = [120] dtype = "float32" min_val = float("-5.25466") @@ -2147,6 +2344,7 @@ class Program_weight_tensor_parameter_196: class Program_weight_tensor_parameter_197: name = "parameter_197" + original_name = "batch_norm_13.w_0" shape = [120] dtype = "float32" min_val = float("0.296514") @@ -2158,6 +2356,7 @@ class Program_weight_tensor_parameter_197: class Program_weight_tensor_parameter_198: name = "parameter_198" + original_name = "batch_norm_13.w_2" shape = [120] dtype = "float32" min_val = float("9.31996") @@ -2169,6 +2368,7 @@ class Program_weight_tensor_parameter_198: class Program_weight_tensor_parameter_199: name = "parameter_199" + original_name = "batch_norm_13.w_1" shape = [120] dtype = "float32" min_val = float("-4.6802e-06") @@ -2180,6 +2380,7 @@ class Program_weight_tensor_parameter_199: class Program_weight_tensor_parameter_200: name = "parameter_200" + original_name = "conv2d_15.w_0" shape = [120, 40, 1, 1] dtype = "float32" min_val = float("-0.869297") @@ -2191,6 +2392,7 @@ class Program_weight_tensor_parameter_200: class Program_weight_tensor_parameter_201: name = "parameter_201" + original_name = "batch_norm_12.b_0" shape = [40] dtype = "float32" min_val = float("0") @@ -2200,6 +2402,7 @@ class Program_weight_tensor_parameter_201: class Program_weight_tensor_parameter_202: name = "parameter_202" + original_name = "batch_norm_12.w_0" shape = [40] dtype = "float32" min_val = float("0") @@ -2209,6 +2412,7 @@ class Program_weight_tensor_parameter_202: class Program_weight_tensor_parameter_203: name = "parameter_203" + original_name = "batch_norm_12.w_2" shape = [40] dtype = "float32" min_val = float("0") @@ -2218,6 +2422,7 @@ class Program_weight_tensor_parameter_203: class Program_weight_tensor_parameter_204: name = "parameter_204" + original_name = "batch_norm_12.w_1" shape = [40] dtype = "float32" min_val = float("0") @@ -2227,6 +2432,7 @@ class Program_weight_tensor_parameter_204: class Program_weight_tensor_parameter_205: name = "parameter_205" + original_name = "conv2d_14.w_0" shape = [40, 72, 1, 1] dtype = "float32" min_val = float("-1.74471") @@ -2238,6 +2444,7 @@ class Program_weight_tensor_parameter_205: class Program_weight_tensor_parameter_206: name = "parameter_206" + original_name = "conv2d_13.b_0" shape = [72] dtype = "float32" min_val = float("-0.195335") @@ -2249,6 +2456,7 @@ class Program_weight_tensor_parameter_206: class Program_weight_tensor_parameter_207: name = "parameter_207" + original_name = "conv2d_13.w_0" shape = [72, 18, 1, 1] dtype = "float32" min_val = float("-9.85057e-13") @@ -2260,6 +2468,7 @@ class Program_weight_tensor_parameter_207: class Program_weight_tensor_parameter_208: name = "parameter_208" + original_name = "conv2d_12.b_0" shape = [18] dtype = "float32" min_val = float("0") @@ -2269,6 +2478,7 @@ class Program_weight_tensor_parameter_208: class Program_weight_tensor_parameter_209: name = "parameter_209" + original_name = "conv2d_12.w_0" shape = [18, 72, 1, 1] dtype = "float32" min_val = float("-2.42846e-12") @@ -2280,6 +2490,7 @@ class Program_weight_tensor_parameter_209: class Program_weight_tensor_parameter_210: name = "parameter_210" + original_name = "batch_norm_11.b_0" shape = [72] dtype = "float32" min_val = float("-1.26211") @@ -2291,6 +2502,7 @@ class Program_weight_tensor_parameter_210: class Program_weight_tensor_parameter_211: name = "parameter_211" + original_name = "batch_norm_11.w_0" shape = [72] dtype = "float32" min_val = float("1.59861") @@ -2302,6 +2514,7 @@ class Program_weight_tensor_parameter_211: class Program_weight_tensor_parameter_212: name = "parameter_212" + original_name = "batch_norm_11.w_2" shape = [72] dtype = "float32" min_val = float("0.280344") @@ -2313,6 +2526,7 @@ class Program_weight_tensor_parameter_212: class Program_weight_tensor_parameter_213: name = "parameter_213" + original_name = "batch_norm_11.w_1" shape = [72] dtype = "float32" min_val = float("-12.0204") @@ -2324,6 +2538,7 @@ class Program_weight_tensor_parameter_213: class Program_weight_tensor_parameter_214: name = "parameter_214" + original_name = "conv2d_11.w_0" shape = [72, 1, 5, 5] dtype = "float32" min_val = float("-0.972593") @@ -2335,6 +2550,7 @@ class Program_weight_tensor_parameter_214: class Program_weight_tensor_parameter_215: name = "parameter_215" + original_name = "batch_norm_10.b_0" shape = [72] dtype = "float32" min_val = float("-3.99948") @@ -2346,6 +2562,7 @@ class Program_weight_tensor_parameter_215: class Program_weight_tensor_parameter_216: name = "parameter_216" + original_name = "batch_norm_10.w_0" shape = [72] dtype = "float32" min_val = float("1.01887") @@ -2357,6 +2574,7 @@ class Program_weight_tensor_parameter_216: class Program_weight_tensor_parameter_217: name = "parameter_217" + original_name = "batch_norm_10.w_2" shape = [72] dtype = "float32" min_val = float("16.5667") @@ -2368,6 +2586,7 @@ class Program_weight_tensor_parameter_217: class Program_weight_tensor_parameter_218: name = "parameter_218" + original_name = "batch_norm_10.w_1" shape = [72] dtype = "float32" min_val = float("-5.38923e-05") @@ -2379,6 +2598,7 @@ class Program_weight_tensor_parameter_218: class Program_weight_tensor_parameter_219: name = "parameter_219" + original_name = "conv2d_10.w_0" shape = [72, 24, 1, 1] dtype = "float32" min_val = float("-1.88725") @@ -2390,6 +2610,7 @@ class Program_weight_tensor_parameter_219: class Program_weight_tensor_parameter_220: name = "parameter_220" + original_name = "batch_norm_9.b_0" shape = [24] dtype = "float32" min_val = float("0") @@ -2399,6 +2620,7 @@ class Program_weight_tensor_parameter_220: class Program_weight_tensor_parameter_221: name = "parameter_221" + original_name = "batch_norm_9.w_0" shape = [24] dtype = "float32" min_val = float("0") @@ -2408,6 +2630,7 @@ class Program_weight_tensor_parameter_221: class Program_weight_tensor_parameter_222: name = "parameter_222" + original_name = "batch_norm_9.w_2" shape = [24] dtype = "float32" min_val = float("0") @@ -2417,6 +2640,7 @@ class Program_weight_tensor_parameter_222: class Program_weight_tensor_parameter_223: name = "parameter_223" + original_name = "batch_norm_9.w_1" shape = [24] dtype = "float32" min_val = float("0") @@ -2426,6 +2650,7 @@ class Program_weight_tensor_parameter_223: class Program_weight_tensor_parameter_224: name = "parameter_224" + original_name = "conv2d_9.w_0" shape = [24, 72, 1, 1] dtype = "float32" min_val = float("-0.841199") @@ -2437,6 +2662,7 @@ class Program_weight_tensor_parameter_224: class Program_weight_tensor_parameter_225: name = "parameter_225" + original_name = "batch_norm_8.b_0" shape = [72] dtype = "float32" min_val = float("-5.34795") @@ -2448,6 +2674,7 @@ class Program_weight_tensor_parameter_225: class Program_weight_tensor_parameter_226: name = "parameter_226" + original_name = "batch_norm_8.w_0" shape = [72] dtype = "float32" min_val = float("1.06751") @@ -2459,6 +2686,7 @@ class Program_weight_tensor_parameter_226: class Program_weight_tensor_parameter_227: name = "parameter_227" + original_name = "batch_norm_8.w_2" shape = [72] dtype = "float32" min_val = float("0.00601087") @@ -2470,6 +2698,7 @@ class Program_weight_tensor_parameter_227: class Program_weight_tensor_parameter_228: name = "parameter_228" + original_name = "batch_norm_8.w_1" shape = [72] dtype = "float32" min_val = float("-2.91857") @@ -2481,6 +2710,7 @@ class Program_weight_tensor_parameter_228: class Program_weight_tensor_parameter_229: name = "parameter_229" + original_name = "conv2d_8.w_0" shape = [72, 1, 3, 3] dtype = "float32" min_val = float("-1.39507") @@ -2492,6 +2722,7 @@ class Program_weight_tensor_parameter_229: class Program_weight_tensor_parameter_230: name = "parameter_230" + original_name = "batch_norm_7.b_0" shape = [72] dtype = "float32" min_val = float("-4.71653") @@ -2503,6 +2734,7 @@ class Program_weight_tensor_parameter_230: class Program_weight_tensor_parameter_231: name = "parameter_231" + original_name = "batch_norm_7.w_0" shape = [72] dtype = "float32" min_val = float("0.577871") @@ -2514,6 +2746,7 @@ class Program_weight_tensor_parameter_231: class Program_weight_tensor_parameter_232: name = "parameter_232" + original_name = "batch_norm_7.w_2" shape = [72] dtype = "float32" min_val = float("9.52495") @@ -2525,6 +2758,7 @@ class Program_weight_tensor_parameter_232: class Program_weight_tensor_parameter_233: name = "parameter_233" + original_name = "batch_norm_7.w_1" shape = [72] dtype = "float32" min_val = float("-2.01434e-05") @@ -2536,6 +2770,7 @@ class Program_weight_tensor_parameter_233: class Program_weight_tensor_parameter_234: name = "parameter_234" + original_name = "conv2d_7.w_0" shape = [72, 24, 1, 1] dtype = "float32" min_val = float("-1.26761") @@ -2547,6 +2782,7 @@ class Program_weight_tensor_parameter_234: class Program_weight_tensor_parameter_235: name = "parameter_235" + original_name = "batch_norm_6.b_0" shape = [24] dtype = "float32" min_val = float("0") @@ -2556,6 +2792,7 @@ class Program_weight_tensor_parameter_235: class Program_weight_tensor_parameter_236: name = "parameter_236" + original_name = "batch_norm_6.w_0" shape = [24] dtype = "float32" min_val = float("0") @@ -2565,6 +2802,7 @@ class Program_weight_tensor_parameter_236: class Program_weight_tensor_parameter_237: name = "parameter_237" + original_name = "batch_norm_6.w_2" shape = [24] dtype = "float32" min_val = float("0") @@ -2574,6 +2812,7 @@ class Program_weight_tensor_parameter_237: class Program_weight_tensor_parameter_238: name = "parameter_238" + original_name = "batch_norm_6.w_1" shape = [24] dtype = "float32" min_val = float("0") @@ -2583,6 +2822,7 @@ class Program_weight_tensor_parameter_238: class Program_weight_tensor_parameter_239: name = "parameter_239" + original_name = "conv2d_6.w_0" shape = [24, 64, 1, 1] dtype = "float32" min_val = float("-1.92209") @@ -2594,6 +2834,7 @@ class Program_weight_tensor_parameter_239: class Program_weight_tensor_parameter_240: name = "parameter_240" + original_name = "batch_norm_5.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -2603,6 +2844,7 @@ class Program_weight_tensor_parameter_240: class Program_weight_tensor_parameter_241: name = "parameter_241" + original_name = "batch_norm_5.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -2612,6 +2854,7 @@ class Program_weight_tensor_parameter_241: class Program_weight_tensor_parameter_242: name = "parameter_242" + original_name = "batch_norm_5.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -2621,6 +2864,7 @@ class Program_weight_tensor_parameter_242: class Program_weight_tensor_parameter_243: name = "parameter_243" + original_name = "batch_norm_5.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -2630,6 +2874,7 @@ class Program_weight_tensor_parameter_243: class Program_weight_tensor_parameter_244: name = "parameter_244" + original_name = "conv2d_5.w_0" shape = [64, 1, 3, 3] dtype = "float32" min_val = float("-0.906474") @@ -2641,6 +2886,7 @@ class Program_weight_tensor_parameter_244: class Program_weight_tensor_parameter_245: name = "parameter_245" + original_name = "batch_norm_4.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -2650,6 +2896,7 @@ class Program_weight_tensor_parameter_245: class Program_weight_tensor_parameter_246: name = "parameter_246" + original_name = "batch_norm_4.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -2659,6 +2906,7 @@ class Program_weight_tensor_parameter_246: class Program_weight_tensor_parameter_247: name = "parameter_247" + original_name = "batch_norm_4.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -2668,6 +2916,7 @@ class Program_weight_tensor_parameter_247: class Program_weight_tensor_parameter_248: name = "parameter_248" + original_name = "batch_norm_4.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -2677,6 +2926,7 @@ class Program_weight_tensor_parameter_248: class Program_weight_tensor_parameter_249: name = "parameter_249" + original_name = "conv2d_4.w_0" shape = [64, 16, 1, 1] dtype = "float32" min_val = float("-2.05556") @@ -2688,6 +2938,7 @@ class Program_weight_tensor_parameter_249: class Program_weight_tensor_parameter_250: name = "parameter_250" + original_name = "batch_norm_3.b_0" shape = [16] dtype = "float32" min_val = float("0") @@ -2697,6 +2948,7 @@ class Program_weight_tensor_parameter_250: class Program_weight_tensor_parameter_251: name = "parameter_251" + original_name = "batch_norm_3.w_0" shape = [16] dtype = "float32" min_val = float("0") @@ -2706,6 +2958,7 @@ class Program_weight_tensor_parameter_251: class Program_weight_tensor_parameter_252: name = "parameter_252" + original_name = "batch_norm_3.w_2" shape = [16] dtype = "float32" min_val = float("0") @@ -2715,6 +2968,7 @@ class Program_weight_tensor_parameter_252: class Program_weight_tensor_parameter_253: name = "parameter_253" + original_name = "batch_norm_3.w_1" shape = [16] dtype = "float32" min_val = float("0") @@ -2724,6 +2978,7 @@ class Program_weight_tensor_parameter_253: class Program_weight_tensor_parameter_254: name = "parameter_254" + original_name = "conv2d_3.w_0" shape = [16, 16, 1, 1] dtype = "float32" min_val = float("-1.14987") @@ -2735,6 +2990,7 @@ class Program_weight_tensor_parameter_254: class Program_weight_tensor_parameter_255: name = "parameter_255" + original_name = "batch_norm_2.b_0" shape = [16] dtype = "float32" min_val = float("0") @@ -2744,6 +3000,7 @@ class Program_weight_tensor_parameter_255: class Program_weight_tensor_parameter_256: name = "parameter_256" + original_name = "batch_norm_2.w_0" shape = [16] dtype = "float32" min_val = float("0") @@ -2753,6 +3010,7 @@ class Program_weight_tensor_parameter_256: class Program_weight_tensor_parameter_257: name = "parameter_257" + original_name = "batch_norm_2.w_2" shape = [16] dtype = "float32" min_val = float("0") @@ -2762,6 +3020,7 @@ class Program_weight_tensor_parameter_257: class Program_weight_tensor_parameter_258: name = "parameter_258" + original_name = "batch_norm_2.w_1" shape = [16] dtype = "float32" min_val = float("0") @@ -2771,6 +3030,7 @@ class Program_weight_tensor_parameter_258: class Program_weight_tensor_parameter_259: name = "parameter_259" + original_name = "conv2d_2.w_0" shape = [16, 1, 3, 3] dtype = "float32" min_val = float("-1.30216") @@ -2782,6 +3042,7 @@ class Program_weight_tensor_parameter_259: class Program_weight_tensor_parameter_260: name = "parameter_260" + original_name = "batch_norm_1.b_0" shape = [16] dtype = "float32" min_val = float("0") @@ -2791,6 +3052,7 @@ class Program_weight_tensor_parameter_260: class Program_weight_tensor_parameter_261: name = "parameter_261" + original_name = "batch_norm_1.w_0" shape = [16] dtype = "float32" min_val = float("0") @@ -2800,6 +3062,7 @@ class Program_weight_tensor_parameter_261: class Program_weight_tensor_parameter_262: name = "parameter_262" + original_name = "batch_norm_1.w_2" shape = [16] dtype = "float32" min_val = float("0") @@ -2809,6 +3072,7 @@ class Program_weight_tensor_parameter_262: class Program_weight_tensor_parameter_263: name = "parameter_263" + original_name = "batch_norm_1.w_1" shape = [16] dtype = "float32" min_val = float("0") @@ -2818,6 +3082,7 @@ class Program_weight_tensor_parameter_263: class Program_weight_tensor_parameter_264: name = "parameter_264" + original_name = "conv2d_1.w_0" shape = [16, 16, 1, 1] dtype = "float32" min_val = float("-1.29094") @@ -2829,6 +3094,7 @@ class Program_weight_tensor_parameter_264: class Program_weight_tensor_parameter_265: name = "parameter_265" + original_name = "batch_norm_0.b_0" shape = [16] dtype = "float32" min_val = float("0") @@ -2838,6 +3104,7 @@ class Program_weight_tensor_parameter_265: class Program_weight_tensor_parameter_266: name = "parameter_266" + original_name = "batch_norm_0.w_0" shape = [16] dtype = "float32" min_val = float("0") @@ -2847,6 +3114,7 @@ class Program_weight_tensor_parameter_266: class Program_weight_tensor_parameter_267: name = "parameter_267" + original_name = "batch_norm_0.w_2" shape = [16] dtype = "float32" min_val = float("0") @@ -2856,6 +3124,7 @@ class Program_weight_tensor_parameter_267: class Program_weight_tensor_parameter_268: name = "parameter_268" + original_name = "batch_norm_0.w_1" shape = [16] dtype = "float32" min_val = float("0") @@ -2865,6 +3134,7 @@ class Program_weight_tensor_parameter_268: class Program_weight_tensor_parameter_269: name = "parameter_269" + original_name = "conv2d_0.w_0" shape = [16, 3, 3, 3] dtype = "float32" min_val = float("-2.92224") diff --git a/paddle_samples/PaddleX/MobileNetV3_large_x1_25/subgraph_1/input_meta.py b/paddle_samples/PaddleX/MobileNetV3_large_x1_25/subgraph_1/input_meta.py index 795f9bcfe..6d7dac1c9 100644 --- a/paddle_samples/PaddleX/MobileNetV3_large_x1_25/subgraph_1/input_meta.py +++ b/paddle_samples/PaddleX/MobileNetV3_large_x1_25/subgraph_1/input_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_data_0: name = "data_0" + original_name = "var_0" shape = [128, 3, 224, 224] dtype = "float32" min_val = float("-2.1179") diff --git a/paddle_samples/PaddleX/MobileNetV3_large_x1_25/subgraph_1/weight_meta.py b/paddle_samples/PaddleX/MobileNetV3_large_x1_25/subgraph_1/weight_meta.py index 5fc50839d..292db129e 100644 --- a/paddle_samples/PaddleX/MobileNetV3_large_x1_25/subgraph_1/weight_meta.py +++ b/paddle_samples/PaddleX/MobileNetV3_large_x1_25/subgraph_1/weight_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_parameter_0: name = "parameter_0" + original_name = "linear_0.b_0" shape = [102] dtype = "float32" data = None @@ -7,6 +8,7 @@ class Program_weight_tensor_parameter_0: class Program_weight_tensor_parameter_1: name = "parameter_1" + original_name = "linear_0.w_0" shape = [1280, 102] dtype = "float32" min_val = float("-0.0658899") @@ -18,6 +20,7 @@ class Program_weight_tensor_parameter_1: class Program_weight_tensor_parameter_2: name = "parameter_2" + original_name = "conv2d_63.w_0" shape = [1280, 1200, 1, 1] dtype = "float32" min_val = float("-0.327378") @@ -29,6 +32,7 @@ class Program_weight_tensor_parameter_2: class Program_weight_tensor_parameter_3: name = "parameter_3" + original_name = "batch_norm_46.b_0" shape = [1200] dtype = "float32" min_val = float("-8.47213") @@ -40,6 +44,7 @@ class Program_weight_tensor_parameter_3: class Program_weight_tensor_parameter_4: name = "parameter_4" + original_name = "batch_norm_46.w_0" shape = [1200] dtype = "float32" min_val = float("0.918406") @@ -51,6 +56,7 @@ class Program_weight_tensor_parameter_4: class Program_weight_tensor_parameter_5: name = "parameter_5" + original_name = "batch_norm_46.w_2" shape = [1200] dtype = "float32" min_val = float("4.16248") @@ -62,6 +68,7 @@ class Program_weight_tensor_parameter_5: class Program_weight_tensor_parameter_6: name = "parameter_6" + original_name = "batch_norm_46.w_1" shape = [1200] dtype = "float32" min_val = float("-1.56611e-05") @@ -73,6 +80,7 @@ class Program_weight_tensor_parameter_6: class Program_weight_tensor_parameter_7: name = "parameter_7" + original_name = "conv2d_62.w_0" shape = [1200, 200, 1, 1] dtype = "float32" min_val = float("-0.374435") @@ -84,6 +92,7 @@ class Program_weight_tensor_parameter_7: class Program_weight_tensor_parameter_8: name = "parameter_8" + original_name = "batch_norm_45.b_0" shape = [200] dtype = "float32" min_val = float("-2.7289e-06") @@ -95,6 +104,7 @@ class Program_weight_tensor_parameter_8: class Program_weight_tensor_parameter_9: name = "parameter_9" + original_name = "batch_norm_45.w_0" shape = [200] dtype = "float32" min_val = float("1.54917") @@ -106,6 +116,7 @@ class Program_weight_tensor_parameter_9: class Program_weight_tensor_parameter_10: name = "parameter_10" + original_name = "batch_norm_45.w_2" shape = [200] dtype = "float32" min_val = float("0.00874101") @@ -117,6 +128,7 @@ class Program_weight_tensor_parameter_10: class Program_weight_tensor_parameter_11: name = "parameter_11" + original_name = "batch_norm_45.w_1" shape = [200] dtype = "float32" min_val = float("-0.268502") @@ -128,6 +140,7 @@ class Program_weight_tensor_parameter_11: class Program_weight_tensor_parameter_12: name = "parameter_12" + original_name = "conv2d_61.w_0" shape = [200, 1200, 1, 1] dtype = "float32" min_val = float("-0.315683") @@ -139,6 +152,7 @@ class Program_weight_tensor_parameter_12: class Program_weight_tensor_parameter_13: name = "parameter_13" + original_name = "conv2d_60.b_0" shape = [1200] dtype = "float32" min_val = float("-0.154426") @@ -150,6 +164,7 @@ class Program_weight_tensor_parameter_13: class Program_weight_tensor_parameter_14: name = "parameter_14" + original_name = "conv2d_60.w_0" shape = [1200, 300, 1, 1] dtype = "float32" min_val = float("-0.332669") @@ -161,6 +176,7 @@ class Program_weight_tensor_parameter_14: class Program_weight_tensor_parameter_15: name = "parameter_15" + original_name = "conv2d_59.b_0" shape = [300] dtype = "float32" min_val = float("-0.0405183") @@ -172,6 +188,7 @@ class Program_weight_tensor_parameter_15: class Program_weight_tensor_parameter_16: name = "parameter_16" + original_name = "conv2d_59.w_0" shape = [300, 1200, 1, 1] dtype = "float32" min_val = float("-0.199494") @@ -183,6 +200,7 @@ class Program_weight_tensor_parameter_16: class Program_weight_tensor_parameter_17: name = "parameter_17" + original_name = "batch_norm_44.b_0" shape = [1200] dtype = "float32" min_val = float("-3.94283") @@ -194,6 +212,7 @@ class Program_weight_tensor_parameter_17: class Program_weight_tensor_parameter_18: name = "parameter_18" + original_name = "batch_norm_44.w_0" shape = [1200] dtype = "float32" min_val = float("0.0412022") @@ -205,6 +224,7 @@ class Program_weight_tensor_parameter_18: class Program_weight_tensor_parameter_19: name = "parameter_19" + original_name = "batch_norm_44.w_2" shape = [1200] dtype = "float32" min_val = float("3.64079e-05") @@ -216,6 +236,7 @@ class Program_weight_tensor_parameter_19: class Program_weight_tensor_parameter_20: name = "parameter_20" + original_name = "batch_norm_44.w_1" shape = [1200] dtype = "float32" min_val = float("-0.848527") @@ -227,6 +248,7 @@ class Program_weight_tensor_parameter_20: class Program_weight_tensor_parameter_21: name = "parameter_21" + original_name = "conv2d_58.w_0" shape = [1200, 1, 5, 5] dtype = "float32" min_val = float("-0.47925") @@ -238,6 +260,7 @@ class Program_weight_tensor_parameter_21: class Program_weight_tensor_parameter_22: name = "parameter_22" + original_name = "batch_norm_43.b_0" shape = [1200] dtype = "float32" min_val = float("-3.9802") @@ -249,6 +272,7 @@ class Program_weight_tensor_parameter_22: class Program_weight_tensor_parameter_23: name = "parameter_23" + original_name = "batch_norm_43.w_0" shape = [1200] dtype = "float32" min_val = float("-1.17098") @@ -260,6 +284,7 @@ class Program_weight_tensor_parameter_23: class Program_weight_tensor_parameter_24: name = "parameter_24" + original_name = "batch_norm_43.w_2" shape = [1200] dtype = "float32" min_val = float("1.02315") @@ -271,6 +296,7 @@ class Program_weight_tensor_parameter_24: class Program_weight_tensor_parameter_25: name = "parameter_25" + original_name = "batch_norm_43.w_1" shape = [1200] dtype = "float32" min_val = float("-8.79847e-06") @@ -282,6 +308,7 @@ class Program_weight_tensor_parameter_25: class Program_weight_tensor_parameter_26: name = "parameter_26" + original_name = "conv2d_57.w_0" shape = [1200, 200, 1, 1] dtype = "float32" min_val = float("-0.367223") @@ -293,6 +320,7 @@ class Program_weight_tensor_parameter_26: class Program_weight_tensor_parameter_27: name = "parameter_27" + original_name = "batch_norm_42.b_0" shape = [200] dtype = "float32" min_val = float("-2.95943e-06") @@ -304,6 +332,7 @@ class Program_weight_tensor_parameter_27: class Program_weight_tensor_parameter_28: name = "parameter_28" + original_name = "batch_norm_42.w_0" shape = [200] dtype = "float32" min_val = float("0.722258") @@ -315,6 +344,7 @@ class Program_weight_tensor_parameter_28: class Program_weight_tensor_parameter_29: name = "parameter_29" + original_name = "batch_norm_42.w_2" shape = [200] dtype = "float32" min_val = float("0.0129515") @@ -326,6 +356,7 @@ class Program_weight_tensor_parameter_29: class Program_weight_tensor_parameter_30: name = "parameter_30" + original_name = "batch_norm_42.w_1" shape = [200] dtype = "float32" min_val = float("-0.150664") @@ -337,6 +368,7 @@ class Program_weight_tensor_parameter_30: class Program_weight_tensor_parameter_31: name = "parameter_31" + original_name = "conv2d_56.w_0" shape = [200, 1200, 1, 1] dtype = "float32" min_val = float("-0.22769") @@ -348,6 +380,7 @@ class Program_weight_tensor_parameter_31: class Program_weight_tensor_parameter_32: name = "parameter_32" + original_name = "conv2d_55.b_0" shape = [1200] dtype = "float32" min_val = float("-0.107227") @@ -359,6 +392,7 @@ class Program_weight_tensor_parameter_32: class Program_weight_tensor_parameter_33: name = "parameter_33" + original_name = "conv2d_55.w_0" shape = [1200, 300, 1, 1] dtype = "float32" min_val = float("-0.229842") @@ -370,6 +404,7 @@ class Program_weight_tensor_parameter_33: class Program_weight_tensor_parameter_34: name = "parameter_34" + original_name = "conv2d_54.b_0" shape = [300] dtype = "float32" min_val = float("-0.0119445") @@ -381,6 +416,7 @@ class Program_weight_tensor_parameter_34: class Program_weight_tensor_parameter_35: name = "parameter_35" + original_name = "conv2d_54.w_0" shape = [300, 1200, 1, 1] dtype = "float32" min_val = float("-0.319578") @@ -392,6 +428,7 @@ class Program_weight_tensor_parameter_35: class Program_weight_tensor_parameter_36: name = "parameter_36" + original_name = "batch_norm_41.b_0" shape = [1200] dtype = "float32" min_val = float("-3.35187") @@ -403,6 +440,7 @@ class Program_weight_tensor_parameter_36: class Program_weight_tensor_parameter_37: name = "parameter_37" + original_name = "batch_norm_41.w_0" shape = [1200] dtype = "float32" min_val = float("0.179612") @@ -414,6 +452,7 @@ class Program_weight_tensor_parameter_37: class Program_weight_tensor_parameter_38: name = "parameter_38" + original_name = "batch_norm_41.w_2" shape = [1200] dtype = "float32" min_val = float("6.79583e-05") @@ -425,6 +464,7 @@ class Program_weight_tensor_parameter_38: class Program_weight_tensor_parameter_39: name = "parameter_39" + original_name = "batch_norm_41.w_1" shape = [1200] dtype = "float32" min_val = float("-1.08391") @@ -436,6 +476,7 @@ class Program_weight_tensor_parameter_39: class Program_weight_tensor_parameter_40: name = "parameter_40" + original_name = "conv2d_53.w_0" shape = [1200, 1, 5, 5] dtype = "float32" min_val = float("-0.398362") @@ -447,6 +488,7 @@ class Program_weight_tensor_parameter_40: class Program_weight_tensor_parameter_41: name = "parameter_41" + original_name = "batch_norm_40.b_0" shape = [1200] dtype = "float32" min_val = float("-3.63041") @@ -458,6 +500,7 @@ class Program_weight_tensor_parameter_41: class Program_weight_tensor_parameter_42: name = "parameter_42" + original_name = "batch_norm_40.w_0" shape = [1200] dtype = "float32" min_val = float("-0.878308") @@ -469,6 +512,7 @@ class Program_weight_tensor_parameter_42: class Program_weight_tensor_parameter_43: name = "parameter_43" + original_name = "batch_norm_40.w_2" shape = [1200] dtype = "float32" min_val = float("1.00267") @@ -480,6 +524,7 @@ class Program_weight_tensor_parameter_43: class Program_weight_tensor_parameter_44: name = "parameter_44" + original_name = "batch_norm_40.w_1" shape = [1200] dtype = "float32" min_val = float("-3.53517e-06") @@ -491,6 +536,7 @@ class Program_weight_tensor_parameter_44: class Program_weight_tensor_parameter_45: name = "parameter_45" + original_name = "conv2d_52.w_0" shape = [1200, 200, 1, 1] dtype = "float32" min_val = float("-0.279446") @@ -502,6 +548,7 @@ class Program_weight_tensor_parameter_45: class Program_weight_tensor_parameter_46: name = "parameter_46" + original_name = "batch_norm_39.b_0" shape = [200] dtype = "float32" min_val = float("-3.03498e-06") @@ -513,6 +560,7 @@ class Program_weight_tensor_parameter_46: class Program_weight_tensor_parameter_47: name = "parameter_47" + original_name = "batch_norm_39.w_0" shape = [200] dtype = "float32" min_val = float("1.83466") @@ -524,6 +572,7 @@ class Program_weight_tensor_parameter_47: class Program_weight_tensor_parameter_48: name = "parameter_48" + original_name = "batch_norm_39.w_2" shape = [200] dtype = "float32" min_val = float("0.0567872") @@ -535,6 +584,7 @@ class Program_weight_tensor_parameter_48: class Program_weight_tensor_parameter_49: name = "parameter_49" + original_name = "batch_norm_39.w_1" shape = [200] dtype = "float32" min_val = float("-0.586249") @@ -546,6 +596,7 @@ class Program_weight_tensor_parameter_49: class Program_weight_tensor_parameter_50: name = "parameter_50" + original_name = "conv2d_51.w_0" shape = [200, 840, 1, 1] dtype = "float32" min_val = float("-0.413673") @@ -557,6 +608,7 @@ class Program_weight_tensor_parameter_50: class Program_weight_tensor_parameter_51: name = "parameter_51" + original_name = "conv2d_50.b_0" shape = [840] dtype = "float32" min_val = float("-0.169062") @@ -568,6 +620,7 @@ class Program_weight_tensor_parameter_51: class Program_weight_tensor_parameter_52: name = "parameter_52" + original_name = "conv2d_50.w_0" shape = [840, 210, 1, 1] dtype = "float32" min_val = float("-0.559601") @@ -579,6 +632,7 @@ class Program_weight_tensor_parameter_52: class Program_weight_tensor_parameter_53: name = "parameter_53" + original_name = "conv2d_49.b_0" shape = [210] dtype = "float32" min_val = float("-0.020372") @@ -590,6 +644,7 @@ class Program_weight_tensor_parameter_53: class Program_weight_tensor_parameter_54: name = "parameter_54" + original_name = "conv2d_49.w_0" shape = [210, 840, 1, 1] dtype = "float32" min_val = float("-0.212645") @@ -601,6 +656,7 @@ class Program_weight_tensor_parameter_54: class Program_weight_tensor_parameter_55: name = "parameter_55" + original_name = "batch_norm_38.b_0" shape = [840] dtype = "float32" min_val = float("-4.06355") @@ -612,6 +668,7 @@ class Program_weight_tensor_parameter_55: class Program_weight_tensor_parameter_56: name = "parameter_56" + original_name = "batch_norm_38.w_0" shape = [840] dtype = "float32" min_val = float("0.279675") @@ -623,6 +680,7 @@ class Program_weight_tensor_parameter_56: class Program_weight_tensor_parameter_57: name = "parameter_57" + original_name = "batch_norm_38.w_2" shape = [840] dtype = "float32" min_val = float("8.25449e-05") @@ -634,6 +692,7 @@ class Program_weight_tensor_parameter_57: class Program_weight_tensor_parameter_58: name = "parameter_58" + original_name = "batch_norm_38.w_1" shape = [840] dtype = "float32" min_val = float("-8.97924") @@ -645,6 +704,7 @@ class Program_weight_tensor_parameter_58: class Program_weight_tensor_parameter_59: name = "parameter_59" + original_name = "conv2d_48.w_0" shape = [840, 1, 5, 5] dtype = "float32" min_val = float("-0.326246") @@ -656,6 +716,7 @@ class Program_weight_tensor_parameter_59: class Program_weight_tensor_parameter_60: name = "parameter_60" + original_name = "batch_norm_37.b_0" shape = [840] dtype = "float32" min_val = float("-3.71091") @@ -667,6 +728,7 @@ class Program_weight_tensor_parameter_60: class Program_weight_tensor_parameter_61: name = "parameter_61" + original_name = "batch_norm_37.w_0" shape = [840] dtype = "float32" min_val = float("-1.15194") @@ -678,6 +740,7 @@ class Program_weight_tensor_parameter_61: class Program_weight_tensor_parameter_62: name = "parameter_62" + original_name = "batch_norm_37.w_2" shape = [840] dtype = "float32" min_val = float("2.11107") @@ -689,6 +752,7 @@ class Program_weight_tensor_parameter_62: class Program_weight_tensor_parameter_63: name = "parameter_63" + original_name = "batch_norm_37.w_1" shape = [840] dtype = "float32" min_val = float("-2.61852e-06") @@ -700,6 +764,7 @@ class Program_weight_tensor_parameter_63: class Program_weight_tensor_parameter_64: name = "parameter_64" + original_name = "conv2d_47.w_0" shape = [840, 144, 1, 1] dtype = "float32" min_val = float("-0.413557") @@ -711,6 +776,7 @@ class Program_weight_tensor_parameter_64: class Program_weight_tensor_parameter_65: name = "parameter_65" + original_name = "batch_norm_36.b_0" shape = [144] dtype = "float32" min_val = float("-1.77911e-06") @@ -722,6 +788,7 @@ class Program_weight_tensor_parameter_65: class Program_weight_tensor_parameter_66: name = "parameter_66" + original_name = "batch_norm_36.w_0" shape = [144] dtype = "float32" min_val = float("0.756636") @@ -733,6 +800,7 @@ class Program_weight_tensor_parameter_66: class Program_weight_tensor_parameter_67: name = "parameter_67" + original_name = "batch_norm_36.w_2" shape = [144] dtype = "float32" min_val = float("0.00757259") @@ -744,6 +812,7 @@ class Program_weight_tensor_parameter_67: class Program_weight_tensor_parameter_68: name = "parameter_68" + original_name = "batch_norm_36.w_1" shape = [144] dtype = "float32" min_val = float("-0.214596") @@ -755,6 +824,7 @@ class Program_weight_tensor_parameter_68: class Program_weight_tensor_parameter_69: name = "parameter_69" + original_name = "conv2d_46.w_0" shape = [144, 840, 1, 1] dtype = "float32" min_val = float("-0.326582") @@ -766,6 +836,7 @@ class Program_weight_tensor_parameter_69: class Program_weight_tensor_parameter_70: name = "parameter_70" + original_name = "conv2d_45.b_0" shape = [840] dtype = "float32" min_val = float("-0.0964536") @@ -777,6 +848,7 @@ class Program_weight_tensor_parameter_70: class Program_weight_tensor_parameter_71: name = "parameter_71" + original_name = "conv2d_45.w_0" shape = [840, 210, 1, 1] dtype = "float32" min_val = float("-0.278363") @@ -788,6 +860,7 @@ class Program_weight_tensor_parameter_71: class Program_weight_tensor_parameter_72: name = "parameter_72" + original_name = "conv2d_44.b_0" shape = [210] dtype = "float32" min_val = float("-0.00741078") @@ -799,6 +872,7 @@ class Program_weight_tensor_parameter_72: class Program_weight_tensor_parameter_73: name = "parameter_73" + original_name = "conv2d_44.w_0" shape = [210, 840, 1, 1] dtype = "float32" min_val = float("-0.35691") @@ -810,6 +884,7 @@ class Program_weight_tensor_parameter_73: class Program_weight_tensor_parameter_74: name = "parameter_74" + original_name = "batch_norm_35.b_0" shape = [840] dtype = "float32" min_val = float("-4.56636") @@ -821,6 +896,7 @@ class Program_weight_tensor_parameter_74: class Program_weight_tensor_parameter_75: name = "parameter_75" + original_name = "batch_norm_35.w_0" shape = [840] dtype = "float32" min_val = float("0.247327") @@ -832,6 +908,7 @@ class Program_weight_tensor_parameter_75: class Program_weight_tensor_parameter_76: name = "parameter_76" + original_name = "batch_norm_35.w_2" shape = [840] dtype = "float32" min_val = float("1.5411e-05") @@ -843,6 +920,7 @@ class Program_weight_tensor_parameter_76: class Program_weight_tensor_parameter_77: name = "parameter_77" + original_name = "batch_norm_35.w_1" shape = [840] dtype = "float32" min_val = float("-0.931002") @@ -854,6 +932,7 @@ class Program_weight_tensor_parameter_77: class Program_weight_tensor_parameter_78: name = "parameter_78" + original_name = "conv2d_43.w_0" shape = [840, 1, 3, 3] dtype = "float32" min_val = float("-0.39484") @@ -865,6 +944,7 @@ class Program_weight_tensor_parameter_78: class Program_weight_tensor_parameter_79: name = "parameter_79" + original_name = "batch_norm_34.b_0" shape = [840] dtype = "float32" min_val = float("-4.87563") @@ -876,6 +956,7 @@ class Program_weight_tensor_parameter_79: class Program_weight_tensor_parameter_80: name = "parameter_80" + original_name = "batch_norm_34.w_0" shape = [840] dtype = "float32" min_val = float("-0.787364") @@ -887,6 +968,7 @@ class Program_weight_tensor_parameter_80: class Program_weight_tensor_parameter_81: name = "parameter_81" + original_name = "batch_norm_34.w_2" shape = [840] dtype = "float32" min_val = float("1.17297") @@ -898,6 +980,7 @@ class Program_weight_tensor_parameter_81: class Program_weight_tensor_parameter_82: name = "parameter_82" + original_name = "batch_norm_34.w_1" shape = [840] dtype = "float32" min_val = float("-1.15897e-06") @@ -909,6 +992,7 @@ class Program_weight_tensor_parameter_82: class Program_weight_tensor_parameter_83: name = "parameter_83" + original_name = "conv2d_42.w_0" shape = [840, 144, 1, 1] dtype = "float32" min_val = float("-0.303051") @@ -920,6 +1004,7 @@ class Program_weight_tensor_parameter_83: class Program_weight_tensor_parameter_84: name = "parameter_84" + original_name = "batch_norm_33.b_0" shape = [144] dtype = "float32" min_val = float("-1.57352e-06") @@ -931,6 +1016,7 @@ class Program_weight_tensor_parameter_84: class Program_weight_tensor_parameter_85: name = "parameter_85" + original_name = "batch_norm_33.w_0" shape = [144] dtype = "float32" min_val = float("0.872746") @@ -942,6 +1028,7 @@ class Program_weight_tensor_parameter_85: class Program_weight_tensor_parameter_86: name = "parameter_86" + original_name = "batch_norm_33.w_2" shape = [144] dtype = "float32" min_val = float("0.0204485") @@ -953,6 +1040,7 @@ class Program_weight_tensor_parameter_86: class Program_weight_tensor_parameter_87: name = "parameter_87" + original_name = "batch_norm_33.w_1" shape = [144] dtype = "float32" min_val = float("-0.474355") @@ -964,6 +1052,7 @@ class Program_weight_tensor_parameter_87: class Program_weight_tensor_parameter_88: name = "parameter_88" + original_name = "conv2d_41.w_0" shape = [144, 600, 1, 1] dtype = "float32" min_val = float("-0.554162") @@ -975,6 +1064,7 @@ class Program_weight_tensor_parameter_88: class Program_weight_tensor_parameter_89: name = "parameter_89" + original_name = "conv2d_40.b_0" shape = [600] dtype = "float32" min_val = float("-0.141447") @@ -986,6 +1076,7 @@ class Program_weight_tensor_parameter_89: class Program_weight_tensor_parameter_90: name = "parameter_90" + original_name = "conv2d_40.w_0" shape = [600, 150, 1, 1] dtype = "float32" min_val = float("-0.476892") @@ -997,6 +1088,7 @@ class Program_weight_tensor_parameter_90: class Program_weight_tensor_parameter_91: name = "parameter_91" + original_name = "conv2d_39.b_0" shape = [150] dtype = "float32" min_val = float("-0.0376507") @@ -1008,6 +1100,7 @@ class Program_weight_tensor_parameter_91: class Program_weight_tensor_parameter_92: name = "parameter_92" + original_name = "conv2d_39.w_0" shape = [150, 600, 1, 1] dtype = "float32" min_val = float("-0.659015") @@ -1019,6 +1112,7 @@ class Program_weight_tensor_parameter_92: class Program_weight_tensor_parameter_93: name = "parameter_93" + original_name = "batch_norm_32.b_0" shape = [600] dtype = "float32" min_val = float("-3.06399") @@ -1030,6 +1124,7 @@ class Program_weight_tensor_parameter_93: class Program_weight_tensor_parameter_94: name = "parameter_94" + original_name = "batch_norm_32.w_0" shape = [600] dtype = "float32" min_val = float("0.262209") @@ -1041,6 +1136,7 @@ class Program_weight_tensor_parameter_94: class Program_weight_tensor_parameter_95: name = "parameter_95" + original_name = "batch_norm_32.w_2" shape = [600] dtype = "float32" min_val = float("2.38655e-05") @@ -1052,6 +1148,7 @@ class Program_weight_tensor_parameter_95: class Program_weight_tensor_parameter_96: name = "parameter_96" + original_name = "batch_norm_32.w_1" shape = [600] dtype = "float32" min_val = float("-1.42803") @@ -1063,6 +1160,7 @@ class Program_weight_tensor_parameter_96: class Program_weight_tensor_parameter_97: name = "parameter_97" + original_name = "conv2d_38.w_0" shape = [600, 1, 3, 3] dtype = "float32" min_val = float("-0.601882") @@ -1074,6 +1172,7 @@ class Program_weight_tensor_parameter_97: class Program_weight_tensor_parameter_98: name = "parameter_98" + original_name = "batch_norm_31.b_0" shape = [600] dtype = "float32" min_val = float("-3.32715") @@ -1085,6 +1184,7 @@ class Program_weight_tensor_parameter_98: class Program_weight_tensor_parameter_99: name = "parameter_99" + original_name = "batch_norm_31.w_0" shape = [600] dtype = "float32" min_val = float("-1.19622") @@ -1096,6 +1196,7 @@ class Program_weight_tensor_parameter_99: class Program_weight_tensor_parameter_100: name = "parameter_100" + original_name = "batch_norm_31.w_2" shape = [600] dtype = "float32" min_val = float("4.19978") @@ -1107,6 +1208,7 @@ class Program_weight_tensor_parameter_100: class Program_weight_tensor_parameter_101: name = "parameter_101" + original_name = "batch_norm_31.w_1" shape = [600] dtype = "float32" min_val = float("-4.9151e-06") @@ -1118,6 +1220,7 @@ class Program_weight_tensor_parameter_101: class Program_weight_tensor_parameter_102: name = "parameter_102" + original_name = "conv2d_37.w_0" shape = [600, 104, 1, 1] dtype = "float32" min_val = float("-0.613991") @@ -1129,6 +1232,7 @@ class Program_weight_tensor_parameter_102: class Program_weight_tensor_parameter_103: name = "parameter_103" + original_name = "batch_norm_30.b_0" shape = [104] dtype = "float32" min_val = float("-3.3619e-06") @@ -1140,6 +1244,7 @@ class Program_weight_tensor_parameter_103: class Program_weight_tensor_parameter_104: name = "parameter_104" + original_name = "batch_norm_30.w_0" shape = [104] dtype = "float32" min_val = float("0.187303") @@ -1151,6 +1256,7 @@ class Program_weight_tensor_parameter_104: class Program_weight_tensor_parameter_105: name = "parameter_105" + original_name = "batch_norm_30.w_2" shape = [104] dtype = "float32" min_val = float("0.0253125") @@ -1162,6 +1268,7 @@ class Program_weight_tensor_parameter_105: class Program_weight_tensor_parameter_106: name = "parameter_106" + original_name = "batch_norm_30.w_1" shape = [104] dtype = "float32" min_val = float("-1.45084") @@ -1173,6 +1280,7 @@ class Program_weight_tensor_parameter_106: class Program_weight_tensor_parameter_107: name = "parameter_107" + original_name = "conv2d_36.w_0" shape = [104, 232, 1, 1] dtype = "float32" min_val = float("-0.300331") @@ -1184,6 +1292,7 @@ class Program_weight_tensor_parameter_107: class Program_weight_tensor_parameter_108: name = "parameter_108" + original_name = "batch_norm_29.b_0" shape = [232] dtype = "float32" min_val = float("-3.69618") @@ -1195,6 +1304,7 @@ class Program_weight_tensor_parameter_108: class Program_weight_tensor_parameter_109: name = "parameter_109" + original_name = "batch_norm_29.w_0" shape = [232] dtype = "float32" min_val = float("0.289353") @@ -1206,6 +1316,7 @@ class Program_weight_tensor_parameter_109: class Program_weight_tensor_parameter_110: name = "parameter_110" + original_name = "batch_norm_29.w_2" shape = [232] dtype = "float32" min_val = float("7.77651e-05") @@ -1217,6 +1328,7 @@ class Program_weight_tensor_parameter_110: class Program_weight_tensor_parameter_111: name = "parameter_111" + original_name = "batch_norm_29.w_1" shape = [232] dtype = "float32" min_val = float("-1.09885") @@ -1228,6 +1340,7 @@ class Program_weight_tensor_parameter_111: class Program_weight_tensor_parameter_112: name = "parameter_112" + original_name = "conv2d_35.w_0" shape = [232, 1, 3, 3] dtype = "float32" min_val = float("-0.344719") @@ -1239,6 +1352,7 @@ class Program_weight_tensor_parameter_112: class Program_weight_tensor_parameter_113: name = "parameter_113" + original_name = "batch_norm_28.b_0" shape = [232] dtype = "float32" min_val = float("-2.80875") @@ -1250,6 +1364,7 @@ class Program_weight_tensor_parameter_113: class Program_weight_tensor_parameter_114: name = "parameter_114" + original_name = "batch_norm_28.w_0" shape = [232] dtype = "float32" min_val = float("-0.388781") @@ -1261,6 +1376,7 @@ class Program_weight_tensor_parameter_114: class Program_weight_tensor_parameter_115: name = "parameter_115" + original_name = "batch_norm_28.w_2" shape = [232] dtype = "float32" min_val = float("2.52689") @@ -1272,6 +1388,7 @@ class Program_weight_tensor_parameter_115: class Program_weight_tensor_parameter_116: name = "parameter_116" + original_name = "batch_norm_28.w_1" shape = [232] dtype = "float32" min_val = float("-2.63888e-06") @@ -1283,6 +1400,7 @@ class Program_weight_tensor_parameter_116: class Program_weight_tensor_parameter_117: name = "parameter_117" + original_name = "conv2d_34.w_0" shape = [232, 104, 1, 1] dtype = "float32" min_val = float("-0.518282") @@ -1294,6 +1412,7 @@ class Program_weight_tensor_parameter_117: class Program_weight_tensor_parameter_118: name = "parameter_118" + original_name = "batch_norm_27.b_0" shape = [104] dtype = "float32" min_val = float("-1.98292e-06") @@ -1305,6 +1424,7 @@ class Program_weight_tensor_parameter_118: class Program_weight_tensor_parameter_119: name = "parameter_119" + original_name = "batch_norm_27.w_0" shape = [104] dtype = "float32" min_val = float("0.241394") @@ -1316,6 +1436,7 @@ class Program_weight_tensor_parameter_119: class Program_weight_tensor_parameter_120: name = "parameter_120" + original_name = "batch_norm_27.w_2" shape = [104] dtype = "float32" min_val = float("0.00447902") @@ -1327,6 +1448,7 @@ class Program_weight_tensor_parameter_120: class Program_weight_tensor_parameter_121: name = "parameter_121" + original_name = "batch_norm_27.w_1" shape = [104] dtype = "float32" min_val = float("-2.20325") @@ -1338,6 +1460,7 @@ class Program_weight_tensor_parameter_121: class Program_weight_tensor_parameter_122: name = "parameter_122" + original_name = "conv2d_33.w_0" shape = [104, 232, 1, 1] dtype = "float32" min_val = float("-0.417899") @@ -1349,6 +1472,7 @@ class Program_weight_tensor_parameter_122: class Program_weight_tensor_parameter_123: name = "parameter_123" + original_name = "batch_norm_26.b_0" shape = [232] dtype = "float32" min_val = float("-2.5632") @@ -1360,6 +1484,7 @@ class Program_weight_tensor_parameter_123: class Program_weight_tensor_parameter_124: name = "parameter_124" + original_name = "batch_norm_26.w_0" shape = [232] dtype = "float32" min_val = float("0.149454") @@ -1371,6 +1496,7 @@ class Program_weight_tensor_parameter_124: class Program_weight_tensor_parameter_125: name = "parameter_125" + original_name = "batch_norm_26.w_2" shape = [232] dtype = "float32" min_val = float("4.52899e-05") @@ -1382,6 +1508,7 @@ class Program_weight_tensor_parameter_125: class Program_weight_tensor_parameter_126: name = "parameter_126" + original_name = "batch_norm_26.w_1" shape = [232] dtype = "float32" min_val = float("-1.45975") @@ -1393,6 +1520,7 @@ class Program_weight_tensor_parameter_126: class Program_weight_tensor_parameter_127: name = "parameter_127" + original_name = "conv2d_32.w_0" shape = [232, 1, 3, 3] dtype = "float32" min_val = float("-0.4791") @@ -1404,6 +1532,7 @@ class Program_weight_tensor_parameter_127: class Program_weight_tensor_parameter_128: name = "parameter_128" + original_name = "batch_norm_25.b_0" shape = [232] dtype = "float32" min_val = float("-2.7871") @@ -1415,6 +1544,7 @@ class Program_weight_tensor_parameter_128: class Program_weight_tensor_parameter_129: name = "parameter_129" + original_name = "batch_norm_25.w_0" shape = [232] dtype = "float32" min_val = float("-0.194254") @@ -1426,6 +1556,7 @@ class Program_weight_tensor_parameter_129: class Program_weight_tensor_parameter_130: name = "parameter_130" + original_name = "batch_norm_25.w_2" shape = [232] dtype = "float32" min_val = float("2.10394") @@ -1437,6 +1568,7 @@ class Program_weight_tensor_parameter_130: class Program_weight_tensor_parameter_131: name = "parameter_131" + original_name = "batch_norm_25.w_1" shape = [232] dtype = "float32" min_val = float("-2.80471e-06") @@ -1448,6 +1580,7 @@ class Program_weight_tensor_parameter_131: class Program_weight_tensor_parameter_132: name = "parameter_132" + original_name = "conv2d_31.w_0" shape = [232, 104, 1, 1] dtype = "float32" min_val = float("-0.514306") @@ -1459,6 +1592,7 @@ class Program_weight_tensor_parameter_132: class Program_weight_tensor_parameter_133: name = "parameter_133" + original_name = "batch_norm_24.b_0" shape = [104] dtype = "float32" min_val = float("-1.63269e-06") @@ -1470,6 +1604,7 @@ class Program_weight_tensor_parameter_133: class Program_weight_tensor_parameter_134: name = "parameter_134" + original_name = "batch_norm_24.w_0" shape = [104] dtype = "float32" min_val = float("0.272789") @@ -1481,6 +1616,7 @@ class Program_weight_tensor_parameter_134: class Program_weight_tensor_parameter_135: name = "parameter_135" + original_name = "batch_norm_24.w_2" shape = [104] dtype = "float32" min_val = float("0.00259953") @@ -1492,6 +1628,7 @@ class Program_weight_tensor_parameter_135: class Program_weight_tensor_parameter_136: name = "parameter_136" + original_name = "batch_norm_24.w_1" shape = [104] dtype = "float32" min_val = float("-1.25404") @@ -1503,6 +1640,7 @@ class Program_weight_tensor_parameter_136: class Program_weight_tensor_parameter_137: name = "parameter_137" + original_name = "conv2d_30.w_0" shape = [104, 248, 1, 1] dtype = "float32" min_val = float("-0.356707") @@ -1514,6 +1652,7 @@ class Program_weight_tensor_parameter_137: class Program_weight_tensor_parameter_138: name = "parameter_138" + original_name = "batch_norm_23.b_0" shape = [248] dtype = "float32" min_val = float("-3.18014") @@ -1525,6 +1664,7 @@ class Program_weight_tensor_parameter_138: class Program_weight_tensor_parameter_139: name = "parameter_139" + original_name = "batch_norm_23.w_0" shape = [248] dtype = "float32" min_val = float("0.123325") @@ -1536,6 +1676,7 @@ class Program_weight_tensor_parameter_139: class Program_weight_tensor_parameter_140: name = "parameter_140" + original_name = "batch_norm_23.w_2" shape = [248] dtype = "float32" min_val = float("6.44418e-05") @@ -1547,6 +1688,7 @@ class Program_weight_tensor_parameter_140: class Program_weight_tensor_parameter_141: name = "parameter_141" + original_name = "batch_norm_23.w_1" shape = [248] dtype = "float32" min_val = float("-3.33253") @@ -1558,6 +1700,7 @@ class Program_weight_tensor_parameter_141: class Program_weight_tensor_parameter_142: name = "parameter_142" + original_name = "conv2d_29.w_0" shape = [248, 1, 3, 3] dtype = "float32" min_val = float("-0.437799") @@ -1569,6 +1712,7 @@ class Program_weight_tensor_parameter_142: class Program_weight_tensor_parameter_143: name = "parameter_143" + original_name = "batch_norm_22.b_0" shape = [248] dtype = "float32" min_val = float("-2.70179") @@ -1580,6 +1724,7 @@ class Program_weight_tensor_parameter_143: class Program_weight_tensor_parameter_144: name = "parameter_144" + original_name = "batch_norm_22.w_0" shape = [248] dtype = "float32" min_val = float("-0.166877") @@ -1591,6 +1736,7 @@ class Program_weight_tensor_parameter_144: class Program_weight_tensor_parameter_145: name = "parameter_145" + original_name = "batch_norm_22.w_2" shape = [248] dtype = "float32" min_val = float("1.32327") @@ -1602,6 +1748,7 @@ class Program_weight_tensor_parameter_145: class Program_weight_tensor_parameter_146: name = "parameter_146" + original_name = "batch_norm_22.w_1" shape = [248] dtype = "float32" min_val = float("-1.02555e-06") @@ -1613,6 +1760,7 @@ class Program_weight_tensor_parameter_146: class Program_weight_tensor_parameter_147: name = "parameter_147" + original_name = "conv2d_28.w_0" shape = [248, 104, 1, 1] dtype = "float32" min_val = float("-0.412158") @@ -1624,6 +1772,7 @@ class Program_weight_tensor_parameter_147: class Program_weight_tensor_parameter_148: name = "parameter_148" + original_name = "batch_norm_21.b_0" shape = [104] dtype = "float32" min_val = float("-1.81845e-06") @@ -1635,6 +1784,7 @@ class Program_weight_tensor_parameter_148: class Program_weight_tensor_parameter_149: name = "parameter_149" + original_name = "batch_norm_21.w_0" shape = [104] dtype = "float32" min_val = float("0.391092") @@ -1646,6 +1796,7 @@ class Program_weight_tensor_parameter_149: class Program_weight_tensor_parameter_150: name = "parameter_150" + original_name = "batch_norm_21.w_2" shape = [104] dtype = "float32" min_val = float("0.106507") @@ -1657,6 +1808,7 @@ class Program_weight_tensor_parameter_150: class Program_weight_tensor_parameter_151: name = "parameter_151" + original_name = "batch_norm_21.w_1" shape = [104] dtype = "float32" min_val = float("-6.17622") @@ -1668,6 +1820,7 @@ class Program_weight_tensor_parameter_151: class Program_weight_tensor_parameter_152: name = "parameter_152" + original_name = "conv2d_27.w_0" shape = [104, 304, 1, 1] dtype = "float32" min_val = float("-0.692236") @@ -1679,6 +1832,7 @@ class Program_weight_tensor_parameter_152: class Program_weight_tensor_parameter_153: name = "parameter_153" + original_name = "batch_norm_20.b_0" shape = [304] dtype = "float32" min_val = float("-2.6818") @@ -1690,6 +1844,7 @@ class Program_weight_tensor_parameter_153: class Program_weight_tensor_parameter_154: name = "parameter_154" + original_name = "batch_norm_20.w_0" shape = [304] dtype = "float32" min_val = float("0.514729") @@ -1701,6 +1856,7 @@ class Program_weight_tensor_parameter_154: class Program_weight_tensor_parameter_155: name = "parameter_155" + original_name = "batch_norm_20.w_2" shape = [304] dtype = "float32" min_val = float("8.93677e-05") @@ -1712,6 +1868,7 @@ class Program_weight_tensor_parameter_155: class Program_weight_tensor_parameter_156: name = "parameter_156" + original_name = "batch_norm_20.w_1" shape = [304] dtype = "float32" min_val = float("-1.82051") @@ -1723,6 +1880,7 @@ class Program_weight_tensor_parameter_156: class Program_weight_tensor_parameter_157: name = "parameter_157" + original_name = "conv2d_26.w_0" shape = [304, 1, 3, 3] dtype = "float32" min_val = float("-0.259323") @@ -1734,6 +1892,7 @@ class Program_weight_tensor_parameter_157: class Program_weight_tensor_parameter_158: name = "parameter_158" + original_name = "batch_norm_19.b_0" shape = [304] dtype = "float32" min_val = float("-3.13881") @@ -1745,6 +1904,7 @@ class Program_weight_tensor_parameter_158: class Program_weight_tensor_parameter_159: name = "parameter_159" + original_name = "batch_norm_19.w_0" shape = [304] dtype = "float32" min_val = float("-0.0978672") @@ -1756,6 +1916,7 @@ class Program_weight_tensor_parameter_159: class Program_weight_tensor_parameter_160: name = "parameter_160" + original_name = "batch_norm_19.w_2" shape = [304] dtype = "float32" min_val = float("5.05677") @@ -1767,6 +1928,7 @@ class Program_weight_tensor_parameter_160: class Program_weight_tensor_parameter_161: name = "parameter_161" + original_name = "batch_norm_19.w_1" shape = [304] dtype = "float32" min_val = float("-4.8677e-06") @@ -1778,6 +1940,7 @@ class Program_weight_tensor_parameter_161: class Program_weight_tensor_parameter_162: name = "parameter_162" + original_name = "conv2d_25.w_0" shape = [304, 48, 1, 1] dtype = "float32" min_val = float("-0.748927") @@ -1789,6 +1952,7 @@ class Program_weight_tensor_parameter_162: class Program_weight_tensor_parameter_163: name = "parameter_163" + original_name = "batch_norm_18.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -1798,6 +1962,7 @@ class Program_weight_tensor_parameter_163: class Program_weight_tensor_parameter_164: name = "parameter_164" + original_name = "batch_norm_18.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -1807,6 +1972,7 @@ class Program_weight_tensor_parameter_164: class Program_weight_tensor_parameter_165: name = "parameter_165" + original_name = "batch_norm_18.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -1816,6 +1982,7 @@ class Program_weight_tensor_parameter_165: class Program_weight_tensor_parameter_166: name = "parameter_166" + original_name = "batch_norm_18.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -1825,6 +1992,7 @@ class Program_weight_tensor_parameter_166: class Program_weight_tensor_parameter_167: name = "parameter_167" + original_name = "conv2d_24.w_0" shape = [48, 152, 1, 1] dtype = "float32" min_val = float("-0.689141") @@ -1836,6 +2004,7 @@ class Program_weight_tensor_parameter_167: class Program_weight_tensor_parameter_168: name = "parameter_168" + original_name = "conv2d_23.b_0" shape = [152] dtype = "float32" min_val = float("-0.114034") @@ -1847,6 +2016,7 @@ class Program_weight_tensor_parameter_168: class Program_weight_tensor_parameter_169: name = "parameter_169" + original_name = "conv2d_23.w_0" shape = [152, 38, 1, 1] dtype = "float32" min_val = float("-0.398375") @@ -1858,6 +2028,7 @@ class Program_weight_tensor_parameter_169: class Program_weight_tensor_parameter_170: name = "parameter_170" + original_name = "conv2d_22.b_0" shape = [38] dtype = "float32" min_val = float("0") @@ -1867,6 +2038,7 @@ class Program_weight_tensor_parameter_170: class Program_weight_tensor_parameter_171: name = "parameter_171" + original_name = "conv2d_22.w_0" shape = [38, 152, 1, 1] dtype = "float32" min_val = float("-0.584985") @@ -1878,6 +2050,7 @@ class Program_weight_tensor_parameter_171: class Program_weight_tensor_parameter_172: name = "parameter_172" + original_name = "batch_norm_17.b_0" shape = [152] dtype = "float32" min_val = float("-3.53141") @@ -1889,6 +2062,7 @@ class Program_weight_tensor_parameter_172: class Program_weight_tensor_parameter_173: name = "parameter_173" + original_name = "batch_norm_17.w_0" shape = [152] dtype = "float32" min_val = float("0.403764") @@ -1900,6 +2074,7 @@ class Program_weight_tensor_parameter_173: class Program_weight_tensor_parameter_174: name = "parameter_174" + original_name = "batch_norm_17.w_2" shape = [152] dtype = "float32" min_val = float("0.00203221") @@ -1911,6 +2086,7 @@ class Program_weight_tensor_parameter_174: class Program_weight_tensor_parameter_175: name = "parameter_175" + original_name = "batch_norm_17.w_1" shape = [152] dtype = "float32" min_val = float("-1.75633") @@ -1922,6 +2098,7 @@ class Program_weight_tensor_parameter_175: class Program_weight_tensor_parameter_176: name = "parameter_176" + original_name = "conv2d_21.w_0" shape = [152, 1, 5, 5] dtype = "float32" min_val = float("-0.514027") @@ -1933,6 +2110,7 @@ class Program_weight_tensor_parameter_176: class Program_weight_tensor_parameter_177: name = "parameter_177" + original_name = "batch_norm_16.b_0" shape = [152] dtype = "float32" min_val = float("-2.21937") @@ -1944,6 +2122,7 @@ class Program_weight_tensor_parameter_177: class Program_weight_tensor_parameter_178: name = "parameter_178" + original_name = "batch_norm_16.w_0" shape = [152] dtype = "float32" min_val = float("0.535888") @@ -1955,6 +2134,7 @@ class Program_weight_tensor_parameter_178: class Program_weight_tensor_parameter_179: name = "parameter_179" + original_name = "batch_norm_16.w_2" shape = [152] dtype = "float32" min_val = float("4.60597") @@ -1966,6 +2146,7 @@ class Program_weight_tensor_parameter_179: class Program_weight_tensor_parameter_180: name = "parameter_180" + original_name = "batch_norm_16.w_1" shape = [152] dtype = "float32" min_val = float("-2.45762e-06") @@ -1977,6 +2158,7 @@ class Program_weight_tensor_parameter_180: class Program_weight_tensor_parameter_181: name = "parameter_181" + original_name = "conv2d_20.w_0" shape = [152, 48, 1, 1] dtype = "float32" min_val = float("-0.598598") @@ -1988,6 +2170,7 @@ class Program_weight_tensor_parameter_181: class Program_weight_tensor_parameter_182: name = "parameter_182" + original_name = "batch_norm_15.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -1997,6 +2180,7 @@ class Program_weight_tensor_parameter_182: class Program_weight_tensor_parameter_183: name = "parameter_183" + original_name = "batch_norm_15.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -2006,6 +2190,7 @@ class Program_weight_tensor_parameter_183: class Program_weight_tensor_parameter_184: name = "parameter_184" + original_name = "batch_norm_15.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -2015,6 +2200,7 @@ class Program_weight_tensor_parameter_184: class Program_weight_tensor_parameter_185: name = "parameter_185" + original_name = "batch_norm_15.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -2024,6 +2210,7 @@ class Program_weight_tensor_parameter_185: class Program_weight_tensor_parameter_186: name = "parameter_186" + original_name = "conv2d_19.w_0" shape = [48, 152, 1, 1] dtype = "float32" min_val = float("-0.737259") @@ -2035,6 +2222,7 @@ class Program_weight_tensor_parameter_186: class Program_weight_tensor_parameter_187: name = "parameter_187" + original_name = "conv2d_18.b_0" shape = [152] dtype = "float32" min_val = float("-0.111732") @@ -2046,6 +2234,7 @@ class Program_weight_tensor_parameter_187: class Program_weight_tensor_parameter_188: name = "parameter_188" + original_name = "conv2d_18.w_0" shape = [152, 38, 1, 1] dtype = "float32" min_val = float("-0.561793") @@ -2057,6 +2246,7 @@ class Program_weight_tensor_parameter_188: class Program_weight_tensor_parameter_189: name = "parameter_189" + original_name = "conv2d_17.b_0" shape = [38] dtype = "float32" min_val = float("0") @@ -2066,6 +2256,7 @@ class Program_weight_tensor_parameter_189: class Program_weight_tensor_parameter_190: name = "parameter_190" + original_name = "conv2d_17.w_0" shape = [38, 152, 1, 1] dtype = "float32" min_val = float("-0.704741") @@ -2077,6 +2268,7 @@ class Program_weight_tensor_parameter_190: class Program_weight_tensor_parameter_191: name = "parameter_191" + original_name = "batch_norm_14.b_0" shape = [152] dtype = "float32" min_val = float("-3.12786") @@ -2088,6 +2280,7 @@ class Program_weight_tensor_parameter_191: class Program_weight_tensor_parameter_192: name = "parameter_192" + original_name = "batch_norm_14.w_0" shape = [152] dtype = "float32" min_val = float("0.458884") @@ -2099,6 +2292,7 @@ class Program_weight_tensor_parameter_192: class Program_weight_tensor_parameter_193: name = "parameter_193" + original_name = "batch_norm_14.w_2" shape = [152] dtype = "float32" min_val = float("0.00591609") @@ -2110,6 +2304,7 @@ class Program_weight_tensor_parameter_193: class Program_weight_tensor_parameter_194: name = "parameter_194" + original_name = "batch_norm_14.w_1" shape = [152] dtype = "float32" min_val = float("-7.96077") @@ -2121,6 +2316,7 @@ class Program_weight_tensor_parameter_194: class Program_weight_tensor_parameter_195: name = "parameter_195" + original_name = "conv2d_16.w_0" shape = [152, 1, 5, 5] dtype = "float32" min_val = float("-0.617416") @@ -2132,6 +2328,7 @@ class Program_weight_tensor_parameter_195: class Program_weight_tensor_parameter_196: name = "parameter_196" + original_name = "batch_norm_13.b_0" shape = [152] dtype = "float32" min_val = float("-2.75796") @@ -2143,6 +2340,7 @@ class Program_weight_tensor_parameter_196: class Program_weight_tensor_parameter_197: name = "parameter_197" + original_name = "batch_norm_13.w_0" shape = [152] dtype = "float32" min_val = float("0.222253") @@ -2154,6 +2352,7 @@ class Program_weight_tensor_parameter_197: class Program_weight_tensor_parameter_198: name = "parameter_198" + original_name = "batch_norm_13.w_2" shape = [152] dtype = "float32" min_val = float("2.87489") @@ -2165,6 +2364,7 @@ class Program_weight_tensor_parameter_198: class Program_weight_tensor_parameter_199: name = "parameter_199" + original_name = "batch_norm_13.w_1" shape = [152] dtype = "float32" min_val = float("-1.98527e-06") @@ -2176,6 +2376,7 @@ class Program_weight_tensor_parameter_199: class Program_weight_tensor_parameter_200: name = "parameter_200" + original_name = "conv2d_15.w_0" shape = [152, 48, 1, 1] dtype = "float32" min_val = float("-0.527137") @@ -2187,6 +2388,7 @@ class Program_weight_tensor_parameter_200: class Program_weight_tensor_parameter_201: name = "parameter_201" + original_name = "batch_norm_12.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -2196,6 +2398,7 @@ class Program_weight_tensor_parameter_201: class Program_weight_tensor_parameter_202: name = "parameter_202" + original_name = "batch_norm_12.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -2205,6 +2408,7 @@ class Program_weight_tensor_parameter_202: class Program_weight_tensor_parameter_203: name = "parameter_203" + original_name = "batch_norm_12.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -2214,6 +2418,7 @@ class Program_weight_tensor_parameter_203: class Program_weight_tensor_parameter_204: name = "parameter_204" + original_name = "batch_norm_12.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -2223,6 +2428,7 @@ class Program_weight_tensor_parameter_204: class Program_weight_tensor_parameter_205: name = "parameter_205" + original_name = "conv2d_14.w_0" shape = [48, 88, 1, 1] dtype = "float32" min_val = float("-1.52493") @@ -2234,6 +2440,7 @@ class Program_weight_tensor_parameter_205: class Program_weight_tensor_parameter_206: name = "parameter_206" + original_name = "conv2d_13.b_0" shape = [88] dtype = "float32" min_val = float("-0.0106954") @@ -2245,6 +2452,7 @@ class Program_weight_tensor_parameter_206: class Program_weight_tensor_parameter_207: name = "parameter_207" + original_name = "conv2d_13.w_0" shape = [88, 22, 1, 1] dtype = "float32" min_val = float("-0.478849") @@ -2256,6 +2464,7 @@ class Program_weight_tensor_parameter_207: class Program_weight_tensor_parameter_208: name = "parameter_208" + original_name = "conv2d_12.b_0" shape = [22] dtype = "float32" min_val = float("0") @@ -2265,6 +2474,7 @@ class Program_weight_tensor_parameter_208: class Program_weight_tensor_parameter_209: name = "parameter_209" + original_name = "conv2d_12.w_0" shape = [22, 88, 1, 1] dtype = "float32" min_val = float("-0.402646") @@ -2276,6 +2486,7 @@ class Program_weight_tensor_parameter_209: class Program_weight_tensor_parameter_210: name = "parameter_210" + original_name = "batch_norm_11.b_0" shape = [88] dtype = "float32" min_val = float("-1.33649") @@ -2287,6 +2498,7 @@ class Program_weight_tensor_parameter_210: class Program_weight_tensor_parameter_211: name = "parameter_211" + original_name = "batch_norm_11.w_0" shape = [88] dtype = "float32" min_val = float("0.931654") @@ -2298,6 +2510,7 @@ class Program_weight_tensor_parameter_211: class Program_weight_tensor_parameter_212: name = "parameter_212" + original_name = "batch_norm_11.w_2" shape = [88] dtype = "float32" min_val = float("0.119521") @@ -2309,6 +2522,7 @@ class Program_weight_tensor_parameter_212: class Program_weight_tensor_parameter_213: name = "parameter_213" + original_name = "batch_norm_11.w_1" shape = [88] dtype = "float32" min_val = float("-4.85419") @@ -2320,6 +2534,7 @@ class Program_weight_tensor_parameter_213: class Program_weight_tensor_parameter_214: name = "parameter_214" + original_name = "conv2d_11.w_0" shape = [88, 1, 5, 5] dtype = "float32" min_val = float("-0.566227") @@ -2331,6 +2546,7 @@ class Program_weight_tensor_parameter_214: class Program_weight_tensor_parameter_215: name = "parameter_215" + original_name = "batch_norm_10.b_0" shape = [88] dtype = "float32" min_val = float("-3.04724") @@ -2342,6 +2558,7 @@ class Program_weight_tensor_parameter_215: class Program_weight_tensor_parameter_216: name = "parameter_216" + original_name = "batch_norm_10.w_0" shape = [88] dtype = "float32" min_val = float("0.924231") @@ -2353,6 +2570,7 @@ class Program_weight_tensor_parameter_216: class Program_weight_tensor_parameter_217: name = "parameter_217" + original_name = "batch_norm_10.w_2" shape = [88] dtype = "float32" min_val = float("7.4731") @@ -2364,6 +2582,7 @@ class Program_weight_tensor_parameter_217: class Program_weight_tensor_parameter_218: name = "parameter_218" + original_name = "batch_norm_10.w_1" shape = [88] dtype = "float32" min_val = float("-1.30035e-05") @@ -2375,6 +2594,7 @@ class Program_weight_tensor_parameter_218: class Program_weight_tensor_parameter_219: name = "parameter_219" + original_name = "conv2d_10.w_0" shape = [88, 32, 1, 1] dtype = "float32" min_val = float("-1.10964") @@ -2386,6 +2606,7 @@ class Program_weight_tensor_parameter_219: class Program_weight_tensor_parameter_220: name = "parameter_220" + original_name = "batch_norm_9.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -2395,6 +2616,7 @@ class Program_weight_tensor_parameter_220: class Program_weight_tensor_parameter_221: name = "parameter_221" + original_name = "batch_norm_9.w_0" shape = [32] dtype = "float32" min_val = float("0") @@ -2404,6 +2626,7 @@ class Program_weight_tensor_parameter_221: class Program_weight_tensor_parameter_222: name = "parameter_222" + original_name = "batch_norm_9.w_2" shape = [32] dtype = "float32" min_val = float("0") @@ -2413,6 +2636,7 @@ class Program_weight_tensor_parameter_222: class Program_weight_tensor_parameter_223: name = "parameter_223" + original_name = "batch_norm_9.w_1" shape = [32] dtype = "float32" min_val = float("0") @@ -2422,6 +2646,7 @@ class Program_weight_tensor_parameter_223: class Program_weight_tensor_parameter_224: name = "parameter_224" + original_name = "conv2d_9.w_0" shape = [32, 88, 1, 1] dtype = "float32" min_val = float("-0.818492") @@ -2433,6 +2658,7 @@ class Program_weight_tensor_parameter_224: class Program_weight_tensor_parameter_225: name = "parameter_225" + original_name = "batch_norm_8.b_0" shape = [88] dtype = "float32" min_val = float("-3.16676") @@ -2444,6 +2670,7 @@ class Program_weight_tensor_parameter_225: class Program_weight_tensor_parameter_226: name = "parameter_226" + original_name = "batch_norm_8.w_0" shape = [88] dtype = "float32" min_val = float("0.875496") @@ -2455,6 +2682,7 @@ class Program_weight_tensor_parameter_226: class Program_weight_tensor_parameter_227: name = "parameter_227" + original_name = "batch_norm_8.w_2" shape = [88] dtype = "float32" min_val = float("0.00267744") @@ -2466,6 +2694,7 @@ class Program_weight_tensor_parameter_227: class Program_weight_tensor_parameter_228: name = "parameter_228" + original_name = "batch_norm_8.w_1" shape = [88] dtype = "float32" min_val = float("-8.65966") @@ -2477,6 +2706,7 @@ class Program_weight_tensor_parameter_228: class Program_weight_tensor_parameter_229: name = "parameter_229" + original_name = "conv2d_8.w_0" shape = [88, 1, 3, 3] dtype = "float32" min_val = float("-0.815429") @@ -2488,6 +2718,7 @@ class Program_weight_tensor_parameter_229: class Program_weight_tensor_parameter_230: name = "parameter_230" + original_name = "batch_norm_7.b_0" shape = [88] dtype = "float32" min_val = float("-2.01146") @@ -2499,6 +2730,7 @@ class Program_weight_tensor_parameter_230: class Program_weight_tensor_parameter_231: name = "parameter_231" + original_name = "batch_norm_7.w_0" shape = [88] dtype = "float32" min_val = float("0.402047") @@ -2510,6 +2742,7 @@ class Program_weight_tensor_parameter_231: class Program_weight_tensor_parameter_232: name = "parameter_232" + original_name = "batch_norm_7.w_2" shape = [88] dtype = "float32" min_val = float("2.36857") @@ -2521,6 +2754,7 @@ class Program_weight_tensor_parameter_232: class Program_weight_tensor_parameter_233: name = "parameter_233" + original_name = "batch_norm_7.w_1" shape = [88] dtype = "float32" min_val = float("-6.14738e-06") @@ -2532,6 +2766,7 @@ class Program_weight_tensor_parameter_233: class Program_weight_tensor_parameter_234: name = "parameter_234" + original_name = "conv2d_7.w_0" shape = [88, 32, 1, 1] dtype = "float32" min_val = float("-0.587429") @@ -2543,6 +2778,7 @@ class Program_weight_tensor_parameter_234: class Program_weight_tensor_parameter_235: name = "parameter_235" + original_name = "batch_norm_6.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -2552,6 +2788,7 @@ class Program_weight_tensor_parameter_235: class Program_weight_tensor_parameter_236: name = "parameter_236" + original_name = "batch_norm_6.w_0" shape = [32] dtype = "float32" min_val = float("0") @@ -2561,6 +2798,7 @@ class Program_weight_tensor_parameter_236: class Program_weight_tensor_parameter_237: name = "parameter_237" + original_name = "batch_norm_6.w_2" shape = [32] dtype = "float32" min_val = float("0") @@ -2570,6 +2808,7 @@ class Program_weight_tensor_parameter_237: class Program_weight_tensor_parameter_238: name = "parameter_238" + original_name = "batch_norm_6.w_1" shape = [32] dtype = "float32" min_val = float("0") @@ -2579,6 +2818,7 @@ class Program_weight_tensor_parameter_238: class Program_weight_tensor_parameter_239: name = "parameter_239" + original_name = "conv2d_6.w_0" shape = [32, 80, 1, 1] dtype = "float32" min_val = float("-1.12874") @@ -2590,6 +2830,7 @@ class Program_weight_tensor_parameter_239: class Program_weight_tensor_parameter_240: name = "parameter_240" + original_name = "batch_norm_5.b_0" shape = [80] dtype = "float32" min_val = float("-1.95446") @@ -2601,6 +2842,7 @@ class Program_weight_tensor_parameter_240: class Program_weight_tensor_parameter_241: name = "parameter_241" + original_name = "batch_norm_5.w_0" shape = [80] dtype = "float32" min_val = float("1.32848") @@ -2612,6 +2854,7 @@ class Program_weight_tensor_parameter_241: class Program_weight_tensor_parameter_242: name = "parameter_242" + original_name = "batch_norm_5.w_2" shape = [80] dtype = "float32" min_val = float("0.0302486") @@ -2623,6 +2866,7 @@ class Program_weight_tensor_parameter_242: class Program_weight_tensor_parameter_243: name = "parameter_243" + original_name = "batch_norm_5.w_1" shape = [80] dtype = "float32" min_val = float("-4.78155") @@ -2634,6 +2878,7 @@ class Program_weight_tensor_parameter_243: class Program_weight_tensor_parameter_244: name = "parameter_244" + original_name = "conv2d_5.w_0" shape = [80, 1, 3, 3] dtype = "float32" min_val = float("-0.453712") @@ -2645,6 +2890,7 @@ class Program_weight_tensor_parameter_244: class Program_weight_tensor_parameter_245: name = "parameter_245" + original_name = "batch_norm_4.b_0" shape = [80] dtype = "float32" min_val = float("-4.71815") @@ -2656,6 +2902,7 @@ class Program_weight_tensor_parameter_245: class Program_weight_tensor_parameter_246: name = "parameter_246" + original_name = "batch_norm_4.w_0" shape = [80] dtype = "float32" min_val = float("0.165056") @@ -2667,6 +2914,7 @@ class Program_weight_tensor_parameter_246: class Program_weight_tensor_parameter_247: name = "parameter_247" + original_name = "batch_norm_4.w_2" shape = [80] dtype = "float32" min_val = float("0.914846") @@ -2678,6 +2926,7 @@ class Program_weight_tensor_parameter_247: class Program_weight_tensor_parameter_248: name = "parameter_248" + original_name = "batch_norm_4.w_1" shape = [80] dtype = "float32" min_val = float("-4.30035") @@ -2689,6 +2938,7 @@ class Program_weight_tensor_parameter_248: class Program_weight_tensor_parameter_249: name = "parameter_249" + original_name = "conv2d_4.w_0" shape = [80, 24, 1, 1] dtype = "float32" min_val = float("-0.893577") @@ -2700,6 +2950,7 @@ class Program_weight_tensor_parameter_249: class Program_weight_tensor_parameter_250: name = "parameter_250" + original_name = "batch_norm_3.b_0" shape = [24] dtype = "float32" min_val = float("0") @@ -2709,6 +2960,7 @@ class Program_weight_tensor_parameter_250: class Program_weight_tensor_parameter_251: name = "parameter_251" + original_name = "batch_norm_3.w_0" shape = [24] dtype = "float32" min_val = float("0") @@ -2718,6 +2970,7 @@ class Program_weight_tensor_parameter_251: class Program_weight_tensor_parameter_252: name = "parameter_252" + original_name = "batch_norm_3.w_2" shape = [24] dtype = "float32" min_val = float("0") @@ -2727,6 +2980,7 @@ class Program_weight_tensor_parameter_252: class Program_weight_tensor_parameter_253: name = "parameter_253" + original_name = "batch_norm_3.w_1" shape = [24] dtype = "float32" min_val = float("0") @@ -2736,6 +2990,7 @@ class Program_weight_tensor_parameter_253: class Program_weight_tensor_parameter_254: name = "parameter_254" + original_name = "conv2d_3.w_0" shape = [24, 24, 1, 1] dtype = "float32" min_val = float("-0.852191") @@ -2747,6 +3002,7 @@ class Program_weight_tensor_parameter_254: class Program_weight_tensor_parameter_255: name = "parameter_255" + original_name = "batch_norm_2.b_0" shape = [24] dtype = "float32" min_val = float("0") @@ -2756,6 +3012,7 @@ class Program_weight_tensor_parameter_255: class Program_weight_tensor_parameter_256: name = "parameter_256" + original_name = "batch_norm_2.w_0" shape = [24] dtype = "float32" min_val = float("0") @@ -2765,6 +3022,7 @@ class Program_weight_tensor_parameter_256: class Program_weight_tensor_parameter_257: name = "parameter_257" + original_name = "batch_norm_2.w_2" shape = [24] dtype = "float32" min_val = float("0") @@ -2774,6 +3032,7 @@ class Program_weight_tensor_parameter_257: class Program_weight_tensor_parameter_258: name = "parameter_258" + original_name = "batch_norm_2.w_1" shape = [24] dtype = "float32" min_val = float("0") @@ -2783,6 +3042,7 @@ class Program_weight_tensor_parameter_258: class Program_weight_tensor_parameter_259: name = "parameter_259" + original_name = "conv2d_2.w_0" shape = [24, 1, 3, 3] dtype = "float32" min_val = float("-0.988106") @@ -2794,6 +3054,7 @@ class Program_weight_tensor_parameter_259: class Program_weight_tensor_parameter_260: name = "parameter_260" + original_name = "batch_norm_1.b_0" shape = [24] dtype = "float32" min_val = float("0") @@ -2803,6 +3064,7 @@ class Program_weight_tensor_parameter_260: class Program_weight_tensor_parameter_261: name = "parameter_261" + original_name = "batch_norm_1.w_0" shape = [24] dtype = "float32" min_val = float("0") @@ -2812,6 +3074,7 @@ class Program_weight_tensor_parameter_261: class Program_weight_tensor_parameter_262: name = "parameter_262" + original_name = "batch_norm_1.w_2" shape = [24] dtype = "float32" min_val = float("0") @@ -2821,6 +3084,7 @@ class Program_weight_tensor_parameter_262: class Program_weight_tensor_parameter_263: name = "parameter_263" + original_name = "batch_norm_1.w_1" shape = [24] dtype = "float32" min_val = float("0") @@ -2830,6 +3094,7 @@ class Program_weight_tensor_parameter_263: class Program_weight_tensor_parameter_264: name = "parameter_264" + original_name = "conv2d_1.w_0" shape = [24, 24, 1, 1] dtype = "float32" min_val = float("-1.21566") @@ -2841,6 +3106,7 @@ class Program_weight_tensor_parameter_264: class Program_weight_tensor_parameter_265: name = "parameter_265" + original_name = "batch_norm_0.b_0" shape = [24] dtype = "float32" min_val = float("0") @@ -2850,6 +3116,7 @@ class Program_weight_tensor_parameter_265: class Program_weight_tensor_parameter_266: name = "parameter_266" + original_name = "batch_norm_0.w_0" shape = [24] dtype = "float32" min_val = float("0") @@ -2859,6 +3126,7 @@ class Program_weight_tensor_parameter_266: class Program_weight_tensor_parameter_267: name = "parameter_267" + original_name = "batch_norm_0.w_2" shape = [24] dtype = "float32" min_val = float("0") @@ -2868,6 +3136,7 @@ class Program_weight_tensor_parameter_267: class Program_weight_tensor_parameter_268: name = "parameter_268" + original_name = "batch_norm_0.w_1" shape = [24] dtype = "float32" min_val = float("0") @@ -2877,6 +3146,7 @@ class Program_weight_tensor_parameter_268: class Program_weight_tensor_parameter_269: name = "parameter_269" + original_name = "conv2d_0.w_0" shape = [24, 3, 3, 3] dtype = "float32" min_val = float("-1.24032") diff --git a/paddle_samples/PaddleX/MobileNetV3_small_x0_35/subgraph_0/input_meta.py b/paddle_samples/PaddleX/MobileNetV3_small_x0_35/subgraph_0/input_meta.py index 4c1d87b63..7cf866357 100644 --- a/paddle_samples/PaddleX/MobileNetV3_small_x0_35/subgraph_0/input_meta.py +++ b/paddle_samples/PaddleX/MobileNetV3_small_x0_35/subgraph_0/input_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_data_0: name = "data_0" + original_name = "var_162" shape = [252, 3, 224, 224] dtype = "float32" min_val = float("-2.1179") diff --git a/paddle_samples/PaddleX/MobileNetV3_small_x0_35/subgraph_0/weight_meta.py b/paddle_samples/PaddleX/MobileNetV3_small_x0_35/subgraph_0/weight_meta.py index 9a3ab113c..b705f7940 100644 --- a/paddle_samples/PaddleX/MobileNetV3_small_x0_35/subgraph_0/weight_meta.py +++ b/paddle_samples/PaddleX/MobileNetV3_small_x0_35/subgraph_0/weight_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_parameter_0: name = "parameter_0" + original_name = "linear_0.b_0" shape = [102] dtype = "float32" min_val = float("-0.000591077") @@ -11,6 +12,7 @@ class Program_weight_tensor_parameter_0: class Program_weight_tensor_parameter_1: name = "parameter_1" + original_name = "linear_0.w_0" shape = [1280, 102] dtype = "float32" min_val = float("-0.0661976") @@ -22,6 +24,7 @@ class Program_weight_tensor_parameter_1: class Program_weight_tensor_parameter_2: name = "parameter_2" + original_name = "conv2d_53.w_0" shape = [1280, 200, 1, 1] dtype = "float32" min_val = float("-1.7897") @@ -33,6 +36,7 @@ class Program_weight_tensor_parameter_2: class Program_weight_tensor_parameter_3: name = "parameter_3" + original_name = "batch_norm_34.b_0" shape = [200] dtype = "float32" min_val = float("-14.7998") @@ -44,6 +48,7 @@ class Program_weight_tensor_parameter_3: class Program_weight_tensor_parameter_4: name = "parameter_4" + original_name = "batch_norm_34.w_0" shape = [200] dtype = "float32" min_val = float("2.13906") @@ -55,6 +60,7 @@ class Program_weight_tensor_parameter_4: class Program_weight_tensor_parameter_5: name = "parameter_5" + original_name = "batch_norm_34.w_2" shape = [200] dtype = "float32" min_val = float("350.577") @@ -66,6 +72,7 @@ class Program_weight_tensor_parameter_5: class Program_weight_tensor_parameter_6: name = "parameter_6" + original_name = "batch_norm_34.w_1" shape = [200] dtype = "float32" min_val = float("-3.91488e-05") @@ -77,6 +84,7 @@ class Program_weight_tensor_parameter_6: class Program_weight_tensor_parameter_7: name = "parameter_7" + original_name = "conv2d_52.w_0" shape = [200, 32, 1, 1] dtype = "float32" min_val = float("-1.53445") @@ -88,6 +96,7 @@ class Program_weight_tensor_parameter_7: class Program_weight_tensor_parameter_8: name = "parameter_8" + original_name = "batch_norm_33.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -97,6 +106,7 @@ class Program_weight_tensor_parameter_8: class Program_weight_tensor_parameter_9: name = "parameter_9" + original_name = "batch_norm_33.w_0" shape = [32] dtype = "float32" min_val = float("0") @@ -106,6 +116,7 @@ class Program_weight_tensor_parameter_9: class Program_weight_tensor_parameter_10: name = "parameter_10" + original_name = "batch_norm_33.w_2" shape = [32] dtype = "float32" min_val = float("0") @@ -115,6 +126,7 @@ class Program_weight_tensor_parameter_10: class Program_weight_tensor_parameter_11: name = "parameter_11" + original_name = "batch_norm_33.w_1" shape = [32] dtype = "float32" min_val = float("0") @@ -124,6 +136,7 @@ class Program_weight_tensor_parameter_11: class Program_weight_tensor_parameter_12: name = "parameter_12" + original_name = "conv2d_51.w_0" shape = [32, 200, 1, 1] dtype = "float32" min_val = float("-1.37395") @@ -135,6 +148,7 @@ class Program_weight_tensor_parameter_12: class Program_weight_tensor_parameter_13: name = "parameter_13" + original_name = "conv2d_50.b_0" shape = [200] dtype = "float32" min_val = float("-0.760746") @@ -146,6 +160,7 @@ class Program_weight_tensor_parameter_13: class Program_weight_tensor_parameter_14: name = "parameter_14" + original_name = "conv2d_50.w_0" shape = [200, 50, 1, 1] dtype = "float32" min_val = float("-1.20617") @@ -157,6 +172,7 @@ class Program_weight_tensor_parameter_14: class Program_weight_tensor_parameter_15: name = "parameter_15" + original_name = "conv2d_49.b_0" shape = [50] dtype = "float32" min_val = float("0") @@ -166,6 +182,7 @@ class Program_weight_tensor_parameter_15: class Program_weight_tensor_parameter_16: name = "parameter_16" + original_name = "conv2d_49.w_0" shape = [50, 200, 1, 1] dtype = "float32" min_val = float("-1.51989") @@ -177,6 +194,7 @@ class Program_weight_tensor_parameter_16: class Program_weight_tensor_parameter_17: name = "parameter_17" + original_name = "batch_norm_32.b_0" shape = [200] dtype = "float32" min_val = float("-3.26573") @@ -188,6 +206,7 @@ class Program_weight_tensor_parameter_17: class Program_weight_tensor_parameter_18: name = "parameter_18" + original_name = "batch_norm_32.w_0" shape = [200] dtype = "float32" min_val = float("0.254814") @@ -199,6 +218,7 @@ class Program_weight_tensor_parameter_18: class Program_weight_tensor_parameter_19: name = "parameter_19" + original_name = "batch_norm_32.w_2" shape = [200] dtype = "float32" min_val = float("0.000303796") @@ -210,6 +230,7 @@ class Program_weight_tensor_parameter_19: class Program_weight_tensor_parameter_20: name = "parameter_20" + original_name = "batch_norm_32.w_1" shape = [200] dtype = "float32" min_val = float("-22.0479") @@ -221,6 +242,7 @@ class Program_weight_tensor_parameter_20: class Program_weight_tensor_parameter_21: name = "parameter_21" + original_name = "conv2d_48.w_0" shape = [200, 1, 5, 5] dtype = "float32" min_val = float("-1.84688") @@ -232,6 +254,7 @@ class Program_weight_tensor_parameter_21: class Program_weight_tensor_parameter_22: name = "parameter_22" + original_name = "batch_norm_31.b_0" shape = [200] dtype = "float32" min_val = float("-3.74953") @@ -243,6 +266,7 @@ class Program_weight_tensor_parameter_22: class Program_weight_tensor_parameter_23: name = "parameter_23" + original_name = "batch_norm_31.w_0" shape = [200] dtype = "float32" min_val = float("-3.73384") @@ -254,6 +278,7 @@ class Program_weight_tensor_parameter_23: class Program_weight_tensor_parameter_24: name = "parameter_24" + original_name = "batch_norm_31.w_2" shape = [200] dtype = "float32" min_val = float("60.1739") @@ -265,6 +290,7 @@ class Program_weight_tensor_parameter_24: class Program_weight_tensor_parameter_25: name = "parameter_25" + original_name = "batch_norm_31.w_1" shape = [200] dtype = "float32" min_val = float("-3.31935e-05") @@ -276,6 +302,7 @@ class Program_weight_tensor_parameter_25: class Program_weight_tensor_parameter_26: name = "parameter_26" + original_name = "conv2d_47.w_0" shape = [200, 32, 1, 1] dtype = "float32" min_val = float("-1.17825") @@ -287,6 +314,7 @@ class Program_weight_tensor_parameter_26: class Program_weight_tensor_parameter_27: name = "parameter_27" + original_name = "batch_norm_30.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -296,6 +324,7 @@ class Program_weight_tensor_parameter_27: class Program_weight_tensor_parameter_28: name = "parameter_28" + original_name = "batch_norm_30.w_0" shape = [32] dtype = "float32" min_val = float("0") @@ -305,6 +334,7 @@ class Program_weight_tensor_parameter_28: class Program_weight_tensor_parameter_29: name = "parameter_29" + original_name = "batch_norm_30.w_2" shape = [32] dtype = "float32" min_val = float("0") @@ -314,6 +344,7 @@ class Program_weight_tensor_parameter_29: class Program_weight_tensor_parameter_30: name = "parameter_30" + original_name = "batch_norm_30.w_1" shape = [32] dtype = "float32" min_val = float("0") @@ -323,6 +354,7 @@ class Program_weight_tensor_parameter_30: class Program_weight_tensor_parameter_31: name = "parameter_31" + original_name = "conv2d_46.w_0" shape = [32, 200, 1, 1] dtype = "float32" min_val = float("-1.09165") @@ -334,6 +366,7 @@ class Program_weight_tensor_parameter_31: class Program_weight_tensor_parameter_32: name = "parameter_32" + original_name = "conv2d_45.b_0" shape = [200] dtype = "float32" min_val = float("-0.756993") @@ -345,6 +378,7 @@ class Program_weight_tensor_parameter_32: class Program_weight_tensor_parameter_33: name = "parameter_33" + original_name = "conv2d_45.w_0" shape = [200, 50, 1, 1] dtype = "float32" min_val = float("-1.35507") @@ -356,6 +390,7 @@ class Program_weight_tensor_parameter_33: class Program_weight_tensor_parameter_34: name = "parameter_34" + original_name = "conv2d_44.b_0" shape = [50] dtype = "float32" min_val = float("0") @@ -365,6 +400,7 @@ class Program_weight_tensor_parameter_34: class Program_weight_tensor_parameter_35: name = "parameter_35" + original_name = "conv2d_44.w_0" shape = [50, 200, 1, 1] dtype = "float32" min_val = float("-1.59292") @@ -376,6 +412,7 @@ class Program_weight_tensor_parameter_35: class Program_weight_tensor_parameter_36: name = "parameter_36" + original_name = "batch_norm_29.b_0" shape = [200] dtype = "float32" min_val = float("-4.71951") @@ -387,6 +424,7 @@ class Program_weight_tensor_parameter_36: class Program_weight_tensor_parameter_37: name = "parameter_37" + original_name = "batch_norm_29.w_0" shape = [200] dtype = "float32" min_val = float("0.399404") @@ -398,6 +436,7 @@ class Program_weight_tensor_parameter_37: class Program_weight_tensor_parameter_38: name = "parameter_38" + original_name = "batch_norm_29.w_2" shape = [200] dtype = "float32" min_val = float("0.000537403") @@ -409,6 +448,7 @@ class Program_weight_tensor_parameter_38: class Program_weight_tensor_parameter_39: name = "parameter_39" + original_name = "batch_norm_29.w_1" shape = [200] dtype = "float32" min_val = float("-15.9929") @@ -420,6 +460,7 @@ class Program_weight_tensor_parameter_39: class Program_weight_tensor_parameter_40: name = "parameter_40" + original_name = "conv2d_43.w_0" shape = [200, 1, 5, 5] dtype = "float32" min_val = float("-2.15566") @@ -431,6 +472,7 @@ class Program_weight_tensor_parameter_40: class Program_weight_tensor_parameter_41: name = "parameter_41" + original_name = "batch_norm_28.b_0" shape = [200] dtype = "float32" min_val = float("-2.69705") @@ -442,6 +484,7 @@ class Program_weight_tensor_parameter_41: class Program_weight_tensor_parameter_42: name = "parameter_42" + original_name = "batch_norm_28.w_0" shape = [200] dtype = "float32" min_val = float("-0.515722") @@ -453,6 +496,7 @@ class Program_weight_tensor_parameter_42: class Program_weight_tensor_parameter_43: name = "parameter_43" + original_name = "batch_norm_28.w_2" shape = [200] dtype = "float32" min_val = float("34.1294") @@ -464,6 +508,7 @@ class Program_weight_tensor_parameter_43: class Program_weight_tensor_parameter_44: name = "parameter_44" + original_name = "batch_norm_28.w_1" shape = [200] dtype = "float32" min_val = float("-1.67736e-05") @@ -475,6 +520,7 @@ class Program_weight_tensor_parameter_44: class Program_weight_tensor_parameter_45: name = "parameter_45" + original_name = "conv2d_42.w_0" shape = [200, 32, 1, 1] dtype = "float32" min_val = float("-1.38004") @@ -486,6 +532,7 @@ class Program_weight_tensor_parameter_45: class Program_weight_tensor_parameter_46: name = "parameter_46" + original_name = "batch_norm_27.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -495,6 +542,7 @@ class Program_weight_tensor_parameter_46: class Program_weight_tensor_parameter_47: name = "parameter_47" + original_name = "batch_norm_27.w_0" shape = [32] dtype = "float32" min_val = float("0") @@ -504,6 +552,7 @@ class Program_weight_tensor_parameter_47: class Program_weight_tensor_parameter_48: name = "parameter_48" + original_name = "batch_norm_27.w_2" shape = [32] dtype = "float32" min_val = float("0") @@ -513,6 +562,7 @@ class Program_weight_tensor_parameter_48: class Program_weight_tensor_parameter_49: name = "parameter_49" + original_name = "batch_norm_27.w_1" shape = [32] dtype = "float32" min_val = float("0") @@ -522,6 +572,7 @@ class Program_weight_tensor_parameter_49: class Program_weight_tensor_parameter_50: name = "parameter_50" + original_name = "conv2d_41.w_0" shape = [32, 104, 1, 1] dtype = "float32" min_val = float("-1.82206") @@ -533,6 +584,7 @@ class Program_weight_tensor_parameter_50: class Program_weight_tensor_parameter_51: name = "parameter_51" + original_name = "conv2d_40.b_0" shape = [104] dtype = "float32" min_val = float("-0.863671") @@ -544,6 +596,7 @@ class Program_weight_tensor_parameter_51: class Program_weight_tensor_parameter_52: name = "parameter_52" + original_name = "conv2d_40.w_0" shape = [104, 26, 1, 1] dtype = "float32" min_val = float("-6.48105e-06") @@ -555,6 +608,7 @@ class Program_weight_tensor_parameter_52: class Program_weight_tensor_parameter_53: name = "parameter_53" + original_name = "conv2d_39.b_0" shape = [26] dtype = "float32" min_val = float("0") @@ -564,6 +618,7 @@ class Program_weight_tensor_parameter_53: class Program_weight_tensor_parameter_54: name = "parameter_54" + original_name = "conv2d_39.w_0" shape = [26, 104, 1, 1] dtype = "float32" min_val = float("-9.18523e-06") @@ -575,6 +630,7 @@ class Program_weight_tensor_parameter_54: class Program_weight_tensor_parameter_55: name = "parameter_55" + original_name = "batch_norm_26.b_0" shape = [104] dtype = "float32" min_val = float("-1.45437") @@ -586,6 +642,7 @@ class Program_weight_tensor_parameter_55: class Program_weight_tensor_parameter_56: name = "parameter_56" + original_name = "batch_norm_26.w_0" shape = [104] dtype = "float32" min_val = float("0.518953") @@ -597,6 +654,7 @@ class Program_weight_tensor_parameter_56: class Program_weight_tensor_parameter_57: name = "parameter_57" + original_name = "batch_norm_26.w_2" shape = [104] dtype = "float32" min_val = float("0.0210462") @@ -608,6 +666,7 @@ class Program_weight_tensor_parameter_57: class Program_weight_tensor_parameter_58: name = "parameter_58" + original_name = "batch_norm_26.w_1" shape = [104] dtype = "float32" min_val = float("-108.809") @@ -619,6 +678,7 @@ class Program_weight_tensor_parameter_58: class Program_weight_tensor_parameter_59: name = "parameter_59" + original_name = "conv2d_38.w_0" shape = [104, 1, 5, 5] dtype = "float32" min_val = float("-0.835291") @@ -630,6 +690,7 @@ class Program_weight_tensor_parameter_59: class Program_weight_tensor_parameter_60: name = "parameter_60" + original_name = "batch_norm_25.b_0" shape = [104] dtype = "float32" min_val = float("-4.98332") @@ -641,6 +702,7 @@ class Program_weight_tensor_parameter_60: class Program_weight_tensor_parameter_61: name = "parameter_61" + original_name = "batch_norm_25.w_0" shape = [104] dtype = "float32" min_val = float("0.305118") @@ -652,6 +714,7 @@ class Program_weight_tensor_parameter_61: class Program_weight_tensor_parameter_62: name = "parameter_62" + original_name = "batch_norm_25.w_2" shape = [104] dtype = "float32" min_val = float("107.365") @@ -663,6 +726,7 @@ class Program_weight_tensor_parameter_62: class Program_weight_tensor_parameter_63: name = "parameter_63" + original_name = "batch_norm_25.w_1" shape = [104] dtype = "float32" min_val = float("-9.44327e-05") @@ -674,6 +738,7 @@ class Program_weight_tensor_parameter_63: class Program_weight_tensor_parameter_64: name = "parameter_64" + original_name = "conv2d_37.w_0" shape = [104, 16, 1, 1] dtype = "float32" min_val = float("-2.0615") @@ -685,6 +750,7 @@ class Program_weight_tensor_parameter_64: class Program_weight_tensor_parameter_65: name = "parameter_65" + original_name = "batch_norm_24.b_0" shape = [16] dtype = "float32" min_val = float("0") @@ -694,6 +760,7 @@ class Program_weight_tensor_parameter_65: class Program_weight_tensor_parameter_66: name = "parameter_66" + original_name = "batch_norm_24.w_0" shape = [16] dtype = "float32" min_val = float("0") @@ -703,6 +770,7 @@ class Program_weight_tensor_parameter_66: class Program_weight_tensor_parameter_67: name = "parameter_67" + original_name = "batch_norm_24.w_2" shape = [16] dtype = "float32" min_val = float("0") @@ -712,6 +780,7 @@ class Program_weight_tensor_parameter_67: class Program_weight_tensor_parameter_68: name = "parameter_68" + original_name = "batch_norm_24.w_1" shape = [16] dtype = "float32" min_val = float("0") @@ -721,6 +790,7 @@ class Program_weight_tensor_parameter_68: class Program_weight_tensor_parameter_69: name = "parameter_69" + original_name = "conv2d_36.w_0" shape = [16, 48, 1, 1] dtype = "float32" min_val = float("-1.79193") @@ -732,6 +802,7 @@ class Program_weight_tensor_parameter_69: class Program_weight_tensor_parameter_70: name = "parameter_70" + original_name = "conv2d_35.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -741,6 +812,7 @@ class Program_weight_tensor_parameter_70: class Program_weight_tensor_parameter_71: name = "parameter_71" + original_name = "conv2d_35.w_0" shape = [48, 12, 1, 1] dtype = "float32" min_val = float("-1.2635") @@ -752,6 +824,7 @@ class Program_weight_tensor_parameter_71: class Program_weight_tensor_parameter_72: name = "parameter_72" + original_name = "conv2d_34.b_0" shape = [12] dtype = "float32" min_val = float("0") @@ -761,6 +834,7 @@ class Program_weight_tensor_parameter_72: class Program_weight_tensor_parameter_73: name = "parameter_73" + original_name = "conv2d_34.w_0" shape = [12, 48, 1, 1] dtype = "float32" min_val = float("-1.16864") @@ -772,6 +846,7 @@ class Program_weight_tensor_parameter_73: class Program_weight_tensor_parameter_74: name = "parameter_74" + original_name = "batch_norm_23.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -781,6 +856,7 @@ class Program_weight_tensor_parameter_74: class Program_weight_tensor_parameter_75: name = "parameter_75" + original_name = "batch_norm_23.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -790,6 +866,7 @@ class Program_weight_tensor_parameter_75: class Program_weight_tensor_parameter_76: name = "parameter_76" + original_name = "batch_norm_23.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -799,6 +876,7 @@ class Program_weight_tensor_parameter_76: class Program_weight_tensor_parameter_77: name = "parameter_77" + original_name = "batch_norm_23.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -808,6 +886,7 @@ class Program_weight_tensor_parameter_77: class Program_weight_tensor_parameter_78: name = "parameter_78" + original_name = "conv2d_33.w_0" shape = [48, 1, 5, 5] dtype = "float32" min_val = float("-1.98015") @@ -819,6 +898,7 @@ class Program_weight_tensor_parameter_78: class Program_weight_tensor_parameter_79: name = "parameter_79" + original_name = "batch_norm_22.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -828,6 +908,7 @@ class Program_weight_tensor_parameter_79: class Program_weight_tensor_parameter_80: name = "parameter_80" + original_name = "batch_norm_22.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -837,6 +918,7 @@ class Program_weight_tensor_parameter_80: class Program_weight_tensor_parameter_81: name = "parameter_81" + original_name = "batch_norm_22.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -846,6 +928,7 @@ class Program_weight_tensor_parameter_81: class Program_weight_tensor_parameter_82: name = "parameter_82" + original_name = "batch_norm_22.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -855,6 +938,7 @@ class Program_weight_tensor_parameter_82: class Program_weight_tensor_parameter_83: name = "parameter_83" + original_name = "conv2d_32.w_0" shape = [48, 16, 1, 1] dtype = "float32" min_val = float("-1.53451") @@ -866,6 +950,7 @@ class Program_weight_tensor_parameter_83: class Program_weight_tensor_parameter_84: name = "parameter_84" + original_name = "batch_norm_21.b_0" shape = [16] dtype = "float32" min_val = float("0") @@ -875,6 +960,7 @@ class Program_weight_tensor_parameter_84: class Program_weight_tensor_parameter_85: name = "parameter_85" + original_name = "batch_norm_21.w_0" shape = [16] dtype = "float32" min_val = float("0") @@ -884,6 +970,7 @@ class Program_weight_tensor_parameter_85: class Program_weight_tensor_parameter_86: name = "parameter_86" + original_name = "batch_norm_21.w_2" shape = [16] dtype = "float32" min_val = float("0") @@ -893,6 +980,7 @@ class Program_weight_tensor_parameter_86: class Program_weight_tensor_parameter_87: name = "parameter_87" + original_name = "batch_norm_21.w_1" shape = [16] dtype = "float32" min_val = float("0") @@ -902,6 +990,7 @@ class Program_weight_tensor_parameter_87: class Program_weight_tensor_parameter_88: name = "parameter_88" + original_name = "conv2d_31.w_0" shape = [16, 40, 1, 1] dtype = "float32" min_val = float("-1.70614") @@ -913,6 +1002,7 @@ class Program_weight_tensor_parameter_88: class Program_weight_tensor_parameter_89: name = "parameter_89" + original_name = "conv2d_30.b_0" shape = [40] dtype = "float32" min_val = float("0") @@ -922,6 +1012,7 @@ class Program_weight_tensor_parameter_89: class Program_weight_tensor_parameter_90: name = "parameter_90" + original_name = "conv2d_30.w_0" shape = [40, 10, 1, 1] dtype = "float32" min_val = float("-1.28617") @@ -933,6 +1024,7 @@ class Program_weight_tensor_parameter_90: class Program_weight_tensor_parameter_91: name = "parameter_91" + original_name = "conv2d_29.b_0" shape = [10] dtype = "float32" min_val = float("0") @@ -942,6 +1034,7 @@ class Program_weight_tensor_parameter_91: class Program_weight_tensor_parameter_92: name = "parameter_92" + original_name = "conv2d_29.w_0" shape = [10, 40, 1, 1] dtype = "float32" min_val = float("-2.04638") @@ -953,6 +1046,7 @@ class Program_weight_tensor_parameter_92: class Program_weight_tensor_parameter_93: name = "parameter_93" + original_name = "batch_norm_20.b_0" shape = [40] dtype = "float32" min_val = float("0") @@ -962,6 +1056,7 @@ class Program_weight_tensor_parameter_93: class Program_weight_tensor_parameter_94: name = "parameter_94" + original_name = "batch_norm_20.w_0" shape = [40] dtype = "float32" min_val = float("0") @@ -971,6 +1066,7 @@ class Program_weight_tensor_parameter_94: class Program_weight_tensor_parameter_95: name = "parameter_95" + original_name = "batch_norm_20.w_2" shape = [40] dtype = "float32" min_val = float("0") @@ -980,6 +1076,7 @@ class Program_weight_tensor_parameter_95: class Program_weight_tensor_parameter_96: name = "parameter_96" + original_name = "batch_norm_20.w_1" shape = [40] dtype = "float32" min_val = float("0") @@ -989,6 +1086,7 @@ class Program_weight_tensor_parameter_96: class Program_weight_tensor_parameter_97: name = "parameter_97" + original_name = "conv2d_28.w_0" shape = [40, 1, 5, 5] dtype = "float32" min_val = float("-1.83334") @@ -1000,6 +1098,7 @@ class Program_weight_tensor_parameter_97: class Program_weight_tensor_parameter_98: name = "parameter_98" + original_name = "batch_norm_19.b_0" shape = [40] dtype = "float32" min_val = float("0") @@ -1009,6 +1108,7 @@ class Program_weight_tensor_parameter_98: class Program_weight_tensor_parameter_99: name = "parameter_99" + original_name = "batch_norm_19.w_0" shape = [40] dtype = "float32" min_val = float("0") @@ -1018,6 +1118,7 @@ class Program_weight_tensor_parameter_99: class Program_weight_tensor_parameter_100: name = "parameter_100" + original_name = "batch_norm_19.w_2" shape = [40] dtype = "float32" min_val = float("0") @@ -1027,6 +1128,7 @@ class Program_weight_tensor_parameter_100: class Program_weight_tensor_parameter_101: name = "parameter_101" + original_name = "batch_norm_19.w_1" shape = [40] dtype = "float32" min_val = float("0") @@ -1036,6 +1138,7 @@ class Program_weight_tensor_parameter_101: class Program_weight_tensor_parameter_102: name = "parameter_102" + original_name = "conv2d_27.w_0" shape = [40, 16, 1, 1] dtype = "float32" min_val = float("-1.44659") @@ -1047,6 +1150,7 @@ class Program_weight_tensor_parameter_102: class Program_weight_tensor_parameter_103: name = "parameter_103" + original_name = "batch_norm_18.b_0" shape = [16] dtype = "float32" min_val = float("0") @@ -1056,6 +1160,7 @@ class Program_weight_tensor_parameter_103: class Program_weight_tensor_parameter_104: name = "parameter_104" + original_name = "batch_norm_18.w_0" shape = [16] dtype = "float32" min_val = float("0") @@ -1065,6 +1170,7 @@ class Program_weight_tensor_parameter_104: class Program_weight_tensor_parameter_105: name = "parameter_105" + original_name = "batch_norm_18.w_2" shape = [16] dtype = "float32" min_val = float("0") @@ -1074,6 +1180,7 @@ class Program_weight_tensor_parameter_105: class Program_weight_tensor_parameter_106: name = "parameter_106" + original_name = "batch_norm_18.w_1" shape = [16] dtype = "float32" min_val = float("0") @@ -1083,6 +1190,7 @@ class Program_weight_tensor_parameter_106: class Program_weight_tensor_parameter_107: name = "parameter_107" + original_name = "conv2d_26.w_0" shape = [16, 88, 1, 1] dtype = "float32" min_val = float("-1.62167") @@ -1094,6 +1202,7 @@ class Program_weight_tensor_parameter_107: class Program_weight_tensor_parameter_108: name = "parameter_108" + original_name = "conv2d_25.b_0" shape = [88] dtype = "float32" min_val = float("-0.714293") @@ -1105,6 +1214,7 @@ class Program_weight_tensor_parameter_108: class Program_weight_tensor_parameter_109: name = "parameter_109" + original_name = "conv2d_25.w_0" shape = [88, 22, 1, 1] dtype = "float32" min_val = float("-1.32424") @@ -1116,6 +1226,7 @@ class Program_weight_tensor_parameter_109: class Program_weight_tensor_parameter_110: name = "parameter_110" + original_name = "conv2d_24.b_0" shape = [22] dtype = "float32" min_val = float("0") @@ -1125,6 +1236,7 @@ class Program_weight_tensor_parameter_110: class Program_weight_tensor_parameter_111: name = "parameter_111" + original_name = "conv2d_24.w_0" shape = [22, 88, 1, 1] dtype = "float32" min_val = float("-1.81078") @@ -1136,6 +1248,7 @@ class Program_weight_tensor_parameter_111: class Program_weight_tensor_parameter_112: name = "parameter_112" + original_name = "batch_norm_17.b_0" shape = [88] dtype = "float32" min_val = float("-2.13498") @@ -1147,6 +1260,7 @@ class Program_weight_tensor_parameter_112: class Program_weight_tensor_parameter_113: name = "parameter_113" + original_name = "batch_norm_17.w_0" shape = [88] dtype = "float32" min_val = float("0.596636") @@ -1158,6 +1272,7 @@ class Program_weight_tensor_parameter_113: class Program_weight_tensor_parameter_114: name = "parameter_114" + original_name = "batch_norm_17.w_2" shape = [88] dtype = "float32" min_val = float("0.00302702") @@ -1169,6 +1284,7 @@ class Program_weight_tensor_parameter_114: class Program_weight_tensor_parameter_115: name = "parameter_115" + original_name = "batch_norm_17.w_1" shape = [88] dtype = "float32" min_val = float("-6.7775") @@ -1180,6 +1296,7 @@ class Program_weight_tensor_parameter_115: class Program_weight_tensor_parameter_116: name = "parameter_116" + original_name = "conv2d_23.w_0" shape = [88, 1, 5, 5] dtype = "float32" min_val = float("-2.40667") @@ -1191,6 +1308,7 @@ class Program_weight_tensor_parameter_116: class Program_weight_tensor_parameter_117: name = "parameter_117" + original_name = "batch_norm_16.b_0" shape = [88] dtype = "float32" min_val = float("-2.05945") @@ -1202,6 +1320,7 @@ class Program_weight_tensor_parameter_117: class Program_weight_tensor_parameter_118: name = "parameter_118" + original_name = "batch_norm_16.w_0" shape = [88] dtype = "float32" min_val = float("0.0731721") @@ -1213,6 +1332,7 @@ class Program_weight_tensor_parameter_118: class Program_weight_tensor_parameter_119: name = "parameter_119" + original_name = "batch_norm_16.w_2" shape = [88] dtype = "float32" min_val = float("97.4765") @@ -1224,6 +1344,7 @@ class Program_weight_tensor_parameter_119: class Program_weight_tensor_parameter_120: name = "parameter_120" + original_name = "batch_norm_16.w_1" shape = [88] dtype = "float32" min_val = float("-2.54512e-05") @@ -1235,6 +1356,7 @@ class Program_weight_tensor_parameter_120: class Program_weight_tensor_parameter_121: name = "parameter_121" + original_name = "conv2d_22.w_0" shape = [88, 16, 1, 1] dtype = "float32" min_val = float("-1.3194") @@ -1246,6 +1368,7 @@ class Program_weight_tensor_parameter_121: class Program_weight_tensor_parameter_122: name = "parameter_122" + original_name = "batch_norm_15.b_0" shape = [16] dtype = "float32" min_val = float("0") @@ -1255,6 +1378,7 @@ class Program_weight_tensor_parameter_122: class Program_weight_tensor_parameter_123: name = "parameter_123" + original_name = "batch_norm_15.w_0" shape = [16] dtype = "float32" min_val = float("0") @@ -1264,6 +1388,7 @@ class Program_weight_tensor_parameter_123: class Program_weight_tensor_parameter_124: name = "parameter_124" + original_name = "batch_norm_15.w_2" shape = [16] dtype = "float32" min_val = float("0") @@ -1273,6 +1398,7 @@ class Program_weight_tensor_parameter_124: class Program_weight_tensor_parameter_125: name = "parameter_125" + original_name = "batch_norm_15.w_1" shape = [16] dtype = "float32" min_val = float("0") @@ -1282,6 +1408,7 @@ class Program_weight_tensor_parameter_125: class Program_weight_tensor_parameter_126: name = "parameter_126" + original_name = "conv2d_21.w_0" shape = [16, 88, 1, 1] dtype = "float32" min_val = float("-2.33937") @@ -1293,6 +1420,7 @@ class Program_weight_tensor_parameter_126: class Program_weight_tensor_parameter_127: name = "parameter_127" + original_name = "conv2d_20.b_0" shape = [88] dtype = "float32" min_val = float("-0.740746") @@ -1304,6 +1432,7 @@ class Program_weight_tensor_parameter_127: class Program_weight_tensor_parameter_128: name = "parameter_128" + original_name = "conv2d_20.w_0" shape = [88, 22, 1, 1] dtype = "float32" min_val = float("-1.51619") @@ -1315,6 +1444,7 @@ class Program_weight_tensor_parameter_128: class Program_weight_tensor_parameter_129: name = "parameter_129" + original_name = "conv2d_19.b_0" shape = [22] dtype = "float32" min_val = float("0") @@ -1324,6 +1454,7 @@ class Program_weight_tensor_parameter_129: class Program_weight_tensor_parameter_130: name = "parameter_130" + original_name = "conv2d_19.w_0" shape = [22, 88, 1, 1] dtype = "float32" min_val = float("-2.62225") @@ -1335,6 +1466,7 @@ class Program_weight_tensor_parameter_130: class Program_weight_tensor_parameter_131: name = "parameter_131" + original_name = "batch_norm_14.b_0" shape = [88] dtype = "float32" min_val = float("-5.89573") @@ -1346,6 +1478,7 @@ class Program_weight_tensor_parameter_131: class Program_weight_tensor_parameter_132: name = "parameter_132" + original_name = "batch_norm_14.w_0" shape = [88] dtype = "float32" min_val = float("0.255124") @@ -1357,6 +1490,7 @@ class Program_weight_tensor_parameter_132: class Program_weight_tensor_parameter_133: name = "parameter_133" + original_name = "batch_norm_14.w_2" shape = [88] dtype = "float32" min_val = float("0.00149338") @@ -1368,6 +1502,7 @@ class Program_weight_tensor_parameter_133: class Program_weight_tensor_parameter_134: name = "parameter_134" + original_name = "batch_norm_14.w_1" shape = [88] dtype = "float32" min_val = float("-12.688") @@ -1379,6 +1514,7 @@ class Program_weight_tensor_parameter_134: class Program_weight_tensor_parameter_135: name = "parameter_135" + original_name = "conv2d_18.w_0" shape = [88, 1, 5, 5] dtype = "float32" min_val = float("-2.69026") @@ -1390,6 +1526,7 @@ class Program_weight_tensor_parameter_135: class Program_weight_tensor_parameter_136: name = "parameter_136" + original_name = "batch_norm_13.b_0" shape = [88] dtype = "float32" min_val = float("-2.15838") @@ -1401,6 +1538,7 @@ class Program_weight_tensor_parameter_136: class Program_weight_tensor_parameter_137: name = "parameter_137" + original_name = "batch_norm_13.w_0" shape = [88] dtype = "float32" min_val = float("-3.28194") @@ -1412,6 +1550,7 @@ class Program_weight_tensor_parameter_137: class Program_weight_tensor_parameter_138: name = "parameter_138" + original_name = "batch_norm_13.w_2" shape = [88] dtype = "float32" min_val = float("58.6424") @@ -1423,6 +1562,7 @@ class Program_weight_tensor_parameter_138: class Program_weight_tensor_parameter_139: name = "parameter_139" + original_name = "batch_norm_13.w_1" shape = [88] dtype = "float32" min_val = float("-1.84776e-05") @@ -1434,6 +1574,7 @@ class Program_weight_tensor_parameter_139: class Program_weight_tensor_parameter_140: name = "parameter_140" + original_name = "conv2d_17.w_0" shape = [88, 16, 1, 1] dtype = "float32" min_val = float("-2.04321") @@ -1445,6 +1586,7 @@ class Program_weight_tensor_parameter_140: class Program_weight_tensor_parameter_141: name = "parameter_141" + original_name = "batch_norm_12.b_0" shape = [16] dtype = "float32" min_val = float("0") @@ -1454,6 +1596,7 @@ class Program_weight_tensor_parameter_141: class Program_weight_tensor_parameter_142: name = "parameter_142" + original_name = "batch_norm_12.w_0" shape = [16] dtype = "float32" min_val = float("0") @@ -1463,6 +1606,7 @@ class Program_weight_tensor_parameter_142: class Program_weight_tensor_parameter_143: name = "parameter_143" + original_name = "batch_norm_12.w_2" shape = [16] dtype = "float32" min_val = float("0") @@ -1472,6 +1616,7 @@ class Program_weight_tensor_parameter_143: class Program_weight_tensor_parameter_144: name = "parameter_144" + original_name = "batch_norm_12.w_1" shape = [16] dtype = "float32" min_val = float("0") @@ -1481,6 +1626,7 @@ class Program_weight_tensor_parameter_144: class Program_weight_tensor_parameter_145: name = "parameter_145" + original_name = "conv2d_16.w_0" shape = [16, 32, 1, 1] dtype = "float32" min_val = float("-3.00913") @@ -1492,6 +1638,7 @@ class Program_weight_tensor_parameter_145: class Program_weight_tensor_parameter_146: name = "parameter_146" + original_name = "conv2d_15.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -1501,6 +1648,7 @@ class Program_weight_tensor_parameter_146: class Program_weight_tensor_parameter_147: name = "parameter_147" + original_name = "conv2d_15.w_0" shape = [32, 8, 1, 1] dtype = "float32" min_val = float("-4.03416e-07") @@ -1512,6 +1660,7 @@ class Program_weight_tensor_parameter_147: class Program_weight_tensor_parameter_148: name = "parameter_148" + original_name = "conv2d_14.b_0" shape = [8] dtype = "float32" min_val = float("0") @@ -1521,6 +1670,7 @@ class Program_weight_tensor_parameter_148: class Program_weight_tensor_parameter_149: name = "parameter_149" + original_name = "conv2d_14.w_0" shape = [8, 32, 1, 1] dtype = "float32" min_val = float("-3.19053e-07") @@ -1532,6 +1682,7 @@ class Program_weight_tensor_parameter_149: class Program_weight_tensor_parameter_150: name = "parameter_150" + original_name = "batch_norm_11.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -1541,6 +1692,7 @@ class Program_weight_tensor_parameter_150: class Program_weight_tensor_parameter_151: name = "parameter_151" + original_name = "batch_norm_11.w_0" shape = [32] dtype = "float32" min_val = float("0") @@ -1550,6 +1702,7 @@ class Program_weight_tensor_parameter_151: class Program_weight_tensor_parameter_152: name = "parameter_152" + original_name = "batch_norm_11.w_2" shape = [32] dtype = "float32" min_val = float("0") @@ -1559,6 +1712,7 @@ class Program_weight_tensor_parameter_152: class Program_weight_tensor_parameter_153: name = "parameter_153" + original_name = "batch_norm_11.w_1" shape = [32] dtype = "float32" min_val = float("0") @@ -1568,6 +1722,7 @@ class Program_weight_tensor_parameter_153: class Program_weight_tensor_parameter_154: name = "parameter_154" + original_name = "conv2d_13.w_0" shape = [32, 1, 5, 5] dtype = "float32" min_val = float("-1.49799") @@ -1579,6 +1734,7 @@ class Program_weight_tensor_parameter_154: class Program_weight_tensor_parameter_155: name = "parameter_155" + original_name = "batch_norm_10.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -1588,6 +1744,7 @@ class Program_weight_tensor_parameter_155: class Program_weight_tensor_parameter_156: name = "parameter_156" + original_name = "batch_norm_10.w_0" shape = [32] dtype = "float32" min_val = float("0") @@ -1597,6 +1754,7 @@ class Program_weight_tensor_parameter_156: class Program_weight_tensor_parameter_157: name = "parameter_157" + original_name = "batch_norm_10.w_2" shape = [32] dtype = "float32" min_val = float("0") @@ -1606,6 +1764,7 @@ class Program_weight_tensor_parameter_157: class Program_weight_tensor_parameter_158: name = "parameter_158" + original_name = "batch_norm_10.w_1" shape = [32] dtype = "float32" min_val = float("0") @@ -1615,6 +1774,7 @@ class Program_weight_tensor_parameter_158: class Program_weight_tensor_parameter_159: name = "parameter_159" + original_name = "conv2d_12.w_0" shape = [32, 8, 1, 1] dtype = "float32" min_val = float("-2.71526") @@ -1626,6 +1786,7 @@ class Program_weight_tensor_parameter_159: class Program_weight_tensor_parameter_160: name = "parameter_160" + original_name = "batch_norm_9.b_0" shape = [8] dtype = "float32" min_val = float("0") @@ -1635,6 +1796,7 @@ class Program_weight_tensor_parameter_160: class Program_weight_tensor_parameter_161: name = "parameter_161" + original_name = "batch_norm_9.w_0" shape = [8] dtype = "float32" min_val = float("0") @@ -1644,6 +1806,7 @@ class Program_weight_tensor_parameter_161: class Program_weight_tensor_parameter_162: name = "parameter_162" + original_name = "batch_norm_9.w_2" shape = [8] dtype = "float32" min_val = float("0") @@ -1653,6 +1816,7 @@ class Program_weight_tensor_parameter_162: class Program_weight_tensor_parameter_163: name = "parameter_163" + original_name = "batch_norm_9.w_1" shape = [8] dtype = "float32" min_val = float("0") @@ -1662,6 +1826,7 @@ class Program_weight_tensor_parameter_163: class Program_weight_tensor_parameter_164: name = "parameter_164" + original_name = "conv2d_11.w_0" shape = [8, 32, 1, 1] dtype = "float32" min_val = float("-3.06071") @@ -1673,6 +1838,7 @@ class Program_weight_tensor_parameter_164: class Program_weight_tensor_parameter_165: name = "parameter_165" + original_name = "batch_norm_8.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -1682,6 +1848,7 @@ class Program_weight_tensor_parameter_165: class Program_weight_tensor_parameter_166: name = "parameter_166" + original_name = "batch_norm_8.w_0" shape = [32] dtype = "float32" min_val = float("0") @@ -1691,6 +1858,7 @@ class Program_weight_tensor_parameter_166: class Program_weight_tensor_parameter_167: name = "parameter_167" + original_name = "batch_norm_8.w_2" shape = [32] dtype = "float32" min_val = float("0") @@ -1700,6 +1868,7 @@ class Program_weight_tensor_parameter_167: class Program_weight_tensor_parameter_168: name = "parameter_168" + original_name = "batch_norm_8.w_1" shape = [32] dtype = "float32" min_val = float("0") @@ -1709,6 +1878,7 @@ class Program_weight_tensor_parameter_168: class Program_weight_tensor_parameter_169: name = "parameter_169" + original_name = "conv2d_10.w_0" shape = [32, 1, 3, 3] dtype = "float32" min_val = float("-4.69753") @@ -1720,6 +1890,7 @@ class Program_weight_tensor_parameter_169: class Program_weight_tensor_parameter_170: name = "parameter_170" + original_name = "batch_norm_7.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -1729,6 +1900,7 @@ class Program_weight_tensor_parameter_170: class Program_weight_tensor_parameter_171: name = "parameter_171" + original_name = "batch_norm_7.w_0" shape = [32] dtype = "float32" min_val = float("0") @@ -1738,6 +1910,7 @@ class Program_weight_tensor_parameter_171: class Program_weight_tensor_parameter_172: name = "parameter_172" + original_name = "batch_norm_7.w_2" shape = [32] dtype = "float32" min_val = float("0") @@ -1747,6 +1920,7 @@ class Program_weight_tensor_parameter_172: class Program_weight_tensor_parameter_173: name = "parameter_173" + original_name = "batch_norm_7.w_1" shape = [32] dtype = "float32" min_val = float("0") @@ -1756,6 +1930,7 @@ class Program_weight_tensor_parameter_173: class Program_weight_tensor_parameter_174: name = "parameter_174" + original_name = "conv2d_9.w_0" shape = [32, 8, 1, 1] dtype = "float32" min_val = float("-4.68863") @@ -1767,6 +1942,7 @@ class Program_weight_tensor_parameter_174: class Program_weight_tensor_parameter_175: name = "parameter_175" + original_name = "batch_norm_6.b_0" shape = [8] dtype = "float32" min_val = float("0") @@ -1776,6 +1952,7 @@ class Program_weight_tensor_parameter_175: class Program_weight_tensor_parameter_176: name = "parameter_176" + original_name = "batch_norm_6.w_0" shape = [8] dtype = "float32" min_val = float("0") @@ -1785,6 +1962,7 @@ class Program_weight_tensor_parameter_176: class Program_weight_tensor_parameter_177: name = "parameter_177" + original_name = "batch_norm_6.w_2" shape = [8] dtype = "float32" min_val = float("0") @@ -1794,6 +1972,7 @@ class Program_weight_tensor_parameter_177: class Program_weight_tensor_parameter_178: name = "parameter_178" + original_name = "batch_norm_6.w_1" shape = [8] dtype = "float32" min_val = float("0") @@ -1803,6 +1982,7 @@ class Program_weight_tensor_parameter_178: class Program_weight_tensor_parameter_179: name = "parameter_179" + original_name = "conv2d_8.w_0" shape = [8, 24, 1, 1] dtype = "float32" min_val = float("-6.00688") @@ -1814,6 +1994,7 @@ class Program_weight_tensor_parameter_179: class Program_weight_tensor_parameter_180: name = "parameter_180" + original_name = "batch_norm_5.b_0" shape = [24] dtype = "float32" min_val = float("0") @@ -1823,6 +2004,7 @@ class Program_weight_tensor_parameter_180: class Program_weight_tensor_parameter_181: name = "parameter_181" + original_name = "batch_norm_5.w_0" shape = [24] dtype = "float32" min_val = float("0") @@ -1832,6 +2014,7 @@ class Program_weight_tensor_parameter_181: class Program_weight_tensor_parameter_182: name = "parameter_182" + original_name = "batch_norm_5.w_2" shape = [24] dtype = "float32" min_val = float("0") @@ -1841,6 +2024,7 @@ class Program_weight_tensor_parameter_182: class Program_weight_tensor_parameter_183: name = "parameter_183" + original_name = "batch_norm_5.w_1" shape = [24] dtype = "float32" min_val = float("0") @@ -1850,6 +2034,7 @@ class Program_weight_tensor_parameter_183: class Program_weight_tensor_parameter_184: name = "parameter_184" + original_name = "conv2d_7.w_0" shape = [24, 1, 3, 3] dtype = "float32" min_val = float("-2.93928") @@ -1861,6 +2046,7 @@ class Program_weight_tensor_parameter_184: class Program_weight_tensor_parameter_185: name = "parameter_185" + original_name = "batch_norm_4.b_0" shape = [24] dtype = "float32" min_val = float("0") @@ -1870,6 +2056,7 @@ class Program_weight_tensor_parameter_185: class Program_weight_tensor_parameter_186: name = "parameter_186" + original_name = "batch_norm_4.w_0" shape = [24] dtype = "float32" min_val = float("0") @@ -1879,6 +2066,7 @@ class Program_weight_tensor_parameter_186: class Program_weight_tensor_parameter_187: name = "parameter_187" + original_name = "batch_norm_4.w_2" shape = [24] dtype = "float32" min_val = float("0") @@ -1888,6 +2076,7 @@ class Program_weight_tensor_parameter_187: class Program_weight_tensor_parameter_188: name = "parameter_188" + original_name = "batch_norm_4.w_1" shape = [24] dtype = "float32" min_val = float("0") @@ -1897,6 +2086,7 @@ class Program_weight_tensor_parameter_188: class Program_weight_tensor_parameter_189: name = "parameter_189" + original_name = "conv2d_6.w_0" shape = [24, 8, 1, 1] dtype = "float32" min_val = float("-3.36198") @@ -1908,6 +2098,7 @@ class Program_weight_tensor_parameter_189: class Program_weight_tensor_parameter_190: name = "parameter_190" + original_name = "batch_norm_3.b_0" shape = [8] dtype = "float32" min_val = float("0") @@ -1917,6 +2108,7 @@ class Program_weight_tensor_parameter_190: class Program_weight_tensor_parameter_191: name = "parameter_191" + original_name = "batch_norm_3.w_0" shape = [8] dtype = "float32" min_val = float("0") @@ -1926,6 +2118,7 @@ class Program_weight_tensor_parameter_191: class Program_weight_tensor_parameter_192: name = "parameter_192" + original_name = "batch_norm_3.w_2" shape = [8] dtype = "float32" min_val = float("0") @@ -1935,6 +2128,7 @@ class Program_weight_tensor_parameter_192: class Program_weight_tensor_parameter_193: name = "parameter_193" + original_name = "batch_norm_3.w_1" shape = [8] dtype = "float32" min_val = float("0") @@ -1944,6 +2138,7 @@ class Program_weight_tensor_parameter_193: class Program_weight_tensor_parameter_194: name = "parameter_194" + original_name = "conv2d_5.w_0" shape = [8, 8, 1, 1] dtype = "float32" min_val = float("0") @@ -1953,6 +2148,7 @@ class Program_weight_tensor_parameter_194: class Program_weight_tensor_parameter_195: name = "parameter_195" + original_name = "conv2d_4.b_0" shape = [8] dtype = "float32" min_val = float("0") @@ -1962,6 +2158,7 @@ class Program_weight_tensor_parameter_195: class Program_weight_tensor_parameter_196: name = "parameter_196" + original_name = "conv2d_4.w_0" shape = [8, 2, 1, 1] dtype = "float32" min_val = float("0") @@ -1971,6 +2168,7 @@ class Program_weight_tensor_parameter_196: class Program_weight_tensor_parameter_197: name = "parameter_197" + original_name = "conv2d_3.b_0" shape = [2] dtype = "float32" min_val = float("0") @@ -1980,6 +2178,7 @@ class Program_weight_tensor_parameter_197: class Program_weight_tensor_parameter_198: name = "parameter_198" + original_name = "conv2d_3.w_0" shape = [2, 8, 1, 1] dtype = "float32" min_val = float("0") @@ -1989,6 +2188,7 @@ class Program_weight_tensor_parameter_198: class Program_weight_tensor_parameter_199: name = "parameter_199" + original_name = "batch_norm_2.b_0" shape = [8] dtype = "float32" min_val = float("0") @@ -1998,6 +2198,7 @@ class Program_weight_tensor_parameter_199: class Program_weight_tensor_parameter_200: name = "parameter_200" + original_name = "batch_norm_2.w_0" shape = [8] dtype = "float32" min_val = float("0") @@ -2007,6 +2208,7 @@ class Program_weight_tensor_parameter_200: class Program_weight_tensor_parameter_201: name = "parameter_201" + original_name = "batch_norm_2.w_2" shape = [8] dtype = "float32" min_val = float("0") @@ -2016,6 +2218,7 @@ class Program_weight_tensor_parameter_201: class Program_weight_tensor_parameter_202: name = "parameter_202" + original_name = "batch_norm_2.w_1" shape = [8] dtype = "float32" min_val = float("0") @@ -2025,6 +2228,7 @@ class Program_weight_tensor_parameter_202: class Program_weight_tensor_parameter_203: name = "parameter_203" + original_name = "conv2d_2.w_0" shape = [8, 1, 3, 3] dtype = "float32" min_val = float("-2.31362") @@ -2036,6 +2240,7 @@ class Program_weight_tensor_parameter_203: class Program_weight_tensor_parameter_204: name = "parameter_204" + original_name = "batch_norm_1.b_0" shape = [8] dtype = "float32" min_val = float("0") @@ -2045,6 +2250,7 @@ class Program_weight_tensor_parameter_204: class Program_weight_tensor_parameter_205: name = "parameter_205" + original_name = "batch_norm_1.w_0" shape = [8] dtype = "float32" min_val = float("0") @@ -2054,6 +2260,7 @@ class Program_weight_tensor_parameter_205: class Program_weight_tensor_parameter_206: name = "parameter_206" + original_name = "batch_norm_1.w_2" shape = [8] dtype = "float32" min_val = float("0") @@ -2063,6 +2270,7 @@ class Program_weight_tensor_parameter_206: class Program_weight_tensor_parameter_207: name = "parameter_207" + original_name = "batch_norm_1.w_1" shape = [8] dtype = "float32" min_val = float("0") @@ -2072,6 +2280,7 @@ class Program_weight_tensor_parameter_207: class Program_weight_tensor_parameter_208: name = "parameter_208" + original_name = "conv2d_1.w_0" shape = [8, 8, 1, 1] dtype = "float32" min_val = float("0") @@ -2081,6 +2290,7 @@ class Program_weight_tensor_parameter_208: class Program_weight_tensor_parameter_209: name = "parameter_209" + original_name = "batch_norm_0.b_0" shape = [8] dtype = "float32" min_val = float("0") @@ -2090,6 +2300,7 @@ class Program_weight_tensor_parameter_209: class Program_weight_tensor_parameter_210: name = "parameter_210" + original_name = "batch_norm_0.w_0" shape = [8] dtype = "float32" min_val = float("0") @@ -2099,6 +2310,7 @@ class Program_weight_tensor_parameter_210: class Program_weight_tensor_parameter_211: name = "parameter_211" + original_name = "batch_norm_0.w_2" shape = [8] dtype = "float32" min_val = float("0") @@ -2108,6 +2320,7 @@ class Program_weight_tensor_parameter_211: class Program_weight_tensor_parameter_212: name = "parameter_212" + original_name = "batch_norm_0.w_1" shape = [8] dtype = "float32" min_val = float("0") @@ -2117,6 +2330,7 @@ class Program_weight_tensor_parameter_212: class Program_weight_tensor_parameter_213: name = "parameter_213" + original_name = "conv2d_0.w_0" shape = [8, 3, 3, 3] dtype = "float32" min_val = float("-3.08279") diff --git a/paddle_samples/PaddleX/MobileNetV3_small_x0_5/subgraph_0/input_meta.py b/paddle_samples/PaddleX/MobileNetV3_small_x0_5/subgraph_0/input_meta.py index 1c7927023..bcf17dbfb 100644 --- a/paddle_samples/PaddleX/MobileNetV3_small_x0_5/subgraph_0/input_meta.py +++ b/paddle_samples/PaddleX/MobileNetV3_small_x0_5/subgraph_0/input_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_data_0: name = "data_0" + original_name = "var_0" shape = [256, 3, 224, 224] dtype = "float32" min_val = float("-2.1179") diff --git a/paddle_samples/PaddleX/MobileNetV3_small_x0_5/subgraph_0/weight_meta.py b/paddle_samples/PaddleX/MobileNetV3_small_x0_5/subgraph_0/weight_meta.py index d70880968..0399b945d 100644 --- a/paddle_samples/PaddleX/MobileNetV3_small_x0_5/subgraph_0/weight_meta.py +++ b/paddle_samples/PaddleX/MobileNetV3_small_x0_5/subgraph_0/weight_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_parameter_0: name = "parameter_0" + original_name = "linear_0.b_0" shape = [102] dtype = "float32" data = None @@ -7,6 +8,7 @@ class Program_weight_tensor_parameter_0: class Program_weight_tensor_parameter_1: name = "parameter_1" + original_name = "linear_0.w_0" shape = [1280, 102] dtype = "float32" min_val = float("-0.0658891") @@ -18,6 +20,7 @@ class Program_weight_tensor_parameter_1: class Program_weight_tensor_parameter_2: name = "parameter_2" + original_name = "conv2d_53.w_0" shape = [1280, 288, 1, 1] dtype = "float32" min_val = float("-1.6974") @@ -29,6 +32,7 @@ class Program_weight_tensor_parameter_2: class Program_weight_tensor_parameter_3: name = "parameter_3" + original_name = "batch_norm_34.b_0" shape = [288] dtype = "float32" min_val = float("-10.5296") @@ -40,6 +44,7 @@ class Program_weight_tensor_parameter_3: class Program_weight_tensor_parameter_4: name = "parameter_4" + original_name = "batch_norm_34.w_0" shape = [288] dtype = "float32" min_val = float("1.90569") @@ -51,6 +56,7 @@ class Program_weight_tensor_parameter_4: class Program_weight_tensor_parameter_5: name = "parameter_5" + original_name = "batch_norm_34.w_2" shape = [288] dtype = "float32" min_val = float("125.187") @@ -62,6 +68,7 @@ class Program_weight_tensor_parameter_5: class Program_weight_tensor_parameter_6: name = "parameter_6" + original_name = "batch_norm_34.w_1" shape = [288] dtype = "float32" min_val = float("-1.68418e-05") @@ -73,6 +80,7 @@ class Program_weight_tensor_parameter_6: class Program_weight_tensor_parameter_7: name = "parameter_7" + original_name = "conv2d_52.w_0" shape = [288, 48, 1, 1] dtype = "float32" min_val = float("-1.14679") @@ -84,6 +92,7 @@ class Program_weight_tensor_parameter_7: class Program_weight_tensor_parameter_8: name = "parameter_8" + original_name = "batch_norm_33.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -93,6 +102,7 @@ class Program_weight_tensor_parameter_8: class Program_weight_tensor_parameter_9: name = "parameter_9" + original_name = "batch_norm_33.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -102,6 +112,7 @@ class Program_weight_tensor_parameter_9: class Program_weight_tensor_parameter_10: name = "parameter_10" + original_name = "batch_norm_33.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -111,6 +122,7 @@ class Program_weight_tensor_parameter_10: class Program_weight_tensor_parameter_11: name = "parameter_11" + original_name = "batch_norm_33.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -120,6 +132,7 @@ class Program_weight_tensor_parameter_11: class Program_weight_tensor_parameter_12: name = "parameter_12" + original_name = "conv2d_51.w_0" shape = [48, 288, 1, 1] dtype = "float32" min_val = float("-0.933478") @@ -131,6 +144,7 @@ class Program_weight_tensor_parameter_12: class Program_weight_tensor_parameter_13: name = "parameter_13" + original_name = "conv2d_50.b_0" shape = [288] dtype = "float32" min_val = float("-1.36382") @@ -142,6 +156,7 @@ class Program_weight_tensor_parameter_13: class Program_weight_tensor_parameter_14: name = "parameter_14" + original_name = "conv2d_50.w_0" shape = [288, 72, 1, 1] dtype = "float32" min_val = float("-1.40328") @@ -153,6 +168,7 @@ class Program_weight_tensor_parameter_14: class Program_weight_tensor_parameter_15: name = "parameter_15" + original_name = "conv2d_49.b_0" shape = [72] dtype = "float32" min_val = float("-0.27153") @@ -164,6 +180,7 @@ class Program_weight_tensor_parameter_15: class Program_weight_tensor_parameter_16: name = "parameter_16" + original_name = "conv2d_49.w_0" shape = [72, 288, 1, 1] dtype = "float32" min_val = float("-1.12863") @@ -175,6 +192,7 @@ class Program_weight_tensor_parameter_16: class Program_weight_tensor_parameter_17: name = "parameter_17" + original_name = "batch_norm_32.b_0" shape = [288] dtype = "float32" min_val = float("-4.32512") @@ -186,6 +204,7 @@ class Program_weight_tensor_parameter_17: class Program_weight_tensor_parameter_18: name = "parameter_18" + original_name = "batch_norm_32.w_0" shape = [288] dtype = "float32" min_val = float("0.310574") @@ -197,6 +216,7 @@ class Program_weight_tensor_parameter_18: class Program_weight_tensor_parameter_19: name = "parameter_19" + original_name = "batch_norm_32.w_2" shape = [288] dtype = "float32" min_val = float("0.000203857") @@ -208,6 +228,7 @@ class Program_weight_tensor_parameter_19: class Program_weight_tensor_parameter_20: name = "parameter_20" + original_name = "batch_norm_32.w_1" shape = [288] dtype = "float32" min_val = float("-7.21686") @@ -219,6 +240,7 @@ class Program_weight_tensor_parameter_20: class Program_weight_tensor_parameter_21: name = "parameter_21" + original_name = "conv2d_48.w_0" shape = [288, 1, 5, 5] dtype = "float32" min_val = float("-1.21039") @@ -230,6 +252,7 @@ class Program_weight_tensor_parameter_21: class Program_weight_tensor_parameter_22: name = "parameter_22" + original_name = "batch_norm_31.b_0" shape = [288] dtype = "float32" min_val = float("-2.63704") @@ -241,6 +264,7 @@ class Program_weight_tensor_parameter_22: class Program_weight_tensor_parameter_23: name = "parameter_23" + original_name = "batch_norm_31.w_0" shape = [288] dtype = "float32" min_val = float("-0.600559") @@ -252,6 +276,7 @@ class Program_weight_tensor_parameter_23: class Program_weight_tensor_parameter_24: name = "parameter_24" + original_name = "batch_norm_31.w_2" shape = [288] dtype = "float32" min_val = float("26.3714") @@ -263,6 +288,7 @@ class Program_weight_tensor_parameter_24: class Program_weight_tensor_parameter_25: name = "parameter_25" + original_name = "batch_norm_31.w_1" shape = [288] dtype = "float32" min_val = float("-8.58776e-06") @@ -274,6 +300,7 @@ class Program_weight_tensor_parameter_25: class Program_weight_tensor_parameter_26: name = "parameter_26" + original_name = "conv2d_47.w_0" shape = [288, 48, 1, 1] dtype = "float32" min_val = float("-0.840793") @@ -285,6 +312,7 @@ class Program_weight_tensor_parameter_26: class Program_weight_tensor_parameter_27: name = "parameter_27" + original_name = "batch_norm_30.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -294,6 +322,7 @@ class Program_weight_tensor_parameter_27: class Program_weight_tensor_parameter_28: name = "parameter_28" + original_name = "batch_norm_30.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -303,6 +332,7 @@ class Program_weight_tensor_parameter_28: class Program_weight_tensor_parameter_29: name = "parameter_29" + original_name = "batch_norm_30.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -312,6 +342,7 @@ class Program_weight_tensor_parameter_29: class Program_weight_tensor_parameter_30: name = "parameter_30" + original_name = "batch_norm_30.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -321,6 +352,7 @@ class Program_weight_tensor_parameter_30: class Program_weight_tensor_parameter_31: name = "parameter_31" + original_name = "conv2d_46.w_0" shape = [48, 288, 1, 1] dtype = "float32" min_val = float("-0.895955") @@ -332,6 +364,7 @@ class Program_weight_tensor_parameter_31: class Program_weight_tensor_parameter_32: name = "parameter_32" + original_name = "conv2d_45.b_0" shape = [288] dtype = "float32" min_val = float("-0.585305") @@ -343,6 +376,7 @@ class Program_weight_tensor_parameter_32: class Program_weight_tensor_parameter_33: name = "parameter_33" + original_name = "conv2d_45.w_0" shape = [288, 72, 1, 1] dtype = "float32" min_val = float("-0.927613") @@ -354,6 +388,7 @@ class Program_weight_tensor_parameter_33: class Program_weight_tensor_parameter_34: name = "parameter_34" + original_name = "conv2d_44.b_0" shape = [72] dtype = "float32" min_val = float("-0.143163") @@ -365,6 +400,7 @@ class Program_weight_tensor_parameter_34: class Program_weight_tensor_parameter_35: name = "parameter_35" + original_name = "conv2d_44.w_0" shape = [72, 288, 1, 1] dtype = "float32" min_val = float("-0.974921") @@ -376,6 +412,7 @@ class Program_weight_tensor_parameter_35: class Program_weight_tensor_parameter_36: name = "parameter_36" + original_name = "batch_norm_29.b_0" shape = [288] dtype = "float32" min_val = float("-2.16943") @@ -387,6 +424,7 @@ class Program_weight_tensor_parameter_36: class Program_weight_tensor_parameter_37: name = "parameter_37" + original_name = "batch_norm_29.w_0" shape = [288] dtype = "float32" min_val = float("0.254105") @@ -398,6 +436,7 @@ class Program_weight_tensor_parameter_37: class Program_weight_tensor_parameter_38: name = "parameter_38" + original_name = "batch_norm_29.w_2" shape = [288] dtype = "float32" min_val = float("0.000198955") @@ -409,6 +448,7 @@ class Program_weight_tensor_parameter_38: class Program_weight_tensor_parameter_39: name = "parameter_39" + original_name = "batch_norm_29.w_1" shape = [288] dtype = "float32" min_val = float("-6.84623") @@ -420,6 +460,7 @@ class Program_weight_tensor_parameter_39: class Program_weight_tensor_parameter_40: name = "parameter_40" + original_name = "conv2d_43.w_0" shape = [288, 1, 5, 5] dtype = "float32" min_val = float("-1.49186") @@ -431,6 +472,7 @@ class Program_weight_tensor_parameter_40: class Program_weight_tensor_parameter_41: name = "parameter_41" + original_name = "batch_norm_28.b_0" shape = [288] dtype = "float32" min_val = float("-2.64335") @@ -442,6 +484,7 @@ class Program_weight_tensor_parameter_41: class Program_weight_tensor_parameter_42: name = "parameter_42" + original_name = "batch_norm_28.w_0" shape = [288] dtype = "float32" min_val = float("-0.0684158") @@ -453,6 +496,7 @@ class Program_weight_tensor_parameter_42: class Program_weight_tensor_parameter_43: name = "parameter_43" + original_name = "batch_norm_28.w_2" shape = [288] dtype = "float32" min_val = float("18.3165") @@ -464,6 +508,7 @@ class Program_weight_tensor_parameter_43: class Program_weight_tensor_parameter_44: name = "parameter_44" + original_name = "batch_norm_28.w_1" shape = [288] dtype = "float32" min_val = float("-5.01497e-06") @@ -475,6 +520,7 @@ class Program_weight_tensor_parameter_44: class Program_weight_tensor_parameter_45: name = "parameter_45" + original_name = "conv2d_42.w_0" shape = [288, 48, 1, 1] dtype = "float32" min_val = float("-1.14101") @@ -486,6 +532,7 @@ class Program_weight_tensor_parameter_45: class Program_weight_tensor_parameter_46: name = "parameter_46" + original_name = "batch_norm_27.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -495,6 +542,7 @@ class Program_weight_tensor_parameter_46: class Program_weight_tensor_parameter_47: name = "parameter_47" + original_name = "batch_norm_27.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -504,6 +552,7 @@ class Program_weight_tensor_parameter_47: class Program_weight_tensor_parameter_48: name = "parameter_48" + original_name = "batch_norm_27.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -513,6 +562,7 @@ class Program_weight_tensor_parameter_48: class Program_weight_tensor_parameter_49: name = "parameter_49" + original_name = "batch_norm_27.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -522,6 +572,7 @@ class Program_weight_tensor_parameter_49: class Program_weight_tensor_parameter_50: name = "parameter_50" + original_name = "conv2d_41.w_0" shape = [48, 144, 1, 1] dtype = "float32" min_val = float("-1.37054") @@ -533,6 +584,7 @@ class Program_weight_tensor_parameter_50: class Program_weight_tensor_parameter_51: name = "parameter_51" + original_name = "conv2d_40.b_0" shape = [144] dtype = "float32" min_val = float("-0.510106") @@ -544,6 +596,7 @@ class Program_weight_tensor_parameter_51: class Program_weight_tensor_parameter_52: name = "parameter_52" + original_name = "conv2d_40.w_0" shape = [144, 36, 1, 1] dtype = "float32" min_val = float("-1.19322") @@ -555,6 +608,7 @@ class Program_weight_tensor_parameter_52: class Program_weight_tensor_parameter_53: name = "parameter_53" + original_name = "conv2d_39.b_0" shape = [36] dtype = "float32" min_val = float("0") @@ -564,6 +618,7 @@ class Program_weight_tensor_parameter_53: class Program_weight_tensor_parameter_54: name = "parameter_54" + original_name = "conv2d_39.w_0" shape = [36, 144, 1, 1] dtype = "float32" min_val = float("-1.3518") @@ -575,6 +630,7 @@ class Program_weight_tensor_parameter_54: class Program_weight_tensor_parameter_55: name = "parameter_55" + original_name = "batch_norm_26.b_0" shape = [144] dtype = "float32" min_val = float("-1.44607") @@ -586,6 +642,7 @@ class Program_weight_tensor_parameter_55: class Program_weight_tensor_parameter_56: name = "parameter_56" + original_name = "batch_norm_26.w_0" shape = [144] dtype = "float32" min_val = float("0.270351") @@ -597,6 +654,7 @@ class Program_weight_tensor_parameter_56: class Program_weight_tensor_parameter_57: name = "parameter_57" + original_name = "batch_norm_26.w_2" shape = [144] dtype = "float32" min_val = float("0.00205673") @@ -608,6 +666,7 @@ class Program_weight_tensor_parameter_57: class Program_weight_tensor_parameter_58: name = "parameter_58" + original_name = "batch_norm_26.w_1" shape = [144] dtype = "float32" min_val = float("-71.0695") @@ -619,6 +678,7 @@ class Program_weight_tensor_parameter_58: class Program_weight_tensor_parameter_59: name = "parameter_59" + original_name = "conv2d_38.w_0" shape = [144, 1, 5, 5] dtype = "float32" min_val = float("-0.610512") @@ -630,6 +690,7 @@ class Program_weight_tensor_parameter_59: class Program_weight_tensor_parameter_60: name = "parameter_60" + original_name = "batch_norm_25.b_0" shape = [144] dtype = "float32" min_val = float("-2.35964") @@ -641,6 +702,7 @@ class Program_weight_tensor_parameter_60: class Program_weight_tensor_parameter_61: name = "parameter_61" + original_name = "batch_norm_25.w_0" shape = [144] dtype = "float32" min_val = float("-0.0668238") @@ -652,6 +714,7 @@ class Program_weight_tensor_parameter_61: class Program_weight_tensor_parameter_62: name = "parameter_62" + original_name = "batch_norm_25.w_2" shape = [144] dtype = "float32" min_val = float("137.252") @@ -663,6 +726,7 @@ class Program_weight_tensor_parameter_62: class Program_weight_tensor_parameter_63: name = "parameter_63" + original_name = "batch_norm_25.w_1" shape = [144] dtype = "float32" min_val = float("-3.02961e-05") @@ -674,6 +738,7 @@ class Program_weight_tensor_parameter_63: class Program_weight_tensor_parameter_64: name = "parameter_64" + original_name = "conv2d_37.w_0" shape = [144, 24, 1, 1] dtype = "float32" min_val = float("-1.40953") @@ -685,6 +750,7 @@ class Program_weight_tensor_parameter_64: class Program_weight_tensor_parameter_65: name = "parameter_65" + original_name = "batch_norm_24.b_0" shape = [24] dtype = "float32" min_val = float("0") @@ -694,6 +760,7 @@ class Program_weight_tensor_parameter_65: class Program_weight_tensor_parameter_66: name = "parameter_66" + original_name = "batch_norm_24.w_0" shape = [24] dtype = "float32" min_val = float("0") @@ -703,6 +770,7 @@ class Program_weight_tensor_parameter_66: class Program_weight_tensor_parameter_67: name = "parameter_67" + original_name = "batch_norm_24.w_2" shape = [24] dtype = "float32" min_val = float("0") @@ -712,6 +780,7 @@ class Program_weight_tensor_parameter_67: class Program_weight_tensor_parameter_68: name = "parameter_68" + original_name = "batch_norm_24.w_1" shape = [24] dtype = "float32" min_val = float("0") @@ -721,6 +790,7 @@ class Program_weight_tensor_parameter_68: class Program_weight_tensor_parameter_69: name = "parameter_69" + original_name = "conv2d_36.w_0" shape = [24, 72, 1, 1] dtype = "float32" min_val = float("-1.32838") @@ -732,6 +802,7 @@ class Program_weight_tensor_parameter_69: class Program_weight_tensor_parameter_70: name = "parameter_70" + original_name = "conv2d_35.b_0" shape = [72] dtype = "float32" min_val = float("-0.637925") @@ -743,6 +814,7 @@ class Program_weight_tensor_parameter_70: class Program_weight_tensor_parameter_71: name = "parameter_71" + original_name = "conv2d_35.w_0" shape = [72, 18, 1, 1] dtype = "float32" min_val = float("-0.999096") @@ -754,6 +826,7 @@ class Program_weight_tensor_parameter_71: class Program_weight_tensor_parameter_72: name = "parameter_72" + original_name = "conv2d_34.b_0" shape = [18] dtype = "float32" min_val = float("0") @@ -763,6 +836,7 @@ class Program_weight_tensor_parameter_72: class Program_weight_tensor_parameter_73: name = "parameter_73" + original_name = "conv2d_34.w_0" shape = [18, 72, 1, 1] dtype = "float32" min_val = float("-1.651") @@ -774,6 +848,7 @@ class Program_weight_tensor_parameter_73: class Program_weight_tensor_parameter_74: name = "parameter_74" + original_name = "batch_norm_23.b_0" shape = [72] dtype = "float32" min_val = float("-2.85473") @@ -785,6 +860,7 @@ class Program_weight_tensor_parameter_74: class Program_weight_tensor_parameter_75: name = "parameter_75" + original_name = "batch_norm_23.w_0" shape = [72] dtype = "float32" min_val = float("0.377459") @@ -796,6 +872,7 @@ class Program_weight_tensor_parameter_75: class Program_weight_tensor_parameter_76: name = "parameter_76" + original_name = "batch_norm_23.w_2" shape = [72] dtype = "float32" min_val = float("0.000481612") @@ -807,6 +884,7 @@ class Program_weight_tensor_parameter_76: class Program_weight_tensor_parameter_77: name = "parameter_77" + original_name = "batch_norm_23.w_1" shape = [72] dtype = "float32" min_val = float("-8.27771") @@ -818,6 +896,7 @@ class Program_weight_tensor_parameter_77: class Program_weight_tensor_parameter_78: name = "parameter_78" + original_name = "conv2d_33.w_0" shape = [72, 1, 5, 5] dtype = "float32" min_val = float("-1.21258") @@ -829,6 +908,7 @@ class Program_weight_tensor_parameter_78: class Program_weight_tensor_parameter_79: name = "parameter_79" + original_name = "batch_norm_22.b_0" shape = [72] dtype = "float32" min_val = float("-1.86341") @@ -840,6 +920,7 @@ class Program_weight_tensor_parameter_79: class Program_weight_tensor_parameter_80: name = "parameter_80" + original_name = "batch_norm_22.w_0" shape = [72] dtype = "float32" min_val = float("0.0441213") @@ -851,6 +932,7 @@ class Program_weight_tensor_parameter_80: class Program_weight_tensor_parameter_81: name = "parameter_81" + original_name = "batch_norm_22.w_2" shape = [72] dtype = "float32" min_val = float("74.5441") @@ -862,6 +944,7 @@ class Program_weight_tensor_parameter_81: class Program_weight_tensor_parameter_82: name = "parameter_82" + original_name = "batch_norm_22.w_1" shape = [72] dtype = "float32" min_val = float("-2.87214e-05") @@ -873,6 +956,7 @@ class Program_weight_tensor_parameter_82: class Program_weight_tensor_parameter_83: name = "parameter_83" + original_name = "conv2d_32.w_0" shape = [72, 24, 1, 1] dtype = "float32" min_val = float("-1.19446") @@ -884,6 +968,7 @@ class Program_weight_tensor_parameter_83: class Program_weight_tensor_parameter_84: name = "parameter_84" + original_name = "batch_norm_21.b_0" shape = [24] dtype = "float32" min_val = float("0") @@ -893,6 +978,7 @@ class Program_weight_tensor_parameter_84: class Program_weight_tensor_parameter_85: name = "parameter_85" + original_name = "batch_norm_21.w_0" shape = [24] dtype = "float32" min_val = float("0") @@ -902,6 +988,7 @@ class Program_weight_tensor_parameter_85: class Program_weight_tensor_parameter_86: name = "parameter_86" + original_name = "batch_norm_21.w_2" shape = [24] dtype = "float32" min_val = float("0") @@ -911,6 +998,7 @@ class Program_weight_tensor_parameter_86: class Program_weight_tensor_parameter_87: name = "parameter_87" + original_name = "batch_norm_21.w_1" shape = [24] dtype = "float32" min_val = float("0") @@ -920,6 +1008,7 @@ class Program_weight_tensor_parameter_87: class Program_weight_tensor_parameter_88: name = "parameter_88" + original_name = "conv2d_31.w_0" shape = [24, 64, 1, 1] dtype = "float32" min_val = float("-1.70071") @@ -931,6 +1020,7 @@ class Program_weight_tensor_parameter_88: class Program_weight_tensor_parameter_89: name = "parameter_89" + original_name = "conv2d_30.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -940,6 +1030,7 @@ class Program_weight_tensor_parameter_89: class Program_weight_tensor_parameter_90: name = "parameter_90" + original_name = "conv2d_30.w_0" shape = [64, 16, 1, 1] dtype = "float32" min_val = float("-0.942652") @@ -951,6 +1042,7 @@ class Program_weight_tensor_parameter_90: class Program_weight_tensor_parameter_91: name = "parameter_91" + original_name = "conv2d_29.b_0" shape = [16] dtype = "float32" min_val = float("0") @@ -960,6 +1052,7 @@ class Program_weight_tensor_parameter_91: class Program_weight_tensor_parameter_92: name = "parameter_92" + original_name = "conv2d_29.w_0" shape = [16, 64, 1, 1] dtype = "float32" min_val = float("-1.56198") @@ -971,6 +1064,7 @@ class Program_weight_tensor_parameter_92: class Program_weight_tensor_parameter_93: name = "parameter_93" + original_name = "batch_norm_20.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -980,6 +1074,7 @@ class Program_weight_tensor_parameter_93: class Program_weight_tensor_parameter_94: name = "parameter_94" + original_name = "batch_norm_20.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -989,6 +1084,7 @@ class Program_weight_tensor_parameter_94: class Program_weight_tensor_parameter_95: name = "parameter_95" + original_name = "batch_norm_20.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -998,6 +1094,7 @@ class Program_weight_tensor_parameter_95: class Program_weight_tensor_parameter_96: name = "parameter_96" + original_name = "batch_norm_20.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -1007,6 +1104,7 @@ class Program_weight_tensor_parameter_96: class Program_weight_tensor_parameter_97: name = "parameter_97" + original_name = "conv2d_28.w_0" shape = [64, 1, 5, 5] dtype = "float32" min_val = float("-1.68602") @@ -1018,6 +1116,7 @@ class Program_weight_tensor_parameter_97: class Program_weight_tensor_parameter_98: name = "parameter_98" + original_name = "batch_norm_19.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -1027,6 +1126,7 @@ class Program_weight_tensor_parameter_98: class Program_weight_tensor_parameter_99: name = "parameter_99" + original_name = "batch_norm_19.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -1036,6 +1136,7 @@ class Program_weight_tensor_parameter_99: class Program_weight_tensor_parameter_100: name = "parameter_100" + original_name = "batch_norm_19.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -1045,6 +1146,7 @@ class Program_weight_tensor_parameter_100: class Program_weight_tensor_parameter_101: name = "parameter_101" + original_name = "batch_norm_19.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -1054,6 +1156,7 @@ class Program_weight_tensor_parameter_101: class Program_weight_tensor_parameter_102: name = "parameter_102" + original_name = "conv2d_27.w_0" shape = [64, 24, 1, 1] dtype = "float32" min_val = float("-1.05643") @@ -1065,6 +1168,7 @@ class Program_weight_tensor_parameter_102: class Program_weight_tensor_parameter_103: name = "parameter_103" + original_name = "batch_norm_18.b_0" shape = [24] dtype = "float32" min_val = float("0") @@ -1074,6 +1178,7 @@ class Program_weight_tensor_parameter_103: class Program_weight_tensor_parameter_104: name = "parameter_104" + original_name = "batch_norm_18.w_0" shape = [24] dtype = "float32" min_val = float("0") @@ -1083,6 +1188,7 @@ class Program_weight_tensor_parameter_104: class Program_weight_tensor_parameter_105: name = "parameter_105" + original_name = "batch_norm_18.w_2" shape = [24] dtype = "float32" min_val = float("0") @@ -1092,6 +1198,7 @@ class Program_weight_tensor_parameter_105: class Program_weight_tensor_parameter_106: name = "parameter_106" + original_name = "batch_norm_18.w_1" shape = [24] dtype = "float32" min_val = float("0") @@ -1101,6 +1208,7 @@ class Program_weight_tensor_parameter_106: class Program_weight_tensor_parameter_107: name = "parameter_107" + original_name = "conv2d_26.w_0" shape = [24, 120, 1, 1] dtype = "float32" min_val = float("-1.30649") @@ -1112,6 +1220,7 @@ class Program_weight_tensor_parameter_107: class Program_weight_tensor_parameter_108: name = "parameter_108" + original_name = "conv2d_25.b_0" shape = [120] dtype = "float32" min_val = float("-0.508291") @@ -1123,6 +1232,7 @@ class Program_weight_tensor_parameter_108: class Program_weight_tensor_parameter_109: name = "parameter_109" + original_name = "conv2d_25.w_0" shape = [120, 30, 1, 1] dtype = "float32" min_val = float("-1.00308") @@ -1134,6 +1244,7 @@ class Program_weight_tensor_parameter_109: class Program_weight_tensor_parameter_110: name = "parameter_110" + original_name = "conv2d_24.b_0" shape = [30] dtype = "float32" min_val = float("0") @@ -1143,6 +1254,7 @@ class Program_weight_tensor_parameter_110: class Program_weight_tensor_parameter_111: name = "parameter_111" + original_name = "conv2d_24.w_0" shape = [30, 120, 1, 1] dtype = "float32" min_val = float("-1.79959") @@ -1154,6 +1266,7 @@ class Program_weight_tensor_parameter_111: class Program_weight_tensor_parameter_112: name = "parameter_112" + original_name = "batch_norm_17.b_0" shape = [120] dtype = "float32" min_val = float("-2.20993") @@ -1165,6 +1278,7 @@ class Program_weight_tensor_parameter_112: class Program_weight_tensor_parameter_113: name = "parameter_113" + original_name = "batch_norm_17.w_0" shape = [120] dtype = "float32" min_val = float("0.262753") @@ -1176,6 +1290,7 @@ class Program_weight_tensor_parameter_113: class Program_weight_tensor_parameter_114: name = "parameter_114" + original_name = "batch_norm_17.w_2" shape = [120] dtype = "float32" min_val = float("0.00125888") @@ -1187,6 +1302,7 @@ class Program_weight_tensor_parameter_114: class Program_weight_tensor_parameter_115: name = "parameter_115" + original_name = "batch_norm_17.w_1" shape = [120] dtype = "float32" min_val = float("-10.0124") @@ -1198,6 +1314,7 @@ class Program_weight_tensor_parameter_115: class Program_weight_tensor_parameter_116: name = "parameter_116" + original_name = "conv2d_23.w_0" shape = [120, 1, 5, 5] dtype = "float32" min_val = float("-1.95545") @@ -1209,6 +1326,7 @@ class Program_weight_tensor_parameter_116: class Program_weight_tensor_parameter_117: name = "parameter_117" + original_name = "batch_norm_16.b_0" shape = [120] dtype = "float32" min_val = float("-2.43797") @@ -1220,6 +1338,7 @@ class Program_weight_tensor_parameter_117: class Program_weight_tensor_parameter_118: name = "parameter_118" + original_name = "batch_norm_16.w_0" shape = [120] dtype = "float32" min_val = float("-0.184583") @@ -1231,6 +1350,7 @@ class Program_weight_tensor_parameter_118: class Program_weight_tensor_parameter_119: name = "parameter_119" + original_name = "batch_norm_16.w_2" shape = [120] dtype = "float32" min_val = float("47.0463") @@ -1242,6 +1362,7 @@ class Program_weight_tensor_parameter_119: class Program_weight_tensor_parameter_120: name = "parameter_120" + original_name = "batch_norm_16.w_1" shape = [120] dtype = "float32" min_val = float("-1.5566e-05") @@ -1253,6 +1374,7 @@ class Program_weight_tensor_parameter_120: class Program_weight_tensor_parameter_121: name = "parameter_121" + original_name = "conv2d_22.w_0" shape = [120, 24, 1, 1] dtype = "float32" min_val = float("-1.30152") @@ -1264,6 +1386,7 @@ class Program_weight_tensor_parameter_121: class Program_weight_tensor_parameter_122: name = "parameter_122" + original_name = "batch_norm_15.b_0" shape = [24] dtype = "float32" min_val = float("0") @@ -1273,6 +1396,7 @@ class Program_weight_tensor_parameter_122: class Program_weight_tensor_parameter_123: name = "parameter_123" + original_name = "batch_norm_15.w_0" shape = [24] dtype = "float32" min_val = float("0") @@ -1282,6 +1406,7 @@ class Program_weight_tensor_parameter_123: class Program_weight_tensor_parameter_124: name = "parameter_124" + original_name = "batch_norm_15.w_2" shape = [24] dtype = "float32" min_val = float("0") @@ -1291,6 +1416,7 @@ class Program_weight_tensor_parameter_124: class Program_weight_tensor_parameter_125: name = "parameter_125" + original_name = "batch_norm_15.w_1" shape = [24] dtype = "float32" min_val = float("0") @@ -1300,6 +1426,7 @@ class Program_weight_tensor_parameter_125: class Program_weight_tensor_parameter_126: name = "parameter_126" + original_name = "conv2d_21.w_0" shape = [24, 120, 1, 1] dtype = "float32" min_val = float("-1.9089") @@ -1311,6 +1438,7 @@ class Program_weight_tensor_parameter_126: class Program_weight_tensor_parameter_127: name = "parameter_127" + original_name = "conv2d_20.b_0" shape = [120] dtype = "float32" min_val = float("-0.697517") @@ -1322,6 +1450,7 @@ class Program_weight_tensor_parameter_127: class Program_weight_tensor_parameter_128: name = "parameter_128" + original_name = "conv2d_20.w_0" shape = [120, 30, 1, 1] dtype = "float32" min_val = float("-1.02935") @@ -1333,6 +1462,7 @@ class Program_weight_tensor_parameter_128: class Program_weight_tensor_parameter_129: name = "parameter_129" + original_name = "conv2d_19.b_0" shape = [30] dtype = "float32" min_val = float("0") @@ -1342,6 +1472,7 @@ class Program_weight_tensor_parameter_129: class Program_weight_tensor_parameter_130: name = "parameter_130" + original_name = "conv2d_19.w_0" shape = [30, 120, 1, 1] dtype = "float32" min_val = float("-1.70665") @@ -1353,6 +1484,7 @@ class Program_weight_tensor_parameter_130: class Program_weight_tensor_parameter_131: name = "parameter_131" + original_name = "batch_norm_14.b_0" shape = [120] dtype = "float32" min_val = float("-2.15707") @@ -1364,6 +1496,7 @@ class Program_weight_tensor_parameter_131: class Program_weight_tensor_parameter_132: name = "parameter_132" + original_name = "batch_norm_14.w_0" shape = [120] dtype = "float32" min_val = float("0.274386") @@ -1375,6 +1508,7 @@ class Program_weight_tensor_parameter_132: class Program_weight_tensor_parameter_133: name = "parameter_133" + original_name = "batch_norm_14.w_2" shape = [120] dtype = "float32" min_val = float("0.00109957") @@ -1386,6 +1520,7 @@ class Program_weight_tensor_parameter_133: class Program_weight_tensor_parameter_134: name = "parameter_134" + original_name = "batch_norm_14.w_1" shape = [120] dtype = "float32" min_val = float("-38.5948") @@ -1397,6 +1532,7 @@ class Program_weight_tensor_parameter_134: class Program_weight_tensor_parameter_135: name = "parameter_135" + original_name = "conv2d_18.w_0" shape = [120, 1, 5, 5] dtype = "float32" min_val = float("-2.18487") @@ -1408,6 +1544,7 @@ class Program_weight_tensor_parameter_135: class Program_weight_tensor_parameter_136: name = "parameter_136" + original_name = "batch_norm_13.b_0" shape = [120] dtype = "float32" min_val = float("-2.17401") @@ -1419,6 +1556,7 @@ class Program_weight_tensor_parameter_136: class Program_weight_tensor_parameter_137: name = "parameter_137" + original_name = "batch_norm_13.w_0" shape = [120] dtype = "float32" min_val = float("-6.70388") @@ -1430,6 +1568,7 @@ class Program_weight_tensor_parameter_137: class Program_weight_tensor_parameter_138: name = "parameter_138" + original_name = "batch_norm_13.w_2" shape = [120] dtype = "float32" min_val = float("25.6074") @@ -1441,6 +1580,7 @@ class Program_weight_tensor_parameter_138: class Program_weight_tensor_parameter_139: name = "parameter_139" + original_name = "batch_norm_13.w_1" shape = [120] dtype = "float32" min_val = float("-1.39421e-05") @@ -1452,6 +1592,7 @@ class Program_weight_tensor_parameter_139: class Program_weight_tensor_parameter_140: name = "parameter_140" + original_name = "conv2d_17.w_0" shape = [120, 24, 1, 1] dtype = "float32" min_val = float("-1.82846") @@ -1463,6 +1604,7 @@ class Program_weight_tensor_parameter_140: class Program_weight_tensor_parameter_141: name = "parameter_141" + original_name = "batch_norm_12.b_0" shape = [24] dtype = "float32" min_val = float("0") @@ -1472,6 +1614,7 @@ class Program_weight_tensor_parameter_141: class Program_weight_tensor_parameter_142: name = "parameter_142" + original_name = "batch_norm_12.w_0" shape = [24] dtype = "float32" min_val = float("0") @@ -1481,6 +1624,7 @@ class Program_weight_tensor_parameter_142: class Program_weight_tensor_parameter_143: name = "parameter_143" + original_name = "batch_norm_12.w_2" shape = [24] dtype = "float32" min_val = float("0") @@ -1490,6 +1634,7 @@ class Program_weight_tensor_parameter_143: class Program_weight_tensor_parameter_144: name = "parameter_144" + original_name = "batch_norm_12.w_1" shape = [24] dtype = "float32" min_val = float("0") @@ -1499,6 +1644,7 @@ class Program_weight_tensor_parameter_144: class Program_weight_tensor_parameter_145: name = "parameter_145" + original_name = "conv2d_16.w_0" shape = [24, 48, 1, 1] dtype = "float32" min_val = float("-2.15534") @@ -1510,6 +1656,7 @@ class Program_weight_tensor_parameter_145: class Program_weight_tensor_parameter_146: name = "parameter_146" + original_name = "conv2d_15.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -1519,6 +1666,7 @@ class Program_weight_tensor_parameter_146: class Program_weight_tensor_parameter_147: name = "parameter_147" + original_name = "conv2d_15.w_0" shape = [48, 12, 1, 1] dtype = "float32" min_val = float("-1.12337e-05") @@ -1530,6 +1678,7 @@ class Program_weight_tensor_parameter_147: class Program_weight_tensor_parameter_148: name = "parameter_148" + original_name = "conv2d_14.b_0" shape = [12] dtype = "float32" min_val = float("0") @@ -1539,6 +1688,7 @@ class Program_weight_tensor_parameter_148: class Program_weight_tensor_parameter_149: name = "parameter_149" + original_name = "conv2d_14.w_0" shape = [12, 48, 1, 1] dtype = "float32" min_val = float("-4.26393e-06") @@ -1550,6 +1700,7 @@ class Program_weight_tensor_parameter_149: class Program_weight_tensor_parameter_150: name = "parameter_150" + original_name = "batch_norm_11.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -1559,6 +1710,7 @@ class Program_weight_tensor_parameter_150: class Program_weight_tensor_parameter_151: name = "parameter_151" + original_name = "batch_norm_11.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -1568,6 +1720,7 @@ class Program_weight_tensor_parameter_151: class Program_weight_tensor_parameter_152: name = "parameter_152" + original_name = "batch_norm_11.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -1577,6 +1730,7 @@ class Program_weight_tensor_parameter_152: class Program_weight_tensor_parameter_153: name = "parameter_153" + original_name = "batch_norm_11.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -1586,6 +1740,7 @@ class Program_weight_tensor_parameter_153: class Program_weight_tensor_parameter_154: name = "parameter_154" + original_name = "conv2d_13.w_0" shape = [48, 1, 5, 5] dtype = "float32" min_val = float("-1.10372") @@ -1597,6 +1752,7 @@ class Program_weight_tensor_parameter_154: class Program_weight_tensor_parameter_155: name = "parameter_155" + original_name = "batch_norm_10.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -1606,6 +1762,7 @@ class Program_weight_tensor_parameter_155: class Program_weight_tensor_parameter_156: name = "parameter_156" + original_name = "batch_norm_10.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -1615,6 +1772,7 @@ class Program_weight_tensor_parameter_156: class Program_weight_tensor_parameter_157: name = "parameter_157" + original_name = "batch_norm_10.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -1624,6 +1782,7 @@ class Program_weight_tensor_parameter_157: class Program_weight_tensor_parameter_158: name = "parameter_158" + original_name = "batch_norm_10.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -1633,6 +1792,7 @@ class Program_weight_tensor_parameter_158: class Program_weight_tensor_parameter_159: name = "parameter_159" + original_name = "conv2d_12.w_0" shape = [48, 16, 1, 1] dtype = "float32" min_val = float("-2.49708") @@ -1644,6 +1804,7 @@ class Program_weight_tensor_parameter_159: class Program_weight_tensor_parameter_160: name = "parameter_160" + original_name = "batch_norm_9.b_0" shape = [16] dtype = "float32" min_val = float("0") @@ -1653,6 +1814,7 @@ class Program_weight_tensor_parameter_160: class Program_weight_tensor_parameter_161: name = "parameter_161" + original_name = "batch_norm_9.w_0" shape = [16] dtype = "float32" min_val = float("0") @@ -1662,6 +1824,7 @@ class Program_weight_tensor_parameter_161: class Program_weight_tensor_parameter_162: name = "parameter_162" + original_name = "batch_norm_9.w_2" shape = [16] dtype = "float32" min_val = float("0") @@ -1671,6 +1834,7 @@ class Program_weight_tensor_parameter_162: class Program_weight_tensor_parameter_163: name = "parameter_163" + original_name = "batch_norm_9.w_1" shape = [16] dtype = "float32" min_val = float("0") @@ -1680,6 +1844,7 @@ class Program_weight_tensor_parameter_163: class Program_weight_tensor_parameter_164: name = "parameter_164" + original_name = "conv2d_11.w_0" shape = [16, 48, 1, 1] dtype = "float32" min_val = float("-1.85897") @@ -1691,6 +1856,7 @@ class Program_weight_tensor_parameter_164: class Program_weight_tensor_parameter_165: name = "parameter_165" + original_name = "batch_norm_8.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -1700,6 +1866,7 @@ class Program_weight_tensor_parameter_165: class Program_weight_tensor_parameter_166: name = "parameter_166" + original_name = "batch_norm_8.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -1709,6 +1876,7 @@ class Program_weight_tensor_parameter_166: class Program_weight_tensor_parameter_167: name = "parameter_167" + original_name = "batch_norm_8.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -1718,6 +1886,7 @@ class Program_weight_tensor_parameter_167: class Program_weight_tensor_parameter_168: name = "parameter_168" + original_name = "batch_norm_8.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -1727,6 +1896,7 @@ class Program_weight_tensor_parameter_168: class Program_weight_tensor_parameter_169: name = "parameter_169" + original_name = "conv2d_10.w_0" shape = [48, 1, 3, 3] dtype = "float32" min_val = float("-2.00754") @@ -1738,6 +1908,7 @@ class Program_weight_tensor_parameter_169: class Program_weight_tensor_parameter_170: name = "parameter_170" + original_name = "batch_norm_7.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -1747,6 +1918,7 @@ class Program_weight_tensor_parameter_170: class Program_weight_tensor_parameter_171: name = "parameter_171" + original_name = "batch_norm_7.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -1756,6 +1928,7 @@ class Program_weight_tensor_parameter_171: class Program_weight_tensor_parameter_172: name = "parameter_172" + original_name = "batch_norm_7.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -1765,6 +1938,7 @@ class Program_weight_tensor_parameter_172: class Program_weight_tensor_parameter_173: name = "parameter_173" + original_name = "batch_norm_7.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -1774,6 +1948,7 @@ class Program_weight_tensor_parameter_173: class Program_weight_tensor_parameter_174: name = "parameter_174" + original_name = "conv2d_9.w_0" shape = [48, 16, 1, 1] dtype = "float32" min_val = float("-1.96768") @@ -1785,6 +1960,7 @@ class Program_weight_tensor_parameter_174: class Program_weight_tensor_parameter_175: name = "parameter_175" + original_name = "batch_norm_6.b_0" shape = [16] dtype = "float32" min_val = float("0") @@ -1794,6 +1970,7 @@ class Program_weight_tensor_parameter_175: class Program_weight_tensor_parameter_176: name = "parameter_176" + original_name = "batch_norm_6.w_0" shape = [16] dtype = "float32" min_val = float("0") @@ -1803,6 +1980,7 @@ class Program_weight_tensor_parameter_176: class Program_weight_tensor_parameter_177: name = "parameter_177" + original_name = "batch_norm_6.w_2" shape = [16] dtype = "float32" min_val = float("0") @@ -1812,6 +1990,7 @@ class Program_weight_tensor_parameter_177: class Program_weight_tensor_parameter_178: name = "parameter_178" + original_name = "batch_norm_6.w_1" shape = [16] dtype = "float32" min_val = float("0") @@ -1821,6 +2000,7 @@ class Program_weight_tensor_parameter_178: class Program_weight_tensor_parameter_179: name = "parameter_179" + original_name = "conv2d_8.w_0" shape = [16, 40, 1, 1] dtype = "float32" min_val = float("-2.17898") @@ -1832,6 +2012,7 @@ class Program_weight_tensor_parameter_179: class Program_weight_tensor_parameter_180: name = "parameter_180" + original_name = "batch_norm_5.b_0" shape = [40] dtype = "float32" min_val = float("0") @@ -1841,6 +2022,7 @@ class Program_weight_tensor_parameter_180: class Program_weight_tensor_parameter_181: name = "parameter_181" + original_name = "batch_norm_5.w_0" shape = [40] dtype = "float32" min_val = float("0") @@ -1850,6 +2032,7 @@ class Program_weight_tensor_parameter_181: class Program_weight_tensor_parameter_182: name = "parameter_182" + original_name = "batch_norm_5.w_2" shape = [40] dtype = "float32" min_val = float("0") @@ -1859,6 +2042,7 @@ class Program_weight_tensor_parameter_182: class Program_weight_tensor_parameter_183: name = "parameter_183" + original_name = "batch_norm_5.w_1" shape = [40] dtype = "float32" min_val = float("0") @@ -1868,6 +2052,7 @@ class Program_weight_tensor_parameter_183: class Program_weight_tensor_parameter_184: name = "parameter_184" + original_name = "conv2d_7.w_0" shape = [40, 1, 3, 3] dtype = "float32" min_val = float("-2.61841") @@ -1879,6 +2064,7 @@ class Program_weight_tensor_parameter_184: class Program_weight_tensor_parameter_185: name = "parameter_185" + original_name = "batch_norm_4.b_0" shape = [40] dtype = "float32" min_val = float("0") @@ -1888,6 +2074,7 @@ class Program_weight_tensor_parameter_185: class Program_weight_tensor_parameter_186: name = "parameter_186" + original_name = "batch_norm_4.w_0" shape = [40] dtype = "float32" min_val = float("0") @@ -1897,6 +2084,7 @@ class Program_weight_tensor_parameter_186: class Program_weight_tensor_parameter_187: name = "parameter_187" + original_name = "batch_norm_4.w_2" shape = [40] dtype = "float32" min_val = float("0") @@ -1906,6 +2094,7 @@ class Program_weight_tensor_parameter_187: class Program_weight_tensor_parameter_188: name = "parameter_188" + original_name = "batch_norm_4.w_1" shape = [40] dtype = "float32" min_val = float("0") @@ -1915,6 +2104,7 @@ class Program_weight_tensor_parameter_188: class Program_weight_tensor_parameter_189: name = "parameter_189" + original_name = "conv2d_6.w_0" shape = [40, 8, 1, 1] dtype = "float32" min_val = float("-3.48694") @@ -1926,6 +2116,7 @@ class Program_weight_tensor_parameter_189: class Program_weight_tensor_parameter_190: name = "parameter_190" + original_name = "batch_norm_3.b_0" shape = [8] dtype = "float32" min_val = float("0") @@ -1935,6 +2126,7 @@ class Program_weight_tensor_parameter_190: class Program_weight_tensor_parameter_191: name = "parameter_191" + original_name = "batch_norm_3.w_0" shape = [8] dtype = "float32" min_val = float("0") @@ -1944,6 +2136,7 @@ class Program_weight_tensor_parameter_191: class Program_weight_tensor_parameter_192: name = "parameter_192" + original_name = "batch_norm_3.w_2" shape = [8] dtype = "float32" min_val = float("0") @@ -1953,6 +2146,7 @@ class Program_weight_tensor_parameter_192: class Program_weight_tensor_parameter_193: name = "parameter_193" + original_name = "batch_norm_3.w_1" shape = [8] dtype = "float32" min_val = float("0") @@ -1962,6 +2156,7 @@ class Program_weight_tensor_parameter_193: class Program_weight_tensor_parameter_194: name = "parameter_194" + original_name = "conv2d_5.w_0" shape = [8, 8, 1, 1] dtype = "float32" min_val = float("0") @@ -1971,6 +2166,7 @@ class Program_weight_tensor_parameter_194: class Program_weight_tensor_parameter_195: name = "parameter_195" + original_name = "conv2d_4.b_0" shape = [8] dtype = "float32" min_val = float("0") @@ -1980,6 +2176,7 @@ class Program_weight_tensor_parameter_195: class Program_weight_tensor_parameter_196: name = "parameter_196" + original_name = "conv2d_4.w_0" shape = [8, 2, 1, 1] dtype = "float32" min_val = float("0") @@ -1989,6 +2186,7 @@ class Program_weight_tensor_parameter_196: class Program_weight_tensor_parameter_197: name = "parameter_197" + original_name = "conv2d_3.b_0" shape = [2] dtype = "float32" min_val = float("0") @@ -1998,6 +2196,7 @@ class Program_weight_tensor_parameter_197: class Program_weight_tensor_parameter_198: name = "parameter_198" + original_name = "conv2d_3.w_0" shape = [2, 8, 1, 1] dtype = "float32" min_val = float("0") @@ -2007,6 +2206,7 @@ class Program_weight_tensor_parameter_198: class Program_weight_tensor_parameter_199: name = "parameter_199" + original_name = "batch_norm_2.b_0" shape = [8] dtype = "float32" min_val = float("0") @@ -2016,6 +2216,7 @@ class Program_weight_tensor_parameter_199: class Program_weight_tensor_parameter_200: name = "parameter_200" + original_name = "batch_norm_2.w_0" shape = [8] dtype = "float32" min_val = float("0") @@ -2025,6 +2226,7 @@ class Program_weight_tensor_parameter_200: class Program_weight_tensor_parameter_201: name = "parameter_201" + original_name = "batch_norm_2.w_2" shape = [8] dtype = "float32" min_val = float("0") @@ -2034,6 +2236,7 @@ class Program_weight_tensor_parameter_201: class Program_weight_tensor_parameter_202: name = "parameter_202" + original_name = "batch_norm_2.w_1" shape = [8] dtype = "float32" min_val = float("0") @@ -2043,6 +2246,7 @@ class Program_weight_tensor_parameter_202: class Program_weight_tensor_parameter_203: name = "parameter_203" + original_name = "conv2d_2.w_0" shape = [8, 1, 3, 3] dtype = "float32" min_val = float("-1.4107") @@ -2054,6 +2258,7 @@ class Program_weight_tensor_parameter_203: class Program_weight_tensor_parameter_204: name = "parameter_204" + original_name = "batch_norm_1.b_0" shape = [8] dtype = "float32" min_val = float("0") @@ -2063,6 +2268,7 @@ class Program_weight_tensor_parameter_204: class Program_weight_tensor_parameter_205: name = "parameter_205" + original_name = "batch_norm_1.w_0" shape = [8] dtype = "float32" min_val = float("0") @@ -2072,6 +2278,7 @@ class Program_weight_tensor_parameter_205: class Program_weight_tensor_parameter_206: name = "parameter_206" + original_name = "batch_norm_1.w_2" shape = [8] dtype = "float32" min_val = float("0") @@ -2081,6 +2288,7 @@ class Program_weight_tensor_parameter_206: class Program_weight_tensor_parameter_207: name = "parameter_207" + original_name = "batch_norm_1.w_1" shape = [8] dtype = "float32" min_val = float("0") @@ -2090,6 +2298,7 @@ class Program_weight_tensor_parameter_207: class Program_weight_tensor_parameter_208: name = "parameter_208" + original_name = "conv2d_1.w_0" shape = [8, 8, 1, 1] dtype = "float32" min_val = float("0") @@ -2099,6 +2308,7 @@ class Program_weight_tensor_parameter_208: class Program_weight_tensor_parameter_209: name = "parameter_209" + original_name = "batch_norm_0.b_0" shape = [8] dtype = "float32" min_val = float("0") @@ -2108,6 +2318,7 @@ class Program_weight_tensor_parameter_209: class Program_weight_tensor_parameter_210: name = "parameter_210" + original_name = "batch_norm_0.w_0" shape = [8] dtype = "float32" min_val = float("0") @@ -2117,6 +2328,7 @@ class Program_weight_tensor_parameter_210: class Program_weight_tensor_parameter_211: name = "parameter_211" + original_name = "batch_norm_0.w_2" shape = [8] dtype = "float32" min_val = float("0") @@ -2126,6 +2338,7 @@ class Program_weight_tensor_parameter_211: class Program_weight_tensor_parameter_212: name = "parameter_212" + original_name = "batch_norm_0.w_1" shape = [8] dtype = "float32" min_val = float("0") @@ -2135,6 +2348,7 @@ class Program_weight_tensor_parameter_212: class Program_weight_tensor_parameter_213: name = "parameter_213" + original_name = "conv2d_0.w_0" shape = [8, 3, 3, 3] dtype = "float32" min_val = float("-6.0502") diff --git a/paddle_samples/PaddleX/MobileNetV3_small_x0_75/subgraph_0/input_meta.py b/paddle_samples/PaddleX/MobileNetV3_small_x0_75/subgraph_0/input_meta.py index 4c1d87b63..459245741 100644 --- a/paddle_samples/PaddleX/MobileNetV3_small_x0_75/subgraph_0/input_meta.py +++ b/paddle_samples/PaddleX/MobileNetV3_small_x0_75/subgraph_0/input_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_data_0: name = "data_0" + original_name = "var_161" shape = [252, 3, 224, 224] dtype = "float32" min_val = float("-2.1179") diff --git a/paddle_samples/PaddleX/MobileNetV3_small_x0_75/subgraph_0/weight_meta.py b/paddle_samples/PaddleX/MobileNetV3_small_x0_75/subgraph_0/weight_meta.py index 209c1405c..90006d51f 100644 --- a/paddle_samples/PaddleX/MobileNetV3_small_x0_75/subgraph_0/weight_meta.py +++ b/paddle_samples/PaddleX/MobileNetV3_small_x0_75/subgraph_0/weight_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_parameter_0: name = "parameter_0" + original_name = "linear_0.b_0" shape = [102] dtype = "float32" min_val = float("-0.00061221") @@ -11,6 +12,7 @@ class Program_weight_tensor_parameter_0: class Program_weight_tensor_parameter_1: name = "parameter_1" + original_name = "linear_0.w_0" shape = [1280, 102] dtype = "float32" min_val = float("-0.0661416") @@ -22,6 +24,7 @@ class Program_weight_tensor_parameter_1: class Program_weight_tensor_parameter_2: name = "parameter_2" + original_name = "conv2d_53.w_0" shape = [1280, 432, 1, 1] dtype = "float32" min_val = float("-0.630878") @@ -33,6 +36,7 @@ class Program_weight_tensor_parameter_2: class Program_weight_tensor_parameter_3: name = "parameter_3" + original_name = "batch_norm_34.b_0" shape = [432] dtype = "float32" min_val = float("-12.9619") @@ -44,6 +48,7 @@ class Program_weight_tensor_parameter_3: class Program_weight_tensor_parameter_4: name = "parameter_4" + original_name = "batch_norm_34.w_0" shape = [432] dtype = "float32" min_val = float("2.59718") @@ -55,6 +60,7 @@ class Program_weight_tensor_parameter_4: class Program_weight_tensor_parameter_5: name = "parameter_5" + original_name = "batch_norm_34.w_2" shape = [432] dtype = "float32" min_val = float("40.4058") @@ -66,6 +72,7 @@ class Program_weight_tensor_parameter_5: class Program_weight_tensor_parameter_6: name = "parameter_6" + original_name = "batch_norm_34.w_1" shape = [432] dtype = "float32" min_val = float("-1.51113e-05") @@ -77,6 +84,7 @@ class Program_weight_tensor_parameter_6: class Program_weight_tensor_parameter_7: name = "parameter_7" + original_name = "conv2d_52.w_0" shape = [432, 72, 1, 1] dtype = "float32" min_val = float("-0.939511") @@ -88,6 +96,7 @@ class Program_weight_tensor_parameter_7: class Program_weight_tensor_parameter_8: name = "parameter_8" + original_name = "batch_norm_33.b_0" shape = [72] dtype = "float32" min_val = float("-2.50928e-06") @@ -99,6 +108,7 @@ class Program_weight_tensor_parameter_8: class Program_weight_tensor_parameter_9: name = "parameter_9" + original_name = "batch_norm_33.w_0" shape = [72] dtype = "float32" min_val = float("2.38866") @@ -110,6 +120,7 @@ class Program_weight_tensor_parameter_9: class Program_weight_tensor_parameter_10: name = "parameter_10" + original_name = "batch_norm_33.w_2" shape = [72] dtype = "float32" min_val = float("0.0858519") @@ -121,6 +132,7 @@ class Program_weight_tensor_parameter_10: class Program_weight_tensor_parameter_11: name = "parameter_11" + original_name = "batch_norm_33.w_1" shape = [72] dtype = "float32" min_val = float("-0.651469") @@ -132,6 +144,7 @@ class Program_weight_tensor_parameter_11: class Program_weight_tensor_parameter_12: name = "parameter_12" + original_name = "conv2d_51.w_0" shape = [72, 432, 1, 1] dtype = "float32" min_val = float("-0.62298") @@ -143,6 +156,7 @@ class Program_weight_tensor_parameter_12: class Program_weight_tensor_parameter_13: name = "parameter_13" + original_name = "conv2d_50.b_0" shape = [432] dtype = "float32" min_val = float("-0.654824") @@ -154,6 +168,7 @@ class Program_weight_tensor_parameter_13: class Program_weight_tensor_parameter_14: name = "parameter_14" + original_name = "conv2d_50.w_0" shape = [432, 108, 1, 1] dtype = "float32" min_val = float("-0.790912") @@ -165,6 +180,7 @@ class Program_weight_tensor_parameter_14: class Program_weight_tensor_parameter_15: name = "parameter_15" + original_name = "conv2d_49.b_0" shape = [108] dtype = "float32" min_val = float("-0.0875705") @@ -176,6 +192,7 @@ class Program_weight_tensor_parameter_15: class Program_weight_tensor_parameter_16: name = "parameter_16" + original_name = "conv2d_49.w_0" shape = [108, 432, 1, 1] dtype = "float32" min_val = float("-0.541094") @@ -187,6 +204,7 @@ class Program_weight_tensor_parameter_16: class Program_weight_tensor_parameter_17: name = "parameter_17" + original_name = "batch_norm_32.b_0" shape = [432] dtype = "float32" min_val = float("-4.56081") @@ -198,6 +216,7 @@ class Program_weight_tensor_parameter_17: class Program_weight_tensor_parameter_18: name = "parameter_18" + original_name = "batch_norm_32.w_0" shape = [432] dtype = "float32" min_val = float("0.193385") @@ -209,6 +228,7 @@ class Program_weight_tensor_parameter_18: class Program_weight_tensor_parameter_19: name = "parameter_19" + original_name = "batch_norm_32.w_2" shape = [432] dtype = "float32" min_val = float("0.000107906") @@ -220,6 +240,7 @@ class Program_weight_tensor_parameter_19: class Program_weight_tensor_parameter_20: name = "parameter_20" + original_name = "batch_norm_32.w_1" shape = [432] dtype = "float32" min_val = float("-4.25679") @@ -231,6 +252,7 @@ class Program_weight_tensor_parameter_20: class Program_weight_tensor_parameter_21: name = "parameter_21" + original_name = "conv2d_48.w_0" shape = [432, 1, 5, 5] dtype = "float32" min_val = float("-0.775915") @@ -242,6 +264,7 @@ class Program_weight_tensor_parameter_21: class Program_weight_tensor_parameter_22: name = "parameter_22" + original_name = "batch_norm_31.b_0" shape = [432] dtype = "float32" min_val = float("-4.01881") @@ -253,6 +276,7 @@ class Program_weight_tensor_parameter_22: class Program_weight_tensor_parameter_23: name = "parameter_23" + original_name = "batch_norm_31.w_0" shape = [432] dtype = "float32" min_val = float("-1.3499") @@ -264,6 +288,7 @@ class Program_weight_tensor_parameter_23: class Program_weight_tensor_parameter_24: name = "parameter_24" + original_name = "batch_norm_31.w_2" shape = [432] dtype = "float32" min_val = float("9.968") @@ -275,6 +300,7 @@ class Program_weight_tensor_parameter_24: class Program_weight_tensor_parameter_25: name = "parameter_25" + original_name = "batch_norm_31.w_1" shape = [432] dtype = "float32" min_val = float("-9.38527e-06") @@ -286,6 +312,7 @@ class Program_weight_tensor_parameter_25: class Program_weight_tensor_parameter_26: name = "parameter_26" + original_name = "conv2d_47.w_0" shape = [432, 72, 1, 1] dtype = "float32" min_val = float("-0.546759") @@ -297,6 +324,7 @@ class Program_weight_tensor_parameter_26: class Program_weight_tensor_parameter_27: name = "parameter_27" + original_name = "batch_norm_30.b_0" shape = [72] dtype = "float32" min_val = float("-3.22649e-06") @@ -308,6 +336,7 @@ class Program_weight_tensor_parameter_27: class Program_weight_tensor_parameter_28: name = "parameter_28" + original_name = "batch_norm_30.w_0" shape = [72] dtype = "float32" min_val = float("1.17798") @@ -319,6 +348,7 @@ class Program_weight_tensor_parameter_28: class Program_weight_tensor_parameter_29: name = "parameter_29" + original_name = "batch_norm_30.w_2" shape = [72] dtype = "float32" min_val = float("0.0739239") @@ -330,6 +360,7 @@ class Program_weight_tensor_parameter_29: class Program_weight_tensor_parameter_30: name = "parameter_30" + original_name = "batch_norm_30.w_1" shape = [72] dtype = "float32" min_val = float("-0.577555") @@ -341,6 +372,7 @@ class Program_weight_tensor_parameter_30: class Program_weight_tensor_parameter_31: name = "parameter_31" + original_name = "conv2d_46.w_0" shape = [72, 432, 1, 1] dtype = "float32" min_val = float("-0.576954") @@ -352,6 +384,7 @@ class Program_weight_tensor_parameter_31: class Program_weight_tensor_parameter_32: name = "parameter_32" + original_name = "conv2d_45.b_0" shape = [432] dtype = "float32" min_val = float("-0.344611") @@ -363,6 +396,7 @@ class Program_weight_tensor_parameter_32: class Program_weight_tensor_parameter_33: name = "parameter_33" + original_name = "conv2d_45.w_0" shape = [432, 108, 1, 1] dtype = "float32" min_val = float("-0.616197") @@ -374,6 +408,7 @@ class Program_weight_tensor_parameter_33: class Program_weight_tensor_parameter_34: name = "parameter_34" + original_name = "conv2d_44.b_0" shape = [108] dtype = "float32" min_val = float("-0.0995143") @@ -385,6 +420,7 @@ class Program_weight_tensor_parameter_34: class Program_weight_tensor_parameter_35: name = "parameter_35" + original_name = "conv2d_44.w_0" shape = [108, 432, 1, 1] dtype = "float32" min_val = float("-0.841889") @@ -396,6 +432,7 @@ class Program_weight_tensor_parameter_35: class Program_weight_tensor_parameter_36: name = "parameter_36" + original_name = "batch_norm_29.b_0" shape = [432] dtype = "float32" min_val = float("-5.90644") @@ -407,6 +444,7 @@ class Program_weight_tensor_parameter_36: class Program_weight_tensor_parameter_37: name = "parameter_37" + original_name = "batch_norm_29.w_0" shape = [432] dtype = "float32" min_val = float("0.280337") @@ -418,6 +456,7 @@ class Program_weight_tensor_parameter_37: class Program_weight_tensor_parameter_38: name = "parameter_38" + original_name = "batch_norm_29.w_2" shape = [432] dtype = "float32" min_val = float("0.000130561") @@ -429,6 +468,7 @@ class Program_weight_tensor_parameter_38: class Program_weight_tensor_parameter_39: name = "parameter_39" + original_name = "batch_norm_29.w_1" shape = [432] dtype = "float32" min_val = float("-4.49545") @@ -440,6 +480,7 @@ class Program_weight_tensor_parameter_39: class Program_weight_tensor_parameter_40: name = "parameter_40" + original_name = "conv2d_43.w_0" shape = [432, 1, 5, 5] dtype = "float32" min_val = float("-0.874044") @@ -451,6 +492,7 @@ class Program_weight_tensor_parameter_40: class Program_weight_tensor_parameter_41: name = "parameter_41" + original_name = "batch_norm_28.b_0" shape = [432] dtype = "float32" min_val = float("-3.15393") @@ -462,6 +504,7 @@ class Program_weight_tensor_parameter_41: class Program_weight_tensor_parameter_42: name = "parameter_42" + original_name = "batch_norm_28.w_0" shape = [432] dtype = "float32" min_val = float("-0.868398") @@ -473,6 +516,7 @@ class Program_weight_tensor_parameter_42: class Program_weight_tensor_parameter_43: name = "parameter_43" + original_name = "batch_norm_28.w_2" shape = [432] dtype = "float32" min_val = float("6.68664") @@ -484,6 +528,7 @@ class Program_weight_tensor_parameter_43: class Program_weight_tensor_parameter_44: name = "parameter_44" + original_name = "batch_norm_28.w_1" shape = [432] dtype = "float32" min_val = float("-5.08751e-06") @@ -495,6 +540,7 @@ class Program_weight_tensor_parameter_44: class Program_weight_tensor_parameter_45: name = "parameter_45" + original_name = "conv2d_42.w_0" shape = [432, 72, 1, 1] dtype = "float32" min_val = float("-0.600019") @@ -506,6 +552,7 @@ class Program_weight_tensor_parameter_45: class Program_weight_tensor_parameter_46: name = "parameter_46" + original_name = "batch_norm_27.b_0" shape = [72] dtype = "float32" min_val = float("-3.29585e-06") @@ -517,6 +564,7 @@ class Program_weight_tensor_parameter_46: class Program_weight_tensor_parameter_47: name = "parameter_47" + original_name = "batch_norm_27.w_0" shape = [72] dtype = "float32" min_val = float("2.83293") @@ -528,6 +576,7 @@ class Program_weight_tensor_parameter_47: class Program_weight_tensor_parameter_48: name = "parameter_48" + original_name = "batch_norm_27.w_2" shape = [72] dtype = "float32" min_val = float("0.0957908") @@ -539,6 +588,7 @@ class Program_weight_tensor_parameter_48: class Program_weight_tensor_parameter_49: name = "parameter_49" + original_name = "batch_norm_27.w_1" shape = [72] dtype = "float32" min_val = float("-1.17811") @@ -550,6 +600,7 @@ class Program_weight_tensor_parameter_49: class Program_weight_tensor_parameter_50: name = "parameter_50" + original_name = "conv2d_41.w_0" shape = [72, 216, 1, 1] dtype = "float32" min_val = float("-0.921338") @@ -561,6 +612,7 @@ class Program_weight_tensor_parameter_50: class Program_weight_tensor_parameter_51: name = "parameter_51" + original_name = "conv2d_40.b_0" shape = [216] dtype = "float32" min_val = float("-0.428496") @@ -572,6 +624,7 @@ class Program_weight_tensor_parameter_51: class Program_weight_tensor_parameter_52: name = "parameter_52" + original_name = "conv2d_40.w_0" shape = [216, 54, 1, 1] dtype = "float32" min_val = float("-0.901557") @@ -583,6 +636,7 @@ class Program_weight_tensor_parameter_52: class Program_weight_tensor_parameter_53: name = "parameter_53" + original_name = "conv2d_39.b_0" shape = [54] dtype = "float32" min_val = float("0") @@ -592,6 +646,7 @@ class Program_weight_tensor_parameter_53: class Program_weight_tensor_parameter_54: name = "parameter_54" + original_name = "conv2d_39.w_0" shape = [54, 216, 1, 1] dtype = "float32" min_val = float("-0.654308") @@ -603,6 +658,7 @@ class Program_weight_tensor_parameter_54: class Program_weight_tensor_parameter_55: name = "parameter_55" + original_name = "batch_norm_26.b_0" shape = [216] dtype = "float32" min_val = float("-2.28304") @@ -614,6 +670,7 @@ class Program_weight_tensor_parameter_55: class Program_weight_tensor_parameter_56: name = "parameter_56" + original_name = "batch_norm_26.w_0" shape = [216] dtype = "float32" min_val = float("-0.824314") @@ -625,6 +682,7 @@ class Program_weight_tensor_parameter_56: class Program_weight_tensor_parameter_57: name = "parameter_57" + original_name = "batch_norm_26.w_2" shape = [216] dtype = "float32" min_val = float("0.000216886") @@ -636,6 +694,7 @@ class Program_weight_tensor_parameter_57: class Program_weight_tensor_parameter_58: name = "parameter_58" + original_name = "batch_norm_26.w_1" shape = [216] dtype = "float32" min_val = float("-13.7393") @@ -647,6 +706,7 @@ class Program_weight_tensor_parameter_58: class Program_weight_tensor_parameter_59: name = "parameter_59" + original_name = "conv2d_38.w_0" shape = [216, 1, 5, 5] dtype = "float32" min_val = float("-0.470519") @@ -658,6 +718,7 @@ class Program_weight_tensor_parameter_59: class Program_weight_tensor_parameter_60: name = "parameter_60" + original_name = "batch_norm_25.b_0" shape = [216] dtype = "float32" min_val = float("-2.43602") @@ -669,6 +730,7 @@ class Program_weight_tensor_parameter_60: class Program_weight_tensor_parameter_61: name = "parameter_61" + original_name = "batch_norm_25.w_0" shape = [216] dtype = "float32" min_val = float("-1.39866") @@ -680,6 +742,7 @@ class Program_weight_tensor_parameter_61: class Program_weight_tensor_parameter_62: name = "parameter_62" + original_name = "batch_norm_25.w_2" shape = [216] dtype = "float32" min_val = float("19.2104") @@ -691,6 +754,7 @@ class Program_weight_tensor_parameter_62: class Program_weight_tensor_parameter_63: name = "parameter_63" + original_name = "batch_norm_25.w_1" shape = [216] dtype = "float32" min_val = float("-1.05871e-05") @@ -702,6 +766,7 @@ class Program_weight_tensor_parameter_63: class Program_weight_tensor_parameter_64: name = "parameter_64" + original_name = "conv2d_37.w_0" shape = [216, 40, 1, 1] dtype = "float32" min_val = float("-0.809628") @@ -713,6 +778,7 @@ class Program_weight_tensor_parameter_64: class Program_weight_tensor_parameter_65: name = "parameter_65" + original_name = "batch_norm_24.b_0" shape = [40] dtype = "float32" min_val = float("0") @@ -722,6 +788,7 @@ class Program_weight_tensor_parameter_65: class Program_weight_tensor_parameter_66: name = "parameter_66" + original_name = "batch_norm_24.w_0" shape = [40] dtype = "float32" min_val = float("0") @@ -731,6 +798,7 @@ class Program_weight_tensor_parameter_66: class Program_weight_tensor_parameter_67: name = "parameter_67" + original_name = "batch_norm_24.w_2" shape = [40] dtype = "float32" min_val = float("0") @@ -740,6 +808,7 @@ class Program_weight_tensor_parameter_67: class Program_weight_tensor_parameter_68: name = "parameter_68" + original_name = "batch_norm_24.w_1" shape = [40] dtype = "float32" min_val = float("0") @@ -749,6 +818,7 @@ class Program_weight_tensor_parameter_68: class Program_weight_tensor_parameter_69: name = "parameter_69" + original_name = "conv2d_36.w_0" shape = [40, 112, 1, 1] dtype = "float32" min_val = float("-0.668136") @@ -760,6 +830,7 @@ class Program_weight_tensor_parameter_69: class Program_weight_tensor_parameter_70: name = "parameter_70" + original_name = "conv2d_35.b_0" shape = [112] dtype = "float32" min_val = float("-0.346814") @@ -771,6 +842,7 @@ class Program_weight_tensor_parameter_70: class Program_weight_tensor_parameter_71: name = "parameter_71" + original_name = "conv2d_35.w_0" shape = [112, 28, 1, 1] dtype = "float32" min_val = float("-0.67531") @@ -782,6 +854,7 @@ class Program_weight_tensor_parameter_71: class Program_weight_tensor_parameter_72: name = "parameter_72" + original_name = "conv2d_34.b_0" shape = [28] dtype = "float32" min_val = float("0") @@ -791,6 +864,7 @@ class Program_weight_tensor_parameter_72: class Program_weight_tensor_parameter_73: name = "parameter_73" + original_name = "conv2d_34.w_0" shape = [28, 112, 1, 1] dtype = "float32" min_val = float("-0.823942") @@ -802,6 +876,7 @@ class Program_weight_tensor_parameter_73: class Program_weight_tensor_parameter_74: name = "parameter_74" + original_name = "batch_norm_23.b_0" shape = [112] dtype = "float32" min_val = float("-4.33983") @@ -813,6 +888,7 @@ class Program_weight_tensor_parameter_74: class Program_weight_tensor_parameter_75: name = "parameter_75" + original_name = "batch_norm_23.w_0" shape = [112] dtype = "float32" min_val = float("0.296614") @@ -824,6 +900,7 @@ class Program_weight_tensor_parameter_75: class Program_weight_tensor_parameter_76: name = "parameter_76" + original_name = "batch_norm_23.w_2" shape = [112] dtype = "float32" min_val = float("0.000296525") @@ -835,6 +912,7 @@ class Program_weight_tensor_parameter_76: class Program_weight_tensor_parameter_77: name = "parameter_77" + original_name = "batch_norm_23.w_1" shape = [112] dtype = "float32" min_val = float("-3.93031") @@ -846,6 +924,7 @@ class Program_weight_tensor_parameter_77: class Program_weight_tensor_parameter_78: name = "parameter_78" + original_name = "conv2d_33.w_0" shape = [112, 1, 5, 5] dtype = "float32" min_val = float("-1.03791") @@ -857,6 +936,7 @@ class Program_weight_tensor_parameter_78: class Program_weight_tensor_parameter_79: name = "parameter_79" + original_name = "batch_norm_22.b_0" shape = [112] dtype = "float32" min_val = float("-2.98963") @@ -868,6 +948,7 @@ class Program_weight_tensor_parameter_79: class Program_weight_tensor_parameter_80: name = "parameter_80" + original_name = "batch_norm_22.w_0" shape = [112] dtype = "float32" min_val = float("-1.552") @@ -879,6 +960,7 @@ class Program_weight_tensor_parameter_80: class Program_weight_tensor_parameter_81: name = "parameter_81" + original_name = "batch_norm_22.w_2" shape = [112] dtype = "float32" min_val = float("13.699") @@ -890,6 +972,7 @@ class Program_weight_tensor_parameter_81: class Program_weight_tensor_parameter_82: name = "parameter_82" + original_name = "batch_norm_22.w_1" shape = [112] dtype = "float32" min_val = float("-3.60683e-06") @@ -901,6 +984,7 @@ class Program_weight_tensor_parameter_82: class Program_weight_tensor_parameter_83: name = "parameter_83" + original_name = "conv2d_32.w_0" shape = [112, 40, 1, 1] dtype = "float32" min_val = float("-0.828591") @@ -912,6 +996,7 @@ class Program_weight_tensor_parameter_83: class Program_weight_tensor_parameter_84: name = "parameter_84" + original_name = "batch_norm_21.b_0" shape = [40] dtype = "float32" min_val = float("0") @@ -921,6 +1006,7 @@ class Program_weight_tensor_parameter_84: class Program_weight_tensor_parameter_85: name = "parameter_85" + original_name = "batch_norm_21.w_0" shape = [40] dtype = "float32" min_val = float("0") @@ -930,6 +1016,7 @@ class Program_weight_tensor_parameter_85: class Program_weight_tensor_parameter_86: name = "parameter_86" + original_name = "batch_norm_21.w_2" shape = [40] dtype = "float32" min_val = float("0") @@ -939,6 +1026,7 @@ class Program_weight_tensor_parameter_86: class Program_weight_tensor_parameter_87: name = "parameter_87" + original_name = "batch_norm_21.w_1" shape = [40] dtype = "float32" min_val = float("0") @@ -948,6 +1036,7 @@ class Program_weight_tensor_parameter_87: class Program_weight_tensor_parameter_88: name = "parameter_88" + original_name = "conv2d_31.w_0" shape = [40, 88, 1, 1] dtype = "float32" min_val = float("-1.09306") @@ -959,6 +1048,7 @@ class Program_weight_tensor_parameter_88: class Program_weight_tensor_parameter_89: name = "parameter_89" + original_name = "conv2d_30.b_0" shape = [88] dtype = "float32" min_val = float("0.00445048") @@ -970,6 +1060,7 @@ class Program_weight_tensor_parameter_89: class Program_weight_tensor_parameter_90: name = "parameter_90" + original_name = "conv2d_30.w_0" shape = [88, 22, 1, 1] dtype = "float32" min_val = float("-3.74814e-12") @@ -981,6 +1072,7 @@ class Program_weight_tensor_parameter_90: class Program_weight_tensor_parameter_91: name = "parameter_91" + original_name = "conv2d_29.b_0" shape = [22] dtype = "float32" min_val = float("0") @@ -990,6 +1082,7 @@ class Program_weight_tensor_parameter_91: class Program_weight_tensor_parameter_92: name = "parameter_92" + original_name = "conv2d_29.w_0" shape = [22, 88, 1, 1] dtype = "float32" min_val = float("-0.00189723") @@ -1001,6 +1094,7 @@ class Program_weight_tensor_parameter_92: class Program_weight_tensor_parameter_93: name = "parameter_93" + original_name = "batch_norm_20.b_0" shape = [88] dtype = "float32" min_val = float("-1.4829") @@ -1012,6 +1106,7 @@ class Program_weight_tensor_parameter_93: class Program_weight_tensor_parameter_94: name = "parameter_94" + original_name = "batch_norm_20.w_0" shape = [88] dtype = "float32" min_val = float("-2.22526") @@ -1023,6 +1118,7 @@ class Program_weight_tensor_parameter_94: class Program_weight_tensor_parameter_95: name = "parameter_95" + original_name = "batch_norm_20.w_2" shape = [88] dtype = "float32" min_val = float("0.00595107") @@ -1034,6 +1130,7 @@ class Program_weight_tensor_parameter_95: class Program_weight_tensor_parameter_96: name = "parameter_96" + original_name = "batch_norm_20.w_1" shape = [88] dtype = "float32" min_val = float("-30.2279") @@ -1045,6 +1142,7 @@ class Program_weight_tensor_parameter_96: class Program_weight_tensor_parameter_97: name = "parameter_97" + original_name = "conv2d_28.w_0" shape = [88, 1, 5, 5] dtype = "float32" min_val = float("-1.83987") @@ -1056,6 +1154,7 @@ class Program_weight_tensor_parameter_97: class Program_weight_tensor_parameter_98: name = "parameter_98" + original_name = "batch_norm_19.b_0" shape = [88] dtype = "float32" min_val = float("-1.30172") @@ -1067,6 +1166,7 @@ class Program_weight_tensor_parameter_98: class Program_weight_tensor_parameter_99: name = "parameter_99" + original_name = "batch_norm_19.w_0" shape = [88] dtype = "float32" min_val = float("-0.964801") @@ -1078,6 +1178,7 @@ class Program_weight_tensor_parameter_99: class Program_weight_tensor_parameter_100: name = "parameter_100" + original_name = "batch_norm_19.w_2" shape = [88] dtype = "float32" min_val = float("35.8749") @@ -1089,6 +1190,7 @@ class Program_weight_tensor_parameter_100: class Program_weight_tensor_parameter_101: name = "parameter_101" + original_name = "batch_norm_19.w_1" shape = [88] dtype = "float32" min_val = float("-1.5043e-05") @@ -1100,6 +1202,7 @@ class Program_weight_tensor_parameter_101: class Program_weight_tensor_parameter_102: name = "parameter_102" + original_name = "conv2d_27.w_0" shape = [88, 32, 1, 1] dtype = "float32" min_val = float("-1.37431") @@ -1111,6 +1214,7 @@ class Program_weight_tensor_parameter_102: class Program_weight_tensor_parameter_103: name = "parameter_103" + original_name = "batch_norm_18.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -1120,6 +1224,7 @@ class Program_weight_tensor_parameter_103: class Program_weight_tensor_parameter_104: name = "parameter_104" + original_name = "batch_norm_18.w_0" shape = [32] dtype = "float32" min_val = float("0") @@ -1129,6 +1234,7 @@ class Program_weight_tensor_parameter_104: class Program_weight_tensor_parameter_105: name = "parameter_105" + original_name = "batch_norm_18.w_2" shape = [32] dtype = "float32" min_val = float("0") @@ -1138,6 +1244,7 @@ class Program_weight_tensor_parameter_105: class Program_weight_tensor_parameter_106: name = "parameter_106" + original_name = "batch_norm_18.w_1" shape = [32] dtype = "float32" min_val = float("0") @@ -1147,6 +1254,7 @@ class Program_weight_tensor_parameter_106: class Program_weight_tensor_parameter_107: name = "parameter_107" + original_name = "conv2d_26.w_0" shape = [32, 184, 1, 1] dtype = "float32" min_val = float("-0.848754") @@ -1158,6 +1266,7 @@ class Program_weight_tensor_parameter_107: class Program_weight_tensor_parameter_108: name = "parameter_108" + original_name = "conv2d_25.b_0" shape = [184] dtype = "float32" min_val = float("-0.311169") @@ -1169,6 +1278,7 @@ class Program_weight_tensor_parameter_108: class Program_weight_tensor_parameter_109: name = "parameter_109" + original_name = "conv2d_25.w_0" shape = [184, 46, 1, 1] dtype = "float32" min_val = float("-0.879188") @@ -1180,6 +1290,7 @@ class Program_weight_tensor_parameter_109: class Program_weight_tensor_parameter_110: name = "parameter_110" + original_name = "conv2d_24.b_0" shape = [46] dtype = "float32" min_val = float("0") @@ -1189,6 +1300,7 @@ class Program_weight_tensor_parameter_110: class Program_weight_tensor_parameter_111: name = "parameter_111" + original_name = "conv2d_24.w_0" shape = [46, 184, 1, 1] dtype = "float32" min_val = float("-0.985228") @@ -1200,6 +1312,7 @@ class Program_weight_tensor_parameter_111: class Program_weight_tensor_parameter_112: name = "parameter_112" + original_name = "batch_norm_17.b_0" shape = [184] dtype = "float32" min_val = float("-5.76922") @@ -1211,6 +1324,7 @@ class Program_weight_tensor_parameter_112: class Program_weight_tensor_parameter_113: name = "parameter_113" + original_name = "batch_norm_17.w_0" shape = [184] dtype = "float32" min_val = float("0.210366") @@ -1222,6 +1336,7 @@ class Program_weight_tensor_parameter_113: class Program_weight_tensor_parameter_114: name = "parameter_114" + original_name = "batch_norm_17.w_2" shape = [184] dtype = "float32" min_val = float("0.000133428") @@ -1233,6 +1348,7 @@ class Program_weight_tensor_parameter_114: class Program_weight_tensor_parameter_115: name = "parameter_115" + original_name = "batch_norm_17.w_1" shape = [184] dtype = "float32" min_val = float("-4.79156") @@ -1244,6 +1360,7 @@ class Program_weight_tensor_parameter_115: class Program_weight_tensor_parameter_116: name = "parameter_116" + original_name = "conv2d_23.w_0" shape = [184, 1, 5, 5] dtype = "float32" min_val = float("-0.943832") @@ -1255,6 +1372,7 @@ class Program_weight_tensor_parameter_116: class Program_weight_tensor_parameter_117: name = "parameter_117" + original_name = "batch_norm_16.b_0" shape = [184] dtype = "float32" min_val = float("-2.27438") @@ -1266,6 +1384,7 @@ class Program_weight_tensor_parameter_117: class Program_weight_tensor_parameter_118: name = "parameter_118" + original_name = "batch_norm_16.w_0" shape = [184] dtype = "float32" min_val = float("-1.10513") @@ -1277,6 +1396,7 @@ class Program_weight_tensor_parameter_118: class Program_weight_tensor_parameter_119: name = "parameter_119" + original_name = "batch_norm_16.w_2" shape = [184] dtype = "float32" min_val = float("14.7272") @@ -1288,6 +1408,7 @@ class Program_weight_tensor_parameter_119: class Program_weight_tensor_parameter_120: name = "parameter_120" + original_name = "batch_norm_16.w_1" shape = [184] dtype = "float32" min_val = float("-9.01963e-06") @@ -1299,6 +1420,7 @@ class Program_weight_tensor_parameter_120: class Program_weight_tensor_parameter_121: name = "parameter_121" + original_name = "conv2d_22.w_0" shape = [184, 32, 1, 1] dtype = "float32" min_val = float("-0.741532") @@ -1310,6 +1432,7 @@ class Program_weight_tensor_parameter_121: class Program_weight_tensor_parameter_122: name = "parameter_122" + original_name = "batch_norm_15.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -1319,6 +1442,7 @@ class Program_weight_tensor_parameter_122: class Program_weight_tensor_parameter_123: name = "parameter_123" + original_name = "batch_norm_15.w_0" shape = [32] dtype = "float32" min_val = float("0") @@ -1328,6 +1452,7 @@ class Program_weight_tensor_parameter_123: class Program_weight_tensor_parameter_124: name = "parameter_124" + original_name = "batch_norm_15.w_2" shape = [32] dtype = "float32" min_val = float("0") @@ -1337,6 +1462,7 @@ class Program_weight_tensor_parameter_124: class Program_weight_tensor_parameter_125: name = "parameter_125" + original_name = "batch_norm_15.w_1" shape = [32] dtype = "float32" min_val = float("0") @@ -1346,6 +1472,7 @@ class Program_weight_tensor_parameter_125: class Program_weight_tensor_parameter_126: name = "parameter_126" + original_name = "conv2d_21.w_0" shape = [32, 184, 1, 1] dtype = "float32" min_val = float("-0.751599") @@ -1357,6 +1484,7 @@ class Program_weight_tensor_parameter_126: class Program_weight_tensor_parameter_127: name = "parameter_127" + original_name = "conv2d_20.b_0" shape = [184] dtype = "float32" min_val = float("-0.309715") @@ -1368,6 +1496,7 @@ class Program_weight_tensor_parameter_127: class Program_weight_tensor_parameter_128: name = "parameter_128" + original_name = "conv2d_20.w_0" shape = [184, 46, 1, 1] dtype = "float32" min_val = float("-0.757533") @@ -1379,6 +1508,7 @@ class Program_weight_tensor_parameter_128: class Program_weight_tensor_parameter_129: name = "parameter_129" + original_name = "conv2d_19.b_0" shape = [46] dtype = "float32" min_val = float("0") @@ -1388,6 +1518,7 @@ class Program_weight_tensor_parameter_129: class Program_weight_tensor_parameter_130: name = "parameter_130" + original_name = "conv2d_19.w_0" shape = [46, 184, 1, 1] dtype = "float32" min_val = float("-1.2046") @@ -1399,6 +1530,7 @@ class Program_weight_tensor_parameter_130: class Program_weight_tensor_parameter_131: name = "parameter_131" + original_name = "batch_norm_14.b_0" shape = [184] dtype = "float32" min_val = float("-2.76734") @@ -1410,6 +1542,7 @@ class Program_weight_tensor_parameter_131: class Program_weight_tensor_parameter_132: name = "parameter_132" + original_name = "batch_norm_14.w_0" shape = [184] dtype = "float32" min_val = float("0.193493") @@ -1421,6 +1554,7 @@ class Program_weight_tensor_parameter_132: class Program_weight_tensor_parameter_133: name = "parameter_133" + original_name = "batch_norm_14.w_2" shape = [184] dtype = "float32" min_val = float("0.000160967") @@ -1432,6 +1566,7 @@ class Program_weight_tensor_parameter_133: class Program_weight_tensor_parameter_134: name = "parameter_134" + original_name = "batch_norm_14.w_1" shape = [184] dtype = "float32" min_val = float("-20.4374") @@ -1443,6 +1578,7 @@ class Program_weight_tensor_parameter_134: class Program_weight_tensor_parameter_135: name = "parameter_135" + original_name = "conv2d_18.w_0" shape = [184, 1, 5, 5] dtype = "float32" min_val = float("-1.27149") @@ -1454,6 +1590,7 @@ class Program_weight_tensor_parameter_135: class Program_weight_tensor_parameter_136: name = "parameter_136" + original_name = "batch_norm_13.b_0" shape = [184] dtype = "float32" min_val = float("-2.57174") @@ -1465,6 +1602,7 @@ class Program_weight_tensor_parameter_136: class Program_weight_tensor_parameter_137: name = "parameter_137" + original_name = "batch_norm_13.w_0" shape = [184] dtype = "float32" min_val = float("-0.137371") @@ -1476,6 +1614,7 @@ class Program_weight_tensor_parameter_137: class Program_weight_tensor_parameter_138: name = "parameter_138" + original_name = "batch_norm_13.w_2" shape = [184] dtype = "float32" min_val = float("9.69646") @@ -1487,6 +1626,7 @@ class Program_weight_tensor_parameter_138: class Program_weight_tensor_parameter_139: name = "parameter_139" + original_name = "batch_norm_13.w_1" shape = [184] dtype = "float32" min_val = float("-6.17536e-06") @@ -1498,6 +1638,7 @@ class Program_weight_tensor_parameter_139: class Program_weight_tensor_parameter_140: name = "parameter_140" + original_name = "conv2d_17.w_0" shape = [184, 32, 1, 1] dtype = "float32" min_val = float("-0.917385") @@ -1509,6 +1650,7 @@ class Program_weight_tensor_parameter_140: class Program_weight_tensor_parameter_141: name = "parameter_141" + original_name = "batch_norm_12.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -1518,6 +1660,7 @@ class Program_weight_tensor_parameter_141: class Program_weight_tensor_parameter_142: name = "parameter_142" + original_name = "batch_norm_12.w_0" shape = [32] dtype = "float32" min_val = float("0") @@ -1527,6 +1670,7 @@ class Program_weight_tensor_parameter_142: class Program_weight_tensor_parameter_143: name = "parameter_143" + original_name = "batch_norm_12.w_2" shape = [32] dtype = "float32" min_val = float("0") @@ -1536,6 +1680,7 @@ class Program_weight_tensor_parameter_143: class Program_weight_tensor_parameter_144: name = "parameter_144" + original_name = "batch_norm_12.w_1" shape = [32] dtype = "float32" min_val = float("0") @@ -1545,6 +1690,7 @@ class Program_weight_tensor_parameter_144: class Program_weight_tensor_parameter_145: name = "parameter_145" + original_name = "conv2d_16.w_0" shape = [32, 72, 1, 1] dtype = "float32" min_val = float("-1.29855") @@ -1556,6 +1702,7 @@ class Program_weight_tensor_parameter_145: class Program_weight_tensor_parameter_146: name = "parameter_146" + original_name = "conv2d_15.b_0" shape = [72] dtype = "float32" min_val = float("-1.08368") @@ -1567,6 +1714,7 @@ class Program_weight_tensor_parameter_146: class Program_weight_tensor_parameter_147: name = "parameter_147" + original_name = "conv2d_15.w_0" shape = [72, 18, 1, 1] dtype = "float32" min_val = float("-4.23336e-13") @@ -1578,6 +1726,7 @@ class Program_weight_tensor_parameter_147: class Program_weight_tensor_parameter_148: name = "parameter_148" + original_name = "conv2d_14.b_0" shape = [18] dtype = "float32" min_val = float("0") @@ -1587,6 +1736,7 @@ class Program_weight_tensor_parameter_148: class Program_weight_tensor_parameter_149: name = "parameter_149" + original_name = "conv2d_14.w_0" shape = [18, 72, 1, 1] dtype = "float32" min_val = float("-5.7522e-13") @@ -1598,6 +1748,7 @@ class Program_weight_tensor_parameter_149: class Program_weight_tensor_parameter_150: name = "parameter_150" + original_name = "batch_norm_11.b_0" shape = [72] dtype = "float32" min_val = float("-1.20414") @@ -1609,6 +1760,7 @@ class Program_weight_tensor_parameter_150: class Program_weight_tensor_parameter_151: name = "parameter_151" + original_name = "batch_norm_11.w_0" shape = [72] dtype = "float32" min_val = float("0.462512") @@ -1620,6 +1772,7 @@ class Program_weight_tensor_parameter_151: class Program_weight_tensor_parameter_152: name = "parameter_152" + original_name = "batch_norm_11.w_2" shape = [72] dtype = "float32" min_val = float("0.0171048") @@ -1631,6 +1784,7 @@ class Program_weight_tensor_parameter_152: class Program_weight_tensor_parameter_153: name = "parameter_153" + original_name = "batch_norm_11.w_1" shape = [72] dtype = "float32" min_val = float("-22.6788") @@ -1642,6 +1796,7 @@ class Program_weight_tensor_parameter_153: class Program_weight_tensor_parameter_154: name = "parameter_154" + original_name = "conv2d_13.w_0" shape = [72, 1, 5, 5] dtype = "float32" min_val = float("-0.652797") @@ -1653,6 +1808,7 @@ class Program_weight_tensor_parameter_154: class Program_weight_tensor_parameter_155: name = "parameter_155" + original_name = "batch_norm_10.b_0" shape = [72] dtype = "float32" min_val = float("-1.69181") @@ -1664,6 +1820,7 @@ class Program_weight_tensor_parameter_155: class Program_weight_tensor_parameter_156: name = "parameter_156" + original_name = "batch_norm_10.w_0" shape = [72] dtype = "float32" min_val = float("0.321626") @@ -1675,6 +1832,7 @@ class Program_weight_tensor_parameter_156: class Program_weight_tensor_parameter_157: name = "parameter_157" + original_name = "batch_norm_10.w_2" shape = [72] dtype = "float32" min_val = float("38.7628") @@ -1686,6 +1844,7 @@ class Program_weight_tensor_parameter_157: class Program_weight_tensor_parameter_158: name = "parameter_158" + original_name = "batch_norm_10.w_1" shape = [72] dtype = "float32" min_val = float("-3.25434e-05") @@ -1697,6 +1856,7 @@ class Program_weight_tensor_parameter_158: class Program_weight_tensor_parameter_159: name = "parameter_159" + original_name = "conv2d_12.w_0" shape = [72, 24, 1, 1] dtype = "float32" min_val = float("-1.49767") @@ -1708,6 +1868,7 @@ class Program_weight_tensor_parameter_159: class Program_weight_tensor_parameter_160: name = "parameter_160" + original_name = "batch_norm_9.b_0" shape = [24] dtype = "float32" min_val = float("0") @@ -1717,6 +1878,7 @@ class Program_weight_tensor_parameter_160: class Program_weight_tensor_parameter_161: name = "parameter_161" + original_name = "batch_norm_9.w_0" shape = [24] dtype = "float32" min_val = float("0") @@ -1726,6 +1888,7 @@ class Program_weight_tensor_parameter_161: class Program_weight_tensor_parameter_162: name = "parameter_162" + original_name = "batch_norm_9.w_2" shape = [24] dtype = "float32" min_val = float("0") @@ -1735,6 +1898,7 @@ class Program_weight_tensor_parameter_162: class Program_weight_tensor_parameter_163: name = "parameter_163" + original_name = "batch_norm_9.w_1" shape = [24] dtype = "float32" min_val = float("0") @@ -1744,6 +1908,7 @@ class Program_weight_tensor_parameter_163: class Program_weight_tensor_parameter_164: name = "parameter_164" + original_name = "conv2d_11.w_0" shape = [24, 64, 1, 1] dtype = "float32" min_val = float("-1.30064") @@ -1755,6 +1920,7 @@ class Program_weight_tensor_parameter_164: class Program_weight_tensor_parameter_165: name = "parameter_165" + original_name = "batch_norm_8.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -1764,6 +1930,7 @@ class Program_weight_tensor_parameter_165: class Program_weight_tensor_parameter_166: name = "parameter_166" + original_name = "batch_norm_8.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -1773,6 +1940,7 @@ class Program_weight_tensor_parameter_166: class Program_weight_tensor_parameter_167: name = "parameter_167" + original_name = "batch_norm_8.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -1782,6 +1950,7 @@ class Program_weight_tensor_parameter_167: class Program_weight_tensor_parameter_168: name = "parameter_168" + original_name = "batch_norm_8.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -1791,6 +1960,7 @@ class Program_weight_tensor_parameter_168: class Program_weight_tensor_parameter_169: name = "parameter_169" + original_name = "conv2d_10.w_0" shape = [64, 1, 3, 3] dtype = "float32" min_val = float("-1.50124") @@ -1802,6 +1972,7 @@ class Program_weight_tensor_parameter_169: class Program_weight_tensor_parameter_170: name = "parameter_170" + original_name = "batch_norm_7.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -1811,6 +1982,7 @@ class Program_weight_tensor_parameter_170: class Program_weight_tensor_parameter_171: name = "parameter_171" + original_name = "batch_norm_7.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -1820,6 +1992,7 @@ class Program_weight_tensor_parameter_171: class Program_weight_tensor_parameter_172: name = "parameter_172" + original_name = "batch_norm_7.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -1829,6 +2002,7 @@ class Program_weight_tensor_parameter_172: class Program_weight_tensor_parameter_173: name = "parameter_173" + original_name = "batch_norm_7.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -1838,6 +2012,7 @@ class Program_weight_tensor_parameter_173: class Program_weight_tensor_parameter_174: name = "parameter_174" + original_name = "conv2d_9.w_0" shape = [64, 24, 1, 1] dtype = "float32" min_val = float("-1.56376") @@ -1849,6 +2024,7 @@ class Program_weight_tensor_parameter_174: class Program_weight_tensor_parameter_175: name = "parameter_175" + original_name = "batch_norm_6.b_0" shape = [24] dtype = "float32" min_val = float("0") @@ -1858,6 +2034,7 @@ class Program_weight_tensor_parameter_175: class Program_weight_tensor_parameter_176: name = "parameter_176" + original_name = "batch_norm_6.w_0" shape = [24] dtype = "float32" min_val = float("0") @@ -1867,6 +2044,7 @@ class Program_weight_tensor_parameter_176: class Program_weight_tensor_parameter_177: name = "parameter_177" + original_name = "batch_norm_6.w_2" shape = [24] dtype = "float32" min_val = float("0") @@ -1876,6 +2054,7 @@ class Program_weight_tensor_parameter_177: class Program_weight_tensor_parameter_178: name = "parameter_178" + original_name = "batch_norm_6.w_1" shape = [24] dtype = "float32" min_val = float("0") @@ -1885,6 +2064,7 @@ class Program_weight_tensor_parameter_178: class Program_weight_tensor_parameter_179: name = "parameter_179" + original_name = "conv2d_8.w_0" shape = [24, 56, 1, 1] dtype = "float32" min_val = float("-1.28929") @@ -1896,6 +2076,7 @@ class Program_weight_tensor_parameter_179: class Program_weight_tensor_parameter_180: name = "parameter_180" + original_name = "batch_norm_5.b_0" shape = [56] dtype = "float32" min_val = float("0") @@ -1905,6 +2086,7 @@ class Program_weight_tensor_parameter_180: class Program_weight_tensor_parameter_181: name = "parameter_181" + original_name = "batch_norm_5.w_0" shape = [56] dtype = "float32" min_val = float("0") @@ -1914,6 +2096,7 @@ class Program_weight_tensor_parameter_181: class Program_weight_tensor_parameter_182: name = "parameter_182" + original_name = "batch_norm_5.w_2" shape = [56] dtype = "float32" min_val = float("0") @@ -1923,6 +2106,7 @@ class Program_weight_tensor_parameter_182: class Program_weight_tensor_parameter_183: name = "parameter_183" + original_name = "batch_norm_5.w_1" shape = [56] dtype = "float32" min_val = float("0") @@ -1932,6 +2116,7 @@ class Program_weight_tensor_parameter_183: class Program_weight_tensor_parameter_184: name = "parameter_184" + original_name = "conv2d_7.w_0" shape = [56, 1, 3, 3] dtype = "float32" min_val = float("-1.62851") @@ -1943,6 +2128,7 @@ class Program_weight_tensor_parameter_184: class Program_weight_tensor_parameter_185: name = "parameter_185" + original_name = "batch_norm_4.b_0" shape = [56] dtype = "float32" min_val = float("0") @@ -1952,6 +2138,7 @@ class Program_weight_tensor_parameter_185: class Program_weight_tensor_parameter_186: name = "parameter_186" + original_name = "batch_norm_4.w_0" shape = [56] dtype = "float32" min_val = float("0") @@ -1961,6 +2148,7 @@ class Program_weight_tensor_parameter_186: class Program_weight_tensor_parameter_187: name = "parameter_187" + original_name = "batch_norm_4.w_2" shape = [56] dtype = "float32" min_val = float("0") @@ -1970,6 +2158,7 @@ class Program_weight_tensor_parameter_187: class Program_weight_tensor_parameter_188: name = "parameter_188" + original_name = "batch_norm_4.w_1" shape = [56] dtype = "float32" min_val = float("0") @@ -1979,6 +2168,7 @@ class Program_weight_tensor_parameter_188: class Program_weight_tensor_parameter_189: name = "parameter_189" + original_name = "conv2d_6.w_0" shape = [56, 16, 1, 1] dtype = "float32" min_val = float("-1.68026") @@ -1990,6 +2180,7 @@ class Program_weight_tensor_parameter_189: class Program_weight_tensor_parameter_190: name = "parameter_190" + original_name = "batch_norm_3.b_0" shape = [16] dtype = "float32" min_val = float("0") @@ -1999,6 +2190,7 @@ class Program_weight_tensor_parameter_190: class Program_weight_tensor_parameter_191: name = "parameter_191" + original_name = "batch_norm_3.w_0" shape = [16] dtype = "float32" min_val = float("0") @@ -2008,6 +2200,7 @@ class Program_weight_tensor_parameter_191: class Program_weight_tensor_parameter_192: name = "parameter_192" + original_name = "batch_norm_3.w_2" shape = [16] dtype = "float32" min_val = float("0") @@ -2017,6 +2210,7 @@ class Program_weight_tensor_parameter_192: class Program_weight_tensor_parameter_193: name = "parameter_193" + original_name = "batch_norm_3.w_1" shape = [16] dtype = "float32" min_val = float("0") @@ -2026,6 +2220,7 @@ class Program_weight_tensor_parameter_193: class Program_weight_tensor_parameter_194: name = "parameter_194" + original_name = "conv2d_5.w_0" shape = [16, 16, 1, 1] dtype = "float32" min_val = float("-1.94951") @@ -2037,6 +2232,7 @@ class Program_weight_tensor_parameter_194: class Program_weight_tensor_parameter_195: name = "parameter_195" + original_name = "conv2d_4.b_0" shape = [16] dtype = "float32" min_val = float("0") @@ -2046,6 +2242,7 @@ class Program_weight_tensor_parameter_195: class Program_weight_tensor_parameter_196: name = "parameter_196" + original_name = "conv2d_4.w_0" shape = [16, 4, 1, 1] dtype = "float32" min_val = float("0") @@ -2055,6 +2252,7 @@ class Program_weight_tensor_parameter_196: class Program_weight_tensor_parameter_197: name = "parameter_197" + original_name = "conv2d_3.b_0" shape = [4] dtype = "float32" min_val = float("0") @@ -2064,6 +2262,7 @@ class Program_weight_tensor_parameter_197: class Program_weight_tensor_parameter_198: name = "parameter_198" + original_name = "conv2d_3.w_0" shape = [4, 16, 1, 1] dtype = "float32" min_val = float("0") @@ -2073,6 +2272,7 @@ class Program_weight_tensor_parameter_198: class Program_weight_tensor_parameter_199: name = "parameter_199" + original_name = "batch_norm_2.b_0" shape = [16] dtype = "float32" min_val = float("0") @@ -2082,6 +2282,7 @@ class Program_weight_tensor_parameter_199: class Program_weight_tensor_parameter_200: name = "parameter_200" + original_name = "batch_norm_2.w_0" shape = [16] dtype = "float32" min_val = float("0") @@ -2091,6 +2292,7 @@ class Program_weight_tensor_parameter_200: class Program_weight_tensor_parameter_201: name = "parameter_201" + original_name = "batch_norm_2.w_2" shape = [16] dtype = "float32" min_val = float("0") @@ -2100,6 +2302,7 @@ class Program_weight_tensor_parameter_201: class Program_weight_tensor_parameter_202: name = "parameter_202" + original_name = "batch_norm_2.w_1" shape = [16] dtype = "float32" min_val = float("0") @@ -2109,6 +2312,7 @@ class Program_weight_tensor_parameter_202: class Program_weight_tensor_parameter_203: name = "parameter_203" + original_name = "conv2d_2.w_0" shape = [16, 1, 3, 3] dtype = "float32" min_val = float("-1.11817") @@ -2120,6 +2324,7 @@ class Program_weight_tensor_parameter_203: class Program_weight_tensor_parameter_204: name = "parameter_204" + original_name = "batch_norm_1.b_0" shape = [16] dtype = "float32" min_val = float("0") @@ -2129,6 +2334,7 @@ class Program_weight_tensor_parameter_204: class Program_weight_tensor_parameter_205: name = "parameter_205" + original_name = "batch_norm_1.w_0" shape = [16] dtype = "float32" min_val = float("0") @@ -2138,6 +2344,7 @@ class Program_weight_tensor_parameter_205: class Program_weight_tensor_parameter_206: name = "parameter_206" + original_name = "batch_norm_1.w_2" shape = [16] dtype = "float32" min_val = float("0") @@ -2147,6 +2354,7 @@ class Program_weight_tensor_parameter_206: class Program_weight_tensor_parameter_207: name = "parameter_207" + original_name = "batch_norm_1.w_1" shape = [16] dtype = "float32" min_val = float("0") @@ -2156,6 +2364,7 @@ class Program_weight_tensor_parameter_207: class Program_weight_tensor_parameter_208: name = "parameter_208" + original_name = "conv2d_1.w_0" shape = [16, 16, 1, 1] dtype = "float32" min_val = float("-1.97312") @@ -2167,6 +2376,7 @@ class Program_weight_tensor_parameter_208: class Program_weight_tensor_parameter_209: name = "parameter_209" + original_name = "batch_norm_0.b_0" shape = [16] dtype = "float32" min_val = float("0") @@ -2176,6 +2386,7 @@ class Program_weight_tensor_parameter_209: class Program_weight_tensor_parameter_210: name = "parameter_210" + original_name = "batch_norm_0.w_0" shape = [16] dtype = "float32" min_val = float("0") @@ -2185,6 +2396,7 @@ class Program_weight_tensor_parameter_210: class Program_weight_tensor_parameter_211: name = "parameter_211" + original_name = "batch_norm_0.w_2" shape = [16] dtype = "float32" min_val = float("0") @@ -2194,6 +2406,7 @@ class Program_weight_tensor_parameter_211: class Program_weight_tensor_parameter_212: name = "parameter_212" + original_name = "batch_norm_0.w_1" shape = [16] dtype = "float32" min_val = float("0") @@ -2203,6 +2416,7 @@ class Program_weight_tensor_parameter_212: class Program_weight_tensor_parameter_213: name = "parameter_213" + original_name = "conv2d_0.w_0" shape = [16, 3, 3, 3] dtype = "float32" min_val = float("-3.65141") diff --git a/paddle_samples/PaddleX/MobileNetV3_small_x1_0/subgraph_1/input_meta.py b/paddle_samples/PaddleX/MobileNetV3_small_x1_0/subgraph_1/input_meta.py index 4c1d87b63..459245741 100644 --- a/paddle_samples/PaddleX/MobileNetV3_small_x1_0/subgraph_1/input_meta.py +++ b/paddle_samples/PaddleX/MobileNetV3_small_x1_0/subgraph_1/input_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_data_0: name = "data_0" + original_name = "var_161" shape = [252, 3, 224, 224] dtype = "float32" min_val = float("-2.1179") diff --git a/paddle_samples/PaddleX/MobileNetV3_small_x1_0/subgraph_1/weight_meta.py b/paddle_samples/PaddleX/MobileNetV3_small_x1_0/subgraph_1/weight_meta.py index aa8c07e18..11aed612b 100644 --- a/paddle_samples/PaddleX/MobileNetV3_small_x1_0/subgraph_1/weight_meta.py +++ b/paddle_samples/PaddleX/MobileNetV3_small_x1_0/subgraph_1/weight_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_parameter_0: name = "parameter_0" + original_name = "linear_0.b_0" shape = [102] dtype = "float32" min_val = float("-0.000680778") @@ -11,6 +12,7 @@ class Program_weight_tensor_parameter_0: class Program_weight_tensor_parameter_1: name = "parameter_1" + original_name = "linear_0.w_0" shape = [1280, 102] dtype = "float32" min_val = float("-0.066166") @@ -22,6 +24,7 @@ class Program_weight_tensor_parameter_1: class Program_weight_tensor_parameter_2: name = "parameter_2" + original_name = "conv2d_53.w_0" shape = [1280, 576, 1, 1] dtype = "float32" min_val = float("-0.621002") @@ -33,6 +36,7 @@ class Program_weight_tensor_parameter_2: class Program_weight_tensor_parameter_3: name = "parameter_3" + original_name = "batch_norm_34.b_0" shape = [576] dtype = "float32" min_val = float("-10.4918") @@ -44,6 +48,7 @@ class Program_weight_tensor_parameter_3: class Program_weight_tensor_parameter_4: name = "parameter_4" + original_name = "batch_norm_34.w_0" shape = [576] dtype = "float32" min_val = float("2.17729") @@ -55,6 +60,7 @@ class Program_weight_tensor_parameter_4: class Program_weight_tensor_parameter_5: name = "parameter_5" + original_name = "batch_norm_34.w_2" shape = [576] dtype = "float32" min_val = float("29.0597") @@ -66,6 +72,7 @@ class Program_weight_tensor_parameter_5: class Program_weight_tensor_parameter_6: name = "parameter_6" + original_name = "batch_norm_34.w_1" shape = [576] dtype = "float32" min_val = float("-2.227e-05") @@ -77,6 +84,7 @@ class Program_weight_tensor_parameter_6: class Program_weight_tensor_parameter_7: name = "parameter_7" + original_name = "conv2d_52.w_0" shape = [576, 96, 1, 1] dtype = "float32" min_val = float("-0.911471") @@ -88,6 +96,7 @@ class Program_weight_tensor_parameter_7: class Program_weight_tensor_parameter_8: name = "parameter_8" + original_name = "batch_norm_33.b_0" shape = [96] dtype = "float32" min_val = float("-2.2843e-06") @@ -99,6 +108,7 @@ class Program_weight_tensor_parameter_8: class Program_weight_tensor_parameter_9: name = "parameter_9" + original_name = "batch_norm_33.w_0" shape = [96] dtype = "float32" min_val = float("2.13079") @@ -110,6 +120,7 @@ class Program_weight_tensor_parameter_9: class Program_weight_tensor_parameter_10: name = "parameter_10" + original_name = "batch_norm_33.w_2" shape = [96] dtype = "float32" min_val = float("0.0881524") @@ -121,6 +132,7 @@ class Program_weight_tensor_parameter_10: class Program_weight_tensor_parameter_11: name = "parameter_11" + original_name = "batch_norm_33.w_1" shape = [96] dtype = "float32" min_val = float("-0.492459") @@ -132,6 +144,7 @@ class Program_weight_tensor_parameter_11: class Program_weight_tensor_parameter_12: name = "parameter_12" + original_name = "conv2d_51.w_0" shape = [96, 576, 1, 1] dtype = "float32" min_val = float("-0.520394") @@ -143,6 +156,7 @@ class Program_weight_tensor_parameter_12: class Program_weight_tensor_parameter_13: name = "parameter_13" + original_name = "conv2d_50.b_0" shape = [576] dtype = "float32" min_val = float("-0.939092") @@ -154,6 +168,7 @@ class Program_weight_tensor_parameter_13: class Program_weight_tensor_parameter_14: name = "parameter_14" + original_name = "conv2d_50.w_0" shape = [576, 144, 1, 1] dtype = "float32" min_val = float("-0.665781") @@ -165,6 +180,7 @@ class Program_weight_tensor_parameter_14: class Program_weight_tensor_parameter_15: name = "parameter_15" + original_name = "conv2d_49.b_0" shape = [144] dtype = "float32" min_val = float("-0.0864439") @@ -176,6 +192,7 @@ class Program_weight_tensor_parameter_15: class Program_weight_tensor_parameter_16: name = "parameter_16" + original_name = "conv2d_49.w_0" shape = [144, 576, 1, 1] dtype = "float32" min_val = float("-0.401818") @@ -187,6 +204,7 @@ class Program_weight_tensor_parameter_16: class Program_weight_tensor_parameter_17: name = "parameter_17" + original_name = "batch_norm_32.b_0" shape = [576] dtype = "float32" min_val = float("-3.70529") @@ -198,6 +216,7 @@ class Program_weight_tensor_parameter_17: class Program_weight_tensor_parameter_18: name = "parameter_18" + original_name = "batch_norm_32.w_0" shape = [576] dtype = "float32" min_val = float("0.15635") @@ -209,6 +228,7 @@ class Program_weight_tensor_parameter_18: class Program_weight_tensor_parameter_19: name = "parameter_19" + original_name = "batch_norm_32.w_2" shape = [576] dtype = "float32" min_val = float("0.000122316") @@ -220,6 +240,7 @@ class Program_weight_tensor_parameter_19: class Program_weight_tensor_parameter_20: name = "parameter_20" + original_name = "batch_norm_32.w_1" shape = [576] dtype = "float32" min_val = float("-3.58164") @@ -231,6 +252,7 @@ class Program_weight_tensor_parameter_20: class Program_weight_tensor_parameter_21: name = "parameter_21" + original_name = "conv2d_48.w_0" shape = [576, 1, 5, 5] dtype = "float32" min_val = float("-0.685521") @@ -242,6 +264,7 @@ class Program_weight_tensor_parameter_21: class Program_weight_tensor_parameter_22: name = "parameter_22" + original_name = "batch_norm_31.b_0" shape = [576] dtype = "float32" min_val = float("-3.7139") @@ -253,6 +276,7 @@ class Program_weight_tensor_parameter_22: class Program_weight_tensor_parameter_23: name = "parameter_23" + original_name = "batch_norm_31.w_0" shape = [576] dtype = "float32" min_val = float("-0.923995") @@ -264,6 +288,7 @@ class Program_weight_tensor_parameter_23: class Program_weight_tensor_parameter_24: name = "parameter_24" + original_name = "batch_norm_31.w_2" shape = [576] dtype = "float32" min_val = float("7.86915") @@ -275,6 +300,7 @@ class Program_weight_tensor_parameter_24: class Program_weight_tensor_parameter_25: name = "parameter_25" + original_name = "batch_norm_31.w_1" shape = [576] dtype = "float32" min_val = float("-8.30168e-06") @@ -286,6 +312,7 @@ class Program_weight_tensor_parameter_25: class Program_weight_tensor_parameter_26: name = "parameter_26" + original_name = "conv2d_47.w_0" shape = [576, 96, 1, 1] dtype = "float32" min_val = float("-0.524471") @@ -297,6 +324,7 @@ class Program_weight_tensor_parameter_26: class Program_weight_tensor_parameter_27: name = "parameter_27" + original_name = "batch_norm_30.b_0" shape = [96] dtype = "float32" min_val = float("-2.37051e-06") @@ -308,6 +336,7 @@ class Program_weight_tensor_parameter_27: class Program_weight_tensor_parameter_28: name = "parameter_28" + original_name = "batch_norm_30.w_0" shape = [96] dtype = "float32" min_val = float("1.07015") @@ -319,6 +348,7 @@ class Program_weight_tensor_parameter_28: class Program_weight_tensor_parameter_29: name = "parameter_29" + original_name = "batch_norm_30.w_2" shape = [96] dtype = "float32" min_val = float("0.0698189") @@ -330,6 +360,7 @@ class Program_weight_tensor_parameter_29: class Program_weight_tensor_parameter_30: name = "parameter_30" + original_name = "batch_norm_30.w_1" shape = [96] dtype = "float32" min_val = float("-0.406691") @@ -341,6 +372,7 @@ class Program_weight_tensor_parameter_30: class Program_weight_tensor_parameter_31: name = "parameter_31" + original_name = "conv2d_46.w_0" shape = [96, 576, 1, 1] dtype = "float32" min_val = float("-0.738727") @@ -352,6 +384,7 @@ class Program_weight_tensor_parameter_31: class Program_weight_tensor_parameter_32: name = "parameter_32" + original_name = "conv2d_45.b_0" shape = [576] dtype = "float32" min_val = float("-0.272441") @@ -363,6 +396,7 @@ class Program_weight_tensor_parameter_32: class Program_weight_tensor_parameter_33: name = "parameter_33" + original_name = "conv2d_45.w_0" shape = [576, 144, 1, 1] dtype = "float32" min_val = float("-0.566104") @@ -374,6 +408,7 @@ class Program_weight_tensor_parameter_33: class Program_weight_tensor_parameter_34: name = "parameter_34" + original_name = "conv2d_44.b_0" shape = [144] dtype = "float32" min_val = float("-0.0736194") @@ -385,6 +420,7 @@ class Program_weight_tensor_parameter_34: class Program_weight_tensor_parameter_35: name = "parameter_35" + original_name = "conv2d_44.w_0" shape = [144, 576, 1, 1] dtype = "float32" min_val = float("-0.727416") @@ -396,6 +432,7 @@ class Program_weight_tensor_parameter_35: class Program_weight_tensor_parameter_36: name = "parameter_36" + original_name = "batch_norm_29.b_0" shape = [576] dtype = "float32" min_val = float("-3.72924") @@ -407,6 +444,7 @@ class Program_weight_tensor_parameter_36: class Program_weight_tensor_parameter_37: name = "parameter_37" + original_name = "batch_norm_29.w_0" shape = [576] dtype = "float32" min_val = float("0.289524") @@ -418,6 +456,7 @@ class Program_weight_tensor_parameter_37: class Program_weight_tensor_parameter_38: name = "parameter_38" + original_name = "batch_norm_29.w_2" shape = [576] dtype = "float32" min_val = float("0.000136088") @@ -429,6 +468,7 @@ class Program_weight_tensor_parameter_38: class Program_weight_tensor_parameter_39: name = "parameter_39" + original_name = "batch_norm_29.w_1" shape = [576] dtype = "float32" min_val = float("-2.8639") @@ -440,6 +480,7 @@ class Program_weight_tensor_parameter_39: class Program_weight_tensor_parameter_40: name = "parameter_40" + original_name = "conv2d_43.w_0" shape = [576, 1, 5, 5] dtype = "float32" min_val = float("-0.792041") @@ -451,6 +492,7 @@ class Program_weight_tensor_parameter_40: class Program_weight_tensor_parameter_41: name = "parameter_41" + original_name = "batch_norm_28.b_0" shape = [576] dtype = "float32" min_val = float("-3.85997") @@ -462,6 +504,7 @@ class Program_weight_tensor_parameter_41: class Program_weight_tensor_parameter_42: name = "parameter_42" + original_name = "batch_norm_28.w_0" shape = [576] dtype = "float32" min_val = float("-0.135053") @@ -473,6 +516,7 @@ class Program_weight_tensor_parameter_42: class Program_weight_tensor_parameter_43: name = "parameter_43" + original_name = "batch_norm_28.w_2" shape = [576] dtype = "float32" min_val = float("5.88739") @@ -484,6 +528,7 @@ class Program_weight_tensor_parameter_43: class Program_weight_tensor_parameter_44: name = "parameter_44" + original_name = "batch_norm_28.w_1" shape = [576] dtype = "float32" min_val = float("-4.41462e-06") @@ -495,6 +540,7 @@ class Program_weight_tensor_parameter_44: class Program_weight_tensor_parameter_45: name = "parameter_45" + original_name = "conv2d_42.w_0" shape = [576, 96, 1, 1] dtype = "float32" min_val = float("-0.486147") @@ -506,6 +552,7 @@ class Program_weight_tensor_parameter_45: class Program_weight_tensor_parameter_46: name = "parameter_46" + original_name = "batch_norm_27.b_0" shape = [96] dtype = "float32" min_val = float("-2.87088e-06") @@ -517,6 +564,7 @@ class Program_weight_tensor_parameter_46: class Program_weight_tensor_parameter_47: name = "parameter_47" + original_name = "batch_norm_27.w_0" shape = [96] dtype = "float32" min_val = float("2.4647") @@ -528,6 +576,7 @@ class Program_weight_tensor_parameter_47: class Program_weight_tensor_parameter_48: name = "parameter_48" + original_name = "batch_norm_27.w_2" shape = [96] dtype = "float32" min_val = float("0.0618971") @@ -539,6 +588,7 @@ class Program_weight_tensor_parameter_48: class Program_weight_tensor_parameter_49: name = "parameter_49" + original_name = "batch_norm_27.w_1" shape = [96] dtype = "float32" min_val = float("-0.81071") @@ -550,6 +600,7 @@ class Program_weight_tensor_parameter_49: class Program_weight_tensor_parameter_50: name = "parameter_50" + original_name = "conv2d_41.w_0" shape = [96, 288, 1, 1] dtype = "float32" min_val = float("-0.773263") @@ -561,6 +612,7 @@ class Program_weight_tensor_parameter_50: class Program_weight_tensor_parameter_51: name = "parameter_51" + original_name = "conv2d_40.b_0" shape = [288] dtype = "float32" min_val = float("-0.394609") @@ -572,6 +624,7 @@ class Program_weight_tensor_parameter_51: class Program_weight_tensor_parameter_52: name = "parameter_52" + original_name = "conv2d_40.w_0" shape = [288, 72, 1, 1] dtype = "float32" min_val = float("-0.876898") @@ -583,6 +636,7 @@ class Program_weight_tensor_parameter_52: class Program_weight_tensor_parameter_53: name = "parameter_53" + original_name = "conv2d_39.b_0" shape = [72] dtype = "float32" min_val = float("-0.0436365") @@ -594,6 +648,7 @@ class Program_weight_tensor_parameter_53: class Program_weight_tensor_parameter_54: name = "parameter_54" + original_name = "conv2d_39.w_0" shape = [72, 288, 1, 1] dtype = "float32" min_val = float("-0.630901") @@ -605,6 +660,7 @@ class Program_weight_tensor_parameter_54: class Program_weight_tensor_parameter_55: name = "parameter_55" + original_name = "batch_norm_26.b_0" shape = [288] dtype = "float32" min_val = float("-2.36598") @@ -616,6 +672,7 @@ class Program_weight_tensor_parameter_55: class Program_weight_tensor_parameter_56: name = "parameter_56" + original_name = "batch_norm_26.w_0" shape = [288] dtype = "float32" min_val = float("-0.659041") @@ -627,6 +684,7 @@ class Program_weight_tensor_parameter_56: class Program_weight_tensor_parameter_57: name = "parameter_57" + original_name = "batch_norm_26.w_2" shape = [288] dtype = "float32" min_val = float("0.000124935") @@ -638,6 +696,7 @@ class Program_weight_tensor_parameter_57: class Program_weight_tensor_parameter_58: name = "parameter_58" + original_name = "batch_norm_26.w_1" shape = [288] dtype = "float32" min_val = float("-13.035") @@ -649,6 +708,7 @@ class Program_weight_tensor_parameter_58: class Program_weight_tensor_parameter_59: name = "parameter_59" + original_name = "conv2d_38.w_0" shape = [288, 1, 5, 5] dtype = "float32" min_val = float("-0.424371") @@ -660,6 +720,7 @@ class Program_weight_tensor_parameter_59: class Program_weight_tensor_parameter_60: name = "parameter_60" + original_name = "batch_norm_25.b_0" shape = [288] dtype = "float32" min_val = float("-2.7018") @@ -671,6 +732,7 @@ class Program_weight_tensor_parameter_60: class Program_weight_tensor_parameter_61: name = "parameter_61" + original_name = "batch_norm_25.w_0" shape = [288] dtype = "float32" min_val = float("-2.55368") @@ -682,6 +744,7 @@ class Program_weight_tensor_parameter_61: class Program_weight_tensor_parameter_62: name = "parameter_62" + original_name = "batch_norm_25.w_2" shape = [288] dtype = "float32" min_val = float("10.5683") @@ -693,6 +756,7 @@ class Program_weight_tensor_parameter_62: class Program_weight_tensor_parameter_63: name = "parameter_63" + original_name = "batch_norm_25.w_1" shape = [288] dtype = "float32" min_val = float("-9.45429e-06") @@ -704,6 +768,7 @@ class Program_weight_tensor_parameter_63: class Program_weight_tensor_parameter_64: name = "parameter_64" + original_name = "conv2d_37.w_0" shape = [288, 48, 1, 1] dtype = "float32" min_val = float("-0.766378") @@ -715,6 +780,7 @@ class Program_weight_tensor_parameter_64: class Program_weight_tensor_parameter_65: name = "parameter_65" + original_name = "batch_norm_24.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -724,6 +790,7 @@ class Program_weight_tensor_parameter_65: class Program_weight_tensor_parameter_66: name = "parameter_66" + original_name = "batch_norm_24.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -733,6 +800,7 @@ class Program_weight_tensor_parameter_66: class Program_weight_tensor_parameter_67: name = "parameter_67" + original_name = "batch_norm_24.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -742,6 +810,7 @@ class Program_weight_tensor_parameter_67: class Program_weight_tensor_parameter_68: name = "parameter_68" + original_name = "batch_norm_24.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -751,6 +820,7 @@ class Program_weight_tensor_parameter_68: class Program_weight_tensor_parameter_69: name = "parameter_69" + original_name = "conv2d_36.w_0" shape = [48, 144, 1, 1] dtype = "float32" min_val = float("-0.658908") @@ -762,6 +832,7 @@ class Program_weight_tensor_parameter_69: class Program_weight_tensor_parameter_70: name = "parameter_70" + original_name = "conv2d_35.b_0" shape = [144] dtype = "float32" min_val = float("-0.333388") @@ -773,6 +844,7 @@ class Program_weight_tensor_parameter_70: class Program_weight_tensor_parameter_71: name = "parameter_71" + original_name = "conv2d_35.w_0" shape = [144, 36, 1, 1] dtype = "float32" min_val = float("-0.653218") @@ -784,6 +856,7 @@ class Program_weight_tensor_parameter_71: class Program_weight_tensor_parameter_72: name = "parameter_72" + original_name = "conv2d_34.b_0" shape = [36] dtype = "float32" min_val = float("0") @@ -793,6 +866,7 @@ class Program_weight_tensor_parameter_72: class Program_weight_tensor_parameter_73: name = "parameter_73" + original_name = "conv2d_34.w_0" shape = [36, 144, 1, 1] dtype = "float32" min_val = float("-0.836522") @@ -804,6 +878,7 @@ class Program_weight_tensor_parameter_73: class Program_weight_tensor_parameter_74: name = "parameter_74" + original_name = "batch_norm_23.b_0" shape = [144] dtype = "float32" min_val = float("-4.12061") @@ -815,6 +890,7 @@ class Program_weight_tensor_parameter_74: class Program_weight_tensor_parameter_75: name = "parameter_75" + original_name = "batch_norm_23.w_0" shape = [144] dtype = "float32" min_val = float("0.304347") @@ -826,6 +902,7 @@ class Program_weight_tensor_parameter_75: class Program_weight_tensor_parameter_76: name = "parameter_76" + original_name = "batch_norm_23.w_2" shape = [144] dtype = "float32" min_val = float("0.00109112") @@ -837,6 +914,7 @@ class Program_weight_tensor_parameter_76: class Program_weight_tensor_parameter_77: name = "parameter_77" + original_name = "batch_norm_23.w_1" shape = [144] dtype = "float32" min_val = float("-3.49106") @@ -848,6 +926,7 @@ class Program_weight_tensor_parameter_77: class Program_weight_tensor_parameter_78: name = "parameter_78" + original_name = "conv2d_33.w_0" shape = [144, 1, 5, 5] dtype = "float32" min_val = float("-0.893889") @@ -859,6 +938,7 @@ class Program_weight_tensor_parameter_78: class Program_weight_tensor_parameter_79: name = "parameter_79" + original_name = "batch_norm_22.b_0" shape = [144] dtype = "float32" min_val = float("-1.93896") @@ -870,6 +950,7 @@ class Program_weight_tensor_parameter_79: class Program_weight_tensor_parameter_80: name = "parameter_80" + original_name = "batch_norm_22.w_0" shape = [144] dtype = "float32" min_val = float("0.0907257") @@ -881,6 +962,7 @@ class Program_weight_tensor_parameter_80: class Program_weight_tensor_parameter_81: name = "parameter_81" + original_name = "batch_norm_22.w_2" shape = [144] dtype = "float32" min_val = float("13.3597") @@ -892,6 +974,7 @@ class Program_weight_tensor_parameter_81: class Program_weight_tensor_parameter_82: name = "parameter_82" + original_name = "batch_norm_22.w_1" shape = [144] dtype = "float32" min_val = float("-3.85523e-06") @@ -903,6 +986,7 @@ class Program_weight_tensor_parameter_82: class Program_weight_tensor_parameter_83: name = "parameter_83" + original_name = "conv2d_32.w_0" shape = [144, 48, 1, 1] dtype = "float32" min_val = float("-0.648127") @@ -914,6 +998,7 @@ class Program_weight_tensor_parameter_83: class Program_weight_tensor_parameter_84: name = "parameter_84" + original_name = "batch_norm_21.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -923,6 +1008,7 @@ class Program_weight_tensor_parameter_84: class Program_weight_tensor_parameter_85: name = "parameter_85" + original_name = "batch_norm_21.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -932,6 +1018,7 @@ class Program_weight_tensor_parameter_85: class Program_weight_tensor_parameter_86: name = "parameter_86" + original_name = "batch_norm_21.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -941,6 +1028,7 @@ class Program_weight_tensor_parameter_86: class Program_weight_tensor_parameter_87: name = "parameter_87" + original_name = "batch_norm_21.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -950,6 +1038,7 @@ class Program_weight_tensor_parameter_87: class Program_weight_tensor_parameter_88: name = "parameter_88" + original_name = "conv2d_31.w_0" shape = [48, 120, 1, 1] dtype = "float32" min_val = float("-0.793744") @@ -961,6 +1050,7 @@ class Program_weight_tensor_parameter_88: class Program_weight_tensor_parameter_89: name = "parameter_89" + original_name = "conv2d_30.b_0" shape = [120] dtype = "float32" min_val = float("-0.646494") @@ -972,6 +1062,7 @@ class Program_weight_tensor_parameter_89: class Program_weight_tensor_parameter_90: name = "parameter_90" + original_name = "conv2d_30.w_0" shape = [120, 30, 1, 1] dtype = "float32" min_val = float("-1.42327e-12") @@ -983,6 +1074,7 @@ class Program_weight_tensor_parameter_90: class Program_weight_tensor_parameter_91: name = "parameter_91" + original_name = "conv2d_29.b_0" shape = [30] dtype = "float32" min_val = float("0") @@ -992,6 +1084,7 @@ class Program_weight_tensor_parameter_91: class Program_weight_tensor_parameter_92: name = "parameter_92" + original_name = "conv2d_29.w_0" shape = [30, 120, 1, 1] dtype = "float32" min_val = float("-2.21195e-12") @@ -1003,6 +1096,7 @@ class Program_weight_tensor_parameter_92: class Program_weight_tensor_parameter_93: name = "parameter_93" + original_name = "batch_norm_20.b_0" shape = [120] dtype = "float32" min_val = float("-1.37631") @@ -1014,6 +1108,7 @@ class Program_weight_tensor_parameter_93: class Program_weight_tensor_parameter_94: name = "parameter_94" + original_name = "batch_norm_20.w_0" shape = [120] dtype = "float32" min_val = float("-1.00863") @@ -1025,6 +1120,7 @@ class Program_weight_tensor_parameter_94: class Program_weight_tensor_parameter_95: name = "parameter_95" + original_name = "batch_norm_20.w_2" shape = [120] dtype = "float32" min_val = float("0.000292613") @@ -1036,6 +1132,7 @@ class Program_weight_tensor_parameter_95: class Program_weight_tensor_parameter_96: name = "parameter_96" + original_name = "batch_norm_20.w_1" shape = [120] dtype = "float32" min_val = float("-7.74041") @@ -1047,6 +1144,7 @@ class Program_weight_tensor_parameter_96: class Program_weight_tensor_parameter_97: name = "parameter_97" + original_name = "conv2d_28.w_0" shape = [120, 1, 5, 5] dtype = "float32" min_val = float("-1.5518") @@ -1058,6 +1156,7 @@ class Program_weight_tensor_parameter_97: class Program_weight_tensor_parameter_98: name = "parameter_98" + original_name = "batch_norm_19.b_0" shape = [120] dtype = "float32" min_val = float("-1.55136") @@ -1069,6 +1168,7 @@ class Program_weight_tensor_parameter_98: class Program_weight_tensor_parameter_99: name = "parameter_99" + original_name = "batch_norm_19.w_0" shape = [120] dtype = "float32" min_val = float("-1.76578") @@ -1080,6 +1180,7 @@ class Program_weight_tensor_parameter_99: class Program_weight_tensor_parameter_100: name = "parameter_100" + original_name = "batch_norm_19.w_2" shape = [120] dtype = "float32" min_val = float("26.9582") @@ -1091,6 +1192,7 @@ class Program_weight_tensor_parameter_100: class Program_weight_tensor_parameter_101: name = "parameter_101" + original_name = "batch_norm_19.w_1" shape = [120] dtype = "float32" min_val = float("-2.06196e-05") @@ -1102,6 +1204,7 @@ class Program_weight_tensor_parameter_101: class Program_weight_tensor_parameter_102: name = "parameter_102" + original_name = "conv2d_27.w_0" shape = [120, 40, 1, 1] dtype = "float32" min_val = float("-1.26656") @@ -1113,6 +1216,7 @@ class Program_weight_tensor_parameter_102: class Program_weight_tensor_parameter_103: name = "parameter_103" + original_name = "batch_norm_18.b_0" shape = [40] dtype = "float32" min_val = float("0") @@ -1122,6 +1226,7 @@ class Program_weight_tensor_parameter_103: class Program_weight_tensor_parameter_104: name = "parameter_104" + original_name = "batch_norm_18.w_0" shape = [40] dtype = "float32" min_val = float("0") @@ -1131,6 +1236,7 @@ class Program_weight_tensor_parameter_104: class Program_weight_tensor_parameter_105: name = "parameter_105" + original_name = "batch_norm_18.w_2" shape = [40] dtype = "float32" min_val = float("0") @@ -1140,6 +1246,7 @@ class Program_weight_tensor_parameter_105: class Program_weight_tensor_parameter_106: name = "parameter_106" + original_name = "batch_norm_18.w_1" shape = [40] dtype = "float32" min_val = float("0") @@ -1149,6 +1256,7 @@ class Program_weight_tensor_parameter_106: class Program_weight_tensor_parameter_107: name = "parameter_107" + original_name = "conv2d_26.w_0" shape = [40, 240, 1, 1] dtype = "float32" min_val = float("-0.73243") @@ -1160,6 +1268,7 @@ class Program_weight_tensor_parameter_107: class Program_weight_tensor_parameter_108: name = "parameter_108" + original_name = "conv2d_25.b_0" shape = [240] dtype = "float32" min_val = float("-0.285765") @@ -1171,6 +1280,7 @@ class Program_weight_tensor_parameter_108: class Program_weight_tensor_parameter_109: name = "parameter_109" + original_name = "conv2d_25.w_0" shape = [240, 60, 1, 1] dtype = "float32" min_val = float("-0.729173") @@ -1182,6 +1292,7 @@ class Program_weight_tensor_parameter_109: class Program_weight_tensor_parameter_110: name = "parameter_110" + original_name = "conv2d_24.b_0" shape = [60] dtype = "float32" min_val = float("0") @@ -1191,6 +1302,7 @@ class Program_weight_tensor_parameter_110: class Program_weight_tensor_parameter_111: name = "parameter_111" + original_name = "conv2d_24.w_0" shape = [60, 240, 1, 1] dtype = "float32" min_val = float("-1.02194") @@ -1202,6 +1314,7 @@ class Program_weight_tensor_parameter_111: class Program_weight_tensor_parameter_112: name = "parameter_112" + original_name = "batch_norm_17.b_0" shape = [240] dtype = "float32" min_val = float("-4.56095") @@ -1213,6 +1326,7 @@ class Program_weight_tensor_parameter_112: class Program_weight_tensor_parameter_113: name = "parameter_113" + original_name = "batch_norm_17.w_0" shape = [240] dtype = "float32" min_val = float("0.151793") @@ -1224,6 +1338,7 @@ class Program_weight_tensor_parameter_113: class Program_weight_tensor_parameter_114: name = "parameter_114" + original_name = "batch_norm_17.w_2" shape = [240] dtype = "float32" min_val = float("0.000194449") @@ -1235,6 +1350,7 @@ class Program_weight_tensor_parameter_114: class Program_weight_tensor_parameter_115: name = "parameter_115" + original_name = "batch_norm_17.w_1" shape = [240] dtype = "float32" min_val = float("-5.24683") @@ -1246,6 +1362,7 @@ class Program_weight_tensor_parameter_115: class Program_weight_tensor_parameter_116: name = "parameter_116" + original_name = "conv2d_23.w_0" shape = [240, 1, 5, 5] dtype = "float32" min_val = float("-0.900123") @@ -1257,6 +1374,7 @@ class Program_weight_tensor_parameter_116: class Program_weight_tensor_parameter_117: name = "parameter_117" + original_name = "batch_norm_16.b_0" shape = [240] dtype = "float32" min_val = float("-3.78027") @@ -1268,6 +1386,7 @@ class Program_weight_tensor_parameter_117: class Program_weight_tensor_parameter_118: name = "parameter_118" + original_name = "batch_norm_16.w_0" shape = [240] dtype = "float32" min_val = float("0.0345772") @@ -1279,6 +1398,7 @@ class Program_weight_tensor_parameter_118: class Program_weight_tensor_parameter_119: name = "parameter_119" + original_name = "batch_norm_16.w_2" shape = [240] dtype = "float32" min_val = float("7.47576") @@ -1290,6 +1410,7 @@ class Program_weight_tensor_parameter_119: class Program_weight_tensor_parameter_120: name = "parameter_120" + original_name = "batch_norm_16.w_1" shape = [240] dtype = "float32" min_val = float("-7.6614e-06") @@ -1301,6 +1422,7 @@ class Program_weight_tensor_parameter_120: class Program_weight_tensor_parameter_121: name = "parameter_121" + original_name = "conv2d_22.w_0" shape = [240, 40, 1, 1] dtype = "float32" min_val = float("-1.39554") @@ -1312,6 +1434,7 @@ class Program_weight_tensor_parameter_121: class Program_weight_tensor_parameter_122: name = "parameter_122" + original_name = "batch_norm_15.b_0" shape = [40] dtype = "float32" min_val = float("0") @@ -1321,6 +1444,7 @@ class Program_weight_tensor_parameter_122: class Program_weight_tensor_parameter_123: name = "parameter_123" + original_name = "batch_norm_15.w_0" shape = [40] dtype = "float32" min_val = float("0") @@ -1330,6 +1454,7 @@ class Program_weight_tensor_parameter_123: class Program_weight_tensor_parameter_124: name = "parameter_124" + original_name = "batch_norm_15.w_2" shape = [40] dtype = "float32" min_val = float("0") @@ -1339,6 +1464,7 @@ class Program_weight_tensor_parameter_124: class Program_weight_tensor_parameter_125: name = "parameter_125" + original_name = "batch_norm_15.w_1" shape = [40] dtype = "float32" min_val = float("0") @@ -1348,6 +1474,7 @@ class Program_weight_tensor_parameter_125: class Program_weight_tensor_parameter_126: name = "parameter_126" + original_name = "conv2d_21.w_0" shape = [40, 240, 1, 1] dtype = "float32" min_val = float("-0.917397") @@ -1359,6 +1486,7 @@ class Program_weight_tensor_parameter_126: class Program_weight_tensor_parameter_127: name = "parameter_127" + original_name = "conv2d_20.b_0" shape = [240] dtype = "float32" min_val = float("-0.282661") @@ -1370,6 +1498,7 @@ class Program_weight_tensor_parameter_127: class Program_weight_tensor_parameter_128: name = "parameter_128" + original_name = "conv2d_20.w_0" shape = [240, 60, 1, 1] dtype = "float32" min_val = float("-0.621251") @@ -1381,6 +1510,7 @@ class Program_weight_tensor_parameter_128: class Program_weight_tensor_parameter_129: name = "parameter_129" + original_name = "conv2d_19.b_0" shape = [60] dtype = "float32" min_val = float("0") @@ -1390,6 +1520,7 @@ class Program_weight_tensor_parameter_129: class Program_weight_tensor_parameter_130: name = "parameter_130" + original_name = "conv2d_19.w_0" shape = [60, 240, 1, 1] dtype = "float32" min_val = float("-0.872053") @@ -1401,6 +1532,7 @@ class Program_weight_tensor_parameter_130: class Program_weight_tensor_parameter_131: name = "parameter_131" + original_name = "batch_norm_14.b_0" shape = [240] dtype = "float32" min_val = float("-2.46527") @@ -1412,6 +1544,7 @@ class Program_weight_tensor_parameter_131: class Program_weight_tensor_parameter_132: name = "parameter_132" + original_name = "batch_norm_14.w_0" shape = [240] dtype = "float32" min_val = float("0.195196") @@ -1423,6 +1556,7 @@ class Program_weight_tensor_parameter_132: class Program_weight_tensor_parameter_133: name = "parameter_133" + original_name = "batch_norm_14.w_2" shape = [240] dtype = "float32" min_val = float("0.000135199") @@ -1434,6 +1568,7 @@ class Program_weight_tensor_parameter_133: class Program_weight_tensor_parameter_134: name = "parameter_134" + original_name = "batch_norm_14.w_1" shape = [240] dtype = "float32" min_val = float("-19.0963") @@ -1445,6 +1580,7 @@ class Program_weight_tensor_parameter_134: class Program_weight_tensor_parameter_135: name = "parameter_135" + original_name = "conv2d_18.w_0" shape = [240, 1, 5, 5] dtype = "float32" min_val = float("-1.08565") @@ -1456,6 +1592,7 @@ class Program_weight_tensor_parameter_135: class Program_weight_tensor_parameter_136: name = "parameter_136" + original_name = "batch_norm_13.b_0" shape = [240] dtype = "float32" min_val = float("-1.93029") @@ -1467,6 +1604,7 @@ class Program_weight_tensor_parameter_136: class Program_weight_tensor_parameter_137: name = "parameter_137" + original_name = "batch_norm_13.w_0" shape = [240] dtype = "float32" min_val = float("-0.443253") @@ -1478,6 +1616,7 @@ class Program_weight_tensor_parameter_137: class Program_weight_tensor_parameter_138: name = "parameter_138" + original_name = "batch_norm_13.w_2" shape = [240] dtype = "float32" min_val = float("9.1708") @@ -1489,6 +1628,7 @@ class Program_weight_tensor_parameter_138: class Program_weight_tensor_parameter_139: name = "parameter_139" + original_name = "batch_norm_13.w_1" shape = [240] dtype = "float32" min_val = float("-5.41307e-06") @@ -1500,6 +1640,7 @@ class Program_weight_tensor_parameter_139: class Program_weight_tensor_parameter_140: name = "parameter_140" + original_name = "conv2d_17.w_0" shape = [240, 40, 1, 1] dtype = "float32" min_val = float("-0.624959") @@ -1511,6 +1652,7 @@ class Program_weight_tensor_parameter_140: class Program_weight_tensor_parameter_141: name = "parameter_141" + original_name = "batch_norm_12.b_0" shape = [40] dtype = "float32" min_val = float("0") @@ -1520,6 +1662,7 @@ class Program_weight_tensor_parameter_141: class Program_weight_tensor_parameter_142: name = "parameter_142" + original_name = "batch_norm_12.w_0" shape = [40] dtype = "float32" min_val = float("0") @@ -1529,6 +1672,7 @@ class Program_weight_tensor_parameter_142: class Program_weight_tensor_parameter_143: name = "parameter_143" + original_name = "batch_norm_12.w_2" shape = [40] dtype = "float32" min_val = float("0") @@ -1538,6 +1682,7 @@ class Program_weight_tensor_parameter_143: class Program_weight_tensor_parameter_144: name = "parameter_144" + original_name = "batch_norm_12.w_1" shape = [40] dtype = "float32" min_val = float("0") @@ -1547,6 +1692,7 @@ class Program_weight_tensor_parameter_144: class Program_weight_tensor_parameter_145: name = "parameter_145" + original_name = "conv2d_16.w_0" shape = [40, 96, 1, 1] dtype = "float32" min_val = float("-0.960479") @@ -1558,6 +1704,7 @@ class Program_weight_tensor_parameter_145: class Program_weight_tensor_parameter_146: name = "parameter_146" + original_name = "conv2d_15.b_0" shape = [96] dtype = "float32" min_val = float("0.00556343") @@ -1569,6 +1716,7 @@ class Program_weight_tensor_parameter_146: class Program_weight_tensor_parameter_147: name = "parameter_147" + original_name = "conv2d_15.w_0" shape = [96, 24, 1, 1] dtype = "float32" min_val = float("-2.33486e-09") @@ -1580,6 +1728,7 @@ class Program_weight_tensor_parameter_147: class Program_weight_tensor_parameter_148: name = "parameter_148" + original_name = "conv2d_14.b_0" shape = [24] dtype = "float32" min_val = float("0") @@ -1589,6 +1738,7 @@ class Program_weight_tensor_parameter_148: class Program_weight_tensor_parameter_149: name = "parameter_149" + original_name = "conv2d_14.w_0" shape = [24, 96, 1, 1] dtype = "float32" min_val = float("-0.00298537") @@ -1600,6 +1750,7 @@ class Program_weight_tensor_parameter_149: class Program_weight_tensor_parameter_150: name = "parameter_150" + original_name = "batch_norm_11.b_0" shape = [96] dtype = "float32" min_val = float("-1.21169") @@ -1611,6 +1762,7 @@ class Program_weight_tensor_parameter_150: class Program_weight_tensor_parameter_151: name = "parameter_151" + original_name = "batch_norm_11.w_0" shape = [96] dtype = "float32" min_val = float("0.42486") @@ -1622,6 +1774,7 @@ class Program_weight_tensor_parameter_151: class Program_weight_tensor_parameter_152: name = "parameter_152" + original_name = "batch_norm_11.w_2" shape = [96] dtype = "float32" min_val = float("0.000749812") @@ -1633,6 +1786,7 @@ class Program_weight_tensor_parameter_152: class Program_weight_tensor_parameter_153: name = "parameter_153" + original_name = "batch_norm_11.w_1" shape = [96] dtype = "float32" min_val = float("-11.4467") @@ -1644,6 +1798,7 @@ class Program_weight_tensor_parameter_153: class Program_weight_tensor_parameter_154: name = "parameter_154" + original_name = "conv2d_13.w_0" shape = [96, 1, 5, 5] dtype = "float32" min_val = float("-0.621281") @@ -1655,6 +1810,7 @@ class Program_weight_tensor_parameter_154: class Program_weight_tensor_parameter_155: name = "parameter_155" + original_name = "batch_norm_10.b_0" shape = [96] dtype = "float32" min_val = float("-2.63607") @@ -1666,6 +1822,7 @@ class Program_weight_tensor_parameter_155: class Program_weight_tensor_parameter_156: name = "parameter_156" + original_name = "batch_norm_10.w_0" shape = [96] dtype = "float32" min_val = float("0.060942") @@ -1677,6 +1834,7 @@ class Program_weight_tensor_parameter_156: class Program_weight_tensor_parameter_157: name = "parameter_157" + original_name = "batch_norm_10.w_2" shape = [96] dtype = "float32" min_val = float("26.1813") @@ -1688,6 +1846,7 @@ class Program_weight_tensor_parameter_157: class Program_weight_tensor_parameter_158: name = "parameter_158" + original_name = "batch_norm_10.w_1" shape = [96] dtype = "float32" min_val = float("-3.60324e-05") @@ -1699,6 +1858,7 @@ class Program_weight_tensor_parameter_158: class Program_weight_tensor_parameter_159: name = "parameter_159" + original_name = "conv2d_12.w_0" shape = [96, 24, 1, 1] dtype = "float32" min_val = float("-1.48159") @@ -1710,6 +1870,7 @@ class Program_weight_tensor_parameter_159: class Program_weight_tensor_parameter_160: name = "parameter_160" + original_name = "batch_norm_9.b_0" shape = [24] dtype = "float32" min_val = float("0") @@ -1719,6 +1880,7 @@ class Program_weight_tensor_parameter_160: class Program_weight_tensor_parameter_161: name = "parameter_161" + original_name = "batch_norm_9.w_0" shape = [24] dtype = "float32" min_val = float("0") @@ -1728,6 +1890,7 @@ class Program_weight_tensor_parameter_161: class Program_weight_tensor_parameter_162: name = "parameter_162" + original_name = "batch_norm_9.w_2" shape = [24] dtype = "float32" min_val = float("0") @@ -1737,6 +1900,7 @@ class Program_weight_tensor_parameter_162: class Program_weight_tensor_parameter_163: name = "parameter_163" + original_name = "batch_norm_9.w_1" shape = [24] dtype = "float32" min_val = float("0") @@ -1746,6 +1910,7 @@ class Program_weight_tensor_parameter_163: class Program_weight_tensor_parameter_164: name = "parameter_164" + original_name = "conv2d_11.w_0" shape = [24, 88, 1, 1] dtype = "float32" min_val = float("-1.47738") @@ -1757,6 +1922,7 @@ class Program_weight_tensor_parameter_164: class Program_weight_tensor_parameter_165: name = "parameter_165" + original_name = "batch_norm_8.b_0" shape = [88] dtype = "float32" min_val = float("-6.41054") @@ -1768,6 +1934,7 @@ class Program_weight_tensor_parameter_165: class Program_weight_tensor_parameter_166: name = "parameter_166" + original_name = "batch_norm_8.w_0" shape = [88] dtype = "float32" min_val = float("0.863438") @@ -1779,6 +1946,7 @@ class Program_weight_tensor_parameter_166: class Program_weight_tensor_parameter_167: name = "parameter_167" + original_name = "batch_norm_8.w_2" shape = [88] dtype = "float32" min_val = float("0.00985231") @@ -1790,6 +1958,7 @@ class Program_weight_tensor_parameter_167: class Program_weight_tensor_parameter_168: name = "parameter_168" + original_name = "batch_norm_8.w_1" shape = [88] dtype = "float32" min_val = float("-4.8238") @@ -1801,6 +1970,7 @@ class Program_weight_tensor_parameter_168: class Program_weight_tensor_parameter_169: name = "parameter_169" + original_name = "conv2d_10.w_0" shape = [88, 1, 3, 3] dtype = "float32" min_val = float("-1.34106") @@ -1812,6 +1982,7 @@ class Program_weight_tensor_parameter_169: class Program_weight_tensor_parameter_170: name = "parameter_170" + original_name = "batch_norm_7.b_0" shape = [88] dtype = "float32" min_val = float("-3.26577") @@ -1823,6 +1994,7 @@ class Program_weight_tensor_parameter_170: class Program_weight_tensor_parameter_171: name = "parameter_171" + original_name = "batch_norm_7.w_0" shape = [88] dtype = "float32" min_val = float("0.51747") @@ -1834,6 +2006,7 @@ class Program_weight_tensor_parameter_171: class Program_weight_tensor_parameter_172: name = "parameter_172" + original_name = "batch_norm_7.w_2" shape = [88] dtype = "float32" min_val = float("9.47278") @@ -1845,6 +2018,7 @@ class Program_weight_tensor_parameter_172: class Program_weight_tensor_parameter_173: name = "parameter_173" + original_name = "batch_norm_7.w_1" shape = [88] dtype = "float32" min_val = float("-1.64261e-05") @@ -1856,6 +2030,7 @@ class Program_weight_tensor_parameter_173: class Program_weight_tensor_parameter_174: name = "parameter_174" + original_name = "conv2d_9.w_0" shape = [88, 24, 1, 1] dtype = "float32" min_val = float("-1.2909") @@ -1867,6 +2042,7 @@ class Program_weight_tensor_parameter_174: class Program_weight_tensor_parameter_175: name = "parameter_175" + original_name = "batch_norm_6.b_0" shape = [24] dtype = "float32" min_val = float("0") @@ -1876,6 +2052,7 @@ class Program_weight_tensor_parameter_175: class Program_weight_tensor_parameter_176: name = "parameter_176" + original_name = "batch_norm_6.w_0" shape = [24] dtype = "float32" min_val = float("0") @@ -1885,6 +2062,7 @@ class Program_weight_tensor_parameter_176: class Program_weight_tensor_parameter_177: name = "parameter_177" + original_name = "batch_norm_6.w_2" shape = [24] dtype = "float32" min_val = float("0") @@ -1894,6 +2072,7 @@ class Program_weight_tensor_parameter_177: class Program_weight_tensor_parameter_178: name = "parameter_178" + original_name = "batch_norm_6.w_1" shape = [24] dtype = "float32" min_val = float("0") @@ -1903,6 +2082,7 @@ class Program_weight_tensor_parameter_178: class Program_weight_tensor_parameter_179: name = "parameter_179" + original_name = "conv2d_8.w_0" shape = [24, 72, 1, 1] dtype = "float32" min_val = float("-1.72477") @@ -1914,6 +2094,7 @@ class Program_weight_tensor_parameter_179: class Program_weight_tensor_parameter_180: name = "parameter_180" + original_name = "batch_norm_5.b_0" shape = [72] dtype = "float32" min_val = float("-1.88791") @@ -1925,6 +2106,7 @@ class Program_weight_tensor_parameter_180: class Program_weight_tensor_parameter_181: name = "parameter_181" + original_name = "batch_norm_5.w_0" shape = [72] dtype = "float32" min_val = float("1.10046") @@ -1936,6 +2118,7 @@ class Program_weight_tensor_parameter_181: class Program_weight_tensor_parameter_182: name = "parameter_182" + original_name = "batch_norm_5.w_2" shape = [72] dtype = "float32" min_val = float("0.0240755") @@ -1947,6 +2130,7 @@ class Program_weight_tensor_parameter_182: class Program_weight_tensor_parameter_183: name = "parameter_183" + original_name = "batch_norm_5.w_1" shape = [72] dtype = "float32" min_val = float("-10.9978") @@ -1958,6 +2142,7 @@ class Program_weight_tensor_parameter_183: class Program_weight_tensor_parameter_184: name = "parameter_184" + original_name = "conv2d_7.w_0" shape = [72, 1, 3, 3] dtype = "float32" min_val = float("-1.47549") @@ -1969,6 +2154,7 @@ class Program_weight_tensor_parameter_184: class Program_weight_tensor_parameter_185: name = "parameter_185" + original_name = "batch_norm_4.b_0" shape = [72] dtype = "float32" min_val = float("-5.40437") @@ -1980,6 +2166,7 @@ class Program_weight_tensor_parameter_185: class Program_weight_tensor_parameter_186: name = "parameter_186" + original_name = "batch_norm_4.w_0" shape = [72] dtype = "float32" min_val = float("0.200701") @@ -1991,6 +2178,7 @@ class Program_weight_tensor_parameter_186: class Program_weight_tensor_parameter_187: name = "parameter_187" + original_name = "batch_norm_4.w_2" shape = [72] dtype = "float32" min_val = float("13.8218") @@ -2002,6 +2190,7 @@ class Program_weight_tensor_parameter_187: class Program_weight_tensor_parameter_188: name = "parameter_188" + original_name = "batch_norm_4.w_1" shape = [72] dtype = "float32" min_val = float("-0.00011272") @@ -2013,6 +2202,7 @@ class Program_weight_tensor_parameter_188: class Program_weight_tensor_parameter_189: name = "parameter_189" + original_name = "conv2d_6.w_0" shape = [72, 16, 1, 1] dtype = "float32" min_val = float("-1.62457") @@ -2024,6 +2214,7 @@ class Program_weight_tensor_parameter_189: class Program_weight_tensor_parameter_190: name = "parameter_190" + original_name = "batch_norm_3.b_0" shape = [16] dtype = "float32" min_val = float("0") @@ -2033,6 +2224,7 @@ class Program_weight_tensor_parameter_190: class Program_weight_tensor_parameter_191: name = "parameter_191" + original_name = "batch_norm_3.w_0" shape = [16] dtype = "float32" min_val = float("0") @@ -2042,6 +2234,7 @@ class Program_weight_tensor_parameter_191: class Program_weight_tensor_parameter_192: name = "parameter_192" + original_name = "batch_norm_3.w_2" shape = [16] dtype = "float32" min_val = float("0") @@ -2051,6 +2244,7 @@ class Program_weight_tensor_parameter_192: class Program_weight_tensor_parameter_193: name = "parameter_193" + original_name = "batch_norm_3.w_1" shape = [16] dtype = "float32" min_val = float("0") @@ -2060,6 +2254,7 @@ class Program_weight_tensor_parameter_193: class Program_weight_tensor_parameter_194: name = "parameter_194" + original_name = "conv2d_5.w_0" shape = [16, 16, 1, 1] dtype = "float32" min_val = float("-1.68584") @@ -2071,6 +2266,7 @@ class Program_weight_tensor_parameter_194: class Program_weight_tensor_parameter_195: name = "parameter_195" + original_name = "conv2d_4.b_0" shape = [16] dtype = "float32" min_val = float("0") @@ -2080,6 +2276,7 @@ class Program_weight_tensor_parameter_195: class Program_weight_tensor_parameter_196: name = "parameter_196" + original_name = "conv2d_4.w_0" shape = [16, 4, 1, 1] dtype = "float32" min_val = float("0") @@ -2089,6 +2286,7 @@ class Program_weight_tensor_parameter_196: class Program_weight_tensor_parameter_197: name = "parameter_197" + original_name = "conv2d_3.b_0" shape = [4] dtype = "float32" min_val = float("0") @@ -2098,6 +2296,7 @@ class Program_weight_tensor_parameter_197: class Program_weight_tensor_parameter_198: name = "parameter_198" + original_name = "conv2d_3.w_0" shape = [4, 16, 1, 1] dtype = "float32" min_val = float("0") @@ -2107,6 +2306,7 @@ class Program_weight_tensor_parameter_198: class Program_weight_tensor_parameter_199: name = "parameter_199" + original_name = "batch_norm_2.b_0" shape = [16] dtype = "float32" min_val = float("0") @@ -2116,6 +2316,7 @@ class Program_weight_tensor_parameter_199: class Program_weight_tensor_parameter_200: name = "parameter_200" + original_name = "batch_norm_2.w_0" shape = [16] dtype = "float32" min_val = float("0") @@ -2125,6 +2326,7 @@ class Program_weight_tensor_parameter_200: class Program_weight_tensor_parameter_201: name = "parameter_201" + original_name = "batch_norm_2.w_2" shape = [16] dtype = "float32" min_val = float("0") @@ -2134,6 +2336,7 @@ class Program_weight_tensor_parameter_201: class Program_weight_tensor_parameter_202: name = "parameter_202" + original_name = "batch_norm_2.w_1" shape = [16] dtype = "float32" min_val = float("0") @@ -2143,6 +2346,7 @@ class Program_weight_tensor_parameter_202: class Program_weight_tensor_parameter_203: name = "parameter_203" + original_name = "conv2d_2.w_0" shape = [16, 1, 3, 3] dtype = "float32" min_val = float("-1.95237") @@ -2154,6 +2358,7 @@ class Program_weight_tensor_parameter_203: class Program_weight_tensor_parameter_204: name = "parameter_204" + original_name = "batch_norm_1.b_0" shape = [16] dtype = "float32" min_val = float("0") @@ -2163,6 +2368,7 @@ class Program_weight_tensor_parameter_204: class Program_weight_tensor_parameter_205: name = "parameter_205" + original_name = "batch_norm_1.w_0" shape = [16] dtype = "float32" min_val = float("0") @@ -2172,6 +2378,7 @@ class Program_weight_tensor_parameter_205: class Program_weight_tensor_parameter_206: name = "parameter_206" + original_name = "batch_norm_1.w_2" shape = [16] dtype = "float32" min_val = float("0") @@ -2181,6 +2388,7 @@ class Program_weight_tensor_parameter_206: class Program_weight_tensor_parameter_207: name = "parameter_207" + original_name = "batch_norm_1.w_1" shape = [16] dtype = "float32" min_val = float("0") @@ -2190,6 +2398,7 @@ class Program_weight_tensor_parameter_207: class Program_weight_tensor_parameter_208: name = "parameter_208" + original_name = "conv2d_1.w_0" shape = [16, 16, 1, 1] dtype = "float32" min_val = float("-2.4205") @@ -2201,6 +2410,7 @@ class Program_weight_tensor_parameter_208: class Program_weight_tensor_parameter_209: name = "parameter_209" + original_name = "batch_norm_0.b_0" shape = [16] dtype = "float32" min_val = float("0") @@ -2210,6 +2420,7 @@ class Program_weight_tensor_parameter_209: class Program_weight_tensor_parameter_210: name = "parameter_210" + original_name = "batch_norm_0.w_0" shape = [16] dtype = "float32" min_val = float("0") @@ -2219,6 +2430,7 @@ class Program_weight_tensor_parameter_210: class Program_weight_tensor_parameter_211: name = "parameter_211" + original_name = "batch_norm_0.w_2" shape = [16] dtype = "float32" min_val = float("0") @@ -2228,6 +2440,7 @@ class Program_weight_tensor_parameter_211: class Program_weight_tensor_parameter_212: name = "parameter_212" + original_name = "batch_norm_0.w_1" shape = [16] dtype = "float32" min_val = float("0") @@ -2237,6 +2450,7 @@ class Program_weight_tensor_parameter_212: class Program_weight_tensor_parameter_213: name = "parameter_213" + original_name = "conv2d_0.w_0" shape = [16, 3, 3, 3] dtype = "float32" min_val = float("-3.03761") diff --git a/paddle_samples/PaddleX/MobileNetV3_small_x1_25/subgraph_0/input_meta.py b/paddle_samples/PaddleX/MobileNetV3_small_x1_25/subgraph_0/input_meta.py index 4c1d87b63..459245741 100644 --- a/paddle_samples/PaddleX/MobileNetV3_small_x1_25/subgraph_0/input_meta.py +++ b/paddle_samples/PaddleX/MobileNetV3_small_x1_25/subgraph_0/input_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_data_0: name = "data_0" + original_name = "var_161" shape = [252, 3, 224, 224] dtype = "float32" min_val = float("-2.1179") diff --git a/paddle_samples/PaddleX/MobileNetV3_small_x1_25/subgraph_0/weight_meta.py b/paddle_samples/PaddleX/MobileNetV3_small_x1_25/subgraph_0/weight_meta.py index 03933f9d7..842a1fd46 100644 --- a/paddle_samples/PaddleX/MobileNetV3_small_x1_25/subgraph_0/weight_meta.py +++ b/paddle_samples/PaddleX/MobileNetV3_small_x1_25/subgraph_0/weight_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_parameter_0: name = "parameter_0" + original_name = "linear_0.b_0" shape = [102] dtype = "float32" min_val = float("-0.000564915") @@ -11,6 +12,7 @@ class Program_weight_tensor_parameter_0: class Program_weight_tensor_parameter_1: name = "parameter_1" + original_name = "linear_0.w_0" shape = [1280, 102] dtype = "float32" min_val = float("-0.0661541") @@ -22,6 +24,7 @@ class Program_weight_tensor_parameter_1: class Program_weight_tensor_parameter_2: name = "parameter_2" + original_name = "conv2d_53.w_0" shape = [1280, 720, 1, 1] dtype = "float32" min_val = float("-0.437942") @@ -33,6 +36,7 @@ class Program_weight_tensor_parameter_2: class Program_weight_tensor_parameter_3: name = "parameter_3" + original_name = "batch_norm_34.b_0" shape = [720] dtype = "float32" min_val = float("-8.91355") @@ -44,6 +48,7 @@ class Program_weight_tensor_parameter_3: class Program_weight_tensor_parameter_4: name = "parameter_4" + original_name = "batch_norm_34.w_0" shape = [720] dtype = "float32" min_val = float("1.54916") @@ -55,6 +60,7 @@ class Program_weight_tensor_parameter_4: class Program_weight_tensor_parameter_5: name = "parameter_5" + original_name = "batch_norm_34.w_2" shape = [720] dtype = "float32" min_val = float("20.6431") @@ -66,6 +72,7 @@ class Program_weight_tensor_parameter_5: class Program_weight_tensor_parameter_6: name = "parameter_6" + original_name = "batch_norm_34.w_1" shape = [720] dtype = "float32" min_val = float("-2.1023e-05") @@ -77,6 +84,7 @@ class Program_weight_tensor_parameter_6: class Program_weight_tensor_parameter_7: name = "parameter_7" + original_name = "conv2d_52.w_0" shape = [720, 120, 1, 1] dtype = "float32" min_val = float("-0.740783") @@ -88,6 +96,7 @@ class Program_weight_tensor_parameter_7: class Program_weight_tensor_parameter_8: name = "parameter_8" + original_name = "batch_norm_33.b_0" shape = [120] dtype = "float32" min_val = float("-4.14127e-06") @@ -99,6 +108,7 @@ class Program_weight_tensor_parameter_8: class Program_weight_tensor_parameter_9: name = "parameter_9" + original_name = "batch_norm_33.w_0" shape = [120] dtype = "float32" min_val = float("1.8289") @@ -110,6 +120,7 @@ class Program_weight_tensor_parameter_9: class Program_weight_tensor_parameter_10: name = "parameter_10" + original_name = "batch_norm_33.w_2" shape = [120] dtype = "float32" min_val = float("0.0459655") @@ -121,6 +132,7 @@ class Program_weight_tensor_parameter_10: class Program_weight_tensor_parameter_11: name = "parameter_11" + original_name = "batch_norm_33.w_1" shape = [120] dtype = "float32" min_val = float("-0.409518") @@ -132,6 +144,7 @@ class Program_weight_tensor_parameter_11: class Program_weight_tensor_parameter_12: name = "parameter_12" + original_name = "conv2d_51.w_0" shape = [120, 720, 1, 1] dtype = "float32" min_val = float("-0.569827") @@ -143,6 +156,7 @@ class Program_weight_tensor_parameter_12: class Program_weight_tensor_parameter_13: name = "parameter_13" + original_name = "conv2d_50.b_0" shape = [720] dtype = "float32" min_val = float("-1.11454") @@ -154,6 +168,7 @@ class Program_weight_tensor_parameter_13: class Program_weight_tensor_parameter_14: name = "parameter_14" + original_name = "conv2d_50.w_0" shape = [720, 180, 1, 1] dtype = "float32" min_val = float("-0.580675") @@ -165,6 +180,7 @@ class Program_weight_tensor_parameter_14: class Program_weight_tensor_parameter_15: name = "parameter_15" + original_name = "conv2d_49.b_0" shape = [180] dtype = "float32" min_val = float("-0.0891055") @@ -176,6 +192,7 @@ class Program_weight_tensor_parameter_15: class Program_weight_tensor_parameter_16: name = "parameter_16" + original_name = "conv2d_49.w_0" shape = [180, 720, 1, 1] dtype = "float32" min_val = float("-0.346485") @@ -187,6 +204,7 @@ class Program_weight_tensor_parameter_16: class Program_weight_tensor_parameter_17: name = "parameter_17" + original_name = "batch_norm_32.b_0" shape = [720] dtype = "float32" min_val = float("-4.54869") @@ -198,6 +216,7 @@ class Program_weight_tensor_parameter_17: class Program_weight_tensor_parameter_18: name = "parameter_18" + original_name = "batch_norm_32.w_0" shape = [720] dtype = "float32" min_val = float("0.117211") @@ -209,6 +228,7 @@ class Program_weight_tensor_parameter_18: class Program_weight_tensor_parameter_19: name = "parameter_19" + original_name = "batch_norm_32.w_2" shape = [720] dtype = "float32" min_val = float("8.04682e-05") @@ -220,6 +240,7 @@ class Program_weight_tensor_parameter_19: class Program_weight_tensor_parameter_20: name = "parameter_20" + original_name = "batch_norm_32.w_1" shape = [720] dtype = "float32" min_val = float("-2.36345") @@ -231,6 +252,7 @@ class Program_weight_tensor_parameter_20: class Program_weight_tensor_parameter_21: name = "parameter_21" + original_name = "conv2d_48.w_0" shape = [720, 1, 5, 5] dtype = "float32" min_val = float("-0.669574") @@ -242,6 +264,7 @@ class Program_weight_tensor_parameter_21: class Program_weight_tensor_parameter_22: name = "parameter_22" + original_name = "batch_norm_31.b_0" shape = [720] dtype = "float32" min_val = float("-3.78383") @@ -253,6 +276,7 @@ class Program_weight_tensor_parameter_22: class Program_weight_tensor_parameter_23: name = "parameter_23" + original_name = "batch_norm_31.w_0" shape = [720] dtype = "float32" min_val = float("-0.92228") @@ -264,6 +288,7 @@ class Program_weight_tensor_parameter_23: class Program_weight_tensor_parameter_24: name = "parameter_24" + original_name = "batch_norm_31.w_2" shape = [720] dtype = "float32" min_val = float("4.4774") @@ -275,6 +300,7 @@ class Program_weight_tensor_parameter_24: class Program_weight_tensor_parameter_25: name = "parameter_25" + original_name = "batch_norm_31.w_1" shape = [720] dtype = "float32" min_val = float("-8.47465e-06") @@ -286,6 +312,7 @@ class Program_weight_tensor_parameter_25: class Program_weight_tensor_parameter_26: name = "parameter_26" + original_name = "conv2d_47.w_0" shape = [720, 120, 1, 1] dtype = "float32" min_val = float("-0.49802") @@ -297,6 +324,7 @@ class Program_weight_tensor_parameter_26: class Program_weight_tensor_parameter_27: name = "parameter_27" + original_name = "batch_norm_30.b_0" shape = [120] dtype = "float32" min_val = float("-3.77936e-06") @@ -308,6 +336,7 @@ class Program_weight_tensor_parameter_27: class Program_weight_tensor_parameter_28: name = "parameter_28" + original_name = "batch_norm_30.w_0" shape = [120] dtype = "float32" min_val = float("1.19334") @@ -319,6 +348,7 @@ class Program_weight_tensor_parameter_28: class Program_weight_tensor_parameter_29: name = "parameter_29" + original_name = "batch_norm_30.w_2" shape = [120] dtype = "float32" min_val = float("0.0648105") @@ -330,6 +360,7 @@ class Program_weight_tensor_parameter_29: class Program_weight_tensor_parameter_30: name = "parameter_30" + original_name = "batch_norm_30.w_1" shape = [120] dtype = "float32" min_val = float("-0.458339") @@ -341,6 +372,7 @@ class Program_weight_tensor_parameter_30: class Program_weight_tensor_parameter_31: name = "parameter_31" + original_name = "conv2d_46.w_0" shape = [120, 720, 1, 1] dtype = "float32" min_val = float("-0.463052") @@ -352,6 +384,7 @@ class Program_weight_tensor_parameter_31: class Program_weight_tensor_parameter_32: name = "parameter_32" + original_name = "conv2d_45.b_0" shape = [720] dtype = "float32" min_val = float("-0.248793") @@ -363,6 +396,7 @@ class Program_weight_tensor_parameter_32: class Program_weight_tensor_parameter_33: name = "parameter_33" + original_name = "conv2d_45.w_0" shape = [720, 180, 1, 1] dtype = "float32" min_val = float("-0.509203") @@ -374,6 +408,7 @@ class Program_weight_tensor_parameter_33: class Program_weight_tensor_parameter_34: name = "parameter_34" + original_name = "conv2d_44.b_0" shape = [180] dtype = "float32" min_val = float("-0.0540427") @@ -385,6 +420,7 @@ class Program_weight_tensor_parameter_34: class Program_weight_tensor_parameter_35: name = "parameter_35" + original_name = "conv2d_44.w_0" shape = [180, 720, 1, 1] dtype = "float32" min_val = float("-0.516009") @@ -396,6 +432,7 @@ class Program_weight_tensor_parameter_35: class Program_weight_tensor_parameter_36: name = "parameter_36" + original_name = "batch_norm_29.b_0" shape = [720] dtype = "float32" min_val = float("-4.33386") @@ -407,6 +444,7 @@ class Program_weight_tensor_parameter_36: class Program_weight_tensor_parameter_37: name = "parameter_37" + original_name = "batch_norm_29.w_0" shape = [720] dtype = "float32" min_val = float("0.29142") @@ -418,6 +456,7 @@ class Program_weight_tensor_parameter_37: class Program_weight_tensor_parameter_38: name = "parameter_38" + original_name = "batch_norm_29.w_2" shape = [720] dtype = "float32" min_val = float("0.000109091") @@ -429,6 +468,7 @@ class Program_weight_tensor_parameter_38: class Program_weight_tensor_parameter_39: name = "parameter_39" + original_name = "batch_norm_29.w_1" shape = [720] dtype = "float32" min_val = float("-2.97226") @@ -440,6 +480,7 @@ class Program_weight_tensor_parameter_39: class Program_weight_tensor_parameter_40: name = "parameter_40" + original_name = "conv2d_43.w_0" shape = [720, 1, 5, 5] dtype = "float32" min_val = float("-0.718452") @@ -451,6 +492,7 @@ class Program_weight_tensor_parameter_40: class Program_weight_tensor_parameter_41: name = "parameter_41" + original_name = "batch_norm_28.b_0" shape = [720] dtype = "float32" min_val = float("-3.22995") @@ -462,6 +504,7 @@ class Program_weight_tensor_parameter_41: class Program_weight_tensor_parameter_42: name = "parameter_42" + original_name = "batch_norm_28.w_0" shape = [720] dtype = "float32" min_val = float("-0.843472") @@ -473,6 +516,7 @@ class Program_weight_tensor_parameter_42: class Program_weight_tensor_parameter_43: name = "parameter_43" + original_name = "batch_norm_28.w_2" shape = [720] dtype = "float32" min_val = float("4.56524") @@ -484,6 +528,7 @@ class Program_weight_tensor_parameter_43: class Program_weight_tensor_parameter_44: name = "parameter_44" + original_name = "batch_norm_28.w_1" shape = [720] dtype = "float32" min_val = float("-4.2903e-06") @@ -495,6 +540,7 @@ class Program_weight_tensor_parameter_44: class Program_weight_tensor_parameter_45: name = "parameter_45" + original_name = "conv2d_42.w_0" shape = [720, 120, 1, 1] dtype = "float32" min_val = float("-0.517121") @@ -506,6 +552,7 @@ class Program_weight_tensor_parameter_45: class Program_weight_tensor_parameter_46: name = "parameter_46" + original_name = "batch_norm_27.b_0" shape = [120] dtype = "float32" min_val = float("-2.94702e-06") @@ -517,6 +564,7 @@ class Program_weight_tensor_parameter_46: class Program_weight_tensor_parameter_47: name = "parameter_47" + original_name = "batch_norm_27.w_0" shape = [120] dtype = "float32" min_val = float("2.20509") @@ -528,6 +576,7 @@ class Program_weight_tensor_parameter_47: class Program_weight_tensor_parameter_48: name = "parameter_48" + original_name = "batch_norm_27.w_2" shape = [120] dtype = "float32" min_val = float("0.0804236") @@ -539,6 +588,7 @@ class Program_weight_tensor_parameter_48: class Program_weight_tensor_parameter_49: name = "parameter_49" + original_name = "batch_norm_27.w_1" shape = [120] dtype = "float32" min_val = float("-0.776124") @@ -550,6 +600,7 @@ class Program_weight_tensor_parameter_49: class Program_weight_tensor_parameter_50: name = "parameter_50" + original_name = "conv2d_41.w_0" shape = [120, 360, 1, 1] dtype = "float32" min_val = float("-0.623831") @@ -561,6 +612,7 @@ class Program_weight_tensor_parameter_50: class Program_weight_tensor_parameter_51: name = "parameter_51" + original_name = "conv2d_40.b_0" shape = [360] dtype = "float32" min_val = float("-0.38039") @@ -572,6 +624,7 @@ class Program_weight_tensor_parameter_51: class Program_weight_tensor_parameter_52: name = "parameter_52" + original_name = "conv2d_40.w_0" shape = [360, 90, 1, 1] dtype = "float32" min_val = float("-0.801583") @@ -583,6 +636,7 @@ class Program_weight_tensor_parameter_52: class Program_weight_tensor_parameter_53: name = "parameter_53" + original_name = "conv2d_39.b_0" shape = [90] dtype = "float32" min_val = float("-0.037677") @@ -594,6 +648,7 @@ class Program_weight_tensor_parameter_53: class Program_weight_tensor_parameter_54: name = "parameter_54" + original_name = "conv2d_39.w_0" shape = [90, 360, 1, 1] dtype = "float32" min_val = float("-0.666341") @@ -605,6 +660,7 @@ class Program_weight_tensor_parameter_54: class Program_weight_tensor_parameter_55: name = "parameter_55" + original_name = "batch_norm_26.b_0" shape = [360] dtype = "float32" min_val = float("-1.57703") @@ -616,6 +672,7 @@ class Program_weight_tensor_parameter_55: class Program_weight_tensor_parameter_56: name = "parameter_56" + original_name = "batch_norm_26.w_0" shape = [360] dtype = "float32" min_val = float("-0.778081") @@ -627,6 +684,7 @@ class Program_weight_tensor_parameter_56: class Program_weight_tensor_parameter_57: name = "parameter_57" + original_name = "batch_norm_26.w_2" shape = [360] dtype = "float32" min_val = float("0.000159973") @@ -638,6 +696,7 @@ class Program_weight_tensor_parameter_57: class Program_weight_tensor_parameter_58: name = "parameter_58" + original_name = "batch_norm_26.w_1" shape = [360] dtype = "float32" min_val = float("-7.46841") @@ -649,6 +708,7 @@ class Program_weight_tensor_parameter_58: class Program_weight_tensor_parameter_59: name = "parameter_59" + original_name = "conv2d_38.w_0" shape = [360, 1, 5, 5] dtype = "float32" min_val = float("-0.461165") @@ -660,6 +720,7 @@ class Program_weight_tensor_parameter_59: class Program_weight_tensor_parameter_60: name = "parameter_60" + original_name = "batch_norm_25.b_0" shape = [360] dtype = "float32" min_val = float("-3.62349") @@ -671,6 +732,7 @@ class Program_weight_tensor_parameter_60: class Program_weight_tensor_parameter_61: name = "parameter_61" + original_name = "batch_norm_25.w_0" shape = [360] dtype = "float32" min_val = float("-0.916083") @@ -682,6 +744,7 @@ class Program_weight_tensor_parameter_61: class Program_weight_tensor_parameter_62: name = "parameter_62" + original_name = "batch_norm_25.w_2" shape = [360] dtype = "float32" min_val = float("14.0582") @@ -693,6 +756,7 @@ class Program_weight_tensor_parameter_62: class Program_weight_tensor_parameter_63: name = "parameter_63" + original_name = "batch_norm_25.w_1" shape = [360] dtype = "float32" min_val = float("-9.35827e-06") @@ -704,6 +768,7 @@ class Program_weight_tensor_parameter_63: class Program_weight_tensor_parameter_64: name = "parameter_64" + original_name = "conv2d_37.w_0" shape = [360, 64, 1, 1] dtype = "float32" min_val = float("-0.805866") @@ -715,6 +780,7 @@ class Program_weight_tensor_parameter_64: class Program_weight_tensor_parameter_65: name = "parameter_65" + original_name = "batch_norm_24.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -724,6 +790,7 @@ class Program_weight_tensor_parameter_65: class Program_weight_tensor_parameter_66: name = "parameter_66" + original_name = "batch_norm_24.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -733,6 +800,7 @@ class Program_weight_tensor_parameter_66: class Program_weight_tensor_parameter_67: name = "parameter_67" + original_name = "batch_norm_24.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -742,6 +810,7 @@ class Program_weight_tensor_parameter_67: class Program_weight_tensor_parameter_68: name = "parameter_68" + original_name = "batch_norm_24.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -751,6 +820,7 @@ class Program_weight_tensor_parameter_68: class Program_weight_tensor_parameter_69: name = "parameter_69" + original_name = "conv2d_36.w_0" shape = [64, 184, 1, 1] dtype = "float32" min_val = float("-0.51435") @@ -762,6 +832,7 @@ class Program_weight_tensor_parameter_69: class Program_weight_tensor_parameter_70: name = "parameter_70" + original_name = "conv2d_35.b_0" shape = [184] dtype = "float32" min_val = float("-0.273427") @@ -773,6 +844,7 @@ class Program_weight_tensor_parameter_70: class Program_weight_tensor_parameter_71: name = "parameter_71" + original_name = "conv2d_35.w_0" shape = [184, 46, 1, 1] dtype = "float32" min_val = float("-0.75633") @@ -784,6 +856,7 @@ class Program_weight_tensor_parameter_71: class Program_weight_tensor_parameter_72: name = "parameter_72" + original_name = "conv2d_34.b_0" shape = [46] dtype = "float32" min_val = float("0") @@ -793,6 +866,7 @@ class Program_weight_tensor_parameter_72: class Program_weight_tensor_parameter_73: name = "parameter_73" + original_name = "conv2d_34.w_0" shape = [46, 184, 1, 1] dtype = "float32" min_val = float("-0.840016") @@ -804,6 +878,7 @@ class Program_weight_tensor_parameter_73: class Program_weight_tensor_parameter_74: name = "parameter_74" + original_name = "batch_norm_23.b_0" shape = [184] dtype = "float32" min_val = float("-4.6796") @@ -815,6 +890,7 @@ class Program_weight_tensor_parameter_74: class Program_weight_tensor_parameter_75: name = "parameter_75" + original_name = "batch_norm_23.w_0" shape = [184] dtype = "float32" min_val = float("0.304559") @@ -826,6 +902,7 @@ class Program_weight_tensor_parameter_75: class Program_weight_tensor_parameter_76: name = "parameter_76" + original_name = "batch_norm_23.w_2" shape = [184] dtype = "float32" min_val = float("0.000149817") @@ -837,6 +914,7 @@ class Program_weight_tensor_parameter_76: class Program_weight_tensor_parameter_77: name = "parameter_77" + original_name = "batch_norm_23.w_1" shape = [184] dtype = "float32" min_val = float("-2.72082") @@ -848,6 +926,7 @@ class Program_weight_tensor_parameter_77: class Program_weight_tensor_parameter_78: name = "parameter_78" + original_name = "conv2d_33.w_0" shape = [184, 1, 5, 5] dtype = "float32" min_val = float("-0.718155") @@ -859,6 +938,7 @@ class Program_weight_tensor_parameter_78: class Program_weight_tensor_parameter_79: name = "parameter_79" + original_name = "batch_norm_22.b_0" shape = [184] dtype = "float32" min_val = float("-2.70614") @@ -870,6 +950,7 @@ class Program_weight_tensor_parameter_79: class Program_weight_tensor_parameter_80: name = "parameter_80" + original_name = "batch_norm_22.w_0" shape = [184] dtype = "float32" min_val = float("-0.513869") @@ -881,6 +962,7 @@ class Program_weight_tensor_parameter_80: class Program_weight_tensor_parameter_81: name = "parameter_81" + original_name = "batch_norm_22.w_2" shape = [184] dtype = "float32" min_val = float("8.25469") @@ -892,6 +974,7 @@ class Program_weight_tensor_parameter_81: class Program_weight_tensor_parameter_82: name = "parameter_82" + original_name = "batch_norm_22.w_1" shape = [184] dtype = "float32" min_val = float("-3.75064e-06") @@ -903,6 +986,7 @@ class Program_weight_tensor_parameter_82: class Program_weight_tensor_parameter_83: name = "parameter_83" + original_name = "conv2d_32.w_0" shape = [184, 64, 1, 1] dtype = "float32" min_val = float("-0.604") @@ -914,6 +998,7 @@ class Program_weight_tensor_parameter_83: class Program_weight_tensor_parameter_84: name = "parameter_84" + original_name = "batch_norm_21.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -923,6 +1008,7 @@ class Program_weight_tensor_parameter_84: class Program_weight_tensor_parameter_85: name = "parameter_85" + original_name = "batch_norm_21.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -932,6 +1018,7 @@ class Program_weight_tensor_parameter_85: class Program_weight_tensor_parameter_86: name = "parameter_86" + original_name = "batch_norm_21.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -941,6 +1028,7 @@ class Program_weight_tensor_parameter_86: class Program_weight_tensor_parameter_87: name = "parameter_87" + original_name = "batch_norm_21.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -950,6 +1038,7 @@ class Program_weight_tensor_parameter_87: class Program_weight_tensor_parameter_88: name = "parameter_88" + original_name = "conv2d_31.w_0" shape = [64, 152, 1, 1] dtype = "float32" min_val = float("-0.764386") @@ -961,6 +1050,7 @@ class Program_weight_tensor_parameter_88: class Program_weight_tensor_parameter_89: name = "parameter_89" + original_name = "conv2d_30.b_0" shape = [152] dtype = "float32" min_val = float("-0.434361") @@ -972,6 +1062,7 @@ class Program_weight_tensor_parameter_89: class Program_weight_tensor_parameter_90: name = "parameter_90" + original_name = "conv2d_30.w_0" shape = [152, 38, 1, 1] dtype = "float32" min_val = float("-2.1721e-10") @@ -983,6 +1074,7 @@ class Program_weight_tensor_parameter_90: class Program_weight_tensor_parameter_91: name = "parameter_91" + original_name = "conv2d_29.b_0" shape = [38] dtype = "float32" min_val = float("0") @@ -992,6 +1084,7 @@ class Program_weight_tensor_parameter_91: class Program_weight_tensor_parameter_92: name = "parameter_92" + original_name = "conv2d_29.w_0" shape = [38, 152, 1, 1] dtype = "float32" min_val = float("-1.6868e-10") @@ -1003,6 +1096,7 @@ class Program_weight_tensor_parameter_92: class Program_weight_tensor_parameter_93: name = "parameter_93" + original_name = "batch_norm_20.b_0" shape = [152] dtype = "float32" min_val = float("-1.54243") @@ -1014,6 +1108,7 @@ class Program_weight_tensor_parameter_93: class Program_weight_tensor_parameter_94: name = "parameter_94" + original_name = "batch_norm_20.w_0" shape = [152] dtype = "float32" min_val = float("0.459492") @@ -1025,6 +1120,7 @@ class Program_weight_tensor_parameter_94: class Program_weight_tensor_parameter_95: name = "parameter_95" + original_name = "batch_norm_20.w_2" shape = [152] dtype = "float32" min_val = float("0.000214618") @@ -1036,6 +1132,7 @@ class Program_weight_tensor_parameter_95: class Program_weight_tensor_parameter_96: name = "parameter_96" + original_name = "batch_norm_20.w_1" shape = [152] dtype = "float32" min_val = float("-15.3893") @@ -1047,6 +1144,7 @@ class Program_weight_tensor_parameter_96: class Program_weight_tensor_parameter_97: name = "parameter_97" + original_name = "conv2d_28.w_0" shape = [152, 1, 5, 5] dtype = "float32" min_val = float("-1.48243") @@ -1058,6 +1156,7 @@ class Program_weight_tensor_parameter_97: class Program_weight_tensor_parameter_98: name = "parameter_98" + original_name = "batch_norm_19.b_0" shape = [152] dtype = "float32" min_val = float("-1.43233") @@ -1069,6 +1168,7 @@ class Program_weight_tensor_parameter_98: class Program_weight_tensor_parameter_99: name = "parameter_99" + original_name = "batch_norm_19.w_0" shape = [152] dtype = "float32" min_val = float("0.0306071") @@ -1080,6 +1180,7 @@ class Program_weight_tensor_parameter_99: class Program_weight_tensor_parameter_100: name = "parameter_100" + original_name = "batch_norm_19.w_2" shape = [152] dtype = "float32" min_val = float("14.0016") @@ -1091,6 +1192,7 @@ class Program_weight_tensor_parameter_100: class Program_weight_tensor_parameter_101: name = "parameter_101" + original_name = "batch_norm_19.w_1" shape = [152] dtype = "float32" min_val = float("-7.16259e-06") @@ -1102,6 +1204,7 @@ class Program_weight_tensor_parameter_101: class Program_weight_tensor_parameter_102: name = "parameter_102" + original_name = "conv2d_27.w_0" shape = [152, 48, 1, 1] dtype = "float32" min_val = float("-0.867461") @@ -1113,6 +1216,7 @@ class Program_weight_tensor_parameter_102: class Program_weight_tensor_parameter_103: name = "parameter_103" + original_name = "batch_norm_18.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -1122,6 +1226,7 @@ class Program_weight_tensor_parameter_103: class Program_weight_tensor_parameter_104: name = "parameter_104" + original_name = "batch_norm_18.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -1131,6 +1236,7 @@ class Program_weight_tensor_parameter_104: class Program_weight_tensor_parameter_105: name = "parameter_105" + original_name = "batch_norm_18.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -1140,6 +1246,7 @@ class Program_weight_tensor_parameter_105: class Program_weight_tensor_parameter_106: name = "parameter_106" + original_name = "batch_norm_18.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -1149,6 +1256,7 @@ class Program_weight_tensor_parameter_106: class Program_weight_tensor_parameter_107: name = "parameter_107" + original_name = "conv2d_26.w_0" shape = [48, 304, 1, 1] dtype = "float32" min_val = float("-0.65836") @@ -1160,6 +1268,7 @@ class Program_weight_tensor_parameter_107: class Program_weight_tensor_parameter_108: name = "parameter_108" + original_name = "conv2d_25.b_0" shape = [304] dtype = "float32" min_val = float("-0.19574") @@ -1171,6 +1280,7 @@ class Program_weight_tensor_parameter_108: class Program_weight_tensor_parameter_109: name = "parameter_109" + original_name = "conv2d_25.w_0" shape = [304, 76, 1, 1] dtype = "float32" min_val = float("-0.677007") @@ -1182,6 +1292,7 @@ class Program_weight_tensor_parameter_109: class Program_weight_tensor_parameter_110: name = "parameter_110" + original_name = "conv2d_24.b_0" shape = [76] dtype = "float32" min_val = float("-0.0568169") @@ -1193,6 +1304,7 @@ class Program_weight_tensor_parameter_110: class Program_weight_tensor_parameter_111: name = "parameter_111" + original_name = "conv2d_24.w_0" shape = [76, 304, 1, 1] dtype = "float32" min_val = float("-0.784074") @@ -1204,6 +1316,7 @@ class Program_weight_tensor_parameter_111: class Program_weight_tensor_parameter_112: name = "parameter_112" + original_name = "batch_norm_17.b_0" shape = [304] dtype = "float32" min_val = float("-3.57414") @@ -1215,6 +1328,7 @@ class Program_weight_tensor_parameter_112: class Program_weight_tensor_parameter_113: name = "parameter_113" + original_name = "batch_norm_17.w_0" shape = [304] dtype = "float32" min_val = float("0.105077") @@ -1226,6 +1340,7 @@ class Program_weight_tensor_parameter_113: class Program_weight_tensor_parameter_114: name = "parameter_114" + original_name = "batch_norm_17.w_2" shape = [304] dtype = "float32" min_val = float("8.77487e-05") @@ -1237,6 +1352,7 @@ class Program_weight_tensor_parameter_114: class Program_weight_tensor_parameter_115: name = "parameter_115" + original_name = "batch_norm_17.w_1" shape = [304] dtype = "float32" min_val = float("-6.64572") @@ -1248,6 +1364,7 @@ class Program_weight_tensor_parameter_115: class Program_weight_tensor_parameter_116: name = "parameter_116" + original_name = "conv2d_23.w_0" shape = [304, 1, 5, 5] dtype = "float32" min_val = float("-0.788266") @@ -1259,6 +1376,7 @@ class Program_weight_tensor_parameter_116: class Program_weight_tensor_parameter_117: name = "parameter_117" + original_name = "batch_norm_16.b_0" shape = [304] dtype = "float32" min_val = float("-2.55454") @@ -1270,6 +1388,7 @@ class Program_weight_tensor_parameter_117: class Program_weight_tensor_parameter_118: name = "parameter_118" + original_name = "batch_norm_16.w_0" shape = [304] dtype = "float32" min_val = float("-1.63364") @@ -1281,6 +1400,7 @@ class Program_weight_tensor_parameter_118: class Program_weight_tensor_parameter_119: name = "parameter_119" + original_name = "batch_norm_16.w_2" shape = [304] dtype = "float32" min_val = float("8.88593") @@ -1292,6 +1412,7 @@ class Program_weight_tensor_parameter_119: class Program_weight_tensor_parameter_120: name = "parameter_120" + original_name = "batch_norm_16.w_1" shape = [304] dtype = "float32" min_val = float("-3.65139e-06") @@ -1303,6 +1424,7 @@ class Program_weight_tensor_parameter_120: class Program_weight_tensor_parameter_121: name = "parameter_121" + original_name = "conv2d_22.w_0" shape = [304, 48, 1, 1] dtype = "float32" min_val = float("-0.627049") @@ -1314,6 +1436,7 @@ class Program_weight_tensor_parameter_121: class Program_weight_tensor_parameter_122: name = "parameter_122" + original_name = "batch_norm_15.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -1323,6 +1446,7 @@ class Program_weight_tensor_parameter_122: class Program_weight_tensor_parameter_123: name = "parameter_123" + original_name = "batch_norm_15.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -1332,6 +1456,7 @@ class Program_weight_tensor_parameter_123: class Program_weight_tensor_parameter_124: name = "parameter_124" + original_name = "batch_norm_15.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -1341,6 +1466,7 @@ class Program_weight_tensor_parameter_124: class Program_weight_tensor_parameter_125: name = "parameter_125" + original_name = "batch_norm_15.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -1350,6 +1476,7 @@ class Program_weight_tensor_parameter_125: class Program_weight_tensor_parameter_126: name = "parameter_126" + original_name = "conv2d_21.w_0" shape = [48, 304, 1, 1] dtype = "float32" min_val = float("-0.764018") @@ -1361,6 +1488,7 @@ class Program_weight_tensor_parameter_126: class Program_weight_tensor_parameter_127: name = "parameter_127" + original_name = "conv2d_20.b_0" shape = [304] dtype = "float32" min_val = float("-0.225229") @@ -1372,6 +1500,7 @@ class Program_weight_tensor_parameter_127: class Program_weight_tensor_parameter_128: name = "parameter_128" + original_name = "conv2d_20.w_0" shape = [304, 76, 1, 1] dtype = "float32" min_val = float("-0.611921") @@ -1383,6 +1512,7 @@ class Program_weight_tensor_parameter_128: class Program_weight_tensor_parameter_129: name = "parameter_129" + original_name = "conv2d_19.b_0" shape = [76] dtype = "float32" min_val = float("-0.0642324") @@ -1394,6 +1524,7 @@ class Program_weight_tensor_parameter_129: class Program_weight_tensor_parameter_130: name = "parameter_130" + original_name = "conv2d_19.w_0" shape = [76, 304, 1, 1] dtype = "float32" min_val = float("-0.721367") @@ -1405,6 +1536,7 @@ class Program_weight_tensor_parameter_130: class Program_weight_tensor_parameter_131: name = "parameter_131" + original_name = "batch_norm_14.b_0" shape = [304] dtype = "float32" min_val = float("-2.97452") @@ -1416,6 +1548,7 @@ class Program_weight_tensor_parameter_131: class Program_weight_tensor_parameter_132: name = "parameter_132" + original_name = "batch_norm_14.w_0" shape = [304] dtype = "float32" min_val = float("0.254985") @@ -1427,6 +1560,7 @@ class Program_weight_tensor_parameter_132: class Program_weight_tensor_parameter_133: name = "parameter_133" + original_name = "batch_norm_14.w_2" shape = [304] dtype = "float32" min_val = float("0.000128187") @@ -1438,6 +1572,7 @@ class Program_weight_tensor_parameter_133: class Program_weight_tensor_parameter_134: name = "parameter_134" + original_name = "batch_norm_14.w_1" shape = [304] dtype = "float32" min_val = float("-17.8536") @@ -1449,6 +1584,7 @@ class Program_weight_tensor_parameter_134: class Program_weight_tensor_parameter_135: name = "parameter_135" + original_name = "conv2d_18.w_0" shape = [304, 1, 5, 5] dtype = "float32" min_val = float("-0.973936") @@ -1460,6 +1596,7 @@ class Program_weight_tensor_parameter_135: class Program_weight_tensor_parameter_136: name = "parameter_136" + original_name = "batch_norm_13.b_0" shape = [304] dtype = "float32" min_val = float("-2.73801") @@ -1471,6 +1608,7 @@ class Program_weight_tensor_parameter_136: class Program_weight_tensor_parameter_137: name = "parameter_137" + original_name = "batch_norm_13.w_0" shape = [304] dtype = "float32" min_val = float("-0.156431") @@ -1482,6 +1620,7 @@ class Program_weight_tensor_parameter_137: class Program_weight_tensor_parameter_138: name = "parameter_138" + original_name = "batch_norm_13.w_2" shape = [304] dtype = "float32" min_val = float("5.76018") @@ -1493,6 +1632,7 @@ class Program_weight_tensor_parameter_138: class Program_weight_tensor_parameter_139: name = "parameter_139" + original_name = "batch_norm_13.w_1" shape = [304] dtype = "float32" min_val = float("-2.33887e-06") @@ -1504,6 +1644,7 @@ class Program_weight_tensor_parameter_139: class Program_weight_tensor_parameter_140: name = "parameter_140" + original_name = "conv2d_17.w_0" shape = [304, 48, 1, 1] dtype = "float32" min_val = float("-0.675324") @@ -1515,6 +1656,7 @@ class Program_weight_tensor_parameter_140: class Program_weight_tensor_parameter_141: name = "parameter_141" + original_name = "batch_norm_12.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -1524,6 +1666,7 @@ class Program_weight_tensor_parameter_141: class Program_weight_tensor_parameter_142: name = "parameter_142" + original_name = "batch_norm_12.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -1533,6 +1676,7 @@ class Program_weight_tensor_parameter_142: class Program_weight_tensor_parameter_143: name = "parameter_143" + original_name = "batch_norm_12.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -1542,6 +1686,7 @@ class Program_weight_tensor_parameter_143: class Program_weight_tensor_parameter_144: name = "parameter_144" + original_name = "batch_norm_12.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -1551,6 +1696,7 @@ class Program_weight_tensor_parameter_144: class Program_weight_tensor_parameter_145: name = "parameter_145" + original_name = "conv2d_16.w_0" shape = [48, 120, 1, 1] dtype = "float32" min_val = float("-0.896881") @@ -1562,6 +1708,7 @@ class Program_weight_tensor_parameter_145: class Program_weight_tensor_parameter_146: name = "parameter_146" + original_name = "conv2d_15.b_0" shape = [120] dtype = "float32" min_val = float("0.00440334") @@ -1573,6 +1720,7 @@ class Program_weight_tensor_parameter_146: class Program_weight_tensor_parameter_147: name = "parameter_147" + original_name = "conv2d_15.w_0" shape = [120, 30, 1, 1] dtype = "float32" min_val = float("-4.77811e-12") @@ -1584,6 +1732,7 @@ class Program_weight_tensor_parameter_147: class Program_weight_tensor_parameter_148: name = "parameter_148" + original_name = "conv2d_14.b_0" shape = [30] dtype = "float32" min_val = float("0") @@ -1593,6 +1742,7 @@ class Program_weight_tensor_parameter_148: class Program_weight_tensor_parameter_149: name = "parameter_149" + original_name = "conv2d_14.w_0" shape = [30, 120, 1, 1] dtype = "float32" min_val = float("-0.00324109") @@ -1604,6 +1754,7 @@ class Program_weight_tensor_parameter_149: class Program_weight_tensor_parameter_150: name = "parameter_150" + original_name = "batch_norm_11.b_0" shape = [120] dtype = "float32" min_val = float("-1.39434") @@ -1615,6 +1766,7 @@ class Program_weight_tensor_parameter_150: class Program_weight_tensor_parameter_151: name = "parameter_151" + original_name = "batch_norm_11.w_0" shape = [120] dtype = "float32" min_val = float("0.528989") @@ -1626,6 +1778,7 @@ class Program_weight_tensor_parameter_151: class Program_weight_tensor_parameter_152: name = "parameter_152" + original_name = "batch_norm_11.w_2" shape = [120] dtype = "float32" min_val = float("0.000406801") @@ -1637,6 +1790,7 @@ class Program_weight_tensor_parameter_152: class Program_weight_tensor_parameter_153: name = "parameter_153" + original_name = "batch_norm_11.w_1" shape = [120] dtype = "float32" min_val = float("-9.27086") @@ -1648,6 +1802,7 @@ class Program_weight_tensor_parameter_153: class Program_weight_tensor_parameter_154: name = "parameter_154" + original_name = "conv2d_13.w_0" shape = [120, 1, 5, 5] dtype = "float32" min_val = float("-0.560434") @@ -1659,6 +1814,7 @@ class Program_weight_tensor_parameter_154: class Program_weight_tensor_parameter_155: name = "parameter_155" + original_name = "batch_norm_10.b_0" shape = [120] dtype = "float32" min_val = float("-2.62925") @@ -1670,6 +1826,7 @@ class Program_weight_tensor_parameter_155: class Program_weight_tensor_parameter_156: name = "parameter_156" + original_name = "batch_norm_10.w_0" shape = [120] dtype = "float32" min_val = float("0.0286211") @@ -1681,6 +1838,7 @@ class Program_weight_tensor_parameter_156: class Program_weight_tensor_parameter_157: name = "parameter_157" + original_name = "batch_norm_10.w_2" shape = [120] dtype = "float32" min_val = float("11.0363") @@ -1692,6 +1850,7 @@ class Program_weight_tensor_parameter_157: class Program_weight_tensor_parameter_158: name = "parameter_158" + original_name = "batch_norm_10.w_1" shape = [120] dtype = "float32" min_val = float("-3.2851e-05") @@ -1703,6 +1862,7 @@ class Program_weight_tensor_parameter_158: class Program_weight_tensor_parameter_159: name = "parameter_159" + original_name = "conv2d_12.w_0" shape = [120, 32, 1, 1] dtype = "float32" min_val = float("-1.3847") @@ -1714,6 +1874,7 @@ class Program_weight_tensor_parameter_159: class Program_weight_tensor_parameter_160: name = "parameter_160" + original_name = "batch_norm_9.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -1723,6 +1884,7 @@ class Program_weight_tensor_parameter_160: class Program_weight_tensor_parameter_161: name = "parameter_161" + original_name = "batch_norm_9.w_0" shape = [32] dtype = "float32" min_val = float("0") @@ -1732,6 +1894,7 @@ class Program_weight_tensor_parameter_161: class Program_weight_tensor_parameter_162: name = "parameter_162" + original_name = "batch_norm_9.w_2" shape = [32] dtype = "float32" min_val = float("0") @@ -1741,6 +1904,7 @@ class Program_weight_tensor_parameter_162: class Program_weight_tensor_parameter_163: name = "parameter_163" + original_name = "batch_norm_9.w_1" shape = [32] dtype = "float32" min_val = float("0") @@ -1750,6 +1914,7 @@ class Program_weight_tensor_parameter_163: class Program_weight_tensor_parameter_164: name = "parameter_164" + original_name = "conv2d_11.w_0" shape = [32, 112, 1, 1] dtype = "float32" min_val = float("-0.983969") @@ -1761,6 +1926,7 @@ class Program_weight_tensor_parameter_164: class Program_weight_tensor_parameter_165: name = "parameter_165" + original_name = "batch_norm_8.b_0" shape = [112] dtype = "float32" min_val = float("-3.74459") @@ -1772,6 +1938,7 @@ class Program_weight_tensor_parameter_165: class Program_weight_tensor_parameter_166: name = "parameter_166" + original_name = "batch_norm_8.w_0" shape = [112] dtype = "float32" min_val = float("1.00009") @@ -1783,6 +1950,7 @@ class Program_weight_tensor_parameter_166: class Program_weight_tensor_parameter_167: name = "parameter_167" + original_name = "batch_norm_8.w_2" shape = [112] dtype = "float32" min_val = float("0.0199893") @@ -1794,6 +1962,7 @@ class Program_weight_tensor_parameter_167: class Program_weight_tensor_parameter_168: name = "parameter_168" + original_name = "batch_norm_8.w_1" shape = [112] dtype = "float32" min_val = float("-6.19623") @@ -1805,6 +1974,7 @@ class Program_weight_tensor_parameter_168: class Program_weight_tensor_parameter_169: name = "parameter_169" + original_name = "conv2d_10.w_0" shape = [112, 1, 3, 3] dtype = "float32" min_val = float("-1.34511") @@ -1816,6 +1986,7 @@ class Program_weight_tensor_parameter_169: class Program_weight_tensor_parameter_170: name = "parameter_170" + original_name = "batch_norm_7.b_0" shape = [112] dtype = "float32" min_val = float("-2.52403") @@ -1827,6 +1998,7 @@ class Program_weight_tensor_parameter_170: class Program_weight_tensor_parameter_171: name = "parameter_171" + original_name = "batch_norm_7.w_0" shape = [112] dtype = "float32" min_val = float("0.559407") @@ -1838,6 +2010,7 @@ class Program_weight_tensor_parameter_171: class Program_weight_tensor_parameter_172: name = "parameter_172" + original_name = "batch_norm_7.w_2" shape = [112] dtype = "float32" min_val = float("5.14614") @@ -1849,6 +2022,7 @@ class Program_weight_tensor_parameter_172: class Program_weight_tensor_parameter_173: name = "parameter_173" + original_name = "batch_norm_7.w_1" shape = [112] dtype = "float32" min_val = float("-1.80583e-05") @@ -1860,6 +2034,7 @@ class Program_weight_tensor_parameter_173: class Program_weight_tensor_parameter_174: name = "parameter_174" + original_name = "conv2d_9.w_0" shape = [112, 32, 1, 1] dtype = "float32" min_val = float("-0.933187") @@ -1871,6 +2046,7 @@ class Program_weight_tensor_parameter_174: class Program_weight_tensor_parameter_175: name = "parameter_175" + original_name = "batch_norm_6.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -1880,6 +2056,7 @@ class Program_weight_tensor_parameter_175: class Program_weight_tensor_parameter_176: name = "parameter_176" + original_name = "batch_norm_6.w_0" shape = [32] dtype = "float32" min_val = float("0") @@ -1889,6 +2066,7 @@ class Program_weight_tensor_parameter_176: class Program_weight_tensor_parameter_177: name = "parameter_177" + original_name = "batch_norm_6.w_2" shape = [32] dtype = "float32" min_val = float("0") @@ -1898,6 +2076,7 @@ class Program_weight_tensor_parameter_177: class Program_weight_tensor_parameter_178: name = "parameter_178" + original_name = "batch_norm_6.w_1" shape = [32] dtype = "float32" min_val = float("0") @@ -1907,6 +2086,7 @@ class Program_weight_tensor_parameter_178: class Program_weight_tensor_parameter_179: name = "parameter_179" + original_name = "conv2d_8.w_0" shape = [32, 88, 1, 1] dtype = "float32" min_val = float("-1.38126") @@ -1918,6 +2098,7 @@ class Program_weight_tensor_parameter_179: class Program_weight_tensor_parameter_180: name = "parameter_180" + original_name = "batch_norm_5.b_0" shape = [88] dtype = "float32" min_val = float("-5.56045") @@ -1929,6 +2110,7 @@ class Program_weight_tensor_parameter_180: class Program_weight_tensor_parameter_181: name = "parameter_181" + original_name = "batch_norm_5.w_0" shape = [88] dtype = "float32" min_val = float("1.29995") @@ -1940,6 +2122,7 @@ class Program_weight_tensor_parameter_181: class Program_weight_tensor_parameter_182: name = "parameter_182" + original_name = "batch_norm_5.w_2" shape = [88] dtype = "float32" min_val = float("0.119252") @@ -1951,6 +2134,7 @@ class Program_weight_tensor_parameter_182: class Program_weight_tensor_parameter_183: name = "parameter_183" + original_name = "batch_norm_5.w_1" shape = [88] dtype = "float32" min_val = float("-5.49727") @@ -1962,6 +2146,7 @@ class Program_weight_tensor_parameter_183: class Program_weight_tensor_parameter_184: name = "parameter_184" + original_name = "conv2d_7.w_0" shape = [88, 1, 3, 3] dtype = "float32" min_val = float("-0.996255") @@ -1973,6 +2158,7 @@ class Program_weight_tensor_parameter_184: class Program_weight_tensor_parameter_185: name = "parameter_185" + original_name = "batch_norm_4.b_0" shape = [88] dtype = "float32" min_val = float("-3.40537") @@ -1984,6 +2170,7 @@ class Program_weight_tensor_parameter_185: class Program_weight_tensor_parameter_186: name = "parameter_186" + original_name = "batch_norm_4.w_0" shape = [88] dtype = "float32" min_val = float("0.315448") @@ -1995,6 +2182,7 @@ class Program_weight_tensor_parameter_186: class Program_weight_tensor_parameter_187: name = "parameter_187" + original_name = "batch_norm_4.w_2" shape = [88] dtype = "float32" min_val = float("18.578") @@ -2006,6 +2194,7 @@ class Program_weight_tensor_parameter_187: class Program_weight_tensor_parameter_188: name = "parameter_188" + original_name = "batch_norm_4.w_1" shape = [88] dtype = "float32" min_val = float("-6.35897e-05") @@ -2017,6 +2206,7 @@ class Program_weight_tensor_parameter_188: class Program_weight_tensor_parameter_189: name = "parameter_189" + original_name = "conv2d_6.w_0" shape = [88, 24, 1, 1] dtype = "float32" min_val = float("-1.35631") @@ -2028,6 +2218,7 @@ class Program_weight_tensor_parameter_189: class Program_weight_tensor_parameter_190: name = "parameter_190" + original_name = "batch_norm_3.b_0" shape = [24] dtype = "float32" min_val = float("0") @@ -2037,6 +2228,7 @@ class Program_weight_tensor_parameter_190: class Program_weight_tensor_parameter_191: name = "parameter_191" + original_name = "batch_norm_3.w_0" shape = [24] dtype = "float32" min_val = float("0") @@ -2046,6 +2238,7 @@ class Program_weight_tensor_parameter_191: class Program_weight_tensor_parameter_192: name = "parameter_192" + original_name = "batch_norm_3.w_2" shape = [24] dtype = "float32" min_val = float("0") @@ -2055,6 +2248,7 @@ class Program_weight_tensor_parameter_192: class Program_weight_tensor_parameter_193: name = "parameter_193" + original_name = "batch_norm_3.w_1" shape = [24] dtype = "float32" min_val = float("0") @@ -2064,6 +2258,7 @@ class Program_weight_tensor_parameter_193: class Program_weight_tensor_parameter_194: name = "parameter_194" + original_name = "conv2d_5.w_0" shape = [24, 24, 1, 1] dtype = "float32" min_val = float("-1.52298") @@ -2075,6 +2270,7 @@ class Program_weight_tensor_parameter_194: class Program_weight_tensor_parameter_195: name = "parameter_195" + original_name = "conv2d_4.b_0" shape = [24] dtype = "float32" min_val = float("0") @@ -2084,6 +2280,7 @@ class Program_weight_tensor_parameter_195: class Program_weight_tensor_parameter_196: name = "parameter_196" + original_name = "conv2d_4.w_0" shape = [24, 6, 1, 1] dtype = "float32" min_val = float("-2.14411e-13") @@ -2095,6 +2292,7 @@ class Program_weight_tensor_parameter_196: class Program_weight_tensor_parameter_197: name = "parameter_197" + original_name = "conv2d_3.b_0" shape = [6] dtype = "float32" min_val = float("0") @@ -2104,6 +2302,7 @@ class Program_weight_tensor_parameter_197: class Program_weight_tensor_parameter_198: name = "parameter_198" + original_name = "conv2d_3.w_0" shape = [6, 24, 1, 1] dtype = "float32" min_val = float("-2.20555e-13") @@ -2115,6 +2314,7 @@ class Program_weight_tensor_parameter_198: class Program_weight_tensor_parameter_199: name = "parameter_199" + original_name = "batch_norm_2.b_0" shape = [24] dtype = "float32" min_val = float("0") @@ -2124,6 +2324,7 @@ class Program_weight_tensor_parameter_199: class Program_weight_tensor_parameter_200: name = "parameter_200" + original_name = "batch_norm_2.w_0" shape = [24] dtype = "float32" min_val = float("0") @@ -2133,6 +2334,7 @@ class Program_weight_tensor_parameter_200: class Program_weight_tensor_parameter_201: name = "parameter_201" + original_name = "batch_norm_2.w_2" shape = [24] dtype = "float32" min_val = float("0") @@ -2142,6 +2344,7 @@ class Program_weight_tensor_parameter_201: class Program_weight_tensor_parameter_202: name = "parameter_202" + original_name = "batch_norm_2.w_1" shape = [24] dtype = "float32" min_val = float("0") @@ -2151,6 +2354,7 @@ class Program_weight_tensor_parameter_202: class Program_weight_tensor_parameter_203: name = "parameter_203" + original_name = "conv2d_2.w_0" shape = [24, 1, 3, 3] dtype = "float32" min_val = float("-2.78632") @@ -2162,6 +2366,7 @@ class Program_weight_tensor_parameter_203: class Program_weight_tensor_parameter_204: name = "parameter_204" + original_name = "batch_norm_1.b_0" shape = [24] dtype = "float32" min_val = float("0") @@ -2171,6 +2376,7 @@ class Program_weight_tensor_parameter_204: class Program_weight_tensor_parameter_205: name = "parameter_205" + original_name = "batch_norm_1.w_0" shape = [24] dtype = "float32" min_val = float("0") @@ -2180,6 +2386,7 @@ class Program_weight_tensor_parameter_205: class Program_weight_tensor_parameter_206: name = "parameter_206" + original_name = "batch_norm_1.w_2" shape = [24] dtype = "float32" min_val = float("0") @@ -2189,6 +2396,7 @@ class Program_weight_tensor_parameter_206: class Program_weight_tensor_parameter_207: name = "parameter_207" + original_name = "batch_norm_1.w_1" shape = [24] dtype = "float32" min_val = float("0") @@ -2198,6 +2406,7 @@ class Program_weight_tensor_parameter_207: class Program_weight_tensor_parameter_208: name = "parameter_208" + original_name = "conv2d_1.w_0" shape = [24, 24, 1, 1] dtype = "float32" min_val = float("-2.22343") @@ -2209,6 +2418,7 @@ class Program_weight_tensor_parameter_208: class Program_weight_tensor_parameter_209: name = "parameter_209" + original_name = "batch_norm_0.b_0" shape = [24] dtype = "float32" min_val = float("0") @@ -2218,6 +2428,7 @@ class Program_weight_tensor_parameter_209: class Program_weight_tensor_parameter_210: name = "parameter_210" + original_name = "batch_norm_0.w_0" shape = [24] dtype = "float32" min_val = float("0") @@ -2227,6 +2438,7 @@ class Program_weight_tensor_parameter_210: class Program_weight_tensor_parameter_211: name = "parameter_211" + original_name = "batch_norm_0.w_2" shape = [24] dtype = "float32" min_val = float("0") @@ -2236,6 +2448,7 @@ class Program_weight_tensor_parameter_211: class Program_weight_tensor_parameter_212: name = "parameter_212" + original_name = "batch_norm_0.w_1" shape = [24] dtype = "float32" min_val = float("0") @@ -2245,6 +2458,7 @@ class Program_weight_tensor_parameter_212: class Program_weight_tensor_parameter_213: name = "parameter_213" + original_name = "conv2d_0.w_0" shape = [24, 3, 3, 3] dtype = "float32" min_val = float("-2.5086") diff --git a/paddle_samples/PaddleX/MobileNetV4_conv_large/subgraph_0/input_meta.py b/paddle_samples/PaddleX/MobileNetV4_conv_large/subgraph_0/input_meta.py index bdf948710..aa6eac72a 100644 --- a/paddle_samples/PaddleX/MobileNetV4_conv_large/subgraph_0/input_meta.py +++ b/paddle_samples/PaddleX/MobileNetV4_conv_large/subgraph_0/input_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_data_0: name = "data_0" + original_name = "var_0" shape = [32, 3, 384, 384] dtype = "float32" min_val = float("-4.59204") diff --git a/paddle_samples/PaddleX/MobileNetV4_conv_large/subgraph_0/weight_meta.py b/paddle_samples/PaddleX/MobileNetV4_conv_large/subgraph_0/weight_meta.py index 74c8de58c..d0674c6ad 100644 --- a/paddle_samples/PaddleX/MobileNetV4_conv_large/subgraph_0/weight_meta.py +++ b/paddle_samples/PaddleX/MobileNetV4_conv_large/subgraph_0/weight_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_parameter_0: name = "parameter_0" + original_name = "linear_0.b_0" shape = [102] dtype = "float32" data = None @@ -7,6 +8,7 @@ class Program_weight_tensor_parameter_0: class Program_weight_tensor_parameter_1: name = "parameter_1" + original_name = "linear_0.w_0" shape = [1280, 102] dtype = "float32" min_val = float("-0.0658882") @@ -18,6 +20,7 @@ class Program_weight_tensor_parameter_1: class Program_weight_tensor_parameter_2: name = "parameter_2" + original_name = "batch_norm_101.b_0" shape = [1280] dtype = "float32" min_val = float("-7.28647") @@ -29,6 +32,7 @@ class Program_weight_tensor_parameter_2: class Program_weight_tensor_parameter_3: name = "parameter_3" + original_name = "batch_norm_101.w_0" shape = [1280] dtype = "float32" min_val = float("0.952601") @@ -40,6 +44,7 @@ class Program_weight_tensor_parameter_3: class Program_weight_tensor_parameter_4: name = "parameter_4" + original_name = "batch_norm_101.w_2" shape = [1280] dtype = "float32" min_val = float("3.99122e-05") @@ -51,6 +56,7 @@ class Program_weight_tensor_parameter_4: class Program_weight_tensor_parameter_5: name = "parameter_5" + original_name = "batch_norm_101.w_1" shape = [1280] dtype = "float32" min_val = float("-0.0053019") @@ -62,6 +68,7 @@ class Program_weight_tensor_parameter_5: class Program_weight_tensor_parameter_6: name = "parameter_6" + original_name = "conv2d_101.w_0" shape = [1280, 960, 1, 1] dtype = "float32" min_val = float("-0.232122") @@ -73,6 +80,7 @@ class Program_weight_tensor_parameter_6: class Program_weight_tensor_parameter_7: name = "parameter_7" + original_name = "batch_norm_100.b_0" shape = [960] dtype = "float32" min_val = float("-10.6415") @@ -84,6 +92,7 @@ class Program_weight_tensor_parameter_7: class Program_weight_tensor_parameter_8: name = "parameter_8" + original_name = "batch_norm_100.w_0" shape = [960] dtype = "float32" min_val = float("0.077667") @@ -95,6 +104,7 @@ class Program_weight_tensor_parameter_8: class Program_weight_tensor_parameter_9: name = "parameter_9" + original_name = "batch_norm_100.w_2" shape = [960] dtype = "float32" min_val = float("0.508822") @@ -106,6 +116,7 @@ class Program_weight_tensor_parameter_9: class Program_weight_tensor_parameter_10: name = "parameter_10" + original_name = "batch_norm_100.w_1" shape = [960] dtype = "float32" min_val = float("-5.92246") @@ -117,6 +128,7 @@ class Program_weight_tensor_parameter_10: class Program_weight_tensor_parameter_11: name = "parameter_11" + original_name = "conv2d_100.w_0" shape = [960, 512, 1, 1] dtype = "float32" min_val = float("-0.352171") @@ -128,6 +140,7 @@ class Program_weight_tensor_parameter_11: class Program_weight_tensor_parameter_12: name = "parameter_12" + original_name = "batch_norm_99.b_0" shape = [512] dtype = "float32" min_val = float("-1.53792") @@ -139,6 +152,7 @@ class Program_weight_tensor_parameter_12: class Program_weight_tensor_parameter_13: name = "parameter_13" + original_name = "batch_norm_99.w_0" shape = [512] dtype = "float32" min_val = float("0.498182") @@ -150,6 +164,7 @@ class Program_weight_tensor_parameter_13: class Program_weight_tensor_parameter_14: name = "parameter_14" + original_name = "batch_norm_99.w_2" shape = [512] dtype = "float32" min_val = float("0.000474777") @@ -161,6 +176,7 @@ class Program_weight_tensor_parameter_14: class Program_weight_tensor_parameter_15: name = "parameter_15" + original_name = "batch_norm_99.w_1" shape = [512] dtype = "float32" min_val = float("-0.0279064") @@ -172,6 +188,7 @@ class Program_weight_tensor_parameter_15: class Program_weight_tensor_parameter_16: name = "parameter_16" + original_name = "conv2d_99.w_0" shape = [512, 2048, 1, 1] dtype = "float32" min_val = float("-0.324183") @@ -183,6 +200,7 @@ class Program_weight_tensor_parameter_16: class Program_weight_tensor_parameter_17: name = "parameter_17" + original_name = "batch_norm_98.b_0" shape = [2048] dtype = "float32" min_val = float("-10.1298") @@ -194,6 +212,7 @@ class Program_weight_tensor_parameter_17: class Program_weight_tensor_parameter_18: name = "parameter_18" + original_name = "batch_norm_98.w_0" shape = [2048] dtype = "float32" min_val = float("0.00437991") @@ -205,6 +224,7 @@ class Program_weight_tensor_parameter_18: class Program_weight_tensor_parameter_19: name = "parameter_19" + original_name = "batch_norm_98.w_2" shape = [2048] dtype = "float32" min_val = float("2.87546e-42") @@ -216,6 +236,7 @@ class Program_weight_tensor_parameter_19: class Program_weight_tensor_parameter_20: name = "parameter_20" + original_name = "batch_norm_98.w_1" shape = [2048] dtype = "float32" min_val = float("-1.0254") @@ -227,6 +248,7 @@ class Program_weight_tensor_parameter_20: class Program_weight_tensor_parameter_21: name = "parameter_21" + original_name = "conv2d_98.w_0" shape = [2048, 512, 1, 1] dtype = "float32" min_val = float("-0.333301") @@ -238,6 +260,7 @@ class Program_weight_tensor_parameter_21: class Program_weight_tensor_parameter_22: name = "parameter_22" + original_name = "batch_norm_97.b_0" shape = [512] dtype = "float32" min_val = float("-2.5288") @@ -249,6 +272,7 @@ class Program_weight_tensor_parameter_22: class Program_weight_tensor_parameter_23: name = "parameter_23" + original_name = "batch_norm_97.w_0" shape = [512] dtype = "float32" min_val = float("0.307202") @@ -260,6 +284,7 @@ class Program_weight_tensor_parameter_23: class Program_weight_tensor_parameter_24: name = "parameter_24" + original_name = "batch_norm_97.w_2" shape = [512] dtype = "float32" min_val = float("0.00118714") @@ -271,6 +296,7 @@ class Program_weight_tensor_parameter_24: class Program_weight_tensor_parameter_25: name = "parameter_25" + original_name = "batch_norm_97.w_1" shape = [512] dtype = "float32" min_val = float("-3.41808") @@ -282,6 +308,7 @@ class Program_weight_tensor_parameter_25: class Program_weight_tensor_parameter_26: name = "parameter_26" + original_name = "conv2d_97.w_0" shape = [512, 1, 5, 5] dtype = "float32" min_val = float("-0.241987") @@ -293,6 +320,7 @@ class Program_weight_tensor_parameter_26: class Program_weight_tensor_parameter_27: name = "parameter_27" + original_name = "batch_norm_96.b_0" shape = [512] dtype = "float32" min_val = float("-1.44063") @@ -304,6 +332,7 @@ class Program_weight_tensor_parameter_27: class Program_weight_tensor_parameter_28: name = "parameter_28" + original_name = "batch_norm_96.w_0" shape = [512] dtype = "float32" min_val = float("-0.0420744") @@ -315,6 +344,7 @@ class Program_weight_tensor_parameter_28: class Program_weight_tensor_parameter_29: name = "parameter_29" + original_name = "batch_norm_96.w_2" shape = [512] dtype = "float32" min_val = float("0.000310175") @@ -326,6 +356,7 @@ class Program_weight_tensor_parameter_29: class Program_weight_tensor_parameter_30: name = "parameter_30" + original_name = "batch_norm_96.w_1" shape = [512] dtype = "float32" min_val = float("-0.0469773") @@ -337,6 +368,7 @@ class Program_weight_tensor_parameter_30: class Program_weight_tensor_parameter_31: name = "parameter_31" + original_name = "conv2d_96.w_0" shape = [512, 2048, 1, 1] dtype = "float32" min_val = float("-0.220279") @@ -348,6 +380,7 @@ class Program_weight_tensor_parameter_31: class Program_weight_tensor_parameter_32: name = "parameter_32" + original_name = "batch_norm_95.b_0" shape = [2048] dtype = "float32" min_val = float("-5.85875") @@ -359,6 +392,7 @@ class Program_weight_tensor_parameter_32: class Program_weight_tensor_parameter_33: name = "parameter_33" + original_name = "batch_norm_95.w_0" shape = [2048] dtype = "float32" min_val = float("-0.0575405") @@ -370,6 +404,7 @@ class Program_weight_tensor_parameter_33: class Program_weight_tensor_parameter_34: name = "parameter_34" + original_name = "batch_norm_95.w_2" shape = [2048] dtype = "float32" min_val = float("2.75495e-42") @@ -381,6 +416,7 @@ class Program_weight_tensor_parameter_34: class Program_weight_tensor_parameter_35: name = "parameter_35" + original_name = "batch_norm_95.w_1" shape = [2048] dtype = "float32" min_val = float("-2.32512") @@ -392,6 +428,7 @@ class Program_weight_tensor_parameter_35: class Program_weight_tensor_parameter_36: name = "parameter_36" + original_name = "conv2d_95.w_0" shape = [2048, 512, 1, 1] dtype = "float32" min_val = float("-0.221471") @@ -403,6 +440,7 @@ class Program_weight_tensor_parameter_36: class Program_weight_tensor_parameter_37: name = "parameter_37" + original_name = "batch_norm_94.b_0" shape = [512] dtype = "float32" min_val = float("-3.05348") @@ -414,6 +452,7 @@ class Program_weight_tensor_parameter_37: class Program_weight_tensor_parameter_38: name = "parameter_38" + original_name = "batch_norm_94.w_0" shape = [512] dtype = "float32" min_val = float("0.430806") @@ -425,6 +464,7 @@ class Program_weight_tensor_parameter_38: class Program_weight_tensor_parameter_39: name = "parameter_39" + original_name = "batch_norm_94.w_2" shape = [512] dtype = "float32" min_val = float("0.00117162") @@ -436,6 +476,7 @@ class Program_weight_tensor_parameter_39: class Program_weight_tensor_parameter_40: name = "parameter_40" + original_name = "batch_norm_94.w_1" shape = [512] dtype = "float32" min_val = float("-3.11543") @@ -447,6 +488,7 @@ class Program_weight_tensor_parameter_40: class Program_weight_tensor_parameter_41: name = "parameter_41" + original_name = "conv2d_94.w_0" shape = [512, 1, 5, 5] dtype = "float32" min_val = float("-0.261002") @@ -458,6 +500,7 @@ class Program_weight_tensor_parameter_41: class Program_weight_tensor_parameter_42: name = "parameter_42" + original_name = "batch_norm_93.b_0" shape = [512] dtype = "float32" min_val = float("-1.64797") @@ -469,6 +512,7 @@ class Program_weight_tensor_parameter_42: class Program_weight_tensor_parameter_43: name = "parameter_43" + original_name = "batch_norm_93.w_0" shape = [512] dtype = "float32" min_val = float("-0.153304") @@ -480,6 +524,7 @@ class Program_weight_tensor_parameter_43: class Program_weight_tensor_parameter_44: name = "parameter_44" + original_name = "batch_norm_93.w_2" shape = [512] dtype = "float32" min_val = float("0.000121534") @@ -491,6 +536,7 @@ class Program_weight_tensor_parameter_44: class Program_weight_tensor_parameter_45: name = "parameter_45" + original_name = "batch_norm_93.w_1" shape = [512] dtype = "float32" min_val = float("-0.0273515") @@ -502,6 +548,7 @@ class Program_weight_tensor_parameter_45: class Program_weight_tensor_parameter_46: name = "parameter_46" + original_name = "conv2d_93.w_0" shape = [512, 2048, 1, 1] dtype = "float32" min_val = float("-0.247926") @@ -513,6 +560,7 @@ class Program_weight_tensor_parameter_46: class Program_weight_tensor_parameter_47: name = "parameter_47" + original_name = "batch_norm_92.b_0" shape = [2048] dtype = "float32" min_val = float("-3.78403") @@ -524,6 +572,7 @@ class Program_weight_tensor_parameter_47: class Program_weight_tensor_parameter_48: name = "parameter_48" + original_name = "batch_norm_92.w_0" shape = [2048] dtype = "float32" min_val = float("0.0246116") @@ -535,6 +584,7 @@ class Program_weight_tensor_parameter_48: class Program_weight_tensor_parameter_49: name = "parameter_49" + original_name = "batch_norm_92.w_2" shape = [2048] dtype = "float32" min_val = float("2.89368e-42") @@ -546,6 +596,7 @@ class Program_weight_tensor_parameter_49: class Program_weight_tensor_parameter_50: name = "parameter_50" + original_name = "batch_norm_92.w_1" shape = [2048] dtype = "float32" min_val = float("-2.00046") @@ -557,6 +608,7 @@ class Program_weight_tensor_parameter_50: class Program_weight_tensor_parameter_51: name = "parameter_51" + original_name = "conv2d_92.w_0" shape = [2048, 512, 1, 1] dtype = "float32" min_val = float("-0.240553") @@ -568,6 +620,7 @@ class Program_weight_tensor_parameter_51: class Program_weight_tensor_parameter_52: name = "parameter_52" + original_name = "batch_norm_91.b_0" shape = [512] dtype = "float32" min_val = float("-3.41733") @@ -579,6 +632,7 @@ class Program_weight_tensor_parameter_52: class Program_weight_tensor_parameter_53: name = "parameter_53" + original_name = "batch_norm_91.w_0" shape = [512] dtype = "float32" min_val = float("0.365644") @@ -590,6 +644,7 @@ class Program_weight_tensor_parameter_53: class Program_weight_tensor_parameter_54: name = "parameter_54" + original_name = "batch_norm_91.w_2" shape = [512] dtype = "float32" min_val = float("0.000291028") @@ -601,6 +656,7 @@ class Program_weight_tensor_parameter_54: class Program_weight_tensor_parameter_55: name = "parameter_55" + original_name = "batch_norm_91.w_1" shape = [512] dtype = "float32" min_val = float("-2.15924") @@ -612,6 +668,7 @@ class Program_weight_tensor_parameter_55: class Program_weight_tensor_parameter_56: name = "parameter_56" + original_name = "conv2d_91.w_0" shape = [512, 1, 5, 5] dtype = "float32" min_val = float("-0.216383") @@ -623,6 +680,7 @@ class Program_weight_tensor_parameter_56: class Program_weight_tensor_parameter_57: name = "parameter_57" + original_name = "batch_norm_90.b_0" shape = [512] dtype = "float32" min_val = float("-1.73698") @@ -634,6 +692,7 @@ class Program_weight_tensor_parameter_57: class Program_weight_tensor_parameter_58: name = "parameter_58" + original_name = "batch_norm_90.w_0" shape = [512] dtype = "float32" min_val = float("-0.407294") @@ -645,6 +704,7 @@ class Program_weight_tensor_parameter_58: class Program_weight_tensor_parameter_59: name = "parameter_59" + original_name = "batch_norm_90.w_2" shape = [512] dtype = "float32" min_val = float("0.0353802") @@ -656,6 +716,7 @@ class Program_weight_tensor_parameter_59: class Program_weight_tensor_parameter_60: name = "parameter_60" + original_name = "batch_norm_90.w_1" shape = [512] dtype = "float32" min_val = float("-1.98256") @@ -667,6 +728,7 @@ class Program_weight_tensor_parameter_60: class Program_weight_tensor_parameter_61: name = "parameter_61" + original_name = "conv2d_90.w_0" shape = [512, 2048, 1, 1] dtype = "float32" min_val = float("-0.27489") @@ -678,6 +740,7 @@ class Program_weight_tensor_parameter_61: class Program_weight_tensor_parameter_62: name = "parameter_62" + original_name = "batch_norm_89.b_0" shape = [2048] dtype = "float32" min_val = float("-4.16353") @@ -689,6 +752,7 @@ class Program_weight_tensor_parameter_62: class Program_weight_tensor_parameter_63: name = "parameter_63" + original_name = "batch_norm_89.w_0" shape = [2048] dtype = "float32" min_val = float("-0.669085") @@ -700,6 +764,7 @@ class Program_weight_tensor_parameter_63: class Program_weight_tensor_parameter_64: name = "parameter_64" + original_name = "batch_norm_89.w_2" shape = [2048] dtype = "float32" min_val = float("1.98564e-42") @@ -711,6 +776,7 @@ class Program_weight_tensor_parameter_64: class Program_weight_tensor_parameter_65: name = "parameter_65" + original_name = "batch_norm_89.w_1" shape = [2048] dtype = "float32" min_val = float("-0.52688") @@ -722,6 +788,7 @@ class Program_weight_tensor_parameter_65: class Program_weight_tensor_parameter_66: name = "parameter_66" + original_name = "conv2d_89.w_0" shape = [2048, 1, 5, 5] dtype = "float32" min_val = float("-0.398878") @@ -733,6 +800,7 @@ class Program_weight_tensor_parameter_66: class Program_weight_tensor_parameter_67: name = "parameter_67" + original_name = "batch_norm_88.b_0" shape = [2048] dtype = "float32" min_val = float("-1.80841") @@ -744,6 +812,7 @@ class Program_weight_tensor_parameter_67: class Program_weight_tensor_parameter_68: name = "parameter_68" + original_name = "batch_norm_88.w_0" shape = [2048] dtype = "float32" min_val = float("-0.0214885") @@ -755,6 +824,7 @@ class Program_weight_tensor_parameter_68: class Program_weight_tensor_parameter_69: name = "parameter_69" + original_name = "batch_norm_88.w_2" shape = [2048] dtype = "float32" min_val = float("2.49011e-42") @@ -766,6 +836,7 @@ class Program_weight_tensor_parameter_69: class Program_weight_tensor_parameter_70: name = "parameter_70" + original_name = "batch_norm_88.w_1" shape = [2048] dtype = "float32" min_val = float("-1.17569") @@ -777,6 +848,7 @@ class Program_weight_tensor_parameter_70: class Program_weight_tensor_parameter_71: name = "parameter_71" + original_name = "conv2d_88.w_0" shape = [2048, 512, 1, 1] dtype = "float32" min_val = float("-0.3545") @@ -788,6 +860,7 @@ class Program_weight_tensor_parameter_71: class Program_weight_tensor_parameter_72: name = "parameter_72" + original_name = "batch_norm_87.b_0" shape = [512] dtype = "float32" min_val = float("-2.30176") @@ -799,6 +872,7 @@ class Program_weight_tensor_parameter_72: class Program_weight_tensor_parameter_73: name = "parameter_73" + original_name = "batch_norm_87.w_0" shape = [512] dtype = "float32" min_val = float("0.36444") @@ -810,6 +884,7 @@ class Program_weight_tensor_parameter_73: class Program_weight_tensor_parameter_74: name = "parameter_74" + original_name = "batch_norm_87.w_2" shape = [512] dtype = "float32" min_val = float("0.000624975") @@ -821,6 +896,7 @@ class Program_weight_tensor_parameter_74: class Program_weight_tensor_parameter_75: name = "parameter_75" + original_name = "batch_norm_87.w_1" shape = [512] dtype = "float32" min_val = float("-2.01929") @@ -832,6 +908,7 @@ class Program_weight_tensor_parameter_75: class Program_weight_tensor_parameter_76: name = "parameter_76" + original_name = "conv2d_87.w_0" shape = [512, 1, 5, 5] dtype = "float32" min_val = float("-0.220292") @@ -843,6 +920,7 @@ class Program_weight_tensor_parameter_76: class Program_weight_tensor_parameter_77: name = "parameter_77" + original_name = "batch_norm_86.b_0" shape = [512] dtype = "float32" min_val = float("-2.58435") @@ -854,6 +932,7 @@ class Program_weight_tensor_parameter_77: class Program_weight_tensor_parameter_78: name = "parameter_78" + original_name = "batch_norm_86.w_0" shape = [512] dtype = "float32" min_val = float("0.010893") @@ -865,6 +944,7 @@ class Program_weight_tensor_parameter_78: class Program_weight_tensor_parameter_79: name = "parameter_79" + original_name = "batch_norm_86.w_2" shape = [512] dtype = "float32" min_val = float("0.0252294") @@ -876,6 +956,7 @@ class Program_weight_tensor_parameter_79: class Program_weight_tensor_parameter_80: name = "parameter_80" + original_name = "batch_norm_86.w_1" shape = [512] dtype = "float32" min_val = float("-2.28022") @@ -887,6 +968,7 @@ class Program_weight_tensor_parameter_80: class Program_weight_tensor_parameter_81: name = "parameter_81" + original_name = "conv2d_86.w_0" shape = [512, 2048, 1, 1] dtype = "float32" min_val = float("-0.504555") @@ -898,6 +980,7 @@ class Program_weight_tensor_parameter_81: class Program_weight_tensor_parameter_82: name = "parameter_82" + original_name = "batch_norm_85.b_0" shape = [2048] dtype = "float32" min_val = float("-5.14937") @@ -909,6 +992,7 @@ class Program_weight_tensor_parameter_82: class Program_weight_tensor_parameter_83: name = "parameter_83" + original_name = "batch_norm_85.w_0" shape = [2048] dtype = "float32" min_val = float("-0.0283303") @@ -920,6 +1004,7 @@ class Program_weight_tensor_parameter_83: class Program_weight_tensor_parameter_84: name = "parameter_84" + original_name = "batch_norm_85.w_2" shape = [2048] dtype = "float32" min_val = float("1.93239e-42") @@ -931,6 +1016,7 @@ class Program_weight_tensor_parameter_84: class Program_weight_tensor_parameter_85: name = "parameter_85" + original_name = "batch_norm_85.w_1" shape = [2048] dtype = "float32" min_val = float("-0.228844") @@ -942,6 +1028,7 @@ class Program_weight_tensor_parameter_85: class Program_weight_tensor_parameter_86: name = "parameter_86" + original_name = "conv2d_85.w_0" shape = [2048, 1, 3, 3] dtype = "float32" min_val = float("-0.0958151") @@ -953,6 +1040,7 @@ class Program_weight_tensor_parameter_86: class Program_weight_tensor_parameter_87: name = "parameter_87" + original_name = "batch_norm_84.b_0" shape = [2048] dtype = "float32" min_val = float("-2.18023") @@ -964,6 +1052,7 @@ class Program_weight_tensor_parameter_87: class Program_weight_tensor_parameter_88: name = "parameter_88" + original_name = "batch_norm_84.w_0" shape = [2048] dtype = "float32" min_val = float("0.0115061") @@ -975,6 +1064,7 @@ class Program_weight_tensor_parameter_88: class Program_weight_tensor_parameter_89: name = "parameter_89" + original_name = "batch_norm_84.w_2" shape = [2048] dtype = "float32" min_val = float("2.39062e-42") @@ -986,6 +1076,7 @@ class Program_weight_tensor_parameter_89: class Program_weight_tensor_parameter_90: name = "parameter_90" + original_name = "batch_norm_84.w_1" shape = [2048] dtype = "float32" min_val = float("-1.21131") @@ -997,6 +1088,7 @@ class Program_weight_tensor_parameter_90: class Program_weight_tensor_parameter_91: name = "parameter_91" + original_name = "conv2d_84.w_0" shape = [2048, 512, 1, 1] dtype = "float32" min_val = float("-0.470783") @@ -1008,6 +1100,7 @@ class Program_weight_tensor_parameter_91: class Program_weight_tensor_parameter_92: name = "parameter_92" + original_name = "batch_norm_83.b_0" shape = [512] dtype = "float32" min_val = float("-2.54673") @@ -1019,6 +1112,7 @@ class Program_weight_tensor_parameter_92: class Program_weight_tensor_parameter_93: name = "parameter_93" + original_name = "batch_norm_83.w_0" shape = [512] dtype = "float32" min_val = float("0.283696") @@ -1030,6 +1124,7 @@ class Program_weight_tensor_parameter_93: class Program_weight_tensor_parameter_94: name = "parameter_94" + original_name = "batch_norm_83.w_2" shape = [512] dtype = "float32" min_val = float("0.000194946") @@ -1041,6 +1136,7 @@ class Program_weight_tensor_parameter_94: class Program_weight_tensor_parameter_95: name = "parameter_95" + original_name = "batch_norm_83.w_1" shape = [512] dtype = "float32" min_val = float("-2.14935") @@ -1052,6 +1148,7 @@ class Program_weight_tensor_parameter_95: class Program_weight_tensor_parameter_96: name = "parameter_96" + original_name = "conv2d_83.w_0" shape = [512, 1, 5, 5] dtype = "float32" min_val = float("-0.236161") @@ -1063,6 +1160,7 @@ class Program_weight_tensor_parameter_96: class Program_weight_tensor_parameter_97: name = "parameter_97" + original_name = "batch_norm_82.b_0" shape = [512] dtype = "float32" min_val = float("-2.48249") @@ -1074,6 +1172,7 @@ class Program_weight_tensor_parameter_97: class Program_weight_tensor_parameter_98: name = "parameter_98" + original_name = "batch_norm_82.w_0" shape = [512] dtype = "float32" min_val = float("-0.17464") @@ -1085,6 +1184,7 @@ class Program_weight_tensor_parameter_98: class Program_weight_tensor_parameter_99: name = "parameter_99" + original_name = "batch_norm_82.w_2" shape = [512] dtype = "float32" min_val = float("6.93179e-05") @@ -1096,6 +1196,7 @@ class Program_weight_tensor_parameter_99: class Program_weight_tensor_parameter_100: name = "parameter_100" + original_name = "batch_norm_82.w_1" shape = [512] dtype = "float32" min_val = float("-0.0341773") @@ -1107,6 +1208,7 @@ class Program_weight_tensor_parameter_100: class Program_weight_tensor_parameter_101: name = "parameter_101" + original_name = "conv2d_82.w_0" shape = [512, 2048, 1, 1] dtype = "float32" min_val = float("-0.187782") @@ -1118,6 +1220,7 @@ class Program_weight_tensor_parameter_101: class Program_weight_tensor_parameter_102: name = "parameter_102" + original_name = "batch_norm_81.b_0" shape = [2048] dtype = "float32" min_val = float("-3.38872") @@ -1129,6 +1232,7 @@ class Program_weight_tensor_parameter_102: class Program_weight_tensor_parameter_103: name = "parameter_103" + original_name = "batch_norm_81.w_0" shape = [2048] dtype = "float32" min_val = float("0.0109399") @@ -1140,6 +1244,7 @@ class Program_weight_tensor_parameter_103: class Program_weight_tensor_parameter_104: name = "parameter_104" + original_name = "batch_norm_81.w_2" shape = [2048] dtype = "float32" min_val = float("2.86986e-42") @@ -1151,6 +1256,7 @@ class Program_weight_tensor_parameter_104: class Program_weight_tensor_parameter_105: name = "parameter_105" + original_name = "batch_norm_81.w_1" shape = [2048] dtype = "float32" min_val = float("-1.85067") @@ -1162,6 +1268,7 @@ class Program_weight_tensor_parameter_105: class Program_weight_tensor_parameter_106: name = "parameter_106" + original_name = "conv2d_81.w_0" shape = [2048, 512, 1, 1] dtype = "float32" min_val = float("-0.266474") @@ -1173,6 +1280,7 @@ class Program_weight_tensor_parameter_106: class Program_weight_tensor_parameter_107: name = "parameter_107" + original_name = "batch_norm_80.b_0" shape = [512] dtype = "float32" min_val = float("-4.725") @@ -1184,6 +1292,7 @@ class Program_weight_tensor_parameter_107: class Program_weight_tensor_parameter_108: name = "parameter_108" + original_name = "batch_norm_80.w_0" shape = [512] dtype = "float32" min_val = float("-0.440625") @@ -1195,6 +1304,7 @@ class Program_weight_tensor_parameter_108: class Program_weight_tensor_parameter_109: name = "parameter_109" + original_name = "batch_norm_80.w_2" shape = [512] dtype = "float32" min_val = float("0.000168878") @@ -1206,6 +1316,7 @@ class Program_weight_tensor_parameter_109: class Program_weight_tensor_parameter_110: name = "parameter_110" + original_name = "batch_norm_80.w_1" shape = [512] dtype = "float32" min_val = float("-1.79333") @@ -1217,6 +1328,7 @@ class Program_weight_tensor_parameter_110: class Program_weight_tensor_parameter_111: name = "parameter_111" + original_name = "conv2d_80.w_0" shape = [512, 1, 5, 5] dtype = "float32" min_val = float("-0.233553") @@ -1228,6 +1340,7 @@ class Program_weight_tensor_parameter_111: class Program_weight_tensor_parameter_112: name = "parameter_112" + original_name = "batch_norm_79.b_0" shape = [512] dtype = "float32" min_val = float("-1.56354") @@ -1239,6 +1352,7 @@ class Program_weight_tensor_parameter_112: class Program_weight_tensor_parameter_113: name = "parameter_113" + original_name = "batch_norm_79.w_0" shape = [512] dtype = "float32" min_val = float("-0.00918385") @@ -1250,6 +1364,7 @@ class Program_weight_tensor_parameter_113: class Program_weight_tensor_parameter_114: name = "parameter_114" + original_name = "batch_norm_79.w_2" shape = [512] dtype = "float32" min_val = float("0.000108985") @@ -1261,6 +1376,7 @@ class Program_weight_tensor_parameter_114: class Program_weight_tensor_parameter_115: name = "parameter_115" + original_name = "batch_norm_79.w_1" shape = [512] dtype = "float32" min_val = float("-0.00958077") @@ -1272,6 +1388,7 @@ class Program_weight_tensor_parameter_115: class Program_weight_tensor_parameter_116: name = "parameter_116" + original_name = "conv2d_79.w_0" shape = [512, 2048, 1, 1] dtype = "float32" min_val = float("-0.118448") @@ -1283,6 +1400,7 @@ class Program_weight_tensor_parameter_116: class Program_weight_tensor_parameter_117: name = "parameter_117" + original_name = "batch_norm_78.b_0" shape = [2048] dtype = "float32" min_val = float("-3.84187") @@ -1294,6 +1412,7 @@ class Program_weight_tensor_parameter_117: class Program_weight_tensor_parameter_118: name = "parameter_118" + original_name = "batch_norm_78.w_0" shape = [2048] dtype = "float32" min_val = float("0.00396764") @@ -1305,6 +1424,7 @@ class Program_weight_tensor_parameter_118: class Program_weight_tensor_parameter_119: name = "parameter_119" + original_name = "batch_norm_78.w_2" shape = [2048] dtype = "float32" min_val = float("2.88808e-42") @@ -1316,6 +1436,7 @@ class Program_weight_tensor_parameter_119: class Program_weight_tensor_parameter_120: name = "parameter_120" + original_name = "batch_norm_78.w_1" shape = [2048] dtype = "float32" min_val = float("-1.51991") @@ -1327,6 +1448,7 @@ class Program_weight_tensor_parameter_120: class Program_weight_tensor_parameter_121: name = "parameter_121" + original_name = "conv2d_78.w_0" shape = [2048, 512, 1, 1] dtype = "float32" min_val = float("-0.167228") @@ -1338,6 +1460,7 @@ class Program_weight_tensor_parameter_121: class Program_weight_tensor_parameter_122: name = "parameter_122" + original_name = "batch_norm_77.b_0" shape = [512] dtype = "float32" min_val = float("-3.28978") @@ -1349,6 +1472,7 @@ class Program_weight_tensor_parameter_122: class Program_weight_tensor_parameter_123: name = "parameter_123" + original_name = "batch_norm_77.w_0" shape = [512] dtype = "float32" min_val = float("-0.545562") @@ -1360,6 +1484,7 @@ class Program_weight_tensor_parameter_123: class Program_weight_tensor_parameter_124: name = "parameter_124" + original_name = "batch_norm_77.w_2" shape = [512] dtype = "float32" min_val = float("0.000149431") @@ -1371,6 +1496,7 @@ class Program_weight_tensor_parameter_124: class Program_weight_tensor_parameter_125: name = "parameter_125" + original_name = "batch_norm_77.w_1" shape = [512] dtype = "float32" min_val = float("-1.77509") @@ -1382,6 +1508,7 @@ class Program_weight_tensor_parameter_125: class Program_weight_tensor_parameter_126: name = "parameter_126" + original_name = "conv2d_77.w_0" shape = [512, 1, 5, 5] dtype = "float32" min_val = float("-0.234007") @@ -1393,6 +1520,7 @@ class Program_weight_tensor_parameter_126: class Program_weight_tensor_parameter_127: name = "parameter_127" + original_name = "batch_norm_76.b_0" shape = [512] dtype = "float32" min_val = float("-1.70643") @@ -1404,6 +1532,7 @@ class Program_weight_tensor_parameter_127: class Program_weight_tensor_parameter_128: name = "parameter_128" + original_name = "batch_norm_76.w_0" shape = [512] dtype = "float32" min_val = float("-0.000734476") @@ -1415,6 +1544,7 @@ class Program_weight_tensor_parameter_128: class Program_weight_tensor_parameter_129: name = "parameter_129" + original_name = "batch_norm_76.w_2" shape = [512] dtype = "float32" min_val = float("0.0466089") @@ -1426,6 +1556,7 @@ class Program_weight_tensor_parameter_129: class Program_weight_tensor_parameter_130: name = "parameter_130" + original_name = "batch_norm_76.w_1" shape = [512] dtype = "float32" min_val = float("-4.25865") @@ -1437,6 +1568,7 @@ class Program_weight_tensor_parameter_130: class Program_weight_tensor_parameter_131: name = "parameter_131" + original_name = "conv2d_76.w_0" shape = [512, 2048, 1, 1] dtype = "float32" min_val = float("-0.285626") @@ -1448,6 +1580,7 @@ class Program_weight_tensor_parameter_131: class Program_weight_tensor_parameter_132: name = "parameter_132" + original_name = "batch_norm_75.b_0" shape = [2048] dtype = "float32" min_val = float("-6.76864") @@ -1459,6 +1592,7 @@ class Program_weight_tensor_parameter_132: class Program_weight_tensor_parameter_133: name = "parameter_133" + original_name = "batch_norm_75.w_0" shape = [2048] dtype = "float32" min_val = float("-0.0506871") @@ -1470,6 +1604,7 @@ class Program_weight_tensor_parameter_133: class Program_weight_tensor_parameter_134: name = "parameter_134" + original_name = "batch_norm_75.w_2" shape = [2048] dtype = "float32" min_val = float("1.97863e-42") @@ -1481,6 +1616,7 @@ class Program_weight_tensor_parameter_134: class Program_weight_tensor_parameter_135: name = "parameter_135" + original_name = "batch_norm_75.w_1" shape = [2048] dtype = "float32" min_val = float("-0.224808") @@ -1492,6 +1628,7 @@ class Program_weight_tensor_parameter_135: class Program_weight_tensor_parameter_136: name = "parameter_136" + original_name = "conv2d_75.w_0" shape = [2048, 1, 3, 3] dtype = "float32" min_val = float("-0.0933703") @@ -1503,6 +1640,7 @@ class Program_weight_tensor_parameter_136: class Program_weight_tensor_parameter_137: name = "parameter_137" + original_name = "batch_norm_74.b_0" shape = [2048] dtype = "float32" min_val = float("-2.14057") @@ -1514,6 +1652,7 @@ class Program_weight_tensor_parameter_137: class Program_weight_tensor_parameter_138: name = "parameter_138" + original_name = "batch_norm_74.w_0" shape = [2048] dtype = "float32" min_val = float("0.0205888") @@ -1525,6 +1664,7 @@ class Program_weight_tensor_parameter_138: class Program_weight_tensor_parameter_139: name = "parameter_139" + original_name = "batch_norm_74.w_2" shape = [2048] dtype = "float32" min_val = float("2.45227e-42") @@ -1536,6 +1676,7 @@ class Program_weight_tensor_parameter_139: class Program_weight_tensor_parameter_140: name = "parameter_140" + original_name = "batch_norm_74.w_1" shape = [2048] dtype = "float32" min_val = float("-1.1837") @@ -1547,6 +1688,7 @@ class Program_weight_tensor_parameter_140: class Program_weight_tensor_parameter_141: name = "parameter_141" + original_name = "conv2d_74.w_0" shape = [2048, 512, 1, 1] dtype = "float32" min_val = float("-0.36499") @@ -1558,6 +1700,7 @@ class Program_weight_tensor_parameter_141: class Program_weight_tensor_parameter_142: name = "parameter_142" + original_name = "batch_norm_73.b_0" shape = [512] dtype = "float32" min_val = float("-2.67926") @@ -1569,6 +1712,7 @@ class Program_weight_tensor_parameter_142: class Program_weight_tensor_parameter_143: name = "parameter_143" + original_name = "batch_norm_73.w_0" shape = [512] dtype = "float32" min_val = float("0.321339") @@ -1580,6 +1724,7 @@ class Program_weight_tensor_parameter_143: class Program_weight_tensor_parameter_144: name = "parameter_144" + original_name = "batch_norm_73.w_2" shape = [512] dtype = "float32" min_val = float("0.000119375") @@ -1591,6 +1736,7 @@ class Program_weight_tensor_parameter_144: class Program_weight_tensor_parameter_145: name = "parameter_145" + original_name = "batch_norm_73.w_1" shape = [512] dtype = "float32" min_val = float("-1.45572") @@ -1602,6 +1748,7 @@ class Program_weight_tensor_parameter_145: class Program_weight_tensor_parameter_146: name = "parameter_146" + original_name = "conv2d_73.w_0" shape = [512, 1, 5, 5] dtype = "float32" min_val = float("-0.278055") @@ -1613,6 +1760,7 @@ class Program_weight_tensor_parameter_146: class Program_weight_tensor_parameter_147: name = "parameter_147" + original_name = "batch_norm_72.b_0" shape = [512] dtype = "float32" min_val = float("-1.20274") @@ -1624,6 +1772,7 @@ class Program_weight_tensor_parameter_147: class Program_weight_tensor_parameter_148: name = "parameter_148" + original_name = "batch_norm_72.w_0" shape = [512] dtype = "float32" min_val = float("-0.0321616") @@ -1635,6 +1784,7 @@ class Program_weight_tensor_parameter_148: class Program_weight_tensor_parameter_149: name = "parameter_149" + original_name = "batch_norm_72.w_2" shape = [512] dtype = "float32" min_val = float("0.000949724") @@ -1646,6 +1796,7 @@ class Program_weight_tensor_parameter_149: class Program_weight_tensor_parameter_150: name = "parameter_150" + original_name = "batch_norm_72.w_1" shape = [512] dtype = "float32" min_val = float("-0.0595983") @@ -1657,6 +1808,7 @@ class Program_weight_tensor_parameter_150: class Program_weight_tensor_parameter_151: name = "parameter_151" + original_name = "conv2d_72.w_0" shape = [512, 2048, 1, 1] dtype = "float32" min_val = float("-0.227093") @@ -1668,6 +1820,7 @@ class Program_weight_tensor_parameter_151: class Program_weight_tensor_parameter_152: name = "parameter_152" + original_name = "batch_norm_71.b_0" shape = [2048] dtype = "float32" min_val = float("-3.28666") @@ -1679,6 +1832,7 @@ class Program_weight_tensor_parameter_152: class Program_weight_tensor_parameter_153: name = "parameter_153" + original_name = "batch_norm_71.w_0" shape = [2048] dtype = "float32" min_val = float("0.0202884") @@ -1690,6 +1844,7 @@ class Program_weight_tensor_parameter_153: class Program_weight_tensor_parameter_154: name = "parameter_154" + original_name = "batch_norm_71.w_2" shape = [2048] dtype = "float32" min_val = float("2.78718e-42") @@ -1701,6 +1856,7 @@ class Program_weight_tensor_parameter_154: class Program_weight_tensor_parameter_155: name = "parameter_155" + original_name = "batch_norm_71.w_1" shape = [2048] dtype = "float32" min_val = float("-1.23462") @@ -1712,6 +1868,7 @@ class Program_weight_tensor_parameter_155: class Program_weight_tensor_parameter_156: name = "parameter_156" + original_name = "conv2d_71.w_0" shape = [2048, 512, 1, 1] dtype = "float32" min_val = float("-0.181611") @@ -1723,6 +1880,7 @@ class Program_weight_tensor_parameter_156: class Program_weight_tensor_parameter_157: name = "parameter_157" + original_name = "batch_norm_70.b_0" shape = [512] dtype = "float32" min_val = float("-2.52768") @@ -1734,6 +1892,7 @@ class Program_weight_tensor_parameter_157: class Program_weight_tensor_parameter_158: name = "parameter_158" + original_name = "batch_norm_70.w_0" shape = [512] dtype = "float32" min_val = float("-0.125384") @@ -1745,6 +1904,7 @@ class Program_weight_tensor_parameter_158: class Program_weight_tensor_parameter_159: name = "parameter_159" + original_name = "batch_norm_70.w_2" shape = [512] dtype = "float32" min_val = float("0.000123029") @@ -1756,6 +1916,7 @@ class Program_weight_tensor_parameter_159: class Program_weight_tensor_parameter_160: name = "parameter_160" + original_name = "batch_norm_70.w_1" shape = [512] dtype = "float32" min_val = float("-1.89873") @@ -1767,6 +1928,7 @@ class Program_weight_tensor_parameter_160: class Program_weight_tensor_parameter_161: name = "parameter_161" + original_name = "conv2d_70.w_0" shape = [512, 1, 5, 5] dtype = "float32" min_val = float("-0.250515") @@ -1778,6 +1940,7 @@ class Program_weight_tensor_parameter_161: class Program_weight_tensor_parameter_162: name = "parameter_162" + original_name = "batch_norm_69.b_0" shape = [512] dtype = "float32" min_val = float("-1.66793") @@ -1789,6 +1952,7 @@ class Program_weight_tensor_parameter_162: class Program_weight_tensor_parameter_163: name = "parameter_163" + original_name = "batch_norm_69.w_0" shape = [512] dtype = "float32" min_val = float("-0.0165403") @@ -1800,6 +1964,7 @@ class Program_weight_tensor_parameter_163: class Program_weight_tensor_parameter_164: name = "parameter_164" + original_name = "batch_norm_69.w_2" shape = [512] dtype = "float32" min_val = float("0.13219") @@ -1811,6 +1976,7 @@ class Program_weight_tensor_parameter_164: class Program_weight_tensor_parameter_165: name = "parameter_165" + original_name = "batch_norm_69.w_1" shape = [512] dtype = "float32" min_val = float("-3.2286") @@ -1822,6 +1988,7 @@ class Program_weight_tensor_parameter_165: class Program_weight_tensor_parameter_166: name = "parameter_166" + original_name = "conv2d_69.w_0" shape = [512, 2048, 1, 1] dtype = "float32" min_val = float("-0.238172") @@ -1833,6 +2000,7 @@ class Program_weight_tensor_parameter_166: class Program_weight_tensor_parameter_167: name = "parameter_167" + original_name = "batch_norm_68.b_0" shape = [2048] dtype = "float32" min_val = float("-5.47701") @@ -1844,6 +2012,7 @@ class Program_weight_tensor_parameter_167: class Program_weight_tensor_parameter_168: name = "parameter_168" + original_name = "batch_norm_68.w_0" shape = [2048] dtype = "float32" min_val = float("0.016515") @@ -1855,6 +2024,7 @@ class Program_weight_tensor_parameter_168: class Program_weight_tensor_parameter_169: name = "parameter_169" + original_name = "batch_norm_68.w_2" shape = [2048] dtype = "float32" min_val = float("1.98844e-42") @@ -1866,6 +2036,7 @@ class Program_weight_tensor_parameter_169: class Program_weight_tensor_parameter_170: name = "parameter_170" + original_name = "batch_norm_68.w_1" shape = [2048] dtype = "float32" min_val = float("-0.381359") @@ -1877,6 +2048,7 @@ class Program_weight_tensor_parameter_170: class Program_weight_tensor_parameter_171: name = "parameter_171" + original_name = "conv2d_68.w_0" shape = [2048, 1, 5, 5] dtype = "float32" min_val = float("-0.246563") @@ -1888,6 +2060,7 @@ class Program_weight_tensor_parameter_171: class Program_weight_tensor_parameter_172: name = "parameter_172" + original_name = "batch_norm_67.b_0" shape = [2048] dtype = "float32" min_val = float("-1.84924") @@ -1899,6 +2072,7 @@ class Program_weight_tensor_parameter_172: class Program_weight_tensor_parameter_173: name = "parameter_173" + original_name = "batch_norm_67.w_0" shape = [2048] dtype = "float32" min_val = float("0.0136606") @@ -1910,6 +2084,7 @@ class Program_weight_tensor_parameter_173: class Program_weight_tensor_parameter_174: name = "parameter_174" + original_name = "batch_norm_67.w_2" shape = [2048] dtype = "float32" min_val = float("2.48871e-42") @@ -1921,6 +2096,7 @@ class Program_weight_tensor_parameter_174: class Program_weight_tensor_parameter_175: name = "parameter_175" + original_name = "batch_norm_67.w_1" shape = [2048] dtype = "float32" min_val = float("-1.21993") @@ -1932,6 +2108,7 @@ class Program_weight_tensor_parameter_175: class Program_weight_tensor_parameter_176: name = "parameter_176" + original_name = "conv2d_67.w_0" shape = [2048, 512, 1, 1] dtype = "float32" min_val = float("-0.242585") @@ -1943,6 +2120,7 @@ class Program_weight_tensor_parameter_176: class Program_weight_tensor_parameter_177: name = "parameter_177" + original_name = "batch_norm_66.b_0" shape = [512] dtype = "float32" min_val = float("-2.27451") @@ -1954,6 +2132,7 @@ class Program_weight_tensor_parameter_177: class Program_weight_tensor_parameter_178: name = "parameter_178" + original_name = "batch_norm_66.w_0" shape = [512] dtype = "float32" min_val = float("0.325418") @@ -1965,6 +2144,7 @@ class Program_weight_tensor_parameter_178: class Program_weight_tensor_parameter_179: name = "parameter_179" + original_name = "batch_norm_66.w_2" shape = [512] dtype = "float32" min_val = float("8.22079e-05") @@ -1976,6 +2156,7 @@ class Program_weight_tensor_parameter_179: class Program_weight_tensor_parameter_180: name = "parameter_180" + original_name = "batch_norm_66.w_1" shape = [512] dtype = "float32" min_val = float("-1.95296") @@ -1987,6 +2168,7 @@ class Program_weight_tensor_parameter_180: class Program_weight_tensor_parameter_181: name = "parameter_181" + original_name = "conv2d_66.w_0" shape = [512, 1, 5, 5] dtype = "float32" min_val = float("-0.254229") @@ -1998,6 +2180,7 @@ class Program_weight_tensor_parameter_181: class Program_weight_tensor_parameter_182: name = "parameter_182" + original_name = "batch_norm_65.b_0" shape = [512] dtype = "float32" min_val = float("-1.49786") @@ -2009,6 +2192,7 @@ class Program_weight_tensor_parameter_182: class Program_weight_tensor_parameter_183: name = "parameter_183" + original_name = "batch_norm_65.w_0" shape = [512] dtype = "float32" min_val = float("-0.0419769") @@ -2020,6 +2204,7 @@ class Program_weight_tensor_parameter_183: class Program_weight_tensor_parameter_184: name = "parameter_184" + original_name = "batch_norm_65.w_2" shape = [512] dtype = "float32" min_val = float("0.114558") @@ -2031,6 +2216,7 @@ class Program_weight_tensor_parameter_184: class Program_weight_tensor_parameter_185: name = "parameter_185" + original_name = "batch_norm_65.w_1" shape = [512] dtype = "float32" min_val = float("-2.41654") @@ -2042,6 +2228,7 @@ class Program_weight_tensor_parameter_185: class Program_weight_tensor_parameter_186: name = "parameter_186" + original_name = "conv2d_65.w_0" shape = [512, 2048, 1, 1] dtype = "float32" min_val = float("-0.232467") @@ -2053,6 +2240,7 @@ class Program_weight_tensor_parameter_186: class Program_weight_tensor_parameter_187: name = "parameter_187" + original_name = "batch_norm_64.b_0" shape = [2048] dtype = "float32" min_val = float("-6.32564") @@ -2064,6 +2252,7 @@ class Program_weight_tensor_parameter_187: class Program_weight_tensor_parameter_188: name = "parameter_188" + original_name = "batch_norm_64.w_0" shape = [2048] dtype = "float32" min_val = float("-0.0309241") @@ -2075,6 +2264,7 @@ class Program_weight_tensor_parameter_188: class Program_weight_tensor_parameter_189: name = "parameter_189" + original_name = "batch_norm_64.w_2" shape = [2048] dtype = "float32" min_val = float("1.98424e-42") @@ -2086,6 +2276,7 @@ class Program_weight_tensor_parameter_189: class Program_weight_tensor_parameter_190: name = "parameter_190" + original_name = "batch_norm_64.w_1" shape = [2048] dtype = "float32" min_val = float("-0.452017") @@ -2097,6 +2288,7 @@ class Program_weight_tensor_parameter_190: class Program_weight_tensor_parameter_191: name = "parameter_191" + original_name = "conv2d_64.w_0" shape = [2048, 1, 5, 5] dtype = "float32" min_val = float("-0.0917296") @@ -2108,6 +2300,7 @@ class Program_weight_tensor_parameter_191: class Program_weight_tensor_parameter_192: name = "parameter_192" + original_name = "batch_norm_63.b_0" shape = [2048] dtype = "float32" min_val = float("-1.50111") @@ -2119,6 +2312,7 @@ class Program_weight_tensor_parameter_192: class Program_weight_tensor_parameter_193: name = "parameter_193" + original_name = "batch_norm_63.w_0" shape = [2048] dtype = "float32" min_val = float("0.0685336") @@ -2130,6 +2324,7 @@ class Program_weight_tensor_parameter_193: class Program_weight_tensor_parameter_194: name = "parameter_194" + original_name = "batch_norm_63.w_2" shape = [2048] dtype = "float32" min_val = float("2.4845e-42") @@ -2141,6 +2336,7 @@ class Program_weight_tensor_parameter_194: class Program_weight_tensor_parameter_195: name = "parameter_195" + original_name = "batch_norm_63.w_1" shape = [2048] dtype = "float32" min_val = float("-1.14895") @@ -2152,6 +2348,7 @@ class Program_weight_tensor_parameter_195: class Program_weight_tensor_parameter_196: name = "parameter_196" + original_name = "conv2d_63.w_0" shape = [2048, 512, 1, 1] dtype = "float32" min_val = float("-0.279298") @@ -2163,6 +2360,7 @@ class Program_weight_tensor_parameter_196: class Program_weight_tensor_parameter_197: name = "parameter_197" + original_name = "batch_norm_62.b_0" shape = [512] dtype = "float32" min_val = float("-1.97274") @@ -2174,6 +2372,7 @@ class Program_weight_tensor_parameter_197: class Program_weight_tensor_parameter_198: name = "parameter_198" + original_name = "batch_norm_62.w_0" shape = [512] dtype = "float32" min_val = float("0.42346") @@ -2185,6 +2384,7 @@ class Program_weight_tensor_parameter_198: class Program_weight_tensor_parameter_199: name = "parameter_199" + original_name = "batch_norm_62.w_2" shape = [512] dtype = "float32" min_val = float("9.22635e-05") @@ -2196,6 +2396,7 @@ class Program_weight_tensor_parameter_199: class Program_weight_tensor_parameter_200: name = "parameter_200" + original_name = "batch_norm_62.w_1" shape = [512] dtype = "float32" min_val = float("-2.08117") @@ -2207,6 +2408,7 @@ class Program_weight_tensor_parameter_200: class Program_weight_tensor_parameter_201: name = "parameter_201" + original_name = "conv2d_62.w_0" shape = [512, 1, 5, 5] dtype = "float32" min_val = float("-0.216559") @@ -2218,6 +2420,7 @@ class Program_weight_tensor_parameter_201: class Program_weight_tensor_parameter_202: name = "parameter_202" + original_name = "batch_norm_61.b_0" shape = [512] dtype = "float32" min_val = float("-1.3417") @@ -2229,6 +2432,7 @@ class Program_weight_tensor_parameter_202: class Program_weight_tensor_parameter_203: name = "parameter_203" + original_name = "batch_norm_61.w_0" shape = [512] dtype = "float32" min_val = float("-0.164658") @@ -2240,6 +2444,7 @@ class Program_weight_tensor_parameter_203: class Program_weight_tensor_parameter_204: name = "parameter_204" + original_name = "batch_norm_61.w_2" shape = [512] dtype = "float32" min_val = float("0.0591604") @@ -2251,6 +2456,7 @@ class Program_weight_tensor_parameter_204: class Program_weight_tensor_parameter_205: name = "parameter_205" + original_name = "batch_norm_61.w_1" shape = [512] dtype = "float32" min_val = float("-1.21905") @@ -2262,6 +2468,7 @@ class Program_weight_tensor_parameter_205: class Program_weight_tensor_parameter_206: name = "parameter_206" + original_name = "conv2d_61.w_0" shape = [512, 2048, 1, 1] dtype = "float32" min_val = float("-0.314264") @@ -2273,6 +2480,7 @@ class Program_weight_tensor_parameter_206: class Program_weight_tensor_parameter_207: name = "parameter_207" + original_name = "batch_norm_60.b_0" shape = [2048] dtype = "float32" min_val = float("-6.96426") @@ -2284,6 +2492,7 @@ class Program_weight_tensor_parameter_207: class Program_weight_tensor_parameter_208: name = "parameter_208" + original_name = "batch_norm_60.w_0" shape = [2048] dtype = "float32" min_val = float("0.0131657") @@ -2295,6 +2504,7 @@ class Program_weight_tensor_parameter_208: class Program_weight_tensor_parameter_209: name = "parameter_209" + original_name = "batch_norm_60.w_2" shape = [2048] dtype = "float32" min_val = float("1.95481e-42") @@ -2306,6 +2516,7 @@ class Program_weight_tensor_parameter_209: class Program_weight_tensor_parameter_210: name = "parameter_210" + original_name = "batch_norm_60.w_1" shape = [2048] dtype = "float32" min_val = float("-0.523446") @@ -2317,6 +2528,7 @@ class Program_weight_tensor_parameter_210: class Program_weight_tensor_parameter_211: name = "parameter_211" + original_name = "conv2d_60.w_0" shape = [2048, 1, 5, 5] dtype = "float32" min_val = float("-0.160422") @@ -2328,6 +2540,7 @@ class Program_weight_tensor_parameter_211: class Program_weight_tensor_parameter_212: name = "parameter_212" + original_name = "batch_norm_59.b_0" shape = [2048] dtype = "float32" min_val = float("-1.52852") @@ -2339,6 +2552,7 @@ class Program_weight_tensor_parameter_212: class Program_weight_tensor_parameter_213: name = "parameter_213" + original_name = "batch_norm_59.w_0" shape = [2048] dtype = "float32" min_val = float("0.0320967") @@ -2350,6 +2564,7 @@ class Program_weight_tensor_parameter_213: class Program_weight_tensor_parameter_214: name = "parameter_214" + original_name = "batch_norm_59.w_2" shape = [2048] dtype = "float32" min_val = float("2.47189e-42") @@ -2361,6 +2576,7 @@ class Program_weight_tensor_parameter_214: class Program_weight_tensor_parameter_215: name = "parameter_215" + original_name = "batch_norm_59.w_1" shape = [2048] dtype = "float32" min_val = float("-1.22121") @@ -2372,6 +2588,7 @@ class Program_weight_tensor_parameter_215: class Program_weight_tensor_parameter_216: name = "parameter_216" + original_name = "conv2d_59.w_0" shape = [2048, 512, 1, 1] dtype = "float32" min_val = float("-0.355567") @@ -2383,6 +2600,7 @@ class Program_weight_tensor_parameter_216: class Program_weight_tensor_parameter_217: name = "parameter_217" + original_name = "batch_norm_58.b_0" shape = [512] dtype = "float32" min_val = float("-2.40458") @@ -2394,6 +2612,7 @@ class Program_weight_tensor_parameter_217: class Program_weight_tensor_parameter_218: name = "parameter_218" + original_name = "batch_norm_58.w_0" shape = [512] dtype = "float32" min_val = float("-0.0143071") @@ -2405,6 +2624,7 @@ class Program_weight_tensor_parameter_218: class Program_weight_tensor_parameter_219: name = "parameter_219" + original_name = "batch_norm_58.w_2" shape = [512] dtype = "float32" min_val = float("4.04268e-05") @@ -2416,6 +2636,7 @@ class Program_weight_tensor_parameter_219: class Program_weight_tensor_parameter_220: name = "parameter_220" + original_name = "batch_norm_58.w_1" shape = [512] dtype = "float32" min_val = float("-1.92304") @@ -2427,6 +2648,7 @@ class Program_weight_tensor_parameter_220: class Program_weight_tensor_parameter_221: name = "parameter_221" + original_name = "conv2d_58.w_0" shape = [512, 1, 5, 5] dtype = "float32" min_val = float("-0.224531") @@ -2438,6 +2660,7 @@ class Program_weight_tensor_parameter_221: class Program_weight_tensor_parameter_222: name = "parameter_222" + original_name = "batch_norm_57.b_0" shape = [512] dtype = "float32" min_val = float("-8.78578") @@ -2449,6 +2672,7 @@ class Program_weight_tensor_parameter_222: class Program_weight_tensor_parameter_223: name = "parameter_223" + original_name = "batch_norm_57.w_0" shape = [512] dtype = "float32" min_val = float("0.071092") @@ -2460,6 +2684,7 @@ class Program_weight_tensor_parameter_223: class Program_weight_tensor_parameter_224: name = "parameter_224" + original_name = "batch_norm_57.w_2" shape = [512] dtype = "float32" min_val = float("0.103099") @@ -2471,6 +2696,7 @@ class Program_weight_tensor_parameter_224: class Program_weight_tensor_parameter_225: name = "parameter_225" + original_name = "batch_norm_57.w_1" shape = [512] dtype = "float32" min_val = float("-3.47004") @@ -2482,6 +2708,7 @@ class Program_weight_tensor_parameter_225: class Program_weight_tensor_parameter_226: name = "parameter_226" + original_name = "conv2d_57.w_0" shape = [512, 768, 1, 1] dtype = "float32" min_val = float("-0.283037") @@ -2493,6 +2720,7 @@ class Program_weight_tensor_parameter_226: class Program_weight_tensor_parameter_227: name = "parameter_227" + original_name = "batch_norm_56.b_0" shape = [768] dtype = "float32" min_val = float("-10.1643") @@ -2504,6 +2732,7 @@ class Program_weight_tensor_parameter_227: class Program_weight_tensor_parameter_228: name = "parameter_228" + original_name = "batch_norm_56.w_0" shape = [768] dtype = "float32" min_val = float("0.564954") @@ -2515,6 +2744,7 @@ class Program_weight_tensor_parameter_228: class Program_weight_tensor_parameter_229: name = "parameter_229" + original_name = "batch_norm_56.w_2" shape = [768] dtype = "float32" min_val = float("2.95113e-42") @@ -2526,6 +2756,7 @@ class Program_weight_tensor_parameter_229: class Program_weight_tensor_parameter_230: name = "parameter_230" + original_name = "batch_norm_56.w_1" shape = [768] dtype = "float32" min_val = float("-0.851233") @@ -2537,6 +2768,7 @@ class Program_weight_tensor_parameter_230: class Program_weight_tensor_parameter_231: name = "parameter_231" + original_name = "conv2d_56.w_0" shape = [768, 1, 5, 5] dtype = "float32" min_val = float("-0.0892376") @@ -2548,6 +2780,7 @@ class Program_weight_tensor_parameter_231: class Program_weight_tensor_parameter_232: name = "parameter_232" + original_name = "batch_norm_55.b_0" shape = [768] dtype = "float32" min_val = float("-1.36733") @@ -2559,6 +2792,7 @@ class Program_weight_tensor_parameter_232: class Program_weight_tensor_parameter_233: name = "parameter_233" + original_name = "batch_norm_55.w_0" shape = [768] dtype = "float32" min_val = float("0.156658") @@ -2570,6 +2804,7 @@ class Program_weight_tensor_parameter_233: class Program_weight_tensor_parameter_234: name = "parameter_234" + original_name = "batch_norm_55.w_2" shape = [768] dtype = "float32" min_val = float("3.33649e-42") @@ -2581,6 +2816,7 @@ class Program_weight_tensor_parameter_234: class Program_weight_tensor_parameter_235: name = "parameter_235" + original_name = "batch_norm_55.w_1" shape = [768] dtype = "float32" min_val = float("-0.737875") @@ -2592,6 +2828,7 @@ class Program_weight_tensor_parameter_235: class Program_weight_tensor_parameter_236: name = "parameter_236" + original_name = "conv2d_55.w_0" shape = [768, 192, 1, 1] dtype = "float32" min_val = float("-0.247549") @@ -2603,6 +2840,7 @@ class Program_weight_tensor_parameter_236: class Program_weight_tensor_parameter_237: name = "parameter_237" + original_name = "batch_norm_54.b_0" shape = [192] dtype = "float32" min_val = float("-2.71942") @@ -2614,6 +2852,7 @@ class Program_weight_tensor_parameter_237: class Program_weight_tensor_parameter_238: name = "parameter_238" + original_name = "batch_norm_54.w_0" shape = [192] dtype = "float32" min_val = float("0.346059") @@ -2625,6 +2864,7 @@ class Program_weight_tensor_parameter_238: class Program_weight_tensor_parameter_239: name = "parameter_239" + original_name = "batch_norm_54.w_2" shape = [192] dtype = "float32" min_val = float("0.0556382") @@ -2636,6 +2876,7 @@ class Program_weight_tensor_parameter_239: class Program_weight_tensor_parameter_240: name = "parameter_240" + original_name = "batch_norm_54.w_1" shape = [192] dtype = "float32" min_val = float("-2.05304") @@ -2647,6 +2888,7 @@ class Program_weight_tensor_parameter_240: class Program_weight_tensor_parameter_241: name = "parameter_241" + original_name = "conv2d_54.w_0" shape = [192, 1, 5, 5] dtype = "float32" min_val = float("-0.227452") @@ -2658,6 +2900,7 @@ class Program_weight_tensor_parameter_241: class Program_weight_tensor_parameter_242: name = "parameter_242" + original_name = "batch_norm_53.b_0" shape = [192] dtype = "float32" min_val = float("-1.07126") @@ -2669,6 +2912,7 @@ class Program_weight_tensor_parameter_242: class Program_weight_tensor_parameter_243: name = "parameter_243" + original_name = "batch_norm_53.w_0" shape = [192] dtype = "float32" min_val = float("-0.21205") @@ -2680,6 +2924,7 @@ class Program_weight_tensor_parameter_243: class Program_weight_tensor_parameter_244: name = "parameter_244" + original_name = "batch_norm_53.w_2" shape = [192] dtype = "float32" min_val = float("0.00144158") @@ -2691,6 +2936,7 @@ class Program_weight_tensor_parameter_244: class Program_weight_tensor_parameter_245: name = "parameter_245" + original_name = "batch_norm_53.w_1" shape = [192] dtype = "float32" min_val = float("-0.106335") @@ -2702,6 +2948,7 @@ class Program_weight_tensor_parameter_245: class Program_weight_tensor_parameter_246: name = "parameter_246" + original_name = "conv2d_53.w_0" shape = [192, 768, 1, 1] dtype = "float32" min_val = float("-0.222737") @@ -2713,6 +2960,7 @@ class Program_weight_tensor_parameter_246: class Program_weight_tensor_parameter_247: name = "parameter_247" + original_name = "batch_norm_52.b_0" shape = [768] dtype = "float32" min_val = float("-4.94233") @@ -2724,6 +2972,7 @@ class Program_weight_tensor_parameter_247: class Program_weight_tensor_parameter_248: name = "parameter_248" + original_name = "batch_norm_52.w_0" shape = [768] dtype = "float32" min_val = float("0.027454") @@ -2735,6 +2984,7 @@ class Program_weight_tensor_parameter_248: class Program_weight_tensor_parameter_249: name = "parameter_249" + original_name = "batch_norm_52.w_2" shape = [768] dtype = "float32" min_val = float("2.88527e-42") @@ -2746,6 +2996,7 @@ class Program_weight_tensor_parameter_249: class Program_weight_tensor_parameter_250: name = "parameter_250" + original_name = "batch_norm_52.w_1" shape = [768] dtype = "float32" min_val = float("-0.803496") @@ -2757,6 +3008,7 @@ class Program_weight_tensor_parameter_250: class Program_weight_tensor_parameter_251: name = "parameter_251" + original_name = "conv2d_52.w_0" shape = [768, 192, 1, 1] dtype = "float32" min_val = float("-0.161753") @@ -2768,6 +3020,7 @@ class Program_weight_tensor_parameter_251: class Program_weight_tensor_parameter_252: name = "parameter_252" + original_name = "batch_norm_51.b_0" shape = [192] dtype = "float32" min_val = float("-2.37032") @@ -2779,6 +3032,7 @@ class Program_weight_tensor_parameter_252: class Program_weight_tensor_parameter_253: name = "parameter_253" + original_name = "batch_norm_51.w_0" shape = [192] dtype = "float32" min_val = float("0.299585") @@ -2790,6 +3044,7 @@ class Program_weight_tensor_parameter_253: class Program_weight_tensor_parameter_254: name = "parameter_254" + original_name = "batch_norm_51.w_2" shape = [192] dtype = "float32" min_val = float("0.0157586") @@ -2801,6 +3056,7 @@ class Program_weight_tensor_parameter_254: class Program_weight_tensor_parameter_255: name = "parameter_255" + original_name = "batch_norm_51.w_1" shape = [192] dtype = "float32" min_val = float("-1.83533") @@ -2812,6 +3068,7 @@ class Program_weight_tensor_parameter_255: class Program_weight_tensor_parameter_256: name = "parameter_256" + original_name = "conv2d_51.w_0" shape = [192, 1, 3, 3] dtype = "float32" min_val = float("-0.156214") @@ -2823,6 +3080,7 @@ class Program_weight_tensor_parameter_256: class Program_weight_tensor_parameter_257: name = "parameter_257" + original_name = "batch_norm_50.b_0" shape = [192] dtype = "float32" min_val = float("-2.77558") @@ -2834,6 +3092,7 @@ class Program_weight_tensor_parameter_257: class Program_weight_tensor_parameter_258: name = "parameter_258" + original_name = "batch_norm_50.w_0" shape = [192] dtype = "float32" min_val = float("-0.393058") @@ -2845,6 +3104,7 @@ class Program_weight_tensor_parameter_258: class Program_weight_tensor_parameter_259: name = "parameter_259" + original_name = "batch_norm_50.w_2" shape = [192] dtype = "float32" min_val = float("0.0362127") @@ -2856,6 +3116,7 @@ class Program_weight_tensor_parameter_259: class Program_weight_tensor_parameter_260: name = "parameter_260" + original_name = "batch_norm_50.w_1" shape = [192] dtype = "float32" min_val = float("-1.43532") @@ -2867,6 +3128,7 @@ class Program_weight_tensor_parameter_260: class Program_weight_tensor_parameter_261: name = "parameter_261" + original_name = "conv2d_50.w_0" shape = [192, 768, 1, 1] dtype = "float32" min_val = float("-0.177188") @@ -2878,6 +3140,7 @@ class Program_weight_tensor_parameter_261: class Program_weight_tensor_parameter_262: name = "parameter_262" + original_name = "batch_norm_49.b_0" shape = [768] dtype = "float32" min_val = float("-4.81109") @@ -2889,6 +3152,7 @@ class Program_weight_tensor_parameter_262: class Program_weight_tensor_parameter_263: name = "parameter_263" + original_name = "batch_norm_49.w_0" shape = [768] dtype = "float32" min_val = float("0.0715181") @@ -2900,6 +3164,7 @@ class Program_weight_tensor_parameter_263: class Program_weight_tensor_parameter_264: name = "parameter_264" + original_name = "batch_norm_49.w_2" shape = [768] dtype = "float32" min_val = float("2.02628e-42") @@ -2911,6 +3176,7 @@ class Program_weight_tensor_parameter_264: class Program_weight_tensor_parameter_265: name = "parameter_265" + original_name = "batch_norm_49.w_1" shape = [768] dtype = "float32" min_val = float("-0.181838") @@ -2922,6 +3188,7 @@ class Program_weight_tensor_parameter_265: class Program_weight_tensor_parameter_266: name = "parameter_266" + original_name = "conv2d_49.w_0" shape = [768, 1, 3, 3] dtype = "float32" min_val = float("-0.144689") @@ -2933,6 +3200,7 @@ class Program_weight_tensor_parameter_266: class Program_weight_tensor_parameter_267: name = "parameter_267" + original_name = "batch_norm_48.b_0" shape = [768] dtype = "float32" min_val = float("-1.74935") @@ -2944,6 +3212,7 @@ class Program_weight_tensor_parameter_267: class Program_weight_tensor_parameter_268: name = "parameter_268" + original_name = "batch_norm_48.w_0" shape = [768] dtype = "float32" min_val = float("0.101392") @@ -2955,6 +3224,7 @@ class Program_weight_tensor_parameter_268: class Program_weight_tensor_parameter_269: name = "parameter_269" + original_name = "batch_norm_48.w_2" shape = [768] dtype = "float32" min_val = float("2.43406e-42") @@ -2966,6 +3236,7 @@ class Program_weight_tensor_parameter_269: class Program_weight_tensor_parameter_270: name = "parameter_270" + original_name = "batch_norm_48.w_1" shape = [768] dtype = "float32" min_val = float("-0.64342") @@ -2977,6 +3248,7 @@ class Program_weight_tensor_parameter_270: class Program_weight_tensor_parameter_271: name = "parameter_271" + original_name = "conv2d_48.w_0" shape = [768, 192, 1, 1] dtype = "float32" min_val = float("-0.189628") @@ -2988,6 +3260,7 @@ class Program_weight_tensor_parameter_271: class Program_weight_tensor_parameter_272: name = "parameter_272" + original_name = "batch_norm_47.b_0" shape = [192] dtype = "float32" min_val = float("-2.29124") @@ -2999,6 +3272,7 @@ class Program_weight_tensor_parameter_272: class Program_weight_tensor_parameter_273: name = "parameter_273" + original_name = "batch_norm_47.w_0" shape = [192] dtype = "float32" min_val = float("-0.389975") @@ -3010,6 +3284,7 @@ class Program_weight_tensor_parameter_273: class Program_weight_tensor_parameter_274: name = "parameter_274" + original_name = "batch_norm_47.w_2" shape = [192] dtype = "float32" min_val = float("0.0820127") @@ -3021,6 +3296,7 @@ class Program_weight_tensor_parameter_274: class Program_weight_tensor_parameter_275: name = "parameter_275" + original_name = "batch_norm_47.w_1" shape = [192] dtype = "float32" min_val = float("-0.838424") @@ -3032,6 +3308,7 @@ class Program_weight_tensor_parameter_275: class Program_weight_tensor_parameter_276: name = "parameter_276" + original_name = "conv2d_47.w_0" shape = [192, 1, 5, 5] dtype = "float32" min_val = float("-0.254345") @@ -3043,6 +3320,7 @@ class Program_weight_tensor_parameter_276: class Program_weight_tensor_parameter_277: name = "parameter_277" + original_name = "batch_norm_46.b_0" shape = [192] dtype = "float32" min_val = float("-3.82579") @@ -3054,6 +3332,7 @@ class Program_weight_tensor_parameter_277: class Program_weight_tensor_parameter_278: name = "parameter_278" + original_name = "batch_norm_46.w_0" shape = [192] dtype = "float32" min_val = float("-0.141922") @@ -3065,6 +3344,7 @@ class Program_weight_tensor_parameter_278: class Program_weight_tensor_parameter_279: name = "parameter_279" + original_name = "batch_norm_46.w_2" shape = [192] dtype = "float32" min_val = float("0.042376") @@ -3076,6 +3356,7 @@ class Program_weight_tensor_parameter_279: class Program_weight_tensor_parameter_280: name = "parameter_280" + original_name = "batch_norm_46.w_1" shape = [192] dtype = "float32" min_val = float("-1.16666") @@ -3087,6 +3368,7 @@ class Program_weight_tensor_parameter_280: class Program_weight_tensor_parameter_281: name = "parameter_281" + original_name = "conv2d_46.w_0" shape = [192, 768, 1, 1] dtype = "float32" min_val = float("-0.19765") @@ -3098,6 +3380,7 @@ class Program_weight_tensor_parameter_281: class Program_weight_tensor_parameter_282: name = "parameter_282" + original_name = "batch_norm_45.b_0" shape = [768] dtype = "float32" min_val = float("-5.55847") @@ -3109,6 +3392,7 @@ class Program_weight_tensor_parameter_282: class Program_weight_tensor_parameter_283: name = "parameter_283" + original_name = "batch_norm_45.w_0" shape = [768] dtype = "float32" min_val = float("-0.00147949") @@ -3120,6 +3404,7 @@ class Program_weight_tensor_parameter_283: class Program_weight_tensor_parameter_284: name = "parameter_284" + original_name = "batch_norm_45.w_2" shape = [768] dtype = "float32" min_val = float("1.98144e-42") @@ -3131,6 +3416,7 @@ class Program_weight_tensor_parameter_284: class Program_weight_tensor_parameter_285: name = "parameter_285" + original_name = "batch_norm_45.w_1" shape = [768] dtype = "float32" min_val = float("-0.200372") @@ -3142,6 +3428,7 @@ class Program_weight_tensor_parameter_285: class Program_weight_tensor_parameter_286: name = "parameter_286" + original_name = "conv2d_45.w_0" shape = [768, 1, 3, 3] dtype = "float32" min_val = float("-0.11578") @@ -3153,6 +3440,7 @@ class Program_weight_tensor_parameter_286: class Program_weight_tensor_parameter_287: name = "parameter_287" + original_name = "batch_norm_44.b_0" shape = [768] dtype = "float32" min_val = float("-1.33364") @@ -3164,6 +3452,7 @@ class Program_weight_tensor_parameter_287: class Program_weight_tensor_parameter_288: name = "parameter_288" + original_name = "batch_norm_44.w_0" shape = [768] dtype = "float32" min_val = float("0.119102") @@ -3175,6 +3464,7 @@ class Program_weight_tensor_parameter_288: class Program_weight_tensor_parameter_289: name = "parameter_289" + original_name = "batch_norm_44.w_2" shape = [768] dtype = "float32" min_val = float("2.47189e-42") @@ -3186,6 +3476,7 @@ class Program_weight_tensor_parameter_289: class Program_weight_tensor_parameter_290: name = "parameter_290" + original_name = "batch_norm_44.w_1" shape = [768] dtype = "float32" min_val = float("-0.66603") @@ -3197,6 +3488,7 @@ class Program_weight_tensor_parameter_290: class Program_weight_tensor_parameter_291: name = "parameter_291" + original_name = "conv2d_44.w_0" shape = [768, 192, 1, 1] dtype = "float32" min_val = float("-0.157796") @@ -3208,6 +3500,7 @@ class Program_weight_tensor_parameter_291: class Program_weight_tensor_parameter_292: name = "parameter_292" + original_name = "batch_norm_43.b_0" shape = [192] dtype = "float32" min_val = float("-1.97234") @@ -3219,6 +3512,7 @@ class Program_weight_tensor_parameter_292: class Program_weight_tensor_parameter_293: name = "parameter_293" + original_name = "batch_norm_43.w_0" shape = [192] dtype = "float32" min_val = float("-0.176767") @@ -3230,6 +3524,7 @@ class Program_weight_tensor_parameter_293: class Program_weight_tensor_parameter_294: name = "parameter_294" + original_name = "batch_norm_43.w_2" shape = [192] dtype = "float32" min_val = float("0.0455728") @@ -3241,6 +3536,7 @@ class Program_weight_tensor_parameter_294: class Program_weight_tensor_parameter_295: name = "parameter_295" + original_name = "batch_norm_43.w_1" shape = [192] dtype = "float32" min_val = float("-2.18087") @@ -3252,6 +3548,7 @@ class Program_weight_tensor_parameter_295: class Program_weight_tensor_parameter_296: name = "parameter_296" + original_name = "conv2d_43.w_0" shape = [192, 1, 5, 5] dtype = "float32" min_val = float("-0.248283") @@ -3263,6 +3560,7 @@ class Program_weight_tensor_parameter_296: class Program_weight_tensor_parameter_297: name = "parameter_297" + original_name = "batch_norm_42.b_0" shape = [192] dtype = "float32" min_val = float("-2.70189") @@ -3274,6 +3572,7 @@ class Program_weight_tensor_parameter_297: class Program_weight_tensor_parameter_298: name = "parameter_298" + original_name = "batch_norm_42.w_0" shape = [192] dtype = "float32" min_val = float("-0.266124") @@ -3285,6 +3584,7 @@ class Program_weight_tensor_parameter_298: class Program_weight_tensor_parameter_299: name = "parameter_299" + original_name = "batch_norm_42.w_2" shape = [192] dtype = "float32" min_val = float("0.0171015") @@ -3296,6 +3596,7 @@ class Program_weight_tensor_parameter_299: class Program_weight_tensor_parameter_300: name = "parameter_300" + original_name = "batch_norm_42.w_1" shape = [192] dtype = "float32" min_val = float("-0.880041") @@ -3307,6 +3608,7 @@ class Program_weight_tensor_parameter_300: class Program_weight_tensor_parameter_301: name = "parameter_301" + original_name = "conv2d_42.w_0" shape = [192, 768, 1, 1] dtype = "float32" min_val = float("-0.187466") @@ -3318,6 +3620,7 @@ class Program_weight_tensor_parameter_301: class Program_weight_tensor_parameter_302: name = "parameter_302" + original_name = "batch_norm_41.b_0" shape = [768] dtype = "float32" min_val = float("-5.0374") @@ -3329,6 +3632,7 @@ class Program_weight_tensor_parameter_302: class Program_weight_tensor_parameter_303: name = "parameter_303" + original_name = "batch_norm_41.w_0" shape = [768] dtype = "float32" min_val = float("-0.0173342") @@ -3340,6 +3644,7 @@ class Program_weight_tensor_parameter_303: class Program_weight_tensor_parameter_304: name = "parameter_304" + original_name = "batch_norm_41.w_2" shape = [768] dtype = "float32" min_val = float("2.10755e-42") @@ -3351,6 +3656,7 @@ class Program_weight_tensor_parameter_304: class Program_weight_tensor_parameter_305: name = "parameter_305" + original_name = "batch_norm_41.w_1" shape = [768] dtype = "float32" min_val = float("-0.251727") @@ -3362,6 +3668,7 @@ class Program_weight_tensor_parameter_305: class Program_weight_tensor_parameter_306: name = "parameter_306" + original_name = "conv2d_41.w_0" shape = [768, 1, 3, 3] dtype = "float32" min_val = float("-0.113377") @@ -3373,6 +3680,7 @@ class Program_weight_tensor_parameter_306: class Program_weight_tensor_parameter_307: name = "parameter_307" + original_name = "batch_norm_40.b_0" shape = [768] dtype = "float32" min_val = float("-1.4694") @@ -3384,6 +3692,7 @@ class Program_weight_tensor_parameter_307: class Program_weight_tensor_parameter_308: name = "parameter_308" + original_name = "batch_norm_40.w_0" shape = [768] dtype = "float32" min_val = float("0.0864175") @@ -3395,6 +3704,7 @@ class Program_weight_tensor_parameter_308: class Program_weight_tensor_parameter_309: name = "parameter_309" + original_name = "batch_norm_40.w_2" shape = [768] dtype = "float32" min_val = float("2.47329e-42") @@ -3406,6 +3716,7 @@ class Program_weight_tensor_parameter_309: class Program_weight_tensor_parameter_310: name = "parameter_310" + original_name = "batch_norm_40.w_1" shape = [768] dtype = "float32" min_val = float("-0.634681") @@ -3417,6 +3728,7 @@ class Program_weight_tensor_parameter_310: class Program_weight_tensor_parameter_311: name = "parameter_311" + original_name = "conv2d_40.w_0" shape = [768, 192, 1, 1] dtype = "float32" min_val = float("-0.234014") @@ -3428,6 +3740,7 @@ class Program_weight_tensor_parameter_311: class Program_weight_tensor_parameter_312: name = "parameter_312" + original_name = "batch_norm_39.b_0" shape = [192] dtype = "float32" min_val = float("-1.71944") @@ -3439,6 +3752,7 @@ class Program_weight_tensor_parameter_312: class Program_weight_tensor_parameter_313: name = "parameter_313" + original_name = "batch_norm_39.w_0" shape = [192] dtype = "float32" min_val = float("0.179929") @@ -3450,6 +3764,7 @@ class Program_weight_tensor_parameter_313: class Program_weight_tensor_parameter_314: name = "parameter_314" + original_name = "batch_norm_39.w_2" shape = [192] dtype = "float32" min_val = float("0.0108757") @@ -3461,6 +3776,7 @@ class Program_weight_tensor_parameter_314: class Program_weight_tensor_parameter_315: name = "parameter_315" + original_name = "batch_norm_39.w_1" shape = [192] dtype = "float32" min_val = float("-1.60283") @@ -3472,6 +3788,7 @@ class Program_weight_tensor_parameter_315: class Program_weight_tensor_parameter_316: name = "parameter_316" + original_name = "conv2d_39.w_0" shape = [192, 1, 5, 5] dtype = "float32" min_val = float("-0.254901") @@ -3483,6 +3800,7 @@ class Program_weight_tensor_parameter_316: class Program_weight_tensor_parameter_317: name = "parameter_317" + original_name = "batch_norm_38.b_0" shape = [192] dtype = "float32" min_val = float("-2.03767") @@ -3494,6 +3812,7 @@ class Program_weight_tensor_parameter_317: class Program_weight_tensor_parameter_318: name = "parameter_318" + original_name = "batch_norm_38.w_0" shape = [192] dtype = "float32" min_val = float("-0.130529") @@ -3505,6 +3824,7 @@ class Program_weight_tensor_parameter_318: class Program_weight_tensor_parameter_319: name = "parameter_319" + original_name = "batch_norm_38.w_2" shape = [192] dtype = "float32" min_val = float("0.0311346") @@ -3516,6 +3836,7 @@ class Program_weight_tensor_parameter_319: class Program_weight_tensor_parameter_320: name = "parameter_320" + original_name = "batch_norm_38.w_1" shape = [192] dtype = "float32" min_val = float("-0.798163") @@ -3527,6 +3848,7 @@ class Program_weight_tensor_parameter_320: class Program_weight_tensor_parameter_321: name = "parameter_321" + original_name = "conv2d_38.w_0" shape = [192, 768, 1, 1] dtype = "float32" min_val = float("-0.178118") @@ -3538,6 +3860,7 @@ class Program_weight_tensor_parameter_321: class Program_weight_tensor_parameter_322: name = "parameter_322" + original_name = "batch_norm_37.b_0" shape = [768] dtype = "float32" min_val = float("-11.2585") @@ -3549,6 +3872,7 @@ class Program_weight_tensor_parameter_322: class Program_weight_tensor_parameter_323: name = "parameter_323" + original_name = "batch_norm_37.w_0" shape = [768] dtype = "float32" min_val = float("0.0593271") @@ -3560,6 +3884,7 @@ class Program_weight_tensor_parameter_323: class Program_weight_tensor_parameter_324: name = "parameter_324" + original_name = "batch_norm_37.w_2" shape = [768] dtype = "float32" min_val = float("2.02908e-42") @@ -3571,6 +3896,7 @@ class Program_weight_tensor_parameter_324: class Program_weight_tensor_parameter_325: name = "parameter_325" + original_name = "batch_norm_37.w_1" shape = [768] dtype = "float32" min_val = float("-0.269337") @@ -3582,6 +3908,7 @@ class Program_weight_tensor_parameter_325: class Program_weight_tensor_parameter_326: name = "parameter_326" + original_name = "conv2d_37.w_0" shape = [768, 1, 3, 3] dtype = "float32" min_val = float("-0.0539496") @@ -3593,6 +3920,7 @@ class Program_weight_tensor_parameter_326: class Program_weight_tensor_parameter_327: name = "parameter_327" + original_name = "batch_norm_36.b_0" shape = [768] dtype = "float32" min_val = float("-1.76569") @@ -3604,6 +3932,7 @@ class Program_weight_tensor_parameter_327: class Program_weight_tensor_parameter_328: name = "parameter_328" + original_name = "batch_norm_36.w_0" shape = [768] dtype = "float32" min_val = float("0.0366888") @@ -3615,6 +3944,7 @@ class Program_weight_tensor_parameter_328: class Program_weight_tensor_parameter_329: name = "parameter_329" + original_name = "batch_norm_36.w_2" shape = [768] dtype = "float32" min_val = float("2.47189e-42") @@ -3626,6 +3956,7 @@ class Program_weight_tensor_parameter_329: class Program_weight_tensor_parameter_330: name = "parameter_330" + original_name = "batch_norm_36.w_1" shape = [768] dtype = "float32" min_val = float("-0.551648") @@ -3637,6 +3968,7 @@ class Program_weight_tensor_parameter_330: class Program_weight_tensor_parameter_331: name = "parameter_331" + original_name = "conv2d_36.w_0" shape = [768, 192, 1, 1] dtype = "float32" min_val = float("-0.214242") @@ -3648,6 +3980,7 @@ class Program_weight_tensor_parameter_331: class Program_weight_tensor_parameter_332: name = "parameter_332" + original_name = "batch_norm_35.b_0" shape = [192] dtype = "float32" min_val = float("-2.00186") @@ -3659,6 +3992,7 @@ class Program_weight_tensor_parameter_332: class Program_weight_tensor_parameter_333: name = "parameter_333" + original_name = "batch_norm_35.w_0" shape = [192] dtype = "float32" min_val = float("-0.158289") @@ -3670,6 +4004,7 @@ class Program_weight_tensor_parameter_333: class Program_weight_tensor_parameter_334: name = "parameter_334" + original_name = "batch_norm_35.w_2" shape = [192] dtype = "float32" min_val = float("0.00339113") @@ -3681,6 +4016,7 @@ class Program_weight_tensor_parameter_334: class Program_weight_tensor_parameter_335: name = "parameter_335" + original_name = "batch_norm_35.w_1" shape = [192] dtype = "float32" min_val = float("-0.941446") @@ -3692,6 +4028,7 @@ class Program_weight_tensor_parameter_335: class Program_weight_tensor_parameter_336: name = "parameter_336" + original_name = "conv2d_35.w_0" shape = [192, 1, 5, 5] dtype = "float32" min_val = float("-0.21561") @@ -3703,6 +4040,7 @@ class Program_weight_tensor_parameter_336: class Program_weight_tensor_parameter_337: name = "parameter_337" + original_name = "batch_norm_34.b_0" shape = [192] dtype = "float32" min_val = float("-1.42927") @@ -3714,6 +4052,7 @@ class Program_weight_tensor_parameter_337: class Program_weight_tensor_parameter_338: name = "parameter_338" + original_name = "batch_norm_34.w_0" shape = [192] dtype = "float32" min_val = float("-0.130843") @@ -3725,6 +4064,7 @@ class Program_weight_tensor_parameter_338: class Program_weight_tensor_parameter_339: name = "parameter_339" + original_name = "batch_norm_34.w_2" shape = [192] dtype = "float32" min_val = float("0.0165322") @@ -3736,6 +4076,7 @@ class Program_weight_tensor_parameter_339: class Program_weight_tensor_parameter_340: name = "parameter_340" + original_name = "batch_norm_34.w_1" shape = [192] dtype = "float32" min_val = float("-1.15061") @@ -3747,6 +4088,7 @@ class Program_weight_tensor_parameter_340: class Program_weight_tensor_parameter_341: name = "parameter_341" + original_name = "conv2d_34.w_0" shape = [192, 768, 1, 1] dtype = "float32" min_val = float("-0.265387") @@ -3758,6 +4100,7 @@ class Program_weight_tensor_parameter_341: class Program_weight_tensor_parameter_342: name = "parameter_342" + original_name = "batch_norm_33.b_0" shape = [768] dtype = "float32" min_val = float("-7.40707") @@ -3769,6 +4112,7 @@ class Program_weight_tensor_parameter_342: class Program_weight_tensor_parameter_343: name = "parameter_343" + original_name = "batch_norm_33.w_0" shape = [768] dtype = "float32" min_val = float("0.00688959") @@ -3780,6 +4124,7 @@ class Program_weight_tensor_parameter_343: class Program_weight_tensor_parameter_344: name = "parameter_344" + original_name = "batch_norm_33.w_2" shape = [768] dtype = "float32" min_val = float("2.12577e-42") @@ -3791,6 +4136,7 @@ class Program_weight_tensor_parameter_344: class Program_weight_tensor_parameter_345: name = "parameter_345" + original_name = "batch_norm_33.w_1" shape = [768] dtype = "float32" min_val = float("-0.259311") @@ -3802,6 +4148,7 @@ class Program_weight_tensor_parameter_345: class Program_weight_tensor_parameter_346: name = "parameter_346" + original_name = "conv2d_33.w_0" shape = [768, 1, 3, 3] dtype = "float32" min_val = float("-0.087711") @@ -3813,6 +4160,7 @@ class Program_weight_tensor_parameter_346: class Program_weight_tensor_parameter_347: name = "parameter_347" + original_name = "batch_norm_32.b_0" shape = [768] dtype = "float32" min_val = float("-1.45382") @@ -3824,6 +4172,7 @@ class Program_weight_tensor_parameter_347: class Program_weight_tensor_parameter_348: name = "parameter_348" + original_name = "batch_norm_32.w_0" shape = [768] dtype = "float32" min_val = float("0.110399") @@ -3835,6 +4184,7 @@ class Program_weight_tensor_parameter_348: class Program_weight_tensor_parameter_349: name = "parameter_349" + original_name = "batch_norm_32.w_2" shape = [768] dtype = "float32" min_val = float("2.46769e-42") @@ -3846,6 +4196,7 @@ class Program_weight_tensor_parameter_349: class Program_weight_tensor_parameter_350: name = "parameter_350" + original_name = "batch_norm_32.w_1" shape = [768] dtype = "float32" min_val = float("-0.601292") @@ -3857,6 +4208,7 @@ class Program_weight_tensor_parameter_350: class Program_weight_tensor_parameter_351: name = "parameter_351" + original_name = "conv2d_32.w_0" shape = [768, 192, 1, 1] dtype = "float32" min_val = float("-0.213222") @@ -3868,6 +4220,7 @@ class Program_weight_tensor_parameter_351: class Program_weight_tensor_parameter_352: name = "parameter_352" + original_name = "batch_norm_31.b_0" shape = [192] dtype = "float32" min_val = float("-2.69105") @@ -3879,6 +4232,7 @@ class Program_weight_tensor_parameter_352: class Program_weight_tensor_parameter_353: name = "parameter_353" + original_name = "batch_norm_31.w_0" shape = [192] dtype = "float32" min_val = float("0.301758") @@ -3890,6 +4244,7 @@ class Program_weight_tensor_parameter_353: class Program_weight_tensor_parameter_354: name = "parameter_354" + original_name = "batch_norm_31.w_2" shape = [192] dtype = "float32" min_val = float("0.000806838") @@ -3901,6 +4256,7 @@ class Program_weight_tensor_parameter_354: class Program_weight_tensor_parameter_355: name = "parameter_355" + original_name = "batch_norm_31.w_1" shape = [192] dtype = "float32" min_val = float("-1.19393") @@ -3912,6 +4268,7 @@ class Program_weight_tensor_parameter_355: class Program_weight_tensor_parameter_356: name = "parameter_356" + original_name = "conv2d_31.w_0" shape = [192, 1, 5, 5] dtype = "float32" min_val = float("-0.292655") @@ -3923,6 +4280,7 @@ class Program_weight_tensor_parameter_356: class Program_weight_tensor_parameter_357: name = "parameter_357" + original_name = "batch_norm_30.b_0" shape = [192] dtype = "float32" min_val = float("-1.48282") @@ -3934,6 +4292,7 @@ class Program_weight_tensor_parameter_357: class Program_weight_tensor_parameter_358: name = "parameter_358" + original_name = "batch_norm_30.w_0" shape = [192] dtype = "float32" min_val = float("-0.0668591") @@ -3945,6 +4304,7 @@ class Program_weight_tensor_parameter_358: class Program_weight_tensor_parameter_359: name = "parameter_359" + original_name = "batch_norm_30.w_2" shape = [192] dtype = "float32" min_val = float("0.00957581") @@ -3956,6 +4316,7 @@ class Program_weight_tensor_parameter_359: class Program_weight_tensor_parameter_360: name = "parameter_360" + original_name = "batch_norm_30.w_1" shape = [192] dtype = "float32" min_val = float("-0.507373") @@ -3967,6 +4328,7 @@ class Program_weight_tensor_parameter_360: class Program_weight_tensor_parameter_361: name = "parameter_361" + original_name = "conv2d_30.w_0" shape = [192, 768, 1, 1] dtype = "float32" min_val = float("-0.207309") @@ -3978,6 +4340,7 @@ class Program_weight_tensor_parameter_361: class Program_weight_tensor_parameter_362: name = "parameter_362" + original_name = "batch_norm_29.b_0" shape = [768] dtype = "float32" min_val = float("-11.029") @@ -3989,6 +4352,7 @@ class Program_weight_tensor_parameter_362: class Program_weight_tensor_parameter_363: name = "parameter_363" + original_name = "batch_norm_29.w_0" shape = [768] dtype = "float32" min_val = float("-0.00847595") @@ -4000,6 +4364,7 @@ class Program_weight_tensor_parameter_363: class Program_weight_tensor_parameter_364: name = "parameter_364" + original_name = "batch_norm_29.w_2" shape = [768] dtype = "float32" min_val = float("1.9394e-42") @@ -4011,6 +4376,7 @@ class Program_weight_tensor_parameter_364: class Program_weight_tensor_parameter_365: name = "parameter_365" + original_name = "batch_norm_29.w_1" shape = [768] dtype = "float32" min_val = float("-0.608728") @@ -4022,6 +4388,7 @@ class Program_weight_tensor_parameter_365: class Program_weight_tensor_parameter_366: name = "parameter_366" + original_name = "conv2d_29.w_0" shape = [768, 1, 5, 5] dtype = "float32" min_val = float("-0.196994") @@ -4033,6 +4400,7 @@ class Program_weight_tensor_parameter_366: class Program_weight_tensor_parameter_367: name = "parameter_367" + original_name = "batch_norm_28.b_0" shape = [768] dtype = "float32" min_val = float("-1.64133") @@ -4044,6 +4412,7 @@ class Program_weight_tensor_parameter_367: class Program_weight_tensor_parameter_368: name = "parameter_368" + original_name = "batch_norm_28.w_0" shape = [768] dtype = "float32" min_val = float("0.125451") @@ -4055,6 +4424,7 @@ class Program_weight_tensor_parameter_368: class Program_weight_tensor_parameter_369: name = "parameter_369" + original_name = "batch_norm_28.w_2" shape = [768] dtype = "float32" min_val = float("2.43686e-42") @@ -4066,6 +4436,7 @@ class Program_weight_tensor_parameter_369: class Program_weight_tensor_parameter_370: name = "parameter_370" + original_name = "batch_norm_28.w_1" shape = [768] dtype = "float32" min_val = float("-0.796606") @@ -4077,6 +4448,7 @@ class Program_weight_tensor_parameter_370: class Program_weight_tensor_parameter_371: name = "parameter_371" + original_name = "conv2d_28.w_0" shape = [768, 192, 1, 1] dtype = "float32" min_val = float("-0.222814") @@ -4088,6 +4460,7 @@ class Program_weight_tensor_parameter_371: class Program_weight_tensor_parameter_372: name = "parameter_372" + original_name = "batch_norm_27.b_0" shape = [192] dtype = "float32" min_val = float("-1.90915") @@ -4099,6 +4472,7 @@ class Program_weight_tensor_parameter_372: class Program_weight_tensor_parameter_373: name = "parameter_373" + original_name = "batch_norm_27.w_0" shape = [192] dtype = "float32" min_val = float("-0.247599") @@ -4110,6 +4484,7 @@ class Program_weight_tensor_parameter_373: class Program_weight_tensor_parameter_374: name = "parameter_374" + original_name = "batch_norm_27.w_2" shape = [192] dtype = "float32" min_val = float("0.000218568") @@ -4121,6 +4496,7 @@ class Program_weight_tensor_parameter_374: class Program_weight_tensor_parameter_375: name = "parameter_375" + original_name = "batch_norm_27.w_1" shape = [192] dtype = "float32" min_val = float("-0.673977") @@ -4132,6 +4508,7 @@ class Program_weight_tensor_parameter_375: class Program_weight_tensor_parameter_376: name = "parameter_376" + original_name = "conv2d_27.w_0" shape = [192, 1, 3, 3] dtype = "float32" min_val = float("-0.143129") @@ -4143,6 +4520,7 @@ class Program_weight_tensor_parameter_376: class Program_weight_tensor_parameter_377: name = "parameter_377" + original_name = "batch_norm_26.b_0" shape = [192] dtype = "float32" min_val = float("-1.07745") @@ -4154,6 +4532,7 @@ class Program_weight_tensor_parameter_377: class Program_weight_tensor_parameter_378: name = "parameter_378" + original_name = "batch_norm_26.w_0" shape = [192] dtype = "float32" min_val = float("0.0899232") @@ -4165,6 +4544,7 @@ class Program_weight_tensor_parameter_378: class Program_weight_tensor_parameter_379: name = "parameter_379" + original_name = "batch_norm_26.w_2" shape = [192] dtype = "float32" min_val = float("0.00912112") @@ -4176,6 +4556,7 @@ class Program_weight_tensor_parameter_379: class Program_weight_tensor_parameter_380: name = "parameter_380" + original_name = "batch_norm_26.w_1" shape = [192] dtype = "float32" min_val = float("-0.31149") @@ -4187,6 +4568,7 @@ class Program_weight_tensor_parameter_380: class Program_weight_tensor_parameter_381: name = "parameter_381" + original_name = "conv2d_26.w_0" shape = [192, 768, 1, 1] dtype = "float32" min_val = float("-0.193763") @@ -4198,6 +4580,7 @@ class Program_weight_tensor_parameter_381: class Program_weight_tensor_parameter_382: name = "parameter_382" + original_name = "batch_norm_25.b_0" shape = [768] dtype = "float32" min_val = float("-6.97645") @@ -4209,6 +4592,7 @@ class Program_weight_tensor_parameter_382: class Program_weight_tensor_parameter_383: name = "parameter_383" + original_name = "batch_norm_25.w_0" shape = [768] dtype = "float32" min_val = float("0.108562") @@ -4220,6 +4604,7 @@ class Program_weight_tensor_parameter_383: class Program_weight_tensor_parameter_384: name = "parameter_384" + original_name = "batch_norm_25.w_2" shape = [768] dtype = "float32" min_val = float("2.01787e-42") @@ -4231,6 +4616,7 @@ class Program_weight_tensor_parameter_384: class Program_weight_tensor_parameter_385: name = "parameter_385" + original_name = "batch_norm_25.w_1" shape = [768] dtype = "float32" min_val = float("-0.334285") @@ -4242,6 +4628,7 @@ class Program_weight_tensor_parameter_385: class Program_weight_tensor_parameter_386: name = "parameter_386" + original_name = "conv2d_25.w_0" shape = [768, 1, 3, 3] dtype = "float32" min_val = float("-0.22188") @@ -4253,6 +4640,7 @@ class Program_weight_tensor_parameter_386: class Program_weight_tensor_parameter_387: name = "parameter_387" + original_name = "batch_norm_24.b_0" shape = [768] dtype = "float32" min_val = float("-1.48744") @@ -4264,6 +4652,7 @@ class Program_weight_tensor_parameter_387: class Program_weight_tensor_parameter_388: name = "parameter_388" + original_name = "batch_norm_24.w_0" shape = [768] dtype = "float32" min_val = float("0.116942") @@ -4275,6 +4664,7 @@ class Program_weight_tensor_parameter_388: class Program_weight_tensor_parameter_389: name = "parameter_389" + original_name = "batch_norm_24.w_2" shape = [768] dtype = "float32" min_val = float("2.51813e-42") @@ -4286,6 +4676,7 @@ class Program_weight_tensor_parameter_389: class Program_weight_tensor_parameter_390: name = "parameter_390" + original_name = "batch_norm_24.w_1" shape = [768] dtype = "float32" min_val = float("-0.731045") @@ -4297,6 +4688,7 @@ class Program_weight_tensor_parameter_390: class Program_weight_tensor_parameter_391: name = "parameter_391" + original_name = "conv2d_24.w_0" shape = [768, 192, 1, 1] dtype = "float32" min_val = float("-0.207637") @@ -4308,6 +4700,7 @@ class Program_weight_tensor_parameter_391: class Program_weight_tensor_parameter_392: name = "parameter_392" + original_name = "batch_norm_23.b_0" shape = [192] dtype = "float32" min_val = float("-2.53921") @@ -4319,6 +4712,7 @@ class Program_weight_tensor_parameter_392: class Program_weight_tensor_parameter_393: name = "parameter_393" + original_name = "batch_norm_23.w_0" shape = [192] dtype = "float32" min_val = float("0.187758") @@ -4330,6 +4724,7 @@ class Program_weight_tensor_parameter_393: class Program_weight_tensor_parameter_394: name = "parameter_394" + original_name = "batch_norm_23.w_2" shape = [192] dtype = "float32" min_val = float("4.99664e-05") @@ -4341,6 +4736,7 @@ class Program_weight_tensor_parameter_394: class Program_weight_tensor_parameter_395: name = "parameter_395" + original_name = "batch_norm_23.w_1" shape = [192] dtype = "float32" min_val = float("-0.861841") @@ -4352,6 +4748,7 @@ class Program_weight_tensor_parameter_395: class Program_weight_tensor_parameter_396: name = "parameter_396" + original_name = "conv2d_23.w_0" shape = [192, 1, 3, 3] dtype = "float32" min_val = float("-0.240399") @@ -4363,6 +4760,7 @@ class Program_weight_tensor_parameter_396: class Program_weight_tensor_parameter_397: name = "parameter_397" + original_name = "batch_norm_22.b_0" shape = [192] dtype = "float32" min_val = float("-0.903546") @@ -4374,6 +4772,7 @@ class Program_weight_tensor_parameter_397: class Program_weight_tensor_parameter_398: name = "parameter_398" + original_name = "batch_norm_22.w_0" shape = [192] dtype = "float32" min_val = float("-0.141176") @@ -4385,6 +4784,7 @@ class Program_weight_tensor_parameter_398: class Program_weight_tensor_parameter_399: name = "parameter_399" + original_name = "batch_norm_22.w_2" shape = [192] dtype = "float32" min_val = float("0.0132501") @@ -4396,6 +4796,7 @@ class Program_weight_tensor_parameter_399: class Program_weight_tensor_parameter_400: name = "parameter_400" + original_name = "batch_norm_22.w_1" shape = [192] dtype = "float32" min_val = float("-0.442671") @@ -4407,6 +4808,7 @@ class Program_weight_tensor_parameter_400: class Program_weight_tensor_parameter_401: name = "parameter_401" + original_name = "conv2d_22.w_0" shape = [192, 768, 1, 1] dtype = "float32" min_val = float("-0.21894") @@ -4418,6 +4820,7 @@ class Program_weight_tensor_parameter_401: class Program_weight_tensor_parameter_402: name = "parameter_402" + original_name = "batch_norm_21.b_0" shape = [768] dtype = "float32" min_val = float("-6.50449") @@ -4429,6 +4832,7 @@ class Program_weight_tensor_parameter_402: class Program_weight_tensor_parameter_403: name = "parameter_403" + original_name = "batch_norm_21.w_0" shape = [768] dtype = "float32" min_val = float("0.0512646") @@ -4440,6 +4844,7 @@ class Program_weight_tensor_parameter_403: class Program_weight_tensor_parameter_404: name = "parameter_404" + original_name = "batch_norm_21.w_2" shape = [768] dtype = "float32" min_val = float("1.99125e-42") @@ -4451,6 +4856,7 @@ class Program_weight_tensor_parameter_404: class Program_weight_tensor_parameter_405: name = "parameter_405" + original_name = "batch_norm_21.w_1" shape = [768] dtype = "float32" min_val = float("-0.241305") @@ -4462,6 +4868,7 @@ class Program_weight_tensor_parameter_405: class Program_weight_tensor_parameter_406: name = "parameter_406" + original_name = "conv2d_21.w_0" shape = [768, 1, 3, 3] dtype = "float32" min_val = float("-0.128292") @@ -4473,6 +4880,7 @@ class Program_weight_tensor_parameter_406: class Program_weight_tensor_parameter_407: name = "parameter_407" + original_name = "batch_norm_20.b_0" shape = [768] dtype = "float32" min_val = float("-1.50322") @@ -4484,6 +4892,7 @@ class Program_weight_tensor_parameter_407: class Program_weight_tensor_parameter_408: name = "parameter_408" + original_name = "batch_norm_20.w_0" shape = [768] dtype = "float32" min_val = float("0.13239") @@ -4495,6 +4904,7 @@ class Program_weight_tensor_parameter_408: class Program_weight_tensor_parameter_409: name = "parameter_409" + original_name = "batch_norm_20.w_2" shape = [768] dtype = "float32" min_val = float("2.49151e-42") @@ -4506,6 +4916,7 @@ class Program_weight_tensor_parameter_409: class Program_weight_tensor_parameter_410: name = "parameter_410" + original_name = "batch_norm_20.w_1" shape = [768] dtype = "float32" min_val = float("-0.629727") @@ -4517,6 +4928,7 @@ class Program_weight_tensor_parameter_410: class Program_weight_tensor_parameter_411: name = "parameter_411" + original_name = "conv2d_20.w_0" shape = [768, 192, 1, 1] dtype = "float32" min_val = float("-0.262071") @@ -4528,6 +4940,7 @@ class Program_weight_tensor_parameter_411: class Program_weight_tensor_parameter_412: name = "parameter_412" + original_name = "batch_norm_19.b_0" shape = [192] dtype = "float32" min_val = float("-2.05552") @@ -4539,6 +4952,7 @@ class Program_weight_tensor_parameter_412: class Program_weight_tensor_parameter_413: name = "parameter_413" + original_name = "batch_norm_19.w_0" shape = [192] dtype = "float32" min_val = float("-0.339596") @@ -4550,6 +4964,7 @@ class Program_weight_tensor_parameter_413: class Program_weight_tensor_parameter_414: name = "parameter_414" + original_name = "batch_norm_19.w_2" shape = [192] dtype = "float32" min_val = float("1.78489e-05") @@ -4561,6 +4976,7 @@ class Program_weight_tensor_parameter_414: class Program_weight_tensor_parameter_415: name = "parameter_415" + original_name = "batch_norm_19.w_1" shape = [192] dtype = "float32" min_val = float("-0.785407") @@ -4572,6 +4988,7 @@ class Program_weight_tensor_parameter_415: class Program_weight_tensor_parameter_416: name = "parameter_416" + original_name = "conv2d_19.w_0" shape = [192, 1, 3, 3] dtype = "float32" min_val = float("-0.246862") @@ -4583,6 +5000,7 @@ class Program_weight_tensor_parameter_416: class Program_weight_tensor_parameter_417: name = "parameter_417" + original_name = "batch_norm_18.b_0" shape = [192] dtype = "float32" min_val = float("-1.1344") @@ -4594,6 +5012,7 @@ class Program_weight_tensor_parameter_417: class Program_weight_tensor_parameter_418: name = "parameter_418" + original_name = "batch_norm_18.w_0" shape = [192] dtype = "float32" min_val = float("-0.172301") @@ -4605,6 +5024,7 @@ class Program_weight_tensor_parameter_418: class Program_weight_tensor_parameter_419: name = "parameter_419" + original_name = "batch_norm_18.w_2" shape = [192] dtype = "float32" min_val = float("0.0111597") @@ -4616,6 +5036,7 @@ class Program_weight_tensor_parameter_419: class Program_weight_tensor_parameter_420: name = "parameter_420" + original_name = "batch_norm_18.w_1" shape = [192] dtype = "float32" min_val = float("-0.300227") @@ -4627,6 +5048,7 @@ class Program_weight_tensor_parameter_420: class Program_weight_tensor_parameter_421: name = "parameter_421" + original_name = "conv2d_18.w_0" shape = [192, 768, 1, 1] dtype = "float32" min_val = float("-0.193986") @@ -4638,6 +5060,7 @@ class Program_weight_tensor_parameter_421: class Program_weight_tensor_parameter_422: name = "parameter_422" + original_name = "batch_norm_17.b_0" shape = [768] dtype = "float32" min_val = float("-7.01598") @@ -4649,6 +5072,7 @@ class Program_weight_tensor_parameter_422: class Program_weight_tensor_parameter_423: name = "parameter_423" + original_name = "batch_norm_17.w_0" shape = [768] dtype = "float32" min_val = float("0.0197177") @@ -4660,6 +5084,7 @@ class Program_weight_tensor_parameter_423: class Program_weight_tensor_parameter_424: name = "parameter_424" + original_name = "batch_norm_17.w_2" shape = [768] dtype = "float32" min_val = float("2.07392e-42") @@ -4671,6 +5096,7 @@ class Program_weight_tensor_parameter_424: class Program_weight_tensor_parameter_425: name = "parameter_425" + original_name = "batch_norm_17.w_1" shape = [768] dtype = "float32" min_val = float("-0.292343") @@ -4682,6 +5108,7 @@ class Program_weight_tensor_parameter_425: class Program_weight_tensor_parameter_426: name = "parameter_426" + original_name = "conv2d_17.w_0" shape = [768, 1, 3, 3] dtype = "float32" min_val = float("-0.0972677") @@ -4693,6 +5120,7 @@ class Program_weight_tensor_parameter_426: class Program_weight_tensor_parameter_427: name = "parameter_427" + original_name = "batch_norm_16.b_0" shape = [768] dtype = "float32" min_val = float("-1.43359") @@ -4704,6 +5132,7 @@ class Program_weight_tensor_parameter_427: class Program_weight_tensor_parameter_428: name = "parameter_428" + original_name = "batch_norm_16.w_0" shape = [768] dtype = "float32" min_val = float("-0.00379258") @@ -4715,6 +5144,7 @@ class Program_weight_tensor_parameter_428: class Program_weight_tensor_parameter_429: name = "parameter_429" + original_name = "batch_norm_16.w_2" shape = [768] dtype = "float32" min_val = float("2.53215e-42") @@ -4726,6 +5156,7 @@ class Program_weight_tensor_parameter_429: class Program_weight_tensor_parameter_430: name = "parameter_430" + original_name = "batch_norm_16.w_1" shape = [768] dtype = "float32" min_val = float("-0.653836") @@ -4737,6 +5168,7 @@ class Program_weight_tensor_parameter_430: class Program_weight_tensor_parameter_431: name = "parameter_431" + original_name = "conv2d_16.w_0" shape = [768, 192, 1, 1] dtype = "float32" min_val = float("-0.259385") @@ -4748,6 +5180,7 @@ class Program_weight_tensor_parameter_431: class Program_weight_tensor_parameter_432: name = "parameter_432" + original_name = "batch_norm_15.b_0" shape = [192] dtype = "float32" min_val = float("-1.99208") @@ -4759,6 +5192,7 @@ class Program_weight_tensor_parameter_432: class Program_weight_tensor_parameter_433: name = "parameter_433" + original_name = "batch_norm_15.w_0" shape = [192] dtype = "float32" min_val = float("-0.257855") @@ -4770,6 +5204,7 @@ class Program_weight_tensor_parameter_433: class Program_weight_tensor_parameter_434: name = "parameter_434" + original_name = "batch_norm_15.w_2" shape = [192] dtype = "float32" min_val = float("6.10813e-06") @@ -4781,6 +5216,7 @@ class Program_weight_tensor_parameter_434: class Program_weight_tensor_parameter_435: name = "parameter_435" + original_name = "batch_norm_15.w_1" shape = [192] dtype = "float32" min_val = float("-0.417457") @@ -4792,6 +5228,7 @@ class Program_weight_tensor_parameter_435: class Program_weight_tensor_parameter_436: name = "parameter_436" + original_name = "conv2d_15.w_0" shape = [192, 1, 3, 3] dtype = "float32" min_val = float("-0.219741") @@ -4803,6 +5240,7 @@ class Program_weight_tensor_parameter_436: class Program_weight_tensor_parameter_437: name = "parameter_437" + original_name = "batch_norm_14.b_0" shape = [192] dtype = "float32" min_val = float("-5.73288") @@ -4814,6 +5252,7 @@ class Program_weight_tensor_parameter_437: class Program_weight_tensor_parameter_438: name = "parameter_438" + original_name = "batch_norm_14.w_0" shape = [192] dtype = "float32" min_val = float("0.0277693") @@ -4825,6 +5264,7 @@ class Program_weight_tensor_parameter_438: class Program_weight_tensor_parameter_439: name = "parameter_439" + original_name = "batch_norm_14.w_2" shape = [192] dtype = "float32" min_val = float("0.032407") @@ -4836,6 +5276,7 @@ class Program_weight_tensor_parameter_439: class Program_weight_tensor_parameter_440: name = "parameter_440" + original_name = "batch_norm_14.w_1" shape = [192] dtype = "float32" min_val = float("-2.23754") @@ -4847,6 +5288,7 @@ class Program_weight_tensor_parameter_440: class Program_weight_tensor_parameter_441: name = "parameter_441" + original_name = "conv2d_14.w_0" shape = [192, 384, 1, 1] dtype = "float32" min_val = float("-0.274936") @@ -4858,6 +5300,7 @@ class Program_weight_tensor_parameter_441: class Program_weight_tensor_parameter_442: name = "parameter_442" + original_name = "batch_norm_13.b_0" shape = [384] dtype = "float32" min_val = float("-2.1721") @@ -4869,6 +5312,7 @@ class Program_weight_tensor_parameter_442: class Program_weight_tensor_parameter_443: name = "parameter_443" + original_name = "batch_norm_13.w_0" shape = [384] dtype = "float32" min_val = float("0.403777") @@ -4880,6 +5324,7 @@ class Program_weight_tensor_parameter_443: class Program_weight_tensor_parameter_444: name = "parameter_444" + original_name = "batch_norm_13.w_2" shape = [384] dtype = "float32" min_val = float("1.93558e-05") @@ -4891,6 +5336,7 @@ class Program_weight_tensor_parameter_444: class Program_weight_tensor_parameter_445: name = "parameter_445" + original_name = "batch_norm_13.w_1" shape = [384] dtype = "float32" min_val = float("-0.499382") @@ -4902,6 +5348,7 @@ class Program_weight_tensor_parameter_445: class Program_weight_tensor_parameter_446: name = "parameter_446" + original_name = "conv2d_13.w_0" shape = [384, 1, 5, 5] dtype = "float32" min_val = float("-0.0706211") @@ -4913,6 +5360,7 @@ class Program_weight_tensor_parameter_446: class Program_weight_tensor_parameter_447: name = "parameter_447" + original_name = "batch_norm_12.b_0" shape = [384] dtype = "float32" min_val = float("-1.40649") @@ -4924,6 +5372,7 @@ class Program_weight_tensor_parameter_447: class Program_weight_tensor_parameter_448: name = "parameter_448" + original_name = "batch_norm_12.w_0" shape = [384] dtype = "float32" min_val = float("0.317081") @@ -4935,6 +5384,7 @@ class Program_weight_tensor_parameter_448: class Program_weight_tensor_parameter_449: name = "parameter_449" + original_name = "batch_norm_12.w_2" shape = [384] dtype = "float32" min_val = float("0.031582") @@ -4946,6 +5396,7 @@ class Program_weight_tensor_parameter_449: class Program_weight_tensor_parameter_450: name = "parameter_450" + original_name = "batch_norm_12.w_1" shape = [384] dtype = "float32" min_val = float("-0.350639") @@ -4957,6 +5408,7 @@ class Program_weight_tensor_parameter_450: class Program_weight_tensor_parameter_451: name = "parameter_451" + original_name = "conv2d_12.w_0" shape = [384, 96, 1, 1] dtype = "float32" min_val = float("-0.382398") @@ -4968,6 +5420,7 @@ class Program_weight_tensor_parameter_451: class Program_weight_tensor_parameter_452: name = "parameter_452" + original_name = "batch_norm_11.b_0" shape = [96] dtype = "float32" min_val = float("-1.79") @@ -4979,6 +5432,7 @@ class Program_weight_tensor_parameter_452: class Program_weight_tensor_parameter_453: name = "parameter_453" + original_name = "batch_norm_11.w_0" shape = [96] dtype = "float32" min_val = float("0.405604") @@ -4990,6 +5444,7 @@ class Program_weight_tensor_parameter_453: class Program_weight_tensor_parameter_454: name = "parameter_454" + original_name = "batch_norm_11.w_2" shape = [96] dtype = "float32" min_val = float("0.00111677") @@ -5001,6 +5456,7 @@ class Program_weight_tensor_parameter_454: class Program_weight_tensor_parameter_455: name = "parameter_455" + original_name = "batch_norm_11.w_1" shape = [96] dtype = "float32" min_val = float("-0.623896") @@ -5012,6 +5468,7 @@ class Program_weight_tensor_parameter_455: class Program_weight_tensor_parameter_456: name = "parameter_456" + original_name = "conv2d_11.w_0" shape = [96, 1, 3, 3] dtype = "float32" min_val = float("-0.135227") @@ -5023,6 +5480,7 @@ class Program_weight_tensor_parameter_456: class Program_weight_tensor_parameter_457: name = "parameter_457" + original_name = "batch_norm_10.b_0" shape = [96] dtype = "float32" min_val = float("-1.76525") @@ -5034,6 +5492,7 @@ class Program_weight_tensor_parameter_457: class Program_weight_tensor_parameter_458: name = "parameter_458" + original_name = "batch_norm_10.w_0" shape = [96] dtype = "float32" min_val = float("-0.651984") @@ -5045,6 +5504,7 @@ class Program_weight_tensor_parameter_458: class Program_weight_tensor_parameter_459: name = "parameter_459" + original_name = "batch_norm_10.w_2" shape = [96] dtype = "float32" min_val = float("0.0266692") @@ -5056,6 +5516,7 @@ class Program_weight_tensor_parameter_459: class Program_weight_tensor_parameter_460: name = "parameter_460" + original_name = "batch_norm_10.w_1" shape = [96] dtype = "float32" min_val = float("-1.20065") @@ -5067,6 +5528,7 @@ class Program_weight_tensor_parameter_460: class Program_weight_tensor_parameter_461: name = "parameter_461" + original_name = "conv2d_10.w_0" shape = [96, 384, 1, 1] dtype = "float32" min_val = float("-0.149334") @@ -5078,6 +5540,7 @@ class Program_weight_tensor_parameter_461: class Program_weight_tensor_parameter_462: name = "parameter_462" + original_name = "batch_norm_9.b_0" shape = [384] dtype = "float32" min_val = float("-5.8773") @@ -5089,6 +5552,7 @@ class Program_weight_tensor_parameter_462: class Program_weight_tensor_parameter_463: name = "parameter_463" + original_name = "batch_norm_9.w_0" shape = [384] dtype = "float32" min_val = float("0.341224") @@ -5100,6 +5564,7 @@ class Program_weight_tensor_parameter_463: class Program_weight_tensor_parameter_464: name = "parameter_464" + original_name = "batch_norm_9.w_2" shape = [384] dtype = "float32" min_val = float("2.57559e-42") @@ -5111,6 +5576,7 @@ class Program_weight_tensor_parameter_464: class Program_weight_tensor_parameter_465: name = "parameter_465" + original_name = "batch_norm_9.w_1" shape = [384] dtype = "float32" min_val = float("-0.21283") @@ -5122,6 +5588,7 @@ class Program_weight_tensor_parameter_465: class Program_weight_tensor_parameter_466: name = "parameter_466" + original_name = "conv2d_9.w_0" shape = [384, 1, 3, 3] dtype = "float32" min_val = float("-0.101289") @@ -5133,6 +5600,7 @@ class Program_weight_tensor_parameter_466: class Program_weight_tensor_parameter_467: name = "parameter_467" + original_name = "batch_norm_8.b_0" shape = [384] dtype = "float32" min_val = float("-1.62755") @@ -5144,6 +5612,7 @@ class Program_weight_tensor_parameter_467: class Program_weight_tensor_parameter_468: name = "parameter_468" + original_name = "batch_norm_8.w_0" shape = [384] dtype = "float32" min_val = float("0.211333") @@ -5155,6 +5624,7 @@ class Program_weight_tensor_parameter_468: class Program_weight_tensor_parameter_469: name = "parameter_469" + original_name = "batch_norm_8.w_2" shape = [384] dtype = "float32" min_val = float("2.73253e-42") @@ -5166,6 +5636,7 @@ class Program_weight_tensor_parameter_469: class Program_weight_tensor_parameter_470: name = "parameter_470" + original_name = "batch_norm_8.w_1" shape = [384] dtype = "float32" min_val = float("-0.327269") @@ -5177,6 +5648,7 @@ class Program_weight_tensor_parameter_470: class Program_weight_tensor_parameter_471: name = "parameter_471" + original_name = "conv2d_8.w_0" shape = [384, 96, 1, 1] dtype = "float32" min_val = float("-0.180699") @@ -5188,6 +5660,7 @@ class Program_weight_tensor_parameter_471: class Program_weight_tensor_parameter_472: name = "parameter_472" + original_name = "batch_norm_7.b_0" shape = [96] dtype = "float32" min_val = float("-2.10623") @@ -5199,6 +5672,7 @@ class Program_weight_tensor_parameter_472: class Program_weight_tensor_parameter_473: name = "parameter_473" + original_name = "batch_norm_7.w_0" shape = [96] dtype = "float32" min_val = float("-0.115533") @@ -5210,6 +5684,7 @@ class Program_weight_tensor_parameter_473: class Program_weight_tensor_parameter_474: name = "parameter_474" + original_name = "batch_norm_7.w_2" shape = [96] dtype = "float32" min_val = float("7.67166e-05") @@ -5221,6 +5696,7 @@ class Program_weight_tensor_parameter_474: class Program_weight_tensor_parameter_475: name = "parameter_475" + original_name = "batch_norm_7.w_1" shape = [96] dtype = "float32" min_val = float("-0.314452") @@ -5232,6 +5708,7 @@ class Program_weight_tensor_parameter_475: class Program_weight_tensor_parameter_476: name = "parameter_476" + original_name = "conv2d_7.w_0" shape = [96, 1, 3, 3] dtype = "float32" min_val = float("-0.176988") @@ -5243,6 +5720,7 @@ class Program_weight_tensor_parameter_476: class Program_weight_tensor_parameter_477: name = "parameter_477" + original_name = "batch_norm_6.b_0" shape = [96] dtype = "float32" min_val = float("-3.05917") @@ -5254,6 +5732,7 @@ class Program_weight_tensor_parameter_477: class Program_weight_tensor_parameter_478: name = "parameter_478" + original_name = "batch_norm_6.w_0" shape = [96] dtype = "float32" min_val = float("0.147973") @@ -5265,6 +5744,7 @@ class Program_weight_tensor_parameter_478: class Program_weight_tensor_parameter_479: name = "parameter_479" + original_name = "batch_norm_6.w_2" shape = [96] dtype = "float32" min_val = float("0.0108142") @@ -5276,6 +5756,7 @@ class Program_weight_tensor_parameter_479: class Program_weight_tensor_parameter_480: name = "parameter_480" + original_name = "batch_norm_6.w_1" shape = [96] dtype = "float32" min_val = float("-0.935357") @@ -5287,6 +5768,7 @@ class Program_weight_tensor_parameter_480: class Program_weight_tensor_parameter_481: name = "parameter_481" + original_name = "conv2d_6.w_0" shape = [96, 192, 1, 1] dtype = "float32" min_val = float("-0.244285") @@ -5298,6 +5780,7 @@ class Program_weight_tensor_parameter_481: class Program_weight_tensor_parameter_482: name = "parameter_482" + original_name = "batch_norm_5.b_0" shape = [192] dtype = "float32" min_val = float("-1.92164") @@ -5309,6 +5792,7 @@ class Program_weight_tensor_parameter_482: class Program_weight_tensor_parameter_483: name = "parameter_483" + original_name = "batch_norm_5.w_0" shape = [192] dtype = "float32" min_val = float("0.294063") @@ -5320,6 +5804,7 @@ class Program_weight_tensor_parameter_483: class Program_weight_tensor_parameter_484: name = "parameter_484" + original_name = "batch_norm_5.w_2" shape = [192] dtype = "float32" min_val = float("0.000212842") @@ -5331,6 +5816,7 @@ class Program_weight_tensor_parameter_484: class Program_weight_tensor_parameter_485: name = "parameter_485" + original_name = "batch_norm_5.w_1" shape = [192] dtype = "float32" min_val = float("-0.414551") @@ -5342,6 +5828,7 @@ class Program_weight_tensor_parameter_485: class Program_weight_tensor_parameter_486: name = "parameter_486" + original_name = "conv2d_5.w_0" shape = [192, 1, 5, 5] dtype = "float32" min_val = float("-0.0929958") @@ -5353,6 +5840,7 @@ class Program_weight_tensor_parameter_486: class Program_weight_tensor_parameter_487: name = "parameter_487" + original_name = "batch_norm_4.b_0" shape = [192] dtype = "float32" min_val = float("-1.4598") @@ -5364,6 +5852,7 @@ class Program_weight_tensor_parameter_487: class Program_weight_tensor_parameter_488: name = "parameter_488" + original_name = "batch_norm_4.w_0" shape = [192] dtype = "float32" min_val = float("0.482704") @@ -5375,6 +5864,7 @@ class Program_weight_tensor_parameter_488: class Program_weight_tensor_parameter_489: name = "parameter_489" + original_name = "batch_norm_4.w_2" shape = [192] dtype = "float32" min_val = float("0.00296479") @@ -5386,6 +5876,7 @@ class Program_weight_tensor_parameter_489: class Program_weight_tensor_parameter_490: name = "parameter_490" + original_name = "batch_norm_4.w_1" shape = [192] dtype = "float32" min_val = float("-0.285905") @@ -5397,6 +5888,7 @@ class Program_weight_tensor_parameter_490: class Program_weight_tensor_parameter_491: name = "parameter_491" + original_name = "conv2d_4.w_0" shape = [192, 48, 1, 1] dtype = "float32" min_val = float("-0.457061") @@ -5408,6 +5900,7 @@ class Program_weight_tensor_parameter_491: class Program_weight_tensor_parameter_492: name = "parameter_492" + original_name = "batch_norm_3.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -5417,6 +5910,7 @@ class Program_weight_tensor_parameter_492: class Program_weight_tensor_parameter_493: name = "parameter_493" + original_name = "batch_norm_3.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -5426,6 +5920,7 @@ class Program_weight_tensor_parameter_493: class Program_weight_tensor_parameter_494: name = "parameter_494" + original_name = "batch_norm_3.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -5435,6 +5930,7 @@ class Program_weight_tensor_parameter_494: class Program_weight_tensor_parameter_495: name = "parameter_495" + original_name = "batch_norm_3.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -5444,6 +5940,7 @@ class Program_weight_tensor_parameter_495: class Program_weight_tensor_parameter_496: name = "parameter_496" + original_name = "conv2d_3.w_0" shape = [48, 1, 3, 3] dtype = "float32" min_val = float("-0.172858") @@ -5455,6 +5952,7 @@ class Program_weight_tensor_parameter_496: class Program_weight_tensor_parameter_497: name = "parameter_497" + original_name = "batch_norm_2.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -5464,6 +5962,7 @@ class Program_weight_tensor_parameter_497: class Program_weight_tensor_parameter_498: name = "parameter_498" + original_name = "batch_norm_2.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -5473,6 +5972,7 @@ class Program_weight_tensor_parameter_498: class Program_weight_tensor_parameter_499: name = "parameter_499" + original_name = "batch_norm_2.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -5482,6 +5982,7 @@ class Program_weight_tensor_parameter_499: class Program_weight_tensor_parameter_500: name = "parameter_500" + original_name = "batch_norm_2.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -5491,6 +5992,7 @@ class Program_weight_tensor_parameter_500: class Program_weight_tensor_parameter_501: name = "parameter_501" + original_name = "conv2d_2.w_0" shape = [48, 96, 1, 1] dtype = "float32" min_val = float("-0.14205") @@ -5502,6 +6004,7 @@ class Program_weight_tensor_parameter_501: class Program_weight_tensor_parameter_502: name = "parameter_502" + original_name = "batch_norm_1.b_0" shape = [96] dtype = "float32" min_val = float("-5.87499") @@ -5513,6 +6016,7 @@ class Program_weight_tensor_parameter_502: class Program_weight_tensor_parameter_503: name = "parameter_503" + original_name = "batch_norm_1.w_0" shape = [96] dtype = "float32" min_val = float("0.445028") @@ -5524,6 +6028,7 @@ class Program_weight_tensor_parameter_503: class Program_weight_tensor_parameter_504: name = "parameter_504" + original_name = "batch_norm_1.w_2" shape = [96] dtype = "float32" min_val = float("0.0159042") @@ -5535,6 +6040,7 @@ class Program_weight_tensor_parameter_504: class Program_weight_tensor_parameter_505: name = "parameter_505" + original_name = "batch_norm_1.w_1" shape = [96] dtype = "float32" min_val = float("-0.896093") @@ -5546,6 +6052,7 @@ class Program_weight_tensor_parameter_505: class Program_weight_tensor_parameter_506: name = "parameter_506" + original_name = "conv2d_1.w_0" shape = [96, 24, 3, 3] dtype = "float32" min_val = float("-0.235484") @@ -5557,6 +6064,7 @@ class Program_weight_tensor_parameter_506: class Program_weight_tensor_parameter_507: name = "parameter_507" + original_name = "batch_norm_0.b_0" shape = [24] dtype = "float32" min_val = float("0") @@ -5566,6 +6074,7 @@ class Program_weight_tensor_parameter_507: class Program_weight_tensor_parameter_508: name = "parameter_508" + original_name = "batch_norm_0.w_0" shape = [24] dtype = "float32" min_val = float("0") @@ -5575,6 +6084,7 @@ class Program_weight_tensor_parameter_508: class Program_weight_tensor_parameter_509: name = "parameter_509" + original_name = "batch_norm_0.w_2" shape = [24] dtype = "float32" min_val = float("0") @@ -5584,6 +6094,7 @@ class Program_weight_tensor_parameter_509: class Program_weight_tensor_parameter_510: name = "parameter_510" + original_name = "batch_norm_0.w_1" shape = [24] dtype = "float32" min_val = float("0") @@ -5593,6 +6104,7 @@ class Program_weight_tensor_parameter_510: class Program_weight_tensor_parameter_511: name = "parameter_511" + original_name = "conv2d_0.w_0" shape = [24, 3, 3, 3] dtype = "float32" min_val = float("-0.0669035") diff --git a/paddle_samples/PaddleX/MobileNetV4_conv_medium/subgraph_2/input_meta.py b/paddle_samples/PaddleX/MobileNetV4_conv_medium/subgraph_2/input_meta.py index a1aecd876..335545f59 100644 --- a/paddle_samples/PaddleX/MobileNetV4_conv_medium/subgraph_2/input_meta.py +++ b/paddle_samples/PaddleX/MobileNetV4_conv_medium/subgraph_2/input_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_data_0: name = "data_0" + original_name = "var_0" shape = [128, 3, 256, 256] dtype = "float32" min_val = float("-5.14588") diff --git a/paddle_samples/PaddleX/MobileNetV4_conv_medium/subgraph_2/weight_meta.py b/paddle_samples/PaddleX/MobileNetV4_conv_medium/subgraph_2/weight_meta.py index 636621da4..b65cbadfa 100644 --- a/paddle_samples/PaddleX/MobileNetV4_conv_medium/subgraph_2/weight_meta.py +++ b/paddle_samples/PaddleX/MobileNetV4_conv_medium/subgraph_2/weight_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_parameter_0: name = "parameter_0" + original_name = "linear_0.b_0" shape = [102] dtype = "float32" data = None @@ -7,6 +8,7 @@ class Program_weight_tensor_parameter_0: class Program_weight_tensor_parameter_1: name = "parameter_1" + original_name = "linear_0.w_0" shape = [1280, 102] dtype = "float32" min_val = float("-0.065887") @@ -18,6 +20,7 @@ class Program_weight_tensor_parameter_1: class Program_weight_tensor_parameter_2: name = "parameter_2" + original_name = "batch_norm_76.b_0" shape = [1280] dtype = "float32" min_val = float("-2.39253") @@ -29,6 +32,7 @@ class Program_weight_tensor_parameter_2: class Program_weight_tensor_parameter_3: name = "parameter_3" + original_name = "batch_norm_76.w_0" shape = [1280] dtype = "float32" min_val = float("0.816504") @@ -40,6 +44,7 @@ class Program_weight_tensor_parameter_3: class Program_weight_tensor_parameter_4: name = "parameter_4" + original_name = "batch_norm_76.w_2" shape = [1280] dtype = "float32" min_val = float("0.000847673") @@ -51,6 +56,7 @@ class Program_weight_tensor_parameter_4: class Program_weight_tensor_parameter_5: name = "parameter_5" + original_name = "batch_norm_76.w_1" shape = [1280] dtype = "float32" min_val = float("-0.0806118") @@ -62,6 +68,7 @@ class Program_weight_tensor_parameter_5: class Program_weight_tensor_parameter_6: name = "parameter_6" + original_name = "conv2d_76.w_0" shape = [1280, 960, 1, 1] dtype = "float32" min_val = float("-0.202999") @@ -73,6 +80,7 @@ class Program_weight_tensor_parameter_6: class Program_weight_tensor_parameter_7: name = "parameter_7" + original_name = "batch_norm_75.b_0" shape = [960] dtype = "float32" min_val = float("-4.12888") @@ -84,6 +92,7 @@ class Program_weight_tensor_parameter_7: class Program_weight_tensor_parameter_8: name = "parameter_8" + original_name = "batch_norm_75.w_0" shape = [960] dtype = "float32" min_val = float("0.304459") @@ -95,6 +104,7 @@ class Program_weight_tensor_parameter_8: class Program_weight_tensor_parameter_9: name = "parameter_9" + original_name = "batch_norm_75.w_2" shape = [960] dtype = "float32" min_val = float("2.83472") @@ -106,6 +116,7 @@ class Program_weight_tensor_parameter_9: class Program_weight_tensor_parameter_10: name = "parameter_10" + original_name = "batch_norm_75.w_1" shape = [960] dtype = "float32" min_val = float("-4.51892") @@ -117,6 +128,7 @@ class Program_weight_tensor_parameter_10: class Program_weight_tensor_parameter_11: name = "parameter_11" + original_name = "conv2d_75.w_0" shape = [960, 256, 1, 1] dtype = "float32" min_val = float("-0.410201") @@ -128,6 +140,7 @@ class Program_weight_tensor_parameter_11: class Program_weight_tensor_parameter_12: name = "parameter_12" + original_name = "batch_norm_74.b_0" shape = [256] dtype = "float32" min_val = float("-2.25962") @@ -139,6 +152,7 @@ class Program_weight_tensor_parameter_12: class Program_weight_tensor_parameter_13: name = "parameter_13" + original_name = "batch_norm_74.w_0" shape = [256] dtype = "float32" min_val = float("0.541") @@ -150,6 +164,7 @@ class Program_weight_tensor_parameter_13: class Program_weight_tensor_parameter_14: name = "parameter_14" + original_name = "batch_norm_74.w_2" shape = [256] dtype = "float32" min_val = float("0.0020107") @@ -161,6 +176,7 @@ class Program_weight_tensor_parameter_14: class Program_weight_tensor_parameter_15: name = "parameter_15" + original_name = "batch_norm_74.w_1" shape = [256] dtype = "float32" min_val = float("-0.0800153") @@ -172,6 +188,7 @@ class Program_weight_tensor_parameter_15: class Program_weight_tensor_parameter_16: name = "parameter_16" + original_name = "conv2d_74.w_0" shape = [256, 512, 1, 1] dtype = "float32" min_val = float("-0.286493") @@ -183,6 +200,7 @@ class Program_weight_tensor_parameter_16: class Program_weight_tensor_parameter_17: name = "parameter_17" + original_name = "batch_norm_73.b_0" shape = [512] dtype = "float32" min_val = float("-7.81166") @@ -194,6 +212,7 @@ class Program_weight_tensor_parameter_17: class Program_weight_tensor_parameter_18: name = "parameter_18" + original_name = "batch_norm_73.w_0" shape = [512] dtype = "float32" min_val = float("0.22782") @@ -205,6 +224,7 @@ class Program_weight_tensor_parameter_18: class Program_weight_tensor_parameter_19: name = "parameter_19" + original_name = "batch_norm_73.w_2" shape = [512] dtype = "float32" min_val = float("6.39103e-19") @@ -216,6 +236,7 @@ class Program_weight_tensor_parameter_19: class Program_weight_tensor_parameter_20: name = "parameter_20" + original_name = "batch_norm_73.w_1" shape = [512] dtype = "float32" min_val = float("-1.47201") @@ -227,6 +248,7 @@ class Program_weight_tensor_parameter_20: class Program_weight_tensor_parameter_21: name = "parameter_21" + original_name = "conv2d_73.w_0" shape = [512, 256, 1, 1] dtype = "float32" min_val = float("-0.274376") @@ -238,6 +260,7 @@ class Program_weight_tensor_parameter_21: class Program_weight_tensor_parameter_22: name = "parameter_22" + original_name = "batch_norm_72.b_0" shape = [256] dtype = "float32" min_val = float("-3.28114") @@ -249,6 +272,7 @@ class Program_weight_tensor_parameter_22: class Program_weight_tensor_parameter_23: name = "parameter_23" + original_name = "batch_norm_72.w_0" shape = [256] dtype = "float32" min_val = float("0.759356") @@ -260,6 +284,7 @@ class Program_weight_tensor_parameter_23: class Program_weight_tensor_parameter_24: name = "parameter_24" + original_name = "batch_norm_72.w_2" shape = [256] dtype = "float32" min_val = float("0.201307") @@ -271,6 +296,7 @@ class Program_weight_tensor_parameter_24: class Program_weight_tensor_parameter_25: name = "parameter_25" + original_name = "batch_norm_72.w_1" shape = [256] dtype = "float32" min_val = float("-4.84508") @@ -282,6 +308,7 @@ class Program_weight_tensor_parameter_25: class Program_weight_tensor_parameter_26: name = "parameter_26" + original_name = "conv2d_72.w_0" shape = [256, 1, 5, 5] dtype = "float32" min_val = float("-0.335452") @@ -293,6 +320,7 @@ class Program_weight_tensor_parameter_26: class Program_weight_tensor_parameter_27: name = "parameter_27" + original_name = "batch_norm_71.b_0" shape = [256] dtype = "float32" min_val = float("-1.14467") @@ -304,6 +332,7 @@ class Program_weight_tensor_parameter_27: class Program_weight_tensor_parameter_28: name = "parameter_28" + original_name = "batch_norm_71.w_0" shape = [256] dtype = "float32" min_val = float("0.670602") @@ -315,6 +344,7 @@ class Program_weight_tensor_parameter_28: class Program_weight_tensor_parameter_29: name = "parameter_29" + original_name = "batch_norm_71.w_2" shape = [256] dtype = "float32" min_val = float("0.000170708") @@ -326,6 +356,7 @@ class Program_weight_tensor_parameter_29: class Program_weight_tensor_parameter_30: name = "parameter_30" + original_name = "batch_norm_71.w_1" shape = [256] dtype = "float32" min_val = float("-0.0121758") @@ -337,6 +368,7 @@ class Program_weight_tensor_parameter_30: class Program_weight_tensor_parameter_31: name = "parameter_31" + original_name = "conv2d_71.w_0" shape = [256, 1024, 1, 1] dtype = "float32" min_val = float("-0.327775") @@ -348,6 +380,7 @@ class Program_weight_tensor_parameter_31: class Program_weight_tensor_parameter_32: name = "parameter_32" + original_name = "batch_norm_70.b_0" shape = [1024] dtype = "float32" min_val = float("-3.96506") @@ -359,6 +392,7 @@ class Program_weight_tensor_parameter_32: class Program_weight_tensor_parameter_33: name = "parameter_33" + original_name = "batch_norm_70.w_0" shape = [1024] dtype = "float32" min_val = float("0.0285159") @@ -370,6 +404,7 @@ class Program_weight_tensor_parameter_33: class Program_weight_tensor_parameter_34: name = "parameter_34" + original_name = "batch_norm_70.w_2" shape = [1024] dtype = "float32" min_val = float("3.50885e-42") @@ -381,6 +416,7 @@ class Program_weight_tensor_parameter_34: class Program_weight_tensor_parameter_35: name = "parameter_35" + original_name = "batch_norm_70.w_1" shape = [1024] dtype = "float32" min_val = float("-4.76175") @@ -392,6 +428,7 @@ class Program_weight_tensor_parameter_35: class Program_weight_tensor_parameter_36: name = "parameter_36" + original_name = "conv2d_70.w_0" shape = [1024, 256, 1, 1] dtype = "float32" min_val = float("-0.339514") @@ -403,6 +440,7 @@ class Program_weight_tensor_parameter_36: class Program_weight_tensor_parameter_37: name = "parameter_37" + original_name = "batch_norm_69.b_0" shape = [256] dtype = "float32" min_val = float("-0.816218") @@ -414,6 +452,7 @@ class Program_weight_tensor_parameter_37: class Program_weight_tensor_parameter_38: name = "parameter_38" + original_name = "batch_norm_69.w_0" shape = [256] dtype = "float32" min_val = float("0.598678") @@ -425,6 +464,7 @@ class Program_weight_tensor_parameter_38: class Program_weight_tensor_parameter_39: name = "parameter_39" + original_name = "batch_norm_69.w_2" shape = [256] dtype = "float32" min_val = float("0.00033605") @@ -436,6 +476,7 @@ class Program_weight_tensor_parameter_39: class Program_weight_tensor_parameter_40: name = "parameter_40" + original_name = "batch_norm_69.w_1" shape = [256] dtype = "float32" min_val = float("-0.0257608") @@ -447,6 +488,7 @@ class Program_weight_tensor_parameter_40: class Program_weight_tensor_parameter_41: name = "parameter_41" + original_name = "conv2d_69.w_0" shape = [256, 1024, 1, 1] dtype = "float32" min_val = float("-0.186477") @@ -458,6 +500,7 @@ class Program_weight_tensor_parameter_41: class Program_weight_tensor_parameter_42: name = "parameter_42" + original_name = "batch_norm_68.b_0" shape = [1024] dtype = "float32" min_val = float("-3.95164") @@ -469,6 +512,7 @@ class Program_weight_tensor_parameter_42: class Program_weight_tensor_parameter_43: name = "parameter_43" + original_name = "batch_norm_68.w_0" shape = [1024] dtype = "float32" min_val = float("0.0696335") @@ -480,6 +524,7 @@ class Program_weight_tensor_parameter_43: class Program_weight_tensor_parameter_44: name = "parameter_44" + original_name = "batch_norm_68.w_2" shape = [1024] dtype = "float32" min_val = float("1.21243e-27") @@ -491,6 +536,7 @@ class Program_weight_tensor_parameter_44: class Program_weight_tensor_parameter_45: name = "parameter_45" + original_name = "batch_norm_68.w_1" shape = [1024] dtype = "float32" min_val = float("-4.34205") @@ -502,6 +548,7 @@ class Program_weight_tensor_parameter_45: class Program_weight_tensor_parameter_46: name = "parameter_46" + original_name = "conv2d_68.w_0" shape = [1024, 256, 1, 1] dtype = "float32" min_val = float("-0.305797") @@ -513,6 +560,7 @@ class Program_weight_tensor_parameter_46: class Program_weight_tensor_parameter_47: name = "parameter_47" + original_name = "batch_norm_67.b_0" shape = [256] dtype = "float32" min_val = float("-1.98968") @@ -524,6 +572,7 @@ class Program_weight_tensor_parameter_47: class Program_weight_tensor_parameter_48: name = "parameter_48" + original_name = "batch_norm_67.w_0" shape = [256] dtype = "float32" min_val = float("0.482527") @@ -535,6 +584,7 @@ class Program_weight_tensor_parameter_48: class Program_weight_tensor_parameter_49: name = "parameter_49" + original_name = "batch_norm_67.w_2" shape = [256] dtype = "float32" min_val = float("0.163203") @@ -546,6 +596,7 @@ class Program_weight_tensor_parameter_49: class Program_weight_tensor_parameter_50: name = "parameter_50" + original_name = "batch_norm_67.w_1" shape = [256] dtype = "float32" min_val = float("-1.83639") @@ -557,6 +608,7 @@ class Program_weight_tensor_parameter_50: class Program_weight_tensor_parameter_51: name = "parameter_51" + original_name = "conv2d_67.w_0" shape = [256, 1024, 1, 1] dtype = "float32" min_val = float("-0.293587") @@ -568,6 +620,7 @@ class Program_weight_tensor_parameter_51: class Program_weight_tensor_parameter_52: name = "parameter_52" + original_name = "batch_norm_66.b_0" shape = [1024] dtype = "float32" min_val = float("-7.28465") @@ -579,6 +632,7 @@ class Program_weight_tensor_parameter_52: class Program_weight_tensor_parameter_53: name = "parameter_53" + original_name = "batch_norm_66.w_0" shape = [1024] dtype = "float32" min_val = float("0.0108447") @@ -590,6 +644,7 @@ class Program_weight_tensor_parameter_53: class Program_weight_tensor_parameter_54: name = "parameter_54" + original_name = "batch_norm_66.w_2" shape = [1024] dtype = "float32" min_val = float("2.34017e-42") @@ -601,6 +656,7 @@ class Program_weight_tensor_parameter_54: class Program_weight_tensor_parameter_55: name = "parameter_55" + original_name = "batch_norm_66.w_1" shape = [1024] dtype = "float32" min_val = float("-0.734679") @@ -612,6 +668,7 @@ class Program_weight_tensor_parameter_55: class Program_weight_tensor_parameter_56: name = "parameter_56" + original_name = "conv2d_66.w_0" shape = [1024, 1, 5, 5] dtype = "float32" min_val = float("-0.325517") @@ -623,6 +680,7 @@ class Program_weight_tensor_parameter_56: class Program_weight_tensor_parameter_57: name = "parameter_57" + original_name = "batch_norm_65.b_0" shape = [1024] dtype = "float32" min_val = float("-2.97498") @@ -634,6 +692,7 @@ class Program_weight_tensor_parameter_57: class Program_weight_tensor_parameter_58: name = "parameter_58" + original_name = "batch_norm_65.w_0" shape = [1024] dtype = "float32" min_val = float("0.0512698") @@ -645,6 +704,7 @@ class Program_weight_tensor_parameter_58: class Program_weight_tensor_parameter_59: name = "parameter_59" + original_name = "batch_norm_65.w_2" shape = [1024] dtype = "float32" min_val = float("3.39955e-42") @@ -656,6 +716,7 @@ class Program_weight_tensor_parameter_59: class Program_weight_tensor_parameter_60: name = "parameter_60" + original_name = "batch_norm_65.w_1" shape = [1024] dtype = "float32" min_val = float("-1.33231") @@ -667,6 +728,7 @@ class Program_weight_tensor_parameter_60: class Program_weight_tensor_parameter_61: name = "parameter_61" + original_name = "conv2d_65.w_0" shape = [1024, 256, 1, 1] dtype = "float32" min_val = float("-0.333331") @@ -678,6 +740,7 @@ class Program_weight_tensor_parameter_61: class Program_weight_tensor_parameter_62: name = "parameter_62" + original_name = "batch_norm_64.b_0" shape = [256] dtype = "float32" min_val = float("-2.5208") @@ -689,6 +752,7 @@ class Program_weight_tensor_parameter_62: class Program_weight_tensor_parameter_63: name = "parameter_63" + original_name = "batch_norm_64.w_0" shape = [256] dtype = "float32" min_val = float("0.875024") @@ -700,6 +764,7 @@ class Program_weight_tensor_parameter_63: class Program_weight_tensor_parameter_64: name = "parameter_64" + original_name = "batch_norm_64.w_2" shape = [256] dtype = "float32" min_val = float("0.00986637") @@ -711,6 +776,7 @@ class Program_weight_tensor_parameter_64: class Program_weight_tensor_parameter_65: name = "parameter_65" + original_name = "batch_norm_64.w_1" shape = [256] dtype = "float32" min_val = float("-5.39897") @@ -722,6 +788,7 @@ class Program_weight_tensor_parameter_65: class Program_weight_tensor_parameter_66: name = "parameter_66" + original_name = "conv2d_64.w_0" shape = [256, 1, 5, 5] dtype = "float32" min_val = float("-0.354987") @@ -733,6 +800,7 @@ class Program_weight_tensor_parameter_66: class Program_weight_tensor_parameter_67: name = "parameter_67" + original_name = "batch_norm_63.b_0" shape = [256] dtype = "float32" min_val = float("-1.51872") @@ -744,6 +812,7 @@ class Program_weight_tensor_parameter_67: class Program_weight_tensor_parameter_68: name = "parameter_68" + original_name = "batch_norm_63.w_0" shape = [256] dtype = "float32" min_val = float("0.24764") @@ -755,6 +824,7 @@ class Program_weight_tensor_parameter_68: class Program_weight_tensor_parameter_69: name = "parameter_69" + original_name = "batch_norm_63.w_2" shape = [256] dtype = "float32" min_val = float("0.113296") @@ -766,6 +836,7 @@ class Program_weight_tensor_parameter_69: class Program_weight_tensor_parameter_70: name = "parameter_70" + original_name = "batch_norm_63.w_1" shape = [256] dtype = "float32" min_val = float("-2.26608") @@ -777,6 +848,7 @@ class Program_weight_tensor_parameter_70: class Program_weight_tensor_parameter_71: name = "parameter_71" + original_name = "conv2d_63.w_0" shape = [256, 512, 1, 1] dtype = "float32" min_val = float("-0.22915") @@ -788,6 +860,7 @@ class Program_weight_tensor_parameter_71: class Program_weight_tensor_parameter_72: name = "parameter_72" + original_name = "batch_norm_62.b_0" shape = [512] dtype = "float32" min_val = float("-6.24833") @@ -799,6 +872,7 @@ class Program_weight_tensor_parameter_72: class Program_weight_tensor_parameter_73: name = "parameter_73" + original_name = "batch_norm_62.w_0" shape = [512] dtype = "float32" min_val = float("0.121069") @@ -810,6 +884,7 @@ class Program_weight_tensor_parameter_73: class Program_weight_tensor_parameter_74: name = "parameter_74" + original_name = "batch_norm_62.w_2" shape = [512] dtype = "float32" min_val = float("2.51813e-42") @@ -821,6 +896,7 @@ class Program_weight_tensor_parameter_74: class Program_weight_tensor_parameter_75: name = "parameter_75" + original_name = "batch_norm_62.w_1" shape = [512] dtype = "float32" min_val = float("-0.507624") @@ -832,6 +908,7 @@ class Program_weight_tensor_parameter_75: class Program_weight_tensor_parameter_76: name = "parameter_76" + original_name = "conv2d_62.w_0" shape = [512, 1, 5, 5] dtype = "float32" min_val = float("-0.134429") @@ -843,6 +920,7 @@ class Program_weight_tensor_parameter_76: class Program_weight_tensor_parameter_77: name = "parameter_77" + original_name = "batch_norm_61.b_0" shape = [512] dtype = "float32" min_val = float("-2.01705") @@ -854,6 +932,7 @@ class Program_weight_tensor_parameter_77: class Program_weight_tensor_parameter_78: name = "parameter_78" + original_name = "batch_norm_61.w_0" shape = [512] dtype = "float32" min_val = float("0.0642323") @@ -865,6 +944,7 @@ class Program_weight_tensor_parameter_78: class Program_weight_tensor_parameter_79: name = "parameter_79" + original_name = "batch_norm_61.w_2" shape = [512] dtype = "float32" min_val = float("3.50885e-42") @@ -876,6 +956,7 @@ class Program_weight_tensor_parameter_79: class Program_weight_tensor_parameter_80: name = "parameter_80" + original_name = "batch_norm_61.w_1" shape = [512] dtype = "float32" min_val = float("-1.39433") @@ -887,6 +968,7 @@ class Program_weight_tensor_parameter_80: class Program_weight_tensor_parameter_81: name = "parameter_81" + original_name = "conv2d_61.w_0" shape = [512, 256, 1, 1] dtype = "float32" min_val = float("-0.342027") @@ -898,6 +980,7 @@ class Program_weight_tensor_parameter_81: class Program_weight_tensor_parameter_82: name = "parameter_82" + original_name = "batch_norm_60.b_0" shape = [256] dtype = "float32" min_val = float("-2.34808") @@ -909,6 +992,7 @@ class Program_weight_tensor_parameter_82: class Program_weight_tensor_parameter_83: name = "parameter_83" + original_name = "batch_norm_60.w_0" shape = [256] dtype = "float32" min_val = float("0.680999") @@ -920,6 +1004,7 @@ class Program_weight_tensor_parameter_83: class Program_weight_tensor_parameter_84: name = "parameter_84" + original_name = "batch_norm_60.w_2" shape = [256] dtype = "float32" min_val = float("0.000286501") @@ -931,6 +1016,7 @@ class Program_weight_tensor_parameter_84: class Program_weight_tensor_parameter_85: name = "parameter_85" + original_name = "batch_norm_60.w_1" shape = [256] dtype = "float32" min_val = float("-1.51416") @@ -942,6 +1028,7 @@ class Program_weight_tensor_parameter_85: class Program_weight_tensor_parameter_86: name = "parameter_86" + original_name = "conv2d_60.w_0" shape = [256, 1, 3, 3] dtype = "float32" min_val = float("-0.2387") @@ -953,6 +1040,7 @@ class Program_weight_tensor_parameter_86: class Program_weight_tensor_parameter_87: name = "parameter_87" + original_name = "batch_norm_59.b_0" shape = [256] dtype = "float32" min_val = float("-1.07136") @@ -964,6 +1052,7 @@ class Program_weight_tensor_parameter_87: class Program_weight_tensor_parameter_88: name = "parameter_88" + original_name = "batch_norm_59.w_0" shape = [256] dtype = "float32" min_val = float("0.0872126") @@ -975,6 +1064,7 @@ class Program_weight_tensor_parameter_88: class Program_weight_tensor_parameter_89: name = "parameter_89" + original_name = "batch_norm_59.w_2" shape = [256] dtype = "float32" min_val = float("0.000819224") @@ -986,6 +1076,7 @@ class Program_weight_tensor_parameter_89: class Program_weight_tensor_parameter_90: name = "parameter_90" + original_name = "batch_norm_59.w_1" shape = [256] dtype = "float32" min_val = float("-0.026104") @@ -997,6 +1088,7 @@ class Program_weight_tensor_parameter_90: class Program_weight_tensor_parameter_91: name = "parameter_91" + original_name = "conv2d_59.w_0" shape = [256, 1024, 1, 1] dtype = "float32" min_val = float("-0.205036") @@ -1008,6 +1100,7 @@ class Program_weight_tensor_parameter_91: class Program_weight_tensor_parameter_92: name = "parameter_92" + original_name = "batch_norm_58.b_0" shape = [1024] dtype = "float32" min_val = float("-4.12463") @@ -1019,6 +1112,7 @@ class Program_weight_tensor_parameter_92: class Program_weight_tensor_parameter_93: name = "parameter_93" + original_name = "batch_norm_58.w_0" shape = [1024] dtype = "float32" min_val = float("0.113134") @@ -1030,6 +1124,7 @@ class Program_weight_tensor_parameter_93: class Program_weight_tensor_parameter_94: name = "parameter_94" + original_name = "batch_norm_58.w_2" shape = [1024] dtype = "float32" min_val = float("3.50885e-42") @@ -1041,6 +1136,7 @@ class Program_weight_tensor_parameter_94: class Program_weight_tensor_parameter_95: name = "parameter_95" + original_name = "batch_norm_58.w_1" shape = [1024] dtype = "float32" min_val = float("-1.19719") @@ -1052,6 +1148,7 @@ class Program_weight_tensor_parameter_95: class Program_weight_tensor_parameter_96: name = "parameter_96" + original_name = "conv2d_58.w_0" shape = [1024, 256, 1, 1] dtype = "float32" min_val = float("-0.154685") @@ -1063,6 +1160,7 @@ class Program_weight_tensor_parameter_96: class Program_weight_tensor_parameter_97: name = "parameter_97" + original_name = "batch_norm_57.b_0" shape = [256] dtype = "float32" min_val = float("-1.98714") @@ -1074,6 +1172,7 @@ class Program_weight_tensor_parameter_97: class Program_weight_tensor_parameter_98: name = "parameter_98" + original_name = "batch_norm_57.w_0" shape = [256] dtype = "float32" min_val = float("0.667212") @@ -1085,6 +1184,7 @@ class Program_weight_tensor_parameter_98: class Program_weight_tensor_parameter_99: name = "parameter_99" + original_name = "batch_norm_57.w_2" shape = [256] dtype = "float32" min_val = float("0.000229811") @@ -1096,6 +1196,7 @@ class Program_weight_tensor_parameter_99: class Program_weight_tensor_parameter_100: name = "parameter_100" + original_name = "batch_norm_57.w_1" shape = [256] dtype = "float32" min_val = float("-0.967291") @@ -1107,6 +1208,7 @@ class Program_weight_tensor_parameter_100: class Program_weight_tensor_parameter_101: name = "parameter_101" + original_name = "conv2d_57.w_0" shape = [256, 1, 3, 3] dtype = "float32" min_val = float("-0.207575") @@ -1118,6 +1220,7 @@ class Program_weight_tensor_parameter_101: class Program_weight_tensor_parameter_102: name = "parameter_102" + original_name = "batch_norm_56.b_0" shape = [256] dtype = "float32" min_val = float("-0.859875") @@ -1129,6 +1232,7 @@ class Program_weight_tensor_parameter_102: class Program_weight_tensor_parameter_103: name = "parameter_103" + original_name = "batch_norm_56.w_0" shape = [256] dtype = "float32" min_val = float("0.0753627") @@ -1140,6 +1244,7 @@ class Program_weight_tensor_parameter_103: class Program_weight_tensor_parameter_104: name = "parameter_104" + original_name = "batch_norm_56.w_2" shape = [256] dtype = "float32" min_val = float("0.00444596") @@ -1151,6 +1256,7 @@ class Program_weight_tensor_parameter_104: class Program_weight_tensor_parameter_105: name = "parameter_105" + original_name = "batch_norm_56.w_1" shape = [256] dtype = "float32" min_val = float("-0.102455") @@ -1162,6 +1268,7 @@ class Program_weight_tensor_parameter_105: class Program_weight_tensor_parameter_106: name = "parameter_106" + original_name = "conv2d_56.w_0" shape = [256, 1024, 1, 1] dtype = "float32" min_val = float("-0.174062") @@ -1173,6 +1280,7 @@ class Program_weight_tensor_parameter_106: class Program_weight_tensor_parameter_107: name = "parameter_107" + original_name = "batch_norm_55.b_0" shape = [1024] dtype = "float32" min_val = float("-5.69847") @@ -1184,6 +1292,7 @@ class Program_weight_tensor_parameter_107: class Program_weight_tensor_parameter_108: name = "parameter_108" + original_name = "batch_norm_55.w_0" shape = [1024] dtype = "float32" min_val = float("0.193728") @@ -1195,6 +1304,7 @@ class Program_weight_tensor_parameter_108: class Program_weight_tensor_parameter_109: name = "parameter_109" + original_name = "batch_norm_55.w_2" shape = [1024] dtype = "float32" min_val = float("3.50885e-42") @@ -1206,6 +1316,7 @@ class Program_weight_tensor_parameter_109: class Program_weight_tensor_parameter_110: name = "parameter_110" + original_name = "batch_norm_55.w_1" shape = [1024] dtype = "float32" min_val = float("-3.2393") @@ -1217,6 +1328,7 @@ class Program_weight_tensor_parameter_110: class Program_weight_tensor_parameter_111: name = "parameter_111" + original_name = "conv2d_55.w_0" shape = [1024, 256, 1, 1] dtype = "float32" min_val = float("-0.215234") @@ -1228,6 +1340,7 @@ class Program_weight_tensor_parameter_111: class Program_weight_tensor_parameter_112: name = "parameter_112" + original_name = "batch_norm_54.b_0" shape = [256] dtype = "float32" min_val = float("-1.50762") @@ -1239,6 +1352,7 @@ class Program_weight_tensor_parameter_112: class Program_weight_tensor_parameter_113: name = "parameter_113" + original_name = "batch_norm_54.w_0" shape = [256] dtype = "float32" min_val = float("0.061722") @@ -1250,6 +1364,7 @@ class Program_weight_tensor_parameter_113: class Program_weight_tensor_parameter_114: name = "parameter_114" + original_name = "batch_norm_54.w_2" shape = [256] dtype = "float32" min_val = float("0.27692") @@ -1261,6 +1376,7 @@ class Program_weight_tensor_parameter_114: class Program_weight_tensor_parameter_115: name = "parameter_115" + original_name = "batch_norm_54.w_1" shape = [256] dtype = "float32" min_val = float("-5.01889") @@ -1272,6 +1388,7 @@ class Program_weight_tensor_parameter_115: class Program_weight_tensor_parameter_116: name = "parameter_116" + original_name = "conv2d_54.w_0" shape = [256, 1024, 1, 1] dtype = "float32" min_val = float("-0.217436") @@ -1283,6 +1400,7 @@ class Program_weight_tensor_parameter_116: class Program_weight_tensor_parameter_117: name = "parameter_117" + original_name = "batch_norm_53.b_0" shape = [1024] dtype = "float32" min_val = float("-6.91335") @@ -1294,6 +1412,7 @@ class Program_weight_tensor_parameter_117: class Program_weight_tensor_parameter_118: name = "parameter_118" + original_name = "batch_norm_53.w_0" shape = [1024] dtype = "float32" min_val = float("0.0338038") @@ -1305,6 +1424,7 @@ class Program_weight_tensor_parameter_118: class Program_weight_tensor_parameter_119: name = "parameter_119" + original_name = "batch_norm_53.w_2" shape = [1024] dtype = "float32" min_val = float("2.3766e-42") @@ -1316,6 +1436,7 @@ class Program_weight_tensor_parameter_119: class Program_weight_tensor_parameter_120: name = "parameter_120" + original_name = "batch_norm_53.w_1" shape = [1024] dtype = "float32" min_val = float("-0.675852") @@ -1327,6 +1448,7 @@ class Program_weight_tensor_parameter_120: class Program_weight_tensor_parameter_121: name = "parameter_121" + original_name = "conv2d_53.w_0" shape = [1024, 1, 5, 5] dtype = "float32" min_val = float("-0.136097") @@ -1338,6 +1460,7 @@ class Program_weight_tensor_parameter_121: class Program_weight_tensor_parameter_122: name = "parameter_122" + original_name = "batch_norm_52.b_0" shape = [1024] dtype = "float32" min_val = float("-1.93199") @@ -1349,6 +1472,7 @@ class Program_weight_tensor_parameter_122: class Program_weight_tensor_parameter_123: name = "parameter_123" + original_name = "batch_norm_52.w_0" shape = [1024] dtype = "float32" min_val = float("0.0668295") @@ -1360,6 +1484,7 @@ class Program_weight_tensor_parameter_123: class Program_weight_tensor_parameter_124: name = "parameter_124" + original_name = "batch_norm_52.w_2" shape = [1024] dtype = "float32" min_val = float("3.4542e-42") @@ -1371,6 +1496,7 @@ class Program_weight_tensor_parameter_124: class Program_weight_tensor_parameter_125: name = "parameter_125" + original_name = "batch_norm_52.w_1" shape = [1024] dtype = "float32" min_val = float("-1.42938") @@ -1382,6 +1508,7 @@ class Program_weight_tensor_parameter_125: class Program_weight_tensor_parameter_126: name = "parameter_126" + original_name = "conv2d_52.w_0" shape = [1024, 256, 1, 1] dtype = "float32" min_val = float("-0.222955") @@ -1393,6 +1520,7 @@ class Program_weight_tensor_parameter_126: class Program_weight_tensor_parameter_127: name = "parameter_127" + original_name = "batch_norm_51.b_0" shape = [256] dtype = "float32" min_val = float("-2.92007") @@ -1404,6 +1532,7 @@ class Program_weight_tensor_parameter_127: class Program_weight_tensor_parameter_128: name = "parameter_128" + original_name = "batch_norm_51.w_0" shape = [256] dtype = "float32" min_val = float("0.831741") @@ -1415,6 +1544,7 @@ class Program_weight_tensor_parameter_128: class Program_weight_tensor_parameter_129: name = "parameter_129" + original_name = "batch_norm_51.w_2" shape = [256] dtype = "float32" min_val = float("1.67467e-05") @@ -1426,6 +1556,7 @@ class Program_weight_tensor_parameter_129: class Program_weight_tensor_parameter_130: name = "parameter_130" + original_name = "batch_norm_51.w_1" shape = [256] dtype = "float32" min_val = float("-0.971199") @@ -1437,6 +1568,7 @@ class Program_weight_tensor_parameter_130: class Program_weight_tensor_parameter_131: name = "parameter_131" + original_name = "conv2d_51.w_0" shape = [256, 1, 3, 3] dtype = "float32" min_val = float("-0.235344") @@ -1448,6 +1580,7 @@ class Program_weight_tensor_parameter_131: class Program_weight_tensor_parameter_132: name = "parameter_132" + original_name = "batch_norm_50.b_0" shape = [256] dtype = "float32" min_val = float("-1.0719") @@ -1459,6 +1592,7 @@ class Program_weight_tensor_parameter_132: class Program_weight_tensor_parameter_133: name = "parameter_133" + original_name = "batch_norm_50.w_0" shape = [256] dtype = "float32" min_val = float("-0.00596973") @@ -1470,6 +1604,7 @@ class Program_weight_tensor_parameter_133: class Program_weight_tensor_parameter_134: name = "parameter_134" + original_name = "batch_norm_50.w_2" shape = [256] dtype = "float32" min_val = float("0.182694") @@ -1481,6 +1616,7 @@ class Program_weight_tensor_parameter_134: class Program_weight_tensor_parameter_135: name = "parameter_135" + original_name = "batch_norm_50.w_1" shape = [256] dtype = "float32" min_val = float("-1.78319") @@ -1492,6 +1628,7 @@ class Program_weight_tensor_parameter_135: class Program_weight_tensor_parameter_136: name = "parameter_136" + original_name = "conv2d_50.w_0" shape = [256, 1024, 1, 1] dtype = "float32" min_val = float("-0.254084") @@ -1503,6 +1640,7 @@ class Program_weight_tensor_parameter_136: class Program_weight_tensor_parameter_137: name = "parameter_137" + original_name = "batch_norm_49.b_0" shape = [1024] dtype = "float32" min_val = float("-10.3783") @@ -1514,6 +1652,7 @@ class Program_weight_tensor_parameter_137: class Program_weight_tensor_parameter_138: name = "parameter_138" + original_name = "batch_norm_49.w_0" shape = [1024] dtype = "float32" min_val = float("0.171525") @@ -1525,6 +1664,7 @@ class Program_weight_tensor_parameter_138: class Program_weight_tensor_parameter_139: name = "parameter_139" + original_name = "batch_norm_49.w_2" shape = [1024] dtype = "float32" min_val = float("2.42425e-42") @@ -1536,6 +1676,7 @@ class Program_weight_tensor_parameter_139: class Program_weight_tensor_parameter_140: name = "parameter_140" + original_name = "batch_norm_49.w_1" shape = [1024] dtype = "float32" min_val = float("-0.604826") @@ -1547,6 +1688,7 @@ class Program_weight_tensor_parameter_140: class Program_weight_tensor_parameter_141: name = "parameter_141" + original_name = "conv2d_49.w_0" shape = [1024, 1, 5, 5] dtype = "float32" min_val = float("-0.133567") @@ -1558,6 +1700,7 @@ class Program_weight_tensor_parameter_141: class Program_weight_tensor_parameter_142: name = "parameter_142" + original_name = "batch_norm_48.b_0" shape = [1024] dtype = "float32" min_val = float("-1.96827") @@ -1569,6 +1712,7 @@ class Program_weight_tensor_parameter_142: class Program_weight_tensor_parameter_143: name = "parameter_143" + original_name = "batch_norm_48.w_0" shape = [1024] dtype = "float32" min_val = float("0.11464") @@ -1580,6 +1724,7 @@ class Program_weight_tensor_parameter_143: class Program_weight_tensor_parameter_144: name = "parameter_144" + original_name = "batch_norm_48.w_2" shape = [1024] dtype = "float32" min_val = float("3.47382e-42") @@ -1591,6 +1736,7 @@ class Program_weight_tensor_parameter_144: class Program_weight_tensor_parameter_145: name = "parameter_145" + original_name = "batch_norm_48.w_1" shape = [1024] dtype = "float32" min_val = float("-1.48207") @@ -1602,6 +1748,7 @@ class Program_weight_tensor_parameter_145: class Program_weight_tensor_parameter_146: name = "parameter_146" + original_name = "conv2d_48.w_0" shape = [1024, 256, 1, 1] dtype = "float32" min_val = float("-0.319395") @@ -1613,6 +1760,7 @@ class Program_weight_tensor_parameter_146: class Program_weight_tensor_parameter_147: name = "parameter_147" + original_name = "batch_norm_47.b_0" shape = [256] dtype = "float32" min_val = float("-2.8702") @@ -1624,6 +1772,7 @@ class Program_weight_tensor_parameter_147: class Program_weight_tensor_parameter_148: name = "parameter_148" + original_name = "batch_norm_47.w_0" shape = [256] dtype = "float32" min_val = float("0.715975") @@ -1635,6 +1784,7 @@ class Program_weight_tensor_parameter_148: class Program_weight_tensor_parameter_149: name = "parameter_149" + original_name = "batch_norm_47.w_2" shape = [256] dtype = "float32" min_val = float("4.88071e-05") @@ -1646,6 +1796,7 @@ class Program_weight_tensor_parameter_149: class Program_weight_tensor_parameter_150: name = "parameter_150" + original_name = "batch_norm_47.w_1" shape = [256] dtype = "float32" min_val = float("-1.39141") @@ -1657,6 +1808,7 @@ class Program_weight_tensor_parameter_150: class Program_weight_tensor_parameter_151: name = "parameter_151" + original_name = "conv2d_47.w_0" shape = [256, 1, 3, 3] dtype = "float32" min_val = float("-0.250556") @@ -1668,6 +1820,7 @@ class Program_weight_tensor_parameter_151: class Program_weight_tensor_parameter_152: name = "parameter_152" + original_name = "batch_norm_46.b_0" shape = [256] dtype = "float32" min_val = float("-1.15953") @@ -1679,6 +1832,7 @@ class Program_weight_tensor_parameter_152: class Program_weight_tensor_parameter_153: name = "parameter_153" + original_name = "batch_norm_46.w_0" shape = [256] dtype = "float32" min_val = float("0.0110894") @@ -1690,6 +1844,7 @@ class Program_weight_tensor_parameter_153: class Program_weight_tensor_parameter_154: name = "parameter_154" + original_name = "batch_norm_46.w_2" shape = [256] dtype = "float32" min_val = float("0.189487") @@ -1701,6 +1856,7 @@ class Program_weight_tensor_parameter_154: class Program_weight_tensor_parameter_155: name = "parameter_155" + original_name = "batch_norm_46.w_1" shape = [256] dtype = "float32" min_val = float("-2.09085") @@ -1712,6 +1868,7 @@ class Program_weight_tensor_parameter_155: class Program_weight_tensor_parameter_156: name = "parameter_156" + original_name = "conv2d_46.w_0" shape = [256, 1024, 1, 1] dtype = "float32" min_val = float("-0.19291") @@ -1723,6 +1880,7 @@ class Program_weight_tensor_parameter_156: class Program_weight_tensor_parameter_157: name = "parameter_157" + original_name = "batch_norm_45.b_0" shape = [1024] dtype = "float32" min_val = float("-6.3781") @@ -1734,6 +1892,7 @@ class Program_weight_tensor_parameter_157: class Program_weight_tensor_parameter_158: name = "parameter_158" + original_name = "batch_norm_45.w_0" shape = [1024] dtype = "float32" min_val = float("0.0558595") @@ -1745,6 +1904,7 @@ class Program_weight_tensor_parameter_158: class Program_weight_tensor_parameter_159: name = "parameter_159" + original_name = "batch_norm_45.w_2" shape = [1024] dtype = "float32" min_val = float("2.65966e-42") @@ -1756,6 +1916,7 @@ class Program_weight_tensor_parameter_159: class Program_weight_tensor_parameter_160: name = "parameter_160" + original_name = "batch_norm_45.w_1" shape = [1024] dtype = "float32" min_val = float("-0.777197") @@ -1767,6 +1928,7 @@ class Program_weight_tensor_parameter_160: class Program_weight_tensor_parameter_161: name = "parameter_161" + original_name = "conv2d_45.w_0" shape = [1024, 1, 5, 5] dtype = "float32" min_val = float("-0.13") @@ -1778,6 +1940,7 @@ class Program_weight_tensor_parameter_161: class Program_weight_tensor_parameter_162: name = "parameter_162" + original_name = "batch_norm_44.b_0" shape = [1024] dtype = "float32" min_val = float("-1.84035") @@ -1789,6 +1952,7 @@ class Program_weight_tensor_parameter_162: class Program_weight_tensor_parameter_163: name = "parameter_163" + original_name = "batch_norm_44.w_0" shape = [1024] dtype = "float32" min_val = float("0.139966") @@ -1800,6 +1964,7 @@ class Program_weight_tensor_parameter_163: class Program_weight_tensor_parameter_164: name = "parameter_164" + original_name = "batch_norm_44.w_2" shape = [1024] dtype = "float32" min_val = float("3.43458e-42") @@ -1811,6 +1976,7 @@ class Program_weight_tensor_parameter_164: class Program_weight_tensor_parameter_165: name = "parameter_165" + original_name = "batch_norm_44.w_1" shape = [1024] dtype = "float32" min_val = float("-1.55113") @@ -1822,6 +1988,7 @@ class Program_weight_tensor_parameter_165: class Program_weight_tensor_parameter_166: name = "parameter_166" + original_name = "conv2d_44.w_0" shape = [1024, 256, 1, 1] dtype = "float32" min_val = float("-0.330392") @@ -1833,6 +2000,7 @@ class Program_weight_tensor_parameter_166: class Program_weight_tensor_parameter_167: name = "parameter_167" + original_name = "batch_norm_43.b_0" shape = [256] dtype = "float32" min_val = float("-2.35224") @@ -1844,6 +2012,7 @@ class Program_weight_tensor_parameter_167: class Program_weight_tensor_parameter_168: name = "parameter_168" + original_name = "batch_norm_43.w_0" shape = [256] dtype = "float32" min_val = float("0.658933") @@ -1855,6 +2024,7 @@ class Program_weight_tensor_parameter_168: class Program_weight_tensor_parameter_169: name = "parameter_169" + original_name = "batch_norm_43.w_2" shape = [256] dtype = "float32" min_val = float("2.18747e-05") @@ -1866,6 +2036,7 @@ class Program_weight_tensor_parameter_169: class Program_weight_tensor_parameter_170: name = "parameter_170" + original_name = "batch_norm_43.w_1" shape = [256] dtype = "float32" min_val = float("-2.24777") @@ -1877,6 +2048,7 @@ class Program_weight_tensor_parameter_170: class Program_weight_tensor_parameter_171: name = "parameter_171" + original_name = "conv2d_43.w_0" shape = [256, 1, 5, 5] dtype = "float32" min_val = float("-0.336473") @@ -1888,6 +2060,7 @@ class Program_weight_tensor_parameter_171: class Program_weight_tensor_parameter_172: name = "parameter_172" + original_name = "batch_norm_42.b_0" shape = [256] dtype = "float32" min_val = float("-2.67548") @@ -1899,6 +2072,7 @@ class Program_weight_tensor_parameter_172: class Program_weight_tensor_parameter_173: name = "parameter_173" + original_name = "batch_norm_42.w_0" shape = [256] dtype = "float32" min_val = float("0.0390857") @@ -1910,6 +2084,7 @@ class Program_weight_tensor_parameter_173: class Program_weight_tensor_parameter_174: name = "parameter_174" + original_name = "batch_norm_42.w_2" shape = [256] dtype = "float32" min_val = float("0.495518") @@ -1921,6 +2096,7 @@ class Program_weight_tensor_parameter_174: class Program_weight_tensor_parameter_175: name = "parameter_175" + original_name = "batch_norm_42.w_1" shape = [256] dtype = "float32" min_val = float("-5.2569") @@ -1932,6 +2108,7 @@ class Program_weight_tensor_parameter_175: class Program_weight_tensor_parameter_176: name = "parameter_176" + original_name = "conv2d_42.w_0" shape = [256, 960, 1, 1] dtype = "float32" min_val = float("-0.445156") @@ -1943,6 +2120,7 @@ class Program_weight_tensor_parameter_176: class Program_weight_tensor_parameter_177: name = "parameter_177" + original_name = "batch_norm_41.b_0" shape = [960] dtype = "float32" min_val = float("-4.33607") @@ -1954,6 +2132,7 @@ class Program_weight_tensor_parameter_177: class Program_weight_tensor_parameter_178: name = "parameter_178" + original_name = "batch_norm_41.w_0" shape = [960] dtype = "float32" min_val = float("0.305971") @@ -1965,6 +2144,7 @@ class Program_weight_tensor_parameter_178: class Program_weight_tensor_parameter_179: name = "parameter_179" + original_name = "batch_norm_41.w_2" shape = [960] dtype = "float32" min_val = float("3.50885e-42") @@ -1976,6 +2156,7 @@ class Program_weight_tensor_parameter_179: class Program_weight_tensor_parameter_180: name = "parameter_180" + original_name = "batch_norm_41.w_1" shape = [960] dtype = "float32" min_val = float("-0.766273") @@ -1987,6 +2168,7 @@ class Program_weight_tensor_parameter_180: class Program_weight_tensor_parameter_181: name = "parameter_181" + original_name = "conv2d_41.w_0" shape = [960, 1, 5, 5] dtype = "float32" min_val = float("-0.0825941") @@ -1998,6 +2180,7 @@ class Program_weight_tensor_parameter_181: class Program_weight_tensor_parameter_182: name = "parameter_182" + original_name = "batch_norm_40.b_0" shape = [960] dtype = "float32" min_val = float("-1.56285") @@ -2009,6 +2192,7 @@ class Program_weight_tensor_parameter_182: class Program_weight_tensor_parameter_183: name = "parameter_183" + original_name = "batch_norm_40.w_0" shape = [960] dtype = "float32" min_val = float("0.488292") @@ -2020,6 +2204,7 @@ class Program_weight_tensor_parameter_183: class Program_weight_tensor_parameter_184: name = "parameter_184" + original_name = "batch_norm_40.w_2" shape = [960] dtype = "float32" min_val = float("3.53688e-42") @@ -2031,6 +2216,7 @@ class Program_weight_tensor_parameter_184: class Program_weight_tensor_parameter_185: name = "parameter_185" + original_name = "batch_norm_40.w_1" shape = [960] dtype = "float32" min_val = float("-0.948002") @@ -2042,6 +2228,7 @@ class Program_weight_tensor_parameter_185: class Program_weight_tensor_parameter_186: name = "parameter_186" + original_name = "conv2d_40.w_0" shape = [960, 160, 1, 1] dtype = "float32" min_val = float("-0.506966") @@ -2053,6 +2240,7 @@ class Program_weight_tensor_parameter_186: class Program_weight_tensor_parameter_187: name = "parameter_187" + original_name = "batch_norm_39.b_0" shape = [160] dtype = "float32" min_val = float("-2.54285") @@ -2064,6 +2252,7 @@ class Program_weight_tensor_parameter_187: class Program_weight_tensor_parameter_188: name = "parameter_188" + original_name = "batch_norm_39.w_0" shape = [160] dtype = "float32" min_val = float("0.520903") @@ -2075,6 +2264,7 @@ class Program_weight_tensor_parameter_188: class Program_weight_tensor_parameter_189: name = "parameter_189" + original_name = "batch_norm_39.w_2" shape = [160] dtype = "float32" min_val = float("0.0165522") @@ -2086,6 +2276,7 @@ class Program_weight_tensor_parameter_189: class Program_weight_tensor_parameter_190: name = "parameter_190" + original_name = "batch_norm_39.w_1" shape = [160] dtype = "float32" min_val = float("-8.40565") @@ -2097,6 +2288,7 @@ class Program_weight_tensor_parameter_190: class Program_weight_tensor_parameter_191: name = "parameter_191" + original_name = "conv2d_39.w_0" shape = [160, 1, 5, 5] dtype = "float32" min_val = float("-0.337009") @@ -2108,6 +2300,7 @@ class Program_weight_tensor_parameter_191: class Program_weight_tensor_parameter_192: name = "parameter_192" + original_name = "batch_norm_38.b_0" shape = [160] dtype = "float32" min_val = float("-4.26969") @@ -2119,6 +2312,7 @@ class Program_weight_tensor_parameter_192: class Program_weight_tensor_parameter_193: name = "parameter_193" + original_name = "batch_norm_38.w_0" shape = [160] dtype = "float32" min_val = float("-0.324246") @@ -2130,6 +2324,7 @@ class Program_weight_tensor_parameter_193: class Program_weight_tensor_parameter_194: name = "parameter_194" + original_name = "batch_norm_38.w_2" shape = [160] dtype = "float32" min_val = float("0.00847427") @@ -2141,6 +2336,7 @@ class Program_weight_tensor_parameter_194: class Program_weight_tensor_parameter_195: name = "parameter_195" + original_name = "batch_norm_38.w_1" shape = [160] dtype = "float32" min_val = float("-0.184183") @@ -2152,6 +2348,7 @@ class Program_weight_tensor_parameter_195: class Program_weight_tensor_parameter_196: name = "parameter_196" + original_name = "conv2d_38.w_0" shape = [160, 640, 1, 1] dtype = "float32" min_val = float("-0.150129") @@ -2163,6 +2360,7 @@ class Program_weight_tensor_parameter_196: class Program_weight_tensor_parameter_197: name = "parameter_197" + original_name = "batch_norm_37.b_0" shape = [640] dtype = "float32" min_val = float("-4.11188") @@ -2174,6 +2372,7 @@ class Program_weight_tensor_parameter_197: class Program_weight_tensor_parameter_198: name = "parameter_198" + original_name = "batch_norm_37.w_0" shape = [640] dtype = "float32" min_val = float("0.177202") @@ -2185,6 +2384,7 @@ class Program_weight_tensor_parameter_198: class Program_weight_tensor_parameter_199: name = "parameter_199" + original_name = "batch_norm_37.w_2" shape = [640] dtype = "float32" min_val = float("3.50885e-42") @@ -2196,6 +2396,7 @@ class Program_weight_tensor_parameter_199: class Program_weight_tensor_parameter_200: name = "parameter_200" + original_name = "batch_norm_37.w_1" shape = [640] dtype = "float32" min_val = float("-1.04264") @@ -2207,6 +2408,7 @@ class Program_weight_tensor_parameter_200: class Program_weight_tensor_parameter_201: name = "parameter_201" + original_name = "conv2d_37.w_0" shape = [640, 160, 1, 1] dtype = "float32" min_val = float("-0.221754") @@ -2218,6 +2420,7 @@ class Program_weight_tensor_parameter_201: class Program_weight_tensor_parameter_202: name = "parameter_202" + original_name = "batch_norm_36.b_0" shape = [160] dtype = "float32" min_val = float("-2.61107") @@ -2229,6 +2432,7 @@ class Program_weight_tensor_parameter_202: class Program_weight_tensor_parameter_203: name = "parameter_203" + original_name = "batch_norm_36.w_0" shape = [160] dtype = "float32" min_val = float("-0.206667") @@ -2240,6 +2444,7 @@ class Program_weight_tensor_parameter_203: class Program_weight_tensor_parameter_204: name = "parameter_204" + original_name = "batch_norm_36.w_2" shape = [160] dtype = "float32" min_val = float("0.00290535") @@ -2251,6 +2456,7 @@ class Program_weight_tensor_parameter_204: class Program_weight_tensor_parameter_205: name = "parameter_205" + original_name = "batch_norm_36.w_1" shape = [160] dtype = "float32" min_val = float("-5.81446") @@ -2262,6 +2468,7 @@ class Program_weight_tensor_parameter_205: class Program_weight_tensor_parameter_206: name = "parameter_206" + original_name = "conv2d_36.w_0" shape = [160, 1, 3, 3] dtype = "float32" min_val = float("-0.250073") @@ -2273,6 +2480,7 @@ class Program_weight_tensor_parameter_206: class Program_weight_tensor_parameter_207: name = "parameter_207" + original_name = "batch_norm_35.b_0" shape = [160] dtype = "float32" min_val = float("-2.64034") @@ -2284,6 +2492,7 @@ class Program_weight_tensor_parameter_207: class Program_weight_tensor_parameter_208: name = "parameter_208" + original_name = "batch_norm_35.w_0" shape = [160] dtype = "float32" min_val = float("-0.184872") @@ -2295,6 +2504,7 @@ class Program_weight_tensor_parameter_208: class Program_weight_tensor_parameter_209: name = "parameter_209" + original_name = "batch_norm_35.w_2" shape = [160] dtype = "float32" min_val = float("0.00545003") @@ -2306,6 +2516,7 @@ class Program_weight_tensor_parameter_209: class Program_weight_tensor_parameter_210: name = "parameter_210" + original_name = "batch_norm_35.w_1" shape = [160] dtype = "float32" min_val = float("-0.189513") @@ -2317,6 +2528,7 @@ class Program_weight_tensor_parameter_210: class Program_weight_tensor_parameter_211: name = "parameter_211" + original_name = "conv2d_35.w_0" shape = [160, 320, 1, 1] dtype = "float32" min_val = float("-0.252249") @@ -2328,6 +2540,7 @@ class Program_weight_tensor_parameter_211: class Program_weight_tensor_parameter_212: name = "parameter_212" + original_name = "batch_norm_34.b_0" shape = [320] dtype = "float32" min_val = float("-4.01418") @@ -2339,6 +2552,7 @@ class Program_weight_tensor_parameter_212: class Program_weight_tensor_parameter_213: name = "parameter_213" + original_name = "batch_norm_34.w_0" shape = [320] dtype = "float32" min_val = float("0.45528") @@ -2350,6 +2564,7 @@ class Program_weight_tensor_parameter_213: class Program_weight_tensor_parameter_214: name = "parameter_214" + original_name = "batch_norm_34.w_2" shape = [320] dtype = "float32" min_val = float("6.34629e-31") @@ -2361,6 +2576,7 @@ class Program_weight_tensor_parameter_214: class Program_weight_tensor_parameter_215: name = "parameter_215" + original_name = "batch_norm_34.w_1" shape = [320] dtype = "float32" min_val = float("-4.97406") @@ -2372,6 +2588,7 @@ class Program_weight_tensor_parameter_215: class Program_weight_tensor_parameter_216: name = "parameter_216" + original_name = "conv2d_34.w_0" shape = [320, 160, 1, 1] dtype = "float32" min_val = float("-0.288887") @@ -2383,6 +2600,7 @@ class Program_weight_tensor_parameter_216: class Program_weight_tensor_parameter_217: name = "parameter_217" + original_name = "batch_norm_33.b_0" shape = [160] dtype = "float32" min_val = float("-3.46082") @@ -2394,6 +2612,7 @@ class Program_weight_tensor_parameter_217: class Program_weight_tensor_parameter_218: name = "parameter_218" + original_name = "batch_norm_33.w_0" shape = [160] dtype = "float32" min_val = float("-0.0650295") @@ -2405,6 +2624,7 @@ class Program_weight_tensor_parameter_218: class Program_weight_tensor_parameter_219: name = "parameter_219" + original_name = "batch_norm_33.w_2" shape = [160] dtype = "float32" min_val = float("0.0078126") @@ -2416,6 +2636,7 @@ class Program_weight_tensor_parameter_219: class Program_weight_tensor_parameter_220: name = "parameter_220" + original_name = "batch_norm_33.w_1" shape = [160] dtype = "float32" min_val = float("-0.212787") @@ -2427,6 +2648,7 @@ class Program_weight_tensor_parameter_220: class Program_weight_tensor_parameter_221: name = "parameter_221" + original_name = "conv2d_33.w_0" shape = [160, 640, 1, 1] dtype = "float32" min_val = float("-0.303016") @@ -2438,6 +2660,7 @@ class Program_weight_tensor_parameter_221: class Program_weight_tensor_parameter_222: name = "parameter_222" + original_name = "batch_norm_32.b_0" shape = [640] dtype = "float32" min_val = float("-8.16424") @@ -2449,6 +2672,7 @@ class Program_weight_tensor_parameter_222: class Program_weight_tensor_parameter_223: name = "parameter_223" + original_name = "batch_norm_32.w_0" shape = [640] dtype = "float32" min_val = float("0.0720156") @@ -2460,6 +2684,7 @@ class Program_weight_tensor_parameter_223: class Program_weight_tensor_parameter_224: name = "parameter_224" + original_name = "batch_norm_32.w_2" shape = [640] dtype = "float32" min_val = float("3.50885e-42") @@ -2471,6 +2696,7 @@ class Program_weight_tensor_parameter_224: class Program_weight_tensor_parameter_225: name = "parameter_225" + original_name = "batch_norm_32.w_1" shape = [640] dtype = "float32" min_val = float("-1.12349") @@ -2482,6 +2708,7 @@ class Program_weight_tensor_parameter_225: class Program_weight_tensor_parameter_226: name = "parameter_226" + original_name = "conv2d_32.w_0" shape = [640, 160, 1, 1] dtype = "float32" min_val = float("-0.408777") @@ -2493,6 +2720,7 @@ class Program_weight_tensor_parameter_226: class Program_weight_tensor_parameter_227: name = "parameter_227" + original_name = "batch_norm_31.b_0" shape = [160] dtype = "float32" min_val = float("-2.03558") @@ -2504,6 +2732,7 @@ class Program_weight_tensor_parameter_227: class Program_weight_tensor_parameter_228: name = "parameter_228" + original_name = "batch_norm_31.w_0" shape = [160] dtype = "float32" min_val = float("0.604025") @@ -2515,6 +2744,7 @@ class Program_weight_tensor_parameter_228: class Program_weight_tensor_parameter_229: name = "parameter_229" + original_name = "batch_norm_31.w_2" shape = [160] dtype = "float32" min_val = float("9.91693e-05") @@ -2526,6 +2756,7 @@ class Program_weight_tensor_parameter_229: class Program_weight_tensor_parameter_230: name = "parameter_230" + original_name = "batch_norm_31.w_1" shape = [160] dtype = "float32" min_val = float("-0.844152") @@ -2537,6 +2768,7 @@ class Program_weight_tensor_parameter_230: class Program_weight_tensor_parameter_231: name = "parameter_231" + original_name = "conv2d_31.w_0" shape = [160, 1, 3, 3] dtype = "float32" min_val = float("-0.283086") @@ -2548,6 +2780,7 @@ class Program_weight_tensor_parameter_231: class Program_weight_tensor_parameter_232: name = "parameter_232" + original_name = "batch_norm_30.b_0" shape = [160] dtype = "float32" min_val = float("-4.34597") @@ -2559,6 +2792,7 @@ class Program_weight_tensor_parameter_232: class Program_weight_tensor_parameter_233: name = "parameter_233" + original_name = "batch_norm_30.w_0" shape = [160] dtype = "float32" min_val = float("0.0421184") @@ -2570,6 +2804,7 @@ class Program_weight_tensor_parameter_233: class Program_weight_tensor_parameter_234: name = "parameter_234" + original_name = "batch_norm_30.w_2" shape = [160] dtype = "float32" min_val = float("0.0565186") @@ -2581,6 +2816,7 @@ class Program_weight_tensor_parameter_234: class Program_weight_tensor_parameter_235: name = "parameter_235" + original_name = "batch_norm_30.w_1" shape = [160] dtype = "float32" min_val = float("-1.85269") @@ -2592,6 +2828,7 @@ class Program_weight_tensor_parameter_235: class Program_weight_tensor_parameter_236: name = "parameter_236" + original_name = "conv2d_30.w_0" shape = [160, 640, 1, 1] dtype = "float32" min_val = float("-0.305928") @@ -2603,6 +2840,7 @@ class Program_weight_tensor_parameter_236: class Program_weight_tensor_parameter_237: name = "parameter_237" + original_name = "batch_norm_29.b_0" shape = [640] dtype = "float32" min_val = float("-15.2901") @@ -2614,6 +2852,7 @@ class Program_weight_tensor_parameter_237: class Program_weight_tensor_parameter_238: name = "parameter_238" + original_name = "batch_norm_29.w_0" shape = [640] dtype = "float32" min_val = float("0.00447807") @@ -2625,6 +2864,7 @@ class Program_weight_tensor_parameter_238: class Program_weight_tensor_parameter_239: name = "parameter_239" + original_name = "batch_norm_29.w_2" shape = [640] dtype = "float32" min_val = float("2.60642e-42") @@ -2636,6 +2876,7 @@ class Program_weight_tensor_parameter_239: class Program_weight_tensor_parameter_240: name = "parameter_240" + original_name = "batch_norm_29.w_1" shape = [640] dtype = "float32" min_val = float("-0.396532") @@ -2647,6 +2888,7 @@ class Program_weight_tensor_parameter_240: class Program_weight_tensor_parameter_241: name = "parameter_241" + original_name = "conv2d_29.w_0" shape = [640, 1, 3, 3] dtype = "float32" min_val = float("-0.280622") @@ -2658,6 +2900,7 @@ class Program_weight_tensor_parameter_241: class Program_weight_tensor_parameter_242: name = "parameter_242" + original_name = "batch_norm_28.b_0" shape = [640] dtype = "float32" min_val = float("-1.83315") @@ -2669,6 +2912,7 @@ class Program_weight_tensor_parameter_242: class Program_weight_tensor_parameter_243: name = "parameter_243" + original_name = "batch_norm_28.w_0" shape = [640] dtype = "float32" min_val = float("0.0628849") @@ -2680,6 +2924,7 @@ class Program_weight_tensor_parameter_243: class Program_weight_tensor_parameter_244: name = "parameter_244" + original_name = "batch_norm_28.w_2" shape = [640] dtype = "float32" min_val = float("3.46121e-42") @@ -2691,6 +2936,7 @@ class Program_weight_tensor_parameter_244: class Program_weight_tensor_parameter_245: name = "parameter_245" + original_name = "batch_norm_28.w_1" shape = [640] dtype = "float32" min_val = float("-0.891478") @@ -2702,6 +2948,7 @@ class Program_weight_tensor_parameter_245: class Program_weight_tensor_parameter_246: name = "parameter_246" + original_name = "conv2d_28.w_0" shape = [640, 160, 1, 1] dtype = "float32" min_val = float("-0.306686") @@ -2713,6 +2960,7 @@ class Program_weight_tensor_parameter_246: class Program_weight_tensor_parameter_247: name = "parameter_247" + original_name = "batch_norm_27.b_0" shape = [160] dtype = "float32" min_val = float("-2.54784") @@ -2724,6 +2972,7 @@ class Program_weight_tensor_parameter_247: class Program_weight_tensor_parameter_248: name = "parameter_248" + original_name = "batch_norm_27.w_0" shape = [160] dtype = "float32" min_val = float("-0.128419") @@ -2735,6 +2984,7 @@ class Program_weight_tensor_parameter_248: class Program_weight_tensor_parameter_249: name = "parameter_249" + original_name = "batch_norm_27.w_2" shape = [160] dtype = "float32" min_val = float("0.000457411") @@ -2746,6 +2996,7 @@ class Program_weight_tensor_parameter_249: class Program_weight_tensor_parameter_250: name = "parameter_250" + original_name = "batch_norm_27.w_1" shape = [160] dtype = "float32" min_val = float("-1.13942") @@ -2757,6 +3008,7 @@ class Program_weight_tensor_parameter_250: class Program_weight_tensor_parameter_251: name = "parameter_251" + original_name = "conv2d_27.w_0" shape = [160, 1, 3, 3] dtype = "float32" min_val = float("-0.324471") @@ -2768,6 +3020,7 @@ class Program_weight_tensor_parameter_251: class Program_weight_tensor_parameter_252: name = "parameter_252" + original_name = "batch_norm_26.b_0" shape = [160] dtype = "float32" min_val = float("-3.04951") @@ -2779,6 +3032,7 @@ class Program_weight_tensor_parameter_252: class Program_weight_tensor_parameter_253: name = "parameter_253" + original_name = "batch_norm_26.w_0" shape = [160] dtype = "float32" min_val = float("-0.0836476") @@ -2790,6 +3044,7 @@ class Program_weight_tensor_parameter_253: class Program_weight_tensor_parameter_254: name = "parameter_254" + original_name = "batch_norm_26.w_2" shape = [160] dtype = "float32" min_val = float("0.0431994") @@ -2801,6 +3056,7 @@ class Program_weight_tensor_parameter_254: class Program_weight_tensor_parameter_255: name = "parameter_255" + original_name = "batch_norm_26.w_1" shape = [160] dtype = "float32" min_val = float("-0.712717") @@ -2812,6 +3068,7 @@ class Program_weight_tensor_parameter_255: class Program_weight_tensor_parameter_256: name = "parameter_256" + original_name = "conv2d_26.w_0" shape = [160, 640, 1, 1] dtype = "float32" min_val = float("-0.282634") @@ -2823,6 +3080,7 @@ class Program_weight_tensor_parameter_256: class Program_weight_tensor_parameter_257: name = "parameter_257" + original_name = "batch_norm_25.b_0" shape = [640] dtype = "float32" min_val = float("-18.0885") @@ -2834,6 +3092,7 @@ class Program_weight_tensor_parameter_257: class Program_weight_tensor_parameter_258: name = "parameter_258" + original_name = "batch_norm_25.w_0" shape = [640] dtype = "float32" min_val = float("0.0181023") @@ -2845,6 +3104,7 @@ class Program_weight_tensor_parameter_258: class Program_weight_tensor_parameter_259: name = "parameter_259" + original_name = "batch_norm_25.w_2" shape = [640] dtype = "float32" min_val = float("2.67788e-42") @@ -2856,6 +3116,7 @@ class Program_weight_tensor_parameter_259: class Program_weight_tensor_parameter_260: name = "parameter_260" + original_name = "batch_norm_25.w_1" shape = [640] dtype = "float32" min_val = float("-0.645869") @@ -2867,6 +3128,7 @@ class Program_weight_tensor_parameter_260: class Program_weight_tensor_parameter_261: name = "parameter_261" + original_name = "conv2d_25.w_0" shape = [640, 1, 5, 5] dtype = "float32" min_val = float("-0.173642") @@ -2878,6 +3140,7 @@ class Program_weight_tensor_parameter_261: class Program_weight_tensor_parameter_262: name = "parameter_262" + original_name = "batch_norm_24.b_0" shape = [640] dtype = "float32" min_val = float("-1.61983") @@ -2889,6 +3152,7 @@ class Program_weight_tensor_parameter_262: class Program_weight_tensor_parameter_263: name = "parameter_263" + original_name = "batch_norm_24.w_0" shape = [640] dtype = "float32" min_val = float("0.134278") @@ -2900,6 +3164,7 @@ class Program_weight_tensor_parameter_263: class Program_weight_tensor_parameter_264: name = "parameter_264" + original_name = "batch_norm_24.w_2" shape = [640] dtype = "float32" min_val = float("3.50885e-42") @@ -2911,6 +3176,7 @@ class Program_weight_tensor_parameter_264: class Program_weight_tensor_parameter_265: name = "parameter_265" + original_name = "batch_norm_24.w_1" shape = [640] dtype = "float32" min_val = float("-1.17255") @@ -2922,6 +3188,7 @@ class Program_weight_tensor_parameter_265: class Program_weight_tensor_parameter_266: name = "parameter_266" + original_name = "conv2d_24.w_0" shape = [640, 160, 1, 1] dtype = "float32" min_val = float("-0.289419") @@ -2933,6 +3200,7 @@ class Program_weight_tensor_parameter_266: class Program_weight_tensor_parameter_267: name = "parameter_267" + original_name = "batch_norm_23.b_0" shape = [160] dtype = "float32" min_val = float("-2.05983") @@ -2944,6 +3212,7 @@ class Program_weight_tensor_parameter_267: class Program_weight_tensor_parameter_268: name = "parameter_268" + original_name = "batch_norm_23.w_0" shape = [160] dtype = "float32" min_val = float("0.347568") @@ -2955,6 +3224,7 @@ class Program_weight_tensor_parameter_268: class Program_weight_tensor_parameter_269: name = "parameter_269" + original_name = "batch_norm_23.w_2" shape = [160] dtype = "float32" min_val = float("8.6487e-05") @@ -2966,6 +3236,7 @@ class Program_weight_tensor_parameter_269: class Program_weight_tensor_parameter_270: name = "parameter_270" + original_name = "batch_norm_23.w_1" shape = [160] dtype = "float32" min_val = float("-0.559256") @@ -2977,6 +3248,7 @@ class Program_weight_tensor_parameter_270: class Program_weight_tensor_parameter_271: name = "parameter_271" + original_name = "conv2d_23.w_0" shape = [160, 1, 3, 3] dtype = "float32" min_val = float("-0.303426") @@ -2988,6 +3260,7 @@ class Program_weight_tensor_parameter_271: class Program_weight_tensor_parameter_272: name = "parameter_272" + original_name = "batch_norm_22.b_0" shape = [160] dtype = "float32" min_val = float("-2.65274") @@ -2999,6 +3272,7 @@ class Program_weight_tensor_parameter_272: class Program_weight_tensor_parameter_273: name = "parameter_273" + original_name = "batch_norm_22.w_0" shape = [160] dtype = "float32" min_val = float("0.0256737") @@ -3010,6 +3284,7 @@ class Program_weight_tensor_parameter_273: class Program_weight_tensor_parameter_274: name = "parameter_274" + original_name = "batch_norm_22.w_2" shape = [160] dtype = "float32" min_val = float("0.0687884") @@ -3021,6 +3296,7 @@ class Program_weight_tensor_parameter_274: class Program_weight_tensor_parameter_275: name = "parameter_275" + original_name = "batch_norm_22.w_1" shape = [160] dtype = "float32" min_val = float("-1.19179") @@ -3032,6 +3308,7 @@ class Program_weight_tensor_parameter_275: class Program_weight_tensor_parameter_276: name = "parameter_276" + original_name = "conv2d_22.w_0" shape = [160, 640, 1, 1] dtype = "float32" min_val = float("-0.432576") @@ -3043,6 +3320,7 @@ class Program_weight_tensor_parameter_276: class Program_weight_tensor_parameter_277: name = "parameter_277" + original_name = "batch_norm_21.b_0" shape = [640] dtype = "float32" min_val = float("-7.42145") @@ -3054,6 +3332,7 @@ class Program_weight_tensor_parameter_277: class Program_weight_tensor_parameter_278: name = "parameter_278" + original_name = "batch_norm_21.w_0" shape = [640] dtype = "float32" min_val = float("0.0366009") @@ -3065,6 +3344,7 @@ class Program_weight_tensor_parameter_278: class Program_weight_tensor_parameter_279: name = "parameter_279" + original_name = "batch_norm_21.w_2" shape = [640] dtype = "float32" min_val = float("2.591e-42") @@ -3076,6 +3356,7 @@ class Program_weight_tensor_parameter_279: class Program_weight_tensor_parameter_280: name = "parameter_280" + original_name = "batch_norm_21.w_1" shape = [640] dtype = "float32" min_val = float("-0.774091") @@ -3087,6 +3368,7 @@ class Program_weight_tensor_parameter_280: class Program_weight_tensor_parameter_281: name = "parameter_281" + original_name = "conv2d_21.w_0" shape = [640, 1, 3, 3] dtype = "float32" min_val = float("-0.286938") @@ -3098,6 +3380,7 @@ class Program_weight_tensor_parameter_281: class Program_weight_tensor_parameter_282: name = "parameter_282" + original_name = "batch_norm_20.b_0" shape = [640] dtype = "float32" min_val = float("-1.93225") @@ -3109,6 +3392,7 @@ class Program_weight_tensor_parameter_282: class Program_weight_tensor_parameter_283: name = "parameter_283" + original_name = "batch_norm_20.w_0" shape = [640] dtype = "float32" min_val = float("0.0153546") @@ -3120,6 +3404,7 @@ class Program_weight_tensor_parameter_283: class Program_weight_tensor_parameter_284: name = "parameter_284" + original_name = "batch_norm_20.w_2" shape = [640] dtype = "float32" min_val = float("3.50885e-42") @@ -3131,6 +3416,7 @@ class Program_weight_tensor_parameter_284: class Program_weight_tensor_parameter_285: name = "parameter_285" + original_name = "batch_norm_20.w_1" shape = [640] dtype = "float32" min_val = float("-1.34469") @@ -3142,6 +3428,7 @@ class Program_weight_tensor_parameter_285: class Program_weight_tensor_parameter_286: name = "parameter_286" + original_name = "conv2d_20.w_0" shape = [640, 160, 1, 1] dtype = "float32" min_val = float("-0.387298") @@ -3153,6 +3440,7 @@ class Program_weight_tensor_parameter_286: class Program_weight_tensor_parameter_287: name = "parameter_287" + original_name = "batch_norm_19.b_0" shape = [160] dtype = "float32" min_val = float("-2.41874") @@ -3164,6 +3452,7 @@ class Program_weight_tensor_parameter_287: class Program_weight_tensor_parameter_288: name = "parameter_288" + original_name = "batch_norm_19.w_0" shape = [160] dtype = "float32" min_val = float("0.267499") @@ -3175,6 +3464,7 @@ class Program_weight_tensor_parameter_288: class Program_weight_tensor_parameter_289: name = "parameter_289" + original_name = "batch_norm_19.w_2" shape = [160] dtype = "float32" min_val = float("1.46603e-05") @@ -3186,6 +3476,7 @@ class Program_weight_tensor_parameter_289: class Program_weight_tensor_parameter_290: name = "parameter_290" + original_name = "batch_norm_19.w_1" shape = [160] dtype = "float32" min_val = float("-0.523491") @@ -3197,6 +3488,7 @@ class Program_weight_tensor_parameter_290: class Program_weight_tensor_parameter_291: name = "parameter_291" + original_name = "conv2d_19.w_0" shape = [160, 1, 3, 3] dtype = "float32" min_val = float("-0.272732") @@ -3208,6 +3500,7 @@ class Program_weight_tensor_parameter_291: class Program_weight_tensor_parameter_292: name = "parameter_292" + original_name = "batch_norm_18.b_0" shape = [160] dtype = "float32" min_val = float("-2.41706") @@ -3219,6 +3512,7 @@ class Program_weight_tensor_parameter_292: class Program_weight_tensor_parameter_293: name = "parameter_293" + original_name = "batch_norm_18.w_0" shape = [160] dtype = "float32" min_val = float("-0.115328") @@ -3230,6 +3524,7 @@ class Program_weight_tensor_parameter_293: class Program_weight_tensor_parameter_294: name = "parameter_294" + original_name = "batch_norm_18.w_2" shape = [160] dtype = "float32" min_val = float("0.0433159") @@ -3241,6 +3536,7 @@ class Program_weight_tensor_parameter_294: class Program_weight_tensor_parameter_295: name = "parameter_295" + original_name = "batch_norm_18.w_1" shape = [160] dtype = "float32" min_val = float("-1.6948") @@ -3252,6 +3548,7 @@ class Program_weight_tensor_parameter_295: class Program_weight_tensor_parameter_296: name = "parameter_296" + original_name = "conv2d_18.w_0" shape = [160, 640, 1, 1] dtype = "float32" min_val = float("-0.359081") @@ -3263,6 +3560,7 @@ class Program_weight_tensor_parameter_296: class Program_weight_tensor_parameter_297: name = "parameter_297" + original_name = "batch_norm_17.b_0" shape = [640] dtype = "float32" min_val = float("-10.0311") @@ -3274,6 +3572,7 @@ class Program_weight_tensor_parameter_297: class Program_weight_tensor_parameter_298: name = "parameter_298" + original_name = "batch_norm_17.w_0" shape = [640] dtype = "float32" min_val = float("0.143022") @@ -3285,6 +3584,7 @@ class Program_weight_tensor_parameter_298: class Program_weight_tensor_parameter_299: name = "parameter_299" + original_name = "batch_norm_17.w_2" shape = [640] dtype = "float32" min_val = float("2.61202e-42") @@ -3296,6 +3596,7 @@ class Program_weight_tensor_parameter_299: class Program_weight_tensor_parameter_300: name = "parameter_300" + original_name = "batch_norm_17.w_1" shape = [640] dtype = "float32" min_val = float("-0.411761") @@ -3307,6 +3608,7 @@ class Program_weight_tensor_parameter_300: class Program_weight_tensor_parameter_301: name = "parameter_301" + original_name = "conv2d_17.w_0" shape = [640, 1, 3, 3] dtype = "float32" min_val = float("-0.187617") @@ -3318,6 +3620,7 @@ class Program_weight_tensor_parameter_301: class Program_weight_tensor_parameter_302: name = "parameter_302" + original_name = "batch_norm_16.b_0" shape = [640] dtype = "float32" min_val = float("-2.17507") @@ -3329,6 +3632,7 @@ class Program_weight_tensor_parameter_302: class Program_weight_tensor_parameter_303: name = "parameter_303" + original_name = "batch_norm_16.w_0" shape = [640] dtype = "float32" min_val = float("-0.0383393") @@ -3340,6 +3644,7 @@ class Program_weight_tensor_parameter_303: class Program_weight_tensor_parameter_304: name = "parameter_304" + original_name = "batch_norm_16.w_2" shape = [640] dtype = "float32" min_val = float("3.48363e-42") @@ -3351,6 +3656,7 @@ class Program_weight_tensor_parameter_304: class Program_weight_tensor_parameter_305: name = "parameter_305" + original_name = "batch_norm_16.w_1" shape = [640] dtype = "float32" min_val = float("-1.06467") @@ -3362,6 +3668,7 @@ class Program_weight_tensor_parameter_305: class Program_weight_tensor_parameter_306: name = "parameter_306" + original_name = "conv2d_16.w_0" shape = [640, 160, 1, 1] dtype = "float32" min_val = float("-0.291123") @@ -3373,6 +3680,7 @@ class Program_weight_tensor_parameter_306: class Program_weight_tensor_parameter_307: name = "parameter_307" + original_name = "batch_norm_15.b_0" shape = [160] dtype = "float32" min_val = float("-2.05643") @@ -3384,6 +3692,7 @@ class Program_weight_tensor_parameter_307: class Program_weight_tensor_parameter_308: name = "parameter_308" + original_name = "batch_norm_15.w_0" shape = [160] dtype = "float32" min_val = float("-0.570453") @@ -3395,6 +3704,7 @@ class Program_weight_tensor_parameter_308: class Program_weight_tensor_parameter_309: name = "parameter_309" + original_name = "batch_norm_15.w_2" shape = [160] dtype = "float32" min_val = float("1.43325e-05") @@ -3406,6 +3716,7 @@ class Program_weight_tensor_parameter_309: class Program_weight_tensor_parameter_310: name = "parameter_310" + original_name = "batch_norm_15.w_1" shape = [160] dtype = "float32" min_val = float("-0.49403") @@ -3417,6 +3728,7 @@ class Program_weight_tensor_parameter_310: class Program_weight_tensor_parameter_311: name = "parameter_311" + original_name = "conv2d_15.w_0" shape = [160, 1, 3, 3] dtype = "float32" min_val = float("-0.332538") @@ -3428,6 +3740,7 @@ class Program_weight_tensor_parameter_311: class Program_weight_tensor_parameter_312: name = "parameter_312" + original_name = "batch_norm_14.b_0" shape = [160] dtype = "float32" min_val = float("-4.26341") @@ -3439,6 +3752,7 @@ class Program_weight_tensor_parameter_312: class Program_weight_tensor_parameter_313: name = "parameter_313" + original_name = "batch_norm_14.w_0" shape = [160] dtype = "float32" min_val = float("0.0185652") @@ -3450,6 +3764,7 @@ class Program_weight_tensor_parameter_313: class Program_weight_tensor_parameter_314: name = "parameter_314" + original_name = "batch_norm_14.w_2" shape = [160] dtype = "float32" min_val = float("0.123508") @@ -3461,6 +3776,7 @@ class Program_weight_tensor_parameter_314: class Program_weight_tensor_parameter_315: name = "parameter_315" + original_name = "batch_norm_14.w_1" shape = [160] dtype = "float32" min_val = float("-3.41777") @@ -3472,6 +3788,7 @@ class Program_weight_tensor_parameter_315: class Program_weight_tensor_parameter_316: name = "parameter_316" + original_name = "conv2d_14.w_0" shape = [160, 480, 1, 1] dtype = "float32" min_val = float("-0.275705") @@ -3483,6 +3800,7 @@ class Program_weight_tensor_parameter_316: class Program_weight_tensor_parameter_317: name = "parameter_317" + original_name = "batch_norm_13.b_0" shape = [480] dtype = "float32" min_val = float("-3.71339") @@ -3494,6 +3812,7 @@ class Program_weight_tensor_parameter_317: class Program_weight_tensor_parameter_318: name = "parameter_318" + original_name = "batch_norm_13.w_0" shape = [480] dtype = "float32" min_val = float("0.488664") @@ -3505,6 +3824,7 @@ class Program_weight_tensor_parameter_318: class Program_weight_tensor_parameter_319: name = "parameter_319" + original_name = "batch_norm_13.w_2" shape = [480] dtype = "float32" min_val = float("2.86005e-42") @@ -3516,6 +3836,7 @@ class Program_weight_tensor_parameter_319: class Program_weight_tensor_parameter_320: name = "parameter_320" + original_name = "batch_norm_13.w_1" shape = [480] dtype = "float32" min_val = float("-1.02632") @@ -3527,6 +3848,7 @@ class Program_weight_tensor_parameter_320: class Program_weight_tensor_parameter_321: name = "parameter_321" + original_name = "conv2d_13.w_0" shape = [480, 1, 5, 5] dtype = "float32" min_val = float("-0.0995059") @@ -3538,6 +3860,7 @@ class Program_weight_tensor_parameter_321: class Program_weight_tensor_parameter_322: name = "parameter_322" + original_name = "batch_norm_12.b_0" shape = [480] dtype = "float32" min_val = float("-1.561") @@ -3549,6 +3872,7 @@ class Program_weight_tensor_parameter_322: class Program_weight_tensor_parameter_323: name = "parameter_323" + original_name = "batch_norm_12.w_0" shape = [480] dtype = "float32" min_val = float("0.172465") @@ -3560,6 +3884,7 @@ class Program_weight_tensor_parameter_323: class Program_weight_tensor_parameter_324: name = "parameter_324" + original_name = "batch_norm_12.w_2" shape = [480] dtype = "float32" min_val = float("3.50885e-42") @@ -3571,6 +3896,7 @@ class Program_weight_tensor_parameter_324: class Program_weight_tensor_parameter_325: name = "parameter_325" + original_name = "batch_norm_12.w_1" shape = [480] dtype = "float32" min_val = float("-0.877657") @@ -3582,6 +3908,7 @@ class Program_weight_tensor_parameter_325: class Program_weight_tensor_parameter_326: name = "parameter_326" + original_name = "conv2d_12.w_0" shape = [480, 80, 1, 1] dtype = "float32" min_val = float("-0.325539") @@ -3593,6 +3920,7 @@ class Program_weight_tensor_parameter_326: class Program_weight_tensor_parameter_327: name = "parameter_327" + original_name = "batch_norm_11.b_0" shape = [80] dtype = "float32" min_val = float("-1.96528") @@ -3604,6 +3932,7 @@ class Program_weight_tensor_parameter_327: class Program_weight_tensor_parameter_328: name = "parameter_328" + original_name = "batch_norm_11.w_0" shape = [80] dtype = "float32" min_val = float("0.571272") @@ -3615,6 +3944,7 @@ class Program_weight_tensor_parameter_328: class Program_weight_tensor_parameter_329: name = "parameter_329" + original_name = "batch_norm_11.w_2" shape = [80] dtype = "float32" min_val = float("0.00450147") @@ -3626,6 +3956,7 @@ class Program_weight_tensor_parameter_329: class Program_weight_tensor_parameter_330: name = "parameter_330" + original_name = "batch_norm_11.w_1" shape = [80] dtype = "float32" min_val = float("-0.610591") @@ -3637,6 +3968,7 @@ class Program_weight_tensor_parameter_330: class Program_weight_tensor_parameter_331: name = "parameter_331" + original_name = "conv2d_11.w_0" shape = [80, 1, 3, 3] dtype = "float32" min_val = float("-0.197854") @@ -3648,6 +3980,7 @@ class Program_weight_tensor_parameter_331: class Program_weight_tensor_parameter_332: name = "parameter_332" + original_name = "batch_norm_10.b_0" shape = [80] dtype = "float32" min_val = float("-2.7216") @@ -3659,6 +3992,7 @@ class Program_weight_tensor_parameter_332: class Program_weight_tensor_parameter_333: name = "parameter_333" + original_name = "batch_norm_10.w_0" shape = [80] dtype = "float32" min_val = float("-0.059897") @@ -3670,6 +4004,7 @@ class Program_weight_tensor_parameter_333: class Program_weight_tensor_parameter_334: name = "parameter_334" + original_name = "batch_norm_10.w_2" shape = [80] dtype = "float32" min_val = float("0.023872") @@ -3681,6 +4016,7 @@ class Program_weight_tensor_parameter_334: class Program_weight_tensor_parameter_335: name = "parameter_335" + original_name = "batch_norm_10.w_1" shape = [80] dtype = "float32" min_val = float("-0.672421") @@ -3692,6 +4028,7 @@ class Program_weight_tensor_parameter_335: class Program_weight_tensor_parameter_336: name = "parameter_336" + original_name = "conv2d_10.w_0" shape = [80, 160, 1, 1] dtype = "float32" min_val = float("-0.218933") @@ -3703,6 +4040,7 @@ class Program_weight_tensor_parameter_336: class Program_weight_tensor_parameter_337: name = "parameter_337" + original_name = "batch_norm_9.b_0" shape = [160] dtype = "float32" min_val = float("-4.77325") @@ -3714,6 +4052,7 @@ class Program_weight_tensor_parameter_337: class Program_weight_tensor_parameter_338: name = "parameter_338" + original_name = "batch_norm_9.w_0" shape = [160] dtype = "float32" min_val = float("0.378542") @@ -3725,6 +4064,7 @@ class Program_weight_tensor_parameter_338: class Program_weight_tensor_parameter_339: name = "parameter_339" + original_name = "batch_norm_9.w_2" shape = [160] dtype = "float32" min_val = float("8.45452e-05") @@ -3736,6 +4076,7 @@ class Program_weight_tensor_parameter_339: class Program_weight_tensor_parameter_340: name = "parameter_340" + original_name = "batch_norm_9.w_1" shape = [160] dtype = "float32" min_val = float("-0.715293") @@ -3747,6 +4088,7 @@ class Program_weight_tensor_parameter_340: class Program_weight_tensor_parameter_341: name = "parameter_341" + original_name = "conv2d_9.w_0" shape = [160, 1, 3, 3] dtype = "float32" min_val = float("-0.184543") @@ -3758,6 +4100,7 @@ class Program_weight_tensor_parameter_341: class Program_weight_tensor_parameter_342: name = "parameter_342" + original_name = "batch_norm_8.b_0" shape = [160] dtype = "float32" min_val = float("-1.32656") @@ -3769,6 +4112,7 @@ class Program_weight_tensor_parameter_342: class Program_weight_tensor_parameter_343: name = "parameter_343" + original_name = "batch_norm_8.w_0" shape = [160] dtype = "float32" min_val = float("0.246593") @@ -3780,6 +4124,7 @@ class Program_weight_tensor_parameter_343: class Program_weight_tensor_parameter_344: name = "parameter_344" + original_name = "batch_norm_8.w_2" shape = [160] dtype = "float32" min_val = float("0.0475178") @@ -3791,6 +4136,7 @@ class Program_weight_tensor_parameter_344: class Program_weight_tensor_parameter_345: name = "parameter_345" + original_name = "batch_norm_8.w_1" shape = [160] dtype = "float32" min_val = float("-0.687169") @@ -3802,6 +4148,7 @@ class Program_weight_tensor_parameter_345: class Program_weight_tensor_parameter_346: name = "parameter_346" + original_name = "conv2d_8.w_0" shape = [160, 80, 1, 1] dtype = "float32" min_val = float("-0.243774") @@ -3813,6 +4160,7 @@ class Program_weight_tensor_parameter_346: class Program_weight_tensor_parameter_347: name = "parameter_347" + original_name = "batch_norm_7.b_0" shape = [80] dtype = "float32" min_val = float("-2.93415") @@ -3824,6 +4172,7 @@ class Program_weight_tensor_parameter_347: class Program_weight_tensor_parameter_348: name = "parameter_348" + original_name = "batch_norm_7.w_0" shape = [80] dtype = "float32" min_val = float("-0.0866071") @@ -3835,6 +4184,7 @@ class Program_weight_tensor_parameter_348: class Program_weight_tensor_parameter_349: name = "parameter_349" + original_name = "batch_norm_7.w_2" shape = [80] dtype = "float32" min_val = float("0.000620872") @@ -3846,6 +4196,7 @@ class Program_weight_tensor_parameter_349: class Program_weight_tensor_parameter_350: name = "parameter_350" + original_name = "batch_norm_7.w_1" shape = [80] dtype = "float32" min_val = float("-0.388984") @@ -3857,6 +4208,7 @@ class Program_weight_tensor_parameter_350: class Program_weight_tensor_parameter_351: name = "parameter_351" + original_name = "conv2d_7.w_0" shape = [80, 1, 3, 3] dtype = "float32" min_val = float("-0.201205") @@ -3868,6 +4220,7 @@ class Program_weight_tensor_parameter_351: class Program_weight_tensor_parameter_352: name = "parameter_352" + original_name = "batch_norm_6.b_0" shape = [80] dtype = "float32" min_val = float("-1.6922") @@ -3879,6 +4232,7 @@ class Program_weight_tensor_parameter_352: class Program_weight_tensor_parameter_353: name = "parameter_353" + original_name = "batch_norm_6.w_0" shape = [80] dtype = "float32" min_val = float("0.474202") @@ -3890,6 +4244,7 @@ class Program_weight_tensor_parameter_353: class Program_weight_tensor_parameter_354: name = "parameter_354" + original_name = "batch_norm_6.w_2" shape = [80] dtype = "float32" min_val = float("0.0472127") @@ -3901,6 +4256,7 @@ class Program_weight_tensor_parameter_354: class Program_weight_tensor_parameter_355: name = "parameter_355" + original_name = "batch_norm_6.w_1" shape = [80] dtype = "float32" min_val = float("-1.87343") @@ -3912,6 +4268,7 @@ class Program_weight_tensor_parameter_355: class Program_weight_tensor_parameter_356: name = "parameter_356" + original_name = "conv2d_6.w_0" shape = [80, 192, 1, 1] dtype = "float32" min_val = float("-0.313394") @@ -3923,6 +4280,7 @@ class Program_weight_tensor_parameter_356: class Program_weight_tensor_parameter_357: name = "parameter_357" + original_name = "batch_norm_5.b_0" shape = [192] dtype = "float32" min_val = float("-1.13416") @@ -3934,6 +4292,7 @@ class Program_weight_tensor_parameter_357: class Program_weight_tensor_parameter_358: name = "parameter_358" + original_name = "batch_norm_5.w_0" shape = [192] dtype = "float32" min_val = float("0.509166") @@ -3945,6 +4304,7 @@ class Program_weight_tensor_parameter_358: class Program_weight_tensor_parameter_359: name = "parameter_359" + original_name = "batch_norm_5.w_2" shape = [192] dtype = "float32" min_val = float("2.20303e-05") @@ -3956,6 +4316,7 @@ class Program_weight_tensor_parameter_359: class Program_weight_tensor_parameter_360: name = "parameter_360" + original_name = "batch_norm_5.w_1" shape = [192] dtype = "float32" min_val = float("-0.994265") @@ -3967,6 +4328,7 @@ class Program_weight_tensor_parameter_360: class Program_weight_tensor_parameter_361: name = "parameter_361" + original_name = "conv2d_5.w_0" shape = [192, 1, 5, 5] dtype = "float32" min_val = float("-0.105825") @@ -3978,6 +4340,7 @@ class Program_weight_tensor_parameter_361: class Program_weight_tensor_parameter_362: name = "parameter_362" + original_name = "batch_norm_4.b_0" shape = [192] dtype = "float32" min_val = float("-1.2668") @@ -3989,6 +4352,7 @@ class Program_weight_tensor_parameter_362: class Program_weight_tensor_parameter_363: name = "parameter_363" + original_name = "batch_norm_4.w_0" shape = [192] dtype = "float32" min_val = float("0.729186") @@ -4000,6 +4364,7 @@ class Program_weight_tensor_parameter_363: class Program_weight_tensor_parameter_364: name = "parameter_364" + original_name = "batch_norm_4.w_2" shape = [192] dtype = "float32" min_val = float("0.0159128") @@ -4011,6 +4376,7 @@ class Program_weight_tensor_parameter_364: class Program_weight_tensor_parameter_365: name = "parameter_365" + original_name = "batch_norm_4.w_1" shape = [192] dtype = "float32" min_val = float("-0.747783") @@ -4022,6 +4388,7 @@ class Program_weight_tensor_parameter_365: class Program_weight_tensor_parameter_366: name = "parameter_366" + original_name = "conv2d_4.w_0" shape = [192, 48, 1, 1] dtype = "float32" min_val = float("-0.554988") @@ -4033,6 +4400,7 @@ class Program_weight_tensor_parameter_366: class Program_weight_tensor_parameter_367: name = "parameter_367" + original_name = "batch_norm_3.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -4042,6 +4410,7 @@ class Program_weight_tensor_parameter_367: class Program_weight_tensor_parameter_368: name = "parameter_368" + original_name = "batch_norm_3.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -4051,6 +4420,7 @@ class Program_weight_tensor_parameter_368: class Program_weight_tensor_parameter_369: name = "parameter_369" + original_name = "batch_norm_3.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -4060,6 +4430,7 @@ class Program_weight_tensor_parameter_369: class Program_weight_tensor_parameter_370: name = "parameter_370" + original_name = "batch_norm_3.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -4069,6 +4440,7 @@ class Program_weight_tensor_parameter_370: class Program_weight_tensor_parameter_371: name = "parameter_371" + original_name = "conv2d_3.w_0" shape = [48, 1, 3, 3] dtype = "float32" min_val = float("-0.229726") @@ -4080,6 +4452,7 @@ class Program_weight_tensor_parameter_371: class Program_weight_tensor_parameter_372: name = "parameter_372" + original_name = "batch_norm_2.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -4089,6 +4462,7 @@ class Program_weight_tensor_parameter_372: class Program_weight_tensor_parameter_373: name = "parameter_373" + original_name = "batch_norm_2.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -4098,6 +4472,7 @@ class Program_weight_tensor_parameter_373: class Program_weight_tensor_parameter_374: name = "parameter_374" + original_name = "batch_norm_2.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -4107,6 +4482,7 @@ class Program_weight_tensor_parameter_374: class Program_weight_tensor_parameter_375: name = "parameter_375" + original_name = "batch_norm_2.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -4116,6 +4492,7 @@ class Program_weight_tensor_parameter_375: class Program_weight_tensor_parameter_376: name = "parameter_376" + original_name = "conv2d_2.w_0" shape = [48, 128, 1, 1] dtype = "float32" min_val = float("-0.29441") @@ -4127,6 +4504,7 @@ class Program_weight_tensor_parameter_376: class Program_weight_tensor_parameter_377: name = "parameter_377" + original_name = "batch_norm_1.b_0" shape = [128] dtype = "float32" min_val = float("-6.15152") @@ -4138,6 +4516,7 @@ class Program_weight_tensor_parameter_377: class Program_weight_tensor_parameter_378: name = "parameter_378" + original_name = "batch_norm_1.w_0" shape = [128] dtype = "float32" min_val = float("0.483426") @@ -4149,6 +4528,7 @@ class Program_weight_tensor_parameter_378: class Program_weight_tensor_parameter_379: name = "parameter_379" + original_name = "batch_norm_1.w_2" shape = [128] dtype = "float32" min_val = float("0.0609025") @@ -4160,6 +4540,7 @@ class Program_weight_tensor_parameter_379: class Program_weight_tensor_parameter_380: name = "parameter_380" + original_name = "batch_norm_1.w_1" shape = [128] dtype = "float32" min_val = float("-2.51899") @@ -4171,6 +4552,7 @@ class Program_weight_tensor_parameter_380: class Program_weight_tensor_parameter_381: name = "parameter_381" + original_name = "conv2d_1.w_0" shape = [128, 32, 3, 3] dtype = "float32" min_val = float("-0.440861") @@ -4182,6 +4564,7 @@ class Program_weight_tensor_parameter_381: class Program_weight_tensor_parameter_382: name = "parameter_382" + original_name = "batch_norm_0.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -4191,6 +4574,7 @@ class Program_weight_tensor_parameter_382: class Program_weight_tensor_parameter_383: name = "parameter_383" + original_name = "batch_norm_0.w_0" shape = [32] dtype = "float32" min_val = float("0") @@ -4200,6 +4584,7 @@ class Program_weight_tensor_parameter_383: class Program_weight_tensor_parameter_384: name = "parameter_384" + original_name = "batch_norm_0.w_2" shape = [32] dtype = "float32" min_val = float("0") @@ -4209,6 +4594,7 @@ class Program_weight_tensor_parameter_384: class Program_weight_tensor_parameter_385: name = "parameter_385" + original_name = "batch_norm_0.w_1" shape = [32] dtype = "float32" min_val = float("0") @@ -4218,6 +4604,7 @@ class Program_weight_tensor_parameter_385: class Program_weight_tensor_parameter_386: name = "parameter_386" + original_name = "conv2d_0.w_0" shape = [32, 3, 3, 3] dtype = "float32" min_val = float("-0.100361") diff --git a/paddle_samples/PaddleX/MobileNetV4_hybrid_large/subgraph_1/input_meta.py b/paddle_samples/PaddleX/MobileNetV4_hybrid_large/subgraph_1/input_meta.py index 979393867..3fe30316c 100644 --- a/paddle_samples/PaddleX/MobileNetV4_hybrid_large/subgraph_1/input_meta.py +++ b/paddle_samples/PaddleX/MobileNetV4_hybrid_large/subgraph_1/input_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_data_0: name = "data_0" + original_name = "param_35" shape = [96] dtype = "float32" min_val = float("-1.72259") @@ -11,6 +12,7 @@ class Program_weight_tensor_data_0: class Program_weight_tensor_data_1: name = "data_1" + original_name = "param_36" shape = [96] dtype = "float32" min_val = float("-8.21448") @@ -22,6 +24,7 @@ class Program_weight_tensor_data_1: class Program_weight_tensor_data_2: name = "data_2" + original_name = "param_37" shape = [192] dtype = "float32" min_val = float("-2.20064") @@ -33,6 +36,7 @@ class Program_weight_tensor_data_2: class Program_weight_tensor_data_3: name = "data_3" + original_name = "param_38" shape = [192] dtype = "float32" min_val = float("-1.49745") @@ -44,6 +48,7 @@ class Program_weight_tensor_data_3: class Program_weight_tensor_data_4: name = "data_4" + original_name = "param_39" shape = [192] dtype = "float32" min_val = float("-1.56831") @@ -55,6 +60,7 @@ class Program_weight_tensor_data_4: class Program_weight_tensor_data_5: name = "data_5" + original_name = "param_40" shape = [192] dtype = "float32" min_val = float("-1.49819") @@ -66,6 +72,7 @@ class Program_weight_tensor_data_5: class Program_weight_tensor_data_6: name = "data_6" + original_name = "param_41" shape = [192] dtype = "float32" min_val = float("-1.51742") @@ -77,6 +84,7 @@ class Program_weight_tensor_data_6: class Program_weight_tensor_data_7: name = "data_7" + original_name = "param_42" shape = [192] dtype = "float32" min_val = float("-1.54471") @@ -88,6 +96,7 @@ class Program_weight_tensor_data_7: class Program_weight_tensor_data_8: name = "data_8" + original_name = "param_43" shape = [192] dtype = "float32" min_val = float("-1.63894") @@ -99,6 +108,7 @@ class Program_weight_tensor_data_8: class Program_weight_tensor_data_9: name = "data_9" + original_name = "param_44" shape = [192] dtype = "float32" min_val = float("-3.48394") @@ -110,6 +120,7 @@ class Program_weight_tensor_data_9: class Program_weight_tensor_data_10: name = "data_10" + original_name = "param_45" shape = [192] dtype = "float32" min_val = float("-1.25641") @@ -121,6 +132,7 @@ class Program_weight_tensor_data_10: class Program_weight_tensor_data_11: name = "data_11" + original_name = "param_46" shape = [192] dtype = "float32" min_val = float("-3.88505") @@ -132,6 +144,7 @@ class Program_weight_tensor_data_11: class Program_weight_tensor_data_12: name = "data_12" + original_name = "param_47" shape = [192] dtype = "float32" min_val = float("-1.84247") @@ -143,6 +156,7 @@ class Program_weight_tensor_data_12: class Program_weight_tensor_data_13: name = "data_13" + original_name = "param_48" shape = [192] dtype = "float32" min_val = float("-6.06403") @@ -154,6 +168,7 @@ class Program_weight_tensor_data_13: class Program_weight_tensor_data_14: name = "data_14" + original_name = "param_49" shape = [192] dtype = "float32" min_val = float("-3.3806") @@ -165,6 +180,7 @@ class Program_weight_tensor_data_14: class Program_weight_tensor_data_15: name = "data_15" + original_name = "param_50" shape = [192] dtype = "float32" min_val = float("-13.6566") @@ -176,6 +192,7 @@ class Program_weight_tensor_data_15: class Program_weight_tensor_data_16: name = "data_16" + original_name = "param_51" shape = [192] dtype = "float32" min_val = float("-7.55277") @@ -187,6 +204,7 @@ class Program_weight_tensor_data_16: class Program_weight_tensor_data_17: name = "data_17" + original_name = "param_52" shape = [512] dtype = "float32" min_val = float("-3.39935") @@ -198,6 +216,7 @@ class Program_weight_tensor_data_17: class Program_weight_tensor_data_18: name = "data_18" + original_name = "param_53" shape = [512] dtype = "float32" min_val = float("-1.62819") @@ -209,6 +228,7 @@ class Program_weight_tensor_data_18: class Program_weight_tensor_data_19: name = "data_19" + original_name = "param_54" shape = [512] dtype = "float32" min_val = float("-1.77072") @@ -220,6 +240,7 @@ class Program_weight_tensor_data_19: class Program_weight_tensor_data_20: name = "data_20" + original_name = "param_55" shape = [512] dtype = "float32" min_val = float("-1.10969") @@ -231,6 +252,7 @@ class Program_weight_tensor_data_20: class Program_weight_tensor_data_21: name = "data_21" + original_name = "param_56" shape = [512] dtype = "float32" min_val = float("-1.79568") @@ -242,6 +264,7 @@ class Program_weight_tensor_data_21: class Program_weight_tensor_data_22: name = "data_22" + original_name = "param_57" shape = [512] dtype = "float32" min_val = float("-1.64805") @@ -253,6 +276,7 @@ class Program_weight_tensor_data_22: class Program_weight_tensor_data_23: name = "data_23" + original_name = "param_58" shape = [512] dtype = "float32" min_val = float("-1.55701") @@ -264,6 +288,7 @@ class Program_weight_tensor_data_23: class Program_weight_tensor_data_24: name = "data_24" + original_name = "param_59" shape = [512] dtype = "float32" min_val = float("-1.38743") @@ -275,6 +300,7 @@ class Program_weight_tensor_data_24: class Program_weight_tensor_data_25: name = "data_25" + original_name = "param_60" shape = [512] dtype = "float32" min_val = float("-1.33237") @@ -286,6 +312,7 @@ class Program_weight_tensor_data_25: class Program_weight_tensor_data_26: name = "data_26" + original_name = "param_61" shape = [512] dtype = "float32" min_val = float("-1.81403") @@ -297,6 +324,7 @@ class Program_weight_tensor_data_26: class Program_weight_tensor_data_27: name = "data_27" + original_name = "param_62" shape = [512] dtype = "float32" min_val = float("-4.8125") @@ -308,6 +336,7 @@ class Program_weight_tensor_data_27: class Program_weight_tensor_data_28: name = "data_28" + original_name = "param_63" shape = [512] dtype = "float32" min_val = float("-1.31754") @@ -319,6 +348,7 @@ class Program_weight_tensor_data_28: class Program_weight_tensor_data_29: name = "data_29" + original_name = "param_64" shape = [512] dtype = "float32" min_val = float("-5.56266") @@ -330,6 +360,7 @@ class Program_weight_tensor_data_29: class Program_weight_tensor_data_30: name = "data_30" + original_name = "param_65" shape = [512] dtype = "float32" min_val = float("-2.15692") @@ -341,6 +372,7 @@ class Program_weight_tensor_data_30: class Program_weight_tensor_data_31: name = "data_31" + original_name = "param_66" shape = [512] dtype = "float32" min_val = float("-9.29469") @@ -352,6 +384,7 @@ class Program_weight_tensor_data_31: class Program_weight_tensor_data_32: name = "data_32" + original_name = "param_67" shape = [512] dtype = "float32" min_val = float("-1.26358") @@ -363,6 +396,7 @@ class Program_weight_tensor_data_32: class Program_weight_tensor_data_33: name = "data_33" + original_name = "param_68" shape = [512] dtype = "float32" min_val = float("-11.0993") @@ -374,6 +408,7 @@ class Program_weight_tensor_data_33: class Program_weight_tensor_data_34: name = "data_34" + original_name = "param_69" shape = [512] dtype = "float32" min_val = float("-3.45103") @@ -385,6 +420,7 @@ class Program_weight_tensor_data_34: class Program_weight_tensor_data_35: name = "data_35" + original_name = "var_805" shape = [28, 3, 384, 384] dtype = "float32" min_val = float("-4.7452") diff --git a/paddle_samples/PaddleX/MobileNetV4_hybrid_large/subgraph_1/weight_meta.py b/paddle_samples/PaddleX/MobileNetV4_hybrid_large/subgraph_1/weight_meta.py index dde800071..07768d69a 100644 --- a/paddle_samples/PaddleX/MobileNetV4_hybrid_large/subgraph_1/weight_meta.py +++ b/paddle_samples/PaddleX/MobileNetV4_hybrid_large/subgraph_1/weight_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_parameter_0: name = "parameter_0" + original_name = "linear_0.b_0" shape = [102] dtype = "float32" min_val = float("-0.357402") @@ -11,6 +12,7 @@ class Program_weight_tensor_parameter_0: class Program_weight_tensor_parameter_1: name = "parameter_1" + original_name = "linear_0.w_0" shape = [1280, 102] dtype = "float32" min_val = float("-0.49603") @@ -22,6 +24,7 @@ class Program_weight_tensor_parameter_1: class Program_weight_tensor_parameter_2: name = "parameter_2" + original_name = "batch_norm_120.b_0" shape = [1280] dtype = "float32" min_val = float("-0.454073") @@ -33,6 +36,7 @@ class Program_weight_tensor_parameter_2: class Program_weight_tensor_parameter_3: name = "parameter_3" + original_name = "batch_norm_120.w_0" shape = [1280] dtype = "float32" min_val = float("-0.177929") @@ -44,6 +48,7 @@ class Program_weight_tensor_parameter_3: class Program_weight_tensor_parameter_4: name = "parameter_4" + original_name = "batch_norm_120.w_2" shape = [1280] dtype = "float32" min_val = float("3.38979") @@ -55,6 +60,7 @@ class Program_weight_tensor_parameter_4: class Program_weight_tensor_parameter_5: name = "parameter_5" + original_name = "batch_norm_120.w_1" shape = [1280] dtype = "float32" min_val = float("-2.73006") @@ -66,6 +72,7 @@ class Program_weight_tensor_parameter_5: class Program_weight_tensor_parameter_6: name = "parameter_6" + original_name = "conv2d_144.w_0" shape = [1280, 960, 1, 1] dtype = "float32" min_val = float("-0.469764") @@ -77,6 +84,7 @@ class Program_weight_tensor_parameter_6: class Program_weight_tensor_parameter_7: name = "parameter_7" + original_name = "batch_norm_119.b_0" shape = [960] dtype = "float32" min_val = float("-6.13028") @@ -88,6 +96,7 @@ class Program_weight_tensor_parameter_7: class Program_weight_tensor_parameter_8: name = "parameter_8" + original_name = "batch_norm_119.w_0" shape = [960] dtype = "float32" min_val = float("0.0960158") @@ -99,6 +108,7 @@ class Program_weight_tensor_parameter_8: class Program_weight_tensor_parameter_9: name = "parameter_9" + original_name = "batch_norm_119.w_2" shape = [960] dtype = "float32" min_val = float("7709340.0") @@ -110,6 +120,7 @@ class Program_weight_tensor_parameter_9: class Program_weight_tensor_parameter_10: name = "parameter_10" + original_name = "batch_norm_119.w_1" shape = [960] dtype = "float32" min_val = float("-9314.34") @@ -121,6 +132,7 @@ class Program_weight_tensor_parameter_10: class Program_weight_tensor_parameter_11: name = "parameter_11" + original_name = "conv2d_143.w_0" shape = [960, 512, 1, 1] dtype = "float32" min_val = float("-0.383309") @@ -132,6 +144,7 @@ class Program_weight_tensor_parameter_11: class Program_weight_tensor_parameter_12: name = "parameter_12" + original_name = "batch_norm_118.b_0" shape = [512] dtype = "float32" min_val = float("-0.962456") @@ -143,6 +156,7 @@ class Program_weight_tensor_parameter_12: class Program_weight_tensor_parameter_13: name = "parameter_13" + original_name = "batch_norm_118.w_0" shape = [512] dtype = "float32" min_val = float("-3.59776") @@ -154,6 +168,7 @@ class Program_weight_tensor_parameter_13: class Program_weight_tensor_parameter_14: name = "parameter_14" + original_name = "batch_norm_118.w_2" shape = [512] dtype = "float32" min_val = float("0.674992") @@ -165,6 +180,7 @@ class Program_weight_tensor_parameter_14: class Program_weight_tensor_parameter_15: name = "parameter_15" + original_name = "batch_norm_118.w_1" shape = [512] dtype = "float32" min_val = float("-1.78075") @@ -176,6 +192,7 @@ class Program_weight_tensor_parameter_15: class Program_weight_tensor_parameter_16: name = "parameter_16" + original_name = "conv2d_142.w_0" shape = [512, 2048, 1, 1] dtype = "float32" min_val = float("-0.320195") @@ -187,6 +204,7 @@ class Program_weight_tensor_parameter_16: class Program_weight_tensor_parameter_17: name = "parameter_17" + original_name = "batch_norm_117.b_0" shape = [2048] dtype = "float32" min_val = float("-11.3964") @@ -198,6 +216,7 @@ class Program_weight_tensor_parameter_17: class Program_weight_tensor_parameter_18: name = "parameter_18" + original_name = "batch_norm_117.w_0" shape = [2048] dtype = "float32" min_val = float("-0.463229") @@ -209,6 +228,7 @@ class Program_weight_tensor_parameter_18: class Program_weight_tensor_parameter_19: name = "parameter_19" + original_name = "batch_norm_117.w_2" shape = [2048] dtype = "float32" min_val = float("0.829625") @@ -220,6 +240,7 @@ class Program_weight_tensor_parameter_19: class Program_weight_tensor_parameter_20: name = "parameter_20" + original_name = "batch_norm_117.w_1" shape = [2048] dtype = "float32" min_val = float("-3.40786") @@ -231,6 +252,7 @@ class Program_weight_tensor_parameter_20: class Program_weight_tensor_parameter_21: name = "parameter_21" + original_name = "conv2d_141.w_0" shape = [2048, 512, 1, 1] dtype = "float32" min_val = float("-0.410549") @@ -242,6 +264,7 @@ class Program_weight_tensor_parameter_21: class Program_weight_tensor_parameter_22: name = "parameter_22" + original_name = "batch_norm_116.b_0" shape = [512] dtype = "float32" min_val = float("-2.19714") @@ -253,6 +276,7 @@ class Program_weight_tensor_parameter_22: class Program_weight_tensor_parameter_23: name = "parameter_23" + original_name = "batch_norm_116.w_0" shape = [512] dtype = "float32" min_val = float("0.531586") @@ -264,6 +288,7 @@ class Program_weight_tensor_parameter_23: class Program_weight_tensor_parameter_24: name = "parameter_24" + original_name = "batch_norm_116.w_2" shape = [512] dtype = "float32" min_val = float("1586.77") @@ -275,6 +300,7 @@ class Program_weight_tensor_parameter_24: class Program_weight_tensor_parameter_25: name = "parameter_25" + original_name = "batch_norm_116.w_1" shape = [512] dtype = "float32" min_val = float("-3260.9") @@ -286,6 +312,7 @@ class Program_weight_tensor_parameter_25: class Program_weight_tensor_parameter_26: name = "parameter_26" + original_name = "conv2d_140.w_0" shape = [512, 1, 5, 5] dtype = "float32" min_val = float("-0.414156") @@ -297,6 +324,7 @@ class Program_weight_tensor_parameter_26: class Program_weight_tensor_parameter_27: name = "parameter_27" + original_name = "conv2d_139.w_0" shape = [512, 512, 1, 1] dtype = "float32" min_val = float("-0.584258") @@ -308,6 +336,7 @@ class Program_weight_tensor_parameter_27: class Program_weight_tensor_parameter_28: name = "parameter_28" + original_name = "conv2d_138.w_0" shape = [64, 512, 1, 1] dtype = "float32" min_val = float("-0.453555") @@ -319,6 +348,7 @@ class Program_weight_tensor_parameter_28: class Program_weight_tensor_parameter_29: name = "parameter_29" + original_name = "conv2d_137.w_0" shape = [64, 512, 1, 1] dtype = "float32" min_val = float("-0.540787") @@ -330,6 +360,7 @@ class Program_weight_tensor_parameter_29: class Program_weight_tensor_parameter_30: name = "parameter_30" + original_name = "conv2d_136.w_0" shape = [512, 512, 1, 1] dtype = "float32" min_val = float("-0.408499") @@ -341,6 +372,7 @@ class Program_weight_tensor_parameter_30: class Program_weight_tensor_parameter_31: name = "parameter_31" + original_name = "batch_norm_115.b_0" shape = [512] dtype = "float32" min_val = float("-3.75258") @@ -352,6 +384,7 @@ class Program_weight_tensor_parameter_31: class Program_weight_tensor_parameter_32: name = "parameter_32" + original_name = "batch_norm_115.w_0" shape = [512] dtype = "float32" min_val = float("-0.956415") @@ -363,6 +396,7 @@ class Program_weight_tensor_parameter_32: class Program_weight_tensor_parameter_33: name = "parameter_33" + original_name = "batch_norm_115.w_2" shape = [512] dtype = "float32" min_val = float("9260.88") @@ -374,6 +408,7 @@ class Program_weight_tensor_parameter_33: class Program_weight_tensor_parameter_34: name = "parameter_34" + original_name = "batch_norm_115.w_1" shape = [512] dtype = "float32" min_val = float("-1008.31") @@ -385,6 +420,7 @@ class Program_weight_tensor_parameter_34: class Program_weight_tensor_parameter_35: name = "parameter_35" + original_name = "batch_norm_114.b_0" shape = [512] dtype = "float32" min_val = float("-1.35796") @@ -396,6 +432,7 @@ class Program_weight_tensor_parameter_35: class Program_weight_tensor_parameter_36: name = "parameter_36" + original_name = "batch_norm_114.w_0" shape = [512] dtype = "float32" min_val = float("-1.44973") @@ -407,6 +444,7 @@ class Program_weight_tensor_parameter_36: class Program_weight_tensor_parameter_37: name = "parameter_37" + original_name = "batch_norm_114.w_2" shape = [512] dtype = "float32" min_val = float("0.666078") @@ -418,6 +456,7 @@ class Program_weight_tensor_parameter_37: class Program_weight_tensor_parameter_38: name = "parameter_38" + original_name = "batch_norm_114.w_1" shape = [512] dtype = "float32" min_val = float("-3.14527") @@ -429,6 +468,7 @@ class Program_weight_tensor_parameter_38: class Program_weight_tensor_parameter_39: name = "parameter_39" + original_name = "conv2d_135.w_0" shape = [512, 2048, 1, 1] dtype = "float32" min_val = float("-0.338959") @@ -440,6 +480,7 @@ class Program_weight_tensor_parameter_39: class Program_weight_tensor_parameter_40: name = "parameter_40" + original_name = "batch_norm_113.b_0" shape = [2048] dtype = "float32" min_val = float("-6.06472") @@ -451,6 +492,7 @@ class Program_weight_tensor_parameter_40: class Program_weight_tensor_parameter_41: name = "parameter_41" + original_name = "batch_norm_113.w_0" shape = [2048] dtype = "float32" min_val = float("-0.89828") @@ -462,6 +504,7 @@ class Program_weight_tensor_parameter_41: class Program_weight_tensor_parameter_42: name = "parameter_42" + original_name = "batch_norm_113.w_2" shape = [2048] dtype = "float32" min_val = float("0.94717") @@ -473,6 +516,7 @@ class Program_weight_tensor_parameter_42: class Program_weight_tensor_parameter_43: name = "parameter_43" + original_name = "batch_norm_113.w_1" shape = [2048] dtype = "float32" min_val = float("-3.91058") @@ -484,6 +528,7 @@ class Program_weight_tensor_parameter_43: class Program_weight_tensor_parameter_44: name = "parameter_44" + original_name = "conv2d_134.w_0" shape = [2048, 512, 1, 1] dtype = "float32" min_val = float("-0.62158") @@ -495,6 +540,7 @@ class Program_weight_tensor_parameter_44: class Program_weight_tensor_parameter_45: name = "parameter_45" + original_name = "batch_norm_112.b_0" shape = [512] dtype = "float32" min_val = float("-2.24567") @@ -506,6 +552,7 @@ class Program_weight_tensor_parameter_45: class Program_weight_tensor_parameter_46: name = "parameter_46" + original_name = "batch_norm_112.w_0" shape = [512] dtype = "float32" min_val = float("0.0959359") @@ -517,6 +564,7 @@ class Program_weight_tensor_parameter_46: class Program_weight_tensor_parameter_47: name = "parameter_47" + original_name = "batch_norm_112.w_2" shape = [512] dtype = "float32" min_val = float("706.006") @@ -528,6 +576,7 @@ class Program_weight_tensor_parameter_47: class Program_weight_tensor_parameter_48: name = "parameter_48" + original_name = "batch_norm_112.w_1" shape = [512] dtype = "float32" min_val = float("-1323.38") @@ -539,6 +588,7 @@ class Program_weight_tensor_parameter_48: class Program_weight_tensor_parameter_49: name = "parameter_49" + original_name = "conv2d_133.w_0" shape = [512, 1, 5, 5] dtype = "float32" min_val = float("-0.438111") @@ -550,6 +600,7 @@ class Program_weight_tensor_parameter_49: class Program_weight_tensor_parameter_50: name = "parameter_50" + original_name = "conv2d_132.w_0" shape = [512, 512, 1, 1] dtype = "float32" min_val = float("-0.506661") @@ -561,6 +612,7 @@ class Program_weight_tensor_parameter_50: class Program_weight_tensor_parameter_51: name = "parameter_51" + original_name = "conv2d_131.w_0" shape = [64, 512, 1, 1] dtype = "float32" min_val = float("-0.438629") @@ -572,6 +624,7 @@ class Program_weight_tensor_parameter_51: class Program_weight_tensor_parameter_52: name = "parameter_52" + original_name = "conv2d_130.w_0" shape = [64, 512, 1, 1] dtype = "float32" min_val = float("-0.414467") @@ -583,6 +636,7 @@ class Program_weight_tensor_parameter_52: class Program_weight_tensor_parameter_53: name = "parameter_53" + original_name = "conv2d_129.w_0" shape = [512, 512, 1, 1] dtype = "float32" min_val = float("-0.444289") @@ -594,6 +648,7 @@ class Program_weight_tensor_parameter_53: class Program_weight_tensor_parameter_54: name = "parameter_54" + original_name = "batch_norm_111.b_0" shape = [512] dtype = "float32" min_val = float("-2.17078") @@ -605,6 +660,7 @@ class Program_weight_tensor_parameter_54: class Program_weight_tensor_parameter_55: name = "parameter_55" + original_name = "batch_norm_111.w_0" shape = [512] dtype = "float32" min_val = float("-1.82933") @@ -616,6 +672,7 @@ class Program_weight_tensor_parameter_55: class Program_weight_tensor_parameter_56: name = "parameter_56" + original_name = "batch_norm_111.w_2" shape = [512] dtype = "float32" min_val = float("1295.85") @@ -627,6 +684,7 @@ class Program_weight_tensor_parameter_56: class Program_weight_tensor_parameter_57: name = "parameter_57" + original_name = "batch_norm_111.w_1" shape = [512] dtype = "float32" min_val = float("-920.86") @@ -638,6 +696,7 @@ class Program_weight_tensor_parameter_57: class Program_weight_tensor_parameter_58: name = "parameter_58" + original_name = "batch_norm_110.b_0" shape = [512] dtype = "float32" min_val = float("-1.60566") @@ -649,6 +708,7 @@ class Program_weight_tensor_parameter_58: class Program_weight_tensor_parameter_59: name = "parameter_59" + original_name = "batch_norm_110.w_0" shape = [512] dtype = "float32" min_val = float("-1.51442") @@ -660,6 +720,7 @@ class Program_weight_tensor_parameter_59: class Program_weight_tensor_parameter_60: name = "parameter_60" + original_name = "batch_norm_110.w_2" shape = [512] dtype = "float32" min_val = float("0.695805") @@ -671,6 +732,7 @@ class Program_weight_tensor_parameter_60: class Program_weight_tensor_parameter_61: name = "parameter_61" + original_name = "batch_norm_110.w_1" shape = [512] dtype = "float32" min_val = float("-2.62419") @@ -682,6 +744,7 @@ class Program_weight_tensor_parameter_61: class Program_weight_tensor_parameter_62: name = "parameter_62" + original_name = "conv2d_128.w_0" shape = [512, 2048, 1, 1] dtype = "float32" min_val = float("-0.410223") @@ -693,6 +756,7 @@ class Program_weight_tensor_parameter_62: class Program_weight_tensor_parameter_63: name = "parameter_63" + original_name = "batch_norm_109.b_0" shape = [2048] dtype = "float32" min_val = float("-5.61688") @@ -704,6 +768,7 @@ class Program_weight_tensor_parameter_63: class Program_weight_tensor_parameter_64: name = "parameter_64" + original_name = "batch_norm_109.w_0" shape = [2048] dtype = "float32" min_val = float("-0.607127") @@ -715,6 +780,7 @@ class Program_weight_tensor_parameter_64: class Program_weight_tensor_parameter_65: name = "parameter_65" + original_name = "batch_norm_109.w_2" shape = [2048] dtype = "float32" min_val = float("0.514599") @@ -726,6 +792,7 @@ class Program_weight_tensor_parameter_65: class Program_weight_tensor_parameter_66: name = "parameter_66" + original_name = "batch_norm_109.w_1" shape = [2048] dtype = "float32" min_val = float("-3.58126") @@ -737,6 +804,7 @@ class Program_weight_tensor_parameter_66: class Program_weight_tensor_parameter_67: name = "parameter_67" + original_name = "conv2d_127.w_0" shape = [2048, 512, 1, 1] dtype = "float32" min_val = float("-0.367865") @@ -748,6 +816,7 @@ class Program_weight_tensor_parameter_67: class Program_weight_tensor_parameter_68: name = "parameter_68" + original_name = "batch_norm_108.b_0" shape = [512] dtype = "float32" min_val = float("-2.75248") @@ -759,6 +828,7 @@ class Program_weight_tensor_parameter_68: class Program_weight_tensor_parameter_69: name = "parameter_69" + original_name = "batch_norm_108.w_0" shape = [512] dtype = "float32" min_val = float("-0.563789") @@ -770,6 +840,7 @@ class Program_weight_tensor_parameter_69: class Program_weight_tensor_parameter_70: name = "parameter_70" + original_name = "batch_norm_108.w_2" shape = [512] dtype = "float32" min_val = float("1661.79") @@ -781,6 +852,7 @@ class Program_weight_tensor_parameter_70: class Program_weight_tensor_parameter_71: name = "parameter_71" + original_name = "batch_norm_108.w_1" shape = [512] dtype = "float32" min_val = float("-1151.92") @@ -792,6 +864,7 @@ class Program_weight_tensor_parameter_71: class Program_weight_tensor_parameter_72: name = "parameter_72" + original_name = "conv2d_126.w_0" shape = [512, 1, 5, 5] dtype = "float32" min_val = float("-0.525116") @@ -803,6 +876,7 @@ class Program_weight_tensor_parameter_72: class Program_weight_tensor_parameter_73: name = "parameter_73" + original_name = "conv2d_125.w_0" shape = [512, 512, 1, 1] dtype = "float32" min_val = float("-0.475913") @@ -814,6 +888,7 @@ class Program_weight_tensor_parameter_73: class Program_weight_tensor_parameter_74: name = "parameter_74" + original_name = "conv2d_124.w_0" shape = [64, 512, 1, 1] dtype = "float32" min_val = float("-0.423448") @@ -825,6 +900,7 @@ class Program_weight_tensor_parameter_74: class Program_weight_tensor_parameter_75: name = "parameter_75" + original_name = "conv2d_123.w_0" shape = [64, 512, 1, 1] dtype = "float32" min_val = float("-0.398432") @@ -836,6 +912,7 @@ class Program_weight_tensor_parameter_75: class Program_weight_tensor_parameter_76: name = "parameter_76" + original_name = "conv2d_122.w_0" shape = [512, 512, 1, 1] dtype = "float32" min_val = float("-0.412834") @@ -847,6 +924,7 @@ class Program_weight_tensor_parameter_76: class Program_weight_tensor_parameter_77: name = "parameter_77" + original_name = "batch_norm_107.b_0" shape = [512] dtype = "float32" min_val = float("-1.12798") @@ -858,6 +936,7 @@ class Program_weight_tensor_parameter_77: class Program_weight_tensor_parameter_78: name = "parameter_78" + original_name = "batch_norm_107.w_0" shape = [512] dtype = "float32" min_val = float("-1.01369") @@ -869,6 +948,7 @@ class Program_weight_tensor_parameter_78: class Program_weight_tensor_parameter_79: name = "parameter_79" + original_name = "batch_norm_107.w_2" shape = [512] dtype = "float32" min_val = float("234.368") @@ -880,6 +960,7 @@ class Program_weight_tensor_parameter_79: class Program_weight_tensor_parameter_80: name = "parameter_80" + original_name = "batch_norm_107.w_1" shape = [512] dtype = "float32" min_val = float("-847.132") @@ -891,6 +972,7 @@ class Program_weight_tensor_parameter_80: class Program_weight_tensor_parameter_81: name = "parameter_81" + original_name = "batch_norm_106.b_0" shape = [512] dtype = "float32" min_val = float("-1.60609") @@ -902,6 +984,7 @@ class Program_weight_tensor_parameter_81: class Program_weight_tensor_parameter_82: name = "parameter_82" + original_name = "batch_norm_106.w_0" shape = [512] dtype = "float32" min_val = float("-1.37373") @@ -913,6 +996,7 @@ class Program_weight_tensor_parameter_82: class Program_weight_tensor_parameter_83: name = "parameter_83" + original_name = "batch_norm_106.w_2" shape = [512] dtype = "float32" min_val = float("1.54885") @@ -924,6 +1008,7 @@ class Program_weight_tensor_parameter_83: class Program_weight_tensor_parameter_84: name = "parameter_84" + original_name = "batch_norm_106.w_1" shape = [512] dtype = "float32" min_val = float("-3.70179") @@ -935,6 +1020,7 @@ class Program_weight_tensor_parameter_84: class Program_weight_tensor_parameter_85: name = "parameter_85" + original_name = "conv2d_121.w_0" shape = [512, 2048, 1, 1] dtype = "float32" min_val = float("-0.419649") @@ -946,6 +1032,7 @@ class Program_weight_tensor_parameter_85: class Program_weight_tensor_parameter_86: name = "parameter_86" + original_name = "batch_norm_105.b_0" shape = [2048] dtype = "float32" min_val = float("-5.85258") @@ -957,6 +1044,7 @@ class Program_weight_tensor_parameter_86: class Program_weight_tensor_parameter_87: name = "parameter_87" + original_name = "batch_norm_105.w_0" shape = [2048] dtype = "float32" min_val = float("-0.134552") @@ -968,6 +1056,7 @@ class Program_weight_tensor_parameter_87: class Program_weight_tensor_parameter_88: name = "parameter_88" + original_name = "batch_norm_105.w_2" shape = [2048] dtype = "float32" min_val = float("1.03467") @@ -979,6 +1068,7 @@ class Program_weight_tensor_parameter_88: class Program_weight_tensor_parameter_89: name = "parameter_89" + original_name = "batch_norm_105.w_1" shape = [2048] dtype = "float32" min_val = float("-4.03127") @@ -990,6 +1080,7 @@ class Program_weight_tensor_parameter_89: class Program_weight_tensor_parameter_90: name = "parameter_90" + original_name = "conv2d_120.w_0" shape = [2048, 512, 1, 1] dtype = "float32" min_val = float("-0.402273") @@ -1001,6 +1092,7 @@ class Program_weight_tensor_parameter_90: class Program_weight_tensor_parameter_91: name = "parameter_91" + original_name = "batch_norm_104.b_0" shape = [512] dtype = "float32" min_val = float("-2.41456") @@ -1012,6 +1104,7 @@ class Program_weight_tensor_parameter_91: class Program_weight_tensor_parameter_92: name = "parameter_92" + original_name = "batch_norm_104.w_0" shape = [512] dtype = "float32" min_val = float("-1.03844") @@ -1023,6 +1116,7 @@ class Program_weight_tensor_parameter_92: class Program_weight_tensor_parameter_93: name = "parameter_93" + original_name = "batch_norm_104.w_2" shape = [512] dtype = "float32" min_val = float("390.948") @@ -1034,6 +1128,7 @@ class Program_weight_tensor_parameter_93: class Program_weight_tensor_parameter_94: name = "parameter_94" + original_name = "batch_norm_104.w_1" shape = [512] dtype = "float32" min_val = float("-1210.98") @@ -1045,6 +1140,7 @@ class Program_weight_tensor_parameter_94: class Program_weight_tensor_parameter_95: name = "parameter_95" + original_name = "conv2d_119.w_0" shape = [512, 1, 5, 5] dtype = "float32" min_val = float("-0.410539") @@ -1056,6 +1152,7 @@ class Program_weight_tensor_parameter_95: class Program_weight_tensor_parameter_96: name = "parameter_96" + original_name = "conv2d_118.w_0" shape = [512, 512, 1, 1] dtype = "float32" min_val = float("-0.475912") @@ -1067,6 +1164,7 @@ class Program_weight_tensor_parameter_96: class Program_weight_tensor_parameter_97: name = "parameter_97" + original_name = "conv2d_117.w_0" shape = [64, 512, 1, 1] dtype = "float32" min_val = float("-0.457541") @@ -1078,6 +1176,7 @@ class Program_weight_tensor_parameter_97: class Program_weight_tensor_parameter_98: name = "parameter_98" + original_name = "conv2d_116.w_0" shape = [64, 512, 1, 1] dtype = "float32" min_val = float("-0.515426") @@ -1089,6 +1188,7 @@ class Program_weight_tensor_parameter_98: class Program_weight_tensor_parameter_99: name = "parameter_99" + original_name = "conv2d_115.w_0" shape = [512, 512, 1, 1] dtype = "float32" min_val = float("-0.469156") @@ -1100,6 +1200,7 @@ class Program_weight_tensor_parameter_99: class Program_weight_tensor_parameter_100: name = "parameter_100" + original_name = "batch_norm_103.b_0" shape = [512] dtype = "float32" min_val = float("-1.27434") @@ -1111,6 +1212,7 @@ class Program_weight_tensor_parameter_100: class Program_weight_tensor_parameter_101: name = "parameter_101" + original_name = "batch_norm_103.w_0" shape = [512] dtype = "float32" min_val = float("-1.06869") @@ -1122,6 +1224,7 @@ class Program_weight_tensor_parameter_101: class Program_weight_tensor_parameter_102: name = "parameter_102" + original_name = "batch_norm_103.w_2" shape = [512] dtype = "float32" min_val = float("0.0296135") @@ -1133,6 +1236,7 @@ class Program_weight_tensor_parameter_102: class Program_weight_tensor_parameter_103: name = "parameter_103" + original_name = "batch_norm_103.w_1" shape = [512] dtype = "float32" min_val = float("-5.84435") @@ -1144,6 +1248,7 @@ class Program_weight_tensor_parameter_103: class Program_weight_tensor_parameter_104: name = "parameter_104" + original_name = "batch_norm_102.b_0" shape = [512] dtype = "float32" min_val = float("-0.956299") @@ -1155,6 +1260,7 @@ class Program_weight_tensor_parameter_104: class Program_weight_tensor_parameter_105: name = "parameter_105" + original_name = "batch_norm_102.w_0" shape = [512] dtype = "float32" min_val = float("-1.96813") @@ -1166,6 +1272,7 @@ class Program_weight_tensor_parameter_105: class Program_weight_tensor_parameter_106: name = "parameter_106" + original_name = "batch_norm_102.w_2" shape = [512] dtype = "float32" min_val = float("4.46025") @@ -1177,6 +1284,7 @@ class Program_weight_tensor_parameter_106: class Program_weight_tensor_parameter_107: name = "parameter_107" + original_name = "batch_norm_102.w_1" shape = [512] dtype = "float32" min_val = float("-11.4624") @@ -1188,6 +1296,7 @@ class Program_weight_tensor_parameter_107: class Program_weight_tensor_parameter_108: name = "parameter_108" + original_name = "conv2d_114.w_0" shape = [512, 2048, 1, 1] dtype = "float32" min_val = float("-0.455752") @@ -1199,6 +1308,7 @@ class Program_weight_tensor_parameter_108: class Program_weight_tensor_parameter_109: name = "parameter_109" + original_name = "batch_norm_101.b_0" shape = [2048] dtype = "float32" min_val = float("-5.50222") @@ -1210,6 +1320,7 @@ class Program_weight_tensor_parameter_109: class Program_weight_tensor_parameter_110: name = "parameter_110" + original_name = "batch_norm_101.w_0" shape = [2048] dtype = "float32" min_val = float("-1.39695") @@ -1221,6 +1332,7 @@ class Program_weight_tensor_parameter_110: class Program_weight_tensor_parameter_111: name = "parameter_111" + original_name = "batch_norm_101.w_2" shape = [2048] dtype = "float32" min_val = float("1.44334e-43") @@ -1232,6 +1344,7 @@ class Program_weight_tensor_parameter_111: class Program_weight_tensor_parameter_112: name = "parameter_112" + original_name = "batch_norm_101.w_1" shape = [2048] dtype = "float32" min_val = float("-2.83872") @@ -1243,6 +1356,7 @@ class Program_weight_tensor_parameter_112: class Program_weight_tensor_parameter_113: name = "parameter_113" + original_name = "conv2d_113.w_0" shape = [2048, 1, 5, 5] dtype = "float32" min_val = float("-0.486613") @@ -1254,6 +1368,7 @@ class Program_weight_tensor_parameter_113: class Program_weight_tensor_parameter_114: name = "parameter_114" + original_name = "batch_norm_100.b_0" shape = [2048] dtype = "float32" min_val = float("-3.62479") @@ -1265,6 +1380,7 @@ class Program_weight_tensor_parameter_114: class Program_weight_tensor_parameter_115: name = "parameter_115" + original_name = "batch_norm_100.w_0" shape = [2048] dtype = "float32" min_val = float("-1.17727") @@ -1276,6 +1392,7 @@ class Program_weight_tensor_parameter_115: class Program_weight_tensor_parameter_116: name = "parameter_116" + original_name = "batch_norm_100.w_2" shape = [2048] dtype = "float32" min_val = float("1.59748e-43") @@ -1287,6 +1404,7 @@ class Program_weight_tensor_parameter_116: class Program_weight_tensor_parameter_117: name = "parameter_117" + original_name = "batch_norm_100.w_1" shape = [2048] dtype = "float32" min_val = float("-5.98671") @@ -1298,6 +1416,7 @@ class Program_weight_tensor_parameter_117: class Program_weight_tensor_parameter_118: name = "parameter_118" + original_name = "conv2d_112.w_0" shape = [2048, 512, 1, 1] dtype = "float32" min_val = float("-0.464848") @@ -1309,6 +1428,7 @@ class Program_weight_tensor_parameter_118: class Program_weight_tensor_parameter_119: name = "parameter_119" + original_name = "batch_norm_99.b_0" shape = [512] dtype = "float32" min_val = float("-1.88624") @@ -1320,6 +1440,7 @@ class Program_weight_tensor_parameter_119: class Program_weight_tensor_parameter_120: name = "parameter_120" + original_name = "batch_norm_99.w_0" shape = [512] dtype = "float32" min_val = float("0.313269") @@ -1331,6 +1452,7 @@ class Program_weight_tensor_parameter_120: class Program_weight_tensor_parameter_121: name = "parameter_121" + original_name = "batch_norm_99.w_2" shape = [512] dtype = "float32" min_val = float("0.0221007") @@ -1342,6 +1464,7 @@ class Program_weight_tensor_parameter_121: class Program_weight_tensor_parameter_122: name = "parameter_122" + original_name = "batch_norm_99.w_1" shape = [512] dtype = "float32" min_val = float("-9.8854") @@ -1353,6 +1476,7 @@ class Program_weight_tensor_parameter_122: class Program_weight_tensor_parameter_123: name = "parameter_123" + original_name = "conv2d_111.w_0" shape = [512, 1, 5, 5] dtype = "float32" min_val = float("-0.488226") @@ -1364,6 +1488,7 @@ class Program_weight_tensor_parameter_123: class Program_weight_tensor_parameter_124: name = "parameter_124" + original_name = "batch_norm_98.b_0" shape = [512] dtype = "float32" min_val = float("-1.12915") @@ -1375,6 +1500,7 @@ class Program_weight_tensor_parameter_124: class Program_weight_tensor_parameter_125: name = "parameter_125" + original_name = "batch_norm_98.w_0" shape = [512] dtype = "float32" min_val = float("-1.52448") @@ -1386,6 +1512,7 @@ class Program_weight_tensor_parameter_125: class Program_weight_tensor_parameter_126: name = "parameter_126" + original_name = "batch_norm_98.w_2" shape = [512] dtype = "float32" min_val = float("7.73765") @@ -1397,6 +1524,7 @@ class Program_weight_tensor_parameter_126: class Program_weight_tensor_parameter_127: name = "parameter_127" + original_name = "batch_norm_98.w_1" shape = [512] dtype = "float32" min_val = float("-19.7109") @@ -1408,6 +1536,7 @@ class Program_weight_tensor_parameter_127: class Program_weight_tensor_parameter_128: name = "parameter_128" + original_name = "conv2d_110.w_0" shape = [512, 2048, 1, 1] dtype = "float32" min_val = float("-0.484056") @@ -1419,6 +1548,7 @@ class Program_weight_tensor_parameter_128: class Program_weight_tensor_parameter_129: name = "parameter_129" + original_name = "batch_norm_97.b_0" shape = [2048] dtype = "float32" min_val = float("-5.02962") @@ -1430,6 +1560,7 @@ class Program_weight_tensor_parameter_129: class Program_weight_tensor_parameter_130: name = "parameter_130" + original_name = "batch_norm_97.w_0" shape = [2048] dtype = "float32" min_val = float("-1.45022") @@ -1441,6 +1572,7 @@ class Program_weight_tensor_parameter_130: class Program_weight_tensor_parameter_131: name = "parameter_131" + original_name = "batch_norm_97.w_2" shape = [2048] dtype = "float32" min_val = float("1.5134e-43") @@ -1452,6 +1584,7 @@ class Program_weight_tensor_parameter_131: class Program_weight_tensor_parameter_132: name = "parameter_132" + original_name = "batch_norm_97.w_1" shape = [2048] dtype = "float32" min_val = float("-1.83844") @@ -1463,6 +1596,7 @@ class Program_weight_tensor_parameter_132: class Program_weight_tensor_parameter_133: name = "parameter_133" + original_name = "conv2d_109.w_0" shape = [2048, 1, 3, 3] dtype = "float32" min_val = float("-0.433561") @@ -1474,6 +1608,7 @@ class Program_weight_tensor_parameter_133: class Program_weight_tensor_parameter_134: name = "parameter_134" + original_name = "batch_norm_96.b_0" shape = [2048] dtype = "float32" min_val = float("-3.64825") @@ -1485,6 +1620,7 @@ class Program_weight_tensor_parameter_134: class Program_weight_tensor_parameter_135: name = "parameter_135" + original_name = "batch_norm_96.w_0" shape = [2048] dtype = "float32" min_val = float("-0.547827") @@ -1496,6 +1632,7 @@ class Program_weight_tensor_parameter_135: class Program_weight_tensor_parameter_136: name = "parameter_136" + original_name = "batch_norm_96.w_2" shape = [2048] dtype = "float32" min_val = float("1.33123e-43") @@ -1507,6 +1644,7 @@ class Program_weight_tensor_parameter_136: class Program_weight_tensor_parameter_137: name = "parameter_137" + original_name = "batch_norm_96.w_1" shape = [2048] dtype = "float32" min_val = float("-4.02615") @@ -1518,6 +1656,7 @@ class Program_weight_tensor_parameter_137: class Program_weight_tensor_parameter_138: name = "parameter_138" + original_name = "conv2d_108.w_0" shape = [2048, 512, 1, 1] dtype = "float32" min_val = float("-0.49392") @@ -1529,6 +1668,7 @@ class Program_weight_tensor_parameter_138: class Program_weight_tensor_parameter_139: name = "parameter_139" + original_name = "batch_norm_95.b_0" shape = [512] dtype = "float32" min_val = float("-2.16753") @@ -1540,6 +1680,7 @@ class Program_weight_tensor_parameter_139: class Program_weight_tensor_parameter_140: name = "parameter_140" + original_name = "batch_norm_95.w_0" shape = [512] dtype = "float32" min_val = float("-1.49678") @@ -1551,6 +1692,7 @@ class Program_weight_tensor_parameter_140: class Program_weight_tensor_parameter_141: name = "parameter_141" + original_name = "batch_norm_95.w_2" shape = [512] dtype = "float32" min_val = float("0.00641504") @@ -1562,6 +1704,7 @@ class Program_weight_tensor_parameter_141: class Program_weight_tensor_parameter_142: name = "parameter_142" + original_name = "batch_norm_95.w_1" shape = [512] dtype = "float32" min_val = float("-10.4433") @@ -1573,6 +1716,7 @@ class Program_weight_tensor_parameter_142: class Program_weight_tensor_parameter_143: name = "parameter_143" + original_name = "conv2d_107.w_0" shape = [512, 1, 5, 5] dtype = "float32" min_val = float("-0.471907") @@ -1584,6 +1728,7 @@ class Program_weight_tensor_parameter_143: class Program_weight_tensor_parameter_144: name = "parameter_144" + original_name = "batch_norm_94.b_0" shape = [512] dtype = "float32" min_val = float("-1.4023") @@ -1595,6 +1740,7 @@ class Program_weight_tensor_parameter_144: class Program_weight_tensor_parameter_145: name = "parameter_145" + original_name = "batch_norm_94.w_0" shape = [512] dtype = "float32" min_val = float("-1.66582") @@ -1606,6 +1752,7 @@ class Program_weight_tensor_parameter_145: class Program_weight_tensor_parameter_146: name = "parameter_146" + original_name = "batch_norm_94.w_2" shape = [512] dtype = "float32" min_val = float("19.9177") @@ -1617,6 +1764,7 @@ class Program_weight_tensor_parameter_146: class Program_weight_tensor_parameter_147: name = "parameter_147" + original_name = "batch_norm_94.w_1" shape = [512] dtype = "float32" min_val = float("-3.82533") @@ -1628,6 +1776,7 @@ class Program_weight_tensor_parameter_147: class Program_weight_tensor_parameter_148: name = "parameter_148" + original_name = "conv2d_106.w_0" shape = [512, 2048, 1, 1] dtype = "float32" min_val = float("-0.478036") @@ -1639,6 +1788,7 @@ class Program_weight_tensor_parameter_148: class Program_weight_tensor_parameter_149: name = "parameter_149" + original_name = "batch_norm_93.b_0" shape = [2048] dtype = "float32" min_val = float("-5.38028") @@ -1650,6 +1800,7 @@ class Program_weight_tensor_parameter_149: class Program_weight_tensor_parameter_150: name = "parameter_150" + original_name = "batch_norm_93.w_0" shape = [2048] dtype = "float32" min_val = float("0.343269") @@ -1661,6 +1812,7 @@ class Program_weight_tensor_parameter_150: class Program_weight_tensor_parameter_151: name = "parameter_151" + original_name = "batch_norm_93.w_2" shape = [2048] dtype = "float32" min_val = float("2.34957") @@ -1672,6 +1824,7 @@ class Program_weight_tensor_parameter_151: class Program_weight_tensor_parameter_152: name = "parameter_152" + original_name = "batch_norm_93.w_1" shape = [2048] dtype = "float32" min_val = float("-4.5236") @@ -1683,6 +1836,7 @@ class Program_weight_tensor_parameter_152: class Program_weight_tensor_parameter_153: name = "parameter_153" + original_name = "conv2d_105.w_0" shape = [2048, 512, 1, 1] dtype = "float32" min_val = float("-0.433067") @@ -1694,6 +1848,7 @@ class Program_weight_tensor_parameter_153: class Program_weight_tensor_parameter_154: name = "parameter_154" + original_name = "batch_norm_92.b_0" shape = [512] dtype = "float32" min_val = float("-2.34115") @@ -1705,6 +1860,7 @@ class Program_weight_tensor_parameter_154: class Program_weight_tensor_parameter_155: name = "parameter_155" + original_name = "batch_norm_92.w_0" shape = [512] dtype = "float32" min_val = float("-1.02394") @@ -1716,6 +1872,7 @@ class Program_weight_tensor_parameter_155: class Program_weight_tensor_parameter_156: name = "parameter_156" + original_name = "batch_norm_92.w_2" shape = [512] dtype = "float32" min_val = float("0.00962626") @@ -1727,6 +1884,7 @@ class Program_weight_tensor_parameter_156: class Program_weight_tensor_parameter_157: name = "parameter_157" + original_name = "batch_norm_92.w_1" shape = [512] dtype = "float32" min_val = float("-5.61862") @@ -1738,6 +1896,7 @@ class Program_weight_tensor_parameter_157: class Program_weight_tensor_parameter_158: name = "parameter_158" + original_name = "conv2d_104.w_0" shape = [512, 1, 5, 5] dtype = "float32" min_val = float("-0.42638") @@ -1749,6 +1908,7 @@ class Program_weight_tensor_parameter_158: class Program_weight_tensor_parameter_159: name = "parameter_159" + original_name = "batch_norm_91.b_0" shape = [512] dtype = "float32" min_val = float("-1.00955") @@ -1760,6 +1920,7 @@ class Program_weight_tensor_parameter_159: class Program_weight_tensor_parameter_160: name = "parameter_160" + original_name = "batch_norm_91.w_0" shape = [512] dtype = "float32" min_val = float("-1.76475") @@ -1771,6 +1932,7 @@ class Program_weight_tensor_parameter_160: class Program_weight_tensor_parameter_161: name = "parameter_161" + original_name = "batch_norm_91.w_2" shape = [512] dtype = "float32" min_val = float("14.7394") @@ -1782,6 +1944,7 @@ class Program_weight_tensor_parameter_161: class Program_weight_tensor_parameter_162: name = "parameter_162" + original_name = "batch_norm_91.w_1" shape = [512] dtype = "float32" min_val = float("-3.29156") @@ -1793,6 +1956,7 @@ class Program_weight_tensor_parameter_162: class Program_weight_tensor_parameter_163: name = "parameter_163" + original_name = "conv2d_103.w_0" shape = [512, 2048, 1, 1] dtype = "float32" min_val = float("-0.454209") @@ -1804,6 +1968,7 @@ class Program_weight_tensor_parameter_163: class Program_weight_tensor_parameter_164: name = "parameter_164" + original_name = "batch_norm_90.b_0" shape = [2048] dtype = "float32" min_val = float("-5.41036") @@ -1815,6 +1980,7 @@ class Program_weight_tensor_parameter_164: class Program_weight_tensor_parameter_165: name = "parameter_165" + original_name = "batch_norm_90.w_0" shape = [2048] dtype = "float32" min_val = float("0.423083") @@ -1826,6 +1992,7 @@ class Program_weight_tensor_parameter_165: class Program_weight_tensor_parameter_166: name = "parameter_166" + original_name = "batch_norm_90.w_2" shape = [2048] dtype = "float32" min_val = float("2.30694") @@ -1837,6 +2004,7 @@ class Program_weight_tensor_parameter_166: class Program_weight_tensor_parameter_167: name = "parameter_167" + original_name = "batch_norm_90.w_1" shape = [2048] dtype = "float32" min_val = float("-4.62402") @@ -1848,6 +2016,7 @@ class Program_weight_tensor_parameter_167: class Program_weight_tensor_parameter_168: name = "parameter_168" + original_name = "conv2d_102.w_0" shape = [2048, 512, 1, 1] dtype = "float32" min_val = float("-0.423781") @@ -1859,6 +2028,7 @@ class Program_weight_tensor_parameter_168: class Program_weight_tensor_parameter_169: name = "parameter_169" + original_name = "batch_norm_89.b_0" shape = [512] dtype = "float32" min_val = float("-2.39046") @@ -1870,6 +2040,7 @@ class Program_weight_tensor_parameter_169: class Program_weight_tensor_parameter_170: name = "parameter_170" + original_name = "batch_norm_89.w_0" shape = [512] dtype = "float32" min_val = float("-1.26714") @@ -1881,6 +2052,7 @@ class Program_weight_tensor_parameter_170: class Program_weight_tensor_parameter_171: name = "parameter_171" + original_name = "batch_norm_89.w_2" shape = [512] dtype = "float32" min_val = float("0.00394058") @@ -1892,6 +2064,7 @@ class Program_weight_tensor_parameter_171: class Program_weight_tensor_parameter_172: name = "parameter_172" + original_name = "batch_norm_89.w_1" shape = [512] dtype = "float32" min_val = float("-8.20809") @@ -1903,6 +2076,7 @@ class Program_weight_tensor_parameter_172: class Program_weight_tensor_parameter_173: name = "parameter_173" + original_name = "conv2d_101.w_0" shape = [512, 1, 5, 5] dtype = "float32" min_val = float("-0.456644") @@ -1914,6 +2088,7 @@ class Program_weight_tensor_parameter_173: class Program_weight_tensor_parameter_174: name = "parameter_174" + original_name = "batch_norm_88.b_0" shape = [512] dtype = "float32" min_val = float("-1.33591") @@ -1925,6 +2100,7 @@ class Program_weight_tensor_parameter_174: class Program_weight_tensor_parameter_175: name = "parameter_175" + original_name = "batch_norm_88.w_0" shape = [512] dtype = "float32" min_val = float("-1.3714") @@ -1936,6 +2112,7 @@ class Program_weight_tensor_parameter_175: class Program_weight_tensor_parameter_176: name = "parameter_176" + original_name = "batch_norm_88.w_2" shape = [512] dtype = "float32" min_val = float("20.3164") @@ -1947,6 +2124,7 @@ class Program_weight_tensor_parameter_176: class Program_weight_tensor_parameter_177: name = "parameter_177" + original_name = "batch_norm_88.w_1" shape = [512] dtype = "float32" min_val = float("-42.9411") @@ -1958,6 +2136,7 @@ class Program_weight_tensor_parameter_177: class Program_weight_tensor_parameter_178: name = "parameter_178" + original_name = "conv2d_100.w_0" shape = [512, 2048, 1, 1] dtype = "float32" min_val = float("-0.507818") @@ -1969,6 +2148,7 @@ class Program_weight_tensor_parameter_178: class Program_weight_tensor_parameter_179: name = "parameter_179" + original_name = "batch_norm_87.b_0" shape = [2048] dtype = "float32" min_val = float("-4.96889") @@ -1980,6 +2160,7 @@ class Program_weight_tensor_parameter_179: class Program_weight_tensor_parameter_180: name = "parameter_180" + original_name = "batch_norm_87.w_0" shape = [2048] dtype = "float32" min_val = float("-1.40836") @@ -1991,6 +2172,7 @@ class Program_weight_tensor_parameter_180: class Program_weight_tensor_parameter_181: name = "parameter_181" + original_name = "batch_norm_87.w_2" shape = [2048] dtype = "float32" min_val = float("4.25686e-36") @@ -2002,6 +2184,7 @@ class Program_weight_tensor_parameter_181: class Program_weight_tensor_parameter_182: name = "parameter_182" + original_name = "batch_norm_87.w_1" shape = [2048] dtype = "float32" min_val = float("-1.40129") @@ -2013,6 +2196,7 @@ class Program_weight_tensor_parameter_182: class Program_weight_tensor_parameter_183: name = "parameter_183" + original_name = "conv2d_99.w_0" shape = [2048, 1, 3, 3] dtype = "float32" min_val = float("-0.387825") @@ -2024,6 +2208,7 @@ class Program_weight_tensor_parameter_183: class Program_weight_tensor_parameter_184: name = "parameter_184" + original_name = "batch_norm_86.b_0" shape = [2048] dtype = "float32" min_val = float("-3.94359") @@ -2035,6 +2220,7 @@ class Program_weight_tensor_parameter_184: class Program_weight_tensor_parameter_185: name = "parameter_185" + original_name = "batch_norm_86.w_0" shape = [2048] dtype = "float32" min_val = float("-0.664943") @@ -2046,6 +2232,7 @@ class Program_weight_tensor_parameter_185: class Program_weight_tensor_parameter_186: name = "parameter_186" + original_name = "batch_norm_86.w_2" shape = [2048] dtype = "float32" min_val = float("8.17788e-31") @@ -2057,6 +2244,7 @@ class Program_weight_tensor_parameter_186: class Program_weight_tensor_parameter_187: name = "parameter_187" + original_name = "batch_norm_86.w_1" shape = [2048] dtype = "float32" min_val = float("-6.27469") @@ -2068,6 +2256,7 @@ class Program_weight_tensor_parameter_187: class Program_weight_tensor_parameter_188: name = "parameter_188" + original_name = "conv2d_98.w_0" shape = [2048, 512, 1, 1] dtype = "float32" min_val = float("-0.444094") @@ -2079,6 +2268,7 @@ class Program_weight_tensor_parameter_188: class Program_weight_tensor_parameter_189: name = "parameter_189" + original_name = "batch_norm_85.b_0" shape = [512] dtype = "float32" min_val = float("-2.32427") @@ -2090,6 +2280,7 @@ class Program_weight_tensor_parameter_189: class Program_weight_tensor_parameter_190: name = "parameter_190" + original_name = "batch_norm_85.w_0" shape = [512] dtype = "float32" min_val = float("0.0946403") @@ -2101,6 +2292,7 @@ class Program_weight_tensor_parameter_190: class Program_weight_tensor_parameter_191: name = "parameter_191" + original_name = "batch_norm_85.w_2" shape = [512] dtype = "float32" min_val = float("0.00902463") @@ -2112,6 +2304,7 @@ class Program_weight_tensor_parameter_191: class Program_weight_tensor_parameter_192: name = "parameter_192" + original_name = "batch_norm_85.w_1" shape = [512] dtype = "float32" min_val = float("-6.14905") @@ -2123,6 +2316,7 @@ class Program_weight_tensor_parameter_192: class Program_weight_tensor_parameter_193: name = "parameter_193" + original_name = "conv2d_97.w_0" shape = [512, 1, 5, 5] dtype = "float32" min_val = float("-0.491483") @@ -2134,6 +2328,7 @@ class Program_weight_tensor_parameter_193: class Program_weight_tensor_parameter_194: name = "parameter_194" + original_name = "batch_norm_84.b_0" shape = [512] dtype = "float32" min_val = float("-1.22936") @@ -2145,6 +2340,7 @@ class Program_weight_tensor_parameter_194: class Program_weight_tensor_parameter_195: name = "parameter_195" + original_name = "batch_norm_84.w_0" shape = [512] dtype = "float32" min_val = float("-1.5094") @@ -2156,6 +2352,7 @@ class Program_weight_tensor_parameter_195: class Program_weight_tensor_parameter_196: name = "parameter_196" + original_name = "batch_norm_84.w_2" shape = [512] dtype = "float32" min_val = float("7.33899") @@ -2167,6 +2364,7 @@ class Program_weight_tensor_parameter_196: class Program_weight_tensor_parameter_197: name = "parameter_197" + original_name = "batch_norm_84.w_1" shape = [512] dtype = "float32" min_val = float("-3.70676") @@ -2178,6 +2376,7 @@ class Program_weight_tensor_parameter_197: class Program_weight_tensor_parameter_198: name = "parameter_198" + original_name = "conv2d_96.w_0" shape = [512, 2048, 1, 1] dtype = "float32" min_val = float("-0.426828") @@ -2189,6 +2388,7 @@ class Program_weight_tensor_parameter_198: class Program_weight_tensor_parameter_199: name = "parameter_199" + original_name = "batch_norm_83.b_0" shape = [2048] dtype = "float32" min_val = float("-4.8757") @@ -2200,6 +2400,7 @@ class Program_weight_tensor_parameter_199: class Program_weight_tensor_parameter_200: name = "parameter_200" + original_name = "batch_norm_83.w_0" shape = [2048] dtype = "float32" min_val = float("0.387679") @@ -2211,6 +2412,7 @@ class Program_weight_tensor_parameter_200: class Program_weight_tensor_parameter_201: name = "parameter_201" + original_name = "batch_norm_83.w_2" shape = [2048] dtype = "float32" min_val = float("1.99863") @@ -2222,6 +2424,7 @@ class Program_weight_tensor_parameter_201: class Program_weight_tensor_parameter_202: name = "parameter_202" + original_name = "batch_norm_83.w_1" shape = [2048] dtype = "float32" min_val = float("-5.80633") @@ -2233,6 +2436,7 @@ class Program_weight_tensor_parameter_202: class Program_weight_tensor_parameter_203: name = "parameter_203" + original_name = "conv2d_95.w_0" shape = [2048, 512, 1, 1] dtype = "float32" min_val = float("-0.438598") @@ -2244,6 +2448,7 @@ class Program_weight_tensor_parameter_203: class Program_weight_tensor_parameter_204: name = "parameter_204" + original_name = "batch_norm_82.b_0" shape = [512] dtype = "float32" min_val = float("-2.26852") @@ -2255,6 +2460,7 @@ class Program_weight_tensor_parameter_204: class Program_weight_tensor_parameter_205: name = "parameter_205" + original_name = "batch_norm_82.w_0" shape = [512] dtype = "float32" min_val = float("-0.117638") @@ -2266,6 +2472,7 @@ class Program_weight_tensor_parameter_205: class Program_weight_tensor_parameter_206: name = "parameter_206" + original_name = "batch_norm_82.w_2" shape = [512] dtype = "float32" min_val = float("0.00630864") @@ -2277,6 +2484,7 @@ class Program_weight_tensor_parameter_206: class Program_weight_tensor_parameter_207: name = "parameter_207" + original_name = "batch_norm_82.w_1" shape = [512] dtype = "float32" min_val = float("-5.25685") @@ -2288,6 +2496,7 @@ class Program_weight_tensor_parameter_207: class Program_weight_tensor_parameter_208: name = "parameter_208" + original_name = "conv2d_94.w_0" shape = [512, 1, 5, 5] dtype = "float32" min_val = float("-0.49947") @@ -2299,6 +2508,7 @@ class Program_weight_tensor_parameter_208: class Program_weight_tensor_parameter_209: name = "parameter_209" + original_name = "batch_norm_81.b_0" shape = [512] dtype = "float32" min_val = float("-1.06884") @@ -2310,6 +2520,7 @@ class Program_weight_tensor_parameter_209: class Program_weight_tensor_parameter_210: name = "parameter_210" + original_name = "batch_norm_81.w_0" shape = [512] dtype = "float32" min_val = float("-1.19829") @@ -2321,6 +2532,7 @@ class Program_weight_tensor_parameter_210: class Program_weight_tensor_parameter_211: name = "parameter_211" + original_name = "batch_norm_81.w_2" shape = [512] dtype = "float32" min_val = float("7.89059") @@ -2332,6 +2544,7 @@ class Program_weight_tensor_parameter_211: class Program_weight_tensor_parameter_212: name = "parameter_212" + original_name = "batch_norm_81.w_1" shape = [512] dtype = "float32" min_val = float("-20.1616") @@ -2343,6 +2556,7 @@ class Program_weight_tensor_parameter_212: class Program_weight_tensor_parameter_213: name = "parameter_213" + original_name = "conv2d_93.w_0" shape = [512, 2048, 1, 1] dtype = "float32" min_val = float("-0.466889") @@ -2354,6 +2568,7 @@ class Program_weight_tensor_parameter_213: class Program_weight_tensor_parameter_214: name = "parameter_214" + original_name = "batch_norm_80.b_0" shape = [2048] dtype = "float32" min_val = float("-7.40577") @@ -2365,6 +2580,7 @@ class Program_weight_tensor_parameter_214: class Program_weight_tensor_parameter_215: name = "parameter_215" + original_name = "batch_norm_80.w_0" shape = [2048] dtype = "float32" min_val = float("-1.21291") @@ -2376,6 +2592,7 @@ class Program_weight_tensor_parameter_215: class Program_weight_tensor_parameter_216: name = "parameter_216" + original_name = "batch_norm_80.w_2" shape = [2048] dtype = "float32" min_val = float("9.55327e-26") @@ -2387,6 +2604,7 @@ class Program_weight_tensor_parameter_216: class Program_weight_tensor_parameter_217: name = "parameter_217" + original_name = "batch_norm_80.w_1" shape = [2048] dtype = "float32" min_val = float("-4.10631") @@ -2398,6 +2616,7 @@ class Program_weight_tensor_parameter_217: class Program_weight_tensor_parameter_218: name = "parameter_218" + original_name = "conv2d_92.w_0" shape = [2048, 1, 5, 5] dtype = "float32" min_val = float("-0.458278") @@ -2409,6 +2628,7 @@ class Program_weight_tensor_parameter_218: class Program_weight_tensor_parameter_219: name = "parameter_219" + original_name = "batch_norm_79.b_0" shape = [2048] dtype = "float32" min_val = float("-3.58972") @@ -2420,6 +2640,7 @@ class Program_weight_tensor_parameter_219: class Program_weight_tensor_parameter_220: name = "parameter_220" + original_name = "batch_norm_79.w_0" shape = [2048] dtype = "float32" min_val = float("-1.86209") @@ -2431,6 +2652,7 @@ class Program_weight_tensor_parameter_220: class Program_weight_tensor_parameter_221: name = "parameter_221" + original_name = "batch_norm_79.w_2" shape = [2048] dtype = "float32" min_val = float("4.02771e-26") @@ -2442,6 +2664,7 @@ class Program_weight_tensor_parameter_221: class Program_weight_tensor_parameter_222: name = "parameter_222" + original_name = "batch_norm_79.w_1" shape = [2048] dtype = "float32" min_val = float("-5.3815") @@ -2453,6 +2676,7 @@ class Program_weight_tensor_parameter_222: class Program_weight_tensor_parameter_223: name = "parameter_223" + original_name = "conv2d_91.w_0" shape = [2048, 512, 1, 1] dtype = "float32" min_val = float("-0.444699") @@ -2464,6 +2688,7 @@ class Program_weight_tensor_parameter_223: class Program_weight_tensor_parameter_224: name = "parameter_224" + original_name = "batch_norm_78.b_0" shape = [512] dtype = "float32" min_val = float("-2.49326") @@ -2475,6 +2700,7 @@ class Program_weight_tensor_parameter_224: class Program_weight_tensor_parameter_225: name = "parameter_225" + original_name = "batch_norm_78.w_0" shape = [512] dtype = "float32" min_val = float("0.278529") @@ -2486,6 +2712,7 @@ class Program_weight_tensor_parameter_225: class Program_weight_tensor_parameter_226: name = "parameter_226" + original_name = "batch_norm_78.w_2" shape = [512] dtype = "float32" min_val = float("0.00188405") @@ -2497,6 +2724,7 @@ class Program_weight_tensor_parameter_226: class Program_weight_tensor_parameter_227: name = "parameter_227" + original_name = "batch_norm_78.w_1" shape = [512] dtype = "float32" min_val = float("-5.83867") @@ -2508,6 +2736,7 @@ class Program_weight_tensor_parameter_227: class Program_weight_tensor_parameter_228: name = "parameter_228" + original_name = "conv2d_90.w_0" shape = [512, 1, 5, 5] dtype = "float32" min_val = float("-0.540702") @@ -2519,6 +2748,7 @@ class Program_weight_tensor_parameter_228: class Program_weight_tensor_parameter_229: name = "parameter_229" + original_name = "batch_norm_77.b_0" shape = [512] dtype = "float32" min_val = float("-1.41189") @@ -2530,6 +2760,7 @@ class Program_weight_tensor_parameter_229: class Program_weight_tensor_parameter_230: name = "parameter_230" + original_name = "batch_norm_77.w_0" shape = [512] dtype = "float32" min_val = float("-1.14888") @@ -2541,6 +2772,7 @@ class Program_weight_tensor_parameter_230: class Program_weight_tensor_parameter_231: name = "parameter_231" + original_name = "batch_norm_77.w_2" shape = [512] dtype = "float32" min_val = float("1.87107") @@ -2552,6 +2784,7 @@ class Program_weight_tensor_parameter_231: class Program_weight_tensor_parameter_232: name = "parameter_232" + original_name = "batch_norm_77.w_1" shape = [512] dtype = "float32" min_val = float("-6.84815") @@ -2563,6 +2796,7 @@ class Program_weight_tensor_parameter_232: class Program_weight_tensor_parameter_233: name = "parameter_233" + original_name = "conv2d_89.w_0" shape = [512, 2048, 1, 1] dtype = "float32" min_val = float("-0.485521") @@ -2574,6 +2808,7 @@ class Program_weight_tensor_parameter_233: class Program_weight_tensor_parameter_234: name = "parameter_234" + original_name = "batch_norm_76.b_0" shape = [2048] dtype = "float32" min_val = float("-4.93584") @@ -2585,6 +2820,7 @@ class Program_weight_tensor_parameter_234: class Program_weight_tensor_parameter_235: name = "parameter_235" + original_name = "batch_norm_76.w_0" shape = [2048] dtype = "float32" min_val = float("-1.47443") @@ -2596,6 +2832,7 @@ class Program_weight_tensor_parameter_235: class Program_weight_tensor_parameter_236: name = "parameter_236" + original_name = "batch_norm_76.w_2" shape = [2048] dtype = "float32" min_val = float("1.99334e-36") @@ -2607,6 +2844,7 @@ class Program_weight_tensor_parameter_236: class Program_weight_tensor_parameter_237: name = "parameter_237" + original_name = "batch_norm_76.w_1" shape = [2048] dtype = "float32" min_val = float("-2.75626") @@ -2618,6 +2856,7 @@ class Program_weight_tensor_parameter_237: class Program_weight_tensor_parameter_238: name = "parameter_238" + original_name = "conv2d_88.w_0" shape = [2048, 1, 5, 5] dtype = "float32" min_val = float("-0.485618") @@ -2629,6 +2868,7 @@ class Program_weight_tensor_parameter_238: class Program_weight_tensor_parameter_239: name = "parameter_239" + original_name = "batch_norm_75.b_0" shape = [2048] dtype = "float32" min_val = float("-3.40111") @@ -2640,6 +2880,7 @@ class Program_weight_tensor_parameter_239: class Program_weight_tensor_parameter_240: name = "parameter_240" + original_name = "batch_norm_75.w_0" shape = [2048] dtype = "float32" min_val = float("-0.488094") @@ -2651,6 +2892,7 @@ class Program_weight_tensor_parameter_240: class Program_weight_tensor_parameter_241: name = "parameter_241" + original_name = "batch_norm_75.w_2" shape = [2048] dtype = "float32" min_val = float("2.65128e-33") @@ -2662,6 +2904,7 @@ class Program_weight_tensor_parameter_241: class Program_weight_tensor_parameter_242: name = "parameter_242" + original_name = "batch_norm_75.w_1" shape = [2048] dtype = "float32" min_val = float("-4.32325") @@ -2673,6 +2916,7 @@ class Program_weight_tensor_parameter_242: class Program_weight_tensor_parameter_243: name = "parameter_243" + original_name = "conv2d_87.w_0" shape = [2048, 512, 1, 1] dtype = "float32" min_val = float("-0.45358") @@ -2684,6 +2928,7 @@ class Program_weight_tensor_parameter_243: class Program_weight_tensor_parameter_244: name = "parameter_244" + original_name = "batch_norm_74.b_0" shape = [512] dtype = "float32" min_val = float("-1.99927") @@ -2695,6 +2940,7 @@ class Program_weight_tensor_parameter_244: class Program_weight_tensor_parameter_245: name = "parameter_245" + original_name = "batch_norm_74.w_0" shape = [512] dtype = "float32" min_val = float("0.243056") @@ -2706,6 +2952,7 @@ class Program_weight_tensor_parameter_245: class Program_weight_tensor_parameter_246: name = "parameter_246" + original_name = "batch_norm_74.w_2" shape = [512] dtype = "float32" min_val = float("6.77768e-05") @@ -2717,6 +2964,7 @@ class Program_weight_tensor_parameter_246: class Program_weight_tensor_parameter_247: name = "parameter_247" + original_name = "batch_norm_74.w_1" shape = [512] dtype = "float32" min_val = float("-4.20579") @@ -2728,6 +2976,7 @@ class Program_weight_tensor_parameter_247: class Program_weight_tensor_parameter_248: name = "parameter_248" + original_name = "conv2d_86.w_0" shape = [512, 1, 5, 5] dtype = "float32" min_val = float("-0.554609") @@ -2739,6 +2988,7 @@ class Program_weight_tensor_parameter_248: class Program_weight_tensor_parameter_249: name = "parameter_249" + original_name = "batch_norm_73.b_0" shape = [512] dtype = "float32" min_val = float("-1.36245") @@ -2750,6 +3000,7 @@ class Program_weight_tensor_parameter_249: class Program_weight_tensor_parameter_250: name = "parameter_250" + original_name = "batch_norm_73.w_0" shape = [512] dtype = "float32" min_val = float("-1.00164") @@ -2761,6 +3012,7 @@ class Program_weight_tensor_parameter_250: class Program_weight_tensor_parameter_251: name = "parameter_251" + original_name = "batch_norm_73.w_2" shape = [512] dtype = "float32" min_val = float("2.97034") @@ -2772,6 +3024,7 @@ class Program_weight_tensor_parameter_251: class Program_weight_tensor_parameter_252: name = "parameter_252" + original_name = "batch_norm_73.w_1" shape = [512] dtype = "float32" min_val = float("-6.71058") @@ -2783,6 +3036,7 @@ class Program_weight_tensor_parameter_252: class Program_weight_tensor_parameter_253: name = "parameter_253" + original_name = "conv2d_85.w_0" shape = [512, 2048, 1, 1] dtype = "float32" min_val = float("-0.552885") @@ -2794,6 +3048,7 @@ class Program_weight_tensor_parameter_253: class Program_weight_tensor_parameter_254: name = "parameter_254" + original_name = "batch_norm_72.b_0" shape = [2048] dtype = "float32" min_val = float("-8.30279") @@ -2805,6 +3060,7 @@ class Program_weight_tensor_parameter_254: class Program_weight_tensor_parameter_255: name = "parameter_255" + original_name = "batch_norm_72.w_0" shape = [2048] dtype = "float32" min_val = float("-1.05531") @@ -2816,6 +3072,7 @@ class Program_weight_tensor_parameter_255: class Program_weight_tensor_parameter_256: name = "parameter_256" + original_name = "batch_norm_72.w_2" shape = [2048] dtype = "float32" min_val = float("7.07747e-35") @@ -2827,6 +3084,7 @@ class Program_weight_tensor_parameter_256: class Program_weight_tensor_parameter_257: name = "parameter_257" + original_name = "batch_norm_72.w_1" shape = [2048] dtype = "float32" min_val = float("-1.87778") @@ -2838,6 +3096,7 @@ class Program_weight_tensor_parameter_257: class Program_weight_tensor_parameter_258: name = "parameter_258" + original_name = "conv2d_84.w_0" shape = [2048, 1, 5, 5] dtype = "float32" min_val = float("-0.425218") @@ -2849,6 +3108,7 @@ class Program_weight_tensor_parameter_258: class Program_weight_tensor_parameter_259: name = "parameter_259" + original_name = "batch_norm_71.b_0" shape = [2048] dtype = "float32" min_val = float("-3.90161") @@ -2860,6 +3120,7 @@ class Program_weight_tensor_parameter_259: class Program_weight_tensor_parameter_260: name = "parameter_260" + original_name = "batch_norm_71.w_0" shape = [2048] dtype = "float32" min_val = float("-0.773258") @@ -2871,6 +3132,7 @@ class Program_weight_tensor_parameter_260: class Program_weight_tensor_parameter_261: name = "parameter_261" + original_name = "batch_norm_71.w_2" shape = [2048] dtype = "float32" min_val = float("8.90707e-41") @@ -2882,6 +3144,7 @@ class Program_weight_tensor_parameter_261: class Program_weight_tensor_parameter_262: name = "parameter_262" + original_name = "batch_norm_71.w_1" shape = [2048] dtype = "float32" min_val = float("-4.4795") @@ -2893,6 +3156,7 @@ class Program_weight_tensor_parameter_262: class Program_weight_tensor_parameter_263: name = "parameter_263" + original_name = "conv2d_83.w_0" shape = [2048, 512, 1, 1] dtype = "float32" min_val = float("-0.460649") @@ -2904,6 +3168,7 @@ class Program_weight_tensor_parameter_263: class Program_weight_tensor_parameter_264: name = "parameter_264" + original_name = "batch_norm_70.b_0" shape = [512] dtype = "float32" min_val = float("-2.12008") @@ -2915,6 +3180,7 @@ class Program_weight_tensor_parameter_264: class Program_weight_tensor_parameter_265: name = "parameter_265" + original_name = "batch_norm_70.w_0" shape = [512] dtype = "float32" min_val = float("-0.466914") @@ -2926,6 +3192,7 @@ class Program_weight_tensor_parameter_265: class Program_weight_tensor_parameter_266: name = "parameter_266" + original_name = "batch_norm_70.w_2" shape = [512] dtype = "float32" min_val = float("9.12548e-07") @@ -2937,6 +3204,7 @@ class Program_weight_tensor_parameter_266: class Program_weight_tensor_parameter_267: name = "parameter_267" + original_name = "batch_norm_70.w_1" shape = [512] dtype = "float32" min_val = float("-20.1154") @@ -2948,6 +3216,7 @@ class Program_weight_tensor_parameter_267: class Program_weight_tensor_parameter_268: name = "parameter_268" + original_name = "conv2d_82.w_0" shape = [512, 1, 5, 5] dtype = "float32" min_val = float("-0.487653") @@ -2959,6 +3228,7 @@ class Program_weight_tensor_parameter_268: class Program_weight_tensor_parameter_269: name = "parameter_269" + original_name = "batch_norm_69.b_0" shape = [512] dtype = "float32" min_val = float("-3.24428") @@ -2970,6 +3240,7 @@ class Program_weight_tensor_parameter_269: class Program_weight_tensor_parameter_270: name = "parameter_270" + original_name = "batch_norm_69.w_0" shape = [512] dtype = "float32" min_val = float("0.0403706") @@ -2981,6 +3252,7 @@ class Program_weight_tensor_parameter_270: class Program_weight_tensor_parameter_271: name = "parameter_271" + original_name = "batch_norm_69.w_2" shape = [512] dtype = "float32" min_val = float("6.3279") @@ -2992,6 +3264,7 @@ class Program_weight_tensor_parameter_271: class Program_weight_tensor_parameter_272: name = "parameter_272" + original_name = "batch_norm_69.w_1" shape = [512] dtype = "float32" min_val = float("-27.1171") @@ -3003,6 +3276,7 @@ class Program_weight_tensor_parameter_272: class Program_weight_tensor_parameter_273: name = "parameter_273" + original_name = "conv2d_81.w_0" shape = [512, 768, 1, 1] dtype = "float32" min_val = float("-0.478877") @@ -3014,6 +3288,7 @@ class Program_weight_tensor_parameter_273: class Program_weight_tensor_parameter_274: name = "parameter_274" + original_name = "batch_norm_68.b_0" shape = [768] dtype = "float32" min_val = float("-6.59439") @@ -3025,6 +3300,7 @@ class Program_weight_tensor_parameter_274: class Program_weight_tensor_parameter_275: name = "parameter_275" + original_name = "batch_norm_68.w_0" shape = [768] dtype = "float32" min_val = float("0.335293") @@ -3036,6 +3312,7 @@ class Program_weight_tensor_parameter_275: class Program_weight_tensor_parameter_276: name = "parameter_276" + original_name = "batch_norm_68.w_2" shape = [768] dtype = "float32" min_val = float("0.000494923") @@ -3047,6 +3324,7 @@ class Program_weight_tensor_parameter_276: class Program_weight_tensor_parameter_277: name = "parameter_277" + original_name = "batch_norm_68.w_1" shape = [768] dtype = "float32" min_val = float("-4.62851") @@ -3058,6 +3336,7 @@ class Program_weight_tensor_parameter_277: class Program_weight_tensor_parameter_278: name = "parameter_278" + original_name = "conv2d_80.w_0" shape = [768, 1, 5, 5] dtype = "float32" min_val = float("-0.421024") @@ -3069,6 +3348,7 @@ class Program_weight_tensor_parameter_278: class Program_weight_tensor_parameter_279: name = "parameter_279" + original_name = "batch_norm_67.b_0" shape = [768] dtype = "float32" min_val = float("-2.70588") @@ -3080,6 +3360,7 @@ class Program_weight_tensor_parameter_279: class Program_weight_tensor_parameter_280: name = "parameter_280" + original_name = "batch_norm_67.w_0" shape = [768] dtype = "float32" min_val = float("-0.223399") @@ -3091,6 +3372,7 @@ class Program_weight_tensor_parameter_280: class Program_weight_tensor_parameter_281: name = "parameter_281" + original_name = "batch_norm_67.w_2" shape = [768] dtype = "float32" min_val = float("1.29695") @@ -3102,6 +3384,7 @@ class Program_weight_tensor_parameter_281: class Program_weight_tensor_parameter_282: name = "parameter_282" + original_name = "batch_norm_67.w_1" shape = [768] dtype = "float32" min_val = float("-3.55317") @@ -3113,6 +3396,7 @@ class Program_weight_tensor_parameter_282: class Program_weight_tensor_parameter_283: name = "parameter_283" + original_name = "conv2d_79.w_0" shape = [768, 192, 1, 1] dtype = "float32" min_val = float("-0.471231") @@ -3124,6 +3408,7 @@ class Program_weight_tensor_parameter_283: class Program_weight_tensor_parameter_284: name = "parameter_284" + original_name = "batch_norm_66.b_0" shape = [192] dtype = "float32" min_val = float("-2.41908") @@ -3135,6 +3420,7 @@ class Program_weight_tensor_parameter_284: class Program_weight_tensor_parameter_285: name = "parameter_285" + original_name = "batch_norm_66.w_0" shape = [192] dtype = "float32" min_val = float("0.0618396") @@ -3146,6 +3432,7 @@ class Program_weight_tensor_parameter_285: class Program_weight_tensor_parameter_286: name = "parameter_286" + original_name = "batch_norm_66.w_2" shape = [192] dtype = "float32" min_val = float("3722.26") @@ -3157,6 +3444,7 @@ class Program_weight_tensor_parameter_286: class Program_weight_tensor_parameter_287: name = "parameter_287" + original_name = "batch_norm_66.w_1" shape = [192] dtype = "float32" min_val = float("-1638.4") @@ -3168,6 +3456,7 @@ class Program_weight_tensor_parameter_287: class Program_weight_tensor_parameter_288: name = "parameter_288" + original_name = "conv2d_78.w_0" shape = [192, 1, 5, 5] dtype = "float32" min_val = float("-0.474769") @@ -3179,6 +3468,7 @@ class Program_weight_tensor_parameter_288: class Program_weight_tensor_parameter_289: name = "parameter_289" + original_name = "batch_norm_65.b_0" shape = [192] dtype = "float32" min_val = float("-2.28192") @@ -3190,6 +3480,7 @@ class Program_weight_tensor_parameter_289: class Program_weight_tensor_parameter_290: name = "parameter_290" + original_name = "batch_norm_65.w_0" shape = [192] dtype = "float32" min_val = float("-1.5784") @@ -3201,6 +3492,7 @@ class Program_weight_tensor_parameter_290: class Program_weight_tensor_parameter_291: name = "parameter_291" + original_name = "batch_norm_65.w_2" shape = [192] dtype = "float32" min_val = float("0.359874") @@ -3212,6 +3504,7 @@ class Program_weight_tensor_parameter_291: class Program_weight_tensor_parameter_292: name = "parameter_292" + original_name = "batch_norm_65.w_1" shape = [192] dtype = "float32" min_val = float("-3.11921") @@ -3223,6 +3516,7 @@ class Program_weight_tensor_parameter_292: class Program_weight_tensor_parameter_293: name = "parameter_293" + original_name = "conv2d_77.w_0" shape = [192, 768, 1, 1] dtype = "float32" min_val = float("-0.447465") @@ -3234,6 +3528,7 @@ class Program_weight_tensor_parameter_293: class Program_weight_tensor_parameter_294: name = "parameter_294" + original_name = "batch_norm_64.b_0" shape = [768] dtype = "float32" min_val = float("-5.54216") @@ -3245,6 +3540,7 @@ class Program_weight_tensor_parameter_294: class Program_weight_tensor_parameter_295: name = "parameter_295" + original_name = "batch_norm_64.w_0" shape = [768] dtype = "float32" min_val = float("0.287344") @@ -3256,6 +3552,7 @@ class Program_weight_tensor_parameter_295: class Program_weight_tensor_parameter_296: name = "parameter_296" + original_name = "batch_norm_64.w_2" shape = [768] dtype = "float32" min_val = float("0.573726") @@ -3267,6 +3564,7 @@ class Program_weight_tensor_parameter_296: class Program_weight_tensor_parameter_297: name = "parameter_297" + original_name = "batch_norm_64.w_1" shape = [768] dtype = "float32" min_val = float("-2.8216") @@ -3278,6 +3576,7 @@ class Program_weight_tensor_parameter_297: class Program_weight_tensor_parameter_298: name = "parameter_298" + original_name = "conv2d_76.w_0" shape = [768, 192, 1, 1] dtype = "float32" min_val = float("-0.397202") @@ -3289,6 +3588,7 @@ class Program_weight_tensor_parameter_298: class Program_weight_tensor_parameter_299: name = "parameter_299" + original_name = "batch_norm_63.b_0" shape = [192] dtype = "float32" min_val = float("-2.50051") @@ -3300,6 +3600,7 @@ class Program_weight_tensor_parameter_299: class Program_weight_tensor_parameter_300: name = "parameter_300" + original_name = "batch_norm_63.w_0" shape = [192] dtype = "float32" min_val = float("-0.913315") @@ -3311,6 +3612,7 @@ class Program_weight_tensor_parameter_300: class Program_weight_tensor_parameter_301: name = "parameter_301" + original_name = "batch_norm_63.w_2" shape = [192] dtype = "float32" min_val = float("127.515") @@ -3322,6 +3624,7 @@ class Program_weight_tensor_parameter_301: class Program_weight_tensor_parameter_302: name = "parameter_302" + original_name = "batch_norm_63.w_1" shape = [192] dtype = "float32" min_val = float("-432.775") @@ -3333,6 +3636,7 @@ class Program_weight_tensor_parameter_302: class Program_weight_tensor_parameter_303: name = "parameter_303" + original_name = "conv2d_75.w_0" shape = [192, 1, 3, 3] dtype = "float32" min_val = float("-0.423785") @@ -3344,6 +3648,7 @@ class Program_weight_tensor_parameter_303: class Program_weight_tensor_parameter_304: name = "parameter_304" + original_name = "conv2d_74.w_0" shape = [192, 384, 1, 1] dtype = "float32" min_val = float("-0.478038") @@ -3355,6 +3660,7 @@ class Program_weight_tensor_parameter_304: class Program_weight_tensor_parameter_305: name = "parameter_305" + original_name = "conv2d_73.w_0" shape = [48, 192, 1, 1] dtype = "float32" min_val = float("-0.457019") @@ -3366,6 +3672,7 @@ class Program_weight_tensor_parameter_305: class Program_weight_tensor_parameter_306: name = "parameter_306" + original_name = "batch_norm_62.b_0" shape = [192] dtype = "float32" min_val = float("-2.25012") @@ -3377,6 +3684,7 @@ class Program_weight_tensor_parameter_306: class Program_weight_tensor_parameter_307: name = "parameter_307" + original_name = "batch_norm_62.w_0" shape = [192] dtype = "float32" min_val = float("-4.04102") @@ -3388,6 +3696,7 @@ class Program_weight_tensor_parameter_307: class Program_weight_tensor_parameter_308: name = "parameter_308" + original_name = "batch_norm_62.w_2" shape = [192] dtype = "float32" min_val = float("0.00023628") @@ -3399,6 +3708,7 @@ class Program_weight_tensor_parameter_308: class Program_weight_tensor_parameter_309: name = "parameter_309" + original_name = "batch_norm_62.w_1" shape = [192] dtype = "float32" min_val = float("-0.391862") @@ -3410,6 +3720,7 @@ class Program_weight_tensor_parameter_309: class Program_weight_tensor_parameter_310: name = "parameter_310" + original_name = "conv2d_71.w_0" shape = [192, 1, 3, 3] dtype = "float32" min_val = float("-0.413886") @@ -3421,6 +3732,7 @@ class Program_weight_tensor_parameter_310: class Program_weight_tensor_parameter_311: name = "parameter_311" + original_name = "conv2d_72.w_0" shape = [48, 192, 1, 1] dtype = "float32" min_val = float("-0.432733") @@ -3432,6 +3744,7 @@ class Program_weight_tensor_parameter_311: class Program_weight_tensor_parameter_312: name = "parameter_312" + original_name = "batch_norm_61.b_0" shape = [192] dtype = "float32" min_val = float("-1.46094") @@ -3443,6 +3756,7 @@ class Program_weight_tensor_parameter_312: class Program_weight_tensor_parameter_313: name = "parameter_313" + original_name = "batch_norm_61.w_0" shape = [192] dtype = "float32" min_val = float("-1.96441") @@ -3454,6 +3768,7 @@ class Program_weight_tensor_parameter_313: class Program_weight_tensor_parameter_314: name = "parameter_314" + original_name = "batch_norm_61.w_2" shape = [192] dtype = "float32" min_val = float("0.0005277") @@ -3465,6 +3780,7 @@ class Program_weight_tensor_parameter_314: class Program_weight_tensor_parameter_315: name = "parameter_315" + original_name = "batch_norm_61.w_1" shape = [192] dtype = "float32" min_val = float("-0.527653") @@ -3476,6 +3792,7 @@ class Program_weight_tensor_parameter_315: class Program_weight_tensor_parameter_316: name = "parameter_316" + original_name = "conv2d_70.w_0" shape = [192, 1, 3, 3] dtype = "float32" min_val = float("-0.410584") @@ -3487,6 +3804,7 @@ class Program_weight_tensor_parameter_316: class Program_weight_tensor_parameter_317: name = "parameter_317" + original_name = "conv2d_69.w_0" shape = [384, 192, 1, 1] dtype = "float32" min_val = float("-0.472236") @@ -3498,6 +3816,7 @@ class Program_weight_tensor_parameter_317: class Program_weight_tensor_parameter_318: name = "parameter_318" + original_name = "batch_norm_60.b_0" shape = [192] dtype = "float32" min_val = float("-1.78079") @@ -3509,6 +3828,7 @@ class Program_weight_tensor_parameter_318: class Program_weight_tensor_parameter_319: name = "parameter_319" + original_name = "batch_norm_60.w_0" shape = [192] dtype = "float32" min_val = float("-0.323185") @@ -3520,6 +3840,7 @@ class Program_weight_tensor_parameter_319: class Program_weight_tensor_parameter_320: name = "parameter_320" + original_name = "batch_norm_60.w_2" shape = [192] dtype = "float32" min_val = float("1023.0") @@ -3531,6 +3852,7 @@ class Program_weight_tensor_parameter_320: class Program_weight_tensor_parameter_321: name = "parameter_321" + original_name = "batch_norm_60.w_1" shape = [192] dtype = "float32" min_val = float("-555.545") @@ -3542,6 +3864,7 @@ class Program_weight_tensor_parameter_321: class Program_weight_tensor_parameter_322: name = "parameter_322" + original_name = "batch_norm_59.b_0" shape = [192] dtype = "float32" min_val = float("-3.21842") @@ -3553,6 +3876,7 @@ class Program_weight_tensor_parameter_322: class Program_weight_tensor_parameter_323: name = "parameter_323" + original_name = "batch_norm_59.w_0" shape = [192] dtype = "float32" min_val = float("-0.938294") @@ -3564,6 +3888,7 @@ class Program_weight_tensor_parameter_323: class Program_weight_tensor_parameter_324: name = "parameter_324" + original_name = "batch_norm_59.w_2" shape = [192] dtype = "float32" min_val = float("0.770831") @@ -3575,6 +3900,7 @@ class Program_weight_tensor_parameter_324: class Program_weight_tensor_parameter_325: name = "parameter_325" + original_name = "batch_norm_59.w_1" shape = [192] dtype = "float32" min_val = float("-8.14141") @@ -3586,6 +3912,7 @@ class Program_weight_tensor_parameter_325: class Program_weight_tensor_parameter_326: name = "parameter_326" + original_name = "conv2d_68.w_0" shape = [192, 768, 1, 1] dtype = "float32" min_val = float("-0.461058") @@ -3597,6 +3924,7 @@ class Program_weight_tensor_parameter_326: class Program_weight_tensor_parameter_327: name = "parameter_327" + original_name = "batch_norm_58.b_0" shape = [768] dtype = "float32" min_val = float("-5.16451") @@ -3608,6 +3936,7 @@ class Program_weight_tensor_parameter_327: class Program_weight_tensor_parameter_328: name = "parameter_328" + original_name = "batch_norm_58.w_0" shape = [768] dtype = "float32" min_val = float("-1.34065") @@ -3619,6 +3948,7 @@ class Program_weight_tensor_parameter_328: class Program_weight_tensor_parameter_329: name = "parameter_329" + original_name = "batch_norm_58.w_2" shape = [768] dtype = "float32" min_val = float("1.8297e-39") @@ -3630,6 +3960,7 @@ class Program_weight_tensor_parameter_329: class Program_weight_tensor_parameter_330: name = "parameter_330" + original_name = "batch_norm_58.w_1" shape = [768] dtype = "float32" min_val = float("-1.26748") @@ -3641,6 +3972,7 @@ class Program_weight_tensor_parameter_330: class Program_weight_tensor_parameter_331: name = "parameter_331" + original_name = "conv2d_67.w_0" shape = [768, 1, 3, 3] dtype = "float32" min_val = float("-0.397212") @@ -3652,6 +3984,7 @@ class Program_weight_tensor_parameter_331: class Program_weight_tensor_parameter_332: name = "parameter_332" + original_name = "batch_norm_57.b_0" shape = [768] dtype = "float32" min_val = float("-3.98242") @@ -3663,6 +3996,7 @@ class Program_weight_tensor_parameter_332: class Program_weight_tensor_parameter_333: name = "parameter_333" + original_name = "batch_norm_57.w_0" shape = [768] dtype = "float32" min_val = float("-0.214746") @@ -3674,6 +4008,7 @@ class Program_weight_tensor_parameter_333: class Program_weight_tensor_parameter_334: name = "parameter_334" + original_name = "batch_norm_57.w_2" shape = [768] dtype = "float32" min_val = float("2.59633e-41") @@ -3685,6 +4020,7 @@ class Program_weight_tensor_parameter_334: class Program_weight_tensor_parameter_335: name = "parameter_335" + original_name = "batch_norm_57.w_1" shape = [768] dtype = "float32" min_val = float("-3.12353") @@ -3696,6 +4032,7 @@ class Program_weight_tensor_parameter_335: class Program_weight_tensor_parameter_336: name = "parameter_336" + original_name = "conv2d_66.w_0" shape = [768, 192, 1, 1] dtype = "float32" min_val = float("-0.467739") @@ -3707,6 +4044,7 @@ class Program_weight_tensor_parameter_336: class Program_weight_tensor_parameter_337: name = "parameter_337" + original_name = "batch_norm_56.b_0" shape = [192] dtype = "float32" min_val = float("-1.92106") @@ -3718,6 +4056,7 @@ class Program_weight_tensor_parameter_337: class Program_weight_tensor_parameter_338: name = "parameter_338" + original_name = "batch_norm_56.w_0" shape = [192] dtype = "float32" min_val = float("-0.190248") @@ -3729,6 +4068,7 @@ class Program_weight_tensor_parameter_338: class Program_weight_tensor_parameter_339: name = "parameter_339" + original_name = "batch_norm_56.w_2" shape = [192] dtype = "float32" min_val = float("1787.49") @@ -3740,6 +4080,7 @@ class Program_weight_tensor_parameter_339: class Program_weight_tensor_parameter_340: name = "parameter_340" + original_name = "batch_norm_56.w_1" shape = [192] dtype = "float32" min_val = float("-1531.17") @@ -3751,6 +4092,7 @@ class Program_weight_tensor_parameter_340: class Program_weight_tensor_parameter_341: name = "parameter_341" + original_name = "conv2d_65.w_0" shape = [192, 1, 5, 5] dtype = "float32" min_val = float("-0.647436") @@ -3762,6 +4104,7 @@ class Program_weight_tensor_parameter_341: class Program_weight_tensor_parameter_342: name = "parameter_342" + original_name = "conv2d_64.w_0" shape = [192, 384, 1, 1] dtype = "float32" min_val = float("-0.466684") @@ -3773,6 +4116,7 @@ class Program_weight_tensor_parameter_342: class Program_weight_tensor_parameter_343: name = "parameter_343" + original_name = "conv2d_63.w_0" shape = [48, 192, 1, 1] dtype = "float32" min_val = float("-0.44953") @@ -3784,6 +4128,7 @@ class Program_weight_tensor_parameter_343: class Program_weight_tensor_parameter_344: name = "parameter_344" + original_name = "batch_norm_55.b_0" shape = [192] dtype = "float32" min_val = float("-1.24602") @@ -3795,6 +4140,7 @@ class Program_weight_tensor_parameter_344: class Program_weight_tensor_parameter_345: name = "parameter_345" + original_name = "batch_norm_55.w_0" shape = [192] dtype = "float32" min_val = float("-4.41686") @@ -3806,6 +4152,7 @@ class Program_weight_tensor_parameter_345: class Program_weight_tensor_parameter_346: name = "parameter_346" + original_name = "batch_norm_55.w_2" shape = [192] dtype = "float32" min_val = float("0.00236472") @@ -3817,6 +4164,7 @@ class Program_weight_tensor_parameter_346: class Program_weight_tensor_parameter_347: name = "parameter_347" + original_name = "batch_norm_55.w_1" shape = [192] dtype = "float32" min_val = float("-0.907748") @@ -3828,6 +4176,7 @@ class Program_weight_tensor_parameter_347: class Program_weight_tensor_parameter_348: name = "parameter_348" + original_name = "conv2d_61.w_0" shape = [192, 1, 3, 3] dtype = "float32" min_val = float("-0.374802") @@ -3839,6 +4188,7 @@ class Program_weight_tensor_parameter_348: class Program_weight_tensor_parameter_349: name = "parameter_349" + original_name = "conv2d_62.w_0" shape = [48, 192, 1, 1] dtype = "float32" min_val = float("-0.50296") @@ -3850,6 +4200,7 @@ class Program_weight_tensor_parameter_349: class Program_weight_tensor_parameter_350: name = "parameter_350" + original_name = "batch_norm_54.b_0" shape = [192] dtype = "float32" min_val = float("-2.22689") @@ -3861,6 +4212,7 @@ class Program_weight_tensor_parameter_350: class Program_weight_tensor_parameter_351: name = "parameter_351" + original_name = "batch_norm_54.w_0" shape = [192] dtype = "float32" min_val = float("-1.8986") @@ -3872,6 +4224,7 @@ class Program_weight_tensor_parameter_351: class Program_weight_tensor_parameter_352: name = "parameter_352" + original_name = "batch_norm_54.w_2" shape = [192] dtype = "float32" min_val = float("0.00441886") @@ -3883,6 +4236,7 @@ class Program_weight_tensor_parameter_352: class Program_weight_tensor_parameter_353: name = "parameter_353" + original_name = "batch_norm_54.w_1" shape = [192] dtype = "float32" min_val = float("-2.84547") @@ -3894,6 +4248,7 @@ class Program_weight_tensor_parameter_353: class Program_weight_tensor_parameter_354: name = "parameter_354" + original_name = "conv2d_60.w_0" shape = [192, 1, 3, 3] dtype = "float32" min_val = float("-0.403637") @@ -3905,6 +4260,7 @@ class Program_weight_tensor_parameter_354: class Program_weight_tensor_parameter_355: name = "parameter_355" + original_name = "conv2d_59.w_0" shape = [384, 192, 1, 1] dtype = "float32" min_val = float("-0.478642") @@ -3916,6 +4272,7 @@ class Program_weight_tensor_parameter_355: class Program_weight_tensor_parameter_356: name = "parameter_356" + original_name = "batch_norm_53.b_0" shape = [192] dtype = "float32" min_val = float("-5.86816") @@ -3927,6 +4284,7 @@ class Program_weight_tensor_parameter_356: class Program_weight_tensor_parameter_357: name = "parameter_357" + original_name = "batch_norm_53.w_0" shape = [192] dtype = "float32" min_val = float("-0.287234") @@ -3938,6 +4296,7 @@ class Program_weight_tensor_parameter_357: class Program_weight_tensor_parameter_358: name = "parameter_358" + original_name = "batch_norm_53.w_2" shape = [192] dtype = "float32" min_val = float("128.269") @@ -3949,6 +4308,7 @@ class Program_weight_tensor_parameter_358: class Program_weight_tensor_parameter_359: name = "parameter_359" + original_name = "batch_norm_53.w_1" shape = [192] dtype = "float32" min_val = float("-532.749") @@ -3960,6 +4320,7 @@ class Program_weight_tensor_parameter_359: class Program_weight_tensor_parameter_360: name = "parameter_360" + original_name = "batch_norm_52.b_0" shape = [192] dtype = "float32" min_val = float("-1.0663") @@ -3971,6 +4332,7 @@ class Program_weight_tensor_parameter_360: class Program_weight_tensor_parameter_361: name = "parameter_361" + original_name = "batch_norm_52.w_0" shape = [192] dtype = "float32" min_val = float("-0.744302") @@ -3982,6 +4344,7 @@ class Program_weight_tensor_parameter_361: class Program_weight_tensor_parameter_362: name = "parameter_362" + original_name = "batch_norm_52.w_2" shape = [192] dtype = "float32" min_val = float("0.701872") @@ -3993,6 +4356,7 @@ class Program_weight_tensor_parameter_362: class Program_weight_tensor_parameter_363: name = "parameter_363" + original_name = "batch_norm_52.w_1" shape = [192] dtype = "float32" min_val = float("-6.95306") @@ -4004,6 +4368,7 @@ class Program_weight_tensor_parameter_363: class Program_weight_tensor_parameter_364: name = "parameter_364" + original_name = "conv2d_58.w_0" shape = [192, 768, 1, 1] dtype = "float32" min_val = float("-0.443977") @@ -4015,6 +4380,7 @@ class Program_weight_tensor_parameter_364: class Program_weight_tensor_parameter_365: name = "parameter_365" + original_name = "batch_norm_51.b_0" shape = [768] dtype = "float32" min_val = float("-5.68141") @@ -4026,6 +4392,7 @@ class Program_weight_tensor_parameter_365: class Program_weight_tensor_parameter_366: name = "parameter_366" + original_name = "batch_norm_51.w_0" shape = [768] dtype = "float32" min_val = float("-1.7286") @@ -4037,6 +4404,7 @@ class Program_weight_tensor_parameter_366: class Program_weight_tensor_parameter_367: name = "parameter_367" + original_name = "batch_norm_51.w_2" shape = [768] dtype = "float32" min_val = float("1.37327e-43") @@ -4048,6 +4416,7 @@ class Program_weight_tensor_parameter_367: class Program_weight_tensor_parameter_368: name = "parameter_368" + original_name = "batch_norm_51.w_1" shape = [768] dtype = "float32" min_val = float("-1.90516") @@ -4059,6 +4428,7 @@ class Program_weight_tensor_parameter_368: class Program_weight_tensor_parameter_369: name = "parameter_369" + original_name = "conv2d_57.w_0" shape = [768, 1, 3, 3] dtype = "float32" min_val = float("-0.444534") @@ -4070,6 +4440,7 @@ class Program_weight_tensor_parameter_369: class Program_weight_tensor_parameter_370: name = "parameter_370" + original_name = "batch_norm_50.b_0" shape = [768] dtype = "float32" min_val = float("-3.02302") @@ -4081,6 +4452,7 @@ class Program_weight_tensor_parameter_370: class Program_weight_tensor_parameter_371: name = "parameter_371" + original_name = "batch_norm_50.w_0" shape = [768] dtype = "float32" min_val = float("-0.390728") @@ -4092,6 +4464,7 @@ class Program_weight_tensor_parameter_371: class Program_weight_tensor_parameter_372: name = "parameter_372" + original_name = "batch_norm_50.w_2" shape = [768] dtype = "float32" min_val = float("1.52742e-43") @@ -4103,6 +4476,7 @@ class Program_weight_tensor_parameter_372: class Program_weight_tensor_parameter_373: name = "parameter_373" + original_name = "batch_norm_50.w_1" shape = [768] dtype = "float32" min_val = float("-4.70545") @@ -4114,6 +4488,7 @@ class Program_weight_tensor_parameter_373: class Program_weight_tensor_parameter_374: name = "parameter_374" + original_name = "conv2d_56.w_0" shape = [768, 192, 1, 1] dtype = "float32" min_val = float("-0.661591") @@ -4125,6 +4500,7 @@ class Program_weight_tensor_parameter_374: class Program_weight_tensor_parameter_375: name = "parameter_375" + original_name = "batch_norm_49.b_0" shape = [192] dtype = "float32" min_val = float("-2.17452") @@ -4136,6 +4512,7 @@ class Program_weight_tensor_parameter_375: class Program_weight_tensor_parameter_376: name = "parameter_376" + original_name = "batch_norm_49.w_0" shape = [192] dtype = "float32" min_val = float("-0.529652") @@ -4147,6 +4524,7 @@ class Program_weight_tensor_parameter_376: class Program_weight_tensor_parameter_377: name = "parameter_377" + original_name = "batch_norm_49.w_2" shape = [192] dtype = "float32" min_val = float("381.721") @@ -4158,6 +4536,7 @@ class Program_weight_tensor_parameter_377: class Program_weight_tensor_parameter_378: name = "parameter_378" + original_name = "batch_norm_49.w_1" shape = [192] dtype = "float32" min_val = float("-1226.1") @@ -4169,6 +4548,7 @@ class Program_weight_tensor_parameter_378: class Program_weight_tensor_parameter_379: name = "parameter_379" + original_name = "conv2d_55.w_0" shape = [192, 1, 5, 5] dtype = "float32" min_val = float("-0.553414") @@ -4180,6 +4560,7 @@ class Program_weight_tensor_parameter_379: class Program_weight_tensor_parameter_380: name = "parameter_380" + original_name = "conv2d_54.w_0" shape = [192, 384, 1, 1] dtype = "float32" min_val = float("-0.479818") @@ -4191,6 +4572,7 @@ class Program_weight_tensor_parameter_380: class Program_weight_tensor_parameter_381: name = "parameter_381" + original_name = "conv2d_53.w_0" shape = [48, 192, 1, 1] dtype = "float32" min_val = float("-0.477917") @@ -4202,6 +4584,7 @@ class Program_weight_tensor_parameter_381: class Program_weight_tensor_parameter_382: name = "parameter_382" + original_name = "batch_norm_48.b_0" shape = [192] dtype = "float32" min_val = float("-1.91865") @@ -4213,6 +4596,7 @@ class Program_weight_tensor_parameter_382: class Program_weight_tensor_parameter_383: name = "parameter_383" + original_name = "batch_norm_48.w_0" shape = [192] dtype = "float32" min_val = float("-1.73885") @@ -4224,6 +4608,7 @@ class Program_weight_tensor_parameter_383: class Program_weight_tensor_parameter_384: name = "parameter_384" + original_name = "batch_norm_48.w_2" shape = [192] dtype = "float32" min_val = float("0.000258152") @@ -4235,6 +4620,7 @@ class Program_weight_tensor_parameter_384: class Program_weight_tensor_parameter_385: name = "parameter_385" + original_name = "batch_norm_48.w_1" shape = [192] dtype = "float32" min_val = float("-3.27257") @@ -4246,6 +4632,7 @@ class Program_weight_tensor_parameter_385: class Program_weight_tensor_parameter_386: name = "parameter_386" + original_name = "conv2d_51.w_0" shape = [192, 1, 3, 3] dtype = "float32" min_val = float("-0.408844") @@ -4257,6 +4644,7 @@ class Program_weight_tensor_parameter_386: class Program_weight_tensor_parameter_387: name = "parameter_387" + original_name = "conv2d_52.w_0" shape = [48, 192, 1, 1] dtype = "float32" min_val = float("-0.452689") @@ -4268,6 +4656,7 @@ class Program_weight_tensor_parameter_387: class Program_weight_tensor_parameter_388: name = "parameter_388" + original_name = "batch_norm_47.b_0" shape = [192] dtype = "float32" min_val = float("-2.45729") @@ -4279,6 +4668,7 @@ class Program_weight_tensor_parameter_388: class Program_weight_tensor_parameter_389: name = "parameter_389" + original_name = "batch_norm_47.w_0" shape = [192] dtype = "float32" min_val = float("-1.13226") @@ -4290,6 +4680,7 @@ class Program_weight_tensor_parameter_389: class Program_weight_tensor_parameter_390: name = "parameter_390" + original_name = "batch_norm_47.w_2" shape = [192] dtype = "float32" min_val = float("0.000196489") @@ -4301,6 +4692,7 @@ class Program_weight_tensor_parameter_390: class Program_weight_tensor_parameter_391: name = "parameter_391" + original_name = "batch_norm_47.w_1" shape = [192] dtype = "float32" min_val = float("-1.14781") @@ -4312,6 +4704,7 @@ class Program_weight_tensor_parameter_391: class Program_weight_tensor_parameter_392: name = "parameter_392" + original_name = "conv2d_50.w_0" shape = [192, 1, 3, 3] dtype = "float32" min_val = float("-0.41948") @@ -4323,6 +4716,7 @@ class Program_weight_tensor_parameter_392: class Program_weight_tensor_parameter_393: name = "parameter_393" + original_name = "conv2d_49.w_0" shape = [384, 192, 1, 1] dtype = "float32" min_val = float("-0.530547") @@ -4334,6 +4728,7 @@ class Program_weight_tensor_parameter_393: class Program_weight_tensor_parameter_394: name = "parameter_394" + original_name = "batch_norm_46.b_0" shape = [192] dtype = "float32" min_val = float("-4.37711") @@ -4345,6 +4740,7 @@ class Program_weight_tensor_parameter_394: class Program_weight_tensor_parameter_395: name = "parameter_395" + original_name = "batch_norm_46.w_0" shape = [192] dtype = "float32" min_val = float("-1.65701") @@ -4356,6 +4752,7 @@ class Program_weight_tensor_parameter_395: class Program_weight_tensor_parameter_396: name = "parameter_396" + original_name = "batch_norm_46.w_2" shape = [192] dtype = "float32" min_val = float("40.0531") @@ -4367,6 +4764,7 @@ class Program_weight_tensor_parameter_396: class Program_weight_tensor_parameter_397: name = "parameter_397" + original_name = "batch_norm_46.w_1" shape = [192] dtype = "float32" min_val = float("-516.648") @@ -4378,6 +4776,7 @@ class Program_weight_tensor_parameter_397: class Program_weight_tensor_parameter_398: name = "parameter_398" + original_name = "batch_norm_45.b_0" shape = [192] dtype = "float32" min_val = float("-0.819501") @@ -4389,6 +4788,7 @@ class Program_weight_tensor_parameter_398: class Program_weight_tensor_parameter_399: name = "parameter_399" + original_name = "batch_norm_45.w_0" shape = [192] dtype = "float32" min_val = float("-0.737765") @@ -4400,6 +4800,7 @@ class Program_weight_tensor_parameter_399: class Program_weight_tensor_parameter_400: name = "parameter_400" + original_name = "batch_norm_45.w_2" shape = [192] dtype = "float32" min_val = float("0.463353") @@ -4411,6 +4812,7 @@ class Program_weight_tensor_parameter_400: class Program_weight_tensor_parameter_401: name = "parameter_401" + original_name = "batch_norm_45.w_1" shape = [192] dtype = "float32" min_val = float("-7.67427") @@ -4422,6 +4824,7 @@ class Program_weight_tensor_parameter_401: class Program_weight_tensor_parameter_402: name = "parameter_402" + original_name = "conv2d_48.w_0" shape = [192, 768, 1, 1] dtype = "float32" min_val = float("-0.467904") @@ -4433,6 +4836,7 @@ class Program_weight_tensor_parameter_402: class Program_weight_tensor_parameter_403: name = "parameter_403" + original_name = "batch_norm_44.b_0" shape = [768] dtype = "float32" min_val = float("-7.28441") @@ -4444,6 +4848,7 @@ class Program_weight_tensor_parameter_403: class Program_weight_tensor_parameter_404: name = "parameter_404" + original_name = "batch_norm_44.w_0" shape = [768] dtype = "float32" min_val = float("-2.42037") @@ -4455,6 +4860,7 @@ class Program_weight_tensor_parameter_404: class Program_weight_tensor_parameter_405: name = "parameter_405" + original_name = "batch_norm_44.w_2" shape = [768] dtype = "float32" min_val = float("3.76669e-42") @@ -4466,6 +4872,7 @@ class Program_weight_tensor_parameter_405: class Program_weight_tensor_parameter_406: name = "parameter_406" + original_name = "batch_norm_44.w_1" shape = [768] dtype = "float32" min_val = float("-1.40191") @@ -4477,6 +4884,7 @@ class Program_weight_tensor_parameter_406: class Program_weight_tensor_parameter_407: name = "parameter_407" + original_name = "conv2d_47.w_0" shape = [768, 1, 3, 3] dtype = "float32" min_val = float("-0.447939") @@ -4488,6 +4896,7 @@ class Program_weight_tensor_parameter_407: class Program_weight_tensor_parameter_408: name = "parameter_408" + original_name = "batch_norm_43.b_0" shape = [768] dtype = "float32" min_val = float("-4.12202") @@ -4499,6 +4908,7 @@ class Program_weight_tensor_parameter_408: class Program_weight_tensor_parameter_409: name = "parameter_409" + original_name = "batch_norm_43.w_0" shape = [768] dtype = "float32" min_val = float("-0.47747") @@ -4510,6 +4920,7 @@ class Program_weight_tensor_parameter_409: class Program_weight_tensor_parameter_410: name = "parameter_410" + original_name = "batch_norm_43.w_2" shape = [768] dtype = "float32" min_val = float("9.11825e-42") @@ -4521,6 +4932,7 @@ class Program_weight_tensor_parameter_410: class Program_weight_tensor_parameter_411: name = "parameter_411" + original_name = "batch_norm_43.w_1" shape = [768] dtype = "float32" min_val = float("-3.2664") @@ -4532,6 +4944,7 @@ class Program_weight_tensor_parameter_411: class Program_weight_tensor_parameter_412: name = "parameter_412" + original_name = "conv2d_46.w_0" shape = [768, 192, 1, 1] dtype = "float32" min_val = float("-0.455392") @@ -4543,6 +4956,7 @@ class Program_weight_tensor_parameter_412: class Program_weight_tensor_parameter_413: name = "parameter_413" + original_name = "batch_norm_42.b_0" shape = [192] dtype = "float32" min_val = float("-1.95172") @@ -4554,6 +4968,7 @@ class Program_weight_tensor_parameter_413: class Program_weight_tensor_parameter_414: name = "parameter_414" + original_name = "batch_norm_42.w_0" shape = [192] dtype = "float32" min_val = float("-0.084367") @@ -4565,6 +4980,7 @@ class Program_weight_tensor_parameter_414: class Program_weight_tensor_parameter_415: name = "parameter_415" + original_name = "batch_norm_42.w_2" shape = [192] dtype = "float32" min_val = float("31.543") @@ -4576,6 +4992,7 @@ class Program_weight_tensor_parameter_415: class Program_weight_tensor_parameter_416: name = "parameter_416" + original_name = "batch_norm_42.w_1" shape = [192] dtype = "float32" min_val = float("-2024.8") @@ -4587,6 +5004,7 @@ class Program_weight_tensor_parameter_416: class Program_weight_tensor_parameter_417: name = "parameter_417" + original_name = "conv2d_45.w_0" shape = [192, 1, 5, 5] dtype = "float32" min_val = float("-0.415789") @@ -4598,6 +5016,7 @@ class Program_weight_tensor_parameter_417: class Program_weight_tensor_parameter_418: name = "parameter_418" + original_name = "conv2d_44.w_0" shape = [192, 384, 1, 1] dtype = "float32" min_val = float("-0.490187") @@ -4609,6 +5028,7 @@ class Program_weight_tensor_parameter_418: class Program_weight_tensor_parameter_419: name = "parameter_419" + original_name = "conv2d_43.w_0" shape = [48, 192, 1, 1] dtype = "float32" min_val = float("-0.484937") @@ -4620,6 +5040,7 @@ class Program_weight_tensor_parameter_419: class Program_weight_tensor_parameter_420: name = "parameter_420" + original_name = "batch_norm_41.b_0" shape = [192] dtype = "float32" min_val = float("-2.02801") @@ -4631,6 +5052,7 @@ class Program_weight_tensor_parameter_420: class Program_weight_tensor_parameter_421: name = "parameter_421" + original_name = "batch_norm_41.w_0" shape = [192] dtype = "float32" min_val = float("-1.44342") @@ -4642,6 +5064,7 @@ class Program_weight_tensor_parameter_421: class Program_weight_tensor_parameter_422: name = "parameter_422" + original_name = "batch_norm_41.w_2" shape = [192] dtype = "float32" min_val = float("0.000426982") @@ -4653,6 +5076,7 @@ class Program_weight_tensor_parameter_422: class Program_weight_tensor_parameter_423: name = "parameter_423" + original_name = "batch_norm_41.w_1" shape = [192] dtype = "float32" min_val = float("-3.05074") @@ -4664,6 +5088,7 @@ class Program_weight_tensor_parameter_423: class Program_weight_tensor_parameter_424: name = "parameter_424" + original_name = "conv2d_41.w_0" shape = [192, 1, 3, 3] dtype = "float32" min_val = float("-0.412146") @@ -4675,6 +5100,7 @@ class Program_weight_tensor_parameter_424: class Program_weight_tensor_parameter_425: name = "parameter_425" + original_name = "conv2d_42.w_0" shape = [48, 192, 1, 1] dtype = "float32" min_val = float("-0.492958") @@ -4686,6 +5112,7 @@ class Program_weight_tensor_parameter_425: class Program_weight_tensor_parameter_426: name = "parameter_426" + original_name = "batch_norm_40.b_0" shape = [192] dtype = "float32" min_val = float("-2.05529") @@ -4697,6 +5124,7 @@ class Program_weight_tensor_parameter_426: class Program_weight_tensor_parameter_427: name = "parameter_427" + original_name = "batch_norm_40.w_0" shape = [192] dtype = "float32" min_val = float("-1.4329") @@ -4708,6 +5136,7 @@ class Program_weight_tensor_parameter_427: class Program_weight_tensor_parameter_428: name = "parameter_428" + original_name = "batch_norm_40.w_2" shape = [192] dtype = "float32" min_val = float("0.000565339") @@ -4719,6 +5148,7 @@ class Program_weight_tensor_parameter_428: class Program_weight_tensor_parameter_429: name = "parameter_429" + original_name = "batch_norm_40.w_1" shape = [192] dtype = "float32" min_val = float("-3.46282") @@ -4730,6 +5160,7 @@ class Program_weight_tensor_parameter_429: class Program_weight_tensor_parameter_430: name = "parameter_430" + original_name = "conv2d_40.w_0" shape = [192, 1, 3, 3] dtype = "float32" min_val = float("-0.448159") @@ -4741,6 +5172,7 @@ class Program_weight_tensor_parameter_430: class Program_weight_tensor_parameter_431: name = "parameter_431" + original_name = "conv2d_39.w_0" shape = [384, 192, 1, 1] dtype = "float32" min_val = float("-0.481493") @@ -4752,6 +5184,7 @@ class Program_weight_tensor_parameter_431: class Program_weight_tensor_parameter_432: name = "parameter_432" + original_name = "batch_norm_39.b_0" shape = [192] dtype = "float32" min_val = float("-6.10758") @@ -4763,6 +5196,7 @@ class Program_weight_tensor_parameter_432: class Program_weight_tensor_parameter_433: name = "parameter_433" + original_name = "batch_norm_39.w_0" shape = [192] dtype = "float32" min_val = float("-1.7707") @@ -4774,6 +5208,7 @@ class Program_weight_tensor_parameter_433: class Program_weight_tensor_parameter_434: name = "parameter_434" + original_name = "batch_norm_39.w_2" shape = [192] dtype = "float32" min_val = float("0.0409091") @@ -4785,6 +5220,7 @@ class Program_weight_tensor_parameter_434: class Program_weight_tensor_parameter_435: name = "parameter_435" + original_name = "batch_norm_39.w_1" shape = [192] dtype = "float32" min_val = float("-4.29209") @@ -4796,6 +5232,7 @@ class Program_weight_tensor_parameter_435: class Program_weight_tensor_parameter_436: name = "parameter_436" + original_name = "batch_norm_38.b_0" shape = [192] dtype = "float32" min_val = float("-0.816021") @@ -4807,6 +5244,7 @@ class Program_weight_tensor_parameter_436: class Program_weight_tensor_parameter_437: name = "parameter_437" + original_name = "batch_norm_38.w_0" shape = [192] dtype = "float32" min_val = float("-0.756494") @@ -4818,6 +5256,7 @@ class Program_weight_tensor_parameter_437: class Program_weight_tensor_parameter_438: name = "parameter_438" + original_name = "batch_norm_38.w_2" shape = [192] dtype = "float32" min_val = float("2.27109") @@ -4829,6 +5268,7 @@ class Program_weight_tensor_parameter_438: class Program_weight_tensor_parameter_439: name = "parameter_439" + original_name = "batch_norm_38.w_1" shape = [192] dtype = "float32" min_val = float("-14.719") @@ -4840,6 +5280,7 @@ class Program_weight_tensor_parameter_439: class Program_weight_tensor_parameter_440: name = "parameter_440" + original_name = "conv2d_38.w_0" shape = [192, 768, 1, 1] dtype = "float32" min_val = float("-0.542244") @@ -4851,6 +5292,7 @@ class Program_weight_tensor_parameter_440: class Program_weight_tensor_parameter_441: name = "parameter_441" + original_name = "batch_norm_37.b_0" shape = [768] dtype = "float32" min_val = float("-9.53836") @@ -4862,6 +5304,7 @@ class Program_weight_tensor_parameter_441: class Program_weight_tensor_parameter_442: name = "parameter_442" + original_name = "batch_norm_37.w_0" shape = [768] dtype = "float32" min_val = float("-1.56386") @@ -4873,6 +5316,7 @@ class Program_weight_tensor_parameter_442: class Program_weight_tensor_parameter_443: name = "parameter_443" + original_name = "batch_norm_37.w_2" shape = [768] dtype = "float32" min_val = float("8.54075e-36") @@ -4884,6 +5328,7 @@ class Program_weight_tensor_parameter_443: class Program_weight_tensor_parameter_444: name = "parameter_444" + original_name = "batch_norm_37.w_1" shape = [768] dtype = "float32" min_val = float("-2.73997") @@ -4895,6 +5340,7 @@ class Program_weight_tensor_parameter_444: class Program_weight_tensor_parameter_445: name = "parameter_445" + original_name = "conv2d_37.w_0" shape = [768, 1, 3, 3] dtype = "float32" min_val = float("-0.513387") @@ -4906,6 +5352,7 @@ class Program_weight_tensor_parameter_445: class Program_weight_tensor_parameter_446: name = "parameter_446" + original_name = "batch_norm_36.b_0" shape = [768] dtype = "float32" min_val = float("-4.64562") @@ -4917,6 +5364,7 @@ class Program_weight_tensor_parameter_446: class Program_weight_tensor_parameter_447: name = "parameter_447" + original_name = "batch_norm_36.w_0" shape = [768] dtype = "float32" min_val = float("-0.198908") @@ -4928,6 +5376,7 @@ class Program_weight_tensor_parameter_447: class Program_weight_tensor_parameter_448: name = "parameter_448" + original_name = "batch_norm_36.w_2" shape = [768] dtype = "float32" min_val = float("1.1851e-38") @@ -4939,6 +5388,7 @@ class Program_weight_tensor_parameter_448: class Program_weight_tensor_parameter_449: name = "parameter_449" + original_name = "batch_norm_36.w_1" shape = [768] dtype = "float32" min_val = float("-4.14052") @@ -4950,6 +5400,7 @@ class Program_weight_tensor_parameter_449: class Program_weight_tensor_parameter_450: name = "parameter_450" + original_name = "conv2d_36.w_0" shape = [768, 192, 1, 1] dtype = "float32" min_val = float("-0.572608") @@ -4961,6 +5412,7 @@ class Program_weight_tensor_parameter_450: class Program_weight_tensor_parameter_451: name = "parameter_451" + original_name = "batch_norm_35.b_0" shape = [192] dtype = "float32" min_val = float("-1.92174") @@ -4972,6 +5424,7 @@ class Program_weight_tensor_parameter_451: class Program_weight_tensor_parameter_452: name = "parameter_452" + original_name = "batch_norm_35.w_0" shape = [192] dtype = "float32" min_val = float("-1.04842") @@ -4983,6 +5436,7 @@ class Program_weight_tensor_parameter_452: class Program_weight_tensor_parameter_453: name = "parameter_453" + original_name = "batch_norm_35.w_2" shape = [192] dtype = "float32" min_val = float("0.00994838") @@ -4994,6 +5448,7 @@ class Program_weight_tensor_parameter_453: class Program_weight_tensor_parameter_454: name = "parameter_454" + original_name = "batch_norm_35.w_1" shape = [192] dtype = "float32" min_val = float("-9.25415") @@ -5005,6 +5460,7 @@ class Program_weight_tensor_parameter_454: class Program_weight_tensor_parameter_455: name = "parameter_455" + original_name = "conv2d_35.w_0" shape = [192, 1, 5, 5] dtype = "float32" min_val = float("-0.543549") @@ -5016,6 +5472,7 @@ class Program_weight_tensor_parameter_455: class Program_weight_tensor_parameter_456: name = "parameter_456" + original_name = "batch_norm_34.b_0" shape = [192] dtype = "float32" min_val = float("-1.33325") @@ -5027,6 +5484,7 @@ class Program_weight_tensor_parameter_456: class Program_weight_tensor_parameter_457: name = "parameter_457" + original_name = "batch_norm_34.w_0" shape = [192] dtype = "float32" min_val = float("-1.41299") @@ -5038,6 +5496,7 @@ class Program_weight_tensor_parameter_457: class Program_weight_tensor_parameter_458: name = "parameter_458" + original_name = "batch_norm_34.w_2" shape = [192] dtype = "float32" min_val = float("4.02859") @@ -5049,6 +5508,7 @@ class Program_weight_tensor_parameter_458: class Program_weight_tensor_parameter_459: name = "parameter_459" + original_name = "batch_norm_34.w_1" shape = [192] dtype = "float32" min_val = float("-10.6764") @@ -5060,6 +5520,7 @@ class Program_weight_tensor_parameter_459: class Program_weight_tensor_parameter_460: name = "parameter_460" + original_name = "conv2d_34.w_0" shape = [192, 768, 1, 1] dtype = "float32" min_val = float("-0.490705") @@ -5071,6 +5532,7 @@ class Program_weight_tensor_parameter_460: class Program_weight_tensor_parameter_461: name = "parameter_461" + original_name = "batch_norm_33.b_0" shape = [768] dtype = "float32" min_val = float("-8.4247") @@ -5082,6 +5544,7 @@ class Program_weight_tensor_parameter_461: class Program_weight_tensor_parameter_462: name = "parameter_462" + original_name = "batch_norm_33.w_0" shape = [768] dtype = "float32" min_val = float("-1.99672") @@ -5093,6 +5556,7 @@ class Program_weight_tensor_parameter_462: class Program_weight_tensor_parameter_463: name = "parameter_463" + original_name = "batch_norm_33.w_2" shape = [768] dtype = "float32" min_val = float("3.44719e-43") @@ -5104,6 +5568,7 @@ class Program_weight_tensor_parameter_463: class Program_weight_tensor_parameter_464: name = "parameter_464" + original_name = "batch_norm_33.w_1" shape = [768] dtype = "float32" min_val = float("-2.15877") @@ -5115,6 +5580,7 @@ class Program_weight_tensor_parameter_464: class Program_weight_tensor_parameter_465: name = "parameter_465" + original_name = "conv2d_33.w_0" shape = [768, 1, 3, 3] dtype = "float32" min_val = float("-0.463297") @@ -5126,6 +5592,7 @@ class Program_weight_tensor_parameter_465: class Program_weight_tensor_parameter_466: name = "parameter_466" + original_name = "batch_norm_32.b_0" shape = [768] dtype = "float32" min_val = float("-4.1458") @@ -5137,6 +5604,7 @@ class Program_weight_tensor_parameter_466: class Program_weight_tensor_parameter_467: name = "parameter_467" + original_name = "batch_norm_32.w_0" shape = [768] dtype = "float32" min_val = float("-0.520692") @@ -5148,6 +5616,7 @@ class Program_weight_tensor_parameter_467: class Program_weight_tensor_parameter_468: name = "parameter_468" + original_name = "batch_norm_32.w_2" shape = [768] dtype = "float32" min_val = float("5.56282e-37") @@ -5159,6 +5628,7 @@ class Program_weight_tensor_parameter_468: class Program_weight_tensor_parameter_469: name = "parameter_469" + original_name = "batch_norm_32.w_1" shape = [768] dtype = "float32" min_val = float("-3.66969") @@ -5170,6 +5640,7 @@ class Program_weight_tensor_parameter_469: class Program_weight_tensor_parameter_470: name = "parameter_470" + original_name = "conv2d_32.w_0" shape = [768, 192, 1, 1] dtype = "float32" min_val = float("-0.492571") @@ -5181,6 +5652,7 @@ class Program_weight_tensor_parameter_470: class Program_weight_tensor_parameter_471: name = "parameter_471" + original_name = "batch_norm_31.b_0" shape = [192] dtype = "float32" min_val = float("-1.99078") @@ -5192,6 +5664,7 @@ class Program_weight_tensor_parameter_471: class Program_weight_tensor_parameter_472: name = "parameter_472" + original_name = "batch_norm_31.w_0" shape = [192] dtype = "float32" min_val = float("-0.160778") @@ -5203,6 +5676,7 @@ class Program_weight_tensor_parameter_472: class Program_weight_tensor_parameter_473: name = "parameter_473" + original_name = "batch_norm_31.w_2" shape = [192] dtype = "float32" min_val = float("0.00493484") @@ -5214,6 +5688,7 @@ class Program_weight_tensor_parameter_473: class Program_weight_tensor_parameter_474: name = "parameter_474" + original_name = "batch_norm_31.w_1" shape = [192] dtype = "float32" min_val = float("-5.43589") @@ -5225,6 +5700,7 @@ class Program_weight_tensor_parameter_474: class Program_weight_tensor_parameter_475: name = "parameter_475" + original_name = "conv2d_31.w_0" shape = [192, 1, 5, 5] dtype = "float32" min_val = float("-0.599106") @@ -5236,6 +5712,7 @@ class Program_weight_tensor_parameter_475: class Program_weight_tensor_parameter_476: name = "parameter_476" + original_name = "batch_norm_30.b_0" shape = [192] dtype = "float32" min_val = float("-1.1603") @@ -5247,6 +5724,7 @@ class Program_weight_tensor_parameter_476: class Program_weight_tensor_parameter_477: name = "parameter_477" + original_name = "batch_norm_30.w_0" shape = [192] dtype = "float32" min_val = float("-0.940788") @@ -5258,6 +5736,7 @@ class Program_weight_tensor_parameter_477: class Program_weight_tensor_parameter_478: name = "parameter_478" + original_name = "batch_norm_30.w_2" shape = [192] dtype = "float32" min_val = float("2.16801") @@ -5269,6 +5748,7 @@ class Program_weight_tensor_parameter_478: class Program_weight_tensor_parameter_479: name = "parameter_479" + original_name = "batch_norm_30.w_1" shape = [192] dtype = "float32" min_val = float("-5.925") @@ -5280,6 +5760,7 @@ class Program_weight_tensor_parameter_479: class Program_weight_tensor_parameter_480: name = "parameter_480" + original_name = "conv2d_30.w_0" shape = [192, 768, 1, 1] dtype = "float32" min_val = float("-0.545164") @@ -5291,6 +5772,7 @@ class Program_weight_tensor_parameter_480: class Program_weight_tensor_parameter_481: name = "parameter_481" + original_name = "batch_norm_29.b_0" shape = [768] dtype = "float32" min_val = float("-12.2198") @@ -5302,6 +5784,7 @@ class Program_weight_tensor_parameter_481: class Program_weight_tensor_parameter_482: name = "parameter_482" + original_name = "batch_norm_29.w_0" shape = [768] dtype = "float32" min_val = float("-0.529818") @@ -5313,6 +5796,7 @@ class Program_weight_tensor_parameter_482: class Program_weight_tensor_parameter_483: name = "parameter_483" + original_name = "batch_norm_29.w_2" shape = [768] dtype = "float32" min_val = float("7.85072e-25") @@ -5324,6 +5808,7 @@ class Program_weight_tensor_parameter_483: class Program_weight_tensor_parameter_484: name = "parameter_484" + original_name = "batch_norm_29.w_1" shape = [768] dtype = "float32" min_val = float("-4.21349") @@ -5335,6 +5820,7 @@ class Program_weight_tensor_parameter_484: class Program_weight_tensor_parameter_485: name = "parameter_485" + original_name = "conv2d_29.w_0" shape = [768, 1, 5, 5] dtype = "float32" min_val = float("-0.45907") @@ -5346,6 +5832,7 @@ class Program_weight_tensor_parameter_485: class Program_weight_tensor_parameter_486: name = "parameter_486" + original_name = "batch_norm_28.b_0" shape = [768] dtype = "float32" min_val = float("-5.38858") @@ -5357,6 +5844,7 @@ class Program_weight_tensor_parameter_486: class Program_weight_tensor_parameter_487: name = "parameter_487" + original_name = "batch_norm_28.w_0" shape = [768] dtype = "float32" min_val = float("-0.296403") @@ -5368,6 +5856,7 @@ class Program_weight_tensor_parameter_487: class Program_weight_tensor_parameter_488: name = "parameter_488" + original_name = "batch_norm_28.w_2" shape = [768] dtype = "float32" min_val = float("1.20145e-19") @@ -5379,6 +5868,7 @@ class Program_weight_tensor_parameter_488: class Program_weight_tensor_parameter_489: name = "parameter_489" + original_name = "batch_norm_28.w_1" shape = [768] dtype = "float32" min_val = float("-3.00182") @@ -5390,6 +5880,7 @@ class Program_weight_tensor_parameter_489: class Program_weight_tensor_parameter_490: name = "parameter_490" + original_name = "conv2d_28.w_0" shape = [768, 192, 1, 1] dtype = "float32" min_val = float("-0.518669") @@ -5401,6 +5892,7 @@ class Program_weight_tensor_parameter_490: class Program_weight_tensor_parameter_491: name = "parameter_491" + original_name = "batch_norm_27.b_0" shape = [192] dtype = "float32" min_val = float("-2.27634") @@ -5412,6 +5904,7 @@ class Program_weight_tensor_parameter_491: class Program_weight_tensor_parameter_492: name = "parameter_492" + original_name = "batch_norm_27.w_0" shape = [192] dtype = "float32" min_val = float("-0.347139") @@ -5423,6 +5916,7 @@ class Program_weight_tensor_parameter_492: class Program_weight_tensor_parameter_493: name = "parameter_493" + original_name = "batch_norm_27.w_2" shape = [192] dtype = "float32" min_val = float("1.7587e-05") @@ -5434,6 +5928,7 @@ class Program_weight_tensor_parameter_493: class Program_weight_tensor_parameter_494: name = "parameter_494" + original_name = "batch_norm_27.w_1" shape = [192] dtype = "float32" min_val = float("-1.85069") @@ -5445,6 +5940,7 @@ class Program_weight_tensor_parameter_494: class Program_weight_tensor_parameter_495: name = "parameter_495" + original_name = "conv2d_27.w_0" shape = [192, 1, 3, 3] dtype = "float32" min_val = float("-0.43043") @@ -5456,6 +5952,7 @@ class Program_weight_tensor_parameter_495: class Program_weight_tensor_parameter_496: name = "parameter_496" + original_name = "batch_norm_26.b_0" shape = [192] dtype = "float32" min_val = float("-1.21364") @@ -5467,6 +5964,7 @@ class Program_weight_tensor_parameter_496: class Program_weight_tensor_parameter_497: name = "parameter_497" + original_name = "batch_norm_26.w_0" shape = [192] dtype = "float32" min_val = float("-0.924213") @@ -5478,6 +5976,7 @@ class Program_weight_tensor_parameter_497: class Program_weight_tensor_parameter_498: name = "parameter_498" + original_name = "batch_norm_26.w_2" shape = [192] dtype = "float32" min_val = float("2.17548") @@ -5489,6 +5988,7 @@ class Program_weight_tensor_parameter_498: class Program_weight_tensor_parameter_499: name = "parameter_499" + original_name = "batch_norm_26.w_1" shape = [192] dtype = "float32" min_val = float("-9.06843") @@ -5500,6 +6000,7 @@ class Program_weight_tensor_parameter_499: class Program_weight_tensor_parameter_500: name = "parameter_500" + original_name = "conv2d_26.w_0" shape = [192, 768, 1, 1] dtype = "float32" min_val = float("-0.498693") @@ -5511,6 +6012,7 @@ class Program_weight_tensor_parameter_500: class Program_weight_tensor_parameter_501: name = "parameter_501" + original_name = "batch_norm_25.b_0" shape = [768] dtype = "float32" min_val = float("-6.55987") @@ -5522,6 +6024,7 @@ class Program_weight_tensor_parameter_501: class Program_weight_tensor_parameter_502: name = "parameter_502" + original_name = "batch_norm_25.w_0" shape = [768] dtype = "float32" min_val = float("-0.13457") @@ -5533,6 +6036,7 @@ class Program_weight_tensor_parameter_502: class Program_weight_tensor_parameter_503: name = "parameter_503" + original_name = "batch_norm_25.w_2" shape = [768] dtype = "float32" min_val = float("6.36246e-23") @@ -5544,6 +6048,7 @@ class Program_weight_tensor_parameter_503: class Program_weight_tensor_parameter_504: name = "parameter_504" + original_name = "batch_norm_25.w_1" shape = [768] dtype = "float32" min_val = float("-1.57153") @@ -5555,6 +6060,7 @@ class Program_weight_tensor_parameter_504: class Program_weight_tensor_parameter_505: name = "parameter_505" + original_name = "conv2d_25.w_0" shape = [768, 1, 3, 3] dtype = "float32" min_val = float("-0.424978") @@ -5566,6 +6072,7 @@ class Program_weight_tensor_parameter_505: class Program_weight_tensor_parameter_506: name = "parameter_506" + original_name = "batch_norm_24.b_0" shape = [768] dtype = "float32" min_val = float("-2.73652") @@ -5577,6 +6084,7 @@ class Program_weight_tensor_parameter_506: class Program_weight_tensor_parameter_507: name = "parameter_507" + original_name = "batch_norm_24.w_0" shape = [768] dtype = "float32" min_val = float("-0.211764") @@ -5588,6 +6096,7 @@ class Program_weight_tensor_parameter_507: class Program_weight_tensor_parameter_508: name = "parameter_508" + original_name = "batch_norm_24.w_2" shape = [768] dtype = "float32" min_val = float("7.27395e-28") @@ -5599,6 +6108,7 @@ class Program_weight_tensor_parameter_508: class Program_weight_tensor_parameter_509: name = "parameter_509" + original_name = "batch_norm_24.w_1" shape = [768] dtype = "float32" min_val = float("-3.04354") @@ -5610,6 +6120,7 @@ class Program_weight_tensor_parameter_509: class Program_weight_tensor_parameter_510: name = "parameter_510" + original_name = "conv2d_24.w_0" shape = [768, 192, 1, 1] dtype = "float32" min_val = float("-0.527209") @@ -5621,6 +6132,7 @@ class Program_weight_tensor_parameter_510: class Program_weight_tensor_parameter_511: name = "parameter_511" + original_name = "batch_norm_23.b_0" shape = [192] dtype = "float32" min_val = float("-2.11347") @@ -5632,6 +6144,7 @@ class Program_weight_tensor_parameter_511: class Program_weight_tensor_parameter_512: name = "parameter_512" + original_name = "batch_norm_23.w_0" shape = [192] dtype = "float32" min_val = float("-0.542576") @@ -5643,6 +6156,7 @@ class Program_weight_tensor_parameter_512: class Program_weight_tensor_parameter_513: name = "parameter_513" + original_name = "batch_norm_23.w_2" shape = [192] dtype = "float32" min_val = float("1.95379e-05") @@ -5654,6 +6168,7 @@ class Program_weight_tensor_parameter_513: class Program_weight_tensor_parameter_514: name = "parameter_514" + original_name = "batch_norm_23.w_1" shape = [192] dtype = "float32" min_val = float("-2.05102") @@ -5665,6 +6180,7 @@ class Program_weight_tensor_parameter_514: class Program_weight_tensor_parameter_515: name = "parameter_515" + original_name = "conv2d_23.w_0" shape = [192, 1, 3, 3] dtype = "float32" min_val = float("-0.423914") @@ -5676,6 +6192,7 @@ class Program_weight_tensor_parameter_515: class Program_weight_tensor_parameter_516: name = "parameter_516" + original_name = "batch_norm_22.b_0" shape = [192] dtype = "float32" min_val = float("-1.6225") @@ -5687,6 +6204,7 @@ class Program_weight_tensor_parameter_516: class Program_weight_tensor_parameter_517: name = "parameter_517" + original_name = "batch_norm_22.w_0" shape = [192] dtype = "float32" min_val = float("-0.310347") @@ -5698,6 +6216,7 @@ class Program_weight_tensor_parameter_517: class Program_weight_tensor_parameter_518: name = "parameter_518" + original_name = "batch_norm_22.w_2" shape = [192] dtype = "float32" min_val = float("0.570746") @@ -5709,6 +6228,7 @@ class Program_weight_tensor_parameter_518: class Program_weight_tensor_parameter_519: name = "parameter_519" + original_name = "batch_norm_22.w_1" shape = [192] dtype = "float32" min_val = float("-6.54649") @@ -5720,6 +6240,7 @@ class Program_weight_tensor_parameter_519: class Program_weight_tensor_parameter_520: name = "parameter_520" + original_name = "conv2d_22.w_0" shape = [192, 768, 1, 1] dtype = "float32" min_val = float("-0.511516") @@ -5731,6 +6252,7 @@ class Program_weight_tensor_parameter_520: class Program_weight_tensor_parameter_521: name = "parameter_521" + original_name = "batch_norm_21.b_0" shape = [768] dtype = "float32" min_val = float("-5.99127") @@ -5742,6 +6264,7 @@ class Program_weight_tensor_parameter_521: class Program_weight_tensor_parameter_522: name = "parameter_522" + original_name = "batch_norm_21.w_0" shape = [768] dtype = "float32" min_val = float("-1.23483") @@ -5753,6 +6276,7 @@ class Program_weight_tensor_parameter_522: class Program_weight_tensor_parameter_523: name = "parameter_523" + original_name = "batch_norm_21.w_2" shape = [768] dtype = "float32" min_val = float("1.91319e-41") @@ -5764,6 +6288,7 @@ class Program_weight_tensor_parameter_523: class Program_weight_tensor_parameter_524: name = "parameter_524" + original_name = "batch_norm_21.w_1" shape = [768] dtype = "float32" min_val = float("-1.65103") @@ -5775,6 +6300,7 @@ class Program_weight_tensor_parameter_524: class Program_weight_tensor_parameter_525: name = "parameter_525" + original_name = "conv2d_21.w_0" shape = [768, 1, 3, 3] dtype = "float32" min_val = float("-0.5176") @@ -5786,6 +6312,7 @@ class Program_weight_tensor_parameter_525: class Program_weight_tensor_parameter_526: name = "parameter_526" + original_name = "batch_norm_20.b_0" shape = [768] dtype = "float32" min_val = float("-3.58609") @@ -5797,6 +6324,7 @@ class Program_weight_tensor_parameter_526: class Program_weight_tensor_parameter_527: name = "parameter_527" + original_name = "batch_norm_20.w_0" shape = [768] dtype = "float32" min_val = float("-0.25698") @@ -5808,6 +6336,7 @@ class Program_weight_tensor_parameter_527: class Program_weight_tensor_parameter_528: name = "parameter_528" + original_name = "batch_norm_20.w_2" shape = [768] dtype = "float32" min_val = float("7.64536e-38") @@ -5819,6 +6348,7 @@ class Program_weight_tensor_parameter_528: class Program_weight_tensor_parameter_529: name = "parameter_529" + original_name = "batch_norm_20.w_1" shape = [768] dtype = "float32" min_val = float("-3.77586") @@ -5830,6 +6360,7 @@ class Program_weight_tensor_parameter_529: class Program_weight_tensor_parameter_530: name = "parameter_530" + original_name = "conv2d_20.w_0" shape = [768, 192, 1, 1] dtype = "float32" min_val = float("-0.521256") @@ -5841,6 +6372,7 @@ class Program_weight_tensor_parameter_530: class Program_weight_tensor_parameter_531: name = "parameter_531" + original_name = "batch_norm_19.b_0" shape = [192] dtype = "float32" min_val = float("-2.21649") @@ -5852,6 +6384,7 @@ class Program_weight_tensor_parameter_531: class Program_weight_tensor_parameter_532: name = "parameter_532" + original_name = "batch_norm_19.w_0" shape = [192] dtype = "float32" min_val = float("-2.86445") @@ -5863,6 +6396,7 @@ class Program_weight_tensor_parameter_532: class Program_weight_tensor_parameter_533: name = "parameter_533" + original_name = "batch_norm_19.w_2" shape = [192] dtype = "float32" min_val = float("5.1687e-06") @@ -5874,6 +6408,7 @@ class Program_weight_tensor_parameter_533: class Program_weight_tensor_parameter_534: name = "parameter_534" + original_name = "batch_norm_19.w_1" shape = [192] dtype = "float32" min_val = float("-3.25287") @@ -5885,6 +6420,7 @@ class Program_weight_tensor_parameter_534: class Program_weight_tensor_parameter_535: name = "parameter_535" + original_name = "conv2d_19.w_0" shape = [192, 1, 3, 3] dtype = "float32" min_val = float("-0.477088") @@ -5896,6 +6432,7 @@ class Program_weight_tensor_parameter_535: class Program_weight_tensor_parameter_536: name = "parameter_536" + original_name = "batch_norm_18.b_0" shape = [192] dtype = "float32" min_val = float("-1.84585") @@ -5907,6 +6444,7 @@ class Program_weight_tensor_parameter_536: class Program_weight_tensor_parameter_537: name = "parameter_537" + original_name = "batch_norm_18.w_0" shape = [192] dtype = "float32" min_val = float("-0.438473") @@ -5918,6 +6456,7 @@ class Program_weight_tensor_parameter_537: class Program_weight_tensor_parameter_538: name = "parameter_538" + original_name = "batch_norm_18.w_2" shape = [192] dtype = "float32" min_val = float("1.0467") @@ -5929,6 +6468,7 @@ class Program_weight_tensor_parameter_538: class Program_weight_tensor_parameter_539: name = "parameter_539" + original_name = "batch_norm_18.w_1" shape = [192] dtype = "float32" min_val = float("-7.51575") @@ -5940,6 +6480,7 @@ class Program_weight_tensor_parameter_539: class Program_weight_tensor_parameter_540: name = "parameter_540" + original_name = "conv2d_18.w_0" shape = [192, 768, 1, 1] dtype = "float32" min_val = float("-0.496174") @@ -5951,6 +6492,7 @@ class Program_weight_tensor_parameter_540: class Program_weight_tensor_parameter_541: name = "parameter_541" + original_name = "batch_norm_17.b_0" shape = [768] dtype = "float32" min_val = float("-5.24263") @@ -5962,6 +6504,7 @@ class Program_weight_tensor_parameter_541: class Program_weight_tensor_parameter_542: name = "parameter_542" + original_name = "batch_norm_17.w_0" shape = [768] dtype = "float32" min_val = float("-0.331109") @@ -5973,6 +6516,7 @@ class Program_weight_tensor_parameter_542: class Program_weight_tensor_parameter_543: name = "parameter_543" + original_name = "batch_norm_17.w_2" shape = [768] dtype = "float32" min_val = float("3.21712e-06") @@ -5984,6 +6528,7 @@ class Program_weight_tensor_parameter_543: class Program_weight_tensor_parameter_544: name = "parameter_544" + original_name = "batch_norm_17.w_1" shape = [768] dtype = "float32" min_val = float("-1.96805") @@ -5995,6 +6540,7 @@ class Program_weight_tensor_parameter_544: class Program_weight_tensor_parameter_545: name = "parameter_545" + original_name = "conv2d_17.w_0" shape = [768, 1, 3, 3] dtype = "float32" min_val = float("-0.485981") @@ -6006,6 +6552,7 @@ class Program_weight_tensor_parameter_545: class Program_weight_tensor_parameter_546: name = "parameter_546" + original_name = "batch_norm_16.b_0" shape = [768] dtype = "float32" min_val = float("-4.16252") @@ -6017,6 +6564,7 @@ class Program_weight_tensor_parameter_546: class Program_weight_tensor_parameter_547: name = "parameter_547" + original_name = "batch_norm_16.w_0" shape = [768] dtype = "float32" min_val = float("-0.282117") @@ -6028,6 +6576,7 @@ class Program_weight_tensor_parameter_547: class Program_weight_tensor_parameter_548: name = "parameter_548" + original_name = "batch_norm_16.w_2" shape = [768] dtype = "float32" min_val = float("2.39344") @@ -6039,6 +6588,7 @@ class Program_weight_tensor_parameter_548: class Program_weight_tensor_parameter_549: name = "parameter_549" + original_name = "batch_norm_16.w_1" shape = [768] dtype = "float32" min_val = float("-3.07932") @@ -6050,6 +6600,7 @@ class Program_weight_tensor_parameter_549: class Program_weight_tensor_parameter_550: name = "parameter_550" + original_name = "conv2d_16.w_0" shape = [768, 192, 1, 1] dtype = "float32" min_val = float("-0.496565") @@ -6061,6 +6612,7 @@ class Program_weight_tensor_parameter_550: class Program_weight_tensor_parameter_551: name = "parameter_551" + original_name = "batch_norm_15.b_0" shape = [192] dtype = "float32" min_val = float("-1.83653") @@ -6072,6 +6624,7 @@ class Program_weight_tensor_parameter_551: class Program_weight_tensor_parameter_552: name = "parameter_552" + original_name = "batch_norm_15.w_0" shape = [192] dtype = "float32" min_val = float("-0.122656") @@ -6083,6 +6636,7 @@ class Program_weight_tensor_parameter_552: class Program_weight_tensor_parameter_553: name = "parameter_553" + original_name = "batch_norm_15.w_2" shape = [192] dtype = "float32" min_val = float("6.31298e-07") @@ -6094,6 +6648,7 @@ class Program_weight_tensor_parameter_553: class Program_weight_tensor_parameter_554: name = "parameter_554" + original_name = "batch_norm_15.w_1" shape = [192] dtype = "float32" min_val = float("-1.89782") @@ -6105,6 +6660,7 @@ class Program_weight_tensor_parameter_554: class Program_weight_tensor_parameter_555: name = "parameter_555" + original_name = "conv2d_15.w_0" shape = [192, 1, 3, 3] dtype = "float32" min_val = float("-0.424793") @@ -6116,6 +6672,7 @@ class Program_weight_tensor_parameter_555: class Program_weight_tensor_parameter_556: name = "parameter_556" + original_name = "batch_norm_14.b_0" shape = [192] dtype = "float32" min_val = float("-1.82409") @@ -6127,6 +6684,7 @@ class Program_weight_tensor_parameter_556: class Program_weight_tensor_parameter_557: name = "parameter_557" + original_name = "batch_norm_14.w_0" shape = [192] dtype = "float32" min_val = float("-0.199537") @@ -6138,6 +6696,7 @@ class Program_weight_tensor_parameter_557: class Program_weight_tensor_parameter_558: name = "parameter_558" + original_name = "batch_norm_14.w_2" shape = [192] dtype = "float32" min_val = float("2.88966") @@ -6149,6 +6708,7 @@ class Program_weight_tensor_parameter_558: class Program_weight_tensor_parameter_559: name = "parameter_559" + original_name = "batch_norm_14.w_1" shape = [192] dtype = "float32" min_val = float("-14.5398") @@ -6160,6 +6720,7 @@ class Program_weight_tensor_parameter_559: class Program_weight_tensor_parameter_560: name = "parameter_560" + original_name = "conv2d_14.w_0" shape = [192, 384, 1, 1] dtype = "float32" min_val = float("-0.580096") @@ -6171,6 +6732,7 @@ class Program_weight_tensor_parameter_560: class Program_weight_tensor_parameter_561: name = "parameter_561" + original_name = "batch_norm_13.b_0" shape = [384] dtype = "float32" min_val = float("-1.11562") @@ -6182,6 +6744,7 @@ class Program_weight_tensor_parameter_561: class Program_weight_tensor_parameter_562: name = "parameter_562" + original_name = "batch_norm_13.w_0" shape = [384] dtype = "float32" min_val = float("-0.00346239") @@ -6193,6 +6756,7 @@ class Program_weight_tensor_parameter_562: class Program_weight_tensor_parameter_563: name = "parameter_563" + original_name = "batch_norm_13.w_2" shape = [384] dtype = "float32" min_val = float("0.000217224") @@ -6204,6 +6768,7 @@ class Program_weight_tensor_parameter_563: class Program_weight_tensor_parameter_564: name = "parameter_564" + original_name = "batch_norm_13.w_1" shape = [384] dtype = "float32" min_val = float("-4.41873") @@ -6215,6 +6780,7 @@ class Program_weight_tensor_parameter_564: class Program_weight_tensor_parameter_565: name = "parameter_565" + original_name = "conv2d_13.w_0" shape = [384, 1, 5, 5] dtype = "float32" min_val = float("-0.487166") @@ -6226,6 +6792,7 @@ class Program_weight_tensor_parameter_565: class Program_weight_tensor_parameter_566: name = "parameter_566" + original_name = "batch_norm_12.b_0" shape = [384] dtype = "float32" min_val = float("-1.95041") @@ -6237,6 +6804,7 @@ class Program_weight_tensor_parameter_566: class Program_weight_tensor_parameter_567: name = "parameter_567" + original_name = "batch_norm_12.w_0" shape = [384] dtype = "float32" min_val = float("-0.245005") @@ -6248,6 +6816,7 @@ class Program_weight_tensor_parameter_567: class Program_weight_tensor_parameter_568: name = "parameter_568" + original_name = "batch_norm_12.w_2" shape = [384] dtype = "float32" min_val = float("1.67049") @@ -6259,6 +6828,7 @@ class Program_weight_tensor_parameter_568: class Program_weight_tensor_parameter_569: name = "parameter_569" + original_name = "batch_norm_12.w_1" shape = [384] dtype = "float32" min_val = float("-2.07271") @@ -6270,6 +6840,7 @@ class Program_weight_tensor_parameter_569: class Program_weight_tensor_parameter_570: name = "parameter_570" + original_name = "conv2d_12.w_0" shape = [384, 96, 1, 1] dtype = "float32" min_val = float("-0.496373") @@ -6281,6 +6852,7 @@ class Program_weight_tensor_parameter_570: class Program_weight_tensor_parameter_571: name = "parameter_571" + original_name = "batch_norm_11.b_0" shape = [96] dtype = "float32" min_val = float("-1.8081") @@ -6292,6 +6864,7 @@ class Program_weight_tensor_parameter_571: class Program_weight_tensor_parameter_572: name = "parameter_572" + original_name = "batch_norm_11.w_0" shape = [96] dtype = "float32" min_val = float("0.205714") @@ -6303,6 +6876,7 @@ class Program_weight_tensor_parameter_572: class Program_weight_tensor_parameter_573: name = "parameter_573" + original_name = "batch_norm_11.w_2" shape = [96] dtype = "float32" min_val = float("0.081127") @@ -6314,6 +6888,7 @@ class Program_weight_tensor_parameter_573: class Program_weight_tensor_parameter_574: name = "parameter_574" + original_name = "batch_norm_11.w_1" shape = [96] dtype = "float32" min_val = float("-3.92299") @@ -6325,6 +6900,7 @@ class Program_weight_tensor_parameter_574: class Program_weight_tensor_parameter_575: name = "parameter_575" + original_name = "conv2d_11.w_0" shape = [96, 1, 3, 3] dtype = "float32" min_val = float("-0.384673") @@ -6336,6 +6912,7 @@ class Program_weight_tensor_parameter_575: class Program_weight_tensor_parameter_576: name = "parameter_576" + original_name = "batch_norm_10.b_0" shape = [96] dtype = "float32" min_val = float("-1.54719") @@ -6347,6 +6924,7 @@ class Program_weight_tensor_parameter_576: class Program_weight_tensor_parameter_577: name = "parameter_577" + original_name = "batch_norm_10.w_0" shape = [96] dtype = "float32" min_val = float("-0.250768") @@ -6358,6 +6936,7 @@ class Program_weight_tensor_parameter_577: class Program_weight_tensor_parameter_578: name = "parameter_578" + original_name = "batch_norm_10.w_2" shape = [96] dtype = "float32" min_val = float("3.4037") @@ -6369,6 +6948,7 @@ class Program_weight_tensor_parameter_578: class Program_weight_tensor_parameter_579: name = "parameter_579" + original_name = "batch_norm_10.w_1" shape = [96] dtype = "float32" min_val = float("-3.66091") @@ -6380,6 +6960,7 @@ class Program_weight_tensor_parameter_579: class Program_weight_tensor_parameter_580: name = "parameter_580" + original_name = "conv2d_10.w_0" shape = [96, 384, 1, 1] dtype = "float32" min_val = float("-0.472517") @@ -6391,6 +6972,7 @@ class Program_weight_tensor_parameter_580: class Program_weight_tensor_parameter_581: name = "parameter_581" + original_name = "batch_norm_9.b_0" shape = [384] dtype = "float32" min_val = float("-7.35909") @@ -6402,6 +6984,7 @@ class Program_weight_tensor_parameter_581: class Program_weight_tensor_parameter_582: name = "parameter_582" + original_name = "batch_norm_9.w_0" shape = [384] dtype = "float32" min_val = float("0.112189") @@ -6413,6 +6996,7 @@ class Program_weight_tensor_parameter_582: class Program_weight_tensor_parameter_583: name = "parameter_583" + original_name = "batch_norm_9.w_2" shape = [384] dtype = "float32" min_val = float("0.000130086") @@ -6424,6 +7008,7 @@ class Program_weight_tensor_parameter_583: class Program_weight_tensor_parameter_584: name = "parameter_584" + original_name = "batch_norm_9.w_1" shape = [384] dtype = "float32" min_val = float("-2.20332") @@ -6435,6 +7020,7 @@ class Program_weight_tensor_parameter_584: class Program_weight_tensor_parameter_585: name = "parameter_585" + original_name = "conv2d_9.w_0" shape = [384, 1, 3, 3] dtype = "float32" min_val = float("-0.443053") @@ -6446,6 +7032,7 @@ class Program_weight_tensor_parameter_585: class Program_weight_tensor_parameter_586: name = "parameter_586" + original_name = "batch_norm_8.b_0" shape = [384] dtype = "float32" min_val = float("-3.35246") @@ -6457,6 +7044,7 @@ class Program_weight_tensor_parameter_586: class Program_weight_tensor_parameter_587: name = "parameter_587" + original_name = "batch_norm_8.w_0" shape = [384] dtype = "float32" min_val = float("-0.223193") @@ -6468,6 +7056,7 @@ class Program_weight_tensor_parameter_587: class Program_weight_tensor_parameter_588: name = "parameter_588" + original_name = "batch_norm_8.w_2" shape = [384] dtype = "float32" min_val = float("1.22727") @@ -6479,6 +7068,7 @@ class Program_weight_tensor_parameter_588: class Program_weight_tensor_parameter_589: name = "parameter_589" + original_name = "batch_norm_8.w_1" shape = [384] dtype = "float32" min_val = float("-2.52434") @@ -6490,6 +7080,7 @@ class Program_weight_tensor_parameter_589: class Program_weight_tensor_parameter_590: name = "parameter_590" + original_name = "conv2d_8.w_0" shape = [384, 96, 1, 1] dtype = "float32" min_val = float("-0.449402") @@ -6501,6 +7092,7 @@ class Program_weight_tensor_parameter_590: class Program_weight_tensor_parameter_591: name = "parameter_591" + original_name = "batch_norm_7.b_0" shape = [96] dtype = "float32" min_val = float("-1.74517") @@ -6512,6 +7104,7 @@ class Program_weight_tensor_parameter_591: class Program_weight_tensor_parameter_592: name = "parameter_592" + original_name = "batch_norm_7.w_0" shape = [96] dtype = "float32" min_val = float("-0.00521074") @@ -6523,6 +7116,7 @@ class Program_weight_tensor_parameter_592: class Program_weight_tensor_parameter_593: name = "parameter_593" + original_name = "batch_norm_7.w_2" shape = [96] dtype = "float32" min_val = float("0.00123213") @@ -6534,6 +7128,7 @@ class Program_weight_tensor_parameter_593: class Program_weight_tensor_parameter_594: name = "parameter_594" + original_name = "batch_norm_7.w_1" shape = [96] dtype = "float32" min_val = float("-0.90895") @@ -6545,6 +7140,7 @@ class Program_weight_tensor_parameter_594: class Program_weight_tensor_parameter_595: name = "parameter_595" + original_name = "conv2d_7.w_0" shape = [96, 1, 3, 3] dtype = "float32" min_val = float("-0.386994") @@ -6556,6 +7152,7 @@ class Program_weight_tensor_parameter_595: class Program_weight_tensor_parameter_596: name = "parameter_596" + original_name = "batch_norm_6.b_0" shape = [96] dtype = "float32" min_val = float("-2.74739") @@ -6567,6 +7164,7 @@ class Program_weight_tensor_parameter_596: class Program_weight_tensor_parameter_597: name = "parameter_597" + original_name = "batch_norm_6.w_0" shape = [96] dtype = "float32" min_val = float("0.433307") @@ -6578,6 +7176,7 @@ class Program_weight_tensor_parameter_597: class Program_weight_tensor_parameter_598: name = "parameter_598" + original_name = "batch_norm_6.w_2" shape = [96] dtype = "float32" min_val = float("2.96416") @@ -6589,6 +7188,7 @@ class Program_weight_tensor_parameter_598: class Program_weight_tensor_parameter_599: name = "parameter_599" + original_name = "batch_norm_6.w_1" shape = [96] dtype = "float32" min_val = float("-10.1096") @@ -6600,6 +7200,7 @@ class Program_weight_tensor_parameter_599: class Program_weight_tensor_parameter_600: name = "parameter_600" + original_name = "conv2d_6.w_0" shape = [96, 192, 1, 1] dtype = "float32" min_val = float("-0.459703") @@ -6611,6 +7212,7 @@ class Program_weight_tensor_parameter_600: class Program_weight_tensor_parameter_601: name = "parameter_601" + original_name = "batch_norm_5.b_0" shape = [192] dtype = "float32" min_val = float("-4.75128") @@ -6622,6 +7224,7 @@ class Program_weight_tensor_parameter_601: class Program_weight_tensor_parameter_602: name = "parameter_602" + original_name = "batch_norm_5.w_0" shape = [192] dtype = "float32" min_val = float("0.147693") @@ -6633,6 +7236,7 @@ class Program_weight_tensor_parameter_602: class Program_weight_tensor_parameter_603: name = "parameter_603" + original_name = "batch_norm_5.w_2" shape = [192] dtype = "float32" min_val = float("0.00304562") @@ -6644,6 +7248,7 @@ class Program_weight_tensor_parameter_603: class Program_weight_tensor_parameter_604: name = "parameter_604" + original_name = "batch_norm_5.w_1" shape = [192] dtype = "float32" min_val = float("-10.2496") @@ -6655,6 +7260,7 @@ class Program_weight_tensor_parameter_604: class Program_weight_tensor_parameter_605: name = "parameter_605" + original_name = "conv2d_5.w_0" shape = [192, 1, 5, 5] dtype = "float32" min_val = float("-0.466539") @@ -6666,6 +7272,7 @@ class Program_weight_tensor_parameter_605: class Program_weight_tensor_parameter_606: name = "parameter_606" + original_name = "batch_norm_4.b_0" shape = [192] dtype = "float32" min_val = float("-3.36419") @@ -6677,6 +7284,7 @@ class Program_weight_tensor_parameter_606: class Program_weight_tensor_parameter_607: name = "parameter_607" + original_name = "batch_norm_4.w_0" shape = [192] dtype = "float32" min_val = float("0.0623472") @@ -6688,6 +7296,7 @@ class Program_weight_tensor_parameter_607: class Program_weight_tensor_parameter_608: name = "parameter_608" + original_name = "batch_norm_4.w_2" shape = [192] dtype = "float32" min_val = float("0.722838") @@ -6699,6 +7308,7 @@ class Program_weight_tensor_parameter_608: class Program_weight_tensor_parameter_609: name = "parameter_609" + original_name = "batch_norm_4.w_1" shape = [192] dtype = "float32" min_val = float("-1.59005") @@ -6710,6 +7320,7 @@ class Program_weight_tensor_parameter_609: class Program_weight_tensor_parameter_610: name = "parameter_610" + original_name = "conv2d_4.w_0" shape = [192, 48, 1, 1] dtype = "float32" min_val = float("-0.895115") @@ -6721,6 +7332,7 @@ class Program_weight_tensor_parameter_610: class Program_weight_tensor_parameter_611: name = "parameter_611" + original_name = "batch_norm_3.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -6730,6 +7342,7 @@ class Program_weight_tensor_parameter_611: class Program_weight_tensor_parameter_612: name = "parameter_612" + original_name = "batch_norm_3.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -6739,6 +7352,7 @@ class Program_weight_tensor_parameter_612: class Program_weight_tensor_parameter_613: name = "parameter_613" + original_name = "batch_norm_3.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -6748,6 +7362,7 @@ class Program_weight_tensor_parameter_613: class Program_weight_tensor_parameter_614: name = "parameter_614" + original_name = "batch_norm_3.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -6757,6 +7372,7 @@ class Program_weight_tensor_parameter_614: class Program_weight_tensor_parameter_615: name = "parameter_615" + original_name = "conv2d_3.w_0" shape = [48, 1, 3, 3] dtype = "float32" min_val = float("-0.422758") @@ -6768,6 +7384,7 @@ class Program_weight_tensor_parameter_615: class Program_weight_tensor_parameter_616: name = "parameter_616" + original_name = "batch_norm_2.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -6777,6 +7394,7 @@ class Program_weight_tensor_parameter_616: class Program_weight_tensor_parameter_617: name = "parameter_617" + original_name = "batch_norm_2.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -6786,6 +7404,7 @@ class Program_weight_tensor_parameter_617: class Program_weight_tensor_parameter_618: name = "parameter_618" + original_name = "batch_norm_2.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -6795,6 +7414,7 @@ class Program_weight_tensor_parameter_618: class Program_weight_tensor_parameter_619: name = "parameter_619" + original_name = "batch_norm_2.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -6804,6 +7424,7 @@ class Program_weight_tensor_parameter_619: class Program_weight_tensor_parameter_620: name = "parameter_620" + original_name = "conv2d_2.w_0" shape = [48, 96, 1, 1] dtype = "float32" min_val = float("-0.426748") @@ -6815,6 +7436,7 @@ class Program_weight_tensor_parameter_620: class Program_weight_tensor_parameter_621: name = "parameter_621" + original_name = "batch_norm_1.b_0" shape = [96] dtype = "float32" min_val = float("-3.70158") @@ -6826,6 +7448,7 @@ class Program_weight_tensor_parameter_621: class Program_weight_tensor_parameter_622: name = "parameter_622" + original_name = "batch_norm_1.w_0" shape = [96] dtype = "float32" min_val = float("1.88944") @@ -6837,6 +7460,7 @@ class Program_weight_tensor_parameter_622: class Program_weight_tensor_parameter_623: name = "parameter_623" + original_name = "batch_norm_1.w_2" shape = [96] dtype = "float32" min_val = float("23.2183") @@ -6848,6 +7472,7 @@ class Program_weight_tensor_parameter_623: class Program_weight_tensor_parameter_624: name = "parameter_624" + original_name = "batch_norm_1.w_1" shape = [96] dtype = "float32" min_val = float("-19.4311") @@ -6859,6 +7484,7 @@ class Program_weight_tensor_parameter_624: class Program_weight_tensor_parameter_625: name = "parameter_625" + original_name = "conv2d_1.w_0" shape = [96, 24, 3, 3] dtype = "float32" min_val = float("-0.499336") @@ -6870,6 +7496,7 @@ class Program_weight_tensor_parameter_625: class Program_weight_tensor_parameter_626: name = "parameter_626" + original_name = "batch_norm_0.b_0" shape = [24] dtype = "float32" min_val = float("0") @@ -6879,6 +7506,7 @@ class Program_weight_tensor_parameter_626: class Program_weight_tensor_parameter_627: name = "parameter_627" + original_name = "batch_norm_0.w_0" shape = [24] dtype = "float32" min_val = float("0") @@ -6888,6 +7516,7 @@ class Program_weight_tensor_parameter_627: class Program_weight_tensor_parameter_628: name = "parameter_628" + original_name = "batch_norm_0.w_2" shape = [24] dtype = "float32" min_val = float("0") @@ -6897,6 +7526,7 @@ class Program_weight_tensor_parameter_628: class Program_weight_tensor_parameter_629: name = "parameter_629" + original_name = "batch_norm_0.w_1" shape = [24] dtype = "float32" min_val = float("0") @@ -6906,6 +7536,7 @@ class Program_weight_tensor_parameter_629: class Program_weight_tensor_parameter_630: name = "parameter_630" + original_name = "conv2d_0.w_0" shape = [24, 3, 3, 3] dtype = "float32" min_val = float("-0.35795") diff --git a/paddle_samples/PaddleX/MobileNetV4_hybrid_medium/subgraph_2/input_meta.py b/paddle_samples/PaddleX/MobileNetV4_hybrid_medium/subgraph_2/input_meta.py index 5e78194d9..b5da25e78 100644 --- a/paddle_samples/PaddleX/MobileNetV4_hybrid_medium/subgraph_2/input_meta.py +++ b/paddle_samples/PaddleX/MobileNetV4_hybrid_medium/subgraph_2/input_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_data_0: name = "data_0" + original_name = "param_0" shape = [80] dtype = "float32" min_val = float("-2.5852") @@ -11,6 +12,7 @@ class Program_weight_tensor_data_0: class Program_weight_tensor_data_1: name = "data_1" + original_name = "param_1" shape = [80] dtype = "float32" min_val = float("-1.29432") @@ -22,6 +24,7 @@ class Program_weight_tensor_data_1: class Program_weight_tensor_data_2: name = "data_2" + original_name = "param_10" shape = [160] dtype = "float32" min_val = float("-4.8749") @@ -33,6 +36,7 @@ class Program_weight_tensor_data_2: class Program_weight_tensor_data_3: name = "data_3" + original_name = "param_11" shape = [160] dtype = "float32" min_val = float("-2.0034") @@ -44,6 +48,7 @@ class Program_weight_tensor_data_3: class Program_weight_tensor_data_4: name = "data_4" + original_name = "param_12" shape = [160] dtype = "float32" min_val = float("-4.04529") @@ -55,6 +60,7 @@ class Program_weight_tensor_data_4: class Program_weight_tensor_data_5: name = "data_5" + original_name = "param_13" shape = [160] dtype = "float32" min_val = float("-4.84156") @@ -66,6 +72,7 @@ class Program_weight_tensor_data_5: class Program_weight_tensor_data_6: name = "data_6" + original_name = "param_14" shape = [256] dtype = "float32" min_val = float("-1.53345") @@ -77,6 +84,7 @@ class Program_weight_tensor_data_6: class Program_weight_tensor_data_7: name = "data_7" + original_name = "param_15" shape = [256] dtype = "float32" min_val = float("-0.553647") @@ -88,6 +96,7 @@ class Program_weight_tensor_data_7: class Program_weight_tensor_data_8: name = "data_8" + original_name = "param_16" shape = [256] dtype = "float32" min_val = float("-0.490158") @@ -99,6 +108,7 @@ class Program_weight_tensor_data_8: class Program_weight_tensor_data_9: name = "data_9" + original_name = "param_17" shape = [256] dtype = "float32" min_val = float("-0.681105") @@ -110,6 +120,7 @@ class Program_weight_tensor_data_9: class Program_weight_tensor_data_10: name = "data_10" + original_name = "param_18" shape = [256] dtype = "float32" min_val = float("-0.547094") @@ -121,6 +132,7 @@ class Program_weight_tensor_data_10: class Program_weight_tensor_data_11: name = "data_11" + original_name = "param_19" shape = [256] dtype = "float32" min_val = float("-0.576047") @@ -132,6 +144,7 @@ class Program_weight_tensor_data_11: class Program_weight_tensor_data_12: name = "data_12" + original_name = "param_2" shape = [160] dtype = "float32" min_val = float("-0.897128") @@ -143,6 +156,7 @@ class Program_weight_tensor_data_12: class Program_weight_tensor_data_13: name = "data_13" + original_name = "param_20" shape = [256] dtype = "float32" min_val = float("-0.81347") @@ -154,6 +168,7 @@ class Program_weight_tensor_data_13: class Program_weight_tensor_data_14: name = "data_14" + original_name = "param_21" shape = [256] dtype = "float32" min_val = float("-0.687271") @@ -165,6 +180,7 @@ class Program_weight_tensor_data_14: class Program_weight_tensor_data_15: name = "data_15" + original_name = "param_22" shape = [256] dtype = "float32" min_val = float("-2.84591") @@ -176,6 +192,7 @@ class Program_weight_tensor_data_15: class Program_weight_tensor_data_16: name = "data_16" + original_name = "param_23" shape = [256] dtype = "float32" min_val = float("-0.570242") @@ -187,6 +204,7 @@ class Program_weight_tensor_data_16: class Program_weight_tensor_data_17: name = "data_17" + original_name = "param_24" shape = [256] dtype = "float32" min_val = float("-2.15215") @@ -198,6 +216,7 @@ class Program_weight_tensor_data_17: class Program_weight_tensor_data_18: name = "data_18" + original_name = "param_25" shape = [256] dtype = "float32" min_val = float("-0.686385") @@ -209,6 +228,7 @@ class Program_weight_tensor_data_18: class Program_weight_tensor_data_19: name = "data_19" + original_name = "param_26" shape = [256] dtype = "float32" min_val = float("-2.18056") @@ -220,6 +240,7 @@ class Program_weight_tensor_data_19: class Program_weight_tensor_data_20: name = "data_20" + original_name = "param_27" shape = [256] dtype = "float32" min_val = float("-0.691547") @@ -231,6 +252,7 @@ class Program_weight_tensor_data_20: class Program_weight_tensor_data_21: name = "data_21" + original_name = "param_28" shape = [256] dtype = "float32" min_val = float("-3.34237") @@ -242,6 +264,7 @@ class Program_weight_tensor_data_21: class Program_weight_tensor_data_22: name = "data_22" + original_name = "param_29" shape = [256] dtype = "float32" min_val = float("-1.68244") @@ -253,6 +276,7 @@ class Program_weight_tensor_data_22: class Program_weight_tensor_data_23: name = "data_23" + original_name = "param_3" shape = [160] dtype = "float32" min_val = float("-0.714844") @@ -264,6 +288,7 @@ class Program_weight_tensor_data_23: class Program_weight_tensor_data_24: name = "data_24" + original_name = "param_4" shape = [160] dtype = "float32" min_val = float("-0.626585") @@ -275,6 +300,7 @@ class Program_weight_tensor_data_24: class Program_weight_tensor_data_25: name = "data_25" + original_name = "param_5" shape = [160] dtype = "float32" min_val = float("-0.756078") @@ -286,6 +312,7 @@ class Program_weight_tensor_data_25: class Program_weight_tensor_data_26: name = "data_26" + original_name = "param_6" shape = [160] dtype = "float32" min_val = float("-1.52779") @@ -297,6 +324,7 @@ class Program_weight_tensor_data_26: class Program_weight_tensor_data_27: name = "data_27" + original_name = "param_7" shape = [160] dtype = "float32" min_val = float("-0.966885") @@ -308,6 +336,7 @@ class Program_weight_tensor_data_27: class Program_weight_tensor_data_28: name = "data_28" + original_name = "param_8" shape = [160] dtype = "float32" min_val = float("-4.08722") @@ -319,6 +348,7 @@ class Program_weight_tensor_data_28: class Program_weight_tensor_data_29: name = "data_29" + original_name = "param_9" shape = [160] dtype = "float32" min_val = float("-1.051") @@ -330,6 +360,7 @@ class Program_weight_tensor_data_29: class Program_weight_tensor_data_30: name = "data_30" + original_name = "var_0" shape = [64, 3, 224, 224] dtype = "float32" min_val = float("-3.87123") diff --git a/paddle_samples/PaddleX/MobileNetV4_hybrid_medium/subgraph_2/weight_meta.py b/paddle_samples/PaddleX/MobileNetV4_hybrid_medium/subgraph_2/weight_meta.py index 6cff57404..d50eb4868 100644 --- a/paddle_samples/PaddleX/MobileNetV4_hybrid_medium/subgraph_2/weight_meta.py +++ b/paddle_samples/PaddleX/MobileNetV4_hybrid_medium/subgraph_2/weight_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_parameter_0: name = "parameter_0" + original_name = "linear_0.b_0" shape = [102] dtype = "float32" data = None @@ -7,6 +8,7 @@ class Program_weight_tensor_parameter_0: class Program_weight_tensor_parameter_1: name = "parameter_1" + original_name = "linear_0.w_0" shape = [1280, 102] dtype = "float32" min_val = float("-0.0658901") @@ -18,6 +20,7 @@ class Program_weight_tensor_parameter_1: class Program_weight_tensor_parameter_2: name = "parameter_2" + original_name = "batch_norm_95.b_0" shape = [1280] dtype = "float32" min_val = float("-1.65592") @@ -29,6 +32,7 @@ class Program_weight_tensor_parameter_2: class Program_weight_tensor_parameter_3: name = "parameter_3" + original_name = "batch_norm_95.w_0" shape = [1280] dtype = "float32" min_val = float("0.816223") @@ -40,6 +44,7 @@ class Program_weight_tensor_parameter_3: class Program_weight_tensor_parameter_4: name = "parameter_4" + original_name = "batch_norm_95.w_2" shape = [1280] dtype = "float32" min_val = float("0.000284435") @@ -51,6 +56,7 @@ class Program_weight_tensor_parameter_4: class Program_weight_tensor_parameter_5: name = "parameter_5" + original_name = "batch_norm_95.w_1" shape = [1280] dtype = "float32" min_val = float("-0.0203485") @@ -62,6 +68,7 @@ class Program_weight_tensor_parameter_5: class Program_weight_tensor_parameter_6: name = "parameter_6" + original_name = "conv2d_119.w_0" shape = [1280, 960, 1, 1] dtype = "float32" min_val = float("-0.197958") @@ -73,6 +80,7 @@ class Program_weight_tensor_parameter_6: class Program_weight_tensor_parameter_7: name = "parameter_7" + original_name = "batch_norm_94.b_0" shape = [960] dtype = "float32" min_val = float("-2.93121") @@ -84,6 +92,7 @@ class Program_weight_tensor_parameter_7: class Program_weight_tensor_parameter_8: name = "parameter_8" + original_name = "batch_norm_94.w_0" shape = [960] dtype = "float32" min_val = float("0.147799") @@ -95,6 +104,7 @@ class Program_weight_tensor_parameter_8: class Program_weight_tensor_parameter_9: name = "parameter_9" + original_name = "batch_norm_94.w_2" shape = [960] dtype = "float32" min_val = float("0.355746") @@ -106,6 +116,7 @@ class Program_weight_tensor_parameter_9: class Program_weight_tensor_parameter_10: name = "parameter_10" + original_name = "batch_norm_94.w_1" shape = [960] dtype = "float32" min_val = float("-1.09557") @@ -117,6 +128,7 @@ class Program_weight_tensor_parameter_10: class Program_weight_tensor_parameter_11: name = "parameter_11" + original_name = "conv2d_118.w_0" shape = [960, 256, 1, 1] dtype = "float32" min_val = float("-0.218184") @@ -128,6 +140,7 @@ class Program_weight_tensor_parameter_11: class Program_weight_tensor_parameter_12: name = "parameter_12" + original_name = "batch_norm_93.b_0" shape = [256] dtype = "float32" min_val = float("-1.1945") @@ -139,6 +152,7 @@ class Program_weight_tensor_parameter_12: class Program_weight_tensor_parameter_13: name = "parameter_13" + original_name = "batch_norm_93.w_0" shape = [256] dtype = "float32" min_val = float("-0.693552") @@ -150,6 +164,7 @@ class Program_weight_tensor_parameter_13: class Program_weight_tensor_parameter_14: name = "parameter_14" + original_name = "batch_norm_93.w_2" shape = [256] dtype = "float32" min_val = float("0.000545829") @@ -161,6 +176,7 @@ class Program_weight_tensor_parameter_14: class Program_weight_tensor_parameter_15: name = "parameter_15" + original_name = "batch_norm_93.w_1" shape = [256] dtype = "float32" min_val = float("-0.0510152") @@ -172,6 +188,7 @@ class Program_weight_tensor_parameter_15: class Program_weight_tensor_parameter_16: name = "parameter_16" + original_name = "conv2d_117.w_0" shape = [256, 1024, 1, 1] dtype = "float32" min_val = float("-0.193144") @@ -183,6 +200,7 @@ class Program_weight_tensor_parameter_16: class Program_weight_tensor_parameter_17: name = "parameter_17" + original_name = "batch_norm_92.b_0" shape = [1024] dtype = "float32" min_val = float("-6.24724") @@ -194,6 +212,7 @@ class Program_weight_tensor_parameter_17: class Program_weight_tensor_parameter_18: name = "parameter_18" + original_name = "batch_norm_92.w_0" shape = [1024] dtype = "float32" min_val = float("0.142254") @@ -205,6 +224,7 @@ class Program_weight_tensor_parameter_18: class Program_weight_tensor_parameter_19: name = "parameter_19" + original_name = "batch_norm_92.w_2" shape = [1024] dtype = "float32" min_val = float("0.0946919") @@ -216,6 +236,7 @@ class Program_weight_tensor_parameter_19: class Program_weight_tensor_parameter_20: name = "parameter_20" + original_name = "batch_norm_92.w_1" shape = [1024] dtype = "float32" min_val = float("-0.535466") @@ -227,6 +248,7 @@ class Program_weight_tensor_parameter_20: class Program_weight_tensor_parameter_21: name = "parameter_21" + original_name = "conv2d_116.w_0" shape = [1024, 256, 1, 1] dtype = "float32" min_val = float("-0.272254") @@ -238,6 +260,7 @@ class Program_weight_tensor_parameter_21: class Program_weight_tensor_parameter_22: name = "parameter_22" + original_name = "batch_norm_91.b_0" shape = [256] dtype = "float32" min_val = float("-1.65454") @@ -249,6 +272,7 @@ class Program_weight_tensor_parameter_22: class Program_weight_tensor_parameter_23: name = "parameter_23" + original_name = "batch_norm_91.w_0" shape = [256] dtype = "float32" min_val = float("0.660011") @@ -260,6 +284,7 @@ class Program_weight_tensor_parameter_23: class Program_weight_tensor_parameter_24: name = "parameter_24" + original_name = "batch_norm_91.w_2" shape = [256] dtype = "float32" min_val = float("0.00260837") @@ -271,6 +296,7 @@ class Program_weight_tensor_parameter_24: class Program_weight_tensor_parameter_25: name = "parameter_25" + original_name = "batch_norm_91.w_1" shape = [256] dtype = "float32" min_val = float("-1.15289") @@ -282,6 +308,7 @@ class Program_weight_tensor_parameter_25: class Program_weight_tensor_parameter_26: name = "parameter_26" + original_name = "conv2d_115.w_0" shape = [256, 1, 5, 5] dtype = "float32" min_val = float("-0.355846") @@ -293,6 +320,7 @@ class Program_weight_tensor_parameter_26: class Program_weight_tensor_parameter_27: name = "parameter_27" + original_name = "conv2d_114.w_0" shape = [256, 256, 1, 1] dtype = "float32" min_val = float("-0.172969") @@ -304,6 +332,7 @@ class Program_weight_tensor_parameter_27: class Program_weight_tensor_parameter_28: name = "parameter_28" + original_name = "conv2d_113.w_0" shape = [64, 256, 1, 1] dtype = "float32" min_val = float("-0.20635") @@ -315,6 +344,7 @@ class Program_weight_tensor_parameter_28: class Program_weight_tensor_parameter_29: name = "parameter_29" + original_name = "conv2d_112.w_0" shape = [64, 256, 1, 1] dtype = "float32" min_val = float("-0.222786") @@ -326,6 +356,7 @@ class Program_weight_tensor_parameter_29: class Program_weight_tensor_parameter_30: name = "parameter_30" + original_name = "conv2d_111.w_0" shape = [256, 256, 1, 1] dtype = "float32" min_val = float("-0.302765") @@ -337,6 +368,7 @@ class Program_weight_tensor_parameter_30: class Program_weight_tensor_parameter_31: name = "parameter_31" + original_name = "batch_norm_90.b_0" shape = [256] dtype = "float32" min_val = float("-1.97352") @@ -348,6 +380,7 @@ class Program_weight_tensor_parameter_31: class Program_weight_tensor_parameter_32: name = "parameter_32" + original_name = "batch_norm_90.w_0" shape = [256] dtype = "float32" min_val = float("0.333598") @@ -359,6 +392,7 @@ class Program_weight_tensor_parameter_32: class Program_weight_tensor_parameter_33: name = "parameter_33" + original_name = "batch_norm_90.w_2" shape = [256] dtype = "float32" min_val = float("0.0711613") @@ -370,6 +404,7 @@ class Program_weight_tensor_parameter_33: class Program_weight_tensor_parameter_34: name = "parameter_34" + original_name = "batch_norm_90.w_1" shape = [256] dtype = "float32" min_val = float("-3.37049") @@ -381,6 +416,7 @@ class Program_weight_tensor_parameter_34: class Program_weight_tensor_parameter_35: name = "parameter_35" + original_name = "batch_norm_89.b_0" shape = [256] dtype = "float32" min_val = float("-0.859424") @@ -392,6 +428,7 @@ class Program_weight_tensor_parameter_35: class Program_weight_tensor_parameter_36: name = "parameter_36" + original_name = "batch_norm_89.w_0" shape = [256] dtype = "float32" min_val = float("-0.691708") @@ -403,6 +440,7 @@ class Program_weight_tensor_parameter_36: class Program_weight_tensor_parameter_37: name = "parameter_37" + original_name = "batch_norm_89.w_2" shape = [256] dtype = "float32" min_val = float("7.43688e-05") @@ -414,6 +452,7 @@ class Program_weight_tensor_parameter_37: class Program_weight_tensor_parameter_38: name = "parameter_38" + original_name = "batch_norm_89.w_1" shape = [256] dtype = "float32" min_val = float("-0.00654145") @@ -425,6 +464,7 @@ class Program_weight_tensor_parameter_38: class Program_weight_tensor_parameter_39: name = "parameter_39" + original_name = "conv2d_110.w_0" shape = [256, 1024, 1, 1] dtype = "float32" min_val = float("-0.243312") @@ -436,6 +476,7 @@ class Program_weight_tensor_parameter_39: class Program_weight_tensor_parameter_40: name = "parameter_40" + original_name = "batch_norm_88.b_0" shape = [1024] dtype = "float32" min_val = float("-3.00884") @@ -447,6 +488,7 @@ class Program_weight_tensor_parameter_40: class Program_weight_tensor_parameter_41: name = "parameter_41" + original_name = "batch_norm_88.w_0" shape = [1024] dtype = "float32" min_val = float("0.0745732") @@ -458,6 +500,7 @@ class Program_weight_tensor_parameter_41: class Program_weight_tensor_parameter_42: name = "parameter_42" + original_name = "batch_norm_88.w_2" shape = [1024] dtype = "float32" min_val = float("0.105582") @@ -469,6 +512,7 @@ class Program_weight_tensor_parameter_42: class Program_weight_tensor_parameter_43: name = "parameter_43" + original_name = "batch_norm_88.w_1" shape = [1024] dtype = "float32" min_val = float("-0.79354") @@ -480,6 +524,7 @@ class Program_weight_tensor_parameter_43: class Program_weight_tensor_parameter_44: name = "parameter_44" + original_name = "conv2d_109.w_0" shape = [1024, 256, 1, 1] dtype = "float32" min_val = float("-0.27873") @@ -491,6 +536,7 @@ class Program_weight_tensor_parameter_44: class Program_weight_tensor_parameter_45: name = "parameter_45" + original_name = "batch_norm_87.b_0" shape = [256] dtype = "float32" min_val = float("-1.31225") @@ -502,6 +548,7 @@ class Program_weight_tensor_parameter_45: class Program_weight_tensor_parameter_46: name = "parameter_46" + original_name = "batch_norm_87.w_0" shape = [256] dtype = "float32" min_val = float("0.442031") @@ -513,6 +560,7 @@ class Program_weight_tensor_parameter_46: class Program_weight_tensor_parameter_47: name = "parameter_47" + original_name = "batch_norm_87.w_2" shape = [256] dtype = "float32" min_val = float("0.00138184") @@ -524,6 +572,7 @@ class Program_weight_tensor_parameter_47: class Program_weight_tensor_parameter_48: name = "parameter_48" + original_name = "batch_norm_87.w_1" shape = [256] dtype = "float32" min_val = float("-1.29325") @@ -535,6 +584,7 @@ class Program_weight_tensor_parameter_48: class Program_weight_tensor_parameter_49: name = "parameter_49" + original_name = "conv2d_108.w_0" shape = [256, 1, 5, 5] dtype = "float32" min_val = float("-0.350082") @@ -546,6 +596,7 @@ class Program_weight_tensor_parameter_49: class Program_weight_tensor_parameter_50: name = "parameter_50" + original_name = "conv2d_107.w_0" shape = [256, 256, 1, 1] dtype = "float32" min_val = float("-0.166751") @@ -557,6 +608,7 @@ class Program_weight_tensor_parameter_50: class Program_weight_tensor_parameter_51: name = "parameter_51" + original_name = "conv2d_106.w_0" shape = [64, 256, 1, 1] dtype = "float32" min_val = float("-0.161451") @@ -568,6 +620,7 @@ class Program_weight_tensor_parameter_51: class Program_weight_tensor_parameter_52: name = "parameter_52" + original_name = "conv2d_105.w_0" shape = [64, 256, 1, 1] dtype = "float32" min_val = float("-0.192372") @@ -579,6 +632,7 @@ class Program_weight_tensor_parameter_52: class Program_weight_tensor_parameter_53: name = "parameter_53" + original_name = "conv2d_104.w_0" shape = [256, 256, 1, 1] dtype = "float32" min_val = float("-0.174397") @@ -590,6 +644,7 @@ class Program_weight_tensor_parameter_53: class Program_weight_tensor_parameter_54: name = "parameter_54" + original_name = "batch_norm_86.b_0" shape = [256] dtype = "float32" min_val = float("-1.08441") @@ -601,6 +656,7 @@ class Program_weight_tensor_parameter_54: class Program_weight_tensor_parameter_55: name = "parameter_55" + original_name = "batch_norm_86.w_0" shape = [256] dtype = "float32" min_val = float("0.90466") @@ -612,6 +668,7 @@ class Program_weight_tensor_parameter_55: class Program_weight_tensor_parameter_56: name = "parameter_56" + original_name = "batch_norm_86.w_2" shape = [256] dtype = "float32" min_val = float("0.0121626") @@ -623,6 +680,7 @@ class Program_weight_tensor_parameter_56: class Program_weight_tensor_parameter_57: name = "parameter_57" + original_name = "batch_norm_86.w_1" shape = [256] dtype = "float32" min_val = float("-2.80479") @@ -634,6 +692,7 @@ class Program_weight_tensor_parameter_57: class Program_weight_tensor_parameter_58: name = "parameter_58" + original_name = "batch_norm_85.b_0" shape = [256] dtype = "float32" min_val = float("-0.984466") @@ -645,6 +704,7 @@ class Program_weight_tensor_parameter_58: class Program_weight_tensor_parameter_59: name = "parameter_59" + original_name = "batch_norm_85.w_0" shape = [256] dtype = "float32" min_val = float("0.603443") @@ -656,6 +716,7 @@ class Program_weight_tensor_parameter_59: class Program_weight_tensor_parameter_60: name = "parameter_60" + original_name = "batch_norm_85.w_2" shape = [256] dtype = "float32" min_val = float("0.0666545") @@ -667,6 +728,7 @@ class Program_weight_tensor_parameter_60: class Program_weight_tensor_parameter_61: name = "parameter_61" + original_name = "batch_norm_85.w_1" shape = [256] dtype = "float32" min_val = float("-1.21938") @@ -678,6 +740,7 @@ class Program_weight_tensor_parameter_61: class Program_weight_tensor_parameter_62: name = "parameter_62" + original_name = "conv2d_103.w_0" shape = [256, 1024, 1, 1] dtype = "float32" min_val = float("-0.172177") @@ -689,6 +752,7 @@ class Program_weight_tensor_parameter_62: class Program_weight_tensor_parameter_63: name = "parameter_63" + original_name = "batch_norm_84.b_0" shape = [1024] dtype = "float32" min_val = float("-5.14751") @@ -700,6 +764,7 @@ class Program_weight_tensor_parameter_63: class Program_weight_tensor_parameter_64: name = "parameter_64" + original_name = "batch_norm_84.w_0" shape = [1024] dtype = "float32" min_val = float("0.382758") @@ -711,6 +776,7 @@ class Program_weight_tensor_parameter_64: class Program_weight_tensor_parameter_65: name = "parameter_65" + original_name = "batch_norm_84.w_2" shape = [1024] dtype = "float32" min_val = float("3.15292e-42") @@ -722,6 +788,7 @@ class Program_weight_tensor_parameter_65: class Program_weight_tensor_parameter_66: name = "parameter_66" + original_name = "batch_norm_84.w_1" shape = [1024] dtype = "float32" min_val = float("-0.459247") @@ -733,6 +800,7 @@ class Program_weight_tensor_parameter_66: class Program_weight_tensor_parameter_67: name = "parameter_67" + original_name = "conv2d_102.w_0" shape = [1024, 1, 5, 5] dtype = "float32" min_val = float("-0.404487") @@ -744,6 +812,7 @@ class Program_weight_tensor_parameter_67: class Program_weight_tensor_parameter_68: name = "parameter_68" + original_name = "batch_norm_83.b_0" shape = [1024] dtype = "float32" min_val = float("-1.98727") @@ -755,6 +824,7 @@ class Program_weight_tensor_parameter_68: class Program_weight_tensor_parameter_69: name = "parameter_69" + original_name = "batch_norm_83.w_0" shape = [1024] dtype = "float32" min_val = float("0.187662") @@ -766,6 +836,7 @@ class Program_weight_tensor_parameter_69: class Program_weight_tensor_parameter_70: name = "parameter_70" + original_name = "batch_norm_83.w_2" shape = [1024] dtype = "float32" min_val = float("7.0886e-25") @@ -777,6 +848,7 @@ class Program_weight_tensor_parameter_70: class Program_weight_tensor_parameter_71: name = "parameter_71" + original_name = "batch_norm_83.w_1" shape = [1024] dtype = "float32" min_val = float("-0.607002") @@ -788,6 +860,7 @@ class Program_weight_tensor_parameter_71: class Program_weight_tensor_parameter_72: name = "parameter_72" + original_name = "conv2d_101.w_0" shape = [1024, 256, 1, 1] dtype = "float32" min_val = float("-0.341424") @@ -799,6 +872,7 @@ class Program_weight_tensor_parameter_72: class Program_weight_tensor_parameter_73: name = "parameter_73" + original_name = "batch_norm_82.b_0" shape = [256] dtype = "float32" min_val = float("-1.11685") @@ -810,6 +884,7 @@ class Program_weight_tensor_parameter_73: class Program_weight_tensor_parameter_74: name = "parameter_74" + original_name = "batch_norm_82.w_0" shape = [256] dtype = "float32" min_val = float("0.623382") @@ -821,6 +896,7 @@ class Program_weight_tensor_parameter_74: class Program_weight_tensor_parameter_75: name = "parameter_75" + original_name = "batch_norm_82.w_2" shape = [256] dtype = "float32" min_val = float("0.000348899") @@ -832,6 +908,7 @@ class Program_weight_tensor_parameter_75: class Program_weight_tensor_parameter_76: name = "parameter_76" + original_name = "batch_norm_82.w_1" shape = [256] dtype = "float32" min_val = float("-1.24805") @@ -843,6 +920,7 @@ class Program_weight_tensor_parameter_76: class Program_weight_tensor_parameter_77: name = "parameter_77" + original_name = "conv2d_100.w_0" shape = [256, 1, 5, 5] dtype = "float32" min_val = float("-0.414819") @@ -854,6 +932,7 @@ class Program_weight_tensor_parameter_77: class Program_weight_tensor_parameter_78: name = "parameter_78" + original_name = "conv2d_99.w_0" shape = [256, 256, 1, 1] dtype = "float32" min_val = float("-0.139159") @@ -865,6 +944,7 @@ class Program_weight_tensor_parameter_78: class Program_weight_tensor_parameter_79: name = "parameter_79" + original_name = "conv2d_98.w_0" shape = [64, 256, 1, 1] dtype = "float32" min_val = float("-0.129259") @@ -876,6 +956,7 @@ class Program_weight_tensor_parameter_79: class Program_weight_tensor_parameter_80: name = "parameter_80" + original_name = "conv2d_97.w_0" shape = [64, 256, 1, 1] dtype = "float32" min_val = float("-0.184316") @@ -887,6 +968,7 @@ class Program_weight_tensor_parameter_80: class Program_weight_tensor_parameter_81: name = "parameter_81" + original_name = "conv2d_96.w_0" shape = [256, 256, 1, 1] dtype = "float32" min_val = float("-0.153751") @@ -898,6 +980,7 @@ class Program_weight_tensor_parameter_81: class Program_weight_tensor_parameter_82: name = "parameter_82" + original_name = "batch_norm_81.b_0" shape = [256] dtype = "float32" min_val = float("-1.67684") @@ -909,6 +992,7 @@ class Program_weight_tensor_parameter_82: class Program_weight_tensor_parameter_83: name = "parameter_83" + original_name = "batch_norm_81.w_0" shape = [256] dtype = "float32" min_val = float("0.795951") @@ -920,6 +1004,7 @@ class Program_weight_tensor_parameter_83: class Program_weight_tensor_parameter_84: name = "parameter_84" + original_name = "batch_norm_81.w_2" shape = [256] dtype = "float32" min_val = float("0.0089074") @@ -931,6 +1016,7 @@ class Program_weight_tensor_parameter_84: class Program_weight_tensor_parameter_85: name = "parameter_85" + original_name = "batch_norm_81.w_1" shape = [256] dtype = "float32" min_val = float("-2.28253") @@ -942,6 +1028,7 @@ class Program_weight_tensor_parameter_85: class Program_weight_tensor_parameter_86: name = "parameter_86" + original_name = "batch_norm_80.b_0" shape = [256] dtype = "float32" min_val = float("-0.771481") @@ -953,6 +1040,7 @@ class Program_weight_tensor_parameter_86: class Program_weight_tensor_parameter_87: name = "parameter_87" + original_name = "batch_norm_80.w_0" shape = [256] dtype = "float32" min_val = float("0.381269") @@ -964,6 +1052,7 @@ class Program_weight_tensor_parameter_87: class Program_weight_tensor_parameter_88: name = "parameter_88" + original_name = "batch_norm_80.w_2" shape = [256] dtype = "float32" min_val = float("0.000268346") @@ -975,6 +1064,7 @@ class Program_weight_tensor_parameter_88: class Program_weight_tensor_parameter_89: name = "parameter_89" + original_name = "batch_norm_80.w_1" shape = [256] dtype = "float32" min_val = float("-0.0191594") @@ -986,6 +1076,7 @@ class Program_weight_tensor_parameter_89: class Program_weight_tensor_parameter_90: name = "parameter_90" + original_name = "conv2d_95.w_0" shape = [256, 1024, 1, 1] dtype = "float32" min_val = float("-0.112692") @@ -997,6 +1088,7 @@ class Program_weight_tensor_parameter_90: class Program_weight_tensor_parameter_91: name = "parameter_91" + original_name = "batch_norm_79.b_0" shape = [1024] dtype = "float32" min_val = float("-2.89252") @@ -1008,6 +1100,7 @@ class Program_weight_tensor_parameter_91: class Program_weight_tensor_parameter_92: name = "parameter_92" + original_name = "batch_norm_79.w_0" shape = [1024] dtype = "float32" min_val = float("0.126879") @@ -1019,6 +1112,7 @@ class Program_weight_tensor_parameter_92: class Program_weight_tensor_parameter_93: name = "parameter_93" + original_name = "batch_norm_79.w_2" shape = [1024] dtype = "float32" min_val = float("0.148967") @@ -1030,6 +1124,7 @@ class Program_weight_tensor_parameter_93: class Program_weight_tensor_parameter_94: name = "parameter_94" + original_name = "batch_norm_79.w_1" shape = [1024] dtype = "float32" min_val = float("-0.900129") @@ -1041,6 +1136,7 @@ class Program_weight_tensor_parameter_94: class Program_weight_tensor_parameter_95: name = "parameter_95" + original_name = "conv2d_94.w_0" shape = [1024, 256, 1, 1] dtype = "float32" min_val = float("-0.15241") @@ -1052,6 +1148,7 @@ class Program_weight_tensor_parameter_95: class Program_weight_tensor_parameter_96: name = "parameter_96" + original_name = "batch_norm_78.b_0" shape = [256] dtype = "float32" min_val = float("-1.94687") @@ -1063,6 +1160,7 @@ class Program_weight_tensor_parameter_96: class Program_weight_tensor_parameter_97: name = "parameter_97" + original_name = "batch_norm_78.w_0" shape = [256] dtype = "float32" min_val = float("0.289202") @@ -1074,6 +1172,7 @@ class Program_weight_tensor_parameter_97: class Program_weight_tensor_parameter_98: name = "parameter_98" + original_name = "batch_norm_78.w_2" shape = [256] dtype = "float32" min_val = float("0.000210387") @@ -1085,6 +1184,7 @@ class Program_weight_tensor_parameter_98: class Program_weight_tensor_parameter_99: name = "parameter_99" + original_name = "batch_norm_78.w_1" shape = [256] dtype = "float32" min_val = float("-0.531232") @@ -1096,6 +1196,7 @@ class Program_weight_tensor_parameter_99: class Program_weight_tensor_parameter_100: name = "parameter_100" + original_name = "conv2d_93.w_0" shape = [256, 1, 3, 3] dtype = "float32" min_val = float("-0.243897") @@ -1107,6 +1208,7 @@ class Program_weight_tensor_parameter_100: class Program_weight_tensor_parameter_101: name = "parameter_101" + original_name = "conv2d_92.w_0" shape = [256, 256, 1, 1] dtype = "float32" min_val = float("-0.128302") @@ -1118,6 +1220,7 @@ class Program_weight_tensor_parameter_101: class Program_weight_tensor_parameter_102: name = "parameter_102" + original_name = "conv2d_91.w_0" shape = [64, 256, 1, 1] dtype = "float32" min_val = float("-0.145898") @@ -1129,6 +1232,7 @@ class Program_weight_tensor_parameter_102: class Program_weight_tensor_parameter_103: name = "parameter_103" + original_name = "conv2d_90.w_0" shape = [64, 256, 1, 1] dtype = "float32" min_val = float("-0.210526") @@ -1140,6 +1244,7 @@ class Program_weight_tensor_parameter_103: class Program_weight_tensor_parameter_104: name = "parameter_104" + original_name = "conv2d_89.w_0" shape = [256, 256, 1, 1] dtype = "float32" min_val = float("-0.168526") @@ -1151,6 +1256,7 @@ class Program_weight_tensor_parameter_104: class Program_weight_tensor_parameter_105: name = "parameter_105" + original_name = "batch_norm_77.b_0" shape = [256] dtype = "float32" min_val = float("-1.15202") @@ -1162,6 +1268,7 @@ class Program_weight_tensor_parameter_105: class Program_weight_tensor_parameter_106: name = "parameter_106" + original_name = "batch_norm_77.w_0" shape = [256] dtype = "float32" min_val = float("0.608753") @@ -1173,6 +1280,7 @@ class Program_weight_tensor_parameter_106: class Program_weight_tensor_parameter_107: name = "parameter_107" + original_name = "batch_norm_77.w_2" shape = [256] dtype = "float32" min_val = float("0.00700626") @@ -1184,6 +1292,7 @@ class Program_weight_tensor_parameter_107: class Program_weight_tensor_parameter_108: name = "parameter_108" + original_name = "batch_norm_77.w_1" shape = [256] dtype = "float32" min_val = float("-2.32886") @@ -1195,6 +1304,7 @@ class Program_weight_tensor_parameter_108: class Program_weight_tensor_parameter_109: name = "parameter_109" + original_name = "batch_norm_76.b_0" shape = [256] dtype = "float32" min_val = float("-0.697839") @@ -1206,6 +1316,7 @@ class Program_weight_tensor_parameter_109: class Program_weight_tensor_parameter_110: name = "parameter_110" + original_name = "batch_norm_76.w_0" shape = [256] dtype = "float32" min_val = float("-0.654048") @@ -1217,6 +1328,7 @@ class Program_weight_tensor_parameter_110: class Program_weight_tensor_parameter_111: name = "parameter_111" + original_name = "batch_norm_76.w_2" shape = [256] dtype = "float32" min_val = float("0.000767348") @@ -1228,6 +1340,7 @@ class Program_weight_tensor_parameter_111: class Program_weight_tensor_parameter_112: name = "parameter_112" + original_name = "batch_norm_76.w_1" shape = [256] dtype = "float32" min_val = float("-0.0335569") @@ -1239,6 +1352,7 @@ class Program_weight_tensor_parameter_112: class Program_weight_tensor_parameter_113: name = "parameter_113" + original_name = "conv2d_88.w_0" shape = [256, 1024, 1, 1] dtype = "float32" min_val = float("-0.119917") @@ -1250,6 +1364,7 @@ class Program_weight_tensor_parameter_113: class Program_weight_tensor_parameter_114: name = "parameter_114" + original_name = "batch_norm_75.b_0" shape = [1024] dtype = "float32" min_val = float("-2.73234") @@ -1261,6 +1376,7 @@ class Program_weight_tensor_parameter_114: class Program_weight_tensor_parameter_115: name = "parameter_115" + original_name = "batch_norm_75.w_0" shape = [1024] dtype = "float32" min_val = float("0.209932") @@ -1272,6 +1388,7 @@ class Program_weight_tensor_parameter_115: class Program_weight_tensor_parameter_116: name = "parameter_116" + original_name = "batch_norm_75.w_2" shape = [1024] dtype = "float32" min_val = float("9.25652e-22") @@ -1283,6 +1400,7 @@ class Program_weight_tensor_parameter_116: class Program_weight_tensor_parameter_117: name = "parameter_117" + original_name = "batch_norm_75.w_1" shape = [1024] dtype = "float32" min_val = float("-0.657359") @@ -1294,6 +1412,7 @@ class Program_weight_tensor_parameter_117: class Program_weight_tensor_parameter_118: name = "parameter_118" + original_name = "conv2d_87.w_0" shape = [1024, 256, 1, 1] dtype = "float32" min_val = float("-0.168564") @@ -1305,6 +1424,7 @@ class Program_weight_tensor_parameter_118: class Program_weight_tensor_parameter_119: name = "parameter_119" + original_name = "batch_norm_74.b_0" shape = [256] dtype = "float32" min_val = float("-0.636822") @@ -1316,6 +1436,7 @@ class Program_weight_tensor_parameter_119: class Program_weight_tensor_parameter_120: name = "parameter_120" + original_name = "batch_norm_74.w_0" shape = [256] dtype = "float32" min_val = float("-0.677514") @@ -1327,6 +1448,7 @@ class Program_weight_tensor_parameter_120: class Program_weight_tensor_parameter_121: name = "parameter_121" + original_name = "batch_norm_74.w_2" shape = [256] dtype = "float32" min_val = float("0.00165207") @@ -1338,6 +1460,7 @@ class Program_weight_tensor_parameter_121: class Program_weight_tensor_parameter_122: name = "parameter_122" + original_name = "batch_norm_74.w_1" shape = [256] dtype = "float32" min_val = float("-0.0799585") @@ -1349,6 +1472,7 @@ class Program_weight_tensor_parameter_122: class Program_weight_tensor_parameter_123: name = "parameter_123" + original_name = "conv2d_86.w_0" shape = [256, 512, 1, 1] dtype = "float32" min_val = float("-0.206253") @@ -1360,6 +1484,7 @@ class Program_weight_tensor_parameter_123: class Program_weight_tensor_parameter_124: name = "parameter_124" + original_name = "batch_norm_73.b_0" shape = [512] dtype = "float32" min_val = float("-2.55909") @@ -1371,6 +1496,7 @@ class Program_weight_tensor_parameter_124: class Program_weight_tensor_parameter_125: name = "parameter_125" + original_name = "batch_norm_73.w_0" shape = [512] dtype = "float32" min_val = float("0.322778") @@ -1382,6 +1508,7 @@ class Program_weight_tensor_parameter_125: class Program_weight_tensor_parameter_126: name = "parameter_126" + original_name = "batch_norm_73.w_2" shape = [512] dtype = "float32" min_val = float("0.15549") @@ -1393,6 +1520,7 @@ class Program_weight_tensor_parameter_126: class Program_weight_tensor_parameter_127: name = "parameter_127" + original_name = "batch_norm_73.w_1" shape = [512] dtype = "float32" min_val = float("-0.545164") @@ -1404,6 +1532,7 @@ class Program_weight_tensor_parameter_127: class Program_weight_tensor_parameter_128: name = "parameter_128" + original_name = "conv2d_85.w_0" shape = [512, 256, 1, 1] dtype = "float32" min_val = float("-0.271539") @@ -1415,6 +1544,7 @@ class Program_weight_tensor_parameter_128: class Program_weight_tensor_parameter_129: name = "parameter_129" + original_name = "batch_norm_72.b_0" shape = [256] dtype = "float32" min_val = float("-0.803985") @@ -1426,6 +1556,7 @@ class Program_weight_tensor_parameter_129: class Program_weight_tensor_parameter_130: name = "parameter_130" + original_name = "batch_norm_72.w_0" shape = [256] dtype = "float32" min_val = float("0.327692") @@ -1437,6 +1568,7 @@ class Program_weight_tensor_parameter_130: class Program_weight_tensor_parameter_131: name = "parameter_131" + original_name = "batch_norm_72.w_2" shape = [256] dtype = "float32" min_val = float("0.0459146") @@ -1448,6 +1580,7 @@ class Program_weight_tensor_parameter_131: class Program_weight_tensor_parameter_132: name = "parameter_132" + original_name = "batch_norm_72.w_1" shape = [256] dtype = "float32" min_val = float("-1.07324") @@ -1459,6 +1592,7 @@ class Program_weight_tensor_parameter_132: class Program_weight_tensor_parameter_133: name = "parameter_133" + original_name = "conv2d_84.w_0" shape = [256, 512, 1, 1] dtype = "float32" min_val = float("-0.180474") @@ -1470,6 +1604,7 @@ class Program_weight_tensor_parameter_133: class Program_weight_tensor_parameter_134: name = "parameter_134" + original_name = "batch_norm_71.b_0" shape = [512] dtype = "float32" min_val = float("-4.63246") @@ -1481,6 +1616,7 @@ class Program_weight_tensor_parameter_134: class Program_weight_tensor_parameter_135: name = "parameter_135" + original_name = "batch_norm_71.w_0" shape = [512] dtype = "float32" min_val = float("0.296114") @@ -1492,6 +1628,7 @@ class Program_weight_tensor_parameter_135: class Program_weight_tensor_parameter_136: name = "parameter_136" + original_name = "batch_norm_71.w_2" shape = [512] dtype = "float32" min_val = float("2.54476e-42") @@ -1503,6 +1640,7 @@ class Program_weight_tensor_parameter_136: class Program_weight_tensor_parameter_137: name = "parameter_137" + original_name = "batch_norm_71.w_1" shape = [512] dtype = "float32" min_val = float("-0.286801") @@ -1514,6 +1652,7 @@ class Program_weight_tensor_parameter_137: class Program_weight_tensor_parameter_138: name = "parameter_138" + original_name = "conv2d_83.w_0" shape = [512, 1, 5, 5] dtype = "float32" min_val = float("-0.267966") @@ -1525,6 +1664,7 @@ class Program_weight_tensor_parameter_138: class Program_weight_tensor_parameter_139: name = "parameter_139" + original_name = "batch_norm_70.b_0" shape = [512] dtype = "float32" min_val = float("-1.60863") @@ -1536,6 +1676,7 @@ class Program_weight_tensor_parameter_139: class Program_weight_tensor_parameter_140: name = "parameter_140" + original_name = "batch_norm_70.w_0" shape = [512] dtype = "float32" min_val = float("0.0986059") @@ -1547,6 +1688,7 @@ class Program_weight_tensor_parameter_140: class Program_weight_tensor_parameter_141: name = "parameter_141" + original_name = "batch_norm_70.w_2" shape = [512] dtype = "float32" min_val = float("3.15377e-27") @@ -1558,6 +1700,7 @@ class Program_weight_tensor_parameter_141: class Program_weight_tensor_parameter_142: name = "parameter_142" + original_name = "batch_norm_70.w_1" shape = [512] dtype = "float32" min_val = float("-0.570178") @@ -1569,6 +1712,7 @@ class Program_weight_tensor_parameter_142: class Program_weight_tensor_parameter_143: name = "parameter_143" + original_name = "conv2d_82.w_0" shape = [512, 256, 1, 1] dtype = "float32" min_val = float("-0.385187") @@ -1580,6 +1724,7 @@ class Program_weight_tensor_parameter_143: class Program_weight_tensor_parameter_144: name = "parameter_144" + original_name = "batch_norm_69.b_0" shape = [256] dtype = "float32" min_val = float("-1.83475") @@ -1591,6 +1736,7 @@ class Program_weight_tensor_parameter_144: class Program_weight_tensor_parameter_145: name = "parameter_145" + original_name = "batch_norm_69.w_0" shape = [256] dtype = "float32" min_val = float("0.553193") @@ -1602,6 +1748,7 @@ class Program_weight_tensor_parameter_145: class Program_weight_tensor_parameter_146: name = "parameter_146" + original_name = "batch_norm_69.w_2" shape = [256] dtype = "float32" min_val = float("2.1683e-05") @@ -1613,6 +1760,7 @@ class Program_weight_tensor_parameter_146: class Program_weight_tensor_parameter_147: name = "parameter_147" + original_name = "batch_norm_69.w_1" shape = [256] dtype = "float32" min_val = float("-0.158419") @@ -1624,6 +1772,7 @@ class Program_weight_tensor_parameter_147: class Program_weight_tensor_parameter_148: name = "parameter_148" + original_name = "conv2d_81.w_0" shape = [256, 1, 3, 3] dtype = "float32" min_val = float("-0.230699") @@ -1635,6 +1784,7 @@ class Program_weight_tensor_parameter_148: class Program_weight_tensor_parameter_149: name = "parameter_149" + original_name = "batch_norm_68.b_0" shape = [256] dtype = "float32" min_val = float("-0.528457") @@ -1646,6 +1796,7 @@ class Program_weight_tensor_parameter_149: class Program_weight_tensor_parameter_150: name = "parameter_150" + original_name = "batch_norm_68.w_0" shape = [256] dtype = "float32" min_val = float("-0.513843") @@ -1657,6 +1808,7 @@ class Program_weight_tensor_parameter_150: class Program_weight_tensor_parameter_151: name = "parameter_151" + original_name = "batch_norm_68.w_2" shape = [256] dtype = "float32" min_val = float("0.00252618") @@ -1668,6 +1820,7 @@ class Program_weight_tensor_parameter_151: class Program_weight_tensor_parameter_152: name = "parameter_152" + original_name = "batch_norm_68.w_1" shape = [256] dtype = "float32" min_val = float("-0.10581") @@ -1679,6 +1832,7 @@ class Program_weight_tensor_parameter_152: class Program_weight_tensor_parameter_153: name = "parameter_153" + original_name = "conv2d_80.w_0" shape = [256, 512, 1, 1] dtype = "float32" min_val = float("-0.181626") @@ -1690,6 +1844,7 @@ class Program_weight_tensor_parameter_153: class Program_weight_tensor_parameter_154: name = "parameter_154" + original_name = "batch_norm_67.b_0" shape = [512] dtype = "float32" min_val = float("-2.26588") @@ -1701,6 +1856,7 @@ class Program_weight_tensor_parameter_154: class Program_weight_tensor_parameter_155: name = "parameter_155" + original_name = "batch_norm_67.w_0" shape = [512] dtype = "float32" min_val = float("0.465619") @@ -1712,6 +1868,7 @@ class Program_weight_tensor_parameter_155: class Program_weight_tensor_parameter_156: name = "parameter_156" + original_name = "batch_norm_67.w_2" shape = [512] dtype = "float32" min_val = float("0.115567") @@ -1723,6 +1880,7 @@ class Program_weight_tensor_parameter_156: class Program_weight_tensor_parameter_157: name = "parameter_157" + original_name = "batch_norm_67.w_1" shape = [512] dtype = "float32" min_val = float("-0.46384") @@ -1734,6 +1892,7 @@ class Program_weight_tensor_parameter_157: class Program_weight_tensor_parameter_158: name = "parameter_158" + original_name = "conv2d_79.w_0" shape = [512, 256, 1, 1] dtype = "float32" min_val = float("-0.156073") @@ -1745,6 +1904,7 @@ class Program_weight_tensor_parameter_158: class Program_weight_tensor_parameter_159: name = "parameter_159" + original_name = "batch_norm_66.b_0" shape = [256] dtype = "float32" min_val = float("-1.13297") @@ -1756,6 +1916,7 @@ class Program_weight_tensor_parameter_159: class Program_weight_tensor_parameter_160: name = "parameter_160" + original_name = "batch_norm_66.w_0" shape = [256] dtype = "float32" min_val = float("-0.528192") @@ -1767,6 +1928,7 @@ class Program_weight_tensor_parameter_160: class Program_weight_tensor_parameter_161: name = "parameter_161" + original_name = "batch_norm_66.w_2" shape = [256] dtype = "float32" min_val = float("0.101011") @@ -1778,6 +1940,7 @@ class Program_weight_tensor_parameter_161: class Program_weight_tensor_parameter_162: name = "parameter_162" + original_name = "batch_norm_66.w_1" shape = [256] dtype = "float32" min_val = float("-1.85329") @@ -1789,6 +1952,7 @@ class Program_weight_tensor_parameter_162: class Program_weight_tensor_parameter_163: name = "parameter_163" + original_name = "conv2d_78.w_0" shape = [256, 1024, 1, 1] dtype = "float32" min_val = float("-0.255615") @@ -1800,6 +1964,7 @@ class Program_weight_tensor_parameter_163: class Program_weight_tensor_parameter_164: name = "parameter_164" + original_name = "batch_norm_65.b_0" shape = [1024] dtype = "float32" min_val = float("-4.94413") @@ -1811,6 +1976,7 @@ class Program_weight_tensor_parameter_164: class Program_weight_tensor_parameter_165: name = "parameter_165" + original_name = "batch_norm_65.w_0" shape = [1024] dtype = "float32" min_val = float("0.287252") @@ -1822,6 +1988,7 @@ class Program_weight_tensor_parameter_165: class Program_weight_tensor_parameter_166: name = "parameter_166" + original_name = "batch_norm_65.w_2" shape = [1024] dtype = "float32" min_val = float("3.4486e-42") @@ -1833,6 +2000,7 @@ class Program_weight_tensor_parameter_166: class Program_weight_tensor_parameter_167: name = "parameter_167" + original_name = "batch_norm_65.w_1" shape = [1024] dtype = "float32" min_val = float("-0.444477") @@ -1844,6 +2012,7 @@ class Program_weight_tensor_parameter_167: class Program_weight_tensor_parameter_168: name = "parameter_168" + original_name = "conv2d_77.w_0" shape = [1024, 1, 5, 5] dtype = "float32" min_val = float("-0.498273") @@ -1855,6 +2024,7 @@ class Program_weight_tensor_parameter_168: class Program_weight_tensor_parameter_169: name = "parameter_169" + original_name = "batch_norm_64.b_0" shape = [1024] dtype = "float32" min_val = float("-1.57031") @@ -1866,6 +2036,7 @@ class Program_weight_tensor_parameter_169: class Program_weight_tensor_parameter_170: name = "parameter_170" + original_name = "batch_norm_64.w_0" shape = [1024] dtype = "float32" min_val = float("0.157467") @@ -1877,6 +2048,7 @@ class Program_weight_tensor_parameter_170: class Program_weight_tensor_parameter_171: name = "parameter_171" + original_name = "batch_norm_64.w_2" shape = [1024] dtype = "float32" min_val = float("1.09255e-24") @@ -1888,6 +2060,7 @@ class Program_weight_tensor_parameter_171: class Program_weight_tensor_parameter_172: name = "parameter_172" + original_name = "batch_norm_64.w_1" shape = [1024] dtype = "float32" min_val = float("-0.69224") @@ -1899,6 +2072,7 @@ class Program_weight_tensor_parameter_172: class Program_weight_tensor_parameter_173: name = "parameter_173" + original_name = "conv2d_76.w_0" shape = [1024, 256, 1, 1] dtype = "float32" min_val = float("-0.187303") @@ -1910,6 +2084,7 @@ class Program_weight_tensor_parameter_173: class Program_weight_tensor_parameter_174: name = "parameter_174" + original_name = "batch_norm_63.b_0" shape = [256] dtype = "float32" min_val = float("-1.315") @@ -1921,6 +2096,7 @@ class Program_weight_tensor_parameter_174: class Program_weight_tensor_parameter_175: name = "parameter_175" + original_name = "batch_norm_63.w_0" shape = [256] dtype = "float32" min_val = float("0.526333") @@ -1932,6 +2108,7 @@ class Program_weight_tensor_parameter_175: class Program_weight_tensor_parameter_176: name = "parameter_176" + original_name = "batch_norm_63.w_2" shape = [256] dtype = "float32" min_val = float("5.99995e-06") @@ -1943,6 +2120,7 @@ class Program_weight_tensor_parameter_176: class Program_weight_tensor_parameter_177: name = "parameter_177" + original_name = "batch_norm_63.w_1" shape = [256] dtype = "float32" min_val = float("-0.0940835") @@ -1954,6 +2132,7 @@ class Program_weight_tensor_parameter_177: class Program_weight_tensor_parameter_178: name = "parameter_178" + original_name = "conv2d_75.w_0" shape = [256, 1, 3, 3] dtype = "float32" min_val = float("-0.228679") @@ -1965,6 +2144,7 @@ class Program_weight_tensor_parameter_178: class Program_weight_tensor_parameter_179: name = "parameter_179" + original_name = "batch_norm_62.b_0" shape = [256] dtype = "float32" min_val = float("-0.94596") @@ -1976,6 +2156,7 @@ class Program_weight_tensor_parameter_179: class Program_weight_tensor_parameter_180: name = "parameter_180" + original_name = "batch_norm_62.w_0" shape = [256] dtype = "float32" min_val = float("-0.411029") @@ -1987,6 +2168,7 @@ class Program_weight_tensor_parameter_180: class Program_weight_tensor_parameter_181: name = "parameter_181" + original_name = "batch_norm_62.w_2" shape = [256] dtype = "float32" min_val = float("0.0785286") @@ -1998,6 +2180,7 @@ class Program_weight_tensor_parameter_181: class Program_weight_tensor_parameter_182: name = "parameter_182" + original_name = "batch_norm_62.w_1" shape = [256] dtype = "float32" min_val = float("-1.3079") @@ -2009,6 +2192,7 @@ class Program_weight_tensor_parameter_182: class Program_weight_tensor_parameter_183: name = "parameter_183" + original_name = "conv2d_74.w_0" shape = [256, 1024, 1, 1] dtype = "float32" min_val = float("-0.283292") @@ -2020,6 +2204,7 @@ class Program_weight_tensor_parameter_183: class Program_weight_tensor_parameter_184: name = "parameter_184" + original_name = "batch_norm_61.b_0" shape = [1024] dtype = "float32" min_val = float("-5.60632") @@ -2031,6 +2216,7 @@ class Program_weight_tensor_parameter_184: class Program_weight_tensor_parameter_185: name = "parameter_185" + original_name = "batch_norm_61.w_0" shape = [1024] dtype = "float32" min_val = float("0.19851") @@ -2042,6 +2228,7 @@ class Program_weight_tensor_parameter_185: class Program_weight_tensor_parameter_186: name = "parameter_186" + original_name = "batch_norm_61.w_2" shape = [1024] dtype = "float32" min_val = float("3.27904e-42") @@ -2053,6 +2240,7 @@ class Program_weight_tensor_parameter_186: class Program_weight_tensor_parameter_187: name = "parameter_187" + original_name = "batch_norm_61.w_1" shape = [1024] dtype = "float32" min_val = float("-0.313936") @@ -2064,6 +2252,7 @@ class Program_weight_tensor_parameter_187: class Program_weight_tensor_parameter_188: name = "parameter_188" + original_name = "conv2d_73.w_0" shape = [1024, 1, 5, 5] dtype = "float32" min_val = float("-0.214485") @@ -2075,6 +2264,7 @@ class Program_weight_tensor_parameter_188: class Program_weight_tensor_parameter_189: name = "parameter_189" + original_name = "batch_norm_60.b_0" shape = [1024] dtype = "float32" min_val = float("-1.28421") @@ -2086,6 +2276,7 @@ class Program_weight_tensor_parameter_189: class Program_weight_tensor_parameter_190: name = "parameter_190" + original_name = "batch_norm_60.w_0" shape = [1024] dtype = "float32" min_val = float("0.123831") @@ -2097,6 +2288,7 @@ class Program_weight_tensor_parameter_190: class Program_weight_tensor_parameter_191: name = "parameter_191" + original_name = "batch_norm_60.w_2" shape = [1024] dtype = "float32" min_val = float("1.30021e-25") @@ -2108,6 +2300,7 @@ class Program_weight_tensor_parameter_191: class Program_weight_tensor_parameter_192: name = "parameter_192" + original_name = "batch_norm_60.w_1" shape = [1024] dtype = "float32" min_val = float("-0.729659") @@ -2119,6 +2312,7 @@ class Program_weight_tensor_parameter_192: class Program_weight_tensor_parameter_193: name = "parameter_193" + original_name = "conv2d_72.w_0" shape = [1024, 256, 1, 1] dtype = "float32" min_val = float("-0.226382") @@ -2130,6 +2324,7 @@ class Program_weight_tensor_parameter_193: class Program_weight_tensor_parameter_194: name = "parameter_194" + original_name = "batch_norm_59.b_0" shape = [256] dtype = "float32" min_val = float("-1.38132") @@ -2141,6 +2336,7 @@ class Program_weight_tensor_parameter_194: class Program_weight_tensor_parameter_195: name = "parameter_195" + original_name = "batch_norm_59.w_0" shape = [256] dtype = "float32" min_val = float("0.488798") @@ -2152,6 +2348,7 @@ class Program_weight_tensor_parameter_195: class Program_weight_tensor_parameter_196: name = "parameter_196" + original_name = "batch_norm_59.w_2" shape = [256] dtype = "float32" min_val = float("6.86075e-06") @@ -2163,6 +2360,7 @@ class Program_weight_tensor_parameter_196: class Program_weight_tensor_parameter_197: name = "parameter_197" + original_name = "batch_norm_59.w_1" shape = [256] dtype = "float32" min_val = float("-0.103602") @@ -2174,6 +2372,7 @@ class Program_weight_tensor_parameter_197: class Program_weight_tensor_parameter_198: name = "parameter_198" + original_name = "conv2d_71.w_0" shape = [256, 1, 3, 3] dtype = "float32" min_val = float("-0.236959") @@ -2185,6 +2384,7 @@ class Program_weight_tensor_parameter_198: class Program_weight_tensor_parameter_199: name = "parameter_199" + original_name = "batch_norm_58.b_0" shape = [256] dtype = "float32" min_val = float("-0.919447") @@ -2196,6 +2396,7 @@ class Program_weight_tensor_parameter_199: class Program_weight_tensor_parameter_200: name = "parameter_200" + original_name = "batch_norm_58.w_0" shape = [256] dtype = "float32" min_val = float("-0.532961") @@ -2207,6 +2408,7 @@ class Program_weight_tensor_parameter_200: class Program_weight_tensor_parameter_201: name = "parameter_201" + original_name = "batch_norm_58.w_2" shape = [256] dtype = "float32" min_val = float("0.116891") @@ -2218,6 +2420,7 @@ class Program_weight_tensor_parameter_201: class Program_weight_tensor_parameter_202: name = "parameter_202" + original_name = "batch_norm_58.w_1" shape = [256] dtype = "float32" min_val = float("-1.67192") @@ -2229,6 +2432,7 @@ class Program_weight_tensor_parameter_202: class Program_weight_tensor_parameter_203: name = "parameter_203" + original_name = "conv2d_70.w_0" shape = [256, 1024, 1, 1] dtype = "float32" min_val = float("-0.18377") @@ -2240,6 +2444,7 @@ class Program_weight_tensor_parameter_203: class Program_weight_tensor_parameter_204: name = "parameter_204" + original_name = "batch_norm_57.b_0" shape = [1024] dtype = "float32" min_val = float("-4.42359") @@ -2251,6 +2456,7 @@ class Program_weight_tensor_parameter_204: class Program_weight_tensor_parameter_205: name = "parameter_205" + original_name = "batch_norm_57.w_0" shape = [1024] dtype = "float32" min_val = float("0.273494") @@ -2262,6 +2468,7 @@ class Program_weight_tensor_parameter_205: class Program_weight_tensor_parameter_206: name = "parameter_206" + original_name = "batch_norm_57.w_2" shape = [1024] dtype = "float32" min_val = float("3.37293e-42") @@ -2273,6 +2480,7 @@ class Program_weight_tensor_parameter_206: class Program_weight_tensor_parameter_207: name = "parameter_207" + original_name = "batch_norm_57.w_1" shape = [1024] dtype = "float32" min_val = float("-0.330261") @@ -2284,6 +2492,7 @@ class Program_weight_tensor_parameter_207: class Program_weight_tensor_parameter_208: name = "parameter_208" + original_name = "conv2d_69.w_0" shape = [1024, 1, 5, 5] dtype = "float32" min_val = float("-0.142099") @@ -2295,6 +2504,7 @@ class Program_weight_tensor_parameter_208: class Program_weight_tensor_parameter_209: name = "parameter_209" + original_name = "batch_norm_56.b_0" shape = [1024] dtype = "float32" min_val = float("-1.05062") @@ -2306,6 +2516,7 @@ class Program_weight_tensor_parameter_209: class Program_weight_tensor_parameter_210: name = "parameter_210" + original_name = "batch_norm_56.w_0" shape = [1024] dtype = "float32" min_val = float("0.150814") @@ -2317,6 +2528,7 @@ class Program_weight_tensor_parameter_210: class Program_weight_tensor_parameter_211: name = "parameter_211" + original_name = "batch_norm_56.w_2" shape = [1024] dtype = "float32" min_val = float("8.50161e-26") @@ -2328,6 +2540,7 @@ class Program_weight_tensor_parameter_211: class Program_weight_tensor_parameter_212: name = "parameter_212" + original_name = "batch_norm_56.w_1" shape = [1024] dtype = "float32" min_val = float("-0.478875") @@ -2339,6 +2552,7 @@ class Program_weight_tensor_parameter_212: class Program_weight_tensor_parameter_213: name = "parameter_213" + original_name = "conv2d_68.w_0" shape = [1024, 256, 1, 1] dtype = "float32" min_val = float("-0.331815") @@ -2350,6 +2564,7 @@ class Program_weight_tensor_parameter_213: class Program_weight_tensor_parameter_214: name = "parameter_214" + original_name = "batch_norm_55.b_0" shape = [256] dtype = "float32" min_val = float("-1.35245") @@ -2361,6 +2576,7 @@ class Program_weight_tensor_parameter_214: class Program_weight_tensor_parameter_215: name = "parameter_215" + original_name = "batch_norm_55.w_0" shape = [256] dtype = "float32" min_val = float("0.404272") @@ -2372,6 +2588,7 @@ class Program_weight_tensor_parameter_215: class Program_weight_tensor_parameter_216: name = "parameter_216" + original_name = "batch_norm_55.w_2" shape = [256] dtype = "float32" min_val = float("1.42812e-05") @@ -2383,6 +2600,7 @@ class Program_weight_tensor_parameter_216: class Program_weight_tensor_parameter_217: name = "parameter_217" + original_name = "batch_norm_55.w_1" shape = [256] dtype = "float32" min_val = float("-0.245008") @@ -2394,6 +2612,7 @@ class Program_weight_tensor_parameter_217: class Program_weight_tensor_parameter_218: name = "parameter_218" + original_name = "conv2d_67.w_0" shape = [256, 1, 5, 5] dtype = "float32" min_val = float("-0.276888") @@ -2405,6 +2624,7 @@ class Program_weight_tensor_parameter_218: class Program_weight_tensor_parameter_219: name = "parameter_219" + original_name = "batch_norm_54.b_0" shape = [256] dtype = "float32" min_val = float("-2.04761") @@ -2416,6 +2636,7 @@ class Program_weight_tensor_parameter_219: class Program_weight_tensor_parameter_220: name = "parameter_220" + original_name = "batch_norm_54.w_0" shape = [256] dtype = "float32" min_val = float("0.310984") @@ -2427,6 +2648,7 @@ class Program_weight_tensor_parameter_220: class Program_weight_tensor_parameter_221: name = "parameter_221" + original_name = "batch_norm_54.w_2" shape = [256] dtype = "float32" min_val = float("0.245499") @@ -2438,6 +2660,7 @@ class Program_weight_tensor_parameter_221: class Program_weight_tensor_parameter_222: name = "parameter_222" + original_name = "batch_norm_54.w_1" shape = [256] dtype = "float32" min_val = float("-2.63479") @@ -2449,6 +2672,7 @@ class Program_weight_tensor_parameter_222: class Program_weight_tensor_parameter_223: name = "parameter_223" + original_name = "conv2d_66.w_0" shape = [256, 960, 1, 1] dtype = "float32" min_val = float("-0.47287") @@ -2460,6 +2684,7 @@ class Program_weight_tensor_parameter_223: class Program_weight_tensor_parameter_224: name = "parameter_224" + original_name = "batch_norm_53.b_0" shape = [960] dtype = "float32" min_val = float("-4.45904") @@ -2471,6 +2696,7 @@ class Program_weight_tensor_parameter_224: class Program_weight_tensor_parameter_225: name = "parameter_225" + original_name = "batch_norm_53.w_0" shape = [960] dtype = "float32" min_val = float("0.441641") @@ -2482,6 +2708,7 @@ class Program_weight_tensor_parameter_225: class Program_weight_tensor_parameter_226: name = "parameter_226" + original_name = "batch_norm_53.w_2" shape = [960] dtype = "float32" min_val = float("3.14311e-42") @@ -2493,6 +2720,7 @@ class Program_weight_tensor_parameter_226: class Program_weight_tensor_parameter_227: name = "parameter_227" + original_name = "batch_norm_53.w_1" shape = [960] dtype = "float32" min_val = float("-0.463321") @@ -2504,6 +2732,7 @@ class Program_weight_tensor_parameter_227: class Program_weight_tensor_parameter_228: name = "parameter_228" + original_name = "conv2d_65.w_0" shape = [960, 1, 5, 5] dtype = "float32" min_val = float("-0.0961546") @@ -2515,6 +2744,7 @@ class Program_weight_tensor_parameter_228: class Program_weight_tensor_parameter_229: name = "parameter_229" + original_name = "batch_norm_52.b_0" shape = [960] dtype = "float32" min_val = float("-1.29981") @@ -2526,6 +2756,7 @@ class Program_weight_tensor_parameter_229: class Program_weight_tensor_parameter_230: name = "parameter_230" + original_name = "batch_norm_52.w_0" shape = [960] dtype = "float32" min_val = float("0.149151") @@ -2537,6 +2768,7 @@ class Program_weight_tensor_parameter_230: class Program_weight_tensor_parameter_231: name = "parameter_231" + original_name = "batch_norm_52.w_2" shape = [960] dtype = "float32" min_val = float("1.7194e-25") @@ -2548,6 +2780,7 @@ class Program_weight_tensor_parameter_231: class Program_weight_tensor_parameter_232: name = "parameter_232" + original_name = "batch_norm_52.w_1" shape = [960] dtype = "float32" min_val = float("-0.479108") @@ -2559,6 +2792,7 @@ class Program_weight_tensor_parameter_232: class Program_weight_tensor_parameter_233: name = "parameter_233" + original_name = "conv2d_64.w_0" shape = [960, 160, 1, 1] dtype = "float32" min_val = float("-0.28934") @@ -2570,6 +2804,7 @@ class Program_weight_tensor_parameter_233: class Program_weight_tensor_parameter_234: name = "parameter_234" + original_name = "batch_norm_51.b_0" shape = [160] dtype = "float32" min_val = float("-0.994556") @@ -2581,6 +2816,7 @@ class Program_weight_tensor_parameter_234: class Program_weight_tensor_parameter_235: name = "parameter_235" + original_name = "batch_norm_51.w_0" shape = [160] dtype = "float32" min_val = float("0.379542") @@ -2592,6 +2828,7 @@ class Program_weight_tensor_parameter_235: class Program_weight_tensor_parameter_236: name = "parameter_236" + original_name = "batch_norm_51.w_2" shape = [160] dtype = "float32" min_val = float("0.000806212") @@ -2603,6 +2840,7 @@ class Program_weight_tensor_parameter_236: class Program_weight_tensor_parameter_237: name = "parameter_237" + original_name = "batch_norm_51.w_1" shape = [160] dtype = "float32" min_val = float("-1.81682") @@ -2614,6 +2852,7 @@ class Program_weight_tensor_parameter_237: class Program_weight_tensor_parameter_238: name = "parameter_238" + original_name = "conv2d_63.w_0" shape = [160, 1, 5, 5] dtype = "float32" min_val = float("-0.418787") @@ -2625,6 +2864,7 @@ class Program_weight_tensor_parameter_238: class Program_weight_tensor_parameter_239: name = "parameter_239" + original_name = "batch_norm_50.b_0" shape = [160] dtype = "float32" min_val = float("-2.50409") @@ -2636,6 +2876,7 @@ class Program_weight_tensor_parameter_239: class Program_weight_tensor_parameter_240: name = "parameter_240" + original_name = "batch_norm_50.w_0" shape = [160] dtype = "float32" min_val = float("-0.347695") @@ -2647,6 +2888,7 @@ class Program_weight_tensor_parameter_240: class Program_weight_tensor_parameter_241: name = "parameter_241" + original_name = "batch_norm_50.w_2" shape = [160] dtype = "float32" min_val = float("0.0051228") @@ -2658,6 +2900,7 @@ class Program_weight_tensor_parameter_241: class Program_weight_tensor_parameter_242: name = "parameter_242" + original_name = "batch_norm_50.w_1" shape = [160] dtype = "float32" min_val = float("-0.291155") @@ -2669,6 +2912,7 @@ class Program_weight_tensor_parameter_242: class Program_weight_tensor_parameter_243: name = "parameter_243" + original_name = "conv2d_62.w_0" shape = [160, 640, 1, 1] dtype = "float32" min_val = float("-0.258174") @@ -2680,6 +2924,7 @@ class Program_weight_tensor_parameter_243: class Program_weight_tensor_parameter_244: name = "parameter_244" + original_name = "batch_norm_49.b_0" shape = [640] dtype = "float32" min_val = float("-2.79675") @@ -2691,6 +2936,7 @@ class Program_weight_tensor_parameter_244: class Program_weight_tensor_parameter_245: name = "parameter_245" + original_name = "batch_norm_49.w_0" shape = [640] dtype = "float32" min_val = float("0.332761") @@ -2702,6 +2948,7 @@ class Program_weight_tensor_parameter_245: class Program_weight_tensor_parameter_246: name = "parameter_246" + original_name = "batch_norm_49.w_2" shape = [640] dtype = "float32" min_val = float("0.0577768") @@ -2713,6 +2960,7 @@ class Program_weight_tensor_parameter_246: class Program_weight_tensor_parameter_247: name = "parameter_247" + original_name = "batch_norm_49.w_1" shape = [640] dtype = "float32" min_val = float("-0.468454") @@ -2724,6 +2972,7 @@ class Program_weight_tensor_parameter_247: class Program_weight_tensor_parameter_248: name = "parameter_248" + original_name = "conv2d_61.w_0" shape = [640, 160, 1, 1] dtype = "float32" min_val = float("-0.332917") @@ -2735,6 +2984,7 @@ class Program_weight_tensor_parameter_248: class Program_weight_tensor_parameter_249: name = "parameter_249" + original_name = "batch_norm_48.b_0" shape = [160] dtype = "float32" min_val = float("-1.28488") @@ -2746,6 +2996,7 @@ class Program_weight_tensor_parameter_249: class Program_weight_tensor_parameter_250: name = "parameter_250" + original_name = "batch_norm_48.w_0" shape = [160] dtype = "float32" min_val = float("0.307933") @@ -2757,6 +3008,7 @@ class Program_weight_tensor_parameter_250: class Program_weight_tensor_parameter_251: name = "parameter_251" + original_name = "batch_norm_48.w_2" shape = [160] dtype = "float32" min_val = float("0.000500866") @@ -2768,6 +3020,7 @@ class Program_weight_tensor_parameter_251: class Program_weight_tensor_parameter_252: name = "parameter_252" + original_name = "batch_norm_48.w_1" shape = [160] dtype = "float32" min_val = float("-1.32769") @@ -2779,6 +3032,7 @@ class Program_weight_tensor_parameter_252: class Program_weight_tensor_parameter_253: name = "parameter_253" + original_name = "conv2d_60.w_0" shape = [160, 1, 3, 3] dtype = "float32" min_val = float("-0.318323") @@ -2790,6 +3044,7 @@ class Program_weight_tensor_parameter_253: class Program_weight_tensor_parameter_254: name = "parameter_254" + original_name = "conv2d_59.w_0" shape = [160, 256, 1, 1] dtype = "float32" min_val = float("-0.175727") @@ -2801,6 +3056,7 @@ class Program_weight_tensor_parameter_254: class Program_weight_tensor_parameter_255: name = "parameter_255" + original_name = "conv2d_58.w_0" shape = [64, 160, 1, 1] dtype = "float32" min_val = float("-0.159131") @@ -2812,6 +3068,7 @@ class Program_weight_tensor_parameter_255: class Program_weight_tensor_parameter_256: name = "parameter_256" + original_name = "batch_norm_47.b_0" shape = [160] dtype = "float32" min_val = float("-0.810279") @@ -2823,6 +3080,7 @@ class Program_weight_tensor_parameter_256: class Program_weight_tensor_parameter_257: name = "parameter_257" + original_name = "batch_norm_47.w_0" shape = [160] dtype = "float32" min_val = float("-2.13414") @@ -2834,6 +3092,7 @@ class Program_weight_tensor_parameter_257: class Program_weight_tensor_parameter_258: name = "parameter_258" + original_name = "batch_norm_47.w_2" shape = [160] dtype = "float32" min_val = float("4.13041e-08") @@ -2845,6 +3104,7 @@ class Program_weight_tensor_parameter_258: class Program_weight_tensor_parameter_259: name = "parameter_259" + original_name = "batch_norm_47.w_1" shape = [160] dtype = "float32" min_val = float("-0.106136") @@ -2856,6 +3116,7 @@ class Program_weight_tensor_parameter_259: class Program_weight_tensor_parameter_260: name = "parameter_260" + original_name = "conv2d_56.w_0" shape = [160, 1, 3, 3] dtype = "float32" min_val = float("-0.078817") @@ -2867,6 +3128,7 @@ class Program_weight_tensor_parameter_260: class Program_weight_tensor_parameter_261: name = "parameter_261" + original_name = "conv2d_57.w_0" shape = [64, 160, 1, 1] dtype = "float32" min_val = float("-0.199013") @@ -2878,6 +3140,7 @@ class Program_weight_tensor_parameter_261: class Program_weight_tensor_parameter_262: name = "parameter_262" + original_name = "batch_norm_46.b_0" shape = [160] dtype = "float32" min_val = float("-1.29938") @@ -2889,6 +3152,7 @@ class Program_weight_tensor_parameter_262: class Program_weight_tensor_parameter_263: name = "parameter_263" + original_name = "batch_norm_46.w_0" shape = [160] dtype = "float32" min_val = float("-1.49607") @@ -2900,6 +3164,7 @@ class Program_weight_tensor_parameter_263: class Program_weight_tensor_parameter_264: name = "parameter_264" + original_name = "batch_norm_46.w_2" shape = [160] dtype = "float32" min_val = float("9.42994e-08") @@ -2911,6 +3176,7 @@ class Program_weight_tensor_parameter_264: class Program_weight_tensor_parameter_265: name = "parameter_265" + original_name = "batch_norm_46.w_1" shape = [160] dtype = "float32" min_val = float("-0.668775") @@ -2922,6 +3188,7 @@ class Program_weight_tensor_parameter_265: class Program_weight_tensor_parameter_266: name = "parameter_266" + original_name = "conv2d_55.w_0" shape = [160, 1, 3, 3] dtype = "float32" min_val = float("-0.0755167") @@ -2933,6 +3200,7 @@ class Program_weight_tensor_parameter_266: class Program_weight_tensor_parameter_267: name = "parameter_267" + original_name = "conv2d_54.w_0" shape = [256, 160, 1, 1] dtype = "float32" min_val = float("-0.268973") @@ -2944,6 +3212,7 @@ class Program_weight_tensor_parameter_267: class Program_weight_tensor_parameter_268: name = "parameter_268" + original_name = "batch_norm_45.b_0" shape = [160] dtype = "float32" min_val = float("-2.8005") @@ -2955,6 +3224,7 @@ class Program_weight_tensor_parameter_268: class Program_weight_tensor_parameter_269: name = "parameter_269" + original_name = "batch_norm_45.w_0" shape = [160] dtype = "float32" min_val = float("-1.81616") @@ -2966,6 +3236,7 @@ class Program_weight_tensor_parameter_269: class Program_weight_tensor_parameter_270: name = "parameter_270" + original_name = "batch_norm_45.w_2" shape = [160] dtype = "float32" min_val = float("0.0344565") @@ -2977,6 +3248,7 @@ class Program_weight_tensor_parameter_270: class Program_weight_tensor_parameter_271: name = "parameter_271" + original_name = "batch_norm_45.w_1" shape = [160] dtype = "float32" min_val = float("-6.96094") @@ -2988,6 +3260,7 @@ class Program_weight_tensor_parameter_271: class Program_weight_tensor_parameter_272: name = "parameter_272" + original_name = "batch_norm_44.b_0" shape = [160] dtype = "float32" min_val = float("-1.69813") @@ -2999,6 +3272,7 @@ class Program_weight_tensor_parameter_272: class Program_weight_tensor_parameter_273: name = "parameter_273" + original_name = "batch_norm_44.w_0" shape = [160] dtype = "float32" min_val = float("0.217909") @@ -3010,6 +3284,7 @@ class Program_weight_tensor_parameter_273: class Program_weight_tensor_parameter_274: name = "parameter_274" + original_name = "batch_norm_44.w_2" shape = [160] dtype = "float32" min_val = float("0.0255919") @@ -3021,6 +3296,7 @@ class Program_weight_tensor_parameter_274: class Program_weight_tensor_parameter_275: name = "parameter_275" + original_name = "batch_norm_44.w_1" shape = [160] dtype = "float32" min_val = float("-0.820557") @@ -3032,6 +3308,7 @@ class Program_weight_tensor_parameter_275: class Program_weight_tensor_parameter_276: name = "parameter_276" + original_name = "conv2d_53.w_0" shape = [160, 640, 1, 1] dtype = "float32" min_val = float("-0.223839") @@ -3043,6 +3320,7 @@ class Program_weight_tensor_parameter_276: class Program_weight_tensor_parameter_277: name = "parameter_277" + original_name = "batch_norm_43.b_0" shape = [640] dtype = "float32" min_val = float("-2.5213") @@ -3054,6 +3332,7 @@ class Program_weight_tensor_parameter_277: class Program_weight_tensor_parameter_278: name = "parameter_278" + original_name = "batch_norm_43.w_0" shape = [640] dtype = "float32" min_val = float("0.114301") @@ -3065,6 +3344,7 @@ class Program_weight_tensor_parameter_278: class Program_weight_tensor_parameter_279: name = "parameter_279" + original_name = "batch_norm_43.w_2" shape = [640] dtype = "float32" min_val = float("3.024e-42") @@ -3076,6 +3356,7 @@ class Program_weight_tensor_parameter_279: class Program_weight_tensor_parameter_280: name = "parameter_280" + original_name = "batch_norm_43.w_1" shape = [640] dtype = "float32" min_val = float("-0.272962") @@ -3087,6 +3368,7 @@ class Program_weight_tensor_parameter_280: class Program_weight_tensor_parameter_281: name = "parameter_281" + original_name = "conv2d_52.w_0" shape = [640, 1, 3, 3] dtype = "float32" min_val = float("-0.225781") @@ -3098,6 +3380,7 @@ class Program_weight_tensor_parameter_281: class Program_weight_tensor_parameter_282: name = "parameter_282" + original_name = "batch_norm_42.b_0" shape = [640] dtype = "float32" min_val = float("-1.08465") @@ -3109,6 +3392,7 @@ class Program_weight_tensor_parameter_282: class Program_weight_tensor_parameter_283: name = "parameter_283" + original_name = "batch_norm_42.w_0" shape = [640] dtype = "float32" min_val = float("0.135105") @@ -3120,6 +3404,7 @@ class Program_weight_tensor_parameter_283: class Program_weight_tensor_parameter_284: name = "parameter_284" + original_name = "batch_norm_42.w_2" shape = [640] dtype = "float32" min_val = float("3.28602e-26") @@ -3131,6 +3416,7 @@ class Program_weight_tensor_parameter_284: class Program_weight_tensor_parameter_285: name = "parameter_285" + original_name = "batch_norm_42.w_1" shape = [640] dtype = "float32" min_val = float("-0.403123") @@ -3142,6 +3428,7 @@ class Program_weight_tensor_parameter_285: class Program_weight_tensor_parameter_286: name = "parameter_286" + original_name = "conv2d_51.w_0" shape = [640, 160, 1, 1] dtype = "float32" min_val = float("-0.297655") @@ -3153,6 +3440,7 @@ class Program_weight_tensor_parameter_286: class Program_weight_tensor_parameter_287: name = "parameter_287" + original_name = "batch_norm_41.b_0" shape = [160] dtype = "float32" min_val = float("-1.36381") @@ -3164,6 +3452,7 @@ class Program_weight_tensor_parameter_287: class Program_weight_tensor_parameter_288: name = "parameter_288" + original_name = "batch_norm_41.w_0" shape = [160] dtype = "float32" min_val = float("0.300887") @@ -3175,6 +3464,7 @@ class Program_weight_tensor_parameter_288: class Program_weight_tensor_parameter_289: name = "parameter_289" + original_name = "batch_norm_41.w_2" shape = [160] dtype = "float32" min_val = float("5.71306e-05") @@ -3186,6 +3476,7 @@ class Program_weight_tensor_parameter_289: class Program_weight_tensor_parameter_290: name = "parameter_290" + original_name = "batch_norm_41.w_1" shape = [160] dtype = "float32" min_val = float("-1.50147") @@ -3197,6 +3488,7 @@ class Program_weight_tensor_parameter_290: class Program_weight_tensor_parameter_291: name = "parameter_291" + original_name = "conv2d_50.w_0" shape = [160, 1, 3, 3] dtype = "float32" min_val = float("-0.280749") @@ -3208,6 +3500,7 @@ class Program_weight_tensor_parameter_291: class Program_weight_tensor_parameter_292: name = "parameter_292" + original_name = "conv2d_49.w_0" shape = [160, 256, 1, 1] dtype = "float32" min_val = float("-0.232644") @@ -3219,6 +3512,7 @@ class Program_weight_tensor_parameter_292: class Program_weight_tensor_parameter_293: name = "parameter_293" + original_name = "conv2d_48.w_0" shape = [64, 160, 1, 1] dtype = "float32" min_val = float("-0.259151") @@ -3230,6 +3524,7 @@ class Program_weight_tensor_parameter_293: class Program_weight_tensor_parameter_294: name = "parameter_294" + original_name = "batch_norm_40.b_0" shape = [160] dtype = "float32" min_val = float("-0.742698") @@ -3241,6 +3536,7 @@ class Program_weight_tensor_parameter_294: class Program_weight_tensor_parameter_295: name = "parameter_295" + original_name = "batch_norm_40.w_0" shape = [160] dtype = "float32" min_val = float("-2.24404") @@ -3252,6 +3548,7 @@ class Program_weight_tensor_parameter_295: class Program_weight_tensor_parameter_296: name = "parameter_296" + original_name = "batch_norm_40.w_2" shape = [160] dtype = "float32" min_val = float("7.46714e-05") @@ -3263,6 +3560,7 @@ class Program_weight_tensor_parameter_296: class Program_weight_tensor_parameter_297: name = "parameter_297" + original_name = "batch_norm_40.w_1" shape = [160] dtype = "float32" min_val = float("-0.287931") @@ -3274,6 +3572,7 @@ class Program_weight_tensor_parameter_297: class Program_weight_tensor_parameter_298: name = "parameter_298" + original_name = "conv2d_46.w_0" shape = [160, 1, 3, 3] dtype = "float32" min_val = float("-0.114036") @@ -3285,6 +3584,7 @@ class Program_weight_tensor_parameter_298: class Program_weight_tensor_parameter_299: name = "parameter_299" + original_name = "conv2d_47.w_0" shape = [64, 160, 1, 1] dtype = "float32" min_val = float("-0.22939") @@ -3296,6 +3596,7 @@ class Program_weight_tensor_parameter_299: class Program_weight_tensor_parameter_300: name = "parameter_300" + original_name = "batch_norm_39.b_0" shape = [160] dtype = "float32" min_val = float("-1.61864") @@ -3307,6 +3608,7 @@ class Program_weight_tensor_parameter_300: class Program_weight_tensor_parameter_301: name = "parameter_301" + original_name = "batch_norm_39.w_0" shape = [160] dtype = "float32" min_val = float("-0.436589") @@ -3318,6 +3620,7 @@ class Program_weight_tensor_parameter_301: class Program_weight_tensor_parameter_302: name = "parameter_302" + original_name = "batch_norm_39.w_2" shape = [160] dtype = "float32" min_val = float("9.06621e-05") @@ -3329,6 +3632,7 @@ class Program_weight_tensor_parameter_302: class Program_weight_tensor_parameter_303: name = "parameter_303" + original_name = "batch_norm_39.w_1" shape = [160] dtype = "float32" min_val = float("-0.614662") @@ -3340,6 +3644,7 @@ class Program_weight_tensor_parameter_303: class Program_weight_tensor_parameter_304: name = "parameter_304" + original_name = "conv2d_45.w_0" shape = [160, 1, 3, 3] dtype = "float32" min_val = float("-0.103601") @@ -3351,6 +3656,7 @@ class Program_weight_tensor_parameter_304: class Program_weight_tensor_parameter_305: name = "parameter_305" + original_name = "conv2d_44.w_0" shape = [256, 160, 1, 1] dtype = "float32" min_val = float("-0.226674") @@ -3362,6 +3668,7 @@ class Program_weight_tensor_parameter_305: class Program_weight_tensor_parameter_306: name = "parameter_306" + original_name = "batch_norm_38.b_0" shape = [160] dtype = "float32" min_val = float("-2.93067") @@ -3373,6 +3680,7 @@ class Program_weight_tensor_parameter_306: class Program_weight_tensor_parameter_307: name = "parameter_307" + original_name = "batch_norm_38.w_0" shape = [160] dtype = "float32" min_val = float("-0.11005") @@ -3384,6 +3692,7 @@ class Program_weight_tensor_parameter_307: class Program_weight_tensor_parameter_308: name = "parameter_308" + original_name = "batch_norm_38.w_2" shape = [160] dtype = "float32" min_val = float("0.0346149") @@ -3395,6 +3704,7 @@ class Program_weight_tensor_parameter_308: class Program_weight_tensor_parameter_309: name = "parameter_309" + original_name = "batch_norm_38.w_1" shape = [160] dtype = "float32" min_val = float("-3.16387") @@ -3406,6 +3716,7 @@ class Program_weight_tensor_parameter_309: class Program_weight_tensor_parameter_310: name = "parameter_310" + original_name = "batch_norm_37.b_0" shape = [160] dtype = "float32" min_val = float("-2.04077") @@ -3417,6 +3728,7 @@ class Program_weight_tensor_parameter_310: class Program_weight_tensor_parameter_311: name = "parameter_311" + original_name = "batch_norm_37.w_0" shape = [160] dtype = "float32" min_val = float("-0.00528844") @@ -3428,6 +3740,7 @@ class Program_weight_tensor_parameter_311: class Program_weight_tensor_parameter_312: name = "parameter_312" + original_name = "batch_norm_37.w_2" shape = [160] dtype = "float32" min_val = float("0.00423533") @@ -3439,6 +3752,7 @@ class Program_weight_tensor_parameter_312: class Program_weight_tensor_parameter_313: name = "parameter_313" + original_name = "batch_norm_37.w_1" shape = [160] dtype = "float32" min_val = float("-0.186867") @@ -3450,6 +3764,7 @@ class Program_weight_tensor_parameter_313: class Program_weight_tensor_parameter_314: name = "parameter_314" + original_name = "conv2d_43.w_0" shape = [160, 640, 1, 1] dtype = "float32" min_val = float("-0.35406") @@ -3461,6 +3776,7 @@ class Program_weight_tensor_parameter_314: class Program_weight_tensor_parameter_315: name = "parameter_315" + original_name = "batch_norm_36.b_0" shape = [640] dtype = "float32" min_val = float("-3.63355") @@ -3472,6 +3788,7 @@ class Program_weight_tensor_parameter_315: class Program_weight_tensor_parameter_316: name = "parameter_316" + original_name = "batch_norm_36.w_0" shape = [640] dtype = "float32" min_val = float("0.22692") @@ -3483,6 +3800,7 @@ class Program_weight_tensor_parameter_316: class Program_weight_tensor_parameter_317: name = "parameter_317" + original_name = "batch_norm_36.w_2" shape = [640] dtype = "float32" min_val = float("1.80244e-07") @@ -3494,6 +3812,7 @@ class Program_weight_tensor_parameter_317: class Program_weight_tensor_parameter_318: name = "parameter_318" + original_name = "batch_norm_36.w_1" shape = [640] dtype = "float32" min_val = float("-0.49831") @@ -3505,6 +3824,7 @@ class Program_weight_tensor_parameter_318: class Program_weight_tensor_parameter_319: name = "parameter_319" + original_name = "conv2d_42.w_0" shape = [640, 160, 1, 1] dtype = "float32" min_val = float("-0.402277") @@ -3516,6 +3836,7 @@ class Program_weight_tensor_parameter_319: class Program_weight_tensor_parameter_320: name = "parameter_320" + original_name = "batch_norm_35.b_0" shape = [160] dtype = "float32" min_val = float("-1.28533") @@ -3527,6 +3848,7 @@ class Program_weight_tensor_parameter_320: class Program_weight_tensor_parameter_321: name = "parameter_321" + original_name = "batch_norm_35.w_0" shape = [160] dtype = "float32" min_val = float("0.0855914") @@ -3538,6 +3860,7 @@ class Program_weight_tensor_parameter_321: class Program_weight_tensor_parameter_322: name = "parameter_322" + original_name = "batch_norm_35.w_2" shape = [160] dtype = "float32" min_val = float("4.09332e-05") @@ -3549,6 +3872,7 @@ class Program_weight_tensor_parameter_322: class Program_weight_tensor_parameter_323: name = "parameter_323" + original_name = "batch_norm_35.w_1" shape = [160] dtype = "float32" min_val = float("-0.244023") @@ -3560,6 +3884,7 @@ class Program_weight_tensor_parameter_323: class Program_weight_tensor_parameter_324: name = "parameter_324" + original_name = "conv2d_41.w_0" shape = [160, 1, 3, 3] dtype = "float32" min_val = float("-0.271397") @@ -3571,6 +3896,7 @@ class Program_weight_tensor_parameter_324: class Program_weight_tensor_parameter_325: name = "parameter_325" + original_name = "conv2d_40.w_0" shape = [160, 256, 1, 1] dtype = "float32" min_val = float("-0.162459") @@ -3582,6 +3908,7 @@ class Program_weight_tensor_parameter_325: class Program_weight_tensor_parameter_326: name = "parameter_326" + original_name = "conv2d_39.w_0" shape = [64, 160, 1, 1] dtype = "float32" min_val = float("-0.171835") @@ -3593,6 +3920,7 @@ class Program_weight_tensor_parameter_326: class Program_weight_tensor_parameter_327: name = "parameter_327" + original_name = "batch_norm_34.b_0" shape = [160] dtype = "float32" min_val = float("-1.11268") @@ -3604,6 +3932,7 @@ class Program_weight_tensor_parameter_327: class Program_weight_tensor_parameter_328: name = "parameter_328" + original_name = "batch_norm_34.w_0" shape = [160] dtype = "float32" min_val = float("-0.437674") @@ -3615,6 +3944,7 @@ class Program_weight_tensor_parameter_328: class Program_weight_tensor_parameter_329: name = "parameter_329" + original_name = "batch_norm_34.w_2" shape = [160] dtype = "float32" min_val = float("0.000188232") @@ -3626,6 +3956,7 @@ class Program_weight_tensor_parameter_329: class Program_weight_tensor_parameter_330: name = "parameter_330" + original_name = "batch_norm_34.w_1" shape = [160] dtype = "float32" min_val = float("-0.290062") @@ -3637,6 +3968,7 @@ class Program_weight_tensor_parameter_330: class Program_weight_tensor_parameter_331: name = "parameter_331" + original_name = "conv2d_37.w_0" shape = [160, 1, 3, 3] dtype = "float32" min_val = float("-0.102145") @@ -3648,6 +3980,7 @@ class Program_weight_tensor_parameter_331: class Program_weight_tensor_parameter_332: name = "parameter_332" + original_name = "conv2d_38.w_0" shape = [64, 160, 1, 1] dtype = "float32" min_val = float("-0.182986") @@ -3659,6 +3992,7 @@ class Program_weight_tensor_parameter_332: class Program_weight_tensor_parameter_333: name = "parameter_333" + original_name = "batch_norm_33.b_0" shape = [160] dtype = "float32" min_val = float("-1.61702") @@ -3670,6 +4004,7 @@ class Program_weight_tensor_parameter_333: class Program_weight_tensor_parameter_334: name = "parameter_334" + original_name = "batch_norm_33.w_0" shape = [160] dtype = "float32" min_val = float("-1.41779") @@ -3681,6 +4016,7 @@ class Program_weight_tensor_parameter_334: class Program_weight_tensor_parameter_335: name = "parameter_335" + original_name = "batch_norm_33.w_2" shape = [160] dtype = "float32" min_val = float("0.000206087") @@ -3692,6 +4028,7 @@ class Program_weight_tensor_parameter_335: class Program_weight_tensor_parameter_336: name = "parameter_336" + original_name = "batch_norm_33.w_1" shape = [160] dtype = "float32" min_val = float("-0.385786") @@ -3703,6 +4040,7 @@ class Program_weight_tensor_parameter_336: class Program_weight_tensor_parameter_337: name = "parameter_337" + original_name = "conv2d_36.w_0" shape = [160, 1, 3, 3] dtype = "float32" min_val = float("-0.0948321") @@ -3714,6 +4052,7 @@ class Program_weight_tensor_parameter_337: class Program_weight_tensor_parameter_338: name = "parameter_338" + original_name = "conv2d_35.w_0" shape = [256, 160, 1, 1] dtype = "float32" min_val = float("-0.203254") @@ -3725,6 +4064,7 @@ class Program_weight_tensor_parameter_338: class Program_weight_tensor_parameter_339: name = "parameter_339" + original_name = "batch_norm_32.b_0" shape = [160] dtype = "float32" min_val = float("-3.9754") @@ -3736,6 +4076,7 @@ class Program_weight_tensor_parameter_339: class Program_weight_tensor_parameter_340: name = "parameter_340" + original_name = "batch_norm_32.w_0" shape = [160] dtype = "float32" min_val = float("-0.0126311") @@ -3747,6 +4088,7 @@ class Program_weight_tensor_parameter_340: class Program_weight_tensor_parameter_341: name = "parameter_341" + original_name = "batch_norm_32.w_2" shape = [160] dtype = "float32" min_val = float("0.0127402") @@ -3758,6 +4100,7 @@ class Program_weight_tensor_parameter_341: class Program_weight_tensor_parameter_342: name = "parameter_342" + original_name = "batch_norm_32.w_1" shape = [160] dtype = "float32" min_val = float("-1.63408") @@ -3769,6 +4112,7 @@ class Program_weight_tensor_parameter_342: class Program_weight_tensor_parameter_343: name = "parameter_343" + original_name = "batch_norm_31.b_0" shape = [160] dtype = "float32" min_val = float("-1.65281") @@ -3780,6 +4124,7 @@ class Program_weight_tensor_parameter_343: class Program_weight_tensor_parameter_344: name = "parameter_344" + original_name = "batch_norm_31.w_0" shape = [160] dtype = "float32" min_val = float("-0.0249717") @@ -3791,6 +4136,7 @@ class Program_weight_tensor_parameter_344: class Program_weight_tensor_parameter_345: name = "parameter_345" + original_name = "batch_norm_31.w_2" shape = [160] dtype = "float32" min_val = float("0.0352698") @@ -3802,6 +4148,7 @@ class Program_weight_tensor_parameter_345: class Program_weight_tensor_parameter_346: name = "parameter_346" + original_name = "batch_norm_31.w_1" shape = [160] dtype = "float32" min_val = float("-0.718378") @@ -3813,6 +4160,7 @@ class Program_weight_tensor_parameter_346: class Program_weight_tensor_parameter_347: name = "parameter_347" + original_name = "conv2d_34.w_0" shape = [160, 640, 1, 1] dtype = "float32" min_val = float("-0.259631") @@ -3824,6 +4172,7 @@ class Program_weight_tensor_parameter_347: class Program_weight_tensor_parameter_348: name = "parameter_348" + original_name = "batch_norm_30.b_0" shape = [640] dtype = "float32" min_val = float("-3.21629") @@ -3835,6 +4184,7 @@ class Program_weight_tensor_parameter_348: class Program_weight_tensor_parameter_349: name = "parameter_349" + original_name = "batch_norm_30.w_0" shape = [640] dtype = "float32" min_val = float("0.29673") @@ -3846,6 +4196,7 @@ class Program_weight_tensor_parameter_349: class Program_weight_tensor_parameter_350: name = "parameter_350" + original_name = "batch_norm_30.w_2" shape = [640] dtype = "float32" min_val = float("3.50885e-42") @@ -3857,6 +4208,7 @@ class Program_weight_tensor_parameter_350: class Program_weight_tensor_parameter_351: name = "parameter_351" + original_name = "batch_norm_30.w_1" shape = [640] dtype = "float32" min_val = float("-0.164406") @@ -3868,6 +4220,7 @@ class Program_weight_tensor_parameter_351: class Program_weight_tensor_parameter_352: name = "parameter_352" + original_name = "conv2d_33.w_0" shape = [640, 1, 3, 3] dtype = "float32" min_val = float("-0.202809") @@ -3879,6 +4232,7 @@ class Program_weight_tensor_parameter_352: class Program_weight_tensor_parameter_353: name = "parameter_353" + original_name = "batch_norm_29.b_0" shape = [640] dtype = "float32" min_val = float("-1.46714") @@ -3890,6 +4244,7 @@ class Program_weight_tensor_parameter_353: class Program_weight_tensor_parameter_354: name = "parameter_354" + original_name = "batch_norm_29.w_0" shape = [640] dtype = "float32" min_val = float("0.202068") @@ -3901,6 +4256,7 @@ class Program_weight_tensor_parameter_354: class Program_weight_tensor_parameter_355: name = "parameter_355" + original_name = "batch_norm_29.w_2" shape = [640] dtype = "float32" min_val = float("4.9806e-26") @@ -3912,6 +4268,7 @@ class Program_weight_tensor_parameter_355: class Program_weight_tensor_parameter_356: name = "parameter_356" + original_name = "batch_norm_29.w_1" shape = [640] dtype = "float32" min_val = float("-0.496652") @@ -3923,6 +4280,7 @@ class Program_weight_tensor_parameter_356: class Program_weight_tensor_parameter_357: name = "parameter_357" + original_name = "conv2d_32.w_0" shape = [640, 160, 1, 1] dtype = "float32" min_val = float("-0.252023") @@ -3934,6 +4292,7 @@ class Program_weight_tensor_parameter_357: class Program_weight_tensor_parameter_358: name = "parameter_358" + original_name = "batch_norm_28.b_0" shape = [160] dtype = "float32" min_val = float("-1.37069") @@ -3945,6 +4304,7 @@ class Program_weight_tensor_parameter_358: class Program_weight_tensor_parameter_359: name = "parameter_359" + original_name = "batch_norm_28.w_0" shape = [160] dtype = "float32" min_val = float("0.110753") @@ -3956,6 +4316,7 @@ class Program_weight_tensor_parameter_359: class Program_weight_tensor_parameter_360: name = "parameter_360" + original_name = "batch_norm_28.w_2" shape = [160] dtype = "float32" min_val = float("3.50051e-05") @@ -3967,6 +4328,7 @@ class Program_weight_tensor_parameter_360: class Program_weight_tensor_parameter_361: name = "parameter_361" + original_name = "batch_norm_28.w_1" shape = [160] dtype = "float32" min_val = float("-0.289872") @@ -3978,6 +4340,7 @@ class Program_weight_tensor_parameter_361: class Program_weight_tensor_parameter_362: name = "parameter_362" + original_name = "conv2d_31.w_0" shape = [160, 1, 3, 3] dtype = "float32" min_val = float("-0.28049") @@ -3989,6 +4352,7 @@ class Program_weight_tensor_parameter_362: class Program_weight_tensor_parameter_363: name = "parameter_363" + original_name = "conv2d_30.w_0" shape = [160, 256, 1, 1] dtype = "float32" min_val = float("-0.212941") @@ -4000,6 +4364,7 @@ class Program_weight_tensor_parameter_363: class Program_weight_tensor_parameter_364: name = "parameter_364" + original_name = "conv2d_29.w_0" shape = [64, 160, 1, 1] dtype = "float32" min_val = float("-0.136518") @@ -4011,6 +4376,7 @@ class Program_weight_tensor_parameter_364: class Program_weight_tensor_parameter_365: name = "parameter_365" + original_name = "batch_norm_27.b_0" shape = [160] dtype = "float32" min_val = float("-0.88704") @@ -4022,6 +4388,7 @@ class Program_weight_tensor_parameter_365: class Program_weight_tensor_parameter_366: name = "parameter_366" + original_name = "batch_norm_27.w_0" shape = [160] dtype = "float32" min_val = float("-1.17055") @@ -4033,6 +4400,7 @@ class Program_weight_tensor_parameter_366: class Program_weight_tensor_parameter_367: name = "parameter_367" + original_name = "batch_norm_27.w_2" shape = [160] dtype = "float32" min_val = float("7.55707e-07") @@ -4044,6 +4412,7 @@ class Program_weight_tensor_parameter_367: class Program_weight_tensor_parameter_368: name = "parameter_368" + original_name = "batch_norm_27.w_1" shape = [160] dtype = "float32" min_val = float("-0.369075") @@ -4055,6 +4424,7 @@ class Program_weight_tensor_parameter_368: class Program_weight_tensor_parameter_369: name = "parameter_369" + original_name = "conv2d_27.w_0" shape = [160, 1, 3, 3] dtype = "float32" min_val = float("-0.0654211") @@ -4066,6 +4436,7 @@ class Program_weight_tensor_parameter_369: class Program_weight_tensor_parameter_370: name = "parameter_370" + original_name = "conv2d_28.w_0" shape = [64, 160, 1, 1] dtype = "float32" min_val = float("-0.200566") @@ -4077,6 +4448,7 @@ class Program_weight_tensor_parameter_370: class Program_weight_tensor_parameter_371: name = "parameter_371" + original_name = "batch_norm_26.b_0" shape = [160] dtype = "float32" min_val = float("-1.31181") @@ -4088,6 +4460,7 @@ class Program_weight_tensor_parameter_371: class Program_weight_tensor_parameter_372: name = "parameter_372" + original_name = "batch_norm_26.w_0" shape = [160] dtype = "float32" min_val = float("-1.59337") @@ -4099,6 +4472,7 @@ class Program_weight_tensor_parameter_372: class Program_weight_tensor_parameter_373: name = "parameter_373" + original_name = "batch_norm_26.w_2" shape = [160] dtype = "float32" min_val = float("5.81298e-07") @@ -4110,6 +4484,7 @@ class Program_weight_tensor_parameter_373: class Program_weight_tensor_parameter_374: name = "parameter_374" + original_name = "batch_norm_26.w_1" shape = [160] dtype = "float32" min_val = float("-0.329884") @@ -4121,6 +4496,7 @@ class Program_weight_tensor_parameter_374: class Program_weight_tensor_parameter_375: name = "parameter_375" + original_name = "conv2d_26.w_0" shape = [160, 1, 3, 3] dtype = "float32" min_val = float("-0.0663044") @@ -4132,6 +4508,7 @@ class Program_weight_tensor_parameter_375: class Program_weight_tensor_parameter_376: name = "parameter_376" + original_name = "conv2d_25.w_0" shape = [256, 160, 1, 1] dtype = "float32" min_val = float("-0.195392") @@ -4143,6 +4520,7 @@ class Program_weight_tensor_parameter_376: class Program_weight_tensor_parameter_377: name = "parameter_377" + original_name = "batch_norm_25.b_0" shape = [160] dtype = "float32" min_val = float("-2.6699") @@ -4154,6 +4532,7 @@ class Program_weight_tensor_parameter_377: class Program_weight_tensor_parameter_378: name = "parameter_378" + original_name = "batch_norm_25.w_0" shape = [160] dtype = "float32" min_val = float("-0.70583") @@ -4165,6 +4544,7 @@ class Program_weight_tensor_parameter_378: class Program_weight_tensor_parameter_379: name = "parameter_379" + original_name = "batch_norm_25.w_2" shape = [160] dtype = "float32" min_val = float("0.00078917") @@ -4176,6 +4556,7 @@ class Program_weight_tensor_parameter_379: class Program_weight_tensor_parameter_380: name = "parameter_380" + original_name = "batch_norm_25.w_1" shape = [160] dtype = "float32" min_val = float("-0.995739") @@ -4187,6 +4568,7 @@ class Program_weight_tensor_parameter_380: class Program_weight_tensor_parameter_381: name = "parameter_381" + original_name = "batch_norm_24.b_0" shape = [160] dtype = "float32" min_val = float("-1.85966") @@ -4198,6 +4580,7 @@ class Program_weight_tensor_parameter_381: class Program_weight_tensor_parameter_382: name = "parameter_382" + original_name = "batch_norm_24.w_0" shape = [160] dtype = "float32" min_val = float("0.167319") @@ -4209,6 +4592,7 @@ class Program_weight_tensor_parameter_382: class Program_weight_tensor_parameter_383: name = "parameter_383" + original_name = "batch_norm_24.w_2" shape = [160] dtype = "float32" min_val = float("0.0305356") @@ -4220,6 +4604,7 @@ class Program_weight_tensor_parameter_383: class Program_weight_tensor_parameter_384: name = "parameter_384" + original_name = "batch_norm_24.w_1" shape = [160] dtype = "float32" min_val = float("-0.613947") @@ -4231,6 +4616,7 @@ class Program_weight_tensor_parameter_384: class Program_weight_tensor_parameter_385: name = "parameter_385" + original_name = "conv2d_24.w_0" shape = [160, 640, 1, 1] dtype = "float32" min_val = float("-0.220669") @@ -4242,6 +4628,7 @@ class Program_weight_tensor_parameter_385: class Program_weight_tensor_parameter_386: name = "parameter_386" + original_name = "batch_norm_23.b_0" shape = [640] dtype = "float32" min_val = float("-4.31418") @@ -4253,6 +4640,7 @@ class Program_weight_tensor_parameter_386: class Program_weight_tensor_parameter_387: name = "parameter_387" + original_name = "batch_norm_23.w_0" shape = [640] dtype = "float32" min_val = float("0.125845") @@ -4264,6 +4652,7 @@ class Program_weight_tensor_parameter_387: class Program_weight_tensor_parameter_388: name = "parameter_388" + original_name = "batch_norm_23.w_2" shape = [640] dtype = "float32" min_val = float("2.804e-42") @@ -4275,6 +4664,7 @@ class Program_weight_tensor_parameter_388: class Program_weight_tensor_parameter_389: name = "parameter_389" + original_name = "batch_norm_23.w_1" shape = [640] dtype = "float32" min_val = float("-0.457522") @@ -4286,6 +4676,7 @@ class Program_weight_tensor_parameter_389: class Program_weight_tensor_parameter_390: name = "parameter_390" + original_name = "conv2d_23.w_0" shape = [640, 1, 5, 5] dtype = "float32" min_val = float("-0.128151") @@ -4297,6 +4688,7 @@ class Program_weight_tensor_parameter_390: class Program_weight_tensor_parameter_391: name = "parameter_391" + original_name = "batch_norm_22.b_0" shape = [640] dtype = "float32" min_val = float("-1.03593") @@ -4308,6 +4700,7 @@ class Program_weight_tensor_parameter_391: class Program_weight_tensor_parameter_392: name = "parameter_392" + original_name = "batch_norm_22.w_0" shape = [640] dtype = "float32" min_val = float("0.142163") @@ -4319,6 +4712,7 @@ class Program_weight_tensor_parameter_392: class Program_weight_tensor_parameter_393: name = "parameter_393" + original_name = "batch_norm_22.w_2" shape = [640] dtype = "float32" min_val = float("7.52699e-27") @@ -4330,6 +4724,7 @@ class Program_weight_tensor_parameter_393: class Program_weight_tensor_parameter_394: name = "parameter_394" + original_name = "batch_norm_22.w_1" shape = [640] dtype = "float32" min_val = float("-0.553602") @@ -4341,6 +4736,7 @@ class Program_weight_tensor_parameter_394: class Program_weight_tensor_parameter_395: name = "parameter_395" + original_name = "conv2d_22.w_0" shape = [640, 160, 1, 1] dtype = "float32" min_val = float("-0.334139") @@ -4352,6 +4748,7 @@ class Program_weight_tensor_parameter_395: class Program_weight_tensor_parameter_396: name = "parameter_396" + original_name = "batch_norm_21.b_0" shape = [160] dtype = "float32" min_val = float("-1.33482") @@ -4363,6 +4760,7 @@ class Program_weight_tensor_parameter_396: class Program_weight_tensor_parameter_397: name = "parameter_397" + original_name = "batch_norm_21.w_0" shape = [160] dtype = "float32" min_val = float("0.272243") @@ -4374,6 +4772,7 @@ class Program_weight_tensor_parameter_397: class Program_weight_tensor_parameter_398: name = "parameter_398" + original_name = "batch_norm_21.w_2" shape = [160] dtype = "float32" min_val = float("2.18665e-06") @@ -4385,6 +4784,7 @@ class Program_weight_tensor_parameter_398: class Program_weight_tensor_parameter_399: name = "parameter_399" + original_name = "batch_norm_21.w_1" shape = [160] dtype = "float32" min_val = float("-0.219335") @@ -4396,6 +4796,7 @@ class Program_weight_tensor_parameter_399: class Program_weight_tensor_parameter_400: name = "parameter_400" + original_name = "conv2d_21.w_0" shape = [160, 1, 3, 3] dtype = "float32" min_val = float("-0.244518") @@ -4407,6 +4808,7 @@ class Program_weight_tensor_parameter_400: class Program_weight_tensor_parameter_401: name = "parameter_401" + original_name = "batch_norm_20.b_0" shape = [160] dtype = "float32" min_val = float("-1.40237") @@ -4418,6 +4820,7 @@ class Program_weight_tensor_parameter_401: class Program_weight_tensor_parameter_402: name = "parameter_402" + original_name = "batch_norm_20.w_0" shape = [160] dtype = "float32" min_val = float("-0.0870024") @@ -4429,6 +4832,7 @@ class Program_weight_tensor_parameter_402: class Program_weight_tensor_parameter_403: name = "parameter_403" + original_name = "batch_norm_20.w_2" shape = [160] dtype = "float32" min_val = float("0.0462044") @@ -4440,6 +4844,7 @@ class Program_weight_tensor_parameter_403: class Program_weight_tensor_parameter_404: name = "parameter_404" + original_name = "batch_norm_20.w_1" shape = [160] dtype = "float32" min_val = float("-0.570962") @@ -4451,6 +4856,7 @@ class Program_weight_tensor_parameter_404: class Program_weight_tensor_parameter_405: name = "parameter_405" + original_name = "conv2d_20.w_0" shape = [160, 640, 1, 1] dtype = "float32" min_val = float("-0.455014") @@ -4462,6 +4868,7 @@ class Program_weight_tensor_parameter_405: class Program_weight_tensor_parameter_406: name = "parameter_406" + original_name = "batch_norm_19.b_0" shape = [640] dtype = "float32" min_val = float("-3.30986") @@ -4473,6 +4880,7 @@ class Program_weight_tensor_parameter_406: class Program_weight_tensor_parameter_407: name = "parameter_407" + original_name = "batch_norm_19.w_0" shape = [640] dtype = "float32" min_val = float("0.19459") @@ -4484,6 +4892,7 @@ class Program_weight_tensor_parameter_407: class Program_weight_tensor_parameter_408: name = "parameter_408" + original_name = "batch_norm_19.w_2" shape = [640] dtype = "float32" min_val = float("3.03241e-42") @@ -4495,6 +4904,7 @@ class Program_weight_tensor_parameter_408: class Program_weight_tensor_parameter_409: name = "parameter_409" + original_name = "batch_norm_19.w_1" shape = [640] dtype = "float32" min_val = float("-0.255099") @@ -4506,6 +4916,7 @@ class Program_weight_tensor_parameter_409: class Program_weight_tensor_parameter_410: name = "parameter_410" + original_name = "conv2d_19.w_0" shape = [640, 1, 3, 3] dtype = "float32" min_val = float("-0.101531") @@ -4517,6 +4928,7 @@ class Program_weight_tensor_parameter_410: class Program_weight_tensor_parameter_411: name = "parameter_411" + original_name = "batch_norm_18.b_0" shape = [640] dtype = "float32" min_val = float("-1.04057") @@ -4528,6 +4940,7 @@ class Program_weight_tensor_parameter_411: class Program_weight_tensor_parameter_412: name = "parameter_412" + original_name = "batch_norm_18.w_0" shape = [640] dtype = "float32" min_val = float("0.116878") @@ -4539,6 +4952,7 @@ class Program_weight_tensor_parameter_412: class Program_weight_tensor_parameter_413: name = "parameter_413" + original_name = "batch_norm_18.w_2" shape = [640] dtype = "float32" min_val = float("6.63601e-27") @@ -4550,6 +4964,7 @@ class Program_weight_tensor_parameter_413: class Program_weight_tensor_parameter_414: name = "parameter_414" + original_name = "batch_norm_18.w_1" shape = [640] dtype = "float32" min_val = float("-0.463547") @@ -4561,6 +4976,7 @@ class Program_weight_tensor_parameter_414: class Program_weight_tensor_parameter_415: name = "parameter_415" + original_name = "conv2d_18.w_0" shape = [640, 160, 1, 1] dtype = "float32" min_val = float("-0.178339") @@ -4572,6 +4988,7 @@ class Program_weight_tensor_parameter_415: class Program_weight_tensor_parameter_416: name = "parameter_416" + original_name = "batch_norm_17.b_0" shape = [160] dtype = "float32" min_val = float("-1.19267") @@ -4583,6 +5000,7 @@ class Program_weight_tensor_parameter_416: class Program_weight_tensor_parameter_417: name = "parameter_417" + original_name = "batch_norm_17.w_0" shape = [160] dtype = "float32" min_val = float("-0.0838629") @@ -4594,6 +5012,7 @@ class Program_weight_tensor_parameter_417: class Program_weight_tensor_parameter_418: name = "parameter_418" + original_name = "batch_norm_17.w_2" shape = [160] dtype = "float32" min_val = float("1.44851e-06") @@ -4605,6 +5024,7 @@ class Program_weight_tensor_parameter_418: class Program_weight_tensor_parameter_419: name = "parameter_419" + original_name = "batch_norm_17.w_1" shape = [160] dtype = "float32" min_val = float("-0.184071") @@ -4616,6 +5036,7 @@ class Program_weight_tensor_parameter_419: class Program_weight_tensor_parameter_420: name = "parameter_420" + original_name = "conv2d_17.w_0" shape = [160, 1, 3, 3] dtype = "float32" min_val = float("-0.264274") @@ -4627,6 +5048,7 @@ class Program_weight_tensor_parameter_420: class Program_weight_tensor_parameter_421: name = "parameter_421" + original_name = "batch_norm_16.b_0" shape = [160] dtype = "float32" min_val = float("-0.599252") @@ -4638,6 +5060,7 @@ class Program_weight_tensor_parameter_421: class Program_weight_tensor_parameter_422: name = "parameter_422" + original_name = "batch_norm_16.w_0" shape = [160] dtype = "float32" min_val = float("-0.658567") @@ -4649,6 +5072,7 @@ class Program_weight_tensor_parameter_422: class Program_weight_tensor_parameter_423: name = "parameter_423" + original_name = "batch_norm_16.w_2" shape = [160] dtype = "float32" min_val = float("0.00204519") @@ -4660,6 +5084,7 @@ class Program_weight_tensor_parameter_423: class Program_weight_tensor_parameter_424: name = "parameter_424" + original_name = "batch_norm_16.w_1" shape = [160] dtype = "float32" min_val = float("-0.111787") @@ -4671,6 +5096,7 @@ class Program_weight_tensor_parameter_424: class Program_weight_tensor_parameter_425: name = "parameter_425" + original_name = "conv2d_16.w_0" shape = [160, 320, 1, 1] dtype = "float32" min_val = float("-0.248832") @@ -4682,6 +5108,7 @@ class Program_weight_tensor_parameter_425: class Program_weight_tensor_parameter_426: name = "parameter_426" + original_name = "batch_norm_15.b_0" shape = [320] dtype = "float32" min_val = float("-2.71934") @@ -4693,6 +5120,7 @@ class Program_weight_tensor_parameter_426: class Program_weight_tensor_parameter_427: name = "parameter_427" + original_name = "batch_norm_15.w_0" shape = [320] dtype = "float32" min_val = float("0.474345") @@ -4704,6 +5132,7 @@ class Program_weight_tensor_parameter_427: class Program_weight_tensor_parameter_428: name = "parameter_428" + original_name = "batch_norm_15.w_2" shape = [320] dtype = "float32" min_val = float("0.0218904") @@ -4715,6 +5144,7 @@ class Program_weight_tensor_parameter_428: class Program_weight_tensor_parameter_429: name = "parameter_429" + original_name = "batch_norm_15.w_1" shape = [320] dtype = "float32" min_val = float("-0.575006") @@ -4726,6 +5156,7 @@ class Program_weight_tensor_parameter_429: class Program_weight_tensor_parameter_430: name = "parameter_430" + original_name = "conv2d_15.w_0" shape = [320, 160, 1, 1] dtype = "float32" min_val = float("-0.312325") @@ -4737,6 +5168,7 @@ class Program_weight_tensor_parameter_430: class Program_weight_tensor_parameter_431: name = "parameter_431" + original_name = "batch_norm_14.b_0" shape = [160] dtype = "float32" min_val = float("-2.75148") @@ -4748,6 +5180,7 @@ class Program_weight_tensor_parameter_431: class Program_weight_tensor_parameter_432: name = "parameter_432" + original_name = "batch_norm_14.w_0" shape = [160] dtype = "float32" min_val = float("0.108617") @@ -4759,6 +5192,7 @@ class Program_weight_tensor_parameter_432: class Program_weight_tensor_parameter_433: name = "parameter_433" + original_name = "batch_norm_14.w_2" shape = [160] dtype = "float32" min_val = float("0.0644808") @@ -4770,6 +5204,7 @@ class Program_weight_tensor_parameter_433: class Program_weight_tensor_parameter_434: name = "parameter_434" + original_name = "batch_norm_14.w_1" shape = [160] dtype = "float32" min_val = float("-2.13859") @@ -4781,6 +5216,7 @@ class Program_weight_tensor_parameter_434: class Program_weight_tensor_parameter_435: name = "parameter_435" + original_name = "conv2d_14.w_0" shape = [160, 480, 1, 1] dtype = "float32" min_val = float("-0.237586") @@ -4792,6 +5228,7 @@ class Program_weight_tensor_parameter_435: class Program_weight_tensor_parameter_436: name = "parameter_436" + original_name = "batch_norm_13.b_0" shape = [480] dtype = "float32" min_val = float("-1.09317") @@ -4803,6 +5240,7 @@ class Program_weight_tensor_parameter_436: class Program_weight_tensor_parameter_437: name = "parameter_437" + original_name = "batch_norm_13.w_0" shape = [480] dtype = "float32" min_val = float("0.329786") @@ -4814,6 +5252,7 @@ class Program_weight_tensor_parameter_437: class Program_weight_tensor_parameter_438: name = "parameter_438" + original_name = "batch_norm_13.w_2" shape = [480] dtype = "float32" min_val = float("2.66405e-05") @@ -4825,6 +5264,7 @@ class Program_weight_tensor_parameter_438: class Program_weight_tensor_parameter_439: name = "parameter_439" + original_name = "batch_norm_13.w_1" shape = [480] dtype = "float32" min_val = float("-0.476364") @@ -4836,6 +5276,7 @@ class Program_weight_tensor_parameter_439: class Program_weight_tensor_parameter_440: name = "parameter_440" + original_name = "conv2d_13.w_0" shape = [480, 1, 5, 5] dtype = "float32" min_val = float("-0.0752012") @@ -4847,6 +5288,7 @@ class Program_weight_tensor_parameter_440: class Program_weight_tensor_parameter_441: name = "parameter_441" + original_name = "batch_norm_12.b_0" shape = [480] dtype = "float32" min_val = float("-0.979684") @@ -4858,6 +5300,7 @@ class Program_weight_tensor_parameter_441: class Program_weight_tensor_parameter_442: name = "parameter_442" + original_name = "batch_norm_12.w_0" shape = [480] dtype = "float32" min_val = float("0.238952") @@ -4869,6 +5312,7 @@ class Program_weight_tensor_parameter_442: class Program_weight_tensor_parameter_443: name = "parameter_443" + original_name = "batch_norm_12.w_2" shape = [480] dtype = "float32" min_val = float("0.030082") @@ -4880,6 +5324,7 @@ class Program_weight_tensor_parameter_443: class Program_weight_tensor_parameter_444: name = "parameter_444" + original_name = "batch_norm_12.w_1" shape = [480] dtype = "float32" min_val = float("-0.366016") @@ -4891,6 +5336,7 @@ class Program_weight_tensor_parameter_444: class Program_weight_tensor_parameter_445: name = "parameter_445" + original_name = "conv2d_12.w_0" shape = [480, 80, 1, 1] dtype = "float32" min_val = float("-0.279988") @@ -4902,6 +5348,7 @@ class Program_weight_tensor_parameter_445: class Program_weight_tensor_parameter_446: name = "parameter_446" + original_name = "batch_norm_11.b_0" shape = [80] dtype = "float32" min_val = float("-1.41794") @@ -4913,6 +5360,7 @@ class Program_weight_tensor_parameter_446: class Program_weight_tensor_parameter_447: name = "parameter_447" + original_name = "batch_norm_11.w_0" shape = [80] dtype = "float32" min_val = float("0.499816") @@ -4924,6 +5372,7 @@ class Program_weight_tensor_parameter_447: class Program_weight_tensor_parameter_448: name = "parameter_448" + original_name = "batch_norm_11.w_2" shape = [80] dtype = "float32" min_val = float("0.000309496") @@ -4935,6 +5384,7 @@ class Program_weight_tensor_parameter_448: class Program_weight_tensor_parameter_449: name = "parameter_449" + original_name = "batch_norm_11.w_1" shape = [80] dtype = "float32" min_val = float("-0.178853") @@ -4946,6 +5396,7 @@ class Program_weight_tensor_parameter_449: class Program_weight_tensor_parameter_450: name = "parameter_450" + original_name = "conv2d_11.w_0" shape = [80, 1, 3, 3] dtype = "float32" min_val = float("-0.143886") @@ -4957,6 +5408,7 @@ class Program_weight_tensor_parameter_450: class Program_weight_tensor_parameter_451: name = "parameter_451" + original_name = "batch_norm_10.b_0" shape = [80] dtype = "float32" min_val = float("-1.50653") @@ -4968,6 +5420,7 @@ class Program_weight_tensor_parameter_451: class Program_weight_tensor_parameter_452: name = "parameter_452" + original_name = "batch_norm_10.w_0" shape = [80] dtype = "float32" min_val = float("-0.00309445") @@ -4979,6 +5432,7 @@ class Program_weight_tensor_parameter_452: class Program_weight_tensor_parameter_453: name = "parameter_453" + original_name = "batch_norm_10.w_2" shape = [80] dtype = "float32" min_val = float("0.0157422") @@ -4990,6 +5444,7 @@ class Program_weight_tensor_parameter_453: class Program_weight_tensor_parameter_454: name = "parameter_454" + original_name = "batch_norm_10.w_1" shape = [80] dtype = "float32" min_val = float("-1.03112") @@ -5001,6 +5456,7 @@ class Program_weight_tensor_parameter_454: class Program_weight_tensor_parameter_455: name = "parameter_455" + original_name = "conv2d_10.w_0" shape = [80, 160, 1, 1] dtype = "float32" min_val = float("-0.23413") @@ -5012,6 +5468,7 @@ class Program_weight_tensor_parameter_455: class Program_weight_tensor_parameter_456: name = "parameter_456" + original_name = "batch_norm_9.b_0" shape = [160] dtype = "float32" min_val = float("-3.0925") @@ -5023,6 +5480,7 @@ class Program_weight_tensor_parameter_456: class Program_weight_tensor_parameter_457: name = "parameter_457" + original_name = "batch_norm_9.w_0" shape = [160] dtype = "float32" min_val = float("0.329961") @@ -5034,6 +5492,7 @@ class Program_weight_tensor_parameter_457: class Program_weight_tensor_parameter_458: name = "parameter_458" + original_name = "batch_norm_9.w_2" shape = [160] dtype = "float32" min_val = float("1.03073e-05") @@ -5045,6 +5504,7 @@ class Program_weight_tensor_parameter_458: class Program_weight_tensor_parameter_459: name = "parameter_459" + original_name = "batch_norm_9.w_1" shape = [160] dtype = "float32" min_val = float("-0.184871") @@ -5056,6 +5516,7 @@ class Program_weight_tensor_parameter_459: class Program_weight_tensor_parameter_460: name = "parameter_460" + original_name = "conv2d_9.w_0" shape = [160, 1, 3, 3] dtype = "float32" min_val = float("-0.189446") @@ -5067,6 +5528,7 @@ class Program_weight_tensor_parameter_460: class Program_weight_tensor_parameter_461: name = "parameter_461" + original_name = "batch_norm_8.b_0" shape = [160] dtype = "float32" min_val = float("-0.970376") @@ -5078,6 +5540,7 @@ class Program_weight_tensor_parameter_461: class Program_weight_tensor_parameter_462: name = "parameter_462" + original_name = "batch_norm_8.w_0" shape = [160] dtype = "float32" min_val = float("0.213367") @@ -5089,6 +5552,7 @@ class Program_weight_tensor_parameter_462: class Program_weight_tensor_parameter_463: name = "parameter_463" + original_name = "batch_norm_8.w_2" shape = [160] dtype = "float32" min_val = float("0.0208283") @@ -5100,6 +5564,7 @@ class Program_weight_tensor_parameter_463: class Program_weight_tensor_parameter_464: name = "parameter_464" + original_name = "batch_norm_8.w_1" shape = [160] dtype = "float32" min_val = float("-0.350586") @@ -5111,6 +5576,7 @@ class Program_weight_tensor_parameter_464: class Program_weight_tensor_parameter_465: name = "parameter_465" + original_name = "conv2d_8.w_0" shape = [160, 80, 1, 1] dtype = "float32" min_val = float("-0.275416") @@ -5122,6 +5588,7 @@ class Program_weight_tensor_parameter_465: class Program_weight_tensor_parameter_466: name = "parameter_466" + original_name = "batch_norm_7.b_0" shape = [80] dtype = "float32" min_val = float("-1.38748") @@ -5133,6 +5600,7 @@ class Program_weight_tensor_parameter_466: class Program_weight_tensor_parameter_467: name = "parameter_467" + original_name = "batch_norm_7.w_0" shape = [80] dtype = "float32" min_val = float("-0.107812") @@ -5144,6 +5612,7 @@ class Program_weight_tensor_parameter_467: class Program_weight_tensor_parameter_468: name = "parameter_468" + original_name = "batch_norm_7.w_2" shape = [80] dtype = "float32" min_val = float("2.53158e-05") @@ -5155,6 +5624,7 @@ class Program_weight_tensor_parameter_468: class Program_weight_tensor_parameter_469: name = "parameter_469" + original_name = "batch_norm_7.w_1" shape = [80] dtype = "float32" min_val = float("-0.117618") @@ -5166,6 +5636,7 @@ class Program_weight_tensor_parameter_469: class Program_weight_tensor_parameter_470: name = "parameter_470" + original_name = "conv2d_7.w_0" shape = [80, 1, 3, 3] dtype = "float32" min_val = float("-0.257925") @@ -5177,6 +5648,7 @@ class Program_weight_tensor_parameter_470: class Program_weight_tensor_parameter_471: name = "parameter_471" + original_name = "batch_norm_6.b_0" shape = [80] dtype = "float32" min_val = float("-2.31921") @@ -5188,6 +5660,7 @@ class Program_weight_tensor_parameter_471: class Program_weight_tensor_parameter_472: name = "parameter_472" + original_name = "batch_norm_6.w_0" shape = [80] dtype = "float32" min_val = float("0.389264") @@ -5199,6 +5672,7 @@ class Program_weight_tensor_parameter_472: class Program_weight_tensor_parameter_473: name = "parameter_473" + original_name = "batch_norm_6.w_2" shape = [80] dtype = "float32" min_val = float("0.0218512") @@ -5210,6 +5684,7 @@ class Program_weight_tensor_parameter_473: class Program_weight_tensor_parameter_474: name = "parameter_474" + original_name = "batch_norm_6.w_1" shape = [80] dtype = "float32" min_val = float("-1.33321") @@ -5221,6 +5696,7 @@ class Program_weight_tensor_parameter_474: class Program_weight_tensor_parameter_475: name = "parameter_475" + original_name = "conv2d_6.w_0" shape = [80, 192, 1, 1] dtype = "float32" min_val = float("-0.205083") @@ -5232,6 +5708,7 @@ class Program_weight_tensor_parameter_475: class Program_weight_tensor_parameter_476: name = "parameter_476" + original_name = "batch_norm_5.b_0" shape = [192] dtype = "float32" min_val = float("-0.738548") @@ -5243,6 +5720,7 @@ class Program_weight_tensor_parameter_476: class Program_weight_tensor_parameter_477: name = "parameter_477" + original_name = "batch_norm_5.w_0" shape = [192] dtype = "float32" min_val = float("0.316142") @@ -5254,6 +5732,7 @@ class Program_weight_tensor_parameter_477: class Program_weight_tensor_parameter_478: name = "parameter_478" + original_name = "batch_norm_5.w_2" shape = [192] dtype = "float32" min_val = float("3.60325e-06") @@ -5265,6 +5744,7 @@ class Program_weight_tensor_parameter_478: class Program_weight_tensor_parameter_479: name = "parameter_479" + original_name = "batch_norm_5.w_1" shape = [192] dtype = "float32" min_val = float("-0.318309") @@ -5276,6 +5756,7 @@ class Program_weight_tensor_parameter_479: class Program_weight_tensor_parameter_480: name = "parameter_480" + original_name = "conv2d_5.w_0" shape = [192, 1, 5, 5] dtype = "float32" min_val = float("-0.091625") @@ -5287,6 +5768,7 @@ class Program_weight_tensor_parameter_480: class Program_weight_tensor_parameter_481: name = "parameter_481" + original_name = "batch_norm_4.b_0" shape = [192] dtype = "float32" min_val = float("-0.824922") @@ -5298,6 +5780,7 @@ class Program_weight_tensor_parameter_481: class Program_weight_tensor_parameter_482: name = "parameter_482" + original_name = "batch_norm_4.w_0" shape = [192] dtype = "float32" min_val = float("0.302154") @@ -5309,6 +5792,7 @@ class Program_weight_tensor_parameter_482: class Program_weight_tensor_parameter_483: name = "parameter_483" + original_name = "batch_norm_4.w_2" shape = [192] dtype = "float32" min_val = float("0.00394685") @@ -5320,6 +5804,7 @@ class Program_weight_tensor_parameter_483: class Program_weight_tensor_parameter_484: name = "parameter_484" + original_name = "batch_norm_4.w_1" shape = [192] dtype = "float32" min_val = float("-0.229786") @@ -5331,6 +5816,7 @@ class Program_weight_tensor_parameter_484: class Program_weight_tensor_parameter_485: name = "parameter_485" + original_name = "conv2d_4.w_0" shape = [192, 48, 1, 1] dtype = "float32" min_val = float("-0.306529") @@ -5342,6 +5828,7 @@ class Program_weight_tensor_parameter_485: class Program_weight_tensor_parameter_486: name = "parameter_486" + original_name = "batch_norm_3.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -5351,6 +5838,7 @@ class Program_weight_tensor_parameter_486: class Program_weight_tensor_parameter_487: name = "parameter_487" + original_name = "batch_norm_3.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -5360,6 +5848,7 @@ class Program_weight_tensor_parameter_487: class Program_weight_tensor_parameter_488: name = "parameter_488" + original_name = "batch_norm_3.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -5369,6 +5858,7 @@ class Program_weight_tensor_parameter_488: class Program_weight_tensor_parameter_489: name = "parameter_489" + original_name = "batch_norm_3.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -5378,6 +5868,7 @@ class Program_weight_tensor_parameter_489: class Program_weight_tensor_parameter_490: name = "parameter_490" + original_name = "conv2d_3.w_0" shape = [48, 1, 3, 3] dtype = "float32" min_val = float("-0.191253") @@ -5389,6 +5880,7 @@ class Program_weight_tensor_parameter_490: class Program_weight_tensor_parameter_491: name = "parameter_491" + original_name = "batch_norm_2.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -5398,6 +5890,7 @@ class Program_weight_tensor_parameter_491: class Program_weight_tensor_parameter_492: name = "parameter_492" + original_name = "batch_norm_2.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -5407,6 +5900,7 @@ class Program_weight_tensor_parameter_492: class Program_weight_tensor_parameter_493: name = "parameter_493" + original_name = "batch_norm_2.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -5416,6 +5910,7 @@ class Program_weight_tensor_parameter_493: class Program_weight_tensor_parameter_494: name = "parameter_494" + original_name = "batch_norm_2.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -5425,6 +5920,7 @@ class Program_weight_tensor_parameter_494: class Program_weight_tensor_parameter_495: name = "parameter_495" + original_name = "conv2d_2.w_0" shape = [48, 128, 1, 1] dtype = "float32" min_val = float("-0.200893") @@ -5436,6 +5932,7 @@ class Program_weight_tensor_parameter_495: class Program_weight_tensor_parameter_496: name = "parameter_496" + original_name = "batch_norm_1.b_0" shape = [128] dtype = "float32" min_val = float("-6.12301") @@ -5447,6 +5944,7 @@ class Program_weight_tensor_parameter_496: class Program_weight_tensor_parameter_497: name = "parameter_497" + original_name = "batch_norm_1.w_0" shape = [128] dtype = "float32" min_val = float("0.342235") @@ -5458,6 +5956,7 @@ class Program_weight_tensor_parameter_497: class Program_weight_tensor_parameter_498: name = "parameter_498" + original_name = "batch_norm_1.w_2" shape = [128] dtype = "float32" min_val = float("0.03344") @@ -5469,6 +5968,7 @@ class Program_weight_tensor_parameter_498: class Program_weight_tensor_parameter_499: name = "parameter_499" + original_name = "batch_norm_1.w_1" shape = [128] dtype = "float32" min_val = float("-3.84651") @@ -5480,6 +5980,7 @@ class Program_weight_tensor_parameter_499: class Program_weight_tensor_parameter_500: name = "parameter_500" + original_name = "conv2d_1.w_0" shape = [128, 32, 3, 3] dtype = "float32" min_val = float("-0.627198") @@ -5491,6 +5992,7 @@ class Program_weight_tensor_parameter_500: class Program_weight_tensor_parameter_501: name = "parameter_501" + original_name = "batch_norm_0.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -5500,6 +6002,7 @@ class Program_weight_tensor_parameter_501: class Program_weight_tensor_parameter_502: name = "parameter_502" + original_name = "batch_norm_0.w_0" shape = [32] dtype = "float32" min_val = float("0") @@ -5509,6 +6012,7 @@ class Program_weight_tensor_parameter_502: class Program_weight_tensor_parameter_503: name = "parameter_503" + original_name = "batch_norm_0.w_2" shape = [32] dtype = "float32" min_val = float("0") @@ -5518,6 +6022,7 @@ class Program_weight_tensor_parameter_503: class Program_weight_tensor_parameter_504: name = "parameter_504" + original_name = "batch_norm_0.w_1" shape = [32] dtype = "float32" min_val = float("0") @@ -5527,6 +6032,7 @@ class Program_weight_tensor_parameter_504: class Program_weight_tensor_parameter_505: name = "parameter_505" + original_name = "conv2d_0.w_0" shape = [32, 3, 3, 3] dtype = "float32" min_val = float("-0.0781542") diff --git a/paddle_samples/PaddleX/PP-HGNetV2-B0_ML/subgraph_1/input_meta.py b/paddle_samples/PaddleX/PP-HGNetV2-B0_ML/subgraph_1/input_meta.py index 1f0813456..72a9ed0b1 100644 --- a/paddle_samples/PaddleX/PP-HGNetV2-B0_ML/subgraph_1/input_meta.py +++ b/paddle_samples/PaddleX/PP-HGNetV2-B0_ML/subgraph_1/input_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_data_0: name = "data_0" + original_name = "param_100" shape = [1] dtype = "float32" data = [0.000197422] @@ -7,6 +8,7 @@ class Program_weight_tensor_data_0: class Program_weight_tensor_data_1: name = "data_1" + original_name = "param_101" shape = [1] dtype = "float32" data = [0.354466] @@ -14,6 +16,7 @@ class Program_weight_tensor_data_1: class Program_weight_tensor_data_2: name = "data_2" + original_name = "param_102" shape = [1] dtype = "float32" data = [-0.227031] @@ -21,6 +24,7 @@ class Program_weight_tensor_data_2: class Program_weight_tensor_data_3: name = "data_3" + original_name = "param_103" shape = [1] dtype = "float32" data = [0.279167] @@ -28,6 +32,7 @@ class Program_weight_tensor_data_3: class Program_weight_tensor_data_4: name = "data_4" + original_name = "param_104" shape = [1] dtype = "float32" data = [-0.000192061] @@ -35,6 +40,7 @@ class Program_weight_tensor_data_4: class Program_weight_tensor_data_5: name = "data_5" + original_name = "param_105" shape = [1] dtype = "float32" data = [0.264018] @@ -42,6 +48,7 @@ class Program_weight_tensor_data_5: class Program_weight_tensor_data_6: name = "data_6" + original_name = "param_106" shape = [1] dtype = "float32" data = [-4.8865e-05] @@ -49,6 +56,7 @@ class Program_weight_tensor_data_6: class Program_weight_tensor_data_7: name = "data_7" + original_name = "param_107" shape = [1] dtype = "float32" data = [0.316896] @@ -56,6 +64,7 @@ class Program_weight_tensor_data_7: class Program_weight_tensor_data_8: name = "data_8" + original_name = "param_108" shape = [1] dtype = "float32" data = [6.64035e-06] @@ -63,6 +72,7 @@ class Program_weight_tensor_data_8: class Program_weight_tensor_data_9: name = "data_9" + original_name = "param_109" shape = [1] dtype = "float32" data = [0.654354] @@ -70,6 +80,7 @@ class Program_weight_tensor_data_9: class Program_weight_tensor_data_10: name = "data_10" + original_name = "param_110" shape = [1] dtype = "float32" data = [0.000656549] @@ -77,6 +88,7 @@ class Program_weight_tensor_data_10: class Program_weight_tensor_data_11: name = "data_11" + original_name = "param_111" shape = [1] dtype = "float32" data = [1.68149] @@ -84,6 +96,7 @@ class Program_weight_tensor_data_11: class Program_weight_tensor_data_12: name = "data_12" + original_name = "param_112" shape = [1] dtype = "float32" data = [-0.22685] @@ -91,6 +104,7 @@ class Program_weight_tensor_data_12: class Program_weight_tensor_data_13: name = "data_13" + original_name = "param_113" shape = [1] dtype = "float32" data = [1.35568] @@ -98,6 +112,7 @@ class Program_weight_tensor_data_13: class Program_weight_tensor_data_14: name = "data_14" + original_name = "param_114" shape = [1] dtype = "float32" data = [-6.80522e-05] @@ -105,6 +120,7 @@ class Program_weight_tensor_data_14: class Program_weight_tensor_data_15: name = "data_15" + original_name = "param_115" shape = [1] dtype = "float32" data = [1.45413] @@ -112,6 +128,7 @@ class Program_weight_tensor_data_15: class Program_weight_tensor_data_16: name = "data_16" + original_name = "param_116" shape = [1] dtype = "float32" data = [0.00011129] @@ -119,6 +136,7 @@ class Program_weight_tensor_data_16: class Program_weight_tensor_data_17: name = "data_17" + original_name = "param_117" shape = [1] dtype = "float32" data = [1.94575] @@ -126,6 +144,7 @@ class Program_weight_tensor_data_17: class Program_weight_tensor_data_18: name = "data_18" + original_name = "param_118" shape = [1] dtype = "float32" data = [1.42343e-05] @@ -133,6 +152,7 @@ class Program_weight_tensor_data_18: class Program_weight_tensor_data_19: name = "data_19" + original_name = "param_119" shape = [1] dtype = "float32" data = [0.651347] @@ -140,6 +160,7 @@ class Program_weight_tensor_data_19: class Program_weight_tensor_data_20: name = "data_20" + original_name = "param_120" shape = [1] dtype = "float32" data = [0.0108459] @@ -147,6 +168,7 @@ class Program_weight_tensor_data_20: class Program_weight_tensor_data_21: name = "data_21" + original_name = "param_121" shape = [1] dtype = "float32" data = [3.90235] @@ -154,6 +176,7 @@ class Program_weight_tensor_data_21: class Program_weight_tensor_data_22: name = "data_22" + original_name = "param_122" shape = [1] dtype = "float32" data = [-0.406396] @@ -161,6 +184,7 @@ class Program_weight_tensor_data_22: class Program_weight_tensor_data_23: name = "data_23" + original_name = "param_123" shape = [1] dtype = "float32" data = [2.09851] @@ -168,6 +192,7 @@ class Program_weight_tensor_data_23: class Program_weight_tensor_data_24: name = "data_24" + original_name = "param_124" shape = [1] dtype = "float32" data = [-0.44432] @@ -175,6 +200,7 @@ class Program_weight_tensor_data_24: class Program_weight_tensor_data_25: name = "data_25" + original_name = "param_125" shape = [33, 768] dtype = "float32" min_val = float("-0.0865443") @@ -186,6 +212,7 @@ class Program_weight_tensor_data_25: class Program_weight_tensor_data_26: name = "data_26" + original_name = "param_63" shape = [1] dtype = "float32" data = [0.165453] @@ -193,6 +220,7 @@ class Program_weight_tensor_data_26: class Program_weight_tensor_data_27: name = "data_27" + original_name = "param_64" shape = [1] dtype = "float32" data = [-1.4634] @@ -200,6 +228,7 @@ class Program_weight_tensor_data_27: class Program_weight_tensor_data_28: name = "data_28" + original_name = "param_65" shape = [1] dtype = "float32" data = [0.168553] @@ -207,6 +236,7 @@ class Program_weight_tensor_data_28: class Program_weight_tensor_data_29: name = "data_29" + original_name = "param_66" shape = [1] dtype = "float32" data = [-0.637126] @@ -214,6 +244,7 @@ class Program_weight_tensor_data_29: class Program_weight_tensor_data_30: name = "data_30" + original_name = "param_67" shape = [1] dtype = "float32" data = [0.123213] @@ -221,6 +252,7 @@ class Program_weight_tensor_data_30: class Program_weight_tensor_data_31: name = "data_31" + original_name = "param_68" shape = [1] dtype = "float32" data = [-0.0597448] @@ -228,6 +260,7 @@ class Program_weight_tensor_data_31: class Program_weight_tensor_data_32: name = "data_32" + original_name = "param_69" shape = [1] dtype = "float32" data = [0.703493] @@ -235,6 +268,7 @@ class Program_weight_tensor_data_32: class Program_weight_tensor_data_33: name = "data_33" + original_name = "param_70" shape = [1] dtype = "float32" data = [0.000180886] @@ -242,6 +276,7 @@ class Program_weight_tensor_data_33: class Program_weight_tensor_data_34: name = "data_34" + original_name = "param_71" shape = [1] dtype = "float32" data = [0.304602] @@ -249,6 +284,7 @@ class Program_weight_tensor_data_34: class Program_weight_tensor_data_35: name = "data_35" + original_name = "param_72" shape = [1] dtype = "float32" data = [-0.970265] @@ -256,6 +292,7 @@ class Program_weight_tensor_data_35: class Program_weight_tensor_data_36: name = "data_36" + original_name = "param_73" shape = [1] dtype = "float32" data = [0.347885] @@ -263,6 +300,7 @@ class Program_weight_tensor_data_36: class Program_weight_tensor_data_37: name = "data_37" + original_name = "param_74" shape = [1] dtype = "float32" data = [-1.32624] @@ -270,6 +308,7 @@ class Program_weight_tensor_data_37: class Program_weight_tensor_data_38: name = "data_38" + original_name = "param_75" shape = [1] dtype = "float32" data = [0.421036] @@ -277,6 +316,7 @@ class Program_weight_tensor_data_38: class Program_weight_tensor_data_39: name = "data_39" + original_name = "param_76" shape = [1] dtype = "float32" data = [-0.383203] @@ -284,6 +324,7 @@ class Program_weight_tensor_data_39: class Program_weight_tensor_data_40: name = "data_40" + original_name = "param_77" shape = [1] dtype = "float32" data = [0.372754] @@ -291,6 +332,7 @@ class Program_weight_tensor_data_40: class Program_weight_tensor_data_41: name = "data_41" + original_name = "param_78" shape = [1] dtype = "float32" data = [4.73484e-05] @@ -298,6 +340,7 @@ class Program_weight_tensor_data_41: class Program_weight_tensor_data_42: name = "data_42" + original_name = "param_79" shape = [1] dtype = "float32" data = [0.662699] @@ -305,6 +348,7 @@ class Program_weight_tensor_data_42: class Program_weight_tensor_data_43: name = "data_43" + original_name = "param_80" shape = [1] dtype = "float32" data = [3.04501e-05] @@ -312,6 +356,7 @@ class Program_weight_tensor_data_43: class Program_weight_tensor_data_44: name = "data_44" + original_name = "param_81" shape = [1] dtype = "float32" data = [0.46966] @@ -319,6 +364,7 @@ class Program_weight_tensor_data_44: class Program_weight_tensor_data_45: name = "data_45" + original_name = "param_82" shape = [1] dtype = "float32" data = [-0.524609] @@ -326,6 +372,7 @@ class Program_weight_tensor_data_45: class Program_weight_tensor_data_46: name = "data_46" + original_name = "param_83" shape = [1] dtype = "float32" data = [1.22375] @@ -333,6 +380,7 @@ class Program_weight_tensor_data_46: class Program_weight_tensor_data_47: name = "data_47" + original_name = "param_84" shape = [1] dtype = "float32" data = [-0.937789] @@ -340,6 +388,7 @@ class Program_weight_tensor_data_47: class Program_weight_tensor_data_48: name = "data_48" + original_name = "param_85" shape = [1] dtype = "float32" data = [1.27285] @@ -347,6 +396,7 @@ class Program_weight_tensor_data_48: class Program_weight_tensor_data_49: name = "data_49" + original_name = "param_86" shape = [1] dtype = "float32" data = [-0.33691] @@ -354,6 +404,7 @@ class Program_weight_tensor_data_49: class Program_weight_tensor_data_50: name = "data_50" + original_name = "param_87" shape = [1] dtype = "float32" data = [1.41884] @@ -361,6 +412,7 @@ class Program_weight_tensor_data_50: class Program_weight_tensor_data_51: name = "data_51" + original_name = "param_88" shape = [1] dtype = "float32" data = [0.0001068] @@ -368,6 +420,7 @@ class Program_weight_tensor_data_51: class Program_weight_tensor_data_52: name = "data_52" + original_name = "param_89" shape = [1] dtype = "float32" data = [0.656008] @@ -375,6 +428,7 @@ class Program_weight_tensor_data_52: class Program_weight_tensor_data_53: name = "data_53" + original_name = "param_90" shape = [1] dtype = "float32" data = [-3.33434e-05] @@ -382,6 +436,7 @@ class Program_weight_tensor_data_53: class Program_weight_tensor_data_54: name = "data_54" + original_name = "param_91" shape = [1] dtype = "float32" data = [0.666857] @@ -389,6 +444,7 @@ class Program_weight_tensor_data_54: class Program_weight_tensor_data_55: name = "data_55" + original_name = "param_92" shape = [1] dtype = "float32" data = [-0.227735] @@ -396,6 +452,7 @@ class Program_weight_tensor_data_55: class Program_weight_tensor_data_56: name = "data_56" + original_name = "param_93" shape = [1] dtype = "float32" data = [1.34855] @@ -403,6 +460,7 @@ class Program_weight_tensor_data_56: class Program_weight_tensor_data_57: name = "data_57" + original_name = "param_94" shape = [1] dtype = "float32" data = [-9.74106e-05] @@ -410,6 +468,7 @@ class Program_weight_tensor_data_57: class Program_weight_tensor_data_58: name = "data_58" + original_name = "param_95" shape = [1] dtype = "float32" data = [1.26776] @@ -417,6 +476,7 @@ class Program_weight_tensor_data_58: class Program_weight_tensor_data_59: name = "data_59" + original_name = "param_96" shape = [1] dtype = "float32" data = [2.57887e-06] @@ -424,6 +484,7 @@ class Program_weight_tensor_data_59: class Program_weight_tensor_data_60: name = "data_60" + original_name = "param_97" shape = [1] dtype = "float32" data = [1.6243] @@ -431,6 +492,7 @@ class Program_weight_tensor_data_60: class Program_weight_tensor_data_61: name = "data_61" + original_name = "param_98" shape = [1] dtype = "float32" data = [7.92396e-06] @@ -438,6 +500,7 @@ class Program_weight_tensor_data_61: class Program_weight_tensor_data_62: name = "data_62" + original_name = "param_99" shape = [1] dtype = "float32" data = [0.654671] @@ -445,6 +508,7 @@ class Program_weight_tensor_data_62: class Program_weight_tensor_data_63: name = "data_63" + original_name = "var_202" shape = [55, 3, 448, 448] dtype = "float32" min_val = float("-2.1179") diff --git a/paddle_samples/PaddleX/PP-HGNetV2-B0_ML/subgraph_1/weight_meta.py b/paddle_samples/PaddleX/PP-HGNetV2-B0_ML/subgraph_1/weight_meta.py index 10ead40f4..1e1726f17 100644 --- a/paddle_samples/PaddleX/PP-HGNetV2-B0_ML/subgraph_1/weight_meta.py +++ b/paddle_samples/PaddleX/PP-HGNetV2-B0_ML/subgraph_1/weight_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_parameter_0: name = "parameter_0" + original_name = "conv2d_44.b_0" shape = [33] dtype = "float32" min_val = float("0") @@ -9,6 +10,7 @@ class Program_weight_tensor_parameter_0: class Program_weight_tensor_parameter_1: name = "parameter_1" + original_name = "conv2d_44.w_0" shape = [33, 768, 1, 1] dtype = "float32" min_val = float("-0.187735") @@ -20,6 +22,7 @@ class Program_weight_tensor_parameter_1: class Program_weight_tensor_parameter_2: name = "parameter_2" + original_name = "layer_norm_2.b_0" shape = [768] dtype = "float32" min_val = float("-0.00838926") @@ -31,6 +34,7 @@ class Program_weight_tensor_parameter_2: class Program_weight_tensor_parameter_3: name = "parameter_3" + original_name = "layer_norm_2.w_0" shape = [768] dtype = "float32" min_val = float("0.93939") @@ -42,6 +46,7 @@ class Program_weight_tensor_parameter_3: class Program_weight_tensor_parameter_4: name = "parameter_4" + original_name = "linear_10.b_0" shape = [768] dtype = "float32" min_val = float("-0.0280762") @@ -53,6 +58,7 @@ class Program_weight_tensor_parameter_4: class Program_weight_tensor_parameter_5: name = "parameter_5" + original_name = "linear_10.w_0" shape = [2048, 768] dtype = "float32" min_val = float("-0.109483") @@ -64,6 +70,7 @@ class Program_weight_tensor_parameter_5: class Program_weight_tensor_parameter_6: name = "parameter_6" + original_name = "linear_9.b_0" shape = [2048] dtype = "float32" min_val = float("-0.0123772") @@ -75,6 +82,7 @@ class Program_weight_tensor_parameter_6: class Program_weight_tensor_parameter_7: name = "parameter_7" + original_name = "linear_9.w_0" shape = [768, 2048] dtype = "float32" min_val = float("-0.0789205") @@ -86,6 +94,7 @@ class Program_weight_tensor_parameter_7: class Program_weight_tensor_parameter_8: name = "parameter_8" + original_name = "layer_norm_1.b_0" shape = [768] dtype = "float32" min_val = float("-0.0832909") @@ -97,6 +106,7 @@ class Program_weight_tensor_parameter_8: class Program_weight_tensor_parameter_9: name = "parameter_9" + original_name = "layer_norm_1.w_0" shape = [768] dtype = "float32" min_val = float("0.979165") @@ -108,6 +118,7 @@ class Program_weight_tensor_parameter_9: class Program_weight_tensor_parameter_10: name = "parameter_10" + original_name = "linear_8.b_0" shape = [768] dtype = "float32" min_val = float("-0.0499441") @@ -119,6 +130,7 @@ class Program_weight_tensor_parameter_10: class Program_weight_tensor_parameter_11: name = "parameter_11" + original_name = "linear_8.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.125395") @@ -130,6 +142,7 @@ class Program_weight_tensor_parameter_11: class Program_weight_tensor_parameter_12: name = "parameter_12" + original_name = "linear_7.b_0" shape = [768] dtype = "float32" min_val = float("-0.010729") @@ -141,6 +154,7 @@ class Program_weight_tensor_parameter_12: class Program_weight_tensor_parameter_13: name = "parameter_13" + original_name = "linear_7.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.0851121") @@ -152,6 +166,7 @@ class Program_weight_tensor_parameter_13: class Program_weight_tensor_parameter_14: name = "parameter_14" + original_name = "linear_6.b_0" shape = [768] dtype = "float32" min_val = float("-0.0137759") @@ -163,6 +178,7 @@ class Program_weight_tensor_parameter_14: class Program_weight_tensor_parameter_15: name = "parameter_15" + original_name = "linear_6.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.0823939") @@ -174,6 +190,7 @@ class Program_weight_tensor_parameter_15: class Program_weight_tensor_parameter_16: name = "parameter_16" + original_name = "linear_5.b_0" shape = [768] dtype = "float32" min_val = float("-0.0320041") @@ -185,6 +202,7 @@ class Program_weight_tensor_parameter_16: class Program_weight_tensor_parameter_17: name = "parameter_17" + original_name = "linear_5.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.0779226") @@ -196,6 +214,7 @@ class Program_weight_tensor_parameter_17: class Program_weight_tensor_parameter_18: name = "parameter_18" + original_name = "layer_norm_0.b_0" shape = [768] dtype = "float32" min_val = float("-0.0561806") @@ -207,6 +226,7 @@ class Program_weight_tensor_parameter_18: class Program_weight_tensor_parameter_19: name = "parameter_19" + original_name = "layer_norm_0.w_0" shape = [768] dtype = "float32" min_val = float("0.971001") @@ -218,6 +238,7 @@ class Program_weight_tensor_parameter_19: class Program_weight_tensor_parameter_20: name = "parameter_20" + original_name = "conv2d_43.b_0" shape = [768] dtype = "float32" min_val = float("-0.00583616") @@ -229,6 +250,7 @@ class Program_weight_tensor_parameter_20: class Program_weight_tensor_parameter_21: name = "parameter_21" + original_name = "conv2d_43.w_0" shape = [768, 2048, 1, 1] dtype = "float32" min_val = float("-0.149168") @@ -240,6 +262,7 @@ class Program_weight_tensor_parameter_21: class Program_weight_tensor_parameter_22: name = "parameter_22" + original_name = "conv2d_42.w_0" shape = [2048, 1024, 1, 1] dtype = "float32" min_val = float("-1.13303") @@ -251,6 +274,7 @@ class Program_weight_tensor_parameter_22: class Program_weight_tensor_parameter_23: name = "parameter_23" + original_name = "batch_norm2d_41.b_0" shape = [1024] dtype = "float32" min_val = float("-13.1868") @@ -262,6 +286,7 @@ class Program_weight_tensor_parameter_23: class Program_weight_tensor_parameter_24: name = "parameter_24" + original_name = "batch_norm2d_41.w_0" shape = [1024] dtype = "float32" min_val = float("0.36901") @@ -273,6 +298,7 @@ class Program_weight_tensor_parameter_24: class Program_weight_tensor_parameter_25: name = "parameter_25" + original_name = "batch_norm2d_41.w_2" shape = [1024] dtype = "float32" min_val = float("0.458813") @@ -284,6 +310,7 @@ class Program_weight_tensor_parameter_25: class Program_weight_tensor_parameter_26: name = "parameter_26" + original_name = "batch_norm2d_41.w_1" shape = [1024] dtype = "float32" min_val = float("-2.54986") @@ -295,6 +322,7 @@ class Program_weight_tensor_parameter_26: class Program_weight_tensor_parameter_27: name = "parameter_27" + original_name = "conv2d_41.w_0" shape = [1024, 512, 1, 1] dtype = "float32" min_val = float("-1.16548") @@ -306,6 +334,7 @@ class Program_weight_tensor_parameter_27: class Program_weight_tensor_parameter_28: name = "parameter_28" + original_name = "batch_norm2d_40.b_0" shape = [512] dtype = "float32" min_val = float("-10.9274") @@ -317,6 +346,7 @@ class Program_weight_tensor_parameter_28: class Program_weight_tensor_parameter_29: name = "parameter_29" + original_name = "batch_norm2d_40.w_0" shape = [512] dtype = "float32" min_val = float("1.08943") @@ -328,6 +358,7 @@ class Program_weight_tensor_parameter_29: class Program_weight_tensor_parameter_30: name = "parameter_30" + original_name = "batch_norm2d_40.w_2" shape = [512] dtype = "float32" min_val = float("4.87137") @@ -339,6 +370,7 @@ class Program_weight_tensor_parameter_30: class Program_weight_tensor_parameter_31: name = "parameter_31" + original_name = "batch_norm2d_40.w_1" shape = [512] dtype = "float32" min_val = float("-11.4022") @@ -350,6 +382,7 @@ class Program_weight_tensor_parameter_31: class Program_weight_tensor_parameter_32: name = "parameter_32" + original_name = "conv2d_40.w_0" shape = [512, 896, 1, 1] dtype = "float32" min_val = float("-0.913472") @@ -361,6 +394,7 @@ class Program_weight_tensor_parameter_32: class Program_weight_tensor_parameter_33: name = "parameter_33" + original_name = "batch_norm2d_39.b_0" shape = [128] dtype = "float32" min_val = float("-7.49825") @@ -372,6 +406,7 @@ class Program_weight_tensor_parameter_33: class Program_weight_tensor_parameter_34: name = "parameter_34" + original_name = "batch_norm2d_39.w_0" shape = [128] dtype = "float32" min_val = float("0.861344") @@ -383,6 +418,7 @@ class Program_weight_tensor_parameter_34: class Program_weight_tensor_parameter_35: name = "parameter_35" + original_name = "batch_norm2d_39.w_2" shape = [128] dtype = "float32" min_val = float("0.241215") @@ -394,6 +430,7 @@ class Program_weight_tensor_parameter_35: class Program_weight_tensor_parameter_36: name = "parameter_36" + original_name = "batch_norm2d_39.w_1" shape = [128] dtype = "float32" min_val = float("-4.52067") @@ -405,6 +442,7 @@ class Program_weight_tensor_parameter_36: class Program_weight_tensor_parameter_37: name = "parameter_37" + original_name = "conv2d_39.w_0" shape = [128, 1, 5, 5] dtype = "float32" min_val = float("-0.611937") @@ -416,6 +454,7 @@ class Program_weight_tensor_parameter_37: class Program_weight_tensor_parameter_38: name = "parameter_38" + original_name = "batch_norm2d_38.b_0" shape = [128] dtype = "float32" min_val = float("-2.95745") @@ -427,6 +466,7 @@ class Program_weight_tensor_parameter_38: class Program_weight_tensor_parameter_39: name = "parameter_39" + original_name = "batch_norm2d_38.w_0" shape = [128] dtype = "float32" min_val = float("0.715911") @@ -438,6 +478,7 @@ class Program_weight_tensor_parameter_39: class Program_weight_tensor_parameter_40: name = "parameter_40" + original_name = "batch_norm2d_38.w_2" shape = [128] dtype = "float32" min_val = float("0.588251") @@ -449,6 +490,7 @@ class Program_weight_tensor_parameter_40: class Program_weight_tensor_parameter_41: name = "parameter_41" + original_name = "batch_norm2d_38.w_1" shape = [128] dtype = "float32" min_val = float("-7.72584") @@ -460,6 +502,7 @@ class Program_weight_tensor_parameter_41: class Program_weight_tensor_parameter_42: name = "parameter_42" + original_name = "conv2d_38.w_0" shape = [128, 128, 1, 1] dtype = "float32" min_val = float("-0.701822") @@ -471,6 +514,7 @@ class Program_weight_tensor_parameter_42: class Program_weight_tensor_parameter_43: name = "parameter_43" + original_name = "batch_norm2d_37.b_0" shape = [128] dtype = "float32" min_val = float("-6.70145") @@ -482,6 +526,7 @@ class Program_weight_tensor_parameter_43: class Program_weight_tensor_parameter_44: name = "parameter_44" + original_name = "batch_norm2d_37.w_0" shape = [128] dtype = "float32" min_val = float("1.03681") @@ -493,6 +538,7 @@ class Program_weight_tensor_parameter_44: class Program_weight_tensor_parameter_45: name = "parameter_45" + original_name = "batch_norm2d_37.w_2" shape = [128] dtype = "float32" min_val = float("0.113906") @@ -504,6 +550,7 @@ class Program_weight_tensor_parameter_45: class Program_weight_tensor_parameter_46: name = "parameter_46" + original_name = "batch_norm2d_37.w_1" shape = [128] dtype = "float32" min_val = float("-2.7207") @@ -515,6 +562,7 @@ class Program_weight_tensor_parameter_46: class Program_weight_tensor_parameter_47: name = "parameter_47" + original_name = "conv2d_37.w_0" shape = [128, 1, 5, 5] dtype = "float32" min_val = float("-0.85633") @@ -526,6 +574,7 @@ class Program_weight_tensor_parameter_47: class Program_weight_tensor_parameter_48: name = "parameter_48" + original_name = "batch_norm2d_36.b_0" shape = [128] dtype = "float32" min_val = float("-1.90118") @@ -537,6 +586,7 @@ class Program_weight_tensor_parameter_48: class Program_weight_tensor_parameter_49: name = "parameter_49" + original_name = "batch_norm2d_36.w_0" shape = [128] dtype = "float32" min_val = float("0.469957") @@ -548,6 +598,7 @@ class Program_weight_tensor_parameter_49: class Program_weight_tensor_parameter_50: name = "parameter_50" + original_name = "batch_norm2d_36.w_2" shape = [128] dtype = "float32" min_val = float("1.25307") @@ -559,6 +610,7 @@ class Program_weight_tensor_parameter_50: class Program_weight_tensor_parameter_51: name = "parameter_51" + original_name = "batch_norm2d_36.w_1" shape = [128] dtype = "float32" min_val = float("-8.48682") @@ -570,6 +622,7 @@ class Program_weight_tensor_parameter_51: class Program_weight_tensor_parameter_52: name = "parameter_52" + original_name = "conv2d_36.w_0" shape = [128, 128, 1, 1] dtype = "float32" min_val = float("-0.908453") @@ -581,6 +634,7 @@ class Program_weight_tensor_parameter_52: class Program_weight_tensor_parameter_53: name = "parameter_53" + original_name = "batch_norm2d_35.b_0" shape = [128] dtype = "float32" min_val = float("-8.19504") @@ -592,6 +646,7 @@ class Program_weight_tensor_parameter_53: class Program_weight_tensor_parameter_54: name = "parameter_54" + original_name = "batch_norm2d_35.w_0" shape = [128] dtype = "float32" min_val = float("1.19499") @@ -603,6 +658,7 @@ class Program_weight_tensor_parameter_54: class Program_weight_tensor_parameter_55: name = "parameter_55" + original_name = "batch_norm2d_35.w_2" shape = [128] dtype = "float32" min_val = float("0.308033") @@ -614,6 +670,7 @@ class Program_weight_tensor_parameter_55: class Program_weight_tensor_parameter_56: name = "parameter_56" + original_name = "batch_norm2d_35.w_1" shape = [128] dtype = "float32" min_val = float("-1.90061") @@ -625,6 +682,7 @@ class Program_weight_tensor_parameter_56: class Program_weight_tensor_parameter_57: name = "parameter_57" + original_name = "conv2d_35.w_0" shape = [128, 1, 5, 5] dtype = "float32" min_val = float("-0.718004") @@ -636,6 +694,7 @@ class Program_weight_tensor_parameter_57: class Program_weight_tensor_parameter_58: name = "parameter_58" + original_name = "batch_norm2d_34.b_0" shape = [128] dtype = "float32" min_val = float("-2.20069") @@ -647,6 +706,7 @@ class Program_weight_tensor_parameter_58: class Program_weight_tensor_parameter_59: name = "parameter_59" + original_name = "batch_norm2d_34.w_0" shape = [128] dtype = "float32" min_val = float("0.972093") @@ -658,6 +718,7 @@ class Program_weight_tensor_parameter_59: class Program_weight_tensor_parameter_60: name = "parameter_60" + original_name = "batch_norm2d_34.w_2" shape = [128] dtype = "float32" min_val = float("4.7045") @@ -669,6 +730,7 @@ class Program_weight_tensor_parameter_60: class Program_weight_tensor_parameter_61: name = "parameter_61" + original_name = "batch_norm2d_34.w_1" shape = [128] dtype = "float32" min_val = float("-0.0209562") @@ -680,6 +742,7 @@ class Program_weight_tensor_parameter_61: class Program_weight_tensor_parameter_62: name = "parameter_62" + original_name = "conv2d_34.w_0" shape = [128, 512, 1, 1] dtype = "float32" min_val = float("-0.636454") @@ -691,6 +754,7 @@ class Program_weight_tensor_parameter_62: class Program_weight_tensor_parameter_63: name = "parameter_63" + original_name = "batch_norm2d_33.b_0" shape = [512] dtype = "float32" min_val = float("-0.0115022") @@ -702,6 +766,7 @@ class Program_weight_tensor_parameter_63: class Program_weight_tensor_parameter_64: name = "parameter_64" + original_name = "batch_norm2d_33.w_0" shape = [512] dtype = "float32" min_val = float("1.34916") @@ -713,6 +778,7 @@ class Program_weight_tensor_parameter_64: class Program_weight_tensor_parameter_65: name = "parameter_65" + original_name = "batch_norm2d_33.w_2" shape = [512] dtype = "float32" min_val = float("0.0054595") @@ -724,6 +790,7 @@ class Program_weight_tensor_parameter_65: class Program_weight_tensor_parameter_66: name = "parameter_66" + original_name = "batch_norm2d_33.w_1" shape = [512] dtype = "float32" min_val = float("-4.00115") @@ -735,6 +802,7 @@ class Program_weight_tensor_parameter_66: class Program_weight_tensor_parameter_67: name = "parameter_67" + original_name = "conv2d_33.w_0" shape = [512, 1, 3, 3] dtype = "float32" min_val = float("-0.357989") @@ -746,6 +814,7 @@ class Program_weight_tensor_parameter_67: class Program_weight_tensor_parameter_68: name = "parameter_68" + original_name = "batch_norm2d_32.b_0" shape = [512] dtype = "float32" min_val = float("-7.77926") @@ -757,6 +826,7 @@ class Program_weight_tensor_parameter_68: class Program_weight_tensor_parameter_69: name = "parameter_69" + original_name = "batch_norm2d_32.w_0" shape = [512] dtype = "float32" min_val = float("0.166236") @@ -768,6 +838,7 @@ class Program_weight_tensor_parameter_69: class Program_weight_tensor_parameter_70: name = "parameter_70" + original_name = "batch_norm2d_32.w_2" shape = [512] dtype = "float32" min_val = float("0.200817") @@ -779,6 +850,7 @@ class Program_weight_tensor_parameter_70: class Program_weight_tensor_parameter_71: name = "parameter_71" + original_name = "batch_norm2d_32.w_1" shape = [512] dtype = "float32" min_val = float("-2.32542") @@ -790,6 +862,7 @@ class Program_weight_tensor_parameter_71: class Program_weight_tensor_parameter_72: name = "parameter_72" + original_name = "conv2d_32.w_0" shape = [512, 256, 1, 1] dtype = "float32" min_val = float("-0.597668") @@ -801,6 +874,7 @@ class Program_weight_tensor_parameter_72: class Program_weight_tensor_parameter_73: name = "parameter_73" + original_name = "batch_norm2d_31.b_0" shape = [256] dtype = "float32" min_val = float("-9.64372") @@ -812,6 +886,7 @@ class Program_weight_tensor_parameter_73: class Program_weight_tensor_parameter_74: name = "parameter_74" + original_name = "batch_norm2d_31.w_0" shape = [256] dtype = "float32" min_val = float("1.22708") @@ -823,6 +898,7 @@ class Program_weight_tensor_parameter_74: class Program_weight_tensor_parameter_75: name = "parameter_75" + original_name = "batch_norm2d_31.w_2" shape = [256] dtype = "float32" min_val = float("0.158594") @@ -834,6 +910,7 @@ class Program_weight_tensor_parameter_75: class Program_weight_tensor_parameter_76: name = "parameter_76" + original_name = "batch_norm2d_31.w_1" shape = [256] dtype = "float32" min_val = float("-1.9897") @@ -845,6 +922,7 @@ class Program_weight_tensor_parameter_76: class Program_weight_tensor_parameter_77: name = "parameter_77" + original_name = "conv2d_31.w_0" shape = [256, 704, 1, 1] dtype = "float32" min_val = float("-0.850675") @@ -856,6 +934,7 @@ class Program_weight_tensor_parameter_77: class Program_weight_tensor_parameter_78: name = "parameter_78" + original_name = "batch_norm2d_30.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -865,6 +944,7 @@ class Program_weight_tensor_parameter_78: class Program_weight_tensor_parameter_79: name = "parameter_79" + original_name = "batch_norm2d_30.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -874,6 +954,7 @@ class Program_weight_tensor_parameter_79: class Program_weight_tensor_parameter_80: name = "parameter_80" + original_name = "batch_norm2d_30.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -883,6 +964,7 @@ class Program_weight_tensor_parameter_80: class Program_weight_tensor_parameter_81: name = "parameter_81" + original_name = "batch_norm2d_30.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -892,6 +974,7 @@ class Program_weight_tensor_parameter_81: class Program_weight_tensor_parameter_82: name = "parameter_82" + original_name = "conv2d_30.w_0" shape = [64, 1, 5, 5] dtype = "float32" min_val = float("-0.645889") @@ -903,6 +986,7 @@ class Program_weight_tensor_parameter_82: class Program_weight_tensor_parameter_83: name = "parameter_83" + original_name = "batch_norm2d_29.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -912,6 +996,7 @@ class Program_weight_tensor_parameter_83: class Program_weight_tensor_parameter_84: name = "parameter_84" + original_name = "batch_norm2d_29.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -921,6 +1006,7 @@ class Program_weight_tensor_parameter_84: class Program_weight_tensor_parameter_85: name = "parameter_85" + original_name = "batch_norm2d_29.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -930,6 +1016,7 @@ class Program_weight_tensor_parameter_85: class Program_weight_tensor_parameter_86: name = "parameter_86" + original_name = "batch_norm2d_29.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -939,6 +1026,7 @@ class Program_weight_tensor_parameter_86: class Program_weight_tensor_parameter_87: name = "parameter_87" + original_name = "conv2d_29.w_0" shape = [64, 64, 1, 1] dtype = "float32" min_val = float("-0.581291") @@ -950,6 +1038,7 @@ class Program_weight_tensor_parameter_87: class Program_weight_tensor_parameter_88: name = "parameter_88" + original_name = "batch_norm2d_28.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -959,6 +1048,7 @@ class Program_weight_tensor_parameter_88: class Program_weight_tensor_parameter_89: name = "parameter_89" + original_name = "batch_norm2d_28.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -968,6 +1058,7 @@ class Program_weight_tensor_parameter_89: class Program_weight_tensor_parameter_90: name = "parameter_90" + original_name = "batch_norm2d_28.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -977,6 +1068,7 @@ class Program_weight_tensor_parameter_90: class Program_weight_tensor_parameter_91: name = "parameter_91" + original_name = "batch_norm2d_28.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -986,6 +1078,7 @@ class Program_weight_tensor_parameter_91: class Program_weight_tensor_parameter_92: name = "parameter_92" + original_name = "conv2d_28.w_0" shape = [64, 1, 5, 5] dtype = "float32" min_val = float("-0.675192") @@ -997,6 +1090,7 @@ class Program_weight_tensor_parameter_92: class Program_weight_tensor_parameter_93: name = "parameter_93" + original_name = "batch_norm2d_27.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -1006,6 +1100,7 @@ class Program_weight_tensor_parameter_93: class Program_weight_tensor_parameter_94: name = "parameter_94" + original_name = "batch_norm2d_27.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -1015,6 +1110,7 @@ class Program_weight_tensor_parameter_94: class Program_weight_tensor_parameter_95: name = "parameter_95" + original_name = "batch_norm2d_27.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -1024,6 +1120,7 @@ class Program_weight_tensor_parameter_95: class Program_weight_tensor_parameter_96: name = "parameter_96" + original_name = "batch_norm2d_27.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -1033,6 +1130,7 @@ class Program_weight_tensor_parameter_96: class Program_weight_tensor_parameter_97: name = "parameter_97" + original_name = "conv2d_27.w_0" shape = [64, 64, 1, 1] dtype = "float32" min_val = float("-0.753258") @@ -1044,6 +1142,7 @@ class Program_weight_tensor_parameter_97: class Program_weight_tensor_parameter_98: name = "parameter_98" + original_name = "batch_norm2d_26.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -1053,6 +1152,7 @@ class Program_weight_tensor_parameter_98: class Program_weight_tensor_parameter_99: name = "parameter_99" + original_name = "batch_norm2d_26.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -1062,6 +1162,7 @@ class Program_weight_tensor_parameter_99: class Program_weight_tensor_parameter_100: name = "parameter_100" + original_name = "batch_norm2d_26.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -1071,6 +1172,7 @@ class Program_weight_tensor_parameter_100: class Program_weight_tensor_parameter_101: name = "parameter_101" + original_name = "batch_norm2d_26.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -1080,6 +1182,7 @@ class Program_weight_tensor_parameter_101: class Program_weight_tensor_parameter_102: name = "parameter_102" + original_name = "conv2d_26.w_0" shape = [64, 1, 5, 5] dtype = "float32" min_val = float("-1.1167") @@ -1091,6 +1194,7 @@ class Program_weight_tensor_parameter_102: class Program_weight_tensor_parameter_103: name = "parameter_103" + original_name = "batch_norm2d_25.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -1100,6 +1204,7 @@ class Program_weight_tensor_parameter_103: class Program_weight_tensor_parameter_104: name = "parameter_104" + original_name = "batch_norm2d_25.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -1109,6 +1214,7 @@ class Program_weight_tensor_parameter_104: class Program_weight_tensor_parameter_105: name = "parameter_105" + original_name = "batch_norm2d_25.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -1118,6 +1224,7 @@ class Program_weight_tensor_parameter_105: class Program_weight_tensor_parameter_106: name = "parameter_106" + original_name = "batch_norm2d_25.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -1127,6 +1234,7 @@ class Program_weight_tensor_parameter_106: class Program_weight_tensor_parameter_107: name = "parameter_107" + original_name = "conv2d_25.w_0" shape = [64, 512, 1, 1] dtype = "float32" min_val = float("-1.07484") @@ -1138,6 +1246,7 @@ class Program_weight_tensor_parameter_107: class Program_weight_tensor_parameter_108: name = "parameter_108" + original_name = "batch_norm2d_24.b_0" shape = [512] dtype = "float32" min_val = float("-11.7337") @@ -1149,6 +1258,7 @@ class Program_weight_tensor_parameter_108: class Program_weight_tensor_parameter_109: name = "parameter_109" + original_name = "batch_norm2d_24.w_0" shape = [512] dtype = "float32" min_val = float("0.611141") @@ -1160,6 +1270,7 @@ class Program_weight_tensor_parameter_109: class Program_weight_tensor_parameter_110: name = "parameter_110" + original_name = "batch_norm2d_24.w_2" shape = [512] dtype = "float32" min_val = float("0.575679") @@ -1171,6 +1282,7 @@ class Program_weight_tensor_parameter_110: class Program_weight_tensor_parameter_111: name = "parameter_111" + original_name = "batch_norm2d_24.w_1" shape = [512] dtype = "float32" min_val = float("-5.24368") @@ -1182,6 +1294,7 @@ class Program_weight_tensor_parameter_111: class Program_weight_tensor_parameter_112: name = "parameter_112" + original_name = "conv2d_24.w_0" shape = [512, 256, 1, 1] dtype = "float32" min_val = float("-1.40225") @@ -1193,6 +1306,7 @@ class Program_weight_tensor_parameter_112: class Program_weight_tensor_parameter_113: name = "parameter_113" + original_name = "batch_norm2d_23.b_0" shape = [256] dtype = "float32" min_val = float("-16.1964") @@ -1204,6 +1318,7 @@ class Program_weight_tensor_parameter_113: class Program_weight_tensor_parameter_114: name = "parameter_114" + original_name = "batch_norm2d_23.w_0" shape = [256] dtype = "float32" min_val = float("1.43554") @@ -1215,6 +1330,7 @@ class Program_weight_tensor_parameter_114: class Program_weight_tensor_parameter_115: name = "parameter_115" + original_name = "batch_norm2d_23.w_2" shape = [256] dtype = "float32" min_val = float("6.85103") @@ -1226,6 +1342,7 @@ class Program_weight_tensor_parameter_115: class Program_weight_tensor_parameter_116: name = "parameter_116" + original_name = "batch_norm2d_23.w_1" shape = [256] dtype = "float32" min_val = float("-16.69") @@ -1237,6 +1354,7 @@ class Program_weight_tensor_parameter_116: class Program_weight_tensor_parameter_117: name = "parameter_117" + original_name = "conv2d_23.w_0" shape = [256, 448, 1, 1] dtype = "float32" min_val = float("-1.12622") @@ -1248,6 +1366,7 @@ class Program_weight_tensor_parameter_117: class Program_weight_tensor_parameter_118: name = "parameter_118" + original_name = "batch_norm2d_22.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -1257,6 +1376,7 @@ class Program_weight_tensor_parameter_118: class Program_weight_tensor_parameter_119: name = "parameter_119" + original_name = "batch_norm2d_22.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -1266,6 +1386,7 @@ class Program_weight_tensor_parameter_119: class Program_weight_tensor_parameter_120: name = "parameter_120" + original_name = "batch_norm2d_22.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -1275,6 +1396,7 @@ class Program_weight_tensor_parameter_120: class Program_weight_tensor_parameter_121: name = "parameter_121" + original_name = "batch_norm2d_22.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -1284,6 +1406,7 @@ class Program_weight_tensor_parameter_121: class Program_weight_tensor_parameter_122: name = "parameter_122" + original_name = "conv2d_22.w_0" shape = [64, 1, 5, 5] dtype = "float32" min_val = float("-0.472804") @@ -1295,6 +1418,7 @@ class Program_weight_tensor_parameter_122: class Program_weight_tensor_parameter_123: name = "parameter_123" + original_name = "batch_norm2d_21.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -1304,6 +1428,7 @@ class Program_weight_tensor_parameter_123: class Program_weight_tensor_parameter_124: name = "parameter_124" + original_name = "batch_norm2d_21.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -1313,6 +1438,7 @@ class Program_weight_tensor_parameter_124: class Program_weight_tensor_parameter_125: name = "parameter_125" + original_name = "batch_norm2d_21.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -1322,6 +1448,7 @@ class Program_weight_tensor_parameter_125: class Program_weight_tensor_parameter_126: name = "parameter_126" + original_name = "batch_norm2d_21.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -1331,6 +1458,7 @@ class Program_weight_tensor_parameter_126: class Program_weight_tensor_parameter_127: name = "parameter_127" + original_name = "conv2d_21.w_0" shape = [64, 64, 1, 1] dtype = "float32" min_val = float("-0.886575") @@ -1342,6 +1470,7 @@ class Program_weight_tensor_parameter_127: class Program_weight_tensor_parameter_128: name = "parameter_128" + original_name = "batch_norm2d_20.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -1351,6 +1480,7 @@ class Program_weight_tensor_parameter_128: class Program_weight_tensor_parameter_129: name = "parameter_129" + original_name = "batch_norm2d_20.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -1360,6 +1490,7 @@ class Program_weight_tensor_parameter_129: class Program_weight_tensor_parameter_130: name = "parameter_130" + original_name = "batch_norm2d_20.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -1369,6 +1500,7 @@ class Program_weight_tensor_parameter_130: class Program_weight_tensor_parameter_131: name = "parameter_131" + original_name = "batch_norm2d_20.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -1378,6 +1510,7 @@ class Program_weight_tensor_parameter_131: class Program_weight_tensor_parameter_132: name = "parameter_132" + original_name = "conv2d_20.w_0" shape = [64, 1, 5, 5] dtype = "float32" min_val = float("-1.35696") @@ -1389,6 +1522,7 @@ class Program_weight_tensor_parameter_132: class Program_weight_tensor_parameter_133: name = "parameter_133" + original_name = "batch_norm2d_19.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -1398,6 +1532,7 @@ class Program_weight_tensor_parameter_133: class Program_weight_tensor_parameter_134: name = "parameter_134" + original_name = "batch_norm2d_19.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -1407,6 +1542,7 @@ class Program_weight_tensor_parameter_134: class Program_weight_tensor_parameter_135: name = "parameter_135" + original_name = "batch_norm2d_19.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -1416,6 +1552,7 @@ class Program_weight_tensor_parameter_135: class Program_weight_tensor_parameter_136: name = "parameter_136" + original_name = "batch_norm2d_19.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -1425,6 +1562,7 @@ class Program_weight_tensor_parameter_136: class Program_weight_tensor_parameter_137: name = "parameter_137" + original_name = "conv2d_19.w_0" shape = [64, 64, 1, 1] dtype = "float32" min_val = float("-1.34598") @@ -1436,6 +1574,7 @@ class Program_weight_tensor_parameter_137: class Program_weight_tensor_parameter_138: name = "parameter_138" + original_name = "batch_norm2d_18.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -1445,6 +1584,7 @@ class Program_weight_tensor_parameter_138: class Program_weight_tensor_parameter_139: name = "parameter_139" + original_name = "batch_norm2d_18.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -1454,6 +1594,7 @@ class Program_weight_tensor_parameter_139: class Program_weight_tensor_parameter_140: name = "parameter_140" + original_name = "batch_norm2d_18.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -1463,6 +1604,7 @@ class Program_weight_tensor_parameter_140: class Program_weight_tensor_parameter_141: name = "parameter_141" + original_name = "batch_norm2d_18.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -1472,6 +1614,7 @@ class Program_weight_tensor_parameter_141: class Program_weight_tensor_parameter_142: name = "parameter_142" + original_name = "conv2d_18.w_0" shape = [64, 1, 5, 5] dtype = "float32" min_val = float("-1.23043") @@ -1483,6 +1626,7 @@ class Program_weight_tensor_parameter_142: class Program_weight_tensor_parameter_143: name = "parameter_143" + original_name = "batch_norm2d_17.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -1492,6 +1636,7 @@ class Program_weight_tensor_parameter_143: class Program_weight_tensor_parameter_144: name = "parameter_144" + original_name = "batch_norm2d_17.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -1501,6 +1646,7 @@ class Program_weight_tensor_parameter_144: class Program_weight_tensor_parameter_145: name = "parameter_145" + original_name = "batch_norm2d_17.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -1510,6 +1656,7 @@ class Program_weight_tensor_parameter_145: class Program_weight_tensor_parameter_146: name = "parameter_146" + original_name = "batch_norm2d_17.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -1519,6 +1666,7 @@ class Program_weight_tensor_parameter_146: class Program_weight_tensor_parameter_147: name = "parameter_147" + original_name = "conv2d_17.w_0" shape = [64, 256, 1, 1] dtype = "float32" min_val = float("-1.0854") @@ -1530,6 +1678,7 @@ class Program_weight_tensor_parameter_147: class Program_weight_tensor_parameter_148: name = "parameter_148" + original_name = "batch_norm2d_16.b_0" shape = [256] dtype = "float32" min_val = float("-0.0122923") @@ -1541,6 +1690,7 @@ class Program_weight_tensor_parameter_148: class Program_weight_tensor_parameter_149: name = "parameter_149" + original_name = "batch_norm2d_16.w_0" shape = [256] dtype = "float32" min_val = float("1.75624") @@ -1552,6 +1702,7 @@ class Program_weight_tensor_parameter_149: class Program_weight_tensor_parameter_150: name = "parameter_150" + original_name = "batch_norm2d_16.w_2" shape = [256] dtype = "float32" min_val = float("0.000276555") @@ -1563,6 +1714,7 @@ class Program_weight_tensor_parameter_150: class Program_weight_tensor_parameter_151: name = "parameter_151" + original_name = "batch_norm2d_16.w_1" shape = [256] dtype = "float32" min_val = float("-2.1523") @@ -1574,6 +1726,7 @@ class Program_weight_tensor_parameter_151: class Program_weight_tensor_parameter_152: name = "parameter_152" + original_name = "conv2d_16.w_0" shape = [256, 1, 3, 3] dtype = "float32" min_val = float("-0.356065") @@ -1585,6 +1738,7 @@ class Program_weight_tensor_parameter_152: class Program_weight_tensor_parameter_153: name = "parameter_153" + original_name = "batch_norm2d_15.b_0" shape = [256] dtype = "float32" min_val = float("-4.03256") @@ -1596,6 +1750,7 @@ class Program_weight_tensor_parameter_153: class Program_weight_tensor_parameter_154: name = "parameter_154" + original_name = "batch_norm2d_15.w_0" shape = [256] dtype = "float32" min_val = float("0.0511606") @@ -1607,6 +1762,7 @@ class Program_weight_tensor_parameter_154: class Program_weight_tensor_parameter_155: name = "parameter_155" + original_name = "batch_norm2d_15.w_2" shape = [256] dtype = "float32" min_val = float("1.06249") @@ -1618,6 +1774,7 @@ class Program_weight_tensor_parameter_155: class Program_weight_tensor_parameter_156: name = "parameter_156" + original_name = "batch_norm2d_15.w_1" shape = [256] dtype = "float32" min_val = float("-7.52144") @@ -1629,6 +1786,7 @@ class Program_weight_tensor_parameter_156: class Program_weight_tensor_parameter_157: name = "parameter_157" + original_name = "conv2d_15.w_0" shape = [256, 128, 1, 1] dtype = "float32" min_val = float("-1.16676") @@ -1640,6 +1798,7 @@ class Program_weight_tensor_parameter_157: class Program_weight_tensor_parameter_158: name = "parameter_158" + original_name = "batch_norm2d_14.b_0" shape = [128] dtype = "float32" min_val = float("-5.05021") @@ -1651,6 +1810,7 @@ class Program_weight_tensor_parameter_158: class Program_weight_tensor_parameter_159: name = "parameter_159" + original_name = "batch_norm2d_14.w_0" shape = [128] dtype = "float32" min_val = float("1.77083") @@ -1662,6 +1822,7 @@ class Program_weight_tensor_parameter_159: class Program_weight_tensor_parameter_160: name = "parameter_160" + original_name = "batch_norm2d_14.w_2" shape = [128] dtype = "float32" min_val = float("9.64208") @@ -1673,6 +1834,7 @@ class Program_weight_tensor_parameter_160: class Program_weight_tensor_parameter_161: name = "parameter_161" + original_name = "batch_norm2d_14.w_1" shape = [128] dtype = "float32" min_val = float("-12.4181") @@ -1684,6 +1846,7 @@ class Program_weight_tensor_parameter_161: class Program_weight_tensor_parameter_162: name = "parameter_162" + original_name = "conv2d_14.w_0" shape = [128, 160, 1, 1] dtype = "float32" min_val = float("-1.47651") @@ -1695,6 +1858,7 @@ class Program_weight_tensor_parameter_162: class Program_weight_tensor_parameter_163: name = "parameter_163" + original_name = "batch_norm2d_13.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -1704,6 +1868,7 @@ class Program_weight_tensor_parameter_163: class Program_weight_tensor_parameter_164: name = "parameter_164" + original_name = "batch_norm2d_13.w_0" shape = [32] dtype = "float32" min_val = float("0") @@ -1713,6 +1878,7 @@ class Program_weight_tensor_parameter_164: class Program_weight_tensor_parameter_165: name = "parameter_165" + original_name = "batch_norm2d_13.w_2" shape = [32] dtype = "float32" min_val = float("0") @@ -1722,6 +1888,7 @@ class Program_weight_tensor_parameter_165: class Program_weight_tensor_parameter_166: name = "parameter_166" + original_name = "batch_norm2d_13.w_1" shape = [32] dtype = "float32" min_val = float("0") @@ -1731,6 +1898,7 @@ class Program_weight_tensor_parameter_166: class Program_weight_tensor_parameter_167: name = "parameter_167" + original_name = "conv2d_13.w_0" shape = [32, 32, 3, 3] dtype = "float32" min_val = float("-0.660673") @@ -1742,6 +1910,7 @@ class Program_weight_tensor_parameter_167: class Program_weight_tensor_parameter_168: name = "parameter_168" + original_name = "batch_norm2d_12.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -1751,6 +1920,7 @@ class Program_weight_tensor_parameter_168: class Program_weight_tensor_parameter_169: name = "parameter_169" + original_name = "batch_norm2d_12.w_0" shape = [32] dtype = "float32" min_val = float("0") @@ -1760,6 +1930,7 @@ class Program_weight_tensor_parameter_169: class Program_weight_tensor_parameter_170: name = "parameter_170" + original_name = "batch_norm2d_12.w_2" shape = [32] dtype = "float32" min_val = float("0") @@ -1769,6 +1940,7 @@ class Program_weight_tensor_parameter_170: class Program_weight_tensor_parameter_171: name = "parameter_171" + original_name = "batch_norm2d_12.w_1" shape = [32] dtype = "float32" min_val = float("0") @@ -1778,6 +1950,7 @@ class Program_weight_tensor_parameter_171: class Program_weight_tensor_parameter_172: name = "parameter_172" + original_name = "conv2d_12.w_0" shape = [32, 32, 3, 3] dtype = "float32" min_val = float("-0.98024") @@ -1789,6 +1962,7 @@ class Program_weight_tensor_parameter_172: class Program_weight_tensor_parameter_173: name = "parameter_173" + original_name = "batch_norm2d_11.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -1798,6 +1972,7 @@ class Program_weight_tensor_parameter_173: class Program_weight_tensor_parameter_174: name = "parameter_174" + original_name = "batch_norm2d_11.w_0" shape = [32] dtype = "float32" min_val = float("0") @@ -1807,6 +1982,7 @@ class Program_weight_tensor_parameter_174: class Program_weight_tensor_parameter_175: name = "parameter_175" + original_name = "batch_norm2d_11.w_2" shape = [32] dtype = "float32" min_val = float("0") @@ -1816,6 +1992,7 @@ class Program_weight_tensor_parameter_175: class Program_weight_tensor_parameter_176: name = "parameter_176" + original_name = "batch_norm2d_11.w_1" shape = [32] dtype = "float32" min_val = float("0") @@ -1825,6 +2002,7 @@ class Program_weight_tensor_parameter_176: class Program_weight_tensor_parameter_177: name = "parameter_177" + original_name = "conv2d_11.w_0" shape = [32, 64, 3, 3] dtype = "float32" min_val = float("-1.54695") @@ -1836,6 +2014,7 @@ class Program_weight_tensor_parameter_177: class Program_weight_tensor_parameter_178: name = "parameter_178" + original_name = "batch_norm2d_10.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -1845,6 +2024,7 @@ class Program_weight_tensor_parameter_178: class Program_weight_tensor_parameter_179: name = "parameter_179" + original_name = "batch_norm2d_10.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -1854,6 +2034,7 @@ class Program_weight_tensor_parameter_179: class Program_weight_tensor_parameter_180: name = "parameter_180" + original_name = "batch_norm2d_10.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -1863,6 +2044,7 @@ class Program_weight_tensor_parameter_180: class Program_weight_tensor_parameter_181: name = "parameter_181" + original_name = "batch_norm2d_10.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -1872,6 +2054,7 @@ class Program_weight_tensor_parameter_181: class Program_weight_tensor_parameter_182: name = "parameter_182" + original_name = "conv2d_10.w_0" shape = [64, 1, 3, 3] dtype = "float32" min_val = float("-0.71408") @@ -1883,6 +2066,7 @@ class Program_weight_tensor_parameter_182: class Program_weight_tensor_parameter_183: name = "parameter_183" + original_name = "batch_norm2d_9.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -1892,6 +2076,7 @@ class Program_weight_tensor_parameter_183: class Program_weight_tensor_parameter_184: name = "parameter_184" + original_name = "batch_norm2d_9.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -1901,6 +2086,7 @@ class Program_weight_tensor_parameter_184: class Program_weight_tensor_parameter_185: name = "parameter_185" + original_name = "batch_norm2d_9.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -1910,6 +2096,7 @@ class Program_weight_tensor_parameter_185: class Program_weight_tensor_parameter_186: name = "parameter_186" + original_name = "batch_norm2d_9.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -1919,6 +2106,7 @@ class Program_weight_tensor_parameter_186: class Program_weight_tensor_parameter_187: name = "parameter_187" + original_name = "conv2d_9.w_0" shape = [64, 32, 1, 1] dtype = "float32" min_val = float("-1.96355") @@ -1930,6 +2118,7 @@ class Program_weight_tensor_parameter_187: class Program_weight_tensor_parameter_188: name = "parameter_188" + original_name = "batch_norm2d_8.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -1939,6 +2128,7 @@ class Program_weight_tensor_parameter_188: class Program_weight_tensor_parameter_189: name = "parameter_189" + original_name = "batch_norm2d_8.w_0" shape = [32] dtype = "float32" min_val = float("0") @@ -1948,6 +2138,7 @@ class Program_weight_tensor_parameter_189: class Program_weight_tensor_parameter_190: name = "parameter_190" + original_name = "batch_norm2d_8.w_2" shape = [32] dtype = "float32" min_val = float("0") @@ -1957,6 +2148,7 @@ class Program_weight_tensor_parameter_190: class Program_weight_tensor_parameter_191: name = "parameter_191" + original_name = "batch_norm2d_8.w_1" shape = [32] dtype = "float32" min_val = float("0") @@ -1966,6 +2158,7 @@ class Program_weight_tensor_parameter_191: class Program_weight_tensor_parameter_192: name = "parameter_192" + original_name = "conv2d_8.w_0" shape = [32, 64, 1, 1] dtype = "float32" min_val = float("-2.21591") @@ -1977,6 +2170,7 @@ class Program_weight_tensor_parameter_192: class Program_weight_tensor_parameter_193: name = "parameter_193" + original_name = "batch_norm2d_7.b_0" shape = [16] dtype = "float32" min_val = float("0") @@ -1986,6 +2180,7 @@ class Program_weight_tensor_parameter_193: class Program_weight_tensor_parameter_194: name = "parameter_194" + original_name = "batch_norm2d_7.w_0" shape = [16] dtype = "float32" min_val = float("0") @@ -1995,6 +2190,7 @@ class Program_weight_tensor_parameter_194: class Program_weight_tensor_parameter_195: name = "parameter_195" + original_name = "batch_norm2d_7.w_2" shape = [16] dtype = "float32" min_val = float("0") @@ -2004,6 +2200,7 @@ class Program_weight_tensor_parameter_195: class Program_weight_tensor_parameter_196: name = "parameter_196" + original_name = "batch_norm2d_7.w_1" shape = [16] dtype = "float32" min_val = float("0") @@ -2013,6 +2210,7 @@ class Program_weight_tensor_parameter_196: class Program_weight_tensor_parameter_197: name = "parameter_197" + original_name = "conv2d_7.w_0" shape = [16, 16, 3, 3] dtype = "float32" min_val = float("-1.14047") @@ -2024,6 +2222,7 @@ class Program_weight_tensor_parameter_197: class Program_weight_tensor_parameter_198: name = "parameter_198" + original_name = "batch_norm2d_6.b_0" shape = [16] dtype = "float32" min_val = float("0") @@ -2033,6 +2232,7 @@ class Program_weight_tensor_parameter_198: class Program_weight_tensor_parameter_199: name = "parameter_199" + original_name = "batch_norm2d_6.w_0" shape = [16] dtype = "float32" min_val = float("0") @@ -2042,6 +2242,7 @@ class Program_weight_tensor_parameter_199: class Program_weight_tensor_parameter_200: name = "parameter_200" + original_name = "batch_norm2d_6.w_2" shape = [16] dtype = "float32" min_val = float("0") @@ -2051,6 +2252,7 @@ class Program_weight_tensor_parameter_200: class Program_weight_tensor_parameter_201: name = "parameter_201" + original_name = "batch_norm2d_6.w_1" shape = [16] dtype = "float32" min_val = float("0") @@ -2060,6 +2262,7 @@ class Program_weight_tensor_parameter_201: class Program_weight_tensor_parameter_202: name = "parameter_202" + original_name = "conv2d_6.w_0" shape = [16, 16, 3, 3] dtype = "float32" min_val = float("-1.10075") @@ -2071,6 +2274,7 @@ class Program_weight_tensor_parameter_202: class Program_weight_tensor_parameter_203: name = "parameter_203" + original_name = "batch_norm2d_5.b_0" shape = [16] dtype = "float32" min_val = float("0") @@ -2080,6 +2284,7 @@ class Program_weight_tensor_parameter_203: class Program_weight_tensor_parameter_204: name = "parameter_204" + original_name = "batch_norm2d_5.w_0" shape = [16] dtype = "float32" min_val = float("0") @@ -2089,6 +2294,7 @@ class Program_weight_tensor_parameter_204: class Program_weight_tensor_parameter_205: name = "parameter_205" + original_name = "batch_norm2d_5.w_2" shape = [16] dtype = "float32" min_val = float("0") @@ -2098,6 +2304,7 @@ class Program_weight_tensor_parameter_205: class Program_weight_tensor_parameter_206: name = "parameter_206" + original_name = "batch_norm2d_5.w_1" shape = [16] dtype = "float32" min_val = float("0") @@ -2107,6 +2314,7 @@ class Program_weight_tensor_parameter_206: class Program_weight_tensor_parameter_207: name = "parameter_207" + original_name = "conv2d_5.w_0" shape = [16, 16, 3, 3] dtype = "float32" min_val = float("-1.99491") @@ -2118,6 +2326,7 @@ class Program_weight_tensor_parameter_207: class Program_weight_tensor_parameter_208: name = "parameter_208" + original_name = "batch_norm2d_4.b_0" shape = [16] dtype = "float32" min_val = float("0") @@ -2127,6 +2336,7 @@ class Program_weight_tensor_parameter_208: class Program_weight_tensor_parameter_209: name = "parameter_209" + original_name = "batch_norm2d_4.w_0" shape = [16] dtype = "float32" min_val = float("0") @@ -2136,6 +2346,7 @@ class Program_weight_tensor_parameter_209: class Program_weight_tensor_parameter_210: name = "parameter_210" + original_name = "batch_norm2d_4.w_2" shape = [16] dtype = "float32" min_val = float("0") @@ -2145,6 +2356,7 @@ class Program_weight_tensor_parameter_210: class Program_weight_tensor_parameter_211: name = "parameter_211" + original_name = "batch_norm2d_4.w_1" shape = [16] dtype = "float32" min_val = float("0") @@ -2154,6 +2366,7 @@ class Program_weight_tensor_parameter_211: class Program_weight_tensor_parameter_212: name = "parameter_212" + original_name = "conv2d_4.w_0" shape = [16, 16, 1, 1] dtype = "float32" min_val = float("-2.41969") @@ -2165,6 +2378,7 @@ class Program_weight_tensor_parameter_212: class Program_weight_tensor_parameter_213: name = "parameter_213" + original_name = "batch_norm2d_3.b_0" shape = [16] dtype = "float32" min_val = float("0") @@ -2174,6 +2388,7 @@ class Program_weight_tensor_parameter_213: class Program_weight_tensor_parameter_214: name = "parameter_214" + original_name = "batch_norm2d_3.w_0" shape = [16] dtype = "float32" min_val = float("0") @@ -2183,6 +2398,7 @@ class Program_weight_tensor_parameter_214: class Program_weight_tensor_parameter_215: name = "parameter_215" + original_name = "batch_norm2d_3.w_2" shape = [16] dtype = "float32" min_val = float("0") @@ -2192,6 +2408,7 @@ class Program_weight_tensor_parameter_215: class Program_weight_tensor_parameter_216: name = "parameter_216" + original_name = "batch_norm2d_3.w_1" shape = [16] dtype = "float32" min_val = float("0") @@ -2201,6 +2418,7 @@ class Program_weight_tensor_parameter_216: class Program_weight_tensor_parameter_217: name = "parameter_217" + original_name = "conv2d_3.w_0" shape = [16, 32, 3, 3] dtype = "float32" min_val = float("-1.57878") @@ -2212,6 +2430,7 @@ class Program_weight_tensor_parameter_217: class Program_weight_tensor_parameter_218: name = "parameter_218" + original_name = "batch_norm2d_2.b_0" shape = [16] dtype = "float32" min_val = float("0") @@ -2221,6 +2440,7 @@ class Program_weight_tensor_parameter_218: class Program_weight_tensor_parameter_219: name = "parameter_219" + original_name = "batch_norm2d_2.w_0" shape = [16] dtype = "float32" min_val = float("0") @@ -2230,6 +2450,7 @@ class Program_weight_tensor_parameter_219: class Program_weight_tensor_parameter_220: name = "parameter_220" + original_name = "batch_norm2d_2.w_2" shape = [16] dtype = "float32" min_val = float("0") @@ -2239,6 +2460,7 @@ class Program_weight_tensor_parameter_220: class Program_weight_tensor_parameter_221: name = "parameter_221" + original_name = "batch_norm2d_2.w_1" shape = [16] dtype = "float32" min_val = float("0") @@ -2248,6 +2470,7 @@ class Program_weight_tensor_parameter_221: class Program_weight_tensor_parameter_222: name = "parameter_222" + original_name = "conv2d_2.w_0" shape = [16, 8, 2, 2] dtype = "float32" min_val = float("-1.1654") @@ -2259,6 +2482,7 @@ class Program_weight_tensor_parameter_222: class Program_weight_tensor_parameter_223: name = "parameter_223" + original_name = "batch_norm2d_1.b_0" shape = [8] dtype = "float32" min_val = float("0") @@ -2268,6 +2492,7 @@ class Program_weight_tensor_parameter_223: class Program_weight_tensor_parameter_224: name = "parameter_224" + original_name = "batch_norm2d_1.w_0" shape = [8] dtype = "float32" min_val = float("0") @@ -2277,6 +2502,7 @@ class Program_weight_tensor_parameter_224: class Program_weight_tensor_parameter_225: name = "parameter_225" + original_name = "batch_norm2d_1.w_2" shape = [8] dtype = "float32" min_val = float("0") @@ -2286,6 +2512,7 @@ class Program_weight_tensor_parameter_225: class Program_weight_tensor_parameter_226: name = "parameter_226" + original_name = "batch_norm2d_1.w_1" shape = [8] dtype = "float32" min_val = float("0") @@ -2295,6 +2522,7 @@ class Program_weight_tensor_parameter_226: class Program_weight_tensor_parameter_227: name = "parameter_227" + original_name = "conv2d_1.w_0" shape = [8, 16, 2, 2] dtype = "float32" min_val = float("-1.20487") @@ -2306,6 +2534,7 @@ class Program_weight_tensor_parameter_227: class Program_weight_tensor_parameter_228: name = "parameter_228" + original_name = "batch_norm2d_0.b_0" shape = [16] dtype = "float32" min_val = float("0") @@ -2315,6 +2544,7 @@ class Program_weight_tensor_parameter_228: class Program_weight_tensor_parameter_229: name = "parameter_229" + original_name = "batch_norm2d_0.w_0" shape = [16] dtype = "float32" min_val = float("0") @@ -2324,6 +2554,7 @@ class Program_weight_tensor_parameter_229: class Program_weight_tensor_parameter_230: name = "parameter_230" + original_name = "batch_norm2d_0.w_2" shape = [16] dtype = "float32" min_val = float("0") @@ -2333,6 +2564,7 @@ class Program_weight_tensor_parameter_230: class Program_weight_tensor_parameter_231: name = "parameter_231" + original_name = "batch_norm2d_0.w_1" shape = [16] dtype = "float32" min_val = float("0") @@ -2342,6 +2574,7 @@ class Program_weight_tensor_parameter_231: class Program_weight_tensor_parameter_232: name = "parameter_232" + original_name = "conv2d_0.w_0" shape = [16, 3, 3, 3] dtype = "float32" min_val = float("-2.15343") diff --git a/paddle_samples/PaddleX/PP-HGNetV2-B0_ML/subgraph_2/input_meta.py b/paddle_samples/PaddleX/PP-HGNetV2-B0_ML/subgraph_2/input_meta.py index 7a3571791..71e43c183 100644 --- a/paddle_samples/PaddleX/PP-HGNetV2-B0_ML/subgraph_2/input_meta.py +++ b/paddle_samples/PaddleX/PP-HGNetV2-B0_ML/subgraph_2/input_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_data_0: name = "data_0" + original_name = "param_0" shape = [1] dtype = "float32" data = [0.16544] @@ -7,6 +8,7 @@ class Program_weight_tensor_data_0: class Program_weight_tensor_data_1: name = "data_1" + original_name = "param_1" shape = [1] dtype = "float32" data = [-1.4634] @@ -14,6 +16,7 @@ class Program_weight_tensor_data_1: class Program_weight_tensor_data_2: name = "data_2" + original_name = "param_10" shape = [1] dtype = "float32" data = [0.347854] @@ -21,6 +24,7 @@ class Program_weight_tensor_data_2: class Program_weight_tensor_data_3: name = "data_3" + original_name = "param_11" shape = [1] dtype = "float32" data = [-1.32627] @@ -28,6 +32,7 @@ class Program_weight_tensor_data_3: class Program_weight_tensor_data_4: name = "data_4" + original_name = "param_12" shape = [1] dtype = "float32" data = [0.421071] @@ -35,6 +40,7 @@ class Program_weight_tensor_data_4: class Program_weight_tensor_data_5: name = "data_5" + original_name = "param_13" shape = [1] dtype = "float32" data = [-0.383221] @@ -42,6 +48,7 @@ class Program_weight_tensor_data_5: class Program_weight_tensor_data_6: name = "data_6" + original_name = "param_14" shape = [1] dtype = "float32" data = [0.37269] @@ -49,6 +56,7 @@ class Program_weight_tensor_data_6: class Program_weight_tensor_data_7: name = "data_7" + original_name = "param_15" shape = [1] dtype = "float32" data = [6.41386e-05] @@ -56,6 +64,7 @@ class Program_weight_tensor_data_7: class Program_weight_tensor_data_8: name = "data_8" + original_name = "param_16" shape = [1] dtype = "float32" data = [0.662726] @@ -63,6 +72,7 @@ class Program_weight_tensor_data_8: class Program_weight_tensor_data_9: name = "data_9" + original_name = "param_17" shape = [1] dtype = "float32" data = [4.34099e-05] @@ -70,6 +80,7 @@ class Program_weight_tensor_data_9: class Program_weight_tensor_data_10: name = "data_10" + original_name = "param_18" shape = [1] dtype = "float32" data = [0.469679] @@ -77,6 +88,7 @@ class Program_weight_tensor_data_10: class Program_weight_tensor_data_11: name = "data_11" + original_name = "param_19" shape = [1] dtype = "float32" data = [-0.524591] @@ -84,6 +96,7 @@ class Program_weight_tensor_data_11: class Program_weight_tensor_data_12: name = "data_12" + original_name = "param_2" shape = [1] dtype = "float32" data = [0.168583] @@ -91,6 +104,7 @@ class Program_weight_tensor_data_12: class Program_weight_tensor_data_13: name = "data_13" + original_name = "param_20" shape = [1] dtype = "float32" data = [1.22378] @@ -98,6 +112,7 @@ class Program_weight_tensor_data_13: class Program_weight_tensor_data_14: name = "data_14" + original_name = "param_21" shape = [1] dtype = "float32" data = [-0.937791] @@ -105,6 +120,7 @@ class Program_weight_tensor_data_14: class Program_weight_tensor_data_15: name = "data_15" + original_name = "param_22" shape = [1] dtype = "float32" data = [1.27281] @@ -112,6 +128,7 @@ class Program_weight_tensor_data_15: class Program_weight_tensor_data_16: name = "data_16" + original_name = "param_23" shape = [1] dtype = "float32" data = [-0.336926] @@ -119,6 +136,7 @@ class Program_weight_tensor_data_16: class Program_weight_tensor_data_17: name = "data_17" + original_name = "param_24" shape = [1] dtype = "float32" data = [1.41884] @@ -126,6 +144,7 @@ class Program_weight_tensor_data_17: class Program_weight_tensor_data_18: name = "data_18" + original_name = "param_25" shape = [1] dtype = "float32" data = [8.68882e-05] @@ -133,6 +152,7 @@ class Program_weight_tensor_data_18: class Program_weight_tensor_data_19: name = "data_19" + original_name = "param_26" shape = [1] dtype = "float32" data = [0.656092] @@ -140,6 +160,7 @@ class Program_weight_tensor_data_19: class Program_weight_tensor_data_20: name = "data_20" + original_name = "param_27" shape = [1] dtype = "float32" data = [-2.70586e-05] @@ -147,6 +168,7 @@ class Program_weight_tensor_data_20: class Program_weight_tensor_data_21: name = "data_21" + original_name = "param_28" shape = [1] dtype = "float32" data = [0.666898] @@ -154,6 +176,7 @@ class Program_weight_tensor_data_21: class Program_weight_tensor_data_22: name = "data_22" + original_name = "param_29" shape = [1] dtype = "float32" data = [-0.227693] @@ -161,6 +184,7 @@ class Program_weight_tensor_data_22: class Program_weight_tensor_data_23: name = "data_23" + original_name = "param_3" shape = [1] dtype = "float32" data = [-0.637096] @@ -168,6 +192,7 @@ class Program_weight_tensor_data_23: class Program_weight_tensor_data_24: name = "data_24" + original_name = "param_30" shape = [1] dtype = "float32" data = [1.34858] @@ -175,6 +200,7 @@ class Program_weight_tensor_data_24: class Program_weight_tensor_data_25: name = "data_25" + original_name = "param_31" shape = [1] dtype = "float32" data = [-7.77697e-05] @@ -182,6 +208,7 @@ class Program_weight_tensor_data_25: class Program_weight_tensor_data_26: name = "data_26" + original_name = "param_32" shape = [1] dtype = "float32" data = [1.26777] @@ -189,6 +216,7 @@ class Program_weight_tensor_data_26: class Program_weight_tensor_data_27: name = "data_27" + original_name = "param_33" shape = [1] dtype = "float32" data = [-3.16482e-05] @@ -196,6 +224,7 @@ class Program_weight_tensor_data_27: class Program_weight_tensor_data_28: name = "data_28" + original_name = "param_34" shape = [1] dtype = "float32" data = [1.62428] @@ -203,6 +232,7 @@ class Program_weight_tensor_data_28: class Program_weight_tensor_data_29: name = "data_29" + original_name = "param_35" shape = [1] dtype = "float32" data = [2.42148e-05] @@ -210,6 +240,7 @@ class Program_weight_tensor_data_29: class Program_weight_tensor_data_30: name = "data_30" + original_name = "param_36" shape = [1] dtype = "float32" data = [0.654679] @@ -217,6 +248,7 @@ class Program_weight_tensor_data_30: class Program_weight_tensor_data_31: name = "data_31" + original_name = "param_37" shape = [1] dtype = "float32" data = [0.000159355] @@ -224,6 +256,7 @@ class Program_weight_tensor_data_31: class Program_weight_tensor_data_32: name = "data_32" + original_name = "param_38" shape = [1] dtype = "float32" data = [0.354411] @@ -231,6 +264,7 @@ class Program_weight_tensor_data_32: class Program_weight_tensor_data_33: name = "data_33" + original_name = "param_39" shape = [1] dtype = "float32" data = [-0.227028] @@ -238,6 +272,7 @@ class Program_weight_tensor_data_33: class Program_weight_tensor_data_34: name = "data_34" + original_name = "param_4" shape = [1] dtype = "float32" data = [0.123237] @@ -245,6 +280,7 @@ class Program_weight_tensor_data_34: class Program_weight_tensor_data_35: name = "data_35" + original_name = "param_40" shape = [1] dtype = "float32" data = [0.279174] @@ -252,6 +288,7 @@ class Program_weight_tensor_data_35: class Program_weight_tensor_data_36: name = "data_36" + original_name = "param_41" shape = [1] dtype = "float32" data = [-0.000178568] @@ -259,6 +296,7 @@ class Program_weight_tensor_data_36: class Program_weight_tensor_data_37: name = "data_37" + original_name = "param_42" shape = [1] dtype = "float32" data = [0.264008] @@ -266,6 +304,7 @@ class Program_weight_tensor_data_37: class Program_weight_tensor_data_38: name = "data_38" + original_name = "param_43" shape = [1] dtype = "float32" data = [-4.15996e-05] @@ -273,6 +312,7 @@ class Program_weight_tensor_data_38: class Program_weight_tensor_data_39: name = "data_39" + original_name = "param_44" shape = [1] dtype = "float32" data = [0.316964] @@ -280,6 +320,7 @@ class Program_weight_tensor_data_39: class Program_weight_tensor_data_40: name = "data_40" + original_name = "param_45" shape = [1] dtype = "float32" data = [1.00781e-05] @@ -287,6 +328,7 @@ class Program_weight_tensor_data_40: class Program_weight_tensor_data_41: name = "data_41" + original_name = "param_46" shape = [1] dtype = "float32" data = [0.654381] @@ -294,6 +336,7 @@ class Program_weight_tensor_data_41: class Program_weight_tensor_data_42: name = "data_42" + original_name = "param_47" shape = [1] dtype = "float32" data = [0.000633419] @@ -301,6 +344,7 @@ class Program_weight_tensor_data_42: class Program_weight_tensor_data_43: name = "data_43" + original_name = "param_48" shape = [1] dtype = "float32" data = [1.6815] @@ -308,6 +352,7 @@ class Program_weight_tensor_data_43: class Program_weight_tensor_data_44: name = "data_44" + original_name = "param_49" shape = [1] dtype = "float32" data = [-0.226847] @@ -315,6 +360,7 @@ class Program_weight_tensor_data_44: class Program_weight_tensor_data_45: name = "data_45" + original_name = "param_5" shape = [1] dtype = "float32" data = [-0.0597621] @@ -322,6 +368,7 @@ class Program_weight_tensor_data_45: class Program_weight_tensor_data_46: name = "data_46" + original_name = "param_50" shape = [1] dtype = "float32" data = [1.35565] @@ -329,6 +376,7 @@ class Program_weight_tensor_data_46: class Program_weight_tensor_data_47: name = "data_47" + original_name = "param_51" shape = [1] dtype = "float32" data = [-2.50515e-05] @@ -336,6 +384,7 @@ class Program_weight_tensor_data_47: class Program_weight_tensor_data_48: name = "data_48" + original_name = "param_52" shape = [1] dtype = "float32" data = [1.45413] @@ -343,6 +392,7 @@ class Program_weight_tensor_data_48: class Program_weight_tensor_data_49: name = "data_49" + original_name = "param_53" shape = [1] dtype = "float32" data = [0.000101959] @@ -350,6 +400,7 @@ class Program_weight_tensor_data_49: class Program_weight_tensor_data_50: name = "data_50" + original_name = "param_54" shape = [1] dtype = "float32" data = [1.94576] @@ -357,6 +408,7 @@ class Program_weight_tensor_data_50: class Program_weight_tensor_data_51: name = "data_51" + original_name = "param_55" shape = [1] dtype = "float32" data = [5.18194e-06] @@ -364,6 +416,7 @@ class Program_weight_tensor_data_51: class Program_weight_tensor_data_52: name = "data_52" + original_name = "param_56" shape = [1] dtype = "float32" data = [0.651345] @@ -371,6 +424,7 @@ class Program_weight_tensor_data_52: class Program_weight_tensor_data_53: name = "data_53" + original_name = "param_57" shape = [1] dtype = "float32" data = [0.0108328] @@ -378,6 +432,7 @@ class Program_weight_tensor_data_53: class Program_weight_tensor_data_54: name = "data_54" + original_name = "param_58" shape = [1] dtype = "float32" data = [3.90236] @@ -385,6 +440,7 @@ class Program_weight_tensor_data_54: class Program_weight_tensor_data_55: name = "data_55" + original_name = "param_59" shape = [1] dtype = "float32" data = [-0.406518] @@ -392,6 +448,7 @@ class Program_weight_tensor_data_55: class Program_weight_tensor_data_56: name = "data_56" + original_name = "param_6" shape = [1] dtype = "float32" data = [0.703457] @@ -399,6 +456,7 @@ class Program_weight_tensor_data_56: class Program_weight_tensor_data_57: name = "data_57" + original_name = "param_60" shape = [1] dtype = "float32" data = [2.09855] @@ -406,6 +464,7 @@ class Program_weight_tensor_data_57: class Program_weight_tensor_data_58: name = "data_58" + original_name = "param_61" shape = [1] dtype = "float32" data = [-0.444284] @@ -413,6 +472,7 @@ class Program_weight_tensor_data_58: class Program_weight_tensor_data_59: name = "data_59" + original_name = "param_62" shape = [33, 768] dtype = "float32" min_val = float("-0.0865443") @@ -424,6 +484,7 @@ class Program_weight_tensor_data_59: class Program_weight_tensor_data_60: name = "data_60" + original_name = "param_7" shape = [1] dtype = "float32" data = [0.000148769] @@ -431,6 +492,7 @@ class Program_weight_tensor_data_60: class Program_weight_tensor_data_61: name = "data_61" + original_name = "param_8" shape = [1] dtype = "float32" data = [0.304639] @@ -438,6 +500,7 @@ class Program_weight_tensor_data_61: class Program_weight_tensor_data_62: name = "data_62" + original_name = "param_9" shape = [1] dtype = "float32" data = [-0.970244] @@ -445,6 +508,7 @@ class Program_weight_tensor_data_62: class Program_weight_tensor_data_63: name = "data_63" + original_name = "var_0" shape = [64, 3, 448, 448] dtype = "float32" min_val = float("-2.1179") diff --git a/paddle_samples/PaddleX/PP-HGNetV2-B0_ML/subgraph_2/weight_meta.py b/paddle_samples/PaddleX/PP-HGNetV2-B0_ML/subgraph_2/weight_meta.py index 4af8ebd30..03031c04e 100644 --- a/paddle_samples/PaddleX/PP-HGNetV2-B0_ML/subgraph_2/weight_meta.py +++ b/paddle_samples/PaddleX/PP-HGNetV2-B0_ML/subgraph_2/weight_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_parameter_0: name = "parameter_0" + original_name = "conv2d_44.b_0" shape = [33] dtype = "float32" min_val = float("0") @@ -9,6 +10,7 @@ class Program_weight_tensor_parameter_0: class Program_weight_tensor_parameter_1: name = "parameter_1" + original_name = "conv2d_44.w_0" shape = [33, 768, 1, 1] dtype = "float32" min_val = float("-0.18959") @@ -20,6 +22,7 @@ class Program_weight_tensor_parameter_1: class Program_weight_tensor_parameter_2: name = "parameter_2" + original_name = "layer_norm_2.b_0" shape = [768] dtype = "float32" min_val = float("-0.00634373") @@ -31,6 +34,7 @@ class Program_weight_tensor_parameter_2: class Program_weight_tensor_parameter_3: name = "parameter_3" + original_name = "layer_norm_2.w_0" shape = [768] dtype = "float32" min_val = float("0.942345") @@ -42,6 +46,7 @@ class Program_weight_tensor_parameter_3: class Program_weight_tensor_parameter_4: name = "parameter_4" + original_name = "linear_10.b_0" shape = [768] dtype = "float32" min_val = float("-0.0213794") @@ -53,6 +58,7 @@ class Program_weight_tensor_parameter_4: class Program_weight_tensor_parameter_5: name = "parameter_5" + original_name = "linear_10.w_0" shape = [2048, 768] dtype = "float32" min_val = float("-0.109624") @@ -64,6 +70,7 @@ class Program_weight_tensor_parameter_5: class Program_weight_tensor_parameter_6: name = "parameter_6" + original_name = "linear_9.b_0" shape = [2048] dtype = "float32" min_val = float("-0.012339") @@ -75,6 +82,7 @@ class Program_weight_tensor_parameter_6: class Program_weight_tensor_parameter_7: name = "parameter_7" + original_name = "linear_9.w_0" shape = [768, 2048] dtype = "float32" min_val = float("-0.0762291") @@ -86,6 +94,7 @@ class Program_weight_tensor_parameter_7: class Program_weight_tensor_parameter_8: name = "parameter_8" + original_name = "layer_norm_1.b_0" shape = [768] dtype = "float32" min_val = float("-0.073914") @@ -97,6 +106,7 @@ class Program_weight_tensor_parameter_8: class Program_weight_tensor_parameter_9: name = "parameter_9" + original_name = "layer_norm_1.w_0" shape = [768] dtype = "float32" min_val = float("0.981155") @@ -108,6 +118,7 @@ class Program_weight_tensor_parameter_9: class Program_weight_tensor_parameter_10: name = "parameter_10" + original_name = "linear_8.b_0" shape = [768] dtype = "float32" min_val = float("-0.0438467") @@ -119,6 +130,7 @@ class Program_weight_tensor_parameter_10: class Program_weight_tensor_parameter_11: name = "parameter_11" + original_name = "linear_8.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.112544") @@ -130,6 +142,7 @@ class Program_weight_tensor_parameter_11: class Program_weight_tensor_parameter_12: name = "parameter_12" + original_name = "linear_7.b_0" shape = [768] dtype = "float32" min_val = float("-0.0104531") @@ -141,6 +154,7 @@ class Program_weight_tensor_parameter_12: class Program_weight_tensor_parameter_13: name = "parameter_13" + original_name = "linear_7.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.0842362") @@ -152,6 +166,7 @@ class Program_weight_tensor_parameter_13: class Program_weight_tensor_parameter_14: name = "parameter_14" + original_name = "linear_6.b_0" shape = [768] dtype = "float32" min_val = float("-0.015012") @@ -163,6 +178,7 @@ class Program_weight_tensor_parameter_14: class Program_weight_tensor_parameter_15: name = "parameter_15" + original_name = "linear_6.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.0814508") @@ -174,6 +190,7 @@ class Program_weight_tensor_parameter_15: class Program_weight_tensor_parameter_16: name = "parameter_16" + original_name = "linear_5.b_0" shape = [768] dtype = "float32" min_val = float("-0.0291279") @@ -185,6 +202,7 @@ class Program_weight_tensor_parameter_16: class Program_weight_tensor_parameter_17: name = "parameter_17" + original_name = "linear_5.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.0755716") @@ -196,6 +214,7 @@ class Program_weight_tensor_parameter_17: class Program_weight_tensor_parameter_18: name = "parameter_18" + original_name = "layer_norm_0.b_0" shape = [768] dtype = "float32" min_val = float("-0.0490417") @@ -207,6 +226,7 @@ class Program_weight_tensor_parameter_18: class Program_weight_tensor_parameter_19: name = "parameter_19" + original_name = "layer_norm_0.w_0" shape = [768] dtype = "float32" min_val = float("0.974511") @@ -218,6 +238,7 @@ class Program_weight_tensor_parameter_19: class Program_weight_tensor_parameter_20: name = "parameter_20" + original_name = "conv2d_43.b_0" shape = [768] dtype = "float32" min_val = float("-0.00610541") @@ -229,6 +250,7 @@ class Program_weight_tensor_parameter_20: class Program_weight_tensor_parameter_21: name = "parameter_21" + original_name = "conv2d_43.w_0" shape = [768, 2048, 1, 1] dtype = "float32" min_val = float("-0.147691") @@ -240,6 +262,7 @@ class Program_weight_tensor_parameter_21: class Program_weight_tensor_parameter_22: name = "parameter_22" + original_name = "conv2d_42.w_0" shape = [2048, 1024, 1, 1] dtype = "float32" min_val = float("-1.13196") @@ -251,6 +274,7 @@ class Program_weight_tensor_parameter_22: class Program_weight_tensor_parameter_23: name = "parameter_23" + original_name = "batch_norm2d_41.b_0" shape = [1024] dtype = "float32" min_val = float("-13.185") @@ -262,6 +286,7 @@ class Program_weight_tensor_parameter_23: class Program_weight_tensor_parameter_24: name = "parameter_24" + original_name = "batch_norm2d_41.w_0" shape = [1024] dtype = "float32" min_val = float("0.368556") @@ -273,6 +298,7 @@ class Program_weight_tensor_parameter_24: class Program_weight_tensor_parameter_25: name = "parameter_25" + original_name = "batch_norm2d_41.w_2" shape = [1024] dtype = "float32" min_val = float("0.446079") @@ -284,6 +310,7 @@ class Program_weight_tensor_parameter_25: class Program_weight_tensor_parameter_26: name = "parameter_26" + original_name = "batch_norm2d_41.w_1" shape = [1024] dtype = "float32" min_val = float("-2.50152") @@ -295,6 +322,7 @@ class Program_weight_tensor_parameter_26: class Program_weight_tensor_parameter_27: name = "parameter_27" + original_name = "conv2d_41.w_0" shape = [1024, 512, 1, 1] dtype = "float32" min_val = float("-1.16367") @@ -306,6 +334,7 @@ class Program_weight_tensor_parameter_27: class Program_weight_tensor_parameter_28: name = "parameter_28" + original_name = "batch_norm2d_40.b_0" shape = [512] dtype = "float32" min_val = float("-10.9259") @@ -317,6 +346,7 @@ class Program_weight_tensor_parameter_28: class Program_weight_tensor_parameter_29: name = "parameter_29" + original_name = "batch_norm2d_40.w_0" shape = [512] dtype = "float32" min_val = float("1.09012") @@ -328,6 +358,7 @@ class Program_weight_tensor_parameter_29: class Program_weight_tensor_parameter_30: name = "parameter_30" + original_name = "batch_norm2d_40.w_2" shape = [512] dtype = "float32" min_val = float("5.77559") @@ -339,6 +370,7 @@ class Program_weight_tensor_parameter_30: class Program_weight_tensor_parameter_31: name = "parameter_31" + original_name = "batch_norm2d_40.w_1" shape = [512] dtype = "float32" min_val = float("-11.2684") @@ -350,6 +382,7 @@ class Program_weight_tensor_parameter_31: class Program_weight_tensor_parameter_32: name = "parameter_32" + original_name = "conv2d_40.w_0" shape = [512, 896, 1, 1] dtype = "float32" min_val = float("-0.919393") @@ -361,6 +394,7 @@ class Program_weight_tensor_parameter_32: class Program_weight_tensor_parameter_33: name = "parameter_33" + original_name = "batch_norm2d_39.b_0" shape = [128] dtype = "float32" min_val = float("-7.49733") @@ -372,6 +406,7 @@ class Program_weight_tensor_parameter_33: class Program_weight_tensor_parameter_34: name = "parameter_34" + original_name = "batch_norm2d_39.w_0" shape = [128] dtype = "float32" min_val = float("0.860746") @@ -383,6 +418,7 @@ class Program_weight_tensor_parameter_34: class Program_weight_tensor_parameter_35: name = "parameter_35" + original_name = "batch_norm2d_39.w_2" shape = [128] dtype = "float32" min_val = float("0.235889") @@ -394,6 +430,7 @@ class Program_weight_tensor_parameter_35: class Program_weight_tensor_parameter_36: name = "parameter_36" + original_name = "batch_norm2d_39.w_1" shape = [128] dtype = "float32" min_val = float("-4.4692") @@ -405,6 +442,7 @@ class Program_weight_tensor_parameter_36: class Program_weight_tensor_parameter_37: name = "parameter_37" + original_name = "conv2d_39.w_0" shape = [128, 1, 5, 5] dtype = "float32" min_val = float("-0.610755") @@ -416,6 +454,7 @@ class Program_weight_tensor_parameter_37: class Program_weight_tensor_parameter_38: name = "parameter_38" + original_name = "batch_norm2d_38.b_0" shape = [128] dtype = "float32" min_val = float("-2.95335") @@ -427,6 +466,7 @@ class Program_weight_tensor_parameter_38: class Program_weight_tensor_parameter_39: name = "parameter_39" + original_name = "batch_norm2d_38.w_0" shape = [128] dtype = "float32" min_val = float("0.726973") @@ -438,6 +478,7 @@ class Program_weight_tensor_parameter_39: class Program_weight_tensor_parameter_40: name = "parameter_40" + original_name = "batch_norm2d_38.w_2" shape = [128] dtype = "float32" min_val = float("0.889493") @@ -449,6 +490,7 @@ class Program_weight_tensor_parameter_40: class Program_weight_tensor_parameter_41: name = "parameter_41" + original_name = "batch_norm2d_38.w_1" shape = [128] dtype = "float32" min_val = float("-7.73106") @@ -460,6 +502,7 @@ class Program_weight_tensor_parameter_41: class Program_weight_tensor_parameter_42: name = "parameter_42" + original_name = "conv2d_38.w_0" shape = [128, 128, 1, 1] dtype = "float32" min_val = float("-0.700852") @@ -471,6 +514,7 @@ class Program_weight_tensor_parameter_42: class Program_weight_tensor_parameter_43: name = "parameter_43" + original_name = "batch_norm2d_37.b_0" shape = [128] dtype = "float32" min_val = float("-6.70058") @@ -482,6 +526,7 @@ class Program_weight_tensor_parameter_43: class Program_weight_tensor_parameter_44: name = "parameter_44" + original_name = "batch_norm2d_37.w_0" shape = [128] dtype = "float32" min_val = float("1.04038") @@ -493,6 +538,7 @@ class Program_weight_tensor_parameter_44: class Program_weight_tensor_parameter_45: name = "parameter_45" + original_name = "batch_norm2d_37.w_2" shape = [128] dtype = "float32" min_val = float("0.113784") @@ -504,6 +550,7 @@ class Program_weight_tensor_parameter_45: class Program_weight_tensor_parameter_46: name = "parameter_46" + original_name = "batch_norm2d_37.w_1" shape = [128] dtype = "float32" min_val = float("-2.74669") @@ -515,6 +562,7 @@ class Program_weight_tensor_parameter_46: class Program_weight_tensor_parameter_47: name = "parameter_47" + original_name = "conv2d_37.w_0" shape = [128, 1, 5, 5] dtype = "float32" min_val = float("-0.8572") @@ -526,6 +574,7 @@ class Program_weight_tensor_parameter_47: class Program_weight_tensor_parameter_48: name = "parameter_48" + original_name = "batch_norm2d_36.b_0" shape = [128] dtype = "float32" min_val = float("-1.90555") @@ -537,6 +586,7 @@ class Program_weight_tensor_parameter_48: class Program_weight_tensor_parameter_49: name = "parameter_49" + original_name = "batch_norm2d_36.w_0" shape = [128] dtype = "float32" min_val = float("0.47071") @@ -548,6 +598,7 @@ class Program_weight_tensor_parameter_49: class Program_weight_tensor_parameter_50: name = "parameter_50" + original_name = "batch_norm2d_36.w_2" shape = [128] dtype = "float32" min_val = float("1.17201") @@ -559,6 +610,7 @@ class Program_weight_tensor_parameter_50: class Program_weight_tensor_parameter_51: name = "parameter_51" + original_name = "batch_norm2d_36.w_1" shape = [128] dtype = "float32" min_val = float("-8.44452") @@ -570,6 +622,7 @@ class Program_weight_tensor_parameter_51: class Program_weight_tensor_parameter_52: name = "parameter_52" + original_name = "conv2d_36.w_0" shape = [128, 128, 1, 1] dtype = "float32" min_val = float("-0.908032") @@ -581,6 +634,7 @@ class Program_weight_tensor_parameter_52: class Program_weight_tensor_parameter_53: name = "parameter_53" + original_name = "batch_norm2d_35.b_0" shape = [128] dtype = "float32" min_val = float("-8.19392") @@ -592,6 +646,7 @@ class Program_weight_tensor_parameter_53: class Program_weight_tensor_parameter_54: name = "parameter_54" + original_name = "batch_norm2d_35.w_0" shape = [128] dtype = "float32" min_val = float("1.20021") @@ -603,6 +658,7 @@ class Program_weight_tensor_parameter_54: class Program_weight_tensor_parameter_55: name = "parameter_55" + original_name = "batch_norm2d_35.w_2" shape = [128] dtype = "float32" min_val = float("0.357326") @@ -614,6 +670,7 @@ class Program_weight_tensor_parameter_55: class Program_weight_tensor_parameter_56: name = "parameter_56" + original_name = "batch_norm2d_35.w_1" shape = [128] dtype = "float32" min_val = float("-1.88256") @@ -625,6 +682,7 @@ class Program_weight_tensor_parameter_56: class Program_weight_tensor_parameter_57: name = "parameter_57" + original_name = "conv2d_35.w_0" shape = [128, 1, 5, 5] dtype = "float32" min_val = float("-0.717986") @@ -636,6 +694,7 @@ class Program_weight_tensor_parameter_57: class Program_weight_tensor_parameter_58: name = "parameter_58" + original_name = "batch_norm2d_34.b_0" shape = [128] dtype = "float32" min_val = float("-2.19606") @@ -647,6 +706,7 @@ class Program_weight_tensor_parameter_58: class Program_weight_tensor_parameter_59: name = "parameter_59" + original_name = "batch_norm2d_34.w_0" shape = [128] dtype = "float32" min_val = float("0.972198") @@ -658,6 +718,7 @@ class Program_weight_tensor_parameter_59: class Program_weight_tensor_parameter_60: name = "parameter_60" + original_name = "batch_norm2d_34.w_2" shape = [128] dtype = "float32" min_val = float("4.47207") @@ -669,6 +730,7 @@ class Program_weight_tensor_parameter_60: class Program_weight_tensor_parameter_61: name = "parameter_61" + original_name = "batch_norm2d_34.w_1" shape = [128] dtype = "float32" min_val = float("-0.0152718") @@ -680,6 +742,7 @@ class Program_weight_tensor_parameter_61: class Program_weight_tensor_parameter_62: name = "parameter_62" + original_name = "conv2d_34.w_0" shape = [128, 512, 1, 1] dtype = "float32" min_val = float("-0.634672") @@ -691,6 +754,7 @@ class Program_weight_tensor_parameter_62: class Program_weight_tensor_parameter_63: name = "parameter_63" + original_name = "batch_norm2d_33.b_0" shape = [512] dtype = "float32" min_val = float("-0.0106532") @@ -702,6 +766,7 @@ class Program_weight_tensor_parameter_63: class Program_weight_tensor_parameter_64: name = "parameter_64" + original_name = "batch_norm2d_33.w_0" shape = [512] dtype = "float32" min_val = float("1.34923") @@ -713,6 +778,7 @@ class Program_weight_tensor_parameter_64: class Program_weight_tensor_parameter_65: name = "parameter_65" + original_name = "batch_norm2d_33.w_2" shape = [512] dtype = "float32" min_val = float("0.00769316") @@ -724,6 +790,7 @@ class Program_weight_tensor_parameter_65: class Program_weight_tensor_parameter_66: name = "parameter_66" + original_name = "batch_norm2d_33.w_1" shape = [512] dtype = "float32" min_val = float("-3.96765") @@ -735,6 +802,7 @@ class Program_weight_tensor_parameter_66: class Program_weight_tensor_parameter_67: name = "parameter_67" + original_name = "conv2d_33.w_0" shape = [512, 1, 3, 3] dtype = "float32" min_val = float("-0.357953") @@ -746,6 +814,7 @@ class Program_weight_tensor_parameter_67: class Program_weight_tensor_parameter_68: name = "parameter_68" + original_name = "batch_norm2d_32.b_0" shape = [512] dtype = "float32" min_val = float("-7.78069") @@ -757,6 +826,7 @@ class Program_weight_tensor_parameter_68: class Program_weight_tensor_parameter_69: name = "parameter_69" + original_name = "batch_norm2d_32.w_0" shape = [512] dtype = "float32" min_val = float("0.173312") @@ -768,6 +838,7 @@ class Program_weight_tensor_parameter_69: class Program_weight_tensor_parameter_70: name = "parameter_70" + original_name = "batch_norm2d_32.w_2" shape = [512] dtype = "float32" min_val = float("0.262339") @@ -779,6 +850,7 @@ class Program_weight_tensor_parameter_70: class Program_weight_tensor_parameter_71: name = "parameter_71" + original_name = "batch_norm2d_32.w_1" shape = [512] dtype = "float32" min_val = float("-2.38926") @@ -790,6 +862,7 @@ class Program_weight_tensor_parameter_71: class Program_weight_tensor_parameter_72: name = "parameter_72" + original_name = "conv2d_32.w_0" shape = [512, 256, 1, 1] dtype = "float32" min_val = float("-0.602725") @@ -801,6 +874,7 @@ class Program_weight_tensor_parameter_72: class Program_weight_tensor_parameter_73: name = "parameter_73" + original_name = "batch_norm2d_31.b_0" shape = [256] dtype = "float32" min_val = float("-9.64223") @@ -812,6 +886,7 @@ class Program_weight_tensor_parameter_73: class Program_weight_tensor_parameter_74: name = "parameter_74" + original_name = "batch_norm2d_31.w_0" shape = [256] dtype = "float32" min_val = float("1.22837") @@ -823,6 +898,7 @@ class Program_weight_tensor_parameter_74: class Program_weight_tensor_parameter_75: name = "parameter_75" + original_name = "batch_norm2d_31.w_2" shape = [256] dtype = "float32" min_val = float("0.157536") @@ -834,6 +910,7 @@ class Program_weight_tensor_parameter_75: class Program_weight_tensor_parameter_76: name = "parameter_76" + original_name = "batch_norm2d_31.w_1" shape = [256] dtype = "float32" min_val = float("-1.99012") @@ -845,6 +922,7 @@ class Program_weight_tensor_parameter_76: class Program_weight_tensor_parameter_77: name = "parameter_77" + original_name = "conv2d_31.w_0" shape = [256, 704, 1, 1] dtype = "float32" min_val = float("-0.852449") @@ -856,6 +934,7 @@ class Program_weight_tensor_parameter_77: class Program_weight_tensor_parameter_78: name = "parameter_78" + original_name = "batch_norm2d_30.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -865,6 +944,7 @@ class Program_weight_tensor_parameter_78: class Program_weight_tensor_parameter_79: name = "parameter_79" + original_name = "batch_norm2d_30.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -874,6 +954,7 @@ class Program_weight_tensor_parameter_79: class Program_weight_tensor_parameter_80: name = "parameter_80" + original_name = "batch_norm2d_30.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -883,6 +964,7 @@ class Program_weight_tensor_parameter_80: class Program_weight_tensor_parameter_81: name = "parameter_81" + original_name = "batch_norm2d_30.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -892,6 +974,7 @@ class Program_weight_tensor_parameter_81: class Program_weight_tensor_parameter_82: name = "parameter_82" + original_name = "conv2d_30.w_0" shape = [64, 1, 5, 5] dtype = "float32" min_val = float("-0.646274") @@ -903,6 +986,7 @@ class Program_weight_tensor_parameter_82: class Program_weight_tensor_parameter_83: name = "parameter_83" + original_name = "batch_norm2d_29.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -912,6 +996,7 @@ class Program_weight_tensor_parameter_83: class Program_weight_tensor_parameter_84: name = "parameter_84" + original_name = "batch_norm2d_29.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -921,6 +1006,7 @@ class Program_weight_tensor_parameter_84: class Program_weight_tensor_parameter_85: name = "parameter_85" + original_name = "batch_norm2d_29.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -930,6 +1016,7 @@ class Program_weight_tensor_parameter_85: class Program_weight_tensor_parameter_86: name = "parameter_86" + original_name = "batch_norm2d_29.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -939,6 +1026,7 @@ class Program_weight_tensor_parameter_86: class Program_weight_tensor_parameter_87: name = "parameter_87" + original_name = "conv2d_29.w_0" shape = [64, 64, 1, 1] dtype = "float32" min_val = float("-0.581811") @@ -950,6 +1038,7 @@ class Program_weight_tensor_parameter_87: class Program_weight_tensor_parameter_88: name = "parameter_88" + original_name = "batch_norm2d_28.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -959,6 +1048,7 @@ class Program_weight_tensor_parameter_88: class Program_weight_tensor_parameter_89: name = "parameter_89" + original_name = "batch_norm2d_28.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -968,6 +1058,7 @@ class Program_weight_tensor_parameter_89: class Program_weight_tensor_parameter_90: name = "parameter_90" + original_name = "batch_norm2d_28.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -977,6 +1068,7 @@ class Program_weight_tensor_parameter_90: class Program_weight_tensor_parameter_91: name = "parameter_91" + original_name = "batch_norm2d_28.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -986,6 +1078,7 @@ class Program_weight_tensor_parameter_91: class Program_weight_tensor_parameter_92: name = "parameter_92" + original_name = "conv2d_28.w_0" shape = [64, 1, 5, 5] dtype = "float32" min_val = float("-0.674616") @@ -997,6 +1090,7 @@ class Program_weight_tensor_parameter_92: class Program_weight_tensor_parameter_93: name = "parameter_93" + original_name = "batch_norm2d_27.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -1006,6 +1100,7 @@ class Program_weight_tensor_parameter_93: class Program_weight_tensor_parameter_94: name = "parameter_94" + original_name = "batch_norm2d_27.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -1015,6 +1110,7 @@ class Program_weight_tensor_parameter_94: class Program_weight_tensor_parameter_95: name = "parameter_95" + original_name = "batch_norm2d_27.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -1024,6 +1120,7 @@ class Program_weight_tensor_parameter_95: class Program_weight_tensor_parameter_96: name = "parameter_96" + original_name = "batch_norm2d_27.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -1033,6 +1130,7 @@ class Program_weight_tensor_parameter_96: class Program_weight_tensor_parameter_97: name = "parameter_97" + original_name = "conv2d_27.w_0" shape = [64, 64, 1, 1] dtype = "float32" min_val = float("-0.753959") @@ -1044,6 +1142,7 @@ class Program_weight_tensor_parameter_97: class Program_weight_tensor_parameter_98: name = "parameter_98" + original_name = "batch_norm2d_26.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -1053,6 +1152,7 @@ class Program_weight_tensor_parameter_98: class Program_weight_tensor_parameter_99: name = "parameter_99" + original_name = "batch_norm2d_26.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -1062,6 +1162,7 @@ class Program_weight_tensor_parameter_99: class Program_weight_tensor_parameter_100: name = "parameter_100" + original_name = "batch_norm2d_26.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -1071,6 +1172,7 @@ class Program_weight_tensor_parameter_100: class Program_weight_tensor_parameter_101: name = "parameter_101" + original_name = "batch_norm2d_26.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -1080,6 +1182,7 @@ class Program_weight_tensor_parameter_101: class Program_weight_tensor_parameter_102: name = "parameter_102" + original_name = "conv2d_26.w_0" shape = [64, 1, 5, 5] dtype = "float32" min_val = float("-1.11574") @@ -1091,6 +1194,7 @@ class Program_weight_tensor_parameter_102: class Program_weight_tensor_parameter_103: name = "parameter_103" + original_name = "batch_norm2d_25.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -1100,6 +1204,7 @@ class Program_weight_tensor_parameter_103: class Program_weight_tensor_parameter_104: name = "parameter_104" + original_name = "batch_norm2d_25.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -1109,6 +1214,7 @@ class Program_weight_tensor_parameter_104: class Program_weight_tensor_parameter_105: name = "parameter_105" + original_name = "batch_norm2d_25.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -1118,6 +1224,7 @@ class Program_weight_tensor_parameter_105: class Program_weight_tensor_parameter_106: name = "parameter_106" + original_name = "batch_norm2d_25.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -1127,6 +1234,7 @@ class Program_weight_tensor_parameter_106: class Program_weight_tensor_parameter_107: name = "parameter_107" + original_name = "conv2d_25.w_0" shape = [64, 512, 1, 1] dtype = "float32" min_val = float("-1.07302") @@ -1138,6 +1246,7 @@ class Program_weight_tensor_parameter_107: class Program_weight_tensor_parameter_108: name = "parameter_108" + original_name = "batch_norm2d_24.b_0" shape = [512] dtype = "float32" min_val = float("-11.7347") @@ -1149,6 +1258,7 @@ class Program_weight_tensor_parameter_108: class Program_weight_tensor_parameter_109: name = "parameter_109" + original_name = "batch_norm2d_24.w_0" shape = [512] dtype = "float32" min_val = float("0.609352") @@ -1160,6 +1270,7 @@ class Program_weight_tensor_parameter_109: class Program_weight_tensor_parameter_110: name = "parameter_110" + original_name = "batch_norm2d_24.w_2" shape = [512] dtype = "float32" min_val = float("0.585545") @@ -1171,6 +1282,7 @@ class Program_weight_tensor_parameter_110: class Program_weight_tensor_parameter_111: name = "parameter_111" + original_name = "batch_norm2d_24.w_1" shape = [512] dtype = "float32" min_val = float("-5.33348") @@ -1182,6 +1294,7 @@ class Program_weight_tensor_parameter_111: class Program_weight_tensor_parameter_112: name = "parameter_112" + original_name = "conv2d_24.w_0" shape = [512, 256, 1, 1] dtype = "float32" min_val = float("-1.40269") @@ -1193,6 +1306,7 @@ class Program_weight_tensor_parameter_112: class Program_weight_tensor_parameter_113: name = "parameter_113" + original_name = "batch_norm2d_23.b_0" shape = [256] dtype = "float32" min_val = float("-16.202") @@ -1204,6 +1318,7 @@ class Program_weight_tensor_parameter_113: class Program_weight_tensor_parameter_114: name = "parameter_114" + original_name = "batch_norm2d_23.w_0" shape = [256] dtype = "float32" min_val = float("1.43543") @@ -1215,6 +1330,7 @@ class Program_weight_tensor_parameter_114: class Program_weight_tensor_parameter_115: name = "parameter_115" + original_name = "batch_norm2d_23.w_2" shape = [256] dtype = "float32" min_val = float("6.98125") @@ -1226,6 +1342,7 @@ class Program_weight_tensor_parameter_115: class Program_weight_tensor_parameter_116: name = "parameter_116" + original_name = "batch_norm2d_23.w_1" shape = [256] dtype = "float32" min_val = float("-16.3757") @@ -1237,6 +1354,7 @@ class Program_weight_tensor_parameter_116: class Program_weight_tensor_parameter_117: name = "parameter_117" + original_name = "conv2d_23.w_0" shape = [256, 448, 1, 1] dtype = "float32" min_val = float("-1.12619") @@ -1248,6 +1366,7 @@ class Program_weight_tensor_parameter_117: class Program_weight_tensor_parameter_118: name = "parameter_118" + original_name = "batch_norm2d_22.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -1257,6 +1376,7 @@ class Program_weight_tensor_parameter_118: class Program_weight_tensor_parameter_119: name = "parameter_119" + original_name = "batch_norm2d_22.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -1266,6 +1386,7 @@ class Program_weight_tensor_parameter_119: class Program_weight_tensor_parameter_120: name = "parameter_120" + original_name = "batch_norm2d_22.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -1275,6 +1396,7 @@ class Program_weight_tensor_parameter_120: class Program_weight_tensor_parameter_121: name = "parameter_121" + original_name = "batch_norm2d_22.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -1284,6 +1406,7 @@ class Program_weight_tensor_parameter_121: class Program_weight_tensor_parameter_122: name = "parameter_122" + original_name = "conv2d_22.w_0" shape = [64, 1, 5, 5] dtype = "float32" min_val = float("-0.472954") @@ -1295,6 +1418,7 @@ class Program_weight_tensor_parameter_122: class Program_weight_tensor_parameter_123: name = "parameter_123" + original_name = "batch_norm2d_21.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -1304,6 +1428,7 @@ class Program_weight_tensor_parameter_123: class Program_weight_tensor_parameter_124: name = "parameter_124" + original_name = "batch_norm2d_21.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -1313,6 +1438,7 @@ class Program_weight_tensor_parameter_124: class Program_weight_tensor_parameter_125: name = "parameter_125" + original_name = "batch_norm2d_21.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -1322,6 +1448,7 @@ class Program_weight_tensor_parameter_125: class Program_weight_tensor_parameter_126: name = "parameter_126" + original_name = "batch_norm2d_21.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -1331,6 +1458,7 @@ class Program_weight_tensor_parameter_126: class Program_weight_tensor_parameter_127: name = "parameter_127" + original_name = "conv2d_21.w_0" shape = [64, 64, 1, 1] dtype = "float32" min_val = float("-0.880354") @@ -1342,6 +1470,7 @@ class Program_weight_tensor_parameter_127: class Program_weight_tensor_parameter_128: name = "parameter_128" + original_name = "batch_norm2d_20.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -1351,6 +1480,7 @@ class Program_weight_tensor_parameter_128: class Program_weight_tensor_parameter_129: name = "parameter_129" + original_name = "batch_norm2d_20.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -1360,6 +1490,7 @@ class Program_weight_tensor_parameter_129: class Program_weight_tensor_parameter_130: name = "parameter_130" + original_name = "batch_norm2d_20.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -1369,6 +1500,7 @@ class Program_weight_tensor_parameter_130: class Program_weight_tensor_parameter_131: name = "parameter_131" + original_name = "batch_norm2d_20.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -1378,6 +1510,7 @@ class Program_weight_tensor_parameter_131: class Program_weight_tensor_parameter_132: name = "parameter_132" + original_name = "conv2d_20.w_0" shape = [64, 1, 5, 5] dtype = "float32" min_val = float("-1.35858") @@ -1389,6 +1522,7 @@ class Program_weight_tensor_parameter_132: class Program_weight_tensor_parameter_133: name = "parameter_133" + original_name = "batch_norm2d_19.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -1398,6 +1532,7 @@ class Program_weight_tensor_parameter_133: class Program_weight_tensor_parameter_134: name = "parameter_134" + original_name = "batch_norm2d_19.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -1407,6 +1542,7 @@ class Program_weight_tensor_parameter_134: class Program_weight_tensor_parameter_135: name = "parameter_135" + original_name = "batch_norm2d_19.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -1416,6 +1552,7 @@ class Program_weight_tensor_parameter_135: class Program_weight_tensor_parameter_136: name = "parameter_136" + original_name = "batch_norm2d_19.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -1425,6 +1562,7 @@ class Program_weight_tensor_parameter_136: class Program_weight_tensor_parameter_137: name = "parameter_137" + original_name = "conv2d_19.w_0" shape = [64, 64, 1, 1] dtype = "float32" min_val = float("-1.34413") @@ -1436,6 +1574,7 @@ class Program_weight_tensor_parameter_137: class Program_weight_tensor_parameter_138: name = "parameter_138" + original_name = "batch_norm2d_18.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -1445,6 +1584,7 @@ class Program_weight_tensor_parameter_138: class Program_weight_tensor_parameter_139: name = "parameter_139" + original_name = "batch_norm2d_18.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -1454,6 +1594,7 @@ class Program_weight_tensor_parameter_139: class Program_weight_tensor_parameter_140: name = "parameter_140" + original_name = "batch_norm2d_18.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -1463,6 +1604,7 @@ class Program_weight_tensor_parameter_140: class Program_weight_tensor_parameter_141: name = "parameter_141" + original_name = "batch_norm2d_18.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -1472,6 +1614,7 @@ class Program_weight_tensor_parameter_141: class Program_weight_tensor_parameter_142: name = "parameter_142" + original_name = "conv2d_18.w_0" shape = [64, 1, 5, 5] dtype = "float32" min_val = float("-1.22412") @@ -1483,6 +1626,7 @@ class Program_weight_tensor_parameter_142: class Program_weight_tensor_parameter_143: name = "parameter_143" + original_name = "batch_norm2d_17.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -1492,6 +1636,7 @@ class Program_weight_tensor_parameter_143: class Program_weight_tensor_parameter_144: name = "parameter_144" + original_name = "batch_norm2d_17.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -1501,6 +1646,7 @@ class Program_weight_tensor_parameter_144: class Program_weight_tensor_parameter_145: name = "parameter_145" + original_name = "batch_norm2d_17.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -1510,6 +1656,7 @@ class Program_weight_tensor_parameter_145: class Program_weight_tensor_parameter_146: name = "parameter_146" + original_name = "batch_norm2d_17.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -1519,6 +1666,7 @@ class Program_weight_tensor_parameter_146: class Program_weight_tensor_parameter_147: name = "parameter_147" + original_name = "conv2d_17.w_0" shape = [64, 256, 1, 1] dtype = "float32" min_val = float("-1.09013") @@ -1530,6 +1678,7 @@ class Program_weight_tensor_parameter_147: class Program_weight_tensor_parameter_148: name = "parameter_148" + original_name = "batch_norm2d_16.b_0" shape = [256] dtype = "float32" min_val = float("-0.0122304") @@ -1541,6 +1690,7 @@ class Program_weight_tensor_parameter_148: class Program_weight_tensor_parameter_149: name = "parameter_149" + original_name = "batch_norm2d_16.w_0" shape = [256] dtype = "float32" min_val = float("1.75814") @@ -1552,6 +1702,7 @@ class Program_weight_tensor_parameter_149: class Program_weight_tensor_parameter_150: name = "parameter_150" + original_name = "batch_norm2d_16.w_2" shape = [256] dtype = "float32" min_val = float("0.000270616") @@ -1563,6 +1714,7 @@ class Program_weight_tensor_parameter_150: class Program_weight_tensor_parameter_151: name = "parameter_151" + original_name = "batch_norm2d_16.w_1" shape = [256] dtype = "float32" min_val = float("-2.14963") @@ -1574,6 +1726,7 @@ class Program_weight_tensor_parameter_151: class Program_weight_tensor_parameter_152: name = "parameter_152" + original_name = "conv2d_16.w_0" shape = [256, 1, 3, 3] dtype = "float32" min_val = float("-0.35146") @@ -1585,6 +1738,7 @@ class Program_weight_tensor_parameter_152: class Program_weight_tensor_parameter_153: name = "parameter_153" + original_name = "batch_norm2d_15.b_0" shape = [256] dtype = "float32" min_val = float("-4.03611") @@ -1596,6 +1750,7 @@ class Program_weight_tensor_parameter_153: class Program_weight_tensor_parameter_154: name = "parameter_154" + original_name = "batch_norm2d_15.w_0" shape = [256] dtype = "float32" min_val = float("0.0496178") @@ -1607,6 +1762,7 @@ class Program_weight_tensor_parameter_154: class Program_weight_tensor_parameter_155: name = "parameter_155" + original_name = "batch_norm2d_15.w_2" shape = [256] dtype = "float32" min_val = float("1.13721") @@ -1618,6 +1774,7 @@ class Program_weight_tensor_parameter_155: class Program_weight_tensor_parameter_156: name = "parameter_156" + original_name = "batch_norm2d_15.w_1" shape = [256] dtype = "float32" min_val = float("-7.52889") @@ -1629,6 +1786,7 @@ class Program_weight_tensor_parameter_156: class Program_weight_tensor_parameter_157: name = "parameter_157" + original_name = "conv2d_15.w_0" shape = [256, 128, 1, 1] dtype = "float32" min_val = float("-1.16416") @@ -1640,6 +1798,7 @@ class Program_weight_tensor_parameter_157: class Program_weight_tensor_parameter_158: name = "parameter_158" + original_name = "batch_norm2d_14.b_0" shape = [128] dtype = "float32" min_val = float("-5.04793") @@ -1651,6 +1810,7 @@ class Program_weight_tensor_parameter_158: class Program_weight_tensor_parameter_159: name = "parameter_159" + original_name = "batch_norm2d_14.w_0" shape = [128] dtype = "float32" min_val = float("1.77336") @@ -1662,6 +1822,7 @@ class Program_weight_tensor_parameter_159: class Program_weight_tensor_parameter_160: name = "parameter_160" + original_name = "batch_norm2d_14.w_2" shape = [128] dtype = "float32" min_val = float("9.50644") @@ -1673,6 +1834,7 @@ class Program_weight_tensor_parameter_160: class Program_weight_tensor_parameter_161: name = "parameter_161" + original_name = "batch_norm2d_14.w_1" shape = [128] dtype = "float32" min_val = float("-12.1181") @@ -1684,6 +1846,7 @@ class Program_weight_tensor_parameter_161: class Program_weight_tensor_parameter_162: name = "parameter_162" + original_name = "conv2d_14.w_0" shape = [128, 160, 1, 1] dtype = "float32" min_val = float("-1.48061") @@ -1695,6 +1858,7 @@ class Program_weight_tensor_parameter_162: class Program_weight_tensor_parameter_163: name = "parameter_163" + original_name = "batch_norm2d_13.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -1704,6 +1868,7 @@ class Program_weight_tensor_parameter_163: class Program_weight_tensor_parameter_164: name = "parameter_164" + original_name = "batch_norm2d_13.w_0" shape = [32] dtype = "float32" min_val = float("0") @@ -1713,6 +1878,7 @@ class Program_weight_tensor_parameter_164: class Program_weight_tensor_parameter_165: name = "parameter_165" + original_name = "batch_norm2d_13.w_2" shape = [32] dtype = "float32" min_val = float("0") @@ -1722,6 +1888,7 @@ class Program_weight_tensor_parameter_165: class Program_weight_tensor_parameter_166: name = "parameter_166" + original_name = "batch_norm2d_13.w_1" shape = [32] dtype = "float32" min_val = float("0") @@ -1731,6 +1898,7 @@ class Program_weight_tensor_parameter_166: class Program_weight_tensor_parameter_167: name = "parameter_167" + original_name = "conv2d_13.w_0" shape = [32, 32, 3, 3] dtype = "float32" min_val = float("-0.661935") @@ -1742,6 +1910,7 @@ class Program_weight_tensor_parameter_167: class Program_weight_tensor_parameter_168: name = "parameter_168" + original_name = "batch_norm2d_12.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -1751,6 +1920,7 @@ class Program_weight_tensor_parameter_168: class Program_weight_tensor_parameter_169: name = "parameter_169" + original_name = "batch_norm2d_12.w_0" shape = [32] dtype = "float32" min_val = float("0") @@ -1760,6 +1930,7 @@ class Program_weight_tensor_parameter_169: class Program_weight_tensor_parameter_170: name = "parameter_170" + original_name = "batch_norm2d_12.w_2" shape = [32] dtype = "float32" min_val = float("0") @@ -1769,6 +1940,7 @@ class Program_weight_tensor_parameter_170: class Program_weight_tensor_parameter_171: name = "parameter_171" + original_name = "batch_norm2d_12.w_1" shape = [32] dtype = "float32" min_val = float("0") @@ -1778,6 +1950,7 @@ class Program_weight_tensor_parameter_171: class Program_weight_tensor_parameter_172: name = "parameter_172" + original_name = "conv2d_12.w_0" shape = [32, 32, 3, 3] dtype = "float32" min_val = float("-0.979117") @@ -1789,6 +1962,7 @@ class Program_weight_tensor_parameter_172: class Program_weight_tensor_parameter_173: name = "parameter_173" + original_name = "batch_norm2d_11.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -1798,6 +1972,7 @@ class Program_weight_tensor_parameter_173: class Program_weight_tensor_parameter_174: name = "parameter_174" + original_name = "batch_norm2d_11.w_0" shape = [32] dtype = "float32" min_val = float("0") @@ -1807,6 +1982,7 @@ class Program_weight_tensor_parameter_174: class Program_weight_tensor_parameter_175: name = "parameter_175" + original_name = "batch_norm2d_11.w_2" shape = [32] dtype = "float32" min_val = float("0") @@ -1816,6 +1992,7 @@ class Program_weight_tensor_parameter_175: class Program_weight_tensor_parameter_176: name = "parameter_176" + original_name = "batch_norm2d_11.w_1" shape = [32] dtype = "float32" min_val = float("0") @@ -1825,6 +2002,7 @@ class Program_weight_tensor_parameter_176: class Program_weight_tensor_parameter_177: name = "parameter_177" + original_name = "conv2d_11.w_0" shape = [32, 64, 3, 3] dtype = "float32" min_val = float("-1.5457") @@ -1836,6 +2014,7 @@ class Program_weight_tensor_parameter_177: class Program_weight_tensor_parameter_178: name = "parameter_178" + original_name = "batch_norm2d_10.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -1845,6 +2024,7 @@ class Program_weight_tensor_parameter_178: class Program_weight_tensor_parameter_179: name = "parameter_179" + original_name = "batch_norm2d_10.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -1854,6 +2034,7 @@ class Program_weight_tensor_parameter_179: class Program_weight_tensor_parameter_180: name = "parameter_180" + original_name = "batch_norm2d_10.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -1863,6 +2044,7 @@ class Program_weight_tensor_parameter_180: class Program_weight_tensor_parameter_181: name = "parameter_181" + original_name = "batch_norm2d_10.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -1872,6 +2054,7 @@ class Program_weight_tensor_parameter_181: class Program_weight_tensor_parameter_182: name = "parameter_182" + original_name = "conv2d_10.w_0" shape = [64, 1, 3, 3] dtype = "float32" min_val = float("-0.714508") @@ -1883,6 +2066,7 @@ class Program_weight_tensor_parameter_182: class Program_weight_tensor_parameter_183: name = "parameter_183" + original_name = "batch_norm2d_9.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -1892,6 +2076,7 @@ class Program_weight_tensor_parameter_183: class Program_weight_tensor_parameter_184: name = "parameter_184" + original_name = "batch_norm2d_9.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -1901,6 +2086,7 @@ class Program_weight_tensor_parameter_184: class Program_weight_tensor_parameter_185: name = "parameter_185" + original_name = "batch_norm2d_9.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -1910,6 +2096,7 @@ class Program_weight_tensor_parameter_185: class Program_weight_tensor_parameter_186: name = "parameter_186" + original_name = "batch_norm2d_9.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -1919,6 +2106,7 @@ class Program_weight_tensor_parameter_186: class Program_weight_tensor_parameter_187: name = "parameter_187" + original_name = "conv2d_9.w_0" shape = [64, 32, 1, 1] dtype = "float32" min_val = float("-1.96281") @@ -1930,6 +2118,7 @@ class Program_weight_tensor_parameter_187: class Program_weight_tensor_parameter_188: name = "parameter_188" + original_name = "batch_norm2d_8.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -1939,6 +2128,7 @@ class Program_weight_tensor_parameter_188: class Program_weight_tensor_parameter_189: name = "parameter_189" + original_name = "batch_norm2d_8.w_0" shape = [32] dtype = "float32" min_val = float("0") @@ -1948,6 +2138,7 @@ class Program_weight_tensor_parameter_189: class Program_weight_tensor_parameter_190: name = "parameter_190" + original_name = "batch_norm2d_8.w_2" shape = [32] dtype = "float32" min_val = float("0") @@ -1957,6 +2148,7 @@ class Program_weight_tensor_parameter_190: class Program_weight_tensor_parameter_191: name = "parameter_191" + original_name = "batch_norm2d_8.w_1" shape = [32] dtype = "float32" min_val = float("0") @@ -1966,6 +2158,7 @@ class Program_weight_tensor_parameter_191: class Program_weight_tensor_parameter_192: name = "parameter_192" + original_name = "conv2d_8.w_0" shape = [32, 64, 1, 1] dtype = "float32" min_val = float("-2.21205") @@ -1977,6 +2170,7 @@ class Program_weight_tensor_parameter_192: class Program_weight_tensor_parameter_193: name = "parameter_193" + original_name = "batch_norm2d_7.b_0" shape = [16] dtype = "float32" min_val = float("0") @@ -1986,6 +2180,7 @@ class Program_weight_tensor_parameter_193: class Program_weight_tensor_parameter_194: name = "parameter_194" + original_name = "batch_norm2d_7.w_0" shape = [16] dtype = "float32" min_val = float("0") @@ -1995,6 +2190,7 @@ class Program_weight_tensor_parameter_194: class Program_weight_tensor_parameter_195: name = "parameter_195" + original_name = "batch_norm2d_7.w_2" shape = [16] dtype = "float32" min_val = float("0") @@ -2004,6 +2200,7 @@ class Program_weight_tensor_parameter_195: class Program_weight_tensor_parameter_196: name = "parameter_196" + original_name = "batch_norm2d_7.w_1" shape = [16] dtype = "float32" min_val = float("0") @@ -2013,6 +2210,7 @@ class Program_weight_tensor_parameter_196: class Program_weight_tensor_parameter_197: name = "parameter_197" + original_name = "conv2d_7.w_0" shape = [16, 16, 3, 3] dtype = "float32" min_val = float("-1.13765") @@ -2024,6 +2222,7 @@ class Program_weight_tensor_parameter_197: class Program_weight_tensor_parameter_198: name = "parameter_198" + original_name = "batch_norm2d_6.b_0" shape = [16] dtype = "float32" min_val = float("0") @@ -2033,6 +2232,7 @@ class Program_weight_tensor_parameter_198: class Program_weight_tensor_parameter_199: name = "parameter_199" + original_name = "batch_norm2d_6.w_0" shape = [16] dtype = "float32" min_val = float("0") @@ -2042,6 +2242,7 @@ class Program_weight_tensor_parameter_199: class Program_weight_tensor_parameter_200: name = "parameter_200" + original_name = "batch_norm2d_6.w_2" shape = [16] dtype = "float32" min_val = float("0") @@ -2051,6 +2252,7 @@ class Program_weight_tensor_parameter_200: class Program_weight_tensor_parameter_201: name = "parameter_201" + original_name = "batch_norm2d_6.w_1" shape = [16] dtype = "float32" min_val = float("0") @@ -2060,6 +2262,7 @@ class Program_weight_tensor_parameter_201: class Program_weight_tensor_parameter_202: name = "parameter_202" + original_name = "conv2d_6.w_0" shape = [16, 16, 3, 3] dtype = "float32" min_val = float("-1.10058") @@ -2071,6 +2274,7 @@ class Program_weight_tensor_parameter_202: class Program_weight_tensor_parameter_203: name = "parameter_203" + original_name = "batch_norm2d_5.b_0" shape = [16] dtype = "float32" min_val = float("0") @@ -2080,6 +2284,7 @@ class Program_weight_tensor_parameter_203: class Program_weight_tensor_parameter_204: name = "parameter_204" + original_name = "batch_norm2d_5.w_0" shape = [16] dtype = "float32" min_val = float("0") @@ -2089,6 +2294,7 @@ class Program_weight_tensor_parameter_204: class Program_weight_tensor_parameter_205: name = "parameter_205" + original_name = "batch_norm2d_5.w_2" shape = [16] dtype = "float32" min_val = float("0") @@ -2098,6 +2304,7 @@ class Program_weight_tensor_parameter_205: class Program_weight_tensor_parameter_206: name = "parameter_206" + original_name = "batch_norm2d_5.w_1" shape = [16] dtype = "float32" min_val = float("0") @@ -2107,6 +2314,7 @@ class Program_weight_tensor_parameter_206: class Program_weight_tensor_parameter_207: name = "parameter_207" + original_name = "conv2d_5.w_0" shape = [16, 16, 3, 3] dtype = "float32" min_val = float("-1.9931") @@ -2118,6 +2326,7 @@ class Program_weight_tensor_parameter_207: class Program_weight_tensor_parameter_208: name = "parameter_208" + original_name = "batch_norm2d_4.b_0" shape = [16] dtype = "float32" min_val = float("0") @@ -2127,6 +2336,7 @@ class Program_weight_tensor_parameter_208: class Program_weight_tensor_parameter_209: name = "parameter_209" + original_name = "batch_norm2d_4.w_0" shape = [16] dtype = "float32" min_val = float("0") @@ -2136,6 +2346,7 @@ class Program_weight_tensor_parameter_209: class Program_weight_tensor_parameter_210: name = "parameter_210" + original_name = "batch_norm2d_4.w_2" shape = [16] dtype = "float32" min_val = float("0") @@ -2145,6 +2356,7 @@ class Program_weight_tensor_parameter_210: class Program_weight_tensor_parameter_211: name = "parameter_211" + original_name = "batch_norm2d_4.w_1" shape = [16] dtype = "float32" min_val = float("0") @@ -2154,6 +2366,7 @@ class Program_weight_tensor_parameter_211: class Program_weight_tensor_parameter_212: name = "parameter_212" + original_name = "conv2d_4.w_0" shape = [16, 16, 1, 1] dtype = "float32" min_val = float("-2.42264") @@ -2165,6 +2378,7 @@ class Program_weight_tensor_parameter_212: class Program_weight_tensor_parameter_213: name = "parameter_213" + original_name = "batch_norm2d_3.b_0" shape = [16] dtype = "float32" min_val = float("0") @@ -2174,6 +2388,7 @@ class Program_weight_tensor_parameter_213: class Program_weight_tensor_parameter_214: name = "parameter_214" + original_name = "batch_norm2d_3.w_0" shape = [16] dtype = "float32" min_val = float("0") @@ -2183,6 +2398,7 @@ class Program_weight_tensor_parameter_214: class Program_weight_tensor_parameter_215: name = "parameter_215" + original_name = "batch_norm2d_3.w_2" shape = [16] dtype = "float32" min_val = float("0") @@ -2192,6 +2408,7 @@ class Program_weight_tensor_parameter_215: class Program_weight_tensor_parameter_216: name = "parameter_216" + original_name = "batch_norm2d_3.w_1" shape = [16] dtype = "float32" min_val = float("0") @@ -2201,6 +2418,7 @@ class Program_weight_tensor_parameter_216: class Program_weight_tensor_parameter_217: name = "parameter_217" + original_name = "conv2d_3.w_0" shape = [16, 32, 3, 3] dtype = "float32" min_val = float("-1.58105") @@ -2212,6 +2430,7 @@ class Program_weight_tensor_parameter_217: class Program_weight_tensor_parameter_218: name = "parameter_218" + original_name = "batch_norm2d_2.b_0" shape = [16] dtype = "float32" min_val = float("0") @@ -2221,6 +2440,7 @@ class Program_weight_tensor_parameter_218: class Program_weight_tensor_parameter_219: name = "parameter_219" + original_name = "batch_norm2d_2.w_0" shape = [16] dtype = "float32" min_val = float("0") @@ -2230,6 +2450,7 @@ class Program_weight_tensor_parameter_219: class Program_weight_tensor_parameter_220: name = "parameter_220" + original_name = "batch_norm2d_2.w_2" shape = [16] dtype = "float32" min_val = float("0") @@ -2239,6 +2460,7 @@ class Program_weight_tensor_parameter_220: class Program_weight_tensor_parameter_221: name = "parameter_221" + original_name = "batch_norm2d_2.w_1" shape = [16] dtype = "float32" min_val = float("0") @@ -2248,6 +2470,7 @@ class Program_weight_tensor_parameter_221: class Program_weight_tensor_parameter_222: name = "parameter_222" + original_name = "conv2d_2.w_0" shape = [16, 8, 2, 2] dtype = "float32" min_val = float("-1.16564") @@ -2259,6 +2482,7 @@ class Program_weight_tensor_parameter_222: class Program_weight_tensor_parameter_223: name = "parameter_223" + original_name = "batch_norm2d_1.b_0" shape = [8] dtype = "float32" min_val = float("0") @@ -2268,6 +2492,7 @@ class Program_weight_tensor_parameter_223: class Program_weight_tensor_parameter_224: name = "parameter_224" + original_name = "batch_norm2d_1.w_0" shape = [8] dtype = "float32" min_val = float("0") @@ -2277,6 +2502,7 @@ class Program_weight_tensor_parameter_224: class Program_weight_tensor_parameter_225: name = "parameter_225" + original_name = "batch_norm2d_1.w_2" shape = [8] dtype = "float32" min_val = float("0") @@ -2286,6 +2512,7 @@ class Program_weight_tensor_parameter_225: class Program_weight_tensor_parameter_226: name = "parameter_226" + original_name = "batch_norm2d_1.w_1" shape = [8] dtype = "float32" min_val = float("0") @@ -2295,6 +2522,7 @@ class Program_weight_tensor_parameter_226: class Program_weight_tensor_parameter_227: name = "parameter_227" + original_name = "conv2d_1.w_0" shape = [8, 16, 2, 2] dtype = "float32" min_val = float("-1.20596") @@ -2306,6 +2534,7 @@ class Program_weight_tensor_parameter_227: class Program_weight_tensor_parameter_228: name = "parameter_228" + original_name = "batch_norm2d_0.b_0" shape = [16] dtype = "float32" min_val = float("0") @@ -2315,6 +2544,7 @@ class Program_weight_tensor_parameter_228: class Program_weight_tensor_parameter_229: name = "parameter_229" + original_name = "batch_norm2d_0.w_0" shape = [16] dtype = "float32" min_val = float("0") @@ -2324,6 +2554,7 @@ class Program_weight_tensor_parameter_229: class Program_weight_tensor_parameter_230: name = "parameter_230" + original_name = "batch_norm2d_0.w_2" shape = [16] dtype = "float32" min_val = float("0") @@ -2333,6 +2564,7 @@ class Program_weight_tensor_parameter_230: class Program_weight_tensor_parameter_231: name = "parameter_231" + original_name = "batch_norm2d_0.w_1" shape = [16] dtype = "float32" min_val = float("0") @@ -2342,6 +2574,7 @@ class Program_weight_tensor_parameter_231: class Program_weight_tensor_parameter_232: name = "parameter_232" + original_name = "conv2d_0.w_0" shape = [16, 3, 3, 3] dtype = "float32" min_val = float("-2.15329") diff --git a/paddle_samples/PaddleX/PP-HGNetV2-B4_ML/subgraph_0/input_meta.py b/paddle_samples/PaddleX/PP-HGNetV2-B4_ML/subgraph_0/input_meta.py index b5fc033ed..efed7799c 100644 --- a/paddle_samples/PaddleX/PP-HGNetV2-B4_ML/subgraph_0/input_meta.py +++ b/paddle_samples/PaddleX/PP-HGNetV2-B4_ML/subgraph_0/input_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_data_0: name = "data_0" + original_name = "param_0" shape = [33, 768] dtype = "float32" min_val = float("-0.0865409") @@ -11,6 +12,7 @@ class Program_weight_tensor_data_0: class Program_weight_tensor_data_1: name = "data_1" + original_name = "var_0" shape = [32, 3, 448, 448] dtype = "float32" min_val = float("-2.1179") diff --git a/paddle_samples/PaddleX/PP-HGNetV2-B4_ML/subgraph_0/weight_meta.py b/paddle_samples/PaddleX/PP-HGNetV2-B4_ML/subgraph_0/weight_meta.py index 4d13b75a8..c3fead8e1 100644 --- a/paddle_samples/PaddleX/PP-HGNetV2-B4_ML/subgraph_0/weight_meta.py +++ b/paddle_samples/PaddleX/PP-HGNetV2-B4_ML/subgraph_0/weight_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_parameter_0: name = "parameter_0" + original_name = "conv2d_82.b_0" shape = [33] dtype = "float32" min_val = float("0") @@ -9,6 +10,7 @@ class Program_weight_tensor_parameter_0: class Program_weight_tensor_parameter_1: name = "parameter_1" + original_name = "conv2d_82.w_0" shape = [33, 768, 1, 1] dtype = "float32" min_val = float("-0.206653") @@ -20,6 +22,7 @@ class Program_weight_tensor_parameter_1: class Program_weight_tensor_parameter_2: name = "parameter_2" + original_name = "layer_norm_2.b_0" shape = [768] dtype = "float32" min_val = float("-0.00234522") @@ -31,6 +34,7 @@ class Program_weight_tensor_parameter_2: class Program_weight_tensor_parameter_3: name = "parameter_3" + original_name = "layer_norm_2.w_0" shape = [768] dtype = "float32" min_val = float("0.95996") @@ -42,6 +46,7 @@ class Program_weight_tensor_parameter_3: class Program_weight_tensor_parameter_4: name = "parameter_4" + original_name = "linear_10.b_0" shape = [768] dtype = "float32" min_val = float("-0.0117901") @@ -53,6 +58,7 @@ class Program_weight_tensor_parameter_4: class Program_weight_tensor_parameter_5: name = "parameter_5" + original_name = "linear_10.w_0" shape = [2048, 768] dtype = "float32" min_val = float("-0.0637204") @@ -64,6 +70,7 @@ class Program_weight_tensor_parameter_5: class Program_weight_tensor_parameter_6: name = "parameter_6" + original_name = "linear_9.b_0" shape = [2048] dtype = "float32" min_val = float("-0.010954") @@ -75,6 +82,7 @@ class Program_weight_tensor_parameter_6: class Program_weight_tensor_parameter_7: name = "parameter_7" + original_name = "linear_9.w_0" shape = [768, 2048] dtype = "float32" min_val = float("-0.0591675") @@ -86,6 +94,7 @@ class Program_weight_tensor_parameter_7: class Program_weight_tensor_parameter_8: name = "parameter_8" + original_name = "layer_norm_1.b_0" shape = [768] dtype = "float32" min_val = float("-0.0353723") @@ -97,6 +106,7 @@ class Program_weight_tensor_parameter_8: class Program_weight_tensor_parameter_9: name = "parameter_9" + original_name = "layer_norm_1.w_0" shape = [768] dtype = "float32" min_val = float("0.99269") @@ -108,6 +118,7 @@ class Program_weight_tensor_parameter_9: class Program_weight_tensor_parameter_10: name = "parameter_10" + original_name = "linear_8.b_0" shape = [768] dtype = "float32" min_val = float("-0.0263088") @@ -119,6 +130,7 @@ class Program_weight_tensor_parameter_10: class Program_weight_tensor_parameter_11: name = "parameter_11" + original_name = "linear_8.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.0930652") @@ -130,6 +142,7 @@ class Program_weight_tensor_parameter_11: class Program_weight_tensor_parameter_12: name = "parameter_12" + original_name = "linear_7.b_0" shape = [768] dtype = "float32" min_val = float("-0.00426818") @@ -141,6 +154,7 @@ class Program_weight_tensor_parameter_12: class Program_weight_tensor_parameter_13: name = "parameter_13" + original_name = "linear_7.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.0756318") @@ -152,6 +166,7 @@ class Program_weight_tensor_parameter_13: class Program_weight_tensor_parameter_14: name = "parameter_14" + original_name = "linear_6.b_0" shape = [768] dtype = "float32" min_val = float("-0.010354") @@ -163,6 +178,7 @@ class Program_weight_tensor_parameter_14: class Program_weight_tensor_parameter_15: name = "parameter_15" + original_name = "linear_6.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.0742996") @@ -174,6 +190,7 @@ class Program_weight_tensor_parameter_15: class Program_weight_tensor_parameter_16: name = "parameter_16" + original_name = "linear_5.b_0" shape = [768] dtype = "float32" min_val = float("-0.0114138") @@ -185,6 +202,7 @@ class Program_weight_tensor_parameter_16: class Program_weight_tensor_parameter_17: name = "parameter_17" + original_name = "linear_5.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.0733468") @@ -196,6 +214,7 @@ class Program_weight_tensor_parameter_17: class Program_weight_tensor_parameter_18: name = "parameter_18" + original_name = "layer_norm_0.b_0" shape = [768] dtype = "float32" min_val = float("-0.0289076") @@ -207,6 +226,7 @@ class Program_weight_tensor_parameter_18: class Program_weight_tensor_parameter_19: name = "parameter_19" + original_name = "layer_norm_0.w_0" shape = [768] dtype = "float32" min_val = float("0.988295") @@ -218,6 +238,7 @@ class Program_weight_tensor_parameter_19: class Program_weight_tensor_parameter_20: name = "parameter_20" + original_name = "conv2d_81.b_0" shape = [768] dtype = "float32" min_val = float("-0.00359367") @@ -229,6 +250,7 @@ class Program_weight_tensor_parameter_20: class Program_weight_tensor_parameter_21: name = "parameter_21" + original_name = "conv2d_81.w_0" shape = [768, 2048, 1, 1] dtype = "float32" min_val = float("-0.151871") @@ -240,6 +262,7 @@ class Program_weight_tensor_parameter_21: class Program_weight_tensor_parameter_22: name = "parameter_22" + original_name = "conv2d_80.w_0" shape = [2048, 2048, 1, 1] dtype = "float32" min_val = float("-0.303594") @@ -251,6 +274,7 @@ class Program_weight_tensor_parameter_22: class Program_weight_tensor_parameter_23: name = "parameter_23" + original_name = "batch_norm2d_79.b_0" shape = [2048] dtype = "float32" min_val = float("-9.09639") @@ -262,6 +286,7 @@ class Program_weight_tensor_parameter_23: class Program_weight_tensor_parameter_24: name = "parameter_24" + original_name = "batch_norm2d_79.w_0" shape = [2048] dtype = "float32" min_val = float("2.20235") @@ -273,6 +298,7 @@ class Program_weight_tensor_parameter_24: class Program_weight_tensor_parameter_25: name = "parameter_25" + original_name = "batch_norm2d_79.w_2" shape = [2048] dtype = "float32" min_val = float("0.0239894") @@ -284,6 +310,7 @@ class Program_weight_tensor_parameter_25: class Program_weight_tensor_parameter_26: name = "parameter_26" + original_name = "batch_norm2d_79.w_1" shape = [2048] dtype = "float32" min_val = float("-0.941602") @@ -295,6 +322,7 @@ class Program_weight_tensor_parameter_26: class Program_weight_tensor_parameter_27: name = "parameter_27" + original_name = "conv2d_79.w_0" shape = [2048, 1024, 1, 1] dtype = "float32" min_val = float("-0.480226") @@ -306,6 +334,7 @@ class Program_weight_tensor_parameter_27: class Program_weight_tensor_parameter_28: name = "parameter_28" + original_name = "batch_norm2d_78.b_0" shape = [1024] dtype = "float32" min_val = float("-5.73131") @@ -317,6 +346,7 @@ class Program_weight_tensor_parameter_28: class Program_weight_tensor_parameter_29: name = "parameter_29" + original_name = "batch_norm2d_78.w_0" shape = [1024] dtype = "float32" min_val = float("0.604286") @@ -328,6 +358,7 @@ class Program_weight_tensor_parameter_29: class Program_weight_tensor_parameter_30: name = "parameter_30" + original_name = "batch_norm2d_78.w_2" shape = [1024] dtype = "float32" min_val = float("0.251613") @@ -339,6 +370,7 @@ class Program_weight_tensor_parameter_30: class Program_weight_tensor_parameter_31: name = "parameter_31" + original_name = "batch_norm2d_78.w_1" shape = [1024] dtype = "float32" min_val = float("-2.16008") @@ -350,6 +382,7 @@ class Program_weight_tensor_parameter_31: class Program_weight_tensor_parameter_32: name = "parameter_32" + original_name = "conv2d_78.w_0" shape = [1024, 3328, 1, 1] dtype = "float32" min_val = float("-0.555938") @@ -361,6 +394,7 @@ class Program_weight_tensor_parameter_32: class Program_weight_tensor_parameter_33: name = "parameter_33" + original_name = "batch_norm2d_77.b_0" shape = [384] dtype = "float32" min_val = float("-3.73784") @@ -372,6 +406,7 @@ class Program_weight_tensor_parameter_33: class Program_weight_tensor_parameter_34: name = "parameter_34" + original_name = "batch_norm2d_77.w_0" shape = [384] dtype = "float32" min_val = float("0.694568") @@ -383,6 +418,7 @@ class Program_weight_tensor_parameter_34: class Program_weight_tensor_parameter_35: name = "parameter_35" + original_name = "batch_norm2d_77.w_2" shape = [384] dtype = "float32" min_val = float("0.00909812") @@ -394,6 +430,7 @@ class Program_weight_tensor_parameter_35: class Program_weight_tensor_parameter_36: name = "parameter_36" + original_name = "batch_norm2d_77.w_1" shape = [384] dtype = "float32" min_val = float("-0.360724") @@ -405,6 +442,7 @@ class Program_weight_tensor_parameter_36: class Program_weight_tensor_parameter_37: name = "parameter_37" + original_name = "conv2d_77.w_0" shape = [384, 1, 5, 5] dtype = "float32" min_val = float("-0.238502") @@ -416,6 +454,7 @@ class Program_weight_tensor_parameter_37: class Program_weight_tensor_parameter_38: name = "parameter_38" + original_name = "batch_norm2d_76.b_0" shape = [384] dtype = "float32" min_val = float("-1.01471") @@ -427,6 +466,7 @@ class Program_weight_tensor_parameter_38: class Program_weight_tensor_parameter_39: name = "parameter_39" + original_name = "batch_norm2d_76.w_0" shape = [384] dtype = "float32" min_val = float("0.54682") @@ -438,6 +478,7 @@ class Program_weight_tensor_parameter_39: class Program_weight_tensor_parameter_40: name = "parameter_40" + original_name = "batch_norm2d_76.w_2" shape = [384] dtype = "float32" min_val = float("0.0087312") @@ -449,6 +490,7 @@ class Program_weight_tensor_parameter_40: class Program_weight_tensor_parameter_41: name = "parameter_41" + original_name = "batch_norm2d_76.w_1" shape = [384] dtype = "float32" min_val = float("-0.679546") @@ -460,6 +502,7 @@ class Program_weight_tensor_parameter_41: class Program_weight_tensor_parameter_42: name = "parameter_42" + original_name = "conv2d_76.w_0" shape = [384, 384, 1, 1] dtype = "float32" min_val = float("-0.379942") @@ -471,6 +514,7 @@ class Program_weight_tensor_parameter_42: class Program_weight_tensor_parameter_43: name = "parameter_43" + original_name = "batch_norm2d_75.b_0" shape = [384] dtype = "float32" min_val = float("-3.03852") @@ -482,6 +526,7 @@ class Program_weight_tensor_parameter_43: class Program_weight_tensor_parameter_44: name = "parameter_44" + original_name = "batch_norm2d_75.w_0" shape = [384] dtype = "float32" min_val = float("0.477033") @@ -493,6 +538,7 @@ class Program_weight_tensor_parameter_44: class Program_weight_tensor_parameter_45: name = "parameter_45" + original_name = "batch_norm2d_75.w_2" shape = [384] dtype = "float32" min_val = float("0.00486645") @@ -504,6 +550,7 @@ class Program_weight_tensor_parameter_45: class Program_weight_tensor_parameter_46: name = "parameter_46" + original_name = "batch_norm2d_75.w_1" shape = [384] dtype = "float32" min_val = float("-0.672571") @@ -515,6 +562,7 @@ class Program_weight_tensor_parameter_46: class Program_weight_tensor_parameter_47: name = "parameter_47" + original_name = "conv2d_75.w_0" shape = [384, 1, 5, 5] dtype = "float32" min_val = float("-0.190216") @@ -526,6 +574,7 @@ class Program_weight_tensor_parameter_47: class Program_weight_tensor_parameter_48: name = "parameter_48" + original_name = "batch_norm2d_74.b_0" shape = [384] dtype = "float32" min_val = float("-1.03953") @@ -537,6 +586,7 @@ class Program_weight_tensor_parameter_48: class Program_weight_tensor_parameter_49: name = "parameter_49" + original_name = "batch_norm2d_74.w_0" shape = [384] dtype = "float32" min_val = float("0.397401") @@ -548,6 +598,7 @@ class Program_weight_tensor_parameter_49: class Program_weight_tensor_parameter_50: name = "parameter_50" + original_name = "batch_norm2d_74.w_2" shape = [384] dtype = "float32" min_val = float("0.0115856") @@ -559,6 +610,7 @@ class Program_weight_tensor_parameter_50: class Program_weight_tensor_parameter_51: name = "parameter_51" + original_name = "batch_norm2d_74.w_1" shape = [384] dtype = "float32" min_val = float("-0.432512") @@ -570,6 +622,7 @@ class Program_weight_tensor_parameter_51: class Program_weight_tensor_parameter_52: name = "parameter_52" + original_name = "conv2d_74.w_0" shape = [384, 384, 1, 1] dtype = "float32" min_val = float("-0.144805") @@ -581,6 +634,7 @@ class Program_weight_tensor_parameter_52: class Program_weight_tensor_parameter_53: name = "parameter_53" + original_name = "batch_norm2d_73.b_0" shape = [384] dtype = "float32" min_val = float("-2.72879") @@ -592,6 +646,7 @@ class Program_weight_tensor_parameter_53: class Program_weight_tensor_parameter_54: name = "parameter_54" + original_name = "batch_norm2d_73.w_0" shape = [384] dtype = "float32" min_val = float("0.470188") @@ -603,6 +658,7 @@ class Program_weight_tensor_parameter_54: class Program_weight_tensor_parameter_55: name = "parameter_55" + original_name = "batch_norm2d_73.w_2" shape = [384] dtype = "float32" min_val = float("0.00382899") @@ -614,6 +670,7 @@ class Program_weight_tensor_parameter_55: class Program_weight_tensor_parameter_56: name = "parameter_56" + original_name = "batch_norm2d_73.w_1" shape = [384] dtype = "float32" min_val = float("-0.775394") @@ -625,6 +682,7 @@ class Program_weight_tensor_parameter_56: class Program_weight_tensor_parameter_57: name = "parameter_57" + original_name = "conv2d_73.w_0" shape = [384, 1, 5, 5] dtype = "float32" min_val = float("-0.246246") @@ -636,6 +694,7 @@ class Program_weight_tensor_parameter_57: class Program_weight_tensor_parameter_58: name = "parameter_58" + original_name = "batch_norm2d_72.b_0" shape = [384] dtype = "float32" min_val = float("-1.33719") @@ -647,6 +706,7 @@ class Program_weight_tensor_parameter_58: class Program_weight_tensor_parameter_59: name = "parameter_59" + original_name = "batch_norm2d_72.w_0" shape = [384] dtype = "float32" min_val = float("0.452236") @@ -658,6 +718,7 @@ class Program_weight_tensor_parameter_59: class Program_weight_tensor_parameter_60: name = "parameter_60" + original_name = "batch_norm2d_72.w_2" shape = [384] dtype = "float32" min_val = float("0.0181699") @@ -669,6 +730,7 @@ class Program_weight_tensor_parameter_60: class Program_weight_tensor_parameter_61: name = "parameter_61" + original_name = "batch_norm2d_72.w_1" shape = [384] dtype = "float32" min_val = float("-0.998168") @@ -680,6 +742,7 @@ class Program_weight_tensor_parameter_61: class Program_weight_tensor_parameter_62: name = "parameter_62" + original_name = "conv2d_72.w_0" shape = [384, 384, 1, 1] dtype = "float32" min_val = float("-0.279953") @@ -691,6 +754,7 @@ class Program_weight_tensor_parameter_62: class Program_weight_tensor_parameter_63: name = "parameter_63" + original_name = "batch_norm2d_71.b_0" shape = [384] dtype = "float32" min_val = float("-3.32388") @@ -702,6 +766,7 @@ class Program_weight_tensor_parameter_63: class Program_weight_tensor_parameter_64: name = "parameter_64" + original_name = "batch_norm2d_71.w_0" shape = [384] dtype = "float32" min_val = float("0.51119") @@ -713,6 +778,7 @@ class Program_weight_tensor_parameter_64: class Program_weight_tensor_parameter_65: name = "parameter_65" + original_name = "batch_norm2d_71.w_2" shape = [384] dtype = "float32" min_val = float("0.00814008") @@ -724,6 +790,7 @@ class Program_weight_tensor_parameter_65: class Program_weight_tensor_parameter_66: name = "parameter_66" + original_name = "batch_norm2d_71.w_1" shape = [384] dtype = "float32" min_val = float("-0.882301") @@ -735,6 +802,7 @@ class Program_weight_tensor_parameter_66: class Program_weight_tensor_parameter_67: name = "parameter_67" + original_name = "conv2d_71.w_0" shape = [384, 1, 5, 5] dtype = "float32" min_val = float("-0.263098") @@ -746,6 +814,7 @@ class Program_weight_tensor_parameter_67: class Program_weight_tensor_parameter_68: name = "parameter_68" + original_name = "batch_norm2d_70.b_0" shape = [384] dtype = "float32" min_val = float("-1.3028") @@ -757,6 +826,7 @@ class Program_weight_tensor_parameter_68: class Program_weight_tensor_parameter_69: name = "parameter_69" + original_name = "batch_norm2d_70.w_0" shape = [384] dtype = "float32" min_val = float("0.403633") @@ -768,6 +838,7 @@ class Program_weight_tensor_parameter_69: class Program_weight_tensor_parameter_70: name = "parameter_70" + original_name = "batch_norm2d_70.w_2" shape = [384] dtype = "float32" min_val = float("0.0203441") @@ -779,6 +850,7 @@ class Program_weight_tensor_parameter_70: class Program_weight_tensor_parameter_71: name = "parameter_71" + original_name = "batch_norm2d_70.w_1" shape = [384] dtype = "float32" min_val = float("-1.16273") @@ -790,6 +862,7 @@ class Program_weight_tensor_parameter_71: class Program_weight_tensor_parameter_72: name = "parameter_72" + original_name = "conv2d_70.w_0" shape = [384, 384, 1, 1] dtype = "float32" min_val = float("-0.412967") @@ -801,6 +874,7 @@ class Program_weight_tensor_parameter_72: class Program_weight_tensor_parameter_73: name = "parameter_73" + original_name = "batch_norm2d_69.b_0" shape = [384] dtype = "float32" min_val = float("-3.69561") @@ -812,6 +886,7 @@ class Program_weight_tensor_parameter_73: class Program_weight_tensor_parameter_74: name = "parameter_74" + original_name = "batch_norm2d_69.w_0" shape = [384] dtype = "float32" min_val = float("0.451449") @@ -823,6 +898,7 @@ class Program_weight_tensor_parameter_74: class Program_weight_tensor_parameter_75: name = "parameter_75" + original_name = "batch_norm2d_69.w_2" shape = [384] dtype = "float32" min_val = float("0.0158173") @@ -834,6 +910,7 @@ class Program_weight_tensor_parameter_75: class Program_weight_tensor_parameter_76: name = "parameter_76" + original_name = "batch_norm2d_69.w_1" shape = [384] dtype = "float32" min_val = float("-1.09525") @@ -845,6 +922,7 @@ class Program_weight_tensor_parameter_76: class Program_weight_tensor_parameter_77: name = "parameter_77" + original_name = "conv2d_69.w_0" shape = [384, 1, 5, 5] dtype = "float32" min_val = float("-0.276942") @@ -856,6 +934,7 @@ class Program_weight_tensor_parameter_77: class Program_weight_tensor_parameter_78: name = "parameter_78" + original_name = "batch_norm2d_68.b_0" shape = [384] dtype = "float32" min_val = float("-1.68842") @@ -867,6 +946,7 @@ class Program_weight_tensor_parameter_78: class Program_weight_tensor_parameter_79: name = "parameter_79" + original_name = "batch_norm2d_68.w_0" shape = [384] dtype = "float32" min_val = float("0.436377") @@ -878,6 +958,7 @@ class Program_weight_tensor_parameter_79: class Program_weight_tensor_parameter_80: name = "parameter_80" + original_name = "batch_norm2d_68.w_2" shape = [384] dtype = "float32" min_val = float("0.0421427") @@ -889,6 +970,7 @@ class Program_weight_tensor_parameter_80: class Program_weight_tensor_parameter_81: name = "parameter_81" + original_name = "batch_norm2d_68.w_1" shape = [384] dtype = "float32" min_val = float("-1.59478") @@ -900,6 +982,7 @@ class Program_weight_tensor_parameter_81: class Program_weight_tensor_parameter_82: name = "parameter_82" + original_name = "conv2d_68.w_0" shape = [384, 384, 1, 1] dtype = "float32" min_val = float("-0.400071") @@ -911,6 +994,7 @@ class Program_weight_tensor_parameter_82: class Program_weight_tensor_parameter_83: name = "parameter_83" + original_name = "batch_norm2d_67.b_0" shape = [384] dtype = "float32" min_val = float("-6.32674") @@ -922,6 +1006,7 @@ class Program_weight_tensor_parameter_83: class Program_weight_tensor_parameter_84: name = "parameter_84" + original_name = "batch_norm2d_67.w_0" shape = [384] dtype = "float32" min_val = float("0.611355") @@ -933,6 +1018,7 @@ class Program_weight_tensor_parameter_84: class Program_weight_tensor_parameter_85: name = "parameter_85" + original_name = "batch_norm2d_67.w_2" shape = [384] dtype = "float32" min_val = float("0.0273804") @@ -944,6 +1030,7 @@ class Program_weight_tensor_parameter_85: class Program_weight_tensor_parameter_86: name = "parameter_86" + original_name = "batch_norm2d_67.w_1" shape = [384] dtype = "float32" min_val = float("-0.486724") @@ -955,6 +1042,7 @@ class Program_weight_tensor_parameter_86: class Program_weight_tensor_parameter_87: name = "parameter_87" + original_name = "conv2d_67.w_0" shape = [384, 1, 5, 5] dtype = "float32" min_val = float("-0.397212") @@ -966,6 +1054,7 @@ class Program_weight_tensor_parameter_87: class Program_weight_tensor_parameter_88: name = "parameter_88" + original_name = "batch_norm2d_66.b_0" shape = [384] dtype = "float32" min_val = float("-1.32699") @@ -977,6 +1066,7 @@ class Program_weight_tensor_parameter_88: class Program_weight_tensor_parameter_89: name = "parameter_89" + original_name = "batch_norm2d_66.w_0" shape = [384] dtype = "float32" min_val = float("0.529452") @@ -988,6 +1078,7 @@ class Program_weight_tensor_parameter_89: class Program_weight_tensor_parameter_90: name = "parameter_90" + original_name = "batch_norm2d_66.w_2" shape = [384] dtype = "float32" min_val = float("0.147408") @@ -999,6 +1090,7 @@ class Program_weight_tensor_parameter_90: class Program_weight_tensor_parameter_91: name = "parameter_91" + original_name = "batch_norm2d_66.w_1" shape = [384] dtype = "float32" min_val = float("-0.00843827") @@ -1010,6 +1102,7 @@ class Program_weight_tensor_parameter_91: class Program_weight_tensor_parameter_92: name = "parameter_92" + original_name = "conv2d_66.w_0" shape = [384, 1024, 1, 1] dtype = "float32" min_val = float("-0.435623") @@ -1021,6 +1114,7 @@ class Program_weight_tensor_parameter_92: class Program_weight_tensor_parameter_93: name = "parameter_93" + original_name = "batch_norm2d_65.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0114276") @@ -1032,6 +1126,7 @@ class Program_weight_tensor_parameter_93: class Program_weight_tensor_parameter_94: name = "parameter_94" + original_name = "batch_norm2d_65.w_0" shape = [1024] dtype = "float32" min_val = float("0.650006") @@ -1043,6 +1138,7 @@ class Program_weight_tensor_parameter_94: class Program_weight_tensor_parameter_95: name = "parameter_95" + original_name = "batch_norm2d_65.w_2" shape = [1024] dtype = "float32" min_val = float("0.00242073") @@ -1054,6 +1150,7 @@ class Program_weight_tensor_parameter_95: class Program_weight_tensor_parameter_96: name = "parameter_96" + original_name = "batch_norm2d_65.w_1" shape = [1024] dtype = "float32" min_val = float("-0.779907") @@ -1065,6 +1162,7 @@ class Program_weight_tensor_parameter_96: class Program_weight_tensor_parameter_97: name = "parameter_97" + original_name = "conv2d_65.w_0" shape = [1024, 1, 3, 3] dtype = "float32" min_val = float("-0.0951603") @@ -1076,6 +1174,7 @@ class Program_weight_tensor_parameter_97: class Program_weight_tensor_parameter_98: name = "parameter_98" + original_name = "batch_norm2d_64.b_0" shape = [1024] dtype = "float32" min_val = float("-4.62928") @@ -1087,6 +1186,7 @@ class Program_weight_tensor_parameter_98: class Program_weight_tensor_parameter_99: name = "parameter_99" + original_name = "batch_norm2d_64.w_0" shape = [1024] dtype = "float32" min_val = float("0.0406095") @@ -1098,6 +1198,7 @@ class Program_weight_tensor_parameter_99: class Program_weight_tensor_parameter_100: name = "parameter_100" + original_name = "batch_norm2d_64.w_2" shape = [1024] dtype = "float32" min_val = float("5.60519e-45") @@ -1109,6 +1210,7 @@ class Program_weight_tensor_parameter_100: class Program_weight_tensor_parameter_101: name = "parameter_101" + original_name = "batch_norm2d_64.w_1" shape = [1024] dtype = "float32" min_val = float("-0.808307") @@ -1120,6 +1222,7 @@ class Program_weight_tensor_parameter_101: class Program_weight_tensor_parameter_102: name = "parameter_102" + original_name = "conv2d_64.w_0" shape = [1024, 512, 1, 1] dtype = "float32" min_val = float("-0.219781") @@ -1131,6 +1234,7 @@ class Program_weight_tensor_parameter_102: class Program_weight_tensor_parameter_103: name = "parameter_103" + original_name = "batch_norm2d_63.b_0" shape = [512] dtype = "float32" min_val = float("-5.64037") @@ -1142,6 +1246,7 @@ class Program_weight_tensor_parameter_103: class Program_weight_tensor_parameter_104: name = "parameter_104" + original_name = "batch_norm2d_63.w_0" shape = [512] dtype = "float32" min_val = float("0.666518") @@ -1153,6 +1258,7 @@ class Program_weight_tensor_parameter_104: class Program_weight_tensor_parameter_105: name = "parameter_105" + original_name = "batch_norm2d_63.w_2" shape = [512] dtype = "float32" min_val = float("0.0820819") @@ -1164,6 +1270,7 @@ class Program_weight_tensor_parameter_105: class Program_weight_tensor_parameter_106: name = "parameter_106" + original_name = "batch_norm2d_63.w_1" shape = [512] dtype = "float32" min_val = float("-1.43632") @@ -1175,6 +1282,7 @@ class Program_weight_tensor_parameter_106: class Program_weight_tensor_parameter_107: name = "parameter_107" + original_name = "conv2d_63.w_0" shape = [512, 2176, 1, 1] dtype = "float32" min_val = float("-0.270045") @@ -1186,6 +1294,7 @@ class Program_weight_tensor_parameter_107: class Program_weight_tensor_parameter_108: name = "parameter_108" + original_name = "batch_norm2d_62.b_0" shape = [192] dtype = "float32" min_val = float("-2.50479") @@ -1197,6 +1306,7 @@ class Program_weight_tensor_parameter_108: class Program_weight_tensor_parameter_109: name = "parameter_109" + original_name = "batch_norm2d_62.w_0" shape = [192] dtype = "float32" min_val = float("0.398613") @@ -1208,6 +1318,7 @@ class Program_weight_tensor_parameter_109: class Program_weight_tensor_parameter_110: name = "parameter_110" + original_name = "batch_norm2d_62.w_2" shape = [192] dtype = "float32" min_val = float("0.00415056") @@ -1219,6 +1330,7 @@ class Program_weight_tensor_parameter_110: class Program_weight_tensor_parameter_111: name = "parameter_111" + original_name = "batch_norm2d_62.w_1" shape = [192] dtype = "float32" min_val = float("-0.9102") @@ -1230,6 +1342,7 @@ class Program_weight_tensor_parameter_111: class Program_weight_tensor_parameter_112: name = "parameter_112" + original_name = "conv2d_62.w_0" shape = [192, 1, 5, 5] dtype = "float32" min_val = float("-0.211267") @@ -1241,6 +1354,7 @@ class Program_weight_tensor_parameter_112: class Program_weight_tensor_parameter_113: name = "parameter_113" + original_name = "batch_norm2d_61.b_0" shape = [192] dtype = "float32" min_val = float("-1.3447") @@ -1252,6 +1366,7 @@ class Program_weight_tensor_parameter_113: class Program_weight_tensor_parameter_114: name = "parameter_114" + original_name = "batch_norm2d_61.w_0" shape = [192] dtype = "float32" min_val = float("0.298437") @@ -1263,6 +1378,7 @@ class Program_weight_tensor_parameter_114: class Program_weight_tensor_parameter_115: name = "parameter_115" + original_name = "batch_norm2d_61.w_2" shape = [192] dtype = "float32" min_val = float("0.00299239") @@ -1274,6 +1390,7 @@ class Program_weight_tensor_parameter_115: class Program_weight_tensor_parameter_116: name = "parameter_116" + original_name = "batch_norm2d_61.w_1" shape = [192] dtype = "float32" min_val = float("-0.258833") @@ -1285,6 +1402,7 @@ class Program_weight_tensor_parameter_116: class Program_weight_tensor_parameter_117: name = "parameter_117" + original_name = "conv2d_61.w_0" shape = [192, 192, 1, 1] dtype = "float32" min_val = float("-0.16194") @@ -1296,6 +1414,7 @@ class Program_weight_tensor_parameter_117: class Program_weight_tensor_parameter_118: name = "parameter_118" + original_name = "batch_norm2d_60.b_0" shape = [192] dtype = "float32" min_val = float("-2.31545") @@ -1307,6 +1426,7 @@ class Program_weight_tensor_parameter_118: class Program_weight_tensor_parameter_119: name = "parameter_119" + original_name = "batch_norm2d_60.w_0" shape = [192] dtype = "float32" min_val = float("0.308319") @@ -1318,6 +1438,7 @@ class Program_weight_tensor_parameter_119: class Program_weight_tensor_parameter_120: name = "parameter_120" + original_name = "batch_norm2d_60.w_2" shape = [192] dtype = "float32" min_val = float("0.0147881") @@ -1329,6 +1450,7 @@ class Program_weight_tensor_parameter_120: class Program_weight_tensor_parameter_121: name = "parameter_121" + original_name = "batch_norm2d_60.w_1" shape = [192] dtype = "float32" min_val = float("-0.380377") @@ -1340,6 +1462,7 @@ class Program_weight_tensor_parameter_121: class Program_weight_tensor_parameter_122: name = "parameter_122" + original_name = "conv2d_60.w_0" shape = [192, 1, 5, 5] dtype = "float32" min_val = float("-0.235803") @@ -1351,6 +1474,7 @@ class Program_weight_tensor_parameter_122: class Program_weight_tensor_parameter_123: name = "parameter_123" + original_name = "batch_norm2d_59.b_0" shape = [192] dtype = "float32" min_val = float("-0.937606") @@ -1362,6 +1486,7 @@ class Program_weight_tensor_parameter_123: class Program_weight_tensor_parameter_124: name = "parameter_124" + original_name = "batch_norm2d_59.w_0" shape = [192] dtype = "float32" min_val = float("0.364889") @@ -1373,6 +1498,7 @@ class Program_weight_tensor_parameter_124: class Program_weight_tensor_parameter_125: name = "parameter_125" + original_name = "batch_norm2d_59.w_2" shape = [192] dtype = "float32" min_val = float("0.00674709") @@ -1384,6 +1510,7 @@ class Program_weight_tensor_parameter_125: class Program_weight_tensor_parameter_126: name = "parameter_126" + original_name = "batch_norm2d_59.w_1" shape = [192] dtype = "float32" min_val = float("-0.336115") @@ -1395,6 +1522,7 @@ class Program_weight_tensor_parameter_126: class Program_weight_tensor_parameter_127: name = "parameter_127" + original_name = "conv2d_59.w_0" shape = [192, 192, 1, 1] dtype = "float32" min_val = float("-0.226788") @@ -1406,6 +1534,7 @@ class Program_weight_tensor_parameter_127: class Program_weight_tensor_parameter_128: name = "parameter_128" + original_name = "batch_norm2d_58.b_0" shape = [192] dtype = "float32" min_val = float("-2.09417") @@ -1417,6 +1546,7 @@ class Program_weight_tensor_parameter_128: class Program_weight_tensor_parameter_129: name = "parameter_129" + original_name = "batch_norm2d_58.w_0" shape = [192] dtype = "float32" min_val = float("0.271442") @@ -1428,6 +1558,7 @@ class Program_weight_tensor_parameter_129: class Program_weight_tensor_parameter_130: name = "parameter_130" + original_name = "batch_norm2d_58.w_2" shape = [192] dtype = "float32" min_val = float("0.0170696") @@ -1439,6 +1570,7 @@ class Program_weight_tensor_parameter_130: class Program_weight_tensor_parameter_131: name = "parameter_131" + original_name = "batch_norm2d_58.w_1" shape = [192] dtype = "float32" min_val = float("-1.34848") @@ -1450,6 +1582,7 @@ class Program_weight_tensor_parameter_131: class Program_weight_tensor_parameter_132: name = "parameter_132" + original_name = "conv2d_58.w_0" shape = [192, 1, 5, 5] dtype = "float32" min_val = float("-0.204048") @@ -1461,6 +1594,7 @@ class Program_weight_tensor_parameter_132: class Program_weight_tensor_parameter_133: name = "parameter_133" + original_name = "batch_norm2d_57.b_0" shape = [192] dtype = "float32" min_val = float("-1.65647") @@ -1472,6 +1606,7 @@ class Program_weight_tensor_parameter_133: class Program_weight_tensor_parameter_134: name = "parameter_134" + original_name = "batch_norm2d_57.w_0" shape = [192] dtype = "float32" min_val = float("0.286769") @@ -1483,6 +1618,7 @@ class Program_weight_tensor_parameter_134: class Program_weight_tensor_parameter_135: name = "parameter_135" + original_name = "batch_norm2d_57.w_2" shape = [192] dtype = "float32" min_val = float("0.00697477") @@ -1494,6 +1630,7 @@ class Program_weight_tensor_parameter_135: class Program_weight_tensor_parameter_136: name = "parameter_136" + original_name = "batch_norm2d_57.w_1" shape = [192] dtype = "float32" min_val = float("-0.555885") @@ -1505,6 +1642,7 @@ class Program_weight_tensor_parameter_136: class Program_weight_tensor_parameter_137: name = "parameter_137" + original_name = "conv2d_57.w_0" shape = [192, 192, 1, 1] dtype = "float32" min_val = float("-0.201031") @@ -1516,6 +1654,7 @@ class Program_weight_tensor_parameter_137: class Program_weight_tensor_parameter_138: name = "parameter_138" + original_name = "batch_norm2d_56.b_0" shape = [192] dtype = "float32" min_val = float("-2.5939") @@ -1527,6 +1666,7 @@ class Program_weight_tensor_parameter_138: class Program_weight_tensor_parameter_139: name = "parameter_139" + original_name = "batch_norm2d_56.w_0" shape = [192] dtype = "float32" min_val = float("0.308342") @@ -1538,6 +1678,7 @@ class Program_weight_tensor_parameter_139: class Program_weight_tensor_parameter_140: name = "parameter_140" + original_name = "batch_norm2d_56.w_2" shape = [192] dtype = "float32" min_val = float("0.00858353") @@ -1549,6 +1690,7 @@ class Program_weight_tensor_parameter_140: class Program_weight_tensor_parameter_141: name = "parameter_141" + original_name = "batch_norm2d_56.w_1" shape = [192] dtype = "float32" min_val = float("-0.32159") @@ -1560,6 +1702,7 @@ class Program_weight_tensor_parameter_141: class Program_weight_tensor_parameter_142: name = "parameter_142" + original_name = "conv2d_56.w_0" shape = [192, 1, 5, 5] dtype = "float32" min_val = float("-0.280539") @@ -1571,6 +1714,7 @@ class Program_weight_tensor_parameter_142: class Program_weight_tensor_parameter_143: name = "parameter_143" + original_name = "batch_norm2d_55.b_0" shape = [192] dtype = "float32" min_val = float("-1.06875") @@ -1582,6 +1726,7 @@ class Program_weight_tensor_parameter_143: class Program_weight_tensor_parameter_144: name = "parameter_144" + original_name = "batch_norm2d_55.w_0" shape = [192] dtype = "float32" min_val = float("0.466195") @@ -1593,6 +1738,7 @@ class Program_weight_tensor_parameter_144: class Program_weight_tensor_parameter_145: name = "parameter_145" + original_name = "batch_norm2d_55.w_2" shape = [192] dtype = "float32" min_val = float("0.00758869") @@ -1604,6 +1750,7 @@ class Program_weight_tensor_parameter_145: class Program_weight_tensor_parameter_146: name = "parameter_146" + original_name = "batch_norm2d_55.w_1" shape = [192] dtype = "float32" min_val = float("-0.455754") @@ -1615,6 +1762,7 @@ class Program_weight_tensor_parameter_146: class Program_weight_tensor_parameter_147: name = "parameter_147" + original_name = "conv2d_55.w_0" shape = [192, 192, 1, 1] dtype = "float32" min_val = float("-0.211669") @@ -1626,6 +1774,7 @@ class Program_weight_tensor_parameter_147: class Program_weight_tensor_parameter_148: name = "parameter_148" + original_name = "batch_norm2d_54.b_0" shape = [192] dtype = "float32" min_val = float("-3.19724") @@ -1637,6 +1786,7 @@ class Program_weight_tensor_parameter_148: class Program_weight_tensor_parameter_149: name = "parameter_149" + original_name = "batch_norm2d_54.w_0" shape = [192] dtype = "float32" min_val = float("0.336724") @@ -1648,6 +1798,7 @@ class Program_weight_tensor_parameter_149: class Program_weight_tensor_parameter_150: name = "parameter_150" + original_name = "batch_norm2d_54.w_2" shape = [192] dtype = "float32" min_val = float("0.0183511") @@ -1659,6 +1810,7 @@ class Program_weight_tensor_parameter_150: class Program_weight_tensor_parameter_151: name = "parameter_151" + original_name = "batch_norm2d_54.w_1" shape = [192] dtype = "float32" min_val = float("-0.573257") @@ -1670,6 +1822,7 @@ class Program_weight_tensor_parameter_151: class Program_weight_tensor_parameter_152: name = "parameter_152" + original_name = "conv2d_54.w_0" shape = [192, 1, 5, 5] dtype = "float32" min_val = float("-0.293574") @@ -1681,6 +1834,7 @@ class Program_weight_tensor_parameter_152: class Program_weight_tensor_parameter_153: name = "parameter_153" + original_name = "batch_norm2d_53.b_0" shape = [192] dtype = "float32" min_val = float("-1.15804") @@ -1692,6 +1846,7 @@ class Program_weight_tensor_parameter_153: class Program_weight_tensor_parameter_154: name = "parameter_154" + original_name = "batch_norm2d_53.w_0" shape = [192] dtype = "float32" min_val = float("0.464899") @@ -1703,6 +1858,7 @@ class Program_weight_tensor_parameter_154: class Program_weight_tensor_parameter_155: name = "parameter_155" + original_name = "batch_norm2d_53.w_2" shape = [192] dtype = "float32" min_val = float("0.0104417") @@ -1714,6 +1870,7 @@ class Program_weight_tensor_parameter_155: class Program_weight_tensor_parameter_156: name = "parameter_156" + original_name = "batch_norm2d_53.w_1" shape = [192] dtype = "float32" min_val = float("-1.01624") @@ -1725,6 +1882,7 @@ class Program_weight_tensor_parameter_156: class Program_weight_tensor_parameter_157: name = "parameter_157" + original_name = "conv2d_53.w_0" shape = [192, 192, 1, 1] dtype = "float32" min_val = float("-0.339683") @@ -1736,6 +1894,7 @@ class Program_weight_tensor_parameter_157: class Program_weight_tensor_parameter_158: name = "parameter_158" + original_name = "batch_norm2d_52.b_0" shape = [192] dtype = "float32" min_val = float("-4.67339") @@ -1747,6 +1906,7 @@ class Program_weight_tensor_parameter_158: class Program_weight_tensor_parameter_159: name = "parameter_159" + original_name = "batch_norm2d_52.w_0" shape = [192] dtype = "float32" min_val = float("0.449954") @@ -1758,6 +1918,7 @@ class Program_weight_tensor_parameter_159: class Program_weight_tensor_parameter_160: name = "parameter_160" + original_name = "batch_norm2d_52.w_2" shape = [192] dtype = "float32" min_val = float("0.0258744") @@ -1769,6 +1930,7 @@ class Program_weight_tensor_parameter_160: class Program_weight_tensor_parameter_161: name = "parameter_161" + original_name = "batch_norm2d_52.w_1" shape = [192] dtype = "float32" min_val = float("-0.869984") @@ -1780,6 +1942,7 @@ class Program_weight_tensor_parameter_161: class Program_weight_tensor_parameter_162: name = "parameter_162" + original_name = "conv2d_52.w_0" shape = [192, 1, 5, 5] dtype = "float32" min_val = float("-0.374118") @@ -1791,6 +1954,7 @@ class Program_weight_tensor_parameter_162: class Program_weight_tensor_parameter_163: name = "parameter_163" + original_name = "batch_norm2d_51.b_0" shape = [192] dtype = "float32" min_val = float("-1.51992") @@ -1802,6 +1966,7 @@ class Program_weight_tensor_parameter_163: class Program_weight_tensor_parameter_164: name = "parameter_164" + original_name = "batch_norm2d_51.w_0" shape = [192] dtype = "float32" min_val = float("0.690793") @@ -1813,6 +1978,7 @@ class Program_weight_tensor_parameter_164: class Program_weight_tensor_parameter_165: name = "parameter_165" + original_name = "batch_norm2d_51.w_2" shape = [192] dtype = "float32" min_val = float("0.0596562") @@ -1824,6 +1990,7 @@ class Program_weight_tensor_parameter_165: class Program_weight_tensor_parameter_166: name = "parameter_166" + original_name = "batch_norm2d_51.w_1" shape = [192] dtype = "float32" min_val = float("-1.26098") @@ -1835,6 +2002,7 @@ class Program_weight_tensor_parameter_166: class Program_weight_tensor_parameter_167: name = "parameter_167" + original_name = "conv2d_51.w_0" shape = [192, 1024, 1, 1] dtype = "float32" min_val = float("-0.223709") @@ -1846,6 +2014,7 @@ class Program_weight_tensor_parameter_167: class Program_weight_tensor_parameter_168: name = "parameter_168" + original_name = "batch_norm2d_50.b_0" shape = [1024] dtype = "float32" min_val = float("-3.73301") @@ -1857,6 +2026,7 @@ class Program_weight_tensor_parameter_168: class Program_weight_tensor_parameter_169: name = "parameter_169" + original_name = "batch_norm2d_50.w_0" shape = [1024] dtype = "float32" min_val = float("0.0417816") @@ -1868,6 +2038,7 @@ class Program_weight_tensor_parameter_169: class Program_weight_tensor_parameter_170: name = "parameter_170" + original_name = "batch_norm2d_50.w_2" shape = [1024] dtype = "float32" min_val = float("5.60519e-45") @@ -1879,6 +2050,7 @@ class Program_weight_tensor_parameter_170: class Program_weight_tensor_parameter_171: name = "parameter_171" + original_name = "batch_norm2d_50.w_1" shape = [1024] dtype = "float32" min_val = float("-0.849446") @@ -1890,6 +2062,7 @@ class Program_weight_tensor_parameter_171: class Program_weight_tensor_parameter_172: name = "parameter_172" + original_name = "conv2d_50.w_0" shape = [1024, 512, 1, 1] dtype = "float32" min_val = float("-0.20057") @@ -1901,6 +2074,7 @@ class Program_weight_tensor_parameter_172: class Program_weight_tensor_parameter_173: name = "parameter_173" + original_name = "batch_norm2d_49.b_0" shape = [512] dtype = "float32" min_val = float("-5.20439") @@ -1912,6 +2086,7 @@ class Program_weight_tensor_parameter_173: class Program_weight_tensor_parameter_174: name = "parameter_174" + original_name = "batch_norm2d_49.w_0" shape = [512] dtype = "float32" min_val = float("0.67268") @@ -1923,6 +2098,7 @@ class Program_weight_tensor_parameter_174: class Program_weight_tensor_parameter_175: name = "parameter_175" + original_name = "batch_norm2d_49.w_2" shape = [512] dtype = "float32" min_val = float("0.0816975") @@ -1934,6 +2110,7 @@ class Program_weight_tensor_parameter_175: class Program_weight_tensor_parameter_176: name = "parameter_176" + original_name = "batch_norm2d_49.w_1" shape = [512] dtype = "float32" min_val = float("-1.44029") @@ -1945,6 +2122,7 @@ class Program_weight_tensor_parameter_176: class Program_weight_tensor_parameter_177: name = "parameter_177" + original_name = "conv2d_49.w_0" shape = [512, 2176, 1, 1] dtype = "float32" min_val = float("-0.402532") @@ -1956,6 +2134,7 @@ class Program_weight_tensor_parameter_177: class Program_weight_tensor_parameter_178: name = "parameter_178" + original_name = "batch_norm2d_48.b_0" shape = [192] dtype = "float32" min_val = float("-2.05521") @@ -1967,6 +2146,7 @@ class Program_weight_tensor_parameter_178: class Program_weight_tensor_parameter_179: name = "parameter_179" + original_name = "batch_norm2d_48.w_0" shape = [192] dtype = "float32" min_val = float("0.422052") @@ -1978,6 +2158,7 @@ class Program_weight_tensor_parameter_179: class Program_weight_tensor_parameter_180: name = "parameter_180" + original_name = "batch_norm2d_48.w_2" shape = [192] dtype = "float32" min_val = float("0.00345166") @@ -1989,6 +2170,7 @@ class Program_weight_tensor_parameter_180: class Program_weight_tensor_parameter_181: name = "parameter_181" + original_name = "batch_norm2d_48.w_1" shape = [192] dtype = "float32" min_val = float("-0.716531") @@ -2000,6 +2182,7 @@ class Program_weight_tensor_parameter_181: class Program_weight_tensor_parameter_182: name = "parameter_182" + original_name = "conv2d_48.w_0" shape = [192, 1, 5, 5] dtype = "float32" min_val = float("-0.230589") @@ -2011,6 +2194,7 @@ class Program_weight_tensor_parameter_182: class Program_weight_tensor_parameter_183: name = "parameter_183" + original_name = "batch_norm2d_47.b_0" shape = [192] dtype = "float32" min_val = float("-1.46055") @@ -2022,6 +2206,7 @@ class Program_weight_tensor_parameter_183: class Program_weight_tensor_parameter_184: name = "parameter_184" + original_name = "batch_norm2d_47.w_0" shape = [192] dtype = "float32" min_val = float("0.35268") @@ -2033,6 +2218,7 @@ class Program_weight_tensor_parameter_184: class Program_weight_tensor_parameter_185: name = "parameter_185" + original_name = "batch_norm2d_47.w_2" shape = [192] dtype = "float32" min_val = float("0.0107789") @@ -2044,6 +2230,7 @@ class Program_weight_tensor_parameter_185: class Program_weight_tensor_parameter_186: name = "parameter_186" + original_name = "batch_norm2d_47.w_1" shape = [192] dtype = "float32" min_val = float("-0.385797") @@ -2055,6 +2242,7 @@ class Program_weight_tensor_parameter_186: class Program_weight_tensor_parameter_187: name = "parameter_187" + original_name = "conv2d_47.w_0" shape = [192, 192, 1, 1] dtype = "float32" min_val = float("-0.454384") @@ -2066,6 +2254,7 @@ class Program_weight_tensor_parameter_187: class Program_weight_tensor_parameter_188: name = "parameter_188" + original_name = "batch_norm2d_46.b_0" shape = [192] dtype = "float32" min_val = float("-2.36204") @@ -2077,6 +2266,7 @@ class Program_weight_tensor_parameter_188: class Program_weight_tensor_parameter_189: name = "parameter_189" + original_name = "batch_norm2d_46.w_0" shape = [192] dtype = "float32" min_val = float("0.39148") @@ -2088,6 +2278,7 @@ class Program_weight_tensor_parameter_189: class Program_weight_tensor_parameter_190: name = "parameter_190" + original_name = "batch_norm2d_46.w_2" shape = [192] dtype = "float32" min_val = float("0.00806842") @@ -2099,6 +2290,7 @@ class Program_weight_tensor_parameter_190: class Program_weight_tensor_parameter_191: name = "parameter_191" + original_name = "batch_norm2d_46.w_1" shape = [192] dtype = "float32" min_val = float("-0.411915") @@ -2110,6 +2302,7 @@ class Program_weight_tensor_parameter_191: class Program_weight_tensor_parameter_192: name = "parameter_192" + original_name = "conv2d_46.w_0" shape = [192, 1, 5, 5] dtype = "float32" min_val = float("-0.223618") @@ -2121,6 +2314,7 @@ class Program_weight_tensor_parameter_192: class Program_weight_tensor_parameter_193: name = "parameter_193" + original_name = "batch_norm2d_45.b_0" shape = [192] dtype = "float32" min_val = float("-1.16163") @@ -2132,6 +2326,7 @@ class Program_weight_tensor_parameter_193: class Program_weight_tensor_parameter_194: name = "parameter_194" + original_name = "batch_norm2d_45.w_0" shape = [192] dtype = "float32" min_val = float("0.524949") @@ -2143,6 +2338,7 @@ class Program_weight_tensor_parameter_194: class Program_weight_tensor_parameter_195: name = "parameter_195" + original_name = "batch_norm2d_45.w_2" shape = [192] dtype = "float32" min_val = float("0.0125172") @@ -2154,6 +2350,7 @@ class Program_weight_tensor_parameter_195: class Program_weight_tensor_parameter_196: name = "parameter_196" + original_name = "batch_norm2d_45.w_1" shape = [192] dtype = "float32" min_val = float("-0.386308") @@ -2165,6 +2362,7 @@ class Program_weight_tensor_parameter_196: class Program_weight_tensor_parameter_197: name = "parameter_197" + original_name = "conv2d_45.w_0" shape = [192, 192, 1, 1] dtype = "float32" min_val = float("-0.203661") @@ -2176,6 +2374,7 @@ class Program_weight_tensor_parameter_197: class Program_weight_tensor_parameter_198: name = "parameter_198" + original_name = "batch_norm2d_44.b_0" shape = [192] dtype = "float32" min_val = float("-1.9807") @@ -2187,6 +2386,7 @@ class Program_weight_tensor_parameter_198: class Program_weight_tensor_parameter_199: name = "parameter_199" + original_name = "batch_norm2d_44.w_0" shape = [192] dtype = "float32" min_val = float("0.358479") @@ -2198,6 +2398,7 @@ class Program_weight_tensor_parameter_199: class Program_weight_tensor_parameter_200: name = "parameter_200" + original_name = "batch_norm2d_44.w_2" shape = [192] dtype = "float32" min_val = float("0.0196223") @@ -2209,6 +2410,7 @@ class Program_weight_tensor_parameter_200: class Program_weight_tensor_parameter_201: name = "parameter_201" + original_name = "batch_norm2d_44.w_1" shape = [192] dtype = "float32" min_val = float("-0.99819") @@ -2220,6 +2422,7 @@ class Program_weight_tensor_parameter_201: class Program_weight_tensor_parameter_202: name = "parameter_202" + original_name = "conv2d_44.w_0" shape = [192, 1, 5, 5] dtype = "float32" min_val = float("-0.261121") @@ -2231,6 +2434,7 @@ class Program_weight_tensor_parameter_202: class Program_weight_tensor_parameter_203: name = "parameter_203" + original_name = "batch_norm2d_43.b_0" shape = [192] dtype = "float32" min_val = float("-1.84882") @@ -2242,6 +2446,7 @@ class Program_weight_tensor_parameter_203: class Program_weight_tensor_parameter_204: name = "parameter_204" + original_name = "batch_norm2d_43.w_0" shape = [192] dtype = "float32" min_val = float("0.28487") @@ -2253,6 +2458,7 @@ class Program_weight_tensor_parameter_204: class Program_weight_tensor_parameter_205: name = "parameter_205" + original_name = "batch_norm2d_43.w_2" shape = [192] dtype = "float32" min_val = float("0.01478") @@ -2264,6 +2470,7 @@ class Program_weight_tensor_parameter_205: class Program_weight_tensor_parameter_206: name = "parameter_206" + original_name = "batch_norm2d_43.w_1" shape = [192] dtype = "float32" min_val = float("-0.672234") @@ -2275,6 +2482,7 @@ class Program_weight_tensor_parameter_206: class Program_weight_tensor_parameter_207: name = "parameter_207" + original_name = "conv2d_43.w_0" shape = [192, 192, 1, 1] dtype = "float32" min_val = float("-0.250961") @@ -2286,6 +2494,7 @@ class Program_weight_tensor_parameter_207: class Program_weight_tensor_parameter_208: name = "parameter_208" + original_name = "batch_norm2d_42.b_0" shape = [192] dtype = "float32" min_val = float("-2.2666") @@ -2297,6 +2506,7 @@ class Program_weight_tensor_parameter_208: class Program_weight_tensor_parameter_209: name = "parameter_209" + original_name = "batch_norm2d_42.w_0" shape = [192] dtype = "float32" min_val = float("0.383407") @@ -2308,6 +2518,7 @@ class Program_weight_tensor_parameter_209: class Program_weight_tensor_parameter_210: name = "parameter_210" + original_name = "batch_norm2d_42.w_2" shape = [192] dtype = "float32" min_val = float("0.0225413") @@ -2319,6 +2530,7 @@ class Program_weight_tensor_parameter_210: class Program_weight_tensor_parameter_211: name = "parameter_211" + original_name = "batch_norm2d_42.w_1" shape = [192] dtype = "float32" min_val = float("-1.98574") @@ -2330,6 +2542,7 @@ class Program_weight_tensor_parameter_211: class Program_weight_tensor_parameter_212: name = "parameter_212" + original_name = "conv2d_42.w_0" shape = [192, 1, 5, 5] dtype = "float32" min_val = float("-0.258795") @@ -2341,6 +2554,7 @@ class Program_weight_tensor_parameter_212: class Program_weight_tensor_parameter_213: name = "parameter_213" + original_name = "batch_norm2d_41.b_0" shape = [192] dtype = "float32" min_val = float("-2.226") @@ -2352,6 +2566,7 @@ class Program_weight_tensor_parameter_213: class Program_weight_tensor_parameter_214: name = "parameter_214" + original_name = "batch_norm2d_41.w_0" shape = [192] dtype = "float32" min_val = float("0.402497") @@ -2363,6 +2578,7 @@ class Program_weight_tensor_parameter_214: class Program_weight_tensor_parameter_215: name = "parameter_215" + original_name = "batch_norm2d_41.w_2" shape = [192] dtype = "float32" min_val = float("0.0156417") @@ -2374,6 +2590,7 @@ class Program_weight_tensor_parameter_215: class Program_weight_tensor_parameter_216: name = "parameter_216" + original_name = "batch_norm2d_41.w_1" shape = [192] dtype = "float32" min_val = float("-0.767177") @@ -2385,6 +2602,7 @@ class Program_weight_tensor_parameter_216: class Program_weight_tensor_parameter_217: name = "parameter_217" + original_name = "conv2d_41.w_0" shape = [192, 192, 1, 1] dtype = "float32" min_val = float("-0.323398") @@ -2396,6 +2614,7 @@ class Program_weight_tensor_parameter_217: class Program_weight_tensor_parameter_218: name = "parameter_218" + original_name = "batch_norm2d_40.b_0" shape = [192] dtype = "float32" min_val = float("-2.69867") @@ -2407,6 +2626,7 @@ class Program_weight_tensor_parameter_218: class Program_weight_tensor_parameter_219: name = "parameter_219" + original_name = "batch_norm2d_40.w_0" shape = [192] dtype = "float32" min_val = float("0.351054") @@ -2418,6 +2638,7 @@ class Program_weight_tensor_parameter_219: class Program_weight_tensor_parameter_220: name = "parameter_220" + original_name = "batch_norm2d_40.w_2" shape = [192] dtype = "float32" min_val = float("0.01948") @@ -2429,6 +2650,7 @@ class Program_weight_tensor_parameter_220: class Program_weight_tensor_parameter_221: name = "parameter_221" + original_name = "batch_norm2d_40.w_1" shape = [192] dtype = "float32" min_val = float("-0.456054") @@ -2440,6 +2662,7 @@ class Program_weight_tensor_parameter_221: class Program_weight_tensor_parameter_222: name = "parameter_222" + original_name = "conv2d_40.w_0" shape = [192, 1, 5, 5] dtype = "float32" min_val = float("-0.345438") @@ -2451,6 +2674,7 @@ class Program_weight_tensor_parameter_222: class Program_weight_tensor_parameter_223: name = "parameter_223" + original_name = "batch_norm2d_39.b_0" shape = [192] dtype = "float32" min_val = float("-1.11282") @@ -2462,6 +2686,7 @@ class Program_weight_tensor_parameter_223: class Program_weight_tensor_parameter_224: name = "parameter_224" + original_name = "batch_norm2d_39.w_0" shape = [192] dtype = "float32" min_val = float("0.639075") @@ -2473,6 +2698,7 @@ class Program_weight_tensor_parameter_224: class Program_weight_tensor_parameter_225: name = "parameter_225" + original_name = "batch_norm2d_39.w_2" shape = [192] dtype = "float32" min_val = float("0.0252763") @@ -2484,6 +2710,7 @@ class Program_weight_tensor_parameter_225: class Program_weight_tensor_parameter_226: name = "parameter_226" + original_name = "batch_norm2d_39.w_1" shape = [192] dtype = "float32" min_val = float("-1.10519") @@ -2495,6 +2722,7 @@ class Program_weight_tensor_parameter_226: class Program_weight_tensor_parameter_227: name = "parameter_227" + original_name = "conv2d_39.w_0" shape = [192, 192, 1, 1] dtype = "float32" min_val = float("-0.348284") @@ -2506,6 +2734,7 @@ class Program_weight_tensor_parameter_227: class Program_weight_tensor_parameter_228: name = "parameter_228" + original_name = "batch_norm2d_38.b_0" shape = [192] dtype = "float32" min_val = float("-2.9211") @@ -2517,6 +2746,7 @@ class Program_weight_tensor_parameter_228: class Program_weight_tensor_parameter_229: name = "parameter_229" + original_name = "batch_norm2d_38.w_0" shape = [192] dtype = "float32" min_val = float("0.436481") @@ -2528,6 +2758,7 @@ class Program_weight_tensor_parameter_229: class Program_weight_tensor_parameter_230: name = "parameter_230" + original_name = "batch_norm2d_38.w_2" shape = [192] dtype = "float32" min_val = float("0.0180733") @@ -2539,6 +2770,7 @@ class Program_weight_tensor_parameter_230: class Program_weight_tensor_parameter_231: name = "parameter_231" + original_name = "batch_norm2d_38.w_1" shape = [192] dtype = "float32" min_val = float("-0.489004") @@ -2550,6 +2782,7 @@ class Program_weight_tensor_parameter_231: class Program_weight_tensor_parameter_232: name = "parameter_232" + original_name = "conv2d_38.w_0" shape = [192, 1, 5, 5] dtype = "float32" min_val = float("-0.526221") @@ -2561,6 +2794,7 @@ class Program_weight_tensor_parameter_232: class Program_weight_tensor_parameter_233: name = "parameter_233" + original_name = "batch_norm2d_37.b_0" shape = [192] dtype = "float32" min_val = float("-1.26838") @@ -2572,6 +2806,7 @@ class Program_weight_tensor_parameter_233: class Program_weight_tensor_parameter_234: name = "parameter_234" + original_name = "batch_norm2d_37.w_0" shape = [192] dtype = "float32" min_val = float("0.534668") @@ -2583,6 +2818,7 @@ class Program_weight_tensor_parameter_234: class Program_weight_tensor_parameter_235: name = "parameter_235" + original_name = "batch_norm2d_37.w_2" shape = [192] dtype = "float32" min_val = float("0.0333058") @@ -2594,6 +2830,7 @@ class Program_weight_tensor_parameter_235: class Program_weight_tensor_parameter_236: name = "parameter_236" + original_name = "batch_norm2d_37.w_1" shape = [192] dtype = "float32" min_val = float("-0.613532") @@ -2605,6 +2842,7 @@ class Program_weight_tensor_parameter_236: class Program_weight_tensor_parameter_237: name = "parameter_237" + original_name = "conv2d_37.w_0" shape = [192, 1024, 1, 1] dtype = "float32" min_val = float("-0.417509") @@ -2616,6 +2854,7 @@ class Program_weight_tensor_parameter_237: class Program_weight_tensor_parameter_238: name = "parameter_238" + original_name = "batch_norm2d_36.b_0" shape = [1024] dtype = "float32" min_val = float("-5.02682") @@ -2627,6 +2866,7 @@ class Program_weight_tensor_parameter_238: class Program_weight_tensor_parameter_239: name = "parameter_239" + original_name = "batch_norm2d_36.w_0" shape = [1024] dtype = "float32" min_val = float("0.270514") @@ -2638,6 +2878,7 @@ class Program_weight_tensor_parameter_239: class Program_weight_tensor_parameter_240: name = "parameter_240" + original_name = "batch_norm2d_36.w_2" shape = [1024] dtype = "float32" min_val = float("0.109789") @@ -2649,6 +2890,7 @@ class Program_weight_tensor_parameter_240: class Program_weight_tensor_parameter_241: name = "parameter_241" + original_name = "batch_norm2d_36.w_1" shape = [1024] dtype = "float32" min_val = float("-1.88012") @@ -2660,6 +2902,7 @@ class Program_weight_tensor_parameter_241: class Program_weight_tensor_parameter_242: name = "parameter_242" + original_name = "conv2d_36.w_0" shape = [1024, 512, 1, 1] dtype = "float32" min_val = float("-0.4898") @@ -2671,6 +2914,7 @@ class Program_weight_tensor_parameter_242: class Program_weight_tensor_parameter_243: name = "parameter_243" + original_name = "batch_norm2d_35.b_0" shape = [512] dtype = "float32" min_val = float("-5.98402") @@ -2682,6 +2926,7 @@ class Program_weight_tensor_parameter_243: class Program_weight_tensor_parameter_244: name = "parameter_244" + original_name = "batch_norm2d_35.w_0" shape = [512] dtype = "float32" min_val = float("0.880964") @@ -2693,6 +2938,7 @@ class Program_weight_tensor_parameter_244: class Program_weight_tensor_parameter_245: name = "parameter_245" + original_name = "batch_norm2d_35.w_2" shape = [512] dtype = "float32" min_val = float("0.495448") @@ -2704,6 +2950,7 @@ class Program_weight_tensor_parameter_245: class Program_weight_tensor_parameter_246: name = "parameter_246" + original_name = "batch_norm2d_35.w_1" shape = [512] dtype = "float32" min_val = float("-2.87319") @@ -2715,6 +2962,7 @@ class Program_weight_tensor_parameter_246: class Program_weight_tensor_parameter_247: name = "parameter_247" + original_name = "conv2d_35.w_0" shape = [512, 1664, 1, 1] dtype = "float32" min_val = float("-0.434164") @@ -2726,6 +2974,7 @@ class Program_weight_tensor_parameter_247: class Program_weight_tensor_parameter_248: name = "parameter_248" + original_name = "batch_norm2d_34.b_0" shape = [192] dtype = "float32" min_val = float("-4.99742") @@ -2737,6 +2986,7 @@ class Program_weight_tensor_parameter_248: class Program_weight_tensor_parameter_249: name = "parameter_249" + original_name = "batch_norm2d_34.w_0" shape = [192] dtype = "float32" min_val = float("0.832045") @@ -2748,6 +2998,7 @@ class Program_weight_tensor_parameter_249: class Program_weight_tensor_parameter_250: name = "parameter_250" + original_name = "batch_norm2d_34.w_2" shape = [192] dtype = "float32" min_val = float("0.00507067") @@ -2759,6 +3010,7 @@ class Program_weight_tensor_parameter_250: class Program_weight_tensor_parameter_251: name = "parameter_251" + original_name = "batch_norm2d_34.w_1" shape = [192] dtype = "float32" min_val = float("-0.627523") @@ -2770,6 +3022,7 @@ class Program_weight_tensor_parameter_251: class Program_weight_tensor_parameter_252: name = "parameter_252" + original_name = "conv2d_34.w_0" shape = [192, 1, 5, 5] dtype = "float32" min_val = float("-0.172479") @@ -2781,6 +3034,7 @@ class Program_weight_tensor_parameter_252: class Program_weight_tensor_parameter_253: name = "parameter_253" + original_name = "batch_norm2d_33.b_0" shape = [192] dtype = "float32" min_val = float("-1.39675") @@ -2792,6 +3046,7 @@ class Program_weight_tensor_parameter_253: class Program_weight_tensor_parameter_254: name = "parameter_254" + original_name = "batch_norm2d_33.w_0" shape = [192] dtype = "float32" min_val = float("0.19963") @@ -2803,6 +3058,7 @@ class Program_weight_tensor_parameter_254: class Program_weight_tensor_parameter_255: name = "parameter_255" + original_name = "batch_norm2d_33.w_2" shape = [192] dtype = "float32" min_val = float("0.0109137") @@ -2814,6 +3070,7 @@ class Program_weight_tensor_parameter_255: class Program_weight_tensor_parameter_256: name = "parameter_256" + original_name = "batch_norm2d_33.w_1" shape = [192] dtype = "float32" min_val = float("-0.600008") @@ -2825,6 +3082,7 @@ class Program_weight_tensor_parameter_256: class Program_weight_tensor_parameter_257: name = "parameter_257" + original_name = "conv2d_33.w_0" shape = [192, 192, 1, 1] dtype = "float32" min_val = float("-0.241584") @@ -2836,6 +3094,7 @@ class Program_weight_tensor_parameter_257: class Program_weight_tensor_parameter_258: name = "parameter_258" + original_name = "batch_norm2d_32.b_0" shape = [192] dtype = "float32" min_val = float("-2.1379") @@ -2847,6 +3106,7 @@ class Program_weight_tensor_parameter_258: class Program_weight_tensor_parameter_259: name = "parameter_259" + original_name = "batch_norm2d_32.w_0" shape = [192] dtype = "float32" min_val = float("0.528826") @@ -2858,6 +3118,7 @@ class Program_weight_tensor_parameter_259: class Program_weight_tensor_parameter_260: name = "parameter_260" + original_name = "batch_norm2d_32.w_2" shape = [192] dtype = "float32" min_val = float("0.00510014") @@ -2869,6 +3130,7 @@ class Program_weight_tensor_parameter_260: class Program_weight_tensor_parameter_261: name = "parameter_261" + original_name = "batch_norm2d_32.w_1" shape = [192] dtype = "float32" min_val = float("-0.947804") @@ -2880,6 +3142,7 @@ class Program_weight_tensor_parameter_261: class Program_weight_tensor_parameter_262: name = "parameter_262" + original_name = "conv2d_32.w_0" shape = [192, 1, 5, 5] dtype = "float32" min_val = float("-0.159607") @@ -2891,6 +3154,7 @@ class Program_weight_tensor_parameter_262: class Program_weight_tensor_parameter_263: name = "parameter_263" + original_name = "batch_norm2d_31.b_0" shape = [192] dtype = "float32" min_val = float("-0.903667") @@ -2902,6 +3166,7 @@ class Program_weight_tensor_parameter_263: class Program_weight_tensor_parameter_264: name = "parameter_264" + original_name = "batch_norm2d_31.w_0" shape = [192] dtype = "float32" min_val = float("0.204626") @@ -2913,6 +3178,7 @@ class Program_weight_tensor_parameter_264: class Program_weight_tensor_parameter_265: name = "parameter_265" + original_name = "batch_norm2d_31.w_2" shape = [192] dtype = "float32" min_val = float("0.0188151") @@ -2924,6 +3190,7 @@ class Program_weight_tensor_parameter_265: class Program_weight_tensor_parameter_266: name = "parameter_266" + original_name = "batch_norm2d_31.w_1" shape = [192] dtype = "float32" min_val = float("-0.855231") @@ -2935,6 +3202,7 @@ class Program_weight_tensor_parameter_266: class Program_weight_tensor_parameter_267: name = "parameter_267" + original_name = "conv2d_31.w_0" shape = [192, 192, 1, 1] dtype = "float32" min_val = float("-0.258655") @@ -2946,6 +3214,7 @@ class Program_weight_tensor_parameter_267: class Program_weight_tensor_parameter_268: name = "parameter_268" + original_name = "batch_norm2d_30.b_0" shape = [192] dtype = "float32" min_val = float("-2.45539") @@ -2957,6 +3226,7 @@ class Program_weight_tensor_parameter_268: class Program_weight_tensor_parameter_269: name = "parameter_269" + original_name = "batch_norm2d_30.w_0" shape = [192] dtype = "float32" min_val = float("0.460274") @@ -2968,6 +3238,7 @@ class Program_weight_tensor_parameter_269: class Program_weight_tensor_parameter_270: name = "parameter_270" + original_name = "batch_norm2d_30.w_2" shape = [192] dtype = "float32" min_val = float("0.00957304") @@ -2979,6 +3250,7 @@ class Program_weight_tensor_parameter_270: class Program_weight_tensor_parameter_271: name = "parameter_271" + original_name = "batch_norm2d_30.w_1" shape = [192] dtype = "float32" min_val = float("-2.59197") @@ -2990,6 +3262,7 @@ class Program_weight_tensor_parameter_271: class Program_weight_tensor_parameter_272: name = "parameter_272" + original_name = "conv2d_30.w_0" shape = [192, 1, 5, 5] dtype = "float32" min_val = float("-0.237418") @@ -3001,6 +3274,7 @@ class Program_weight_tensor_parameter_272: class Program_weight_tensor_parameter_273: name = "parameter_273" + original_name = "batch_norm2d_29.b_0" shape = [192] dtype = "float32" min_val = float("-1.29569") @@ -3012,6 +3286,7 @@ class Program_weight_tensor_parameter_273: class Program_weight_tensor_parameter_274: name = "parameter_274" + original_name = "batch_norm2d_29.w_0" shape = [192] dtype = "float32" min_val = float("0.184573") @@ -3023,6 +3298,7 @@ class Program_weight_tensor_parameter_274: class Program_weight_tensor_parameter_275: name = "parameter_275" + original_name = "batch_norm2d_29.w_2" shape = [192] dtype = "float32" min_val = float("0.0265302") @@ -3034,6 +3310,7 @@ class Program_weight_tensor_parameter_275: class Program_weight_tensor_parameter_276: name = "parameter_276" + original_name = "batch_norm2d_29.w_1" shape = [192] dtype = "float32" min_val = float("-0.733232") @@ -3045,6 +3322,7 @@ class Program_weight_tensor_parameter_276: class Program_weight_tensor_parameter_277: name = "parameter_277" + original_name = "conv2d_29.w_0" shape = [192, 192, 1, 1] dtype = "float32" min_val = float("-0.322992") @@ -3056,6 +3334,7 @@ class Program_weight_tensor_parameter_277: class Program_weight_tensor_parameter_278: name = "parameter_278" + original_name = "batch_norm2d_28.b_0" shape = [192] dtype = "float32" min_val = float("-2.88415") @@ -3067,6 +3346,7 @@ class Program_weight_tensor_parameter_278: class Program_weight_tensor_parameter_279: name = "parameter_279" + original_name = "batch_norm2d_28.w_0" shape = [192] dtype = "float32" min_val = float("0.520259") @@ -3078,6 +3358,7 @@ class Program_weight_tensor_parameter_279: class Program_weight_tensor_parameter_280: name = "parameter_280" + original_name = "batch_norm2d_28.w_2" shape = [192] dtype = "float32" min_val = float("0.0138317") @@ -3089,6 +3370,7 @@ class Program_weight_tensor_parameter_280: class Program_weight_tensor_parameter_281: name = "parameter_281" + original_name = "batch_norm2d_28.w_1" shape = [192] dtype = "float32" min_val = float("-0.830472") @@ -3100,6 +3382,7 @@ class Program_weight_tensor_parameter_281: class Program_weight_tensor_parameter_282: name = "parameter_282" + original_name = "conv2d_28.w_0" shape = [192, 1, 5, 5] dtype = "float32" min_val = float("-0.232179") @@ -3111,6 +3394,7 @@ class Program_weight_tensor_parameter_282: class Program_weight_tensor_parameter_283: name = "parameter_283" + original_name = "batch_norm2d_27.b_0" shape = [192] dtype = "float32" min_val = float("-1.5225") @@ -3122,6 +3406,7 @@ class Program_weight_tensor_parameter_283: class Program_weight_tensor_parameter_284: name = "parameter_284" + original_name = "batch_norm2d_27.w_0" shape = [192] dtype = "float32" min_val = float("0.358956") @@ -3133,6 +3418,7 @@ class Program_weight_tensor_parameter_284: class Program_weight_tensor_parameter_285: name = "parameter_285" + original_name = "batch_norm2d_27.w_2" shape = [192] dtype = "float32" min_val = float("0.0463884") @@ -3144,6 +3430,7 @@ class Program_weight_tensor_parameter_285: class Program_weight_tensor_parameter_286: name = "parameter_286" + original_name = "batch_norm2d_27.w_1" shape = [192] dtype = "float32" min_val = float("-1.0478") @@ -3155,6 +3442,7 @@ class Program_weight_tensor_parameter_286: class Program_weight_tensor_parameter_287: name = "parameter_287" + original_name = "conv2d_27.w_0" shape = [192, 192, 1, 1] dtype = "float32" min_val = float("-0.280877") @@ -3166,6 +3454,7 @@ class Program_weight_tensor_parameter_287: class Program_weight_tensor_parameter_288: name = "parameter_288" + original_name = "batch_norm2d_26.b_0" shape = [192] dtype = "float32" min_val = float("-2.79101") @@ -3177,6 +3466,7 @@ class Program_weight_tensor_parameter_288: class Program_weight_tensor_parameter_289: name = "parameter_289" + original_name = "batch_norm2d_26.w_0" shape = [192] dtype = "float32" min_val = float("0.581983") @@ -3188,6 +3478,7 @@ class Program_weight_tensor_parameter_289: class Program_weight_tensor_parameter_290: name = "parameter_290" + original_name = "batch_norm2d_26.w_2" shape = [192] dtype = "float32" min_val = float("0.01811") @@ -3199,6 +3490,7 @@ class Program_weight_tensor_parameter_290: class Program_weight_tensor_parameter_291: name = "parameter_291" + original_name = "batch_norm2d_26.w_1" shape = [192] dtype = "float32" min_val = float("-0.858431") @@ -3210,6 +3502,7 @@ class Program_weight_tensor_parameter_291: class Program_weight_tensor_parameter_292: name = "parameter_292" + original_name = "conv2d_26.w_0" shape = [192, 1, 5, 5] dtype = "float32" min_val = float("-0.40775") @@ -3221,6 +3514,7 @@ class Program_weight_tensor_parameter_292: class Program_weight_tensor_parameter_293: name = "parameter_293" + original_name = "batch_norm2d_25.b_0" shape = [192] dtype = "float32" min_val = float("-1.25238") @@ -3232,6 +3526,7 @@ class Program_weight_tensor_parameter_293: class Program_weight_tensor_parameter_294: name = "parameter_294" + original_name = "batch_norm2d_25.w_0" shape = [192] dtype = "float32" min_val = float("0.552681") @@ -3243,6 +3538,7 @@ class Program_weight_tensor_parameter_294: class Program_weight_tensor_parameter_295: name = "parameter_295" + original_name = "batch_norm2d_25.w_2" shape = [192] dtype = "float32" min_val = float("0.0839605") @@ -3254,6 +3550,7 @@ class Program_weight_tensor_parameter_295: class Program_weight_tensor_parameter_296: name = "parameter_296" + original_name = "batch_norm2d_25.w_1" shape = [192] dtype = "float32" min_val = float("-2.04912") @@ -3265,6 +3562,7 @@ class Program_weight_tensor_parameter_296: class Program_weight_tensor_parameter_297: name = "parameter_297" + original_name = "conv2d_25.w_0" shape = [192, 192, 1, 1] dtype = "float32" min_val = float("-0.400584") @@ -3276,6 +3574,7 @@ class Program_weight_tensor_parameter_297: class Program_weight_tensor_parameter_298: name = "parameter_298" + original_name = "batch_norm2d_24.b_0" shape = [192] dtype = "float32" min_val = float("-2.66666") @@ -3287,6 +3586,7 @@ class Program_weight_tensor_parameter_298: class Program_weight_tensor_parameter_299: name = "parameter_299" + original_name = "batch_norm2d_24.w_0" shape = [192] dtype = "float32" min_val = float("0.925967") @@ -3298,6 +3598,7 @@ class Program_weight_tensor_parameter_299: class Program_weight_tensor_parameter_300: name = "parameter_300" + original_name = "batch_norm2d_24.w_2" shape = [192] dtype = "float32" min_val = float("0.0233999") @@ -3309,6 +3610,7 @@ class Program_weight_tensor_parameter_300: class Program_weight_tensor_parameter_301: name = "parameter_301" + original_name = "batch_norm2d_24.w_1" shape = [192] dtype = "float32" min_val = float("-0.581718") @@ -3320,6 +3622,7 @@ class Program_weight_tensor_parameter_301: class Program_weight_tensor_parameter_302: name = "parameter_302" + original_name = "conv2d_24.w_0" shape = [192, 1, 5, 5] dtype = "float32" min_val = float("-0.553781") @@ -3331,6 +3634,7 @@ class Program_weight_tensor_parameter_302: class Program_weight_tensor_parameter_303: name = "parameter_303" + original_name = "batch_norm2d_23.b_0" shape = [192] dtype = "float32" min_val = float("-1.37004") @@ -3342,6 +3646,7 @@ class Program_weight_tensor_parameter_303: class Program_weight_tensor_parameter_304: name = "parameter_304" + original_name = "batch_norm2d_23.w_0" shape = [192] dtype = "float32" min_val = float("0.633523") @@ -3353,6 +3658,7 @@ class Program_weight_tensor_parameter_304: class Program_weight_tensor_parameter_305: name = "parameter_305" + original_name = "batch_norm2d_23.w_2" shape = [192] dtype = "float32" min_val = float("0.436681") @@ -3364,6 +3670,7 @@ class Program_weight_tensor_parameter_305: class Program_weight_tensor_parameter_306: name = "parameter_306" + original_name = "batch_norm2d_23.w_1" shape = [192] dtype = "float32" min_val = float("-0.00776186") @@ -3375,6 +3682,7 @@ class Program_weight_tensor_parameter_306: class Program_weight_tensor_parameter_307: name = "parameter_307" + original_name = "conv2d_23.w_0" shape = [192, 512, 1, 1] dtype = "float32" min_val = float("-0.554423") @@ -3386,6 +3694,7 @@ class Program_weight_tensor_parameter_307: class Program_weight_tensor_parameter_308: name = "parameter_308" + original_name = "batch_norm2d_22.b_0" shape = [512] dtype = "float32" min_val = float("-0.00996946") @@ -3397,6 +3706,7 @@ class Program_weight_tensor_parameter_308: class Program_weight_tensor_parameter_309: name = "parameter_309" + original_name = "batch_norm2d_22.w_0" shape = [512] dtype = "float32" min_val = float("0.773066") @@ -3408,6 +3718,7 @@ class Program_weight_tensor_parameter_309: class Program_weight_tensor_parameter_310: name = "parameter_310" + original_name = "batch_norm2d_22.w_2" shape = [512] dtype = "float32" min_val = float("0.000145489") @@ -3419,6 +3730,7 @@ class Program_weight_tensor_parameter_310: class Program_weight_tensor_parameter_311: name = "parameter_311" + original_name = "batch_norm2d_22.w_1" shape = [512] dtype = "float32" min_val = float("-0.862025") @@ -3430,6 +3742,7 @@ class Program_weight_tensor_parameter_311: class Program_weight_tensor_parameter_312: name = "parameter_312" + original_name = "conv2d_22.w_0" shape = [512, 1, 3, 3] dtype = "float32" min_val = float("-0.169825") @@ -3441,6 +3754,7 @@ class Program_weight_tensor_parameter_312: class Program_weight_tensor_parameter_313: name = "parameter_313" + original_name = "batch_norm2d_21.b_0" shape = [512] dtype = "float32" min_val = float("-2.49442") @@ -3452,6 +3766,7 @@ class Program_weight_tensor_parameter_313: class Program_weight_tensor_parameter_314: name = "parameter_314" + original_name = "batch_norm2d_21.w_0" shape = [512] dtype = "float32" min_val = float("0.416714") @@ -3463,6 +3778,7 @@ class Program_weight_tensor_parameter_314: class Program_weight_tensor_parameter_315: name = "parameter_315" + original_name = "batch_norm2d_21.w_2" shape = [512] dtype = "float32" min_val = float("0.140153") @@ -3474,6 +3790,7 @@ class Program_weight_tensor_parameter_315: class Program_weight_tensor_parameter_316: name = "parameter_316" + original_name = "batch_norm2d_21.w_1" shape = [512] dtype = "float32" min_val = float("-1.49346") @@ -3485,6 +3802,7 @@ class Program_weight_tensor_parameter_316: class Program_weight_tensor_parameter_317: name = "parameter_317" + original_name = "conv2d_21.w_0" shape = [512, 256, 1, 1] dtype = "float32" min_val = float("-0.482404") @@ -3496,6 +3814,7 @@ class Program_weight_tensor_parameter_317: class Program_weight_tensor_parameter_318: name = "parameter_318" + original_name = "batch_norm2d_20.b_0" shape = [256] dtype = "float32" min_val = float("-3.49485") @@ -3507,6 +3826,7 @@ class Program_weight_tensor_parameter_318: class Program_weight_tensor_parameter_319: name = "parameter_319" + original_name = "batch_norm2d_20.w_0" shape = [256] dtype = "float32" min_val = float("0.915684") @@ -3518,6 +3838,7 @@ class Program_weight_tensor_parameter_319: class Program_weight_tensor_parameter_320: name = "parameter_320" + original_name = "batch_norm2d_20.w_2" shape = [256] dtype = "float32" min_val = float("0.496253") @@ -3529,6 +3850,7 @@ class Program_weight_tensor_parameter_320: class Program_weight_tensor_parameter_321: name = "parameter_321" + original_name = "batch_norm2d_20.w_1" shape = [256] dtype = "float32" min_val = float("-2.12194") @@ -3540,6 +3862,7 @@ class Program_weight_tensor_parameter_321: class Program_weight_tensor_parameter_322: name = "parameter_322" + original_name = "conv2d_20.w_0" shape = [256, 704, 1, 1] dtype = "float32" min_val = float("-0.698997") @@ -3551,6 +3874,7 @@ class Program_weight_tensor_parameter_322: class Program_weight_tensor_parameter_323: name = "parameter_323" + original_name = "batch_norm2d_19.b_0" shape = [96] dtype = "float32" min_val = float("-3.9611") @@ -3562,6 +3886,7 @@ class Program_weight_tensor_parameter_323: class Program_weight_tensor_parameter_324: name = "parameter_324" + original_name = "batch_norm2d_19.w_0" shape = [96] dtype = "float32" min_val = float("1.23599") @@ -3573,6 +3898,7 @@ class Program_weight_tensor_parameter_324: class Program_weight_tensor_parameter_325: name = "parameter_325" + original_name = "batch_norm2d_19.w_2" shape = [96] dtype = "float32" min_val = float("0.161442") @@ -3584,6 +3910,7 @@ class Program_weight_tensor_parameter_325: class Program_weight_tensor_parameter_326: name = "parameter_326" + original_name = "batch_norm2d_19.w_1" shape = [96] dtype = "float32" min_val = float("-2.16612") @@ -3595,6 +3922,7 @@ class Program_weight_tensor_parameter_326: class Program_weight_tensor_parameter_327: name = "parameter_327" + original_name = "conv2d_19.w_0" shape = [96, 96, 3, 3] dtype = "float32" min_val = float("-0.14703") @@ -3606,6 +3934,7 @@ class Program_weight_tensor_parameter_327: class Program_weight_tensor_parameter_328: name = "parameter_328" + original_name = "batch_norm2d_18.b_0" shape = [96] dtype = "float32" min_val = float("-4.84151") @@ -3617,6 +3946,7 @@ class Program_weight_tensor_parameter_328: class Program_weight_tensor_parameter_329: name = "parameter_329" + original_name = "batch_norm2d_18.w_0" shape = [96] dtype = "float32" min_val = float("0.904071") @@ -3628,6 +3958,7 @@ class Program_weight_tensor_parameter_329: class Program_weight_tensor_parameter_330: name = "parameter_330" + original_name = "batch_norm2d_18.w_2" shape = [96] dtype = "float32" min_val = float("0.244585") @@ -3639,6 +3970,7 @@ class Program_weight_tensor_parameter_330: class Program_weight_tensor_parameter_331: name = "parameter_331" + original_name = "batch_norm2d_18.w_1" shape = [96] dtype = "float32" min_val = float("-1.6295") @@ -3650,6 +3982,7 @@ class Program_weight_tensor_parameter_331: class Program_weight_tensor_parameter_332: name = "parameter_332" + original_name = "conv2d_18.w_0" shape = [96, 96, 3, 3] dtype = "float32" min_val = float("-0.151504") @@ -3661,6 +3994,7 @@ class Program_weight_tensor_parameter_332: class Program_weight_tensor_parameter_333: name = "parameter_333" + original_name = "batch_norm2d_17.b_0" shape = [96] dtype = "float32" min_val = float("-4.26368") @@ -3672,6 +4006,7 @@ class Program_weight_tensor_parameter_333: class Program_weight_tensor_parameter_334: name = "parameter_334" + original_name = "batch_norm2d_17.w_0" shape = [96] dtype = "float32" min_val = float("0.809512") @@ -3683,6 +4018,7 @@ class Program_weight_tensor_parameter_334: class Program_weight_tensor_parameter_335: name = "parameter_335" + original_name = "batch_norm2d_17.w_2" shape = [96] dtype = "float32" min_val = float("0.264339") @@ -3694,6 +4030,7 @@ class Program_weight_tensor_parameter_335: class Program_weight_tensor_parameter_336: name = "parameter_336" + original_name = "batch_norm2d_17.w_1" shape = [96] dtype = "float32" min_val = float("-3.58589") @@ -3705,6 +4042,7 @@ class Program_weight_tensor_parameter_336: class Program_weight_tensor_parameter_337: name = "parameter_337" + original_name = "conv2d_17.w_0" shape = [96, 96, 3, 3] dtype = "float32" min_val = float("-0.221973") @@ -3716,6 +4054,7 @@ class Program_weight_tensor_parameter_337: class Program_weight_tensor_parameter_338: name = "parameter_338" + original_name = "batch_norm2d_16.b_0" shape = [96] dtype = "float32" min_val = float("-3.91405") @@ -3727,6 +4066,7 @@ class Program_weight_tensor_parameter_338: class Program_weight_tensor_parameter_339: name = "parameter_339" + original_name = "batch_norm2d_16.w_0" shape = [96] dtype = "float32" min_val = float("1.11039") @@ -3738,6 +4078,7 @@ class Program_weight_tensor_parameter_339: class Program_weight_tensor_parameter_340: name = "parameter_340" + original_name = "batch_norm2d_16.w_2" shape = [96] dtype = "float32" min_val = float("0.553601") @@ -3749,6 +4090,7 @@ class Program_weight_tensor_parameter_340: class Program_weight_tensor_parameter_341: name = "parameter_341" + original_name = "batch_norm2d_16.w_1" shape = [96] dtype = "float32" min_val = float("-4.09024") @@ -3760,6 +4102,7 @@ class Program_weight_tensor_parameter_341: class Program_weight_tensor_parameter_342: name = "parameter_342" + original_name = "conv2d_16.w_0" shape = [96, 96, 3, 3] dtype = "float32" min_val = float("-0.213268") @@ -3771,6 +4114,7 @@ class Program_weight_tensor_parameter_342: class Program_weight_tensor_parameter_343: name = "parameter_343" + original_name = "batch_norm2d_15.b_0" shape = [96] dtype = "float32" min_val = float("-8.29257") @@ -3782,6 +4126,7 @@ class Program_weight_tensor_parameter_343: class Program_weight_tensor_parameter_344: name = "parameter_344" + original_name = "batch_norm2d_15.w_0" shape = [96] dtype = "float32" min_val = float("1.1315") @@ -3793,6 +4138,7 @@ class Program_weight_tensor_parameter_344: class Program_weight_tensor_parameter_345: name = "parameter_345" + original_name = "batch_norm2d_15.w_2" shape = [96] dtype = "float32" min_val = float("0.71146") @@ -3804,6 +4150,7 @@ class Program_weight_tensor_parameter_345: class Program_weight_tensor_parameter_346: name = "parameter_346" + original_name = "batch_norm2d_15.w_1" shape = [96] dtype = "float32" min_val = float("-5.39858") @@ -3815,6 +4162,7 @@ class Program_weight_tensor_parameter_346: class Program_weight_tensor_parameter_347: name = "parameter_347" + original_name = "conv2d_15.w_0" shape = [96, 96, 3, 3] dtype = "float32" min_val = float("-0.358326") @@ -3826,6 +4174,7 @@ class Program_weight_tensor_parameter_347: class Program_weight_tensor_parameter_348: name = "parameter_348" + original_name = "batch_norm2d_14.b_0" shape = [96] dtype = "float32" min_val = float("-5.24145") @@ -3837,6 +4186,7 @@ class Program_weight_tensor_parameter_348: class Program_weight_tensor_parameter_349: name = "parameter_349" + original_name = "batch_norm2d_14.w_0" shape = [96] dtype = "float32" min_val = float("1.18748") @@ -3848,6 +4198,7 @@ class Program_weight_tensor_parameter_349: class Program_weight_tensor_parameter_350: name = "parameter_350" + original_name = "batch_norm2d_14.w_2" shape = [96] dtype = "float32" min_val = float("1.29853") @@ -3859,6 +4210,7 @@ class Program_weight_tensor_parameter_350: class Program_weight_tensor_parameter_351: name = "parameter_351" + original_name = "batch_norm2d_14.w_1" shape = [96] dtype = "float32" min_val = float("-3.27913") @@ -3870,6 +4222,7 @@ class Program_weight_tensor_parameter_351: class Program_weight_tensor_parameter_352: name = "parameter_352" + original_name = "conv2d_14.w_0" shape = [96, 128, 3, 3] dtype = "float32" min_val = float("-0.400767") @@ -3881,6 +4234,7 @@ class Program_weight_tensor_parameter_352: class Program_weight_tensor_parameter_353: name = "parameter_353" + original_name = "batch_norm2d_13.b_0" shape = [128] dtype = "float32" min_val = float("-2.79295") @@ -3892,6 +4246,7 @@ class Program_weight_tensor_parameter_353: class Program_weight_tensor_parameter_354: name = "parameter_354" + original_name = "batch_norm2d_13.w_0" shape = [128] dtype = "float32" min_val = float("1.10058") @@ -3903,6 +4258,7 @@ class Program_weight_tensor_parameter_354: class Program_weight_tensor_parameter_355: name = "parameter_355" + original_name = "batch_norm2d_13.w_2" shape = [128] dtype = "float32" min_val = float("0.00512487") @@ -3914,6 +4270,7 @@ class Program_weight_tensor_parameter_355: class Program_weight_tensor_parameter_356: name = "parameter_356" + original_name = "batch_norm2d_13.w_1" shape = [128] dtype = "float32" min_val = float("-1.7968") @@ -3925,6 +4282,7 @@ class Program_weight_tensor_parameter_356: class Program_weight_tensor_parameter_357: name = "parameter_357" + original_name = "conv2d_13.w_0" shape = [128, 1, 3, 3] dtype = "float32" min_val = float("-0.236893") @@ -3936,6 +4294,7 @@ class Program_weight_tensor_parameter_357: class Program_weight_tensor_parameter_358: name = "parameter_358" + original_name = "batch_norm2d_12.b_0" shape = [128] dtype = "float32" min_val = float("-2.23026") @@ -3947,6 +4306,7 @@ class Program_weight_tensor_parameter_358: class Program_weight_tensor_parameter_359: name = "parameter_359" + original_name = "batch_norm2d_12.w_0" shape = [128] dtype = "float32" min_val = float("0.527108") @@ -3958,6 +4318,7 @@ class Program_weight_tensor_parameter_359: class Program_weight_tensor_parameter_360: name = "parameter_360" + original_name = "batch_norm2d_12.w_2" shape = [128] dtype = "float32" min_val = float("0.353561") @@ -3969,6 +4330,7 @@ class Program_weight_tensor_parameter_360: class Program_weight_tensor_parameter_361: name = "parameter_361" + original_name = "batch_norm2d_12.w_1" shape = [128] dtype = "float32" min_val = float("-3.60869") @@ -3980,6 +4342,7 @@ class Program_weight_tensor_parameter_361: class Program_weight_tensor_parameter_362: name = "parameter_362" + original_name = "conv2d_12.w_0" shape = [128, 64, 1, 1] dtype = "float32" min_val = float("-0.77328") @@ -3991,6 +4354,7 @@ class Program_weight_tensor_parameter_362: class Program_weight_tensor_parameter_363: name = "parameter_363" + original_name = "batch_norm2d_11.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -4000,6 +4364,7 @@ class Program_weight_tensor_parameter_363: class Program_weight_tensor_parameter_364: name = "parameter_364" + original_name = "batch_norm2d_11.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -4009,6 +4374,7 @@ class Program_weight_tensor_parameter_364: class Program_weight_tensor_parameter_365: name = "parameter_365" + original_name = "batch_norm2d_11.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -4018,6 +4384,7 @@ class Program_weight_tensor_parameter_365: class Program_weight_tensor_parameter_366: name = "parameter_366" + original_name = "batch_norm2d_11.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -4027,6 +4394,7 @@ class Program_weight_tensor_parameter_366: class Program_weight_tensor_parameter_367: name = "parameter_367" + original_name = "conv2d_11.w_0" shape = [64, 336, 1, 1] dtype = "float32" min_val = float("-0.728813") @@ -4038,6 +4406,7 @@ class Program_weight_tensor_parameter_367: class Program_weight_tensor_parameter_368: name = "parameter_368" + original_name = "batch_norm2d_10.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -4047,6 +4416,7 @@ class Program_weight_tensor_parameter_368: class Program_weight_tensor_parameter_369: name = "parameter_369" + original_name = "batch_norm2d_10.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -4056,6 +4426,7 @@ class Program_weight_tensor_parameter_369: class Program_weight_tensor_parameter_370: name = "parameter_370" + original_name = "batch_norm2d_10.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -4065,6 +4436,7 @@ class Program_weight_tensor_parameter_370: class Program_weight_tensor_parameter_371: name = "parameter_371" + original_name = "batch_norm2d_10.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -4074,6 +4446,7 @@ class Program_weight_tensor_parameter_371: class Program_weight_tensor_parameter_372: name = "parameter_372" + original_name = "conv2d_10.w_0" shape = [48, 48, 3, 3] dtype = "float32" min_val = float("-0.199324") @@ -4085,6 +4458,7 @@ class Program_weight_tensor_parameter_372: class Program_weight_tensor_parameter_373: name = "parameter_373" + original_name = "batch_norm2d_9.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -4094,6 +4468,7 @@ class Program_weight_tensor_parameter_373: class Program_weight_tensor_parameter_374: name = "parameter_374" + original_name = "batch_norm2d_9.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -4103,6 +4478,7 @@ class Program_weight_tensor_parameter_374: class Program_weight_tensor_parameter_375: name = "parameter_375" + original_name = "batch_norm2d_9.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -4112,6 +4488,7 @@ class Program_weight_tensor_parameter_375: class Program_weight_tensor_parameter_376: name = "parameter_376" + original_name = "batch_norm2d_9.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -4121,6 +4498,7 @@ class Program_weight_tensor_parameter_376: class Program_weight_tensor_parameter_377: name = "parameter_377" + original_name = "conv2d_9.w_0" shape = [48, 48, 3, 3] dtype = "float32" min_val = float("-0.317018") @@ -4132,6 +4510,7 @@ class Program_weight_tensor_parameter_377: class Program_weight_tensor_parameter_378: name = "parameter_378" + original_name = "batch_norm2d_8.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -4141,6 +4520,7 @@ class Program_weight_tensor_parameter_378: class Program_weight_tensor_parameter_379: name = "parameter_379" + original_name = "batch_norm2d_8.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -4150,6 +4530,7 @@ class Program_weight_tensor_parameter_379: class Program_weight_tensor_parameter_380: name = "parameter_380" + original_name = "batch_norm2d_8.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -4159,6 +4540,7 @@ class Program_weight_tensor_parameter_380: class Program_weight_tensor_parameter_381: name = "parameter_381" + original_name = "batch_norm2d_8.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -4168,6 +4550,7 @@ class Program_weight_tensor_parameter_381: class Program_weight_tensor_parameter_382: name = "parameter_382" + original_name = "conv2d_8.w_0" shape = [48, 48, 3, 3] dtype = "float32" min_val = float("-0.331156") @@ -4179,6 +4562,7 @@ class Program_weight_tensor_parameter_382: class Program_weight_tensor_parameter_383: name = "parameter_383" + original_name = "batch_norm2d_7.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -4188,6 +4572,7 @@ class Program_weight_tensor_parameter_383: class Program_weight_tensor_parameter_384: name = "parameter_384" + original_name = "batch_norm2d_7.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -4197,6 +4582,7 @@ class Program_weight_tensor_parameter_384: class Program_weight_tensor_parameter_385: name = "parameter_385" + original_name = "batch_norm2d_7.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -4206,6 +4592,7 @@ class Program_weight_tensor_parameter_385: class Program_weight_tensor_parameter_386: name = "parameter_386" + original_name = "batch_norm2d_7.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -4215,6 +4602,7 @@ class Program_weight_tensor_parameter_386: class Program_weight_tensor_parameter_387: name = "parameter_387" + original_name = "conv2d_7.w_0" shape = [48, 48, 3, 3] dtype = "float32" min_val = float("-0.469539") @@ -4226,6 +4614,7 @@ class Program_weight_tensor_parameter_387: class Program_weight_tensor_parameter_388: name = "parameter_388" + original_name = "batch_norm2d_6.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -4235,6 +4624,7 @@ class Program_weight_tensor_parameter_388: class Program_weight_tensor_parameter_389: name = "parameter_389" + original_name = "batch_norm2d_6.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -4244,6 +4634,7 @@ class Program_weight_tensor_parameter_389: class Program_weight_tensor_parameter_390: name = "parameter_390" + original_name = "batch_norm2d_6.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -4253,6 +4644,7 @@ class Program_weight_tensor_parameter_390: class Program_weight_tensor_parameter_391: name = "parameter_391" + original_name = "batch_norm2d_6.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -4262,6 +4654,7 @@ class Program_weight_tensor_parameter_391: class Program_weight_tensor_parameter_392: name = "parameter_392" + original_name = "conv2d_6.w_0" shape = [48, 48, 3, 3] dtype = "float32" min_val = float("-0.495964") @@ -4273,6 +4666,7 @@ class Program_weight_tensor_parameter_392: class Program_weight_tensor_parameter_393: name = "parameter_393" + original_name = "batch_norm2d_5.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -4282,6 +4676,7 @@ class Program_weight_tensor_parameter_393: class Program_weight_tensor_parameter_394: name = "parameter_394" + original_name = "batch_norm2d_5.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -4291,6 +4686,7 @@ class Program_weight_tensor_parameter_394: class Program_weight_tensor_parameter_395: name = "parameter_395" + original_name = "batch_norm2d_5.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -4300,6 +4696,7 @@ class Program_weight_tensor_parameter_395: class Program_weight_tensor_parameter_396: name = "parameter_396" + original_name = "batch_norm2d_5.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -4309,6 +4706,7 @@ class Program_weight_tensor_parameter_396: class Program_weight_tensor_parameter_397: name = "parameter_397" + original_name = "conv2d_5.w_0" shape = [48, 48, 3, 3] dtype = "float32" min_val = float("-0.789162") @@ -4320,6 +4718,7 @@ class Program_weight_tensor_parameter_397: class Program_weight_tensor_parameter_398: name = "parameter_398" + original_name = "batch_norm2d_4.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -4329,6 +4728,7 @@ class Program_weight_tensor_parameter_398: class Program_weight_tensor_parameter_399: name = "parameter_399" + original_name = "batch_norm2d_4.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -4338,6 +4738,7 @@ class Program_weight_tensor_parameter_399: class Program_weight_tensor_parameter_400: name = "parameter_400" + original_name = "batch_norm2d_4.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -4347,6 +4748,7 @@ class Program_weight_tensor_parameter_400: class Program_weight_tensor_parameter_401: name = "parameter_401" + original_name = "batch_norm2d_4.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -4356,6 +4758,7 @@ class Program_weight_tensor_parameter_401: class Program_weight_tensor_parameter_402: name = "parameter_402" + original_name = "conv2d_4.w_0" shape = [48, 32, 1, 1] dtype = "float32" min_val = float("-0.711784") @@ -4367,6 +4770,7 @@ class Program_weight_tensor_parameter_402: class Program_weight_tensor_parameter_403: name = "parameter_403" + original_name = "batch_norm2d_3.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -4376,6 +4780,7 @@ class Program_weight_tensor_parameter_403: class Program_weight_tensor_parameter_404: name = "parameter_404" + original_name = "batch_norm2d_3.w_0" shape = [32] dtype = "float32" min_val = float("0") @@ -4385,6 +4790,7 @@ class Program_weight_tensor_parameter_404: class Program_weight_tensor_parameter_405: name = "parameter_405" + original_name = "batch_norm2d_3.w_2" shape = [32] dtype = "float32" min_val = float("0") @@ -4394,6 +4800,7 @@ class Program_weight_tensor_parameter_405: class Program_weight_tensor_parameter_406: name = "parameter_406" + original_name = "batch_norm2d_3.w_1" shape = [32] dtype = "float32" min_val = float("0") @@ -4403,6 +4810,7 @@ class Program_weight_tensor_parameter_406: class Program_weight_tensor_parameter_407: name = "parameter_407" + original_name = "conv2d_3.w_0" shape = [32, 64, 3, 3] dtype = "float32" min_val = float("-0.751864") @@ -4414,6 +4822,7 @@ class Program_weight_tensor_parameter_407: class Program_weight_tensor_parameter_408: name = "parameter_408" + original_name = "batch_norm2d_2.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -4423,6 +4832,7 @@ class Program_weight_tensor_parameter_408: class Program_weight_tensor_parameter_409: name = "parameter_409" + original_name = "batch_norm2d_2.w_0" shape = [32] dtype = "float32" min_val = float("0") @@ -4432,6 +4842,7 @@ class Program_weight_tensor_parameter_409: class Program_weight_tensor_parameter_410: name = "parameter_410" + original_name = "batch_norm2d_2.w_2" shape = [32] dtype = "float32" min_val = float("0") @@ -4441,6 +4852,7 @@ class Program_weight_tensor_parameter_410: class Program_weight_tensor_parameter_411: name = "parameter_411" + original_name = "batch_norm2d_2.w_1" shape = [32] dtype = "float32" min_val = float("0") @@ -4450,6 +4862,7 @@ class Program_weight_tensor_parameter_411: class Program_weight_tensor_parameter_412: name = "parameter_412" + original_name = "conv2d_2.w_0" shape = [32, 16, 2, 2] dtype = "float32" min_val = float("-0.356884") @@ -4461,6 +4874,7 @@ class Program_weight_tensor_parameter_412: class Program_weight_tensor_parameter_413: name = "parameter_413" + original_name = "batch_norm2d_1.b_0" shape = [16] dtype = "float32" min_val = float("0") @@ -4470,6 +4884,7 @@ class Program_weight_tensor_parameter_413: class Program_weight_tensor_parameter_414: name = "parameter_414" + original_name = "batch_norm2d_1.w_0" shape = [16] dtype = "float32" min_val = float("0") @@ -4479,6 +4894,7 @@ class Program_weight_tensor_parameter_414: class Program_weight_tensor_parameter_415: name = "parameter_415" + original_name = "batch_norm2d_1.w_2" shape = [16] dtype = "float32" min_val = float("0") @@ -4488,6 +4904,7 @@ class Program_weight_tensor_parameter_415: class Program_weight_tensor_parameter_416: name = "parameter_416" + original_name = "batch_norm2d_1.w_1" shape = [16] dtype = "float32" min_val = float("0") @@ -4497,6 +4914,7 @@ class Program_weight_tensor_parameter_416: class Program_weight_tensor_parameter_417: name = "parameter_417" + original_name = "conv2d_1.w_0" shape = [16, 32, 2, 2] dtype = "float32" min_val = float("-0.827632") @@ -4508,6 +4926,7 @@ class Program_weight_tensor_parameter_417: class Program_weight_tensor_parameter_418: name = "parameter_418" + original_name = "batch_norm2d_0.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -4517,6 +4936,7 @@ class Program_weight_tensor_parameter_418: class Program_weight_tensor_parameter_419: name = "parameter_419" + original_name = "batch_norm2d_0.w_0" shape = [32] dtype = "float32" min_val = float("0") @@ -4526,6 +4946,7 @@ class Program_weight_tensor_parameter_419: class Program_weight_tensor_parameter_420: name = "parameter_420" + original_name = "batch_norm2d_0.w_2" shape = [32] dtype = "float32" min_val = float("0") @@ -4535,6 +4956,7 @@ class Program_weight_tensor_parameter_420: class Program_weight_tensor_parameter_421: name = "parameter_421" + original_name = "batch_norm2d_0.w_1" shape = [32] dtype = "float32" min_val = float("0") @@ -4544,6 +4966,7 @@ class Program_weight_tensor_parameter_421: class Program_weight_tensor_parameter_422: name = "parameter_422" + original_name = "conv2d_0.w_0" shape = [32, 3, 3, 3] dtype = "float32" min_val = float("-1.10689") diff --git a/paddle_samples/PaddleX/PP-HGNetV2-B6_ML/subgraph_1/input_meta.py b/paddle_samples/PaddleX/PP-HGNetV2-B6_ML/subgraph_1/input_meta.py index b0fcc874c..1f12c26a9 100644 --- a/paddle_samples/PaddleX/PP-HGNetV2-B6_ML/subgraph_1/input_meta.py +++ b/paddle_samples/PaddleX/PP-HGNetV2-B6_ML/subgraph_1/input_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_data_0: name = "data_0" + original_name = "param_0" shape = [33, 768] dtype = "float32" min_val = float("-0.0865465") @@ -11,6 +12,7 @@ class Program_weight_tensor_data_0: class Program_weight_tensor_data_1: name = "data_1" + original_name = "var_0" shape = [8, 3, 448, 448] dtype = "float32" min_val = float("-2.1179") diff --git a/paddle_samples/PaddleX/PP-HGNetV2-B6_ML/subgraph_1/weight_meta.py b/paddle_samples/PaddleX/PP-HGNetV2-B6_ML/subgraph_1/weight_meta.py index 192a1f6e9..304417471 100644 --- a/paddle_samples/PaddleX/PP-HGNetV2-B6_ML/subgraph_1/weight_meta.py +++ b/paddle_samples/PaddleX/PP-HGNetV2-B6_ML/subgraph_1/weight_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_parameter_0: name = "parameter_0" + original_name = "conv2d_176.b_0" shape = [33] dtype = "float32" min_val = float("0") @@ -9,6 +10,7 @@ class Program_weight_tensor_parameter_0: class Program_weight_tensor_parameter_1: name = "parameter_1" + original_name = "conv2d_176.w_0" shape = [33, 768, 1, 1] dtype = "float32" min_val = float("-0.209522") @@ -20,6 +22,7 @@ class Program_weight_tensor_parameter_1: class Program_weight_tensor_parameter_2: name = "parameter_2" + original_name = "layer_norm_2.b_0" shape = [768] dtype = "float32" min_val = float("-0.00217718") @@ -31,6 +34,7 @@ class Program_weight_tensor_parameter_2: class Program_weight_tensor_parameter_3: name = "parameter_3" + original_name = "layer_norm_2.w_0" shape = [768] dtype = "float32" min_val = float("0.970031") @@ -42,6 +46,7 @@ class Program_weight_tensor_parameter_3: class Program_weight_tensor_parameter_4: name = "parameter_4" + original_name = "linear_10.b_0" shape = [768] dtype = "float32" min_val = float("-0.0251424") @@ -53,6 +58,7 @@ class Program_weight_tensor_parameter_4: class Program_weight_tensor_parameter_5: name = "parameter_5" + original_name = "linear_10.w_0" shape = [2048, 768] dtype = "float32" min_val = float("-0.0643041") @@ -64,6 +70,7 @@ class Program_weight_tensor_parameter_5: class Program_weight_tensor_parameter_6: name = "parameter_6" + original_name = "linear_9.b_0" shape = [2048] dtype = "float32" min_val = float("-0.00757047") @@ -75,6 +82,7 @@ class Program_weight_tensor_parameter_6: class Program_weight_tensor_parameter_7: name = "parameter_7" + original_name = "linear_9.w_0" shape = [768, 2048] dtype = "float32" min_val = float("-0.0578989") @@ -86,6 +94,7 @@ class Program_weight_tensor_parameter_7: class Program_weight_tensor_parameter_8: name = "parameter_8" + original_name = "layer_norm_1.b_0" shape = [768] dtype = "float32" min_val = float("-0.0317476") @@ -97,6 +106,7 @@ class Program_weight_tensor_parameter_8: class Program_weight_tensor_parameter_9: name = "parameter_9" + original_name = "layer_norm_1.w_0" shape = [768] dtype = "float32" min_val = float("0.993735") @@ -108,6 +118,7 @@ class Program_weight_tensor_parameter_9: class Program_weight_tensor_parameter_10: name = "parameter_10" + original_name = "linear_8.b_0" shape = [768] dtype = "float32" min_val = float("-0.0204261") @@ -119,6 +130,7 @@ class Program_weight_tensor_parameter_10: class Program_weight_tensor_parameter_11: name = "parameter_11" + original_name = "linear_8.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.0921004") @@ -130,6 +142,7 @@ class Program_weight_tensor_parameter_11: class Program_weight_tensor_parameter_12: name = "parameter_12" + original_name = "linear_7.b_0" shape = [768] dtype = "float32" min_val = float("-0.0054675") @@ -141,6 +154,7 @@ class Program_weight_tensor_parameter_12: class Program_weight_tensor_parameter_13: name = "parameter_13" + original_name = "linear_7.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.072349") @@ -152,6 +166,7 @@ class Program_weight_tensor_parameter_13: class Program_weight_tensor_parameter_14: name = "parameter_14" + original_name = "linear_6.b_0" shape = [768] dtype = "float32" min_val = float("-0.00843113") @@ -163,6 +178,7 @@ class Program_weight_tensor_parameter_14: class Program_weight_tensor_parameter_15: name = "parameter_15" + original_name = "linear_6.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.0726317") @@ -174,6 +190,7 @@ class Program_weight_tensor_parameter_15: class Program_weight_tensor_parameter_16: name = "parameter_16" + original_name = "linear_5.b_0" shape = [768] dtype = "float32" min_val = float("-0.0123917") @@ -185,6 +202,7 @@ class Program_weight_tensor_parameter_16: class Program_weight_tensor_parameter_17: name = "parameter_17" + original_name = "linear_5.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.0713264") @@ -196,6 +214,7 @@ class Program_weight_tensor_parameter_17: class Program_weight_tensor_parameter_18: name = "parameter_18" + original_name = "layer_norm_0.b_0" shape = [768] dtype = "float32" min_val = float("-0.0288016") @@ -207,6 +226,7 @@ class Program_weight_tensor_parameter_18: class Program_weight_tensor_parameter_19: name = "parameter_19" + original_name = "layer_norm_0.w_0" shape = [768] dtype = "float32" min_val = float("0.993175") @@ -218,6 +238,7 @@ class Program_weight_tensor_parameter_19: class Program_weight_tensor_parameter_20: name = "parameter_20" + original_name = "conv2d_175.b_0" shape = [768] dtype = "float32" min_val = float("-0.00494243") @@ -229,6 +250,7 @@ class Program_weight_tensor_parameter_20: class Program_weight_tensor_parameter_21: name = "parameter_21" + original_name = "conv2d_175.w_0" shape = [768, 2048, 1, 1] dtype = "float32" min_val = float("-0.152863") @@ -240,6 +262,7 @@ class Program_weight_tensor_parameter_21: class Program_weight_tensor_parameter_22: name = "parameter_22" + original_name = "conv2d_174.w_0" shape = [2048, 2048, 1, 1] dtype = "float32" min_val = float("-0.372811") @@ -251,6 +274,7 @@ class Program_weight_tensor_parameter_22: class Program_weight_tensor_parameter_23: name = "parameter_23" + original_name = "batch_norm2d_173.b_0" shape = [2048] dtype = "float32" min_val = float("-6.01464") @@ -262,6 +286,7 @@ class Program_weight_tensor_parameter_23: class Program_weight_tensor_parameter_24: name = "parameter_24" + original_name = "batch_norm2d_173.w_0" shape = [2048] dtype = "float32" min_val = float("0.133301") @@ -273,6 +298,7 @@ class Program_weight_tensor_parameter_24: class Program_weight_tensor_parameter_25: name = "parameter_25" + original_name = "batch_norm2d_173.w_2" shape = [2048] dtype = "float32" min_val = float("5.60519e-45") @@ -284,6 +310,7 @@ class Program_weight_tensor_parameter_25: class Program_weight_tensor_parameter_26: name = "parameter_26" + original_name = "batch_norm2d_173.w_1" shape = [2048] dtype = "float32" min_val = float("-1.41134") @@ -295,6 +322,7 @@ class Program_weight_tensor_parameter_26: class Program_weight_tensor_parameter_27: name = "parameter_27" + original_name = "conv2d_173.w_0" shape = [2048, 1024, 1, 1] dtype = "float32" min_val = float("-0.42585") @@ -306,6 +334,7 @@ class Program_weight_tensor_parameter_27: class Program_weight_tensor_parameter_28: name = "parameter_28" + original_name = "batch_norm2d_172.b_0" shape = [1024] dtype = "float32" min_val = float("-4.02861") @@ -317,6 +346,7 @@ class Program_weight_tensor_parameter_28: class Program_weight_tensor_parameter_29: name = "parameter_29" + original_name = "batch_norm2d_172.w_0" shape = [1024] dtype = "float32" min_val = float("0.325068") @@ -328,6 +358,7 @@ class Program_weight_tensor_parameter_29: class Program_weight_tensor_parameter_30: name = "parameter_30" + original_name = "batch_norm2d_172.w_2" shape = [1024] dtype = "float32" min_val = float("0.308111") @@ -339,6 +370,7 @@ class Program_weight_tensor_parameter_30: class Program_weight_tensor_parameter_31: name = "parameter_31" + original_name = "batch_norm2d_172.w_1" shape = [1024] dtype = "float32" min_val = float("-3.17776") @@ -350,6 +382,7 @@ class Program_weight_tensor_parameter_31: class Program_weight_tensor_parameter_32: name = "parameter_32" + original_name = "conv2d_172.w_0" shape = [1024, 6656, 1, 1] dtype = "float32" min_val = float("-0.327148") @@ -361,6 +394,7 @@ class Program_weight_tensor_parameter_32: class Program_weight_tensor_parameter_33: name = "parameter_33" + original_name = "batch_norm2d_171.b_0" shape = [768] dtype = "float32" min_val = float("-2.3486") @@ -372,6 +406,7 @@ class Program_weight_tensor_parameter_33: class Program_weight_tensor_parameter_34: name = "parameter_34" + original_name = "batch_norm2d_171.w_0" shape = [768] dtype = "float32" min_val = float("0.479371") @@ -383,6 +418,7 @@ class Program_weight_tensor_parameter_34: class Program_weight_tensor_parameter_35: name = "parameter_35" + original_name = "batch_norm2d_171.w_2" shape = [768] dtype = "float32" min_val = float("5.60519e-45") @@ -394,6 +430,7 @@ class Program_weight_tensor_parameter_35: class Program_weight_tensor_parameter_36: name = "parameter_36" + original_name = "batch_norm2d_171.w_1" shape = [768] dtype = "float32" min_val = float("-0.298986") @@ -405,6 +442,7 @@ class Program_weight_tensor_parameter_36: class Program_weight_tensor_parameter_37: name = "parameter_37" + original_name = "conv2d_171.w_0" shape = [768, 1, 5, 5] dtype = "float32" min_val = float("-0.251464") @@ -416,6 +454,7 @@ class Program_weight_tensor_parameter_37: class Program_weight_tensor_parameter_38: name = "parameter_38" + original_name = "batch_norm2d_170.b_0" shape = [768] dtype = "float32" min_val = float("-0.918817") @@ -427,6 +466,7 @@ class Program_weight_tensor_parameter_38: class Program_weight_tensor_parameter_39: name = "parameter_39" + original_name = "batch_norm2d_170.w_0" shape = [768] dtype = "float32" min_val = float("0.502767") @@ -438,6 +478,7 @@ class Program_weight_tensor_parameter_39: class Program_weight_tensor_parameter_40: name = "parameter_40" + original_name = "batch_norm2d_170.w_2" shape = [768] dtype = "float32" min_val = float("5.60519e-45") @@ -449,6 +490,7 @@ class Program_weight_tensor_parameter_40: class Program_weight_tensor_parameter_41: name = "parameter_41" + original_name = "batch_norm2d_170.w_1" shape = [768] dtype = "float32" min_val = float("-0.101915") @@ -460,6 +502,7 @@ class Program_weight_tensor_parameter_41: class Program_weight_tensor_parameter_42: name = "parameter_42" + original_name = "conv2d_170.w_0" shape = [768, 768, 1, 1] dtype = "float32" min_val = float("-0.323817") @@ -471,6 +514,7 @@ class Program_weight_tensor_parameter_42: class Program_weight_tensor_parameter_43: name = "parameter_43" + original_name = "batch_norm2d_169.b_0" shape = [768] dtype = "float32" min_val = float("-8.83608") @@ -482,6 +526,7 @@ class Program_weight_tensor_parameter_43: class Program_weight_tensor_parameter_44: name = "parameter_44" + original_name = "batch_norm2d_169.w_0" shape = [768] dtype = "float32" min_val = float("0.152653") @@ -493,6 +538,7 @@ class Program_weight_tensor_parameter_44: class Program_weight_tensor_parameter_45: name = "parameter_45" + original_name = "batch_norm2d_169.w_2" shape = [768] dtype = "float32" min_val = float("5.60519e-45") @@ -504,6 +550,7 @@ class Program_weight_tensor_parameter_45: class Program_weight_tensor_parameter_46: name = "parameter_46" + original_name = "batch_norm2d_169.w_1" shape = [768] dtype = "float32" min_val = float("-0.451977") @@ -515,6 +562,7 @@ class Program_weight_tensor_parameter_46: class Program_weight_tensor_parameter_47: name = "parameter_47" + original_name = "conv2d_169.w_0" shape = [768, 1, 5, 5] dtype = "float32" min_val = float("-0.205461") @@ -526,6 +574,7 @@ class Program_weight_tensor_parameter_47: class Program_weight_tensor_parameter_48: name = "parameter_48" + original_name = "batch_norm2d_168.b_0" shape = [768] dtype = "float32" min_val = float("-1.41417") @@ -537,6 +586,7 @@ class Program_weight_tensor_parameter_48: class Program_weight_tensor_parameter_49: name = "parameter_49" + original_name = "batch_norm2d_168.w_0" shape = [768] dtype = "float32" min_val = float("0.365643") @@ -548,6 +598,7 @@ class Program_weight_tensor_parameter_49: class Program_weight_tensor_parameter_50: name = "parameter_50" + original_name = "batch_norm2d_168.w_2" shape = [768] dtype = "float32" min_val = float("5.60519e-45") @@ -559,6 +610,7 @@ class Program_weight_tensor_parameter_50: class Program_weight_tensor_parameter_51: name = "parameter_51" + original_name = "batch_norm2d_168.w_1" shape = [768] dtype = "float32" min_val = float("-0.442412") @@ -570,6 +622,7 @@ class Program_weight_tensor_parameter_51: class Program_weight_tensor_parameter_52: name = "parameter_52" + original_name = "conv2d_168.w_0" shape = [768, 768, 1, 1] dtype = "float32" min_val = float("-0.287176") @@ -581,6 +634,7 @@ class Program_weight_tensor_parameter_52: class Program_weight_tensor_parameter_53: name = "parameter_53" + original_name = "batch_norm2d_167.b_0" shape = [768] dtype = "float32" min_val = float("-5.71936") @@ -592,6 +646,7 @@ class Program_weight_tensor_parameter_53: class Program_weight_tensor_parameter_54: name = "parameter_54" + original_name = "batch_norm2d_167.w_0" shape = [768] dtype = "float32" min_val = float("0.33474") @@ -603,6 +658,7 @@ class Program_weight_tensor_parameter_54: class Program_weight_tensor_parameter_55: name = "parameter_55" + original_name = "batch_norm2d_167.w_2" shape = [768] dtype = "float32" min_val = float("0.00184821") @@ -614,6 +670,7 @@ class Program_weight_tensor_parameter_55: class Program_weight_tensor_parameter_56: name = "parameter_56" + original_name = "batch_norm2d_167.w_1" shape = [768] dtype = "float32" min_val = float("-0.351071") @@ -625,6 +682,7 @@ class Program_weight_tensor_parameter_56: class Program_weight_tensor_parameter_57: name = "parameter_57" + original_name = "conv2d_167.w_0" shape = [768, 1, 5, 5] dtype = "float32" min_val = float("-0.198617") @@ -636,6 +694,7 @@ class Program_weight_tensor_parameter_57: class Program_weight_tensor_parameter_58: name = "parameter_58" + original_name = "batch_norm2d_166.b_0" shape = [768] dtype = "float32" min_val = float("-1.05679") @@ -647,6 +706,7 @@ class Program_weight_tensor_parameter_58: class Program_weight_tensor_parameter_59: name = "parameter_59" + original_name = "batch_norm2d_166.w_0" shape = [768] dtype = "float32" min_val = float("0.391581") @@ -658,6 +718,7 @@ class Program_weight_tensor_parameter_59: class Program_weight_tensor_parameter_60: name = "parameter_60" + original_name = "batch_norm2d_166.w_2" shape = [768] dtype = "float32" min_val = float("0.00171475") @@ -669,6 +730,7 @@ class Program_weight_tensor_parameter_60: class Program_weight_tensor_parameter_61: name = "parameter_61" + original_name = "batch_norm2d_166.w_1" shape = [768] dtype = "float32" min_val = float("-0.635231") @@ -680,6 +742,7 @@ class Program_weight_tensor_parameter_61: class Program_weight_tensor_parameter_62: name = "parameter_62" + original_name = "conv2d_166.w_0" shape = [768, 768, 1, 1] dtype = "float32" min_val = float("-0.151538") @@ -691,6 +754,7 @@ class Program_weight_tensor_parameter_62: class Program_weight_tensor_parameter_63: name = "parameter_63" + original_name = "batch_norm2d_165.b_0" shape = [768] dtype = "float32" min_val = float("-2.5448") @@ -702,6 +766,7 @@ class Program_weight_tensor_parameter_63: class Program_weight_tensor_parameter_64: name = "parameter_64" + original_name = "batch_norm2d_165.w_0" shape = [768] dtype = "float32" min_val = float("0.463124") @@ -713,6 +778,7 @@ class Program_weight_tensor_parameter_64: class Program_weight_tensor_parameter_65: name = "parameter_65" + original_name = "batch_norm2d_165.w_2" shape = [768] dtype = "float32" min_val = float("0.0010807") @@ -724,6 +790,7 @@ class Program_weight_tensor_parameter_65: class Program_weight_tensor_parameter_66: name = "parameter_66" + original_name = "batch_norm2d_165.w_1" shape = [768] dtype = "float32" min_val = float("-1.37708") @@ -735,6 +802,7 @@ class Program_weight_tensor_parameter_66: class Program_weight_tensor_parameter_67: name = "parameter_67" + original_name = "conv2d_165.w_0" shape = [768, 1, 5, 5] dtype = "float32" min_val = float("-0.265143") @@ -746,6 +814,7 @@ class Program_weight_tensor_parameter_67: class Program_weight_tensor_parameter_68: name = "parameter_68" + original_name = "batch_norm2d_164.b_0" shape = [768] dtype = "float32" min_val = float("-1.71842") @@ -757,6 +826,7 @@ class Program_weight_tensor_parameter_68: class Program_weight_tensor_parameter_69: name = "parameter_69" + original_name = "batch_norm2d_164.w_0" shape = [768] dtype = "float32" min_val = float("0.0397149") @@ -768,6 +838,7 @@ class Program_weight_tensor_parameter_69: class Program_weight_tensor_parameter_70: name = "parameter_70" + original_name = "batch_norm2d_164.w_2" shape = [768] dtype = "float32" min_val = float("0.0038165") @@ -779,6 +850,7 @@ class Program_weight_tensor_parameter_70: class Program_weight_tensor_parameter_71: name = "parameter_71" + original_name = "batch_norm2d_164.w_1" shape = [768] dtype = "float32" min_val = float("-1.02487") @@ -790,6 +862,7 @@ class Program_weight_tensor_parameter_71: class Program_weight_tensor_parameter_72: name = "parameter_72" + original_name = "conv2d_164.w_0" shape = [768, 768, 1, 1] dtype = "float32" min_val = float("-0.205006") @@ -801,6 +874,7 @@ class Program_weight_tensor_parameter_72: class Program_weight_tensor_parameter_73: name = "parameter_73" + original_name = "batch_norm2d_163.b_0" shape = [768] dtype = "float32" min_val = float("-2.94539") @@ -812,6 +886,7 @@ class Program_weight_tensor_parameter_73: class Program_weight_tensor_parameter_74: name = "parameter_74" + original_name = "batch_norm2d_163.w_0" shape = [768] dtype = "float32" min_val = float("0.363255") @@ -823,6 +898,7 @@ class Program_weight_tensor_parameter_74: class Program_weight_tensor_parameter_75: name = "parameter_75" + original_name = "batch_norm2d_163.w_2" shape = [768] dtype = "float32" min_val = float("0.00922154") @@ -834,6 +910,7 @@ class Program_weight_tensor_parameter_75: class Program_weight_tensor_parameter_76: name = "parameter_76" + original_name = "batch_norm2d_163.w_1" shape = [768] dtype = "float32" min_val = float("-0.438635") @@ -845,6 +922,7 @@ class Program_weight_tensor_parameter_76: class Program_weight_tensor_parameter_77: name = "parameter_77" + original_name = "conv2d_163.w_0" shape = [768, 1, 5, 5] dtype = "float32" min_val = float("-0.275653") @@ -856,6 +934,7 @@ class Program_weight_tensor_parameter_77: class Program_weight_tensor_parameter_78: name = "parameter_78" + original_name = "batch_norm2d_162.b_0" shape = [768] dtype = "float32" min_val = float("-1.30798") @@ -867,6 +946,7 @@ class Program_weight_tensor_parameter_78: class Program_weight_tensor_parameter_79: name = "parameter_79" + original_name = "batch_norm2d_162.w_0" shape = [768] dtype = "float32" min_val = float("0.406763") @@ -878,6 +958,7 @@ class Program_weight_tensor_parameter_79: class Program_weight_tensor_parameter_80: name = "parameter_80" + original_name = "batch_norm2d_162.w_2" shape = [768] dtype = "float32" min_val = float("0.0249607") @@ -889,6 +970,7 @@ class Program_weight_tensor_parameter_80: class Program_weight_tensor_parameter_81: name = "parameter_81" + original_name = "batch_norm2d_162.w_1" shape = [768] dtype = "float32" min_val = float("-1.95478") @@ -900,6 +982,7 @@ class Program_weight_tensor_parameter_81: class Program_weight_tensor_parameter_82: name = "parameter_82" + original_name = "conv2d_162.w_0" shape = [768, 768, 1, 1] dtype = "float32" min_val = float("-0.302541") @@ -911,6 +994,7 @@ class Program_weight_tensor_parameter_82: class Program_weight_tensor_parameter_83: name = "parameter_83" + original_name = "batch_norm2d_161.b_0" shape = [768] dtype = "float32" min_val = float("-3.77211") @@ -922,6 +1006,7 @@ class Program_weight_tensor_parameter_83: class Program_weight_tensor_parameter_84: name = "parameter_84" + original_name = "batch_norm2d_161.w_0" shape = [768] dtype = "float32" min_val = float("0.750337") @@ -933,6 +1018,7 @@ class Program_weight_tensor_parameter_84: class Program_weight_tensor_parameter_85: name = "parameter_85" + original_name = "batch_norm2d_161.w_2" shape = [768] dtype = "float32" min_val = float("0.0175462") @@ -944,6 +1030,7 @@ class Program_weight_tensor_parameter_85: class Program_weight_tensor_parameter_86: name = "parameter_86" + original_name = "batch_norm2d_161.w_1" shape = [768] dtype = "float32" min_val = float("-0.524633") @@ -955,6 +1042,7 @@ class Program_weight_tensor_parameter_86: class Program_weight_tensor_parameter_87: name = "parameter_87" + original_name = "conv2d_161.w_0" shape = [768, 1, 5, 5] dtype = "float32" min_val = float("-0.324986") @@ -966,6 +1054,7 @@ class Program_weight_tensor_parameter_87: class Program_weight_tensor_parameter_88: name = "parameter_88" + original_name = "batch_norm2d_160.b_0" shape = [768] dtype = "float32" min_val = float("-1.10973") @@ -977,6 +1066,7 @@ class Program_weight_tensor_parameter_88: class Program_weight_tensor_parameter_89: name = "parameter_89" + original_name = "batch_norm2d_160.w_0" shape = [768] dtype = "float32" min_val = float("0.712228") @@ -988,6 +1078,7 @@ class Program_weight_tensor_parameter_89: class Program_weight_tensor_parameter_90: name = "parameter_90" + original_name = "batch_norm2d_160.w_2" shape = [768] dtype = "float32" min_val = float("0.0840615") @@ -999,6 +1090,7 @@ class Program_weight_tensor_parameter_90: class Program_weight_tensor_parameter_91: name = "parameter_91" + original_name = "batch_norm2d_160.w_1" shape = [768] dtype = "float32" min_val = float("-1.72316") @@ -1010,6 +1102,7 @@ class Program_weight_tensor_parameter_91: class Program_weight_tensor_parameter_92: name = "parameter_92" + original_name = "conv2d_160.w_0" shape = [768, 2048, 1, 1] dtype = "float32" min_val = float("-0.358792") @@ -1021,6 +1114,7 @@ class Program_weight_tensor_parameter_92: class Program_weight_tensor_parameter_93: name = "parameter_93" + original_name = "batch_norm2d_159.b_0" shape = [2048] dtype = "float32" min_val = float("-5.99702") @@ -1032,6 +1126,7 @@ class Program_weight_tensor_parameter_93: class Program_weight_tensor_parameter_94: name = "parameter_94" + original_name = "batch_norm2d_159.w_0" shape = [2048] dtype = "float32" min_val = float("0.0147206") @@ -1043,6 +1138,7 @@ class Program_weight_tensor_parameter_94: class Program_weight_tensor_parameter_95: name = "parameter_95" + original_name = "batch_norm2d_159.w_2" shape = [2048] dtype = "float32" min_val = float("5.60519e-45") @@ -1054,6 +1150,7 @@ class Program_weight_tensor_parameter_95: class Program_weight_tensor_parameter_96: name = "parameter_96" + original_name = "batch_norm2d_159.w_1" shape = [2048] dtype = "float32" min_val = float("-1.29306") @@ -1065,6 +1162,7 @@ class Program_weight_tensor_parameter_96: class Program_weight_tensor_parameter_97: name = "parameter_97" + original_name = "conv2d_159.w_0" shape = [2048, 1024, 1, 1] dtype = "float32" min_val = float("-0.465767") @@ -1076,6 +1174,7 @@ class Program_weight_tensor_parameter_97: class Program_weight_tensor_parameter_98: name = "parameter_98" + original_name = "batch_norm2d_158.b_0" shape = [1024] dtype = "float32" min_val = float("-3.51816") @@ -1087,6 +1186,7 @@ class Program_weight_tensor_parameter_98: class Program_weight_tensor_parameter_99: name = "parameter_99" + original_name = "batch_norm2d_158.w_0" shape = [1024] dtype = "float32" min_val = float("0.508807") @@ -1098,6 +1198,7 @@ class Program_weight_tensor_parameter_99: class Program_weight_tensor_parameter_100: name = "parameter_100" + original_name = "batch_norm2d_158.w_2" shape = [1024] dtype = "float32" min_val = float("0.166438") @@ -1109,6 +1210,7 @@ class Program_weight_tensor_parameter_100: class Program_weight_tensor_parameter_101: name = "parameter_101" + original_name = "batch_norm2d_158.w_1" shape = [1024] dtype = "float32" min_val = float("-2.66633") @@ -1120,6 +1222,7 @@ class Program_weight_tensor_parameter_101: class Program_weight_tensor_parameter_102: name = "parameter_102" + original_name = "conv2d_158.w_0" shape = [1024, 6656, 1, 1] dtype = "float32" min_val = float("-0.204425") @@ -1131,6 +1234,7 @@ class Program_weight_tensor_parameter_102: class Program_weight_tensor_parameter_103: name = "parameter_103" + original_name = "batch_norm2d_157.b_0" shape = [768] dtype = "float32" min_val = float("-1.03806") @@ -1142,6 +1246,7 @@ class Program_weight_tensor_parameter_103: class Program_weight_tensor_parameter_104: name = "parameter_104" + original_name = "batch_norm2d_157.w_0" shape = [768] dtype = "float32" min_val = float("0.256805") @@ -1153,6 +1258,7 @@ class Program_weight_tensor_parameter_104: class Program_weight_tensor_parameter_105: name = "parameter_105" + original_name = "batch_norm2d_157.w_2" shape = [768] dtype = "float32" min_val = float("5.60519e-45") @@ -1164,6 +1270,7 @@ class Program_weight_tensor_parameter_105: class Program_weight_tensor_parameter_106: name = "parameter_106" + original_name = "batch_norm2d_157.w_1" shape = [768] dtype = "float32" min_val = float("-0.206218") @@ -1175,6 +1282,7 @@ class Program_weight_tensor_parameter_106: class Program_weight_tensor_parameter_107: name = "parameter_107" + original_name = "conv2d_157.w_0" shape = [768, 1, 5, 5] dtype = "float32" min_val = float("-0.12513") @@ -1186,6 +1294,7 @@ class Program_weight_tensor_parameter_107: class Program_weight_tensor_parameter_108: name = "parameter_108" + original_name = "batch_norm2d_156.b_0" shape = [768] dtype = "float32" min_val = float("-0.929181") @@ -1197,6 +1306,7 @@ class Program_weight_tensor_parameter_108: class Program_weight_tensor_parameter_109: name = "parameter_109" + original_name = "batch_norm2d_156.w_0" shape = [768] dtype = "float32" min_val = float("0.465713") @@ -1208,6 +1318,7 @@ class Program_weight_tensor_parameter_109: class Program_weight_tensor_parameter_110: name = "parameter_110" + original_name = "batch_norm2d_156.w_2" shape = [768] dtype = "float32" min_val = float("5.60519e-45") @@ -1219,6 +1330,7 @@ class Program_weight_tensor_parameter_110: class Program_weight_tensor_parameter_111: name = "parameter_111" + original_name = "batch_norm2d_156.w_1" shape = [768] dtype = "float32" min_val = float("-0.224384") @@ -1230,6 +1342,7 @@ class Program_weight_tensor_parameter_111: class Program_weight_tensor_parameter_112: name = "parameter_112" + original_name = "conv2d_156.w_0" shape = [768, 768, 1, 1] dtype = "float32" min_val = float("-0.104393") @@ -1241,6 +1354,7 @@ class Program_weight_tensor_parameter_112: class Program_weight_tensor_parameter_113: name = "parameter_113" + original_name = "batch_norm2d_155.b_0" shape = [768] dtype = "float32" min_val = float("-1.29183") @@ -1252,6 +1366,7 @@ class Program_weight_tensor_parameter_113: class Program_weight_tensor_parameter_114: name = "parameter_114" + original_name = "batch_norm2d_155.w_0" shape = [768] dtype = "float32" min_val = float("0.116249") @@ -1263,6 +1378,7 @@ class Program_weight_tensor_parameter_114: class Program_weight_tensor_parameter_115: name = "parameter_115" + original_name = "batch_norm2d_155.w_2" shape = [768] dtype = "float32" min_val = float("5.60519e-45") @@ -1274,6 +1390,7 @@ class Program_weight_tensor_parameter_115: class Program_weight_tensor_parameter_116: name = "parameter_116" + original_name = "batch_norm2d_155.w_1" shape = [768] dtype = "float32" min_val = float("-0.395086") @@ -1285,6 +1402,7 @@ class Program_weight_tensor_parameter_116: class Program_weight_tensor_parameter_117: name = "parameter_117" + original_name = "conv2d_155.w_0" shape = [768, 1, 5, 5] dtype = "float32" min_val = float("-0.116556") @@ -1296,6 +1414,7 @@ class Program_weight_tensor_parameter_117: class Program_weight_tensor_parameter_118: name = "parameter_118" + original_name = "batch_norm2d_154.b_0" shape = [768] dtype = "float32" min_val = float("-1.17311") @@ -1307,6 +1426,7 @@ class Program_weight_tensor_parameter_118: class Program_weight_tensor_parameter_119: name = "parameter_119" + original_name = "batch_norm2d_154.w_0" shape = [768] dtype = "float32" min_val = float("0.346779") @@ -1318,6 +1438,7 @@ class Program_weight_tensor_parameter_119: class Program_weight_tensor_parameter_120: name = "parameter_120" + original_name = "batch_norm2d_154.w_2" shape = [768] dtype = "float32" min_val = float("5.60519e-45") @@ -1329,6 +1450,7 @@ class Program_weight_tensor_parameter_120: class Program_weight_tensor_parameter_121: name = "parameter_121" + original_name = "batch_norm2d_154.w_1" shape = [768] dtype = "float32" min_val = float("-0.130081") @@ -1340,6 +1462,7 @@ class Program_weight_tensor_parameter_121: class Program_weight_tensor_parameter_122: name = "parameter_122" + original_name = "conv2d_154.w_0" shape = [768, 768, 1, 1] dtype = "float32" min_val = float("-0.110931") @@ -1351,6 +1474,7 @@ class Program_weight_tensor_parameter_122: class Program_weight_tensor_parameter_123: name = "parameter_123" + original_name = "batch_norm2d_153.b_0" shape = [768] dtype = "float32" min_val = float("-1.49092") @@ -1362,6 +1486,7 @@ class Program_weight_tensor_parameter_123: class Program_weight_tensor_parameter_124: name = "parameter_124" + original_name = "batch_norm2d_153.w_0" shape = [768] dtype = "float32" min_val = float("0.10543") @@ -1373,6 +1498,7 @@ class Program_weight_tensor_parameter_124: class Program_weight_tensor_parameter_125: name = "parameter_125" + original_name = "batch_norm2d_153.w_2" shape = [768] dtype = "float32" min_val = float("5.60519e-45") @@ -1384,6 +1510,7 @@ class Program_weight_tensor_parameter_125: class Program_weight_tensor_parameter_126: name = "parameter_126" + original_name = "batch_norm2d_153.w_1" shape = [768] dtype = "float32" min_val = float("-0.37451") @@ -1395,6 +1522,7 @@ class Program_weight_tensor_parameter_126: class Program_weight_tensor_parameter_127: name = "parameter_127" + original_name = "conv2d_153.w_0" shape = [768, 1, 5, 5] dtype = "float32" min_val = float("-0.201903") @@ -1406,6 +1534,7 @@ class Program_weight_tensor_parameter_127: class Program_weight_tensor_parameter_128: name = "parameter_128" + original_name = "batch_norm2d_152.b_0" shape = [768] dtype = "float32" min_val = float("-1.18527") @@ -1417,6 +1546,7 @@ class Program_weight_tensor_parameter_128: class Program_weight_tensor_parameter_129: name = "parameter_129" + original_name = "batch_norm2d_152.w_0" shape = [768] dtype = "float32" min_val = float("0.390502") @@ -1428,6 +1558,7 @@ class Program_weight_tensor_parameter_129: class Program_weight_tensor_parameter_130: name = "parameter_130" + original_name = "batch_norm2d_152.w_2" shape = [768] dtype = "float32" min_val = float("5.60519e-45") @@ -1439,6 +1570,7 @@ class Program_weight_tensor_parameter_130: class Program_weight_tensor_parameter_131: name = "parameter_131" + original_name = "batch_norm2d_152.w_1" shape = [768] dtype = "float32" min_val = float("-0.361126") @@ -1450,6 +1582,7 @@ class Program_weight_tensor_parameter_131: class Program_weight_tensor_parameter_132: name = "parameter_132" + original_name = "conv2d_152.w_0" shape = [768, 768, 1, 1] dtype = "float32" min_val = float("-0.126468") @@ -1461,6 +1594,7 @@ class Program_weight_tensor_parameter_132: class Program_weight_tensor_parameter_133: name = "parameter_133" + original_name = "batch_norm2d_151.b_0" shape = [768] dtype = "float32" min_val = float("-2.27293") @@ -1472,6 +1606,7 @@ class Program_weight_tensor_parameter_133: class Program_weight_tensor_parameter_134: name = "parameter_134" + original_name = "batch_norm2d_151.w_0" shape = [768] dtype = "float32" min_val = float("0.441175") @@ -1483,6 +1618,7 @@ class Program_weight_tensor_parameter_134: class Program_weight_tensor_parameter_135: name = "parameter_135" + original_name = "batch_norm2d_151.w_2" shape = [768] dtype = "float32" min_val = float("0.00407724") @@ -1494,6 +1630,7 @@ class Program_weight_tensor_parameter_135: class Program_weight_tensor_parameter_136: name = "parameter_136" + original_name = "batch_norm2d_151.w_1" shape = [768] dtype = "float32" min_val = float("-0.22989") @@ -1505,6 +1642,7 @@ class Program_weight_tensor_parameter_136: class Program_weight_tensor_parameter_137: name = "parameter_137" + original_name = "conv2d_151.w_0" shape = [768, 1, 5, 5] dtype = "float32" min_val = float("-0.245312") @@ -1516,6 +1654,7 @@ class Program_weight_tensor_parameter_137: class Program_weight_tensor_parameter_138: name = "parameter_138" + original_name = "batch_norm2d_150.b_0" shape = [768] dtype = "float32" min_val = float("-1.04341") @@ -1527,6 +1666,7 @@ class Program_weight_tensor_parameter_138: class Program_weight_tensor_parameter_139: name = "parameter_139" + original_name = "batch_norm2d_150.w_0" shape = [768] dtype = "float32" min_val = float("0.507539") @@ -1538,6 +1678,7 @@ class Program_weight_tensor_parameter_139: class Program_weight_tensor_parameter_140: name = "parameter_140" + original_name = "batch_norm2d_150.w_2" shape = [768] dtype = "float32" min_val = float("0.00799475") @@ -1549,6 +1690,7 @@ class Program_weight_tensor_parameter_140: class Program_weight_tensor_parameter_141: name = "parameter_141" + original_name = "batch_norm2d_150.w_1" shape = [768] dtype = "float32" min_val = float("-0.495059") @@ -1560,6 +1702,7 @@ class Program_weight_tensor_parameter_141: class Program_weight_tensor_parameter_142: name = "parameter_142" + original_name = "conv2d_150.w_0" shape = [768, 768, 1, 1] dtype = "float32" min_val = float("-0.180137") @@ -1571,6 +1714,7 @@ class Program_weight_tensor_parameter_142: class Program_weight_tensor_parameter_143: name = "parameter_143" + original_name = "batch_norm2d_149.b_0" shape = [768] dtype = "float32" min_val = float("-2.55579") @@ -1582,6 +1726,7 @@ class Program_weight_tensor_parameter_143: class Program_weight_tensor_parameter_144: name = "parameter_144" + original_name = "batch_norm2d_149.w_0" shape = [768] dtype = "float32" min_val = float("0.371534") @@ -1593,6 +1738,7 @@ class Program_weight_tensor_parameter_144: class Program_weight_tensor_parameter_145: name = "parameter_145" + original_name = "batch_norm2d_149.w_2" shape = [768] dtype = "float32" min_val = float("0.00714039") @@ -1604,6 +1750,7 @@ class Program_weight_tensor_parameter_145: class Program_weight_tensor_parameter_146: name = "parameter_146" + original_name = "batch_norm2d_149.w_1" shape = [768] dtype = "float32" min_val = float("-0.292576") @@ -1615,6 +1762,7 @@ class Program_weight_tensor_parameter_146: class Program_weight_tensor_parameter_147: name = "parameter_147" + original_name = "conv2d_149.w_0" shape = [768, 1, 5, 5] dtype = "float32" min_val = float("-0.232555") @@ -1626,6 +1774,7 @@ class Program_weight_tensor_parameter_147: class Program_weight_tensor_parameter_148: name = "parameter_148" + original_name = "batch_norm2d_148.b_0" shape = [768] dtype = "float32" min_val = float("-1.02556") @@ -1637,6 +1786,7 @@ class Program_weight_tensor_parameter_148: class Program_weight_tensor_parameter_149: name = "parameter_149" + original_name = "batch_norm2d_148.w_0" shape = [768] dtype = "float32" min_val = float("0.556788") @@ -1648,6 +1798,7 @@ class Program_weight_tensor_parameter_149: class Program_weight_tensor_parameter_150: name = "parameter_150" + original_name = "batch_norm2d_148.w_2" shape = [768] dtype = "float32" min_val = float("0.0233231") @@ -1659,6 +1810,7 @@ class Program_weight_tensor_parameter_150: class Program_weight_tensor_parameter_151: name = "parameter_151" + original_name = "batch_norm2d_148.w_1" shape = [768] dtype = "float32" min_val = float("-1.28125") @@ -1670,6 +1822,7 @@ class Program_weight_tensor_parameter_151: class Program_weight_tensor_parameter_152: name = "parameter_152" + original_name = "conv2d_148.w_0" shape = [768, 768, 1, 1] dtype = "float32" min_val = float("-0.187883") @@ -1681,6 +1834,7 @@ class Program_weight_tensor_parameter_152: class Program_weight_tensor_parameter_153: name = "parameter_153" + original_name = "batch_norm2d_147.b_0" shape = [768] dtype = "float32" min_val = float("-3.33703") @@ -1692,6 +1846,7 @@ class Program_weight_tensor_parameter_153: class Program_weight_tensor_parameter_154: name = "parameter_154" + original_name = "batch_norm2d_147.w_0" shape = [768] dtype = "float32" min_val = float("0.586882") @@ -1703,6 +1858,7 @@ class Program_weight_tensor_parameter_154: class Program_weight_tensor_parameter_155: name = "parameter_155" + original_name = "batch_norm2d_147.w_2" shape = [768] dtype = "float32" min_val = float("0.00976779") @@ -1714,6 +1870,7 @@ class Program_weight_tensor_parameter_155: class Program_weight_tensor_parameter_156: name = "parameter_156" + original_name = "batch_norm2d_147.w_1" shape = [768] dtype = "float32" min_val = float("-0.339943") @@ -1725,6 +1882,7 @@ class Program_weight_tensor_parameter_156: class Program_weight_tensor_parameter_157: name = "parameter_157" + original_name = "conv2d_147.w_0" shape = [768, 1, 5, 5] dtype = "float32" min_val = float("-0.220188") @@ -1736,6 +1894,7 @@ class Program_weight_tensor_parameter_157: class Program_weight_tensor_parameter_158: name = "parameter_158" + original_name = "batch_norm2d_146.b_0" shape = [768] dtype = "float32" min_val = float("-1.15383") @@ -1747,6 +1906,7 @@ class Program_weight_tensor_parameter_158: class Program_weight_tensor_parameter_159: name = "parameter_159" + original_name = "batch_norm2d_146.w_0" shape = [768] dtype = "float32" min_val = float("0.658801") @@ -1758,6 +1918,7 @@ class Program_weight_tensor_parameter_159: class Program_weight_tensor_parameter_160: name = "parameter_160" + original_name = "batch_norm2d_146.w_2" shape = [768] dtype = "float32" min_val = float("0.0494411") @@ -1769,6 +1930,7 @@ class Program_weight_tensor_parameter_160: class Program_weight_tensor_parameter_161: name = "parameter_161" + original_name = "batch_norm2d_146.w_1" shape = [768] dtype = "float32" min_val = float("-0.482039") @@ -1780,6 +1942,7 @@ class Program_weight_tensor_parameter_161: class Program_weight_tensor_parameter_162: name = "parameter_162" + original_name = "conv2d_146.w_0" shape = [768, 2048, 1, 1] dtype = "float32" min_val = float("-0.216685") @@ -1791,6 +1954,7 @@ class Program_weight_tensor_parameter_162: class Program_weight_tensor_parameter_163: name = "parameter_163" + original_name = "batch_norm2d_145.b_0" shape = [2048] dtype = "float32" min_val = float("-6.32689") @@ -1802,6 +1966,7 @@ class Program_weight_tensor_parameter_163: class Program_weight_tensor_parameter_164: name = "parameter_164" + original_name = "batch_norm2d_145.w_0" shape = [2048] dtype = "float32" min_val = float("-0.0740162") @@ -1813,6 +1978,7 @@ class Program_weight_tensor_parameter_164: class Program_weight_tensor_parameter_165: name = "parameter_165" + original_name = "batch_norm2d_145.w_2" shape = [2048] dtype = "float32" min_val = float("5.60519e-45") @@ -1824,6 +1990,7 @@ class Program_weight_tensor_parameter_165: class Program_weight_tensor_parameter_166: name = "parameter_166" + original_name = "batch_norm2d_145.w_1" shape = [2048] dtype = "float32" min_val = float("-2.49716") @@ -1835,6 +2002,7 @@ class Program_weight_tensor_parameter_166: class Program_weight_tensor_parameter_167: name = "parameter_167" + original_name = "conv2d_145.w_0" shape = [2048, 1024, 1, 1] dtype = "float32" min_val = float("-0.652681") @@ -1846,6 +2014,7 @@ class Program_weight_tensor_parameter_167: class Program_weight_tensor_parameter_168: name = "parameter_168" + original_name = "batch_norm2d_144.b_0" shape = [1024] dtype = "float32" min_val = float("-3.93375") @@ -1857,6 +2026,7 @@ class Program_weight_tensor_parameter_168: class Program_weight_tensor_parameter_169: name = "parameter_169" + original_name = "batch_norm2d_144.w_0" shape = [1024] dtype = "float32" min_val = float("0.876041") @@ -1868,6 +2038,7 @@ class Program_weight_tensor_parameter_169: class Program_weight_tensor_parameter_170: name = "parameter_170" + original_name = "batch_norm2d_144.w_2" shape = [1024] dtype = "float32" min_val = float("0.363466") @@ -1879,6 +2050,7 @@ class Program_weight_tensor_parameter_170: class Program_weight_tensor_parameter_171: name = "parameter_171" + original_name = "batch_norm2d_144.w_1" shape = [1024] dtype = "float32" min_val = float("-1.55899") @@ -1890,6 +2062,7 @@ class Program_weight_tensor_parameter_171: class Program_weight_tensor_parameter_172: name = "parameter_172" + original_name = "conv2d_144.w_0" shape = [1024, 5632, 1, 1] dtype = "float32" min_val = float("-0.639391") @@ -1901,6 +2074,7 @@ class Program_weight_tensor_parameter_172: class Program_weight_tensor_parameter_173: name = "parameter_173" + original_name = "batch_norm2d_143.b_0" shape = [768] dtype = "float32" min_val = float("-1.92231") @@ -1912,6 +2086,7 @@ class Program_weight_tensor_parameter_173: class Program_weight_tensor_parameter_174: name = "parameter_174" + original_name = "batch_norm2d_143.w_0" shape = [768] dtype = "float32" min_val = float("0.453156") @@ -1923,6 +2098,7 @@ class Program_weight_tensor_parameter_174: class Program_weight_tensor_parameter_175: name = "parameter_175" + original_name = "batch_norm2d_143.w_2" shape = [768] dtype = "float32" min_val = float("0.00365113") @@ -1934,6 +2110,7 @@ class Program_weight_tensor_parameter_175: class Program_weight_tensor_parameter_176: name = "parameter_176" + original_name = "batch_norm2d_143.w_1" shape = [768] dtype = "float32" min_val = float("-0.242685") @@ -1945,6 +2122,7 @@ class Program_weight_tensor_parameter_176: class Program_weight_tensor_parameter_177: name = "parameter_177" + original_name = "conv2d_143.w_0" shape = [768, 1, 5, 5] dtype = "float32" min_val = float("-0.169156") @@ -1956,6 +2134,7 @@ class Program_weight_tensor_parameter_177: class Program_weight_tensor_parameter_178: name = "parameter_178" + original_name = "batch_norm2d_142.b_0" shape = [768] dtype = "float32" min_val = float("-0.92928") @@ -1967,6 +2146,7 @@ class Program_weight_tensor_parameter_178: class Program_weight_tensor_parameter_179: name = "parameter_179" + original_name = "batch_norm2d_142.w_0" shape = [768] dtype = "float32" min_val = float("0.501483") @@ -1978,6 +2158,7 @@ class Program_weight_tensor_parameter_179: class Program_weight_tensor_parameter_180: name = "parameter_180" + original_name = "batch_norm2d_142.w_2" shape = [768] dtype = "float32" min_val = float("0.00146172") @@ -1989,6 +2170,7 @@ class Program_weight_tensor_parameter_180: class Program_weight_tensor_parameter_181: name = "parameter_181" + original_name = "batch_norm2d_142.w_1" shape = [768] dtype = "float32" min_val = float("-0.15135") @@ -2000,6 +2182,7 @@ class Program_weight_tensor_parameter_181: class Program_weight_tensor_parameter_182: name = "parameter_182" + original_name = "conv2d_142.w_0" shape = [768, 768, 1, 1] dtype = "float32" min_val = float("-0.177133") @@ -2011,6 +2194,7 @@ class Program_weight_tensor_parameter_182: class Program_weight_tensor_parameter_183: name = "parameter_183" + original_name = "batch_norm2d_141.b_0" shape = [768] dtype = "float32" min_val = float("-2.3279") @@ -2022,6 +2206,7 @@ class Program_weight_tensor_parameter_183: class Program_weight_tensor_parameter_184: name = "parameter_184" + original_name = "batch_norm2d_141.w_0" shape = [768] dtype = "float32" min_val = float("0.217294") @@ -2033,6 +2218,7 @@ class Program_weight_tensor_parameter_184: class Program_weight_tensor_parameter_185: name = "parameter_185" + original_name = "batch_norm2d_141.w_2" shape = [768] dtype = "float32" min_val = float("9.95879e-34") @@ -2044,6 +2230,7 @@ class Program_weight_tensor_parameter_185: class Program_weight_tensor_parameter_186: name = "parameter_186" + original_name = "batch_norm2d_141.w_1" shape = [768] dtype = "float32" min_val = float("-0.664225") @@ -2055,6 +2242,7 @@ class Program_weight_tensor_parameter_186: class Program_weight_tensor_parameter_187: name = "parameter_187" + original_name = "conv2d_141.w_0" shape = [768, 1, 5, 5] dtype = "float32" min_val = float("-0.152236") @@ -2066,6 +2254,7 @@ class Program_weight_tensor_parameter_187: class Program_weight_tensor_parameter_188: name = "parameter_188" + original_name = "batch_norm2d_140.b_0" shape = [768] dtype = "float32" min_val = float("-1.26207") @@ -2077,6 +2266,7 @@ class Program_weight_tensor_parameter_188: class Program_weight_tensor_parameter_189: name = "parameter_189" + original_name = "batch_norm2d_140.w_0" shape = [768] dtype = "float32" min_val = float("0.412249") @@ -2088,6 +2278,7 @@ class Program_weight_tensor_parameter_189: class Program_weight_tensor_parameter_190: name = "parameter_190" + original_name = "batch_norm2d_140.w_2" shape = [768] dtype = "float32" min_val = float("5.60519e-45") @@ -2099,6 +2290,7 @@ class Program_weight_tensor_parameter_190: class Program_weight_tensor_parameter_191: name = "parameter_191" + original_name = "batch_norm2d_140.w_1" shape = [768] dtype = "float32" min_val = float("-0.315726") @@ -2110,6 +2302,7 @@ class Program_weight_tensor_parameter_191: class Program_weight_tensor_parameter_192: name = "parameter_192" + original_name = "conv2d_140.w_0" shape = [768, 768, 1, 1] dtype = "float32" min_val = float("-0.173836") @@ -2121,6 +2314,7 @@ class Program_weight_tensor_parameter_192: class Program_weight_tensor_parameter_193: name = "parameter_193" + original_name = "batch_norm2d_139.b_0" shape = [768] dtype = "float32" min_val = float("-2.31713") @@ -2132,6 +2326,7 @@ class Program_weight_tensor_parameter_193: class Program_weight_tensor_parameter_194: name = "parameter_194" + original_name = "batch_norm2d_139.w_0" shape = [768] dtype = "float32" min_val = float("0.316561") @@ -2143,6 +2338,7 @@ class Program_weight_tensor_parameter_194: class Program_weight_tensor_parameter_195: name = "parameter_195" + original_name = "batch_norm2d_139.w_2" shape = [768] dtype = "float32" min_val = float("5.60519e-45") @@ -2154,6 +2350,7 @@ class Program_weight_tensor_parameter_195: class Program_weight_tensor_parameter_196: name = "parameter_196" + original_name = "batch_norm2d_139.w_1" shape = [768] dtype = "float32" min_val = float("-0.428089") @@ -2165,6 +2362,7 @@ class Program_weight_tensor_parameter_196: class Program_weight_tensor_parameter_197: name = "parameter_197" + original_name = "conv2d_139.w_0" shape = [768, 1, 5, 5] dtype = "float32" min_val = float("-0.208446") @@ -2176,6 +2374,7 @@ class Program_weight_tensor_parameter_197: class Program_weight_tensor_parameter_198: name = "parameter_198" + original_name = "batch_norm2d_138.b_0" shape = [768] dtype = "float32" min_val = float("-1.11621") @@ -2187,6 +2386,7 @@ class Program_weight_tensor_parameter_198: class Program_weight_tensor_parameter_199: name = "parameter_199" + original_name = "batch_norm2d_138.w_0" shape = [768] dtype = "float32" min_val = float("0.393363") @@ -2198,6 +2398,7 @@ class Program_weight_tensor_parameter_199: class Program_weight_tensor_parameter_200: name = "parameter_200" + original_name = "batch_norm2d_138.w_2" shape = [768] dtype = "float32" min_val = float("5.60519e-45") @@ -2209,6 +2410,7 @@ class Program_weight_tensor_parameter_200: class Program_weight_tensor_parameter_201: name = "parameter_201" + original_name = "batch_norm2d_138.w_1" shape = [768] dtype = "float32" min_val = float("-0.458386") @@ -2220,6 +2422,7 @@ class Program_weight_tensor_parameter_201: class Program_weight_tensor_parameter_202: name = "parameter_202" + original_name = "conv2d_138.w_0" shape = [768, 768, 1, 1] dtype = "float32" min_val = float("-0.227384") @@ -2231,6 +2434,7 @@ class Program_weight_tensor_parameter_202: class Program_weight_tensor_parameter_203: name = "parameter_203" + original_name = "batch_norm2d_137.b_0" shape = [768] dtype = "float32" min_val = float("-2.91242") @@ -2242,6 +2446,7 @@ class Program_weight_tensor_parameter_203: class Program_weight_tensor_parameter_204: name = "parameter_204" + original_name = "batch_norm2d_137.w_0" shape = [768] dtype = "float32" min_val = float("0.392892") @@ -2253,6 +2458,7 @@ class Program_weight_tensor_parameter_204: class Program_weight_tensor_parameter_205: name = "parameter_205" + original_name = "batch_norm2d_137.w_2" shape = [768] dtype = "float32" min_val = float("0.00587871") @@ -2264,6 +2470,7 @@ class Program_weight_tensor_parameter_205: class Program_weight_tensor_parameter_206: name = "parameter_206" + original_name = "batch_norm2d_137.w_1" shape = [768] dtype = "float32" min_val = float("-0.472055") @@ -2275,6 +2482,7 @@ class Program_weight_tensor_parameter_206: class Program_weight_tensor_parameter_207: name = "parameter_207" + original_name = "conv2d_137.w_0" shape = [768, 1, 5, 5] dtype = "float32" min_val = float("-0.213137") @@ -2286,6 +2494,7 @@ class Program_weight_tensor_parameter_207: class Program_weight_tensor_parameter_208: name = "parameter_208" + original_name = "batch_norm2d_136.b_0" shape = [768] dtype = "float32" min_val = float("-1.32434") @@ -2297,6 +2506,7 @@ class Program_weight_tensor_parameter_208: class Program_weight_tensor_parameter_209: name = "parameter_209" + original_name = "batch_norm2d_136.w_0" shape = [768] dtype = "float32" min_val = float("0.47526") @@ -2308,6 +2518,7 @@ class Program_weight_tensor_parameter_209: class Program_weight_tensor_parameter_210: name = "parameter_210" + original_name = "batch_norm2d_136.w_2" shape = [768] dtype = "float32" min_val = float("0.00996572") @@ -2319,6 +2530,7 @@ class Program_weight_tensor_parameter_210: class Program_weight_tensor_parameter_211: name = "parameter_211" + original_name = "batch_norm2d_136.w_1" shape = [768] dtype = "float32" min_val = float("-0.623457") @@ -2330,6 +2542,7 @@ class Program_weight_tensor_parameter_211: class Program_weight_tensor_parameter_212: name = "parameter_212" + original_name = "conv2d_136.w_0" shape = [768, 768, 1, 1] dtype = "float32" min_val = float("-0.223325") @@ -2341,6 +2554,7 @@ class Program_weight_tensor_parameter_212: class Program_weight_tensor_parameter_213: name = "parameter_213" + original_name = "batch_norm2d_135.b_0" shape = [768] dtype = "float32" min_val = float("-3.59462") @@ -2352,6 +2566,7 @@ class Program_weight_tensor_parameter_213: class Program_weight_tensor_parameter_214: name = "parameter_214" + original_name = "batch_norm2d_135.w_0" shape = [768] dtype = "float32" min_val = float("0.381844") @@ -2363,6 +2578,7 @@ class Program_weight_tensor_parameter_214: class Program_weight_tensor_parameter_215: name = "parameter_215" + original_name = "batch_norm2d_135.w_2" shape = [768] dtype = "float32" min_val = float("0.00921308") @@ -2374,6 +2590,7 @@ class Program_weight_tensor_parameter_215: class Program_weight_tensor_parameter_216: name = "parameter_216" + original_name = "batch_norm2d_135.w_1" shape = [768] dtype = "float32" min_val = float("-0.541482") @@ -2385,6 +2602,7 @@ class Program_weight_tensor_parameter_216: class Program_weight_tensor_parameter_217: name = "parameter_217" + original_name = "conv2d_135.w_0" shape = [768, 1, 5, 5] dtype = "float32" min_val = float("-0.291026") @@ -2396,6 +2614,7 @@ class Program_weight_tensor_parameter_217: class Program_weight_tensor_parameter_218: name = "parameter_218" + original_name = "batch_norm2d_134.b_0" shape = [768] dtype = "float32" min_val = float("-1.39482") @@ -2407,6 +2626,7 @@ class Program_weight_tensor_parameter_218: class Program_weight_tensor_parameter_219: name = "parameter_219" + original_name = "batch_norm2d_134.w_0" shape = [768] dtype = "float32" min_val = float("0.281948") @@ -2418,6 +2638,7 @@ class Program_weight_tensor_parameter_219: class Program_weight_tensor_parameter_220: name = "parameter_220" + original_name = "batch_norm2d_134.w_2" shape = [768] dtype = "float32" min_val = float("0.0315367") @@ -2429,6 +2650,7 @@ class Program_weight_tensor_parameter_220: class Program_weight_tensor_parameter_221: name = "parameter_221" + original_name = "batch_norm2d_134.w_1" shape = [768] dtype = "float32" min_val = float("-1.0785") @@ -2440,6 +2662,7 @@ class Program_weight_tensor_parameter_221: class Program_weight_tensor_parameter_222: name = "parameter_222" + original_name = "conv2d_134.w_0" shape = [768, 768, 1, 1] dtype = "float32" min_val = float("-0.46614") @@ -2451,6 +2674,7 @@ class Program_weight_tensor_parameter_222: class Program_weight_tensor_parameter_223: name = "parameter_223" + original_name = "batch_norm2d_133.b_0" shape = [768] dtype = "float32" min_val = float("-5.76533") @@ -2462,6 +2686,7 @@ class Program_weight_tensor_parameter_223: class Program_weight_tensor_parameter_224: name = "parameter_224" + original_name = "batch_norm2d_133.w_0" shape = [768] dtype = "float32" min_val = float("0.662891") @@ -2473,6 +2698,7 @@ class Program_weight_tensor_parameter_224: class Program_weight_tensor_parameter_225: name = "parameter_225" + original_name = "batch_norm2d_133.w_2" shape = [768] dtype = "float32" min_val = float("0.0168725") @@ -2484,6 +2710,7 @@ class Program_weight_tensor_parameter_225: class Program_weight_tensor_parameter_226: name = "parameter_226" + original_name = "batch_norm2d_133.w_1" shape = [768] dtype = "float32" min_val = float("-0.704863") @@ -2495,6 +2722,7 @@ class Program_weight_tensor_parameter_226: class Program_weight_tensor_parameter_227: name = "parameter_227" + original_name = "conv2d_133.w_0" shape = [768, 1, 5, 5] dtype = "float32" min_val = float("-0.292723") @@ -2506,6 +2734,7 @@ class Program_weight_tensor_parameter_227: class Program_weight_tensor_parameter_228: name = "parameter_228" + original_name = "batch_norm2d_132.b_0" shape = [768] dtype = "float32" min_val = float("-1.07657") @@ -2517,6 +2746,7 @@ class Program_weight_tensor_parameter_228: class Program_weight_tensor_parameter_229: name = "parameter_229" + original_name = "batch_norm2d_132.w_0" shape = [768] dtype = "float32" min_val = float("0.358854") @@ -2528,6 +2758,7 @@ class Program_weight_tensor_parameter_229: class Program_weight_tensor_parameter_230: name = "parameter_230" + original_name = "batch_norm2d_132.w_2" shape = [768] dtype = "float32" min_val = float("0.15943") @@ -2539,6 +2770,7 @@ class Program_weight_tensor_parameter_230: class Program_weight_tensor_parameter_231: name = "parameter_231" + original_name = "batch_norm2d_132.w_1" shape = [768] dtype = "float32" min_val = float("-0.00752778") @@ -2550,6 +2782,7 @@ class Program_weight_tensor_parameter_231: class Program_weight_tensor_parameter_232: name = "parameter_232" + original_name = "conv2d_132.w_0" shape = [768, 1024, 1, 1] dtype = "float32" min_val = float("-0.464909") @@ -2561,6 +2794,7 @@ class Program_weight_tensor_parameter_232: class Program_weight_tensor_parameter_233: name = "parameter_233" + original_name = "batch_norm2d_131.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0129033") @@ -2572,6 +2806,7 @@ class Program_weight_tensor_parameter_233: class Program_weight_tensor_parameter_234: name = "parameter_234" + original_name = "batch_norm2d_131.w_0" shape = [1024] dtype = "float32" min_val = float("0.473081") @@ -2583,6 +2818,7 @@ class Program_weight_tensor_parameter_234: class Program_weight_tensor_parameter_235: name = "parameter_235" + original_name = "batch_norm2d_131.w_2" shape = [1024] dtype = "float32" min_val = float("0.00468078") @@ -2594,6 +2830,7 @@ class Program_weight_tensor_parameter_235: class Program_weight_tensor_parameter_236: name = "parameter_236" + original_name = "batch_norm2d_131.w_1" shape = [1024] dtype = "float32" min_val = float("-0.928586") @@ -2605,6 +2842,7 @@ class Program_weight_tensor_parameter_236: class Program_weight_tensor_parameter_237: name = "parameter_237" + original_name = "conv2d_131.w_0" shape = [1024, 1, 3, 3] dtype = "float32" min_val = float("-0.111346") @@ -2616,6 +2854,7 @@ class Program_weight_tensor_parameter_237: class Program_weight_tensor_parameter_238: name = "parameter_238" + original_name = "batch_norm2d_130.b_0" shape = [1024] dtype = "float32" min_val = float("-2.85412") @@ -2627,6 +2866,7 @@ class Program_weight_tensor_parameter_238: class Program_weight_tensor_parameter_239: name = "parameter_239" + original_name = "batch_norm2d_130.w_0" shape = [1024] dtype = "float32" min_val = float("-0.00307949") @@ -2638,6 +2878,7 @@ class Program_weight_tensor_parameter_239: class Program_weight_tensor_parameter_240: name = "parameter_240" + original_name = "batch_norm2d_130.w_2" shape = [1024] dtype = "float32" min_val = float("5.60519e-45") @@ -2649,6 +2890,7 @@ class Program_weight_tensor_parameter_240: class Program_weight_tensor_parameter_241: name = "parameter_241" + original_name = "batch_norm2d_130.w_1" shape = [1024] dtype = "float32" min_val = float("-0.726516") @@ -2660,6 +2902,7 @@ class Program_weight_tensor_parameter_241: class Program_weight_tensor_parameter_242: name = "parameter_242" + original_name = "conv2d_130.w_0" shape = [1024, 512, 1, 1] dtype = "float32" min_val = float("-0.155727") @@ -2671,6 +2914,7 @@ class Program_weight_tensor_parameter_242: class Program_weight_tensor_parameter_243: name = "parameter_243" + original_name = "batch_norm2d_129.b_0" shape = [512] dtype = "float32" min_val = float("-3.77873") @@ -2682,6 +2926,7 @@ class Program_weight_tensor_parameter_243: class Program_weight_tensor_parameter_244: name = "parameter_244" + original_name = "batch_norm2d_129.w_0" shape = [512] dtype = "float32" min_val = float("0.373207") @@ -2693,6 +2938,7 @@ class Program_weight_tensor_parameter_244: class Program_weight_tensor_parameter_245: name = "parameter_245" + original_name = "batch_norm2d_129.w_2" shape = [512] dtype = "float32" min_val = float("0.181144") @@ -2704,6 +2950,7 @@ class Program_weight_tensor_parameter_245: class Program_weight_tensor_parameter_246: name = "parameter_246" + original_name = "batch_norm2d_129.w_1" shape = [512] dtype = "float32" min_val = float("-1.6385") @@ -2715,6 +2962,7 @@ class Program_weight_tensor_parameter_246: class Program_weight_tensor_parameter_247: name = "parameter_247" + original_name = "conv2d_129.w_0" shape = [512, 3328, 1, 1] dtype = "float32" min_val = float("-0.188065") @@ -2726,6 +2974,7 @@ class Program_weight_tensor_parameter_247: class Program_weight_tensor_parameter_248: name = "parameter_248" + original_name = "batch_norm2d_128.b_0" shape = [384] dtype = "float32" min_val = float("-0.625452") @@ -2737,6 +2986,7 @@ class Program_weight_tensor_parameter_248: class Program_weight_tensor_parameter_249: name = "parameter_249" + original_name = "batch_norm2d_128.w_0" shape = [384] dtype = "float32" min_val = float("0.581948") @@ -2748,6 +2998,7 @@ class Program_weight_tensor_parameter_249: class Program_weight_tensor_parameter_250: name = "parameter_250" + original_name = "batch_norm2d_128.w_2" shape = [384] dtype = "float32" min_val = float("5.60519e-45") @@ -2759,6 +3010,7 @@ class Program_weight_tensor_parameter_250: class Program_weight_tensor_parameter_251: name = "parameter_251" + original_name = "batch_norm2d_128.w_1" shape = [384] dtype = "float32" min_val = float("-0.0692315") @@ -2770,6 +3022,7 @@ class Program_weight_tensor_parameter_251: class Program_weight_tensor_parameter_252: name = "parameter_252" + original_name = "conv2d_128.w_0" shape = [384, 1, 5, 5] dtype = "float32" min_val = float("-0.0676291") @@ -2781,6 +3034,7 @@ class Program_weight_tensor_parameter_252: class Program_weight_tensor_parameter_253: name = "parameter_253" + original_name = "batch_norm2d_127.b_0" shape = [384] dtype = "float32" min_val = float("-0.426511") @@ -2792,6 +3046,7 @@ class Program_weight_tensor_parameter_253: class Program_weight_tensor_parameter_254: name = "parameter_254" + original_name = "batch_norm2d_127.w_0" shape = [384] dtype = "float32" min_val = float("0.906661") @@ -2803,6 +3058,7 @@ class Program_weight_tensor_parameter_254: class Program_weight_tensor_parameter_255: name = "parameter_255" + original_name = "batch_norm2d_127.w_2" shape = [384] dtype = "float32" min_val = float("5.60519e-45") @@ -2814,6 +3070,7 @@ class Program_weight_tensor_parameter_255: class Program_weight_tensor_parameter_256: name = "parameter_256" + original_name = "batch_norm2d_127.w_1" shape = [384] dtype = "float32" min_val = float("-0.304779") @@ -2825,6 +3082,7 @@ class Program_weight_tensor_parameter_256: class Program_weight_tensor_parameter_257: name = "parameter_257" + original_name = "conv2d_127.w_0" shape = [384, 384, 1, 1] dtype = "float32" min_val = float("-0.0672616") @@ -2836,6 +3094,7 @@ class Program_weight_tensor_parameter_257: class Program_weight_tensor_parameter_258: name = "parameter_258" + original_name = "batch_norm2d_126.b_0" shape = [384] dtype = "float32" min_val = float("-0.60566") @@ -2847,6 +3106,7 @@ class Program_weight_tensor_parameter_258: class Program_weight_tensor_parameter_259: name = "parameter_259" + original_name = "batch_norm2d_126.w_0" shape = [384] dtype = "float32" min_val = float("0.593388") @@ -2858,6 +3118,7 @@ class Program_weight_tensor_parameter_259: class Program_weight_tensor_parameter_260: name = "parameter_260" + original_name = "batch_norm2d_126.w_2" shape = [384] dtype = "float32" min_val = float("5.60519e-45") @@ -2869,6 +3130,7 @@ class Program_weight_tensor_parameter_260: class Program_weight_tensor_parameter_261: name = "parameter_261" + original_name = "batch_norm2d_126.w_1" shape = [384] dtype = "float32" min_val = float("-0.311524") @@ -2880,6 +3142,7 @@ class Program_weight_tensor_parameter_261: class Program_weight_tensor_parameter_262: name = "parameter_262" + original_name = "conv2d_126.w_0" shape = [384, 1, 5, 5] dtype = "float32" min_val = float("-0.0998878") @@ -2891,6 +3154,7 @@ class Program_weight_tensor_parameter_262: class Program_weight_tensor_parameter_263: name = "parameter_263" + original_name = "batch_norm2d_125.b_0" shape = [384] dtype = "float32" min_val = float("-1.02076") @@ -2902,6 +3166,7 @@ class Program_weight_tensor_parameter_263: class Program_weight_tensor_parameter_264: name = "parameter_264" + original_name = "batch_norm2d_125.w_0" shape = [384] dtype = "float32" min_val = float("0.162312") @@ -2913,6 +3178,7 @@ class Program_weight_tensor_parameter_264: class Program_weight_tensor_parameter_265: name = "parameter_265" + original_name = "batch_norm2d_125.w_2" shape = [384] dtype = "float32" min_val = float("5.60519e-45") @@ -2924,6 +3190,7 @@ class Program_weight_tensor_parameter_265: class Program_weight_tensor_parameter_266: name = "parameter_266" + original_name = "batch_norm2d_125.w_1" shape = [384] dtype = "float32" min_val = float("-0.14117") @@ -2935,6 +3202,7 @@ class Program_weight_tensor_parameter_266: class Program_weight_tensor_parameter_267: name = "parameter_267" + original_name = "conv2d_125.w_0" shape = [384, 384, 1, 1] dtype = "float32" min_val = float("-0.0811716") @@ -2946,6 +3214,7 @@ class Program_weight_tensor_parameter_267: class Program_weight_tensor_parameter_268: name = "parameter_268" + original_name = "batch_norm2d_124.b_0" shape = [384] dtype = "float32" min_val = float("-1.38525") @@ -2957,6 +3226,7 @@ class Program_weight_tensor_parameter_268: class Program_weight_tensor_parameter_269: name = "parameter_269" + original_name = "batch_norm2d_124.w_0" shape = [384] dtype = "float32" min_val = float("0.575772") @@ -2968,6 +3238,7 @@ class Program_weight_tensor_parameter_269: class Program_weight_tensor_parameter_270: name = "parameter_270" + original_name = "batch_norm2d_124.w_2" shape = [384] dtype = "float32" min_val = float("5.60519e-45") @@ -2979,6 +3250,7 @@ class Program_weight_tensor_parameter_270: class Program_weight_tensor_parameter_271: name = "parameter_271" + original_name = "batch_norm2d_124.w_1" shape = [384] dtype = "float32" min_val = float("-0.291001") @@ -2990,6 +3262,7 @@ class Program_weight_tensor_parameter_271: class Program_weight_tensor_parameter_272: name = "parameter_272" + original_name = "conv2d_124.w_0" shape = [384, 1, 5, 5] dtype = "float32" min_val = float("-0.113337") @@ -3001,6 +3274,7 @@ class Program_weight_tensor_parameter_272: class Program_weight_tensor_parameter_273: name = "parameter_273" + original_name = "batch_norm2d_123.b_0" shape = [384] dtype = "float32" min_val = float("-1.22999") @@ -3012,6 +3286,7 @@ class Program_weight_tensor_parameter_273: class Program_weight_tensor_parameter_274: name = "parameter_274" + original_name = "batch_norm2d_123.w_0" shape = [384] dtype = "float32" min_val = float("0.32223") @@ -3023,6 +3298,7 @@ class Program_weight_tensor_parameter_274: class Program_weight_tensor_parameter_275: name = "parameter_275" + original_name = "batch_norm2d_123.w_2" shape = [384] dtype = "float32" min_val = float("5.60519e-45") @@ -3034,6 +3310,7 @@ class Program_weight_tensor_parameter_275: class Program_weight_tensor_parameter_276: name = "parameter_276" + original_name = "batch_norm2d_123.w_1" shape = [384] dtype = "float32" min_val = float("-0.3615") @@ -3045,6 +3322,7 @@ class Program_weight_tensor_parameter_276: class Program_weight_tensor_parameter_277: name = "parameter_277" + original_name = "conv2d_123.w_0" shape = [384, 384, 1, 1] dtype = "float32" min_val = float("-0.136468") @@ -3056,6 +3334,7 @@ class Program_weight_tensor_parameter_277: class Program_weight_tensor_parameter_278: name = "parameter_278" + original_name = "batch_norm2d_122.b_0" shape = [384] dtype = "float32" min_val = float("-2.46979") @@ -3067,6 +3346,7 @@ class Program_weight_tensor_parameter_278: class Program_weight_tensor_parameter_279: name = "parameter_279" + original_name = "batch_norm2d_122.w_0" shape = [384] dtype = "float32" min_val = float("0.378817") @@ -3078,6 +3358,7 @@ class Program_weight_tensor_parameter_279: class Program_weight_tensor_parameter_280: name = "parameter_280" + original_name = "batch_norm2d_122.w_2" shape = [384] dtype = "float32" min_val = float("0.0042351") @@ -3089,6 +3370,7 @@ class Program_weight_tensor_parameter_280: class Program_weight_tensor_parameter_281: name = "parameter_281" + original_name = "batch_norm2d_122.w_1" shape = [384] dtype = "float32" min_val = float("-0.220201") @@ -3100,6 +3382,7 @@ class Program_weight_tensor_parameter_281: class Program_weight_tensor_parameter_282: name = "parameter_282" + original_name = "conv2d_122.w_0" shape = [384, 1, 5, 5] dtype = "float32" min_val = float("-0.202081") @@ -3111,6 +3394,7 @@ class Program_weight_tensor_parameter_282: class Program_weight_tensor_parameter_283: name = "parameter_283" + original_name = "batch_norm2d_121.b_0" shape = [384] dtype = "float32" min_val = float("-0.730035") @@ -3122,6 +3406,7 @@ class Program_weight_tensor_parameter_283: class Program_weight_tensor_parameter_284: name = "parameter_284" + original_name = "batch_norm2d_121.w_0" shape = [384] dtype = "float32" min_val = float("0.566291") @@ -3133,6 +3418,7 @@ class Program_weight_tensor_parameter_284: class Program_weight_tensor_parameter_285: name = "parameter_285" + original_name = "batch_norm2d_121.w_2" shape = [384] dtype = "float32" min_val = float("0.00296517") @@ -3144,6 +3430,7 @@ class Program_weight_tensor_parameter_285: class Program_weight_tensor_parameter_286: name = "parameter_286" + original_name = "batch_norm2d_121.w_1" shape = [384] dtype = "float32" min_val = float("-0.141511") @@ -3155,6 +3442,7 @@ class Program_weight_tensor_parameter_286: class Program_weight_tensor_parameter_287: name = "parameter_287" + original_name = "conv2d_121.w_0" shape = [384, 384, 1, 1] dtype = "float32" min_val = float("-0.101246") @@ -3166,6 +3454,7 @@ class Program_weight_tensor_parameter_287: class Program_weight_tensor_parameter_288: name = "parameter_288" + original_name = "batch_norm2d_120.b_0" shape = [384] dtype = "float32" min_val = float("-2.79953") @@ -3177,6 +3466,7 @@ class Program_weight_tensor_parameter_288: class Program_weight_tensor_parameter_289: name = "parameter_289" + original_name = "batch_norm2d_120.w_0" shape = [384] dtype = "float32" min_val = float("0.467627") @@ -3188,6 +3478,7 @@ class Program_weight_tensor_parameter_289: class Program_weight_tensor_parameter_290: name = "parameter_290" + original_name = "batch_norm2d_120.w_2" shape = [384] dtype = "float32" min_val = float("0.00496943") @@ -3199,6 +3490,7 @@ class Program_weight_tensor_parameter_290: class Program_weight_tensor_parameter_291: name = "parameter_291" + original_name = "batch_norm2d_120.w_1" shape = [384] dtype = "float32" min_val = float("-0.287122") @@ -3210,6 +3502,7 @@ class Program_weight_tensor_parameter_291: class Program_weight_tensor_parameter_292: name = "parameter_292" + original_name = "conv2d_120.w_0" shape = [384, 1, 5, 5] dtype = "float32" min_val = float("-0.162244") @@ -3221,6 +3514,7 @@ class Program_weight_tensor_parameter_292: class Program_weight_tensor_parameter_293: name = "parameter_293" + original_name = "batch_norm2d_119.b_0" shape = [384] dtype = "float32" min_val = float("-0.928211") @@ -3232,6 +3526,7 @@ class Program_weight_tensor_parameter_293: class Program_weight_tensor_parameter_294: name = "parameter_294" + original_name = "batch_norm2d_119.w_0" shape = [384] dtype = "float32" min_val = float("0.428318") @@ -3243,6 +3538,7 @@ class Program_weight_tensor_parameter_294: class Program_weight_tensor_parameter_295: name = "parameter_295" + original_name = "batch_norm2d_119.w_2" shape = [384] dtype = "float32" min_val = float("0.0153149") @@ -3254,6 +3550,7 @@ class Program_weight_tensor_parameter_295: class Program_weight_tensor_parameter_296: name = "parameter_296" + original_name = "batch_norm2d_119.w_1" shape = [384] dtype = "float32" min_val = float("-1.06413") @@ -3265,6 +3562,7 @@ class Program_weight_tensor_parameter_296: class Program_weight_tensor_parameter_297: name = "parameter_297" + original_name = "conv2d_119.w_0" shape = [384, 384, 1, 1] dtype = "float32" min_val = float("-0.147346") @@ -3276,6 +3574,7 @@ class Program_weight_tensor_parameter_297: class Program_weight_tensor_parameter_298: name = "parameter_298" + original_name = "batch_norm2d_118.b_0" shape = [384] dtype = "float32" min_val = float("-3.06186") @@ -3287,6 +3586,7 @@ class Program_weight_tensor_parameter_298: class Program_weight_tensor_parameter_299: name = "parameter_299" + original_name = "batch_norm2d_118.w_0" shape = [384] dtype = "float32" min_val = float("0.820392") @@ -3298,6 +3598,7 @@ class Program_weight_tensor_parameter_299: class Program_weight_tensor_parameter_300: name = "parameter_300" + original_name = "batch_norm2d_118.w_2" shape = [384] dtype = "float32" min_val = float("0.0166377") @@ -3309,6 +3610,7 @@ class Program_weight_tensor_parameter_300: class Program_weight_tensor_parameter_301: name = "parameter_301" + original_name = "batch_norm2d_118.w_1" shape = [384] dtype = "float32" min_val = float("-0.341227") @@ -3320,6 +3622,7 @@ class Program_weight_tensor_parameter_301: class Program_weight_tensor_parameter_302: name = "parameter_302" + original_name = "conv2d_118.w_0" shape = [384, 1, 5, 5] dtype = "float32" min_val = float("-0.195981") @@ -3331,6 +3634,7 @@ class Program_weight_tensor_parameter_302: class Program_weight_tensor_parameter_303: name = "parameter_303" + original_name = "batch_norm2d_117.b_0" shape = [384] dtype = "float32" min_val = float("-0.946056") @@ -3342,6 +3646,7 @@ class Program_weight_tensor_parameter_303: class Program_weight_tensor_parameter_304: name = "parameter_304" + original_name = "batch_norm2d_117.w_0" shape = [384] dtype = "float32" min_val = float("0.509163") @@ -3353,6 +3658,7 @@ class Program_weight_tensor_parameter_304: class Program_weight_tensor_parameter_305: name = "parameter_305" + original_name = "batch_norm2d_117.w_2" shape = [384] dtype = "float32" min_val = float("0.0789388") @@ -3364,6 +3670,7 @@ class Program_weight_tensor_parameter_305: class Program_weight_tensor_parameter_306: name = "parameter_306" + original_name = "batch_norm2d_117.w_1" shape = [384] dtype = "float32" min_val = float("-1.17216") @@ -3375,6 +3682,7 @@ class Program_weight_tensor_parameter_306: class Program_weight_tensor_parameter_307: name = "parameter_307" + original_name = "conv2d_117.w_0" shape = [384, 1024, 1, 1] dtype = "float32" min_val = float("-0.170829") @@ -3386,6 +3694,7 @@ class Program_weight_tensor_parameter_307: class Program_weight_tensor_parameter_308: name = "parameter_308" + original_name = "batch_norm2d_116.b_0" shape = [1024] dtype = "float32" min_val = float("-2.78924") @@ -3397,6 +3706,7 @@ class Program_weight_tensor_parameter_308: class Program_weight_tensor_parameter_309: name = "parameter_309" + original_name = "batch_norm2d_116.w_0" shape = [1024] dtype = "float32" min_val = float("0.00921835") @@ -3408,6 +3718,7 @@ class Program_weight_tensor_parameter_309: class Program_weight_tensor_parameter_310: name = "parameter_310" + original_name = "batch_norm2d_116.w_2" shape = [1024] dtype = "float32" min_val = float("5.60519e-45") @@ -3419,6 +3730,7 @@ class Program_weight_tensor_parameter_310: class Program_weight_tensor_parameter_311: name = "parameter_311" + original_name = "batch_norm2d_116.w_1" shape = [1024] dtype = "float32" min_val = float("-0.45613") @@ -3430,6 +3742,7 @@ class Program_weight_tensor_parameter_311: class Program_weight_tensor_parameter_312: name = "parameter_312" + original_name = "conv2d_116.w_0" shape = [1024, 512, 1, 1] dtype = "float32" min_val = float("-0.204749") @@ -3441,6 +3754,7 @@ class Program_weight_tensor_parameter_312: class Program_weight_tensor_parameter_313: name = "parameter_313" + original_name = "batch_norm2d_115.b_0" shape = [512] dtype = "float32" min_val = float("-4.50565") @@ -3452,6 +3766,7 @@ class Program_weight_tensor_parameter_313: class Program_weight_tensor_parameter_314: name = "parameter_314" + original_name = "batch_norm2d_115.w_0" shape = [512] dtype = "float32" min_val = float("0.430967") @@ -3463,6 +3778,7 @@ class Program_weight_tensor_parameter_314: class Program_weight_tensor_parameter_315: name = "parameter_315" + original_name = "batch_norm2d_115.w_2" shape = [512] dtype = "float32" min_val = float("0.102188") @@ -3474,6 +3790,7 @@ class Program_weight_tensor_parameter_315: class Program_weight_tensor_parameter_316: name = "parameter_316" + original_name = "batch_norm2d_115.w_1" shape = [512] dtype = "float32" min_val = float("-3.34031") @@ -3485,6 +3802,7 @@ class Program_weight_tensor_parameter_316: class Program_weight_tensor_parameter_317: name = "parameter_317" + original_name = "conv2d_115.w_0" shape = [512, 3328, 1, 1] dtype = "float32" min_val = float("-0.2428") @@ -3496,6 +3814,7 @@ class Program_weight_tensor_parameter_317: class Program_weight_tensor_parameter_318: name = "parameter_318" + original_name = "batch_norm2d_114.b_0" shape = [384] dtype = "float32" min_val = float("-1.23258") @@ -3507,6 +3826,7 @@ class Program_weight_tensor_parameter_318: class Program_weight_tensor_parameter_319: name = "parameter_319" + original_name = "batch_norm2d_114.w_0" shape = [384] dtype = "float32" min_val = float("0.553337") @@ -3518,6 +3838,7 @@ class Program_weight_tensor_parameter_319: class Program_weight_tensor_parameter_320: name = "parameter_320" + original_name = "batch_norm2d_114.w_2" shape = [384] dtype = "float32" min_val = float("5.60519e-45") @@ -3529,6 +3850,7 @@ class Program_weight_tensor_parameter_320: class Program_weight_tensor_parameter_321: name = "parameter_321" + original_name = "batch_norm2d_114.w_1" shape = [384] dtype = "float32" min_val = float("-0.352255") @@ -3540,6 +3862,7 @@ class Program_weight_tensor_parameter_321: class Program_weight_tensor_parameter_322: name = "parameter_322" + original_name = "conv2d_114.w_0" shape = [384, 1, 5, 5] dtype = "float32" min_val = float("-0.0844072") @@ -3551,6 +3874,7 @@ class Program_weight_tensor_parameter_322: class Program_weight_tensor_parameter_323: name = "parameter_323" + original_name = "batch_norm2d_113.b_0" shape = [384] dtype = "float32" min_val = float("-0.915401") @@ -3562,6 +3886,7 @@ class Program_weight_tensor_parameter_323: class Program_weight_tensor_parameter_324: name = "parameter_324" + original_name = "batch_norm2d_113.w_0" shape = [384] dtype = "float32" min_val = float("0.473174") @@ -3573,6 +3898,7 @@ class Program_weight_tensor_parameter_324: class Program_weight_tensor_parameter_325: name = "parameter_325" + original_name = "batch_norm2d_113.w_2" shape = [384] dtype = "float32" min_val = float("5.60519e-45") @@ -3584,6 +3910,7 @@ class Program_weight_tensor_parameter_325: class Program_weight_tensor_parameter_326: name = "parameter_326" + original_name = "batch_norm2d_113.w_1" shape = [384] dtype = "float32" min_val = float("-0.23617") @@ -3595,6 +3922,7 @@ class Program_weight_tensor_parameter_326: class Program_weight_tensor_parameter_327: name = "parameter_327" + original_name = "conv2d_113.w_0" shape = [384, 384, 1, 1] dtype = "float32" min_val = float("-0.216627") @@ -3606,6 +3934,7 @@ class Program_weight_tensor_parameter_327: class Program_weight_tensor_parameter_328: name = "parameter_328" + original_name = "batch_norm2d_112.b_0" shape = [384] dtype = "float32" min_val = float("-0.472409") @@ -3617,6 +3946,7 @@ class Program_weight_tensor_parameter_328: class Program_weight_tensor_parameter_329: name = "parameter_329" + original_name = "batch_norm2d_112.w_0" shape = [384] dtype = "float32" min_val = float("0.267028") @@ -3628,6 +3958,7 @@ class Program_weight_tensor_parameter_329: class Program_weight_tensor_parameter_330: name = "parameter_330" + original_name = "batch_norm2d_112.w_2" shape = [384] dtype = "float32" min_val = float("5.60519e-45") @@ -3639,6 +3970,7 @@ class Program_weight_tensor_parameter_330: class Program_weight_tensor_parameter_331: name = "parameter_331" + original_name = "batch_norm2d_112.w_1" shape = [384] dtype = "float32" min_val = float("-0.0810151") @@ -3650,6 +3982,7 @@ class Program_weight_tensor_parameter_331: class Program_weight_tensor_parameter_332: name = "parameter_332" + original_name = "conv2d_112.w_0" shape = [384, 1, 5, 5] dtype = "float32" min_val = float("-0.112119") @@ -3661,6 +3994,7 @@ class Program_weight_tensor_parameter_332: class Program_weight_tensor_parameter_333: name = "parameter_333" + original_name = "batch_norm2d_111.b_0" shape = [384] dtype = "float32" min_val = float("-0.84325") @@ -3672,6 +4006,7 @@ class Program_weight_tensor_parameter_333: class Program_weight_tensor_parameter_334: name = "parameter_334" + original_name = "batch_norm2d_111.w_0" shape = [384] dtype = "float32" min_val = float("0.246913") @@ -3683,6 +4018,7 @@ class Program_weight_tensor_parameter_334: class Program_weight_tensor_parameter_335: name = "parameter_335" + original_name = "batch_norm2d_111.w_2" shape = [384] dtype = "float32" min_val = float("5.60519e-45") @@ -3694,6 +4030,7 @@ class Program_weight_tensor_parameter_335: class Program_weight_tensor_parameter_336: name = "parameter_336" + original_name = "batch_norm2d_111.w_1" shape = [384] dtype = "float32" min_val = float("-0.444659") @@ -3705,6 +4042,7 @@ class Program_weight_tensor_parameter_336: class Program_weight_tensor_parameter_337: name = "parameter_337" + original_name = "conv2d_111.w_0" shape = [384, 384, 1, 1] dtype = "float32" min_val = float("-0.251538") @@ -3716,6 +4054,7 @@ class Program_weight_tensor_parameter_337: class Program_weight_tensor_parameter_338: name = "parameter_338" + original_name = "batch_norm2d_110.b_0" shape = [384] dtype = "float32" min_val = float("-1.04733") @@ -3727,6 +4066,7 @@ class Program_weight_tensor_parameter_338: class Program_weight_tensor_parameter_339: name = "parameter_339" + original_name = "batch_norm2d_110.w_0" shape = [384] dtype = "float32" min_val = float("0.331571") @@ -3738,6 +4078,7 @@ class Program_weight_tensor_parameter_339: class Program_weight_tensor_parameter_340: name = "parameter_340" + original_name = "batch_norm2d_110.w_2" shape = [384] dtype = "float32" min_val = float("5.60519e-45") @@ -3749,6 +4090,7 @@ class Program_weight_tensor_parameter_340: class Program_weight_tensor_parameter_341: name = "parameter_341" + original_name = "batch_norm2d_110.w_1" shape = [384] dtype = "float32" min_val = float("-0.41206") @@ -3760,6 +4102,7 @@ class Program_weight_tensor_parameter_341: class Program_weight_tensor_parameter_342: name = "parameter_342" + original_name = "conv2d_110.w_0" shape = [384, 1, 5, 5] dtype = "float32" min_val = float("-0.135509") @@ -3771,6 +4114,7 @@ class Program_weight_tensor_parameter_342: class Program_weight_tensor_parameter_343: name = "parameter_343" + original_name = "batch_norm2d_109.b_0" shape = [384] dtype = "float32" min_val = float("-1.0108") @@ -3782,6 +4126,7 @@ class Program_weight_tensor_parameter_343: class Program_weight_tensor_parameter_344: name = "parameter_344" + original_name = "batch_norm2d_109.w_0" shape = [384] dtype = "float32" min_val = float("0.311787") @@ -3793,6 +4138,7 @@ class Program_weight_tensor_parameter_344: class Program_weight_tensor_parameter_345: name = "parameter_345" + original_name = "batch_norm2d_109.w_2" shape = [384] dtype = "float32" min_val = float("5.60519e-45") @@ -3804,6 +4150,7 @@ class Program_weight_tensor_parameter_345: class Program_weight_tensor_parameter_346: name = "parameter_346" + original_name = "batch_norm2d_109.w_1" shape = [384] dtype = "float32" min_val = float("-0.236032") @@ -3815,6 +4162,7 @@ class Program_weight_tensor_parameter_346: class Program_weight_tensor_parameter_347: name = "parameter_347" + original_name = "conv2d_109.w_0" shape = [384, 384, 1, 1] dtype = "float32" min_val = float("-0.149638") @@ -3826,6 +4174,7 @@ class Program_weight_tensor_parameter_347: class Program_weight_tensor_parameter_348: name = "parameter_348" + original_name = "batch_norm2d_108.b_0" shape = [384] dtype = "float32" min_val = float("-1.9204") @@ -3837,6 +4186,7 @@ class Program_weight_tensor_parameter_348: class Program_weight_tensor_parameter_349: name = "parameter_349" + original_name = "batch_norm2d_108.w_0" shape = [384] dtype = "float32" min_val = float("0.40205") @@ -3848,6 +4198,7 @@ class Program_weight_tensor_parameter_349: class Program_weight_tensor_parameter_350: name = "parameter_350" + original_name = "batch_norm2d_108.w_2" shape = [384] dtype = "float32" min_val = float("5.60519e-45") @@ -3859,6 +4210,7 @@ class Program_weight_tensor_parameter_350: class Program_weight_tensor_parameter_351: name = "parameter_351" + original_name = "batch_norm2d_108.w_1" shape = [384] dtype = "float32" min_val = float("-0.306255") @@ -3870,6 +4222,7 @@ class Program_weight_tensor_parameter_351: class Program_weight_tensor_parameter_352: name = "parameter_352" + original_name = "conv2d_108.w_0" shape = [384, 1, 5, 5] dtype = "float32" min_val = float("-0.114348") @@ -3881,6 +4234,7 @@ class Program_weight_tensor_parameter_352: class Program_weight_tensor_parameter_353: name = "parameter_353" + original_name = "batch_norm2d_107.b_0" shape = [384] dtype = "float32" min_val = float("-0.896639") @@ -3892,6 +4246,7 @@ class Program_weight_tensor_parameter_353: class Program_weight_tensor_parameter_354: name = "parameter_354" + original_name = "batch_norm2d_107.w_0" shape = [384] dtype = "float32" min_val = float("0.545342") @@ -3903,6 +4258,7 @@ class Program_weight_tensor_parameter_354: class Program_weight_tensor_parameter_355: name = "parameter_355" + original_name = "batch_norm2d_107.w_2" shape = [384] dtype = "float32" min_val = float("5.60519e-45") @@ -3914,6 +4270,7 @@ class Program_weight_tensor_parameter_355: class Program_weight_tensor_parameter_356: name = "parameter_356" + original_name = "batch_norm2d_107.w_1" shape = [384] dtype = "float32" min_val = float("-0.206759") @@ -3925,6 +4282,7 @@ class Program_weight_tensor_parameter_356: class Program_weight_tensor_parameter_357: name = "parameter_357" + original_name = "conv2d_107.w_0" shape = [384, 384, 1, 1] dtype = "float32" min_val = float("-0.200029") @@ -3936,6 +4294,7 @@ class Program_weight_tensor_parameter_357: class Program_weight_tensor_parameter_358: name = "parameter_358" + original_name = "batch_norm2d_106.b_0" shape = [384] dtype = "float32" min_val = float("-2.24976") @@ -3947,6 +4306,7 @@ class Program_weight_tensor_parameter_358: class Program_weight_tensor_parameter_359: name = "parameter_359" + original_name = "batch_norm2d_106.w_0" shape = [384] dtype = "float32" min_val = float("0.37159") @@ -3958,6 +4318,7 @@ class Program_weight_tensor_parameter_359: class Program_weight_tensor_parameter_360: name = "parameter_360" + original_name = "batch_norm2d_106.w_2" shape = [384] dtype = "float32" min_val = float("0.00634831") @@ -3969,6 +4330,7 @@ class Program_weight_tensor_parameter_360: class Program_weight_tensor_parameter_361: name = "parameter_361" + original_name = "batch_norm2d_106.w_1" shape = [384] dtype = "float32" min_val = float("-0.31471") @@ -3980,6 +4342,7 @@ class Program_weight_tensor_parameter_361: class Program_weight_tensor_parameter_362: name = "parameter_362" + original_name = "conv2d_106.w_0" shape = [384, 1, 5, 5] dtype = "float32" min_val = float("-0.16588") @@ -3991,6 +4354,7 @@ class Program_weight_tensor_parameter_362: class Program_weight_tensor_parameter_363: name = "parameter_363" + original_name = "batch_norm2d_105.b_0" shape = [384] dtype = "float32" min_val = float("-0.887521") @@ -4002,6 +4366,7 @@ class Program_weight_tensor_parameter_363: class Program_weight_tensor_parameter_364: name = "parameter_364" + original_name = "batch_norm2d_105.w_0" shape = [384] dtype = "float32" min_val = float("0.548948") @@ -4013,6 +4378,7 @@ class Program_weight_tensor_parameter_364: class Program_weight_tensor_parameter_365: name = "parameter_365" + original_name = "batch_norm2d_105.w_2" shape = [384] dtype = "float32" min_val = float("0.0147972") @@ -4024,6 +4390,7 @@ class Program_weight_tensor_parameter_365: class Program_weight_tensor_parameter_366: name = "parameter_366" + original_name = "batch_norm2d_105.w_1" shape = [384] dtype = "float32" min_val = float("-0.663477") @@ -4035,6 +4402,7 @@ class Program_weight_tensor_parameter_366: class Program_weight_tensor_parameter_367: name = "parameter_367" + original_name = "conv2d_105.w_0" shape = [384, 384, 1, 1] dtype = "float32" min_val = float("-0.429613") @@ -4046,6 +4414,7 @@ class Program_weight_tensor_parameter_367: class Program_weight_tensor_parameter_368: name = "parameter_368" + original_name = "batch_norm2d_104.b_0" shape = [384] dtype = "float32" min_val = float("-3.25349") @@ -4057,6 +4426,7 @@ class Program_weight_tensor_parameter_368: class Program_weight_tensor_parameter_369: name = "parameter_369" + original_name = "batch_norm2d_104.w_0" shape = [384] dtype = "float32" min_val = float("0.532176") @@ -4068,6 +4438,7 @@ class Program_weight_tensor_parameter_369: class Program_weight_tensor_parameter_370: name = "parameter_370" + original_name = "batch_norm2d_104.w_2" shape = [384] dtype = "float32" min_val = float("0.00753172") @@ -4079,6 +4450,7 @@ class Program_weight_tensor_parameter_370: class Program_weight_tensor_parameter_371: name = "parameter_371" + original_name = "batch_norm2d_104.w_1" shape = [384] dtype = "float32" min_val = float("-0.30962") @@ -4090,6 +4462,7 @@ class Program_weight_tensor_parameter_371: class Program_weight_tensor_parameter_372: name = "parameter_372" + original_name = "conv2d_104.w_0" shape = [384, 1, 5, 5] dtype = "float32" min_val = float("-0.316701") @@ -4101,6 +4474,7 @@ class Program_weight_tensor_parameter_372: class Program_weight_tensor_parameter_373: name = "parameter_373" + original_name = "batch_norm2d_103.b_0" shape = [384] dtype = "float32" min_val = float("-1.11574") @@ -4112,6 +4486,7 @@ class Program_weight_tensor_parameter_373: class Program_weight_tensor_parameter_374: name = "parameter_374" + original_name = "batch_norm2d_103.w_0" shape = [384] dtype = "float32" min_val = float("0.234984") @@ -4123,6 +4498,7 @@ class Program_weight_tensor_parameter_374: class Program_weight_tensor_parameter_375: name = "parameter_375" + original_name = "batch_norm2d_103.w_2" shape = [384] dtype = "float32" min_val = float("0.0373457") @@ -4134,6 +4510,7 @@ class Program_weight_tensor_parameter_375: class Program_weight_tensor_parameter_376: name = "parameter_376" + original_name = "batch_norm2d_103.w_1" shape = [384] dtype = "float32" min_val = float("-0.864794") @@ -4145,6 +4522,7 @@ class Program_weight_tensor_parameter_376: class Program_weight_tensor_parameter_377: name = "parameter_377" + original_name = "conv2d_103.w_0" shape = [384, 1024, 1, 1] dtype = "float32" min_val = float("-0.226582") @@ -4156,6 +4534,7 @@ class Program_weight_tensor_parameter_377: class Program_weight_tensor_parameter_378: name = "parameter_378" + original_name = "batch_norm2d_102.b_0" shape = [1024] dtype = "float32" min_val = float("-2.78604") @@ -4167,6 +4546,7 @@ class Program_weight_tensor_parameter_378: class Program_weight_tensor_parameter_379: name = "parameter_379" + original_name = "batch_norm2d_102.w_0" shape = [1024] dtype = "float32" min_val = float("0.0309497") @@ -4178,6 +4558,7 @@ class Program_weight_tensor_parameter_379: class Program_weight_tensor_parameter_380: name = "parameter_380" + original_name = "batch_norm2d_102.w_2" shape = [1024] dtype = "float32" min_val = float("5.60519e-45") @@ -4189,6 +4570,7 @@ class Program_weight_tensor_parameter_380: class Program_weight_tensor_parameter_381: name = "parameter_381" + original_name = "batch_norm2d_102.w_1" shape = [1024] dtype = "float32" min_val = float("-0.391039") @@ -4200,6 +4582,7 @@ class Program_weight_tensor_parameter_381: class Program_weight_tensor_parameter_382: name = "parameter_382" + original_name = "conv2d_102.w_0" shape = [1024, 512, 1, 1] dtype = "float32" min_val = float("-0.261529") @@ -4211,6 +4594,7 @@ class Program_weight_tensor_parameter_382: class Program_weight_tensor_parameter_383: name = "parameter_383" + original_name = "batch_norm2d_101.b_0" shape = [512] dtype = "float32" min_val = float("-4.1697") @@ -4222,6 +4606,7 @@ class Program_weight_tensor_parameter_383: class Program_weight_tensor_parameter_384: name = "parameter_384" + original_name = "batch_norm2d_101.w_0" shape = [512] dtype = "float32" min_val = float("0.492087") @@ -4233,6 +4618,7 @@ class Program_weight_tensor_parameter_384: class Program_weight_tensor_parameter_385: name = "parameter_385" + original_name = "batch_norm2d_101.w_2" shape = [512] dtype = "float32" min_val = float("0.115103") @@ -4244,6 +4630,7 @@ class Program_weight_tensor_parameter_385: class Program_weight_tensor_parameter_386: name = "parameter_386" + original_name = "batch_norm2d_101.w_1" shape = [512] dtype = "float32" min_val = float("-1.21915") @@ -4255,6 +4642,7 @@ class Program_weight_tensor_parameter_386: class Program_weight_tensor_parameter_387: name = "parameter_387" + original_name = "conv2d_101.w_0" shape = [512, 3328, 1, 1] dtype = "float32" min_val = float("-0.20937") @@ -4266,6 +4654,7 @@ class Program_weight_tensor_parameter_387: class Program_weight_tensor_parameter_388: name = "parameter_388" + original_name = "batch_norm2d_100.b_0" shape = [384] dtype = "float32" min_val = float("-0.762217") @@ -4277,6 +4666,7 @@ class Program_weight_tensor_parameter_388: class Program_weight_tensor_parameter_389: name = "parameter_389" + original_name = "batch_norm2d_100.w_0" shape = [384] dtype = "float32" min_val = float("0.53178") @@ -4288,6 +4678,7 @@ class Program_weight_tensor_parameter_389: class Program_weight_tensor_parameter_390: name = "parameter_390" + original_name = "batch_norm2d_100.w_2" shape = [384] dtype = "float32" min_val = float("5.60519e-45") @@ -4299,6 +4690,7 @@ class Program_weight_tensor_parameter_390: class Program_weight_tensor_parameter_391: name = "parameter_391" + original_name = "batch_norm2d_100.w_1" shape = [384] dtype = "float32" min_val = float("-0.232276") @@ -4310,6 +4702,7 @@ class Program_weight_tensor_parameter_391: class Program_weight_tensor_parameter_392: name = "parameter_392" + original_name = "conv2d_100.w_0" shape = [384, 1, 5, 5] dtype = "float32" min_val = float("-0.0710113") @@ -4321,6 +4714,7 @@ class Program_weight_tensor_parameter_392: class Program_weight_tensor_parameter_393: name = "parameter_393" + original_name = "batch_norm2d_99.b_0" shape = [384] dtype = "float32" min_val = float("-0.820711") @@ -4332,6 +4726,7 @@ class Program_weight_tensor_parameter_393: class Program_weight_tensor_parameter_394: name = "parameter_394" + original_name = "batch_norm2d_99.w_0" shape = [384] dtype = "float32" min_val = float("0.598571") @@ -4343,6 +4738,7 @@ class Program_weight_tensor_parameter_394: class Program_weight_tensor_parameter_395: name = "parameter_395" + original_name = "batch_norm2d_99.w_2" shape = [384] dtype = "float32" min_val = float("5.60519e-45") @@ -4354,6 +4750,7 @@ class Program_weight_tensor_parameter_395: class Program_weight_tensor_parameter_396: name = "parameter_396" + original_name = "batch_norm2d_99.w_1" shape = [384] dtype = "float32" min_val = float("-0.293963") @@ -4365,6 +4762,7 @@ class Program_weight_tensor_parameter_396: class Program_weight_tensor_parameter_397: name = "parameter_397" + original_name = "conv2d_99.w_0" shape = [384, 384, 1, 1] dtype = "float32" min_val = float("-0.14405") @@ -4376,6 +4774,7 @@ class Program_weight_tensor_parameter_397: class Program_weight_tensor_parameter_398: name = "parameter_398" + original_name = "batch_norm2d_98.b_0" shape = [384] dtype = "float32" min_val = float("-1.98677") @@ -4387,6 +4786,7 @@ class Program_weight_tensor_parameter_398: class Program_weight_tensor_parameter_399: name = "parameter_399" + original_name = "batch_norm2d_98.w_0" shape = [384] dtype = "float32" min_val = float("0.267524") @@ -4398,6 +4798,7 @@ class Program_weight_tensor_parameter_399: class Program_weight_tensor_parameter_400: name = "parameter_400" + original_name = "batch_norm2d_98.w_2" shape = [384] dtype = "float32" min_val = float("5.60519e-45") @@ -4409,6 +4810,7 @@ class Program_weight_tensor_parameter_400: class Program_weight_tensor_parameter_401: name = "parameter_401" + original_name = "batch_norm2d_98.w_1" shape = [384] dtype = "float32" min_val = float("-0.142137") @@ -4420,6 +4822,7 @@ class Program_weight_tensor_parameter_401: class Program_weight_tensor_parameter_402: name = "parameter_402" + original_name = "conv2d_98.w_0" shape = [384, 1, 5, 5] dtype = "float32" min_val = float("-0.130652") @@ -4431,6 +4834,7 @@ class Program_weight_tensor_parameter_402: class Program_weight_tensor_parameter_403: name = "parameter_403" + original_name = "batch_norm2d_97.b_0" shape = [384] dtype = "float32" min_val = float("-0.650621") @@ -4442,6 +4846,7 @@ class Program_weight_tensor_parameter_403: class Program_weight_tensor_parameter_404: name = "parameter_404" + original_name = "batch_norm2d_97.w_0" shape = [384] dtype = "float32" min_val = float("0.761444") @@ -4453,6 +4858,7 @@ class Program_weight_tensor_parameter_404: class Program_weight_tensor_parameter_405: name = "parameter_405" + original_name = "batch_norm2d_97.w_2" shape = [384] dtype = "float32" min_val = float("5.60519e-45") @@ -4464,6 +4870,7 @@ class Program_weight_tensor_parameter_405: class Program_weight_tensor_parameter_406: name = "parameter_406" + original_name = "batch_norm2d_97.w_1" shape = [384] dtype = "float32" min_val = float("-0.127518") @@ -4475,6 +4882,7 @@ class Program_weight_tensor_parameter_406: class Program_weight_tensor_parameter_407: name = "parameter_407" + original_name = "conv2d_97.w_0" shape = [384, 384, 1, 1] dtype = "float32" min_val = float("-0.0986328") @@ -4486,6 +4894,7 @@ class Program_weight_tensor_parameter_407: class Program_weight_tensor_parameter_408: name = "parameter_408" + original_name = "batch_norm2d_96.b_0" shape = [384] dtype = "float32" min_val = float("-1.47185") @@ -4497,6 +4906,7 @@ class Program_weight_tensor_parameter_408: class Program_weight_tensor_parameter_409: name = "parameter_409" + original_name = "batch_norm2d_96.w_0" shape = [384] dtype = "float32" min_val = float("0.433124") @@ -4508,6 +4918,7 @@ class Program_weight_tensor_parameter_409: class Program_weight_tensor_parameter_410: name = "parameter_410" + original_name = "batch_norm2d_96.w_2" shape = [384] dtype = "float32" min_val = float("5.60519e-45") @@ -4519,6 +4930,7 @@ class Program_weight_tensor_parameter_410: class Program_weight_tensor_parameter_411: name = "parameter_411" + original_name = "batch_norm2d_96.w_1" shape = [384] dtype = "float32" min_val = float("-0.260155") @@ -4530,6 +4942,7 @@ class Program_weight_tensor_parameter_411: class Program_weight_tensor_parameter_412: name = "parameter_412" + original_name = "conv2d_96.w_0" shape = [384, 1, 5, 5] dtype = "float32" min_val = float("-0.0992104") @@ -4541,6 +4954,7 @@ class Program_weight_tensor_parameter_412: class Program_weight_tensor_parameter_413: name = "parameter_413" + original_name = "batch_norm2d_95.b_0" shape = [384] dtype = "float32" min_val = float("-0.907741") @@ -4552,6 +4966,7 @@ class Program_weight_tensor_parameter_413: class Program_weight_tensor_parameter_414: name = "parameter_414" + original_name = "batch_norm2d_95.w_0" shape = [384] dtype = "float32" min_val = float("0.455687") @@ -4563,6 +4978,7 @@ class Program_weight_tensor_parameter_414: class Program_weight_tensor_parameter_415: name = "parameter_415" + original_name = "batch_norm2d_95.w_2" shape = [384] dtype = "float32" min_val = float("5.60519e-45") @@ -4574,6 +4990,7 @@ class Program_weight_tensor_parameter_415: class Program_weight_tensor_parameter_416: name = "parameter_416" + original_name = "batch_norm2d_95.w_1" shape = [384] dtype = "float32" min_val = float("-0.255064") @@ -4585,6 +5002,7 @@ class Program_weight_tensor_parameter_416: class Program_weight_tensor_parameter_417: name = "parameter_417" + original_name = "conv2d_95.w_0" shape = [384, 384, 1, 1] dtype = "float32" min_val = float("-0.0971502") @@ -4596,6 +5014,7 @@ class Program_weight_tensor_parameter_417: class Program_weight_tensor_parameter_418: name = "parameter_418" + original_name = "batch_norm2d_94.b_0" shape = [384] dtype = "float32" min_val = float("-2.22373") @@ -4607,6 +5026,7 @@ class Program_weight_tensor_parameter_418: class Program_weight_tensor_parameter_419: name = "parameter_419" + original_name = "batch_norm2d_94.w_0" shape = [384] dtype = "float32" min_val = float("0.48269") @@ -4618,6 +5038,7 @@ class Program_weight_tensor_parameter_419: class Program_weight_tensor_parameter_420: name = "parameter_420" + original_name = "batch_norm2d_94.w_2" shape = [384] dtype = "float32" min_val = float("0.00412925") @@ -4629,6 +5050,7 @@ class Program_weight_tensor_parameter_420: class Program_weight_tensor_parameter_421: name = "parameter_421" + original_name = "batch_norm2d_94.w_1" shape = [384] dtype = "float32" min_val = float("-0.446876") @@ -4640,6 +5062,7 @@ class Program_weight_tensor_parameter_421: class Program_weight_tensor_parameter_422: name = "parameter_422" + original_name = "conv2d_94.w_0" shape = [384, 1, 5, 5] dtype = "float32" min_val = float("-0.149308") @@ -4651,6 +5074,7 @@ class Program_weight_tensor_parameter_422: class Program_weight_tensor_parameter_423: name = "parameter_423" + original_name = "batch_norm2d_93.b_0" shape = [384] dtype = "float32" min_val = float("-1.06877") @@ -4662,6 +5086,7 @@ class Program_weight_tensor_parameter_423: class Program_weight_tensor_parameter_424: name = "parameter_424" + original_name = "batch_norm2d_93.w_0" shape = [384] dtype = "float32" min_val = float("0.287864") @@ -4673,6 +5098,7 @@ class Program_weight_tensor_parameter_424: class Program_weight_tensor_parameter_425: name = "parameter_425" + original_name = "batch_norm2d_93.w_2" shape = [384] dtype = "float32" min_val = float("0.00199839") @@ -4684,6 +5110,7 @@ class Program_weight_tensor_parameter_425: class Program_weight_tensor_parameter_426: name = "parameter_426" + original_name = "batch_norm2d_93.w_1" shape = [384] dtype = "float32" min_val = float("-0.204615") @@ -4695,6 +5122,7 @@ class Program_weight_tensor_parameter_426: class Program_weight_tensor_parameter_427: name = "parameter_427" + original_name = "conv2d_93.w_0" shape = [384, 384, 1, 1] dtype = "float32" min_val = float("-0.117092") @@ -4706,6 +5134,7 @@ class Program_weight_tensor_parameter_427: class Program_weight_tensor_parameter_428: name = "parameter_428" + original_name = "batch_norm2d_92.b_0" shape = [384] dtype = "float32" min_val = float("-2.84955") @@ -4717,6 +5146,7 @@ class Program_weight_tensor_parameter_428: class Program_weight_tensor_parameter_429: name = "parameter_429" + original_name = "batch_norm2d_92.w_0" shape = [384] dtype = "float32" min_val = float("0.533484") @@ -4728,6 +5158,7 @@ class Program_weight_tensor_parameter_429: class Program_weight_tensor_parameter_430: name = "parameter_430" + original_name = "batch_norm2d_92.w_2" shape = [384] dtype = "float32" min_val = float("0.00688426") @@ -4739,6 +5170,7 @@ class Program_weight_tensor_parameter_430: class Program_weight_tensor_parameter_431: name = "parameter_431" + original_name = "batch_norm2d_92.w_1" shape = [384] dtype = "float32" min_val = float("-0.255715") @@ -4750,6 +5182,7 @@ class Program_weight_tensor_parameter_431: class Program_weight_tensor_parameter_432: name = "parameter_432" + original_name = "conv2d_92.w_0" shape = [384, 1, 5, 5] dtype = "float32" min_val = float("-0.17124") @@ -4761,6 +5194,7 @@ class Program_weight_tensor_parameter_432: class Program_weight_tensor_parameter_433: name = "parameter_433" + original_name = "batch_norm2d_91.b_0" shape = [384] dtype = "float32" min_val = float("-0.839826") @@ -4772,6 +5206,7 @@ class Program_weight_tensor_parameter_433: class Program_weight_tensor_parameter_434: name = "parameter_434" + original_name = "batch_norm2d_91.w_0" shape = [384] dtype = "float32" min_val = float("0.62843") @@ -4783,6 +5218,7 @@ class Program_weight_tensor_parameter_434: class Program_weight_tensor_parameter_435: name = "parameter_435" + original_name = "batch_norm2d_91.w_2" shape = [384] dtype = "float32" min_val = float("0.00879723") @@ -4794,6 +5230,7 @@ class Program_weight_tensor_parameter_435: class Program_weight_tensor_parameter_436: name = "parameter_436" + original_name = "batch_norm2d_91.w_1" shape = [384] dtype = "float32" min_val = float("-0.466733") @@ -4805,6 +5242,7 @@ class Program_weight_tensor_parameter_436: class Program_weight_tensor_parameter_437: name = "parameter_437" + original_name = "conv2d_91.w_0" shape = [384, 384, 1, 1] dtype = "float32" min_val = float("-0.109293") @@ -4816,6 +5254,7 @@ class Program_weight_tensor_parameter_437: class Program_weight_tensor_parameter_438: name = "parameter_438" + original_name = "batch_norm2d_90.b_0" shape = [384] dtype = "float32" min_val = float("-3.00598") @@ -4827,6 +5266,7 @@ class Program_weight_tensor_parameter_438: class Program_weight_tensor_parameter_439: name = "parameter_439" + original_name = "batch_norm2d_90.w_0" shape = [384] dtype = "float32" min_val = float("0.671459") @@ -4838,6 +5278,7 @@ class Program_weight_tensor_parameter_439: class Program_weight_tensor_parameter_440: name = "parameter_440" + original_name = "batch_norm2d_90.w_2" shape = [384] dtype = "float32" min_val = float("0.0103478") @@ -4849,6 +5290,7 @@ class Program_weight_tensor_parameter_440: class Program_weight_tensor_parameter_441: name = "parameter_441" + original_name = "batch_norm2d_90.w_1" shape = [384] dtype = "float32" min_val = float("-0.338717") @@ -4860,6 +5302,7 @@ class Program_weight_tensor_parameter_441: class Program_weight_tensor_parameter_442: name = "parameter_442" + original_name = "conv2d_90.w_0" shape = [384, 1, 5, 5] dtype = "float32" min_val = float("-0.262848") @@ -4871,6 +5314,7 @@ class Program_weight_tensor_parameter_442: class Program_weight_tensor_parameter_443: name = "parameter_443" + original_name = "batch_norm2d_89.b_0" shape = [384] dtype = "float32" min_val = float("-0.942343") @@ -4882,6 +5326,7 @@ class Program_weight_tensor_parameter_443: class Program_weight_tensor_parameter_444: name = "parameter_444" + original_name = "batch_norm2d_89.w_0" shape = [384] dtype = "float32" min_val = float("0.536809") @@ -4893,6 +5338,7 @@ class Program_weight_tensor_parameter_444: class Program_weight_tensor_parameter_445: name = "parameter_445" + original_name = "batch_norm2d_89.w_2" shape = [384] dtype = "float32" min_val = float("0.0562657") @@ -4904,6 +5350,7 @@ class Program_weight_tensor_parameter_445: class Program_weight_tensor_parameter_446: name = "parameter_446" + original_name = "batch_norm2d_89.w_1" shape = [384] dtype = "float32" min_val = float("-0.854075") @@ -4915,6 +5362,7 @@ class Program_weight_tensor_parameter_446: class Program_weight_tensor_parameter_447: name = "parameter_447" + original_name = "conv2d_89.w_0" shape = [384, 1024, 1, 1] dtype = "float32" min_val = float("-0.170074") @@ -4926,6 +5374,7 @@ class Program_weight_tensor_parameter_447: class Program_weight_tensor_parameter_448: name = "parameter_448" + original_name = "batch_norm2d_88.b_0" shape = [1024] dtype = "float32" min_val = float("-2.18276") @@ -4937,6 +5386,7 @@ class Program_weight_tensor_parameter_448: class Program_weight_tensor_parameter_449: name = "parameter_449" + original_name = "batch_norm2d_88.w_0" shape = [1024] dtype = "float32" min_val = float("0.0387132") @@ -4948,6 +5398,7 @@ class Program_weight_tensor_parameter_449: class Program_weight_tensor_parameter_450: name = "parameter_450" + original_name = "batch_norm2d_88.w_2" shape = [1024] dtype = "float32" min_val = float("5.60519e-45") @@ -4959,6 +5410,7 @@ class Program_weight_tensor_parameter_450: class Program_weight_tensor_parameter_451: name = "parameter_451" + original_name = "batch_norm2d_88.w_1" shape = [1024] dtype = "float32" min_val = float("-0.427382") @@ -4970,6 +5422,7 @@ class Program_weight_tensor_parameter_451: class Program_weight_tensor_parameter_452: name = "parameter_452" + original_name = "conv2d_88.w_0" shape = [1024, 512, 1, 1] dtype = "float32" min_val = float("-0.300386") @@ -4981,6 +5434,7 @@ class Program_weight_tensor_parameter_452: class Program_weight_tensor_parameter_453: name = "parameter_453" + original_name = "batch_norm2d_87.b_0" shape = [512] dtype = "float32" min_val = float("-4.71042") @@ -4992,6 +5446,7 @@ class Program_weight_tensor_parameter_453: class Program_weight_tensor_parameter_454: name = "parameter_454" + original_name = "batch_norm2d_87.w_0" shape = [512] dtype = "float32" min_val = float("0.53064") @@ -5003,6 +5458,7 @@ class Program_weight_tensor_parameter_454: class Program_weight_tensor_parameter_455: name = "parameter_455" + original_name = "batch_norm2d_87.w_2" shape = [512] dtype = "float32" min_val = float("0.102476") @@ -5014,6 +5470,7 @@ class Program_weight_tensor_parameter_455: class Program_weight_tensor_parameter_456: name = "parameter_456" + original_name = "batch_norm2d_87.w_1" shape = [512] dtype = "float32" min_val = float("-2.2863") @@ -5025,6 +5482,7 @@ class Program_weight_tensor_parameter_456: class Program_weight_tensor_parameter_457: name = "parameter_457" + original_name = "conv2d_87.w_0" shape = [512, 3328, 1, 1] dtype = "float32" min_val = float("-0.29429") @@ -5036,6 +5494,7 @@ class Program_weight_tensor_parameter_457: class Program_weight_tensor_parameter_458: name = "parameter_458" + original_name = "batch_norm2d_86.b_0" shape = [384] dtype = "float32" min_val = float("-1.34148") @@ -5047,6 +5506,7 @@ class Program_weight_tensor_parameter_458: class Program_weight_tensor_parameter_459: name = "parameter_459" + original_name = "batch_norm2d_86.w_0" shape = [384] dtype = "float32" min_val = float("0.500453") @@ -5058,6 +5518,7 @@ class Program_weight_tensor_parameter_459: class Program_weight_tensor_parameter_460: name = "parameter_460" + original_name = "batch_norm2d_86.w_2" shape = [384] dtype = "float32" min_val = float("5.60519e-45") @@ -5069,6 +5530,7 @@ class Program_weight_tensor_parameter_460: class Program_weight_tensor_parameter_461: name = "parameter_461" + original_name = "batch_norm2d_86.w_1" shape = [384] dtype = "float32" min_val = float("-0.338465") @@ -5080,6 +5542,7 @@ class Program_weight_tensor_parameter_461: class Program_weight_tensor_parameter_462: name = "parameter_462" + original_name = "conv2d_86.w_0" shape = [384, 1, 5, 5] dtype = "float32" min_val = float("-0.141599") @@ -5091,6 +5554,7 @@ class Program_weight_tensor_parameter_462: class Program_weight_tensor_parameter_463: name = "parameter_463" + original_name = "batch_norm2d_85.b_0" shape = [384] dtype = "float32" min_val = float("-1.03902") @@ -5102,6 +5566,7 @@ class Program_weight_tensor_parameter_463: class Program_weight_tensor_parameter_464: name = "parameter_464" + original_name = "batch_norm2d_85.w_0" shape = [384] dtype = "float32" min_val = float("0.491913") @@ -5113,6 +5578,7 @@ class Program_weight_tensor_parameter_464: class Program_weight_tensor_parameter_465: name = "parameter_465" + original_name = "batch_norm2d_85.w_2" shape = [384] dtype = "float32" min_val = float("5.60519e-45") @@ -5124,6 +5590,7 @@ class Program_weight_tensor_parameter_465: class Program_weight_tensor_parameter_466: name = "parameter_466" + original_name = "batch_norm2d_85.w_1" shape = [384] dtype = "float32" min_val = float("-0.308923") @@ -5135,6 +5602,7 @@ class Program_weight_tensor_parameter_466: class Program_weight_tensor_parameter_467: name = "parameter_467" + original_name = "conv2d_85.w_0" shape = [384, 384, 1, 1] dtype = "float32" min_val = float("-0.115652") @@ -5146,6 +5614,7 @@ class Program_weight_tensor_parameter_467: class Program_weight_tensor_parameter_468: name = "parameter_468" + original_name = "batch_norm2d_84.b_0" shape = [384] dtype = "float32" min_val = float("-1.76324") @@ -5157,6 +5626,7 @@ class Program_weight_tensor_parameter_468: class Program_weight_tensor_parameter_469: name = "parameter_469" + original_name = "batch_norm2d_84.w_0" shape = [384] dtype = "float32" min_val = float("0.348053") @@ -5168,6 +5638,7 @@ class Program_weight_tensor_parameter_469: class Program_weight_tensor_parameter_470: name = "parameter_470" + original_name = "batch_norm2d_84.w_2" shape = [384] dtype = "float32" min_val = float("0.00560796") @@ -5179,6 +5650,7 @@ class Program_weight_tensor_parameter_470: class Program_weight_tensor_parameter_471: name = "parameter_471" + original_name = "batch_norm2d_84.w_1" shape = [384] dtype = "float32" min_val = float("-0.63911") @@ -5190,6 +5662,7 @@ class Program_weight_tensor_parameter_471: class Program_weight_tensor_parameter_472: name = "parameter_472" + original_name = "conv2d_84.w_0" shape = [384, 1, 5, 5] dtype = "float32" min_val = float("-0.175919") @@ -5201,6 +5674,7 @@ class Program_weight_tensor_parameter_472: class Program_weight_tensor_parameter_473: name = "parameter_473" + original_name = "batch_norm2d_83.b_0" shape = [384] dtype = "float32" min_val = float("-0.881921") @@ -5212,6 +5686,7 @@ class Program_weight_tensor_parameter_473: class Program_weight_tensor_parameter_474: name = "parameter_474" + original_name = "batch_norm2d_83.w_0" shape = [384] dtype = "float32" min_val = float("0.249422") @@ -5223,6 +5698,7 @@ class Program_weight_tensor_parameter_474: class Program_weight_tensor_parameter_475: name = "parameter_475" + original_name = "batch_norm2d_83.w_2" shape = [384] dtype = "float32" min_val = float("0.00467576") @@ -5234,6 +5710,7 @@ class Program_weight_tensor_parameter_475: class Program_weight_tensor_parameter_476: name = "parameter_476" + original_name = "batch_norm2d_83.w_1" shape = [384] dtype = "float32" min_val = float("-0.239437") @@ -5245,6 +5722,7 @@ class Program_weight_tensor_parameter_476: class Program_weight_tensor_parameter_477: name = "parameter_477" + original_name = "conv2d_83.w_0" shape = [384, 384, 1, 1] dtype = "float32" min_val = float("-0.157234") @@ -5256,6 +5734,7 @@ class Program_weight_tensor_parameter_477: class Program_weight_tensor_parameter_478: name = "parameter_478" + original_name = "batch_norm2d_82.b_0" shape = [384] dtype = "float32" min_val = float("-1.98106") @@ -5267,6 +5746,7 @@ class Program_weight_tensor_parameter_478: class Program_weight_tensor_parameter_479: name = "parameter_479" + original_name = "batch_norm2d_82.w_0" shape = [384] dtype = "float32" min_val = float("0.380012") @@ -5278,6 +5758,7 @@ class Program_weight_tensor_parameter_479: class Program_weight_tensor_parameter_480: name = "parameter_480" + original_name = "batch_norm2d_82.w_2" shape = [384] dtype = "float32" min_val = float("0.0030561") @@ -5289,6 +5770,7 @@ class Program_weight_tensor_parameter_480: class Program_weight_tensor_parameter_481: name = "parameter_481" + original_name = "batch_norm2d_82.w_1" shape = [384] dtype = "float32" min_val = float("-0.652693") @@ -5300,6 +5782,7 @@ class Program_weight_tensor_parameter_481: class Program_weight_tensor_parameter_482: name = "parameter_482" + original_name = "conv2d_82.w_0" shape = [384, 1, 5, 5] dtype = "float32" min_val = float("-0.170925") @@ -5311,6 +5794,7 @@ class Program_weight_tensor_parameter_482: class Program_weight_tensor_parameter_483: name = "parameter_483" + original_name = "batch_norm2d_81.b_0" shape = [384] dtype = "float32" min_val = float("-1.00525") @@ -5322,6 +5806,7 @@ class Program_weight_tensor_parameter_483: class Program_weight_tensor_parameter_484: name = "parameter_484" + original_name = "batch_norm2d_81.w_0" shape = [384] dtype = "float32" min_val = float("0.2207") @@ -5333,6 +5818,7 @@ class Program_weight_tensor_parameter_484: class Program_weight_tensor_parameter_485: name = "parameter_485" + original_name = "batch_norm2d_81.w_2" shape = [384] dtype = "float32" min_val = float("0.00534858") @@ -5344,6 +5830,7 @@ class Program_weight_tensor_parameter_485: class Program_weight_tensor_parameter_486: name = "parameter_486" + original_name = "batch_norm2d_81.w_1" shape = [384] dtype = "float32" min_val = float("-0.348074") @@ -5355,6 +5842,7 @@ class Program_weight_tensor_parameter_486: class Program_weight_tensor_parameter_487: name = "parameter_487" + original_name = "conv2d_81.w_0" shape = [384, 384, 1, 1] dtype = "float32" min_val = float("-0.178638") @@ -5366,6 +5854,7 @@ class Program_weight_tensor_parameter_487: class Program_weight_tensor_parameter_488: name = "parameter_488" + original_name = "batch_norm2d_80.b_0" shape = [384] dtype = "float32" min_val = float("-2.01382") @@ -5377,6 +5866,7 @@ class Program_weight_tensor_parameter_488: class Program_weight_tensor_parameter_489: name = "parameter_489" + original_name = "batch_norm2d_80.w_0" shape = [384] dtype = "float32" min_val = float("0.317967") @@ -5388,6 +5878,7 @@ class Program_weight_tensor_parameter_489: class Program_weight_tensor_parameter_490: name = "parameter_490" + original_name = "batch_norm2d_80.w_2" shape = [384] dtype = "float32" min_val = float("0.00852514") @@ -5399,6 +5890,7 @@ class Program_weight_tensor_parameter_490: class Program_weight_tensor_parameter_491: name = "parameter_491" + original_name = "batch_norm2d_80.w_1" shape = [384] dtype = "float32" min_val = float("-0.318876") @@ -5410,6 +5902,7 @@ class Program_weight_tensor_parameter_491: class Program_weight_tensor_parameter_492: name = "parameter_492" + original_name = "conv2d_80.w_0" shape = [384, 1, 5, 5] dtype = "float32" min_val = float("-0.188809") @@ -5421,6 +5914,7 @@ class Program_weight_tensor_parameter_492: class Program_weight_tensor_parameter_493: name = "parameter_493" + original_name = "batch_norm2d_79.b_0" shape = [384] dtype = "float32" min_val = float("-1.04312") @@ -5432,6 +5926,7 @@ class Program_weight_tensor_parameter_493: class Program_weight_tensor_parameter_494: name = "parameter_494" + original_name = "batch_norm2d_79.w_0" shape = [384] dtype = "float32" min_val = float("0.420293") @@ -5443,6 +5938,7 @@ class Program_weight_tensor_parameter_494: class Program_weight_tensor_parameter_495: name = "parameter_495" + original_name = "batch_norm2d_79.w_2" shape = [384] dtype = "float32" min_val = float("0.00911839") @@ -5454,6 +5950,7 @@ class Program_weight_tensor_parameter_495: class Program_weight_tensor_parameter_496: name = "parameter_496" + original_name = "batch_norm2d_79.w_1" shape = [384] dtype = "float32" min_val = float("-0.371294") @@ -5465,6 +5962,7 @@ class Program_weight_tensor_parameter_496: class Program_weight_tensor_parameter_497: name = "parameter_497" + original_name = "conv2d_79.w_0" shape = [384, 384, 1, 1] dtype = "float32" min_val = float("-0.151291") @@ -5476,6 +5974,7 @@ class Program_weight_tensor_parameter_497: class Program_weight_tensor_parameter_498: name = "parameter_498" + original_name = "batch_norm2d_78.b_0" shape = [384] dtype = "float32" min_val = float("-2.31122") @@ -5487,6 +5986,7 @@ class Program_weight_tensor_parameter_498: class Program_weight_tensor_parameter_499: name = "parameter_499" + original_name = "batch_norm2d_78.w_0" shape = [384] dtype = "float32" min_val = float("0.417603") @@ -5498,6 +5998,7 @@ class Program_weight_tensor_parameter_499: class Program_weight_tensor_parameter_500: name = "parameter_500" + original_name = "batch_norm2d_78.w_2" shape = [384] dtype = "float32" min_val = float("0.00809077") @@ -5509,6 +6010,7 @@ class Program_weight_tensor_parameter_500: class Program_weight_tensor_parameter_501: name = "parameter_501" + original_name = "batch_norm2d_78.w_1" shape = [384] dtype = "float32" min_val = float("-0.606186") @@ -5520,6 +6022,7 @@ class Program_weight_tensor_parameter_501: class Program_weight_tensor_parameter_502: name = "parameter_502" + original_name = "conv2d_78.w_0" shape = [384, 1, 5, 5] dtype = "float32" min_val = float("-0.220619") @@ -5531,6 +6034,7 @@ class Program_weight_tensor_parameter_502: class Program_weight_tensor_parameter_503: name = "parameter_503" + original_name = "batch_norm2d_77.b_0" shape = [384] dtype = "float32" min_val = float("-1.37562") @@ -5542,6 +6046,7 @@ class Program_weight_tensor_parameter_503: class Program_weight_tensor_parameter_504: name = "parameter_504" + original_name = "batch_norm2d_77.w_0" shape = [384] dtype = "float32" min_val = float("0.199467") @@ -5553,6 +6058,7 @@ class Program_weight_tensor_parameter_504: class Program_weight_tensor_parameter_505: name = "parameter_505" + original_name = "batch_norm2d_77.w_2" shape = [384] dtype = "float32" min_val = float("0.0108931") @@ -5564,6 +6070,7 @@ class Program_weight_tensor_parameter_505: class Program_weight_tensor_parameter_506: name = "parameter_506" + original_name = "batch_norm2d_77.w_1" shape = [384] dtype = "float32" min_val = float("-0.591521") @@ -5575,6 +6082,7 @@ class Program_weight_tensor_parameter_506: class Program_weight_tensor_parameter_507: name = "parameter_507" + original_name = "conv2d_77.w_0" shape = [384, 384, 1, 1] dtype = "float32" min_val = float("-0.224681") @@ -5586,6 +6094,7 @@ class Program_weight_tensor_parameter_507: class Program_weight_tensor_parameter_508: name = "parameter_508" + original_name = "batch_norm2d_76.b_0" shape = [384] dtype = "float32" min_val = float("-6.35966") @@ -5597,6 +6106,7 @@ class Program_weight_tensor_parameter_508: class Program_weight_tensor_parameter_509: name = "parameter_509" + original_name = "batch_norm2d_76.w_0" shape = [384] dtype = "float32" min_val = float("0.496937") @@ -5608,6 +6118,7 @@ class Program_weight_tensor_parameter_509: class Program_weight_tensor_parameter_510: name = "parameter_510" + original_name = "batch_norm2d_76.w_2" shape = [384] dtype = "float32" min_val = float("0.0110934") @@ -5619,6 +6130,7 @@ class Program_weight_tensor_parameter_510: class Program_weight_tensor_parameter_511: name = "parameter_511" + original_name = "batch_norm2d_76.w_1" shape = [384] dtype = "float32" min_val = float("-2.24162") @@ -5630,6 +6142,7 @@ class Program_weight_tensor_parameter_511: class Program_weight_tensor_parameter_512: name = "parameter_512" + original_name = "conv2d_76.w_0" shape = [384, 1, 5, 5] dtype = "float32" min_val = float("-0.388421") @@ -5641,6 +6154,7 @@ class Program_weight_tensor_parameter_512: class Program_weight_tensor_parameter_513: name = "parameter_513" + original_name = "batch_norm2d_75.b_0" shape = [384] dtype = "float32" min_val = float("-2.37145") @@ -5652,6 +6166,7 @@ class Program_weight_tensor_parameter_513: class Program_weight_tensor_parameter_514: name = "parameter_514" + original_name = "batch_norm2d_75.w_0" shape = [384] dtype = "float32" min_val = float("0.339681") @@ -5663,6 +6178,7 @@ class Program_weight_tensor_parameter_514: class Program_weight_tensor_parameter_515: name = "parameter_515" + original_name = "batch_norm2d_75.w_2" shape = [384] dtype = "float32" min_val = float("0.0363021") @@ -5674,6 +6190,7 @@ class Program_weight_tensor_parameter_515: class Program_weight_tensor_parameter_516: name = "parameter_516" + original_name = "batch_norm2d_75.w_1" shape = [384] dtype = "float32" min_val = float("-1.61959") @@ -5685,6 +6202,7 @@ class Program_weight_tensor_parameter_516: class Program_weight_tensor_parameter_517: name = "parameter_517" + original_name = "conv2d_75.w_0" shape = [384, 1024, 1, 1] dtype = "float32" min_val = float("-0.471833") @@ -5696,6 +6214,7 @@ class Program_weight_tensor_parameter_517: class Program_weight_tensor_parameter_518: name = "parameter_518" + original_name = "batch_norm2d_74.b_0" shape = [1024] dtype = "float32" min_val = float("-3.35267") @@ -5707,6 +6226,7 @@ class Program_weight_tensor_parameter_518: class Program_weight_tensor_parameter_519: name = "parameter_519" + original_name = "batch_norm2d_74.w_0" shape = [1024] dtype = "float32" min_val = float("0.0126463") @@ -5718,6 +6238,7 @@ class Program_weight_tensor_parameter_519: class Program_weight_tensor_parameter_520: name = "parameter_520" + original_name = "batch_norm2d_74.w_2" shape = [1024] dtype = "float32" min_val = float("5.60519e-45") @@ -5729,6 +6250,7 @@ class Program_weight_tensor_parameter_520: class Program_weight_tensor_parameter_521: name = "parameter_521" + original_name = "batch_norm2d_74.w_1" shape = [1024] dtype = "float32" min_val = float("-0.389728") @@ -5740,6 +6262,7 @@ class Program_weight_tensor_parameter_521: class Program_weight_tensor_parameter_522: name = "parameter_522" + original_name = "conv2d_74.w_0" shape = [1024, 512, 1, 1] dtype = "float32" min_val = float("-0.231918") @@ -5751,6 +6274,7 @@ class Program_weight_tensor_parameter_522: class Program_weight_tensor_parameter_523: name = "parameter_523" + original_name = "batch_norm2d_73.b_0" shape = [512] dtype = "float32" min_val = float("-9.15063") @@ -5762,6 +6286,7 @@ class Program_weight_tensor_parameter_523: class Program_weight_tensor_parameter_524: name = "parameter_524" + original_name = "batch_norm2d_73.w_0" shape = [512] dtype = "float32" min_val = float("0.028597") @@ -5773,6 +6298,7 @@ class Program_weight_tensor_parameter_524: class Program_weight_tensor_parameter_525: name = "parameter_525" + original_name = "batch_norm2d_73.w_2" shape = [512] dtype = "float32" min_val = float("5.60519e-45") @@ -5784,6 +6310,7 @@ class Program_weight_tensor_parameter_525: class Program_weight_tensor_parameter_526: name = "parameter_526" + original_name = "batch_norm2d_73.w_1" shape = [512] dtype = "float32" min_val = float("-1.25272") @@ -5795,6 +6322,7 @@ class Program_weight_tensor_parameter_526: class Program_weight_tensor_parameter_527: name = "parameter_527" + original_name = "conv2d_73.w_0" shape = [512, 3328, 1, 1] dtype = "float32" min_val = float("-0.231655") @@ -5806,6 +6334,7 @@ class Program_weight_tensor_parameter_527: class Program_weight_tensor_parameter_528: name = "parameter_528" + original_name = "batch_norm2d_72.b_0" shape = [384] dtype = "float32" min_val = float("-1.98589") @@ -5817,6 +6346,7 @@ class Program_weight_tensor_parameter_528: class Program_weight_tensor_parameter_529: name = "parameter_529" + original_name = "batch_norm2d_72.w_0" shape = [384] dtype = "float32" min_val = float("0.21834") @@ -5828,6 +6358,7 @@ class Program_weight_tensor_parameter_529: class Program_weight_tensor_parameter_530: name = "parameter_530" + original_name = "batch_norm2d_72.w_2" shape = [384] dtype = "float32" min_val = float("1.81452e-37") @@ -5839,6 +6370,7 @@ class Program_weight_tensor_parameter_530: class Program_weight_tensor_parameter_531: name = "parameter_531" + original_name = "batch_norm2d_72.w_1" shape = [384] dtype = "float32" min_val = float("-0.358196") @@ -5850,6 +6382,7 @@ class Program_weight_tensor_parameter_531: class Program_weight_tensor_parameter_532: name = "parameter_532" + original_name = "conv2d_72.w_0" shape = [384, 1, 5, 5] dtype = "float32" min_val = float("-0.133851") @@ -5861,6 +6394,7 @@ class Program_weight_tensor_parameter_532: class Program_weight_tensor_parameter_533: name = "parameter_533" + original_name = "batch_norm2d_71.b_0" shape = [384] dtype = "float32" min_val = float("-1.03175") @@ -5872,6 +6406,7 @@ class Program_weight_tensor_parameter_533: class Program_weight_tensor_parameter_534: name = "parameter_534" + original_name = "batch_norm2d_71.w_0" shape = [384] dtype = "float32" min_val = float("0.473426") @@ -5883,6 +6418,7 @@ class Program_weight_tensor_parameter_534: class Program_weight_tensor_parameter_535: name = "parameter_535" + original_name = "batch_norm2d_71.w_2" shape = [384] dtype = "float32" min_val = float("5.60519e-45") @@ -5894,6 +6430,7 @@ class Program_weight_tensor_parameter_535: class Program_weight_tensor_parameter_536: name = "parameter_536" + original_name = "batch_norm2d_71.w_1" shape = [384] dtype = "float32" min_val = float("-0.200421") @@ -5905,6 +6442,7 @@ class Program_weight_tensor_parameter_536: class Program_weight_tensor_parameter_537: name = "parameter_537" + original_name = "conv2d_71.w_0" shape = [384, 384, 1, 1] dtype = "float32" min_val = float("-0.216636") @@ -5916,6 +6454,7 @@ class Program_weight_tensor_parameter_537: class Program_weight_tensor_parameter_538: name = "parameter_538" + original_name = "batch_norm2d_70.b_0" shape = [384] dtype = "float32" min_val = float("-1.56062") @@ -5927,6 +6466,7 @@ class Program_weight_tensor_parameter_538: class Program_weight_tensor_parameter_539: name = "parameter_539" + original_name = "batch_norm2d_70.w_0" shape = [384] dtype = "float32" min_val = float("0.196924") @@ -5938,6 +6478,7 @@ class Program_weight_tensor_parameter_539: class Program_weight_tensor_parameter_540: name = "parameter_540" + original_name = "batch_norm2d_70.w_2" shape = [384] dtype = "float32" min_val = float("1.58481e-15") @@ -5949,6 +6490,7 @@ class Program_weight_tensor_parameter_540: class Program_weight_tensor_parameter_541: name = "parameter_541" + original_name = "batch_norm2d_70.w_1" shape = [384] dtype = "float32" min_val = float("-0.363056") @@ -5960,6 +6502,7 @@ class Program_weight_tensor_parameter_541: class Program_weight_tensor_parameter_542: name = "parameter_542" + original_name = "conv2d_70.w_0" shape = [384, 1, 5, 5] dtype = "float32" min_val = float("-0.169159") @@ -5971,6 +6514,7 @@ class Program_weight_tensor_parameter_542: class Program_weight_tensor_parameter_543: name = "parameter_543" + original_name = "batch_norm2d_69.b_0" shape = [384] dtype = "float32" min_val = float("-0.953001") @@ -5982,6 +6526,7 @@ class Program_weight_tensor_parameter_543: class Program_weight_tensor_parameter_544: name = "parameter_544" + original_name = "batch_norm2d_69.w_0" shape = [384] dtype = "float32" min_val = float("0.471718") @@ -5993,6 +6538,7 @@ class Program_weight_tensor_parameter_544: class Program_weight_tensor_parameter_545: name = "parameter_545" + original_name = "batch_norm2d_69.w_2" shape = [384] dtype = "float32" min_val = float("9.92336e-13") @@ -6004,6 +6550,7 @@ class Program_weight_tensor_parameter_545: class Program_weight_tensor_parameter_546: name = "parameter_546" + original_name = "batch_norm2d_69.w_1" shape = [384] dtype = "float32" min_val = float("-0.142154") @@ -6015,6 +6562,7 @@ class Program_weight_tensor_parameter_546: class Program_weight_tensor_parameter_547: name = "parameter_547" + original_name = "conv2d_69.w_0" shape = [384, 384, 1, 1] dtype = "float32" min_val = float("-0.157455") @@ -6026,6 +6574,7 @@ class Program_weight_tensor_parameter_547: class Program_weight_tensor_parameter_548: name = "parameter_548" + original_name = "batch_norm2d_68.b_0" shape = [384] dtype = "float32" min_val = float("-2.63571") @@ -6037,6 +6586,7 @@ class Program_weight_tensor_parameter_548: class Program_weight_tensor_parameter_549: name = "parameter_549" + original_name = "batch_norm2d_68.w_0" shape = [384] dtype = "float32" min_val = float("0.378889") @@ -6048,6 +6598,7 @@ class Program_weight_tensor_parameter_549: class Program_weight_tensor_parameter_550: name = "parameter_550" + original_name = "batch_norm2d_68.w_2" shape = [384] dtype = "float32" min_val = float("0.00343905") @@ -6059,6 +6610,7 @@ class Program_weight_tensor_parameter_550: class Program_weight_tensor_parameter_551: name = "parameter_551" + original_name = "batch_norm2d_68.w_1" shape = [384] dtype = "float32" min_val = float("-0.394638") @@ -6070,6 +6622,7 @@ class Program_weight_tensor_parameter_551: class Program_weight_tensor_parameter_552: name = "parameter_552" + original_name = "conv2d_68.w_0" shape = [384, 1, 5, 5] dtype = "float32" min_val = float("-0.156648") @@ -6081,6 +6634,7 @@ class Program_weight_tensor_parameter_552: class Program_weight_tensor_parameter_553: name = "parameter_553" + original_name = "batch_norm2d_67.b_0" shape = [384] dtype = "float32" min_val = float("-0.970117") @@ -6092,6 +6646,7 @@ class Program_weight_tensor_parameter_553: class Program_weight_tensor_parameter_554: name = "parameter_554" + original_name = "batch_norm2d_67.w_0" shape = [384] dtype = "float32" min_val = float("0.2957") @@ -6103,6 +6658,7 @@ class Program_weight_tensor_parameter_554: class Program_weight_tensor_parameter_555: name = "parameter_555" + original_name = "batch_norm2d_67.w_2" shape = [384] dtype = "float32" min_val = float("0.00280248") @@ -6114,6 +6670,7 @@ class Program_weight_tensor_parameter_555: class Program_weight_tensor_parameter_556: name = "parameter_556" + original_name = "batch_norm2d_67.w_1" shape = [384] dtype = "float32" min_val = float("-0.345818") @@ -6125,6 +6682,7 @@ class Program_weight_tensor_parameter_556: class Program_weight_tensor_parameter_557: name = "parameter_557" + original_name = "conv2d_67.w_0" shape = [384, 384, 1, 1] dtype = "float32" min_val = float("-0.3764") @@ -6136,6 +6694,7 @@ class Program_weight_tensor_parameter_557: class Program_weight_tensor_parameter_558: name = "parameter_558" + original_name = "batch_norm2d_66.b_0" shape = [384] dtype = "float32" min_val = float("-6.62258") @@ -6147,6 +6706,7 @@ class Program_weight_tensor_parameter_558: class Program_weight_tensor_parameter_559: name = "parameter_559" + original_name = "batch_norm2d_66.w_0" shape = [384] dtype = "float32" min_val = float("0.34901") @@ -6158,6 +6718,7 @@ class Program_weight_tensor_parameter_559: class Program_weight_tensor_parameter_560: name = "parameter_560" + original_name = "batch_norm2d_66.w_2" shape = [384] dtype = "float32" min_val = float("5.60519e-45") @@ -6169,6 +6730,7 @@ class Program_weight_tensor_parameter_560: class Program_weight_tensor_parameter_561: name = "parameter_561" + original_name = "batch_norm2d_66.w_1" shape = [384] dtype = "float32" min_val = float("-0.719388") @@ -6180,6 +6742,7 @@ class Program_weight_tensor_parameter_561: class Program_weight_tensor_parameter_562: name = "parameter_562" + original_name = "conv2d_66.w_0" shape = [384, 1, 5, 5] dtype = "float32" min_val = float("-0.177844") @@ -6191,6 +6754,7 @@ class Program_weight_tensor_parameter_562: class Program_weight_tensor_parameter_563: name = "parameter_563" + original_name = "batch_norm2d_65.b_0" shape = [384] dtype = "float32" min_val = float("-1.3928") @@ -6202,6 +6766,7 @@ class Program_weight_tensor_parameter_563: class Program_weight_tensor_parameter_564: name = "parameter_564" + original_name = "batch_norm2d_65.w_0" shape = [384] dtype = "float32" min_val = float("0.346649") @@ -6213,6 +6778,7 @@ class Program_weight_tensor_parameter_564: class Program_weight_tensor_parameter_565: name = "parameter_565" + original_name = "batch_norm2d_65.w_2" shape = [384] dtype = "float32" min_val = float("5.60519e-45") @@ -6224,6 +6790,7 @@ class Program_weight_tensor_parameter_565: class Program_weight_tensor_parameter_566: name = "parameter_566" + original_name = "batch_norm2d_65.w_1" shape = [384] dtype = "float32" min_val = float("-0.576899") @@ -6235,6 +6802,7 @@ class Program_weight_tensor_parameter_566: class Program_weight_tensor_parameter_567: name = "parameter_567" + original_name = "conv2d_65.w_0" shape = [384, 384, 1, 1] dtype = "float32" min_val = float("-0.191076") @@ -6246,6 +6814,7 @@ class Program_weight_tensor_parameter_567: class Program_weight_tensor_parameter_568: name = "parameter_568" + original_name = "batch_norm2d_64.b_0" shape = [384] dtype = "float32" min_val = float("-2.13564") @@ -6257,6 +6826,7 @@ class Program_weight_tensor_parameter_568: class Program_weight_tensor_parameter_569: name = "parameter_569" + original_name = "batch_norm2d_64.w_0" shape = [384] dtype = "float32" min_val = float("0.435743") @@ -6268,6 +6838,7 @@ class Program_weight_tensor_parameter_569: class Program_weight_tensor_parameter_570: name = "parameter_570" + original_name = "batch_norm2d_64.w_2" shape = [384] dtype = "float32" min_val = float("0.00666269") @@ -6279,6 +6850,7 @@ class Program_weight_tensor_parameter_570: class Program_weight_tensor_parameter_571: name = "parameter_571" + original_name = "batch_norm2d_64.w_1" shape = [384] dtype = "float32" min_val = float("-0.622384") @@ -6290,6 +6862,7 @@ class Program_weight_tensor_parameter_571: class Program_weight_tensor_parameter_572: name = "parameter_572" + original_name = "conv2d_64.w_0" shape = [384, 1, 5, 5] dtype = "float32" min_val = float("-0.340903") @@ -6301,6 +6874,7 @@ class Program_weight_tensor_parameter_572: class Program_weight_tensor_parameter_573: name = "parameter_573" + original_name = "batch_norm2d_63.b_0" shape = [384] dtype = "float32" min_val = float("-1.14627") @@ -6312,6 +6886,7 @@ class Program_weight_tensor_parameter_573: class Program_weight_tensor_parameter_574: name = "parameter_574" + original_name = "batch_norm2d_63.w_0" shape = [384] dtype = "float32" min_val = float("0.232981") @@ -6323,6 +6898,7 @@ class Program_weight_tensor_parameter_574: class Program_weight_tensor_parameter_575: name = "parameter_575" + original_name = "batch_norm2d_63.w_2" shape = [384] dtype = "float32" min_val = float("0.0178398") @@ -6334,6 +6910,7 @@ class Program_weight_tensor_parameter_575: class Program_weight_tensor_parameter_576: name = "parameter_576" + original_name = "batch_norm2d_63.w_1" shape = [384] dtype = "float32" min_val = float("-0.761854") @@ -6345,6 +6922,7 @@ class Program_weight_tensor_parameter_576: class Program_weight_tensor_parameter_577: name = "parameter_577" + original_name = "conv2d_63.w_0" shape = [384, 384, 1, 1] dtype = "float32" min_val = float("-0.216271") @@ -6356,6 +6934,7 @@ class Program_weight_tensor_parameter_577: class Program_weight_tensor_parameter_578: name = "parameter_578" + original_name = "batch_norm2d_62.b_0" shape = [384] dtype = "float32" min_val = float("-3.82078") @@ -6367,6 +6946,7 @@ class Program_weight_tensor_parameter_578: class Program_weight_tensor_parameter_579: name = "parameter_579" + original_name = "batch_norm2d_62.w_0" shape = [384] dtype = "float32" min_val = float("0.126715") @@ -6378,6 +6958,7 @@ class Program_weight_tensor_parameter_579: class Program_weight_tensor_parameter_580: name = "parameter_580" + original_name = "batch_norm2d_62.w_2" shape = [384] dtype = "float32" min_val = float("5.60519e-45") @@ -6389,6 +6970,7 @@ class Program_weight_tensor_parameter_580: class Program_weight_tensor_parameter_581: name = "parameter_581" + original_name = "batch_norm2d_62.w_1" shape = [384] dtype = "float32" min_val = float("-1.57553") @@ -6400,6 +6982,7 @@ class Program_weight_tensor_parameter_581: class Program_weight_tensor_parameter_582: name = "parameter_582" + original_name = "conv2d_62.w_0" shape = [384, 1, 5, 5] dtype = "float32" min_val = float("-0.315842") @@ -6411,6 +6994,7 @@ class Program_weight_tensor_parameter_582: class Program_weight_tensor_parameter_583: name = "parameter_583" + original_name = "batch_norm2d_61.b_0" shape = [384] dtype = "float32" min_val = float("-1.12464") @@ -6422,6 +7006,7 @@ class Program_weight_tensor_parameter_583: class Program_weight_tensor_parameter_584: name = "parameter_584" + original_name = "batch_norm2d_61.w_0" shape = [384] dtype = "float32" min_val = float("0.461449") @@ -6433,6 +7018,7 @@ class Program_weight_tensor_parameter_584: class Program_weight_tensor_parameter_585: name = "parameter_585" + original_name = "batch_norm2d_61.w_2" shape = [384] dtype = "float32" min_val = float("5.60519e-45") @@ -6444,6 +7030,7 @@ class Program_weight_tensor_parameter_585: class Program_weight_tensor_parameter_586: name = "parameter_586" + original_name = "batch_norm2d_61.w_1" shape = [384] dtype = "float32" min_val = float("-0.782302") @@ -6455,6 +7042,7 @@ class Program_weight_tensor_parameter_586: class Program_weight_tensor_parameter_587: name = "parameter_587" + original_name = "conv2d_61.w_0" shape = [384, 1024, 1, 1] dtype = "float32" min_val = float("-0.224179") @@ -6466,6 +7054,7 @@ class Program_weight_tensor_parameter_587: class Program_weight_tensor_parameter_588: name = "parameter_588" + original_name = "batch_norm2d_60.b_0" shape = [1024] dtype = "float32" min_val = float("-3.93203") @@ -6477,6 +7066,7 @@ class Program_weight_tensor_parameter_588: class Program_weight_tensor_parameter_589: name = "parameter_589" + original_name = "batch_norm2d_60.w_0" shape = [1024] dtype = "float32" min_val = float("0.156353") @@ -6488,6 +7078,7 @@ class Program_weight_tensor_parameter_589: class Program_weight_tensor_parameter_590: name = "parameter_590" + original_name = "batch_norm2d_60.w_2" shape = [1024] dtype = "float32" min_val = float("5.60519e-45") @@ -6499,6 +7090,7 @@ class Program_weight_tensor_parameter_590: class Program_weight_tensor_parameter_591: name = "parameter_591" + original_name = "batch_norm2d_60.w_1" shape = [1024] dtype = "float32" min_val = float("-1.86108") @@ -6510,6 +7102,7 @@ class Program_weight_tensor_parameter_591: class Program_weight_tensor_parameter_592: name = "parameter_592" + original_name = "conv2d_60.w_0" shape = [1024, 512, 1, 1] dtype = "float32" min_val = float("-0.363978") @@ -6521,6 +7114,7 @@ class Program_weight_tensor_parameter_592: class Program_weight_tensor_parameter_593: name = "parameter_593" + original_name = "batch_norm2d_59.b_0" shape = [512] dtype = "float32" min_val = float("-5.97118") @@ -6532,6 +7126,7 @@ class Program_weight_tensor_parameter_593: class Program_weight_tensor_parameter_594: name = "parameter_594" + original_name = "batch_norm2d_59.w_0" shape = [512] dtype = "float32" min_val = float("0.0980281") @@ -6543,6 +7138,7 @@ class Program_weight_tensor_parameter_594: class Program_weight_tensor_parameter_595: name = "parameter_595" + original_name = "batch_norm2d_59.w_2" shape = [512] dtype = "float32" min_val = float("5.60519e-45") @@ -6554,6 +7150,7 @@ class Program_weight_tensor_parameter_595: class Program_weight_tensor_parameter_596: name = "parameter_596" + original_name = "batch_norm2d_59.w_1" shape = [512] dtype = "float32" min_val = float("-2.02944") @@ -6565,6 +7162,7 @@ class Program_weight_tensor_parameter_596: class Program_weight_tensor_parameter_597: name = "parameter_597" + original_name = "conv2d_59.w_0" shape = [512, 2816, 1, 1] dtype = "float32" min_val = float("-0.445954") @@ -6576,6 +7174,7 @@ class Program_weight_tensor_parameter_597: class Program_weight_tensor_parameter_598: name = "parameter_598" + original_name = "batch_norm2d_58.b_0" shape = [384] dtype = "float32" min_val = float("-5.0879") @@ -6587,6 +7186,7 @@ class Program_weight_tensor_parameter_598: class Program_weight_tensor_parameter_599: name = "parameter_599" + original_name = "batch_norm2d_58.w_0" shape = [384] dtype = "float32" min_val = float("0.0632834") @@ -6598,6 +7198,7 @@ class Program_weight_tensor_parameter_599: class Program_weight_tensor_parameter_600: name = "parameter_600" + original_name = "batch_norm2d_58.w_2" shape = [384] dtype = "float32" min_val = float("5.60519e-45") @@ -6609,6 +7210,7 @@ class Program_weight_tensor_parameter_600: class Program_weight_tensor_parameter_601: name = "parameter_601" + original_name = "batch_norm2d_58.w_1" shape = [384] dtype = "float32" min_val = float("-0.694305") @@ -6620,6 +7222,7 @@ class Program_weight_tensor_parameter_601: class Program_weight_tensor_parameter_602: name = "parameter_602" + original_name = "conv2d_58.w_0" shape = [384, 1, 5, 5] dtype = "float32" min_val = float("-0.153213") @@ -6631,6 +7234,7 @@ class Program_weight_tensor_parameter_602: class Program_weight_tensor_parameter_603: name = "parameter_603" + original_name = "batch_norm2d_57.b_0" shape = [384] dtype = "float32" min_val = float("-1.47387") @@ -6642,6 +7246,7 @@ class Program_weight_tensor_parameter_603: class Program_weight_tensor_parameter_604: name = "parameter_604" + original_name = "batch_norm2d_57.w_0" shape = [384] dtype = "float32" min_val = float("0.252332") @@ -6653,6 +7258,7 @@ class Program_weight_tensor_parameter_604: class Program_weight_tensor_parameter_605: name = "parameter_605" + original_name = "batch_norm2d_57.w_2" shape = [384] dtype = "float32" min_val = float("5.60519e-45") @@ -6664,6 +7270,7 @@ class Program_weight_tensor_parameter_605: class Program_weight_tensor_parameter_606: name = "parameter_606" + original_name = "batch_norm2d_57.w_1" shape = [384] dtype = "float32" min_val = float("-0.439715") @@ -6675,6 +7282,7 @@ class Program_weight_tensor_parameter_606: class Program_weight_tensor_parameter_607: name = "parameter_607" + original_name = "conv2d_57.w_0" shape = [384, 384, 1, 1] dtype = "float32" min_val = float("-0.283704") @@ -6686,6 +7294,7 @@ class Program_weight_tensor_parameter_607: class Program_weight_tensor_parameter_608: name = "parameter_608" + original_name = "batch_norm2d_56.b_0" shape = [384] dtype = "float32" min_val = float("-9.62862") @@ -6697,6 +7306,7 @@ class Program_weight_tensor_parameter_608: class Program_weight_tensor_parameter_609: name = "parameter_609" + original_name = "batch_norm2d_56.w_0" shape = [384] dtype = "float32" min_val = float("0.0292768") @@ -6708,6 +7318,7 @@ class Program_weight_tensor_parameter_609: class Program_weight_tensor_parameter_610: name = "parameter_610" + original_name = "batch_norm2d_56.w_2" shape = [384] dtype = "float32" min_val = float("5.60519e-45") @@ -6719,6 +7330,7 @@ class Program_weight_tensor_parameter_610: class Program_weight_tensor_parameter_611: name = "parameter_611" + original_name = "batch_norm2d_56.w_1" shape = [384] dtype = "float32" min_val = float("-1.64476") @@ -6730,6 +7342,7 @@ class Program_weight_tensor_parameter_611: class Program_weight_tensor_parameter_612: name = "parameter_612" + original_name = "conv2d_56.w_0" shape = [384, 1, 5, 5] dtype = "float32" min_val = float("-0.211994") @@ -6741,6 +7354,7 @@ class Program_weight_tensor_parameter_612: class Program_weight_tensor_parameter_613: name = "parameter_613" + original_name = "batch_norm2d_55.b_0" shape = [384] dtype = "float32" min_val = float("-2.62235") @@ -6752,6 +7366,7 @@ class Program_weight_tensor_parameter_613: class Program_weight_tensor_parameter_614: name = "parameter_614" + original_name = "batch_norm2d_55.w_0" shape = [384] dtype = "float32" min_val = float("0.139145") @@ -6763,6 +7378,7 @@ class Program_weight_tensor_parameter_614: class Program_weight_tensor_parameter_615: name = "parameter_615" + original_name = "batch_norm2d_55.w_2" shape = [384] dtype = "float32" min_val = float("5.60519e-45") @@ -6774,6 +7390,7 @@ class Program_weight_tensor_parameter_615: class Program_weight_tensor_parameter_616: name = "parameter_616" + original_name = "batch_norm2d_55.w_1" shape = [384] dtype = "float32" min_val = float("-0.399427") @@ -6785,6 +7402,7 @@ class Program_weight_tensor_parameter_616: class Program_weight_tensor_parameter_617: name = "parameter_617" + original_name = "conv2d_55.w_0" shape = [384, 384, 1, 1] dtype = "float32" min_val = float("-0.333681") @@ -6796,6 +7414,7 @@ class Program_weight_tensor_parameter_617: class Program_weight_tensor_parameter_618: name = "parameter_618" + original_name = "batch_norm2d_54.b_0" shape = [384] dtype = "float32" min_val = float("-6.1496") @@ -6807,6 +7426,7 @@ class Program_weight_tensor_parameter_618: class Program_weight_tensor_parameter_619: name = "parameter_619" + original_name = "batch_norm2d_54.w_0" shape = [384] dtype = "float32" min_val = float("-0.176633") @@ -6818,6 +7438,7 @@ class Program_weight_tensor_parameter_619: class Program_weight_tensor_parameter_620: name = "parameter_620" + original_name = "batch_norm2d_54.w_2" shape = [384] dtype = "float32" min_val = float("5.60519e-45") @@ -6829,6 +7450,7 @@ class Program_weight_tensor_parameter_620: class Program_weight_tensor_parameter_621: name = "parameter_621" + original_name = "batch_norm2d_54.w_1" shape = [384] dtype = "float32" min_val = float("-0.660444") @@ -6840,6 +7462,7 @@ class Program_weight_tensor_parameter_621: class Program_weight_tensor_parameter_622: name = "parameter_622" + original_name = "conv2d_54.w_0" shape = [384, 1, 5, 5] dtype = "float32" min_val = float("-0.18887") @@ -6851,6 +7474,7 @@ class Program_weight_tensor_parameter_622: class Program_weight_tensor_parameter_623: name = "parameter_623" + original_name = "batch_norm2d_53.b_0" shape = [384] dtype = "float32" min_val = float("-1.09631") @@ -6862,6 +7486,7 @@ class Program_weight_tensor_parameter_623: class Program_weight_tensor_parameter_624: name = "parameter_624" + original_name = "batch_norm2d_53.w_0" shape = [384] dtype = "float32" min_val = float("0.285513") @@ -6873,6 +7498,7 @@ class Program_weight_tensor_parameter_624: class Program_weight_tensor_parameter_625: name = "parameter_625" + original_name = "batch_norm2d_53.w_2" shape = [384] dtype = "float32" min_val = float("5.60519e-45") @@ -6884,6 +7510,7 @@ class Program_weight_tensor_parameter_625: class Program_weight_tensor_parameter_626: name = "parameter_626" + original_name = "batch_norm2d_53.w_1" shape = [384] dtype = "float32" min_val = float("-0.539631") @@ -6895,6 +7522,7 @@ class Program_weight_tensor_parameter_626: class Program_weight_tensor_parameter_627: name = "parameter_627" + original_name = "conv2d_53.w_0" shape = [384, 384, 1, 1] dtype = "float32" min_val = float("-0.294913") @@ -6906,6 +7534,7 @@ class Program_weight_tensor_parameter_627: class Program_weight_tensor_parameter_628: name = "parameter_628" + original_name = "batch_norm2d_52.b_0" shape = [384] dtype = "float32" min_val = float("-3.10695") @@ -6917,6 +7546,7 @@ class Program_weight_tensor_parameter_628: class Program_weight_tensor_parameter_629: name = "parameter_629" + original_name = "batch_norm2d_52.w_0" shape = [384] dtype = "float32" min_val = float("-0.0097252") @@ -6928,6 +7558,7 @@ class Program_weight_tensor_parameter_629: class Program_weight_tensor_parameter_630: name = "parameter_630" + original_name = "batch_norm2d_52.w_2" shape = [384] dtype = "float32" min_val = float("5.60519e-45") @@ -6939,6 +7570,7 @@ class Program_weight_tensor_parameter_630: class Program_weight_tensor_parameter_631: name = "parameter_631" + original_name = "batch_norm2d_52.w_1" shape = [384] dtype = "float32" min_val = float("-1.00865") @@ -6950,6 +7582,7 @@ class Program_weight_tensor_parameter_631: class Program_weight_tensor_parameter_632: name = "parameter_632" + original_name = "conv2d_52.w_0" shape = [384, 1, 5, 5] dtype = "float32" min_val = float("-0.209722") @@ -6961,6 +7594,7 @@ class Program_weight_tensor_parameter_632: class Program_weight_tensor_parameter_633: name = "parameter_633" + original_name = "batch_norm2d_51.b_0" shape = [384] dtype = "float32" min_val = float("-1.70204") @@ -6972,6 +7606,7 @@ class Program_weight_tensor_parameter_633: class Program_weight_tensor_parameter_634: name = "parameter_634" + original_name = "batch_norm2d_51.w_0" shape = [384] dtype = "float32" min_val = float("0.319168") @@ -6983,6 +7618,7 @@ class Program_weight_tensor_parameter_634: class Program_weight_tensor_parameter_635: name = "parameter_635" + original_name = "batch_norm2d_51.w_2" shape = [384] dtype = "float32" min_val = float("5.60519e-45") @@ -6994,6 +7630,7 @@ class Program_weight_tensor_parameter_635: class Program_weight_tensor_parameter_636: name = "parameter_636" + original_name = "batch_norm2d_51.w_1" shape = [384] dtype = "float32" min_val = float("-0.635501") @@ -7005,6 +7642,7 @@ class Program_weight_tensor_parameter_636: class Program_weight_tensor_parameter_637: name = "parameter_637" + original_name = "conv2d_51.w_0" shape = [384, 384, 1, 1] dtype = "float32" min_val = float("-0.253966") @@ -7016,6 +7654,7 @@ class Program_weight_tensor_parameter_637: class Program_weight_tensor_parameter_638: name = "parameter_638" + original_name = "batch_norm2d_50.b_0" shape = [384] dtype = "float32" min_val = float("-4.30995") @@ -7027,6 +7666,7 @@ class Program_weight_tensor_parameter_638: class Program_weight_tensor_parameter_639: name = "parameter_639" + original_name = "batch_norm2d_50.w_0" shape = [384] dtype = "float32" min_val = float("0.104233") @@ -7038,6 +7678,7 @@ class Program_weight_tensor_parameter_639: class Program_weight_tensor_parameter_640: name = "parameter_640" + original_name = "batch_norm2d_50.w_2" shape = [384] dtype = "float32" min_val = float("5.60519e-45") @@ -7049,6 +7690,7 @@ class Program_weight_tensor_parameter_640: class Program_weight_tensor_parameter_641: name = "parameter_641" + original_name = "batch_norm2d_50.w_1" shape = [384] dtype = "float32" min_val = float("-0.79973") @@ -7060,6 +7702,7 @@ class Program_weight_tensor_parameter_641: class Program_weight_tensor_parameter_642: name = "parameter_642" + original_name = "conv2d_50.w_0" shape = [384, 1, 5, 5] dtype = "float32" min_val = float("-0.271974") @@ -7071,6 +7714,7 @@ class Program_weight_tensor_parameter_642: class Program_weight_tensor_parameter_643: name = "parameter_643" + original_name = "batch_norm2d_49.b_0" shape = [384] dtype = "float32" min_val = float("-1.70073") @@ -7082,6 +7726,7 @@ class Program_weight_tensor_parameter_643: class Program_weight_tensor_parameter_644: name = "parameter_644" + original_name = "batch_norm2d_49.w_0" shape = [384] dtype = "float32" min_val = float("0.573543") @@ -7093,6 +7738,7 @@ class Program_weight_tensor_parameter_644: class Program_weight_tensor_parameter_645: name = "parameter_645" + original_name = "batch_norm2d_49.w_2" shape = [384] dtype = "float32" min_val = float("5.60519e-45") @@ -7104,6 +7750,7 @@ class Program_weight_tensor_parameter_645: class Program_weight_tensor_parameter_646: name = "parameter_646" + original_name = "batch_norm2d_49.w_1" shape = [384] dtype = "float32" min_val = float("-1.06207") @@ -7115,6 +7762,7 @@ class Program_weight_tensor_parameter_646: class Program_weight_tensor_parameter_647: name = "parameter_647" + original_name = "conv2d_49.w_0" shape = [384, 384, 1, 1] dtype = "float32" min_val = float("-0.356749") @@ -7126,6 +7774,7 @@ class Program_weight_tensor_parameter_647: class Program_weight_tensor_parameter_648: name = "parameter_648" + original_name = "batch_norm2d_48.b_0" shape = [384] dtype = "float32" min_val = float("-5.57508") @@ -7137,6 +7786,7 @@ class Program_weight_tensor_parameter_648: class Program_weight_tensor_parameter_649: name = "parameter_649" + original_name = "batch_norm2d_48.w_0" shape = [384] dtype = "float32" min_val = float("-2.42817") @@ -7148,6 +7798,7 @@ class Program_weight_tensor_parameter_649: class Program_weight_tensor_parameter_650: name = "parameter_650" + original_name = "batch_norm2d_48.w_2" shape = [384] dtype = "float32" min_val = float("5.60519e-45") @@ -7159,6 +7810,7 @@ class Program_weight_tensor_parameter_650: class Program_weight_tensor_parameter_651: name = "parameter_651" + original_name = "batch_norm2d_48.w_1" shape = [384] dtype = "float32" min_val = float("-0.937796") @@ -7170,6 +7822,7 @@ class Program_weight_tensor_parameter_651: class Program_weight_tensor_parameter_652: name = "parameter_652" + original_name = "conv2d_48.w_0" shape = [384, 1, 5, 5] dtype = "float32" min_val = float("-0.418238") @@ -7181,6 +7834,7 @@ class Program_weight_tensor_parameter_652: class Program_weight_tensor_parameter_653: name = "parameter_653" + original_name = "batch_norm2d_47.b_0" shape = [384] dtype = "float32" min_val = float("-1.99761") @@ -7192,6 +7846,7 @@ class Program_weight_tensor_parameter_653: class Program_weight_tensor_parameter_654: name = "parameter_654" + original_name = "batch_norm2d_47.w_0" shape = [384] dtype = "float32" min_val = float("0.626215") @@ -7203,6 +7858,7 @@ class Program_weight_tensor_parameter_654: class Program_weight_tensor_parameter_655: name = "parameter_655" + original_name = "batch_norm2d_47.w_2" shape = [384] dtype = "float32" min_val = float("5.60519e-45") @@ -7214,6 +7870,7 @@ class Program_weight_tensor_parameter_655: class Program_weight_tensor_parameter_656: name = "parameter_656" + original_name = "batch_norm2d_47.w_1" shape = [384] dtype = "float32" min_val = float("-0.00476672") @@ -7225,6 +7882,7 @@ class Program_weight_tensor_parameter_656: class Program_weight_tensor_parameter_657: name = "parameter_657" + original_name = "conv2d_47.w_0" shape = [384, 512, 1, 1] dtype = "float32" min_val = float("-0.41565") @@ -7236,6 +7894,7 @@ class Program_weight_tensor_parameter_657: class Program_weight_tensor_parameter_658: name = "parameter_658" + original_name = "batch_norm2d_46.b_0" shape = [512] dtype = "float32" min_val = float("-0.00997735") @@ -7247,6 +7906,7 @@ class Program_weight_tensor_parameter_658: class Program_weight_tensor_parameter_659: name = "parameter_659" + original_name = "batch_norm2d_46.w_0" shape = [512] dtype = "float32" min_val = float("0.812687") @@ -7258,6 +7918,7 @@ class Program_weight_tensor_parameter_659: class Program_weight_tensor_parameter_660: name = "parameter_660" + original_name = "batch_norm2d_46.w_2" shape = [512] dtype = "float32" min_val = float("0.00222622") @@ -7269,6 +7930,7 @@ class Program_weight_tensor_parameter_660: class Program_weight_tensor_parameter_661: name = "parameter_661" + original_name = "batch_norm2d_46.w_1" shape = [512] dtype = "float32" min_val = float("-1.03105") @@ -7280,6 +7942,7 @@ class Program_weight_tensor_parameter_661: class Program_weight_tensor_parameter_662: name = "parameter_662" + original_name = "conv2d_46.w_0" shape = [512, 1, 3, 3] dtype = "float32" min_val = float("-0.212113") @@ -7291,6 +7954,7 @@ class Program_weight_tensor_parameter_662: class Program_weight_tensor_parameter_663: name = "parameter_663" + original_name = "batch_norm2d_45.b_0" shape = [512] dtype = "float32" min_val = float("-4.03815") @@ -7302,6 +7966,7 @@ class Program_weight_tensor_parameter_663: class Program_weight_tensor_parameter_664: name = "parameter_664" + original_name = "batch_norm2d_45.w_0" shape = [512] dtype = "float32" min_val = float("-0.629023") @@ -7313,6 +7978,7 @@ class Program_weight_tensor_parameter_664: class Program_weight_tensor_parameter_665: name = "parameter_665" + original_name = "batch_norm2d_45.w_2" shape = [512] dtype = "float32" min_val = float("5.60519e-45") @@ -7324,6 +7990,7 @@ class Program_weight_tensor_parameter_665: class Program_weight_tensor_parameter_666: name = "parameter_666" + original_name = "batch_norm2d_45.w_1" shape = [512] dtype = "float32" min_val = float("-0.819093") @@ -7335,6 +8002,7 @@ class Program_weight_tensor_parameter_666: class Program_weight_tensor_parameter_667: name = "parameter_667" + original_name = "conv2d_45.w_0" shape = [512, 256, 1, 1] dtype = "float32" min_val = float("-0.241372") @@ -7346,6 +8014,7 @@ class Program_weight_tensor_parameter_667: class Program_weight_tensor_parameter_668: name = "parameter_668" + original_name = "batch_norm2d_44.b_0" shape = [256] dtype = "float32" min_val = float("-4.48499") @@ -7357,6 +8026,7 @@ class Program_weight_tensor_parameter_668: class Program_weight_tensor_parameter_669: name = "parameter_669" + original_name = "batch_norm2d_44.w_0" shape = [256] dtype = "float32" min_val = float("-0.102443") @@ -7368,6 +8038,7 @@ class Program_weight_tensor_parameter_669: class Program_weight_tensor_parameter_670: name = "parameter_670" + original_name = "batch_norm2d_44.w_2" shape = [256] dtype = "float32" min_val = float("5.60519e-45") @@ -7379,6 +8050,7 @@ class Program_weight_tensor_parameter_670: class Program_weight_tensor_parameter_671: name = "parameter_671" + original_name = "batch_norm2d_44.w_1" shape = [256] dtype = "float32" min_val = float("-0.918398") @@ -7390,6 +8062,7 @@ class Program_weight_tensor_parameter_671: class Program_weight_tensor_parameter_672: name = "parameter_672" + original_name = "conv2d_44.w_0" shape = [256, 1664, 1, 1] dtype = "float32" min_val = float("-0.224272") @@ -7401,6 +8074,7 @@ class Program_weight_tensor_parameter_672: class Program_weight_tensor_parameter_673: name = "parameter_673" + original_name = "batch_norm2d_43.b_0" shape = [192] dtype = "float32" min_val = float("-1.12611") @@ -7412,6 +8086,7 @@ class Program_weight_tensor_parameter_673: class Program_weight_tensor_parameter_674: name = "parameter_674" + original_name = "batch_norm2d_43.w_0" shape = [192] dtype = "float32" min_val = float("0.337233") @@ -7423,6 +8098,7 @@ class Program_weight_tensor_parameter_674: class Program_weight_tensor_parameter_675: name = "parameter_675" + original_name = "batch_norm2d_43.w_2" shape = [192] dtype = "float32" min_val = float("5.60519e-45") @@ -7434,6 +8110,7 @@ class Program_weight_tensor_parameter_675: class Program_weight_tensor_parameter_676: name = "parameter_676" + original_name = "batch_norm2d_43.w_1" shape = [192] dtype = "float32" min_val = float("-0.404381") @@ -7445,6 +8122,7 @@ class Program_weight_tensor_parameter_676: class Program_weight_tensor_parameter_677: name = "parameter_677" + original_name = "conv2d_43.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.0655201") @@ -7456,6 +8134,7 @@ class Program_weight_tensor_parameter_677: class Program_weight_tensor_parameter_678: name = "parameter_678" + original_name = "batch_norm2d_42.b_0" shape = [192] dtype = "float32" min_val = float("-1.47831") @@ -7467,6 +8146,7 @@ class Program_weight_tensor_parameter_678: class Program_weight_tensor_parameter_679: name = "parameter_679" + original_name = "batch_norm2d_42.w_0" shape = [192] dtype = "float32" min_val = float("0.229018") @@ -7478,6 +8158,7 @@ class Program_weight_tensor_parameter_679: class Program_weight_tensor_parameter_680: name = "parameter_680" + original_name = "batch_norm2d_42.w_2" shape = [192] dtype = "float32" min_val = float("5.60519e-45") @@ -7489,6 +8170,7 @@ class Program_weight_tensor_parameter_680: class Program_weight_tensor_parameter_681: name = "parameter_681" + original_name = "batch_norm2d_42.w_1" shape = [192] dtype = "float32" min_val = float("-0.635557") @@ -7500,6 +8182,7 @@ class Program_weight_tensor_parameter_681: class Program_weight_tensor_parameter_682: name = "parameter_682" + original_name = "conv2d_42.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.077972") @@ -7511,6 +8194,7 @@ class Program_weight_tensor_parameter_682: class Program_weight_tensor_parameter_683: name = "parameter_683" + original_name = "batch_norm2d_41.b_0" shape = [192] dtype = "float32" min_val = float("-1.87841") @@ -7522,6 +8206,7 @@ class Program_weight_tensor_parameter_683: class Program_weight_tensor_parameter_684: name = "parameter_684" + original_name = "batch_norm2d_41.w_0" shape = [192] dtype = "float32" min_val = float("0.297936") @@ -7533,6 +8218,7 @@ class Program_weight_tensor_parameter_684: class Program_weight_tensor_parameter_685: name = "parameter_685" + original_name = "batch_norm2d_41.w_2" shape = [192] dtype = "float32" min_val = float("0.0143504") @@ -7544,6 +8230,7 @@ class Program_weight_tensor_parameter_685: class Program_weight_tensor_parameter_686: name = "parameter_686" + original_name = "batch_norm2d_41.w_1" shape = [192] dtype = "float32" min_val = float("-0.522048") @@ -7555,6 +8242,7 @@ class Program_weight_tensor_parameter_686: class Program_weight_tensor_parameter_687: name = "parameter_687" + original_name = "conv2d_41.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.124926") @@ -7566,6 +8254,7 @@ class Program_weight_tensor_parameter_687: class Program_weight_tensor_parameter_688: name = "parameter_688" + original_name = "batch_norm2d_40.b_0" shape = [192] dtype = "float32" min_val = float("-3.1668") @@ -7577,6 +8266,7 @@ class Program_weight_tensor_parameter_688: class Program_weight_tensor_parameter_689: name = "parameter_689" + original_name = "batch_norm2d_40.w_0" shape = [192] dtype = "float32" min_val = float("0.626449") @@ -7588,6 +8278,7 @@ class Program_weight_tensor_parameter_689: class Program_weight_tensor_parameter_690: name = "parameter_690" + original_name = "batch_norm2d_40.w_2" shape = [192] dtype = "float32" min_val = float("0.0678091") @@ -7599,6 +8290,7 @@ class Program_weight_tensor_parameter_690: class Program_weight_tensor_parameter_691: name = "parameter_691" + original_name = "batch_norm2d_40.w_1" shape = [192] dtype = "float32" min_val = float("-1.06661") @@ -7610,6 +8302,7 @@ class Program_weight_tensor_parameter_691: class Program_weight_tensor_parameter_692: name = "parameter_692" + original_name = "conv2d_40.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.0915296") @@ -7621,6 +8314,7 @@ class Program_weight_tensor_parameter_692: class Program_weight_tensor_parameter_693: name = "parameter_693" + original_name = "batch_norm2d_39.b_0" shape = [192] dtype = "float32" min_val = float("-3.576") @@ -7632,6 +8326,7 @@ class Program_weight_tensor_parameter_693: class Program_weight_tensor_parameter_694: name = "parameter_694" + original_name = "batch_norm2d_39.w_0" shape = [192] dtype = "float32" min_val = float("0.71965") @@ -7643,6 +8338,7 @@ class Program_weight_tensor_parameter_694: class Program_weight_tensor_parameter_695: name = "parameter_695" + original_name = "batch_norm2d_39.w_2" shape = [192] dtype = "float32" min_val = float("0.122374") @@ -7654,6 +8350,7 @@ class Program_weight_tensor_parameter_695: class Program_weight_tensor_parameter_696: name = "parameter_696" + original_name = "batch_norm2d_39.w_1" shape = [192] dtype = "float32" min_val = float("-2.06392") @@ -7665,6 +8362,7 @@ class Program_weight_tensor_parameter_696: class Program_weight_tensor_parameter_697: name = "parameter_697" + original_name = "conv2d_39.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.164833") @@ -7676,6 +8374,7 @@ class Program_weight_tensor_parameter_697: class Program_weight_tensor_parameter_698: name = "parameter_698" + original_name = "batch_norm2d_38.b_0" shape = [192] dtype = "float32" min_val = float("-3.78456") @@ -7687,6 +8386,7 @@ class Program_weight_tensor_parameter_698: class Program_weight_tensor_parameter_699: name = "parameter_699" + original_name = "batch_norm2d_38.w_0" shape = [192] dtype = "float32" min_val = float("0.792114") @@ -7698,6 +8398,7 @@ class Program_weight_tensor_parameter_699: class Program_weight_tensor_parameter_700: name = "parameter_700" + original_name = "batch_norm2d_38.w_2" shape = [192] dtype = "float32" min_val = float("0.297992") @@ -7709,6 +8410,7 @@ class Program_weight_tensor_parameter_700: class Program_weight_tensor_parameter_701: name = "parameter_701" + original_name = "batch_norm2d_38.w_1" shape = [192] dtype = "float32" min_val = float("-4.70438") @@ -7720,6 +8422,7 @@ class Program_weight_tensor_parameter_701: class Program_weight_tensor_parameter_702: name = "parameter_702" + original_name = "conv2d_38.w_0" shape = [192, 512, 3, 3] dtype = "float32" min_val = float("-0.151072") @@ -7731,6 +8434,7 @@ class Program_weight_tensor_parameter_702: class Program_weight_tensor_parameter_703: name = "parameter_703" + original_name = "batch_norm2d_37.b_0" shape = [512] dtype = "float32" min_val = float("-2.42872") @@ -7742,6 +8446,7 @@ class Program_weight_tensor_parameter_703: class Program_weight_tensor_parameter_704: name = "parameter_704" + original_name = "batch_norm2d_37.w_0" shape = [512] dtype = "float32" min_val = float("-0.137592") @@ -7753,6 +8458,7 @@ class Program_weight_tensor_parameter_704: class Program_weight_tensor_parameter_705: name = "parameter_705" + original_name = "batch_norm2d_37.w_2" shape = [512] dtype = "float32" min_val = float("5.60519e-45") @@ -7764,6 +8470,7 @@ class Program_weight_tensor_parameter_705: class Program_weight_tensor_parameter_706: name = "parameter_706" + original_name = "batch_norm2d_37.w_1" shape = [512] dtype = "float32" min_val = float("-1.13769") @@ -7775,6 +8482,7 @@ class Program_weight_tensor_parameter_706: class Program_weight_tensor_parameter_707: name = "parameter_707" + original_name = "conv2d_37.w_0" shape = [512, 256, 1, 1] dtype = "float32" min_val = float("-0.223396") @@ -7786,6 +8494,7 @@ class Program_weight_tensor_parameter_707: class Program_weight_tensor_parameter_708: name = "parameter_708" + original_name = "batch_norm2d_36.b_0" shape = [256] dtype = "float32" min_val = float("-3.60353") @@ -7797,6 +8506,7 @@ class Program_weight_tensor_parameter_708: class Program_weight_tensor_parameter_709: name = "parameter_709" + original_name = "batch_norm2d_36.w_0" shape = [256] dtype = "float32" min_val = float("-0.0420502") @@ -7808,6 +8518,7 @@ class Program_weight_tensor_parameter_709: class Program_weight_tensor_parameter_710: name = "parameter_710" + original_name = "batch_norm2d_36.w_2" shape = [256] dtype = "float32" min_val = float("5.60519e-45") @@ -7819,6 +8530,7 @@ class Program_weight_tensor_parameter_710: class Program_weight_tensor_parameter_711: name = "parameter_711" + original_name = "batch_norm2d_36.w_1" shape = [256] dtype = "float32" min_val = float("-1.06356") @@ -7830,6 +8542,7 @@ class Program_weight_tensor_parameter_711: class Program_weight_tensor_parameter_712: name = "parameter_712" + original_name = "conv2d_36.w_0" shape = [256, 1664, 1, 1] dtype = "float32" min_val = float("-0.141315") @@ -7841,6 +8554,7 @@ class Program_weight_tensor_parameter_712: class Program_weight_tensor_parameter_713: name = "parameter_713" + original_name = "batch_norm2d_35.b_0" shape = [192] dtype = "float32" min_val = float("-1.75499") @@ -7852,6 +8566,7 @@ class Program_weight_tensor_parameter_713: class Program_weight_tensor_parameter_714: name = "parameter_714" + original_name = "batch_norm2d_35.w_0" shape = [192] dtype = "float32" min_val = float("0.813071") @@ -7863,6 +8578,7 @@ class Program_weight_tensor_parameter_714: class Program_weight_tensor_parameter_715: name = "parameter_715" + original_name = "batch_norm2d_35.w_2" shape = [192] dtype = "float32" min_val = float("0.0497467") @@ -7874,6 +8590,7 @@ class Program_weight_tensor_parameter_715: class Program_weight_tensor_parameter_716: name = "parameter_716" + original_name = "batch_norm2d_35.w_1" shape = [192] dtype = "float32" min_val = float("-0.895396") @@ -7885,6 +8602,7 @@ class Program_weight_tensor_parameter_716: class Program_weight_tensor_parameter_717: name = "parameter_717" + original_name = "conv2d_35.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.0602233") @@ -7896,6 +8614,7 @@ class Program_weight_tensor_parameter_717: class Program_weight_tensor_parameter_718: name = "parameter_718" + original_name = "batch_norm2d_34.b_0" shape = [192] dtype = "float32" min_val = float("-2.55453") @@ -7907,6 +8626,7 @@ class Program_weight_tensor_parameter_718: class Program_weight_tensor_parameter_719: name = "parameter_719" + original_name = "batch_norm2d_34.w_0" shape = [192] dtype = "float32" min_val = float("-1.23245") @@ -7918,6 +8638,7 @@ class Program_weight_tensor_parameter_719: class Program_weight_tensor_parameter_720: name = "parameter_720" + original_name = "batch_norm2d_34.w_2" shape = [192] dtype = "float32" min_val = float("5.60519e-45") @@ -7929,6 +8650,7 @@ class Program_weight_tensor_parameter_720: class Program_weight_tensor_parameter_721: name = "parameter_721" + original_name = "batch_norm2d_34.w_1" shape = [192] dtype = "float32" min_val = float("-0.876927") @@ -7940,6 +8662,7 @@ class Program_weight_tensor_parameter_721: class Program_weight_tensor_parameter_722: name = "parameter_722" + original_name = "conv2d_34.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.0750585") @@ -7951,6 +8674,7 @@ class Program_weight_tensor_parameter_722: class Program_weight_tensor_parameter_723: name = "parameter_723" + original_name = "batch_norm2d_33.b_0" shape = [192] dtype = "float32" min_val = float("-3.03205") @@ -7962,6 +8686,7 @@ class Program_weight_tensor_parameter_723: class Program_weight_tensor_parameter_724: name = "parameter_724" + original_name = "batch_norm2d_33.w_0" shape = [192] dtype = "float32" min_val = float("-0.033131") @@ -7973,6 +8698,7 @@ class Program_weight_tensor_parameter_724: class Program_weight_tensor_parameter_725: name = "parameter_725" + original_name = "batch_norm2d_33.w_2" shape = [192] dtype = "float32" min_val = float("5.60519e-45") @@ -7984,6 +8710,7 @@ class Program_weight_tensor_parameter_725: class Program_weight_tensor_parameter_726: name = "parameter_726" + original_name = "batch_norm2d_33.w_1" shape = [192] dtype = "float32" min_val = float("-1.20121") @@ -7995,6 +8722,7 @@ class Program_weight_tensor_parameter_726: class Program_weight_tensor_parameter_727: name = "parameter_727" + original_name = "conv2d_33.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.0979813") @@ -8006,6 +8734,7 @@ class Program_weight_tensor_parameter_727: class Program_weight_tensor_parameter_728: name = "parameter_728" + original_name = "batch_norm2d_32.b_0" shape = [192] dtype = "float32" min_val = float("-3.10571") @@ -8017,6 +8746,7 @@ class Program_weight_tensor_parameter_728: class Program_weight_tensor_parameter_729: name = "parameter_729" + original_name = "batch_norm2d_32.w_0" shape = [192] dtype = "float32" min_val = float("-1.59511") @@ -8028,6 +8758,7 @@ class Program_weight_tensor_parameter_729: class Program_weight_tensor_parameter_730: name = "parameter_730" + original_name = "batch_norm2d_32.w_2" shape = [192] dtype = "float32" min_val = float("5.60519e-45") @@ -8039,6 +8770,7 @@ class Program_weight_tensor_parameter_730: class Program_weight_tensor_parameter_731: name = "parameter_731" + original_name = "batch_norm2d_32.w_1" shape = [192] dtype = "float32" min_val = float("-1.45372") @@ -8050,6 +8782,7 @@ class Program_weight_tensor_parameter_731: class Program_weight_tensor_parameter_732: name = "parameter_732" + original_name = "conv2d_32.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.098868") @@ -8061,6 +8794,7 @@ class Program_weight_tensor_parameter_732: class Program_weight_tensor_parameter_733: name = "parameter_733" + original_name = "batch_norm2d_31.b_0" shape = [192] dtype = "float32" min_val = float("-3.34655") @@ -8072,6 +8806,7 @@ class Program_weight_tensor_parameter_733: class Program_weight_tensor_parameter_734: name = "parameter_734" + original_name = "batch_norm2d_31.w_0" shape = [192] dtype = "float32" min_val = float("-0.0916034") @@ -8083,6 +8818,7 @@ class Program_weight_tensor_parameter_734: class Program_weight_tensor_parameter_735: name = "parameter_735" + original_name = "batch_norm2d_31.w_2" shape = [192] dtype = "float32" min_val = float("5.60519e-45") @@ -8094,6 +8830,7 @@ class Program_weight_tensor_parameter_735: class Program_weight_tensor_parameter_736: name = "parameter_736" + original_name = "batch_norm2d_31.w_1" shape = [192] dtype = "float32" min_val = float("-1.8631") @@ -8105,6 +8842,7 @@ class Program_weight_tensor_parameter_736: class Program_weight_tensor_parameter_737: name = "parameter_737" + original_name = "conv2d_31.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.0986614") @@ -8116,6 +8854,7 @@ class Program_weight_tensor_parameter_737: class Program_weight_tensor_parameter_738: name = "parameter_738" + original_name = "batch_norm2d_30.b_0" shape = [192] dtype = "float32" min_val = float("-4.11574") @@ -8127,6 +8866,7 @@ class Program_weight_tensor_parameter_738: class Program_weight_tensor_parameter_739: name = "parameter_739" + original_name = "batch_norm2d_30.w_0" shape = [192] dtype = "float32" min_val = float("0.0284117") @@ -8138,6 +8878,7 @@ class Program_weight_tensor_parameter_739: class Program_weight_tensor_parameter_740: name = "parameter_740" + original_name = "batch_norm2d_30.w_2" shape = [192] dtype = "float32" min_val = float("5.60519e-45") @@ -8149,6 +8890,7 @@ class Program_weight_tensor_parameter_740: class Program_weight_tensor_parameter_741: name = "parameter_741" + original_name = "batch_norm2d_30.w_1" shape = [192] dtype = "float32" min_val = float("-3.40615") @@ -8160,6 +8902,7 @@ class Program_weight_tensor_parameter_741: class Program_weight_tensor_parameter_742: name = "parameter_742" + original_name = "conv2d_30.w_0" shape = [192, 512, 3, 3] dtype = "float32" min_val = float("-0.139136") @@ -8171,6 +8914,7 @@ class Program_weight_tensor_parameter_742: class Program_weight_tensor_parameter_743: name = "parameter_743" + original_name = "batch_norm2d_29.b_0" shape = [512] dtype = "float32" min_val = float("-6.66784") @@ -8182,6 +8926,7 @@ class Program_weight_tensor_parameter_743: class Program_weight_tensor_parameter_744: name = "parameter_744" + original_name = "batch_norm2d_29.w_0" shape = [512] dtype = "float32" min_val = float("-0.110185") @@ -8193,6 +8938,7 @@ class Program_weight_tensor_parameter_744: class Program_weight_tensor_parameter_745: name = "parameter_745" + original_name = "batch_norm2d_29.w_2" shape = [512] dtype = "float32" min_val = float("5.60519e-45") @@ -8204,6 +8950,7 @@ class Program_weight_tensor_parameter_745: class Program_weight_tensor_parameter_746: name = "parameter_746" + original_name = "batch_norm2d_29.w_1" shape = [512] dtype = "float32" min_val = float("-5.46432") @@ -8215,6 +8962,7 @@ class Program_weight_tensor_parameter_746: class Program_weight_tensor_parameter_747: name = "parameter_747" + original_name = "conv2d_29.w_0" shape = [512, 256, 1, 1] dtype = "float32" min_val = float("-0.539237") @@ -8226,6 +8974,7 @@ class Program_weight_tensor_parameter_747: class Program_weight_tensor_parameter_748: name = "parameter_748" + original_name = "batch_norm2d_28.b_0" shape = [256] dtype = "float32" min_val = float("-4.26231") @@ -8237,6 +8986,7 @@ class Program_weight_tensor_parameter_748: class Program_weight_tensor_parameter_749: name = "parameter_749" + original_name = "batch_norm2d_28.w_0" shape = [256] dtype = "float32" min_val = float("-1.48916") @@ -8248,6 +8998,7 @@ class Program_weight_tensor_parameter_749: class Program_weight_tensor_parameter_750: name = "parameter_750" + original_name = "batch_norm2d_28.w_2" shape = [256] dtype = "float32" min_val = float("5.60519e-45") @@ -8259,6 +9010,7 @@ class Program_weight_tensor_parameter_750: class Program_weight_tensor_parameter_751: name = "parameter_751" + original_name = "batch_norm2d_28.w_1" shape = [256] dtype = "float32" min_val = float("-2.0028") @@ -8270,6 +9022,7 @@ class Program_weight_tensor_parameter_751: class Program_weight_tensor_parameter_752: name = "parameter_752" + original_name = "conv2d_28.w_0" shape = [256, 1344, 1, 1] dtype = "float32" min_val = float("-0.666644") @@ -8281,6 +9034,7 @@ class Program_weight_tensor_parameter_752: class Program_weight_tensor_parameter_753: name = "parameter_753" + original_name = "batch_norm2d_27.b_0" shape = [192] dtype = "float32" min_val = float("-1.59416") @@ -8292,6 +9046,7 @@ class Program_weight_tensor_parameter_753: class Program_weight_tensor_parameter_754: name = "parameter_754" + original_name = "batch_norm2d_27.w_0" shape = [192] dtype = "float32" min_val = float("-0.0122209") @@ -8303,6 +9058,7 @@ class Program_weight_tensor_parameter_754: class Program_weight_tensor_parameter_755: name = "parameter_755" + original_name = "batch_norm2d_27.w_2" shape = [192] dtype = "float32" min_val = float("5.60519e-45") @@ -8314,6 +9070,7 @@ class Program_weight_tensor_parameter_755: class Program_weight_tensor_parameter_756: name = "parameter_756" + original_name = "batch_norm2d_27.w_1" shape = [192] dtype = "float32" min_val = float("-0.583403") @@ -8325,6 +9082,7 @@ class Program_weight_tensor_parameter_756: class Program_weight_tensor_parameter_757: name = "parameter_757" + original_name = "conv2d_27.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.106728") @@ -8336,6 +9094,7 @@ class Program_weight_tensor_parameter_757: class Program_weight_tensor_parameter_758: name = "parameter_758" + original_name = "batch_norm2d_26.b_0" shape = [192] dtype = "float32" min_val = float("-2.83914") @@ -8347,6 +9106,7 @@ class Program_weight_tensor_parameter_758: class Program_weight_tensor_parameter_759: name = "parameter_759" + original_name = "batch_norm2d_26.w_0" shape = [192] dtype = "float32" min_val = float("-0.70529") @@ -8358,6 +9118,7 @@ class Program_weight_tensor_parameter_759: class Program_weight_tensor_parameter_760: name = "parameter_760" + original_name = "batch_norm2d_26.w_2" shape = [192] dtype = "float32" min_val = float("5.60519e-45") @@ -8369,6 +9130,7 @@ class Program_weight_tensor_parameter_760: class Program_weight_tensor_parameter_761: name = "parameter_761" + original_name = "batch_norm2d_26.w_1" shape = [192] dtype = "float32" min_val = float("-1.49712") @@ -8380,6 +9142,7 @@ class Program_weight_tensor_parameter_761: class Program_weight_tensor_parameter_762: name = "parameter_762" + original_name = "conv2d_26.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.100349") @@ -8391,6 +9154,7 @@ class Program_weight_tensor_parameter_762: class Program_weight_tensor_parameter_763: name = "parameter_763" + original_name = "batch_norm2d_25.b_0" shape = [192] dtype = "float32" min_val = float("-3.07128") @@ -8402,6 +9166,7 @@ class Program_weight_tensor_parameter_763: class Program_weight_tensor_parameter_764: name = "parameter_764" + original_name = "batch_norm2d_25.w_0" shape = [192] dtype = "float32" min_val = float("0.0178483") @@ -8413,6 +9178,7 @@ class Program_weight_tensor_parameter_764: class Program_weight_tensor_parameter_765: name = "parameter_765" + original_name = "batch_norm2d_25.w_2" shape = [192] dtype = "float32" min_val = float("5.60519e-45") @@ -8424,6 +9190,7 @@ class Program_weight_tensor_parameter_765: class Program_weight_tensor_parameter_766: name = "parameter_766" + original_name = "batch_norm2d_25.w_1" shape = [192] dtype = "float32" min_val = float("-1.06337") @@ -8435,6 +9202,7 @@ class Program_weight_tensor_parameter_766: class Program_weight_tensor_parameter_767: name = "parameter_767" + original_name = "conv2d_25.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.200282") @@ -8446,6 +9214,7 @@ class Program_weight_tensor_parameter_767: class Program_weight_tensor_parameter_768: name = "parameter_768" + original_name = "batch_norm2d_24.b_0" shape = [192] dtype = "float32" min_val = float("-3.14343") @@ -8457,6 +9226,7 @@ class Program_weight_tensor_parameter_768: class Program_weight_tensor_parameter_769: name = "parameter_769" + original_name = "batch_norm2d_24.w_0" shape = [192] dtype = "float32" min_val = float("-1.30078") @@ -8468,6 +9238,7 @@ class Program_weight_tensor_parameter_769: class Program_weight_tensor_parameter_770: name = "parameter_770" + original_name = "batch_norm2d_24.w_2" shape = [192] dtype = "float32" min_val = float("5.60519e-45") @@ -8479,6 +9250,7 @@ class Program_weight_tensor_parameter_770: class Program_weight_tensor_parameter_771: name = "parameter_771" + original_name = "batch_norm2d_24.w_1" shape = [192] dtype = "float32" min_val = float("-1.70024") @@ -8490,6 +9262,7 @@ class Program_weight_tensor_parameter_771: class Program_weight_tensor_parameter_772: name = "parameter_772" + original_name = "conv2d_24.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.136157") @@ -8501,6 +9274,7 @@ class Program_weight_tensor_parameter_772: class Program_weight_tensor_parameter_773: name = "parameter_773" + original_name = "batch_norm2d_23.b_0" shape = [192] dtype = "float32" min_val = float("-4.18016") @@ -8512,6 +9286,7 @@ class Program_weight_tensor_parameter_773: class Program_weight_tensor_parameter_774: name = "parameter_774" + original_name = "batch_norm2d_23.w_0" shape = [192] dtype = "float32" min_val = float("-0.0391092") @@ -8523,6 +9298,7 @@ class Program_weight_tensor_parameter_774: class Program_weight_tensor_parameter_775: name = "parameter_775" + original_name = "batch_norm2d_23.w_2" shape = [192] dtype = "float32" min_val = float("5.60519e-45") @@ -8534,6 +9310,7 @@ class Program_weight_tensor_parameter_775: class Program_weight_tensor_parameter_776: name = "parameter_776" + original_name = "batch_norm2d_23.w_1" shape = [192] dtype = "float32" min_val = float("-2.95315") @@ -8545,6 +9322,7 @@ class Program_weight_tensor_parameter_776: class Program_weight_tensor_parameter_777: name = "parameter_777" + original_name = "conv2d_23.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.201697") @@ -8556,6 +9334,7 @@ class Program_weight_tensor_parameter_777: class Program_weight_tensor_parameter_778: name = "parameter_778" + original_name = "batch_norm2d_22.b_0" shape = [192] dtype = "float32" min_val = float("-4.56663") @@ -8567,6 +9346,7 @@ class Program_weight_tensor_parameter_778: class Program_weight_tensor_parameter_779: name = "parameter_779" + original_name = "batch_norm2d_22.w_0" shape = [192] dtype = "float32" min_val = float("0.179058") @@ -8578,6 +9358,7 @@ class Program_weight_tensor_parameter_779: class Program_weight_tensor_parameter_780: name = "parameter_780" + original_name = "batch_norm2d_22.w_2" shape = [192] dtype = "float32" min_val = float("5.60519e-45") @@ -8589,6 +9370,7 @@ class Program_weight_tensor_parameter_780: class Program_weight_tensor_parameter_781: name = "parameter_781" + original_name = "batch_norm2d_22.w_1" shape = [192] dtype = "float32" min_val = float("-3.45587") @@ -8600,6 +9382,7 @@ class Program_weight_tensor_parameter_781: class Program_weight_tensor_parameter_782: name = "parameter_782" + original_name = "conv2d_22.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.328479") @@ -8611,6 +9394,7 @@ class Program_weight_tensor_parameter_782: class Program_weight_tensor_parameter_783: name = "parameter_783" + original_name = "batch_norm2d_21.b_0" shape = [192] dtype = "float32" min_val = float("-1.91921") @@ -8622,6 +9406,7 @@ class Program_weight_tensor_parameter_783: class Program_weight_tensor_parameter_784: name = "parameter_784" + original_name = "batch_norm2d_21.w_0" shape = [192] dtype = "float32" min_val = float("-1.66302") @@ -8633,6 +9418,7 @@ class Program_weight_tensor_parameter_784: class Program_weight_tensor_parameter_785: name = "parameter_785" + original_name = "batch_norm2d_21.w_2" shape = [192] dtype = "float32" min_val = float("5.60519e-45") @@ -8644,6 +9430,7 @@ class Program_weight_tensor_parameter_785: class Program_weight_tensor_parameter_786: name = "parameter_786" + original_name = "batch_norm2d_21.w_1" shape = [192] dtype = "float32" min_val = float("-2.31508") @@ -8655,6 +9442,7 @@ class Program_weight_tensor_parameter_786: class Program_weight_tensor_parameter_787: name = "parameter_787" + original_name = "conv2d_21.w_0" shape = [192, 1, 3, 3] dtype = "float32" min_val = float("-0.577024") @@ -8666,6 +9454,7 @@ class Program_weight_tensor_parameter_787: class Program_weight_tensor_parameter_788: name = "parameter_788" + original_name = "batch_norm2d_20.b_0" shape = [192] dtype = "float32" min_val = float("-4.0118") @@ -8677,6 +9466,7 @@ class Program_weight_tensor_parameter_788: class Program_weight_tensor_parameter_789: name = "parameter_789" + original_name = "batch_norm2d_20.w_0" shape = [192] dtype = "float32" min_val = float("-5.32204") @@ -8688,6 +9478,7 @@ class Program_weight_tensor_parameter_789: class Program_weight_tensor_parameter_790: name = "parameter_790" + original_name = "batch_norm2d_20.w_2" shape = [192] dtype = "float32" min_val = float("5.60519e-45") @@ -8699,6 +9490,7 @@ class Program_weight_tensor_parameter_790: class Program_weight_tensor_parameter_791: name = "parameter_791" + original_name = "batch_norm2d_20.w_1" shape = [192] dtype = "float32" min_val = float("-1.18109") @@ -8710,6 +9502,7 @@ class Program_weight_tensor_parameter_791: class Program_weight_tensor_parameter_792: name = "parameter_792" + original_name = "conv2d_20.w_0" shape = [192, 96, 1, 1] dtype = "float32" min_val = float("-0.339843") @@ -8721,6 +9514,7 @@ class Program_weight_tensor_parameter_792: class Program_weight_tensor_parameter_793: name = "parameter_793" + original_name = "batch_norm2d_19.b_0" shape = [96] dtype = "float32" min_val = float("-4.78472") @@ -8732,6 +9526,7 @@ class Program_weight_tensor_parameter_793: class Program_weight_tensor_parameter_794: name = "parameter_794" + original_name = "batch_norm2d_19.w_0" shape = [96] dtype = "float32" min_val = float("0.149152") @@ -8743,6 +9538,7 @@ class Program_weight_tensor_parameter_794: class Program_weight_tensor_parameter_795: name = "parameter_795" + original_name = "batch_norm2d_19.w_2" shape = [96] dtype = "float32" min_val = float("5.60519e-45") @@ -8754,6 +9550,7 @@ class Program_weight_tensor_parameter_795: class Program_weight_tensor_parameter_796: name = "parameter_796" + original_name = "batch_norm2d_19.w_1" shape = [96] dtype = "float32" min_val = float("-1.7431") @@ -8765,6 +9562,7 @@ class Program_weight_tensor_parameter_796: class Program_weight_tensor_parameter_797: name = "parameter_797" + original_name = "conv2d_19.w_0" shape = [96, 768, 1, 1] dtype = "float32" min_val = float("-0.212997") @@ -8776,6 +9574,7 @@ class Program_weight_tensor_parameter_797: class Program_weight_tensor_parameter_798: name = "parameter_798" + original_name = "batch_norm2d_18.b_0" shape = [96] dtype = "float32" min_val = float("-3.29619") @@ -8787,6 +9586,7 @@ class Program_weight_tensor_parameter_798: class Program_weight_tensor_parameter_799: name = "parameter_799" + original_name = "batch_norm2d_18.w_0" shape = [96] dtype = "float32" min_val = float("1.10442") @@ -8798,6 +9598,7 @@ class Program_weight_tensor_parameter_799: class Program_weight_tensor_parameter_800: name = "parameter_800" + original_name = "batch_norm2d_18.w_2" shape = [96] dtype = "float32" min_val = float("0.0885448") @@ -8809,6 +9610,7 @@ class Program_weight_tensor_parameter_800: class Program_weight_tensor_parameter_801: name = "parameter_801" + original_name = "batch_norm2d_18.w_1" shape = [96] dtype = "float32" min_val = float("-0.965519") @@ -8820,6 +9622,7 @@ class Program_weight_tensor_parameter_801: class Program_weight_tensor_parameter_802: name = "parameter_802" + original_name = "conv2d_18.w_0" shape = [96, 96, 3, 3] dtype = "float32" min_val = float("-0.0878876") @@ -8831,6 +9634,7 @@ class Program_weight_tensor_parameter_802: class Program_weight_tensor_parameter_803: name = "parameter_803" + original_name = "batch_norm2d_17.b_0" shape = [96] dtype = "float32" min_val = float("-3.78168") @@ -8842,6 +9646,7 @@ class Program_weight_tensor_parameter_803: class Program_weight_tensor_parameter_804: name = "parameter_804" + original_name = "batch_norm2d_17.w_0" shape = [96] dtype = "float32" min_val = float("0.657292") @@ -8853,6 +9658,7 @@ class Program_weight_tensor_parameter_804: class Program_weight_tensor_parameter_805: name = "parameter_805" + original_name = "batch_norm2d_17.w_2" shape = [96] dtype = "float32" min_val = float("5.60519e-45") @@ -8864,6 +9670,7 @@ class Program_weight_tensor_parameter_805: class Program_weight_tensor_parameter_806: name = "parameter_806" + original_name = "batch_norm2d_17.w_1" shape = [96] dtype = "float32" min_val = float("-1.82934") @@ -8875,6 +9682,7 @@ class Program_weight_tensor_parameter_806: class Program_weight_tensor_parameter_807: name = "parameter_807" + original_name = "conv2d_17.w_0" shape = [96, 96, 3, 3] dtype = "float32" min_val = float("-0.122299") @@ -8886,6 +9694,7 @@ class Program_weight_tensor_parameter_807: class Program_weight_tensor_parameter_808: name = "parameter_808" + original_name = "batch_norm2d_16.b_0" shape = [96] dtype = "float32" min_val = float("-4.10839") @@ -8897,6 +9706,7 @@ class Program_weight_tensor_parameter_808: class Program_weight_tensor_parameter_809: name = "parameter_809" + original_name = "batch_norm2d_16.w_0" shape = [96] dtype = "float32" min_val = float("0.835909") @@ -8908,6 +9718,7 @@ class Program_weight_tensor_parameter_809: class Program_weight_tensor_parameter_810: name = "parameter_810" + original_name = "batch_norm2d_16.w_2" shape = [96] dtype = "float32" min_val = float("0.32802") @@ -8919,6 +9730,7 @@ class Program_weight_tensor_parameter_810: class Program_weight_tensor_parameter_811: name = "parameter_811" + original_name = "batch_norm2d_16.w_1" shape = [96] dtype = "float32" min_val = float("-2.33553") @@ -8930,6 +9742,7 @@ class Program_weight_tensor_parameter_811: class Program_weight_tensor_parameter_812: name = "parameter_812" + original_name = "conv2d_16.w_0" shape = [96, 96, 3, 3] dtype = "float32" min_val = float("-0.144295") @@ -8941,6 +9754,7 @@ class Program_weight_tensor_parameter_812: class Program_weight_tensor_parameter_813: name = "parameter_813" + original_name = "batch_norm2d_15.b_0" shape = [96] dtype = "float32" min_val = float("-3.65341") @@ -8952,6 +9766,7 @@ class Program_weight_tensor_parameter_813: class Program_weight_tensor_parameter_814: name = "parameter_814" + original_name = "batch_norm2d_15.w_0" shape = [96] dtype = "float32" min_val = float("0.292533") @@ -8963,6 +9778,7 @@ class Program_weight_tensor_parameter_814: class Program_weight_tensor_parameter_815: name = "parameter_815" + original_name = "batch_norm2d_15.w_2" shape = [96] dtype = "float32" min_val = float("5.60519e-45") @@ -8974,6 +9790,7 @@ class Program_weight_tensor_parameter_815: class Program_weight_tensor_parameter_816: name = "parameter_816" + original_name = "batch_norm2d_15.w_1" shape = [96] dtype = "float32" min_val = float("-2.09968") @@ -8985,6 +9802,7 @@ class Program_weight_tensor_parameter_816: class Program_weight_tensor_parameter_817: name = "parameter_817" + original_name = "conv2d_15.w_0" shape = [96, 96, 3, 3] dtype = "float32" min_val = float("-0.179792") @@ -8996,6 +9814,7 @@ class Program_weight_tensor_parameter_817: class Program_weight_tensor_parameter_818: name = "parameter_818" + original_name = "batch_norm2d_14.b_0" shape = [96] dtype = "float32" min_val = float("-3.07563") @@ -9007,6 +9826,7 @@ class Program_weight_tensor_parameter_818: class Program_weight_tensor_parameter_819: name = "parameter_819" + original_name = "batch_norm2d_14.w_0" shape = [96] dtype = "float32" min_val = float("1.09966") @@ -9018,6 +9838,7 @@ class Program_weight_tensor_parameter_819: class Program_weight_tensor_parameter_820: name = "parameter_820" + original_name = "batch_norm2d_14.w_2" shape = [96] dtype = "float32" min_val = float("0.543494") @@ -9029,6 +9850,7 @@ class Program_weight_tensor_parameter_820: class Program_weight_tensor_parameter_821: name = "parameter_821" + original_name = "batch_norm2d_14.w_1" shape = [96] dtype = "float32" min_val = float("-2.40007") @@ -9040,6 +9862,7 @@ class Program_weight_tensor_parameter_821: class Program_weight_tensor_parameter_822: name = "parameter_822" + original_name = "conv2d_14.w_0" shape = [96, 96, 3, 3] dtype = "float32" min_val = float("-0.201563") @@ -9051,6 +9874,7 @@ class Program_weight_tensor_parameter_822: class Program_weight_tensor_parameter_823: name = "parameter_823" + original_name = "batch_norm2d_13.b_0" shape = [96] dtype = "float32" min_val = float("-5.94967") @@ -9062,6 +9886,7 @@ class Program_weight_tensor_parameter_823: class Program_weight_tensor_parameter_824: name = "parameter_824" + original_name = "batch_norm2d_13.w_0" shape = [96] dtype = "float32" min_val = float("0.589796") @@ -9073,6 +9898,7 @@ class Program_weight_tensor_parameter_824: class Program_weight_tensor_parameter_825: name = "parameter_825" + original_name = "batch_norm2d_13.w_2" shape = [96] dtype = "float32" min_val = float("0.271463") @@ -9084,6 +9910,7 @@ class Program_weight_tensor_parameter_825: class Program_weight_tensor_parameter_826: name = "parameter_826" + original_name = "batch_norm2d_13.w_1" shape = [96] dtype = "float32" min_val = float("-4.85095") @@ -9095,6 +9922,7 @@ class Program_weight_tensor_parameter_826: class Program_weight_tensor_parameter_827: name = "parameter_827" + original_name = "conv2d_13.w_0" shape = [96, 192, 3, 3] dtype = "float32" min_val = float("-0.295496") @@ -9106,6 +9934,7 @@ class Program_weight_tensor_parameter_827: class Program_weight_tensor_parameter_828: name = "parameter_828" + original_name = "batch_norm2d_12.b_0" shape = [192] dtype = "float32" min_val = float("-6.23769") @@ -9117,6 +9946,7 @@ class Program_weight_tensor_parameter_828: class Program_weight_tensor_parameter_829: name = "parameter_829" + original_name = "batch_norm2d_12.w_0" shape = [192] dtype = "float32" min_val = float("-3.40597") @@ -9128,6 +9958,7 @@ class Program_weight_tensor_parameter_829: class Program_weight_tensor_parameter_830: name = "parameter_830" + original_name = "batch_norm2d_12.w_2" shape = [192] dtype = "float32" min_val = float("5.60519e-45") @@ -9139,6 +9970,7 @@ class Program_weight_tensor_parameter_830: class Program_weight_tensor_parameter_831: name = "parameter_831" + original_name = "batch_norm2d_12.w_1" shape = [192] dtype = "float32" min_val = float("-7.92672") @@ -9150,6 +9982,7 @@ class Program_weight_tensor_parameter_831: class Program_weight_tensor_parameter_832: name = "parameter_832" + original_name = "conv2d_12.w_0" shape = [192, 96, 1, 1] dtype = "float32" min_val = float("-0.5598") @@ -9161,6 +9994,7 @@ class Program_weight_tensor_parameter_832: class Program_weight_tensor_parameter_833: name = "parameter_833" + original_name = "batch_norm2d_11.b_0" shape = [96] dtype = "float32" min_val = float("-5.45669") @@ -9172,6 +10006,7 @@ class Program_weight_tensor_parameter_833: class Program_weight_tensor_parameter_834: name = "parameter_834" + original_name = "batch_norm2d_11.w_0" shape = [96] dtype = "float32" min_val = float("-0.0517751") @@ -9183,6 +10018,7 @@ class Program_weight_tensor_parameter_834: class Program_weight_tensor_parameter_835: name = "parameter_835" + original_name = "batch_norm2d_11.w_2" shape = [96] dtype = "float32" min_val = float("5.60519e-45") @@ -9194,6 +10030,7 @@ class Program_weight_tensor_parameter_835: class Program_weight_tensor_parameter_836: name = "parameter_836" + original_name = "batch_norm2d_11.w_1" shape = [96] dtype = "float32" min_val = float("-2.68274") @@ -9205,6 +10042,7 @@ class Program_weight_tensor_parameter_836: class Program_weight_tensor_parameter_837: name = "parameter_837" + original_name = "conv2d_11.w_0" shape = [96, 672, 1, 1] dtype = "float32" min_val = float("-0.511724") @@ -9216,6 +10054,7 @@ class Program_weight_tensor_parameter_837: class Program_weight_tensor_parameter_838: name = "parameter_838" + original_name = "batch_norm2d_10.b_0" shape = [96] dtype = "float32" min_val = float("-1.333") @@ -9227,6 +10066,7 @@ class Program_weight_tensor_parameter_838: class Program_weight_tensor_parameter_839: name = "parameter_839" + original_name = "batch_norm2d_10.w_0" shape = [96] dtype = "float32" min_val = float("-0.212284") @@ -9238,6 +10078,7 @@ class Program_weight_tensor_parameter_839: class Program_weight_tensor_parameter_840: name = "parameter_840" + original_name = "batch_norm2d_10.w_2" shape = [96] dtype = "float32" min_val = float("5.60519e-45") @@ -9249,6 +10090,7 @@ class Program_weight_tensor_parameter_840: class Program_weight_tensor_parameter_841: name = "parameter_841" + original_name = "batch_norm2d_10.w_1" shape = [96] dtype = "float32" min_val = float("-0.663517") @@ -9260,6 +10102,7 @@ class Program_weight_tensor_parameter_841: class Program_weight_tensor_parameter_842: name = "parameter_842" + original_name = "conv2d_10.w_0" shape = [96, 96, 3, 3] dtype = "float32" min_val = float("-0.0908706") @@ -9271,6 +10114,7 @@ class Program_weight_tensor_parameter_842: class Program_weight_tensor_parameter_843: name = "parameter_843" + original_name = "batch_norm2d_9.b_0" shape = [96] dtype = "float32" min_val = float("-1.54108") @@ -9282,6 +10126,7 @@ class Program_weight_tensor_parameter_843: class Program_weight_tensor_parameter_844: name = "parameter_844" + original_name = "batch_norm2d_9.w_0" shape = [96] dtype = "float32" min_val = float("-0.0465636") @@ -9293,6 +10138,7 @@ class Program_weight_tensor_parameter_844: class Program_weight_tensor_parameter_845: name = "parameter_845" + original_name = "batch_norm2d_9.w_2" shape = [96] dtype = "float32" min_val = float("5.60519e-45") @@ -9304,6 +10150,7 @@ class Program_weight_tensor_parameter_845: class Program_weight_tensor_parameter_846: name = "parameter_846" + original_name = "batch_norm2d_9.w_1" shape = [96] dtype = "float32" min_val = float("-1.48957") @@ -9315,6 +10162,7 @@ class Program_weight_tensor_parameter_846: class Program_weight_tensor_parameter_847: name = "parameter_847" + original_name = "conv2d_9.w_0" shape = [96, 96, 3, 3] dtype = "float32" min_val = float("-0.104713") @@ -9326,6 +10174,7 @@ class Program_weight_tensor_parameter_847: class Program_weight_tensor_parameter_848: name = "parameter_848" + original_name = "batch_norm2d_8.b_0" shape = [96] dtype = "float32" min_val = float("-1.88164") @@ -9337,6 +10186,7 @@ class Program_weight_tensor_parameter_848: class Program_weight_tensor_parameter_849: name = "parameter_849" + original_name = "batch_norm2d_8.w_0" shape = [96] dtype = "float32" min_val = float("-1.07452") @@ -9348,6 +10198,7 @@ class Program_weight_tensor_parameter_849: class Program_weight_tensor_parameter_850: name = "parameter_850" + original_name = "batch_norm2d_8.w_2" shape = [96] dtype = "float32" min_val = float("5.60519e-45") @@ -9359,6 +10210,7 @@ class Program_weight_tensor_parameter_850: class Program_weight_tensor_parameter_851: name = "parameter_851" + original_name = "batch_norm2d_8.w_1" shape = [96] dtype = "float32" min_val = float("-0.897463") @@ -9370,6 +10222,7 @@ class Program_weight_tensor_parameter_851: class Program_weight_tensor_parameter_852: name = "parameter_852" + original_name = "conv2d_8.w_0" shape = [96, 96, 3, 3] dtype = "float32" min_val = float("-0.21468") @@ -9381,6 +10234,7 @@ class Program_weight_tensor_parameter_852: class Program_weight_tensor_parameter_853: name = "parameter_853" + original_name = "batch_norm2d_7.b_0" shape = [96] dtype = "float32" min_val = float("-7.38151") @@ -9392,6 +10246,7 @@ class Program_weight_tensor_parameter_853: class Program_weight_tensor_parameter_854: name = "parameter_854" + original_name = "batch_norm2d_7.w_0" shape = [96] dtype = "float32" min_val = float("-3.61004") @@ -9403,6 +10258,7 @@ class Program_weight_tensor_parameter_854: class Program_weight_tensor_parameter_855: name = "parameter_855" + original_name = "batch_norm2d_7.w_2" shape = [96] dtype = "float32" min_val = float("5.60519e-45") @@ -9414,6 +10270,7 @@ class Program_weight_tensor_parameter_855: class Program_weight_tensor_parameter_856: name = "parameter_856" + original_name = "batch_norm2d_7.w_1" shape = [96] dtype = "float32" min_val = float("-1.64917") @@ -9425,6 +10282,7 @@ class Program_weight_tensor_parameter_856: class Program_weight_tensor_parameter_857: name = "parameter_857" + original_name = "conv2d_7.w_0" shape = [96, 96, 3, 3] dtype = "float32" min_val = float("-0.200674") @@ -9436,6 +10294,7 @@ class Program_weight_tensor_parameter_857: class Program_weight_tensor_parameter_858: name = "parameter_858" + original_name = "batch_norm2d_6.b_0" shape = [96] dtype = "float32" min_val = float("-3.48645") @@ -9447,6 +10306,7 @@ class Program_weight_tensor_parameter_858: class Program_weight_tensor_parameter_859: name = "parameter_859" + original_name = "batch_norm2d_6.w_0" shape = [96] dtype = "float32" min_val = float("-2.52693") @@ -9458,6 +10318,7 @@ class Program_weight_tensor_parameter_859: class Program_weight_tensor_parameter_860: name = "parameter_860" + original_name = "batch_norm2d_6.w_2" shape = [96] dtype = "float32" min_val = float("5.60519e-45") @@ -9469,6 +10330,7 @@ class Program_weight_tensor_parameter_860: class Program_weight_tensor_parameter_861: name = "parameter_861" + original_name = "batch_norm2d_6.w_1" shape = [96] dtype = "float32" min_val = float("-1.47867") @@ -9480,6 +10342,7 @@ class Program_weight_tensor_parameter_861: class Program_weight_tensor_parameter_862: name = "parameter_862" + original_name = "conv2d_6.w_0" shape = [96, 96, 3, 3] dtype = "float32" min_val = float("-0.334288") @@ -9491,6 +10354,7 @@ class Program_weight_tensor_parameter_862: class Program_weight_tensor_parameter_863: name = "parameter_863" + original_name = "batch_norm2d_5.b_0" shape = [96] dtype = "float32" min_val = float("-4.9204") @@ -9502,6 +10366,7 @@ class Program_weight_tensor_parameter_863: class Program_weight_tensor_parameter_864: name = "parameter_864" + original_name = "batch_norm2d_5.w_0" shape = [96] dtype = "float32" min_val = float("-0.260154") @@ -9513,6 +10378,7 @@ class Program_weight_tensor_parameter_864: class Program_weight_tensor_parameter_865: name = "parameter_865" + original_name = "batch_norm2d_5.w_2" shape = [96] dtype = "float32" min_val = float("5.60519e-45") @@ -9524,6 +10390,7 @@ class Program_weight_tensor_parameter_865: class Program_weight_tensor_parameter_866: name = "parameter_866" + original_name = "batch_norm2d_5.w_1" shape = [96] dtype = "float32" min_val = float("-10.9535") @@ -9535,6 +10402,7 @@ class Program_weight_tensor_parameter_866: class Program_weight_tensor_parameter_867: name = "parameter_867" + original_name = "conv2d_5.w_0" shape = [96, 96, 3, 3] dtype = "float32" min_val = float("-0.382021") @@ -9546,6 +10414,7 @@ class Program_weight_tensor_parameter_867: class Program_weight_tensor_parameter_868: name = "parameter_868" + original_name = "batch_norm2d_4.b_0" shape = [96] dtype = "float32" min_val = float("-5.77021") @@ -9557,6 +10426,7 @@ class Program_weight_tensor_parameter_868: class Program_weight_tensor_parameter_869: name = "parameter_869" + original_name = "batch_norm2d_4.w_0" shape = [96] dtype = "float32" min_val = float("0.534115") @@ -9568,6 +10438,7 @@ class Program_weight_tensor_parameter_869: class Program_weight_tensor_parameter_870: name = "parameter_870" + original_name = "batch_norm2d_4.w_2" shape = [96] dtype = "float32" min_val = float("5.60519e-45") @@ -9579,6 +10450,7 @@ class Program_weight_tensor_parameter_870: class Program_weight_tensor_parameter_871: name = "parameter_871" + original_name = "batch_norm2d_4.w_1" shape = [96] dtype = "float32" min_val = float("-8.14601") @@ -9590,6 +10462,7 @@ class Program_weight_tensor_parameter_871: class Program_weight_tensor_parameter_872: name = "parameter_872" + original_name = "conv2d_4.w_0" shape = [96, 48, 1, 1] dtype = "float32" min_val = float("-0.539821") @@ -9601,6 +10474,7 @@ class Program_weight_tensor_parameter_872: class Program_weight_tensor_parameter_873: name = "parameter_873" + original_name = "batch_norm2d_3.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -9610,6 +10484,7 @@ class Program_weight_tensor_parameter_873: class Program_weight_tensor_parameter_874: name = "parameter_874" + original_name = "batch_norm2d_3.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -9619,6 +10494,7 @@ class Program_weight_tensor_parameter_874: class Program_weight_tensor_parameter_875: name = "parameter_875" + original_name = "batch_norm2d_3.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -9628,6 +10504,7 @@ class Program_weight_tensor_parameter_875: class Program_weight_tensor_parameter_876: name = "parameter_876" + original_name = "batch_norm2d_3.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -9637,6 +10514,7 @@ class Program_weight_tensor_parameter_876: class Program_weight_tensor_parameter_877: name = "parameter_877" + original_name = "conv2d_3.w_0" shape = [48, 96, 3, 3] dtype = "float32" min_val = float("-0.646409") @@ -9648,6 +10526,7 @@ class Program_weight_tensor_parameter_877: class Program_weight_tensor_parameter_878: name = "parameter_878" + original_name = "batch_norm2d_2.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -9657,6 +10536,7 @@ class Program_weight_tensor_parameter_878: class Program_weight_tensor_parameter_879: name = "parameter_879" + original_name = "batch_norm2d_2.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -9666,6 +10546,7 @@ class Program_weight_tensor_parameter_879: class Program_weight_tensor_parameter_880: name = "parameter_880" + original_name = "batch_norm2d_2.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -9675,6 +10556,7 @@ class Program_weight_tensor_parameter_880: class Program_weight_tensor_parameter_881: name = "parameter_881" + original_name = "batch_norm2d_2.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -9684,6 +10566,7 @@ class Program_weight_tensor_parameter_881: class Program_weight_tensor_parameter_882: name = "parameter_882" + original_name = "conv2d_2.w_0" shape = [48, 24, 2, 2] dtype = "float32" min_val = float("-0.431558") @@ -9695,6 +10578,7 @@ class Program_weight_tensor_parameter_882: class Program_weight_tensor_parameter_883: name = "parameter_883" + original_name = "batch_norm2d_1.b_0" shape = [24] dtype = "float32" min_val = float("0") @@ -9704,6 +10588,7 @@ class Program_weight_tensor_parameter_883: class Program_weight_tensor_parameter_884: name = "parameter_884" + original_name = "batch_norm2d_1.w_0" shape = [24] dtype = "float32" min_val = float("0") @@ -9713,6 +10598,7 @@ class Program_weight_tensor_parameter_884: class Program_weight_tensor_parameter_885: name = "parameter_885" + original_name = "batch_norm2d_1.w_2" shape = [24] dtype = "float32" min_val = float("0") @@ -9722,6 +10608,7 @@ class Program_weight_tensor_parameter_885: class Program_weight_tensor_parameter_886: name = "parameter_886" + original_name = "batch_norm2d_1.w_1" shape = [24] dtype = "float32" min_val = float("0") @@ -9731,6 +10618,7 @@ class Program_weight_tensor_parameter_886: class Program_weight_tensor_parameter_887: name = "parameter_887" + original_name = "conv2d_1.w_0" shape = [24, 48, 2, 2] dtype = "float32" min_val = float("-0.866387") @@ -9742,6 +10630,7 @@ class Program_weight_tensor_parameter_887: class Program_weight_tensor_parameter_888: name = "parameter_888" + original_name = "batch_norm2d_0.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -9751,6 +10640,7 @@ class Program_weight_tensor_parameter_888: class Program_weight_tensor_parameter_889: name = "parameter_889" + original_name = "batch_norm2d_0.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -9760,6 +10650,7 @@ class Program_weight_tensor_parameter_889: class Program_weight_tensor_parameter_890: name = "parameter_890" + original_name = "batch_norm2d_0.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -9769,6 +10660,7 @@ class Program_weight_tensor_parameter_890: class Program_weight_tensor_parameter_891: name = "parameter_891" + original_name = "batch_norm2d_0.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -9778,6 +10670,7 @@ class Program_weight_tensor_parameter_891: class Program_weight_tensor_parameter_892: name = "parameter_892" + original_name = "conv2d_0.w_0" shape = [48, 3, 3, 3] dtype = "float32" min_val = float("-0.870867") diff --git a/paddle_samples/PaddleX/PP-LCNet_x1_0_ML/subgraph_0/input_meta.py b/paddle_samples/PaddleX/PP-LCNet_x1_0_ML/subgraph_0/input_meta.py index e4033b4ea..5a911434e 100644 --- a/paddle_samples/PaddleX/PP-LCNet_x1_0_ML/subgraph_0/input_meta.py +++ b/paddle_samples/PaddleX/PP-LCNet_x1_0_ML/subgraph_0/input_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_data_0: name = "data_0" + original_name = "param_0" shape = [33, 768] dtype = "float32" min_val = float("-0.0865401") @@ -11,6 +12,7 @@ class Program_weight_tensor_data_0: class Program_weight_tensor_data_1: name = "data_1" + original_name = "var_0" shape = [32, 3, 448, 448] dtype = "float32" min_val = float("-2.1179") diff --git a/paddle_samples/PaddleX/PP-LCNet_x1_0_ML/subgraph_0/weight_meta.py b/paddle_samples/PaddleX/PP-LCNet_x1_0_ML/subgraph_0/weight_meta.py index 3d96d1a44..71df44a09 100644 --- a/paddle_samples/PaddleX/PP-LCNet_x1_0_ML/subgraph_0/weight_meta.py +++ b/paddle_samples/PaddleX/PP-LCNet_x1_0_ML/subgraph_0/weight_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_parameter_0: name = "parameter_0" + original_name = "conv2d_33.b_0" shape = [33] dtype = "float32" min_val = float("0") @@ -9,6 +10,7 @@ class Program_weight_tensor_parameter_0: class Program_weight_tensor_parameter_1: name = "parameter_1" + original_name = "conv2d_33.w_0" shape = [33, 768, 1, 1] dtype = "float32" min_val = float("-0.214921") @@ -20,6 +22,7 @@ class Program_weight_tensor_parameter_1: class Program_weight_tensor_parameter_2: name = "parameter_2" + original_name = "layer_norm_2.b_0" shape = [768] dtype = "float32" min_val = float("-0.0064855") @@ -31,6 +34,7 @@ class Program_weight_tensor_parameter_2: class Program_weight_tensor_parameter_3: name = "parameter_3" + original_name = "layer_norm_2.w_0" shape = [768] dtype = "float32" min_val = float("0.920068") @@ -42,6 +46,7 @@ class Program_weight_tensor_parameter_3: class Program_weight_tensor_parameter_4: name = "parameter_4" + original_name = "linear_10.b_0" shape = [768] dtype = "float32" min_val = float("-0.0258869") @@ -53,6 +58,7 @@ class Program_weight_tensor_parameter_4: class Program_weight_tensor_parameter_5: name = "parameter_5" + original_name = "linear_10.w_0" shape = [2048, 768] dtype = "float32" min_val = float("-0.133631") @@ -64,6 +70,7 @@ class Program_weight_tensor_parameter_5: class Program_weight_tensor_parameter_6: name = "parameter_6" + original_name = "linear_9.b_0" shape = [2048] dtype = "float32" min_val = float("-0.0204863") @@ -75,6 +82,7 @@ class Program_weight_tensor_parameter_6: class Program_weight_tensor_parameter_7: name = "parameter_7" + original_name = "linear_9.w_0" shape = [768, 2048] dtype = "float32" min_val = float("-0.0705326") @@ -86,6 +94,7 @@ class Program_weight_tensor_parameter_7: class Program_weight_tensor_parameter_8: name = "parameter_8" + original_name = "layer_norm_1.b_0" shape = [768] dtype = "float32" min_val = float("-0.0784116") @@ -97,6 +106,7 @@ class Program_weight_tensor_parameter_8: class Program_weight_tensor_parameter_9: name = "parameter_9" + original_name = "layer_norm_1.w_0" shape = [768] dtype = "float32" min_val = float("0.983765") @@ -108,6 +118,7 @@ class Program_weight_tensor_parameter_9: class Program_weight_tensor_parameter_10: name = "parameter_10" + original_name = "linear_8.b_0" shape = [768] dtype = "float32" min_val = float("-0.0465604") @@ -119,6 +130,7 @@ class Program_weight_tensor_parameter_10: class Program_weight_tensor_parameter_11: name = "parameter_11" + original_name = "linear_8.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.175042") @@ -130,6 +142,7 @@ class Program_weight_tensor_parameter_11: class Program_weight_tensor_parameter_12: name = "parameter_12" + original_name = "linear_7.b_0" shape = [768] dtype = "float32" min_val = float("-0.00762") @@ -141,6 +154,7 @@ class Program_weight_tensor_parameter_12: class Program_weight_tensor_parameter_13: name = "parameter_13" + original_name = "linear_7.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.0872142") @@ -152,6 +166,7 @@ class Program_weight_tensor_parameter_13: class Program_weight_tensor_parameter_14: name = "parameter_14" + original_name = "linear_6.b_0" shape = [768] dtype = "float32" min_val = float("-0.0155351") @@ -163,6 +178,7 @@ class Program_weight_tensor_parameter_14: class Program_weight_tensor_parameter_15: name = "parameter_15" + original_name = "linear_6.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.0931939") @@ -174,6 +190,7 @@ class Program_weight_tensor_parameter_15: class Program_weight_tensor_parameter_16: name = "parameter_16" + original_name = "linear_5.b_0" shape = [768] dtype = "float32" min_val = float("-0.0357969") @@ -185,6 +202,7 @@ class Program_weight_tensor_parameter_16: class Program_weight_tensor_parameter_17: name = "parameter_17" + original_name = "linear_5.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.0834058") @@ -196,6 +214,7 @@ class Program_weight_tensor_parameter_17: class Program_weight_tensor_parameter_18: name = "parameter_18" + original_name = "layer_norm_0.b_0" shape = [768] dtype = "float32" min_val = float("-0.0832166") @@ -207,6 +226,7 @@ class Program_weight_tensor_parameter_18: class Program_weight_tensor_parameter_19: name = "parameter_19" + original_name = "layer_norm_0.w_0" shape = [768] dtype = "float32" min_val = float("0.979694") @@ -218,6 +238,7 @@ class Program_weight_tensor_parameter_19: class Program_weight_tensor_parameter_20: name = "parameter_20" + original_name = "conv2d_32.b_0" shape = [768] dtype = "float32" min_val = float("-0.00614192") @@ -229,6 +250,7 @@ class Program_weight_tensor_parameter_20: class Program_weight_tensor_parameter_21: name = "parameter_21" + original_name = "conv2d_32.w_0" shape = [768, 1280, 1, 1] dtype = "float32" min_val = float("-0.183057") @@ -240,6 +262,7 @@ class Program_weight_tensor_parameter_21: class Program_weight_tensor_parameter_22: name = "parameter_22" + original_name = "conv2d_31.w_0" shape = [1280, 512, 1, 1] dtype = "float32" min_val = float("-0.635346") @@ -251,6 +274,7 @@ class Program_weight_tensor_parameter_22: class Program_weight_tensor_parameter_23: name = "parameter_23" + original_name = "batch_norm2d_26.b_0" shape = [512] dtype = "float32" min_val = float("-5.09144") @@ -262,6 +286,7 @@ class Program_weight_tensor_parameter_23: class Program_weight_tensor_parameter_24: name = "parameter_24" + original_name = "batch_norm2d_26.w_0" shape = [512] dtype = "float32" min_val = float("2.51098") @@ -273,6 +298,7 @@ class Program_weight_tensor_parameter_24: class Program_weight_tensor_parameter_25: name = "parameter_25" + original_name = "batch_norm2d_26.w_2" shape = [512] dtype = "float32" min_val = float("0.0212376") @@ -284,6 +310,7 @@ class Program_weight_tensor_parameter_25: class Program_weight_tensor_parameter_26: name = "parameter_26" + original_name = "batch_norm2d_26.w_1" shape = [512] dtype = "float32" min_val = float("-0.704243") @@ -295,6 +322,7 @@ class Program_weight_tensor_parameter_26: class Program_weight_tensor_parameter_27: name = "parameter_27" + original_name = "conv2d_30.w_0" shape = [512, 512, 1, 1] dtype = "float32" min_val = float("-0.587353") @@ -306,6 +334,7 @@ class Program_weight_tensor_parameter_27: class Program_weight_tensor_parameter_28: name = "parameter_28" + original_name = "conv2d_29.b_0" shape = [512] dtype = "float32" min_val = float("-0.216613") @@ -317,6 +346,7 @@ class Program_weight_tensor_parameter_28: class Program_weight_tensor_parameter_29: name = "parameter_29" + original_name = "conv2d_29.w_0" shape = [512, 128, 1, 1] dtype = "float32" min_val = float("-0.517766") @@ -328,6 +358,7 @@ class Program_weight_tensor_parameter_29: class Program_weight_tensor_parameter_30: name = "parameter_30" + original_name = "conv2d_28.b_0" shape = [128] dtype = "float32" min_val = float("-0.0701944") @@ -339,6 +370,7 @@ class Program_weight_tensor_parameter_30: class Program_weight_tensor_parameter_31: name = "parameter_31" + original_name = "conv2d_28.w_0" shape = [128, 512, 1, 1] dtype = "float32" min_val = float("-0.476772") @@ -350,6 +382,7 @@ class Program_weight_tensor_parameter_31: class Program_weight_tensor_parameter_32: name = "parameter_32" + original_name = "batch_norm2d_25.b_0" shape = [512] dtype = "float32" min_val = float("-3.24175") @@ -361,6 +394,7 @@ class Program_weight_tensor_parameter_32: class Program_weight_tensor_parameter_33: name = "parameter_33" + original_name = "batch_norm2d_25.w_0" shape = [512] dtype = "float32" min_val = float("0.174115") @@ -372,6 +406,7 @@ class Program_weight_tensor_parameter_33: class Program_weight_tensor_parameter_34: name = "parameter_34" + original_name = "batch_norm2d_25.w_2" shape = [512] dtype = "float32" min_val = float("1.86581e-05") @@ -383,6 +418,7 @@ class Program_weight_tensor_parameter_34: class Program_weight_tensor_parameter_35: name = "parameter_35" + original_name = "batch_norm2d_25.w_1" shape = [512] dtype = "float32" min_val = float("-1.29441") @@ -394,6 +430,7 @@ class Program_weight_tensor_parameter_35: class Program_weight_tensor_parameter_36: name = "parameter_36" + original_name = "conv2d_27.w_0" shape = [512, 1, 5, 5] dtype = "float32" min_val = float("-0.315521") @@ -405,6 +442,7 @@ class Program_weight_tensor_parameter_36: class Program_weight_tensor_parameter_37: name = "parameter_37" + original_name = "batch_norm2d_24.b_0" shape = [512] dtype = "float32" min_val = float("-2.72401") @@ -416,6 +454,7 @@ class Program_weight_tensor_parameter_37: class Program_weight_tensor_parameter_38: name = "parameter_38" + original_name = "batch_norm2d_24.w_0" shape = [512] dtype = "float32" min_val = float("-0.883896") @@ -427,6 +466,7 @@ class Program_weight_tensor_parameter_38: class Program_weight_tensor_parameter_39: name = "parameter_39" + original_name = "batch_norm2d_24.w_2" shape = [512] dtype = "float32" min_val = float("0.026373") @@ -438,6 +478,7 @@ class Program_weight_tensor_parameter_39: class Program_weight_tensor_parameter_40: name = "parameter_40" + original_name = "batch_norm2d_24.w_1" shape = [512] dtype = "float32" min_val = float("-0.926518") @@ -449,6 +490,7 @@ class Program_weight_tensor_parameter_40: class Program_weight_tensor_parameter_41: name = "parameter_41" + original_name = "conv2d_26.w_0" shape = [512, 256, 1, 1] dtype = "float32" min_val = float("-0.61167") @@ -460,6 +502,7 @@ class Program_weight_tensor_parameter_41: class Program_weight_tensor_parameter_42: name = "parameter_42" + original_name = "conv2d_25.b_0" shape = [256] dtype = "float32" min_val = float("-0.28404") @@ -471,6 +514,7 @@ class Program_weight_tensor_parameter_42: class Program_weight_tensor_parameter_43: name = "parameter_43" + original_name = "conv2d_25.w_0" shape = [256, 64, 1, 1] dtype = "float32" min_val = float("-0.911054") @@ -482,6 +526,7 @@ class Program_weight_tensor_parameter_43: class Program_weight_tensor_parameter_44: name = "parameter_44" + original_name = "conv2d_24.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -491,6 +536,7 @@ class Program_weight_tensor_parameter_44: class Program_weight_tensor_parameter_45: name = "parameter_45" + original_name = "conv2d_24.w_0" shape = [64, 256, 1, 1] dtype = "float32" min_val = float("-0.774262") @@ -502,6 +548,7 @@ class Program_weight_tensor_parameter_45: class Program_weight_tensor_parameter_46: name = "parameter_46" + original_name = "batch_norm2d_23.b_0" shape = [256] dtype = "float32" min_val = float("-1.77904") @@ -513,6 +560,7 @@ class Program_weight_tensor_parameter_46: class Program_weight_tensor_parameter_47: name = "parameter_47" + original_name = "batch_norm2d_23.w_0" shape = [256] dtype = "float32" min_val = float("0.359308") @@ -524,6 +572,7 @@ class Program_weight_tensor_parameter_47: class Program_weight_tensor_parameter_48: name = "parameter_48" + original_name = "batch_norm2d_23.w_2" shape = [256] dtype = "float32" min_val = float("8.36395e-05") @@ -535,6 +584,7 @@ class Program_weight_tensor_parameter_48: class Program_weight_tensor_parameter_49: name = "parameter_49" + original_name = "batch_norm2d_23.w_1" shape = [256] dtype = "float32" min_val = float("-7.01996") @@ -546,6 +596,7 @@ class Program_weight_tensor_parameter_49: class Program_weight_tensor_parameter_50: name = "parameter_50" + original_name = "conv2d_23.w_0" shape = [256, 1, 5, 5] dtype = "float32" min_val = float("-0.303061") @@ -557,6 +608,7 @@ class Program_weight_tensor_parameter_50: class Program_weight_tensor_parameter_51: name = "parameter_51" + original_name = "batch_norm2d_22.b_0" shape = [256] dtype = "float32" min_val = float("-3.23843") @@ -568,6 +620,7 @@ class Program_weight_tensor_parameter_51: class Program_weight_tensor_parameter_52: name = "parameter_52" + original_name = "batch_norm2d_22.w_0" shape = [256] dtype = "float32" min_val = float("-0.38494") @@ -579,6 +632,7 @@ class Program_weight_tensor_parameter_52: class Program_weight_tensor_parameter_53: name = "parameter_53" + original_name = "batch_norm2d_22.w_2" shape = [256] dtype = "float32" min_val = float("0.963713") @@ -590,6 +644,7 @@ class Program_weight_tensor_parameter_53: class Program_weight_tensor_parameter_54: name = "parameter_54" + original_name = "batch_norm2d_22.w_1" shape = [256] dtype = "float32" min_val = float("-5.3764") @@ -601,6 +656,7 @@ class Program_weight_tensor_parameter_54: class Program_weight_tensor_parameter_55: name = "parameter_55" + original_name = "conv2d_22.w_0" shape = [256, 256, 1, 1] dtype = "float32" min_val = float("-0.764621") @@ -612,6 +668,7 @@ class Program_weight_tensor_parameter_55: class Program_weight_tensor_parameter_56: name = "parameter_56" + original_name = "batch_norm2d_21.b_0" shape = [256] dtype = "float32" min_val = float("-4.37345") @@ -623,6 +680,7 @@ class Program_weight_tensor_parameter_56: class Program_weight_tensor_parameter_57: name = "parameter_57" + original_name = "batch_norm2d_21.w_0" shape = [256] dtype = "float32" min_val = float("0.528463") @@ -634,6 +692,7 @@ class Program_weight_tensor_parameter_57: class Program_weight_tensor_parameter_58: name = "parameter_58" + original_name = "batch_norm2d_21.w_2" shape = [256] dtype = "float32" min_val = float("9.47117e-05") @@ -645,6 +704,7 @@ class Program_weight_tensor_parameter_58: class Program_weight_tensor_parameter_59: name = "parameter_59" + original_name = "batch_norm2d_21.w_1" shape = [256] dtype = "float32" min_val = float("-1.15169") @@ -656,6 +716,7 @@ class Program_weight_tensor_parameter_59: class Program_weight_tensor_parameter_60: name = "parameter_60" + original_name = "conv2d_21.w_0" shape = [256, 1, 5, 5] dtype = "float32" min_val = float("-0.990117") @@ -667,6 +728,7 @@ class Program_weight_tensor_parameter_60: class Program_weight_tensor_parameter_61: name = "parameter_61" + original_name = "batch_norm2d_20.b_0" shape = [256] dtype = "float32" min_val = float("-2.87827") @@ -678,6 +740,7 @@ class Program_weight_tensor_parameter_61: class Program_weight_tensor_parameter_62: name = "parameter_62" + original_name = "batch_norm2d_20.w_0" shape = [256] dtype = "float32" min_val = float("0.0259372") @@ -689,6 +752,7 @@ class Program_weight_tensor_parameter_62: class Program_weight_tensor_parameter_63: name = "parameter_63" + original_name = "batch_norm2d_20.w_2" shape = [256] dtype = "float32" min_val = float("0.643314") @@ -700,6 +764,7 @@ class Program_weight_tensor_parameter_63: class Program_weight_tensor_parameter_64: name = "parameter_64" + original_name = "batch_norm2d_20.w_1" shape = [256] dtype = "float32" min_val = float("-4.6576") @@ -711,6 +776,7 @@ class Program_weight_tensor_parameter_64: class Program_weight_tensor_parameter_65: name = "parameter_65" + original_name = "conv2d_20.w_0" shape = [256, 256, 1, 1] dtype = "float32" min_val = float("-0.865708") @@ -722,6 +788,7 @@ class Program_weight_tensor_parameter_65: class Program_weight_tensor_parameter_66: name = "parameter_66" + original_name = "batch_norm2d_19.b_0" shape = [256] dtype = "float32" min_val = float("-4.33513") @@ -733,6 +800,7 @@ class Program_weight_tensor_parameter_66: class Program_weight_tensor_parameter_67: name = "parameter_67" + original_name = "batch_norm2d_19.w_0" shape = [256] dtype = "float32" min_val = float("0.429251") @@ -744,6 +812,7 @@ class Program_weight_tensor_parameter_67: class Program_weight_tensor_parameter_68: name = "parameter_68" + original_name = "batch_norm2d_19.w_2" shape = [256] dtype = "float32" min_val = float("5.80614e-05") @@ -755,6 +824,7 @@ class Program_weight_tensor_parameter_68: class Program_weight_tensor_parameter_69: name = "parameter_69" + original_name = "batch_norm2d_19.w_1" shape = [256] dtype = "float32" min_val = float("-10.6876") @@ -766,6 +836,7 @@ class Program_weight_tensor_parameter_69: class Program_weight_tensor_parameter_70: name = "parameter_70" + original_name = "conv2d_19.w_0" shape = [256, 1, 5, 5] dtype = "float32" min_val = float("-0.966461") @@ -777,6 +848,7 @@ class Program_weight_tensor_parameter_70: class Program_weight_tensor_parameter_71: name = "parameter_71" + original_name = "batch_norm2d_18.b_0" shape = [256] dtype = "float32" min_val = float("-4.22825") @@ -788,6 +860,7 @@ class Program_weight_tensor_parameter_71: class Program_weight_tensor_parameter_72: name = "parameter_72" + original_name = "batch_norm2d_18.w_0" shape = [256] dtype = "float32" min_val = float("-0.988813") @@ -799,6 +872,7 @@ class Program_weight_tensor_parameter_72: class Program_weight_tensor_parameter_73: name = "parameter_73" + original_name = "batch_norm2d_18.w_2" shape = [256] dtype = "float32" min_val = float("0.593346") @@ -810,6 +884,7 @@ class Program_weight_tensor_parameter_73: class Program_weight_tensor_parameter_74: name = "parameter_74" + original_name = "batch_norm2d_18.w_1" shape = [256] dtype = "float32" min_val = float("-6.62936") @@ -821,6 +896,7 @@ class Program_weight_tensor_parameter_74: class Program_weight_tensor_parameter_75: name = "parameter_75" + original_name = "conv2d_18.w_0" shape = [256, 256, 1, 1] dtype = "float32" min_val = float("-0.761926") @@ -832,6 +908,7 @@ class Program_weight_tensor_parameter_75: class Program_weight_tensor_parameter_76: name = "parameter_76" + original_name = "batch_norm2d_17.b_0" shape = [256] dtype = "float32" min_val = float("-1.74212") @@ -843,6 +920,7 @@ class Program_weight_tensor_parameter_76: class Program_weight_tensor_parameter_77: name = "parameter_77" + original_name = "batch_norm2d_17.w_0" shape = [256] dtype = "float32" min_val = float("0.455477") @@ -854,6 +932,7 @@ class Program_weight_tensor_parameter_77: class Program_weight_tensor_parameter_78: name = "parameter_78" + original_name = "batch_norm2d_17.w_2" shape = [256] dtype = "float32" min_val = float("1.29547e-05") @@ -865,6 +944,7 @@ class Program_weight_tensor_parameter_78: class Program_weight_tensor_parameter_79: name = "parameter_79" + original_name = "batch_norm2d_17.w_1" shape = [256] dtype = "float32" min_val = float("-2.45374") @@ -876,6 +956,7 @@ class Program_weight_tensor_parameter_79: class Program_weight_tensor_parameter_80: name = "parameter_80" + original_name = "conv2d_17.w_0" shape = [256, 1, 5, 5] dtype = "float32" min_val = float("-1.09005") @@ -887,6 +968,7 @@ class Program_weight_tensor_parameter_80: class Program_weight_tensor_parameter_81: name = "parameter_81" + original_name = "batch_norm2d_16.b_0" shape = [256] dtype = "float32" min_val = float("-3.95596") @@ -898,6 +980,7 @@ class Program_weight_tensor_parameter_81: class Program_weight_tensor_parameter_82: name = "parameter_82" + original_name = "batch_norm2d_16.w_0" shape = [256] dtype = "float32" min_val = float("-0.620373") @@ -909,6 +992,7 @@ class Program_weight_tensor_parameter_82: class Program_weight_tensor_parameter_83: name = "parameter_83" + original_name = "batch_norm2d_16.w_2" shape = [256] dtype = "float32" min_val = float("0.832338") @@ -920,6 +1004,7 @@ class Program_weight_tensor_parameter_83: class Program_weight_tensor_parameter_84: name = "parameter_84" + original_name = "batch_norm2d_16.w_1" shape = [256] dtype = "float32" min_val = float("-6.69774") @@ -931,6 +1016,7 @@ class Program_weight_tensor_parameter_84: class Program_weight_tensor_parameter_85: name = "parameter_85" + original_name = "conv2d_16.w_0" shape = [256, 256, 1, 1] dtype = "float32" min_val = float("-0.954536") @@ -942,6 +1028,7 @@ class Program_weight_tensor_parameter_85: class Program_weight_tensor_parameter_86: name = "parameter_86" + original_name = "batch_norm2d_15.b_0" shape = [256] dtype = "float32" min_val = float("-2.27024") @@ -953,6 +1040,7 @@ class Program_weight_tensor_parameter_86: class Program_weight_tensor_parameter_87: name = "parameter_87" + original_name = "batch_norm2d_15.w_0" shape = [256] dtype = "float32" min_val = float("0.533904") @@ -964,6 +1052,7 @@ class Program_weight_tensor_parameter_87: class Program_weight_tensor_parameter_88: name = "parameter_88" + original_name = "batch_norm2d_15.w_2" shape = [256] dtype = "float32" min_val = float("4.04275e-05") @@ -975,6 +1064,7 @@ class Program_weight_tensor_parameter_88: class Program_weight_tensor_parameter_89: name = "parameter_89" + original_name = "batch_norm2d_15.w_1" shape = [256] dtype = "float32" min_val = float("-3.09706") @@ -986,6 +1076,7 @@ class Program_weight_tensor_parameter_89: class Program_weight_tensor_parameter_90: name = "parameter_90" + original_name = "conv2d_15.w_0" shape = [256, 1, 5, 5] dtype = "float32" min_val = float("-0.910621") @@ -997,6 +1088,7 @@ class Program_weight_tensor_parameter_90: class Program_weight_tensor_parameter_91: name = "parameter_91" + original_name = "batch_norm2d_14.b_0" shape = [256] dtype = "float32" min_val = float("-2.24081") @@ -1008,6 +1100,7 @@ class Program_weight_tensor_parameter_91: class Program_weight_tensor_parameter_92: name = "parameter_92" + original_name = "batch_norm2d_14.w_0" shape = [256] dtype = "float32" min_val = float("-0.0698286") @@ -1019,6 +1112,7 @@ class Program_weight_tensor_parameter_92: class Program_weight_tensor_parameter_93: name = "parameter_93" + original_name = "batch_norm2d_14.w_2" shape = [256] dtype = "float32" min_val = float("0.859497") @@ -1030,6 +1124,7 @@ class Program_weight_tensor_parameter_93: class Program_weight_tensor_parameter_94: name = "parameter_94" + original_name = "batch_norm2d_14.w_1" shape = [256] dtype = "float32" min_val = float("-5.39083") @@ -1041,6 +1136,7 @@ class Program_weight_tensor_parameter_94: class Program_weight_tensor_parameter_95: name = "parameter_95" + original_name = "conv2d_14.w_0" shape = [256, 256, 1, 1] dtype = "float32" min_val = float("-1.38839") @@ -1052,6 +1148,7 @@ class Program_weight_tensor_parameter_95: class Program_weight_tensor_parameter_96: name = "parameter_96" + original_name = "batch_norm2d_13.b_0" shape = [256] dtype = "float32" min_val = float("-1.66791") @@ -1063,6 +1160,7 @@ class Program_weight_tensor_parameter_96: class Program_weight_tensor_parameter_97: name = "parameter_97" + original_name = "batch_norm2d_13.w_0" shape = [256] dtype = "float32" min_val = float("0.525063") @@ -1074,6 +1172,7 @@ class Program_weight_tensor_parameter_97: class Program_weight_tensor_parameter_98: name = "parameter_98" + original_name = "batch_norm2d_13.w_2" shape = [256] dtype = "float32" min_val = float("8.30418e-05") @@ -1085,6 +1184,7 @@ class Program_weight_tensor_parameter_98: class Program_weight_tensor_parameter_99: name = "parameter_99" + original_name = "batch_norm2d_13.w_1" shape = [256] dtype = "float32" min_val = float("-3.93807") @@ -1096,6 +1196,7 @@ class Program_weight_tensor_parameter_99: class Program_weight_tensor_parameter_100: name = "parameter_100" + original_name = "conv2d_13.w_0" shape = [256, 1, 5, 5] dtype = "float32" min_val = float("-1.07621") @@ -1107,6 +1208,7 @@ class Program_weight_tensor_parameter_100: class Program_weight_tensor_parameter_101: name = "parameter_101" + original_name = "batch_norm2d_12.b_0" shape = [256] dtype = "float32" min_val = float("-2.3307") @@ -1118,6 +1220,7 @@ class Program_weight_tensor_parameter_101: class Program_weight_tensor_parameter_102: name = "parameter_102" + original_name = "batch_norm2d_12.w_0" shape = [256] dtype = "float32" min_val = float("-0.0894802") @@ -1129,6 +1232,7 @@ class Program_weight_tensor_parameter_102: class Program_weight_tensor_parameter_103: name = "parameter_103" + original_name = "batch_norm2d_12.w_2" shape = [256] dtype = "float32" min_val = float("0.268346") @@ -1140,6 +1244,7 @@ class Program_weight_tensor_parameter_103: class Program_weight_tensor_parameter_104: name = "parameter_104" + original_name = "batch_norm2d_12.w_1" shape = [256] dtype = "float32" min_val = float("-10.2208") @@ -1151,6 +1256,7 @@ class Program_weight_tensor_parameter_104: class Program_weight_tensor_parameter_105: name = "parameter_105" + original_name = "conv2d_12.w_0" shape = [256, 128, 1, 1] dtype = "float32" min_val = float("-1.14056") @@ -1162,6 +1268,7 @@ class Program_weight_tensor_parameter_105: class Program_weight_tensor_parameter_106: name = "parameter_106" + original_name = "batch_norm2d_11.b_0" shape = [128] dtype = "float32" min_val = float("-1.17191") @@ -1173,6 +1280,7 @@ class Program_weight_tensor_parameter_106: class Program_weight_tensor_parameter_107: name = "parameter_107" + original_name = "batch_norm2d_11.w_0" shape = [128] dtype = "float32" min_val = float("0.416415") @@ -1184,6 +1292,7 @@ class Program_weight_tensor_parameter_107: class Program_weight_tensor_parameter_108: name = "parameter_108" + original_name = "batch_norm2d_11.w_2" shape = [128] dtype = "float32" min_val = float("4.23945e-05") @@ -1195,6 +1304,7 @@ class Program_weight_tensor_parameter_108: class Program_weight_tensor_parameter_109: name = "parameter_109" + original_name = "batch_norm2d_11.w_1" shape = [128] dtype = "float32" min_val = float("-3.19435") @@ -1206,6 +1316,7 @@ class Program_weight_tensor_parameter_109: class Program_weight_tensor_parameter_110: name = "parameter_110" + original_name = "conv2d_11.w_0" shape = [128, 1, 3, 3] dtype = "float32" min_val = float("-0.407171") @@ -1217,6 +1328,7 @@ class Program_weight_tensor_parameter_110: class Program_weight_tensor_parameter_111: name = "parameter_111" + original_name = "batch_norm2d_10.b_0" shape = [128] dtype = "float32" min_val = float("-1.98784") @@ -1228,6 +1340,7 @@ class Program_weight_tensor_parameter_111: class Program_weight_tensor_parameter_112: name = "parameter_112" + original_name = "batch_norm2d_10.w_0" shape = [128] dtype = "float32" min_val = float("-0.494363") @@ -1239,6 +1352,7 @@ class Program_weight_tensor_parameter_112: class Program_weight_tensor_parameter_113: name = "parameter_113" + original_name = "batch_norm2d_10.w_2" shape = [128] dtype = "float32" min_val = float("0.896345") @@ -1250,6 +1364,7 @@ class Program_weight_tensor_parameter_113: class Program_weight_tensor_parameter_114: name = "parameter_114" + original_name = "batch_norm2d_10.w_1" shape = [128] dtype = "float32" min_val = float("-9.10655") @@ -1261,6 +1376,7 @@ class Program_weight_tensor_parameter_114: class Program_weight_tensor_parameter_115: name = "parameter_115" + original_name = "conv2d_10.w_0" shape = [128, 128, 1, 1] dtype = "float32" min_val = float("-1.17162") @@ -1272,6 +1388,7 @@ class Program_weight_tensor_parameter_115: class Program_weight_tensor_parameter_116: name = "parameter_116" + original_name = "batch_norm2d_9.b_0" shape = [128] dtype = "float32" min_val = float("-2.79875") @@ -1283,6 +1400,7 @@ class Program_weight_tensor_parameter_116: class Program_weight_tensor_parameter_117: name = "parameter_117" + original_name = "batch_norm2d_9.w_0" shape = [128] dtype = "float32" min_val = float("0.768268") @@ -1294,6 +1412,7 @@ class Program_weight_tensor_parameter_117: class Program_weight_tensor_parameter_118: name = "parameter_118" + original_name = "batch_norm2d_9.w_2" shape = [128] dtype = "float32" min_val = float("3.49606e-05") @@ -1305,6 +1424,7 @@ class Program_weight_tensor_parameter_118: class Program_weight_tensor_parameter_119: name = "parameter_119" + original_name = "batch_norm2d_9.w_1" shape = [128] dtype = "float32" min_val = float("-5.99844") @@ -1316,6 +1436,7 @@ class Program_weight_tensor_parameter_119: class Program_weight_tensor_parameter_120: name = "parameter_120" + original_name = "conv2d_9.w_0" shape = [128, 1, 3, 3] dtype = "float32" min_val = float("-1.06449") @@ -1327,6 +1448,7 @@ class Program_weight_tensor_parameter_120: class Program_weight_tensor_parameter_121: name = "parameter_121" + original_name = "batch_norm2d_8.b_0" shape = [128] dtype = "float32" min_val = float("-1.48153") @@ -1338,6 +1460,7 @@ class Program_weight_tensor_parameter_121: class Program_weight_tensor_parameter_122: name = "parameter_122" + original_name = "batch_norm2d_8.w_0" shape = [128] dtype = "float32" min_val = float("-0.391881") @@ -1349,6 +1472,7 @@ class Program_weight_tensor_parameter_122: class Program_weight_tensor_parameter_123: name = "parameter_123" + original_name = "batch_norm2d_8.w_2" shape = [128] dtype = "float32" min_val = float("0.195711") @@ -1360,6 +1484,7 @@ class Program_weight_tensor_parameter_123: class Program_weight_tensor_parameter_124: name = "parameter_124" + original_name = "batch_norm2d_8.w_1" shape = [128] dtype = "float32" min_val = float("-9.83676") @@ -1371,6 +1496,7 @@ class Program_weight_tensor_parameter_124: class Program_weight_tensor_parameter_125: name = "parameter_125" + original_name = "conv2d_8.w_0" shape = [128, 64, 1, 1] dtype = "float32" min_val = float("-1.09441") @@ -1382,6 +1508,7 @@ class Program_weight_tensor_parameter_125: class Program_weight_tensor_parameter_126: name = "parameter_126" + original_name = "batch_norm2d_7.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -1391,6 +1518,7 @@ class Program_weight_tensor_parameter_126: class Program_weight_tensor_parameter_127: name = "parameter_127" + original_name = "batch_norm2d_7.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -1400,6 +1528,7 @@ class Program_weight_tensor_parameter_127: class Program_weight_tensor_parameter_128: name = "parameter_128" + original_name = "batch_norm2d_7.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -1409,6 +1538,7 @@ class Program_weight_tensor_parameter_128: class Program_weight_tensor_parameter_129: name = "parameter_129" + original_name = "batch_norm2d_7.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -1418,6 +1548,7 @@ class Program_weight_tensor_parameter_129: class Program_weight_tensor_parameter_130: name = "parameter_130" + original_name = "conv2d_7.w_0" shape = [64, 1, 3, 3] dtype = "float32" min_val = float("-0.382338") @@ -1429,6 +1560,7 @@ class Program_weight_tensor_parameter_130: class Program_weight_tensor_parameter_131: name = "parameter_131" + original_name = "batch_norm2d_6.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -1438,6 +1570,7 @@ class Program_weight_tensor_parameter_131: class Program_weight_tensor_parameter_132: name = "parameter_132" + original_name = "batch_norm2d_6.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -1447,6 +1580,7 @@ class Program_weight_tensor_parameter_132: class Program_weight_tensor_parameter_133: name = "parameter_133" + original_name = "batch_norm2d_6.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -1456,6 +1590,7 @@ class Program_weight_tensor_parameter_133: class Program_weight_tensor_parameter_134: name = "parameter_134" + original_name = "batch_norm2d_6.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -1465,6 +1600,7 @@ class Program_weight_tensor_parameter_134: class Program_weight_tensor_parameter_135: name = "parameter_135" + original_name = "conv2d_6.w_0" shape = [64, 64, 1, 1] dtype = "float32" min_val = float("-1.48215") @@ -1476,6 +1612,7 @@ class Program_weight_tensor_parameter_135: class Program_weight_tensor_parameter_136: name = "parameter_136" + original_name = "batch_norm2d_5.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -1485,6 +1622,7 @@ class Program_weight_tensor_parameter_136: class Program_weight_tensor_parameter_137: name = "parameter_137" + original_name = "batch_norm2d_5.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -1494,6 +1632,7 @@ class Program_weight_tensor_parameter_137: class Program_weight_tensor_parameter_138: name = "parameter_138" + original_name = "batch_norm2d_5.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -1503,6 +1642,7 @@ class Program_weight_tensor_parameter_138: class Program_weight_tensor_parameter_139: name = "parameter_139" + original_name = "batch_norm2d_5.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -1512,6 +1652,7 @@ class Program_weight_tensor_parameter_139: class Program_weight_tensor_parameter_140: name = "parameter_140" + original_name = "conv2d_5.w_0" shape = [64, 1, 3, 3] dtype = "float32" min_val = float("-1.25083") @@ -1523,6 +1664,7 @@ class Program_weight_tensor_parameter_140: class Program_weight_tensor_parameter_141: name = "parameter_141" + original_name = "batch_norm2d_4.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -1532,6 +1674,7 @@ class Program_weight_tensor_parameter_141: class Program_weight_tensor_parameter_142: name = "parameter_142" + original_name = "batch_norm2d_4.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -1541,6 +1684,7 @@ class Program_weight_tensor_parameter_142: class Program_weight_tensor_parameter_143: name = "parameter_143" + original_name = "batch_norm2d_4.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -1550,6 +1694,7 @@ class Program_weight_tensor_parameter_143: class Program_weight_tensor_parameter_144: name = "parameter_144" + original_name = "batch_norm2d_4.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -1559,6 +1704,7 @@ class Program_weight_tensor_parameter_144: class Program_weight_tensor_parameter_145: name = "parameter_145" + original_name = "conv2d_4.w_0" shape = [64, 32, 1, 1] dtype = "float32" min_val = float("-1.93171") @@ -1570,6 +1716,7 @@ class Program_weight_tensor_parameter_145: class Program_weight_tensor_parameter_146: name = "parameter_146" + original_name = "batch_norm2d_3.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -1579,6 +1726,7 @@ class Program_weight_tensor_parameter_146: class Program_weight_tensor_parameter_147: name = "parameter_147" + original_name = "batch_norm2d_3.w_0" shape = [32] dtype = "float32" min_val = float("0") @@ -1588,6 +1736,7 @@ class Program_weight_tensor_parameter_147: class Program_weight_tensor_parameter_148: name = "parameter_148" + original_name = "batch_norm2d_3.w_2" shape = [32] dtype = "float32" min_val = float("0") @@ -1597,6 +1746,7 @@ class Program_weight_tensor_parameter_148: class Program_weight_tensor_parameter_149: name = "parameter_149" + original_name = "batch_norm2d_3.w_1" shape = [32] dtype = "float32" min_val = float("0") @@ -1606,6 +1756,7 @@ class Program_weight_tensor_parameter_149: class Program_weight_tensor_parameter_150: name = "parameter_150" + original_name = "conv2d_3.w_0" shape = [32, 1, 3, 3] dtype = "float32" min_val = float("-0.640915") @@ -1617,6 +1768,7 @@ class Program_weight_tensor_parameter_150: class Program_weight_tensor_parameter_151: name = "parameter_151" + original_name = "batch_norm2d_2.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -1626,6 +1778,7 @@ class Program_weight_tensor_parameter_151: class Program_weight_tensor_parameter_152: name = "parameter_152" + original_name = "batch_norm2d_2.w_0" shape = [32] dtype = "float32" min_val = float("0") @@ -1635,6 +1788,7 @@ class Program_weight_tensor_parameter_152: class Program_weight_tensor_parameter_153: name = "parameter_153" + original_name = "batch_norm2d_2.w_2" shape = [32] dtype = "float32" min_val = float("0") @@ -1644,6 +1798,7 @@ class Program_weight_tensor_parameter_153: class Program_weight_tensor_parameter_154: name = "parameter_154" + original_name = "batch_norm2d_2.w_1" shape = [32] dtype = "float32" min_val = float("0") @@ -1653,6 +1808,7 @@ class Program_weight_tensor_parameter_154: class Program_weight_tensor_parameter_155: name = "parameter_155" + original_name = "conv2d_2.w_0" shape = [32, 16, 1, 1] dtype = "float32" min_val = float("-1.65439") @@ -1664,6 +1820,7 @@ class Program_weight_tensor_parameter_155: class Program_weight_tensor_parameter_156: name = "parameter_156" + original_name = "batch_norm2d_1.b_0" shape = [16] dtype = "float32" min_val = float("0") @@ -1673,6 +1830,7 @@ class Program_weight_tensor_parameter_156: class Program_weight_tensor_parameter_157: name = "parameter_157" + original_name = "batch_norm2d_1.w_0" shape = [16] dtype = "float32" min_val = float("0") @@ -1682,6 +1840,7 @@ class Program_weight_tensor_parameter_157: class Program_weight_tensor_parameter_158: name = "parameter_158" + original_name = "batch_norm2d_1.w_2" shape = [16] dtype = "float32" min_val = float("0") @@ -1691,6 +1850,7 @@ class Program_weight_tensor_parameter_158: class Program_weight_tensor_parameter_159: name = "parameter_159" + original_name = "batch_norm2d_1.w_1" shape = [16] dtype = "float32" min_val = float("0") @@ -1700,6 +1860,7 @@ class Program_weight_tensor_parameter_159: class Program_weight_tensor_parameter_160: name = "parameter_160" + original_name = "conv2d_1.w_0" shape = [16, 1, 3, 3] dtype = "float32" min_val = float("-2.75536") @@ -1711,6 +1872,7 @@ class Program_weight_tensor_parameter_160: class Program_weight_tensor_parameter_161: name = "parameter_161" + original_name = "batch_norm2d_0.b_0" shape = [16] dtype = "float32" min_val = float("0") @@ -1720,6 +1882,7 @@ class Program_weight_tensor_parameter_161: class Program_weight_tensor_parameter_162: name = "parameter_162" + original_name = "batch_norm2d_0.w_0" shape = [16] dtype = "float32" min_val = float("0") @@ -1729,6 +1892,7 @@ class Program_weight_tensor_parameter_162: class Program_weight_tensor_parameter_163: name = "parameter_163" + original_name = "batch_norm2d_0.w_2" shape = [16] dtype = "float32" min_val = float("0") @@ -1738,6 +1902,7 @@ class Program_weight_tensor_parameter_163: class Program_weight_tensor_parameter_164: name = "parameter_164" + original_name = "batch_norm2d_0.w_1" shape = [16] dtype = "float32" min_val = float("0") @@ -1747,6 +1912,7 @@ class Program_weight_tensor_parameter_164: class Program_weight_tensor_parameter_165: name = "parameter_165" + original_name = "conv2d_0.w_0" shape = [16, 3, 3, 3] dtype = "float32" min_val = float("-1.39179") diff --git a/paddle_samples/PaddleX/PP-LiteSeg-B/subgraph_5/input_meta.py b/paddle_samples/PaddleX/PP-LiteSeg-B/subgraph_5/input_meta.py index 1c2795ba3..89de0d4f7 100644 --- a/paddle_samples/PaddleX/PP-LiteSeg-B/subgraph_5/input_meta.py +++ b/paddle_samples/PaddleX/PP-LiteSeg-B/subgraph_5/input_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_data_0: name = "data_0" + original_name = "param_3" shape = [1] dtype = "float32" data = [1.0] @@ -7,6 +8,7 @@ class Program_weight_tensor_data_0: class Program_weight_tensor_data_1: name = "data_1" + original_name = "param_4" shape = [1] dtype = "float32" data = [1.0] @@ -14,6 +16,7 @@ class Program_weight_tensor_data_1: class Program_weight_tensor_data_2: name = "data_2" + original_name = "param_5" shape = [1] dtype = "float32" data = [1.0] @@ -21,6 +24,7 @@ class Program_weight_tensor_data_2: class Program_weight_tensor_data_3: name = "data_3" + original_name = "var_435" shape = [2, 256, 64, 128] dtype = "float32" max_val = float("9.51792") @@ -31,6 +35,7 @@ class Program_weight_tensor_data_3: class Program_weight_tensor_data_4: name = "data_4" + original_name = "var_436" shape = [2, 512, 32, 64] dtype = "float32" max_val = float("9.8545") @@ -41,6 +46,7 @@ class Program_weight_tensor_data_4: class Program_weight_tensor_data_5: name = "data_5" + original_name = "var_437" shape = [2, 1024, 16, 32] dtype = "float32" max_val = float("3.7237") diff --git a/paddle_samples/PaddleX/PP-LiteSeg-B/subgraph_5/weight_meta.py b/paddle_samples/PaddleX/PP-LiteSeg-B/subgraph_5/weight_meta.py index b1f1991cd..f65f62377 100644 --- a/paddle_samples/PaddleX/PP-LiteSeg-B/subgraph_5/weight_meta.py +++ b/paddle_samples/PaddleX/PP-LiteSeg-B/subgraph_5/weight_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_parameter_0: name = "parameter_0" + original_name = "batch_norm2d_58.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -9,6 +10,7 @@ class Program_weight_tensor_parameter_0: class Program_weight_tensor_parameter_1: name = "parameter_1" + original_name = "batch_norm2d_58.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -18,6 +20,7 @@ class Program_weight_tensor_parameter_1: class Program_weight_tensor_parameter_2: name = "parameter_2" + original_name = "batch_norm2d_58.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -27,6 +30,7 @@ class Program_weight_tensor_parameter_2: class Program_weight_tensor_parameter_3: name = "parameter_3" + original_name = "batch_norm2d_58.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -36,6 +40,7 @@ class Program_weight_tensor_parameter_3: class Program_weight_tensor_parameter_4: name = "parameter_4" + original_name = "conv2d_58.w_0" shape = [64, 96, 3, 3] dtype = "float32" min_val = float("-0.19692") @@ -47,6 +52,7 @@ class Program_weight_tensor_parameter_4: class Program_weight_tensor_parameter_5: name = "parameter_5" + original_name = "batch_norm2d_60.b_0" shape = [1] dtype = "float32" min_val = float("0") @@ -56,6 +62,7 @@ class Program_weight_tensor_parameter_5: class Program_weight_tensor_parameter_6: name = "parameter_6" + original_name = "batch_norm2d_60.w_0" shape = [1] dtype = "float32" min_val = float("0") @@ -65,6 +72,7 @@ class Program_weight_tensor_parameter_6: class Program_weight_tensor_parameter_7: name = "parameter_7" + original_name = "batch_norm2d_60.w_2" shape = [1] dtype = "float32" min_val = float("0") @@ -74,6 +82,7 @@ class Program_weight_tensor_parameter_7: class Program_weight_tensor_parameter_8: name = "parameter_8" + original_name = "batch_norm2d_60.w_1" shape = [1] dtype = "float32" min_val = float("0") @@ -83,6 +92,7 @@ class Program_weight_tensor_parameter_8: class Program_weight_tensor_parameter_9: name = "parameter_9" + original_name = "conv2d_60.w_0" shape = [1, 2, 3, 3] dtype = "float32" min_val = float("0") @@ -92,6 +102,7 @@ class Program_weight_tensor_parameter_9: class Program_weight_tensor_parameter_10: name = "parameter_10" + original_name = "batch_norm2d_59.b_0" shape = [2] dtype = "float32" min_val = float("0") @@ -101,6 +112,7 @@ class Program_weight_tensor_parameter_10: class Program_weight_tensor_parameter_11: name = "parameter_11" + original_name = "batch_norm2d_59.w_0" shape = [2] dtype = "float32" min_val = float("0") @@ -110,6 +122,7 @@ class Program_weight_tensor_parameter_11: class Program_weight_tensor_parameter_12: name = "parameter_12" + original_name = "batch_norm2d_59.w_2" shape = [2] dtype = "float32" min_val = float("0") @@ -119,6 +132,7 @@ class Program_weight_tensor_parameter_12: class Program_weight_tensor_parameter_13: name = "parameter_13" + original_name = "batch_norm2d_59.w_1" shape = [2] dtype = "float32" min_val = float("0") @@ -128,6 +142,7 @@ class Program_weight_tensor_parameter_13: class Program_weight_tensor_parameter_14: name = "parameter_14" + original_name = "conv2d_59.w_0" shape = [2, 4, 3, 3] dtype = "float32" min_val = float("-0.391965") @@ -139,6 +154,7 @@ class Program_weight_tensor_parameter_14: class Program_weight_tensor_parameter_15: name = "parameter_15" + original_name = "batch_norm2d_57.b_0" shape = [96] dtype = "float32" data = None @@ -146,6 +162,7 @@ class Program_weight_tensor_parameter_15: class Program_weight_tensor_parameter_16: name = "parameter_16" + original_name = "batch_norm2d_57.w_0" shape = [96] dtype = "float32" min_val = float("1.0") @@ -156,6 +173,7 @@ class Program_weight_tensor_parameter_16: class Program_weight_tensor_parameter_17: name = "parameter_17" + original_name = "batch_norm2d_57.w_2" shape = [96] dtype = "float32" min_val = float("1.0") @@ -166,6 +184,7 @@ class Program_weight_tensor_parameter_17: class Program_weight_tensor_parameter_18: name = "parameter_18" + original_name = "batch_norm2d_57.w_1" shape = [96] dtype = "float32" data = None @@ -173,6 +192,7 @@ class Program_weight_tensor_parameter_18: class Program_weight_tensor_parameter_19: name = "parameter_19" + original_name = "conv2d_57.w_0" shape = [96, 256, 3, 3] dtype = "float32" min_val = float("-0.135008") @@ -184,6 +204,7 @@ class Program_weight_tensor_parameter_19: class Program_weight_tensor_parameter_20: name = "parameter_20" + original_name = "batch_norm2d_62.b_0" shape = [96] dtype = "float32" data = None @@ -191,6 +212,7 @@ class Program_weight_tensor_parameter_20: class Program_weight_tensor_parameter_21: name = "parameter_21" + original_name = "batch_norm2d_62.w_0" shape = [96] dtype = "float32" min_val = float("1.0") @@ -201,6 +223,7 @@ class Program_weight_tensor_parameter_21: class Program_weight_tensor_parameter_22: name = "parameter_22" + original_name = "batch_norm2d_62.w_2" shape = [96] dtype = "float32" min_val = float("1.0") @@ -211,6 +234,7 @@ class Program_weight_tensor_parameter_22: class Program_weight_tensor_parameter_23: name = "parameter_23" + original_name = "batch_norm2d_62.w_1" shape = [96] dtype = "float32" data = None @@ -218,6 +242,7 @@ class Program_weight_tensor_parameter_23: class Program_weight_tensor_parameter_24: name = "parameter_24" + original_name = "conv2d_62.w_0" shape = [96, 128, 3, 3] dtype = "float32" min_val = float("-0.171958") @@ -229,6 +254,7 @@ class Program_weight_tensor_parameter_24: class Program_weight_tensor_parameter_25: name = "parameter_25" + original_name = "batch_norm2d_64.b_0" shape = [1] dtype = "float32" min_val = float("0") @@ -238,6 +264,7 @@ class Program_weight_tensor_parameter_25: class Program_weight_tensor_parameter_26: name = "parameter_26" + original_name = "batch_norm2d_64.w_0" shape = [1] dtype = "float32" min_val = float("0") @@ -247,6 +274,7 @@ class Program_weight_tensor_parameter_26: class Program_weight_tensor_parameter_27: name = "parameter_27" + original_name = "batch_norm2d_64.w_2" shape = [1] dtype = "float32" min_val = float("0") @@ -256,6 +284,7 @@ class Program_weight_tensor_parameter_27: class Program_weight_tensor_parameter_28: name = "parameter_28" + original_name = "batch_norm2d_64.w_1" shape = [1] dtype = "float32" min_val = float("0") @@ -265,6 +294,7 @@ class Program_weight_tensor_parameter_28: class Program_weight_tensor_parameter_29: name = "parameter_29" + original_name = "conv2d_64.w_0" shape = [1, 2, 3, 3] dtype = "float32" min_val = float("0") @@ -274,6 +304,7 @@ class Program_weight_tensor_parameter_29: class Program_weight_tensor_parameter_30: name = "parameter_30" + original_name = "batch_norm2d_63.b_0" shape = [2] dtype = "float32" min_val = float("0") @@ -283,6 +314,7 @@ class Program_weight_tensor_parameter_30: class Program_weight_tensor_parameter_31: name = "parameter_31" + original_name = "batch_norm2d_63.w_0" shape = [2] dtype = "float32" min_val = float("0") @@ -292,6 +324,7 @@ class Program_weight_tensor_parameter_31: class Program_weight_tensor_parameter_32: name = "parameter_32" + original_name = "batch_norm2d_63.w_2" shape = [2] dtype = "float32" min_val = float("0") @@ -301,6 +334,7 @@ class Program_weight_tensor_parameter_32: class Program_weight_tensor_parameter_33: name = "parameter_33" + original_name = "batch_norm2d_63.w_1" shape = [2] dtype = "float32" min_val = float("0") @@ -310,6 +344,7 @@ class Program_weight_tensor_parameter_33: class Program_weight_tensor_parameter_34: name = "parameter_34" + original_name = "conv2d_63.w_0" shape = [2, 4, 3, 3] dtype = "float32" min_val = float("-0.486806") @@ -321,6 +356,7 @@ class Program_weight_tensor_parameter_34: class Program_weight_tensor_parameter_35: name = "parameter_35" + original_name = "batch_norm2d_61.b_0" shape = [128] dtype = "float32" data = None @@ -328,6 +364,7 @@ class Program_weight_tensor_parameter_35: class Program_weight_tensor_parameter_36: name = "parameter_36" + original_name = "batch_norm2d_61.w_0" shape = [128] dtype = "float32" min_val = float("1.0") @@ -338,6 +375,7 @@ class Program_weight_tensor_parameter_36: class Program_weight_tensor_parameter_37: name = "parameter_37" + original_name = "batch_norm2d_61.w_2" shape = [128] dtype = "float32" min_val = float("1.0") @@ -348,6 +386,7 @@ class Program_weight_tensor_parameter_37: class Program_weight_tensor_parameter_38: name = "parameter_38" + original_name = "batch_norm2d_61.w_1" shape = [128] dtype = "float32" data = None @@ -355,6 +394,7 @@ class Program_weight_tensor_parameter_38: class Program_weight_tensor_parameter_39: name = "parameter_39" + original_name = "conv2d_61.w_0" shape = [128, 512, 3, 3] dtype = "float32" min_val = float("-0.100363") @@ -366,6 +406,7 @@ class Program_weight_tensor_parameter_39: class Program_weight_tensor_parameter_40: name = "parameter_40" + original_name = "batch_norm2d_66.b_0" shape = [128] dtype = "float32" data = None @@ -373,6 +414,7 @@ class Program_weight_tensor_parameter_40: class Program_weight_tensor_parameter_41: name = "parameter_41" + original_name = "batch_norm2d_66.w_0" shape = [128] dtype = "float32" min_val = float("1.0") @@ -383,6 +425,7 @@ class Program_weight_tensor_parameter_41: class Program_weight_tensor_parameter_42: name = "parameter_42" + original_name = "batch_norm2d_66.w_2" shape = [128] dtype = "float32" min_val = float("1.0") @@ -393,6 +436,7 @@ class Program_weight_tensor_parameter_42: class Program_weight_tensor_parameter_43: name = "parameter_43" + original_name = "batch_norm2d_66.w_1" shape = [128] dtype = "float32" data = None @@ -400,6 +444,7 @@ class Program_weight_tensor_parameter_43: class Program_weight_tensor_parameter_44: name = "parameter_44" + original_name = "conv2d_66.w_0" shape = [128, 128, 3, 3] dtype = "float32" min_val = float("-0.179717") @@ -411,6 +456,7 @@ class Program_weight_tensor_parameter_44: class Program_weight_tensor_parameter_45: name = "parameter_45" + original_name = "batch_norm2d_68.b_0" shape = [1] dtype = "float32" min_val = float("0") @@ -420,6 +466,7 @@ class Program_weight_tensor_parameter_45: class Program_weight_tensor_parameter_46: name = "parameter_46" + original_name = "batch_norm2d_68.w_0" shape = [1] dtype = "float32" min_val = float("0") @@ -429,6 +476,7 @@ class Program_weight_tensor_parameter_46: class Program_weight_tensor_parameter_47: name = "parameter_47" + original_name = "batch_norm2d_68.w_2" shape = [1] dtype = "float32" min_val = float("0") @@ -438,6 +486,7 @@ class Program_weight_tensor_parameter_47: class Program_weight_tensor_parameter_48: name = "parameter_48" + original_name = "batch_norm2d_68.w_1" shape = [1] dtype = "float32" min_val = float("0") @@ -447,6 +496,7 @@ class Program_weight_tensor_parameter_48: class Program_weight_tensor_parameter_49: name = "parameter_49" + original_name = "conv2d_68.w_0" shape = [1, 2, 3, 3] dtype = "float32" min_val = float("0") @@ -456,6 +506,7 @@ class Program_weight_tensor_parameter_49: class Program_weight_tensor_parameter_50: name = "parameter_50" + original_name = "batch_norm2d_67.b_0" shape = [2] dtype = "float32" min_val = float("0") @@ -465,6 +516,7 @@ class Program_weight_tensor_parameter_50: class Program_weight_tensor_parameter_51: name = "parameter_51" + original_name = "batch_norm2d_67.w_0" shape = [2] dtype = "float32" min_val = float("0") @@ -474,6 +526,7 @@ class Program_weight_tensor_parameter_51: class Program_weight_tensor_parameter_52: name = "parameter_52" + original_name = "batch_norm2d_67.w_2" shape = [2] dtype = "float32" min_val = float("0") @@ -483,6 +536,7 @@ class Program_weight_tensor_parameter_52: class Program_weight_tensor_parameter_53: name = "parameter_53" + original_name = "batch_norm2d_67.w_1" shape = [2] dtype = "float32" min_val = float("0") @@ -492,6 +546,7 @@ class Program_weight_tensor_parameter_53: class Program_weight_tensor_parameter_54: name = "parameter_54" + original_name = "conv2d_67.w_0" shape = [2, 4, 3, 3] dtype = "float32" min_val = float("-0.453525") @@ -503,6 +558,7 @@ class Program_weight_tensor_parameter_54: class Program_weight_tensor_parameter_55: name = "parameter_55" + original_name = "batch_norm2d_65.b_0" shape = [128] dtype = "float32" data = None @@ -510,6 +566,7 @@ class Program_weight_tensor_parameter_55: class Program_weight_tensor_parameter_56: name = "parameter_56" + original_name = "batch_norm2d_65.w_0" shape = [128] dtype = "float32" min_val = float("1.0") @@ -520,6 +577,7 @@ class Program_weight_tensor_parameter_56: class Program_weight_tensor_parameter_57: name = "parameter_57" + original_name = "batch_norm2d_65.w_2" shape = [128] dtype = "float32" min_val = float("1.0") @@ -530,6 +588,7 @@ class Program_weight_tensor_parameter_57: class Program_weight_tensor_parameter_58: name = "parameter_58" + original_name = "batch_norm2d_65.w_1" shape = [128] dtype = "float32" data = None @@ -537,6 +596,7 @@ class Program_weight_tensor_parameter_58: class Program_weight_tensor_parameter_59: name = "parameter_59" + original_name = "conv2d_65.w_0" shape = [128, 1024, 3, 3] dtype = "float32" min_val = float("-0.0689184") @@ -548,6 +608,7 @@ class Program_weight_tensor_parameter_59: class Program_weight_tensor_parameter_60: name = "parameter_60" + original_name = "batch_norm2d_56.b_0" shape = [128] dtype = "float32" data = None @@ -555,6 +616,7 @@ class Program_weight_tensor_parameter_60: class Program_weight_tensor_parameter_61: name = "parameter_61" + original_name = "batch_norm2d_56.w_0" shape = [128] dtype = "float32" min_val = float("1.0") @@ -565,6 +627,7 @@ class Program_weight_tensor_parameter_61: class Program_weight_tensor_parameter_62: name = "parameter_62" + original_name = "batch_norm2d_56.w_2" shape = [128] dtype = "float32" min_val = float("1.0") @@ -575,6 +638,7 @@ class Program_weight_tensor_parameter_62: class Program_weight_tensor_parameter_63: name = "parameter_63" + original_name = "batch_norm2d_56.w_1" shape = [128] dtype = "float32" data = None @@ -582,6 +646,7 @@ class Program_weight_tensor_parameter_63: class Program_weight_tensor_parameter_64: name = "parameter_64" + original_name = "conv2d_56.b_0" shape = [128] dtype = "float32" data = None @@ -589,6 +654,7 @@ class Program_weight_tensor_parameter_64: class Program_weight_tensor_parameter_65: name = "parameter_65" + original_name = "conv2d_56.w_0" shape = [128, 128, 3, 3] dtype = "float32" min_val = float("-0.194398") @@ -600,6 +666,7 @@ class Program_weight_tensor_parameter_65: class Program_weight_tensor_parameter_66: name = "parameter_66" + original_name = "batch_norm2d_55.b_0" shape = [128] dtype = "float32" data = None @@ -607,6 +674,7 @@ class Program_weight_tensor_parameter_66: class Program_weight_tensor_parameter_67: name = "parameter_67" + original_name = "batch_norm2d_55.w_0" shape = [128] dtype = "float32" min_val = float("1.0") @@ -617,6 +685,7 @@ class Program_weight_tensor_parameter_67: class Program_weight_tensor_parameter_68: name = "parameter_68" + original_name = "batch_norm2d_55.w_2" shape = [128] dtype = "float32" min_val = float("1.0") @@ -627,6 +696,7 @@ class Program_weight_tensor_parameter_68: class Program_weight_tensor_parameter_69: name = "parameter_69" + original_name = "batch_norm2d_55.w_1" shape = [128] dtype = "float32" data = None @@ -634,6 +704,7 @@ class Program_weight_tensor_parameter_69: class Program_weight_tensor_parameter_70: name = "parameter_70" + original_name = "conv2d_55.b_0" shape = [128] dtype = "float32" data = None @@ -641,6 +712,7 @@ class Program_weight_tensor_parameter_70: class Program_weight_tensor_parameter_71: name = "parameter_71" + original_name = "conv2d_55.w_0" shape = [128, 1024, 1, 1] dtype = "float32" min_val = float("-0.182235") @@ -652,6 +724,7 @@ class Program_weight_tensor_parameter_71: class Program_weight_tensor_parameter_72: name = "parameter_72" + original_name = "batch_norm2d_54.b_0" shape = [128] dtype = "float32" data = None @@ -659,6 +732,7 @@ class Program_weight_tensor_parameter_72: class Program_weight_tensor_parameter_73: name = "parameter_73" + original_name = "batch_norm2d_54.w_0" shape = [128] dtype = "float32" min_val = float("1.0") @@ -669,6 +743,7 @@ class Program_weight_tensor_parameter_73: class Program_weight_tensor_parameter_74: name = "parameter_74" + original_name = "batch_norm2d_54.w_2" shape = [128] dtype = "float32" min_val = float("1.0") @@ -679,6 +754,7 @@ class Program_weight_tensor_parameter_74: class Program_weight_tensor_parameter_75: name = "parameter_75" + original_name = "batch_norm2d_54.w_1" shape = [128] dtype = "float32" data = None @@ -686,6 +762,7 @@ class Program_weight_tensor_parameter_75: class Program_weight_tensor_parameter_76: name = "parameter_76" + original_name = "conv2d_54.b_0" shape = [128] dtype = "float32" data = None @@ -693,6 +770,7 @@ class Program_weight_tensor_parameter_76: class Program_weight_tensor_parameter_77: name = "parameter_77" + original_name = "conv2d_54.w_0" shape = [128, 1024, 1, 1] dtype = "float32" min_val = float("-0.211263") @@ -704,6 +782,7 @@ class Program_weight_tensor_parameter_77: class Program_weight_tensor_parameter_78: name = "parameter_78" + original_name = "batch_norm2d_53.b_0" shape = [128] dtype = "float32" data = None @@ -711,6 +790,7 @@ class Program_weight_tensor_parameter_78: class Program_weight_tensor_parameter_79: name = "parameter_79" + original_name = "batch_norm2d_53.w_0" shape = [128] dtype = "float32" min_val = float("1.0") @@ -721,6 +801,7 @@ class Program_weight_tensor_parameter_79: class Program_weight_tensor_parameter_80: name = "parameter_80" + original_name = "batch_norm2d_53.w_2" shape = [128] dtype = "float32" min_val = float("1.0") @@ -731,6 +812,7 @@ class Program_weight_tensor_parameter_80: class Program_weight_tensor_parameter_81: name = "parameter_81" + original_name = "batch_norm2d_53.w_1" shape = [128] dtype = "float32" data = None @@ -738,6 +820,7 @@ class Program_weight_tensor_parameter_81: class Program_weight_tensor_parameter_82: name = "parameter_82" + original_name = "conv2d_53.b_0" shape = [128] dtype = "float32" data = None @@ -745,6 +828,7 @@ class Program_weight_tensor_parameter_82: class Program_weight_tensor_parameter_83: name = "parameter_83" + original_name = "conv2d_53.w_0" shape = [128, 1024, 1, 1] dtype = "float32" min_val = float("-0.20468") diff --git a/paddle_samples/PaddleX/PP-OCRv4_mobile_seal_det/input_meta.py b/paddle_samples/PaddleX/PP-OCRv4_mobile_seal_det/input_meta.py index f3891fde1..69536f63e 100644 --- a/paddle_samples/PaddleX/PP-OCRv4_mobile_seal_det/input_meta.py +++ b/paddle_samples/PaddleX/PP-OCRv4_mobile_seal_det/input_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_data_0: name = "data_0" + original_name = "param_104" shape = [1] dtype = "float32" data = [1.07226] @@ -7,6 +8,7 @@ class Program_weight_tensor_data_0: class Program_weight_tensor_data_1: name = "data_1" + original_name = "param_105" shape = [1] dtype = "float32" data = [0.510215] @@ -14,6 +16,7 @@ class Program_weight_tensor_data_1: class Program_weight_tensor_data_2: name = "data_2" + original_name = "param_106" shape = [1] dtype = "float32" data = [0.0700188] @@ -21,6 +24,7 @@ class Program_weight_tensor_data_2: class Program_weight_tensor_data_3: name = "data_3" + original_name = "param_107" shape = [1] dtype = "float32" data = [0.00173821] @@ -28,6 +32,7 @@ class Program_weight_tensor_data_3: class Program_weight_tensor_data_4: name = "data_4" + original_name = "param_108" shape = [1] dtype = "float32" data = [1.90512] @@ -35,6 +40,7 @@ class Program_weight_tensor_data_4: class Program_weight_tensor_data_5: name = "data_5" + original_name = "param_109" shape = [1] dtype = "float32" data = [0.77264] @@ -42,6 +48,7 @@ class Program_weight_tensor_data_5: class Program_weight_tensor_data_6: name = "data_6" + original_name = "param_110" shape = [1] dtype = "float32" data = [0.649909] @@ -49,6 +56,7 @@ class Program_weight_tensor_data_6: class Program_weight_tensor_data_7: name = "data_7" + original_name = "param_111" shape = [1] dtype = "float32" data = [-0.523152] @@ -56,6 +64,7 @@ class Program_weight_tensor_data_7: class Program_weight_tensor_data_8: name = "data_8" + original_name = "param_112" shape = [1] dtype = "float32" data = [0.0715766] @@ -63,6 +72,7 @@ class Program_weight_tensor_data_8: class Program_weight_tensor_data_9: name = "data_9" + original_name = "param_113" shape = [1] dtype = "float32" data = [-0.000209593] @@ -70,6 +80,7 @@ class Program_weight_tensor_data_9: class Program_weight_tensor_data_10: name = "data_10" + original_name = "param_114" shape = [1] dtype = "float32" data = [0.98377] @@ -77,6 +88,7 @@ class Program_weight_tensor_data_10: class Program_weight_tensor_data_11: name = "data_11" + original_name = "param_115" shape = [1] dtype = "float32" data = [1.77638] @@ -84,6 +96,7 @@ class Program_weight_tensor_data_11: class Program_weight_tensor_data_12: name = "data_12" + original_name = "param_116" shape = [1] dtype = "float32" data = [0.437085] @@ -91,6 +104,7 @@ class Program_weight_tensor_data_12: class Program_weight_tensor_data_13: name = "data_13" + original_name = "param_117" shape = [1] dtype = "float32" data = [-0.844384] @@ -98,6 +112,7 @@ class Program_weight_tensor_data_13: class Program_weight_tensor_data_14: name = "data_14" + original_name = "param_118" shape = [1] dtype = "float32" data = [2.64405] @@ -105,6 +120,7 @@ class Program_weight_tensor_data_14: class Program_weight_tensor_data_15: name = "data_15" + original_name = "param_119" shape = [1] dtype = "float32" data = [0.410898] @@ -112,6 +128,7 @@ class Program_weight_tensor_data_15: class Program_weight_tensor_data_16: name = "data_16" + original_name = "param_120" shape = [1] dtype = "float32" data = [0.0685827] @@ -119,6 +136,7 @@ class Program_weight_tensor_data_16: class Program_weight_tensor_data_17: name = "data_17" + original_name = "param_121" shape = [1] dtype = "float32" data = [0.000203539] @@ -126,6 +144,7 @@ class Program_weight_tensor_data_17: class Program_weight_tensor_data_18: name = "data_18" + original_name = "param_122" shape = [1] dtype = "float32" data = [1.2546] @@ -133,6 +152,7 @@ class Program_weight_tensor_data_18: class Program_weight_tensor_data_19: name = "data_19" + original_name = "param_123" shape = [1] dtype = "float32" data = [0.900154] @@ -140,6 +160,7 @@ class Program_weight_tensor_data_19: class Program_weight_tensor_data_20: name = "data_20" + original_name = "param_124" shape = [1] dtype = "float32" data = [0.875604] @@ -147,6 +168,7 @@ class Program_weight_tensor_data_20: class Program_weight_tensor_data_21: name = "data_21" + original_name = "param_125" shape = [1] dtype = "float32" data = [-0.246236] @@ -154,6 +176,7 @@ class Program_weight_tensor_data_21: class Program_weight_tensor_data_22: name = "data_22" + original_name = "param_126" shape = [1] dtype = "float32" data = [0.0863976] @@ -161,6 +184,7 @@ class Program_weight_tensor_data_22: class Program_weight_tensor_data_23: name = "data_23" + original_name = "param_127" shape = [1] dtype = "float32" data = [0.000341887] @@ -168,6 +192,7 @@ class Program_weight_tensor_data_23: class Program_weight_tensor_data_24: name = "data_24" + original_name = "param_128" shape = [1] dtype = "float32" data = [0.639156] @@ -175,6 +200,7 @@ class Program_weight_tensor_data_24: class Program_weight_tensor_data_25: name = "data_25" + original_name = "param_129" shape = [1] dtype = "float32" data = [1.95213] @@ -182,6 +208,7 @@ class Program_weight_tensor_data_25: class Program_weight_tensor_data_26: name = "data_26" + original_name = "param_130" shape = [1] dtype = "float32" data = [0.567057] @@ -189,6 +216,7 @@ class Program_weight_tensor_data_26: class Program_weight_tensor_data_27: name = "data_27" + original_name = "param_131" shape = [1] dtype = "float32" data = [-0.975948] @@ -196,6 +224,7 @@ class Program_weight_tensor_data_27: class Program_weight_tensor_data_28: name = "data_28" + original_name = "param_132" shape = [1] dtype = "float32" data = [1.4956] @@ -203,6 +232,7 @@ class Program_weight_tensor_data_28: class Program_weight_tensor_data_29: name = "data_29" + original_name = "param_133" shape = [1] dtype = "float32" data = [0.563418] @@ -210,6 +240,7 @@ class Program_weight_tensor_data_29: class Program_weight_tensor_data_30: name = "data_30" + original_name = "param_134" shape = [1] dtype = "float32" data = [0.140527] @@ -217,6 +248,7 @@ class Program_weight_tensor_data_30: class Program_weight_tensor_data_31: name = "data_31" + original_name = "param_135" shape = [1] dtype = "float32" data = [0.000263839] @@ -224,6 +256,7 @@ class Program_weight_tensor_data_31: class Program_weight_tensor_data_32: name = "data_32" + original_name = "param_136" shape = [1] dtype = "float32" data = [0.809868] @@ -231,6 +264,7 @@ class Program_weight_tensor_data_32: class Program_weight_tensor_data_33: name = "data_33" + original_name = "param_137" shape = [1] dtype = "float32" data = [-0.0863734] @@ -238,6 +272,7 @@ class Program_weight_tensor_data_33: class Program_weight_tensor_data_34: name = "data_34" + original_name = "param_138" shape = [1] dtype = "float32" data = [0.843443] @@ -245,6 +280,7 @@ class Program_weight_tensor_data_34: class Program_weight_tensor_data_35: name = "data_35" + original_name = "param_139" shape = [1] dtype = "float32" data = [0.22632] @@ -252,6 +288,7 @@ class Program_weight_tensor_data_35: class Program_weight_tensor_data_36: name = "data_36" + original_name = "param_140" shape = [1] dtype = "float32" data = [0.120963] @@ -259,6 +296,7 @@ class Program_weight_tensor_data_36: class Program_weight_tensor_data_37: name = "data_37" + original_name = "param_141" shape = [1] dtype = "float32" data = [0.000562135] @@ -266,6 +304,7 @@ class Program_weight_tensor_data_37: class Program_weight_tensor_data_38: name = "data_38" + original_name = "param_142" shape = [1] dtype = "float32" data = [0.796834] @@ -273,6 +312,7 @@ class Program_weight_tensor_data_38: class Program_weight_tensor_data_39: name = "data_39" + original_name = "param_143" shape = [1] dtype = "float32" data = [1.35969] @@ -280,6 +320,7 @@ class Program_weight_tensor_data_39: class Program_weight_tensor_data_40: name = "data_40" + original_name = "param_144" shape = [1] dtype = "float32" data = [0.784661] @@ -287,6 +328,7 @@ class Program_weight_tensor_data_40: class Program_weight_tensor_data_41: name = "data_41" + original_name = "param_145" shape = [1] dtype = "float32" data = [-0.712838] @@ -294,6 +336,7 @@ class Program_weight_tensor_data_41: class Program_weight_tensor_data_42: name = "data_42" + original_name = "param_146" shape = [1] dtype = "float32" data = [1.36056] @@ -301,6 +344,7 @@ class Program_weight_tensor_data_42: class Program_weight_tensor_data_43: name = "data_43" + original_name = "param_147" shape = [1] dtype = "float32" data = [0.455125] @@ -308,6 +352,7 @@ class Program_weight_tensor_data_43: class Program_weight_tensor_data_44: name = "data_44" + original_name = "param_148" shape = [1] dtype = "float32" data = [0.090437] @@ -315,6 +360,7 @@ class Program_weight_tensor_data_44: class Program_weight_tensor_data_45: name = "data_45" + original_name = "param_149" shape = [1] dtype = "float32" data = [-0.000373527] @@ -322,6 +368,7 @@ class Program_weight_tensor_data_45: class Program_weight_tensor_data_46: name = "data_46" + original_name = "param_150" shape = [1] dtype = "float32" data = [1.08924] @@ -329,6 +376,7 @@ class Program_weight_tensor_data_46: class Program_weight_tensor_data_47: name = "data_47" + original_name = "param_151" shape = [1] dtype = "float32" data = [0.486609] @@ -336,6 +384,7 @@ class Program_weight_tensor_data_47: class Program_weight_tensor_data_48: name = "data_48" + original_name = "param_152" shape = [1] dtype = "float32" data = [0.896556] @@ -343,6 +392,7 @@ class Program_weight_tensor_data_48: class Program_weight_tensor_data_49: name = "data_49" + original_name = "param_153" shape = [1] dtype = "float32" data = [-0.141479] @@ -350,6 +400,7 @@ class Program_weight_tensor_data_49: class Program_weight_tensor_data_50: name = "data_50" + original_name = "param_154" shape = [1] dtype = "float32" data = [1.31986] @@ -357,6 +408,7 @@ class Program_weight_tensor_data_50: class Program_weight_tensor_data_51: name = "data_51" + original_name = "param_155" shape = [1] dtype = "float32" data = [1.0159] @@ -364,6 +416,7 @@ class Program_weight_tensor_data_51: class Program_weight_tensor_data_52: name = "data_52" + original_name = "param_156" shape = [1] dtype = "float32" data = [0.0977035] @@ -371,6 +424,7 @@ class Program_weight_tensor_data_52: class Program_weight_tensor_data_53: name = "data_53" + original_name = "param_157" shape = [1] dtype = "float32" data = [0.00030766] @@ -378,6 +432,7 @@ class Program_weight_tensor_data_53: class Program_weight_tensor_data_54: name = "data_54" + original_name = "param_158" shape = [1] dtype = "float32" data = [0.989657] @@ -385,6 +440,7 @@ class Program_weight_tensor_data_54: class Program_weight_tensor_data_55: name = "data_55" + original_name = "param_159" shape = [1] dtype = "float32" data = [0.0787528] @@ -392,6 +448,7 @@ class Program_weight_tensor_data_55: class Program_weight_tensor_data_56: name = "data_56" + original_name = "param_160" shape = [1] dtype = "float32" data = [0.911533] @@ -399,6 +456,7 @@ class Program_weight_tensor_data_56: class Program_weight_tensor_data_57: name = "data_57" + original_name = "param_161" shape = [1] dtype = "float32" data = [0.0661269] @@ -406,6 +464,7 @@ class Program_weight_tensor_data_57: class Program_weight_tensor_data_58: name = "data_58" + original_name = "param_162" shape = [1] dtype = "float32" data = [1.20921] @@ -413,6 +472,7 @@ class Program_weight_tensor_data_58: class Program_weight_tensor_data_59: name = "data_59" + original_name = "param_163" shape = [1] dtype = "float32" data = [1.00555] @@ -420,6 +480,7 @@ class Program_weight_tensor_data_59: class Program_weight_tensor_data_60: name = "data_60" + original_name = "param_164" shape = [1] dtype = "float32" data = [0.107168] @@ -427,6 +488,7 @@ class Program_weight_tensor_data_60: class Program_weight_tensor_data_61: name = "data_61" + original_name = "param_165" shape = [1] dtype = "float32" data = [-0.000856931] @@ -434,6 +496,7 @@ class Program_weight_tensor_data_61: class Program_weight_tensor_data_62: name = "data_62" + original_name = "param_166" shape = [1] dtype = "float32" data = [1.0968] @@ -441,6 +504,7 @@ class Program_weight_tensor_data_62: class Program_weight_tensor_data_63: name = "data_63" + original_name = "param_167" shape = [1] dtype = "float32" data = [-0.110075] @@ -448,6 +512,7 @@ class Program_weight_tensor_data_63: class Program_weight_tensor_data_64: name = "data_64" + original_name = "param_168" shape = [1] dtype = "float32" data = [0.891515] @@ -455,6 +520,7 @@ class Program_weight_tensor_data_64: class Program_weight_tensor_data_65: name = "data_65" + original_name = "param_169" shape = [1] dtype = "float32" data = [0.185332] @@ -462,6 +528,7 @@ class Program_weight_tensor_data_65: class Program_weight_tensor_data_66: name = "data_66" + original_name = "param_170" shape = [1] dtype = "float32" data = [1.15191] @@ -469,6 +536,7 @@ class Program_weight_tensor_data_66: class Program_weight_tensor_data_67: name = "data_67" + original_name = "param_171" shape = [1] dtype = "float32" data = [1.23246] @@ -476,6 +544,7 @@ class Program_weight_tensor_data_67: class Program_weight_tensor_data_68: name = "data_68" + original_name = "param_172" shape = [1] dtype = "float32" data = [0.19435] @@ -483,6 +552,7 @@ class Program_weight_tensor_data_68: class Program_weight_tensor_data_69: name = "data_69" + original_name = "param_173" shape = [1] dtype = "float32" data = [-0.000450665] @@ -490,6 +560,7 @@ class Program_weight_tensor_data_69: class Program_weight_tensor_data_70: name = "data_70" + original_name = "param_174" shape = [1] dtype = "float32" data = [0.674214] @@ -497,6 +568,7 @@ class Program_weight_tensor_data_70: class Program_weight_tensor_data_71: name = "data_71" + original_name = "param_175" shape = [1] dtype = "float32" data = [-0.900158] @@ -504,6 +576,7 @@ class Program_weight_tensor_data_71: class Program_weight_tensor_data_72: name = "data_72" + original_name = "param_176" shape = [1] dtype = "float32" data = [0.971572] @@ -511,6 +584,7 @@ class Program_weight_tensor_data_72: class Program_weight_tensor_data_73: name = "data_73" + original_name = "param_177" shape = [1] dtype = "float32" data = [0.224685] @@ -518,6 +592,7 @@ class Program_weight_tensor_data_73: class Program_weight_tensor_data_74: name = "data_74" + original_name = "param_178" shape = [1] dtype = "float32" data = [0.596819] @@ -525,6 +600,7 @@ class Program_weight_tensor_data_74: class Program_weight_tensor_data_75: name = "data_75" + original_name = "param_179" shape = [1] dtype = "float32" data = [-0.0953855] @@ -532,6 +608,7 @@ class Program_weight_tensor_data_75: class Program_weight_tensor_data_76: name = "data_76" + original_name = "param_180" shape = [1] dtype = "float32" data = [1.2003] @@ -539,6 +616,7 @@ class Program_weight_tensor_data_76: class Program_weight_tensor_data_77: name = "data_77" + original_name = "param_181" shape = [1] dtype = "float32" data = [-0.0407014] @@ -546,6 +624,7 @@ class Program_weight_tensor_data_77: class Program_weight_tensor_data_78: name = "data_78" + original_name = "param_182" shape = [1] dtype = "float32" data = [1.07966] @@ -553,6 +632,7 @@ class Program_weight_tensor_data_78: class Program_weight_tensor_data_79: name = "data_79" + original_name = "param_183" shape = [1] dtype = "float32" data = [0.15161] @@ -560,6 +640,7 @@ class Program_weight_tensor_data_79: class Program_weight_tensor_data_80: name = "data_80" + original_name = "param_184" shape = [1] dtype = "float32" data = [1.40233] @@ -567,6 +648,7 @@ class Program_weight_tensor_data_80: class Program_weight_tensor_data_81: name = "data_81" + original_name = "param_185" shape = [1] dtype = "float32" data = [0.920292] @@ -574,6 +656,7 @@ class Program_weight_tensor_data_81: class Program_weight_tensor_data_82: name = "data_82" + original_name = "param_186" shape = [1] dtype = "float32" data = [0.984614] @@ -581,6 +664,7 @@ class Program_weight_tensor_data_82: class Program_weight_tensor_data_83: name = "data_83" + original_name = "param_187" shape = [1] dtype = "float32" data = [-0.914484] @@ -588,6 +672,7 @@ class Program_weight_tensor_data_83: class Program_weight_tensor_data_84: name = "data_84" + original_name = "param_188" shape = [1] dtype = "float32" data = [0.942324] @@ -595,6 +680,7 @@ class Program_weight_tensor_data_84: class Program_weight_tensor_data_85: name = "data_85" + original_name = "param_189" shape = [1] dtype = "float32" data = [-0.600987] @@ -602,6 +688,7 @@ class Program_weight_tensor_data_85: class Program_weight_tensor_data_86: name = "data_86" + original_name = "param_190" shape = [1] dtype = "float32" data = [1.02852] @@ -609,6 +696,7 @@ class Program_weight_tensor_data_86: class Program_weight_tensor_data_87: name = "data_87" + original_name = "param_191" shape = [1] dtype = "float32" data = [0.271914] @@ -616,6 +704,7 @@ class Program_weight_tensor_data_87: class Program_weight_tensor_data_88: name = "data_88" + original_name = "param_192" shape = [1] dtype = "float32" data = [1.34126] @@ -623,6 +712,7 @@ class Program_weight_tensor_data_88: class Program_weight_tensor_data_89: name = "data_89" + original_name = "param_193" shape = [1] dtype = "float32" data = [0.831191] @@ -630,6 +720,7 @@ class Program_weight_tensor_data_89: class Program_weight_tensor_data_90: name = "data_90" + original_name = "param_194" shape = [1] dtype = "float32" data = [0.120061] @@ -637,6 +728,7 @@ class Program_weight_tensor_data_90: class Program_weight_tensor_data_91: name = "data_91" + original_name = "param_195" shape = [1] dtype = "float32" data = [-0.000653583] @@ -644,6 +736,7 @@ class Program_weight_tensor_data_91: class Program_weight_tensor_data_92: name = "data_92" + original_name = "param_196" shape = [1] dtype = "float32" data = [0.895926] @@ -651,6 +744,7 @@ class Program_weight_tensor_data_92: class Program_weight_tensor_data_93: name = "data_93" + original_name = "param_197" shape = [1] dtype = "float32" data = [-0.318049] @@ -658,6 +752,7 @@ class Program_weight_tensor_data_93: class Program_weight_tensor_data_94: name = "data_94" + original_name = "param_198" shape = [1] dtype = "float32" data = [1.09926] @@ -665,6 +760,7 @@ class Program_weight_tensor_data_94: class Program_weight_tensor_data_95: name = "data_95" + original_name = "param_199" shape = [1] dtype = "float32" data = [0.214898] @@ -672,6 +768,7 @@ class Program_weight_tensor_data_95: class Program_weight_tensor_data_96: name = "data_96" + original_name = "param_200" shape = [1] dtype = "float32" data = [1.33878] @@ -679,6 +776,7 @@ class Program_weight_tensor_data_96: class Program_weight_tensor_data_97: name = "data_97" + original_name = "param_201" shape = [1] dtype = "float32" data = [0.519713] @@ -686,6 +784,7 @@ class Program_weight_tensor_data_97: class Program_weight_tensor_data_98: name = "data_98" + original_name = "param_202" shape = [1] dtype = "float32" data = [0.188471] @@ -693,6 +792,7 @@ class Program_weight_tensor_data_98: class Program_weight_tensor_data_99: name = "data_99" + original_name = "param_203" shape = [1] dtype = "float32" data = [0.000376625] @@ -700,6 +800,7 @@ class Program_weight_tensor_data_99: class Program_weight_tensor_data_100: name = "data_100" + original_name = "param_204" shape = [1] dtype = "float32" data = [0.558112] @@ -707,6 +808,7 @@ class Program_weight_tensor_data_100: class Program_weight_tensor_data_101: name = "data_101" + original_name = "param_205" shape = [1] dtype = "float32" data = [-1.12189] @@ -714,6 +816,7 @@ class Program_weight_tensor_data_101: class Program_weight_tensor_data_102: name = "data_102" + original_name = "param_206" shape = [1] dtype = "float32" data = [5.14948] @@ -721,6 +824,7 @@ class Program_weight_tensor_data_102: class Program_weight_tensor_data_103: name = "data_103" + original_name = "param_207" shape = [1] dtype = "float32" data = [1.11505] @@ -728,6 +832,7 @@ class Program_weight_tensor_data_103: class Program_weight_tensor_data_104: name = "data_104" + original_name = "var_661" shape = [6, 3, 640, 640] dtype = "float32" min_val = float("-2.1179") diff --git a/paddle_samples/PaddleX/PP-OCRv4_mobile_seal_det/weight_meta.py b/paddle_samples/PaddleX/PP-OCRv4_mobile_seal_det/weight_meta.py index caa0dd401..69124e295 100644 --- a/paddle_samples/PaddleX/PP-OCRv4_mobile_seal_det/weight_meta.py +++ b/paddle_samples/PaddleX/PP-OCRv4_mobile_seal_det/weight_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_parameter_0: name = "parameter_0" + original_name = "conv2d_transpose_3.b_0" shape = [1] dtype = "float32" min_val = float("0") @@ -9,6 +10,7 @@ class Program_weight_tensor_parameter_0: class Program_weight_tensor_parameter_1: name = "parameter_1" + original_name = "conv2d_transpose_3.w_0" shape = [24, 1, 2, 2] dtype = "float32" min_val = float("-1.19684") @@ -20,6 +22,7 @@ class Program_weight_tensor_parameter_1: class Program_weight_tensor_parameter_2: name = "parameter_2" + original_name = "batch_norm_3.b_0" shape = [24] dtype = "float32" min_val = float("0") @@ -29,6 +32,7 @@ class Program_weight_tensor_parameter_2: class Program_weight_tensor_parameter_3: name = "parameter_3" + original_name = "batch_norm_3.w_0" shape = [24] dtype = "float32" min_val = float("0") @@ -38,6 +42,7 @@ class Program_weight_tensor_parameter_3: class Program_weight_tensor_parameter_4: name = "parameter_4" + original_name = "batch_norm_3.w_2" shape = [24] dtype = "float32" min_val = float("0") @@ -47,6 +52,7 @@ class Program_weight_tensor_parameter_4: class Program_weight_tensor_parameter_5: name = "parameter_5" + original_name = "batch_norm_3.w_1" shape = [24] dtype = "float32" min_val = float("0") @@ -56,6 +62,7 @@ class Program_weight_tensor_parameter_5: class Program_weight_tensor_parameter_6: name = "parameter_6" + original_name = "conv2d_transpose_2.b_0" shape = [24] dtype = "float32" min_val = float("0") @@ -65,6 +72,7 @@ class Program_weight_tensor_parameter_6: class Program_weight_tensor_parameter_7: name = "parameter_7" + original_name = "conv2d_transpose_2.w_0" shape = [24, 24, 2, 2] dtype = "float32" min_val = float("-0.696789") @@ -76,6 +84,7 @@ class Program_weight_tensor_parameter_7: class Program_weight_tensor_parameter_8: name = "parameter_8" + original_name = "batch_norm_2.b_0" shape = [24] dtype = "float32" min_val = float("0") @@ -85,6 +94,7 @@ class Program_weight_tensor_parameter_8: class Program_weight_tensor_parameter_9: name = "parameter_9" + original_name = "batch_norm_2.w_0" shape = [24] dtype = "float32" min_val = float("0") @@ -94,6 +104,7 @@ class Program_weight_tensor_parameter_9: class Program_weight_tensor_parameter_10: name = "parameter_10" + original_name = "batch_norm_2.w_2" shape = [24] dtype = "float32" min_val = float("0") @@ -103,6 +114,7 @@ class Program_weight_tensor_parameter_10: class Program_weight_tensor_parameter_11: name = "parameter_11" + original_name = "batch_norm_2.w_1" shape = [24] dtype = "float32" min_val = float("0") @@ -112,6 +124,7 @@ class Program_weight_tensor_parameter_11: class Program_weight_tensor_parameter_12: name = "parameter_12" + original_name = "conv2d_160.w_0" shape = [24, 96, 3, 3] dtype = "float32" min_val = float("-0.530696") @@ -123,6 +136,7 @@ class Program_weight_tensor_parameter_12: class Program_weight_tensor_parameter_13: name = "parameter_13" + original_name = "conv2d_transpose_1.b_0" shape = [1] dtype = "float32" min_val = float("0") @@ -132,6 +146,7 @@ class Program_weight_tensor_parameter_13: class Program_weight_tensor_parameter_14: name = "parameter_14" + original_name = "conv2d_transpose_1.w_0" shape = [24, 1, 2, 2] dtype = "float32" min_val = float("-2.53851") @@ -143,6 +158,7 @@ class Program_weight_tensor_parameter_14: class Program_weight_tensor_parameter_15: name = "parameter_15" + original_name = "batch_norm_1.b_0" shape = [24] dtype = "float32" min_val = float("0") @@ -152,6 +168,7 @@ class Program_weight_tensor_parameter_15: class Program_weight_tensor_parameter_16: name = "parameter_16" + original_name = "batch_norm_1.w_0" shape = [24] dtype = "float32" min_val = float("0") @@ -161,6 +178,7 @@ class Program_weight_tensor_parameter_16: class Program_weight_tensor_parameter_17: name = "parameter_17" + original_name = "batch_norm_1.w_2" shape = [24] dtype = "float32" min_val = float("0") @@ -170,6 +188,7 @@ class Program_weight_tensor_parameter_17: class Program_weight_tensor_parameter_18: name = "parameter_18" + original_name = "batch_norm_1.w_1" shape = [24] dtype = "float32" min_val = float("0") @@ -179,6 +198,7 @@ class Program_weight_tensor_parameter_18: class Program_weight_tensor_parameter_19: name = "parameter_19" + original_name = "conv2d_transpose_0.b_0" shape = [24] dtype = "float32" min_val = float("0") @@ -188,6 +208,7 @@ class Program_weight_tensor_parameter_19: class Program_weight_tensor_parameter_20: name = "parameter_20" + original_name = "conv2d_transpose_0.w_0" shape = [24, 24, 2, 2] dtype = "float32" min_val = float("-0.582271") @@ -199,6 +220,7 @@ class Program_weight_tensor_parameter_20: class Program_weight_tensor_parameter_21: name = "parameter_21" + original_name = "batch_norm_0.b_0" shape = [24] dtype = "float32" min_val = float("0") @@ -208,6 +230,7 @@ class Program_weight_tensor_parameter_21: class Program_weight_tensor_parameter_22: name = "parameter_22" + original_name = "batch_norm_0.w_0" shape = [24] dtype = "float32" min_val = float("0") @@ -217,6 +240,7 @@ class Program_weight_tensor_parameter_22: class Program_weight_tensor_parameter_23: name = "parameter_23" + original_name = "batch_norm_0.w_2" shape = [24] dtype = "float32" min_val = float("0") @@ -226,6 +250,7 @@ class Program_weight_tensor_parameter_23: class Program_weight_tensor_parameter_24: name = "parameter_24" + original_name = "batch_norm_0.w_1" shape = [24] dtype = "float32" min_val = float("0") @@ -235,6 +260,7 @@ class Program_weight_tensor_parameter_24: class Program_weight_tensor_parameter_25: name = "parameter_25" + original_name = "conv2d_159.w_0" shape = [24, 96, 3, 3] dtype = "float32" min_val = float("-0.421387") @@ -246,6 +272,7 @@ class Program_weight_tensor_parameter_25: class Program_weight_tensor_parameter_26: name = "parameter_26" + original_name = "conv2d_140.b_0" shape = [24] dtype = "float32" min_val = float("0") @@ -255,6 +282,7 @@ class Program_weight_tensor_parameter_26: class Program_weight_tensor_parameter_27: name = "parameter_27" + original_name = "conv2d_140.w_0" shape = [24, 6, 1, 1] dtype = "float32" min_val = float("-0.0377433") @@ -266,6 +294,7 @@ class Program_weight_tensor_parameter_27: class Program_weight_tensor_parameter_28: name = "parameter_28" + original_name = "conv2d_139.b_0" shape = [6] dtype = "float32" min_val = float("0") @@ -275,6 +304,7 @@ class Program_weight_tensor_parameter_28: class Program_weight_tensor_parameter_29: name = "parameter_29" + original_name = "conv2d_139.w_0" shape = [6, 24, 1, 1] dtype = "float32" min_val = float("-0.163102") @@ -286,6 +316,7 @@ class Program_weight_tensor_parameter_29: class Program_weight_tensor_parameter_30: name = "parameter_30" + original_name = "conv2d_138.w_0" shape = [24, 96, 3, 3] dtype = "float32" min_val = float("-0.31352") @@ -297,6 +328,7 @@ class Program_weight_tensor_parameter_30: class Program_weight_tensor_parameter_31: name = "parameter_31" + original_name = "conv2d_146.b_0" shape = [24] dtype = "float32" min_val = float("0") @@ -306,6 +338,7 @@ class Program_weight_tensor_parameter_31: class Program_weight_tensor_parameter_32: name = "parameter_32" + original_name = "conv2d_146.w_0" shape = [24, 6, 1, 1] dtype = "float32" min_val = float("-0.0728546") @@ -317,6 +350,7 @@ class Program_weight_tensor_parameter_32: class Program_weight_tensor_parameter_33: name = "parameter_33" + original_name = "conv2d_145.b_0" shape = [6] dtype = "float32" min_val = float("0") @@ -326,6 +360,7 @@ class Program_weight_tensor_parameter_33: class Program_weight_tensor_parameter_34: name = "parameter_34" + original_name = "conv2d_145.w_0" shape = [6, 24, 1, 1] dtype = "float32" min_val = float("-0.364883") @@ -337,6 +372,7 @@ class Program_weight_tensor_parameter_34: class Program_weight_tensor_parameter_35: name = "parameter_35" + original_name = "conv2d_144.w_0" shape = [24, 96, 3, 3] dtype = "float32" min_val = float("-0.378681") @@ -348,6 +384,7 @@ class Program_weight_tensor_parameter_35: class Program_weight_tensor_parameter_36: name = "parameter_36" + original_name = "conv2d_152.b_0" shape = [24] dtype = "float32" min_val = float("0") @@ -357,6 +394,7 @@ class Program_weight_tensor_parameter_36: class Program_weight_tensor_parameter_37: name = "parameter_37" + original_name = "conv2d_152.w_0" shape = [24, 6, 1, 1] dtype = "float32" min_val = float("-0.214365") @@ -368,6 +406,7 @@ class Program_weight_tensor_parameter_37: class Program_weight_tensor_parameter_38: name = "parameter_38" + original_name = "conv2d_151.b_0" shape = [6] dtype = "float32" min_val = float("0") @@ -377,6 +416,7 @@ class Program_weight_tensor_parameter_38: class Program_weight_tensor_parameter_39: name = "parameter_39" + original_name = "conv2d_151.w_0" shape = [6, 24, 1, 1] dtype = "float32" min_val = float("-0.524883") @@ -388,6 +428,7 @@ class Program_weight_tensor_parameter_39: class Program_weight_tensor_parameter_40: name = "parameter_40" + original_name = "conv2d_150.w_0" shape = [24, 96, 3, 3] dtype = "float32" min_val = float("-0.440288") @@ -399,6 +440,7 @@ class Program_weight_tensor_parameter_40: class Program_weight_tensor_parameter_41: name = "parameter_41" + original_name = "conv2d_158.b_0" shape = [24] dtype = "float32" min_val = float("0") @@ -408,6 +450,7 @@ class Program_weight_tensor_parameter_41: class Program_weight_tensor_parameter_42: name = "parameter_42" + original_name = "conv2d_158.w_0" shape = [24, 6, 1, 1] dtype = "float32" min_val = float("-0.258871") @@ -419,6 +462,7 @@ class Program_weight_tensor_parameter_42: class Program_weight_tensor_parameter_43: name = "parameter_43" + original_name = "conv2d_157.b_0" shape = [6] dtype = "float32" min_val = float("0") @@ -428,6 +472,7 @@ class Program_weight_tensor_parameter_43: class Program_weight_tensor_parameter_44: name = "parameter_44" + original_name = "conv2d_157.w_0" shape = [6, 24, 1, 1] dtype = "float32" min_val = float("-0.464688") @@ -439,6 +484,7 @@ class Program_weight_tensor_parameter_44: class Program_weight_tensor_parameter_45: name = "parameter_45" + original_name = "conv2d_156.w_0" shape = [24, 96, 3, 3] dtype = "float32" min_val = float("-0.39125") @@ -450,6 +496,7 @@ class Program_weight_tensor_parameter_45: class Program_weight_tensor_parameter_46: name = "parameter_46" + original_name = "conv2d_137.b_0" shape = [96] dtype = "float32" min_val = float("-0.00360104") @@ -461,6 +508,7 @@ class Program_weight_tensor_parameter_46: class Program_weight_tensor_parameter_47: name = "parameter_47" + original_name = "conv2d_137.w_0" shape = [96, 24, 1, 1] dtype = "float32" min_val = float("-0.0694112") @@ -472,6 +520,7 @@ class Program_weight_tensor_parameter_47: class Program_weight_tensor_parameter_48: name = "parameter_48" + original_name = "conv2d_136.b_0" shape = [24] dtype = "float32" min_val = float("0") @@ -481,6 +530,7 @@ class Program_weight_tensor_parameter_48: class Program_weight_tensor_parameter_49: name = "parameter_49" + original_name = "conv2d_136.w_0" shape = [24, 96, 1, 1] dtype = "float32" min_val = float("-0.196071") @@ -492,6 +542,7 @@ class Program_weight_tensor_parameter_49: class Program_weight_tensor_parameter_50: name = "parameter_50" + original_name = "conv2d_135.w_0" shape = [96, 12, 1, 1] dtype = "float32" min_val = float("-0.438535") @@ -503,6 +554,7 @@ class Program_weight_tensor_parameter_50: class Program_weight_tensor_parameter_51: name = "parameter_51" + original_name = "conv2d_143.b_0" shape = [96] dtype = "float32" min_val = float("-0.00975305") @@ -514,6 +566,7 @@ class Program_weight_tensor_parameter_51: class Program_weight_tensor_parameter_52: name = "parameter_52" + original_name = "conv2d_143.w_0" shape = [96, 24, 1, 1] dtype = "float32" min_val = float("-0.515387") @@ -525,6 +578,7 @@ class Program_weight_tensor_parameter_52: class Program_weight_tensor_parameter_53: name = "parameter_53" + original_name = "conv2d_142.b_0" shape = [24] dtype = "float32" min_val = float("0") @@ -534,6 +588,7 @@ class Program_weight_tensor_parameter_53: class Program_weight_tensor_parameter_54: name = "parameter_54" + original_name = "conv2d_142.w_0" shape = [24, 96, 1, 1] dtype = "float32" min_val = float("-0.291381") @@ -545,6 +600,7 @@ class Program_weight_tensor_parameter_54: class Program_weight_tensor_parameter_55: name = "parameter_55" + original_name = "conv2d_141.w_0" shape = [96, 18, 1, 1] dtype = "float32" min_val = float("-0.60152") @@ -556,6 +612,7 @@ class Program_weight_tensor_parameter_55: class Program_weight_tensor_parameter_56: name = "parameter_56" + original_name = "conv2d_149.b_0" shape = [96] dtype = "float32" min_val = float("-0.100249") @@ -567,6 +624,7 @@ class Program_weight_tensor_parameter_56: class Program_weight_tensor_parameter_57: name = "parameter_57" + original_name = "conv2d_149.w_0" shape = [96, 24, 1, 1] dtype = "float32" min_val = float("-0.486452") @@ -578,6 +636,7 @@ class Program_weight_tensor_parameter_57: class Program_weight_tensor_parameter_58: name = "parameter_58" + original_name = "conv2d_148.b_0" shape = [24] dtype = "float32" min_val = float("0") @@ -587,6 +646,7 @@ class Program_weight_tensor_parameter_58: class Program_weight_tensor_parameter_59: name = "parameter_59" + original_name = "conv2d_148.w_0" shape = [24, 96, 1, 1] dtype = "float32" min_val = float("-0.769329") @@ -598,6 +658,7 @@ class Program_weight_tensor_parameter_59: class Program_weight_tensor_parameter_60: name = "parameter_60" + original_name = "conv2d_147.w_0" shape = [96, 42, 1, 1] dtype = "float32" min_val = float("-0.667384") @@ -609,6 +670,7 @@ class Program_weight_tensor_parameter_60: class Program_weight_tensor_parameter_61: name = "parameter_61" + original_name = "conv2d_155.b_0" shape = [96] dtype = "float32" min_val = float("-0.101202") @@ -620,6 +682,7 @@ class Program_weight_tensor_parameter_61: class Program_weight_tensor_parameter_62: name = "parameter_62" + original_name = "conv2d_155.w_0" shape = [96, 24, 1, 1] dtype = "float32" min_val = float("-0.949587") @@ -631,6 +694,7 @@ class Program_weight_tensor_parameter_62: class Program_weight_tensor_parameter_63: name = "parameter_63" + original_name = "conv2d_154.b_0" shape = [24] dtype = "float32" min_val = float("0") @@ -640,6 +704,7 @@ class Program_weight_tensor_parameter_63: class Program_weight_tensor_parameter_64: name = "parameter_64" + original_name = "conv2d_154.w_0" shape = [24, 96, 1, 1] dtype = "float32" min_val = float("-0.439529") @@ -651,6 +716,7 @@ class Program_weight_tensor_parameter_64: class Program_weight_tensor_parameter_65: name = "parameter_65" + original_name = "conv2d_153.w_0" shape = [96, 360, 1, 1] dtype = "float32" min_val = float("-0.368024") @@ -662,6 +728,7 @@ class Program_weight_tensor_parameter_65: class Program_weight_tensor_parameter_66: name = "parameter_66" + original_name = "conv2d_134.b_0" shape = [360] dtype = "float32" min_val = float("-0.257234") @@ -673,6 +740,7 @@ class Program_weight_tensor_parameter_66: class Program_weight_tensor_parameter_67: name = "parameter_67" + original_name = "conv2d_134.w_0" shape = [360, 384, 1, 1] dtype = "float32" min_val = float("-0.504584") @@ -684,6 +752,7 @@ class Program_weight_tensor_parameter_67: class Program_weight_tensor_parameter_68: name = "parameter_68" + original_name = "conv2d_133.b_0" shape = [42] dtype = "float32" min_val = float("0") @@ -693,6 +762,7 @@ class Program_weight_tensor_parameter_68: class Program_weight_tensor_parameter_69: name = "parameter_69" + original_name = "conv2d_133.w_0" shape = [42, 192, 1, 1] dtype = "float32" min_val = float("-0.61724") @@ -704,6 +774,7 @@ class Program_weight_tensor_parameter_69: class Program_weight_tensor_parameter_70: name = "parameter_70" + original_name = "conv2d_132.b_0" shape = [18] dtype = "float32" min_val = float("0") @@ -713,6 +784,7 @@ class Program_weight_tensor_parameter_70: class Program_weight_tensor_parameter_71: name = "parameter_71" + original_name = "conv2d_132.w_0" shape = [18, 96, 1, 1] dtype = "float32" min_val = float("-0.5711") @@ -724,6 +796,7 @@ class Program_weight_tensor_parameter_71: class Program_weight_tensor_parameter_72: name = "parameter_72" + original_name = "conv2d_131.b_0" shape = [12] dtype = "float32" min_val = float("0") @@ -733,6 +806,7 @@ class Program_weight_tensor_parameter_72: class Program_weight_tensor_parameter_73: name = "parameter_73" + original_name = "conv2d_131.w_0" shape = [12, 48, 1, 1] dtype = "float32" min_val = float("-0.467618") @@ -744,6 +818,7 @@ class Program_weight_tensor_parameter_73: class Program_weight_tensor_parameter_74: name = "parameter_74" + original_name = "batch_norm2d_145.b_0" shape = [384] dtype = "float32" min_val = float("-0.410025") @@ -755,6 +830,7 @@ class Program_weight_tensor_parameter_74: class Program_weight_tensor_parameter_75: name = "parameter_75" + original_name = "batch_norm2d_145.w_0" shape = [384] dtype = "float32" min_val = float("0.253726") @@ -766,6 +842,7 @@ class Program_weight_tensor_parameter_75: class Program_weight_tensor_parameter_76: name = "parameter_76" + original_name = "batch_norm2d_145.w_2" shape = [384] dtype = "float32" min_val = float("0.028662") @@ -777,6 +854,7 @@ class Program_weight_tensor_parameter_76: class Program_weight_tensor_parameter_77: name = "parameter_77" + original_name = "batch_norm2d_145.w_1" shape = [384] dtype = "float32" min_val = float("-0.93198") @@ -788,6 +866,7 @@ class Program_weight_tensor_parameter_77: class Program_weight_tensor_parameter_78: name = "parameter_78" + original_name = "conv2d_130.w_0" shape = [384, 384, 1, 1] dtype = "float32" min_val = float("-0.727177") @@ -799,6 +878,7 @@ class Program_weight_tensor_parameter_78: class Program_weight_tensor_parameter_79: name = "parameter_79" + original_name = "batch_norm2d_144.b_0" shape = [384] dtype = "float32" min_val = float("-0.410024") @@ -810,6 +890,7 @@ class Program_weight_tensor_parameter_79: class Program_weight_tensor_parameter_80: name = "parameter_80" + original_name = "batch_norm2d_144.w_0" shape = [384] dtype = "float32" min_val = float("0.214943") @@ -821,6 +902,7 @@ class Program_weight_tensor_parameter_80: class Program_weight_tensor_parameter_81: name = "parameter_81" + original_name = "batch_norm2d_144.w_2" shape = [384] dtype = "float32" min_val = float("0.0348518") @@ -832,6 +914,7 @@ class Program_weight_tensor_parameter_81: class Program_weight_tensor_parameter_82: name = "parameter_82" + original_name = "batch_norm2d_144.w_1" shape = [384] dtype = "float32" min_val = float("-0.95969") @@ -843,6 +926,7 @@ class Program_weight_tensor_parameter_82: class Program_weight_tensor_parameter_83: name = "parameter_83" + original_name = "conv2d_129.w_0" shape = [384, 384, 1, 1] dtype = "float32" min_val = float("-0.716284") @@ -854,6 +938,7 @@ class Program_weight_tensor_parameter_83: class Program_weight_tensor_parameter_84: name = "parameter_84" + original_name = "batch_norm2d_143.b_0" shape = [384] dtype = "float32" min_val = float("-0.410025") @@ -865,6 +950,7 @@ class Program_weight_tensor_parameter_84: class Program_weight_tensor_parameter_85: name = "parameter_85" + original_name = "batch_norm2d_143.w_0" shape = [384] dtype = "float32" min_val = float("0.243453") @@ -876,6 +962,7 @@ class Program_weight_tensor_parameter_85: class Program_weight_tensor_parameter_86: name = "parameter_86" + original_name = "batch_norm2d_143.w_2" shape = [384] dtype = "float32" min_val = float("0.0288336") @@ -887,6 +974,7 @@ class Program_weight_tensor_parameter_86: class Program_weight_tensor_parameter_87: name = "parameter_87" + original_name = "batch_norm2d_143.w_1" shape = [384] dtype = "float32" min_val = float("-0.906602") @@ -898,6 +986,7 @@ class Program_weight_tensor_parameter_87: class Program_weight_tensor_parameter_88: name = "parameter_88" + original_name = "conv2d_128.w_0" shape = [384, 384, 1, 1] dtype = "float32" min_val = float("-0.709712") @@ -909,6 +998,7 @@ class Program_weight_tensor_parameter_88: class Program_weight_tensor_parameter_89: name = "parameter_89" + original_name = "batch_norm2d_142.b_0" shape = [384] dtype = "float32" min_val = float("-0.410025") @@ -920,6 +1010,7 @@ class Program_weight_tensor_parameter_89: class Program_weight_tensor_parameter_90: name = "parameter_90" + original_name = "batch_norm2d_142.w_0" shape = [384] dtype = "float32" min_val = float("0.257591") @@ -931,6 +1022,7 @@ class Program_weight_tensor_parameter_90: class Program_weight_tensor_parameter_91: name = "parameter_91" + original_name = "batch_norm2d_142.w_2" shape = [384] dtype = "float32" min_val = float("0.0296982") @@ -942,6 +1034,7 @@ class Program_weight_tensor_parameter_91: class Program_weight_tensor_parameter_92: name = "parameter_92" + original_name = "batch_norm2d_142.w_1" shape = [384] dtype = "float32" min_val = float("-0.956832") @@ -953,6 +1046,7 @@ class Program_weight_tensor_parameter_92: class Program_weight_tensor_parameter_93: name = "parameter_93" + original_name = "conv2d_127.w_0" shape = [384, 384, 1, 1] dtype = "float32" min_val = float("-0.681882") @@ -964,6 +1058,7 @@ class Program_weight_tensor_parameter_93: class Program_weight_tensor_parameter_94: name = "parameter_94" + original_name = "batch_norm2d_141.b_0" shape = [384] dtype = "float32" min_val = float("-0.384184") @@ -975,6 +1070,7 @@ class Program_weight_tensor_parameter_94: class Program_weight_tensor_parameter_95: name = "parameter_95" + original_name = "batch_norm2d_141.w_0" shape = [384] dtype = "float32" min_val = float("-0.0480784") @@ -986,6 +1082,7 @@ class Program_weight_tensor_parameter_95: class Program_weight_tensor_parameter_96: name = "parameter_96" + original_name = "batch_norm2d_141.w_2" shape = [384] dtype = "float32" min_val = float("0.00407949") @@ -997,6 +1094,7 @@ class Program_weight_tensor_parameter_96: class Program_weight_tensor_parameter_97: name = "parameter_97" + original_name = "batch_norm2d_141.w_1" shape = [384] dtype = "float32" min_val = float("-0.0371368") @@ -1008,6 +1106,7 @@ class Program_weight_tensor_parameter_97: class Program_weight_tensor_parameter_98: name = "parameter_98" + original_name = "batch_norm2d_139.b_0" shape = [384] dtype = "float32" min_val = float("-0.315969") @@ -1019,6 +1118,7 @@ class Program_weight_tensor_parameter_98: class Program_weight_tensor_parameter_99: name = "parameter_99" + original_name = "batch_norm2d_139.w_0" shape = [384] dtype = "float32" min_val = float("-0.0521014") @@ -1030,6 +1130,7 @@ class Program_weight_tensor_parameter_99: class Program_weight_tensor_parameter_100: name = "parameter_100" + original_name = "batch_norm2d_139.w_2" shape = [384] dtype = "float32" min_val = float("0.000254626") @@ -1041,6 +1142,7 @@ class Program_weight_tensor_parameter_100: class Program_weight_tensor_parameter_101: name = "parameter_101" + original_name = "batch_norm2d_139.w_1" shape = [384] dtype = "float32" min_val = float("-1.32444") @@ -1052,6 +1154,7 @@ class Program_weight_tensor_parameter_101: class Program_weight_tensor_parameter_102: name = "parameter_102" + original_name = "conv2d_125.w_0" shape = [384, 1, 5, 5] dtype = "float32" min_val = float("-0.411714") @@ -1063,6 +1166,7 @@ class Program_weight_tensor_parameter_102: class Program_weight_tensor_parameter_103: name = "parameter_103" + original_name = "batch_norm2d_138.b_0" shape = [384] dtype = "float32" min_val = float("-0.315969") @@ -1074,6 +1178,7 @@ class Program_weight_tensor_parameter_103: class Program_weight_tensor_parameter_104: name = "parameter_104" + original_name = "batch_norm2d_138.w_0" shape = [384] dtype = "float32" min_val = float("-0.18291") @@ -1085,6 +1190,7 @@ class Program_weight_tensor_parameter_104: class Program_weight_tensor_parameter_105: name = "parameter_105" + original_name = "batch_norm2d_138.w_2" shape = [384] dtype = "float32" min_val = float("0.000262387") @@ -1096,6 +1202,7 @@ class Program_weight_tensor_parameter_105: class Program_weight_tensor_parameter_106: name = "parameter_106" + original_name = "batch_norm2d_138.w_1" shape = [384] dtype = "float32" min_val = float("-1.32584") @@ -1107,6 +1214,7 @@ class Program_weight_tensor_parameter_106: class Program_weight_tensor_parameter_107: name = "parameter_107" + original_name = "conv2d_124.w_0" shape = [384, 1, 5, 5] dtype = "float32" min_val = float("-0.434561") @@ -1118,6 +1226,7 @@ class Program_weight_tensor_parameter_107: class Program_weight_tensor_parameter_108: name = "parameter_108" + original_name = "batch_norm2d_137.b_0" shape = [384] dtype = "float32" min_val = float("-0.315969") @@ -1129,6 +1238,7 @@ class Program_weight_tensor_parameter_108: class Program_weight_tensor_parameter_109: name = "parameter_109" + original_name = "batch_norm2d_137.w_0" shape = [384] dtype = "float32" min_val = float("-0.00909981") @@ -1140,6 +1250,7 @@ class Program_weight_tensor_parameter_109: class Program_weight_tensor_parameter_110: name = "parameter_110" + original_name = "batch_norm2d_137.w_2" shape = [384] dtype = "float32" min_val = float("0.000254556") @@ -1151,6 +1262,7 @@ class Program_weight_tensor_parameter_110: class Program_weight_tensor_parameter_111: name = "parameter_111" + original_name = "batch_norm2d_137.w_1" shape = [384] dtype = "float32" min_val = float("-1.32461") @@ -1162,6 +1274,7 @@ class Program_weight_tensor_parameter_111: class Program_weight_tensor_parameter_112: name = "parameter_112" + original_name = "conv2d_123.w_0" shape = [384, 1, 5, 5] dtype = "float32" min_val = float("-0.414958") @@ -1173,6 +1286,7 @@ class Program_weight_tensor_parameter_112: class Program_weight_tensor_parameter_113: name = "parameter_113" + original_name = "batch_norm2d_136.b_0" shape = [384] dtype = "float32" min_val = float("-0.315969") @@ -1184,6 +1298,7 @@ class Program_weight_tensor_parameter_113: class Program_weight_tensor_parameter_114: name = "parameter_114" + original_name = "batch_norm2d_136.w_0" shape = [384] dtype = "float32" min_val = float("-0.00412843") @@ -1195,6 +1310,7 @@ class Program_weight_tensor_parameter_114: class Program_weight_tensor_parameter_115: name = "parameter_115" + original_name = "batch_norm2d_136.w_2" shape = [384] dtype = "float32" min_val = float("0.000262601") @@ -1206,6 +1322,7 @@ class Program_weight_tensor_parameter_115: class Program_weight_tensor_parameter_116: name = "parameter_116" + original_name = "batch_norm2d_136.w_1" shape = [384] dtype = "float32" min_val = float("-1.3302") @@ -1217,6 +1334,7 @@ class Program_weight_tensor_parameter_116: class Program_weight_tensor_parameter_117: name = "parameter_117" + original_name = "conv2d_122.w_0" shape = [384, 1, 5, 5] dtype = "float32" min_val = float("-0.403001") @@ -1228,6 +1346,7 @@ class Program_weight_tensor_parameter_117: class Program_weight_tensor_parameter_118: name = "parameter_118" + original_name = "batch_norm2d_140.b_0" shape = [384] dtype = "float32" min_val = float("-0.315969") @@ -1239,6 +1358,7 @@ class Program_weight_tensor_parameter_118: class Program_weight_tensor_parameter_119: name = "parameter_119" + original_name = "batch_norm2d_140.w_0" shape = [384] dtype = "float32" min_val = float("-0.224045") @@ -1250,6 +1370,7 @@ class Program_weight_tensor_parameter_119: class Program_weight_tensor_parameter_120: name = "parameter_120" + original_name = "batch_norm2d_140.w_2" shape = [384] dtype = "float32" min_val = float("5.5736e-11") @@ -1261,6 +1382,7 @@ class Program_weight_tensor_parameter_120: class Program_weight_tensor_parameter_121: name = "parameter_121" + original_name = "batch_norm2d_140.w_1" shape = [384] dtype = "float32" min_val = float("-0.0550974") @@ -1272,6 +1394,7 @@ class Program_weight_tensor_parameter_121: class Program_weight_tensor_parameter_122: name = "parameter_122" + original_name = "conv2d_126.w_0" shape = [384, 1, 1, 1] dtype = "float32" min_val = float("-0.170961") @@ -1283,6 +1406,7 @@ class Program_weight_tensor_parameter_122: class Program_weight_tensor_parameter_123: name = "parameter_123" + original_name = "batch_norm2d_135.b_0" shape = [384] dtype = "float32" min_val = float("-0.27779") @@ -1294,6 +1418,7 @@ class Program_weight_tensor_parameter_123: class Program_weight_tensor_parameter_124: name = "parameter_124" + original_name = "batch_norm2d_135.w_0" shape = [384] dtype = "float32" min_val = float("-0.159442") @@ -1305,6 +1430,7 @@ class Program_weight_tensor_parameter_124: class Program_weight_tensor_parameter_125: name = "parameter_125" + original_name = "batch_norm2d_135.w_2" shape = [384] dtype = "float32" min_val = float("0.00143262") @@ -1316,6 +1442,7 @@ class Program_weight_tensor_parameter_125: class Program_weight_tensor_parameter_126: name = "parameter_126" + original_name = "batch_norm2d_135.w_1" shape = [384] dtype = "float32" min_val = float("-0.184301") @@ -1327,6 +1454,7 @@ class Program_weight_tensor_parameter_126: class Program_weight_tensor_parameter_127: name = "parameter_127" + original_name = "batch_norm2d_134.b_0" shape = [384] dtype = "float32" min_val = float("-0.301801") @@ -1338,6 +1466,7 @@ class Program_weight_tensor_parameter_127: class Program_weight_tensor_parameter_128: name = "parameter_128" + original_name = "batch_norm2d_134.w_0" shape = [384] dtype = "float32" min_val = float("-0.00517258") @@ -1349,6 +1478,7 @@ class Program_weight_tensor_parameter_128: class Program_weight_tensor_parameter_129: name = "parameter_129" + original_name = "batch_norm2d_134.w_2" shape = [384] dtype = "float32" min_val = float("0.0142084") @@ -1360,6 +1490,7 @@ class Program_weight_tensor_parameter_129: class Program_weight_tensor_parameter_130: name = "parameter_130" + original_name = "batch_norm2d_134.w_1" shape = [384] dtype = "float32" min_val = float("-0.946833") @@ -1371,6 +1502,7 @@ class Program_weight_tensor_parameter_130: class Program_weight_tensor_parameter_131: name = "parameter_131" + original_name = "conv2d_121.w_0" shape = [384, 384, 1, 1] dtype = "float32" min_val = float("-0.755152") @@ -1382,6 +1514,7 @@ class Program_weight_tensor_parameter_131: class Program_weight_tensor_parameter_132: name = "parameter_132" + original_name = "batch_norm2d_133.b_0" shape = [384] dtype = "float32" min_val = float("-0.301801") @@ -1393,6 +1526,7 @@ class Program_weight_tensor_parameter_132: class Program_weight_tensor_parameter_133: name = "parameter_133" + original_name = "batch_norm2d_133.w_0" shape = [384] dtype = "float32" min_val = float("-0.00348972") @@ -1404,6 +1538,7 @@ class Program_weight_tensor_parameter_133: class Program_weight_tensor_parameter_134: name = "parameter_134" + original_name = "batch_norm2d_133.w_2" shape = [384] dtype = "float32" min_val = float("0.00636127") @@ -1415,6 +1550,7 @@ class Program_weight_tensor_parameter_134: class Program_weight_tensor_parameter_135: name = "parameter_135" + original_name = "batch_norm2d_133.w_1" shape = [384] dtype = "float32" min_val = float("-0.94308") @@ -1426,6 +1562,7 @@ class Program_weight_tensor_parameter_135: class Program_weight_tensor_parameter_136: name = "parameter_136" + original_name = "conv2d_120.w_0" shape = [384, 384, 1, 1] dtype = "float32" min_val = float("-0.762526") @@ -1437,6 +1574,7 @@ class Program_weight_tensor_parameter_136: class Program_weight_tensor_parameter_137: name = "parameter_137" + original_name = "batch_norm2d_132.b_0" shape = [384] dtype = "float32" min_val = float("-0.301801") @@ -1448,6 +1586,7 @@ class Program_weight_tensor_parameter_137: class Program_weight_tensor_parameter_138: name = "parameter_138" + original_name = "batch_norm2d_132.w_0" shape = [384] dtype = "float32" min_val = float("-0.00140934") @@ -1459,6 +1598,7 @@ class Program_weight_tensor_parameter_138: class Program_weight_tensor_parameter_139: name = "parameter_139" + original_name = "batch_norm2d_132.w_2" shape = [384] dtype = "float32" min_val = float("0.00538692") @@ -1470,6 +1610,7 @@ class Program_weight_tensor_parameter_139: class Program_weight_tensor_parameter_140: name = "parameter_140" + original_name = "batch_norm2d_132.w_1" shape = [384] dtype = "float32" min_val = float("-0.914784") @@ -1481,6 +1622,7 @@ class Program_weight_tensor_parameter_140: class Program_weight_tensor_parameter_141: name = "parameter_141" + original_name = "conv2d_119.w_0" shape = [384, 384, 1, 1] dtype = "float32" min_val = float("-0.621265") @@ -1492,6 +1634,7 @@ class Program_weight_tensor_parameter_141: class Program_weight_tensor_parameter_142: name = "parameter_142" + original_name = "batch_norm2d_131.b_0" shape = [384] dtype = "float32" min_val = float("-0.301801") @@ -1503,6 +1646,7 @@ class Program_weight_tensor_parameter_142: class Program_weight_tensor_parameter_143: name = "parameter_143" + original_name = "batch_norm2d_131.w_0" shape = [384] dtype = "float32" min_val = float("6.27642e-06") @@ -1514,6 +1658,7 @@ class Program_weight_tensor_parameter_143: class Program_weight_tensor_parameter_144: name = "parameter_144" + original_name = "batch_norm2d_131.w_2" shape = [384] dtype = "float32" min_val = float("0.00728078") @@ -1525,6 +1670,7 @@ class Program_weight_tensor_parameter_144: class Program_weight_tensor_parameter_145: name = "parameter_145" + original_name = "batch_norm2d_131.w_1" shape = [384] dtype = "float32" min_val = float("-0.949119") @@ -1536,6 +1682,7 @@ class Program_weight_tensor_parameter_145: class Program_weight_tensor_parameter_146: name = "parameter_146" + original_name = "conv2d_118.w_0" shape = [384, 384, 1, 1] dtype = "float32" min_val = float("-0.758421") @@ -1547,6 +1694,7 @@ class Program_weight_tensor_parameter_146: class Program_weight_tensor_parameter_147: name = "parameter_147" + original_name = "batch_norm2d_130.b_0" shape = [384] dtype = "float32" min_val = float("-0.256711") @@ -1558,6 +1706,7 @@ class Program_weight_tensor_parameter_147: class Program_weight_tensor_parameter_148: name = "parameter_148" + original_name = "batch_norm2d_130.w_0" shape = [384] dtype = "float32" min_val = float("-0.156985") @@ -1569,6 +1718,7 @@ class Program_weight_tensor_parameter_148: class Program_weight_tensor_parameter_149: name = "parameter_149" + original_name = "batch_norm2d_130.w_2" shape = [384] dtype = "float32" min_val = float("0.00224831") @@ -1580,6 +1730,7 @@ class Program_weight_tensor_parameter_149: class Program_weight_tensor_parameter_150: name = "parameter_150" + original_name = "batch_norm2d_130.w_1" shape = [384] dtype = "float32" min_val = float("-0.0202904") @@ -1591,6 +1742,7 @@ class Program_weight_tensor_parameter_150: class Program_weight_tensor_parameter_151: name = "parameter_151" + original_name = "batch_norm2d_128.b_0" shape = [384] dtype = "float32" min_val = float("-0.323027") @@ -1602,6 +1754,7 @@ class Program_weight_tensor_parameter_151: class Program_weight_tensor_parameter_152: name = "parameter_152" + original_name = "batch_norm2d_128.w_0" shape = [384] dtype = "float32" min_val = float("-0.306507") @@ -1613,6 +1766,7 @@ class Program_weight_tensor_parameter_152: class Program_weight_tensor_parameter_153: name = "parameter_153" + original_name = "batch_norm2d_128.w_2" shape = [384] dtype = "float32" min_val = float("0.000329539") @@ -1624,6 +1778,7 @@ class Program_weight_tensor_parameter_153: class Program_weight_tensor_parameter_154: name = "parameter_154" + original_name = "batch_norm2d_128.w_1" shape = [384] dtype = "float32" min_val = float("-1.17758") @@ -1635,6 +1790,7 @@ class Program_weight_tensor_parameter_154: class Program_weight_tensor_parameter_155: name = "parameter_155" + original_name = "conv2d_116.w_0" shape = [384, 1, 5, 5] dtype = "float32" min_val = float("-0.469001") @@ -1646,6 +1802,7 @@ class Program_weight_tensor_parameter_155: class Program_weight_tensor_parameter_156: name = "parameter_156" + original_name = "batch_norm2d_127.b_0" shape = [384] dtype = "float32" min_val = float("-0.323027") @@ -1657,6 +1814,7 @@ class Program_weight_tensor_parameter_156: class Program_weight_tensor_parameter_157: name = "parameter_157" + original_name = "batch_norm2d_127.w_0" shape = [384] dtype = "float32" min_val = float("-0.0161042") @@ -1668,6 +1826,7 @@ class Program_weight_tensor_parameter_157: class Program_weight_tensor_parameter_158: name = "parameter_158" + original_name = "batch_norm2d_127.w_2" shape = [384] dtype = "float32" min_val = float("0.000995517") @@ -1679,6 +1838,7 @@ class Program_weight_tensor_parameter_158: class Program_weight_tensor_parameter_159: name = "parameter_159" + original_name = "batch_norm2d_127.w_1" shape = [384] dtype = "float32" min_val = float("-1.18387") @@ -1690,6 +1850,7 @@ class Program_weight_tensor_parameter_159: class Program_weight_tensor_parameter_160: name = "parameter_160" + original_name = "conv2d_115.w_0" shape = [384, 1, 5, 5] dtype = "float32" min_val = float("-0.50121") @@ -1701,6 +1862,7 @@ class Program_weight_tensor_parameter_160: class Program_weight_tensor_parameter_161: name = "parameter_161" + original_name = "batch_norm2d_126.b_0" shape = [384] dtype = "float32" min_val = float("-0.323027") @@ -1712,6 +1874,7 @@ class Program_weight_tensor_parameter_161: class Program_weight_tensor_parameter_162: name = "parameter_162" + original_name = "batch_norm2d_126.w_0" shape = [384] dtype = "float32" min_val = float("-0.00928076") @@ -1723,6 +1886,7 @@ class Program_weight_tensor_parameter_162: class Program_weight_tensor_parameter_163: name = "parameter_163" + original_name = "batch_norm2d_126.w_2" shape = [384] dtype = "float32" min_val = float("0.000642092") @@ -1734,6 +1898,7 @@ class Program_weight_tensor_parameter_163: class Program_weight_tensor_parameter_164: name = "parameter_164" + original_name = "batch_norm2d_126.w_1" shape = [384] dtype = "float32" min_val = float("-1.17635") @@ -1745,6 +1910,7 @@ class Program_weight_tensor_parameter_164: class Program_weight_tensor_parameter_165: name = "parameter_165" + original_name = "conv2d_114.w_0" shape = [384, 1, 5, 5] dtype = "float32" min_val = float("-0.499429") @@ -1756,6 +1922,7 @@ class Program_weight_tensor_parameter_165: class Program_weight_tensor_parameter_166: name = "parameter_166" + original_name = "batch_norm2d_125.b_0" shape = [384] dtype = "float32" min_val = float("-0.323027") @@ -1767,6 +1934,7 @@ class Program_weight_tensor_parameter_166: class Program_weight_tensor_parameter_167: name = "parameter_167" + original_name = "batch_norm2d_125.w_0" shape = [384] dtype = "float32" min_val = float("-0.0876685") @@ -1778,6 +1946,7 @@ class Program_weight_tensor_parameter_167: class Program_weight_tensor_parameter_168: name = "parameter_168" + original_name = "batch_norm2d_125.w_2" shape = [384] dtype = "float32" min_val = float("0.000534702") @@ -1789,6 +1958,7 @@ class Program_weight_tensor_parameter_168: class Program_weight_tensor_parameter_169: name = "parameter_169" + original_name = "batch_norm2d_125.w_1" shape = [384] dtype = "float32" min_val = float("-1.177") @@ -1800,6 +1970,7 @@ class Program_weight_tensor_parameter_169: class Program_weight_tensor_parameter_170: name = "parameter_170" + original_name = "conv2d_113.w_0" shape = [384, 1, 5, 5] dtype = "float32" min_val = float("-0.499381") @@ -1811,6 +1982,7 @@ class Program_weight_tensor_parameter_170: class Program_weight_tensor_parameter_171: name = "parameter_171" + original_name = "batch_norm2d_129.b_0" shape = [384] dtype = "float32" min_val = float("-0.323027") @@ -1822,6 +1994,7 @@ class Program_weight_tensor_parameter_171: class Program_weight_tensor_parameter_172: name = "parameter_172" + original_name = "batch_norm2d_129.w_0" shape = [384] dtype = "float32" min_val = float("-0.232808") @@ -1833,6 +2006,7 @@ class Program_weight_tensor_parameter_172: class Program_weight_tensor_parameter_173: name = "parameter_173" + original_name = "batch_norm2d_129.w_2" shape = [384] dtype = "float32" min_val = float("5.57038e-10") @@ -1844,6 +2018,7 @@ class Program_weight_tensor_parameter_173: class Program_weight_tensor_parameter_174: name = "parameter_174" + original_name = "batch_norm2d_129.w_1" shape = [384] dtype = "float32" min_val = float("-0.0452907") @@ -1855,6 +2030,7 @@ class Program_weight_tensor_parameter_174: class Program_weight_tensor_parameter_175: name = "parameter_175" + original_name = "conv2d_117.w_0" shape = [384, 1, 1, 1] dtype = "float32" min_val = float("-0.125777") @@ -1866,6 +2042,7 @@ class Program_weight_tensor_parameter_175: class Program_weight_tensor_parameter_176: name = "parameter_176" + original_name = "batch_norm2d_124.b_0" shape = [384] dtype = "float32" min_val = float("-0.302161") @@ -1877,6 +2054,7 @@ class Program_weight_tensor_parameter_176: class Program_weight_tensor_parameter_177: name = "parameter_177" + original_name = "batch_norm2d_124.w_0" shape = [384] dtype = "float32" min_val = float("-0.209853") @@ -1888,6 +2066,7 @@ class Program_weight_tensor_parameter_177: class Program_weight_tensor_parameter_178: name = "parameter_178" + original_name = "batch_norm2d_124.w_2" shape = [384] dtype = "float32" min_val = float("0.00494122") @@ -1899,6 +2078,7 @@ class Program_weight_tensor_parameter_178: class Program_weight_tensor_parameter_179: name = "parameter_179" + original_name = "batch_norm2d_124.w_1" shape = [384] dtype = "float32" min_val = float("-0.0743273") @@ -1910,6 +2090,7 @@ class Program_weight_tensor_parameter_179: class Program_weight_tensor_parameter_180: name = "parameter_180" + original_name = "batch_norm2d_123.b_0" shape = [384] dtype = "float32" min_val = float("-0.312754") @@ -1921,6 +2102,7 @@ class Program_weight_tensor_parameter_180: class Program_weight_tensor_parameter_181: name = "parameter_181" + original_name = "batch_norm2d_123.w_0" shape = [384] dtype = "float32" min_val = float("0.00389017") @@ -1932,6 +2114,7 @@ class Program_weight_tensor_parameter_181: class Program_weight_tensor_parameter_182: name = "parameter_182" + original_name = "batch_norm2d_123.w_2" shape = [384] dtype = "float32" min_val = float("0.347776") @@ -1943,6 +2126,7 @@ class Program_weight_tensor_parameter_182: class Program_weight_tensor_parameter_183: name = "parameter_183" + original_name = "batch_norm2d_123.w_1" shape = [384] dtype = "float32" min_val = float("-0.89453") @@ -1954,6 +2138,7 @@ class Program_weight_tensor_parameter_183: class Program_weight_tensor_parameter_184: name = "parameter_184" + original_name = "conv2d_112.w_0" shape = [384, 384, 1, 1] dtype = "float32" min_val = float("-0.584804") @@ -1965,6 +2150,7 @@ class Program_weight_tensor_parameter_184: class Program_weight_tensor_parameter_185: name = "parameter_185" + original_name = "batch_norm2d_122.b_0" shape = [384] dtype = "float32" min_val = float("-0.312754") @@ -1976,6 +2162,7 @@ class Program_weight_tensor_parameter_185: class Program_weight_tensor_parameter_186: name = "parameter_186" + original_name = "batch_norm2d_122.w_0" shape = [384] dtype = "float32" min_val = float("-0.000706829") @@ -1987,6 +2174,7 @@ class Program_weight_tensor_parameter_186: class Program_weight_tensor_parameter_187: name = "parameter_187" + original_name = "batch_norm2d_122.w_2" shape = [384] dtype = "float32" min_val = float("0.215972") @@ -1998,6 +2186,7 @@ class Program_weight_tensor_parameter_187: class Program_weight_tensor_parameter_188: name = "parameter_188" + original_name = "batch_norm2d_122.w_1" shape = [384] dtype = "float32" min_val = float("-0.967287") @@ -2009,6 +2198,7 @@ class Program_weight_tensor_parameter_188: class Program_weight_tensor_parameter_189: name = "parameter_189" + original_name = "conv2d_111.w_0" shape = [384, 384, 1, 1] dtype = "float32" min_val = float("-0.593506") @@ -2020,6 +2210,7 @@ class Program_weight_tensor_parameter_189: class Program_weight_tensor_parameter_190: name = "parameter_190" + original_name = "batch_norm2d_121.b_0" shape = [384] dtype = "float32" min_val = float("-0.312754") @@ -2031,6 +2222,7 @@ class Program_weight_tensor_parameter_190: class Program_weight_tensor_parameter_191: name = "parameter_191" + original_name = "batch_norm2d_121.w_0" shape = [384] dtype = "float32" min_val = float("-0.000188664") @@ -2042,6 +2234,7 @@ class Program_weight_tensor_parameter_191: class Program_weight_tensor_parameter_192: name = "parameter_192" + original_name = "batch_norm2d_121.w_2" shape = [384] dtype = "float32" min_val = float("0.29037") @@ -2053,6 +2246,7 @@ class Program_weight_tensor_parameter_192: class Program_weight_tensor_parameter_193: name = "parameter_193" + original_name = "batch_norm2d_121.w_1" shape = [384] dtype = "float32" min_val = float("-0.94837") @@ -2064,6 +2258,7 @@ class Program_weight_tensor_parameter_193: class Program_weight_tensor_parameter_194: name = "parameter_194" + original_name = "conv2d_110.w_0" shape = [384, 384, 1, 1] dtype = "float32" min_val = float("-0.612275") @@ -2075,6 +2270,7 @@ class Program_weight_tensor_parameter_194: class Program_weight_tensor_parameter_195: name = "parameter_195" + original_name = "batch_norm2d_120.b_0" shape = [384] dtype = "float32" min_val = float("-0.312754") @@ -2086,6 +2282,7 @@ class Program_weight_tensor_parameter_195: class Program_weight_tensor_parameter_196: name = "parameter_196" + original_name = "batch_norm2d_120.w_0" shape = [384] dtype = "float32" min_val = float("-0.00379736") @@ -2097,6 +2294,7 @@ class Program_weight_tensor_parameter_196: class Program_weight_tensor_parameter_197: name = "parameter_197" + original_name = "batch_norm2d_120.w_2" shape = [384] dtype = "float32" min_val = float("0.163463") @@ -2108,6 +2306,7 @@ class Program_weight_tensor_parameter_197: class Program_weight_tensor_parameter_198: name = "parameter_198" + original_name = "batch_norm2d_120.w_1" shape = [384] dtype = "float32" min_val = float("-0.871506") @@ -2119,6 +2318,7 @@ class Program_weight_tensor_parameter_198: class Program_weight_tensor_parameter_199: name = "parameter_199" + original_name = "conv2d_109.w_0" shape = [384, 384, 1, 1] dtype = "float32" min_val = float("-0.631798") @@ -2130,6 +2330,7 @@ class Program_weight_tensor_parameter_199: class Program_weight_tensor_parameter_200: name = "parameter_200" + original_name = "batch_norm2d_119.b_0" shape = [384] dtype = "float32" min_val = float("-0.303134") @@ -2141,6 +2342,7 @@ class Program_weight_tensor_parameter_200: class Program_weight_tensor_parameter_201: name = "parameter_201" + original_name = "batch_norm2d_119.w_0" shape = [384] dtype = "float32" min_val = float("-0.125409") @@ -2152,6 +2354,7 @@ class Program_weight_tensor_parameter_201: class Program_weight_tensor_parameter_202: name = "parameter_202" + original_name = "batch_norm2d_119.w_2" shape = [384] dtype = "float32" min_val = float("3.27586e-05") @@ -2163,6 +2366,7 @@ class Program_weight_tensor_parameter_202: class Program_weight_tensor_parameter_203: name = "parameter_203" + original_name = "batch_norm2d_119.w_1" shape = [384] dtype = "float32" min_val = float("-0.934814") @@ -2174,6 +2378,7 @@ class Program_weight_tensor_parameter_203: class Program_weight_tensor_parameter_204: name = "parameter_204" + original_name = "conv2d_108.b_0" shape = [384] dtype = "float32" min_val = float("-0.351636") @@ -2185,6 +2390,7 @@ class Program_weight_tensor_parameter_204: class Program_weight_tensor_parameter_205: name = "parameter_205" + original_name = "conv2d_108.w_0" shape = [384, 96, 1, 1] dtype = "float32" min_val = float("-0.896196") @@ -2196,6 +2402,7 @@ class Program_weight_tensor_parameter_205: class Program_weight_tensor_parameter_206: name = "parameter_206" + original_name = "conv2d_107.b_0" shape = [96] dtype = "float32" min_val = float("-0.212827") @@ -2207,6 +2414,7 @@ class Program_weight_tensor_parameter_206: class Program_weight_tensor_parameter_207: name = "parameter_207" + original_name = "conv2d_107.w_0" shape = [96, 384, 1, 1] dtype = "float32" min_val = float("-1.24274") @@ -2218,6 +2426,7 @@ class Program_weight_tensor_parameter_207: class Program_weight_tensor_parameter_208: name = "parameter_208" + original_name = "batch_norm2d_117.b_0" shape = [384] dtype = "float32" min_val = float("-0.281138") @@ -2229,6 +2438,7 @@ class Program_weight_tensor_parameter_208: class Program_weight_tensor_parameter_209: name = "parameter_209" + original_name = "batch_norm2d_117.w_0" shape = [384] dtype = "float32" min_val = float("-0.0177038") @@ -2240,6 +2450,7 @@ class Program_weight_tensor_parameter_209: class Program_weight_tensor_parameter_210: name = "parameter_210" + original_name = "batch_norm2d_117.w_2" shape = [384] dtype = "float32" min_val = float("0.000701383") @@ -2251,6 +2462,7 @@ class Program_weight_tensor_parameter_210: class Program_weight_tensor_parameter_211: name = "parameter_211" + original_name = "batch_norm2d_117.w_1" shape = [384] dtype = "float32" min_val = float("-5.31026") @@ -2262,6 +2474,7 @@ class Program_weight_tensor_parameter_211: class Program_weight_tensor_parameter_212: name = "parameter_212" + original_name = "conv2d_105.w_0" shape = [384, 1, 5, 5] dtype = "float32" min_val = float("-0.477105") @@ -2273,6 +2486,7 @@ class Program_weight_tensor_parameter_212: class Program_weight_tensor_parameter_213: name = "parameter_213" + original_name = "batch_norm2d_116.b_0" shape = [384] dtype = "float32" min_val = float("-0.281138") @@ -2284,6 +2498,7 @@ class Program_weight_tensor_parameter_213: class Program_weight_tensor_parameter_214: name = "parameter_214" + original_name = "batch_norm2d_116.w_0" shape = [384] dtype = "float32" min_val = float("-0.150741") @@ -2295,6 +2510,7 @@ class Program_weight_tensor_parameter_214: class Program_weight_tensor_parameter_215: name = "parameter_215" + original_name = "batch_norm2d_116.w_2" shape = [384] dtype = "float32" min_val = float("0.000706115") @@ -2306,6 +2522,7 @@ class Program_weight_tensor_parameter_215: class Program_weight_tensor_parameter_216: name = "parameter_216" + original_name = "batch_norm2d_116.w_1" shape = [384] dtype = "float32" min_val = float("-4.21819") @@ -2317,6 +2534,7 @@ class Program_weight_tensor_parameter_216: class Program_weight_tensor_parameter_217: name = "parameter_217" + original_name = "conv2d_104.w_0" shape = [384, 1, 5, 5] dtype = "float32" min_val = float("-0.476284") @@ -2328,6 +2546,7 @@ class Program_weight_tensor_parameter_217: class Program_weight_tensor_parameter_218: name = "parameter_218" + original_name = "batch_norm2d_115.b_0" shape = [384] dtype = "float32" min_val = float("-0.281138") @@ -2339,6 +2558,7 @@ class Program_weight_tensor_parameter_218: class Program_weight_tensor_parameter_219: name = "parameter_219" + original_name = "batch_norm2d_115.w_0" shape = [384] dtype = "float32" min_val = float("-0.142942") @@ -2350,6 +2570,7 @@ class Program_weight_tensor_parameter_219: class Program_weight_tensor_parameter_220: name = "parameter_220" + original_name = "batch_norm2d_115.w_2" shape = [384] dtype = "float32" min_val = float("0.000701905") @@ -2361,6 +2582,7 @@ class Program_weight_tensor_parameter_220: class Program_weight_tensor_parameter_221: name = "parameter_221" + original_name = "batch_norm2d_115.w_1" shape = [384] dtype = "float32" min_val = float("-5.33") @@ -2372,6 +2594,7 @@ class Program_weight_tensor_parameter_221: class Program_weight_tensor_parameter_222: name = "parameter_222" + original_name = "conv2d_103.w_0" shape = [384, 1, 5, 5] dtype = "float32" min_val = float("-0.476901") @@ -2383,6 +2606,7 @@ class Program_weight_tensor_parameter_222: class Program_weight_tensor_parameter_223: name = "parameter_223" + original_name = "batch_norm2d_114.b_0" shape = [384] dtype = "float32" min_val = float("-0.281138") @@ -2394,6 +2618,7 @@ class Program_weight_tensor_parameter_223: class Program_weight_tensor_parameter_224: name = "parameter_224" + original_name = "batch_norm2d_114.w_0" shape = [384] dtype = "float32" min_val = float("-0.00594849") @@ -2405,6 +2630,7 @@ class Program_weight_tensor_parameter_224: class Program_weight_tensor_parameter_225: name = "parameter_225" + original_name = "batch_norm2d_114.w_2" shape = [384] dtype = "float32" min_val = float("0.00071371") @@ -2416,6 +2642,7 @@ class Program_weight_tensor_parameter_225: class Program_weight_tensor_parameter_226: name = "parameter_226" + original_name = "batch_norm2d_114.w_1" shape = [384] dtype = "float32" min_val = float("-4.22103") @@ -2427,6 +2654,7 @@ class Program_weight_tensor_parameter_226: class Program_weight_tensor_parameter_227: name = "parameter_227" + original_name = "conv2d_102.w_0" shape = [384, 1, 5, 5] dtype = "float32" min_val = float("-0.475724") @@ -2438,6 +2666,7 @@ class Program_weight_tensor_parameter_227: class Program_weight_tensor_parameter_228: name = "parameter_228" + original_name = "batch_norm2d_118.b_0" shape = [384] dtype = "float32" min_val = float("-0.281138") @@ -2449,6 +2678,7 @@ class Program_weight_tensor_parameter_228: class Program_weight_tensor_parameter_229: name = "parameter_229" + original_name = "batch_norm2d_118.w_0" shape = [384] dtype = "float32" min_val = float("-0.45451") @@ -2460,6 +2690,7 @@ class Program_weight_tensor_parameter_229: class Program_weight_tensor_parameter_230: name = "parameter_230" + original_name = "batch_norm2d_118.w_2" shape = [384] dtype = "float32" min_val = float("3.97754e-09") @@ -2471,6 +2702,7 @@ class Program_weight_tensor_parameter_230: class Program_weight_tensor_parameter_231: name = "parameter_231" + original_name = "batch_norm2d_118.w_1" shape = [384] dtype = "float32" min_val = float("-0.10556") @@ -2482,6 +2714,7 @@ class Program_weight_tensor_parameter_231: class Program_weight_tensor_parameter_232: name = "parameter_232" + original_name = "conv2d_106.w_0" shape = [384, 1, 1, 1] dtype = "float32" min_val = float("-0.130368") @@ -2493,6 +2726,7 @@ class Program_weight_tensor_parameter_232: class Program_weight_tensor_parameter_233: name = "parameter_233" + original_name = "batch_norm2d_113.b_0" shape = [384] dtype = "float32" min_val = float("-0.271179") @@ -2504,6 +2738,7 @@ class Program_weight_tensor_parameter_233: class Program_weight_tensor_parameter_234: name = "parameter_234" + original_name = "batch_norm2d_113.w_0" shape = [384] dtype = "float32" min_val = float("-0.326484") @@ -2515,6 +2750,7 @@ class Program_weight_tensor_parameter_234: class Program_weight_tensor_parameter_235: name = "parameter_235" + original_name = "batch_norm2d_113.w_2" shape = [384] dtype = "float32" min_val = float("0.00176364") @@ -2526,6 +2762,7 @@ class Program_weight_tensor_parameter_235: class Program_weight_tensor_parameter_236: name = "parameter_236" + original_name = "batch_norm2d_113.w_1" shape = [384] dtype = "float32" min_val = float("-0.218312") @@ -2537,6 +2774,7 @@ class Program_weight_tensor_parameter_236: class Program_weight_tensor_parameter_237: name = "parameter_237" + original_name = "batch_norm2d_112.b_0" shape = [384] dtype = "float32" min_val = float("-0.464706") @@ -2548,6 +2786,7 @@ class Program_weight_tensor_parameter_237: class Program_weight_tensor_parameter_238: name = "parameter_238" + original_name = "batch_norm2d_112.w_0" shape = [384] dtype = "float32" min_val = float("-0.00399007") @@ -2559,6 +2798,7 @@ class Program_weight_tensor_parameter_238: class Program_weight_tensor_parameter_239: name = "parameter_239" + original_name = "batch_norm2d_112.w_2" shape = [384] dtype = "float32" min_val = float("0.0312595") @@ -2570,6 +2810,7 @@ class Program_weight_tensor_parameter_239: class Program_weight_tensor_parameter_240: name = "parameter_240" + original_name = "batch_norm2d_112.w_1" shape = [384] dtype = "float32" min_val = float("-0.654959") @@ -2581,6 +2822,7 @@ class Program_weight_tensor_parameter_240: class Program_weight_tensor_parameter_241: name = "parameter_241" + original_name = "conv2d_101.w_0" shape = [384, 192, 1, 1] dtype = "float32" min_val = float("-0.635369") @@ -2592,6 +2834,7 @@ class Program_weight_tensor_parameter_241: class Program_weight_tensor_parameter_242: name = "parameter_242" + original_name = "batch_norm2d_111.b_0" shape = [384] dtype = "float32" min_val = float("-0.464706") @@ -2603,6 +2846,7 @@ class Program_weight_tensor_parameter_242: class Program_weight_tensor_parameter_243: name = "parameter_243" + original_name = "batch_norm2d_111.w_0" shape = [384] dtype = "float32" min_val = float("-0.00583191") @@ -2614,6 +2858,7 @@ class Program_weight_tensor_parameter_243: class Program_weight_tensor_parameter_244: name = "parameter_244" + original_name = "batch_norm2d_111.w_2" shape = [384] dtype = "float32" min_val = float("0.0706834") @@ -2625,6 +2870,7 @@ class Program_weight_tensor_parameter_244: class Program_weight_tensor_parameter_245: name = "parameter_245" + original_name = "batch_norm2d_111.w_1" shape = [384] dtype = "float32" min_val = float("-0.648639") @@ -2636,6 +2882,7 @@ class Program_weight_tensor_parameter_245: class Program_weight_tensor_parameter_246: name = "parameter_246" + original_name = "conv2d_100.w_0" shape = [384, 192, 1, 1] dtype = "float32" min_val = float("-0.581688") @@ -2647,6 +2894,7 @@ class Program_weight_tensor_parameter_246: class Program_weight_tensor_parameter_247: name = "parameter_247" + original_name = "batch_norm2d_110.b_0" shape = [384] dtype = "float32" min_val = float("-0.464706") @@ -2658,6 +2906,7 @@ class Program_weight_tensor_parameter_247: class Program_weight_tensor_parameter_248: name = "parameter_248" + original_name = "batch_norm2d_110.w_0" shape = [384] dtype = "float32" min_val = float("-0.00349256") @@ -2669,6 +2918,7 @@ class Program_weight_tensor_parameter_248: class Program_weight_tensor_parameter_249: name = "parameter_249" + original_name = "batch_norm2d_110.w_2" shape = [384] dtype = "float32" min_val = float("0.0550379") @@ -2680,6 +2930,7 @@ class Program_weight_tensor_parameter_249: class Program_weight_tensor_parameter_250: name = "parameter_250" + original_name = "batch_norm2d_110.w_1" shape = [384] dtype = "float32" min_val = float("-0.655926") @@ -2691,6 +2942,7 @@ class Program_weight_tensor_parameter_250: class Program_weight_tensor_parameter_251: name = "parameter_251" + original_name = "conv2d_99.w_0" shape = [384, 192, 1, 1] dtype = "float32" min_val = float("-0.622883") @@ -2702,6 +2954,7 @@ class Program_weight_tensor_parameter_251: class Program_weight_tensor_parameter_252: name = "parameter_252" + original_name = "batch_norm2d_109.b_0" shape = [384] dtype = "float32" min_val = float("-0.464706") @@ -2713,6 +2966,7 @@ class Program_weight_tensor_parameter_252: class Program_weight_tensor_parameter_253: name = "parameter_253" + original_name = "batch_norm2d_109.w_0" shape = [384] dtype = "float32" min_val = float("-0.00147187") @@ -2724,6 +2978,7 @@ class Program_weight_tensor_parameter_253: class Program_weight_tensor_parameter_254: name = "parameter_254" + original_name = "batch_norm2d_109.w_2" shape = [384] dtype = "float32" min_val = float("0.0547428") @@ -2735,6 +2990,7 @@ class Program_weight_tensor_parameter_254: class Program_weight_tensor_parameter_255: name = "parameter_255" + original_name = "batch_norm2d_109.w_1" shape = [384] dtype = "float32" min_val = float("-0.611931") @@ -2746,6 +3002,7 @@ class Program_weight_tensor_parameter_255: class Program_weight_tensor_parameter_256: name = "parameter_256" + original_name = "conv2d_98.w_0" shape = [384, 192, 1, 1] dtype = "float32" min_val = float("-0.555636") @@ -2757,6 +3014,7 @@ class Program_weight_tensor_parameter_256: class Program_weight_tensor_parameter_257: name = "parameter_257" + original_name = "conv2d_97.b_0" shape = [192] dtype = "float32" min_val = float("-0.253323") @@ -2768,6 +3026,7 @@ class Program_weight_tensor_parameter_257: class Program_weight_tensor_parameter_258: name = "parameter_258" + original_name = "conv2d_97.w_0" shape = [192, 48, 1, 1] dtype = "float32" min_val = float("-0.586485") @@ -2779,6 +3038,7 @@ class Program_weight_tensor_parameter_258: class Program_weight_tensor_parameter_259: name = "parameter_259" + original_name = "conv2d_96.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -2788,6 +3048,7 @@ class Program_weight_tensor_parameter_259: class Program_weight_tensor_parameter_260: name = "parameter_260" + original_name = "conv2d_96.w_0" shape = [48, 192, 1, 1] dtype = "float32" min_val = float("-0.734674") @@ -2799,6 +3060,7 @@ class Program_weight_tensor_parameter_260: class Program_weight_tensor_parameter_261: name = "parameter_261" + original_name = "batch_norm2d_107.b_0" shape = [192] dtype = "float32" min_val = float("-0.291474") @@ -2810,6 +3072,7 @@ class Program_weight_tensor_parameter_261: class Program_weight_tensor_parameter_262: name = "parameter_262" + original_name = "batch_norm2d_107.w_0" shape = [192] dtype = "float32" min_val = float("-0.00184506") @@ -2821,6 +3084,7 @@ class Program_weight_tensor_parameter_262: class Program_weight_tensor_parameter_263: name = "parameter_263" + original_name = "batch_norm2d_107.w_2" shape = [192] dtype = "float32" min_val = float("0.000667279") @@ -2832,6 +3096,7 @@ class Program_weight_tensor_parameter_263: class Program_weight_tensor_parameter_264: name = "parameter_264" + original_name = "batch_norm2d_107.w_1" shape = [192] dtype = "float32" min_val = float("-0.331656") @@ -2843,6 +3108,7 @@ class Program_weight_tensor_parameter_264: class Program_weight_tensor_parameter_265: name = "parameter_265" + original_name = "conv2d_94.w_0" shape = [192, 1, 5, 5] dtype = "float32" min_val = float("-0.345684") @@ -2854,6 +3120,7 @@ class Program_weight_tensor_parameter_265: class Program_weight_tensor_parameter_266: name = "parameter_266" + original_name = "batch_norm2d_106.b_0" shape = [192] dtype = "float32" min_val = float("-0.291474") @@ -2865,6 +3132,7 @@ class Program_weight_tensor_parameter_266: class Program_weight_tensor_parameter_267: name = "parameter_267" + original_name = "batch_norm2d_106.w_0" shape = [192] dtype = "float32" min_val = float("-0.00720705") @@ -2876,6 +3144,7 @@ class Program_weight_tensor_parameter_267: class Program_weight_tensor_parameter_268: name = "parameter_268" + original_name = "batch_norm2d_106.w_2" shape = [192] dtype = "float32" min_val = float("0.0017789") @@ -2887,6 +3156,7 @@ class Program_weight_tensor_parameter_268: class Program_weight_tensor_parameter_269: name = "parameter_269" + original_name = "batch_norm2d_106.w_1" shape = [192] dtype = "float32" min_val = float("-0.422027") @@ -2898,6 +3168,7 @@ class Program_weight_tensor_parameter_269: class Program_weight_tensor_parameter_270: name = "parameter_270" + original_name = "conv2d_93.w_0" shape = [192, 1, 5, 5] dtype = "float32" min_val = float("-0.296363") @@ -2909,6 +3180,7 @@ class Program_weight_tensor_parameter_270: class Program_weight_tensor_parameter_271: name = "parameter_271" + original_name = "batch_norm2d_105.b_0" shape = [192] dtype = "float32" min_val = float("-0.291474") @@ -2920,6 +3192,7 @@ class Program_weight_tensor_parameter_271: class Program_weight_tensor_parameter_272: name = "parameter_272" + original_name = "batch_norm2d_105.w_0" shape = [192] dtype = "float32" min_val = float("-0.00406078") @@ -2931,6 +3204,7 @@ class Program_weight_tensor_parameter_272: class Program_weight_tensor_parameter_273: name = "parameter_273" + original_name = "batch_norm2d_105.w_2" shape = [192] dtype = "float32" min_val = float("0.00329588") @@ -2942,6 +3216,7 @@ class Program_weight_tensor_parameter_273: class Program_weight_tensor_parameter_274: name = "parameter_274" + original_name = "batch_norm2d_105.w_1" shape = [192] dtype = "float32" min_val = float("-0.421413") @@ -2953,6 +3228,7 @@ class Program_weight_tensor_parameter_274: class Program_weight_tensor_parameter_275: name = "parameter_275" + original_name = "conv2d_92.w_0" shape = [192, 1, 5, 5] dtype = "float32" min_val = float("-0.294495") @@ -2964,6 +3240,7 @@ class Program_weight_tensor_parameter_275: class Program_weight_tensor_parameter_276: name = "parameter_276" + original_name = "batch_norm2d_104.b_0" shape = [192] dtype = "float32" min_val = float("-0.291474") @@ -2975,6 +3252,7 @@ class Program_weight_tensor_parameter_276: class Program_weight_tensor_parameter_277: name = "parameter_277" + original_name = "batch_norm2d_104.w_0" shape = [192] dtype = "float32" min_val = float("-0.0139822") @@ -2986,6 +3264,7 @@ class Program_weight_tensor_parameter_277: class Program_weight_tensor_parameter_278: name = "parameter_278" + original_name = "batch_norm2d_104.w_2" shape = [192] dtype = "float32" min_val = float("0.000501186") @@ -2997,6 +3276,7 @@ class Program_weight_tensor_parameter_278: class Program_weight_tensor_parameter_279: name = "parameter_279" + original_name = "batch_norm2d_104.w_1" shape = [192] dtype = "float32" min_val = float("-0.423032") @@ -3008,6 +3288,7 @@ class Program_weight_tensor_parameter_279: class Program_weight_tensor_parameter_280: name = "parameter_280" + original_name = "conv2d_91.w_0" shape = [192, 1, 5, 5] dtype = "float32" min_val = float("-0.342233") @@ -3019,6 +3300,7 @@ class Program_weight_tensor_parameter_280: class Program_weight_tensor_parameter_281: name = "parameter_281" + original_name = "batch_norm2d_108.b_0" shape = [192] dtype = "float32" min_val = float("-0.291474") @@ -3030,6 +3312,7 @@ class Program_weight_tensor_parameter_281: class Program_weight_tensor_parameter_282: name = "parameter_282" + original_name = "batch_norm2d_108.w_0" shape = [192] dtype = "float32" min_val = float("-0.181217") @@ -3041,6 +3324,7 @@ class Program_weight_tensor_parameter_282: class Program_weight_tensor_parameter_283: name = "parameter_283" + original_name = "batch_norm2d_108.w_2" shape = [192] dtype = "float32" min_val = float("1.6884e-10") @@ -3052,6 +3336,7 @@ class Program_weight_tensor_parameter_283: class Program_weight_tensor_parameter_284: name = "parameter_284" + original_name = "batch_norm2d_108.w_1" shape = [192] dtype = "float32" min_val = float("-0.0301091") @@ -3063,6 +3348,7 @@ class Program_weight_tensor_parameter_284: class Program_weight_tensor_parameter_285: name = "parameter_285" + original_name = "conv2d_95.w_0" shape = [192, 1, 1, 1] dtype = "float32" min_val = float("-0.132698") @@ -3074,6 +3360,7 @@ class Program_weight_tensor_parameter_285: class Program_weight_tensor_parameter_286: name = "parameter_286" + original_name = "batch_norm2d_103.b_0" shape = [192] dtype = "float32" min_val = float("-0.299652") @@ -3085,6 +3372,7 @@ class Program_weight_tensor_parameter_286: class Program_weight_tensor_parameter_287: name = "parameter_287" + original_name = "batch_norm2d_103.w_0" shape = [192] dtype = "float32" min_val = float("-0.00325594") @@ -3096,6 +3384,7 @@ class Program_weight_tensor_parameter_287: class Program_weight_tensor_parameter_288: name = "parameter_288" + original_name = "batch_norm2d_103.w_2" shape = [192] dtype = "float32" min_val = float("0.0428087") @@ -3107,6 +3396,7 @@ class Program_weight_tensor_parameter_288: class Program_weight_tensor_parameter_289: name = "parameter_289" + original_name = "batch_norm2d_103.w_1" shape = [192] dtype = "float32" min_val = float("-2.13946") @@ -3118,6 +3408,7 @@ class Program_weight_tensor_parameter_289: class Program_weight_tensor_parameter_290: name = "parameter_290" + original_name = "conv2d_90.w_0" shape = [192, 192, 1, 1] dtype = "float32" min_val = float("-0.862979") @@ -3129,6 +3420,7 @@ class Program_weight_tensor_parameter_290: class Program_weight_tensor_parameter_291: name = "parameter_291" + original_name = "batch_norm2d_102.b_0" shape = [192] dtype = "float32" min_val = float("-0.299652") @@ -3140,6 +3432,7 @@ class Program_weight_tensor_parameter_291: class Program_weight_tensor_parameter_292: name = "parameter_292" + original_name = "batch_norm2d_102.w_0" shape = [192] dtype = "float32" min_val = float("0.109786") @@ -3151,6 +3444,7 @@ class Program_weight_tensor_parameter_292: class Program_weight_tensor_parameter_293: name = "parameter_293" + original_name = "batch_norm2d_102.w_2" shape = [192] dtype = "float32" min_val = float("0.0905589") @@ -3162,6 +3456,7 @@ class Program_weight_tensor_parameter_293: class Program_weight_tensor_parameter_294: name = "parameter_294" + original_name = "batch_norm2d_102.w_1" shape = [192] dtype = "float32" min_val = float("-1.97884") @@ -3173,6 +3468,7 @@ class Program_weight_tensor_parameter_294: class Program_weight_tensor_parameter_295: name = "parameter_295" + original_name = "conv2d_89.w_0" shape = [192, 192, 1, 1] dtype = "float32" min_val = float("-0.984592") @@ -3184,6 +3480,7 @@ class Program_weight_tensor_parameter_295: class Program_weight_tensor_parameter_296: name = "parameter_296" + original_name = "batch_norm2d_101.b_0" shape = [192] dtype = "float32" min_val = float("-0.299652") @@ -3195,6 +3492,7 @@ class Program_weight_tensor_parameter_296: class Program_weight_tensor_parameter_297: name = "parameter_297" + original_name = "batch_norm2d_101.w_0" shape = [192] dtype = "float32" min_val = float("0.0908175") @@ -3206,6 +3504,7 @@ class Program_weight_tensor_parameter_297: class Program_weight_tensor_parameter_298: name = "parameter_298" + original_name = "batch_norm2d_101.w_2" shape = [192] dtype = "float32" min_val = float("0.0974024") @@ -3217,6 +3516,7 @@ class Program_weight_tensor_parameter_298: class Program_weight_tensor_parameter_299: name = "parameter_299" + original_name = "batch_norm2d_101.w_1" shape = [192] dtype = "float32" min_val = float("-1.97119") @@ -3228,6 +3528,7 @@ class Program_weight_tensor_parameter_299: class Program_weight_tensor_parameter_300: name = "parameter_300" + original_name = "conv2d_88.w_0" shape = [192, 192, 1, 1] dtype = "float32" min_val = float("-0.983374") @@ -3239,6 +3540,7 @@ class Program_weight_tensor_parameter_300: class Program_weight_tensor_parameter_301: name = "parameter_301" + original_name = "batch_norm2d_100.b_0" shape = [192] dtype = "float32" min_val = float("-0.299652") @@ -3250,6 +3552,7 @@ class Program_weight_tensor_parameter_301: class Program_weight_tensor_parameter_302: name = "parameter_302" + original_name = "batch_norm2d_100.w_0" shape = [192] dtype = "float32" min_val = float("0.0943865") @@ -3261,6 +3564,7 @@ class Program_weight_tensor_parameter_302: class Program_weight_tensor_parameter_303: name = "parameter_303" + original_name = "batch_norm2d_100.w_2" shape = [192] dtype = "float32" min_val = float("0.0985271") @@ -3272,6 +3576,7 @@ class Program_weight_tensor_parameter_303: class Program_weight_tensor_parameter_304: name = "parameter_304" + original_name = "batch_norm2d_100.w_1" shape = [192] dtype = "float32" min_val = float("-2.0113") @@ -3283,6 +3588,7 @@ class Program_weight_tensor_parameter_304: class Program_weight_tensor_parameter_305: name = "parameter_305" + original_name = "conv2d_87.w_0" shape = [192, 192, 1, 1] dtype = "float32" min_val = float("-0.931198") @@ -3294,6 +3600,7 @@ class Program_weight_tensor_parameter_305: class Program_weight_tensor_parameter_306: name = "parameter_306" + original_name = "batch_norm2d_99.b_0" shape = [192] dtype = "float32" min_val = float("-0.285114") @@ -3305,6 +3612,7 @@ class Program_weight_tensor_parameter_306: class Program_weight_tensor_parameter_307: name = "parameter_307" + original_name = "batch_norm2d_99.w_0" shape = [192] dtype = "float32" min_val = float("-0.12942") @@ -3316,6 +3624,7 @@ class Program_weight_tensor_parameter_307: class Program_weight_tensor_parameter_308: name = "parameter_308" + original_name = "batch_norm2d_99.w_2" shape = [192] dtype = "float32" min_val = float("0.00401344") @@ -3327,6 +3636,7 @@ class Program_weight_tensor_parameter_308: class Program_weight_tensor_parameter_309: name = "parameter_309" + original_name = "batch_norm2d_99.w_1" shape = [192] dtype = "float32" min_val = float("-0.0317386") @@ -3338,6 +3648,7 @@ class Program_weight_tensor_parameter_309: class Program_weight_tensor_parameter_310: name = "parameter_310" + original_name = "batch_norm2d_97.b_0" shape = [192] dtype = "float32" min_val = float("-0.413204") @@ -3349,6 +3660,7 @@ class Program_weight_tensor_parameter_310: class Program_weight_tensor_parameter_311: name = "parameter_311" + original_name = "batch_norm2d_97.w_0" shape = [192] dtype = "float32" min_val = float("-0.00684977") @@ -3360,6 +3672,7 @@ class Program_weight_tensor_parameter_311: class Program_weight_tensor_parameter_312: name = "parameter_312" + original_name = "batch_norm2d_97.w_2" shape = [192] dtype = "float32" min_val = float("0.0035388") @@ -3371,6 +3684,7 @@ class Program_weight_tensor_parameter_312: class Program_weight_tensor_parameter_313: name = "parameter_313" + original_name = "batch_norm2d_97.w_1" shape = [192] dtype = "float32" min_val = float("-1.95306") @@ -3382,6 +3696,7 @@ class Program_weight_tensor_parameter_313: class Program_weight_tensor_parameter_314: name = "parameter_314" + original_name = "conv2d_85.w_0" shape = [192, 1, 5, 5] dtype = "float32" min_val = float("-0.467915") @@ -3393,6 +3708,7 @@ class Program_weight_tensor_parameter_314: class Program_weight_tensor_parameter_315: name = "parameter_315" + original_name = "batch_norm2d_96.b_0" shape = [192] dtype = "float32" min_val = float("-0.413204") @@ -3404,6 +3720,7 @@ class Program_weight_tensor_parameter_315: class Program_weight_tensor_parameter_316: name = "parameter_316" + original_name = "batch_norm2d_96.w_0" shape = [192] dtype = "float32" min_val = float("-0.110382") @@ -3415,6 +3732,7 @@ class Program_weight_tensor_parameter_316: class Program_weight_tensor_parameter_317: name = "parameter_317" + original_name = "batch_norm2d_96.w_2" shape = [192] dtype = "float32" min_val = float("0.00243392") @@ -3426,6 +3744,7 @@ class Program_weight_tensor_parameter_317: class Program_weight_tensor_parameter_318: name = "parameter_318" + original_name = "batch_norm2d_96.w_1" shape = [192] dtype = "float32" min_val = float("-2.14365") @@ -3437,6 +3756,7 @@ class Program_weight_tensor_parameter_318: class Program_weight_tensor_parameter_319: name = "parameter_319" + original_name = "conv2d_84.w_0" shape = [192, 1, 5, 5] dtype = "float32" min_val = float("-0.492824") @@ -3448,6 +3768,7 @@ class Program_weight_tensor_parameter_319: class Program_weight_tensor_parameter_320: name = "parameter_320" + original_name = "batch_norm2d_95.b_0" shape = [192] dtype = "float32" min_val = float("-0.413204") @@ -3459,6 +3780,7 @@ class Program_weight_tensor_parameter_320: class Program_weight_tensor_parameter_321: name = "parameter_321" + original_name = "batch_norm2d_95.w_0" shape = [192] dtype = "float32" min_val = float("-0.165145") @@ -3470,6 +3792,7 @@ class Program_weight_tensor_parameter_321: class Program_weight_tensor_parameter_322: name = "parameter_322" + original_name = "batch_norm2d_95.w_2" shape = [192] dtype = "float32" min_val = float("0.0025902") @@ -3481,6 +3804,7 @@ class Program_weight_tensor_parameter_322: class Program_weight_tensor_parameter_323: name = "parameter_323" + original_name = "batch_norm2d_95.w_1" shape = [192] dtype = "float32" min_val = float("-1.9607") @@ -3492,6 +3816,7 @@ class Program_weight_tensor_parameter_323: class Program_weight_tensor_parameter_324: name = "parameter_324" + original_name = "conv2d_83.w_0" shape = [192, 1, 5, 5] dtype = "float32" min_val = float("-0.466522") @@ -3503,6 +3828,7 @@ class Program_weight_tensor_parameter_324: class Program_weight_tensor_parameter_325: name = "parameter_325" + original_name = "batch_norm2d_94.b_0" shape = [192] dtype = "float32" min_val = float("-0.413204") @@ -3514,6 +3840,7 @@ class Program_weight_tensor_parameter_325: class Program_weight_tensor_parameter_326: name = "parameter_326" + original_name = "batch_norm2d_94.w_0" shape = [192] dtype = "float32" min_val = float("-0.236653") @@ -3525,6 +3852,7 @@ class Program_weight_tensor_parameter_326: class Program_weight_tensor_parameter_327: name = "parameter_327" + original_name = "batch_norm2d_94.w_2" shape = [192] dtype = "float32" min_val = float("0.00143369") @@ -3536,6 +3864,7 @@ class Program_weight_tensor_parameter_327: class Program_weight_tensor_parameter_328: name = "parameter_328" + original_name = "batch_norm2d_94.w_1" shape = [192] dtype = "float32" min_val = float("-2.30899") @@ -3547,6 +3876,7 @@ class Program_weight_tensor_parameter_328: class Program_weight_tensor_parameter_329: name = "parameter_329" + original_name = "conv2d_82.w_0" shape = [192, 1, 5, 5] dtype = "float32" min_val = float("-0.464186") @@ -3558,6 +3888,7 @@ class Program_weight_tensor_parameter_329: class Program_weight_tensor_parameter_330: name = "parameter_330" + original_name = "batch_norm2d_98.b_0" shape = [192] dtype = "float32" min_val = float("-0.413204") @@ -3569,6 +3900,7 @@ class Program_weight_tensor_parameter_330: class Program_weight_tensor_parameter_331: name = "parameter_331" + original_name = "batch_norm2d_98.w_0" shape = [192] dtype = "float32" min_val = float("-0.328232") @@ -3580,6 +3912,7 @@ class Program_weight_tensor_parameter_331: class Program_weight_tensor_parameter_332: name = "parameter_332" + original_name = "batch_norm2d_98.w_2" shape = [192] dtype = "float32" min_val = float("1.02311e-09") @@ -3591,6 +3924,7 @@ class Program_weight_tensor_parameter_332: class Program_weight_tensor_parameter_333: name = "parameter_333" + original_name = "batch_norm2d_98.w_1" shape = [192] dtype = "float32" min_val = float("-0.0828295") @@ -3602,6 +3936,7 @@ class Program_weight_tensor_parameter_333: class Program_weight_tensor_parameter_334: name = "parameter_334" + original_name = "conv2d_86.w_0" shape = [192, 1, 1, 1] dtype = "float32" min_val = float("-0.129465") @@ -3613,6 +3948,7 @@ class Program_weight_tensor_parameter_334: class Program_weight_tensor_parameter_335: name = "parameter_335" + original_name = "batch_norm2d_93.b_0" shape = [192] dtype = "float32" min_val = float("-0.39302") @@ -3624,6 +3960,7 @@ class Program_weight_tensor_parameter_335: class Program_weight_tensor_parameter_336: name = "parameter_336" + original_name = "batch_norm2d_93.w_0" shape = [192] dtype = "float32" min_val = float("-0.273974") @@ -3635,6 +3972,7 @@ class Program_weight_tensor_parameter_336: class Program_weight_tensor_parameter_337: name = "parameter_337" + original_name = "batch_norm2d_93.w_2" shape = [192] dtype = "float32" min_val = float("0.0151236") @@ -3646,6 +3984,7 @@ class Program_weight_tensor_parameter_337: class Program_weight_tensor_parameter_338: name = "parameter_338" + original_name = "batch_norm2d_93.w_1" shape = [192] dtype = "float32" min_val = float("-0.0999204") @@ -3657,6 +3996,7 @@ class Program_weight_tensor_parameter_338: class Program_weight_tensor_parameter_339: name = "parameter_339" + original_name = "batch_norm2d_92.b_0" shape = [192] dtype = "float32" min_val = float("-0.281058") @@ -3668,6 +4008,7 @@ class Program_weight_tensor_parameter_339: class Program_weight_tensor_parameter_340: name = "parameter_340" + original_name = "batch_norm2d_92.w_0" shape = [192] dtype = "float32" min_val = float("0.00359232") @@ -3679,6 +4020,7 @@ class Program_weight_tensor_parameter_340: class Program_weight_tensor_parameter_341: name = "parameter_341" + original_name = "batch_norm2d_92.w_2" shape = [192] dtype = "float32" min_val = float("0.00364222") @@ -3690,6 +4032,7 @@ class Program_weight_tensor_parameter_341: class Program_weight_tensor_parameter_342: name = "parameter_342" + original_name = "batch_norm2d_92.w_1" shape = [192] dtype = "float32" min_val = float("-0.975287") @@ -3701,6 +4044,7 @@ class Program_weight_tensor_parameter_342: class Program_weight_tensor_parameter_343: name = "parameter_343" + original_name = "conv2d_81.w_0" shape = [192, 192, 1, 1] dtype = "float32" min_val = float("-1.08004") @@ -3712,6 +4056,7 @@ class Program_weight_tensor_parameter_343: class Program_weight_tensor_parameter_344: name = "parameter_344" + original_name = "batch_norm2d_91.b_0" shape = [192] dtype = "float32" min_val = float("-0.281058") @@ -3723,6 +4068,7 @@ class Program_weight_tensor_parameter_344: class Program_weight_tensor_parameter_345: name = "parameter_345" + original_name = "batch_norm2d_91.w_0" shape = [192] dtype = "float32" min_val = float("0.0232939") @@ -3734,6 +4080,7 @@ class Program_weight_tensor_parameter_345: class Program_weight_tensor_parameter_346: name = "parameter_346" + original_name = "batch_norm2d_91.w_2" shape = [192] dtype = "float32" min_val = float("0.0195967") @@ -3745,6 +4092,7 @@ class Program_weight_tensor_parameter_346: class Program_weight_tensor_parameter_347: name = "parameter_347" + original_name = "batch_norm2d_91.w_1" shape = [192] dtype = "float32" min_val = float("-0.909193") @@ -3756,6 +4104,7 @@ class Program_weight_tensor_parameter_347: class Program_weight_tensor_parameter_348: name = "parameter_348" + original_name = "conv2d_80.w_0" shape = [192, 192, 1, 1] dtype = "float32" min_val = float("-1.04611") @@ -3767,6 +4116,7 @@ class Program_weight_tensor_parameter_348: class Program_weight_tensor_parameter_349: name = "parameter_349" + original_name = "batch_norm2d_90.b_0" shape = [192] dtype = "float32" min_val = float("-0.281058") @@ -3778,6 +4128,7 @@ class Program_weight_tensor_parameter_349: class Program_weight_tensor_parameter_350: name = "parameter_350" + original_name = "batch_norm2d_90.w_0" shape = [192] dtype = "float32" min_val = float("-0.00375904") @@ -3789,6 +4140,7 @@ class Program_weight_tensor_parameter_350: class Program_weight_tensor_parameter_351: name = "parameter_351" + original_name = "batch_norm2d_90.w_2" shape = [192] dtype = "float32" min_val = float("0.00467438") @@ -3800,6 +4152,7 @@ class Program_weight_tensor_parameter_351: class Program_weight_tensor_parameter_352: name = "parameter_352" + original_name = "batch_norm2d_90.w_1" shape = [192] dtype = "float32" min_val = float("-0.96383") @@ -3811,6 +4164,7 @@ class Program_weight_tensor_parameter_352: class Program_weight_tensor_parameter_353: name = "parameter_353" + original_name = "conv2d_79.w_0" shape = [192, 192, 1, 1] dtype = "float32" min_val = float("-1.22625") @@ -3822,6 +4176,7 @@ class Program_weight_tensor_parameter_353: class Program_weight_tensor_parameter_354: name = "parameter_354" + original_name = "batch_norm2d_89.b_0" shape = [192] dtype = "float32" min_val = float("-0.281058") @@ -3833,6 +4188,7 @@ class Program_weight_tensor_parameter_354: class Program_weight_tensor_parameter_355: name = "parameter_355" + original_name = "batch_norm2d_89.w_0" shape = [192] dtype = "float32" min_val = float("-0.00316328") @@ -3844,6 +4200,7 @@ class Program_weight_tensor_parameter_355: class Program_weight_tensor_parameter_356: name = "parameter_356" + original_name = "batch_norm2d_89.w_2" shape = [192] dtype = "float32" min_val = float("0.00469967") @@ -3855,6 +4212,7 @@ class Program_weight_tensor_parameter_356: class Program_weight_tensor_parameter_357: name = "parameter_357" + original_name = "batch_norm2d_89.w_1" shape = [192] dtype = "float32" min_val = float("-1.00874") @@ -3866,6 +4224,7 @@ class Program_weight_tensor_parameter_357: class Program_weight_tensor_parameter_358: name = "parameter_358" + original_name = "conv2d_78.w_0" shape = [192, 192, 1, 1] dtype = "float32" min_val = float("-1.38267") @@ -3877,6 +4236,7 @@ class Program_weight_tensor_parameter_358: class Program_weight_tensor_parameter_359: name = "parameter_359" + original_name = "batch_norm2d_88.b_0" shape = [192] dtype = "float32" min_val = float("-0.26102") @@ -3888,6 +4248,7 @@ class Program_weight_tensor_parameter_359: class Program_weight_tensor_parameter_360: name = "parameter_360" + original_name = "batch_norm2d_88.w_0" shape = [192] dtype = "float32" min_val = float("-0.0962581") @@ -3899,6 +4260,7 @@ class Program_weight_tensor_parameter_360: class Program_weight_tensor_parameter_361: name = "parameter_361" + original_name = "batch_norm2d_88.w_2" shape = [192] dtype = "float32" min_val = float("0.00155914") @@ -3910,6 +4272,7 @@ class Program_weight_tensor_parameter_361: class Program_weight_tensor_parameter_362: name = "parameter_362" + original_name = "batch_norm2d_88.w_1" shape = [192] dtype = "float32" min_val = float("-0.0153234") @@ -3921,6 +4284,7 @@ class Program_weight_tensor_parameter_362: class Program_weight_tensor_parameter_363: name = "parameter_363" + original_name = "batch_norm2d_86.b_0" shape = [192] dtype = "float32" min_val = float("-0.340167") @@ -3932,6 +4296,7 @@ class Program_weight_tensor_parameter_363: class Program_weight_tensor_parameter_364: name = "parameter_364" + original_name = "batch_norm2d_86.w_0" shape = [192] dtype = "float32" min_val = float("-0.076142") @@ -3943,6 +4308,7 @@ class Program_weight_tensor_parameter_364: class Program_weight_tensor_parameter_365: name = "parameter_365" + original_name = "batch_norm2d_86.w_2" shape = [192] dtype = "float32" min_val = float("0.00399197") @@ -3954,6 +4320,7 @@ class Program_weight_tensor_parameter_365: class Program_weight_tensor_parameter_366: name = "parameter_366" + original_name = "batch_norm2d_86.w_1" shape = [192] dtype = "float32" min_val = float("-2.23523") @@ -3965,6 +4332,7 @@ class Program_weight_tensor_parameter_366: class Program_weight_tensor_parameter_367: name = "parameter_367" + original_name = "conv2d_76.w_0" shape = [192, 1, 5, 5] dtype = "float32" min_val = float("-0.442368") @@ -3976,6 +4344,7 @@ class Program_weight_tensor_parameter_367: class Program_weight_tensor_parameter_368: name = "parameter_368" + original_name = "batch_norm2d_85.b_0" shape = [192] dtype = "float32" min_val = float("-0.340167") @@ -3987,6 +4356,7 @@ class Program_weight_tensor_parameter_368: class Program_weight_tensor_parameter_369: name = "parameter_369" + original_name = "batch_norm2d_85.w_0" shape = [192] dtype = "float32" min_val = float("-0.0912104") @@ -3998,6 +4368,7 @@ class Program_weight_tensor_parameter_369: class Program_weight_tensor_parameter_370: name = "parameter_370" + original_name = "batch_norm2d_85.w_2" shape = [192] dtype = "float32" min_val = float("0.00531813") @@ -4009,6 +4380,7 @@ class Program_weight_tensor_parameter_370: class Program_weight_tensor_parameter_371: name = "parameter_371" + original_name = "batch_norm2d_85.w_1" shape = [192] dtype = "float32" min_val = float("-2.23566") @@ -4020,6 +4392,7 @@ class Program_weight_tensor_parameter_371: class Program_weight_tensor_parameter_372: name = "parameter_372" + original_name = "conv2d_75.w_0" shape = [192, 1, 5, 5] dtype = "float32" min_val = float("-0.443193") @@ -4031,6 +4404,7 @@ class Program_weight_tensor_parameter_372: class Program_weight_tensor_parameter_373: name = "parameter_373" + original_name = "batch_norm2d_84.b_0" shape = [192] dtype = "float32" min_val = float("-0.340167") @@ -4042,6 +4416,7 @@ class Program_weight_tensor_parameter_373: class Program_weight_tensor_parameter_374: name = "parameter_374" + original_name = "batch_norm2d_84.w_0" shape = [192] dtype = "float32" min_val = float("-0.0736194") @@ -4053,6 +4428,7 @@ class Program_weight_tensor_parameter_374: class Program_weight_tensor_parameter_375: name = "parameter_375" + original_name = "batch_norm2d_84.w_2" shape = [192] dtype = "float32" min_val = float("0.00279182") @@ -4064,6 +4440,7 @@ class Program_weight_tensor_parameter_375: class Program_weight_tensor_parameter_376: name = "parameter_376" + original_name = "batch_norm2d_84.w_1" shape = [192] dtype = "float32" min_val = float("-2.78086") @@ -4075,6 +4452,7 @@ class Program_weight_tensor_parameter_376: class Program_weight_tensor_parameter_377: name = "parameter_377" + original_name = "conv2d_74.w_0" shape = [192, 1, 5, 5] dtype = "float32" min_val = float("-0.438884") @@ -4086,6 +4464,7 @@ class Program_weight_tensor_parameter_377: class Program_weight_tensor_parameter_378: name = "parameter_378" + original_name = "batch_norm2d_83.b_0" shape = [192] dtype = "float32" min_val = float("-0.340167") @@ -4097,6 +4476,7 @@ class Program_weight_tensor_parameter_378: class Program_weight_tensor_parameter_379: name = "parameter_379" + original_name = "batch_norm2d_83.w_0" shape = [192] dtype = "float32" min_val = float("-0.0783052") @@ -4108,6 +4488,7 @@ class Program_weight_tensor_parameter_379: class Program_weight_tensor_parameter_380: name = "parameter_380" + original_name = "batch_norm2d_83.w_2" shape = [192] dtype = "float32" min_val = float("0.00088884") @@ -4119,6 +4500,7 @@ class Program_weight_tensor_parameter_380: class Program_weight_tensor_parameter_381: name = "parameter_381" + original_name = "batch_norm2d_83.w_1" shape = [192] dtype = "float32" min_val = float("-2.24286") @@ -4130,6 +4512,7 @@ class Program_weight_tensor_parameter_381: class Program_weight_tensor_parameter_382: name = "parameter_382" + original_name = "conv2d_73.w_0" shape = [192, 1, 5, 5] dtype = "float32" min_val = float("-0.441587") @@ -4141,6 +4524,7 @@ class Program_weight_tensor_parameter_382: class Program_weight_tensor_parameter_383: name = "parameter_383" + original_name = "batch_norm2d_87.b_0" shape = [192] dtype = "float32" min_val = float("-0.340167") @@ -4152,6 +4536,7 @@ class Program_weight_tensor_parameter_383: class Program_weight_tensor_parameter_384: name = "parameter_384" + original_name = "batch_norm2d_87.w_0" shape = [192] dtype = "float32" min_val = float("-0.284923") @@ -4163,6 +4548,7 @@ class Program_weight_tensor_parameter_384: class Program_weight_tensor_parameter_385: name = "parameter_385" + original_name = "batch_norm2d_87.w_2" shape = [192] dtype = "float32" min_val = float("6.72208e-09") @@ -4174,6 +4560,7 @@ class Program_weight_tensor_parameter_385: class Program_weight_tensor_parameter_386: name = "parameter_386" + original_name = "batch_norm2d_87.w_1" shape = [192] dtype = "float32" min_val = float("-0.0819488") @@ -4185,6 +4572,7 @@ class Program_weight_tensor_parameter_386: class Program_weight_tensor_parameter_387: name = "parameter_387" + original_name = "conv2d_77.w_0" shape = [192, 1, 1, 1] dtype = "float32" min_val = float("-0.121247") @@ -4196,6 +4584,7 @@ class Program_weight_tensor_parameter_387: class Program_weight_tensor_parameter_388: name = "parameter_388" + original_name = "batch_norm2d_82.b_0" shape = [192] dtype = "float32" min_val = float("-0.313307") @@ -4207,6 +4596,7 @@ class Program_weight_tensor_parameter_388: class Program_weight_tensor_parameter_389: name = "parameter_389" + original_name = "batch_norm2d_82.w_0" shape = [192] dtype = "float32" min_val = float("-0.333448") @@ -4218,6 +4608,7 @@ class Program_weight_tensor_parameter_389: class Program_weight_tensor_parameter_390: name = "parameter_390" + original_name = "batch_norm2d_82.w_2" shape = [192] dtype = "float32" min_val = float("0.0108921") @@ -4229,6 +4620,7 @@ class Program_weight_tensor_parameter_390: class Program_weight_tensor_parameter_391: name = "parameter_391" + original_name = "batch_norm2d_82.w_1" shape = [192] dtype = "float32" min_val = float("-0.183107") @@ -4240,6 +4632,7 @@ class Program_weight_tensor_parameter_391: class Program_weight_tensor_parameter_392: name = "parameter_392" + original_name = "batch_norm2d_81.b_0" shape = [192] dtype = "float32" min_val = float("-0.349742") @@ -4251,6 +4644,7 @@ class Program_weight_tensor_parameter_392: class Program_weight_tensor_parameter_393: name = "parameter_393" + original_name = "batch_norm2d_81.w_0" shape = [192] dtype = "float32" min_val = float("0.00214999") @@ -4262,6 +4656,7 @@ class Program_weight_tensor_parameter_393: class Program_weight_tensor_parameter_394: name = "parameter_394" + original_name = "batch_norm2d_81.w_2" shape = [192] dtype = "float32" min_val = float("0.00675447") @@ -4273,6 +4668,7 @@ class Program_weight_tensor_parameter_394: class Program_weight_tensor_parameter_395: name = "parameter_395" + original_name = "batch_norm2d_81.w_1" shape = [192] dtype = "float32" min_val = float("-0.640104") @@ -4284,6 +4680,7 @@ class Program_weight_tensor_parameter_395: class Program_weight_tensor_parameter_396: name = "parameter_396" + original_name = "conv2d_72.w_0" shape = [192, 192, 1, 1] dtype = "float32" min_val = float("-0.643694") @@ -4295,6 +4692,7 @@ class Program_weight_tensor_parameter_396: class Program_weight_tensor_parameter_397: name = "parameter_397" + original_name = "batch_norm2d_80.b_0" shape = [192] dtype = "float32" min_val = float("-0.349742") @@ -4306,6 +4704,7 @@ class Program_weight_tensor_parameter_397: class Program_weight_tensor_parameter_398: name = "parameter_398" + original_name = "batch_norm2d_80.w_0" shape = [192] dtype = "float32" min_val = float("-0.00108123") @@ -4317,6 +4716,7 @@ class Program_weight_tensor_parameter_398: class Program_weight_tensor_parameter_399: name = "parameter_399" + original_name = "batch_norm2d_80.w_2" shape = [192] dtype = "float32" min_val = float("0.013718") @@ -4328,6 +4728,7 @@ class Program_weight_tensor_parameter_399: class Program_weight_tensor_parameter_400: name = "parameter_400" + original_name = "batch_norm2d_80.w_1" shape = [192] dtype = "float32" min_val = float("-0.617877") @@ -4339,6 +4740,7 @@ class Program_weight_tensor_parameter_400: class Program_weight_tensor_parameter_401: name = "parameter_401" + original_name = "conv2d_71.w_0" shape = [192, 192, 1, 1] dtype = "float32" min_val = float("-0.917863") @@ -4350,6 +4752,7 @@ class Program_weight_tensor_parameter_401: class Program_weight_tensor_parameter_402: name = "parameter_402" + original_name = "batch_norm2d_79.b_0" shape = [192] dtype = "float32" min_val = float("-0.349742") @@ -4361,6 +4764,7 @@ class Program_weight_tensor_parameter_402: class Program_weight_tensor_parameter_403: name = "parameter_403" + original_name = "batch_norm2d_79.w_0" shape = [192] dtype = "float32" min_val = float("-0.00113501") @@ -4372,6 +4776,7 @@ class Program_weight_tensor_parameter_403: class Program_weight_tensor_parameter_404: name = "parameter_404" + original_name = "batch_norm2d_79.w_2" shape = [192] dtype = "float32" min_val = float("0.00749564") @@ -4383,6 +4788,7 @@ class Program_weight_tensor_parameter_404: class Program_weight_tensor_parameter_405: name = "parameter_405" + original_name = "batch_norm2d_79.w_1" shape = [192] dtype = "float32" min_val = float("-0.609732") @@ -4394,6 +4800,7 @@ class Program_weight_tensor_parameter_405: class Program_weight_tensor_parameter_406: name = "parameter_406" + original_name = "conv2d_70.w_0" shape = [192, 192, 1, 1] dtype = "float32" min_val = float("-0.925507") @@ -4405,6 +4812,7 @@ class Program_weight_tensor_parameter_406: class Program_weight_tensor_parameter_407: name = "parameter_407" + original_name = "batch_norm2d_78.b_0" shape = [192] dtype = "float32" min_val = float("-0.349742") @@ -4416,6 +4824,7 @@ class Program_weight_tensor_parameter_407: class Program_weight_tensor_parameter_408: name = "parameter_408" + original_name = "batch_norm2d_78.w_0" shape = [192] dtype = "float32" min_val = float("-0.000522815") @@ -4427,6 +4836,7 @@ class Program_weight_tensor_parameter_408: class Program_weight_tensor_parameter_409: name = "parameter_409" + original_name = "batch_norm2d_78.w_2" shape = [192] dtype = "float32" min_val = float("0.0172134") @@ -4438,6 +4848,7 @@ class Program_weight_tensor_parameter_409: class Program_weight_tensor_parameter_410: name = "parameter_410" + original_name = "batch_norm2d_78.w_1" shape = [192] dtype = "float32" min_val = float("-0.614245") @@ -4449,6 +4860,7 @@ class Program_weight_tensor_parameter_410: class Program_weight_tensor_parameter_411: name = "parameter_411" + original_name = "conv2d_69.w_0" shape = [192, 192, 1, 1] dtype = "float32" min_val = float("-0.879897") @@ -4460,6 +4872,7 @@ class Program_weight_tensor_parameter_411: class Program_weight_tensor_parameter_412: name = "parameter_412" + original_name = "batch_norm2d_77.b_0" shape = [192] dtype = "float32" min_val = float("-0.31027") @@ -4471,6 +4884,7 @@ class Program_weight_tensor_parameter_412: class Program_weight_tensor_parameter_413: name = "parameter_413" + original_name = "batch_norm2d_77.w_0" shape = [192] dtype = "float32" min_val = float("-0.169074") @@ -4482,6 +4896,7 @@ class Program_weight_tensor_parameter_413: class Program_weight_tensor_parameter_414: name = "parameter_414" + original_name = "batch_norm2d_77.w_2" shape = [192] dtype = "float32" min_val = float("0.00245429") @@ -4493,6 +4908,7 @@ class Program_weight_tensor_parameter_414: class Program_weight_tensor_parameter_415: name = "parameter_415" + original_name = "batch_norm2d_77.w_1" shape = [192] dtype = "float32" min_val = float("-0.00665941") @@ -4504,6 +4920,7 @@ class Program_weight_tensor_parameter_415: class Program_weight_tensor_parameter_416: name = "parameter_416" + original_name = "batch_norm2d_75.b_0" shape = [192] dtype = "float32" min_val = float("-0.319308") @@ -4515,6 +4932,7 @@ class Program_weight_tensor_parameter_416: class Program_weight_tensor_parameter_417: name = "parameter_417" + original_name = "batch_norm2d_75.w_0" shape = [192] dtype = "float32" min_val = float("-0.103055") @@ -4526,6 +4944,7 @@ class Program_weight_tensor_parameter_417: class Program_weight_tensor_parameter_418: name = "parameter_418" + original_name = "batch_norm2d_75.w_2" shape = [192] dtype = "float32" min_val = float("0.00298544") @@ -4537,6 +4956,7 @@ class Program_weight_tensor_parameter_418: class Program_weight_tensor_parameter_419: name = "parameter_419" + original_name = "batch_norm2d_75.w_1" shape = [192] dtype = "float32" min_val = float("-2.30607") @@ -4548,6 +4968,7 @@ class Program_weight_tensor_parameter_419: class Program_weight_tensor_parameter_420: name = "parameter_420" + original_name = "conv2d_67.w_0" shape = [192, 1, 5, 5] dtype = "float32" min_val = float("-0.653657") @@ -4559,6 +4980,7 @@ class Program_weight_tensor_parameter_420: class Program_weight_tensor_parameter_421: name = "parameter_421" + original_name = "batch_norm2d_74.b_0" shape = [192] dtype = "float32" min_val = float("-0.319308") @@ -4570,6 +4992,7 @@ class Program_weight_tensor_parameter_421: class Program_weight_tensor_parameter_422: name = "parameter_422" + original_name = "batch_norm2d_74.w_0" shape = [192] dtype = "float32" min_val = float("-0.115907") @@ -4581,6 +5004,7 @@ class Program_weight_tensor_parameter_422: class Program_weight_tensor_parameter_423: name = "parameter_423" + original_name = "batch_norm2d_74.w_2" shape = [192] dtype = "float32" min_val = float("0.00282473") @@ -4592,6 +5016,7 @@ class Program_weight_tensor_parameter_423: class Program_weight_tensor_parameter_424: name = "parameter_424" + original_name = "batch_norm2d_74.w_1" shape = [192] dtype = "float32" min_val = float("-2.30584") @@ -4603,6 +5028,7 @@ class Program_weight_tensor_parameter_424: class Program_weight_tensor_parameter_425: name = "parameter_425" + original_name = "conv2d_66.w_0" shape = [192, 1, 5, 5] dtype = "float32" min_val = float("-0.662079") @@ -4614,6 +5040,7 @@ class Program_weight_tensor_parameter_425: class Program_weight_tensor_parameter_426: name = "parameter_426" + original_name = "batch_norm2d_73.b_0" shape = [192] dtype = "float32" min_val = float("-0.319308") @@ -4625,6 +5052,7 @@ class Program_weight_tensor_parameter_426: class Program_weight_tensor_parameter_427: name = "parameter_427" + original_name = "batch_norm2d_73.w_0" shape = [192] dtype = "float32" min_val = float("-0.152383") @@ -4636,6 +5064,7 @@ class Program_weight_tensor_parameter_427: class Program_weight_tensor_parameter_428: name = "parameter_428" + original_name = "batch_norm2d_73.w_2" shape = [192] dtype = "float32" min_val = float("0.00329976") @@ -4647,6 +5076,7 @@ class Program_weight_tensor_parameter_428: class Program_weight_tensor_parameter_429: name = "parameter_429" + original_name = "batch_norm2d_73.w_1" shape = [192] dtype = "float32" min_val = float("-2.30501") @@ -4658,6 +5088,7 @@ class Program_weight_tensor_parameter_429: class Program_weight_tensor_parameter_430: name = "parameter_430" + original_name = "conv2d_65.w_0" shape = [192, 1, 5, 5] dtype = "float32" min_val = float("-0.572808") @@ -4669,6 +5100,7 @@ class Program_weight_tensor_parameter_430: class Program_weight_tensor_parameter_431: name = "parameter_431" + original_name = "batch_norm2d_72.b_0" shape = [192] dtype = "float32" min_val = float("-0.319308") @@ -4680,6 +5112,7 @@ class Program_weight_tensor_parameter_431: class Program_weight_tensor_parameter_432: name = "parameter_432" + original_name = "batch_norm2d_72.w_0" shape = [192] dtype = "float32" min_val = float("-0.117028") @@ -4691,6 +5124,7 @@ class Program_weight_tensor_parameter_432: class Program_weight_tensor_parameter_433: name = "parameter_433" + original_name = "batch_norm2d_72.w_2" shape = [192] dtype = "float32" min_val = float("0.00577935") @@ -4702,6 +5136,7 @@ class Program_weight_tensor_parameter_433: class Program_weight_tensor_parameter_434: name = "parameter_434" + original_name = "batch_norm2d_72.w_1" shape = [192] dtype = "float32" min_val = float("-2.28919") @@ -4713,6 +5148,7 @@ class Program_weight_tensor_parameter_434: class Program_weight_tensor_parameter_435: name = "parameter_435" + original_name = "conv2d_64.w_0" shape = [192, 1, 5, 5] dtype = "float32" min_val = float("-0.651471") @@ -4724,6 +5160,7 @@ class Program_weight_tensor_parameter_435: class Program_weight_tensor_parameter_436: name = "parameter_436" + original_name = "batch_norm2d_76.b_0" shape = [192] dtype = "float32" min_val = float("-0.319308") @@ -4735,6 +5172,7 @@ class Program_weight_tensor_parameter_436: class Program_weight_tensor_parameter_437: name = "parameter_437" + original_name = "batch_norm2d_76.w_0" shape = [192] dtype = "float32" min_val = float("-0.362885") @@ -4746,6 +5184,7 @@ class Program_weight_tensor_parameter_437: class Program_weight_tensor_parameter_438: name = "parameter_438" + original_name = "batch_norm2d_76.w_2" shape = [192] dtype = "float32" min_val = float("2.96811e-08") @@ -4757,6 +5196,7 @@ class Program_weight_tensor_parameter_438: class Program_weight_tensor_parameter_439: name = "parameter_439" + original_name = "batch_norm2d_76.w_1" shape = [192] dtype = "float32" min_val = float("-0.123637") @@ -4768,6 +5208,7 @@ class Program_weight_tensor_parameter_439: class Program_weight_tensor_parameter_440: name = "parameter_440" + original_name = "conv2d_68.w_0" shape = [192, 1, 1, 1] dtype = "float32" min_val = float("-0.118954") @@ -4779,6 +5220,7 @@ class Program_weight_tensor_parameter_440: class Program_weight_tensor_parameter_441: name = "parameter_441" + original_name = "batch_norm2d_71.b_0" shape = [192] dtype = "float32" min_val = float("-0.290638") @@ -4790,6 +5232,7 @@ class Program_weight_tensor_parameter_441: class Program_weight_tensor_parameter_442: name = "parameter_442" + original_name = "batch_norm2d_71.w_0" shape = [192] dtype = "float32" min_val = float("-0.249002") @@ -4801,6 +5244,7 @@ class Program_weight_tensor_parameter_442: class Program_weight_tensor_parameter_443: name = "parameter_443" + original_name = "batch_norm2d_71.w_2" shape = [192] dtype = "float32" min_val = float("0.0404722") @@ -4812,6 +5256,7 @@ class Program_weight_tensor_parameter_443: class Program_weight_tensor_parameter_444: name = "parameter_444" + original_name = "batch_norm2d_71.w_1" shape = [192] dtype = "float32" min_val = float("-0.326944") @@ -4823,6 +5268,7 @@ class Program_weight_tensor_parameter_444: class Program_weight_tensor_parameter_445: name = "parameter_445" + original_name = "batch_norm2d_70.b_0" shape = [192] dtype = "float32" min_val = float("-0.344441") @@ -4834,6 +5280,7 @@ class Program_weight_tensor_parameter_445: class Program_weight_tensor_parameter_446: name = "parameter_446" + original_name = "batch_norm2d_70.w_0" shape = [192] dtype = "float32" min_val = float("-0.00352983") @@ -4845,6 +5292,7 @@ class Program_weight_tensor_parameter_446: class Program_weight_tensor_parameter_447: name = "parameter_447" + original_name = "batch_norm2d_70.w_2" shape = [192] dtype = "float32" min_val = float("0.00437324") @@ -4856,6 +5304,7 @@ class Program_weight_tensor_parameter_447: class Program_weight_tensor_parameter_448: name = "parameter_448" + original_name = "batch_norm2d_70.w_1" shape = [192] dtype = "float32" min_val = float("-0.599229") @@ -4867,6 +5316,7 @@ class Program_weight_tensor_parameter_448: class Program_weight_tensor_parameter_449: name = "parameter_449" + original_name = "conv2d_63.w_0" shape = [192, 192, 1, 1] dtype = "float32" min_val = float("-0.94821") @@ -4878,6 +5328,7 @@ class Program_weight_tensor_parameter_449: class Program_weight_tensor_parameter_450: name = "parameter_450" + original_name = "batch_norm2d_69.b_0" shape = [192] dtype = "float32" min_val = float("-0.344441") @@ -4889,6 +5340,7 @@ class Program_weight_tensor_parameter_450: class Program_weight_tensor_parameter_451: name = "parameter_451" + original_name = "batch_norm2d_69.w_0" shape = [192] dtype = "float32" min_val = float("-0.00158593") @@ -4900,6 +5352,7 @@ class Program_weight_tensor_parameter_451: class Program_weight_tensor_parameter_452: name = "parameter_452" + original_name = "batch_norm2d_69.w_2" shape = [192] dtype = "float32" min_val = float("0.00282787") @@ -4911,6 +5364,7 @@ class Program_weight_tensor_parameter_452: class Program_weight_tensor_parameter_453: name = "parameter_453" + original_name = "batch_norm2d_69.w_1" shape = [192] dtype = "float32" min_val = float("-0.552572") @@ -4922,6 +5376,7 @@ class Program_weight_tensor_parameter_453: class Program_weight_tensor_parameter_454: name = "parameter_454" + original_name = "conv2d_62.w_0" shape = [192, 192, 1, 1] dtype = "float32" min_val = float("-1.34154") @@ -4933,6 +5388,7 @@ class Program_weight_tensor_parameter_454: class Program_weight_tensor_parameter_455: name = "parameter_455" + original_name = "batch_norm2d_68.b_0" shape = [192] dtype = "float32" min_val = float("-0.344441") @@ -4944,6 +5400,7 @@ class Program_weight_tensor_parameter_455: class Program_weight_tensor_parameter_456: name = "parameter_456" + original_name = "batch_norm2d_68.w_0" shape = [192] dtype = "float32" min_val = float("0.000603294") @@ -4955,6 +5412,7 @@ class Program_weight_tensor_parameter_456: class Program_weight_tensor_parameter_457: name = "parameter_457" + original_name = "batch_norm2d_68.w_2" shape = [192] dtype = "float32" min_val = float("0.00708626") @@ -4966,6 +5424,7 @@ class Program_weight_tensor_parameter_457: class Program_weight_tensor_parameter_458: name = "parameter_458" + original_name = "batch_norm2d_68.w_1" shape = [192] dtype = "float32" min_val = float("-0.56978") @@ -4977,6 +5436,7 @@ class Program_weight_tensor_parameter_458: class Program_weight_tensor_parameter_459: name = "parameter_459" + original_name = "conv2d_61.w_0" shape = [192, 192, 1, 1] dtype = "float32" min_val = float("-0.966182") @@ -4988,6 +5448,7 @@ class Program_weight_tensor_parameter_459: class Program_weight_tensor_parameter_460: name = "parameter_460" + original_name = "batch_norm2d_67.b_0" shape = [192] dtype = "float32" min_val = float("-0.344441") @@ -4999,6 +5460,7 @@ class Program_weight_tensor_parameter_460: class Program_weight_tensor_parameter_461: name = "parameter_461" + original_name = "batch_norm2d_67.w_0" shape = [192] dtype = "float32" min_val = float("-0.00156661") @@ -5010,6 +5472,7 @@ class Program_weight_tensor_parameter_461: class Program_weight_tensor_parameter_462: name = "parameter_462" + original_name = "batch_norm2d_67.w_2" shape = [192] dtype = "float32" min_val = float("0.00284229") @@ -5021,6 +5484,7 @@ class Program_weight_tensor_parameter_462: class Program_weight_tensor_parameter_463: name = "parameter_463" + original_name = "batch_norm2d_67.w_1" shape = [192] dtype = "float32" min_val = float("-0.530726") @@ -5032,6 +5496,7 @@ class Program_weight_tensor_parameter_463: class Program_weight_tensor_parameter_464: name = "parameter_464" + original_name = "conv2d_60.w_0" shape = [192, 192, 1, 1] dtype = "float32" min_val = float("-1.26732") @@ -5043,6 +5508,7 @@ class Program_weight_tensor_parameter_464: class Program_weight_tensor_parameter_465: name = "parameter_465" + original_name = "batch_norm2d_66.b_0" shape = [192] dtype = "float32" min_val = float("-0.307977") @@ -5054,6 +5520,7 @@ class Program_weight_tensor_parameter_465: class Program_weight_tensor_parameter_466: name = "parameter_466" + original_name = "batch_norm2d_66.w_0" shape = [192] dtype = "float32" min_val = float("-0.173151") @@ -5065,6 +5532,7 @@ class Program_weight_tensor_parameter_466: class Program_weight_tensor_parameter_467: name = "parameter_467" + original_name = "batch_norm2d_66.w_2" shape = [192] dtype = "float32" min_val = float("0.000742376") @@ -5076,6 +5544,7 @@ class Program_weight_tensor_parameter_467: class Program_weight_tensor_parameter_468: name = "parameter_468" + original_name = "batch_norm2d_66.w_1" shape = [192] dtype = "float32" min_val = float("-0.016225") @@ -5087,6 +5556,7 @@ class Program_weight_tensor_parameter_468: class Program_weight_tensor_parameter_469: name = "parameter_469" + original_name = "batch_norm2d_64.b_0" shape = [192] dtype = "float32" min_val = float("-0.292671") @@ -5098,6 +5568,7 @@ class Program_weight_tensor_parameter_469: class Program_weight_tensor_parameter_470: name = "parameter_470" + original_name = "batch_norm2d_64.w_0" shape = [192] dtype = "float32" min_val = float("-0.084764") @@ -5109,6 +5580,7 @@ class Program_weight_tensor_parameter_470: class Program_weight_tensor_parameter_471: name = "parameter_471" + original_name = "batch_norm2d_64.w_2" shape = [192] dtype = "float32" min_val = float("0.00620152") @@ -5120,6 +5592,7 @@ class Program_weight_tensor_parameter_471: class Program_weight_tensor_parameter_472: name = "parameter_472" + original_name = "batch_norm2d_64.w_1" shape = [192] dtype = "float32" min_val = float("-2.56617") @@ -5131,6 +5604,7 @@ class Program_weight_tensor_parameter_472: class Program_weight_tensor_parameter_473: name = "parameter_473" + original_name = "conv2d_58.w_0" shape = [192, 1, 5, 5] dtype = "float32" min_val = float("-0.577181") @@ -5142,6 +5616,7 @@ class Program_weight_tensor_parameter_473: class Program_weight_tensor_parameter_474: name = "parameter_474" + original_name = "batch_norm2d_63.b_0" shape = [192] dtype = "float32" min_val = float("-0.292671") @@ -5153,6 +5628,7 @@ class Program_weight_tensor_parameter_474: class Program_weight_tensor_parameter_475: name = "parameter_475" + original_name = "batch_norm2d_63.w_0" shape = [192] dtype = "float32" min_val = float("-0.0758604") @@ -5164,6 +5640,7 @@ class Program_weight_tensor_parameter_475: class Program_weight_tensor_parameter_476: name = "parameter_476" + original_name = "batch_norm2d_63.w_2" shape = [192] dtype = "float32" min_val = float("0.0054682") @@ -5175,6 +5652,7 @@ class Program_weight_tensor_parameter_476: class Program_weight_tensor_parameter_477: name = "parameter_477" + original_name = "batch_norm2d_63.w_1" shape = [192] dtype = "float32" min_val = float("-2.57785") @@ -5186,6 +5664,7 @@ class Program_weight_tensor_parameter_477: class Program_weight_tensor_parameter_478: name = "parameter_478" + original_name = "conv2d_57.w_0" shape = [192, 1, 5, 5] dtype = "float32" min_val = float("-0.582548") @@ -5197,6 +5676,7 @@ class Program_weight_tensor_parameter_478: class Program_weight_tensor_parameter_479: name = "parameter_479" + original_name = "batch_norm2d_62.b_0" shape = [192] dtype = "float32" min_val = float("-0.292671") @@ -5208,6 +5688,7 @@ class Program_weight_tensor_parameter_479: class Program_weight_tensor_parameter_480: name = "parameter_480" + original_name = "batch_norm2d_62.w_0" shape = [192] dtype = "float32" min_val = float("-0.12214") @@ -5219,6 +5700,7 @@ class Program_weight_tensor_parameter_480: class Program_weight_tensor_parameter_481: name = "parameter_481" + original_name = "batch_norm2d_62.w_2" shape = [192] dtype = "float32" min_val = float("0.00529705") @@ -5230,6 +5712,7 @@ class Program_weight_tensor_parameter_481: class Program_weight_tensor_parameter_482: name = "parameter_482" + original_name = "batch_norm2d_62.w_1" shape = [192] dtype = "float32" min_val = float("-2.57168") @@ -5241,6 +5724,7 @@ class Program_weight_tensor_parameter_482: class Program_weight_tensor_parameter_483: name = "parameter_483" + original_name = "conv2d_56.w_0" shape = [192, 1, 5, 5] dtype = "float32" min_val = float("-0.628033") @@ -5252,6 +5736,7 @@ class Program_weight_tensor_parameter_483: class Program_weight_tensor_parameter_484: name = "parameter_484" + original_name = "batch_norm2d_61.b_0" shape = [192] dtype = "float32" min_val = float("-0.292671") @@ -5263,6 +5748,7 @@ class Program_weight_tensor_parameter_484: class Program_weight_tensor_parameter_485: name = "parameter_485" + original_name = "batch_norm2d_61.w_0" shape = [192] dtype = "float32" min_val = float("-0.0858055") @@ -5274,6 +5760,7 @@ class Program_weight_tensor_parameter_485: class Program_weight_tensor_parameter_486: name = "parameter_486" + original_name = "batch_norm2d_61.w_2" shape = [192] dtype = "float32" min_val = float("0.00778062") @@ -5285,6 +5772,7 @@ class Program_weight_tensor_parameter_486: class Program_weight_tensor_parameter_487: name = "parameter_487" + original_name = "batch_norm2d_61.w_1" shape = [192] dtype = "float32" min_val = float("-2.56683") @@ -5296,6 +5784,7 @@ class Program_weight_tensor_parameter_487: class Program_weight_tensor_parameter_488: name = "parameter_488" + original_name = "conv2d_55.w_0" shape = [192, 1, 5, 5] dtype = "float32" min_val = float("-0.737106") @@ -5307,6 +5796,7 @@ class Program_weight_tensor_parameter_488: class Program_weight_tensor_parameter_489: name = "parameter_489" + original_name = "batch_norm2d_65.b_0" shape = [192] dtype = "float32" min_val = float("-0.292671") @@ -5318,6 +5808,7 @@ class Program_weight_tensor_parameter_489: class Program_weight_tensor_parameter_490: name = "parameter_490" + original_name = "batch_norm2d_65.w_0" shape = [192] dtype = "float32" min_val = float("-0.420175") @@ -5329,6 +5820,7 @@ class Program_weight_tensor_parameter_490: class Program_weight_tensor_parameter_491: name = "parameter_491" + original_name = "batch_norm2d_65.w_2" shape = [192] dtype = "float32" min_val = float("3.38127e-09") @@ -5340,6 +5832,7 @@ class Program_weight_tensor_parameter_491: class Program_weight_tensor_parameter_492: name = "parameter_492" + original_name = "batch_norm2d_65.w_1" shape = [192] dtype = "float32" min_val = float("-0.0938896") @@ -5351,6 +5844,7 @@ class Program_weight_tensor_parameter_492: class Program_weight_tensor_parameter_493: name = "parameter_493" + original_name = "conv2d_59.w_0" shape = [192, 1, 1, 1] dtype = "float32" min_val = float("-0.119602") @@ -5362,6 +5856,7 @@ class Program_weight_tensor_parameter_493: class Program_weight_tensor_parameter_494: name = "parameter_494" + original_name = "batch_norm2d_60.b_0" shape = [192] dtype = "float32" min_val = float("-0.261161") @@ -5373,6 +5868,7 @@ class Program_weight_tensor_parameter_494: class Program_weight_tensor_parameter_495: name = "parameter_495" + original_name = "batch_norm2d_60.w_0" shape = [192] dtype = "float32" min_val = float("-0.493433") @@ -5384,6 +5880,7 @@ class Program_weight_tensor_parameter_495: class Program_weight_tensor_parameter_496: name = "parameter_496" + original_name = "batch_norm2d_60.w_2" shape = [192] dtype = "float32" min_val = float("0.0183105") @@ -5395,6 +5892,7 @@ class Program_weight_tensor_parameter_496: class Program_weight_tensor_parameter_497: name = "parameter_497" + original_name = "batch_norm2d_60.w_1" shape = [192] dtype = "float32" min_val = float("-0.775015") @@ -5406,6 +5904,7 @@ class Program_weight_tensor_parameter_497: class Program_weight_tensor_parameter_498: name = "parameter_498" + original_name = "batch_norm2d_59.b_0" shape = [192] dtype = "float32" min_val = float("-0.600327") @@ -5417,6 +5916,7 @@ class Program_weight_tensor_parameter_498: class Program_weight_tensor_parameter_499: name = "parameter_499" + original_name = "batch_norm2d_59.w_0" shape = [192] dtype = "float32" min_val = float("0.00482509") @@ -5428,6 +5928,7 @@ class Program_weight_tensor_parameter_499: class Program_weight_tensor_parameter_500: name = "parameter_500" + original_name = "batch_norm2d_59.w_2" shape = [192] dtype = "float32" min_val = float("0.0160027") @@ -5439,6 +5940,7 @@ class Program_weight_tensor_parameter_500: class Program_weight_tensor_parameter_501: name = "parameter_501" + original_name = "batch_norm2d_59.w_1" shape = [192] dtype = "float32" min_val = float("-0.00461249") @@ -5450,6 +5952,7 @@ class Program_weight_tensor_parameter_501: class Program_weight_tensor_parameter_502: name = "parameter_502" + original_name = "conv2d_54.w_0" shape = [192, 96, 1, 1] dtype = "float32" min_val = float("-1.25071") @@ -5461,6 +5964,7 @@ class Program_weight_tensor_parameter_502: class Program_weight_tensor_parameter_503: name = "parameter_503" + original_name = "batch_norm2d_58.b_0" shape = [192] dtype = "float32" min_val = float("-0.600327") @@ -5472,6 +5976,7 @@ class Program_weight_tensor_parameter_503: class Program_weight_tensor_parameter_504: name = "parameter_504" + original_name = "batch_norm2d_58.w_0" shape = [192] dtype = "float32" min_val = float("-0.00373012") @@ -5483,6 +5988,7 @@ class Program_weight_tensor_parameter_504: class Program_weight_tensor_parameter_505: name = "parameter_505" + original_name = "batch_norm2d_58.w_2" shape = [192] dtype = "float32" min_val = float("0.00953556") @@ -5494,6 +6000,7 @@ class Program_weight_tensor_parameter_505: class Program_weight_tensor_parameter_506: name = "parameter_506" + original_name = "batch_norm2d_58.w_1" shape = [192] dtype = "float32" min_val = float("-0.00442949") @@ -5505,6 +6012,7 @@ class Program_weight_tensor_parameter_506: class Program_weight_tensor_parameter_507: name = "parameter_507" + original_name = "conv2d_53.w_0" shape = [192, 96, 1, 1] dtype = "float32" min_val = float("-1.13422") @@ -5516,6 +6024,7 @@ class Program_weight_tensor_parameter_507: class Program_weight_tensor_parameter_508: name = "parameter_508" + original_name = "batch_norm2d_57.b_0" shape = [192] dtype = "float32" min_val = float("-0.600327") @@ -5527,6 +6036,7 @@ class Program_weight_tensor_parameter_508: class Program_weight_tensor_parameter_509: name = "parameter_509" + original_name = "batch_norm2d_57.w_0" shape = [192] dtype = "float32" min_val = float("0.00151174") @@ -5538,6 +6048,7 @@ class Program_weight_tensor_parameter_509: class Program_weight_tensor_parameter_510: name = "parameter_510" + original_name = "batch_norm2d_57.w_2" shape = [192] dtype = "float32" min_val = float("0.0188113") @@ -5549,6 +6060,7 @@ class Program_weight_tensor_parameter_510: class Program_weight_tensor_parameter_511: name = "parameter_511" + original_name = "batch_norm2d_57.w_1" shape = [192] dtype = "float32" min_val = float("-0.00448969") @@ -5560,6 +6072,7 @@ class Program_weight_tensor_parameter_511: class Program_weight_tensor_parameter_512: name = "parameter_512" + original_name = "conv2d_52.w_0" shape = [192, 96, 1, 1] dtype = "float32" min_val = float("-1.05159") @@ -5571,6 +6084,7 @@ class Program_weight_tensor_parameter_512: class Program_weight_tensor_parameter_513: name = "parameter_513" + original_name = "batch_norm2d_56.b_0" shape = [192] dtype = "float32" min_val = float("-0.600327") @@ -5582,6 +6096,7 @@ class Program_weight_tensor_parameter_513: class Program_weight_tensor_parameter_514: name = "parameter_514" + original_name = "batch_norm2d_56.w_0" shape = [192] dtype = "float32" min_val = float("-0.00307998") @@ -5593,6 +6108,7 @@ class Program_weight_tensor_parameter_514: class Program_weight_tensor_parameter_515: name = "parameter_515" + original_name = "batch_norm2d_56.w_2" shape = [192] dtype = "float32" min_val = float("0.0122514") @@ -5604,6 +6120,7 @@ class Program_weight_tensor_parameter_515: class Program_weight_tensor_parameter_516: name = "parameter_516" + original_name = "batch_norm2d_56.w_1" shape = [192] dtype = "float32" min_val = float("-0.00480251") @@ -5615,6 +6132,7 @@ class Program_weight_tensor_parameter_516: class Program_weight_tensor_parameter_517: name = "parameter_517" + original_name = "conv2d_51.w_0" shape = [192, 96, 1, 1] dtype = "float32" min_val = float("-1.00798") @@ -5626,6 +6144,7 @@ class Program_weight_tensor_parameter_517: class Program_weight_tensor_parameter_518: name = "parameter_518" + original_name = "batch_norm2d_54.b_0" shape = [96] dtype = "float32" min_val = float("-0.00487491") @@ -5637,6 +6156,7 @@ class Program_weight_tensor_parameter_518: class Program_weight_tensor_parameter_519: name = "parameter_519" + original_name = "batch_norm2d_54.w_0" shape = [96] dtype = "float32" min_val = float("-0.213919") @@ -5648,6 +6168,7 @@ class Program_weight_tensor_parameter_519: class Program_weight_tensor_parameter_520: name = "parameter_520" + original_name = "batch_norm2d_54.w_2" shape = [96] dtype = "float32" min_val = float("0.000760404") @@ -5659,6 +6180,7 @@ class Program_weight_tensor_parameter_520: class Program_weight_tensor_parameter_521: name = "parameter_521" + original_name = "batch_norm2d_54.w_1" shape = [96] dtype = "float32" min_val = float("-2.39336") @@ -5670,6 +6192,7 @@ class Program_weight_tensor_parameter_521: class Program_weight_tensor_parameter_522: name = "parameter_522" + original_name = "conv2d_49.w_0" shape = [96, 1, 3, 3] dtype = "float32" min_val = float("-0.275574") @@ -5681,6 +6204,7 @@ class Program_weight_tensor_parameter_522: class Program_weight_tensor_parameter_523: name = "parameter_523" + original_name = "batch_norm2d_53.b_0" shape = [96] dtype = "float32" min_val = float("-0.00487491") @@ -5692,6 +6216,7 @@ class Program_weight_tensor_parameter_523: class Program_weight_tensor_parameter_524: name = "parameter_524" + original_name = "batch_norm2d_53.w_0" shape = [96] dtype = "float32" min_val = float("-0.0319166") @@ -5703,6 +6228,7 @@ class Program_weight_tensor_parameter_524: class Program_weight_tensor_parameter_525: name = "parameter_525" + original_name = "batch_norm2d_53.w_2" shape = [96] dtype = "float32" min_val = float("0.00058296") @@ -5714,6 +6240,7 @@ class Program_weight_tensor_parameter_525: class Program_weight_tensor_parameter_526: name = "parameter_526" + original_name = "batch_norm2d_53.w_1" shape = [96] dtype = "float32" min_val = float("-2.39625") @@ -5725,6 +6252,7 @@ class Program_weight_tensor_parameter_526: class Program_weight_tensor_parameter_527: name = "parameter_527" + original_name = "conv2d_48.w_0" shape = [96, 1, 3, 3] dtype = "float32" min_val = float("-0.278141") @@ -5736,6 +6264,7 @@ class Program_weight_tensor_parameter_527: class Program_weight_tensor_parameter_528: name = "parameter_528" + original_name = "batch_norm2d_52.b_0" shape = [96] dtype = "float32" min_val = float("-0.00487491") @@ -5747,6 +6276,7 @@ class Program_weight_tensor_parameter_528: class Program_weight_tensor_parameter_529: name = "parameter_529" + original_name = "batch_norm2d_52.w_0" shape = [96] dtype = "float32" min_val = float("0.10515") @@ -5758,6 +6288,7 @@ class Program_weight_tensor_parameter_529: class Program_weight_tensor_parameter_530: name = "parameter_530" + original_name = "batch_norm2d_52.w_2" shape = [96] dtype = "float32" min_val = float("0.000402816") @@ -5769,6 +6300,7 @@ class Program_weight_tensor_parameter_530: class Program_weight_tensor_parameter_531: name = "parameter_531" + original_name = "batch_norm2d_52.w_1" shape = [96] dtype = "float32" min_val = float("-2.39868") @@ -5780,6 +6312,7 @@ class Program_weight_tensor_parameter_531: class Program_weight_tensor_parameter_532: name = "parameter_532" + original_name = "conv2d_47.w_0" shape = [96, 1, 3, 3] dtype = "float32" min_val = float("-0.27058") @@ -5791,6 +6324,7 @@ class Program_weight_tensor_parameter_532: class Program_weight_tensor_parameter_533: name = "parameter_533" + original_name = "batch_norm2d_51.b_0" shape = [96] dtype = "float32" min_val = float("-0.00487491") @@ -5802,6 +6336,7 @@ class Program_weight_tensor_parameter_533: class Program_weight_tensor_parameter_534: name = "parameter_534" + original_name = "batch_norm2d_51.w_0" shape = [96] dtype = "float32" min_val = float("-0.0200657") @@ -5813,6 +6348,7 @@ class Program_weight_tensor_parameter_534: class Program_weight_tensor_parameter_535: name = "parameter_535" + original_name = "batch_norm2d_51.w_2" shape = [96] dtype = "float32" min_val = float("0.00058465") @@ -5824,6 +6360,7 @@ class Program_weight_tensor_parameter_535: class Program_weight_tensor_parameter_536: name = "parameter_536" + original_name = "batch_norm2d_51.w_1" shape = [96] dtype = "float32" min_val = float("-2.39635") @@ -5835,6 +6372,7 @@ class Program_weight_tensor_parameter_536: class Program_weight_tensor_parameter_537: name = "parameter_537" + original_name = "conv2d_46.w_0" shape = [96, 1, 3, 3] dtype = "float32" min_val = float("-0.28698") @@ -5846,6 +6384,7 @@ class Program_weight_tensor_parameter_537: class Program_weight_tensor_parameter_538: name = "parameter_538" + original_name = "batch_norm2d_55.b_0" shape = [96] dtype = "float32" min_val = float("-0.00487491") @@ -5857,6 +6396,7 @@ class Program_weight_tensor_parameter_538: class Program_weight_tensor_parameter_539: name = "parameter_539" + original_name = "batch_norm2d_55.w_0" shape = [96] dtype = "float32" min_val = float("-0.511124") @@ -5868,6 +6408,7 @@ class Program_weight_tensor_parameter_539: class Program_weight_tensor_parameter_540: name = "parameter_540" + original_name = "batch_norm2d_55.w_2" shape = [96] dtype = "float32" min_val = float("7.04499e-10") @@ -5879,6 +6420,7 @@ class Program_weight_tensor_parameter_540: class Program_weight_tensor_parameter_541: name = "parameter_541" + original_name = "batch_norm2d_55.w_1" shape = [96] dtype = "float32" min_val = float("-0.197718") @@ -5890,6 +6432,7 @@ class Program_weight_tensor_parameter_541: class Program_weight_tensor_parameter_542: name = "parameter_542" + original_name = "conv2d_50.w_0" shape = [96, 1, 1, 1] dtype = "float32" min_val = float("-0.133818") @@ -5901,6 +6444,7 @@ class Program_weight_tensor_parameter_542: class Program_weight_tensor_parameter_543: name = "parameter_543" + original_name = "batch_norm2d_50.b_0" shape = [96] dtype = "float32" min_val = float("-0.320182") @@ -5912,6 +6456,7 @@ class Program_weight_tensor_parameter_543: class Program_weight_tensor_parameter_544: name = "parameter_544" + original_name = "batch_norm2d_50.w_0" shape = [96] dtype = "float32" min_val = float("0.00204699") @@ -5923,6 +6468,7 @@ class Program_weight_tensor_parameter_544: class Program_weight_tensor_parameter_545: name = "parameter_545" + original_name = "batch_norm2d_50.w_2" shape = [96] dtype = "float32" min_val = float("0.0331995") @@ -5934,6 +6480,7 @@ class Program_weight_tensor_parameter_545: class Program_weight_tensor_parameter_546: name = "parameter_546" + original_name = "batch_norm2d_50.w_1" shape = [96] dtype = "float32" min_val = float("-1.26158") @@ -5945,6 +6492,7 @@ class Program_weight_tensor_parameter_546: class Program_weight_tensor_parameter_547: name = "parameter_547" + original_name = "conv2d_45.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.737922") @@ -5956,6 +6504,7 @@ class Program_weight_tensor_parameter_547: class Program_weight_tensor_parameter_548: name = "parameter_548" + original_name = "batch_norm2d_49.b_0" shape = [96] dtype = "float32" min_val = float("-0.320182") @@ -5967,6 +6516,7 @@ class Program_weight_tensor_parameter_548: class Program_weight_tensor_parameter_549: name = "parameter_549" + original_name = "batch_norm2d_49.w_0" shape = [96] dtype = "float32" min_val = float("-0.00100992") @@ -5978,6 +6528,7 @@ class Program_weight_tensor_parameter_549: class Program_weight_tensor_parameter_550: name = "parameter_550" + original_name = "batch_norm2d_49.w_2" shape = [96] dtype = "float32" min_val = float("0.0225888") @@ -5989,6 +6540,7 @@ class Program_weight_tensor_parameter_550: class Program_weight_tensor_parameter_551: name = "parameter_551" + original_name = "batch_norm2d_49.w_1" shape = [96] dtype = "float32" min_val = float("-1.35972") @@ -6000,6 +6552,7 @@ class Program_weight_tensor_parameter_551: class Program_weight_tensor_parameter_552: name = "parameter_552" + original_name = "conv2d_44.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.860871") @@ -6011,6 +6564,7 @@ class Program_weight_tensor_parameter_552: class Program_weight_tensor_parameter_553: name = "parameter_553" + original_name = "batch_norm2d_48.b_0" shape = [96] dtype = "float32" min_val = float("-0.320182") @@ -6022,6 +6576,7 @@ class Program_weight_tensor_parameter_553: class Program_weight_tensor_parameter_554: name = "parameter_554" + original_name = "batch_norm2d_48.w_0" shape = [96] dtype = "float32" min_val = float("0.000303357") @@ -6033,6 +6588,7 @@ class Program_weight_tensor_parameter_554: class Program_weight_tensor_parameter_555: name = "parameter_555" + original_name = "batch_norm2d_48.w_2" shape = [96] dtype = "float32" min_val = float("0.0106636") @@ -6044,6 +6600,7 @@ class Program_weight_tensor_parameter_555: class Program_weight_tensor_parameter_556: name = "parameter_556" + original_name = "batch_norm2d_48.w_1" shape = [96] dtype = "float32" min_val = float("-1.29585") @@ -6055,6 +6612,7 @@ class Program_weight_tensor_parameter_556: class Program_weight_tensor_parameter_557: name = "parameter_557" + original_name = "conv2d_43.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.867308") @@ -6066,6 +6624,7 @@ class Program_weight_tensor_parameter_557: class Program_weight_tensor_parameter_558: name = "parameter_558" + original_name = "batch_norm2d_47.b_0" shape = [96] dtype = "float32" min_val = float("-0.320182") @@ -6077,6 +6636,7 @@ class Program_weight_tensor_parameter_558: class Program_weight_tensor_parameter_559: name = "parameter_559" + original_name = "batch_norm2d_47.w_0" shape = [96] dtype = "float32" min_val = float("0.0503656") @@ -6088,6 +6648,7 @@ class Program_weight_tensor_parameter_559: class Program_weight_tensor_parameter_560: name = "parameter_560" + original_name = "batch_norm2d_47.w_2" shape = [96] dtype = "float32" min_val = float("0.0424927") @@ -6099,6 +6660,7 @@ class Program_weight_tensor_parameter_560: class Program_weight_tensor_parameter_561: name = "parameter_561" + original_name = "batch_norm2d_47.w_1" shape = [96] dtype = "float32" min_val = float("-1.18138") @@ -6110,6 +6672,7 @@ class Program_weight_tensor_parameter_561: class Program_weight_tensor_parameter_562: name = "parameter_562" + original_name = "conv2d_42.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.73066") @@ -6121,6 +6684,7 @@ class Program_weight_tensor_parameter_562: class Program_weight_tensor_parameter_563: name = "parameter_563" + original_name = "batch_norm2d_46.b_0" shape = [96] dtype = "float32" min_val = float("-0.306649") @@ -6132,6 +6696,7 @@ class Program_weight_tensor_parameter_563: class Program_weight_tensor_parameter_564: name = "parameter_564" + original_name = "batch_norm2d_46.w_0" shape = [96] dtype = "float32" min_val = float("-0.175034") @@ -6143,6 +6708,7 @@ class Program_weight_tensor_parameter_564: class Program_weight_tensor_parameter_565: name = "parameter_565" + original_name = "batch_norm2d_46.w_2" shape = [96] dtype = "float32" min_val = float("0.00383813") @@ -6154,6 +6720,7 @@ class Program_weight_tensor_parameter_565: class Program_weight_tensor_parameter_566: name = "parameter_566" + original_name = "batch_norm2d_46.w_1" shape = [96] dtype = "float32" min_val = float("0.00448567") @@ -6165,6 +6732,7 @@ class Program_weight_tensor_parameter_566: class Program_weight_tensor_parameter_567: name = "parameter_567" + original_name = "batch_norm2d_44.b_0" shape = [96] dtype = "float32" min_val = float("-0.193968") @@ -6176,6 +6744,7 @@ class Program_weight_tensor_parameter_567: class Program_weight_tensor_parameter_568: name = "parameter_568" + original_name = "batch_norm2d_44.w_0" shape = [96] dtype = "float32" min_val = float("-0.130107") @@ -6187,6 +6756,7 @@ class Program_weight_tensor_parameter_568: class Program_weight_tensor_parameter_569: name = "parameter_569" + original_name = "batch_norm2d_44.w_2" shape = [96] dtype = "float32" min_val = float("0.000400882") @@ -6198,6 +6768,7 @@ class Program_weight_tensor_parameter_569: class Program_weight_tensor_parameter_570: name = "parameter_570" + original_name = "batch_norm2d_44.w_1" shape = [96] dtype = "float32" min_val = float("-0.560565") @@ -6209,6 +6780,7 @@ class Program_weight_tensor_parameter_570: class Program_weight_tensor_parameter_571: name = "parameter_571" + original_name = "conv2d_40.w_0" shape = [96, 1, 3, 3] dtype = "float32" min_val = float("-1.11675") @@ -6220,6 +6792,7 @@ class Program_weight_tensor_parameter_571: class Program_weight_tensor_parameter_572: name = "parameter_572" + original_name = "batch_norm2d_43.b_0" shape = [96] dtype = "float32" min_val = float("-0.193968") @@ -6231,6 +6804,7 @@ class Program_weight_tensor_parameter_572: class Program_weight_tensor_parameter_573: name = "parameter_573" + original_name = "batch_norm2d_43.w_0" shape = [96] dtype = "float32" min_val = float("-0.036535") @@ -6242,6 +6816,7 @@ class Program_weight_tensor_parameter_573: class Program_weight_tensor_parameter_574: name = "parameter_574" + original_name = "batch_norm2d_43.w_2" shape = [96] dtype = "float32" min_val = float("0.000660416") @@ -6253,6 +6828,7 @@ class Program_weight_tensor_parameter_574: class Program_weight_tensor_parameter_575: name = "parameter_575" + original_name = "batch_norm2d_43.w_1" shape = [96] dtype = "float32" min_val = float("-0.560288") @@ -6264,6 +6840,7 @@ class Program_weight_tensor_parameter_575: class Program_weight_tensor_parameter_576: name = "parameter_576" + original_name = "conv2d_39.w_0" shape = [96, 1, 3, 3] dtype = "float32" min_val = float("-0.669519") @@ -6275,6 +6852,7 @@ class Program_weight_tensor_parameter_576: class Program_weight_tensor_parameter_577: name = "parameter_577" + original_name = "batch_norm2d_42.b_0" shape = [96] dtype = "float32" min_val = float("-0.193968") @@ -6286,6 +6864,7 @@ class Program_weight_tensor_parameter_577: class Program_weight_tensor_parameter_578: name = "parameter_578" + original_name = "batch_norm2d_42.w_0" shape = [96] dtype = "float32" min_val = float("-0.135336") @@ -6297,6 +6876,7 @@ class Program_weight_tensor_parameter_578: class Program_weight_tensor_parameter_579: name = "parameter_579" + original_name = "batch_norm2d_42.w_2" shape = [96] dtype = "float32" min_val = float("0.000461872") @@ -6308,6 +6888,7 @@ class Program_weight_tensor_parameter_579: class Program_weight_tensor_parameter_580: name = "parameter_580" + original_name = "batch_norm2d_42.w_1" shape = [96] dtype = "float32" min_val = float("-0.56238") @@ -6319,6 +6900,7 @@ class Program_weight_tensor_parameter_580: class Program_weight_tensor_parameter_581: name = "parameter_581" + original_name = "conv2d_38.w_0" shape = [96, 1, 3, 3] dtype = "float32" min_val = float("-0.673625") @@ -6330,6 +6912,7 @@ class Program_weight_tensor_parameter_581: class Program_weight_tensor_parameter_582: name = "parameter_582" + original_name = "batch_norm2d_41.b_0" shape = [96] dtype = "float32" min_val = float("-0.193968") @@ -6341,6 +6924,7 @@ class Program_weight_tensor_parameter_582: class Program_weight_tensor_parameter_583: name = "parameter_583" + original_name = "batch_norm2d_41.w_0" shape = [96] dtype = "float32" min_val = float("-0.000247392") @@ -6352,6 +6936,7 @@ class Program_weight_tensor_parameter_583: class Program_weight_tensor_parameter_584: name = "parameter_584" + original_name = "batch_norm2d_41.w_2" shape = [96] dtype = "float32" min_val = float("0.000417419") @@ -6363,6 +6948,7 @@ class Program_weight_tensor_parameter_584: class Program_weight_tensor_parameter_585: name = "parameter_585" + original_name = "batch_norm2d_41.w_1" shape = [96] dtype = "float32" min_val = float("-0.565229") @@ -6374,6 +6960,7 @@ class Program_weight_tensor_parameter_585: class Program_weight_tensor_parameter_586: name = "parameter_586" + original_name = "conv2d_37.w_0" shape = [96, 1, 3, 3] dtype = "float32" min_val = float("-0.560745") @@ -6385,6 +6972,7 @@ class Program_weight_tensor_parameter_586: class Program_weight_tensor_parameter_587: name = "parameter_587" + original_name = "batch_norm2d_45.b_0" shape = [96] dtype = "float32" min_val = float("-0.193968") @@ -6396,6 +6984,7 @@ class Program_weight_tensor_parameter_587: class Program_weight_tensor_parameter_588: name = "parameter_588" + original_name = "batch_norm2d_45.w_0" shape = [96] dtype = "float32" min_val = float("-0.320329") @@ -6407,6 +6996,7 @@ class Program_weight_tensor_parameter_588: class Program_weight_tensor_parameter_589: name = "parameter_589" + original_name = "batch_norm2d_45.w_2" shape = [96] dtype = "float32" min_val = float("1.34288e-08") @@ -6418,6 +7008,7 @@ class Program_weight_tensor_parameter_589: class Program_weight_tensor_parameter_590: name = "parameter_590" + original_name = "batch_norm2d_45.w_1" shape = [96] dtype = "float32" min_val = float("-0.0951603") @@ -6429,6 +7020,7 @@ class Program_weight_tensor_parameter_590: class Program_weight_tensor_parameter_591: name = "parameter_591" + original_name = "conv2d_41.w_0" shape = [96, 1, 1, 1] dtype = "float32" min_val = float("-0.156096") @@ -6440,6 +7032,7 @@ class Program_weight_tensor_parameter_591: class Program_weight_tensor_parameter_592: name = "parameter_592" + original_name = "batch_norm2d_40.b_0" shape = [96] dtype = "float32" min_val = float("-0.179074") @@ -6451,6 +7044,7 @@ class Program_weight_tensor_parameter_592: class Program_weight_tensor_parameter_593: name = "parameter_593" + original_name = "batch_norm2d_40.w_0" shape = [96] dtype = "float32" min_val = float("-0.336282") @@ -6462,6 +7056,7 @@ class Program_weight_tensor_parameter_593: class Program_weight_tensor_parameter_594: name = "parameter_594" + original_name = "batch_norm2d_40.w_2" shape = [96] dtype = "float32" min_val = float("0.0112225") @@ -6473,6 +7068,7 @@ class Program_weight_tensor_parameter_594: class Program_weight_tensor_parameter_595: name = "parameter_595" + original_name = "batch_norm2d_40.w_1" shape = [96] dtype = "float32" min_val = float("-0.955063") @@ -6484,6 +7080,7 @@ class Program_weight_tensor_parameter_595: class Program_weight_tensor_parameter_596: name = "parameter_596" + original_name = "batch_norm2d_39.b_0" shape = [96] dtype = "float32" min_val = float("-0.831751") @@ -6495,6 +7092,7 @@ class Program_weight_tensor_parameter_596: class Program_weight_tensor_parameter_597: name = "parameter_597" + original_name = "batch_norm2d_39.w_0" shape = [96] dtype = "float32" min_val = float("0.00521394") @@ -6506,6 +7104,7 @@ class Program_weight_tensor_parameter_597: class Program_weight_tensor_parameter_598: name = "parameter_598" + original_name = "batch_norm2d_39.w_2" shape = [96] dtype = "float32" min_val = float("0.00720984") @@ -6517,6 +7116,7 @@ class Program_weight_tensor_parameter_598: class Program_weight_tensor_parameter_599: name = "parameter_599" + original_name = "batch_norm2d_39.w_1" shape = [96] dtype = "float32" min_val = float("-0.00372417") @@ -6528,6 +7128,7 @@ class Program_weight_tensor_parameter_599: class Program_weight_tensor_parameter_600: name = "parameter_600" + original_name = "conv2d_36.w_0" shape = [96, 48, 1, 1] dtype = "float32" min_val = float("-1.03014") @@ -6539,6 +7140,7 @@ class Program_weight_tensor_parameter_600: class Program_weight_tensor_parameter_601: name = "parameter_601" + original_name = "batch_norm2d_38.b_0" shape = [96] dtype = "float32" min_val = float("-0.831751") @@ -6550,6 +7152,7 @@ class Program_weight_tensor_parameter_601: class Program_weight_tensor_parameter_602: name = "parameter_602" + original_name = "batch_norm2d_38.w_0" shape = [96] dtype = "float32" min_val = float("-0.00501337") @@ -6561,6 +7164,7 @@ class Program_weight_tensor_parameter_602: class Program_weight_tensor_parameter_603: name = "parameter_603" + original_name = "batch_norm2d_38.w_2" shape = [96] dtype = "float32" min_val = float("0.00518392") @@ -6572,6 +7176,7 @@ class Program_weight_tensor_parameter_603: class Program_weight_tensor_parameter_604: name = "parameter_604" + original_name = "batch_norm2d_38.w_1" shape = [96] dtype = "float32" min_val = float("-0.00370593") @@ -6583,6 +7188,7 @@ class Program_weight_tensor_parameter_604: class Program_weight_tensor_parameter_605: name = "parameter_605" + original_name = "conv2d_35.w_0" shape = [96, 48, 1, 1] dtype = "float32" min_val = float("-0.939445") @@ -6594,6 +7200,7 @@ class Program_weight_tensor_parameter_605: class Program_weight_tensor_parameter_606: name = "parameter_606" + original_name = "batch_norm2d_37.b_0" shape = [96] dtype = "float32" min_val = float("-0.831751") @@ -6605,6 +7212,7 @@ class Program_weight_tensor_parameter_606: class Program_weight_tensor_parameter_607: name = "parameter_607" + original_name = "batch_norm2d_37.w_0" shape = [96] dtype = "float32" min_val = float("-0.0163726") @@ -6616,6 +7224,7 @@ class Program_weight_tensor_parameter_607: class Program_weight_tensor_parameter_608: name = "parameter_608" + original_name = "batch_norm2d_37.w_2" shape = [96] dtype = "float32" min_val = float("0.0103846") @@ -6627,6 +7236,7 @@ class Program_weight_tensor_parameter_608: class Program_weight_tensor_parameter_609: name = "parameter_609" + original_name = "batch_norm2d_37.w_1" shape = [96] dtype = "float32" min_val = float("-0.00429068") @@ -6638,6 +7248,7 @@ class Program_weight_tensor_parameter_609: class Program_weight_tensor_parameter_610: name = "parameter_610" + original_name = "conv2d_34.w_0" shape = [96, 48, 1, 1] dtype = "float32" min_val = float("-0.8523") @@ -6649,6 +7260,7 @@ class Program_weight_tensor_parameter_610: class Program_weight_tensor_parameter_611: name = "parameter_611" + original_name = "batch_norm2d_36.b_0" shape = [96] dtype = "float32" min_val = float("-0.831751") @@ -6660,6 +7272,7 @@ class Program_weight_tensor_parameter_611: class Program_weight_tensor_parameter_612: name = "parameter_612" + original_name = "batch_norm2d_36.w_0" shape = [96] dtype = "float32" min_val = float("-0.00191146") @@ -6671,6 +7284,7 @@ class Program_weight_tensor_parameter_612: class Program_weight_tensor_parameter_613: name = "parameter_613" + original_name = "batch_norm2d_36.w_2" shape = [96] dtype = "float32" min_val = float("0.00396639") @@ -6682,6 +7296,7 @@ class Program_weight_tensor_parameter_613: class Program_weight_tensor_parameter_614: name = "parameter_614" + original_name = "batch_norm2d_36.w_1" shape = [96] dtype = "float32" min_val = float("-0.00375176") @@ -6693,6 +7308,7 @@ class Program_weight_tensor_parameter_614: class Program_weight_tensor_parameter_615: name = "parameter_615" + original_name = "conv2d_33.w_0" shape = [96, 48, 1, 1] dtype = "float32" min_val = float("-0.865664") @@ -6704,6 +7320,7 @@ class Program_weight_tensor_parameter_615: class Program_weight_tensor_parameter_616: name = "parameter_616" + original_name = "batch_norm2d_34.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -6713,6 +7330,7 @@ class Program_weight_tensor_parameter_616: class Program_weight_tensor_parameter_617: name = "parameter_617" + original_name = "batch_norm2d_34.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -6722,6 +7340,7 @@ class Program_weight_tensor_parameter_617: class Program_weight_tensor_parameter_618: name = "parameter_618" + original_name = "batch_norm2d_34.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -6731,6 +7350,7 @@ class Program_weight_tensor_parameter_618: class Program_weight_tensor_parameter_619: name = "parameter_619" + original_name = "batch_norm2d_34.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -6740,6 +7360,7 @@ class Program_weight_tensor_parameter_619: class Program_weight_tensor_parameter_620: name = "parameter_620" + original_name = "conv2d_31.w_0" shape = [48, 1, 3, 3] dtype = "float32" min_val = float("-0.290552") @@ -6751,6 +7372,7 @@ class Program_weight_tensor_parameter_620: class Program_weight_tensor_parameter_621: name = "parameter_621" + original_name = "batch_norm2d_33.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -6760,6 +7382,7 @@ class Program_weight_tensor_parameter_621: class Program_weight_tensor_parameter_622: name = "parameter_622" + original_name = "batch_norm2d_33.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -6769,6 +7392,7 @@ class Program_weight_tensor_parameter_622: class Program_weight_tensor_parameter_623: name = "parameter_623" + original_name = "batch_norm2d_33.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -6778,6 +7402,7 @@ class Program_weight_tensor_parameter_623: class Program_weight_tensor_parameter_624: name = "parameter_624" + original_name = "batch_norm2d_33.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -6787,6 +7412,7 @@ class Program_weight_tensor_parameter_624: class Program_weight_tensor_parameter_625: name = "parameter_625" + original_name = "conv2d_30.w_0" shape = [48, 1, 3, 3] dtype = "float32" min_val = float("-0.291619") @@ -6798,6 +7424,7 @@ class Program_weight_tensor_parameter_625: class Program_weight_tensor_parameter_626: name = "parameter_626" + original_name = "batch_norm2d_32.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -6807,6 +7434,7 @@ class Program_weight_tensor_parameter_626: class Program_weight_tensor_parameter_627: name = "parameter_627" + original_name = "batch_norm2d_32.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -6816,6 +7444,7 @@ class Program_weight_tensor_parameter_627: class Program_weight_tensor_parameter_628: name = "parameter_628" + original_name = "batch_norm2d_32.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -6825,6 +7454,7 @@ class Program_weight_tensor_parameter_628: class Program_weight_tensor_parameter_629: name = "parameter_629" + original_name = "batch_norm2d_32.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -6834,6 +7464,7 @@ class Program_weight_tensor_parameter_629: class Program_weight_tensor_parameter_630: name = "parameter_630" + original_name = "conv2d_29.w_0" shape = [48, 1, 3, 3] dtype = "float32" min_val = float("-0.285038") @@ -6845,6 +7476,7 @@ class Program_weight_tensor_parameter_630: class Program_weight_tensor_parameter_631: name = "parameter_631" + original_name = "batch_norm2d_31.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -6854,6 +7486,7 @@ class Program_weight_tensor_parameter_631: class Program_weight_tensor_parameter_632: name = "parameter_632" + original_name = "batch_norm2d_31.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -6863,6 +7496,7 @@ class Program_weight_tensor_parameter_632: class Program_weight_tensor_parameter_633: name = "parameter_633" + original_name = "batch_norm2d_31.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -6872,6 +7506,7 @@ class Program_weight_tensor_parameter_633: class Program_weight_tensor_parameter_634: name = "parameter_634" + original_name = "batch_norm2d_31.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -6881,6 +7516,7 @@ class Program_weight_tensor_parameter_634: class Program_weight_tensor_parameter_635: name = "parameter_635" + original_name = "conv2d_28.w_0" shape = [48, 1, 3, 3] dtype = "float32" min_val = float("-0.28803") @@ -6892,6 +7528,7 @@ class Program_weight_tensor_parameter_635: class Program_weight_tensor_parameter_636: name = "parameter_636" + original_name = "batch_norm2d_35.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -6901,6 +7538,7 @@ class Program_weight_tensor_parameter_636: class Program_weight_tensor_parameter_637: name = "parameter_637" + original_name = "batch_norm2d_35.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -6910,6 +7548,7 @@ class Program_weight_tensor_parameter_637: class Program_weight_tensor_parameter_638: name = "parameter_638" + original_name = "batch_norm2d_35.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -6919,6 +7558,7 @@ class Program_weight_tensor_parameter_638: class Program_weight_tensor_parameter_639: name = "parameter_639" + original_name = "batch_norm2d_35.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -6928,6 +7568,7 @@ class Program_weight_tensor_parameter_639: class Program_weight_tensor_parameter_640: name = "parameter_640" + original_name = "conv2d_32.w_0" shape = [48, 1, 1, 1] dtype = "float32" min_val = float("0") @@ -6937,6 +7578,7 @@ class Program_weight_tensor_parameter_640: class Program_weight_tensor_parameter_641: name = "parameter_641" + original_name = "batch_norm2d_30.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -6946,6 +7588,7 @@ class Program_weight_tensor_parameter_641: class Program_weight_tensor_parameter_642: name = "parameter_642" + original_name = "batch_norm2d_30.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -6955,6 +7598,7 @@ class Program_weight_tensor_parameter_642: class Program_weight_tensor_parameter_643: name = "parameter_643" + original_name = "batch_norm2d_30.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -6964,6 +7608,7 @@ class Program_weight_tensor_parameter_643: class Program_weight_tensor_parameter_644: name = "parameter_644" + original_name = "batch_norm2d_30.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -6973,6 +7618,7 @@ class Program_weight_tensor_parameter_644: class Program_weight_tensor_parameter_645: name = "parameter_645" + original_name = "conv2d_27.w_0" shape = [48, 48, 1, 1] dtype = "float32" min_val = float("-0.942761") @@ -6984,6 +7630,7 @@ class Program_weight_tensor_parameter_645: class Program_weight_tensor_parameter_646: name = "parameter_646" + original_name = "batch_norm2d_29.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -6993,6 +7640,7 @@ class Program_weight_tensor_parameter_646: class Program_weight_tensor_parameter_647: name = "parameter_647" + original_name = "batch_norm2d_29.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -7002,6 +7650,7 @@ class Program_weight_tensor_parameter_647: class Program_weight_tensor_parameter_648: name = "parameter_648" + original_name = "batch_norm2d_29.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -7011,6 +7660,7 @@ class Program_weight_tensor_parameter_648: class Program_weight_tensor_parameter_649: name = "parameter_649" + original_name = "batch_norm2d_29.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -7020,6 +7670,7 @@ class Program_weight_tensor_parameter_649: class Program_weight_tensor_parameter_650: name = "parameter_650" + original_name = "conv2d_26.w_0" shape = [48, 48, 1, 1] dtype = "float32" min_val = float("-1.08695") @@ -7031,6 +7682,7 @@ class Program_weight_tensor_parameter_650: class Program_weight_tensor_parameter_651: name = "parameter_651" + original_name = "batch_norm2d_28.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -7040,6 +7692,7 @@ class Program_weight_tensor_parameter_651: class Program_weight_tensor_parameter_652: name = "parameter_652" + original_name = "batch_norm2d_28.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -7049,6 +7702,7 @@ class Program_weight_tensor_parameter_652: class Program_weight_tensor_parameter_653: name = "parameter_653" + original_name = "batch_norm2d_28.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -7058,6 +7712,7 @@ class Program_weight_tensor_parameter_653: class Program_weight_tensor_parameter_654: name = "parameter_654" + original_name = "batch_norm2d_28.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -7067,6 +7722,7 @@ class Program_weight_tensor_parameter_654: class Program_weight_tensor_parameter_655: name = "parameter_655" + original_name = "conv2d_25.w_0" shape = [48, 48, 1, 1] dtype = "float32" min_val = float("-1.05153") @@ -7078,6 +7734,7 @@ class Program_weight_tensor_parameter_655: class Program_weight_tensor_parameter_656: name = "parameter_656" + original_name = "batch_norm2d_27.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -7087,6 +7744,7 @@ class Program_weight_tensor_parameter_656: class Program_weight_tensor_parameter_657: name = "parameter_657" + original_name = "batch_norm2d_27.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -7096,6 +7754,7 @@ class Program_weight_tensor_parameter_657: class Program_weight_tensor_parameter_658: name = "parameter_658" + original_name = "batch_norm2d_27.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -7105,6 +7764,7 @@ class Program_weight_tensor_parameter_658: class Program_weight_tensor_parameter_659: name = "parameter_659" + original_name = "batch_norm2d_27.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -7114,6 +7774,7 @@ class Program_weight_tensor_parameter_659: class Program_weight_tensor_parameter_660: name = "parameter_660" + original_name = "conv2d_24.w_0" shape = [48, 48, 1, 1] dtype = "float32" min_val = float("-0.892054") @@ -7125,6 +7786,7 @@ class Program_weight_tensor_parameter_660: class Program_weight_tensor_parameter_661: name = "parameter_661" + original_name = "batch_norm2d_26.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -7134,6 +7796,7 @@ class Program_weight_tensor_parameter_661: class Program_weight_tensor_parameter_662: name = "parameter_662" + original_name = "batch_norm2d_26.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -7143,6 +7806,7 @@ class Program_weight_tensor_parameter_662: class Program_weight_tensor_parameter_663: name = "parameter_663" + original_name = "batch_norm2d_26.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -7152,6 +7816,7 @@ class Program_weight_tensor_parameter_663: class Program_weight_tensor_parameter_664: name = "parameter_664" + original_name = "batch_norm2d_26.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -7161,6 +7826,7 @@ class Program_weight_tensor_parameter_664: class Program_weight_tensor_parameter_665: name = "parameter_665" + original_name = "batch_norm2d_24.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -7170,6 +7836,7 @@ class Program_weight_tensor_parameter_665: class Program_weight_tensor_parameter_666: name = "parameter_666" + original_name = "batch_norm2d_24.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -7179,6 +7846,7 @@ class Program_weight_tensor_parameter_666: class Program_weight_tensor_parameter_667: name = "parameter_667" + original_name = "batch_norm2d_24.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -7188,6 +7856,7 @@ class Program_weight_tensor_parameter_667: class Program_weight_tensor_parameter_668: name = "parameter_668" + original_name = "batch_norm2d_24.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -7197,6 +7866,7 @@ class Program_weight_tensor_parameter_668: class Program_weight_tensor_parameter_669: name = "parameter_669" + original_name = "conv2d_22.w_0" shape = [48, 1, 3, 3] dtype = "float32" min_val = float("-0.595766") @@ -7208,6 +7878,7 @@ class Program_weight_tensor_parameter_669: class Program_weight_tensor_parameter_670: name = "parameter_670" + original_name = "batch_norm2d_23.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -7217,6 +7888,7 @@ class Program_weight_tensor_parameter_670: class Program_weight_tensor_parameter_671: name = "parameter_671" + original_name = "batch_norm2d_23.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -7226,6 +7898,7 @@ class Program_weight_tensor_parameter_671: class Program_weight_tensor_parameter_672: name = "parameter_672" + original_name = "batch_norm2d_23.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -7235,6 +7908,7 @@ class Program_weight_tensor_parameter_672: class Program_weight_tensor_parameter_673: name = "parameter_673" + original_name = "batch_norm2d_23.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -7244,6 +7918,7 @@ class Program_weight_tensor_parameter_673: class Program_weight_tensor_parameter_674: name = "parameter_674" + original_name = "conv2d_21.w_0" shape = [48, 1, 3, 3] dtype = "float32" min_val = float("-0.57054") @@ -7255,6 +7930,7 @@ class Program_weight_tensor_parameter_674: class Program_weight_tensor_parameter_675: name = "parameter_675" + original_name = "batch_norm2d_22.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -7264,6 +7940,7 @@ class Program_weight_tensor_parameter_675: class Program_weight_tensor_parameter_676: name = "parameter_676" + original_name = "batch_norm2d_22.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -7273,6 +7950,7 @@ class Program_weight_tensor_parameter_676: class Program_weight_tensor_parameter_677: name = "parameter_677" + original_name = "batch_norm2d_22.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -7282,6 +7960,7 @@ class Program_weight_tensor_parameter_677: class Program_weight_tensor_parameter_678: name = "parameter_678" + original_name = "batch_norm2d_22.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -7291,6 +7970,7 @@ class Program_weight_tensor_parameter_678: class Program_weight_tensor_parameter_679: name = "parameter_679" + original_name = "conv2d_20.w_0" shape = [48, 1, 3, 3] dtype = "float32" min_val = float("-0.568308") @@ -7302,6 +7982,7 @@ class Program_weight_tensor_parameter_679: class Program_weight_tensor_parameter_680: name = "parameter_680" + original_name = "batch_norm2d_21.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -7311,6 +7992,7 @@ class Program_weight_tensor_parameter_680: class Program_weight_tensor_parameter_681: name = "parameter_681" + original_name = "batch_norm2d_21.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -7320,6 +8002,7 @@ class Program_weight_tensor_parameter_681: class Program_weight_tensor_parameter_682: name = "parameter_682" + original_name = "batch_norm2d_21.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -7329,6 +8012,7 @@ class Program_weight_tensor_parameter_682: class Program_weight_tensor_parameter_683: name = "parameter_683" + original_name = "batch_norm2d_21.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -7338,6 +8022,7 @@ class Program_weight_tensor_parameter_683: class Program_weight_tensor_parameter_684: name = "parameter_684" + original_name = "conv2d_19.w_0" shape = [48, 1, 3, 3] dtype = "float32" min_val = float("-0.621412") @@ -7349,6 +8034,7 @@ class Program_weight_tensor_parameter_684: class Program_weight_tensor_parameter_685: name = "parameter_685" + original_name = "batch_norm2d_25.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -7358,6 +8044,7 @@ class Program_weight_tensor_parameter_685: class Program_weight_tensor_parameter_686: name = "parameter_686" + original_name = "batch_norm2d_25.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -7367,6 +8054,7 @@ class Program_weight_tensor_parameter_686: class Program_weight_tensor_parameter_687: name = "parameter_687" + original_name = "batch_norm2d_25.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -7376,6 +8064,7 @@ class Program_weight_tensor_parameter_687: class Program_weight_tensor_parameter_688: name = "parameter_688" + original_name = "batch_norm2d_25.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -7385,6 +8074,7 @@ class Program_weight_tensor_parameter_688: class Program_weight_tensor_parameter_689: name = "parameter_689" + original_name = "conv2d_23.w_0" shape = [48, 1, 1, 1] dtype = "float32" min_val = float("0") @@ -7394,6 +8084,7 @@ class Program_weight_tensor_parameter_689: class Program_weight_tensor_parameter_690: name = "parameter_690" + original_name = "batch_norm2d_20.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -7403,6 +8094,7 @@ class Program_weight_tensor_parameter_690: class Program_weight_tensor_parameter_691: name = "parameter_691" + original_name = "batch_norm2d_20.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -7412,6 +8104,7 @@ class Program_weight_tensor_parameter_691: class Program_weight_tensor_parameter_692: name = "parameter_692" + original_name = "batch_norm2d_20.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -7421,6 +8114,7 @@ class Program_weight_tensor_parameter_692: class Program_weight_tensor_parameter_693: name = "parameter_693" + original_name = "batch_norm2d_20.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -7430,6 +8124,7 @@ class Program_weight_tensor_parameter_693: class Program_weight_tensor_parameter_694: name = "parameter_694" + original_name = "batch_norm2d_19.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -7439,6 +8134,7 @@ class Program_weight_tensor_parameter_694: class Program_weight_tensor_parameter_695: name = "parameter_695" + original_name = "batch_norm2d_19.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -7448,6 +8144,7 @@ class Program_weight_tensor_parameter_695: class Program_weight_tensor_parameter_696: name = "parameter_696" + original_name = "batch_norm2d_19.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -7457,6 +8154,7 @@ class Program_weight_tensor_parameter_696: class Program_weight_tensor_parameter_697: name = "parameter_697" + original_name = "batch_norm2d_19.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -7466,6 +8164,7 @@ class Program_weight_tensor_parameter_697: class Program_weight_tensor_parameter_698: name = "parameter_698" + original_name = "conv2d_18.w_0" shape = [48, 32, 1, 1] dtype = "float32" min_val = float("-0.935603") @@ -7477,6 +8176,7 @@ class Program_weight_tensor_parameter_698: class Program_weight_tensor_parameter_699: name = "parameter_699" + original_name = "batch_norm2d_18.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -7486,6 +8186,7 @@ class Program_weight_tensor_parameter_699: class Program_weight_tensor_parameter_700: name = "parameter_700" + original_name = "batch_norm2d_18.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -7495,6 +8196,7 @@ class Program_weight_tensor_parameter_700: class Program_weight_tensor_parameter_701: name = "parameter_701" + original_name = "batch_norm2d_18.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -7504,6 +8206,7 @@ class Program_weight_tensor_parameter_701: class Program_weight_tensor_parameter_702: name = "parameter_702" + original_name = "batch_norm2d_18.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -7513,6 +8216,7 @@ class Program_weight_tensor_parameter_702: class Program_weight_tensor_parameter_703: name = "parameter_703" + original_name = "conv2d_17.w_0" shape = [48, 32, 1, 1] dtype = "float32" min_val = float("-1.15907") @@ -7524,6 +8228,7 @@ class Program_weight_tensor_parameter_703: class Program_weight_tensor_parameter_704: name = "parameter_704" + original_name = "batch_norm2d_17.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -7533,6 +8238,7 @@ class Program_weight_tensor_parameter_704: class Program_weight_tensor_parameter_705: name = "parameter_705" + original_name = "batch_norm2d_17.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -7542,6 +8248,7 @@ class Program_weight_tensor_parameter_705: class Program_weight_tensor_parameter_706: name = "parameter_706" + original_name = "batch_norm2d_17.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -7551,6 +8258,7 @@ class Program_weight_tensor_parameter_706: class Program_weight_tensor_parameter_707: name = "parameter_707" + original_name = "batch_norm2d_17.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -7560,6 +8268,7 @@ class Program_weight_tensor_parameter_707: class Program_weight_tensor_parameter_708: name = "parameter_708" + original_name = "conv2d_16.w_0" shape = [48, 32, 1, 1] dtype = "float32" min_val = float("-1.20425") @@ -7571,6 +8280,7 @@ class Program_weight_tensor_parameter_708: class Program_weight_tensor_parameter_709: name = "parameter_709" + original_name = "batch_norm2d_16.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -7580,6 +8290,7 @@ class Program_weight_tensor_parameter_709: class Program_weight_tensor_parameter_710: name = "parameter_710" + original_name = "batch_norm2d_16.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -7589,6 +8300,7 @@ class Program_weight_tensor_parameter_710: class Program_weight_tensor_parameter_711: name = "parameter_711" + original_name = "batch_norm2d_16.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -7598,6 +8310,7 @@ class Program_weight_tensor_parameter_711: class Program_weight_tensor_parameter_712: name = "parameter_712" + original_name = "batch_norm2d_16.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -7607,6 +8320,7 @@ class Program_weight_tensor_parameter_712: class Program_weight_tensor_parameter_713: name = "parameter_713" + original_name = "conv2d_15.w_0" shape = [48, 32, 1, 1] dtype = "float32" min_val = float("-1.16617") @@ -7618,6 +8332,7 @@ class Program_weight_tensor_parameter_713: class Program_weight_tensor_parameter_714: name = "parameter_714" + original_name = "batch_norm2d_14.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -7627,6 +8342,7 @@ class Program_weight_tensor_parameter_714: class Program_weight_tensor_parameter_715: name = "parameter_715" + original_name = "batch_norm2d_14.w_0" shape = [32] dtype = "float32" min_val = float("0") @@ -7636,6 +8352,7 @@ class Program_weight_tensor_parameter_715: class Program_weight_tensor_parameter_716: name = "parameter_716" + original_name = "batch_norm2d_14.w_2" shape = [32] dtype = "float32" min_val = float("0") @@ -7645,6 +8362,7 @@ class Program_weight_tensor_parameter_716: class Program_weight_tensor_parameter_717: name = "parameter_717" + original_name = "batch_norm2d_14.w_1" shape = [32] dtype = "float32" min_val = float("0") @@ -7654,6 +8372,7 @@ class Program_weight_tensor_parameter_717: class Program_weight_tensor_parameter_718: name = "parameter_718" + original_name = "conv2d_13.w_0" shape = [32, 1, 3, 3] dtype = "float32" min_val = float("-0.536174") @@ -7665,6 +8384,7 @@ class Program_weight_tensor_parameter_718: class Program_weight_tensor_parameter_719: name = "parameter_719" + original_name = "batch_norm2d_13.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -7674,6 +8394,7 @@ class Program_weight_tensor_parameter_719: class Program_weight_tensor_parameter_720: name = "parameter_720" + original_name = "batch_norm2d_13.w_0" shape = [32] dtype = "float32" min_val = float("0") @@ -7683,6 +8404,7 @@ class Program_weight_tensor_parameter_720: class Program_weight_tensor_parameter_721: name = "parameter_721" + original_name = "batch_norm2d_13.w_2" shape = [32] dtype = "float32" min_val = float("0") @@ -7692,6 +8414,7 @@ class Program_weight_tensor_parameter_721: class Program_weight_tensor_parameter_722: name = "parameter_722" + original_name = "batch_norm2d_13.w_1" shape = [32] dtype = "float32" min_val = float("0") @@ -7701,6 +8424,7 @@ class Program_weight_tensor_parameter_722: class Program_weight_tensor_parameter_723: name = "parameter_723" + original_name = "conv2d_12.w_0" shape = [32, 1, 3, 3] dtype = "float32" min_val = float("-0.489") @@ -7712,6 +8436,7 @@ class Program_weight_tensor_parameter_723: class Program_weight_tensor_parameter_724: name = "parameter_724" + original_name = "batch_norm2d_12.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -7721,6 +8446,7 @@ class Program_weight_tensor_parameter_724: class Program_weight_tensor_parameter_725: name = "parameter_725" + original_name = "batch_norm2d_12.w_0" shape = [32] dtype = "float32" min_val = float("0") @@ -7730,6 +8456,7 @@ class Program_weight_tensor_parameter_725: class Program_weight_tensor_parameter_726: name = "parameter_726" + original_name = "batch_norm2d_12.w_2" shape = [32] dtype = "float32" min_val = float("0") @@ -7739,6 +8466,7 @@ class Program_weight_tensor_parameter_726: class Program_weight_tensor_parameter_727: name = "parameter_727" + original_name = "batch_norm2d_12.w_1" shape = [32] dtype = "float32" min_val = float("0") @@ -7748,6 +8476,7 @@ class Program_weight_tensor_parameter_727: class Program_weight_tensor_parameter_728: name = "parameter_728" + original_name = "conv2d_11.w_0" shape = [32, 1, 3, 3] dtype = "float32" min_val = float("-0.514415") @@ -7759,6 +8488,7 @@ class Program_weight_tensor_parameter_728: class Program_weight_tensor_parameter_729: name = "parameter_729" + original_name = "batch_norm2d_11.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -7768,6 +8498,7 @@ class Program_weight_tensor_parameter_729: class Program_weight_tensor_parameter_730: name = "parameter_730" + original_name = "batch_norm2d_11.w_0" shape = [32] dtype = "float32" min_val = float("0") @@ -7777,6 +8508,7 @@ class Program_weight_tensor_parameter_730: class Program_weight_tensor_parameter_731: name = "parameter_731" + original_name = "batch_norm2d_11.w_2" shape = [32] dtype = "float32" min_val = float("0") @@ -7786,6 +8518,7 @@ class Program_weight_tensor_parameter_731: class Program_weight_tensor_parameter_732: name = "parameter_732" + original_name = "batch_norm2d_11.w_1" shape = [32] dtype = "float32" min_val = float("0") @@ -7795,6 +8528,7 @@ class Program_weight_tensor_parameter_732: class Program_weight_tensor_parameter_733: name = "parameter_733" + original_name = "conv2d_10.w_0" shape = [32, 1, 3, 3] dtype = "float32" min_val = float("-0.498247") @@ -7806,6 +8540,7 @@ class Program_weight_tensor_parameter_733: class Program_weight_tensor_parameter_734: name = "parameter_734" + original_name = "batch_norm2d_15.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -7815,6 +8550,7 @@ class Program_weight_tensor_parameter_734: class Program_weight_tensor_parameter_735: name = "parameter_735" + original_name = "batch_norm2d_15.w_0" shape = [32] dtype = "float32" min_val = float("0") @@ -7824,6 +8560,7 @@ class Program_weight_tensor_parameter_735: class Program_weight_tensor_parameter_736: name = "parameter_736" + original_name = "batch_norm2d_15.w_2" shape = [32] dtype = "float32" min_val = float("0") @@ -7833,6 +8570,7 @@ class Program_weight_tensor_parameter_736: class Program_weight_tensor_parameter_737: name = "parameter_737" + original_name = "batch_norm2d_15.w_1" shape = [32] dtype = "float32" min_val = float("0") @@ -7842,6 +8580,7 @@ class Program_weight_tensor_parameter_737: class Program_weight_tensor_parameter_738: name = "parameter_738" + original_name = "conv2d_14.w_0" shape = [32, 1, 1, 1] dtype = "float32" min_val = float("0") @@ -7851,6 +8590,7 @@ class Program_weight_tensor_parameter_738: class Program_weight_tensor_parameter_739: name = "parameter_739" + original_name = "batch_norm2d_10.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -7860,6 +8600,7 @@ class Program_weight_tensor_parameter_739: class Program_weight_tensor_parameter_740: name = "parameter_740" + original_name = "batch_norm2d_10.w_0" shape = [32] dtype = "float32" min_val = float("0") @@ -7869,6 +8610,7 @@ class Program_weight_tensor_parameter_740: class Program_weight_tensor_parameter_741: name = "parameter_741" + original_name = "batch_norm2d_10.w_2" shape = [32] dtype = "float32" min_val = float("0") @@ -7878,6 +8620,7 @@ class Program_weight_tensor_parameter_741: class Program_weight_tensor_parameter_742: name = "parameter_742" + original_name = "batch_norm2d_10.w_1" shape = [32] dtype = "float32" min_val = float("0") @@ -7887,6 +8630,7 @@ class Program_weight_tensor_parameter_742: class Program_weight_tensor_parameter_743: name = "parameter_743" + original_name = "conv2d_9.w_0" shape = [32, 16, 1, 1] dtype = "float32" min_val = float("-1.53916") @@ -7898,6 +8642,7 @@ class Program_weight_tensor_parameter_743: class Program_weight_tensor_parameter_744: name = "parameter_744" + original_name = "batch_norm2d_9.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -7907,6 +8652,7 @@ class Program_weight_tensor_parameter_744: class Program_weight_tensor_parameter_745: name = "parameter_745" + original_name = "batch_norm2d_9.w_0" shape = [32] dtype = "float32" min_val = float("0") @@ -7916,6 +8662,7 @@ class Program_weight_tensor_parameter_745: class Program_weight_tensor_parameter_746: name = "parameter_746" + original_name = "batch_norm2d_9.w_2" shape = [32] dtype = "float32" min_val = float("0") @@ -7925,6 +8672,7 @@ class Program_weight_tensor_parameter_746: class Program_weight_tensor_parameter_747: name = "parameter_747" + original_name = "batch_norm2d_9.w_1" shape = [32] dtype = "float32" min_val = float("0") @@ -7934,6 +8682,7 @@ class Program_weight_tensor_parameter_747: class Program_weight_tensor_parameter_748: name = "parameter_748" + original_name = "conv2d_8.w_0" shape = [32, 16, 1, 1] dtype = "float32" min_val = float("-1.69203") @@ -7945,6 +8694,7 @@ class Program_weight_tensor_parameter_748: class Program_weight_tensor_parameter_749: name = "parameter_749" + original_name = "batch_norm2d_8.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -7954,6 +8704,7 @@ class Program_weight_tensor_parameter_749: class Program_weight_tensor_parameter_750: name = "parameter_750" + original_name = "batch_norm2d_8.w_0" shape = [32] dtype = "float32" min_val = float("0") @@ -7963,6 +8714,7 @@ class Program_weight_tensor_parameter_750: class Program_weight_tensor_parameter_751: name = "parameter_751" + original_name = "batch_norm2d_8.w_2" shape = [32] dtype = "float32" min_val = float("0") @@ -7972,6 +8724,7 @@ class Program_weight_tensor_parameter_751: class Program_weight_tensor_parameter_752: name = "parameter_752" + original_name = "batch_norm2d_8.w_1" shape = [32] dtype = "float32" min_val = float("0") @@ -7981,6 +8734,7 @@ class Program_weight_tensor_parameter_752: class Program_weight_tensor_parameter_753: name = "parameter_753" + original_name = "conv2d_7.w_0" shape = [32, 16, 1, 1] dtype = "float32" min_val = float("-1.75469") @@ -7992,6 +8746,7 @@ class Program_weight_tensor_parameter_753: class Program_weight_tensor_parameter_754: name = "parameter_754" + original_name = "batch_norm2d_7.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -8001,6 +8756,7 @@ class Program_weight_tensor_parameter_754: class Program_weight_tensor_parameter_755: name = "parameter_755" + original_name = "batch_norm2d_7.w_0" shape = [32] dtype = "float32" min_val = float("0") @@ -8010,6 +8766,7 @@ class Program_weight_tensor_parameter_755: class Program_weight_tensor_parameter_756: name = "parameter_756" + original_name = "batch_norm2d_7.w_2" shape = [32] dtype = "float32" min_val = float("0") @@ -8019,6 +8776,7 @@ class Program_weight_tensor_parameter_756: class Program_weight_tensor_parameter_757: name = "parameter_757" + original_name = "batch_norm2d_7.w_1" shape = [32] dtype = "float32" min_val = float("0") @@ -8028,6 +8786,7 @@ class Program_weight_tensor_parameter_757: class Program_weight_tensor_parameter_758: name = "parameter_758" + original_name = "conv2d_6.w_0" shape = [32, 16, 1, 1] dtype = "float32" min_val = float("-1.58513") @@ -8039,6 +8798,7 @@ class Program_weight_tensor_parameter_758: class Program_weight_tensor_parameter_759: name = "parameter_759" + original_name = "batch_norm2d_5.b_0" shape = [16] dtype = "float32" min_val = float("0") @@ -8048,6 +8808,7 @@ class Program_weight_tensor_parameter_759: class Program_weight_tensor_parameter_760: name = "parameter_760" + original_name = "batch_norm2d_5.w_0" shape = [16] dtype = "float32" min_val = float("0") @@ -8057,6 +8818,7 @@ class Program_weight_tensor_parameter_760: class Program_weight_tensor_parameter_761: name = "parameter_761" + original_name = "batch_norm2d_5.w_2" shape = [16] dtype = "float32" min_val = float("0") @@ -8066,6 +8828,7 @@ class Program_weight_tensor_parameter_761: class Program_weight_tensor_parameter_762: name = "parameter_762" + original_name = "batch_norm2d_5.w_1" shape = [16] dtype = "float32" min_val = float("0") @@ -8075,6 +8838,7 @@ class Program_weight_tensor_parameter_762: class Program_weight_tensor_parameter_763: name = "parameter_763" + original_name = "conv2d_4.w_0" shape = [16, 1, 3, 3] dtype = "float32" min_val = float("-1.62152") @@ -8086,6 +8850,7 @@ class Program_weight_tensor_parameter_763: class Program_weight_tensor_parameter_764: name = "parameter_764" + original_name = "batch_norm2d_4.b_0" shape = [16] dtype = "float32" min_val = float("0") @@ -8095,6 +8860,7 @@ class Program_weight_tensor_parameter_764: class Program_weight_tensor_parameter_765: name = "parameter_765" + original_name = "batch_norm2d_4.w_0" shape = [16] dtype = "float32" min_val = float("0") @@ -8104,6 +8870,7 @@ class Program_weight_tensor_parameter_765: class Program_weight_tensor_parameter_766: name = "parameter_766" + original_name = "batch_norm2d_4.w_2" shape = [16] dtype = "float32" min_val = float("0") @@ -8113,6 +8880,7 @@ class Program_weight_tensor_parameter_766: class Program_weight_tensor_parameter_767: name = "parameter_767" + original_name = "batch_norm2d_4.w_1" shape = [16] dtype = "float32" min_val = float("0") @@ -8122,6 +8890,7 @@ class Program_weight_tensor_parameter_767: class Program_weight_tensor_parameter_768: name = "parameter_768" + original_name = "conv2d_3.w_0" shape = [16, 1, 3, 3] dtype = "float32" min_val = float("-1.63931") @@ -8133,6 +8902,7 @@ class Program_weight_tensor_parameter_768: class Program_weight_tensor_parameter_769: name = "parameter_769" + original_name = "batch_norm2d_3.b_0" shape = [16] dtype = "float32" min_val = float("0") @@ -8142,6 +8912,7 @@ class Program_weight_tensor_parameter_769: class Program_weight_tensor_parameter_770: name = "parameter_770" + original_name = "batch_norm2d_3.w_0" shape = [16] dtype = "float32" min_val = float("0") @@ -8151,6 +8922,7 @@ class Program_weight_tensor_parameter_770: class Program_weight_tensor_parameter_771: name = "parameter_771" + original_name = "batch_norm2d_3.w_2" shape = [16] dtype = "float32" min_val = float("0") @@ -8160,6 +8932,7 @@ class Program_weight_tensor_parameter_771: class Program_weight_tensor_parameter_772: name = "parameter_772" + original_name = "batch_norm2d_3.w_1" shape = [16] dtype = "float32" min_val = float("0") @@ -8169,6 +8942,7 @@ class Program_weight_tensor_parameter_772: class Program_weight_tensor_parameter_773: name = "parameter_773" + original_name = "conv2d_2.w_0" shape = [16, 1, 3, 3] dtype = "float32" min_val = float("-1.64363") @@ -8180,6 +8954,7 @@ class Program_weight_tensor_parameter_773: class Program_weight_tensor_parameter_774: name = "parameter_774" + original_name = "batch_norm2d_2.b_0" shape = [16] dtype = "float32" min_val = float("0") @@ -8189,6 +8964,7 @@ class Program_weight_tensor_parameter_774: class Program_weight_tensor_parameter_775: name = "parameter_775" + original_name = "batch_norm2d_2.w_0" shape = [16] dtype = "float32" min_val = float("0") @@ -8198,6 +8974,7 @@ class Program_weight_tensor_parameter_775: class Program_weight_tensor_parameter_776: name = "parameter_776" + original_name = "batch_norm2d_2.w_2" shape = [16] dtype = "float32" min_val = float("0") @@ -8207,6 +8984,7 @@ class Program_weight_tensor_parameter_776: class Program_weight_tensor_parameter_777: name = "parameter_777" + original_name = "batch_norm2d_2.w_1" shape = [16] dtype = "float32" min_val = float("0") @@ -8216,6 +8994,7 @@ class Program_weight_tensor_parameter_777: class Program_weight_tensor_parameter_778: name = "parameter_778" + original_name = "conv2d_1.w_0" shape = [16, 1, 3, 3] dtype = "float32" min_val = float("-1.65374") @@ -8227,6 +9006,7 @@ class Program_weight_tensor_parameter_778: class Program_weight_tensor_parameter_779: name = "parameter_779" + original_name = "batch_norm2d_6.b_0" shape = [16] dtype = "float32" min_val = float("0") @@ -8236,6 +9016,7 @@ class Program_weight_tensor_parameter_779: class Program_weight_tensor_parameter_780: name = "parameter_780" + original_name = "batch_norm2d_6.w_0" shape = [16] dtype = "float32" min_val = float("0") @@ -8245,6 +9026,7 @@ class Program_weight_tensor_parameter_780: class Program_weight_tensor_parameter_781: name = "parameter_781" + original_name = "batch_norm2d_6.w_2" shape = [16] dtype = "float32" min_val = float("0") @@ -8254,6 +9036,7 @@ class Program_weight_tensor_parameter_781: class Program_weight_tensor_parameter_782: name = "parameter_782" + original_name = "batch_norm2d_6.w_1" shape = [16] dtype = "float32" min_val = float("0") @@ -8263,6 +9046,7 @@ class Program_weight_tensor_parameter_782: class Program_weight_tensor_parameter_783: name = "parameter_783" + original_name = "conv2d_5.w_0" shape = [16, 1, 1, 1] dtype = "float32" min_val = float("0") @@ -8272,6 +9056,7 @@ class Program_weight_tensor_parameter_783: class Program_weight_tensor_parameter_784: name = "parameter_784" + original_name = "batch_norm2d_1.b_0" shape = [16] dtype = "float32" min_val = float("0") @@ -8281,6 +9066,7 @@ class Program_weight_tensor_parameter_784: class Program_weight_tensor_parameter_785: name = "parameter_785" + original_name = "batch_norm2d_1.w_0" shape = [16] dtype = "float32" min_val = float("0") @@ -8290,6 +9076,7 @@ class Program_weight_tensor_parameter_785: class Program_weight_tensor_parameter_786: name = "parameter_786" + original_name = "batch_norm2d_1.w_2" shape = [16] dtype = "float32" min_val = float("0") @@ -8299,6 +9086,7 @@ class Program_weight_tensor_parameter_786: class Program_weight_tensor_parameter_787: name = "parameter_787" + original_name = "batch_norm2d_1.w_1" shape = [16] dtype = "float32" min_val = float("0") @@ -8308,6 +9096,7 @@ class Program_weight_tensor_parameter_787: class Program_weight_tensor_parameter_788: name = "parameter_788" + original_name = "batch_norm2d_0.b_0" shape = [16] dtype = "float32" min_val = float("0") @@ -8317,6 +9106,7 @@ class Program_weight_tensor_parameter_788: class Program_weight_tensor_parameter_789: name = "parameter_789" + original_name = "batch_norm2d_0.w_0" shape = [16] dtype = "float32" min_val = float("0") @@ -8326,6 +9116,7 @@ class Program_weight_tensor_parameter_789: class Program_weight_tensor_parameter_790: name = "parameter_790" + original_name = "batch_norm2d_0.w_2" shape = [16] dtype = "float32" min_val = float("0") @@ -8335,6 +9126,7 @@ class Program_weight_tensor_parameter_790: class Program_weight_tensor_parameter_791: name = "parameter_791" + original_name = "batch_norm2d_0.w_1" shape = [16] dtype = "float32" min_val = float("0") @@ -8344,6 +9136,7 @@ class Program_weight_tensor_parameter_791: class Program_weight_tensor_parameter_792: name = "parameter_792" + original_name = "conv2d_0.w_0" shape = [16, 3, 3, 3] dtype = "float32" min_val = float("-1.86399") diff --git a/paddle_samples/PaddleX/PP-OCRv4_server_det/subgraph_0/input_meta.py b/paddle_samples/PaddleX/PP-OCRv4_server_det/subgraph_0/input_meta.py index f648c954c..356931a55 100644 --- a/paddle_samples/PaddleX/PP-OCRv4_server_det/subgraph_0/input_meta.py +++ b/paddle_samples/PaddleX/PP-OCRv4_server_det/subgraph_0/input_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_data_0: name = "data_0" + original_name = "var_0" shape = [4, 3, 640, 640] dtype = "float32" min_val = float("-2.1179") diff --git a/paddle_samples/PaddleX/PP-OCRv4_server_det/subgraph_0/weight_meta.py b/paddle_samples/PaddleX/PP-OCRv4_server_det/subgraph_0/weight_meta.py index a63e97011..bba477498 100644 --- a/paddle_samples/PaddleX/PP-OCRv4_server_det/subgraph_0/weight_meta.py +++ b/paddle_samples/PaddleX/PP-OCRv4_server_det/subgraph_0/weight_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_parameter_0: name = "parameter_0" + original_name = "conv2d_transpose_3.b_0" shape = [1] dtype = "float32" min_val = float("0") @@ -9,6 +10,7 @@ class Program_weight_tensor_parameter_0: class Program_weight_tensor_parameter_1: name = "parameter_1" + original_name = "conv2d_transpose_3.w_0" shape = [64, 1, 2, 2] dtype = "float32" min_val = float("-1.39557") @@ -20,6 +22,7 @@ class Program_weight_tensor_parameter_1: class Program_weight_tensor_parameter_2: name = "parameter_2" + original_name = "batch_norm_3.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -29,6 +32,7 @@ class Program_weight_tensor_parameter_2: class Program_weight_tensor_parameter_3: name = "parameter_3" + original_name = "batch_norm_3.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -38,6 +42,7 @@ class Program_weight_tensor_parameter_3: class Program_weight_tensor_parameter_4: name = "parameter_4" + original_name = "batch_norm_3.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -47,6 +52,7 @@ class Program_weight_tensor_parameter_4: class Program_weight_tensor_parameter_5: name = "parameter_5" + original_name = "batch_norm_3.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -56,6 +62,7 @@ class Program_weight_tensor_parameter_5: class Program_weight_tensor_parameter_6: name = "parameter_6" + original_name = "conv2d_transpose_2.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -65,6 +72,7 @@ class Program_weight_tensor_parameter_6: class Program_weight_tensor_parameter_7: name = "parameter_7" + original_name = "conv2d_transpose_2.w_0" shape = [64, 64, 2, 2] dtype = "float32" min_val = float("-1.36424") @@ -76,6 +84,7 @@ class Program_weight_tensor_parameter_7: class Program_weight_tensor_parameter_8: name = "parameter_8" + original_name = "batch_norm_2.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -85,6 +94,7 @@ class Program_weight_tensor_parameter_8: class Program_weight_tensor_parameter_9: name = "parameter_9" + original_name = "batch_norm_2.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -94,6 +104,7 @@ class Program_weight_tensor_parameter_9: class Program_weight_tensor_parameter_10: name = "parameter_10" + original_name = "batch_norm_2.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -103,6 +114,7 @@ class Program_weight_tensor_parameter_10: class Program_weight_tensor_parameter_11: name = "parameter_11" + original_name = "batch_norm_2.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -112,6 +124,7 @@ class Program_weight_tensor_parameter_11: class Program_weight_tensor_parameter_12: name = "parameter_12" + original_name = "conv2d_106.w_0" shape = [64, 256, 3, 3] dtype = "float32" min_val = float("-1.08913") @@ -123,6 +136,7 @@ class Program_weight_tensor_parameter_12: class Program_weight_tensor_parameter_13: name = "parameter_13" + original_name = "conv2d_108.b_0" shape = [1] dtype = "float32" min_val = float("0") @@ -132,6 +146,7 @@ class Program_weight_tensor_parameter_13: class Program_weight_tensor_parameter_14: name = "parameter_14" + original_name = "conv2d_108.w_0" shape = [1, 64, 1, 1] dtype = "float32" min_val = float("0") @@ -141,6 +156,7 @@ class Program_weight_tensor_parameter_14: class Program_weight_tensor_parameter_15: name = "parameter_15" + original_name = "batch_norm_4.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -150,6 +166,7 @@ class Program_weight_tensor_parameter_15: class Program_weight_tensor_parameter_16: name = "parameter_16" + original_name = "batch_norm_4.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -159,6 +176,7 @@ class Program_weight_tensor_parameter_16: class Program_weight_tensor_parameter_17: name = "parameter_17" + original_name = "batch_norm_4.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -168,6 +186,7 @@ class Program_weight_tensor_parameter_17: class Program_weight_tensor_parameter_18: name = "parameter_18" + original_name = "batch_norm_4.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -177,6 +196,7 @@ class Program_weight_tensor_parameter_18: class Program_weight_tensor_parameter_19: name = "parameter_19" + original_name = "conv2d_107.w_0" shape = [64, 65, 3, 3] dtype = "float32" min_val = float("-1.3832") @@ -188,6 +208,7 @@ class Program_weight_tensor_parameter_19: class Program_weight_tensor_parameter_20: name = "parameter_20" + original_name = "conv2d_transpose_1.b_0" shape = [1] dtype = "float32" min_val = float("0") @@ -197,6 +218,7 @@ class Program_weight_tensor_parameter_20: class Program_weight_tensor_parameter_21: name = "parameter_21" + original_name = "conv2d_transpose_1.w_0" shape = [64, 1, 2, 2] dtype = "float32" min_val = float("-1.80665") @@ -208,6 +230,7 @@ class Program_weight_tensor_parameter_21: class Program_weight_tensor_parameter_22: name = "parameter_22" + original_name = "batch_norm_1.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -217,6 +240,7 @@ class Program_weight_tensor_parameter_22: class Program_weight_tensor_parameter_23: name = "parameter_23" + original_name = "batch_norm_1.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -226,6 +250,7 @@ class Program_weight_tensor_parameter_23: class Program_weight_tensor_parameter_24: name = "parameter_24" + original_name = "batch_norm_1.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -235,6 +260,7 @@ class Program_weight_tensor_parameter_24: class Program_weight_tensor_parameter_25: name = "parameter_25" + original_name = "batch_norm_1.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -244,6 +270,7 @@ class Program_weight_tensor_parameter_25: class Program_weight_tensor_parameter_26: name = "parameter_26" + original_name = "conv2d_transpose_0.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -253,6 +280,7 @@ class Program_weight_tensor_parameter_26: class Program_weight_tensor_parameter_27: name = "parameter_27" + original_name = "conv2d_transpose_0.w_0" shape = [64, 64, 2, 2] dtype = "float32" min_val = float("-1.07545") @@ -264,6 +292,7 @@ class Program_weight_tensor_parameter_27: class Program_weight_tensor_parameter_28: name = "parameter_28" + original_name = "batch_norm_0.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -273,6 +302,7 @@ class Program_weight_tensor_parameter_28: class Program_weight_tensor_parameter_29: name = "parameter_29" + original_name = "batch_norm_0.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -282,6 +312,7 @@ class Program_weight_tensor_parameter_29: class Program_weight_tensor_parameter_30: name = "parameter_30" + original_name = "batch_norm_0.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -291,6 +322,7 @@ class Program_weight_tensor_parameter_30: class Program_weight_tensor_parameter_31: name = "parameter_31" + original_name = "batch_norm_0.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -300,6 +332,7 @@ class Program_weight_tensor_parameter_31: class Program_weight_tensor_parameter_32: name = "parameter_32" + original_name = "conv2d_105.w_0" shape = [64, 256, 3, 3] dtype = "float32" min_val = float("-1.35647") @@ -311,6 +344,7 @@ class Program_weight_tensor_parameter_32: class Program_weight_tensor_parameter_33: name = "parameter_33" + original_name = "batch_norm2d_41.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -320,6 +354,7 @@ class Program_weight_tensor_parameter_33: class Program_weight_tensor_parameter_34: name = "parameter_34" + original_name = "batch_norm2d_41.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -329,6 +364,7 @@ class Program_weight_tensor_parameter_34: class Program_weight_tensor_parameter_35: name = "parameter_35" + original_name = "batch_norm2d_41.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -338,6 +374,7 @@ class Program_weight_tensor_parameter_35: class Program_weight_tensor_parameter_36: name = "parameter_36" + original_name = "batch_norm2d_41.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -347,6 +384,7 @@ class Program_weight_tensor_parameter_36: class Program_weight_tensor_parameter_37: name = "parameter_37" + original_name = "conv2d_62.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -356,6 +394,7 @@ class Program_weight_tensor_parameter_37: class Program_weight_tensor_parameter_38: name = "parameter_38" + original_name = "conv2d_62.w_0" shape = [64, 32, 1, 1] dtype = "float32" min_val = float("-6.71668e-39") @@ -366,6 +405,7 @@ class Program_weight_tensor_parameter_38: class Program_weight_tensor_parameter_39: name = "parameter_39" + original_name = "conv2d_68.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -375,6 +415,7 @@ class Program_weight_tensor_parameter_39: class Program_weight_tensor_parameter_40: name = "parameter_40" + original_name = "conv2d_68.w_0" shape = [32, 32, 1, 3] dtype = "float32" min_val = float("-6.71203e-39") @@ -385,6 +426,7 @@ class Program_weight_tensor_parameter_40: class Program_weight_tensor_parameter_41: name = "parameter_41" + original_name = "conv2d_65.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -394,6 +436,7 @@ class Program_weight_tensor_parameter_41: class Program_weight_tensor_parameter_42: name = "parameter_42" + original_name = "conv2d_65.w_0" shape = [32, 32, 3, 1] dtype = "float32" min_val = float("-6.71687e-39") @@ -404,6 +447,7 @@ class Program_weight_tensor_parameter_42: class Program_weight_tensor_parameter_43: name = "parameter_43" + original_name = "conv2d_71.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -413,6 +457,7 @@ class Program_weight_tensor_parameter_43: class Program_weight_tensor_parameter_44: name = "parameter_44" + original_name = "conv2d_71.w_0" shape = [32, 32, 3, 3] dtype = "float32" min_val = float("-6.7148e-39") @@ -423,6 +468,7 @@ class Program_weight_tensor_parameter_44: class Program_weight_tensor_parameter_45: name = "parameter_45" + original_name = "conv2d_67.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -432,6 +478,7 @@ class Program_weight_tensor_parameter_45: class Program_weight_tensor_parameter_46: name = "parameter_46" + original_name = "conv2d_67.w_0" shape = [32, 32, 1, 5] dtype = "float32" min_val = float("-6.71596e-39") @@ -442,6 +489,7 @@ class Program_weight_tensor_parameter_46: class Program_weight_tensor_parameter_47: name = "parameter_47" + original_name = "conv2d_64.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -451,6 +499,7 @@ class Program_weight_tensor_parameter_47: class Program_weight_tensor_parameter_48: name = "parameter_48" + original_name = "conv2d_64.w_0" shape = [32, 32, 5, 1] dtype = "float32" min_val = float("-6.71097e-39") @@ -461,6 +510,7 @@ class Program_weight_tensor_parameter_48: class Program_weight_tensor_parameter_49: name = "parameter_49" + original_name = "conv2d_70.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -470,6 +520,7 @@ class Program_weight_tensor_parameter_49: class Program_weight_tensor_parameter_50: name = "parameter_50" + original_name = "conv2d_70.w_0" shape = [32, 32, 5, 5] dtype = "float32" min_val = float("-6.71651e-39") @@ -480,6 +531,7 @@ class Program_weight_tensor_parameter_50: class Program_weight_tensor_parameter_51: name = "parameter_51" + original_name = "conv2d_66.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -489,6 +541,7 @@ class Program_weight_tensor_parameter_51: class Program_weight_tensor_parameter_52: name = "parameter_52" + original_name = "conv2d_66.w_0" shape = [32, 32, 1, 7] dtype = "float32" min_val = float("-6.70365e-39") @@ -499,6 +552,7 @@ class Program_weight_tensor_parameter_52: class Program_weight_tensor_parameter_53: name = "parameter_53" + original_name = "conv2d_63.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -508,6 +562,7 @@ class Program_weight_tensor_parameter_53: class Program_weight_tensor_parameter_54: name = "parameter_54" + original_name = "conv2d_63.w_0" shape = [32, 32, 7, 1] dtype = "float32" min_val = float("-6.71122e-39") @@ -518,6 +573,7 @@ class Program_weight_tensor_parameter_54: class Program_weight_tensor_parameter_55: name = "parameter_55" + original_name = "conv2d_69.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -527,6 +583,7 @@ class Program_weight_tensor_parameter_55: class Program_weight_tensor_parameter_56: name = "parameter_56" + original_name = "conv2d_69.w_0" shape = [32, 32, 7, 7] dtype = "float32" min_val = float("-6.71658e-39") @@ -537,6 +594,7 @@ class Program_weight_tensor_parameter_56: class Program_weight_tensor_parameter_57: name = "parameter_57" + original_name = "conv2d_61.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -546,6 +604,7 @@ class Program_weight_tensor_parameter_57: class Program_weight_tensor_parameter_58: name = "parameter_58" + original_name = "conv2d_61.w_0" shape = [32, 64, 1, 1] dtype = "float32" min_val = float("-6.69613e-39") @@ -556,6 +615,7 @@ class Program_weight_tensor_parameter_58: class Program_weight_tensor_parameter_59: name = "parameter_59" + original_name = "batch_norm2d_42.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -565,6 +625,7 @@ class Program_weight_tensor_parameter_59: class Program_weight_tensor_parameter_60: name = "parameter_60" + original_name = "batch_norm2d_42.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -574,6 +635,7 @@ class Program_weight_tensor_parameter_60: class Program_weight_tensor_parameter_61: name = "parameter_61" + original_name = "batch_norm2d_42.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -583,6 +645,7 @@ class Program_weight_tensor_parameter_61: class Program_weight_tensor_parameter_62: name = "parameter_62" + original_name = "batch_norm2d_42.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -592,6 +655,7 @@ class Program_weight_tensor_parameter_62: class Program_weight_tensor_parameter_63: name = "parameter_63" + original_name = "conv2d_73.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -601,6 +665,7 @@ class Program_weight_tensor_parameter_63: class Program_weight_tensor_parameter_64: name = "parameter_64" + original_name = "conv2d_73.w_0" shape = [64, 32, 1, 1] dtype = "float32" min_val = float("-6.71322e-39") @@ -611,6 +676,7 @@ class Program_weight_tensor_parameter_64: class Program_weight_tensor_parameter_65: name = "parameter_65" + original_name = "conv2d_79.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -620,6 +686,7 @@ class Program_weight_tensor_parameter_65: class Program_weight_tensor_parameter_66: name = "parameter_66" + original_name = "conv2d_79.w_0" shape = [32, 32, 1, 3] dtype = "float32" min_val = float("-6.70394e-39") @@ -630,6 +697,7 @@ class Program_weight_tensor_parameter_66: class Program_weight_tensor_parameter_67: name = "parameter_67" + original_name = "conv2d_76.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -639,6 +707,7 @@ class Program_weight_tensor_parameter_67: class Program_weight_tensor_parameter_68: name = "parameter_68" + original_name = "conv2d_76.w_0" shape = [32, 32, 3, 1] dtype = "float32" min_val = float("-6.69191e-39") @@ -649,6 +718,7 @@ class Program_weight_tensor_parameter_68: class Program_weight_tensor_parameter_69: name = "parameter_69" + original_name = "conv2d_82.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -658,6 +728,7 @@ class Program_weight_tensor_parameter_69: class Program_weight_tensor_parameter_70: name = "parameter_70" + original_name = "conv2d_82.w_0" shape = [32, 32, 3, 3] dtype = "float32" min_val = float("-6.71599e-39") @@ -668,6 +739,7 @@ class Program_weight_tensor_parameter_70: class Program_weight_tensor_parameter_71: name = "parameter_71" + original_name = "conv2d_78.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -677,6 +749,7 @@ class Program_weight_tensor_parameter_71: class Program_weight_tensor_parameter_72: name = "parameter_72" + original_name = "conv2d_78.w_0" shape = [32, 32, 1, 5] dtype = "float32" min_val = float("-6.71367e-39") @@ -687,6 +760,7 @@ class Program_weight_tensor_parameter_72: class Program_weight_tensor_parameter_73: name = "parameter_73" + original_name = "conv2d_75.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -696,6 +770,7 @@ class Program_weight_tensor_parameter_73: class Program_weight_tensor_parameter_74: name = "parameter_74" + original_name = "conv2d_75.w_0" shape = [32, 32, 5, 1] dtype = "float32" min_val = float("-6.71343e-39") @@ -706,6 +781,7 @@ class Program_weight_tensor_parameter_74: class Program_weight_tensor_parameter_75: name = "parameter_75" + original_name = "conv2d_81.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -715,6 +791,7 @@ class Program_weight_tensor_parameter_75: class Program_weight_tensor_parameter_76: name = "parameter_76" + original_name = "conv2d_81.w_0" shape = [32, 32, 5, 5] dtype = "float32" min_val = float("-6.71671e-39") @@ -725,6 +802,7 @@ class Program_weight_tensor_parameter_76: class Program_weight_tensor_parameter_77: name = "parameter_77" + original_name = "conv2d_77.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -734,6 +812,7 @@ class Program_weight_tensor_parameter_77: class Program_weight_tensor_parameter_78: name = "parameter_78" + original_name = "conv2d_77.w_0" shape = [32, 32, 1, 7] dtype = "float32" min_val = float("-6.71534e-39") @@ -744,6 +823,7 @@ class Program_weight_tensor_parameter_78: class Program_weight_tensor_parameter_79: name = "parameter_79" + original_name = "conv2d_74.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -753,6 +833,7 @@ class Program_weight_tensor_parameter_79: class Program_weight_tensor_parameter_80: name = "parameter_80" + original_name = "conv2d_74.w_0" shape = [32, 32, 7, 1] dtype = "float32" min_val = float("-6.71371e-39") @@ -763,6 +844,7 @@ class Program_weight_tensor_parameter_80: class Program_weight_tensor_parameter_81: name = "parameter_81" + original_name = "conv2d_80.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -772,6 +854,7 @@ class Program_weight_tensor_parameter_81: class Program_weight_tensor_parameter_82: name = "parameter_82" + original_name = "conv2d_80.w_0" shape = [32, 32, 7, 7] dtype = "float32" min_val = float("-6.71683e-39") @@ -782,6 +865,7 @@ class Program_weight_tensor_parameter_82: class Program_weight_tensor_parameter_83: name = "parameter_83" + original_name = "conv2d_72.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -791,6 +875,7 @@ class Program_weight_tensor_parameter_83: class Program_weight_tensor_parameter_84: name = "parameter_84" + original_name = "conv2d_72.w_0" shape = [32, 64, 1, 1] dtype = "float32" min_val = float("-6.70856e-39") @@ -801,6 +886,7 @@ class Program_weight_tensor_parameter_84: class Program_weight_tensor_parameter_85: name = "parameter_85" + original_name = "batch_norm2d_43.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -810,6 +896,7 @@ class Program_weight_tensor_parameter_85: class Program_weight_tensor_parameter_86: name = "parameter_86" + original_name = "batch_norm2d_43.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -819,6 +906,7 @@ class Program_weight_tensor_parameter_86: class Program_weight_tensor_parameter_87: name = "parameter_87" + original_name = "batch_norm2d_43.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -828,6 +916,7 @@ class Program_weight_tensor_parameter_87: class Program_weight_tensor_parameter_88: name = "parameter_88" + original_name = "batch_norm2d_43.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -837,6 +926,7 @@ class Program_weight_tensor_parameter_88: class Program_weight_tensor_parameter_89: name = "parameter_89" + original_name = "conv2d_84.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -846,6 +936,7 @@ class Program_weight_tensor_parameter_89: class Program_weight_tensor_parameter_90: name = "parameter_90" + original_name = "conv2d_84.w_0" shape = [64, 32, 1, 1] dtype = "float32" min_val = float("-6.71587e-39") @@ -856,6 +947,7 @@ class Program_weight_tensor_parameter_90: class Program_weight_tensor_parameter_91: name = "parameter_91" + original_name = "conv2d_90.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -865,6 +957,7 @@ class Program_weight_tensor_parameter_91: class Program_weight_tensor_parameter_92: name = "parameter_92" + original_name = "conv2d_90.w_0" shape = [32, 32, 1, 3] dtype = "float32" min_val = float("-6.71173e-39") @@ -875,6 +968,7 @@ class Program_weight_tensor_parameter_92: class Program_weight_tensor_parameter_93: name = "parameter_93" + original_name = "conv2d_87.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -884,6 +978,7 @@ class Program_weight_tensor_parameter_93: class Program_weight_tensor_parameter_94: name = "parameter_94" + original_name = "conv2d_87.w_0" shape = [32, 32, 3, 1] dtype = "float32" min_val = float("-6.71611e-39") @@ -894,6 +989,7 @@ class Program_weight_tensor_parameter_94: class Program_weight_tensor_parameter_95: name = "parameter_95" + original_name = "conv2d_93.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -903,6 +999,7 @@ class Program_weight_tensor_parameter_95: class Program_weight_tensor_parameter_96: name = "parameter_96" + original_name = "conv2d_93.w_0" shape = [32, 32, 3, 3] dtype = "float32" min_val = float("-6.71303e-39") @@ -913,6 +1010,7 @@ class Program_weight_tensor_parameter_96: class Program_weight_tensor_parameter_97: name = "parameter_97" + original_name = "conv2d_89.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -922,6 +1020,7 @@ class Program_weight_tensor_parameter_97: class Program_weight_tensor_parameter_98: name = "parameter_98" + original_name = "conv2d_89.w_0" shape = [32, 32, 1, 5] dtype = "float32" min_val = float("-6.71598e-39") @@ -932,6 +1031,7 @@ class Program_weight_tensor_parameter_98: class Program_weight_tensor_parameter_99: name = "parameter_99" + original_name = "conv2d_86.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -941,6 +1041,7 @@ class Program_weight_tensor_parameter_99: class Program_weight_tensor_parameter_100: name = "parameter_100" + original_name = "conv2d_86.w_0" shape = [32, 32, 5, 1] dtype = "float32" min_val = float("-6.71491e-39") @@ -951,6 +1052,7 @@ class Program_weight_tensor_parameter_100: class Program_weight_tensor_parameter_101: name = "parameter_101" + original_name = "conv2d_92.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -960,6 +1062,7 @@ class Program_weight_tensor_parameter_101: class Program_weight_tensor_parameter_102: name = "parameter_102" + original_name = "conv2d_92.w_0" shape = [32, 32, 5, 5] dtype = "float32" min_val = float("-6.71617e-39") @@ -970,6 +1073,7 @@ class Program_weight_tensor_parameter_102: class Program_weight_tensor_parameter_103: name = "parameter_103" + original_name = "conv2d_88.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -979,6 +1083,7 @@ class Program_weight_tensor_parameter_103: class Program_weight_tensor_parameter_104: name = "parameter_104" + original_name = "conv2d_88.w_0" shape = [32, 32, 1, 7] dtype = "float32" min_val = float("-6.71613e-39") @@ -989,6 +1094,7 @@ class Program_weight_tensor_parameter_104: class Program_weight_tensor_parameter_105: name = "parameter_105" + original_name = "conv2d_85.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -998,6 +1104,7 @@ class Program_weight_tensor_parameter_105: class Program_weight_tensor_parameter_106: name = "parameter_106" + original_name = "conv2d_85.w_0" shape = [32, 32, 7, 1] dtype = "float32" min_val = float("-6.71627e-39") @@ -1008,6 +1115,7 @@ class Program_weight_tensor_parameter_106: class Program_weight_tensor_parameter_107: name = "parameter_107" + original_name = "conv2d_91.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -1017,6 +1125,7 @@ class Program_weight_tensor_parameter_107: class Program_weight_tensor_parameter_108: name = "parameter_108" + original_name = "conv2d_91.w_0" shape = [32, 32, 7, 7] dtype = "float32" min_val = float("-6.71689e-39") @@ -1027,6 +1136,7 @@ class Program_weight_tensor_parameter_108: class Program_weight_tensor_parameter_109: name = "parameter_109" + original_name = "conv2d_83.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -1036,6 +1146,7 @@ class Program_weight_tensor_parameter_109: class Program_weight_tensor_parameter_110: name = "parameter_110" + original_name = "conv2d_83.w_0" shape = [32, 64, 1, 1] dtype = "float32" min_val = float("-6.71032e-39") @@ -1046,6 +1157,7 @@ class Program_weight_tensor_parameter_110: class Program_weight_tensor_parameter_111: name = "parameter_111" + original_name = "batch_norm2d_44.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -1055,6 +1167,7 @@ class Program_weight_tensor_parameter_111: class Program_weight_tensor_parameter_112: name = "parameter_112" + original_name = "batch_norm2d_44.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -1064,6 +1177,7 @@ class Program_weight_tensor_parameter_112: class Program_weight_tensor_parameter_113: name = "parameter_113" + original_name = "batch_norm2d_44.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -1073,6 +1187,7 @@ class Program_weight_tensor_parameter_113: class Program_weight_tensor_parameter_114: name = "parameter_114" + original_name = "batch_norm2d_44.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -1082,6 +1197,7 @@ class Program_weight_tensor_parameter_114: class Program_weight_tensor_parameter_115: name = "parameter_115" + original_name = "conv2d_95.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -1091,6 +1207,7 @@ class Program_weight_tensor_parameter_115: class Program_weight_tensor_parameter_116: name = "parameter_116" + original_name = "conv2d_95.w_0" shape = [64, 32, 1, 1] dtype = "float32" min_val = float("-6.71249e-39") @@ -1101,6 +1218,7 @@ class Program_weight_tensor_parameter_116: class Program_weight_tensor_parameter_117: name = "parameter_117" + original_name = "conv2d_101.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -1110,6 +1228,7 @@ class Program_weight_tensor_parameter_117: class Program_weight_tensor_parameter_118: name = "parameter_118" + original_name = "conv2d_101.w_0" shape = [32, 32, 1, 3] dtype = "float32" min_val = float("-6.70783e-39") @@ -1120,6 +1239,7 @@ class Program_weight_tensor_parameter_118: class Program_weight_tensor_parameter_119: name = "parameter_119" + original_name = "conv2d_98.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -1129,6 +1249,7 @@ class Program_weight_tensor_parameter_119: class Program_weight_tensor_parameter_120: name = "parameter_120" + original_name = "conv2d_98.w_0" shape = [32, 32, 3, 1] dtype = "float32" min_val = float("-6.71476e-39") @@ -1139,6 +1260,7 @@ class Program_weight_tensor_parameter_120: class Program_weight_tensor_parameter_121: name = "parameter_121" + original_name = "conv2d_104.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -1148,6 +1270,7 @@ class Program_weight_tensor_parameter_121: class Program_weight_tensor_parameter_122: name = "parameter_122" + original_name = "conv2d_104.w_0" shape = [32, 32, 3, 3] dtype = "float32" min_val = float("-6.71564e-39") @@ -1158,6 +1281,7 @@ class Program_weight_tensor_parameter_122: class Program_weight_tensor_parameter_123: name = "parameter_123" + original_name = "conv2d_100.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -1167,6 +1291,7 @@ class Program_weight_tensor_parameter_123: class Program_weight_tensor_parameter_124: name = "parameter_124" + original_name = "conv2d_100.w_0" shape = [32, 32, 1, 5] dtype = "float32" min_val = float("-6.71264e-39") @@ -1177,6 +1302,7 @@ class Program_weight_tensor_parameter_124: class Program_weight_tensor_parameter_125: name = "parameter_125" + original_name = "conv2d_97.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -1186,6 +1312,7 @@ class Program_weight_tensor_parameter_125: class Program_weight_tensor_parameter_126: name = "parameter_126" + original_name = "conv2d_97.w_0" shape = [32, 32, 5, 1] dtype = "float32" min_val = float("-6.70838e-39") @@ -1196,6 +1323,7 @@ class Program_weight_tensor_parameter_126: class Program_weight_tensor_parameter_127: name = "parameter_127" + original_name = "conv2d_103.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -1205,6 +1333,7 @@ class Program_weight_tensor_parameter_127: class Program_weight_tensor_parameter_128: name = "parameter_128" + original_name = "conv2d_103.w_0" shape = [32, 32, 5, 5] dtype = "float32" min_val = float("-6.71654e-39") @@ -1215,6 +1344,7 @@ class Program_weight_tensor_parameter_128: class Program_weight_tensor_parameter_129: name = "parameter_129" + original_name = "conv2d_99.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -1224,6 +1354,7 @@ class Program_weight_tensor_parameter_129: class Program_weight_tensor_parameter_130: name = "parameter_130" + original_name = "conv2d_99.w_0" shape = [32, 32, 1, 7] dtype = "float32" min_val = float("-6.71313e-39") @@ -1234,6 +1365,7 @@ class Program_weight_tensor_parameter_130: class Program_weight_tensor_parameter_131: name = "parameter_131" + original_name = "conv2d_96.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -1243,6 +1375,7 @@ class Program_weight_tensor_parameter_131: class Program_weight_tensor_parameter_132: name = "parameter_132" + original_name = "conv2d_96.w_0" shape = [32, 32, 7, 1] dtype = "float32" min_val = float("-6.71566e-39") @@ -1253,6 +1386,7 @@ class Program_weight_tensor_parameter_132: class Program_weight_tensor_parameter_133: name = "parameter_133" + original_name = "conv2d_102.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -1262,6 +1396,7 @@ class Program_weight_tensor_parameter_133: class Program_weight_tensor_parameter_134: name = "parameter_134" + original_name = "conv2d_102.w_0" shape = [32, 32, 7, 7] dtype = "float32" min_val = float("-6.71645e-39") @@ -1272,6 +1407,7 @@ class Program_weight_tensor_parameter_134: class Program_weight_tensor_parameter_135: name = "parameter_135" + original_name = "conv2d_94.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -1281,6 +1417,7 @@ class Program_weight_tensor_parameter_135: class Program_weight_tensor_parameter_136: name = "parameter_136" + original_name = "conv2d_94.w_0" shape = [32, 64, 1, 1] dtype = "float32" min_val = float("-6.70284e-39") @@ -1291,6 +1428,7 @@ class Program_weight_tensor_parameter_136: class Program_weight_tensor_parameter_137: name = "parameter_137" + original_name = "conv2d_60.w_0" shape = [64, 64, 9, 9] dtype = "float32" min_val = float("-1.13239") @@ -1302,6 +1440,7 @@ class Program_weight_tensor_parameter_137: class Program_weight_tensor_parameter_138: name = "parameter_138" + original_name = "conv2d_56.w_0" shape = [64, 64, 9, 9] dtype = "float32" min_val = float("-1.27343") @@ -1313,6 +1452,7 @@ class Program_weight_tensor_parameter_138: class Program_weight_tensor_parameter_139: name = "parameter_139" + original_name = "conv2d_52.w_0" shape = [64, 64, 9, 9] dtype = "float32" min_val = float("-0.797558") @@ -1324,6 +1464,7 @@ class Program_weight_tensor_parameter_139: class Program_weight_tensor_parameter_140: name = "parameter_140" + original_name = "conv2d_48.w_0" shape = [64, 64, 9, 9] dtype = "float32" min_val = float("-1.24834") @@ -1335,6 +1476,7 @@ class Program_weight_tensor_parameter_140: class Program_weight_tensor_parameter_141: name = "parameter_141" + original_name = "conv2d_59.w_0" shape = [64, 64, 3, 3] dtype = "float32" min_val = float("-0.526092") @@ -1346,6 +1488,7 @@ class Program_weight_tensor_parameter_141: class Program_weight_tensor_parameter_142: name = "parameter_142" + original_name = "conv2d_55.w_0" shape = [64, 64, 3, 3] dtype = "float32" min_val = float("-0.364849") @@ -1357,6 +1500,7 @@ class Program_weight_tensor_parameter_142: class Program_weight_tensor_parameter_143: name = "parameter_143" + original_name = "conv2d_51.w_0" shape = [64, 64, 3, 3] dtype = "float32" min_val = float("-0.873638") @@ -1368,6 +1512,7 @@ class Program_weight_tensor_parameter_143: class Program_weight_tensor_parameter_144: name = "parameter_144" + original_name = "conv2d_47.w_0" shape = [64, 256, 9, 9] dtype = "float32" min_val = float("-0.911977") @@ -1379,6 +1524,7 @@ class Program_weight_tensor_parameter_144: class Program_weight_tensor_parameter_145: name = "parameter_145" + original_name = "conv2d_50.w_0" shape = [64, 256, 9, 9] dtype = "float32" min_val = float("-0.197539") @@ -1390,6 +1536,7 @@ class Program_weight_tensor_parameter_145: class Program_weight_tensor_parameter_146: name = "parameter_146" + original_name = "conv2d_54.w_0" shape = [64, 256, 9, 9] dtype = "float32" min_val = float("-0.43803") @@ -1401,6 +1548,7 @@ class Program_weight_tensor_parameter_146: class Program_weight_tensor_parameter_147: name = "parameter_147" + original_name = "conv2d_58.w_0" shape = [64, 256, 9, 9] dtype = "float32" min_val = float("-0.0913674") @@ -1412,6 +1560,7 @@ class Program_weight_tensor_parameter_147: class Program_weight_tensor_parameter_148: name = "parameter_148" + original_name = "conv2d_46.w_0" shape = [256, 256, 1, 1] dtype = "float32" min_val = float("-2.74111") @@ -1423,6 +1572,7 @@ class Program_weight_tensor_parameter_148: class Program_weight_tensor_parameter_149: name = "parameter_149" + original_name = "conv2d_49.w_0" shape = [256, 512, 1, 1] dtype = "float32" min_val = float("-1.05964") @@ -1434,6 +1584,7 @@ class Program_weight_tensor_parameter_149: class Program_weight_tensor_parameter_150: name = "parameter_150" + original_name = "conv2d_53.w_0" shape = [256, 768, 1, 1] dtype = "float32" min_val = float("-1.10705") @@ -1445,6 +1596,7 @@ class Program_weight_tensor_parameter_150: class Program_weight_tensor_parameter_151: name = "parameter_151" + original_name = "conv2d_57.w_0" shape = [256, 1024, 1, 1] dtype = "float32" min_val = float("-0.778035") @@ -1456,6 +1608,7 @@ class Program_weight_tensor_parameter_151: class Program_weight_tensor_parameter_152: name = "parameter_152" + original_name = "conv2d_45.b_0" shape = [1024] dtype = "float32" min_val = float("-0.210059") @@ -1467,6 +1620,7 @@ class Program_weight_tensor_parameter_152: class Program_weight_tensor_parameter_153: name = "parameter_153" + original_name = "conv2d_45.w_0" shape = [1024, 1024, 1, 1] dtype = "float32" min_val = float("-1.85846") @@ -1478,6 +1632,7 @@ class Program_weight_tensor_parameter_153: class Program_weight_tensor_parameter_154: name = "parameter_154" + original_name = "batch_norm2d_40.b_0" shape = [1024] dtype = "float32" min_val = float("-16.7598") @@ -1489,6 +1644,7 @@ class Program_weight_tensor_parameter_154: class Program_weight_tensor_parameter_155: name = "parameter_155" + original_name = "batch_norm2d_40.w_0" shape = [1024] dtype = "float32" min_val = float("2.46105") @@ -1500,6 +1656,7 @@ class Program_weight_tensor_parameter_155: class Program_weight_tensor_parameter_156: name = "parameter_156" + original_name = "batch_norm2d_40.w_2" shape = [1024] dtype = "float32" min_val = float("5.60519e-45") @@ -1511,6 +1668,7 @@ class Program_weight_tensor_parameter_156: class Program_weight_tensor_parameter_157: name = "parameter_157" + original_name = "batch_norm2d_40.w_1" shape = [1024] dtype = "float32" min_val = float("-10.8745") @@ -1522,6 +1680,7 @@ class Program_weight_tensor_parameter_157: class Program_weight_tensor_parameter_158: name = "parameter_158" + original_name = "conv2d_44.w_0" shape = [1024, 2112, 1, 1] dtype = "float32" min_val = float("-1.22208") @@ -1533,6 +1692,7 @@ class Program_weight_tensor_parameter_158: class Program_weight_tensor_parameter_159: name = "parameter_159" + original_name = "batch_norm2d_39.b_0" shape = [224] dtype = "float32" min_val = float("-5.91798") @@ -1544,6 +1704,7 @@ class Program_weight_tensor_parameter_159: class Program_weight_tensor_parameter_160: name = "parameter_160" + original_name = "batch_norm2d_39.w_0" shape = [224] dtype = "float32" min_val = float("-0.000620959") @@ -1555,6 +1716,7 @@ class Program_weight_tensor_parameter_160: class Program_weight_tensor_parameter_161: name = "parameter_161" + original_name = "batch_norm2d_39.w_2" shape = [224] dtype = "float32" min_val = float("5.60519e-45") @@ -1566,6 +1728,7 @@ class Program_weight_tensor_parameter_161: class Program_weight_tensor_parameter_162: name = "parameter_162" + original_name = "batch_norm2d_39.w_1" shape = [224] dtype = "float32" min_val = float("-0.808777") @@ -1577,6 +1740,7 @@ class Program_weight_tensor_parameter_162: class Program_weight_tensor_parameter_163: name = "parameter_163" + original_name = "conv2d_43.w_0" shape = [224, 224, 3, 3] dtype = "float32" min_val = float("-0.538942") @@ -1588,6 +1752,7 @@ class Program_weight_tensor_parameter_163: class Program_weight_tensor_parameter_164: name = "parameter_164" + original_name = "batch_norm2d_38.b_0" shape = [224] dtype = "float32" min_val = float("-5.11691") @@ -1599,6 +1764,7 @@ class Program_weight_tensor_parameter_164: class Program_weight_tensor_parameter_165: name = "parameter_165" + original_name = "batch_norm2d_38.w_0" shape = [224] dtype = "float32" min_val = float("0.000165796") @@ -1610,6 +1776,7 @@ class Program_weight_tensor_parameter_165: class Program_weight_tensor_parameter_166: name = "parameter_166" + original_name = "batch_norm2d_38.w_2" shape = [224] dtype = "float32" min_val = float("5.60519e-45") @@ -1621,6 +1788,7 @@ class Program_weight_tensor_parameter_166: class Program_weight_tensor_parameter_167: name = "parameter_167" + original_name = "batch_norm2d_38.w_1" shape = [224] dtype = "float32" min_val = float("-2.21212") @@ -1632,6 +1800,7 @@ class Program_weight_tensor_parameter_167: class Program_weight_tensor_parameter_168: name = "parameter_168" + original_name = "conv2d_42.w_0" shape = [224, 224, 3, 3] dtype = "float32" min_val = float("-0.468705") @@ -1643,6 +1812,7 @@ class Program_weight_tensor_parameter_168: class Program_weight_tensor_parameter_169: name = "parameter_169" + original_name = "batch_norm2d_37.b_0" shape = [224] dtype = "float32" min_val = float("-5.48244") @@ -1654,6 +1824,7 @@ class Program_weight_tensor_parameter_169: class Program_weight_tensor_parameter_170: name = "parameter_170" + original_name = "batch_norm2d_37.w_0" shape = [224] dtype = "float32" min_val = float("0.00247109") @@ -1665,6 +1836,7 @@ class Program_weight_tensor_parameter_170: class Program_weight_tensor_parameter_171: name = "parameter_171" + original_name = "batch_norm2d_37.w_2" shape = [224] dtype = "float32" min_val = float("5.60519e-45") @@ -1676,6 +1848,7 @@ class Program_weight_tensor_parameter_171: class Program_weight_tensor_parameter_172: name = "parameter_172" + original_name = "batch_norm2d_37.w_1" shape = [224] dtype = "float32" min_val = float("-4.85184") @@ -1687,6 +1860,7 @@ class Program_weight_tensor_parameter_172: class Program_weight_tensor_parameter_173: name = "parameter_173" + original_name = "conv2d_41.w_0" shape = [224, 224, 3, 3] dtype = "float32" min_val = float("-0.473897") @@ -1698,6 +1872,7 @@ class Program_weight_tensor_parameter_173: class Program_weight_tensor_parameter_174: name = "parameter_174" + original_name = "batch_norm2d_36.b_0" shape = [224] dtype = "float32" min_val = float("-5.7064") @@ -1709,6 +1884,7 @@ class Program_weight_tensor_parameter_174: class Program_weight_tensor_parameter_175: name = "parameter_175" + original_name = "batch_norm2d_36.w_0" shape = [224] dtype = "float32" min_val = float("-0.000526838") @@ -1720,6 +1896,7 @@ class Program_weight_tensor_parameter_175: class Program_weight_tensor_parameter_176: name = "parameter_176" + original_name = "batch_norm2d_36.w_2" shape = [224] dtype = "float32" min_val = float("5.60519e-45") @@ -1731,6 +1908,7 @@ class Program_weight_tensor_parameter_176: class Program_weight_tensor_parameter_177: name = "parameter_177" + original_name = "batch_norm2d_36.w_1" shape = [224] dtype = "float32" min_val = float("-7.54952") @@ -1742,6 +1920,7 @@ class Program_weight_tensor_parameter_177: class Program_weight_tensor_parameter_178: name = "parameter_178" + original_name = "conv2d_40.w_0" shape = [224, 224, 3, 3] dtype = "float32" min_val = float("-0.600621") @@ -1753,6 +1932,7 @@ class Program_weight_tensor_parameter_178: class Program_weight_tensor_parameter_179: name = "parameter_179" + original_name = "batch_norm2d_35.b_0" shape = [224] dtype = "float32" min_val = float("-6.21687") @@ -1764,6 +1944,7 @@ class Program_weight_tensor_parameter_179: class Program_weight_tensor_parameter_180: name = "parameter_180" + original_name = "batch_norm2d_35.w_0" shape = [224] dtype = "float32" min_val = float("-0.000430432") @@ -1775,6 +1956,7 @@ class Program_weight_tensor_parameter_180: class Program_weight_tensor_parameter_181: name = "parameter_181" + original_name = "batch_norm2d_35.w_2" shape = [224] dtype = "float32" min_val = float("5.60519e-45") @@ -1786,6 +1968,7 @@ class Program_weight_tensor_parameter_181: class Program_weight_tensor_parameter_182: name = "parameter_182" + original_name = "batch_norm2d_35.w_1" shape = [224] dtype = "float32" min_val = float("-16.3168") @@ -1797,6 +1980,7 @@ class Program_weight_tensor_parameter_182: class Program_weight_tensor_parameter_183: name = "parameter_183" + original_name = "conv2d_39.w_0" shape = [224, 224, 3, 3] dtype = "float32" min_val = float("-0.889788") @@ -1808,6 +1992,7 @@ class Program_weight_tensor_parameter_183: class Program_weight_tensor_parameter_184: name = "parameter_184" + original_name = "batch_norm2d_34.b_0" shape = [224] dtype = "float32" min_val = float("-10.5963") @@ -1819,6 +2004,7 @@ class Program_weight_tensor_parameter_184: class Program_weight_tensor_parameter_185: name = "parameter_185" + original_name = "batch_norm2d_34.w_0" shape = [224] dtype = "float32" min_val = float("-0.000383113") @@ -1830,6 +2016,7 @@ class Program_weight_tensor_parameter_185: class Program_weight_tensor_parameter_186: name = "parameter_186" + original_name = "batch_norm2d_34.w_2" shape = [224] dtype = "float32" min_val = float("5.60519e-45") @@ -1841,6 +2028,7 @@ class Program_weight_tensor_parameter_186: class Program_weight_tensor_parameter_187: name = "parameter_187" + original_name = "batch_norm2d_34.w_1" shape = [224] dtype = "float32" min_val = float("-38.5701") @@ -1852,6 +2040,7 @@ class Program_weight_tensor_parameter_187: class Program_weight_tensor_parameter_188: name = "parameter_188" + original_name = "conv2d_38.w_0" shape = [224, 768, 3, 3] dtype = "float32" min_val = float("-1.17797") @@ -1863,6 +2052,7 @@ class Program_weight_tensor_parameter_188: class Program_weight_tensor_parameter_189: name = "parameter_189" + original_name = "batch_norm2d_33.b_0" shape = [768] dtype = "float32" min_val = float("-2.56599") @@ -1874,6 +2064,7 @@ class Program_weight_tensor_parameter_189: class Program_weight_tensor_parameter_190: name = "parameter_190" + original_name = "batch_norm2d_33.w_0" shape = [768] dtype = "float32" min_val = float("-0.0167131") @@ -1885,6 +2076,7 @@ class Program_weight_tensor_parameter_190: class Program_weight_tensor_parameter_191: name = "parameter_191" + original_name = "batch_norm2d_33.w_2" shape = [768] dtype = "float32" min_val = float("5.60519e-45") @@ -1896,6 +2088,7 @@ class Program_weight_tensor_parameter_191: class Program_weight_tensor_parameter_192: name = "parameter_192" + original_name = "batch_norm2d_33.w_1" shape = [768] dtype = "float32" min_val = float("-0.865777") @@ -1907,6 +2100,7 @@ class Program_weight_tensor_parameter_192: class Program_weight_tensor_parameter_193: name = "parameter_193" + original_name = "conv2d_37.w_0" shape = [768, 1, 3, 3] dtype = "float32" min_val = float("-0.583948") @@ -1918,6 +2112,7 @@ class Program_weight_tensor_parameter_193: class Program_weight_tensor_parameter_194: name = "parameter_194" + original_name = "conv2d_36.b_0" shape = [768] dtype = "float32" min_val = float("-0.511435") @@ -1929,6 +2124,7 @@ class Program_weight_tensor_parameter_194: class Program_weight_tensor_parameter_195: name = "parameter_195" + original_name = "conv2d_36.w_0" shape = [768, 768, 1, 1] dtype = "float32" min_val = float("-1.77359") @@ -1940,6 +2136,7 @@ class Program_weight_tensor_parameter_195: class Program_weight_tensor_parameter_196: name = "parameter_196" + original_name = "batch_norm2d_32.b_0" shape = [768] dtype = "float32" min_val = float("-12.9412") @@ -1951,6 +2148,7 @@ class Program_weight_tensor_parameter_196: class Program_weight_tensor_parameter_197: name = "parameter_197" + original_name = "batch_norm2d_32.w_0" shape = [768] dtype = "float32" min_val = float("0.0351857") @@ -1962,6 +2160,7 @@ class Program_weight_tensor_parameter_197: class Program_weight_tensor_parameter_198: name = "parameter_198" + original_name = "batch_norm2d_32.w_2" shape = [768] dtype = "float32" min_val = float("5.60519e-45") @@ -1973,6 +2172,7 @@ class Program_weight_tensor_parameter_198: class Program_weight_tensor_parameter_199: name = "parameter_199" + original_name = "batch_norm2d_32.w_1" shape = [768] dtype = "float32" min_val = float("-5.99008") @@ -1984,6 +2184,7 @@ class Program_weight_tensor_parameter_199: class Program_weight_tensor_parameter_200: name = "parameter_200" + original_name = "conv2d_35.w_0" shape = [768, 1920, 1, 1] dtype = "float32" min_val = float("-1.49255") @@ -1995,6 +2196,7 @@ class Program_weight_tensor_parameter_200: class Program_weight_tensor_parameter_201: name = "parameter_201" + original_name = "batch_norm2d_31.b_0" shape = [192] dtype = "float32" min_val = float("-3.24497") @@ -2006,6 +2208,7 @@ class Program_weight_tensor_parameter_201: class Program_weight_tensor_parameter_202: name = "parameter_202" + original_name = "batch_norm2d_31.w_0" shape = [192] dtype = "float32" min_val = float("0.0236468") @@ -2017,6 +2220,7 @@ class Program_weight_tensor_parameter_202: class Program_weight_tensor_parameter_203: name = "parameter_203" + original_name = "batch_norm2d_31.w_2" shape = [192] dtype = "float32" min_val = float("5.60519e-45") @@ -2028,6 +2232,7 @@ class Program_weight_tensor_parameter_203: class Program_weight_tensor_parameter_204: name = "parameter_204" + original_name = "batch_norm2d_31.w_1" shape = [192] dtype = "float32" min_val = float("-2.30075") @@ -2039,6 +2244,7 @@ class Program_weight_tensor_parameter_204: class Program_weight_tensor_parameter_205: name = "parameter_205" + original_name = "conv2d_34.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.801362") @@ -2050,6 +2256,7 @@ class Program_weight_tensor_parameter_205: class Program_weight_tensor_parameter_206: name = "parameter_206" + original_name = "batch_norm2d_30.b_0" shape = [192] dtype = "float32" min_val = float("-2.64017") @@ -2061,6 +2268,7 @@ class Program_weight_tensor_parameter_206: class Program_weight_tensor_parameter_207: name = "parameter_207" + original_name = "batch_norm2d_30.w_0" shape = [192] dtype = "float32" min_val = float("0.000159657") @@ -2072,6 +2280,7 @@ class Program_weight_tensor_parameter_207: class Program_weight_tensor_parameter_208: name = "parameter_208" + original_name = "batch_norm2d_30.w_2" shape = [192] dtype = "float32" min_val = float("5.60519e-45") @@ -2083,6 +2292,7 @@ class Program_weight_tensor_parameter_208: class Program_weight_tensor_parameter_209: name = "parameter_209" + original_name = "batch_norm2d_30.w_1" shape = [192] dtype = "float32" min_val = float("-4.48323") @@ -2094,6 +2304,7 @@ class Program_weight_tensor_parameter_209: class Program_weight_tensor_parameter_210: name = "parameter_210" + original_name = "conv2d_33.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.863938") @@ -2105,6 +2316,7 @@ class Program_weight_tensor_parameter_210: class Program_weight_tensor_parameter_211: name = "parameter_211" + original_name = "batch_norm2d_29.b_0" shape = [192] dtype = "float32" min_val = float("-3.9513") @@ -2116,6 +2328,7 @@ class Program_weight_tensor_parameter_211: class Program_weight_tensor_parameter_212: name = "parameter_212" + original_name = "batch_norm2d_29.w_0" shape = [192] dtype = "float32" min_val = float("0.0279247") @@ -2127,6 +2340,7 @@ class Program_weight_tensor_parameter_212: class Program_weight_tensor_parameter_213: name = "parameter_213" + original_name = "batch_norm2d_29.w_2" shape = [192] dtype = "float32" min_val = float("5.60519e-45") @@ -2138,6 +2352,7 @@ class Program_weight_tensor_parameter_213: class Program_weight_tensor_parameter_214: name = "parameter_214" + original_name = "batch_norm2d_29.w_1" shape = [192] dtype = "float32" min_val = float("-4.40015") @@ -2149,6 +2364,7 @@ class Program_weight_tensor_parameter_214: class Program_weight_tensor_parameter_215: name = "parameter_215" + original_name = "conv2d_32.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.924322") @@ -2160,6 +2376,7 @@ class Program_weight_tensor_parameter_215: class Program_weight_tensor_parameter_216: name = "parameter_216" + original_name = "batch_norm2d_28.b_0" shape = [192] dtype = "float32" min_val = float("-3.97523") @@ -2171,6 +2388,7 @@ class Program_weight_tensor_parameter_216: class Program_weight_tensor_parameter_217: name = "parameter_217" + original_name = "batch_norm2d_28.w_0" shape = [192] dtype = "float32" min_val = float("-0.00527014") @@ -2182,6 +2400,7 @@ class Program_weight_tensor_parameter_217: class Program_weight_tensor_parameter_218: name = "parameter_218" + original_name = "batch_norm2d_28.w_2" shape = [192] dtype = "float32" min_val = float("5.60519e-45") @@ -2193,6 +2412,7 @@ class Program_weight_tensor_parameter_218: class Program_weight_tensor_parameter_219: name = "parameter_219" + original_name = "batch_norm2d_28.w_1" shape = [192] dtype = "float32" min_val = float("-9.99067") @@ -2204,6 +2424,7 @@ class Program_weight_tensor_parameter_219: class Program_weight_tensor_parameter_220: name = "parameter_220" + original_name = "conv2d_31.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.861987") @@ -2215,6 +2436,7 @@ class Program_weight_tensor_parameter_220: class Program_weight_tensor_parameter_221: name = "parameter_221" + original_name = "batch_norm2d_27.b_0" shape = [192] dtype = "float32" min_val = float("-3.9103") @@ -2226,6 +2448,7 @@ class Program_weight_tensor_parameter_221: class Program_weight_tensor_parameter_222: name = "parameter_222" + original_name = "batch_norm2d_27.w_0" shape = [192] dtype = "float32" min_val = float("0.0167882") @@ -2237,6 +2460,7 @@ class Program_weight_tensor_parameter_222: class Program_weight_tensor_parameter_223: name = "parameter_223" + original_name = "batch_norm2d_27.w_2" shape = [192] dtype = "float32" min_val = float("5.60519e-45") @@ -2248,6 +2472,7 @@ class Program_weight_tensor_parameter_223: class Program_weight_tensor_parameter_224: name = "parameter_224" + original_name = "batch_norm2d_27.w_1" shape = [192] dtype = "float32" min_val = float("-10.0219") @@ -2259,6 +2484,7 @@ class Program_weight_tensor_parameter_224: class Program_weight_tensor_parameter_225: name = "parameter_225" + original_name = "conv2d_30.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.932144") @@ -2270,6 +2496,7 @@ class Program_weight_tensor_parameter_225: class Program_weight_tensor_parameter_226: name = "parameter_226" + original_name = "batch_norm2d_26.b_0" shape = [192] dtype = "float32" min_val = float("-5.89608") @@ -2281,6 +2508,7 @@ class Program_weight_tensor_parameter_226: class Program_weight_tensor_parameter_227: name = "parameter_227" + original_name = "batch_norm2d_26.w_0" shape = [192] dtype = "float32" min_val = float("0.00167011") @@ -2292,6 +2520,7 @@ class Program_weight_tensor_parameter_227: class Program_weight_tensor_parameter_228: name = "parameter_228" + original_name = "batch_norm2d_26.w_2" shape = [192] dtype = "float32" min_val = float("5.60519e-45") @@ -2303,6 +2532,7 @@ class Program_weight_tensor_parameter_228: class Program_weight_tensor_parameter_229: name = "parameter_229" + original_name = "batch_norm2d_26.w_1" shape = [192] dtype = "float32" min_val = float("-16.6163") @@ -2314,6 +2544,7 @@ class Program_weight_tensor_parameter_229: class Program_weight_tensor_parameter_230: name = "parameter_230" + original_name = "conv2d_29.w_0" shape = [192, 768, 3, 3] dtype = "float32" min_val = float("-0.886818") @@ -2325,6 +2556,7 @@ class Program_weight_tensor_parameter_230: class Program_weight_tensor_parameter_231: name = "parameter_231" + original_name = "conv2d_28.b_0" shape = [768] dtype = "float32" min_val = float("-0.284795") @@ -2336,6 +2568,7 @@ class Program_weight_tensor_parameter_231: class Program_weight_tensor_parameter_232: name = "parameter_232" + original_name = "conv2d_28.w_0" shape = [768, 768, 1, 1] dtype = "float32" min_val = float("-2.49026") @@ -2347,6 +2580,7 @@ class Program_weight_tensor_parameter_232: class Program_weight_tensor_parameter_233: name = "parameter_233" + original_name = "batch_norm2d_25.b_0" shape = [768] dtype = "float32" min_val = float("-13.6146") @@ -2358,6 +2592,7 @@ class Program_weight_tensor_parameter_233: class Program_weight_tensor_parameter_234: name = "parameter_234" + original_name = "batch_norm2d_25.w_0" shape = [768] dtype = "float32" min_val = float("-0.00392536") @@ -2369,6 +2604,7 @@ class Program_weight_tensor_parameter_234: class Program_weight_tensor_parameter_235: name = "parameter_235" + original_name = "batch_norm2d_25.w_2" shape = [768] dtype = "float32" min_val = float("5.60519e-45") @@ -2380,6 +2616,7 @@ class Program_weight_tensor_parameter_235: class Program_weight_tensor_parameter_236: name = "parameter_236" + original_name = "batch_norm2d_25.w_1" shape = [768] dtype = "float32" min_val = float("-24.9875") @@ -2391,6 +2628,7 @@ class Program_weight_tensor_parameter_236: class Program_weight_tensor_parameter_237: name = "parameter_237" + original_name = "conv2d_27.w_0" shape = [768, 1664, 1, 1] dtype = "float32" min_val = float("-1.63401") @@ -2402,6 +2640,7 @@ class Program_weight_tensor_parameter_237: class Program_weight_tensor_parameter_238: name = "parameter_238" + original_name = "batch_norm2d_24.b_0" shape = [192] dtype = "float32" min_val = float("-6.71259") @@ -2413,6 +2652,7 @@ class Program_weight_tensor_parameter_238: class Program_weight_tensor_parameter_239: name = "parameter_239" + original_name = "batch_norm2d_24.w_0" shape = [192] dtype = "float32" min_val = float("0.55321") @@ -2424,6 +2664,7 @@ class Program_weight_tensor_parameter_239: class Program_weight_tensor_parameter_240: name = "parameter_240" + original_name = "batch_norm2d_24.w_2" shape = [192] dtype = "float32" min_val = float("5.60519e-45") @@ -2435,6 +2676,7 @@ class Program_weight_tensor_parameter_240: class Program_weight_tensor_parameter_241: name = "parameter_241" + original_name = "batch_norm2d_24.w_1" shape = [192] dtype = "float32" min_val = float("-11.1197") @@ -2446,6 +2688,7 @@ class Program_weight_tensor_parameter_241: class Program_weight_tensor_parameter_242: name = "parameter_242" + original_name = "conv2d_26.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.980829") @@ -2457,6 +2700,7 @@ class Program_weight_tensor_parameter_242: class Program_weight_tensor_parameter_243: name = "parameter_243" + original_name = "batch_norm2d_23.b_0" shape = [192] dtype = "float32" min_val = float("-6.04808") @@ -2468,6 +2712,7 @@ class Program_weight_tensor_parameter_243: class Program_weight_tensor_parameter_244: name = "parameter_244" + original_name = "batch_norm2d_23.w_0" shape = [192] dtype = "float32" min_val = float("-0.000255708") @@ -2479,6 +2724,7 @@ class Program_weight_tensor_parameter_244: class Program_weight_tensor_parameter_245: name = "parameter_245" + original_name = "batch_norm2d_23.w_2" shape = [192] dtype = "float32" min_val = float("5.60519e-45") @@ -2490,6 +2736,7 @@ class Program_weight_tensor_parameter_245: class Program_weight_tensor_parameter_246: name = "parameter_246" + original_name = "batch_norm2d_23.w_1" shape = [192] dtype = "float32" min_val = float("-28.2937") @@ -2501,6 +2748,7 @@ class Program_weight_tensor_parameter_246: class Program_weight_tensor_parameter_247: name = "parameter_247" + original_name = "conv2d_25.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-1.0646") @@ -2512,6 +2760,7 @@ class Program_weight_tensor_parameter_247: class Program_weight_tensor_parameter_248: name = "parameter_248" + original_name = "batch_norm2d_22.b_0" shape = [192] dtype = "float32" min_val = float("-5.57104") @@ -2523,6 +2772,7 @@ class Program_weight_tensor_parameter_248: class Program_weight_tensor_parameter_249: name = "parameter_249" + original_name = "batch_norm2d_22.w_0" shape = [192] dtype = "float32" min_val = float("0.232818") @@ -2534,6 +2784,7 @@ class Program_weight_tensor_parameter_249: class Program_weight_tensor_parameter_250: name = "parameter_250" + original_name = "batch_norm2d_22.w_2" shape = [192] dtype = "float32" min_val = float("5.60519e-45") @@ -2545,6 +2796,7 @@ class Program_weight_tensor_parameter_250: class Program_weight_tensor_parameter_251: name = "parameter_251" + original_name = "batch_norm2d_22.w_1" shape = [192] dtype = "float32" min_val = float("-13.933") @@ -2556,6 +2808,7 @@ class Program_weight_tensor_parameter_251: class Program_weight_tensor_parameter_252: name = "parameter_252" + original_name = "conv2d_24.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-1.2683") @@ -2567,6 +2820,7 @@ class Program_weight_tensor_parameter_252: class Program_weight_tensor_parameter_253: name = "parameter_253" + original_name = "batch_norm2d_21.b_0" shape = [192] dtype = "float32" min_val = float("-4.89989") @@ -2578,6 +2832,7 @@ class Program_weight_tensor_parameter_253: class Program_weight_tensor_parameter_254: name = "parameter_254" + original_name = "batch_norm2d_21.w_0" shape = [192] dtype = "float32" min_val = float("-0.000456049") @@ -2589,6 +2844,7 @@ class Program_weight_tensor_parameter_254: class Program_weight_tensor_parameter_255: name = "parameter_255" + original_name = "batch_norm2d_21.w_2" shape = [192] dtype = "float32" min_val = float("5.60519e-45") @@ -2600,6 +2856,7 @@ class Program_weight_tensor_parameter_255: class Program_weight_tensor_parameter_256: name = "parameter_256" + original_name = "batch_norm2d_21.w_1" shape = [192] dtype = "float32" min_val = float("-30.1144") @@ -2611,6 +2868,7 @@ class Program_weight_tensor_parameter_256: class Program_weight_tensor_parameter_257: name = "parameter_257" + original_name = "conv2d_23.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-1.30699") @@ -2622,6 +2880,7 @@ class Program_weight_tensor_parameter_257: class Program_weight_tensor_parameter_258: name = "parameter_258" + original_name = "batch_norm2d_20.b_0" shape = [192] dtype = "float32" min_val = float("-6.92109") @@ -2633,6 +2892,7 @@ class Program_weight_tensor_parameter_258: class Program_weight_tensor_parameter_259: name = "parameter_259" + original_name = "batch_norm2d_20.w_0" shape = [192] dtype = "float32" min_val = float("0.354047") @@ -2644,6 +2904,7 @@ class Program_weight_tensor_parameter_259: class Program_weight_tensor_parameter_260: name = "parameter_260" + original_name = "batch_norm2d_20.w_2" shape = [192] dtype = "float32" min_val = float("5.60519e-45") @@ -2655,6 +2916,7 @@ class Program_weight_tensor_parameter_260: class Program_weight_tensor_parameter_261: name = "parameter_261" + original_name = "batch_norm2d_20.w_1" shape = [192] dtype = "float32" min_val = float("-31.7516") @@ -2666,6 +2928,7 @@ class Program_weight_tensor_parameter_261: class Program_weight_tensor_parameter_262: name = "parameter_262" + original_name = "conv2d_22.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-1.25587") @@ -2677,6 +2940,7 @@ class Program_weight_tensor_parameter_262: class Program_weight_tensor_parameter_263: name = "parameter_263" + original_name = "batch_norm2d_19.b_0" shape = [192] dtype = "float32" min_val = float("-10.1758") @@ -2688,6 +2952,7 @@ class Program_weight_tensor_parameter_263: class Program_weight_tensor_parameter_264: name = "parameter_264" + original_name = "batch_norm2d_19.w_0" shape = [192] dtype = "float32" min_val = float("0.443323") @@ -2699,6 +2964,7 @@ class Program_weight_tensor_parameter_264: class Program_weight_tensor_parameter_265: name = "parameter_265" + original_name = "batch_norm2d_19.w_2" shape = [192] dtype = "float32" min_val = float("5.60519e-45") @@ -2710,6 +2976,7 @@ class Program_weight_tensor_parameter_265: class Program_weight_tensor_parameter_266: name = "parameter_266" + original_name = "batch_norm2d_19.w_1" shape = [192] dtype = "float32" min_val = float("-59.0023") @@ -2721,6 +2988,7 @@ class Program_weight_tensor_parameter_266: class Program_weight_tensor_parameter_267: name = "parameter_267" + original_name = "conv2d_21.w_0" shape = [192, 512, 3, 3] dtype = "float32" min_val = float("-1.53004") @@ -2732,6 +3000,7 @@ class Program_weight_tensor_parameter_267: class Program_weight_tensor_parameter_268: name = "parameter_268" + original_name = "batch_norm2d_18.b_0" shape = [512] dtype = "float32" min_val = float("-3.50587") @@ -2743,6 +3012,7 @@ class Program_weight_tensor_parameter_268: class Program_weight_tensor_parameter_269: name = "parameter_269" + original_name = "batch_norm2d_18.w_0" shape = [512] dtype = "float32" min_val = float("-0.0095385") @@ -2754,6 +3024,7 @@ class Program_weight_tensor_parameter_269: class Program_weight_tensor_parameter_270: name = "parameter_270" + original_name = "batch_norm2d_18.w_2" shape = [512] dtype = "float32" min_val = float("5.60519e-45") @@ -2765,6 +3036,7 @@ class Program_weight_tensor_parameter_270: class Program_weight_tensor_parameter_271: name = "parameter_271" + original_name = "batch_norm2d_18.w_1" shape = [512] dtype = "float32" min_val = float("-0.86961") @@ -2776,6 +3048,7 @@ class Program_weight_tensor_parameter_271: class Program_weight_tensor_parameter_272: name = "parameter_272" + original_name = "conv2d_20.w_0" shape = [512, 1, 3, 3] dtype = "float32" min_val = float("-0.644153") @@ -2787,6 +3060,7 @@ class Program_weight_tensor_parameter_272: class Program_weight_tensor_parameter_273: name = "parameter_273" + original_name = "conv2d_19.b_0" shape = [512] dtype = "float32" min_val = float("-0.35841") @@ -2798,6 +3072,7 @@ class Program_weight_tensor_parameter_273: class Program_weight_tensor_parameter_274: name = "parameter_274" + original_name = "conv2d_19.w_0" shape = [512, 512, 1, 1] dtype = "float32" min_val = float("-1.75336") @@ -2809,6 +3084,7 @@ class Program_weight_tensor_parameter_274: class Program_weight_tensor_parameter_275: name = "parameter_275" + original_name = "batch_norm2d_17.b_0" shape = [512] dtype = "float32" min_val = float("-14.7244") @@ -2820,6 +3096,7 @@ class Program_weight_tensor_parameter_275: class Program_weight_tensor_parameter_276: name = "parameter_276" + original_name = "batch_norm2d_17.w_0" shape = [512] dtype = "float32" min_val = float("0.0720109") @@ -2831,6 +3108,7 @@ class Program_weight_tensor_parameter_276: class Program_weight_tensor_parameter_277: name = "parameter_277" + original_name = "batch_norm2d_17.w_2" shape = [512] dtype = "float32" min_val = float("5.60519e-45") @@ -2842,6 +3120,7 @@ class Program_weight_tensor_parameter_277: class Program_weight_tensor_parameter_278: name = "parameter_278" + original_name = "batch_norm2d_17.w_1" shape = [512] dtype = "float32" min_val = float("-21.4041") @@ -2853,6 +3132,7 @@ class Program_weight_tensor_parameter_278: class Program_weight_tensor_parameter_279: name = "parameter_279" + original_name = "conv2d_18.w_0" shape = [512, 1216, 1, 1] dtype = "float32" min_val = float("-1.50117") @@ -2864,6 +3144,7 @@ class Program_weight_tensor_parameter_279: class Program_weight_tensor_parameter_280: name = "parameter_280" + original_name = "batch_norm2d_16.b_0" shape = [160] dtype = "float32" min_val = float("-5.22508") @@ -2875,6 +3156,7 @@ class Program_weight_tensor_parameter_280: class Program_weight_tensor_parameter_281: name = "parameter_281" + original_name = "batch_norm2d_16.w_0" shape = [160] dtype = "float32" min_val = float("1.40309") @@ -2886,6 +3168,7 @@ class Program_weight_tensor_parameter_281: class Program_weight_tensor_parameter_282: name = "parameter_282" + original_name = "batch_norm2d_16.w_2" shape = [160] dtype = "float32" min_val = float("2.57694") @@ -2897,6 +3180,7 @@ class Program_weight_tensor_parameter_282: class Program_weight_tensor_parameter_283: name = "parameter_283" + original_name = "batch_norm2d_16.w_1" shape = [160] dtype = "float32" min_val = float("-16.9196") @@ -2908,6 +3192,7 @@ class Program_weight_tensor_parameter_283: class Program_weight_tensor_parameter_284: name = "parameter_284" + original_name = "conv2d_17.w_0" shape = [160, 160, 3, 3] dtype = "float32" min_val = float("-1.03959") @@ -2919,6 +3204,7 @@ class Program_weight_tensor_parameter_284: class Program_weight_tensor_parameter_285: name = "parameter_285" + original_name = "batch_norm2d_15.b_0" shape = [160] dtype = "float32" min_val = float("-6.12315") @@ -2930,6 +3216,7 @@ class Program_weight_tensor_parameter_285: class Program_weight_tensor_parameter_286: name = "parameter_286" + original_name = "batch_norm2d_15.w_0" shape = [160] dtype = "float32" min_val = float("0.554147") @@ -2941,6 +3228,7 @@ class Program_weight_tensor_parameter_286: class Program_weight_tensor_parameter_287: name = "parameter_287" + original_name = "batch_norm2d_15.w_2" shape = [160] dtype = "float32" min_val = float("5.60519e-45") @@ -2952,6 +3240,7 @@ class Program_weight_tensor_parameter_287: class Program_weight_tensor_parameter_288: name = "parameter_288" + original_name = "batch_norm2d_15.w_1" shape = [160] dtype = "float32" min_val = float("-21.6958") @@ -2963,6 +3252,7 @@ class Program_weight_tensor_parameter_288: class Program_weight_tensor_parameter_289: name = "parameter_289" + original_name = "conv2d_16.w_0" shape = [160, 160, 3, 3] dtype = "float32" min_val = float("-1.20371") @@ -2974,6 +3264,7 @@ class Program_weight_tensor_parameter_289: class Program_weight_tensor_parameter_290: name = "parameter_290" + original_name = "batch_norm2d_14.b_0" shape = [160] dtype = "float32" min_val = float("-6.35881") @@ -2985,6 +3276,7 @@ class Program_weight_tensor_parameter_290: class Program_weight_tensor_parameter_291: name = "parameter_291" + original_name = "batch_norm2d_14.w_0" shape = [160] dtype = "float32" min_val = float("0.945639") @@ -2996,6 +3288,7 @@ class Program_weight_tensor_parameter_291: class Program_weight_tensor_parameter_292: name = "parameter_292" + original_name = "batch_norm2d_14.w_2" shape = [160] dtype = "float32" min_val = float("5.60519e-45") @@ -3007,6 +3300,7 @@ class Program_weight_tensor_parameter_292: class Program_weight_tensor_parameter_293: name = "parameter_293" + original_name = "batch_norm2d_14.w_1" shape = [160] dtype = "float32" min_val = float("-18.6845") @@ -3018,6 +3312,7 @@ class Program_weight_tensor_parameter_293: class Program_weight_tensor_parameter_294: name = "parameter_294" + original_name = "conv2d_15.w_0" shape = [160, 160, 3, 3] dtype = "float32" min_val = float("-1.16422") @@ -3029,6 +3324,7 @@ class Program_weight_tensor_parameter_294: class Program_weight_tensor_parameter_295: name = "parameter_295" + original_name = "batch_norm2d_13.b_0" shape = [160] dtype = "float32" min_val = float("-6.40283") @@ -3040,6 +3336,7 @@ class Program_weight_tensor_parameter_295: class Program_weight_tensor_parameter_296: name = "parameter_296" + original_name = "batch_norm2d_13.w_0" shape = [160] dtype = "float32" min_val = float("0.26433") @@ -3051,6 +3348,7 @@ class Program_weight_tensor_parameter_296: class Program_weight_tensor_parameter_297: name = "parameter_297" + original_name = "batch_norm2d_13.w_2" shape = [160] dtype = "float32" min_val = float("5.60519e-45") @@ -3062,6 +3360,7 @@ class Program_weight_tensor_parameter_297: class Program_weight_tensor_parameter_298: name = "parameter_298" + original_name = "batch_norm2d_13.w_1" shape = [160] dtype = "float32" min_val = float("-51.5202") @@ -3073,6 +3372,7 @@ class Program_weight_tensor_parameter_298: class Program_weight_tensor_parameter_299: name = "parameter_299" + original_name = "conv2d_14.w_0" shape = [160, 160, 3, 3] dtype = "float32" min_val = float("-1.19949") @@ -3084,6 +3384,7 @@ class Program_weight_tensor_parameter_299: class Program_weight_tensor_parameter_300: name = "parameter_300" + original_name = "batch_norm2d_12.b_0" shape = [160] dtype = "float32" min_val = float("-5.6566") @@ -3095,6 +3396,7 @@ class Program_weight_tensor_parameter_300: class Program_weight_tensor_parameter_301: name = "parameter_301" + original_name = "batch_norm2d_12.w_0" shape = [160] dtype = "float32" min_val = float("0.67174") @@ -3106,6 +3408,7 @@ class Program_weight_tensor_parameter_301: class Program_weight_tensor_parameter_302: name = "parameter_302" + original_name = "batch_norm2d_12.w_2" shape = [160] dtype = "float32" min_val = float("5.2232") @@ -3117,6 +3420,7 @@ class Program_weight_tensor_parameter_302: class Program_weight_tensor_parameter_303: name = "parameter_303" + original_name = "batch_norm2d_12.w_1" shape = [160] dtype = "float32" min_val = float("-44.9604") @@ -3128,6 +3432,7 @@ class Program_weight_tensor_parameter_303: class Program_weight_tensor_parameter_304: name = "parameter_304" + original_name = "conv2d_13.w_0" shape = [160, 160, 3, 3] dtype = "float32" min_val = float("-1.16682") @@ -3139,6 +3444,7 @@ class Program_weight_tensor_parameter_304: class Program_weight_tensor_parameter_305: name = "parameter_305" + original_name = "batch_norm2d_11.b_0" shape = [160] dtype = "float32" min_val = float("-9.91221") @@ -3150,6 +3456,7 @@ class Program_weight_tensor_parameter_305: class Program_weight_tensor_parameter_306: name = "parameter_306" + original_name = "batch_norm2d_11.w_0" shape = [160] dtype = "float32" min_val = float("0.991402") @@ -3161,6 +3468,7 @@ class Program_weight_tensor_parameter_306: class Program_weight_tensor_parameter_307: name = "parameter_307" + original_name = "batch_norm2d_11.w_2" shape = [160] dtype = "float32" min_val = float("32.0167") @@ -3172,6 +3480,7 @@ class Program_weight_tensor_parameter_307: class Program_weight_tensor_parameter_308: name = "parameter_308" + original_name = "batch_norm2d_11.w_1" shape = [160] dtype = "float32" min_val = float("-41.1972") @@ -3183,6 +3492,7 @@ class Program_weight_tensor_parameter_308: class Program_weight_tensor_parameter_309: name = "parameter_309" + original_name = "conv2d_12.w_0" shape = [160, 256, 3, 3] dtype = "float32" min_val = float("-1.41866") @@ -3194,6 +3504,7 @@ class Program_weight_tensor_parameter_309: class Program_weight_tensor_parameter_310: name = "parameter_310" + original_name = "batch_norm2d_10.b_0" shape = [256] dtype = "float32" min_val = float("-2.66602") @@ -3205,6 +3516,7 @@ class Program_weight_tensor_parameter_310: class Program_weight_tensor_parameter_311: name = "parameter_311" + original_name = "batch_norm2d_10.w_0" shape = [256] dtype = "float32" min_val = float("-0.0206668") @@ -3216,6 +3528,7 @@ class Program_weight_tensor_parameter_311: class Program_weight_tensor_parameter_312: name = "parameter_312" + original_name = "batch_norm2d_10.w_2" shape = [256] dtype = "float32" min_val = float("5.60519e-45") @@ -3227,6 +3540,7 @@ class Program_weight_tensor_parameter_312: class Program_weight_tensor_parameter_313: name = "parameter_313" + original_name = "batch_norm2d_10.w_1" shape = [256] dtype = "float32" min_val = float("-0.74794") @@ -3238,6 +3552,7 @@ class Program_weight_tensor_parameter_313: class Program_weight_tensor_parameter_314: name = "parameter_314" + original_name = "conv2d_11.w_0" shape = [256, 1, 3, 3] dtype = "float32" min_val = float("-0.772341") @@ -3249,6 +3564,7 @@ class Program_weight_tensor_parameter_314: class Program_weight_tensor_parameter_315: name = "parameter_315" + original_name = "conv2d_10.b_0" shape = [256] dtype = "float32" min_val = float("-0.391773") @@ -3260,6 +3576,7 @@ class Program_weight_tensor_parameter_315: class Program_weight_tensor_parameter_316: name = "parameter_316" + original_name = "conv2d_10.w_0" shape = [256, 256, 1, 1] dtype = "float32" min_val = float("-2.09665") @@ -3271,6 +3588,7 @@ class Program_weight_tensor_parameter_316: class Program_weight_tensor_parameter_317: name = "parameter_317" + original_name = "batch_norm2d_9.b_0" shape = [256] dtype = "float32" min_val = float("-10.5484") @@ -3282,6 +3600,7 @@ class Program_weight_tensor_parameter_317: class Program_weight_tensor_parameter_318: name = "parameter_318" + original_name = "batch_norm2d_9.w_0" shape = [256] dtype = "float32" min_val = float("0.00168943") @@ -3293,6 +3612,7 @@ class Program_weight_tensor_parameter_318: class Program_weight_tensor_parameter_319: name = "parameter_319" + original_name = "batch_norm2d_9.w_2" shape = [256] dtype = "float32" min_val = float("5.60519e-45") @@ -3304,6 +3624,7 @@ class Program_weight_tensor_parameter_319: class Program_weight_tensor_parameter_320: name = "parameter_320" + original_name = "batch_norm2d_9.w_1" shape = [256] dtype = "float32" min_val = float("-25.8101") @@ -3315,6 +3636,7 @@ class Program_weight_tensor_parameter_320: class Program_weight_tensor_parameter_321: name = "parameter_321" + original_name = "conv2d_9.w_0" shape = [256, 896, 1, 1] dtype = "float32" min_val = float("-2.11031") @@ -3326,6 +3648,7 @@ class Program_weight_tensor_parameter_321: class Program_weight_tensor_parameter_322: name = "parameter_322" + original_name = "batch_norm2d_8.b_0" shape = [128] dtype = "float32" min_val = float("-6.48896") @@ -3337,6 +3660,7 @@ class Program_weight_tensor_parameter_322: class Program_weight_tensor_parameter_323: name = "parameter_323" + original_name = "batch_norm2d_8.w_0" shape = [128] dtype = "float32" min_val = float("1.37771") @@ -3348,6 +3672,7 @@ class Program_weight_tensor_parameter_323: class Program_weight_tensor_parameter_324: name = "parameter_324" + original_name = "batch_norm2d_8.w_2" shape = [128] dtype = "float32" min_val = float("3.49155") @@ -3359,6 +3684,7 @@ class Program_weight_tensor_parameter_324: class Program_weight_tensor_parameter_325: name = "parameter_325" + original_name = "batch_norm2d_8.w_1" shape = [128] dtype = "float32" min_val = float("-16.4816") @@ -3370,6 +3696,7 @@ class Program_weight_tensor_parameter_325: class Program_weight_tensor_parameter_326: name = "parameter_326" + original_name = "conv2d_8.w_0" shape = [128, 128, 3, 3] dtype = "float32" min_val = float("-0.860862") @@ -3381,6 +3708,7 @@ class Program_weight_tensor_parameter_326: class Program_weight_tensor_parameter_327: name = "parameter_327" + original_name = "batch_norm2d_7.b_0" shape = [128] dtype = "float32" min_val = float("-8.94491") @@ -3392,6 +3720,7 @@ class Program_weight_tensor_parameter_327: class Program_weight_tensor_parameter_328: name = "parameter_328" + original_name = "batch_norm2d_7.w_0" shape = [128] dtype = "float32" min_val = float("1.31775") @@ -3403,6 +3732,7 @@ class Program_weight_tensor_parameter_328: class Program_weight_tensor_parameter_329: name = "parameter_329" + original_name = "batch_norm2d_7.w_2" shape = [128] dtype = "float32" min_val = float("8.06828") @@ -3414,6 +3744,7 @@ class Program_weight_tensor_parameter_329: class Program_weight_tensor_parameter_330: name = "parameter_330" + original_name = "batch_norm2d_7.w_1" shape = [128] dtype = "float32" min_val = float("-31.6809") @@ -3425,6 +3756,7 @@ class Program_weight_tensor_parameter_330: class Program_weight_tensor_parameter_331: name = "parameter_331" + original_name = "conv2d_7.w_0" shape = [128, 128, 3, 3] dtype = "float32" min_val = float("-1.0386") @@ -3436,6 +3768,7 @@ class Program_weight_tensor_parameter_331: class Program_weight_tensor_parameter_332: name = "parameter_332" + original_name = "batch_norm2d_6.b_0" shape = [128] dtype = "float32" min_val = float("-6.10058") @@ -3447,6 +3780,7 @@ class Program_weight_tensor_parameter_332: class Program_weight_tensor_parameter_333: name = "parameter_333" + original_name = "batch_norm2d_6.w_0" shape = [128] dtype = "float32" min_val = float("1.17255") @@ -3458,6 +3792,7 @@ class Program_weight_tensor_parameter_333: class Program_weight_tensor_parameter_334: name = "parameter_334" + original_name = "batch_norm2d_6.w_2" shape = [128] dtype = "float32" min_val = float("36.7004") @@ -3469,6 +3804,7 @@ class Program_weight_tensor_parameter_334: class Program_weight_tensor_parameter_335: name = "parameter_335" + original_name = "batch_norm2d_6.w_1" shape = [128] dtype = "float32" min_val = float("-48.2903") @@ -3480,6 +3816,7 @@ class Program_weight_tensor_parameter_335: class Program_weight_tensor_parameter_336: name = "parameter_336" + original_name = "conv2d_6.w_0" shape = [128, 128, 3, 3] dtype = "float32" min_val = float("-1.22005") @@ -3491,6 +3828,7 @@ class Program_weight_tensor_parameter_336: class Program_weight_tensor_parameter_337: name = "parameter_337" + original_name = "batch_norm2d_5.b_0" shape = [128] dtype = "float32" min_val = float("-4.64819") @@ -3502,6 +3840,7 @@ class Program_weight_tensor_parameter_337: class Program_weight_tensor_parameter_338: name = "parameter_338" + original_name = "batch_norm2d_5.w_0" shape = [128] dtype = "float32" min_val = float("1.01447") @@ -3513,6 +3852,7 @@ class Program_weight_tensor_parameter_338: class Program_weight_tensor_parameter_339: name = "parameter_339" + original_name = "batch_norm2d_5.w_2" shape = [128] dtype = "float32" min_val = float("24.298") @@ -3524,6 +3864,7 @@ class Program_weight_tensor_parameter_339: class Program_weight_tensor_parameter_340: name = "parameter_340" + original_name = "batch_norm2d_5.w_1" shape = [128] dtype = "float32" min_val = float("-28.8593") @@ -3535,6 +3876,7 @@ class Program_weight_tensor_parameter_340: class Program_weight_tensor_parameter_341: name = "parameter_341" + original_name = "conv2d_5.w_0" shape = [128, 128, 3, 3] dtype = "float32" min_val = float("-1.40406") @@ -3546,6 +3888,7 @@ class Program_weight_tensor_parameter_341: class Program_weight_tensor_parameter_342: name = "parameter_342" + original_name = "batch_norm2d_4.b_0" shape = [128] dtype = "float32" min_val = float("-9.50367") @@ -3557,6 +3900,7 @@ class Program_weight_tensor_parameter_342: class Program_weight_tensor_parameter_343: name = "parameter_343" + original_name = "batch_norm2d_4.w_0" shape = [128] dtype = "float32" min_val = float("1.07334") @@ -3568,6 +3912,7 @@ class Program_weight_tensor_parameter_343: class Program_weight_tensor_parameter_344: name = "parameter_344" + original_name = "batch_norm2d_4.w_2" shape = [128] dtype = "float32" min_val = float("0.802514") @@ -3579,6 +3924,7 @@ class Program_weight_tensor_parameter_344: class Program_weight_tensor_parameter_345: name = "parameter_345" + original_name = "batch_norm2d_4.w_1" shape = [128] dtype = "float32" min_val = float("-34.9073") @@ -3590,6 +3936,7 @@ class Program_weight_tensor_parameter_345: class Program_weight_tensor_parameter_346: name = "parameter_346" + original_name = "conv2d_4.w_0" shape = [128, 128, 3, 3] dtype = "float32" min_val = float("-1.50788") @@ -3601,6 +3948,7 @@ class Program_weight_tensor_parameter_346: class Program_weight_tensor_parameter_347: name = "parameter_347" + original_name = "batch_norm2d_3.b_0" shape = [128] dtype = "float32" min_val = float("-7.06689") @@ -3612,6 +3960,7 @@ class Program_weight_tensor_parameter_347: class Program_weight_tensor_parameter_348: name = "parameter_348" + original_name = "batch_norm2d_3.w_0" shape = [128] dtype = "float32" min_val = float("1.17741") @@ -3623,6 +3972,7 @@ class Program_weight_tensor_parameter_348: class Program_weight_tensor_parameter_349: name = "parameter_349" + original_name = "batch_norm2d_3.w_2" shape = [128] dtype = "float32" min_val = float("5.60519e-45") @@ -3634,6 +3984,7 @@ class Program_weight_tensor_parameter_349: class Program_weight_tensor_parameter_350: name = "parameter_350" + original_name = "batch_norm2d_3.w_1" shape = [128] dtype = "float32" min_val = float("-40.1239") @@ -3645,6 +3996,7 @@ class Program_weight_tensor_parameter_350: class Program_weight_tensor_parameter_351: name = "parameter_351" + original_name = "conv2d_3.w_0" shape = [128, 128, 3, 3] dtype = "float32" min_val = float("-1.66284") @@ -3656,6 +4008,7 @@ class Program_weight_tensor_parameter_351: class Program_weight_tensor_parameter_352: name = "parameter_352" + original_name = "batch_norm2d_2.b_0" shape = [128] dtype = "float32" min_val = float("-1.61595") @@ -3667,6 +4020,7 @@ class Program_weight_tensor_parameter_352: class Program_weight_tensor_parameter_353: name = "parameter_353" + original_name = "batch_norm2d_2.w_0" shape = [128] dtype = "float32" min_val = float("0.729716") @@ -3678,6 +4032,7 @@ class Program_weight_tensor_parameter_353: class Program_weight_tensor_parameter_354: name = "parameter_354" + original_name = "batch_norm2d_2.w_2" shape = [128] dtype = "float32" min_val = float("66.9467") @@ -3689,6 +4044,7 @@ class Program_weight_tensor_parameter_354: class Program_weight_tensor_parameter_355: name = "parameter_355" + original_name = "batch_norm2d_2.w_1" shape = [128] dtype = "float32" min_val = float("-76.2154") @@ -3700,6 +4056,7 @@ class Program_weight_tensor_parameter_355: class Program_weight_tensor_parameter_356: name = "parameter_356" + original_name = "conv2d_2.w_0" shape = [128, 64, 3, 3] dtype = "float32" min_val = float("-2.16327") @@ -3711,6 +4068,7 @@ class Program_weight_tensor_parameter_356: class Program_weight_tensor_parameter_357: name = "parameter_357" + original_name = "batch_norm2d_1.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -3720,6 +4078,7 @@ class Program_weight_tensor_parameter_357: class Program_weight_tensor_parameter_358: name = "parameter_358" + original_name = "batch_norm2d_1.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -3729,6 +4088,7 @@ class Program_weight_tensor_parameter_358: class Program_weight_tensor_parameter_359: name = "parameter_359" + original_name = "batch_norm2d_1.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -3738,6 +4098,7 @@ class Program_weight_tensor_parameter_359: class Program_weight_tensor_parameter_360: name = "parameter_360" + original_name = "batch_norm2d_1.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -3747,6 +4108,7 @@ class Program_weight_tensor_parameter_360: class Program_weight_tensor_parameter_361: name = "parameter_361" + original_name = "conv2d_1.w_0" shape = [64, 64, 3, 3] dtype = "float32" min_val = float("-2.69851") @@ -3758,6 +4120,7 @@ class Program_weight_tensor_parameter_361: class Program_weight_tensor_parameter_362: name = "parameter_362" + original_name = "batch_norm2d_0.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -3767,6 +4130,7 @@ class Program_weight_tensor_parameter_362: class Program_weight_tensor_parameter_363: name = "parameter_363" + original_name = "batch_norm2d_0.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -3776,6 +4140,7 @@ class Program_weight_tensor_parameter_363: class Program_weight_tensor_parameter_364: name = "parameter_364" + original_name = "batch_norm2d_0.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -3785,6 +4150,7 @@ class Program_weight_tensor_parameter_364: class Program_weight_tensor_parameter_365: name = "parameter_365" + original_name = "batch_norm2d_0.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -3794,6 +4160,7 @@ class Program_weight_tensor_parameter_365: class Program_weight_tensor_parameter_366: name = "parameter_366" + original_name = "conv2d_0.w_0" shape = [64, 3, 3, 3] dtype = "float32" min_val = float("-0.803985") diff --git a/paddle_samples/PaddleX/PP-OCRv4_server_det/subgraph_1/input_meta.py b/paddle_samples/PaddleX/PP-OCRv4_server_det/subgraph_1/input_meta.py index dd0e33770..47287f82d 100644 --- a/paddle_samples/PaddleX/PP-OCRv4_server_det/subgraph_1/input_meta.py +++ b/paddle_samples/PaddleX/PP-OCRv4_server_det/subgraph_1/input_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_data_0: name = "data_0" + original_name = "var_301" shape = [1, 3, 640, 1024] dtype = "float32" min_val = float("-2.1179") diff --git a/paddle_samples/PaddleX/PP-OCRv4_server_det/subgraph_1/weight_meta.py b/paddle_samples/PaddleX/PP-OCRv4_server_det/subgraph_1/weight_meta.py index c2f03a73e..ab238bbf8 100644 --- a/paddle_samples/PaddleX/PP-OCRv4_server_det/subgraph_1/weight_meta.py +++ b/paddle_samples/PaddleX/PP-OCRv4_server_det/subgraph_1/weight_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_parameter_0: name = "parameter_0" + original_name = "conv2d_108.b_0" shape = [1] dtype = "float32" min_val = float("0") @@ -9,6 +10,7 @@ class Program_weight_tensor_parameter_0: class Program_weight_tensor_parameter_1: name = "parameter_1" + original_name = "conv2d_108.w_0" shape = [1, 64, 1, 1] dtype = "float32" min_val = float("0") @@ -18,6 +20,7 @@ class Program_weight_tensor_parameter_1: class Program_weight_tensor_parameter_2: name = "parameter_2" + original_name = "batch_norm_4.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -27,6 +30,7 @@ class Program_weight_tensor_parameter_2: class Program_weight_tensor_parameter_3: name = "parameter_3" + original_name = "batch_norm_4.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -36,6 +40,7 @@ class Program_weight_tensor_parameter_3: class Program_weight_tensor_parameter_4: name = "parameter_4" + original_name = "batch_norm_4.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -45,6 +50,7 @@ class Program_weight_tensor_parameter_4: class Program_weight_tensor_parameter_5: name = "parameter_5" + original_name = "batch_norm_4.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -54,6 +60,7 @@ class Program_weight_tensor_parameter_5: class Program_weight_tensor_parameter_6: name = "parameter_6" + original_name = "conv2d_107.w_0" shape = [64, 65, 3, 3] dtype = "float32" min_val = float("-1.38082") @@ -65,6 +72,7 @@ class Program_weight_tensor_parameter_6: class Program_weight_tensor_parameter_7: name = "parameter_7" + original_name = "conv2d_transpose_1.b_0" shape = [1] dtype = "float32" min_val = float("0") @@ -74,6 +82,7 @@ class Program_weight_tensor_parameter_7: class Program_weight_tensor_parameter_8: name = "parameter_8" + original_name = "conv2d_transpose_1.w_0" shape = [64, 1, 2, 2] dtype = "float32" min_val = float("-1.8089") @@ -85,6 +94,7 @@ class Program_weight_tensor_parameter_8: class Program_weight_tensor_parameter_9: name = "parameter_9" + original_name = "batch_norm_1.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -94,6 +104,7 @@ class Program_weight_tensor_parameter_9: class Program_weight_tensor_parameter_10: name = "parameter_10" + original_name = "batch_norm_1.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -103,6 +114,7 @@ class Program_weight_tensor_parameter_10: class Program_weight_tensor_parameter_11: name = "parameter_11" + original_name = "batch_norm_1.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -112,6 +124,7 @@ class Program_weight_tensor_parameter_11: class Program_weight_tensor_parameter_12: name = "parameter_12" + original_name = "batch_norm_1.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -121,6 +134,7 @@ class Program_weight_tensor_parameter_12: class Program_weight_tensor_parameter_13: name = "parameter_13" + original_name = "conv2d_transpose_0.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -130,6 +144,7 @@ class Program_weight_tensor_parameter_13: class Program_weight_tensor_parameter_14: name = "parameter_14" + original_name = "conv2d_transpose_0.w_0" shape = [64, 64, 2, 2] dtype = "float32" min_val = float("-1.08046") @@ -141,6 +156,7 @@ class Program_weight_tensor_parameter_14: class Program_weight_tensor_parameter_15: name = "parameter_15" + original_name = "batch_norm_0.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -150,6 +166,7 @@ class Program_weight_tensor_parameter_15: class Program_weight_tensor_parameter_16: name = "parameter_16" + original_name = "batch_norm_0.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -159,6 +176,7 @@ class Program_weight_tensor_parameter_16: class Program_weight_tensor_parameter_17: name = "parameter_17" + original_name = "batch_norm_0.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -168,6 +186,7 @@ class Program_weight_tensor_parameter_17: class Program_weight_tensor_parameter_18: name = "parameter_18" + original_name = "batch_norm_0.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -177,6 +196,7 @@ class Program_weight_tensor_parameter_18: class Program_weight_tensor_parameter_19: name = "parameter_19" + original_name = "conv2d_105.w_0" shape = [64, 256, 3, 3] dtype = "float32" min_val = float("-1.35645") @@ -188,6 +208,7 @@ class Program_weight_tensor_parameter_19: class Program_weight_tensor_parameter_20: name = "parameter_20" + original_name = "batch_norm2d_41.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -197,6 +218,7 @@ class Program_weight_tensor_parameter_20: class Program_weight_tensor_parameter_21: name = "parameter_21" + original_name = "batch_norm2d_41.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -206,6 +228,7 @@ class Program_weight_tensor_parameter_21: class Program_weight_tensor_parameter_22: name = "parameter_22" + original_name = "batch_norm2d_41.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -215,6 +238,7 @@ class Program_weight_tensor_parameter_22: class Program_weight_tensor_parameter_23: name = "parameter_23" + original_name = "batch_norm2d_41.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -224,6 +248,7 @@ class Program_weight_tensor_parameter_23: class Program_weight_tensor_parameter_24: name = "parameter_24" + original_name = "conv2d_62.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -233,6 +258,7 @@ class Program_weight_tensor_parameter_24: class Program_weight_tensor_parameter_25: name = "parameter_25" + original_name = "conv2d_62.w_0" shape = [64, 32, 1, 1] dtype = "float32" min_val = float("-6.30279e-39") @@ -243,6 +269,7 @@ class Program_weight_tensor_parameter_25: class Program_weight_tensor_parameter_26: name = "parameter_26" + original_name = "conv2d_68.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -252,6 +279,7 @@ class Program_weight_tensor_parameter_26: class Program_weight_tensor_parameter_27: name = "parameter_27" + original_name = "conv2d_68.w_0" shape = [32, 32, 1, 3] dtype = "float32" min_val = float("-6.30299e-39") @@ -262,6 +290,7 @@ class Program_weight_tensor_parameter_27: class Program_weight_tensor_parameter_28: name = "parameter_28" + original_name = "conv2d_65.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -271,6 +300,7 @@ class Program_weight_tensor_parameter_28: class Program_weight_tensor_parameter_29: name = "parameter_29" + original_name = "conv2d_65.w_0" shape = [32, 32, 3, 1] dtype = "float32" min_val = float("-6.30543e-39") @@ -281,6 +311,7 @@ class Program_weight_tensor_parameter_29: class Program_weight_tensor_parameter_30: name = "parameter_30" + original_name = "conv2d_71.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -290,6 +321,7 @@ class Program_weight_tensor_parameter_30: class Program_weight_tensor_parameter_31: name = "parameter_31" + original_name = "conv2d_71.w_0" shape = [32, 32, 3, 3] dtype = "float32" min_val = float("-6.30575e-39") @@ -300,6 +332,7 @@ class Program_weight_tensor_parameter_31: class Program_weight_tensor_parameter_32: name = "parameter_32" + original_name = "conv2d_67.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -309,6 +342,7 @@ class Program_weight_tensor_parameter_32: class Program_weight_tensor_parameter_33: name = "parameter_33" + original_name = "conv2d_67.w_0" shape = [32, 32, 1, 5] dtype = "float32" min_val = float("-6.30217e-39") @@ -319,6 +353,7 @@ class Program_weight_tensor_parameter_33: class Program_weight_tensor_parameter_34: name = "parameter_34" + original_name = "conv2d_64.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -328,6 +363,7 @@ class Program_weight_tensor_parameter_34: class Program_weight_tensor_parameter_35: name = "parameter_35" + original_name = "conv2d_64.w_0" shape = [32, 32, 5, 1] dtype = "float32" min_val = float("-6.30561e-39") @@ -338,6 +374,7 @@ class Program_weight_tensor_parameter_35: class Program_weight_tensor_parameter_36: name = "parameter_36" + original_name = "conv2d_70.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -347,6 +384,7 @@ class Program_weight_tensor_parameter_36: class Program_weight_tensor_parameter_37: name = "parameter_37" + original_name = "conv2d_70.w_0" shape = [32, 32, 5, 5] dtype = "float32" min_val = float("-6.30544e-39") @@ -357,6 +395,7 @@ class Program_weight_tensor_parameter_37: class Program_weight_tensor_parameter_38: name = "parameter_38" + original_name = "conv2d_66.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -366,6 +405,7 @@ class Program_weight_tensor_parameter_38: class Program_weight_tensor_parameter_39: name = "parameter_39" + original_name = "conv2d_66.w_0" shape = [32, 32, 1, 7] dtype = "float32" min_val = float("-6.30538e-39") @@ -376,6 +416,7 @@ class Program_weight_tensor_parameter_39: class Program_weight_tensor_parameter_40: name = "parameter_40" + original_name = "conv2d_63.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -385,6 +426,7 @@ class Program_weight_tensor_parameter_40: class Program_weight_tensor_parameter_41: name = "parameter_41" + original_name = "conv2d_63.w_0" shape = [32, 32, 7, 1] dtype = "float32" min_val = float("-6.30581e-39") @@ -395,6 +437,7 @@ class Program_weight_tensor_parameter_41: class Program_weight_tensor_parameter_42: name = "parameter_42" + original_name = "conv2d_69.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -404,6 +447,7 @@ class Program_weight_tensor_parameter_42: class Program_weight_tensor_parameter_43: name = "parameter_43" + original_name = "conv2d_69.w_0" shape = [32, 32, 7, 7] dtype = "float32" min_val = float("-6.30572e-39") @@ -414,6 +458,7 @@ class Program_weight_tensor_parameter_43: class Program_weight_tensor_parameter_44: name = "parameter_44" + original_name = "conv2d_61.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -423,6 +468,7 @@ class Program_weight_tensor_parameter_44: class Program_weight_tensor_parameter_45: name = "parameter_45" + original_name = "conv2d_61.w_0" shape = [32, 64, 1, 1] dtype = "float32" min_val = float("-6.30522e-39") @@ -433,6 +479,7 @@ class Program_weight_tensor_parameter_45: class Program_weight_tensor_parameter_46: name = "parameter_46" + original_name = "batch_norm2d_42.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -442,6 +489,7 @@ class Program_weight_tensor_parameter_46: class Program_weight_tensor_parameter_47: name = "parameter_47" + original_name = "batch_norm2d_42.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -451,6 +499,7 @@ class Program_weight_tensor_parameter_47: class Program_weight_tensor_parameter_48: name = "parameter_48" + original_name = "batch_norm2d_42.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -460,6 +509,7 @@ class Program_weight_tensor_parameter_48: class Program_weight_tensor_parameter_49: name = "parameter_49" + original_name = "batch_norm2d_42.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -469,6 +519,7 @@ class Program_weight_tensor_parameter_49: class Program_weight_tensor_parameter_50: name = "parameter_50" + original_name = "conv2d_73.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -478,6 +529,7 @@ class Program_weight_tensor_parameter_50: class Program_weight_tensor_parameter_51: name = "parameter_51" + original_name = "conv2d_73.w_0" shape = [64, 32, 1, 1] dtype = "float32" min_val = float("-6.30238e-39") @@ -488,6 +540,7 @@ class Program_weight_tensor_parameter_51: class Program_weight_tensor_parameter_52: name = "parameter_52" + original_name = "conv2d_79.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -497,6 +550,7 @@ class Program_weight_tensor_parameter_52: class Program_weight_tensor_parameter_53: name = "parameter_53" + original_name = "conv2d_79.w_0" shape = [32, 32, 1, 3] dtype = "float32" min_val = float("-6.30524e-39") @@ -507,6 +561,7 @@ class Program_weight_tensor_parameter_53: class Program_weight_tensor_parameter_54: name = "parameter_54" + original_name = "conv2d_76.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -516,6 +571,7 @@ class Program_weight_tensor_parameter_54: class Program_weight_tensor_parameter_55: name = "parameter_55" + original_name = "conv2d_76.w_0" shape = [32, 32, 3, 1] dtype = "float32" min_val = float("-6.30276e-39") @@ -526,6 +582,7 @@ class Program_weight_tensor_parameter_55: class Program_weight_tensor_parameter_56: name = "parameter_56" + original_name = "conv2d_82.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -535,6 +592,7 @@ class Program_weight_tensor_parameter_56: class Program_weight_tensor_parameter_57: name = "parameter_57" + original_name = "conv2d_82.w_0" shape = [32, 32, 3, 3] dtype = "float32" min_val = float("-6.30567e-39") @@ -545,6 +603,7 @@ class Program_weight_tensor_parameter_57: class Program_weight_tensor_parameter_58: name = "parameter_58" + original_name = "conv2d_78.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -554,6 +613,7 @@ class Program_weight_tensor_parameter_58: class Program_weight_tensor_parameter_59: name = "parameter_59" + original_name = "conv2d_78.w_0" shape = [32, 32, 1, 5] dtype = "float32" min_val = float("-6.305e-39") @@ -564,6 +624,7 @@ class Program_weight_tensor_parameter_59: class Program_weight_tensor_parameter_60: name = "parameter_60" + original_name = "conv2d_75.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -573,6 +634,7 @@ class Program_weight_tensor_parameter_60: class Program_weight_tensor_parameter_61: name = "parameter_61" + original_name = "conv2d_75.w_0" shape = [32, 32, 5, 1] dtype = "float32" min_val = float("-6.30161e-39") @@ -583,6 +645,7 @@ class Program_weight_tensor_parameter_61: class Program_weight_tensor_parameter_62: name = "parameter_62" + original_name = "conv2d_81.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -592,6 +655,7 @@ class Program_weight_tensor_parameter_62: class Program_weight_tensor_parameter_63: name = "parameter_63" + original_name = "conv2d_81.w_0" shape = [32, 32, 5, 5] dtype = "float32" min_val = float("-6.30475e-39") @@ -602,6 +666,7 @@ class Program_weight_tensor_parameter_63: class Program_weight_tensor_parameter_64: name = "parameter_64" + original_name = "conv2d_77.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -611,6 +676,7 @@ class Program_weight_tensor_parameter_64: class Program_weight_tensor_parameter_65: name = "parameter_65" + original_name = "conv2d_77.w_0" shape = [32, 32, 1, 7] dtype = "float32" min_val = float("-6.30366e-39") @@ -621,6 +687,7 @@ class Program_weight_tensor_parameter_65: class Program_weight_tensor_parameter_66: name = "parameter_66" + original_name = "conv2d_74.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -630,6 +697,7 @@ class Program_weight_tensor_parameter_66: class Program_weight_tensor_parameter_67: name = "parameter_67" + original_name = "conv2d_74.w_0" shape = [32, 32, 7, 1] dtype = "float32" min_val = float("-6.30561e-39") @@ -640,6 +708,7 @@ class Program_weight_tensor_parameter_67: class Program_weight_tensor_parameter_68: name = "parameter_68" + original_name = "conv2d_80.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -649,6 +718,7 @@ class Program_weight_tensor_parameter_68: class Program_weight_tensor_parameter_69: name = "parameter_69" + original_name = "conv2d_80.w_0" shape = [32, 32, 7, 7] dtype = "float32" min_val = float("-6.30584e-39") @@ -659,6 +729,7 @@ class Program_weight_tensor_parameter_69: class Program_weight_tensor_parameter_70: name = "parameter_70" + original_name = "conv2d_72.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -668,6 +739,7 @@ class Program_weight_tensor_parameter_70: class Program_weight_tensor_parameter_71: name = "parameter_71" + original_name = "conv2d_72.w_0" shape = [32, 64, 1, 1] dtype = "float32" min_val = float("-6.30468e-39") @@ -678,6 +750,7 @@ class Program_weight_tensor_parameter_71: class Program_weight_tensor_parameter_72: name = "parameter_72" + original_name = "batch_norm2d_43.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -687,6 +760,7 @@ class Program_weight_tensor_parameter_72: class Program_weight_tensor_parameter_73: name = "parameter_73" + original_name = "batch_norm2d_43.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -696,6 +770,7 @@ class Program_weight_tensor_parameter_73: class Program_weight_tensor_parameter_74: name = "parameter_74" + original_name = "batch_norm2d_43.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -705,6 +780,7 @@ class Program_weight_tensor_parameter_74: class Program_weight_tensor_parameter_75: name = "parameter_75" + original_name = "batch_norm2d_43.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -714,6 +790,7 @@ class Program_weight_tensor_parameter_75: class Program_weight_tensor_parameter_76: name = "parameter_76" + original_name = "conv2d_84.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -723,6 +800,7 @@ class Program_weight_tensor_parameter_76: class Program_weight_tensor_parameter_77: name = "parameter_77" + original_name = "conv2d_84.w_0" shape = [64, 32, 1, 1] dtype = "float32" min_val = float("-6.30322e-39") @@ -733,6 +811,7 @@ class Program_weight_tensor_parameter_77: class Program_weight_tensor_parameter_78: name = "parameter_78" + original_name = "conv2d_90.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -742,6 +821,7 @@ class Program_weight_tensor_parameter_78: class Program_weight_tensor_parameter_79: name = "parameter_79" + original_name = "conv2d_90.w_0" shape = [32, 32, 1, 3] dtype = "float32" min_val = float("-6.30507e-39") @@ -752,6 +832,7 @@ class Program_weight_tensor_parameter_79: class Program_weight_tensor_parameter_80: name = "parameter_80" + original_name = "conv2d_87.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -761,6 +842,7 @@ class Program_weight_tensor_parameter_80: class Program_weight_tensor_parameter_81: name = "parameter_81" + original_name = "conv2d_87.w_0" shape = [32, 32, 3, 1] dtype = "float32" min_val = float("-6.30562e-39") @@ -771,6 +853,7 @@ class Program_weight_tensor_parameter_81: class Program_weight_tensor_parameter_82: name = "parameter_82" + original_name = "conv2d_93.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -780,6 +863,7 @@ class Program_weight_tensor_parameter_82: class Program_weight_tensor_parameter_83: name = "parameter_83" + original_name = "conv2d_93.w_0" shape = [32, 32, 3, 3] dtype = "float32" min_val = float("-6.30014e-39") @@ -790,6 +874,7 @@ class Program_weight_tensor_parameter_83: class Program_weight_tensor_parameter_84: name = "parameter_84" + original_name = "conv2d_89.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -799,6 +884,7 @@ class Program_weight_tensor_parameter_84: class Program_weight_tensor_parameter_85: name = "parameter_85" + original_name = "conv2d_89.w_0" shape = [32, 32, 1, 5] dtype = "float32" min_val = float("-6.29705e-39") @@ -809,6 +895,7 @@ class Program_weight_tensor_parameter_85: class Program_weight_tensor_parameter_86: name = "parameter_86" + original_name = "conv2d_86.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -818,6 +905,7 @@ class Program_weight_tensor_parameter_86: class Program_weight_tensor_parameter_87: name = "parameter_87" + original_name = "conv2d_86.w_0" shape = [32, 32, 5, 1] dtype = "float32" min_val = float("-6.29741e-39") @@ -828,6 +916,7 @@ class Program_weight_tensor_parameter_87: class Program_weight_tensor_parameter_88: name = "parameter_88" + original_name = "conv2d_92.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -837,6 +926,7 @@ class Program_weight_tensor_parameter_88: class Program_weight_tensor_parameter_89: name = "parameter_89" + original_name = "conv2d_92.w_0" shape = [32, 32, 5, 5] dtype = "float32" min_val = float("-6.30438e-39") @@ -847,6 +937,7 @@ class Program_weight_tensor_parameter_89: class Program_weight_tensor_parameter_90: name = "parameter_90" + original_name = "conv2d_88.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -856,6 +947,7 @@ class Program_weight_tensor_parameter_90: class Program_weight_tensor_parameter_91: name = "parameter_91" + original_name = "conv2d_88.w_0" shape = [32, 32, 1, 7] dtype = "float32" min_val = float("-6.30576e-39") @@ -866,6 +958,7 @@ class Program_weight_tensor_parameter_91: class Program_weight_tensor_parameter_92: name = "parameter_92" + original_name = "conv2d_85.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -875,6 +968,7 @@ class Program_weight_tensor_parameter_92: class Program_weight_tensor_parameter_93: name = "parameter_93" + original_name = "conv2d_85.w_0" shape = [32, 32, 7, 1] dtype = "float32" min_val = float("-6.30319e-39") @@ -885,6 +979,7 @@ class Program_weight_tensor_parameter_93: class Program_weight_tensor_parameter_94: name = "parameter_94" + original_name = "conv2d_91.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -894,6 +989,7 @@ class Program_weight_tensor_parameter_94: class Program_weight_tensor_parameter_95: name = "parameter_95" + original_name = "conv2d_91.w_0" shape = [32, 32, 7, 7] dtype = "float32" min_val = float("-6.30579e-39") @@ -904,6 +1000,7 @@ class Program_weight_tensor_parameter_95: class Program_weight_tensor_parameter_96: name = "parameter_96" + original_name = "conv2d_83.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -913,6 +1010,7 @@ class Program_weight_tensor_parameter_96: class Program_weight_tensor_parameter_97: name = "parameter_97" + original_name = "conv2d_83.w_0" shape = [32, 64, 1, 1] dtype = "float32" min_val = float("-6.29611e-39") @@ -923,6 +1021,7 @@ class Program_weight_tensor_parameter_97: class Program_weight_tensor_parameter_98: name = "parameter_98" + original_name = "batch_norm2d_44.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -932,6 +1031,7 @@ class Program_weight_tensor_parameter_98: class Program_weight_tensor_parameter_99: name = "parameter_99" + original_name = "batch_norm2d_44.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -941,6 +1041,7 @@ class Program_weight_tensor_parameter_99: class Program_weight_tensor_parameter_100: name = "parameter_100" + original_name = "batch_norm2d_44.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -950,6 +1051,7 @@ class Program_weight_tensor_parameter_100: class Program_weight_tensor_parameter_101: name = "parameter_101" + original_name = "batch_norm2d_44.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -959,6 +1061,7 @@ class Program_weight_tensor_parameter_101: class Program_weight_tensor_parameter_102: name = "parameter_102" + original_name = "conv2d_95.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -968,6 +1071,7 @@ class Program_weight_tensor_parameter_102: class Program_weight_tensor_parameter_103: name = "parameter_103" + original_name = "conv2d_95.w_0" shape = [64, 32, 1, 1] dtype = "float32" min_val = float("-6.30495e-39") @@ -978,6 +1082,7 @@ class Program_weight_tensor_parameter_103: class Program_weight_tensor_parameter_104: name = "parameter_104" + original_name = "conv2d_101.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -987,6 +1092,7 @@ class Program_weight_tensor_parameter_104: class Program_weight_tensor_parameter_105: name = "parameter_105" + original_name = "conv2d_101.w_0" shape = [32, 32, 1, 3] dtype = "float32" min_val = float("-6.30416e-39") @@ -997,6 +1103,7 @@ class Program_weight_tensor_parameter_105: class Program_weight_tensor_parameter_106: name = "parameter_106" + original_name = "conv2d_98.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -1006,6 +1113,7 @@ class Program_weight_tensor_parameter_106: class Program_weight_tensor_parameter_107: name = "parameter_107" + original_name = "conv2d_98.w_0" shape = [32, 32, 3, 1] dtype = "float32" min_val = float("-6.29944e-39") @@ -1016,6 +1124,7 @@ class Program_weight_tensor_parameter_107: class Program_weight_tensor_parameter_108: name = "parameter_108" + original_name = "conv2d_104.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -1025,6 +1134,7 @@ class Program_weight_tensor_parameter_108: class Program_weight_tensor_parameter_109: name = "parameter_109" + original_name = "conv2d_104.w_0" shape = [32, 32, 3, 3] dtype = "float32" min_val = float("-6.30584e-39") @@ -1035,6 +1145,7 @@ class Program_weight_tensor_parameter_109: class Program_weight_tensor_parameter_110: name = "parameter_110" + original_name = "conv2d_100.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -1044,6 +1155,7 @@ class Program_weight_tensor_parameter_110: class Program_weight_tensor_parameter_111: name = "parameter_111" + original_name = "conv2d_100.w_0" shape = [32, 32, 1, 5] dtype = "float32" min_val = float("-6.30117e-39") @@ -1054,6 +1166,7 @@ class Program_weight_tensor_parameter_111: class Program_weight_tensor_parameter_112: name = "parameter_112" + original_name = "conv2d_97.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -1063,6 +1176,7 @@ class Program_weight_tensor_parameter_112: class Program_weight_tensor_parameter_113: name = "parameter_113" + original_name = "conv2d_97.w_0" shape = [32, 32, 5, 1] dtype = "float32" min_val = float("-6.30388e-39") @@ -1073,6 +1187,7 @@ class Program_weight_tensor_parameter_113: class Program_weight_tensor_parameter_114: name = "parameter_114" + original_name = "conv2d_103.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -1082,6 +1197,7 @@ class Program_weight_tensor_parameter_114: class Program_weight_tensor_parameter_115: name = "parameter_115" + original_name = "conv2d_103.w_0" shape = [32, 32, 5, 5] dtype = "float32" min_val = float("-6.30414e-39") @@ -1092,6 +1208,7 @@ class Program_weight_tensor_parameter_115: class Program_weight_tensor_parameter_116: name = "parameter_116" + original_name = "conv2d_99.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -1101,6 +1218,7 @@ class Program_weight_tensor_parameter_116: class Program_weight_tensor_parameter_117: name = "parameter_117" + original_name = "conv2d_99.w_0" shape = [32, 32, 1, 7] dtype = "float32" min_val = float("-6.30514e-39") @@ -1111,6 +1229,7 @@ class Program_weight_tensor_parameter_117: class Program_weight_tensor_parameter_118: name = "parameter_118" + original_name = "conv2d_96.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -1120,6 +1239,7 @@ class Program_weight_tensor_parameter_118: class Program_weight_tensor_parameter_119: name = "parameter_119" + original_name = "conv2d_96.w_0" shape = [32, 32, 7, 1] dtype = "float32" min_val = float("-6.30558e-39") @@ -1130,6 +1250,7 @@ class Program_weight_tensor_parameter_119: class Program_weight_tensor_parameter_120: name = "parameter_120" + original_name = "conv2d_102.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -1139,6 +1260,7 @@ class Program_weight_tensor_parameter_120: class Program_weight_tensor_parameter_121: name = "parameter_121" + original_name = "conv2d_102.w_0" shape = [32, 32, 7, 7] dtype = "float32" min_val = float("-6.30555e-39") @@ -1149,6 +1271,7 @@ class Program_weight_tensor_parameter_121: class Program_weight_tensor_parameter_122: name = "parameter_122" + original_name = "conv2d_94.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -1158,6 +1281,7 @@ class Program_weight_tensor_parameter_122: class Program_weight_tensor_parameter_123: name = "parameter_123" + original_name = "conv2d_94.w_0" shape = [32, 64, 1, 1] dtype = "float32" min_val = float("-6.30261e-39") @@ -1168,6 +1292,7 @@ class Program_weight_tensor_parameter_123: class Program_weight_tensor_parameter_124: name = "parameter_124" + original_name = "conv2d_60.w_0" shape = [64, 64, 9, 9] dtype = "float32" min_val = float("-1.13071") @@ -1179,6 +1304,7 @@ class Program_weight_tensor_parameter_124: class Program_weight_tensor_parameter_125: name = "parameter_125" + original_name = "conv2d_56.w_0" shape = [64, 64, 9, 9] dtype = "float32" min_val = float("-1.27211") @@ -1190,6 +1316,7 @@ class Program_weight_tensor_parameter_125: class Program_weight_tensor_parameter_126: name = "parameter_126" + original_name = "conv2d_52.w_0" shape = [64, 64, 9, 9] dtype = "float32" min_val = float("-0.796373") @@ -1201,6 +1328,7 @@ class Program_weight_tensor_parameter_126: class Program_weight_tensor_parameter_127: name = "parameter_127" + original_name = "conv2d_48.w_0" shape = [64, 64, 9, 9] dtype = "float32" min_val = float("-1.24486") @@ -1212,6 +1340,7 @@ class Program_weight_tensor_parameter_127: class Program_weight_tensor_parameter_128: name = "parameter_128" + original_name = "conv2d_59.w_0" shape = [64, 64, 3, 3] dtype = "float32" min_val = float("-0.526123") @@ -1223,6 +1352,7 @@ class Program_weight_tensor_parameter_128: class Program_weight_tensor_parameter_129: name = "parameter_129" + original_name = "conv2d_55.w_0" shape = [64, 64, 3, 3] dtype = "float32" min_val = float("-0.361634") @@ -1234,6 +1364,7 @@ class Program_weight_tensor_parameter_129: class Program_weight_tensor_parameter_130: name = "parameter_130" + original_name = "conv2d_51.w_0" shape = [64, 64, 3, 3] dtype = "float32" min_val = float("-0.873763") @@ -1245,6 +1376,7 @@ class Program_weight_tensor_parameter_130: class Program_weight_tensor_parameter_131: name = "parameter_131" + original_name = "conv2d_47.w_0" shape = [64, 256, 9, 9] dtype = "float32" min_val = float("-0.91112") @@ -1256,6 +1388,7 @@ class Program_weight_tensor_parameter_131: class Program_weight_tensor_parameter_132: name = "parameter_132" + original_name = "conv2d_50.w_0" shape = [64, 256, 9, 9] dtype = "float32" min_val = float("-0.195519") @@ -1267,6 +1400,7 @@ class Program_weight_tensor_parameter_132: class Program_weight_tensor_parameter_133: name = "parameter_133" + original_name = "conv2d_54.w_0" shape = [64, 256, 9, 9] dtype = "float32" min_val = float("-0.440963") @@ -1278,6 +1412,7 @@ class Program_weight_tensor_parameter_133: class Program_weight_tensor_parameter_134: name = "parameter_134" + original_name = "conv2d_58.w_0" shape = [64, 256, 9, 9] dtype = "float32" min_val = float("-0.0913491") @@ -1289,6 +1424,7 @@ class Program_weight_tensor_parameter_134: class Program_weight_tensor_parameter_135: name = "parameter_135" + original_name = "conv2d_46.w_0" shape = [256, 256, 1, 1] dtype = "float32" min_val = float("-2.74007") @@ -1300,6 +1436,7 @@ class Program_weight_tensor_parameter_135: class Program_weight_tensor_parameter_136: name = "parameter_136" + original_name = "conv2d_49.w_0" shape = [256, 512, 1, 1] dtype = "float32" min_val = float("-1.05984") @@ -1311,6 +1448,7 @@ class Program_weight_tensor_parameter_136: class Program_weight_tensor_parameter_137: name = "parameter_137" + original_name = "conv2d_53.w_0" shape = [256, 768, 1, 1] dtype = "float32" min_val = float("-1.1037") @@ -1322,6 +1460,7 @@ class Program_weight_tensor_parameter_137: class Program_weight_tensor_parameter_138: name = "parameter_138" + original_name = "conv2d_57.w_0" shape = [256, 1024, 1, 1] dtype = "float32" min_val = float("-0.778606") @@ -1333,6 +1472,7 @@ class Program_weight_tensor_parameter_138: class Program_weight_tensor_parameter_139: name = "parameter_139" + original_name = "conv2d_45.b_0" shape = [1024] dtype = "float32" min_val = float("-0.212113") @@ -1344,6 +1484,7 @@ class Program_weight_tensor_parameter_139: class Program_weight_tensor_parameter_140: name = "parameter_140" + original_name = "conv2d_45.w_0" shape = [1024, 1024, 1, 1] dtype = "float32" min_val = float("-1.85646") @@ -1355,6 +1496,7 @@ class Program_weight_tensor_parameter_140: class Program_weight_tensor_parameter_141: name = "parameter_141" + original_name = "batch_norm2d_40.b_0" shape = [1024] dtype = "float32" min_val = float("-16.759") @@ -1366,6 +1508,7 @@ class Program_weight_tensor_parameter_141: class Program_weight_tensor_parameter_142: name = "parameter_142" + original_name = "batch_norm2d_40.w_0" shape = [1024] dtype = "float32" min_val = float("2.45943") @@ -1377,6 +1520,7 @@ class Program_weight_tensor_parameter_142: class Program_weight_tensor_parameter_143: name = "parameter_143" + original_name = "batch_norm2d_40.w_2" shape = [1024] dtype = "float32" min_val = float("5.60519e-45") @@ -1388,6 +1532,7 @@ class Program_weight_tensor_parameter_143: class Program_weight_tensor_parameter_144: name = "parameter_144" + original_name = "batch_norm2d_40.w_1" shape = [1024] dtype = "float32" min_val = float("-10.9073") @@ -1399,6 +1544,7 @@ class Program_weight_tensor_parameter_144: class Program_weight_tensor_parameter_145: name = "parameter_145" + original_name = "conv2d_44.w_0" shape = [1024, 2112, 1, 1] dtype = "float32" min_val = float("-1.22029") @@ -1410,6 +1556,7 @@ class Program_weight_tensor_parameter_145: class Program_weight_tensor_parameter_146: name = "parameter_146" + original_name = "batch_norm2d_39.b_0" shape = [224] dtype = "float32" min_val = float("-5.91798") @@ -1421,6 +1568,7 @@ class Program_weight_tensor_parameter_146: class Program_weight_tensor_parameter_147: name = "parameter_147" + original_name = "batch_norm2d_39.w_0" shape = [224] dtype = "float32" min_val = float("-0.0011185") @@ -1432,6 +1580,7 @@ class Program_weight_tensor_parameter_147: class Program_weight_tensor_parameter_148: name = "parameter_148" + original_name = "batch_norm2d_39.w_2" shape = [224] dtype = "float32" min_val = float("5.60519e-45") @@ -1443,6 +1592,7 @@ class Program_weight_tensor_parameter_148: class Program_weight_tensor_parameter_149: name = "parameter_149" + original_name = "batch_norm2d_39.w_1" shape = [224] dtype = "float32" min_val = float("-0.766669") @@ -1454,6 +1604,7 @@ class Program_weight_tensor_parameter_149: class Program_weight_tensor_parameter_150: name = "parameter_150" + original_name = "conv2d_43.w_0" shape = [224, 224, 3, 3] dtype = "float32" min_val = float("-0.538801") @@ -1465,6 +1616,7 @@ class Program_weight_tensor_parameter_150: class Program_weight_tensor_parameter_151: name = "parameter_151" + original_name = "batch_norm2d_38.b_0" shape = [224] dtype = "float32" min_val = float("-5.11582") @@ -1476,6 +1628,7 @@ class Program_weight_tensor_parameter_151: class Program_weight_tensor_parameter_152: name = "parameter_152" + original_name = "batch_norm2d_38.w_0" shape = [224] dtype = "float32" min_val = float("0.000165796") @@ -1487,6 +1640,7 @@ class Program_weight_tensor_parameter_152: class Program_weight_tensor_parameter_153: name = "parameter_153" + original_name = "batch_norm2d_38.w_2" shape = [224] dtype = "float32" min_val = float("5.60519e-45") @@ -1498,6 +1652,7 @@ class Program_weight_tensor_parameter_153: class Program_weight_tensor_parameter_154: name = "parameter_154" + original_name = "batch_norm2d_38.w_1" shape = [224] dtype = "float32" min_val = float("-2.15633") @@ -1509,6 +1664,7 @@ class Program_weight_tensor_parameter_154: class Program_weight_tensor_parameter_155: name = "parameter_155" + original_name = "conv2d_42.w_0" shape = [224, 224, 3, 3] dtype = "float32" min_val = float("-0.469498") @@ -1520,6 +1676,7 @@ class Program_weight_tensor_parameter_155: class Program_weight_tensor_parameter_156: name = "parameter_156" + original_name = "batch_norm2d_37.b_0" shape = [224] dtype = "float32" min_val = float("-5.48079") @@ -1531,6 +1688,7 @@ class Program_weight_tensor_parameter_156: class Program_weight_tensor_parameter_157: name = "parameter_157" + original_name = "batch_norm2d_37.w_0" shape = [224] dtype = "float32" min_val = float("0.00178436") @@ -1542,6 +1700,7 @@ class Program_weight_tensor_parameter_157: class Program_weight_tensor_parameter_158: name = "parameter_158" + original_name = "batch_norm2d_37.w_2" shape = [224] dtype = "float32" min_val = float("5.60519e-45") @@ -1553,6 +1712,7 @@ class Program_weight_tensor_parameter_158: class Program_weight_tensor_parameter_159: name = "parameter_159" + original_name = "batch_norm2d_37.w_1" shape = [224] dtype = "float32" min_val = float("-4.80434") @@ -1564,6 +1724,7 @@ class Program_weight_tensor_parameter_159: class Program_weight_tensor_parameter_160: name = "parameter_160" + original_name = "conv2d_41.w_0" shape = [224, 224, 3, 3] dtype = "float32" min_val = float("-0.473288") @@ -1575,6 +1736,7 @@ class Program_weight_tensor_parameter_160: class Program_weight_tensor_parameter_161: name = "parameter_161" + original_name = "batch_norm2d_36.b_0" shape = [224] dtype = "float32" min_val = float("-5.70744") @@ -1586,6 +1748,7 @@ class Program_weight_tensor_parameter_161: class Program_weight_tensor_parameter_162: name = "parameter_162" + original_name = "batch_norm2d_36.w_0" shape = [224] dtype = "float32" min_val = float("0.00223806") @@ -1597,6 +1760,7 @@ class Program_weight_tensor_parameter_162: class Program_weight_tensor_parameter_163: name = "parameter_163" + original_name = "batch_norm2d_36.w_2" shape = [224] dtype = "float32" min_val = float("5.60519e-45") @@ -1608,6 +1772,7 @@ class Program_weight_tensor_parameter_163: class Program_weight_tensor_parameter_164: name = "parameter_164" + original_name = "batch_norm2d_36.w_1" shape = [224] dtype = "float32" min_val = float("-7.42173") @@ -1619,6 +1784,7 @@ class Program_weight_tensor_parameter_164: class Program_weight_tensor_parameter_165: name = "parameter_165" + original_name = "conv2d_40.w_0" shape = [224, 224, 3, 3] dtype = "float32" min_val = float("-0.601781") @@ -1630,6 +1796,7 @@ class Program_weight_tensor_parameter_165: class Program_weight_tensor_parameter_166: name = "parameter_166" + original_name = "batch_norm2d_35.b_0" shape = [224] dtype = "float32" min_val = float("-6.21687") @@ -1641,6 +1808,7 @@ class Program_weight_tensor_parameter_166: class Program_weight_tensor_parameter_167: name = "parameter_167" + original_name = "batch_norm2d_35.w_0" shape = [224] dtype = "float32" min_val = float("-0.000430432") @@ -1652,6 +1820,7 @@ class Program_weight_tensor_parameter_167: class Program_weight_tensor_parameter_168: name = "parameter_168" + original_name = "batch_norm2d_35.w_2" shape = [224] dtype = "float32" min_val = float("5.60519e-45") @@ -1663,6 +1832,7 @@ class Program_weight_tensor_parameter_168: class Program_weight_tensor_parameter_169: name = "parameter_169" + original_name = "batch_norm2d_35.w_1" shape = [224] dtype = "float32" min_val = float("-16.155") @@ -1674,6 +1844,7 @@ class Program_weight_tensor_parameter_169: class Program_weight_tensor_parameter_170: name = "parameter_170" + original_name = "conv2d_39.w_0" shape = [224, 224, 3, 3] dtype = "float32" min_val = float("-0.88932") @@ -1685,6 +1856,7 @@ class Program_weight_tensor_parameter_170: class Program_weight_tensor_parameter_171: name = "parameter_171" + original_name = "batch_norm2d_34.b_0" shape = [224] dtype = "float32" min_val = float("-10.5971") @@ -1696,6 +1868,7 @@ class Program_weight_tensor_parameter_171: class Program_weight_tensor_parameter_172: name = "parameter_172" + original_name = "batch_norm2d_34.w_0" shape = [224] dtype = "float32" min_val = float("-0.000383113") @@ -1707,6 +1880,7 @@ class Program_weight_tensor_parameter_172: class Program_weight_tensor_parameter_173: name = "parameter_173" + original_name = "batch_norm2d_34.w_2" shape = [224] dtype = "float32" min_val = float("5.60519e-45") @@ -1718,6 +1892,7 @@ class Program_weight_tensor_parameter_173: class Program_weight_tensor_parameter_174: name = "parameter_174" + original_name = "batch_norm2d_34.w_1" shape = [224] dtype = "float32" min_val = float("-38.4428") @@ -1729,6 +1904,7 @@ class Program_weight_tensor_parameter_174: class Program_weight_tensor_parameter_175: name = "parameter_175" + original_name = "conv2d_38.w_0" shape = [224, 768, 3, 3] dtype = "float32" min_val = float("-1.17794") @@ -1740,6 +1916,7 @@ class Program_weight_tensor_parameter_175: class Program_weight_tensor_parameter_176: name = "parameter_176" + original_name = "batch_norm2d_33.b_0" shape = [768] dtype = "float32" min_val = float("-2.56353") @@ -1751,6 +1928,7 @@ class Program_weight_tensor_parameter_176: class Program_weight_tensor_parameter_177: name = "parameter_177" + original_name = "batch_norm2d_33.w_0" shape = [768] dtype = "float32" min_val = float("-0.0176625") @@ -1762,6 +1940,7 @@ class Program_weight_tensor_parameter_177: class Program_weight_tensor_parameter_178: name = "parameter_178" + original_name = "batch_norm2d_33.w_2" shape = [768] dtype = "float32" min_val = float("5.60519e-45") @@ -1773,6 +1952,7 @@ class Program_weight_tensor_parameter_178: class Program_weight_tensor_parameter_179: name = "parameter_179" + original_name = "batch_norm2d_33.w_1" shape = [768] dtype = "float32" min_val = float("-0.914731") @@ -1784,6 +1964,7 @@ class Program_weight_tensor_parameter_179: class Program_weight_tensor_parameter_180: name = "parameter_180" + original_name = "conv2d_37.w_0" shape = [768, 1, 3, 3] dtype = "float32" min_val = float("-0.585581") @@ -1795,6 +1976,7 @@ class Program_weight_tensor_parameter_180: class Program_weight_tensor_parameter_181: name = "parameter_181" + original_name = "conv2d_36.b_0" shape = [768] dtype = "float32" min_val = float("-0.510395") @@ -1806,6 +1988,7 @@ class Program_weight_tensor_parameter_181: class Program_weight_tensor_parameter_182: name = "parameter_182" + original_name = "conv2d_36.w_0" shape = [768, 768, 1, 1] dtype = "float32" min_val = float("-1.7736") @@ -1817,6 +2000,7 @@ class Program_weight_tensor_parameter_182: class Program_weight_tensor_parameter_183: name = "parameter_183" + original_name = "batch_norm2d_32.b_0" shape = [768] dtype = "float32" min_val = float("-12.9396") @@ -1828,6 +2012,7 @@ class Program_weight_tensor_parameter_183: class Program_weight_tensor_parameter_184: name = "parameter_184" + original_name = "batch_norm2d_32.w_0" shape = [768] dtype = "float32" min_val = float("0.0351857") @@ -1839,6 +2024,7 @@ class Program_weight_tensor_parameter_184: class Program_weight_tensor_parameter_185: name = "parameter_185" + original_name = "batch_norm2d_32.w_2" shape = [768] dtype = "float32" min_val = float("5.60519e-45") @@ -1850,6 +2036,7 @@ class Program_weight_tensor_parameter_185: class Program_weight_tensor_parameter_186: name = "parameter_186" + original_name = "batch_norm2d_32.w_1" shape = [768] dtype = "float32" min_val = float("-5.87422") @@ -1861,6 +2048,7 @@ class Program_weight_tensor_parameter_186: class Program_weight_tensor_parameter_187: name = "parameter_187" + original_name = "conv2d_35.w_0" shape = [768, 1920, 1, 1] dtype = "float32" min_val = float("-1.48946") @@ -1872,6 +2060,7 @@ class Program_weight_tensor_parameter_187: class Program_weight_tensor_parameter_188: name = "parameter_188" + original_name = "batch_norm2d_31.b_0" shape = [192] dtype = "float32" min_val = float("-3.24736") @@ -1883,6 +2072,7 @@ class Program_weight_tensor_parameter_188: class Program_weight_tensor_parameter_189: name = "parameter_189" + original_name = "batch_norm2d_31.w_0" shape = [192] dtype = "float32" min_val = float("0.0236468") @@ -1894,6 +2084,7 @@ class Program_weight_tensor_parameter_189: class Program_weight_tensor_parameter_190: name = "parameter_190" + original_name = "batch_norm2d_31.w_2" shape = [192] dtype = "float32" min_val = float("5.60519e-45") @@ -1905,6 +2096,7 @@ class Program_weight_tensor_parameter_190: class Program_weight_tensor_parameter_191: name = "parameter_191" + original_name = "batch_norm2d_31.w_1" shape = [192] dtype = "float32" min_val = float("-2.40757") @@ -1916,6 +2108,7 @@ class Program_weight_tensor_parameter_191: class Program_weight_tensor_parameter_192: name = "parameter_192" + original_name = "conv2d_34.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.801684") @@ -1927,6 +2120,7 @@ class Program_weight_tensor_parameter_192: class Program_weight_tensor_parameter_193: name = "parameter_193" + original_name = "batch_norm2d_30.b_0" shape = [192] dtype = "float32" min_val = float("-2.64017") @@ -1938,6 +2132,7 @@ class Program_weight_tensor_parameter_193: class Program_weight_tensor_parameter_194: name = "parameter_194" + original_name = "batch_norm2d_30.w_0" shape = [192] dtype = "float32" min_val = float("0.000159657") @@ -1949,6 +2144,7 @@ class Program_weight_tensor_parameter_194: class Program_weight_tensor_parameter_195: name = "parameter_195" + original_name = "batch_norm2d_30.w_2" shape = [192] dtype = "float32" min_val = float("5.60519e-45") @@ -1960,6 +2156,7 @@ class Program_weight_tensor_parameter_195: class Program_weight_tensor_parameter_196: name = "parameter_196" + original_name = "batch_norm2d_30.w_1" shape = [192] dtype = "float32" min_val = float("-4.54397") @@ -1971,6 +2168,7 @@ class Program_weight_tensor_parameter_196: class Program_weight_tensor_parameter_197: name = "parameter_197" + original_name = "conv2d_33.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.865569") @@ -1982,6 +2180,7 @@ class Program_weight_tensor_parameter_197: class Program_weight_tensor_parameter_198: name = "parameter_198" + original_name = "batch_norm2d_29.b_0" shape = [192] dtype = "float32" min_val = float("-3.9513") @@ -1993,6 +2192,7 @@ class Program_weight_tensor_parameter_198: class Program_weight_tensor_parameter_199: name = "parameter_199" + original_name = "batch_norm2d_29.w_0" shape = [192] dtype = "float32" min_val = float("0.0279247") @@ -2004,6 +2204,7 @@ class Program_weight_tensor_parameter_199: class Program_weight_tensor_parameter_200: name = "parameter_200" + original_name = "batch_norm2d_29.w_2" shape = [192] dtype = "float32" min_val = float("5.60519e-45") @@ -2015,6 +2216,7 @@ class Program_weight_tensor_parameter_200: class Program_weight_tensor_parameter_201: name = "parameter_201" + original_name = "batch_norm2d_29.w_1" shape = [192] dtype = "float32" min_val = float("-4.41101") @@ -2026,6 +2228,7 @@ class Program_weight_tensor_parameter_201: class Program_weight_tensor_parameter_202: name = "parameter_202" + original_name = "conv2d_32.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.922849") @@ -2037,6 +2240,7 @@ class Program_weight_tensor_parameter_202: class Program_weight_tensor_parameter_203: name = "parameter_203" + original_name = "batch_norm2d_28.b_0" shape = [192] dtype = "float32" min_val = float("-3.97438") @@ -2048,6 +2252,7 @@ class Program_weight_tensor_parameter_203: class Program_weight_tensor_parameter_204: name = "parameter_204" + original_name = "batch_norm2d_28.w_0" shape = [192] dtype = "float32" min_val = float("-0.00527014") @@ -2059,6 +2264,7 @@ class Program_weight_tensor_parameter_204: class Program_weight_tensor_parameter_205: name = "parameter_205" + original_name = "batch_norm2d_28.w_2" shape = [192] dtype = "float32" min_val = float("5.60519e-45") @@ -2070,6 +2276,7 @@ class Program_weight_tensor_parameter_205: class Program_weight_tensor_parameter_206: name = "parameter_206" + original_name = "batch_norm2d_28.w_1" shape = [192] dtype = "float32" min_val = float("-10.0256") @@ -2081,6 +2288,7 @@ class Program_weight_tensor_parameter_206: class Program_weight_tensor_parameter_207: name = "parameter_207" + original_name = "conv2d_31.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.864141") @@ -2092,6 +2300,7 @@ class Program_weight_tensor_parameter_207: class Program_weight_tensor_parameter_208: name = "parameter_208" + original_name = "batch_norm2d_27.b_0" shape = [192] dtype = "float32" min_val = float("-3.9103") @@ -2103,6 +2312,7 @@ class Program_weight_tensor_parameter_208: class Program_weight_tensor_parameter_209: name = "parameter_209" + original_name = "batch_norm2d_27.w_0" shape = [192] dtype = "float32" min_val = float("0.0167882") @@ -2114,6 +2324,7 @@ class Program_weight_tensor_parameter_209: class Program_weight_tensor_parameter_210: name = "parameter_210" + original_name = "batch_norm2d_27.w_2" shape = [192] dtype = "float32" min_val = float("5.60519e-45") @@ -2125,6 +2336,7 @@ class Program_weight_tensor_parameter_210: class Program_weight_tensor_parameter_211: name = "parameter_211" + original_name = "batch_norm2d_27.w_1" shape = [192] dtype = "float32" min_val = float("-9.96308") @@ -2136,6 +2348,7 @@ class Program_weight_tensor_parameter_211: class Program_weight_tensor_parameter_212: name = "parameter_212" + original_name = "conv2d_30.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.932015") @@ -2147,6 +2360,7 @@ class Program_weight_tensor_parameter_212: class Program_weight_tensor_parameter_213: name = "parameter_213" + original_name = "batch_norm2d_26.b_0" shape = [192] dtype = "float32" min_val = float("-5.89813") @@ -2158,6 +2372,7 @@ class Program_weight_tensor_parameter_213: class Program_weight_tensor_parameter_214: name = "parameter_214" + original_name = "batch_norm2d_26.w_0" shape = [192] dtype = "float32" min_val = float("0.00167011") @@ -2169,6 +2384,7 @@ class Program_weight_tensor_parameter_214: class Program_weight_tensor_parameter_215: name = "parameter_215" + original_name = "batch_norm2d_26.w_2" shape = [192] dtype = "float32" min_val = float("5.60519e-45") @@ -2180,6 +2396,7 @@ class Program_weight_tensor_parameter_215: class Program_weight_tensor_parameter_216: name = "parameter_216" + original_name = "batch_norm2d_26.w_1" shape = [192] dtype = "float32" min_val = float("-15.9911") @@ -2191,6 +2408,7 @@ class Program_weight_tensor_parameter_216: class Program_weight_tensor_parameter_217: name = "parameter_217" + original_name = "conv2d_29.w_0" shape = [192, 768, 3, 3] dtype = "float32" min_val = float("-0.88755") @@ -2202,6 +2420,7 @@ class Program_weight_tensor_parameter_217: class Program_weight_tensor_parameter_218: name = "parameter_218" + original_name = "conv2d_28.b_0" shape = [768] dtype = "float32" min_val = float("-0.285088") @@ -2213,6 +2432,7 @@ class Program_weight_tensor_parameter_218: class Program_weight_tensor_parameter_219: name = "parameter_219" + original_name = "conv2d_28.w_0" shape = [768, 768, 1, 1] dtype = "float32" min_val = float("-2.49224") @@ -2224,6 +2444,7 @@ class Program_weight_tensor_parameter_219: class Program_weight_tensor_parameter_220: name = "parameter_220" + original_name = "batch_norm2d_25.b_0" shape = [768] dtype = "float32" min_val = float("-13.6166") @@ -2235,6 +2456,7 @@ class Program_weight_tensor_parameter_220: class Program_weight_tensor_parameter_221: name = "parameter_221" + original_name = "batch_norm2d_25.w_0" shape = [768] dtype = "float32" min_val = float("-0.00392536") @@ -2246,6 +2468,7 @@ class Program_weight_tensor_parameter_221: class Program_weight_tensor_parameter_222: name = "parameter_222" + original_name = "batch_norm2d_25.w_2" shape = [768] dtype = "float32" min_val = float("5.60519e-45") @@ -2257,6 +2480,7 @@ class Program_weight_tensor_parameter_222: class Program_weight_tensor_parameter_223: name = "parameter_223" + original_name = "batch_norm2d_25.w_1" shape = [768] dtype = "float32" min_val = float("-24.8115") @@ -2268,6 +2492,7 @@ class Program_weight_tensor_parameter_223: class Program_weight_tensor_parameter_224: name = "parameter_224" + original_name = "conv2d_27.w_0" shape = [768, 1664, 1, 1] dtype = "float32" min_val = float("-1.63676") @@ -2279,6 +2504,7 @@ class Program_weight_tensor_parameter_224: class Program_weight_tensor_parameter_225: name = "parameter_225" + original_name = "batch_norm2d_24.b_0" shape = [192] dtype = "float32" min_val = float("-6.71315") @@ -2290,6 +2516,7 @@ class Program_weight_tensor_parameter_225: class Program_weight_tensor_parameter_226: name = "parameter_226" + original_name = "batch_norm2d_24.w_0" shape = [192] dtype = "float32" min_val = float("0.55321") @@ -2301,6 +2528,7 @@ class Program_weight_tensor_parameter_226: class Program_weight_tensor_parameter_227: name = "parameter_227" + original_name = "batch_norm2d_24.w_2" shape = [192] dtype = "float32" min_val = float("5.60519e-45") @@ -2312,6 +2540,7 @@ class Program_weight_tensor_parameter_227: class Program_weight_tensor_parameter_228: name = "parameter_228" + original_name = "batch_norm2d_24.w_1" shape = [192] dtype = "float32" min_val = float("-10.8853") @@ -2323,6 +2552,7 @@ class Program_weight_tensor_parameter_228: class Program_weight_tensor_parameter_229: name = "parameter_229" + original_name = "conv2d_26.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.981468") @@ -2334,6 +2564,7 @@ class Program_weight_tensor_parameter_229: class Program_weight_tensor_parameter_230: name = "parameter_230" + original_name = "batch_norm2d_23.b_0" shape = [192] dtype = "float32" min_val = float("-6.04808") @@ -2345,6 +2576,7 @@ class Program_weight_tensor_parameter_230: class Program_weight_tensor_parameter_231: name = "parameter_231" + original_name = "batch_norm2d_23.w_0" shape = [192] dtype = "float32" min_val = float("-0.000255708") @@ -2356,6 +2588,7 @@ class Program_weight_tensor_parameter_231: class Program_weight_tensor_parameter_232: name = "parameter_232" + original_name = "batch_norm2d_23.w_2" shape = [192] dtype = "float32" min_val = float("5.60519e-45") @@ -2367,6 +2600,7 @@ class Program_weight_tensor_parameter_232: class Program_weight_tensor_parameter_233: name = "parameter_233" + original_name = "batch_norm2d_23.w_1" shape = [192] dtype = "float32" min_val = float("-28.2793") @@ -2378,6 +2612,7 @@ class Program_weight_tensor_parameter_233: class Program_weight_tensor_parameter_234: name = "parameter_234" + original_name = "conv2d_25.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-1.06357") @@ -2389,6 +2624,7 @@ class Program_weight_tensor_parameter_234: class Program_weight_tensor_parameter_235: name = "parameter_235" + original_name = "batch_norm2d_22.b_0" shape = [192] dtype = "float32" min_val = float("-5.5729") @@ -2400,6 +2636,7 @@ class Program_weight_tensor_parameter_235: class Program_weight_tensor_parameter_236: name = "parameter_236" + original_name = "batch_norm2d_22.w_0" shape = [192] dtype = "float32" min_val = float("0.232818") @@ -2411,6 +2648,7 @@ class Program_weight_tensor_parameter_236: class Program_weight_tensor_parameter_237: name = "parameter_237" + original_name = "batch_norm2d_22.w_2" shape = [192] dtype = "float32" min_val = float("5.60519e-45") @@ -2422,6 +2660,7 @@ class Program_weight_tensor_parameter_237: class Program_weight_tensor_parameter_238: name = "parameter_238" + original_name = "batch_norm2d_22.w_1" shape = [192] dtype = "float32" min_val = float("-13.7212") @@ -2433,6 +2672,7 @@ class Program_weight_tensor_parameter_238: class Program_weight_tensor_parameter_239: name = "parameter_239" + original_name = "conv2d_24.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-1.26709") @@ -2444,6 +2684,7 @@ class Program_weight_tensor_parameter_239: class Program_weight_tensor_parameter_240: name = "parameter_240" + original_name = "batch_norm2d_21.b_0" shape = [192] dtype = "float32" min_val = float("-4.89989") @@ -2455,6 +2696,7 @@ class Program_weight_tensor_parameter_240: class Program_weight_tensor_parameter_241: name = "parameter_241" + original_name = "batch_norm2d_21.w_0" shape = [192] dtype = "float32" min_val = float("-0.000873162") @@ -2466,6 +2708,7 @@ class Program_weight_tensor_parameter_241: class Program_weight_tensor_parameter_242: name = "parameter_242" + original_name = "batch_norm2d_21.w_2" shape = [192] dtype = "float32" min_val = float("5.60519e-45") @@ -2477,6 +2720,7 @@ class Program_weight_tensor_parameter_242: class Program_weight_tensor_parameter_243: name = "parameter_243" + original_name = "batch_norm2d_21.w_1" shape = [192] dtype = "float32" min_val = float("-30.036") @@ -2488,6 +2732,7 @@ class Program_weight_tensor_parameter_243: class Program_weight_tensor_parameter_244: name = "parameter_244" + original_name = "conv2d_23.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-1.30847") @@ -2499,6 +2744,7 @@ class Program_weight_tensor_parameter_244: class Program_weight_tensor_parameter_245: name = "parameter_245" + original_name = "batch_norm2d_20.b_0" shape = [192] dtype = "float32" min_val = float("-6.91983") @@ -2510,6 +2756,7 @@ class Program_weight_tensor_parameter_245: class Program_weight_tensor_parameter_246: name = "parameter_246" + original_name = "batch_norm2d_20.w_0" shape = [192] dtype = "float32" min_val = float("0.352431") @@ -2521,6 +2768,7 @@ class Program_weight_tensor_parameter_246: class Program_weight_tensor_parameter_247: name = "parameter_247" + original_name = "batch_norm2d_20.w_2" shape = [192] dtype = "float32" min_val = float("5.60519e-45") @@ -2532,6 +2780,7 @@ class Program_weight_tensor_parameter_247: class Program_weight_tensor_parameter_248: name = "parameter_248" + original_name = "batch_norm2d_20.w_1" shape = [192] dtype = "float32" min_val = float("-31.7956") @@ -2543,6 +2792,7 @@ class Program_weight_tensor_parameter_248: class Program_weight_tensor_parameter_249: name = "parameter_249" + original_name = "conv2d_22.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-1.25657") @@ -2554,6 +2804,7 @@ class Program_weight_tensor_parameter_249: class Program_weight_tensor_parameter_250: name = "parameter_250" + original_name = "batch_norm2d_19.b_0" shape = [192] dtype = "float32" min_val = float("-10.1759") @@ -2565,6 +2816,7 @@ class Program_weight_tensor_parameter_250: class Program_weight_tensor_parameter_251: name = "parameter_251" + original_name = "batch_norm2d_19.w_0" shape = [192] dtype = "float32" min_val = float("0.440332") @@ -2576,6 +2828,7 @@ class Program_weight_tensor_parameter_251: class Program_weight_tensor_parameter_252: name = "parameter_252" + original_name = "batch_norm2d_19.w_2" shape = [192] dtype = "float32" min_val = float("5.60519e-45") @@ -2587,6 +2840,7 @@ class Program_weight_tensor_parameter_252: class Program_weight_tensor_parameter_253: name = "parameter_253" + original_name = "batch_norm2d_19.w_1" shape = [192] dtype = "float32" min_val = float("-58.6427") @@ -2598,6 +2852,7 @@ class Program_weight_tensor_parameter_253: class Program_weight_tensor_parameter_254: name = "parameter_254" + original_name = "conv2d_21.w_0" shape = [192, 512, 3, 3] dtype = "float32" min_val = float("-1.5312") @@ -2609,6 +2864,7 @@ class Program_weight_tensor_parameter_254: class Program_weight_tensor_parameter_255: name = "parameter_255" + original_name = "batch_norm2d_18.b_0" shape = [512] dtype = "float32" min_val = float("-3.50597") @@ -2620,6 +2876,7 @@ class Program_weight_tensor_parameter_255: class Program_weight_tensor_parameter_256: name = "parameter_256" + original_name = "batch_norm2d_18.w_0" shape = [512] dtype = "float32" min_val = float("-0.00907414") @@ -2631,6 +2888,7 @@ class Program_weight_tensor_parameter_256: class Program_weight_tensor_parameter_257: name = "parameter_257" + original_name = "batch_norm2d_18.w_2" shape = [512] dtype = "float32" min_val = float("5.60519e-45") @@ -2642,6 +2900,7 @@ class Program_weight_tensor_parameter_257: class Program_weight_tensor_parameter_258: name = "parameter_258" + original_name = "batch_norm2d_18.w_1" shape = [512] dtype = "float32" min_val = float("-0.855364") @@ -2653,6 +2912,7 @@ class Program_weight_tensor_parameter_258: class Program_weight_tensor_parameter_259: name = "parameter_259" + original_name = "conv2d_20.w_0" shape = [512, 1, 3, 3] dtype = "float32" min_val = float("-0.644089") @@ -2664,6 +2924,7 @@ class Program_weight_tensor_parameter_259: class Program_weight_tensor_parameter_260: name = "parameter_260" + original_name = "conv2d_19.b_0" shape = [512] dtype = "float32" min_val = float("-0.355783") @@ -2675,6 +2936,7 @@ class Program_weight_tensor_parameter_260: class Program_weight_tensor_parameter_261: name = "parameter_261" + original_name = "conv2d_19.w_0" shape = [512, 512, 1, 1] dtype = "float32" min_val = float("-1.75084") @@ -2686,6 +2948,7 @@ class Program_weight_tensor_parameter_261: class Program_weight_tensor_parameter_262: name = "parameter_262" + original_name = "batch_norm2d_17.b_0" shape = [512] dtype = "float32" min_val = float("-14.7231") @@ -2697,6 +2960,7 @@ class Program_weight_tensor_parameter_262: class Program_weight_tensor_parameter_263: name = "parameter_263" + original_name = "batch_norm2d_17.w_0" shape = [512] dtype = "float32" min_val = float("0.0742969") @@ -2708,6 +2972,7 @@ class Program_weight_tensor_parameter_263: class Program_weight_tensor_parameter_264: name = "parameter_264" + original_name = "batch_norm2d_17.w_2" shape = [512] dtype = "float32" min_val = float("5.60519e-45") @@ -2719,6 +2984,7 @@ class Program_weight_tensor_parameter_264: class Program_weight_tensor_parameter_265: name = "parameter_265" + original_name = "batch_norm2d_17.w_1" shape = [512] dtype = "float32" min_val = float("-21.1524") @@ -2730,6 +2996,7 @@ class Program_weight_tensor_parameter_265: class Program_weight_tensor_parameter_266: name = "parameter_266" + original_name = "conv2d_18.w_0" shape = [512, 1216, 1, 1] dtype = "float32" min_val = float("-1.50017") @@ -2741,6 +3008,7 @@ class Program_weight_tensor_parameter_266: class Program_weight_tensor_parameter_267: name = "parameter_267" + original_name = "batch_norm2d_16.b_0" shape = [160] dtype = "float32" min_val = float("-5.22583") @@ -2752,6 +3020,7 @@ class Program_weight_tensor_parameter_267: class Program_weight_tensor_parameter_268: name = "parameter_268" + original_name = "batch_norm2d_16.w_0" shape = [160] dtype = "float32" min_val = float("1.40603") @@ -2763,6 +3032,7 @@ class Program_weight_tensor_parameter_268: class Program_weight_tensor_parameter_269: name = "parameter_269" + original_name = "batch_norm2d_16.w_2" shape = [160] dtype = "float32" min_val = float("2.00589") @@ -2774,6 +3044,7 @@ class Program_weight_tensor_parameter_269: class Program_weight_tensor_parameter_270: name = "parameter_270" + original_name = "batch_norm2d_16.w_1" shape = [160] dtype = "float32" min_val = float("-17.5281") @@ -2785,6 +3056,7 @@ class Program_weight_tensor_parameter_270: class Program_weight_tensor_parameter_271: name = "parameter_271" + original_name = "conv2d_17.w_0" shape = [160, 160, 3, 3] dtype = "float32" min_val = float("-1.04462") @@ -2796,6 +3068,7 @@ class Program_weight_tensor_parameter_271: class Program_weight_tensor_parameter_272: name = "parameter_272" + original_name = "batch_norm2d_15.b_0" shape = [160] dtype = "float32" min_val = float("-6.12277") @@ -2807,6 +3080,7 @@ class Program_weight_tensor_parameter_272: class Program_weight_tensor_parameter_273: name = "parameter_273" + original_name = "batch_norm2d_15.w_0" shape = [160] dtype = "float32" min_val = float("0.552904") @@ -2818,6 +3092,7 @@ class Program_weight_tensor_parameter_273: class Program_weight_tensor_parameter_274: name = "parameter_274" + original_name = "batch_norm2d_15.w_2" shape = [160] dtype = "float32" min_val = float("5.60519e-45") @@ -2829,6 +3104,7 @@ class Program_weight_tensor_parameter_274: class Program_weight_tensor_parameter_275: name = "parameter_275" + original_name = "batch_norm2d_15.w_1" shape = [160] dtype = "float32" min_val = float("-21.34") @@ -2840,6 +3116,7 @@ class Program_weight_tensor_parameter_275: class Program_weight_tensor_parameter_276: name = "parameter_276" + original_name = "conv2d_16.w_0" shape = [160, 160, 3, 3] dtype = "float32" min_val = float("-1.20127") @@ -2851,6 +3128,7 @@ class Program_weight_tensor_parameter_276: class Program_weight_tensor_parameter_277: name = "parameter_277" + original_name = "batch_norm2d_14.b_0" shape = [160] dtype = "float32" min_val = float("-6.35455") @@ -2862,6 +3140,7 @@ class Program_weight_tensor_parameter_277: class Program_weight_tensor_parameter_278: name = "parameter_278" + original_name = "batch_norm2d_14.w_0" shape = [160] dtype = "float32" min_val = float("0.942598") @@ -2873,6 +3152,7 @@ class Program_weight_tensor_parameter_278: class Program_weight_tensor_parameter_279: name = "parameter_279" + original_name = "batch_norm2d_14.w_2" shape = [160] dtype = "float32" min_val = float("5.60519e-45") @@ -2884,6 +3164,7 @@ class Program_weight_tensor_parameter_279: class Program_weight_tensor_parameter_280: name = "parameter_280" + original_name = "batch_norm2d_14.w_1" shape = [160] dtype = "float32" min_val = float("-19.1812") @@ -2895,6 +3176,7 @@ class Program_weight_tensor_parameter_280: class Program_weight_tensor_parameter_281: name = "parameter_281" + original_name = "conv2d_15.w_0" shape = [160, 160, 3, 3] dtype = "float32" min_val = float("-1.16529") @@ -2906,6 +3188,7 @@ class Program_weight_tensor_parameter_281: class Program_weight_tensor_parameter_282: name = "parameter_282" + original_name = "batch_norm2d_13.b_0" shape = [160] dtype = "float32" min_val = float("-6.40401") @@ -2917,6 +3200,7 @@ class Program_weight_tensor_parameter_282: class Program_weight_tensor_parameter_283: name = "parameter_283" + original_name = "batch_norm2d_13.w_0" shape = [160] dtype = "float32" min_val = float("0.26433") @@ -2928,6 +3212,7 @@ class Program_weight_tensor_parameter_283: class Program_weight_tensor_parameter_284: name = "parameter_284" + original_name = "batch_norm2d_13.w_2" shape = [160] dtype = "float32" min_val = float("5.60519e-45") @@ -2939,6 +3224,7 @@ class Program_weight_tensor_parameter_284: class Program_weight_tensor_parameter_285: name = "parameter_285" + original_name = "batch_norm2d_13.w_1" shape = [160] dtype = "float32" min_val = float("-51.1503") @@ -2950,6 +3236,7 @@ class Program_weight_tensor_parameter_285: class Program_weight_tensor_parameter_286: name = "parameter_286" + original_name = "conv2d_14.w_0" shape = [160, 160, 3, 3] dtype = "float32" min_val = float("-1.19937") @@ -2961,6 +3248,7 @@ class Program_weight_tensor_parameter_286: class Program_weight_tensor_parameter_287: name = "parameter_287" + original_name = "batch_norm2d_12.b_0" shape = [160] dtype = "float32" min_val = float("-5.6566") @@ -2972,6 +3260,7 @@ class Program_weight_tensor_parameter_287: class Program_weight_tensor_parameter_288: name = "parameter_288" + original_name = "batch_norm2d_12.w_0" shape = [160] dtype = "float32" min_val = float("0.670406") @@ -2983,6 +3272,7 @@ class Program_weight_tensor_parameter_288: class Program_weight_tensor_parameter_289: name = "parameter_289" + original_name = "batch_norm2d_12.w_2" shape = [160] dtype = "float32" min_val = float("5.52133") @@ -2994,6 +3284,7 @@ class Program_weight_tensor_parameter_289: class Program_weight_tensor_parameter_290: name = "parameter_290" + original_name = "batch_norm2d_12.w_1" shape = [160] dtype = "float32" min_val = float("-44.1152") @@ -3005,6 +3296,7 @@ class Program_weight_tensor_parameter_290: class Program_weight_tensor_parameter_291: name = "parameter_291" + original_name = "conv2d_13.w_0" shape = [160, 160, 3, 3] dtype = "float32" min_val = float("-1.16644") @@ -3016,6 +3308,7 @@ class Program_weight_tensor_parameter_291: class Program_weight_tensor_parameter_292: name = "parameter_292" + original_name = "batch_norm2d_11.b_0" shape = [160] dtype = "float32" min_val = float("-9.91221") @@ -3027,6 +3320,7 @@ class Program_weight_tensor_parameter_292: class Program_weight_tensor_parameter_293: name = "parameter_293" + original_name = "batch_norm2d_11.w_0" shape = [160] dtype = "float32" min_val = float("0.992803") @@ -3038,6 +3332,7 @@ class Program_weight_tensor_parameter_293: class Program_weight_tensor_parameter_294: name = "parameter_294" + original_name = "batch_norm2d_11.w_2" shape = [160] dtype = "float32" min_val = float("21.3369") @@ -3049,6 +3344,7 @@ class Program_weight_tensor_parameter_294: class Program_weight_tensor_parameter_295: name = "parameter_295" + original_name = "batch_norm2d_11.w_1" shape = [160] dtype = "float32" min_val = float("-41.2149") @@ -3060,6 +3356,7 @@ class Program_weight_tensor_parameter_295: class Program_weight_tensor_parameter_296: name = "parameter_296" + original_name = "conv2d_12.w_0" shape = [160, 256, 3, 3] dtype = "float32" min_val = float("-1.4183") @@ -3071,6 +3368,7 @@ class Program_weight_tensor_parameter_296: class Program_weight_tensor_parameter_297: name = "parameter_297" + original_name = "batch_norm2d_10.b_0" shape = [256] dtype = "float32" min_val = float("-2.66667") @@ -3082,6 +3380,7 @@ class Program_weight_tensor_parameter_297: class Program_weight_tensor_parameter_298: name = "parameter_298" + original_name = "batch_norm2d_10.w_0" shape = [256] dtype = "float32" min_val = float("-0.0187076") @@ -3093,6 +3392,7 @@ class Program_weight_tensor_parameter_298: class Program_weight_tensor_parameter_299: name = "parameter_299" + original_name = "batch_norm2d_10.w_2" shape = [256] dtype = "float32" min_val = float("5.60519e-45") @@ -3104,6 +3404,7 @@ class Program_weight_tensor_parameter_299: class Program_weight_tensor_parameter_300: name = "parameter_300" + original_name = "batch_norm2d_10.w_1" shape = [256] dtype = "float32" min_val = float("-0.859712") @@ -3115,6 +3416,7 @@ class Program_weight_tensor_parameter_300: class Program_weight_tensor_parameter_301: name = "parameter_301" + original_name = "conv2d_11.w_0" shape = [256, 1, 3, 3] dtype = "float32" min_val = float("-0.773494") @@ -3126,6 +3428,7 @@ class Program_weight_tensor_parameter_301: class Program_weight_tensor_parameter_302: name = "parameter_302" + original_name = "conv2d_10.b_0" shape = [256] dtype = "float32" min_val = float("-0.391566") @@ -3137,6 +3440,7 @@ class Program_weight_tensor_parameter_302: class Program_weight_tensor_parameter_303: name = "parameter_303" + original_name = "conv2d_10.w_0" shape = [256, 256, 1, 1] dtype = "float32" min_val = float("-2.0974") @@ -3148,6 +3452,7 @@ class Program_weight_tensor_parameter_303: class Program_weight_tensor_parameter_304: name = "parameter_304" + original_name = "batch_norm2d_9.b_0" shape = [256] dtype = "float32" min_val = float("-10.5491") @@ -3159,6 +3464,7 @@ class Program_weight_tensor_parameter_304: class Program_weight_tensor_parameter_305: name = "parameter_305" + original_name = "batch_norm2d_9.w_0" shape = [256] dtype = "float32" min_val = float("0.00168943") @@ -3170,6 +3476,7 @@ class Program_weight_tensor_parameter_305: class Program_weight_tensor_parameter_306: name = "parameter_306" + original_name = "batch_norm2d_9.w_2" shape = [256] dtype = "float32" min_val = float("5.60519e-45") @@ -3181,6 +3488,7 @@ class Program_weight_tensor_parameter_306: class Program_weight_tensor_parameter_307: name = "parameter_307" + original_name = "batch_norm2d_9.w_1" shape = [256] dtype = "float32" min_val = float("-24.8729") @@ -3192,6 +3500,7 @@ class Program_weight_tensor_parameter_307: class Program_weight_tensor_parameter_308: name = "parameter_308" + original_name = "conv2d_9.w_0" shape = [256, 896, 1, 1] dtype = "float32" min_val = float("-2.10959") @@ -3203,6 +3512,7 @@ class Program_weight_tensor_parameter_308: class Program_weight_tensor_parameter_309: name = "parameter_309" + original_name = "batch_norm2d_8.b_0" shape = [128] dtype = "float32" min_val = float("-6.48719") @@ -3214,6 +3524,7 @@ class Program_weight_tensor_parameter_309: class Program_weight_tensor_parameter_310: name = "parameter_310" + original_name = "batch_norm2d_8.w_0" shape = [128] dtype = "float32" min_val = float("1.37939") @@ -3225,6 +3536,7 @@ class Program_weight_tensor_parameter_310: class Program_weight_tensor_parameter_311: name = "parameter_311" + original_name = "batch_norm2d_8.w_2" shape = [128] dtype = "float32" min_val = float("2.95682") @@ -3236,6 +3548,7 @@ class Program_weight_tensor_parameter_311: class Program_weight_tensor_parameter_312: name = "parameter_312" + original_name = "batch_norm2d_8.w_1" shape = [128] dtype = "float32" min_val = float("-16.9052") @@ -3247,6 +3560,7 @@ class Program_weight_tensor_parameter_312: class Program_weight_tensor_parameter_313: name = "parameter_313" + original_name = "conv2d_8.w_0" shape = [128, 128, 3, 3] dtype = "float32" min_val = float("-0.863172") @@ -3258,6 +3572,7 @@ class Program_weight_tensor_parameter_313: class Program_weight_tensor_parameter_314: name = "parameter_314" + original_name = "batch_norm2d_7.b_0" shape = [128] dtype = "float32" min_val = float("-8.94302") @@ -3269,6 +3584,7 @@ class Program_weight_tensor_parameter_314: class Program_weight_tensor_parameter_315: name = "parameter_315" + original_name = "batch_norm2d_7.w_0" shape = [128] dtype = "float32" min_val = float("1.31524") @@ -3280,6 +3596,7 @@ class Program_weight_tensor_parameter_315: class Program_weight_tensor_parameter_316: name = "parameter_316" + original_name = "batch_norm2d_7.w_2" shape = [128] dtype = "float32" min_val = float("9.8959") @@ -3291,6 +3608,7 @@ class Program_weight_tensor_parameter_316: class Program_weight_tensor_parameter_317: name = "parameter_317" + original_name = "batch_norm2d_7.w_1" shape = [128] dtype = "float32" min_val = float("-31.9293") @@ -3302,6 +3620,7 @@ class Program_weight_tensor_parameter_317: class Program_weight_tensor_parameter_318: name = "parameter_318" + original_name = "conv2d_7.w_0" shape = [128, 128, 3, 3] dtype = "float32" min_val = float("-1.03811") @@ -3313,6 +3632,7 @@ class Program_weight_tensor_parameter_318: class Program_weight_tensor_parameter_319: name = "parameter_319" + original_name = "batch_norm2d_6.b_0" shape = [128] dtype = "float32" min_val = float("-6.0996") @@ -3324,6 +3644,7 @@ class Program_weight_tensor_parameter_319: class Program_weight_tensor_parameter_320: name = "parameter_320" + original_name = "batch_norm2d_6.w_0" shape = [128] dtype = "float32" min_val = float("1.17167") @@ -3335,6 +3656,7 @@ class Program_weight_tensor_parameter_320: class Program_weight_tensor_parameter_321: name = "parameter_321" + original_name = "batch_norm2d_6.w_2" shape = [128] dtype = "float32" min_val = float("21.7332") @@ -3346,6 +3668,7 @@ class Program_weight_tensor_parameter_321: class Program_weight_tensor_parameter_322: name = "parameter_322" + original_name = "batch_norm2d_6.w_1" shape = [128] dtype = "float32" min_val = float("-47.7969") @@ -3357,6 +3680,7 @@ class Program_weight_tensor_parameter_322: class Program_weight_tensor_parameter_323: name = "parameter_323" + original_name = "conv2d_6.w_0" shape = [128, 128, 3, 3] dtype = "float32" min_val = float("-1.21833") @@ -3368,6 +3692,7 @@ class Program_weight_tensor_parameter_323: class Program_weight_tensor_parameter_324: name = "parameter_324" + original_name = "batch_norm2d_5.b_0" shape = [128] dtype = "float32" min_val = float("-4.65014") @@ -3379,6 +3704,7 @@ class Program_weight_tensor_parameter_324: class Program_weight_tensor_parameter_325: name = "parameter_325" + original_name = "batch_norm2d_5.w_0" shape = [128] dtype = "float32" min_val = float("1.01169") @@ -3390,6 +3716,7 @@ class Program_weight_tensor_parameter_325: class Program_weight_tensor_parameter_326: name = "parameter_326" + original_name = "batch_norm2d_5.w_2" shape = [128] dtype = "float32" min_val = float("44.3861") @@ -3401,6 +3728,7 @@ class Program_weight_tensor_parameter_326: class Program_weight_tensor_parameter_327: name = "parameter_327" + original_name = "batch_norm2d_5.w_1" shape = [128] dtype = "float32" min_val = float("-29.819") @@ -3412,6 +3740,7 @@ class Program_weight_tensor_parameter_327: class Program_weight_tensor_parameter_328: name = "parameter_328" + original_name = "conv2d_5.w_0" shape = [128, 128, 3, 3] dtype = "float32" min_val = float("-1.40128") @@ -3423,6 +3752,7 @@ class Program_weight_tensor_parameter_328: class Program_weight_tensor_parameter_329: name = "parameter_329" + original_name = "batch_norm2d_4.b_0" shape = [128] dtype = "float32" min_val = float("-9.50361") @@ -3434,6 +3764,7 @@ class Program_weight_tensor_parameter_329: class Program_weight_tensor_parameter_330: name = "parameter_330" + original_name = "batch_norm2d_4.w_0" shape = [128] dtype = "float32" min_val = float("1.07329") @@ -3445,6 +3776,7 @@ class Program_weight_tensor_parameter_330: class Program_weight_tensor_parameter_331: name = "parameter_331" + original_name = "batch_norm2d_4.w_2" shape = [128] dtype = "float32" min_val = float("0.851679") @@ -3456,6 +3788,7 @@ class Program_weight_tensor_parameter_331: class Program_weight_tensor_parameter_332: name = "parameter_332" + original_name = "batch_norm2d_4.w_1" shape = [128] dtype = "float32" min_val = float("-34.1777") @@ -3467,6 +3800,7 @@ class Program_weight_tensor_parameter_332: class Program_weight_tensor_parameter_333: name = "parameter_333" + original_name = "conv2d_4.w_0" shape = [128, 128, 3, 3] dtype = "float32" min_val = float("-1.51044") @@ -3478,6 +3812,7 @@ class Program_weight_tensor_parameter_333: class Program_weight_tensor_parameter_334: name = "parameter_334" + original_name = "batch_norm2d_3.b_0" shape = [128] dtype = "float32" min_val = float("-7.06732") @@ -3489,6 +3824,7 @@ class Program_weight_tensor_parameter_334: class Program_weight_tensor_parameter_335: name = "parameter_335" + original_name = "batch_norm2d_3.w_0" shape = [128] dtype = "float32" min_val = float("1.17554") @@ -3500,6 +3836,7 @@ class Program_weight_tensor_parameter_335: class Program_weight_tensor_parameter_336: name = "parameter_336" + original_name = "batch_norm2d_3.w_2" shape = [128] dtype = "float32" min_val = float("5.60519e-45") @@ -3511,6 +3848,7 @@ class Program_weight_tensor_parameter_336: class Program_weight_tensor_parameter_337: name = "parameter_337" + original_name = "batch_norm2d_3.w_1" shape = [128] dtype = "float32" min_val = float("-43.5263") @@ -3522,6 +3860,7 @@ class Program_weight_tensor_parameter_337: class Program_weight_tensor_parameter_338: name = "parameter_338" + original_name = "conv2d_3.w_0" shape = [128, 128, 3, 3] dtype = "float32" min_val = float("-1.66397") @@ -3533,6 +3872,7 @@ class Program_weight_tensor_parameter_338: class Program_weight_tensor_parameter_339: name = "parameter_339" + original_name = "batch_norm2d_2.b_0" shape = [128] dtype = "float32" min_val = float("-1.614") @@ -3544,6 +3884,7 @@ class Program_weight_tensor_parameter_339: class Program_weight_tensor_parameter_340: name = "parameter_340" + original_name = "batch_norm2d_2.w_0" shape = [128] dtype = "float32" min_val = float("0.726712") @@ -3555,6 +3896,7 @@ class Program_weight_tensor_parameter_340: class Program_weight_tensor_parameter_341: name = "parameter_341" + original_name = "batch_norm2d_2.w_2" shape = [128] dtype = "float32" min_val = float("55.6017") @@ -3566,6 +3908,7 @@ class Program_weight_tensor_parameter_341: class Program_weight_tensor_parameter_342: name = "parameter_342" + original_name = "batch_norm2d_2.w_1" shape = [128] dtype = "float32" min_val = float("-76.7242") @@ -3577,6 +3920,7 @@ class Program_weight_tensor_parameter_342: class Program_weight_tensor_parameter_343: name = "parameter_343" + original_name = "conv2d_2.w_0" shape = [128, 64, 3, 3] dtype = "float32" min_val = float("-2.16191") @@ -3588,6 +3932,7 @@ class Program_weight_tensor_parameter_343: class Program_weight_tensor_parameter_344: name = "parameter_344" + original_name = "batch_norm2d_1.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -3597,6 +3942,7 @@ class Program_weight_tensor_parameter_344: class Program_weight_tensor_parameter_345: name = "parameter_345" + original_name = "batch_norm2d_1.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -3606,6 +3952,7 @@ class Program_weight_tensor_parameter_345: class Program_weight_tensor_parameter_346: name = "parameter_346" + original_name = "batch_norm2d_1.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -3615,6 +3962,7 @@ class Program_weight_tensor_parameter_346: class Program_weight_tensor_parameter_347: name = "parameter_347" + original_name = "batch_norm2d_1.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -3624,6 +3972,7 @@ class Program_weight_tensor_parameter_347: class Program_weight_tensor_parameter_348: name = "parameter_348" + original_name = "conv2d_1.w_0" shape = [64, 64, 3, 3] dtype = "float32" min_val = float("-2.69629") @@ -3635,6 +3984,7 @@ class Program_weight_tensor_parameter_348: class Program_weight_tensor_parameter_349: name = "parameter_349" + original_name = "batch_norm2d_0.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -3644,6 +3994,7 @@ class Program_weight_tensor_parameter_349: class Program_weight_tensor_parameter_350: name = "parameter_350" + original_name = "batch_norm2d_0.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -3653,6 +4004,7 @@ class Program_weight_tensor_parameter_350: class Program_weight_tensor_parameter_351: name = "parameter_351" + original_name = "batch_norm2d_0.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -3662,6 +4014,7 @@ class Program_weight_tensor_parameter_351: class Program_weight_tensor_parameter_352: name = "parameter_352" + original_name = "batch_norm2d_0.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -3671,6 +4024,7 @@ class Program_weight_tensor_parameter_352: class Program_weight_tensor_parameter_353: name = "parameter_353" + original_name = "conv2d_0.w_0" shape = [64, 3, 3, 3] dtype = "float32" min_val = float("-0.805409") diff --git a/paddle_samples/PaddleX/PP-OCRv4_server_rec/subgraph_0/input_meta.py b/paddle_samples/PaddleX/PP-OCRv4_server_rec/subgraph_0/input_meta.py index e8cc5cbd4..a06f8ac17 100644 --- a/paddle_samples/PaddleX/PP-OCRv4_server_rec/subgraph_0/input_meta.py +++ b/paddle_samples/PaddleX/PP-OCRv4_server_rec/subgraph_0/input_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_data_0: name = "data_0" + original_name = "var_1130" shape = [8, 3, 48, 320] dtype = "float32" min_val = float("-1.0") diff --git a/paddle_samples/PaddleX/PP-OCRv4_server_rec/subgraph_0/weight_meta.py b/paddle_samples/PaddleX/PP-OCRv4_server_rec/subgraph_0/weight_meta.py index e4cd4220c..0ca79855b 100644 --- a/paddle_samples/PaddleX/PP-OCRv4_server_rec/subgraph_0/weight_meta.py +++ b/paddle_samples/PaddleX/PP-OCRv4_server_rec/subgraph_0/weight_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_parameter_0: name = "parameter_0" + original_name = "conv2d_46.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0106389") @@ -11,6 +12,7 @@ class Program_weight_tensor_parameter_0: class Program_weight_tensor_parameter_1: name = "parameter_1" + original_name = "conv2d_46.w_0" shape = [1024, 1024, 1, 1] dtype = "float32" min_val = float("-0.346015") @@ -22,6 +24,7 @@ class Program_weight_tensor_parameter_1: class Program_weight_tensor_parameter_2: name = "parameter_2" + original_name = "batch_norm2d_41.b_0" shape = [1024] dtype = "float32" min_val = float("-4.61392") @@ -33,6 +36,7 @@ class Program_weight_tensor_parameter_2: class Program_weight_tensor_parameter_3: name = "parameter_3" + original_name = "batch_norm2d_41.w_0" shape = [1024] dtype = "float32" min_val = float("0.0784453") @@ -44,6 +48,7 @@ class Program_weight_tensor_parameter_3: class Program_weight_tensor_parameter_4: name = "parameter_4" + original_name = "batch_norm2d_41.w_2" shape = [1024] dtype = "float32" min_val = float("5.60519e-45") @@ -55,6 +60,7 @@ class Program_weight_tensor_parameter_4: class Program_weight_tensor_parameter_5: name = "parameter_5" + original_name = "batch_norm2d_41.w_1" shape = [1024] dtype = "float32" min_val = float("-3.95853") @@ -66,6 +72,7 @@ class Program_weight_tensor_parameter_5: class Program_weight_tensor_parameter_6: name = "parameter_6" + original_name = "conv2d_45.w_0" shape = [1024, 2112, 1, 1] dtype = "float32" min_val = float("-0.62983") @@ -77,6 +84,7 @@ class Program_weight_tensor_parameter_6: class Program_weight_tensor_parameter_7: name = "parameter_7" + original_name = "batch_norm2d_40.b_0" shape = [224] dtype = "float32" min_val = float("-5.43051") @@ -88,6 +96,7 @@ class Program_weight_tensor_parameter_7: class Program_weight_tensor_parameter_8: name = "parameter_8" + original_name = "batch_norm2d_40.w_0" shape = [224] dtype = "float32" min_val = float("0.128111") @@ -99,6 +108,7 @@ class Program_weight_tensor_parameter_8: class Program_weight_tensor_parameter_9: name = "parameter_9" + original_name = "batch_norm2d_40.w_2" shape = [224] dtype = "float32" min_val = float("5.60519e-45") @@ -110,6 +120,7 @@ class Program_weight_tensor_parameter_9: class Program_weight_tensor_parameter_10: name = "parameter_10" + original_name = "batch_norm2d_40.w_1" shape = [224] dtype = "float32" min_val = float("-0.424584") @@ -121,6 +132,7 @@ class Program_weight_tensor_parameter_10: class Program_weight_tensor_parameter_11: name = "parameter_11" + original_name = "conv2d_44.w_0" shape = [224, 224, 3, 3] dtype = "float32" min_val = float("-0.255276") @@ -132,6 +144,7 @@ class Program_weight_tensor_parameter_11: class Program_weight_tensor_parameter_12: name = "parameter_12" + original_name = "batch_norm2d_39.b_0" shape = [224] dtype = "float32" min_val = float("-2.51503") @@ -143,6 +156,7 @@ class Program_weight_tensor_parameter_12: class Program_weight_tensor_parameter_13: name = "parameter_13" + original_name = "batch_norm2d_39.w_0" shape = [224] dtype = "float32" min_val = float("0.146254") @@ -154,6 +168,7 @@ class Program_weight_tensor_parameter_13: class Program_weight_tensor_parameter_14: name = "parameter_14" + original_name = "batch_norm2d_39.w_2" shape = [224] dtype = "float32" min_val = float("5.60519e-45") @@ -165,6 +180,7 @@ class Program_weight_tensor_parameter_14: class Program_weight_tensor_parameter_15: name = "parameter_15" + original_name = "batch_norm2d_39.w_1" shape = [224] dtype = "float32" min_val = float("-1.37687") @@ -176,6 +192,7 @@ class Program_weight_tensor_parameter_15: class Program_weight_tensor_parameter_16: name = "parameter_16" + original_name = "conv2d_43.w_0" shape = [224, 224, 3, 3] dtype = "float32" min_val = float("-0.281075") @@ -187,6 +204,7 @@ class Program_weight_tensor_parameter_16: class Program_weight_tensor_parameter_17: name = "parameter_17" + original_name = "batch_norm2d_38.b_0" shape = [224] dtype = "float32" min_val = float("-3.0557") @@ -198,6 +216,7 @@ class Program_weight_tensor_parameter_17: class Program_weight_tensor_parameter_18: name = "parameter_18" + original_name = "batch_norm2d_38.w_0" shape = [224] dtype = "float32" min_val = float("0.0807891") @@ -209,6 +228,7 @@ class Program_weight_tensor_parameter_18: class Program_weight_tensor_parameter_19: name = "parameter_19" + original_name = "batch_norm2d_38.w_2" shape = [224] dtype = "float32" min_val = float("5.60519e-45") @@ -220,6 +240,7 @@ class Program_weight_tensor_parameter_19: class Program_weight_tensor_parameter_20: name = "parameter_20" + original_name = "batch_norm2d_38.w_1" shape = [224] dtype = "float32" min_val = float("-0.449152") @@ -231,6 +252,7 @@ class Program_weight_tensor_parameter_20: class Program_weight_tensor_parameter_21: name = "parameter_21" + original_name = "conv2d_42.w_0" shape = [224, 224, 3, 3] dtype = "float32" min_val = float("-0.356335") @@ -242,6 +264,7 @@ class Program_weight_tensor_parameter_21: class Program_weight_tensor_parameter_22: name = "parameter_22" + original_name = "batch_norm2d_37.b_0" shape = [224] dtype = "float32" min_val = float("-2.35244") @@ -253,6 +276,7 @@ class Program_weight_tensor_parameter_22: class Program_weight_tensor_parameter_23: name = "parameter_23" + original_name = "batch_norm2d_37.w_0" shape = [224] dtype = "float32" min_val = float("0.144582") @@ -264,6 +288,7 @@ class Program_weight_tensor_parameter_23: class Program_weight_tensor_parameter_24: name = "parameter_24" + original_name = "batch_norm2d_37.w_2" shape = [224] dtype = "float32" min_val = float("5.60519e-45") @@ -275,6 +300,7 @@ class Program_weight_tensor_parameter_24: class Program_weight_tensor_parameter_25: name = "parameter_25" + original_name = "batch_norm2d_37.w_1" shape = [224] dtype = "float32" min_val = float("-0.796251") @@ -286,6 +312,7 @@ class Program_weight_tensor_parameter_25: class Program_weight_tensor_parameter_26: name = "parameter_26" + original_name = "conv2d_41.w_0" shape = [224, 224, 3, 3] dtype = "float32" min_val = float("-0.296061") @@ -297,6 +324,7 @@ class Program_weight_tensor_parameter_26: class Program_weight_tensor_parameter_27: name = "parameter_27" + original_name = "batch_norm2d_36.b_0" shape = [224] dtype = "float32" min_val = float("-1.86946") @@ -308,6 +336,7 @@ class Program_weight_tensor_parameter_27: class Program_weight_tensor_parameter_28: name = "parameter_28" + original_name = "batch_norm2d_36.w_0" shape = [224] dtype = "float32" min_val = float("0.154221") @@ -319,6 +348,7 @@ class Program_weight_tensor_parameter_28: class Program_weight_tensor_parameter_29: name = "parameter_29" + original_name = "batch_norm2d_36.w_2" shape = [224] dtype = "float32" min_val = float("5.60519e-45") @@ -330,6 +360,7 @@ class Program_weight_tensor_parameter_29: class Program_weight_tensor_parameter_30: name = "parameter_30" + original_name = "batch_norm2d_36.w_1" shape = [224] dtype = "float32" min_val = float("-0.213286") @@ -341,6 +372,7 @@ class Program_weight_tensor_parameter_30: class Program_weight_tensor_parameter_31: name = "parameter_31" + original_name = "conv2d_40.w_0" shape = [224, 224, 3, 3] dtype = "float32" min_val = float("-0.284469") @@ -352,6 +384,7 @@ class Program_weight_tensor_parameter_31: class Program_weight_tensor_parameter_32: name = "parameter_32" + original_name = "batch_norm2d_35.b_0" shape = [224] dtype = "float32" min_val = float("-3.86036") @@ -363,6 +396,7 @@ class Program_weight_tensor_parameter_32: class Program_weight_tensor_parameter_33: name = "parameter_33" + original_name = "batch_norm2d_35.w_0" shape = [224] dtype = "float32" min_val = float("0.0661827") @@ -374,6 +408,7 @@ class Program_weight_tensor_parameter_33: class Program_weight_tensor_parameter_34: name = "parameter_34" + original_name = "batch_norm2d_35.w_2" shape = [224] dtype = "float32" min_val = float("5.60519e-45") @@ -385,6 +420,7 @@ class Program_weight_tensor_parameter_34: class Program_weight_tensor_parameter_35: name = "parameter_35" + original_name = "batch_norm2d_35.w_1" shape = [224] dtype = "float32" min_val = float("-5.42916") @@ -396,6 +432,7 @@ class Program_weight_tensor_parameter_35: class Program_weight_tensor_parameter_36: name = "parameter_36" + original_name = "conv2d_39.w_0" shape = [224, 768, 3, 3] dtype = "float32" min_val = float("-0.608597") @@ -407,6 +444,7 @@ class Program_weight_tensor_parameter_36: class Program_weight_tensor_parameter_37: name = "parameter_37" + original_name = "batch_norm2d_34.b_0" shape = [768] dtype = "float32" min_val = float("-3.47093") @@ -418,6 +456,7 @@ class Program_weight_tensor_parameter_37: class Program_weight_tensor_parameter_38: name = "parameter_38" + original_name = "batch_norm2d_34.w_0" shape = [768] dtype = "float32" min_val = float("-0.0509393") @@ -429,6 +468,7 @@ class Program_weight_tensor_parameter_38: class Program_weight_tensor_parameter_39: name = "parameter_39" + original_name = "batch_norm2d_34.w_2" shape = [768] dtype = "float32" min_val = float("5.60519e-45") @@ -440,6 +480,7 @@ class Program_weight_tensor_parameter_39: class Program_weight_tensor_parameter_40: name = "parameter_40" + original_name = "batch_norm2d_34.w_1" shape = [768] dtype = "float32" min_val = float("-0.172708") @@ -451,6 +492,7 @@ class Program_weight_tensor_parameter_40: class Program_weight_tensor_parameter_41: name = "parameter_41" + original_name = "conv2d_38.w_0" shape = [768, 1, 3, 3] dtype = "float32" min_val = float("-0.251818") @@ -462,6 +504,7 @@ class Program_weight_tensor_parameter_41: class Program_weight_tensor_parameter_42: name = "parameter_42" + original_name = "conv2d_37.b_0" shape = [768] dtype = "float32" min_val = float("-0.0431036") @@ -473,6 +516,7 @@ class Program_weight_tensor_parameter_42: class Program_weight_tensor_parameter_43: name = "parameter_43" + original_name = "conv2d_37.w_0" shape = [768, 768, 1, 1] dtype = "float32" min_val = float("-0.584867") @@ -484,6 +528,7 @@ class Program_weight_tensor_parameter_43: class Program_weight_tensor_parameter_44: name = "parameter_44" + original_name = "batch_norm2d_33.b_0" shape = [768] dtype = "float32" min_val = float("-9.56825") @@ -495,6 +540,7 @@ class Program_weight_tensor_parameter_44: class Program_weight_tensor_parameter_45: name = "parameter_45" + original_name = "batch_norm2d_33.w_0" shape = [768] dtype = "float32" min_val = float("0.0115258") @@ -506,6 +552,7 @@ class Program_weight_tensor_parameter_45: class Program_weight_tensor_parameter_46: name = "parameter_46" + original_name = "batch_norm2d_33.w_2" shape = [768] dtype = "float32" min_val = float("5.60519e-45") @@ -517,6 +564,7 @@ class Program_weight_tensor_parameter_46: class Program_weight_tensor_parameter_47: name = "parameter_47" + original_name = "batch_norm2d_33.w_1" shape = [768] dtype = "float32" min_val = float("-0.605102") @@ -528,6 +576,7 @@ class Program_weight_tensor_parameter_47: class Program_weight_tensor_parameter_48: name = "parameter_48" + original_name = "conv2d_36.w_0" shape = [768, 1920, 1, 1] dtype = "float32" min_val = float("-0.368661") @@ -539,6 +588,7 @@ class Program_weight_tensor_parameter_48: class Program_weight_tensor_parameter_49: name = "parameter_49" + original_name = "batch_norm2d_32.b_0" shape = [192] dtype = "float32" min_val = float("-2.36697") @@ -550,6 +600,7 @@ class Program_weight_tensor_parameter_49: class Program_weight_tensor_parameter_50: name = "parameter_50" + original_name = "batch_norm2d_32.w_0" shape = [192] dtype = "float32" min_val = float("0.169035") @@ -561,6 +612,7 @@ class Program_weight_tensor_parameter_50: class Program_weight_tensor_parameter_51: name = "parameter_51" + original_name = "batch_norm2d_32.w_2" shape = [192] dtype = "float32" min_val = float("5.60519e-45") @@ -572,6 +624,7 @@ class Program_weight_tensor_parameter_51: class Program_weight_tensor_parameter_52: name = "parameter_52" + original_name = "batch_norm2d_32.w_1" shape = [192] dtype = "float32" min_val = float("-0.40271") @@ -583,6 +636,7 @@ class Program_weight_tensor_parameter_52: class Program_weight_tensor_parameter_53: name = "parameter_53" + original_name = "conv2d_35.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.242957") @@ -594,6 +648,7 @@ class Program_weight_tensor_parameter_53: class Program_weight_tensor_parameter_54: name = "parameter_54" + original_name = "batch_norm2d_31.b_0" shape = [192] dtype = "float32" min_val = float("-2.02743") @@ -605,6 +660,7 @@ class Program_weight_tensor_parameter_54: class Program_weight_tensor_parameter_55: name = "parameter_55" + original_name = "batch_norm2d_31.w_0" shape = [192] dtype = "float32" min_val = float("0.0121464") @@ -616,6 +672,7 @@ class Program_weight_tensor_parameter_55: class Program_weight_tensor_parameter_56: name = "parameter_56" + original_name = "batch_norm2d_31.w_2" shape = [192] dtype = "float32" min_val = float("5.60519e-45") @@ -627,6 +684,7 @@ class Program_weight_tensor_parameter_56: class Program_weight_tensor_parameter_57: name = "parameter_57" + original_name = "batch_norm2d_31.w_1" shape = [192] dtype = "float32" min_val = float("-0.696058") @@ -638,6 +696,7 @@ class Program_weight_tensor_parameter_57: class Program_weight_tensor_parameter_58: name = "parameter_58" + original_name = "conv2d_34.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.350282") @@ -649,6 +708,7 @@ class Program_weight_tensor_parameter_58: class Program_weight_tensor_parameter_59: name = "parameter_59" + original_name = "batch_norm2d_30.b_0" shape = [192] dtype = "float32" min_val = float("-3.07083") @@ -660,6 +720,7 @@ class Program_weight_tensor_parameter_59: class Program_weight_tensor_parameter_60: name = "parameter_60" + original_name = "batch_norm2d_30.w_0" shape = [192] dtype = "float32" min_val = float("0.0725845") @@ -671,6 +732,7 @@ class Program_weight_tensor_parameter_60: class Program_weight_tensor_parameter_61: name = "parameter_61" + original_name = "batch_norm2d_30.w_2" shape = [192] dtype = "float32" min_val = float("5.60519e-45") @@ -682,6 +744,7 @@ class Program_weight_tensor_parameter_61: class Program_weight_tensor_parameter_62: name = "parameter_62" + original_name = "batch_norm2d_30.w_1" shape = [192] dtype = "float32" min_val = float("-0.977669") @@ -693,6 +756,7 @@ class Program_weight_tensor_parameter_62: class Program_weight_tensor_parameter_63: name = "parameter_63" + original_name = "conv2d_33.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.430979") @@ -704,6 +768,7 @@ class Program_weight_tensor_parameter_63: class Program_weight_tensor_parameter_64: name = "parameter_64" + original_name = "batch_norm2d_29.b_0" shape = [192] dtype = "float32" min_val = float("-3.22475") @@ -715,6 +780,7 @@ class Program_weight_tensor_parameter_64: class Program_weight_tensor_parameter_65: name = "parameter_65" + original_name = "batch_norm2d_29.w_0" shape = [192] dtype = "float32" min_val = float("2.65214e-05") @@ -726,6 +792,7 @@ class Program_weight_tensor_parameter_65: class Program_weight_tensor_parameter_66: name = "parameter_66" + original_name = "batch_norm2d_29.w_2" shape = [192] dtype = "float32" min_val = float("5.60519e-45") @@ -737,6 +804,7 @@ class Program_weight_tensor_parameter_66: class Program_weight_tensor_parameter_67: name = "parameter_67" + original_name = "batch_norm2d_29.w_1" shape = [192] dtype = "float32" min_val = float("-0.815296") @@ -748,6 +816,7 @@ class Program_weight_tensor_parameter_67: class Program_weight_tensor_parameter_68: name = "parameter_68" + original_name = "conv2d_32.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.443708") @@ -759,6 +828,7 @@ class Program_weight_tensor_parameter_68: class Program_weight_tensor_parameter_69: name = "parameter_69" + original_name = "batch_norm2d_28.b_0" shape = [192] dtype = "float32" min_val = float("-4.90014") @@ -770,6 +840,7 @@ class Program_weight_tensor_parameter_69: class Program_weight_tensor_parameter_70: name = "parameter_70" + original_name = "batch_norm2d_28.w_0" shape = [192] dtype = "float32" min_val = float("0.154663") @@ -781,6 +852,7 @@ class Program_weight_tensor_parameter_70: class Program_weight_tensor_parameter_71: name = "parameter_71" + original_name = "batch_norm2d_28.w_2" shape = [192] dtype = "float32" min_val = float("5.60519e-45") @@ -792,6 +864,7 @@ class Program_weight_tensor_parameter_71: class Program_weight_tensor_parameter_72: name = "parameter_72" + original_name = "batch_norm2d_28.w_1" shape = [192] dtype = "float32" min_val = float("-0.528588") @@ -803,6 +876,7 @@ class Program_weight_tensor_parameter_72: class Program_weight_tensor_parameter_73: name = "parameter_73" + original_name = "conv2d_31.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.296865") @@ -814,6 +888,7 @@ class Program_weight_tensor_parameter_73: class Program_weight_tensor_parameter_74: name = "parameter_74" + original_name = "batch_norm2d_27.b_0" shape = [192] dtype = "float32" min_val = float("-3.20422") @@ -825,6 +900,7 @@ class Program_weight_tensor_parameter_74: class Program_weight_tensor_parameter_75: name = "parameter_75" + original_name = "batch_norm2d_27.w_0" shape = [192] dtype = "float32" min_val = float("0.141583") @@ -836,6 +912,7 @@ class Program_weight_tensor_parameter_75: class Program_weight_tensor_parameter_76: name = "parameter_76" + original_name = "batch_norm2d_27.w_2" shape = [192] dtype = "float32" min_val = float("5.60519e-45") @@ -847,6 +924,7 @@ class Program_weight_tensor_parameter_76: class Program_weight_tensor_parameter_77: name = "parameter_77" + original_name = "batch_norm2d_27.w_1" shape = [192] dtype = "float32" min_val = float("-0.723996") @@ -858,6 +936,7 @@ class Program_weight_tensor_parameter_77: class Program_weight_tensor_parameter_78: name = "parameter_78" + original_name = "conv2d_30.w_0" shape = [192, 768, 3, 3] dtype = "float32" min_val = float("-0.601036") @@ -869,6 +948,7 @@ class Program_weight_tensor_parameter_78: class Program_weight_tensor_parameter_79: name = "parameter_79" + original_name = "conv2d_29.b_0" shape = [768] dtype = "float32" min_val = float("-0.0266875") @@ -880,6 +960,7 @@ class Program_weight_tensor_parameter_79: class Program_weight_tensor_parameter_80: name = "parameter_80" + original_name = "conv2d_29.w_0" shape = [768, 768, 1, 1] dtype = "float32" min_val = float("-0.720064") @@ -891,6 +972,7 @@ class Program_weight_tensor_parameter_80: class Program_weight_tensor_parameter_81: name = "parameter_81" + original_name = "batch_norm2d_26.b_0" shape = [768] dtype = "float32" min_val = float("-7.00944") @@ -902,6 +984,7 @@ class Program_weight_tensor_parameter_81: class Program_weight_tensor_parameter_82: name = "parameter_82" + original_name = "batch_norm2d_26.w_0" shape = [768] dtype = "float32" min_val = float("0.01545") @@ -913,6 +996,7 @@ class Program_weight_tensor_parameter_82: class Program_weight_tensor_parameter_83: name = "parameter_83" + original_name = "batch_norm2d_26.w_2" shape = [768] dtype = "float32" min_val = float("5.60519e-45") @@ -924,6 +1008,7 @@ class Program_weight_tensor_parameter_83: class Program_weight_tensor_parameter_84: name = "parameter_84" + original_name = "batch_norm2d_26.w_1" shape = [768] dtype = "float32" min_val = float("-1.437") @@ -935,6 +1020,7 @@ class Program_weight_tensor_parameter_84: class Program_weight_tensor_parameter_85: name = "parameter_85" + original_name = "conv2d_28.w_0" shape = [768, 1664, 1, 1] dtype = "float32" min_val = float("-0.375854") @@ -946,6 +1032,7 @@ class Program_weight_tensor_parameter_85: class Program_weight_tensor_parameter_86: name = "parameter_86" + original_name = "batch_norm2d_25.b_0" shape = [192] dtype = "float32" min_val = float("-4.71894") @@ -957,6 +1044,7 @@ class Program_weight_tensor_parameter_86: class Program_weight_tensor_parameter_87: name = "parameter_87" + original_name = "batch_norm2d_25.w_0" shape = [192] dtype = "float32" min_val = float("0.178074") @@ -968,6 +1056,7 @@ class Program_weight_tensor_parameter_87: class Program_weight_tensor_parameter_88: name = "parameter_88" + original_name = "batch_norm2d_25.w_2" shape = [192] dtype = "float32" min_val = float("5.60519e-45") @@ -979,6 +1068,7 @@ class Program_weight_tensor_parameter_88: class Program_weight_tensor_parameter_89: name = "parameter_89" + original_name = "batch_norm2d_25.w_1" shape = [192] dtype = "float32" min_val = float("-0.467628") @@ -990,6 +1080,7 @@ class Program_weight_tensor_parameter_89: class Program_weight_tensor_parameter_90: name = "parameter_90" + original_name = "conv2d_27.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.218353") @@ -1001,6 +1092,7 @@ class Program_weight_tensor_parameter_90: class Program_weight_tensor_parameter_91: name = "parameter_91" + original_name = "batch_norm2d_24.b_0" shape = [192] dtype = "float32" min_val = float("-1.89299") @@ -1012,6 +1104,7 @@ class Program_weight_tensor_parameter_91: class Program_weight_tensor_parameter_92: name = "parameter_92" + original_name = "batch_norm2d_24.w_0" shape = [192] dtype = "float32" min_val = float("0.0900117") @@ -1023,6 +1116,7 @@ class Program_weight_tensor_parameter_92: class Program_weight_tensor_parameter_93: name = "parameter_93" + original_name = "batch_norm2d_24.w_2" shape = [192] dtype = "float32" min_val = float("5.60519e-45") @@ -1034,6 +1128,7 @@ class Program_weight_tensor_parameter_93: class Program_weight_tensor_parameter_94: name = "parameter_94" + original_name = "batch_norm2d_24.w_1" shape = [192] dtype = "float32" min_val = float("-1.1023") @@ -1045,6 +1140,7 @@ class Program_weight_tensor_parameter_94: class Program_weight_tensor_parameter_95: name = "parameter_95" + original_name = "conv2d_26.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.239786") @@ -1056,6 +1152,7 @@ class Program_weight_tensor_parameter_95: class Program_weight_tensor_parameter_96: name = "parameter_96" + original_name = "batch_norm2d_23.b_0" shape = [192] dtype = "float32" min_val = float("-2.39657") @@ -1067,6 +1164,7 @@ class Program_weight_tensor_parameter_96: class Program_weight_tensor_parameter_97: name = "parameter_97" + original_name = "batch_norm2d_23.w_0" shape = [192] dtype = "float32" min_val = float("0.183894") @@ -1078,6 +1176,7 @@ class Program_weight_tensor_parameter_97: class Program_weight_tensor_parameter_98: name = "parameter_98" + original_name = "batch_norm2d_23.w_2" shape = [192] dtype = "float32" min_val = float("5.60519e-45") @@ -1089,6 +1188,7 @@ class Program_weight_tensor_parameter_98: class Program_weight_tensor_parameter_99: name = "parameter_99" + original_name = "batch_norm2d_23.w_1" shape = [192] dtype = "float32" min_val = float("-0.88404") @@ -1100,6 +1200,7 @@ class Program_weight_tensor_parameter_99: class Program_weight_tensor_parameter_100: name = "parameter_100" + original_name = "conv2d_25.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.316945") @@ -1111,6 +1212,7 @@ class Program_weight_tensor_parameter_100: class Program_weight_tensor_parameter_101: name = "parameter_101" + original_name = "batch_norm2d_22.b_0" shape = [192] dtype = "float32" min_val = float("-2.8033") @@ -1122,6 +1224,7 @@ class Program_weight_tensor_parameter_101: class Program_weight_tensor_parameter_102: name = "parameter_102" + original_name = "batch_norm2d_22.w_0" shape = [192] dtype = "float32" min_val = float("0.0752399") @@ -1133,6 +1236,7 @@ class Program_weight_tensor_parameter_102: class Program_weight_tensor_parameter_103: name = "parameter_103" + original_name = "batch_norm2d_22.w_2" shape = [192] dtype = "float32" min_val = float("5.60519e-45") @@ -1144,6 +1248,7 @@ class Program_weight_tensor_parameter_103: class Program_weight_tensor_parameter_104: name = "parameter_104" + original_name = "batch_norm2d_22.w_1" shape = [192] dtype = "float32" min_val = float("-1.41745") @@ -1155,6 +1260,7 @@ class Program_weight_tensor_parameter_104: class Program_weight_tensor_parameter_105: name = "parameter_105" + original_name = "conv2d_24.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.23435") @@ -1166,6 +1272,7 @@ class Program_weight_tensor_parameter_105: class Program_weight_tensor_parameter_106: name = "parameter_106" + original_name = "batch_norm2d_21.b_0" shape = [192] dtype = "float32" min_val = float("-2.90627") @@ -1177,6 +1284,7 @@ class Program_weight_tensor_parameter_106: class Program_weight_tensor_parameter_107: name = "parameter_107" + original_name = "batch_norm2d_21.w_0" shape = [192] dtype = "float32" min_val = float("0.226439") @@ -1188,6 +1296,7 @@ class Program_weight_tensor_parameter_107: class Program_weight_tensor_parameter_108: name = "parameter_108" + original_name = "batch_norm2d_21.w_2" shape = [192] dtype = "float32" min_val = float("5.60519e-45") @@ -1199,6 +1308,7 @@ class Program_weight_tensor_parameter_108: class Program_weight_tensor_parameter_109: name = "parameter_109" + original_name = "batch_norm2d_21.w_1" shape = [192] dtype = "float32" min_val = float("-1.28534") @@ -1210,6 +1320,7 @@ class Program_weight_tensor_parameter_109: class Program_weight_tensor_parameter_110: name = "parameter_110" + original_name = "conv2d_23.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.269045") @@ -1221,6 +1332,7 @@ class Program_weight_tensor_parameter_110: class Program_weight_tensor_parameter_111: name = "parameter_111" + original_name = "batch_norm2d_20.b_0" shape = [192] dtype = "float32" min_val = float("-3.15092") @@ -1232,6 +1344,7 @@ class Program_weight_tensor_parameter_111: class Program_weight_tensor_parameter_112: name = "parameter_112" + original_name = "batch_norm2d_20.w_0" shape = [192] dtype = "float32" min_val = float("0.0845418") @@ -1243,6 +1356,7 @@ class Program_weight_tensor_parameter_112: class Program_weight_tensor_parameter_113: name = "parameter_113" + original_name = "batch_norm2d_20.w_2" shape = [192] dtype = "float32" min_val = float("5.60519e-45") @@ -1254,6 +1368,7 @@ class Program_weight_tensor_parameter_113: class Program_weight_tensor_parameter_114: name = "parameter_114" + original_name = "batch_norm2d_20.w_1" shape = [192] dtype = "float32" min_val = float("-2.38707") @@ -1265,6 +1380,7 @@ class Program_weight_tensor_parameter_114: class Program_weight_tensor_parameter_115: name = "parameter_115" + original_name = "conv2d_22.w_0" shape = [192, 512, 3, 3] dtype = "float32" min_val = float("-0.287364") @@ -1276,6 +1392,7 @@ class Program_weight_tensor_parameter_115: class Program_weight_tensor_parameter_116: name = "parameter_116" + original_name = "batch_norm2d_19.b_0" shape = [512] dtype = "float32" min_val = float("-1.15487") @@ -1287,6 +1404,7 @@ class Program_weight_tensor_parameter_116: class Program_weight_tensor_parameter_117: name = "parameter_117" + original_name = "batch_norm2d_19.w_0" shape = [512] dtype = "float32" min_val = float("-0.0100505") @@ -1298,6 +1416,7 @@ class Program_weight_tensor_parameter_117: class Program_weight_tensor_parameter_118: name = "parameter_118" + original_name = "batch_norm2d_19.w_2" shape = [512] dtype = "float32" min_val = float("5.60519e-45") @@ -1309,6 +1428,7 @@ class Program_weight_tensor_parameter_118: class Program_weight_tensor_parameter_119: name = "parameter_119" + original_name = "batch_norm2d_19.w_1" shape = [512] dtype = "float32" min_val = float("-0.529125") @@ -1320,6 +1440,7 @@ class Program_weight_tensor_parameter_119: class Program_weight_tensor_parameter_120: name = "parameter_120" + original_name = "conv2d_21.w_0" shape = [512, 1, 3, 3] dtype = "float32" min_val = float("-0.309818") @@ -1331,6 +1452,7 @@ class Program_weight_tensor_parameter_120: class Program_weight_tensor_parameter_121: name = "parameter_121" + original_name = "conv2d_20.b_0" shape = [512] dtype = "float32" min_val = float("-0.0344602") @@ -1342,6 +1464,7 @@ class Program_weight_tensor_parameter_121: class Program_weight_tensor_parameter_122: name = "parameter_122" + original_name = "conv2d_20.w_0" shape = [512, 512, 1, 1] dtype = "float32" min_val = float("-0.666568") @@ -1353,6 +1476,7 @@ class Program_weight_tensor_parameter_122: class Program_weight_tensor_parameter_123: name = "parameter_123" + original_name = "batch_norm2d_18.b_0" shape = [512] dtype = "float32" min_val = float("-10.3485") @@ -1364,6 +1488,7 @@ class Program_weight_tensor_parameter_123: class Program_weight_tensor_parameter_124: name = "parameter_124" + original_name = "batch_norm2d_18.w_0" shape = [512] dtype = "float32" min_val = float("0.0705023") @@ -1375,6 +1500,7 @@ class Program_weight_tensor_parameter_124: class Program_weight_tensor_parameter_125: name = "parameter_125" + original_name = "batch_norm2d_18.w_2" shape = [512] dtype = "float32" min_val = float("5.60519e-45") @@ -1386,6 +1512,7 @@ class Program_weight_tensor_parameter_125: class Program_weight_tensor_parameter_126: name = "parameter_126" + original_name = "batch_norm2d_18.w_1" shape = [512] dtype = "float32" min_val = float("-1.29794") @@ -1397,6 +1524,7 @@ class Program_weight_tensor_parameter_126: class Program_weight_tensor_parameter_127: name = "parameter_127" + original_name = "conv2d_19.w_0" shape = [512, 1216, 1, 1] dtype = "float32" min_val = float("-0.28685") @@ -1408,6 +1536,7 @@ class Program_weight_tensor_parameter_127: class Program_weight_tensor_parameter_128: name = "parameter_128" + original_name = "batch_norm2d_17.b_0" shape = [160] dtype = "float32" min_val = float("-3.65985") @@ -1419,6 +1548,7 @@ class Program_weight_tensor_parameter_128: class Program_weight_tensor_parameter_129: name = "parameter_129" + original_name = "batch_norm2d_17.w_0" shape = [160] dtype = "float32" min_val = float("0.142092") @@ -1430,6 +1560,7 @@ class Program_weight_tensor_parameter_129: class Program_weight_tensor_parameter_130: name = "parameter_130" + original_name = "batch_norm2d_17.w_2" shape = [160] dtype = "float32" min_val = float("5.60519e-45") @@ -1441,6 +1572,7 @@ class Program_weight_tensor_parameter_130: class Program_weight_tensor_parameter_131: name = "parameter_131" + original_name = "batch_norm2d_17.w_1" shape = [160] dtype = "float32" min_val = float("-0.761527") @@ -1452,6 +1584,7 @@ class Program_weight_tensor_parameter_131: class Program_weight_tensor_parameter_132: name = "parameter_132" + original_name = "conv2d_18.w_0" shape = [160, 160, 3, 3] dtype = "float32" min_val = float("-0.531788") @@ -1463,6 +1596,7 @@ class Program_weight_tensor_parameter_132: class Program_weight_tensor_parameter_133: name = "parameter_133" + original_name = "batch_norm2d_16.b_0" shape = [160] dtype = "float32" min_val = float("-2.98266") @@ -1474,6 +1608,7 @@ class Program_weight_tensor_parameter_133: class Program_weight_tensor_parameter_134: name = "parameter_134" + original_name = "batch_norm2d_16.w_0" shape = [160] dtype = "float32" min_val = float("0.00086597") @@ -1485,6 +1620,7 @@ class Program_weight_tensor_parameter_134: class Program_weight_tensor_parameter_135: name = "parameter_135" + original_name = "batch_norm2d_16.w_2" shape = [160] dtype = "float32" min_val = float("5.60519e-45") @@ -1496,6 +1632,7 @@ class Program_weight_tensor_parameter_135: class Program_weight_tensor_parameter_136: name = "parameter_136" + original_name = "batch_norm2d_16.w_1" shape = [160] dtype = "float32" min_val = float("-1.098") @@ -1507,6 +1644,7 @@ class Program_weight_tensor_parameter_136: class Program_weight_tensor_parameter_137: name = "parameter_137" + original_name = "conv2d_17.w_0" shape = [160, 160, 3, 3] dtype = "float32" min_val = float("-0.439071") @@ -1518,6 +1656,7 @@ class Program_weight_tensor_parameter_137: class Program_weight_tensor_parameter_138: name = "parameter_138" + original_name = "batch_norm2d_15.b_0" shape = [160] dtype = "float32" min_val = float("-2.53758") @@ -1529,6 +1668,7 @@ class Program_weight_tensor_parameter_138: class Program_weight_tensor_parameter_139: name = "parameter_139" + original_name = "batch_norm2d_15.w_0" shape = [160] dtype = "float32" min_val = float("0.178299") @@ -1540,6 +1680,7 @@ class Program_weight_tensor_parameter_139: class Program_weight_tensor_parameter_140: name = "parameter_140" + original_name = "batch_norm2d_15.w_2" shape = [160] dtype = "float32" min_val = float("5.60519e-45") @@ -1551,6 +1692,7 @@ class Program_weight_tensor_parameter_140: class Program_weight_tensor_parameter_141: name = "parameter_141" + original_name = "batch_norm2d_15.w_1" shape = [160] dtype = "float32" min_val = float("-1.16357") @@ -1562,6 +1704,7 @@ class Program_weight_tensor_parameter_141: class Program_weight_tensor_parameter_142: name = "parameter_142" + original_name = "conv2d_16.w_0" shape = [160, 160, 3, 3] dtype = "float32" min_val = float("-0.227142") @@ -1573,6 +1716,7 @@ class Program_weight_tensor_parameter_142: class Program_weight_tensor_parameter_143: name = "parameter_143" + original_name = "batch_norm2d_14.b_0" shape = [160] dtype = "float32" min_val = float("-2.52695") @@ -1584,6 +1728,7 @@ class Program_weight_tensor_parameter_143: class Program_weight_tensor_parameter_144: name = "parameter_144" + original_name = "batch_norm2d_14.w_0" shape = [160] dtype = "float32" min_val = float("0.174552") @@ -1595,6 +1740,7 @@ class Program_weight_tensor_parameter_144: class Program_weight_tensor_parameter_145: name = "parameter_145" + original_name = "batch_norm2d_14.w_2" shape = [160] dtype = "float32" min_val = float("5.60519e-45") @@ -1606,6 +1752,7 @@ class Program_weight_tensor_parameter_145: class Program_weight_tensor_parameter_146: name = "parameter_146" + original_name = "batch_norm2d_14.w_1" shape = [160] dtype = "float32" min_val = float("-1.5033") @@ -1617,6 +1764,7 @@ class Program_weight_tensor_parameter_146: class Program_weight_tensor_parameter_147: name = "parameter_147" + original_name = "conv2d_15.w_0" shape = [160, 160, 3, 3] dtype = "float32" min_val = float("-0.193468") @@ -1628,6 +1776,7 @@ class Program_weight_tensor_parameter_147: class Program_weight_tensor_parameter_148: name = "parameter_148" + original_name = "batch_norm2d_13.b_0" shape = [160] dtype = "float32" min_val = float("-2.1706") @@ -1639,6 +1788,7 @@ class Program_weight_tensor_parameter_148: class Program_weight_tensor_parameter_149: name = "parameter_149" + original_name = "batch_norm2d_13.w_0" shape = [160] dtype = "float32" min_val = float("0.177653") @@ -1650,6 +1800,7 @@ class Program_weight_tensor_parameter_149: class Program_weight_tensor_parameter_150: name = "parameter_150" + original_name = "batch_norm2d_13.w_2" shape = [160] dtype = "float32" min_val = float("5.60519e-45") @@ -1661,6 +1812,7 @@ class Program_weight_tensor_parameter_150: class Program_weight_tensor_parameter_151: name = "parameter_151" + original_name = "batch_norm2d_13.w_1" shape = [160] dtype = "float32" min_val = float("-1.04302") @@ -1672,6 +1824,7 @@ class Program_weight_tensor_parameter_151: class Program_weight_tensor_parameter_152: name = "parameter_152" + original_name = "conv2d_14.w_0" shape = [160, 160, 3, 3] dtype = "float32" min_val = float("-0.500709") @@ -1683,6 +1836,7 @@ class Program_weight_tensor_parameter_152: class Program_weight_tensor_parameter_153: name = "parameter_153" + original_name = "batch_norm2d_12.b_0" shape = [160] dtype = "float32" min_val = float("-2.24115") @@ -1694,6 +1848,7 @@ class Program_weight_tensor_parameter_153: class Program_weight_tensor_parameter_154: name = "parameter_154" + original_name = "batch_norm2d_12.w_0" shape = [160] dtype = "float32" min_val = float("0.132613") @@ -1705,6 +1860,7 @@ class Program_weight_tensor_parameter_154: class Program_weight_tensor_parameter_155: name = "parameter_155" + original_name = "batch_norm2d_12.w_2" shape = [160] dtype = "float32" min_val = float("5.60519e-45") @@ -1716,6 +1872,7 @@ class Program_weight_tensor_parameter_155: class Program_weight_tensor_parameter_156: name = "parameter_156" + original_name = "batch_norm2d_12.w_1" shape = [160] dtype = "float32" min_val = float("-1.80896") @@ -1727,6 +1884,7 @@ class Program_weight_tensor_parameter_156: class Program_weight_tensor_parameter_157: name = "parameter_157" + original_name = "conv2d_13.w_0" shape = [160, 256, 3, 3] dtype = "float32" min_val = float("-0.372421") @@ -1738,6 +1896,7 @@ class Program_weight_tensor_parameter_157: class Program_weight_tensor_parameter_158: name = "parameter_158" + original_name = "batch_norm2d_11.b_0" shape = [256] dtype = "float32" min_val = float("-0.741466") @@ -1749,6 +1908,7 @@ class Program_weight_tensor_parameter_158: class Program_weight_tensor_parameter_159: name = "parameter_159" + original_name = "batch_norm2d_11.w_0" shape = [256] dtype = "float32" min_val = float("-0.0193116") @@ -1760,6 +1920,7 @@ class Program_weight_tensor_parameter_159: class Program_weight_tensor_parameter_160: name = "parameter_160" + original_name = "batch_norm2d_11.w_2" shape = [256] dtype = "float32" min_val = float("5.60519e-45") @@ -1771,6 +1932,7 @@ class Program_weight_tensor_parameter_160: class Program_weight_tensor_parameter_161: name = "parameter_161" + original_name = "batch_norm2d_11.w_1" shape = [256] dtype = "float32" min_val = float("-0.550272") @@ -1782,6 +1944,7 @@ class Program_weight_tensor_parameter_161: class Program_weight_tensor_parameter_162: name = "parameter_162" + original_name = "conv2d_12.w_0" shape = [256, 1, 3, 3] dtype = "float32" min_val = float("-0.430381") @@ -1793,6 +1956,7 @@ class Program_weight_tensor_parameter_162: class Program_weight_tensor_parameter_163: name = "parameter_163" + original_name = "conv2d_11.b_0" shape = [256] dtype = "float32" min_val = float("-0.050004") @@ -1804,6 +1968,7 @@ class Program_weight_tensor_parameter_163: class Program_weight_tensor_parameter_164: name = "parameter_164" + original_name = "conv2d_11.w_0" shape = [256, 256, 1, 1] dtype = "float32" min_val = float("-1.28239") @@ -1815,6 +1980,7 @@ class Program_weight_tensor_parameter_164: class Program_weight_tensor_parameter_165: name = "parameter_165" + original_name = "batch_norm2d_10.b_0" shape = [256] dtype = "float32" min_val = float("-5.85304") @@ -1826,6 +1992,7 @@ class Program_weight_tensor_parameter_165: class Program_weight_tensor_parameter_166: name = "parameter_166" + original_name = "batch_norm2d_10.w_0" shape = [256] dtype = "float32" min_val = float("0.0357678") @@ -1837,6 +2004,7 @@ class Program_weight_tensor_parameter_166: class Program_weight_tensor_parameter_167: name = "parameter_167" + original_name = "batch_norm2d_10.w_2" shape = [256] dtype = "float32" min_val = float("5.60519e-45") @@ -1848,6 +2016,7 @@ class Program_weight_tensor_parameter_167: class Program_weight_tensor_parameter_168: name = "parameter_168" + original_name = "batch_norm2d_10.w_1" shape = [256] dtype = "float32" min_val = float("-1.22782") @@ -1859,6 +2028,7 @@ class Program_weight_tensor_parameter_168: class Program_weight_tensor_parameter_169: name = "parameter_169" + original_name = "conv2d_10.w_0" shape = [256, 896, 1, 1] dtype = "float32" min_val = float("-0.599782") @@ -1870,6 +2040,7 @@ class Program_weight_tensor_parameter_169: class Program_weight_tensor_parameter_170: name = "parameter_170" + original_name = "batch_norm2d_9.b_0" shape = [128] dtype = "float32" min_val = float("-4.31096") @@ -1881,6 +2052,7 @@ class Program_weight_tensor_parameter_170: class Program_weight_tensor_parameter_171: name = "parameter_171" + original_name = "batch_norm2d_9.w_0" shape = [128] dtype = "float32" min_val = float("0.16275") @@ -1892,6 +2064,7 @@ class Program_weight_tensor_parameter_171: class Program_weight_tensor_parameter_172: name = "parameter_172" + original_name = "batch_norm2d_9.w_2" shape = [128] dtype = "float32" min_val = float("5.60519e-45") @@ -1903,6 +2076,7 @@ class Program_weight_tensor_parameter_172: class Program_weight_tensor_parameter_173: name = "parameter_173" + original_name = "batch_norm2d_9.w_1" shape = [128] dtype = "float32" min_val = float("-0.662831") @@ -1914,6 +2088,7 @@ class Program_weight_tensor_parameter_173: class Program_weight_tensor_parameter_174: name = "parameter_174" + original_name = "conv2d_9.w_0" shape = [128, 128, 3, 3] dtype = "float32" min_val = float("-0.236725") @@ -1925,6 +2100,7 @@ class Program_weight_tensor_parameter_174: class Program_weight_tensor_parameter_175: name = "parameter_175" + original_name = "batch_norm2d_8.b_0" shape = [128] dtype = "float32" min_val = float("-3.02228") @@ -1936,6 +2112,7 @@ class Program_weight_tensor_parameter_175: class Program_weight_tensor_parameter_176: name = "parameter_176" + original_name = "batch_norm2d_8.w_0" shape = [128] dtype = "float32" min_val = float("0.174719") @@ -1947,6 +2124,7 @@ class Program_weight_tensor_parameter_176: class Program_weight_tensor_parameter_177: name = "parameter_177" + original_name = "batch_norm2d_8.w_2" shape = [128] dtype = "float32" min_val = float("5.60519e-45") @@ -1958,6 +2136,7 @@ class Program_weight_tensor_parameter_177: class Program_weight_tensor_parameter_178: name = "parameter_178" + original_name = "batch_norm2d_8.w_1" shape = [128] dtype = "float32" min_val = float("-1.32987") @@ -1969,6 +2148,7 @@ class Program_weight_tensor_parameter_178: class Program_weight_tensor_parameter_179: name = "parameter_179" + original_name = "conv2d_8.w_0" shape = [128, 128, 3, 3] dtype = "float32" min_val = float("-0.250536") @@ -1980,6 +2160,7 @@ class Program_weight_tensor_parameter_179: class Program_weight_tensor_parameter_180: name = "parameter_180" + original_name = "batch_norm2d_7.b_0" shape = [128] dtype = "float32" min_val = float("-2.11769") @@ -1991,6 +2172,7 @@ class Program_weight_tensor_parameter_180: class Program_weight_tensor_parameter_181: name = "parameter_181" + original_name = "batch_norm2d_7.w_0" shape = [128] dtype = "float32" min_val = float("0.18906") @@ -2002,6 +2184,7 @@ class Program_weight_tensor_parameter_181: class Program_weight_tensor_parameter_182: name = "parameter_182" + original_name = "batch_norm2d_7.w_2" shape = [128] dtype = "float32" min_val = float("5.60519e-45") @@ -2013,6 +2196,7 @@ class Program_weight_tensor_parameter_182: class Program_weight_tensor_parameter_183: name = "parameter_183" + original_name = "batch_norm2d_7.w_1" shape = [128] dtype = "float32" min_val = float("-1.10771") @@ -2024,6 +2208,7 @@ class Program_weight_tensor_parameter_183: class Program_weight_tensor_parameter_184: name = "parameter_184" + original_name = "conv2d_7.w_0" shape = [128, 128, 3, 3] dtype = "float32" min_val = float("-0.365977") @@ -2035,6 +2220,7 @@ class Program_weight_tensor_parameter_184: class Program_weight_tensor_parameter_185: name = "parameter_185" + original_name = "batch_norm2d_6.b_0" shape = [128] dtype = "float32" min_val = float("-1.87659") @@ -2046,6 +2232,7 @@ class Program_weight_tensor_parameter_185: class Program_weight_tensor_parameter_186: name = "parameter_186" + original_name = "batch_norm2d_6.w_0" shape = [128] dtype = "float32" min_val = float("0.160651") @@ -2057,6 +2244,7 @@ class Program_weight_tensor_parameter_186: class Program_weight_tensor_parameter_187: name = "parameter_187" + original_name = "batch_norm2d_6.w_2" shape = [128] dtype = "float32" min_val = float("5.60519e-45") @@ -2068,6 +2256,7 @@ class Program_weight_tensor_parameter_187: class Program_weight_tensor_parameter_188: name = "parameter_188" + original_name = "batch_norm2d_6.w_1" shape = [128] dtype = "float32" min_val = float("-1.84041") @@ -2079,6 +2268,7 @@ class Program_weight_tensor_parameter_188: class Program_weight_tensor_parameter_189: name = "parameter_189" + original_name = "conv2d_6.w_0" shape = [128, 128, 3, 3] dtype = "float32" min_val = float("-0.269367") @@ -2090,6 +2280,7 @@ class Program_weight_tensor_parameter_189: class Program_weight_tensor_parameter_190: name = "parameter_190" + original_name = "batch_norm2d_5.b_0" shape = [128] dtype = "float32" min_val = float("-1.79534") @@ -2101,6 +2292,7 @@ class Program_weight_tensor_parameter_190: class Program_weight_tensor_parameter_191: name = "parameter_191" + original_name = "batch_norm2d_5.w_0" shape = [128] dtype = "float32" min_val = float("0.119348") @@ -2112,6 +2304,7 @@ class Program_weight_tensor_parameter_191: class Program_weight_tensor_parameter_192: name = "parameter_192" + original_name = "batch_norm2d_5.w_2" shape = [128] dtype = "float32" min_val = float("5.60519e-45") @@ -2123,6 +2316,7 @@ class Program_weight_tensor_parameter_192: class Program_weight_tensor_parameter_193: name = "parameter_193" + original_name = "batch_norm2d_5.w_1" shape = [128] dtype = "float32" min_val = float("-1.30478") @@ -2134,6 +2328,7 @@ class Program_weight_tensor_parameter_193: class Program_weight_tensor_parameter_194: name = "parameter_194" + original_name = "conv2d_5.w_0" shape = [128, 128, 3, 3] dtype = "float32" min_val = float("-0.265752") @@ -2145,6 +2340,7 @@ class Program_weight_tensor_parameter_194: class Program_weight_tensor_parameter_195: name = "parameter_195" + original_name = "batch_norm2d_4.b_0" shape = [128] dtype = "float32" min_val = float("-1.66729") @@ -2156,6 +2352,7 @@ class Program_weight_tensor_parameter_195: class Program_weight_tensor_parameter_196: name = "parameter_196" + original_name = "batch_norm2d_4.w_0" shape = [128] dtype = "float32" min_val = float("0.0633489") @@ -2167,6 +2364,7 @@ class Program_weight_tensor_parameter_196: class Program_weight_tensor_parameter_197: name = "parameter_197" + original_name = "batch_norm2d_4.w_2" shape = [128] dtype = "float32" min_val = float("5.60519e-45") @@ -2178,6 +2376,7 @@ class Program_weight_tensor_parameter_197: class Program_weight_tensor_parameter_198: name = "parameter_198" + original_name = "batch_norm2d_4.w_1" shape = [128] dtype = "float32" min_val = float("-2.24189") @@ -2189,6 +2388,7 @@ class Program_weight_tensor_parameter_198: class Program_weight_tensor_parameter_199: name = "parameter_199" + original_name = "conv2d_4.w_0" shape = [128, 128, 3, 3] dtype = "float32" min_val = float("-0.386929") @@ -2200,6 +2400,7 @@ class Program_weight_tensor_parameter_199: class Program_weight_tensor_parameter_200: name = "parameter_200" + original_name = "batch_norm2d_3.b_0" shape = [128] dtype = "float32" min_val = float("-1.1244") @@ -2211,6 +2412,7 @@ class Program_weight_tensor_parameter_200: class Program_weight_tensor_parameter_201: name = "parameter_201" + original_name = "batch_norm2d_3.w_0" shape = [128] dtype = "float32" min_val = float("-0.0059161") @@ -2222,6 +2424,7 @@ class Program_weight_tensor_parameter_201: class Program_weight_tensor_parameter_202: name = "parameter_202" + original_name = "batch_norm2d_3.w_2" shape = [128] dtype = "float32" min_val = float("5.60519e-45") @@ -2233,6 +2436,7 @@ class Program_weight_tensor_parameter_202: class Program_weight_tensor_parameter_203: name = "parameter_203" + original_name = "batch_norm2d_3.w_1" shape = [128] dtype = "float32" min_val = float("-0.401874") @@ -2244,6 +2448,7 @@ class Program_weight_tensor_parameter_203: class Program_weight_tensor_parameter_204: name = "parameter_204" + original_name = "conv2d_3.w_0" shape = [128, 1, 3, 3] dtype = "float32" min_val = float("-0.334231") @@ -2255,6 +2460,7 @@ class Program_weight_tensor_parameter_204: class Program_weight_tensor_parameter_205: name = "parameter_205" + original_name = "batch_norm2d_2.b_0" shape = [128] dtype = "float32" min_val = float("-0.948793") @@ -2266,6 +2472,7 @@ class Program_weight_tensor_parameter_205: class Program_weight_tensor_parameter_206: name = "parameter_206" + original_name = "batch_norm2d_2.w_0" shape = [128] dtype = "float32" min_val = float("-0.00931573") @@ -2277,6 +2484,7 @@ class Program_weight_tensor_parameter_206: class Program_weight_tensor_parameter_207: name = "parameter_207" + original_name = "batch_norm2d_2.w_2" shape = [128] dtype = "float32" min_val = float("5.60519e-45") @@ -2288,6 +2496,7 @@ class Program_weight_tensor_parameter_207: class Program_weight_tensor_parameter_208: name = "parameter_208" + original_name = "batch_norm2d_2.w_1" shape = [128] dtype = "float32" min_val = float("-1.81529") @@ -2299,6 +2508,7 @@ class Program_weight_tensor_parameter_208: class Program_weight_tensor_parameter_209: name = "parameter_209" + original_name = "conv2d_2.w_0" shape = [128, 64, 3, 3] dtype = "float32" min_val = float("-0.418763") @@ -2310,6 +2520,7 @@ class Program_weight_tensor_parameter_209: class Program_weight_tensor_parameter_210: name = "parameter_210" + original_name = "batch_norm2d_1.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -2319,6 +2530,7 @@ class Program_weight_tensor_parameter_210: class Program_weight_tensor_parameter_211: name = "parameter_211" + original_name = "batch_norm2d_1.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -2328,6 +2540,7 @@ class Program_weight_tensor_parameter_211: class Program_weight_tensor_parameter_212: name = "parameter_212" + original_name = "batch_norm2d_1.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -2337,6 +2550,7 @@ class Program_weight_tensor_parameter_212: class Program_weight_tensor_parameter_213: name = "parameter_213" + original_name = "batch_norm2d_1.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -2346,6 +2560,7 @@ class Program_weight_tensor_parameter_213: class Program_weight_tensor_parameter_214: name = "parameter_214" + original_name = "conv2d_1.w_0" shape = [64, 64, 3, 3] dtype = "float32" min_val = float("-0.579162") @@ -2357,6 +2572,7 @@ class Program_weight_tensor_parameter_214: class Program_weight_tensor_parameter_215: name = "parameter_215" + original_name = "batch_norm2d_0.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -2366,6 +2582,7 @@ class Program_weight_tensor_parameter_215: class Program_weight_tensor_parameter_216: name = "parameter_216" + original_name = "batch_norm2d_0.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -2375,6 +2592,7 @@ class Program_weight_tensor_parameter_216: class Program_weight_tensor_parameter_217: name = "parameter_217" + original_name = "batch_norm2d_0.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -2384,6 +2602,7 @@ class Program_weight_tensor_parameter_217: class Program_weight_tensor_parameter_218: name = "parameter_218" + original_name = "batch_norm2d_0.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -2393,6 +2612,7 @@ class Program_weight_tensor_parameter_218: class Program_weight_tensor_parameter_219: name = "parameter_219" + original_name = "conv2d_0.w_0" shape = [64, 3, 3, 3] dtype = "float32" min_val = float("-0.760448") diff --git a/paddle_samples/PaddleX/PP-OCRv4_server_rec/subgraph_1/input_meta.py b/paddle_samples/PaddleX/PP-OCRv4_server_rec/subgraph_1/input_meta.py index 822fa52d2..c862612ef 100644 --- a/paddle_samples/PaddleX/PP-OCRv4_server_rec/subgraph_1/input_meta.py +++ b/paddle_samples/PaddleX/PP-OCRv4_server_rec/subgraph_1/input_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_data_0: name = "data_0" + original_name = "var_868" shape = [5, 1024, 1, 40] dtype = "float32" max_val = float("17.9388") @@ -10,6 +11,7 @@ class Program_weight_tensor_data_0: class Program_weight_tensor_data_1: name = "data_1" + original_name = "var_870" shape = [5, 25] dtype = "int64" min_val = 0 @@ -19,6 +21,7 @@ class Program_weight_tensor_data_1: class Program_weight_tensor_data_2: name = "data_2" + original_name = "var_871" shape = [5] dtype = "int64" data = [5, 4, 14, 16, 3] @@ -26,6 +29,7 @@ class Program_weight_tensor_data_2: class Program_weight_tensor_data_3: name = "data_3" + original_name = "var_883" shape = [5000, 1, 384] dtype = "float32" min_val = float("-1.0") diff --git a/paddle_samples/PaddleX/PP-OCRv4_server_rec/subgraph_1/weight_meta.py b/paddle_samples/PaddleX/PP-OCRv4_server_rec/subgraph_1/weight_meta.py index ca8ed42a3..1c1ce2e52 100644 --- a/paddle_samples/PaddleX/PP-OCRv4_server_rec/subgraph_1/weight_meta.py +++ b/paddle_samples/PaddleX/PP-OCRv4_server_rec/subgraph_1/weight_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_parameter_0: name = "parameter_0" + original_name = "linear_38.w_0" shape = [384, 6629] dtype = "float32" min_val = float("-1.32918") @@ -11,6 +12,7 @@ class Program_weight_tensor_parameter_0: class Program_weight_tensor_parameter_1: name = "parameter_1" + original_name = "layer_norm_16.b_0" shape = [384] dtype = "float32" min_val = float("-2.00494") @@ -22,6 +24,7 @@ class Program_weight_tensor_parameter_1: class Program_weight_tensor_parameter_2: name = "parameter_2" + original_name = "layer_norm_16.w_0" shape = [384] dtype = "float32" min_val = float("-0.000245228") @@ -33,6 +36,7 @@ class Program_weight_tensor_parameter_2: class Program_weight_tensor_parameter_3: name = "parameter_3" + original_name = "linear_37.b_0" shape = [384] dtype = "float32" min_val = float("-0.0518493") @@ -44,6 +48,7 @@ class Program_weight_tensor_parameter_3: class Program_weight_tensor_parameter_4: name = "parameter_4" + original_name = "linear_37.w_0" shape = [1536, 384] dtype = "float32" min_val = float("-0.227294") @@ -55,6 +60,7 @@ class Program_weight_tensor_parameter_4: class Program_weight_tensor_parameter_5: name = "parameter_5" + original_name = "linear_36.b_0" shape = [1536] dtype = "float32" min_val = float("-0.200106") @@ -66,6 +72,7 @@ class Program_weight_tensor_parameter_5: class Program_weight_tensor_parameter_6: name = "parameter_6" + original_name = "linear_36.w_0" shape = [384, 1536] dtype = "float32" min_val = float("-0.957273") @@ -77,6 +84,7 @@ class Program_weight_tensor_parameter_6: class Program_weight_tensor_parameter_7: name = "parameter_7" + original_name = "layer_norm_15.b_0" shape = [384] dtype = "float32" min_val = float("-0.0491315") @@ -88,6 +96,7 @@ class Program_weight_tensor_parameter_7: class Program_weight_tensor_parameter_8: name = "parameter_8" + original_name = "layer_norm_15.w_0" shape = [384] dtype = "float32" min_val = float("-0.0866739") @@ -99,6 +108,7 @@ class Program_weight_tensor_parameter_8: class Program_weight_tensor_parameter_9: name = "parameter_9" + original_name = "linear_35.b_0" shape = [384] dtype = "float32" min_val = float("-0.0762894") @@ -110,6 +120,7 @@ class Program_weight_tensor_parameter_9: class Program_weight_tensor_parameter_10: name = "parameter_10" + original_name = "linear_35.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.264504") @@ -121,6 +132,7 @@ class Program_weight_tensor_parameter_10: class Program_weight_tensor_parameter_11: name = "parameter_11" + original_name = "linear_34.b_0" shape = [768] dtype = "float32" min_val = float("-0.0725435") @@ -132,6 +144,7 @@ class Program_weight_tensor_parameter_11: class Program_weight_tensor_parameter_12: name = "parameter_12" + original_name = "linear_34.w_0" shape = [384, 768] dtype = "float32" min_val = float("-0.701239") @@ -143,6 +156,7 @@ class Program_weight_tensor_parameter_12: class Program_weight_tensor_parameter_13: name = "parameter_13" + original_name = "linear_33.b_0" shape = [384] dtype = "float32" min_val = float("-0.176879") @@ -154,6 +168,7 @@ class Program_weight_tensor_parameter_13: class Program_weight_tensor_parameter_14: name = "parameter_14" + original_name = "linear_33.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.619542") @@ -165,6 +180,7 @@ class Program_weight_tensor_parameter_14: class Program_weight_tensor_parameter_15: name = "parameter_15" + original_name = "layer_norm_14.b_0" shape = [384] dtype = "float32" min_val = float("-1.64474") @@ -176,6 +192,7 @@ class Program_weight_tensor_parameter_15: class Program_weight_tensor_parameter_16: name = "parameter_16" + original_name = "layer_norm_14.w_0" shape = [384] dtype = "float32" min_val = float("-0.291678") @@ -187,6 +204,7 @@ class Program_weight_tensor_parameter_16: class Program_weight_tensor_parameter_17: name = "parameter_17" + original_name = "linear_32.b_0" shape = [384] dtype = "float32" min_val = float("-0.0726199") @@ -198,6 +216,7 @@ class Program_weight_tensor_parameter_17: class Program_weight_tensor_parameter_18: name = "parameter_18" + original_name = "linear_32.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.310778") @@ -209,6 +228,7 @@ class Program_weight_tensor_parameter_18: class Program_weight_tensor_parameter_19: name = "parameter_19" + original_name = "linear_31.b_0" shape = [1152] dtype = "float32" min_val = float("-0.275863") @@ -220,6 +240,7 @@ class Program_weight_tensor_parameter_19: class Program_weight_tensor_parameter_20: name = "parameter_20" + original_name = "linear_31.w_0" shape = [384, 1152] dtype = "float32" min_val = float("-0.508469") @@ -231,6 +252,7 @@ class Program_weight_tensor_parameter_20: class Program_weight_tensor_parameter_21: name = "parameter_21" + original_name = "layer_norm_13.b_0" shape = [384] dtype = "float32" min_val = float("-0.304213") @@ -242,6 +264,7 @@ class Program_weight_tensor_parameter_21: class Program_weight_tensor_parameter_22: name = "parameter_22" + original_name = "layer_norm_13.w_0" shape = [384] dtype = "float32" min_val = float("-0.00319213") @@ -253,6 +276,7 @@ class Program_weight_tensor_parameter_22: class Program_weight_tensor_parameter_23: name = "parameter_23" + original_name = "linear_30.b_0" shape = [384] dtype = "float32" min_val = float("-0.0102057") @@ -264,6 +288,7 @@ class Program_weight_tensor_parameter_23: class Program_weight_tensor_parameter_24: name = "parameter_24" + original_name = "linear_30.w_0" shape = [1536, 384] dtype = "float32" min_val = float("-0.252119") @@ -275,6 +300,7 @@ class Program_weight_tensor_parameter_24: class Program_weight_tensor_parameter_25: name = "parameter_25" + original_name = "linear_29.b_0" shape = [1536] dtype = "float32" min_val = float("-0.157487") @@ -286,6 +312,7 @@ class Program_weight_tensor_parameter_25: class Program_weight_tensor_parameter_26: name = "parameter_26" + original_name = "linear_29.w_0" shape = [384, 1536] dtype = "float32" min_val = float("-0.369529") @@ -297,6 +324,7 @@ class Program_weight_tensor_parameter_26: class Program_weight_tensor_parameter_27: name = "parameter_27" + original_name = "layer_norm_12.b_0" shape = [384] dtype = "float32" min_val = float("-0.284967") @@ -308,6 +336,7 @@ class Program_weight_tensor_parameter_27: class Program_weight_tensor_parameter_28: name = "parameter_28" + original_name = "layer_norm_12.w_0" shape = [384] dtype = "float32" min_val = float("-4.96736e-05") @@ -319,6 +348,7 @@ class Program_weight_tensor_parameter_28: class Program_weight_tensor_parameter_29: name = "parameter_29" + original_name = "linear_28.b_0" shape = [384] dtype = "float32" min_val = float("-0.114198") @@ -330,6 +360,7 @@ class Program_weight_tensor_parameter_29: class Program_weight_tensor_parameter_30: name = "parameter_30" + original_name = "linear_28.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.30081") @@ -341,6 +372,7 @@ class Program_weight_tensor_parameter_30: class Program_weight_tensor_parameter_31: name = "parameter_31" + original_name = "linear_27.b_0" shape = [768] dtype = "float32" min_val = float("-0.141038") @@ -352,6 +384,7 @@ class Program_weight_tensor_parameter_31: class Program_weight_tensor_parameter_32: name = "parameter_32" + original_name = "linear_27.w_0" shape = [384, 768] dtype = "float32" min_val = float("-0.66392") @@ -363,6 +396,7 @@ class Program_weight_tensor_parameter_32: class Program_weight_tensor_parameter_33: name = "parameter_33" + original_name = "linear_26.b_0" shape = [384] dtype = "float32" min_val = float("-0.18186") @@ -374,6 +408,7 @@ class Program_weight_tensor_parameter_33: class Program_weight_tensor_parameter_34: name = "parameter_34" + original_name = "linear_26.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.641299") @@ -385,6 +420,7 @@ class Program_weight_tensor_parameter_34: class Program_weight_tensor_parameter_35: name = "parameter_35" + original_name = "layer_norm_11.b_0" shape = [384] dtype = "float32" min_val = float("-0.904482") @@ -396,6 +432,7 @@ class Program_weight_tensor_parameter_35: class Program_weight_tensor_parameter_36: name = "parameter_36" + original_name = "layer_norm_11.w_0" shape = [384] dtype = "float32" min_val = float("-0.135072") @@ -407,6 +444,7 @@ class Program_weight_tensor_parameter_36: class Program_weight_tensor_parameter_37: name = "parameter_37" + original_name = "linear_25.b_0" shape = [384] dtype = "float32" min_val = float("-0.0547508") @@ -418,6 +456,7 @@ class Program_weight_tensor_parameter_37: class Program_weight_tensor_parameter_38: name = "parameter_38" + original_name = "linear_25.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.429037") @@ -429,6 +468,7 @@ class Program_weight_tensor_parameter_38: class Program_weight_tensor_parameter_39: name = "parameter_39" + original_name = "linear_24.b_0" shape = [1152] dtype = "float32" min_val = float("-0.22592") @@ -440,6 +480,7 @@ class Program_weight_tensor_parameter_39: class Program_weight_tensor_parameter_40: name = "parameter_40" + original_name = "linear_24.w_0" shape = [384, 1152] dtype = "float32" min_val = float("-0.753119") @@ -451,6 +492,7 @@ class Program_weight_tensor_parameter_40: class Program_weight_tensor_parameter_41: name = "parameter_41" + original_name = "layer_norm_10.b_0" shape = [384] dtype = "float32" min_val = float("-0.88727") @@ -462,6 +504,7 @@ class Program_weight_tensor_parameter_41: class Program_weight_tensor_parameter_42: name = "parameter_42" + original_name = "layer_norm_10.w_0" shape = [384] dtype = "float32" min_val = float("-0.00890233") @@ -473,6 +516,7 @@ class Program_weight_tensor_parameter_42: class Program_weight_tensor_parameter_43: name = "parameter_43" + original_name = "linear_23.b_0" shape = [384] dtype = "float32" min_val = float("-0.0250739") @@ -484,6 +528,7 @@ class Program_weight_tensor_parameter_43: class Program_weight_tensor_parameter_44: name = "parameter_44" + original_name = "linear_23.w_0" shape = [1536, 384] dtype = "float32" min_val = float("-0.300596") @@ -495,6 +540,7 @@ class Program_weight_tensor_parameter_44: class Program_weight_tensor_parameter_45: name = "parameter_45" + original_name = "linear_22.b_0" shape = [1536] dtype = "float32" min_val = float("-0.168316") @@ -506,6 +552,7 @@ class Program_weight_tensor_parameter_45: class Program_weight_tensor_parameter_46: name = "parameter_46" + original_name = "linear_22.w_0" shape = [384, 1536] dtype = "float32" min_val = float("-0.2671") @@ -517,6 +564,7 @@ class Program_weight_tensor_parameter_46: class Program_weight_tensor_parameter_47: name = "parameter_47" + original_name = "layer_norm_9.b_0" shape = [384] dtype = "float32" min_val = float("-0.358253") @@ -528,6 +576,7 @@ class Program_weight_tensor_parameter_47: class Program_weight_tensor_parameter_48: name = "parameter_48" + original_name = "layer_norm_9.w_0" shape = [384] dtype = "float32" min_val = float("-0.0111031") @@ -539,6 +588,7 @@ class Program_weight_tensor_parameter_48: class Program_weight_tensor_parameter_49: name = "parameter_49" + original_name = "linear_21.b_0" shape = [384] dtype = "float32" min_val = float("-0.0907258") @@ -550,6 +600,7 @@ class Program_weight_tensor_parameter_49: class Program_weight_tensor_parameter_50: name = "parameter_50" + original_name = "linear_21.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.296141") @@ -561,6 +612,7 @@ class Program_weight_tensor_parameter_50: class Program_weight_tensor_parameter_51: name = "parameter_51" + original_name = "linear_20.b_0" shape = [768] dtype = "float32" min_val = float("-0.246955") @@ -572,6 +624,7 @@ class Program_weight_tensor_parameter_51: class Program_weight_tensor_parameter_52: name = "parameter_52" + original_name = "linear_20.w_0" shape = [384, 768] dtype = "float32" min_val = float("-0.55802") @@ -583,6 +636,7 @@ class Program_weight_tensor_parameter_52: class Program_weight_tensor_parameter_53: name = "parameter_53" + original_name = "linear_19.b_0" shape = [384] dtype = "float32" min_val = float("-0.227319") @@ -594,6 +648,7 @@ class Program_weight_tensor_parameter_53: class Program_weight_tensor_parameter_54: name = "parameter_54" + original_name = "linear_19.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.458955") @@ -605,6 +660,7 @@ class Program_weight_tensor_parameter_54: class Program_weight_tensor_parameter_55: name = "parameter_55" + original_name = "layer_norm_8.b_0" shape = [384] dtype = "float32" min_val = float("-0.314201") @@ -616,6 +672,7 @@ class Program_weight_tensor_parameter_55: class Program_weight_tensor_parameter_56: name = "parameter_56" + original_name = "layer_norm_8.w_0" shape = [384] dtype = "float32" min_val = float("-0.0335737") @@ -627,6 +684,7 @@ class Program_weight_tensor_parameter_56: class Program_weight_tensor_parameter_57: name = "parameter_57" + original_name = "linear_18.b_0" shape = [384] dtype = "float32" min_val = float("-0.0935431") @@ -638,6 +696,7 @@ class Program_weight_tensor_parameter_57: class Program_weight_tensor_parameter_58: name = "parameter_58" + original_name = "linear_18.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.316009") @@ -649,6 +708,7 @@ class Program_weight_tensor_parameter_58: class Program_weight_tensor_parameter_59: name = "parameter_59" + original_name = "linear_17.b_0" shape = [1152] dtype = "float32" min_val = float("-0.239339") @@ -660,6 +720,7 @@ class Program_weight_tensor_parameter_59: class Program_weight_tensor_parameter_60: name = "parameter_60" + original_name = "linear_17.w_0" shape = [384, 1152] dtype = "float32" min_val = float("-0.461366") @@ -671,6 +732,7 @@ class Program_weight_tensor_parameter_60: class Program_weight_tensor_parameter_61: name = "parameter_61" + original_name = "layer_norm_7.b_0" shape = [384] dtype = "float32" min_val = float("-0.182215") @@ -682,6 +744,7 @@ class Program_weight_tensor_parameter_61: class Program_weight_tensor_parameter_62: name = "parameter_62" + original_name = "layer_norm_7.w_0" shape = [384] dtype = "float32" min_val = float("-0.00580322") @@ -693,6 +756,7 @@ class Program_weight_tensor_parameter_62: class Program_weight_tensor_parameter_63: name = "parameter_63" + original_name = "linear_16.b_0" shape = [384] dtype = "float32" min_val = float("-0.0179786") @@ -704,6 +768,7 @@ class Program_weight_tensor_parameter_63: class Program_weight_tensor_parameter_64: name = "parameter_64" + original_name = "linear_16.w_0" shape = [1536, 384] dtype = "float32" min_val = float("-0.321224") @@ -715,6 +780,7 @@ class Program_weight_tensor_parameter_64: class Program_weight_tensor_parameter_65: name = "parameter_65" + original_name = "linear_15.b_0" shape = [1536] dtype = "float32" min_val = float("-0.25248") @@ -726,6 +792,7 @@ class Program_weight_tensor_parameter_65: class Program_weight_tensor_parameter_66: name = "parameter_66" + original_name = "linear_15.w_0" shape = [384, 1536] dtype = "float32" min_val = float("-0.246037") @@ -737,6 +804,7 @@ class Program_weight_tensor_parameter_66: class Program_weight_tensor_parameter_67: name = "parameter_67" + original_name = "layer_norm_6.b_0" shape = [384] dtype = "float32" min_val = float("-0.692917") @@ -748,6 +816,7 @@ class Program_weight_tensor_parameter_67: class Program_weight_tensor_parameter_68: name = "parameter_68" + original_name = "layer_norm_6.w_0" shape = [384] dtype = "float32" min_val = float("-0.00846829") @@ -759,6 +828,7 @@ class Program_weight_tensor_parameter_68: class Program_weight_tensor_parameter_69: name = "parameter_69" + original_name = "linear_14.b_0" shape = [384] dtype = "float32" min_val = float("-0.0622596") @@ -770,6 +840,7 @@ class Program_weight_tensor_parameter_69: class Program_weight_tensor_parameter_70: name = "parameter_70" + original_name = "linear_14.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.297331") @@ -781,6 +852,7 @@ class Program_weight_tensor_parameter_70: class Program_weight_tensor_parameter_71: name = "parameter_71" + original_name = "linear_13.b_0" shape = [768] dtype = "float32" min_val = float("-0.286059") @@ -792,6 +864,7 @@ class Program_weight_tensor_parameter_71: class Program_weight_tensor_parameter_72: name = "parameter_72" + original_name = "linear_13.w_0" shape = [384, 768] dtype = "float32" min_val = float("-0.554386") @@ -803,6 +876,7 @@ class Program_weight_tensor_parameter_72: class Program_weight_tensor_parameter_73: name = "parameter_73" + original_name = "linear_12.b_0" shape = [384] dtype = "float32" min_val = float("-0.303832") @@ -814,6 +888,7 @@ class Program_weight_tensor_parameter_73: class Program_weight_tensor_parameter_74: name = "parameter_74" + original_name = "linear_12.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.366028") @@ -825,6 +900,7 @@ class Program_weight_tensor_parameter_74: class Program_weight_tensor_parameter_75: name = "parameter_75" + original_name = "layer_norm_5.b_0" shape = [384] dtype = "float32" min_val = float("-0.254323") @@ -836,6 +912,7 @@ class Program_weight_tensor_parameter_75: class Program_weight_tensor_parameter_76: name = "parameter_76" + original_name = "layer_norm_5.w_0" shape = [384] dtype = "float32" min_val = float("-0.108733") @@ -847,6 +924,7 @@ class Program_weight_tensor_parameter_76: class Program_weight_tensor_parameter_77: name = "parameter_77" + original_name = "linear_11.b_0" shape = [384] dtype = "float32" min_val = float("-0.165999") @@ -858,6 +936,7 @@ class Program_weight_tensor_parameter_77: class Program_weight_tensor_parameter_78: name = "parameter_78" + original_name = "linear_11.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.298527") @@ -869,6 +948,7 @@ class Program_weight_tensor_parameter_78: class Program_weight_tensor_parameter_79: name = "parameter_79" + original_name = "linear_10.b_0" shape = [1152] dtype = "float32" min_val = float("-0.0263861") @@ -880,6 +960,7 @@ class Program_weight_tensor_parameter_79: class Program_weight_tensor_parameter_80: name = "parameter_80" + original_name = "linear_10.w_0" shape = [384, 1152] dtype = "float32" min_val = float("-0.245316") @@ -891,6 +972,7 @@ class Program_weight_tensor_parameter_80: class Program_weight_tensor_parameter_81: name = "parameter_81" + original_name = "embedding_0.w_0" shape = [6629, 384] dtype = "float32" min_val = float("-0.389145") @@ -902,6 +984,7 @@ class Program_weight_tensor_parameter_81: class Program_weight_tensor_parameter_82: name = "parameter_82" + original_name = "linear_9.w_0" shape = [1024, 384] dtype = "float32" min_val = float("-1.23423") diff --git a/paddle_samples/PaddleX/PP-OCRv4_server_rec/subgraph_4/input_meta.py b/paddle_samples/PaddleX/PP-OCRv4_server_rec/subgraph_4/input_meta.py index 447a44198..963aedb9e 100644 --- a/paddle_samples/PaddleX/PP-OCRv4_server_rec/subgraph_4/input_meta.py +++ b/paddle_samples/PaddleX/PP-OCRv4_server_rec/subgraph_4/input_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_data_0: name = "data_0" + original_name = "var_309" shape = [4, 1024, 1, 40] dtype = "float32" max_val = float("24.2002") @@ -10,6 +11,7 @@ class Program_weight_tensor_data_0: class Program_weight_tensor_data_1: name = "data_1" + original_name = "var_311" shape = [4, 25] dtype = "int64" min_val = 0 @@ -19,6 +21,7 @@ class Program_weight_tensor_data_1: class Program_weight_tensor_data_2: name = "data_2" + original_name = "var_312" shape = [4] dtype = "int64" data = [7, 4, 3, 12] @@ -26,6 +29,7 @@ class Program_weight_tensor_data_2: class Program_weight_tensor_data_3: name = "data_3" + original_name = "var_324" shape = [5000, 1, 384] dtype = "float32" min_val = float("-1.0") diff --git a/paddle_samples/PaddleX/PP-OCRv4_server_rec/subgraph_4/weight_meta.py b/paddle_samples/PaddleX/PP-OCRv4_server_rec/subgraph_4/weight_meta.py index ca8ed42a3..1c1ce2e52 100644 --- a/paddle_samples/PaddleX/PP-OCRv4_server_rec/subgraph_4/weight_meta.py +++ b/paddle_samples/PaddleX/PP-OCRv4_server_rec/subgraph_4/weight_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_parameter_0: name = "parameter_0" + original_name = "linear_38.w_0" shape = [384, 6629] dtype = "float32" min_val = float("-1.32918") @@ -11,6 +12,7 @@ class Program_weight_tensor_parameter_0: class Program_weight_tensor_parameter_1: name = "parameter_1" + original_name = "layer_norm_16.b_0" shape = [384] dtype = "float32" min_val = float("-2.00494") @@ -22,6 +24,7 @@ class Program_weight_tensor_parameter_1: class Program_weight_tensor_parameter_2: name = "parameter_2" + original_name = "layer_norm_16.w_0" shape = [384] dtype = "float32" min_val = float("-0.000245228") @@ -33,6 +36,7 @@ class Program_weight_tensor_parameter_2: class Program_weight_tensor_parameter_3: name = "parameter_3" + original_name = "linear_37.b_0" shape = [384] dtype = "float32" min_val = float("-0.0518493") @@ -44,6 +48,7 @@ class Program_weight_tensor_parameter_3: class Program_weight_tensor_parameter_4: name = "parameter_4" + original_name = "linear_37.w_0" shape = [1536, 384] dtype = "float32" min_val = float("-0.227294") @@ -55,6 +60,7 @@ class Program_weight_tensor_parameter_4: class Program_weight_tensor_parameter_5: name = "parameter_5" + original_name = "linear_36.b_0" shape = [1536] dtype = "float32" min_val = float("-0.200106") @@ -66,6 +72,7 @@ class Program_weight_tensor_parameter_5: class Program_weight_tensor_parameter_6: name = "parameter_6" + original_name = "linear_36.w_0" shape = [384, 1536] dtype = "float32" min_val = float("-0.957273") @@ -77,6 +84,7 @@ class Program_weight_tensor_parameter_6: class Program_weight_tensor_parameter_7: name = "parameter_7" + original_name = "layer_norm_15.b_0" shape = [384] dtype = "float32" min_val = float("-0.0491315") @@ -88,6 +96,7 @@ class Program_weight_tensor_parameter_7: class Program_weight_tensor_parameter_8: name = "parameter_8" + original_name = "layer_norm_15.w_0" shape = [384] dtype = "float32" min_val = float("-0.0866739") @@ -99,6 +108,7 @@ class Program_weight_tensor_parameter_8: class Program_weight_tensor_parameter_9: name = "parameter_9" + original_name = "linear_35.b_0" shape = [384] dtype = "float32" min_val = float("-0.0762894") @@ -110,6 +120,7 @@ class Program_weight_tensor_parameter_9: class Program_weight_tensor_parameter_10: name = "parameter_10" + original_name = "linear_35.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.264504") @@ -121,6 +132,7 @@ class Program_weight_tensor_parameter_10: class Program_weight_tensor_parameter_11: name = "parameter_11" + original_name = "linear_34.b_0" shape = [768] dtype = "float32" min_val = float("-0.0725435") @@ -132,6 +144,7 @@ class Program_weight_tensor_parameter_11: class Program_weight_tensor_parameter_12: name = "parameter_12" + original_name = "linear_34.w_0" shape = [384, 768] dtype = "float32" min_val = float("-0.701239") @@ -143,6 +156,7 @@ class Program_weight_tensor_parameter_12: class Program_weight_tensor_parameter_13: name = "parameter_13" + original_name = "linear_33.b_0" shape = [384] dtype = "float32" min_val = float("-0.176879") @@ -154,6 +168,7 @@ class Program_weight_tensor_parameter_13: class Program_weight_tensor_parameter_14: name = "parameter_14" + original_name = "linear_33.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.619542") @@ -165,6 +180,7 @@ class Program_weight_tensor_parameter_14: class Program_weight_tensor_parameter_15: name = "parameter_15" + original_name = "layer_norm_14.b_0" shape = [384] dtype = "float32" min_val = float("-1.64474") @@ -176,6 +192,7 @@ class Program_weight_tensor_parameter_15: class Program_weight_tensor_parameter_16: name = "parameter_16" + original_name = "layer_norm_14.w_0" shape = [384] dtype = "float32" min_val = float("-0.291678") @@ -187,6 +204,7 @@ class Program_weight_tensor_parameter_16: class Program_weight_tensor_parameter_17: name = "parameter_17" + original_name = "linear_32.b_0" shape = [384] dtype = "float32" min_val = float("-0.0726199") @@ -198,6 +216,7 @@ class Program_weight_tensor_parameter_17: class Program_weight_tensor_parameter_18: name = "parameter_18" + original_name = "linear_32.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.310778") @@ -209,6 +228,7 @@ class Program_weight_tensor_parameter_18: class Program_weight_tensor_parameter_19: name = "parameter_19" + original_name = "linear_31.b_0" shape = [1152] dtype = "float32" min_val = float("-0.275863") @@ -220,6 +240,7 @@ class Program_weight_tensor_parameter_19: class Program_weight_tensor_parameter_20: name = "parameter_20" + original_name = "linear_31.w_0" shape = [384, 1152] dtype = "float32" min_val = float("-0.508469") @@ -231,6 +252,7 @@ class Program_weight_tensor_parameter_20: class Program_weight_tensor_parameter_21: name = "parameter_21" + original_name = "layer_norm_13.b_0" shape = [384] dtype = "float32" min_val = float("-0.304213") @@ -242,6 +264,7 @@ class Program_weight_tensor_parameter_21: class Program_weight_tensor_parameter_22: name = "parameter_22" + original_name = "layer_norm_13.w_0" shape = [384] dtype = "float32" min_val = float("-0.00319213") @@ -253,6 +276,7 @@ class Program_weight_tensor_parameter_22: class Program_weight_tensor_parameter_23: name = "parameter_23" + original_name = "linear_30.b_0" shape = [384] dtype = "float32" min_val = float("-0.0102057") @@ -264,6 +288,7 @@ class Program_weight_tensor_parameter_23: class Program_weight_tensor_parameter_24: name = "parameter_24" + original_name = "linear_30.w_0" shape = [1536, 384] dtype = "float32" min_val = float("-0.252119") @@ -275,6 +300,7 @@ class Program_weight_tensor_parameter_24: class Program_weight_tensor_parameter_25: name = "parameter_25" + original_name = "linear_29.b_0" shape = [1536] dtype = "float32" min_val = float("-0.157487") @@ -286,6 +312,7 @@ class Program_weight_tensor_parameter_25: class Program_weight_tensor_parameter_26: name = "parameter_26" + original_name = "linear_29.w_0" shape = [384, 1536] dtype = "float32" min_val = float("-0.369529") @@ -297,6 +324,7 @@ class Program_weight_tensor_parameter_26: class Program_weight_tensor_parameter_27: name = "parameter_27" + original_name = "layer_norm_12.b_0" shape = [384] dtype = "float32" min_val = float("-0.284967") @@ -308,6 +336,7 @@ class Program_weight_tensor_parameter_27: class Program_weight_tensor_parameter_28: name = "parameter_28" + original_name = "layer_norm_12.w_0" shape = [384] dtype = "float32" min_val = float("-4.96736e-05") @@ -319,6 +348,7 @@ class Program_weight_tensor_parameter_28: class Program_weight_tensor_parameter_29: name = "parameter_29" + original_name = "linear_28.b_0" shape = [384] dtype = "float32" min_val = float("-0.114198") @@ -330,6 +360,7 @@ class Program_weight_tensor_parameter_29: class Program_weight_tensor_parameter_30: name = "parameter_30" + original_name = "linear_28.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.30081") @@ -341,6 +372,7 @@ class Program_weight_tensor_parameter_30: class Program_weight_tensor_parameter_31: name = "parameter_31" + original_name = "linear_27.b_0" shape = [768] dtype = "float32" min_val = float("-0.141038") @@ -352,6 +384,7 @@ class Program_weight_tensor_parameter_31: class Program_weight_tensor_parameter_32: name = "parameter_32" + original_name = "linear_27.w_0" shape = [384, 768] dtype = "float32" min_val = float("-0.66392") @@ -363,6 +396,7 @@ class Program_weight_tensor_parameter_32: class Program_weight_tensor_parameter_33: name = "parameter_33" + original_name = "linear_26.b_0" shape = [384] dtype = "float32" min_val = float("-0.18186") @@ -374,6 +408,7 @@ class Program_weight_tensor_parameter_33: class Program_weight_tensor_parameter_34: name = "parameter_34" + original_name = "linear_26.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.641299") @@ -385,6 +420,7 @@ class Program_weight_tensor_parameter_34: class Program_weight_tensor_parameter_35: name = "parameter_35" + original_name = "layer_norm_11.b_0" shape = [384] dtype = "float32" min_val = float("-0.904482") @@ -396,6 +432,7 @@ class Program_weight_tensor_parameter_35: class Program_weight_tensor_parameter_36: name = "parameter_36" + original_name = "layer_norm_11.w_0" shape = [384] dtype = "float32" min_val = float("-0.135072") @@ -407,6 +444,7 @@ class Program_weight_tensor_parameter_36: class Program_weight_tensor_parameter_37: name = "parameter_37" + original_name = "linear_25.b_0" shape = [384] dtype = "float32" min_val = float("-0.0547508") @@ -418,6 +456,7 @@ class Program_weight_tensor_parameter_37: class Program_weight_tensor_parameter_38: name = "parameter_38" + original_name = "linear_25.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.429037") @@ -429,6 +468,7 @@ class Program_weight_tensor_parameter_38: class Program_weight_tensor_parameter_39: name = "parameter_39" + original_name = "linear_24.b_0" shape = [1152] dtype = "float32" min_val = float("-0.22592") @@ -440,6 +480,7 @@ class Program_weight_tensor_parameter_39: class Program_weight_tensor_parameter_40: name = "parameter_40" + original_name = "linear_24.w_0" shape = [384, 1152] dtype = "float32" min_val = float("-0.753119") @@ -451,6 +492,7 @@ class Program_weight_tensor_parameter_40: class Program_weight_tensor_parameter_41: name = "parameter_41" + original_name = "layer_norm_10.b_0" shape = [384] dtype = "float32" min_val = float("-0.88727") @@ -462,6 +504,7 @@ class Program_weight_tensor_parameter_41: class Program_weight_tensor_parameter_42: name = "parameter_42" + original_name = "layer_norm_10.w_0" shape = [384] dtype = "float32" min_val = float("-0.00890233") @@ -473,6 +516,7 @@ class Program_weight_tensor_parameter_42: class Program_weight_tensor_parameter_43: name = "parameter_43" + original_name = "linear_23.b_0" shape = [384] dtype = "float32" min_val = float("-0.0250739") @@ -484,6 +528,7 @@ class Program_weight_tensor_parameter_43: class Program_weight_tensor_parameter_44: name = "parameter_44" + original_name = "linear_23.w_0" shape = [1536, 384] dtype = "float32" min_val = float("-0.300596") @@ -495,6 +540,7 @@ class Program_weight_tensor_parameter_44: class Program_weight_tensor_parameter_45: name = "parameter_45" + original_name = "linear_22.b_0" shape = [1536] dtype = "float32" min_val = float("-0.168316") @@ -506,6 +552,7 @@ class Program_weight_tensor_parameter_45: class Program_weight_tensor_parameter_46: name = "parameter_46" + original_name = "linear_22.w_0" shape = [384, 1536] dtype = "float32" min_val = float("-0.2671") @@ -517,6 +564,7 @@ class Program_weight_tensor_parameter_46: class Program_weight_tensor_parameter_47: name = "parameter_47" + original_name = "layer_norm_9.b_0" shape = [384] dtype = "float32" min_val = float("-0.358253") @@ -528,6 +576,7 @@ class Program_weight_tensor_parameter_47: class Program_weight_tensor_parameter_48: name = "parameter_48" + original_name = "layer_norm_9.w_0" shape = [384] dtype = "float32" min_val = float("-0.0111031") @@ -539,6 +588,7 @@ class Program_weight_tensor_parameter_48: class Program_weight_tensor_parameter_49: name = "parameter_49" + original_name = "linear_21.b_0" shape = [384] dtype = "float32" min_val = float("-0.0907258") @@ -550,6 +600,7 @@ class Program_weight_tensor_parameter_49: class Program_weight_tensor_parameter_50: name = "parameter_50" + original_name = "linear_21.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.296141") @@ -561,6 +612,7 @@ class Program_weight_tensor_parameter_50: class Program_weight_tensor_parameter_51: name = "parameter_51" + original_name = "linear_20.b_0" shape = [768] dtype = "float32" min_val = float("-0.246955") @@ -572,6 +624,7 @@ class Program_weight_tensor_parameter_51: class Program_weight_tensor_parameter_52: name = "parameter_52" + original_name = "linear_20.w_0" shape = [384, 768] dtype = "float32" min_val = float("-0.55802") @@ -583,6 +636,7 @@ class Program_weight_tensor_parameter_52: class Program_weight_tensor_parameter_53: name = "parameter_53" + original_name = "linear_19.b_0" shape = [384] dtype = "float32" min_val = float("-0.227319") @@ -594,6 +648,7 @@ class Program_weight_tensor_parameter_53: class Program_weight_tensor_parameter_54: name = "parameter_54" + original_name = "linear_19.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.458955") @@ -605,6 +660,7 @@ class Program_weight_tensor_parameter_54: class Program_weight_tensor_parameter_55: name = "parameter_55" + original_name = "layer_norm_8.b_0" shape = [384] dtype = "float32" min_val = float("-0.314201") @@ -616,6 +672,7 @@ class Program_weight_tensor_parameter_55: class Program_weight_tensor_parameter_56: name = "parameter_56" + original_name = "layer_norm_8.w_0" shape = [384] dtype = "float32" min_val = float("-0.0335737") @@ -627,6 +684,7 @@ class Program_weight_tensor_parameter_56: class Program_weight_tensor_parameter_57: name = "parameter_57" + original_name = "linear_18.b_0" shape = [384] dtype = "float32" min_val = float("-0.0935431") @@ -638,6 +696,7 @@ class Program_weight_tensor_parameter_57: class Program_weight_tensor_parameter_58: name = "parameter_58" + original_name = "linear_18.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.316009") @@ -649,6 +708,7 @@ class Program_weight_tensor_parameter_58: class Program_weight_tensor_parameter_59: name = "parameter_59" + original_name = "linear_17.b_0" shape = [1152] dtype = "float32" min_val = float("-0.239339") @@ -660,6 +720,7 @@ class Program_weight_tensor_parameter_59: class Program_weight_tensor_parameter_60: name = "parameter_60" + original_name = "linear_17.w_0" shape = [384, 1152] dtype = "float32" min_val = float("-0.461366") @@ -671,6 +732,7 @@ class Program_weight_tensor_parameter_60: class Program_weight_tensor_parameter_61: name = "parameter_61" + original_name = "layer_norm_7.b_0" shape = [384] dtype = "float32" min_val = float("-0.182215") @@ -682,6 +744,7 @@ class Program_weight_tensor_parameter_61: class Program_weight_tensor_parameter_62: name = "parameter_62" + original_name = "layer_norm_7.w_0" shape = [384] dtype = "float32" min_val = float("-0.00580322") @@ -693,6 +756,7 @@ class Program_weight_tensor_parameter_62: class Program_weight_tensor_parameter_63: name = "parameter_63" + original_name = "linear_16.b_0" shape = [384] dtype = "float32" min_val = float("-0.0179786") @@ -704,6 +768,7 @@ class Program_weight_tensor_parameter_63: class Program_weight_tensor_parameter_64: name = "parameter_64" + original_name = "linear_16.w_0" shape = [1536, 384] dtype = "float32" min_val = float("-0.321224") @@ -715,6 +780,7 @@ class Program_weight_tensor_parameter_64: class Program_weight_tensor_parameter_65: name = "parameter_65" + original_name = "linear_15.b_0" shape = [1536] dtype = "float32" min_val = float("-0.25248") @@ -726,6 +792,7 @@ class Program_weight_tensor_parameter_65: class Program_weight_tensor_parameter_66: name = "parameter_66" + original_name = "linear_15.w_0" shape = [384, 1536] dtype = "float32" min_val = float("-0.246037") @@ -737,6 +804,7 @@ class Program_weight_tensor_parameter_66: class Program_weight_tensor_parameter_67: name = "parameter_67" + original_name = "layer_norm_6.b_0" shape = [384] dtype = "float32" min_val = float("-0.692917") @@ -748,6 +816,7 @@ class Program_weight_tensor_parameter_67: class Program_weight_tensor_parameter_68: name = "parameter_68" + original_name = "layer_norm_6.w_0" shape = [384] dtype = "float32" min_val = float("-0.00846829") @@ -759,6 +828,7 @@ class Program_weight_tensor_parameter_68: class Program_weight_tensor_parameter_69: name = "parameter_69" + original_name = "linear_14.b_0" shape = [384] dtype = "float32" min_val = float("-0.0622596") @@ -770,6 +840,7 @@ class Program_weight_tensor_parameter_69: class Program_weight_tensor_parameter_70: name = "parameter_70" + original_name = "linear_14.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.297331") @@ -781,6 +852,7 @@ class Program_weight_tensor_parameter_70: class Program_weight_tensor_parameter_71: name = "parameter_71" + original_name = "linear_13.b_0" shape = [768] dtype = "float32" min_val = float("-0.286059") @@ -792,6 +864,7 @@ class Program_weight_tensor_parameter_71: class Program_weight_tensor_parameter_72: name = "parameter_72" + original_name = "linear_13.w_0" shape = [384, 768] dtype = "float32" min_val = float("-0.554386") @@ -803,6 +876,7 @@ class Program_weight_tensor_parameter_72: class Program_weight_tensor_parameter_73: name = "parameter_73" + original_name = "linear_12.b_0" shape = [384] dtype = "float32" min_val = float("-0.303832") @@ -814,6 +888,7 @@ class Program_weight_tensor_parameter_73: class Program_weight_tensor_parameter_74: name = "parameter_74" + original_name = "linear_12.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.366028") @@ -825,6 +900,7 @@ class Program_weight_tensor_parameter_74: class Program_weight_tensor_parameter_75: name = "parameter_75" + original_name = "layer_norm_5.b_0" shape = [384] dtype = "float32" min_val = float("-0.254323") @@ -836,6 +912,7 @@ class Program_weight_tensor_parameter_75: class Program_weight_tensor_parameter_76: name = "parameter_76" + original_name = "layer_norm_5.w_0" shape = [384] dtype = "float32" min_val = float("-0.108733") @@ -847,6 +924,7 @@ class Program_weight_tensor_parameter_76: class Program_weight_tensor_parameter_77: name = "parameter_77" + original_name = "linear_11.b_0" shape = [384] dtype = "float32" min_val = float("-0.165999") @@ -858,6 +936,7 @@ class Program_weight_tensor_parameter_77: class Program_weight_tensor_parameter_78: name = "parameter_78" + original_name = "linear_11.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.298527") @@ -869,6 +948,7 @@ class Program_weight_tensor_parameter_78: class Program_weight_tensor_parameter_79: name = "parameter_79" + original_name = "linear_10.b_0" shape = [1152] dtype = "float32" min_val = float("-0.0263861") @@ -880,6 +960,7 @@ class Program_weight_tensor_parameter_79: class Program_weight_tensor_parameter_80: name = "parameter_80" + original_name = "linear_10.w_0" shape = [384, 1152] dtype = "float32" min_val = float("-0.245316") @@ -891,6 +972,7 @@ class Program_weight_tensor_parameter_80: class Program_weight_tensor_parameter_81: name = "parameter_81" + original_name = "embedding_0.w_0" shape = [6629, 384] dtype = "float32" min_val = float("-0.389145") @@ -902,6 +984,7 @@ class Program_weight_tensor_parameter_81: class Program_weight_tensor_parameter_82: name = "parameter_82" + original_name = "linear_9.w_0" shape = [1024, 384] dtype = "float32" min_val = float("-1.23423") diff --git a/paddle_samples/PaddleX/PP-OCRv4_server_seal_det/input_meta.py b/paddle_samples/PaddleX/PP-OCRv4_server_seal_det/input_meta.py index d3649d2e0..a05227057 100644 --- a/paddle_samples/PaddleX/PP-OCRv4_server_seal_det/input_meta.py +++ b/paddle_samples/PaddleX/PP-OCRv4_server_seal_det/input_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_data_0: name = "data_0" + original_name = "var_0" shape = [8, 3, 640, 640] dtype = "float32" min_val = float("-2.1179") diff --git a/paddle_samples/PaddleX/PP-OCRv4_server_seal_det/weight_meta.py b/paddle_samples/PaddleX/PP-OCRv4_server_seal_det/weight_meta.py index cf444ac19..814b42986 100644 --- a/paddle_samples/PaddleX/PP-OCRv4_server_seal_det/weight_meta.py +++ b/paddle_samples/PaddleX/PP-OCRv4_server_seal_det/weight_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_parameter_0: name = "parameter_0" + original_name = "conv2d_transpose_3.b_0" shape = [1] dtype = "float32" min_val = float("0") @@ -9,6 +10,7 @@ class Program_weight_tensor_parameter_0: class Program_weight_tensor_parameter_1: name = "parameter_1" + original_name = "conv2d_transpose_3.w_0" shape = [64, 1, 2, 2] dtype = "float32" min_val = float("-1.16822") @@ -20,6 +22,7 @@ class Program_weight_tensor_parameter_1: class Program_weight_tensor_parameter_2: name = "parameter_2" + original_name = "batch_norm_3.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -29,6 +32,7 @@ class Program_weight_tensor_parameter_2: class Program_weight_tensor_parameter_3: name = "parameter_3" + original_name = "batch_norm_3.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -38,6 +42,7 @@ class Program_weight_tensor_parameter_3: class Program_weight_tensor_parameter_4: name = "parameter_4" + original_name = "batch_norm_3.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -47,6 +52,7 @@ class Program_weight_tensor_parameter_4: class Program_weight_tensor_parameter_5: name = "parameter_5" + original_name = "batch_norm_3.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -56,6 +62,7 @@ class Program_weight_tensor_parameter_5: class Program_weight_tensor_parameter_6: name = "parameter_6" + original_name = "conv2d_transpose_2.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -65,6 +72,7 @@ class Program_weight_tensor_parameter_6: class Program_weight_tensor_parameter_7: name = "parameter_7" + original_name = "conv2d_transpose_2.w_0" shape = [64, 64, 2, 2] dtype = "float32" min_val = float("-1.20277") @@ -76,6 +84,7 @@ class Program_weight_tensor_parameter_7: class Program_weight_tensor_parameter_8: name = "parameter_8" + original_name = "batch_norm_2.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -85,6 +94,7 @@ class Program_weight_tensor_parameter_8: class Program_weight_tensor_parameter_9: name = "parameter_9" + original_name = "batch_norm_2.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -94,6 +104,7 @@ class Program_weight_tensor_parameter_9: class Program_weight_tensor_parameter_10: name = "parameter_10" + original_name = "batch_norm_2.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -103,6 +114,7 @@ class Program_weight_tensor_parameter_10: class Program_weight_tensor_parameter_11: name = "parameter_11" + original_name = "batch_norm_2.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -112,6 +124,7 @@ class Program_weight_tensor_parameter_11: class Program_weight_tensor_parameter_12: name = "parameter_12" + original_name = "conv2d_106.w_0" shape = [64, 256, 3, 3] dtype = "float32" min_val = float("-0.73936") @@ -123,6 +136,7 @@ class Program_weight_tensor_parameter_12: class Program_weight_tensor_parameter_13: name = "parameter_13" + original_name = "conv2d_108.b_0" shape = [1] dtype = "float32" min_val = float("0") @@ -132,6 +146,7 @@ class Program_weight_tensor_parameter_13: class Program_weight_tensor_parameter_14: name = "parameter_14" + original_name = "conv2d_108.w_0" shape = [1, 64, 1, 1] dtype = "float32" min_val = float("0") @@ -141,6 +156,7 @@ class Program_weight_tensor_parameter_14: class Program_weight_tensor_parameter_15: name = "parameter_15" + original_name = "batch_norm_4.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -150,6 +166,7 @@ class Program_weight_tensor_parameter_15: class Program_weight_tensor_parameter_16: name = "parameter_16" + original_name = "batch_norm_4.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -159,6 +176,7 @@ class Program_weight_tensor_parameter_16: class Program_weight_tensor_parameter_17: name = "parameter_17" + original_name = "batch_norm_4.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -168,6 +186,7 @@ class Program_weight_tensor_parameter_17: class Program_weight_tensor_parameter_18: name = "parameter_18" + original_name = "batch_norm_4.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -177,6 +196,7 @@ class Program_weight_tensor_parameter_18: class Program_weight_tensor_parameter_19: name = "parameter_19" + original_name = "conv2d_107.w_0" shape = [64, 65, 3, 3] dtype = "float32" min_val = float("-1.2239") @@ -188,6 +208,7 @@ class Program_weight_tensor_parameter_19: class Program_weight_tensor_parameter_20: name = "parameter_20" + original_name = "conv2d_transpose_1.b_0" shape = [1] dtype = "float32" min_val = float("0") @@ -197,6 +218,7 @@ class Program_weight_tensor_parameter_20: class Program_weight_tensor_parameter_21: name = "parameter_21" + original_name = "conv2d_transpose_1.w_0" shape = [64, 1, 2, 2] dtype = "float32" min_val = float("-1.14801") @@ -208,6 +230,7 @@ class Program_weight_tensor_parameter_21: class Program_weight_tensor_parameter_22: name = "parameter_22" + original_name = "batch_norm_1.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -217,6 +240,7 @@ class Program_weight_tensor_parameter_22: class Program_weight_tensor_parameter_23: name = "parameter_23" + original_name = "batch_norm_1.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -226,6 +250,7 @@ class Program_weight_tensor_parameter_23: class Program_weight_tensor_parameter_24: name = "parameter_24" + original_name = "batch_norm_1.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -235,6 +260,7 @@ class Program_weight_tensor_parameter_24: class Program_weight_tensor_parameter_25: name = "parameter_25" + original_name = "batch_norm_1.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -244,6 +270,7 @@ class Program_weight_tensor_parameter_25: class Program_weight_tensor_parameter_26: name = "parameter_26" + original_name = "conv2d_transpose_0.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -253,6 +280,7 @@ class Program_weight_tensor_parameter_26: class Program_weight_tensor_parameter_27: name = "parameter_27" + original_name = "conv2d_transpose_0.w_0" shape = [64, 64, 2, 2] dtype = "float32" min_val = float("-0.525336") @@ -264,6 +292,7 @@ class Program_weight_tensor_parameter_27: class Program_weight_tensor_parameter_28: name = "parameter_28" + original_name = "batch_norm_0.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -273,6 +302,7 @@ class Program_weight_tensor_parameter_28: class Program_weight_tensor_parameter_29: name = "parameter_29" + original_name = "batch_norm_0.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -282,6 +312,7 @@ class Program_weight_tensor_parameter_29: class Program_weight_tensor_parameter_30: name = "parameter_30" + original_name = "batch_norm_0.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -291,6 +322,7 @@ class Program_weight_tensor_parameter_30: class Program_weight_tensor_parameter_31: name = "parameter_31" + original_name = "batch_norm_0.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -300,6 +332,7 @@ class Program_weight_tensor_parameter_31: class Program_weight_tensor_parameter_32: name = "parameter_32" + original_name = "conv2d_105.w_0" shape = [64, 256, 3, 3] dtype = "float32" min_val = float("-0.486898") @@ -311,6 +344,7 @@ class Program_weight_tensor_parameter_32: class Program_weight_tensor_parameter_33: name = "parameter_33" + original_name = "batch_norm2d_41.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -320,6 +354,7 @@ class Program_weight_tensor_parameter_33: class Program_weight_tensor_parameter_34: name = "parameter_34" + original_name = "batch_norm2d_41.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -329,6 +364,7 @@ class Program_weight_tensor_parameter_34: class Program_weight_tensor_parameter_35: name = "parameter_35" + original_name = "batch_norm2d_41.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -338,6 +374,7 @@ class Program_weight_tensor_parameter_35: class Program_weight_tensor_parameter_36: name = "parameter_36" + original_name = "batch_norm2d_41.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -347,6 +384,7 @@ class Program_weight_tensor_parameter_36: class Program_weight_tensor_parameter_37: name = "parameter_37" + original_name = "conv2d_62.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -356,6 +394,7 @@ class Program_weight_tensor_parameter_37: class Program_weight_tensor_parameter_38: name = "parameter_38" + original_name = "conv2d_62.w_0" shape = [64, 32, 1, 1] dtype = "float32" min_val = float("-6.65054e-39") @@ -366,6 +405,7 @@ class Program_weight_tensor_parameter_38: class Program_weight_tensor_parameter_39: name = "parameter_39" + original_name = "conv2d_68.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -375,6 +415,7 @@ class Program_weight_tensor_parameter_39: class Program_weight_tensor_parameter_40: name = "parameter_40" + original_name = "conv2d_68.w_0" shape = [32, 32, 1, 3] dtype = "float32" min_val = float("-6.65236e-39") @@ -385,6 +426,7 @@ class Program_weight_tensor_parameter_40: class Program_weight_tensor_parameter_41: name = "parameter_41" + original_name = "conv2d_65.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -394,6 +436,7 @@ class Program_weight_tensor_parameter_41: class Program_weight_tensor_parameter_42: name = "parameter_42" + original_name = "conv2d_65.w_0" shape = [32, 32, 3, 1] dtype = "float32" min_val = float("-6.64247e-39") @@ -404,6 +447,7 @@ class Program_weight_tensor_parameter_42: class Program_weight_tensor_parameter_43: name = "parameter_43" + original_name = "conv2d_71.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -413,6 +457,7 @@ class Program_weight_tensor_parameter_43: class Program_weight_tensor_parameter_44: name = "parameter_44" + original_name = "conv2d_71.w_0" shape = [32, 32, 3, 3] dtype = "float32" min_val = float("-6.65532e-39") @@ -423,6 +468,7 @@ class Program_weight_tensor_parameter_44: class Program_weight_tensor_parameter_45: name = "parameter_45" + original_name = "conv2d_67.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -432,6 +478,7 @@ class Program_weight_tensor_parameter_45: class Program_weight_tensor_parameter_46: name = "parameter_46" + original_name = "conv2d_67.w_0" shape = [32, 32, 1, 5] dtype = "float32" min_val = float("-6.65569e-39") @@ -442,6 +489,7 @@ class Program_weight_tensor_parameter_46: class Program_weight_tensor_parameter_47: name = "parameter_47" + original_name = "conv2d_64.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -451,6 +499,7 @@ class Program_weight_tensor_parameter_47: class Program_weight_tensor_parameter_48: name = "parameter_48" + original_name = "conv2d_64.w_0" shape = [32, 32, 5, 1] dtype = "float32" min_val = float("-6.65616e-39") @@ -461,6 +510,7 @@ class Program_weight_tensor_parameter_48: class Program_weight_tensor_parameter_49: name = "parameter_49" + original_name = "conv2d_70.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -470,6 +520,7 @@ class Program_weight_tensor_parameter_49: class Program_weight_tensor_parameter_50: name = "parameter_50" + original_name = "conv2d_70.w_0" shape = [32, 32, 5, 5] dtype = "float32" min_val = float("-6.65694e-39") @@ -480,6 +531,7 @@ class Program_weight_tensor_parameter_50: class Program_weight_tensor_parameter_51: name = "parameter_51" + original_name = "conv2d_66.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -489,6 +541,7 @@ class Program_weight_tensor_parameter_51: class Program_weight_tensor_parameter_52: name = "parameter_52" + original_name = "conv2d_66.w_0" shape = [32, 32, 1, 7] dtype = "float32" min_val = float("-6.65593e-39") @@ -499,6 +552,7 @@ class Program_weight_tensor_parameter_52: class Program_weight_tensor_parameter_53: name = "parameter_53" + original_name = "conv2d_63.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -508,6 +562,7 @@ class Program_weight_tensor_parameter_53: class Program_weight_tensor_parameter_54: name = "parameter_54" + original_name = "conv2d_63.w_0" shape = [32, 32, 7, 1] dtype = "float32" min_val = float("-6.65725e-39") @@ -518,6 +573,7 @@ class Program_weight_tensor_parameter_54: class Program_weight_tensor_parameter_55: name = "parameter_55" + original_name = "conv2d_69.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -527,6 +583,7 @@ class Program_weight_tensor_parameter_55: class Program_weight_tensor_parameter_56: name = "parameter_56" + original_name = "conv2d_69.w_0" shape = [32, 32, 7, 7] dtype = "float32" min_val = float("-6.65748e-39") @@ -537,6 +594,7 @@ class Program_weight_tensor_parameter_56: class Program_weight_tensor_parameter_57: name = "parameter_57" + original_name = "conv2d_61.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -546,6 +604,7 @@ class Program_weight_tensor_parameter_57: class Program_weight_tensor_parameter_58: name = "parameter_58" + original_name = "conv2d_61.w_0" shape = [32, 64, 1, 1] dtype = "float32" min_val = float("-6.65427e-39") @@ -556,6 +615,7 @@ class Program_weight_tensor_parameter_58: class Program_weight_tensor_parameter_59: name = "parameter_59" + original_name = "batch_norm2d_42.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -565,6 +625,7 @@ class Program_weight_tensor_parameter_59: class Program_weight_tensor_parameter_60: name = "parameter_60" + original_name = "batch_norm2d_42.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -574,6 +635,7 @@ class Program_weight_tensor_parameter_60: class Program_weight_tensor_parameter_61: name = "parameter_61" + original_name = "batch_norm2d_42.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -583,6 +645,7 @@ class Program_weight_tensor_parameter_61: class Program_weight_tensor_parameter_62: name = "parameter_62" + original_name = "batch_norm2d_42.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -592,6 +655,7 @@ class Program_weight_tensor_parameter_62: class Program_weight_tensor_parameter_63: name = "parameter_63" + original_name = "conv2d_73.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -601,6 +665,7 @@ class Program_weight_tensor_parameter_63: class Program_weight_tensor_parameter_64: name = "parameter_64" + original_name = "conv2d_73.w_0" shape = [64, 32, 1, 1] dtype = "float32" min_val = float("-6.64285e-39") @@ -611,6 +676,7 @@ class Program_weight_tensor_parameter_64: class Program_weight_tensor_parameter_65: name = "parameter_65" + original_name = "conv2d_79.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -620,6 +686,7 @@ class Program_weight_tensor_parameter_65: class Program_weight_tensor_parameter_66: name = "parameter_66" + original_name = "conv2d_79.w_0" shape = [32, 32, 1, 3] dtype = "float32" min_val = float("-6.65134e-39") @@ -630,6 +697,7 @@ class Program_weight_tensor_parameter_66: class Program_weight_tensor_parameter_67: name = "parameter_67" + original_name = "conv2d_76.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -639,6 +707,7 @@ class Program_weight_tensor_parameter_67: class Program_weight_tensor_parameter_68: name = "parameter_68" + original_name = "conv2d_76.w_0" shape = [32, 32, 3, 1] dtype = "float32" min_val = float("-6.65222e-39") @@ -649,6 +718,7 @@ class Program_weight_tensor_parameter_68: class Program_weight_tensor_parameter_69: name = "parameter_69" + original_name = "conv2d_82.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -658,6 +728,7 @@ class Program_weight_tensor_parameter_69: class Program_weight_tensor_parameter_70: name = "parameter_70" + original_name = "conv2d_82.w_0" shape = [32, 32, 3, 3] dtype = "float32" min_val = float("-6.65397e-39") @@ -668,6 +739,7 @@ class Program_weight_tensor_parameter_70: class Program_weight_tensor_parameter_71: name = "parameter_71" + original_name = "conv2d_78.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -677,6 +749,7 @@ class Program_weight_tensor_parameter_71: class Program_weight_tensor_parameter_72: name = "parameter_72" + original_name = "conv2d_78.w_0" shape = [32, 32, 1, 5] dtype = "float32" min_val = float("-6.65696e-39") @@ -687,6 +760,7 @@ class Program_weight_tensor_parameter_72: class Program_weight_tensor_parameter_73: name = "parameter_73" + original_name = "conv2d_75.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -696,6 +770,7 @@ class Program_weight_tensor_parameter_73: class Program_weight_tensor_parameter_74: name = "parameter_74" + original_name = "conv2d_75.w_0" shape = [32, 32, 5, 1] dtype = "float32" min_val = float("-6.64865e-39") @@ -706,6 +781,7 @@ class Program_weight_tensor_parameter_74: class Program_weight_tensor_parameter_75: name = "parameter_75" + original_name = "conv2d_81.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -715,6 +791,7 @@ class Program_weight_tensor_parameter_75: class Program_weight_tensor_parameter_76: name = "parameter_76" + original_name = "conv2d_81.w_0" shape = [32, 32, 5, 5] dtype = "float32" min_val = float("-6.65717e-39") @@ -725,6 +802,7 @@ class Program_weight_tensor_parameter_76: class Program_weight_tensor_parameter_77: name = "parameter_77" + original_name = "conv2d_77.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -734,6 +812,7 @@ class Program_weight_tensor_parameter_77: class Program_weight_tensor_parameter_78: name = "parameter_78" + original_name = "conv2d_77.w_0" shape = [32, 32, 1, 7] dtype = "float32" min_val = float("-6.65748e-39") @@ -744,6 +823,7 @@ class Program_weight_tensor_parameter_78: class Program_weight_tensor_parameter_79: name = "parameter_79" + original_name = "conv2d_74.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -753,6 +833,7 @@ class Program_weight_tensor_parameter_79: class Program_weight_tensor_parameter_80: name = "parameter_80" + original_name = "conv2d_74.w_0" shape = [32, 32, 7, 1] dtype = "float32" min_val = float("-6.65479e-39") @@ -763,6 +844,7 @@ class Program_weight_tensor_parameter_80: class Program_weight_tensor_parameter_81: name = "parameter_81" + original_name = "conv2d_80.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -772,6 +854,7 @@ class Program_weight_tensor_parameter_81: class Program_weight_tensor_parameter_82: name = "parameter_82" + original_name = "conv2d_80.w_0" shape = [32, 32, 7, 7] dtype = "float32" min_val = float("-6.65693e-39") @@ -782,6 +865,7 @@ class Program_weight_tensor_parameter_82: class Program_weight_tensor_parameter_83: name = "parameter_83" + original_name = "conv2d_72.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -791,6 +875,7 @@ class Program_weight_tensor_parameter_83: class Program_weight_tensor_parameter_84: name = "parameter_84" + original_name = "conv2d_72.w_0" shape = [32, 64, 1, 1] dtype = "float32" min_val = float("-6.6525e-39") @@ -801,6 +886,7 @@ class Program_weight_tensor_parameter_84: class Program_weight_tensor_parameter_85: name = "parameter_85" + original_name = "batch_norm2d_43.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -810,6 +896,7 @@ class Program_weight_tensor_parameter_85: class Program_weight_tensor_parameter_86: name = "parameter_86" + original_name = "batch_norm2d_43.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -819,6 +906,7 @@ class Program_weight_tensor_parameter_86: class Program_weight_tensor_parameter_87: name = "parameter_87" + original_name = "batch_norm2d_43.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -828,6 +916,7 @@ class Program_weight_tensor_parameter_87: class Program_weight_tensor_parameter_88: name = "parameter_88" + original_name = "batch_norm2d_43.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -837,6 +926,7 @@ class Program_weight_tensor_parameter_88: class Program_weight_tensor_parameter_89: name = "parameter_89" + original_name = "conv2d_84.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -846,6 +936,7 @@ class Program_weight_tensor_parameter_89: class Program_weight_tensor_parameter_90: name = "parameter_90" + original_name = "conv2d_84.w_0" shape = [64, 32, 1, 1] dtype = "float32" min_val = float("-6.65357e-39") @@ -856,6 +947,7 @@ class Program_weight_tensor_parameter_90: class Program_weight_tensor_parameter_91: name = "parameter_91" + original_name = "conv2d_90.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -865,6 +957,7 @@ class Program_weight_tensor_parameter_91: class Program_weight_tensor_parameter_92: name = "parameter_92" + original_name = "conv2d_90.w_0" shape = [32, 32, 1, 3] dtype = "float32" min_val = float("-6.65638e-39") @@ -875,6 +968,7 @@ class Program_weight_tensor_parameter_92: class Program_weight_tensor_parameter_93: name = "parameter_93" + original_name = "conv2d_87.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -884,6 +978,7 @@ class Program_weight_tensor_parameter_93: class Program_weight_tensor_parameter_94: name = "parameter_94" + original_name = "conv2d_87.w_0" shape = [32, 32, 3, 1] dtype = "float32" min_val = float("-6.65094e-39") @@ -894,6 +989,7 @@ class Program_weight_tensor_parameter_94: class Program_weight_tensor_parameter_95: name = "parameter_95" + original_name = "conv2d_93.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -903,6 +999,7 @@ class Program_weight_tensor_parameter_95: class Program_weight_tensor_parameter_96: name = "parameter_96" + original_name = "conv2d_93.w_0" shape = [32, 32, 3, 3] dtype = "float32" min_val = float("-6.65647e-39") @@ -913,6 +1010,7 @@ class Program_weight_tensor_parameter_96: class Program_weight_tensor_parameter_97: name = "parameter_97" + original_name = "conv2d_89.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -922,6 +1020,7 @@ class Program_weight_tensor_parameter_97: class Program_weight_tensor_parameter_98: name = "parameter_98" + original_name = "conv2d_89.w_0" shape = [32, 32, 1, 5] dtype = "float32" min_val = float("-6.65711e-39") @@ -932,6 +1031,7 @@ class Program_weight_tensor_parameter_98: class Program_weight_tensor_parameter_99: name = "parameter_99" + original_name = "conv2d_86.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -941,6 +1041,7 @@ class Program_weight_tensor_parameter_99: class Program_weight_tensor_parameter_100: name = "parameter_100" + original_name = "conv2d_86.w_0" shape = [32, 32, 5, 1] dtype = "float32" min_val = float("-6.64361e-39") @@ -951,6 +1052,7 @@ class Program_weight_tensor_parameter_100: class Program_weight_tensor_parameter_101: name = "parameter_101" + original_name = "conv2d_92.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -960,6 +1062,7 @@ class Program_weight_tensor_parameter_101: class Program_weight_tensor_parameter_102: name = "parameter_102" + original_name = "conv2d_92.w_0" shape = [32, 32, 5, 5] dtype = "float32" min_val = float("-6.65734e-39") @@ -970,6 +1073,7 @@ class Program_weight_tensor_parameter_102: class Program_weight_tensor_parameter_103: name = "parameter_103" + original_name = "conv2d_88.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -979,6 +1083,7 @@ class Program_weight_tensor_parameter_103: class Program_weight_tensor_parameter_104: name = "parameter_104" + original_name = "conv2d_88.w_0" shape = [32, 32, 1, 7] dtype = "float32" min_val = float("-6.65637e-39") @@ -989,6 +1094,7 @@ class Program_weight_tensor_parameter_104: class Program_weight_tensor_parameter_105: name = "parameter_105" + original_name = "conv2d_85.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -998,6 +1104,7 @@ class Program_weight_tensor_parameter_105: class Program_weight_tensor_parameter_106: name = "parameter_106" + original_name = "conv2d_85.w_0" shape = [32, 32, 7, 1] dtype = "float32" min_val = float("-6.65683e-39") @@ -1008,6 +1115,7 @@ class Program_weight_tensor_parameter_106: class Program_weight_tensor_parameter_107: name = "parameter_107" + original_name = "conv2d_91.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -1017,6 +1125,7 @@ class Program_weight_tensor_parameter_107: class Program_weight_tensor_parameter_108: name = "parameter_108" + original_name = "conv2d_91.w_0" shape = [32, 32, 7, 7] dtype = "float32" min_val = float("-6.65734e-39") @@ -1027,6 +1136,7 @@ class Program_weight_tensor_parameter_108: class Program_weight_tensor_parameter_109: name = "parameter_109" + original_name = "conv2d_83.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -1036,6 +1146,7 @@ class Program_weight_tensor_parameter_109: class Program_weight_tensor_parameter_110: name = "parameter_110" + original_name = "conv2d_83.w_0" shape = [32, 64, 1, 1] dtype = "float32" min_val = float("-6.6536e-39") @@ -1046,6 +1157,7 @@ class Program_weight_tensor_parameter_110: class Program_weight_tensor_parameter_111: name = "parameter_111" + original_name = "batch_norm2d_44.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -1055,6 +1167,7 @@ class Program_weight_tensor_parameter_111: class Program_weight_tensor_parameter_112: name = "parameter_112" + original_name = "batch_norm2d_44.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -1064,6 +1177,7 @@ class Program_weight_tensor_parameter_112: class Program_weight_tensor_parameter_113: name = "parameter_113" + original_name = "batch_norm2d_44.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -1073,6 +1187,7 @@ class Program_weight_tensor_parameter_113: class Program_weight_tensor_parameter_114: name = "parameter_114" + original_name = "batch_norm2d_44.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -1082,6 +1197,7 @@ class Program_weight_tensor_parameter_114: class Program_weight_tensor_parameter_115: name = "parameter_115" + original_name = "conv2d_95.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -1091,6 +1207,7 @@ class Program_weight_tensor_parameter_115: class Program_weight_tensor_parameter_116: name = "parameter_116" + original_name = "conv2d_95.w_0" shape = [64, 32, 1, 1] dtype = "float32" min_val = float("-6.65418e-39") @@ -1101,6 +1218,7 @@ class Program_weight_tensor_parameter_116: class Program_weight_tensor_parameter_117: name = "parameter_117" + original_name = "conv2d_101.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -1110,6 +1228,7 @@ class Program_weight_tensor_parameter_117: class Program_weight_tensor_parameter_118: name = "parameter_118" + original_name = "conv2d_101.w_0" shape = [32, 32, 1, 3] dtype = "float32" min_val = float("-6.65685e-39") @@ -1120,6 +1239,7 @@ class Program_weight_tensor_parameter_118: class Program_weight_tensor_parameter_119: name = "parameter_119" + original_name = "conv2d_98.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -1129,6 +1249,7 @@ class Program_weight_tensor_parameter_119: class Program_weight_tensor_parameter_120: name = "parameter_120" + original_name = "conv2d_98.w_0" shape = [32, 32, 3, 1] dtype = "float32" min_val = float("-6.64128e-39") @@ -1139,6 +1260,7 @@ class Program_weight_tensor_parameter_120: class Program_weight_tensor_parameter_121: name = "parameter_121" + original_name = "conv2d_104.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -1148,6 +1270,7 @@ class Program_weight_tensor_parameter_121: class Program_weight_tensor_parameter_122: name = "parameter_122" + original_name = "conv2d_104.w_0" shape = [32, 32, 3, 3] dtype = "float32" min_val = float("-6.65739e-39") @@ -1158,6 +1281,7 @@ class Program_weight_tensor_parameter_122: class Program_weight_tensor_parameter_123: name = "parameter_123" + original_name = "conv2d_100.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -1167,6 +1291,7 @@ class Program_weight_tensor_parameter_123: class Program_weight_tensor_parameter_124: name = "parameter_124" + original_name = "conv2d_100.w_0" shape = [32, 32, 1, 5] dtype = "float32" min_val = float("-6.65712e-39") @@ -1177,6 +1302,7 @@ class Program_weight_tensor_parameter_124: class Program_weight_tensor_parameter_125: name = "parameter_125" + original_name = "conv2d_97.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -1186,6 +1312,7 @@ class Program_weight_tensor_parameter_125: class Program_weight_tensor_parameter_126: name = "parameter_126" + original_name = "conv2d_97.w_0" shape = [32, 32, 5, 1] dtype = "float32" min_val = float("-6.65622e-39") @@ -1196,6 +1323,7 @@ class Program_weight_tensor_parameter_126: class Program_weight_tensor_parameter_127: name = "parameter_127" + original_name = "conv2d_103.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -1205,6 +1333,7 @@ class Program_weight_tensor_parameter_127: class Program_weight_tensor_parameter_128: name = "parameter_128" + original_name = "conv2d_103.w_0" shape = [32, 32, 5, 5] dtype = "float32" min_val = float("-6.65701e-39") @@ -1215,6 +1344,7 @@ class Program_weight_tensor_parameter_128: class Program_weight_tensor_parameter_129: name = "parameter_129" + original_name = "conv2d_99.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -1224,6 +1354,7 @@ class Program_weight_tensor_parameter_129: class Program_weight_tensor_parameter_130: name = "parameter_130" + original_name = "conv2d_99.w_0" shape = [32, 32, 1, 7] dtype = "float32" min_val = float("-6.65653e-39") @@ -1234,6 +1365,7 @@ class Program_weight_tensor_parameter_130: class Program_weight_tensor_parameter_131: name = "parameter_131" + original_name = "conv2d_96.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -1243,6 +1375,7 @@ class Program_weight_tensor_parameter_131: class Program_weight_tensor_parameter_132: name = "parameter_132" + original_name = "conv2d_96.w_0" shape = [32, 32, 7, 1] dtype = "float32" min_val = float("-6.65454e-39") @@ -1253,6 +1386,7 @@ class Program_weight_tensor_parameter_132: class Program_weight_tensor_parameter_133: name = "parameter_133" + original_name = "conv2d_102.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -1262,6 +1396,7 @@ class Program_weight_tensor_parameter_133: class Program_weight_tensor_parameter_134: name = "parameter_134" + original_name = "conv2d_102.w_0" shape = [32, 32, 7, 7] dtype = "float32" min_val = float("-6.65707e-39") @@ -1272,6 +1407,7 @@ class Program_weight_tensor_parameter_134: class Program_weight_tensor_parameter_135: name = "parameter_135" + original_name = "conv2d_94.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -1281,6 +1417,7 @@ class Program_weight_tensor_parameter_135: class Program_weight_tensor_parameter_136: name = "parameter_136" + original_name = "conv2d_94.w_0" shape = [32, 64, 1, 1] dtype = "float32" min_val = float("-6.65348e-39") @@ -1291,6 +1428,7 @@ class Program_weight_tensor_parameter_136: class Program_weight_tensor_parameter_137: name = "parameter_137" + original_name = "conv2d_60.w_0" shape = [64, 64, 9, 9] dtype = "float32" min_val = float("-0.635527") @@ -1302,6 +1440,7 @@ class Program_weight_tensor_parameter_137: class Program_weight_tensor_parameter_138: name = "parameter_138" + original_name = "conv2d_56.w_0" shape = [64, 64, 9, 9] dtype = "float32" min_val = float("-0.629213") @@ -1313,6 +1452,7 @@ class Program_weight_tensor_parameter_138: class Program_weight_tensor_parameter_139: name = "parameter_139" + original_name = "conv2d_52.w_0" shape = [64, 64, 9, 9] dtype = "float32" min_val = float("-0.552784") @@ -1324,6 +1464,7 @@ class Program_weight_tensor_parameter_139: class Program_weight_tensor_parameter_140: name = "parameter_140" + original_name = "conv2d_48.w_0" shape = [64, 64, 9, 9] dtype = "float32" min_val = float("-0.349797") @@ -1335,6 +1476,7 @@ class Program_weight_tensor_parameter_140: class Program_weight_tensor_parameter_141: name = "parameter_141" + original_name = "conv2d_59.w_0" shape = [64, 64, 3, 3] dtype = "float32" min_val = float("-0.354901") @@ -1346,6 +1488,7 @@ class Program_weight_tensor_parameter_141: class Program_weight_tensor_parameter_142: name = "parameter_142" + original_name = "conv2d_55.w_0" shape = [64, 64, 3, 3] dtype = "float32" min_val = float("-0.260383") @@ -1357,6 +1500,7 @@ class Program_weight_tensor_parameter_142: class Program_weight_tensor_parameter_143: name = "parameter_143" + original_name = "conv2d_51.w_0" shape = [64, 64, 3, 3] dtype = "float32" min_val = float("-0.789898") @@ -1368,6 +1512,7 @@ class Program_weight_tensor_parameter_143: class Program_weight_tensor_parameter_144: name = "parameter_144" + original_name = "conv2d_47.w_0" shape = [64, 256, 9, 9] dtype = "float32" min_val = float("-0.334457") @@ -1379,6 +1524,7 @@ class Program_weight_tensor_parameter_144: class Program_weight_tensor_parameter_145: name = "parameter_145" + original_name = "conv2d_50.w_0" shape = [64, 256, 9, 9] dtype = "float32" min_val = float("-0.341071") @@ -1390,6 +1536,7 @@ class Program_weight_tensor_parameter_145: class Program_weight_tensor_parameter_146: name = "parameter_146" + original_name = "conv2d_54.w_0" shape = [64, 256, 9, 9] dtype = "float32" min_val = float("-0.375846") @@ -1401,6 +1548,7 @@ class Program_weight_tensor_parameter_146: class Program_weight_tensor_parameter_147: name = "parameter_147" + original_name = "conv2d_58.w_0" shape = [64, 256, 9, 9] dtype = "float32" min_val = float("-0.141057") @@ -1412,6 +1560,7 @@ class Program_weight_tensor_parameter_147: class Program_weight_tensor_parameter_148: name = "parameter_148" + original_name = "conv2d_46.w_0" shape = [256, 256, 1, 1] dtype = "float32" min_val = float("-0.69766") @@ -1423,6 +1572,7 @@ class Program_weight_tensor_parameter_148: class Program_weight_tensor_parameter_149: name = "parameter_149" + original_name = "conv2d_49.w_0" shape = [256, 512, 1, 1] dtype = "float32" min_val = float("-0.408204") @@ -1434,6 +1584,7 @@ class Program_weight_tensor_parameter_149: class Program_weight_tensor_parameter_150: name = "parameter_150" + original_name = "conv2d_53.w_0" shape = [256, 768, 1, 1] dtype = "float32" min_val = float("-1.10542") @@ -1445,6 +1596,7 @@ class Program_weight_tensor_parameter_150: class Program_weight_tensor_parameter_151: name = "parameter_151" + original_name = "conv2d_57.w_0" shape = [256, 1024, 1, 1] dtype = "float32" min_val = float("-0.8864") @@ -1456,6 +1608,7 @@ class Program_weight_tensor_parameter_151: class Program_weight_tensor_parameter_152: name = "parameter_152" + original_name = "conv2d_45.b_0" shape = [1024] dtype = "float32" min_val = float("-0.0438569") @@ -1467,6 +1620,7 @@ class Program_weight_tensor_parameter_152: class Program_weight_tensor_parameter_153: name = "parameter_153" + original_name = "conv2d_45.w_0" shape = [1024, 1024, 1, 1] dtype = "float32" min_val = float("-1.32631") @@ -1478,6 +1632,7 @@ class Program_weight_tensor_parameter_153: class Program_weight_tensor_parameter_154: name = "parameter_154" + original_name = "batch_norm2d_40.b_0" shape = [1024] dtype = "float32" min_val = float("-16.7742") @@ -1489,6 +1644,7 @@ class Program_weight_tensor_parameter_154: class Program_weight_tensor_parameter_155: name = "parameter_155" + original_name = "batch_norm2d_40.w_0" shape = [1024] dtype = "float32" min_val = float("2.15554") @@ -1500,6 +1656,7 @@ class Program_weight_tensor_parameter_155: class Program_weight_tensor_parameter_156: name = "parameter_156" + original_name = "batch_norm2d_40.w_2" shape = [1024] dtype = "float32" min_val = float("5.60519e-45") @@ -1511,6 +1668,7 @@ class Program_weight_tensor_parameter_156: class Program_weight_tensor_parameter_157: name = "parameter_157" + original_name = "batch_norm2d_40.w_1" shape = [1024] dtype = "float32" min_val = float("-16.4235") @@ -1522,6 +1680,7 @@ class Program_weight_tensor_parameter_157: class Program_weight_tensor_parameter_158: name = "parameter_158" + original_name = "conv2d_44.w_0" shape = [1024, 2112, 1, 1] dtype = "float32" min_val = float("-1.07314") @@ -1533,6 +1692,7 @@ class Program_weight_tensor_parameter_158: class Program_weight_tensor_parameter_159: name = "parameter_159" + original_name = "batch_norm2d_39.b_0" shape = [224] dtype = "float32" min_val = float("-3.63551") @@ -1544,6 +1704,7 @@ class Program_weight_tensor_parameter_159: class Program_weight_tensor_parameter_160: name = "parameter_160" + original_name = "batch_norm2d_39.w_0" shape = [224] dtype = "float32" min_val = float("0.671715") @@ -1555,6 +1716,7 @@ class Program_weight_tensor_parameter_160: class Program_weight_tensor_parameter_161: name = "parameter_161" + original_name = "batch_norm2d_39.w_2" shape = [224] dtype = "float32" min_val = float("5.60519e-45") @@ -1566,6 +1728,7 @@ class Program_weight_tensor_parameter_161: class Program_weight_tensor_parameter_162: name = "parameter_162" + original_name = "batch_norm2d_39.w_1" shape = [224] dtype = "float32" min_val = float("-7.65448") @@ -1577,6 +1740,7 @@ class Program_weight_tensor_parameter_162: class Program_weight_tensor_parameter_163: name = "parameter_163" + original_name = "conv2d_43.w_0" shape = [224, 224, 3, 3] dtype = "float32" min_val = float("-0.45728") @@ -1588,6 +1752,7 @@ class Program_weight_tensor_parameter_163: class Program_weight_tensor_parameter_164: name = "parameter_164" + original_name = "batch_norm2d_38.b_0" shape = [224] dtype = "float32" min_val = float("-3.0171") @@ -1599,6 +1764,7 @@ class Program_weight_tensor_parameter_164: class Program_weight_tensor_parameter_165: name = "parameter_165" + original_name = "batch_norm2d_38.w_0" shape = [224] dtype = "float32" min_val = float("0.242337") @@ -1610,6 +1776,7 @@ class Program_weight_tensor_parameter_165: class Program_weight_tensor_parameter_166: name = "parameter_166" + original_name = "batch_norm2d_38.w_2" shape = [224] dtype = "float32" min_val = float("5.60519e-45") @@ -1621,6 +1788,7 @@ class Program_weight_tensor_parameter_166: class Program_weight_tensor_parameter_167: name = "parameter_167" + original_name = "batch_norm2d_38.w_1" shape = [224] dtype = "float32" min_val = float("-4.80738") @@ -1632,6 +1800,7 @@ class Program_weight_tensor_parameter_167: class Program_weight_tensor_parameter_168: name = "parameter_168" + original_name = "conv2d_42.w_0" shape = [224, 224, 3, 3] dtype = "float32" min_val = float("-0.433406") @@ -1643,6 +1812,7 @@ class Program_weight_tensor_parameter_168: class Program_weight_tensor_parameter_169: name = "parameter_169" + original_name = "batch_norm2d_37.b_0" shape = [224] dtype = "float32" min_val = float("-3.33922") @@ -1654,6 +1824,7 @@ class Program_weight_tensor_parameter_169: class Program_weight_tensor_parameter_170: name = "parameter_170" + original_name = "batch_norm2d_37.w_0" shape = [224] dtype = "float32" min_val = float("0.208764") @@ -1665,6 +1836,7 @@ class Program_weight_tensor_parameter_170: class Program_weight_tensor_parameter_171: name = "parameter_171" + original_name = "batch_norm2d_37.w_2" shape = [224] dtype = "float32" min_val = float("5.60519e-45") @@ -1676,6 +1848,7 @@ class Program_weight_tensor_parameter_171: class Program_weight_tensor_parameter_172: name = "parameter_172" + original_name = "batch_norm2d_37.w_1" shape = [224] dtype = "float32" min_val = float("-6.96837") @@ -1687,6 +1860,7 @@ class Program_weight_tensor_parameter_172: class Program_weight_tensor_parameter_173: name = "parameter_173" + original_name = "conv2d_41.w_0" shape = [224, 224, 3, 3] dtype = "float32" min_val = float("-0.466328") @@ -1698,6 +1872,7 @@ class Program_weight_tensor_parameter_173: class Program_weight_tensor_parameter_174: name = "parameter_174" + original_name = "batch_norm2d_36.b_0" shape = [224] dtype = "float32" min_val = float("-4.14337") @@ -1709,6 +1884,7 @@ class Program_weight_tensor_parameter_174: class Program_weight_tensor_parameter_175: name = "parameter_175" + original_name = "batch_norm2d_36.w_0" shape = [224] dtype = "float32" min_val = float("0.253652") @@ -1720,6 +1896,7 @@ class Program_weight_tensor_parameter_175: class Program_weight_tensor_parameter_176: name = "parameter_176" + original_name = "batch_norm2d_36.w_2" shape = [224] dtype = "float32" min_val = float("5.60519e-45") @@ -1731,6 +1908,7 @@ class Program_weight_tensor_parameter_176: class Program_weight_tensor_parameter_177: name = "parameter_177" + original_name = "batch_norm2d_36.w_1" shape = [224] dtype = "float32" min_val = float("-7.19195") @@ -1742,6 +1920,7 @@ class Program_weight_tensor_parameter_177: class Program_weight_tensor_parameter_178: name = "parameter_178" + original_name = "conv2d_40.w_0" shape = [224, 224, 3, 3] dtype = "float32" min_val = float("-0.524839") @@ -1753,6 +1932,7 @@ class Program_weight_tensor_parameter_178: class Program_weight_tensor_parameter_179: name = "parameter_179" + original_name = "batch_norm2d_35.b_0" shape = [224] dtype = "float32" min_val = float("-5.58568") @@ -1764,6 +1944,7 @@ class Program_weight_tensor_parameter_179: class Program_weight_tensor_parameter_180: name = "parameter_180" + original_name = "batch_norm2d_35.w_0" shape = [224] dtype = "float32" min_val = float("0.469686") @@ -1775,6 +1956,7 @@ class Program_weight_tensor_parameter_180: class Program_weight_tensor_parameter_181: name = "parameter_181" + original_name = "batch_norm2d_35.w_2" shape = [224] dtype = "float32" min_val = float("5.60519e-45") @@ -1786,6 +1968,7 @@ class Program_weight_tensor_parameter_181: class Program_weight_tensor_parameter_182: name = "parameter_182" + original_name = "batch_norm2d_35.w_1" shape = [224] dtype = "float32" min_val = float("-11.2245") @@ -1797,6 +1980,7 @@ class Program_weight_tensor_parameter_182: class Program_weight_tensor_parameter_183: name = "parameter_183" + original_name = "conv2d_39.w_0" shape = [224, 224, 3, 3] dtype = "float32" min_val = float("-0.572262") @@ -1808,6 +1992,7 @@ class Program_weight_tensor_parameter_183: class Program_weight_tensor_parameter_184: name = "parameter_184" + original_name = "batch_norm2d_34.b_0" shape = [224] dtype = "float32" min_val = float("-10.3708") @@ -1819,6 +2004,7 @@ class Program_weight_tensor_parameter_184: class Program_weight_tensor_parameter_185: name = "parameter_185" + original_name = "batch_norm2d_34.w_0" shape = [224] dtype = "float32" min_val = float("0.323924") @@ -1830,6 +2016,7 @@ class Program_weight_tensor_parameter_185: class Program_weight_tensor_parameter_186: name = "parameter_186" + original_name = "batch_norm2d_34.w_2" shape = [224] dtype = "float32" min_val = float("5.60519e-45") @@ -1841,6 +2028,7 @@ class Program_weight_tensor_parameter_186: class Program_weight_tensor_parameter_187: name = "parameter_187" + original_name = "batch_norm2d_34.w_1" shape = [224] dtype = "float32" min_val = float("-10.4313") @@ -1852,6 +2040,7 @@ class Program_weight_tensor_parameter_187: class Program_weight_tensor_parameter_188: name = "parameter_188" + original_name = "conv2d_38.w_0" shape = [224, 768, 3, 3] dtype = "float32" min_val = float("-0.487017") @@ -1863,6 +2052,7 @@ class Program_weight_tensor_parameter_188: class Program_weight_tensor_parameter_189: name = "parameter_189" + original_name = "batch_norm2d_33.b_0" shape = [768] dtype = "float32" min_val = float("-2.27951") @@ -1874,6 +2064,7 @@ class Program_weight_tensor_parameter_189: class Program_weight_tensor_parameter_190: name = "parameter_190" + original_name = "batch_norm2d_33.w_0" shape = [768] dtype = "float32" min_val = float("-0.00295642") @@ -1885,6 +2076,7 @@ class Program_weight_tensor_parameter_190: class Program_weight_tensor_parameter_191: name = "parameter_191" + original_name = "batch_norm2d_33.w_2" shape = [768] dtype = "float32" min_val = float("5.60519e-45") @@ -1896,6 +2088,7 @@ class Program_weight_tensor_parameter_191: class Program_weight_tensor_parameter_192: name = "parameter_192" + original_name = "batch_norm2d_33.w_1" shape = [768] dtype = "float32" min_val = float("-1.20276") @@ -1907,6 +2100,7 @@ class Program_weight_tensor_parameter_192: class Program_weight_tensor_parameter_193: name = "parameter_193" + original_name = "conv2d_37.w_0" shape = [768, 1, 3, 3] dtype = "float32" min_val = float("-0.384348") @@ -1918,6 +2112,7 @@ class Program_weight_tensor_parameter_193: class Program_weight_tensor_parameter_194: name = "parameter_194" + original_name = "conv2d_36.b_0" shape = [768] dtype = "float32" min_val = float("-0.168911") @@ -1929,6 +2124,7 @@ class Program_weight_tensor_parameter_194: class Program_weight_tensor_parameter_195: name = "parameter_195" + original_name = "conv2d_36.w_0" shape = [768, 768, 1, 1] dtype = "float32" min_val = float("-1.26229") @@ -1940,6 +2136,7 @@ class Program_weight_tensor_parameter_195: class Program_weight_tensor_parameter_196: name = "parameter_196" + original_name = "batch_norm2d_32.b_0" shape = [768] dtype = "float32" min_val = float("-12.1332") @@ -1951,6 +2148,7 @@ class Program_weight_tensor_parameter_196: class Program_weight_tensor_parameter_197: name = "parameter_197" + original_name = "batch_norm2d_32.w_0" shape = [768] dtype = "float32" min_val = float("0.0351857") @@ -1962,6 +2160,7 @@ class Program_weight_tensor_parameter_197: class Program_weight_tensor_parameter_198: name = "parameter_198" + original_name = "batch_norm2d_32.w_2" shape = [768] dtype = "float32" min_val = float("5.60519e-45") @@ -1973,6 +2172,7 @@ class Program_weight_tensor_parameter_198: class Program_weight_tensor_parameter_199: name = "parameter_199" + original_name = "batch_norm2d_32.w_1" shape = [768] dtype = "float32" min_val = float("-5.75011") @@ -1984,6 +2184,7 @@ class Program_weight_tensor_parameter_199: class Program_weight_tensor_parameter_200: name = "parameter_200" + original_name = "conv2d_35.w_0" shape = [768, 1920, 1, 1] dtype = "float32" min_val = float("-1.42284") @@ -1995,6 +2196,7 @@ class Program_weight_tensor_parameter_200: class Program_weight_tensor_parameter_201: name = "parameter_201" + original_name = "batch_norm2d_31.b_0" shape = [192] dtype = "float32" min_val = float("-2.9769") @@ -2006,6 +2208,7 @@ class Program_weight_tensor_parameter_201: class Program_weight_tensor_parameter_202: name = "parameter_202" + original_name = "batch_norm2d_31.w_0" shape = [192] dtype = "float32" min_val = float("0.271253") @@ -2017,6 +2220,7 @@ class Program_weight_tensor_parameter_202: class Program_weight_tensor_parameter_203: name = "parameter_203" + original_name = "batch_norm2d_31.w_2" shape = [192] dtype = "float32" min_val = float("5.60519e-45") @@ -2028,6 +2232,7 @@ class Program_weight_tensor_parameter_203: class Program_weight_tensor_parameter_204: name = "parameter_204" + original_name = "batch_norm2d_31.w_1" shape = [192] dtype = "float32" min_val = float("-3.67047") @@ -2039,6 +2244,7 @@ class Program_weight_tensor_parameter_204: class Program_weight_tensor_parameter_205: name = "parameter_205" + original_name = "conv2d_34.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.80071") @@ -2050,6 +2256,7 @@ class Program_weight_tensor_parameter_205: class Program_weight_tensor_parameter_206: name = "parameter_206" + original_name = "batch_norm2d_30.b_0" shape = [192] dtype = "float32" min_val = float("-2.39577") @@ -2061,6 +2268,7 @@ class Program_weight_tensor_parameter_206: class Program_weight_tensor_parameter_207: name = "parameter_207" + original_name = "batch_norm2d_30.w_0" shape = [192] dtype = "float32" min_val = float("0.00056985") @@ -2072,6 +2280,7 @@ class Program_weight_tensor_parameter_207: class Program_weight_tensor_parameter_208: name = "parameter_208" + original_name = "batch_norm2d_30.w_2" shape = [192] dtype = "float32" min_val = float("5.60519e-45") @@ -2083,6 +2292,7 @@ class Program_weight_tensor_parameter_208: class Program_weight_tensor_parameter_209: name = "parameter_209" + original_name = "batch_norm2d_30.w_1" shape = [192] dtype = "float32" min_val = float("-2.01196") @@ -2094,6 +2304,7 @@ class Program_weight_tensor_parameter_209: class Program_weight_tensor_parameter_210: name = "parameter_210" + original_name = "conv2d_33.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.624228") @@ -2105,6 +2316,7 @@ class Program_weight_tensor_parameter_210: class Program_weight_tensor_parameter_211: name = "parameter_211" + original_name = "batch_norm2d_29.b_0" shape = [192] dtype = "float32" min_val = float("-4.17926") @@ -2116,6 +2328,7 @@ class Program_weight_tensor_parameter_211: class Program_weight_tensor_parameter_212: name = "parameter_212" + original_name = "batch_norm2d_29.w_0" shape = [192] dtype = "float32" min_val = float("-0.000950568") @@ -2127,6 +2340,7 @@ class Program_weight_tensor_parameter_212: class Program_weight_tensor_parameter_213: name = "parameter_213" + original_name = "batch_norm2d_29.w_2" shape = [192] dtype = "float32" min_val = float("5.60519e-45") @@ -2138,6 +2352,7 @@ class Program_weight_tensor_parameter_213: class Program_weight_tensor_parameter_214: name = "parameter_214" + original_name = "batch_norm2d_29.w_1" shape = [192] dtype = "float32" min_val = float("-3.21464") @@ -2149,6 +2364,7 @@ class Program_weight_tensor_parameter_214: class Program_weight_tensor_parameter_215: name = "parameter_215" + original_name = "conv2d_32.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.541526") @@ -2160,6 +2376,7 @@ class Program_weight_tensor_parameter_215: class Program_weight_tensor_parameter_216: name = "parameter_216" + original_name = "batch_norm2d_28.b_0" shape = [192] dtype = "float32" min_val = float("-4.30215") @@ -2171,6 +2388,7 @@ class Program_weight_tensor_parameter_216: class Program_weight_tensor_parameter_217: name = "parameter_217" + original_name = "batch_norm2d_28.w_0" shape = [192] dtype = "float32" min_val = float("0.000491233") @@ -2182,6 +2400,7 @@ class Program_weight_tensor_parameter_217: class Program_weight_tensor_parameter_218: name = "parameter_218" + original_name = "batch_norm2d_28.w_2" shape = [192] dtype = "float32" min_val = float("5.60519e-45") @@ -2193,6 +2412,7 @@ class Program_weight_tensor_parameter_218: class Program_weight_tensor_parameter_219: name = "parameter_219" + original_name = "batch_norm2d_28.w_1" shape = [192] dtype = "float32" min_val = float("-3.42498") @@ -2204,6 +2424,7 @@ class Program_weight_tensor_parameter_219: class Program_weight_tensor_parameter_220: name = "parameter_220" + original_name = "conv2d_31.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.484216") @@ -2215,6 +2436,7 @@ class Program_weight_tensor_parameter_220: class Program_weight_tensor_parameter_221: name = "parameter_221" + original_name = "batch_norm2d_27.b_0" shape = [192] dtype = "float32" min_val = float("-4.21069") @@ -2226,6 +2448,7 @@ class Program_weight_tensor_parameter_221: class Program_weight_tensor_parameter_222: name = "parameter_222" + original_name = "batch_norm2d_27.w_0" shape = [192] dtype = "float32" min_val = float("0.0267023") @@ -2237,6 +2460,7 @@ class Program_weight_tensor_parameter_222: class Program_weight_tensor_parameter_223: name = "parameter_223" + original_name = "batch_norm2d_27.w_2" shape = [192] dtype = "float32" min_val = float("5.60519e-45") @@ -2248,6 +2472,7 @@ class Program_weight_tensor_parameter_223: class Program_weight_tensor_parameter_224: name = "parameter_224" + original_name = "batch_norm2d_27.w_1" shape = [192] dtype = "float32" min_val = float("-3.84438") @@ -2259,6 +2484,7 @@ class Program_weight_tensor_parameter_224: class Program_weight_tensor_parameter_225: name = "parameter_225" + original_name = "conv2d_30.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.840071") @@ -2270,6 +2496,7 @@ class Program_weight_tensor_parameter_225: class Program_weight_tensor_parameter_226: name = "parameter_226" + original_name = "batch_norm2d_26.b_0" shape = [192] dtype = "float32" min_val = float("-5.8111") @@ -2281,6 +2508,7 @@ class Program_weight_tensor_parameter_226: class Program_weight_tensor_parameter_227: name = "parameter_227" + original_name = "batch_norm2d_26.w_0" shape = [192] dtype = "float32" min_val = float("0.00973616") @@ -2292,6 +2520,7 @@ class Program_weight_tensor_parameter_227: class Program_weight_tensor_parameter_228: name = "parameter_228" + original_name = "batch_norm2d_26.w_2" shape = [192] dtype = "float32" min_val = float("5.60519e-45") @@ -2303,6 +2532,7 @@ class Program_weight_tensor_parameter_228: class Program_weight_tensor_parameter_229: name = "parameter_229" + original_name = "batch_norm2d_26.w_1" shape = [192] dtype = "float32" min_val = float("-4.99117") @@ -2314,6 +2544,7 @@ class Program_weight_tensor_parameter_229: class Program_weight_tensor_parameter_230: name = "parameter_230" + original_name = "conv2d_29.w_0" shape = [192, 768, 3, 3] dtype = "float32" min_val = float("-0.549061") @@ -2325,6 +2556,7 @@ class Program_weight_tensor_parameter_230: class Program_weight_tensor_parameter_231: name = "parameter_231" + original_name = "conv2d_28.b_0" shape = [768] dtype = "float32" min_val = float("-0.257451") @@ -2336,6 +2568,7 @@ class Program_weight_tensor_parameter_231: class Program_weight_tensor_parameter_232: name = "parameter_232" + original_name = "conv2d_28.w_0" shape = [768, 768, 1, 1] dtype = "float32" min_val = float("-1.64889") @@ -2347,6 +2580,7 @@ class Program_weight_tensor_parameter_232: class Program_weight_tensor_parameter_233: name = "parameter_233" + original_name = "batch_norm2d_25.b_0" shape = [768] dtype = "float32" min_val = float("-13.0182") @@ -2358,6 +2592,7 @@ class Program_weight_tensor_parameter_233: class Program_weight_tensor_parameter_234: name = "parameter_234" + original_name = "batch_norm2d_25.w_0" shape = [768] dtype = "float32" min_val = float("0.0103833") @@ -2369,6 +2604,7 @@ class Program_weight_tensor_parameter_234: class Program_weight_tensor_parameter_235: name = "parameter_235" + original_name = "batch_norm2d_25.w_2" shape = [768] dtype = "float32" min_val = float("5.60519e-45") @@ -2380,6 +2616,7 @@ class Program_weight_tensor_parameter_235: class Program_weight_tensor_parameter_236: name = "parameter_236" + original_name = "batch_norm2d_25.w_1" shape = [768] dtype = "float32" min_val = float("-11.3288") @@ -2391,6 +2628,7 @@ class Program_weight_tensor_parameter_236: class Program_weight_tensor_parameter_237: name = "parameter_237" + original_name = "conv2d_27.w_0" shape = [768, 1664, 1, 1] dtype = "float32" min_val = float("-0.912853") @@ -2402,6 +2640,7 @@ class Program_weight_tensor_parameter_237: class Program_weight_tensor_parameter_238: name = "parameter_238" + original_name = "batch_norm2d_24.b_0" shape = [192] dtype = "float32" min_val = float("-5.64347") @@ -2413,6 +2652,7 @@ class Program_weight_tensor_parameter_238: class Program_weight_tensor_parameter_239: name = "parameter_239" + original_name = "batch_norm2d_24.w_0" shape = [192] dtype = "float32" min_val = float("0.119573") @@ -2424,6 +2664,7 @@ class Program_weight_tensor_parameter_239: class Program_weight_tensor_parameter_240: name = "parameter_240" + original_name = "batch_norm2d_24.w_2" shape = [192] dtype = "float32" min_val = float("5.60519e-45") @@ -2435,6 +2676,7 @@ class Program_weight_tensor_parameter_240: class Program_weight_tensor_parameter_241: name = "parameter_241" + original_name = "batch_norm2d_24.w_1" shape = [192] dtype = "float32" min_val = float("-5.9052") @@ -2446,6 +2688,7 @@ class Program_weight_tensor_parameter_241: class Program_weight_tensor_parameter_242: name = "parameter_242" + original_name = "conv2d_26.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.4142") @@ -2457,6 +2700,7 @@ class Program_weight_tensor_parameter_242: class Program_weight_tensor_parameter_243: name = "parameter_243" + original_name = "batch_norm2d_23.b_0" shape = [192] dtype = "float32" min_val = float("-5.60742") @@ -2468,6 +2712,7 @@ class Program_weight_tensor_parameter_243: class Program_weight_tensor_parameter_244: name = "parameter_244" + original_name = "batch_norm2d_23.w_0" shape = [192] dtype = "float32" min_val = float("0.122475") @@ -2479,6 +2724,7 @@ class Program_weight_tensor_parameter_244: class Program_weight_tensor_parameter_245: name = "parameter_245" + original_name = "batch_norm2d_23.w_2" shape = [192] dtype = "float32" min_val = float("5.60519e-45") @@ -2490,6 +2736,7 @@ class Program_weight_tensor_parameter_245: class Program_weight_tensor_parameter_246: name = "parameter_246" + original_name = "batch_norm2d_23.w_1" shape = [192] dtype = "float32" min_val = float("-8.08676") @@ -2501,6 +2748,7 @@ class Program_weight_tensor_parameter_246: class Program_weight_tensor_parameter_247: name = "parameter_247" + original_name = "conv2d_25.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.439962") @@ -2512,6 +2760,7 @@ class Program_weight_tensor_parameter_247: class Program_weight_tensor_parameter_248: name = "parameter_248" + original_name = "batch_norm2d_22.b_0" shape = [192] dtype = "float32" min_val = float("-5.06254") @@ -2523,6 +2772,7 @@ class Program_weight_tensor_parameter_248: class Program_weight_tensor_parameter_249: name = "parameter_249" + original_name = "batch_norm2d_22.w_0" shape = [192] dtype = "float32" min_val = float("0.00260581") @@ -2534,6 +2784,7 @@ class Program_weight_tensor_parameter_249: class Program_weight_tensor_parameter_250: name = "parameter_250" + original_name = "batch_norm2d_22.w_2" shape = [192] dtype = "float32" min_val = float("5.60519e-45") @@ -2545,6 +2796,7 @@ class Program_weight_tensor_parameter_250: class Program_weight_tensor_parameter_251: name = "parameter_251" + original_name = "batch_norm2d_22.w_1" shape = [192] dtype = "float32" min_val = float("-5.45661") @@ -2556,6 +2808,7 @@ class Program_weight_tensor_parameter_251: class Program_weight_tensor_parameter_252: name = "parameter_252" + original_name = "conv2d_24.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.650638") @@ -2567,6 +2820,7 @@ class Program_weight_tensor_parameter_252: class Program_weight_tensor_parameter_253: name = "parameter_253" + original_name = "batch_norm2d_21.b_0" shape = [192] dtype = "float32" min_val = float("-4.65491") @@ -2578,6 +2832,7 @@ class Program_weight_tensor_parameter_253: class Program_weight_tensor_parameter_254: name = "parameter_254" + original_name = "batch_norm2d_21.w_0" shape = [192] dtype = "float32" min_val = float("-0.000694999") @@ -2589,6 +2844,7 @@ class Program_weight_tensor_parameter_254: class Program_weight_tensor_parameter_255: name = "parameter_255" + original_name = "batch_norm2d_21.w_2" shape = [192] dtype = "float32" min_val = float("5.60519e-45") @@ -2600,6 +2856,7 @@ class Program_weight_tensor_parameter_255: class Program_weight_tensor_parameter_256: name = "parameter_256" + original_name = "batch_norm2d_21.w_1" shape = [192] dtype = "float32" min_val = float("-11.1529") @@ -2611,6 +2868,7 @@ class Program_weight_tensor_parameter_256: class Program_weight_tensor_parameter_257: name = "parameter_257" + original_name = "conv2d_23.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.435511") @@ -2622,6 +2880,7 @@ class Program_weight_tensor_parameter_257: class Program_weight_tensor_parameter_258: name = "parameter_258" + original_name = "batch_norm2d_20.b_0" shape = [192] dtype = "float32" min_val = float("-6.64039") @@ -2633,6 +2892,7 @@ class Program_weight_tensor_parameter_258: class Program_weight_tensor_parameter_259: name = "parameter_259" + original_name = "batch_norm2d_20.w_0" shape = [192] dtype = "float32" min_val = float("0.698244") @@ -2644,6 +2904,7 @@ class Program_weight_tensor_parameter_259: class Program_weight_tensor_parameter_260: name = "parameter_260" + original_name = "batch_norm2d_20.w_2" shape = [192] dtype = "float32" min_val = float("5.60519e-45") @@ -2655,6 +2916,7 @@ class Program_weight_tensor_parameter_260: class Program_weight_tensor_parameter_261: name = "parameter_261" + original_name = "batch_norm2d_20.w_1" shape = [192] dtype = "float32" min_val = float("-14.4534") @@ -2666,6 +2928,7 @@ class Program_weight_tensor_parameter_261: class Program_weight_tensor_parameter_262: name = "parameter_262" + original_name = "conv2d_22.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.669143") @@ -2677,6 +2940,7 @@ class Program_weight_tensor_parameter_262: class Program_weight_tensor_parameter_263: name = "parameter_263" + original_name = "batch_norm2d_19.b_0" shape = [192] dtype = "float32" min_val = float("-10.7379") @@ -2688,6 +2952,7 @@ class Program_weight_tensor_parameter_263: class Program_weight_tensor_parameter_264: name = "parameter_264" + original_name = "batch_norm2d_19.w_0" shape = [192] dtype = "float32" min_val = float("0.827384") @@ -2699,6 +2964,7 @@ class Program_weight_tensor_parameter_264: class Program_weight_tensor_parameter_265: name = "parameter_265" + original_name = "batch_norm2d_19.w_2" shape = [192] dtype = "float32" min_val = float("5.60519e-45") @@ -2710,6 +2976,7 @@ class Program_weight_tensor_parameter_265: class Program_weight_tensor_parameter_266: name = "parameter_266" + original_name = "batch_norm2d_19.w_1" shape = [192] dtype = "float32" min_val = float("-25.8756") @@ -2721,6 +2988,7 @@ class Program_weight_tensor_parameter_266: class Program_weight_tensor_parameter_267: name = "parameter_267" + original_name = "conv2d_21.w_0" shape = [192, 512, 3, 3] dtype = "float32" min_val = float("-0.621902") @@ -2732,6 +3000,7 @@ class Program_weight_tensor_parameter_267: class Program_weight_tensor_parameter_268: name = "parameter_268" + original_name = "batch_norm2d_18.b_0" shape = [512] dtype = "float32" min_val = float("-3.45605") @@ -2743,6 +3012,7 @@ class Program_weight_tensor_parameter_268: class Program_weight_tensor_parameter_269: name = "parameter_269" + original_name = "batch_norm2d_18.w_0" shape = [512] dtype = "float32" min_val = float("-6.01897e-05") @@ -2754,6 +3024,7 @@ class Program_weight_tensor_parameter_269: class Program_weight_tensor_parameter_270: name = "parameter_270" + original_name = "batch_norm2d_18.w_2" shape = [512] dtype = "float32" min_val = float("5.60519e-45") @@ -2765,6 +3036,7 @@ class Program_weight_tensor_parameter_270: class Program_weight_tensor_parameter_271: name = "parameter_271" + original_name = "batch_norm2d_18.w_1" shape = [512] dtype = "float32" min_val = float("-0.417436") @@ -2776,6 +3048,7 @@ class Program_weight_tensor_parameter_271: class Program_weight_tensor_parameter_272: name = "parameter_272" + original_name = "conv2d_20.w_0" shape = [512, 1, 3, 3] dtype = "float32" min_val = float("-0.300424") @@ -2787,6 +3060,7 @@ class Program_weight_tensor_parameter_272: class Program_weight_tensor_parameter_273: name = "parameter_273" + original_name = "conv2d_19.b_0" shape = [512] dtype = "float32" min_val = float("-0.110146") @@ -2798,6 +3072,7 @@ class Program_weight_tensor_parameter_273: class Program_weight_tensor_parameter_274: name = "parameter_274" + original_name = "conv2d_19.w_0" shape = [512, 512, 1, 1] dtype = "float32" min_val = float("-2.02461") @@ -2809,6 +3084,7 @@ class Program_weight_tensor_parameter_274: class Program_weight_tensor_parameter_275: name = "parameter_275" + original_name = "batch_norm2d_17.b_0" shape = [512] dtype = "float32" min_val = float("-14.854") @@ -2820,6 +3096,7 @@ class Program_weight_tensor_parameter_275: class Program_weight_tensor_parameter_276: name = "parameter_276" + original_name = "batch_norm2d_17.w_0" shape = [512] dtype = "float32" min_val = float("0.0454745") @@ -2831,6 +3108,7 @@ class Program_weight_tensor_parameter_276: class Program_weight_tensor_parameter_277: name = "parameter_277" + original_name = "batch_norm2d_17.w_2" shape = [512] dtype = "float32" min_val = float("5.60519e-45") @@ -2842,6 +3120,7 @@ class Program_weight_tensor_parameter_277: class Program_weight_tensor_parameter_278: name = "parameter_278" + original_name = "batch_norm2d_17.w_1" shape = [512] dtype = "float32" min_val = float("-9.85744") @@ -2853,6 +3132,7 @@ class Program_weight_tensor_parameter_278: class Program_weight_tensor_parameter_279: name = "parameter_279" + original_name = "conv2d_18.w_0" shape = [512, 1216, 1, 1] dtype = "float32" min_val = float("-0.72981") @@ -2864,6 +3144,7 @@ class Program_weight_tensor_parameter_279: class Program_weight_tensor_parameter_280: name = "parameter_280" + original_name = "batch_norm2d_16.b_0" shape = [160] dtype = "float32" min_val = float("-5.38556") @@ -2875,6 +3156,7 @@ class Program_weight_tensor_parameter_280: class Program_weight_tensor_parameter_281: name = "parameter_281" + original_name = "batch_norm2d_16.w_0" shape = [160] dtype = "float32" min_val = float("1.12475") @@ -2886,6 +3168,7 @@ class Program_weight_tensor_parameter_281: class Program_weight_tensor_parameter_282: name = "parameter_282" + original_name = "batch_norm2d_16.w_2" shape = [160] dtype = "float32" min_val = float("5.60519e-45") @@ -2897,6 +3180,7 @@ class Program_weight_tensor_parameter_282: class Program_weight_tensor_parameter_283: name = "parameter_283" + original_name = "batch_norm2d_16.w_1" shape = [160] dtype = "float32" min_val = float("-11.2921") @@ -2908,6 +3192,7 @@ class Program_weight_tensor_parameter_283: class Program_weight_tensor_parameter_284: name = "parameter_284" + original_name = "conv2d_17.w_0" shape = [160, 160, 3, 3] dtype = "float32" min_val = float("-0.318814") @@ -2919,6 +3204,7 @@ class Program_weight_tensor_parameter_284: class Program_weight_tensor_parameter_285: name = "parameter_285" + original_name = "batch_norm2d_15.b_0" shape = [160] dtype = "float32" min_val = float("-7.00077") @@ -2930,6 +3216,7 @@ class Program_weight_tensor_parameter_285: class Program_weight_tensor_parameter_286: name = "parameter_286" + original_name = "batch_norm2d_15.w_0" shape = [160] dtype = "float32" min_val = float("0.265856") @@ -2941,6 +3228,7 @@ class Program_weight_tensor_parameter_286: class Program_weight_tensor_parameter_287: name = "parameter_287" + original_name = "batch_norm2d_15.w_2" shape = [160] dtype = "float32" min_val = float("5.60519e-45") @@ -2952,6 +3240,7 @@ class Program_weight_tensor_parameter_287: class Program_weight_tensor_parameter_288: name = "parameter_288" + original_name = "batch_norm2d_15.w_1" shape = [160] dtype = "float32" min_val = float("-5.80709") @@ -2963,6 +3252,7 @@ class Program_weight_tensor_parameter_288: class Program_weight_tensor_parameter_289: name = "parameter_289" + original_name = "conv2d_16.w_0" shape = [160, 160, 3, 3] dtype = "float32" min_val = float("-0.526021") @@ -2974,6 +3264,7 @@ class Program_weight_tensor_parameter_289: class Program_weight_tensor_parameter_290: name = "parameter_290" + original_name = "batch_norm2d_14.b_0" shape = [160] dtype = "float32" min_val = float("-5.7591") @@ -2985,6 +3276,7 @@ class Program_weight_tensor_parameter_290: class Program_weight_tensor_parameter_291: name = "parameter_291" + original_name = "batch_norm2d_14.w_0" shape = [160] dtype = "float32" min_val = float("0.743675") @@ -2996,6 +3288,7 @@ class Program_weight_tensor_parameter_291: class Program_weight_tensor_parameter_292: name = "parameter_292" + original_name = "batch_norm2d_14.w_2" shape = [160] dtype = "float32" min_val = float("5.60519e-45") @@ -3007,6 +3300,7 @@ class Program_weight_tensor_parameter_292: class Program_weight_tensor_parameter_293: name = "parameter_293" + original_name = "batch_norm2d_14.w_1" shape = [160] dtype = "float32" min_val = float("-12.1801") @@ -3018,6 +3312,7 @@ class Program_weight_tensor_parameter_293: class Program_weight_tensor_parameter_294: name = "parameter_294" + original_name = "conv2d_15.w_0" shape = [160, 160, 3, 3] dtype = "float32" min_val = float("-0.353851") @@ -3029,6 +3324,7 @@ class Program_weight_tensor_parameter_294: class Program_weight_tensor_parameter_295: name = "parameter_295" + original_name = "batch_norm2d_13.b_0" shape = [160] dtype = "float32" min_val = float("-5.6647") @@ -3040,6 +3336,7 @@ class Program_weight_tensor_parameter_295: class Program_weight_tensor_parameter_296: name = "parameter_296" + original_name = "batch_norm2d_13.w_0" shape = [160] dtype = "float32" min_val = float("0.00114078") @@ -3051,6 +3348,7 @@ class Program_weight_tensor_parameter_296: class Program_weight_tensor_parameter_297: name = "parameter_297" + original_name = "batch_norm2d_13.w_2" shape = [160] dtype = "float32" min_val = float("5.60519e-45") @@ -3062,6 +3360,7 @@ class Program_weight_tensor_parameter_297: class Program_weight_tensor_parameter_298: name = "parameter_298" + original_name = "batch_norm2d_13.w_1" shape = [160] dtype = "float32" min_val = float("-22.0715") @@ -3073,6 +3372,7 @@ class Program_weight_tensor_parameter_298: class Program_weight_tensor_parameter_299: name = "parameter_299" + original_name = "conv2d_14.w_0" shape = [160, 160, 3, 3] dtype = "float32" min_val = float("-0.513872") @@ -3084,6 +3384,7 @@ class Program_weight_tensor_parameter_299: class Program_weight_tensor_parameter_300: name = "parameter_300" + original_name = "batch_norm2d_12.b_0" shape = [160] dtype = "float32" min_val = float("-5.74121") @@ -3095,6 +3396,7 @@ class Program_weight_tensor_parameter_300: class Program_weight_tensor_parameter_301: name = "parameter_301" + original_name = "batch_norm2d_12.w_0" shape = [160] dtype = "float32" min_val = float("0.388887") @@ -3106,6 +3408,7 @@ class Program_weight_tensor_parameter_301: class Program_weight_tensor_parameter_302: name = "parameter_302" + original_name = "batch_norm2d_12.w_2" shape = [160] dtype = "float32" min_val = float("5.60519e-45") @@ -3117,6 +3420,7 @@ class Program_weight_tensor_parameter_302: class Program_weight_tensor_parameter_303: name = "parameter_303" + original_name = "batch_norm2d_12.w_1" shape = [160] dtype = "float32" min_val = float("-18.0207") @@ -3128,6 +3432,7 @@ class Program_weight_tensor_parameter_303: class Program_weight_tensor_parameter_304: name = "parameter_304" + original_name = "conv2d_13.w_0" shape = [160, 160, 3, 3] dtype = "float32" min_val = float("-0.532287") @@ -3139,6 +3444,7 @@ class Program_weight_tensor_parameter_304: class Program_weight_tensor_parameter_305: name = "parameter_305" + original_name = "batch_norm2d_11.b_0" shape = [160] dtype = "float32" min_val = float("-9.62261") @@ -3150,6 +3456,7 @@ class Program_weight_tensor_parameter_305: class Program_weight_tensor_parameter_306: name = "parameter_306" + original_name = "batch_norm2d_11.w_0" shape = [160] dtype = "float32" min_val = float("0.551166") @@ -3161,6 +3468,7 @@ class Program_weight_tensor_parameter_306: class Program_weight_tensor_parameter_307: name = "parameter_307" + original_name = "batch_norm2d_11.w_2" shape = [160] dtype = "float32" min_val = float("5.60519e-45") @@ -3172,6 +3480,7 @@ class Program_weight_tensor_parameter_307: class Program_weight_tensor_parameter_308: name = "parameter_308" + original_name = "batch_norm2d_11.w_1" shape = [160] dtype = "float32" min_val = float("-16.2092") @@ -3183,6 +3492,7 @@ class Program_weight_tensor_parameter_308: class Program_weight_tensor_parameter_309: name = "parameter_309" + original_name = "conv2d_12.w_0" shape = [160, 256, 3, 3] dtype = "float32" min_val = float("-0.686342") @@ -3194,6 +3504,7 @@ class Program_weight_tensor_parameter_309: class Program_weight_tensor_parameter_310: name = "parameter_310" + original_name = "batch_norm2d_10.b_0" shape = [256] dtype = "float32" min_val = float("-2.54056") @@ -3205,6 +3516,7 @@ class Program_weight_tensor_parameter_310: class Program_weight_tensor_parameter_311: name = "parameter_311" + original_name = "batch_norm2d_10.w_0" shape = [256] dtype = "float32" min_val = float("-0.00014248") @@ -3216,6 +3528,7 @@ class Program_weight_tensor_parameter_311: class Program_weight_tensor_parameter_312: name = "parameter_312" + original_name = "batch_norm2d_10.w_2" shape = [256] dtype = "float32" min_val = float("5.60519e-45") @@ -3227,6 +3540,7 @@ class Program_weight_tensor_parameter_312: class Program_weight_tensor_parameter_313: name = "parameter_313" + original_name = "batch_norm2d_10.w_1" shape = [256] dtype = "float32" min_val = float("-0.697656") @@ -3238,6 +3552,7 @@ class Program_weight_tensor_parameter_313: class Program_weight_tensor_parameter_314: name = "parameter_314" + original_name = "conv2d_11.w_0" shape = [256, 1, 3, 3] dtype = "float32" min_val = float("-0.349396") @@ -3249,6 +3564,7 @@ class Program_weight_tensor_parameter_314: class Program_weight_tensor_parameter_315: name = "parameter_315" + original_name = "conv2d_10.b_0" shape = [256] dtype = "float32" min_val = float("-0.151844") @@ -3260,6 +3576,7 @@ class Program_weight_tensor_parameter_315: class Program_weight_tensor_parameter_316: name = "parameter_316" + original_name = "conv2d_10.w_0" shape = [256, 256, 1, 1] dtype = "float32" min_val = float("-2.22047") @@ -3271,6 +3588,7 @@ class Program_weight_tensor_parameter_316: class Program_weight_tensor_parameter_317: name = "parameter_317" + original_name = "batch_norm2d_9.b_0" shape = [256] dtype = "float32" min_val = float("-10.7762") @@ -3282,6 +3600,7 @@ class Program_weight_tensor_parameter_317: class Program_weight_tensor_parameter_318: name = "parameter_318" + original_name = "batch_norm2d_9.w_0" shape = [256] dtype = "float32" min_val = float("-0.000974758") @@ -3293,6 +3612,7 @@ class Program_weight_tensor_parameter_318: class Program_weight_tensor_parameter_319: name = "parameter_319" + original_name = "batch_norm2d_9.w_2" shape = [256] dtype = "float32" min_val = float("5.60519e-45") @@ -3304,6 +3624,7 @@ class Program_weight_tensor_parameter_319: class Program_weight_tensor_parameter_320: name = "parameter_320" + original_name = "batch_norm2d_9.w_1" shape = [256] dtype = "float32" min_val = float("-10.5129") @@ -3315,6 +3636,7 @@ class Program_weight_tensor_parameter_320: class Program_weight_tensor_parameter_321: name = "parameter_321" + original_name = "conv2d_9.w_0" shape = [256, 896, 1, 1] dtype = "float32" min_val = float("-1.34067") @@ -3326,6 +3648,7 @@ class Program_weight_tensor_parameter_321: class Program_weight_tensor_parameter_322: name = "parameter_322" + original_name = "batch_norm2d_8.b_0" shape = [128] dtype = "float32" min_val = float("-6.77381") @@ -3337,6 +3660,7 @@ class Program_weight_tensor_parameter_322: class Program_weight_tensor_parameter_323: name = "parameter_323" + original_name = "batch_norm2d_8.w_0" shape = [128] dtype = "float32" min_val = float("0.960223") @@ -3348,6 +3672,7 @@ class Program_weight_tensor_parameter_323: class Program_weight_tensor_parameter_324: name = "parameter_324" + original_name = "batch_norm2d_8.w_2" shape = [128] dtype = "float32" min_val = float("5.60519e-45") @@ -3359,6 +3684,7 @@ class Program_weight_tensor_parameter_324: class Program_weight_tensor_parameter_325: name = "parameter_325" + original_name = "batch_norm2d_8.w_1" shape = [128] dtype = "float32" min_val = float("-2.99585") @@ -3370,6 +3696,7 @@ class Program_weight_tensor_parameter_325: class Program_weight_tensor_parameter_326: name = "parameter_326" + original_name = "conv2d_8.w_0" shape = [128, 128, 3, 3] dtype = "float32" min_val = float("-0.292693") @@ -3381,6 +3708,7 @@ class Program_weight_tensor_parameter_326: class Program_weight_tensor_parameter_327: name = "parameter_327" + original_name = "batch_norm2d_7.b_0" shape = [128] dtype = "float32" min_val = float("-8.99448") @@ -3392,6 +3720,7 @@ class Program_weight_tensor_parameter_327: class Program_weight_tensor_parameter_328: name = "parameter_328" + original_name = "batch_norm2d_7.w_0" shape = [128] dtype = "float32" min_val = float("0.380296") @@ -3403,6 +3732,7 @@ class Program_weight_tensor_parameter_328: class Program_weight_tensor_parameter_329: name = "parameter_329" + original_name = "batch_norm2d_7.w_2" shape = [128] dtype = "float32" min_val = float("5.60519e-45") @@ -3414,6 +3744,7 @@ class Program_weight_tensor_parameter_329: class Program_weight_tensor_parameter_330: name = "parameter_330" + original_name = "batch_norm2d_7.w_1" shape = [128] dtype = "float32" min_val = float("-6.62991") @@ -3425,6 +3756,7 @@ class Program_weight_tensor_parameter_330: class Program_weight_tensor_parameter_331: name = "parameter_331" + original_name = "conv2d_7.w_0" shape = [128, 128, 3, 3] dtype = "float32" min_val = float("-0.356885") @@ -3436,6 +3768,7 @@ class Program_weight_tensor_parameter_331: class Program_weight_tensor_parameter_332: name = "parameter_332" + original_name = "batch_norm2d_6.b_0" shape = [128] dtype = "float32" min_val = float("-6.2876") @@ -3447,6 +3780,7 @@ class Program_weight_tensor_parameter_332: class Program_weight_tensor_parameter_333: name = "parameter_333" + original_name = "batch_norm2d_6.w_0" shape = [128] dtype = "float32" min_val = float("0.405473") @@ -3458,6 +3792,7 @@ class Program_weight_tensor_parameter_333: class Program_weight_tensor_parameter_334: name = "parameter_334" + original_name = "batch_norm2d_6.w_2" shape = [128] dtype = "float32" min_val = float("5.60519e-45") @@ -3469,6 +3804,7 @@ class Program_weight_tensor_parameter_334: class Program_weight_tensor_parameter_335: name = "parameter_335" + original_name = "batch_norm2d_6.w_1" shape = [128] dtype = "float32" min_val = float("-22.0119") @@ -3480,6 +3816,7 @@ class Program_weight_tensor_parameter_335: class Program_weight_tensor_parameter_336: name = "parameter_336" + original_name = "conv2d_6.w_0" shape = [128, 128, 3, 3] dtype = "float32" min_val = float("-0.497324") @@ -3491,6 +3828,7 @@ class Program_weight_tensor_parameter_336: class Program_weight_tensor_parameter_337: name = "parameter_337" + original_name = "batch_norm2d_5.b_0" shape = [128] dtype = "float32" min_val = float("-4.65093") @@ -3502,6 +3840,7 @@ class Program_weight_tensor_parameter_337: class Program_weight_tensor_parameter_338: name = "parameter_338" + original_name = "batch_norm2d_5.w_0" shape = [128] dtype = "float32" min_val = float("0.434265") @@ -3513,6 +3852,7 @@ class Program_weight_tensor_parameter_338: class Program_weight_tensor_parameter_339: name = "parameter_339" + original_name = "batch_norm2d_5.w_2" shape = [128] dtype = "float32" min_val = float("5.60519e-45") @@ -3524,6 +3864,7 @@ class Program_weight_tensor_parameter_339: class Program_weight_tensor_parameter_340: name = "parameter_340" + original_name = "batch_norm2d_5.w_1" shape = [128] dtype = "float32" min_val = float("-9.39086") @@ -3535,6 +3876,7 @@ class Program_weight_tensor_parameter_340: class Program_weight_tensor_parameter_341: name = "parameter_341" + original_name = "conv2d_5.w_0" shape = [128, 128, 3, 3] dtype = "float32" min_val = float("-0.473818") @@ -3546,6 +3888,7 @@ class Program_weight_tensor_parameter_341: class Program_weight_tensor_parameter_342: name = "parameter_342" + original_name = "batch_norm2d_4.b_0" shape = [128] dtype = "float32" min_val = float("-9.0258") @@ -3557,6 +3900,7 @@ class Program_weight_tensor_parameter_342: class Program_weight_tensor_parameter_343: name = "parameter_343" + original_name = "batch_norm2d_4.w_0" shape = [128] dtype = "float32" min_val = float("-0.00169905") @@ -3568,6 +3912,7 @@ class Program_weight_tensor_parameter_343: class Program_weight_tensor_parameter_344: name = "parameter_344" + original_name = "batch_norm2d_4.w_2" shape = [128] dtype = "float32" min_val = float("5.60519e-45") @@ -3579,6 +3924,7 @@ class Program_weight_tensor_parameter_344: class Program_weight_tensor_parameter_345: name = "parameter_345" + original_name = "batch_norm2d_4.w_1" shape = [128] dtype = "float32" min_val = float("-18.7924") @@ -3590,6 +3936,7 @@ class Program_weight_tensor_parameter_345: class Program_weight_tensor_parameter_346: name = "parameter_346" + original_name = "conv2d_4.w_0" shape = [128, 128, 3, 3] dtype = "float32" min_val = float("-0.538366") @@ -3601,6 +3948,7 @@ class Program_weight_tensor_parameter_346: class Program_weight_tensor_parameter_347: name = "parameter_347" + original_name = "batch_norm2d_3.b_0" shape = [128] dtype = "float32" min_val = float("-7.2894") @@ -3612,6 +3960,7 @@ class Program_weight_tensor_parameter_347: class Program_weight_tensor_parameter_348: name = "parameter_348" + original_name = "batch_norm2d_3.w_0" shape = [128] dtype = "float32" min_val = float("0.318356") @@ -3623,6 +3972,7 @@ class Program_weight_tensor_parameter_348: class Program_weight_tensor_parameter_349: name = "parameter_349" + original_name = "batch_norm2d_3.w_2" shape = [128] dtype = "float32" min_val = float("5.60519e-45") @@ -3634,6 +3984,7 @@ class Program_weight_tensor_parameter_349: class Program_weight_tensor_parameter_350: name = "parameter_350" + original_name = "batch_norm2d_3.w_1" shape = [128] dtype = "float32" min_val = float("-22.4318") @@ -3645,6 +3996,7 @@ class Program_weight_tensor_parameter_350: class Program_weight_tensor_parameter_351: name = "parameter_351" + original_name = "conv2d_3.w_0" shape = [128, 128, 3, 3] dtype = "float32" min_val = float("-0.776036") @@ -3656,6 +4008,7 @@ class Program_weight_tensor_parameter_351: class Program_weight_tensor_parameter_352: name = "parameter_352" + original_name = "batch_norm2d_2.b_0" shape = [128] dtype = "float32" min_val = float("-1.60328") @@ -3667,6 +4020,7 @@ class Program_weight_tensor_parameter_352: class Program_weight_tensor_parameter_353: name = "parameter_353" + original_name = "batch_norm2d_2.w_0" shape = [128] dtype = "float32" min_val = float("0.0485431") @@ -3678,6 +4032,7 @@ class Program_weight_tensor_parameter_353: class Program_weight_tensor_parameter_354: name = "parameter_354" + original_name = "batch_norm2d_2.w_2" shape = [128] dtype = "float32" min_val = float("0.0667826") @@ -3689,6 +4044,7 @@ class Program_weight_tensor_parameter_354: class Program_weight_tensor_parameter_355: name = "parameter_355" + original_name = "batch_norm2d_2.w_1" shape = [128] dtype = "float32" min_val = float("-22.6417") @@ -3700,6 +4056,7 @@ class Program_weight_tensor_parameter_355: class Program_weight_tensor_parameter_356: name = "parameter_356" + original_name = "conv2d_2.w_0" shape = [128, 64, 3, 3] dtype = "float32" min_val = float("-0.950136") @@ -3711,6 +4068,7 @@ class Program_weight_tensor_parameter_356: class Program_weight_tensor_parameter_357: name = "parameter_357" + original_name = "batch_norm2d_1.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -3720,6 +4078,7 @@ class Program_weight_tensor_parameter_357: class Program_weight_tensor_parameter_358: name = "parameter_358" + original_name = "batch_norm2d_1.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -3729,6 +4088,7 @@ class Program_weight_tensor_parameter_358: class Program_weight_tensor_parameter_359: name = "parameter_359" + original_name = "batch_norm2d_1.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -3738,6 +4098,7 @@ class Program_weight_tensor_parameter_359: class Program_weight_tensor_parameter_360: name = "parameter_360" + original_name = "batch_norm2d_1.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -3747,6 +4108,7 @@ class Program_weight_tensor_parameter_360: class Program_weight_tensor_parameter_361: name = "parameter_361" + original_name = "conv2d_1.w_0" shape = [64, 64, 3, 3] dtype = "float32" min_val = float("-1.28267") @@ -3758,6 +4120,7 @@ class Program_weight_tensor_parameter_361: class Program_weight_tensor_parameter_362: name = "parameter_362" + original_name = "batch_norm2d_0.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -3767,6 +4130,7 @@ class Program_weight_tensor_parameter_362: class Program_weight_tensor_parameter_363: name = "parameter_363" + original_name = "batch_norm2d_0.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -3776,6 +4140,7 @@ class Program_weight_tensor_parameter_363: class Program_weight_tensor_parameter_364: name = "parameter_364" + original_name = "batch_norm2d_0.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -3785,6 +4150,7 @@ class Program_weight_tensor_parameter_364: class Program_weight_tensor_parameter_365: name = "parameter_365" + original_name = "batch_norm2d_0.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -3794,6 +4160,7 @@ class Program_weight_tensor_parameter_365: class Program_weight_tensor_parameter_366: name = "parameter_366" + original_name = "conv2d_0.w_0" shape = [64, 3, 3, 3] dtype = "float32" min_val = float("-0.710112") diff --git a/paddle_samples/PaddleX/PP-ShiTuV2_det/subgraph_1/input_meta.py b/paddle_samples/PaddleX/PP-ShiTuV2_det/subgraph_1/input_meta.py index 16cf942ac..c87f211e7 100644 --- a/paddle_samples/PaddleX/PP-ShiTuV2_det/subgraph_1/input_meta.py +++ b/paddle_samples/PaddleX/PP-ShiTuV2_det/subgraph_1/input_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_data_0: name = "data_0" + original_name = "var_6700" shape = [8, 3, 640, 640] dtype = "float32" min_val = float("-2.1179") @@ -11,6 +12,7 @@ class Program_weight_tensor_data_0: class Program_weight_tensor_data_1: name = "data_1" + original_name = "var_6912" shape = [8500, 2] dtype = "float32" min_val = float("0.5") @@ -22,6 +24,7 @@ class Program_weight_tensor_data_1: class Program_weight_tensor_data_2: name = "data_2" + original_name = "var_6913" shape = [8500, 1] dtype = "float32" min_val = float("8.0") @@ -33,6 +36,7 @@ class Program_weight_tensor_data_2: class Program_weight_tensor_data_3: name = "data_3" + original_name = "var_6946" shape = [8] dtype = "float32" data = [0.0, 1.00026, 2.00052, 3.00079, 4.00105, 5.00131, 6.00157, 7.00183] diff --git a/paddle_samples/PaddleX/PP-ShiTuV2_det/subgraph_1/weight_meta.py b/paddle_samples/PaddleX/PP-ShiTuV2_det/subgraph_1/weight_meta.py index 1f5149559..00e09a96c 100644 --- a/paddle_samples/PaddleX/PP-ShiTuV2_det/subgraph_1/weight_meta.py +++ b/paddle_samples/PaddleX/PP-ShiTuV2_det/subgraph_1/weight_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_parameter_0: name = "parameter_0" + original_name = "conv2d_103.b_0" shape = [33] dtype = "float32" min_val = float("0") @@ -9,96 +10,127 @@ class Program_weight_tensor_parameter_0: class Program_weight_tensor_parameter_1: name = "parameter_1" + original_name = "conv2d_103.w_0" shape = [33, 128, 1, 1] dtype = "float32" - min_val = float("-0.5") - max_val = float("0.5") + min_val = float("-23.2679") + max_val = float("5.52957") + mean = float("-0.103645") + std = float("1.06012") data = None class Program_weight_tensor_parameter_2: name = "parameter_2" + original_name = "batch_norm2d_93.b_0" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-9.44689") + max_val = float("28.8858") + mean = float("-0.086711") + std = float("4.3485") data = None class Program_weight_tensor_parameter_3: name = "parameter_3" + original_name = "batch_norm2d_93.w_0" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-8.25383") + max_val = float("19.3763") + mean = float("0.697591") + std = float("3.62338") data = None class Program_weight_tensor_parameter_4: name = "parameter_4" + original_name = "batch_norm2d_93.w_2" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("27.5368") + max_val = float("101359.0") + mean = float("5488.19") + std = float("14115.3") data = None class Program_weight_tensor_parameter_5: name = "parameter_5" + original_name = "batch_norm2d_93.w_1" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-348.55") + max_val = float("383.928") + mean = float("2.32895") + std = float("65.0104") data = None class Program_weight_tensor_parameter_6: name = "parameter_6" + original_name = "conv2d_97.w_0" shape = [128, 128, 1, 1] dtype = "float32" - min_val = float("-0.5") - max_val = float("0.5") + min_val = float("-17.8787") + max_val = float("18.8271") + mean = float("0.00311308") + std = float("0.815553") data = None class Program_weight_tensor_parameter_7: name = "parameter_7" + original_name = "batch_norm2d_92.b_0" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-6.12415") + max_val = float("10.327") + mean = float("0.368743") + std = float("1.63624") data = None class Program_weight_tensor_parameter_8: name = "parameter_8" + original_name = "batch_norm2d_92.w_0" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-18.3997") + max_val = float("12.8783") + mean = float("1.09473") + std = float("3.17309") data = None class Program_weight_tensor_parameter_9: name = "parameter_9" + original_name = "batch_norm2d_92.w_2" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("1.90338e-05") + max_val = float("101966.0") + mean = float("1223.35") + std = float("9196.41") data = None class Program_weight_tensor_parameter_10: name = "parameter_10" + original_name = "batch_norm2d_92.w_1" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-71.4263") + max_val = float("82.1673") + mean = float("1.04932") + std = float("14.0907") data = None class Program_weight_tensor_parameter_11: name = "parameter_11" + original_name = "conv2d_96.w_0" shape = [128, 1, 5, 5] dtype = "float32" min_val = float("-6.93173") @@ -110,177 +142,235 @@ class Program_weight_tensor_parameter_11: class Program_weight_tensor_parameter_12: name = "parameter_12" + original_name = "batch_norm2d_91.b_0" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-13.911") + max_val = float("14.8464") + mean = float("0.00232437") + std = float("2.28411") data = None class Program_weight_tensor_parameter_13: name = "parameter_13" + original_name = "batch_norm2d_91.w_0" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-24.3725") + max_val = float("11.7775") + mean = float("0.964567") + std = float("2.89227") data = None class Program_weight_tensor_parameter_14: name = "parameter_14" + original_name = "batch_norm2d_91.w_2" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("19.8364") + max_val = float("67885.5") + mean = float("1729.25") + std = float("6620.55") data = None class Program_weight_tensor_parameter_15: name = "parameter_15" + original_name = "batch_norm2d_91.w_1" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-78.8594") + max_val = float("88.0705") + mean = float("-0.478662") + std = float("20.4114") data = None class Program_weight_tensor_parameter_16: name = "parameter_16" + original_name = "conv2d_95.w_0" shape = [128, 128, 1, 1] dtype = "float32" - min_val = float("-0.5") - max_val = float("0.5") + min_val = float("-13.1537") + max_val = float("20.4486") + mean = float("0.00395922") + std = float("0.728075") data = None class Program_weight_tensor_parameter_17: name = "parameter_17" + original_name = "batch_norm2d_90.b_0" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-4.40036") + max_val = float("2.91727") + mean = float("0.0107575") + std = float("1.22564") data = None class Program_weight_tensor_parameter_18: name = "parameter_18" + original_name = "batch_norm2d_90.w_0" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-16.5993") + max_val = float("11.6119") + mean = float("0.758058") + std = float("3.40494") data = None class Program_weight_tensor_parameter_19: name = "parameter_19" + original_name = "batch_norm2d_90.w_2" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("0.00344497") + max_val = float("21277.2") + mean = float("392.752") + std = float("1969.61") data = None class Program_weight_tensor_parameter_20: name = "parameter_20" + original_name = "batch_norm2d_90.w_1" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-36.1668") + max_val = float("135.041") + mean = float("-0.0150359") + std = float("14.7424") data = None class Program_weight_tensor_parameter_21: name = "parameter_21" + original_name = "conv2d_94.w_0" shape = [128, 1, 5, 5] dtype = "float32" - min_val = float("-0.5") - max_val = float("0.5") + min_val = float("-10.038") + max_val = float("6.96991") + mean = float("-0.0426436") + std = float("0.898701") data = None class Program_weight_tensor_parameter_22: name = "parameter_22" + original_name = "batch_norm2d_89.b_0" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-13.883") + max_val = float("3.98622") + mean = float("-0.129747") + std = float("1.98716") data = None class Program_weight_tensor_parameter_23: name = "parameter_23" + original_name = "batch_norm2d_89.w_0" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-7.05052") + max_val = float("25.728") + mean = float("1.48559") + std = float("2.75163") data = None class Program_weight_tensor_parameter_24: name = "parameter_24" + original_name = "batch_norm2d_89.w_2" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("19.9804") + max_val = float("21819.4") + mean = float("1027.41") + std = float("2403.57") data = None class Program_weight_tensor_parameter_25: name = "parameter_25" + original_name = "batch_norm2d_89.w_1" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-45.336") + max_val = float("43.4259") + mean = float("2.48744") + std = float("13.3475") data = None class Program_weight_tensor_parameter_26: name = "parameter_26" + original_name = "conv2d_93.w_0" shape = [128, 128, 1, 1] dtype = "float32" - min_val = float("-0.5") - max_val = float("0.5") + min_val = float("-9.19836") + max_val = float("11.9814") + mean = float("-0.00328695") + std = float("0.621386") data = None class Program_weight_tensor_parameter_27: name = "parameter_27" + original_name = "batch_norm2d_88.b_0" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-2.79735") + max_val = float("4.02966") + mean = float("0.108958") + std = float("1.06723") data = None class Program_weight_tensor_parameter_28: name = "parameter_28" + original_name = "batch_norm2d_88.w_0" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-8.20665") + max_val = float("10.8662") + mean = float("1.02705") + std = float("2.23558") data = None class Program_weight_tensor_parameter_29: name = "parameter_29" + original_name = "batch_norm2d_88.w_2" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("0.00604247") + max_val = float("18291.6") + mean = float("291.972") + std = float("1703.65") data = None class Program_weight_tensor_parameter_30: name = "parameter_30" + original_name = "batch_norm2d_88.w_1" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-73.2998") + max_val = float("177.569") + mean = float("-0.0510375") + std = float("19.7218") data = None class Program_weight_tensor_parameter_31: name = "parameter_31" + original_name = "conv2d_92.w_0" shape = [128, 1, 5, 5] dtype = "float32" min_val = float("-7.06043") @@ -292,96 +382,127 @@ class Program_weight_tensor_parameter_31: class Program_weight_tensor_parameter_32: name = "parameter_32" + original_name = "batch_norm2d_87.b_0" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-15.4828") + max_val = float("5.07715") + mean = float("0.154783") + std = float("1.75707") data = None class Program_weight_tensor_parameter_33: name = "parameter_33" + original_name = "batch_norm2d_87.w_0" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-7.76485") + max_val = float("5.04566") + mean = float("1.10151") + std = float("1.48781") data = None class Program_weight_tensor_parameter_34: name = "parameter_34" + original_name = "batch_norm2d_87.w_2" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("13.3475") + max_val = float("46424.9") + mean = float("969.637") + std = float("4205.52") data = None class Program_weight_tensor_parameter_35: name = "parameter_35" + original_name = "batch_norm2d_87.w_1" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-86.1271") + max_val = float("37.2109") + mean = float("-1.96438") + std = float("17.3468") data = None class Program_weight_tensor_parameter_36: name = "parameter_36" + original_name = "conv2d_91.w_0" shape = [128, 128, 1, 1] dtype = "float32" - min_val = float("-0.5") - max_val = float("0.5") + min_val = float("-9.85211") + max_val = float("15.267") + mean = float("-0.0319715") + std = float("0.567769") data = None class Program_weight_tensor_parameter_37: name = "parameter_37" + original_name = "batch_norm2d_86.b_0" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-3.62098") + max_val = float("5.00576") + mean = float("-0.0100362") + std = float("1.25985") data = None class Program_weight_tensor_parameter_38: name = "parameter_38" + original_name = "batch_norm2d_86.w_0" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-3.41748") + max_val = float("14.553") + mean = float("1.46587") + std = float("2.09719") data = None class Program_weight_tensor_parameter_39: name = "parameter_39" + original_name = "batch_norm2d_86.w_2" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("0.156019") + max_val = float("79843.1") + mean = float("1260.04") + std = float("7312.93") data = None class Program_weight_tensor_parameter_40: name = "parameter_40" + original_name = "batch_norm2d_86.w_1" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-93.9435") + max_val = float("195.31") + mean = float("-1.08895") + std = float("24.8723") data = None class Program_weight_tensor_parameter_41: name = "parameter_41" + original_name = "conv2d_90.w_0" shape = [128, 1, 5, 5] dtype = "float32" - min_val = float("-0.5") - max_val = float("0.5") + min_val = float("-6.84996") + max_val = float("11.0312") + mean = float("-0.00683052") + std = float("1.02627") data = None class Program_weight_tensor_parameter_42: name = "parameter_42" + original_name = "conv2d_102.b_0" shape = [33] dtype = "float32" min_val = float("0") @@ -391,375 +512,499 @@ class Program_weight_tensor_parameter_42: class Program_weight_tensor_parameter_43: name = "parameter_43" + original_name = "conv2d_102.w_0" shape = [33, 128, 1, 1] dtype = "float32" - min_val = float("-0.5") - max_val = float("0.5") + min_val = float("-442.261") + max_val = float("29.5582") + mean = float("-0.748782") + std = float("10.2722") data = None class Program_weight_tensor_parameter_44: name = "parameter_44" + original_name = "batch_norm2d_85.b_0" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-128.453") + max_val = float("128.458") + mean = float("-19.2707") + std = float("38.4629") data = None class Program_weight_tensor_parameter_45: name = "parameter_45" + original_name = "batch_norm2d_85.w_0" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-171.061") + max_val = float("123.793") + mean = float("2.34575") + std = float("40.9887") data = None class Program_weight_tensor_parameter_46: name = "parameter_46" + original_name = "batch_norm2d_85.w_2" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("58247.6") + max_val = float("20544400000.0") + mean = float("675237000.0") + std = float("2138970000.0") data = None class Program_weight_tensor_parameter_47: name = "parameter_47" + original_name = "batch_norm2d_85.w_1" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-36045.8") + max_val = float("40594.9") + mean = float("-441.179") + std = float("9183.83") data = None class Program_weight_tensor_parameter_48: name = "parameter_48" + original_name = "conv2d_89.w_0" shape = [128, 128, 1, 1] dtype = "float32" - min_val = float("-0.5") - max_val = float("0.5") + min_val = float("-402.401") + max_val = float("590.416") + mean = float("-0.216676") + std = float("12.7821") data = None class Program_weight_tensor_parameter_49: name = "parameter_49" + original_name = "batch_norm2d_84.b_0" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-43.8456") + max_val = float("60.0145") + mean = float("2.08177") + std = float("14.533") data = None class Program_weight_tensor_parameter_50: name = "parameter_50" + original_name = "batch_norm2d_84.w_0" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-215.889") + max_val = float("245.955") + mean = float("-0.402725") + std = float("42.6923") data = None class Program_weight_tensor_parameter_51: name = "parameter_51" + original_name = "batch_norm2d_84.w_2" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("0.000582512") + max_val = float("3249090000.0") + mean = float("60770300.0") + std = float("377372000.0") data = None class Program_weight_tensor_parameter_52: name = "parameter_52" + original_name = "batch_norm2d_84.w_1" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-12740.5") + max_val = float("12205.8") + mean = float("420.62") + std = float("2425.21") data = None class Program_weight_tensor_parameter_53: name = "parameter_53" + original_name = "conv2d_88.w_0" shape = [128, 1, 5, 5] dtype = "float32" - min_val = float("-0.5") - max_val = float("0.5") + min_val = float("-225.295") + max_val = float("160.234") + mean = float("0.718652") + std = float("12.0568") data = None class Program_weight_tensor_parameter_54: name = "parameter_54" + original_name = "batch_norm2d_83.b_0" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-165.229") + max_val = float("98.7489") + mean = float("-2.39266") + std = float("27.7159") data = None class Program_weight_tensor_parameter_55: name = "parameter_55" + original_name = "batch_norm2d_83.w_0" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-440.613") + max_val = float("192.461") + mean = float("-3.13911") + std = float("53.579") data = None class Program_weight_tensor_parameter_56: name = "parameter_56" + original_name = "batch_norm2d_83.w_2" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("2857.87") + max_val = float("412905000.0") + mean = float("15818900.0") + std = float("54657900.0") data = None class Program_weight_tensor_parameter_57: name = "parameter_57" + original_name = "batch_norm2d_83.w_1" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-32178.6") + max_val = float("17526.5") + mean = float("-726.813") + std = float("5036.54") data = None class Program_weight_tensor_parameter_58: name = "parameter_58" + original_name = "conv2d_87.w_0" shape = [128, 128, 1, 1] dtype = "float32" - min_val = float("-0.5") - max_val = float("0.5") + min_val = float("-222.106") + max_val = float("184.113") + mean = float("-0.212476") + std = float("6.73944") data = None class Program_weight_tensor_parameter_59: name = "parameter_59" + original_name = "batch_norm2d_82.b_0" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-42.2108") + max_val = float("80.0762") + mean = float("0.384325") + std = float("13.0137") data = None class Program_weight_tensor_parameter_60: name = "parameter_60" + original_name = "batch_norm2d_82.w_0" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-119.168") + max_val = float("76.3692") + mean = float("-0.512938") + std = float("24.178") data = None class Program_weight_tensor_parameter_61: name = "parameter_61" + original_name = "batch_norm2d_82.w_2" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("0.00900684") + max_val = float("3556900000.0") + mean = float("43511400.0") + std = float("330022000.0") data = None class Program_weight_tensor_parameter_62: name = "parameter_62" + original_name = "batch_norm2d_82.w_1" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-7513.39") + max_val = float("11736.8") + mean = float("35.0078") + std = float("2015.92") data = None class Program_weight_tensor_parameter_63: name = "parameter_63" + original_name = "conv2d_86.w_0" shape = [128, 1, 5, 5] dtype = "float32" - min_val = float("-0.5") - max_val = float("0.5") + min_val = float("-108.332") + max_val = float("93.6729") + mean = float("-0.381156") + std = float("9.67139") data = None class Program_weight_tensor_parameter_64: name = "parameter_64" + original_name = "batch_norm2d_81.b_0" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-120.379") + max_val = float("55.5018") + mean = float("0.1033") + std = float("21.0663") data = None class Program_weight_tensor_parameter_65: name = "parameter_65" + original_name = "batch_norm2d_81.w_0" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-541.963") + max_val = float("68.5113") + mean = float("-9.1398") + std = float("55.5127") data = None class Program_weight_tensor_parameter_66: name = "parameter_66" + original_name = "batch_norm2d_81.w_2" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("1872.14") + max_val = float("47439700.0") + mean = float("2879840.0") + std = float("7270230.0") data = None class Program_weight_tensor_parameter_67: name = "parameter_67" + original_name = "batch_norm2d_81.w_1" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-6632.99") + max_val = float("7118.33") + mean = float("50.1768") + std = float("1655.88") data = None class Program_weight_tensor_parameter_68: name = "parameter_68" + original_name = "conv2d_85.w_0" shape = [128, 128, 1, 1] dtype = "float32" - min_val = float("-0.5") - max_val = float("0.5") + min_val = float("-124.574") + max_val = float("161.432") + mean = float("0.0220628") + std = float("5.32322") data = None class Program_weight_tensor_parameter_69: name = "parameter_69" + original_name = "batch_norm2d_80.b_0" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-63.5623") + max_val = float("80.578") + mean = float("-0.638638") + std = float("15.4666") data = None class Program_weight_tensor_parameter_70: name = "parameter_70" + original_name = "batch_norm2d_80.w_0" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-75.8899") + max_val = float("111.098") + mean = float("-1.81645") + std = float("19.2609") data = None class Program_weight_tensor_parameter_71: name = "parameter_71" + original_name = "batch_norm2d_80.w_2" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("0.219354") + max_val = float("10336600000.0") + mean = float("92021900.0") + std = float("912436000.0") data = None class Program_weight_tensor_parameter_72: name = "parameter_72" + original_name = "batch_norm2d_80.w_1" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-65518.1") + max_val = float("14121.3") + mean = float("-386.052") + std = float("6243.63") data = None class Program_weight_tensor_parameter_73: name = "parameter_73" + original_name = "conv2d_84.w_0" shape = [128, 1, 5, 5] dtype = "float32" - min_val = float("-0.5") - max_val = float("0.5") + min_val = float("-153.114") + max_val = float("241.766") + mean = float("0.298693") + std = float("9.93815") data = None class Program_weight_tensor_parameter_74: name = "parameter_74" + original_name = "batch_norm2d_79.b_0" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-166.143") + max_val = float("305.91") + mean = float("0.962712") + std = float("37.9996") data = None class Program_weight_tensor_parameter_75: name = "parameter_75" + original_name = "batch_norm2d_79.w_0" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-203.995") + max_val = float("1262.01") + mean = float("9.292") + std = float("114.669") data = None class Program_weight_tensor_parameter_76: name = "parameter_76" + original_name = "batch_norm2d_79.w_2" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("9659.32") + max_val = float("2853920000.0") + mean = float("67779800.0") + std = float("311856000.0") data = None class Program_weight_tensor_parameter_77: name = "parameter_77" + original_name = "batch_norm2d_79.w_1" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-15644.3") + max_val = float("56290.6") + mean = float("390.564") + std = float("7221.66") data = None class Program_weight_tensor_parameter_78: name = "parameter_78" + original_name = "conv2d_83.w_0" shape = [128, 128, 1, 1] dtype = "float32" - min_val = float("-0.5") - max_val = float("0.5") + min_val = float("-288.428") + max_val = float("272.699") + mean = float("0.162479") + std = float("7.34509") data = None class Program_weight_tensor_parameter_79: name = "parameter_79" + original_name = "batch_norm2d_78.b_0" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-73.7613") + max_val = float("85.9011") + mean = float("1.63348") + std = float("20.1408") data = None class Program_weight_tensor_parameter_80: name = "parameter_80" + original_name = "batch_norm2d_78.w_0" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-223.75") + max_val = float("449.956") + mean = float("4.01356") + std = float("58.707") data = None class Program_weight_tensor_parameter_81: name = "parameter_81" + original_name = "batch_norm2d_78.w_2" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("0.00619647") + max_val = float("8816430000000.0") + mean = float("70110200000.0") + std = float("776208000000.0") data = None class Program_weight_tensor_parameter_82: name = "parameter_82" + original_name = "batch_norm2d_78.w_1" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-121377.0") + max_val = float("1669060.0") + mean = float("15548.7") + std = float("148946.0") data = None class Program_weight_tensor_parameter_83: name = "parameter_83" + original_name = "conv2d_82.w_0" shape = [128, 1, 5, 5] dtype = "float32" - min_val = float("-0.5") - max_val = float("0.5") + min_val = float("-472.185") + max_val = float("574.109") + mean = float("1.12515") + std = float("31.1575") data = None class Program_weight_tensor_parameter_84: name = "parameter_84" + original_name = "conv2d_101.b_0" shape = [33] dtype = "float32" min_val = float("0") @@ -769,375 +1014,499 @@ class Program_weight_tensor_parameter_84: class Program_weight_tensor_parameter_85: name = "parameter_85" + original_name = "conv2d_101.w_0" shape = [33, 128, 1, 1] dtype = "float32" - min_val = float("-0.5") - max_val = float("0.5") + min_val = float("-823874.0") + max_val = float("13.6964") + mean = float("-628.361") + std = float("14963.0") data = None class Program_weight_tensor_parameter_86: name = "parameter_86" + original_name = "batch_norm2d_77.b_0" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-93729.2") + max_val = float("245650.0") + mean = float("5390.82") + std = float("32965.1") data = None class Program_weight_tensor_parameter_87: name = "parameter_87" + original_name = "batch_norm2d_77.w_0" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-29877.5") + max_val = float("252064.0") + mean = float("3525.51") + std = float("26071.8") data = None class Program_weight_tensor_parameter_88: name = "parameter_88" + original_name = "batch_norm2d_77.w_2" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("1256560000.0") + max_val = float("2.75381e+20") + mean = float("6.70377e+18") + std = float("inf") data = None class Program_weight_tensor_parameter_89: name = "parameter_89" + original_name = "batch_norm2d_77.w_1" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-18448400000.0") + max_val = float("13078200000.0") + mean = float("-31256100.0") + std = float("2219470000.0") data = None class Program_weight_tensor_parameter_90: name = "parameter_90" + original_name = "conv2d_81.w_0" shape = [128, 128, 1, 1] dtype = "float32" - min_val = float("-0.5") - max_val = float("0.5") + min_val = float("-210775.0") + max_val = float("255860.0") + mean = float("5.87248") + std = float("6313.0") data = None class Program_weight_tensor_parameter_91: name = "parameter_91" + original_name = "batch_norm2d_76.b_0" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-274863.0") + max_val = float("103392.0") + mean = float("-3753.37") + std = float("32628.5") data = None class Program_weight_tensor_parameter_92: name = "parameter_92" + original_name = "batch_norm2d_76.w_0" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-116014.0") + max_val = float("105656.0") + mean = float("-1460.97") + std = float("22152.4") data = None class Program_weight_tensor_parameter_93: name = "parameter_93" + original_name = "batch_norm2d_76.w_2" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("1.40568e-05") + max_val = float("4.05002e+20") + mean = float("1.03691e+19") + std = float("inf") data = None class Program_weight_tensor_parameter_94: name = "parameter_94" + original_name = "batch_norm2d_76.w_1" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-33518000000.0") + max_val = float("9649450000.0") + mean = float("-209845000.0") + std = float("3196110000.0") data = None class Program_weight_tensor_parameter_95: name = "parameter_95" + original_name = "conv2d_80.w_0" shape = [128, 1, 5, 5] dtype = "float32" - min_val = float("-0.5") - max_val = float("0.5") + min_val = float("-79093.5") + max_val = float("99306.2") + mean = float("82.9882") + std = float("8433.19") data = None class Program_weight_tensor_parameter_96: name = "parameter_96" + original_name = "batch_norm2d_75.b_0" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-232696.0") + max_val = float("114264.0") + mean = float("-3885.22") + std = float("30265.4") data = None class Program_weight_tensor_parameter_97: name = "parameter_97" + original_name = "batch_norm2d_75.w_0" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-136351.0") + max_val = float("303355.0") + mean = float("2293.62") + std = float("32246.7") data = None class Program_weight_tensor_parameter_98: name = "parameter_98" + original_name = "batch_norm2d_75.w_2" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("37547100000.0") + max_val = float("2.3025e+21") + mean = float("4.36871e+19") + std = float("inf") data = None class Program_weight_tensor_parameter_99: name = "parameter_99" + original_name = "batch_norm2d_75.w_1" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-3777800000.0") + max_val = float("25793100000.0") + mean = float("647896000.0") + std = float("3109750000.0") data = None class Program_weight_tensor_parameter_100: name = "parameter_100" + original_name = "conv2d_79.w_0" shape = [128, 128, 1, 1] dtype = "float32" - min_val = float("-0.5") - max_val = float("0.5") + min_val = float("-153705.0") + max_val = float("870333.0") + mean = float("170.504") + std = float("7840.61") data = None class Program_weight_tensor_parameter_101: name = "parameter_101" + original_name = "batch_norm2d_74.b_0" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-76325.2") + max_val = float("18022.2") + mean = float("-865.982") + std = float("8585.91") data = None class Program_weight_tensor_parameter_102: name = "parameter_102" + original_name = "batch_norm2d_74.w_0" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-410378.0") + max_val = float("166971.0") + mean = float("-3527.54") + std = float("46161.8") data = None class Program_weight_tensor_parameter_103: name = "parameter_103" + original_name = "batch_norm2d_74.w_2" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("4.68528e-06") + max_val = float("6.52868e+22") + mean = float("5.83095e+20") + std = float("inf") data = None class Program_weight_tensor_parameter_104: name = "parameter_104" + original_name = "batch_norm2d_74.w_1" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-36490500000.0") + max_val = float("482921000000.0") + mean = float("3732870000.0") + std = float("42791300000.0") data = None class Program_weight_tensor_parameter_105: name = "parameter_105" + original_name = "conv2d_78.w_0" shape = [128, 1, 5, 5] dtype = "float32" - min_val = float("-0.5") - max_val = float("0.5") + min_val = float("-42075.6") + max_val = float("904287.0") + mean = float("2982.04") + std = float("34371.4") data = None class Program_weight_tensor_parameter_106: name = "parameter_106" + original_name = "batch_norm2d_73.b_0" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-41159.6") + max_val = float("97743.5") + mean = float("2018.04") + std = float("14221.8") data = None class Program_weight_tensor_parameter_107: name = "parameter_107" + original_name = "batch_norm2d_73.w_0" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-86577.1") + max_val = float("244368.0") + mean = float("3856.38") + std = float("31713.6") data = None class Program_weight_tensor_parameter_108: name = "parameter_108" + original_name = "batch_norm2d_73.w_2" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("39533200000.0") + max_val = float("2.72523e+20") + mean = float("6.79091e+18") + std = float("inf") data = None class Program_weight_tensor_parameter_109: name = "parameter_109" + original_name = "batch_norm2d_73.w_1" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-4719580000.0") + max_val = float("9380190000.0") + mean = float("36117900.0") + std = float("1408010000.0") data = None class Program_weight_tensor_parameter_110: name = "parameter_110" + original_name = "conv2d_77.w_0" shape = [128, 128, 1, 1] dtype = "float32" - min_val = float("-0.5") - max_val = float("0.5") + min_val = float("-131862.0") + max_val = float("155690.0") + mean = float("-3.24126") + std = float("3304.22") data = None class Program_weight_tensor_parameter_111: name = "parameter_111" + original_name = "batch_norm2d_72.b_0" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-22895.6") + max_val = float("37366.4") + mean = float("53.4149") + std = float("6580.69") data = None class Program_weight_tensor_parameter_112: name = "parameter_112" + original_name = "batch_norm2d_72.w_0" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-84375.4") + max_val = float("99002.5") + mean = float("1751.88") + std = float("16543.5") data = None class Program_weight_tensor_parameter_113: name = "parameter_113" + original_name = "batch_norm2d_72.w_2" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("4.3711e-05") + max_val = float("6.69868e+20") + mean = float("1.90444e+19") + std = float("inf") data = None class Program_weight_tensor_parameter_114: name = "parameter_114" + original_name = "batch_norm2d_72.w_1" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-32210600000.0") + max_val = float("19296400000.0") + mean = float("-222346000.0") + std = float("3712630000.0") data = None class Program_weight_tensor_parameter_115: name = "parameter_115" + original_name = "conv2d_76.w_0" shape = [128, 1, 5, 5] dtype = "float32" - min_val = float("-0.5") - max_val = float("0.5") + min_val = float("-97902.7") + max_val = float("208884.0") + mean = float("201.586") + std = float("10257.7") data = None class Program_weight_tensor_parameter_116: name = "parameter_116" + original_name = "batch_norm2d_71.b_0" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-179354.0") + max_val = float("94265.3") + mean = float("305.636") + std = float("22342.2") data = None class Program_weight_tensor_parameter_117: name = "parameter_117" + original_name = "batch_norm2d_71.w_0" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-487725.0") + max_val = float("75441.7") + mean = float("-6466.6") + std = float("48156.5") data = None class Program_weight_tensor_parameter_118: name = "parameter_118" + original_name = "batch_norm2d_71.w_2" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("131404000000.0") + max_val = float("6.8141e+21") + mean = float("1.01263e+20") + std = float("inf") data = None class Program_weight_tensor_parameter_119: name = "parameter_119" + original_name = "batch_norm2d_71.w_1" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-44285100000.0") + max_val = float("3857900000.0") + mean = float("-921017000.0") + std = float("5922380000.0") data = None class Program_weight_tensor_parameter_120: name = "parameter_120" + original_name = "conv2d_75.w_0" shape = [128, 128, 1, 1] dtype = "float32" - min_val = float("-0.5") - max_val = float("0.5") + min_val = float("-792833.0") + max_val = float("106772.0") + mean = float("-149.312") + std = float("9197.89") data = None class Program_weight_tensor_parameter_121: name = "parameter_121" + original_name = "batch_norm2d_70.b_0" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-45871.1") + max_val = float("37039.7") + mean = float("-1274.91") + std = float("9259.08") data = None class Program_weight_tensor_parameter_122: name = "parameter_122" + original_name = "batch_norm2d_70.w_0" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-125005.0") + max_val = float("240761.0") + mean = float("-612.89") + std = float("33942.6") data = None class Program_weight_tensor_parameter_123: name = "parameter_123" + original_name = "batch_norm2d_70.w_2" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("3.88607e-06") + max_val = float("3.93957e+20") + mean = float("1.01325e+19") + std = float("inf") data = None class Program_weight_tensor_parameter_124: name = "parameter_124" + original_name = "batch_norm2d_70.w_1" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-30152100000.0") + max_val = float("10858700000.0") + mean = float("-434912000.0") + std = float("4031290000.0") data = None class Program_weight_tensor_parameter_125: name = "parameter_125" + original_name = "conv2d_74.w_0" shape = [128, 1, 5, 5] dtype = "float32" - min_val = float("-0.5") - max_val = float("0.5") + min_val = float("-77094.2") + max_val = float("119918.0") + mean = float("156.161") + std = float("9081.22") data = None class Program_weight_tensor_parameter_126: name = "parameter_126" + original_name = "conv2d_100.b_0" shape = [33] dtype = "float32" min_val = float("0") @@ -1147,600 +1516,799 @@ class Program_weight_tensor_parameter_126: class Program_weight_tensor_parameter_127: name = "parameter_127" + original_name = "conv2d_100.w_0" shape = [33, 128, 1, 1] dtype = "float32" - min_val = float("-0.5") - max_val = float("0.5") + min_val = float("-468.888") + max_val = float("27.5042") + mean = float("-0.569604") + std = float("9.00485") data = None class Program_weight_tensor_parameter_128: name = "parameter_128" + original_name = "batch_norm2d_69.b_0" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-82.0118") + max_val = float("127.528") + mean = float("-5.15686") + std = float("30.2407") data = None class Program_weight_tensor_parameter_129: name = "parameter_129" + original_name = "batch_norm2d_69.w_0" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-51.5873") + max_val = float("166.429") + mean = float("1.47682") + std = float("23.6335") data = None class Program_weight_tensor_parameter_130: name = "parameter_130" + original_name = "batch_norm2d_69.w_2" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("3228.36") + max_val = float("101589000.0") + mean = float("4930090.0") + std = float("14493200.0") data = None class Program_weight_tensor_parameter_131: name = "parameter_131" + original_name = "batch_norm2d_69.w_1" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-3009.92") + max_val = float("4340.97") + mean = float("-66.7597") + std = float("874.949") data = None class Program_weight_tensor_parameter_132: name = "parameter_132" + original_name = "conv2d_73.w_0" shape = [128, 128, 1, 1] dtype = "float32" - min_val = float("-0.5") - max_val = float("0.5") + min_val = float("-103.697") + max_val = float("99.6534") + mean = float("0.00775166") + std = float("3.89286") data = None class Program_weight_tensor_parameter_133: name = "parameter_133" + original_name = "batch_norm2d_68.b_0" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-15.1906") + max_val = float("21.8346") + mean = float("0.175516") + std = float("5.38915") data = None class Program_weight_tensor_parameter_134: name = "parameter_134" + original_name = "batch_norm2d_68.w_0" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-94.8118") + max_val = float("36.2165") + mean = float("0.88053") + std = float("13.9991") data = None class Program_weight_tensor_parameter_135: name = "parameter_135" + original_name = "batch_norm2d_68.w_2" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("0.003342") + max_val = float("260327000.0") + mean = float("5286810.0") + std = float("29247200.0") data = None class Program_weight_tensor_parameter_136: name = "parameter_136" + original_name = "batch_norm2d_68.w_1" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-3332.74") + max_val = float("28189.5") + mean = float("383.105") + std = float("2752.61") data = None class Program_weight_tensor_parameter_137: name = "parameter_137" + original_name = "conv2d_72.w_0" shape = [128, 1, 5, 5] dtype = "float32" - min_val = float("-0.5") - max_val = float("0.5") + min_val = float("-73.2403") + max_val = float("40.042") + mean = float("0.103104") + std = float("5.58779") data = None class Program_weight_tensor_parameter_138: name = "parameter_138" + original_name = "batch_norm2d_67.b_0" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-57.6623") + max_val = float("215.244") + mean = float("1.2782") + std = float("24.4074") data = None class Program_weight_tensor_parameter_139: name = "parameter_139" + original_name = "batch_norm2d_67.w_0" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-113.865") + max_val = float("170.717") + mean = float("3.4637") + std = float("24.0477") data = None class Program_weight_tensor_parameter_140: name = "parameter_140" + original_name = "batch_norm2d_67.w_2" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("818.739") + max_val = float("44282300.0") + mean = float("1670460.0") + std = float("6291510.0") data = None class Program_weight_tensor_parameter_141: name = "parameter_141" + original_name = "batch_norm2d_67.w_1" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-2165.3") + max_val = float("3949.8") + mean = float("106.898") + std = float("695.056") data = None class Program_weight_tensor_parameter_142: name = "parameter_142" + original_name = "conv2d_71.w_0" shape = [128, 128, 1, 1] dtype = "float32" - min_val = float("-0.5") - max_val = float("0.5") + min_val = float("-65.2065") + max_val = float("77.8282") + mean = float("0.0883105") + std = float("3.09941") data = None class Program_weight_tensor_parameter_143: name = "parameter_143" + original_name = "batch_norm2d_66.b_0" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-19.8648") + max_val = float("14.432") + mean = float("-0.271648") + std = float("3.96903") data = None class Program_weight_tensor_parameter_144: name = "parameter_144" + original_name = "batch_norm2d_66.w_0" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-36.0747") + max_val = float("66.766") + mean = float("2.29129") + std = float("11.3359") data = None class Program_weight_tensor_parameter_145: name = "parameter_145" + original_name = "batch_norm2d_66.w_2" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("0.00133377") + max_val = float("6346100000.0") + mean = float("54220400.0") + std = float("558920000.0") data = None class Program_weight_tensor_parameter_146: name = "parameter_146" + original_name = "batch_norm2d_66.w_1" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-8008.63") + max_val = float("41689.7") + mean = float("686.51") + std = float("4509.7") data = None class Program_weight_tensor_parameter_147: name = "parameter_147" + original_name = "conv2d_70.w_0" shape = [128, 1, 5, 5] dtype = "float32" - min_val = float("-0.5") - max_val = float("0.5") + min_val = float("-36.7345") + max_val = float("58.2389") + mean = float("0.11292") + std = float("5.37781") data = None class Program_weight_tensor_parameter_148: name = "parameter_148" + original_name = "batch_norm2d_65.b_0" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-40.4075") + max_val = float("62.0484") + mean = float("1.86309") + std = float("11.8903") data = None class Program_weight_tensor_parameter_149: name = "parameter_149" + original_name = "batch_norm2d_65.w_0" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-76.1972") + max_val = float("526.74") + mean = float("6.85971") + std = float("49.8881") data = None class Program_weight_tensor_parameter_150: name = "parameter_150" + original_name = "batch_norm2d_65.w_2" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("473.663") + max_val = float("19018900.0") + mean = float("1142690.0") + std = float("3006780.0") data = None class Program_weight_tensor_parameter_151: name = "parameter_151" + original_name = "batch_norm2d_65.w_1" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-2675.49") + max_val = float("5130.18") + mean = float("-65.0968") + std = float("773.894") data = None class Program_weight_tensor_parameter_152: name = "parameter_152" + original_name = "conv2d_69.w_0" shape = [128, 128, 1, 1] dtype = "float32" - min_val = float("-0.5") - max_val = float("0.5") + min_val = float("-51.7617") + max_val = float("102.068") + mean = float("-0.0261933") + std = float("2.53584") data = None class Program_weight_tensor_parameter_153: name = "parameter_153" + original_name = "batch_norm2d_64.b_0" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-16.4991") + max_val = float("32.76") + mean = float("0.506381") + std = float("5.43") data = None class Program_weight_tensor_parameter_154: name = "parameter_154" + original_name = "batch_norm2d_64.w_0" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-24.5138") + max_val = float("91.9106") + mean = float("2.81984") + std = float("11.0954") data = None class Program_weight_tensor_parameter_155: name = "parameter_155" + original_name = "batch_norm2d_64.w_2" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("0.00890485") + max_val = float("924151000.0") + mean = float("10677200.0") + std = float("84650300.0") data = None class Program_weight_tensor_parameter_156: name = "parameter_156" + original_name = "batch_norm2d_64.w_1" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-9411.82") + max_val = float("17154.5") + mean = float("235.661") + std = float("2131.56") data = None class Program_weight_tensor_parameter_157: name = "parameter_157" + original_name = "conv2d_68.w_0" shape = [128, 1, 5, 5] dtype = "float32" - min_val = float("-0.5") - max_val = float("0.5") + min_val = float("-18.8293") + max_val = float("62.2737") + mean = float("0.344756") + std = float("4.71471") data = None class Program_weight_tensor_parameter_158: name = "parameter_158" + original_name = "batch_norm2d_63.b_0" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-55.2683") + max_val = float("158.105") + mean = float("2.12354") + std = float("18.1255") data = None class Program_weight_tensor_parameter_159: name = "parameter_159" + original_name = "batch_norm2d_63.w_0" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-91.5053") + max_val = float("177.076") + mean = float("3.11946") + std = float("26.193") data = None class Program_weight_tensor_parameter_160: name = "parameter_160" + original_name = "batch_norm2d_63.w_2" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("287.46") + max_val = float("42264800.0") + mean = float("830575.0") + std = float("4131720.0") data = None class Program_weight_tensor_parameter_161: name = "parameter_161" + original_name = "batch_norm2d_63.w_1" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-7298.6") + max_val = float("12899.9") + mean = float("184.813") + std = float("1576.31") data = None class Program_weight_tensor_parameter_162: name = "parameter_162" + original_name = "conv2d_67.w_0" shape = [128, 128, 1, 1] dtype = "float32" - min_val = float("-0.5") - max_val = float("0.5") + min_val = float("-103.91") + max_val = float("190.112") + mean = float("0.169758") + std = float("4.2492") data = None class Program_weight_tensor_parameter_163: name = "parameter_163" + original_name = "batch_norm2d_62.b_0" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-32.0046") + max_val = float("41.8268") + mean = float("0.450423") + std = float("9.05095") data = None class Program_weight_tensor_parameter_164: name = "parameter_164" + original_name = "batch_norm2d_62.w_0" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-36.0147") + max_val = float("60.5265") + mean = float("1.98405") + std = float("14.2687") data = None class Program_weight_tensor_parameter_165: name = "parameter_165" + original_name = "batch_norm2d_62.w_2" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("0.000128292") + max_val = float("7539160000000.0") + mean = float("120339000000.0") + std = float("872477000000.0") data = None class Program_weight_tensor_parameter_166: name = "parameter_166" + original_name = "batch_norm2d_62.w_1" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-5927410.0") + max_val = float("3255460.0") + mean = float("-21456.2") + std = float("605831.0") data = None class Program_weight_tensor_parameter_167: name = "parameter_167" + original_name = "conv2d_66.w_0" shape = [128, 1, 5, 5] dtype = "float32" - min_val = float("-0.5") - max_val = float("0.5") + min_val = float("-48.7218") + max_val = float("39.6093") + mean = float("0.0303446") + std = float("4.60074") data = None class Program_weight_tensor_parameter_168: name = "parameter_168" + original_name = "batch_norm2d_33.b_0" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-4.87282") + max_val = float("9.13118") + mean = float("0.255483") + std = float("1.57109") data = None class Program_weight_tensor_parameter_169: name = "parameter_169" + original_name = "batch_norm2d_33.w_0" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-9.06693") + max_val = float("6.04652") + mean = float("0.101314") + std = float("2.22412") data = None class Program_weight_tensor_parameter_170: name = "parameter_170" + original_name = "batch_norm2d_33.w_2" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("0.893517") + max_val = float("8213.64") + mean = float("306.764") + std = float("839.585") data = None class Program_weight_tensor_parameter_171: name = "parameter_171" + original_name = "batch_norm2d_33.w_1" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-83.3458") + max_val = float("34.4687") + mean = float("-1.76345") + std = float("14.1703") data = None class Program_weight_tensor_parameter_172: name = "parameter_172" + original_name = "conv2d_37.w_0" shape = [128, 128, 1, 1] dtype = "float32" - min_val = float("-0.5") - max_val = float("0.5") + min_val = float("-12.8666") + max_val = float("9.25836") + mean = float("-0.0020712") + std = float("0.438125") data = None class Program_weight_tensor_parameter_173: name = "parameter_173" + original_name = "batch_norm2d_32.b_0" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-0.923822") + max_val = float("8.06789") + mean = float("0.0833533") + std = float("0.807324") data = None class Program_weight_tensor_parameter_174: name = "parameter_174" + original_name = "batch_norm2d_32.w_0" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-6.61104") + max_val = float("9.9265") + mean = float("0.253523") + std = float("1.53549") data = None class Program_weight_tensor_parameter_175: name = "parameter_175" + original_name = "batch_norm2d_32.w_2" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("0.000156346") + max_val = float("391558000.0") + mean = float("3072800.0") + std = float("34472500.0") data = None class Program_weight_tensor_parameter_176: name = "parameter_176" + original_name = "batch_norm2d_32.w_1" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-1911.58") + max_val = float("10317.2") + mean = float("69.531") + std = float("929.701") data = None class Program_weight_tensor_parameter_177: name = "parameter_177" + original_name = "conv2d_36.w_0" shape = [128, 1, 5, 5] dtype = "float32" - min_val = float("-0.5") - max_val = float("0.5") + min_val = float("-9.02725") + max_val = float("20.3279") + mean = float("-0.00291877") + std = float("0.962834") data = None class Program_weight_tensor_parameter_178: name = "parameter_178" + original_name = "batch_norm2d_31.b_0" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-9.29749") + max_val = float("3.43862") + mean = float("-0.0410556") + std = float("1.33901") data = None class Program_weight_tensor_parameter_179: name = "parameter_179" + original_name = "batch_norm2d_31.w_0" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-32.9479") + max_val = float("6.66634") + mean = float("-0.0304042") + std = float("3.75495") data = None class Program_weight_tensor_parameter_180: name = "parameter_180" + original_name = "batch_norm2d_31.w_2" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("2.3579") + max_val = float("2470.7") + mean = float("95.5496") + std = float("260.919") data = None class Program_weight_tensor_parameter_181: name = "parameter_181" + original_name = "batch_norm2d_31.w_1" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-30.8253") + max_val = float("28.8304") + mean = float("0.435417") + std = float("8.50347") data = None class Program_weight_tensor_parameter_182: name = "parameter_182" + original_name = "conv2d_35.w_0" shape = [128, 128, 1, 1] dtype = "float32" - min_val = float("-0.5") - max_val = float("0.5") + min_val = float("-9.81677") + max_val = float("10.8001") + mean = float("0.000582034") + std = float("0.387895") data = None class Program_weight_tensor_parameter_183: name = "parameter_183" + original_name = "batch_norm2d_30.b_0" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-1.97589") + max_val = float("3.28159") + mean = float("0.0251685") + std = float("0.588334") data = None class Program_weight_tensor_parameter_184: name = "parameter_184" + original_name = "batch_norm2d_30.w_0" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-16.0092") + max_val = float("9.1929") + mean = float("-0.372784") + std = float("2.03039") data = None class Program_weight_tensor_parameter_185: name = "parameter_185" + original_name = "batch_norm2d_30.w_2" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("5.86101e-05") + max_val = float("12261400000000.0") + mean = float("105002000000.0") + std = float("1081960000000.0") data = None class Program_weight_tensor_parameter_186: name = "parameter_186" + original_name = "batch_norm2d_30.w_1" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-528399.0") + max_val = float("1345590.0") + mean = float("-1085.89") + std = float("134658.0") data = None class Program_weight_tensor_parameter_187: name = "parameter_187" + original_name = "conv2d_34.w_0" shape = [128, 1, 5, 5] dtype = "float32" - min_val = float("-0.5") - max_val = float("0.5") + min_val = float("-7.94225") + max_val = float("25.4521") + mean = float("0.00719978") + std = float("1.01759") data = None class Program_weight_tensor_parameter_188: name = "parameter_188" + original_name = "batch_norm2d_58.b_0" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-376.422") + max_val = float("894.238") + mean = float("5.5682") + std = float("95.7981") data = None class Program_weight_tensor_parameter_189: name = "parameter_189" + original_name = "batch_norm2d_58.w_0" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-8336.48") + max_val = float("686.181") + mean = float("-60.5026") + std = float("738.032") data = None class Program_weight_tensor_parameter_190: name = "parameter_190" + original_name = "batch_norm2d_58.w_2" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("14777.6") + max_val = float("6550930000.0") + mean = float("152320000.0") + std = float("760843000.0") data = None class Program_weight_tensor_parameter_191: name = "parameter_191" + original_name = "batch_norm2d_58.w_1" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-50080.1") + max_val = float("30125.9") + mean = float("-1057.73") + std = float("6672.84") data = None class Program_weight_tensor_parameter_192: name = "parameter_192" + original_name = "conv2d_62.w_0" shape = [128, 128, 1, 1] dtype = "float32" - min_val = float("-0.5") - max_val = float("0.5") + min_val = float("-362.557") + max_val = float("172.702") + mean = float("-0.327888") + std = float("10.0138") data = None class Program_weight_tensor_parameter_193: name = "parameter_193" + original_name = "batch_norm2d_61.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -1750,6 +2318,7 @@ class Program_weight_tensor_parameter_193: class Program_weight_tensor_parameter_194: name = "parameter_194" + original_name = "batch_norm2d_61.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -1759,6 +2328,7 @@ class Program_weight_tensor_parameter_194: class Program_weight_tensor_parameter_195: name = "parameter_195" + original_name = "batch_norm2d_61.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -1768,6 +2338,7 @@ class Program_weight_tensor_parameter_195: class Program_weight_tensor_parameter_196: name = "parameter_196" + original_name = "batch_norm2d_61.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -1777,15 +2348,19 @@ class Program_weight_tensor_parameter_196: class Program_weight_tensor_parameter_197: name = "parameter_197" + original_name = "conv2d_65.w_0" shape = [64, 64, 1, 1] dtype = "float32" - min_val = float("-0.5") - max_val = float("0.5") + min_val = float("-183.511") + max_val = float("147.697") + mean = float("0.025535") + std = float("7.38341") data = None class Program_weight_tensor_parameter_198: name = "parameter_198" + original_name = "batch_norm2d_60.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -1795,6 +2370,7 @@ class Program_weight_tensor_parameter_198: class Program_weight_tensor_parameter_199: name = "parameter_199" + original_name = "batch_norm2d_60.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -1804,6 +2380,7 @@ class Program_weight_tensor_parameter_199: class Program_weight_tensor_parameter_200: name = "parameter_200" + original_name = "batch_norm2d_60.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -1813,6 +2390,7 @@ class Program_weight_tensor_parameter_200: class Program_weight_tensor_parameter_201: name = "parameter_201" + original_name = "batch_norm2d_60.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -1822,15 +2400,19 @@ class Program_weight_tensor_parameter_201: class Program_weight_tensor_parameter_202: name = "parameter_202" + original_name = "conv2d_64.w_0" shape = [64, 1, 5, 5] dtype = "float32" - min_val = float("-0.5") - max_val = float("0.5") + min_val = float("-374.46") + max_val = float("525.953") + mean = float("1.30091") + std = float("27.3902") data = None class Program_weight_tensor_parameter_203: name = "parameter_203" + original_name = "batch_norm2d_59.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -1840,6 +2422,7 @@ class Program_weight_tensor_parameter_203: class Program_weight_tensor_parameter_204: name = "parameter_204" + original_name = "batch_norm2d_59.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -1849,6 +2432,7 @@ class Program_weight_tensor_parameter_204: class Program_weight_tensor_parameter_205: name = "parameter_205" + original_name = "batch_norm2d_59.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -1858,6 +2442,7 @@ class Program_weight_tensor_parameter_205: class Program_weight_tensor_parameter_206: name = "parameter_206" + original_name = "batch_norm2d_59.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -1867,15 +2452,19 @@ class Program_weight_tensor_parameter_206: class Program_weight_tensor_parameter_207: name = "parameter_207" + original_name = "conv2d_63.w_0" shape = [64, 64, 1, 1] dtype = "float32" - min_val = float("-0.5") - max_val = float("0.5") + min_val = float("-396.785") + max_val = float("315.363") + mean = float("-0.212783") + std = float("12.5802") data = None class Program_weight_tensor_parameter_208: name = "parameter_208" + original_name = "batch_norm2d_56.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -1885,6 +2474,7 @@ class Program_weight_tensor_parameter_208: class Program_weight_tensor_parameter_209: name = "parameter_209" + original_name = "batch_norm2d_56.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -1894,6 +2484,7 @@ class Program_weight_tensor_parameter_209: class Program_weight_tensor_parameter_210: name = "parameter_210" + original_name = "batch_norm2d_56.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -1903,6 +2494,7 @@ class Program_weight_tensor_parameter_210: class Program_weight_tensor_parameter_211: name = "parameter_211" + original_name = "batch_norm2d_56.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -1912,15 +2504,19 @@ class Program_weight_tensor_parameter_211: class Program_weight_tensor_parameter_212: name = "parameter_212" + original_name = "conv2d_60.w_0" shape = [64, 256, 1, 1] dtype = "float32" - min_val = float("-0.5") - max_val = float("0.5") + min_val = float("-147.868") + max_val = float("207.3") + mean = float("-0.0228569") + std = float("4.13461") data = None class Program_weight_tensor_parameter_213: name = "parameter_213" + original_name = "batch_norm2d_57.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -1930,6 +2526,7 @@ class Program_weight_tensor_parameter_213: class Program_weight_tensor_parameter_214: name = "parameter_214" + original_name = "batch_norm2d_57.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -1939,6 +2536,7 @@ class Program_weight_tensor_parameter_214: class Program_weight_tensor_parameter_215: name = "parameter_215" + original_name = "batch_norm2d_57.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -1948,6 +2546,7 @@ class Program_weight_tensor_parameter_215: class Program_weight_tensor_parameter_216: name = "parameter_216" + original_name = "batch_norm2d_57.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -1957,51 +2556,67 @@ class Program_weight_tensor_parameter_216: class Program_weight_tensor_parameter_217: name = "parameter_217" + original_name = "conv2d_61.w_0" shape = [64, 256, 1, 1] dtype = "float32" - min_val = float("-0.5") - max_val = float("0.5") + min_val = float("-431.692") + max_val = float("155.313") + mean = float("-0.20306") + std = float("5.7531") data = None class Program_weight_tensor_parameter_218: name = "parameter_218" + original_name = "batch_norm2d_55.b_0" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-0.758811") + max_val = float("0.309114") + mean = float("-0.0539417") + std = float("0.127762") data = None class Program_weight_tensor_parameter_219: name = "parameter_219" + original_name = "batch_norm2d_55.w_0" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-0.849583") + max_val = float("1.86193") + mean = float("0.169073") + std = float("0.346604") data = None class Program_weight_tensor_parameter_220: name = "parameter_220" + original_name = "batch_norm2d_55.w_2" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("0.000854349") + max_val = float("13.4449") + mean = float("0.555821") + std = float("1.4836") data = None class Program_weight_tensor_parameter_221: name = "parameter_221" + original_name = "batch_norm2d_55.w_1" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-1.06848") + max_val = float("3.94871") + mean = float("-0.0112817") + std = float("0.469857") data = None class Program_weight_tensor_parameter_222: name = "parameter_222" + original_name = "conv2d_59.w_0" shape = [128, 128, 1, 1] dtype = "float32" min_val = float("-1.3054") @@ -2013,42 +2628,55 @@ class Program_weight_tensor_parameter_222: class Program_weight_tensor_parameter_223: name = "parameter_223" + original_name = "batch_norm2d_54.b_0" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-0.229151") + max_val = float("0.383636") + mean = float("-0.00581602") + std = float("0.0761077") data = None class Program_weight_tensor_parameter_224: name = "parameter_224" + original_name = "batch_norm2d_54.w_0" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-1.47951") + max_val = float("2.31176") + mean = float("0.189723") + std = float("0.520794") data = None class Program_weight_tensor_parameter_225: name = "parameter_225" + original_name = "batch_norm2d_54.w_2" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("3.53324e-11") + max_val = float("61754200000.0") + mean = float("882636000.0") + std = float("5662340000.0") data = None class Program_weight_tensor_parameter_226: name = "parameter_226" + original_name = "batch_norm2d_54.w_1" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-94937.4") + max_val = float("440735.0") + mean = float("6123.79") + std = float("45723.3") data = None class Program_weight_tensor_parameter_227: name = "parameter_227" + original_name = "conv2d_58.w_0" shape = [128, 1, 5, 5] dtype = "float32" min_val = float("-1.88336") @@ -2060,51 +2688,67 @@ class Program_weight_tensor_parameter_227: class Program_weight_tensor_parameter_228: name = "parameter_228" + original_name = "batch_norm2d_50.b_0" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-82868.3") + max_val = float("94602.9") + mean = float("-839.968") + std = float("19134.3") data = None class Program_weight_tensor_parameter_229: name = "parameter_229" + original_name = "batch_norm2d_50.w_0" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-65172.4") + max_val = float("190740.0") + mean = float("2172.17") + std = float("23677.1") data = None class Program_weight_tensor_parameter_230: name = "parameter_230" + original_name = "batch_norm2d_50.w_2" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("10737900000.0") + max_val = float("7.12963e+20") + mean = float("9.05393e+18") + std = float("inf") data = None class Program_weight_tensor_parameter_231: name = "parameter_231" + original_name = "batch_norm2d_50.w_1" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-6512240000.0") + max_val = float("32799100000.0") + mean = float("166414000.0") + std = float("3188900000.0") data = None class Program_weight_tensor_parameter_232: name = "parameter_232" + original_name = "conv2d_54.w_0" shape = [128, 128, 1, 1] dtype = "float32" - min_val = float("-0.5") - max_val = float("0.5") + min_val = float("-170848.0") + max_val = float("367603.0") + mean = float("-48.9465") + std = float("5426.47") data = None class Program_weight_tensor_parameter_233: name = "parameter_233" + original_name = "batch_norm2d_53.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -2114,6 +2758,7 @@ class Program_weight_tensor_parameter_233: class Program_weight_tensor_parameter_234: name = "parameter_234" + original_name = "batch_norm2d_53.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -2123,6 +2768,7 @@ class Program_weight_tensor_parameter_234: class Program_weight_tensor_parameter_235: name = "parameter_235" + original_name = "batch_norm2d_53.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -2132,6 +2778,7 @@ class Program_weight_tensor_parameter_235: class Program_weight_tensor_parameter_236: name = "parameter_236" + original_name = "batch_norm2d_53.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -2141,15 +2788,19 @@ class Program_weight_tensor_parameter_236: class Program_weight_tensor_parameter_237: name = "parameter_237" + original_name = "conv2d_57.w_0" shape = [64, 64, 1, 1] dtype = "float32" - min_val = float("-0.5") - max_val = float("0.5") + min_val = float("-201543.0") + max_val = float("126826.0") + mean = float("165.659") + std = float("5500.63") data = None class Program_weight_tensor_parameter_238: name = "parameter_238" + original_name = "batch_norm2d_52.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -2159,6 +2810,7 @@ class Program_weight_tensor_parameter_238: class Program_weight_tensor_parameter_239: name = "parameter_239" + original_name = "batch_norm2d_52.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -2168,6 +2820,7 @@ class Program_weight_tensor_parameter_239: class Program_weight_tensor_parameter_240: name = "parameter_240" + original_name = "batch_norm2d_52.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -2177,6 +2830,7 @@ class Program_weight_tensor_parameter_240: class Program_weight_tensor_parameter_241: name = "parameter_241" + original_name = "batch_norm2d_52.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -2186,15 +2840,19 @@ class Program_weight_tensor_parameter_241: class Program_weight_tensor_parameter_242: name = "parameter_242" + original_name = "conv2d_56.w_0" shape = [64, 1, 5, 5] dtype = "float32" - min_val = float("-0.5") - max_val = float("0.5") + min_val = float("-79066.4") + max_val = float("40183.0") + mean = float("-366.24") + std = float("7227.17") data = None class Program_weight_tensor_parameter_243: name = "parameter_243" + original_name = "batch_norm2d_51.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -2204,6 +2862,7 @@ class Program_weight_tensor_parameter_243: class Program_weight_tensor_parameter_244: name = "parameter_244" + original_name = "batch_norm2d_51.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -2213,6 +2872,7 @@ class Program_weight_tensor_parameter_244: class Program_weight_tensor_parameter_245: name = "parameter_245" + original_name = "batch_norm2d_51.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -2222,6 +2882,7 @@ class Program_weight_tensor_parameter_245: class Program_weight_tensor_parameter_246: name = "parameter_246" + original_name = "batch_norm2d_51.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -2231,15 +2892,19 @@ class Program_weight_tensor_parameter_246: class Program_weight_tensor_parameter_247: name = "parameter_247" + original_name = "conv2d_55.w_0" shape = [64, 64, 1, 1] dtype = "float32" - min_val = float("-0.5") - max_val = float("0.5") + min_val = float("-213890.0") + max_val = float("566581.0") + mean = float("179.605") + std = float("14075.6") data = None class Program_weight_tensor_parameter_248: name = "parameter_248" + original_name = "batch_norm2d_48.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -2249,6 +2914,7 @@ class Program_weight_tensor_parameter_248: class Program_weight_tensor_parameter_249: name = "parameter_249" + original_name = "batch_norm2d_48.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -2258,6 +2924,7 @@ class Program_weight_tensor_parameter_249: class Program_weight_tensor_parameter_250: name = "parameter_250" + original_name = "batch_norm2d_48.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -2267,6 +2934,7 @@ class Program_weight_tensor_parameter_250: class Program_weight_tensor_parameter_251: name = "parameter_251" + original_name = "batch_norm2d_48.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -2276,15 +2944,19 @@ class Program_weight_tensor_parameter_251: class Program_weight_tensor_parameter_252: name = "parameter_252" + original_name = "conv2d_52.w_0" shape = [64, 256, 1, 1] dtype = "float32" - min_val = float("-0.5") - max_val = float("0.5") + min_val = float("-158711.0") + max_val = float("86099.3") + mean = float("-74.8862") + std = float("3135.0") data = None class Program_weight_tensor_parameter_253: name = "parameter_253" + original_name = "batch_norm2d_49.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -2294,6 +2966,7 @@ class Program_weight_tensor_parameter_253: class Program_weight_tensor_parameter_254: name = "parameter_254" + original_name = "batch_norm2d_49.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -2303,6 +2976,7 @@ class Program_weight_tensor_parameter_254: class Program_weight_tensor_parameter_255: name = "parameter_255" + original_name = "batch_norm2d_49.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -2312,6 +2986,7 @@ class Program_weight_tensor_parameter_255: class Program_weight_tensor_parameter_256: name = "parameter_256" + original_name = "batch_norm2d_49.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -2321,150 +2996,199 @@ class Program_weight_tensor_parameter_256: class Program_weight_tensor_parameter_257: name = "parameter_257" + original_name = "conv2d_53.w_0" shape = [64, 256, 1, 1] dtype = "float32" - min_val = float("-0.5") - max_val = float("0.5") + min_val = float("-34433.7") + max_val = float("57521.0") + mean = float("33.2978") + std = float("1164.92") data = None class Program_weight_tensor_parameter_258: name = "parameter_258" + original_name = "batch_norm2d_47.b_0" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-42985.6") + max_val = float("4691.7") + mean = float("-940.023") + std = float("4554.53") data = None class Program_weight_tensor_parameter_259: name = "parameter_259" + original_name = "batch_norm2d_47.w_0" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-69389.2") + max_val = float("66657.7") + mean = float("233.771") + std = float("9886.5") data = None class Program_weight_tensor_parameter_260: name = "parameter_260" + original_name = "batch_norm2d_47.w_2" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("1289980000.0") + max_val = float("1.01952e+19") + mean = float("1.77883e+17") + std = float("1.09482e+18") data = None class Program_weight_tensor_parameter_261: name = "parameter_261" + original_name = "batch_norm2d_47.w_1" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-969451000.0") + max_val = float("1719590000.0") + mean = float("19588100.0") + std = float("230945000.0") data = None class Program_weight_tensor_parameter_262: name = "parameter_262" + original_name = "conv2d_51.w_0" shape = [128, 128, 1, 1] dtype = "float32" - min_val = float("-0.5") - max_val = float("0.5") + min_val = float("-54924.7") + max_val = float("74038.4") + mean = float("-6.87325") + std = float("940.882") data = None class Program_weight_tensor_parameter_263: name = "parameter_263" + original_name = "batch_norm2d_46.b_0" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-46456.9") + max_val = float("51379.1") + mean = float("-414.741") + std = float("6921.25") data = None class Program_weight_tensor_parameter_264: name = "parameter_264" + original_name = "batch_norm2d_46.w_0" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-18536.7") + max_val = float("99124.8") + mean = float("809.038") + std = float("9397.11") data = None class Program_weight_tensor_parameter_265: name = "parameter_265" + original_name = "batch_norm2d_46.w_2" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("0.000180454") + max_val = float("2.04952e+21") + mean = float("1.68005e+19") + std = float("inf") data = None class Program_weight_tensor_parameter_266: name = "parameter_266" + original_name = "batch_norm2d_46.w_1" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-760651000.0") + max_val = float("48053000000.0") + mean = float("482595000.0") + std = float("4368170000.0") data = None class Program_weight_tensor_parameter_267: name = "parameter_267" + original_name = "conv2d_50.w_0" shape = [128, 1, 5, 5] dtype = "float32" - min_val = float("-0.5") - max_val = float("0.5") + min_val = float("-23419.0") + max_val = float("60952.1") + mean = float("306.019") + std = float("3030.82") data = None class Program_weight_tensor_parameter_268: name = "parameter_268" + original_name = "batch_norm2d_42.b_0" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-130090.0") + max_val = float("108204.0") + mean = float("-158.833") + std = float("15935.3") data = None class Program_weight_tensor_parameter_269: name = "parameter_269" + original_name = "batch_norm2d_42.w_0" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-53015.6") + max_val = float("168304.0") + mean = float("-117.703") + std = float("17232.8") data = None class Program_weight_tensor_parameter_270: name = "parameter_270" + original_name = "batch_norm2d_42.w_2" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("29624000000.0") + max_val = float("3.50963e+20") + mean = float("8.11164e+18") + std = float("inf") data = None class Program_weight_tensor_parameter_271: name = "parameter_271" + original_name = "batch_norm2d_42.w_1" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-25856000000.0") + max_val = float("31882100000.0") + mean = float("-168986000.0") + std = float("3911860000.0") data = None class Program_weight_tensor_parameter_272: name = "parameter_272" + original_name = "conv2d_46.w_0" shape = [128, 128, 1, 1] dtype = "float32" - min_val = float("-0.5") - max_val = float("0.5") + min_val = float("-184337.0") + max_val = float("120696.0") + mean = float("5.00414") + std = float("3399.38") data = None class Program_weight_tensor_parameter_273: name = "parameter_273" + original_name = "batch_norm2d_45.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -2474,6 +3198,7 @@ class Program_weight_tensor_parameter_273: class Program_weight_tensor_parameter_274: name = "parameter_274" + original_name = "batch_norm2d_45.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -2483,6 +3208,7 @@ class Program_weight_tensor_parameter_274: class Program_weight_tensor_parameter_275: name = "parameter_275" + original_name = "batch_norm2d_45.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -2492,6 +3218,7 @@ class Program_weight_tensor_parameter_275: class Program_weight_tensor_parameter_276: name = "parameter_276" + original_name = "batch_norm2d_45.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -2501,15 +3228,19 @@ class Program_weight_tensor_parameter_276: class Program_weight_tensor_parameter_277: name = "parameter_277" + original_name = "conv2d_49.w_0" shape = [64, 64, 1, 1] dtype = "float32" - min_val = float("-0.5") - max_val = float("0.5") + min_val = float("-301068.0") + max_val = float("44085.2") + mean = float("-207.207") + std = float("6212.15") data = None class Program_weight_tensor_parameter_278: name = "parameter_278" + original_name = "batch_norm2d_44.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -2519,6 +3250,7 @@ class Program_weight_tensor_parameter_278: class Program_weight_tensor_parameter_279: name = "parameter_279" + original_name = "batch_norm2d_44.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -2528,6 +3260,7 @@ class Program_weight_tensor_parameter_279: class Program_weight_tensor_parameter_280: name = "parameter_280" + original_name = "batch_norm2d_44.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -2537,6 +3270,7 @@ class Program_weight_tensor_parameter_280: class Program_weight_tensor_parameter_281: name = "parameter_281" + original_name = "batch_norm2d_44.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -2546,15 +3280,19 @@ class Program_weight_tensor_parameter_281: class Program_weight_tensor_parameter_282: name = "parameter_282" + original_name = "conv2d_48.w_0" shape = [64, 1, 5, 5] dtype = "float32" - min_val = float("-0.5") - max_val = float("0.5") + min_val = float("-62839.4") + max_val = float("64846.8") + mean = float("-1704.03") + std = float("9405.05") data = None class Program_weight_tensor_parameter_283: name = "parameter_283" + original_name = "batch_norm2d_43.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -2564,6 +3302,7 @@ class Program_weight_tensor_parameter_283: class Program_weight_tensor_parameter_284: name = "parameter_284" + original_name = "batch_norm2d_43.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -2573,6 +3312,7 @@ class Program_weight_tensor_parameter_284: class Program_weight_tensor_parameter_285: name = "parameter_285" + original_name = "batch_norm2d_43.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -2582,6 +3322,7 @@ class Program_weight_tensor_parameter_285: class Program_weight_tensor_parameter_286: name = "parameter_286" + original_name = "batch_norm2d_43.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -2591,15 +3332,19 @@ class Program_weight_tensor_parameter_286: class Program_weight_tensor_parameter_287: name = "parameter_287" + original_name = "conv2d_47.w_0" shape = [64, 64, 1, 1] dtype = "float32" - min_val = float("-0.5") - max_val = float("0.5") + min_val = float("-171028.0") + max_val = float("147214.0") + mean = float("6.04784") + std = float("5460.79") data = None class Program_weight_tensor_parameter_288: name = "parameter_288" + original_name = "batch_norm2d_40.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -2609,6 +3354,7 @@ class Program_weight_tensor_parameter_288: class Program_weight_tensor_parameter_289: name = "parameter_289" + original_name = "batch_norm2d_40.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -2618,6 +3364,7 @@ class Program_weight_tensor_parameter_289: class Program_weight_tensor_parameter_290: name = "parameter_290" + original_name = "batch_norm2d_40.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -2627,6 +3374,7 @@ class Program_weight_tensor_parameter_290: class Program_weight_tensor_parameter_291: name = "parameter_291" + original_name = "batch_norm2d_40.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -2636,15 +3384,19 @@ class Program_weight_tensor_parameter_291: class Program_weight_tensor_parameter_292: name = "parameter_292" + original_name = "conv2d_44.w_0" shape = [64, 256, 1, 1] dtype = "float32" - min_val = float("-0.5") - max_val = float("0.5") + min_val = float("-77807.2") + max_val = float("117252.0") + mean = float("57.8441") + std = float("2647.75") data = None class Program_weight_tensor_parameter_293: name = "parameter_293" + original_name = "batch_norm2d_41.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -2654,6 +3406,7 @@ class Program_weight_tensor_parameter_293: class Program_weight_tensor_parameter_294: name = "parameter_294" + original_name = "batch_norm2d_41.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -2663,6 +3416,7 @@ class Program_weight_tensor_parameter_294: class Program_weight_tensor_parameter_295: name = "parameter_295" + original_name = "batch_norm2d_41.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -2672,6 +3426,7 @@ class Program_weight_tensor_parameter_295: class Program_weight_tensor_parameter_296: name = "parameter_296" + original_name = "batch_norm2d_41.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -2681,60 +3436,79 @@ class Program_weight_tensor_parameter_296: class Program_weight_tensor_parameter_297: name = "parameter_297" + original_name = "conv2d_45.w_0" shape = [64, 256, 1, 1] dtype = "float32" - min_val = float("-0.5") - max_val = float("0.5") + min_val = float("-200717.0") + max_val = float("95087.8") + mean = float("-81.1523") + std = float("3329.94") data = None class Program_weight_tensor_parameter_298: name = "parameter_298" + original_name = "batch_norm2d_36.b_0" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-73962.8") + max_val = float("43163.7") + mean = float("218.145") + std = float("10336.5") data = None class Program_weight_tensor_parameter_299: name = "parameter_299" + original_name = "batch_norm2d_36.w_0" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-101542.0") + max_val = float("130481.0") + mean = float("-213.081") + std = float("19353.6") data = None class Program_weight_tensor_parameter_300: name = "parameter_300" + original_name = "batch_norm2d_36.w_2" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("5133710000.0") + max_val = float("6.86238e+20") + mean = float("1.13767e+19") + std = float("inf") data = None class Program_weight_tensor_parameter_301: name = "parameter_301" + original_name = "batch_norm2d_36.w_1" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-10862500000.0") + max_val = float("21049400000.0") + mean = float("215672000.0") + std = float("2658170000.0") data = None class Program_weight_tensor_parameter_302: name = "parameter_302" + original_name = "conv2d_40.w_0" shape = [128, 128, 1, 1] dtype = "float32" - min_val = float("-0.5") - max_val = float("0.5") + min_val = float("-92814.5") + max_val = float("135289.0") + mean = float("40.2464") + std = float("3240.74") data = None class Program_weight_tensor_parameter_303: name = "parameter_303" + original_name = "batch_norm2d_39.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -2744,6 +3518,7 @@ class Program_weight_tensor_parameter_303: class Program_weight_tensor_parameter_304: name = "parameter_304" + original_name = "batch_norm2d_39.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -2753,6 +3528,7 @@ class Program_weight_tensor_parameter_304: class Program_weight_tensor_parameter_305: name = "parameter_305" + original_name = "batch_norm2d_39.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -2762,6 +3538,7 @@ class Program_weight_tensor_parameter_305: class Program_weight_tensor_parameter_306: name = "parameter_306" + original_name = "batch_norm2d_39.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -2771,15 +3548,19 @@ class Program_weight_tensor_parameter_306: class Program_weight_tensor_parameter_307: name = "parameter_307" + original_name = "conv2d_43.w_0" shape = [64, 64, 1, 1] dtype = "float32" - min_val = float("-0.5") - max_val = float("0.5") + min_val = float("-117059.0") + max_val = float("206924.0") + mean = float("165.362") + std = float("5930.1") data = None class Program_weight_tensor_parameter_308: name = "parameter_308" + original_name = "batch_norm2d_38.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -2789,6 +3570,7 @@ class Program_weight_tensor_parameter_308: class Program_weight_tensor_parameter_309: name = "parameter_309" + original_name = "batch_norm2d_38.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -2798,6 +3580,7 @@ class Program_weight_tensor_parameter_309: class Program_weight_tensor_parameter_310: name = "parameter_310" + original_name = "batch_norm2d_38.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -2807,6 +3590,7 @@ class Program_weight_tensor_parameter_310: class Program_weight_tensor_parameter_311: name = "parameter_311" + original_name = "batch_norm2d_38.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -2816,15 +3600,19 @@ class Program_weight_tensor_parameter_311: class Program_weight_tensor_parameter_312: name = "parameter_312" + original_name = "conv2d_42.w_0" shape = [64, 1, 5, 5] dtype = "float32" - min_val = float("-0.5") - max_val = float("0.5") + min_val = float("-53971.6") + max_val = float("124672.0") + mean = float("1035.04") + std = float("8747.79") data = None class Program_weight_tensor_parameter_313: name = "parameter_313" + original_name = "batch_norm2d_37.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -2834,6 +3622,7 @@ class Program_weight_tensor_parameter_313: class Program_weight_tensor_parameter_314: name = "parameter_314" + original_name = "batch_norm2d_37.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -2843,6 +3632,7 @@ class Program_weight_tensor_parameter_314: class Program_weight_tensor_parameter_315: name = "parameter_315" + original_name = "batch_norm2d_37.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -2852,6 +3642,7 @@ class Program_weight_tensor_parameter_315: class Program_weight_tensor_parameter_316: name = "parameter_316" + original_name = "batch_norm2d_37.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -2861,15 +3652,19 @@ class Program_weight_tensor_parameter_316: class Program_weight_tensor_parameter_317: name = "parameter_317" + original_name = "conv2d_41.w_0" shape = [64, 64, 1, 1] dtype = "float32" - min_val = float("-0.5") - max_val = float("0.5") + min_val = float("-233204.0") + max_val = float("176114.0") + mean = float("-60.6732") + std = float("7319.98") data = None class Program_weight_tensor_parameter_318: name = "parameter_318" + original_name = "batch_norm2d_34.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -2879,6 +3674,7 @@ class Program_weight_tensor_parameter_318: class Program_weight_tensor_parameter_319: name = "parameter_319" + original_name = "batch_norm2d_34.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -2888,6 +3684,7 @@ class Program_weight_tensor_parameter_319: class Program_weight_tensor_parameter_320: name = "parameter_320" + original_name = "batch_norm2d_34.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -2897,6 +3694,7 @@ class Program_weight_tensor_parameter_320: class Program_weight_tensor_parameter_321: name = "parameter_321" + original_name = "batch_norm2d_34.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -2906,15 +3704,19 @@ class Program_weight_tensor_parameter_321: class Program_weight_tensor_parameter_322: name = "parameter_322" + original_name = "conv2d_38.w_0" shape = [64, 256, 1, 1] dtype = "float32" - min_val = float("-0.5") - max_val = float("0.5") + min_val = float("-178268.0") + max_val = float("161180.0") + mean = float("7.65597") + std = float("3433.28") data = None class Program_weight_tensor_parameter_323: name = "parameter_323" + original_name = "batch_norm2d_35.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -2924,6 +3726,7 @@ class Program_weight_tensor_parameter_323: class Program_weight_tensor_parameter_324: name = "parameter_324" + original_name = "batch_norm2d_35.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -2933,6 +3736,7 @@ class Program_weight_tensor_parameter_324: class Program_weight_tensor_parameter_325: name = "parameter_325" + original_name = "batch_norm2d_35.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -2942,6 +3746,7 @@ class Program_weight_tensor_parameter_325: class Program_weight_tensor_parameter_326: name = "parameter_326" + original_name = "batch_norm2d_35.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -2951,195 +3756,259 @@ class Program_weight_tensor_parameter_326: class Program_weight_tensor_parameter_327: name = "parameter_327" + original_name = "conv2d_39.w_0" shape = [64, 256, 1, 1] dtype = "float32" - min_val = float("-0.5") - max_val = float("0.5") + min_val = float("-133557.0") + max_val = float("173236.0") + mean = float("26.4618") + std = float("3282.41") data = None class Program_weight_tensor_parameter_328: name = "parameter_328" + original_name = "batch_norm2d_29.b_0" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-47508.9") + max_val = float("73282.8") + mean = float("1804.49") + std = float("11856.4") data = None class Program_weight_tensor_parameter_329: name = "parameter_329" + original_name = "batch_norm2d_29.w_0" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-305173.0") + max_val = float("73744.0") + mean = float("-2413.64") + std = float("33354.8") data = None class Program_weight_tensor_parameter_330: name = "parameter_330" + original_name = "batch_norm2d_29.w_2" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("34782800000.0") + max_val = float("1.29377e+22") + mean = float("1.87313e+20") + std = float("inf") data = None class Program_weight_tensor_parameter_331: name = "parameter_331" + original_name = "batch_norm2d_29.w_1" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-196325000000.0") + max_val = float("43131900000.0") + mean = float("-956323000.0") + std = float("18669900000.0") data = None class Program_weight_tensor_parameter_332: name = "parameter_332" + original_name = "conv2d_33.w_0" shape = [128, 1280, 1, 1] dtype = "float32" - min_val = float("-0.5") - max_val = float("0.5") + min_val = float("-610360.0") + max_val = float("297877.0") + mean = float("-43.7803") + std = float("4659.28") data = None class Program_weight_tensor_parameter_333: name = "parameter_333" + original_name = "batch_norm2d_28.b_0" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-18920.4") + max_val = float("50131.1") + mean = float("1276.02") + std = float("5869.96") data = None class Program_weight_tensor_parameter_334: name = "parameter_334" + original_name = "batch_norm2d_28.w_0" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-12693.7") + max_val = float("32683.8") + mean = float("33.8802") + std = float("3864.04") data = None class Program_weight_tensor_parameter_335: name = "parameter_335" + original_name = "batch_norm2d_28.w_2" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("12554200000.0") + max_val = float("4.2572e+20") + mean = float("7.03993e+18") + std = float("inf") data = None class Program_weight_tensor_parameter_336: name = "parameter_336" + original_name = "batch_norm2d_28.w_1" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-6273210000.0") + max_val = float("5488470000.0") + mean = float("-65105300.0") + std = float("1089890000.0") data = None class Program_weight_tensor_parameter_337: name = "parameter_337" + original_name = "conv2d_32.w_0" shape = [128, 640, 1, 1] dtype = "float32" - min_val = float("-0.5") - max_val = float("0.5") + min_val = float("-123129.0") + max_val = float("62305.1") + mean = float("-1.43098") + std = float("872.978") data = None class Program_weight_tensor_parameter_338: name = "parameter_338" + original_name = "batch_norm2d_27.b_0" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-58560.4") + max_val = float("43130.9") + mean = float("408.051") + std = float("7528.23") data = None class Program_weight_tensor_parameter_339: name = "parameter_339" + original_name = "batch_norm2d_27.w_0" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-89612.3") + max_val = float("68563.0") + mean = float("-424.56") + std = float("12344.8") data = None class Program_weight_tensor_parameter_340: name = "parameter_340" + original_name = "batch_norm2d_27.w_2" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("20311000000.0") + max_val = float("1.84851e+21") + mean = float("3.34316e+19") + std = float("inf") data = None class Program_weight_tensor_parameter_341: name = "parameter_341" + original_name = "batch_norm2d_27.w_1" shape = [128] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-32024200000.0") + max_val = float("21390200000.0") + mean = float("129123000.0") + std = float("3778280000.0") data = None class Program_weight_tensor_parameter_342: name = "parameter_342" + original_name = "conv2d_31.w_0" shape = [128, 320, 1, 1] dtype = "float32" - min_val = float("-0.5") - max_val = float("0.5") + min_val = float("-190478.0") + max_val = float("154822.0") + mean = float("-12.6005") + std = float("2540.67") data = None class Program_weight_tensor_parameter_343: name = "parameter_343" + original_name = "batch_norm2d_26.b_0" shape = [1280] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-60207.4") + max_val = float("42828.5") + mean = float("-200.78") + std = float("5044.97") data = None class Program_weight_tensor_parameter_344: name = "parameter_344" + original_name = "batch_norm2d_26.w_0" shape = [1280] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-195758.0") + max_val = float("161767.0") + mean = float("-725.617") + std = float("16749.3") data = None class Program_weight_tensor_parameter_345: name = "parameter_345" + original_name = "batch_norm2d_26.w_2" shape = [1280] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("6039170000.0") + max_val = float("6.78926e+20") + mean = float("2.28536e+18") + std = float("inf") data = None class Program_weight_tensor_parameter_346: name = "parameter_346" + original_name = "batch_norm2d_26.w_1" shape = [1280] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-14069300000.0") + max_val = float("7056590000.0") + mean = float("-8576950.0") + std = float("746888000.0") data = None class Program_weight_tensor_parameter_347: name = "parameter_347" + original_name = "conv2d_30.w_0" shape = [1280, 1280, 1, 1] dtype = "float32" - min_val = float("-0.5") - max_val = float("0.5") + min_val = float("-145432.0") + max_val = float("330986.0") + mean = float("-0.255197") + std = float("821.862") data = None class Program_weight_tensor_parameter_348: name = "parameter_348" + original_name = "conv2d_29.b_0" shape = [1280] dtype = "float32" min_val = float("-1.80924") @@ -3151,15 +4020,19 @@ class Program_weight_tensor_parameter_348: class Program_weight_tensor_parameter_349: name = "parameter_349" + original_name = "conv2d_29.w_0" shape = [1280, 320, 1, 1] dtype = "float32" - min_val = float("-0.5") - max_val = float("0.5") + min_val = float("-114.733") + max_val = float("228.459") + mean = float("0.00122015") + std = float("1.07793") data = None class Program_weight_tensor_parameter_350: name = "parameter_350" + original_name = "conv2d_28.b_0" shape = [320] dtype = "float32" min_val = float("-0.820646") @@ -3171,6 +4044,7 @@ class Program_weight_tensor_parameter_350: class Program_weight_tensor_parameter_351: name = "parameter_351" + original_name = "conv2d_28.w_0" shape = [320, 1280, 1, 1] dtype = "float32" min_val = float("-2.08969") @@ -3182,96 +4056,127 @@ class Program_weight_tensor_parameter_351: class Program_weight_tensor_parameter_352: name = "parameter_352" + original_name = "batch_norm2d_25.b_0" shape = [1280] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-92041.1") + max_val = float("59511.4") + mean = float("-289.956") + std = float("5875.43") data = None class Program_weight_tensor_parameter_353: name = "parameter_353" + original_name = "batch_norm2d_25.w_0" shape = [1280] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-113878.0") + max_val = float("141014.0") + mean = float("70.4936") + std = float("8206.11") data = None class Program_weight_tensor_parameter_354: name = "parameter_354" + original_name = "batch_norm2d_25.w_2" shape = [1280] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("3.36739e-06") + max_val = float("6.40361e+20") + mean = float("1.15563e+18") + std = float("inf") data = None class Program_weight_tensor_parameter_355: name = "parameter_355" + original_name = "batch_norm2d_25.w_1" shape = [1280] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-5005020000.0") + max_val = float("11081300000.0") + mean = float("100970.0") + std = float("432630000.0") data = None class Program_weight_tensor_parameter_356: name = "parameter_356" + original_name = "conv2d_27.w_0" shape = [1280, 1, 5, 5] dtype = "float32" - min_val = float("-0.5") - max_val = float("0.5") + min_val = float("-99611.1") + max_val = float("125189.0") + mean = float("-3.56955") + std = float("2867.13") data = None class Program_weight_tensor_parameter_357: name = "parameter_357" + original_name = "batch_norm2d_24.b_0" shape = [1280] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-219047.0") + max_val = float("58324.6") + mean = float("-562.455") + std = float("9867.86") data = None class Program_weight_tensor_parameter_358: name = "parameter_358" + original_name = "batch_norm2d_24.w_0" shape = [1280] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-253196.0") + max_val = float("382351.0") + mean = float("111.635") + std = float("16429.8") data = None class Program_weight_tensor_parameter_359: name = "parameter_359" + original_name = "batch_norm2d_24.w_2" shape = [1280] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("1812810.0") + max_val = float("3.21017e+21") + mean = float("6.6749e+18") + std = float("inf") data = None class Program_weight_tensor_parameter_360: name = "parameter_360" + original_name = "batch_norm2d_24.w_1" shape = [1280] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-16396500000.0") + max_val = float("20826700000.0") + mean = float("36435300.0") + std = float("1141100000.0") data = None class Program_weight_tensor_parameter_361: name = "parameter_361" + original_name = "conv2d_26.w_0" shape = [1280, 640, 1, 1] dtype = "float32" - min_val = float("-0.5") - max_val = float("0.5") + min_val = float("-141872.0") + max_val = float("225461.0") + mean = float("0.881674") + std = float("719.874") data = None class Program_weight_tensor_parameter_362: name = "parameter_362" + original_name = "conv2d_25.b_0" shape = [640] dtype = "float32" min_val = float("-4.8922") @@ -3283,15 +4188,19 @@ class Program_weight_tensor_parameter_362: class Program_weight_tensor_parameter_363: name = "parameter_363" + original_name = "conv2d_25.w_0" shape = [640, 160, 1, 1] dtype = "float32" - min_val = float("-0.5") - max_val = float("0.5") + min_val = float("-144.957") + max_val = float("11.1219") + mean = float("-0.0296543") + std = float("1.14276") data = None class Program_weight_tensor_parameter_364: name = "parameter_364" + original_name = "conv2d_24.b_0" shape = [160] dtype = "float32" min_val = float("-0.202493") @@ -3303,6 +4212,7 @@ class Program_weight_tensor_parameter_364: class Program_weight_tensor_parameter_365: name = "parameter_365" + original_name = "conv2d_24.w_0" shape = [160, 640, 1, 1] dtype = "float32" min_val = float("-1.1422") @@ -3314,996 +4224,1327 @@ class Program_weight_tensor_parameter_365: class Program_weight_tensor_parameter_366: name = "parameter_366" + original_name = "batch_norm2d_23.b_0" shape = [640] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-51267.9") + max_val = float("65772.6") + mean = float("-174.428") + std = float("4167.03") data = None class Program_weight_tensor_parameter_367: name = "parameter_367" + original_name = "batch_norm2d_23.w_0" shape = [640] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-113047.0") + max_val = float("229229.0") + mean = float("-34.7143") + std = float("11243.8") data = None class Program_weight_tensor_parameter_368: name = "parameter_368" + original_name = "batch_norm2d_23.w_2" shape = [640] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("1.61123e-07") + max_val = float("1.73247e+22") + mean = float("2.78801e+19") + std = float("inf") data = None class Program_weight_tensor_parameter_369: name = "parameter_369" + original_name = "batch_norm2d_23.w_1" shape = [640] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-42998100000.0") + max_val = float("39908100000.0") + mean = float("-17340100.0") + std = float("2350230000.0") data = None class Program_weight_tensor_parameter_370: name = "parameter_370" + original_name = "conv2d_23.w_0" shape = [640, 1, 5, 5] dtype = "float32" - min_val = float("-0.5") - max_val = float("0.5") + min_val = float("-79299.7") + max_val = float("316052.0") + mean = float("-84.379") + std = float("3872.66") data = None class Program_weight_tensor_parameter_371: name = "parameter_371" + original_name = "batch_norm2d_22.b_0" shape = [640] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-100022.0") + max_val = float("94709.8") + mean = float("-168.737") + std = float("7862.23") data = None class Program_weight_tensor_parameter_372: name = "parameter_372" + original_name = "batch_norm2d_22.w_0" shape = [640] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-249761.0") + max_val = float("236523.0") + mean = float("-312.544") + std = float("17289.8") data = None class Program_weight_tensor_parameter_373: name = "parameter_373" + original_name = "batch_norm2d_22.w_2" shape = [640] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("1290730000.0") + max_val = float("8.78844e+20") + mean = float("2.72846e+18") + std = float("inf") data = None class Program_weight_tensor_parameter_374: name = "parameter_374" + original_name = "batch_norm2d_22.w_1" shape = [640] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-24967400000.0") + max_val = float("15440700000.0") + mean = float("-87857400.0") + std = float("1512910000.0") data = None class Program_weight_tensor_parameter_375: name = "parameter_375" + original_name = "conv2d_22.w_0" shape = [640, 640, 1, 1] dtype = "float32" - min_val = float("-0.5") - max_val = float("0.5") + min_val = float("-199012.0") + max_val = float("200054.0") + mean = float("-4.14526") + std = float("992.058") data = None class Program_weight_tensor_parameter_376: name = "parameter_376" + original_name = "batch_norm2d_21.b_0" shape = [640] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-82518.6") + max_val = float("79143.0") + mean = float("168.267") + std = float("6067.01") data = None class Program_weight_tensor_parameter_377: name = "parameter_377" + original_name = "batch_norm2d_21.w_0" shape = [640] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-57321.9") + max_val = float("116597.0") + mean = float("387.647") + std = float("9894.24") data = None class Program_weight_tensor_parameter_378: name = "parameter_378" + original_name = "batch_norm2d_21.w_2" shape = [640] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("2.85619e-06") + max_val = float("1.25829e+23") + mean = float("1.97856e+20") + std = float("inf") data = None class Program_weight_tensor_parameter_379: name = "parameter_379" + original_name = "batch_norm2d_21.w_1" shape = [640] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-11284800000.0") + max_val = float("103528000000.0") + mean = float("205898000.0") + std = float("4183220000.0") data = None class Program_weight_tensor_parameter_380: name = "parameter_380" + original_name = "conv2d_21.w_0" shape = [640, 1, 5, 5] dtype = "float32" - min_val = float("-0.5") - max_val = float("0.5") + min_val = float("-50405.0") + max_val = float("61159.5") + mean = float("150.022") + std = float("3343.35") data = None class Program_weight_tensor_parameter_381: name = "parameter_381" + original_name = "batch_norm2d_20.b_0" shape = [640] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-953235.0") + max_val = float("70820.5") + mean = float("-1227.65") + std = float("38205.4") data = None class Program_weight_tensor_parameter_382: name = "parameter_382" + original_name = "batch_norm2d_20.w_0" shape = [640] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-172826.0") + max_val = float("1335040.0") + mean = float("1940.38") + std = float("54381.6") data = None class Program_weight_tensor_parameter_383: name = "parameter_383" + original_name = "batch_norm2d_20.w_2" shape = [640] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("313117000.0") + max_val = float("3.12437e+22") + mean = float("6.42489e+19") + std = float("inf") data = None class Program_weight_tensor_parameter_384: name = "parameter_384" + original_name = "batch_norm2d_20.w_1" shape = [640] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-13702700000.0") + max_val = float("70694600000.0") + mean = float("181567000.0") + std = float("3102160000.0") data = None class Program_weight_tensor_parameter_385: name = "parameter_385" + original_name = "conv2d_20.w_0" shape = [640, 640, 1, 1] dtype = "float32" - min_val = float("-0.5") - max_val = float("0.5") + min_val = float("-470013.0") + max_val = float("995229.0") + mean = float("3.77786") + std = float("2578.84") data = None class Program_weight_tensor_parameter_386: name = "parameter_386" + original_name = "batch_norm2d_19.b_0" shape = [640] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-252028.0") + max_val = float("84774.3") + mean = float("-799.12") + std = float("14776.7") data = None class Program_weight_tensor_parameter_387: name = "parameter_387" + original_name = "batch_norm2d_19.w_0" shape = [640] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-319427.0") + max_val = float("105043.0") + mean = float("-951.734") + std = float("18548.8") data = None class Program_weight_tensor_parameter_388: name = "parameter_388" + original_name = "batch_norm2d_19.w_2" shape = [640] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("7.95316e-07") + max_val = float("2.20108e+21") + mean = float("5.93378e+18") + std = float("inf") data = None class Program_weight_tensor_parameter_389: name = "parameter_389" + original_name = "batch_norm2d_19.w_1" shape = [640] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-14660200000.0") + max_val = float("41376300000.0") + mean = float("3122360.0") + std = float("1908930000.0") data = None class Program_weight_tensor_parameter_390: name = "parameter_390" + original_name = "conv2d_19.w_0" shape = [640, 1, 5, 5] dtype = "float32" - min_val = float("-0.5") - max_val = float("0.5") + min_val = float("-83822.9") + max_val = float("65669.4") + mean = float("14.0029") + std = float("3653.38") data = None class Program_weight_tensor_parameter_391: name = "parameter_391" + original_name = "batch_norm2d_18.b_0" shape = [640] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-116928.0") + max_val = float("169219.0") + mean = float("205.423") + std = float("11805.7") data = None class Program_weight_tensor_parameter_392: name = "parameter_392" + original_name = "batch_norm2d_18.w_0" shape = [640] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-287412.0") + max_val = float("336385.0") + mean = float("1440.21") + std = float("24024.2") data = None class Program_weight_tensor_parameter_393: name = "parameter_393" + original_name = "batch_norm2d_18.w_2" shape = [640] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("1108370000.0") + max_val = float("5.8681e+20") + mean = float("3.0167e+18") + std = float("inf") data = None class Program_weight_tensor_parameter_394: name = "parameter_394" + original_name = "batch_norm2d_18.w_1" shape = [640] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-23696700000.0") + max_val = float("11167400000.0") + mean = float("17698800.0") + std = float("1192040000.0") data = None class Program_weight_tensor_parameter_395: name = "parameter_395" + original_name = "conv2d_18.w_0" shape = [640, 640, 1, 1] dtype = "float32" - min_val = float("-0.5") - max_val = float("0.5") + min_val = float("-189081.0") + max_val = float("159719.0") + mean = float("2.7706") + std = float("1256.91") data = None class Program_weight_tensor_parameter_396: name = "parameter_396" + original_name = "batch_norm2d_17.b_0" shape = [640] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-18760.1") + max_val = float("53626.7") + mean = float("139.094") + std = float("4251.74") data = None class Program_weight_tensor_parameter_397: name = "parameter_397" + original_name = "batch_norm2d_17.w_0" shape = [640] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-36010.7") + max_val = float("152549.0") + mean = float("853.639") + std = float("9022.04") data = None class Program_weight_tensor_parameter_398: name = "parameter_398" + original_name = "batch_norm2d_17.w_2" shape = [640] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("1.64366e-05") + max_val = float("2.62001e+20") + mean = float("1.07981e+18") + std = float("inf") data = None class Program_weight_tensor_parameter_399: name = "parameter_399" + original_name = "batch_norm2d_17.w_1" shape = [640] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-41680700000.0") + max_val = float("3437810000.0") + mean = float("-89032600.0") + std = float("1735380000.0") data = None class Program_weight_tensor_parameter_400: name = "parameter_400" + original_name = "conv2d_17.w_0" shape = [640, 1, 5, 5] dtype = "float32" - min_val = float("-0.5") - max_val = float("0.5") + min_val = float("-55000.7") + max_val = float("63725.1") + mean = float("-44.8327") + std = float("2906.3") data = None class Program_weight_tensor_parameter_401: name = "parameter_401" + original_name = "batch_norm2d_16.b_0" shape = [640] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-107793.0") + max_val = float("173271.0") + mean = float("-155.612") + std = float("10219.1") data = None class Program_weight_tensor_parameter_402: name = "parameter_402" + original_name = "batch_norm2d_16.w_0" shape = [640] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-289542.0") + max_val = float("97698.9") + mean = float("-783.656") + std = float("15184.8") data = None class Program_weight_tensor_parameter_403: name = "parameter_403" + original_name = "batch_norm2d_16.w_2" shape = [640] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("698022000.0") + max_val = float("8.75483e+19") + mean = float("9.29516e+17") + std = float("inf") data = None class Program_weight_tensor_parameter_404: name = "parameter_404" + original_name = "batch_norm2d_16.w_1" shape = [640] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-6894760000.0") + max_val = float("4684640000.0") + mean = float("-33198100.0") + std = float("540811000.0") data = None class Program_weight_tensor_parameter_405: name = "parameter_405" + original_name = "conv2d_16.w_0" shape = [640, 640, 1, 1] dtype = "float32" - min_val = float("-0.5") - max_val = float("0.5") + min_val = float("-110776.0") + max_val = float("115143.0") + mean = float("-3.04249") + std = float("834.264") data = None class Program_weight_tensor_parameter_406: name = "parameter_406" + original_name = "batch_norm2d_15.b_0" shape = [640] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-31439.6") + max_val = float("19646.5") + mean = float("-58.6923") + std = float("3663.18") data = None class Program_weight_tensor_parameter_407: name = "parameter_407" + original_name = "batch_norm2d_15.w_0" shape = [640] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-48381.4") + max_val = float("105450.0") + mean = float("61.0085") + std = float("7173.4") data = None class Program_weight_tensor_parameter_408: name = "parameter_408" + original_name = "batch_norm2d_15.w_2" shape = [640] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("1.1929e-06") + max_val = float("3.93504e+20") + mean = float("1.88509e+18") + std = float("inf") data = None class Program_weight_tensor_parameter_409: name = "parameter_409" + original_name = "batch_norm2d_15.w_1" shape = [640] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-9042340000.0") + max_val = float("16873600000.0") + mean = float("51932200.0") + std = float("1074300000.0") data = None class Program_weight_tensor_parameter_410: name = "parameter_410" + original_name = "conv2d_15.w_0" shape = [640, 1, 5, 5] dtype = "float32" - min_val = float("-0.5") - max_val = float("0.5") + min_val = float("-53214.9") + max_val = float("39237.3") + mean = float("59.6433") + std = float("2571.49") data = None class Program_weight_tensor_parameter_411: name = "parameter_411" + original_name = "batch_norm2d_14.b_0" shape = [640] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-141870.0") + max_val = float("105569.0") + mean = float("-265.588") + std = float("8516.61") data = None class Program_weight_tensor_parameter_412: name = "parameter_412" + original_name = "batch_norm2d_14.w_0" shape = [640] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-256081.0") + max_val = float("219945.0") + mean = float("-298.469") + std = float("17334.8") data = None class Program_weight_tensor_parameter_413: name = "parameter_413" + original_name = "batch_norm2d_14.w_2" shape = [640] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("1423200000.0") + max_val = float("4.42497e+20") + mean = float("1.82884e+18") + std = float("inf") data = None class Program_weight_tensor_parameter_414: name = "parameter_414" + original_name = "batch_norm2d_14.w_1" shape = [640] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-11531700000.0") + max_val = float("10098100000.0") + mean = float("-4629380.0") + std = float("1126350000.0") data = None class Program_weight_tensor_parameter_415: name = "parameter_415" + original_name = "conv2d_14.w_0" shape = [640, 640, 1, 1] dtype = "float32" - min_val = float("-0.5") - max_val = float("0.5") + min_val = float("-166407.0") + max_val = float("134799.0") + mean = float("-0.0640012") + std = float("911.345") data = None class Program_weight_tensor_parameter_416: name = "parameter_416" + original_name = "batch_norm2d_13.b_0" shape = [640] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-98749.1") + max_val = float("49021.0") + mean = float("238.541") + std = float("6233.03") data = None class Program_weight_tensor_parameter_417: name = "parameter_417" + original_name = "batch_norm2d_13.w_0" shape = [640] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-33682.5") + max_val = float("191085.0") + mean = float("1007.18") + std = float("12820.5") data = None class Program_weight_tensor_parameter_418: name = "parameter_418" + original_name = "batch_norm2d_13.w_2" shape = [640] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("0.000284452") + max_val = float("6.8432e+21") + mean = float("1.23178e+19") + std = float("inf") data = None class Program_weight_tensor_parameter_419: name = "parameter_419" + original_name = "batch_norm2d_13.w_1" shape = [640] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-52727100000.0") + max_val = float("32822800000.0") + mean = float("-74949700.0") + std = float("2727810000.0") data = None class Program_weight_tensor_parameter_420: name = "parameter_420" + original_name = "conv2d_13.w_0" shape = [640, 1, 5, 5] dtype = "float32" - min_val = float("-0.5") - max_val = float("0.5") + min_val = float("-66680.5") + max_val = float("84994.8") + mean = float("-73.9363") + std = float("3665.21") data = None class Program_weight_tensor_parameter_421: name = "parameter_421" + original_name = "batch_norm2d_12.b_0" shape = [640] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-77988.1") + max_val = float("562647.0") + mean = float("1436.6") + std = float("23724.0") data = None class Program_weight_tensor_parameter_422: name = "parameter_422" + original_name = "batch_norm2d_12.w_0" shape = [640] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-240995.0") + max_val = float("232443.0") + mean = float("-328.11") + std = float("20135.6") data = None class Program_weight_tensor_parameter_423: name = "parameter_423" + original_name = "batch_norm2d_12.w_2" shape = [640] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("5812160000.0") + max_val = float("8.35144e+20") + mean = float("4.67478e+18") + std = float("inf") data = None class Program_weight_tensor_parameter_424: name = "parameter_424" + original_name = "batch_norm2d_12.w_1" shape = [640] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-21357400000.0") + max_val = float("19106800000.0") + mean = float("-14194600.0") + std = float("1680160000.0") data = None class Program_weight_tensor_parameter_425: name = "parameter_425" + original_name = "conv2d_12.w_0" shape = [640, 320, 1, 1] dtype = "float32" - min_val = float("-0.5") - max_val = float("0.5") + min_val = float("-294149.0") + max_val = float("143696.0") + mean = float("-3.3726") + std = float("2081.93") data = None class Program_weight_tensor_parameter_426: name = "parameter_426" + original_name = "batch_norm2d_11.b_0" shape = [320] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-45313.3") + max_val = float("92783.3") + mean = float("462.242") + std = float("10443.7") data = None class Program_weight_tensor_parameter_427: name = "parameter_427" + original_name = "batch_norm2d_11.w_0" shape = [320] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-169521.0") + max_val = float("157971.0") + mean = float("261.368") + std = float("17818.0") data = None class Program_weight_tensor_parameter_428: name = "parameter_428" + original_name = "batch_norm2d_11.w_2" shape = [320] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("4.15695e-08") + max_val = float("5.23365e+20") + mean = float("3.4446e+18") + std = float("inf") data = None class Program_weight_tensor_parameter_429: name = "parameter_429" + original_name = "batch_norm2d_11.w_1" shape = [320] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-14413500000.0") + max_val = float("10180900000.0") + mean = float("30528800.0") + std = float("1377360000.0") data = None class Program_weight_tensor_parameter_430: name = "parameter_430" + original_name = "conv2d_11.w_0" shape = [320, 1, 3, 3] dtype = "float32" - min_val = float("-0.5") - max_val = float("0.5") + min_val = float("-56482.0") + max_val = float("350087.0") + mean = float("546.868") + std = float("11964.5") data = None class Program_weight_tensor_parameter_431: name = "parameter_431" + original_name = "batch_norm2d_10.b_0" shape = [320] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-734620.0") + max_val = float("137797.0") + mean = float("-4486.18") + std = float("50758.4") data = None class Program_weight_tensor_parameter_432: name = "parameter_432" + original_name = "batch_norm2d_10.w_0" shape = [320] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-989577.0") + max_val = float("202182.0") + mean = float("-7610.08") + std = float("80422.4") data = None class Program_weight_tensor_parameter_433: name = "parameter_433" + original_name = "batch_norm2d_10.w_2" shape = [320] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("1546490000.0") + max_val = float("1.64206e+23") + mean = float("6.03107e+20") + std = float("inf") data = None class Program_weight_tensor_parameter_434: name = "parameter_434" + original_name = "batch_norm2d_10.w_1" shape = [320] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-416065000000.0") + max_val = float("99026300000.0") + mean = float("-1221850000.0") + std = float("24383200000.0") data = None class Program_weight_tensor_parameter_435: name = "parameter_435" + original_name = "conv2d_10.w_0" shape = [320, 320, 1, 1] dtype = "float32" - min_val = float("-0.5") - max_val = float("0.5") + min_val = float("-1694900.0") + max_val = float("438858.0") + mean = float("-53.813") + std = float("7330.09") data = None class Program_weight_tensor_parameter_436: name = "parameter_436" + original_name = "batch_norm2d_9.b_0" shape = [320] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-51494.0") + max_val = float("90543.9") + mean = float("177.611") + std = float("8570.07") data = None class Program_weight_tensor_parameter_437: name = "parameter_437" + original_name = "batch_norm2d_9.w_0" shape = [320] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-157702.0") + max_val = float("482208.0") + mean = float("-130.509") + std = float("33656.5") data = None class Program_weight_tensor_parameter_438: name = "parameter_438" + original_name = "batch_norm2d_9.w_2" shape = [320] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("0.110365") + max_val = float("3.00155e+21") + mean = float("2.22167e+19") + std = float("inf") data = None class Program_weight_tensor_parameter_439: name = "parameter_439" + original_name = "batch_norm2d_9.w_1" shape = [320] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-7671760000.0") + max_val = float("90259200000.0") + mean = float("604382000.0") + std = float("6808880000.0") data = None class Program_weight_tensor_parameter_440: name = "parameter_440" + original_name = "conv2d_9.w_0" shape = [320, 1, 3, 3] dtype = "float32" - min_val = float("-0.5") - max_val = float("0.5") + min_val = float("-97477.5") + max_val = float("109388.0") + mean = float("187.728") + std = float("10174.9") data = None class Program_weight_tensor_parameter_441: name = "parameter_441" + original_name = "batch_norm2d_8.b_0" shape = [320] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-239762.0") + max_val = float("252510.0") + mean = float("943.655") + std = float("28421.7") data = None class Program_weight_tensor_parameter_442: name = "parameter_442" + original_name = "batch_norm2d_8.w_0" shape = [320] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-544097.0") + max_val = float("268352.0") + mean = float("-2726.81") + std = float("51248.6") data = None class Program_weight_tensor_parameter_443: name = "parameter_443" + original_name = "batch_norm2d_8.w_2" shape = [320] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("24902000000.0") + max_val = float("6.12093e+21") + mean = float("4.71114e+19") + std = float("inf") data = None class Program_weight_tensor_parameter_444: name = "parameter_444" + original_name = "batch_norm2d_8.w_1" shape = [320] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-15461400000.0") + max_val = float("57228900000.0") + mean = float("276379000.0") + std = float("4405570000.0") data = None class Program_weight_tensor_parameter_445: name = "parameter_445" + original_name = "conv2d_8.w_0" shape = [320, 160, 1, 1] dtype = "float32" - min_val = float("-0.5") - max_val = float("0.5") + min_val = float("-220564.0") + max_val = float("376610.0") + mean = float("21.0935") + std = float("4018.67") data = None class Program_weight_tensor_parameter_446: name = "parameter_446" + original_name = "batch_norm2d_7.b_0" shape = [160] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-162309.0") + max_val = float("130569.0") + mean = float("-683.445") + std = float("18479.0") data = None class Program_weight_tensor_parameter_447: name = "parameter_447" + original_name = "batch_norm2d_7.w_0" shape = [160] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-179674.0") + max_val = float("77173.1") + mean = float("-4052.22") + std = float("26432.2") data = None class Program_weight_tensor_parameter_448: name = "parameter_448" + original_name = "batch_norm2d_7.w_2" shape = [160] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("0.00149326") + max_val = float("1.16217e+23") + mean = float("7.49654e+20") + std = float("inf") data = None class Program_weight_tensor_parameter_449: name = "parameter_449" + original_name = "batch_norm2d_7.w_1" shape = [160] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-179041000000.0") + max_val = float("5518870000.0") + mean = float("-1501810000.0") + std = float("14392900000.0") data = None class Program_weight_tensor_parameter_450: name = "parameter_450" + original_name = "conv2d_7.w_0" shape = [160, 1, 3, 3] dtype = "float32" - min_val = float("-0.5") - max_val = float("0.5") + min_val = float("-218611.0") + max_val = float("251772.0") + mean = float("-1109.64") + std = float("16572.9") data = None class Program_weight_tensor_parameter_451: name = "parameter_451" + original_name = "batch_norm2d_6.b_0" shape = [160] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-127954.0") + max_val = float("294238.0") + mean = float("4012.73") + std = float("36280.8") data = None class Program_weight_tensor_parameter_452: name = "parameter_452" + original_name = "batch_norm2d_6.w_0" shape = [160] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-1333710.0") + max_val = float("293716.0") + mean = float("-8540.26") + std = float("109498.0") data = None class Program_weight_tensor_parameter_453: name = "parameter_453" + original_name = "batch_norm2d_6.w_2" shape = [160] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("37523100000.0") + max_val = float("7.09867e+21") + mean = float("8.02912e+19") + std = float("inf") data = None class Program_weight_tensor_parameter_454: name = "parameter_454" + original_name = "batch_norm2d_6.w_1" shape = [160] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-70317900000.0") + max_val = float("40189100000.0") + mean = float("-286138000.0") + std = float("7667220000.0") data = None class Program_weight_tensor_parameter_455: name = "parameter_455" + original_name = "conv2d_6.w_0" shape = [160, 160, 1, 1] dtype = "float32" - min_val = float("-0.5") - max_val = float("0.5") + min_val = float("-523848.0") + max_val = float("314533.0") + mean = float("-39.6524") + std = float("7387.07") data = None class Program_weight_tensor_parameter_456: name = "parameter_456" + original_name = "batch_norm2d_5.b_0" shape = [160] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-33114.9") + max_val = float("91148.3") + mean = float("975.587") + std = float("11432.6") data = None class Program_weight_tensor_parameter_457: name = "parameter_457" + original_name = "batch_norm2d_5.w_0" shape = [160] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-196932.0") + max_val = float("171571.0") + mean = float("-4440.13") + std = float("31559.0") data = None class Program_weight_tensor_parameter_458: name = "parameter_458" + original_name = "batch_norm2d_5.w_2" shape = [160] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("5.88099") + max_val = float("4.35754e+23") + mean = float("3.37242e+21") + std = float("inf") data = None class Program_weight_tensor_parameter_459: name = "parameter_459" + original_name = "batch_norm2d_5.w_1" shape = [160] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-1002460000000.0") + max_val = float("138202000000.0") + mean = float("-7099590000.0") + std = float("81248300000.0") data = None class Program_weight_tensor_parameter_460: name = "parameter_460" + original_name = "conv2d_5.w_0" shape = [160, 1, 3, 3] dtype = "float32" - min_val = float("-0.5") - max_val = float("0.5") + min_val = float("-611339.0") + max_val = float("336052.0") + mean = float("-2702.8") + std = float("31899.9") data = None class Program_weight_tensor_parameter_461: name = "parameter_461" + original_name = "batch_norm2d_4.b_0" shape = [160] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-108093.0") + max_val = float("1299840.0") + mean = float("11947.5") + std = float("108050.0") data = None class Program_weight_tensor_parameter_462: name = "parameter_462" + original_name = "batch_norm2d_4.w_0" shape = [160] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-572583.0") + max_val = float("594078.0") + mean = float("-1702.6") + std = float("87472.5") data = None class Program_weight_tensor_parameter_463: name = "parameter_463" + original_name = "batch_norm2d_4.w_2" shape = [160] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("10782800000.0") + max_val = float("8.59511e+21") + mean = float("2.0735e+20") + std = float("inf") data = None class Program_weight_tensor_parameter_464: name = "parameter_464" + original_name = "batch_norm2d_4.w_1" shape = [160] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-55967000000.0") + max_val = float("24376500000.0") + mean = float("-1380990000.0") + std = float("7539840000.0") data = None class Program_weight_tensor_parameter_465: name = "parameter_465" + original_name = "conv2d_4.w_0" shape = [160, 80, 1, 1] dtype = "float32" - min_val = float("-0.5") - max_val = float("0.5") + min_val = float("-369152.0") + max_val = float("205088.0") + mean = float("-163.126") + std = float("8649.37") data = None class Program_weight_tensor_parameter_466: name = "parameter_466" + original_name = "batch_norm2d_3.b_0" shape = [80] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-143820.0") + max_val = float("78462.5") + mean = float("-796.571") + std = float("21777.6") data = None class Program_weight_tensor_parameter_467: name = "parameter_467" + original_name = "batch_norm2d_3.w_0" shape = [80] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-438638.0") + max_val = float("114473.0") + mean = float("-11780.5") + std = float("72033.3") data = None class Program_weight_tensor_parameter_468: name = "parameter_468" + original_name = "batch_norm2d_3.w_2" shape = [80] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("0.00285509") + max_val = float("6.28874e+23") + mean = float("8.72561e+21") + std = float("inf") data = None class Program_weight_tensor_parameter_469: name = "parameter_469" + original_name = "batch_norm2d_3.w_1" shape = [80] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-824200000000.0") + max_val = float("116249000000.0") + mean = float("-10789600000.0") + std = float("93473500000.0") data = None class Program_weight_tensor_parameter_470: name = "parameter_470" + original_name = "conv2d_3.w_0" shape = [80, 1, 3, 3] dtype = "float32" - min_val = float("-0.5") - max_val = float("0.5") + min_val = float("-392158.0") + max_val = float("177629.0") + mean = float("-5138.2") + std = float("47175.8") data = None class Program_weight_tensor_parameter_471: name = "parameter_471" + original_name = "batch_norm2d_2.b_0" shape = [80] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-215737.0") + max_val = float("257631.0") + mean = float("1777.31") + std = float("51261.8") data = None class Program_weight_tensor_parameter_472: name = "parameter_472" + original_name = "batch_norm2d_2.w_0" shape = [80] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-545558.0") + max_val = float("1947410.0") + mean = float("23082.4") + std = float("244212.0") data = None class Program_weight_tensor_parameter_473: name = "parameter_473" + original_name = "batch_norm2d_2.w_2" shape = [80] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("29321900000.0") + max_val = float("7.11039e+23") + mean = float("9.04831e+21") + std = float("inf") data = None class Program_weight_tensor_parameter_474: name = "parameter_474" + original_name = "batch_norm2d_2.w_1" shape = [80] dtype = "float32" - min_val = float("0") - max_val = float("0.5") + min_val = float("-925988000000.0") + max_val = float("53176100000.0") + mean = float("-12275300000.0") + std = float("103743000000.0") data = None class Program_weight_tensor_parameter_475: name = "parameter_475" + original_name = "conv2d_2.w_0" shape = [80, 40, 1, 1] dtype = "float32" - min_val = float("-0.5") - max_val = float("0.5") + min_val = float("-4851060.0") + max_val = float("300462.0") + mean = float("-1559.73") + std = float("87051.1") data = None class Program_weight_tensor_parameter_476: name = "parameter_476" + original_name = "batch_norm2d_1.b_0" shape = [40] dtype = "float32" min_val = float("0") @@ -4313,6 +5554,7 @@ class Program_weight_tensor_parameter_476: class Program_weight_tensor_parameter_477: name = "parameter_477" + original_name = "batch_norm2d_1.w_0" shape = [40] dtype = "float32" min_val = float("0") @@ -4322,6 +5564,7 @@ class Program_weight_tensor_parameter_477: class Program_weight_tensor_parameter_478: name = "parameter_478" + original_name = "batch_norm2d_1.w_2" shape = [40] dtype = "float32" min_val = float("0") @@ -4331,6 +5574,7 @@ class Program_weight_tensor_parameter_478: class Program_weight_tensor_parameter_479: name = "parameter_479" + original_name = "batch_norm2d_1.w_1" shape = [40] dtype = "float32" min_val = float("0") @@ -4340,15 +5584,19 @@ class Program_weight_tensor_parameter_479: class Program_weight_tensor_parameter_480: name = "parameter_480" + original_name = "conv2d_1.w_0" shape = [40, 1, 3, 3] dtype = "float32" - min_val = float("-0.5") - max_val = float("0.5") + min_val = float("-129418.0") + max_val = float("213897.0") + mean = float("-97.4288") + std = float("29963.4") data = None class Program_weight_tensor_parameter_481: name = "parameter_481" + original_name = "batch_norm2d_0.b_0" shape = [40] dtype = "float32" min_val = float("0") @@ -4358,6 +5606,7 @@ class Program_weight_tensor_parameter_481: class Program_weight_tensor_parameter_482: name = "parameter_482" + original_name = "batch_norm2d_0.w_0" shape = [40] dtype = "float32" min_val = float("0") @@ -4367,6 +5616,7 @@ class Program_weight_tensor_parameter_482: class Program_weight_tensor_parameter_483: name = "parameter_483" + original_name = "batch_norm2d_0.w_2" shape = [40] dtype = "float32" min_val = float("0") @@ -4376,6 +5626,7 @@ class Program_weight_tensor_parameter_483: class Program_weight_tensor_parameter_484: name = "parameter_484" + original_name = "batch_norm2d_0.w_1" shape = [40] dtype = "float32" min_val = float("0") @@ -4385,8 +5636,11 @@ class Program_weight_tensor_parameter_484: class Program_weight_tensor_parameter_485: name = "parameter_485" + original_name = "conv2d_0.w_0" shape = [40, 3, 3, 3] dtype = "float32" - min_val = float("-0.5") - max_val = float("0.5") + min_val = float("-267436.0") + max_val = float("254233.0") + mean = float("2987.07") + std = float("32747.5") data = None diff --git a/paddle_samples/PaddleX/PP-ShiTuV2_det/subgraph_5/input_meta.py b/paddle_samples/PaddleX/PP-ShiTuV2_det/subgraph_5/input_meta.py index eb93d7c31..97376148e 100644 --- a/paddle_samples/PaddleX/PP-ShiTuV2_det/subgraph_5/input_meta.py +++ b/paddle_samples/PaddleX/PP-ShiTuV2_det/subgraph_5/input_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_data_0: name = "data_0" + original_name = "var_569" shape = [2, 3, 704, 704] dtype = "float32" min_val = float("-2.82094") diff --git a/paddle_samples/PaddleX/PP-ShiTuV2_det/subgraph_5/weight_meta.py b/paddle_samples/PaddleX/PP-ShiTuV2_det/subgraph_5/weight_meta.py index c013bd537..61862497b 100644 --- a/paddle_samples/PaddleX/PP-ShiTuV2_det/subgraph_5/weight_meta.py +++ b/paddle_samples/PaddleX/PP-ShiTuV2_det/subgraph_5/weight_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_parameter_0: name = "parameter_0" + original_name = "conv2d_103.b_0" shape = [33] dtype = "float32" min_val = float("0") @@ -9,6 +10,7 @@ class Program_weight_tensor_parameter_0: class Program_weight_tensor_parameter_1: name = "parameter_1" + original_name = "conv2d_103.w_0" shape = [33, 128, 1, 1] dtype = "float32" min_val = float("-0.375201") @@ -20,6 +22,7 @@ class Program_weight_tensor_parameter_1: class Program_weight_tensor_parameter_2: name = "parameter_2" + original_name = "batch_norm2d_93.b_0" shape = [128] dtype = "float32" min_val = float("-1.74348") @@ -31,6 +34,7 @@ class Program_weight_tensor_parameter_2: class Program_weight_tensor_parameter_3: name = "parameter_3" + original_name = "batch_norm2d_93.w_0" shape = [128] dtype = "float32" min_val = float("0.514801") @@ -42,6 +46,7 @@ class Program_weight_tensor_parameter_3: class Program_weight_tensor_parameter_4: name = "parameter_4" + original_name = "batch_norm2d_93.w_2" shape = [128] dtype = "float32" min_val = float("0.0297132") @@ -53,6 +58,7 @@ class Program_weight_tensor_parameter_4: class Program_weight_tensor_parameter_5: name = "parameter_5" + original_name = "batch_norm2d_93.w_1" shape = [128] dtype = "float32" min_val = float("-1.33017") @@ -64,6 +70,7 @@ class Program_weight_tensor_parameter_5: class Program_weight_tensor_parameter_6: name = "parameter_6" + original_name = "conv2d_97.w_0" shape = [128, 128, 1, 1] dtype = "float32" min_val = float("-0.422951") @@ -75,6 +82,7 @@ class Program_weight_tensor_parameter_6: class Program_weight_tensor_parameter_7: name = "parameter_7" + original_name = "batch_norm2d_92.b_0" shape = [128] dtype = "float32" min_val = float("-1.2212") @@ -86,6 +94,7 @@ class Program_weight_tensor_parameter_7: class Program_weight_tensor_parameter_8: name = "parameter_8" + original_name = "batch_norm2d_92.w_0" shape = [128] dtype = "float32" min_val = float("0.628216") @@ -97,6 +106,7 @@ class Program_weight_tensor_parameter_8: class Program_weight_tensor_parameter_9: name = "parameter_9" + original_name = "batch_norm2d_92.w_2" shape = [128] dtype = "float32" min_val = float("0.00106378") @@ -108,6 +118,7 @@ class Program_weight_tensor_parameter_9: class Program_weight_tensor_parameter_10: name = "parameter_10" + original_name = "batch_norm2d_92.w_1" shape = [128] dtype = "float32" min_val = float("-0.420077") @@ -119,6 +130,7 @@ class Program_weight_tensor_parameter_10: class Program_weight_tensor_parameter_11: name = "parameter_11" + original_name = "conv2d_96.w_0" shape = [128, 1, 5, 5] dtype = "float32" min_val = float("-0.36121") @@ -130,6 +142,7 @@ class Program_weight_tensor_parameter_11: class Program_weight_tensor_parameter_12: name = "parameter_12" + original_name = "batch_norm2d_91.b_0" shape = [128] dtype = "float32" min_val = float("-3.60292") @@ -141,6 +154,7 @@ class Program_weight_tensor_parameter_12: class Program_weight_tensor_parameter_13: name = "parameter_13" + original_name = "batch_norm2d_91.w_0" shape = [128] dtype = "float32" min_val = float("0.114278") @@ -152,6 +166,7 @@ class Program_weight_tensor_parameter_13: class Program_weight_tensor_parameter_14: name = "parameter_14" + original_name = "batch_norm2d_91.w_2" shape = [128] dtype = "float32" min_val = float("0.0778152") @@ -163,6 +178,7 @@ class Program_weight_tensor_parameter_14: class Program_weight_tensor_parameter_15: name = "parameter_15" + original_name = "batch_norm2d_91.w_1" shape = [128] dtype = "float32" min_val = float("-0.672865") @@ -174,6 +190,7 @@ class Program_weight_tensor_parameter_15: class Program_weight_tensor_parameter_16: name = "parameter_16" + original_name = "conv2d_95.w_0" shape = [128, 128, 1, 1] dtype = "float32" min_val = float("-0.549527") @@ -185,6 +202,7 @@ class Program_weight_tensor_parameter_16: class Program_weight_tensor_parameter_17: name = "parameter_17" + original_name = "batch_norm2d_90.b_0" shape = [128] dtype = "float32" min_val = float("-1.1464") @@ -196,6 +214,7 @@ class Program_weight_tensor_parameter_17: class Program_weight_tensor_parameter_18: name = "parameter_18" + original_name = "batch_norm2d_90.w_0" shape = [128] dtype = "float32" min_val = float("0.614281") @@ -207,6 +226,7 @@ class Program_weight_tensor_parameter_18: class Program_weight_tensor_parameter_19: name = "parameter_19" + original_name = "batch_norm2d_90.w_2" shape = [128] dtype = "float32" min_val = float("0.00433059") @@ -218,6 +238,7 @@ class Program_weight_tensor_parameter_19: class Program_weight_tensor_parameter_20: name = "parameter_20" + original_name = "batch_norm2d_90.w_1" shape = [128] dtype = "float32" min_val = float("-0.676552") @@ -229,6 +250,7 @@ class Program_weight_tensor_parameter_20: class Program_weight_tensor_parameter_21: name = "parameter_21" + original_name = "conv2d_94.w_0" shape = [128, 1, 5, 5] dtype = "float32" min_val = float("-0.310353") @@ -240,6 +262,7 @@ class Program_weight_tensor_parameter_21: class Program_weight_tensor_parameter_22: name = "parameter_22" + original_name = "batch_norm2d_89.b_0" shape = [128] dtype = "float32" min_val = float("-0.860454") @@ -251,6 +274,7 @@ class Program_weight_tensor_parameter_22: class Program_weight_tensor_parameter_23: name = "parameter_23" + original_name = "batch_norm2d_89.w_0" shape = [128] dtype = "float32" min_val = float("0.216525") @@ -262,6 +286,7 @@ class Program_weight_tensor_parameter_23: class Program_weight_tensor_parameter_24: name = "parameter_24" + original_name = "batch_norm2d_89.w_2" shape = [128] dtype = "float32" min_val = float("0.0923019") @@ -273,6 +298,7 @@ class Program_weight_tensor_parameter_24: class Program_weight_tensor_parameter_25: name = "parameter_25" + original_name = "batch_norm2d_89.w_1" shape = [128] dtype = "float32" min_val = float("-0.707531") @@ -284,6 +310,7 @@ class Program_weight_tensor_parameter_25: class Program_weight_tensor_parameter_26: name = "parameter_26" + original_name = "conv2d_93.w_0" shape = [128, 128, 1, 1] dtype = "float32" min_val = float("-0.441662") @@ -295,6 +322,7 @@ class Program_weight_tensor_parameter_26: class Program_weight_tensor_parameter_27: name = "parameter_27" + original_name = "batch_norm2d_88.b_0" shape = [128] dtype = "float32" min_val = float("-0.878364") @@ -306,6 +334,7 @@ class Program_weight_tensor_parameter_27: class Program_weight_tensor_parameter_28: name = "parameter_28" + original_name = "batch_norm2d_88.w_0" shape = [128] dtype = "float32" min_val = float("0.561267") @@ -317,6 +346,7 @@ class Program_weight_tensor_parameter_28: class Program_weight_tensor_parameter_29: name = "parameter_29" + original_name = "batch_norm2d_88.w_2" shape = [128] dtype = "float32" min_val = float("0.00378212") @@ -328,6 +358,7 @@ class Program_weight_tensor_parameter_29: class Program_weight_tensor_parameter_30: name = "parameter_30" + original_name = "batch_norm2d_88.w_1" shape = [128] dtype = "float32" min_val = float("-0.571453") @@ -339,6 +370,7 @@ class Program_weight_tensor_parameter_30: class Program_weight_tensor_parameter_31: name = "parameter_31" + original_name = "conv2d_92.w_0" shape = [128, 1, 5, 5] dtype = "float32" min_val = float("-0.326377") @@ -350,6 +382,7 @@ class Program_weight_tensor_parameter_31: class Program_weight_tensor_parameter_32: name = "parameter_32" + original_name = "batch_norm2d_87.b_0" shape = [128] dtype = "float32" min_val = float("-0.771394") @@ -361,6 +394,7 @@ class Program_weight_tensor_parameter_32: class Program_weight_tensor_parameter_33: name = "parameter_33" + original_name = "batch_norm2d_87.w_0" shape = [128] dtype = "float32" min_val = float("0.303182") @@ -372,6 +406,7 @@ class Program_weight_tensor_parameter_33: class Program_weight_tensor_parameter_34: name = "parameter_34" + original_name = "batch_norm2d_87.w_2" shape = [128] dtype = "float32" min_val = float("0.131878") @@ -383,6 +418,7 @@ class Program_weight_tensor_parameter_34: class Program_weight_tensor_parameter_35: name = "parameter_35" + original_name = "batch_norm2d_87.w_1" shape = [128] dtype = "float32" min_val = float("-0.380765") @@ -394,6 +430,7 @@ class Program_weight_tensor_parameter_35: class Program_weight_tensor_parameter_36: name = "parameter_36" + original_name = "conv2d_91.w_0" shape = [128, 128, 1, 1] dtype = "float32" min_val = float("-0.327775") @@ -405,6 +442,7 @@ class Program_weight_tensor_parameter_36: class Program_weight_tensor_parameter_37: name = "parameter_37" + original_name = "batch_norm2d_86.b_0" shape = [128] dtype = "float32" min_val = float("-1.2202") @@ -416,6 +454,7 @@ class Program_weight_tensor_parameter_37: class Program_weight_tensor_parameter_38: name = "parameter_38" + original_name = "batch_norm2d_86.w_0" shape = [128] dtype = "float32" min_val = float("0.771122") @@ -427,6 +466,7 @@ class Program_weight_tensor_parameter_38: class Program_weight_tensor_parameter_39: name = "parameter_39" + original_name = "batch_norm2d_86.w_2" shape = [128] dtype = "float32" min_val = float("0.000450363") @@ -438,6 +478,7 @@ class Program_weight_tensor_parameter_39: class Program_weight_tensor_parameter_40: name = "parameter_40" + original_name = "batch_norm2d_86.w_1" shape = [128] dtype = "float32" min_val = float("-0.536661") @@ -449,6 +490,7 @@ class Program_weight_tensor_parameter_40: class Program_weight_tensor_parameter_41: name = "parameter_41" + original_name = "conv2d_90.w_0" shape = [128, 1, 5, 5] dtype = "float32" min_val = float("-0.27148") @@ -460,6 +502,7 @@ class Program_weight_tensor_parameter_41: class Program_weight_tensor_parameter_42: name = "parameter_42" + original_name = "conv2d_102.b_0" shape = [33] dtype = "float32" min_val = float("0") @@ -469,6 +512,7 @@ class Program_weight_tensor_parameter_42: class Program_weight_tensor_parameter_43: name = "parameter_43" + original_name = "conv2d_102.w_0" shape = [33, 128, 1, 1] dtype = "float32" min_val = float("-0.45582") @@ -480,6 +524,7 @@ class Program_weight_tensor_parameter_43: class Program_weight_tensor_parameter_44: name = "parameter_44" + original_name = "batch_norm2d_85.b_0" shape = [128] dtype = "float32" min_val = float("-1.73839") @@ -491,6 +536,7 @@ class Program_weight_tensor_parameter_44: class Program_weight_tensor_parameter_45: name = "parameter_45" + original_name = "batch_norm2d_85.w_0" shape = [128] dtype = "float32" min_val = float("0.46275") @@ -502,6 +548,7 @@ class Program_weight_tensor_parameter_45: class Program_weight_tensor_parameter_46: name = "parameter_46" + original_name = "batch_norm2d_85.w_2" shape = [128] dtype = "float32" min_val = float("0.0385639") @@ -513,6 +560,7 @@ class Program_weight_tensor_parameter_46: class Program_weight_tensor_parameter_47: name = "parameter_47" + original_name = "batch_norm2d_85.w_1" shape = [128] dtype = "float32" min_val = float("-1.58375") @@ -524,6 +572,7 @@ class Program_weight_tensor_parameter_47: class Program_weight_tensor_parameter_48: name = "parameter_48" + original_name = "conv2d_89.w_0" shape = [128, 128, 1, 1] dtype = "float32" min_val = float("-0.357583") @@ -535,6 +584,7 @@ class Program_weight_tensor_parameter_48: class Program_weight_tensor_parameter_49: name = "parameter_49" + original_name = "batch_norm2d_84.b_0" shape = [128] dtype = "float32" min_val = float("-0.762939") @@ -546,6 +596,7 @@ class Program_weight_tensor_parameter_49: class Program_weight_tensor_parameter_50: name = "parameter_50" + original_name = "batch_norm2d_84.w_0" shape = [128] dtype = "float32" min_val = float("0.482731") @@ -557,6 +608,7 @@ class Program_weight_tensor_parameter_50: class Program_weight_tensor_parameter_51: name = "parameter_51" + original_name = "batch_norm2d_84.w_2" shape = [128] dtype = "float32" min_val = float("0.000741799") @@ -568,6 +620,7 @@ class Program_weight_tensor_parameter_51: class Program_weight_tensor_parameter_52: name = "parameter_52" + original_name = "batch_norm2d_84.w_1" shape = [128] dtype = "float32" min_val = float("-0.334637") @@ -579,6 +632,7 @@ class Program_weight_tensor_parameter_52: class Program_weight_tensor_parameter_53: name = "parameter_53" + original_name = "conv2d_88.w_0" shape = [128, 1, 5, 5] dtype = "float32" min_val = float("-0.448479") @@ -590,6 +644,7 @@ class Program_weight_tensor_parameter_53: class Program_weight_tensor_parameter_54: name = "parameter_54" + original_name = "batch_norm2d_83.b_0" shape = [128] dtype = "float32" min_val = float("-3.59086") @@ -601,6 +656,7 @@ class Program_weight_tensor_parameter_54: class Program_weight_tensor_parameter_55: name = "parameter_55" + original_name = "batch_norm2d_83.w_0" shape = [128] dtype = "float32" min_val = float("0.165909") @@ -612,6 +668,7 @@ class Program_weight_tensor_parameter_55: class Program_weight_tensor_parameter_56: name = "parameter_56" + original_name = "batch_norm2d_83.w_2" shape = [128] dtype = "float32" min_val = float("0.0620864") @@ -623,6 +680,7 @@ class Program_weight_tensor_parameter_56: class Program_weight_tensor_parameter_57: name = "parameter_57" + original_name = "batch_norm2d_83.w_1" shape = [128] dtype = "float32" min_val = float("-1.25083") @@ -634,6 +692,7 @@ class Program_weight_tensor_parameter_57: class Program_weight_tensor_parameter_58: name = "parameter_58" + original_name = "conv2d_87.w_0" shape = [128, 128, 1, 1] dtype = "float32" min_val = float("-0.412967") @@ -645,6 +704,7 @@ class Program_weight_tensor_parameter_58: class Program_weight_tensor_parameter_59: name = "parameter_59" + original_name = "batch_norm2d_82.b_0" shape = [128] dtype = "float32" min_val = float("-0.637418") @@ -656,6 +716,7 @@ class Program_weight_tensor_parameter_59: class Program_weight_tensor_parameter_60: name = "parameter_60" + original_name = "batch_norm2d_82.w_0" shape = [128] dtype = "float32" min_val = float("0.461447") @@ -667,6 +728,7 @@ class Program_weight_tensor_parameter_60: class Program_weight_tensor_parameter_61: name = "parameter_61" + original_name = "batch_norm2d_82.w_2" shape = [128] dtype = "float32" min_val = float("0.00329002") @@ -678,6 +740,7 @@ class Program_weight_tensor_parameter_61: class Program_weight_tensor_parameter_62: name = "parameter_62" + original_name = "batch_norm2d_82.w_1" shape = [128] dtype = "float32" min_val = float("-0.571658") @@ -689,6 +752,7 @@ class Program_weight_tensor_parameter_62: class Program_weight_tensor_parameter_63: name = "parameter_63" + original_name = "conv2d_86.w_0" shape = [128, 1, 5, 5] dtype = "float32" min_val = float("-0.266141") @@ -700,6 +764,7 @@ class Program_weight_tensor_parameter_63: class Program_weight_tensor_parameter_64: name = "parameter_64" + original_name = "batch_norm2d_81.b_0" shape = [128] dtype = "float32" min_val = float("-0.627466") @@ -711,6 +776,7 @@ class Program_weight_tensor_parameter_64: class Program_weight_tensor_parameter_65: name = "parameter_65" + original_name = "batch_norm2d_81.w_0" shape = [128] dtype = "float32" min_val = float("0.241978") @@ -722,6 +788,7 @@ class Program_weight_tensor_parameter_65: class Program_weight_tensor_parameter_66: name = "parameter_66" + original_name = "batch_norm2d_81.w_2" shape = [128] dtype = "float32" min_val = float("0.0947985") @@ -733,6 +800,7 @@ class Program_weight_tensor_parameter_66: class Program_weight_tensor_parameter_67: name = "parameter_67" + original_name = "batch_norm2d_81.w_1" shape = [128] dtype = "float32" min_val = float("-0.658904") @@ -744,6 +812,7 @@ class Program_weight_tensor_parameter_67: class Program_weight_tensor_parameter_68: name = "parameter_68" + original_name = "conv2d_85.w_0" shape = [128, 128, 1, 1] dtype = "float32" min_val = float("-0.476021") @@ -755,6 +824,7 @@ class Program_weight_tensor_parameter_68: class Program_weight_tensor_parameter_69: name = "parameter_69" + original_name = "batch_norm2d_80.b_0" shape = [128] dtype = "float32" min_val = float("-0.522486") @@ -766,6 +836,7 @@ class Program_weight_tensor_parameter_69: class Program_weight_tensor_parameter_70: name = "parameter_70" + original_name = "batch_norm2d_80.w_0" shape = [128] dtype = "float32" min_val = float("0.603769") @@ -777,6 +848,7 @@ class Program_weight_tensor_parameter_70: class Program_weight_tensor_parameter_71: name = "parameter_71" + original_name = "batch_norm2d_80.w_2" shape = [128] dtype = "float32" min_val = float("0.0032515") @@ -788,6 +860,7 @@ class Program_weight_tensor_parameter_71: class Program_weight_tensor_parameter_72: name = "parameter_72" + original_name = "batch_norm2d_80.w_1" shape = [128] dtype = "float32" min_val = float("-0.653981") @@ -799,6 +872,7 @@ class Program_weight_tensor_parameter_72: class Program_weight_tensor_parameter_73: name = "parameter_73" + original_name = "conv2d_84.w_0" shape = [128, 1, 5, 5] dtype = "float32" min_val = float("-0.247819") @@ -810,6 +884,7 @@ class Program_weight_tensor_parameter_73: class Program_weight_tensor_parameter_74: name = "parameter_74" + original_name = "batch_norm2d_79.b_0" shape = [128] dtype = "float32" min_val = float("-0.84365") @@ -821,6 +896,7 @@ class Program_weight_tensor_parameter_74: class Program_weight_tensor_parameter_75: name = "parameter_75" + original_name = "batch_norm2d_79.w_0" shape = [128] dtype = "float32" min_val = float("0.243127") @@ -832,6 +908,7 @@ class Program_weight_tensor_parameter_75: class Program_weight_tensor_parameter_76: name = "parameter_76" + original_name = "batch_norm2d_79.w_2" shape = [128] dtype = "float32" min_val = float("0.0767344") @@ -843,6 +920,7 @@ class Program_weight_tensor_parameter_76: class Program_weight_tensor_parameter_77: name = "parameter_77" + original_name = "batch_norm2d_79.w_1" shape = [128] dtype = "float32" min_val = float("-0.494532") @@ -854,6 +932,7 @@ class Program_weight_tensor_parameter_77: class Program_weight_tensor_parameter_78: name = "parameter_78" + original_name = "conv2d_83.w_0" shape = [128, 128, 1, 1] dtype = "float32" min_val = float("-0.308412") @@ -865,6 +944,7 @@ class Program_weight_tensor_parameter_78: class Program_weight_tensor_parameter_79: name = "parameter_79" + original_name = "batch_norm2d_78.b_0" shape = [128] dtype = "float32" min_val = float("-0.744444") @@ -876,6 +956,7 @@ class Program_weight_tensor_parameter_79: class Program_weight_tensor_parameter_80: name = "parameter_80" + original_name = "batch_norm2d_78.w_0" shape = [128] dtype = "float32" min_val = float("0.665894") @@ -887,6 +968,7 @@ class Program_weight_tensor_parameter_80: class Program_weight_tensor_parameter_81: name = "parameter_81" + original_name = "batch_norm2d_78.w_2" shape = [128] dtype = "float32" min_val = float("0.000103655") @@ -898,6 +980,7 @@ class Program_weight_tensor_parameter_81: class Program_weight_tensor_parameter_82: name = "parameter_82" + original_name = "batch_norm2d_78.w_1" shape = [128] dtype = "float32" min_val = float("-0.0458635") @@ -909,6 +992,7 @@ class Program_weight_tensor_parameter_82: class Program_weight_tensor_parameter_83: name = "parameter_83" + original_name = "conv2d_82.w_0" shape = [128, 1, 5, 5] dtype = "float32" min_val = float("-0.400488") @@ -920,6 +1004,7 @@ class Program_weight_tensor_parameter_83: class Program_weight_tensor_parameter_84: name = "parameter_84" + original_name = "conv2d_101.b_0" shape = [33] dtype = "float32" min_val = float("0") @@ -929,6 +1014,7 @@ class Program_weight_tensor_parameter_84: class Program_weight_tensor_parameter_85: name = "parameter_85" + original_name = "conv2d_101.w_0" shape = [33, 128, 1, 1] dtype = "float32" min_val = float("-0.371053") @@ -940,6 +1026,7 @@ class Program_weight_tensor_parameter_85: class Program_weight_tensor_parameter_86: name = "parameter_86" + original_name = "batch_norm2d_77.b_0" shape = [128] dtype = "float32" min_val = float("-1.52737") @@ -951,6 +1038,7 @@ class Program_weight_tensor_parameter_86: class Program_weight_tensor_parameter_87: name = "parameter_87" + original_name = "batch_norm2d_77.w_0" shape = [128] dtype = "float32" min_val = float("0.389002") @@ -962,6 +1050,7 @@ class Program_weight_tensor_parameter_87: class Program_weight_tensor_parameter_88: name = "parameter_88" + original_name = "batch_norm2d_77.w_2" shape = [128] dtype = "float32" min_val = float("0.040885") @@ -973,6 +1062,7 @@ class Program_weight_tensor_parameter_88: class Program_weight_tensor_parameter_89: name = "parameter_89" + original_name = "batch_norm2d_77.w_1" shape = [128] dtype = "float32" min_val = float("-1.65759") @@ -984,6 +1074,7 @@ class Program_weight_tensor_parameter_89: class Program_weight_tensor_parameter_90: name = "parameter_90" + original_name = "conv2d_81.w_0" shape = [128, 128, 1, 1] dtype = "float32" min_val = float("-0.481541") @@ -995,6 +1086,7 @@ class Program_weight_tensor_parameter_90: class Program_weight_tensor_parameter_91: name = "parameter_91" + original_name = "batch_norm2d_76.b_0" shape = [128] dtype = "float32" min_val = float("-1.21849") @@ -1006,6 +1098,7 @@ class Program_weight_tensor_parameter_91: class Program_weight_tensor_parameter_92: name = "parameter_92" + original_name = "batch_norm2d_76.w_0" shape = [128] dtype = "float32" min_val = float("0.35011") @@ -1017,6 +1110,7 @@ class Program_weight_tensor_parameter_92: class Program_weight_tensor_parameter_93: name = "parameter_93" + original_name = "batch_norm2d_76.w_2" shape = [128] dtype = "float32" min_val = float("0.000111139") @@ -1028,6 +1122,7 @@ class Program_weight_tensor_parameter_93: class Program_weight_tensor_parameter_94: name = "parameter_94" + original_name = "batch_norm2d_76.w_1" shape = [128] dtype = "float32" min_val = float("-0.662113") @@ -1039,6 +1134,7 @@ class Program_weight_tensor_parameter_94: class Program_weight_tensor_parameter_95: name = "parameter_95" + original_name = "conv2d_80.w_0" shape = [128, 1, 5, 5] dtype = "float32" min_val = float("-0.617492") @@ -1050,6 +1146,7 @@ class Program_weight_tensor_parameter_95: class Program_weight_tensor_parameter_96: name = "parameter_96" + original_name = "batch_norm2d_75.b_0" shape = [128] dtype = "float32" min_val = float("-2.80035") @@ -1061,6 +1158,7 @@ class Program_weight_tensor_parameter_96: class Program_weight_tensor_parameter_97: name = "parameter_97" + original_name = "batch_norm2d_75.w_0" shape = [128] dtype = "float32" min_val = float("0.0419769") @@ -1072,6 +1170,7 @@ class Program_weight_tensor_parameter_97: class Program_weight_tensor_parameter_98: name = "parameter_98" + original_name = "batch_norm2d_75.w_2" shape = [128] dtype = "float32" min_val = float("0.0720405") @@ -1083,6 +1182,7 @@ class Program_weight_tensor_parameter_98: class Program_weight_tensor_parameter_99: name = "parameter_99" + original_name = "batch_norm2d_75.w_1" shape = [128] dtype = "float32" min_val = float("-1.46577") @@ -1094,6 +1194,7 @@ class Program_weight_tensor_parameter_99: class Program_weight_tensor_parameter_100: name = "parameter_100" + original_name = "conv2d_79.w_0" shape = [128, 128, 1, 1] dtype = "float32" min_val = float("-0.505075") @@ -1105,6 +1206,7 @@ class Program_weight_tensor_parameter_100: class Program_weight_tensor_parameter_101: name = "parameter_101" + original_name = "batch_norm2d_74.b_0" shape = [128] dtype = "float32" min_val = float("-0.325808") @@ -1116,6 +1218,7 @@ class Program_weight_tensor_parameter_101: class Program_weight_tensor_parameter_102: name = "parameter_102" + original_name = "batch_norm2d_74.w_0" shape = [128] dtype = "float32" min_val = float("0.439181") @@ -1127,6 +1230,7 @@ class Program_weight_tensor_parameter_102: class Program_weight_tensor_parameter_103: name = "parameter_103" + original_name = "batch_norm2d_74.w_2" shape = [128] dtype = "float32" min_val = float("0.000209082") @@ -1138,6 +1242,7 @@ class Program_weight_tensor_parameter_103: class Program_weight_tensor_parameter_104: name = "parameter_104" + original_name = "batch_norm2d_74.w_1" shape = [128] dtype = "float32" min_val = float("-0.361578") @@ -1149,6 +1254,7 @@ class Program_weight_tensor_parameter_104: class Program_weight_tensor_parameter_105: name = "parameter_105" + original_name = "conv2d_78.w_0" shape = [128, 1, 5, 5] dtype = "float32" min_val = float("-0.480383") @@ -1160,6 +1266,7 @@ class Program_weight_tensor_parameter_105: class Program_weight_tensor_parameter_106: name = "parameter_106" + original_name = "batch_norm2d_73.b_0" shape = [128] dtype = "float32" min_val = float("-0.863449") @@ -1171,6 +1278,7 @@ class Program_weight_tensor_parameter_106: class Program_weight_tensor_parameter_107: name = "parameter_107" + original_name = "batch_norm2d_73.w_0" shape = [128] dtype = "float32" min_val = float("0.0665504") @@ -1182,6 +1290,7 @@ class Program_weight_tensor_parameter_107: class Program_weight_tensor_parameter_108: name = "parameter_108" + original_name = "batch_norm2d_73.w_2" shape = [128] dtype = "float32" min_val = float("0.0719913") @@ -1193,6 +1302,7 @@ class Program_weight_tensor_parameter_108: class Program_weight_tensor_parameter_109: name = "parameter_109" + original_name = "batch_norm2d_73.w_1" shape = [128] dtype = "float32" min_val = float("-0.666007") @@ -1204,6 +1314,7 @@ class Program_weight_tensor_parameter_109: class Program_weight_tensor_parameter_110: name = "parameter_110" + original_name = "conv2d_77.w_0" shape = [128, 128, 1, 1] dtype = "float32" min_val = float("-0.337938") @@ -1215,6 +1326,7 @@ class Program_weight_tensor_parameter_110: class Program_weight_tensor_parameter_111: name = "parameter_111" + original_name = "batch_norm2d_72.b_0" shape = [128] dtype = "float32" min_val = float("-0.439472") @@ -1226,6 +1338,7 @@ class Program_weight_tensor_parameter_111: class Program_weight_tensor_parameter_112: name = "parameter_112" + original_name = "batch_norm2d_72.w_0" shape = [128] dtype = "float32" min_val = float("0.557845") @@ -1237,6 +1350,7 @@ class Program_weight_tensor_parameter_112: class Program_weight_tensor_parameter_113: name = "parameter_113" + original_name = "batch_norm2d_72.w_2" shape = [128] dtype = "float32" min_val = float("0.00188746") @@ -1248,6 +1362,7 @@ class Program_weight_tensor_parameter_113: class Program_weight_tensor_parameter_114: name = "parameter_114" + original_name = "batch_norm2d_72.w_1" shape = [128] dtype = "float32" min_val = float("-0.570214") @@ -1259,6 +1374,7 @@ class Program_weight_tensor_parameter_114: class Program_weight_tensor_parameter_115: name = "parameter_115" + original_name = "conv2d_76.w_0" shape = [128, 1, 5, 5] dtype = "float32" min_val = float("-0.268484") @@ -1270,6 +1386,7 @@ class Program_weight_tensor_parameter_115: class Program_weight_tensor_parameter_116: name = "parameter_116" + original_name = "batch_norm2d_71.b_0" shape = [128] dtype = "float32" min_val = float("-0.666527") @@ -1281,6 +1398,7 @@ class Program_weight_tensor_parameter_116: class Program_weight_tensor_parameter_117: name = "parameter_117" + original_name = "batch_norm2d_71.w_0" shape = [128] dtype = "float32" min_val = float("0.127829") @@ -1292,6 +1410,7 @@ class Program_weight_tensor_parameter_117: class Program_weight_tensor_parameter_118: name = "parameter_118" + original_name = "batch_norm2d_71.w_2" shape = [128] dtype = "float32" min_val = float("0.0736821") @@ -1303,6 +1422,7 @@ class Program_weight_tensor_parameter_118: class Program_weight_tensor_parameter_119: name = "parameter_119" + original_name = "batch_norm2d_71.w_1" shape = [128] dtype = "float32" min_val = float("-0.616939") @@ -1314,6 +1434,7 @@ class Program_weight_tensor_parameter_119: class Program_weight_tensor_parameter_120: name = "parameter_120" + original_name = "conv2d_75.w_0" shape = [128, 128, 1, 1] dtype = "float32" min_val = float("-0.336161") @@ -1325,6 +1446,7 @@ class Program_weight_tensor_parameter_120: class Program_weight_tensor_parameter_121: name = "parameter_121" + original_name = "batch_norm2d_70.b_0" shape = [128] dtype = "float32" min_val = float("-0.483507") @@ -1336,6 +1458,7 @@ class Program_weight_tensor_parameter_121: class Program_weight_tensor_parameter_122: name = "parameter_122" + original_name = "batch_norm2d_70.w_0" shape = [128] dtype = "float32" min_val = float("0.669199") @@ -1347,6 +1470,7 @@ class Program_weight_tensor_parameter_122: class Program_weight_tensor_parameter_123: name = "parameter_123" + original_name = "batch_norm2d_70.w_2" shape = [128] dtype = "float32" min_val = float("6.60628e-05") @@ -1358,6 +1482,7 @@ class Program_weight_tensor_parameter_123: class Program_weight_tensor_parameter_124: name = "parameter_124" + original_name = "batch_norm2d_70.w_1" shape = [128] dtype = "float32" min_val = float("-0.0733793") @@ -1369,6 +1494,7 @@ class Program_weight_tensor_parameter_124: class Program_weight_tensor_parameter_125: name = "parameter_125" + original_name = "conv2d_74.w_0" shape = [128, 1, 5, 5] dtype = "float32" min_val = float("-0.361941") @@ -1380,6 +1506,7 @@ class Program_weight_tensor_parameter_125: class Program_weight_tensor_parameter_126: name = "parameter_126" + original_name = "conv2d_100.b_0" shape = [33] dtype = "float32" min_val = float("0") @@ -1389,6 +1516,7 @@ class Program_weight_tensor_parameter_126: class Program_weight_tensor_parameter_127: name = "parameter_127" + original_name = "conv2d_100.w_0" shape = [33, 128, 1, 1] dtype = "float32" min_val = float("-0.462323") @@ -1400,6 +1528,7 @@ class Program_weight_tensor_parameter_127: class Program_weight_tensor_parameter_128: name = "parameter_128" + original_name = "batch_norm2d_69.b_0" shape = [128] dtype = "float32" min_val = float("-1.83425") @@ -1411,6 +1540,7 @@ class Program_weight_tensor_parameter_128: class Program_weight_tensor_parameter_129: name = "parameter_129" + original_name = "batch_norm2d_69.w_0" shape = [128] dtype = "float32" min_val = float("0.302939") @@ -1422,6 +1552,7 @@ class Program_weight_tensor_parameter_129: class Program_weight_tensor_parameter_130: name = "parameter_130" + original_name = "batch_norm2d_69.w_2" shape = [128] dtype = "float32" min_val = float("0.0336375") @@ -1433,6 +1564,7 @@ class Program_weight_tensor_parameter_130: class Program_weight_tensor_parameter_131: name = "parameter_131" + original_name = "batch_norm2d_69.w_1" shape = [128] dtype = "float32" min_val = float("-1.59633") @@ -1444,6 +1576,7 @@ class Program_weight_tensor_parameter_131: class Program_weight_tensor_parameter_132: name = "parameter_132" + original_name = "conv2d_73.w_0" shape = [128, 128, 1, 1] dtype = "float32" min_val = float("-0.424084") @@ -1455,6 +1588,7 @@ class Program_weight_tensor_parameter_132: class Program_weight_tensor_parameter_133: name = "parameter_133" + original_name = "batch_norm2d_68.b_0" shape = [128] dtype = "float32" min_val = float("-0.885282") @@ -1466,6 +1600,7 @@ class Program_weight_tensor_parameter_133: class Program_weight_tensor_parameter_134: name = "parameter_134" + original_name = "batch_norm2d_68.w_0" shape = [128] dtype = "float32" min_val = float("0.281883") @@ -1477,6 +1612,7 @@ class Program_weight_tensor_parameter_134: class Program_weight_tensor_parameter_135: name = "parameter_135" + original_name = "batch_norm2d_68.w_2" shape = [128] dtype = "float32" min_val = float("0.000127685") @@ -1488,6 +1624,7 @@ class Program_weight_tensor_parameter_135: class Program_weight_tensor_parameter_136: name = "parameter_136" + original_name = "batch_norm2d_68.w_1" shape = [128] dtype = "float32" min_val = float("-0.850447") @@ -1499,6 +1636,7 @@ class Program_weight_tensor_parameter_136: class Program_weight_tensor_parameter_137: name = "parameter_137" + original_name = "conv2d_72.w_0" shape = [128, 1, 5, 5] dtype = "float32" min_val = float("-0.787919") @@ -1510,6 +1648,7 @@ class Program_weight_tensor_parameter_137: class Program_weight_tensor_parameter_138: name = "parameter_138" + original_name = "batch_norm2d_67.b_0" shape = [128] dtype = "float32" min_val = float("-2.40717") @@ -1521,6 +1660,7 @@ class Program_weight_tensor_parameter_138: class Program_weight_tensor_parameter_139: name = "parameter_139" + original_name = "batch_norm2d_67.w_0" shape = [128] dtype = "float32" min_val = float("0.0374566") @@ -1532,6 +1672,7 @@ class Program_weight_tensor_parameter_139: class Program_weight_tensor_parameter_140: name = "parameter_140" + original_name = "batch_norm2d_67.w_2" shape = [128] dtype = "float32" min_val = float("0.0723185") @@ -1543,6 +1684,7 @@ class Program_weight_tensor_parameter_140: class Program_weight_tensor_parameter_141: name = "parameter_141" + original_name = "batch_norm2d_67.w_1" shape = [128] dtype = "float32" min_val = float("-1.04937") @@ -1554,6 +1696,7 @@ class Program_weight_tensor_parameter_141: class Program_weight_tensor_parameter_142: name = "parameter_142" + original_name = "conv2d_71.w_0" shape = [128, 128, 1, 1] dtype = "float32" min_val = float("-0.480066") @@ -1565,6 +1708,7 @@ class Program_weight_tensor_parameter_142: class Program_weight_tensor_parameter_143: name = "parameter_143" + original_name = "batch_norm2d_66.b_0" shape = [128] dtype = "float32" min_val = float("-0.499643") @@ -1576,6 +1720,7 @@ class Program_weight_tensor_parameter_143: class Program_weight_tensor_parameter_144: name = "parameter_144" + original_name = "batch_norm2d_66.w_0" shape = [128] dtype = "float32" min_val = float("0.443765") @@ -1587,6 +1732,7 @@ class Program_weight_tensor_parameter_144: class Program_weight_tensor_parameter_145: name = "parameter_145" + original_name = "batch_norm2d_66.w_2" shape = [128] dtype = "float32" min_val = float("0.000515519") @@ -1598,6 +1744,7 @@ class Program_weight_tensor_parameter_145: class Program_weight_tensor_parameter_146: name = "parameter_146" + original_name = "batch_norm2d_66.w_1" shape = [128] dtype = "float32" min_val = float("-0.33182") @@ -1609,6 +1756,7 @@ class Program_weight_tensor_parameter_146: class Program_weight_tensor_parameter_147: name = "parameter_147" + original_name = "conv2d_70.w_0" shape = [128, 1, 5, 5] dtype = "float32" min_val = float("-0.454176") @@ -1620,6 +1768,7 @@ class Program_weight_tensor_parameter_147: class Program_weight_tensor_parameter_148: name = "parameter_148" + original_name = "batch_norm2d_65.b_0" shape = [128] dtype = "float32" min_val = float("-0.61559") @@ -1631,6 +1780,7 @@ class Program_weight_tensor_parameter_148: class Program_weight_tensor_parameter_149: name = "parameter_149" + original_name = "batch_norm2d_65.w_0" shape = [128] dtype = "float32" min_val = float("0.106835") @@ -1642,6 +1792,7 @@ class Program_weight_tensor_parameter_149: class Program_weight_tensor_parameter_150: name = "parameter_150" + original_name = "batch_norm2d_65.w_2" shape = [128] dtype = "float32" min_val = float("0.0506852") @@ -1653,6 +1804,7 @@ class Program_weight_tensor_parameter_150: class Program_weight_tensor_parameter_151: name = "parameter_151" + original_name = "batch_norm2d_65.w_1" shape = [128] dtype = "float32" min_val = float("-1.04463") @@ -1664,6 +1816,7 @@ class Program_weight_tensor_parameter_151: class Program_weight_tensor_parameter_152: name = "parameter_152" + original_name = "conv2d_69.w_0" shape = [128, 128, 1, 1] dtype = "float32" min_val = float("-0.314917") @@ -1675,6 +1828,7 @@ class Program_weight_tensor_parameter_152: class Program_weight_tensor_parameter_153: name = "parameter_153" + original_name = "batch_norm2d_64.b_0" shape = [128] dtype = "float32" min_val = float("-0.522771") @@ -1686,6 +1840,7 @@ class Program_weight_tensor_parameter_153: class Program_weight_tensor_parameter_154: name = "parameter_154" + original_name = "batch_norm2d_64.w_0" shape = [128] dtype = "float32" min_val = float("0.470558") @@ -1697,6 +1852,7 @@ class Program_weight_tensor_parameter_154: class Program_weight_tensor_parameter_155: name = "parameter_155" + original_name = "batch_norm2d_64.w_2" shape = [128] dtype = "float32" min_val = float("0.00057566") @@ -1708,6 +1864,7 @@ class Program_weight_tensor_parameter_155: class Program_weight_tensor_parameter_156: name = "parameter_156" + original_name = "batch_norm2d_64.w_1" shape = [128] dtype = "float32" min_val = float("-0.522868") @@ -1719,6 +1876,7 @@ class Program_weight_tensor_parameter_156: class Program_weight_tensor_parameter_157: name = "parameter_157" + original_name = "conv2d_68.w_0" shape = [128, 1, 5, 5] dtype = "float32" min_val = float("-0.262746") @@ -1730,6 +1888,7 @@ class Program_weight_tensor_parameter_157: class Program_weight_tensor_parameter_158: name = "parameter_158" + original_name = "batch_norm2d_63.b_0" shape = [128] dtype = "float32" min_val = float("-0.722184") @@ -1741,6 +1900,7 @@ class Program_weight_tensor_parameter_158: class Program_weight_tensor_parameter_159: name = "parameter_159" + original_name = "batch_norm2d_63.w_0" shape = [128] dtype = "float32" min_val = float("0.126979") @@ -1752,6 +1912,7 @@ class Program_weight_tensor_parameter_159: class Program_weight_tensor_parameter_160: name = "parameter_160" + original_name = "batch_norm2d_63.w_2" shape = [128] dtype = "float32" min_val = float("0.054903") @@ -1763,6 +1924,7 @@ class Program_weight_tensor_parameter_160: class Program_weight_tensor_parameter_161: name = "parameter_161" + original_name = "batch_norm2d_63.w_1" shape = [128] dtype = "float32" min_val = float("-0.726827") @@ -1774,6 +1936,7 @@ class Program_weight_tensor_parameter_161: class Program_weight_tensor_parameter_162: name = "parameter_162" + original_name = "conv2d_67.w_0" shape = [128, 128, 1, 1] dtype = "float32" min_val = float("-0.335131") @@ -1785,6 +1948,7 @@ class Program_weight_tensor_parameter_162: class Program_weight_tensor_parameter_163: name = "parameter_163" + original_name = "batch_norm2d_62.b_0" shape = [128] dtype = "float32" min_val = float("-0.857479") @@ -1796,6 +1960,7 @@ class Program_weight_tensor_parameter_163: class Program_weight_tensor_parameter_164: name = "parameter_164" + original_name = "batch_norm2d_62.w_0" shape = [128] dtype = "float32" min_val = float("0.577456") @@ -1807,6 +1972,7 @@ class Program_weight_tensor_parameter_164: class Program_weight_tensor_parameter_165: name = "parameter_165" + original_name = "batch_norm2d_62.w_2" shape = [128] dtype = "float32" min_val = float("4.4668e-05") @@ -1818,6 +1984,7 @@ class Program_weight_tensor_parameter_165: class Program_weight_tensor_parameter_166: name = "parameter_166" + original_name = "batch_norm2d_62.w_1" shape = [128] dtype = "float32" min_val = float("-0.0474587") @@ -1829,6 +1996,7 @@ class Program_weight_tensor_parameter_166: class Program_weight_tensor_parameter_167: name = "parameter_167" + original_name = "conv2d_66.w_0" shape = [128, 1, 5, 5] dtype = "float32" min_val = float("-0.343819") @@ -1840,6 +2008,7 @@ class Program_weight_tensor_parameter_167: class Program_weight_tensor_parameter_168: name = "parameter_168" + original_name = "batch_norm2d_33.b_0" shape = [128] dtype = "float32" min_val = float("-0.109651") @@ -1851,6 +2020,7 @@ class Program_weight_tensor_parameter_168: class Program_weight_tensor_parameter_169: name = "parameter_169" + original_name = "batch_norm2d_33.w_0" shape = [128] dtype = "float32" min_val = float("0.0151343") @@ -1862,6 +2032,7 @@ class Program_weight_tensor_parameter_169: class Program_weight_tensor_parameter_170: name = "parameter_170" + original_name = "batch_norm2d_33.w_2" shape = [128] dtype = "float32" min_val = float("0.00050776") @@ -1873,6 +2044,7 @@ class Program_weight_tensor_parameter_170: class Program_weight_tensor_parameter_171: name = "parameter_171" + original_name = "batch_norm2d_33.w_1" shape = [128] dtype = "float32" min_val = float("-0.0205566") @@ -1884,6 +2056,7 @@ class Program_weight_tensor_parameter_171: class Program_weight_tensor_parameter_172: name = "parameter_172" + original_name = "conv2d_37.w_0" shape = [128, 128, 1, 1] dtype = "float32" min_val = float("-0.307281") @@ -1895,6 +2068,7 @@ class Program_weight_tensor_parameter_172: class Program_weight_tensor_parameter_173: name = "parameter_173" + original_name = "batch_norm2d_32.b_0" shape = [128] dtype = "float32" min_val = float("-0.345297") @@ -1906,6 +2080,7 @@ class Program_weight_tensor_parameter_173: class Program_weight_tensor_parameter_174: name = "parameter_174" + original_name = "batch_norm2d_32.w_0" shape = [128] dtype = "float32" min_val = float("0.0473144") @@ -1917,6 +2092,7 @@ class Program_weight_tensor_parameter_174: class Program_weight_tensor_parameter_175: name = "parameter_175" + original_name = "batch_norm2d_32.w_2" shape = [128] dtype = "float32" min_val = float("7.35355e-06") @@ -1928,6 +2104,7 @@ class Program_weight_tensor_parameter_175: class Program_weight_tensor_parameter_176: name = "parameter_176" + original_name = "batch_norm2d_32.w_1" shape = [128] dtype = "float32" min_val = float("-0.0299703") @@ -1939,6 +2116,7 @@ class Program_weight_tensor_parameter_176: class Program_weight_tensor_parameter_177: name = "parameter_177" + original_name = "conv2d_36.w_0" shape = [128, 1, 5, 5] dtype = "float32" min_val = float("-0.264332") @@ -1950,6 +2128,7 @@ class Program_weight_tensor_parameter_177: class Program_weight_tensor_parameter_178: name = "parameter_178" + original_name = "batch_norm2d_31.b_0" shape = [128] dtype = "float32" min_val = float("-0.103489") @@ -1961,6 +2140,7 @@ class Program_weight_tensor_parameter_178: class Program_weight_tensor_parameter_179: name = "parameter_179" + original_name = "batch_norm2d_31.w_0" shape = [128] dtype = "float32" min_val = float("0.0173989") @@ -1972,6 +2152,7 @@ class Program_weight_tensor_parameter_179: class Program_weight_tensor_parameter_180: name = "parameter_180" + original_name = "batch_norm2d_31.w_2" shape = [128] dtype = "float32" min_val = float("0.000119903") @@ -1983,6 +2164,7 @@ class Program_weight_tensor_parameter_180: class Program_weight_tensor_parameter_181: name = "parameter_181" + original_name = "batch_norm2d_31.w_1" shape = [128] dtype = "float32" min_val = float("-0.00732816") @@ -1994,6 +2176,7 @@ class Program_weight_tensor_parameter_181: class Program_weight_tensor_parameter_182: name = "parameter_182" + original_name = "conv2d_35.w_0" shape = [128, 128, 1, 1] dtype = "float32" min_val = float("-0.345218") @@ -2005,6 +2188,7 @@ class Program_weight_tensor_parameter_182: class Program_weight_tensor_parameter_183: name = "parameter_183" + original_name = "batch_norm2d_30.b_0" shape = [128] dtype = "float32" min_val = float("-0.0587116") @@ -2016,6 +2200,7 @@ class Program_weight_tensor_parameter_183: class Program_weight_tensor_parameter_184: name = "parameter_184" + original_name = "batch_norm2d_30.w_0" shape = [128] dtype = "float32" min_val = float("0.00654403") @@ -2027,6 +2212,7 @@ class Program_weight_tensor_parameter_184: class Program_weight_tensor_parameter_185: name = "parameter_185" + original_name = "batch_norm2d_30.w_2" shape = [128] dtype = "float32" min_val = float("1.21138e-05") @@ -2038,6 +2224,7 @@ class Program_weight_tensor_parameter_185: class Program_weight_tensor_parameter_186: name = "parameter_186" + original_name = "batch_norm2d_30.w_1" shape = [128] dtype = "float32" min_val = float("-0.0534565") @@ -2049,6 +2236,7 @@ class Program_weight_tensor_parameter_186: class Program_weight_tensor_parameter_187: name = "parameter_187" + original_name = "conv2d_34.w_0" shape = [128, 1, 5, 5] dtype = "float32" min_val = float("-0.14609") @@ -2060,6 +2248,7 @@ class Program_weight_tensor_parameter_187: class Program_weight_tensor_parameter_188: name = "parameter_188" + original_name = "batch_norm2d_58.b_0" shape = [128] dtype = "float32" min_val = float("-0.262536") @@ -2071,6 +2260,7 @@ class Program_weight_tensor_parameter_188: class Program_weight_tensor_parameter_189: name = "parameter_189" + original_name = "batch_norm2d_58.w_0" shape = [128] dtype = "float32" min_val = float("0.0517739") @@ -2082,6 +2272,7 @@ class Program_weight_tensor_parameter_189: class Program_weight_tensor_parameter_190: name = "parameter_190" + original_name = "batch_norm2d_58.w_2" shape = [128] dtype = "float32" min_val = float("0.00558142") @@ -2093,6 +2284,7 @@ class Program_weight_tensor_parameter_190: class Program_weight_tensor_parameter_191: name = "parameter_191" + original_name = "batch_norm2d_58.w_1" shape = [128] dtype = "float32" min_val = float("-0.0543098") @@ -2104,6 +2296,7 @@ class Program_weight_tensor_parameter_191: class Program_weight_tensor_parameter_192: name = "parameter_192" + original_name = "conv2d_62.w_0" shape = [128, 128, 1, 1] dtype = "float32" min_val = float("-0.371713") @@ -2115,6 +2308,7 @@ class Program_weight_tensor_parameter_192: class Program_weight_tensor_parameter_193: name = "parameter_193" + original_name = "batch_norm2d_61.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -2124,6 +2318,7 @@ class Program_weight_tensor_parameter_193: class Program_weight_tensor_parameter_194: name = "parameter_194" + original_name = "batch_norm2d_61.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -2133,6 +2328,7 @@ class Program_weight_tensor_parameter_194: class Program_weight_tensor_parameter_195: name = "parameter_195" + original_name = "batch_norm2d_61.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -2142,6 +2338,7 @@ class Program_weight_tensor_parameter_195: class Program_weight_tensor_parameter_196: name = "parameter_196" + original_name = "batch_norm2d_61.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -2151,6 +2348,7 @@ class Program_weight_tensor_parameter_196: class Program_weight_tensor_parameter_197: name = "parameter_197" + original_name = "conv2d_65.w_0" shape = [64, 64, 1, 1] dtype = "float32" min_val = float("-0.408624") @@ -2162,6 +2360,7 @@ class Program_weight_tensor_parameter_197: class Program_weight_tensor_parameter_198: name = "parameter_198" + original_name = "batch_norm2d_60.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -2171,6 +2370,7 @@ class Program_weight_tensor_parameter_198: class Program_weight_tensor_parameter_199: name = "parameter_199" + original_name = "batch_norm2d_60.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -2180,6 +2380,7 @@ class Program_weight_tensor_parameter_199: class Program_weight_tensor_parameter_200: name = "parameter_200" + original_name = "batch_norm2d_60.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -2189,6 +2390,7 @@ class Program_weight_tensor_parameter_200: class Program_weight_tensor_parameter_201: name = "parameter_201" + original_name = "batch_norm2d_60.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -2198,6 +2400,7 @@ class Program_weight_tensor_parameter_201: class Program_weight_tensor_parameter_202: name = "parameter_202" + original_name = "conv2d_64.w_0" shape = [64, 1, 5, 5] dtype = "float32" min_val = float("-0.398122") @@ -2209,6 +2412,7 @@ class Program_weight_tensor_parameter_202: class Program_weight_tensor_parameter_203: name = "parameter_203" + original_name = "batch_norm2d_59.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -2218,6 +2422,7 @@ class Program_weight_tensor_parameter_203: class Program_weight_tensor_parameter_204: name = "parameter_204" + original_name = "batch_norm2d_59.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -2227,6 +2432,7 @@ class Program_weight_tensor_parameter_204: class Program_weight_tensor_parameter_205: name = "parameter_205" + original_name = "batch_norm2d_59.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -2236,6 +2442,7 @@ class Program_weight_tensor_parameter_205: class Program_weight_tensor_parameter_206: name = "parameter_206" + original_name = "batch_norm2d_59.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -2245,6 +2452,7 @@ class Program_weight_tensor_parameter_206: class Program_weight_tensor_parameter_207: name = "parameter_207" + original_name = "conv2d_63.w_0" shape = [64, 64, 1, 1] dtype = "float32" min_val = float("-0.263455") @@ -2256,6 +2464,7 @@ class Program_weight_tensor_parameter_207: class Program_weight_tensor_parameter_208: name = "parameter_208" + original_name = "batch_norm2d_56.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -2265,6 +2474,7 @@ class Program_weight_tensor_parameter_208: class Program_weight_tensor_parameter_209: name = "parameter_209" + original_name = "batch_norm2d_56.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -2274,6 +2484,7 @@ class Program_weight_tensor_parameter_209: class Program_weight_tensor_parameter_210: name = "parameter_210" + original_name = "batch_norm2d_56.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -2283,6 +2494,7 @@ class Program_weight_tensor_parameter_210: class Program_weight_tensor_parameter_211: name = "parameter_211" + original_name = "batch_norm2d_56.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -2292,6 +2504,7 @@ class Program_weight_tensor_parameter_211: class Program_weight_tensor_parameter_212: name = "parameter_212" + original_name = "conv2d_60.w_0" shape = [64, 256, 1, 1] dtype = "float32" min_val = float("-0.330818") @@ -2303,6 +2516,7 @@ class Program_weight_tensor_parameter_212: class Program_weight_tensor_parameter_213: name = "parameter_213" + original_name = "batch_norm2d_57.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -2312,6 +2526,7 @@ class Program_weight_tensor_parameter_213: class Program_weight_tensor_parameter_214: name = "parameter_214" + original_name = "batch_norm2d_57.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -2321,6 +2536,7 @@ class Program_weight_tensor_parameter_214: class Program_weight_tensor_parameter_215: name = "parameter_215" + original_name = "batch_norm2d_57.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -2330,6 +2546,7 @@ class Program_weight_tensor_parameter_215: class Program_weight_tensor_parameter_216: name = "parameter_216" + original_name = "batch_norm2d_57.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -2339,6 +2556,7 @@ class Program_weight_tensor_parameter_216: class Program_weight_tensor_parameter_217: name = "parameter_217" + original_name = "conv2d_61.w_0" shape = [64, 256, 1, 1] dtype = "float32" min_val = float("-0.405559") @@ -2350,6 +2568,7 @@ class Program_weight_tensor_parameter_217: class Program_weight_tensor_parameter_218: name = "parameter_218" + original_name = "batch_norm2d_55.b_0" shape = [128] dtype = "float32" min_val = float("-0.321786") @@ -2361,6 +2580,7 @@ class Program_weight_tensor_parameter_218: class Program_weight_tensor_parameter_219: name = "parameter_219" + original_name = "batch_norm2d_55.w_0" shape = [128] dtype = "float32" min_val = float("0.0220778") @@ -2372,6 +2592,7 @@ class Program_weight_tensor_parameter_219: class Program_weight_tensor_parameter_220: name = "parameter_220" + original_name = "batch_norm2d_55.w_2" shape = [128] dtype = "float32" min_val = float("1.3571e-05") @@ -2383,6 +2604,7 @@ class Program_weight_tensor_parameter_220: class Program_weight_tensor_parameter_221: name = "parameter_221" + original_name = "batch_norm2d_55.w_1" shape = [128] dtype = "float32" min_val = float("-0.0433239") @@ -2394,6 +2616,7 @@ class Program_weight_tensor_parameter_221: class Program_weight_tensor_parameter_222: name = "parameter_222" + original_name = "conv2d_59.w_0" shape = [128, 128, 1, 1] dtype = "float32" min_val = float("-0.319065") @@ -2405,6 +2628,7 @@ class Program_weight_tensor_parameter_222: class Program_weight_tensor_parameter_223: name = "parameter_223" + original_name = "batch_norm2d_54.b_0" shape = [128] dtype = "float32" min_val = float("-0.208316") @@ -2416,6 +2640,7 @@ class Program_weight_tensor_parameter_223: class Program_weight_tensor_parameter_224: name = "parameter_224" + original_name = "batch_norm2d_54.w_0" shape = [128] dtype = "float32" min_val = float("0.000556338") @@ -2427,6 +2652,7 @@ class Program_weight_tensor_parameter_224: class Program_weight_tensor_parameter_225: name = "parameter_225" + original_name = "batch_norm2d_54.w_2" shape = [128] dtype = "float32" min_val = float("2.28275e-09") @@ -2438,6 +2664,7 @@ class Program_weight_tensor_parameter_225: class Program_weight_tensor_parameter_226: name = "parameter_226" + original_name = "batch_norm2d_54.w_1" shape = [128] dtype = "float32" min_val = float("-0.0288547") @@ -2449,6 +2676,7 @@ class Program_weight_tensor_parameter_226: class Program_weight_tensor_parameter_227: name = "parameter_227" + original_name = "conv2d_58.w_0" shape = [128, 1, 5, 5] dtype = "float32" min_val = float("-0.233903") @@ -2460,6 +2688,7 @@ class Program_weight_tensor_parameter_227: class Program_weight_tensor_parameter_228: name = "parameter_228" + original_name = "batch_norm2d_50.b_0" shape = [128] dtype = "float32" min_val = float("-0.18228") @@ -2471,6 +2700,7 @@ class Program_weight_tensor_parameter_228: class Program_weight_tensor_parameter_229: name = "parameter_229" + original_name = "batch_norm2d_50.w_0" shape = [128] dtype = "float32" min_val = float("0.0486596") @@ -2482,6 +2712,7 @@ class Program_weight_tensor_parameter_229: class Program_weight_tensor_parameter_230: name = "parameter_230" + original_name = "batch_norm2d_50.w_2" shape = [128] dtype = "float32" min_val = float("0.00718747") @@ -2493,6 +2724,7 @@ class Program_weight_tensor_parameter_230: class Program_weight_tensor_parameter_231: name = "parameter_231" + original_name = "batch_norm2d_50.w_1" shape = [128] dtype = "float32" min_val = float("-0.0611763") @@ -2504,6 +2736,7 @@ class Program_weight_tensor_parameter_231: class Program_weight_tensor_parameter_232: name = "parameter_232" + original_name = "conv2d_54.w_0" shape = [128, 128, 1, 1] dtype = "float32" min_val = float("-0.316076") @@ -2515,6 +2748,7 @@ class Program_weight_tensor_parameter_232: class Program_weight_tensor_parameter_233: name = "parameter_233" + original_name = "batch_norm2d_53.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -2524,6 +2758,7 @@ class Program_weight_tensor_parameter_233: class Program_weight_tensor_parameter_234: name = "parameter_234" + original_name = "batch_norm2d_53.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -2533,6 +2768,7 @@ class Program_weight_tensor_parameter_234: class Program_weight_tensor_parameter_235: name = "parameter_235" + original_name = "batch_norm2d_53.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -2542,6 +2778,7 @@ class Program_weight_tensor_parameter_235: class Program_weight_tensor_parameter_236: name = "parameter_236" + original_name = "batch_norm2d_53.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -2551,6 +2788,7 @@ class Program_weight_tensor_parameter_236: class Program_weight_tensor_parameter_237: name = "parameter_237" + original_name = "conv2d_57.w_0" shape = [64, 64, 1, 1] dtype = "float32" min_val = float("-0.3675") @@ -2562,6 +2800,7 @@ class Program_weight_tensor_parameter_237: class Program_weight_tensor_parameter_238: name = "parameter_238" + original_name = "batch_norm2d_52.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -2571,6 +2810,7 @@ class Program_weight_tensor_parameter_238: class Program_weight_tensor_parameter_239: name = "parameter_239" + original_name = "batch_norm2d_52.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -2580,6 +2820,7 @@ class Program_weight_tensor_parameter_239: class Program_weight_tensor_parameter_240: name = "parameter_240" + original_name = "batch_norm2d_52.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -2589,6 +2830,7 @@ class Program_weight_tensor_parameter_240: class Program_weight_tensor_parameter_241: name = "parameter_241" + original_name = "batch_norm2d_52.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -2598,6 +2840,7 @@ class Program_weight_tensor_parameter_241: class Program_weight_tensor_parameter_242: name = "parameter_242" + original_name = "conv2d_56.w_0" shape = [64, 1, 5, 5] dtype = "float32" min_val = float("-0.436263") @@ -2609,6 +2852,7 @@ class Program_weight_tensor_parameter_242: class Program_weight_tensor_parameter_243: name = "parameter_243" + original_name = "batch_norm2d_51.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -2618,6 +2862,7 @@ class Program_weight_tensor_parameter_243: class Program_weight_tensor_parameter_244: name = "parameter_244" + original_name = "batch_norm2d_51.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -2627,6 +2872,7 @@ class Program_weight_tensor_parameter_244: class Program_weight_tensor_parameter_245: name = "parameter_245" + original_name = "batch_norm2d_51.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -2636,6 +2882,7 @@ class Program_weight_tensor_parameter_245: class Program_weight_tensor_parameter_246: name = "parameter_246" + original_name = "batch_norm2d_51.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -2645,6 +2892,7 @@ class Program_weight_tensor_parameter_246: class Program_weight_tensor_parameter_247: name = "parameter_247" + original_name = "conv2d_55.w_0" shape = [64, 64, 1, 1] dtype = "float32" min_val = float("-0.304634") @@ -2656,6 +2904,7 @@ class Program_weight_tensor_parameter_247: class Program_weight_tensor_parameter_248: name = "parameter_248" + original_name = "batch_norm2d_48.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -2665,6 +2914,7 @@ class Program_weight_tensor_parameter_248: class Program_weight_tensor_parameter_249: name = "parameter_249" + original_name = "batch_norm2d_48.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -2674,6 +2924,7 @@ class Program_weight_tensor_parameter_249: class Program_weight_tensor_parameter_250: name = "parameter_250" + original_name = "batch_norm2d_48.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -2683,6 +2934,7 @@ class Program_weight_tensor_parameter_250: class Program_weight_tensor_parameter_251: name = "parameter_251" + original_name = "batch_norm2d_48.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -2692,6 +2944,7 @@ class Program_weight_tensor_parameter_251: class Program_weight_tensor_parameter_252: name = "parameter_252" + original_name = "conv2d_52.w_0" shape = [64, 256, 1, 1] dtype = "float32" min_val = float("-0.455253") @@ -2703,6 +2956,7 @@ class Program_weight_tensor_parameter_252: class Program_weight_tensor_parameter_253: name = "parameter_253" + original_name = "batch_norm2d_49.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -2712,6 +2966,7 @@ class Program_weight_tensor_parameter_253: class Program_weight_tensor_parameter_254: name = "parameter_254" + original_name = "batch_norm2d_49.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -2721,6 +2976,7 @@ class Program_weight_tensor_parameter_254: class Program_weight_tensor_parameter_255: name = "parameter_255" + original_name = "batch_norm2d_49.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -2730,6 +2986,7 @@ class Program_weight_tensor_parameter_255: class Program_weight_tensor_parameter_256: name = "parameter_256" + original_name = "batch_norm2d_49.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -2739,6 +2996,7 @@ class Program_weight_tensor_parameter_256: class Program_weight_tensor_parameter_257: name = "parameter_257" + original_name = "conv2d_53.w_0" shape = [64, 256, 1, 1] dtype = "float32" min_val = float("-0.294551") @@ -2750,6 +3008,7 @@ class Program_weight_tensor_parameter_257: class Program_weight_tensor_parameter_258: name = "parameter_258" + original_name = "batch_norm2d_47.b_0" shape = [128] dtype = "float32" min_val = float("-0.510433") @@ -2761,6 +3020,7 @@ class Program_weight_tensor_parameter_258: class Program_weight_tensor_parameter_259: name = "parameter_259" + original_name = "batch_norm2d_47.w_0" shape = [128] dtype = "float32" min_val = float("0.0366756") @@ -2772,6 +3032,7 @@ class Program_weight_tensor_parameter_259: class Program_weight_tensor_parameter_260: name = "parameter_260" + original_name = "batch_norm2d_47.w_2" shape = [128] dtype = "float32" min_val = float("9.01179e-05") @@ -2783,6 +3044,7 @@ class Program_weight_tensor_parameter_260: class Program_weight_tensor_parameter_261: name = "parameter_261" + original_name = "batch_norm2d_47.w_1" shape = [128] dtype = "float32" min_val = float("-0.0472358") @@ -2794,6 +3056,7 @@ class Program_weight_tensor_parameter_261: class Program_weight_tensor_parameter_262: name = "parameter_262" + original_name = "conv2d_51.w_0" shape = [128, 128, 1, 1] dtype = "float32" min_val = float("-0.346032") @@ -2805,6 +3068,7 @@ class Program_weight_tensor_parameter_262: class Program_weight_tensor_parameter_263: name = "parameter_263" + original_name = "batch_norm2d_46.b_0" shape = [128] dtype = "float32" min_val = float("-0.555553") @@ -2816,6 +3080,7 @@ class Program_weight_tensor_parameter_263: class Program_weight_tensor_parameter_264: name = "parameter_264" + original_name = "batch_norm2d_46.w_0" shape = [128] dtype = "float32" min_val = float("0.0493619") @@ -2827,6 +3092,7 @@ class Program_weight_tensor_parameter_264: class Program_weight_tensor_parameter_265: name = "parameter_265" + original_name = "batch_norm2d_46.w_2" shape = [128] dtype = "float32" min_val = float("3.33585e-05") @@ -2838,6 +3104,7 @@ class Program_weight_tensor_parameter_265: class Program_weight_tensor_parameter_266: name = "parameter_266" + original_name = "batch_norm2d_46.w_1" shape = [128] dtype = "float32" min_val = float("-0.0437983") @@ -2849,6 +3116,7 @@ class Program_weight_tensor_parameter_266: class Program_weight_tensor_parameter_267: name = "parameter_267" + original_name = "conv2d_50.w_0" shape = [128, 1, 5, 5] dtype = "float32" min_val = float("-0.260839") @@ -2860,6 +3128,7 @@ class Program_weight_tensor_parameter_267: class Program_weight_tensor_parameter_268: name = "parameter_268" + original_name = "batch_norm2d_42.b_0" shape = [128] dtype = "float32" min_val = float("-0.25283") @@ -2871,6 +3140,7 @@ class Program_weight_tensor_parameter_268: class Program_weight_tensor_parameter_269: name = "parameter_269" + original_name = "batch_norm2d_42.w_0" shape = [128] dtype = "float32" min_val = float("0.0435377") @@ -2882,6 +3152,7 @@ class Program_weight_tensor_parameter_269: class Program_weight_tensor_parameter_270: name = "parameter_270" + original_name = "batch_norm2d_42.w_2" shape = [128] dtype = "float32" min_val = float("0.0142984") @@ -2893,6 +3164,7 @@ class Program_weight_tensor_parameter_270: class Program_weight_tensor_parameter_271: name = "parameter_271" + original_name = "batch_norm2d_42.w_1" shape = [128] dtype = "float32" min_val = float("-0.0913079") @@ -2904,6 +3176,7 @@ class Program_weight_tensor_parameter_271: class Program_weight_tensor_parameter_272: name = "parameter_272" + original_name = "conv2d_46.w_0" shape = [128, 128, 1, 1] dtype = "float32" min_val = float("-0.724668") @@ -2915,6 +3188,7 @@ class Program_weight_tensor_parameter_272: class Program_weight_tensor_parameter_273: name = "parameter_273" + original_name = "batch_norm2d_45.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -2924,6 +3198,7 @@ class Program_weight_tensor_parameter_273: class Program_weight_tensor_parameter_274: name = "parameter_274" + original_name = "batch_norm2d_45.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -2933,6 +3208,7 @@ class Program_weight_tensor_parameter_274: class Program_weight_tensor_parameter_275: name = "parameter_275" + original_name = "batch_norm2d_45.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -2942,6 +3218,7 @@ class Program_weight_tensor_parameter_275: class Program_weight_tensor_parameter_276: name = "parameter_276" + original_name = "batch_norm2d_45.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -2951,6 +3228,7 @@ class Program_weight_tensor_parameter_276: class Program_weight_tensor_parameter_277: name = "parameter_277" + original_name = "conv2d_49.w_0" shape = [64, 64, 1, 1] dtype = "float32" min_val = float("-0.348777") @@ -2962,6 +3240,7 @@ class Program_weight_tensor_parameter_277: class Program_weight_tensor_parameter_278: name = "parameter_278" + original_name = "batch_norm2d_44.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -2971,6 +3250,7 @@ class Program_weight_tensor_parameter_278: class Program_weight_tensor_parameter_279: name = "parameter_279" + original_name = "batch_norm2d_44.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -2980,6 +3260,7 @@ class Program_weight_tensor_parameter_279: class Program_weight_tensor_parameter_280: name = "parameter_280" + original_name = "batch_norm2d_44.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -2989,6 +3270,7 @@ class Program_weight_tensor_parameter_280: class Program_weight_tensor_parameter_281: name = "parameter_281" + original_name = "batch_norm2d_44.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -2998,6 +3280,7 @@ class Program_weight_tensor_parameter_281: class Program_weight_tensor_parameter_282: name = "parameter_282" + original_name = "conv2d_48.w_0" shape = [64, 1, 5, 5] dtype = "float32" min_val = float("-0.48691") @@ -3009,6 +3292,7 @@ class Program_weight_tensor_parameter_282: class Program_weight_tensor_parameter_283: name = "parameter_283" + original_name = "batch_norm2d_43.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -3018,6 +3302,7 @@ class Program_weight_tensor_parameter_283: class Program_weight_tensor_parameter_284: name = "parameter_284" + original_name = "batch_norm2d_43.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -3027,6 +3312,7 @@ class Program_weight_tensor_parameter_284: class Program_weight_tensor_parameter_285: name = "parameter_285" + original_name = "batch_norm2d_43.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -3036,6 +3322,7 @@ class Program_weight_tensor_parameter_285: class Program_weight_tensor_parameter_286: name = "parameter_286" + original_name = "batch_norm2d_43.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -3045,6 +3332,7 @@ class Program_weight_tensor_parameter_286: class Program_weight_tensor_parameter_287: name = "parameter_287" + original_name = "conv2d_47.w_0" shape = [64, 64, 1, 1] dtype = "float32" min_val = float("-0.383156") @@ -3056,6 +3344,7 @@ class Program_weight_tensor_parameter_287: class Program_weight_tensor_parameter_288: name = "parameter_288" + original_name = "batch_norm2d_40.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -3065,6 +3354,7 @@ class Program_weight_tensor_parameter_288: class Program_weight_tensor_parameter_289: name = "parameter_289" + original_name = "batch_norm2d_40.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -3074,6 +3364,7 @@ class Program_weight_tensor_parameter_289: class Program_weight_tensor_parameter_290: name = "parameter_290" + original_name = "batch_norm2d_40.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -3083,6 +3374,7 @@ class Program_weight_tensor_parameter_290: class Program_weight_tensor_parameter_291: name = "parameter_291" + original_name = "batch_norm2d_40.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -3092,6 +3384,7 @@ class Program_weight_tensor_parameter_291: class Program_weight_tensor_parameter_292: name = "parameter_292" + original_name = "conv2d_44.w_0" shape = [64, 256, 1, 1] dtype = "float32" min_val = float("-0.398157") @@ -3103,6 +3396,7 @@ class Program_weight_tensor_parameter_292: class Program_weight_tensor_parameter_293: name = "parameter_293" + original_name = "batch_norm2d_41.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -3112,6 +3406,7 @@ class Program_weight_tensor_parameter_293: class Program_weight_tensor_parameter_294: name = "parameter_294" + original_name = "batch_norm2d_41.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -3121,6 +3416,7 @@ class Program_weight_tensor_parameter_294: class Program_weight_tensor_parameter_295: name = "parameter_295" + original_name = "batch_norm2d_41.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -3130,6 +3426,7 @@ class Program_weight_tensor_parameter_295: class Program_weight_tensor_parameter_296: name = "parameter_296" + original_name = "batch_norm2d_41.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -3139,6 +3436,7 @@ class Program_weight_tensor_parameter_296: class Program_weight_tensor_parameter_297: name = "parameter_297" + original_name = "conv2d_45.w_0" shape = [64, 256, 1, 1] dtype = "float32" min_val = float("-0.34841") @@ -3150,6 +3448,7 @@ class Program_weight_tensor_parameter_297: class Program_weight_tensor_parameter_298: name = "parameter_298" + original_name = "batch_norm2d_36.b_0" shape = [128] dtype = "float32" min_val = float("-0.439457") @@ -3161,6 +3460,7 @@ class Program_weight_tensor_parameter_298: class Program_weight_tensor_parameter_299: name = "parameter_299" + original_name = "batch_norm2d_36.w_0" shape = [128] dtype = "float32" min_val = float("0.150265") @@ -3172,6 +3472,7 @@ class Program_weight_tensor_parameter_299: class Program_weight_tensor_parameter_300: name = "parameter_300" + original_name = "batch_norm2d_36.w_2" shape = [128] dtype = "float32" min_val = float("0.00852005") @@ -3183,6 +3484,7 @@ class Program_weight_tensor_parameter_300: class Program_weight_tensor_parameter_301: name = "parameter_301" + original_name = "batch_norm2d_36.w_1" shape = [128] dtype = "float32" min_val = float("-0.0472771") @@ -3194,6 +3496,7 @@ class Program_weight_tensor_parameter_301: class Program_weight_tensor_parameter_302: name = "parameter_302" + original_name = "conv2d_40.w_0" shape = [128, 128, 1, 1] dtype = "float32" min_val = float("-0.425686") @@ -3205,6 +3508,7 @@ class Program_weight_tensor_parameter_302: class Program_weight_tensor_parameter_303: name = "parameter_303" + original_name = "batch_norm2d_39.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -3214,6 +3518,7 @@ class Program_weight_tensor_parameter_303: class Program_weight_tensor_parameter_304: name = "parameter_304" + original_name = "batch_norm2d_39.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -3223,6 +3528,7 @@ class Program_weight_tensor_parameter_304: class Program_weight_tensor_parameter_305: name = "parameter_305" + original_name = "batch_norm2d_39.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -3232,6 +3538,7 @@ class Program_weight_tensor_parameter_305: class Program_weight_tensor_parameter_306: name = "parameter_306" + original_name = "batch_norm2d_39.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -3241,6 +3548,7 @@ class Program_weight_tensor_parameter_306: class Program_weight_tensor_parameter_307: name = "parameter_307" + original_name = "conv2d_43.w_0" shape = [64, 64, 1, 1] dtype = "float32" min_val = float("-0.296093") @@ -3252,6 +3560,7 @@ class Program_weight_tensor_parameter_307: class Program_weight_tensor_parameter_308: name = "parameter_308" + original_name = "batch_norm2d_38.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -3261,6 +3570,7 @@ class Program_weight_tensor_parameter_308: class Program_weight_tensor_parameter_309: name = "parameter_309" + original_name = "batch_norm2d_38.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -3270,6 +3580,7 @@ class Program_weight_tensor_parameter_309: class Program_weight_tensor_parameter_310: name = "parameter_310" + original_name = "batch_norm2d_38.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -3279,6 +3590,7 @@ class Program_weight_tensor_parameter_310: class Program_weight_tensor_parameter_311: name = "parameter_311" + original_name = "batch_norm2d_38.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -3288,6 +3600,7 @@ class Program_weight_tensor_parameter_311: class Program_weight_tensor_parameter_312: name = "parameter_312" + original_name = "conv2d_42.w_0" shape = [64, 1, 5, 5] dtype = "float32" min_val = float("-0.411226") @@ -3299,6 +3612,7 @@ class Program_weight_tensor_parameter_312: class Program_weight_tensor_parameter_313: name = "parameter_313" + original_name = "batch_norm2d_37.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -3308,6 +3622,7 @@ class Program_weight_tensor_parameter_313: class Program_weight_tensor_parameter_314: name = "parameter_314" + original_name = "batch_norm2d_37.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -3317,6 +3632,7 @@ class Program_weight_tensor_parameter_314: class Program_weight_tensor_parameter_315: name = "parameter_315" + original_name = "batch_norm2d_37.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -3326,6 +3642,7 @@ class Program_weight_tensor_parameter_315: class Program_weight_tensor_parameter_316: name = "parameter_316" + original_name = "batch_norm2d_37.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -3335,6 +3652,7 @@ class Program_weight_tensor_parameter_316: class Program_weight_tensor_parameter_317: name = "parameter_317" + original_name = "conv2d_41.w_0" shape = [64, 64, 1, 1] dtype = "float32" min_val = float("-0.274894") @@ -3346,6 +3664,7 @@ class Program_weight_tensor_parameter_317: class Program_weight_tensor_parameter_318: name = "parameter_318" + original_name = "batch_norm2d_34.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -3355,6 +3674,7 @@ class Program_weight_tensor_parameter_318: class Program_weight_tensor_parameter_319: name = "parameter_319" + original_name = "batch_norm2d_34.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -3364,6 +3684,7 @@ class Program_weight_tensor_parameter_319: class Program_weight_tensor_parameter_320: name = "parameter_320" + original_name = "batch_norm2d_34.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -3373,6 +3694,7 @@ class Program_weight_tensor_parameter_320: class Program_weight_tensor_parameter_321: name = "parameter_321" + original_name = "batch_norm2d_34.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -3382,6 +3704,7 @@ class Program_weight_tensor_parameter_321: class Program_weight_tensor_parameter_322: name = "parameter_322" + original_name = "conv2d_38.w_0" shape = [64, 256, 1, 1] dtype = "float32" min_val = float("-0.29147") @@ -3393,6 +3716,7 @@ class Program_weight_tensor_parameter_322: class Program_weight_tensor_parameter_323: name = "parameter_323" + original_name = "batch_norm2d_35.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -3402,6 +3726,7 @@ class Program_weight_tensor_parameter_323: class Program_weight_tensor_parameter_324: name = "parameter_324" + original_name = "batch_norm2d_35.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -3411,6 +3736,7 @@ class Program_weight_tensor_parameter_324: class Program_weight_tensor_parameter_325: name = "parameter_325" + original_name = "batch_norm2d_35.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -3420,6 +3746,7 @@ class Program_weight_tensor_parameter_325: class Program_weight_tensor_parameter_326: name = "parameter_326" + original_name = "batch_norm2d_35.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -3429,6 +3756,7 @@ class Program_weight_tensor_parameter_326: class Program_weight_tensor_parameter_327: name = "parameter_327" + original_name = "conv2d_39.w_0" shape = [64, 256, 1, 1] dtype = "float32" min_val = float("-0.454853") @@ -3440,6 +3768,7 @@ class Program_weight_tensor_parameter_327: class Program_weight_tensor_parameter_328: name = "parameter_328" + original_name = "batch_norm2d_29.b_0" shape = [128] dtype = "float32" min_val = float("-0.622133") @@ -3451,6 +3780,7 @@ class Program_weight_tensor_parameter_328: class Program_weight_tensor_parameter_329: name = "parameter_329" + original_name = "batch_norm2d_29.w_0" shape = [128] dtype = "float32" min_val = float("0.140385") @@ -3462,6 +3792,7 @@ class Program_weight_tensor_parameter_329: class Program_weight_tensor_parameter_330: name = "parameter_330" + original_name = "batch_norm2d_29.w_2" shape = [128] dtype = "float32" min_val = float("1.65991") @@ -3473,6 +3804,7 @@ class Program_weight_tensor_parameter_330: class Program_weight_tensor_parameter_331: name = "parameter_331" + original_name = "batch_norm2d_29.w_1" shape = [128] dtype = "float32" min_val = float("-2.53177") @@ -3484,6 +3816,7 @@ class Program_weight_tensor_parameter_331: class Program_weight_tensor_parameter_332: name = "parameter_332" + original_name = "conv2d_33.w_0" shape = [128, 1280, 1, 1] dtype = "float32" min_val = float("-0.297541") @@ -3495,6 +3828,7 @@ class Program_weight_tensor_parameter_332: class Program_weight_tensor_parameter_333: name = "parameter_333" + original_name = "batch_norm2d_28.b_0" shape = [128] dtype = "float32" min_val = float("-0.124373") @@ -3506,6 +3840,7 @@ class Program_weight_tensor_parameter_333: class Program_weight_tensor_parameter_334: name = "parameter_334" + original_name = "batch_norm2d_28.w_0" shape = [128] dtype = "float32" min_val = float("0.0647393") @@ -3517,6 +3852,7 @@ class Program_weight_tensor_parameter_334: class Program_weight_tensor_parameter_335: name = "parameter_335" + original_name = "batch_norm2d_28.w_2" shape = [128] dtype = "float32" min_val = float("0.00838456") @@ -3528,6 +3864,7 @@ class Program_weight_tensor_parameter_335: class Program_weight_tensor_parameter_336: name = "parameter_336" + original_name = "batch_norm2d_28.w_1" shape = [128] dtype = "float32" min_val = float("-0.424838") @@ -3539,6 +3876,7 @@ class Program_weight_tensor_parameter_336: class Program_weight_tensor_parameter_337: name = "parameter_337" + original_name = "conv2d_32.w_0" shape = [128, 640, 1, 1] dtype = "float32" min_val = float("-0.226608") @@ -3550,6 +3888,7 @@ class Program_weight_tensor_parameter_337: class Program_weight_tensor_parameter_338: name = "parameter_338" + original_name = "batch_norm2d_27.b_0" shape = [128] dtype = "float32" min_val = float("-0.191131") @@ -3561,6 +3900,7 @@ class Program_weight_tensor_parameter_338: class Program_weight_tensor_parameter_339: name = "parameter_339" + original_name = "batch_norm2d_27.w_0" shape = [128] dtype = "float32" min_val = float("0.0802208") @@ -3572,6 +3912,7 @@ class Program_weight_tensor_parameter_339: class Program_weight_tensor_parameter_340: name = "parameter_340" + original_name = "batch_norm2d_27.w_2" shape = [128] dtype = "float32" min_val = float("0.0042114") @@ -3583,6 +3924,7 @@ class Program_weight_tensor_parameter_340: class Program_weight_tensor_parameter_341: name = "parameter_341" + original_name = "batch_norm2d_27.w_1" shape = [128] dtype = "float32" min_val = float("-1.03688") @@ -3594,6 +3936,7 @@ class Program_weight_tensor_parameter_341: class Program_weight_tensor_parameter_342: name = "parameter_342" + original_name = "conv2d_31.w_0" shape = [128, 320, 1, 1] dtype = "float32" min_val = float("-0.311938") @@ -3605,6 +3948,7 @@ class Program_weight_tensor_parameter_342: class Program_weight_tensor_parameter_343: name = "parameter_343" + original_name = "batch_norm2d_26.b_0" shape = [1280] dtype = "float32" min_val = float("-8.8846") @@ -3616,6 +3960,7 @@ class Program_weight_tensor_parameter_343: class Program_weight_tensor_parameter_344: name = "parameter_344" + original_name = "batch_norm2d_26.w_0" shape = [1280] dtype = "float32" min_val = float("2.05074") @@ -3627,6 +3972,7 @@ class Program_weight_tensor_parameter_344: class Program_weight_tensor_parameter_345: name = "parameter_345" + original_name = "batch_norm2d_26.w_2" shape = [1280] dtype = "float32" min_val = float("0.0569627") @@ -3638,6 +3984,7 @@ class Program_weight_tensor_parameter_345: class Program_weight_tensor_parameter_346: name = "parameter_346" + original_name = "batch_norm2d_26.w_1" shape = [1280] dtype = "float32" min_val = float("-0.659116") @@ -3649,6 +3996,7 @@ class Program_weight_tensor_parameter_346: class Program_weight_tensor_parameter_347: name = "parameter_347" + original_name = "conv2d_30.w_0" shape = [1280, 1280, 1, 1] dtype = "float32" min_val = float("-0.266364") @@ -3660,6 +4008,7 @@ class Program_weight_tensor_parameter_347: class Program_weight_tensor_parameter_348: name = "parameter_348" + original_name = "conv2d_29.b_0" shape = [1280] dtype = "float32" min_val = float("-0.042786") @@ -3671,6 +4020,7 @@ class Program_weight_tensor_parameter_348: class Program_weight_tensor_parameter_349: name = "parameter_349" + original_name = "conv2d_29.w_0" shape = [1280, 320, 1, 1] dtype = "float32" min_val = float("-0.367974") @@ -3682,6 +4032,7 @@ class Program_weight_tensor_parameter_349: class Program_weight_tensor_parameter_350: name = "parameter_350" + original_name = "conv2d_28.b_0" shape = [320] dtype = "float32" min_val = float("-0.0111934") @@ -3693,6 +4044,7 @@ class Program_weight_tensor_parameter_350: class Program_weight_tensor_parameter_351: name = "parameter_351" + original_name = "conv2d_28.w_0" shape = [320, 1280, 1, 1] dtype = "float32" min_val = float("-0.397549") @@ -3704,6 +4056,7 @@ class Program_weight_tensor_parameter_351: class Program_weight_tensor_parameter_352: name = "parameter_352" + original_name = "batch_norm2d_25.b_0" shape = [1280] dtype = "float32" min_val = float("-4.71811") @@ -3715,6 +4068,7 @@ class Program_weight_tensor_parameter_352: class Program_weight_tensor_parameter_353: name = "parameter_353" + original_name = "batch_norm2d_25.w_0" shape = [1280] dtype = "float32" min_val = float("0.441272") @@ -3726,6 +4080,7 @@ class Program_weight_tensor_parameter_353: class Program_weight_tensor_parameter_354: name = "parameter_354" + original_name = "batch_norm2d_25.w_2" shape = [1280] dtype = "float32" min_val = float("0.000109978") @@ -3737,6 +4092,7 @@ class Program_weight_tensor_parameter_354: class Program_weight_tensor_parameter_355: name = "parameter_355" + original_name = "batch_norm2d_25.w_1" shape = [1280] dtype = "float32" min_val = float("-0.844568") @@ -3748,6 +4104,7 @@ class Program_weight_tensor_parameter_355: class Program_weight_tensor_parameter_356: name = "parameter_356" + original_name = "conv2d_27.w_0" shape = [1280, 1, 5, 5] dtype = "float32" min_val = float("-0.426666") @@ -3759,6 +4116,7 @@ class Program_weight_tensor_parameter_356: class Program_weight_tensor_parameter_357: name = "parameter_357" + original_name = "batch_norm2d_24.b_0" shape = [1280] dtype = "float32" min_val = float("-2.84991") @@ -3770,6 +4128,7 @@ class Program_weight_tensor_parameter_357: class Program_weight_tensor_parameter_358: name = "parameter_358" + original_name = "batch_norm2d_24.w_0" shape = [1280] dtype = "float32" min_val = float("-1.14982") @@ -3781,6 +4140,7 @@ class Program_weight_tensor_parameter_358: class Program_weight_tensor_parameter_359: name = "parameter_359" + original_name = "batch_norm2d_24.w_2" shape = [1280] dtype = "float32" min_val = float("0.0464973") @@ -3792,6 +4152,7 @@ class Program_weight_tensor_parameter_359: class Program_weight_tensor_parameter_360: name = "parameter_360" + original_name = "batch_norm2d_24.w_1" shape = [1280] dtype = "float32" min_val = float("-1.05058") @@ -3803,6 +4164,7 @@ class Program_weight_tensor_parameter_360: class Program_weight_tensor_parameter_361: name = "parameter_361" + original_name = "conv2d_26.w_0" shape = [1280, 640, 1, 1] dtype = "float32" min_val = float("-0.534266") @@ -3814,6 +4176,7 @@ class Program_weight_tensor_parameter_361: class Program_weight_tensor_parameter_362: name = "parameter_362" + original_name = "conv2d_25.b_0" shape = [640] dtype = "float32" min_val = float("-0.0719395") @@ -3825,6 +4188,7 @@ class Program_weight_tensor_parameter_362: class Program_weight_tensor_parameter_363: name = "parameter_363" + original_name = "conv2d_25.w_0" shape = [640, 160, 1, 1] dtype = "float32" min_val = float("-0.338524") @@ -3836,6 +4200,7 @@ class Program_weight_tensor_parameter_363: class Program_weight_tensor_parameter_364: name = "parameter_364" + original_name = "conv2d_24.b_0" shape = [160] dtype = "float32" min_val = float("-0.000925034") @@ -3847,6 +4212,7 @@ class Program_weight_tensor_parameter_364: class Program_weight_tensor_parameter_365: name = "parameter_365" + original_name = "conv2d_24.w_0" shape = [160, 640, 1, 1] dtype = "float32" min_val = float("-0.277791") @@ -3858,6 +4224,7 @@ class Program_weight_tensor_parameter_365: class Program_weight_tensor_parameter_366: name = "parameter_366" + original_name = "batch_norm2d_23.b_0" shape = [640] dtype = "float32" min_val = float("-3.81155") @@ -3869,6 +4236,7 @@ class Program_weight_tensor_parameter_366: class Program_weight_tensor_parameter_367: name = "parameter_367" + original_name = "batch_norm2d_23.w_0" shape = [640] dtype = "float32" min_val = float("0.693287") @@ -3880,6 +4248,7 @@ class Program_weight_tensor_parameter_367: class Program_weight_tensor_parameter_368: name = "parameter_368" + original_name = "batch_norm2d_23.w_2" shape = [640] dtype = "float32" min_val = float("0.000126229") @@ -3891,6 +4260,7 @@ class Program_weight_tensor_parameter_368: class Program_weight_tensor_parameter_369: name = "parameter_369" + original_name = "batch_norm2d_23.w_1" shape = [640] dtype = "float32" min_val = float("-1.18432") @@ -3902,6 +4272,7 @@ class Program_weight_tensor_parameter_369: class Program_weight_tensor_parameter_370: name = "parameter_370" + original_name = "conv2d_23.w_0" shape = [640, 1, 5, 5] dtype = "float32" min_val = float("-0.230546") @@ -3913,6 +4284,7 @@ class Program_weight_tensor_parameter_370: class Program_weight_tensor_parameter_371: name = "parameter_371" + original_name = "batch_norm2d_22.b_0" shape = [640] dtype = "float32" min_val = float("-3.05791") @@ -3924,6 +4296,7 @@ class Program_weight_tensor_parameter_371: class Program_weight_tensor_parameter_372: name = "parameter_372" + original_name = "batch_norm2d_22.w_0" shape = [640] dtype = "float32" min_val = float("-1.35402") @@ -3935,6 +4308,7 @@ class Program_weight_tensor_parameter_372: class Program_weight_tensor_parameter_373: name = "parameter_373" + original_name = "batch_norm2d_22.w_2" shape = [640] dtype = "float32" min_val = float("0.233821") @@ -3946,6 +4320,7 @@ class Program_weight_tensor_parameter_373: class Program_weight_tensor_parameter_374: name = "parameter_374" + original_name = "batch_norm2d_22.w_1" shape = [640] dtype = "float32" min_val = float("-5.32728") @@ -3957,6 +4332,7 @@ class Program_weight_tensor_parameter_374: class Program_weight_tensor_parameter_375: name = "parameter_375" + original_name = "conv2d_22.w_0" shape = [640, 640, 1, 1] dtype = "float32" min_val = float("-0.803609") @@ -3968,6 +4344,7 @@ class Program_weight_tensor_parameter_375: class Program_weight_tensor_parameter_376: name = "parameter_376" + original_name = "batch_norm2d_21.b_0" shape = [640] dtype = "float32" min_val = float("-7.57989") @@ -3979,6 +4356,7 @@ class Program_weight_tensor_parameter_376: class Program_weight_tensor_parameter_377: name = "parameter_377" + original_name = "batch_norm2d_21.w_0" shape = [640] dtype = "float32" min_val = float("0.373415") @@ -3990,6 +4368,7 @@ class Program_weight_tensor_parameter_377: class Program_weight_tensor_parameter_378: name = "parameter_378" + original_name = "batch_norm2d_21.w_2" shape = [640] dtype = "float32" min_val = float("0.000184215") @@ -4001,6 +4380,7 @@ class Program_weight_tensor_parameter_378: class Program_weight_tensor_parameter_379: name = "parameter_379" + original_name = "batch_norm2d_21.w_1" shape = [640] dtype = "float32" min_val = float("-1.1274") @@ -4012,6 +4392,7 @@ class Program_weight_tensor_parameter_379: class Program_weight_tensor_parameter_380: name = "parameter_380" + original_name = "conv2d_21.w_0" shape = [640, 1, 5, 5] dtype = "float32" min_val = float("-0.541308") @@ -4023,6 +4404,7 @@ class Program_weight_tensor_parameter_380: class Program_weight_tensor_parameter_381: name = "parameter_381" + original_name = "batch_norm2d_20.b_0" shape = [640] dtype = "float32" min_val = float("-3.19086") @@ -4034,6 +4416,7 @@ class Program_weight_tensor_parameter_381: class Program_weight_tensor_parameter_382: name = "parameter_382" + original_name = "batch_norm2d_20.w_0" shape = [640] dtype = "float32" min_val = float("-0.663839") @@ -4045,6 +4428,7 @@ class Program_weight_tensor_parameter_382: class Program_weight_tensor_parameter_383: name = "parameter_383" + original_name = "batch_norm2d_20.w_2" shape = [640] dtype = "float32" min_val = float("0.269841") @@ -4056,6 +4440,7 @@ class Program_weight_tensor_parameter_383: class Program_weight_tensor_parameter_384: name = "parameter_384" + original_name = "batch_norm2d_20.w_1" shape = [640] dtype = "float32" min_val = float("-4.27379") @@ -4067,6 +4452,7 @@ class Program_weight_tensor_parameter_384: class Program_weight_tensor_parameter_385: name = "parameter_385" + original_name = "conv2d_20.w_0" shape = [640, 640, 1, 1] dtype = "float32" min_val = float("-0.804398") @@ -4078,6 +4464,7 @@ class Program_weight_tensor_parameter_385: class Program_weight_tensor_parameter_386: name = "parameter_386" + original_name = "batch_norm2d_19.b_0" shape = [640] dtype = "float32" min_val = float("-3.4579") @@ -4089,6 +4476,7 @@ class Program_weight_tensor_parameter_386: class Program_weight_tensor_parameter_387: name = "parameter_387" + original_name = "batch_norm2d_19.w_0" shape = [640] dtype = "float32" min_val = float("0.376776") @@ -4100,6 +4488,7 @@ class Program_weight_tensor_parameter_387: class Program_weight_tensor_parameter_388: name = "parameter_388" + original_name = "batch_norm2d_19.w_2" shape = [640] dtype = "float32" min_val = float("0.00013012") @@ -4111,6 +4500,7 @@ class Program_weight_tensor_parameter_388: class Program_weight_tensor_parameter_389: name = "parameter_389" + original_name = "batch_norm2d_19.w_1" shape = [640] dtype = "float32" min_val = float("-2.38378") @@ -4122,6 +4512,7 @@ class Program_weight_tensor_parameter_389: class Program_weight_tensor_parameter_390: name = "parameter_390" + original_name = "conv2d_19.w_0" shape = [640, 1, 5, 5] dtype = "float32" min_val = float("-0.483452") @@ -4133,6 +4524,7 @@ class Program_weight_tensor_parameter_390: class Program_weight_tensor_parameter_391: name = "parameter_391" + original_name = "batch_norm2d_18.b_0" shape = [640] dtype = "float32" min_val = float("-3.61245") @@ -4144,6 +4536,7 @@ class Program_weight_tensor_parameter_391: class Program_weight_tensor_parameter_392: name = "parameter_392" + original_name = "batch_norm2d_18.w_0" shape = [640] dtype = "float32" min_val = float("-0.636083") @@ -4155,6 +4548,7 @@ class Program_weight_tensor_parameter_392: class Program_weight_tensor_parameter_393: name = "parameter_393" + original_name = "batch_norm2d_18.w_2" shape = [640] dtype = "float32" min_val = float("0.26087") @@ -4166,6 +4560,7 @@ class Program_weight_tensor_parameter_393: class Program_weight_tensor_parameter_394: name = "parameter_394" + original_name = "batch_norm2d_18.w_1" shape = [640] dtype = "float32" min_val = float("-4.27354") @@ -4177,6 +4572,7 @@ class Program_weight_tensor_parameter_394: class Program_weight_tensor_parameter_395: name = "parameter_395" + original_name = "conv2d_18.w_0" shape = [640, 640, 1, 1] dtype = "float32" min_val = float("-0.845832") @@ -4188,6 +4584,7 @@ class Program_weight_tensor_parameter_395: class Program_weight_tensor_parameter_396: name = "parameter_396" + original_name = "batch_norm2d_17.b_0" shape = [640] dtype = "float32" min_val = float("-5.466") @@ -4199,6 +4596,7 @@ class Program_weight_tensor_parameter_396: class Program_weight_tensor_parameter_397: name = "parameter_397" + original_name = "batch_norm2d_17.w_0" shape = [640] dtype = "float32" min_val = float("0.339562") @@ -4210,6 +4608,7 @@ class Program_weight_tensor_parameter_397: class Program_weight_tensor_parameter_398: name = "parameter_398" + original_name = "batch_norm2d_17.w_2" shape = [640] dtype = "float32" min_val = float("0.000121584") @@ -4221,6 +4620,7 @@ class Program_weight_tensor_parameter_398: class Program_weight_tensor_parameter_399: name = "parameter_399" + original_name = "batch_norm2d_17.w_1" shape = [640] dtype = "float32" min_val = float("-3.01173") @@ -4232,6 +4632,7 @@ class Program_weight_tensor_parameter_399: class Program_weight_tensor_parameter_400: name = "parameter_400" + original_name = "conv2d_17.w_0" shape = [640, 1, 5, 5] dtype = "float32" min_val = float("-0.539614") @@ -4243,6 +4644,7 @@ class Program_weight_tensor_parameter_400: class Program_weight_tensor_parameter_401: name = "parameter_401" + original_name = "batch_norm2d_16.b_0" shape = [640] dtype = "float32" min_val = float("-2.75492") @@ -4254,6 +4656,7 @@ class Program_weight_tensor_parameter_401: class Program_weight_tensor_parameter_402: name = "parameter_402" + original_name = "batch_norm2d_16.w_0" shape = [640] dtype = "float32" min_val = float("-0.645262") @@ -4265,6 +4668,7 @@ class Program_weight_tensor_parameter_402: class Program_weight_tensor_parameter_403: name = "parameter_403" + original_name = "batch_norm2d_16.w_2" shape = [640] dtype = "float32" min_val = float("0.20784") @@ -4276,6 +4680,7 @@ class Program_weight_tensor_parameter_403: class Program_weight_tensor_parameter_404: name = "parameter_404" + original_name = "batch_norm2d_16.w_1" shape = [640] dtype = "float32" min_val = float("-4.60523") @@ -4287,6 +4692,7 @@ class Program_weight_tensor_parameter_404: class Program_weight_tensor_parameter_405: name = "parameter_405" + original_name = "conv2d_16.w_0" shape = [640, 640, 1, 1] dtype = "float32" min_val = float("-1.09987") @@ -4298,6 +4704,7 @@ class Program_weight_tensor_parameter_405: class Program_weight_tensor_parameter_406: name = "parameter_406" + original_name = "batch_norm2d_15.b_0" shape = [640] dtype = "float32" min_val = float("-3.34751") @@ -4309,6 +4716,7 @@ class Program_weight_tensor_parameter_406: class Program_weight_tensor_parameter_407: name = "parameter_407" + original_name = "batch_norm2d_15.w_0" shape = [640] dtype = "float32" min_val = float("0.35337") @@ -4320,6 +4728,7 @@ class Program_weight_tensor_parameter_407: class Program_weight_tensor_parameter_408: name = "parameter_408" + original_name = "batch_norm2d_15.w_2" shape = [640] dtype = "float32" min_val = float("0.000115833") @@ -4331,6 +4740,7 @@ class Program_weight_tensor_parameter_408: class Program_weight_tensor_parameter_409: name = "parameter_409" + original_name = "batch_norm2d_15.w_1" shape = [640] dtype = "float32" min_val = float("-2.43344") @@ -4342,6 +4752,7 @@ class Program_weight_tensor_parameter_409: class Program_weight_tensor_parameter_410: name = "parameter_410" + original_name = "conv2d_15.w_0" shape = [640, 1, 5, 5] dtype = "float32" min_val = float("-0.820659") @@ -4353,6 +4764,7 @@ class Program_weight_tensor_parameter_410: class Program_weight_tensor_parameter_411: name = "parameter_411" + original_name = "batch_norm2d_14.b_0" shape = [640] dtype = "float32" min_val = float("-4.1585") @@ -4364,6 +4776,7 @@ class Program_weight_tensor_parameter_411: class Program_weight_tensor_parameter_412: name = "parameter_412" + original_name = "batch_norm2d_14.w_0" shape = [640] dtype = "float32" min_val = float("-0.554302") @@ -4375,6 +4788,7 @@ class Program_weight_tensor_parameter_412: class Program_weight_tensor_parameter_413: name = "parameter_413" + original_name = "batch_norm2d_14.w_2" shape = [640] dtype = "float32" min_val = float("0.309801") @@ -4386,6 +4800,7 @@ class Program_weight_tensor_parameter_413: class Program_weight_tensor_parameter_414: name = "parameter_414" + original_name = "batch_norm2d_14.w_1" shape = [640] dtype = "float32" min_val = float("-4.26121") @@ -4397,6 +4812,7 @@ class Program_weight_tensor_parameter_414: class Program_weight_tensor_parameter_415: name = "parameter_415" + original_name = "conv2d_14.w_0" shape = [640, 640, 1, 1] dtype = "float32" min_val = float("-1.0905") @@ -4408,6 +4824,7 @@ class Program_weight_tensor_parameter_415: class Program_weight_tensor_parameter_416: name = "parameter_416" + original_name = "batch_norm2d_13.b_0" shape = [640] dtype = "float32" min_val = float("-4.47224") @@ -4419,6 +4836,7 @@ class Program_weight_tensor_parameter_416: class Program_weight_tensor_parameter_417: name = "parameter_417" + original_name = "batch_norm2d_13.w_0" shape = [640] dtype = "float32" min_val = float("0.379251") @@ -4430,6 +4848,7 @@ class Program_weight_tensor_parameter_417: class Program_weight_tensor_parameter_418: name = "parameter_418" + original_name = "batch_norm2d_13.w_2" shape = [640] dtype = "float32" min_val = float("0.00012172") @@ -4441,6 +4860,7 @@ class Program_weight_tensor_parameter_418: class Program_weight_tensor_parameter_419: name = "parameter_419" + original_name = "batch_norm2d_13.w_1" shape = [640] dtype = "float32" min_val = float("-3.96486") @@ -4452,6 +4872,7 @@ class Program_weight_tensor_parameter_419: class Program_weight_tensor_parameter_420: name = "parameter_420" + original_name = "conv2d_13.w_0" shape = [640, 1, 5, 5] dtype = "float32" min_val = float("-0.574642") @@ -4463,6 +4884,7 @@ class Program_weight_tensor_parameter_420: class Program_weight_tensor_parameter_421: name = "parameter_421" + original_name = "batch_norm2d_12.b_0" shape = [640] dtype = "float32" min_val = float("-2.37543") @@ -4474,6 +4896,7 @@ class Program_weight_tensor_parameter_421: class Program_weight_tensor_parameter_422: name = "parameter_422" + original_name = "batch_norm2d_12.w_0" shape = [640] dtype = "float32" min_val = float("-0.228328") @@ -4485,6 +4908,7 @@ class Program_weight_tensor_parameter_422: class Program_weight_tensor_parameter_423: name = "parameter_423" + original_name = "batch_norm2d_12.w_2" shape = [640] dtype = "float32" min_val = float("0.0976078") @@ -4496,6 +4920,7 @@ class Program_weight_tensor_parameter_423: class Program_weight_tensor_parameter_424: name = "parameter_424" + original_name = "batch_norm2d_12.w_1" shape = [640] dtype = "float32" min_val = float("-3.77979") @@ -4507,6 +4932,7 @@ class Program_weight_tensor_parameter_424: class Program_weight_tensor_parameter_425: name = "parameter_425" + original_name = "conv2d_12.w_0" shape = [640, 320, 1, 1] dtype = "float32" min_val = float("-0.564012") @@ -4518,6 +4944,7 @@ class Program_weight_tensor_parameter_425: class Program_weight_tensor_parameter_426: name = "parameter_426" + original_name = "batch_norm2d_11.b_0" shape = [320] dtype = "float32" min_val = float("-1.48436") @@ -4529,6 +4956,7 @@ class Program_weight_tensor_parameter_426: class Program_weight_tensor_parameter_427: name = "parameter_427" + original_name = "batch_norm2d_11.w_0" shape = [320] dtype = "float32" min_val = float("0.351578") @@ -4540,6 +4968,7 @@ class Program_weight_tensor_parameter_427: class Program_weight_tensor_parameter_428: name = "parameter_428" + original_name = "batch_norm2d_11.w_2" shape = [320] dtype = "float32" min_val = float("9.15431e-05") @@ -4551,6 +4980,7 @@ class Program_weight_tensor_parameter_428: class Program_weight_tensor_parameter_429: name = "parameter_429" + original_name = "batch_norm2d_11.w_1" shape = [320] dtype = "float32" min_val = float("-2.2357") @@ -4562,6 +4992,7 @@ class Program_weight_tensor_parameter_429: class Program_weight_tensor_parameter_430: name = "parameter_430" + original_name = "conv2d_11.w_0" shape = [320, 1, 3, 3] dtype = "float32" min_val = float("-0.26875") @@ -4573,6 +5004,7 @@ class Program_weight_tensor_parameter_430: class Program_weight_tensor_parameter_431: name = "parameter_431" + original_name = "batch_norm2d_10.b_0" shape = [320] dtype = "float32" min_val = float("-2.83582") @@ -4584,6 +5016,7 @@ class Program_weight_tensor_parameter_431: class Program_weight_tensor_parameter_432: name = "parameter_432" + original_name = "batch_norm2d_10.w_0" shape = [320] dtype = "float32" min_val = float("-0.773699") @@ -4595,6 +5028,7 @@ class Program_weight_tensor_parameter_432: class Program_weight_tensor_parameter_433: name = "parameter_433" + original_name = "batch_norm2d_10.w_2" shape = [320] dtype = "float32" min_val = float("0.196441") @@ -4606,6 +5040,7 @@ class Program_weight_tensor_parameter_433: class Program_weight_tensor_parameter_434: name = "parameter_434" + original_name = "batch_norm2d_10.w_1" shape = [320] dtype = "float32" min_val = float("-4.39517") @@ -4617,6 +5052,7 @@ class Program_weight_tensor_parameter_434: class Program_weight_tensor_parameter_435: name = "parameter_435" + original_name = "conv2d_10.w_0" shape = [320, 320, 1, 1] dtype = "float32" min_val = float("-0.641979") @@ -4628,6 +5064,7 @@ class Program_weight_tensor_parameter_435: class Program_weight_tensor_parameter_436: name = "parameter_436" + original_name = "batch_norm2d_9.b_0" shape = [320] dtype = "float32" min_val = float("-1.96779") @@ -4639,6 +5076,7 @@ class Program_weight_tensor_parameter_436: class Program_weight_tensor_parameter_437: name = "parameter_437" + original_name = "batch_norm2d_9.w_0" shape = [320] dtype = "float32" min_val = float("0.328643") @@ -4650,6 +5088,7 @@ class Program_weight_tensor_parameter_437: class Program_weight_tensor_parameter_438: name = "parameter_438" + original_name = "batch_norm2d_9.w_2" shape = [320] dtype = "float32" min_val = float("8.86168e-05") @@ -4661,6 +5100,7 @@ class Program_weight_tensor_parameter_438: class Program_weight_tensor_parameter_439: name = "parameter_439" + original_name = "batch_norm2d_9.w_1" shape = [320] dtype = "float32" min_val = float("-7.02641") @@ -4672,6 +5112,7 @@ class Program_weight_tensor_parameter_439: class Program_weight_tensor_parameter_440: name = "parameter_440" + original_name = "conv2d_9.w_0" shape = [320, 1, 3, 3] dtype = "float32" min_val = float("-0.854933") @@ -4683,6 +5124,7 @@ class Program_weight_tensor_parameter_440: class Program_weight_tensor_parameter_441: name = "parameter_441" + original_name = "batch_norm2d_8.b_0" shape = [320] dtype = "float32" min_val = float("-1.83963") @@ -4694,6 +5136,7 @@ class Program_weight_tensor_parameter_441: class Program_weight_tensor_parameter_442: name = "parameter_442" + original_name = "batch_norm2d_8.w_0" shape = [320] dtype = "float32" min_val = float("-0.543489") @@ -4705,6 +5148,7 @@ class Program_weight_tensor_parameter_442: class Program_weight_tensor_parameter_443: name = "parameter_443" + original_name = "batch_norm2d_8.w_2" shape = [320] dtype = "float32" min_val = float("0.135784") @@ -4716,6 +5160,7 @@ class Program_weight_tensor_parameter_443: class Program_weight_tensor_parameter_444: name = "parameter_444" + original_name = "batch_norm2d_8.w_1" shape = [320] dtype = "float32" min_val = float("-5.16552") @@ -4727,6 +5172,7 @@ class Program_weight_tensor_parameter_444: class Program_weight_tensor_parameter_445: name = "parameter_445" + original_name = "conv2d_8.w_0" shape = [320, 160, 1, 1] dtype = "float32" min_val = float("-0.765441") @@ -4738,6 +5184,7 @@ class Program_weight_tensor_parameter_445: class Program_weight_tensor_parameter_446: name = "parameter_446" + original_name = "batch_norm2d_7.b_0" shape = [160] dtype = "float32" min_val = float("-1.04661") @@ -4749,6 +5196,7 @@ class Program_weight_tensor_parameter_446: class Program_weight_tensor_parameter_447: name = "parameter_447" + original_name = "batch_norm2d_7.w_0" shape = [160] dtype = "float32" min_val = float("0.310809") @@ -4760,6 +5208,7 @@ class Program_weight_tensor_parameter_447: class Program_weight_tensor_parameter_448: name = "parameter_448" + original_name = "batch_norm2d_7.w_2" shape = [160] dtype = "float32" min_val = float("0.000196213") @@ -4771,6 +5220,7 @@ class Program_weight_tensor_parameter_448: class Program_weight_tensor_parameter_449: name = "parameter_449" + original_name = "batch_norm2d_7.w_1" shape = [160] dtype = "float32" min_val = float("-2.19277") @@ -4782,6 +5232,7 @@ class Program_weight_tensor_parameter_449: class Program_weight_tensor_parameter_450: name = "parameter_450" + original_name = "conv2d_7.w_0" shape = [160, 1, 3, 3] dtype = "float32" min_val = float("-0.38736") @@ -4793,6 +5244,7 @@ class Program_weight_tensor_parameter_450: class Program_weight_tensor_parameter_451: name = "parameter_451" + original_name = "batch_norm2d_6.b_0" shape = [160] dtype = "float32" min_val = float("-2.50122") @@ -4804,6 +5256,7 @@ class Program_weight_tensor_parameter_451: class Program_weight_tensor_parameter_452: name = "parameter_452" + original_name = "batch_norm2d_6.w_0" shape = [160] dtype = "float32" min_val = float("-0.0543131") @@ -4815,6 +5268,7 @@ class Program_weight_tensor_parameter_452: class Program_weight_tensor_parameter_453: name = "parameter_453" + original_name = "batch_norm2d_6.w_2" shape = [160] dtype = "float32" min_val = float("1.33736") @@ -4826,6 +5280,7 @@ class Program_weight_tensor_parameter_453: class Program_weight_tensor_parameter_454: name = "parameter_454" + original_name = "batch_norm2d_6.w_1" shape = [160] dtype = "float32" min_val = float("-8.2722") @@ -4837,6 +5292,7 @@ class Program_weight_tensor_parameter_454: class Program_weight_tensor_parameter_455: name = "parameter_455" + original_name = "conv2d_6.w_0" shape = [160, 160, 1, 1] dtype = "float32" min_val = float("-0.792491") @@ -4848,6 +5304,7 @@ class Program_weight_tensor_parameter_455: class Program_weight_tensor_parameter_456: name = "parameter_456" + original_name = "batch_norm2d_5.b_0" shape = [160] dtype = "float32" min_val = float("-2.352") @@ -4859,6 +5316,7 @@ class Program_weight_tensor_parameter_456: class Program_weight_tensor_parameter_457: name = "parameter_457" + original_name = "batch_norm2d_5.w_0" shape = [160] dtype = "float32" min_val = float("0.820069") @@ -4870,6 +5328,7 @@ class Program_weight_tensor_parameter_457: class Program_weight_tensor_parameter_458: name = "parameter_458" + original_name = "batch_norm2d_5.w_2" shape = [160] dtype = "float32" min_val = float("0.000168784") @@ -4881,6 +5340,7 @@ class Program_weight_tensor_parameter_458: class Program_weight_tensor_parameter_459: name = "parameter_459" + original_name = "batch_norm2d_5.w_1" shape = [160] dtype = "float32" min_val = float("-2.00332") @@ -4892,6 +5352,7 @@ class Program_weight_tensor_parameter_459: class Program_weight_tensor_parameter_460: name = "parameter_460" + original_name = "conv2d_5.w_0" shape = [160, 1, 3, 3] dtype = "float32" min_val = float("-0.810507") @@ -4903,6 +5364,7 @@ class Program_weight_tensor_parameter_460: class Program_weight_tensor_parameter_461: name = "parameter_461" + original_name = "batch_norm2d_4.b_0" shape = [160] dtype = "float32" min_val = float("-2.73212") @@ -4914,6 +5376,7 @@ class Program_weight_tensor_parameter_461: class Program_weight_tensor_parameter_462: name = "parameter_462" + original_name = "batch_norm2d_4.w_0" shape = [160] dtype = "float32" min_val = float("-0.62687") @@ -4925,6 +5388,7 @@ class Program_weight_tensor_parameter_462: class Program_weight_tensor_parameter_463: name = "parameter_463" + original_name = "batch_norm2d_4.w_2" shape = [160] dtype = "float32" min_val = float("1.61221") @@ -4936,6 +5400,7 @@ class Program_weight_tensor_parameter_463: class Program_weight_tensor_parameter_464: name = "parameter_464" + original_name = "batch_norm2d_4.w_1" shape = [160] dtype = "float32" min_val = float("-14.4648") @@ -4947,6 +5412,7 @@ class Program_weight_tensor_parameter_464: class Program_weight_tensor_parameter_465: name = "parameter_465" + original_name = "conv2d_4.w_0" shape = [160, 80, 1, 1] dtype = "float32" min_val = float("-0.636448") @@ -4958,6 +5424,7 @@ class Program_weight_tensor_parameter_465: class Program_weight_tensor_parameter_466: name = "parameter_466" + original_name = "batch_norm2d_3.b_0" shape = [80] dtype = "float32" min_val = float("-1.70845") @@ -4969,6 +5436,7 @@ class Program_weight_tensor_parameter_466: class Program_weight_tensor_parameter_467: name = "parameter_467" + original_name = "batch_norm2d_3.w_0" shape = [80] dtype = "float32" min_val = float("1.42176") @@ -4980,6 +5448,7 @@ class Program_weight_tensor_parameter_467: class Program_weight_tensor_parameter_468: name = "parameter_468" + original_name = "batch_norm2d_3.w_2" shape = [80] dtype = "float32" min_val = float("0.000309091") @@ -4991,6 +5460,7 @@ class Program_weight_tensor_parameter_468: class Program_weight_tensor_parameter_469: name = "parameter_469" + original_name = "batch_norm2d_3.w_1" shape = [80] dtype = "float32" min_val = float("-3.61962") @@ -5002,6 +5472,7 @@ class Program_weight_tensor_parameter_469: class Program_weight_tensor_parameter_470: name = "parameter_470" + original_name = "conv2d_3.w_0" shape = [80, 1, 3, 3] dtype = "float32" min_val = float("-0.486495") @@ -5013,6 +5484,7 @@ class Program_weight_tensor_parameter_470: class Program_weight_tensor_parameter_471: name = "parameter_471" + original_name = "batch_norm2d_2.b_0" shape = [80] dtype = "float32" min_val = float("-2.97708") @@ -5024,6 +5496,7 @@ class Program_weight_tensor_parameter_471: class Program_weight_tensor_parameter_472: name = "parameter_472" + original_name = "batch_norm2d_2.w_0" shape = [80] dtype = "float32" min_val = float("-1.5869") @@ -5035,6 +5508,7 @@ class Program_weight_tensor_parameter_472: class Program_weight_tensor_parameter_473: name = "parameter_473" + original_name = "batch_norm2d_2.w_2" shape = [80] dtype = "float32" min_val = float("0.614783") @@ -5046,6 +5520,7 @@ class Program_weight_tensor_parameter_473: class Program_weight_tensor_parameter_474: name = "parameter_474" + original_name = "batch_norm2d_2.w_1" shape = [80] dtype = "float32" min_val = float("-7.66754") @@ -5057,6 +5532,7 @@ class Program_weight_tensor_parameter_474: class Program_weight_tensor_parameter_475: name = "parameter_475" + original_name = "conv2d_2.w_0" shape = [80, 40, 1, 1] dtype = "float32" min_val = float("-1.16253") @@ -5068,6 +5544,7 @@ class Program_weight_tensor_parameter_475: class Program_weight_tensor_parameter_476: name = "parameter_476" + original_name = "batch_norm2d_1.b_0" shape = [40] dtype = "float32" min_val = float("0") @@ -5077,6 +5554,7 @@ class Program_weight_tensor_parameter_476: class Program_weight_tensor_parameter_477: name = "parameter_477" + original_name = "batch_norm2d_1.w_0" shape = [40] dtype = "float32" min_val = float("0") @@ -5086,6 +5564,7 @@ class Program_weight_tensor_parameter_477: class Program_weight_tensor_parameter_478: name = "parameter_478" + original_name = "batch_norm2d_1.w_2" shape = [40] dtype = "float32" min_val = float("0") @@ -5095,6 +5574,7 @@ class Program_weight_tensor_parameter_478: class Program_weight_tensor_parameter_479: name = "parameter_479" + original_name = "batch_norm2d_1.w_1" shape = [40] dtype = "float32" min_val = float("0") @@ -5104,6 +5584,7 @@ class Program_weight_tensor_parameter_479: class Program_weight_tensor_parameter_480: name = "parameter_480" + original_name = "conv2d_1.w_0" shape = [40, 1, 3, 3] dtype = "float32" min_val = float("-1.9714") @@ -5115,6 +5596,7 @@ class Program_weight_tensor_parameter_480: class Program_weight_tensor_parameter_481: name = "parameter_481" + original_name = "batch_norm2d_0.b_0" shape = [40] dtype = "float32" min_val = float("0") @@ -5124,6 +5606,7 @@ class Program_weight_tensor_parameter_481: class Program_weight_tensor_parameter_482: name = "parameter_482" + original_name = "batch_norm2d_0.w_0" shape = [40] dtype = "float32" min_val = float("0") @@ -5133,6 +5616,7 @@ class Program_weight_tensor_parameter_482: class Program_weight_tensor_parameter_483: name = "parameter_483" + original_name = "batch_norm2d_0.w_2" shape = [40] dtype = "float32" min_val = float("0") @@ -5142,6 +5626,7 @@ class Program_weight_tensor_parameter_483: class Program_weight_tensor_parameter_484: name = "parameter_484" + original_name = "batch_norm2d_0.w_1" shape = [40] dtype = "float32" min_val = float("0") @@ -5151,6 +5636,7 @@ class Program_weight_tensor_parameter_484: class Program_weight_tensor_parameter_485: name = "parameter_485" + original_name = "conv2d_0.w_0" shape = [40, 3, 3, 3] dtype = "float32" min_val = float("-0.551297") diff --git a/paddle_samples/PaddleX/PP-ShiTuV2_rec/subgraph_1/input_meta.py b/paddle_samples/PaddleX/PP-ShiTuV2_rec/subgraph_1/input_meta.py index 3624b94dc..88bd32f7d 100644 --- a/paddle_samples/PaddleX/PP-ShiTuV2_rec/subgraph_1/input_meta.py +++ b/paddle_samples/PaddleX/PP-ShiTuV2_rec/subgraph_1/input_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_data_0: name = "data_0" + original_name = "var_0" shape = [128, 3, 224, 224] dtype = "float32" min_val = float("-2.1179") diff --git a/paddle_samples/PaddleX/PP-ShiTuV2_rec/subgraph_1/weight_meta.py b/paddle_samples/PaddleX/PP-ShiTuV2_rec/subgraph_1/weight_meta.py index a76fa5b69..07df28660 100644 --- a/paddle_samples/PaddleX/PP-ShiTuV2_rec/subgraph_1/weight_meta.py +++ b/paddle_samples/PaddleX/PP-ShiTuV2_rec/subgraph_1/weight_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_parameter_0: name = "parameter_0" + original_name = "linear_1.w_0" shape = [512, 159] dtype = "float32" min_val = float("-0.004888") @@ -11,6 +12,7 @@ class Program_weight_tensor_parameter_0: class Program_weight_tensor_parameter_1: name = "parameter_1" + original_name = "batch_norm1d_0.b_0" shape = [512] dtype = "float32" min_val = float("-1.15832e-19") @@ -22,6 +24,7 @@ class Program_weight_tensor_parameter_1: class Program_weight_tensor_parameter_2: name = "parameter_2" + original_name = "batch_norm1d_0.w_0" shape = [512] dtype = "float32" min_val = float("2.69551") @@ -33,6 +36,7 @@ class Program_weight_tensor_parameter_2: class Program_weight_tensor_parameter_3: name = "parameter_3" + original_name = "batch_norm1d_0.w_2" shape = [512] dtype = "float32" min_val = float("0.276074") @@ -44,6 +48,7 @@ class Program_weight_tensor_parameter_3: class Program_weight_tensor_parameter_4: name = "parameter_4" + original_name = "batch_norm1d_0.w_1" shape = [512] dtype = "float32" min_val = float("-0.0428911") @@ -55,6 +60,7 @@ class Program_weight_tensor_parameter_4: class Program_weight_tensor_parameter_5: name = "parameter_5" + original_name = "conv2d_68.b_0" shape = [512] dtype = "float32" min_val = float("-0.167411") @@ -66,6 +72,7 @@ class Program_weight_tensor_parameter_5: class Program_weight_tensor_parameter_6: name = "parameter_6" + original_name = "conv2d_68.w_0" shape = [512, 1024, 1, 1] dtype = "float32" min_val = float("-0.405835") @@ -77,6 +84,7 @@ class Program_weight_tensor_parameter_6: class Program_weight_tensor_parameter_7: name = "parameter_7" + original_name = "batch_norm2d_54.b_0" shape = [1024] dtype = "float32" min_val = float("-12.8937") @@ -88,6 +96,7 @@ class Program_weight_tensor_parameter_7: class Program_weight_tensor_parameter_8: name = "parameter_8" + original_name = "batch_norm2d_54.w_0" shape = [1024] dtype = "float32" min_val = float("1.03265") @@ -99,6 +108,7 @@ class Program_weight_tensor_parameter_8: class Program_weight_tensor_parameter_9: name = "parameter_9" + original_name = "batch_norm2d_54.w_2" shape = [1024] dtype = "float32" min_val = float("5.06125") @@ -110,6 +120,7 @@ class Program_weight_tensor_parameter_9: class Program_weight_tensor_parameter_10: name = "parameter_10" + original_name = "batch_norm2d_54.w_1" shape = [1024] dtype = "float32" min_val = float("-30.3748") @@ -121,6 +132,7 @@ class Program_weight_tensor_parameter_10: class Program_weight_tensor_parameter_11: name = "parameter_11" + original_name = "conv2d_78.w_0" shape = [1024, 1024, 1, 1] dtype = "float32" min_val = float("-0.866283") @@ -132,6 +144,7 @@ class Program_weight_tensor_parameter_11: class Program_weight_tensor_parameter_12: name = "parameter_12" + original_name = "batch_norm2d_53.b_0" shape = [1024] dtype = "float32" min_val = float("-2.446") @@ -143,6 +156,7 @@ class Program_weight_tensor_parameter_12: class Program_weight_tensor_parameter_13: name = "parameter_13" + original_name = "batch_norm2d_53.w_0" shape = [1024] dtype = "float32" min_val = float("-2.19304") @@ -154,6 +168,7 @@ class Program_weight_tensor_parameter_13: class Program_weight_tensor_parameter_14: name = "parameter_14" + original_name = "batch_norm2d_53.w_2" shape = [1024] dtype = "float32" min_val = float("5.60519e-45") @@ -165,6 +180,7 @@ class Program_weight_tensor_parameter_14: class Program_weight_tensor_parameter_15: name = "parameter_15" + original_name = "batch_norm2d_53.w_1" shape = [1024] dtype = "float32" min_val = float("-0.0141175") @@ -176,6 +192,7 @@ class Program_weight_tensor_parameter_15: class Program_weight_tensor_parameter_16: name = "parameter_16" + original_name = "conv2d_76.w_0" shape = [1024, 1, 1, 1] dtype = "float32" min_val = float("-0.1532") @@ -187,6 +204,7 @@ class Program_weight_tensor_parameter_16: class Program_weight_tensor_parameter_17: name = "parameter_17" + original_name = "batch_norm2d_52.b_0" shape = [1024] dtype = "float32" min_val = float("-2.446") @@ -198,6 +216,7 @@ class Program_weight_tensor_parameter_17: class Program_weight_tensor_parameter_18: name = "parameter_18" + original_name = "batch_norm2d_52.w_0" shape = [1024] dtype = "float32" min_val = float("-0.273242") @@ -209,6 +228,7 @@ class Program_weight_tensor_parameter_18: class Program_weight_tensor_parameter_19: name = "parameter_19" + original_name = "batch_norm2d_52.w_2" shape = [1024] dtype = "float32" min_val = float("5.60519e-45") @@ -220,6 +240,7 @@ class Program_weight_tensor_parameter_19: class Program_weight_tensor_parameter_20: name = "parameter_20" + original_name = "batch_norm2d_52.w_1" shape = [1024] dtype = "float32" min_val = float("-0.139547") @@ -231,6 +252,7 @@ class Program_weight_tensor_parameter_20: class Program_weight_tensor_parameter_21: name = "parameter_21" + original_name = "conv2d_75.w_0" shape = [1024, 1, 3, 3] dtype = "float32" min_val = float("-0.241024") @@ -242,6 +264,7 @@ class Program_weight_tensor_parameter_21: class Program_weight_tensor_parameter_22: name = "parameter_22" + original_name = "batch_norm2d_51.b_0" shape = [1024] dtype = "float32" min_val = float("-2.446") @@ -253,6 +276,7 @@ class Program_weight_tensor_parameter_22: class Program_weight_tensor_parameter_23: name = "parameter_23" + original_name = "batch_norm2d_51.w_0" shape = [1024] dtype = "float32" min_val = float("-0.666614") @@ -264,6 +288,7 @@ class Program_weight_tensor_parameter_23: class Program_weight_tensor_parameter_24: name = "parameter_24" + original_name = "batch_norm2d_51.w_2" shape = [1024] dtype = "float32" min_val = float("5.60519e-45") @@ -275,6 +300,7 @@ class Program_weight_tensor_parameter_24: class Program_weight_tensor_parameter_25: name = "parameter_25" + original_name = "batch_norm2d_51.w_1" shape = [1024] dtype = "float32" min_val = float("-0.540386") @@ -286,6 +312,7 @@ class Program_weight_tensor_parameter_25: class Program_weight_tensor_parameter_26: name = "parameter_26" + original_name = "conv2d_74.w_0" shape = [1024, 1, 5, 5] dtype = "float32" min_val = float("-0.288186") @@ -297,6 +324,7 @@ class Program_weight_tensor_parameter_26: class Program_weight_tensor_parameter_27: name = "parameter_27" + original_name = "batch_norm2d_50.b_0" shape = [1024] dtype = "float32" min_val = float("-6.73249") @@ -308,6 +336,7 @@ class Program_weight_tensor_parameter_27: class Program_weight_tensor_parameter_28: name = "parameter_28" + original_name = "batch_norm2d_50.w_0" shape = [1024] dtype = "float32" min_val = float("-0.338143") @@ -319,6 +348,7 @@ class Program_weight_tensor_parameter_28: class Program_weight_tensor_parameter_29: name = "parameter_29" + original_name = "batch_norm2d_50.w_2" shape = [1024] dtype = "float32" min_val = float("5.60519e-45") @@ -330,6 +360,7 @@ class Program_weight_tensor_parameter_29: class Program_weight_tensor_parameter_30: name = "parameter_30" + original_name = "batch_norm2d_50.w_1" shape = [1024] dtype = "float32" min_val = float("-21.4202") @@ -341,6 +372,7 @@ class Program_weight_tensor_parameter_30: class Program_weight_tensor_parameter_31: name = "parameter_31" + original_name = "conv2d_73.w_0" shape = [1024, 512, 1, 1] dtype = "float32" min_val = float("-1.20007") @@ -352,6 +384,7 @@ class Program_weight_tensor_parameter_31: class Program_weight_tensor_parameter_32: name = "parameter_32" + original_name = "batch_norm2d_49.b_0" shape = [512] dtype = "float32" min_val = float("-0.223522") @@ -363,6 +396,7 @@ class Program_weight_tensor_parameter_32: class Program_weight_tensor_parameter_33: name = "parameter_33" + original_name = "batch_norm2d_49.w_0" shape = [512] dtype = "float32" min_val = float("0.356103") @@ -374,6 +408,7 @@ class Program_weight_tensor_parameter_33: class Program_weight_tensor_parameter_34: name = "parameter_34" + original_name = "batch_norm2d_49.w_2" shape = [512] dtype = "float32" min_val = float("0.000108972") @@ -385,6 +420,7 @@ class Program_weight_tensor_parameter_34: class Program_weight_tensor_parameter_35: name = "parameter_35" + original_name = "batch_norm2d_49.w_1" shape = [512] dtype = "float32" min_val = float("-1.70051") @@ -396,6 +432,7 @@ class Program_weight_tensor_parameter_35: class Program_weight_tensor_parameter_36: name = "parameter_36" + original_name = "conv2d_71.w_0" shape = [512, 1, 1, 1] dtype = "float32" min_val = float("-1.4111") @@ -407,6 +444,7 @@ class Program_weight_tensor_parameter_36: class Program_weight_tensor_parameter_37: name = "parameter_37" + original_name = "batch_norm2d_48.b_0" shape = [512] dtype = "float32" min_val = float("-4.88145") @@ -418,6 +456,7 @@ class Program_weight_tensor_parameter_37: class Program_weight_tensor_parameter_38: name = "parameter_38" + original_name = "batch_norm2d_48.w_0" shape = [512] dtype = "float32" min_val = float("-1.28777") @@ -429,6 +468,7 @@ class Program_weight_tensor_parameter_38: class Program_weight_tensor_parameter_39: name = "parameter_39" + original_name = "batch_norm2d_48.w_2" shape = [512] dtype = "float32" min_val = float("0.000104172") @@ -440,6 +480,7 @@ class Program_weight_tensor_parameter_39: class Program_weight_tensor_parameter_40: name = "parameter_40" + original_name = "batch_norm2d_48.w_1" shape = [512] dtype = "float32" min_val = float("-1.87717") @@ -451,6 +492,7 @@ class Program_weight_tensor_parameter_40: class Program_weight_tensor_parameter_41: name = "parameter_41" + original_name = "conv2d_70.w_0" shape = [512, 1, 3, 3] dtype = "float32" min_val = float("-0.423681") @@ -462,6 +504,7 @@ class Program_weight_tensor_parameter_41: class Program_weight_tensor_parameter_42: name = "parameter_42" + original_name = "batch_norm2d_47.b_0" shape = [512] dtype = "float32" min_val = float("-4.88145") @@ -473,6 +516,7 @@ class Program_weight_tensor_parameter_42: class Program_weight_tensor_parameter_43: name = "parameter_43" + original_name = "batch_norm2d_47.w_0" shape = [512] dtype = "float32" min_val = float("0.163248") @@ -484,6 +528,7 @@ class Program_weight_tensor_parameter_43: class Program_weight_tensor_parameter_44: name = "parameter_44" + original_name = "batch_norm2d_47.w_2" shape = [512] dtype = "float32" min_val = float("0.012105") @@ -495,6 +540,7 @@ class Program_weight_tensor_parameter_44: class Program_weight_tensor_parameter_45: name = "parameter_45" + original_name = "batch_norm2d_47.w_1" shape = [512] dtype = "float32" min_val = float("-6.17369") @@ -506,6 +552,7 @@ class Program_weight_tensor_parameter_45: class Program_weight_tensor_parameter_46: name = "parameter_46" + original_name = "conv2d_69.w_0" shape = [512, 1, 5, 5] dtype = "float32" min_val = float("-0.351437") @@ -517,6 +564,7 @@ class Program_weight_tensor_parameter_46: class Program_weight_tensor_parameter_47: name = "parameter_47" + original_name = "batch_norm2d_39.b_0" shape = [512] dtype = "float32" min_val = float("-2.77047") @@ -528,6 +576,7 @@ class Program_weight_tensor_parameter_47: class Program_weight_tensor_parameter_48: name = "parameter_48" + original_name = "batch_norm2d_39.w_0" shape = [512] dtype = "float32" min_val = float("0.462261") @@ -539,6 +588,7 @@ class Program_weight_tensor_parameter_48: class Program_weight_tensor_parameter_49: name = "parameter_49" + original_name = "batch_norm2d_39.w_2" shape = [512] dtype = "float32" min_val = float("2.34792") @@ -550,6 +600,7 @@ class Program_weight_tensor_parameter_49: class Program_weight_tensor_parameter_50: name = "parameter_50" + original_name = "batch_norm2d_39.w_1" shape = [512] dtype = "float32" min_val = float("-5.17683") @@ -561,6 +612,7 @@ class Program_weight_tensor_parameter_50: class Program_weight_tensor_parameter_51: name = "parameter_51" + original_name = "conv2d_57.w_0" shape = [512, 256, 1, 1] dtype = "float32" min_val = float("-0.855038") @@ -572,6 +624,7 @@ class Program_weight_tensor_parameter_51: class Program_weight_tensor_parameter_52: name = "parameter_52" + original_name = "batch_norm2d_38.b_0" shape = [256] dtype = "float32" min_val = float("-5.02305") @@ -583,6 +636,7 @@ class Program_weight_tensor_parameter_52: class Program_weight_tensor_parameter_53: name = "parameter_53" + original_name = "batch_norm2d_38.w_0" shape = [256] dtype = "float32" min_val = float("1.75157") @@ -594,6 +648,7 @@ class Program_weight_tensor_parameter_53: class Program_weight_tensor_parameter_54: name = "parameter_54" + original_name = "batch_norm2d_38.w_2" shape = [256] dtype = "float32" min_val = float("1.04221") @@ -605,6 +660,7 @@ class Program_weight_tensor_parameter_54: class Program_weight_tensor_parameter_55: name = "parameter_55" + original_name = "batch_norm2d_38.w_1" shape = [256] dtype = "float32" min_val = float("-2.26988") @@ -616,6 +672,7 @@ class Program_weight_tensor_parameter_55: class Program_weight_tensor_parameter_56: name = "parameter_56" + original_name = "conv2d_56.w_0" shape = [256, 512, 1, 1] dtype = "float32" min_val = float("-1.30354") @@ -627,6 +684,7 @@ class Program_weight_tensor_parameter_56: class Program_weight_tensor_parameter_57: name = "parameter_57" + original_name = "conv2d_55.b_0" shape = [512] dtype = "float32" min_val = float("-0.14264") @@ -638,6 +696,7 @@ class Program_weight_tensor_parameter_57: class Program_weight_tensor_parameter_58: name = "parameter_58" + original_name = "conv2d_55.w_0" shape = [512, 128, 1, 1] dtype = "float32" min_val = float("-0.690847") @@ -649,6 +708,7 @@ class Program_weight_tensor_parameter_58: class Program_weight_tensor_parameter_59: name = "parameter_59" + original_name = "conv2d_54.b_0" shape = [128] dtype = "float32" min_val = float("-0.0472658") @@ -660,6 +720,7 @@ class Program_weight_tensor_parameter_59: class Program_weight_tensor_parameter_60: name = "parameter_60" + original_name = "conv2d_54.w_0" shape = [128, 512, 1, 1] dtype = "float32" min_val = float("-0.77497") @@ -671,6 +732,7 @@ class Program_weight_tensor_parameter_60: class Program_weight_tensor_parameter_61: name = "parameter_61" + original_name = "batch_norm2d_37.b_0" shape = [512] dtype = "float32" min_val = float("-2.78947") @@ -682,6 +744,7 @@ class Program_weight_tensor_parameter_61: class Program_weight_tensor_parameter_62: name = "parameter_62" + original_name = "batch_norm2d_37.w_0" shape = [512] dtype = "float32" min_val = float("-2.77708") @@ -693,6 +756,7 @@ class Program_weight_tensor_parameter_62: class Program_weight_tensor_parameter_63: name = "parameter_63" + original_name = "batch_norm2d_37.w_2" shape = [512] dtype = "float32" min_val = float("8.48197e-17") @@ -704,6 +768,7 @@ class Program_weight_tensor_parameter_63: class Program_weight_tensor_parameter_64: name = "parameter_64" + original_name = "batch_norm2d_37.w_1" shape = [512] dtype = "float32" min_val = float("-0.20729") @@ -715,6 +780,7 @@ class Program_weight_tensor_parameter_64: class Program_weight_tensor_parameter_65: name = "parameter_65" + original_name = "conv2d_52.w_0" shape = [512, 1, 1, 1] dtype = "float32" min_val = float("-0.116203") @@ -726,6 +792,7 @@ class Program_weight_tensor_parameter_65: class Program_weight_tensor_parameter_66: name = "parameter_66" + original_name = "batch_norm2d_36.b_0" shape = [512] dtype = "float32" min_val = float("-2.78947") @@ -737,6 +804,7 @@ class Program_weight_tensor_parameter_66: class Program_weight_tensor_parameter_67: name = "parameter_67" + original_name = "batch_norm2d_36.w_0" shape = [512] dtype = "float32" min_val = float("-1.27567") @@ -748,6 +816,7 @@ class Program_weight_tensor_parameter_67: class Program_weight_tensor_parameter_68: name = "parameter_68" + original_name = "batch_norm2d_36.w_2" shape = [512] dtype = "float32" min_val = float("2.01505e-05") @@ -759,6 +828,7 @@ class Program_weight_tensor_parameter_68: class Program_weight_tensor_parameter_69: name = "parameter_69" + original_name = "batch_norm2d_36.w_1" shape = [512] dtype = "float32" min_val = float("-2.4575") @@ -770,6 +840,7 @@ class Program_weight_tensor_parameter_69: class Program_weight_tensor_parameter_70: name = "parameter_70" + original_name = "conv2d_51.w_0" shape = [512, 1, 3, 3] dtype = "float32" min_val = float("-0.290013") @@ -781,6 +852,7 @@ class Program_weight_tensor_parameter_70: class Program_weight_tensor_parameter_71: name = "parameter_71" + original_name = "batch_norm2d_35.b_0" shape = [512] dtype = "float32" min_val = float("-2.78947") @@ -792,6 +864,7 @@ class Program_weight_tensor_parameter_71: class Program_weight_tensor_parameter_72: name = "parameter_72" + original_name = "batch_norm2d_35.w_0" shape = [512] dtype = "float32" min_val = float("-0.706791") @@ -803,6 +876,7 @@ class Program_weight_tensor_parameter_72: class Program_weight_tensor_parameter_73: name = "parameter_73" + original_name = "batch_norm2d_35.w_2" shape = [512] dtype = "float32" min_val = float("0.000352087") @@ -814,6 +888,7 @@ class Program_weight_tensor_parameter_73: class Program_weight_tensor_parameter_74: name = "parameter_74" + original_name = "batch_norm2d_35.w_1" shape = [512] dtype = "float32" min_val = float("-4.75424") @@ -825,6 +900,7 @@ class Program_weight_tensor_parameter_74: class Program_weight_tensor_parameter_75: name = "parameter_75" + original_name = "conv2d_50.w_0" shape = [512, 1, 5, 5] dtype = "float32" min_val = float("-0.295999") @@ -836,6 +912,7 @@ class Program_weight_tensor_parameter_75: class Program_weight_tensor_parameter_76: name = "parameter_76" + original_name = "batch_norm2d_34.b_0" shape = [512] dtype = "float32" min_val = float("-3.08347") @@ -847,6 +924,7 @@ class Program_weight_tensor_parameter_76: class Program_weight_tensor_parameter_77: name = "parameter_77" + original_name = "batch_norm2d_34.w_0" shape = [512] dtype = "float32" min_val = float("0.134382") @@ -858,6 +936,7 @@ class Program_weight_tensor_parameter_77: class Program_weight_tensor_parameter_78: name = "parameter_78" + original_name = "batch_norm2d_34.w_2" shape = [512] dtype = "float32" min_val = float("1.04379") @@ -869,6 +948,7 @@ class Program_weight_tensor_parameter_78: class Program_weight_tensor_parameter_79: name = "parameter_79" + original_name = "batch_norm2d_34.w_1" shape = [512] dtype = "float32" min_val = float("-10.5117") @@ -880,6 +960,7 @@ class Program_weight_tensor_parameter_79: class Program_weight_tensor_parameter_80: name = "parameter_80" + original_name = "conv2d_49.w_0" shape = [512, 256, 1, 1] dtype = "float32" min_val = float("-1.31049") @@ -891,6 +972,7 @@ class Program_weight_tensor_parameter_80: class Program_weight_tensor_parameter_81: name = "parameter_81" + original_name = "batch_norm2d_33.b_0" shape = [256] dtype = "float32" min_val = float("-7.04922") @@ -902,6 +984,7 @@ class Program_weight_tensor_parameter_81: class Program_weight_tensor_parameter_82: name = "parameter_82" + original_name = "batch_norm2d_33.w_0" shape = [256] dtype = "float32" min_val = float("1.22799") @@ -913,6 +996,7 @@ class Program_weight_tensor_parameter_82: class Program_weight_tensor_parameter_83: name = "parameter_83" + original_name = "batch_norm2d_33.w_2" shape = [256] dtype = "float32" min_val = float("0.74225") @@ -924,6 +1008,7 @@ class Program_weight_tensor_parameter_83: class Program_weight_tensor_parameter_84: name = "parameter_84" + original_name = "batch_norm2d_33.w_1" shape = [256] dtype = "float32" min_val = float("-3.39829") @@ -935,6 +1020,7 @@ class Program_weight_tensor_parameter_84: class Program_weight_tensor_parameter_85: name = "parameter_85" + original_name = "conv2d_48.w_0" shape = [256, 512, 1, 1] dtype = "float32" min_val = float("-1.45511") @@ -946,6 +1032,7 @@ class Program_weight_tensor_parameter_85: class Program_weight_tensor_parameter_86: name = "parameter_86" + original_name = "conv2d_47.b_0" shape = [512] dtype = "float32" min_val = float("-0.150837") @@ -957,6 +1044,7 @@ class Program_weight_tensor_parameter_86: class Program_weight_tensor_parameter_87: name = "parameter_87" + original_name = "conv2d_47.w_0" shape = [512, 128, 1, 1] dtype = "float32" min_val = float("-1.04837") @@ -968,6 +1056,7 @@ class Program_weight_tensor_parameter_87: class Program_weight_tensor_parameter_88: name = "parameter_88" + original_name = "conv2d_46.b_0" shape = [128] dtype = "float32" min_val = float("-0.0160665") @@ -979,6 +1068,7 @@ class Program_weight_tensor_parameter_88: class Program_weight_tensor_parameter_89: name = "parameter_89" + original_name = "conv2d_46.w_0" shape = [128, 512, 1, 1] dtype = "float32" min_val = float("-1.10044") @@ -990,6 +1080,7 @@ class Program_weight_tensor_parameter_89: class Program_weight_tensor_parameter_90: name = "parameter_90" + original_name = "batch_norm2d_32.b_0" shape = [512] dtype = "float32" min_val = float("-2.56407") @@ -1001,6 +1092,7 @@ class Program_weight_tensor_parameter_90: class Program_weight_tensor_parameter_91: name = "parameter_91" + original_name = "batch_norm2d_32.w_0" shape = [512] dtype = "float32" min_val = float("-3.02065") @@ -1012,6 +1104,7 @@ class Program_weight_tensor_parameter_91: class Program_weight_tensor_parameter_92: name = "parameter_92" + original_name = "batch_norm2d_32.w_2" shape = [512] dtype = "float32" min_val = float("4.06844e-17") @@ -1023,6 +1116,7 @@ class Program_weight_tensor_parameter_92: class Program_weight_tensor_parameter_93: name = "parameter_93" + original_name = "batch_norm2d_32.w_1" shape = [512] dtype = "float32" min_val = float("-0.256646") @@ -1034,6 +1128,7 @@ class Program_weight_tensor_parameter_93: class Program_weight_tensor_parameter_94: name = "parameter_94" + original_name = "conv2d_44.w_0" shape = [512, 1, 1, 1] dtype = "float32" min_val = float("-0.0973175") @@ -1045,6 +1140,7 @@ class Program_weight_tensor_parameter_94: class Program_weight_tensor_parameter_95: name = "parameter_95" + original_name = "batch_norm2d_31.b_0" shape = [512] dtype = "float32" min_val = float("-2.56407") @@ -1056,6 +1152,7 @@ class Program_weight_tensor_parameter_95: class Program_weight_tensor_parameter_96: name = "parameter_96" + original_name = "batch_norm2d_31.w_0" shape = [512] dtype = "float32" min_val = float("-1.51512") @@ -1067,6 +1164,7 @@ class Program_weight_tensor_parameter_96: class Program_weight_tensor_parameter_97: name = "parameter_97" + original_name = "batch_norm2d_31.w_2" shape = [512] dtype = "float32" min_val = float("1.45191e-05") @@ -1078,6 +1176,7 @@ class Program_weight_tensor_parameter_97: class Program_weight_tensor_parameter_98: name = "parameter_98" + original_name = "batch_norm2d_31.w_1" shape = [512] dtype = "float32" min_val = float("-2.04283") @@ -1089,6 +1188,7 @@ class Program_weight_tensor_parameter_98: class Program_weight_tensor_parameter_99: name = "parameter_99" + original_name = "conv2d_43.w_0" shape = [512, 1, 3, 3] dtype = "float32" min_val = float("-0.237834") @@ -1100,6 +1200,7 @@ class Program_weight_tensor_parameter_99: class Program_weight_tensor_parameter_100: name = "parameter_100" + original_name = "batch_norm2d_30.b_0" shape = [512] dtype = "float32" min_val = float("-2.56407") @@ -1111,6 +1212,7 @@ class Program_weight_tensor_parameter_100: class Program_weight_tensor_parameter_101: name = "parameter_101" + original_name = "batch_norm2d_30.w_0" shape = [512] dtype = "float32" min_val = float("-1.01478") @@ -1122,6 +1224,7 @@ class Program_weight_tensor_parameter_101: class Program_weight_tensor_parameter_102: name = "parameter_102" + original_name = "batch_norm2d_30.w_2" shape = [512] dtype = "float32" min_val = float("0.00094865") @@ -1133,6 +1236,7 @@ class Program_weight_tensor_parameter_102: class Program_weight_tensor_parameter_103: name = "parameter_103" + original_name = "batch_norm2d_30.w_1" shape = [512] dtype = "float32" min_val = float("-4.03439") @@ -1144,6 +1248,7 @@ class Program_weight_tensor_parameter_103: class Program_weight_tensor_parameter_104: name = "parameter_104" + original_name = "conv2d_42.w_0" shape = [512, 1, 5, 5] dtype = "float32" min_val = float("-0.340974") @@ -1155,6 +1260,7 @@ class Program_weight_tensor_parameter_104: class Program_weight_tensor_parameter_105: name = "parameter_105" + original_name = "batch_norm2d_29.b_0" shape = [512] dtype = "float32" min_val = float("-2.71515") @@ -1166,6 +1272,7 @@ class Program_weight_tensor_parameter_105: class Program_weight_tensor_parameter_106: name = "parameter_106" + original_name = "batch_norm2d_29.w_0" shape = [512] dtype = "float32" min_val = float("0.174368") @@ -1177,6 +1284,7 @@ class Program_weight_tensor_parameter_106: class Program_weight_tensor_parameter_107: name = "parameter_107" + original_name = "batch_norm2d_29.w_2" shape = [512] dtype = "float32" min_val = float("1.09308") @@ -1188,6 +1296,7 @@ class Program_weight_tensor_parameter_107: class Program_weight_tensor_parameter_108: name = "parameter_108" + original_name = "batch_norm2d_29.w_1" shape = [512] dtype = "float32" min_val = float("-9.22105") @@ -1199,6 +1308,7 @@ class Program_weight_tensor_parameter_108: class Program_weight_tensor_parameter_109: name = "parameter_109" + original_name = "conv2d_41.w_0" shape = [512, 256, 1, 1] dtype = "float32" min_val = float("-1.47501") @@ -1210,6 +1320,7 @@ class Program_weight_tensor_parameter_109: class Program_weight_tensor_parameter_110: name = "parameter_110" + original_name = "batch_norm2d_28.b_0" shape = [256] dtype = "float32" min_val = float("-6.27082") @@ -1221,6 +1332,7 @@ class Program_weight_tensor_parameter_110: class Program_weight_tensor_parameter_111: name = "parameter_111" + original_name = "batch_norm2d_28.w_0" shape = [256] dtype = "float32" min_val = float("1.27357") @@ -1232,6 +1344,7 @@ class Program_weight_tensor_parameter_111: class Program_weight_tensor_parameter_112: name = "parameter_112" + original_name = "batch_norm2d_28.w_2" shape = [256] dtype = "float32" min_val = float("0.529698") @@ -1243,6 +1356,7 @@ class Program_weight_tensor_parameter_112: class Program_weight_tensor_parameter_113: name = "parameter_113" + original_name = "batch_norm2d_28.w_1" shape = [256] dtype = "float32" min_val = float("-3.07509") @@ -1254,6 +1368,7 @@ class Program_weight_tensor_parameter_113: class Program_weight_tensor_parameter_114: name = "parameter_114" + original_name = "conv2d_40.w_0" shape = [256, 512, 1, 1] dtype = "float32" min_val = float("-1.15099") @@ -1265,6 +1380,7 @@ class Program_weight_tensor_parameter_114: class Program_weight_tensor_parameter_115: name = "parameter_115" + original_name = "conv2d_39.b_0" shape = [512] dtype = "float32" min_val = float("-0.104581") @@ -1276,6 +1392,7 @@ class Program_weight_tensor_parameter_115: class Program_weight_tensor_parameter_116: name = "parameter_116" + original_name = "conv2d_39.w_0" shape = [512, 128, 1, 1] dtype = "float32" min_val = float("-0.592596") @@ -1287,6 +1404,7 @@ class Program_weight_tensor_parameter_116: class Program_weight_tensor_parameter_117: name = "parameter_117" + original_name = "conv2d_38.b_0" shape = [128] dtype = "float32" min_val = float("-0.00776211") @@ -1298,6 +1416,7 @@ class Program_weight_tensor_parameter_117: class Program_weight_tensor_parameter_118: name = "parameter_118" + original_name = "conv2d_38.w_0" shape = [128, 512, 1, 1] dtype = "float32" min_val = float("-0.555076") @@ -1309,6 +1428,7 @@ class Program_weight_tensor_parameter_118: class Program_weight_tensor_parameter_119: name = "parameter_119" + original_name = "batch_norm2d_27.b_0" shape = [512] dtype = "float32" min_val = float("-4.50152") @@ -1320,6 +1440,7 @@ class Program_weight_tensor_parameter_119: class Program_weight_tensor_parameter_120: name = "parameter_120" + original_name = "batch_norm2d_27.w_0" shape = [512] dtype = "float32" min_val = float("-3.29262") @@ -1331,6 +1452,7 @@ class Program_weight_tensor_parameter_120: class Program_weight_tensor_parameter_121: name = "parameter_121" + original_name = "batch_norm2d_27.w_2" shape = [512] dtype = "float32" min_val = float("1.94239e-18") @@ -1342,6 +1464,7 @@ class Program_weight_tensor_parameter_121: class Program_weight_tensor_parameter_122: name = "parameter_122" + original_name = "batch_norm2d_27.w_1" shape = [512] dtype = "float32" min_val = float("-0.151156") @@ -1353,6 +1476,7 @@ class Program_weight_tensor_parameter_122: class Program_weight_tensor_parameter_123: name = "parameter_123" + original_name = "conv2d_36.w_0" shape = [512, 1, 1, 1] dtype = "float32" min_val = float("-0.120746") @@ -1364,6 +1488,7 @@ class Program_weight_tensor_parameter_123: class Program_weight_tensor_parameter_124: name = "parameter_124" + original_name = "batch_norm2d_26.b_0" shape = [512] dtype = "float32" min_val = float("-4.50152") @@ -1375,6 +1500,7 @@ class Program_weight_tensor_parameter_124: class Program_weight_tensor_parameter_125: name = "parameter_125" + original_name = "batch_norm2d_26.w_0" shape = [512] dtype = "float32" min_val = float("-0.96103") @@ -1386,6 +1512,7 @@ class Program_weight_tensor_parameter_125: class Program_weight_tensor_parameter_126: name = "parameter_126" + original_name = "batch_norm2d_26.w_2" shape = [512] dtype = "float32" min_val = float("2.41241e-06") @@ -1397,6 +1524,7 @@ class Program_weight_tensor_parameter_126: class Program_weight_tensor_parameter_127: name = "parameter_127" + original_name = "batch_norm2d_26.w_1" shape = [512] dtype = "float32" min_val = float("-1.69269") @@ -1408,6 +1536,7 @@ class Program_weight_tensor_parameter_127: class Program_weight_tensor_parameter_128: name = "parameter_128" + original_name = "conv2d_35.w_0" shape = [512, 1, 3, 3] dtype = "float32" min_val = float("-0.294231") @@ -1419,6 +1548,7 @@ class Program_weight_tensor_parameter_128: class Program_weight_tensor_parameter_129: name = "parameter_129" + original_name = "batch_norm2d_25.b_0" shape = [512] dtype = "float32" min_val = float("-4.50152") @@ -1430,6 +1560,7 @@ class Program_weight_tensor_parameter_129: class Program_weight_tensor_parameter_130: name = "parameter_130" + original_name = "batch_norm2d_25.w_0" shape = [512] dtype = "float32" min_val = float("-0.881879") @@ -1441,6 +1572,7 @@ class Program_weight_tensor_parameter_130: class Program_weight_tensor_parameter_131: name = "parameter_131" + original_name = "batch_norm2d_25.w_2" shape = [512] dtype = "float32" min_val = float("0.000863103") @@ -1452,6 +1584,7 @@ class Program_weight_tensor_parameter_131: class Program_weight_tensor_parameter_132: name = "parameter_132" + original_name = "batch_norm2d_25.w_1" shape = [512] dtype = "float32" min_val = float("-2.49672") @@ -1463,6 +1596,7 @@ class Program_weight_tensor_parameter_132: class Program_weight_tensor_parameter_133: name = "parameter_133" + original_name = "conv2d_34.w_0" shape = [512, 1, 5, 5] dtype = "float32" min_val = float("-0.307159") @@ -1474,6 +1608,7 @@ class Program_weight_tensor_parameter_133: class Program_weight_tensor_parameter_134: name = "parameter_134" + original_name = "batch_norm2d_24.b_0" shape = [512] dtype = "float32" min_val = float("-3.25301") @@ -1485,6 +1620,7 @@ class Program_weight_tensor_parameter_134: class Program_weight_tensor_parameter_135: name = "parameter_135" + original_name = "batch_norm2d_24.w_0" shape = [512] dtype = "float32" min_val = float("0.201117") @@ -1496,6 +1632,7 @@ class Program_weight_tensor_parameter_135: class Program_weight_tensor_parameter_136: name = "parameter_136" + original_name = "batch_norm2d_24.w_2" shape = [512] dtype = "float32" min_val = float("1.14387") @@ -1507,6 +1644,7 @@ class Program_weight_tensor_parameter_136: class Program_weight_tensor_parameter_137: name = "parameter_137" + original_name = "batch_norm2d_24.w_1" shape = [512] dtype = "float32" min_val = float("-5.85754") @@ -1518,6 +1656,7 @@ class Program_weight_tensor_parameter_137: class Program_weight_tensor_parameter_138: name = "parameter_138" + original_name = "conv2d_33.w_0" shape = [512, 256, 1, 1] dtype = "float32" min_val = float("-1.60008") @@ -1529,6 +1668,7 @@ class Program_weight_tensor_parameter_138: class Program_weight_tensor_parameter_139: name = "parameter_139" + original_name = "batch_norm2d_23.b_0" shape = [256] dtype = "float32" min_val = float("-6.55761") @@ -1540,6 +1680,7 @@ class Program_weight_tensor_parameter_139: class Program_weight_tensor_parameter_140: name = "parameter_140" + original_name = "batch_norm2d_23.w_0" shape = [256] dtype = "float32" min_val = float("1.3049") @@ -1551,6 +1692,7 @@ class Program_weight_tensor_parameter_140: class Program_weight_tensor_parameter_141: name = "parameter_141" + original_name = "batch_norm2d_23.w_2" shape = [256] dtype = "float32" min_val = float("1.54461") @@ -1562,6 +1704,7 @@ class Program_weight_tensor_parameter_141: class Program_weight_tensor_parameter_142: name = "parameter_142" + original_name = "batch_norm2d_23.w_1" shape = [256] dtype = "float32" min_val = float("-3.76314") @@ -1573,6 +1716,7 @@ class Program_weight_tensor_parameter_142: class Program_weight_tensor_parameter_143: name = "parameter_143" + original_name = "conv2d_32.w_0" shape = [256, 512, 1, 1] dtype = "float32" min_val = float("-1.0672") @@ -1584,6 +1728,7 @@ class Program_weight_tensor_parameter_143: class Program_weight_tensor_parameter_144: name = "parameter_144" + original_name = "conv2d_31.b_0" shape = [512] dtype = "float32" min_val = float("-0.0926585") @@ -1595,6 +1740,7 @@ class Program_weight_tensor_parameter_144: class Program_weight_tensor_parameter_145: name = "parameter_145" + original_name = "conv2d_31.w_0" shape = [512, 128, 1, 1] dtype = "float32" min_val = float("-0.726161") @@ -1606,6 +1752,7 @@ class Program_weight_tensor_parameter_145: class Program_weight_tensor_parameter_146: name = "parameter_146" + original_name = "conv2d_30.b_0" shape = [128] dtype = "float32" min_val = float("-0.00882057") @@ -1617,6 +1764,7 @@ class Program_weight_tensor_parameter_146: class Program_weight_tensor_parameter_147: name = "parameter_147" + original_name = "conv2d_30.w_0" shape = [128, 512, 1, 1] dtype = "float32" min_val = float("-0.93542") @@ -1628,6 +1776,7 @@ class Program_weight_tensor_parameter_147: class Program_weight_tensor_parameter_148: name = "parameter_148" + original_name = "batch_norm2d_22.b_0" shape = [512] dtype = "float32" min_val = float("-1.82712") @@ -1639,6 +1788,7 @@ class Program_weight_tensor_parameter_148: class Program_weight_tensor_parameter_149: name = "parameter_149" + original_name = "batch_norm2d_22.w_0" shape = [512] dtype = "float32" min_val = float("-3.28826") @@ -1650,6 +1800,7 @@ class Program_weight_tensor_parameter_149: class Program_weight_tensor_parameter_150: name = "parameter_150" + original_name = "batch_norm2d_22.w_2" shape = [512] dtype = "float32" min_val = float("1.32904e-16") @@ -1661,6 +1812,7 @@ class Program_weight_tensor_parameter_150: class Program_weight_tensor_parameter_151: name = "parameter_151" + original_name = "batch_norm2d_22.w_1" shape = [512] dtype = "float32" min_val = float("-0.278824") @@ -1672,6 +1824,7 @@ class Program_weight_tensor_parameter_151: class Program_weight_tensor_parameter_152: name = "parameter_152" + original_name = "conv2d_28.w_0" shape = [512, 1, 1, 1] dtype = "float32" min_val = float("-0.140368") @@ -1683,6 +1836,7 @@ class Program_weight_tensor_parameter_152: class Program_weight_tensor_parameter_153: name = "parameter_153" + original_name = "batch_norm2d_21.b_0" shape = [512] dtype = "float32" min_val = float("-1.82712") @@ -1694,6 +1848,7 @@ class Program_weight_tensor_parameter_153: class Program_weight_tensor_parameter_154: name = "parameter_154" + original_name = "batch_norm2d_21.w_0" shape = [512] dtype = "float32" min_val = float("-0.349138") @@ -1705,6 +1860,7 @@ class Program_weight_tensor_parameter_154: class Program_weight_tensor_parameter_155: name = "parameter_155" + original_name = "batch_norm2d_21.w_2" shape = [512] dtype = "float32" min_val = float("7.03632e-07") @@ -1716,6 +1872,7 @@ class Program_weight_tensor_parameter_155: class Program_weight_tensor_parameter_156: name = "parameter_156" + original_name = "batch_norm2d_21.w_1" shape = [512] dtype = "float32" min_val = float("-2.60065") @@ -1727,6 +1884,7 @@ class Program_weight_tensor_parameter_156: class Program_weight_tensor_parameter_157: name = "parameter_157" + original_name = "conv2d_27.w_0" shape = [512, 1, 3, 3] dtype = "float32" min_val = float("-0.37054") @@ -1738,6 +1896,7 @@ class Program_weight_tensor_parameter_157: class Program_weight_tensor_parameter_158: name = "parameter_158" + original_name = "batch_norm2d_20.b_0" shape = [512] dtype = "float32" min_val = float("-1.82712") @@ -1749,6 +1908,7 @@ class Program_weight_tensor_parameter_158: class Program_weight_tensor_parameter_159: name = "parameter_159" + original_name = "batch_norm2d_20.w_0" shape = [512] dtype = "float32" min_val = float("-0.566326") @@ -1760,6 +1920,7 @@ class Program_weight_tensor_parameter_159: class Program_weight_tensor_parameter_160: name = "parameter_160" + original_name = "batch_norm2d_20.w_2" shape = [512] dtype = "float32" min_val = float("0.00158792") @@ -1771,6 +1932,7 @@ class Program_weight_tensor_parameter_160: class Program_weight_tensor_parameter_161: name = "parameter_161" + original_name = "batch_norm2d_20.w_1" shape = [512] dtype = "float32" min_val = float("-4.98537") @@ -1782,6 +1944,7 @@ class Program_weight_tensor_parameter_161: class Program_weight_tensor_parameter_162: name = "parameter_162" + original_name = "conv2d_26.w_0" shape = [512, 1, 5, 5] dtype = "float32" min_val = float("-0.243933") @@ -1793,6 +1956,7 @@ class Program_weight_tensor_parameter_162: class Program_weight_tensor_parameter_163: name = "parameter_163" + original_name = "batch_norm2d_19.b_0" shape = [512] dtype = "float32" min_val = float("-2.11591") @@ -1804,6 +1968,7 @@ class Program_weight_tensor_parameter_163: class Program_weight_tensor_parameter_164: name = "parameter_164" + original_name = "batch_norm2d_19.w_0" shape = [512] dtype = "float32" min_val = float("0.0820572") @@ -1815,6 +1980,7 @@ class Program_weight_tensor_parameter_164: class Program_weight_tensor_parameter_165: name = "parameter_165" + original_name = "batch_norm2d_19.w_2" shape = [512] dtype = "float32" min_val = float("1.23568") @@ -1826,6 +1992,7 @@ class Program_weight_tensor_parameter_165: class Program_weight_tensor_parameter_166: name = "parameter_166" + original_name = "batch_norm2d_19.w_1" shape = [512] dtype = "float32" min_val = float("-6.28304") @@ -1837,6 +2004,7 @@ class Program_weight_tensor_parameter_166: class Program_weight_tensor_parameter_167: name = "parameter_167" + original_name = "conv2d_25.w_0" shape = [512, 256, 1, 1] dtype = "float32" min_val = float("-1.02249") @@ -1848,6 +2016,7 @@ class Program_weight_tensor_parameter_167: class Program_weight_tensor_parameter_168: name = "parameter_168" + original_name = "batch_norm2d_18.b_0" shape = [256] dtype = "float32" min_val = float("-4.74669") @@ -1859,6 +2028,7 @@ class Program_weight_tensor_parameter_168: class Program_weight_tensor_parameter_169: name = "parameter_169" + original_name = "batch_norm2d_18.w_0" shape = [256] dtype = "float32" min_val = float("1.45577") @@ -1870,6 +2040,7 @@ class Program_weight_tensor_parameter_169: class Program_weight_tensor_parameter_170: name = "parameter_170" + original_name = "batch_norm2d_18.w_2" shape = [256] dtype = "float32" min_val = float("1.53825") @@ -1881,6 +2052,7 @@ class Program_weight_tensor_parameter_170: class Program_weight_tensor_parameter_171: name = "parameter_171" + original_name = "batch_norm2d_18.w_1" shape = [256] dtype = "float32" min_val = float("-3.72747") @@ -1892,6 +2064,7 @@ class Program_weight_tensor_parameter_171: class Program_weight_tensor_parameter_172: name = "parameter_172" + original_name = "conv2d_24.w_0" shape = [256, 512, 1, 1] dtype = "float32" min_val = float("-0.955977") @@ -1903,6 +2076,7 @@ class Program_weight_tensor_parameter_172: class Program_weight_tensor_parameter_173: name = "parameter_173" + original_name = "conv2d_23.b_0" shape = [512] dtype = "float32" min_val = float("-0.0866419") @@ -1914,6 +2088,7 @@ class Program_weight_tensor_parameter_173: class Program_weight_tensor_parameter_174: name = "parameter_174" + original_name = "conv2d_23.w_0" shape = [512, 128, 1, 1] dtype = "float32" min_val = float("-0.763721") @@ -1925,6 +2100,7 @@ class Program_weight_tensor_parameter_174: class Program_weight_tensor_parameter_175: name = "parameter_175" + original_name = "conv2d_22.b_0" shape = [128] dtype = "float32" min_val = float("-0.0152322") @@ -1936,6 +2112,7 @@ class Program_weight_tensor_parameter_175: class Program_weight_tensor_parameter_176: name = "parameter_176" + original_name = "conv2d_22.w_0" shape = [128, 512, 1, 1] dtype = "float32" min_val = float("-0.743588") @@ -1947,6 +2124,7 @@ class Program_weight_tensor_parameter_176: class Program_weight_tensor_parameter_177: name = "parameter_177" + original_name = "batch_norm2d_17.b_0" shape = [512] dtype = "float32" min_val = float("-2.00951") @@ -1958,6 +2136,7 @@ class Program_weight_tensor_parameter_177: class Program_weight_tensor_parameter_178: name = "parameter_178" + original_name = "batch_norm2d_17.w_0" shape = [512] dtype = "float32" min_val = float("-3.33509") @@ -1969,6 +2148,7 @@ class Program_weight_tensor_parameter_178: class Program_weight_tensor_parameter_179: name = "parameter_179" + original_name = "batch_norm2d_17.w_2" shape = [512] dtype = "float32" min_val = float("7.59628e-20") @@ -1980,6 +2160,7 @@ class Program_weight_tensor_parameter_179: class Program_weight_tensor_parameter_180: name = "parameter_180" + original_name = "batch_norm2d_17.w_1" shape = [512] dtype = "float32" min_val = float("-0.213635") @@ -1991,6 +2172,7 @@ class Program_weight_tensor_parameter_180: class Program_weight_tensor_parameter_181: name = "parameter_181" + original_name = "conv2d_20.w_0" shape = [512, 1, 1, 1] dtype = "float32" min_val = float("-0.115733") @@ -2002,6 +2184,7 @@ class Program_weight_tensor_parameter_181: class Program_weight_tensor_parameter_182: name = "parameter_182" + original_name = "batch_norm2d_16.b_0" shape = [512] dtype = "float32" min_val = float("-2.00951") @@ -2013,6 +2196,7 @@ class Program_weight_tensor_parameter_182: class Program_weight_tensor_parameter_183: name = "parameter_183" + original_name = "batch_norm2d_16.w_0" shape = [512] dtype = "float32" min_val = float("-2.07384") @@ -2024,6 +2208,7 @@ class Program_weight_tensor_parameter_183: class Program_weight_tensor_parameter_184: name = "parameter_184" + original_name = "batch_norm2d_16.w_2" shape = [512] dtype = "float32" min_val = float("7.24203e-07") @@ -2035,6 +2220,7 @@ class Program_weight_tensor_parameter_184: class Program_weight_tensor_parameter_185: name = "parameter_185" + original_name = "batch_norm2d_16.w_1" shape = [512] dtype = "float32" min_val = float("-3.87882") @@ -2046,6 +2232,7 @@ class Program_weight_tensor_parameter_185: class Program_weight_tensor_parameter_186: name = "parameter_186" + original_name = "conv2d_19.w_0" shape = [512, 1, 3, 3] dtype = "float32" min_val = float("-0.424429") @@ -2057,6 +2244,7 @@ class Program_weight_tensor_parameter_186: class Program_weight_tensor_parameter_187: name = "parameter_187" + original_name = "batch_norm2d_15.b_0" shape = [512] dtype = "float32" min_val = float("-2.00951") @@ -2068,6 +2256,7 @@ class Program_weight_tensor_parameter_187: class Program_weight_tensor_parameter_188: name = "parameter_188" + original_name = "batch_norm2d_15.w_0" shape = [512] dtype = "float32" min_val = float("-0.520411") @@ -2079,6 +2268,7 @@ class Program_weight_tensor_parameter_188: class Program_weight_tensor_parameter_189: name = "parameter_189" + original_name = "batch_norm2d_15.w_2" shape = [512] dtype = "float32" min_val = float("0.000797748") @@ -2090,6 +2280,7 @@ class Program_weight_tensor_parameter_189: class Program_weight_tensor_parameter_190: name = "parameter_190" + original_name = "batch_norm2d_15.w_1" shape = [512] dtype = "float32" min_val = float("-8.33902") @@ -2101,6 +2292,7 @@ class Program_weight_tensor_parameter_190: class Program_weight_tensor_parameter_191: name = "parameter_191" + original_name = "conv2d_18.w_0" shape = [512, 1, 5, 5] dtype = "float32" min_val = float("-0.338285") @@ -2112,6 +2304,7 @@ class Program_weight_tensor_parameter_191: class Program_weight_tensor_parameter_192: name = "parameter_192" + original_name = "batch_norm2d_14.b_0" shape = [512] dtype = "float32" min_val = float("-3.16709") @@ -2123,6 +2316,7 @@ class Program_weight_tensor_parameter_192: class Program_weight_tensor_parameter_193: name = "parameter_193" + original_name = "batch_norm2d_14.w_0" shape = [512] dtype = "float32" min_val = float("0.129088") @@ -2134,6 +2328,7 @@ class Program_weight_tensor_parameter_193: class Program_weight_tensor_parameter_194: name = "parameter_194" + original_name = "batch_norm2d_14.w_2" shape = [512] dtype = "float32" min_val = float("1.44322") @@ -2145,6 +2340,7 @@ class Program_weight_tensor_parameter_194: class Program_weight_tensor_parameter_195: name = "parameter_195" + original_name = "batch_norm2d_14.w_1" shape = [512] dtype = "float32" min_val = float("-10.6214") @@ -2156,6 +2352,7 @@ class Program_weight_tensor_parameter_195: class Program_weight_tensor_parameter_196: name = "parameter_196" + original_name = "conv2d_17.w_0" shape = [512, 256, 1, 1] dtype = "float32" min_val = float("-1.22785") @@ -2167,6 +2364,7 @@ class Program_weight_tensor_parameter_196: class Program_weight_tensor_parameter_197: name = "parameter_197" + original_name = "batch_norm2d_13.b_0" shape = [256] dtype = "float32" min_val = float("-6.34418") @@ -2178,6 +2376,7 @@ class Program_weight_tensor_parameter_197: class Program_weight_tensor_parameter_198: name = "parameter_198" + original_name = "batch_norm2d_13.w_0" shape = [256] dtype = "float32" min_val = float("1.37056") @@ -2189,6 +2388,7 @@ class Program_weight_tensor_parameter_198: class Program_weight_tensor_parameter_199: name = "parameter_199" + original_name = "batch_norm2d_13.w_2" shape = [256] dtype = "float32" min_val = float("0.670561") @@ -2200,6 +2400,7 @@ class Program_weight_tensor_parameter_199: class Program_weight_tensor_parameter_200: name = "parameter_200" + original_name = "batch_norm2d_13.w_1" shape = [256] dtype = "float32" min_val = float("-5.93789") @@ -2211,6 +2412,7 @@ class Program_weight_tensor_parameter_200: class Program_weight_tensor_parameter_201: name = "parameter_201" + original_name = "conv2d_16.w_0" shape = [256, 256, 1, 1] dtype = "float32" min_val = float("-1.03799") @@ -2222,6 +2424,7 @@ class Program_weight_tensor_parameter_201: class Program_weight_tensor_parameter_202: name = "parameter_202" + original_name = "conv2d_15.b_0" shape = [256] dtype = "float32" min_val = float("-0.113673") @@ -2233,6 +2436,7 @@ class Program_weight_tensor_parameter_202: class Program_weight_tensor_parameter_203: name = "parameter_203" + original_name = "conv2d_15.w_0" shape = [256, 64, 1, 1] dtype = "float32" min_val = float("-0.558348") @@ -2244,6 +2448,7 @@ class Program_weight_tensor_parameter_203: class Program_weight_tensor_parameter_204: name = "parameter_204" + original_name = "conv2d_14.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -2253,6 +2458,7 @@ class Program_weight_tensor_parameter_204: class Program_weight_tensor_parameter_205: name = "parameter_205" + original_name = "conv2d_14.w_0" shape = [64, 256, 1, 1] dtype = "float32" min_val = float("-0.78541") @@ -2264,6 +2470,7 @@ class Program_weight_tensor_parameter_205: class Program_weight_tensor_parameter_206: name = "parameter_206" + original_name = "batch_norm2d_12.b_0" shape = [256] dtype = "float32" min_val = float("-1.90402") @@ -2275,6 +2482,7 @@ class Program_weight_tensor_parameter_206: class Program_weight_tensor_parameter_207: name = "parameter_207" + original_name = "batch_norm2d_12.w_0" shape = [256] dtype = "float32" min_val = float("-0.64051") @@ -2286,6 +2494,7 @@ class Program_weight_tensor_parameter_207: class Program_weight_tensor_parameter_208: name = "parameter_208" + original_name = "batch_norm2d_12.w_2" shape = [256] dtype = "float32" min_val = float("0.000156689") @@ -2297,6 +2506,7 @@ class Program_weight_tensor_parameter_208: class Program_weight_tensor_parameter_209: name = "parameter_209" + original_name = "batch_norm2d_12.w_1" shape = [256] dtype = "float32" min_val = float("-4.58032") @@ -2308,6 +2518,7 @@ class Program_weight_tensor_parameter_209: class Program_weight_tensor_parameter_210: name = "parameter_210" + original_name = "conv2d_12.w_0" shape = [256, 1, 3, 3] dtype = "float32" min_val = float("-0.287251") @@ -2319,6 +2530,7 @@ class Program_weight_tensor_parameter_210: class Program_weight_tensor_parameter_211: name = "parameter_211" + original_name = "batch_norm2d_11.b_0" shape = [256] dtype = "float32" min_val = float("-1.90402") @@ -2330,6 +2542,7 @@ class Program_weight_tensor_parameter_211: class Program_weight_tensor_parameter_212: name = "parameter_212" + original_name = "batch_norm2d_11.w_0" shape = [256] dtype = "float32" min_val = float("-0.399563") @@ -2341,6 +2554,7 @@ class Program_weight_tensor_parameter_212: class Program_weight_tensor_parameter_213: name = "parameter_213" + original_name = "batch_norm2d_11.w_2" shape = [256] dtype = "float32" min_val = float("0.0156294") @@ -2352,6 +2566,7 @@ class Program_weight_tensor_parameter_213: class Program_weight_tensor_parameter_214: name = "parameter_214" + original_name = "batch_norm2d_11.w_1" shape = [256] dtype = "float32" min_val = float("-4.50212") @@ -2363,6 +2578,7 @@ class Program_weight_tensor_parameter_214: class Program_weight_tensor_parameter_215: name = "parameter_215" + original_name = "conv2d_11.w_0" shape = [256, 1, 5, 5] dtype = "float32" min_val = float("-0.339778") @@ -2374,6 +2590,7 @@ class Program_weight_tensor_parameter_215: class Program_weight_tensor_parameter_216: name = "parameter_216" + original_name = "batch_norm2d_10.b_0" shape = [256] dtype = "float32" min_val = float("-3.8278") @@ -2385,6 +2602,7 @@ class Program_weight_tensor_parameter_216: class Program_weight_tensor_parameter_217: name = "parameter_217" + original_name = "batch_norm2d_10.w_0" shape = [256] dtype = "float32" min_val = float("0.449185") @@ -2396,6 +2614,7 @@ class Program_weight_tensor_parameter_217: class Program_weight_tensor_parameter_218: name = "parameter_218" + original_name = "batch_norm2d_10.w_2" shape = [256] dtype = "float32" min_val = float("1.78478") @@ -2407,6 +2626,7 @@ class Program_weight_tensor_parameter_218: class Program_weight_tensor_parameter_219: name = "parameter_219" + original_name = "batch_norm2d_10.w_1" shape = [256] dtype = "float32" min_val = float("-15.3463") @@ -2418,6 +2638,7 @@ class Program_weight_tensor_parameter_219: class Program_weight_tensor_parameter_220: name = "parameter_220" + original_name = "conv2d_10.w_0" shape = [256, 256, 1, 1] dtype = "float32" min_val = float("-1.02007") @@ -2429,6 +2650,7 @@ class Program_weight_tensor_parameter_220: class Program_weight_tensor_parameter_221: name = "parameter_221" + original_name = "batch_norm2d_9.b_0" shape = [256] dtype = "float32" min_val = float("-2.89318") @@ -2440,6 +2662,7 @@ class Program_weight_tensor_parameter_221: class Program_weight_tensor_parameter_222: name = "parameter_222" + original_name = "batch_norm2d_9.w_0" shape = [256] dtype = "float32" min_val = float("1.07422") @@ -2451,6 +2674,7 @@ class Program_weight_tensor_parameter_222: class Program_weight_tensor_parameter_223: name = "parameter_223" + original_name = "batch_norm2d_9.w_2" shape = [256] dtype = "float32" min_val = float("0.0392314") @@ -2462,6 +2686,7 @@ class Program_weight_tensor_parameter_223: class Program_weight_tensor_parameter_224: name = "parameter_224" + original_name = "batch_norm2d_9.w_1" shape = [256] dtype = "float32" min_val = float("-6.80052") @@ -2473,6 +2698,7 @@ class Program_weight_tensor_parameter_224: class Program_weight_tensor_parameter_225: name = "parameter_225" + original_name = "conv2d_9.w_0" shape = [256, 1, 3, 3] dtype = "float32" min_val = float("-1.37071") @@ -2484,6 +2710,7 @@ class Program_weight_tensor_parameter_225: class Program_weight_tensor_parameter_226: name = "parameter_226" + original_name = "batch_norm2d_8.b_0" shape = [256] dtype = "float32" min_val = float("-4.96061") @@ -2495,6 +2722,7 @@ class Program_weight_tensor_parameter_226: class Program_weight_tensor_parameter_227: name = "parameter_227" + original_name = "batch_norm2d_8.w_0" shape = [256] dtype = "float32" min_val = float("0.461848") @@ -2506,6 +2734,7 @@ class Program_weight_tensor_parameter_227: class Program_weight_tensor_parameter_228: name = "parameter_228" + original_name = "batch_norm2d_8.w_2" shape = [256] dtype = "float32" min_val = float("2.76751") @@ -2517,6 +2746,7 @@ class Program_weight_tensor_parameter_228: class Program_weight_tensor_parameter_229: name = "parameter_229" + original_name = "batch_norm2d_8.w_1" shape = [256] dtype = "float32" min_val = float("-22.479") @@ -2528,6 +2758,7 @@ class Program_weight_tensor_parameter_229: class Program_weight_tensor_parameter_230: name = "parameter_230" + original_name = "conv2d_8.w_0" shape = [256, 128, 1, 1] dtype = "float32" min_val = float("-2.14361") @@ -2539,6 +2770,7 @@ class Program_weight_tensor_parameter_230: class Program_weight_tensor_parameter_231: name = "parameter_231" + original_name = "batch_norm2d_7.b_0" shape = [128] dtype = "float32" min_val = float("-3.35061") @@ -2550,6 +2782,7 @@ class Program_weight_tensor_parameter_231: class Program_weight_tensor_parameter_232: name = "parameter_232" + original_name = "batch_norm2d_7.w_0" shape = [128] dtype = "float32" min_val = float("1.33067") @@ -2561,6 +2794,7 @@ class Program_weight_tensor_parameter_232: class Program_weight_tensor_parameter_233: name = "parameter_233" + original_name = "batch_norm2d_7.w_2" shape = [128] dtype = "float32" min_val = float("0.192111") @@ -2572,6 +2806,7 @@ class Program_weight_tensor_parameter_233: class Program_weight_tensor_parameter_234: name = "parameter_234" + original_name = "batch_norm2d_7.w_1" shape = [128] dtype = "float32" min_val = float("-6.39434") @@ -2583,6 +2818,7 @@ class Program_weight_tensor_parameter_234: class Program_weight_tensor_parameter_235: name = "parameter_235" + original_name = "conv2d_7.w_0" shape = [128, 1, 3, 3] dtype = "float32" min_val = float("-0.513908") @@ -2594,6 +2830,7 @@ class Program_weight_tensor_parameter_235: class Program_weight_tensor_parameter_236: name = "parameter_236" + original_name = "batch_norm2d_6.b_0" shape = [128] dtype = "float32" min_val = float("-3.17094") @@ -2605,6 +2842,7 @@ class Program_weight_tensor_parameter_236: class Program_weight_tensor_parameter_237: name = "parameter_237" + original_name = "batch_norm2d_6.w_0" shape = [128] dtype = "float32" min_val = float("0.491935") @@ -2616,6 +2854,7 @@ class Program_weight_tensor_parameter_237: class Program_weight_tensor_parameter_238: name = "parameter_238" + original_name = "batch_norm2d_6.w_2" shape = [128] dtype = "float32" min_val = float("4.87695") @@ -2627,6 +2866,7 @@ class Program_weight_tensor_parameter_238: class Program_weight_tensor_parameter_239: name = "parameter_239" + original_name = "batch_norm2d_6.w_1" shape = [128] dtype = "float32" min_val = float("-26.843") @@ -2638,6 +2878,7 @@ class Program_weight_tensor_parameter_239: class Program_weight_tensor_parameter_240: name = "parameter_240" + original_name = "conv2d_6.w_0" shape = [128, 128, 1, 1] dtype = "float32" min_val = float("-2.01457") @@ -2649,6 +2890,7 @@ class Program_weight_tensor_parameter_240: class Program_weight_tensor_parameter_241: name = "parameter_241" + original_name = "batch_norm2d_5.b_0" shape = [128] dtype = "float32" min_val = float("-2.42436") @@ -2660,6 +2902,7 @@ class Program_weight_tensor_parameter_241: class Program_weight_tensor_parameter_242: name = "parameter_242" + original_name = "batch_norm2d_5.w_0" shape = [128] dtype = "float32" min_val = float("1.32557") @@ -2671,6 +2914,7 @@ class Program_weight_tensor_parameter_242: class Program_weight_tensor_parameter_243: name = "parameter_243" + original_name = "batch_norm2d_5.w_2" shape = [128] dtype = "float32" min_val = float("0.0209423") @@ -2682,6 +2926,7 @@ class Program_weight_tensor_parameter_243: class Program_weight_tensor_parameter_244: name = "parameter_244" + original_name = "batch_norm2d_5.w_1" shape = [128] dtype = "float32" min_val = float("-5.10852") @@ -2693,6 +2938,7 @@ class Program_weight_tensor_parameter_244: class Program_weight_tensor_parameter_245: name = "parameter_245" + original_name = "conv2d_5.w_0" shape = [128, 1, 3, 3] dtype = "float32" min_val = float("-1.36374") @@ -2704,6 +2950,7 @@ class Program_weight_tensor_parameter_245: class Program_weight_tensor_parameter_246: name = "parameter_246" + original_name = "batch_norm2d_4.b_0" shape = [128] dtype = "float32" min_val = float("-2.83123") @@ -2715,6 +2962,7 @@ class Program_weight_tensor_parameter_246: class Program_weight_tensor_parameter_247: name = "parameter_247" + original_name = "batch_norm2d_4.w_0" shape = [128] dtype = "float32" min_val = float("0.574823") @@ -2726,6 +2974,7 @@ class Program_weight_tensor_parameter_247: class Program_weight_tensor_parameter_248: name = "parameter_248" + original_name = "batch_norm2d_4.w_2" shape = [128] dtype = "float32" min_val = float("4.24792") @@ -2737,6 +2986,7 @@ class Program_weight_tensor_parameter_248: class Program_weight_tensor_parameter_249: name = "parameter_249" + original_name = "batch_norm2d_4.w_1" shape = [128] dtype = "float32" min_val = float("-14.5176") @@ -2748,6 +2998,7 @@ class Program_weight_tensor_parameter_249: class Program_weight_tensor_parameter_250: name = "parameter_250" + original_name = "conv2d_4.w_0" shape = [128, 64, 1, 1] dtype = "float32" min_val = float("-1.26239") @@ -2759,6 +3010,7 @@ class Program_weight_tensor_parameter_250: class Program_weight_tensor_parameter_251: name = "parameter_251" + original_name = "batch_norm2d_3.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -2768,6 +3020,7 @@ class Program_weight_tensor_parameter_251: class Program_weight_tensor_parameter_252: name = "parameter_252" + original_name = "batch_norm2d_3.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -2777,6 +3030,7 @@ class Program_weight_tensor_parameter_252: class Program_weight_tensor_parameter_253: name = "parameter_253" + original_name = "batch_norm2d_3.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -2786,6 +3040,7 @@ class Program_weight_tensor_parameter_253: class Program_weight_tensor_parameter_254: name = "parameter_254" + original_name = "batch_norm2d_3.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -2795,6 +3050,7 @@ class Program_weight_tensor_parameter_254: class Program_weight_tensor_parameter_255: name = "parameter_255" + original_name = "conv2d_3.w_0" shape = [64, 1, 3, 3] dtype = "float32" min_val = float("-0.929814") @@ -2806,6 +3062,7 @@ class Program_weight_tensor_parameter_255: class Program_weight_tensor_parameter_256: name = "parameter_256" + original_name = "batch_norm2d_2.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -2815,6 +3072,7 @@ class Program_weight_tensor_parameter_256: class Program_weight_tensor_parameter_257: name = "parameter_257" + original_name = "batch_norm2d_2.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -2824,6 +3082,7 @@ class Program_weight_tensor_parameter_257: class Program_weight_tensor_parameter_258: name = "parameter_258" + original_name = "batch_norm2d_2.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -2833,6 +3092,7 @@ class Program_weight_tensor_parameter_258: class Program_weight_tensor_parameter_259: name = "parameter_259" + original_name = "batch_norm2d_2.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -2842,6 +3102,7 @@ class Program_weight_tensor_parameter_259: class Program_weight_tensor_parameter_260: name = "parameter_260" + original_name = "conv2d_2.w_0" shape = [64, 32, 1, 1] dtype = "float32" min_val = float("-2.07561") @@ -2853,6 +3114,7 @@ class Program_weight_tensor_parameter_260: class Program_weight_tensor_parameter_261: name = "parameter_261" + original_name = "batch_norm2d_1.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -2862,6 +3124,7 @@ class Program_weight_tensor_parameter_261: class Program_weight_tensor_parameter_262: name = "parameter_262" + original_name = "batch_norm2d_1.w_0" shape = [32] dtype = "float32" min_val = float("0") @@ -2871,6 +3134,7 @@ class Program_weight_tensor_parameter_262: class Program_weight_tensor_parameter_263: name = "parameter_263" + original_name = "batch_norm2d_1.w_2" shape = [32] dtype = "float32" min_val = float("0") @@ -2880,6 +3144,7 @@ class Program_weight_tensor_parameter_263: class Program_weight_tensor_parameter_264: name = "parameter_264" + original_name = "batch_norm2d_1.w_1" shape = [32] dtype = "float32" min_val = float("0") @@ -2889,6 +3154,7 @@ class Program_weight_tensor_parameter_264: class Program_weight_tensor_parameter_265: name = "parameter_265" + original_name = "conv2d_1.w_0" shape = [32, 1, 3, 3] dtype = "float32" min_val = float("-2.31529") @@ -2900,6 +3166,7 @@ class Program_weight_tensor_parameter_265: class Program_weight_tensor_parameter_266: name = "parameter_266" + original_name = "batch_norm2d_0.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -2909,6 +3176,7 @@ class Program_weight_tensor_parameter_266: class Program_weight_tensor_parameter_267: name = "parameter_267" + original_name = "batch_norm2d_0.w_0" shape = [32] dtype = "float32" min_val = float("0") @@ -2918,6 +3186,7 @@ class Program_weight_tensor_parameter_267: class Program_weight_tensor_parameter_268: name = "parameter_268" + original_name = "batch_norm2d_0.w_2" shape = [32] dtype = "float32" min_val = float("0") @@ -2927,6 +3196,7 @@ class Program_weight_tensor_parameter_268: class Program_weight_tensor_parameter_269: name = "parameter_269" + original_name = "batch_norm2d_0.w_1" shape = [32] dtype = "float32" min_val = float("0") @@ -2936,6 +3206,7 @@ class Program_weight_tensor_parameter_269: class Program_weight_tensor_parameter_270: name = "parameter_270" + original_name = "conv2d_0.w_0" shape = [32, 3, 3, 3] dtype = "float32" min_val = float("-1.54054") diff --git a/paddle_samples/PaddleX/PP-YOLOE-R-L/subgraph_3/input_meta.py b/paddle_samples/PaddleX/PP-YOLOE-R-L/subgraph_3/input_meta.py index eef1c0cfc..370e425ce 100644 --- a/paddle_samples/PaddleX/PP-YOLOE-R-L/subgraph_3/input_meta.py +++ b/paddle_samples/PaddleX/PP-YOLOE-R-L/subgraph_3/input_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_data_0: name = "data_0" + original_name = "param_66" shape = [1] dtype = "float32" data = [1.0] @@ -7,6 +8,7 @@ class Program_weight_tensor_data_0: class Program_weight_tensor_data_1: name = "data_1" + original_name = "param_67" shape = [1] dtype = "float32" data = [1.0] @@ -14,6 +16,7 @@ class Program_weight_tensor_data_1: class Program_weight_tensor_data_2: name = "data_2" + original_name = "param_68" shape = [1] dtype = "float32" data = [1.0] @@ -21,6 +24,7 @@ class Program_weight_tensor_data_2: class Program_weight_tensor_data_3: name = "data_3" + original_name = "param_69" shape = [1] dtype = "float32" data = [1.0] @@ -28,6 +32,7 @@ class Program_weight_tensor_data_3: class Program_weight_tensor_data_4: name = "data_4" + original_name = "param_70" shape = [1] dtype = "float32" data = [1.0] @@ -35,6 +40,7 @@ class Program_weight_tensor_data_4: class Program_weight_tensor_data_5: name = "data_5" + original_name = "param_71" shape = [1] dtype = "float32" data = [1.0] @@ -42,6 +48,7 @@ class Program_weight_tensor_data_5: class Program_weight_tensor_data_6: name = "data_6" + original_name = "param_72" shape = [1] dtype = "float32" data = [1.0] @@ -49,6 +56,7 @@ class Program_weight_tensor_data_6: class Program_weight_tensor_data_7: name = "data_7" + original_name = "param_73" shape = [1] dtype = "float32" data = [1.0] @@ -56,6 +64,7 @@ class Program_weight_tensor_data_7: class Program_weight_tensor_data_8: name = "data_8" + original_name = "param_74" shape = [1] dtype = "float32" data = [1.0] @@ -63,6 +72,7 @@ class Program_weight_tensor_data_8: class Program_weight_tensor_data_9: name = "data_9" + original_name = "param_75" shape = [1] dtype = "float32" data = [1.0] @@ -70,6 +80,7 @@ class Program_weight_tensor_data_9: class Program_weight_tensor_data_10: name = "data_10" + original_name = "param_76" shape = [1] dtype = "float32" data = [1.0] @@ -77,6 +88,7 @@ class Program_weight_tensor_data_10: class Program_weight_tensor_data_11: name = "data_11" + original_name = "param_77" shape = [1] dtype = "float32" data = [1.0] @@ -84,6 +96,7 @@ class Program_weight_tensor_data_11: class Program_weight_tensor_data_12: name = "data_12" + original_name = "param_78" shape = [1] dtype = "float32" data = [1.0] @@ -91,6 +104,7 @@ class Program_weight_tensor_data_12: class Program_weight_tensor_data_13: name = "data_13" + original_name = "param_79" shape = [1] dtype = "float32" data = [1.0] @@ -98,6 +112,7 @@ class Program_weight_tensor_data_13: class Program_weight_tensor_data_14: name = "data_14" + original_name = "param_80" shape = [1] dtype = "float32" data = [1.0] @@ -105,6 +120,7 @@ class Program_weight_tensor_data_14: class Program_weight_tensor_data_15: name = "data_15" + original_name = "param_81" shape = [1] dtype = "float32" data = [1.0] @@ -112,6 +128,7 @@ class Program_weight_tensor_data_15: class Program_weight_tensor_data_16: name = "data_16" + original_name = "param_82" shape = [1] dtype = "float32" data = [1.0] @@ -119,6 +136,7 @@ class Program_weight_tensor_data_16: class Program_weight_tensor_data_17: name = "data_17" + original_name = "param_83" shape = [1] dtype = "float32" data = [1.0] @@ -126,6 +144,7 @@ class Program_weight_tensor_data_17: class Program_weight_tensor_data_18: name = "data_18" + original_name = "param_84" shape = [1] dtype = "float32" data = [1.0] @@ -133,6 +152,7 @@ class Program_weight_tensor_data_18: class Program_weight_tensor_data_19: name = "data_19" + original_name = "param_85" shape = [1] dtype = "float32" data = [1.0] @@ -140,6 +160,7 @@ class Program_weight_tensor_data_19: class Program_weight_tensor_data_20: name = "data_20" + original_name = "param_86" shape = [1] dtype = "float32" data = [1.0] @@ -147,6 +168,7 @@ class Program_weight_tensor_data_20: class Program_weight_tensor_data_21: name = "data_21" + original_name = "param_87" shape = [1] dtype = "float32" data = [1.0] @@ -154,6 +176,7 @@ class Program_weight_tensor_data_21: class Program_weight_tensor_data_22: name = "data_22" + original_name = "param_88" shape = [1] dtype = "float32" data = [1.0] @@ -161,6 +184,7 @@ class Program_weight_tensor_data_22: class Program_weight_tensor_data_23: name = "data_23" + original_name = "param_89" shape = [1] dtype = "float32" data = [1.0] @@ -168,6 +192,7 @@ class Program_weight_tensor_data_23: class Program_weight_tensor_data_24: name = "data_24" + original_name = "param_90" shape = [1] dtype = "float32" data = [1.0] @@ -175,6 +200,7 @@ class Program_weight_tensor_data_24: class Program_weight_tensor_data_25: name = "data_25" + original_name = "param_91" shape = [1] dtype = "float32" data = [1.0] @@ -182,6 +208,7 @@ class Program_weight_tensor_data_25: class Program_weight_tensor_data_26: name = "data_26" + original_name = "param_92" shape = [1] dtype = "float32" data = [1.0] @@ -189,6 +216,7 @@ class Program_weight_tensor_data_26: class Program_weight_tensor_data_27: name = "data_27" + original_name = "param_93" shape = [1] dtype = "float32" data = [1.0] @@ -196,6 +224,7 @@ class Program_weight_tensor_data_27: class Program_weight_tensor_data_28: name = "data_28" + original_name = "param_94" shape = [1] dtype = "float32" data = [1.0] @@ -203,6 +232,7 @@ class Program_weight_tensor_data_28: class Program_weight_tensor_data_29: name = "data_29" + original_name = "param_95" shape = [1] dtype = "float32" data = [1.0] @@ -210,6 +240,7 @@ class Program_weight_tensor_data_29: class Program_weight_tensor_data_30: name = "data_30" + original_name = "param_96" shape = [1] dtype = "float32" data = [1.0] @@ -217,6 +248,7 @@ class Program_weight_tensor_data_30: class Program_weight_tensor_data_31: name = "data_31" + original_name = "param_97" shape = [1] dtype = "float32" data = [1.0] @@ -224,6 +256,7 @@ class Program_weight_tensor_data_31: class Program_weight_tensor_data_32: name = "data_32" + original_name = "param_98" shape = [1] dtype = "float32" data = [1.0] @@ -231,6 +264,7 @@ class Program_weight_tensor_data_32: class Program_weight_tensor_data_33: name = "data_33" + original_name = "var_1321" shape = [1, 3, 1024, 1024] dtype = "float32" min_val = float("-2.1179") diff --git a/paddle_samples/PaddleX/PP-YOLOE-R-L/subgraph_3/weight_meta.py b/paddle_samples/PaddleX/PP-YOLOE-R-L/subgraph_3/weight_meta.py index a6a19c8c2..1059a040f 100644 --- a/paddle_samples/PaddleX/PP-YOLOE-R-L/subgraph_3/weight_meta.py +++ b/paddle_samples/PaddleX/PP-YOLOE-R-L/subgraph_3/weight_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_parameter_0: name = "parameter_0" + original_name = "batch_norm2d_137.b_0" shape = [768] dtype = "float32" data = None @@ -7,6 +8,7 @@ class Program_weight_tensor_parameter_0: class Program_weight_tensor_parameter_1: name = "parameter_1" + original_name = "batch_norm2d_137.w_0" shape = [768] dtype = "float32" min_val = float("1.0") @@ -17,6 +19,7 @@ class Program_weight_tensor_parameter_1: class Program_weight_tensor_parameter_2: name = "parameter_2" + original_name = "batch_norm2d_137.w_2" shape = [768] dtype = "float32" min_val = float("1.0") @@ -27,6 +30,7 @@ class Program_weight_tensor_parameter_2: class Program_weight_tensor_parameter_3: name = "parameter_3" + original_name = "batch_norm2d_137.w_1" shape = [768] dtype = "float32" data = None @@ -34,6 +38,7 @@ class Program_weight_tensor_parameter_3: class Program_weight_tensor_parameter_4: name = "parameter_4" + original_name = "conv2d_141.w_0" shape = [768, 768, 1, 1] dtype = "float32" min_val = float("-0.225002") @@ -45,6 +50,7 @@ class Program_weight_tensor_parameter_4: class Program_weight_tensor_parameter_5: name = "parameter_5" + original_name = "batch_norm2d_136.b_0" shape = [384] dtype = "float32" data = None @@ -52,6 +58,7 @@ class Program_weight_tensor_parameter_5: class Program_weight_tensor_parameter_6: name = "parameter_6" + original_name = "batch_norm2d_136.w_0" shape = [384] dtype = "float32" min_val = float("1.0") @@ -63,6 +70,7 @@ class Program_weight_tensor_parameter_6: class Program_weight_tensor_parameter_7: name = "parameter_7" + original_name = "batch_norm2d_136.w_2" shape = [384] dtype = "float32" min_val = float("1.0") @@ -74,6 +82,7 @@ class Program_weight_tensor_parameter_7: class Program_weight_tensor_parameter_8: name = "parameter_8" + original_name = "batch_norm2d_136.w_1" shape = [384] dtype = "float32" data = None @@ -81,6 +90,7 @@ class Program_weight_tensor_parameter_8: class Program_weight_tensor_parameter_9: name = "parameter_9" + original_name = "conv2d_140.w_0" shape = [384, 384, 1, 1] dtype = "float32" min_val = float("-0.322942") @@ -92,6 +102,7 @@ class Program_weight_tensor_parameter_9: class Program_weight_tensor_parameter_10: name = "parameter_10" + original_name = "batch_norm2d_135.b_0" shape = [384] dtype = "float32" data = None @@ -99,6 +110,7 @@ class Program_weight_tensor_parameter_10: class Program_weight_tensor_parameter_11: name = "parameter_11" + original_name = "batch_norm2d_135.w_0" shape = [384] dtype = "float32" min_val = float("1.0") @@ -110,6 +122,7 @@ class Program_weight_tensor_parameter_11: class Program_weight_tensor_parameter_12: name = "parameter_12" + original_name = "batch_norm2d_135.w_2" shape = [384] dtype = "float32" min_val = float("1.0") @@ -121,6 +134,7 @@ class Program_weight_tensor_parameter_12: class Program_weight_tensor_parameter_13: name = "parameter_13" + original_name = "batch_norm2d_135.w_1" shape = [384] dtype = "float32" data = None @@ -128,6 +142,7 @@ class Program_weight_tensor_parameter_13: class Program_weight_tensor_parameter_14: name = "parameter_14" + original_name = "conv2d_139.w_0" shape = [384, 384, 3, 3] dtype = "float32" min_val = float("-0.12304") @@ -139,6 +154,7 @@ class Program_weight_tensor_parameter_14: class Program_weight_tensor_parameter_15: name = "parameter_15" + original_name = "batch_norm2d_134.b_0" shape = [384] dtype = "float32" data = None @@ -146,6 +162,7 @@ class Program_weight_tensor_parameter_15: class Program_weight_tensor_parameter_16: name = "parameter_16" + original_name = "batch_norm2d_134.w_0" shape = [384] dtype = "float32" min_val = float("1.0") @@ -157,6 +174,7 @@ class Program_weight_tensor_parameter_16: class Program_weight_tensor_parameter_17: name = "parameter_17" + original_name = "batch_norm2d_134.w_2" shape = [384] dtype = "float32" min_val = float("1.0") @@ -168,6 +186,7 @@ class Program_weight_tensor_parameter_17: class Program_weight_tensor_parameter_18: name = "parameter_18" + original_name = "batch_norm2d_134.w_1" shape = [384] dtype = "float32" data = None @@ -175,6 +194,7 @@ class Program_weight_tensor_parameter_18: class Program_weight_tensor_parameter_19: name = "parameter_19" + original_name = "conv2d_138.w_0" shape = [384, 384, 3, 3] dtype = "float32" min_val = float("-0.119696") @@ -186,6 +206,7 @@ class Program_weight_tensor_parameter_19: class Program_weight_tensor_parameter_20: name = "parameter_20" + original_name = "batch_norm2d_133.b_0" shape = [384] dtype = "float32" data = None @@ -193,6 +214,7 @@ class Program_weight_tensor_parameter_20: class Program_weight_tensor_parameter_21: name = "parameter_21" + original_name = "batch_norm2d_133.w_0" shape = [384] dtype = "float32" min_val = float("1.0") @@ -204,6 +226,7 @@ class Program_weight_tensor_parameter_21: class Program_weight_tensor_parameter_22: name = "parameter_22" + original_name = "batch_norm2d_133.w_2" shape = [384] dtype = "float32" min_val = float("1.0") @@ -215,6 +238,7 @@ class Program_weight_tensor_parameter_22: class Program_weight_tensor_parameter_23: name = "parameter_23" + original_name = "batch_norm2d_133.w_1" shape = [384] dtype = "float32" data = None @@ -222,6 +246,7 @@ class Program_weight_tensor_parameter_23: class Program_weight_tensor_parameter_24: name = "parameter_24" + original_name = "conv2d_137.w_0" shape = [384, 384, 1, 1] dtype = "float32" min_val = float("-0.31449") @@ -233,6 +258,7 @@ class Program_weight_tensor_parameter_24: class Program_weight_tensor_parameter_25: name = "parameter_25" + original_name = "batch_norm2d_132.b_0" shape = [384] dtype = "float32" data = None @@ -240,6 +266,7 @@ class Program_weight_tensor_parameter_25: class Program_weight_tensor_parameter_26: name = "parameter_26" + original_name = "batch_norm2d_132.w_0" shape = [384] dtype = "float32" min_val = float("1.0") @@ -251,6 +278,7 @@ class Program_weight_tensor_parameter_26: class Program_weight_tensor_parameter_27: name = "parameter_27" + original_name = "batch_norm2d_132.w_2" shape = [384] dtype = "float32" min_val = float("1.0") @@ -262,6 +290,7 @@ class Program_weight_tensor_parameter_27: class Program_weight_tensor_parameter_28: name = "parameter_28" + original_name = "batch_norm2d_132.w_1" shape = [384] dtype = "float32" data = None @@ -269,6 +298,7 @@ class Program_weight_tensor_parameter_28: class Program_weight_tensor_parameter_29: name = "parameter_29" + original_name = "conv2d_136.w_0" shape = [384, 384, 3, 3] dtype = "float32" min_val = float("-0.115579") @@ -280,6 +310,7 @@ class Program_weight_tensor_parameter_29: class Program_weight_tensor_parameter_30: name = "parameter_30" + original_name = "batch_norm2d_131.b_0" shape = [384] dtype = "float32" data = None @@ -287,6 +318,7 @@ class Program_weight_tensor_parameter_30: class Program_weight_tensor_parameter_31: name = "parameter_31" + original_name = "batch_norm2d_131.w_0" shape = [384] dtype = "float32" min_val = float("1.0") @@ -298,6 +330,7 @@ class Program_weight_tensor_parameter_31: class Program_weight_tensor_parameter_32: name = "parameter_32" + original_name = "batch_norm2d_131.w_2" shape = [384] dtype = "float32" min_val = float("1.0") @@ -309,6 +342,7 @@ class Program_weight_tensor_parameter_32: class Program_weight_tensor_parameter_33: name = "parameter_33" + original_name = "batch_norm2d_131.w_1" shape = [384] dtype = "float32" data = None @@ -316,6 +350,7 @@ class Program_weight_tensor_parameter_33: class Program_weight_tensor_parameter_34: name = "parameter_34" + original_name = "conv2d_135.w_0" shape = [384, 384, 3, 3] dtype = "float32" min_val = float("-0.113888") @@ -327,6 +362,7 @@ class Program_weight_tensor_parameter_34: class Program_weight_tensor_parameter_35: name = "parameter_35" + original_name = "batch_norm2d_130.b_0" shape = [384] dtype = "float32" data = None @@ -334,6 +370,7 @@ class Program_weight_tensor_parameter_35: class Program_weight_tensor_parameter_36: name = "parameter_36" + original_name = "batch_norm2d_130.w_0" shape = [384] dtype = "float32" min_val = float("1.0") @@ -345,6 +382,7 @@ class Program_weight_tensor_parameter_36: class Program_weight_tensor_parameter_37: name = "parameter_37" + original_name = "batch_norm2d_130.w_2" shape = [384] dtype = "float32" min_val = float("1.0") @@ -356,6 +394,7 @@ class Program_weight_tensor_parameter_37: class Program_weight_tensor_parameter_38: name = "parameter_38" + original_name = "batch_norm2d_130.w_1" shape = [384] dtype = "float32" data = None @@ -363,6 +402,7 @@ class Program_weight_tensor_parameter_38: class Program_weight_tensor_parameter_39: name = "parameter_39" + original_name = "conv2d_134.w_0" shape = [384, 384, 1, 1] dtype = "float32" min_val = float("-0.302931") @@ -374,6 +414,7 @@ class Program_weight_tensor_parameter_39: class Program_weight_tensor_parameter_40: name = "parameter_40" + original_name = "batch_norm2d_129.b_0" shape = [384] dtype = "float32" data = None @@ -381,6 +422,7 @@ class Program_weight_tensor_parameter_40: class Program_weight_tensor_parameter_41: name = "parameter_41" + original_name = "batch_norm2d_129.w_0" shape = [384] dtype = "float32" min_val = float("1.0") @@ -392,6 +434,7 @@ class Program_weight_tensor_parameter_41: class Program_weight_tensor_parameter_42: name = "parameter_42" + original_name = "batch_norm2d_129.w_2" shape = [384] dtype = "float32" min_val = float("1.0") @@ -403,6 +446,7 @@ class Program_weight_tensor_parameter_42: class Program_weight_tensor_parameter_43: name = "parameter_43" + original_name = "batch_norm2d_129.w_1" shape = [384] dtype = "float32" data = None @@ -410,6 +454,7 @@ class Program_weight_tensor_parameter_43: class Program_weight_tensor_parameter_44: name = "parameter_44" + original_name = "conv2d_133.w_0" shape = [384, 384, 3, 3] dtype = "float32" min_val = float("-0.114187") @@ -421,6 +466,7 @@ class Program_weight_tensor_parameter_44: class Program_weight_tensor_parameter_45: name = "parameter_45" + original_name = "batch_norm2d_128.b_0" shape = [384] dtype = "float32" data = None @@ -428,6 +474,7 @@ class Program_weight_tensor_parameter_45: class Program_weight_tensor_parameter_46: name = "parameter_46" + original_name = "batch_norm2d_128.w_0" shape = [384] dtype = "float32" min_val = float("1.0") @@ -439,6 +486,7 @@ class Program_weight_tensor_parameter_46: class Program_weight_tensor_parameter_47: name = "parameter_47" + original_name = "batch_norm2d_128.w_2" shape = [384] dtype = "float32" min_val = float("1.0") @@ -450,6 +498,7 @@ class Program_weight_tensor_parameter_47: class Program_weight_tensor_parameter_48: name = "parameter_48" + original_name = "batch_norm2d_128.w_1" shape = [384] dtype = "float32" data = None @@ -457,6 +506,7 @@ class Program_weight_tensor_parameter_48: class Program_weight_tensor_parameter_49: name = "parameter_49" + original_name = "conv2d_132.w_0" shape = [384, 384, 3, 3] dtype = "float32" min_val = float("-0.118342") @@ -468,6 +518,7 @@ class Program_weight_tensor_parameter_49: class Program_weight_tensor_parameter_50: name = "parameter_50" + original_name = "batch_norm2d_127.b_0" shape = [384] dtype = "float32" data = None @@ -475,6 +526,7 @@ class Program_weight_tensor_parameter_50: class Program_weight_tensor_parameter_51: name = "parameter_51" + original_name = "batch_norm2d_127.w_0" shape = [384] dtype = "float32" min_val = float("1.0") @@ -486,6 +538,7 @@ class Program_weight_tensor_parameter_51: class Program_weight_tensor_parameter_52: name = "parameter_52" + original_name = "batch_norm2d_127.w_2" shape = [384] dtype = "float32" min_val = float("1.0") @@ -497,6 +550,7 @@ class Program_weight_tensor_parameter_52: class Program_weight_tensor_parameter_53: name = "parameter_53" + original_name = "batch_norm2d_127.w_1" shape = [384] dtype = "float32" data = None @@ -504,6 +558,7 @@ class Program_weight_tensor_parameter_53: class Program_weight_tensor_parameter_54: name = "parameter_54" + original_name = "conv2d_131.w_0" shape = [384, 1152, 1, 1] dtype = "float32" min_val = float("-0.191508") @@ -515,6 +570,7 @@ class Program_weight_tensor_parameter_54: class Program_weight_tensor_parameter_55: name = "parameter_55" + original_name = "batch_norm2d_126.b_0" shape = [384] dtype = "float32" data = None @@ -522,6 +578,7 @@ class Program_weight_tensor_parameter_55: class Program_weight_tensor_parameter_56: name = "parameter_56" + original_name = "batch_norm2d_126.w_0" shape = [384] dtype = "float32" min_val = float("1.0") @@ -533,6 +590,7 @@ class Program_weight_tensor_parameter_56: class Program_weight_tensor_parameter_57: name = "parameter_57" + original_name = "batch_norm2d_126.w_2" shape = [384] dtype = "float32" min_val = float("1.0") @@ -544,6 +602,7 @@ class Program_weight_tensor_parameter_57: class Program_weight_tensor_parameter_58: name = "parameter_58" + original_name = "batch_norm2d_126.w_1" shape = [384] dtype = "float32" data = None @@ -551,6 +610,7 @@ class Program_weight_tensor_parameter_58: class Program_weight_tensor_parameter_59: name = "parameter_59" + original_name = "conv2d_130.w_0" shape = [384, 1152, 1, 1] dtype = "float32" min_val = float("-0.181007") @@ -562,6 +622,7 @@ class Program_weight_tensor_parameter_59: class Program_weight_tensor_parameter_60: name = "parameter_60" + original_name = "batch_norm2d_125.b_0" shape = [384] dtype = "float32" data = None @@ -569,6 +630,7 @@ class Program_weight_tensor_parameter_60: class Program_weight_tensor_parameter_61: name = "parameter_61" + original_name = "batch_norm2d_125.w_0" shape = [384] dtype = "float32" min_val = float("1.0") @@ -580,6 +642,7 @@ class Program_weight_tensor_parameter_61: class Program_weight_tensor_parameter_62: name = "parameter_62" + original_name = "batch_norm2d_125.w_2" shape = [384] dtype = "float32" min_val = float("1.0") @@ -591,6 +654,7 @@ class Program_weight_tensor_parameter_62: class Program_weight_tensor_parameter_63: name = "parameter_63" + original_name = "batch_norm2d_125.w_1" shape = [384] dtype = "float32" data = None @@ -598,6 +662,7 @@ class Program_weight_tensor_parameter_63: class Program_weight_tensor_parameter_64: name = "parameter_64" + original_name = "conv2d_129.w_0" shape = [384, 384, 3, 3] dtype = "float32" min_val = float("-0.129977") @@ -609,6 +674,7 @@ class Program_weight_tensor_parameter_64: class Program_weight_tensor_parameter_65: name = "parameter_65" + original_name = "batch_norm2d_124.b_0" shape = [384] dtype = "float32" data = None @@ -616,6 +682,7 @@ class Program_weight_tensor_parameter_65: class Program_weight_tensor_parameter_66: name = "parameter_66" + original_name = "batch_norm2d_124.w_0" shape = [384] dtype = "float32" min_val = float("1.0") @@ -627,6 +694,7 @@ class Program_weight_tensor_parameter_66: class Program_weight_tensor_parameter_67: name = "parameter_67" + original_name = "batch_norm2d_124.w_2" shape = [384] dtype = "float32" min_val = float("1.0") @@ -638,6 +706,7 @@ class Program_weight_tensor_parameter_67: class Program_weight_tensor_parameter_68: name = "parameter_68" + original_name = "batch_norm2d_124.w_1" shape = [384] dtype = "float32" data = None @@ -645,6 +714,7 @@ class Program_weight_tensor_parameter_68: class Program_weight_tensor_parameter_69: name = "parameter_69" + original_name = "conv2d_128.w_0" shape = [384, 384, 1, 1] dtype = "float32" min_val = float("-0.302833") @@ -656,6 +726,7 @@ class Program_weight_tensor_parameter_69: class Program_weight_tensor_parameter_70: name = "parameter_70" + original_name = "batch_norm2d_123.b_0" shape = [192] dtype = "float32" data = None @@ -663,6 +734,7 @@ class Program_weight_tensor_parameter_70: class Program_weight_tensor_parameter_71: name = "parameter_71" + original_name = "batch_norm2d_123.w_0" shape = [192] dtype = "float32" min_val = float("1.0") @@ -673,6 +745,7 @@ class Program_weight_tensor_parameter_71: class Program_weight_tensor_parameter_72: name = "parameter_72" + original_name = "batch_norm2d_123.w_2" shape = [192] dtype = "float32" min_val = float("1.0") @@ -683,6 +756,7 @@ class Program_weight_tensor_parameter_72: class Program_weight_tensor_parameter_73: name = "parameter_73" + original_name = "batch_norm2d_123.w_1" shape = [192] dtype = "float32" data = None @@ -690,6 +764,7 @@ class Program_weight_tensor_parameter_73: class Program_weight_tensor_parameter_74: name = "parameter_74" + original_name = "conv2d_127.w_0" shape = [192, 192, 1, 1] dtype = "float32" min_val = float("-0.400722") @@ -701,6 +776,7 @@ class Program_weight_tensor_parameter_74: class Program_weight_tensor_parameter_75: name = "parameter_75" + original_name = "batch_norm2d_122.b_0" shape = [192] dtype = "float32" data = None @@ -708,6 +784,7 @@ class Program_weight_tensor_parameter_75: class Program_weight_tensor_parameter_76: name = "parameter_76" + original_name = "batch_norm2d_122.w_0" shape = [192] dtype = "float32" min_val = float("1.0") @@ -718,6 +795,7 @@ class Program_weight_tensor_parameter_76: class Program_weight_tensor_parameter_77: name = "parameter_77" + original_name = "batch_norm2d_122.w_2" shape = [192] dtype = "float32" min_val = float("1.0") @@ -728,6 +806,7 @@ class Program_weight_tensor_parameter_77: class Program_weight_tensor_parameter_78: name = "parameter_78" + original_name = "batch_norm2d_122.w_1" shape = [192] dtype = "float32" data = None @@ -735,6 +814,7 @@ class Program_weight_tensor_parameter_78: class Program_weight_tensor_parameter_79: name = "parameter_79" + original_name = "conv2d_126.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.154177") @@ -746,6 +826,7 @@ class Program_weight_tensor_parameter_79: class Program_weight_tensor_parameter_80: name = "parameter_80" + original_name = "batch_norm2d_121.b_0" shape = [192] dtype = "float32" data = None @@ -753,6 +834,7 @@ class Program_weight_tensor_parameter_80: class Program_weight_tensor_parameter_81: name = "parameter_81" + original_name = "batch_norm2d_121.w_0" shape = [192] dtype = "float32" min_val = float("1.0") @@ -763,6 +845,7 @@ class Program_weight_tensor_parameter_81: class Program_weight_tensor_parameter_82: name = "parameter_82" + original_name = "batch_norm2d_121.w_2" shape = [192] dtype = "float32" min_val = float("1.0") @@ -773,6 +856,7 @@ class Program_weight_tensor_parameter_82: class Program_weight_tensor_parameter_83: name = "parameter_83" + original_name = "batch_norm2d_121.w_1" shape = [192] dtype = "float32" data = None @@ -780,6 +864,7 @@ class Program_weight_tensor_parameter_83: class Program_weight_tensor_parameter_84: name = "parameter_84" + original_name = "conv2d_125.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.181651") @@ -791,6 +876,7 @@ class Program_weight_tensor_parameter_84: class Program_weight_tensor_parameter_85: name = "parameter_85" + original_name = "batch_norm2d_120.b_0" shape = [192] dtype = "float32" data = None @@ -798,6 +884,7 @@ class Program_weight_tensor_parameter_85: class Program_weight_tensor_parameter_86: name = "parameter_86" + original_name = "batch_norm2d_120.w_0" shape = [192] dtype = "float32" min_val = float("1.0") @@ -808,6 +895,7 @@ class Program_weight_tensor_parameter_86: class Program_weight_tensor_parameter_87: name = "parameter_87" + original_name = "batch_norm2d_120.w_2" shape = [192] dtype = "float32" min_val = float("1.0") @@ -818,6 +906,7 @@ class Program_weight_tensor_parameter_87: class Program_weight_tensor_parameter_88: name = "parameter_88" + original_name = "batch_norm2d_120.w_1" shape = [192] dtype = "float32" data = None @@ -825,6 +914,7 @@ class Program_weight_tensor_parameter_88: class Program_weight_tensor_parameter_89: name = "parameter_89" + original_name = "conv2d_124.w_0" shape = [192, 192, 1, 1] dtype = "float32" min_val = float("-0.384016") @@ -836,6 +926,7 @@ class Program_weight_tensor_parameter_89: class Program_weight_tensor_parameter_90: name = "parameter_90" + original_name = "batch_norm2d_119.b_0" shape = [192] dtype = "float32" data = None @@ -843,6 +934,7 @@ class Program_weight_tensor_parameter_90: class Program_weight_tensor_parameter_91: name = "parameter_91" + original_name = "batch_norm2d_119.w_0" shape = [192] dtype = "float32" min_val = float("1.0") @@ -853,6 +945,7 @@ class Program_weight_tensor_parameter_91: class Program_weight_tensor_parameter_92: name = "parameter_92" + original_name = "batch_norm2d_119.w_2" shape = [192] dtype = "float32" min_val = float("1.0") @@ -863,6 +956,7 @@ class Program_weight_tensor_parameter_92: class Program_weight_tensor_parameter_93: name = "parameter_93" + original_name = "batch_norm2d_119.w_1" shape = [192] dtype = "float32" data = None @@ -870,6 +964,7 @@ class Program_weight_tensor_parameter_93: class Program_weight_tensor_parameter_94: name = "parameter_94" + original_name = "conv2d_123.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.158839") @@ -881,6 +976,7 @@ class Program_weight_tensor_parameter_94: class Program_weight_tensor_parameter_95: name = "parameter_95" + original_name = "batch_norm2d_118.b_0" shape = [192] dtype = "float32" data = None @@ -888,6 +984,7 @@ class Program_weight_tensor_parameter_95: class Program_weight_tensor_parameter_96: name = "parameter_96" + original_name = "batch_norm2d_118.w_0" shape = [192] dtype = "float32" min_val = float("1.0") @@ -898,6 +995,7 @@ class Program_weight_tensor_parameter_96: class Program_weight_tensor_parameter_97: name = "parameter_97" + original_name = "batch_norm2d_118.w_2" shape = [192] dtype = "float32" min_val = float("1.0") @@ -908,6 +1006,7 @@ class Program_weight_tensor_parameter_97: class Program_weight_tensor_parameter_98: name = "parameter_98" + original_name = "batch_norm2d_118.w_1" shape = [192] dtype = "float32" data = None @@ -915,6 +1014,7 @@ class Program_weight_tensor_parameter_98: class Program_weight_tensor_parameter_99: name = "parameter_99" + original_name = "conv2d_122.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.152118") @@ -926,6 +1026,7 @@ class Program_weight_tensor_parameter_99: class Program_weight_tensor_parameter_100: name = "parameter_100" + original_name = "batch_norm2d_117.b_0" shape = [192] dtype = "float32" data = None @@ -933,6 +1034,7 @@ class Program_weight_tensor_parameter_100: class Program_weight_tensor_parameter_101: name = "parameter_101" + original_name = "batch_norm2d_117.w_0" shape = [192] dtype = "float32" min_val = float("1.0") @@ -943,6 +1045,7 @@ class Program_weight_tensor_parameter_101: class Program_weight_tensor_parameter_102: name = "parameter_102" + original_name = "batch_norm2d_117.w_2" shape = [192] dtype = "float32" min_val = float("1.0") @@ -953,6 +1056,7 @@ class Program_weight_tensor_parameter_102: class Program_weight_tensor_parameter_103: name = "parameter_103" + original_name = "batch_norm2d_117.w_1" shape = [192] dtype = "float32" data = None @@ -960,6 +1064,7 @@ class Program_weight_tensor_parameter_103: class Program_weight_tensor_parameter_104: name = "parameter_104" + original_name = "conv2d_121.w_0" shape = [192, 192, 1, 1] dtype = "float32" min_val = float("-0.412261") @@ -971,6 +1076,7 @@ class Program_weight_tensor_parameter_104: class Program_weight_tensor_parameter_105: name = "parameter_105" + original_name = "batch_norm2d_116.b_0" shape = [192] dtype = "float32" data = None @@ -978,6 +1084,7 @@ class Program_weight_tensor_parameter_105: class Program_weight_tensor_parameter_106: name = "parameter_106" + original_name = "batch_norm2d_116.w_0" shape = [192] dtype = "float32" min_val = float("1.0") @@ -988,6 +1095,7 @@ class Program_weight_tensor_parameter_106: class Program_weight_tensor_parameter_107: name = "parameter_107" + original_name = "batch_norm2d_116.w_2" shape = [192] dtype = "float32" min_val = float("1.0") @@ -998,6 +1106,7 @@ class Program_weight_tensor_parameter_107: class Program_weight_tensor_parameter_108: name = "parameter_108" + original_name = "batch_norm2d_116.w_1" shape = [192] dtype = "float32" data = None @@ -1005,6 +1114,7 @@ class Program_weight_tensor_parameter_108: class Program_weight_tensor_parameter_109: name = "parameter_109" + original_name = "conv2d_120.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.149017") @@ -1016,6 +1126,7 @@ class Program_weight_tensor_parameter_109: class Program_weight_tensor_parameter_110: name = "parameter_110" + original_name = "batch_norm2d_115.b_0" shape = [192] dtype = "float32" data = None @@ -1023,6 +1134,7 @@ class Program_weight_tensor_parameter_110: class Program_weight_tensor_parameter_111: name = "parameter_111" + original_name = "batch_norm2d_115.w_0" shape = [192] dtype = "float32" min_val = float("1.0") @@ -1033,6 +1145,7 @@ class Program_weight_tensor_parameter_111: class Program_weight_tensor_parameter_112: name = "parameter_112" + original_name = "batch_norm2d_115.w_2" shape = [192] dtype = "float32" min_val = float("1.0") @@ -1043,6 +1156,7 @@ class Program_weight_tensor_parameter_112: class Program_weight_tensor_parameter_113: name = "parameter_113" + original_name = "batch_norm2d_115.w_1" shape = [192] dtype = "float32" data = None @@ -1050,6 +1164,7 @@ class Program_weight_tensor_parameter_113: class Program_weight_tensor_parameter_114: name = "parameter_114" + original_name = "conv2d_119.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.171461") @@ -1061,6 +1176,7 @@ class Program_weight_tensor_parameter_114: class Program_weight_tensor_parameter_115: name = "parameter_115" + original_name = "batch_norm2d_114.b_0" shape = [192] dtype = "float32" data = None @@ -1068,6 +1184,7 @@ class Program_weight_tensor_parameter_115: class Program_weight_tensor_parameter_116: name = "parameter_116" + original_name = "batch_norm2d_114.w_0" shape = [192] dtype = "float32" min_val = float("1.0") @@ -1078,6 +1195,7 @@ class Program_weight_tensor_parameter_116: class Program_weight_tensor_parameter_117: name = "parameter_117" + original_name = "batch_norm2d_114.w_2" shape = [192] dtype = "float32" min_val = float("1.0") @@ -1088,6 +1206,7 @@ class Program_weight_tensor_parameter_117: class Program_weight_tensor_parameter_118: name = "parameter_118" + original_name = "batch_norm2d_114.w_1" shape = [192] dtype = "float32" data = None @@ -1095,6 +1214,7 @@ class Program_weight_tensor_parameter_118: class Program_weight_tensor_parameter_119: name = "parameter_119" + original_name = "conv2d_118.w_0" shape = [192, 576, 1, 1] dtype = "float32" min_val = float("-0.255931") @@ -1106,6 +1226,7 @@ class Program_weight_tensor_parameter_119: class Program_weight_tensor_parameter_120: name = "parameter_120" + original_name = "batch_norm2d_113.b_0" shape = [192] dtype = "float32" data = None @@ -1113,6 +1234,7 @@ class Program_weight_tensor_parameter_120: class Program_weight_tensor_parameter_121: name = "parameter_121" + original_name = "batch_norm2d_113.w_0" shape = [192] dtype = "float32" min_val = float("1.0") @@ -1123,6 +1245,7 @@ class Program_weight_tensor_parameter_121: class Program_weight_tensor_parameter_122: name = "parameter_122" + original_name = "batch_norm2d_113.w_2" shape = [192] dtype = "float32" min_val = float("1.0") @@ -1133,6 +1256,7 @@ class Program_weight_tensor_parameter_122: class Program_weight_tensor_parameter_123: name = "parameter_123" + original_name = "batch_norm2d_113.w_1" shape = [192] dtype = "float32" data = None @@ -1140,6 +1264,7 @@ class Program_weight_tensor_parameter_123: class Program_weight_tensor_parameter_124: name = "parameter_124" + original_name = "conv2d_117.w_0" shape = [192, 576, 1, 1] dtype = "float32" min_val = float("-0.286651") @@ -1151,6 +1276,7 @@ class Program_weight_tensor_parameter_124: class Program_weight_tensor_parameter_125: name = "parameter_125" + original_name = "batch_norm2d_112.b_0" shape = [192] dtype = "float32" data = None @@ -1158,6 +1284,7 @@ class Program_weight_tensor_parameter_125: class Program_weight_tensor_parameter_126: name = "parameter_126" + original_name = "batch_norm2d_112.w_0" shape = [192] dtype = "float32" min_val = float("1.0") @@ -1168,6 +1295,7 @@ class Program_weight_tensor_parameter_126: class Program_weight_tensor_parameter_127: name = "parameter_127" + original_name = "batch_norm2d_112.w_2" shape = [192] dtype = "float32" min_val = float("1.0") @@ -1178,6 +1306,7 @@ class Program_weight_tensor_parameter_127: class Program_weight_tensor_parameter_128: name = "parameter_128" + original_name = "batch_norm2d_112.w_1" shape = [192] dtype = "float32" data = None @@ -1185,6 +1314,7 @@ class Program_weight_tensor_parameter_128: class Program_weight_tensor_parameter_129: name = "parameter_129" + original_name = "conv2d_116.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.162603") @@ -1196,6 +1326,7 @@ class Program_weight_tensor_parameter_129: class Program_weight_tensor_parameter_130: name = "parameter_130" + original_name = "batch_norm2d_111.b_0" shape = [192] dtype = "float32" data = None @@ -1203,6 +1334,7 @@ class Program_weight_tensor_parameter_130: class Program_weight_tensor_parameter_131: name = "parameter_131" + original_name = "batch_norm2d_111.w_0" shape = [192] dtype = "float32" min_val = float("1.0") @@ -1213,6 +1345,7 @@ class Program_weight_tensor_parameter_131: class Program_weight_tensor_parameter_132: name = "parameter_132" + original_name = "batch_norm2d_111.w_2" shape = [192] dtype = "float32" min_val = float("1.0") @@ -1223,6 +1356,7 @@ class Program_weight_tensor_parameter_132: class Program_weight_tensor_parameter_133: name = "parameter_133" + original_name = "batch_norm2d_111.w_1" shape = [192] dtype = "float32" data = None @@ -1230,6 +1364,7 @@ class Program_weight_tensor_parameter_133: class Program_weight_tensor_parameter_134: name = "parameter_134" + original_name = "conv2d_115.w_0" shape = [192, 192, 1, 1] dtype = "float32" min_val = float("-0.441582") @@ -1241,6 +1376,7 @@ class Program_weight_tensor_parameter_134: class Program_weight_tensor_parameter_135: name = "parameter_135" + original_name = "batch_norm2d_110.b_0" shape = [96] dtype = "float32" data = None @@ -1248,6 +1384,7 @@ class Program_weight_tensor_parameter_135: class Program_weight_tensor_parameter_136: name = "parameter_136" + original_name = "batch_norm2d_110.w_0" shape = [96] dtype = "float32" min_val = float("1.0") @@ -1258,6 +1395,7 @@ class Program_weight_tensor_parameter_136: class Program_weight_tensor_parameter_137: name = "parameter_137" + original_name = "batch_norm2d_110.w_2" shape = [96] dtype = "float32" min_val = float("1.0") @@ -1268,6 +1406,7 @@ class Program_weight_tensor_parameter_137: class Program_weight_tensor_parameter_138: name = "parameter_138" + original_name = "batch_norm2d_110.w_1" shape = [96] dtype = "float32" data = None @@ -1275,6 +1414,7 @@ class Program_weight_tensor_parameter_138: class Program_weight_tensor_parameter_139: name = "parameter_139" + original_name = "conv2d_114.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.473161") @@ -1286,6 +1426,7 @@ class Program_weight_tensor_parameter_139: class Program_weight_tensor_parameter_140: name = "parameter_140" + original_name = "batch_norm2d_109.b_0" shape = [96] dtype = "float32" data = None @@ -1293,6 +1434,7 @@ class Program_weight_tensor_parameter_140: class Program_weight_tensor_parameter_141: name = "parameter_141" + original_name = "batch_norm2d_109.w_0" shape = [96] dtype = "float32" min_val = float("1.0") @@ -1303,6 +1445,7 @@ class Program_weight_tensor_parameter_141: class Program_weight_tensor_parameter_142: name = "parameter_142" + original_name = "batch_norm2d_109.w_2" shape = [96] dtype = "float32" min_val = float("1.0") @@ -1313,6 +1456,7 @@ class Program_weight_tensor_parameter_142: class Program_weight_tensor_parameter_143: name = "parameter_143" + original_name = "batch_norm2d_109.w_1" shape = [96] dtype = "float32" data = None @@ -1320,6 +1464,7 @@ class Program_weight_tensor_parameter_143: class Program_weight_tensor_parameter_144: name = "parameter_144" + original_name = "conv2d_113.w_0" shape = [96, 96, 3, 3] dtype = "float32" min_val = float("-0.209848") @@ -1331,6 +1476,7 @@ class Program_weight_tensor_parameter_144: class Program_weight_tensor_parameter_145: name = "parameter_145" + original_name = "batch_norm2d_108.b_0" shape = [96] dtype = "float32" data = None @@ -1338,6 +1484,7 @@ class Program_weight_tensor_parameter_145: class Program_weight_tensor_parameter_146: name = "parameter_146" + original_name = "batch_norm2d_108.w_0" shape = [96] dtype = "float32" min_val = float("1.0") @@ -1348,6 +1495,7 @@ class Program_weight_tensor_parameter_146: class Program_weight_tensor_parameter_147: name = "parameter_147" + original_name = "batch_norm2d_108.w_2" shape = [96] dtype = "float32" min_val = float("1.0") @@ -1358,6 +1506,7 @@ class Program_weight_tensor_parameter_147: class Program_weight_tensor_parameter_148: name = "parameter_148" + original_name = "batch_norm2d_108.w_1" shape = [96] dtype = "float32" data = None @@ -1365,6 +1514,7 @@ class Program_weight_tensor_parameter_148: class Program_weight_tensor_parameter_149: name = "parameter_149" + original_name = "conv2d_112.w_0" shape = [96, 96, 3, 3] dtype = "float32" min_val = float("-0.201192") @@ -1376,6 +1526,7 @@ class Program_weight_tensor_parameter_149: class Program_weight_tensor_parameter_150: name = "parameter_150" + original_name = "batch_norm2d_107.b_0" shape = [96] dtype = "float32" data = None @@ -1383,6 +1534,7 @@ class Program_weight_tensor_parameter_150: class Program_weight_tensor_parameter_151: name = "parameter_151" + original_name = "batch_norm2d_107.w_0" shape = [96] dtype = "float32" min_val = float("1.0") @@ -1393,6 +1545,7 @@ class Program_weight_tensor_parameter_151: class Program_weight_tensor_parameter_152: name = "parameter_152" + original_name = "batch_norm2d_107.w_2" shape = [96] dtype = "float32" min_val = float("1.0") @@ -1403,6 +1556,7 @@ class Program_weight_tensor_parameter_152: class Program_weight_tensor_parameter_153: name = "parameter_153" + original_name = "batch_norm2d_107.w_1" shape = [96] dtype = "float32" data = None @@ -1410,6 +1564,7 @@ class Program_weight_tensor_parameter_153: class Program_weight_tensor_parameter_154: name = "parameter_154" + original_name = "conv2d_111.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.570801") @@ -1421,6 +1576,7 @@ class Program_weight_tensor_parameter_154: class Program_weight_tensor_parameter_155: name = "parameter_155" + original_name = "batch_norm2d_106.b_0" shape = [96] dtype = "float32" data = None @@ -1428,6 +1584,7 @@ class Program_weight_tensor_parameter_155: class Program_weight_tensor_parameter_156: name = "parameter_156" + original_name = "batch_norm2d_106.w_0" shape = [96] dtype = "float32" min_val = float("1.0") @@ -1438,6 +1595,7 @@ class Program_weight_tensor_parameter_156: class Program_weight_tensor_parameter_157: name = "parameter_157" + original_name = "batch_norm2d_106.w_2" shape = [96] dtype = "float32" min_val = float("1.0") @@ -1448,6 +1606,7 @@ class Program_weight_tensor_parameter_157: class Program_weight_tensor_parameter_158: name = "parameter_158" + original_name = "batch_norm2d_106.w_1" shape = [96] dtype = "float32" data = None @@ -1455,6 +1614,7 @@ class Program_weight_tensor_parameter_158: class Program_weight_tensor_parameter_159: name = "parameter_159" + original_name = "conv2d_110.w_0" shape = [96, 96, 3, 3] dtype = "float32" min_val = float("-0.209186") @@ -1466,6 +1626,7 @@ class Program_weight_tensor_parameter_159: class Program_weight_tensor_parameter_160: name = "parameter_160" + original_name = "batch_norm2d_105.b_0" shape = [96] dtype = "float32" data = None @@ -1473,6 +1634,7 @@ class Program_weight_tensor_parameter_160: class Program_weight_tensor_parameter_161: name = "parameter_161" + original_name = "batch_norm2d_105.w_0" shape = [96] dtype = "float32" min_val = float("1.0") @@ -1483,6 +1645,7 @@ class Program_weight_tensor_parameter_161: class Program_weight_tensor_parameter_162: name = "parameter_162" + original_name = "batch_norm2d_105.w_2" shape = [96] dtype = "float32" min_val = float("1.0") @@ -1493,6 +1656,7 @@ class Program_weight_tensor_parameter_162: class Program_weight_tensor_parameter_163: name = "parameter_163" + original_name = "batch_norm2d_105.w_1" shape = [96] dtype = "float32" data = None @@ -1500,6 +1664,7 @@ class Program_weight_tensor_parameter_163: class Program_weight_tensor_parameter_164: name = "parameter_164" + original_name = "conv2d_109.w_0" shape = [96, 96, 3, 3] dtype = "float32" min_val = float("-0.211593") @@ -1511,6 +1676,7 @@ class Program_weight_tensor_parameter_164: class Program_weight_tensor_parameter_165: name = "parameter_165" + original_name = "batch_norm2d_104.b_0" shape = [96] dtype = "float32" data = None @@ -1518,6 +1684,7 @@ class Program_weight_tensor_parameter_165: class Program_weight_tensor_parameter_166: name = "parameter_166" + original_name = "batch_norm2d_104.w_0" shape = [96] dtype = "float32" min_val = float("1.0") @@ -1528,6 +1695,7 @@ class Program_weight_tensor_parameter_166: class Program_weight_tensor_parameter_167: name = "parameter_167" + original_name = "batch_norm2d_104.w_2" shape = [96] dtype = "float32" min_val = float("1.0") @@ -1538,6 +1706,7 @@ class Program_weight_tensor_parameter_167: class Program_weight_tensor_parameter_168: name = "parameter_168" + original_name = "batch_norm2d_104.w_1" shape = [96] dtype = "float32" data = None @@ -1545,6 +1714,7 @@ class Program_weight_tensor_parameter_168: class Program_weight_tensor_parameter_169: name = "parameter_169" + original_name = "conv2d_108.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.574778") @@ -1556,6 +1726,7 @@ class Program_weight_tensor_parameter_169: class Program_weight_tensor_parameter_170: name = "parameter_170" + original_name = "batch_norm2d_103.b_0" shape = [96] dtype = "float32" data = None @@ -1563,6 +1734,7 @@ class Program_weight_tensor_parameter_170: class Program_weight_tensor_parameter_171: name = "parameter_171" + original_name = "batch_norm2d_103.w_0" shape = [96] dtype = "float32" min_val = float("1.0") @@ -1573,6 +1745,7 @@ class Program_weight_tensor_parameter_171: class Program_weight_tensor_parameter_172: name = "parameter_172" + original_name = "batch_norm2d_103.w_2" shape = [96] dtype = "float32" min_val = float("1.0") @@ -1583,6 +1756,7 @@ class Program_weight_tensor_parameter_172: class Program_weight_tensor_parameter_173: name = "parameter_173" + original_name = "batch_norm2d_103.w_1" shape = [96] dtype = "float32" data = None @@ -1590,6 +1764,7 @@ class Program_weight_tensor_parameter_173: class Program_weight_tensor_parameter_174: name = "parameter_174" + original_name = "conv2d_107.w_0" shape = [96, 96, 3, 3] dtype = "float32" min_val = float("-0.235471") @@ -1601,6 +1776,7 @@ class Program_weight_tensor_parameter_174: class Program_weight_tensor_parameter_175: name = "parameter_175" + original_name = "batch_norm2d_102.b_0" shape = [96] dtype = "float32" data = None @@ -1608,6 +1784,7 @@ class Program_weight_tensor_parameter_175: class Program_weight_tensor_parameter_176: name = "parameter_176" + original_name = "batch_norm2d_102.w_0" shape = [96] dtype = "float32" min_val = float("1.0") @@ -1618,6 +1795,7 @@ class Program_weight_tensor_parameter_176: class Program_weight_tensor_parameter_177: name = "parameter_177" + original_name = "batch_norm2d_102.w_2" shape = [96] dtype = "float32" min_val = float("1.0") @@ -1628,6 +1806,7 @@ class Program_weight_tensor_parameter_177: class Program_weight_tensor_parameter_178: name = "parameter_178" + original_name = "batch_norm2d_102.w_1" shape = [96] dtype = "float32" data = None @@ -1635,6 +1814,7 @@ class Program_weight_tensor_parameter_178: class Program_weight_tensor_parameter_179: name = "parameter_179" + original_name = "conv2d_106.w_0" shape = [96, 96, 3, 3] dtype = "float32" min_val = float("-0.231455") @@ -1646,6 +1826,7 @@ class Program_weight_tensor_parameter_179: class Program_weight_tensor_parameter_180: name = "parameter_180" + original_name = "batch_norm2d_101.b_0" shape = [96] dtype = "float32" data = None @@ -1653,6 +1834,7 @@ class Program_weight_tensor_parameter_180: class Program_weight_tensor_parameter_181: name = "parameter_181" + original_name = "batch_norm2d_101.w_0" shape = [96] dtype = "float32" min_val = float("1.0") @@ -1663,6 +1845,7 @@ class Program_weight_tensor_parameter_181: class Program_weight_tensor_parameter_182: name = "parameter_182" + original_name = "batch_norm2d_101.w_2" shape = [96] dtype = "float32" min_val = float("1.0") @@ -1673,6 +1856,7 @@ class Program_weight_tensor_parameter_182: class Program_weight_tensor_parameter_183: name = "parameter_183" + original_name = "batch_norm2d_101.w_1" shape = [96] dtype = "float32" data = None @@ -1680,6 +1864,7 @@ class Program_weight_tensor_parameter_183: class Program_weight_tensor_parameter_184: name = "parameter_184" + original_name = "conv2d_105.w_0" shape = [96, 448, 1, 1] dtype = "float32" min_val = float("-0.273341") @@ -1691,6 +1876,7 @@ class Program_weight_tensor_parameter_184: class Program_weight_tensor_parameter_185: name = "parameter_185" + original_name = "batch_norm2d_100.b_0" shape = [96] dtype = "float32" data = None @@ -1698,6 +1884,7 @@ class Program_weight_tensor_parameter_185: class Program_weight_tensor_parameter_186: name = "parameter_186" + original_name = "batch_norm2d_100.w_0" shape = [96] dtype = "float32" min_val = float("1.0") @@ -1708,6 +1895,7 @@ class Program_weight_tensor_parameter_186: class Program_weight_tensor_parameter_187: name = "parameter_187" + original_name = "batch_norm2d_100.w_2" shape = [96] dtype = "float32" min_val = float("1.0") @@ -1718,6 +1906,7 @@ class Program_weight_tensor_parameter_187: class Program_weight_tensor_parameter_188: name = "parameter_188" + original_name = "batch_norm2d_100.w_1" shape = [96] dtype = "float32" data = None @@ -1725,6 +1914,7 @@ class Program_weight_tensor_parameter_188: class Program_weight_tensor_parameter_189: name = "parameter_189" + original_name = "conv2d_104.w_0" shape = [96, 448, 1, 1] dtype = "float32" min_val = float("-0.279151") @@ -1736,6 +1926,7 @@ class Program_weight_tensor_parameter_189: class Program_weight_tensor_parameter_190: name = "parameter_190" + original_name = "batch_norm2d_99.b_0" shape = [192] dtype = "float32" data = None @@ -1743,6 +1934,7 @@ class Program_weight_tensor_parameter_190: class Program_weight_tensor_parameter_191: name = "parameter_191" + original_name = "batch_norm2d_99.w_0" shape = [192] dtype = "float32" min_val = float("1.0") @@ -1753,6 +1945,7 @@ class Program_weight_tensor_parameter_191: class Program_weight_tensor_parameter_192: name = "parameter_192" + original_name = "batch_norm2d_99.w_2" shape = [192] dtype = "float32" min_val = float("1.0") @@ -1763,6 +1956,7 @@ class Program_weight_tensor_parameter_192: class Program_weight_tensor_parameter_193: name = "parameter_193" + original_name = "batch_norm2d_99.w_1" shape = [192] dtype = "float32" data = None @@ -1770,6 +1964,7 @@ class Program_weight_tensor_parameter_193: class Program_weight_tensor_parameter_194: name = "parameter_194" + original_name = "conv2d_103.w_0" shape = [192, 384, 1, 1] dtype = "float32" min_val = float("-0.299474") @@ -1781,6 +1976,7 @@ class Program_weight_tensor_parameter_194: class Program_weight_tensor_parameter_195: name = "parameter_195" + original_name = "batch_norm2d_98.b_0" shape = [384] dtype = "float32" data = None @@ -1788,6 +1984,7 @@ class Program_weight_tensor_parameter_195: class Program_weight_tensor_parameter_196: name = "parameter_196" + original_name = "batch_norm2d_98.w_0" shape = [384] dtype = "float32" min_val = float("1.0") @@ -1799,6 +1996,7 @@ class Program_weight_tensor_parameter_196: class Program_weight_tensor_parameter_197: name = "parameter_197" + original_name = "batch_norm2d_98.w_2" shape = [384] dtype = "float32" min_val = float("1.0") @@ -1810,6 +2008,7 @@ class Program_weight_tensor_parameter_197: class Program_weight_tensor_parameter_198: name = "parameter_198" + original_name = "batch_norm2d_98.w_1" shape = [384] dtype = "float32" data = None @@ -1817,6 +2016,7 @@ class Program_weight_tensor_parameter_198: class Program_weight_tensor_parameter_199: name = "parameter_199" + original_name = "conv2d_102.w_0" shape = [384, 384, 1, 1] dtype = "float32" min_val = float("-0.322599") @@ -1828,6 +2028,7 @@ class Program_weight_tensor_parameter_199: class Program_weight_tensor_parameter_200: name = "parameter_200" + original_name = "batch_norm2d_97.b_0" shape = [192] dtype = "float32" data = None @@ -1835,6 +2036,7 @@ class Program_weight_tensor_parameter_200: class Program_weight_tensor_parameter_201: name = "parameter_201" + original_name = "batch_norm2d_97.w_0" shape = [192] dtype = "float32" min_val = float("1.0") @@ -1845,6 +2047,7 @@ class Program_weight_tensor_parameter_201: class Program_weight_tensor_parameter_202: name = "parameter_202" + original_name = "batch_norm2d_97.w_2" shape = [192] dtype = "float32" min_val = float("1.0") @@ -1855,6 +2058,7 @@ class Program_weight_tensor_parameter_202: class Program_weight_tensor_parameter_203: name = "parameter_203" + original_name = "batch_norm2d_97.w_1" shape = [192] dtype = "float32" data = None @@ -1862,6 +2066,7 @@ class Program_weight_tensor_parameter_203: class Program_weight_tensor_parameter_204: name = "parameter_204" + original_name = "conv2d_101.w_0" shape = [192, 192, 1, 1] dtype = "float32" min_val = float("-0.414858") @@ -1873,6 +2078,7 @@ class Program_weight_tensor_parameter_204: class Program_weight_tensor_parameter_205: name = "parameter_205" + original_name = "batch_norm2d_96.b_0" shape = [192] dtype = "float32" data = None @@ -1880,6 +2086,7 @@ class Program_weight_tensor_parameter_205: class Program_weight_tensor_parameter_206: name = "parameter_206" + original_name = "batch_norm2d_96.w_0" shape = [192] dtype = "float32" min_val = float("1.0") @@ -1890,6 +2097,7 @@ class Program_weight_tensor_parameter_206: class Program_weight_tensor_parameter_207: name = "parameter_207" + original_name = "batch_norm2d_96.w_2" shape = [192] dtype = "float32" min_val = float("1.0") @@ -1900,6 +2108,7 @@ class Program_weight_tensor_parameter_207: class Program_weight_tensor_parameter_208: name = "parameter_208" + original_name = "batch_norm2d_96.w_1" shape = [192] dtype = "float32" data = None @@ -1907,6 +2116,7 @@ class Program_weight_tensor_parameter_208: class Program_weight_tensor_parameter_209: name = "parameter_209" + original_name = "conv2d_100.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.187654") @@ -1918,6 +2128,7 @@ class Program_weight_tensor_parameter_209: class Program_weight_tensor_parameter_210: name = "parameter_210" + original_name = "batch_norm2d_95.b_0" shape = [192] dtype = "float32" data = None @@ -1925,6 +2136,7 @@ class Program_weight_tensor_parameter_210: class Program_weight_tensor_parameter_211: name = "parameter_211" + original_name = "batch_norm2d_95.w_0" shape = [192] dtype = "float32" min_val = float("1.0") @@ -1935,6 +2147,7 @@ class Program_weight_tensor_parameter_211: class Program_weight_tensor_parameter_212: name = "parameter_212" + original_name = "batch_norm2d_95.w_2" shape = [192] dtype = "float32" min_val = float("1.0") @@ -1945,6 +2158,7 @@ class Program_weight_tensor_parameter_212: class Program_weight_tensor_parameter_213: name = "parameter_213" + original_name = "batch_norm2d_95.w_1" shape = [192] dtype = "float32" data = None @@ -1952,6 +2166,7 @@ class Program_weight_tensor_parameter_213: class Program_weight_tensor_parameter_214: name = "parameter_214" + original_name = "conv2d_99.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.159537") @@ -1963,6 +2178,7 @@ class Program_weight_tensor_parameter_214: class Program_weight_tensor_parameter_215: name = "parameter_215" + original_name = "batch_norm2d_94.b_0" shape = [192] dtype = "float32" data = None @@ -1970,6 +2186,7 @@ class Program_weight_tensor_parameter_215: class Program_weight_tensor_parameter_216: name = "parameter_216" + original_name = "batch_norm2d_94.w_0" shape = [192] dtype = "float32" min_val = float("1.0") @@ -1980,6 +2197,7 @@ class Program_weight_tensor_parameter_216: class Program_weight_tensor_parameter_217: name = "parameter_217" + original_name = "batch_norm2d_94.w_2" shape = [192] dtype = "float32" min_val = float("1.0") @@ -1990,6 +2208,7 @@ class Program_weight_tensor_parameter_217: class Program_weight_tensor_parameter_218: name = "parameter_218" + original_name = "batch_norm2d_94.w_1" shape = [192] dtype = "float32" data = None @@ -1997,6 +2216,7 @@ class Program_weight_tensor_parameter_218: class Program_weight_tensor_parameter_219: name = "parameter_219" + original_name = "conv2d_98.w_0" shape = [192, 192, 1, 1] dtype = "float32" min_val = float("-0.409606") @@ -2008,6 +2228,7 @@ class Program_weight_tensor_parameter_219: class Program_weight_tensor_parameter_220: name = "parameter_220" + original_name = "batch_norm2d_93.b_0" shape = [192] dtype = "float32" data = None @@ -2015,6 +2236,7 @@ class Program_weight_tensor_parameter_220: class Program_weight_tensor_parameter_221: name = "parameter_221" + original_name = "batch_norm2d_93.w_0" shape = [192] dtype = "float32" min_val = float("1.0") @@ -2025,6 +2247,7 @@ class Program_weight_tensor_parameter_221: class Program_weight_tensor_parameter_222: name = "parameter_222" + original_name = "batch_norm2d_93.w_2" shape = [192] dtype = "float32" min_val = float("1.0") @@ -2035,6 +2258,7 @@ class Program_weight_tensor_parameter_222: class Program_weight_tensor_parameter_223: name = "parameter_223" + original_name = "batch_norm2d_93.w_1" shape = [192] dtype = "float32" data = None @@ -2042,6 +2266,7 @@ class Program_weight_tensor_parameter_223: class Program_weight_tensor_parameter_224: name = "parameter_224" + original_name = "conv2d_97.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.158137") @@ -2053,6 +2278,7 @@ class Program_weight_tensor_parameter_224: class Program_weight_tensor_parameter_225: name = "parameter_225" + original_name = "batch_norm2d_92.b_0" shape = [192] dtype = "float32" data = None @@ -2060,6 +2286,7 @@ class Program_weight_tensor_parameter_225: class Program_weight_tensor_parameter_226: name = "parameter_226" + original_name = "batch_norm2d_92.w_0" shape = [192] dtype = "float32" min_val = float("1.0") @@ -2070,6 +2297,7 @@ class Program_weight_tensor_parameter_226: class Program_weight_tensor_parameter_227: name = "parameter_227" + original_name = "batch_norm2d_92.w_2" shape = [192] dtype = "float32" min_val = float("1.0") @@ -2080,6 +2308,7 @@ class Program_weight_tensor_parameter_227: class Program_weight_tensor_parameter_228: name = "parameter_228" + original_name = "batch_norm2d_92.w_1" shape = [192] dtype = "float32" data = None @@ -2087,6 +2316,7 @@ class Program_weight_tensor_parameter_228: class Program_weight_tensor_parameter_229: name = "parameter_229" + original_name = "conv2d_96.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.163613") @@ -2098,6 +2328,7 @@ class Program_weight_tensor_parameter_229: class Program_weight_tensor_parameter_230: name = "parameter_230" + original_name = "batch_norm2d_91.b_0" shape = [192] dtype = "float32" data = None @@ -2105,6 +2336,7 @@ class Program_weight_tensor_parameter_230: class Program_weight_tensor_parameter_231: name = "parameter_231" + original_name = "batch_norm2d_91.w_0" shape = [192] dtype = "float32" min_val = float("1.0") @@ -2115,6 +2347,7 @@ class Program_weight_tensor_parameter_231: class Program_weight_tensor_parameter_232: name = "parameter_232" + original_name = "batch_norm2d_91.w_2" shape = [192] dtype = "float32" min_val = float("1.0") @@ -2125,6 +2358,7 @@ class Program_weight_tensor_parameter_232: class Program_weight_tensor_parameter_233: name = "parameter_233" + original_name = "batch_norm2d_91.w_1" shape = [192] dtype = "float32" data = None @@ -2132,6 +2366,7 @@ class Program_weight_tensor_parameter_233: class Program_weight_tensor_parameter_234: name = "parameter_234" + original_name = "conv2d_95.w_0" shape = [192, 192, 1, 1] dtype = "float32" min_val = float("-0.39617") @@ -2143,6 +2378,7 @@ class Program_weight_tensor_parameter_234: class Program_weight_tensor_parameter_235: name = "parameter_235" + original_name = "batch_norm2d_90.b_0" shape = [192] dtype = "float32" data = None @@ -2150,6 +2386,7 @@ class Program_weight_tensor_parameter_235: class Program_weight_tensor_parameter_236: name = "parameter_236" + original_name = "batch_norm2d_90.w_0" shape = [192] dtype = "float32" min_val = float("1.0") @@ -2160,6 +2397,7 @@ class Program_weight_tensor_parameter_236: class Program_weight_tensor_parameter_237: name = "parameter_237" + original_name = "batch_norm2d_90.w_2" shape = [192] dtype = "float32" min_val = float("1.0") @@ -2170,6 +2408,7 @@ class Program_weight_tensor_parameter_237: class Program_weight_tensor_parameter_238: name = "parameter_238" + original_name = "batch_norm2d_90.w_1" shape = [192] dtype = "float32" data = None @@ -2177,6 +2416,7 @@ class Program_weight_tensor_parameter_238: class Program_weight_tensor_parameter_239: name = "parameter_239" + original_name = "conv2d_94.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.152365") @@ -2188,6 +2428,7 @@ class Program_weight_tensor_parameter_239: class Program_weight_tensor_parameter_240: name = "parameter_240" + original_name = "batch_norm2d_89.b_0" shape = [192] dtype = "float32" data = None @@ -2195,6 +2436,7 @@ class Program_weight_tensor_parameter_240: class Program_weight_tensor_parameter_241: name = "parameter_241" + original_name = "batch_norm2d_89.w_0" shape = [192] dtype = "float32" min_val = float("1.0") @@ -2205,6 +2447,7 @@ class Program_weight_tensor_parameter_241: class Program_weight_tensor_parameter_242: name = "parameter_242" + original_name = "batch_norm2d_89.w_2" shape = [192] dtype = "float32" min_val = float("1.0") @@ -2215,6 +2458,7 @@ class Program_weight_tensor_parameter_242: class Program_weight_tensor_parameter_243: name = "parameter_243" + original_name = "batch_norm2d_89.w_1" shape = [192] dtype = "float32" data = None @@ -2222,6 +2466,7 @@ class Program_weight_tensor_parameter_243: class Program_weight_tensor_parameter_244: name = "parameter_244" + original_name = "conv2d_93.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.141975") @@ -2233,6 +2478,7 @@ class Program_weight_tensor_parameter_244: class Program_weight_tensor_parameter_245: name = "parameter_245" + original_name = "batch_norm2d_88.b_0" shape = [192] dtype = "float32" data = None @@ -2240,6 +2486,7 @@ class Program_weight_tensor_parameter_245: class Program_weight_tensor_parameter_246: name = "parameter_246" + original_name = "batch_norm2d_88.w_0" shape = [192] dtype = "float32" min_val = float("1.0") @@ -2250,6 +2497,7 @@ class Program_weight_tensor_parameter_246: class Program_weight_tensor_parameter_247: name = "parameter_247" + original_name = "batch_norm2d_88.w_2" shape = [192] dtype = "float32" min_val = float("1.0") @@ -2260,6 +2508,7 @@ class Program_weight_tensor_parameter_247: class Program_weight_tensor_parameter_248: name = "parameter_248" + original_name = "batch_norm2d_88.w_1" shape = [192] dtype = "float32" data = None @@ -2267,6 +2516,7 @@ class Program_weight_tensor_parameter_248: class Program_weight_tensor_parameter_249: name = "parameter_249" + original_name = "conv2d_92.w_0" shape = [192, 896, 1, 1] dtype = "float32" min_val = float("-0.218845") @@ -2278,6 +2528,7 @@ class Program_weight_tensor_parameter_249: class Program_weight_tensor_parameter_250: name = "parameter_250" + original_name = "batch_norm2d_87.b_0" shape = [192] dtype = "float32" data = None @@ -2285,6 +2536,7 @@ class Program_weight_tensor_parameter_250: class Program_weight_tensor_parameter_251: name = "parameter_251" + original_name = "batch_norm2d_87.w_0" shape = [192] dtype = "float32" min_val = float("1.0") @@ -2295,6 +2547,7 @@ class Program_weight_tensor_parameter_251: class Program_weight_tensor_parameter_252: name = "parameter_252" + original_name = "batch_norm2d_87.w_2" shape = [192] dtype = "float32" min_val = float("1.0") @@ -2305,6 +2558,7 @@ class Program_weight_tensor_parameter_252: class Program_weight_tensor_parameter_253: name = "parameter_253" + original_name = "batch_norm2d_87.w_1" shape = [192] dtype = "float32" data = None @@ -2312,6 +2566,7 @@ class Program_weight_tensor_parameter_253: class Program_weight_tensor_parameter_254: name = "parameter_254" + original_name = "conv2d_91.w_0" shape = [192, 896, 1, 1] dtype = "float32" min_val = float("-0.236832") @@ -2323,6 +2578,7 @@ class Program_weight_tensor_parameter_254: class Program_weight_tensor_parameter_255: name = "parameter_255" + original_name = "batch_norm2d_86.b_0" shape = [384] dtype = "float32" data = None @@ -2330,6 +2586,7 @@ class Program_weight_tensor_parameter_255: class Program_weight_tensor_parameter_256: name = "parameter_256" + original_name = "batch_norm2d_86.w_0" shape = [384] dtype = "float32" min_val = float("1.0") @@ -2341,6 +2598,7 @@ class Program_weight_tensor_parameter_256: class Program_weight_tensor_parameter_257: name = "parameter_257" + original_name = "batch_norm2d_86.w_2" shape = [384] dtype = "float32" min_val = float("1.0") @@ -2352,6 +2610,7 @@ class Program_weight_tensor_parameter_257: class Program_weight_tensor_parameter_258: name = "parameter_258" + original_name = "batch_norm2d_86.w_1" shape = [384] dtype = "float32" data = None @@ -2359,6 +2618,7 @@ class Program_weight_tensor_parameter_258: class Program_weight_tensor_parameter_259: name = "parameter_259" + original_name = "conv2d_90.w_0" shape = [384, 768, 1, 1] dtype = "float32" min_val = float("-0.232199") @@ -2370,6 +2630,7 @@ class Program_weight_tensor_parameter_259: class Program_weight_tensor_parameter_260: name = "parameter_260" + original_name = "batch_norm2d_85.b_0" shape = [768] dtype = "float32" data = None @@ -2377,6 +2638,7 @@ class Program_weight_tensor_parameter_260: class Program_weight_tensor_parameter_261: name = "parameter_261" + original_name = "batch_norm2d_85.w_0" shape = [768] dtype = "float32" min_val = float("1.0") @@ -2387,6 +2649,7 @@ class Program_weight_tensor_parameter_261: class Program_weight_tensor_parameter_262: name = "parameter_262" + original_name = "batch_norm2d_85.w_2" shape = [768] dtype = "float32" min_val = float("1.0") @@ -2397,6 +2660,7 @@ class Program_weight_tensor_parameter_262: class Program_weight_tensor_parameter_263: name = "parameter_263" + original_name = "batch_norm2d_85.w_1" shape = [768] dtype = "float32" data = None @@ -2404,6 +2668,7 @@ class Program_weight_tensor_parameter_263: class Program_weight_tensor_parameter_264: name = "parameter_264" + original_name = "conv2d_89.w_0" shape = [768, 768, 1, 1] dtype = "float32" min_val = float("-0.248379") @@ -2415,6 +2680,7 @@ class Program_weight_tensor_parameter_264: class Program_weight_tensor_parameter_265: name = "parameter_265" + original_name = "batch_norm2d_84.b_0" shape = [384] dtype = "float32" data = None @@ -2422,6 +2688,7 @@ class Program_weight_tensor_parameter_265: class Program_weight_tensor_parameter_266: name = "parameter_266" + original_name = "batch_norm2d_84.w_0" shape = [384] dtype = "float32" min_val = float("1.0") @@ -2433,6 +2700,7 @@ class Program_weight_tensor_parameter_266: class Program_weight_tensor_parameter_267: name = "parameter_267" + original_name = "batch_norm2d_84.w_2" shape = [384] dtype = "float32" min_val = float("1.0") @@ -2444,6 +2712,7 @@ class Program_weight_tensor_parameter_267: class Program_weight_tensor_parameter_268: name = "parameter_268" + original_name = "batch_norm2d_84.w_1" shape = [384] dtype = "float32" data = None @@ -2451,6 +2720,7 @@ class Program_weight_tensor_parameter_268: class Program_weight_tensor_parameter_269: name = "parameter_269" + original_name = "conv2d_88.w_0" shape = [384, 384, 1, 1] dtype = "float32" min_val = float("-0.30031") @@ -2462,6 +2732,7 @@ class Program_weight_tensor_parameter_269: class Program_weight_tensor_parameter_270: name = "parameter_270" + original_name = "batch_norm2d_83.b_0" shape = [384] dtype = "float32" data = None @@ -2469,6 +2740,7 @@ class Program_weight_tensor_parameter_270: class Program_weight_tensor_parameter_271: name = "parameter_271" + original_name = "batch_norm2d_83.w_0" shape = [384] dtype = "float32" min_val = float("1.0") @@ -2480,6 +2752,7 @@ class Program_weight_tensor_parameter_271: class Program_weight_tensor_parameter_272: name = "parameter_272" + original_name = "batch_norm2d_83.w_2" shape = [384] dtype = "float32" min_val = float("1.0") @@ -2491,6 +2764,7 @@ class Program_weight_tensor_parameter_272: class Program_weight_tensor_parameter_273: name = "parameter_273" + original_name = "batch_norm2d_83.w_1" shape = [384] dtype = "float32" data = None @@ -2498,6 +2772,7 @@ class Program_weight_tensor_parameter_273: class Program_weight_tensor_parameter_274: name = "parameter_274" + original_name = "conv2d_87.w_0" shape = [384, 384, 3, 3] dtype = "float32" min_val = float("-0.123888") @@ -2509,6 +2784,7 @@ class Program_weight_tensor_parameter_274: class Program_weight_tensor_parameter_275: name = "parameter_275" + original_name = "batch_norm2d_82.b_0" shape = [384] dtype = "float32" data = None @@ -2516,6 +2792,7 @@ class Program_weight_tensor_parameter_275: class Program_weight_tensor_parameter_276: name = "parameter_276" + original_name = "batch_norm2d_82.w_0" shape = [384] dtype = "float32" min_val = float("1.0") @@ -2527,6 +2804,7 @@ class Program_weight_tensor_parameter_276: class Program_weight_tensor_parameter_277: name = "parameter_277" + original_name = "batch_norm2d_82.w_2" shape = [384] dtype = "float32" min_val = float("1.0") @@ -2538,6 +2816,7 @@ class Program_weight_tensor_parameter_277: class Program_weight_tensor_parameter_278: name = "parameter_278" + original_name = "batch_norm2d_82.w_1" shape = [384] dtype = "float32" data = None @@ -2545,6 +2824,7 @@ class Program_weight_tensor_parameter_278: class Program_weight_tensor_parameter_279: name = "parameter_279" + original_name = "conv2d_86.w_0" shape = [384, 384, 3, 3] dtype = "float32" min_val = float("-0.117054") @@ -2556,6 +2836,7 @@ class Program_weight_tensor_parameter_279: class Program_weight_tensor_parameter_280: name = "parameter_280" + original_name = "batch_norm2d_81.b_0" shape = [384] dtype = "float32" data = None @@ -2563,6 +2844,7 @@ class Program_weight_tensor_parameter_280: class Program_weight_tensor_parameter_281: name = "parameter_281" + original_name = "batch_norm2d_81.w_0" shape = [384] dtype = "float32" min_val = float("1.0") @@ -2574,6 +2856,7 @@ class Program_weight_tensor_parameter_281: class Program_weight_tensor_parameter_282: name = "parameter_282" + original_name = "batch_norm2d_81.w_2" shape = [384] dtype = "float32" min_val = float("1.0") @@ -2585,6 +2868,7 @@ class Program_weight_tensor_parameter_282: class Program_weight_tensor_parameter_283: name = "parameter_283" + original_name = "batch_norm2d_81.w_1" shape = [384] dtype = "float32" data = None @@ -2592,6 +2876,7 @@ class Program_weight_tensor_parameter_283: class Program_weight_tensor_parameter_284: name = "parameter_284" + original_name = "conv2d_85.w_0" shape = [384, 1536, 1, 1] dtype = "float32" min_val = float("-0.171136") @@ -2603,6 +2888,7 @@ class Program_weight_tensor_parameter_284: class Program_weight_tensor_parameter_285: name = "parameter_285" + original_name = "batch_norm2d_80.b_0" shape = [384] dtype = "float32" data = None @@ -2610,6 +2896,7 @@ class Program_weight_tensor_parameter_285: class Program_weight_tensor_parameter_286: name = "parameter_286" + original_name = "batch_norm2d_80.w_0" shape = [384] dtype = "float32" min_val = float("1.0") @@ -2621,6 +2908,7 @@ class Program_weight_tensor_parameter_286: class Program_weight_tensor_parameter_287: name = "parameter_287" + original_name = "batch_norm2d_80.w_2" shape = [384] dtype = "float32" min_val = float("1.0") @@ -2632,6 +2920,7 @@ class Program_weight_tensor_parameter_287: class Program_weight_tensor_parameter_288: name = "parameter_288" + original_name = "batch_norm2d_80.w_1" shape = [384] dtype = "float32" data = None @@ -2639,6 +2928,7 @@ class Program_weight_tensor_parameter_288: class Program_weight_tensor_parameter_289: name = "parameter_289" + original_name = "conv2d_84.w_0" shape = [384, 384, 1, 1] dtype = "float32" min_val = float("-0.381462") @@ -2650,6 +2940,7 @@ class Program_weight_tensor_parameter_289: class Program_weight_tensor_parameter_290: name = "parameter_290" + original_name = "batch_norm2d_79.b_0" shape = [384] dtype = "float32" data = None @@ -2657,6 +2948,7 @@ class Program_weight_tensor_parameter_290: class Program_weight_tensor_parameter_291: name = "parameter_291" + original_name = "batch_norm2d_79.w_0" shape = [384] dtype = "float32" min_val = float("1.0") @@ -2668,6 +2960,7 @@ class Program_weight_tensor_parameter_291: class Program_weight_tensor_parameter_292: name = "parameter_292" + original_name = "batch_norm2d_79.w_2" shape = [384] dtype = "float32" min_val = float("1.0") @@ -2679,6 +2972,7 @@ class Program_weight_tensor_parameter_292: class Program_weight_tensor_parameter_293: name = "parameter_293" + original_name = "batch_norm2d_79.w_1" shape = [384] dtype = "float32" data = None @@ -2686,6 +2980,7 @@ class Program_weight_tensor_parameter_293: class Program_weight_tensor_parameter_294: name = "parameter_294" + original_name = "conv2d_83.w_0" shape = [384, 384, 3, 3] dtype = "float32" min_val = float("-0.113052") @@ -2697,6 +2992,7 @@ class Program_weight_tensor_parameter_294: class Program_weight_tensor_parameter_295: name = "parameter_295" + original_name = "batch_norm2d_78.b_0" shape = [384] dtype = "float32" data = None @@ -2704,6 +3000,7 @@ class Program_weight_tensor_parameter_295: class Program_weight_tensor_parameter_296: name = "parameter_296" + original_name = "batch_norm2d_78.w_0" shape = [384] dtype = "float32" min_val = float("1.0") @@ -2715,6 +3012,7 @@ class Program_weight_tensor_parameter_296: class Program_weight_tensor_parameter_297: name = "parameter_297" + original_name = "batch_norm2d_78.w_2" shape = [384] dtype = "float32" min_val = float("1.0") @@ -2726,6 +3024,7 @@ class Program_weight_tensor_parameter_297: class Program_weight_tensor_parameter_298: name = "parameter_298" + original_name = "batch_norm2d_78.w_1" shape = [384] dtype = "float32" data = None @@ -2733,6 +3032,7 @@ class Program_weight_tensor_parameter_298: class Program_weight_tensor_parameter_299: name = "parameter_299" + original_name = "conv2d_82.w_0" shape = [384, 384, 3, 3] dtype = "float32" min_val = float("-0.122883") @@ -2744,6 +3044,7 @@ class Program_weight_tensor_parameter_299: class Program_weight_tensor_parameter_300: name = "parameter_300" + original_name = "batch_norm2d_77.b_0" shape = [384] dtype = "float32" data = None @@ -2751,6 +3052,7 @@ class Program_weight_tensor_parameter_300: class Program_weight_tensor_parameter_301: name = "parameter_301" + original_name = "batch_norm2d_77.w_0" shape = [384] dtype = "float32" min_val = float("1.0") @@ -2762,6 +3064,7 @@ class Program_weight_tensor_parameter_301: class Program_weight_tensor_parameter_302: name = "parameter_302" + original_name = "batch_norm2d_77.w_2" shape = [384] dtype = "float32" min_val = float("1.0") @@ -2773,6 +3076,7 @@ class Program_weight_tensor_parameter_302: class Program_weight_tensor_parameter_303: name = "parameter_303" + original_name = "batch_norm2d_77.w_1" shape = [384] dtype = "float32" data = None @@ -2780,6 +3084,7 @@ class Program_weight_tensor_parameter_303: class Program_weight_tensor_parameter_304: name = "parameter_304" + original_name = "conv2d_81.w_0" shape = [384, 384, 1, 1] dtype = "float32" min_val = float("-0.308828") @@ -2791,6 +3096,7 @@ class Program_weight_tensor_parameter_304: class Program_weight_tensor_parameter_305: name = "parameter_305" + original_name = "batch_norm2d_76.b_0" shape = [384] dtype = "float32" data = None @@ -2798,6 +3104,7 @@ class Program_weight_tensor_parameter_305: class Program_weight_tensor_parameter_306: name = "parameter_306" + original_name = "batch_norm2d_76.w_0" shape = [384] dtype = "float32" min_val = float("1.0") @@ -2809,6 +3116,7 @@ class Program_weight_tensor_parameter_306: class Program_weight_tensor_parameter_307: name = "parameter_307" + original_name = "batch_norm2d_76.w_2" shape = [384] dtype = "float32" min_val = float("1.0") @@ -2820,6 +3128,7 @@ class Program_weight_tensor_parameter_307: class Program_weight_tensor_parameter_308: name = "parameter_308" + original_name = "batch_norm2d_76.w_1" shape = [384] dtype = "float32" data = None @@ -2827,6 +3136,7 @@ class Program_weight_tensor_parameter_308: class Program_weight_tensor_parameter_309: name = "parameter_309" + original_name = "conv2d_80.w_0" shape = [384, 384, 3, 3] dtype = "float32" min_val = float("-0.109728") @@ -2838,6 +3148,7 @@ class Program_weight_tensor_parameter_309: class Program_weight_tensor_parameter_310: name = "parameter_310" + original_name = "batch_norm2d_75.b_0" shape = [384] dtype = "float32" data = None @@ -2845,6 +3156,7 @@ class Program_weight_tensor_parameter_310: class Program_weight_tensor_parameter_311: name = "parameter_311" + original_name = "batch_norm2d_75.w_0" shape = [384] dtype = "float32" min_val = float("1.0") @@ -2856,6 +3168,7 @@ class Program_weight_tensor_parameter_311: class Program_weight_tensor_parameter_312: name = "parameter_312" + original_name = "batch_norm2d_75.w_2" shape = [384] dtype = "float32" min_val = float("1.0") @@ -2867,6 +3180,7 @@ class Program_weight_tensor_parameter_312: class Program_weight_tensor_parameter_313: name = "parameter_313" + original_name = "batch_norm2d_75.w_1" shape = [384] dtype = "float32" data = None @@ -2874,6 +3188,7 @@ class Program_weight_tensor_parameter_313: class Program_weight_tensor_parameter_314: name = "parameter_314" + original_name = "conv2d_79.w_0" shape = [384, 384, 3, 3] dtype = "float32" min_val = float("-0.111787") @@ -2885,6 +3200,7 @@ class Program_weight_tensor_parameter_314: class Program_weight_tensor_parameter_315: name = "parameter_315" + original_name = "batch_norm2d_74.b_0" shape = [384] dtype = "float32" data = None @@ -2892,6 +3208,7 @@ class Program_weight_tensor_parameter_315: class Program_weight_tensor_parameter_316: name = "parameter_316" + original_name = "batch_norm2d_74.w_0" shape = [384] dtype = "float32" min_val = float("1.0") @@ -2903,6 +3220,7 @@ class Program_weight_tensor_parameter_316: class Program_weight_tensor_parameter_317: name = "parameter_317" + original_name = "batch_norm2d_74.w_2" shape = [384] dtype = "float32" min_val = float("1.0") @@ -2914,6 +3232,7 @@ class Program_weight_tensor_parameter_317: class Program_weight_tensor_parameter_318: name = "parameter_318" + original_name = "batch_norm2d_74.w_1" shape = [384] dtype = "float32" data = None @@ -2921,6 +3240,7 @@ class Program_weight_tensor_parameter_318: class Program_weight_tensor_parameter_319: name = "parameter_319" + original_name = "conv2d_78.w_0" shape = [384, 1024, 1, 1] dtype = "float32" min_val = float("-0.223901") @@ -2932,6 +3252,7 @@ class Program_weight_tensor_parameter_319: class Program_weight_tensor_parameter_320: name = "parameter_320" + original_name = "batch_norm2d_73.b_0" shape = [384] dtype = "float32" data = None @@ -2939,6 +3260,7 @@ class Program_weight_tensor_parameter_320: class Program_weight_tensor_parameter_321: name = "parameter_321" + original_name = "batch_norm2d_73.w_0" shape = [384] dtype = "float32" min_val = float("1.0") @@ -2950,6 +3272,7 @@ class Program_weight_tensor_parameter_321: class Program_weight_tensor_parameter_322: name = "parameter_322" + original_name = "batch_norm2d_73.w_2" shape = [384] dtype = "float32" min_val = float("1.0") @@ -2961,6 +3284,7 @@ class Program_weight_tensor_parameter_322: class Program_weight_tensor_parameter_323: name = "parameter_323" + original_name = "batch_norm2d_73.w_1" shape = [384] dtype = "float32" data = None @@ -2968,6 +3292,7 @@ class Program_weight_tensor_parameter_323: class Program_weight_tensor_parameter_324: name = "parameter_324" + original_name = "conv2d_77.w_0" shape = [384, 1024, 1, 1] dtype = "float32" min_val = float("-0.204276") @@ -2979,6 +3304,7 @@ class Program_weight_tensor_parameter_324: class Program_weight_tensor_parameter_325: name = "parameter_325" + original_name = "batch_norm2d_72.b_0" shape = [1024] dtype = "float32" min_val = float("-3.74704") @@ -2990,6 +3316,7 @@ class Program_weight_tensor_parameter_325: class Program_weight_tensor_parameter_326: name = "parameter_326" + original_name = "batch_norm2d_72.w_0" shape = [1024] dtype = "float32" min_val = float("1.59825") @@ -3001,6 +3328,7 @@ class Program_weight_tensor_parameter_326: class Program_weight_tensor_parameter_327: name = "parameter_327" + original_name = "batch_norm2d_72.w_2" shape = [1024] dtype = "float32" min_val = float("0.350855") @@ -3012,6 +3340,7 @@ class Program_weight_tensor_parameter_327: class Program_weight_tensor_parameter_328: name = "parameter_328" + original_name = "batch_norm2d_72.w_1" shape = [1024] dtype = "float32" min_val = float("-0.827043") @@ -3023,6 +3352,7 @@ class Program_weight_tensor_parameter_328: class Program_weight_tensor_parameter_329: name = "parameter_329" + original_name = "conv2d_76.w_0" shape = [1024, 768, 1, 1] dtype = "float32" min_val = float("-0.630881") @@ -3034,6 +3364,7 @@ class Program_weight_tensor_parameter_329: class Program_weight_tensor_parameter_330: name = "parameter_330" + original_name = "conv2d_75.b_0" shape = [768] dtype = "float32" min_val = float("-1.84619") @@ -3045,6 +3376,7 @@ class Program_weight_tensor_parameter_330: class Program_weight_tensor_parameter_331: name = "parameter_331" + original_name = "conv2d_75.w_0" shape = [768, 768, 1, 1] dtype = "float32" min_val = float("-0.750361") @@ -3056,6 +3388,7 @@ class Program_weight_tensor_parameter_331: class Program_weight_tensor_parameter_332: name = "parameter_332" + original_name = "batch_norm2d_71.b_0" shape = [384] dtype = "float32" min_val = float("-1.7711") @@ -3067,6 +3400,7 @@ class Program_weight_tensor_parameter_332: class Program_weight_tensor_parameter_333: name = "parameter_333" + original_name = "batch_norm2d_71.w_0" shape = [384] dtype = "float32" min_val = float("0.181179") @@ -3078,6 +3412,7 @@ class Program_weight_tensor_parameter_333: class Program_weight_tensor_parameter_334: name = "parameter_334" + original_name = "batch_norm2d_71.w_2" shape = [384] dtype = "float32" min_val = float("0.00704936") @@ -3089,6 +3424,7 @@ class Program_weight_tensor_parameter_334: class Program_weight_tensor_parameter_335: name = "parameter_335" + original_name = "batch_norm2d_71.w_1" shape = [384] dtype = "float32" min_val = float("-1.19051") @@ -3100,6 +3436,7 @@ class Program_weight_tensor_parameter_335: class Program_weight_tensor_parameter_336: name = "parameter_336" + original_name = "conv2d_74.w_0" shape = [384, 384, 1, 1] dtype = "float32" min_val = float("-0.584765") @@ -3111,6 +3448,7 @@ class Program_weight_tensor_parameter_336: class Program_weight_tensor_parameter_337: name = "parameter_337" + original_name = "batch_norm2d_70.b_0" shape = [384] dtype = "float32" min_val = float("-1.7711") @@ -3122,6 +3460,7 @@ class Program_weight_tensor_parameter_337: class Program_weight_tensor_parameter_338: name = "parameter_338" + original_name = "batch_norm2d_70.w_0" shape = [384] dtype = "float32" min_val = float("0.255804") @@ -3133,6 +3472,7 @@ class Program_weight_tensor_parameter_338: class Program_weight_tensor_parameter_339: name = "parameter_339" + original_name = "batch_norm2d_70.w_2" shape = [384] dtype = "float32" min_val = float("0.0698564") @@ -3144,6 +3484,7 @@ class Program_weight_tensor_parameter_339: class Program_weight_tensor_parameter_340: name = "parameter_340" + original_name = "batch_norm2d_70.w_1" shape = [384] dtype = "float32" min_val = float("-1.27275") @@ -3155,6 +3496,7 @@ class Program_weight_tensor_parameter_340: class Program_weight_tensor_parameter_341: name = "parameter_341" + original_name = "conv2d_73.w_0" shape = [384, 384, 3, 3] dtype = "float32" min_val = float("-0.304078") @@ -3166,6 +3508,7 @@ class Program_weight_tensor_parameter_341: class Program_weight_tensor_parameter_342: name = "parameter_342" + original_name = "batch_norm2d_69.b_0" shape = [384] dtype = "float32" min_val = float("-2.55839") @@ -3177,6 +3520,7 @@ class Program_weight_tensor_parameter_342: class Program_weight_tensor_parameter_343: name = "parameter_343" + original_name = "batch_norm2d_69.w_0" shape = [384] dtype = "float32" min_val = float("0.53528") @@ -3188,6 +3532,7 @@ class Program_weight_tensor_parameter_343: class Program_weight_tensor_parameter_344: name = "parameter_344" + original_name = "batch_norm2d_69.w_2" shape = [384] dtype = "float32" min_val = float("3.97934") @@ -3199,6 +3544,7 @@ class Program_weight_tensor_parameter_344: class Program_weight_tensor_parameter_345: name = "parameter_345" + original_name = "batch_norm2d_69.w_1" shape = [384] dtype = "float32" min_val = float("-4.97699") @@ -3210,6 +3556,7 @@ class Program_weight_tensor_parameter_345: class Program_weight_tensor_parameter_346: name = "parameter_346" + original_name = "conv2d_72.w_0" shape = [384, 384, 3, 3] dtype = "float32" min_val = float("-0.196336") @@ -3221,6 +3568,7 @@ class Program_weight_tensor_parameter_346: class Program_weight_tensor_parameter_347: name = "parameter_347" + original_name = "batch_norm2d_68.b_0" shape = [384] dtype = "float32" min_val = float("-1.95623") @@ -3232,6 +3580,7 @@ class Program_weight_tensor_parameter_347: class Program_weight_tensor_parameter_348: name = "parameter_348" + original_name = "batch_norm2d_68.w_0" shape = [384] dtype = "float32" min_val = float("0.136667") @@ -3243,6 +3592,7 @@ class Program_weight_tensor_parameter_348: class Program_weight_tensor_parameter_349: name = "parameter_349" + original_name = "batch_norm2d_68.w_2" shape = [384] dtype = "float32" min_val = float("0.0126792") @@ -3254,6 +3604,7 @@ class Program_weight_tensor_parameter_349: class Program_weight_tensor_parameter_350: name = "parameter_350" + original_name = "batch_norm2d_68.w_1" shape = [384] dtype = "float32" min_val = float("-0.846751") @@ -3265,6 +3616,7 @@ class Program_weight_tensor_parameter_350: class Program_weight_tensor_parameter_351: name = "parameter_351" + original_name = "conv2d_71.w_0" shape = [384, 384, 1, 1] dtype = "float32" min_val = float("-0.395") @@ -3276,6 +3628,7 @@ class Program_weight_tensor_parameter_351: class Program_weight_tensor_parameter_352: name = "parameter_352" + original_name = "batch_norm2d_67.b_0" shape = [384] dtype = "float32" min_val = float("-1.95623") @@ -3287,6 +3640,7 @@ class Program_weight_tensor_parameter_352: class Program_weight_tensor_parameter_353: name = "parameter_353" + original_name = "batch_norm2d_67.w_0" shape = [384] dtype = "float32" min_val = float("0.476039") @@ -3298,6 +3652,7 @@ class Program_weight_tensor_parameter_353: class Program_weight_tensor_parameter_354: name = "parameter_354" + original_name = "batch_norm2d_67.w_2" shape = [384] dtype = "float32" min_val = float("0.0993289") @@ -3309,6 +3664,7 @@ class Program_weight_tensor_parameter_354: class Program_weight_tensor_parameter_355: name = "parameter_355" + original_name = "batch_norm2d_67.w_1" shape = [384] dtype = "float32" min_val = float("-2.91041") @@ -3320,6 +3676,7 @@ class Program_weight_tensor_parameter_355: class Program_weight_tensor_parameter_356: name = "parameter_356" + original_name = "conv2d_70.w_0" shape = [384, 384, 3, 3] dtype = "float32" min_val = float("-0.548844") @@ -3331,6 +3688,7 @@ class Program_weight_tensor_parameter_356: class Program_weight_tensor_parameter_357: name = "parameter_357" + original_name = "batch_norm2d_66.b_0" shape = [384] dtype = "float32" min_val = float("-2.4119") @@ -3342,6 +3700,7 @@ class Program_weight_tensor_parameter_357: class Program_weight_tensor_parameter_358: name = "parameter_358" + original_name = "batch_norm2d_66.w_0" shape = [384] dtype = "float32" min_val = float("0.617352") @@ -3353,6 +3712,7 @@ class Program_weight_tensor_parameter_358: class Program_weight_tensor_parameter_359: name = "parameter_359" + original_name = "batch_norm2d_66.w_2" shape = [384] dtype = "float32" min_val = float("3.32112") @@ -3364,6 +3724,7 @@ class Program_weight_tensor_parameter_359: class Program_weight_tensor_parameter_360: name = "parameter_360" + original_name = "batch_norm2d_66.w_1" shape = [384] dtype = "float32" min_val = float("-6.04079") @@ -3375,6 +3736,7 @@ class Program_weight_tensor_parameter_360: class Program_weight_tensor_parameter_361: name = "parameter_361" + original_name = "conv2d_69.w_0" shape = [384, 384, 3, 3] dtype = "float32" min_val = float("-0.18178") @@ -3386,6 +3748,7 @@ class Program_weight_tensor_parameter_361: class Program_weight_tensor_parameter_362: name = "parameter_362" + original_name = "batch_norm2d_65.b_0" shape = [384] dtype = "float32" min_val = float("-1.87273") @@ -3397,6 +3760,7 @@ class Program_weight_tensor_parameter_362: class Program_weight_tensor_parameter_363: name = "parameter_363" + original_name = "batch_norm2d_65.w_0" shape = [384] dtype = "float32" min_val = float("0.0435078") @@ -3408,6 +3772,7 @@ class Program_weight_tensor_parameter_363: class Program_weight_tensor_parameter_364: name = "parameter_364" + original_name = "batch_norm2d_65.w_2" shape = [384] dtype = "float32" min_val = float("0.00471868") @@ -3419,6 +3784,7 @@ class Program_weight_tensor_parameter_364: class Program_weight_tensor_parameter_365: name = "parameter_365" + original_name = "batch_norm2d_65.w_1" shape = [384] dtype = "float32" min_val = float("-1.03153") @@ -3430,6 +3796,7 @@ class Program_weight_tensor_parameter_365: class Program_weight_tensor_parameter_366: name = "parameter_366" + original_name = "conv2d_68.w_0" shape = [384, 384, 1, 1] dtype = "float32" min_val = float("-0.387213") @@ -3441,6 +3808,7 @@ class Program_weight_tensor_parameter_366: class Program_weight_tensor_parameter_367: name = "parameter_367" + original_name = "batch_norm2d_64.b_0" shape = [384] dtype = "float32" min_val = float("-1.87273") @@ -3452,6 +3820,7 @@ class Program_weight_tensor_parameter_367: class Program_weight_tensor_parameter_368: name = "parameter_368" + original_name = "batch_norm2d_64.w_0" shape = [384] dtype = "float32" min_val = float("0.455155") @@ -3463,6 +3832,7 @@ class Program_weight_tensor_parameter_368: class Program_weight_tensor_parameter_369: name = "parameter_369" + original_name = "batch_norm2d_64.w_2" shape = [384] dtype = "float32" min_val = float("0.120887") @@ -3474,6 +3844,7 @@ class Program_weight_tensor_parameter_369: class Program_weight_tensor_parameter_370: name = "parameter_370" + original_name = "batch_norm2d_64.w_1" shape = [384] dtype = "float32" min_val = float("-1.6152") @@ -3485,6 +3856,7 @@ class Program_weight_tensor_parameter_370: class Program_weight_tensor_parameter_371: name = "parameter_371" + original_name = "conv2d_67.w_0" shape = [384, 384, 3, 3] dtype = "float32" min_val = float("-0.353474") @@ -3496,6 +3868,7 @@ class Program_weight_tensor_parameter_371: class Program_weight_tensor_parameter_372: name = "parameter_372" + original_name = "batch_norm2d_63.b_0" shape = [384] dtype = "float32" min_val = float("-2.14605") @@ -3507,6 +3880,7 @@ class Program_weight_tensor_parameter_372: class Program_weight_tensor_parameter_373: name = "parameter_373" + original_name = "batch_norm2d_63.w_0" shape = [384] dtype = "float32" min_val = float("0.774001") @@ -3518,6 +3892,7 @@ class Program_weight_tensor_parameter_373: class Program_weight_tensor_parameter_374: name = "parameter_374" + original_name = "batch_norm2d_63.w_2" shape = [384] dtype = "float32" min_val = float("2.64922") @@ -3529,6 +3904,7 @@ class Program_weight_tensor_parameter_374: class Program_weight_tensor_parameter_375: name = "parameter_375" + original_name = "batch_norm2d_63.w_1" shape = [384] dtype = "float32" min_val = float("-5.25263") @@ -3540,6 +3916,7 @@ class Program_weight_tensor_parameter_375: class Program_weight_tensor_parameter_376: name = "parameter_376" + original_name = "conv2d_66.w_0" shape = [384, 384, 3, 3] dtype = "float32" min_val = float("-0.206734") @@ -3551,6 +3928,7 @@ class Program_weight_tensor_parameter_376: class Program_weight_tensor_parameter_377: name = "parameter_377" + original_name = "batch_norm2d_62.b_0" shape = [384] dtype = "float32" min_val = float("-2.91797") @@ -3562,6 +3940,7 @@ class Program_weight_tensor_parameter_377: class Program_weight_tensor_parameter_378: name = "parameter_378" + original_name = "batch_norm2d_62.w_0" shape = [384] dtype = "float32" min_val = float("0.867847") @@ -3573,6 +3952,7 @@ class Program_weight_tensor_parameter_378: class Program_weight_tensor_parameter_379: name = "parameter_379" + original_name = "batch_norm2d_62.w_2" shape = [384] dtype = "float32" min_val = float("0.262423") @@ -3584,6 +3964,7 @@ class Program_weight_tensor_parameter_379: class Program_weight_tensor_parameter_380: name = "parameter_380" + original_name = "batch_norm2d_62.w_1" shape = [384] dtype = "float32" min_val = float("-2.60546") @@ -3595,6 +3976,7 @@ class Program_weight_tensor_parameter_380: class Program_weight_tensor_parameter_381: name = "parameter_381" + original_name = "conv2d_65.w_0" shape = [384, 768, 1, 1] dtype = "float32" min_val = float("-0.567572") @@ -3606,6 +3988,7 @@ class Program_weight_tensor_parameter_381: class Program_weight_tensor_parameter_382: name = "parameter_382" + original_name = "batch_norm2d_61.b_0" shape = [384] dtype = "float32" min_val = float("-2.23836") @@ -3617,6 +4000,7 @@ class Program_weight_tensor_parameter_382: class Program_weight_tensor_parameter_383: name = "parameter_383" + original_name = "batch_norm2d_61.w_0" shape = [384] dtype = "float32" min_val = float("0.651326") @@ -3628,6 +4012,7 @@ class Program_weight_tensor_parameter_383: class Program_weight_tensor_parameter_384: name = "parameter_384" + original_name = "batch_norm2d_61.w_2" shape = [384] dtype = "float32" min_val = float("0.286499") @@ -3639,6 +4024,7 @@ class Program_weight_tensor_parameter_384: class Program_weight_tensor_parameter_385: name = "parameter_385" + original_name = "batch_norm2d_61.w_1" shape = [384] dtype = "float32" min_val = float("-0.166745") @@ -3650,6 +4036,7 @@ class Program_weight_tensor_parameter_385: class Program_weight_tensor_parameter_386: name = "parameter_386" + original_name = "conv2d_64.w_0" shape = [384, 768, 1, 1] dtype = "float32" min_val = float("-1.33249") @@ -3661,6 +4048,7 @@ class Program_weight_tensor_parameter_386: class Program_weight_tensor_parameter_387: name = "parameter_387" + original_name = "batch_norm2d_60.b_0" shape = [768] dtype = "float32" min_val = float("-2.40044") @@ -3672,6 +4060,7 @@ class Program_weight_tensor_parameter_387: class Program_weight_tensor_parameter_388: name = "parameter_388" + original_name = "batch_norm2d_60.w_0" shape = [768] dtype = "float32" min_val = float("0.507329") @@ -3683,6 +4072,7 @@ class Program_weight_tensor_parameter_388: class Program_weight_tensor_parameter_389: name = "parameter_389" + original_name = "batch_norm2d_60.w_2" shape = [768] dtype = "float32" min_val = float("0.726932") @@ -3694,6 +4084,7 @@ class Program_weight_tensor_parameter_389: class Program_weight_tensor_parameter_390: name = "parameter_390" + original_name = "batch_norm2d_60.w_1" shape = [768] dtype = "float32" min_val = float("-1.00842") @@ -3705,6 +4096,7 @@ class Program_weight_tensor_parameter_390: class Program_weight_tensor_parameter_391: name = "parameter_391" + original_name = "conv2d_63.w_0" shape = [768, 512, 3, 3] dtype = "float32" min_val = float("-0.539869") @@ -3716,6 +4108,7 @@ class Program_weight_tensor_parameter_391: class Program_weight_tensor_parameter_392: name = "parameter_392" + original_name = "batch_norm2d_59.b_0" shape = [512] dtype = "float32" min_val = float("-3.38221") @@ -3727,6 +4120,7 @@ class Program_weight_tensor_parameter_392: class Program_weight_tensor_parameter_393: name = "parameter_393" + original_name = "batch_norm2d_59.w_0" shape = [512] dtype = "float32" min_val = float("0.311062") @@ -3738,6 +4132,7 @@ class Program_weight_tensor_parameter_393: class Program_weight_tensor_parameter_394: name = "parameter_394" + original_name = "batch_norm2d_59.w_2" shape = [512] dtype = "float32" min_val = float("1.06478") @@ -3749,6 +4144,7 @@ class Program_weight_tensor_parameter_394: class Program_weight_tensor_parameter_395: name = "parameter_395" + original_name = "batch_norm2d_59.w_1" shape = [512] dtype = "float32" min_val = float("-1.49863") @@ -3760,6 +4156,7 @@ class Program_weight_tensor_parameter_395: class Program_weight_tensor_parameter_396: name = "parameter_396" + original_name = "conv2d_62.w_0" shape = [512, 384, 1, 1] dtype = "float32" min_val = float("-0.760857") @@ -3771,6 +4168,7 @@ class Program_weight_tensor_parameter_396: class Program_weight_tensor_parameter_397: name = "parameter_397" + original_name = "conv2d_61.b_0" shape = [384] dtype = "float32" min_val = float("-0.507828") @@ -3782,6 +4180,7 @@ class Program_weight_tensor_parameter_397: class Program_weight_tensor_parameter_398: name = "parameter_398" + original_name = "conv2d_61.w_0" shape = [384, 384, 1, 1] dtype = "float32" min_val = float("-1.34497") @@ -3793,6 +4192,7 @@ class Program_weight_tensor_parameter_398: class Program_weight_tensor_parameter_399: name = "parameter_399" + original_name = "batch_norm2d_58.b_0" shape = [192] dtype = "float32" min_val = float("-2.01897") @@ -3804,6 +4204,7 @@ class Program_weight_tensor_parameter_399: class Program_weight_tensor_parameter_400: name = "parameter_400" + original_name = "batch_norm2d_58.w_0" shape = [192] dtype = "float32" min_val = float("0.0468001") @@ -3815,6 +4216,7 @@ class Program_weight_tensor_parameter_400: class Program_weight_tensor_parameter_401: name = "parameter_401" + original_name = "batch_norm2d_58.w_2" shape = [192] dtype = "float32" min_val = float("0.00434022") @@ -3826,6 +4228,7 @@ class Program_weight_tensor_parameter_401: class Program_weight_tensor_parameter_402: name = "parameter_402" + original_name = "batch_norm2d_58.w_1" shape = [192] dtype = "float32" min_val = float("-0.282835") @@ -3837,6 +4240,7 @@ class Program_weight_tensor_parameter_402: class Program_weight_tensor_parameter_403: name = "parameter_403" + original_name = "conv2d_60.w_0" shape = [192, 192, 1, 1] dtype = "float32" min_val = float("-0.559419") @@ -3848,6 +4252,7 @@ class Program_weight_tensor_parameter_403: class Program_weight_tensor_parameter_404: name = "parameter_404" + original_name = "batch_norm2d_57.b_0" shape = [192] dtype = "float32" min_val = float("-2.01897") @@ -3859,6 +4264,7 @@ class Program_weight_tensor_parameter_404: class Program_weight_tensor_parameter_405: name = "parameter_405" + original_name = "batch_norm2d_57.w_0" shape = [192] dtype = "float32" min_val = float("0.292172") @@ -3870,6 +4276,7 @@ class Program_weight_tensor_parameter_405: class Program_weight_tensor_parameter_406: name = "parameter_406" + original_name = "batch_norm2d_57.w_2" shape = [192] dtype = "float32" min_val = float("0.0869552") @@ -3881,6 +4288,7 @@ class Program_weight_tensor_parameter_406: class Program_weight_tensor_parameter_407: name = "parameter_407" + original_name = "batch_norm2d_57.w_1" shape = [192] dtype = "float32" min_val = float("-0.311857") @@ -3892,6 +4300,7 @@ class Program_weight_tensor_parameter_407: class Program_weight_tensor_parameter_408: name = "parameter_408" + original_name = "conv2d_59.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.253119") @@ -3903,6 +4312,7 @@ class Program_weight_tensor_parameter_408: class Program_weight_tensor_parameter_409: name = "parameter_409" + original_name = "batch_norm2d_56.b_0" shape = [192] dtype = "float32" min_val = float("-2.84761") @@ -3914,6 +4324,7 @@ class Program_weight_tensor_parameter_409: class Program_weight_tensor_parameter_410: name = "parameter_410" + original_name = "batch_norm2d_56.w_0" shape = [192] dtype = "float32" min_val = float("0.617317") @@ -3925,6 +4336,7 @@ class Program_weight_tensor_parameter_410: class Program_weight_tensor_parameter_411: name = "parameter_411" + original_name = "batch_norm2d_56.w_2" shape = [192] dtype = "float32" min_val = float("4.84834") @@ -3936,6 +4348,7 @@ class Program_weight_tensor_parameter_411: class Program_weight_tensor_parameter_412: name = "parameter_412" + original_name = "batch_norm2d_56.w_1" shape = [192] dtype = "float32" min_val = float("-6.62629") @@ -3947,6 +4360,7 @@ class Program_weight_tensor_parameter_412: class Program_weight_tensor_parameter_413: name = "parameter_413" + original_name = "conv2d_58.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.485569") @@ -3958,6 +4372,7 @@ class Program_weight_tensor_parameter_413: class Program_weight_tensor_parameter_414: name = "parameter_414" + original_name = "batch_norm2d_55.b_0" shape = [192] dtype = "float32" min_val = float("-1.93093") @@ -3969,6 +4384,7 @@ class Program_weight_tensor_parameter_414: class Program_weight_tensor_parameter_415: name = "parameter_415" + original_name = "batch_norm2d_55.w_0" shape = [192] dtype = "float32" min_val = float("0.0292327") @@ -3980,6 +4396,7 @@ class Program_weight_tensor_parameter_415: class Program_weight_tensor_parameter_416: name = "parameter_416" + original_name = "batch_norm2d_55.w_2" shape = [192] dtype = "float32" min_val = float("0.00463541") @@ -3991,6 +4408,7 @@ class Program_weight_tensor_parameter_416: class Program_weight_tensor_parameter_417: name = "parameter_417" + original_name = "batch_norm2d_55.w_1" shape = [192] dtype = "float32" min_val = float("-0.267673") @@ -4002,6 +4420,7 @@ class Program_weight_tensor_parameter_417: class Program_weight_tensor_parameter_418: name = "parameter_418" + original_name = "conv2d_57.w_0" shape = [192, 192, 1, 1] dtype = "float32" min_val = float("-0.372544") @@ -4013,6 +4432,7 @@ class Program_weight_tensor_parameter_418: class Program_weight_tensor_parameter_419: name = "parameter_419" + original_name = "batch_norm2d_54.b_0" shape = [192] dtype = "float32" min_val = float("-1.93093") @@ -4024,6 +4444,7 @@ class Program_weight_tensor_parameter_419: class Program_weight_tensor_parameter_420: name = "parameter_420" + original_name = "batch_norm2d_54.w_0" shape = [192] dtype = "float32" min_val = float("0.4465") @@ -4035,6 +4456,7 @@ class Program_weight_tensor_parameter_420: class Program_weight_tensor_parameter_421: name = "parameter_421" + original_name = "batch_norm2d_54.w_2" shape = [192] dtype = "float32" min_val = float("0.160429") @@ -4046,6 +4468,7 @@ class Program_weight_tensor_parameter_421: class Program_weight_tensor_parameter_422: name = "parameter_422" + original_name = "batch_norm2d_54.w_1" shape = [192] dtype = "float32" min_val = float("-0.62254") @@ -4057,6 +4480,7 @@ class Program_weight_tensor_parameter_422: class Program_weight_tensor_parameter_423: name = "parameter_423" + original_name = "conv2d_56.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.28053") @@ -4068,6 +4492,7 @@ class Program_weight_tensor_parameter_423: class Program_weight_tensor_parameter_424: name = "parameter_424" + original_name = "batch_norm2d_53.b_0" shape = [192] dtype = "float32" min_val = float("-2.45161") @@ -4079,6 +4504,7 @@ class Program_weight_tensor_parameter_424: class Program_weight_tensor_parameter_425: name = "parameter_425" + original_name = "batch_norm2d_53.w_0" shape = [192] dtype = "float32" min_val = float("0.665839") @@ -4090,6 +4516,7 @@ class Program_weight_tensor_parameter_425: class Program_weight_tensor_parameter_426: name = "parameter_426" + original_name = "batch_norm2d_53.w_2" shape = [192] dtype = "float32" min_val = float("3.70249") @@ -4101,6 +4528,7 @@ class Program_weight_tensor_parameter_426: class Program_weight_tensor_parameter_427: name = "parameter_427" + original_name = "batch_norm2d_53.w_1" shape = [192] dtype = "float32" min_val = float("-8.16516") @@ -4112,6 +4540,7 @@ class Program_weight_tensor_parameter_427: class Program_weight_tensor_parameter_428: name = "parameter_428" + original_name = "conv2d_55.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.418041") @@ -4123,6 +4552,7 @@ class Program_weight_tensor_parameter_428: class Program_weight_tensor_parameter_429: name = "parameter_429" + original_name = "batch_norm2d_52.b_0" shape = [192] dtype = "float32" min_val = float("-1.75197") @@ -4134,6 +4564,7 @@ class Program_weight_tensor_parameter_429: class Program_weight_tensor_parameter_430: name = "parameter_430" + original_name = "batch_norm2d_52.w_0" shape = [192] dtype = "float32" min_val = float("0.00390728") @@ -4145,6 +4576,7 @@ class Program_weight_tensor_parameter_430: class Program_weight_tensor_parameter_431: name = "parameter_431" + original_name = "batch_norm2d_52.w_2" shape = [192] dtype = "float32" min_val = float("0.00212122") @@ -4156,6 +4588,7 @@ class Program_weight_tensor_parameter_431: class Program_weight_tensor_parameter_432: name = "parameter_432" + original_name = "batch_norm2d_52.w_1" shape = [192] dtype = "float32" min_val = float("-0.236415") @@ -4167,6 +4600,7 @@ class Program_weight_tensor_parameter_432: class Program_weight_tensor_parameter_433: name = "parameter_433" + original_name = "conv2d_54.w_0" shape = [192, 192, 1, 1] dtype = "float32" min_val = float("-0.367892") @@ -4178,6 +4612,7 @@ class Program_weight_tensor_parameter_433: class Program_weight_tensor_parameter_434: name = "parameter_434" + original_name = "batch_norm2d_51.b_0" shape = [192] dtype = "float32" min_val = float("-1.75197") @@ -4189,6 +4624,7 @@ class Program_weight_tensor_parameter_434: class Program_weight_tensor_parameter_435: name = "parameter_435" + original_name = "batch_norm2d_51.w_0" shape = [192] dtype = "float32" min_val = float("0.409451") @@ -4200,6 +4636,7 @@ class Program_weight_tensor_parameter_435: class Program_weight_tensor_parameter_436: name = "parameter_436" + original_name = "batch_norm2d_51.w_2" shape = [192] dtype = "float32" min_val = float("0.140448") @@ -4211,6 +4648,7 @@ class Program_weight_tensor_parameter_436: class Program_weight_tensor_parameter_437: name = "parameter_437" + original_name = "batch_norm2d_51.w_1" shape = [192] dtype = "float32" min_val = float("-0.355816") @@ -4222,6 +4660,7 @@ class Program_weight_tensor_parameter_437: class Program_weight_tensor_parameter_438: name = "parameter_438" + original_name = "conv2d_53.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.356702") @@ -4233,6 +4672,7 @@ class Program_weight_tensor_parameter_438: class Program_weight_tensor_parameter_439: name = "parameter_439" + original_name = "batch_norm2d_50.b_0" shape = [192] dtype = "float32" min_val = float("-2.43882") @@ -4244,6 +4684,7 @@ class Program_weight_tensor_parameter_439: class Program_weight_tensor_parameter_440: name = "parameter_440" + original_name = "batch_norm2d_50.w_0" shape = [192] dtype = "float32" min_val = float("0.704566") @@ -4255,6 +4696,7 @@ class Program_weight_tensor_parameter_440: class Program_weight_tensor_parameter_441: name = "parameter_441" + original_name = "batch_norm2d_50.w_2" shape = [192] dtype = "float32" min_val = float("2.71863") @@ -4266,6 +4708,7 @@ class Program_weight_tensor_parameter_441: class Program_weight_tensor_parameter_442: name = "parameter_442" + original_name = "batch_norm2d_50.w_1" shape = [192] dtype = "float32" min_val = float("-7.84718") @@ -4277,6 +4720,7 @@ class Program_weight_tensor_parameter_442: class Program_weight_tensor_parameter_443: name = "parameter_443" + original_name = "conv2d_52.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.260597") @@ -4288,6 +4732,7 @@ class Program_weight_tensor_parameter_443: class Program_weight_tensor_parameter_444: name = "parameter_444" + original_name = "batch_norm2d_49.b_0" shape = [192] dtype = "float32" min_val = float("-2.0766") @@ -4299,6 +4744,7 @@ class Program_weight_tensor_parameter_444: class Program_weight_tensor_parameter_445: name = "parameter_445" + original_name = "batch_norm2d_49.w_0" shape = [192] dtype = "float32" min_val = float("-0.0011458") @@ -4310,6 +4756,7 @@ class Program_weight_tensor_parameter_445: class Program_weight_tensor_parameter_446: name = "parameter_446" + original_name = "batch_norm2d_49.w_2" shape = [192] dtype = "float32" min_val = float("0.000773168") @@ -4321,6 +4768,7 @@ class Program_weight_tensor_parameter_446: class Program_weight_tensor_parameter_447: name = "parameter_447" + original_name = "batch_norm2d_49.w_1" shape = [192] dtype = "float32" min_val = float("-0.235196") @@ -4332,6 +4780,7 @@ class Program_weight_tensor_parameter_447: class Program_weight_tensor_parameter_448: name = "parameter_448" + original_name = "conv2d_51.w_0" shape = [192, 192, 1, 1] dtype = "float32" min_val = float("-0.231821") @@ -4343,6 +4792,7 @@ class Program_weight_tensor_parameter_448: class Program_weight_tensor_parameter_449: name = "parameter_449" + original_name = "batch_norm2d_48.b_0" shape = [192] dtype = "float32" min_val = float("-2.0766") @@ -4354,6 +4804,7 @@ class Program_weight_tensor_parameter_449: class Program_weight_tensor_parameter_450: name = "parameter_450" + original_name = "batch_norm2d_48.w_0" shape = [192] dtype = "float32" min_val = float("0.391129") @@ -4365,6 +4816,7 @@ class Program_weight_tensor_parameter_450: class Program_weight_tensor_parameter_451: name = "parameter_451" + original_name = "batch_norm2d_48.w_2" shape = [192] dtype = "float32" min_val = float("0.153988") @@ -4376,6 +4828,7 @@ class Program_weight_tensor_parameter_451: class Program_weight_tensor_parameter_452: name = "parameter_452" + original_name = "batch_norm2d_48.w_1" shape = [192] dtype = "float32" min_val = float("-0.425378") @@ -4387,6 +4840,7 @@ class Program_weight_tensor_parameter_452: class Program_weight_tensor_parameter_453: name = "parameter_453" + original_name = "conv2d_50.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.342234") @@ -4398,6 +4852,7 @@ class Program_weight_tensor_parameter_453: class Program_weight_tensor_parameter_454: name = "parameter_454" + original_name = "batch_norm2d_47.b_0" shape = [192] dtype = "float32" min_val = float("-2.68478") @@ -4409,6 +4864,7 @@ class Program_weight_tensor_parameter_454: class Program_weight_tensor_parameter_455: name = "parameter_455" + original_name = "batch_norm2d_47.w_0" shape = [192] dtype = "float32" min_val = float("0.729013") @@ -4420,6 +4876,7 @@ class Program_weight_tensor_parameter_455: class Program_weight_tensor_parameter_456: name = "parameter_456" + original_name = "batch_norm2d_47.w_2" shape = [192] dtype = "float32" min_val = float("1.95681") @@ -4431,6 +4888,7 @@ class Program_weight_tensor_parameter_456: class Program_weight_tensor_parameter_457: name = "parameter_457" + original_name = "batch_norm2d_47.w_1" shape = [192] dtype = "float32" min_val = float("-9.70608") @@ -4442,6 +4900,7 @@ class Program_weight_tensor_parameter_457: class Program_weight_tensor_parameter_458: name = "parameter_458" + original_name = "conv2d_49.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.298277") @@ -4453,6 +4912,7 @@ class Program_weight_tensor_parameter_458: class Program_weight_tensor_parameter_459: name = "parameter_459" + original_name = "batch_norm2d_46.b_0" shape = [192] dtype = "float32" min_val = float("-1.22542") @@ -4464,6 +4924,7 @@ class Program_weight_tensor_parameter_459: class Program_weight_tensor_parameter_460: name = "parameter_460" + original_name = "batch_norm2d_46.w_0" shape = [192] dtype = "float32" min_val = float("-0.00595314") @@ -4475,6 +4936,7 @@ class Program_weight_tensor_parameter_460: class Program_weight_tensor_parameter_461: name = "parameter_461" + original_name = "batch_norm2d_46.w_2" shape = [192] dtype = "float32" min_val = float("0.0013768") @@ -4486,6 +4948,7 @@ class Program_weight_tensor_parameter_461: class Program_weight_tensor_parameter_462: name = "parameter_462" + original_name = "batch_norm2d_46.w_1" shape = [192] dtype = "float32" min_val = float("-0.2057") @@ -4497,6 +4960,7 @@ class Program_weight_tensor_parameter_462: class Program_weight_tensor_parameter_463: name = "parameter_463" + original_name = "conv2d_48.w_0" shape = [192, 192, 1, 1] dtype = "float32" min_val = float("-0.280158") @@ -4508,6 +4972,7 @@ class Program_weight_tensor_parameter_463: class Program_weight_tensor_parameter_464: name = "parameter_464" + original_name = "batch_norm2d_45.b_0" shape = [192] dtype = "float32" min_val = float("-1.22542") @@ -4519,6 +4984,7 @@ class Program_weight_tensor_parameter_464: class Program_weight_tensor_parameter_465: name = "parameter_465" + original_name = "batch_norm2d_45.w_0" shape = [192] dtype = "float32" min_val = float("0.400501") @@ -4530,6 +4996,7 @@ class Program_weight_tensor_parameter_465: class Program_weight_tensor_parameter_466: name = "parameter_466" + original_name = "batch_norm2d_45.w_2" shape = [192] dtype = "float32" min_val = float("0.134699") @@ -4541,6 +5008,7 @@ class Program_weight_tensor_parameter_466: class Program_weight_tensor_parameter_467: name = "parameter_467" + original_name = "batch_norm2d_45.w_1" shape = [192] dtype = "float32" min_val = float("-0.478965") @@ -4552,6 +5020,7 @@ class Program_weight_tensor_parameter_467: class Program_weight_tensor_parameter_468: name = "parameter_468" + original_name = "conv2d_47.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.357121") @@ -4563,6 +5032,7 @@ class Program_weight_tensor_parameter_468: class Program_weight_tensor_parameter_469: name = "parameter_469" + original_name = "batch_norm2d_44.b_0" shape = [192] dtype = "float32" min_val = float("-2.4647") @@ -4574,6 +5044,7 @@ class Program_weight_tensor_parameter_469: class Program_weight_tensor_parameter_470: name = "parameter_470" + original_name = "batch_norm2d_44.w_0" shape = [192] dtype = "float32" min_val = float("0.696898") @@ -4585,6 +5056,7 @@ class Program_weight_tensor_parameter_470: class Program_weight_tensor_parameter_471: name = "parameter_471" + original_name = "batch_norm2d_44.w_2" shape = [192] dtype = "float32" min_val = float("1.42719") @@ -4596,6 +5068,7 @@ class Program_weight_tensor_parameter_471: class Program_weight_tensor_parameter_472: name = "parameter_472" + original_name = "batch_norm2d_44.w_1" shape = [192] dtype = "float32" min_val = float("-4.3222") @@ -4607,6 +5080,7 @@ class Program_weight_tensor_parameter_472: class Program_weight_tensor_parameter_473: name = "parameter_473" + original_name = "conv2d_46.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.497696") @@ -4618,6 +5092,7 @@ class Program_weight_tensor_parameter_473: class Program_weight_tensor_parameter_474: name = "parameter_474" + original_name = "batch_norm2d_43.b_0" shape = [192] dtype = "float32" min_val = float("-1.23483") @@ -4629,6 +5104,7 @@ class Program_weight_tensor_parameter_474: class Program_weight_tensor_parameter_475: name = "parameter_475" + original_name = "batch_norm2d_43.w_0" shape = [192] dtype = "float32" min_val = float("0.0145354") @@ -4640,6 +5116,7 @@ class Program_weight_tensor_parameter_475: class Program_weight_tensor_parameter_476: name = "parameter_476" + original_name = "batch_norm2d_43.w_2" shape = [192] dtype = "float32" min_val = float("0.0048331") @@ -4651,6 +5128,7 @@ class Program_weight_tensor_parameter_476: class Program_weight_tensor_parameter_477: name = "parameter_477" + original_name = "batch_norm2d_43.w_1" shape = [192] dtype = "float32" min_val = float("-0.530999") @@ -4662,6 +5140,7 @@ class Program_weight_tensor_parameter_477: class Program_weight_tensor_parameter_478: name = "parameter_478" + original_name = "conv2d_45.w_0" shape = [192, 192, 1, 1] dtype = "float32" min_val = float("-0.487726") @@ -4673,6 +5152,7 @@ class Program_weight_tensor_parameter_478: class Program_weight_tensor_parameter_479: name = "parameter_479" + original_name = "batch_norm2d_42.b_0" shape = [192] dtype = "float32" min_val = float("-1.23483") @@ -4684,6 +5164,7 @@ class Program_weight_tensor_parameter_479: class Program_weight_tensor_parameter_480: name = "parameter_480" + original_name = "batch_norm2d_42.w_0" shape = [192] dtype = "float32" min_val = float("0.34022") @@ -4695,6 +5176,7 @@ class Program_weight_tensor_parameter_480: class Program_weight_tensor_parameter_481: name = "parameter_481" + original_name = "batch_norm2d_42.w_2" shape = [192] dtype = "float32" min_val = float("0.212544") @@ -4706,6 +5188,7 @@ class Program_weight_tensor_parameter_481: class Program_weight_tensor_parameter_482: name = "parameter_482" + original_name = "batch_norm2d_42.w_1" shape = [192] dtype = "float32" min_val = float("-0.759339") @@ -4717,6 +5200,7 @@ class Program_weight_tensor_parameter_482: class Program_weight_tensor_parameter_483: name = "parameter_483" + original_name = "conv2d_44.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.38033") @@ -4728,6 +5212,7 @@ class Program_weight_tensor_parameter_483: class Program_weight_tensor_parameter_484: name = "parameter_484" + original_name = "batch_norm2d_41.b_0" shape = [192] dtype = "float32" min_val = float("-1.83112") @@ -4739,6 +5224,7 @@ class Program_weight_tensor_parameter_484: class Program_weight_tensor_parameter_485: name = "parameter_485" + original_name = "batch_norm2d_41.w_0" shape = [192] dtype = "float32" min_val = float("0.819725") @@ -4750,6 +5236,7 @@ class Program_weight_tensor_parameter_485: class Program_weight_tensor_parameter_486: name = "parameter_486" + original_name = "batch_norm2d_41.w_2" shape = [192] dtype = "float32" min_val = float("0.891588") @@ -4761,6 +5248,7 @@ class Program_weight_tensor_parameter_486: class Program_weight_tensor_parameter_487: name = "parameter_487" + original_name = "batch_norm2d_41.w_1" shape = [192] dtype = "float32" min_val = float("-4.72978") @@ -4772,6 +5260,7 @@ class Program_weight_tensor_parameter_487: class Program_weight_tensor_parameter_488: name = "parameter_488" + original_name = "conv2d_43.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.65171") @@ -4783,6 +5272,7 @@ class Program_weight_tensor_parameter_488: class Program_weight_tensor_parameter_489: name = "parameter_489" + original_name = "batch_norm2d_40.b_0" shape = [192] dtype = "float32" min_val = float("-2.86404") @@ -4794,6 +5284,7 @@ class Program_weight_tensor_parameter_489: class Program_weight_tensor_parameter_490: name = "parameter_490" + original_name = "batch_norm2d_40.w_0" shape = [192] dtype = "float32" min_val = float("0.500034") @@ -4805,6 +5296,7 @@ class Program_weight_tensor_parameter_490: class Program_weight_tensor_parameter_491: name = "parameter_491" + original_name = "batch_norm2d_40.w_2" shape = [192] dtype = "float32" min_val = float("0.644561") @@ -4816,6 +5308,7 @@ class Program_weight_tensor_parameter_491: class Program_weight_tensor_parameter_492: name = "parameter_492" + original_name = "batch_norm2d_40.w_1" shape = [192] dtype = "float32" min_val = float("-1.2704") @@ -4827,6 +5320,7 @@ class Program_weight_tensor_parameter_492: class Program_weight_tensor_parameter_493: name = "parameter_493" + original_name = "conv2d_42.w_0" shape = [192, 384, 1, 1] dtype = "float32" min_val = float("-0.982015") @@ -4838,6 +5332,7 @@ class Program_weight_tensor_parameter_493: class Program_weight_tensor_parameter_494: name = "parameter_494" + original_name = "batch_norm2d_39.b_0" shape = [192] dtype = "float32" min_val = float("-3.0935") @@ -4849,6 +5344,7 @@ class Program_weight_tensor_parameter_494: class Program_weight_tensor_parameter_495: name = "parameter_495" + original_name = "batch_norm2d_39.w_0" shape = [192] dtype = "float32" min_val = float("0.875539") @@ -4860,6 +5356,7 @@ class Program_weight_tensor_parameter_495: class Program_weight_tensor_parameter_496: name = "parameter_496" + original_name = "batch_norm2d_39.w_2" shape = [192] dtype = "float32" min_val = float("0.762109") @@ -4871,6 +5368,7 @@ class Program_weight_tensor_parameter_496: class Program_weight_tensor_parameter_497: name = "parameter_497" + original_name = "batch_norm2d_39.w_1" shape = [192] dtype = "float32" min_val = float("-1.00415") @@ -4882,6 +5380,7 @@ class Program_weight_tensor_parameter_497: class Program_weight_tensor_parameter_498: name = "parameter_498" + original_name = "conv2d_41.w_0" shape = [192, 384, 1, 1] dtype = "float32" min_val = float("-0.985177") @@ -4893,6 +5392,7 @@ class Program_weight_tensor_parameter_498: class Program_weight_tensor_parameter_499: name = "parameter_499" + original_name = "batch_norm2d_38.b_0" shape = [384] dtype = "float32" min_val = float("-2.90784") @@ -4904,6 +5404,7 @@ class Program_weight_tensor_parameter_499: class Program_weight_tensor_parameter_500: name = "parameter_500" + original_name = "batch_norm2d_38.w_0" shape = [384] dtype = "float32" min_val = float("0.648698") @@ -4915,6 +5416,7 @@ class Program_weight_tensor_parameter_500: class Program_weight_tensor_parameter_501: name = "parameter_501" + original_name = "batch_norm2d_38.w_2" shape = [384] dtype = "float32" min_val = float("0.59614") @@ -4926,6 +5428,7 @@ class Program_weight_tensor_parameter_501: class Program_weight_tensor_parameter_502: name = "parameter_502" + original_name = "batch_norm2d_38.w_1" shape = [384] dtype = "float32" min_val = float("-2.10662") @@ -4937,6 +5440,7 @@ class Program_weight_tensor_parameter_502: class Program_weight_tensor_parameter_503: name = "parameter_503" + original_name = "conv2d_40.w_0" shape = [384, 256, 3, 3] dtype = "float32" min_val = float("-0.655717") @@ -4948,6 +5452,7 @@ class Program_weight_tensor_parameter_503: class Program_weight_tensor_parameter_504: name = "parameter_504" + original_name = "batch_norm2d_37.b_0" shape = [256] dtype = "float32" min_val = float("-1.98829") @@ -4959,6 +5464,7 @@ class Program_weight_tensor_parameter_504: class Program_weight_tensor_parameter_505: name = "parameter_505" + original_name = "batch_norm2d_37.w_0" shape = [256] dtype = "float32" min_val = float("0.40595") @@ -4970,6 +5476,7 @@ class Program_weight_tensor_parameter_505: class Program_weight_tensor_parameter_506: name = "parameter_506" + original_name = "batch_norm2d_37.w_2" shape = [256] dtype = "float32" min_val = float("0.758135") @@ -4981,6 +5488,7 @@ class Program_weight_tensor_parameter_506: class Program_weight_tensor_parameter_507: name = "parameter_507" + original_name = "batch_norm2d_37.w_1" shape = [256] dtype = "float32" min_val = float("-1.75746") @@ -4992,6 +5500,7 @@ class Program_weight_tensor_parameter_507: class Program_weight_tensor_parameter_508: name = "parameter_508" + original_name = "conv2d_39.w_0" shape = [256, 192, 1, 1] dtype = "float32" min_val = float("-1.45832") @@ -5003,6 +5512,7 @@ class Program_weight_tensor_parameter_508: class Program_weight_tensor_parameter_509: name = "parameter_509" + original_name = "conv2d_38.b_0" shape = [192] dtype = "float32" min_val = float("-0.515235") @@ -5014,6 +5524,7 @@ class Program_weight_tensor_parameter_509: class Program_weight_tensor_parameter_510: name = "parameter_510" + original_name = "conv2d_38.w_0" shape = [192, 192, 1, 1] dtype = "float32" min_val = float("-1.91635") @@ -5025,6 +5536,7 @@ class Program_weight_tensor_parameter_510: class Program_weight_tensor_parameter_511: name = "parameter_511" + original_name = "batch_norm2d_36.b_0" shape = [96] dtype = "float32" min_val = float("-2.05715") @@ -5036,6 +5548,7 @@ class Program_weight_tensor_parameter_511: class Program_weight_tensor_parameter_512: name = "parameter_512" + original_name = "batch_norm2d_36.w_0" shape = [96] dtype = "float32" min_val = float("0.143618") @@ -5047,6 +5560,7 @@ class Program_weight_tensor_parameter_512: class Program_weight_tensor_parameter_513: name = "parameter_513" + original_name = "batch_norm2d_36.w_2" shape = [96] dtype = "float32" min_val = float("0.00942651") @@ -5058,6 +5572,7 @@ class Program_weight_tensor_parameter_513: class Program_weight_tensor_parameter_514: name = "parameter_514" + original_name = "batch_norm2d_36.w_1" shape = [96] dtype = "float32" min_val = float("-0.193494") @@ -5069,6 +5584,7 @@ class Program_weight_tensor_parameter_514: class Program_weight_tensor_parameter_515: name = "parameter_515" + original_name = "conv2d_37.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.68938") @@ -5080,6 +5596,7 @@ class Program_weight_tensor_parameter_515: class Program_weight_tensor_parameter_516: name = "parameter_516" + original_name = "batch_norm2d_35.b_0" shape = [96] dtype = "float32" min_val = float("-2.05715") @@ -5091,6 +5608,7 @@ class Program_weight_tensor_parameter_516: class Program_weight_tensor_parameter_517: name = "parameter_517" + original_name = "batch_norm2d_35.w_0" shape = [96] dtype = "float32" min_val = float("0.308202") @@ -5102,6 +5620,7 @@ class Program_weight_tensor_parameter_517: class Program_weight_tensor_parameter_518: name = "parameter_518" + original_name = "batch_norm2d_35.w_2" shape = [96] dtype = "float32" min_val = float("0.0612546") @@ -5113,6 +5632,7 @@ class Program_weight_tensor_parameter_518: class Program_weight_tensor_parameter_519: name = "parameter_519" + original_name = "batch_norm2d_35.w_1" shape = [96] dtype = "float32" min_val = float("-0.413394") @@ -5124,6 +5644,7 @@ class Program_weight_tensor_parameter_519: class Program_weight_tensor_parameter_520: name = "parameter_520" + original_name = "conv2d_36.w_0" shape = [96, 96, 3, 3] dtype = "float32" min_val = float("-0.440687") @@ -5135,6 +5656,7 @@ class Program_weight_tensor_parameter_520: class Program_weight_tensor_parameter_521: name = "parameter_521" + original_name = "batch_norm2d_34.b_0" shape = [96] dtype = "float32" min_val = float("-2.3818") @@ -5146,6 +5668,7 @@ class Program_weight_tensor_parameter_521: class Program_weight_tensor_parameter_522: name = "parameter_522" + original_name = "batch_norm2d_34.w_0" shape = [96] dtype = "float32" min_val = float("0.517878") @@ -5157,6 +5680,7 @@ class Program_weight_tensor_parameter_522: class Program_weight_tensor_parameter_523: name = "parameter_523" + original_name = "batch_norm2d_34.w_2" shape = [96] dtype = "float32" min_val = float("3.14105") @@ -5168,6 +5692,7 @@ class Program_weight_tensor_parameter_523: class Program_weight_tensor_parameter_524: name = "parameter_524" + original_name = "batch_norm2d_34.w_1" shape = [96] dtype = "float32" min_val = float("-11.2555") @@ -5179,6 +5704,7 @@ class Program_weight_tensor_parameter_524: class Program_weight_tensor_parameter_525: name = "parameter_525" + original_name = "conv2d_35.w_0" shape = [96, 96, 3, 3] dtype = "float32" min_val = float("-1.11358") @@ -5190,6 +5716,7 @@ class Program_weight_tensor_parameter_525: class Program_weight_tensor_parameter_526: name = "parameter_526" + original_name = "batch_norm2d_33.b_0" shape = [96] dtype = "float32" min_val = float("-1.43264") @@ -5201,6 +5728,7 @@ class Program_weight_tensor_parameter_526: class Program_weight_tensor_parameter_527: name = "parameter_527" + original_name = "batch_norm2d_33.w_0" shape = [96] dtype = "float32" min_val = float("0.0612513") @@ -5212,6 +5740,7 @@ class Program_weight_tensor_parameter_527: class Program_weight_tensor_parameter_528: name = "parameter_528" + original_name = "batch_norm2d_33.w_2" shape = [96] dtype = "float32" min_val = float("0.0120131") @@ -5223,6 +5752,7 @@ class Program_weight_tensor_parameter_528: class Program_weight_tensor_parameter_529: name = "parameter_529" + original_name = "batch_norm2d_33.w_1" shape = [96] dtype = "float32" min_val = float("-0.1799") @@ -5234,6 +5764,7 @@ class Program_weight_tensor_parameter_529: class Program_weight_tensor_parameter_530: name = "parameter_530" + original_name = "conv2d_34.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.56327") @@ -5245,6 +5776,7 @@ class Program_weight_tensor_parameter_530: class Program_weight_tensor_parameter_531: name = "parameter_531" + original_name = "batch_norm2d_32.b_0" shape = [96] dtype = "float32" min_val = float("-1.43264") @@ -5256,6 +5788,7 @@ class Program_weight_tensor_parameter_531: class Program_weight_tensor_parameter_532: name = "parameter_532" + original_name = "batch_norm2d_32.w_0" shape = [96] dtype = "float32" min_val = float("0.345133") @@ -5267,6 +5800,7 @@ class Program_weight_tensor_parameter_532: class Program_weight_tensor_parameter_533: name = "parameter_533" + original_name = "batch_norm2d_32.w_2" shape = [96] dtype = "float32" min_val = float("0.143109") @@ -5278,6 +5812,7 @@ class Program_weight_tensor_parameter_533: class Program_weight_tensor_parameter_534: name = "parameter_534" + original_name = "batch_norm2d_32.w_1" shape = [96] dtype = "float32" min_val = float("-0.465589") @@ -5289,6 +5824,7 @@ class Program_weight_tensor_parameter_534: class Program_weight_tensor_parameter_535: name = "parameter_535" + original_name = "conv2d_33.w_0" shape = [96, 96, 3, 3] dtype = "float32" min_val = float("-0.44419") @@ -5300,6 +5836,7 @@ class Program_weight_tensor_parameter_535: class Program_weight_tensor_parameter_536: name = "parameter_536" + original_name = "batch_norm2d_31.b_0" shape = [96] dtype = "float32" min_val = float("-3.33855") @@ -5311,6 +5848,7 @@ class Program_weight_tensor_parameter_536: class Program_weight_tensor_parameter_537: name = "parameter_537" + original_name = "batch_norm2d_31.w_0" shape = [96] dtype = "float32" min_val = float("0.53089") @@ -5322,6 +5860,7 @@ class Program_weight_tensor_parameter_537: class Program_weight_tensor_parameter_538: name = "parameter_538" + original_name = "batch_norm2d_31.w_2" shape = [96] dtype = "float32" min_val = float("2.23028") @@ -5333,6 +5872,7 @@ class Program_weight_tensor_parameter_538: class Program_weight_tensor_parameter_539: name = "parameter_539" + original_name = "batch_norm2d_31.w_1" shape = [96] dtype = "float32" min_val = float("-6.68954") @@ -5344,6 +5884,7 @@ class Program_weight_tensor_parameter_539: class Program_weight_tensor_parameter_540: name = "parameter_540" + original_name = "conv2d_32.w_0" shape = [96, 96, 3, 3] dtype = "float32" min_val = float("-0.895325") @@ -5355,6 +5896,7 @@ class Program_weight_tensor_parameter_540: class Program_weight_tensor_parameter_541: name = "parameter_541" + original_name = "batch_norm2d_30.b_0" shape = [96] dtype = "float32" min_val = float("-1.27494") @@ -5366,6 +5908,7 @@ class Program_weight_tensor_parameter_541: class Program_weight_tensor_parameter_542: name = "parameter_542" + original_name = "batch_norm2d_30.w_0" shape = [96] dtype = "float32" min_val = float("0.0455835") @@ -5377,6 +5920,7 @@ class Program_weight_tensor_parameter_542: class Program_weight_tensor_parameter_543: name = "parameter_543" + original_name = "batch_norm2d_30.w_2" shape = [96] dtype = "float32" min_val = float("0.0122159") @@ -5388,6 +5932,7 @@ class Program_weight_tensor_parameter_543: class Program_weight_tensor_parameter_544: name = "parameter_544" + original_name = "batch_norm2d_30.w_1" shape = [96] dtype = "float32" min_val = float("-0.184219") @@ -5399,6 +5944,7 @@ class Program_weight_tensor_parameter_544: class Program_weight_tensor_parameter_545: name = "parameter_545" + original_name = "conv2d_31.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.712045") @@ -5410,6 +5956,7 @@ class Program_weight_tensor_parameter_545: class Program_weight_tensor_parameter_546: name = "parameter_546" + original_name = "batch_norm2d_29.b_0" shape = [96] dtype = "float32" min_val = float("-1.27494") @@ -5421,6 +5968,7 @@ class Program_weight_tensor_parameter_546: class Program_weight_tensor_parameter_547: name = "parameter_547" + original_name = "batch_norm2d_29.w_0" shape = [96] dtype = "float32" min_val = float("0.294566") @@ -5432,6 +5980,7 @@ class Program_weight_tensor_parameter_547: class Program_weight_tensor_parameter_548: name = "parameter_548" + original_name = "batch_norm2d_29.w_2" shape = [96] dtype = "float32" min_val = float("0.165557") @@ -5443,6 +5992,7 @@ class Program_weight_tensor_parameter_548: class Program_weight_tensor_parameter_549: name = "parameter_549" + original_name = "batch_norm2d_29.w_1" shape = [96] dtype = "float32" min_val = float("-0.341849") @@ -5454,6 +6004,7 @@ class Program_weight_tensor_parameter_549: class Program_weight_tensor_parameter_550: name = "parameter_550" + original_name = "conv2d_30.w_0" shape = [96, 96, 3, 3] dtype = "float32" min_val = float("-0.497612") @@ -5465,6 +6016,7 @@ class Program_weight_tensor_parameter_550: class Program_weight_tensor_parameter_551: name = "parameter_551" + original_name = "batch_norm2d_28.b_0" shape = [96] dtype = "float32" min_val = float("-3.61693") @@ -5476,6 +6028,7 @@ class Program_weight_tensor_parameter_551: class Program_weight_tensor_parameter_552: name = "parameter_552" + original_name = "batch_norm2d_28.w_0" shape = [96] dtype = "float32" min_val = float("0.560756") @@ -5487,6 +6040,7 @@ class Program_weight_tensor_parameter_552: class Program_weight_tensor_parameter_553: name = "parameter_553" + original_name = "batch_norm2d_28.w_2" shape = [96] dtype = "float32" min_val = float("1.75486") @@ -5498,6 +6052,7 @@ class Program_weight_tensor_parameter_553: class Program_weight_tensor_parameter_554: name = "parameter_554" + original_name = "batch_norm2d_28.w_1" shape = [96] dtype = "float32" min_val = float("-5.71005") @@ -5509,6 +6064,7 @@ class Program_weight_tensor_parameter_554: class Program_weight_tensor_parameter_555: name = "parameter_555" + original_name = "conv2d_29.w_0" shape = [96, 96, 3, 3] dtype = "float32" min_val = float("-0.800917") @@ -5520,6 +6076,7 @@ class Program_weight_tensor_parameter_555: class Program_weight_tensor_parameter_556: name = "parameter_556" + original_name = "batch_norm2d_27.b_0" shape = [96] dtype = "float32" min_val = float("-0.883924") @@ -5531,6 +6088,7 @@ class Program_weight_tensor_parameter_556: class Program_weight_tensor_parameter_557: name = "parameter_557" + original_name = "batch_norm2d_27.w_0" shape = [96] dtype = "float32" min_val = float("0.0635187") @@ -5542,6 +6100,7 @@ class Program_weight_tensor_parameter_557: class Program_weight_tensor_parameter_558: name = "parameter_558" + original_name = "batch_norm2d_27.w_2" shape = [96] dtype = "float32" min_val = float("0.00987779") @@ -5553,6 +6112,7 @@ class Program_weight_tensor_parameter_558: class Program_weight_tensor_parameter_559: name = "parameter_559" + original_name = "batch_norm2d_27.w_1" shape = [96] dtype = "float32" min_val = float("-0.191291") @@ -5564,6 +6124,7 @@ class Program_weight_tensor_parameter_559: class Program_weight_tensor_parameter_560: name = "parameter_560" + original_name = "conv2d_28.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.690611") @@ -5575,6 +6136,7 @@ class Program_weight_tensor_parameter_560: class Program_weight_tensor_parameter_561: name = "parameter_561" + original_name = "batch_norm2d_26.b_0" shape = [96] dtype = "float32" min_val = float("-0.883924") @@ -5586,6 +6148,7 @@ class Program_weight_tensor_parameter_561: class Program_weight_tensor_parameter_562: name = "parameter_562" + original_name = "batch_norm2d_26.w_0" shape = [96] dtype = "float32" min_val = float("0.196107") @@ -5597,6 +6160,7 @@ class Program_weight_tensor_parameter_562: class Program_weight_tensor_parameter_563: name = "parameter_563" + original_name = "batch_norm2d_26.w_2" shape = [96] dtype = "float32" min_val = float("0.124698") @@ -5608,6 +6172,7 @@ class Program_weight_tensor_parameter_563: class Program_weight_tensor_parameter_564: name = "parameter_564" + original_name = "batch_norm2d_26.w_1" shape = [96] dtype = "float32" min_val = float("-0.536893") @@ -5619,6 +6184,7 @@ class Program_weight_tensor_parameter_564: class Program_weight_tensor_parameter_565: name = "parameter_565" + original_name = "conv2d_27.w_0" shape = [96, 96, 3, 3] dtype = "float32" min_val = float("-0.460783") @@ -5630,6 +6196,7 @@ class Program_weight_tensor_parameter_565: class Program_weight_tensor_parameter_566: name = "parameter_566" + original_name = "batch_norm2d_25.b_0" shape = [96] dtype = "float32" min_val = float("-2.61271") @@ -5641,6 +6208,7 @@ class Program_weight_tensor_parameter_566: class Program_weight_tensor_parameter_567: name = "parameter_567" + original_name = "batch_norm2d_25.w_0" shape = [96] dtype = "float32" min_val = float("0.555948") @@ -5652,6 +6220,7 @@ class Program_weight_tensor_parameter_567: class Program_weight_tensor_parameter_568: name = "parameter_568" + original_name = "batch_norm2d_25.w_2" shape = [96] dtype = "float32" min_val = float("1.34698") @@ -5663,6 +6232,7 @@ class Program_weight_tensor_parameter_568: class Program_weight_tensor_parameter_569: name = "parameter_569" + original_name = "batch_norm2d_25.w_1" shape = [96] dtype = "float32" min_val = float("-6.05567") @@ -5674,6 +6244,7 @@ class Program_weight_tensor_parameter_569: class Program_weight_tensor_parameter_570: name = "parameter_570" + original_name = "conv2d_26.w_0" shape = [96, 96, 3, 3] dtype = "float32" min_val = float("-0.570501") @@ -5685,6 +6256,7 @@ class Program_weight_tensor_parameter_570: class Program_weight_tensor_parameter_571: name = "parameter_571" + original_name = "batch_norm2d_24.b_0" shape = [96] dtype = "float32" min_val = float("-1.02922") @@ -5696,6 +6268,7 @@ class Program_weight_tensor_parameter_571: class Program_weight_tensor_parameter_572: name = "parameter_572" + original_name = "batch_norm2d_24.w_0" shape = [96] dtype = "float32" min_val = float("0.0878477") @@ -5707,6 +6280,7 @@ class Program_weight_tensor_parameter_572: class Program_weight_tensor_parameter_573: name = "parameter_573" + original_name = "batch_norm2d_24.w_2" shape = [96] dtype = "float32" min_val = float("0.0219696") @@ -5718,6 +6292,7 @@ class Program_weight_tensor_parameter_573: class Program_weight_tensor_parameter_574: name = "parameter_574" + original_name = "batch_norm2d_24.w_1" shape = [96] dtype = "float32" min_val = float("-0.183021") @@ -5729,6 +6304,7 @@ class Program_weight_tensor_parameter_574: class Program_weight_tensor_parameter_575: name = "parameter_575" + original_name = "conv2d_25.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.42972") @@ -5740,6 +6316,7 @@ class Program_weight_tensor_parameter_575: class Program_weight_tensor_parameter_576: name = "parameter_576" + original_name = "batch_norm2d_23.b_0" shape = [96] dtype = "float32" min_val = float("-1.02923") @@ -5751,6 +6328,7 @@ class Program_weight_tensor_parameter_576: class Program_weight_tensor_parameter_577: name = "parameter_577" + original_name = "batch_norm2d_23.w_0" shape = [96] dtype = "float32" min_val = float("0.324905") @@ -5762,6 +6340,7 @@ class Program_weight_tensor_parameter_577: class Program_weight_tensor_parameter_578: name = "parameter_578" + original_name = "batch_norm2d_23.w_2" shape = [96] dtype = "float32" min_val = float("0.243143") @@ -5773,6 +6352,7 @@ class Program_weight_tensor_parameter_578: class Program_weight_tensor_parameter_579: name = "parameter_579" + original_name = "batch_norm2d_23.w_1" shape = [96] dtype = "float32" min_val = float("-0.382214") @@ -5784,6 +6364,7 @@ class Program_weight_tensor_parameter_579: class Program_weight_tensor_parameter_580: name = "parameter_580" + original_name = "conv2d_24.w_0" shape = [96, 96, 3, 3] dtype = "float32" min_val = float("-0.862711") @@ -5795,6 +6376,7 @@ class Program_weight_tensor_parameter_580: class Program_weight_tensor_parameter_581: name = "parameter_581" + original_name = "batch_norm2d_22.b_0" shape = [96] dtype = "float32" min_val = float("-3.53042") @@ -5806,6 +6388,7 @@ class Program_weight_tensor_parameter_581: class Program_weight_tensor_parameter_582: name = "parameter_582" + original_name = "batch_norm2d_22.w_0" shape = [96] dtype = "float32" min_val = float("0.642452") @@ -5817,6 +6400,7 @@ class Program_weight_tensor_parameter_582: class Program_weight_tensor_parameter_583: name = "parameter_583" + original_name = "batch_norm2d_22.w_2" shape = [96] dtype = "float32" min_val = float("1.07356") @@ -5828,6 +6412,7 @@ class Program_weight_tensor_parameter_583: class Program_weight_tensor_parameter_584: name = "parameter_584" + original_name = "batch_norm2d_22.w_1" shape = [96] dtype = "float32" min_val = float("-4.34406") @@ -5839,6 +6424,7 @@ class Program_weight_tensor_parameter_584: class Program_weight_tensor_parameter_585: name = "parameter_585" + original_name = "conv2d_23.w_0" shape = [96, 96, 3, 3] dtype = "float32" min_val = float("-0.503152") @@ -5850,6 +6436,7 @@ class Program_weight_tensor_parameter_585: class Program_weight_tensor_parameter_586: name = "parameter_586" + original_name = "batch_norm2d_21.b_0" shape = [96] dtype = "float32" min_val = float("-0.681237") @@ -5861,6 +6448,7 @@ class Program_weight_tensor_parameter_586: class Program_weight_tensor_parameter_587: name = "parameter_587" + original_name = "batch_norm2d_21.w_0" shape = [96] dtype = "float32" min_val = float("0.0912332") @@ -5872,6 +6460,7 @@ class Program_weight_tensor_parameter_587: class Program_weight_tensor_parameter_588: name = "parameter_588" + original_name = "batch_norm2d_21.w_2" shape = [96] dtype = "float32" min_val = float("0.0386177") @@ -5883,6 +6472,7 @@ class Program_weight_tensor_parameter_588: class Program_weight_tensor_parameter_589: name = "parameter_589" + original_name = "batch_norm2d_21.w_1" shape = [96] dtype = "float32" min_val = float("-0.274091") @@ -5894,6 +6484,7 @@ class Program_weight_tensor_parameter_589: class Program_weight_tensor_parameter_590: name = "parameter_590" + original_name = "conv2d_22.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.47366") @@ -5905,6 +6496,7 @@ class Program_weight_tensor_parameter_590: class Program_weight_tensor_parameter_591: name = "parameter_591" + original_name = "batch_norm2d_20.b_0" shape = [96] dtype = "float32" min_val = float("-0.681237") @@ -5916,6 +6508,7 @@ class Program_weight_tensor_parameter_591: class Program_weight_tensor_parameter_592: name = "parameter_592" + original_name = "batch_norm2d_20.w_0" shape = [96] dtype = "float32" min_val = float("0.275505") @@ -5927,6 +6520,7 @@ class Program_weight_tensor_parameter_592: class Program_weight_tensor_parameter_593: name = "parameter_593" + original_name = "batch_norm2d_20.w_2" shape = [96] dtype = "float32" min_val = float("0.482727") @@ -5938,6 +6532,7 @@ class Program_weight_tensor_parameter_593: class Program_weight_tensor_parameter_594: name = "parameter_594" + original_name = "batch_norm2d_20.w_1" shape = [96] dtype = "float32" min_val = float("-1.52674") @@ -5949,6 +6544,7 @@ class Program_weight_tensor_parameter_594: class Program_weight_tensor_parameter_595: name = "parameter_595" + original_name = "conv2d_21.w_0" shape = [96, 96, 3, 3] dtype = "float32" min_val = float("-0.63607") @@ -5960,6 +6556,7 @@ class Program_weight_tensor_parameter_595: class Program_weight_tensor_parameter_596: name = "parameter_596" + original_name = "batch_norm2d_19.b_0" shape = [96] dtype = "float32" min_val = float("-2.46451") @@ -5971,6 +6568,7 @@ class Program_weight_tensor_parameter_596: class Program_weight_tensor_parameter_597: name = "parameter_597" + original_name = "batch_norm2d_19.w_0" shape = [96] dtype = "float32" min_val = float("0.859783") @@ -5982,6 +6580,7 @@ class Program_weight_tensor_parameter_597: class Program_weight_tensor_parameter_598: name = "parameter_598" + original_name = "batch_norm2d_19.w_2" shape = [96] dtype = "float32" min_val = float("0.795885") @@ -5993,6 +6592,7 @@ class Program_weight_tensor_parameter_598: class Program_weight_tensor_parameter_599: name = "parameter_599" + original_name = "batch_norm2d_19.w_1" shape = [96] dtype = "float32" min_val = float("-5.34734") @@ -6004,6 +6604,7 @@ class Program_weight_tensor_parameter_599: class Program_weight_tensor_parameter_600: name = "parameter_600" + original_name = "conv2d_20.w_0" shape = [96, 96, 3, 3] dtype = "float32" min_val = float("-0.637024") @@ -6015,6 +6616,7 @@ class Program_weight_tensor_parameter_600: class Program_weight_tensor_parameter_601: name = "parameter_601" + original_name = "batch_norm2d_18.b_0" shape = [96] dtype = "float32" min_val = float("-3.39272") @@ -6026,6 +6628,7 @@ class Program_weight_tensor_parameter_601: class Program_weight_tensor_parameter_602: name = "parameter_602" + original_name = "batch_norm2d_18.w_0" shape = [96] dtype = "float32" min_val = float("0.395125") @@ -6037,6 +6640,7 @@ class Program_weight_tensor_parameter_602: class Program_weight_tensor_parameter_603: name = "parameter_603" + original_name = "batch_norm2d_18.w_2" shape = [96] dtype = "float32" min_val = float("0.520837") @@ -6048,6 +6652,7 @@ class Program_weight_tensor_parameter_603: class Program_weight_tensor_parameter_604: name = "parameter_604" + original_name = "batch_norm2d_18.w_1" shape = [96] dtype = "float32" min_val = float("-1.68717") @@ -6059,6 +6664,7 @@ class Program_weight_tensor_parameter_604: class Program_weight_tensor_parameter_605: name = "parameter_605" + original_name = "conv2d_19.w_0" shape = [96, 192, 1, 1] dtype = "float32" min_val = float("-1.22045") @@ -6070,6 +6676,7 @@ class Program_weight_tensor_parameter_605: class Program_weight_tensor_parameter_606: name = "parameter_606" + original_name = "batch_norm2d_17.b_0" shape = [96] dtype = "float32" min_val = float("-5.00988") @@ -6081,6 +6688,7 @@ class Program_weight_tensor_parameter_606: class Program_weight_tensor_parameter_607: name = "parameter_607" + original_name = "batch_norm2d_17.w_0" shape = [96] dtype = "float32" min_val = float("0.640985") @@ -6092,6 +6700,7 @@ class Program_weight_tensor_parameter_607: class Program_weight_tensor_parameter_608: name = "parameter_608" + original_name = "batch_norm2d_17.w_2" shape = [96] dtype = "float32" min_val = float("0.650263") @@ -6103,6 +6712,7 @@ class Program_weight_tensor_parameter_608: class Program_weight_tensor_parameter_609: name = "parameter_609" + original_name = "batch_norm2d_17.w_1" shape = [96] dtype = "float32" min_val = float("-1.2689") @@ -6114,6 +6724,7 @@ class Program_weight_tensor_parameter_609: class Program_weight_tensor_parameter_610: name = "parameter_610" + original_name = "conv2d_18.w_0" shape = [96, 192, 1, 1] dtype = "float32" min_val = float("-1.09266") @@ -6125,6 +6736,7 @@ class Program_weight_tensor_parameter_610: class Program_weight_tensor_parameter_611: name = "parameter_611" + original_name = "batch_norm2d_16.b_0" shape = [192] dtype = "float32" min_val = float("-2.25095") @@ -6136,6 +6748,7 @@ class Program_weight_tensor_parameter_611: class Program_weight_tensor_parameter_612: name = "parameter_612" + original_name = "batch_norm2d_16.w_0" shape = [192] dtype = "float32" min_val = float("0.639639") @@ -6147,6 +6760,7 @@ class Program_weight_tensor_parameter_612: class Program_weight_tensor_parameter_613: name = "parameter_613" + original_name = "batch_norm2d_16.w_2" shape = [192] dtype = "float32" min_val = float("0.882003") @@ -6158,6 +6772,7 @@ class Program_weight_tensor_parameter_613: class Program_weight_tensor_parameter_614: name = "parameter_614" + original_name = "batch_norm2d_16.w_1" shape = [192] dtype = "float32" min_val = float("-2.85994") @@ -6169,6 +6784,7 @@ class Program_weight_tensor_parameter_614: class Program_weight_tensor_parameter_615: name = "parameter_615" + original_name = "conv2d_17.w_0" shape = [192, 128, 3, 3] dtype = "float32" min_val = float("-0.839395") @@ -6180,6 +6796,7 @@ class Program_weight_tensor_parameter_615: class Program_weight_tensor_parameter_616: name = "parameter_616" + original_name = "batch_norm2d_15.b_0" shape = [128] dtype = "float32" min_val = float("-2.74747") @@ -6191,6 +6808,7 @@ class Program_weight_tensor_parameter_616: class Program_weight_tensor_parameter_617: name = "parameter_617" + original_name = "batch_norm2d_15.w_0" shape = [128] dtype = "float32" min_val = float("0.352065") @@ -6202,6 +6820,7 @@ class Program_weight_tensor_parameter_617: class Program_weight_tensor_parameter_618: name = "parameter_618" + original_name = "batch_norm2d_15.w_2" shape = [128] dtype = "float32" min_val = float("0.407101") @@ -6213,6 +6832,7 @@ class Program_weight_tensor_parameter_618: class Program_weight_tensor_parameter_619: name = "parameter_619" + original_name = "batch_norm2d_15.w_1" shape = [128] dtype = "float32" min_val = float("-1.02626") @@ -6224,6 +6844,7 @@ class Program_weight_tensor_parameter_619: class Program_weight_tensor_parameter_620: name = "parameter_620" + original_name = "conv2d_16.w_0" shape = [128, 96, 1, 1] dtype = "float32" min_val = float("-1.79941") @@ -6235,6 +6856,7 @@ class Program_weight_tensor_parameter_620: class Program_weight_tensor_parameter_621: name = "parameter_621" + original_name = "conv2d_15.b_0" shape = [96] dtype = "float32" min_val = float("-0.721414") @@ -6246,6 +6868,7 @@ class Program_weight_tensor_parameter_621: class Program_weight_tensor_parameter_622: name = "parameter_622" + original_name = "conv2d_15.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-2.39477") @@ -6257,6 +6880,7 @@ class Program_weight_tensor_parameter_622: class Program_weight_tensor_parameter_623: name = "parameter_623" + original_name = "batch_norm2d_14.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -6266,6 +6890,7 @@ class Program_weight_tensor_parameter_623: class Program_weight_tensor_parameter_624: name = "parameter_624" + original_name = "batch_norm2d_14.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -6275,6 +6900,7 @@ class Program_weight_tensor_parameter_624: class Program_weight_tensor_parameter_625: name = "parameter_625" + original_name = "batch_norm2d_14.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -6284,6 +6910,7 @@ class Program_weight_tensor_parameter_625: class Program_weight_tensor_parameter_626: name = "parameter_626" + original_name = "batch_norm2d_14.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -6293,6 +6920,7 @@ class Program_weight_tensor_parameter_626: class Program_weight_tensor_parameter_627: name = "parameter_627" + original_name = "conv2d_14.w_0" shape = [48, 48, 1, 1] dtype = "float32" min_val = float("-0.500373") @@ -6304,6 +6932,7 @@ class Program_weight_tensor_parameter_627: class Program_weight_tensor_parameter_628: name = "parameter_628" + original_name = "batch_norm2d_13.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -6313,6 +6942,7 @@ class Program_weight_tensor_parameter_628: class Program_weight_tensor_parameter_629: name = "parameter_629" + original_name = "batch_norm2d_13.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -6322,6 +6952,7 @@ class Program_weight_tensor_parameter_629: class Program_weight_tensor_parameter_630: name = "parameter_630" + original_name = "batch_norm2d_13.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -6331,6 +6962,7 @@ class Program_weight_tensor_parameter_630: class Program_weight_tensor_parameter_631: name = "parameter_631" + original_name = "batch_norm2d_13.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -6340,6 +6972,7 @@ class Program_weight_tensor_parameter_631: class Program_weight_tensor_parameter_632: name = "parameter_632" + original_name = "conv2d_13.w_0" shape = [48, 48, 3, 3] dtype = "float32" min_val = float("-0.391989") @@ -6351,6 +6984,7 @@ class Program_weight_tensor_parameter_632: class Program_weight_tensor_parameter_633: name = "parameter_633" + original_name = "batch_norm2d_12.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -6360,6 +6994,7 @@ class Program_weight_tensor_parameter_633: class Program_weight_tensor_parameter_634: name = "parameter_634" + original_name = "batch_norm2d_12.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -6369,6 +7004,7 @@ class Program_weight_tensor_parameter_634: class Program_weight_tensor_parameter_635: name = "parameter_635" + original_name = "batch_norm2d_12.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -6378,6 +7014,7 @@ class Program_weight_tensor_parameter_635: class Program_weight_tensor_parameter_636: name = "parameter_636" + original_name = "batch_norm2d_12.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -6387,6 +7024,7 @@ class Program_weight_tensor_parameter_636: class Program_weight_tensor_parameter_637: name = "parameter_637" + original_name = "conv2d_12.w_0" shape = [48, 48, 3, 3] dtype = "float32" min_val = float("-0.563675") @@ -6398,6 +7036,7 @@ class Program_weight_tensor_parameter_637: class Program_weight_tensor_parameter_638: name = "parameter_638" + original_name = "batch_norm2d_11.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -6407,6 +7046,7 @@ class Program_weight_tensor_parameter_638: class Program_weight_tensor_parameter_639: name = "parameter_639" + original_name = "batch_norm2d_11.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -6416,6 +7056,7 @@ class Program_weight_tensor_parameter_639: class Program_weight_tensor_parameter_640: name = "parameter_640" + original_name = "batch_norm2d_11.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -6425,6 +7066,7 @@ class Program_weight_tensor_parameter_640: class Program_weight_tensor_parameter_641: name = "parameter_641" + original_name = "batch_norm2d_11.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -6434,6 +7076,7 @@ class Program_weight_tensor_parameter_641: class Program_weight_tensor_parameter_642: name = "parameter_642" + original_name = "conv2d_11.w_0" shape = [48, 48, 1, 1] dtype = "float32" min_val = float("-0.528835") @@ -6445,6 +7088,7 @@ class Program_weight_tensor_parameter_642: class Program_weight_tensor_parameter_643: name = "parameter_643" + original_name = "batch_norm2d_10.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -6454,6 +7098,7 @@ class Program_weight_tensor_parameter_643: class Program_weight_tensor_parameter_644: name = "parameter_644" + original_name = "batch_norm2d_10.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -6463,6 +7108,7 @@ class Program_weight_tensor_parameter_644: class Program_weight_tensor_parameter_645: name = "parameter_645" + original_name = "batch_norm2d_10.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -6472,6 +7118,7 @@ class Program_weight_tensor_parameter_645: class Program_weight_tensor_parameter_646: name = "parameter_646" + original_name = "batch_norm2d_10.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -6481,6 +7128,7 @@ class Program_weight_tensor_parameter_646: class Program_weight_tensor_parameter_647: name = "parameter_647" + original_name = "conv2d_10.w_0" shape = [48, 48, 3, 3] dtype = "float32" min_val = float("-0.399429") @@ -6492,6 +7140,7 @@ class Program_weight_tensor_parameter_647: class Program_weight_tensor_parameter_648: name = "parameter_648" + original_name = "batch_norm2d_9.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -6501,6 +7150,7 @@ class Program_weight_tensor_parameter_648: class Program_weight_tensor_parameter_649: name = "parameter_649" + original_name = "batch_norm2d_9.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -6510,6 +7160,7 @@ class Program_weight_tensor_parameter_649: class Program_weight_tensor_parameter_650: name = "parameter_650" + original_name = "batch_norm2d_9.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -6519,6 +7170,7 @@ class Program_weight_tensor_parameter_650: class Program_weight_tensor_parameter_651: name = "parameter_651" + original_name = "batch_norm2d_9.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -6528,6 +7180,7 @@ class Program_weight_tensor_parameter_651: class Program_weight_tensor_parameter_652: name = "parameter_652" + original_name = "conv2d_9.w_0" shape = [48, 48, 3, 3] dtype = "float32" min_val = float("-0.555547") @@ -6539,6 +7192,7 @@ class Program_weight_tensor_parameter_652: class Program_weight_tensor_parameter_653: name = "parameter_653" + original_name = "batch_norm2d_8.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -6548,6 +7202,7 @@ class Program_weight_tensor_parameter_653: class Program_weight_tensor_parameter_654: name = "parameter_654" + original_name = "batch_norm2d_8.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -6557,6 +7212,7 @@ class Program_weight_tensor_parameter_654: class Program_weight_tensor_parameter_655: name = "parameter_655" + original_name = "batch_norm2d_8.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -6566,6 +7222,7 @@ class Program_weight_tensor_parameter_655: class Program_weight_tensor_parameter_656: name = "parameter_656" + original_name = "batch_norm2d_8.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -6575,6 +7232,7 @@ class Program_weight_tensor_parameter_656: class Program_weight_tensor_parameter_657: name = "parameter_657" + original_name = "conv2d_8.w_0" shape = [48, 48, 1, 1] dtype = "float32" min_val = float("-0.851714") @@ -6586,6 +7244,7 @@ class Program_weight_tensor_parameter_657: class Program_weight_tensor_parameter_658: name = "parameter_658" + original_name = "batch_norm2d_7.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -6595,6 +7254,7 @@ class Program_weight_tensor_parameter_658: class Program_weight_tensor_parameter_659: name = "parameter_659" + original_name = "batch_norm2d_7.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -6604,6 +7264,7 @@ class Program_weight_tensor_parameter_659: class Program_weight_tensor_parameter_660: name = "parameter_660" + original_name = "batch_norm2d_7.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -6613,6 +7274,7 @@ class Program_weight_tensor_parameter_660: class Program_weight_tensor_parameter_661: name = "parameter_661" + original_name = "batch_norm2d_7.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -6622,6 +7284,7 @@ class Program_weight_tensor_parameter_661: class Program_weight_tensor_parameter_662: name = "parameter_662" + original_name = "conv2d_7.w_0" shape = [48, 48, 3, 3] dtype = "float32" min_val = float("-0.609572") @@ -6633,6 +7296,7 @@ class Program_weight_tensor_parameter_662: class Program_weight_tensor_parameter_663: name = "parameter_663" + original_name = "batch_norm2d_6.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -6642,6 +7306,7 @@ class Program_weight_tensor_parameter_663: class Program_weight_tensor_parameter_664: name = "parameter_664" + original_name = "batch_norm2d_6.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -6651,6 +7316,7 @@ class Program_weight_tensor_parameter_664: class Program_weight_tensor_parameter_665: name = "parameter_665" + original_name = "batch_norm2d_6.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -6660,6 +7326,7 @@ class Program_weight_tensor_parameter_665: class Program_weight_tensor_parameter_666: name = "parameter_666" + original_name = "batch_norm2d_6.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -6669,6 +7336,7 @@ class Program_weight_tensor_parameter_666: class Program_weight_tensor_parameter_667: name = "parameter_667" + original_name = "conv2d_6.w_0" shape = [48, 48, 3, 3] dtype = "float32" min_val = float("-0.796944") @@ -6680,6 +7348,7 @@ class Program_weight_tensor_parameter_667: class Program_weight_tensor_parameter_668: name = "parameter_668" + original_name = "batch_norm2d_5.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -6689,6 +7358,7 @@ class Program_weight_tensor_parameter_668: class Program_weight_tensor_parameter_669: name = "parameter_669" + original_name = "batch_norm2d_5.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -6698,6 +7368,7 @@ class Program_weight_tensor_parameter_669: class Program_weight_tensor_parameter_670: name = "parameter_670" + original_name = "batch_norm2d_5.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -6707,6 +7378,7 @@ class Program_weight_tensor_parameter_670: class Program_weight_tensor_parameter_671: name = "parameter_671" + original_name = "batch_norm2d_5.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -6716,6 +7388,7 @@ class Program_weight_tensor_parameter_671: class Program_weight_tensor_parameter_672: name = "parameter_672" + original_name = "conv2d_5.w_0" shape = [48, 96, 1, 1] dtype = "float32" min_val = float("-1.66802") @@ -6727,6 +7400,7 @@ class Program_weight_tensor_parameter_672: class Program_weight_tensor_parameter_673: name = "parameter_673" + original_name = "batch_norm2d_4.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -6736,6 +7410,7 @@ class Program_weight_tensor_parameter_673: class Program_weight_tensor_parameter_674: name = "parameter_674" + original_name = "batch_norm2d_4.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -6745,6 +7420,7 @@ class Program_weight_tensor_parameter_674: class Program_weight_tensor_parameter_675: name = "parameter_675" + original_name = "batch_norm2d_4.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -6754,6 +7430,7 @@ class Program_weight_tensor_parameter_675: class Program_weight_tensor_parameter_676: name = "parameter_676" + original_name = "batch_norm2d_4.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -6763,6 +7440,7 @@ class Program_weight_tensor_parameter_676: class Program_weight_tensor_parameter_677: name = "parameter_677" + original_name = "conv2d_4.w_0" shape = [48, 96, 1, 1] dtype = "float32" min_val = float("-1.06553") @@ -6774,6 +7452,7 @@ class Program_weight_tensor_parameter_677: class Program_weight_tensor_parameter_678: name = "parameter_678" + original_name = "batch_norm2d_3.b_0" shape = [96] dtype = "float32" min_val = float("-3.6652") @@ -6785,6 +7464,7 @@ class Program_weight_tensor_parameter_678: class Program_weight_tensor_parameter_679: name = "parameter_679" + original_name = "batch_norm2d_3.w_0" shape = [96] dtype = "float32" min_val = float("0.972278") @@ -6796,6 +7476,7 @@ class Program_weight_tensor_parameter_679: class Program_weight_tensor_parameter_680: name = "parameter_680" + original_name = "batch_norm2d_3.w_2" shape = [96] dtype = "float32" min_val = float("24.1873") @@ -6807,6 +7488,7 @@ class Program_weight_tensor_parameter_680: class Program_weight_tensor_parameter_681: name = "parameter_681" + original_name = "batch_norm2d_3.w_1" shape = [96] dtype = "float32" min_val = float("-14.6306") @@ -6818,6 +7500,7 @@ class Program_weight_tensor_parameter_681: class Program_weight_tensor_parameter_682: name = "parameter_682" + original_name = "conv2d_3.w_0" shape = [96, 64, 3, 3] dtype = "float32" min_val = float("-1.17881") @@ -6829,6 +7512,7 @@ class Program_weight_tensor_parameter_682: class Program_weight_tensor_parameter_683: name = "parameter_683" + original_name = "batch_norm2d_2.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -6838,6 +7522,7 @@ class Program_weight_tensor_parameter_683: class Program_weight_tensor_parameter_684: name = "parameter_684" + original_name = "batch_norm2d_2.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -6847,6 +7532,7 @@ class Program_weight_tensor_parameter_684: class Program_weight_tensor_parameter_685: name = "parameter_685" + original_name = "batch_norm2d_2.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -6856,6 +7542,7 @@ class Program_weight_tensor_parameter_685: class Program_weight_tensor_parameter_686: name = "parameter_686" + original_name = "batch_norm2d_2.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -6865,6 +7552,7 @@ class Program_weight_tensor_parameter_686: class Program_weight_tensor_parameter_687: name = "parameter_687" + original_name = "conv2d_2.w_0" shape = [64, 32, 3, 3] dtype = "float32" min_val = float("-1.49789") @@ -6876,6 +7564,7 @@ class Program_weight_tensor_parameter_687: class Program_weight_tensor_parameter_688: name = "parameter_688" + original_name = "batch_norm2d_1.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -6885,6 +7574,7 @@ class Program_weight_tensor_parameter_688: class Program_weight_tensor_parameter_689: name = "parameter_689" + original_name = "batch_norm2d_1.w_0" shape = [32] dtype = "float32" min_val = float("0") @@ -6894,6 +7584,7 @@ class Program_weight_tensor_parameter_689: class Program_weight_tensor_parameter_690: name = "parameter_690" + original_name = "batch_norm2d_1.w_2" shape = [32] dtype = "float32" min_val = float("0") @@ -6903,6 +7594,7 @@ class Program_weight_tensor_parameter_690: class Program_weight_tensor_parameter_691: name = "parameter_691" + original_name = "batch_norm2d_1.w_1" shape = [32] dtype = "float32" min_val = float("0") @@ -6912,6 +7604,7 @@ class Program_weight_tensor_parameter_691: class Program_weight_tensor_parameter_692: name = "parameter_692" + original_name = "conv2d_1.w_0" shape = [32, 32, 3, 3] dtype = "float32" min_val = float("-2.24893") @@ -6923,6 +7616,7 @@ class Program_weight_tensor_parameter_692: class Program_weight_tensor_parameter_693: name = "parameter_693" + original_name = "batch_norm2d_0.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -6932,6 +7626,7 @@ class Program_weight_tensor_parameter_693: class Program_weight_tensor_parameter_694: name = "parameter_694" + original_name = "batch_norm2d_0.w_0" shape = [32] dtype = "float32" min_val = float("0") @@ -6941,6 +7636,7 @@ class Program_weight_tensor_parameter_694: class Program_weight_tensor_parameter_695: name = "parameter_695" + original_name = "batch_norm2d_0.w_2" shape = [32] dtype = "float32" min_val = float("0") @@ -6950,6 +7646,7 @@ class Program_weight_tensor_parameter_695: class Program_weight_tensor_parameter_696: name = "parameter_696" + original_name = "batch_norm2d_0.w_1" shape = [32] dtype = "float32" min_val = float("0") @@ -6959,6 +7656,7 @@ class Program_weight_tensor_parameter_696: class Program_weight_tensor_parameter_697: name = "parameter_697" + original_name = "conv2d_0.w_0" shape = [32, 3, 3, 3] dtype = "float32" min_val = float("-2.75119") diff --git a/paddle_samples/PaddleX/PP-YOLOE_plus-L/subgraph_4/input_meta.py b/paddle_samples/PaddleX/PP-YOLOE_plus-L/subgraph_4/input_meta.py index 7738acf83..514d7c75b 100644 --- a/paddle_samples/PaddleX/PP-YOLOE_plus-L/subgraph_4/input_meta.py +++ b/paddle_samples/PaddleX/PP-YOLOE_plus-L/subgraph_4/input_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_data_0: name = "data_0" + original_name = "param_36" shape = [1] dtype = "float32" data = [1.0] @@ -7,6 +8,7 @@ class Program_weight_tensor_data_0: class Program_weight_tensor_data_1: name = "data_1" + original_name = "param_37" shape = [1] dtype = "float32" data = [1.0] @@ -14,6 +16,7 @@ class Program_weight_tensor_data_1: class Program_weight_tensor_data_2: name = "data_2" + original_name = "param_38" shape = [1] dtype = "float32" data = [1.0] @@ -21,6 +24,7 @@ class Program_weight_tensor_data_2: class Program_weight_tensor_data_3: name = "data_3" + original_name = "param_39" shape = [1] dtype = "float32" data = [1.0] @@ -28,6 +32,7 @@ class Program_weight_tensor_data_3: class Program_weight_tensor_data_4: name = "data_4" + original_name = "param_40" shape = [1] dtype = "float32" data = [1.0] @@ -35,6 +40,7 @@ class Program_weight_tensor_data_4: class Program_weight_tensor_data_5: name = "data_5" + original_name = "param_41" shape = [1] dtype = "float32" data = [1.0] @@ -42,6 +48,7 @@ class Program_weight_tensor_data_5: class Program_weight_tensor_data_6: name = "data_6" + original_name = "param_42" shape = [1] dtype = "float32" data = [1.0] @@ -49,6 +56,7 @@ class Program_weight_tensor_data_6: class Program_weight_tensor_data_7: name = "data_7" + original_name = "param_43" shape = [1] dtype = "float32" data = [1.0] @@ -56,6 +64,7 @@ class Program_weight_tensor_data_7: class Program_weight_tensor_data_8: name = "data_8" + original_name = "param_44" shape = [1] dtype = "float32" data = [1.0] @@ -63,6 +72,7 @@ class Program_weight_tensor_data_8: class Program_weight_tensor_data_9: name = "data_9" + original_name = "param_45" shape = [1] dtype = "float32" data = [1.0] @@ -70,6 +80,7 @@ class Program_weight_tensor_data_9: class Program_weight_tensor_data_10: name = "data_10" + original_name = "param_46" shape = [1] dtype = "float32" data = [1.0] @@ -77,6 +88,7 @@ class Program_weight_tensor_data_10: class Program_weight_tensor_data_11: name = "data_11" + original_name = "param_47" shape = [1] dtype = "float32" data = [1.0] @@ -84,6 +96,7 @@ class Program_weight_tensor_data_11: class Program_weight_tensor_data_12: name = "data_12" + original_name = "param_48" shape = [1] dtype = "float32" data = [1.0] @@ -91,6 +104,7 @@ class Program_weight_tensor_data_12: class Program_weight_tensor_data_13: name = "data_13" + original_name = "param_49" shape = [1] dtype = "float32" data = [1.0] @@ -98,6 +112,7 @@ class Program_weight_tensor_data_13: class Program_weight_tensor_data_14: name = "data_14" + original_name = "param_50" shape = [1] dtype = "float32" data = [1.0] @@ -105,6 +120,7 @@ class Program_weight_tensor_data_14: class Program_weight_tensor_data_15: name = "data_15" + original_name = "param_51" shape = [1] dtype = "float32" data = [1.0] @@ -112,6 +128,7 @@ class Program_weight_tensor_data_15: class Program_weight_tensor_data_16: name = "data_16" + original_name = "param_52" shape = [1] dtype = "float32" data = [1.0] @@ -119,6 +136,7 @@ class Program_weight_tensor_data_16: class Program_weight_tensor_data_17: name = "data_17" + original_name = "param_53" shape = [1] dtype = "float32" data = [1.0] @@ -126,6 +144,7 @@ class Program_weight_tensor_data_17: class Program_weight_tensor_data_18: name = "data_18" + original_name = "var_1382" shape = [8, 3, 448, 448] dtype = "float32" max_val = float("1.0") diff --git a/paddle_samples/PaddleX/PP-YOLOE_plus-L/subgraph_4/weight_meta.py b/paddle_samples/PaddleX/PP-YOLOE_plus-L/subgraph_4/weight_meta.py index 3e3459082..1407be49a 100644 --- a/paddle_samples/PaddleX/PP-YOLOE_plus-L/subgraph_4/weight_meta.py +++ b/paddle_samples/PaddleX/PP-YOLOE_plus-L/subgraph_4/weight_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_parameter_0: name = "parameter_0" + original_name = "batch_norm2d_137.b_0" shape = [768] dtype = "float32" min_val = float("-0.174892") @@ -11,6 +12,7 @@ class Program_weight_tensor_parameter_0: class Program_weight_tensor_parameter_1: name = "parameter_1" + original_name = "batch_norm2d_137.w_0" shape = [768] dtype = "float32" min_val = float("0.937556") @@ -22,6 +24,7 @@ class Program_weight_tensor_parameter_1: class Program_weight_tensor_parameter_2: name = "parameter_2" + original_name = "batch_norm2d_137.w_2" shape = [768] dtype = "float32" min_val = float("0.000993105") @@ -33,6 +36,7 @@ class Program_weight_tensor_parameter_2: class Program_weight_tensor_parameter_3: name = "parameter_3" + original_name = "batch_norm2d_137.w_1" shape = [768] dtype = "float32" min_val = float("-0.148256") @@ -44,6 +48,7 @@ class Program_weight_tensor_parameter_3: class Program_weight_tensor_parameter_4: name = "parameter_4" + original_name = "conv2d_141.w_0" shape = [768, 768, 1, 1] dtype = "float32" min_val = float("-0.0483446") @@ -55,6 +60,7 @@ class Program_weight_tensor_parameter_4: class Program_weight_tensor_parameter_5: name = "parameter_5" + original_name = "batch_norm2d_136.b_0" shape = [384] dtype = "float32" min_val = float("-0.138311") @@ -66,6 +72,7 @@ class Program_weight_tensor_parameter_5: class Program_weight_tensor_parameter_6: name = "parameter_6" + original_name = "batch_norm2d_136.w_0" shape = [384] dtype = "float32" min_val = float("0.94769") @@ -77,6 +84,7 @@ class Program_weight_tensor_parameter_6: class Program_weight_tensor_parameter_7: name = "parameter_7" + original_name = "batch_norm2d_136.w_2" shape = [384] dtype = "float32" min_val = float("0.000385121") @@ -88,6 +96,7 @@ class Program_weight_tensor_parameter_7: class Program_weight_tensor_parameter_8: name = "parameter_8" + original_name = "batch_norm2d_136.w_1" shape = [384] dtype = "float32" min_val = float("-0.056999") @@ -99,6 +108,7 @@ class Program_weight_tensor_parameter_8: class Program_weight_tensor_parameter_9: name = "parameter_9" + original_name = "conv2d_140.w_0" shape = [384, 384, 1, 1] dtype = "float32" min_val = float("-0.0299557") @@ -110,6 +120,7 @@ class Program_weight_tensor_parameter_9: class Program_weight_tensor_parameter_10: name = "parameter_10" + original_name = "batch_norm2d_135.b_0" shape = [384] dtype = "float32" min_val = float("-0.138311") @@ -121,6 +132,7 @@ class Program_weight_tensor_parameter_10: class Program_weight_tensor_parameter_11: name = "parameter_11" + original_name = "batch_norm2d_135.w_0" shape = [384] dtype = "float32" min_val = float("0.96766") @@ -132,6 +144,7 @@ class Program_weight_tensor_parameter_11: class Program_weight_tensor_parameter_12: name = "parameter_12" + original_name = "batch_norm2d_135.w_2" shape = [384] dtype = "float32" min_val = float("0.00123062") @@ -143,6 +156,7 @@ class Program_weight_tensor_parameter_12: class Program_weight_tensor_parameter_13: name = "parameter_13" + original_name = "batch_norm2d_135.w_1" shape = [384] dtype = "float32" min_val = float("-0.111547") @@ -154,6 +168,7 @@ class Program_weight_tensor_parameter_13: class Program_weight_tensor_parameter_14: name = "parameter_14" + original_name = "conv2d_139.w_0" shape = [384, 384, 3, 3] dtype = "float32" min_val = float("-0.0252101") @@ -165,6 +180,7 @@ class Program_weight_tensor_parameter_14: class Program_weight_tensor_parameter_15: name = "parameter_15" + original_name = "batch_norm2d_134.b_0" shape = [384] dtype = "float32" min_val = float("-0.166952") @@ -176,6 +192,7 @@ class Program_weight_tensor_parameter_15: class Program_weight_tensor_parameter_16: name = "parameter_16" + original_name = "batch_norm2d_134.w_0" shape = [384] dtype = "float32" min_val = float("0.975112") @@ -187,6 +204,7 @@ class Program_weight_tensor_parameter_16: class Program_weight_tensor_parameter_17: name = "parameter_17" + original_name = "batch_norm2d_134.w_2" shape = [384] dtype = "float32" min_val = float("0.00332451") @@ -198,6 +216,7 @@ class Program_weight_tensor_parameter_17: class Program_weight_tensor_parameter_18: name = "parameter_18" + original_name = "batch_norm2d_134.w_1" shape = [384] dtype = "float32" min_val = float("-0.183511") @@ -209,6 +228,7 @@ class Program_weight_tensor_parameter_18: class Program_weight_tensor_parameter_19: name = "parameter_19" + original_name = "conv2d_138.w_0" shape = [384, 384, 3, 3] dtype = "float32" min_val = float("-0.0246938") @@ -220,6 +240,7 @@ class Program_weight_tensor_parameter_19: class Program_weight_tensor_parameter_20: name = "parameter_20" + original_name = "batch_norm2d_133.b_0" shape = [384] dtype = "float32" min_val = float("-0.101971") @@ -231,6 +252,7 @@ class Program_weight_tensor_parameter_20: class Program_weight_tensor_parameter_21: name = "parameter_21" + original_name = "batch_norm2d_133.w_0" shape = [384] dtype = "float32" min_val = float("0.947094") @@ -242,6 +264,7 @@ class Program_weight_tensor_parameter_21: class Program_weight_tensor_parameter_22: name = "parameter_22" + original_name = "batch_norm2d_133.w_2" shape = [384] dtype = "float32" min_val = float("0.000608792") @@ -253,6 +276,7 @@ class Program_weight_tensor_parameter_22: class Program_weight_tensor_parameter_23: name = "parameter_23" + original_name = "batch_norm2d_133.w_1" shape = [384] dtype = "float32" min_val = float("-0.0673723") @@ -264,6 +288,7 @@ class Program_weight_tensor_parameter_23: class Program_weight_tensor_parameter_24: name = "parameter_24" + original_name = "conv2d_137.w_0" shape = [384, 384, 1, 1] dtype = "float32" min_val = float("-0.023885") @@ -275,6 +300,7 @@ class Program_weight_tensor_parameter_24: class Program_weight_tensor_parameter_25: name = "parameter_25" + original_name = "batch_norm2d_132.b_0" shape = [384] dtype = "float32" min_val = float("-0.101971") @@ -286,6 +312,7 @@ class Program_weight_tensor_parameter_25: class Program_weight_tensor_parameter_26: name = "parameter_26" + original_name = "batch_norm2d_132.w_0" shape = [384] dtype = "float32" min_val = float("0.959612") @@ -297,6 +324,7 @@ class Program_weight_tensor_parameter_26: class Program_weight_tensor_parameter_27: name = "parameter_27" + original_name = "batch_norm2d_132.w_2" shape = [384] dtype = "float32" min_val = float("0.00180756") @@ -308,6 +336,7 @@ class Program_weight_tensor_parameter_27: class Program_weight_tensor_parameter_28: name = "parameter_28" + original_name = "batch_norm2d_132.w_1" shape = [384] dtype = "float32" min_val = float("-0.110394") @@ -319,6 +348,7 @@ class Program_weight_tensor_parameter_28: class Program_weight_tensor_parameter_29: name = "parameter_29" + original_name = "conv2d_136.w_0" shape = [384, 384, 3, 3] dtype = "float32" min_val = float("-0.0305313") @@ -330,6 +360,7 @@ class Program_weight_tensor_parameter_29: class Program_weight_tensor_parameter_30: name = "parameter_30" + original_name = "batch_norm2d_131.b_0" shape = [384] dtype = "float32" min_val = float("-0.0876863") @@ -341,6 +372,7 @@ class Program_weight_tensor_parameter_30: class Program_weight_tensor_parameter_31: name = "parameter_31" + original_name = "batch_norm2d_131.w_0" shape = [384] dtype = "float32" min_val = float("0.935469") @@ -352,6 +384,7 @@ class Program_weight_tensor_parameter_31: class Program_weight_tensor_parameter_32: name = "parameter_32" + original_name = "batch_norm2d_131.w_2" shape = [384] dtype = "float32" min_val = float("0.00350475") @@ -363,6 +396,7 @@ class Program_weight_tensor_parameter_32: class Program_weight_tensor_parameter_33: name = "parameter_33" + original_name = "batch_norm2d_131.w_1" shape = [384] dtype = "float32" min_val = float("-0.111072") @@ -374,6 +408,7 @@ class Program_weight_tensor_parameter_33: class Program_weight_tensor_parameter_34: name = "parameter_34" + original_name = "conv2d_135.w_0" shape = [384, 384, 3, 3] dtype = "float32" min_val = float("-0.03558") @@ -385,6 +420,7 @@ class Program_weight_tensor_parameter_34: class Program_weight_tensor_parameter_35: name = "parameter_35" + original_name = "batch_norm2d_130.b_0" shape = [384] dtype = "float32" min_val = float("-0.112996") @@ -396,6 +432,7 @@ class Program_weight_tensor_parameter_35: class Program_weight_tensor_parameter_36: name = "parameter_36" + original_name = "batch_norm2d_130.w_0" shape = [384] dtype = "float32" min_val = float("0.930402") @@ -407,6 +444,7 @@ class Program_weight_tensor_parameter_36: class Program_weight_tensor_parameter_37: name = "parameter_37" + original_name = "batch_norm2d_130.w_2" shape = [384] dtype = "float32" min_val = float("0.000834798") @@ -418,6 +456,7 @@ class Program_weight_tensor_parameter_37: class Program_weight_tensor_parameter_38: name = "parameter_38" + original_name = "batch_norm2d_130.w_1" shape = [384] dtype = "float32" min_val = float("-0.0318765") @@ -429,6 +468,7 @@ class Program_weight_tensor_parameter_38: class Program_weight_tensor_parameter_39: name = "parameter_39" + original_name = "conv2d_134.w_0" shape = [384, 384, 1, 1] dtype = "float32" min_val = float("-0.0246045") @@ -440,6 +480,7 @@ class Program_weight_tensor_parameter_39: class Program_weight_tensor_parameter_40: name = "parameter_40" + original_name = "batch_norm2d_129.b_0" shape = [384] dtype = "float32" min_val = float("-0.112996") @@ -451,6 +492,7 @@ class Program_weight_tensor_parameter_40: class Program_weight_tensor_parameter_41: name = "parameter_41" + original_name = "batch_norm2d_129.w_0" shape = [384] dtype = "float32" min_val = float("0.981669") @@ -462,6 +504,7 @@ class Program_weight_tensor_parameter_41: class Program_weight_tensor_parameter_42: name = "parameter_42" + original_name = "batch_norm2d_129.w_2" shape = [384] dtype = "float32" min_val = float("0.00248832") @@ -473,6 +516,7 @@ class Program_weight_tensor_parameter_42: class Program_weight_tensor_parameter_43: name = "parameter_43" + original_name = "batch_norm2d_129.w_1" shape = [384] dtype = "float32" min_val = float("-0.119239") @@ -484,6 +528,7 @@ class Program_weight_tensor_parameter_43: class Program_weight_tensor_parameter_44: name = "parameter_44" + original_name = "conv2d_133.w_0" shape = [384, 384, 3, 3] dtype = "float32" min_val = float("-0.0269935") @@ -495,6 +540,7 @@ class Program_weight_tensor_parameter_44: class Program_weight_tensor_parameter_45: name = "parameter_45" + original_name = "batch_norm2d_128.b_0" shape = [384] dtype = "float32" min_val = float("-0.104569") @@ -506,6 +552,7 @@ class Program_weight_tensor_parameter_45: class Program_weight_tensor_parameter_46: name = "parameter_46" + original_name = "batch_norm2d_128.w_0" shape = [384] dtype = "float32" min_val = float("0.946498") @@ -517,6 +564,7 @@ class Program_weight_tensor_parameter_46: class Program_weight_tensor_parameter_47: name = "parameter_47" + original_name = "batch_norm2d_128.w_2" shape = [384] dtype = "float32" min_val = float("0.00353592") @@ -528,6 +576,7 @@ class Program_weight_tensor_parameter_47: class Program_weight_tensor_parameter_48: name = "parameter_48" + original_name = "batch_norm2d_128.w_1" shape = [384] dtype = "float32" min_val = float("-0.105865") @@ -539,6 +588,7 @@ class Program_weight_tensor_parameter_48: class Program_weight_tensor_parameter_49: name = "parameter_49" + original_name = "conv2d_132.w_0" shape = [384, 384, 3, 3] dtype = "float32" min_val = float("-0.019792") @@ -550,6 +600,7 @@ class Program_weight_tensor_parameter_49: class Program_weight_tensor_parameter_50: name = "parameter_50" + original_name = "batch_norm2d_127.b_0" shape = [384] dtype = "float32" min_val = float("-0.103835") @@ -561,6 +612,7 @@ class Program_weight_tensor_parameter_50: class Program_weight_tensor_parameter_51: name = "parameter_51" + original_name = "batch_norm2d_127.w_0" shape = [384] dtype = "float32" min_val = float("0.974516") @@ -572,6 +624,7 @@ class Program_weight_tensor_parameter_51: class Program_weight_tensor_parameter_52: name = "parameter_52" + original_name = "batch_norm2d_127.w_2" shape = [384] dtype = "float32" min_val = float("0.00205733") @@ -583,6 +636,7 @@ class Program_weight_tensor_parameter_52: class Program_weight_tensor_parameter_53: name = "parameter_53" + original_name = "batch_norm2d_127.w_1" shape = [384] dtype = "float32" min_val = float("-0.0618369") @@ -594,6 +648,7 @@ class Program_weight_tensor_parameter_53: class Program_weight_tensor_parameter_54: name = "parameter_54" + original_name = "conv2d_131.w_0" shape = [384, 1152, 1, 1] dtype = "float32" min_val = float("-0.0510992") @@ -605,6 +660,7 @@ class Program_weight_tensor_parameter_54: class Program_weight_tensor_parameter_55: name = "parameter_55" + original_name = "batch_norm2d_126.b_0" shape = [384] dtype = "float32" min_val = float("-0.0409565") @@ -616,6 +672,7 @@ class Program_weight_tensor_parameter_55: class Program_weight_tensor_parameter_56: name = "parameter_56" + original_name = "batch_norm2d_126.w_0" shape = [384] dtype = "float32" min_val = float("0.962593") @@ -627,6 +684,7 @@ class Program_weight_tensor_parameter_56: class Program_weight_tensor_parameter_57: name = "parameter_57" + original_name = "batch_norm2d_126.w_2" shape = [384] dtype = "float32" min_val = float("0.00152991") @@ -638,6 +696,7 @@ class Program_weight_tensor_parameter_57: class Program_weight_tensor_parameter_58: name = "parameter_58" + original_name = "batch_norm2d_126.w_1" shape = [384] dtype = "float32" min_val = float("-0.0602306") @@ -649,6 +708,7 @@ class Program_weight_tensor_parameter_58: class Program_weight_tensor_parameter_59: name = "parameter_59" + original_name = "conv2d_130.w_0" shape = [384, 1152, 1, 1] dtype = "float32" min_val = float("-0.0415169") @@ -660,6 +720,7 @@ class Program_weight_tensor_parameter_59: class Program_weight_tensor_parameter_60: name = "parameter_60" + original_name = "batch_norm2d_125.b_0" shape = [384] dtype = "float32" min_val = float("-0.0516134") @@ -671,6 +732,7 @@ class Program_weight_tensor_parameter_60: class Program_weight_tensor_parameter_61: name = "parameter_61" + original_name = "batch_norm2d_125.w_0" shape = [384] dtype = "float32" min_val = float("0.988237") @@ -682,6 +744,7 @@ class Program_weight_tensor_parameter_61: class Program_weight_tensor_parameter_62: name = "parameter_62" + original_name = "batch_norm2d_125.w_2" shape = [384] dtype = "float32" min_val = float("0.00262887") @@ -693,6 +756,7 @@ class Program_weight_tensor_parameter_62: class Program_weight_tensor_parameter_63: name = "parameter_63" + original_name = "batch_norm2d_125.w_1" shape = [384] dtype = "float32" min_val = float("-0.255923") @@ -704,6 +768,7 @@ class Program_weight_tensor_parameter_63: class Program_weight_tensor_parameter_64: name = "parameter_64" + original_name = "conv2d_129.w_0" shape = [384, 384, 3, 3] dtype = "float32" min_val = float("-0.0248827") @@ -715,6 +780,7 @@ class Program_weight_tensor_parameter_64: class Program_weight_tensor_parameter_65: name = "parameter_65" + original_name = "batch_norm2d_124.b_0" shape = [384] dtype = "float32" min_val = float("-0.222232") @@ -726,6 +792,7 @@ class Program_weight_tensor_parameter_65: class Program_weight_tensor_parameter_66: name = "parameter_66" + original_name = "batch_norm2d_124.w_0" shape = [384] dtype = "float32" min_val = float("0.924739") @@ -737,6 +804,7 @@ class Program_weight_tensor_parameter_66: class Program_weight_tensor_parameter_67: name = "parameter_67" + original_name = "batch_norm2d_124.w_2" shape = [384] dtype = "float32" min_val = float("0.00274326") @@ -748,6 +816,7 @@ class Program_weight_tensor_parameter_67: class Program_weight_tensor_parameter_68: name = "parameter_68" + original_name = "batch_norm2d_124.w_1" shape = [384] dtype = "float32" min_val = float("-0.156631") @@ -759,6 +828,7 @@ class Program_weight_tensor_parameter_68: class Program_weight_tensor_parameter_69: name = "parameter_69" + original_name = "conv2d_128.w_0" shape = [384, 384, 1, 1] dtype = "float32" min_val = float("-0.0874231") @@ -770,6 +840,7 @@ class Program_weight_tensor_parameter_69: class Program_weight_tensor_parameter_70: name = "parameter_70" + original_name = "batch_norm2d_123.b_0" shape = [192] dtype = "float32" min_val = float("-0.164047") @@ -781,6 +852,7 @@ class Program_weight_tensor_parameter_70: class Program_weight_tensor_parameter_71: name = "parameter_71" + original_name = "batch_norm2d_123.w_0" shape = [192] dtype = "float32" min_val = float("0.844262") @@ -792,6 +864,7 @@ class Program_weight_tensor_parameter_71: class Program_weight_tensor_parameter_72: name = "parameter_72" + original_name = "batch_norm2d_123.w_2" shape = [192] dtype = "float32" min_val = float("0.000590004") @@ -803,6 +876,7 @@ class Program_weight_tensor_parameter_72: class Program_weight_tensor_parameter_73: name = "parameter_73" + original_name = "batch_norm2d_123.w_1" shape = [192] dtype = "float32" min_val = float("-0.0504661") @@ -814,6 +888,7 @@ class Program_weight_tensor_parameter_73: class Program_weight_tensor_parameter_74: name = "parameter_74" + original_name = "conv2d_127.w_0" shape = [192, 192, 1, 1] dtype = "float32" min_val = float("-0.0501829") @@ -825,6 +900,7 @@ class Program_weight_tensor_parameter_74: class Program_weight_tensor_parameter_75: name = "parameter_75" + original_name = "batch_norm2d_122.b_0" shape = [192] dtype = "float32" min_val = float("-0.164047") @@ -836,6 +912,7 @@ class Program_weight_tensor_parameter_75: class Program_weight_tensor_parameter_76: name = "parameter_76" + original_name = "batch_norm2d_122.w_0" shape = [192] dtype = "float32" min_val = float("0.734277") @@ -847,6 +924,7 @@ class Program_weight_tensor_parameter_76: class Program_weight_tensor_parameter_77: name = "parameter_77" + original_name = "batch_norm2d_122.w_2" shape = [192] dtype = "float32" min_val = float("0.00271339") @@ -858,6 +936,7 @@ class Program_weight_tensor_parameter_77: class Program_weight_tensor_parameter_78: name = "parameter_78" + original_name = "batch_norm2d_122.w_1" shape = [192] dtype = "float32" min_val = float("-0.159305") @@ -869,6 +948,7 @@ class Program_weight_tensor_parameter_78: class Program_weight_tensor_parameter_79: name = "parameter_79" + original_name = "conv2d_126.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.0443179") @@ -880,6 +960,7 @@ class Program_weight_tensor_parameter_79: class Program_weight_tensor_parameter_80: name = "parameter_80" + original_name = "batch_norm2d_121.b_0" shape = [192] dtype = "float32" min_val = float("-0.188321") @@ -891,6 +972,7 @@ class Program_weight_tensor_parameter_80: class Program_weight_tensor_parameter_81: name = "parameter_81" + original_name = "batch_norm2d_121.w_0" shape = [192] dtype = "float32" min_val = float("0.89851") @@ -902,6 +984,7 @@ class Program_weight_tensor_parameter_81: class Program_weight_tensor_parameter_82: name = "parameter_82" + original_name = "batch_norm2d_121.w_2" shape = [192] dtype = "float32" min_val = float("0.00548443") @@ -913,6 +996,7 @@ class Program_weight_tensor_parameter_82: class Program_weight_tensor_parameter_83: name = "parameter_83" + original_name = "batch_norm2d_121.w_1" shape = [192] dtype = "float32" min_val = float("-0.233482") @@ -924,6 +1008,7 @@ class Program_weight_tensor_parameter_83: class Program_weight_tensor_parameter_84: name = "parameter_84" + original_name = "conv2d_125.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.0579255") @@ -935,6 +1020,7 @@ class Program_weight_tensor_parameter_84: class Program_weight_tensor_parameter_85: name = "parameter_85" + original_name = "batch_norm2d_120.b_0" shape = [192] dtype = "float32" min_val = float("-0.188276") @@ -946,6 +1032,7 @@ class Program_weight_tensor_parameter_85: class Program_weight_tensor_parameter_86: name = "parameter_86" + original_name = "batch_norm2d_120.w_0" shape = [192] dtype = "float32" min_val = float("0.923547") @@ -957,6 +1044,7 @@ class Program_weight_tensor_parameter_86: class Program_weight_tensor_parameter_87: name = "parameter_87" + original_name = "batch_norm2d_120.w_2" shape = [192] dtype = "float32" min_val = float("0.00150378") @@ -968,6 +1056,7 @@ class Program_weight_tensor_parameter_87: class Program_weight_tensor_parameter_88: name = "parameter_88" + original_name = "batch_norm2d_120.w_1" shape = [192] dtype = "float32" min_val = float("-0.0484967") @@ -979,6 +1068,7 @@ class Program_weight_tensor_parameter_88: class Program_weight_tensor_parameter_89: name = "parameter_89" + original_name = "conv2d_124.w_0" shape = [192, 192, 1, 1] dtype = "float32" min_val = float("-0.043172") @@ -990,6 +1080,7 @@ class Program_weight_tensor_parameter_89: class Program_weight_tensor_parameter_90: name = "parameter_90" + original_name = "batch_norm2d_119.b_0" shape = [192] dtype = "float32" min_val = float("-0.188276") @@ -1001,6 +1092,7 @@ class Program_weight_tensor_parameter_90: class Program_weight_tensor_parameter_91: name = "parameter_91" + original_name = "batch_norm2d_119.w_0" shape = [192] dtype = "float32" min_val = float("0.969747") @@ -1012,6 +1104,7 @@ class Program_weight_tensor_parameter_91: class Program_weight_tensor_parameter_92: name = "parameter_92" + original_name = "batch_norm2d_119.w_2" shape = [192] dtype = "float32" min_val = float("0.00270009") @@ -1023,6 +1116,7 @@ class Program_weight_tensor_parameter_92: class Program_weight_tensor_parameter_93: name = "parameter_93" + original_name = "batch_norm2d_119.w_1" shape = [192] dtype = "float32" min_val = float("-0.104084") @@ -1034,6 +1128,7 @@ class Program_weight_tensor_parameter_93: class Program_weight_tensor_parameter_94: name = "parameter_94" + original_name = "conv2d_123.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.0430966") @@ -1045,6 +1140,7 @@ class Program_weight_tensor_parameter_94: class Program_weight_tensor_parameter_95: name = "parameter_95" + original_name = "batch_norm2d_118.b_0" shape = [192] dtype = "float32" min_val = float("-0.186552") @@ -1056,6 +1152,7 @@ class Program_weight_tensor_parameter_95: class Program_weight_tensor_parameter_96: name = "parameter_96" + original_name = "batch_norm2d_118.w_0" shape = [192] dtype = "float32" min_val = float("0.882116") @@ -1067,6 +1164,7 @@ class Program_weight_tensor_parameter_96: class Program_weight_tensor_parameter_97: name = "parameter_97" + original_name = "batch_norm2d_118.w_2" shape = [192] dtype = "float32" min_val = float("0.00590687") @@ -1078,6 +1176,7 @@ class Program_weight_tensor_parameter_97: class Program_weight_tensor_parameter_98: name = "parameter_98" + original_name = "batch_norm2d_118.w_1" shape = [192] dtype = "float32" min_val = float("-0.0728663") @@ -1089,6 +1188,7 @@ class Program_weight_tensor_parameter_98: class Program_weight_tensor_parameter_99: name = "parameter_99" + original_name = "conv2d_122.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.0453925") @@ -1100,6 +1200,7 @@ class Program_weight_tensor_parameter_99: class Program_weight_tensor_parameter_100: name = "parameter_100" + original_name = "batch_norm2d_117.b_0" shape = [192] dtype = "float32" min_val = float("-0.223283") @@ -1111,6 +1212,7 @@ class Program_weight_tensor_parameter_100: class Program_weight_tensor_parameter_101: name = "parameter_101" + original_name = "batch_norm2d_117.w_0" shape = [192] dtype = "float32" min_val = float("0.903255") @@ -1122,6 +1224,7 @@ class Program_weight_tensor_parameter_101: class Program_weight_tensor_parameter_102: name = "parameter_102" + original_name = "batch_norm2d_117.w_2" shape = [192] dtype = "float32" min_val = float("0.0019003") @@ -1133,6 +1236,7 @@ class Program_weight_tensor_parameter_102: class Program_weight_tensor_parameter_103: name = "parameter_103" + original_name = "batch_norm2d_117.w_1" shape = [192] dtype = "float32" min_val = float("-0.0416717") @@ -1144,6 +1248,7 @@ class Program_weight_tensor_parameter_103: class Program_weight_tensor_parameter_104: name = "parameter_104" + original_name = "conv2d_121.w_0" shape = [192, 192, 1, 1] dtype = "float32" min_val = float("-0.0389131") @@ -1155,6 +1260,7 @@ class Program_weight_tensor_parameter_104: class Program_weight_tensor_parameter_105: name = "parameter_105" + original_name = "batch_norm2d_116.b_0" shape = [192] dtype = "float32" min_val = float("-0.223283") @@ -1166,6 +1272,7 @@ class Program_weight_tensor_parameter_105: class Program_weight_tensor_parameter_106: name = "parameter_106" + original_name = "batch_norm2d_116.w_0" shape = [192] dtype = "float32" min_val = float("0.947687") @@ -1177,6 +1284,7 @@ class Program_weight_tensor_parameter_106: class Program_weight_tensor_parameter_107: name = "parameter_107" + original_name = "batch_norm2d_116.w_2" shape = [192] dtype = "float32" min_val = float("0.00463295") @@ -1188,6 +1296,7 @@ class Program_weight_tensor_parameter_107: class Program_weight_tensor_parameter_108: name = "parameter_108" + original_name = "batch_norm2d_116.w_1" shape = [192] dtype = "float32" min_val = float("-0.09989") @@ -1199,6 +1308,7 @@ class Program_weight_tensor_parameter_108: class Program_weight_tensor_parameter_109: name = "parameter_109" + original_name = "conv2d_120.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.0473608") @@ -1210,6 +1320,7 @@ class Program_weight_tensor_parameter_109: class Program_weight_tensor_parameter_110: name = "parameter_110" + original_name = "batch_norm2d_115.b_0" shape = [192] dtype = "float32" min_val = float("-0.228002") @@ -1221,6 +1332,7 @@ class Program_weight_tensor_parameter_110: class Program_weight_tensor_parameter_111: name = "parameter_111" + original_name = "batch_norm2d_115.w_0" shape = [192] dtype = "float32" min_val = float("0.888077") @@ -1232,6 +1344,7 @@ class Program_weight_tensor_parameter_111: class Program_weight_tensor_parameter_112: name = "parameter_112" + original_name = "batch_norm2d_115.w_2" shape = [192] dtype = "float32" min_val = float("0.00606638") @@ -1243,6 +1356,7 @@ class Program_weight_tensor_parameter_112: class Program_weight_tensor_parameter_113: name = "parameter_113" + original_name = "batch_norm2d_115.w_1" shape = [192] dtype = "float32" min_val = float("-0.145865") @@ -1254,6 +1368,7 @@ class Program_weight_tensor_parameter_113: class Program_weight_tensor_parameter_114: name = "parameter_114" + original_name = "conv2d_119.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.0418221") @@ -1265,6 +1380,7 @@ class Program_weight_tensor_parameter_114: class Program_weight_tensor_parameter_115: name = "parameter_115" + original_name = "batch_norm2d_114.b_0" shape = [192] dtype = "float32" min_val = float("-0.192804") @@ -1276,6 +1392,7 @@ class Program_weight_tensor_parameter_115: class Program_weight_tensor_parameter_116: name = "parameter_116" + original_name = "batch_norm2d_114.w_0" shape = [192] dtype = "float32" min_val = float("0.925751") @@ -1287,6 +1404,7 @@ class Program_weight_tensor_parameter_116: class Program_weight_tensor_parameter_117: name = "parameter_117" + original_name = "batch_norm2d_114.w_2" shape = [192] dtype = "float32" min_val = float("0.00312917") @@ -1298,6 +1416,7 @@ class Program_weight_tensor_parameter_117: class Program_weight_tensor_parameter_118: name = "parameter_118" + original_name = "batch_norm2d_114.w_1" shape = [192] dtype = "float32" min_val = float("-0.0749669") @@ -1309,6 +1428,7 @@ class Program_weight_tensor_parameter_118: class Program_weight_tensor_parameter_119: name = "parameter_119" + original_name = "conv2d_118.w_0" shape = [192, 576, 1, 1] dtype = "float32" min_val = float("-0.0569455") @@ -1320,6 +1440,7 @@ class Program_weight_tensor_parameter_119: class Program_weight_tensor_parameter_120: name = "parameter_120" + original_name = "batch_norm2d_113.b_0" shape = [192] dtype = "float32" min_val = float("-0.0978959") @@ -1331,6 +1452,7 @@ class Program_weight_tensor_parameter_120: class Program_weight_tensor_parameter_121: name = "parameter_121" + original_name = "batch_norm2d_113.w_0" shape = [192] dtype = "float32" min_val = float("0.925028") @@ -1342,6 +1464,7 @@ class Program_weight_tensor_parameter_121: class Program_weight_tensor_parameter_122: name = "parameter_122" + original_name = "batch_norm2d_113.w_2" shape = [192] dtype = "float32" min_val = float("0.00234222") @@ -1353,6 +1476,7 @@ class Program_weight_tensor_parameter_122: class Program_weight_tensor_parameter_123: name = "parameter_123" + original_name = "batch_norm2d_113.w_1" shape = [192] dtype = "float32" min_val = float("-0.0571601") @@ -1364,6 +1488,7 @@ class Program_weight_tensor_parameter_123: class Program_weight_tensor_parameter_124: name = "parameter_124" + original_name = "conv2d_117.w_0" shape = [192, 576, 1, 1] dtype = "float32" min_val = float("-0.0968447") @@ -1375,6 +1500,7 @@ class Program_weight_tensor_parameter_124: class Program_weight_tensor_parameter_125: name = "parameter_125" + original_name = "batch_norm2d_112.b_0" shape = [192] dtype = "float32" min_val = float("-0.155666") @@ -1386,6 +1512,7 @@ class Program_weight_tensor_parameter_125: class Program_weight_tensor_parameter_126: name = "parameter_126" + original_name = "batch_norm2d_112.w_0" shape = [192] dtype = "float32" min_val = float("0.921162") @@ -1397,6 +1524,7 @@ class Program_weight_tensor_parameter_126: class Program_weight_tensor_parameter_127: name = "parameter_127" + original_name = "batch_norm2d_112.w_2" shape = [192] dtype = "float32" min_val = float("0.00411353") @@ -1408,6 +1536,7 @@ class Program_weight_tensor_parameter_127: class Program_weight_tensor_parameter_128: name = "parameter_128" + original_name = "batch_norm2d_112.w_1" shape = [192] dtype = "float32" min_val = float("-0.431603") @@ -1419,6 +1548,7 @@ class Program_weight_tensor_parameter_128: class Program_weight_tensor_parameter_129: name = "parameter_129" + original_name = "conv2d_116.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.0481316") @@ -1430,6 +1560,7 @@ class Program_weight_tensor_parameter_129: class Program_weight_tensor_parameter_130: name = "parameter_130" + original_name = "batch_norm2d_111.b_0" shape = [192] dtype = "float32" min_val = float("-0.548219") @@ -1441,6 +1572,7 @@ class Program_weight_tensor_parameter_130: class Program_weight_tensor_parameter_131: name = "parameter_131" + original_name = "batch_norm2d_111.w_0" shape = [192] dtype = "float32" min_val = float("0.546543") @@ -1452,6 +1584,7 @@ class Program_weight_tensor_parameter_131: class Program_weight_tensor_parameter_132: name = "parameter_132" + original_name = "batch_norm2d_111.w_2" shape = [192] dtype = "float32" min_val = float("0.00264418") @@ -1463,6 +1596,7 @@ class Program_weight_tensor_parameter_132: class Program_weight_tensor_parameter_133: name = "parameter_133" + original_name = "batch_norm2d_111.w_1" shape = [192] dtype = "float32" min_val = float("-0.205861") @@ -1474,6 +1608,7 @@ class Program_weight_tensor_parameter_133: class Program_weight_tensor_parameter_134: name = "parameter_134" + original_name = "conv2d_115.w_0" shape = [192, 192, 1, 1] dtype = "float32" min_val = float("-0.122667") @@ -1485,6 +1620,7 @@ class Program_weight_tensor_parameter_134: class Program_weight_tensor_parameter_135: name = "parameter_135" + original_name = "batch_norm2d_110.b_0" shape = [96] dtype = "float32" min_val = float("-0.454892") @@ -1496,6 +1632,7 @@ class Program_weight_tensor_parameter_135: class Program_weight_tensor_parameter_136: name = "parameter_136" + original_name = "batch_norm2d_110.w_0" shape = [96] dtype = "float32" min_val = float("0.766222") @@ -1507,6 +1644,7 @@ class Program_weight_tensor_parameter_136: class Program_weight_tensor_parameter_137: name = "parameter_137" + original_name = "batch_norm2d_110.w_2" shape = [96] dtype = "float32" min_val = float("0.00171231") @@ -1518,6 +1656,7 @@ class Program_weight_tensor_parameter_137: class Program_weight_tensor_parameter_138: name = "parameter_138" + original_name = "batch_norm2d_110.w_1" shape = [96] dtype = "float32" min_val = float("-0.0394539") @@ -1529,6 +1668,7 @@ class Program_weight_tensor_parameter_138: class Program_weight_tensor_parameter_139: name = "parameter_139" + original_name = "conv2d_114.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.0689192") @@ -1540,6 +1680,7 @@ class Program_weight_tensor_parameter_139: class Program_weight_tensor_parameter_140: name = "parameter_140" + original_name = "batch_norm2d_109.b_0" shape = [96] dtype = "float32" min_val = float("-0.454892") @@ -1551,6 +1692,7 @@ class Program_weight_tensor_parameter_140: class Program_weight_tensor_parameter_141: name = "parameter_141" + original_name = "batch_norm2d_109.w_0" shape = [96] dtype = "float32" min_val = float("0.510462") @@ -1562,6 +1704,7 @@ class Program_weight_tensor_parameter_141: class Program_weight_tensor_parameter_142: name = "parameter_142" + original_name = "batch_norm2d_109.w_2" shape = [96] dtype = "float32" min_val = float("0.00451032") @@ -1573,6 +1716,7 @@ class Program_weight_tensor_parameter_142: class Program_weight_tensor_parameter_143: name = "parameter_143" + original_name = "batch_norm2d_109.w_1" shape = [96] dtype = "float32" min_val = float("-0.300223") @@ -1584,6 +1728,7 @@ class Program_weight_tensor_parameter_143: class Program_weight_tensor_parameter_144: name = "parameter_144" + original_name = "conv2d_113.w_0" shape = [96, 96, 3, 3] dtype = "float32" min_val = float("-0.0746669") @@ -1595,6 +1740,7 @@ class Program_weight_tensor_parameter_144: class Program_weight_tensor_parameter_145: name = "parameter_145" + original_name = "batch_norm2d_108.b_0" shape = [96] dtype = "float32" min_val = float("-0.698004") @@ -1606,6 +1752,7 @@ class Program_weight_tensor_parameter_145: class Program_weight_tensor_parameter_146: name = "parameter_146" + original_name = "batch_norm2d_108.w_0" shape = [96] dtype = "float32" min_val = float("0.728018") @@ -1617,6 +1764,7 @@ class Program_weight_tensor_parameter_146: class Program_weight_tensor_parameter_147: name = "parameter_147" + original_name = "batch_norm2d_108.w_2" shape = [96] dtype = "float32" min_val = float("0.00625557") @@ -1628,6 +1776,7 @@ class Program_weight_tensor_parameter_147: class Program_weight_tensor_parameter_148: name = "parameter_148" + original_name = "batch_norm2d_108.w_1" shape = [96] dtype = "float32" min_val = float("-0.16835") @@ -1639,6 +1788,7 @@ class Program_weight_tensor_parameter_148: class Program_weight_tensor_parameter_149: name = "parameter_149" + original_name = "conv2d_112.w_0" shape = [96, 96, 3, 3] dtype = "float32" min_val = float("-0.0901918") @@ -1650,6 +1800,7 @@ class Program_weight_tensor_parameter_149: class Program_weight_tensor_parameter_150: name = "parameter_150" + original_name = "batch_norm2d_107.b_0" shape = [96] dtype = "float32" min_val = float("-0.356561") @@ -1661,6 +1812,7 @@ class Program_weight_tensor_parameter_150: class Program_weight_tensor_parameter_151: name = "parameter_151" + original_name = "batch_norm2d_107.w_0" shape = [96] dtype = "float32" min_val = float("0.63226") @@ -1672,6 +1824,7 @@ class Program_weight_tensor_parameter_151: class Program_weight_tensor_parameter_152: name = "parameter_152" + original_name = "batch_norm2d_107.w_2" shape = [96] dtype = "float32" min_val = float("0.00246361") @@ -1683,6 +1836,7 @@ class Program_weight_tensor_parameter_152: class Program_weight_tensor_parameter_153: name = "parameter_153" + original_name = "batch_norm2d_107.w_1" shape = [96] dtype = "float32" min_val = float("-0.0391778") @@ -1694,6 +1848,7 @@ class Program_weight_tensor_parameter_153: class Program_weight_tensor_parameter_154: name = "parameter_154" + original_name = "conv2d_111.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.0529296") @@ -1705,6 +1860,7 @@ class Program_weight_tensor_parameter_154: class Program_weight_tensor_parameter_155: name = "parameter_155" + original_name = "batch_norm2d_106.b_0" shape = [96] dtype = "float32" min_val = float("-0.356561") @@ -1716,6 +1872,7 @@ class Program_weight_tensor_parameter_155: class Program_weight_tensor_parameter_156: name = "parameter_156" + original_name = "batch_norm2d_106.w_0" shape = [96] dtype = "float32" min_val = float("0.816664") @@ -1727,6 +1884,7 @@ class Program_weight_tensor_parameter_156: class Program_weight_tensor_parameter_157: name = "parameter_157" + original_name = "batch_norm2d_106.w_2" shape = [96] dtype = "float32" min_val = float("0.00554989") @@ -1738,6 +1896,7 @@ class Program_weight_tensor_parameter_157: class Program_weight_tensor_parameter_158: name = "parameter_158" + original_name = "batch_norm2d_106.w_1" shape = [96] dtype = "float32" min_val = float("-0.0892734") @@ -1749,6 +1908,7 @@ class Program_weight_tensor_parameter_158: class Program_weight_tensor_parameter_159: name = "parameter_159" + original_name = "conv2d_110.w_0" shape = [96, 96, 3, 3] dtype = "float32" min_val = float("-0.0686605") @@ -1760,6 +1920,7 @@ class Program_weight_tensor_parameter_159: class Program_weight_tensor_parameter_160: name = "parameter_160" + original_name = "batch_norm2d_105.b_0" shape = [96] dtype = "float32" min_val = float("-0.479114") @@ -1771,6 +1932,7 @@ class Program_weight_tensor_parameter_160: class Program_weight_tensor_parameter_161: name = "parameter_161" + original_name = "batch_norm2d_105.w_0" shape = [96] dtype = "float32" min_val = float("0.779583") @@ -1782,6 +1944,7 @@ class Program_weight_tensor_parameter_161: class Program_weight_tensor_parameter_162: name = "parameter_162" + original_name = "batch_norm2d_105.w_2" shape = [96] dtype = "float32" min_val = float("0.00461085") @@ -1793,6 +1956,7 @@ class Program_weight_tensor_parameter_162: class Program_weight_tensor_parameter_163: name = "parameter_163" + original_name = "batch_norm2d_105.w_1" shape = [96] dtype = "float32" min_val = float("-0.130604") @@ -1804,6 +1968,7 @@ class Program_weight_tensor_parameter_163: class Program_weight_tensor_parameter_164: name = "parameter_164" + original_name = "conv2d_109.w_0" shape = [96, 96, 3, 3] dtype = "float32" min_val = float("-0.0836645") @@ -1815,6 +1980,7 @@ class Program_weight_tensor_parameter_164: class Program_weight_tensor_parameter_165: name = "parameter_165" + original_name = "batch_norm2d_104.b_0" shape = [96] dtype = "float32" min_val = float("-0.481327") @@ -1826,6 +1992,7 @@ class Program_weight_tensor_parameter_165: class Program_weight_tensor_parameter_166: name = "parameter_166" + original_name = "batch_norm2d_104.w_0" shape = [96] dtype = "float32" min_val = float("0.72772") @@ -1837,6 +2004,7 @@ class Program_weight_tensor_parameter_166: class Program_weight_tensor_parameter_167: name = "parameter_167" + original_name = "batch_norm2d_104.w_2" shape = [96] dtype = "float32" min_val = float("0.00471793") @@ -1848,6 +2016,7 @@ class Program_weight_tensor_parameter_167: class Program_weight_tensor_parameter_168: name = "parameter_168" + original_name = "batch_norm2d_104.w_1" shape = [96] dtype = "float32" min_val = float("-0.0437622") @@ -1859,6 +2028,7 @@ class Program_weight_tensor_parameter_168: class Program_weight_tensor_parameter_169: name = "parameter_169" + original_name = "conv2d_108.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.0828667") @@ -1870,6 +2040,7 @@ class Program_weight_tensor_parameter_169: class Program_weight_tensor_parameter_170: name = "parameter_170" + original_name = "batch_norm2d_103.b_0" shape = [96] dtype = "float32" min_val = float("-0.481327") @@ -1881,6 +2052,7 @@ class Program_weight_tensor_parameter_170: class Program_weight_tensor_parameter_171: name = "parameter_171" + original_name = "batch_norm2d_103.w_0" shape = [96] dtype = "float32" min_val = float("0.769747") @@ -1892,6 +2064,7 @@ class Program_weight_tensor_parameter_171: class Program_weight_tensor_parameter_172: name = "parameter_172" + original_name = "batch_norm2d_103.w_2" shape = [96] dtype = "float32" min_val = float("0.00991252") @@ -1903,6 +2076,7 @@ class Program_weight_tensor_parameter_172: class Program_weight_tensor_parameter_173: name = "parameter_173" + original_name = "batch_norm2d_103.w_1" shape = [96] dtype = "float32" min_val = float("-0.116497") @@ -1914,6 +2088,7 @@ class Program_weight_tensor_parameter_173: class Program_weight_tensor_parameter_174: name = "parameter_174" + original_name = "conv2d_107.w_0" shape = [96, 96, 3, 3] dtype = "float32" min_val = float("-0.098636") @@ -1925,6 +2100,7 @@ class Program_weight_tensor_parameter_174: class Program_weight_tensor_parameter_175: name = "parameter_175" + original_name = "batch_norm2d_102.b_0" shape = [96] dtype = "float32" min_val = float("-0.554545") @@ -1936,6 +2112,7 @@ class Program_weight_tensor_parameter_175: class Program_weight_tensor_parameter_176: name = "parameter_176" + original_name = "batch_norm2d_102.w_0" shape = [96] dtype = "float32" min_val = float("0.755258") @@ -1947,6 +2124,7 @@ class Program_weight_tensor_parameter_176: class Program_weight_tensor_parameter_177: name = "parameter_177" + original_name = "batch_norm2d_102.w_2" shape = [96] dtype = "float32" min_val = float("0.00741183") @@ -1958,6 +2136,7 @@ class Program_weight_tensor_parameter_177: class Program_weight_tensor_parameter_178: name = "parameter_178" + original_name = "batch_norm2d_102.w_1" shape = [96] dtype = "float32" min_val = float("-0.113194") @@ -1969,6 +2148,7 @@ class Program_weight_tensor_parameter_178: class Program_weight_tensor_parameter_179: name = "parameter_179" + original_name = "conv2d_106.w_0" shape = [96, 96, 3, 3] dtype = "float32" min_val = float("-0.117137") @@ -1980,6 +2160,7 @@ class Program_weight_tensor_parameter_179: class Program_weight_tensor_parameter_180: name = "parameter_180" + original_name = "batch_norm2d_101.b_0" shape = [96] dtype = "float32" min_val = float("-0.610344") @@ -1991,6 +2172,7 @@ class Program_weight_tensor_parameter_180: class Program_weight_tensor_parameter_181: name = "parameter_181" + original_name = "batch_norm2d_101.w_0" shape = [96] dtype = "float32" min_val = float("0.658569") @@ -2002,6 +2184,7 @@ class Program_weight_tensor_parameter_181: class Program_weight_tensor_parameter_182: name = "parameter_182" + original_name = "batch_norm2d_101.w_2" shape = [96] dtype = "float32" min_val = float("0.00753543") @@ -2013,6 +2196,7 @@ class Program_weight_tensor_parameter_182: class Program_weight_tensor_parameter_183: name = "parameter_183" + original_name = "batch_norm2d_101.w_1" shape = [96] dtype = "float32" min_val = float("-0.0903701") @@ -2024,6 +2208,7 @@ class Program_weight_tensor_parameter_183: class Program_weight_tensor_parameter_184: name = "parameter_184" + original_name = "conv2d_105.w_0" shape = [96, 448, 1, 1] dtype = "float32" min_val = float("-0.130957") @@ -2035,6 +2220,7 @@ class Program_weight_tensor_parameter_184: class Program_weight_tensor_parameter_185: name = "parameter_185" + original_name = "batch_norm2d_100.b_0" shape = [96] dtype = "float32" min_val = float("-0.0946454") @@ -2046,6 +2232,7 @@ class Program_weight_tensor_parameter_185: class Program_weight_tensor_parameter_186: name = "parameter_186" + original_name = "batch_norm2d_100.w_0" shape = [96] dtype = "float32" min_val = float("0.711917") @@ -2057,6 +2244,7 @@ class Program_weight_tensor_parameter_186: class Program_weight_tensor_parameter_187: name = "parameter_187" + original_name = "batch_norm2d_100.w_2" shape = [96] dtype = "float32" min_val = float("0.00157056") @@ -2068,6 +2256,7 @@ class Program_weight_tensor_parameter_187: class Program_weight_tensor_parameter_188: name = "parameter_188" + original_name = "batch_norm2d_100.w_1" shape = [96] dtype = "float32" min_val = float("-0.0600397") @@ -2079,6 +2268,7 @@ class Program_weight_tensor_parameter_188: class Program_weight_tensor_parameter_189: name = "parameter_189" + original_name = "conv2d_104.w_0" shape = [96, 448, 1, 1] dtype = "float32" min_val = float("-0.0876125") @@ -2090,6 +2280,7 @@ class Program_weight_tensor_parameter_189: class Program_weight_tensor_parameter_190: name = "parameter_190" + original_name = "batch_norm2d_99.b_0" shape = [192] dtype = "float32" min_val = float("-0.28994") @@ -2101,6 +2292,7 @@ class Program_weight_tensor_parameter_190: class Program_weight_tensor_parameter_191: name = "parameter_191" + original_name = "batch_norm2d_99.w_0" shape = [192] dtype = "float32" min_val = float("0.675984") @@ -2112,6 +2304,7 @@ class Program_weight_tensor_parameter_191: class Program_weight_tensor_parameter_192: name = "parameter_192" + original_name = "batch_norm2d_99.w_2" shape = [192] dtype = "float32" min_val = float("0.00621495") @@ -2123,6 +2316,7 @@ class Program_weight_tensor_parameter_192: class Program_weight_tensor_parameter_193: name = "parameter_193" + original_name = "batch_norm2d_99.w_1" shape = [192] dtype = "float32" min_val = float("-0.146238") @@ -2134,6 +2328,7 @@ class Program_weight_tensor_parameter_193: class Program_weight_tensor_parameter_194: name = "parameter_194" + original_name = "conv2d_103.w_0" shape = [192, 384, 1, 1] dtype = "float32" min_val = float("-0.103707") @@ -2145,6 +2340,7 @@ class Program_weight_tensor_parameter_194: class Program_weight_tensor_parameter_195: name = "parameter_195" + original_name = "batch_norm2d_98.b_0" shape = [384] dtype = "float32" min_val = float("-0.196718") @@ -2156,6 +2352,7 @@ class Program_weight_tensor_parameter_195: class Program_weight_tensor_parameter_196: name = "parameter_196" + original_name = "batch_norm2d_98.w_0" shape = [384] dtype = "float32" min_val = float("0.871884") @@ -2167,6 +2364,7 @@ class Program_weight_tensor_parameter_196: class Program_weight_tensor_parameter_197: name = "parameter_197" + original_name = "batch_norm2d_98.w_2" shape = [384] dtype = "float32" min_val = float("0.00568249") @@ -2178,6 +2376,7 @@ class Program_weight_tensor_parameter_197: class Program_weight_tensor_parameter_198: name = "parameter_198" + original_name = "batch_norm2d_98.w_1" shape = [384] dtype = "float32" min_val = float("-0.243456") @@ -2189,6 +2388,7 @@ class Program_weight_tensor_parameter_198: class Program_weight_tensor_parameter_199: name = "parameter_199" + original_name = "conv2d_102.w_0" shape = [384, 384, 1, 1] dtype = "float32" min_val = float("-0.155294") @@ -2200,6 +2400,7 @@ class Program_weight_tensor_parameter_199: class Program_weight_tensor_parameter_200: name = "parameter_200" + original_name = "batch_norm2d_97.b_0" shape = [192] dtype = "float32" min_val = float("-0.172574") @@ -2211,6 +2412,7 @@ class Program_weight_tensor_parameter_200: class Program_weight_tensor_parameter_201: name = "parameter_201" + original_name = "batch_norm2d_97.w_0" shape = [192] dtype = "float32" min_val = float("0.886587") @@ -2222,6 +2424,7 @@ class Program_weight_tensor_parameter_201: class Program_weight_tensor_parameter_202: name = "parameter_202" + original_name = "batch_norm2d_97.w_2" shape = [192] dtype = "float32" min_val = float("0.00369063") @@ -2233,6 +2436,7 @@ class Program_weight_tensor_parameter_202: class Program_weight_tensor_parameter_203: name = "parameter_203" + original_name = "batch_norm2d_97.w_1" shape = [192] dtype = "float32" min_val = float("-0.0644225") @@ -2244,6 +2448,7 @@ class Program_weight_tensor_parameter_203: class Program_weight_tensor_parameter_204: name = "parameter_204" + original_name = "conv2d_101.w_0" shape = [192, 192, 1, 1] dtype = "float32" min_val = float("-0.0502286") @@ -2255,6 +2460,7 @@ class Program_weight_tensor_parameter_204: class Program_weight_tensor_parameter_205: name = "parameter_205" + original_name = "batch_norm2d_96.b_0" shape = [192] dtype = "float32" min_val = float("-0.172574") @@ -2266,6 +2472,7 @@ class Program_weight_tensor_parameter_205: class Program_weight_tensor_parameter_206: name = "parameter_206" + original_name = "batch_norm2d_96.w_0" shape = [192] dtype = "float32" min_val = float("0.945305") @@ -2277,6 +2484,7 @@ class Program_weight_tensor_parameter_206: class Program_weight_tensor_parameter_207: name = "parameter_207" + original_name = "batch_norm2d_96.w_2" shape = [192] dtype = "float32" min_val = float("0.0098824") @@ -2288,6 +2496,7 @@ class Program_weight_tensor_parameter_207: class Program_weight_tensor_parameter_208: name = "parameter_208" + original_name = "batch_norm2d_96.w_1" shape = [192] dtype = "float32" min_val = float("-0.137861") @@ -2299,6 +2508,7 @@ class Program_weight_tensor_parameter_208: class Program_weight_tensor_parameter_209: name = "parameter_209" + original_name = "conv2d_100.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.0481666") @@ -2310,6 +2520,7 @@ class Program_weight_tensor_parameter_209: class Program_weight_tensor_parameter_210: name = "parameter_210" + original_name = "batch_norm2d_95.b_0" shape = [192] dtype = "float32" min_val = float("-0.210876") @@ -2321,6 +2532,7 @@ class Program_weight_tensor_parameter_210: class Program_weight_tensor_parameter_211: name = "parameter_211" + original_name = "batch_norm2d_95.w_0" shape = [192] dtype = "float32" min_val = float("0.939491") @@ -2332,6 +2544,7 @@ class Program_weight_tensor_parameter_211: class Program_weight_tensor_parameter_212: name = "parameter_212" + original_name = "batch_norm2d_95.w_2" shape = [192] dtype = "float32" min_val = float("0.0216612") @@ -2343,6 +2556,7 @@ class Program_weight_tensor_parameter_212: class Program_weight_tensor_parameter_213: name = "parameter_213" + original_name = "batch_norm2d_95.w_1" shape = [192] dtype = "float32" min_val = float("-0.165667") @@ -2354,6 +2568,7 @@ class Program_weight_tensor_parameter_213: class Program_weight_tensor_parameter_214: name = "parameter_214" + original_name = "conv2d_99.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.0665545") @@ -2365,6 +2580,7 @@ class Program_weight_tensor_parameter_214: class Program_weight_tensor_parameter_215: name = "parameter_215" + original_name = "batch_norm2d_94.b_0" shape = [192] dtype = "float32" min_val = float("-0.190391") @@ -2376,6 +2592,7 @@ class Program_weight_tensor_parameter_215: class Program_weight_tensor_parameter_216: name = "parameter_216" + original_name = "batch_norm2d_94.w_0" shape = [192] dtype = "float32" min_val = float("0.944709") @@ -2387,6 +2604,7 @@ class Program_weight_tensor_parameter_216: class Program_weight_tensor_parameter_217: name = "parameter_217" + original_name = "batch_norm2d_94.w_2" shape = [192] dtype = "float32" min_val = float("0.00209797") @@ -2398,6 +2616,7 @@ class Program_weight_tensor_parameter_217: class Program_weight_tensor_parameter_218: name = "parameter_218" + original_name = "batch_norm2d_94.w_1" shape = [192] dtype = "float32" min_val = float("-0.0731497") @@ -2409,6 +2628,7 @@ class Program_weight_tensor_parameter_218: class Program_weight_tensor_parameter_219: name = "parameter_219" + original_name = "conv2d_98.w_0" shape = [192, 192, 1, 1] dtype = "float32" min_val = float("-0.0318287") @@ -2420,6 +2640,7 @@ class Program_weight_tensor_parameter_219: class Program_weight_tensor_parameter_220: name = "parameter_220" + original_name = "batch_norm2d_93.b_0" shape = [192] dtype = "float32" min_val = float("-0.190391") @@ -2431,6 +2652,7 @@ class Program_weight_tensor_parameter_220: class Program_weight_tensor_parameter_221: name = "parameter_221" + original_name = "batch_norm2d_93.w_0" shape = [192] dtype = "float32" min_val = float("0.954247") @@ -2442,6 +2664,7 @@ class Program_weight_tensor_parameter_221: class Program_weight_tensor_parameter_222: name = "parameter_222" + original_name = "batch_norm2d_93.w_2" shape = [192] dtype = "float32" min_val = float("0.00638622") @@ -2453,6 +2676,7 @@ class Program_weight_tensor_parameter_222: class Program_weight_tensor_parameter_223: name = "parameter_223" + original_name = "batch_norm2d_93.w_1" shape = [192] dtype = "float32" min_val = float("-0.13966") @@ -2464,6 +2688,7 @@ class Program_weight_tensor_parameter_223: class Program_weight_tensor_parameter_224: name = "parameter_224" + original_name = "conv2d_97.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.0503067") @@ -2475,6 +2700,7 @@ class Program_weight_tensor_parameter_224: class Program_weight_tensor_parameter_225: name = "parameter_225" + original_name = "batch_norm2d_92.b_0" shape = [192] dtype = "float32" min_val = float("-0.224111") @@ -2486,6 +2712,7 @@ class Program_weight_tensor_parameter_225: class Program_weight_tensor_parameter_226: name = "parameter_226" + original_name = "batch_norm2d_92.w_0" shape = [192] dtype = "float32" min_val = float("0.947988") @@ -2497,6 +2724,7 @@ class Program_weight_tensor_parameter_226: class Program_weight_tensor_parameter_227: name = "parameter_227" + original_name = "batch_norm2d_92.w_2" shape = [192] dtype = "float32" min_val = float("0.0202697") @@ -2508,6 +2736,7 @@ class Program_weight_tensor_parameter_227: class Program_weight_tensor_parameter_228: name = "parameter_228" + original_name = "batch_norm2d_92.w_1" shape = [192] dtype = "float32" min_val = float("-0.221703") @@ -2519,6 +2748,7 @@ class Program_weight_tensor_parameter_228: class Program_weight_tensor_parameter_229: name = "parameter_229" + original_name = "conv2d_96.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.0627662") @@ -2530,6 +2760,7 @@ class Program_weight_tensor_parameter_229: class Program_weight_tensor_parameter_230: name = "parameter_230" + original_name = "batch_norm2d_91.b_0" shape = [192] dtype = "float32" min_val = float("-0.149871") @@ -2541,6 +2772,7 @@ class Program_weight_tensor_parameter_230: class Program_weight_tensor_parameter_231: name = "parameter_231" + original_name = "batch_norm2d_91.w_0" shape = [192] dtype = "float32" min_val = float("0.933681") @@ -2552,6 +2784,7 @@ class Program_weight_tensor_parameter_231: class Program_weight_tensor_parameter_232: name = "parameter_232" + original_name = "batch_norm2d_91.w_2" shape = [192] dtype = "float32" min_val = float("0.00185054") @@ -2563,6 +2796,7 @@ class Program_weight_tensor_parameter_232: class Program_weight_tensor_parameter_233: name = "parameter_233" + original_name = "batch_norm2d_91.w_1" shape = [192] dtype = "float32" min_val = float("-0.0635071") @@ -2574,6 +2808,7 @@ class Program_weight_tensor_parameter_233: class Program_weight_tensor_parameter_234: name = "parameter_234" + original_name = "conv2d_95.w_0" shape = [192, 192, 1, 1] dtype = "float32" min_val = float("-0.0294449") @@ -2585,6 +2820,7 @@ class Program_weight_tensor_parameter_234: class Program_weight_tensor_parameter_235: name = "parameter_235" + original_name = "batch_norm2d_90.b_0" shape = [192] dtype = "float32" min_val = float("-0.149871") @@ -2596,6 +2832,7 @@ class Program_weight_tensor_parameter_235: class Program_weight_tensor_parameter_236: name = "parameter_236" + original_name = "batch_norm2d_90.w_0" shape = [192] dtype = "float32" min_val = float("0.937556") @@ -2607,6 +2844,7 @@ class Program_weight_tensor_parameter_236: class Program_weight_tensor_parameter_237: name = "parameter_237" + original_name = "batch_norm2d_90.w_2" shape = [192] dtype = "float32" min_val = float("0.0065578") @@ -2618,6 +2856,7 @@ class Program_weight_tensor_parameter_237: class Program_weight_tensor_parameter_238: name = "parameter_238" + original_name = "batch_norm2d_90.w_1" shape = [192] dtype = "float32" min_val = float("-0.17874") @@ -2629,6 +2868,7 @@ class Program_weight_tensor_parameter_238: class Program_weight_tensor_parameter_239: name = "parameter_239" + original_name = "conv2d_94.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.0381431") @@ -2640,6 +2880,7 @@ class Program_weight_tensor_parameter_239: class Program_weight_tensor_parameter_240: name = "parameter_240" + original_name = "batch_norm2d_89.b_0" shape = [192] dtype = "float32" min_val = float("-0.28186") @@ -2651,6 +2892,7 @@ class Program_weight_tensor_parameter_240: class Program_weight_tensor_parameter_241: name = "parameter_241" + original_name = "batch_norm2d_89.w_0" shape = [192] dtype = "float32" min_val = float("0.943815") @@ -2662,6 +2904,7 @@ class Program_weight_tensor_parameter_241: class Program_weight_tensor_parameter_242: name = "parameter_242" + original_name = "batch_norm2d_89.w_2" shape = [192] dtype = "float32" min_val = float("0.00906398") @@ -2673,6 +2916,7 @@ class Program_weight_tensor_parameter_242: class Program_weight_tensor_parameter_243: name = "parameter_243" + original_name = "batch_norm2d_89.w_1" shape = [192] dtype = "float32" min_val = float("-0.218708") @@ -2684,6 +2928,7 @@ class Program_weight_tensor_parameter_243: class Program_weight_tensor_parameter_244: name = "parameter_244" + original_name = "conv2d_93.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.053249") @@ -2695,6 +2940,7 @@ class Program_weight_tensor_parameter_244: class Program_weight_tensor_parameter_245: name = "parameter_245" + original_name = "batch_norm2d_88.b_0" shape = [192] dtype = "float32" min_val = float("-0.249578") @@ -2706,6 +2952,7 @@ class Program_weight_tensor_parameter_245: class Program_weight_tensor_parameter_246: name = "parameter_246" + original_name = "batch_norm2d_88.w_0" shape = [192] dtype = "float32" min_val = float("0.914307") @@ -2717,6 +2964,7 @@ class Program_weight_tensor_parameter_246: class Program_weight_tensor_parameter_247: name = "parameter_247" + original_name = "batch_norm2d_88.w_2" shape = [192] dtype = "float32" min_val = float("0.00415368") @@ -2728,6 +2976,7 @@ class Program_weight_tensor_parameter_247: class Program_weight_tensor_parameter_248: name = "parameter_248" + original_name = "batch_norm2d_88.w_1" shape = [192] dtype = "float32" min_val = float("-0.11797") @@ -2739,6 +2988,7 @@ class Program_weight_tensor_parameter_248: class Program_weight_tensor_parameter_249: name = "parameter_249" + original_name = "conv2d_92.w_0" shape = [192, 896, 1, 1] dtype = "float32" min_val = float("-0.0596061") @@ -2750,6 +3000,7 @@ class Program_weight_tensor_parameter_249: class Program_weight_tensor_parameter_250: name = "parameter_250" + original_name = "batch_norm2d_87.b_0" shape = [192] dtype = "float32" min_val = float("-0.174062") @@ -2761,6 +3012,7 @@ class Program_weight_tensor_parameter_250: class Program_weight_tensor_parameter_251: name = "parameter_251" + original_name = "batch_norm2d_87.w_0" shape = [192] dtype = "float32" min_val = float("0.949294") @@ -2772,6 +3024,7 @@ class Program_weight_tensor_parameter_251: class Program_weight_tensor_parameter_252: name = "parameter_252" + original_name = "batch_norm2d_87.w_2" shape = [192] dtype = "float32" min_val = float("0.00454854") @@ -2783,6 +3036,7 @@ class Program_weight_tensor_parameter_252: class Program_weight_tensor_parameter_253: name = "parameter_253" + original_name = "batch_norm2d_87.w_1" shape = [192] dtype = "float32" min_val = float("-0.0931261") @@ -2794,6 +3048,7 @@ class Program_weight_tensor_parameter_253: class Program_weight_tensor_parameter_254: name = "parameter_254" + original_name = "conv2d_91.w_0" shape = [192, 896, 1, 1] dtype = "float32" min_val = float("-0.0577965") @@ -2805,6 +3060,7 @@ class Program_weight_tensor_parameter_254: class Program_weight_tensor_parameter_255: name = "parameter_255" + original_name = "batch_norm2d_86.b_0" shape = [384] dtype = "float32" min_val = float("-0.243157") @@ -2816,6 +3072,7 @@ class Program_weight_tensor_parameter_255: class Program_weight_tensor_parameter_256: name = "parameter_256" + original_name = "batch_norm2d_86.w_0" shape = [384] dtype = "float32" min_val = float("0.819523") @@ -2827,6 +3084,7 @@ class Program_weight_tensor_parameter_256: class Program_weight_tensor_parameter_257: name = "parameter_257" + original_name = "batch_norm2d_86.w_2" shape = [384] dtype = "float32" min_val = float("0.00653635") @@ -2838,6 +3096,7 @@ class Program_weight_tensor_parameter_257: class Program_weight_tensor_parameter_258: name = "parameter_258" + original_name = "batch_norm2d_86.w_1" shape = [384] dtype = "float32" min_val = float("-0.0841299") @@ -2849,6 +3108,7 @@ class Program_weight_tensor_parameter_258: class Program_weight_tensor_parameter_259: name = "parameter_259" + original_name = "conv2d_90.w_0" shape = [384, 768, 1, 1] dtype = "float32" min_val = float("-0.036508") @@ -2860,6 +3120,7 @@ class Program_weight_tensor_parameter_259: class Program_weight_tensor_parameter_260: name = "parameter_260" + original_name = "batch_norm2d_85.b_0" shape = [768] dtype = "float32" min_val = float("-0.101546") @@ -2871,6 +3132,7 @@ class Program_weight_tensor_parameter_260: class Program_weight_tensor_parameter_261: name = "parameter_261" + original_name = "batch_norm2d_85.w_0" shape = [768] dtype = "float32" min_val = float("0.953949") @@ -2882,6 +3144,7 @@ class Program_weight_tensor_parameter_261: class Program_weight_tensor_parameter_262: name = "parameter_262" + original_name = "batch_norm2d_85.w_2" shape = [768] dtype = "float32" min_val = float("0.00467604") @@ -2893,6 +3156,7 @@ class Program_weight_tensor_parameter_262: class Program_weight_tensor_parameter_263: name = "parameter_263" + original_name = "batch_norm2d_85.w_1" shape = [768] dtype = "float32" min_val = float("-0.0978559") @@ -2904,6 +3168,7 @@ class Program_weight_tensor_parameter_263: class Program_weight_tensor_parameter_264: name = "parameter_264" + original_name = "conv2d_89.w_0" shape = [768, 768, 1, 1] dtype = "float32" min_val = float("-0.0460146") @@ -2915,6 +3180,7 @@ class Program_weight_tensor_parameter_264: class Program_weight_tensor_parameter_265: name = "parameter_265" + original_name = "batch_norm2d_84.b_0" shape = [384] dtype = "float32" min_val = float("-0.151408") @@ -2926,6 +3192,7 @@ class Program_weight_tensor_parameter_265: class Program_weight_tensor_parameter_266: name = "parameter_266" + original_name = "batch_norm2d_84.w_0" shape = [384] dtype = "float32" min_val = float("0.891922") @@ -2937,6 +3204,7 @@ class Program_weight_tensor_parameter_266: class Program_weight_tensor_parameter_267: name = "parameter_267" + original_name = "batch_norm2d_84.w_2" shape = [384] dtype = "float32" min_val = float("0.00286638") @@ -2948,6 +3216,7 @@ class Program_weight_tensor_parameter_267: class Program_weight_tensor_parameter_268: name = "parameter_268" + original_name = "batch_norm2d_84.w_1" shape = [384] dtype = "float32" min_val = float("-0.064554") @@ -2959,6 +3228,7 @@ class Program_weight_tensor_parameter_268: class Program_weight_tensor_parameter_269: name = "parameter_269" + original_name = "conv2d_88.w_0" shape = [384, 384, 1, 1] dtype = "float32" min_val = float("-0.0308451") @@ -2970,6 +3240,7 @@ class Program_weight_tensor_parameter_269: class Program_weight_tensor_parameter_270: name = "parameter_270" + original_name = "batch_norm2d_83.b_0" shape = [384] dtype = "float32" min_val = float("-0.151408") @@ -2981,6 +3252,7 @@ class Program_weight_tensor_parameter_270: class Program_weight_tensor_parameter_271: name = "parameter_271" + original_name = "batch_norm2d_83.w_0" shape = [384] dtype = "float32" min_val = float("0.886572") @@ -2992,6 +3264,7 @@ class Program_weight_tensor_parameter_271: class Program_weight_tensor_parameter_272: name = "parameter_272" + original_name = "batch_norm2d_83.w_2" shape = [384] dtype = "float32" min_val = float("0.0148773") @@ -3003,6 +3276,7 @@ class Program_weight_tensor_parameter_272: class Program_weight_tensor_parameter_273: name = "parameter_273" + original_name = "batch_norm2d_83.w_1" shape = [384] dtype = "float32" min_val = float("-0.214635") @@ -3014,6 +3288,7 @@ class Program_weight_tensor_parameter_273: class Program_weight_tensor_parameter_274: name = "parameter_274" + original_name = "conv2d_87.w_0" shape = [384, 384, 3, 3] dtype = "float32" min_val = float("-0.0333005") @@ -3025,6 +3300,7 @@ class Program_weight_tensor_parameter_274: class Program_weight_tensor_parameter_275: name = "parameter_275" + original_name = "batch_norm2d_82.b_0" shape = [384] dtype = "float32" min_val = float("-0.0780192") @@ -3036,6 +3312,7 @@ class Program_weight_tensor_parameter_275: class Program_weight_tensor_parameter_276: name = "parameter_276" + original_name = "batch_norm2d_82.w_0" shape = [384] dtype = "float32" min_val = float("0.91913") @@ -3047,6 +3324,7 @@ class Program_weight_tensor_parameter_276: class Program_weight_tensor_parameter_277: name = "parameter_277" + original_name = "batch_norm2d_82.w_2" shape = [384] dtype = "float32" min_val = float("0.0103669") @@ -3058,6 +3336,7 @@ class Program_weight_tensor_parameter_277: class Program_weight_tensor_parameter_278: name = "parameter_278" + original_name = "batch_norm2d_82.w_1" shape = [384] dtype = "float32" min_val = float("-0.152271") @@ -3069,6 +3348,7 @@ class Program_weight_tensor_parameter_278: class Program_weight_tensor_parameter_279: name = "parameter_279" + original_name = "conv2d_86.w_0" shape = [384, 384, 3, 3] dtype = "float32" min_val = float("-0.022605") @@ -3080,6 +3360,7 @@ class Program_weight_tensor_parameter_279: class Program_weight_tensor_parameter_280: name = "parameter_280" + original_name = "batch_norm2d_81.b_0" shape = [384] dtype = "float32" min_val = float("-0.0692737") @@ -3091,6 +3372,7 @@ class Program_weight_tensor_parameter_280: class Program_weight_tensor_parameter_281: name = "parameter_281" + original_name = "batch_norm2d_81.w_0" shape = [384] dtype = "float32" min_val = float("0.946121") @@ -3102,6 +3384,7 @@ class Program_weight_tensor_parameter_281: class Program_weight_tensor_parameter_282: name = "parameter_282" + original_name = "batch_norm2d_81.w_2" shape = [384] dtype = "float32" min_val = float("0.038754") @@ -3113,6 +3396,7 @@ class Program_weight_tensor_parameter_282: class Program_weight_tensor_parameter_283: name = "parameter_283" + original_name = "batch_norm2d_81.w_1" shape = [384] dtype = "float32" min_val = float("-1.36389") @@ -3124,6 +3408,7 @@ class Program_weight_tensor_parameter_283: class Program_weight_tensor_parameter_284: name = "parameter_284" + original_name = "conv2d_85.w_0" shape = [384, 1536, 1, 1] dtype = "float32" min_val = float("-0.0364844") @@ -3135,6 +3420,7 @@ class Program_weight_tensor_parameter_284: class Program_weight_tensor_parameter_285: name = "parameter_285" + original_name = "batch_norm2d_80.b_0" shape = [384] dtype = "float32" min_val = float("-0.0175083") @@ -3146,6 +3432,7 @@ class Program_weight_tensor_parameter_285: class Program_weight_tensor_parameter_286: name = "parameter_286" + original_name = "batch_norm2d_80.w_0" shape = [384] dtype = "float32" min_val = float("0.969747") @@ -3157,6 +3444,7 @@ class Program_weight_tensor_parameter_286: class Program_weight_tensor_parameter_287: name = "parameter_287" + original_name = "batch_norm2d_80.w_2" shape = [384] dtype = "float32" min_val = float("0.00151688") @@ -3168,6 +3456,7 @@ class Program_weight_tensor_parameter_287: class Program_weight_tensor_parameter_288: name = "parameter_288" + original_name = "batch_norm2d_80.w_1" shape = [384] dtype = "float32" min_val = float("-0.082494") @@ -3179,6 +3468,7 @@ class Program_weight_tensor_parameter_288: class Program_weight_tensor_parameter_289: name = "parameter_289" + original_name = "conv2d_84.w_0" shape = [384, 384, 1, 1] dtype = "float32" min_val = float("-0.0288604") @@ -3190,6 +3480,7 @@ class Program_weight_tensor_parameter_289: class Program_weight_tensor_parameter_290: name = "parameter_290" + original_name = "batch_norm2d_79.b_0" shape = [384] dtype = "float32" min_val = float("-0.0175083") @@ -3201,6 +3492,7 @@ class Program_weight_tensor_parameter_290: class Program_weight_tensor_parameter_291: name = "parameter_291" + original_name = "batch_norm2d_79.w_0" shape = [384] dtype = "float32" min_val = float("0.972131") @@ -3212,6 +3504,7 @@ class Program_weight_tensor_parameter_291: class Program_weight_tensor_parameter_292: name = "parameter_292" + original_name = "batch_norm2d_79.w_2" shape = [384] dtype = "float32" min_val = float("0.00691705") @@ -3223,6 +3516,7 @@ class Program_weight_tensor_parameter_292: class Program_weight_tensor_parameter_293: name = "parameter_293" + original_name = "batch_norm2d_79.w_1" shape = [384] dtype = "float32" min_val = float("-0.247766") @@ -3234,6 +3528,7 @@ class Program_weight_tensor_parameter_293: class Program_weight_tensor_parameter_294: name = "parameter_294" + original_name = "conv2d_83.w_0" shape = [384, 384, 3, 3] dtype = "float32" min_val = float("-0.0229517") @@ -3245,6 +3540,7 @@ class Program_weight_tensor_parameter_294: class Program_weight_tensor_parameter_295: name = "parameter_295" + original_name = "batch_norm2d_78.b_0" shape = [384] dtype = "float32" min_val = float("-0.0465124") @@ -3256,6 +3552,7 @@ class Program_weight_tensor_parameter_295: class Program_weight_tensor_parameter_296: name = "parameter_296" + original_name = "batch_norm2d_78.w_0" shape = [384] dtype = "float32" min_val = float("0.955738") @@ -3267,6 +3564,7 @@ class Program_weight_tensor_parameter_296: class Program_weight_tensor_parameter_297: name = "parameter_297" + original_name = "batch_norm2d_78.w_2" shape = [384] dtype = "float32" min_val = float("0.0314414") @@ -3278,6 +3576,7 @@ class Program_weight_tensor_parameter_297: class Program_weight_tensor_parameter_298: name = "parameter_298" + original_name = "batch_norm2d_78.w_1" shape = [384] dtype = "float32" min_val = float("-0.812785") @@ -3289,6 +3588,7 @@ class Program_weight_tensor_parameter_298: class Program_weight_tensor_parameter_299: name = "parameter_299" + original_name = "conv2d_82.w_0" shape = [384, 384, 3, 3] dtype = "float32" min_val = float("-0.0225096") @@ -3300,6 +3600,7 @@ class Program_weight_tensor_parameter_299: class Program_weight_tensor_parameter_300: name = "parameter_300" + original_name = "batch_norm2d_77.b_0" shape = [384] dtype = "float32" min_val = float("-0.033777") @@ -3311,6 +3612,7 @@ class Program_weight_tensor_parameter_300: class Program_weight_tensor_parameter_301: name = "parameter_301" + original_name = "batch_norm2d_77.w_0" shape = [384] dtype = "float32" min_val = float("0.984352") @@ -3322,6 +3624,7 @@ class Program_weight_tensor_parameter_301: class Program_weight_tensor_parameter_302: name = "parameter_302" + original_name = "batch_norm2d_77.w_2" shape = [384] dtype = "float32" min_val = float("0.00111383") @@ -3333,6 +3636,7 @@ class Program_weight_tensor_parameter_302: class Program_weight_tensor_parameter_303: name = "parameter_303" + original_name = "batch_norm2d_77.w_1" shape = [384] dtype = "float32" min_val = float("-0.0572523") @@ -3344,6 +3648,7 @@ class Program_weight_tensor_parameter_303: class Program_weight_tensor_parameter_304: name = "parameter_304" + original_name = "conv2d_81.w_0" shape = [384, 384, 1, 1] dtype = "float32" min_val = float("-0.0207385") @@ -3355,6 +3660,7 @@ class Program_weight_tensor_parameter_304: class Program_weight_tensor_parameter_305: name = "parameter_305" + original_name = "batch_norm2d_76.b_0" shape = [384] dtype = "float32" min_val = float("-0.033777") @@ -3366,6 +3672,7 @@ class Program_weight_tensor_parameter_305: class Program_weight_tensor_parameter_306: name = "parameter_306" + original_name = "batch_norm2d_76.w_0" shape = [384] dtype = "float32" min_val = float("0.982265") @@ -3377,6 +3684,7 @@ class Program_weight_tensor_parameter_306: class Program_weight_tensor_parameter_307: name = "parameter_307" + original_name = "batch_norm2d_76.w_2" shape = [384] dtype = "float32" min_val = float("0.00461328") @@ -3388,6 +3696,7 @@ class Program_weight_tensor_parameter_307: class Program_weight_tensor_parameter_308: name = "parameter_308" + original_name = "batch_norm2d_76.w_1" shape = [384] dtype = "float32" min_val = float("-0.141128") @@ -3399,6 +3708,7 @@ class Program_weight_tensor_parameter_308: class Program_weight_tensor_parameter_309: name = "parameter_309" + original_name = "conv2d_80.w_0" shape = [384, 384, 3, 3] dtype = "float32" min_val = float("-0.0123945") @@ -3410,6 +3720,7 @@ class Program_weight_tensor_parameter_309: class Program_weight_tensor_parameter_310: name = "parameter_310" + original_name = "batch_norm2d_75.b_0" shape = [384] dtype = "float32" min_val = float("-0.0511097") @@ -3421,6 +3732,7 @@ class Program_weight_tensor_parameter_310: class Program_weight_tensor_parameter_311: name = "parameter_311" + original_name = "batch_norm2d_75.w_0" shape = [384] dtype = "float32" min_val = float("0.976304") @@ -3432,6 +3744,7 @@ class Program_weight_tensor_parameter_311: class Program_weight_tensor_parameter_312: name = "parameter_312" + original_name = "batch_norm2d_75.w_2" shape = [384] dtype = "float32" min_val = float("0.00669036") @@ -3443,6 +3756,7 @@ class Program_weight_tensor_parameter_312: class Program_weight_tensor_parameter_313: name = "parameter_313" + original_name = "batch_norm2d_75.w_1" shape = [384] dtype = "float32" min_val = float("-0.11881") @@ -3454,6 +3768,7 @@ class Program_weight_tensor_parameter_313: class Program_weight_tensor_parameter_314: name = "parameter_314" + original_name = "conv2d_79.w_0" shape = [384, 384, 3, 3] dtype = "float32" min_val = float("-0.0142638") @@ -3465,6 +3780,7 @@ class Program_weight_tensor_parameter_314: class Program_weight_tensor_parameter_315: name = "parameter_315" + original_name = "batch_norm2d_74.b_0" shape = [384] dtype = "float32" min_val = float("-0.0675065") @@ -3476,6 +3792,7 @@ class Program_weight_tensor_parameter_315: class Program_weight_tensor_parameter_316: name = "parameter_316" + original_name = "batch_norm2d_74.w_0" shape = [384] dtype = "float32" min_val = float("0.981929") @@ -3487,6 +3804,7 @@ class Program_weight_tensor_parameter_316: class Program_weight_tensor_parameter_317: name = "parameter_317" + original_name = "batch_norm2d_74.w_2" shape = [384] dtype = "float32" min_val = float("0.00777962") @@ -3498,6 +3816,7 @@ class Program_weight_tensor_parameter_317: class Program_weight_tensor_parameter_318: name = "parameter_318" + original_name = "batch_norm2d_74.w_1" shape = [384] dtype = "float32" min_val = float("-0.101238") @@ -3509,6 +3828,7 @@ class Program_weight_tensor_parameter_318: class Program_weight_tensor_parameter_319: name = "parameter_319" + original_name = "conv2d_78.w_0" shape = [384, 1024, 1, 1] dtype = "float32" min_val = float("-0.0171645") @@ -3520,6 +3840,7 @@ class Program_weight_tensor_parameter_319: class Program_weight_tensor_parameter_320: name = "parameter_320" + original_name = "batch_norm2d_73.b_0" shape = [384] dtype = "float32" min_val = float("-0.0229577") @@ -3531,6 +3852,7 @@ class Program_weight_tensor_parameter_320: class Program_weight_tensor_parameter_321: name = "parameter_321" + original_name = "batch_norm2d_73.w_0" shape = [384] dtype = "float32" min_val = float("0.992697") @@ -3542,6 +3864,7 @@ class Program_weight_tensor_parameter_321: class Program_weight_tensor_parameter_322: name = "parameter_322" + original_name = "batch_norm2d_73.w_2" shape = [384] dtype = "float32" min_val = float("0.0119054") @@ -3553,6 +3876,7 @@ class Program_weight_tensor_parameter_322: class Program_weight_tensor_parameter_323: name = "parameter_323" + original_name = "batch_norm2d_73.w_1" shape = [384] dtype = "float32" min_val = float("-0.0851231") @@ -3564,6 +3888,7 @@ class Program_weight_tensor_parameter_323: class Program_weight_tensor_parameter_324: name = "parameter_324" + original_name = "conv2d_77.w_0" shape = [384, 1024, 1, 1] dtype = "float32" min_val = float("-0.0427528") @@ -3575,6 +3900,7 @@ class Program_weight_tensor_parameter_324: class Program_weight_tensor_parameter_325: name = "parameter_325" + original_name = "batch_norm2d_72.b_0" shape = [1024] dtype = "float32" min_val = float("-3.75857") @@ -3586,6 +3912,7 @@ class Program_weight_tensor_parameter_325: class Program_weight_tensor_parameter_326: name = "parameter_326" + original_name = "batch_norm2d_72.w_0" shape = [1024] dtype = "float32" min_val = float("1.61878") @@ -3597,6 +3924,7 @@ class Program_weight_tensor_parameter_326: class Program_weight_tensor_parameter_327: name = "parameter_327" + original_name = "batch_norm2d_72.w_2" shape = [1024] dtype = "float32" min_val = float("0.00282903") @@ -3608,6 +3936,7 @@ class Program_weight_tensor_parameter_327: class Program_weight_tensor_parameter_328: name = "parameter_328" + original_name = "batch_norm2d_72.w_1" shape = [1024] dtype = "float32" min_val = float("-0.0984391") @@ -3619,6 +3948,7 @@ class Program_weight_tensor_parameter_328: class Program_weight_tensor_parameter_329: name = "parameter_329" + original_name = "conv2d_76.w_0" shape = [1024, 768, 1, 1] dtype = "float32" min_val = float("-0.0757279") @@ -3630,6 +3960,7 @@ class Program_weight_tensor_parameter_329: class Program_weight_tensor_parameter_330: name = "parameter_330" + original_name = "conv2d_75.b_0" shape = [768] dtype = "float32" min_val = float("-0.0164166") @@ -3641,6 +3972,7 @@ class Program_weight_tensor_parameter_330: class Program_weight_tensor_parameter_331: name = "parameter_331" + original_name = "conv2d_75.w_0" shape = [768, 768, 1, 1] dtype = "float32" min_val = float("-0.085699") @@ -3652,6 +3984,7 @@ class Program_weight_tensor_parameter_331: class Program_weight_tensor_parameter_332: name = "parameter_332" + original_name = "batch_norm2d_71.b_0" shape = [384] dtype = "float32" min_val = float("-1.77377") @@ -3663,6 +3996,7 @@ class Program_weight_tensor_parameter_332: class Program_weight_tensor_parameter_333: name = "parameter_333" + original_name = "batch_norm2d_71.w_0" shape = [384] dtype = "float32" min_val = float("0.188432") @@ -3674,6 +4008,7 @@ class Program_weight_tensor_parameter_333: class Program_weight_tensor_parameter_334: name = "parameter_334" + original_name = "batch_norm2d_71.w_2" shape = [384] dtype = "float32" min_val = float("4.42942e-05") @@ -3685,6 +4020,7 @@ class Program_weight_tensor_parameter_334: class Program_weight_tensor_parameter_335: name = "parameter_335" + original_name = "batch_norm2d_71.w_1" shape = [384] dtype = "float32" min_val = float("-0.0900967") @@ -3696,6 +4032,7 @@ class Program_weight_tensor_parameter_335: class Program_weight_tensor_parameter_336: name = "parameter_336" + original_name = "conv2d_74.w_0" shape = [384, 384, 1, 1] dtype = "float32" min_val = float("-0.0238567") @@ -3707,6 +4044,7 @@ class Program_weight_tensor_parameter_336: class Program_weight_tensor_parameter_337: name = "parameter_337" + original_name = "batch_norm2d_70.b_0" shape = [384] dtype = "float32" min_val = float("-1.77377") @@ -3718,6 +4056,7 @@ class Program_weight_tensor_parameter_337: class Program_weight_tensor_parameter_338: name = "parameter_338" + original_name = "batch_norm2d_70.w_0" shape = [384] dtype = "float32" min_val = float("0.331669") @@ -3729,6 +4068,7 @@ class Program_weight_tensor_parameter_338: class Program_weight_tensor_parameter_339: name = "parameter_339" + original_name = "batch_norm2d_70.w_2" shape = [384] dtype = "float32" min_val = float("0.000339636") @@ -3740,6 +4080,7 @@ class Program_weight_tensor_parameter_339: class Program_weight_tensor_parameter_340: name = "parameter_340" + original_name = "batch_norm2d_70.w_1" shape = [384] dtype = "float32" min_val = float("-0.185857") @@ -3751,6 +4092,7 @@ class Program_weight_tensor_parameter_340: class Program_weight_tensor_parameter_341: name = "parameter_341" + original_name = "conv2d_73.w_0" shape = [384, 384, 3, 3] dtype = "float32" min_val = float("-0.0184368") @@ -3762,6 +4104,7 @@ class Program_weight_tensor_parameter_341: class Program_weight_tensor_parameter_342: name = "parameter_342" + original_name = "batch_norm2d_69.b_0" shape = [384] dtype = "float32" min_val = float("-2.58063") @@ -3773,6 +4116,7 @@ class Program_weight_tensor_parameter_342: class Program_weight_tensor_parameter_343: name = "parameter_343" + original_name = "batch_norm2d_69.w_0" shape = [384] dtype = "float32" min_val = float("0.524143") @@ -3784,6 +4128,7 @@ class Program_weight_tensor_parameter_343: class Program_weight_tensor_parameter_344: name = "parameter_344" + original_name = "batch_norm2d_69.w_2" shape = [384] dtype = "float32" min_val = float("0.0283643") @@ -3795,6 +4140,7 @@ class Program_weight_tensor_parameter_344: class Program_weight_tensor_parameter_345: name = "parameter_345" + original_name = "batch_norm2d_69.w_1" shape = [384] dtype = "float32" min_val = float("-0.790455") @@ -3806,6 +4152,7 @@ class Program_weight_tensor_parameter_345: class Program_weight_tensor_parameter_346: name = "parameter_346" + original_name = "conv2d_72.w_0" shape = [384, 384, 3, 3] dtype = "float32" min_val = float("-0.0295918") @@ -3817,6 +4164,7 @@ class Program_weight_tensor_parameter_346: class Program_weight_tensor_parameter_347: name = "parameter_347" + original_name = "batch_norm2d_68.b_0" shape = [384] dtype = "float32" min_val = float("-1.9389") @@ -3828,6 +4176,7 @@ class Program_weight_tensor_parameter_347: class Program_weight_tensor_parameter_348: name = "parameter_348" + original_name = "batch_norm2d_68.w_0" shape = [384] dtype = "float32" min_val = float("0.164447") @@ -3839,6 +4188,7 @@ class Program_weight_tensor_parameter_348: class Program_weight_tensor_parameter_349: name = "parameter_349" + original_name = "batch_norm2d_68.w_2" shape = [384] dtype = "float32" min_val = float("7.09489e-05") @@ -3850,6 +4200,7 @@ class Program_weight_tensor_parameter_349: class Program_weight_tensor_parameter_350: name = "parameter_350" + original_name = "batch_norm2d_68.w_1" shape = [384] dtype = "float32" min_val = float("-0.0343246") @@ -3861,6 +4212,7 @@ class Program_weight_tensor_parameter_350: class Program_weight_tensor_parameter_351: name = "parameter_351" + original_name = "conv2d_71.w_0" shape = [384, 384, 1, 1] dtype = "float32" min_val = float("-0.0210778") @@ -3872,6 +4224,7 @@ class Program_weight_tensor_parameter_351: class Program_weight_tensor_parameter_352: name = "parameter_352" + original_name = "batch_norm2d_67.b_0" shape = [384] dtype = "float32" min_val = float("-1.9389") @@ -3883,6 +4236,7 @@ class Program_weight_tensor_parameter_352: class Program_weight_tensor_parameter_353: name = "parameter_353" + original_name = "batch_norm2d_67.w_0" shape = [384] dtype = "float32" min_val = float("0.580775") @@ -3894,6 +4248,7 @@ class Program_weight_tensor_parameter_353: class Program_weight_tensor_parameter_354: name = "parameter_354" + original_name = "batch_norm2d_67.w_2" shape = [384] dtype = "float32" min_val = float("0.000843902") @@ -3905,6 +4260,7 @@ class Program_weight_tensor_parameter_354: class Program_weight_tensor_parameter_355: name = "parameter_355" + original_name = "batch_norm2d_67.w_1" shape = [384] dtype = "float32" min_val = float("-0.0632911") @@ -3916,6 +4272,7 @@ class Program_weight_tensor_parameter_355: class Program_weight_tensor_parameter_356: name = "parameter_356" + original_name = "conv2d_70.w_0" shape = [384, 384, 3, 3] dtype = "float32" min_val = float("-0.0186829") @@ -3927,6 +4284,7 @@ class Program_weight_tensor_parameter_356: class Program_weight_tensor_parameter_357: name = "parameter_357" + original_name = "batch_norm2d_66.b_0" shape = [384] dtype = "float32" min_val = float("-2.39583") @@ -3938,6 +4296,7 @@ class Program_weight_tensor_parameter_357: class Program_weight_tensor_parameter_358: name = "parameter_358" + original_name = "batch_norm2d_66.w_0" shape = [384] dtype = "float32" min_val = float("0.460879") @@ -3949,6 +4308,7 @@ class Program_weight_tensor_parameter_358: class Program_weight_tensor_parameter_359: name = "parameter_359" + original_name = "batch_norm2d_66.w_2" shape = [384] dtype = "float32" min_val = float("0.023841") @@ -3960,6 +4320,7 @@ class Program_weight_tensor_parameter_359: class Program_weight_tensor_parameter_360: name = "parameter_360" + original_name = "batch_norm2d_66.w_1" shape = [384] dtype = "float32" min_val = float("-0.608691") @@ -3971,6 +4332,7 @@ class Program_weight_tensor_parameter_360: class Program_weight_tensor_parameter_361: name = "parameter_361" + original_name = "conv2d_69.w_0" shape = [384, 384, 3, 3] dtype = "float32" min_val = float("-0.0202641") @@ -3982,6 +4344,7 @@ class Program_weight_tensor_parameter_361: class Program_weight_tensor_parameter_362: name = "parameter_362" + original_name = "batch_norm2d_65.b_0" shape = [384] dtype = "float32" min_val = float("-1.87571") @@ -3993,6 +4356,7 @@ class Program_weight_tensor_parameter_362: class Program_weight_tensor_parameter_363: name = "parameter_363" + original_name = "batch_norm2d_65.w_0" shape = [384] dtype = "float32" min_val = float("0.0770862") @@ -4004,6 +4368,7 @@ class Program_weight_tensor_parameter_363: class Program_weight_tensor_parameter_364: name = "parameter_364" + original_name = "batch_norm2d_65.w_2" shape = [384] dtype = "float32" min_val = float("5.43266e-05") @@ -4015,6 +4380,7 @@ class Program_weight_tensor_parameter_364: class Program_weight_tensor_parameter_365: name = "parameter_365" + original_name = "batch_norm2d_65.w_1" shape = [384] dtype = "float32" min_val = float("-0.0347425") @@ -4026,6 +4392,7 @@ class Program_weight_tensor_parameter_365: class Program_weight_tensor_parameter_366: name = "parameter_366" + original_name = "conv2d_68.w_0" shape = [384, 384, 1, 1] dtype = "float32" min_val = float("-0.0144415") @@ -4037,6 +4404,7 @@ class Program_weight_tensor_parameter_366: class Program_weight_tensor_parameter_367: name = "parameter_367" + original_name = "batch_norm2d_64.b_0" shape = [384] dtype = "float32" min_val = float("-1.87571") @@ -4048,6 +4416,7 @@ class Program_weight_tensor_parameter_367: class Program_weight_tensor_parameter_368: name = "parameter_368" + original_name = "batch_norm2d_64.w_0" shape = [384] dtype = "float32" min_val = float("0.517586") @@ -4059,6 +4428,7 @@ class Program_weight_tensor_parameter_368: class Program_weight_tensor_parameter_369: name = "parameter_369" + original_name = "batch_norm2d_64.w_2" shape = [384] dtype = "float32" min_val = float("0.000996551") @@ -4070,6 +4440,7 @@ class Program_weight_tensor_parameter_369: class Program_weight_tensor_parameter_370: name = "parameter_370" + original_name = "batch_norm2d_64.w_1" shape = [384] dtype = "float32" min_val = float("-0.154082") @@ -4081,6 +4452,7 @@ class Program_weight_tensor_parameter_370: class Program_weight_tensor_parameter_371: name = "parameter_371" + original_name = "conv2d_67.w_0" shape = [384, 384, 3, 3] dtype = "float32" min_val = float("-0.0199711") @@ -4092,6 +4464,7 @@ class Program_weight_tensor_parameter_371: class Program_weight_tensor_parameter_372: name = "parameter_372" + original_name = "batch_norm2d_63.b_0" shape = [384] dtype = "float32" min_val = float("-2.15499") @@ -4103,6 +4476,7 @@ class Program_weight_tensor_parameter_372: class Program_weight_tensor_parameter_373: name = "parameter_373" + original_name = "batch_norm2d_63.w_0" shape = [384] dtype = "float32" min_val = float("0.711873") @@ -4114,6 +4488,7 @@ class Program_weight_tensor_parameter_373: class Program_weight_tensor_parameter_374: name = "parameter_374" + original_name = "batch_norm2d_63.w_2" shape = [384] dtype = "float32" min_val = float("0.0152961") @@ -4125,6 +4500,7 @@ class Program_weight_tensor_parameter_374: class Program_weight_tensor_parameter_375: name = "parameter_375" + original_name = "batch_norm2d_63.w_1" shape = [384] dtype = "float32" min_val = float("-0.599745") @@ -4136,6 +4512,7 @@ class Program_weight_tensor_parameter_375: class Program_weight_tensor_parameter_376: name = "parameter_376" + original_name = "conv2d_66.w_0" shape = [384, 384, 3, 3] dtype = "float32" min_val = float("-0.0258655") @@ -4147,6 +4524,7 @@ class Program_weight_tensor_parameter_376: class Program_weight_tensor_parameter_377: name = "parameter_377" + original_name = "batch_norm2d_62.b_0" shape = [384] dtype = "float32" min_val = float("-2.92168") @@ -4158,6 +4536,7 @@ class Program_weight_tensor_parameter_377: class Program_weight_tensor_parameter_378: name = "parameter_378" + original_name = "batch_norm2d_62.w_0" shape = [384] dtype = "float32" min_val = float("0.951267") @@ -4169,6 +4548,7 @@ class Program_weight_tensor_parameter_378: class Program_weight_tensor_parameter_379: name = "parameter_379" + original_name = "batch_norm2d_62.w_2" shape = [384] dtype = "float32" min_val = float("0.00163804") @@ -4180,6 +4560,7 @@ class Program_weight_tensor_parameter_379: class Program_weight_tensor_parameter_380: name = "parameter_380" + original_name = "batch_norm2d_62.w_1" shape = [384] dtype = "float32" min_val = float("-0.203596") @@ -4191,6 +4572,7 @@ class Program_weight_tensor_parameter_380: class Program_weight_tensor_parameter_381: name = "parameter_381" + original_name = "conv2d_65.w_0" shape = [384, 768, 1, 1] dtype = "float32" min_val = float("-0.045466") @@ -4202,6 +4584,7 @@ class Program_weight_tensor_parameter_381: class Program_weight_tensor_parameter_382: name = "parameter_382" + original_name = "batch_norm2d_61.b_0" shape = [384] dtype = "float32" min_val = float("-2.2468") @@ -4213,6 +4596,7 @@ class Program_weight_tensor_parameter_382: class Program_weight_tensor_parameter_383: name = "parameter_383" + original_name = "batch_norm2d_61.w_0" shape = [384] dtype = "float32" min_val = float("0.963785") @@ -4224,6 +4608,7 @@ class Program_weight_tensor_parameter_383: class Program_weight_tensor_parameter_384: name = "parameter_384" + original_name = "batch_norm2d_61.w_2" shape = [384] dtype = "float32" min_val = float("0.000409017") @@ -4235,6 +4620,7 @@ class Program_weight_tensor_parameter_384: class Program_weight_tensor_parameter_385: name = "parameter_385" + original_name = "batch_norm2d_61.w_1" shape = [384] dtype = "float32" min_val = float("-0.0495182") @@ -4246,6 +4632,7 @@ class Program_weight_tensor_parameter_385: class Program_weight_tensor_parameter_386: name = "parameter_386" + original_name = "conv2d_64.w_0" shape = [384, 768, 1, 1] dtype = "float32" min_val = float("-0.145268") @@ -4257,6 +4644,7 @@ class Program_weight_tensor_parameter_386: class Program_weight_tensor_parameter_387: name = "parameter_387" + original_name = "batch_norm2d_60.b_0" shape = [768] dtype = "float32" min_val = float("-2.40179") @@ -4268,6 +4656,7 @@ class Program_weight_tensor_parameter_387: class Program_weight_tensor_parameter_388: name = "parameter_388" + original_name = "batch_norm2d_60.w_0" shape = [768] dtype = "float32" min_val = float("0.528614") @@ -4279,6 +4668,7 @@ class Program_weight_tensor_parameter_388: class Program_weight_tensor_parameter_389: name = "parameter_389" + original_name = "batch_norm2d_60.w_2" shape = [768] dtype = "float32" min_val = float("0.00512452") @@ -4290,6 +4680,7 @@ class Program_weight_tensor_parameter_389: class Program_weight_tensor_parameter_390: name = "parameter_390" + original_name = "batch_norm2d_60.w_1" shape = [768] dtype = "float32" min_val = float("-0.181132") @@ -4301,6 +4692,7 @@ class Program_weight_tensor_parameter_390: class Program_weight_tensor_parameter_391: name = "parameter_391" + original_name = "conv2d_63.w_0" shape = [768, 512, 3, 3] dtype = "float32" min_val = float("-0.0589591") @@ -4312,6 +4704,7 @@ class Program_weight_tensor_parameter_391: class Program_weight_tensor_parameter_392: name = "parameter_392" + original_name = "batch_norm2d_59.b_0" shape = [512] dtype = "float32" min_val = float("-3.39016") @@ -4323,6 +4716,7 @@ class Program_weight_tensor_parameter_392: class Program_weight_tensor_parameter_393: name = "parameter_393" + original_name = "batch_norm2d_59.w_0" shape = [512] dtype = "float32" min_val = float("0.522279") @@ -4334,6 +4728,7 @@ class Program_weight_tensor_parameter_393: class Program_weight_tensor_parameter_394: name = "parameter_394" + original_name = "batch_norm2d_59.w_2" shape = [512] dtype = "float32" min_val = float("0.00103462") @@ -4345,6 +4740,7 @@ class Program_weight_tensor_parameter_394: class Program_weight_tensor_parameter_395: name = "parameter_395" + original_name = "batch_norm2d_59.w_1" shape = [512] dtype = "float32" min_val = float("-0.109764") @@ -4356,6 +4752,7 @@ class Program_weight_tensor_parameter_395: class Program_weight_tensor_parameter_396: name = "parameter_396" + original_name = "conv2d_62.w_0" shape = [512, 384, 1, 1] dtype = "float32" min_val = float("-0.330093") @@ -4367,6 +4764,7 @@ class Program_weight_tensor_parameter_396: class Program_weight_tensor_parameter_397: name = "parameter_397" + original_name = "conv2d_61.b_0" shape = [384] dtype = "float32" min_val = float("-0.0107955") @@ -4378,6 +4776,7 @@ class Program_weight_tensor_parameter_397: class Program_weight_tensor_parameter_398: name = "parameter_398" + original_name = "conv2d_61.w_0" shape = [384, 384, 1, 1] dtype = "float32" min_val = float("-0.223299") @@ -4389,6 +4788,7 @@ class Program_weight_tensor_parameter_398: class Program_weight_tensor_parameter_399: name = "parameter_399" + original_name = "batch_norm2d_58.b_0" shape = [192] dtype = "float32" min_val = float("-1.97347") @@ -4400,6 +4800,7 @@ class Program_weight_tensor_parameter_399: class Program_weight_tensor_parameter_400: name = "parameter_400" + original_name = "batch_norm2d_58.w_0" shape = [192] dtype = "float32" min_val = float("0.0526394") @@ -4411,6 +4812,7 @@ class Program_weight_tensor_parameter_400: class Program_weight_tensor_parameter_401: name = "parameter_401" + original_name = "batch_norm2d_58.w_2" shape = [192] dtype = "float32" min_val = float("6.09083e-05") @@ -4422,6 +4824,7 @@ class Program_weight_tensor_parameter_401: class Program_weight_tensor_parameter_402: name = "parameter_402" + original_name = "batch_norm2d_58.w_1" shape = [192] dtype = "float32" min_val = float("-0.0270574") @@ -4433,6 +4836,7 @@ class Program_weight_tensor_parameter_402: class Program_weight_tensor_parameter_403: name = "parameter_403" + original_name = "conv2d_60.w_0" shape = [192, 192, 1, 1] dtype = "float32" min_val = float("-0.0211288") @@ -4444,6 +4848,7 @@ class Program_weight_tensor_parameter_403: class Program_weight_tensor_parameter_404: name = "parameter_404" + original_name = "batch_norm2d_57.b_0" shape = [192] dtype = "float32" min_val = float("-1.97347") @@ -4455,6 +4860,7 @@ class Program_weight_tensor_parameter_404: class Program_weight_tensor_parameter_405: name = "parameter_405" + original_name = "batch_norm2d_57.w_0" shape = [192] dtype = "float32" min_val = float("0.371516") @@ -4466,6 +4872,7 @@ class Program_weight_tensor_parameter_405: class Program_weight_tensor_parameter_406: name = "parameter_406" + original_name = "batch_norm2d_57.w_2" shape = [192] dtype = "float32" min_val = float("0.000794724") @@ -4477,6 +4884,7 @@ class Program_weight_tensor_parameter_406: class Program_weight_tensor_parameter_407: name = "parameter_407" + original_name = "batch_norm2d_57.w_1" shape = [192] dtype = "float32" min_val = float("-0.0718468") @@ -4488,6 +4896,7 @@ class Program_weight_tensor_parameter_407: class Program_weight_tensor_parameter_408: name = "parameter_408" + original_name = "conv2d_59.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.0267308") @@ -4499,6 +4908,7 @@ class Program_weight_tensor_parameter_408: class Program_weight_tensor_parameter_409: name = "parameter_409" + original_name = "batch_norm2d_56.b_0" shape = [192] dtype = "float32" min_val = float("-2.88942") @@ -4510,6 +4920,7 @@ class Program_weight_tensor_parameter_409: class Program_weight_tensor_parameter_410: name = "parameter_410" + original_name = "batch_norm2d_56.w_0" shape = [192] dtype = "float32" min_val = float("0.693744") @@ -4521,6 +4932,7 @@ class Program_weight_tensor_parameter_410: class Program_weight_tensor_parameter_411: name = "parameter_411" + original_name = "batch_norm2d_56.w_2" shape = [192] dtype = "float32" min_val = float("0.0397823") @@ -4532,6 +4944,7 @@ class Program_weight_tensor_parameter_411: class Program_weight_tensor_parameter_412: name = "parameter_412" + original_name = "batch_norm2d_56.w_1" shape = [192] dtype = "float32" min_val = float("-1.78945") @@ -4543,6 +4956,7 @@ class Program_weight_tensor_parameter_412: class Program_weight_tensor_parameter_413: name = "parameter_413" + original_name = "conv2d_58.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.0375591") @@ -4554,6 +4968,7 @@ class Program_weight_tensor_parameter_413: class Program_weight_tensor_parameter_414: name = "parameter_414" + original_name = "batch_norm2d_55.b_0" shape = [192] dtype = "float32" min_val = float("-1.93949") @@ -4565,6 +4980,7 @@ class Program_weight_tensor_parameter_414: class Program_weight_tensor_parameter_415: name = "parameter_415" + original_name = "batch_norm2d_55.w_0" shape = [192] dtype = "float32" min_val = float("0.0469211") @@ -4576,6 +4992,7 @@ class Program_weight_tensor_parameter_415: class Program_weight_tensor_parameter_416: name = "parameter_416" + original_name = "batch_norm2d_55.w_2" shape = [192] dtype = "float32" min_val = float("5.97514e-05") @@ -4587,6 +5004,7 @@ class Program_weight_tensor_parameter_416: class Program_weight_tensor_parameter_417: name = "parameter_417" + original_name = "batch_norm2d_55.w_1" shape = [192] dtype = "float32" min_val = float("-0.0317628") @@ -4598,6 +5016,7 @@ class Program_weight_tensor_parameter_417: class Program_weight_tensor_parameter_418: name = "parameter_418" + original_name = "conv2d_57.w_0" shape = [192, 192, 1, 1] dtype = "float32" min_val = float("-0.0232224") @@ -4609,6 +5028,7 @@ class Program_weight_tensor_parameter_418: class Program_weight_tensor_parameter_419: name = "parameter_419" + original_name = "batch_norm2d_54.b_0" shape = [192] dtype = "float32" min_val = float("-1.93949") @@ -4620,6 +5040,7 @@ class Program_weight_tensor_parameter_419: class Program_weight_tensor_parameter_420: name = "parameter_420" + original_name = "batch_norm2d_54.w_0" shape = [192] dtype = "float32" min_val = float("0.485243") @@ -4631,6 +5052,7 @@ class Program_weight_tensor_parameter_420: class Program_weight_tensor_parameter_421: name = "parameter_421" + original_name = "batch_norm2d_54.w_2" shape = [192] dtype = "float32" min_val = float("0.00155942") @@ -4642,6 +5064,7 @@ class Program_weight_tensor_parameter_421: class Program_weight_tensor_parameter_422: name = "parameter_422" + original_name = "batch_norm2d_54.w_1" shape = [192] dtype = "float32" min_val = float("-0.059897") @@ -4653,6 +5076,7 @@ class Program_weight_tensor_parameter_422: class Program_weight_tensor_parameter_423: name = "parameter_423" + original_name = "conv2d_56.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.021281") @@ -4664,6 +5088,7 @@ class Program_weight_tensor_parameter_423: class Program_weight_tensor_parameter_424: name = "parameter_424" + original_name = "batch_norm2d_53.b_0" shape = [192] dtype = "float32" min_val = float("-2.50671") @@ -4675,6 +5100,7 @@ class Program_weight_tensor_parameter_424: class Program_weight_tensor_parameter_425: name = "parameter_425" + original_name = "batch_norm2d_53.w_0" shape = [192] dtype = "float32" min_val = float("0.654125") @@ -4686,6 +5112,7 @@ class Program_weight_tensor_parameter_425: class Program_weight_tensor_parameter_426: name = "parameter_426" + original_name = "batch_norm2d_53.w_2" shape = [192] dtype = "float32" min_val = float("0.0320525") @@ -4697,6 +5124,7 @@ class Program_weight_tensor_parameter_426: class Program_weight_tensor_parameter_427: name = "parameter_427" + original_name = "batch_norm2d_53.w_1" shape = [192] dtype = "float32" min_val = float("-1.74443") @@ -4708,6 +5136,7 @@ class Program_weight_tensor_parameter_427: class Program_weight_tensor_parameter_428: name = "parameter_428" + original_name = "conv2d_55.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.0357247") @@ -4719,6 +5148,7 @@ class Program_weight_tensor_parameter_428: class Program_weight_tensor_parameter_429: name = "parameter_429" + original_name = "batch_norm2d_52.b_0" shape = [192] dtype = "float32" min_val = float("-1.75648") @@ -4730,6 +5160,7 @@ class Program_weight_tensor_parameter_429: class Program_weight_tensor_parameter_430: name = "parameter_430" + original_name = "batch_norm2d_52.w_0" shape = [192] dtype = "float32" min_val = float("0.00362431") @@ -4741,6 +5172,7 @@ class Program_weight_tensor_parameter_430: class Program_weight_tensor_parameter_431: name = "parameter_431" + original_name = "batch_norm2d_52.w_2" shape = [192] dtype = "float32" min_val = float("1.06024e-06") @@ -4752,6 +5184,7 @@ class Program_weight_tensor_parameter_431: class Program_weight_tensor_parameter_432: name = "parameter_432" + original_name = "batch_norm2d_52.w_1" shape = [192] dtype = "float32" min_val = float("-0.0316145") @@ -4763,6 +5196,7 @@ class Program_weight_tensor_parameter_432: class Program_weight_tensor_parameter_433: name = "parameter_433" + original_name = "conv2d_54.w_0" shape = [192, 192, 1, 1] dtype = "float32" min_val = float("-0.0332344") @@ -4774,6 +5208,7 @@ class Program_weight_tensor_parameter_433: class Program_weight_tensor_parameter_434: name = "parameter_434" + original_name = "batch_norm2d_51.b_0" shape = [192] dtype = "float32" min_val = float("-1.75648") @@ -4785,6 +5220,7 @@ class Program_weight_tensor_parameter_434: class Program_weight_tensor_parameter_435: name = "parameter_435" + original_name = "batch_norm2d_51.w_0" shape = [192] dtype = "float32" min_val = float("0.40635") @@ -4796,6 +5232,7 @@ class Program_weight_tensor_parameter_435: class Program_weight_tensor_parameter_436: name = "parameter_436" + original_name = "batch_norm2d_51.w_2" shape = [192] dtype = "float32" min_val = float("0.00153847") @@ -4807,6 +5244,7 @@ class Program_weight_tensor_parameter_436: class Program_weight_tensor_parameter_437: name = "parameter_437" + original_name = "batch_norm2d_51.w_1" shape = [192] dtype = "float32" min_val = float("-0.0309003") @@ -4818,6 +5256,7 @@ class Program_weight_tensor_parameter_437: class Program_weight_tensor_parameter_438: name = "parameter_438" + original_name = "conv2d_53.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.0284898") @@ -4829,6 +5268,7 @@ class Program_weight_tensor_parameter_438: class Program_weight_tensor_parameter_439: name = "parameter_439" + original_name = "batch_norm2d_50.b_0" shape = [192] dtype = "float32" min_val = float("-2.49598") @@ -4840,6 +5280,7 @@ class Program_weight_tensor_parameter_439: class Program_weight_tensor_parameter_440: name = "parameter_440" + original_name = "batch_norm2d_50.w_0" shape = [192] dtype = "float32" min_val = float("0.655528") @@ -4851,6 +5292,7 @@ class Program_weight_tensor_parameter_440: class Program_weight_tensor_parameter_441: name = "parameter_441" + original_name = "batch_norm2d_50.w_2" shape = [192] dtype = "float32" min_val = float("0.0187082") @@ -4862,6 +5304,7 @@ class Program_weight_tensor_parameter_441: class Program_weight_tensor_parameter_442: name = "parameter_442" + original_name = "batch_norm2d_50.w_1" shape = [192] dtype = "float32" min_val = float("-1.21444") @@ -4873,6 +5316,7 @@ class Program_weight_tensor_parameter_442: class Program_weight_tensor_parameter_443: name = "parameter_443" + original_name = "conv2d_52.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.0372317") @@ -4884,6 +5328,7 @@ class Program_weight_tensor_parameter_443: class Program_weight_tensor_parameter_444: name = "parameter_444" + original_name = "batch_norm2d_49.b_0" shape = [192] dtype = "float32" min_val = float("-2.07869") @@ -4895,6 +5340,7 @@ class Program_weight_tensor_parameter_444: class Program_weight_tensor_parameter_445: name = "parameter_445" + original_name = "batch_norm2d_49.w_0" shape = [192] dtype = "float32" min_val = float("0.000538036") @@ -4906,6 +5352,7 @@ class Program_weight_tensor_parameter_445: class Program_weight_tensor_parameter_446: name = "parameter_446" + original_name = "batch_norm2d_49.w_2" shape = [192] dtype = "float32" min_val = float("6.11339e-08") @@ -4917,6 +5364,7 @@ class Program_weight_tensor_parameter_446: class Program_weight_tensor_parameter_447: name = "parameter_447" + original_name = "batch_norm2d_49.w_1" shape = [192] dtype = "float32" min_val = float("-0.0245402") @@ -4928,6 +5376,7 @@ class Program_weight_tensor_parameter_447: class Program_weight_tensor_parameter_448: name = "parameter_448" + original_name = "conv2d_51.w_0" shape = [192, 192, 1, 1] dtype = "float32" min_val = float("-0.0222475") @@ -4939,6 +5388,7 @@ class Program_weight_tensor_parameter_448: class Program_weight_tensor_parameter_449: name = "parameter_449" + original_name = "batch_norm2d_48.b_0" shape = [192] dtype = "float32" min_val = float("-2.07869") @@ -4950,6 +5400,7 @@ class Program_weight_tensor_parameter_449: class Program_weight_tensor_parameter_450: name = "parameter_450" + original_name = "batch_norm2d_48.w_0" shape = [192] dtype = "float32" min_val = float("0.395056") @@ -4961,6 +5412,7 @@ class Program_weight_tensor_parameter_450: class Program_weight_tensor_parameter_451: name = "parameter_451" + original_name = "batch_norm2d_48.w_2" shape = [192] dtype = "float32" min_val = float("0.00160291") @@ -4972,6 +5424,7 @@ class Program_weight_tensor_parameter_451: class Program_weight_tensor_parameter_452: name = "parameter_452" + original_name = "batch_norm2d_48.w_1" shape = [192] dtype = "float32" min_val = float("-0.0313156") @@ -4983,6 +5436,7 @@ class Program_weight_tensor_parameter_452: class Program_weight_tensor_parameter_453: name = "parameter_453" + original_name = "conv2d_50.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.0299429") @@ -4994,6 +5448,7 @@ class Program_weight_tensor_parameter_453: class Program_weight_tensor_parameter_454: name = "parameter_454" + original_name = "batch_norm2d_47.b_0" shape = [192] dtype = "float32" min_val = float("-2.7392") @@ -5005,6 +5460,7 @@ class Program_weight_tensor_parameter_454: class Program_weight_tensor_parameter_455: name = "parameter_455" + original_name = "batch_norm2d_47.w_0" shape = [192] dtype = "float32" min_val = float("0.761401") @@ -5016,6 +5472,7 @@ class Program_weight_tensor_parameter_455: class Program_weight_tensor_parameter_456: name = "parameter_456" + original_name = "batch_norm2d_47.w_2" shape = [192] dtype = "float32" min_val = float("0.0164018") @@ -5027,6 +5484,7 @@ class Program_weight_tensor_parameter_456: class Program_weight_tensor_parameter_457: name = "parameter_457" + original_name = "batch_norm2d_47.w_1" shape = [192] dtype = "float32" min_val = float("-0.982911") @@ -5038,6 +5496,7 @@ class Program_weight_tensor_parameter_457: class Program_weight_tensor_parameter_458: name = "parameter_458" + original_name = "conv2d_49.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.0516537") @@ -5049,6 +5508,7 @@ class Program_weight_tensor_parameter_458: class Program_weight_tensor_parameter_459: name = "parameter_459" + original_name = "batch_norm2d_46.b_0" shape = [192] dtype = "float32" min_val = float("-1.21222") @@ -5060,6 +5520,7 @@ class Program_weight_tensor_parameter_459: class Program_weight_tensor_parameter_460: name = "parameter_460" + original_name = "batch_norm2d_46.w_0" shape = [192] dtype = "float32" min_val = float("-0.000141087") @@ -5071,6 +5532,7 @@ class Program_weight_tensor_parameter_460: class Program_weight_tensor_parameter_461: name = "parameter_461" + original_name = "batch_norm2d_46.w_2" shape = [192] dtype = "float32" min_val = float("2.02891e-10") @@ -5082,6 +5544,7 @@ class Program_weight_tensor_parameter_461: class Program_weight_tensor_parameter_462: name = "parameter_462" + original_name = "batch_norm2d_46.w_1" shape = [192] dtype = "float32" min_val = float("-0.0400934") @@ -5093,6 +5556,7 @@ class Program_weight_tensor_parameter_462: class Program_weight_tensor_parameter_463: name = "parameter_463" + original_name = "conv2d_48.w_0" shape = [192, 192, 1, 1] dtype = "float32" min_val = float("-0.0361348") @@ -5104,6 +5568,7 @@ class Program_weight_tensor_parameter_463: class Program_weight_tensor_parameter_464: name = "parameter_464" + original_name = "batch_norm2d_45.b_0" shape = [192] dtype = "float32" min_val = float("-1.21222") @@ -5115,6 +5580,7 @@ class Program_weight_tensor_parameter_464: class Program_weight_tensor_parameter_465: name = "parameter_465" + original_name = "batch_norm2d_45.w_0" shape = [192] dtype = "float32" min_val = float("0.38394") @@ -5126,6 +5592,7 @@ class Program_weight_tensor_parameter_465: class Program_weight_tensor_parameter_466: name = "parameter_466" + original_name = "batch_norm2d_45.w_2" shape = [192] dtype = "float32" min_val = float("0.00141501") @@ -5137,6 +5604,7 @@ class Program_weight_tensor_parameter_466: class Program_weight_tensor_parameter_467: name = "parameter_467" + original_name = "batch_norm2d_45.w_1" shape = [192] dtype = "float32" min_val = float("-0.0684229") @@ -5148,6 +5616,7 @@ class Program_weight_tensor_parameter_467: class Program_weight_tensor_parameter_468: name = "parameter_468" + original_name = "conv2d_47.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.0302635") @@ -5159,6 +5628,7 @@ class Program_weight_tensor_parameter_468: class Program_weight_tensor_parameter_469: name = "parameter_469" + original_name = "batch_norm2d_44.b_0" shape = [192] dtype = "float32" min_val = float("-2.48525") @@ -5170,6 +5640,7 @@ class Program_weight_tensor_parameter_469: class Program_weight_tensor_parameter_470: name = "parameter_470" + original_name = "batch_norm2d_44.w_0" shape = [192] dtype = "float32" min_val = float("0.687779") @@ -5181,6 +5652,7 @@ class Program_weight_tensor_parameter_470: class Program_weight_tensor_parameter_471: name = "parameter_471" + original_name = "batch_norm2d_44.w_2" shape = [192] dtype = "float32" min_val = float("0.0103392") @@ -5192,6 +5664,7 @@ class Program_weight_tensor_parameter_471: class Program_weight_tensor_parameter_472: name = "parameter_472" + original_name = "batch_norm2d_44.w_1" shape = [192] dtype = "float32" min_val = float("-0.598739") @@ -5203,6 +5676,7 @@ class Program_weight_tensor_parameter_472: class Program_weight_tensor_parameter_473: name = "parameter_473" + original_name = "conv2d_46.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.053814") @@ -5214,6 +5688,7 @@ class Program_weight_tensor_parameter_473: class Program_weight_tensor_parameter_474: name = "parameter_474" + original_name = "batch_norm2d_43.b_0" shape = [192] dtype = "float32" min_val = float("-1.21759") @@ -5225,6 +5700,7 @@ class Program_weight_tensor_parameter_474: class Program_weight_tensor_parameter_475: name = "parameter_475" + original_name = "batch_norm2d_43.w_0" shape = [192] dtype = "float32" min_val = float("0.00874149") @@ -5236,6 +5712,7 @@ class Program_weight_tensor_parameter_475: class Program_weight_tensor_parameter_476: name = "parameter_476" + original_name = "batch_norm2d_43.w_2" shape = [192] dtype = "float32" min_val = float("1.85538e-05") @@ -5247,6 +5724,7 @@ class Program_weight_tensor_parameter_476: class Program_weight_tensor_parameter_477: name = "parameter_477" + original_name = "batch_norm2d_43.w_1" shape = [192] dtype = "float32" min_val = float("-0.0629882") @@ -5258,6 +5736,7 @@ class Program_weight_tensor_parameter_477: class Program_weight_tensor_parameter_478: name = "parameter_478" + original_name = "conv2d_45.w_0" shape = [192, 192, 1, 1] dtype = "float32" min_val = float("-0.0659294") @@ -5269,6 +5748,7 @@ class Program_weight_tensor_parameter_478: class Program_weight_tensor_parameter_479: name = "parameter_479" + original_name = "batch_norm2d_42.b_0" shape = [192] dtype = "float32" min_val = float("-1.21759") @@ -5280,6 +5760,7 @@ class Program_weight_tensor_parameter_479: class Program_weight_tensor_parameter_480: name = "parameter_480" + original_name = "batch_norm2d_42.w_0" shape = [192] dtype = "float32" min_val = float("0.355067") @@ -5291,6 +5772,7 @@ class Program_weight_tensor_parameter_480: class Program_weight_tensor_parameter_481: name = "parameter_481" + original_name = "batch_norm2d_42.w_2" shape = [192] dtype = "float32" min_val = float("0.00277976") @@ -5302,6 +5784,7 @@ class Program_weight_tensor_parameter_481: class Program_weight_tensor_parameter_482: name = "parameter_482" + original_name = "batch_norm2d_42.w_1" shape = [192] dtype = "float32" min_val = float("-0.0724428") @@ -5313,6 +5796,7 @@ class Program_weight_tensor_parameter_482: class Program_weight_tensor_parameter_483: name = "parameter_483" + original_name = "conv2d_44.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.0699304") @@ -5324,6 +5808,7 @@ class Program_weight_tensor_parameter_483: class Program_weight_tensor_parameter_484: name = "parameter_484" + original_name = "batch_norm2d_41.b_0" shape = [192] dtype = "float32" min_val = float("-1.8763") @@ -5335,6 +5820,7 @@ class Program_weight_tensor_parameter_484: class Program_weight_tensor_parameter_485: name = "parameter_485" + original_name = "batch_norm2d_41.w_0" shape = [192] dtype = "float32" min_val = float("0.790047") @@ -5346,6 +5832,7 @@ class Program_weight_tensor_parameter_485: class Program_weight_tensor_parameter_486: name = "parameter_486" + original_name = "batch_norm2d_41.w_2" shape = [192] dtype = "float32" min_val = float("0.00787833") @@ -5357,6 +5844,7 @@ class Program_weight_tensor_parameter_486: class Program_weight_tensor_parameter_487: name = "parameter_487" + original_name = "batch_norm2d_41.w_1" shape = [192] dtype = "float32" min_val = float("-0.452939") @@ -5368,6 +5856,7 @@ class Program_weight_tensor_parameter_487: class Program_weight_tensor_parameter_488: name = "parameter_488" + original_name = "conv2d_43.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.0608784") @@ -5379,6 +5868,7 @@ class Program_weight_tensor_parameter_488: class Program_weight_tensor_parameter_489: name = "parameter_489" + original_name = "batch_norm2d_40.b_0" shape = [192] dtype = "float32" min_val = float("-2.8608") @@ -5390,6 +5880,7 @@ class Program_weight_tensor_parameter_489: class Program_weight_tensor_parameter_490: name = "parameter_490" + original_name = "batch_norm2d_40.w_0" shape = [192] dtype = "float32" min_val = float("0.485748") @@ -5401,6 +5892,7 @@ class Program_weight_tensor_parameter_490: class Program_weight_tensor_parameter_491: name = "parameter_491" + original_name = "batch_norm2d_40.w_2" shape = [192] dtype = "float32" min_val = float("0.0067126") @@ -5412,6 +5904,7 @@ class Program_weight_tensor_parameter_491: class Program_weight_tensor_parameter_492: name = "parameter_492" + original_name = "batch_norm2d_40.w_1" shape = [192] dtype = "float32" min_val = float("-0.197878") @@ -5423,6 +5916,7 @@ class Program_weight_tensor_parameter_492: class Program_weight_tensor_parameter_493: name = "parameter_493" + original_name = "conv2d_42.w_0" shape = [192, 384, 1, 1] dtype = "float32" min_val = float("-0.0901092") @@ -5434,6 +5928,7 @@ class Program_weight_tensor_parameter_493: class Program_weight_tensor_parameter_494: name = "parameter_494" + original_name = "batch_norm2d_39.b_0" shape = [192] dtype = "float32" min_val = float("-2.96811") @@ -5445,6 +5940,7 @@ class Program_weight_tensor_parameter_494: class Program_weight_tensor_parameter_495: name = "parameter_495" + original_name = "batch_norm2d_39.w_0" shape = [192] dtype = "float32" min_val = float("0.83173") @@ -5456,6 +5952,7 @@ class Program_weight_tensor_parameter_495: class Program_weight_tensor_parameter_496: name = "parameter_496" + original_name = "batch_norm2d_39.w_2" shape = [192] dtype = "float32" min_val = float("0.0026531") @@ -5467,6 +5964,7 @@ class Program_weight_tensor_parameter_496: class Program_weight_tensor_parameter_497: name = "parameter_497" + original_name = "batch_norm2d_39.w_1" shape = [192] dtype = "float32" min_val = float("-0.12448") @@ -5478,6 +5976,7 @@ class Program_weight_tensor_parameter_497: class Program_weight_tensor_parameter_498: name = "parameter_498" + original_name = "conv2d_41.w_0" shape = [192, 384, 1, 1] dtype = "float32" min_val = float("-0.0687437") @@ -5489,6 +5988,7 @@ class Program_weight_tensor_parameter_498: class Program_weight_tensor_parameter_499: name = "parameter_499" + original_name = "batch_norm2d_38.b_0" shape = [384] dtype = "float32" min_val = float("-2.9228") @@ -5500,6 +6000,7 @@ class Program_weight_tensor_parameter_499: class Program_weight_tensor_parameter_500: name = "parameter_500" + original_name = "batch_norm2d_38.w_0" shape = [384] dtype = "float32" min_val = float("0.638427") @@ -5511,6 +6012,7 @@ class Program_weight_tensor_parameter_500: class Program_weight_tensor_parameter_501: name = "parameter_501" + original_name = "batch_norm2d_38.w_2" shape = [384] dtype = "float32" min_val = float("0.00594162") @@ -5522,6 +6024,7 @@ class Program_weight_tensor_parameter_501: class Program_weight_tensor_parameter_502: name = "parameter_502" + original_name = "batch_norm2d_38.w_1" shape = [384] dtype = "float32" min_val = float("-0.173063") @@ -5533,6 +6036,7 @@ class Program_weight_tensor_parameter_502: class Program_weight_tensor_parameter_503: name = "parameter_503" + original_name = "conv2d_40.w_0" shape = [384, 256, 3, 3] dtype = "float32" min_val = float("-0.0649605") @@ -5544,6 +6048,7 @@ class Program_weight_tensor_parameter_503: class Program_weight_tensor_parameter_504: name = "parameter_504" + original_name = "batch_norm2d_37.b_0" shape = [256] dtype = "float32" min_val = float("-2.04411") @@ -5555,6 +6060,7 @@ class Program_weight_tensor_parameter_504: class Program_weight_tensor_parameter_505: name = "parameter_505" + original_name = "batch_norm2d_37.w_0" shape = [256] dtype = "float32" min_val = float("0.528197") @@ -5566,6 +6072,7 @@ class Program_weight_tensor_parameter_505: class Program_weight_tensor_parameter_506: name = "parameter_506" + original_name = "batch_norm2d_37.w_2" shape = [256] dtype = "float32" min_val = float("0.00073195") @@ -5577,6 +6084,7 @@ class Program_weight_tensor_parameter_506: class Program_weight_tensor_parameter_507: name = "parameter_507" + original_name = "batch_norm2d_37.w_1" shape = [256] dtype = "float32" min_val = float("-0.164563") @@ -5588,6 +6096,7 @@ class Program_weight_tensor_parameter_507: class Program_weight_tensor_parameter_508: name = "parameter_508" + original_name = "conv2d_39.w_0" shape = [256, 192, 1, 1] dtype = "float32" min_val = float("-0.165736") @@ -5599,6 +6108,7 @@ class Program_weight_tensor_parameter_508: class Program_weight_tensor_parameter_509: name = "parameter_509" + original_name = "conv2d_38.b_0" shape = [192] dtype = "float32" min_val = float("-0.0132348") @@ -5610,6 +6120,7 @@ class Program_weight_tensor_parameter_509: class Program_weight_tensor_parameter_510: name = "parameter_510" + original_name = "conv2d_38.w_0" shape = [192, 192, 1, 1] dtype = "float32" min_val = float("-0.308917") @@ -5621,6 +6132,7 @@ class Program_weight_tensor_parameter_510: class Program_weight_tensor_parameter_511: name = "parameter_511" + original_name = "batch_norm2d_36.b_0" shape = [96] dtype = "float32" min_val = float("-1.91684") @@ -5632,6 +6144,7 @@ class Program_weight_tensor_parameter_511: class Program_weight_tensor_parameter_512: name = "parameter_512" + original_name = "batch_norm2d_36.w_0" shape = [96] dtype = "float32" min_val = float("0.140578") @@ -5643,6 +6156,7 @@ class Program_weight_tensor_parameter_512: class Program_weight_tensor_parameter_513: name = "parameter_513" + original_name = "batch_norm2d_36.w_2" shape = [96] dtype = "float32" min_val = float("7.2207e-05") @@ -5654,6 +6168,7 @@ class Program_weight_tensor_parameter_513: class Program_weight_tensor_parameter_514: name = "parameter_514" + original_name = "batch_norm2d_36.w_1" shape = [96] dtype = "float32" min_val = float("-0.0374736") @@ -5665,6 +6180,7 @@ class Program_weight_tensor_parameter_514: class Program_weight_tensor_parameter_515: name = "parameter_515" + original_name = "conv2d_37.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.041906") @@ -5676,6 +6192,7 @@ class Program_weight_tensor_parameter_515: class Program_weight_tensor_parameter_516: name = "parameter_516" + original_name = "batch_norm2d_35.b_0" shape = [96] dtype = "float32" min_val = float("-1.91684") @@ -5687,6 +6204,7 @@ class Program_weight_tensor_parameter_516: class Program_weight_tensor_parameter_517: name = "parameter_517" + original_name = "batch_norm2d_35.w_0" shape = [96] dtype = "float32" min_val = float("0.349141") @@ -5698,6 +6216,7 @@ class Program_weight_tensor_parameter_517: class Program_weight_tensor_parameter_518: name = "parameter_518" + original_name = "batch_norm2d_35.w_2" shape = [96] dtype = "float32" min_val = float("0.000395259") @@ -5709,6 +6228,7 @@ class Program_weight_tensor_parameter_518: class Program_weight_tensor_parameter_519: name = "parameter_519" + original_name = "batch_norm2d_35.w_1" shape = [96] dtype = "float32" min_val = float("-0.0894642") @@ -5720,6 +6240,7 @@ class Program_weight_tensor_parameter_519: class Program_weight_tensor_parameter_520: name = "parameter_520" + original_name = "conv2d_36.w_0" shape = [96, 96, 3, 3] dtype = "float32" min_val = float("-0.0396274") @@ -5731,6 +6252,7 @@ class Program_weight_tensor_parameter_520: class Program_weight_tensor_parameter_521: name = "parameter_521" + original_name = "batch_norm2d_34.b_0" shape = [96] dtype = "float32" min_val = float("-2.46379") @@ -5742,6 +6264,7 @@ class Program_weight_tensor_parameter_521: class Program_weight_tensor_parameter_522: name = "parameter_522" + original_name = "batch_norm2d_34.w_0" shape = [96] dtype = "float32" min_val = float("0.531857") @@ -5753,6 +6276,7 @@ class Program_weight_tensor_parameter_522: class Program_weight_tensor_parameter_523: name = "parameter_523" + original_name = "batch_norm2d_34.w_2" shape = [96] dtype = "float32" min_val = float("0.025952") @@ -5764,6 +6288,7 @@ class Program_weight_tensor_parameter_523: class Program_weight_tensor_parameter_524: name = "parameter_524" + original_name = "batch_norm2d_34.w_1" shape = [96] dtype = "float32" min_val = float("-2.5031") @@ -5775,6 +6300,7 @@ class Program_weight_tensor_parameter_524: class Program_weight_tensor_parameter_525: name = "parameter_525" + original_name = "conv2d_35.w_0" shape = [96, 96, 3, 3] dtype = "float32" min_val = float("-0.17314") @@ -5786,6 +6312,7 @@ class Program_weight_tensor_parameter_525: class Program_weight_tensor_parameter_526: name = "parameter_526" + original_name = "batch_norm2d_33.b_0" shape = [96] dtype = "float32" min_val = float("-1.38808") @@ -5797,6 +6324,7 @@ class Program_weight_tensor_parameter_526: class Program_weight_tensor_parameter_527: name = "parameter_527" + original_name = "batch_norm2d_33.w_0" shape = [96] dtype = "float32" min_val = float("0.0455848") @@ -5808,6 +6336,7 @@ class Program_weight_tensor_parameter_527: class Program_weight_tensor_parameter_528: name = "parameter_528" + original_name = "batch_norm2d_33.w_2" shape = [96] dtype = "float32" min_val = float("9.00694e-05") @@ -5819,6 +6348,7 @@ class Program_weight_tensor_parameter_528: class Program_weight_tensor_parameter_529: name = "parameter_529" + original_name = "batch_norm2d_33.w_1" shape = [96] dtype = "float32" min_val = float("-0.0278321") @@ -5830,6 +6360,7 @@ class Program_weight_tensor_parameter_529: class Program_weight_tensor_parameter_530: name = "parameter_530" + original_name = "conv2d_34.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.0497848") @@ -5841,6 +6372,7 @@ class Program_weight_tensor_parameter_530: class Program_weight_tensor_parameter_531: name = "parameter_531" + original_name = "batch_norm2d_32.b_0" shape = [96] dtype = "float32" min_val = float("-1.38808") @@ -5852,6 +6384,7 @@ class Program_weight_tensor_parameter_531: class Program_weight_tensor_parameter_532: name = "parameter_532" + original_name = "batch_norm2d_32.w_0" shape = [96] dtype = "float32" min_val = float("0.368559") @@ -5863,6 +6396,7 @@ class Program_weight_tensor_parameter_532: class Program_weight_tensor_parameter_533: name = "parameter_533" + original_name = "batch_norm2d_32.w_2" shape = [96] dtype = "float32" min_val = float("0.00148637") @@ -5874,6 +6408,7 @@ class Program_weight_tensor_parameter_533: class Program_weight_tensor_parameter_534: name = "parameter_534" + original_name = "batch_norm2d_32.w_1" shape = [96] dtype = "float32" min_val = float("-0.0662393") @@ -5885,6 +6420,7 @@ class Program_weight_tensor_parameter_534: class Program_weight_tensor_parameter_535: name = "parameter_535" + original_name = "conv2d_33.w_0" shape = [96, 96, 3, 3] dtype = "float32" min_val = float("-0.0543383") @@ -5896,6 +6432,7 @@ class Program_weight_tensor_parameter_535: class Program_weight_tensor_parameter_536: name = "parameter_536" + original_name = "batch_norm2d_31.b_0" shape = [96] dtype = "float32" min_val = float("-3.31982") @@ -5907,6 +6444,7 @@ class Program_weight_tensor_parameter_536: class Program_weight_tensor_parameter_537: name = "parameter_537" + original_name = "batch_norm2d_31.w_0" shape = [96] dtype = "float32" min_val = float("0.46978") @@ -5918,6 +6456,7 @@ class Program_weight_tensor_parameter_537: class Program_weight_tensor_parameter_538: name = "parameter_538" + original_name = "batch_norm2d_31.w_2" shape = [96] dtype = "float32" min_val = float("0.0175926") @@ -5929,6 +6468,7 @@ class Program_weight_tensor_parameter_538: class Program_weight_tensor_parameter_539: name = "parameter_539" + original_name = "batch_norm2d_31.w_1" shape = [96] dtype = "float32" min_val = float("-0.65582") @@ -5940,6 +6480,7 @@ class Program_weight_tensor_parameter_539: class Program_weight_tensor_parameter_540: name = "parameter_540" + original_name = "conv2d_32.w_0" shape = [96, 96, 3, 3] dtype = "float32" min_val = float("-0.135994") @@ -5951,6 +6492,7 @@ class Program_weight_tensor_parameter_540: class Program_weight_tensor_parameter_541: name = "parameter_541" + original_name = "batch_norm2d_30.b_0" shape = [96] dtype = "float32" min_val = float("-1.24978") @@ -5962,6 +6504,7 @@ class Program_weight_tensor_parameter_541: class Program_weight_tensor_parameter_542: name = "parameter_542" + original_name = "batch_norm2d_30.w_0" shape = [96] dtype = "float32" min_val = float("0.0246004") @@ -5973,6 +6516,7 @@ class Program_weight_tensor_parameter_542: class Program_weight_tensor_parameter_543: name = "parameter_543" + original_name = "batch_norm2d_30.w_2" shape = [96] dtype = "float32" min_val = float("3.01205e-05") @@ -5984,6 +6528,7 @@ class Program_weight_tensor_parameter_543: class Program_weight_tensor_parameter_544: name = "parameter_544" + original_name = "batch_norm2d_30.w_1" shape = [96] dtype = "float32" min_val = float("-0.0360867") @@ -5995,6 +6540,7 @@ class Program_weight_tensor_parameter_544: class Program_weight_tensor_parameter_545: name = "parameter_545" + original_name = "conv2d_31.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.0411901") @@ -6006,6 +6552,7 @@ class Program_weight_tensor_parameter_545: class Program_weight_tensor_parameter_546: name = "parameter_546" + original_name = "batch_norm2d_29.b_0" shape = [96] dtype = "float32" min_val = float("-1.24978") @@ -6017,6 +6564,7 @@ class Program_weight_tensor_parameter_546: class Program_weight_tensor_parameter_547: name = "parameter_547" + original_name = "batch_norm2d_29.w_0" shape = [96] dtype = "float32" min_val = float("0.316771") @@ -6028,6 +6576,7 @@ class Program_weight_tensor_parameter_547: class Program_weight_tensor_parameter_548: name = "parameter_548" + original_name = "batch_norm2d_29.w_2" shape = [96] dtype = "float32" min_val = float("0.00151664") @@ -6039,6 +6588,7 @@ class Program_weight_tensor_parameter_548: class Program_weight_tensor_parameter_549: name = "parameter_549" + original_name = "batch_norm2d_29.w_1" shape = [96] dtype = "float32" min_val = float("-0.0689274") @@ -6050,6 +6600,7 @@ class Program_weight_tensor_parameter_549: class Program_weight_tensor_parameter_550: name = "parameter_550" + original_name = "conv2d_30.w_0" shape = [96, 96, 3, 3] dtype = "float32" min_val = float("-0.0705669") @@ -6061,6 +6612,7 @@ class Program_weight_tensor_parameter_550: class Program_weight_tensor_parameter_551: name = "parameter_551" + original_name = "batch_norm2d_28.b_0" shape = [96] dtype = "float32" min_val = float("-3.58212") @@ -6072,6 +6624,7 @@ class Program_weight_tensor_parameter_551: class Program_weight_tensor_parameter_552: name = "parameter_552" + original_name = "batch_norm2d_28.w_0" shape = [96] dtype = "float32" min_val = float("0.517852") @@ -6083,6 +6636,7 @@ class Program_weight_tensor_parameter_552: class Program_weight_tensor_parameter_553: name = "parameter_553" + original_name = "batch_norm2d_28.w_2" shape = [96] dtype = "float32" min_val = float("0.0152262") @@ -6094,6 +6648,7 @@ class Program_weight_tensor_parameter_553: class Program_weight_tensor_parameter_554: name = "parameter_554" + original_name = "batch_norm2d_28.w_1" shape = [96] dtype = "float32" min_val = float("-0.696373") @@ -6105,6 +6660,7 @@ class Program_weight_tensor_parameter_554: class Program_weight_tensor_parameter_555: name = "parameter_555" + original_name = "conv2d_29.w_0" shape = [96, 96, 3, 3] dtype = "float32" min_val = float("-0.0861719") @@ -6116,6 +6672,7 @@ class Program_weight_tensor_parameter_555: class Program_weight_tensor_parameter_556: name = "parameter_556" + original_name = "batch_norm2d_27.b_0" shape = [96] dtype = "float32" min_val = float("-0.891952") @@ -6127,6 +6684,7 @@ class Program_weight_tensor_parameter_556: class Program_weight_tensor_parameter_557: name = "parameter_557" + original_name = "batch_norm2d_27.w_0" shape = [96] dtype = "float32" min_val = float("0.0198505") @@ -6138,6 +6696,7 @@ class Program_weight_tensor_parameter_557: class Program_weight_tensor_parameter_558: name = "parameter_558" + original_name = "batch_norm2d_27.w_2" shape = [96] dtype = "float32" min_val = float("2.65875e-05") @@ -6149,6 +6708,7 @@ class Program_weight_tensor_parameter_558: class Program_weight_tensor_parameter_559: name = "parameter_559" + original_name = "batch_norm2d_27.w_1" shape = [96] dtype = "float32" min_val = float("-0.0225542") @@ -6160,6 +6720,7 @@ class Program_weight_tensor_parameter_559: class Program_weight_tensor_parameter_560: name = "parameter_560" + original_name = "conv2d_28.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.0515445") @@ -6171,6 +6732,7 @@ class Program_weight_tensor_parameter_560: class Program_weight_tensor_parameter_561: name = "parameter_561" + original_name = "batch_norm2d_26.b_0" shape = [96] dtype = "float32" min_val = float("-0.891952") @@ -6182,6 +6744,7 @@ class Program_weight_tensor_parameter_561: class Program_weight_tensor_parameter_562: name = "parameter_562" + original_name = "batch_norm2d_26.w_0" shape = [96] dtype = "float32" min_val = float("0.177251") @@ -6193,6 +6756,7 @@ class Program_weight_tensor_parameter_562: class Program_weight_tensor_parameter_563: name = "parameter_563" + original_name = "batch_norm2d_26.w_2" shape = [96] dtype = "float32" min_val = float("0.000954749") @@ -6204,6 +6768,7 @@ class Program_weight_tensor_parameter_563: class Program_weight_tensor_parameter_564: name = "parameter_564" + original_name = "batch_norm2d_26.w_1" shape = [96] dtype = "float32" min_val = float("-0.0437256") @@ -6215,6 +6780,7 @@ class Program_weight_tensor_parameter_564: class Program_weight_tensor_parameter_565: name = "parameter_565" + original_name = "conv2d_27.w_0" shape = [96, 96, 3, 3] dtype = "float32" min_val = float("-0.0615728") @@ -6226,6 +6792,7 @@ class Program_weight_tensor_parameter_565: class Program_weight_tensor_parameter_566: name = "parameter_566" + original_name = "batch_norm2d_25.b_0" shape = [96] dtype = "float32" min_val = float("-2.65693") @@ -6237,6 +6804,7 @@ class Program_weight_tensor_parameter_566: class Program_weight_tensor_parameter_567: name = "parameter_567" + original_name = "batch_norm2d_25.w_0" shape = [96] dtype = "float32" min_val = float("0.514903") @@ -6248,6 +6816,7 @@ class Program_weight_tensor_parameter_567: class Program_weight_tensor_parameter_568: name = "parameter_568" + original_name = "batch_norm2d_25.w_2" shape = [96] dtype = "float32" min_val = float("0.00916572") @@ -6259,6 +6828,7 @@ class Program_weight_tensor_parameter_568: class Program_weight_tensor_parameter_569: name = "parameter_569" + original_name = "batch_norm2d_25.w_1" shape = [96] dtype = "float32" min_val = float("-0.578484") @@ -6270,6 +6840,7 @@ class Program_weight_tensor_parameter_569: class Program_weight_tensor_parameter_570: name = "parameter_570" + original_name = "conv2d_26.w_0" shape = [96, 96, 3, 3] dtype = "float32" min_val = float("-0.0817411") @@ -6281,6 +6852,7 @@ class Program_weight_tensor_parameter_570: class Program_weight_tensor_parameter_571: name = "parameter_571" + original_name = "batch_norm2d_24.b_0" shape = [96] dtype = "float32" min_val = float("-0.977794") @@ -6292,6 +6864,7 @@ class Program_weight_tensor_parameter_571: class Program_weight_tensor_parameter_572: name = "parameter_572" + original_name = "batch_norm2d_24.w_0" shape = [96] dtype = "float32" min_val = float("0.0466995") @@ -6303,6 +6876,7 @@ class Program_weight_tensor_parameter_572: class Program_weight_tensor_parameter_573: name = "parameter_573" + original_name = "batch_norm2d_24.w_2" shape = [96] dtype = "float32" min_val = float("0.000139938") @@ -6314,6 +6888,7 @@ class Program_weight_tensor_parameter_573: class Program_weight_tensor_parameter_574: name = "parameter_574" + original_name = "batch_norm2d_24.w_1" shape = [96] dtype = "float32" min_val = float("-0.0343031") @@ -6325,6 +6900,7 @@ class Program_weight_tensor_parameter_574: class Program_weight_tensor_parameter_575: name = "parameter_575" + original_name = "conv2d_25.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.0659496") @@ -6336,6 +6912,7 @@ class Program_weight_tensor_parameter_575: class Program_weight_tensor_parameter_576: name = "parameter_576" + original_name = "batch_norm2d_23.b_0" shape = [96] dtype = "float32" min_val = float("-0.977794") @@ -6347,6 +6924,7 @@ class Program_weight_tensor_parameter_576: class Program_weight_tensor_parameter_577: name = "parameter_577" + original_name = "batch_norm2d_23.w_0" shape = [96] dtype = "float32" min_val = float("0.245048") @@ -6358,6 +6936,7 @@ class Program_weight_tensor_parameter_577: class Program_weight_tensor_parameter_578: name = "parameter_578" + original_name = "batch_norm2d_23.w_2" shape = [96] dtype = "float32" min_val = float("0.00285907") @@ -6369,6 +6948,7 @@ class Program_weight_tensor_parameter_578: class Program_weight_tensor_parameter_579: name = "parameter_579" + original_name = "batch_norm2d_23.w_1" shape = [96] dtype = "float32" min_val = float("-0.041536") @@ -6380,6 +6960,7 @@ class Program_weight_tensor_parameter_579: class Program_weight_tensor_parameter_580: name = "parameter_580" + original_name = "conv2d_24.w_0" shape = [96, 96, 3, 3] dtype = "float32" min_val = float("-0.0701934") @@ -6391,6 +6972,7 @@ class Program_weight_tensor_parameter_580: class Program_weight_tensor_parameter_581: name = "parameter_581" + original_name = "batch_norm2d_22.b_0" shape = [96] dtype = "float32" min_val = float("-3.46647") @@ -6402,6 +6984,7 @@ class Program_weight_tensor_parameter_581: class Program_weight_tensor_parameter_582: name = "parameter_582" + original_name = "batch_norm2d_22.w_0" shape = [96] dtype = "float32" min_val = float("0.683607") @@ -6413,6 +6996,7 @@ class Program_weight_tensor_parameter_582: class Program_weight_tensor_parameter_583: name = "parameter_583" + original_name = "batch_norm2d_22.w_2" shape = [96] dtype = "float32" min_val = float("0.00759054") @@ -6424,6 +7008,7 @@ class Program_weight_tensor_parameter_583: class Program_weight_tensor_parameter_584: name = "parameter_584" + original_name = "batch_norm2d_22.w_1" shape = [96] dtype = "float32" min_val = float("-0.403653") @@ -6435,6 +7020,7 @@ class Program_weight_tensor_parameter_584: class Program_weight_tensor_parameter_585: name = "parameter_585" + original_name = "conv2d_23.w_0" shape = [96, 96, 3, 3] dtype = "float32" min_val = float("-0.070539") @@ -6446,6 +7032,7 @@ class Program_weight_tensor_parameter_585: class Program_weight_tensor_parameter_586: name = "parameter_586" + original_name = "batch_norm2d_21.b_0" shape = [96] dtype = "float32" min_val = float("-0.625482") @@ -6457,6 +7044,7 @@ class Program_weight_tensor_parameter_586: class Program_weight_tensor_parameter_587: name = "parameter_587" + original_name = "batch_norm2d_21.w_0" shape = [96] dtype = "float32" min_val = float("0.0903974") @@ -6468,6 +7056,7 @@ class Program_weight_tensor_parameter_587: class Program_weight_tensor_parameter_588: name = "parameter_588" + original_name = "batch_norm2d_21.w_2" shape = [96] dtype = "float32" min_val = float("0.00029884") @@ -6479,6 +7068,7 @@ class Program_weight_tensor_parameter_588: class Program_weight_tensor_parameter_589: name = "parameter_589" + original_name = "batch_norm2d_21.w_1" shape = [96] dtype = "float32" min_val = float("-0.0358192") @@ -6490,6 +7080,7 @@ class Program_weight_tensor_parameter_589: class Program_weight_tensor_parameter_590: name = "parameter_590" + original_name = "conv2d_22.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.0909244") @@ -6501,6 +7092,7 @@ class Program_weight_tensor_parameter_590: class Program_weight_tensor_parameter_591: name = "parameter_591" + original_name = "batch_norm2d_20.b_0" shape = [96] dtype = "float32" min_val = float("-0.625482") @@ -6512,6 +7104,7 @@ class Program_weight_tensor_parameter_591: class Program_weight_tensor_parameter_592: name = "parameter_592" + original_name = "batch_norm2d_20.w_0" shape = [96] dtype = "float32" min_val = float("0.209018") @@ -6523,6 +7116,7 @@ class Program_weight_tensor_parameter_592: class Program_weight_tensor_parameter_593: name = "parameter_593" + original_name = "batch_norm2d_20.w_2" shape = [96] dtype = "float32" min_val = float("0.00483437") @@ -6534,6 +7128,7 @@ class Program_weight_tensor_parameter_593: class Program_weight_tensor_parameter_594: name = "parameter_594" + original_name = "batch_norm2d_20.w_1" shape = [96] dtype = "float32" min_val = float("-0.112834") @@ -6545,6 +7140,7 @@ class Program_weight_tensor_parameter_594: class Program_weight_tensor_parameter_595: name = "parameter_595" + original_name = "conv2d_21.w_0" shape = [96, 96, 3, 3] dtype = "float32" min_val = float("-0.0923533") @@ -6556,6 +7152,7 @@ class Program_weight_tensor_parameter_595: class Program_weight_tensor_parameter_596: name = "parameter_596" + original_name = "batch_norm2d_19.b_0" shape = [96] dtype = "float32" min_val = float("-2.41013") @@ -6567,6 +7164,7 @@ class Program_weight_tensor_parameter_596: class Program_weight_tensor_parameter_597: name = "parameter_597" + original_name = "batch_norm2d_19.w_0" shape = [96] dtype = "float32" min_val = float("0.856185") @@ -6578,6 +7176,7 @@ class Program_weight_tensor_parameter_597: class Program_weight_tensor_parameter_598: name = "parameter_598" + original_name = "batch_norm2d_19.w_2" shape = [96] dtype = "float32" min_val = float("0.00590531") @@ -6589,6 +7188,7 @@ class Program_weight_tensor_parameter_598: class Program_weight_tensor_parameter_599: name = "parameter_599" + original_name = "batch_norm2d_19.w_1" shape = [96] dtype = "float32" min_val = float("-0.468657") @@ -6600,6 +7200,7 @@ class Program_weight_tensor_parameter_599: class Program_weight_tensor_parameter_600: name = "parameter_600" + original_name = "conv2d_20.w_0" shape = [96, 96, 3, 3] dtype = "float32" min_val = float("-0.139957") @@ -6611,6 +7212,7 @@ class Program_weight_tensor_parameter_600: class Program_weight_tensor_parameter_601: name = "parameter_601" + original_name = "batch_norm2d_18.b_0" shape = [96] dtype = "float32" min_val = float("-3.16841") @@ -6622,6 +7224,7 @@ class Program_weight_tensor_parameter_601: class Program_weight_tensor_parameter_602: name = "parameter_602" + original_name = "batch_norm2d_18.w_0" shape = [96] dtype = "float32" min_val = float("0.217997") @@ -6633,6 +7236,7 @@ class Program_weight_tensor_parameter_602: class Program_weight_tensor_parameter_603: name = "parameter_603" + original_name = "batch_norm2d_18.w_2" shape = [96] dtype = "float32" min_val = float("0.00570217") @@ -6644,6 +7248,7 @@ class Program_weight_tensor_parameter_603: class Program_weight_tensor_parameter_604: name = "parameter_604" + original_name = "batch_norm2d_18.w_1" shape = [96] dtype = "float32" min_val = float("-0.222949") @@ -6655,6 +7260,7 @@ class Program_weight_tensor_parameter_604: class Program_weight_tensor_parameter_605: name = "parameter_605" + original_name = "conv2d_19.w_0" shape = [96, 192, 1, 1] dtype = "float32" min_val = float("-0.158868") @@ -6666,6 +7272,7 @@ class Program_weight_tensor_parameter_605: class Program_weight_tensor_parameter_606: name = "parameter_606" + original_name = "batch_norm2d_17.b_0" shape = [96] dtype = "float32" min_val = float("-4.92519") @@ -6677,6 +7284,7 @@ class Program_weight_tensor_parameter_606: class Program_weight_tensor_parameter_607: name = "parameter_607" + original_name = "batch_norm2d_17.w_0" shape = [96] dtype = "float32" min_val = float("0.407504") @@ -6688,6 +7296,7 @@ class Program_weight_tensor_parameter_607: class Program_weight_tensor_parameter_608: name = "parameter_608" + original_name = "batch_norm2d_17.w_2" shape = [96] dtype = "float32" min_val = float("0.00267938") @@ -6699,6 +7308,7 @@ class Program_weight_tensor_parameter_608: class Program_weight_tensor_parameter_609: name = "parameter_609" + original_name = "batch_norm2d_17.w_1" shape = [96] dtype = "float32" min_val = float("-0.114101") @@ -6710,6 +7320,7 @@ class Program_weight_tensor_parameter_609: class Program_weight_tensor_parameter_610: name = "parameter_610" + original_name = "conv2d_18.w_0" shape = [96, 192, 1, 1] dtype = "float32" min_val = float("-0.0863343") @@ -6721,6 +7332,7 @@ class Program_weight_tensor_parameter_610: class Program_weight_tensor_parameter_611: name = "parameter_611" + original_name = "batch_norm2d_16.b_0" shape = [192] dtype = "float32" min_val = float("-2.27303") @@ -6732,6 +7344,7 @@ class Program_weight_tensor_parameter_611: class Program_weight_tensor_parameter_612: name = "parameter_612" + original_name = "batch_norm2d_16.w_0" shape = [192] dtype = "float32" min_val = float("0.631744") @@ -6743,6 +7356,7 @@ class Program_weight_tensor_parameter_612: class Program_weight_tensor_parameter_613: name = "parameter_613" + original_name = "batch_norm2d_16.w_2" shape = [192] dtype = "float32" min_val = float("0.0066416") @@ -6754,6 +7368,7 @@ class Program_weight_tensor_parameter_613: class Program_weight_tensor_parameter_614: name = "parameter_614" + original_name = "batch_norm2d_16.w_1" shape = [192] dtype = "float32" min_val = float("-0.372282") @@ -6765,6 +7380,7 @@ class Program_weight_tensor_parameter_614: class Program_weight_tensor_parameter_615: name = "parameter_615" + original_name = "conv2d_17.w_0" shape = [192, 128, 3, 3] dtype = "float32" min_val = float("-0.0730784") @@ -6776,6 +7392,7 @@ class Program_weight_tensor_parameter_615: class Program_weight_tensor_parameter_616: name = "parameter_616" + original_name = "batch_norm2d_15.b_0" shape = [128] dtype = "float32" min_val = float("-2.81073") @@ -6787,6 +7404,7 @@ class Program_weight_tensor_parameter_616: class Program_weight_tensor_parameter_617: name = "parameter_617" + original_name = "batch_norm2d_15.w_0" shape = [128] dtype = "float32" min_val = float("0.305109") @@ -6798,6 +7416,7 @@ class Program_weight_tensor_parameter_617: class Program_weight_tensor_parameter_618: name = "parameter_618" + original_name = "batch_norm2d_15.w_2" shape = [128] dtype = "float32" min_val = float("0.000253069") @@ -6809,6 +7428,7 @@ class Program_weight_tensor_parameter_618: class Program_weight_tensor_parameter_619: name = "parameter_619" + original_name = "batch_norm2d_15.w_1" shape = [128] dtype = "float32" min_val = float("-0.239333") @@ -6820,6 +7440,7 @@ class Program_weight_tensor_parameter_619: class Program_weight_tensor_parameter_620: name = "parameter_620" + original_name = "conv2d_16.w_0" shape = [128, 96, 1, 1] dtype = "float32" min_val = float("-0.147037") @@ -6831,6 +7452,7 @@ class Program_weight_tensor_parameter_620: class Program_weight_tensor_parameter_621: name = "parameter_621" + original_name = "conv2d_15.b_0" shape = [96] dtype = "float32" min_val = float("-0.0165634") @@ -6842,6 +7464,7 @@ class Program_weight_tensor_parameter_621: class Program_weight_tensor_parameter_622: name = "parameter_622" + original_name = "conv2d_15.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.243907") @@ -6853,6 +7476,7 @@ class Program_weight_tensor_parameter_622: class Program_weight_tensor_parameter_623: name = "parameter_623" + original_name = "batch_norm2d_14.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -6862,6 +7486,7 @@ class Program_weight_tensor_parameter_623: class Program_weight_tensor_parameter_624: name = "parameter_624" + original_name = "batch_norm2d_14.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -6871,6 +7496,7 @@ class Program_weight_tensor_parameter_624: class Program_weight_tensor_parameter_625: name = "parameter_625" + original_name = "batch_norm2d_14.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -6880,6 +7506,7 @@ class Program_weight_tensor_parameter_625: class Program_weight_tensor_parameter_626: name = "parameter_626" + original_name = "batch_norm2d_14.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -6889,6 +7516,7 @@ class Program_weight_tensor_parameter_626: class Program_weight_tensor_parameter_627: name = "parameter_627" + original_name = "conv2d_14.w_0" shape = [48, 48, 1, 1] dtype = "float32" min_val = float("-0.0450246") @@ -6900,6 +7528,7 @@ class Program_weight_tensor_parameter_627: class Program_weight_tensor_parameter_628: name = "parameter_628" + original_name = "batch_norm2d_13.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -6909,6 +7538,7 @@ class Program_weight_tensor_parameter_628: class Program_weight_tensor_parameter_629: name = "parameter_629" + original_name = "batch_norm2d_13.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -6918,6 +7548,7 @@ class Program_weight_tensor_parameter_629: class Program_weight_tensor_parameter_630: name = "parameter_630" + original_name = "batch_norm2d_13.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -6927,6 +7558,7 @@ class Program_weight_tensor_parameter_630: class Program_weight_tensor_parameter_631: name = "parameter_631" + original_name = "batch_norm2d_13.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -6936,6 +7568,7 @@ class Program_weight_tensor_parameter_631: class Program_weight_tensor_parameter_632: name = "parameter_632" + original_name = "conv2d_13.w_0" shape = [48, 48, 3, 3] dtype = "float32" min_val = float("-0.0521573") @@ -6947,6 +7580,7 @@ class Program_weight_tensor_parameter_632: class Program_weight_tensor_parameter_633: name = "parameter_633" + original_name = "batch_norm2d_12.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -6956,6 +7590,7 @@ class Program_weight_tensor_parameter_633: class Program_weight_tensor_parameter_634: name = "parameter_634" + original_name = "batch_norm2d_12.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -6965,6 +7600,7 @@ class Program_weight_tensor_parameter_634: class Program_weight_tensor_parameter_635: name = "parameter_635" + original_name = "batch_norm2d_12.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -6974,6 +7610,7 @@ class Program_weight_tensor_parameter_635: class Program_weight_tensor_parameter_636: name = "parameter_636" + original_name = "batch_norm2d_12.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -6983,6 +7620,7 @@ class Program_weight_tensor_parameter_636: class Program_weight_tensor_parameter_637: name = "parameter_637" + original_name = "conv2d_12.w_0" shape = [48, 48, 3, 3] dtype = "float32" min_val = float("-0.0706573") @@ -6994,6 +7632,7 @@ class Program_weight_tensor_parameter_637: class Program_weight_tensor_parameter_638: name = "parameter_638" + original_name = "batch_norm2d_11.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -7003,6 +7642,7 @@ class Program_weight_tensor_parameter_638: class Program_weight_tensor_parameter_639: name = "parameter_639" + original_name = "batch_norm2d_11.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -7012,6 +7652,7 @@ class Program_weight_tensor_parameter_639: class Program_weight_tensor_parameter_640: name = "parameter_640" + original_name = "batch_norm2d_11.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -7021,6 +7662,7 @@ class Program_weight_tensor_parameter_640: class Program_weight_tensor_parameter_641: name = "parameter_641" + original_name = "batch_norm2d_11.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -7030,6 +7672,7 @@ class Program_weight_tensor_parameter_641: class Program_weight_tensor_parameter_642: name = "parameter_642" + original_name = "conv2d_11.w_0" shape = [48, 48, 1, 1] dtype = "float32" min_val = float("-0.0663249") @@ -7041,6 +7684,7 @@ class Program_weight_tensor_parameter_642: class Program_weight_tensor_parameter_643: name = "parameter_643" + original_name = "batch_norm2d_10.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -7050,6 +7694,7 @@ class Program_weight_tensor_parameter_643: class Program_weight_tensor_parameter_644: name = "parameter_644" + original_name = "batch_norm2d_10.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -7059,6 +7704,7 @@ class Program_weight_tensor_parameter_644: class Program_weight_tensor_parameter_645: name = "parameter_645" + original_name = "batch_norm2d_10.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -7068,6 +7714,7 @@ class Program_weight_tensor_parameter_645: class Program_weight_tensor_parameter_646: name = "parameter_646" + original_name = "batch_norm2d_10.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -7077,6 +7724,7 @@ class Program_weight_tensor_parameter_646: class Program_weight_tensor_parameter_647: name = "parameter_647" + original_name = "conv2d_10.w_0" shape = [48, 48, 3, 3] dtype = "float32" min_val = float("-0.0616683") @@ -7088,6 +7736,7 @@ class Program_weight_tensor_parameter_647: class Program_weight_tensor_parameter_648: name = "parameter_648" + original_name = "batch_norm2d_9.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -7097,6 +7746,7 @@ class Program_weight_tensor_parameter_648: class Program_weight_tensor_parameter_649: name = "parameter_649" + original_name = "batch_norm2d_9.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -7106,6 +7756,7 @@ class Program_weight_tensor_parameter_649: class Program_weight_tensor_parameter_650: name = "parameter_650" + original_name = "batch_norm2d_9.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -7115,6 +7766,7 @@ class Program_weight_tensor_parameter_650: class Program_weight_tensor_parameter_651: name = "parameter_651" + original_name = "batch_norm2d_9.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -7124,6 +7776,7 @@ class Program_weight_tensor_parameter_651: class Program_weight_tensor_parameter_652: name = "parameter_652" + original_name = "conv2d_9.w_0" shape = [48, 48, 3, 3] dtype = "float32" min_val = float("-0.0900923") @@ -7135,6 +7788,7 @@ class Program_weight_tensor_parameter_652: class Program_weight_tensor_parameter_653: name = "parameter_653" + original_name = "batch_norm2d_8.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -7144,6 +7798,7 @@ class Program_weight_tensor_parameter_653: class Program_weight_tensor_parameter_654: name = "parameter_654" + original_name = "batch_norm2d_8.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -7153,6 +7808,7 @@ class Program_weight_tensor_parameter_654: class Program_weight_tensor_parameter_655: name = "parameter_655" + original_name = "batch_norm2d_8.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -7162,6 +7818,7 @@ class Program_weight_tensor_parameter_655: class Program_weight_tensor_parameter_656: name = "parameter_656" + original_name = "batch_norm2d_8.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -7171,6 +7828,7 @@ class Program_weight_tensor_parameter_656: class Program_weight_tensor_parameter_657: name = "parameter_657" + original_name = "conv2d_8.w_0" shape = [48, 48, 1, 1] dtype = "float32" min_val = float("-0.079166") @@ -7182,6 +7840,7 @@ class Program_weight_tensor_parameter_657: class Program_weight_tensor_parameter_658: name = "parameter_658" + original_name = "batch_norm2d_7.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -7191,6 +7850,7 @@ class Program_weight_tensor_parameter_658: class Program_weight_tensor_parameter_659: name = "parameter_659" + original_name = "batch_norm2d_7.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -7200,6 +7860,7 @@ class Program_weight_tensor_parameter_659: class Program_weight_tensor_parameter_660: name = "parameter_660" + original_name = "batch_norm2d_7.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -7209,6 +7870,7 @@ class Program_weight_tensor_parameter_660: class Program_weight_tensor_parameter_661: name = "parameter_661" + original_name = "batch_norm2d_7.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -7218,6 +7880,7 @@ class Program_weight_tensor_parameter_661: class Program_weight_tensor_parameter_662: name = "parameter_662" + original_name = "conv2d_7.w_0" shape = [48, 48, 3, 3] dtype = "float32" min_val = float("-0.0574322") @@ -7229,6 +7892,7 @@ class Program_weight_tensor_parameter_662: class Program_weight_tensor_parameter_663: name = "parameter_663" + original_name = "batch_norm2d_6.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -7238,6 +7902,7 @@ class Program_weight_tensor_parameter_663: class Program_weight_tensor_parameter_664: name = "parameter_664" + original_name = "batch_norm2d_6.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -7247,6 +7912,7 @@ class Program_weight_tensor_parameter_664: class Program_weight_tensor_parameter_665: name = "parameter_665" + original_name = "batch_norm2d_6.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -7256,6 +7922,7 @@ class Program_weight_tensor_parameter_665: class Program_weight_tensor_parameter_666: name = "parameter_666" + original_name = "batch_norm2d_6.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -7265,6 +7932,7 @@ class Program_weight_tensor_parameter_666: class Program_weight_tensor_parameter_667: name = "parameter_667" + original_name = "conv2d_6.w_0" shape = [48, 48, 3, 3] dtype = "float32" min_val = float("-0.104177") @@ -7276,6 +7944,7 @@ class Program_weight_tensor_parameter_667: class Program_weight_tensor_parameter_668: name = "parameter_668" + original_name = "batch_norm2d_5.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -7285,6 +7954,7 @@ class Program_weight_tensor_parameter_668: class Program_weight_tensor_parameter_669: name = "parameter_669" + original_name = "batch_norm2d_5.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -7294,6 +7964,7 @@ class Program_weight_tensor_parameter_669: class Program_weight_tensor_parameter_670: name = "parameter_670" + original_name = "batch_norm2d_5.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -7303,6 +7974,7 @@ class Program_weight_tensor_parameter_670: class Program_weight_tensor_parameter_671: name = "parameter_671" + original_name = "batch_norm2d_5.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -7312,6 +7984,7 @@ class Program_weight_tensor_parameter_671: class Program_weight_tensor_parameter_672: name = "parameter_672" + original_name = "conv2d_5.w_0" shape = [48, 96, 1, 1] dtype = "float32" min_val = float("-0.136412") @@ -7323,6 +7996,7 @@ class Program_weight_tensor_parameter_672: class Program_weight_tensor_parameter_673: name = "parameter_673" + original_name = "batch_norm2d_4.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -7332,6 +8006,7 @@ class Program_weight_tensor_parameter_673: class Program_weight_tensor_parameter_674: name = "parameter_674" + original_name = "batch_norm2d_4.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -7341,6 +8016,7 @@ class Program_weight_tensor_parameter_674: class Program_weight_tensor_parameter_675: name = "parameter_675" + original_name = "batch_norm2d_4.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -7350,6 +8026,7 @@ class Program_weight_tensor_parameter_675: class Program_weight_tensor_parameter_676: name = "parameter_676" + original_name = "batch_norm2d_4.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -7359,6 +8036,7 @@ class Program_weight_tensor_parameter_676: class Program_weight_tensor_parameter_677: name = "parameter_677" + original_name = "conv2d_4.w_0" shape = [48, 96, 1, 1] dtype = "float32" min_val = float("-0.101645") @@ -7370,6 +8048,7 @@ class Program_weight_tensor_parameter_677: class Program_weight_tensor_parameter_678: name = "parameter_678" + original_name = "batch_norm2d_3.b_0" shape = [96] dtype = "float32" min_val = float("-3.41513") @@ -7381,6 +8060,7 @@ class Program_weight_tensor_parameter_678: class Program_weight_tensor_parameter_679: name = "parameter_679" + original_name = "batch_norm2d_3.w_0" shape = [96] dtype = "float32" min_val = float("0.872879") @@ -7392,6 +8072,7 @@ class Program_weight_tensor_parameter_679: class Program_weight_tensor_parameter_680: name = "parameter_680" + original_name = "batch_norm2d_3.w_2" shape = [96] dtype = "float32" min_val = float("0.36528") @@ -7403,6 +8084,7 @@ class Program_weight_tensor_parameter_680: class Program_weight_tensor_parameter_681: name = "parameter_681" + original_name = "batch_norm2d_3.w_1" shape = [96] dtype = "float32" min_val = float("-1.13612") @@ -7414,6 +8096,7 @@ class Program_weight_tensor_parameter_681: class Program_weight_tensor_parameter_682: name = "parameter_682" + original_name = "conv2d_3.w_0" shape = [96, 64, 3, 3] dtype = "float32" min_val = float("-0.112294") @@ -7425,6 +8108,7 @@ class Program_weight_tensor_parameter_682: class Program_weight_tensor_parameter_683: name = "parameter_683" + original_name = "batch_norm2d_2.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -7434,6 +8118,7 @@ class Program_weight_tensor_parameter_683: class Program_weight_tensor_parameter_684: name = "parameter_684" + original_name = "batch_norm2d_2.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -7443,6 +8128,7 @@ class Program_weight_tensor_parameter_684: class Program_weight_tensor_parameter_685: name = "parameter_685" + original_name = "batch_norm2d_2.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -7452,6 +8138,7 @@ class Program_weight_tensor_parameter_685: class Program_weight_tensor_parameter_686: name = "parameter_686" + original_name = "batch_norm2d_2.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -7461,6 +8148,7 @@ class Program_weight_tensor_parameter_686: class Program_weight_tensor_parameter_687: name = "parameter_687" + original_name = "conv2d_2.w_0" shape = [64, 32, 3, 3] dtype = "float32" min_val = float("-0.126613") @@ -7472,6 +8160,7 @@ class Program_weight_tensor_parameter_687: class Program_weight_tensor_parameter_688: name = "parameter_688" + original_name = "batch_norm2d_1.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -7481,6 +8170,7 @@ class Program_weight_tensor_parameter_688: class Program_weight_tensor_parameter_689: name = "parameter_689" + original_name = "batch_norm2d_1.w_0" shape = [32] dtype = "float32" min_val = float("0") @@ -7490,6 +8180,7 @@ class Program_weight_tensor_parameter_689: class Program_weight_tensor_parameter_690: name = "parameter_690" + original_name = "batch_norm2d_1.w_2" shape = [32] dtype = "float32" min_val = float("0") @@ -7499,6 +8190,7 @@ class Program_weight_tensor_parameter_690: class Program_weight_tensor_parameter_691: name = "parameter_691" + original_name = "batch_norm2d_1.w_1" shape = [32] dtype = "float32" min_val = float("0") @@ -7508,6 +8200,7 @@ class Program_weight_tensor_parameter_691: class Program_weight_tensor_parameter_692: name = "parameter_692" + original_name = "conv2d_1.w_0" shape = [32, 32, 3, 3] dtype = "float32" min_val = float("-0.253111") @@ -7519,6 +8212,7 @@ class Program_weight_tensor_parameter_692: class Program_weight_tensor_parameter_693: name = "parameter_693" + original_name = "batch_norm2d_0.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -7528,6 +8222,7 @@ class Program_weight_tensor_parameter_693: class Program_weight_tensor_parameter_694: name = "parameter_694" + original_name = "batch_norm2d_0.w_0" shape = [32] dtype = "float32" min_val = float("0") @@ -7537,6 +8232,7 @@ class Program_weight_tensor_parameter_694: class Program_weight_tensor_parameter_695: name = "parameter_695" + original_name = "batch_norm2d_0.w_2" shape = [32] dtype = "float32" min_val = float("0") @@ -7546,6 +8242,7 @@ class Program_weight_tensor_parameter_695: class Program_weight_tensor_parameter_696: name = "parameter_696" + original_name = "batch_norm2d_0.w_1" shape = [32] dtype = "float32" min_val = float("0") @@ -7555,6 +8252,7 @@ class Program_weight_tensor_parameter_696: class Program_weight_tensor_parameter_697: name = "parameter_697" + original_name = "conv2d_0.w_0" shape = [32, 3, 3, 3] dtype = "float32" min_val = float("-0.246483") diff --git a/paddle_samples/PaddleX/PP-YOLOE_plus-M/subgraph_1/input_meta.py b/paddle_samples/PaddleX/PP-YOLOE_plus-M/subgraph_1/input_meta.py index b703bdb55..3a282e48d 100644 --- a/paddle_samples/PaddleX/PP-YOLOE_plus-M/subgraph_1/input_meta.py +++ b/paddle_samples/PaddleX/PP-YOLOE_plus-M/subgraph_1/input_meta.py @@ -99,7 +99,6 @@ class Program_weight_tensor_data_12: original_name = "var_1160" shape = [8, 3, 704, 704] dtype = "float32" - min_val = float("0.0") max_val = float("1.0") mean = float("0.530551") std = float("0.181212") diff --git a/paddle_samples/PaddleX/PP-YOLOE_plus-S/subgraph_2/input_meta.py b/paddle_samples/PaddleX/PP-YOLOE_plus-S/subgraph_2/input_meta.py index 73bf4ff3d..4672bcae8 100644 --- a/paddle_samples/PaddleX/PP-YOLOE_plus-S/subgraph_2/input_meta.py +++ b/paddle_samples/PaddleX/PP-YOLOE_plus-S/subgraph_2/input_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_data_0: name = "data_0" + original_name = "param_12" shape = [1] dtype = "float32" data = [1.0] @@ -7,6 +8,7 @@ class Program_weight_tensor_data_0: class Program_weight_tensor_data_1: name = "data_1" + original_name = "param_13" shape = [1] dtype = "float32" data = [1.0] @@ -14,6 +16,7 @@ class Program_weight_tensor_data_1: class Program_weight_tensor_data_2: name = "data_2" + original_name = "param_14" shape = [1] dtype = "float32" data = [1.0] @@ -21,6 +24,7 @@ class Program_weight_tensor_data_2: class Program_weight_tensor_data_3: name = "data_3" + original_name = "param_15" shape = [1] dtype = "float32" data = [1.0] @@ -28,6 +32,7 @@ class Program_weight_tensor_data_3: class Program_weight_tensor_data_4: name = "data_4" + original_name = "param_16" shape = [1] dtype = "float32" data = [1.0] @@ -35,6 +40,7 @@ class Program_weight_tensor_data_4: class Program_weight_tensor_data_5: name = "data_5" + original_name = "param_17" shape = [1] dtype = "float32" data = [1.0] @@ -42,6 +48,7 @@ class Program_weight_tensor_data_5: class Program_weight_tensor_data_6: name = "data_6" + original_name = "var_938" shape = [8, 3, 608, 608] dtype = "float32" min_val = float("-0.114997") diff --git a/paddle_samples/PaddleX/PP-YOLOE_plus-S/subgraph_2/weight_meta.py b/paddle_samples/PaddleX/PP-YOLOE_plus-S/subgraph_2/weight_meta.py index 967a27eb7..d2a3a33aa 100644 --- a/paddle_samples/PaddleX/PP-YOLOE_plus-S/subgraph_2/weight_meta.py +++ b/paddle_samples/PaddleX/PP-YOLOE_plus-S/subgraph_2/weight_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_parameter_0: name = "parameter_0" + original_name = "batch_norm2d_71.b_0" shape = [384] dtype = "float32" min_val = float("-0.450706") @@ -11,6 +12,7 @@ class Program_weight_tensor_parameter_0: class Program_weight_tensor_parameter_1: name = "parameter_1" + original_name = "batch_norm2d_71.w_0" shape = [384] dtype = "float32" min_val = float("0.925537") @@ -22,6 +24,7 @@ class Program_weight_tensor_parameter_1: class Program_weight_tensor_parameter_2: name = "parameter_2" + original_name = "batch_norm2d_71.w_2" shape = [384] dtype = "float32" min_val = float("0.00224876") @@ -33,6 +36,7 @@ class Program_weight_tensor_parameter_2: class Program_weight_tensor_parameter_3: name = "parameter_3" + original_name = "batch_norm2d_71.w_1" shape = [384] dtype = "float32" min_val = float("-0.159796") @@ -44,6 +48,7 @@ class Program_weight_tensor_parameter_3: class Program_weight_tensor_parameter_4: name = "parameter_4" + original_name = "conv2d_75.w_0" shape = [384, 384, 1, 1] dtype = "float32" min_val = float("-0.0772811") @@ -55,6 +60,7 @@ class Program_weight_tensor_parameter_4: class Program_weight_tensor_parameter_5: name = "parameter_5" + original_name = "batch_norm2d_70.b_0" shape = [192] dtype = "float32" min_val = float("-0.276133") @@ -66,6 +72,7 @@ class Program_weight_tensor_parameter_5: class Program_weight_tensor_parameter_6: name = "parameter_6" + original_name = "batch_norm2d_70.w_0" shape = [192] dtype = "float32" min_val = float("0.877645") @@ -77,6 +84,7 @@ class Program_weight_tensor_parameter_6: class Program_weight_tensor_parameter_7: name = "parameter_7" + original_name = "batch_norm2d_70.w_2" shape = [192] dtype = "float32" min_val = float("0.00112957") @@ -88,6 +96,7 @@ class Program_weight_tensor_parameter_7: class Program_weight_tensor_parameter_8: name = "parameter_8" + original_name = "batch_norm2d_70.w_1" shape = [192] dtype = "float32" min_val = float("-0.0338997") @@ -99,6 +108,7 @@ class Program_weight_tensor_parameter_8: class Program_weight_tensor_parameter_9: name = "parameter_9" + original_name = "conv2d_74.w_0" shape = [192, 192, 1, 1] dtype = "float32" min_val = float("-0.044604") @@ -110,6 +120,7 @@ class Program_weight_tensor_parameter_9: class Program_weight_tensor_parameter_10: name = "parameter_10" + original_name = "batch_norm2d_69.b_0" shape = [192] dtype = "float32" min_val = float("-0.276133") @@ -121,6 +132,7 @@ class Program_weight_tensor_parameter_10: class Program_weight_tensor_parameter_11: name = "parameter_11" + original_name = "batch_norm2d_69.w_0" shape = [192] dtype = "float32" min_val = float("0.959652") @@ -132,6 +144,7 @@ class Program_weight_tensor_parameter_11: class Program_weight_tensor_parameter_12: name = "parameter_12" + original_name = "batch_norm2d_69.w_2" shape = [192] dtype = "float32" min_val = float("0.00208341") @@ -143,6 +156,7 @@ class Program_weight_tensor_parameter_12: class Program_weight_tensor_parameter_13: name = "parameter_13" + original_name = "batch_norm2d_69.w_1" shape = [192] dtype = "float32" min_val = float("-0.127776") @@ -154,6 +168,7 @@ class Program_weight_tensor_parameter_13: class Program_weight_tensor_parameter_14: name = "parameter_14" + original_name = "conv2d_73.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.0607241") @@ -165,6 +180,7 @@ class Program_weight_tensor_parameter_14: class Program_weight_tensor_parameter_15: name = "parameter_15" + original_name = "batch_norm2d_68.b_0" shape = [192] dtype = "float32" min_val = float("-0.358121") @@ -176,6 +192,7 @@ class Program_weight_tensor_parameter_15: class Program_weight_tensor_parameter_16: name = "parameter_16" + original_name = "batch_norm2d_68.w_0" shape = [192] dtype = "float32" min_val = float("0.871088") @@ -187,6 +204,7 @@ class Program_weight_tensor_parameter_16: class Program_weight_tensor_parameter_17: name = "parameter_17" + original_name = "batch_norm2d_68.w_2" shape = [192] dtype = "float32" min_val = float("0.00586839") @@ -198,6 +216,7 @@ class Program_weight_tensor_parameter_17: class Program_weight_tensor_parameter_18: name = "parameter_18" + original_name = "batch_norm2d_68.w_1" shape = [192] dtype = "float32" min_val = float("-0.145118") @@ -209,6 +228,7 @@ class Program_weight_tensor_parameter_18: class Program_weight_tensor_parameter_19: name = "parameter_19" + original_name = "conv2d_72.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.0614253") @@ -220,6 +240,7 @@ class Program_weight_tensor_parameter_19: class Program_weight_tensor_parameter_20: name = "parameter_20" + original_name = "batch_norm2d_67.b_0" shape = [192] dtype = "float32" min_val = float("-0.262068") @@ -231,6 +252,7 @@ class Program_weight_tensor_parameter_20: class Program_weight_tensor_parameter_21: name = "parameter_21" + original_name = "batch_norm2d_67.w_0" shape = [192] dtype = "float32" min_val = float("0.907651") @@ -242,6 +264,7 @@ class Program_weight_tensor_parameter_21: class Program_weight_tensor_parameter_22: name = "parameter_22" + original_name = "batch_norm2d_67.w_2" shape = [192] dtype = "float32" min_val = float("0.00343907") @@ -253,6 +276,7 @@ class Program_weight_tensor_parameter_22: class Program_weight_tensor_parameter_23: name = "parameter_23" + original_name = "batch_norm2d_67.w_1" shape = [192] dtype = "float32" min_val = float("-0.0930804") @@ -264,6 +288,7 @@ class Program_weight_tensor_parameter_23: class Program_weight_tensor_parameter_24: name = "parameter_24" + original_name = "conv2d_71.w_0" shape = [192, 576, 1, 1] dtype = "float32" min_val = float("-0.104555") @@ -275,6 +300,7 @@ class Program_weight_tensor_parameter_24: class Program_weight_tensor_parameter_25: name = "parameter_25" + original_name = "batch_norm2d_66.b_0" shape = [192] dtype = "float32" min_val = float("-0.108168") @@ -286,6 +312,7 @@ class Program_weight_tensor_parameter_25: class Program_weight_tensor_parameter_26: name = "parameter_26" + original_name = "batch_norm2d_66.w_0" shape = [192] dtype = "float32" min_val = float("0.829759") @@ -297,6 +324,7 @@ class Program_weight_tensor_parameter_26: class Program_weight_tensor_parameter_27: name = "parameter_27" + original_name = "batch_norm2d_66.w_2" shape = [192] dtype = "float32" min_val = float("0.00265792") @@ -308,6 +336,7 @@ class Program_weight_tensor_parameter_27: class Program_weight_tensor_parameter_28: name = "parameter_28" + original_name = "batch_norm2d_66.w_1" shape = [192] dtype = "float32" min_val = float("-0.0728807") @@ -319,6 +348,7 @@ class Program_weight_tensor_parameter_28: class Program_weight_tensor_parameter_29: name = "parameter_29" + original_name = "conv2d_70.w_0" shape = [192, 576, 1, 1] dtype = "float32" min_val = float("-0.0458267") @@ -330,6 +360,7 @@ class Program_weight_tensor_parameter_29: class Program_weight_tensor_parameter_30: name = "parameter_30" + original_name = "batch_norm2d_65.b_0" shape = [192] dtype = "float32" min_val = float("-0.182514") @@ -341,6 +372,7 @@ class Program_weight_tensor_parameter_30: class Program_weight_tensor_parameter_31: name = "parameter_31" + original_name = "batch_norm2d_65.w_0" shape = [192] dtype = "float32" min_val = float("0.818331") @@ -352,6 +384,7 @@ class Program_weight_tensor_parameter_31: class Program_weight_tensor_parameter_32: name = "parameter_32" + original_name = "batch_norm2d_65.w_2" shape = [192] dtype = "float32" min_val = float("0.00902086") @@ -363,6 +396,7 @@ class Program_weight_tensor_parameter_32: class Program_weight_tensor_parameter_33: name = "parameter_33" + original_name = "batch_norm2d_65.w_1" shape = [192] dtype = "float32" min_val = float("-0.291809") @@ -374,6 +408,7 @@ class Program_weight_tensor_parameter_33: class Program_weight_tensor_parameter_34: name = "parameter_34" + original_name = "conv2d_69.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.039074") @@ -385,6 +420,7 @@ class Program_weight_tensor_parameter_34: class Program_weight_tensor_parameter_35: name = "parameter_35" + original_name = "batch_norm2d_64.b_0" shape = [192] dtype = "float32" min_val = float("-0.424052") @@ -396,6 +432,7 @@ class Program_weight_tensor_parameter_35: class Program_weight_tensor_parameter_36: name = "parameter_36" + original_name = "batch_norm2d_64.w_0" shape = [192] dtype = "float32" min_val = float("0.684119") @@ -407,6 +444,7 @@ class Program_weight_tensor_parameter_36: class Program_weight_tensor_parameter_37: name = "parameter_37" + original_name = "batch_norm2d_64.w_2" shape = [192] dtype = "float32" min_val = float("0.003583") @@ -418,6 +456,7 @@ class Program_weight_tensor_parameter_37: class Program_weight_tensor_parameter_38: name = "parameter_38" + original_name = "batch_norm2d_64.w_1" shape = [192] dtype = "float32" min_val = float("-0.138562") @@ -429,6 +468,7 @@ class Program_weight_tensor_parameter_38: class Program_weight_tensor_parameter_39: name = "parameter_39" + original_name = "conv2d_68.w_0" shape = [192, 192, 1, 1] dtype = "float32" min_val = float("-0.12775") @@ -440,6 +480,7 @@ class Program_weight_tensor_parameter_39: class Program_weight_tensor_parameter_40: name = "parameter_40" + original_name = "batch_norm2d_63.b_0" shape = [96] dtype = "float32" min_val = float("-0.284378") @@ -451,6 +492,7 @@ class Program_weight_tensor_parameter_40: class Program_weight_tensor_parameter_41: name = "parameter_41" + original_name = "batch_norm2d_63.w_0" shape = [96] dtype = "float32" min_val = float("0.804918") @@ -462,6 +504,7 @@ class Program_weight_tensor_parameter_41: class Program_weight_tensor_parameter_42: name = "parameter_42" + original_name = "batch_norm2d_63.w_2" shape = [96] dtype = "float32" min_val = float("0.00191088") @@ -473,6 +516,7 @@ class Program_weight_tensor_parameter_42: class Program_weight_tensor_parameter_43: name = "parameter_43" + original_name = "batch_norm2d_63.w_1" shape = [96] dtype = "float32" min_val = float("-0.0277279") @@ -484,6 +528,7 @@ class Program_weight_tensor_parameter_43: class Program_weight_tensor_parameter_44: name = "parameter_44" + original_name = "conv2d_67.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.0693136") @@ -495,6 +540,7 @@ class Program_weight_tensor_parameter_44: class Program_weight_tensor_parameter_45: name = "parameter_45" + original_name = "batch_norm2d_62.b_0" shape = [96] dtype = "float32" min_val = float("-0.284378") @@ -506,6 +552,7 @@ class Program_weight_tensor_parameter_45: class Program_weight_tensor_parameter_46: name = "parameter_46" + original_name = "batch_norm2d_62.w_0" shape = [96] dtype = "float32" min_val = float("0.920731") @@ -517,6 +564,7 @@ class Program_weight_tensor_parameter_46: class Program_weight_tensor_parameter_47: name = "parameter_47" + original_name = "batch_norm2d_62.w_2" shape = [96] dtype = "float32" min_val = float("0.00338125") @@ -528,6 +576,7 @@ class Program_weight_tensor_parameter_47: class Program_weight_tensor_parameter_48: name = "parameter_48" + original_name = "batch_norm2d_62.w_1" shape = [96] dtype = "float32" min_val = float("-0.105409") @@ -539,6 +588,7 @@ class Program_weight_tensor_parameter_48: class Program_weight_tensor_parameter_49: name = "parameter_49" + original_name = "conv2d_66.w_0" shape = [96, 96, 3, 3] dtype = "float32" min_val = float("-0.0951475") @@ -550,6 +600,7 @@ class Program_weight_tensor_parameter_49: class Program_weight_tensor_parameter_50: name = "parameter_50" + original_name = "batch_norm2d_61.b_0" shape = [96] dtype = "float32" min_val = float("-0.582433") @@ -561,6 +612,7 @@ class Program_weight_tensor_parameter_50: class Program_weight_tensor_parameter_51: name = "parameter_51" + original_name = "batch_norm2d_61.w_0" shape = [96] dtype = "float32" min_val = float("0.755139") @@ -572,6 +624,7 @@ class Program_weight_tensor_parameter_51: class Program_weight_tensor_parameter_52: name = "parameter_52" + original_name = "batch_norm2d_61.w_2" shape = [96] dtype = "float32" min_val = float("0.0107474") @@ -583,6 +636,7 @@ class Program_weight_tensor_parameter_52: class Program_weight_tensor_parameter_53: name = "parameter_53" + original_name = "batch_norm2d_61.w_1" shape = [96] dtype = "float32" min_val = float("-0.0722031") @@ -594,6 +648,7 @@ class Program_weight_tensor_parameter_53: class Program_weight_tensor_parameter_54: name = "parameter_54" + original_name = "conv2d_65.w_0" shape = [96, 96, 3, 3] dtype = "float32" min_val = float("-0.0916632") @@ -605,6 +660,7 @@ class Program_weight_tensor_parameter_54: class Program_weight_tensor_parameter_55: name = "parameter_55" + original_name = "batch_norm2d_60.b_0" shape = [96] dtype = "float32" min_val = float("-0.479078") @@ -616,6 +672,7 @@ class Program_weight_tensor_parameter_55: class Program_weight_tensor_parameter_56: name = "parameter_56" + original_name = "batch_norm2d_60.w_0" shape = [96] dtype = "float32" min_val = float("0.849031") @@ -627,6 +684,7 @@ class Program_weight_tensor_parameter_56: class Program_weight_tensor_parameter_57: name = "parameter_57" + original_name = "batch_norm2d_60.w_2" shape = [96] dtype = "float32" min_val = float("0.00409541") @@ -638,6 +696,7 @@ class Program_weight_tensor_parameter_57: class Program_weight_tensor_parameter_58: name = "parameter_58" + original_name = "batch_norm2d_60.w_1" shape = [96] dtype = "float32" min_val = float("-0.10468") @@ -649,6 +708,7 @@ class Program_weight_tensor_parameter_58: class Program_weight_tensor_parameter_59: name = "parameter_59" + original_name = "conv2d_64.w_0" shape = [96, 288, 1, 1] dtype = "float32" min_val = float("-0.0855408") @@ -660,6 +720,7 @@ class Program_weight_tensor_parameter_59: class Program_weight_tensor_parameter_60: name = "parameter_60" + original_name = "batch_norm2d_59.b_0" shape = [96] dtype = "float32" min_val = float("-0.130759") @@ -671,6 +732,7 @@ class Program_weight_tensor_parameter_60: class Program_weight_tensor_parameter_61: name = "parameter_61" + original_name = "batch_norm2d_59.w_0" shape = [96] dtype = "float32" min_val = float("0.81062") @@ -682,6 +744,7 @@ class Program_weight_tensor_parameter_61: class Program_weight_tensor_parameter_62: name = "parameter_62" + original_name = "batch_norm2d_59.w_2" shape = [96] dtype = "float32" min_val = float("0.002652") @@ -693,6 +756,7 @@ class Program_weight_tensor_parameter_62: class Program_weight_tensor_parameter_63: name = "parameter_63" + original_name = "batch_norm2d_59.w_1" shape = [96] dtype = "float32" min_val = float("-0.0657455") @@ -704,6 +768,7 @@ class Program_weight_tensor_parameter_63: class Program_weight_tensor_parameter_64: name = "parameter_64" + original_name = "conv2d_63.w_0" shape = [96, 288, 1, 1] dtype = "float32" min_val = float("-0.0936364") @@ -715,6 +780,7 @@ class Program_weight_tensor_parameter_64: class Program_weight_tensor_parameter_65: name = "parameter_65" + original_name = "batch_norm2d_58.b_0" shape = [96] dtype = "float32" min_val = float("-0.269002") @@ -726,6 +792,7 @@ class Program_weight_tensor_parameter_65: class Program_weight_tensor_parameter_66: name = "parameter_66" + original_name = "batch_norm2d_58.w_0" shape = [96] dtype = "float32" min_val = float("0.71414") @@ -737,6 +804,7 @@ class Program_weight_tensor_parameter_66: class Program_weight_tensor_parameter_67: name = "parameter_67" + original_name = "batch_norm2d_58.w_2" shape = [96] dtype = "float32" min_val = float("0.00761574") @@ -748,6 +816,7 @@ class Program_weight_tensor_parameter_67: class Program_weight_tensor_parameter_68: name = "parameter_68" + original_name = "batch_norm2d_58.w_1" shape = [96] dtype = "float32" min_val = float("-0.451267") @@ -759,6 +828,7 @@ class Program_weight_tensor_parameter_68: class Program_weight_tensor_parameter_69: name = "parameter_69" + original_name = "conv2d_62.w_0" shape = [96, 96, 3, 3] dtype = "float32" min_val = float("-0.0502454") @@ -770,6 +840,7 @@ class Program_weight_tensor_parameter_69: class Program_weight_tensor_parameter_70: name = "parameter_70" + original_name = "batch_norm2d_57.b_0" shape = [96] dtype = "float32" min_val = float("-0.71049") @@ -781,6 +852,7 @@ class Program_weight_tensor_parameter_70: class Program_weight_tensor_parameter_71: name = "parameter_71" + original_name = "batch_norm2d_57.w_0" shape = [96] dtype = "float32" min_val = float("0.506139") @@ -792,6 +864,7 @@ class Program_weight_tensor_parameter_71: class Program_weight_tensor_parameter_72: name = "parameter_72" + original_name = "batch_norm2d_57.w_2" shape = [96] dtype = "float32" min_val = float("0.0028799") @@ -803,6 +876,7 @@ class Program_weight_tensor_parameter_72: class Program_weight_tensor_parameter_73: name = "parameter_73" + original_name = "batch_norm2d_57.w_1" shape = [96] dtype = "float32" min_val = float("-0.192297") @@ -814,6 +888,7 @@ class Program_weight_tensor_parameter_73: class Program_weight_tensor_parameter_74: name = "parameter_74" + original_name = "conv2d_61.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.162418") @@ -825,6 +900,7 @@ class Program_weight_tensor_parameter_74: class Program_weight_tensor_parameter_75: name = "parameter_75" + original_name = "batch_norm2d_56.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -834,6 +910,7 @@ class Program_weight_tensor_parameter_75: class Program_weight_tensor_parameter_76: name = "parameter_76" + original_name = "batch_norm2d_56.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -843,6 +920,7 @@ class Program_weight_tensor_parameter_76: class Program_weight_tensor_parameter_77: name = "parameter_77" + original_name = "batch_norm2d_56.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -852,6 +930,7 @@ class Program_weight_tensor_parameter_77: class Program_weight_tensor_parameter_78: name = "parameter_78" + original_name = "batch_norm2d_56.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -861,6 +940,7 @@ class Program_weight_tensor_parameter_78: class Program_weight_tensor_parameter_79: name = "parameter_79" + original_name = "conv2d_60.w_0" shape = [48, 48, 1, 1] dtype = "float32" min_val = float("-0.0654612") @@ -872,6 +952,7 @@ class Program_weight_tensor_parameter_79: class Program_weight_tensor_parameter_80: name = "parameter_80" + original_name = "batch_norm2d_55.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -881,6 +962,7 @@ class Program_weight_tensor_parameter_80: class Program_weight_tensor_parameter_81: name = "parameter_81" + original_name = "batch_norm2d_55.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -890,6 +972,7 @@ class Program_weight_tensor_parameter_81: class Program_weight_tensor_parameter_82: name = "parameter_82" + original_name = "batch_norm2d_55.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -899,6 +982,7 @@ class Program_weight_tensor_parameter_82: class Program_weight_tensor_parameter_83: name = "parameter_83" + original_name = "batch_norm2d_55.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -908,6 +992,7 @@ class Program_weight_tensor_parameter_83: class Program_weight_tensor_parameter_84: name = "parameter_84" + original_name = "conv2d_59.w_0" shape = [48, 48, 3, 3] dtype = "float32" min_val = float("-0.129902") @@ -919,6 +1004,7 @@ class Program_weight_tensor_parameter_84: class Program_weight_tensor_parameter_85: name = "parameter_85" + original_name = "batch_norm2d_54.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -928,6 +1014,7 @@ class Program_weight_tensor_parameter_85: class Program_weight_tensor_parameter_86: name = "parameter_86" + original_name = "batch_norm2d_54.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -937,6 +1024,7 @@ class Program_weight_tensor_parameter_86: class Program_weight_tensor_parameter_87: name = "parameter_87" + original_name = "batch_norm2d_54.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -946,6 +1034,7 @@ class Program_weight_tensor_parameter_87: class Program_weight_tensor_parameter_88: name = "parameter_88" + original_name = "batch_norm2d_54.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -955,6 +1044,7 @@ class Program_weight_tensor_parameter_88: class Program_weight_tensor_parameter_89: name = "parameter_89" + original_name = "conv2d_58.w_0" shape = [48, 48, 3, 3] dtype = "float32" min_val = float("-0.0916908") @@ -966,6 +1056,7 @@ class Program_weight_tensor_parameter_89: class Program_weight_tensor_parameter_90: name = "parameter_90" + original_name = "batch_norm2d_53.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -975,6 +1066,7 @@ class Program_weight_tensor_parameter_90: class Program_weight_tensor_parameter_91: name = "parameter_91" + original_name = "batch_norm2d_53.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -984,6 +1076,7 @@ class Program_weight_tensor_parameter_91: class Program_weight_tensor_parameter_92: name = "parameter_92" + original_name = "batch_norm2d_53.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -993,6 +1086,7 @@ class Program_weight_tensor_parameter_92: class Program_weight_tensor_parameter_93: name = "parameter_93" + original_name = "batch_norm2d_53.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -1002,6 +1096,7 @@ class Program_weight_tensor_parameter_93: class Program_weight_tensor_parameter_94: name = "parameter_94" + original_name = "conv2d_57.w_0" shape = [48, 224, 1, 1] dtype = "float32" min_val = float("-0.16761") @@ -1013,6 +1108,7 @@ class Program_weight_tensor_parameter_94: class Program_weight_tensor_parameter_95: name = "parameter_95" + original_name = "batch_norm2d_52.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -1022,6 +1118,7 @@ class Program_weight_tensor_parameter_95: class Program_weight_tensor_parameter_96: name = "parameter_96" + original_name = "batch_norm2d_52.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -1031,6 +1128,7 @@ class Program_weight_tensor_parameter_96: class Program_weight_tensor_parameter_97: name = "parameter_97" + original_name = "batch_norm2d_52.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -1040,6 +1138,7 @@ class Program_weight_tensor_parameter_97: class Program_weight_tensor_parameter_98: name = "parameter_98" + original_name = "batch_norm2d_52.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -1049,6 +1148,7 @@ class Program_weight_tensor_parameter_98: class Program_weight_tensor_parameter_99: name = "parameter_99" + original_name = "conv2d_56.w_0" shape = [48, 224, 1, 1] dtype = "float32" min_val = float("-0.0769409") @@ -1060,6 +1160,7 @@ class Program_weight_tensor_parameter_99: class Program_weight_tensor_parameter_100: name = "parameter_100" + original_name = "batch_norm2d_51.b_0" shape = [96] dtype = "float32" min_val = float("-0.266698") @@ -1071,6 +1172,7 @@ class Program_weight_tensor_parameter_100: class Program_weight_tensor_parameter_101: name = "parameter_101" + original_name = "batch_norm2d_51.w_0" shape = [96] dtype = "float32" min_val = float("0.569143") @@ -1082,6 +1184,7 @@ class Program_weight_tensor_parameter_101: class Program_weight_tensor_parameter_102: name = "parameter_102" + original_name = "batch_norm2d_51.w_2" shape = [96] dtype = "float32" min_val = float("0.00586312") @@ -1093,6 +1196,7 @@ class Program_weight_tensor_parameter_102: class Program_weight_tensor_parameter_103: name = "parameter_103" + original_name = "batch_norm2d_51.w_1" shape = [96] dtype = "float32" min_val = float("-0.176735") @@ -1104,6 +1208,7 @@ class Program_weight_tensor_parameter_103: class Program_weight_tensor_parameter_104: name = "parameter_104" + original_name = "conv2d_55.w_0" shape = [96, 192, 1, 1] dtype = "float32" min_val = float("-0.0852068") @@ -1115,6 +1220,7 @@ class Program_weight_tensor_parameter_104: class Program_weight_tensor_parameter_105: name = "parameter_105" + original_name = "batch_norm2d_50.b_0" shape = [192] dtype = "float32" min_val = float("-0.397637") @@ -1126,6 +1232,7 @@ class Program_weight_tensor_parameter_105: class Program_weight_tensor_parameter_106: name = "parameter_106" + original_name = "batch_norm2d_50.w_0" shape = [192] dtype = "float32" min_val = float("0.718852") @@ -1137,6 +1244,7 @@ class Program_weight_tensor_parameter_106: class Program_weight_tensor_parameter_107: name = "parameter_107" + original_name = "batch_norm2d_50.w_2" shape = [192] dtype = "float32" min_val = float("0.0065768") @@ -1148,6 +1256,7 @@ class Program_weight_tensor_parameter_107: class Program_weight_tensor_parameter_108: name = "parameter_108" + original_name = "batch_norm2d_50.w_1" shape = [192] dtype = "float32" min_val = float("-0.142548") @@ -1159,6 +1268,7 @@ class Program_weight_tensor_parameter_108: class Program_weight_tensor_parameter_109: name = "parameter_109" + original_name = "conv2d_54.w_0" shape = [192, 192, 1, 1] dtype = "float32" min_val = float("-0.100007") @@ -1170,6 +1280,7 @@ class Program_weight_tensor_parameter_109: class Program_weight_tensor_parameter_110: name = "parameter_110" + original_name = "batch_norm2d_49.b_0" shape = [96] dtype = "float32" min_val = float("-0.321616") @@ -1181,6 +1292,7 @@ class Program_weight_tensor_parameter_110: class Program_weight_tensor_parameter_111: name = "parameter_111" + original_name = "batch_norm2d_49.w_0" shape = [96] dtype = "float32" min_val = float("0.6847") @@ -1192,6 +1304,7 @@ class Program_weight_tensor_parameter_111: class Program_weight_tensor_parameter_112: name = "parameter_112" + original_name = "batch_norm2d_49.w_2" shape = [96] dtype = "float32" min_val = float("0.00334166") @@ -1203,6 +1316,7 @@ class Program_weight_tensor_parameter_112: class Program_weight_tensor_parameter_113: name = "parameter_113" + original_name = "batch_norm2d_49.w_1" shape = [96] dtype = "float32" min_val = float("-0.0465356") @@ -1214,6 +1328,7 @@ class Program_weight_tensor_parameter_113: class Program_weight_tensor_parameter_114: name = "parameter_114" + original_name = "conv2d_53.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.0383231") @@ -1225,6 +1340,7 @@ class Program_weight_tensor_parameter_114: class Program_weight_tensor_parameter_115: name = "parameter_115" + original_name = "batch_norm2d_48.b_0" shape = [96] dtype = "float32" min_val = float("-0.321616") @@ -1236,6 +1352,7 @@ class Program_weight_tensor_parameter_115: class Program_weight_tensor_parameter_116: name = "parameter_116" + original_name = "batch_norm2d_48.w_0" shape = [96] dtype = "float32" min_val = float("0.881977") @@ -1247,6 +1364,7 @@ class Program_weight_tensor_parameter_116: class Program_weight_tensor_parameter_117: name = "parameter_117" + original_name = "batch_norm2d_48.w_2" shape = [96] dtype = "float32" min_val = float("0.00770206") @@ -1258,6 +1376,7 @@ class Program_weight_tensor_parameter_117: class Program_weight_tensor_parameter_118: name = "parameter_118" + original_name = "batch_norm2d_48.w_1" shape = [96] dtype = "float32" min_val = float("-0.0887963") @@ -1269,6 +1388,7 @@ class Program_weight_tensor_parameter_118: class Program_weight_tensor_parameter_119: name = "parameter_119" + original_name = "conv2d_52.w_0" shape = [96, 96, 3, 3] dtype = "float32" min_val = float("-0.127461") @@ -1280,6 +1400,7 @@ class Program_weight_tensor_parameter_119: class Program_weight_tensor_parameter_120: name = "parameter_120" + original_name = "batch_norm2d_47.b_0" shape = [96] dtype = "float32" min_val = float("-0.672352") @@ -1291,6 +1412,7 @@ class Program_weight_tensor_parameter_120: class Program_weight_tensor_parameter_121: name = "parameter_121" + original_name = "batch_norm2d_47.w_0" shape = [96] dtype = "float32" min_val = float("0.641984") @@ -1302,6 +1424,7 @@ class Program_weight_tensor_parameter_121: class Program_weight_tensor_parameter_122: name = "parameter_122" + original_name = "batch_norm2d_47.w_2" shape = [96] dtype = "float32" min_val = float("0.0137156") @@ -1313,6 +1436,7 @@ class Program_weight_tensor_parameter_122: class Program_weight_tensor_parameter_123: name = "parameter_123" + original_name = "batch_norm2d_47.w_1" shape = [96] dtype = "float32" min_val = float("-0.0920378") @@ -1324,6 +1448,7 @@ class Program_weight_tensor_parameter_123: class Program_weight_tensor_parameter_124: name = "parameter_124" + original_name = "conv2d_51.w_0" shape = [96, 96, 3, 3] dtype = "float32" min_val = float("-0.160626") @@ -1335,6 +1460,7 @@ class Program_weight_tensor_parameter_124: class Program_weight_tensor_parameter_125: name = "parameter_125" + original_name = "batch_norm2d_46.b_0" shape = [96] dtype = "float32" min_val = float("-0.61895") @@ -1346,6 +1472,7 @@ class Program_weight_tensor_parameter_125: class Program_weight_tensor_parameter_126: name = "parameter_126" + original_name = "batch_norm2d_46.w_0" shape = [96] dtype = "float32" min_val = float("0.709539") @@ -1357,6 +1484,7 @@ class Program_weight_tensor_parameter_126: class Program_weight_tensor_parameter_127: name = "parameter_127" + original_name = "batch_norm2d_46.w_2" shape = [96] dtype = "float32" min_val = float("0.00553064") @@ -1368,6 +1496,7 @@ class Program_weight_tensor_parameter_127: class Program_weight_tensor_parameter_128: name = "parameter_128" + original_name = "batch_norm2d_46.w_1" shape = [96] dtype = "float32" min_val = float("-0.190973") @@ -1379,6 +1508,7 @@ class Program_weight_tensor_parameter_128: class Program_weight_tensor_parameter_129: name = "parameter_129" + original_name = "conv2d_50.w_0" shape = [96, 448, 1, 1] dtype = "float32" min_val = float("-0.175646") @@ -1390,6 +1520,7 @@ class Program_weight_tensor_parameter_129: class Program_weight_tensor_parameter_130: name = "parameter_130" + original_name = "batch_norm2d_45.b_0" shape = [96] dtype = "float32" min_val = float("-0.182659") @@ -1401,6 +1532,7 @@ class Program_weight_tensor_parameter_130: class Program_weight_tensor_parameter_131: name = "parameter_131" + original_name = "batch_norm2d_45.w_0" shape = [96] dtype = "float32" min_val = float("0.821013") @@ -1412,6 +1544,7 @@ class Program_weight_tensor_parameter_131: class Program_weight_tensor_parameter_132: name = "parameter_132" + original_name = "batch_norm2d_45.w_2" shape = [96] dtype = "float32" min_val = float("0.004441") @@ -1423,6 +1556,7 @@ class Program_weight_tensor_parameter_132: class Program_weight_tensor_parameter_133: name = "parameter_133" + original_name = "batch_norm2d_45.w_1" shape = [96] dtype = "float32" min_val = float("-0.103749") @@ -1434,6 +1568,7 @@ class Program_weight_tensor_parameter_133: class Program_weight_tensor_parameter_134: name = "parameter_134" + original_name = "conv2d_49.w_0" shape = [96, 448, 1, 1] dtype = "float32" min_val = float("-0.236464") @@ -1445,6 +1580,7 @@ class Program_weight_tensor_parameter_134: class Program_weight_tensor_parameter_135: name = "parameter_135" + original_name = "batch_norm2d_44.b_0" shape = [192] dtype = "float32" min_val = float("-0.420244") @@ -1456,6 +1592,7 @@ class Program_weight_tensor_parameter_135: class Program_weight_tensor_parameter_136: name = "parameter_136" + original_name = "batch_norm2d_44.w_0" shape = [192] dtype = "float32" min_val = float("0.679742") @@ -1467,6 +1604,7 @@ class Program_weight_tensor_parameter_136: class Program_weight_tensor_parameter_137: name = "parameter_137" + original_name = "batch_norm2d_44.w_2" shape = [192] dtype = "float32" min_val = float("0.00887022") @@ -1478,6 +1616,7 @@ class Program_weight_tensor_parameter_137: class Program_weight_tensor_parameter_138: name = "parameter_138" + original_name = "batch_norm2d_44.w_1" shape = [192] dtype = "float32" min_val = float("-0.0859951") @@ -1489,6 +1628,7 @@ class Program_weight_tensor_parameter_138: class Program_weight_tensor_parameter_139: name = "parameter_139" + original_name = "conv2d_48.w_0" shape = [192, 384, 1, 1] dtype = "float32" min_val = float("-0.0883702") @@ -1500,6 +1640,7 @@ class Program_weight_tensor_parameter_139: class Program_weight_tensor_parameter_140: name = "parameter_140" + original_name = "batch_norm2d_43.b_0" shape = [384] dtype = "float32" min_val = float("-0.335926") @@ -1511,6 +1652,7 @@ class Program_weight_tensor_parameter_140: class Program_weight_tensor_parameter_141: name = "parameter_141" + original_name = "batch_norm2d_43.w_0" shape = [384] dtype = "float32" min_val = float("0.863976") @@ -1522,6 +1664,7 @@ class Program_weight_tensor_parameter_141: class Program_weight_tensor_parameter_142: name = "parameter_142" + original_name = "batch_norm2d_43.w_2" shape = [384] dtype = "float32" min_val = float("0.00476312") @@ -1533,6 +1676,7 @@ class Program_weight_tensor_parameter_142: class Program_weight_tensor_parameter_143: name = "parameter_143" + original_name = "batch_norm2d_43.w_1" shape = [384] dtype = "float32" min_val = float("-0.0984785") @@ -1544,6 +1688,7 @@ class Program_weight_tensor_parameter_143: class Program_weight_tensor_parameter_144: name = "parameter_144" + original_name = "conv2d_47.w_0" shape = [384, 384, 1, 1] dtype = "float32" min_val = float("-0.0983331") @@ -1555,6 +1700,7 @@ class Program_weight_tensor_parameter_144: class Program_weight_tensor_parameter_145: name = "parameter_145" + original_name = "batch_norm2d_42.b_0" shape = [192] dtype = "float32" min_val = float("-0.328356") @@ -1566,6 +1712,7 @@ class Program_weight_tensor_parameter_145: class Program_weight_tensor_parameter_146: name = "parameter_146" + original_name = "batch_norm2d_42.w_0" shape = [192] dtype = "float32" min_val = float("0.852012") @@ -1577,6 +1724,7 @@ class Program_weight_tensor_parameter_146: class Program_weight_tensor_parameter_147: name = "parameter_147" + original_name = "batch_norm2d_42.w_2" shape = [192] dtype = "float32" min_val = float("0.0255484") @@ -1588,6 +1736,7 @@ class Program_weight_tensor_parameter_147: class Program_weight_tensor_parameter_148: name = "parameter_148" + original_name = "batch_norm2d_42.w_1" shape = [192] dtype = "float32" min_val = float("-0.970626") @@ -1599,6 +1748,7 @@ class Program_weight_tensor_parameter_148: class Program_weight_tensor_parameter_149: name = "parameter_149" + original_name = "conv2d_46.w_0" shape = [192, 768, 1, 1] dtype = "float32" min_val = float("-0.104303") @@ -1610,6 +1760,7 @@ class Program_weight_tensor_parameter_149: class Program_weight_tensor_parameter_150: name = "parameter_150" + original_name = "batch_norm2d_41.b_0" shape = [192] dtype = "float32" min_val = float("-0.0749801") @@ -1621,6 +1772,7 @@ class Program_weight_tensor_parameter_150: class Program_weight_tensor_parameter_151: name = "parameter_151" + original_name = "batch_norm2d_41.w_0" shape = [192] dtype = "float32" min_val = float("0.813562") @@ -1632,6 +1784,7 @@ class Program_weight_tensor_parameter_151: class Program_weight_tensor_parameter_152: name = "parameter_152" + original_name = "batch_norm2d_41.w_2" shape = [192] dtype = "float32" min_val = float("0.00691915") @@ -1643,6 +1796,7 @@ class Program_weight_tensor_parameter_152: class Program_weight_tensor_parameter_153: name = "parameter_153" + original_name = "batch_norm2d_41.w_1" shape = [192] dtype = "float32" min_val = float("-0.0720374") @@ -1654,6 +1808,7 @@ class Program_weight_tensor_parameter_153: class Program_weight_tensor_parameter_154: name = "parameter_154" + original_name = "conv2d_45.w_0" shape = [192, 192, 1, 1] dtype = "float32" min_val = float("-0.0294137") @@ -1665,6 +1820,7 @@ class Program_weight_tensor_parameter_154: class Program_weight_tensor_parameter_155: name = "parameter_155" + original_name = "batch_norm2d_40.b_0" shape = [192] dtype = "float32" min_val = float("-0.0749801") @@ -1676,6 +1832,7 @@ class Program_weight_tensor_parameter_155: class Program_weight_tensor_parameter_156: name = "parameter_156" + original_name = "batch_norm2d_40.w_0" shape = [192] dtype = "float32" min_val = float("0.9") @@ -1687,6 +1844,7 @@ class Program_weight_tensor_parameter_156: class Program_weight_tensor_parameter_157: name = "parameter_157" + original_name = "batch_norm2d_40.w_2" shape = [192] dtype = "float32" min_val = float("0.0172514") @@ -1698,6 +1856,7 @@ class Program_weight_tensor_parameter_157: class Program_weight_tensor_parameter_158: name = "parameter_158" + original_name = "batch_norm2d_40.w_1" shape = [192] dtype = "float32" min_val = float("-0.228306") @@ -1709,6 +1868,7 @@ class Program_weight_tensor_parameter_158: class Program_weight_tensor_parameter_159: name = "parameter_159" + original_name = "conv2d_44.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.0561363") @@ -1720,6 +1880,7 @@ class Program_weight_tensor_parameter_159: class Program_weight_tensor_parameter_160: name = "parameter_160" + original_name = "batch_norm2d_39.b_0" shape = [192] dtype = "float32" min_val = float("-0.270955") @@ -1731,6 +1892,7 @@ class Program_weight_tensor_parameter_160: class Program_weight_tensor_parameter_161: name = "parameter_161" + original_name = "batch_norm2d_39.w_0" shape = [192] dtype = "float32" min_val = float("0.866044") @@ -1742,6 +1904,7 @@ class Program_weight_tensor_parameter_161: class Program_weight_tensor_parameter_162: name = "parameter_162" + original_name = "batch_norm2d_39.w_2" shape = [192] dtype = "float32" min_val = float("0.0203242") @@ -1753,6 +1916,7 @@ class Program_weight_tensor_parameter_162: class Program_weight_tensor_parameter_163: name = "parameter_163" + original_name = "batch_norm2d_39.w_1" shape = [192] dtype = "float32" min_val = float("-0.39533") @@ -1764,6 +1928,7 @@ class Program_weight_tensor_parameter_163: class Program_weight_tensor_parameter_164: name = "parameter_164" + original_name = "conv2d_43.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.0922815") @@ -1775,6 +1940,7 @@ class Program_weight_tensor_parameter_164: class Program_weight_tensor_parameter_165: name = "parameter_165" + original_name = "batch_norm2d_38.b_0" shape = [192] dtype = "float32" min_val = float("-0.295338") @@ -1786,6 +1952,7 @@ class Program_weight_tensor_parameter_165: class Program_weight_tensor_parameter_166: name = "parameter_166" + original_name = "batch_norm2d_38.w_0" shape = [192] dtype = "float32" min_val = float("0.914548") @@ -1797,6 +1964,7 @@ class Program_weight_tensor_parameter_166: class Program_weight_tensor_parameter_167: name = "parameter_167" + original_name = "batch_norm2d_38.w_2" shape = [192] dtype = "float32" min_val = float("0.0032043") @@ -1808,6 +1976,7 @@ class Program_weight_tensor_parameter_167: class Program_weight_tensor_parameter_168: name = "parameter_168" + original_name = "batch_norm2d_38.w_1" shape = [192] dtype = "float32" min_val = float("-0.110933") @@ -1819,6 +1988,7 @@ class Program_weight_tensor_parameter_168: class Program_weight_tensor_parameter_169: name = "parameter_169" + original_name = "conv2d_42.w_0" shape = [192, 512, 1, 1] dtype = "float32" min_val = float("-0.105731") @@ -1830,6 +2000,7 @@ class Program_weight_tensor_parameter_169: class Program_weight_tensor_parameter_170: name = "parameter_170" + original_name = "batch_norm2d_37.b_0" shape = [192] dtype = "float32" min_val = float("-0.120666") @@ -1841,6 +2012,7 @@ class Program_weight_tensor_parameter_170: class Program_weight_tensor_parameter_171: name = "parameter_171" + original_name = "batch_norm2d_37.w_0" shape = [192] dtype = "float32" min_val = float("0.877685") @@ -1852,6 +2024,7 @@ class Program_weight_tensor_parameter_171: class Program_weight_tensor_parameter_172: name = "parameter_172" + original_name = "batch_norm2d_37.w_2" shape = [192] dtype = "float32" min_val = float("0.00222831") @@ -1863,6 +2036,7 @@ class Program_weight_tensor_parameter_172: class Program_weight_tensor_parameter_173: name = "parameter_173" + original_name = "batch_norm2d_37.w_1" shape = [192] dtype = "float32" min_val = float("-0.031455") @@ -1874,6 +2048,7 @@ class Program_weight_tensor_parameter_173: class Program_weight_tensor_parameter_174: name = "parameter_174" + original_name = "conv2d_41.w_0" shape = [192, 512, 1, 1] dtype = "float32" min_val = float("-0.0308802") @@ -1885,6 +2060,7 @@ class Program_weight_tensor_parameter_174: class Program_weight_tensor_parameter_175: name = "parameter_175" + original_name = "batch_norm2d_36.b_0" shape = [512] dtype = "float32" min_val = float("-4.76781") @@ -1896,6 +2072,7 @@ class Program_weight_tensor_parameter_175: class Program_weight_tensor_parameter_176: name = "parameter_176" + original_name = "batch_norm2d_36.w_0" shape = [512] dtype = "float32" min_val = float("2.03338") @@ -1907,6 +2084,7 @@ class Program_weight_tensor_parameter_176: class Program_weight_tensor_parameter_177: name = "parameter_177" + original_name = "batch_norm2d_36.w_2" shape = [512] dtype = "float32" min_val = float("0.010459") @@ -1918,6 +2096,7 @@ class Program_weight_tensor_parameter_177: class Program_weight_tensor_parameter_178: name = "parameter_178" + original_name = "batch_norm2d_36.w_1" shape = [512] dtype = "float32" min_val = float("-0.160667") @@ -1929,6 +2108,7 @@ class Program_weight_tensor_parameter_178: class Program_weight_tensor_parameter_179: name = "parameter_179" + original_name = "conv2d_40.w_0" shape = [512, 384, 1, 1] dtype = "float32" min_val = float("-0.101408") @@ -1940,6 +2120,7 @@ class Program_weight_tensor_parameter_179: class Program_weight_tensor_parameter_180: name = "parameter_180" + original_name = "conv2d_39.b_0" shape = [384] dtype = "float32" min_val = float("-0.027407") @@ -1951,6 +2132,7 @@ class Program_weight_tensor_parameter_180: class Program_weight_tensor_parameter_181: name = "parameter_181" + original_name = "conv2d_39.w_0" shape = [384, 384, 1, 1] dtype = "float32" min_val = float("-0.22661") @@ -1962,6 +2144,7 @@ class Program_weight_tensor_parameter_181: class Program_weight_tensor_parameter_182: name = "parameter_182" + original_name = "batch_norm2d_35.b_0" shape = [192] dtype = "float32" min_val = float("-2.40536") @@ -1973,6 +2156,7 @@ class Program_weight_tensor_parameter_182: class Program_weight_tensor_parameter_183: name = "parameter_183" + original_name = "batch_norm2d_35.w_0" shape = [192] dtype = "float32" min_val = float("0.146531") @@ -1984,6 +2168,7 @@ class Program_weight_tensor_parameter_183: class Program_weight_tensor_parameter_184: name = "parameter_184" + original_name = "batch_norm2d_35.w_2" shape = [192] dtype = "float32" min_val = float("0.000219101") @@ -1995,6 +2180,7 @@ class Program_weight_tensor_parameter_184: class Program_weight_tensor_parameter_185: name = "parameter_185" + original_name = "batch_norm2d_35.w_1" shape = [192] dtype = "float32" min_val = float("-0.0361954") @@ -2006,6 +2192,7 @@ class Program_weight_tensor_parameter_185: class Program_weight_tensor_parameter_186: name = "parameter_186" + original_name = "conv2d_38.w_0" shape = [192, 192, 1, 1] dtype = "float32" min_val = float("-0.0363316") @@ -2017,6 +2204,7 @@ class Program_weight_tensor_parameter_186: class Program_weight_tensor_parameter_187: name = "parameter_187" + original_name = "batch_norm2d_34.b_0" shape = [192] dtype = "float32" min_val = float("-2.40536") @@ -2028,6 +2216,7 @@ class Program_weight_tensor_parameter_187: class Program_weight_tensor_parameter_188: name = "parameter_188" + original_name = "batch_norm2d_34.w_0" shape = [192] dtype = "float32" min_val = float("0.645518") @@ -2039,6 +2228,7 @@ class Program_weight_tensor_parameter_188: class Program_weight_tensor_parameter_189: name = "parameter_189" + original_name = "batch_norm2d_34.w_2" shape = [192] dtype = "float32" min_val = float("0.00250137") @@ -2050,6 +2240,7 @@ class Program_weight_tensor_parameter_189: class Program_weight_tensor_parameter_190: name = "parameter_190" + original_name = "batch_norm2d_34.w_1" shape = [192] dtype = "float32" min_val = float("-0.207353") @@ -2061,6 +2252,7 @@ class Program_weight_tensor_parameter_190: class Program_weight_tensor_parameter_191: name = "parameter_191" + original_name = "conv2d_37.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.0742124") @@ -2072,6 +2264,7 @@ class Program_weight_tensor_parameter_191: class Program_weight_tensor_parameter_192: name = "parameter_192" + original_name = "batch_norm2d_33.b_0" shape = [192] dtype = "float32" min_val = float("-3.28286") @@ -2083,6 +2276,7 @@ class Program_weight_tensor_parameter_192: class Program_weight_tensor_parameter_193: name = "parameter_193" + original_name = "batch_norm2d_33.w_0" shape = [192] dtype = "float32" min_val = float("0.512509") @@ -2094,6 +2288,7 @@ class Program_weight_tensor_parameter_193: class Program_weight_tensor_parameter_194: name = "parameter_194" + original_name = "batch_norm2d_33.w_2" shape = [192] dtype = "float32" min_val = float("0.01008") @@ -2105,6 +2300,7 @@ class Program_weight_tensor_parameter_194: class Program_weight_tensor_parameter_195: name = "parameter_195" + original_name = "batch_norm2d_33.w_1" shape = [192] dtype = "float32" min_val = float("-0.855284") @@ -2116,6 +2312,7 @@ class Program_weight_tensor_parameter_195: class Program_weight_tensor_parameter_196: name = "parameter_196" + original_name = "conv2d_36.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.0638156") @@ -2127,6 +2324,7 @@ class Program_weight_tensor_parameter_196: class Program_weight_tensor_parameter_197: name = "parameter_197" + original_name = "batch_norm2d_32.b_0" shape = [192] dtype = "float32" min_val = float("-3.81222") @@ -2138,6 +2336,7 @@ class Program_weight_tensor_parameter_197: class Program_weight_tensor_parameter_198: name = "parameter_198" + original_name = "batch_norm2d_32.w_0" shape = [192] dtype = "float32" min_val = float("0.699217") @@ -2149,6 +2348,7 @@ class Program_weight_tensor_parameter_198: class Program_weight_tensor_parameter_199: name = "parameter_199" + original_name = "batch_norm2d_32.w_2" shape = [192] dtype = "float32" min_val = float("0.00405284") @@ -2160,6 +2360,7 @@ class Program_weight_tensor_parameter_199: class Program_weight_tensor_parameter_200: name = "parameter_200" + original_name = "batch_norm2d_32.w_1" shape = [192] dtype = "float32" min_val = float("-0.115638") @@ -2171,6 +2372,7 @@ class Program_weight_tensor_parameter_200: class Program_weight_tensor_parameter_201: name = "parameter_201" + original_name = "conv2d_35.w_0" shape = [192, 384, 1, 1] dtype = "float32" min_val = float("-0.101629") @@ -2182,6 +2384,7 @@ class Program_weight_tensor_parameter_201: class Program_weight_tensor_parameter_202: name = "parameter_202" + original_name = "batch_norm2d_31.b_0" shape = [192] dtype = "float32" min_val = float("-2.92399") @@ -2193,6 +2396,7 @@ class Program_weight_tensor_parameter_202: class Program_weight_tensor_parameter_203: name = "parameter_203" + original_name = "batch_norm2d_31.w_0" shape = [192] dtype = "float32" min_val = float("0.777794") @@ -2204,6 +2408,7 @@ class Program_weight_tensor_parameter_203: class Program_weight_tensor_parameter_204: name = "parameter_204" + original_name = "batch_norm2d_31.w_2" shape = [192] dtype = "float32" min_val = float("0.00152121") @@ -2215,6 +2420,7 @@ class Program_weight_tensor_parameter_204: class Program_weight_tensor_parameter_205: name = "parameter_205" + original_name = "batch_norm2d_31.w_1" shape = [192] dtype = "float32" min_val = float("-0.0557254") @@ -2226,6 +2432,7 @@ class Program_weight_tensor_parameter_205: class Program_weight_tensor_parameter_206: name = "parameter_206" + original_name = "conv2d_34.w_0" shape = [192, 384, 1, 1] dtype = "float32" min_val = float("-0.064095") @@ -2237,6 +2444,7 @@ class Program_weight_tensor_parameter_206: class Program_weight_tensor_parameter_207: name = "parameter_207" + original_name = "batch_norm2d_30.b_0" shape = [384] dtype = "float32" min_val = float("-2.76185") @@ -2248,6 +2456,7 @@ class Program_weight_tensor_parameter_207: class Program_weight_tensor_parameter_208: name = "parameter_208" + original_name = "batch_norm2d_30.w_0" shape = [384] dtype = "float32" min_val = float("0.436993") @@ -2259,6 +2468,7 @@ class Program_weight_tensor_parameter_208: class Program_weight_tensor_parameter_209: name = "parameter_209" + original_name = "batch_norm2d_30.w_2" shape = [384] dtype = "float32" min_val = float("0.00444153") @@ -2270,6 +2480,7 @@ class Program_weight_tensor_parameter_209: class Program_weight_tensor_parameter_210: name = "parameter_210" + original_name = "batch_norm2d_30.w_1" shape = [384] dtype = "float32" min_val = float("-0.470345") @@ -2281,6 +2492,7 @@ class Program_weight_tensor_parameter_210: class Program_weight_tensor_parameter_211: name = "parameter_211" + original_name = "conv2d_33.w_0" shape = [384, 256, 3, 3] dtype = "float32" min_val = float("-0.067804") @@ -2292,6 +2504,7 @@ class Program_weight_tensor_parameter_211: class Program_weight_tensor_parameter_212: name = "parameter_212" + original_name = "batch_norm2d_29.b_0" shape = [256] dtype = "float32" min_val = float("-2.80954") @@ -2303,6 +2516,7 @@ class Program_weight_tensor_parameter_212: class Program_weight_tensor_parameter_213: name = "parameter_213" + original_name = "batch_norm2d_29.w_0" shape = [256] dtype = "float32" min_val = float("0.389802") @@ -2314,6 +2528,7 @@ class Program_weight_tensor_parameter_213: class Program_weight_tensor_parameter_214: name = "parameter_214" + original_name = "batch_norm2d_29.w_2" shape = [256] dtype = "float32" min_val = float("0.000375443") @@ -2325,6 +2540,7 @@ class Program_weight_tensor_parameter_214: class Program_weight_tensor_parameter_215: name = "parameter_215" + original_name = "batch_norm2d_29.w_1" shape = [256] dtype = "float32" min_val = float("-0.156638") @@ -2336,6 +2552,7 @@ class Program_weight_tensor_parameter_215: class Program_weight_tensor_parameter_216: name = "parameter_216" + original_name = "conv2d_32.w_0" shape = [256, 192, 1, 1] dtype = "float32" min_val = float("-0.272875") @@ -2347,6 +2564,7 @@ class Program_weight_tensor_parameter_216: class Program_weight_tensor_parameter_217: name = "parameter_217" + original_name = "conv2d_31.b_0" shape = [192] dtype = "float32" min_val = float("-0.0155376") @@ -2358,6 +2576,7 @@ class Program_weight_tensor_parameter_217: class Program_weight_tensor_parameter_218: name = "parameter_218" + original_name = "conv2d_31.w_0" shape = [192, 192, 1, 1] dtype = "float32" min_val = float("-0.184883") @@ -2369,6 +2588,7 @@ class Program_weight_tensor_parameter_218: class Program_weight_tensor_parameter_219: name = "parameter_219" + original_name = "batch_norm2d_28.b_0" shape = [96] dtype = "float32" min_val = float("-2.29568") @@ -2380,6 +2600,7 @@ class Program_weight_tensor_parameter_219: class Program_weight_tensor_parameter_220: name = "parameter_220" + original_name = "batch_norm2d_28.w_0" shape = [96] dtype = "float32" min_val = float("-0.11529") @@ -2391,6 +2612,7 @@ class Program_weight_tensor_parameter_220: class Program_weight_tensor_parameter_221: name = "parameter_221" + original_name = "batch_norm2d_28.w_2" shape = [96] dtype = "float32" min_val = float("1.17632e-10") @@ -2402,6 +2624,7 @@ class Program_weight_tensor_parameter_221: class Program_weight_tensor_parameter_222: name = "parameter_222" + original_name = "batch_norm2d_28.w_1" shape = [96] dtype = "float32" min_val = float("-0.045674") @@ -2413,6 +2636,7 @@ class Program_weight_tensor_parameter_222: class Program_weight_tensor_parameter_223: name = "parameter_223" + original_name = "conv2d_30.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.0465458") @@ -2424,6 +2648,7 @@ class Program_weight_tensor_parameter_223: class Program_weight_tensor_parameter_224: name = "parameter_224" + original_name = "batch_norm2d_27.b_0" shape = [96] dtype = "float32" min_val = float("-2.29568") @@ -2435,6 +2660,7 @@ class Program_weight_tensor_parameter_224: class Program_weight_tensor_parameter_225: name = "parameter_225" + original_name = "batch_norm2d_27.w_0" shape = [96] dtype = "float32" min_val = float("0.472161") @@ -2446,6 +2672,7 @@ class Program_weight_tensor_parameter_225: class Program_weight_tensor_parameter_226: name = "parameter_226" + original_name = "batch_norm2d_27.w_2" shape = [96] dtype = "float32" min_val = float("0.00195915") @@ -2457,6 +2684,7 @@ class Program_weight_tensor_parameter_226: class Program_weight_tensor_parameter_227: name = "parameter_227" + original_name = "batch_norm2d_27.w_1" shape = [96] dtype = "float32" min_val = float("-0.14497") @@ -2468,6 +2696,7 @@ class Program_weight_tensor_parameter_227: class Program_weight_tensor_parameter_228: name = "parameter_228" + original_name = "conv2d_29.w_0" shape = [96, 96, 3, 3] dtype = "float32" min_val = float("-0.143351") @@ -2479,6 +2708,7 @@ class Program_weight_tensor_parameter_228: class Program_weight_tensor_parameter_229: name = "parameter_229" + original_name = "batch_norm2d_26.b_0" shape = [96] dtype = "float32" min_val = float("-2.80238") @@ -2490,6 +2720,7 @@ class Program_weight_tensor_parameter_229: class Program_weight_tensor_parameter_230: name = "parameter_230" + original_name = "batch_norm2d_26.w_0" shape = [96] dtype = "float32" min_val = float("0.367938") @@ -2501,6 +2732,7 @@ class Program_weight_tensor_parameter_230: class Program_weight_tensor_parameter_231: name = "parameter_231" + original_name = "batch_norm2d_26.w_2" shape = [96] dtype = "float32" min_val = float("0.0152909") @@ -2512,6 +2744,7 @@ class Program_weight_tensor_parameter_231: class Program_weight_tensor_parameter_232: name = "parameter_232" + original_name = "batch_norm2d_26.w_1" shape = [96] dtype = "float32" min_val = float("-1.68343") @@ -2523,6 +2756,7 @@ class Program_weight_tensor_parameter_232: class Program_weight_tensor_parameter_233: name = "parameter_233" + original_name = "conv2d_28.w_0" shape = [96, 96, 3, 3] dtype = "float32" min_val = float("-0.0522225") @@ -2534,6 +2768,7 @@ class Program_weight_tensor_parameter_233: class Program_weight_tensor_parameter_234: name = "parameter_234" + original_name = "batch_norm2d_25.b_0" shape = [96] dtype = "float32" min_val = float("-2.54128") @@ -2545,6 +2780,7 @@ class Program_weight_tensor_parameter_234: class Program_weight_tensor_parameter_235: name = "parameter_235" + original_name = "batch_norm2d_25.w_0" shape = [96] dtype = "float32" min_val = float("-0.0971537") @@ -2556,6 +2792,7 @@ class Program_weight_tensor_parameter_235: class Program_weight_tensor_parameter_236: name = "parameter_236" + original_name = "batch_norm2d_25.w_2" shape = [96] dtype = "float32" min_val = float("0.000126891") @@ -2567,6 +2804,7 @@ class Program_weight_tensor_parameter_236: class Program_weight_tensor_parameter_237: name = "parameter_237" + original_name = "batch_norm2d_25.w_1" shape = [96] dtype = "float32" min_val = float("-0.0587587") @@ -2578,6 +2816,7 @@ class Program_weight_tensor_parameter_237: class Program_weight_tensor_parameter_238: name = "parameter_238" + original_name = "conv2d_27.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.119008") @@ -2589,6 +2828,7 @@ class Program_weight_tensor_parameter_238: class Program_weight_tensor_parameter_239: name = "parameter_239" + original_name = "batch_norm2d_24.b_0" shape = [96] dtype = "float32" min_val = float("-2.54128") @@ -2600,6 +2840,7 @@ class Program_weight_tensor_parameter_239: class Program_weight_tensor_parameter_240: name = "parameter_240" + original_name = "batch_norm2d_24.w_0" shape = [96] dtype = "float32" min_val = float("0.393347") @@ -2611,6 +2852,7 @@ class Program_weight_tensor_parameter_240: class Program_weight_tensor_parameter_241: name = "parameter_241" + original_name = "batch_norm2d_24.w_2" shape = [96] dtype = "float32" min_val = float("0.00618077") @@ -2622,6 +2864,7 @@ class Program_weight_tensor_parameter_241: class Program_weight_tensor_parameter_242: name = "parameter_242" + original_name = "batch_norm2d_24.w_1" shape = [96] dtype = "float32" min_val = float("-0.142604") @@ -2633,6 +2876,7 @@ class Program_weight_tensor_parameter_242: class Program_weight_tensor_parameter_243: name = "parameter_243" + original_name = "conv2d_26.w_0" shape = [96, 96, 3, 3] dtype = "float32" min_val = float("-0.0774473") @@ -2644,6 +2888,7 @@ class Program_weight_tensor_parameter_243: class Program_weight_tensor_parameter_244: name = "parameter_244" + original_name = "batch_norm2d_23.b_0" shape = [96] dtype = "float32" min_val = float("-2.062") @@ -2655,6 +2900,7 @@ class Program_weight_tensor_parameter_244: class Program_weight_tensor_parameter_245: name = "parameter_245" + original_name = "batch_norm2d_23.w_0" shape = [96] dtype = "float32" min_val = float("0.441068") @@ -2666,6 +2912,7 @@ class Program_weight_tensor_parameter_245: class Program_weight_tensor_parameter_246: name = "parameter_246" + original_name = "batch_norm2d_23.w_2" shape = [96] dtype = "float32" min_val = float("0.0050545") @@ -2677,6 +2924,7 @@ class Program_weight_tensor_parameter_246: class Program_weight_tensor_parameter_247: name = "parameter_247" + original_name = "batch_norm2d_23.w_1" shape = [96] dtype = "float32" min_val = float("-0.613634") @@ -2688,6 +2936,7 @@ class Program_weight_tensor_parameter_247: class Program_weight_tensor_parameter_248: name = "parameter_248" + original_name = "conv2d_25.w_0" shape = [96, 96, 3, 3] dtype = "float32" min_val = float("-0.135192") @@ -2699,6 +2948,7 @@ class Program_weight_tensor_parameter_248: class Program_weight_tensor_parameter_249: name = "parameter_249" + original_name = "batch_norm2d_22.b_0" shape = [96] dtype = "float32" min_val = float("-1.49598") @@ -2710,6 +2960,7 @@ class Program_weight_tensor_parameter_249: class Program_weight_tensor_parameter_250: name = "parameter_250" + original_name = "batch_norm2d_22.w_0" shape = [96] dtype = "float32" min_val = float("0.273964") @@ -2721,6 +2972,7 @@ class Program_weight_tensor_parameter_250: class Program_weight_tensor_parameter_251: name = "parameter_251" + original_name = "batch_norm2d_22.w_2" shape = [96] dtype = "float32" min_val = float("0.0068113") @@ -2732,6 +2984,7 @@ class Program_weight_tensor_parameter_251: class Program_weight_tensor_parameter_252: name = "parameter_252" + original_name = "batch_norm2d_22.w_1" shape = [96] dtype = "float32" min_val = float("-0.380598") @@ -2743,6 +2996,7 @@ class Program_weight_tensor_parameter_252: class Program_weight_tensor_parameter_253: name = "parameter_253" + original_name = "conv2d_24.w_0" shape = [96, 192, 1, 1] dtype = "float32" min_val = float("-0.148644") @@ -2754,6 +3008,7 @@ class Program_weight_tensor_parameter_253: class Program_weight_tensor_parameter_254: name = "parameter_254" + original_name = "batch_norm2d_21.b_0" shape = [96] dtype = "float32" min_val = float("-2.53532") @@ -2765,6 +3020,7 @@ class Program_weight_tensor_parameter_254: class Program_weight_tensor_parameter_255: name = "parameter_255" + original_name = "batch_norm2d_21.w_0" shape = [96] dtype = "float32" min_val = float("0.394136") @@ -2776,6 +3032,7 @@ class Program_weight_tensor_parameter_255: class Program_weight_tensor_parameter_256: name = "parameter_256" + original_name = "batch_norm2d_21.w_2" shape = [96] dtype = "float32" min_val = float("0.00403087") @@ -2787,6 +3044,7 @@ class Program_weight_tensor_parameter_256: class Program_weight_tensor_parameter_257: name = "parameter_257" + original_name = "batch_norm2d_21.w_1" shape = [96] dtype = "float32" min_val = float("-0.190402") @@ -2798,6 +3056,7 @@ class Program_weight_tensor_parameter_257: class Program_weight_tensor_parameter_258: name = "parameter_258" + original_name = "conv2d_23.w_0" shape = [96, 192, 1, 1] dtype = "float32" min_val = float("-0.0894104") @@ -2809,6 +3068,7 @@ class Program_weight_tensor_parameter_258: class Program_weight_tensor_parameter_259: name = "parameter_259" + original_name = "batch_norm2d_20.b_0" shape = [192] dtype = "float32" min_val = float("-4.57943") @@ -2820,6 +3080,7 @@ class Program_weight_tensor_parameter_259: class Program_weight_tensor_parameter_260: name = "parameter_260" + original_name = "batch_norm2d_20.w_0" shape = [192] dtype = "float32" min_val = float("0.532264") @@ -2831,6 +3092,7 @@ class Program_weight_tensor_parameter_260: class Program_weight_tensor_parameter_261: name = "parameter_261" + original_name = "batch_norm2d_20.w_2" shape = [192] dtype = "float32" min_val = float("0.00481604") @@ -2842,6 +3104,7 @@ class Program_weight_tensor_parameter_261: class Program_weight_tensor_parameter_262: name = "parameter_262" + original_name = "batch_norm2d_20.w_1" shape = [192] dtype = "float32" min_val = float("-0.274056") @@ -2853,6 +3116,7 @@ class Program_weight_tensor_parameter_262: class Program_weight_tensor_parameter_263: name = "parameter_263" + original_name = "conv2d_22.w_0" shape = [192, 128, 3, 3] dtype = "float32" min_val = float("-0.0984626") @@ -2864,6 +3128,7 @@ class Program_weight_tensor_parameter_263: class Program_weight_tensor_parameter_264: name = "parameter_264" + original_name = "batch_norm2d_19.b_0" shape = [128] dtype = "float32" min_val = float("-2.15976") @@ -2875,6 +3140,7 @@ class Program_weight_tensor_parameter_264: class Program_weight_tensor_parameter_265: name = "parameter_265" + original_name = "batch_norm2d_19.w_0" shape = [128] dtype = "float32" min_val = float("0.31086") @@ -2886,6 +3152,7 @@ class Program_weight_tensor_parameter_265: class Program_weight_tensor_parameter_266: name = "parameter_266" + original_name = "batch_norm2d_19.w_2" shape = [128] dtype = "float32" min_val = float("0.000662309") @@ -2897,6 +3164,7 @@ class Program_weight_tensor_parameter_266: class Program_weight_tensor_parameter_267: name = "parameter_267" + original_name = "batch_norm2d_19.w_1" shape = [128] dtype = "float32" min_val = float("-0.276074") @@ -2908,6 +3176,7 @@ class Program_weight_tensor_parameter_267: class Program_weight_tensor_parameter_268: name = "parameter_268" + original_name = "conv2d_21.w_0" shape = [128, 96, 1, 1] dtype = "float32" min_val = float("-0.175655") @@ -2919,6 +3188,7 @@ class Program_weight_tensor_parameter_268: class Program_weight_tensor_parameter_269: name = "parameter_269" + original_name = "conv2d_20.b_0" shape = [96] dtype = "float32" min_val = float("-0.0202559") @@ -2930,6 +3200,7 @@ class Program_weight_tensor_parameter_269: class Program_weight_tensor_parameter_270: name = "parameter_270" + original_name = "conv2d_20.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.286041") @@ -2941,6 +3212,7 @@ class Program_weight_tensor_parameter_270: class Program_weight_tensor_parameter_271: name = "parameter_271" + original_name = "batch_norm2d_18.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -2950,6 +3222,7 @@ class Program_weight_tensor_parameter_271: class Program_weight_tensor_parameter_272: name = "parameter_272" + original_name = "batch_norm2d_18.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -2959,6 +3232,7 @@ class Program_weight_tensor_parameter_272: class Program_weight_tensor_parameter_273: name = "parameter_273" + original_name = "batch_norm2d_18.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -2968,6 +3242,7 @@ class Program_weight_tensor_parameter_273: class Program_weight_tensor_parameter_274: name = "parameter_274" + original_name = "batch_norm2d_18.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -2977,6 +3252,7 @@ class Program_weight_tensor_parameter_274: class Program_weight_tensor_parameter_275: name = "parameter_275" + original_name = "conv2d_19.w_0" shape = [48, 48, 1, 1] dtype = "float32" min_val = float("-0.0596608") @@ -2988,6 +3264,7 @@ class Program_weight_tensor_parameter_275: class Program_weight_tensor_parameter_276: name = "parameter_276" + original_name = "batch_norm2d_17.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -2997,6 +3274,7 @@ class Program_weight_tensor_parameter_276: class Program_weight_tensor_parameter_277: name = "parameter_277" + original_name = "batch_norm2d_17.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -3006,6 +3284,7 @@ class Program_weight_tensor_parameter_277: class Program_weight_tensor_parameter_278: name = "parameter_278" + original_name = "batch_norm2d_17.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -3015,6 +3294,7 @@ class Program_weight_tensor_parameter_278: class Program_weight_tensor_parameter_279: name = "parameter_279" + original_name = "batch_norm2d_17.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -3024,6 +3304,7 @@ class Program_weight_tensor_parameter_279: class Program_weight_tensor_parameter_280: name = "parameter_280" + original_name = "conv2d_18.w_0" shape = [48, 48, 3, 3] dtype = "float32" min_val = float("-0.0891862") @@ -3035,6 +3316,7 @@ class Program_weight_tensor_parameter_280: class Program_weight_tensor_parameter_281: name = "parameter_281" + original_name = "batch_norm2d_16.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -3044,6 +3326,7 @@ class Program_weight_tensor_parameter_281: class Program_weight_tensor_parameter_282: name = "parameter_282" + original_name = "batch_norm2d_16.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -3053,6 +3336,7 @@ class Program_weight_tensor_parameter_282: class Program_weight_tensor_parameter_283: name = "parameter_283" + original_name = "batch_norm2d_16.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -3062,6 +3346,7 @@ class Program_weight_tensor_parameter_283: class Program_weight_tensor_parameter_284: name = "parameter_284" + original_name = "batch_norm2d_16.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -3071,6 +3356,7 @@ class Program_weight_tensor_parameter_284: class Program_weight_tensor_parameter_285: name = "parameter_285" + original_name = "conv2d_17.w_0" shape = [48, 48, 3, 3] dtype = "float32" min_val = float("-0.0785488") @@ -3082,6 +3368,7 @@ class Program_weight_tensor_parameter_285: class Program_weight_tensor_parameter_286: name = "parameter_286" + original_name = "batch_norm2d_15.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -3091,6 +3378,7 @@ class Program_weight_tensor_parameter_286: class Program_weight_tensor_parameter_287: name = "parameter_287" + original_name = "batch_norm2d_15.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -3100,6 +3388,7 @@ class Program_weight_tensor_parameter_287: class Program_weight_tensor_parameter_288: name = "parameter_288" + original_name = "batch_norm2d_15.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -3109,6 +3398,7 @@ class Program_weight_tensor_parameter_288: class Program_weight_tensor_parameter_289: name = "parameter_289" + original_name = "batch_norm2d_15.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -3118,6 +3408,7 @@ class Program_weight_tensor_parameter_289: class Program_weight_tensor_parameter_290: name = "parameter_290" + original_name = "conv2d_16.w_0" shape = [48, 48, 1, 1] dtype = "float32" min_val = float("-0.0640984") @@ -3129,6 +3420,7 @@ class Program_weight_tensor_parameter_290: class Program_weight_tensor_parameter_291: name = "parameter_291" + original_name = "batch_norm2d_14.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -3138,6 +3430,7 @@ class Program_weight_tensor_parameter_291: class Program_weight_tensor_parameter_292: name = "parameter_292" + original_name = "batch_norm2d_14.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -3147,6 +3440,7 @@ class Program_weight_tensor_parameter_292: class Program_weight_tensor_parameter_293: name = "parameter_293" + original_name = "batch_norm2d_14.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -3156,6 +3450,7 @@ class Program_weight_tensor_parameter_293: class Program_weight_tensor_parameter_294: name = "parameter_294" + original_name = "batch_norm2d_14.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -3165,6 +3460,7 @@ class Program_weight_tensor_parameter_294: class Program_weight_tensor_parameter_295: name = "parameter_295" + original_name = "conv2d_15.w_0" shape = [48, 48, 3, 3] dtype = "float32" min_val = float("-0.0652072") @@ -3176,6 +3472,7 @@ class Program_weight_tensor_parameter_295: class Program_weight_tensor_parameter_296: name = "parameter_296" + original_name = "batch_norm2d_13.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -3185,6 +3482,7 @@ class Program_weight_tensor_parameter_296: class Program_weight_tensor_parameter_297: name = "parameter_297" + original_name = "batch_norm2d_13.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -3194,6 +3492,7 @@ class Program_weight_tensor_parameter_297: class Program_weight_tensor_parameter_298: name = "parameter_298" + original_name = "batch_norm2d_13.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -3203,6 +3502,7 @@ class Program_weight_tensor_parameter_298: class Program_weight_tensor_parameter_299: name = "parameter_299" + original_name = "batch_norm2d_13.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -3212,6 +3512,7 @@ class Program_weight_tensor_parameter_299: class Program_weight_tensor_parameter_300: name = "parameter_300" + original_name = "conv2d_14.w_0" shape = [48, 48, 3, 3] dtype = "float32" min_val = float("-0.123766") @@ -3223,6 +3524,7 @@ class Program_weight_tensor_parameter_300: class Program_weight_tensor_parameter_301: name = "parameter_301" + original_name = "batch_norm2d_12.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -3232,6 +3534,7 @@ class Program_weight_tensor_parameter_301: class Program_weight_tensor_parameter_302: name = "parameter_302" + original_name = "batch_norm2d_12.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -3241,6 +3544,7 @@ class Program_weight_tensor_parameter_302: class Program_weight_tensor_parameter_303: name = "parameter_303" + original_name = "batch_norm2d_12.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -3250,6 +3554,7 @@ class Program_weight_tensor_parameter_303: class Program_weight_tensor_parameter_304: name = "parameter_304" + original_name = "batch_norm2d_12.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -3259,6 +3564,7 @@ class Program_weight_tensor_parameter_304: class Program_weight_tensor_parameter_305: name = "parameter_305" + original_name = "conv2d_13.w_0" shape = [48, 96, 1, 1] dtype = "float32" min_val = float("-0.128712") @@ -3270,6 +3576,7 @@ class Program_weight_tensor_parameter_305: class Program_weight_tensor_parameter_306: name = "parameter_306" + original_name = "batch_norm2d_11.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -3279,6 +3586,7 @@ class Program_weight_tensor_parameter_306: class Program_weight_tensor_parameter_307: name = "parameter_307" + original_name = "batch_norm2d_11.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -3288,6 +3596,7 @@ class Program_weight_tensor_parameter_307: class Program_weight_tensor_parameter_308: name = "parameter_308" + original_name = "batch_norm2d_11.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -3297,6 +3606,7 @@ class Program_weight_tensor_parameter_308: class Program_weight_tensor_parameter_309: name = "parameter_309" + original_name = "batch_norm2d_11.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -3306,6 +3616,7 @@ class Program_weight_tensor_parameter_309: class Program_weight_tensor_parameter_310: name = "parameter_310" + original_name = "conv2d_12.w_0" shape = [48, 96, 1, 1] dtype = "float32" min_val = float("-0.135295") @@ -3317,6 +3628,7 @@ class Program_weight_tensor_parameter_310: class Program_weight_tensor_parameter_311: name = "parameter_311" + original_name = "batch_norm2d_10.b_0" shape = [96] dtype = "float32" min_val = float("-3.45574") @@ -3328,6 +3640,7 @@ class Program_weight_tensor_parameter_311: class Program_weight_tensor_parameter_312: name = "parameter_312" + original_name = "batch_norm2d_10.w_0" shape = [96] dtype = "float32" min_val = float("0.533118") @@ -3339,6 +3652,7 @@ class Program_weight_tensor_parameter_312: class Program_weight_tensor_parameter_313: name = "parameter_313" + original_name = "batch_norm2d_10.w_2" shape = [96] dtype = "float32" min_val = float("0.0102217") @@ -3350,6 +3664,7 @@ class Program_weight_tensor_parameter_313: class Program_weight_tensor_parameter_314: name = "parameter_314" + original_name = "batch_norm2d_10.w_1" shape = [96] dtype = "float32" min_val = float("-0.590422") @@ -3361,6 +3676,7 @@ class Program_weight_tensor_parameter_314: class Program_weight_tensor_parameter_315: name = "parameter_315" + original_name = "conv2d_11.w_0" shape = [96, 64, 3, 3] dtype = "float32" min_val = float("-0.101228") @@ -3372,6 +3688,7 @@ class Program_weight_tensor_parameter_315: class Program_weight_tensor_parameter_316: name = "parameter_316" + original_name = "batch_norm2d_9.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -3381,6 +3698,7 @@ class Program_weight_tensor_parameter_316: class Program_weight_tensor_parameter_317: name = "parameter_317" + original_name = "batch_norm2d_9.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -3390,6 +3708,7 @@ class Program_weight_tensor_parameter_317: class Program_weight_tensor_parameter_318: name = "parameter_318" + original_name = "batch_norm2d_9.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -3399,6 +3718,7 @@ class Program_weight_tensor_parameter_318: class Program_weight_tensor_parameter_319: name = "parameter_319" + original_name = "batch_norm2d_9.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -3408,6 +3728,7 @@ class Program_weight_tensor_parameter_319: class Program_weight_tensor_parameter_320: name = "parameter_320" + original_name = "conv2d_10.w_0" shape = [64, 48, 1, 1] dtype = "float32" min_val = float("-0.180249") @@ -3419,6 +3740,7 @@ class Program_weight_tensor_parameter_320: class Program_weight_tensor_parameter_321: name = "parameter_321" + original_name = "conv2d_9.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -3428,6 +3750,7 @@ class Program_weight_tensor_parameter_321: class Program_weight_tensor_parameter_322: name = "parameter_322" + original_name = "conv2d_9.w_0" shape = [48, 48, 1, 1] dtype = "float32" min_val = float("-0.166837") @@ -3439,6 +3762,7 @@ class Program_weight_tensor_parameter_322: class Program_weight_tensor_parameter_323: name = "parameter_323" + original_name = "batch_norm2d_8.b_0" shape = [24] dtype = "float32" min_val = float("0") @@ -3448,6 +3772,7 @@ class Program_weight_tensor_parameter_323: class Program_weight_tensor_parameter_324: name = "parameter_324" + original_name = "batch_norm2d_8.w_0" shape = [24] dtype = "float32" min_val = float("0") @@ -3457,6 +3782,7 @@ class Program_weight_tensor_parameter_324: class Program_weight_tensor_parameter_325: name = "parameter_325" + original_name = "batch_norm2d_8.w_2" shape = [24] dtype = "float32" min_val = float("0") @@ -3466,6 +3792,7 @@ class Program_weight_tensor_parameter_325: class Program_weight_tensor_parameter_326: name = "parameter_326" + original_name = "batch_norm2d_8.w_1" shape = [24] dtype = "float32" min_val = float("0") @@ -3475,6 +3802,7 @@ class Program_weight_tensor_parameter_326: class Program_weight_tensor_parameter_327: name = "parameter_327" + original_name = "conv2d_8.w_0" shape = [24, 24, 1, 1] dtype = "float32" min_val = float("-0.0974583") @@ -3486,6 +3814,7 @@ class Program_weight_tensor_parameter_327: class Program_weight_tensor_parameter_328: name = "parameter_328" + original_name = "batch_norm2d_7.b_0" shape = [24] dtype = "float32" min_val = float("0") @@ -3495,6 +3824,7 @@ class Program_weight_tensor_parameter_328: class Program_weight_tensor_parameter_329: name = "parameter_329" + original_name = "batch_norm2d_7.w_0" shape = [24] dtype = "float32" min_val = float("0") @@ -3504,6 +3834,7 @@ class Program_weight_tensor_parameter_329: class Program_weight_tensor_parameter_330: name = "parameter_330" + original_name = "batch_norm2d_7.w_2" shape = [24] dtype = "float32" min_val = float("0") @@ -3513,6 +3844,7 @@ class Program_weight_tensor_parameter_330: class Program_weight_tensor_parameter_331: name = "parameter_331" + original_name = "batch_norm2d_7.w_1" shape = [24] dtype = "float32" min_val = float("0") @@ -3522,6 +3854,7 @@ class Program_weight_tensor_parameter_331: class Program_weight_tensor_parameter_332: name = "parameter_332" + original_name = "conv2d_7.w_0" shape = [24, 24, 3, 3] dtype = "float32" min_val = float("-0.131819") @@ -3533,6 +3866,7 @@ class Program_weight_tensor_parameter_332: class Program_weight_tensor_parameter_333: name = "parameter_333" + original_name = "batch_norm2d_6.b_0" shape = [24] dtype = "float32" min_val = float("0") @@ -3542,6 +3876,7 @@ class Program_weight_tensor_parameter_333: class Program_weight_tensor_parameter_334: name = "parameter_334" + original_name = "batch_norm2d_6.w_0" shape = [24] dtype = "float32" min_val = float("0") @@ -3551,6 +3886,7 @@ class Program_weight_tensor_parameter_334: class Program_weight_tensor_parameter_335: name = "parameter_335" + original_name = "batch_norm2d_6.w_2" shape = [24] dtype = "float32" min_val = float("0") @@ -3560,6 +3896,7 @@ class Program_weight_tensor_parameter_335: class Program_weight_tensor_parameter_336: name = "parameter_336" + original_name = "batch_norm2d_6.w_1" shape = [24] dtype = "float32" min_val = float("0") @@ -3569,6 +3906,7 @@ class Program_weight_tensor_parameter_336: class Program_weight_tensor_parameter_337: name = "parameter_337" + original_name = "conv2d_6.w_0" shape = [24, 24, 3, 3] dtype = "float32" min_val = float("-0.158151") @@ -3580,6 +3918,7 @@ class Program_weight_tensor_parameter_337: class Program_weight_tensor_parameter_338: name = "parameter_338" + original_name = "batch_norm2d_5.b_0" shape = [24] dtype = "float32" min_val = float("0") @@ -3589,6 +3928,7 @@ class Program_weight_tensor_parameter_338: class Program_weight_tensor_parameter_339: name = "parameter_339" + original_name = "batch_norm2d_5.w_0" shape = [24] dtype = "float32" min_val = float("0") @@ -3598,6 +3938,7 @@ class Program_weight_tensor_parameter_339: class Program_weight_tensor_parameter_340: name = "parameter_340" + original_name = "batch_norm2d_5.w_2" shape = [24] dtype = "float32" min_val = float("0") @@ -3607,6 +3948,7 @@ class Program_weight_tensor_parameter_340: class Program_weight_tensor_parameter_341: name = "parameter_341" + original_name = "batch_norm2d_5.w_1" shape = [24] dtype = "float32" min_val = float("0") @@ -3616,6 +3958,7 @@ class Program_weight_tensor_parameter_341: class Program_weight_tensor_parameter_342: name = "parameter_342" + original_name = "conv2d_5.w_0" shape = [24, 48, 1, 1] dtype = "float32" min_val = float("-0.144738") @@ -3627,6 +3970,7 @@ class Program_weight_tensor_parameter_342: class Program_weight_tensor_parameter_343: name = "parameter_343" + original_name = "batch_norm2d_4.b_0" shape = [24] dtype = "float32" min_val = float("0") @@ -3636,6 +3980,7 @@ class Program_weight_tensor_parameter_343: class Program_weight_tensor_parameter_344: name = "parameter_344" + original_name = "batch_norm2d_4.w_0" shape = [24] dtype = "float32" min_val = float("0") @@ -3645,6 +3990,7 @@ class Program_weight_tensor_parameter_344: class Program_weight_tensor_parameter_345: name = "parameter_345" + original_name = "batch_norm2d_4.w_2" shape = [24] dtype = "float32" min_val = float("0") @@ -3654,6 +4000,7 @@ class Program_weight_tensor_parameter_345: class Program_weight_tensor_parameter_346: name = "parameter_346" + original_name = "batch_norm2d_4.w_1" shape = [24] dtype = "float32" min_val = float("0") @@ -3663,6 +4010,7 @@ class Program_weight_tensor_parameter_346: class Program_weight_tensor_parameter_347: name = "parameter_347" + original_name = "conv2d_4.w_0" shape = [24, 48, 1, 1] dtype = "float32" min_val = float("-0.158658") @@ -3674,6 +4022,7 @@ class Program_weight_tensor_parameter_347: class Program_weight_tensor_parameter_348: name = "parameter_348" + original_name = "batch_norm2d_3.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -3683,6 +4032,7 @@ class Program_weight_tensor_parameter_348: class Program_weight_tensor_parameter_349: name = "parameter_349" + original_name = "batch_norm2d_3.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -3692,6 +4042,7 @@ class Program_weight_tensor_parameter_349: class Program_weight_tensor_parameter_350: name = "parameter_350" + original_name = "batch_norm2d_3.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -3701,6 +4052,7 @@ class Program_weight_tensor_parameter_350: class Program_weight_tensor_parameter_351: name = "parameter_351" + original_name = "batch_norm2d_3.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -3710,6 +4062,7 @@ class Program_weight_tensor_parameter_351: class Program_weight_tensor_parameter_352: name = "parameter_352" + original_name = "conv2d_3.w_0" shape = [48, 32, 3, 3] dtype = "float32" min_val = float("-0.13255") @@ -3721,6 +4074,7 @@ class Program_weight_tensor_parameter_352: class Program_weight_tensor_parameter_353: name = "parameter_353" + original_name = "batch_norm2d_2.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -3730,6 +4084,7 @@ class Program_weight_tensor_parameter_353: class Program_weight_tensor_parameter_354: name = "parameter_354" + original_name = "batch_norm2d_2.w_0" shape = [32] dtype = "float32" min_val = float("0") @@ -3739,6 +4094,7 @@ class Program_weight_tensor_parameter_354: class Program_weight_tensor_parameter_355: name = "parameter_355" + original_name = "batch_norm2d_2.w_2" shape = [32] dtype = "float32" min_val = float("0") @@ -3748,6 +4104,7 @@ class Program_weight_tensor_parameter_355: class Program_weight_tensor_parameter_356: name = "parameter_356" + original_name = "batch_norm2d_2.w_1" shape = [32] dtype = "float32" min_val = float("0") @@ -3757,6 +4114,7 @@ class Program_weight_tensor_parameter_356: class Program_weight_tensor_parameter_357: name = "parameter_357" + original_name = "conv2d_2.w_0" shape = [32, 16, 3, 3] dtype = "float32" min_val = float("-0.232932") @@ -3768,6 +4126,7 @@ class Program_weight_tensor_parameter_357: class Program_weight_tensor_parameter_358: name = "parameter_358" + original_name = "batch_norm2d_1.b_0" shape = [16] dtype = "float32" min_val = float("0") @@ -3777,6 +4136,7 @@ class Program_weight_tensor_parameter_358: class Program_weight_tensor_parameter_359: name = "parameter_359" + original_name = "batch_norm2d_1.w_0" shape = [16] dtype = "float32" min_val = float("0") @@ -3786,6 +4146,7 @@ class Program_weight_tensor_parameter_359: class Program_weight_tensor_parameter_360: name = "parameter_360" + original_name = "batch_norm2d_1.w_2" shape = [16] dtype = "float32" min_val = float("0") @@ -3795,6 +4156,7 @@ class Program_weight_tensor_parameter_360: class Program_weight_tensor_parameter_361: name = "parameter_361" + original_name = "batch_norm2d_1.w_1" shape = [16] dtype = "float32" min_val = float("0") @@ -3804,6 +4166,7 @@ class Program_weight_tensor_parameter_361: class Program_weight_tensor_parameter_362: name = "parameter_362" + original_name = "conv2d_1.w_0" shape = [16, 16, 3, 3] dtype = "float32" min_val = float("-0.248222") @@ -3815,6 +4178,7 @@ class Program_weight_tensor_parameter_362: class Program_weight_tensor_parameter_363: name = "parameter_363" + original_name = "batch_norm2d_0.b_0" shape = [16] dtype = "float32" min_val = float("0") @@ -3824,6 +4188,7 @@ class Program_weight_tensor_parameter_363: class Program_weight_tensor_parameter_364: name = "parameter_364" + original_name = "batch_norm2d_0.w_0" shape = [16] dtype = "float32" min_val = float("0") @@ -3833,6 +4198,7 @@ class Program_weight_tensor_parameter_364: class Program_weight_tensor_parameter_365: name = "parameter_365" + original_name = "batch_norm2d_0.w_2" shape = [16] dtype = "float32" min_val = float("0") @@ -3842,6 +4208,7 @@ class Program_weight_tensor_parameter_365: class Program_weight_tensor_parameter_366: name = "parameter_366" + original_name = "batch_norm2d_0.w_1" shape = [16] dtype = "float32" min_val = float("0") @@ -3851,6 +4218,7 @@ class Program_weight_tensor_parameter_366: class Program_weight_tensor_parameter_367: name = "parameter_367" + original_name = "conv2d_0.w_0" shape = [16, 3, 3, 3] dtype = "float32" min_val = float("-0.181947") diff --git a/paddle_samples/PaddleX/PP-YOLOE_plus-X/subgraph_1/input_meta.py b/paddle_samples/PaddleX/PP-YOLOE_plus-X/subgraph_1/input_meta.py index db4bddbc5..4b99c703f 100644 --- a/paddle_samples/PaddleX/PP-YOLOE_plus-X/subgraph_1/input_meta.py +++ b/paddle_samples/PaddleX/PP-YOLOE_plus-X/subgraph_1/input_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_data_0: name = "data_0" + original_name = "param_48" shape = [1] dtype = "float32" data = [1.0] @@ -7,6 +8,7 @@ class Program_weight_tensor_data_0: class Program_weight_tensor_data_1: name = "data_1" + original_name = "param_49" shape = [1] dtype = "float32" data = [1.0] @@ -14,6 +16,7 @@ class Program_weight_tensor_data_1: class Program_weight_tensor_data_2: name = "data_2" + original_name = "param_50" shape = [1] dtype = "float32" data = [1.0] @@ -21,6 +24,7 @@ class Program_weight_tensor_data_2: class Program_weight_tensor_data_3: name = "data_3" + original_name = "param_51" shape = [1] dtype = "float32" data = [1.0] @@ -28,6 +32,7 @@ class Program_weight_tensor_data_3: class Program_weight_tensor_data_4: name = "data_4" + original_name = "param_52" shape = [1] dtype = "float32" data = [1.0] @@ -35,6 +40,7 @@ class Program_weight_tensor_data_4: class Program_weight_tensor_data_5: name = "data_5" + original_name = "param_53" shape = [1] dtype = "float32" data = [1.0] @@ -42,6 +48,7 @@ class Program_weight_tensor_data_5: class Program_weight_tensor_data_6: name = "data_6" + original_name = "param_54" shape = [1] dtype = "float32" data = [1.0] @@ -49,6 +56,7 @@ class Program_weight_tensor_data_6: class Program_weight_tensor_data_7: name = "data_7" + original_name = "param_55" shape = [1] dtype = "float32" data = [1.0] @@ -56,6 +64,7 @@ class Program_weight_tensor_data_7: class Program_weight_tensor_data_8: name = "data_8" + original_name = "param_56" shape = [1] dtype = "float32" data = [1.0] @@ -63,6 +72,7 @@ class Program_weight_tensor_data_8: class Program_weight_tensor_data_9: name = "data_9" + original_name = "param_57" shape = [1] dtype = "float32" data = [1.0] @@ -70,6 +80,7 @@ class Program_weight_tensor_data_9: class Program_weight_tensor_data_10: name = "data_10" + original_name = "param_58" shape = [1] dtype = "float32" data = [1.0] @@ -77,6 +88,7 @@ class Program_weight_tensor_data_10: class Program_weight_tensor_data_11: name = "data_11" + original_name = "param_59" shape = [1] dtype = "float32" data = [1.0] @@ -84,6 +96,7 @@ class Program_weight_tensor_data_11: class Program_weight_tensor_data_12: name = "data_12" + original_name = "param_60" shape = [1] dtype = "float32" data = [1.0] @@ -91,6 +104,7 @@ class Program_weight_tensor_data_12: class Program_weight_tensor_data_13: name = "data_13" + original_name = "param_61" shape = [1] dtype = "float32" data = [1.0] @@ -98,6 +112,7 @@ class Program_weight_tensor_data_13: class Program_weight_tensor_data_14: name = "data_14" + original_name = "param_62" shape = [1] dtype = "float32" data = [1.0] @@ -105,6 +120,7 @@ class Program_weight_tensor_data_14: class Program_weight_tensor_data_15: name = "data_15" + original_name = "param_63" shape = [1] dtype = "float32" data = [1.0] @@ -112,6 +128,7 @@ class Program_weight_tensor_data_15: class Program_weight_tensor_data_16: name = "data_16" + original_name = "param_64" shape = [1] dtype = "float32" data = [1.0] @@ -119,6 +136,7 @@ class Program_weight_tensor_data_16: class Program_weight_tensor_data_17: name = "data_17" + original_name = "param_65" shape = [1] dtype = "float32" data = [1.0] @@ -126,6 +144,7 @@ class Program_weight_tensor_data_17: class Program_weight_tensor_data_18: name = "data_18" + original_name = "param_66" shape = [1] dtype = "float32" data = [1.0] @@ -133,6 +152,7 @@ class Program_weight_tensor_data_18: class Program_weight_tensor_data_19: name = "data_19" + original_name = "param_67" shape = [1] dtype = "float32" data = [1.0] @@ -140,6 +160,7 @@ class Program_weight_tensor_data_19: class Program_weight_tensor_data_20: name = "data_20" + original_name = "param_68" shape = [1] dtype = "float32" data = [1.0] @@ -147,6 +168,7 @@ class Program_weight_tensor_data_20: class Program_weight_tensor_data_21: name = "data_21" + original_name = "param_69" shape = [1] dtype = "float32" data = [1.0] @@ -154,6 +176,7 @@ class Program_weight_tensor_data_21: class Program_weight_tensor_data_22: name = "data_22" + original_name = "param_70" shape = [1] dtype = "float32" data = [1.0] @@ -161,6 +184,7 @@ class Program_weight_tensor_data_22: class Program_weight_tensor_data_23: name = "data_23" + original_name = "param_71" shape = [1] dtype = "float32" data = [1.0] @@ -168,6 +192,7 @@ class Program_weight_tensor_data_23: class Program_weight_tensor_data_24: name = "data_24" + original_name = "var_1604" shape = [4, 3, 672, 672] dtype = "float32" max_val = float("0.87451") diff --git a/paddle_samples/PaddleX/PP-YOLOE_plus-X/subgraph_1/weight_meta.py b/paddle_samples/PaddleX/PP-YOLOE_plus-X/subgraph_1/weight_meta.py index b6844f2ec..a3c58061e 100644 --- a/paddle_samples/PaddleX/PP-YOLOE_plus-X/subgraph_1/weight_meta.py +++ b/paddle_samples/PaddleX/PP-YOLOE_plus-X/subgraph_1/weight_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_parameter_0: name = "parameter_0" + original_name = "batch_norm2d_170.b_0" shape = [960] dtype = "float32" min_val = float("-0.117783") @@ -11,6 +12,7 @@ class Program_weight_tensor_parameter_0: class Program_weight_tensor_parameter_1: name = "parameter_1" + original_name = "batch_norm2d_170.w_0" shape = [960] dtype = "float32" min_val = float("0.978397") @@ -22,6 +24,7 @@ class Program_weight_tensor_parameter_1: class Program_weight_tensor_parameter_2: name = "parameter_2" + original_name = "batch_norm2d_170.w_2" shape = [960] dtype = "float32" min_val = float("0.00121084") @@ -33,6 +36,7 @@ class Program_weight_tensor_parameter_2: class Program_weight_tensor_parameter_3: name = "parameter_3" + original_name = "batch_norm2d_170.w_1" shape = [960] dtype = "float32" min_val = float("-0.19514") @@ -44,6 +48,7 @@ class Program_weight_tensor_parameter_3: class Program_weight_tensor_parameter_4: name = "parameter_4" + original_name = "conv2d_174.w_0" shape = [960, 960, 1, 1] dtype = "float32" min_val = float("-0.0432682") @@ -55,6 +60,7 @@ class Program_weight_tensor_parameter_4: class Program_weight_tensor_parameter_5: name = "parameter_5" + original_name = "batch_norm2d_169.b_0" shape = [480] dtype = "float32" min_val = float("-0.120025") @@ -66,6 +72,7 @@ class Program_weight_tensor_parameter_5: class Program_weight_tensor_parameter_6: name = "parameter_6" + original_name = "batch_norm2d_169.w_0" shape = [480] dtype = "float32" min_val = float("0.975708") @@ -77,6 +84,7 @@ class Program_weight_tensor_parameter_6: class Program_weight_tensor_parameter_7: name = "parameter_7" + original_name = "batch_norm2d_169.w_2" shape = [480] dtype = "float32" min_val = float("0.000537225") @@ -88,6 +96,7 @@ class Program_weight_tensor_parameter_7: class Program_weight_tensor_parameter_8: name = "parameter_8" + original_name = "batch_norm2d_169.w_1" shape = [480] dtype = "float32" min_val = float("-0.0459159") @@ -99,6 +108,7 @@ class Program_weight_tensor_parameter_8: class Program_weight_tensor_parameter_9: name = "parameter_9" + original_name = "conv2d_173.w_0" shape = [480, 480, 1, 1] dtype = "float32" min_val = float("-0.023804") @@ -110,6 +120,7 @@ class Program_weight_tensor_parameter_9: class Program_weight_tensor_parameter_10: name = "parameter_10" + original_name = "batch_norm2d_168.b_0" shape = [480] dtype = "float32" min_val = float("-0.120025") @@ -121,6 +132,7 @@ class Program_weight_tensor_parameter_10: class Program_weight_tensor_parameter_11: name = "parameter_11" + original_name = "batch_norm2d_168.w_0" shape = [480] dtype = "float32" min_val = float("0.955092") @@ -132,6 +144,7 @@ class Program_weight_tensor_parameter_11: class Program_weight_tensor_parameter_12: name = "parameter_12" + original_name = "batch_norm2d_168.w_2" shape = [480] dtype = "float32" min_val = float("0.00135797") @@ -143,6 +156,7 @@ class Program_weight_tensor_parameter_12: class Program_weight_tensor_parameter_13: name = "parameter_13" + original_name = "batch_norm2d_168.w_1" shape = [480] dtype = "float32" min_val = float("-0.210659") @@ -154,6 +168,7 @@ class Program_weight_tensor_parameter_13: class Program_weight_tensor_parameter_14: name = "parameter_14" + original_name = "conv2d_172.w_0" shape = [480, 480, 3, 3] dtype = "float32" min_val = float("-0.0162804") @@ -165,6 +180,7 @@ class Program_weight_tensor_parameter_14: class Program_weight_tensor_parameter_15: name = "parameter_15" + original_name = "batch_norm2d_167.b_0" shape = [480] dtype = "float32" min_val = float("-0.0802725") @@ -176,6 +192,7 @@ class Program_weight_tensor_parameter_15: class Program_weight_tensor_parameter_16: name = "parameter_16" + original_name = "batch_norm2d_167.w_0" shape = [480] dtype = "float32" min_val = float("0.972429") @@ -187,6 +204,7 @@ class Program_weight_tensor_parameter_16: class Program_weight_tensor_parameter_17: name = "parameter_17" + original_name = "batch_norm2d_167.w_2" shape = [480] dtype = "float32" min_val = float("0.003947") @@ -198,6 +216,7 @@ class Program_weight_tensor_parameter_17: class Program_weight_tensor_parameter_18: name = "parameter_18" + original_name = "batch_norm2d_167.w_1" shape = [480] dtype = "float32" min_val = float("-0.187572") @@ -209,6 +228,7 @@ class Program_weight_tensor_parameter_18: class Program_weight_tensor_parameter_19: name = "parameter_19" + original_name = "conv2d_171.w_0" shape = [480, 480, 3, 3] dtype = "float32" min_val = float("-0.0201355") @@ -220,6 +240,7 @@ class Program_weight_tensor_parameter_19: class Program_weight_tensor_parameter_20: name = "parameter_20" + original_name = "batch_norm2d_166.b_0" shape = [480] dtype = "float32" min_val = float("-0.0466307") @@ -231,6 +252,7 @@ class Program_weight_tensor_parameter_20: class Program_weight_tensor_parameter_21: name = "parameter_21" + original_name = "batch_norm2d_166.w_0" shape = [480] dtype = "float32" min_val = float("0.975112") @@ -242,6 +264,7 @@ class Program_weight_tensor_parameter_21: class Program_weight_tensor_parameter_22: name = "parameter_22" + original_name = "batch_norm2d_166.w_2" shape = [480] dtype = "float32" min_val = float("0.000479061") @@ -253,6 +276,7 @@ class Program_weight_tensor_parameter_22: class Program_weight_tensor_parameter_23: name = "parameter_23" + original_name = "batch_norm2d_166.w_1" shape = [480] dtype = "float32" min_val = float("-0.0572152") @@ -264,6 +288,7 @@ class Program_weight_tensor_parameter_23: class Program_weight_tensor_parameter_24: name = "parameter_24" + original_name = "conv2d_170.w_0" shape = [480, 480, 1, 1] dtype = "float32" min_val = float("-0.0250922") @@ -275,6 +300,7 @@ class Program_weight_tensor_parameter_24: class Program_weight_tensor_parameter_25: name = "parameter_25" + original_name = "batch_norm2d_165.b_0" shape = [480] dtype = "float32" min_val = float("-0.0466307") @@ -286,6 +312,7 @@ class Program_weight_tensor_parameter_25: class Program_weight_tensor_parameter_26: name = "parameter_26" + original_name = "batch_norm2d_165.w_0" shape = [480] dtype = "float32" min_val = float("0.991207") @@ -297,6 +324,7 @@ class Program_weight_tensor_parameter_26: class Program_weight_tensor_parameter_27: name = "parameter_27" + original_name = "batch_norm2d_165.w_2" shape = [480] dtype = "float32" min_val = float("0.00193545") @@ -308,6 +336,7 @@ class Program_weight_tensor_parameter_27: class Program_weight_tensor_parameter_28: name = "parameter_28" + original_name = "batch_norm2d_165.w_1" shape = [480] dtype = "float32" min_val = float("-0.118991") @@ -319,6 +348,7 @@ class Program_weight_tensor_parameter_28: class Program_weight_tensor_parameter_29: name = "parameter_29" + original_name = "conv2d_169.w_0" shape = [480, 480, 3, 3] dtype = "float32" min_val = float("-0.0191837") @@ -330,6 +360,7 @@ class Program_weight_tensor_parameter_29: class Program_weight_tensor_parameter_30: name = "parameter_30" + original_name = "batch_norm2d_164.b_0" shape = [480] dtype = "float32" min_val = float("-0.0496358") @@ -341,6 +372,7 @@ class Program_weight_tensor_parameter_30: class Program_weight_tensor_parameter_31: name = "parameter_31" + original_name = "batch_norm2d_164.w_0" shape = [480] dtype = "float32" min_val = float("0.963711") @@ -352,6 +384,7 @@ class Program_weight_tensor_parameter_31: class Program_weight_tensor_parameter_32: name = "parameter_32" + original_name = "batch_norm2d_164.w_2" shape = [480] dtype = "float32" min_val = float("0.00455154") @@ -363,6 +396,7 @@ class Program_weight_tensor_parameter_32: class Program_weight_tensor_parameter_33: name = "parameter_33" + original_name = "batch_norm2d_164.w_1" shape = [480] dtype = "float32" min_val = float("-0.177271") @@ -374,6 +408,7 @@ class Program_weight_tensor_parameter_33: class Program_weight_tensor_parameter_34: name = "parameter_34" + original_name = "conv2d_168.w_0" shape = [480, 480, 3, 3] dtype = "float32" min_val = float("-0.0191792") @@ -385,6 +420,7 @@ class Program_weight_tensor_parameter_34: class Program_weight_tensor_parameter_35: name = "parameter_35" + original_name = "batch_norm2d_163.b_0" shape = [480] dtype = "float32" min_val = float("-0.0586356") @@ -396,6 +432,7 @@ class Program_weight_tensor_parameter_35: class Program_weight_tensor_parameter_36: name = "parameter_36" + original_name = "batch_norm2d_163.w_0" shape = [480] dtype = "float32" min_val = float("0.969449") @@ -407,6 +444,7 @@ class Program_weight_tensor_parameter_36: class Program_weight_tensor_parameter_37: name = "parameter_37" + original_name = "batch_norm2d_163.w_2" shape = [480] dtype = "float32" min_val = float("0.000629717") @@ -418,6 +456,7 @@ class Program_weight_tensor_parameter_37: class Program_weight_tensor_parameter_38: name = "parameter_38" + original_name = "batch_norm2d_163.w_1" shape = [480] dtype = "float32" min_val = float("-0.0548808") @@ -429,6 +468,7 @@ class Program_weight_tensor_parameter_38: class Program_weight_tensor_parameter_39: name = "parameter_39" + original_name = "conv2d_167.w_0" shape = [480, 480, 1, 1] dtype = "float32" min_val = float("-0.0174508") @@ -440,6 +480,7 @@ class Program_weight_tensor_parameter_39: class Program_weight_tensor_parameter_40: name = "parameter_40" + original_name = "batch_norm2d_162.b_0" shape = [480] dtype = "float32" min_val = float("-0.0586356") @@ -451,6 +492,7 @@ class Program_weight_tensor_parameter_40: class Program_weight_tensor_parameter_41: name = "parameter_41" + original_name = "batch_norm2d_162.w_0" shape = [480] dtype = "float32" min_val = float("0.989717") @@ -462,6 +504,7 @@ class Program_weight_tensor_parameter_41: class Program_weight_tensor_parameter_42: name = "parameter_42" + original_name = "batch_norm2d_162.w_2" shape = [480] dtype = "float32" min_val = float("0.00203401") @@ -473,6 +516,7 @@ class Program_weight_tensor_parameter_42: class Program_weight_tensor_parameter_43: name = "parameter_43" + original_name = "batch_norm2d_162.w_1" shape = [480] dtype = "float32" min_val = float("-0.144746") @@ -484,6 +528,7 @@ class Program_weight_tensor_parameter_43: class Program_weight_tensor_parameter_44: name = "parameter_44" + original_name = "conv2d_166.w_0" shape = [480, 480, 3, 3] dtype = "float32" min_val = float("-0.0187333") @@ -495,6 +540,7 @@ class Program_weight_tensor_parameter_44: class Program_weight_tensor_parameter_45: name = "parameter_45" + original_name = "batch_norm2d_161.b_0" shape = [480] dtype = "float32" min_val = float("-0.0834235") @@ -506,6 +552,7 @@ class Program_weight_tensor_parameter_45: class Program_weight_tensor_parameter_46: name = "parameter_46" + original_name = "batch_norm2d_161.w_0" shape = [480] dtype = "float32" min_val = float("0.967958") @@ -517,6 +564,7 @@ class Program_weight_tensor_parameter_46: class Program_weight_tensor_parameter_47: name = "parameter_47" + original_name = "batch_norm2d_161.w_2" shape = [480] dtype = "float32" min_val = float("0.0037165") @@ -528,6 +576,7 @@ class Program_weight_tensor_parameter_47: class Program_weight_tensor_parameter_48: name = "parameter_48" + original_name = "batch_norm2d_161.w_1" shape = [480] dtype = "float32" min_val = float("-0.186416") @@ -539,6 +588,7 @@ class Program_weight_tensor_parameter_48: class Program_weight_tensor_parameter_49: name = "parameter_49" + original_name = "conv2d_165.w_0" shape = [480, 480, 3, 3] dtype = "float32" min_val = float("-0.0195741") @@ -550,6 +600,7 @@ class Program_weight_tensor_parameter_49: class Program_weight_tensor_parameter_50: name = "parameter_50" + original_name = "batch_norm2d_160.b_0" shape = [480] dtype = "float32" min_val = float("-0.0660395") @@ -561,6 +612,7 @@ class Program_weight_tensor_parameter_50: class Program_weight_tensor_parameter_51: name = "parameter_51" + original_name = "batch_norm2d_160.w_0" shape = [480] dtype = "float32" min_val = float("0.959016") @@ -572,6 +624,7 @@ class Program_weight_tensor_parameter_51: class Program_weight_tensor_parameter_52: name = "parameter_52" + original_name = "batch_norm2d_160.w_2" shape = [480] dtype = "float32" min_val = float("0.000825647") @@ -583,6 +636,7 @@ class Program_weight_tensor_parameter_52: class Program_weight_tensor_parameter_53: name = "parameter_53" + original_name = "batch_norm2d_160.w_1" shape = [480] dtype = "float32" min_val = float("-0.0489866") @@ -594,6 +648,7 @@ class Program_weight_tensor_parameter_53: class Program_weight_tensor_parameter_54: name = "parameter_54" + original_name = "conv2d_164.w_0" shape = [480, 480, 1, 1] dtype = "float32" min_val = float("-0.0169469") @@ -605,6 +660,7 @@ class Program_weight_tensor_parameter_54: class Program_weight_tensor_parameter_55: name = "parameter_55" + original_name = "batch_norm2d_159.b_0" shape = [480] dtype = "float32" min_val = float("-0.0660395") @@ -616,6 +672,7 @@ class Program_weight_tensor_parameter_55: class Program_weight_tensor_parameter_56: name = "parameter_56" + original_name = "batch_norm2d_159.w_0" shape = [480] dtype = "float32" min_val = float("0.990611") @@ -627,6 +684,7 @@ class Program_weight_tensor_parameter_56: class Program_weight_tensor_parameter_57: name = "parameter_57" + original_name = "batch_norm2d_159.w_2" shape = [480] dtype = "float32" min_val = float("0.00316639") @@ -638,6 +696,7 @@ class Program_weight_tensor_parameter_57: class Program_weight_tensor_parameter_58: name = "parameter_58" + original_name = "batch_norm2d_159.w_1" shape = [480] dtype = "float32" min_val = float("-0.123025") @@ -649,6 +708,7 @@ class Program_weight_tensor_parameter_58: class Program_weight_tensor_parameter_59: name = "parameter_59" + original_name = "conv2d_163.w_0" shape = [480, 480, 3, 3] dtype = "float32" min_val = float("-0.0293124") @@ -660,6 +720,7 @@ class Program_weight_tensor_parameter_59: class Program_weight_tensor_parameter_60: name = "parameter_60" + original_name = "batch_norm2d_158.b_0" shape = [480] dtype = "float32" min_val = float("-0.0688525") @@ -671,6 +732,7 @@ class Program_weight_tensor_parameter_60: class Program_weight_tensor_parameter_61: name = "parameter_61" + original_name = "batch_norm2d_158.w_0" shape = [480] dtype = "float32" min_val = float("0.970045") @@ -682,6 +744,7 @@ class Program_weight_tensor_parameter_61: class Program_weight_tensor_parameter_62: name = "parameter_62" + original_name = "batch_norm2d_158.w_2" shape = [480] dtype = "float32" min_val = float("0.00421962") @@ -693,6 +756,7 @@ class Program_weight_tensor_parameter_62: class Program_weight_tensor_parameter_63: name = "parameter_63" + original_name = "batch_norm2d_158.w_1" shape = [480] dtype = "float32" min_val = float("-0.114163") @@ -704,6 +768,7 @@ class Program_weight_tensor_parameter_63: class Program_weight_tensor_parameter_64: name = "parameter_64" + original_name = "conv2d_162.w_0" shape = [480, 480, 3, 3] dtype = "float32" min_val = float("-0.0259895") @@ -715,6 +780,7 @@ class Program_weight_tensor_parameter_64: class Program_weight_tensor_parameter_65: name = "parameter_65" + original_name = "batch_norm2d_157.b_0" shape = [480] dtype = "float32" min_val = float("-0.0445437") @@ -726,6 +792,7 @@ class Program_weight_tensor_parameter_65: class Program_weight_tensor_parameter_66: name = "parameter_66" + original_name = "batch_norm2d_157.w_0" shape = [480] dtype = "float32" min_val = float("0.983756") @@ -737,6 +804,7 @@ class Program_weight_tensor_parameter_66: class Program_weight_tensor_parameter_67: name = "parameter_67" + original_name = "batch_norm2d_157.w_2" shape = [480] dtype = "float32" min_val = float("0.00242431") @@ -748,6 +816,7 @@ class Program_weight_tensor_parameter_67: class Program_weight_tensor_parameter_68: name = "parameter_68" + original_name = "batch_norm2d_157.w_1" shape = [480] dtype = "float32" min_val = float("-0.0728065") @@ -759,6 +828,7 @@ class Program_weight_tensor_parameter_68: class Program_weight_tensor_parameter_69: name = "parameter_69" + original_name = "conv2d_161.w_0" shape = [480, 1440, 1, 1] dtype = "float32" min_val = float("-0.0311823") @@ -770,6 +840,7 @@ class Program_weight_tensor_parameter_69: class Program_weight_tensor_parameter_70: name = "parameter_70" + original_name = "batch_norm2d_156.b_0" shape = [480] dtype = "float32" min_val = float("-0.0199113") @@ -781,6 +852,7 @@ class Program_weight_tensor_parameter_70: class Program_weight_tensor_parameter_71: name = "parameter_71" + original_name = "batch_norm2d_156.w_0" shape = [480] dtype = "float32" min_val = float("0.960507") @@ -792,6 +864,7 @@ class Program_weight_tensor_parameter_71: class Program_weight_tensor_parameter_72: name = "parameter_72" + original_name = "batch_norm2d_156.w_2" shape = [480] dtype = "float32" min_val = float("0.00150293") @@ -803,6 +876,7 @@ class Program_weight_tensor_parameter_72: class Program_weight_tensor_parameter_73: name = "parameter_73" + original_name = "batch_norm2d_156.w_1" shape = [480] dtype = "float32" min_val = float("-0.0642416") @@ -814,6 +888,7 @@ class Program_weight_tensor_parameter_73: class Program_weight_tensor_parameter_74: name = "parameter_74" + original_name = "conv2d_160.w_0" shape = [480, 1440, 1, 1] dtype = "float32" min_val = float("-0.032606") @@ -825,6 +900,7 @@ class Program_weight_tensor_parameter_74: class Program_weight_tensor_parameter_75: name = "parameter_75" + original_name = "batch_norm2d_155.b_0" shape = [480] dtype = "float32" min_val = float("-0.0441952") @@ -836,6 +912,7 @@ class Program_weight_tensor_parameter_75: class Program_weight_tensor_parameter_76: name = "parameter_76" + original_name = "batch_norm2d_155.w_0" shape = [480] dtype = "float32" min_val = float("0.988525") @@ -847,6 +924,7 @@ class Program_weight_tensor_parameter_76: class Program_weight_tensor_parameter_77: name = "parameter_77" + original_name = "batch_norm2d_155.w_2" shape = [480] dtype = "float32" min_val = float("0.0029207") @@ -858,6 +936,7 @@ class Program_weight_tensor_parameter_77: class Program_weight_tensor_parameter_78: name = "parameter_78" + original_name = "batch_norm2d_155.w_1" shape = [480] dtype = "float32" min_val = float("-0.247289") @@ -869,6 +948,7 @@ class Program_weight_tensor_parameter_78: class Program_weight_tensor_parameter_79: name = "parameter_79" + original_name = "conv2d_159.w_0" shape = [480, 480, 3, 3] dtype = "float32" min_val = float("-0.0260306") @@ -880,6 +960,7 @@ class Program_weight_tensor_parameter_79: class Program_weight_tensor_parameter_80: name = "parameter_80" + original_name = "batch_norm2d_154.b_0" shape = [480] dtype = "float32" min_val = float("-0.326578") @@ -891,6 +972,7 @@ class Program_weight_tensor_parameter_80: class Program_weight_tensor_parameter_81: name = "parameter_81" + original_name = "batch_norm2d_154.w_0" shape = [480] dtype = "float32" min_val = float("0.8169") @@ -902,6 +984,7 @@ class Program_weight_tensor_parameter_81: class Program_weight_tensor_parameter_82: name = "parameter_82" + original_name = "batch_norm2d_154.w_2" shape = [480] dtype = "float32" min_val = float("0.00332805") @@ -913,6 +996,7 @@ class Program_weight_tensor_parameter_82: class Program_weight_tensor_parameter_83: name = "parameter_83" + original_name = "batch_norm2d_154.w_1" shape = [480] dtype = "float32" min_val = float("-0.194858") @@ -924,6 +1008,7 @@ class Program_weight_tensor_parameter_83: class Program_weight_tensor_parameter_84: name = "parameter_84" + original_name = "conv2d_158.w_0" shape = [480, 480, 1, 1] dtype = "float32" min_val = float("-0.0773872") @@ -935,6 +1020,7 @@ class Program_weight_tensor_parameter_84: class Program_weight_tensor_parameter_85: name = "parameter_85" + original_name = "batch_norm2d_153.b_0" shape = [240] dtype = "float32" min_val = float("-0.0963525") @@ -946,6 +1032,7 @@ class Program_weight_tensor_parameter_85: class Program_weight_tensor_parameter_86: name = "parameter_86" + original_name = "batch_norm2d_153.w_0" shape = [240] dtype = "float32" min_val = float("0.832936") @@ -957,6 +1044,7 @@ class Program_weight_tensor_parameter_86: class Program_weight_tensor_parameter_87: name = "parameter_87" + original_name = "batch_norm2d_153.w_2" shape = [240] dtype = "float32" min_val = float("0.000528486") @@ -968,6 +1056,7 @@ class Program_weight_tensor_parameter_87: class Program_weight_tensor_parameter_88: name = "parameter_88" + original_name = "batch_norm2d_153.w_1" shape = [240] dtype = "float32" min_val = float("-0.0686565") @@ -979,6 +1068,7 @@ class Program_weight_tensor_parameter_88: class Program_weight_tensor_parameter_89: name = "parameter_89" + original_name = "conv2d_157.w_0" shape = [240, 240, 1, 1] dtype = "float32" min_val = float("-0.0382056") @@ -990,6 +1080,7 @@ class Program_weight_tensor_parameter_89: class Program_weight_tensor_parameter_90: name = "parameter_90" + original_name = "batch_norm2d_152.b_0" shape = [240] dtype = "float32" min_val = float("-0.0963525") @@ -1001,6 +1092,7 @@ class Program_weight_tensor_parameter_90: class Program_weight_tensor_parameter_91: name = "parameter_91" + original_name = "batch_norm2d_152.w_0" shape = [240] dtype = "float32" min_val = float("0.842958") @@ -1012,6 +1104,7 @@ class Program_weight_tensor_parameter_91: class Program_weight_tensor_parameter_92: name = "parameter_92" + original_name = "batch_norm2d_152.w_2" shape = [240] dtype = "float32" min_val = float("0.00282089") @@ -1023,6 +1116,7 @@ class Program_weight_tensor_parameter_92: class Program_weight_tensor_parameter_93: name = "parameter_93" + original_name = "batch_norm2d_152.w_1" shape = [240] dtype = "float32" min_val = float("-0.274015") @@ -1034,6 +1128,7 @@ class Program_weight_tensor_parameter_93: class Program_weight_tensor_parameter_94: name = "parameter_94" + original_name = "conv2d_156.w_0" shape = [240, 240, 3, 3] dtype = "float32" min_val = float("-0.0380686") @@ -1045,6 +1140,7 @@ class Program_weight_tensor_parameter_94: class Program_weight_tensor_parameter_95: name = "parameter_95" + original_name = "batch_norm2d_151.b_0" shape = [240] dtype = "float32" min_val = float("-0.147907") @@ -1056,6 +1152,7 @@ class Program_weight_tensor_parameter_95: class Program_weight_tensor_parameter_96: name = "parameter_96" + original_name = "batch_norm2d_151.w_0" shape = [240] dtype = "float32" min_val = float("0.917288") @@ -1067,6 +1164,7 @@ class Program_weight_tensor_parameter_96: class Program_weight_tensor_parameter_97: name = "parameter_97" + original_name = "batch_norm2d_151.w_2" shape = [240] dtype = "float32" min_val = float("0.00504724") @@ -1078,6 +1176,7 @@ class Program_weight_tensor_parameter_97: class Program_weight_tensor_parameter_98: name = "parameter_98" + original_name = "batch_norm2d_151.w_1" shape = [240] dtype = "float32" min_val = float("-0.226657") @@ -1089,6 +1188,7 @@ class Program_weight_tensor_parameter_98: class Program_weight_tensor_parameter_99: name = "parameter_99" + original_name = "conv2d_155.w_0" shape = [240, 240, 3, 3] dtype = "float32" min_val = float("-0.0464243") @@ -1100,6 +1200,7 @@ class Program_weight_tensor_parameter_99: class Program_weight_tensor_parameter_100: name = "parameter_100" + original_name = "batch_norm2d_150.b_0" shape = [240] dtype = "float32" min_val = float("-0.0796736") @@ -1111,6 +1212,7 @@ class Program_weight_tensor_parameter_100: class Program_weight_tensor_parameter_101: name = "parameter_101" + original_name = "batch_norm2d_150.w_0" shape = [240] dtype = "float32" min_val = float("0.957526") @@ -1122,6 +1224,7 @@ class Program_weight_tensor_parameter_101: class Program_weight_tensor_parameter_102: name = "parameter_102" + original_name = "batch_norm2d_150.w_2" shape = [240] dtype = "float32" min_val = float("0.00116147") @@ -1133,6 +1236,7 @@ class Program_weight_tensor_parameter_102: class Program_weight_tensor_parameter_103: name = "parameter_103" + original_name = "batch_norm2d_150.w_1" shape = [240] dtype = "float32" min_val = float("-0.0411964") @@ -1144,6 +1248,7 @@ class Program_weight_tensor_parameter_103: class Program_weight_tensor_parameter_104: name = "parameter_104" + original_name = "conv2d_154.w_0" shape = [240, 240, 1, 1] dtype = "float32" min_val = float("-0.0355971") @@ -1155,6 +1260,7 @@ class Program_weight_tensor_parameter_104: class Program_weight_tensor_parameter_105: name = "parameter_105" + original_name = "batch_norm2d_149.b_0" shape = [240] dtype = "float32" min_val = float("-0.0796736") @@ -1166,6 +1272,7 @@ class Program_weight_tensor_parameter_105: class Program_weight_tensor_parameter_106: name = "parameter_106" + original_name = "batch_norm2d_149.w_0" shape = [240] dtype = "float32" min_val = float("0.977337") @@ -1177,6 +1284,7 @@ class Program_weight_tensor_parameter_106: class Program_weight_tensor_parameter_107: name = "parameter_107" + original_name = "batch_norm2d_149.w_2" shape = [240] dtype = "float32" min_val = float("0.00298047") @@ -1188,6 +1296,7 @@ class Program_weight_tensor_parameter_107: class Program_weight_tensor_parameter_108: name = "parameter_108" + original_name = "batch_norm2d_149.w_1" shape = [240] dtype = "float32" min_val = float("-0.143174") @@ -1199,6 +1308,7 @@ class Program_weight_tensor_parameter_108: class Program_weight_tensor_parameter_109: name = "parameter_109" + original_name = "conv2d_153.w_0" shape = [240, 240, 3, 3] dtype = "float32" min_val = float("-0.0338693") @@ -1210,6 +1320,7 @@ class Program_weight_tensor_parameter_109: class Program_weight_tensor_parameter_110: name = "parameter_110" + original_name = "batch_norm2d_148.b_0" shape = [240] dtype = "float32" min_val = float("-0.123905") @@ -1221,6 +1332,7 @@ class Program_weight_tensor_parameter_110: class Program_weight_tensor_parameter_111: name = "parameter_111" + original_name = "batch_norm2d_148.w_0" shape = [240] dtype = "float32" min_val = float("0.941431") @@ -1232,6 +1344,7 @@ class Program_weight_tensor_parameter_111: class Program_weight_tensor_parameter_112: name = "parameter_112" + original_name = "batch_norm2d_148.w_2" shape = [240] dtype = "float32" min_val = float("0.00696675") @@ -1243,6 +1356,7 @@ class Program_weight_tensor_parameter_112: class Program_weight_tensor_parameter_113: name = "parameter_113" + original_name = "batch_norm2d_148.w_1" shape = [240] dtype = "float32" min_val = float("-0.156119") @@ -1254,6 +1368,7 @@ class Program_weight_tensor_parameter_113: class Program_weight_tensor_parameter_114: name = "parameter_114" + original_name = "conv2d_152.w_0" shape = [240, 240, 3, 3] dtype = "float32" min_val = float("-0.0491652") @@ -1265,6 +1380,7 @@ class Program_weight_tensor_parameter_114: class Program_weight_tensor_parameter_115: name = "parameter_115" + original_name = "batch_norm2d_147.b_0" shape = [240] dtype = "float32" min_val = float("-0.105518") @@ -1276,6 +1392,7 @@ class Program_weight_tensor_parameter_115: class Program_weight_tensor_parameter_116: name = "parameter_116" + original_name = "batch_norm2d_147.w_0" shape = [240] dtype = "float32" min_val = float("0.953707") @@ -1287,6 +1404,7 @@ class Program_weight_tensor_parameter_116: class Program_weight_tensor_parameter_117: name = "parameter_117" + original_name = "batch_norm2d_147.w_2" shape = [240] dtype = "float32" min_val = float("0.00114972") @@ -1298,6 +1416,7 @@ class Program_weight_tensor_parameter_117: class Program_weight_tensor_parameter_118: name = "parameter_118" + original_name = "batch_norm2d_147.w_1" shape = [240] dtype = "float32" min_val = float("-0.0499622") @@ -1309,6 +1428,7 @@ class Program_weight_tensor_parameter_118: class Program_weight_tensor_parameter_119: name = "parameter_119" + original_name = "conv2d_151.w_0" shape = [240, 240, 1, 1] dtype = "float32" min_val = float("-0.0357876") @@ -1320,6 +1440,7 @@ class Program_weight_tensor_parameter_119: class Program_weight_tensor_parameter_120: name = "parameter_120" + original_name = "batch_norm2d_146.b_0" shape = [240] dtype = "float32" min_val = float("-0.105518") @@ -1331,6 +1452,7 @@ class Program_weight_tensor_parameter_120: class Program_weight_tensor_parameter_121: name = "parameter_121" + original_name = "batch_norm2d_146.w_0" shape = [240] dtype = "float32" min_val = float("0.96915") @@ -1342,6 +1464,7 @@ class Program_weight_tensor_parameter_121: class Program_weight_tensor_parameter_122: name = "parameter_122" + original_name = "batch_norm2d_146.w_2" shape = [240] dtype = "float32" min_val = float("0.00379264") @@ -1353,6 +1476,7 @@ class Program_weight_tensor_parameter_122: class Program_weight_tensor_parameter_123: name = "parameter_123" + original_name = "batch_norm2d_146.w_1" shape = [240] dtype = "float32" min_val = float("-0.139771") @@ -1364,6 +1488,7 @@ class Program_weight_tensor_parameter_123: class Program_weight_tensor_parameter_124: name = "parameter_124" + original_name = "conv2d_150.w_0" shape = [240, 240, 3, 3] dtype = "float32" min_val = float("-0.0343594") @@ -1375,6 +1500,7 @@ class Program_weight_tensor_parameter_124: class Program_weight_tensor_parameter_125: name = "parameter_125" + original_name = "batch_norm2d_145.b_0" shape = [240] dtype = "float32" min_val = float("-0.13142") @@ -1386,6 +1512,7 @@ class Program_weight_tensor_parameter_125: class Program_weight_tensor_parameter_126: name = "parameter_126" + original_name = "batch_norm2d_145.w_0" shape = [240] dtype = "float32" min_val = float("0.915201") @@ -1397,6 +1524,7 @@ class Program_weight_tensor_parameter_126: class Program_weight_tensor_parameter_127: name = "parameter_127" + original_name = "batch_norm2d_145.w_2" shape = [240] dtype = "float32" min_val = float("0.00743392") @@ -1408,6 +1536,7 @@ class Program_weight_tensor_parameter_127: class Program_weight_tensor_parameter_128: name = "parameter_128" + original_name = "batch_norm2d_145.w_1" shape = [240] dtype = "float32" min_val = float("-0.1361") @@ -1419,6 +1548,7 @@ class Program_weight_tensor_parameter_128: class Program_weight_tensor_parameter_129: name = "parameter_129" + original_name = "conv2d_149.w_0" shape = [240, 240, 3, 3] dtype = "float32" min_val = float("-0.043621") @@ -1430,6 +1560,7 @@ class Program_weight_tensor_parameter_129: class Program_weight_tensor_parameter_130: name = "parameter_130" + original_name = "batch_norm2d_144.b_0" shape = [240] dtype = "float32" min_val = float("-0.138967") @@ -1441,6 +1572,7 @@ class Program_weight_tensor_parameter_130: class Program_weight_tensor_parameter_131: name = "parameter_131" + original_name = "batch_norm2d_144.w_0" shape = [240] dtype = "float32" min_val = float("0.946498") @@ -1452,6 +1584,7 @@ class Program_weight_tensor_parameter_131: class Program_weight_tensor_parameter_132: name = "parameter_132" + original_name = "batch_norm2d_144.w_2" shape = [240] dtype = "float32" min_val = float("0.00211459") @@ -1463,6 +1596,7 @@ class Program_weight_tensor_parameter_132: class Program_weight_tensor_parameter_133: name = "parameter_133" + original_name = "batch_norm2d_144.w_1" shape = [240] dtype = "float32" min_val = float("-0.0664631") @@ -1474,6 +1608,7 @@ class Program_weight_tensor_parameter_133: class Program_weight_tensor_parameter_134: name = "parameter_134" + original_name = "conv2d_148.w_0" shape = [240, 240, 1, 1] dtype = "float32" min_val = float("-0.0423829") @@ -1485,6 +1620,7 @@ class Program_weight_tensor_parameter_134: class Program_weight_tensor_parameter_135: name = "parameter_135" + original_name = "batch_norm2d_143.b_0" shape = [240] dtype = "float32" min_val = float("-0.138967") @@ -1496,6 +1632,7 @@ class Program_weight_tensor_parameter_135: class Program_weight_tensor_parameter_136: name = "parameter_136" + original_name = "batch_norm2d_143.w_0" shape = [240] dtype = "float32" min_val = float("0.958712") @@ -1507,6 +1644,7 @@ class Program_weight_tensor_parameter_136: class Program_weight_tensor_parameter_137: name = "parameter_137" + original_name = "batch_norm2d_143.w_2" shape = [240] dtype = "float32" min_val = float("0.0064707") @@ -1518,6 +1656,7 @@ class Program_weight_tensor_parameter_137: class Program_weight_tensor_parameter_138: name = "parameter_138" + original_name = "batch_norm2d_143.w_1" shape = [240] dtype = "float32" min_val = float("-0.199244") @@ -1529,6 +1668,7 @@ class Program_weight_tensor_parameter_138: class Program_weight_tensor_parameter_139: name = "parameter_139" + original_name = "conv2d_147.w_0" shape = [240, 240, 3, 3] dtype = "float32" min_val = float("-0.0406776") @@ -1540,6 +1680,7 @@ class Program_weight_tensor_parameter_139: class Program_weight_tensor_parameter_140: name = "parameter_140" + original_name = "batch_norm2d_142.b_0" shape = [240] dtype = "float32" min_val = float("-0.130819") @@ -1551,6 +1692,7 @@ class Program_weight_tensor_parameter_140: class Program_weight_tensor_parameter_141: name = "parameter_141" + original_name = "batch_norm2d_142.w_0" shape = [240] dtype = "float32" min_val = float("0.943815") @@ -1562,6 +1704,7 @@ class Program_weight_tensor_parameter_141: class Program_weight_tensor_parameter_142: name = "parameter_142" + original_name = "batch_norm2d_142.w_2" shape = [240] dtype = "float32" min_val = float("0.00719668") @@ -1573,6 +1716,7 @@ class Program_weight_tensor_parameter_142: class Program_weight_tensor_parameter_143: name = "parameter_143" + original_name = "batch_norm2d_142.w_1" shape = [240] dtype = "float32" min_val = float("-0.140681") @@ -1584,6 +1728,7 @@ class Program_weight_tensor_parameter_143: class Program_weight_tensor_parameter_144: name = "parameter_144" + original_name = "conv2d_146.w_0" shape = [240, 240, 3, 3] dtype = "float32" min_val = float("-0.0466949") @@ -1595,6 +1740,7 @@ class Program_weight_tensor_parameter_144: class Program_weight_tensor_parameter_145: name = "parameter_145" + original_name = "batch_norm2d_141.b_0" shape = [240] dtype = "float32" min_val = float("-0.0952654") @@ -1606,6 +1752,7 @@ class Program_weight_tensor_parameter_145: class Program_weight_tensor_parameter_146: name = "parameter_146" + original_name = "batch_norm2d_141.w_0" shape = [240] dtype = "float32" min_val = float("0.96086") @@ -1617,6 +1764,7 @@ class Program_weight_tensor_parameter_146: class Program_weight_tensor_parameter_147: name = "parameter_147" + original_name = "batch_norm2d_141.w_2" shape = [240] dtype = "float32" min_val = float("0.00378925") @@ -1628,6 +1776,7 @@ class Program_weight_tensor_parameter_147: class Program_weight_tensor_parameter_148: name = "parameter_148" + original_name = "batch_norm2d_141.w_1" shape = [240] dtype = "float32" min_val = float("-0.0940249") @@ -1639,6 +1788,7 @@ class Program_weight_tensor_parameter_148: class Program_weight_tensor_parameter_149: name = "parameter_149" + original_name = "conv2d_145.w_0" shape = [240, 720, 1, 1] dtype = "float32" min_val = float("-0.0536527") @@ -1650,6 +1800,7 @@ class Program_weight_tensor_parameter_149: class Program_weight_tensor_parameter_150: name = "parameter_150" + original_name = "batch_norm2d_140.b_0" shape = [240] dtype = "float32" min_val = float("-0.0543397") @@ -1661,6 +1812,7 @@ class Program_weight_tensor_parameter_150: class Program_weight_tensor_parameter_151: name = "parameter_151" + original_name = "batch_norm2d_140.w_0" shape = [240] dtype = "float32" min_val = float("0.959016") @@ -1672,6 +1824,7 @@ class Program_weight_tensor_parameter_151: class Program_weight_tensor_parameter_152: name = "parameter_152" + original_name = "batch_norm2d_140.w_2" shape = [240] dtype = "float32" min_val = float("0.00270959") @@ -1683,6 +1836,7 @@ class Program_weight_tensor_parameter_152: class Program_weight_tensor_parameter_153: name = "parameter_153" + original_name = "batch_norm2d_140.w_1" shape = [240] dtype = "float32" min_val = float("-0.0749718") @@ -1694,6 +1848,7 @@ class Program_weight_tensor_parameter_153: class Program_weight_tensor_parameter_154: name = "parameter_154" + original_name = "conv2d_144.w_0" shape = [240, 720, 1, 1] dtype = "float32" min_val = float("-0.0656767") @@ -1705,6 +1860,7 @@ class Program_weight_tensor_parameter_154: class Program_weight_tensor_parameter_155: name = "parameter_155" + original_name = "batch_norm2d_139.b_0" shape = [240] dtype = "float32" min_val = float("-0.0860687") @@ -1716,6 +1872,7 @@ class Program_weight_tensor_parameter_155: class Program_weight_tensor_parameter_156: name = "parameter_156" + original_name = "batch_norm2d_139.w_0" shape = [240] dtype = "float32" min_val = float("0.845751") @@ -1727,6 +1884,7 @@ class Program_weight_tensor_parameter_156: class Program_weight_tensor_parameter_157: name = "parameter_157" + original_name = "batch_norm2d_139.w_2" shape = [240] dtype = "float32" min_val = float("0.00493766") @@ -1738,6 +1896,7 @@ class Program_weight_tensor_parameter_157: class Program_weight_tensor_parameter_158: name = "parameter_158" + original_name = "batch_norm2d_139.w_1" shape = [240] dtype = "float32" min_val = float("-0.534536") @@ -1749,6 +1908,7 @@ class Program_weight_tensor_parameter_158: class Program_weight_tensor_parameter_159: name = "parameter_159" + original_name = "conv2d_143.w_0" shape = [240, 240, 3, 3] dtype = "float32" min_val = float("-0.0465646") @@ -1760,6 +1920,7 @@ class Program_weight_tensor_parameter_159: class Program_weight_tensor_parameter_160: name = "parameter_160" + original_name = "batch_norm2d_138.b_0" shape = [240] dtype = "float32" min_val = float("-0.688753") @@ -1771,6 +1932,7 @@ class Program_weight_tensor_parameter_160: class Program_weight_tensor_parameter_161: name = "parameter_161" + original_name = "batch_norm2d_138.w_0" shape = [240] dtype = "float32" min_val = float("0.534411") @@ -1782,6 +1944,7 @@ class Program_weight_tensor_parameter_161: class Program_weight_tensor_parameter_162: name = "parameter_162" + original_name = "batch_norm2d_138.w_2" shape = [240] dtype = "float32" min_val = float("0.00380324") @@ -1793,6 +1956,7 @@ class Program_weight_tensor_parameter_162: class Program_weight_tensor_parameter_163: name = "parameter_163" + original_name = "batch_norm2d_138.w_1" shape = [240] dtype = "float32" min_val = float("-0.209357") @@ -1804,6 +1968,7 @@ class Program_weight_tensor_parameter_163: class Program_weight_tensor_parameter_164: name = "parameter_164" + original_name = "conv2d_142.w_0" shape = [240, 240, 1, 1] dtype = "float32" min_val = float("-0.132467") @@ -1815,6 +1980,7 @@ class Program_weight_tensor_parameter_164: class Program_weight_tensor_parameter_165: name = "parameter_165" + original_name = "batch_norm2d_137.b_0" shape = [120] dtype = "float32" min_val = float("-0.320602") @@ -1826,6 +1992,7 @@ class Program_weight_tensor_parameter_165: class Program_weight_tensor_parameter_166: name = "parameter_166" + original_name = "batch_norm2d_137.w_0" shape = [120] dtype = "float32" min_val = float("0.865127") @@ -1837,6 +2004,7 @@ class Program_weight_tensor_parameter_166: class Program_weight_tensor_parameter_167: name = "parameter_167" + original_name = "batch_norm2d_137.w_2" shape = [120] dtype = "float32" min_val = float("0.00171871") @@ -1848,6 +2016,7 @@ class Program_weight_tensor_parameter_167: class Program_weight_tensor_parameter_168: name = "parameter_168" + original_name = "batch_norm2d_137.w_1" shape = [120] dtype = "float32" min_val = float("-0.0635781") @@ -1859,6 +2028,7 @@ class Program_weight_tensor_parameter_168: class Program_weight_tensor_parameter_169: name = "parameter_169" + original_name = "conv2d_141.w_0" shape = [120, 120, 1, 1] dtype = "float32" min_val = float("-0.0669268") @@ -1870,6 +2040,7 @@ class Program_weight_tensor_parameter_169: class Program_weight_tensor_parameter_170: name = "parameter_170" + original_name = "batch_norm2d_136.b_0" shape = [120] dtype = "float32" min_val = float("-0.320602") @@ -1881,6 +2052,7 @@ class Program_weight_tensor_parameter_170: class Program_weight_tensor_parameter_171: name = "parameter_171" + original_name = "batch_norm2d_136.w_0" shape = [120] dtype = "float32" min_val = float("0.533957") @@ -1892,6 +2064,7 @@ class Program_weight_tensor_parameter_171: class Program_weight_tensor_parameter_172: name = "parameter_172" + original_name = "batch_norm2d_136.w_2" shape = [120] dtype = "float32" min_val = float("0.00508932") @@ -1903,6 +2076,7 @@ class Program_weight_tensor_parameter_172: class Program_weight_tensor_parameter_173: name = "parameter_173" + original_name = "batch_norm2d_136.w_1" shape = [120] dtype = "float32" min_val = float("-0.215154") @@ -1914,6 +2088,7 @@ class Program_weight_tensor_parameter_173: class Program_weight_tensor_parameter_174: name = "parameter_174" + original_name = "conv2d_140.w_0" shape = [120, 120, 3, 3] dtype = "float32" min_val = float("-0.0598819") @@ -1925,6 +2100,7 @@ class Program_weight_tensor_parameter_174: class Program_weight_tensor_parameter_175: name = "parameter_175" + original_name = "batch_norm2d_135.b_0" shape = [120] dtype = "float32" min_val = float("-0.423726") @@ -1936,6 +2112,7 @@ class Program_weight_tensor_parameter_175: class Program_weight_tensor_parameter_176: name = "parameter_176" + original_name = "batch_norm2d_135.w_0" shape = [120] dtype = "float32" min_val = float("0.665748") @@ -1947,6 +2124,7 @@ class Program_weight_tensor_parameter_176: class Program_weight_tensor_parameter_177: name = "parameter_177" + original_name = "batch_norm2d_135.w_2" shape = [120] dtype = "float32" min_val = float("0.00631202") @@ -1958,6 +2136,7 @@ class Program_weight_tensor_parameter_177: class Program_weight_tensor_parameter_178: name = "parameter_178" + original_name = "batch_norm2d_135.w_1" shape = [120] dtype = "float32" min_val = float("-0.196486") @@ -1969,6 +2148,7 @@ class Program_weight_tensor_parameter_178: class Program_weight_tensor_parameter_179: name = "parameter_179" + original_name = "conv2d_139.w_0" shape = [120, 120, 3, 3] dtype = "float32" min_val = float("-0.0688533") @@ -1980,6 +2160,7 @@ class Program_weight_tensor_parameter_179: class Program_weight_tensor_parameter_180: name = "parameter_180" + original_name = "batch_norm2d_134.b_0" shape = [120] dtype = "float32" min_val = float("-0.195449") @@ -1991,6 +2172,7 @@ class Program_weight_tensor_parameter_180: class Program_weight_tensor_parameter_181: name = "parameter_181" + original_name = "batch_norm2d_134.w_0" shape = [120] dtype = "float32" min_val = float("0.63786") @@ -2002,6 +2184,7 @@ class Program_weight_tensor_parameter_181: class Program_weight_tensor_parameter_182: name = "parameter_182" + original_name = "batch_norm2d_134.w_2" shape = [120] dtype = "float32" min_val = float("0.00281881") @@ -2013,6 +2196,7 @@ class Program_weight_tensor_parameter_182: class Program_weight_tensor_parameter_183: name = "parameter_183" + original_name = "batch_norm2d_134.w_1" shape = [120] dtype = "float32" min_val = float("-0.0516488") @@ -2024,6 +2208,7 @@ class Program_weight_tensor_parameter_183: class Program_weight_tensor_parameter_184: name = "parameter_184" + original_name = "conv2d_138.w_0" shape = [120, 120, 1, 1] dtype = "float32" min_val = float("-0.0690325") @@ -2035,6 +2220,7 @@ class Program_weight_tensor_parameter_184: class Program_weight_tensor_parameter_185: name = "parameter_185" + original_name = "batch_norm2d_133.b_0" shape = [120] dtype = "float32" min_val = float("-0.195449") @@ -2046,6 +2232,7 @@ class Program_weight_tensor_parameter_185: class Program_weight_tensor_parameter_186: name = "parameter_186" + original_name = "batch_norm2d_133.w_0" shape = [120] dtype = "float32" min_val = float("0.820715") @@ -2057,6 +2244,7 @@ class Program_weight_tensor_parameter_186: class Program_weight_tensor_parameter_187: name = "parameter_187" + original_name = "batch_norm2d_133.w_2" shape = [120] dtype = "float32" min_val = float("0.00838187") @@ -2068,6 +2256,7 @@ class Program_weight_tensor_parameter_187: class Program_weight_tensor_parameter_188: name = "parameter_188" + original_name = "batch_norm2d_133.w_1" shape = [120] dtype = "float32" min_val = float("-0.153407") @@ -2079,6 +2268,7 @@ class Program_weight_tensor_parameter_188: class Program_weight_tensor_parameter_189: name = "parameter_189" + original_name = "conv2d_137.w_0" shape = [120, 120, 3, 3] dtype = "float32" min_val = float("-0.0601456") @@ -2090,6 +2280,7 @@ class Program_weight_tensor_parameter_189: class Program_weight_tensor_parameter_190: name = "parameter_190" + original_name = "batch_norm2d_132.b_0" shape = [120] dtype = "float32" min_val = float("-0.248311") @@ -2101,6 +2292,7 @@ class Program_weight_tensor_parameter_190: class Program_weight_tensor_parameter_191: name = "parameter_191" + original_name = "batch_norm2d_132.w_0" shape = [120] dtype = "float32" min_val = float("0.832349") @@ -2112,6 +2304,7 @@ class Program_weight_tensor_parameter_191: class Program_weight_tensor_parameter_192: name = "parameter_192" + original_name = "batch_norm2d_132.w_2" shape = [120] dtype = "float32" min_val = float("0.00659475") @@ -2123,6 +2316,7 @@ class Program_weight_tensor_parameter_192: class Program_weight_tensor_parameter_193: name = "parameter_193" + original_name = "batch_norm2d_132.w_1" shape = [120] dtype = "float32" min_val = float("-0.107366") @@ -2134,6 +2328,7 @@ class Program_weight_tensor_parameter_193: class Program_weight_tensor_parameter_194: name = "parameter_194" + original_name = "conv2d_136.w_0" shape = [120, 120, 3, 3] dtype = "float32" min_val = float("-0.0616609") @@ -2145,6 +2340,7 @@ class Program_weight_tensor_parameter_194: class Program_weight_tensor_parameter_195: name = "parameter_195" + original_name = "batch_norm2d_131.b_0" shape = [120] dtype = "float32" min_val = float("-0.25097") @@ -2156,6 +2352,7 @@ class Program_weight_tensor_parameter_195: class Program_weight_tensor_parameter_196: name = "parameter_196" + original_name = "batch_norm2d_131.w_0" shape = [120] dtype = "float32" min_val = float("0.779191") @@ -2167,6 +2364,7 @@ class Program_weight_tensor_parameter_196: class Program_weight_tensor_parameter_197: name = "parameter_197" + original_name = "batch_norm2d_131.w_2" shape = [120] dtype = "float32" min_val = float("0.00315771") @@ -2178,6 +2376,7 @@ class Program_weight_tensor_parameter_197: class Program_weight_tensor_parameter_198: name = "parameter_198" + original_name = "batch_norm2d_131.w_1" shape = [120] dtype = "float32" min_val = float("-0.0596607") @@ -2189,6 +2388,7 @@ class Program_weight_tensor_parameter_198: class Program_weight_tensor_parameter_199: name = "parameter_199" + original_name = "conv2d_135.w_0" shape = [120, 120, 1, 1] dtype = "float32" min_val = float("-0.0675494") @@ -2200,6 +2400,7 @@ class Program_weight_tensor_parameter_199: class Program_weight_tensor_parameter_200: name = "parameter_200" + original_name = "batch_norm2d_130.b_0" shape = [120] dtype = "float32" min_val = float("-0.25097") @@ -2211,6 +2412,7 @@ class Program_weight_tensor_parameter_200: class Program_weight_tensor_parameter_201: name = "parameter_201" + original_name = "batch_norm2d_130.w_0" shape = [120] dtype = "float32" min_val = float("0.884184") @@ -2222,6 +2424,7 @@ class Program_weight_tensor_parameter_201: class Program_weight_tensor_parameter_202: name = "parameter_202" + original_name = "batch_norm2d_130.w_2" shape = [120] dtype = "float32" min_val = float("0.00697661") @@ -2233,6 +2436,7 @@ class Program_weight_tensor_parameter_202: class Program_weight_tensor_parameter_203: name = "parameter_203" + original_name = "batch_norm2d_130.w_1" shape = [120] dtype = "float32" min_val = float("-0.134494") @@ -2244,6 +2448,7 @@ class Program_weight_tensor_parameter_203: class Program_weight_tensor_parameter_204: name = "parameter_204" + original_name = "conv2d_134.w_0" shape = [120, 120, 3, 3] dtype = "float32" min_val = float("-0.0621776") @@ -2255,6 +2460,7 @@ class Program_weight_tensor_parameter_204: class Program_weight_tensor_parameter_205: name = "parameter_205" + original_name = "batch_norm2d_129.b_0" shape = [120] dtype = "float32" min_val = float("-0.338416") @@ -2266,6 +2472,7 @@ class Program_weight_tensor_parameter_205: class Program_weight_tensor_parameter_206: name = "parameter_206" + original_name = "batch_norm2d_129.w_0" shape = [120] dtype = "float32" min_val = float("0.776745") @@ -2277,6 +2484,7 @@ class Program_weight_tensor_parameter_206: class Program_weight_tensor_parameter_207: name = "parameter_207" + original_name = "batch_norm2d_129.w_2" shape = [120] dtype = "float32" min_val = float("0.00767903") @@ -2288,6 +2496,7 @@ class Program_weight_tensor_parameter_207: class Program_weight_tensor_parameter_208: name = "parameter_208" + original_name = "batch_norm2d_129.w_1" shape = [120] dtype = "float32" min_val = float("-0.100699") @@ -2299,6 +2508,7 @@ class Program_weight_tensor_parameter_208: class Program_weight_tensor_parameter_209: name = "parameter_209" + original_name = "conv2d_133.w_0" shape = [120, 120, 3, 3] dtype = "float32" min_val = float("-0.0822646") @@ -2310,6 +2520,7 @@ class Program_weight_tensor_parameter_209: class Program_weight_tensor_parameter_210: name = "parameter_210" + original_name = "batch_norm2d_128.b_0" shape = [120] dtype = "float32" min_val = float("-0.294204") @@ -2321,6 +2532,7 @@ class Program_weight_tensor_parameter_210: class Program_weight_tensor_parameter_211: name = "parameter_211" + original_name = "batch_norm2d_128.w_0" shape = [120] dtype = "float32" min_val = float("0.760761") @@ -2332,6 +2544,7 @@ class Program_weight_tensor_parameter_211: class Program_weight_tensor_parameter_212: name = "parameter_212" + original_name = "batch_norm2d_128.w_2" shape = [120] dtype = "float32" min_val = float("0.00615239") @@ -2343,6 +2556,7 @@ class Program_weight_tensor_parameter_212: class Program_weight_tensor_parameter_213: name = "parameter_213" + original_name = "batch_norm2d_128.w_1" shape = [120] dtype = "float32" min_val = float("-0.0841368") @@ -2354,6 +2568,7 @@ class Program_weight_tensor_parameter_213: class Program_weight_tensor_parameter_214: name = "parameter_214" + original_name = "conv2d_132.w_0" shape = [120, 120, 1, 1] dtype = "float32" min_val = float("-0.0587426") @@ -2365,6 +2580,7 @@ class Program_weight_tensor_parameter_214: class Program_weight_tensor_parameter_215: name = "parameter_215" + original_name = "batch_norm2d_127.b_0" shape = [120] dtype = "float32" min_val = float("-0.294204") @@ -2376,6 +2592,7 @@ class Program_weight_tensor_parameter_215: class Program_weight_tensor_parameter_216: name = "parameter_216" + original_name = "batch_norm2d_127.w_0" shape = [120] dtype = "float32" min_val = float("0.859114") @@ -2387,6 +2604,7 @@ class Program_weight_tensor_parameter_216: class Program_weight_tensor_parameter_217: name = "parameter_217" + original_name = "batch_norm2d_127.w_2" shape = [120] dtype = "float32" min_val = float("0.01373") @@ -2398,6 +2616,7 @@ class Program_weight_tensor_parameter_217: class Program_weight_tensor_parameter_218: name = "parameter_218" + original_name = "batch_norm2d_127.w_1" shape = [120] dtype = "float32" min_val = float("-0.161049") @@ -2409,6 +2628,7 @@ class Program_weight_tensor_parameter_218: class Program_weight_tensor_parameter_219: name = "parameter_219" + original_name = "conv2d_131.w_0" shape = [120, 120, 3, 3] dtype = "float32" min_val = float("-0.0757779") @@ -2420,6 +2640,7 @@ class Program_weight_tensor_parameter_219: class Program_weight_tensor_parameter_220: name = "parameter_220" + original_name = "batch_norm2d_126.b_0" shape = [120] dtype = "float32" min_val = float("-0.356774") @@ -2431,6 +2652,7 @@ class Program_weight_tensor_parameter_220: class Program_weight_tensor_parameter_221: name = "parameter_221" + original_name = "batch_norm2d_126.w_0" shape = [120] dtype = "float32" min_val = float("0.831744") @@ -2442,6 +2664,7 @@ class Program_weight_tensor_parameter_221: class Program_weight_tensor_parameter_222: name = "parameter_222" + original_name = "batch_norm2d_126.w_2" shape = [120] dtype = "float32" min_val = float("0.0118015") @@ -2453,6 +2676,7 @@ class Program_weight_tensor_parameter_222: class Program_weight_tensor_parameter_223: name = "parameter_223" + original_name = "batch_norm2d_126.w_1" shape = [120] dtype = "float32" min_val = float("-0.141508") @@ -2464,6 +2688,7 @@ class Program_weight_tensor_parameter_223: class Program_weight_tensor_parameter_224: name = "parameter_224" + original_name = "conv2d_130.w_0" shape = [120, 120, 3, 3] dtype = "float32" min_val = float("-0.107276") @@ -2475,6 +2700,7 @@ class Program_weight_tensor_parameter_224: class Program_weight_tensor_parameter_225: name = "parameter_225" + original_name = "batch_norm2d_125.b_0" shape = [120] dtype = "float32" min_val = float("-0.367826") @@ -2486,6 +2712,7 @@ class Program_weight_tensor_parameter_225: class Program_weight_tensor_parameter_226: name = "parameter_226" + original_name = "batch_norm2d_125.w_0" shape = [120] dtype = "float32" min_val = float("0.763084") @@ -2497,6 +2724,7 @@ class Program_weight_tensor_parameter_226: class Program_weight_tensor_parameter_227: name = "parameter_227" + original_name = "batch_norm2d_125.w_2" shape = [120] dtype = "float32" min_val = float("0.00851599") @@ -2508,6 +2736,7 @@ class Program_weight_tensor_parameter_227: class Program_weight_tensor_parameter_228: name = "parameter_228" + original_name = "batch_norm2d_125.w_1" shape = [120] dtype = "float32" min_val = float("-0.220022") @@ -2519,6 +2748,7 @@ class Program_weight_tensor_parameter_228: class Program_weight_tensor_parameter_229: name = "parameter_229" + original_name = "conv2d_129.w_0" shape = [120, 560, 1, 1] dtype = "float32" min_val = float("-0.147478") @@ -2530,6 +2760,7 @@ class Program_weight_tensor_parameter_229: class Program_weight_tensor_parameter_230: name = "parameter_230" + original_name = "batch_norm2d_124.b_0" shape = [120] dtype = "float32" min_val = float("-0.180442") @@ -2541,6 +2772,7 @@ class Program_weight_tensor_parameter_230: class Program_weight_tensor_parameter_231: name = "parameter_231" + original_name = "batch_norm2d_124.w_0" shape = [120] dtype = "float32" min_val = float("0.768106") @@ -2552,6 +2784,7 @@ class Program_weight_tensor_parameter_231: class Program_weight_tensor_parameter_232: name = "parameter_232" + original_name = "batch_norm2d_124.w_2" shape = [120] dtype = "float32" min_val = float("0.00181742") @@ -2563,6 +2796,7 @@ class Program_weight_tensor_parameter_232: class Program_weight_tensor_parameter_233: name = "parameter_233" + original_name = "batch_norm2d_124.w_1" shape = [120] dtype = "float32" min_val = float("-0.100889") @@ -2574,6 +2808,7 @@ class Program_weight_tensor_parameter_233: class Program_weight_tensor_parameter_234: name = "parameter_234" + original_name = "conv2d_128.w_0" shape = [120, 560, 1, 1] dtype = "float32" min_val = float("-0.102841") @@ -2585,6 +2820,7 @@ class Program_weight_tensor_parameter_234: class Program_weight_tensor_parameter_235: name = "parameter_235" + original_name = "batch_norm2d_123.b_0" shape = [240] dtype = "float32" min_val = float("-0.199849") @@ -2596,6 +2832,7 @@ class Program_weight_tensor_parameter_235: class Program_weight_tensor_parameter_236: name = "parameter_236" + original_name = "batch_norm2d_123.w_0" shape = [240] dtype = "float32" min_val = float("0.808675") @@ -2607,6 +2844,7 @@ class Program_weight_tensor_parameter_236: class Program_weight_tensor_parameter_237: name = "parameter_237" + original_name = "batch_norm2d_123.w_2" shape = [240] dtype = "float32" min_val = float("0.00896023") @@ -2618,6 +2856,7 @@ class Program_weight_tensor_parameter_237: class Program_weight_tensor_parameter_238: name = "parameter_238" + original_name = "batch_norm2d_123.w_1" shape = [240] dtype = "float32" min_val = float("-0.117806") @@ -2629,6 +2868,7 @@ class Program_weight_tensor_parameter_238: class Program_weight_tensor_parameter_239: name = "parameter_239" + original_name = "conv2d_127.w_0" shape = [240, 480, 1, 1] dtype = "float32" min_val = float("-0.11245") @@ -2640,6 +2880,7 @@ class Program_weight_tensor_parameter_239: class Program_weight_tensor_parameter_240: name = "parameter_240" + original_name = "batch_norm2d_122.b_0" shape = [480] dtype = "float32" min_val = float("-0.173836") @@ -2651,6 +2892,7 @@ class Program_weight_tensor_parameter_240: class Program_weight_tensor_parameter_241: name = "parameter_241" + original_name = "batch_norm2d_122.w_0" shape = [480] dtype = "float32" min_val = float("0.929581") @@ -2662,6 +2904,7 @@ class Program_weight_tensor_parameter_241: class Program_weight_tensor_parameter_242: name = "parameter_242" + original_name = "batch_norm2d_122.w_2" shape = [480] dtype = "float32" min_val = float("0.00603737") @@ -2673,6 +2916,7 @@ class Program_weight_tensor_parameter_242: class Program_weight_tensor_parameter_243: name = "parameter_243" + original_name = "batch_norm2d_122.w_1" shape = [480] dtype = "float32" min_val = float("-0.163024") @@ -2684,6 +2928,7 @@ class Program_weight_tensor_parameter_243: class Program_weight_tensor_parameter_244: name = "parameter_244" + original_name = "conv2d_126.w_0" shape = [480, 480, 1, 1] dtype = "float32" min_val = float("-0.144912") @@ -2695,6 +2940,7 @@ class Program_weight_tensor_parameter_244: class Program_weight_tensor_parameter_245: name = "parameter_245" + original_name = "batch_norm2d_121.b_0" shape = [240] dtype = "float32" min_val = float("-0.140598") @@ -2706,6 +2952,7 @@ class Program_weight_tensor_parameter_245: class Program_weight_tensor_parameter_246: name = "parameter_246" + original_name = "batch_norm2d_121.w_0" shape = [240] dtype = "float32" min_val = float("0.939344") @@ -2717,6 +2964,7 @@ class Program_weight_tensor_parameter_246: class Program_weight_tensor_parameter_247: name = "parameter_247" + original_name = "batch_norm2d_121.w_2" shape = [240] dtype = "float32" min_val = float("0.00412085") @@ -2728,6 +2976,7 @@ class Program_weight_tensor_parameter_247: class Program_weight_tensor_parameter_248: name = "parameter_248" + original_name = "batch_norm2d_121.w_1" shape = [240] dtype = "float32" min_val = float("-0.0736316") @@ -2739,6 +2988,7 @@ class Program_weight_tensor_parameter_248: class Program_weight_tensor_parameter_249: name = "parameter_249" + original_name = "conv2d_125.w_0" shape = [240, 240, 1, 1] dtype = "float32" min_val = float("-0.0318261") @@ -2750,6 +3000,7 @@ class Program_weight_tensor_parameter_249: class Program_weight_tensor_parameter_250: name = "parameter_250" + original_name = "batch_norm2d_120.b_0" shape = [240] dtype = "float32" min_val = float("-0.140598") @@ -2761,6 +3012,7 @@ class Program_weight_tensor_parameter_250: class Program_weight_tensor_parameter_251: name = "parameter_251" + original_name = "batch_norm2d_120.w_0" shape = [240] dtype = "float32" min_val = float("0.96617") @@ -2772,6 +3024,7 @@ class Program_weight_tensor_parameter_251: class Program_weight_tensor_parameter_252: name = "parameter_252" + original_name = "batch_norm2d_120.w_2" shape = [240] dtype = "float32" min_val = float("0.0148778") @@ -2783,6 +3036,7 @@ class Program_weight_tensor_parameter_252: class Program_weight_tensor_parameter_253: name = "parameter_253" + original_name = "batch_norm2d_120.w_1" shape = [240] dtype = "float32" min_val = float("-0.19403") @@ -2794,6 +3048,7 @@ class Program_weight_tensor_parameter_253: class Program_weight_tensor_parameter_254: name = "parameter_254" + original_name = "conv2d_124.w_0" shape = [240, 240, 3, 3] dtype = "float32" min_val = float("-0.0335272") @@ -2805,6 +3060,7 @@ class Program_weight_tensor_parameter_254: class Program_weight_tensor_parameter_255: name = "parameter_255" + original_name = "batch_norm2d_119.b_0" shape = [240] dtype = "float32" min_val = float("-0.110744") @@ -2816,6 +3072,7 @@ class Program_weight_tensor_parameter_255: class Program_weight_tensor_parameter_256: name = "parameter_256" + original_name = "batch_norm2d_119.w_0" shape = [240] dtype = "float32" min_val = float("0.945008") @@ -2827,6 +3084,7 @@ class Program_weight_tensor_parameter_256: class Program_weight_tensor_parameter_257: name = "parameter_257" + original_name = "batch_norm2d_119.w_2" shape = [240] dtype = "float32" min_val = float("0.0330506") @@ -2838,6 +3096,7 @@ class Program_weight_tensor_parameter_257: class Program_weight_tensor_parameter_258: name = "parameter_258" + original_name = "batch_norm2d_119.w_1" shape = [240] dtype = "float32" min_val = float("-0.29831") @@ -2849,6 +3108,7 @@ class Program_weight_tensor_parameter_258: class Program_weight_tensor_parameter_259: name = "parameter_259" + original_name = "conv2d_123.w_0" shape = [240, 240, 3, 3] dtype = "float32" min_val = float("-0.0428016") @@ -2860,6 +3120,7 @@ class Program_weight_tensor_parameter_259: class Program_weight_tensor_parameter_260: name = "parameter_260" + original_name = "batch_norm2d_118.b_0" shape = [240] dtype = "float32" min_val = float("-0.0744305") @@ -2871,6 +3132,7 @@ class Program_weight_tensor_parameter_260: class Program_weight_tensor_parameter_261: name = "parameter_261" + original_name = "batch_norm2d_118.w_0" shape = [240] dtype = "float32" min_val = float("0.972131") @@ -2882,6 +3144,7 @@ class Program_weight_tensor_parameter_261: class Program_weight_tensor_parameter_262: name = "parameter_262" + original_name = "batch_norm2d_118.w_2" shape = [240] dtype = "float32" min_val = float("0.00149103") @@ -2893,6 +3156,7 @@ class Program_weight_tensor_parameter_262: class Program_weight_tensor_parameter_263: name = "parameter_263" + original_name = "batch_norm2d_118.w_1" shape = [240] dtype = "float32" min_val = float("-0.0547325") @@ -2904,6 +3168,7 @@ class Program_weight_tensor_parameter_263: class Program_weight_tensor_parameter_264: name = "parameter_264" + original_name = "conv2d_122.w_0" shape = [240, 240, 1, 1] dtype = "float32" min_val = float("-0.0327513") @@ -2915,6 +3180,7 @@ class Program_weight_tensor_parameter_264: class Program_weight_tensor_parameter_265: name = "parameter_265" + original_name = "batch_norm2d_117.b_0" shape = [240] dtype = "float32" min_val = float("-0.0744305") @@ -2926,6 +3192,7 @@ class Program_weight_tensor_parameter_265: class Program_weight_tensor_parameter_266: name = "parameter_266" + original_name = "batch_norm2d_117.w_0" shape = [240] dtype = "float32" min_val = float("0.983458") @@ -2937,6 +3204,7 @@ class Program_weight_tensor_parameter_266: class Program_weight_tensor_parameter_267: name = "parameter_267" + original_name = "batch_norm2d_117.w_2" shape = [240] dtype = "float32" min_val = float("0.00471867") @@ -2948,6 +3216,7 @@ class Program_weight_tensor_parameter_267: class Program_weight_tensor_parameter_268: name = "parameter_268" + original_name = "batch_norm2d_117.w_1" shape = [240] dtype = "float32" min_val = float("-0.13207") @@ -2959,6 +3228,7 @@ class Program_weight_tensor_parameter_268: class Program_weight_tensor_parameter_269: name = "parameter_269" + original_name = "conv2d_121.w_0" shape = [240, 240, 3, 3] dtype = "float32" min_val = float("-0.0446192") @@ -2970,6 +3240,7 @@ class Program_weight_tensor_parameter_269: class Program_weight_tensor_parameter_270: name = "parameter_270" + original_name = "batch_norm2d_116.b_0" shape = [240] dtype = "float32" min_val = float("-0.102386") @@ -2981,6 +3252,7 @@ class Program_weight_tensor_parameter_270: class Program_weight_tensor_parameter_271: name = "parameter_271" + original_name = "batch_norm2d_116.w_0" shape = [240] dtype = "float32" min_val = float("0.959992") @@ -2992,6 +3264,7 @@ class Program_weight_tensor_parameter_271: class Program_weight_tensor_parameter_272: name = "parameter_272" + original_name = "batch_norm2d_116.w_2" shape = [240] dtype = "float32" min_val = float("0.0263733") @@ -3003,6 +3276,7 @@ class Program_weight_tensor_parameter_272: class Program_weight_tensor_parameter_273: name = "parameter_273" + original_name = "batch_norm2d_116.w_1" shape = [240] dtype = "float32" min_val = float("-0.412602") @@ -3014,6 +3288,7 @@ class Program_weight_tensor_parameter_273: class Program_weight_tensor_parameter_274: name = "parameter_274" + original_name = "conv2d_120.w_0" shape = [240, 240, 3, 3] dtype = "float32" min_val = float("-0.0312313") @@ -3025,6 +3300,7 @@ class Program_weight_tensor_parameter_274: class Program_weight_tensor_parameter_275: name = "parameter_275" + original_name = "batch_norm2d_115.b_0" shape = [240] dtype = "float32" min_val = float("-0.0765708") @@ -3036,6 +3312,7 @@ class Program_weight_tensor_parameter_275: class Program_weight_tensor_parameter_276: name = "parameter_276" + original_name = "batch_norm2d_115.w_0" shape = [240] dtype = "float32" min_val = float("0.976602") @@ -3047,6 +3324,7 @@ class Program_weight_tensor_parameter_276: class Program_weight_tensor_parameter_277: name = "parameter_277" + original_name = "batch_norm2d_115.w_2" shape = [240] dtype = "float32" min_val = float("0.00179432") @@ -3058,6 +3336,7 @@ class Program_weight_tensor_parameter_277: class Program_weight_tensor_parameter_278: name = "parameter_278" + original_name = "batch_norm2d_115.w_1" shape = [240] dtype = "float32" min_val = float("-0.0873622") @@ -3069,6 +3348,7 @@ class Program_weight_tensor_parameter_278: class Program_weight_tensor_parameter_279: name = "parameter_279" + original_name = "conv2d_119.w_0" shape = [240, 240, 1, 1] dtype = "float32" min_val = float("-0.0269182") @@ -3080,6 +3360,7 @@ class Program_weight_tensor_parameter_279: class Program_weight_tensor_parameter_280: name = "parameter_280" + original_name = "batch_norm2d_114.b_0" shape = [240] dtype = "float32" min_val = float("-0.0765708") @@ -3091,6 +3372,7 @@ class Program_weight_tensor_parameter_280: class Program_weight_tensor_parameter_281: name = "parameter_281" + original_name = "batch_norm2d_114.w_0" shape = [240] dtype = "float32" min_val = float("0.973323") @@ -3102,6 +3384,7 @@ class Program_weight_tensor_parameter_281: class Program_weight_tensor_parameter_282: name = "parameter_282" + original_name = "batch_norm2d_114.w_2" shape = [240] dtype = "float32" min_val = float("0.00674589") @@ -3113,6 +3396,7 @@ class Program_weight_tensor_parameter_282: class Program_weight_tensor_parameter_283: name = "parameter_283" + original_name = "batch_norm2d_114.w_1" shape = [240] dtype = "float32" min_val = float("-0.159769") @@ -3124,6 +3408,7 @@ class Program_weight_tensor_parameter_283: class Program_weight_tensor_parameter_284: name = "parameter_284" + original_name = "conv2d_118.w_0" shape = [240, 240, 3, 3] dtype = "float32" min_val = float("-0.0183669") @@ -3135,6 +3420,7 @@ class Program_weight_tensor_parameter_284: class Program_weight_tensor_parameter_285: name = "parameter_285" + original_name = "batch_norm2d_113.b_0" shape = [240] dtype = "float32" min_val = float("-0.101821") @@ -3146,6 +3432,7 @@ class Program_weight_tensor_parameter_285: class Program_weight_tensor_parameter_286: name = "parameter_286" + original_name = "batch_norm2d_113.w_0" shape = [240] dtype = "float32" min_val = float("0.962505") @@ -3157,6 +3444,7 @@ class Program_weight_tensor_parameter_286: class Program_weight_tensor_parameter_287: name = "parameter_287" + original_name = "batch_norm2d_113.w_2" shape = [240] dtype = "float32" min_val = float("0.0241389") @@ -3168,6 +3456,7 @@ class Program_weight_tensor_parameter_287: class Program_weight_tensor_parameter_288: name = "parameter_288" + original_name = "batch_norm2d_113.w_1" shape = [240] dtype = "float32" min_val = float("-0.467356") @@ -3179,6 +3468,7 @@ class Program_weight_tensor_parameter_288: class Program_weight_tensor_parameter_289: name = "parameter_289" + original_name = "conv2d_117.w_0" shape = [240, 240, 3, 3] dtype = "float32" min_val = float("-0.0219161") @@ -3190,6 +3480,7 @@ class Program_weight_tensor_parameter_289: class Program_weight_tensor_parameter_290: name = "parameter_290" + original_name = "batch_norm2d_112.b_0" shape = [240] dtype = "float32" min_val = float("-0.0842972") @@ -3201,6 +3492,7 @@ class Program_weight_tensor_parameter_290: class Program_weight_tensor_parameter_291: name = "parameter_291" + original_name = "batch_norm2d_112.w_0" shape = [240] dtype = "float32" min_val = float("0.97936") @@ -3212,6 +3504,7 @@ class Program_weight_tensor_parameter_291: class Program_weight_tensor_parameter_292: name = "parameter_292" + original_name = "batch_norm2d_112.w_2" shape = [240] dtype = "float32" min_val = float("0.00164535") @@ -3223,6 +3516,7 @@ class Program_weight_tensor_parameter_292: class Program_weight_tensor_parameter_293: name = "parameter_293" + original_name = "batch_norm2d_112.w_1" shape = [240] dtype = "float32" min_val = float("-0.0716032") @@ -3234,6 +3528,7 @@ class Program_weight_tensor_parameter_293: class Program_weight_tensor_parameter_294: name = "parameter_294" + original_name = "conv2d_116.w_0" shape = [240, 240, 1, 1] dtype = "float32" min_val = float("-0.0295145") @@ -3245,6 +3540,7 @@ class Program_weight_tensor_parameter_294: class Program_weight_tensor_parameter_295: name = "parameter_295" + original_name = "batch_norm2d_111.b_0" shape = [240] dtype = "float32" min_val = float("-0.0842972") @@ -3256,6 +3552,7 @@ class Program_weight_tensor_parameter_295: class Program_weight_tensor_parameter_296: name = "parameter_296" + original_name = "batch_norm2d_111.w_0" shape = [240] dtype = "float32" min_val = float("0.965574") @@ -3267,6 +3564,7 @@ class Program_weight_tensor_parameter_296: class Program_weight_tensor_parameter_297: name = "parameter_297" + original_name = "batch_norm2d_111.w_2" shape = [240] dtype = "float32" min_val = float("0.0074112") @@ -3278,6 +3576,7 @@ class Program_weight_tensor_parameter_297: class Program_weight_tensor_parameter_298: name = "parameter_298" + original_name = "batch_norm2d_111.w_1" shape = [240] dtype = "float32" min_val = float("-0.289886") @@ -3289,6 +3588,7 @@ class Program_weight_tensor_parameter_298: class Program_weight_tensor_parameter_299: name = "parameter_299" + original_name = "conv2d_115.w_0" shape = [240, 240, 3, 3] dtype = "float32" min_val = float("-0.0205579") @@ -3300,6 +3600,7 @@ class Program_weight_tensor_parameter_299: class Program_weight_tensor_parameter_300: name = "parameter_300" + original_name = "batch_norm2d_110.b_0" shape = [240] dtype = "float32" min_val = float("-0.126784") @@ -3311,6 +3612,7 @@ class Program_weight_tensor_parameter_300: class Program_weight_tensor_parameter_301: name = "parameter_301" + original_name = "batch_norm2d_110.w_0" shape = [240] dtype = "float32" min_val = float("0.96645") @@ -3322,6 +3624,7 @@ class Program_weight_tensor_parameter_301: class Program_weight_tensor_parameter_302: name = "parameter_302" + original_name = "batch_norm2d_110.w_2" shape = [240] dtype = "float32" min_val = float("0.0094598") @@ -3333,6 +3636,7 @@ class Program_weight_tensor_parameter_302: class Program_weight_tensor_parameter_303: name = "parameter_303" + original_name = "batch_norm2d_110.w_1" shape = [240] dtype = "float32" min_val = float("-0.162028") @@ -3344,6 +3648,7 @@ class Program_weight_tensor_parameter_303: class Program_weight_tensor_parameter_304: name = "parameter_304" + original_name = "conv2d_114.w_0" shape = [240, 240, 3, 3] dtype = "float32" min_val = float("-0.0431933") @@ -3355,6 +3660,7 @@ class Program_weight_tensor_parameter_304: class Program_weight_tensor_parameter_305: name = "parameter_305" + original_name = "batch_norm2d_109.b_0" shape = [240] dtype = "float32" min_val = float("-0.145043") @@ -3366,6 +3672,7 @@ class Program_weight_tensor_parameter_305: class Program_weight_tensor_parameter_306: name = "parameter_306" + original_name = "batch_norm2d_109.w_0" shape = [240] dtype = "float32" min_val = float("0.941988") @@ -3377,6 +3684,7 @@ class Program_weight_tensor_parameter_306: class Program_weight_tensor_parameter_307: name = "parameter_307" + original_name = "batch_norm2d_109.w_2" shape = [240] dtype = "float32" min_val = float("0.00340711") @@ -3388,6 +3696,7 @@ class Program_weight_tensor_parameter_307: class Program_weight_tensor_parameter_308: name = "parameter_308" + original_name = "batch_norm2d_109.w_1" shape = [240] dtype = "float32" min_val = float("-0.112474") @@ -3399,6 +3708,7 @@ class Program_weight_tensor_parameter_308: class Program_weight_tensor_parameter_309: name = "parameter_309" + original_name = "conv2d_113.w_0" shape = [240, 1120, 1, 1] dtype = "float32" min_val = float("-0.0800266") @@ -3410,6 +3720,7 @@ class Program_weight_tensor_parameter_309: class Program_weight_tensor_parameter_310: name = "parameter_310" + original_name = "batch_norm2d_108.b_0" shape = [240] dtype = "float32" min_val = float("-0.11267") @@ -3421,6 +3732,7 @@ class Program_weight_tensor_parameter_310: class Program_weight_tensor_parameter_311: name = "parameter_311" + original_name = "batch_norm2d_108.w_0" shape = [240] dtype = "float32" min_val = float("0.945604") @@ -3432,6 +3744,7 @@ class Program_weight_tensor_parameter_311: class Program_weight_tensor_parameter_312: name = "parameter_312" + original_name = "batch_norm2d_108.w_2" shape = [240] dtype = "float32" min_val = float("0.00442363") @@ -3443,6 +3756,7 @@ class Program_weight_tensor_parameter_312: class Program_weight_tensor_parameter_313: name = "parameter_313" + original_name = "batch_norm2d_108.w_1" shape = [240] dtype = "float32" min_val = float("-0.0798689") @@ -3454,6 +3768,7 @@ class Program_weight_tensor_parameter_313: class Program_weight_tensor_parameter_314: name = "parameter_314" + original_name = "conv2d_112.w_0" shape = [240, 1120, 1, 1] dtype = "float32" min_val = float("-0.183825") @@ -3465,6 +3780,7 @@ class Program_weight_tensor_parameter_314: class Program_weight_tensor_parameter_315: name = "parameter_315" + original_name = "batch_norm2d_107.b_0" shape = [480] dtype = "float32" min_val = float("-0.161511") @@ -3476,6 +3792,7 @@ class Program_weight_tensor_parameter_315: class Program_weight_tensor_parameter_316: name = "parameter_316" + original_name = "batch_norm2d_107.w_0" shape = [480] dtype = "float32" min_val = float("0.885613") @@ -3487,6 +3804,7 @@ class Program_weight_tensor_parameter_316: class Program_weight_tensor_parameter_317: name = "parameter_317" + original_name = "batch_norm2d_107.w_2" shape = [480] dtype = "float32" min_val = float("0.0113948") @@ -3498,6 +3816,7 @@ class Program_weight_tensor_parameter_317: class Program_weight_tensor_parameter_318: name = "parameter_318" + original_name = "batch_norm2d_107.w_1" shape = [480] dtype = "float32" min_val = float("-0.13326") @@ -3509,6 +3828,7 @@ class Program_weight_tensor_parameter_318: class Program_weight_tensor_parameter_319: name = "parameter_319" + original_name = "conv2d_111.w_0" shape = [480, 960, 1, 1] dtype = "float32" min_val = float("-0.0322665") @@ -3520,6 +3840,7 @@ class Program_weight_tensor_parameter_319: class Program_weight_tensor_parameter_320: name = "parameter_320" + original_name = "batch_norm2d_106.b_0" shape = [960] dtype = "float32" min_val = float("-0.057346") @@ -3531,6 +3852,7 @@ class Program_weight_tensor_parameter_320: class Program_weight_tensor_parameter_321: name = "parameter_321" + original_name = "batch_norm2d_106.w_0" shape = [960] dtype = "float32" min_val = float("0.965554") @@ -3542,6 +3864,7 @@ class Program_weight_tensor_parameter_321: class Program_weight_tensor_parameter_322: name = "parameter_322" + original_name = "batch_norm2d_106.w_2" shape = [960] dtype = "float32" min_val = float("0.00444446") @@ -3553,6 +3876,7 @@ class Program_weight_tensor_parameter_322: class Program_weight_tensor_parameter_323: name = "parameter_323" + original_name = "batch_norm2d_106.w_1" shape = [960] dtype = "float32" min_val = float("-0.11557") @@ -3564,6 +3888,7 @@ class Program_weight_tensor_parameter_323: class Program_weight_tensor_parameter_324: name = "parameter_324" + original_name = "conv2d_110.w_0" shape = [960, 960, 1, 1] dtype = "float32" min_val = float("-0.0401595") @@ -3575,6 +3900,7 @@ class Program_weight_tensor_parameter_324: class Program_weight_tensor_parameter_325: name = "parameter_325" + original_name = "batch_norm2d_105.b_0" shape = [480] dtype = "float32" min_val = float("-0.0754008") @@ -3586,6 +3912,7 @@ class Program_weight_tensor_parameter_325: class Program_weight_tensor_parameter_326: name = "parameter_326" + original_name = "batch_norm2d_105.w_0" shape = [480] dtype = "float32" min_val = float("0.962295") @@ -3597,6 +3924,7 @@ class Program_weight_tensor_parameter_326: class Program_weight_tensor_parameter_327: name = "parameter_327" + original_name = "batch_norm2d_105.w_2" shape = [480] dtype = "float32" min_val = float("0.00316057") @@ -3608,6 +3936,7 @@ class Program_weight_tensor_parameter_327: class Program_weight_tensor_parameter_328: name = "parameter_328" + original_name = "batch_norm2d_105.w_1" shape = [480] dtype = "float32" min_val = float("-0.0536651") @@ -3619,6 +3948,7 @@ class Program_weight_tensor_parameter_328: class Program_weight_tensor_parameter_329: name = "parameter_329" + original_name = "conv2d_109.w_0" shape = [480, 480, 1, 1] dtype = "float32" min_val = float("-0.0275931") @@ -3630,6 +3960,7 @@ class Program_weight_tensor_parameter_329: class Program_weight_tensor_parameter_330: name = "parameter_330" + original_name = "batch_norm2d_104.b_0" shape = [480] dtype = "float32" min_val = float("-0.0754008") @@ -3641,6 +3972,7 @@ class Program_weight_tensor_parameter_330: class Program_weight_tensor_parameter_331: name = "parameter_331" + original_name = "batch_norm2d_104.w_0" shape = [480] dtype = "float32" min_val = float("0.964382") @@ -3652,6 +3984,7 @@ class Program_weight_tensor_parameter_331: class Program_weight_tensor_parameter_332: name = "parameter_332" + original_name = "batch_norm2d_104.w_2" shape = [480] dtype = "float32" min_val = float("0.0241322") @@ -3663,6 +3996,7 @@ class Program_weight_tensor_parameter_332: class Program_weight_tensor_parameter_333: name = "parameter_333" + original_name = "batch_norm2d_104.w_1" shape = [480] dtype = "float32" min_val = float("-0.254908") @@ -3674,6 +4008,7 @@ class Program_weight_tensor_parameter_333: class Program_weight_tensor_parameter_334: name = "parameter_334" + original_name = "conv2d_108.w_0" shape = [480, 480, 3, 3] dtype = "float32" min_val = float("-0.0119617") @@ -3685,6 +4020,7 @@ class Program_weight_tensor_parameter_334: class Program_weight_tensor_parameter_335: name = "parameter_335" + original_name = "batch_norm2d_103.b_0" shape = [480] dtype = "float32" min_val = float("-0.0246241") @@ -3696,6 +4032,7 @@ class Program_weight_tensor_parameter_335: class Program_weight_tensor_parameter_336: name = "parameter_336" + original_name = "batch_norm2d_103.w_0" shape = [480] dtype = "float32" min_val = float("0.977198") @@ -3707,6 +4044,7 @@ class Program_weight_tensor_parameter_336: class Program_weight_tensor_parameter_337: name = "parameter_337" + original_name = "batch_norm2d_103.w_2" shape = [480] dtype = "float32" min_val = float("0.0317041") @@ -3718,6 +4056,7 @@ class Program_weight_tensor_parameter_337: class Program_weight_tensor_parameter_338: name = "parameter_338" + original_name = "batch_norm2d_103.w_1" shape = [480] dtype = "float32" min_val = float("-0.379924") @@ -3729,6 +4068,7 @@ class Program_weight_tensor_parameter_338: class Program_weight_tensor_parameter_339: name = "parameter_339" + original_name = "conv2d_107.w_0" shape = [480, 480, 3, 3] dtype = "float32" min_val = float("-0.0186869") @@ -3740,6 +4080,7 @@ class Program_weight_tensor_parameter_339: class Program_weight_tensor_parameter_340: name = "parameter_340" + original_name = "batch_norm2d_102.b_0" shape = [480] dtype = "float32" min_val = float("-0.0170742") @@ -3751,6 +4092,7 @@ class Program_weight_tensor_parameter_340: class Program_weight_tensor_parameter_341: name = "parameter_341" + original_name = "batch_norm2d_102.w_0" shape = [480] dtype = "float32" min_val = float("0.988226") @@ -3762,6 +4104,7 @@ class Program_weight_tensor_parameter_341: class Program_weight_tensor_parameter_342: name = "parameter_342" + original_name = "batch_norm2d_102.w_2" shape = [480] dtype = "float32" min_val = float("0.000868111") @@ -3773,6 +4116,7 @@ class Program_weight_tensor_parameter_342: class Program_weight_tensor_parameter_343: name = "parameter_343" + original_name = "batch_norm2d_102.w_1" shape = [480] dtype = "float32" min_val = float("-0.0376301") @@ -3784,6 +4128,7 @@ class Program_weight_tensor_parameter_343: class Program_weight_tensor_parameter_344: name = "parameter_344" + original_name = "conv2d_106.w_0" shape = [480, 480, 1, 1] dtype = "float32" min_val = float("-0.0296631") @@ -3795,6 +4140,7 @@ class Program_weight_tensor_parameter_344: class Program_weight_tensor_parameter_345: name = "parameter_345" + original_name = "batch_norm2d_101.b_0" shape = [480] dtype = "float32" min_val = float("-0.0170742") @@ -3806,6 +4152,7 @@ class Program_weight_tensor_parameter_345: class Program_weight_tensor_parameter_346: name = "parameter_346" + original_name = "batch_norm2d_101.w_0" shape = [480] dtype = "float32" min_val = float("0.990015") @@ -3817,6 +4164,7 @@ class Program_weight_tensor_parameter_346: class Program_weight_tensor_parameter_347: name = "parameter_347" + original_name = "batch_norm2d_101.w_2" shape = [480] dtype = "float32" min_val = float("0.00538322") @@ -3828,6 +4176,7 @@ class Program_weight_tensor_parameter_347: class Program_weight_tensor_parameter_348: name = "parameter_348" + original_name = "batch_norm2d_101.w_1" shape = [480] dtype = "float32" min_val = float("-0.130153") @@ -3839,6 +4188,7 @@ class Program_weight_tensor_parameter_348: class Program_weight_tensor_parameter_349: name = "parameter_349" + original_name = "conv2d_105.w_0" shape = [480, 480, 3, 3] dtype = "float32" min_val = float("-0.0161702") @@ -3850,6 +4200,7 @@ class Program_weight_tensor_parameter_349: class Program_weight_tensor_parameter_350: name = "parameter_350" + original_name = "batch_norm2d_100.b_0" shape = [480] dtype = "float32" min_val = float("-0.0229037") @@ -3861,6 +4212,7 @@ class Program_weight_tensor_parameter_350: class Program_weight_tensor_parameter_351: name = "parameter_351" + original_name = "batch_norm2d_100.w_0" shape = [480] dtype = "float32" min_val = float("0.983357") @@ -3872,6 +4224,7 @@ class Program_weight_tensor_parameter_351: class Program_weight_tensor_parameter_352: name = "parameter_352" + original_name = "batch_norm2d_100.w_2" shape = [480] dtype = "float32" min_val = float("0.00804034") @@ -3883,6 +4236,7 @@ class Program_weight_tensor_parameter_352: class Program_weight_tensor_parameter_353: name = "parameter_353" + original_name = "batch_norm2d_100.w_1" shape = [480] dtype = "float32" min_val = float("-0.136793") @@ -3894,6 +4248,7 @@ class Program_weight_tensor_parameter_353: class Program_weight_tensor_parameter_354: name = "parameter_354" + original_name = "conv2d_104.w_0" shape = [480, 480, 3, 3] dtype = "float32" min_val = float("-0.0211452") @@ -3905,6 +4260,7 @@ class Program_weight_tensor_parameter_354: class Program_weight_tensor_parameter_355: name = "parameter_355" + original_name = "batch_norm2d_99.b_0" shape = [480] dtype = "float32" min_val = float("-0.033632") @@ -3916,6 +4272,7 @@ class Program_weight_tensor_parameter_355: class Program_weight_tensor_parameter_356: name = "parameter_356" + original_name = "batch_norm2d_99.w_0" shape = [480] dtype = "float32" min_val = float("0.978987") @@ -3927,6 +4284,7 @@ class Program_weight_tensor_parameter_356: class Program_weight_tensor_parameter_357: name = "parameter_357" + original_name = "batch_norm2d_99.w_2" shape = [480] dtype = "float32" min_val = float("0.0274216") @@ -3938,6 +4296,7 @@ class Program_weight_tensor_parameter_357: class Program_weight_tensor_parameter_358: name = "parameter_358" + original_name = "batch_norm2d_99.w_1" shape = [480] dtype = "float32" min_val = float("-1.60383") @@ -3949,6 +4308,7 @@ class Program_weight_tensor_parameter_358: class Program_weight_tensor_parameter_359: name = "parameter_359" + original_name = "conv2d_103.w_0" shape = [480, 1920, 1, 1] dtype = "float32" min_val = float("-0.0392021") @@ -3960,6 +4320,7 @@ class Program_weight_tensor_parameter_359: class Program_weight_tensor_parameter_360: name = "parameter_360" + original_name = "batch_norm2d_98.b_0" shape = [480] dtype = "float32" min_val = float("-0.00675225") @@ -3971,6 +4332,7 @@ class Program_weight_tensor_parameter_360: class Program_weight_tensor_parameter_361: name = "parameter_361" + original_name = "batch_norm2d_98.w_0" shape = [480] dtype = "float32" min_val = float("0.986417") @@ -3982,6 +4344,7 @@ class Program_weight_tensor_parameter_361: class Program_weight_tensor_parameter_362: name = "parameter_362" + original_name = "batch_norm2d_98.w_2" shape = [480] dtype = "float32" min_val = float("0.00195468") @@ -3993,6 +4356,7 @@ class Program_weight_tensor_parameter_362: class Program_weight_tensor_parameter_363: name = "parameter_363" + original_name = "batch_norm2d_98.w_1" shape = [480] dtype = "float32" min_val = float("-0.0613123") @@ -4004,6 +4368,7 @@ class Program_weight_tensor_parameter_363: class Program_weight_tensor_parameter_364: name = "parameter_364" + original_name = "conv2d_102.w_0" shape = [480, 480, 1, 1] dtype = "float32" min_val = float("-0.0206033") @@ -4015,6 +4380,7 @@ class Program_weight_tensor_parameter_364: class Program_weight_tensor_parameter_365: name = "parameter_365" + original_name = "batch_norm2d_97.b_0" shape = [480] dtype = "float32" min_val = float("-0.00675224") @@ -4026,6 +4392,7 @@ class Program_weight_tensor_parameter_365: class Program_weight_tensor_parameter_366: name = "parameter_366" + original_name = "batch_norm2d_97.w_0" shape = [480] dtype = "float32" min_val = float("0.983756") @@ -4037,6 +4404,7 @@ class Program_weight_tensor_parameter_366: class Program_weight_tensor_parameter_367: name = "parameter_367" + original_name = "batch_norm2d_97.w_2" shape = [480] dtype = "float32" min_val = float("0.00761751") @@ -4048,6 +4416,7 @@ class Program_weight_tensor_parameter_367: class Program_weight_tensor_parameter_368: name = "parameter_368" + original_name = "batch_norm2d_97.w_1" shape = [480] dtype = "float32" min_val = float("-0.249767") @@ -4059,6 +4428,7 @@ class Program_weight_tensor_parameter_368: class Program_weight_tensor_parameter_369: name = "parameter_369" + original_name = "conv2d_101.w_0" shape = [480, 480, 3, 3] dtype = "float32" min_val = float("-0.0168898") @@ -4070,6 +4440,7 @@ class Program_weight_tensor_parameter_369: class Program_weight_tensor_parameter_370: name = "parameter_370" + original_name = "batch_norm2d_96.b_0" shape = [480] dtype = "float32" min_val = float("-0.014117") @@ -4081,6 +4452,7 @@ class Program_weight_tensor_parameter_370: class Program_weight_tensor_parameter_371: name = "parameter_371" + original_name = "batch_norm2d_96.w_0" shape = [480] dtype = "float32" min_val = float("0.982265") @@ -4092,6 +4464,7 @@ class Program_weight_tensor_parameter_371: class Program_weight_tensor_parameter_372: name = "parameter_372" + original_name = "batch_norm2d_96.w_2" shape = [480] dtype = "float32" min_val = float("0.0352619") @@ -4103,6 +4476,7 @@ class Program_weight_tensor_parameter_372: class Program_weight_tensor_parameter_373: name = "parameter_373" + original_name = "batch_norm2d_96.w_1" shape = [480] dtype = "float32" min_val = float("-0.725492") @@ -4114,6 +4488,7 @@ class Program_weight_tensor_parameter_373: class Program_weight_tensor_parameter_374: name = "parameter_374" + original_name = "conv2d_100.w_0" shape = [480, 480, 3, 3] dtype = "float32" min_val = float("-0.0128393") @@ -4125,6 +4500,7 @@ class Program_weight_tensor_parameter_374: class Program_weight_tensor_parameter_375: name = "parameter_375" + original_name = "batch_norm2d_95.b_0" shape = [480] dtype = "float32" min_val = float("-0.0112956") @@ -4136,6 +4512,7 @@ class Program_weight_tensor_parameter_375: class Program_weight_tensor_parameter_376: name = "parameter_376" + original_name = "batch_norm2d_95.w_0" shape = [480] dtype = "float32" min_val = float("0.991803") @@ -4147,6 +4524,7 @@ class Program_weight_tensor_parameter_376: class Program_weight_tensor_parameter_377: name = "parameter_377" + original_name = "batch_norm2d_95.w_2" shape = [480] dtype = "float32" min_val = float("0.000857574") @@ -4158,6 +4536,7 @@ class Program_weight_tensor_parameter_377: class Program_weight_tensor_parameter_378: name = "parameter_378" + original_name = "batch_norm2d_95.w_1" shape = [480] dtype = "float32" min_val = float("-0.0515057") @@ -4169,6 +4548,7 @@ class Program_weight_tensor_parameter_378: class Program_weight_tensor_parameter_379: name = "parameter_379" + original_name = "conv2d_99.w_0" shape = [480, 480, 1, 1] dtype = "float32" min_val = float("-0.0112608") @@ -4180,6 +4560,7 @@ class Program_weight_tensor_parameter_379: class Program_weight_tensor_parameter_380: name = "parameter_380" + original_name = "batch_norm2d_94.b_0" shape = [480] dtype = "float32" min_val = float("-0.0112956") @@ -4191,6 +4572,7 @@ class Program_weight_tensor_parameter_380: class Program_weight_tensor_parameter_381: name = "parameter_381" + original_name = "batch_norm2d_94.w_0" shape = [480] dtype = "float32" min_val = float("0.988823") @@ -4202,6 +4584,7 @@ class Program_weight_tensor_parameter_381: class Program_weight_tensor_parameter_382: name = "parameter_382" + original_name = "batch_norm2d_94.w_2" shape = [480] dtype = "float32" min_val = float("0.00409502") @@ -4213,6 +4596,7 @@ class Program_weight_tensor_parameter_382: class Program_weight_tensor_parameter_383: name = "parameter_383" + original_name = "batch_norm2d_94.w_1" shape = [480] dtype = "float32" min_val = float("-0.145659") @@ -4224,6 +4608,7 @@ class Program_weight_tensor_parameter_383: class Program_weight_tensor_parameter_384: name = "parameter_384" + original_name = "conv2d_98.w_0" shape = [480, 480, 3, 3] dtype = "float32" min_val = float("-0.00630496") @@ -4235,6 +4620,7 @@ class Program_weight_tensor_parameter_384: class Program_weight_tensor_parameter_385: name = "parameter_385" + original_name = "batch_norm2d_93.b_0" shape = [480] dtype = "float32" min_val = float("-0.0161028") @@ -4246,6 +4632,7 @@ class Program_weight_tensor_parameter_385: class Program_weight_tensor_parameter_386: name = "parameter_386" + original_name = "batch_norm2d_93.w_0" shape = [480] dtype = "float32" min_val = float("0.985544") @@ -4257,6 +4644,7 @@ class Program_weight_tensor_parameter_386: class Program_weight_tensor_parameter_387: name = "parameter_387" + original_name = "batch_norm2d_93.w_2" shape = [480] dtype = "float32" min_val = float("0.00665507") @@ -4268,6 +4656,7 @@ class Program_weight_tensor_parameter_387: class Program_weight_tensor_parameter_388: name = "parameter_388" + original_name = "batch_norm2d_93.w_1" shape = [480] dtype = "float32" min_val = float("-0.132361") @@ -4279,6 +4668,7 @@ class Program_weight_tensor_parameter_388: class Program_weight_tensor_parameter_389: name = "parameter_389" + original_name = "conv2d_97.w_0" shape = [480, 480, 3, 3] dtype = "float32" min_val = float("-0.010056") @@ -4290,6 +4680,7 @@ class Program_weight_tensor_parameter_389: class Program_weight_tensor_parameter_390: name = "parameter_390" + original_name = "batch_norm2d_92.b_0" shape = [480] dtype = "float32" min_val = float("-0.0271872") @@ -4301,6 +4692,7 @@ class Program_weight_tensor_parameter_390: class Program_weight_tensor_parameter_391: name = "parameter_391" + original_name = "batch_norm2d_92.w_0" shape = [480] dtype = "float32" min_val = float("0.990909") @@ -4312,6 +4704,7 @@ class Program_weight_tensor_parameter_391: class Program_weight_tensor_parameter_392: name = "parameter_392" + original_name = "batch_norm2d_92.w_2" shape = [480] dtype = "float32" min_val = float("0.00662934") @@ -4323,6 +4716,7 @@ class Program_weight_tensor_parameter_392: class Program_weight_tensor_parameter_393: name = "parameter_393" + original_name = "batch_norm2d_92.w_1" shape = [480] dtype = "float32" min_val = float("-0.106539") @@ -4334,6 +4728,7 @@ class Program_weight_tensor_parameter_393: class Program_weight_tensor_parameter_394: name = "parameter_394" + original_name = "conv2d_96.w_0" shape = [480, 1280, 1, 1] dtype = "float32" min_val = float("-0.0229104") @@ -4345,6 +4740,7 @@ class Program_weight_tensor_parameter_394: class Program_weight_tensor_parameter_395: name = "parameter_395" + original_name = "batch_norm2d_91.b_0" shape = [480] dtype = "float32" min_val = float("-0.0277443") @@ -4356,6 +4752,7 @@ class Program_weight_tensor_parameter_395: class Program_weight_tensor_parameter_396: name = "parameter_396" + original_name = "batch_norm2d_91.w_0" shape = [480] dtype = "float32" min_val = float("0.998188") @@ -4367,6 +4764,7 @@ class Program_weight_tensor_parameter_396: class Program_weight_tensor_parameter_397: name = "parameter_397" + original_name = "batch_norm2d_91.w_2" shape = [480] dtype = "float32" min_val = float("0.0144049") @@ -4378,6 +4776,7 @@ class Program_weight_tensor_parameter_397: class Program_weight_tensor_parameter_398: name = "parameter_398" + original_name = "batch_norm2d_91.w_1" shape = [480] dtype = "float32" min_val = float("-0.103768") @@ -4389,6 +4788,7 @@ class Program_weight_tensor_parameter_398: class Program_weight_tensor_parameter_399: name = "parameter_399" + original_name = "conv2d_95.w_0" shape = [480, 1280, 1, 1] dtype = "float32" min_val = float("-0.0647656") @@ -4400,6 +4800,7 @@ class Program_weight_tensor_parameter_399: class Program_weight_tensor_parameter_400: name = "parameter_400" + original_name = "batch_norm2d_90.b_0" shape = [1280] dtype = "float32" min_val = float("-3.72876") @@ -4411,6 +4812,7 @@ class Program_weight_tensor_parameter_400: class Program_weight_tensor_parameter_401: name = "parameter_401" + original_name = "batch_norm2d_90.w_0" shape = [1280] dtype = "float32" min_val = float("1.31833") @@ -4422,6 +4824,7 @@ class Program_weight_tensor_parameter_401: class Program_weight_tensor_parameter_402: name = "parameter_402" + original_name = "batch_norm2d_90.w_2" shape = [1280] dtype = "float32" min_val = float("0.00397753") @@ -4433,6 +4836,7 @@ class Program_weight_tensor_parameter_402: class Program_weight_tensor_parameter_403: name = "parameter_403" + original_name = "batch_norm2d_90.w_1" shape = [1280] dtype = "float32" min_val = float("-0.135825") @@ -4444,6 +4848,7 @@ class Program_weight_tensor_parameter_403: class Program_weight_tensor_parameter_404: name = "parameter_404" + original_name = "conv2d_94.w_0" shape = [1280, 960, 1, 1] dtype = "float32" min_val = float("-0.0744834") @@ -4455,6 +4860,7 @@ class Program_weight_tensor_parameter_404: class Program_weight_tensor_parameter_405: name = "parameter_405" + original_name = "conv2d_93.b_0" shape = [960] dtype = "float32" min_val = float("-0.0100121") @@ -4466,6 +4872,7 @@ class Program_weight_tensor_parameter_405: class Program_weight_tensor_parameter_406: name = "parameter_406" + original_name = "conv2d_93.w_0" shape = [960, 960, 1, 1] dtype = "float32" min_val = float("-0.1228") @@ -4477,6 +4884,7 @@ class Program_weight_tensor_parameter_406: class Program_weight_tensor_parameter_407: name = "parameter_407" + original_name = "batch_norm2d_89.b_0" shape = [480] dtype = "float32" min_val = float("-1.62891") @@ -4488,6 +4896,7 @@ class Program_weight_tensor_parameter_407: class Program_weight_tensor_parameter_408: name = "parameter_408" + original_name = "batch_norm2d_89.w_0" shape = [480] dtype = "float32" min_val = float("0.163674") @@ -4499,6 +4908,7 @@ class Program_weight_tensor_parameter_408: class Program_weight_tensor_parameter_409: name = "parameter_409" + original_name = "batch_norm2d_89.w_2" shape = [480] dtype = "float32" min_val = float("4.76442e-05") @@ -4510,6 +4920,7 @@ class Program_weight_tensor_parameter_409: class Program_weight_tensor_parameter_410: name = "parameter_410" + original_name = "batch_norm2d_89.w_1" shape = [480] dtype = "float32" min_val = float("-0.0846148") @@ -4521,6 +4932,7 @@ class Program_weight_tensor_parameter_410: class Program_weight_tensor_parameter_411: name = "parameter_411" + original_name = "conv2d_92.w_0" shape = [480, 480, 1, 1] dtype = "float32" min_val = float("-0.0201921") @@ -4532,6 +4944,7 @@ class Program_weight_tensor_parameter_411: class Program_weight_tensor_parameter_412: name = "parameter_412" + original_name = "batch_norm2d_88.b_0" shape = [480] dtype = "float32" min_val = float("-1.62891") @@ -4543,6 +4956,7 @@ class Program_weight_tensor_parameter_412: class Program_weight_tensor_parameter_413: name = "parameter_413" + original_name = "batch_norm2d_88.w_0" shape = [480] dtype = "float32" min_val = float("0.317358") @@ -4554,6 +4968,7 @@ class Program_weight_tensor_parameter_413: class Program_weight_tensor_parameter_414: name = "parameter_414" + original_name = "batch_norm2d_88.w_2" shape = [480] dtype = "float32" min_val = float("0.000403854") @@ -4565,6 +4980,7 @@ class Program_weight_tensor_parameter_414: class Program_weight_tensor_parameter_415: name = "parameter_415" + original_name = "batch_norm2d_88.w_1" shape = [480] dtype = "float32" min_val = float("-0.112821") @@ -4576,6 +4992,7 @@ class Program_weight_tensor_parameter_415: class Program_weight_tensor_parameter_416: name = "parameter_416" + original_name = "conv2d_91.w_0" shape = [480, 480, 3, 3] dtype = "float32" min_val = float("-0.0248061") @@ -4587,6 +5004,7 @@ class Program_weight_tensor_parameter_416: class Program_weight_tensor_parameter_417: name = "parameter_417" + original_name = "batch_norm2d_87.b_0" shape = [480] dtype = "float32" min_val = float("-2.36006") @@ -4598,6 +5016,7 @@ class Program_weight_tensor_parameter_417: class Program_weight_tensor_parameter_418: name = "parameter_418" + original_name = "batch_norm2d_87.w_0" shape = [480] dtype = "float32" min_val = float("0.584352") @@ -4609,6 +5028,7 @@ class Program_weight_tensor_parameter_418: class Program_weight_tensor_parameter_419: name = "parameter_419" + original_name = "batch_norm2d_87.w_2" shape = [480] dtype = "float32" min_val = float("0.0546498") @@ -4620,6 +5040,7 @@ class Program_weight_tensor_parameter_419: class Program_weight_tensor_parameter_420: name = "parameter_420" + original_name = "batch_norm2d_87.w_1" shape = [480] dtype = "float32" min_val = float("-0.97414") @@ -4631,6 +5052,7 @@ class Program_weight_tensor_parameter_420: class Program_weight_tensor_parameter_421: name = "parameter_421" + original_name = "conv2d_90.w_0" shape = [480, 480, 3, 3] dtype = "float32" min_val = float("-0.0218771") @@ -4642,6 +5064,7 @@ class Program_weight_tensor_parameter_421: class Program_weight_tensor_parameter_422: name = "parameter_422" + original_name = "batch_norm2d_86.b_0" shape = [480] dtype = "float32" min_val = float("-1.52459") @@ -4653,6 +5076,7 @@ class Program_weight_tensor_parameter_422: class Program_weight_tensor_parameter_423: name = "parameter_423" + original_name = "batch_norm2d_86.w_0" shape = [480] dtype = "float32" min_val = float("0.274891") @@ -4664,6 +5088,7 @@ class Program_weight_tensor_parameter_423: class Program_weight_tensor_parameter_424: name = "parameter_424" + original_name = "batch_norm2d_86.w_2" shape = [480] dtype = "float32" min_val = float("9.50707e-05") @@ -4675,6 +5100,7 @@ class Program_weight_tensor_parameter_424: class Program_weight_tensor_parameter_425: name = "parameter_425" + original_name = "batch_norm2d_86.w_1" shape = [480] dtype = "float32" min_val = float("-0.0677259") @@ -4686,6 +5112,7 @@ class Program_weight_tensor_parameter_425: class Program_weight_tensor_parameter_426: name = "parameter_426" + original_name = "conv2d_89.w_0" shape = [480, 480, 1, 1] dtype = "float32" min_val = float("-0.022653") @@ -4697,6 +5124,7 @@ class Program_weight_tensor_parameter_426: class Program_weight_tensor_parameter_427: name = "parameter_427" + original_name = "batch_norm2d_85.b_0" shape = [480] dtype = "float32" min_val = float("-1.52459") @@ -4708,6 +5136,7 @@ class Program_weight_tensor_parameter_427: class Program_weight_tensor_parameter_428: name = "parameter_428" + original_name = "batch_norm2d_85.w_0" shape = [480] dtype = "float32" min_val = float("0.470566") @@ -4719,6 +5148,7 @@ class Program_weight_tensor_parameter_428: class Program_weight_tensor_parameter_429: name = "parameter_429" + original_name = "batch_norm2d_85.w_2" shape = [480] dtype = "float32" min_val = float("0.00074604") @@ -4730,6 +5160,7 @@ class Program_weight_tensor_parameter_429: class Program_weight_tensor_parameter_430: name = "parameter_430" + original_name = "batch_norm2d_85.w_1" shape = [480] dtype = "float32" min_val = float("-0.186469") @@ -4741,6 +5172,7 @@ class Program_weight_tensor_parameter_430: class Program_weight_tensor_parameter_431: name = "parameter_431" + original_name = "conv2d_88.w_0" shape = [480, 480, 3, 3] dtype = "float32" min_val = float("-0.0185384") @@ -4752,6 +5184,7 @@ class Program_weight_tensor_parameter_431: class Program_weight_tensor_parameter_432: name = "parameter_432" + original_name = "batch_norm2d_84.b_0" shape = [480] dtype = "float32" min_val = float("-2.09896") @@ -4763,6 +5196,7 @@ class Program_weight_tensor_parameter_432: class Program_weight_tensor_parameter_433: name = "parameter_433" + original_name = "batch_norm2d_84.w_0" shape = [480] dtype = "float32" min_val = float("0.574218") @@ -4774,6 +5208,7 @@ class Program_weight_tensor_parameter_433: class Program_weight_tensor_parameter_434: name = "parameter_434" + original_name = "batch_norm2d_84.w_2" shape = [480] dtype = "float32" min_val = float("0.0440332") @@ -4785,6 +5220,7 @@ class Program_weight_tensor_parameter_434: class Program_weight_tensor_parameter_435: name = "parameter_435" + original_name = "batch_norm2d_84.w_1" shape = [480] dtype = "float32" min_val = float("-1.28055") @@ -4796,6 +5232,7 @@ class Program_weight_tensor_parameter_435: class Program_weight_tensor_parameter_436: name = "parameter_436" + original_name = "conv2d_87.w_0" shape = [480, 480, 3, 3] dtype = "float32" min_val = float("-0.0289033") @@ -4807,6 +5244,7 @@ class Program_weight_tensor_parameter_436: class Program_weight_tensor_parameter_437: name = "parameter_437" + original_name = "batch_norm2d_83.b_0" shape = [480] dtype = "float32" min_val = float("-1.48465") @@ -4818,6 +5256,7 @@ class Program_weight_tensor_parameter_437: class Program_weight_tensor_parameter_438: name = "parameter_438" + original_name = "batch_norm2d_83.w_0" shape = [480] dtype = "float32" min_val = float("0.162109") @@ -4829,6 +5268,7 @@ class Program_weight_tensor_parameter_438: class Program_weight_tensor_parameter_439: name = "parameter_439" + original_name = "batch_norm2d_83.w_2" shape = [480] dtype = "float32" min_val = float("7.97679e-05") @@ -4840,6 +5280,7 @@ class Program_weight_tensor_parameter_439: class Program_weight_tensor_parameter_440: name = "parameter_440" + original_name = "batch_norm2d_83.w_1" shape = [480] dtype = "float32" min_val = float("-0.0420834") @@ -4851,6 +5292,7 @@ class Program_weight_tensor_parameter_440: class Program_weight_tensor_parameter_441: name = "parameter_441" + original_name = "conv2d_86.w_0" shape = [480, 480, 1, 1] dtype = "float32" min_val = float("-0.0187971") @@ -4862,6 +5304,7 @@ class Program_weight_tensor_parameter_441: class Program_weight_tensor_parameter_442: name = "parameter_442" + original_name = "batch_norm2d_82.b_0" shape = [480] dtype = "float32" min_val = float("-1.48465") @@ -4873,6 +5316,7 @@ class Program_weight_tensor_parameter_442: class Program_weight_tensor_parameter_443: name = "parameter_443" + original_name = "batch_norm2d_82.w_0" shape = [480] dtype = "float32" min_val = float("0.505714") @@ -4884,6 +5328,7 @@ class Program_weight_tensor_parameter_443: class Program_weight_tensor_parameter_444: name = "parameter_444" + original_name = "batch_norm2d_82.w_2" shape = [480] dtype = "float32" min_val = float("0.000803707") @@ -4895,6 +5340,7 @@ class Program_weight_tensor_parameter_444: class Program_weight_tensor_parameter_445: name = "parameter_445" + original_name = "batch_norm2d_82.w_1" shape = [480] dtype = "float32" min_val = float("-0.0787019") @@ -4906,6 +5352,7 @@ class Program_weight_tensor_parameter_445: class Program_weight_tensor_parameter_446: name = "parameter_446" + original_name = "conv2d_85.w_0" shape = [480, 480, 3, 3] dtype = "float32" min_val = float("-0.0164505") @@ -4917,6 +5364,7 @@ class Program_weight_tensor_parameter_446: class Program_weight_tensor_parameter_447: name = "parameter_447" + original_name = "batch_norm2d_81.b_0" shape = [480] dtype = "float32" min_val = float("-2.15499") @@ -4928,6 +5376,7 @@ class Program_weight_tensor_parameter_447: class Program_weight_tensor_parameter_448: name = "parameter_448" + original_name = "batch_norm2d_81.w_0" shape = [480] dtype = "float32" min_val = float("0.650522") @@ -4939,6 +5388,7 @@ class Program_weight_tensor_parameter_448: class Program_weight_tensor_parameter_449: name = "parameter_449" + original_name = "batch_norm2d_81.w_2" shape = [480] dtype = "float32" min_val = float("0.0297955") @@ -4950,6 +5400,7 @@ class Program_weight_tensor_parameter_449: class Program_weight_tensor_parameter_450: name = "parameter_450" + original_name = "batch_norm2d_81.w_1" shape = [480] dtype = "float32" min_val = float("-0.848851") @@ -4961,6 +5412,7 @@ class Program_weight_tensor_parameter_450: class Program_weight_tensor_parameter_451: name = "parameter_451" + original_name = "conv2d_84.w_0" shape = [480, 480, 3, 3] dtype = "float32" min_val = float("-0.0266256") @@ -4972,6 +5424,7 @@ class Program_weight_tensor_parameter_451: class Program_weight_tensor_parameter_452: name = "parameter_452" + original_name = "batch_norm2d_80.b_0" shape = [480] dtype = "float32" min_val = float("-1.48346") @@ -4983,6 +5436,7 @@ class Program_weight_tensor_parameter_452: class Program_weight_tensor_parameter_453: name = "parameter_453" + original_name = "batch_norm2d_80.w_0" shape = [480] dtype = "float32" min_val = float("0.105948") @@ -4994,6 +5448,7 @@ class Program_weight_tensor_parameter_453: class Program_weight_tensor_parameter_454: name = "parameter_454" + original_name = "batch_norm2d_80.w_2" shape = [480] dtype = "float32" min_val = float("9.02813e-05") @@ -5005,6 +5460,7 @@ class Program_weight_tensor_parameter_454: class Program_weight_tensor_parameter_455: name = "parameter_455" + original_name = "batch_norm2d_80.w_1" shape = [480] dtype = "float32" min_val = float("-0.0701126") @@ -5016,6 +5472,7 @@ class Program_weight_tensor_parameter_455: class Program_weight_tensor_parameter_456: name = "parameter_456" + original_name = "conv2d_83.w_0" shape = [480, 480, 1, 1] dtype = "float32" min_val = float("-0.0213249") @@ -5027,6 +5484,7 @@ class Program_weight_tensor_parameter_456: class Program_weight_tensor_parameter_457: name = "parameter_457" + original_name = "batch_norm2d_79.b_0" shape = [480] dtype = "float32" min_val = float("-1.48346") @@ -5038,6 +5496,7 @@ class Program_weight_tensor_parameter_457: class Program_weight_tensor_parameter_458: name = "parameter_458" + original_name = "batch_norm2d_79.w_0" shape = [480] dtype = "float32" min_val = float("0.387263") @@ -5049,6 +5508,7 @@ class Program_weight_tensor_parameter_458: class Program_weight_tensor_parameter_459: name = "parameter_459" + original_name = "batch_norm2d_79.w_2" shape = [480] dtype = "float32" min_val = float("0.000961408") @@ -5060,6 +5520,7 @@ class Program_weight_tensor_parameter_459: class Program_weight_tensor_parameter_460: name = "parameter_460" + original_name = "batch_norm2d_79.w_1" shape = [480] dtype = "float32" min_val = float("-0.0638962") @@ -5071,6 +5532,7 @@ class Program_weight_tensor_parameter_460: class Program_weight_tensor_parameter_461: name = "parameter_461" + original_name = "conv2d_82.w_0" shape = [480, 480, 3, 3] dtype = "float32" min_val = float("-0.015977") @@ -5082,6 +5544,7 @@ class Program_weight_tensor_parameter_461: class Program_weight_tensor_parameter_462: name = "parameter_462" + original_name = "batch_norm2d_78.b_0" shape = [480] dtype = "float32" min_val = float("-2.00835") @@ -5093,6 +5556,7 @@ class Program_weight_tensor_parameter_462: class Program_weight_tensor_parameter_463: name = "parameter_463" + original_name = "batch_norm2d_78.w_0" shape = [480] dtype = "float32" min_val = float("0.82039") @@ -5104,6 +5568,7 @@ class Program_weight_tensor_parameter_463: class Program_weight_tensor_parameter_464: name = "parameter_464" + original_name = "batch_norm2d_78.w_2" shape = [480] dtype = "float32" min_val = float("0.026508") @@ -5115,6 +5580,7 @@ class Program_weight_tensor_parameter_464: class Program_weight_tensor_parameter_465: name = "parameter_465" + original_name = "batch_norm2d_78.w_1" shape = [480] dtype = "float32" min_val = float("-0.661746") @@ -5126,6 +5592,7 @@ class Program_weight_tensor_parameter_465: class Program_weight_tensor_parameter_466: name = "parameter_466" + original_name = "conv2d_81.w_0" shape = [480, 480, 3, 3] dtype = "float32" min_val = float("-0.0177882") @@ -5137,6 +5604,7 @@ class Program_weight_tensor_parameter_466: class Program_weight_tensor_parameter_467: name = "parameter_467" + original_name = "batch_norm2d_77.b_0" shape = [480] dtype = "float32" min_val = float("-2.44471") @@ -5148,6 +5616,7 @@ class Program_weight_tensor_parameter_467: class Program_weight_tensor_parameter_468: name = "parameter_468" + original_name = "batch_norm2d_77.w_0" shape = [480] dtype = "float32" min_val = float("1.08286") @@ -5159,6 +5628,7 @@ class Program_weight_tensor_parameter_468: class Program_weight_tensor_parameter_469: name = "parameter_469" + original_name = "batch_norm2d_77.w_2" shape = [480] dtype = "float32" min_val = float("0.00150206") @@ -5170,6 +5640,7 @@ class Program_weight_tensor_parameter_469: class Program_weight_tensor_parameter_470: name = "parameter_470" + original_name = "batch_norm2d_77.w_1" shape = [480] dtype = "float32" min_val = float("-0.210809") @@ -5181,6 +5652,7 @@ class Program_weight_tensor_parameter_470: class Program_weight_tensor_parameter_471: name = "parameter_471" + original_name = "conv2d_80.w_0" shape = [480, 960, 1, 1] dtype = "float32" min_val = float("-0.0589829") @@ -5192,6 +5664,7 @@ class Program_weight_tensor_parameter_471: class Program_weight_tensor_parameter_472: name = "parameter_472" + original_name = "batch_norm2d_76.b_0" shape = [480] dtype = "float32" min_val = float("-2.32072") @@ -5203,6 +5676,7 @@ class Program_weight_tensor_parameter_472: class Program_weight_tensor_parameter_473: name = "parameter_473" + original_name = "batch_norm2d_76.w_0" shape = [480] dtype = "float32" min_val = float("1.37672") @@ -5214,6 +5688,7 @@ class Program_weight_tensor_parameter_473: class Program_weight_tensor_parameter_474: name = "parameter_474" + original_name = "batch_norm2d_76.w_2" shape = [480] dtype = "float32" min_val = float("0.000514538") @@ -5225,6 +5700,7 @@ class Program_weight_tensor_parameter_474: class Program_weight_tensor_parameter_475: name = "parameter_475" + original_name = "batch_norm2d_76.w_1" shape = [480] dtype = "float32" min_val = float("-0.0784749") @@ -5236,6 +5712,7 @@ class Program_weight_tensor_parameter_475: class Program_weight_tensor_parameter_476: name = "parameter_476" + original_name = "conv2d_79.w_0" shape = [480, 960, 1, 1] dtype = "float32" min_val = float("-0.148383") @@ -5247,6 +5724,7 @@ class Program_weight_tensor_parameter_476: class Program_weight_tensor_parameter_477: name = "parameter_477" + original_name = "batch_norm2d_75.b_0" shape = [960] dtype = "float32" min_val = float("-2.0465") @@ -5258,6 +5736,7 @@ class Program_weight_tensor_parameter_477: class Program_weight_tensor_parameter_478: name = "parameter_478" + original_name = "batch_norm2d_75.w_0" shape = [960] dtype = "float32" min_val = float("0.474623") @@ -5269,6 +5748,7 @@ class Program_weight_tensor_parameter_478: class Program_weight_tensor_parameter_479: name = "parameter_479" + original_name = "batch_norm2d_75.w_2" shape = [960] dtype = "float32" min_val = float("0.0052229") @@ -5280,6 +5760,7 @@ class Program_weight_tensor_parameter_479: class Program_weight_tensor_parameter_480: name = "parameter_480" + original_name = "batch_norm2d_75.w_1" shape = [960] dtype = "float32" min_val = float("-0.334956") @@ -5291,6 +5772,7 @@ class Program_weight_tensor_parameter_480: class Program_weight_tensor_parameter_481: name = "parameter_481" + original_name = "conv2d_78.w_0" shape = [960, 640, 3, 3] dtype = "float32" min_val = float("-0.0600408") @@ -5302,6 +5784,7 @@ class Program_weight_tensor_parameter_481: class Program_weight_tensor_parameter_482: name = "parameter_482" + original_name = "batch_norm2d_74.b_0" shape = [640] dtype = "float32" min_val = float("-2.7845") @@ -5313,6 +5796,7 @@ class Program_weight_tensor_parameter_482: class Program_weight_tensor_parameter_483: name = "parameter_483" + original_name = "batch_norm2d_74.w_0" shape = [640] dtype = "float32" min_val = float("0.50092") @@ -5324,6 +5808,7 @@ class Program_weight_tensor_parameter_483: class Program_weight_tensor_parameter_484: name = "parameter_484" + original_name = "batch_norm2d_74.w_2" shape = [640] dtype = "float32" min_val = float("0.00297603") @@ -5335,6 +5820,7 @@ class Program_weight_tensor_parameter_484: class Program_weight_tensor_parameter_485: name = "parameter_485" + original_name = "batch_norm2d_74.w_1" shape = [640] dtype = "float32" min_val = float("-0.135829") @@ -5346,6 +5832,7 @@ class Program_weight_tensor_parameter_485: class Program_weight_tensor_parameter_486: name = "parameter_486" + original_name = "conv2d_77.w_0" shape = [640, 480, 1, 1] dtype = "float32" min_val = float("-0.371604") @@ -5357,6 +5844,7 @@ class Program_weight_tensor_parameter_486: class Program_weight_tensor_parameter_487: name = "parameter_487" + original_name = "conv2d_76.b_0" shape = [480] dtype = "float32" min_val = float("-0.00829471") @@ -5368,6 +5856,7 @@ class Program_weight_tensor_parameter_487: class Program_weight_tensor_parameter_488: name = "parameter_488" + original_name = "conv2d_76.w_0" shape = [480, 480, 1, 1] dtype = "float32" min_val = float("-0.227075") @@ -5379,6 +5868,7 @@ class Program_weight_tensor_parameter_488: class Program_weight_tensor_parameter_489: name = "parameter_489" + original_name = "batch_norm2d_73.b_0" shape = [240] dtype = "float32" min_val = float("-1.72072") @@ -5390,6 +5880,7 @@ class Program_weight_tensor_parameter_489: class Program_weight_tensor_parameter_490: name = "parameter_490" + original_name = "batch_norm2d_73.w_0" shape = [240] dtype = "float32" min_val = float("0.0339229") @@ -5401,6 +5892,7 @@ class Program_weight_tensor_parameter_490: class Program_weight_tensor_parameter_491: name = "parameter_491" + original_name = "batch_norm2d_73.w_2" shape = [240] dtype = "float32" min_val = float("4.70628e-05") @@ -5412,6 +5904,7 @@ class Program_weight_tensor_parameter_491: class Program_weight_tensor_parameter_492: name = "parameter_492" + original_name = "batch_norm2d_73.w_1" shape = [240] dtype = "float32" min_val = float("-0.0453465") @@ -5423,6 +5916,7 @@ class Program_weight_tensor_parameter_492: class Program_weight_tensor_parameter_493: name = "parameter_493" + original_name = "conv2d_75.w_0" shape = [240, 240, 1, 1] dtype = "float32" min_val = float("-0.0213063") @@ -5434,6 +5928,7 @@ class Program_weight_tensor_parameter_493: class Program_weight_tensor_parameter_494: name = "parameter_494" + original_name = "batch_norm2d_72.b_0" shape = [240] dtype = "float32" min_val = float("-1.72072") @@ -5445,6 +5940,7 @@ class Program_weight_tensor_parameter_494: class Program_weight_tensor_parameter_495: name = "parameter_495" + original_name = "batch_norm2d_72.w_0" shape = [240] dtype = "float32" min_val = float("0.354919") @@ -5456,6 +5952,7 @@ class Program_weight_tensor_parameter_495: class Program_weight_tensor_parameter_496: name = "parameter_496" + original_name = "batch_norm2d_72.w_2" shape = [240] dtype = "float32" min_val = float("0.000985034") @@ -5467,6 +5964,7 @@ class Program_weight_tensor_parameter_496: class Program_weight_tensor_parameter_497: name = "parameter_497" + original_name = "batch_norm2d_72.w_1" shape = [240] dtype = "float32" min_val = float("-0.0724016") @@ -5478,6 +5976,7 @@ class Program_weight_tensor_parameter_497: class Program_weight_tensor_parameter_498: name = "parameter_498" + original_name = "conv2d_74.w_0" shape = [240, 240, 3, 3] dtype = "float32" min_val = float("-0.0217163") @@ -5489,6 +5988,7 @@ class Program_weight_tensor_parameter_498: class Program_weight_tensor_parameter_499: name = "parameter_499" + original_name = "batch_norm2d_71.b_0" shape = [240] dtype = "float32" min_val = float("-2.0155") @@ -5500,6 +6000,7 @@ class Program_weight_tensor_parameter_499: class Program_weight_tensor_parameter_500: name = "parameter_500" + original_name = "batch_norm2d_71.w_0" shape = [240] dtype = "float32" min_val = float("0.682425") @@ -5511,6 +6012,7 @@ class Program_weight_tensor_parameter_500: class Program_weight_tensor_parameter_501: name = "parameter_501" + original_name = "batch_norm2d_71.w_2" shape = [240] dtype = "float32" min_val = float("0.0684872") @@ -5522,6 +6024,7 @@ class Program_weight_tensor_parameter_501: class Program_weight_tensor_parameter_502: name = "parameter_502" + original_name = "batch_norm2d_71.w_1" shape = [240] dtype = "float32" min_val = float("-2.62071") @@ -5533,6 +6036,7 @@ class Program_weight_tensor_parameter_502: class Program_weight_tensor_parameter_503: name = "parameter_503" + original_name = "conv2d_73.w_0" shape = [240, 240, 3, 3] dtype = "float32" min_val = float("-0.0216964") @@ -5544,6 +6048,7 @@ class Program_weight_tensor_parameter_503: class Program_weight_tensor_parameter_504: name = "parameter_504" + original_name = "batch_norm2d_70.b_0" shape = [240] dtype = "float32" min_val = float("-1.56274") @@ -5555,6 +6060,7 @@ class Program_weight_tensor_parameter_504: class Program_weight_tensor_parameter_505: name = "parameter_505" + original_name = "batch_norm2d_70.w_0" shape = [240] dtype = "float32" min_val = float("0.00292149") @@ -5566,6 +6072,7 @@ class Program_weight_tensor_parameter_505: class Program_weight_tensor_parameter_506: name = "parameter_506" + original_name = "batch_norm2d_70.w_2" shape = [240] dtype = "float32" min_val = float("2.11748e-06") @@ -5577,6 +6084,7 @@ class Program_weight_tensor_parameter_506: class Program_weight_tensor_parameter_507: name = "parameter_507" + original_name = "batch_norm2d_70.w_1" shape = [240] dtype = "float32" min_val = float("-0.0302319") @@ -5588,6 +6096,7 @@ class Program_weight_tensor_parameter_507: class Program_weight_tensor_parameter_508: name = "parameter_508" + original_name = "conv2d_72.w_0" shape = [240, 240, 1, 1] dtype = "float32" min_val = float("-0.0247917") @@ -5599,6 +6108,7 @@ class Program_weight_tensor_parameter_508: class Program_weight_tensor_parameter_509: name = "parameter_509" + original_name = "batch_norm2d_69.b_0" shape = [240] dtype = "float32" min_val = float("-1.56274") @@ -5610,6 +6120,7 @@ class Program_weight_tensor_parameter_509: class Program_weight_tensor_parameter_510: name = "parameter_510" + original_name = "batch_norm2d_69.w_0" shape = [240] dtype = "float32" min_val = float("0.376677") @@ -5621,6 +6132,7 @@ class Program_weight_tensor_parameter_510: class Program_weight_tensor_parameter_511: name = "parameter_511" + original_name = "batch_norm2d_69.w_2" shape = [240] dtype = "float32" min_val = float("0.000833536") @@ -5632,6 +6144,7 @@ class Program_weight_tensor_parameter_511: class Program_weight_tensor_parameter_512: name = "parameter_512" + original_name = "batch_norm2d_69.w_1" shape = [240] dtype = "float32" min_val = float("-0.0876274") @@ -5643,6 +6156,7 @@ class Program_weight_tensor_parameter_512: class Program_weight_tensor_parameter_513: name = "parameter_513" + original_name = "conv2d_71.w_0" shape = [240, 240, 3, 3] dtype = "float32" min_val = float("-0.0222742") @@ -5654,6 +6168,7 @@ class Program_weight_tensor_parameter_513: class Program_weight_tensor_parameter_514: name = "parameter_514" + original_name = "batch_norm2d_68.b_0" shape = [240] dtype = "float32" min_val = float("-2.26468") @@ -5665,6 +6180,7 @@ class Program_weight_tensor_parameter_514: class Program_weight_tensor_parameter_515: name = "parameter_515" + original_name = "batch_norm2d_68.w_0" shape = [240] dtype = "float32" min_val = float("0.825716") @@ -5676,6 +6192,7 @@ class Program_weight_tensor_parameter_515: class Program_weight_tensor_parameter_516: name = "parameter_516" + original_name = "batch_norm2d_68.w_2" shape = [240] dtype = "float32" min_val = float("0.0526299") @@ -5687,6 +6204,7 @@ class Program_weight_tensor_parameter_516: class Program_weight_tensor_parameter_517: name = "parameter_517" + original_name = "batch_norm2d_68.w_1" shape = [240] dtype = "float32" min_val = float("-1.3719") @@ -5698,6 +6216,7 @@ class Program_weight_tensor_parameter_517: class Program_weight_tensor_parameter_518: name = "parameter_518" + original_name = "conv2d_70.w_0" shape = [240, 240, 3, 3] dtype = "float32" min_val = float("-0.0396919") @@ -5709,6 +6228,7 @@ class Program_weight_tensor_parameter_518: class Program_weight_tensor_parameter_519: name = "parameter_519" + original_name = "batch_norm2d_67.b_0" shape = [240] dtype = "float32" min_val = float("-1.4775") @@ -5720,6 +6240,7 @@ class Program_weight_tensor_parameter_519: class Program_weight_tensor_parameter_520: name = "parameter_520" + original_name = "batch_norm2d_67.w_0" shape = [240] dtype = "float32" min_val = float("0.0597698") @@ -5731,6 +6252,7 @@ class Program_weight_tensor_parameter_520: class Program_weight_tensor_parameter_521: name = "parameter_521" + original_name = "batch_norm2d_67.w_2" shape = [240] dtype = "float32" min_val = float("4.47211e-05") @@ -5742,6 +6264,7 @@ class Program_weight_tensor_parameter_521: class Program_weight_tensor_parameter_522: name = "parameter_522" + original_name = "batch_norm2d_67.w_1" shape = [240] dtype = "float32" min_val = float("-0.034") @@ -5753,6 +6276,7 @@ class Program_weight_tensor_parameter_522: class Program_weight_tensor_parameter_523: name = "parameter_523" + original_name = "conv2d_69.w_0" shape = [240, 240, 1, 1] dtype = "float32" min_val = float("-0.0275889") @@ -5764,6 +6288,7 @@ class Program_weight_tensor_parameter_523: class Program_weight_tensor_parameter_524: name = "parameter_524" + original_name = "batch_norm2d_66.b_0" shape = [240] dtype = "float32" min_val = float("-1.4775") @@ -5775,6 +6300,7 @@ class Program_weight_tensor_parameter_524: class Program_weight_tensor_parameter_525: name = "parameter_525" + original_name = "batch_norm2d_66.w_0" shape = [240] dtype = "float32" min_val = float("0.442697") @@ -5786,6 +6312,7 @@ class Program_weight_tensor_parameter_525: class Program_weight_tensor_parameter_526: name = "parameter_526" + original_name = "batch_norm2d_66.w_2" shape = [240] dtype = "float32" min_val = float("0.00135648") @@ -5797,6 +6324,7 @@ class Program_weight_tensor_parameter_526: class Program_weight_tensor_parameter_527: name = "parameter_527" + original_name = "batch_norm2d_66.w_1" shape = [240] dtype = "float32" min_val = float("-0.116543") @@ -5808,6 +6336,7 @@ class Program_weight_tensor_parameter_527: class Program_weight_tensor_parameter_528: name = "parameter_528" + original_name = "conv2d_68.w_0" shape = [240, 240, 3, 3] dtype = "float32" min_val = float("-0.0231192") @@ -5819,6 +6348,7 @@ class Program_weight_tensor_parameter_528: class Program_weight_tensor_parameter_529: name = "parameter_529" + original_name = "batch_norm2d_65.b_0" shape = [240] dtype = "float32" min_val = float("-2.04888") @@ -5830,6 +6360,7 @@ class Program_weight_tensor_parameter_529: class Program_weight_tensor_parameter_530: name = "parameter_530" + original_name = "batch_norm2d_65.w_0" shape = [240] dtype = "float32" min_val = float("0.815988") @@ -5841,6 +6372,7 @@ class Program_weight_tensor_parameter_530: class Program_weight_tensor_parameter_531: name = "parameter_531" + original_name = "batch_norm2d_65.w_2" shape = [240] dtype = "float32" min_val = float("0.0415805") @@ -5852,6 +6384,7 @@ class Program_weight_tensor_parameter_531: class Program_weight_tensor_parameter_532: name = "parameter_532" + original_name = "batch_norm2d_65.w_1" shape = [240] dtype = "float32" min_val = float("-0.95685") @@ -5863,6 +6396,7 @@ class Program_weight_tensor_parameter_532: class Program_weight_tensor_parameter_533: name = "parameter_533" + original_name = "conv2d_67.w_0" shape = [240, 240, 3, 3] dtype = "float32" min_val = float("-0.0318932") @@ -5874,6 +6408,7 @@ class Program_weight_tensor_parameter_533: class Program_weight_tensor_parameter_534: name = "parameter_534" + original_name = "batch_norm2d_64.b_0" shape = [240] dtype = "float32" min_val = float("-1.6623") @@ -5885,6 +6420,7 @@ class Program_weight_tensor_parameter_534: class Program_weight_tensor_parameter_535: name = "parameter_535" + original_name = "batch_norm2d_64.w_0" shape = [240] dtype = "float32" min_val = float("0.00151445") @@ -5896,6 +6432,7 @@ class Program_weight_tensor_parameter_535: class Program_weight_tensor_parameter_536: name = "parameter_536" + original_name = "batch_norm2d_64.w_2" shape = [240] dtype = "float32" min_val = float("5.14352e-07") @@ -5907,6 +6444,7 @@ class Program_weight_tensor_parameter_536: class Program_weight_tensor_parameter_537: name = "parameter_537" + original_name = "batch_norm2d_64.w_1" shape = [240] dtype = "float32" min_val = float("-0.0441197") @@ -5918,6 +6456,7 @@ class Program_weight_tensor_parameter_537: class Program_weight_tensor_parameter_538: name = "parameter_538" + original_name = "conv2d_66.w_0" shape = [240, 240, 1, 1] dtype = "float32" min_val = float("-0.0202486") @@ -5929,6 +6468,7 @@ class Program_weight_tensor_parameter_538: class Program_weight_tensor_parameter_539: name = "parameter_539" + original_name = "batch_norm2d_63.b_0" shape = [240] dtype = "float32" min_val = float("-1.6623") @@ -5940,6 +6480,7 @@ class Program_weight_tensor_parameter_539: class Program_weight_tensor_parameter_540: name = "parameter_540" + original_name = "batch_norm2d_63.w_0" shape = [240] dtype = "float32" min_val = float("0.397712") @@ -5951,6 +6492,7 @@ class Program_weight_tensor_parameter_540: class Program_weight_tensor_parameter_541: name = "parameter_541" + original_name = "batch_norm2d_63.w_2" shape = [240] dtype = "float32" min_val = float("0.00091409") @@ -5962,6 +6504,7 @@ class Program_weight_tensor_parameter_541: class Program_weight_tensor_parameter_542: name = "parameter_542" + original_name = "batch_norm2d_63.w_1" shape = [240] dtype = "float32" min_val = float("-0.019431") @@ -5973,6 +6516,7 @@ class Program_weight_tensor_parameter_542: class Program_weight_tensor_parameter_543: name = "parameter_543" + original_name = "conv2d_65.w_0" shape = [240, 240, 3, 3] dtype = "float32" min_val = float("-0.0234898") @@ -5984,6 +6528,7 @@ class Program_weight_tensor_parameter_543: class Program_weight_tensor_parameter_544: name = "parameter_544" + original_name = "batch_norm2d_62.b_0" shape = [240] dtype = "float32" min_val = float("-2.42683") @@ -5995,6 +6540,7 @@ class Program_weight_tensor_parameter_544: class Program_weight_tensor_parameter_545: name = "parameter_545" + original_name = "batch_norm2d_62.w_0" shape = [240] dtype = "float32" min_val = float("0.776045") @@ -6006,6 +6552,7 @@ class Program_weight_tensor_parameter_545: class Program_weight_tensor_parameter_546: name = "parameter_546" + original_name = "batch_norm2d_62.w_2" shape = [240] dtype = "float32" min_val = float("0.0314653") @@ -6017,6 +6564,7 @@ class Program_weight_tensor_parameter_546: class Program_weight_tensor_parameter_547: name = "parameter_547" + original_name = "batch_norm2d_62.w_1" shape = [240] dtype = "float32" min_val = float("-1.14936") @@ -6028,6 +6576,7 @@ class Program_weight_tensor_parameter_547: class Program_weight_tensor_parameter_548: name = "parameter_548" + original_name = "conv2d_64.w_0" shape = [240, 240, 3, 3] dtype = "float32" min_val = float("-0.037141") @@ -6039,6 +6588,7 @@ class Program_weight_tensor_parameter_548: class Program_weight_tensor_parameter_549: name = "parameter_549" + original_name = "batch_norm2d_61.b_0" shape = [240] dtype = "float32" min_val = float("-1.42444") @@ -6050,6 +6600,7 @@ class Program_weight_tensor_parameter_549: class Program_weight_tensor_parameter_550: name = "parameter_550" + original_name = "batch_norm2d_61.w_0" shape = [240] dtype = "float32" min_val = float("-0.000935615") @@ -6061,6 +6612,7 @@ class Program_weight_tensor_parameter_550: class Program_weight_tensor_parameter_551: name = "parameter_551" + original_name = "batch_norm2d_61.w_2" shape = [240] dtype = "float32" min_val = float("3.5019e-07") @@ -6072,6 +6624,7 @@ class Program_weight_tensor_parameter_551: class Program_weight_tensor_parameter_552: name = "parameter_552" + original_name = "batch_norm2d_61.w_1" shape = [240] dtype = "float32" min_val = float("-0.0221443") @@ -6083,6 +6636,7 @@ class Program_weight_tensor_parameter_552: class Program_weight_tensor_parameter_553: name = "parameter_553" + original_name = "conv2d_63.w_0" shape = [240, 240, 1, 1] dtype = "float32" min_val = float("-0.0267442") @@ -6094,6 +6648,7 @@ class Program_weight_tensor_parameter_553: class Program_weight_tensor_parameter_554: name = "parameter_554" + original_name = "batch_norm2d_60.b_0" shape = [240] dtype = "float32" min_val = float("-1.42444") @@ -6105,6 +6660,7 @@ class Program_weight_tensor_parameter_554: class Program_weight_tensor_parameter_555: name = "parameter_555" + original_name = "batch_norm2d_60.w_0" shape = [240] dtype = "float32" min_val = float("0.393427") @@ -6116,6 +6672,7 @@ class Program_weight_tensor_parameter_555: class Program_weight_tensor_parameter_556: name = "parameter_556" + original_name = "batch_norm2d_60.w_2" shape = [240] dtype = "float32" min_val = float("0.00112389") @@ -6127,6 +6684,7 @@ class Program_weight_tensor_parameter_556: class Program_weight_tensor_parameter_557: name = "parameter_557" + original_name = "batch_norm2d_60.w_1" shape = [240] dtype = "float32" min_val = float("-0.0558245") @@ -6138,6 +6696,7 @@ class Program_weight_tensor_parameter_557: class Program_weight_tensor_parameter_558: name = "parameter_558" + original_name = "conv2d_62.w_0" shape = [240, 240, 3, 3] dtype = "float32" min_val = float("-0.0261494") @@ -6149,6 +6708,7 @@ class Program_weight_tensor_parameter_558: class Program_weight_tensor_parameter_559: name = "parameter_559" + original_name = "batch_norm2d_59.b_0" shape = [240] dtype = "float32" min_val = float("-2.38271") @@ -6160,6 +6720,7 @@ class Program_weight_tensor_parameter_559: class Program_weight_tensor_parameter_560: name = "parameter_560" + original_name = "batch_norm2d_59.w_0" shape = [240] dtype = "float32" min_val = float("0.73845") @@ -6171,6 +6732,7 @@ class Program_weight_tensor_parameter_560: class Program_weight_tensor_parameter_561: name = "parameter_561" + original_name = "batch_norm2d_59.w_2" shape = [240] dtype = "float32" min_val = float("0.0241186") @@ -6182,6 +6744,7 @@ class Program_weight_tensor_parameter_561: class Program_weight_tensor_parameter_562: name = "parameter_562" + original_name = "batch_norm2d_59.w_1" shape = [240] dtype = "float32" min_val = float("-0.439696") @@ -6193,6 +6756,7 @@ class Program_weight_tensor_parameter_562: class Program_weight_tensor_parameter_563: name = "parameter_563" + original_name = "conv2d_61.w_0" shape = [240, 240, 3, 3] dtype = "float32" min_val = float("-0.0542627") @@ -6204,6 +6768,7 @@ class Program_weight_tensor_parameter_563: class Program_weight_tensor_parameter_564: name = "parameter_564" + original_name = "batch_norm2d_58.b_0" shape = [240] dtype = "float32" min_val = float("-1.35947") @@ -6215,6 +6780,7 @@ class Program_weight_tensor_parameter_564: class Program_weight_tensor_parameter_565: name = "parameter_565" + original_name = "batch_norm2d_58.w_0" shape = [240] dtype = "float32" min_val = float("-0.000573501") @@ -6226,6 +6792,7 @@ class Program_weight_tensor_parameter_565: class Program_weight_tensor_parameter_566: name = "parameter_566" + original_name = "batch_norm2d_58.w_2" shape = [240] dtype = "float32" min_val = float("1.32971e-09") @@ -6237,6 +6804,7 @@ class Program_weight_tensor_parameter_566: class Program_weight_tensor_parameter_567: name = "parameter_567" + original_name = "batch_norm2d_58.w_1" shape = [240] dtype = "float32" min_val = float("-0.0379828") @@ -6248,6 +6816,7 @@ class Program_weight_tensor_parameter_567: class Program_weight_tensor_parameter_568: name = "parameter_568" + original_name = "conv2d_60.w_0" shape = [240, 240, 1, 1] dtype = "float32" min_val = float("-0.0215854") @@ -6259,6 +6828,7 @@ class Program_weight_tensor_parameter_568: class Program_weight_tensor_parameter_569: name = "parameter_569" + original_name = "batch_norm2d_57.b_0" shape = [240] dtype = "float32" min_val = float("-1.35947") @@ -6270,6 +6840,7 @@ class Program_weight_tensor_parameter_569: class Program_weight_tensor_parameter_570: name = "parameter_570" + original_name = "batch_norm2d_57.w_0" shape = [240] dtype = "float32" min_val = float("0.354482") @@ -6281,6 +6852,7 @@ class Program_weight_tensor_parameter_570: class Program_weight_tensor_parameter_571: name = "parameter_571" + original_name = "batch_norm2d_57.w_2" shape = [240] dtype = "float32" min_val = float("0.00156109") @@ -6292,6 +6864,7 @@ class Program_weight_tensor_parameter_571: class Program_weight_tensor_parameter_572: name = "parameter_572" + original_name = "batch_norm2d_57.w_1" shape = [240] dtype = "float32" min_val = float("-0.0373222") @@ -6303,6 +6876,7 @@ class Program_weight_tensor_parameter_572: class Program_weight_tensor_parameter_573: name = "parameter_573" + original_name = "conv2d_59.w_0" shape = [240, 240, 3, 3] dtype = "float32" min_val = float("-0.0214125") @@ -6314,6 +6888,7 @@ class Program_weight_tensor_parameter_573: class Program_weight_tensor_parameter_574: name = "parameter_574" + original_name = "batch_norm2d_56.b_0" shape = [240] dtype = "float32" min_val = float("-2.19791") @@ -6325,6 +6900,7 @@ class Program_weight_tensor_parameter_574: class Program_weight_tensor_parameter_575: name = "parameter_575" + original_name = "batch_norm2d_56.w_0" shape = [240] dtype = "float32" min_val = float("0.72623") @@ -6336,6 +6912,7 @@ class Program_weight_tensor_parameter_575: class Program_weight_tensor_parameter_576: name = "parameter_576" + original_name = "batch_norm2d_56.w_2" shape = [240] dtype = "float32" min_val = float("0.0184826") @@ -6347,6 +6924,7 @@ class Program_weight_tensor_parameter_576: class Program_weight_tensor_parameter_577: name = "parameter_577" + original_name = "batch_norm2d_56.w_1" shape = [240] dtype = "float32" min_val = float("-1.05172") @@ -6358,6 +6936,7 @@ class Program_weight_tensor_parameter_577: class Program_weight_tensor_parameter_578: name = "parameter_578" + original_name = "conv2d_58.w_0" shape = [240, 240, 3, 3] dtype = "float32" min_val = float("-0.0578109") @@ -6369,6 +6948,7 @@ class Program_weight_tensor_parameter_578: class Program_weight_tensor_parameter_579: name = "parameter_579" + original_name = "batch_norm2d_55.b_0" shape = [240] dtype = "float32" min_val = float("-1.06259") @@ -6380,6 +6960,7 @@ class Program_weight_tensor_parameter_579: class Program_weight_tensor_parameter_580: name = "parameter_580" + original_name = "batch_norm2d_55.w_0" shape = [240] dtype = "float32" min_val = float("2.03909e-05") @@ -6391,6 +6972,7 @@ class Program_weight_tensor_parameter_580: class Program_weight_tensor_parameter_581: name = "parameter_581" + original_name = "batch_norm2d_55.w_2" shape = [240] dtype = "float32" min_val = float("1.07072e-10") @@ -6402,6 +6984,7 @@ class Program_weight_tensor_parameter_581: class Program_weight_tensor_parameter_582: name = "parameter_582" + original_name = "batch_norm2d_55.w_1" shape = [240] dtype = "float32" min_val = float("-0.0343338") @@ -6413,6 +6996,7 @@ class Program_weight_tensor_parameter_582: class Program_weight_tensor_parameter_583: name = "parameter_583" + original_name = "conv2d_57.w_0" shape = [240, 240, 1, 1] dtype = "float32" min_val = float("-0.0305483") @@ -6424,6 +7008,7 @@ class Program_weight_tensor_parameter_583: class Program_weight_tensor_parameter_584: name = "parameter_584" + original_name = "batch_norm2d_54.b_0" shape = [240] dtype = "float32" min_val = float("-1.06259") @@ -6435,6 +7020,7 @@ class Program_weight_tensor_parameter_584: class Program_weight_tensor_parameter_585: name = "parameter_585" + original_name = "batch_norm2d_54.w_0" shape = [240] dtype = "float32" min_val = float("0.381036") @@ -6446,6 +7032,7 @@ class Program_weight_tensor_parameter_585: class Program_weight_tensor_parameter_586: name = "parameter_586" + original_name = "batch_norm2d_54.w_2" shape = [240] dtype = "float32" min_val = float("0.00227622") @@ -6457,6 +7044,7 @@ class Program_weight_tensor_parameter_586: class Program_weight_tensor_parameter_587: name = "parameter_587" + original_name = "batch_norm2d_54.w_1" shape = [240] dtype = "float32" min_val = float("-0.0595013") @@ -6468,6 +7056,7 @@ class Program_weight_tensor_parameter_587: class Program_weight_tensor_parameter_588: name = "parameter_588" + original_name = "conv2d_56.w_0" shape = [240, 240, 3, 3] dtype = "float32" min_val = float("-0.0499103") @@ -6479,6 +7068,7 @@ class Program_weight_tensor_parameter_588: class Program_weight_tensor_parameter_589: name = "parameter_589" + original_name = "batch_norm2d_53.b_0" shape = [240] dtype = "float32" min_val = float("-2.88346") @@ -6490,6 +7080,7 @@ class Program_weight_tensor_parameter_589: class Program_weight_tensor_parameter_590: name = "parameter_590" + original_name = "batch_norm2d_53.w_0" shape = [240] dtype = "float32" min_val = float("0.803967") @@ -6501,6 +7092,7 @@ class Program_weight_tensor_parameter_590: class Program_weight_tensor_parameter_591: name = "parameter_591" + original_name = "batch_norm2d_53.w_2" shape = [240] dtype = "float32" min_val = float("0.0130947") @@ -6512,6 +7104,7 @@ class Program_weight_tensor_parameter_591: class Program_weight_tensor_parameter_592: name = "parameter_592" + original_name = "batch_norm2d_53.w_1" shape = [240] dtype = "float32" min_val = float("-1.00946") @@ -6523,6 +7116,7 @@ class Program_weight_tensor_parameter_592: class Program_weight_tensor_parameter_593: name = "parameter_593" + original_name = "conv2d_55.w_0" shape = [240, 240, 3, 3] dtype = "float32" min_val = float("-0.0674478") @@ -6534,6 +7128,7 @@ class Program_weight_tensor_parameter_593: class Program_weight_tensor_parameter_594: name = "parameter_594" + original_name = "batch_norm2d_52.b_0" shape = [240] dtype = "float32" min_val = float("-1.12459") @@ -6545,6 +7140,7 @@ class Program_weight_tensor_parameter_594: class Program_weight_tensor_parameter_595: name = "parameter_595" + original_name = "batch_norm2d_52.w_0" shape = [240] dtype = "float32" min_val = float("0.00382133") @@ -6556,6 +7152,7 @@ class Program_weight_tensor_parameter_595: class Program_weight_tensor_parameter_596: name = "parameter_596" + original_name = "batch_norm2d_52.w_2" shape = [240] dtype = "float32" min_val = float("1.08939e-05") @@ -6567,6 +7164,7 @@ class Program_weight_tensor_parameter_596: class Program_weight_tensor_parameter_597: name = "parameter_597" + original_name = "batch_norm2d_52.w_1" shape = [240] dtype = "float32" min_val = float("-0.0518277") @@ -6578,6 +7176,7 @@ class Program_weight_tensor_parameter_597: class Program_weight_tensor_parameter_598: name = "parameter_598" + original_name = "conv2d_54.w_0" shape = [240, 240, 1, 1] dtype = "float32" min_val = float("-0.0688521") @@ -6589,6 +7188,7 @@ class Program_weight_tensor_parameter_598: class Program_weight_tensor_parameter_599: name = "parameter_599" + original_name = "batch_norm2d_51.b_0" shape = [240] dtype = "float32" min_val = float("-1.12459") @@ -6600,6 +7200,7 @@ class Program_weight_tensor_parameter_599: class Program_weight_tensor_parameter_600: name = "parameter_600" + original_name = "batch_norm2d_51.w_0" shape = [240] dtype = "float32" min_val = float("0.347092") @@ -6611,6 +7212,7 @@ class Program_weight_tensor_parameter_600: class Program_weight_tensor_parameter_601: name = "parameter_601" + original_name = "batch_norm2d_51.w_2" shape = [240] dtype = "float32" min_val = float("0.00366499") @@ -6622,6 +7224,7 @@ class Program_weight_tensor_parameter_601: class Program_weight_tensor_parameter_602: name = "parameter_602" + original_name = "batch_norm2d_51.w_1" shape = [240] dtype = "float32" min_val = float("-0.0948261") @@ -6633,6 +7236,7 @@ class Program_weight_tensor_parameter_602: class Program_weight_tensor_parameter_603: name = "parameter_603" + original_name = "conv2d_53.w_0" shape = [240, 240, 3, 3] dtype = "float32" min_val = float("-0.0552751") @@ -6644,6 +7248,7 @@ class Program_weight_tensor_parameter_603: class Program_weight_tensor_parameter_604: name = "parameter_604" + original_name = "batch_norm2d_50.b_0" shape = [240] dtype = "float32" min_val = float("-1.90909") @@ -6655,6 +7260,7 @@ class Program_weight_tensor_parameter_604: class Program_weight_tensor_parameter_605: name = "parameter_605" + original_name = "batch_norm2d_50.w_0" shape = [240] dtype = "float32" min_val = float("0.810204") @@ -6666,6 +7272,7 @@ class Program_weight_tensor_parameter_605: class Program_weight_tensor_parameter_606: name = "parameter_606" + original_name = "batch_norm2d_50.w_2" shape = [240] dtype = "float32" min_val = float("0.0117367") @@ -6677,6 +7284,7 @@ class Program_weight_tensor_parameter_606: class Program_weight_tensor_parameter_607: name = "parameter_607" + original_name = "batch_norm2d_50.w_1" shape = [240] dtype = "float32" min_val = float("-0.665895") @@ -6688,6 +7296,7 @@ class Program_weight_tensor_parameter_607: class Program_weight_tensor_parameter_608: name = "parameter_608" + original_name = "conv2d_52.w_0" shape = [240, 240, 3, 3] dtype = "float32" min_val = float("-0.0997303") @@ -6699,6 +7308,7 @@ class Program_weight_tensor_parameter_608: class Program_weight_tensor_parameter_609: name = "parameter_609" + original_name = "batch_norm2d_49.b_0" shape = [240] dtype = "float32" min_val = float("-2.43636") @@ -6710,6 +7320,7 @@ class Program_weight_tensor_parameter_609: class Program_weight_tensor_parameter_610: name = "parameter_610" + original_name = "batch_norm2d_49.w_0" shape = [240] dtype = "float32" min_val = float("0.509393") @@ -6721,6 +7332,7 @@ class Program_weight_tensor_parameter_610: class Program_weight_tensor_parameter_611: name = "parameter_611" + original_name = "batch_norm2d_49.w_2" shape = [240] dtype = "float32" min_val = float("0.00852281") @@ -6732,6 +7344,7 @@ class Program_weight_tensor_parameter_611: class Program_weight_tensor_parameter_612: name = "parameter_612" + original_name = "batch_norm2d_49.w_1" shape = [240] dtype = "float32" min_val = float("-0.119529") @@ -6743,6 +7356,7 @@ class Program_weight_tensor_parameter_612: class Program_weight_tensor_parameter_613: name = "parameter_613" + original_name = "conv2d_51.w_0" shape = [240, 480, 1, 1] dtype = "float32" min_val = float("-0.114995") @@ -6754,6 +7368,7 @@ class Program_weight_tensor_parameter_613: class Program_weight_tensor_parameter_614: name = "parameter_614" + original_name = "batch_norm2d_48.b_0" shape = [240] dtype = "float32" min_val = float("-2.46855") @@ -6765,6 +7380,7 @@ class Program_weight_tensor_parameter_614: class Program_weight_tensor_parameter_615: name = "parameter_615" + original_name = "batch_norm2d_48.w_0" shape = [240] dtype = "float32" min_val = float("1.19732") @@ -6776,6 +7392,7 @@ class Program_weight_tensor_parameter_615: class Program_weight_tensor_parameter_616: name = "parameter_616" + original_name = "batch_norm2d_48.w_2" shape = [240] dtype = "float32" min_val = float("0.00435732") @@ -6787,6 +7404,7 @@ class Program_weight_tensor_parameter_616: class Program_weight_tensor_parameter_617: name = "parameter_617" + original_name = "batch_norm2d_48.w_1" shape = [240] dtype = "float32" min_val = float("-0.107848") @@ -6798,6 +7416,7 @@ class Program_weight_tensor_parameter_617: class Program_weight_tensor_parameter_618: name = "parameter_618" + original_name = "conv2d_50.w_0" shape = [240, 480, 1, 1] dtype = "float32" min_val = float("-0.0616602") @@ -6809,6 +7428,7 @@ class Program_weight_tensor_parameter_618: class Program_weight_tensor_parameter_619: name = "parameter_619" + original_name = "batch_norm2d_47.b_0" shape = [480] dtype = "float32" min_val = float("-2.22414") @@ -6820,6 +7440,7 @@ class Program_weight_tensor_parameter_619: class Program_weight_tensor_parameter_620: name = "parameter_620" + original_name = "batch_norm2d_47.w_0" shape = [480] dtype = "float32" min_val = float("0.707847") @@ -6831,6 +7452,7 @@ class Program_weight_tensor_parameter_620: class Program_weight_tensor_parameter_621: name = "parameter_621" + original_name = "batch_norm2d_47.w_2" shape = [480] dtype = "float32" min_val = float("0.00911732") @@ -6842,6 +7464,7 @@ class Program_weight_tensor_parameter_621: class Program_weight_tensor_parameter_622: name = "parameter_622" + original_name = "batch_norm2d_47.w_1" shape = [480] dtype = "float32" min_val = float("-0.149345") @@ -6853,6 +7476,7 @@ class Program_weight_tensor_parameter_622: class Program_weight_tensor_parameter_623: name = "parameter_623" + original_name = "conv2d_49.w_0" shape = [480, 320, 3, 3] dtype = "float32" min_val = float("-0.0673143") @@ -6864,6 +7488,7 @@ class Program_weight_tensor_parameter_623: class Program_weight_tensor_parameter_624: name = "parameter_624" + original_name = "batch_norm2d_46.b_0" shape = [320] dtype = "float32" min_val = float("-2.10684") @@ -6875,6 +7500,7 @@ class Program_weight_tensor_parameter_624: class Program_weight_tensor_parameter_625: name = "parameter_625" + original_name = "batch_norm2d_46.w_0" shape = [320] dtype = "float32" min_val = float("0.509217") @@ -6886,6 +7512,7 @@ class Program_weight_tensor_parameter_625: class Program_weight_tensor_parameter_626: name = "parameter_626" + original_name = "batch_norm2d_46.w_2" shape = [320] dtype = "float32" min_val = float("0.00183519") @@ -6897,6 +7524,7 @@ class Program_weight_tensor_parameter_626: class Program_weight_tensor_parameter_627: name = "parameter_627" + original_name = "batch_norm2d_46.w_1" shape = [320] dtype = "float32" min_val = float("-0.202264") @@ -6908,6 +7536,7 @@ class Program_weight_tensor_parameter_627: class Program_weight_tensor_parameter_628: name = "parameter_628" + original_name = "conv2d_48.w_0" shape = [320, 240, 1, 1] dtype = "float32" min_val = float("-0.154121") @@ -6919,6 +7548,7 @@ class Program_weight_tensor_parameter_628: class Program_weight_tensor_parameter_629: name = "parameter_629" + original_name = "conv2d_47.b_0" shape = [240] dtype = "float32" min_val = float("-0.00957327") @@ -6930,6 +7560,7 @@ class Program_weight_tensor_parameter_629: class Program_weight_tensor_parameter_630: name = "parameter_630" + original_name = "conv2d_47.w_0" shape = [240, 240, 1, 1] dtype = "float32" min_val = float("-0.195899") @@ -6941,6 +7572,7 @@ class Program_weight_tensor_parameter_630: class Program_weight_tensor_parameter_631: name = "parameter_631" + original_name = "batch_norm2d_45.b_0" shape = [120] dtype = "float32" min_val = float("-1.61173") @@ -6952,6 +7584,7 @@ class Program_weight_tensor_parameter_631: class Program_weight_tensor_parameter_632: name = "parameter_632" + original_name = "batch_norm2d_45.w_0" shape = [120] dtype = "float32" min_val = float("0.127172") @@ -6963,6 +7596,7 @@ class Program_weight_tensor_parameter_632: class Program_weight_tensor_parameter_633: name = "parameter_633" + original_name = "batch_norm2d_45.w_2" shape = [120] dtype = "float32" min_val = float("0.000122234") @@ -6974,6 +7608,7 @@ class Program_weight_tensor_parameter_633: class Program_weight_tensor_parameter_634: name = "parameter_634" + original_name = "batch_norm2d_45.w_1" shape = [120] dtype = "float32" min_val = float("-0.045886") @@ -6985,6 +7620,7 @@ class Program_weight_tensor_parameter_634: class Program_weight_tensor_parameter_635: name = "parameter_635" + original_name = "conv2d_46.w_0" shape = [120, 120, 1, 1] dtype = "float32" min_val = float("-0.041441") @@ -6996,6 +7632,7 @@ class Program_weight_tensor_parameter_635: class Program_weight_tensor_parameter_636: name = "parameter_636" + original_name = "batch_norm2d_44.b_0" shape = [120] dtype = "float32" min_val = float("-1.61173") @@ -7007,6 +7644,7 @@ class Program_weight_tensor_parameter_636: class Program_weight_tensor_parameter_637: name = "parameter_637" + original_name = "batch_norm2d_44.w_0" shape = [120] dtype = "float32" min_val = float("0.312032") @@ -7018,6 +7656,7 @@ class Program_weight_tensor_parameter_637: class Program_weight_tensor_parameter_638: name = "parameter_638" + original_name = "batch_norm2d_44.w_2" shape = [120] dtype = "float32" min_val = float("0.000946215") @@ -7029,6 +7668,7 @@ class Program_weight_tensor_parameter_638: class Program_weight_tensor_parameter_639: name = "parameter_639" + original_name = "batch_norm2d_44.w_1" shape = [120] dtype = "float32" min_val = float("-0.0806164") @@ -7040,6 +7680,7 @@ class Program_weight_tensor_parameter_639: class Program_weight_tensor_parameter_640: name = "parameter_640" + original_name = "conv2d_45.w_0" shape = [120, 120, 3, 3] dtype = "float32" min_val = float("-0.0406141") @@ -7051,6 +7692,7 @@ class Program_weight_tensor_parameter_640: class Program_weight_tensor_parameter_641: name = "parameter_641" + original_name = "batch_norm2d_43.b_0" shape = [120] dtype = "float32" min_val = float("-2.09776") @@ -7062,6 +7704,7 @@ class Program_weight_tensor_parameter_641: class Program_weight_tensor_parameter_642: name = "parameter_642" + original_name = "batch_norm2d_43.w_0" shape = [120] dtype = "float32" min_val = float("0.719523") @@ -7073,6 +7716,7 @@ class Program_weight_tensor_parameter_642: class Program_weight_tensor_parameter_643: name = "parameter_643" + original_name = "batch_norm2d_43.w_2" shape = [120] dtype = "float32" min_val = float("0.0459699") @@ -7084,6 +7728,7 @@ class Program_weight_tensor_parameter_643: class Program_weight_tensor_parameter_644: name = "parameter_644" + original_name = "batch_norm2d_43.w_1" shape = [120] dtype = "float32" min_val = float("-2.67426") @@ -7095,6 +7740,7 @@ class Program_weight_tensor_parameter_644: class Program_weight_tensor_parameter_645: name = "parameter_645" + original_name = "conv2d_44.w_0" shape = [120, 120, 3, 3] dtype = "float32" min_val = float("-0.0778378") @@ -7106,6 +7752,7 @@ class Program_weight_tensor_parameter_645: class Program_weight_tensor_parameter_646: name = "parameter_646" + original_name = "batch_norm2d_42.b_0" shape = [120] dtype = "float32" min_val = float("-1.30641") @@ -7117,6 +7764,7 @@ class Program_weight_tensor_parameter_646: class Program_weight_tensor_parameter_647: name = "parameter_647" + original_name = "batch_norm2d_42.w_0" shape = [120] dtype = "float32" min_val = float("0.154756") @@ -7128,6 +7776,7 @@ class Program_weight_tensor_parameter_647: class Program_weight_tensor_parameter_648: name = "parameter_648" + original_name = "batch_norm2d_42.w_2" shape = [120] dtype = "float32" min_val = float("0.00019798") @@ -7139,6 +7788,7 @@ class Program_weight_tensor_parameter_648: class Program_weight_tensor_parameter_649: name = "parameter_649" + original_name = "batch_norm2d_42.w_1" shape = [120] dtype = "float32" min_val = float("-0.0407195") @@ -7150,6 +7800,7 @@ class Program_weight_tensor_parameter_649: class Program_weight_tensor_parameter_650: name = "parameter_650" + original_name = "conv2d_43.w_0" shape = [120, 120, 1, 1] dtype = "float32" min_val = float("-0.0452732") @@ -7161,6 +7812,7 @@ class Program_weight_tensor_parameter_650: class Program_weight_tensor_parameter_651: name = "parameter_651" + original_name = "batch_norm2d_41.b_0" shape = [120] dtype = "float32" min_val = float("-1.30641") @@ -7172,6 +7824,7 @@ class Program_weight_tensor_parameter_651: class Program_weight_tensor_parameter_652: name = "parameter_652" + original_name = "batch_norm2d_41.w_0" shape = [120] dtype = "float32" min_val = float("0.302508") @@ -7183,6 +7836,7 @@ class Program_weight_tensor_parameter_652: class Program_weight_tensor_parameter_653: name = "parameter_653" + original_name = "batch_norm2d_41.w_2" shape = [120] dtype = "float32" min_val = float("0.00108285") @@ -7194,6 +7848,7 @@ class Program_weight_tensor_parameter_653: class Program_weight_tensor_parameter_654: name = "parameter_654" + original_name = "batch_norm2d_41.w_1" shape = [120] dtype = "float32" min_val = float("-0.0639301") @@ -7205,6 +7860,7 @@ class Program_weight_tensor_parameter_654: class Program_weight_tensor_parameter_655: name = "parameter_655" + original_name = "conv2d_42.w_0" shape = [120, 120, 3, 3] dtype = "float32" min_val = float("-0.0330212") @@ -7216,6 +7872,7 @@ class Program_weight_tensor_parameter_655: class Program_weight_tensor_parameter_656: name = "parameter_656" + original_name = "batch_norm2d_40.b_0" shape = [120] dtype = "float32" min_val = float("-2.8918") @@ -7227,6 +7884,7 @@ class Program_weight_tensor_parameter_656: class Program_weight_tensor_parameter_657: name = "parameter_657" + original_name = "batch_norm2d_40.w_0" shape = [120] dtype = "float32" min_val = float("0.679434") @@ -7238,6 +7896,7 @@ class Program_weight_tensor_parameter_657: class Program_weight_tensor_parameter_658: name = "parameter_658" + original_name = "batch_norm2d_40.w_2" shape = [120] dtype = "float32" min_val = float("0.0297261") @@ -7249,6 +7908,7 @@ class Program_weight_tensor_parameter_658: class Program_weight_tensor_parameter_659: name = "parameter_659" + original_name = "batch_norm2d_40.w_1" shape = [120] dtype = "float32" min_val = float("-1.03502") @@ -7260,6 +7920,7 @@ class Program_weight_tensor_parameter_659: class Program_weight_tensor_parameter_660: name = "parameter_660" + original_name = "conv2d_41.w_0" shape = [120, 120, 3, 3] dtype = "float32" min_val = float("-0.057039") @@ -7271,6 +7932,7 @@ class Program_weight_tensor_parameter_660: class Program_weight_tensor_parameter_661: name = "parameter_661" + original_name = "batch_norm2d_39.b_0" shape = [120] dtype = "float32" min_val = float("-1.12876") @@ -7282,6 +7944,7 @@ class Program_weight_tensor_parameter_661: class Program_weight_tensor_parameter_662: name = "parameter_662" + original_name = "batch_norm2d_39.w_0" shape = [120] dtype = "float32" min_val = float("0.11841") @@ -7293,6 +7956,7 @@ class Program_weight_tensor_parameter_662: class Program_weight_tensor_parameter_663: name = "parameter_663" + original_name = "batch_norm2d_39.w_2" shape = [120] dtype = "float32" min_val = float("0.000165367") @@ -7304,6 +7968,7 @@ class Program_weight_tensor_parameter_663: class Program_weight_tensor_parameter_664: name = "parameter_664" + original_name = "batch_norm2d_39.w_1" shape = [120] dtype = "float32" min_val = float("-0.048304") @@ -7315,6 +7980,7 @@ class Program_weight_tensor_parameter_664: class Program_weight_tensor_parameter_665: name = "parameter_665" + original_name = "conv2d_40.w_0" shape = [120, 120, 1, 1] dtype = "float32" min_val = float("-0.045397") @@ -7326,6 +7992,7 @@ class Program_weight_tensor_parameter_665: class Program_weight_tensor_parameter_666: name = "parameter_666" + original_name = "batch_norm2d_38.b_0" shape = [120] dtype = "float32" min_val = float("-1.12876") @@ -7337,6 +8004,7 @@ class Program_weight_tensor_parameter_666: class Program_weight_tensor_parameter_667: name = "parameter_667" + original_name = "batch_norm2d_38.w_0" shape = [120] dtype = "float32" min_val = float("0.341959") @@ -7348,6 +8016,7 @@ class Program_weight_tensor_parameter_667: class Program_weight_tensor_parameter_668: name = "parameter_668" + original_name = "batch_norm2d_38.w_2" shape = [120] dtype = "float32" min_val = float("0.00176778") @@ -7359,6 +8028,7 @@ class Program_weight_tensor_parameter_668: class Program_weight_tensor_parameter_669: name = "parameter_669" + original_name = "batch_norm2d_38.w_1" shape = [120] dtype = "float32" min_val = float("-0.0655687") @@ -7370,6 +8040,7 @@ class Program_weight_tensor_parameter_669: class Program_weight_tensor_parameter_670: name = "parameter_670" + original_name = "conv2d_39.w_0" shape = [120, 120, 3, 3] dtype = "float32" min_val = float("-0.0283787") @@ -7381,6 +8052,7 @@ class Program_weight_tensor_parameter_670: class Program_weight_tensor_parameter_671: name = "parameter_671" + original_name = "batch_norm2d_37.b_0" shape = [120] dtype = "float32" min_val = float("-2.05842") @@ -7392,6 +8064,7 @@ class Program_weight_tensor_parameter_671: class Program_weight_tensor_parameter_672: name = "parameter_672" + original_name = "batch_norm2d_37.w_0" shape = [120] dtype = "float32" min_val = float("0.796267") @@ -7403,6 +8076,7 @@ class Program_weight_tensor_parameter_672: class Program_weight_tensor_parameter_673: name = "parameter_673" + original_name = "batch_norm2d_37.w_2" shape = [120] dtype = "float32" min_val = float("0.0258074") @@ -7414,6 +8088,7 @@ class Program_weight_tensor_parameter_673: class Program_weight_tensor_parameter_674: name = "parameter_674" + original_name = "batch_norm2d_37.w_1" shape = [120] dtype = "float32" min_val = float("-1.59118") @@ -7425,6 +8100,7 @@ class Program_weight_tensor_parameter_674: class Program_weight_tensor_parameter_675: name = "parameter_675" + original_name = "conv2d_38.w_0" shape = [120, 120, 3, 3] dtype = "float32" min_val = float("-0.0452673") @@ -7436,6 +8112,7 @@ class Program_weight_tensor_parameter_675: class Program_weight_tensor_parameter_676: name = "parameter_676" + original_name = "batch_norm2d_36.b_0" shape = [120] dtype = "float32" min_val = float("-1.12638") @@ -7447,6 +8124,7 @@ class Program_weight_tensor_parameter_676: class Program_weight_tensor_parameter_677: name = "parameter_677" + original_name = "batch_norm2d_36.w_0" shape = [120] dtype = "float32" min_val = float("0.108733") @@ -7458,6 +8136,7 @@ class Program_weight_tensor_parameter_677: class Program_weight_tensor_parameter_678: name = "parameter_678" + original_name = "batch_norm2d_36.w_2" shape = [120] dtype = "float32" min_val = float("0.000142339") @@ -7469,6 +8148,7 @@ class Program_weight_tensor_parameter_678: class Program_weight_tensor_parameter_679: name = "parameter_679" + original_name = "batch_norm2d_36.w_1" shape = [120] dtype = "float32" min_val = float("-0.0462951") @@ -7480,6 +8160,7 @@ class Program_weight_tensor_parameter_679: class Program_weight_tensor_parameter_680: name = "parameter_680" + original_name = "conv2d_37.w_0" shape = [120, 120, 1, 1] dtype = "float32" min_val = float("-0.0376082") @@ -7491,6 +8172,7 @@ class Program_weight_tensor_parameter_680: class Program_weight_tensor_parameter_681: name = "parameter_681" + original_name = "batch_norm2d_35.b_0" shape = [120] dtype = "float32" min_val = float("-1.12638") @@ -7502,6 +8184,7 @@ class Program_weight_tensor_parameter_681: class Program_weight_tensor_parameter_682: name = "parameter_682" + original_name = "batch_norm2d_35.w_0" shape = [120] dtype = "float32" min_val = float("0.319747") @@ -7513,6 +8196,7 @@ class Program_weight_tensor_parameter_682: class Program_weight_tensor_parameter_683: name = "parameter_683" + original_name = "batch_norm2d_35.w_2" shape = [120] dtype = "float32" min_val = float("0.00174393") @@ -7524,6 +8208,7 @@ class Program_weight_tensor_parameter_683: class Program_weight_tensor_parameter_684: name = "parameter_684" + original_name = "batch_norm2d_35.w_1" shape = [120] dtype = "float32" min_val = float("-0.0706752") @@ -7535,6 +8220,7 @@ class Program_weight_tensor_parameter_684: class Program_weight_tensor_parameter_685: name = "parameter_685" + original_name = "conv2d_36.w_0" shape = [120, 120, 3, 3] dtype = "float32" min_val = float("-0.0393867") @@ -7546,6 +8232,7 @@ class Program_weight_tensor_parameter_685: class Program_weight_tensor_parameter_686: name = "parameter_686" + original_name = "batch_norm2d_34.b_0" shape = [120] dtype = "float32" min_val = float("-2.89913") @@ -7557,6 +8244,7 @@ class Program_weight_tensor_parameter_686: class Program_weight_tensor_parameter_687: name = "parameter_687" + original_name = "batch_norm2d_34.w_0" shape = [120] dtype = "float32" min_val = float("0.797157") @@ -7568,6 +8256,7 @@ class Program_weight_tensor_parameter_687: class Program_weight_tensor_parameter_688: name = "parameter_688" + original_name = "batch_norm2d_34.w_2" shape = [120] dtype = "float32" min_val = float("0.0176762") @@ -7579,6 +8268,7 @@ class Program_weight_tensor_parameter_688: class Program_weight_tensor_parameter_689: name = "parameter_689" + original_name = "batch_norm2d_34.w_1" shape = [120] dtype = "float32" min_val = float("-0.487844") @@ -7590,6 +8280,7 @@ class Program_weight_tensor_parameter_689: class Program_weight_tensor_parameter_690: name = "parameter_690" + original_name = "conv2d_35.w_0" shape = [120, 120, 3, 3] dtype = "float32" min_val = float("-0.0462502") @@ -7601,6 +8292,7 @@ class Program_weight_tensor_parameter_690: class Program_weight_tensor_parameter_691: name = "parameter_691" + original_name = "batch_norm2d_33.b_0" shape = [120] dtype = "float32" min_val = float("-0.820993") @@ -7612,6 +8304,7 @@ class Program_weight_tensor_parameter_691: class Program_weight_tensor_parameter_692: name = "parameter_692" + original_name = "batch_norm2d_33.w_0" shape = [120] dtype = "float32" min_val = float("0.0586205") @@ -7623,6 +8316,7 @@ class Program_weight_tensor_parameter_692: class Program_weight_tensor_parameter_693: name = "parameter_693" + original_name = "batch_norm2d_33.w_2" shape = [120] dtype = "float32" min_val = float("7.46405e-05") @@ -7634,6 +8328,7 @@ class Program_weight_tensor_parameter_693: class Program_weight_tensor_parameter_694: name = "parameter_694" + original_name = "batch_norm2d_33.w_1" shape = [120] dtype = "float32" min_val = float("-0.0377561") @@ -7645,6 +8340,7 @@ class Program_weight_tensor_parameter_694: class Program_weight_tensor_parameter_695: name = "parameter_695" + original_name = "conv2d_34.w_0" shape = [120, 120, 1, 1] dtype = "float32" min_val = float("-0.0287656") @@ -7656,6 +8352,7 @@ class Program_weight_tensor_parameter_695: class Program_weight_tensor_parameter_696: name = "parameter_696" + original_name = "batch_norm2d_32.b_0" shape = [120] dtype = "float32" min_val = float("-0.820993") @@ -7667,6 +8364,7 @@ class Program_weight_tensor_parameter_696: class Program_weight_tensor_parameter_697: name = "parameter_697" + original_name = "batch_norm2d_32.w_0" shape = [120] dtype = "float32" min_val = float("0.249348") @@ -7678,6 +8376,7 @@ class Program_weight_tensor_parameter_697: class Program_weight_tensor_parameter_698: name = "parameter_698" + original_name = "batch_norm2d_32.w_2" shape = [120] dtype = "float32" min_val = float("0.00162025") @@ -7689,6 +8388,7 @@ class Program_weight_tensor_parameter_698: class Program_weight_tensor_parameter_699: name = "parameter_699" + original_name = "batch_norm2d_32.w_1" shape = [120] dtype = "float32" min_val = float("-0.0708755") @@ -7700,6 +8400,7 @@ class Program_weight_tensor_parameter_699: class Program_weight_tensor_parameter_700: name = "parameter_700" + original_name = "conv2d_33.w_0" shape = [120, 120, 3, 3] dtype = "float32" min_val = float("-0.0317724") @@ -7711,6 +8412,7 @@ class Program_weight_tensor_parameter_700: class Program_weight_tensor_parameter_701: name = "parameter_701" + original_name = "batch_norm2d_31.b_0" shape = [120] dtype = "float32" min_val = float("-2.40775") @@ -7722,6 +8424,7 @@ class Program_weight_tensor_parameter_701: class Program_weight_tensor_parameter_702: name = "parameter_702" + original_name = "batch_norm2d_31.w_0" shape = [120] dtype = "float32" min_val = float("0.697902") @@ -7733,6 +8436,7 @@ class Program_weight_tensor_parameter_702: class Program_weight_tensor_parameter_703: name = "parameter_703" + original_name = "batch_norm2d_31.w_2" shape = [120] dtype = "float32" min_val = float("0.0171996") @@ -7744,6 +8448,7 @@ class Program_weight_tensor_parameter_703: class Program_weight_tensor_parameter_704: name = "parameter_704" + original_name = "batch_norm2d_31.w_1" shape = [120] dtype = "float32" min_val = float("-0.355509") @@ -7755,6 +8460,7 @@ class Program_weight_tensor_parameter_704: class Program_weight_tensor_parameter_705: name = "parameter_705" + original_name = "conv2d_32.w_0" shape = [120, 120, 3, 3] dtype = "float32" min_val = float("-0.0519575") @@ -7766,6 +8472,7 @@ class Program_weight_tensor_parameter_705: class Program_weight_tensor_parameter_706: name = "parameter_706" + original_name = "batch_norm2d_30.b_0" shape = [120] dtype = "float32" min_val = float("-0.760209") @@ -7777,6 +8484,7 @@ class Program_weight_tensor_parameter_706: class Program_weight_tensor_parameter_707: name = "parameter_707" + original_name = "batch_norm2d_30.w_0" shape = [120] dtype = "float32" min_val = float("0.0754274") @@ -7788,6 +8496,7 @@ class Program_weight_tensor_parameter_707: class Program_weight_tensor_parameter_708: name = "parameter_708" + original_name = "batch_norm2d_30.w_2" shape = [120] dtype = "float32" min_val = float("0.00012338") @@ -7799,6 +8508,7 @@ class Program_weight_tensor_parameter_708: class Program_weight_tensor_parameter_709: name = "parameter_709" + original_name = "batch_norm2d_30.w_1" shape = [120] dtype = "float32" min_val = float("-0.0345532") @@ -7810,6 +8520,7 @@ class Program_weight_tensor_parameter_709: class Program_weight_tensor_parameter_710: name = "parameter_710" + original_name = "conv2d_31.w_0" shape = [120, 120, 1, 1] dtype = "float32" min_val = float("-0.0529746") @@ -7821,6 +8532,7 @@ class Program_weight_tensor_parameter_710: class Program_weight_tensor_parameter_711: name = "parameter_711" + original_name = "batch_norm2d_29.b_0" shape = [120] dtype = "float32" min_val = float("-0.760209") @@ -7832,6 +8544,7 @@ class Program_weight_tensor_parameter_711: class Program_weight_tensor_parameter_712: name = "parameter_712" + original_name = "batch_norm2d_29.w_0" shape = [120] dtype = "float32" min_val = float("0.259595") @@ -7843,6 +8556,7 @@ class Program_weight_tensor_parameter_712: class Program_weight_tensor_parameter_713: name = "parameter_713" + original_name = "batch_norm2d_29.w_2" shape = [120] dtype = "float32" min_val = float("0.00241346") @@ -7854,6 +8568,7 @@ class Program_weight_tensor_parameter_713: class Program_weight_tensor_parameter_714: name = "parameter_714" + original_name = "batch_norm2d_29.w_1" shape = [120] dtype = "float32" min_val = float("-0.0572858") @@ -7865,6 +8580,7 @@ class Program_weight_tensor_parameter_714: class Program_weight_tensor_parameter_715: name = "parameter_715" + original_name = "conv2d_30.w_0" shape = [120, 120, 3, 3] dtype = "float32" min_val = float("-0.036664") @@ -7876,6 +8592,7 @@ class Program_weight_tensor_parameter_715: class Program_weight_tensor_parameter_716: name = "parameter_716" + original_name = "batch_norm2d_28.b_0" shape = [120] dtype = "float32" min_val = float("-3.24829") @@ -7887,6 +8604,7 @@ class Program_weight_tensor_parameter_716: class Program_weight_tensor_parameter_717: name = "parameter_717" + original_name = "batch_norm2d_28.w_0" shape = [120] dtype = "float32" min_val = float("0.714517") @@ -7898,6 +8616,7 @@ class Program_weight_tensor_parameter_717: class Program_weight_tensor_parameter_718: name = "parameter_718" + original_name = "batch_norm2d_28.w_2" shape = [120] dtype = "float32" min_val = float("0.0111219") @@ -7909,6 +8628,7 @@ class Program_weight_tensor_parameter_718: class Program_weight_tensor_parameter_719: name = "parameter_719" + original_name = "batch_norm2d_28.w_1" shape = [120] dtype = "float32" min_val = float("-0.646808") @@ -7920,6 +8640,7 @@ class Program_weight_tensor_parameter_719: class Program_weight_tensor_parameter_720: name = "parameter_720" + original_name = "conv2d_29.w_0" shape = [120, 120, 3, 3] dtype = "float32" min_val = float("-0.0612844") @@ -7931,6 +8652,7 @@ class Program_weight_tensor_parameter_720: class Program_weight_tensor_parameter_721: name = "parameter_721" + original_name = "batch_norm2d_27.b_0" shape = [120] dtype = "float32" min_val = float("-0.887183") @@ -7942,6 +8664,7 @@ class Program_weight_tensor_parameter_721: class Program_weight_tensor_parameter_722: name = "parameter_722" + original_name = "batch_norm2d_27.w_0" shape = [120] dtype = "float32" min_val = float("0.0599461") @@ -7953,6 +8676,7 @@ class Program_weight_tensor_parameter_722: class Program_weight_tensor_parameter_723: name = "parameter_723" + original_name = "batch_norm2d_27.w_2" shape = [120] dtype = "float32" min_val = float("0.000118466") @@ -7964,6 +8688,7 @@ class Program_weight_tensor_parameter_723: class Program_weight_tensor_parameter_724: name = "parameter_724" + original_name = "batch_norm2d_27.w_1" shape = [120] dtype = "float32" min_val = float("-0.0269276") @@ -7975,6 +8700,7 @@ class Program_weight_tensor_parameter_724: class Program_weight_tensor_parameter_725: name = "parameter_725" + original_name = "conv2d_28.w_0" shape = [120, 120, 1, 1] dtype = "float32" min_val = float("-0.0356341") @@ -7986,6 +8712,7 @@ class Program_weight_tensor_parameter_725: class Program_weight_tensor_parameter_726: name = "parameter_726" + original_name = "batch_norm2d_26.b_0" shape = [120] dtype = "float32" min_val = float("-0.887183") @@ -7997,6 +8724,7 @@ class Program_weight_tensor_parameter_726: class Program_weight_tensor_parameter_727: name = "parameter_727" + original_name = "batch_norm2d_26.w_0" shape = [120] dtype = "float32" min_val = float("0.252246") @@ -8008,6 +8736,7 @@ class Program_weight_tensor_parameter_727: class Program_weight_tensor_parameter_728: name = "parameter_728" + original_name = "batch_norm2d_26.w_2" shape = [120] dtype = "float32" min_val = float("0.00322862") @@ -8019,6 +8748,7 @@ class Program_weight_tensor_parameter_728: class Program_weight_tensor_parameter_729: name = "parameter_729" + original_name = "batch_norm2d_26.w_1" shape = [120] dtype = "float32" min_val = float("-0.0624692") @@ -8030,6 +8760,7 @@ class Program_weight_tensor_parameter_729: class Program_weight_tensor_parameter_730: name = "parameter_730" + original_name = "conv2d_27.w_0" shape = [120, 120, 3, 3] dtype = "float32" min_val = float("-0.0333751") @@ -8041,6 +8772,7 @@ class Program_weight_tensor_parameter_730: class Program_weight_tensor_parameter_731: name = "parameter_731" + original_name = "batch_norm2d_25.b_0" shape = [120] dtype = "float32" min_val = float("-3.23994") @@ -8052,6 +8784,7 @@ class Program_weight_tensor_parameter_731: class Program_weight_tensor_parameter_732: name = "parameter_732" + original_name = "batch_norm2d_25.w_0" shape = [120] dtype = "float32" min_val = float("0.836157") @@ -8063,6 +8796,7 @@ class Program_weight_tensor_parameter_732: class Program_weight_tensor_parameter_733: name = "parameter_733" + original_name = "batch_norm2d_25.w_2" shape = [120] dtype = "float32" min_val = float("0.00906578") @@ -8074,6 +8808,7 @@ class Program_weight_tensor_parameter_733: class Program_weight_tensor_parameter_734: name = "parameter_734" + original_name = "batch_norm2d_25.w_1" shape = [120] dtype = "float32" min_val = float("-0.403111") @@ -8085,6 +8820,7 @@ class Program_weight_tensor_parameter_734: class Program_weight_tensor_parameter_735: name = "parameter_735" + original_name = "conv2d_26.w_0" shape = [120, 120, 3, 3] dtype = "float32" min_val = float("-0.0679836") @@ -8096,6 +8832,7 @@ class Program_weight_tensor_parameter_735: class Program_weight_tensor_parameter_736: name = "parameter_736" + original_name = "batch_norm2d_24.b_0" shape = [120] dtype = "float32" min_val = float("-0.984259") @@ -8107,6 +8844,7 @@ class Program_weight_tensor_parameter_736: class Program_weight_tensor_parameter_737: name = "parameter_737" + original_name = "batch_norm2d_24.w_0" shape = [120] dtype = "float32" min_val = float("0.0041575") @@ -8118,6 +8856,7 @@ class Program_weight_tensor_parameter_737: class Program_weight_tensor_parameter_738: name = "parameter_738" + original_name = "batch_norm2d_24.w_2" shape = [120] dtype = "float32" min_val = float("3.24611e-06") @@ -8129,6 +8868,7 @@ class Program_weight_tensor_parameter_738: class Program_weight_tensor_parameter_739: name = "parameter_739" + original_name = "batch_norm2d_24.w_1" shape = [120] dtype = "float32" min_val = float("-0.0167298") @@ -8140,6 +8880,7 @@ class Program_weight_tensor_parameter_739: class Program_weight_tensor_parameter_740: name = "parameter_740" + original_name = "conv2d_25.w_0" shape = [120, 120, 1, 1] dtype = "float32" min_val = float("-0.075412") @@ -8151,6 +8892,7 @@ class Program_weight_tensor_parameter_740: class Program_weight_tensor_parameter_741: name = "parameter_741" + original_name = "batch_norm2d_23.b_0" shape = [120] dtype = "float32" min_val = float("-0.984259") @@ -8162,6 +8904,7 @@ class Program_weight_tensor_parameter_741: class Program_weight_tensor_parameter_742: name = "parameter_742" + original_name = "batch_norm2d_23.w_0" shape = [120] dtype = "float32" min_val = float("0.261123") @@ -8173,6 +8916,7 @@ class Program_weight_tensor_parameter_742: class Program_weight_tensor_parameter_743: name = "parameter_743" + original_name = "batch_norm2d_23.w_2" shape = [120] dtype = "float32" min_val = float("0.00595463") @@ -8184,6 +8928,7 @@ class Program_weight_tensor_parameter_743: class Program_weight_tensor_parameter_744: name = "parameter_744" + original_name = "batch_norm2d_23.w_1" shape = [120] dtype = "float32" min_val = float("-0.0597811") @@ -8195,6 +8940,7 @@ class Program_weight_tensor_parameter_744: class Program_weight_tensor_parameter_745: name = "parameter_745" + original_name = "conv2d_24.w_0" shape = [120, 120, 3, 3] dtype = "float32" min_val = float("-0.0413222") @@ -8206,6 +8952,7 @@ class Program_weight_tensor_parameter_745: class Program_weight_tensor_parameter_746: name = "parameter_746" + original_name = "batch_norm2d_22.b_0" shape = [120] dtype = "float32" min_val = float("-2.01669") @@ -8217,6 +8964,7 @@ class Program_weight_tensor_parameter_746: class Program_weight_tensor_parameter_747: name = "parameter_747" + original_name = "batch_norm2d_22.w_0" shape = [120] dtype = "float32" min_val = float("0.423662") @@ -8228,6 +8976,7 @@ class Program_weight_tensor_parameter_747: class Program_weight_tensor_parameter_748: name = "parameter_748" + original_name = "batch_norm2d_22.w_2" shape = [120] dtype = "float32" min_val = float("0.00827892") @@ -8239,6 +8988,7 @@ class Program_weight_tensor_parameter_748: class Program_weight_tensor_parameter_749: name = "parameter_749" + original_name = "batch_norm2d_22.w_1" shape = [120] dtype = "float32" min_val = float("-0.504096") @@ -8250,6 +9000,7 @@ class Program_weight_tensor_parameter_749: class Program_weight_tensor_parameter_750: name = "parameter_750" + original_name = "conv2d_23.w_0" shape = [120, 120, 3, 3] dtype = "float32" min_val = float("-0.116238") @@ -8261,6 +9012,7 @@ class Program_weight_tensor_parameter_750: class Program_weight_tensor_parameter_751: name = "parameter_751" + original_name = "batch_norm2d_21.b_0" shape = [120] dtype = "float32" min_val = float("-1.13353") @@ -8272,6 +9024,7 @@ class Program_weight_tensor_parameter_751: class Program_weight_tensor_parameter_752: name = "parameter_752" + original_name = "batch_norm2d_21.w_0" shape = [120] dtype = "float32" min_val = float("0.361754") @@ -8283,6 +9036,7 @@ class Program_weight_tensor_parameter_752: class Program_weight_tensor_parameter_753: name = "parameter_753" + original_name = "batch_norm2d_21.w_2" shape = [120] dtype = "float32" min_val = float("0.00697088") @@ -8294,6 +9048,7 @@ class Program_weight_tensor_parameter_753: class Program_weight_tensor_parameter_754: name = "parameter_754" + original_name = "batch_norm2d_21.w_1" shape = [120] dtype = "float32" min_val = float("-0.219529") @@ -8305,6 +9060,7 @@ class Program_weight_tensor_parameter_754: class Program_weight_tensor_parameter_755: name = "parameter_755" + original_name = "conv2d_22.w_0" shape = [120, 240, 1, 1] dtype = "float32" min_val = float("-0.127414") @@ -8316,6 +9072,7 @@ class Program_weight_tensor_parameter_755: class Program_weight_tensor_parameter_756: name = "parameter_756" + original_name = "batch_norm2d_20.b_0" shape = [120] dtype = "float32" min_val = float("-3.7228") @@ -8327,6 +9084,7 @@ class Program_weight_tensor_parameter_756: class Program_weight_tensor_parameter_757: name = "parameter_757" + original_name = "batch_norm2d_20.w_0" shape = [120] dtype = "float32" min_val = float("0.699471") @@ -8338,6 +9096,7 @@ class Program_weight_tensor_parameter_757: class Program_weight_tensor_parameter_758: name = "parameter_758" + original_name = "batch_norm2d_20.w_2" shape = [120] dtype = "float32" min_val = float("0.00475616") @@ -8349,6 +9108,7 @@ class Program_weight_tensor_parameter_758: class Program_weight_tensor_parameter_759: name = "parameter_759" + original_name = "batch_norm2d_20.w_1" shape = [120] dtype = "float32" min_val = float("-0.164895") @@ -8360,6 +9120,7 @@ class Program_weight_tensor_parameter_759: class Program_weight_tensor_parameter_760: name = "parameter_760" + original_name = "conv2d_21.w_0" shape = [120, 240, 1, 1] dtype = "float32" min_val = float("-0.101223") @@ -8371,6 +9132,7 @@ class Program_weight_tensor_parameter_760: class Program_weight_tensor_parameter_761: name = "parameter_761" + original_name = "batch_norm2d_19.b_0" shape = [240] dtype = "float32" min_val = float("-2.87978") @@ -8382,6 +9144,7 @@ class Program_weight_tensor_parameter_761: class Program_weight_tensor_parameter_762: name = "parameter_762" + original_name = "batch_norm2d_19.w_0" shape = [240] dtype = "float32" min_val = float("0.621601") @@ -8393,6 +9156,7 @@ class Program_weight_tensor_parameter_762: class Program_weight_tensor_parameter_763: name = "parameter_763" + original_name = "batch_norm2d_19.w_2" shape = [240] dtype = "float32" min_val = float("0.00715788") @@ -8404,6 +9168,7 @@ class Program_weight_tensor_parameter_763: class Program_weight_tensor_parameter_764: name = "parameter_764" + original_name = "batch_norm2d_19.w_1" shape = [240] dtype = "float32" min_val = float("-0.374984") @@ -8415,6 +9180,7 @@ class Program_weight_tensor_parameter_764: class Program_weight_tensor_parameter_765: name = "parameter_765" + original_name = "conv2d_20.w_0" shape = [240, 160, 3, 3] dtype = "float32" min_val = float("-0.0763") @@ -8426,6 +9192,7 @@ class Program_weight_tensor_parameter_765: class Program_weight_tensor_parameter_766: name = "parameter_766" + original_name = "batch_norm2d_18.b_0" shape = [160] dtype = "float32" min_val = float("-1.90297") @@ -8437,6 +9204,7 @@ class Program_weight_tensor_parameter_766: class Program_weight_tensor_parameter_767: name = "parameter_767" + original_name = "batch_norm2d_18.w_0" shape = [160] dtype = "float32" min_val = float("0.445916") @@ -8448,6 +9216,7 @@ class Program_weight_tensor_parameter_767: class Program_weight_tensor_parameter_768: name = "parameter_768" + original_name = "batch_norm2d_18.w_2" shape = [160] dtype = "float32" min_val = float("0.000555214") @@ -8459,6 +9228,7 @@ class Program_weight_tensor_parameter_768: class Program_weight_tensor_parameter_769: name = "parameter_769" + original_name = "batch_norm2d_18.w_1" shape = [160] dtype = "float32" min_val = float("-0.294926") @@ -8470,6 +9240,7 @@ class Program_weight_tensor_parameter_769: class Program_weight_tensor_parameter_770: name = "parameter_770" + original_name = "conv2d_19.w_0" shape = [160, 120, 1, 1] dtype = "float32" min_val = float("-0.243235") @@ -8481,6 +9252,7 @@ class Program_weight_tensor_parameter_770: class Program_weight_tensor_parameter_771: name = "parameter_771" + original_name = "conv2d_18.b_0" shape = [120] dtype = "float32" min_val = float("-0.0131591") @@ -8492,6 +9264,7 @@ class Program_weight_tensor_parameter_771: class Program_weight_tensor_parameter_772: name = "parameter_772" + original_name = "conv2d_18.w_0" shape = [120, 120, 1, 1] dtype = "float32" min_val = float("-0.333481") @@ -8503,6 +9276,7 @@ class Program_weight_tensor_parameter_772: class Program_weight_tensor_parameter_773: name = "parameter_773" + original_name = "batch_norm2d_17.b_0" shape = [60] dtype = "float32" min_val = float("0") @@ -8512,6 +9286,7 @@ class Program_weight_tensor_parameter_773: class Program_weight_tensor_parameter_774: name = "parameter_774" + original_name = "batch_norm2d_17.w_0" shape = [60] dtype = "float32" min_val = float("0") @@ -8521,6 +9296,7 @@ class Program_weight_tensor_parameter_774: class Program_weight_tensor_parameter_775: name = "parameter_775" + original_name = "batch_norm2d_17.w_2" shape = [60] dtype = "float32" min_val = float("0") @@ -8530,6 +9306,7 @@ class Program_weight_tensor_parameter_775: class Program_weight_tensor_parameter_776: name = "parameter_776" + original_name = "batch_norm2d_17.w_1" shape = [60] dtype = "float32" min_val = float("0") @@ -8539,6 +9316,7 @@ class Program_weight_tensor_parameter_776: class Program_weight_tensor_parameter_777: name = "parameter_777" + original_name = "conv2d_17.w_0" shape = [60, 60, 1, 1] dtype = "float32" min_val = float("-0.0488272") @@ -8550,6 +9328,7 @@ class Program_weight_tensor_parameter_777: class Program_weight_tensor_parameter_778: name = "parameter_778" + original_name = "batch_norm2d_16.b_0" shape = [60] dtype = "float32" min_val = float("0") @@ -8559,6 +9338,7 @@ class Program_weight_tensor_parameter_778: class Program_weight_tensor_parameter_779: name = "parameter_779" + original_name = "batch_norm2d_16.w_0" shape = [60] dtype = "float32" min_val = float("0") @@ -8568,6 +9348,7 @@ class Program_weight_tensor_parameter_779: class Program_weight_tensor_parameter_780: name = "parameter_780" + original_name = "batch_norm2d_16.w_2" shape = [60] dtype = "float32" min_val = float("0") @@ -8577,6 +9358,7 @@ class Program_weight_tensor_parameter_780: class Program_weight_tensor_parameter_781: name = "parameter_781" + original_name = "batch_norm2d_16.w_1" shape = [60] dtype = "float32" min_val = float("0") @@ -8586,6 +9368,7 @@ class Program_weight_tensor_parameter_781: class Program_weight_tensor_parameter_782: name = "parameter_782" + original_name = "conv2d_16.w_0" shape = [60, 60, 3, 3] dtype = "float32" min_val = float("-0.049959") @@ -8597,6 +9380,7 @@ class Program_weight_tensor_parameter_782: class Program_weight_tensor_parameter_783: name = "parameter_783" + original_name = "batch_norm2d_15.b_0" shape = [60] dtype = "float32" min_val = float("0") @@ -8606,6 +9390,7 @@ class Program_weight_tensor_parameter_783: class Program_weight_tensor_parameter_784: name = "parameter_784" + original_name = "batch_norm2d_15.w_0" shape = [60] dtype = "float32" min_val = float("0") @@ -8615,6 +9400,7 @@ class Program_weight_tensor_parameter_784: class Program_weight_tensor_parameter_785: name = "parameter_785" + original_name = "batch_norm2d_15.w_2" shape = [60] dtype = "float32" min_val = float("0") @@ -8624,6 +9410,7 @@ class Program_weight_tensor_parameter_785: class Program_weight_tensor_parameter_786: name = "parameter_786" + original_name = "batch_norm2d_15.w_1" shape = [60] dtype = "float32" min_val = float("0") @@ -8633,6 +9420,7 @@ class Program_weight_tensor_parameter_786: class Program_weight_tensor_parameter_787: name = "parameter_787" + original_name = "conv2d_15.w_0" shape = [60, 60, 3, 3] dtype = "float32" min_val = float("-0.0964322") @@ -8644,6 +9432,7 @@ class Program_weight_tensor_parameter_787: class Program_weight_tensor_parameter_788: name = "parameter_788" + original_name = "batch_norm2d_14.b_0" shape = [60] dtype = "float32" min_val = float("0") @@ -8653,6 +9442,7 @@ class Program_weight_tensor_parameter_788: class Program_weight_tensor_parameter_789: name = "parameter_789" + original_name = "batch_norm2d_14.w_0" shape = [60] dtype = "float32" min_val = float("0") @@ -8662,6 +9452,7 @@ class Program_weight_tensor_parameter_789: class Program_weight_tensor_parameter_790: name = "parameter_790" + original_name = "batch_norm2d_14.w_2" shape = [60] dtype = "float32" min_val = float("0") @@ -8671,6 +9462,7 @@ class Program_weight_tensor_parameter_790: class Program_weight_tensor_parameter_791: name = "parameter_791" + original_name = "batch_norm2d_14.w_1" shape = [60] dtype = "float32" min_val = float("0") @@ -8680,6 +9472,7 @@ class Program_weight_tensor_parameter_791: class Program_weight_tensor_parameter_792: name = "parameter_792" + original_name = "conv2d_14.w_0" shape = [60, 60, 1, 1] dtype = "float32" min_val = float("-0.0714922") @@ -8691,6 +9484,7 @@ class Program_weight_tensor_parameter_792: class Program_weight_tensor_parameter_793: name = "parameter_793" + original_name = "batch_norm2d_13.b_0" shape = [60] dtype = "float32" min_val = float("0") @@ -8700,6 +9494,7 @@ class Program_weight_tensor_parameter_793: class Program_weight_tensor_parameter_794: name = "parameter_794" + original_name = "batch_norm2d_13.w_0" shape = [60] dtype = "float32" min_val = float("0") @@ -8709,6 +9504,7 @@ class Program_weight_tensor_parameter_794: class Program_weight_tensor_parameter_795: name = "parameter_795" + original_name = "batch_norm2d_13.w_2" shape = [60] dtype = "float32" min_val = float("0") @@ -8718,6 +9514,7 @@ class Program_weight_tensor_parameter_795: class Program_weight_tensor_parameter_796: name = "parameter_796" + original_name = "batch_norm2d_13.w_1" shape = [60] dtype = "float32" min_val = float("0") @@ -8727,6 +9524,7 @@ class Program_weight_tensor_parameter_796: class Program_weight_tensor_parameter_797: name = "parameter_797" + original_name = "conv2d_13.w_0" shape = [60, 60, 3, 3] dtype = "float32" min_val = float("-0.0699195") @@ -8738,6 +9536,7 @@ class Program_weight_tensor_parameter_797: class Program_weight_tensor_parameter_798: name = "parameter_798" + original_name = "batch_norm2d_12.b_0" shape = [60] dtype = "float32" min_val = float("0") @@ -8747,6 +9546,7 @@ class Program_weight_tensor_parameter_798: class Program_weight_tensor_parameter_799: name = "parameter_799" + original_name = "batch_norm2d_12.w_0" shape = [60] dtype = "float32" min_val = float("0") @@ -8756,6 +9556,7 @@ class Program_weight_tensor_parameter_799: class Program_weight_tensor_parameter_800: name = "parameter_800" + original_name = "batch_norm2d_12.w_2" shape = [60] dtype = "float32" min_val = float("0") @@ -8765,6 +9566,7 @@ class Program_weight_tensor_parameter_800: class Program_weight_tensor_parameter_801: name = "parameter_801" + original_name = "batch_norm2d_12.w_1" shape = [60] dtype = "float32" min_val = float("0") @@ -8774,6 +9576,7 @@ class Program_weight_tensor_parameter_801: class Program_weight_tensor_parameter_802: name = "parameter_802" + original_name = "conv2d_12.w_0" shape = [60, 60, 3, 3] dtype = "float32" min_val = float("-0.0782983") @@ -8785,6 +9588,7 @@ class Program_weight_tensor_parameter_802: class Program_weight_tensor_parameter_803: name = "parameter_803" + original_name = "batch_norm2d_11.b_0" shape = [60] dtype = "float32" min_val = float("0") @@ -8794,6 +9598,7 @@ class Program_weight_tensor_parameter_803: class Program_weight_tensor_parameter_804: name = "parameter_804" + original_name = "batch_norm2d_11.w_0" shape = [60] dtype = "float32" min_val = float("0") @@ -8803,6 +9608,7 @@ class Program_weight_tensor_parameter_804: class Program_weight_tensor_parameter_805: name = "parameter_805" + original_name = "batch_norm2d_11.w_2" shape = [60] dtype = "float32" min_val = float("0") @@ -8812,6 +9618,7 @@ class Program_weight_tensor_parameter_805: class Program_weight_tensor_parameter_806: name = "parameter_806" + original_name = "batch_norm2d_11.w_1" shape = [60] dtype = "float32" min_val = float("0") @@ -8821,6 +9628,7 @@ class Program_weight_tensor_parameter_806: class Program_weight_tensor_parameter_807: name = "parameter_807" + original_name = "conv2d_11.w_0" shape = [60, 60, 1, 1] dtype = "float32" min_val = float("-0.0781044") @@ -8832,6 +9640,7 @@ class Program_weight_tensor_parameter_807: class Program_weight_tensor_parameter_808: name = "parameter_808" + original_name = "batch_norm2d_10.b_0" shape = [60] dtype = "float32" min_val = float("0") @@ -8841,6 +9650,7 @@ class Program_weight_tensor_parameter_808: class Program_weight_tensor_parameter_809: name = "parameter_809" + original_name = "batch_norm2d_10.w_0" shape = [60] dtype = "float32" min_val = float("0") @@ -8850,6 +9660,7 @@ class Program_weight_tensor_parameter_809: class Program_weight_tensor_parameter_810: name = "parameter_810" + original_name = "batch_norm2d_10.w_2" shape = [60] dtype = "float32" min_val = float("0") @@ -8859,6 +9670,7 @@ class Program_weight_tensor_parameter_810: class Program_weight_tensor_parameter_811: name = "parameter_811" + original_name = "batch_norm2d_10.w_1" shape = [60] dtype = "float32" min_val = float("0") @@ -8868,6 +9680,7 @@ class Program_weight_tensor_parameter_811: class Program_weight_tensor_parameter_812: name = "parameter_812" + original_name = "conv2d_10.w_0" shape = [60, 60, 3, 3] dtype = "float32" min_val = float("-0.0736361") @@ -8879,6 +9692,7 @@ class Program_weight_tensor_parameter_812: class Program_weight_tensor_parameter_813: name = "parameter_813" + original_name = "batch_norm2d_9.b_0" shape = [60] dtype = "float32" min_val = float("0") @@ -8888,6 +9702,7 @@ class Program_weight_tensor_parameter_813: class Program_weight_tensor_parameter_814: name = "parameter_814" + original_name = "batch_norm2d_9.w_0" shape = [60] dtype = "float32" min_val = float("0") @@ -8897,6 +9712,7 @@ class Program_weight_tensor_parameter_814: class Program_weight_tensor_parameter_815: name = "parameter_815" + original_name = "batch_norm2d_9.w_2" shape = [60] dtype = "float32" min_val = float("0") @@ -8906,6 +9722,7 @@ class Program_weight_tensor_parameter_815: class Program_weight_tensor_parameter_816: name = "parameter_816" + original_name = "batch_norm2d_9.w_1" shape = [60] dtype = "float32" min_val = float("0") @@ -8915,6 +9732,7 @@ class Program_weight_tensor_parameter_816: class Program_weight_tensor_parameter_817: name = "parameter_817" + original_name = "conv2d_9.w_0" shape = [60, 60, 3, 3] dtype = "float32" min_val = float("-0.0815527") @@ -8926,6 +9744,7 @@ class Program_weight_tensor_parameter_817: class Program_weight_tensor_parameter_818: name = "parameter_818" + original_name = "batch_norm2d_8.b_0" shape = [60] dtype = "float32" min_val = float("0") @@ -8935,6 +9754,7 @@ class Program_weight_tensor_parameter_818: class Program_weight_tensor_parameter_819: name = "parameter_819" + original_name = "batch_norm2d_8.w_0" shape = [60] dtype = "float32" min_val = float("0") @@ -8944,6 +9764,7 @@ class Program_weight_tensor_parameter_819: class Program_weight_tensor_parameter_820: name = "parameter_820" + original_name = "batch_norm2d_8.w_2" shape = [60] dtype = "float32" min_val = float("0") @@ -8953,6 +9774,7 @@ class Program_weight_tensor_parameter_820: class Program_weight_tensor_parameter_821: name = "parameter_821" + original_name = "batch_norm2d_8.w_1" shape = [60] dtype = "float32" min_val = float("0") @@ -8962,6 +9784,7 @@ class Program_weight_tensor_parameter_821: class Program_weight_tensor_parameter_822: name = "parameter_822" + original_name = "conv2d_8.w_0" shape = [60, 60, 1, 1] dtype = "float32" min_val = float("-0.101925") @@ -8973,6 +9796,7 @@ class Program_weight_tensor_parameter_822: class Program_weight_tensor_parameter_823: name = "parameter_823" + original_name = "batch_norm2d_7.b_0" shape = [60] dtype = "float32" min_val = float("0") @@ -8982,6 +9806,7 @@ class Program_weight_tensor_parameter_823: class Program_weight_tensor_parameter_824: name = "parameter_824" + original_name = "batch_norm2d_7.w_0" shape = [60] dtype = "float32" min_val = float("0") @@ -8991,6 +9816,7 @@ class Program_weight_tensor_parameter_824: class Program_weight_tensor_parameter_825: name = "parameter_825" + original_name = "batch_norm2d_7.w_2" shape = [60] dtype = "float32" min_val = float("0") @@ -9000,6 +9826,7 @@ class Program_weight_tensor_parameter_825: class Program_weight_tensor_parameter_826: name = "parameter_826" + original_name = "batch_norm2d_7.w_1" shape = [60] dtype = "float32" min_val = float("0") @@ -9009,6 +9836,7 @@ class Program_weight_tensor_parameter_826: class Program_weight_tensor_parameter_827: name = "parameter_827" + original_name = "conv2d_7.w_0" shape = [60, 60, 3, 3] dtype = "float32" min_val = float("-0.0732117") @@ -9020,6 +9848,7 @@ class Program_weight_tensor_parameter_827: class Program_weight_tensor_parameter_828: name = "parameter_828" + original_name = "batch_norm2d_6.b_0" shape = [60] dtype = "float32" min_val = float("0") @@ -9029,6 +9858,7 @@ class Program_weight_tensor_parameter_828: class Program_weight_tensor_parameter_829: name = "parameter_829" + original_name = "batch_norm2d_6.w_0" shape = [60] dtype = "float32" min_val = float("0") @@ -9038,6 +9868,7 @@ class Program_weight_tensor_parameter_829: class Program_weight_tensor_parameter_830: name = "parameter_830" + original_name = "batch_norm2d_6.w_2" shape = [60] dtype = "float32" min_val = float("0") @@ -9047,6 +9878,7 @@ class Program_weight_tensor_parameter_830: class Program_weight_tensor_parameter_831: name = "parameter_831" + original_name = "batch_norm2d_6.w_1" shape = [60] dtype = "float32" min_val = float("0") @@ -9056,6 +9888,7 @@ class Program_weight_tensor_parameter_831: class Program_weight_tensor_parameter_832: name = "parameter_832" + original_name = "conv2d_6.w_0" shape = [60, 60, 3, 3] dtype = "float32" min_val = float("-0.114812") @@ -9067,6 +9900,7 @@ class Program_weight_tensor_parameter_832: class Program_weight_tensor_parameter_833: name = "parameter_833" + original_name = "batch_norm2d_5.b_0" shape = [60] dtype = "float32" min_val = float("0") @@ -9076,6 +9910,7 @@ class Program_weight_tensor_parameter_833: class Program_weight_tensor_parameter_834: name = "parameter_834" + original_name = "batch_norm2d_5.w_0" shape = [60] dtype = "float32" min_val = float("0") @@ -9085,6 +9920,7 @@ class Program_weight_tensor_parameter_834: class Program_weight_tensor_parameter_835: name = "parameter_835" + original_name = "batch_norm2d_5.w_2" shape = [60] dtype = "float32" min_val = float("0") @@ -9094,6 +9930,7 @@ class Program_weight_tensor_parameter_835: class Program_weight_tensor_parameter_836: name = "parameter_836" + original_name = "batch_norm2d_5.w_1" shape = [60] dtype = "float32" min_val = float("0") @@ -9103,6 +9940,7 @@ class Program_weight_tensor_parameter_836: class Program_weight_tensor_parameter_837: name = "parameter_837" + original_name = "conv2d_5.w_0" shape = [60, 120, 1, 1] dtype = "float32" min_val = float("-0.137762") @@ -9114,6 +9952,7 @@ class Program_weight_tensor_parameter_837: class Program_weight_tensor_parameter_838: name = "parameter_838" + original_name = "batch_norm2d_4.b_0" shape = [60] dtype = "float32" min_val = float("0") @@ -9123,6 +9962,7 @@ class Program_weight_tensor_parameter_838: class Program_weight_tensor_parameter_839: name = "parameter_839" + original_name = "batch_norm2d_4.w_0" shape = [60] dtype = "float32" min_val = float("0") @@ -9132,6 +9972,7 @@ class Program_weight_tensor_parameter_839: class Program_weight_tensor_parameter_840: name = "parameter_840" + original_name = "batch_norm2d_4.w_2" shape = [60] dtype = "float32" min_val = float("0") @@ -9141,6 +9982,7 @@ class Program_weight_tensor_parameter_840: class Program_weight_tensor_parameter_841: name = "parameter_841" + original_name = "batch_norm2d_4.w_1" shape = [60] dtype = "float32" min_val = float("0") @@ -9150,6 +9992,7 @@ class Program_weight_tensor_parameter_841: class Program_weight_tensor_parameter_842: name = "parameter_842" + original_name = "conv2d_4.w_0" shape = [60, 120, 1, 1] dtype = "float32" min_val = float("-0.138787") @@ -9161,6 +10004,7 @@ class Program_weight_tensor_parameter_842: class Program_weight_tensor_parameter_843: name = "parameter_843" + original_name = "batch_norm2d_3.b_0" shape = [120] dtype = "float32" min_val = float("-1.90853") @@ -9172,6 +10016,7 @@ class Program_weight_tensor_parameter_843: class Program_weight_tensor_parameter_844: name = "parameter_844" + original_name = "batch_norm2d_3.w_0" shape = [120] dtype = "float32" min_val = float("1.07034") @@ -9183,6 +10028,7 @@ class Program_weight_tensor_parameter_844: class Program_weight_tensor_parameter_845: name = "parameter_845" + original_name = "batch_norm2d_3.w_2" shape = [120] dtype = "float32" min_val = float("0.347682") @@ -9194,6 +10040,7 @@ class Program_weight_tensor_parameter_845: class Program_weight_tensor_parameter_846: name = "parameter_846" + original_name = "batch_norm2d_3.w_1" shape = [120] dtype = "float32" min_val = float("-1.73874") @@ -9205,6 +10052,7 @@ class Program_weight_tensor_parameter_846: class Program_weight_tensor_parameter_847: name = "parameter_847" + original_name = "conv2d_3.w_0" shape = [120, 80, 3, 3] dtype = "float32" min_val = float("-0.0878422") @@ -9216,6 +10064,7 @@ class Program_weight_tensor_parameter_847: class Program_weight_tensor_parameter_848: name = "parameter_848" + original_name = "batch_norm2d_2.b_0" shape = [80] dtype = "float32" min_val = float("-3.04441") @@ -9227,6 +10076,7 @@ class Program_weight_tensor_parameter_848: class Program_weight_tensor_parameter_849: name = "parameter_849" + original_name = "batch_norm2d_2.w_0" shape = [80] dtype = "float32" min_val = float("1.91145") @@ -9238,6 +10088,7 @@ class Program_weight_tensor_parameter_849: class Program_weight_tensor_parameter_850: name = "parameter_850" + original_name = "batch_norm2d_2.w_2" shape = [80] dtype = "float32" min_val = float("0.299022") @@ -9249,6 +10100,7 @@ class Program_weight_tensor_parameter_850: class Program_weight_tensor_parameter_851: name = "parameter_851" + original_name = "batch_norm2d_2.w_1" shape = [80] dtype = "float32" min_val = float("-1.27152") @@ -9260,6 +10112,7 @@ class Program_weight_tensor_parameter_851: class Program_weight_tensor_parameter_852: name = "parameter_852" + original_name = "conv2d_2.w_0" shape = [80, 40, 3, 3] dtype = "float32" min_val = float("-0.150854") @@ -9271,6 +10124,7 @@ class Program_weight_tensor_parameter_852: class Program_weight_tensor_parameter_853: name = "parameter_853" + original_name = "batch_norm2d_1.b_0" shape = [40] dtype = "float32" min_val = float("0") @@ -9280,6 +10134,7 @@ class Program_weight_tensor_parameter_853: class Program_weight_tensor_parameter_854: name = "parameter_854" + original_name = "batch_norm2d_1.w_0" shape = [40] dtype = "float32" min_val = float("0") @@ -9289,6 +10144,7 @@ class Program_weight_tensor_parameter_854: class Program_weight_tensor_parameter_855: name = "parameter_855" + original_name = "batch_norm2d_1.w_2" shape = [40] dtype = "float32" min_val = float("0") @@ -9298,6 +10154,7 @@ class Program_weight_tensor_parameter_855: class Program_weight_tensor_parameter_856: name = "parameter_856" + original_name = "batch_norm2d_1.w_1" shape = [40] dtype = "float32" min_val = float("0") @@ -9307,6 +10164,7 @@ class Program_weight_tensor_parameter_856: class Program_weight_tensor_parameter_857: name = "parameter_857" + original_name = "conv2d_1.w_0" shape = [40, 40, 3, 3] dtype = "float32" min_val = float("-0.259719") @@ -9318,6 +10176,7 @@ class Program_weight_tensor_parameter_857: class Program_weight_tensor_parameter_858: name = "parameter_858" + original_name = "batch_norm2d_0.b_0" shape = [40] dtype = "float32" min_val = float("0") @@ -9327,6 +10186,7 @@ class Program_weight_tensor_parameter_858: class Program_weight_tensor_parameter_859: name = "parameter_859" + original_name = "batch_norm2d_0.w_0" shape = [40] dtype = "float32" min_val = float("0") @@ -9336,6 +10196,7 @@ class Program_weight_tensor_parameter_859: class Program_weight_tensor_parameter_860: name = "parameter_860" + original_name = "batch_norm2d_0.w_2" shape = [40] dtype = "float32" min_val = float("0") @@ -9345,6 +10206,7 @@ class Program_weight_tensor_parameter_860: class Program_weight_tensor_parameter_861: name = "parameter_861" + original_name = "batch_norm2d_0.w_1" shape = [40] dtype = "float32" min_val = float("0") @@ -9354,6 +10216,7 @@ class Program_weight_tensor_parameter_861: class Program_weight_tensor_parameter_862: name = "parameter_862" + original_name = "conv2d_0.w_0" shape = [40, 3, 3, 3] dtype = "float32" min_val = float("-0.232158") diff --git a/paddle_samples/PaddleX/PP-YOLOE_plus_SOD-L/subgraph_2/input_meta.py b/paddle_samples/PaddleX/PP-YOLOE_plus_SOD-L/subgraph_2/input_meta.py index b94668a44..24c2f4179 100644 --- a/paddle_samples/PaddleX/PP-YOLOE_plus_SOD-L/subgraph_2/input_meta.py +++ b/paddle_samples/PaddleX/PP-YOLOE_plus_SOD-L/subgraph_2/input_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_data_0: name = "data_0" + original_name = "param_78" shape = [1] dtype = "float32" data = [0.699884] @@ -7,6 +8,7 @@ class Program_weight_tensor_data_0: class Program_weight_tensor_data_1: name = "data_1" + original_name = "param_79" shape = [1] dtype = "float32" data = [0.667963] @@ -14,6 +16,7 @@ class Program_weight_tensor_data_1: class Program_weight_tensor_data_2: name = "data_2" + original_name = "param_80" shape = [1] dtype = "float32" data = [0.675792] @@ -21,6 +24,7 @@ class Program_weight_tensor_data_2: class Program_weight_tensor_data_3: name = "data_3" + original_name = "param_81" shape = [1] dtype = "float32" data = [0.676071] @@ -28,6 +32,7 @@ class Program_weight_tensor_data_3: class Program_weight_tensor_data_4: name = "data_4" + original_name = "param_82" shape = [1] dtype = "float32" data = [0.658719] @@ -35,6 +40,7 @@ class Program_weight_tensor_data_4: class Program_weight_tensor_data_5: name = "data_5" + original_name = "param_83" shape = [1] dtype = "float32" data = [0.620637] @@ -42,6 +48,7 @@ class Program_weight_tensor_data_5: class Program_weight_tensor_data_6: name = "data_6" + original_name = "param_84" shape = [1] dtype = "float32" data = [0.637685] @@ -49,6 +56,7 @@ class Program_weight_tensor_data_6: class Program_weight_tensor_data_7: name = "data_7" + original_name = "param_85" shape = [1] dtype = "float32" data = [0.619238] @@ -56,6 +64,7 @@ class Program_weight_tensor_data_7: class Program_weight_tensor_data_8: name = "data_8" + original_name = "param_86" shape = [1] dtype = "float32" data = [0.773168] @@ -63,6 +72,7 @@ class Program_weight_tensor_data_8: class Program_weight_tensor_data_9: name = "data_9" + original_name = "param_87" shape = [1] dtype = "float32" data = [0.635316] @@ -70,6 +80,7 @@ class Program_weight_tensor_data_9: class Program_weight_tensor_data_10: name = "data_10" + original_name = "param_88" shape = [1] dtype = "float32" data = [0.623672] @@ -77,6 +88,7 @@ class Program_weight_tensor_data_10: class Program_weight_tensor_data_11: name = "data_11" + original_name = "param_89" shape = [1] dtype = "float32" data = [0.620323] @@ -84,6 +96,7 @@ class Program_weight_tensor_data_11: class Program_weight_tensor_data_12: name = "data_12" + original_name = "param_90" shape = [1] dtype = "float32" data = [0.621219] @@ -91,6 +104,7 @@ class Program_weight_tensor_data_12: class Program_weight_tensor_data_13: name = "data_13" + original_name = "param_91" shape = [1] dtype = "float32" data = [0.624329] @@ -98,6 +112,7 @@ class Program_weight_tensor_data_13: class Program_weight_tensor_data_14: name = "data_14" + original_name = "param_92" shape = [1] dtype = "float32" data = [0.733117] @@ -105,6 +120,7 @@ class Program_weight_tensor_data_14: class Program_weight_tensor_data_15: name = "data_15" + original_name = "param_93" shape = [1] dtype = "float32" data = [0.557224] @@ -112,6 +128,7 @@ class Program_weight_tensor_data_15: class Program_weight_tensor_data_16: name = "data_16" + original_name = "param_94" shape = [1] dtype = "float32" data = [0.579909] @@ -119,6 +136,7 @@ class Program_weight_tensor_data_16: class Program_weight_tensor_data_17: name = "data_17" + original_name = "param_95" shape = [1] dtype = "float32" data = [0.70327] @@ -126,6 +144,7 @@ class Program_weight_tensor_data_17: class Program_weight_tensor_data_18: name = "data_18" + original_name = "var_3588" shape = [2, 3, 768, 768] dtype = "float32" max_val = float("0.933333") diff --git a/paddle_samples/PaddleX/PP-YOLOE_plus_SOD-L/subgraph_2/weight_meta.py b/paddle_samples/PaddleX/PP-YOLOE_plus_SOD-L/subgraph_2/weight_meta.py index f5b3eb50d..a0e625cb5 100644 --- a/paddle_samples/PaddleX/PP-YOLOE_plus_SOD-L/subgraph_2/weight_meta.py +++ b/paddle_samples/PaddleX/PP-YOLOE_plus_SOD-L/subgraph_2/weight_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_parameter_0: name = "parameter_0" + original_name = "batch_norm2d_72.b_0" shape = [1024] dtype = "float32" min_val = float("-3.7594") @@ -11,6 +12,7 @@ class Program_weight_tensor_parameter_0: class Program_weight_tensor_parameter_1: name = "parameter_1" + original_name = "batch_norm2d_72.w_0" shape = [1024] dtype = "float32" min_val = float("1.61913") @@ -22,6 +24,7 @@ class Program_weight_tensor_parameter_1: class Program_weight_tensor_parameter_2: name = "parameter_2" + original_name = "batch_norm2d_72.w_2" shape = [1024] dtype = "float32" min_val = float("0.00437889") @@ -33,6 +36,7 @@ class Program_weight_tensor_parameter_2: class Program_weight_tensor_parameter_3: name = "parameter_3" + original_name = "batch_norm2d_72.w_1" shape = [1024] dtype = "float32" min_val = float("-0.140825") @@ -44,6 +48,7 @@ class Program_weight_tensor_parameter_3: class Program_weight_tensor_parameter_4: name = "parameter_4" + original_name = "conv2d_76.w_0" shape = [1024, 768, 1, 1] dtype = "float32" min_val = float("-0.0427729") @@ -55,6 +60,7 @@ class Program_weight_tensor_parameter_4: class Program_weight_tensor_parameter_5: name = "parameter_5" + original_name = "conv2d_75.b_0" shape = [768] dtype = "float32" min_val = float("-0.014467") @@ -66,6 +72,7 @@ class Program_weight_tensor_parameter_5: class Program_weight_tensor_parameter_6: name = "parameter_6" + original_name = "conv2d_75.w_0" shape = [768, 768, 1, 1] dtype = "float32" min_val = float("-0.0787519") @@ -77,6 +84,7 @@ class Program_weight_tensor_parameter_6: class Program_weight_tensor_parameter_7: name = "parameter_7" + original_name = "batch_norm2d_71.b_0" shape = [384] dtype = "float32" min_val = float("-1.77402") @@ -88,6 +96,7 @@ class Program_weight_tensor_parameter_7: class Program_weight_tensor_parameter_8: name = "parameter_8" + original_name = "batch_norm2d_71.w_0" shape = [384] dtype = "float32" min_val = float("0.188523") @@ -99,6 +108,7 @@ class Program_weight_tensor_parameter_8: class Program_weight_tensor_parameter_9: name = "parameter_9" + original_name = "batch_norm2d_71.w_2" shape = [384] dtype = "float32" min_val = float("5.23505e-05") @@ -110,6 +120,7 @@ class Program_weight_tensor_parameter_9: class Program_weight_tensor_parameter_10: name = "parameter_10" + original_name = "batch_norm2d_71.w_1" shape = [384] dtype = "float32" min_val = float("-0.0913202") @@ -121,6 +132,7 @@ class Program_weight_tensor_parameter_10: class Program_weight_tensor_parameter_11: name = "parameter_11" + original_name = "conv2d_74.w_0" shape = [384, 384, 1, 1] dtype = "float32" min_val = float("-0.020214") @@ -132,6 +144,7 @@ class Program_weight_tensor_parameter_11: class Program_weight_tensor_parameter_12: name = "parameter_12" + original_name = "batch_norm2d_70.b_0" shape = [384] dtype = "float32" min_val = float("-1.77402") @@ -143,6 +156,7 @@ class Program_weight_tensor_parameter_12: class Program_weight_tensor_parameter_13: name = "parameter_13" + original_name = "batch_norm2d_70.w_0" shape = [384] dtype = "float32" min_val = float("0.334653") @@ -154,6 +168,7 @@ class Program_weight_tensor_parameter_13: class Program_weight_tensor_parameter_14: name = "parameter_14" + original_name = "batch_norm2d_70.w_2" shape = [384] dtype = "float32" min_val = float("0.000593004") @@ -165,6 +180,7 @@ class Program_weight_tensor_parameter_14: class Program_weight_tensor_parameter_15: name = "parameter_15" + original_name = "batch_norm2d_70.w_1" shape = [384] dtype = "float32" min_val = float("-0.227759") @@ -176,6 +192,7 @@ class Program_weight_tensor_parameter_15: class Program_weight_tensor_parameter_16: name = "parameter_16" + original_name = "conv2d_73.w_0" shape = [384, 384, 3, 3] dtype = "float32" min_val = float("-0.0190584") @@ -187,6 +204,7 @@ class Program_weight_tensor_parameter_16: class Program_weight_tensor_parameter_17: name = "parameter_17" + original_name = "batch_norm2d_69.b_0" shape = [384] dtype = "float32" min_val = float("-2.58225") @@ -198,6 +216,7 @@ class Program_weight_tensor_parameter_17: class Program_weight_tensor_parameter_18: name = "parameter_18" + original_name = "batch_norm2d_69.w_0" shape = [384] dtype = "float32" min_val = float("0.52002") @@ -209,6 +228,7 @@ class Program_weight_tensor_parameter_18: class Program_weight_tensor_parameter_19: name = "parameter_19" + original_name = "batch_norm2d_69.w_2" shape = [384] dtype = "float32" min_val = float("0.0404637") @@ -220,6 +240,7 @@ class Program_weight_tensor_parameter_19: class Program_weight_tensor_parameter_20: name = "parameter_20" + original_name = "batch_norm2d_69.w_1" shape = [384] dtype = "float32" min_val = float("-0.9033") @@ -231,6 +252,7 @@ class Program_weight_tensor_parameter_20: class Program_weight_tensor_parameter_21: name = "parameter_21" + original_name = "conv2d_72.w_0" shape = [384, 384, 3, 3] dtype = "float32" min_val = float("-0.0213207") @@ -242,6 +264,7 @@ class Program_weight_tensor_parameter_21: class Program_weight_tensor_parameter_22: name = "parameter_22" + original_name = "batch_norm2d_68.b_0" shape = [384] dtype = "float32" min_val = float("-1.93927") @@ -253,6 +276,7 @@ class Program_weight_tensor_parameter_22: class Program_weight_tensor_parameter_23: name = "parameter_23" + original_name = "batch_norm2d_68.w_0" shape = [384] dtype = "float32" min_val = float("0.163873") @@ -264,6 +288,7 @@ class Program_weight_tensor_parameter_23: class Program_weight_tensor_parameter_24: name = "parameter_24" + original_name = "batch_norm2d_68.w_2" shape = [384] dtype = "float32" min_val = float("7.65946e-05") @@ -275,6 +300,7 @@ class Program_weight_tensor_parameter_24: class Program_weight_tensor_parameter_25: name = "parameter_25" + original_name = "batch_norm2d_68.w_1" shape = [384] dtype = "float32" min_val = float("-0.047045") @@ -286,6 +312,7 @@ class Program_weight_tensor_parameter_25: class Program_weight_tensor_parameter_26: name = "parameter_26" + original_name = "conv2d_71.w_0" shape = [384, 384, 1, 1] dtype = "float32" min_val = float("-0.0246209") @@ -297,6 +324,7 @@ class Program_weight_tensor_parameter_26: class Program_weight_tensor_parameter_27: name = "parameter_27" + original_name = "batch_norm2d_67.b_0" shape = [384] dtype = "float32" min_val = float("-1.93932") @@ -308,6 +336,7 @@ class Program_weight_tensor_parameter_27: class Program_weight_tensor_parameter_28: name = "parameter_28" + original_name = "batch_norm2d_67.w_0" shape = [384] dtype = "float32" min_val = float("0.58315") @@ -319,6 +348,7 @@ class Program_weight_tensor_parameter_28: class Program_weight_tensor_parameter_29: name = "parameter_29" + original_name = "batch_norm2d_67.w_2" shape = [384] dtype = "float32" min_val = float("0.0013599") @@ -330,6 +360,7 @@ class Program_weight_tensor_parameter_29: class Program_weight_tensor_parameter_30: name = "parameter_30" + original_name = "batch_norm2d_67.w_1" shape = [384] dtype = "float32" min_val = float("-0.0821017") @@ -341,6 +372,7 @@ class Program_weight_tensor_parameter_30: class Program_weight_tensor_parameter_31: name = "parameter_31" + original_name = "conv2d_70.w_0" shape = [384, 384, 3, 3] dtype = "float32" min_val = float("-0.017236") @@ -352,6 +384,7 @@ class Program_weight_tensor_parameter_31: class Program_weight_tensor_parameter_32: name = "parameter_32" + original_name = "batch_norm2d_66.b_0" shape = [384] dtype = "float32" min_val = float("-2.39591") @@ -363,6 +396,7 @@ class Program_weight_tensor_parameter_32: class Program_weight_tensor_parameter_33: name = "parameter_33" + original_name = "batch_norm2d_66.w_0" shape = [384] dtype = "float32" min_val = float("0.453112") @@ -374,6 +408,7 @@ class Program_weight_tensor_parameter_33: class Program_weight_tensor_parameter_34: name = "parameter_34" + original_name = "batch_norm2d_66.w_2" shape = [384] dtype = "float32" min_val = float("0.0300933") @@ -385,6 +420,7 @@ class Program_weight_tensor_parameter_34: class Program_weight_tensor_parameter_35: name = "parameter_35" + original_name = "batch_norm2d_66.w_1" shape = [384] dtype = "float32" min_val = float("-0.749117") @@ -396,6 +432,7 @@ class Program_weight_tensor_parameter_35: class Program_weight_tensor_parameter_36: name = "parameter_36" + original_name = "conv2d_69.w_0" shape = [384, 384, 3, 3] dtype = "float32" min_val = float("-0.0259567") @@ -407,6 +444,7 @@ class Program_weight_tensor_parameter_36: class Program_weight_tensor_parameter_37: name = "parameter_37" + original_name = "batch_norm2d_65.b_0" shape = [384] dtype = "float32" min_val = float("-1.8762") @@ -418,6 +456,7 @@ class Program_weight_tensor_parameter_37: class Program_weight_tensor_parameter_38: name = "parameter_38" + original_name = "batch_norm2d_65.w_0" shape = [384] dtype = "float32" min_val = float("0.0773354") @@ -429,6 +468,7 @@ class Program_weight_tensor_parameter_38: class Program_weight_tensor_parameter_39: name = "parameter_39" + original_name = "batch_norm2d_65.w_2" shape = [384] dtype = "float32" min_val = float("6.01092e-05") @@ -440,6 +480,7 @@ class Program_weight_tensor_parameter_39: class Program_weight_tensor_parameter_40: name = "parameter_40" + original_name = "batch_norm2d_65.w_1" shape = [384] dtype = "float32" min_val = float("-0.0475575") @@ -451,6 +492,7 @@ class Program_weight_tensor_parameter_40: class Program_weight_tensor_parameter_41: name = "parameter_41" + original_name = "conv2d_68.w_0" shape = [384, 384, 1, 1] dtype = "float32" min_val = float("-0.0207296") @@ -462,6 +504,7 @@ class Program_weight_tensor_parameter_41: class Program_weight_tensor_parameter_42: name = "parameter_42" + original_name = "batch_norm2d_64.b_0" shape = [384] dtype = "float32" min_val = float("-1.87654") @@ -473,6 +516,7 @@ class Program_weight_tensor_parameter_42: class Program_weight_tensor_parameter_43: name = "parameter_43" + original_name = "batch_norm2d_64.w_0" shape = [384] dtype = "float32" min_val = float("0.521871") @@ -484,6 +528,7 @@ class Program_weight_tensor_parameter_43: class Program_weight_tensor_parameter_44: name = "parameter_44" + original_name = "batch_norm2d_64.w_2" shape = [384] dtype = "float32" min_val = float("0.00177683") @@ -495,6 +540,7 @@ class Program_weight_tensor_parameter_44: class Program_weight_tensor_parameter_45: name = "parameter_45" + original_name = "batch_norm2d_64.w_1" shape = [384] dtype = "float32" min_val = float("-0.209845") @@ -506,6 +552,7 @@ class Program_weight_tensor_parameter_45: class Program_weight_tensor_parameter_46: name = "parameter_46" + original_name = "conv2d_67.w_0" shape = [384, 384, 3, 3] dtype = "float32" min_val = float("-0.0177497") @@ -517,6 +564,7 @@ class Program_weight_tensor_parameter_46: class Program_weight_tensor_parameter_47: name = "parameter_47" + original_name = "batch_norm2d_63.b_0" shape = [384] dtype = "float32" min_val = float("-2.15635") @@ -528,6 +576,7 @@ class Program_weight_tensor_parameter_47: class Program_weight_tensor_parameter_48: name = "parameter_48" + original_name = "batch_norm2d_63.w_0" shape = [384] dtype = "float32" min_val = float("0.706134") @@ -539,6 +588,7 @@ class Program_weight_tensor_parameter_48: class Program_weight_tensor_parameter_49: name = "parameter_49" + original_name = "batch_norm2d_63.w_2" shape = [384] dtype = "float32" min_val = float("0.0216198") @@ -550,6 +600,7 @@ class Program_weight_tensor_parameter_49: class Program_weight_tensor_parameter_50: name = "parameter_50" + original_name = "batch_norm2d_63.w_1" shape = [384] dtype = "float32" min_val = float("-0.694464") @@ -561,6 +612,7 @@ class Program_weight_tensor_parameter_50: class Program_weight_tensor_parameter_51: name = "parameter_51" + original_name = "conv2d_66.w_0" shape = [384, 384, 3, 3] dtype = "float32" min_val = float("-0.0274071") @@ -572,6 +624,7 @@ class Program_weight_tensor_parameter_51: class Program_weight_tensor_parameter_52: name = "parameter_52" + original_name = "batch_norm2d_62.b_0" shape = [384] dtype = "float32" min_val = float("-2.9232") @@ -583,6 +636,7 @@ class Program_weight_tensor_parameter_52: class Program_weight_tensor_parameter_53: name = "parameter_53" + original_name = "batch_norm2d_62.w_0" shape = [384] dtype = "float32" min_val = float("0.953224") @@ -594,6 +648,7 @@ class Program_weight_tensor_parameter_53: class Program_weight_tensor_parameter_54: name = "parameter_54" + original_name = "batch_norm2d_62.w_2" shape = [384] dtype = "float32" min_val = float("0.00275058") @@ -605,6 +660,7 @@ class Program_weight_tensor_parameter_54: class Program_weight_tensor_parameter_55: name = "parameter_55" + original_name = "batch_norm2d_62.w_1" shape = [384] dtype = "float32" min_val = float("-0.249794") @@ -616,6 +672,7 @@ class Program_weight_tensor_parameter_55: class Program_weight_tensor_parameter_56: name = "parameter_56" + original_name = "conv2d_65.w_0" shape = [384, 768, 1, 1] dtype = "float32" min_val = float("-0.0371909") @@ -627,6 +684,7 @@ class Program_weight_tensor_parameter_56: class Program_weight_tensor_parameter_57: name = "parameter_57" + original_name = "batch_norm2d_61.b_0" shape = [384] dtype = "float32" min_val = float("-2.2471") @@ -638,6 +696,7 @@ class Program_weight_tensor_parameter_57: class Program_weight_tensor_parameter_58: name = "parameter_58" + original_name = "batch_norm2d_61.w_0" shape = [384] dtype = "float32" min_val = float("0.965853") @@ -649,6 +708,7 @@ class Program_weight_tensor_parameter_58: class Program_weight_tensor_parameter_59: name = "parameter_59" + original_name = "batch_norm2d_61.w_2" shape = [384] dtype = "float32" min_val = float("0.000799495") @@ -660,6 +720,7 @@ class Program_weight_tensor_parameter_59: class Program_weight_tensor_parameter_60: name = "parameter_60" + original_name = "batch_norm2d_61.w_1" shape = [384] dtype = "float32" min_val = float("-0.0161372") @@ -671,6 +732,7 @@ class Program_weight_tensor_parameter_60: class Program_weight_tensor_parameter_61: name = "parameter_61" + original_name = "conv2d_64.w_0" shape = [384, 768, 1, 1] dtype = "float32" min_val = float("-0.0815437") @@ -682,6 +744,7 @@ class Program_weight_tensor_parameter_61: class Program_weight_tensor_parameter_62: name = "parameter_62" + original_name = "batch_norm2d_60.b_0" shape = [768] dtype = "float32" min_val = float("-2.40199") @@ -693,6 +756,7 @@ class Program_weight_tensor_parameter_62: class Program_weight_tensor_parameter_63: name = "parameter_63" + original_name = "batch_norm2d_60.w_0" shape = [768] dtype = "float32" min_val = float("0.530297") @@ -704,6 +768,7 @@ class Program_weight_tensor_parameter_63: class Program_weight_tensor_parameter_64: name = "parameter_64" + original_name = "batch_norm2d_60.w_2" shape = [768] dtype = "float32" min_val = float("0.00625688") @@ -715,6 +780,7 @@ class Program_weight_tensor_parameter_64: class Program_weight_tensor_parameter_65: name = "parameter_65" + original_name = "batch_norm2d_60.w_1" shape = [768] dtype = "float32" min_val = float("-0.235652") @@ -726,6 +792,7 @@ class Program_weight_tensor_parameter_65: class Program_weight_tensor_parameter_66: name = "parameter_66" + original_name = "conv2d_63.w_0" shape = [768, 512, 3, 3] dtype = "float32" min_val = float("-0.0378314") @@ -737,6 +804,7 @@ class Program_weight_tensor_parameter_66: class Program_weight_tensor_parameter_67: name = "parameter_67" + original_name = "batch_norm2d_59.b_0" shape = [512] dtype = "float32" min_val = float("-3.38998") @@ -748,6 +816,7 @@ class Program_weight_tensor_parameter_67: class Program_weight_tensor_parameter_68: name = "parameter_68" + original_name = "batch_norm2d_59.w_0" shape = [512] dtype = "float32" min_val = float("0.523767") @@ -759,6 +828,7 @@ class Program_weight_tensor_parameter_68: class Program_weight_tensor_parameter_69: name = "parameter_69" + original_name = "batch_norm2d_59.w_2" shape = [512] dtype = "float32" min_val = float("0.00233511") @@ -770,6 +840,7 @@ class Program_weight_tensor_parameter_69: class Program_weight_tensor_parameter_70: name = "parameter_70" + original_name = "batch_norm2d_59.w_1" shape = [512] dtype = "float32" min_val = float("-0.172067") @@ -781,6 +852,7 @@ class Program_weight_tensor_parameter_70: class Program_weight_tensor_parameter_71: name = "parameter_71" + original_name = "conv2d_62.w_0" shape = [512, 384, 1, 1] dtype = "float32" min_val = float("-0.202262") @@ -792,6 +864,7 @@ class Program_weight_tensor_parameter_71: class Program_weight_tensor_parameter_72: name = "parameter_72" + original_name = "conv2d_61.b_0" shape = [384] dtype = "float32" min_val = float("-0.0100703") @@ -803,6 +876,7 @@ class Program_weight_tensor_parameter_72: class Program_weight_tensor_parameter_73: name = "parameter_73" + original_name = "conv2d_61.w_0" shape = [384, 384, 1, 1] dtype = "float32" min_val = float("-0.202729") @@ -814,6 +888,7 @@ class Program_weight_tensor_parameter_73: class Program_weight_tensor_parameter_74: name = "parameter_74" + original_name = "batch_norm2d_58.b_0" shape = [192] dtype = "float32" min_val = float("-1.97045") @@ -825,6 +900,7 @@ class Program_weight_tensor_parameter_74: class Program_weight_tensor_parameter_75: name = "parameter_75" + original_name = "batch_norm2d_58.w_0" shape = [192] dtype = "float32" min_val = float("0.0528864") @@ -836,6 +912,7 @@ class Program_weight_tensor_parameter_75: class Program_weight_tensor_parameter_76: name = "parameter_76" + original_name = "batch_norm2d_58.w_2" shape = [192] dtype = "float32" min_val = float("8.99309e-05") @@ -847,6 +924,7 @@ class Program_weight_tensor_parameter_76: class Program_weight_tensor_parameter_77: name = "parameter_77" + original_name = "batch_norm2d_58.w_1" shape = [192] dtype = "float32" min_val = float("-0.0345233") @@ -858,6 +936,7 @@ class Program_weight_tensor_parameter_77: class Program_weight_tensor_parameter_78: name = "parameter_78" + original_name = "conv2d_60.w_0" shape = [192, 192, 1, 1] dtype = "float32" min_val = float("-0.023487") @@ -869,6 +948,7 @@ class Program_weight_tensor_parameter_78: class Program_weight_tensor_parameter_79: name = "parameter_79" + original_name = "batch_norm2d_57.b_0" shape = [192] dtype = "float32" min_val = float("-1.97037") @@ -880,6 +960,7 @@ class Program_weight_tensor_parameter_79: class Program_weight_tensor_parameter_80: name = "parameter_80" + original_name = "batch_norm2d_57.w_0" shape = [192] dtype = "float32" min_val = float("0.372338") @@ -891,6 +972,7 @@ class Program_weight_tensor_parameter_80: class Program_weight_tensor_parameter_81: name = "parameter_81" + original_name = "batch_norm2d_57.w_2" shape = [192] dtype = "float32" min_val = float("0.00122552") @@ -902,6 +984,7 @@ class Program_weight_tensor_parameter_81: class Program_weight_tensor_parameter_82: name = "parameter_82" + original_name = "batch_norm2d_57.w_1" shape = [192] dtype = "float32" min_val = float("-0.097226") @@ -913,6 +996,7 @@ class Program_weight_tensor_parameter_82: class Program_weight_tensor_parameter_83: name = "parameter_83" + original_name = "conv2d_59.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.0289902") @@ -924,6 +1008,7 @@ class Program_weight_tensor_parameter_83: class Program_weight_tensor_parameter_84: name = "parameter_84" + original_name = "batch_norm2d_56.b_0" shape = [192] dtype = "float32" min_val = float("-2.89065") @@ -935,6 +1020,7 @@ class Program_weight_tensor_parameter_84: class Program_weight_tensor_parameter_85: name = "parameter_85" + original_name = "batch_norm2d_56.w_0" shape = [192] dtype = "float32" min_val = float("0.696524") @@ -946,6 +1032,7 @@ class Program_weight_tensor_parameter_85: class Program_weight_tensor_parameter_86: name = "parameter_86" + original_name = "batch_norm2d_56.w_2" shape = [192] dtype = "float32" min_val = float("0.0626606") @@ -957,6 +1044,7 @@ class Program_weight_tensor_parameter_86: class Program_weight_tensor_parameter_87: name = "parameter_87" + original_name = "batch_norm2d_56.w_1" shape = [192] dtype = "float32" min_val = float("-2.50771") @@ -968,6 +1056,7 @@ class Program_weight_tensor_parameter_87: class Program_weight_tensor_parameter_88: name = "parameter_88" + original_name = "conv2d_58.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.0331927") @@ -979,6 +1068,7 @@ class Program_weight_tensor_parameter_88: class Program_weight_tensor_parameter_89: name = "parameter_89" + original_name = "batch_norm2d_55.b_0" shape = [192] dtype = "float32" min_val = float("-1.9404") @@ -990,6 +1080,7 @@ class Program_weight_tensor_parameter_89: class Program_weight_tensor_parameter_90: name = "parameter_90" + original_name = "batch_norm2d_55.w_0" shape = [192] dtype = "float32" min_val = float("0.0454025") @@ -1001,6 +1092,7 @@ class Program_weight_tensor_parameter_90: class Program_weight_tensor_parameter_91: name = "parameter_91" + original_name = "batch_norm2d_55.w_2" shape = [192] dtype = "float32" min_val = float("7.44986e-05") @@ -1012,6 +1104,7 @@ class Program_weight_tensor_parameter_91: class Program_weight_tensor_parameter_92: name = "parameter_92" + original_name = "batch_norm2d_55.w_1" shape = [192] dtype = "float32" min_val = float("-0.029272") @@ -1023,6 +1116,7 @@ class Program_weight_tensor_parameter_92: class Program_weight_tensor_parameter_93: name = "parameter_93" + original_name = "conv2d_57.w_0" shape = [192, 192, 1, 1] dtype = "float32" min_val = float("-0.0234926") @@ -1034,6 +1128,7 @@ class Program_weight_tensor_parameter_93: class Program_weight_tensor_parameter_94: name = "parameter_94" + original_name = "batch_norm2d_54.b_0" shape = [192] dtype = "float32" min_val = float("-1.94044") @@ -1045,6 +1140,7 @@ class Program_weight_tensor_parameter_94: class Program_weight_tensor_parameter_95: name = "parameter_95" + original_name = "batch_norm2d_54.w_0" shape = [192] dtype = "float32" min_val = float("0.483074") @@ -1056,6 +1152,7 @@ class Program_weight_tensor_parameter_95: class Program_weight_tensor_parameter_96: name = "parameter_96" + original_name = "batch_norm2d_54.w_2" shape = [192] dtype = "float32" min_val = float("0.00270072") @@ -1067,6 +1164,7 @@ class Program_weight_tensor_parameter_96: class Program_weight_tensor_parameter_97: name = "parameter_97" + original_name = "batch_norm2d_54.w_1" shape = [192] dtype = "float32" min_val = float("-0.0923253") @@ -1078,6 +1176,7 @@ class Program_weight_tensor_parameter_97: class Program_weight_tensor_parameter_98: name = "parameter_98" + original_name = "conv2d_56.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.0231072") @@ -1089,6 +1188,7 @@ class Program_weight_tensor_parameter_98: class Program_weight_tensor_parameter_99: name = "parameter_99" + original_name = "batch_norm2d_53.b_0" shape = [192] dtype = "float32" min_val = float("-2.50828") @@ -1100,6 +1200,7 @@ class Program_weight_tensor_parameter_99: class Program_weight_tensor_parameter_100: name = "parameter_100" + original_name = "batch_norm2d_53.w_0" shape = [192] dtype = "float32" min_val = float("0.65494") @@ -1111,6 +1212,7 @@ class Program_weight_tensor_parameter_100: class Program_weight_tensor_parameter_101: name = "parameter_101" + original_name = "batch_norm2d_53.w_2" shape = [192] dtype = "float32" min_val = float("0.0463808") @@ -1122,6 +1224,7 @@ class Program_weight_tensor_parameter_101: class Program_weight_tensor_parameter_102: name = "parameter_102" + original_name = "batch_norm2d_53.w_1" shape = [192] dtype = "float32" min_val = float("-2.14238") @@ -1133,6 +1236,7 @@ class Program_weight_tensor_parameter_102: class Program_weight_tensor_parameter_103: name = "parameter_103" + original_name = "conv2d_55.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.0362254") @@ -1144,6 +1248,7 @@ class Program_weight_tensor_parameter_103: class Program_weight_tensor_parameter_104: name = "parameter_104" + original_name = "batch_norm2d_52.b_0" shape = [192] dtype = "float32" min_val = float("-1.7573") @@ -1155,6 +1260,7 @@ class Program_weight_tensor_parameter_104: class Program_weight_tensor_parameter_105: name = "parameter_105" + original_name = "batch_norm2d_52.w_0" shape = [192] dtype = "float32" min_val = float("0.00295124") @@ -1166,6 +1272,7 @@ class Program_weight_tensor_parameter_105: class Program_weight_tensor_parameter_106: name = "parameter_106" + original_name = "batch_norm2d_52.w_2" shape = [192] dtype = "float32" min_val = float("9.30792e-07") @@ -1177,6 +1284,7 @@ class Program_weight_tensor_parameter_106: class Program_weight_tensor_parameter_107: name = "parameter_107" + original_name = "batch_norm2d_52.w_1" shape = [192] dtype = "float32" min_val = float("-0.0372993") @@ -1188,6 +1296,7 @@ class Program_weight_tensor_parameter_107: class Program_weight_tensor_parameter_108: name = "parameter_108" + original_name = "conv2d_54.w_0" shape = [192, 192, 1, 1] dtype = "float32" min_val = float("-0.0303466") @@ -1199,6 +1308,7 @@ class Program_weight_tensor_parameter_108: class Program_weight_tensor_parameter_109: name = "parameter_109" + original_name = "batch_norm2d_51.b_0" shape = [192] dtype = "float32" min_val = float("-1.7573") @@ -1210,6 +1320,7 @@ class Program_weight_tensor_parameter_109: class Program_weight_tensor_parameter_110: name = "parameter_110" + original_name = "batch_norm2d_51.w_0" shape = [192] dtype = "float32" min_val = float("0.406102") @@ -1221,6 +1332,7 @@ class Program_weight_tensor_parameter_110: class Program_weight_tensor_parameter_111: name = "parameter_111" + original_name = "batch_norm2d_51.w_2" shape = [192] dtype = "float32" min_val = float("0.00267891") @@ -1232,6 +1344,7 @@ class Program_weight_tensor_parameter_111: class Program_weight_tensor_parameter_112: name = "parameter_112" + original_name = "batch_norm2d_51.w_1" shape = [192] dtype = "float32" min_val = float("-0.0635846") @@ -1243,6 +1356,7 @@ class Program_weight_tensor_parameter_112: class Program_weight_tensor_parameter_113: name = "parameter_113" + original_name = "conv2d_53.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.0321474") @@ -1254,6 +1368,7 @@ class Program_weight_tensor_parameter_113: class Program_weight_tensor_parameter_114: name = "parameter_114" + original_name = "batch_norm2d_50.b_0" shape = [192] dtype = "float32" min_val = float("-2.49735") @@ -1265,6 +1380,7 @@ class Program_weight_tensor_parameter_114: class Program_weight_tensor_parameter_115: name = "parameter_115" + original_name = "batch_norm2d_50.w_0" shape = [192] dtype = "float32" min_val = float("0.652126") @@ -1276,6 +1392,7 @@ class Program_weight_tensor_parameter_115: class Program_weight_tensor_parameter_116: name = "parameter_116" + original_name = "batch_norm2d_50.w_2" shape = [192] dtype = "float32" min_val = float("0.0307206") @@ -1287,6 +1404,7 @@ class Program_weight_tensor_parameter_116: class Program_weight_tensor_parameter_117: name = "parameter_117" + original_name = "batch_norm2d_50.w_1" shape = [192] dtype = "float32" min_val = float("-1.51549") @@ -1298,6 +1416,7 @@ class Program_weight_tensor_parameter_117: class Program_weight_tensor_parameter_118: name = "parameter_118" + original_name = "conv2d_52.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.05013") @@ -1309,6 +1428,7 @@ class Program_weight_tensor_parameter_118: class Program_weight_tensor_parameter_119: name = "parameter_119" + original_name = "batch_norm2d_49.b_0" shape = [192] dtype = "float32" min_val = float("-2.07916") @@ -1320,6 +1440,7 @@ class Program_weight_tensor_parameter_119: class Program_weight_tensor_parameter_120: name = "parameter_120" + original_name = "batch_norm2d_49.w_0" shape = [192] dtype = "float32" min_val = float("0.000510371") @@ -1331,6 +1452,7 @@ class Program_weight_tensor_parameter_120: class Program_weight_tensor_parameter_121: name = "parameter_121" + original_name = "batch_norm2d_49.w_2" shape = [192] dtype = "float32" min_val = float("6.27846e-08") @@ -1342,6 +1464,7 @@ class Program_weight_tensor_parameter_121: class Program_weight_tensor_parameter_122: name = "parameter_122" + original_name = "batch_norm2d_49.w_1" shape = [192] dtype = "float32" min_val = float("-0.0197245") @@ -1353,6 +1476,7 @@ class Program_weight_tensor_parameter_122: class Program_weight_tensor_parameter_123: name = "parameter_123" + original_name = "conv2d_51.w_0" shape = [192, 192, 1, 1] dtype = "float32" min_val = float("-0.0202783") @@ -1364,6 +1488,7 @@ class Program_weight_tensor_parameter_123: class Program_weight_tensor_parameter_124: name = "parameter_124" + original_name = "batch_norm2d_48.b_0" shape = [192] dtype = "float32" min_val = float("-2.07922") @@ -1375,6 +1500,7 @@ class Program_weight_tensor_parameter_124: class Program_weight_tensor_parameter_125: name = "parameter_125" + original_name = "batch_norm2d_48.w_0" shape = [192] dtype = "float32" min_val = float("0.396505") @@ -1386,6 +1512,7 @@ class Program_weight_tensor_parameter_125: class Program_weight_tensor_parameter_126: name = "parameter_126" + original_name = "batch_norm2d_48.w_2" shape = [192] dtype = "float32" min_val = float("0.00316561") @@ -1397,6 +1524,7 @@ class Program_weight_tensor_parameter_126: class Program_weight_tensor_parameter_127: name = "parameter_127" + original_name = "batch_norm2d_48.w_1" shape = [192] dtype = "float32" min_val = float("-0.0910185") @@ -1408,6 +1536,7 @@ class Program_weight_tensor_parameter_127: class Program_weight_tensor_parameter_128: name = "parameter_128" + original_name = "conv2d_50.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.0299549") @@ -1419,6 +1548,7 @@ class Program_weight_tensor_parameter_128: class Program_weight_tensor_parameter_129: name = "parameter_129" + original_name = "batch_norm2d_47.b_0" shape = [192] dtype = "float32" min_val = float("-2.74084") @@ -1430,6 +1560,7 @@ class Program_weight_tensor_parameter_129: class Program_weight_tensor_parameter_130: name = "parameter_130" + original_name = "batch_norm2d_47.w_0" shape = [192] dtype = "float32" min_val = float("0.761623") @@ -1441,6 +1572,7 @@ class Program_weight_tensor_parameter_130: class Program_weight_tensor_parameter_131: name = "parameter_131" + original_name = "batch_norm2d_47.w_2" shape = [192] dtype = "float32" min_val = float("0.0268285") @@ -1452,6 +1584,7 @@ class Program_weight_tensor_parameter_131: class Program_weight_tensor_parameter_132: name = "parameter_132" + original_name = "batch_norm2d_47.w_1" shape = [192] dtype = "float32" min_val = float("-1.23693") @@ -1463,6 +1596,7 @@ class Program_weight_tensor_parameter_132: class Program_weight_tensor_parameter_133: name = "parameter_133" + original_name = "conv2d_49.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.0531238") @@ -1474,6 +1608,7 @@ class Program_weight_tensor_parameter_133: class Program_weight_tensor_parameter_134: name = "parameter_134" + original_name = "batch_norm2d_46.b_0" shape = [192] dtype = "float32" min_val = float("-1.21219") @@ -1485,6 +1620,7 @@ class Program_weight_tensor_parameter_134: class Program_weight_tensor_parameter_135: name = "parameter_135" + original_name = "batch_norm2d_46.w_0" shape = [192] dtype = "float32" min_val = float("-9.82711e-05") @@ -1496,6 +1632,7 @@ class Program_weight_tensor_parameter_135: class Program_weight_tensor_parameter_136: name = "parameter_136" + original_name = "batch_norm2d_46.w_2" shape = [192] dtype = "float32" min_val = float("2.25073e-10") @@ -1507,6 +1644,7 @@ class Program_weight_tensor_parameter_136: class Program_weight_tensor_parameter_137: name = "parameter_137" + original_name = "batch_norm2d_46.w_1" shape = [192] dtype = "float32" min_val = float("-0.0493103") @@ -1518,6 +1656,7 @@ class Program_weight_tensor_parameter_137: class Program_weight_tensor_parameter_138: name = "parameter_138" + original_name = "conv2d_48.w_0" shape = [192, 192, 1, 1] dtype = "float32" min_val = float("-0.0342199") @@ -1529,6 +1668,7 @@ class Program_weight_tensor_parameter_138: class Program_weight_tensor_parameter_139: name = "parameter_139" + original_name = "batch_norm2d_45.b_0" shape = [192] dtype = "float32" min_val = float("-1.21223") @@ -1540,6 +1680,7 @@ class Program_weight_tensor_parameter_139: class Program_weight_tensor_parameter_140: name = "parameter_140" + original_name = "batch_norm2d_45.w_0" shape = [192] dtype = "float32" min_val = float("0.382831") @@ -1551,6 +1692,7 @@ class Program_weight_tensor_parameter_140: class Program_weight_tensor_parameter_141: name = "parameter_141" + original_name = "batch_norm2d_45.w_2" shape = [192] dtype = "float32" min_val = float("0.00221295") @@ -1562,6 +1704,7 @@ class Program_weight_tensor_parameter_141: class Program_weight_tensor_parameter_142: name = "parameter_142" + original_name = "batch_norm2d_45.w_1" shape = [192] dtype = "float32" min_val = float("-0.0844719") @@ -1573,6 +1716,7 @@ class Program_weight_tensor_parameter_142: class Program_weight_tensor_parameter_143: name = "parameter_143" + original_name = "conv2d_47.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.0323048") @@ -1584,6 +1728,7 @@ class Program_weight_tensor_parameter_143: class Program_weight_tensor_parameter_144: name = "parameter_144" + original_name = "batch_norm2d_44.b_0" shape = [192] dtype = "float32" min_val = float("-2.48701") @@ -1595,6 +1740,7 @@ class Program_weight_tensor_parameter_144: class Program_weight_tensor_parameter_145: name = "parameter_145" + original_name = "batch_norm2d_44.w_0" shape = [192] dtype = "float32" min_val = float("0.689678") @@ -1606,6 +1752,7 @@ class Program_weight_tensor_parameter_145: class Program_weight_tensor_parameter_146: name = "parameter_146" + original_name = "batch_norm2d_44.w_2" shape = [192] dtype = "float32" min_val = float("0.0183928") @@ -1617,6 +1764,7 @@ class Program_weight_tensor_parameter_146: class Program_weight_tensor_parameter_147: name = "parameter_147" + original_name = "batch_norm2d_44.w_1" shape = [192] dtype = "float32" min_val = float("-0.716469") @@ -1628,6 +1776,7 @@ class Program_weight_tensor_parameter_147: class Program_weight_tensor_parameter_148: name = "parameter_148" + original_name = "conv2d_46.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.0610342") @@ -1639,6 +1788,7 @@ class Program_weight_tensor_parameter_148: class Program_weight_tensor_parameter_149: name = "parameter_149" + original_name = "batch_norm2d_43.b_0" shape = [192] dtype = "float32" min_val = float("-1.21753") @@ -1650,6 +1800,7 @@ class Program_weight_tensor_parameter_149: class Program_weight_tensor_parameter_150: name = "parameter_150" + original_name = "batch_norm2d_43.w_0" shape = [192] dtype = "float32" min_val = float("0.00836385") @@ -1661,6 +1812,7 @@ class Program_weight_tensor_parameter_150: class Program_weight_tensor_parameter_151: name = "parameter_151" + original_name = "batch_norm2d_43.w_2" shape = [192] dtype = "float32" min_val = float("1.96593e-05") @@ -1672,6 +1824,7 @@ class Program_weight_tensor_parameter_151: class Program_weight_tensor_parameter_152: name = "parameter_152" + original_name = "batch_norm2d_43.w_1" shape = [192] dtype = "float32" min_val = float("-0.0656068") @@ -1683,6 +1836,7 @@ class Program_weight_tensor_parameter_152: class Program_weight_tensor_parameter_153: name = "parameter_153" + original_name = "conv2d_45.w_0" shape = [192, 192, 1, 1] dtype = "float32" min_val = float("-0.0600528") @@ -1694,6 +1848,7 @@ class Program_weight_tensor_parameter_153: class Program_weight_tensor_parameter_154: name = "parameter_154" + original_name = "batch_norm2d_42.b_0" shape = [192] dtype = "float32" min_val = float("-1.21747") @@ -1705,6 +1860,7 @@ class Program_weight_tensor_parameter_154: class Program_weight_tensor_parameter_155: name = "parameter_155" + original_name = "batch_norm2d_42.w_0" shape = [192] dtype = "float32" min_val = float("0.354999") @@ -1716,6 +1872,7 @@ class Program_weight_tensor_parameter_155: class Program_weight_tensor_parameter_156: name = "parameter_156" + original_name = "batch_norm2d_42.w_2" shape = [192] dtype = "float32" min_val = float("0.00436972") @@ -1727,6 +1884,7 @@ class Program_weight_tensor_parameter_156: class Program_weight_tensor_parameter_157: name = "parameter_157" + original_name = "batch_norm2d_42.w_1" shape = [192] dtype = "float32" min_val = float("-0.15942") @@ -1738,6 +1896,7 @@ class Program_weight_tensor_parameter_157: class Program_weight_tensor_parameter_158: name = "parameter_158" + original_name = "conv2d_44.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.062497") @@ -1749,6 +1908,7 @@ class Program_weight_tensor_parameter_158: class Program_weight_tensor_parameter_159: name = "parameter_159" + original_name = "batch_norm2d_41.b_0" shape = [192] dtype = "float32" min_val = float("-1.87905") @@ -1760,6 +1920,7 @@ class Program_weight_tensor_parameter_159: class Program_weight_tensor_parameter_160: name = "parameter_160" + original_name = "batch_norm2d_41.w_0" shape = [192] dtype = "float32" min_val = float("0.788784") @@ -1771,6 +1932,7 @@ class Program_weight_tensor_parameter_160: class Program_weight_tensor_parameter_161: name = "parameter_161" + original_name = "batch_norm2d_41.w_2" shape = [192] dtype = "float32" min_val = float("0.0156512") @@ -1782,6 +1944,7 @@ class Program_weight_tensor_parameter_161: class Program_weight_tensor_parameter_162: name = "parameter_162" + original_name = "batch_norm2d_41.w_1" shape = [192] dtype = "float32" min_val = float("-0.689275") @@ -1793,6 +1956,7 @@ class Program_weight_tensor_parameter_162: class Program_weight_tensor_parameter_163: name = "parameter_163" + original_name = "conv2d_43.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.062874") @@ -1804,6 +1968,7 @@ class Program_weight_tensor_parameter_163: class Program_weight_tensor_parameter_164: name = "parameter_164" + original_name = "batch_norm2d_40.b_0" shape = [192] dtype = "float32" min_val = float("-2.86217") @@ -1815,6 +1980,7 @@ class Program_weight_tensor_parameter_164: class Program_weight_tensor_parameter_165: name = "parameter_165" + original_name = "batch_norm2d_40.w_0" shape = [192] dtype = "float32" min_val = float("0.487672") @@ -1826,6 +1992,7 @@ class Program_weight_tensor_parameter_165: class Program_weight_tensor_parameter_166: name = "parameter_166" + original_name = "batch_norm2d_40.w_2" shape = [192] dtype = "float32" min_val = float("0.00975444") @@ -1837,6 +2004,7 @@ class Program_weight_tensor_parameter_166: class Program_weight_tensor_parameter_167: name = "parameter_167" + original_name = "batch_norm2d_40.w_1" shape = [192] dtype = "float32" min_val = float("-0.230409") @@ -1848,6 +2016,7 @@ class Program_weight_tensor_parameter_167: class Program_weight_tensor_parameter_168: name = "parameter_168" + original_name = "conv2d_42.w_0" shape = [192, 384, 1, 1] dtype = "float32" min_val = float("-0.108831") @@ -1859,6 +2028,7 @@ class Program_weight_tensor_parameter_168: class Program_weight_tensor_parameter_169: name = "parameter_169" + original_name = "batch_norm2d_39.b_0" shape = [192] dtype = "float32" min_val = float("-2.96764") @@ -1870,6 +2040,7 @@ class Program_weight_tensor_parameter_169: class Program_weight_tensor_parameter_170: name = "parameter_170" + original_name = "batch_norm2d_39.w_0" shape = [192] dtype = "float32" min_val = float("0.830791") @@ -1881,6 +2052,7 @@ class Program_weight_tensor_parameter_170: class Program_weight_tensor_parameter_171: name = "parameter_171" + original_name = "batch_norm2d_39.w_2" shape = [192] dtype = "float32" min_val = float("0.00614721") @@ -1892,6 +2064,7 @@ class Program_weight_tensor_parameter_171: class Program_weight_tensor_parameter_172: name = "parameter_172" + original_name = "batch_norm2d_39.w_1" shape = [192] dtype = "float32" min_val = float("-0.133354") @@ -1903,6 +2076,7 @@ class Program_weight_tensor_parameter_172: class Program_weight_tensor_parameter_173: name = "parameter_173" + original_name = "conv2d_41.w_0" shape = [192, 384, 1, 1] dtype = "float32" min_val = float("-0.0985625") @@ -1914,6 +2088,7 @@ class Program_weight_tensor_parameter_173: class Program_weight_tensor_parameter_174: name = "parameter_174" + original_name = "batch_norm2d_38.b_0" shape = [384] dtype = "float32" min_val = float("-2.92359") @@ -1925,6 +2100,7 @@ class Program_weight_tensor_parameter_174: class Program_weight_tensor_parameter_175: name = "parameter_175" + original_name = "batch_norm2d_38.w_0" shape = [384] dtype = "float32" min_val = float("0.631853") @@ -1936,6 +2112,7 @@ class Program_weight_tensor_parameter_175: class Program_weight_tensor_parameter_176: name = "parameter_176" + original_name = "batch_norm2d_38.w_2" shape = [384] dtype = "float32" min_val = float("0.0104507") @@ -1947,6 +2124,7 @@ class Program_weight_tensor_parameter_176: class Program_weight_tensor_parameter_177: name = "parameter_177" + original_name = "batch_norm2d_38.w_1" shape = [384] dtype = "float32" min_val = float("-0.269997") @@ -1958,6 +2136,7 @@ class Program_weight_tensor_parameter_177: class Program_weight_tensor_parameter_178: name = "parameter_178" + original_name = "conv2d_40.w_0" shape = [384, 256, 3, 3] dtype = "float32" min_val = float("-0.0753194") @@ -1969,6 +2148,7 @@ class Program_weight_tensor_parameter_178: class Program_weight_tensor_parameter_179: name = "parameter_179" + original_name = "batch_norm2d_37.b_0" shape = [256] dtype = "float32" min_val = float("-2.04502") @@ -1980,6 +2160,7 @@ class Program_weight_tensor_parameter_179: class Program_weight_tensor_parameter_180: name = "parameter_180" + original_name = "batch_norm2d_37.w_0" shape = [256] dtype = "float32" min_val = float("0.517239") @@ -1991,6 +2172,7 @@ class Program_weight_tensor_parameter_180: class Program_weight_tensor_parameter_181: name = "parameter_181" + original_name = "batch_norm2d_37.w_2" shape = [256] dtype = "float32" min_val = float("0.00195657") @@ -2002,6 +2184,7 @@ class Program_weight_tensor_parameter_181: class Program_weight_tensor_parameter_182: name = "parameter_182" + original_name = "batch_norm2d_37.w_1" shape = [256] dtype = "float32" min_val = float("-0.230372") @@ -2013,6 +2196,7 @@ class Program_weight_tensor_parameter_182: class Program_weight_tensor_parameter_183: name = "parameter_183" + original_name = "conv2d_39.w_0" shape = [256, 192, 1, 1] dtype = "float32" min_val = float("-0.206154") @@ -2024,6 +2208,7 @@ class Program_weight_tensor_parameter_183: class Program_weight_tensor_parameter_184: name = "parameter_184" + original_name = "conv2d_38.b_0" shape = [192] dtype = "float32" min_val = float("-0.0139357") @@ -2035,6 +2220,7 @@ class Program_weight_tensor_parameter_184: class Program_weight_tensor_parameter_185: name = "parameter_185" + original_name = "conv2d_38.w_0" shape = [192, 192, 1, 1] dtype = "float32" min_val = float("-0.347135") @@ -2046,6 +2232,7 @@ class Program_weight_tensor_parameter_185: class Program_weight_tensor_parameter_186: name = "parameter_186" + original_name = "batch_norm2d_36.b_0" shape = [96] dtype = "float32" min_val = float("-1.91355") @@ -2057,6 +2244,7 @@ class Program_weight_tensor_parameter_186: class Program_weight_tensor_parameter_187: name = "parameter_187" + original_name = "batch_norm2d_36.w_0" shape = [96] dtype = "float32" min_val = float("0.142427") @@ -2068,6 +2256,7 @@ class Program_weight_tensor_parameter_187: class Program_weight_tensor_parameter_188: name = "parameter_188" + original_name = "batch_norm2d_36.w_2" shape = [96] dtype = "float32" min_val = float("7.85249e-05") @@ -2079,6 +2268,7 @@ class Program_weight_tensor_parameter_188: class Program_weight_tensor_parameter_189: name = "parameter_189" + original_name = "batch_norm2d_36.w_1" shape = [96] dtype = "float32" min_val = float("-0.0546919") @@ -2090,6 +2280,7 @@ class Program_weight_tensor_parameter_189: class Program_weight_tensor_parameter_190: name = "parameter_190" + original_name = "conv2d_37.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.0500852") @@ -2101,6 +2292,7 @@ class Program_weight_tensor_parameter_190: class Program_weight_tensor_parameter_191: name = "parameter_191" + original_name = "batch_norm2d_35.b_0" shape = [96] dtype = "float32" min_val = float("-1.91314") @@ -2112,6 +2304,7 @@ class Program_weight_tensor_parameter_191: class Program_weight_tensor_parameter_192: name = "parameter_192" + original_name = "batch_norm2d_35.w_0" shape = [96] dtype = "float32" min_val = float("0.343774") @@ -2123,6 +2316,7 @@ class Program_weight_tensor_parameter_192: class Program_weight_tensor_parameter_193: name = "parameter_193" + original_name = "batch_norm2d_35.w_2" shape = [96] dtype = "float32" min_val = float("0.000942165") @@ -2134,6 +2328,7 @@ class Program_weight_tensor_parameter_193: class Program_weight_tensor_parameter_194: name = "parameter_194" + original_name = "batch_norm2d_35.w_1" shape = [96] dtype = "float32" min_val = float("-0.137594") @@ -2145,6 +2340,7 @@ class Program_weight_tensor_parameter_194: class Program_weight_tensor_parameter_195: name = "parameter_195" + original_name = "conv2d_36.w_0" shape = [96, 96, 3, 3] dtype = "float32" min_val = float("-0.0398886") @@ -2156,6 +2352,7 @@ class Program_weight_tensor_parameter_195: class Program_weight_tensor_parameter_196: name = "parameter_196" + original_name = "batch_norm2d_34.b_0" shape = [96] dtype = "float32" min_val = float("-2.46605") @@ -2167,6 +2364,7 @@ class Program_weight_tensor_parameter_196: class Program_weight_tensor_parameter_197: name = "parameter_197" + original_name = "batch_norm2d_34.w_0" shape = [96] dtype = "float32" min_val = float("0.542082") @@ -2178,6 +2376,7 @@ class Program_weight_tensor_parameter_197: class Program_weight_tensor_parameter_198: name = "parameter_198" + original_name = "batch_norm2d_34.w_2" shape = [96] dtype = "float32" min_val = float("0.041446") @@ -2189,6 +2388,7 @@ class Program_weight_tensor_parameter_198: class Program_weight_tensor_parameter_199: name = "parameter_199" + original_name = "batch_norm2d_34.w_1" shape = [96] dtype = "float32" min_val = float("-2.80547") @@ -2200,6 +2400,7 @@ class Program_weight_tensor_parameter_199: class Program_weight_tensor_parameter_200: name = "parameter_200" + original_name = "conv2d_35.w_0" shape = [96, 96, 3, 3] dtype = "float32" min_val = float("-0.150203") @@ -2211,6 +2412,7 @@ class Program_weight_tensor_parameter_200: class Program_weight_tensor_parameter_201: name = "parameter_201" + original_name = "batch_norm2d_33.b_0" shape = [96] dtype = "float32" min_val = float("-1.38826") @@ -2222,6 +2424,7 @@ class Program_weight_tensor_parameter_201: class Program_weight_tensor_parameter_202: name = "parameter_202" + original_name = "batch_norm2d_33.w_0" shape = [96] dtype = "float32" min_val = float("0.0453402") @@ -2233,6 +2436,7 @@ class Program_weight_tensor_parameter_202: class Program_weight_tensor_parameter_203: name = "parameter_203" + original_name = "batch_norm2d_33.w_2" shape = [96] dtype = "float32" min_val = float("7.55835e-05") @@ -2244,6 +2448,7 @@ class Program_weight_tensor_parameter_203: class Program_weight_tensor_parameter_204: name = "parameter_204" + original_name = "batch_norm2d_33.w_1" shape = [96] dtype = "float32" min_val = float("-0.048531") @@ -2255,6 +2460,7 @@ class Program_weight_tensor_parameter_204: class Program_weight_tensor_parameter_205: name = "parameter_205" + original_name = "conv2d_34.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.0483138") @@ -2266,6 +2472,7 @@ class Program_weight_tensor_parameter_205: class Program_weight_tensor_parameter_206: name = "parameter_206" + original_name = "batch_norm2d_32.b_0" shape = [96] dtype = "float32" min_val = float("-1.38834") @@ -2277,6 +2484,7 @@ class Program_weight_tensor_parameter_206: class Program_weight_tensor_parameter_207: name = "parameter_207" + original_name = "batch_norm2d_32.w_0" shape = [96] dtype = "float32" min_val = float("0.370504") @@ -2288,6 +2496,7 @@ class Program_weight_tensor_parameter_207: class Program_weight_tensor_parameter_208: name = "parameter_208" + original_name = "batch_norm2d_32.w_2" shape = [96] dtype = "float32" min_val = float("0.00323837") @@ -2299,6 +2508,7 @@ class Program_weight_tensor_parameter_208: class Program_weight_tensor_parameter_209: name = "parameter_209" + original_name = "batch_norm2d_32.w_1" shape = [96] dtype = "float32" min_val = float("-0.0965735") @@ -2310,6 +2520,7 @@ class Program_weight_tensor_parameter_209: class Program_weight_tensor_parameter_210: name = "parameter_210" + original_name = "conv2d_33.w_0" shape = [96, 96, 3, 3] dtype = "float32" min_val = float("-0.058655") @@ -2321,6 +2532,7 @@ class Program_weight_tensor_parameter_210: class Program_weight_tensor_parameter_211: name = "parameter_211" + original_name = "batch_norm2d_31.b_0" shape = [96] dtype = "float32" min_val = float("-3.31955") @@ -2332,6 +2544,7 @@ class Program_weight_tensor_parameter_211: class Program_weight_tensor_parameter_212: name = "parameter_212" + original_name = "batch_norm2d_31.w_0" shape = [96] dtype = "float32" min_val = float("0.473098") @@ -2343,6 +2556,7 @@ class Program_weight_tensor_parameter_212: class Program_weight_tensor_parameter_213: name = "parameter_213" + original_name = "batch_norm2d_31.w_2" shape = [96] dtype = "float32" min_val = float("0.0282091") @@ -2354,6 +2568,7 @@ class Program_weight_tensor_parameter_213: class Program_weight_tensor_parameter_214: name = "parameter_214" + original_name = "batch_norm2d_31.w_1" shape = [96] dtype = "float32" min_val = float("-1.24896") @@ -2365,6 +2580,7 @@ class Program_weight_tensor_parameter_214: class Program_weight_tensor_parameter_215: name = "parameter_215" + original_name = "conv2d_32.w_0" shape = [96, 96, 3, 3] dtype = "float32" min_val = float("-0.147666") @@ -2376,6 +2592,7 @@ class Program_weight_tensor_parameter_215: class Program_weight_tensor_parameter_216: name = "parameter_216" + original_name = "batch_norm2d_30.b_0" shape = [96] dtype = "float32" min_val = float("-1.24956") @@ -2387,6 +2604,7 @@ class Program_weight_tensor_parameter_216: class Program_weight_tensor_parameter_217: name = "parameter_217" + original_name = "batch_norm2d_30.w_0" shape = [96] dtype = "float32" min_val = float("0.0243293") @@ -2398,6 +2616,7 @@ class Program_weight_tensor_parameter_217: class Program_weight_tensor_parameter_218: name = "parameter_218" + original_name = "batch_norm2d_30.w_2" shape = [96] dtype = "float32" min_val = float("6.35226e-05") @@ -2409,6 +2628,7 @@ class Program_weight_tensor_parameter_218: class Program_weight_tensor_parameter_219: name = "parameter_219" + original_name = "batch_norm2d_30.w_1" shape = [96] dtype = "float32" min_val = float("-0.0385319") @@ -2420,6 +2640,7 @@ class Program_weight_tensor_parameter_219: class Program_weight_tensor_parameter_220: name = "parameter_220" + original_name = "conv2d_31.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.0448708") @@ -2431,6 +2652,7 @@ class Program_weight_tensor_parameter_220: class Program_weight_tensor_parameter_221: name = "parameter_221" + original_name = "batch_norm2d_29.b_0" shape = [96] dtype = "float32" min_val = float("-1.24942") @@ -2442,6 +2664,7 @@ class Program_weight_tensor_parameter_221: class Program_weight_tensor_parameter_222: name = "parameter_222" + original_name = "batch_norm2d_29.w_0" shape = [96] dtype = "float32" min_val = float("0.315495") @@ -2453,6 +2676,7 @@ class Program_weight_tensor_parameter_222: class Program_weight_tensor_parameter_223: name = "parameter_223" + original_name = "batch_norm2d_29.w_2" shape = [96] dtype = "float32" min_val = float("0.00338498") @@ -2464,6 +2688,7 @@ class Program_weight_tensor_parameter_223: class Program_weight_tensor_parameter_224: name = "parameter_224" + original_name = "batch_norm2d_29.w_1" shape = [96] dtype = "float32" min_val = float("-0.0546488") @@ -2475,6 +2700,7 @@ class Program_weight_tensor_parameter_224: class Program_weight_tensor_parameter_225: name = "parameter_225" + original_name = "conv2d_30.w_0" shape = [96, 96, 3, 3] dtype = "float32" min_val = float("-0.065253") @@ -2486,6 +2712,7 @@ class Program_weight_tensor_parameter_225: class Program_weight_tensor_parameter_226: name = "parameter_226" + original_name = "batch_norm2d_28.b_0" shape = [96] dtype = "float32" min_val = float("-3.58296") @@ -2497,6 +2724,7 @@ class Program_weight_tensor_parameter_226: class Program_weight_tensor_parameter_227: name = "parameter_227" + original_name = "batch_norm2d_28.w_0" shape = [96] dtype = "float32" min_val = float("0.511106") @@ -2508,6 +2736,7 @@ class Program_weight_tensor_parameter_227: class Program_weight_tensor_parameter_228: name = "parameter_228" + original_name = "batch_norm2d_28.w_2" shape = [96] dtype = "float32" min_val = float("0.0201149") @@ -2519,6 +2748,7 @@ class Program_weight_tensor_parameter_228: class Program_weight_tensor_parameter_229: name = "parameter_229" + original_name = "batch_norm2d_28.w_1" shape = [96] dtype = "float32" min_val = float("-0.822261") @@ -2530,6 +2760,7 @@ class Program_weight_tensor_parameter_229: class Program_weight_tensor_parameter_230: name = "parameter_230" + original_name = "conv2d_29.w_0" shape = [96, 96, 3, 3] dtype = "float32" min_val = float("-0.0973524") @@ -2541,6 +2772,7 @@ class Program_weight_tensor_parameter_230: class Program_weight_tensor_parameter_231: name = "parameter_231" + original_name = "batch_norm2d_27.b_0" shape = [96] dtype = "float32" min_val = float("-0.892064") @@ -2552,6 +2784,7 @@ class Program_weight_tensor_parameter_231: class Program_weight_tensor_parameter_232: name = "parameter_232" + original_name = "batch_norm2d_27.w_0" shape = [96] dtype = "float32" min_val = float("0.0191223") @@ -2563,6 +2796,7 @@ class Program_weight_tensor_parameter_232: class Program_weight_tensor_parameter_233: name = "parameter_233" + original_name = "batch_norm2d_27.w_2" shape = [96] dtype = "float32" min_val = float("4.77273e-05") @@ -2574,6 +2808,7 @@ class Program_weight_tensor_parameter_233: class Program_weight_tensor_parameter_234: name = "parameter_234" + original_name = "batch_norm2d_27.w_1" shape = [96] dtype = "float32" min_val = float("-0.0328433") @@ -2585,6 +2820,7 @@ class Program_weight_tensor_parameter_234: class Program_weight_tensor_parameter_235: name = "parameter_235" + original_name = "conv2d_28.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.0499906") @@ -2596,6 +2832,7 @@ class Program_weight_tensor_parameter_235: class Program_weight_tensor_parameter_236: name = "parameter_236" + original_name = "batch_norm2d_26.b_0" shape = [96] dtype = "float32" min_val = float("-0.891955") @@ -2607,6 +2844,7 @@ class Program_weight_tensor_parameter_236: class Program_weight_tensor_parameter_237: name = "parameter_237" + original_name = "batch_norm2d_26.w_0" shape = [96] dtype = "float32" min_val = float("0.17446") @@ -2618,6 +2856,7 @@ class Program_weight_tensor_parameter_237: class Program_weight_tensor_parameter_238: name = "parameter_238" + original_name = "batch_norm2d_26.w_2" shape = [96] dtype = "float32" min_val = float("0.00236768") @@ -2629,6 +2868,7 @@ class Program_weight_tensor_parameter_238: class Program_weight_tensor_parameter_239: name = "parameter_239" + original_name = "batch_norm2d_26.w_1" shape = [96] dtype = "float32" min_val = float("-0.0582404") @@ -2640,6 +2880,7 @@ class Program_weight_tensor_parameter_239: class Program_weight_tensor_parameter_240: name = "parameter_240" + original_name = "conv2d_27.w_0" shape = [96, 96, 3, 3] dtype = "float32" min_val = float("-0.057305") @@ -2651,6 +2892,7 @@ class Program_weight_tensor_parameter_240: class Program_weight_tensor_parameter_241: name = "parameter_241" + original_name = "batch_norm2d_25.b_0" shape = [96] dtype = "float32" min_val = float("-2.65777") @@ -2662,6 +2904,7 @@ class Program_weight_tensor_parameter_241: class Program_weight_tensor_parameter_242: name = "parameter_242" + original_name = "batch_norm2d_25.w_0" shape = [96] dtype = "float32" min_val = float("0.512951") @@ -2673,6 +2916,7 @@ class Program_weight_tensor_parameter_242: class Program_weight_tensor_parameter_243: name = "parameter_243" + original_name = "batch_norm2d_25.w_2" shape = [96] dtype = "float32" min_val = float("0.0176967") @@ -2684,6 +2928,7 @@ class Program_weight_tensor_parameter_243: class Program_weight_tensor_parameter_244: name = "parameter_244" + original_name = "batch_norm2d_25.w_1" shape = [96] dtype = "float32" min_val = float("-0.761559") @@ -2695,6 +2940,7 @@ class Program_weight_tensor_parameter_244: class Program_weight_tensor_parameter_245: name = "parameter_245" + original_name = "conv2d_26.w_0" shape = [96, 96, 3, 3] dtype = "float32" min_val = float("-0.0738037") @@ -2706,6 +2952,7 @@ class Program_weight_tensor_parameter_245: class Program_weight_tensor_parameter_246: name = "parameter_246" + original_name = "batch_norm2d_24.b_0" shape = [96] dtype = "float32" min_val = float("-0.978262") @@ -2717,6 +2964,7 @@ class Program_weight_tensor_parameter_246: class Program_weight_tensor_parameter_247: name = "parameter_247" + original_name = "batch_norm2d_24.w_0" shape = [96] dtype = "float32" min_val = float("0.0498074") @@ -2728,6 +2976,7 @@ class Program_weight_tensor_parameter_247: class Program_weight_tensor_parameter_248: name = "parameter_248" + original_name = "batch_norm2d_24.w_2" shape = [96] dtype = "float32" min_val = float("0.000180546") @@ -2739,6 +2988,7 @@ class Program_weight_tensor_parameter_248: class Program_weight_tensor_parameter_249: name = "parameter_249" + original_name = "batch_norm2d_24.w_1" shape = [96] dtype = "float32" min_val = float("-0.041806") @@ -2750,6 +3000,7 @@ class Program_weight_tensor_parameter_249: class Program_weight_tensor_parameter_250: name = "parameter_250" + original_name = "conv2d_25.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.0731207") @@ -2761,6 +3012,7 @@ class Program_weight_tensor_parameter_250: class Program_weight_tensor_parameter_251: name = "parameter_251" + original_name = "batch_norm2d_23.b_0" shape = [96] dtype = "float32" min_val = float("-0.978083") @@ -2772,6 +3024,7 @@ class Program_weight_tensor_parameter_251: class Program_weight_tensor_parameter_252: name = "parameter_252" + original_name = "batch_norm2d_23.w_0" shape = [96] dtype = "float32" min_val = float("0.236133") @@ -2783,6 +3036,7 @@ class Program_weight_tensor_parameter_252: class Program_weight_tensor_parameter_253: name = "parameter_253" + original_name = "batch_norm2d_23.w_2" shape = [96] dtype = "float32" min_val = float("0.00612804") @@ -2794,6 +3048,7 @@ class Program_weight_tensor_parameter_253: class Program_weight_tensor_parameter_254: name = "parameter_254" + original_name = "batch_norm2d_23.w_1" shape = [96] dtype = "float32" min_val = float("-0.0713362") @@ -2805,6 +3060,7 @@ class Program_weight_tensor_parameter_254: class Program_weight_tensor_parameter_255: name = "parameter_255" + original_name = "conv2d_24.w_0" shape = [96, 96, 3, 3] dtype = "float32" min_val = float("-0.0654835") @@ -2816,6 +3072,7 @@ class Program_weight_tensor_parameter_255: class Program_weight_tensor_parameter_256: name = "parameter_256" + original_name = "batch_norm2d_22.b_0" shape = [96] dtype = "float32" min_val = float("-3.46434") @@ -2827,6 +3084,7 @@ class Program_weight_tensor_parameter_256: class Program_weight_tensor_parameter_257: name = "parameter_257" + original_name = "batch_norm2d_22.w_0" shape = [96] dtype = "float32" min_val = float("0.686506") @@ -2838,6 +3096,7 @@ class Program_weight_tensor_parameter_257: class Program_weight_tensor_parameter_258: name = "parameter_258" + original_name = "batch_norm2d_22.w_2" shape = [96] dtype = "float32" min_val = float("0.013513") @@ -2849,6 +3108,7 @@ class Program_weight_tensor_parameter_258: class Program_weight_tensor_parameter_259: name = "parameter_259" + original_name = "batch_norm2d_22.w_1" shape = [96] dtype = "float32" min_val = float("-0.482054") @@ -2860,6 +3120,7 @@ class Program_weight_tensor_parameter_259: class Program_weight_tensor_parameter_260: name = "parameter_260" + original_name = "conv2d_23.w_0" shape = [96, 96, 3, 3] dtype = "float32" min_val = float("-0.0824841") @@ -2871,6 +3132,7 @@ class Program_weight_tensor_parameter_260: class Program_weight_tensor_parameter_261: name = "parameter_261" + original_name = "batch_norm2d_21.b_0" shape = [96] dtype = "float32" min_val = float("-0.625302") @@ -2882,6 +3144,7 @@ class Program_weight_tensor_parameter_261: class Program_weight_tensor_parameter_262: name = "parameter_262" + original_name = "batch_norm2d_21.w_0" shape = [96] dtype = "float32" min_val = float("0.0910018") @@ -2893,6 +3156,7 @@ class Program_weight_tensor_parameter_262: class Program_weight_tensor_parameter_263: name = "parameter_263" + original_name = "batch_norm2d_21.w_2" shape = [96] dtype = "float32" min_val = float("0.000380114") @@ -2904,6 +3168,7 @@ class Program_weight_tensor_parameter_263: class Program_weight_tensor_parameter_264: name = "parameter_264" + original_name = "batch_norm2d_21.w_1" shape = [96] dtype = "float32" min_val = float("-0.0360021") @@ -2915,6 +3180,7 @@ class Program_weight_tensor_parameter_264: class Program_weight_tensor_parameter_265: name = "parameter_265" + original_name = "conv2d_22.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.0925016") @@ -2926,6 +3192,7 @@ class Program_weight_tensor_parameter_265: class Program_weight_tensor_parameter_266: name = "parameter_266" + original_name = "batch_norm2d_20.b_0" shape = [96] dtype = "float32" min_val = float("-0.625183") @@ -2937,6 +3204,7 @@ class Program_weight_tensor_parameter_266: class Program_weight_tensor_parameter_267: name = "parameter_267" + original_name = "batch_norm2d_20.w_0" shape = [96] dtype = "float32" min_val = float("0.210658") @@ -2948,6 +3216,7 @@ class Program_weight_tensor_parameter_267: class Program_weight_tensor_parameter_268: name = "parameter_268" + original_name = "batch_norm2d_20.w_2" shape = [96] dtype = "float32" min_val = float("0.0102276") @@ -2959,6 +3228,7 @@ class Program_weight_tensor_parameter_268: class Program_weight_tensor_parameter_269: name = "parameter_269" + original_name = "batch_norm2d_20.w_1" shape = [96] dtype = "float32" min_val = float("-0.108112") @@ -2970,6 +3240,7 @@ class Program_weight_tensor_parameter_269: class Program_weight_tensor_parameter_270: name = "parameter_270" + original_name = "conv2d_21.w_0" shape = [96, 96, 3, 3] dtype = "float32" min_val = float("-0.0885375") @@ -2981,6 +3252,7 @@ class Program_weight_tensor_parameter_270: class Program_weight_tensor_parameter_271: name = "parameter_271" + original_name = "batch_norm2d_19.b_0" shape = [96] dtype = "float32" min_val = float("-2.40893") @@ -2992,6 +3264,7 @@ class Program_weight_tensor_parameter_271: class Program_weight_tensor_parameter_272: name = "parameter_272" + original_name = "batch_norm2d_19.w_0" shape = [96] dtype = "float32" min_val = float("0.853968") @@ -3003,6 +3276,7 @@ class Program_weight_tensor_parameter_272: class Program_weight_tensor_parameter_273: name = "parameter_273" + original_name = "batch_norm2d_19.w_2" shape = [96] dtype = "float32" min_val = float("0.011454") @@ -3014,6 +3288,7 @@ class Program_weight_tensor_parameter_273: class Program_weight_tensor_parameter_274: name = "parameter_274" + original_name = "batch_norm2d_19.w_1" shape = [96] dtype = "float32" min_val = float("-0.570491") @@ -3025,6 +3300,7 @@ class Program_weight_tensor_parameter_274: class Program_weight_tensor_parameter_275: name = "parameter_275" + original_name = "conv2d_20.w_0" shape = [96, 96, 3, 3] dtype = "float32" min_val = float("-0.15411") @@ -3036,6 +3312,7 @@ class Program_weight_tensor_parameter_275: class Program_weight_tensor_parameter_276: name = "parameter_276" + original_name = "batch_norm2d_18.b_0" shape = [96] dtype = "float32" min_val = float("-3.16609") @@ -3047,6 +3324,7 @@ class Program_weight_tensor_parameter_276: class Program_weight_tensor_parameter_277: name = "parameter_277" + original_name = "batch_norm2d_18.w_0" shape = [96] dtype = "float32" min_val = float("0.214988") @@ -3058,6 +3336,7 @@ class Program_weight_tensor_parameter_277: class Program_weight_tensor_parameter_278: name = "parameter_278" + original_name = "batch_norm2d_18.w_2" shape = [96] dtype = "float32" min_val = float("0.00763171") @@ -3069,6 +3348,7 @@ class Program_weight_tensor_parameter_278: class Program_weight_tensor_parameter_279: name = "parameter_279" + original_name = "batch_norm2d_18.w_1" shape = [96] dtype = "float32" min_val = float("-0.271678") @@ -3080,6 +3360,7 @@ class Program_weight_tensor_parameter_279: class Program_weight_tensor_parameter_280: name = "parameter_280" + original_name = "conv2d_19.w_0" shape = [96, 192, 1, 1] dtype = "float32" min_val = float("-0.186901") @@ -3091,6 +3372,7 @@ class Program_weight_tensor_parameter_280: class Program_weight_tensor_parameter_281: name = "parameter_281" + original_name = "batch_norm2d_17.b_0" shape = [96] dtype = "float32" min_val = float("-4.92284") @@ -3102,6 +3384,7 @@ class Program_weight_tensor_parameter_281: class Program_weight_tensor_parameter_282: name = "parameter_282" + original_name = "batch_norm2d_17.w_0" shape = [96] dtype = "float32" min_val = float("0.414126") @@ -3113,6 +3396,7 @@ class Program_weight_tensor_parameter_282: class Program_weight_tensor_parameter_283: name = "parameter_283" + original_name = "batch_norm2d_17.w_2" shape = [96] dtype = "float32" min_val = float("0.00531954") @@ -3124,6 +3408,7 @@ class Program_weight_tensor_parameter_283: class Program_weight_tensor_parameter_284: name = "parameter_284" + original_name = "batch_norm2d_17.w_1" shape = [96] dtype = "float32" min_val = float("-0.17223") @@ -3135,6 +3420,7 @@ class Program_weight_tensor_parameter_284: class Program_weight_tensor_parameter_285: name = "parameter_285" + original_name = "conv2d_18.w_0" shape = [96, 192, 1, 1] dtype = "float32" min_val = float("-0.116975") @@ -3146,6 +3432,7 @@ class Program_weight_tensor_parameter_285: class Program_weight_tensor_parameter_286: name = "parameter_286" + original_name = "batch_norm2d_16.b_0" shape = [192] dtype = "float32" min_val = float("-2.27475") @@ -3157,6 +3444,7 @@ class Program_weight_tensor_parameter_286: class Program_weight_tensor_parameter_287: name = "parameter_287" + original_name = "batch_norm2d_16.w_0" shape = [192] dtype = "float32" min_val = float("0.632268") @@ -3168,6 +3456,7 @@ class Program_weight_tensor_parameter_287: class Program_weight_tensor_parameter_288: name = "parameter_288" + original_name = "batch_norm2d_16.w_2" shape = [192] dtype = "float32" min_val = float("0.0109431") @@ -3179,6 +3468,7 @@ class Program_weight_tensor_parameter_288: class Program_weight_tensor_parameter_289: name = "parameter_289" + original_name = "batch_norm2d_16.w_1" shape = [192] dtype = "float32" min_val = float("-0.576887") @@ -3190,6 +3480,7 @@ class Program_weight_tensor_parameter_289: class Program_weight_tensor_parameter_290: name = "parameter_290" + original_name = "conv2d_17.w_0" shape = [192, 128, 3, 3] dtype = "float32" min_val = float("-0.0856428") @@ -3201,6 +3492,7 @@ class Program_weight_tensor_parameter_290: class Program_weight_tensor_parameter_291: name = "parameter_291" + original_name = "batch_norm2d_15.b_0" shape = [128] dtype = "float32" min_val = float("-2.81597") @@ -3212,6 +3504,7 @@ class Program_weight_tensor_parameter_291: class Program_weight_tensor_parameter_292: name = "parameter_292" + original_name = "batch_norm2d_15.w_0" shape = [128] dtype = "float32" min_val = float("0.311227") @@ -3223,6 +3516,7 @@ class Program_weight_tensor_parameter_292: class Program_weight_tensor_parameter_293: name = "parameter_293" + original_name = "batch_norm2d_15.w_2" shape = [128] dtype = "float32" min_val = float("0.000843216") @@ -3234,6 +3528,7 @@ class Program_weight_tensor_parameter_293: class Program_weight_tensor_parameter_294: name = "parameter_294" + original_name = "batch_norm2d_15.w_1" shape = [128] dtype = "float32" min_val = float("-0.237755") @@ -3245,6 +3540,7 @@ class Program_weight_tensor_parameter_294: class Program_weight_tensor_parameter_295: name = "parameter_295" + original_name = "conv2d_16.w_0" shape = [128, 96, 1, 1] dtype = "float32" min_val = float("-0.171773") @@ -3256,6 +3552,7 @@ class Program_weight_tensor_parameter_295: class Program_weight_tensor_parameter_296: name = "parameter_296" + original_name = "conv2d_15.b_0" shape = [96] dtype = "float32" min_val = float("-0.0180386") @@ -3267,6 +3564,7 @@ class Program_weight_tensor_parameter_296: class Program_weight_tensor_parameter_297: name = "parameter_297" + original_name = "conv2d_15.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.30281") @@ -3278,6 +3576,7 @@ class Program_weight_tensor_parameter_297: class Program_weight_tensor_parameter_298: name = "parameter_298" + original_name = "batch_norm2d_14.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -3287,6 +3586,7 @@ class Program_weight_tensor_parameter_298: class Program_weight_tensor_parameter_299: name = "parameter_299" + original_name = "batch_norm2d_14.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -3296,6 +3596,7 @@ class Program_weight_tensor_parameter_299: class Program_weight_tensor_parameter_300: name = "parameter_300" + original_name = "batch_norm2d_14.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -3305,6 +3606,7 @@ class Program_weight_tensor_parameter_300: class Program_weight_tensor_parameter_301: name = "parameter_301" + original_name = "batch_norm2d_14.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -3314,6 +3616,7 @@ class Program_weight_tensor_parameter_301: class Program_weight_tensor_parameter_302: name = "parameter_302" + original_name = "conv2d_14.w_0" shape = [48, 48, 1, 1] dtype = "float32" min_val = float("-0.0501789") @@ -3325,6 +3628,7 @@ class Program_weight_tensor_parameter_302: class Program_weight_tensor_parameter_303: name = "parameter_303" + original_name = "batch_norm2d_13.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -3334,6 +3638,7 @@ class Program_weight_tensor_parameter_303: class Program_weight_tensor_parameter_304: name = "parameter_304" + original_name = "batch_norm2d_13.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -3343,6 +3648,7 @@ class Program_weight_tensor_parameter_304: class Program_weight_tensor_parameter_305: name = "parameter_305" + original_name = "batch_norm2d_13.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -3352,6 +3658,7 @@ class Program_weight_tensor_parameter_305: class Program_weight_tensor_parameter_306: name = "parameter_306" + original_name = "batch_norm2d_13.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -3361,6 +3668,7 @@ class Program_weight_tensor_parameter_306: class Program_weight_tensor_parameter_307: name = "parameter_307" + original_name = "conv2d_13.w_0" shape = [48, 48, 3, 3] dtype = "float32" min_val = float("-0.0578676") @@ -3372,6 +3680,7 @@ class Program_weight_tensor_parameter_307: class Program_weight_tensor_parameter_308: name = "parameter_308" + original_name = "batch_norm2d_12.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -3381,6 +3690,7 @@ class Program_weight_tensor_parameter_308: class Program_weight_tensor_parameter_309: name = "parameter_309" + original_name = "batch_norm2d_12.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -3390,6 +3700,7 @@ class Program_weight_tensor_parameter_309: class Program_weight_tensor_parameter_310: name = "parameter_310" + original_name = "batch_norm2d_12.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -3399,6 +3710,7 @@ class Program_weight_tensor_parameter_310: class Program_weight_tensor_parameter_311: name = "parameter_311" + original_name = "batch_norm2d_12.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -3408,6 +3720,7 @@ class Program_weight_tensor_parameter_311: class Program_weight_tensor_parameter_312: name = "parameter_312" + original_name = "conv2d_12.w_0" shape = [48, 48, 3, 3] dtype = "float32" min_val = float("-0.0925274") @@ -3419,6 +3732,7 @@ class Program_weight_tensor_parameter_312: class Program_weight_tensor_parameter_313: name = "parameter_313" + original_name = "batch_norm2d_11.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -3428,6 +3742,7 @@ class Program_weight_tensor_parameter_313: class Program_weight_tensor_parameter_314: name = "parameter_314" + original_name = "batch_norm2d_11.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -3437,6 +3752,7 @@ class Program_weight_tensor_parameter_314: class Program_weight_tensor_parameter_315: name = "parameter_315" + original_name = "batch_norm2d_11.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -3446,6 +3762,7 @@ class Program_weight_tensor_parameter_315: class Program_weight_tensor_parameter_316: name = "parameter_316" + original_name = "batch_norm2d_11.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -3455,6 +3772,7 @@ class Program_weight_tensor_parameter_316: class Program_weight_tensor_parameter_317: name = "parameter_317" + original_name = "conv2d_11.w_0" shape = [48, 48, 1, 1] dtype = "float32" min_val = float("-0.0727088") @@ -3466,6 +3784,7 @@ class Program_weight_tensor_parameter_317: class Program_weight_tensor_parameter_318: name = "parameter_318" + original_name = "batch_norm2d_10.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -3475,6 +3794,7 @@ class Program_weight_tensor_parameter_318: class Program_weight_tensor_parameter_319: name = "parameter_319" + original_name = "batch_norm2d_10.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -3484,6 +3804,7 @@ class Program_weight_tensor_parameter_319: class Program_weight_tensor_parameter_320: name = "parameter_320" + original_name = "batch_norm2d_10.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -3493,6 +3814,7 @@ class Program_weight_tensor_parameter_320: class Program_weight_tensor_parameter_321: name = "parameter_321" + original_name = "batch_norm2d_10.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -3502,6 +3824,7 @@ class Program_weight_tensor_parameter_321: class Program_weight_tensor_parameter_322: name = "parameter_322" + original_name = "conv2d_10.w_0" shape = [48, 48, 3, 3] dtype = "float32" min_val = float("-0.0621898") @@ -3513,6 +3836,7 @@ class Program_weight_tensor_parameter_322: class Program_weight_tensor_parameter_323: name = "parameter_323" + original_name = "batch_norm2d_9.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -3522,6 +3846,7 @@ class Program_weight_tensor_parameter_323: class Program_weight_tensor_parameter_324: name = "parameter_324" + original_name = "batch_norm2d_9.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -3531,6 +3856,7 @@ class Program_weight_tensor_parameter_324: class Program_weight_tensor_parameter_325: name = "parameter_325" + original_name = "batch_norm2d_9.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -3540,6 +3866,7 @@ class Program_weight_tensor_parameter_325: class Program_weight_tensor_parameter_326: name = "parameter_326" + original_name = "batch_norm2d_9.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -3549,6 +3876,7 @@ class Program_weight_tensor_parameter_326: class Program_weight_tensor_parameter_327: name = "parameter_327" + original_name = "conv2d_9.w_0" shape = [48, 48, 3, 3] dtype = "float32" min_val = float("-0.11162") @@ -3560,6 +3888,7 @@ class Program_weight_tensor_parameter_327: class Program_weight_tensor_parameter_328: name = "parameter_328" + original_name = "batch_norm2d_8.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -3569,6 +3898,7 @@ class Program_weight_tensor_parameter_328: class Program_weight_tensor_parameter_329: name = "parameter_329" + original_name = "batch_norm2d_8.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -3578,6 +3908,7 @@ class Program_weight_tensor_parameter_329: class Program_weight_tensor_parameter_330: name = "parameter_330" + original_name = "batch_norm2d_8.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -3587,6 +3918,7 @@ class Program_weight_tensor_parameter_330: class Program_weight_tensor_parameter_331: name = "parameter_331" + original_name = "batch_norm2d_8.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -3596,6 +3928,7 @@ class Program_weight_tensor_parameter_331: class Program_weight_tensor_parameter_332: name = "parameter_332" + original_name = "conv2d_8.w_0" shape = [48, 48, 1, 1] dtype = "float32" min_val = float("-0.0944494") @@ -3607,6 +3940,7 @@ class Program_weight_tensor_parameter_332: class Program_weight_tensor_parameter_333: name = "parameter_333" + original_name = "batch_norm2d_7.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -3616,6 +3950,7 @@ class Program_weight_tensor_parameter_333: class Program_weight_tensor_parameter_334: name = "parameter_334" + original_name = "batch_norm2d_7.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -3625,6 +3960,7 @@ class Program_weight_tensor_parameter_334: class Program_weight_tensor_parameter_335: name = "parameter_335" + original_name = "batch_norm2d_7.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -3634,6 +3970,7 @@ class Program_weight_tensor_parameter_335: class Program_weight_tensor_parameter_336: name = "parameter_336" + original_name = "batch_norm2d_7.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -3643,6 +3980,7 @@ class Program_weight_tensor_parameter_336: class Program_weight_tensor_parameter_337: name = "parameter_337" + original_name = "conv2d_7.w_0" shape = [48, 48, 3, 3] dtype = "float32" min_val = float("-0.0691644") @@ -3654,6 +3992,7 @@ class Program_weight_tensor_parameter_337: class Program_weight_tensor_parameter_338: name = "parameter_338" + original_name = "batch_norm2d_6.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -3663,6 +4002,7 @@ class Program_weight_tensor_parameter_338: class Program_weight_tensor_parameter_339: name = "parameter_339" + original_name = "batch_norm2d_6.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -3672,6 +4012,7 @@ class Program_weight_tensor_parameter_339: class Program_weight_tensor_parameter_340: name = "parameter_340" + original_name = "batch_norm2d_6.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -3681,6 +4022,7 @@ class Program_weight_tensor_parameter_340: class Program_weight_tensor_parameter_341: name = "parameter_341" + original_name = "batch_norm2d_6.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -3690,6 +4032,7 @@ class Program_weight_tensor_parameter_341: class Program_weight_tensor_parameter_342: name = "parameter_342" + original_name = "conv2d_6.w_0" shape = [48, 48, 3, 3] dtype = "float32" min_val = float("-0.133213") @@ -3701,6 +4044,7 @@ class Program_weight_tensor_parameter_342: class Program_weight_tensor_parameter_343: name = "parameter_343" + original_name = "batch_norm2d_5.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -3710,6 +4054,7 @@ class Program_weight_tensor_parameter_343: class Program_weight_tensor_parameter_344: name = "parameter_344" + original_name = "batch_norm2d_5.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -3719,6 +4064,7 @@ class Program_weight_tensor_parameter_344: class Program_weight_tensor_parameter_345: name = "parameter_345" + original_name = "batch_norm2d_5.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -3728,6 +4074,7 @@ class Program_weight_tensor_parameter_345: class Program_weight_tensor_parameter_346: name = "parameter_346" + original_name = "batch_norm2d_5.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -3737,6 +4084,7 @@ class Program_weight_tensor_parameter_346: class Program_weight_tensor_parameter_347: name = "parameter_347" + original_name = "conv2d_5.w_0" shape = [48, 96, 1, 1] dtype = "float32" min_val = float("-0.17806") @@ -3748,6 +4096,7 @@ class Program_weight_tensor_parameter_347: class Program_weight_tensor_parameter_348: name = "parameter_348" + original_name = "batch_norm2d_4.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -3757,6 +4106,7 @@ class Program_weight_tensor_parameter_348: class Program_weight_tensor_parameter_349: name = "parameter_349" + original_name = "batch_norm2d_4.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -3766,6 +4116,7 @@ class Program_weight_tensor_parameter_349: class Program_weight_tensor_parameter_350: name = "parameter_350" + original_name = "batch_norm2d_4.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -3775,6 +4126,7 @@ class Program_weight_tensor_parameter_350: class Program_weight_tensor_parameter_351: name = "parameter_351" + original_name = "batch_norm2d_4.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -3784,6 +4136,7 @@ class Program_weight_tensor_parameter_351: class Program_weight_tensor_parameter_352: name = "parameter_352" + original_name = "conv2d_4.w_0" shape = [48, 96, 1, 1] dtype = "float32" min_val = float("-0.135012") @@ -3795,6 +4148,7 @@ class Program_weight_tensor_parameter_352: class Program_weight_tensor_parameter_353: name = "parameter_353" + original_name = "batch_norm2d_3.b_0" shape = [96] dtype = "float32" min_val = float("-3.40701") @@ -3806,6 +4160,7 @@ class Program_weight_tensor_parameter_353: class Program_weight_tensor_parameter_354: name = "parameter_354" + original_name = "batch_norm2d_3.w_0" shape = [96] dtype = "float32" min_val = float("0.865919") @@ -3817,6 +4172,7 @@ class Program_weight_tensor_parameter_354: class Program_weight_tensor_parameter_355: name = "parameter_355" + original_name = "batch_norm2d_3.w_2" shape = [96] dtype = "float32" min_val = float("0.705639") @@ -3828,6 +4184,7 @@ class Program_weight_tensor_parameter_355: class Program_weight_tensor_parameter_356: name = "parameter_356" + original_name = "batch_norm2d_3.w_1" shape = [96] dtype = "float32" min_val = float("-1.47426") @@ -3839,6 +4196,7 @@ class Program_weight_tensor_parameter_356: class Program_weight_tensor_parameter_357: name = "parameter_357" + original_name = "conv2d_3.w_0" shape = [96, 64, 3, 3] dtype = "float32" min_val = float("-0.110689") @@ -3850,6 +4208,7 @@ class Program_weight_tensor_parameter_357: class Program_weight_tensor_parameter_358: name = "parameter_358" + original_name = "batch_norm2d_2.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -3859,6 +4218,7 @@ class Program_weight_tensor_parameter_358: class Program_weight_tensor_parameter_359: name = "parameter_359" + original_name = "batch_norm2d_2.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -3868,6 +4228,7 @@ class Program_weight_tensor_parameter_359: class Program_weight_tensor_parameter_360: name = "parameter_360" + original_name = "batch_norm2d_2.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -3877,6 +4238,7 @@ class Program_weight_tensor_parameter_360: class Program_weight_tensor_parameter_361: name = "parameter_361" + original_name = "batch_norm2d_2.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -3886,6 +4248,7 @@ class Program_weight_tensor_parameter_361: class Program_weight_tensor_parameter_362: name = "parameter_362" + original_name = "conv2d_2.w_0" shape = [64, 32, 3, 3] dtype = "float32" min_val = float("-0.179264") @@ -3897,6 +4260,7 @@ class Program_weight_tensor_parameter_362: class Program_weight_tensor_parameter_363: name = "parameter_363" + original_name = "batch_norm2d_1.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -3906,6 +4270,7 @@ class Program_weight_tensor_parameter_363: class Program_weight_tensor_parameter_364: name = "parameter_364" + original_name = "batch_norm2d_1.w_0" shape = [32] dtype = "float32" min_val = float("0") @@ -3915,6 +4280,7 @@ class Program_weight_tensor_parameter_364: class Program_weight_tensor_parameter_365: name = "parameter_365" + original_name = "batch_norm2d_1.w_2" shape = [32] dtype = "float32" min_val = float("0") @@ -3924,6 +4290,7 @@ class Program_weight_tensor_parameter_365: class Program_weight_tensor_parameter_366: name = "parameter_366" + original_name = "batch_norm2d_1.w_1" shape = [32] dtype = "float32" min_val = float("0") @@ -3933,6 +4300,7 @@ class Program_weight_tensor_parameter_366: class Program_weight_tensor_parameter_367: name = "parameter_367" + original_name = "conv2d_1.w_0" shape = [32, 32, 3, 3] dtype = "float32" min_val = float("-0.347786") @@ -3944,6 +4312,7 @@ class Program_weight_tensor_parameter_367: class Program_weight_tensor_parameter_368: name = "parameter_368" + original_name = "batch_norm2d_0.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -3953,6 +4322,7 @@ class Program_weight_tensor_parameter_368: class Program_weight_tensor_parameter_369: name = "parameter_369" + original_name = "batch_norm2d_0.w_0" shape = [32] dtype = "float32" min_val = float("0") @@ -3962,6 +4332,7 @@ class Program_weight_tensor_parameter_369: class Program_weight_tensor_parameter_370: name = "parameter_370" + original_name = "batch_norm2d_0.w_2" shape = [32] dtype = "float32" min_val = float("0") @@ -3971,6 +4342,7 @@ class Program_weight_tensor_parameter_370: class Program_weight_tensor_parameter_371: name = "parameter_371" + original_name = "batch_norm2d_0.w_1" shape = [32] dtype = "float32" min_val = float("0") @@ -3980,6 +4352,7 @@ class Program_weight_tensor_parameter_371: class Program_weight_tensor_parameter_372: name = "parameter_372" + original_name = "conv2d_0.w_0" shape = [32, 3, 3, 3] dtype = "float32" min_val = float("-0.317155") diff --git a/paddle_samples/PaddleX/PP-YOLOE_plus_SOD-L/subgraph_6/input_meta.py b/paddle_samples/PaddleX/PP-YOLOE_plus_SOD-L/subgraph_6/input_meta.py index 38a68a034..22d2cd848 100644 --- a/paddle_samples/PaddleX/PP-YOLOE_plus_SOD-L/subgraph_6/input_meta.py +++ b/paddle_samples/PaddleX/PP-YOLOE_plus_SOD-L/subgraph_6/input_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_data_0: name = "data_0" + original_name = "param_52" shape = [1] dtype = "float32" data = [0.699884] @@ -7,6 +8,7 @@ class Program_weight_tensor_data_0: class Program_weight_tensor_data_1: name = "data_1" + original_name = "param_53" shape = [1] dtype = "float32" data = [0.667963] @@ -14,6 +16,7 @@ class Program_weight_tensor_data_1: class Program_weight_tensor_data_2: name = "data_2" + original_name = "param_54" shape = [1] dtype = "float32" data = [0.675792] @@ -21,6 +24,7 @@ class Program_weight_tensor_data_2: class Program_weight_tensor_data_3: name = "data_3" + original_name = "param_55" shape = [1] dtype = "float32" data = [0.676071] @@ -28,6 +32,7 @@ class Program_weight_tensor_data_3: class Program_weight_tensor_data_4: name = "data_4" + original_name = "param_56" shape = [1] dtype = "float32" data = [0.658719] @@ -35,6 +40,7 @@ class Program_weight_tensor_data_4: class Program_weight_tensor_data_5: name = "data_5" + original_name = "param_57" shape = [1] dtype = "float32" data = [0.620637] @@ -42,6 +48,7 @@ class Program_weight_tensor_data_5: class Program_weight_tensor_data_6: name = "data_6" + original_name = "param_58" shape = [1] dtype = "float32" data = [0.637685] @@ -49,6 +56,7 @@ class Program_weight_tensor_data_6: class Program_weight_tensor_data_7: name = "data_7" + original_name = "param_59" shape = [1] dtype = "float32" data = [0.619238] @@ -56,6 +64,7 @@ class Program_weight_tensor_data_7: class Program_weight_tensor_data_8: name = "data_8" + original_name = "param_60" shape = [1] dtype = "float32" data = [0.773168] @@ -63,6 +72,7 @@ class Program_weight_tensor_data_8: class Program_weight_tensor_data_9: name = "data_9" + original_name = "param_61" shape = [1] dtype = "float32" data = [0.635316] @@ -70,6 +80,7 @@ class Program_weight_tensor_data_9: class Program_weight_tensor_data_10: name = "data_10" + original_name = "param_62" shape = [1] dtype = "float32" data = [0.623672] @@ -77,6 +88,7 @@ class Program_weight_tensor_data_10: class Program_weight_tensor_data_11: name = "data_11" + original_name = "param_63" shape = [1] dtype = "float32" data = [0.620323] @@ -84,6 +96,7 @@ class Program_weight_tensor_data_11: class Program_weight_tensor_data_12: name = "data_12" + original_name = "param_64" shape = [1] dtype = "float32" data = [0.621219] @@ -91,6 +104,7 @@ class Program_weight_tensor_data_12: class Program_weight_tensor_data_13: name = "data_13" + original_name = "param_65" shape = [1] dtype = "float32" data = [0.624329] @@ -98,6 +112,7 @@ class Program_weight_tensor_data_13: class Program_weight_tensor_data_14: name = "data_14" + original_name = "param_66" shape = [1] dtype = "float32" data = [0.733117] @@ -105,6 +120,7 @@ class Program_weight_tensor_data_14: class Program_weight_tensor_data_15: name = "data_15" + original_name = "param_67" shape = [1] dtype = "float32" data = [0.557224] @@ -112,6 +128,7 @@ class Program_weight_tensor_data_15: class Program_weight_tensor_data_16: name = "data_16" + original_name = "param_68" shape = [1] dtype = "float32" data = [0.579909] @@ -119,6 +136,7 @@ class Program_weight_tensor_data_16: class Program_weight_tensor_data_17: name = "data_17" + original_name = "param_69" shape = [1] dtype = "float32" data = [0.70327] @@ -126,6 +144,7 @@ class Program_weight_tensor_data_17: class Program_weight_tensor_data_18: name = "data_18" + original_name = "param_70" shape = [1024, 3072] dtype = "float32" min_val = float("-0.0319247") @@ -137,6 +156,7 @@ class Program_weight_tensor_data_18: class Program_weight_tensor_data_19: name = "data_19" + original_name = "param_71" shape = [3072] dtype = "float32" min_val = float("-0.000610453") @@ -148,6 +168,7 @@ class Program_weight_tensor_data_19: class Program_weight_tensor_data_20: name = "data_20" + original_name = "param_72" shape = [1024, 3072] dtype = "float32" min_val = float("-0.0310115") @@ -159,6 +180,7 @@ class Program_weight_tensor_data_20: class Program_weight_tensor_data_21: name = "data_21" + original_name = "param_73" shape = [3072] dtype = "float32" min_val = float("-0.000442001") @@ -170,6 +192,7 @@ class Program_weight_tensor_data_21: class Program_weight_tensor_data_22: name = "data_22" + original_name = "param_74" shape = [1024, 3072] dtype = "float32" min_val = float("-0.0310211") @@ -181,6 +204,7 @@ class Program_weight_tensor_data_22: class Program_weight_tensor_data_23: name = "data_23" + original_name = "param_75" shape = [3072] dtype = "float32" min_val = float("-0.00027446") @@ -192,6 +216,7 @@ class Program_weight_tensor_data_23: class Program_weight_tensor_data_24: name = "data_24" + original_name = "param_76" shape = [1024, 3072] dtype = "float32" min_val = float("-0.030934") @@ -203,6 +228,7 @@ class Program_weight_tensor_data_24: class Program_weight_tensor_data_25: name = "data_25" + original_name = "param_77" shape = [3072] dtype = "float32" min_val = float("-0.000277695") @@ -214,6 +240,7 @@ class Program_weight_tensor_data_25: class Program_weight_tensor_data_26: name = "data_26" + original_name = "var_1780" shape = [2, 3, 416, 416] dtype = "float32" max_val = float("1.0") diff --git a/paddle_samples/PaddleX/PP-YOLOE_plus_SOD-L/subgraph_6/weight_meta.py b/paddle_samples/PaddleX/PP-YOLOE_plus_SOD-L/subgraph_6/weight_meta.py index 5dcac1a4a..1d51e39c3 100644 --- a/paddle_samples/PaddleX/PP-YOLOE_plus_SOD-L/subgraph_6/weight_meta.py +++ b/paddle_samples/PaddleX/PP-YOLOE_plus_SOD-L/subgraph_6/weight_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_parameter_0: name = "parameter_0" + original_name = "batch_norm2d_137.b_0" shape = [768] dtype = "float32" min_val = float("-0.175929") @@ -11,6 +12,7 @@ class Program_weight_tensor_parameter_0: class Program_weight_tensor_parameter_1: name = "parameter_1" + original_name = "batch_norm2d_137.w_0" shape = [768] dtype = "float32" min_val = float("0.939955") @@ -22,6 +24,7 @@ class Program_weight_tensor_parameter_1: class Program_weight_tensor_parameter_2: name = "parameter_2" + original_name = "batch_norm2d_137.w_2" shape = [768] dtype = "float32" min_val = float("0.00130768") @@ -33,6 +36,7 @@ class Program_weight_tensor_parameter_2: class Program_weight_tensor_parameter_3: name = "parameter_3" + original_name = "batch_norm2d_137.w_1" shape = [768] dtype = "float32" min_val = float("-0.127586") @@ -44,6 +48,7 @@ class Program_weight_tensor_parameter_3: class Program_weight_tensor_parameter_4: name = "parameter_4" + original_name = "conv2d_141.w_0" shape = [768, 768, 1, 1] dtype = "float32" min_val = float("-0.0515626") @@ -55,6 +60,7 @@ class Program_weight_tensor_parameter_4: class Program_weight_tensor_parameter_5: name = "parameter_5" + original_name = "batch_norm2d_136.b_0" shape = [384] dtype = "float32" min_val = float("-0.141651") @@ -66,6 +72,7 @@ class Program_weight_tensor_parameter_5: class Program_weight_tensor_parameter_6: name = "parameter_6" + original_name = "batch_norm2d_136.w_0" shape = [384] dtype = "float32" min_val = float("0.945806") @@ -77,6 +84,7 @@ class Program_weight_tensor_parameter_6: class Program_weight_tensor_parameter_7: name = "parameter_7" + original_name = "batch_norm2d_136.w_2" shape = [384] dtype = "float32" min_val = float("0.000827797") @@ -88,6 +96,7 @@ class Program_weight_tensor_parameter_7: class Program_weight_tensor_parameter_8: name = "parameter_8" + original_name = "batch_norm2d_136.w_1" shape = [384] dtype = "float32" min_val = float("-0.0553707") @@ -99,6 +108,7 @@ class Program_weight_tensor_parameter_8: class Program_weight_tensor_parameter_9: name = "parameter_9" + original_name = "conv2d_140.w_0" shape = [384, 384, 1, 1] dtype = "float32" min_val = float("-0.0308331") @@ -110,6 +120,7 @@ class Program_weight_tensor_parameter_9: class Program_weight_tensor_parameter_10: name = "parameter_10" + original_name = "batch_norm2d_135.b_0" shape = [384] dtype = "float32" min_val = float("-0.141651") @@ -121,6 +132,7 @@ class Program_weight_tensor_parameter_10: class Program_weight_tensor_parameter_11: name = "parameter_11" + original_name = "batch_norm2d_135.w_0" shape = [384] dtype = "float32" min_val = float("0.968047") @@ -132,6 +144,7 @@ class Program_weight_tensor_parameter_11: class Program_weight_tensor_parameter_12: name = "parameter_12" + original_name = "batch_norm2d_135.w_2" shape = [384] dtype = "float32" min_val = float("0.00219862") @@ -143,6 +156,7 @@ class Program_weight_tensor_parameter_12: class Program_weight_tensor_parameter_13: name = "parameter_13" + original_name = "batch_norm2d_135.w_1" shape = [384] dtype = "float32" min_val = float("-0.173733") @@ -154,6 +168,7 @@ class Program_weight_tensor_parameter_13: class Program_weight_tensor_parameter_14: name = "parameter_14" + original_name = "conv2d_139.w_0" shape = [384, 384, 3, 3] dtype = "float32" min_val = float("-0.0283367") @@ -165,6 +180,7 @@ class Program_weight_tensor_parameter_14: class Program_weight_tensor_parameter_15: name = "parameter_15" + original_name = "batch_norm2d_134.b_0" shape = [384] dtype = "float32" min_val = float("-0.170186") @@ -176,6 +192,7 @@ class Program_weight_tensor_parameter_15: class Program_weight_tensor_parameter_16: name = "parameter_16" + original_name = "batch_norm2d_134.w_0" shape = [384] dtype = "float32" min_val = float("0.975256") @@ -187,6 +204,7 @@ class Program_weight_tensor_parameter_16: class Program_weight_tensor_parameter_17: name = "parameter_17" + original_name = "batch_norm2d_134.w_2" shape = [384] dtype = "float32" min_val = float("0.00658639") @@ -198,6 +216,7 @@ class Program_weight_tensor_parameter_17: class Program_weight_tensor_parameter_18: name = "parameter_18" + original_name = "batch_norm2d_134.w_1" shape = [384] dtype = "float32" min_val = float("-0.245253") @@ -209,6 +228,7 @@ class Program_weight_tensor_parameter_18: class Program_weight_tensor_parameter_19: name = "parameter_19" + original_name = "conv2d_138.w_0" shape = [384, 384, 3, 3] dtype = "float32" min_val = float("-0.0311751") @@ -220,6 +240,7 @@ class Program_weight_tensor_parameter_19: class Program_weight_tensor_parameter_20: name = "parameter_20" + original_name = "batch_norm2d_133.b_0" shape = [384] dtype = "float32" min_val = float("-0.105187") @@ -231,6 +252,7 @@ class Program_weight_tensor_parameter_20: class Program_weight_tensor_parameter_21: name = "parameter_21" + original_name = "batch_norm2d_133.w_0" shape = [384] dtype = "float32" min_val = float("0.945568") @@ -242,6 +264,7 @@ class Program_weight_tensor_parameter_21: class Program_weight_tensor_parameter_22: name = "parameter_22" + original_name = "batch_norm2d_133.w_2" shape = [384] dtype = "float32" min_val = float("0.000652651") @@ -253,6 +276,7 @@ class Program_weight_tensor_parameter_22: class Program_weight_tensor_parameter_23: name = "parameter_23" + original_name = "batch_norm2d_133.w_1" shape = [384] dtype = "float32" min_val = float("-0.077936") @@ -264,6 +288,7 @@ class Program_weight_tensor_parameter_23: class Program_weight_tensor_parameter_24: name = "parameter_24" + original_name = "conv2d_137.w_0" shape = [384, 384, 1, 1] dtype = "float32" min_val = float("-0.0260426") @@ -275,6 +300,7 @@ class Program_weight_tensor_parameter_24: class Program_weight_tensor_parameter_25: name = "parameter_25" + original_name = "batch_norm2d_132.b_0" shape = [384] dtype = "float32" min_val = float("-0.105187") @@ -286,6 +312,7 @@ class Program_weight_tensor_parameter_25: class Program_weight_tensor_parameter_26: name = "parameter_26" + original_name = "batch_norm2d_132.w_0" shape = [384] dtype = "float32" min_val = float("0.959552") @@ -297,6 +324,7 @@ class Program_weight_tensor_parameter_26: class Program_weight_tensor_parameter_27: name = "parameter_27" + original_name = "batch_norm2d_132.w_2" shape = [384] dtype = "float32" min_val = float("0.0024964") @@ -308,6 +336,7 @@ class Program_weight_tensor_parameter_27: class Program_weight_tensor_parameter_28: name = "parameter_28" + original_name = "batch_norm2d_132.w_1" shape = [384] dtype = "float32" min_val = float("-0.160103") @@ -319,6 +348,7 @@ class Program_weight_tensor_parameter_28: class Program_weight_tensor_parameter_29: name = "parameter_29" + original_name = "conv2d_136.w_0" shape = [384, 384, 3, 3] dtype = "float32" min_val = float("-0.034241") @@ -330,6 +360,7 @@ class Program_weight_tensor_parameter_29: class Program_weight_tensor_parameter_30: name = "parameter_30" + original_name = "batch_norm2d_131.b_0" shape = [384] dtype = "float32" min_val = float("-0.0896542") @@ -341,6 +372,7 @@ class Program_weight_tensor_parameter_30: class Program_weight_tensor_parameter_31: name = "parameter_31" + original_name = "batch_norm2d_131.w_0" shape = [384] dtype = "float32" min_val = float("0.933175") @@ -352,6 +384,7 @@ class Program_weight_tensor_parameter_31: class Program_weight_tensor_parameter_32: name = "parameter_32" + original_name = "batch_norm2d_131.w_2" shape = [384] dtype = "float32" min_val = float("0.00448167") @@ -363,6 +396,7 @@ class Program_weight_tensor_parameter_32: class Program_weight_tensor_parameter_33: name = "parameter_33" + original_name = "batch_norm2d_131.w_1" shape = [384] dtype = "float32" min_val = float("-0.168694") @@ -374,6 +408,7 @@ class Program_weight_tensor_parameter_33: class Program_weight_tensor_parameter_34: name = "parameter_34" + original_name = "conv2d_135.w_0" shape = [384, 384, 3, 3] dtype = "float32" min_val = float("-0.0389109") @@ -385,6 +420,7 @@ class Program_weight_tensor_parameter_34: class Program_weight_tensor_parameter_35: name = "parameter_35" + original_name = "batch_norm2d_130.b_0" shape = [384] dtype = "float32" min_val = float("-0.116304") @@ -396,6 +432,7 @@ class Program_weight_tensor_parameter_35: class Program_weight_tensor_parameter_36: name = "parameter_36" + original_name = "batch_norm2d_130.w_0" shape = [384] dtype = "float32" min_val = float("0.929317") @@ -407,6 +444,7 @@ class Program_weight_tensor_parameter_36: class Program_weight_tensor_parameter_37: name = "parameter_37" + original_name = "batch_norm2d_130.w_2" shape = [384] dtype = "float32" min_val = float("0.00119145") @@ -418,6 +456,7 @@ class Program_weight_tensor_parameter_37: class Program_weight_tensor_parameter_38: name = "parameter_38" + original_name = "batch_norm2d_130.w_1" shape = [384] dtype = "float32" min_val = float("-0.0535005") @@ -429,6 +468,7 @@ class Program_weight_tensor_parameter_38: class Program_weight_tensor_parameter_39: name = "parameter_39" + original_name = "conv2d_134.w_0" shape = [384, 384, 1, 1] dtype = "float32" min_val = float("-0.037048") @@ -440,6 +480,7 @@ class Program_weight_tensor_parameter_39: class Program_weight_tensor_parameter_40: name = "parameter_40" + original_name = "batch_norm2d_129.b_0" shape = [384] dtype = "float32" min_val = float("-0.116304") @@ -451,6 +492,7 @@ class Program_weight_tensor_parameter_40: class Program_weight_tensor_parameter_41: name = "parameter_41" + original_name = "batch_norm2d_129.w_0" shape = [384] dtype = "float32" min_val = float("0.98123") @@ -462,6 +504,7 @@ class Program_weight_tensor_parameter_41: class Program_weight_tensor_parameter_42: name = "parameter_42" + original_name = "batch_norm2d_129.w_2" shape = [384] dtype = "float32" min_val = float("0.00438825") @@ -473,6 +516,7 @@ class Program_weight_tensor_parameter_42: class Program_weight_tensor_parameter_43: name = "parameter_43" + original_name = "batch_norm2d_129.w_1" shape = [384] dtype = "float32" min_val = float("-0.158148") @@ -484,6 +528,7 @@ class Program_weight_tensor_parameter_43: class Program_weight_tensor_parameter_44: name = "parameter_44" + original_name = "conv2d_133.w_0" shape = [384, 384, 3, 3] dtype = "float32" min_val = float("-0.0342558") @@ -495,6 +540,7 @@ class Program_weight_tensor_parameter_44: class Program_weight_tensor_parameter_45: name = "parameter_45" + original_name = "batch_norm2d_128.b_0" shape = [384] dtype = "float32" min_val = float("-0.10708") @@ -506,6 +552,7 @@ class Program_weight_tensor_parameter_45: class Program_weight_tensor_parameter_46: name = "parameter_46" + original_name = "batch_norm2d_128.w_0" shape = [384] dtype = "float32" min_val = float("0.944782") @@ -517,6 +564,7 @@ class Program_weight_tensor_parameter_46: class Program_weight_tensor_parameter_47: name = "parameter_47" + original_name = "batch_norm2d_128.w_2" shape = [384] dtype = "float32" min_val = float("0.00510026") @@ -528,6 +576,7 @@ class Program_weight_tensor_parameter_47: class Program_weight_tensor_parameter_48: name = "parameter_48" + original_name = "batch_norm2d_128.w_1" shape = [384] dtype = "float32" min_val = float("-0.145833") @@ -539,6 +588,7 @@ class Program_weight_tensor_parameter_48: class Program_weight_tensor_parameter_49: name = "parameter_49" + original_name = "conv2d_132.w_0" shape = [384, 384, 3, 3] dtype = "float32" min_val = float("-0.0265761") @@ -550,6 +600,7 @@ class Program_weight_tensor_parameter_49: class Program_weight_tensor_parameter_50: name = "parameter_50" + original_name = "batch_norm2d_127.b_0" shape = [384] dtype = "float32" min_val = float("-0.106796") @@ -561,6 +612,7 @@ class Program_weight_tensor_parameter_50: class Program_weight_tensor_parameter_51: name = "parameter_51" + original_name = "batch_norm2d_127.w_0" shape = [384] dtype = "float32" min_val = float("0.973685") @@ -572,6 +624,7 @@ class Program_weight_tensor_parameter_51: class Program_weight_tensor_parameter_52: name = "parameter_52" + original_name = "batch_norm2d_127.w_2" shape = [384] dtype = "float32" min_val = float("0.00211342") @@ -583,6 +636,7 @@ class Program_weight_tensor_parameter_52: class Program_weight_tensor_parameter_53: name = "parameter_53" + original_name = "batch_norm2d_127.w_1" shape = [384] dtype = "float32" min_val = float("-0.0938724") @@ -594,6 +648,7 @@ class Program_weight_tensor_parameter_53: class Program_weight_tensor_parameter_54: name = "parameter_54" + original_name = "conv2d_131.w_0" shape = [384, 1152, 1, 1] dtype = "float32" min_val = float("-0.0600923") @@ -605,6 +660,7 @@ class Program_weight_tensor_parameter_54: class Program_weight_tensor_parameter_55: name = "parameter_55" + original_name = "batch_norm2d_126.b_0" shape = [384] dtype = "float32" min_val = float("-0.0425267") @@ -616,6 +672,7 @@ class Program_weight_tensor_parameter_55: class Program_weight_tensor_parameter_56: name = "parameter_56" + original_name = "batch_norm2d_126.w_0" shape = [384] dtype = "float32" min_val = float("0.959381") @@ -627,6 +684,7 @@ class Program_weight_tensor_parameter_56: class Program_weight_tensor_parameter_57: name = "parameter_57" + original_name = "batch_norm2d_126.w_2" shape = [384] dtype = "float32" min_val = float("0.00177107") @@ -638,6 +696,7 @@ class Program_weight_tensor_parameter_57: class Program_weight_tensor_parameter_58: name = "parameter_58" + original_name = "batch_norm2d_126.w_1" shape = [384] dtype = "float32" min_val = float("-0.101028") @@ -649,6 +708,7 @@ class Program_weight_tensor_parameter_58: class Program_weight_tensor_parameter_59: name = "parameter_59" + original_name = "conv2d_130.w_0" shape = [384, 1152, 1, 1] dtype = "float32" min_val = float("-0.0252393") @@ -660,6 +720,7 @@ class Program_weight_tensor_parameter_59: class Program_weight_tensor_parameter_60: name = "parameter_60" + original_name = "batch_norm2d_125.b_0" shape = [384] dtype = "float32" min_val = float("-0.0530202") @@ -671,6 +732,7 @@ class Program_weight_tensor_parameter_60: class Program_weight_tensor_parameter_61: name = "parameter_61" + original_name = "batch_norm2d_125.w_0" shape = [384] dtype = "float32" min_val = float("0.988638") @@ -682,6 +744,7 @@ class Program_weight_tensor_parameter_61: class Program_weight_tensor_parameter_62: name = "parameter_62" + original_name = "batch_norm2d_125.w_2" shape = [384] dtype = "float32" min_val = float("0.00403058") @@ -693,6 +756,7 @@ class Program_weight_tensor_parameter_62: class Program_weight_tensor_parameter_63: name = "parameter_63" + original_name = "batch_norm2d_125.w_1" shape = [384] dtype = "float32" min_val = float("-0.365522") @@ -704,6 +768,7 @@ class Program_weight_tensor_parameter_63: class Program_weight_tensor_parameter_64: name = "parameter_64" + original_name = "conv2d_129.w_0" shape = [384, 384, 3, 3] dtype = "float32" min_val = float("-0.0201867") @@ -715,6 +780,7 @@ class Program_weight_tensor_parameter_64: class Program_weight_tensor_parameter_65: name = "parameter_65" + original_name = "batch_norm2d_124.b_0" shape = [384] dtype = "float32" min_val = float("-0.22273") @@ -726,6 +792,7 @@ class Program_weight_tensor_parameter_65: class Program_weight_tensor_parameter_66: name = "parameter_66" + original_name = "batch_norm2d_124.w_0" shape = [384] dtype = "float32" min_val = float("0.919294") @@ -737,6 +804,7 @@ class Program_weight_tensor_parameter_66: class Program_weight_tensor_parameter_67: name = "parameter_67" + original_name = "batch_norm2d_124.w_2" shape = [384] dtype = "float32" min_val = float("0.00389442") @@ -748,6 +816,7 @@ class Program_weight_tensor_parameter_67: class Program_weight_tensor_parameter_68: name = "parameter_68" + original_name = "batch_norm2d_124.w_1" shape = [384] dtype = "float32" min_val = float("-0.153266") @@ -759,6 +828,7 @@ class Program_weight_tensor_parameter_68: class Program_weight_tensor_parameter_69: name = "parameter_69" + original_name = "conv2d_128.w_0" shape = [384, 384, 1, 1] dtype = "float32" min_val = float("-0.0888812") @@ -770,6 +840,7 @@ class Program_weight_tensor_parameter_69: class Program_weight_tensor_parameter_70: name = "parameter_70" + original_name = "batch_norm2d_123.b_0" shape = [192] dtype = "float32" min_val = float("-0.166124") @@ -781,6 +852,7 @@ class Program_weight_tensor_parameter_70: class Program_weight_tensor_parameter_71: name = "parameter_71" + original_name = "batch_norm2d_123.w_0" shape = [192] dtype = "float32" min_val = float("0.84107") @@ -792,6 +864,7 @@ class Program_weight_tensor_parameter_71: class Program_weight_tensor_parameter_72: name = "parameter_72" + original_name = "batch_norm2d_123.w_2" shape = [192] dtype = "float32" min_val = float("0.00162489") @@ -803,6 +876,7 @@ class Program_weight_tensor_parameter_72: class Program_weight_tensor_parameter_73: name = "parameter_73" + original_name = "batch_norm2d_123.w_1" shape = [192] dtype = "float32" min_val = float("-0.0629501") @@ -814,6 +888,7 @@ class Program_weight_tensor_parameter_73: class Program_weight_tensor_parameter_74: name = "parameter_74" + original_name = "conv2d_127.w_0" shape = [192, 192, 1, 1] dtype = "float32" min_val = float("-0.0448536") @@ -825,6 +900,7 @@ class Program_weight_tensor_parameter_74: class Program_weight_tensor_parameter_75: name = "parameter_75" + original_name = "batch_norm2d_122.b_0" shape = [192] dtype = "float32" min_val = float("-0.166124") @@ -836,6 +912,7 @@ class Program_weight_tensor_parameter_75: class Program_weight_tensor_parameter_76: name = "parameter_76" + original_name = "batch_norm2d_122.w_0" shape = [192] dtype = "float32" min_val = float("0.729345") @@ -847,6 +924,7 @@ class Program_weight_tensor_parameter_76: class Program_weight_tensor_parameter_77: name = "parameter_77" + original_name = "batch_norm2d_122.w_2" shape = [192] dtype = "float32" min_val = float("0.00562341") @@ -858,6 +936,7 @@ class Program_weight_tensor_parameter_77: class Program_weight_tensor_parameter_78: name = "parameter_78" + original_name = "batch_norm2d_122.w_1" shape = [192] dtype = "float32" min_val = float("-0.191415") @@ -869,6 +948,7 @@ class Program_weight_tensor_parameter_78: class Program_weight_tensor_parameter_79: name = "parameter_79" + original_name = "conv2d_126.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.0346474") @@ -880,6 +960,7 @@ class Program_weight_tensor_parameter_79: class Program_weight_tensor_parameter_80: name = "parameter_80" + original_name = "batch_norm2d_121.b_0" shape = [192] dtype = "float32" min_val = float("-0.191424") @@ -891,6 +972,7 @@ class Program_weight_tensor_parameter_80: class Program_weight_tensor_parameter_81: name = "parameter_81" + original_name = "batch_norm2d_121.w_0" shape = [192] dtype = "float32" min_val = float("0.897189") @@ -902,6 +984,7 @@ class Program_weight_tensor_parameter_81: class Program_weight_tensor_parameter_82: name = "parameter_82" + original_name = "batch_norm2d_121.w_2" shape = [192] dtype = "float32" min_val = float("0.0130131") @@ -913,6 +996,7 @@ class Program_weight_tensor_parameter_82: class Program_weight_tensor_parameter_83: name = "parameter_83" + original_name = "batch_norm2d_121.w_1" shape = [192] dtype = "float32" min_val = float("-0.322411") @@ -924,6 +1008,7 @@ class Program_weight_tensor_parameter_83: class Program_weight_tensor_parameter_84: name = "parameter_84" + original_name = "conv2d_125.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.0436326") @@ -935,6 +1020,7 @@ class Program_weight_tensor_parameter_84: class Program_weight_tensor_parameter_85: name = "parameter_85" + original_name = "batch_norm2d_120.b_0" shape = [192] dtype = "float32" min_val = float("-0.191731") @@ -946,6 +1032,7 @@ class Program_weight_tensor_parameter_85: class Program_weight_tensor_parameter_86: name = "parameter_86" + original_name = "batch_norm2d_120.w_0" shape = [192] dtype = "float32" min_val = float("0.922072") @@ -957,6 +1044,7 @@ class Program_weight_tensor_parameter_86: class Program_weight_tensor_parameter_87: name = "parameter_87" + original_name = "batch_norm2d_120.w_2" shape = [192] dtype = "float32" min_val = float("0.00120452") @@ -968,6 +1056,7 @@ class Program_weight_tensor_parameter_87: class Program_weight_tensor_parameter_88: name = "parameter_88" + original_name = "batch_norm2d_120.w_1" shape = [192] dtype = "float32" min_val = float("-0.0561207") @@ -979,6 +1068,7 @@ class Program_weight_tensor_parameter_88: class Program_weight_tensor_parameter_89: name = "parameter_89" + original_name = "conv2d_124.w_0" shape = [192, 192, 1, 1] dtype = "float32" min_val = float("-0.0389677") @@ -990,6 +1080,7 @@ class Program_weight_tensor_parameter_89: class Program_weight_tensor_parameter_90: name = "parameter_90" + original_name = "batch_norm2d_119.b_0" shape = [192] dtype = "float32" min_val = float("-0.191731") @@ -1001,6 +1092,7 @@ class Program_weight_tensor_parameter_90: class Program_weight_tensor_parameter_91: name = "parameter_91" + original_name = "batch_norm2d_119.w_0" shape = [192] dtype = "float32" min_val = float("0.967917") @@ -1012,6 +1104,7 @@ class Program_weight_tensor_parameter_91: class Program_weight_tensor_parameter_92: name = "parameter_92" + original_name = "batch_norm2d_119.w_2" shape = [192] dtype = "float32" min_val = float("0.00374803") @@ -1023,6 +1116,7 @@ class Program_weight_tensor_parameter_92: class Program_weight_tensor_parameter_93: name = "parameter_93" + original_name = "batch_norm2d_119.w_1" shape = [192] dtype = "float32" min_val = float("-0.141797") @@ -1034,6 +1128,7 @@ class Program_weight_tensor_parameter_93: class Program_weight_tensor_parameter_94: name = "parameter_94" + original_name = "conv2d_123.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.0433361") @@ -1045,6 +1140,7 @@ class Program_weight_tensor_parameter_94: class Program_weight_tensor_parameter_95: name = "parameter_95" + original_name = "batch_norm2d_118.b_0" shape = [192] dtype = "float32" min_val = float("-0.188937") @@ -1056,6 +1152,7 @@ class Program_weight_tensor_parameter_95: class Program_weight_tensor_parameter_96: name = "parameter_96" + original_name = "batch_norm2d_118.w_0" shape = [192] dtype = "float32" min_val = float("0.88236") @@ -1067,6 +1164,7 @@ class Program_weight_tensor_parameter_96: class Program_weight_tensor_parameter_97: name = "parameter_97" + original_name = "batch_norm2d_118.w_2" shape = [192] dtype = "float32" min_val = float("0.00731951") @@ -1078,6 +1176,7 @@ class Program_weight_tensor_parameter_97: class Program_weight_tensor_parameter_98: name = "parameter_98" + original_name = "batch_norm2d_118.w_1" shape = [192] dtype = "float32" min_val = float("-0.105024") @@ -1089,6 +1188,7 @@ class Program_weight_tensor_parameter_98: class Program_weight_tensor_parameter_99: name = "parameter_99" + original_name = "conv2d_122.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.0453344") @@ -1100,6 +1200,7 @@ class Program_weight_tensor_parameter_99: class Program_weight_tensor_parameter_100: name = "parameter_100" + original_name = "batch_norm2d_117.b_0" shape = [192] dtype = "float32" min_val = float("-0.229338") @@ -1111,6 +1212,7 @@ class Program_weight_tensor_parameter_100: class Program_weight_tensor_parameter_101: name = "parameter_101" + original_name = "batch_norm2d_117.w_0" shape = [192] dtype = "float32" min_val = float("0.900655") @@ -1122,6 +1224,7 @@ class Program_weight_tensor_parameter_101: class Program_weight_tensor_parameter_102: name = "parameter_102" + original_name = "batch_norm2d_117.w_2" shape = [192] dtype = "float32" min_val = float("0.00158349") @@ -1133,6 +1236,7 @@ class Program_weight_tensor_parameter_102: class Program_weight_tensor_parameter_103: name = "parameter_103" + original_name = "batch_norm2d_117.w_1" shape = [192] dtype = "float32" min_val = float("-0.0350209") @@ -1144,6 +1248,7 @@ class Program_weight_tensor_parameter_103: class Program_weight_tensor_parameter_104: name = "parameter_104" + original_name = "conv2d_121.w_0" shape = [192, 192, 1, 1] dtype = "float32" min_val = float("-0.0414024") @@ -1155,6 +1260,7 @@ class Program_weight_tensor_parameter_104: class Program_weight_tensor_parameter_105: name = "parameter_105" + original_name = "batch_norm2d_116.b_0" shape = [192] dtype = "float32" min_val = float("-0.229338") @@ -1166,6 +1272,7 @@ class Program_weight_tensor_parameter_105: class Program_weight_tensor_parameter_106: name = "parameter_106" + original_name = "batch_norm2d_116.w_0" shape = [192] dtype = "float32" min_val = float("0.947228") @@ -1177,6 +1284,7 @@ class Program_weight_tensor_parameter_106: class Program_weight_tensor_parameter_107: name = "parameter_107" + original_name = "batch_norm2d_116.w_2" shape = [192] dtype = "float32" min_val = float("0.00636548") @@ -1188,6 +1296,7 @@ class Program_weight_tensor_parameter_107: class Program_weight_tensor_parameter_108: name = "parameter_108" + original_name = "batch_norm2d_116.w_1" shape = [192] dtype = "float32" min_val = float("-0.12433") @@ -1199,6 +1308,7 @@ class Program_weight_tensor_parameter_108: class Program_weight_tensor_parameter_109: name = "parameter_109" + original_name = "conv2d_120.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.0444124") @@ -1210,6 +1320,7 @@ class Program_weight_tensor_parameter_109: class Program_weight_tensor_parameter_110: name = "parameter_110" + original_name = "batch_norm2d_115.b_0" shape = [192] dtype = "float32" min_val = float("-0.234043") @@ -1221,6 +1332,7 @@ class Program_weight_tensor_parameter_110: class Program_weight_tensor_parameter_111: name = "parameter_111" + original_name = "batch_norm2d_115.w_0" shape = [192] dtype = "float32" min_val = float("0.886425") @@ -1232,6 +1344,7 @@ class Program_weight_tensor_parameter_111: class Program_weight_tensor_parameter_112: name = "parameter_112" + original_name = "batch_norm2d_115.w_2" shape = [192] dtype = "float32" min_val = float("0.00720534") @@ -1243,6 +1356,7 @@ class Program_weight_tensor_parameter_112: class Program_weight_tensor_parameter_113: name = "parameter_113" + original_name = "batch_norm2d_115.w_1" shape = [192] dtype = "float32" min_val = float("-0.162095") @@ -1254,6 +1368,7 @@ class Program_weight_tensor_parameter_113: class Program_weight_tensor_parameter_114: name = "parameter_114" + original_name = "conv2d_119.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.0388947") @@ -1265,6 +1380,7 @@ class Program_weight_tensor_parameter_114: class Program_weight_tensor_parameter_115: name = "parameter_115" + original_name = "batch_norm2d_114.b_0" shape = [192] dtype = "float32" min_val = float("-0.200195") @@ -1276,6 +1392,7 @@ class Program_weight_tensor_parameter_115: class Program_weight_tensor_parameter_116: name = "parameter_116" + original_name = "batch_norm2d_114.w_0" shape = [192] dtype = "float32" min_val = float("0.925347") @@ -1287,6 +1404,7 @@ class Program_weight_tensor_parameter_116: class Program_weight_tensor_parameter_117: name = "parameter_117" + original_name = "batch_norm2d_114.w_2" shape = [192] dtype = "float32" min_val = float("0.00390106") @@ -1298,6 +1416,7 @@ class Program_weight_tensor_parameter_117: class Program_weight_tensor_parameter_118: name = "parameter_118" + original_name = "batch_norm2d_114.w_1" shape = [192] dtype = "float32" min_val = float("-0.0919967") @@ -1309,6 +1428,7 @@ class Program_weight_tensor_parameter_118: class Program_weight_tensor_parameter_119: name = "parameter_119" + original_name = "conv2d_118.w_0" shape = [192, 576, 1, 1] dtype = "float32" min_val = float("-0.0594804") @@ -1320,6 +1440,7 @@ class Program_weight_tensor_parameter_119: class Program_weight_tensor_parameter_120: name = "parameter_120" + original_name = "batch_norm2d_113.b_0" shape = [192] dtype = "float32" min_val = float("-0.0998406") @@ -1331,6 +1452,7 @@ class Program_weight_tensor_parameter_120: class Program_weight_tensor_parameter_121: name = "parameter_121" + original_name = "batch_norm2d_113.w_0" shape = [192] dtype = "float32" min_val = float("0.92253") @@ -1342,6 +1464,7 @@ class Program_weight_tensor_parameter_121: class Program_weight_tensor_parameter_122: name = "parameter_122" + original_name = "batch_norm2d_113.w_2" shape = [192] dtype = "float32" min_val = float("0.00364414") @@ -1353,6 +1476,7 @@ class Program_weight_tensor_parameter_122: class Program_weight_tensor_parameter_123: name = "parameter_123" + original_name = "batch_norm2d_113.w_1" shape = [192] dtype = "float32" min_val = float("-0.0792517") @@ -1364,6 +1488,7 @@ class Program_weight_tensor_parameter_123: class Program_weight_tensor_parameter_124: name = "parameter_124" + original_name = "conv2d_117.w_0" shape = [192, 576, 1, 1] dtype = "float32" min_val = float("-0.0668912") @@ -1375,6 +1500,7 @@ class Program_weight_tensor_parameter_124: class Program_weight_tensor_parameter_125: name = "parameter_125" + original_name = "batch_norm2d_112.b_0" shape = [192] dtype = "float32" min_val = float("-0.159157") @@ -1386,6 +1512,7 @@ class Program_weight_tensor_parameter_125: class Program_weight_tensor_parameter_126: name = "parameter_126" + original_name = "batch_norm2d_112.w_0" shape = [192] dtype = "float32" min_val = float("0.921604") @@ -1397,6 +1524,7 @@ class Program_weight_tensor_parameter_126: class Program_weight_tensor_parameter_127: name = "parameter_127" + original_name = "batch_norm2d_112.w_2" shape = [192] dtype = "float32" min_val = float("0.00642947") @@ -1408,6 +1536,7 @@ class Program_weight_tensor_parameter_127: class Program_weight_tensor_parameter_128: name = "parameter_128" + original_name = "batch_norm2d_112.w_1" shape = [192] dtype = "float32" min_val = float("-0.330766") @@ -1419,6 +1548,7 @@ class Program_weight_tensor_parameter_128: class Program_weight_tensor_parameter_129: name = "parameter_129" + original_name = "conv2d_116.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.0500862") @@ -1430,6 +1560,7 @@ class Program_weight_tensor_parameter_129: class Program_weight_tensor_parameter_130: name = "parameter_130" + original_name = "batch_norm2d_111.b_0" shape = [192] dtype = "float32" min_val = float("-0.554317") @@ -1441,6 +1572,7 @@ class Program_weight_tensor_parameter_130: class Program_weight_tensor_parameter_131: name = "parameter_131" + original_name = "batch_norm2d_111.w_0" shape = [192] dtype = "float32" min_val = float("0.546484") @@ -1452,6 +1584,7 @@ class Program_weight_tensor_parameter_131: class Program_weight_tensor_parameter_132: name = "parameter_132" + original_name = "batch_norm2d_111.w_2" shape = [192] dtype = "float32" min_val = float("0.0100657") @@ -1463,6 +1596,7 @@ class Program_weight_tensor_parameter_132: class Program_weight_tensor_parameter_133: name = "parameter_133" + original_name = "batch_norm2d_111.w_1" shape = [192] dtype = "float32" min_val = float("-0.377989") @@ -1474,6 +1608,7 @@ class Program_weight_tensor_parameter_133: class Program_weight_tensor_parameter_134: name = "parameter_134" + original_name = "conv2d_115.w_0" shape = [192, 192, 1, 1] dtype = "float32" min_val = float("-0.153545") @@ -1485,6 +1620,7 @@ class Program_weight_tensor_parameter_134: class Program_weight_tensor_parameter_135: name = "parameter_135" + original_name = "batch_norm2d_110.b_0" shape = [96] dtype = "float32" min_val = float("-0.457832") @@ -1496,6 +1632,7 @@ class Program_weight_tensor_parameter_135: class Program_weight_tensor_parameter_136: name = "parameter_136" + original_name = "batch_norm2d_110.w_0" shape = [96] dtype = "float32" min_val = float("0.761673") @@ -1507,6 +1644,7 @@ class Program_weight_tensor_parameter_136: class Program_weight_tensor_parameter_137: name = "parameter_137" + original_name = "batch_norm2d_110.w_2" shape = [96] dtype = "float32" min_val = float("0.00289003") @@ -1518,6 +1656,7 @@ class Program_weight_tensor_parameter_137: class Program_weight_tensor_parameter_138: name = "parameter_138" + original_name = "batch_norm2d_110.w_1" shape = [96] dtype = "float32" min_val = float("-0.0823141") @@ -1529,6 +1668,7 @@ class Program_weight_tensor_parameter_138: class Program_weight_tensor_parameter_139: name = "parameter_139" + original_name = "conv2d_114.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.0870691") @@ -1540,6 +1680,7 @@ class Program_weight_tensor_parameter_139: class Program_weight_tensor_parameter_140: name = "parameter_140" + original_name = "batch_norm2d_109.b_0" shape = [96] dtype = "float32" min_val = float("-0.457832") @@ -1551,6 +1692,7 @@ class Program_weight_tensor_parameter_140: class Program_weight_tensor_parameter_141: name = "parameter_141" + original_name = "batch_norm2d_109.w_0" shape = [96] dtype = "float32" min_val = float("0.507912") @@ -1562,6 +1704,7 @@ class Program_weight_tensor_parameter_141: class Program_weight_tensor_parameter_142: name = "parameter_142" + original_name = "batch_norm2d_109.w_2" shape = [96] dtype = "float32" min_val = float("0.0126569") @@ -1573,6 +1716,7 @@ class Program_weight_tensor_parameter_142: class Program_weight_tensor_parameter_143: name = "parameter_143" + original_name = "batch_norm2d_109.w_1" shape = [96] dtype = "float32" min_val = float("-0.311825") @@ -1584,6 +1728,7 @@ class Program_weight_tensor_parameter_143: class Program_weight_tensor_parameter_144: name = "parameter_144" + original_name = "conv2d_113.w_0" shape = [96, 96, 3, 3] dtype = "float32" min_val = float("-0.0902007") @@ -1595,6 +1740,7 @@ class Program_weight_tensor_parameter_144: class Program_weight_tensor_parameter_145: name = "parameter_145" + original_name = "batch_norm2d_108.b_0" shape = [96] dtype = "float32" min_val = float("-0.702428") @@ -1606,6 +1752,7 @@ class Program_weight_tensor_parameter_145: class Program_weight_tensor_parameter_146: name = "parameter_146" + original_name = "batch_norm2d_108.w_0" shape = [96] dtype = "float32" min_val = float("0.718272") @@ -1617,6 +1764,7 @@ class Program_weight_tensor_parameter_146: class Program_weight_tensor_parameter_147: name = "parameter_147" + original_name = "batch_norm2d_108.w_2" shape = [96] dtype = "float32" min_val = float("0.0158852") @@ -1628,6 +1776,7 @@ class Program_weight_tensor_parameter_147: class Program_weight_tensor_parameter_148: name = "parameter_148" + original_name = "batch_norm2d_108.w_1" shape = [96] dtype = "float32" min_val = float("-0.218411") @@ -1639,6 +1788,7 @@ class Program_weight_tensor_parameter_148: class Program_weight_tensor_parameter_149: name = "parameter_149" + original_name = "conv2d_112.w_0" shape = [96, 96, 3, 3] dtype = "float32" min_val = float("-0.128433") @@ -1650,6 +1800,7 @@ class Program_weight_tensor_parameter_149: class Program_weight_tensor_parameter_150: name = "parameter_150" + original_name = "batch_norm2d_107.b_0" shape = [96] dtype = "float32" min_val = float("-0.365465") @@ -1661,6 +1812,7 @@ class Program_weight_tensor_parameter_150: class Program_weight_tensor_parameter_151: name = "parameter_151" + original_name = "batch_norm2d_107.w_0" shape = [96] dtype = "float32" min_val = float("0.628614") @@ -1672,6 +1824,7 @@ class Program_weight_tensor_parameter_151: class Program_weight_tensor_parameter_152: name = "parameter_152" + original_name = "batch_norm2d_107.w_2" shape = [96] dtype = "float32" min_val = float("0.00386648") @@ -1683,6 +1836,7 @@ class Program_weight_tensor_parameter_152: class Program_weight_tensor_parameter_153: name = "parameter_153" + original_name = "batch_norm2d_107.w_1" shape = [96] dtype = "float32" min_val = float("-0.0829535") @@ -1694,6 +1848,7 @@ class Program_weight_tensor_parameter_153: class Program_weight_tensor_parameter_154: name = "parameter_154" + original_name = "conv2d_111.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.0826433") @@ -1705,6 +1860,7 @@ class Program_weight_tensor_parameter_154: class Program_weight_tensor_parameter_155: name = "parameter_155" + original_name = "batch_norm2d_106.b_0" shape = [96] dtype = "float32" min_val = float("-0.365465") @@ -1716,6 +1872,7 @@ class Program_weight_tensor_parameter_155: class Program_weight_tensor_parameter_156: name = "parameter_156" + original_name = "batch_norm2d_106.w_0" shape = [96] dtype = "float32" min_val = float("0.808147") @@ -1727,6 +1884,7 @@ class Program_weight_tensor_parameter_156: class Program_weight_tensor_parameter_157: name = "parameter_157" + original_name = "batch_norm2d_106.w_2" shape = [96] dtype = "float32" min_val = float("0.0105243") @@ -1738,6 +1896,7 @@ class Program_weight_tensor_parameter_157: class Program_weight_tensor_parameter_158: name = "parameter_158" + original_name = "batch_norm2d_106.w_1" shape = [96] dtype = "float32" min_val = float("-0.202738") @@ -1749,6 +1908,7 @@ class Program_weight_tensor_parameter_158: class Program_weight_tensor_parameter_159: name = "parameter_159" + original_name = "conv2d_110.w_0" shape = [96, 96, 3, 3] dtype = "float32" min_val = float("-0.0790843") @@ -1760,6 +1920,7 @@ class Program_weight_tensor_parameter_159: class Program_weight_tensor_parameter_160: name = "parameter_160" + original_name = "batch_norm2d_105.b_0" shape = [96] dtype = "float32" min_val = float("-0.488696") @@ -1771,6 +1932,7 @@ class Program_weight_tensor_parameter_160: class Program_weight_tensor_parameter_161: name = "parameter_161" + original_name = "batch_norm2d_105.w_0" shape = [96] dtype = "float32" min_val = float("0.775323") @@ -1782,6 +1944,7 @@ class Program_weight_tensor_parameter_161: class Program_weight_tensor_parameter_162: name = "parameter_162" + original_name = "batch_norm2d_105.w_2" shape = [96] dtype = "float32" min_val = float("0.0130451") @@ -1793,6 +1956,7 @@ class Program_weight_tensor_parameter_162: class Program_weight_tensor_parameter_163: name = "parameter_163" + original_name = "batch_norm2d_105.w_1" shape = [96] dtype = "float32" min_val = float("-0.186613") @@ -1804,6 +1968,7 @@ class Program_weight_tensor_parameter_163: class Program_weight_tensor_parameter_164: name = "parameter_164" + original_name = "conv2d_109.w_0" shape = [96, 96, 3, 3] dtype = "float32" min_val = float("-0.128553") @@ -1815,6 +1980,7 @@ class Program_weight_tensor_parameter_164: class Program_weight_tensor_parameter_165: name = "parameter_165" + original_name = "batch_norm2d_104.b_0" shape = [96] dtype = "float32" min_val = float("-0.492552") @@ -1826,6 +1992,7 @@ class Program_weight_tensor_parameter_165: class Program_weight_tensor_parameter_166: name = "parameter_166" + original_name = "batch_norm2d_104.w_0" shape = [96] dtype = "float32" min_val = float("0.725079") @@ -1837,6 +2004,7 @@ class Program_weight_tensor_parameter_166: class Program_weight_tensor_parameter_167: name = "parameter_167" + original_name = "batch_norm2d_104.w_2" shape = [96] dtype = "float32" min_val = float("0.00612442") @@ -1848,6 +2016,7 @@ class Program_weight_tensor_parameter_167: class Program_weight_tensor_parameter_168: name = "parameter_168" + original_name = "batch_norm2d_104.w_1" shape = [96] dtype = "float32" min_val = float("-0.0660382") @@ -1859,6 +2028,7 @@ class Program_weight_tensor_parameter_168: class Program_weight_tensor_parameter_169: name = "parameter_169" + original_name = "conv2d_108.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.107707") @@ -1870,6 +2040,7 @@ class Program_weight_tensor_parameter_169: class Program_weight_tensor_parameter_170: name = "parameter_170" + original_name = "batch_norm2d_103.b_0" shape = [96] dtype = "float32" min_val = float("-0.492552") @@ -1881,6 +2052,7 @@ class Program_weight_tensor_parameter_170: class Program_weight_tensor_parameter_171: name = "parameter_171" + original_name = "batch_norm2d_103.w_0" shape = [96] dtype = "float32" min_val = float("0.759592") @@ -1892,6 +2064,7 @@ class Program_weight_tensor_parameter_171: class Program_weight_tensor_parameter_172: name = "parameter_172" + original_name = "batch_norm2d_103.w_2" shape = [96] dtype = "float32" min_val = float("0.0172157") @@ -1903,6 +2076,7 @@ class Program_weight_tensor_parameter_172: class Program_weight_tensor_parameter_173: name = "parameter_173" + original_name = "batch_norm2d_103.w_1" shape = [96] dtype = "float32" min_val = float("-0.248811") @@ -1914,6 +2088,7 @@ class Program_weight_tensor_parameter_173: class Program_weight_tensor_parameter_174: name = "parameter_174" + original_name = "conv2d_107.w_0" shape = [96, 96, 3, 3] dtype = "float32" min_val = float("-0.116242") @@ -1925,6 +2100,7 @@ class Program_weight_tensor_parameter_174: class Program_weight_tensor_parameter_175: name = "parameter_175" + original_name = "batch_norm2d_102.b_0" shape = [96] dtype = "float32" min_val = float("-0.567319") @@ -1936,6 +2112,7 @@ class Program_weight_tensor_parameter_175: class Program_weight_tensor_parameter_176: name = "parameter_176" + original_name = "batch_norm2d_102.w_0" shape = [96] dtype = "float32" min_val = float("0.772527") @@ -1947,6 +2124,7 @@ class Program_weight_tensor_parameter_176: class Program_weight_tensor_parameter_177: name = "parameter_177" + original_name = "batch_norm2d_102.w_2" shape = [96] dtype = "float32" min_val = float("0.0160524") @@ -1958,6 +2136,7 @@ class Program_weight_tensor_parameter_177: class Program_weight_tensor_parameter_178: name = "parameter_178" + original_name = "batch_norm2d_102.w_1" shape = [96] dtype = "float32" min_val = float("-0.189984") @@ -1969,6 +2148,7 @@ class Program_weight_tensor_parameter_178: class Program_weight_tensor_parameter_179: name = "parameter_179" + original_name = "conv2d_106.w_0" shape = [96, 96, 3, 3] dtype = "float32" min_val = float("-0.161713") @@ -1980,6 +2160,7 @@ class Program_weight_tensor_parameter_179: class Program_weight_tensor_parameter_180: name = "parameter_180" + original_name = "batch_norm2d_101.b_0" shape = [96] dtype = "float32" min_val = float("-0.627686") @@ -1991,6 +2172,7 @@ class Program_weight_tensor_parameter_180: class Program_weight_tensor_parameter_181: name = "parameter_181" + original_name = "batch_norm2d_101.w_0" shape = [96] dtype = "float32" min_val = float("0.653052") @@ -2002,6 +2184,7 @@ class Program_weight_tensor_parameter_181: class Program_weight_tensor_parameter_182: name = "parameter_182" + original_name = "batch_norm2d_101.w_2" shape = [96] dtype = "float32" min_val = float("0.0123443") @@ -2013,6 +2196,7 @@ class Program_weight_tensor_parameter_182: class Program_weight_tensor_parameter_183: name = "parameter_183" + original_name = "batch_norm2d_101.w_1" shape = [96] dtype = "float32" min_val = float("-0.117179") @@ -2024,6 +2208,7 @@ class Program_weight_tensor_parameter_183: class Program_weight_tensor_parameter_184: name = "parameter_184" + original_name = "conv2d_105.w_0" shape = [96, 448, 1, 1] dtype = "float32" min_val = float("-0.162211") @@ -2035,6 +2220,7 @@ class Program_weight_tensor_parameter_184: class Program_weight_tensor_parameter_185: name = "parameter_185" + original_name = "batch_norm2d_100.b_0" shape = [96] dtype = "float32" min_val = float("-0.0984774") @@ -2046,6 +2232,7 @@ class Program_weight_tensor_parameter_185: class Program_weight_tensor_parameter_186: name = "parameter_186" + original_name = "batch_norm2d_100.w_0" shape = [96] dtype = "float32" min_val = float("0.692561") @@ -2057,6 +2244,7 @@ class Program_weight_tensor_parameter_186: class Program_weight_tensor_parameter_187: name = "parameter_187" + original_name = "batch_norm2d_100.w_2" shape = [96] dtype = "float32" min_val = float("0.00681771") @@ -2068,6 +2256,7 @@ class Program_weight_tensor_parameter_187: class Program_weight_tensor_parameter_188: name = "parameter_188" + original_name = "batch_norm2d_100.w_1" shape = [96] dtype = "float32" min_val = float("-0.133097") @@ -2079,6 +2268,7 @@ class Program_weight_tensor_parameter_188: class Program_weight_tensor_parameter_189: name = "parameter_189" + original_name = "conv2d_104.w_0" shape = [96, 448, 1, 1] dtype = "float32" min_val = float("-0.104838") @@ -2090,6 +2280,7 @@ class Program_weight_tensor_parameter_189: class Program_weight_tensor_parameter_190: name = "parameter_190" + original_name = "batch_norm2d_99.b_0" shape = [192] dtype = "float32" min_val = float("-0.296963") @@ -2101,6 +2292,7 @@ class Program_weight_tensor_parameter_190: class Program_weight_tensor_parameter_191: name = "parameter_191" + original_name = "batch_norm2d_99.w_0" shape = [192] dtype = "float32" min_val = float("0.672164") @@ -2112,6 +2304,7 @@ class Program_weight_tensor_parameter_191: class Program_weight_tensor_parameter_192: name = "parameter_192" + original_name = "batch_norm2d_99.w_2" shape = [192] dtype = "float32" min_val = float("0.0110872") @@ -2123,6 +2316,7 @@ class Program_weight_tensor_parameter_192: class Program_weight_tensor_parameter_193: name = "parameter_193" + original_name = "batch_norm2d_99.w_1" shape = [192] dtype = "float32" min_val = float("-0.151007") @@ -2134,6 +2328,7 @@ class Program_weight_tensor_parameter_193: class Program_weight_tensor_parameter_194: name = "parameter_194" + original_name = "conv2d_103.w_0" shape = [192, 384, 1, 1] dtype = "float32" min_val = float("-0.0904828") @@ -2145,6 +2340,7 @@ class Program_weight_tensor_parameter_194: class Program_weight_tensor_parameter_195: name = "parameter_195" + original_name = "batch_norm2d_98.b_0" shape = [384] dtype = "float32" min_val = float("-0.202322") @@ -2156,6 +2352,7 @@ class Program_weight_tensor_parameter_195: class Program_weight_tensor_parameter_196: name = "parameter_196" + original_name = "batch_norm2d_98.w_0" shape = [384] dtype = "float32" min_val = float("0.872032") @@ -2167,6 +2364,7 @@ class Program_weight_tensor_parameter_196: class Program_weight_tensor_parameter_197: name = "parameter_197" + original_name = "batch_norm2d_98.w_2" shape = [384] dtype = "float32" min_val = float("0.00721296") @@ -2178,6 +2376,7 @@ class Program_weight_tensor_parameter_197: class Program_weight_tensor_parameter_198: name = "parameter_198" + original_name = "batch_norm2d_98.w_1" shape = [384] dtype = "float32" min_val = float("-0.297588") @@ -2189,6 +2388,7 @@ class Program_weight_tensor_parameter_198: class Program_weight_tensor_parameter_199: name = "parameter_199" + original_name = "conv2d_102.w_0" shape = [384, 384, 1, 1] dtype = "float32" min_val = float("-0.102299") @@ -2200,6 +2400,7 @@ class Program_weight_tensor_parameter_199: class Program_weight_tensor_parameter_200: name = "parameter_200" + original_name = "batch_norm2d_97.b_0" shape = [192] dtype = "float32" min_val = float("-0.177096") @@ -2211,6 +2412,7 @@ class Program_weight_tensor_parameter_200: class Program_weight_tensor_parameter_201: name = "parameter_201" + original_name = "batch_norm2d_97.w_0" shape = [192] dtype = "float32" min_val = float("0.884728") @@ -2222,6 +2424,7 @@ class Program_weight_tensor_parameter_201: class Program_weight_tensor_parameter_202: name = "parameter_202" + original_name = "batch_norm2d_97.w_2" shape = [192] dtype = "float32" min_val = float("0.00441221") @@ -2233,6 +2436,7 @@ class Program_weight_tensor_parameter_202: class Program_weight_tensor_parameter_203: name = "parameter_203" + original_name = "batch_norm2d_97.w_1" shape = [192] dtype = "float32" min_val = float("-0.0812598") @@ -2244,6 +2448,7 @@ class Program_weight_tensor_parameter_203: class Program_weight_tensor_parameter_204: name = "parameter_204" + original_name = "conv2d_101.w_0" shape = [192, 192, 1, 1] dtype = "float32" min_val = float("-0.0495063") @@ -2255,6 +2460,7 @@ class Program_weight_tensor_parameter_204: class Program_weight_tensor_parameter_205: name = "parameter_205" + original_name = "batch_norm2d_96.b_0" shape = [192] dtype = "float32" min_val = float("-0.177096") @@ -2266,6 +2472,7 @@ class Program_weight_tensor_parameter_205: class Program_weight_tensor_parameter_206: name = "parameter_206" + original_name = "batch_norm2d_96.w_0" shape = [192] dtype = "float32" min_val = float("0.944815") @@ -2277,6 +2484,7 @@ class Program_weight_tensor_parameter_206: class Program_weight_tensor_parameter_207: name = "parameter_207" + original_name = "batch_norm2d_96.w_2" shape = [192] dtype = "float32" min_val = float("0.016531") @@ -2288,6 +2496,7 @@ class Program_weight_tensor_parameter_207: class Program_weight_tensor_parameter_208: name = "parameter_208" + original_name = "batch_norm2d_96.w_1" shape = [192] dtype = "float32" min_val = float("-0.21924") @@ -2299,6 +2508,7 @@ class Program_weight_tensor_parameter_208: class Program_weight_tensor_parameter_209: name = "parameter_209" + original_name = "conv2d_100.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.0428771") @@ -2310,6 +2520,7 @@ class Program_weight_tensor_parameter_209: class Program_weight_tensor_parameter_210: name = "parameter_210" + original_name = "batch_norm2d_95.b_0" shape = [192] dtype = "float32" min_val = float("-0.216413") @@ -2321,6 +2532,7 @@ class Program_weight_tensor_parameter_210: class Program_weight_tensor_parameter_211: name = "parameter_211" + original_name = "batch_norm2d_95.w_0" shape = [192] dtype = "float32" min_val = float("0.939878") @@ -2332,6 +2544,7 @@ class Program_weight_tensor_parameter_211: class Program_weight_tensor_parameter_212: name = "parameter_212" + original_name = "batch_norm2d_95.w_2" shape = [192] dtype = "float32" min_val = float("0.0381921") @@ -2343,6 +2556,7 @@ class Program_weight_tensor_parameter_212: class Program_weight_tensor_parameter_213: name = "parameter_213" + original_name = "batch_norm2d_95.w_1" shape = [192] dtype = "float32" min_val = float("-0.194913") @@ -2354,6 +2568,7 @@ class Program_weight_tensor_parameter_213: class Program_weight_tensor_parameter_214: name = "parameter_214" + original_name = "conv2d_99.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.0537036") @@ -2365,6 +2580,7 @@ class Program_weight_tensor_parameter_214: class Program_weight_tensor_parameter_215: name = "parameter_215" + original_name = "batch_norm2d_94.b_0" shape = [192] dtype = "float32" min_val = float("-0.196865") @@ -2376,6 +2592,7 @@ class Program_weight_tensor_parameter_215: class Program_weight_tensor_parameter_216: name = "parameter_216" + original_name = "batch_norm2d_94.w_0" shape = [192] dtype = "float32" min_val = float("0.944171") @@ -2387,6 +2604,7 @@ class Program_weight_tensor_parameter_216: class Program_weight_tensor_parameter_217: name = "parameter_217" + original_name = "batch_norm2d_94.w_2" shape = [192] dtype = "float32" min_val = float("0.00232538") @@ -2398,6 +2616,7 @@ class Program_weight_tensor_parameter_217: class Program_weight_tensor_parameter_218: name = "parameter_218" + original_name = "batch_norm2d_94.w_1" shape = [192] dtype = "float32" min_val = float("-0.0893703") @@ -2409,6 +2628,7 @@ class Program_weight_tensor_parameter_218: class Program_weight_tensor_parameter_219: name = "parameter_219" + original_name = "conv2d_98.w_0" shape = [192, 192, 1, 1] dtype = "float32" min_val = float("-0.0308295") @@ -2420,6 +2640,7 @@ class Program_weight_tensor_parameter_219: class Program_weight_tensor_parameter_220: name = "parameter_220" + original_name = "batch_norm2d_93.b_0" shape = [192] dtype = "float32" min_val = float("-0.196865") @@ -2431,6 +2652,7 @@ class Program_weight_tensor_parameter_220: class Program_weight_tensor_parameter_221: name = "parameter_221" + original_name = "batch_norm2d_93.w_0" shape = [192] dtype = "float32" min_val = float("0.953905") @@ -2442,6 +2664,7 @@ class Program_weight_tensor_parameter_221: class Program_weight_tensor_parameter_222: name = "parameter_222" + original_name = "batch_norm2d_93.w_2" shape = [192] dtype = "float32" min_val = float("0.00885181") @@ -2453,6 +2676,7 @@ class Program_weight_tensor_parameter_222: class Program_weight_tensor_parameter_223: name = "parameter_223" + original_name = "batch_norm2d_93.w_1" shape = [192] dtype = "float32" min_val = float("-0.214638") @@ -2464,6 +2688,7 @@ class Program_weight_tensor_parameter_223: class Program_weight_tensor_parameter_224: name = "parameter_224" + original_name = "conv2d_97.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.0395263") @@ -2475,6 +2700,7 @@ class Program_weight_tensor_parameter_224: class Program_weight_tensor_parameter_225: name = "parameter_225" + original_name = "batch_norm2d_92.b_0" shape = [192] dtype = "float32" min_val = float("-0.23254") @@ -2486,6 +2712,7 @@ class Program_weight_tensor_parameter_225: class Program_weight_tensor_parameter_226: name = "parameter_226" + original_name = "batch_norm2d_92.w_0" shape = [192] dtype = "float32" min_val = float("0.94661") @@ -2497,6 +2724,7 @@ class Program_weight_tensor_parameter_226: class Program_weight_tensor_parameter_227: name = "parameter_227" + original_name = "batch_norm2d_92.w_2" shape = [192] dtype = "float32" min_val = float("0.0325254") @@ -2508,6 +2736,7 @@ class Program_weight_tensor_parameter_227: class Program_weight_tensor_parameter_228: name = "parameter_228" + original_name = "batch_norm2d_92.w_1" shape = [192] dtype = "float32" min_val = float("-0.348988") @@ -2519,6 +2748,7 @@ class Program_weight_tensor_parameter_228: class Program_weight_tensor_parameter_229: name = "parameter_229" + original_name = "conv2d_96.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.0523599") @@ -2530,6 +2760,7 @@ class Program_weight_tensor_parameter_229: class Program_weight_tensor_parameter_230: name = "parameter_230" + original_name = "batch_norm2d_91.b_0" shape = [192] dtype = "float32" min_val = float("-0.154829") @@ -2541,6 +2772,7 @@ class Program_weight_tensor_parameter_230: class Program_weight_tensor_parameter_231: name = "parameter_231" + original_name = "batch_norm2d_91.w_0" shape = [192] dtype = "float32" min_val = float("0.932711") @@ -2552,6 +2784,7 @@ class Program_weight_tensor_parameter_231: class Program_weight_tensor_parameter_232: name = "parameter_232" + original_name = "batch_norm2d_91.w_2" shape = [192] dtype = "float32" min_val = float("0.00202058") @@ -2563,6 +2796,7 @@ class Program_weight_tensor_parameter_232: class Program_weight_tensor_parameter_233: name = "parameter_233" + original_name = "batch_norm2d_91.w_1" shape = [192] dtype = "float32" min_val = float("-0.0777916") @@ -2574,6 +2808,7 @@ class Program_weight_tensor_parameter_233: class Program_weight_tensor_parameter_234: name = "parameter_234" + original_name = "conv2d_95.w_0" shape = [192, 192, 1, 1] dtype = "float32" min_val = float("-0.0340527") @@ -2585,6 +2820,7 @@ class Program_weight_tensor_parameter_234: class Program_weight_tensor_parameter_235: name = "parameter_235" + original_name = "batch_norm2d_90.b_0" shape = [192] dtype = "float32" min_val = float("-0.15483") @@ -2596,6 +2832,7 @@ class Program_weight_tensor_parameter_235: class Program_weight_tensor_parameter_236: name = "parameter_236" + original_name = "batch_norm2d_90.w_0" shape = [192] dtype = "float32" min_val = float("0.935817") @@ -2607,6 +2844,7 @@ class Program_weight_tensor_parameter_236: class Program_weight_tensor_parameter_237: name = "parameter_237" + original_name = "batch_norm2d_90.w_2" shape = [192] dtype = "float32" min_val = float("0.00891321") @@ -2618,6 +2856,7 @@ class Program_weight_tensor_parameter_237: class Program_weight_tensor_parameter_238: name = "parameter_238" + original_name = "batch_norm2d_90.w_1" shape = [192] dtype = "float32" min_val = float("-0.258263") @@ -2629,6 +2868,7 @@ class Program_weight_tensor_parameter_238: class Program_weight_tensor_parameter_239: name = "parameter_239" + original_name = "conv2d_94.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.0292394") @@ -2640,6 +2880,7 @@ class Program_weight_tensor_parameter_239: class Program_weight_tensor_parameter_240: name = "parameter_240" + original_name = "batch_norm2d_89.b_0" shape = [192] dtype = "float32" min_val = float("-0.288771") @@ -2651,6 +2892,7 @@ class Program_weight_tensor_parameter_240: class Program_weight_tensor_parameter_241: name = "parameter_241" + original_name = "batch_norm2d_89.w_0" shape = [192] dtype = "float32" min_val = float("0.944044") @@ -2662,6 +2904,7 @@ class Program_weight_tensor_parameter_241: class Program_weight_tensor_parameter_242: name = "parameter_242" + original_name = "batch_norm2d_89.w_2" shape = [192] dtype = "float32" min_val = float("0.0137713") @@ -2673,6 +2916,7 @@ class Program_weight_tensor_parameter_242: class Program_weight_tensor_parameter_243: name = "parameter_243" + original_name = "batch_norm2d_89.w_1" shape = [192] dtype = "float32" min_val = float("-0.354673") @@ -2684,6 +2928,7 @@ class Program_weight_tensor_parameter_243: class Program_weight_tensor_parameter_244: name = "parameter_244" + original_name = "conv2d_93.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.0564756") @@ -2695,6 +2940,7 @@ class Program_weight_tensor_parameter_244: class Program_weight_tensor_parameter_245: name = "parameter_245" + original_name = "batch_norm2d_88.b_0" shape = [192] dtype = "float32" min_val = float("-0.257095") @@ -2706,6 +2952,7 @@ class Program_weight_tensor_parameter_245: class Program_weight_tensor_parameter_246: name = "parameter_246" + original_name = "batch_norm2d_88.w_0" shape = [192] dtype = "float32" min_val = float("0.916459") @@ -2717,6 +2964,7 @@ class Program_weight_tensor_parameter_246: class Program_weight_tensor_parameter_247: name = "parameter_247" + original_name = "batch_norm2d_88.w_2" shape = [192] dtype = "float32" min_val = float("0.00516469") @@ -2728,6 +2976,7 @@ class Program_weight_tensor_parameter_247: class Program_weight_tensor_parameter_248: name = "parameter_248" + original_name = "batch_norm2d_88.w_1" shape = [192] dtype = "float32" min_val = float("-0.127284") @@ -2739,6 +2988,7 @@ class Program_weight_tensor_parameter_248: class Program_weight_tensor_parameter_249: name = "parameter_249" + original_name = "conv2d_92.w_0" shape = [192, 896, 1, 1] dtype = "float32" min_val = float("-0.0721174") @@ -2750,6 +3000,7 @@ class Program_weight_tensor_parameter_249: class Program_weight_tensor_parameter_250: name = "parameter_250" + original_name = "batch_norm2d_87.b_0" shape = [192] dtype = "float32" min_val = float("-0.177705") @@ -2761,6 +3012,7 @@ class Program_weight_tensor_parameter_250: class Program_weight_tensor_parameter_251: name = "parameter_251" + original_name = "batch_norm2d_87.w_0" shape = [192] dtype = "float32" min_val = float("0.954707") @@ -2772,6 +3024,7 @@ class Program_weight_tensor_parameter_251: class Program_weight_tensor_parameter_252: name = "parameter_252" + original_name = "batch_norm2d_87.w_2" shape = [192] dtype = "float32" min_val = float("0.00824461") @@ -2783,6 +3036,7 @@ class Program_weight_tensor_parameter_252: class Program_weight_tensor_parameter_253: name = "parameter_253" + original_name = "batch_norm2d_87.w_1" shape = [192] dtype = "float32" min_val = float("-0.0769287") @@ -2794,6 +3048,7 @@ class Program_weight_tensor_parameter_253: class Program_weight_tensor_parameter_254: name = "parameter_254" + original_name = "conv2d_91.w_0" shape = [192, 896, 1, 1] dtype = "float32" min_val = float("-0.0604261") @@ -2805,6 +3060,7 @@ class Program_weight_tensor_parameter_254: class Program_weight_tensor_parameter_255: name = "parameter_255" + original_name = "batch_norm2d_86.b_0" shape = [384] dtype = "float32" min_val = float("-0.249989") @@ -2816,6 +3072,7 @@ class Program_weight_tensor_parameter_255: class Program_weight_tensor_parameter_256: name = "parameter_256" + original_name = "batch_norm2d_86.w_0" shape = [384] dtype = "float32" min_val = float("0.816049") @@ -2827,6 +3084,7 @@ class Program_weight_tensor_parameter_256: class Program_weight_tensor_parameter_257: name = "parameter_257" + original_name = "batch_norm2d_86.w_2" shape = [384] dtype = "float32" min_val = float("0.0103681") @@ -2838,6 +3096,7 @@ class Program_weight_tensor_parameter_257: class Program_weight_tensor_parameter_258: name = "parameter_258" + original_name = "batch_norm2d_86.w_1" shape = [384] dtype = "float32" min_val = float("-0.16079") @@ -2849,6 +3108,7 @@ class Program_weight_tensor_parameter_258: class Program_weight_tensor_parameter_259: name = "parameter_259" + original_name = "conv2d_90.w_0" shape = [384, 768, 1, 1] dtype = "float32" min_val = float("-0.03192") @@ -2860,6 +3120,7 @@ class Program_weight_tensor_parameter_259: class Program_weight_tensor_parameter_260: name = "parameter_260" + original_name = "batch_norm2d_85.b_0" shape = [768] dtype = "float32" min_val = float("-0.104731") @@ -2871,6 +3132,7 @@ class Program_weight_tensor_parameter_260: class Program_weight_tensor_parameter_261: name = "parameter_261" + original_name = "batch_norm2d_85.w_0" shape = [768] dtype = "float32" min_val = float("0.952515") @@ -2882,6 +3144,7 @@ class Program_weight_tensor_parameter_261: class Program_weight_tensor_parameter_262: name = "parameter_262" + original_name = "batch_norm2d_85.w_2" shape = [768] dtype = "float32" min_val = float("0.00400082") @@ -2893,6 +3156,7 @@ class Program_weight_tensor_parameter_262: class Program_weight_tensor_parameter_263: name = "parameter_263" + original_name = "batch_norm2d_85.w_1" shape = [768] dtype = "float32" min_val = float("-0.108131") @@ -2904,6 +3168,7 @@ class Program_weight_tensor_parameter_263: class Program_weight_tensor_parameter_264: name = "parameter_264" + original_name = "conv2d_89.w_0" shape = [768, 768, 1, 1] dtype = "float32" min_val = float("-0.0561444") @@ -2915,6 +3180,7 @@ class Program_weight_tensor_parameter_264: class Program_weight_tensor_parameter_265: name = "parameter_265" + original_name = "batch_norm2d_84.b_0" shape = [384] dtype = "float32" min_val = float("-0.158352") @@ -2926,6 +3192,7 @@ class Program_weight_tensor_parameter_265: class Program_weight_tensor_parameter_266: name = "parameter_266" + original_name = "batch_norm2d_84.w_0" shape = [384] dtype = "float32" min_val = float("0.888537") @@ -2937,6 +3204,7 @@ class Program_weight_tensor_parameter_266: class Program_weight_tensor_parameter_267: name = "parameter_267" + original_name = "batch_norm2d_84.w_2" shape = [384] dtype = "float32" min_val = float("0.00629694") @@ -2948,6 +3216,7 @@ class Program_weight_tensor_parameter_267: class Program_weight_tensor_parameter_268: name = "parameter_268" + original_name = "batch_norm2d_84.w_1" shape = [384] dtype = "float32" min_val = float("-0.0725783") @@ -2959,6 +3228,7 @@ class Program_weight_tensor_parameter_268: class Program_weight_tensor_parameter_269: name = "parameter_269" + original_name = "conv2d_88.w_0" shape = [384, 384, 1, 1] dtype = "float32" min_val = float("-0.0355484") @@ -2970,6 +3240,7 @@ class Program_weight_tensor_parameter_269: class Program_weight_tensor_parameter_270: name = "parameter_270" + original_name = "batch_norm2d_83.b_0" shape = [384] dtype = "float32" min_val = float("-0.158353") @@ -2981,6 +3252,7 @@ class Program_weight_tensor_parameter_270: class Program_weight_tensor_parameter_271: name = "parameter_271" + original_name = "batch_norm2d_83.w_0" shape = [384] dtype = "float32" min_val = float("0.880933") @@ -2992,6 +3264,7 @@ class Program_weight_tensor_parameter_271: class Program_weight_tensor_parameter_272: name = "parameter_272" + original_name = "batch_norm2d_83.w_2" shape = [384] dtype = "float32" min_val = float("0.0302136") @@ -3003,6 +3276,7 @@ class Program_weight_tensor_parameter_272: class Program_weight_tensor_parameter_273: name = "parameter_273" + original_name = "batch_norm2d_83.w_1" shape = [384] dtype = "float32" min_val = float("-0.280458") @@ -3014,6 +3288,7 @@ class Program_weight_tensor_parameter_273: class Program_weight_tensor_parameter_274: name = "parameter_274" + original_name = "conv2d_87.w_0" shape = [384, 384, 3, 3] dtype = "float32" min_val = float("-0.0402524") @@ -3025,6 +3300,7 @@ class Program_weight_tensor_parameter_274: class Program_weight_tensor_parameter_275: name = "parameter_275" + original_name = "batch_norm2d_82.b_0" shape = [384] dtype = "float32" min_val = float("-0.080174") @@ -3036,6 +3312,7 @@ class Program_weight_tensor_parameter_275: class Program_weight_tensor_parameter_276: name = "parameter_276" + original_name = "batch_norm2d_82.w_0" shape = [384] dtype = "float32" min_val = float("0.920426") @@ -3047,6 +3324,7 @@ class Program_weight_tensor_parameter_276: class Program_weight_tensor_parameter_277: name = "parameter_277" + original_name = "batch_norm2d_82.w_2" shape = [384] dtype = "float32" min_val = float("0.0258173") @@ -3058,6 +3336,7 @@ class Program_weight_tensor_parameter_277: class Program_weight_tensor_parameter_278: name = "parameter_278" + original_name = "batch_norm2d_82.w_1" shape = [384] dtype = "float32" min_val = float("-0.231545") @@ -3069,6 +3348,7 @@ class Program_weight_tensor_parameter_278: class Program_weight_tensor_parameter_279: name = "parameter_279" + original_name = "conv2d_86.w_0" shape = [384, 384, 3, 3] dtype = "float32" min_val = float("-0.023736") @@ -3080,6 +3360,7 @@ class Program_weight_tensor_parameter_279: class Program_weight_tensor_parameter_280: name = "parameter_280" + original_name = "batch_norm2d_81.b_0" shape = [384] dtype = "float32" min_val = float("-0.0734011") @@ -3091,6 +3372,7 @@ class Program_weight_tensor_parameter_280: class Program_weight_tensor_parameter_281: name = "parameter_281" + original_name = "batch_norm2d_81.w_0" shape = [384] dtype = "float32" min_val = float("0.946001") @@ -3102,6 +3384,7 @@ class Program_weight_tensor_parameter_281: class Program_weight_tensor_parameter_282: name = "parameter_282" + original_name = "batch_norm2d_81.w_2" shape = [384] dtype = "float32" min_val = float("0.0574212") @@ -3113,6 +3396,7 @@ class Program_weight_tensor_parameter_282: class Program_weight_tensor_parameter_283: name = "parameter_283" + original_name = "batch_norm2d_81.w_1" shape = [384] dtype = "float32" min_val = float("-1.55759") @@ -3124,6 +3408,7 @@ class Program_weight_tensor_parameter_283: class Program_weight_tensor_parameter_284: name = "parameter_284" + original_name = "conv2d_85.w_0" shape = [384, 1536, 1, 1] dtype = "float32" min_val = float("-0.0464008") @@ -3135,6 +3420,7 @@ class Program_weight_tensor_parameter_284: class Program_weight_tensor_parameter_285: name = "parameter_285" + original_name = "batch_norm2d_80.b_0" shape = [384] dtype = "float32" min_val = float("-0.0183405") @@ -3146,6 +3432,7 @@ class Program_weight_tensor_parameter_285: class Program_weight_tensor_parameter_286: name = "parameter_286" + original_name = "batch_norm2d_80.w_0" shape = [384] dtype = "float32" min_val = float("0.969528") @@ -3157,6 +3444,7 @@ class Program_weight_tensor_parameter_286: class Program_weight_tensor_parameter_287: name = "parameter_287" + original_name = "batch_norm2d_80.w_2" shape = [384] dtype = "float32" min_val = float("0.00270127") @@ -3168,6 +3456,7 @@ class Program_weight_tensor_parameter_287: class Program_weight_tensor_parameter_288: name = "parameter_288" + original_name = "batch_norm2d_80.w_1" shape = [384] dtype = "float32" min_val = float("-0.100294") @@ -3179,6 +3468,7 @@ class Program_weight_tensor_parameter_288: class Program_weight_tensor_parameter_289: name = "parameter_289" + original_name = "conv2d_84.w_0" shape = [384, 384, 1, 1] dtype = "float32" min_val = float("-0.0312313") @@ -3190,6 +3480,7 @@ class Program_weight_tensor_parameter_289: class Program_weight_tensor_parameter_290: name = "parameter_290" + original_name = "batch_norm2d_79.b_0" shape = [384] dtype = "float32" min_val = float("-0.0183405") @@ -3201,6 +3492,7 @@ class Program_weight_tensor_parameter_290: class Program_weight_tensor_parameter_291: name = "parameter_291" + original_name = "batch_norm2d_79.w_0" shape = [384] dtype = "float32" min_val = float("0.971893") @@ -3212,6 +3504,7 @@ class Program_weight_tensor_parameter_291: class Program_weight_tensor_parameter_292: name = "parameter_292" + original_name = "batch_norm2d_79.w_2" shape = [384] dtype = "float32" min_val = float("0.0120607") @@ -3223,6 +3516,7 @@ class Program_weight_tensor_parameter_292: class Program_weight_tensor_parameter_293: name = "parameter_293" + original_name = "batch_norm2d_79.w_1" shape = [384] dtype = "float32" min_val = float("-0.26934") @@ -3234,6 +3528,7 @@ class Program_weight_tensor_parameter_293: class Program_weight_tensor_parameter_294: name = "parameter_294" + original_name = "conv2d_83.w_0" shape = [384, 384, 3, 3] dtype = "float32" min_val = float("-0.0290225") @@ -3245,6 +3540,7 @@ class Program_weight_tensor_parameter_294: class Program_weight_tensor_parameter_295: name = "parameter_295" + original_name = "batch_norm2d_78.b_0" shape = [384] dtype = "float32" min_val = float("-0.0494678") @@ -3256,6 +3552,7 @@ class Program_weight_tensor_parameter_295: class Program_weight_tensor_parameter_296: name = "parameter_296" + original_name = "batch_norm2d_78.w_0" shape = [384] dtype = "float32" min_val = float("0.95427") @@ -3267,6 +3564,7 @@ class Program_weight_tensor_parameter_296: class Program_weight_tensor_parameter_297: name = "parameter_297" + original_name = "batch_norm2d_78.w_2" shape = [384] dtype = "float32" min_val = float("0.0686687") @@ -3278,6 +3576,7 @@ class Program_weight_tensor_parameter_297: class Program_weight_tensor_parameter_298: name = "parameter_298" + original_name = "batch_norm2d_78.w_1" shape = [384] dtype = "float32" min_val = float("-1.18586") @@ -3289,6 +3588,7 @@ class Program_weight_tensor_parameter_298: class Program_weight_tensor_parameter_299: name = "parameter_299" + original_name = "conv2d_82.w_0" shape = [384, 384, 3, 3] dtype = "float32" min_val = float("-0.0230308") @@ -3300,6 +3600,7 @@ class Program_weight_tensor_parameter_299: class Program_weight_tensor_parameter_300: name = "parameter_300" + original_name = "batch_norm2d_77.b_0" shape = [384] dtype = "float32" min_val = float("-0.0358263") @@ -3311,6 +3612,7 @@ class Program_weight_tensor_parameter_300: class Program_weight_tensor_parameter_301: name = "parameter_301" + original_name = "batch_norm2d_77.w_0" shape = [384] dtype = "float32" min_val = float("0.984161") @@ -3322,6 +3624,7 @@ class Program_weight_tensor_parameter_301: class Program_weight_tensor_parameter_302: name = "parameter_302" + original_name = "batch_norm2d_77.w_2" shape = [384] dtype = "float32" min_val = float("0.00175415") @@ -3333,6 +3636,7 @@ class Program_weight_tensor_parameter_302: class Program_weight_tensor_parameter_303: name = "parameter_303" + original_name = "batch_norm2d_77.w_1" shape = [384] dtype = "float32" min_val = float("-0.0809601") @@ -3344,6 +3648,7 @@ class Program_weight_tensor_parameter_303: class Program_weight_tensor_parameter_304: name = "parameter_304" + original_name = "conv2d_81.w_0" shape = [384, 384, 1, 1] dtype = "float32" min_val = float("-0.0193951") @@ -3355,6 +3660,7 @@ class Program_weight_tensor_parameter_304: class Program_weight_tensor_parameter_305: name = "parameter_305" + original_name = "batch_norm2d_76.b_0" shape = [384] dtype = "float32" min_val = float("-0.0358263") @@ -3366,6 +3672,7 @@ class Program_weight_tensor_parameter_305: class Program_weight_tensor_parameter_306: name = "parameter_306" + original_name = "batch_norm2d_76.w_0" shape = [384] dtype = "float32" min_val = float("0.981952") @@ -3377,6 +3684,7 @@ class Program_weight_tensor_parameter_306: class Program_weight_tensor_parameter_307: name = "parameter_307" + original_name = "batch_norm2d_76.w_2" shape = [384] dtype = "float32" min_val = float("0.00890186") @@ -3388,6 +3696,7 @@ class Program_weight_tensor_parameter_307: class Program_weight_tensor_parameter_308: name = "parameter_308" + original_name = "batch_norm2d_76.w_1" shape = [384] dtype = "float32" min_val = float("-0.216942") @@ -3399,6 +3708,7 @@ class Program_weight_tensor_parameter_308: class Program_weight_tensor_parameter_309: name = "parameter_309" + original_name = "conv2d_80.w_0" shape = [384, 384, 3, 3] dtype = "float32" min_val = float("-0.011273") @@ -3410,6 +3720,7 @@ class Program_weight_tensor_parameter_309: class Program_weight_tensor_parameter_310: name = "parameter_310" + original_name = "batch_norm2d_75.b_0" shape = [384] dtype = "float32" min_val = float("-0.0530152") @@ -3421,6 +3732,7 @@ class Program_weight_tensor_parameter_310: class Program_weight_tensor_parameter_311: name = "parameter_311" + original_name = "batch_norm2d_75.w_0" shape = [384] dtype = "float32" min_val = float("0.975684") @@ -3432,6 +3744,7 @@ class Program_weight_tensor_parameter_311: class Program_weight_tensor_parameter_312: name = "parameter_312" + original_name = "batch_norm2d_75.w_2" shape = [384] dtype = "float32" min_val = float("0.0111248") @@ -3443,6 +3756,7 @@ class Program_weight_tensor_parameter_312: class Program_weight_tensor_parameter_313: name = "parameter_313" + original_name = "batch_norm2d_75.w_1" shape = [384] dtype = "float32" min_val = float("-0.17639") @@ -3454,6 +3768,7 @@ class Program_weight_tensor_parameter_313: class Program_weight_tensor_parameter_314: name = "parameter_314" + original_name = "conv2d_79.w_0" shape = [384, 384, 3, 3] dtype = "float32" min_val = float("-0.0145313") @@ -3465,6 +3780,7 @@ class Program_weight_tensor_parameter_314: class Program_weight_tensor_parameter_315: name = "parameter_315" + original_name = "batch_norm2d_74.b_0" shape = [384] dtype = "float32" min_val = float("-0.0699692") @@ -3476,6 +3792,7 @@ class Program_weight_tensor_parameter_315: class Program_weight_tensor_parameter_316: name = "parameter_316" + original_name = "batch_norm2d_74.w_0" shape = [384] dtype = "float32" min_val = float("0.981916") @@ -3487,6 +3804,7 @@ class Program_weight_tensor_parameter_316: class Program_weight_tensor_parameter_317: name = "parameter_317" + original_name = "batch_norm2d_74.w_2" shape = [384] dtype = "float32" min_val = float("0.0069754") @@ -3498,6 +3816,7 @@ class Program_weight_tensor_parameter_317: class Program_weight_tensor_parameter_318: name = "parameter_318" + original_name = "batch_norm2d_74.w_1" shape = [384] dtype = "float32" min_val = float("-0.145777") @@ -3509,6 +3828,7 @@ class Program_weight_tensor_parameter_318: class Program_weight_tensor_parameter_319: name = "parameter_319" + original_name = "conv2d_78.w_0" shape = [384, 1024, 1, 1] dtype = "float32" min_val = float("-0.017823") @@ -3520,6 +3840,7 @@ class Program_weight_tensor_parameter_319: class Program_weight_tensor_parameter_320: name = "parameter_320" + original_name = "batch_norm2d_73.b_0" shape = [384] dtype = "float32" min_val = float("-0.0243502") @@ -3531,6 +3852,7 @@ class Program_weight_tensor_parameter_320: class Program_weight_tensor_parameter_321: name = "parameter_321" + original_name = "batch_norm2d_73.w_0" shape = [384] dtype = "float32" min_val = float("0.994149") @@ -3542,6 +3864,7 @@ class Program_weight_tensor_parameter_321: class Program_weight_tensor_parameter_322: name = "parameter_322" + original_name = "batch_norm2d_73.w_2" shape = [384] dtype = "float32" min_val = float("0.0116675") @@ -3553,6 +3876,7 @@ class Program_weight_tensor_parameter_322: class Program_weight_tensor_parameter_323: name = "parameter_323" + original_name = "batch_norm2d_73.w_1" shape = [384] dtype = "float32" min_val = float("-0.150832") @@ -3564,6 +3888,7 @@ class Program_weight_tensor_parameter_323: class Program_weight_tensor_parameter_324: name = "parameter_324" + original_name = "conv2d_77.w_0" shape = [384, 1024, 1, 1] dtype = "float32" min_val = float("-0.0386333") @@ -3575,6 +3900,7 @@ class Program_weight_tensor_parameter_324: class Program_weight_tensor_parameter_325: name = "parameter_325" + original_name = "layer_norm_1.b_0_deepcopy_47" shape = [1024] dtype = "float32" min_val = float("-2.92289e-10") @@ -3586,6 +3912,7 @@ class Program_weight_tensor_parameter_325: class Program_weight_tensor_parameter_326: name = "parameter_326" + original_name = "layer_norm_1.w_0_deepcopy_46" shape = [1024] dtype = "float32" min_val = float("0.797367") @@ -3597,6 +3924,7 @@ class Program_weight_tensor_parameter_326: class Program_weight_tensor_parameter_327: name = "parameter_327" + original_name = "linear_2.b_0_deepcopy_43" shape = [1024] dtype = "float32" min_val = float("-0.0176922") @@ -3608,6 +3936,7 @@ class Program_weight_tensor_parameter_327: class Program_weight_tensor_parameter_328: name = "parameter_328" + original_name = "linear_2.w_0_deepcopy_42" shape = [2048, 1024] dtype = "float32" min_val = float("-0.0180007") @@ -3619,6 +3948,7 @@ class Program_weight_tensor_parameter_328: class Program_weight_tensor_parameter_329: name = "parameter_329" + original_name = "linear_1.b_0_deepcopy_41" shape = [2048] dtype = "float32" min_val = float("-0.0249658") @@ -3630,6 +3960,7 @@ class Program_weight_tensor_parameter_329: class Program_weight_tensor_parameter_330: name = "parameter_330" + original_name = "linear_1.w_0_deepcopy_40" shape = [1024, 2048] dtype = "float32" min_val = float("-0.0252063") @@ -3641,6 +3972,7 @@ class Program_weight_tensor_parameter_330: class Program_weight_tensor_parameter_331: name = "parameter_331" + original_name = "layer_norm_0.b_0_deepcopy_45" shape = [1024] dtype = "float32" min_val = float("-0.000555217") @@ -3652,6 +3984,7 @@ class Program_weight_tensor_parameter_331: class Program_weight_tensor_parameter_332: name = "parameter_332" + original_name = "layer_norm_0.w_0_deepcopy_44" shape = [1024] dtype = "float32" min_val = float("0.7967") @@ -3663,6 +3996,7 @@ class Program_weight_tensor_parameter_332: class Program_weight_tensor_parameter_333: name = "parameter_333" + original_name = "linear_0.b_0_deepcopy_39" shape = [1024] dtype = "float32" min_val = float("-0.000419711") @@ -3674,6 +4008,7 @@ class Program_weight_tensor_parameter_333: class Program_weight_tensor_parameter_334: name = "parameter_334" + original_name = "linear_0.w_0_deepcopy_38" shape = [1024, 1024] dtype = "float32" min_val = float("-0.0435604") @@ -3685,6 +4020,7 @@ class Program_weight_tensor_parameter_334: class Program_weight_tensor_parameter_335: name = "parameter_335" + original_name = "layer_norm_1.b_0_deepcopy_35" shape = [1024] dtype = "float32" min_val = float("-0.0003762") @@ -3696,6 +4032,7 @@ class Program_weight_tensor_parameter_335: class Program_weight_tensor_parameter_336: name = "parameter_336" + original_name = "layer_norm_1.w_0_deepcopy_34" shape = [1024] dtype = "float32" min_val = float("0.796851") @@ -3707,6 +4044,7 @@ class Program_weight_tensor_parameter_336: class Program_weight_tensor_parameter_337: name = "parameter_337" + original_name = "linear_2.b_0_deepcopy_31" shape = [1024] dtype = "float32" min_val = float("-0.0176874") @@ -3718,6 +4056,7 @@ class Program_weight_tensor_parameter_337: class Program_weight_tensor_parameter_338: name = "parameter_338" + original_name = "linear_2.w_0_deepcopy_30" shape = [2048, 1024] dtype = "float32" min_val = float("-0.0179828") @@ -3729,6 +4068,7 @@ class Program_weight_tensor_parameter_338: class Program_weight_tensor_parameter_339: name = "parameter_339" + original_name = "linear_1.b_0_deepcopy_29" shape = [2048] dtype = "float32" min_val = float("-0.024962") @@ -3740,6 +4080,7 @@ class Program_weight_tensor_parameter_339: class Program_weight_tensor_parameter_340: name = "parameter_340" + original_name = "linear_1.w_0_deepcopy_28" shape = [1024, 2048] dtype = "float32" min_val = float("-0.0251902") @@ -3751,6 +4092,7 @@ class Program_weight_tensor_parameter_340: class Program_weight_tensor_parameter_341: name = "parameter_341" + original_name = "layer_norm_0.b_0_deepcopy_33" shape = [1024] dtype = "float32" min_val = float("-0.000350131") @@ -3762,6 +4104,7 @@ class Program_weight_tensor_parameter_341: class Program_weight_tensor_parameter_342: name = "parameter_342" + original_name = "layer_norm_0.w_0_deepcopy_32" shape = [1024] dtype = "float32" min_val = float("0.797093") @@ -3773,6 +4116,7 @@ class Program_weight_tensor_parameter_342: class Program_weight_tensor_parameter_343: name = "parameter_343" + original_name = "linear_0.b_0_deepcopy_27" shape = [1024] dtype = "float32" min_val = float("-0.000363068") @@ -3784,6 +4128,7 @@ class Program_weight_tensor_parameter_343: class Program_weight_tensor_parameter_344: name = "parameter_344" + original_name = "linear_0.w_0_deepcopy_26" shape = [1024, 1024] dtype = "float32" min_val = float("-0.0434436") @@ -3795,6 +4140,7 @@ class Program_weight_tensor_parameter_344: class Program_weight_tensor_parameter_345: name = "parameter_345" + original_name = "layer_norm_1.b_0_deepcopy_23" shape = [1024] dtype = "float32" min_val = float("-0.000420154") @@ -3806,6 +4152,7 @@ class Program_weight_tensor_parameter_345: class Program_weight_tensor_parameter_346: name = "parameter_346" + original_name = "layer_norm_1.w_0_deepcopy_22" shape = [1024] dtype = "float32" min_val = float("0.797252") @@ -3817,6 +4164,7 @@ class Program_weight_tensor_parameter_346: class Program_weight_tensor_parameter_347: name = "parameter_347" + original_name = "linear_2.b_0_deepcopy_19" shape = [1024] dtype = "float32" min_val = float("-0.0177165") @@ -3828,6 +4176,7 @@ class Program_weight_tensor_parameter_347: class Program_weight_tensor_parameter_348: name = "parameter_348" + original_name = "linear_2.w_0_deepcopy_18" shape = [2048, 1024] dtype = "float32" min_val = float("-0.0179359") @@ -3839,6 +4188,7 @@ class Program_weight_tensor_parameter_348: class Program_weight_tensor_parameter_349: name = "parameter_349" + original_name = "linear_1.b_0_deepcopy_17" shape = [2048] dtype = "float32" min_val = float("-0.0249926") @@ -3850,6 +4200,7 @@ class Program_weight_tensor_parameter_349: class Program_weight_tensor_parameter_350: name = "parameter_350" + original_name = "linear_1.w_0_deepcopy_16" shape = [1024, 2048] dtype = "float32" min_val = float("-0.0251321") @@ -3861,6 +4212,7 @@ class Program_weight_tensor_parameter_350: class Program_weight_tensor_parameter_351: name = "parameter_351" + original_name = "layer_norm_0.b_0_deepcopy_21" shape = [1024] dtype = "float32" min_val = float("-0.000524045") @@ -3872,6 +4224,7 @@ class Program_weight_tensor_parameter_351: class Program_weight_tensor_parameter_352: name = "parameter_352" + original_name = "layer_norm_0.w_0_deepcopy_20" shape = [1024] dtype = "float32" min_val = float("0.797309") @@ -3883,6 +4236,7 @@ class Program_weight_tensor_parameter_352: class Program_weight_tensor_parameter_353: name = "parameter_353" + original_name = "linear_0.b_0_deepcopy_15" shape = [1024] dtype = "float32" min_val = float("-0.000493111") @@ -3894,6 +4248,7 @@ class Program_weight_tensor_parameter_353: class Program_weight_tensor_parameter_354: name = "parameter_354" + original_name = "linear_0.w_0_deepcopy_14" shape = [1024, 1024] dtype = "float32" min_val = float("-0.043449") @@ -3905,6 +4260,7 @@ class Program_weight_tensor_parameter_354: class Program_weight_tensor_parameter_355: name = "parameter_355" + original_name = "layer_norm_1.b_0_deepcopy_11" shape = [1024] dtype = "float32" min_val = float("-0.000657995") @@ -3916,6 +4272,7 @@ class Program_weight_tensor_parameter_355: class Program_weight_tensor_parameter_356: name = "parameter_356" + original_name = "layer_norm_1.w_0_deepcopy_10" shape = [1024] dtype = "float32" min_val = float("0.797424") @@ -3927,6 +4284,7 @@ class Program_weight_tensor_parameter_356: class Program_weight_tensor_parameter_357: name = "parameter_357" + original_name = "linear_2.b_0_deepcopy_7" shape = [1024] dtype = "float32" min_val = float("-0.0177131") @@ -3938,6 +4296,7 @@ class Program_weight_tensor_parameter_357: class Program_weight_tensor_parameter_358: name = "parameter_358" + original_name = "linear_2.w_0_deepcopy_6" shape = [2048, 1024] dtype = "float32" min_val = float("-0.0180121") @@ -3949,6 +4308,7 @@ class Program_weight_tensor_parameter_358: class Program_weight_tensor_parameter_359: name = "parameter_359" + original_name = "linear_1.b_0_deepcopy_5" shape = [2048] dtype = "float32" min_val = float("-0.024929") @@ -3960,6 +4320,7 @@ class Program_weight_tensor_parameter_359: class Program_weight_tensor_parameter_360: name = "parameter_360" + original_name = "linear_1.w_0_deepcopy_4" shape = [1024, 2048] dtype = "float32" min_val = float("-0.0251524") @@ -3971,6 +4332,7 @@ class Program_weight_tensor_parameter_360: class Program_weight_tensor_parameter_361: name = "parameter_361" + original_name = "layer_norm_0.b_0_deepcopy_9" shape = [1024] dtype = "float32" min_val = float("-0.000733503") @@ -3982,6 +4344,7 @@ class Program_weight_tensor_parameter_361: class Program_weight_tensor_parameter_362: name = "parameter_362" + original_name = "layer_norm_0.w_0_deepcopy_8" shape = [1024] dtype = "float32" min_val = float("0.797348") @@ -3993,6 +4356,7 @@ class Program_weight_tensor_parameter_362: class Program_weight_tensor_parameter_363: name = "parameter_363" + original_name = "linear_0.b_0_deepcopy_3" shape = [1024] dtype = "float32" min_val = float("-0.000783512") @@ -4004,6 +4368,7 @@ class Program_weight_tensor_parameter_363: class Program_weight_tensor_parameter_364: name = "parameter_364" + original_name = "linear_0.w_0_deepcopy_2" shape = [1024, 1024] dtype = "float32" min_val = float("-0.0437281") @@ -4015,6 +4380,7 @@ class Program_weight_tensor_parameter_364: class Program_weight_tensor_parameter_365: name = "parameter_365" + original_name = "batch_norm2d_72.b_0" shape = [1024] dtype = "float32" min_val = float("-3.7594") @@ -4026,6 +4392,7 @@ class Program_weight_tensor_parameter_365: class Program_weight_tensor_parameter_366: name = "parameter_366" + original_name = "batch_norm2d_72.w_0" shape = [1024] dtype = "float32" min_val = float("1.61913") @@ -4037,6 +4404,7 @@ class Program_weight_tensor_parameter_366: class Program_weight_tensor_parameter_367: name = "parameter_367" + original_name = "batch_norm2d_72.w_2" shape = [1024] dtype = "float32" min_val = float("0.00445214") @@ -4048,6 +4416,7 @@ class Program_weight_tensor_parameter_367: class Program_weight_tensor_parameter_368: name = "parameter_368" + original_name = "batch_norm2d_72.w_1" shape = [1024] dtype = "float32" min_val = float("-0.140092") @@ -4059,6 +4428,7 @@ class Program_weight_tensor_parameter_368: class Program_weight_tensor_parameter_369: name = "parameter_369" + original_name = "conv2d_76.w_0" shape = [1024, 768, 1, 1] dtype = "float32" min_val = float("-0.0427729") @@ -4070,6 +4440,7 @@ class Program_weight_tensor_parameter_369: class Program_weight_tensor_parameter_370: name = "parameter_370" + original_name = "conv2d_75.b_0" shape = [768] dtype = "float32" min_val = float("-0.014467") @@ -4081,6 +4452,7 @@ class Program_weight_tensor_parameter_370: class Program_weight_tensor_parameter_371: name = "parameter_371" + original_name = "conv2d_75.w_0" shape = [768, 768, 1, 1] dtype = "float32" min_val = float("-0.0787519") @@ -4092,6 +4464,7 @@ class Program_weight_tensor_parameter_371: class Program_weight_tensor_parameter_372: name = "parameter_372" + original_name = "batch_norm2d_71.b_0" shape = [384] dtype = "float32" min_val = float("-1.77402") @@ -4103,6 +4476,7 @@ class Program_weight_tensor_parameter_372: class Program_weight_tensor_parameter_373: name = "parameter_373" + original_name = "batch_norm2d_71.w_0" shape = [384] dtype = "float32" min_val = float("0.188523") @@ -4114,6 +4488,7 @@ class Program_weight_tensor_parameter_373: class Program_weight_tensor_parameter_374: name = "parameter_374" + original_name = "batch_norm2d_71.w_2" shape = [384] dtype = "float32" min_val = float("5.24774e-05") @@ -4125,6 +4500,7 @@ class Program_weight_tensor_parameter_374: class Program_weight_tensor_parameter_375: name = "parameter_375" + original_name = "batch_norm2d_71.w_1" shape = [384] dtype = "float32" min_val = float("-0.0915885") @@ -4136,6 +4512,7 @@ class Program_weight_tensor_parameter_375: class Program_weight_tensor_parameter_376: name = "parameter_376" + original_name = "conv2d_74.w_0" shape = [384, 384, 1, 1] dtype = "float32" min_val = float("-0.020214") @@ -4147,6 +4524,7 @@ class Program_weight_tensor_parameter_376: class Program_weight_tensor_parameter_377: name = "parameter_377" + original_name = "batch_norm2d_70.b_0" shape = [384] dtype = "float32" min_val = float("-1.77402") @@ -4158,6 +4536,7 @@ class Program_weight_tensor_parameter_377: class Program_weight_tensor_parameter_378: name = "parameter_378" + original_name = "batch_norm2d_70.w_0" shape = [384] dtype = "float32" min_val = float("0.334653") @@ -4169,6 +4548,7 @@ class Program_weight_tensor_parameter_378: class Program_weight_tensor_parameter_379: name = "parameter_379" + original_name = "batch_norm2d_70.w_2" shape = [384] dtype = "float32" min_val = float("0.000614651") @@ -4180,6 +4560,7 @@ class Program_weight_tensor_parameter_379: class Program_weight_tensor_parameter_380: name = "parameter_380" + original_name = "batch_norm2d_70.w_1" shape = [384] dtype = "float32" min_val = float("-0.228713") @@ -4191,6 +4572,7 @@ class Program_weight_tensor_parameter_380: class Program_weight_tensor_parameter_381: name = "parameter_381" + original_name = "conv2d_73.w_0" shape = [384, 384, 3, 3] dtype = "float32" min_val = float("-0.0190584") @@ -4202,6 +4584,7 @@ class Program_weight_tensor_parameter_381: class Program_weight_tensor_parameter_382: name = "parameter_382" + original_name = "batch_norm2d_69.b_0" shape = [384] dtype = "float32" min_val = float("-2.58225") @@ -4213,6 +4596,7 @@ class Program_weight_tensor_parameter_382: class Program_weight_tensor_parameter_383: name = "parameter_383" + original_name = "batch_norm2d_69.w_0" shape = [384] dtype = "float32" min_val = float("0.52002") @@ -4224,6 +4608,7 @@ class Program_weight_tensor_parameter_383: class Program_weight_tensor_parameter_384: name = "parameter_384" + original_name = "batch_norm2d_69.w_2" shape = [384] dtype = "float32" min_val = float("0.0409059") @@ -4235,6 +4620,7 @@ class Program_weight_tensor_parameter_384: class Program_weight_tensor_parameter_385: name = "parameter_385" + original_name = "batch_norm2d_69.w_1" shape = [384] dtype = "float32" min_val = float("-0.904392") @@ -4246,6 +4632,7 @@ class Program_weight_tensor_parameter_385: class Program_weight_tensor_parameter_386: name = "parameter_386" + original_name = "conv2d_72.w_0" shape = [384, 384, 3, 3] dtype = "float32" min_val = float("-0.0213207") @@ -4257,6 +4644,7 @@ class Program_weight_tensor_parameter_386: class Program_weight_tensor_parameter_387: name = "parameter_387" + original_name = "batch_norm2d_68.b_0" shape = [384] dtype = "float32" min_val = float("-1.93927") @@ -4268,6 +4656,7 @@ class Program_weight_tensor_parameter_387: class Program_weight_tensor_parameter_388: name = "parameter_388" + original_name = "batch_norm2d_68.w_0" shape = [384] dtype = "float32" min_val = float("0.163873") @@ -4279,6 +4668,7 @@ class Program_weight_tensor_parameter_388: class Program_weight_tensor_parameter_389: name = "parameter_389" + original_name = "batch_norm2d_68.w_2" shape = [384] dtype = "float32" min_val = float("7.74518e-05") @@ -4290,6 +4680,7 @@ class Program_weight_tensor_parameter_389: class Program_weight_tensor_parameter_390: name = "parameter_390" + original_name = "batch_norm2d_68.w_1" shape = [384] dtype = "float32" min_val = float("-0.0472137") @@ -4301,6 +4692,7 @@ class Program_weight_tensor_parameter_390: class Program_weight_tensor_parameter_391: name = "parameter_391" + original_name = "conv2d_71.w_0" shape = [384, 384, 1, 1] dtype = "float32" min_val = float("-0.0246209") @@ -4312,6 +4704,7 @@ class Program_weight_tensor_parameter_391: class Program_weight_tensor_parameter_392: name = "parameter_392" + original_name = "batch_norm2d_67.b_0" shape = [384] dtype = "float32" min_val = float("-1.93932") @@ -4323,6 +4716,7 @@ class Program_weight_tensor_parameter_392: class Program_weight_tensor_parameter_393: name = "parameter_393" + original_name = "batch_norm2d_67.w_0" shape = [384] dtype = "float32" min_val = float("0.58315") @@ -4334,6 +4728,7 @@ class Program_weight_tensor_parameter_393: class Program_weight_tensor_parameter_394: name = "parameter_394" + original_name = "batch_norm2d_67.w_2" shape = [384] dtype = "float32" min_val = float("0.00138889") @@ -4345,6 +4740,7 @@ class Program_weight_tensor_parameter_394: class Program_weight_tensor_parameter_395: name = "parameter_395" + original_name = "batch_norm2d_67.w_1" shape = [384] dtype = "float32" min_val = float("-0.082843") @@ -4356,6 +4752,7 @@ class Program_weight_tensor_parameter_395: class Program_weight_tensor_parameter_396: name = "parameter_396" + original_name = "conv2d_70.w_0" shape = [384, 384, 3, 3] dtype = "float32" min_val = float("-0.017236") @@ -4367,6 +4764,7 @@ class Program_weight_tensor_parameter_396: class Program_weight_tensor_parameter_397: name = "parameter_397" + original_name = "batch_norm2d_66.b_0" shape = [384] dtype = "float32" min_val = float("-2.39591") @@ -4378,6 +4776,7 @@ class Program_weight_tensor_parameter_397: class Program_weight_tensor_parameter_398: name = "parameter_398" + original_name = "batch_norm2d_66.w_0" shape = [384] dtype = "float32" min_val = float("0.453112") @@ -4389,6 +4788,7 @@ class Program_weight_tensor_parameter_398: class Program_weight_tensor_parameter_399: name = "parameter_399" + original_name = "batch_norm2d_66.w_2" shape = [384] dtype = "float32" min_val = float("0.0310517") @@ -4400,6 +4800,7 @@ class Program_weight_tensor_parameter_399: class Program_weight_tensor_parameter_400: name = "parameter_400" + original_name = "batch_norm2d_66.w_1" shape = [384] dtype = "float32" min_val = float("-0.74615") @@ -4411,6 +4812,7 @@ class Program_weight_tensor_parameter_400: class Program_weight_tensor_parameter_401: name = "parameter_401" + original_name = "conv2d_69.w_0" shape = [384, 384, 3, 3] dtype = "float32" min_val = float("-0.0259567") @@ -4422,6 +4824,7 @@ class Program_weight_tensor_parameter_401: class Program_weight_tensor_parameter_402: name = "parameter_402" + original_name = "batch_norm2d_65.b_0" shape = [384] dtype = "float32" min_val = float("-1.8762") @@ -4433,6 +4836,7 @@ class Program_weight_tensor_parameter_402: class Program_weight_tensor_parameter_403: name = "parameter_403" + original_name = "batch_norm2d_65.w_0" shape = [384] dtype = "float32" min_val = float("0.0773354") @@ -4444,6 +4848,7 @@ class Program_weight_tensor_parameter_403: class Program_weight_tensor_parameter_404: name = "parameter_404" + original_name = "batch_norm2d_65.w_2" shape = [384] dtype = "float32" min_val = float("6.06445e-05") @@ -4455,6 +4860,7 @@ class Program_weight_tensor_parameter_404: class Program_weight_tensor_parameter_405: name = "parameter_405" + original_name = "batch_norm2d_65.w_1" shape = [384] dtype = "float32" min_val = float("-0.0476186") @@ -4466,6 +4872,7 @@ class Program_weight_tensor_parameter_405: class Program_weight_tensor_parameter_406: name = "parameter_406" + original_name = "conv2d_68.w_0" shape = [384, 384, 1, 1] dtype = "float32" min_val = float("-0.0207296") @@ -4477,6 +4884,7 @@ class Program_weight_tensor_parameter_406: class Program_weight_tensor_parameter_407: name = "parameter_407" + original_name = "batch_norm2d_64.b_0" shape = [384] dtype = "float32" min_val = float("-1.87654") @@ -4488,6 +4896,7 @@ class Program_weight_tensor_parameter_407: class Program_weight_tensor_parameter_408: name = "parameter_408" + original_name = "batch_norm2d_64.w_0" shape = [384] dtype = "float32" min_val = float("0.521871") @@ -4499,6 +4908,7 @@ class Program_weight_tensor_parameter_408: class Program_weight_tensor_parameter_409: name = "parameter_409" + original_name = "batch_norm2d_64.w_2" shape = [384] dtype = "float32" min_val = float("0.00183356") @@ -4510,6 +4920,7 @@ class Program_weight_tensor_parameter_409: class Program_weight_tensor_parameter_410: name = "parameter_410" + original_name = "batch_norm2d_64.w_1" shape = [384] dtype = "float32" min_val = float("-0.210488") @@ -4521,6 +4932,7 @@ class Program_weight_tensor_parameter_410: class Program_weight_tensor_parameter_411: name = "parameter_411" + original_name = "conv2d_67.w_0" shape = [384, 384, 3, 3] dtype = "float32" min_val = float("-0.0177497") @@ -4532,6 +4944,7 @@ class Program_weight_tensor_parameter_411: class Program_weight_tensor_parameter_412: name = "parameter_412" + original_name = "batch_norm2d_63.b_0" shape = [384] dtype = "float32" min_val = float("-2.15635") @@ -4543,6 +4956,7 @@ class Program_weight_tensor_parameter_412: class Program_weight_tensor_parameter_413: name = "parameter_413" + original_name = "batch_norm2d_63.w_0" shape = [384] dtype = "float32" min_val = float("0.706134") @@ -4554,6 +4968,7 @@ class Program_weight_tensor_parameter_413: class Program_weight_tensor_parameter_414: name = "parameter_414" + original_name = "batch_norm2d_63.w_2" shape = [384] dtype = "float32" min_val = float("0.0221089") @@ -4565,6 +4980,7 @@ class Program_weight_tensor_parameter_414: class Program_weight_tensor_parameter_415: name = "parameter_415" + original_name = "batch_norm2d_63.w_1" shape = [384] dtype = "float32" min_val = float("-0.690683") @@ -4576,6 +4992,7 @@ class Program_weight_tensor_parameter_415: class Program_weight_tensor_parameter_416: name = "parameter_416" + original_name = "conv2d_66.w_0" shape = [384, 384, 3, 3] dtype = "float32" min_val = float("-0.0274071") @@ -4587,6 +5004,7 @@ class Program_weight_tensor_parameter_416: class Program_weight_tensor_parameter_417: name = "parameter_417" + original_name = "batch_norm2d_62.b_0" shape = [384] dtype = "float32" min_val = float("-2.9232") @@ -4598,6 +5016,7 @@ class Program_weight_tensor_parameter_417: class Program_weight_tensor_parameter_418: name = "parameter_418" + original_name = "batch_norm2d_62.w_0" shape = [384] dtype = "float32" min_val = float("0.953224") @@ -4609,6 +5028,7 @@ class Program_weight_tensor_parameter_418: class Program_weight_tensor_parameter_419: name = "parameter_419" + original_name = "batch_norm2d_62.w_2" shape = [384] dtype = "float32" min_val = float("0.00282756") @@ -4620,6 +5040,7 @@ class Program_weight_tensor_parameter_419: class Program_weight_tensor_parameter_420: name = "parameter_420" + original_name = "batch_norm2d_62.w_1" shape = [384] dtype = "float32" min_val = float("-0.250212") @@ -4631,6 +5052,7 @@ class Program_weight_tensor_parameter_420: class Program_weight_tensor_parameter_421: name = "parameter_421" + original_name = "conv2d_65.w_0" shape = [384, 768, 1, 1] dtype = "float32" min_val = float("-0.0371909") @@ -4642,6 +5064,7 @@ class Program_weight_tensor_parameter_421: class Program_weight_tensor_parameter_422: name = "parameter_422" + original_name = "batch_norm2d_61.b_0" shape = [384] dtype = "float32" min_val = float("-2.2471") @@ -4653,6 +5076,7 @@ class Program_weight_tensor_parameter_422: class Program_weight_tensor_parameter_423: name = "parameter_423" + original_name = "batch_norm2d_61.w_0" shape = [384] dtype = "float32" min_val = float("0.965853") @@ -4664,6 +5088,7 @@ class Program_weight_tensor_parameter_423: class Program_weight_tensor_parameter_424: name = "parameter_424" + original_name = "batch_norm2d_61.w_2" shape = [384] dtype = "float32" min_val = float("0.000815531") @@ -4675,6 +5100,7 @@ class Program_weight_tensor_parameter_424: class Program_weight_tensor_parameter_425: name = "parameter_425" + original_name = "batch_norm2d_61.w_1" shape = [384] dtype = "float32" min_val = float("-0.0161045") @@ -4686,6 +5112,7 @@ class Program_weight_tensor_parameter_425: class Program_weight_tensor_parameter_426: name = "parameter_426" + original_name = "conv2d_64.w_0" shape = [384, 768, 1, 1] dtype = "float32" min_val = float("-0.0815437") @@ -4697,6 +5124,7 @@ class Program_weight_tensor_parameter_426: class Program_weight_tensor_parameter_427: name = "parameter_427" + original_name = "batch_norm2d_60.b_0" shape = [768] dtype = "float32" min_val = float("-2.40199") @@ -4708,6 +5136,7 @@ class Program_weight_tensor_parameter_427: class Program_weight_tensor_parameter_428: name = "parameter_428" + original_name = "batch_norm2d_60.w_0" shape = [768] dtype = "float32" min_val = float("0.530297") @@ -4719,6 +5148,7 @@ class Program_weight_tensor_parameter_428: class Program_weight_tensor_parameter_429: name = "parameter_429" + original_name = "batch_norm2d_60.w_2" shape = [768] dtype = "float32" min_val = float("0.00640934") @@ -4730,6 +5160,7 @@ class Program_weight_tensor_parameter_429: class Program_weight_tensor_parameter_430: name = "parameter_430" + original_name = "batch_norm2d_60.w_1" shape = [768] dtype = "float32" min_val = float("-0.235794") @@ -4741,6 +5172,7 @@ class Program_weight_tensor_parameter_430: class Program_weight_tensor_parameter_431: name = "parameter_431" + original_name = "conv2d_63.w_0" shape = [768, 512, 3, 3] dtype = "float32" min_val = float("-0.0378314") @@ -4752,6 +5184,7 @@ class Program_weight_tensor_parameter_431: class Program_weight_tensor_parameter_432: name = "parameter_432" + original_name = "batch_norm2d_59.b_0" shape = [512] dtype = "float32" min_val = float("-3.38998") @@ -4763,6 +5196,7 @@ class Program_weight_tensor_parameter_432: class Program_weight_tensor_parameter_433: name = "parameter_433" + original_name = "batch_norm2d_59.w_0" shape = [512] dtype = "float32" min_val = float("0.523767") @@ -4774,6 +5208,7 @@ class Program_weight_tensor_parameter_433: class Program_weight_tensor_parameter_434: name = "parameter_434" + original_name = "batch_norm2d_59.w_2" shape = [512] dtype = "float32" min_val = float("0.00248322") @@ -4785,6 +5220,7 @@ class Program_weight_tensor_parameter_434: class Program_weight_tensor_parameter_435: name = "parameter_435" + original_name = "batch_norm2d_59.w_1" shape = [512] dtype = "float32" min_val = float("-0.172258") @@ -4796,6 +5232,7 @@ class Program_weight_tensor_parameter_435: class Program_weight_tensor_parameter_436: name = "parameter_436" + original_name = "conv2d_62.w_0" shape = [512, 384, 1, 1] dtype = "float32" min_val = float("-0.202262") @@ -4807,6 +5244,7 @@ class Program_weight_tensor_parameter_436: class Program_weight_tensor_parameter_437: name = "parameter_437" + original_name = "conv2d_61.b_0" shape = [384] dtype = "float32" min_val = float("-0.0100703") @@ -4818,6 +5256,7 @@ class Program_weight_tensor_parameter_437: class Program_weight_tensor_parameter_438: name = "parameter_438" + original_name = "conv2d_61.w_0" shape = [384, 384, 1, 1] dtype = "float32" min_val = float("-0.202729") @@ -4829,6 +5268,7 @@ class Program_weight_tensor_parameter_438: class Program_weight_tensor_parameter_439: name = "parameter_439" + original_name = "batch_norm2d_58.b_0" shape = [192] dtype = "float32" min_val = float("-1.97045") @@ -4840,6 +5280,7 @@ class Program_weight_tensor_parameter_439: class Program_weight_tensor_parameter_440: name = "parameter_440" + original_name = "batch_norm2d_58.w_0" shape = [192] dtype = "float32" min_val = float("0.0528864") @@ -4851,6 +5292,7 @@ class Program_weight_tensor_parameter_440: class Program_weight_tensor_parameter_441: name = "parameter_441" + original_name = "batch_norm2d_58.w_2" shape = [192] dtype = "float32" min_val = float("9.0619e-05") @@ -4862,6 +5304,7 @@ class Program_weight_tensor_parameter_441: class Program_weight_tensor_parameter_442: name = "parameter_442" + original_name = "batch_norm2d_58.w_1" shape = [192] dtype = "float32" min_val = float("-0.0346181") @@ -4873,6 +5316,7 @@ class Program_weight_tensor_parameter_442: class Program_weight_tensor_parameter_443: name = "parameter_443" + original_name = "conv2d_60.w_0" shape = [192, 192, 1, 1] dtype = "float32" min_val = float("-0.023487") @@ -4884,6 +5328,7 @@ class Program_weight_tensor_parameter_443: class Program_weight_tensor_parameter_444: name = "parameter_444" + original_name = "batch_norm2d_57.b_0" shape = [192] dtype = "float32" min_val = float("-1.97037") @@ -4895,6 +5340,7 @@ class Program_weight_tensor_parameter_444: class Program_weight_tensor_parameter_445: name = "parameter_445" + original_name = "batch_norm2d_57.w_0" shape = [192] dtype = "float32" min_val = float("0.372338") @@ -4906,6 +5352,7 @@ class Program_weight_tensor_parameter_445: class Program_weight_tensor_parameter_446: name = "parameter_446" + original_name = "batch_norm2d_57.w_2" shape = [192] dtype = "float32" min_val = float("0.00127295") @@ -4917,6 +5364,7 @@ class Program_weight_tensor_parameter_446: class Program_weight_tensor_parameter_447: name = "parameter_447" + original_name = "batch_norm2d_57.w_1" shape = [192] dtype = "float32" min_val = float("-0.0977349") @@ -4928,6 +5376,7 @@ class Program_weight_tensor_parameter_447: class Program_weight_tensor_parameter_448: name = "parameter_448" + original_name = "conv2d_59.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.0289902") @@ -4939,6 +5388,7 @@ class Program_weight_tensor_parameter_448: class Program_weight_tensor_parameter_449: name = "parameter_449" + original_name = "batch_norm2d_56.b_0" shape = [192] dtype = "float32" min_val = float("-2.89065") @@ -4950,6 +5400,7 @@ class Program_weight_tensor_parameter_449: class Program_weight_tensor_parameter_450: name = "parameter_450" + original_name = "batch_norm2d_56.w_0" shape = [192] dtype = "float32" min_val = float("0.696524") @@ -4961,6 +5412,7 @@ class Program_weight_tensor_parameter_450: class Program_weight_tensor_parameter_451: name = "parameter_451" + original_name = "batch_norm2d_56.w_2" shape = [192] dtype = "float32" min_val = float("0.0632461") @@ -4972,6 +5424,7 @@ class Program_weight_tensor_parameter_451: class Program_weight_tensor_parameter_452: name = "parameter_452" + original_name = "batch_norm2d_56.w_1" shape = [192] dtype = "float32" min_val = float("-2.50976") @@ -4983,6 +5436,7 @@ class Program_weight_tensor_parameter_452: class Program_weight_tensor_parameter_453: name = "parameter_453" + original_name = "conv2d_58.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.0331927") @@ -4994,6 +5448,7 @@ class Program_weight_tensor_parameter_453: class Program_weight_tensor_parameter_454: name = "parameter_454" + original_name = "batch_norm2d_55.b_0" shape = [192] dtype = "float32" min_val = float("-1.9404") @@ -5005,6 +5460,7 @@ class Program_weight_tensor_parameter_454: class Program_weight_tensor_parameter_455: name = "parameter_455" + original_name = "batch_norm2d_55.w_0" shape = [192] dtype = "float32" min_val = float("0.0454025") @@ -5016,6 +5472,7 @@ class Program_weight_tensor_parameter_455: class Program_weight_tensor_parameter_456: name = "parameter_456" + original_name = "batch_norm2d_55.w_2" shape = [192] dtype = "float32" min_val = float("7.46909e-05") @@ -5027,6 +5484,7 @@ class Program_weight_tensor_parameter_456: class Program_weight_tensor_parameter_457: name = "parameter_457" + original_name = "batch_norm2d_55.w_1" shape = [192] dtype = "float32" min_val = float("-0.0293086") @@ -5038,6 +5496,7 @@ class Program_weight_tensor_parameter_457: class Program_weight_tensor_parameter_458: name = "parameter_458" + original_name = "conv2d_57.w_0" shape = [192, 192, 1, 1] dtype = "float32" min_val = float("-0.0234926") @@ -5049,6 +5508,7 @@ class Program_weight_tensor_parameter_458: class Program_weight_tensor_parameter_459: name = "parameter_459" + original_name = "batch_norm2d_54.b_0" shape = [192] dtype = "float32" min_val = float("-1.94044") @@ -5060,6 +5520,7 @@ class Program_weight_tensor_parameter_459: class Program_weight_tensor_parameter_460: name = "parameter_460" + original_name = "batch_norm2d_54.w_0" shape = [192] dtype = "float32" min_val = float("0.483074") @@ -5071,6 +5532,7 @@ class Program_weight_tensor_parameter_460: class Program_weight_tensor_parameter_461: name = "parameter_461" + original_name = "batch_norm2d_54.w_2" shape = [192] dtype = "float32" min_val = float("0.00274472") @@ -5082,6 +5544,7 @@ class Program_weight_tensor_parameter_461: class Program_weight_tensor_parameter_462: name = "parameter_462" + original_name = "batch_norm2d_54.w_1" shape = [192] dtype = "float32" min_val = float("-0.0926083") @@ -5093,6 +5556,7 @@ class Program_weight_tensor_parameter_462: class Program_weight_tensor_parameter_463: name = "parameter_463" + original_name = "conv2d_56.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.0231072") @@ -5104,6 +5568,7 @@ class Program_weight_tensor_parameter_463: class Program_weight_tensor_parameter_464: name = "parameter_464" + original_name = "batch_norm2d_53.b_0" shape = [192] dtype = "float32" min_val = float("-2.50828") @@ -5115,6 +5580,7 @@ class Program_weight_tensor_parameter_464: class Program_weight_tensor_parameter_465: name = "parameter_465" + original_name = "batch_norm2d_53.w_0" shape = [192] dtype = "float32" min_val = float("0.65494") @@ -5126,6 +5592,7 @@ class Program_weight_tensor_parameter_465: class Program_weight_tensor_parameter_466: name = "parameter_466" + original_name = "batch_norm2d_53.w_2" shape = [192] dtype = "float32" min_val = float("0.0467958") @@ -5137,6 +5604,7 @@ class Program_weight_tensor_parameter_466: class Program_weight_tensor_parameter_467: name = "parameter_467" + original_name = "batch_norm2d_53.w_1" shape = [192] dtype = "float32" min_val = float("-2.14487") @@ -5148,6 +5616,7 @@ class Program_weight_tensor_parameter_467: class Program_weight_tensor_parameter_468: name = "parameter_468" + original_name = "conv2d_55.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.0362254") @@ -5159,6 +5628,7 @@ class Program_weight_tensor_parameter_468: class Program_weight_tensor_parameter_469: name = "parameter_469" + original_name = "batch_norm2d_52.b_0" shape = [192] dtype = "float32" min_val = float("-1.7573") @@ -5170,6 +5640,7 @@ class Program_weight_tensor_parameter_469: class Program_weight_tensor_parameter_470: name = "parameter_470" + original_name = "batch_norm2d_52.w_0" shape = [192] dtype = "float32" min_val = float("0.00295124") @@ -5181,6 +5652,7 @@ class Program_weight_tensor_parameter_470: class Program_weight_tensor_parameter_471: name = "parameter_471" + original_name = "batch_norm2d_52.w_2" shape = [192] dtype = "float32" min_val = float("9.28523e-07") @@ -5192,6 +5664,7 @@ class Program_weight_tensor_parameter_471: class Program_weight_tensor_parameter_472: name = "parameter_472" + original_name = "batch_norm2d_52.w_1" shape = [192] dtype = "float32" min_val = float("-0.0373738") @@ -5203,6 +5676,7 @@ class Program_weight_tensor_parameter_472: class Program_weight_tensor_parameter_473: name = "parameter_473" + original_name = "conv2d_54.w_0" shape = [192, 192, 1, 1] dtype = "float32" min_val = float("-0.0303466") @@ -5214,6 +5688,7 @@ class Program_weight_tensor_parameter_473: class Program_weight_tensor_parameter_474: name = "parameter_474" + original_name = "batch_norm2d_51.b_0" shape = [192] dtype = "float32" min_val = float("-1.7573") @@ -5225,6 +5700,7 @@ class Program_weight_tensor_parameter_474: class Program_weight_tensor_parameter_475: name = "parameter_475" + original_name = "batch_norm2d_51.w_0" shape = [192] dtype = "float32" min_val = float("0.406102") @@ -5236,6 +5712,7 @@ class Program_weight_tensor_parameter_475: class Program_weight_tensor_parameter_476: name = "parameter_476" + original_name = "batch_norm2d_51.w_2" shape = [192] dtype = "float32" min_val = float("0.0026697") @@ -5247,6 +5724,7 @@ class Program_weight_tensor_parameter_476: class Program_weight_tensor_parameter_477: name = "parameter_477" + original_name = "batch_norm2d_51.w_1" shape = [192] dtype = "float32" min_val = float("-0.0636079") @@ -5258,6 +5736,7 @@ class Program_weight_tensor_parameter_477: class Program_weight_tensor_parameter_478: name = "parameter_478" + original_name = "conv2d_53.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.0321474") @@ -5269,6 +5748,7 @@ class Program_weight_tensor_parameter_478: class Program_weight_tensor_parameter_479: name = "parameter_479" + original_name = "batch_norm2d_50.b_0" shape = [192] dtype = "float32" min_val = float("-2.49735") @@ -5280,6 +5760,7 @@ class Program_weight_tensor_parameter_479: class Program_weight_tensor_parameter_480: name = "parameter_480" + original_name = "batch_norm2d_50.w_0" shape = [192] dtype = "float32" min_val = float("0.652126") @@ -5291,6 +5772,7 @@ class Program_weight_tensor_parameter_480: class Program_weight_tensor_parameter_481: name = "parameter_481" + original_name = "batch_norm2d_50.w_2" shape = [192] dtype = "float32" min_val = float("0.0309664") @@ -5302,6 +5784,7 @@ class Program_weight_tensor_parameter_481: class Program_weight_tensor_parameter_482: name = "parameter_482" + original_name = "batch_norm2d_50.w_1" shape = [192] dtype = "float32" min_val = float("-1.51706") @@ -5313,6 +5796,7 @@ class Program_weight_tensor_parameter_482: class Program_weight_tensor_parameter_483: name = "parameter_483" + original_name = "conv2d_52.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.05013") @@ -5324,6 +5808,7 @@ class Program_weight_tensor_parameter_483: class Program_weight_tensor_parameter_484: name = "parameter_484" + original_name = "batch_norm2d_49.b_0" shape = [192] dtype = "float32" min_val = float("-2.07916") @@ -5335,6 +5820,7 @@ class Program_weight_tensor_parameter_484: class Program_weight_tensor_parameter_485: name = "parameter_485" + original_name = "batch_norm2d_49.w_0" shape = [192] dtype = "float32" min_val = float("0.000510371") @@ -5346,6 +5832,7 @@ class Program_weight_tensor_parameter_485: class Program_weight_tensor_parameter_486: name = "parameter_486" + original_name = "batch_norm2d_49.w_2" shape = [192] dtype = "float32" min_val = float("6.2328e-08") @@ -5357,6 +5844,7 @@ class Program_weight_tensor_parameter_486: class Program_weight_tensor_parameter_487: name = "parameter_487" + original_name = "batch_norm2d_49.w_1" shape = [192] dtype = "float32" min_val = float("-0.0197338") @@ -5368,6 +5856,7 @@ class Program_weight_tensor_parameter_487: class Program_weight_tensor_parameter_488: name = "parameter_488" + original_name = "conv2d_51.w_0" shape = [192, 192, 1, 1] dtype = "float32" min_val = float("-0.0202783") @@ -5379,6 +5868,7 @@ class Program_weight_tensor_parameter_488: class Program_weight_tensor_parameter_489: name = "parameter_489" + original_name = "batch_norm2d_48.b_0" shape = [192] dtype = "float32" min_val = float("-2.07922") @@ -5390,6 +5880,7 @@ class Program_weight_tensor_parameter_489: class Program_weight_tensor_parameter_490: name = "parameter_490" + original_name = "batch_norm2d_48.w_0" shape = [192] dtype = "float32" min_val = float("0.396505") @@ -5401,6 +5892,7 @@ class Program_weight_tensor_parameter_490: class Program_weight_tensor_parameter_491: name = "parameter_491" + original_name = "batch_norm2d_48.w_2" shape = [192] dtype = "float32" min_val = float("0.00305567") @@ -5412,6 +5904,7 @@ class Program_weight_tensor_parameter_491: class Program_weight_tensor_parameter_492: name = "parameter_492" + original_name = "batch_norm2d_48.w_1" shape = [192] dtype = "float32" min_val = float("-0.0910836") @@ -5423,6 +5916,7 @@ class Program_weight_tensor_parameter_492: class Program_weight_tensor_parameter_493: name = "parameter_493" + original_name = "conv2d_50.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.0299549") @@ -5434,6 +5928,7 @@ class Program_weight_tensor_parameter_493: class Program_weight_tensor_parameter_494: name = "parameter_494" + original_name = "batch_norm2d_47.b_0" shape = [192] dtype = "float32" min_val = float("-2.74084") @@ -5445,6 +5940,7 @@ class Program_weight_tensor_parameter_494: class Program_weight_tensor_parameter_495: name = "parameter_495" + original_name = "batch_norm2d_47.w_0" shape = [192] dtype = "float32" min_val = float("0.761623") @@ -5456,6 +5952,7 @@ class Program_weight_tensor_parameter_495: class Program_weight_tensor_parameter_496: name = "parameter_496" + original_name = "batch_norm2d_47.w_2" shape = [192] dtype = "float32" min_val = float("0.0272966") @@ -5467,6 +5964,7 @@ class Program_weight_tensor_parameter_496: class Program_weight_tensor_parameter_497: name = "parameter_497" + original_name = "batch_norm2d_47.w_1" shape = [192] dtype = "float32" min_val = float("-1.23827") @@ -5478,6 +5976,7 @@ class Program_weight_tensor_parameter_497: class Program_weight_tensor_parameter_498: name = "parameter_498" + original_name = "conv2d_49.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.0531238") @@ -5489,6 +5988,7 @@ class Program_weight_tensor_parameter_498: class Program_weight_tensor_parameter_499: name = "parameter_499" + original_name = "batch_norm2d_46.b_0" shape = [192] dtype = "float32" min_val = float("-1.21219") @@ -5500,6 +6000,7 @@ class Program_weight_tensor_parameter_499: class Program_weight_tensor_parameter_500: name = "parameter_500" + original_name = "batch_norm2d_46.w_0" shape = [192] dtype = "float32" min_val = float("-9.82711e-05") @@ -5511,6 +6012,7 @@ class Program_weight_tensor_parameter_500: class Program_weight_tensor_parameter_501: name = "parameter_501" + original_name = "batch_norm2d_46.w_2" shape = [192] dtype = "float32" min_val = float("2.25949e-10") @@ -5522,6 +6024,7 @@ class Program_weight_tensor_parameter_501: class Program_weight_tensor_parameter_502: name = "parameter_502" + original_name = "batch_norm2d_46.w_1" shape = [192] dtype = "float32" min_val = float("-0.0494718") @@ -5533,6 +6036,7 @@ class Program_weight_tensor_parameter_502: class Program_weight_tensor_parameter_503: name = "parameter_503" + original_name = "conv2d_48.w_0" shape = [192, 192, 1, 1] dtype = "float32" min_val = float("-0.0342199") @@ -5544,6 +6048,7 @@ class Program_weight_tensor_parameter_503: class Program_weight_tensor_parameter_504: name = "parameter_504" + original_name = "batch_norm2d_45.b_0" shape = [192] dtype = "float32" min_val = float("-1.21223") @@ -5555,6 +6060,7 @@ class Program_weight_tensor_parameter_504: class Program_weight_tensor_parameter_505: name = "parameter_505" + original_name = "batch_norm2d_45.w_0" shape = [192] dtype = "float32" min_val = float("0.382831") @@ -5566,6 +6072,7 @@ class Program_weight_tensor_parameter_505: class Program_weight_tensor_parameter_506: name = "parameter_506" + original_name = "batch_norm2d_45.w_2" shape = [192] dtype = "float32" min_val = float("0.00222243") @@ -5577,6 +6084,7 @@ class Program_weight_tensor_parameter_506: class Program_weight_tensor_parameter_507: name = "parameter_507" + original_name = "batch_norm2d_45.w_1" shape = [192] dtype = "float32" min_val = float("-0.0846332") @@ -5588,6 +6096,7 @@ class Program_weight_tensor_parameter_507: class Program_weight_tensor_parameter_508: name = "parameter_508" + original_name = "conv2d_47.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.0323048") @@ -5599,6 +6108,7 @@ class Program_weight_tensor_parameter_508: class Program_weight_tensor_parameter_509: name = "parameter_509" + original_name = "batch_norm2d_44.b_0" shape = [192] dtype = "float32" min_val = float("-2.48701") @@ -5610,6 +6120,7 @@ class Program_weight_tensor_parameter_509: class Program_weight_tensor_parameter_510: name = "parameter_510" + original_name = "batch_norm2d_44.w_0" shape = [192] dtype = "float32" min_val = float("0.689678") @@ -5621,6 +6132,7 @@ class Program_weight_tensor_parameter_510: class Program_weight_tensor_parameter_511: name = "parameter_511" + original_name = "batch_norm2d_44.w_2" shape = [192] dtype = "float32" min_val = float("0.0185507") @@ -5632,6 +6144,7 @@ class Program_weight_tensor_parameter_511: class Program_weight_tensor_parameter_512: name = "parameter_512" + original_name = "batch_norm2d_44.w_1" shape = [192] dtype = "float32" min_val = float("-0.717377") @@ -5643,6 +6156,7 @@ class Program_weight_tensor_parameter_512: class Program_weight_tensor_parameter_513: name = "parameter_513" + original_name = "conv2d_46.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.0610342") @@ -5654,6 +6168,7 @@ class Program_weight_tensor_parameter_513: class Program_weight_tensor_parameter_514: name = "parameter_514" + original_name = "batch_norm2d_43.b_0" shape = [192] dtype = "float32" min_val = float("-1.21753") @@ -5665,6 +6180,7 @@ class Program_weight_tensor_parameter_514: class Program_weight_tensor_parameter_515: name = "parameter_515" + original_name = "batch_norm2d_43.w_0" shape = [192] dtype = "float32" min_val = float("0.00836385") @@ -5676,6 +6192,7 @@ class Program_weight_tensor_parameter_515: class Program_weight_tensor_parameter_516: name = "parameter_516" + original_name = "batch_norm2d_43.w_2" shape = [192] dtype = "float32" min_val = float("1.9816e-05") @@ -5687,6 +6204,7 @@ class Program_weight_tensor_parameter_516: class Program_weight_tensor_parameter_517: name = "parameter_517" + original_name = "batch_norm2d_43.w_1" shape = [192] dtype = "float32" min_val = float("-0.0656722") @@ -5698,6 +6216,7 @@ class Program_weight_tensor_parameter_517: class Program_weight_tensor_parameter_518: name = "parameter_518" + original_name = "conv2d_45.w_0" shape = [192, 192, 1, 1] dtype = "float32" min_val = float("-0.0600528") @@ -5709,6 +6228,7 @@ class Program_weight_tensor_parameter_518: class Program_weight_tensor_parameter_519: name = "parameter_519" + original_name = "batch_norm2d_42.b_0" shape = [192] dtype = "float32" min_val = float("-1.21747") @@ -5720,6 +6240,7 @@ class Program_weight_tensor_parameter_519: class Program_weight_tensor_parameter_520: name = "parameter_520" + original_name = "batch_norm2d_42.w_0" shape = [192] dtype = "float32" min_val = float("0.354999") @@ -5731,6 +6252,7 @@ class Program_weight_tensor_parameter_520: class Program_weight_tensor_parameter_521: name = "parameter_521" + original_name = "batch_norm2d_42.w_2" shape = [192] dtype = "float32" min_val = float("0.00437457") @@ -5742,6 +6264,7 @@ class Program_weight_tensor_parameter_521: class Program_weight_tensor_parameter_522: name = "parameter_522" + original_name = "batch_norm2d_42.w_1" shape = [192] dtype = "float32" min_val = float("-0.159743") @@ -5753,6 +6276,7 @@ class Program_weight_tensor_parameter_522: class Program_weight_tensor_parameter_523: name = "parameter_523" + original_name = "conv2d_44.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.062497") @@ -5764,6 +6288,7 @@ class Program_weight_tensor_parameter_523: class Program_weight_tensor_parameter_524: name = "parameter_524" + original_name = "batch_norm2d_41.b_0" shape = [192] dtype = "float32" min_val = float("-1.87905") @@ -5775,6 +6300,7 @@ class Program_weight_tensor_parameter_524: class Program_weight_tensor_parameter_525: name = "parameter_525" + original_name = "batch_norm2d_41.w_0" shape = [192] dtype = "float32" min_val = float("0.788784") @@ -5786,6 +6312,7 @@ class Program_weight_tensor_parameter_525: class Program_weight_tensor_parameter_526: name = "parameter_526" + original_name = "batch_norm2d_41.w_2" shape = [192] dtype = "float32" min_val = float("0.0159247") @@ -5797,6 +6324,7 @@ class Program_weight_tensor_parameter_526: class Program_weight_tensor_parameter_527: name = "parameter_527" + original_name = "batch_norm2d_41.w_1" shape = [192] dtype = "float32" min_val = float("-0.690131") @@ -5808,6 +6336,7 @@ class Program_weight_tensor_parameter_527: class Program_weight_tensor_parameter_528: name = "parameter_528" + original_name = "conv2d_43.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.062874") @@ -5819,6 +6348,7 @@ class Program_weight_tensor_parameter_528: class Program_weight_tensor_parameter_529: name = "parameter_529" + original_name = "batch_norm2d_40.b_0" shape = [192] dtype = "float32" min_val = float("-2.86217") @@ -5830,6 +6360,7 @@ class Program_weight_tensor_parameter_529: class Program_weight_tensor_parameter_530: name = "parameter_530" + original_name = "batch_norm2d_40.w_0" shape = [192] dtype = "float32" min_val = float("0.487672") @@ -5841,6 +6372,7 @@ class Program_weight_tensor_parameter_530: class Program_weight_tensor_parameter_531: name = "parameter_531" + original_name = "batch_norm2d_40.w_2" shape = [192] dtype = "float32" min_val = float("0.00962562") @@ -5852,6 +6384,7 @@ class Program_weight_tensor_parameter_531: class Program_weight_tensor_parameter_532: name = "parameter_532" + original_name = "batch_norm2d_40.w_1" shape = [192] dtype = "float32" min_val = float("-0.230196") @@ -5863,6 +6396,7 @@ class Program_weight_tensor_parameter_532: class Program_weight_tensor_parameter_533: name = "parameter_533" + original_name = "conv2d_42.w_0" shape = [192, 384, 1, 1] dtype = "float32" min_val = float("-0.108831") @@ -5874,6 +6408,7 @@ class Program_weight_tensor_parameter_533: class Program_weight_tensor_parameter_534: name = "parameter_534" + original_name = "batch_norm2d_39.b_0" shape = [192] dtype = "float32" min_val = float("-2.96764") @@ -5885,6 +6420,7 @@ class Program_weight_tensor_parameter_534: class Program_weight_tensor_parameter_535: name = "parameter_535" + original_name = "batch_norm2d_39.w_0" shape = [192] dtype = "float32" min_val = float("0.830791") @@ -5896,6 +6432,7 @@ class Program_weight_tensor_parameter_535: class Program_weight_tensor_parameter_536: name = "parameter_536" + original_name = "batch_norm2d_39.w_2" shape = [192] dtype = "float32" min_val = float("0.00601536") @@ -5907,6 +6444,7 @@ class Program_weight_tensor_parameter_536: class Program_weight_tensor_parameter_537: name = "parameter_537" + original_name = "batch_norm2d_39.w_1" shape = [192] dtype = "float32" min_val = float("-0.133093") @@ -5918,6 +6456,7 @@ class Program_weight_tensor_parameter_537: class Program_weight_tensor_parameter_538: name = "parameter_538" + original_name = "conv2d_41.w_0" shape = [192, 384, 1, 1] dtype = "float32" min_val = float("-0.0985625") @@ -5929,6 +6468,7 @@ class Program_weight_tensor_parameter_538: class Program_weight_tensor_parameter_539: name = "parameter_539" + original_name = "batch_norm2d_38.b_0" shape = [384] dtype = "float32" min_val = float("-2.92359") @@ -5940,6 +6480,7 @@ class Program_weight_tensor_parameter_539: class Program_weight_tensor_parameter_540: name = "parameter_540" + original_name = "batch_norm2d_38.w_0" shape = [384] dtype = "float32" min_val = float("0.631853") @@ -5951,6 +6492,7 @@ class Program_weight_tensor_parameter_540: class Program_weight_tensor_parameter_541: name = "parameter_541" + original_name = "batch_norm2d_38.w_2" shape = [384] dtype = "float32" min_val = float("0.0103628") @@ -5962,6 +6504,7 @@ class Program_weight_tensor_parameter_541: class Program_weight_tensor_parameter_542: name = "parameter_542" + original_name = "batch_norm2d_38.w_1" shape = [384] dtype = "float32" min_val = float("-0.269684") @@ -5973,6 +6516,7 @@ class Program_weight_tensor_parameter_542: class Program_weight_tensor_parameter_543: name = "parameter_543" + original_name = "conv2d_40.w_0" shape = [384, 256, 3, 3] dtype = "float32" min_val = float("-0.0753194") @@ -5984,6 +6528,7 @@ class Program_weight_tensor_parameter_543: class Program_weight_tensor_parameter_544: name = "parameter_544" + original_name = "batch_norm2d_37.b_0" shape = [256] dtype = "float32" min_val = float("-2.04502") @@ -5995,6 +6540,7 @@ class Program_weight_tensor_parameter_544: class Program_weight_tensor_parameter_545: name = "parameter_545" + original_name = "batch_norm2d_37.w_0" shape = [256] dtype = "float32" min_val = float("0.517239") @@ -6006,6 +6552,7 @@ class Program_weight_tensor_parameter_545: class Program_weight_tensor_parameter_546: name = "parameter_546" + original_name = "batch_norm2d_37.w_2" shape = [256] dtype = "float32" min_val = float("0.00196874") @@ -6017,6 +6564,7 @@ class Program_weight_tensor_parameter_546: class Program_weight_tensor_parameter_547: name = "parameter_547" + original_name = "batch_norm2d_37.w_1" shape = [256] dtype = "float32" min_val = float("-0.230499") @@ -6028,6 +6576,7 @@ class Program_weight_tensor_parameter_547: class Program_weight_tensor_parameter_548: name = "parameter_548" + original_name = "conv2d_39.w_0" shape = [256, 192, 1, 1] dtype = "float32" min_val = float("-0.206154") @@ -6039,6 +6588,7 @@ class Program_weight_tensor_parameter_548: class Program_weight_tensor_parameter_549: name = "parameter_549" + original_name = "conv2d_38.b_0" shape = [192] dtype = "float32" min_val = float("-0.0139357") @@ -6050,6 +6600,7 @@ class Program_weight_tensor_parameter_549: class Program_weight_tensor_parameter_550: name = "parameter_550" + original_name = "conv2d_38.w_0" shape = [192, 192, 1, 1] dtype = "float32" min_val = float("-0.347135") @@ -6061,6 +6612,7 @@ class Program_weight_tensor_parameter_550: class Program_weight_tensor_parameter_551: name = "parameter_551" + original_name = "batch_norm2d_36.b_0" shape = [96] dtype = "float32" min_val = float("-1.91355") @@ -6072,6 +6624,7 @@ class Program_weight_tensor_parameter_551: class Program_weight_tensor_parameter_552: name = "parameter_552" + original_name = "batch_norm2d_36.w_0" shape = [96] dtype = "float32" min_val = float("0.142427") @@ -6083,6 +6636,7 @@ class Program_weight_tensor_parameter_552: class Program_weight_tensor_parameter_553: name = "parameter_553" + original_name = "batch_norm2d_36.w_2" shape = [96] dtype = "float32" min_val = float("7.75639e-05") @@ -6094,6 +6648,7 @@ class Program_weight_tensor_parameter_553: class Program_weight_tensor_parameter_554: name = "parameter_554" + original_name = "batch_norm2d_36.w_1" shape = [96] dtype = "float32" min_val = float("-0.0546921") @@ -6105,6 +6660,7 @@ class Program_weight_tensor_parameter_554: class Program_weight_tensor_parameter_555: name = "parameter_555" + original_name = "conv2d_37.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.0500852") @@ -6116,6 +6672,7 @@ class Program_weight_tensor_parameter_555: class Program_weight_tensor_parameter_556: name = "parameter_556" + original_name = "batch_norm2d_35.b_0" shape = [96] dtype = "float32" min_val = float("-1.91314") @@ -6127,6 +6684,7 @@ class Program_weight_tensor_parameter_556: class Program_weight_tensor_parameter_557: name = "parameter_557" + original_name = "batch_norm2d_35.w_0" shape = [96] dtype = "float32" min_val = float("0.343774") @@ -6138,6 +6696,7 @@ class Program_weight_tensor_parameter_557: class Program_weight_tensor_parameter_558: name = "parameter_558" + original_name = "batch_norm2d_35.w_2" shape = [96] dtype = "float32" min_val = float("0.000976934") @@ -6149,6 +6708,7 @@ class Program_weight_tensor_parameter_558: class Program_weight_tensor_parameter_559: name = "parameter_559" + original_name = "batch_norm2d_35.w_1" shape = [96] dtype = "float32" min_val = float("-0.137727") @@ -6160,6 +6720,7 @@ class Program_weight_tensor_parameter_559: class Program_weight_tensor_parameter_560: name = "parameter_560" + original_name = "conv2d_36.w_0" shape = [96, 96, 3, 3] dtype = "float32" min_val = float("-0.0398886") @@ -6171,6 +6732,7 @@ class Program_weight_tensor_parameter_560: class Program_weight_tensor_parameter_561: name = "parameter_561" + original_name = "batch_norm2d_34.b_0" shape = [96] dtype = "float32" min_val = float("-2.46605") @@ -6182,6 +6744,7 @@ class Program_weight_tensor_parameter_561: class Program_weight_tensor_parameter_562: name = "parameter_562" + original_name = "batch_norm2d_34.w_0" shape = [96] dtype = "float32" min_val = float("0.542082") @@ -6193,6 +6756,7 @@ class Program_weight_tensor_parameter_562: class Program_weight_tensor_parameter_563: name = "parameter_563" + original_name = "batch_norm2d_34.w_2" shape = [96] dtype = "float32" min_val = float("0.0406212") @@ -6204,6 +6768,7 @@ class Program_weight_tensor_parameter_563: class Program_weight_tensor_parameter_564: name = "parameter_564" + original_name = "batch_norm2d_34.w_1" shape = [96] dtype = "float32" min_val = float("-2.80804") @@ -6215,6 +6780,7 @@ class Program_weight_tensor_parameter_564: class Program_weight_tensor_parameter_565: name = "parameter_565" + original_name = "conv2d_35.w_0" shape = [96, 96, 3, 3] dtype = "float32" min_val = float("-0.150203") @@ -6226,6 +6792,7 @@ class Program_weight_tensor_parameter_565: class Program_weight_tensor_parameter_566: name = "parameter_566" + original_name = "batch_norm2d_33.b_0" shape = [96] dtype = "float32" min_val = float("-1.38826") @@ -6237,6 +6804,7 @@ class Program_weight_tensor_parameter_566: class Program_weight_tensor_parameter_567: name = "parameter_567" + original_name = "batch_norm2d_33.w_0" shape = [96] dtype = "float32" min_val = float("0.0453402") @@ -6248,6 +6816,7 @@ class Program_weight_tensor_parameter_567: class Program_weight_tensor_parameter_568: name = "parameter_568" + original_name = "batch_norm2d_33.w_2" shape = [96] dtype = "float32" min_val = float("7.68974e-05") @@ -6259,6 +6828,7 @@ class Program_weight_tensor_parameter_568: class Program_weight_tensor_parameter_569: name = "parameter_569" + original_name = "batch_norm2d_33.w_1" shape = [96] dtype = "float32" min_val = float("-0.0484682") @@ -6270,6 +6840,7 @@ class Program_weight_tensor_parameter_569: class Program_weight_tensor_parameter_570: name = "parameter_570" + original_name = "conv2d_34.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.0483138") @@ -6281,6 +6852,7 @@ class Program_weight_tensor_parameter_570: class Program_weight_tensor_parameter_571: name = "parameter_571" + original_name = "batch_norm2d_32.b_0" shape = [96] dtype = "float32" min_val = float("-1.38834") @@ -6292,6 +6864,7 @@ class Program_weight_tensor_parameter_571: class Program_weight_tensor_parameter_572: name = "parameter_572" + original_name = "batch_norm2d_32.w_0" shape = [96] dtype = "float32" min_val = float("0.370504") @@ -6303,6 +6876,7 @@ class Program_weight_tensor_parameter_572: class Program_weight_tensor_parameter_573: name = "parameter_573" + original_name = "batch_norm2d_32.w_2" shape = [96] dtype = "float32" min_val = float("0.00320483") @@ -6314,6 +6888,7 @@ class Program_weight_tensor_parameter_573: class Program_weight_tensor_parameter_574: name = "parameter_574" + original_name = "batch_norm2d_32.w_1" shape = [96] dtype = "float32" min_val = float("-0.0963095") @@ -6325,6 +6900,7 @@ class Program_weight_tensor_parameter_574: class Program_weight_tensor_parameter_575: name = "parameter_575" + original_name = "conv2d_33.w_0" shape = [96, 96, 3, 3] dtype = "float32" min_val = float("-0.058655") @@ -6336,6 +6912,7 @@ class Program_weight_tensor_parameter_575: class Program_weight_tensor_parameter_576: name = "parameter_576" + original_name = "batch_norm2d_31.b_0" shape = [96] dtype = "float32" min_val = float("-3.31955") @@ -6347,6 +6924,7 @@ class Program_weight_tensor_parameter_576: class Program_weight_tensor_parameter_577: name = "parameter_577" + original_name = "batch_norm2d_31.w_0" shape = [96] dtype = "float32" min_val = float("0.473098") @@ -6358,6 +6936,7 @@ class Program_weight_tensor_parameter_577: class Program_weight_tensor_parameter_578: name = "parameter_578" + original_name = "batch_norm2d_31.w_2" shape = [96] dtype = "float32" min_val = float("0.0285476") @@ -6369,6 +6948,7 @@ class Program_weight_tensor_parameter_578: class Program_weight_tensor_parameter_579: name = "parameter_579" + original_name = "batch_norm2d_31.w_1" shape = [96] dtype = "float32" min_val = float("-1.25068") @@ -6380,6 +6960,7 @@ class Program_weight_tensor_parameter_579: class Program_weight_tensor_parameter_580: name = "parameter_580" + original_name = "conv2d_32.w_0" shape = [96, 96, 3, 3] dtype = "float32" min_val = float("-0.147666") @@ -6391,6 +6972,7 @@ class Program_weight_tensor_parameter_580: class Program_weight_tensor_parameter_581: name = "parameter_581" + original_name = "batch_norm2d_30.b_0" shape = [96] dtype = "float32" min_val = float("-1.24956") @@ -6402,6 +6984,7 @@ class Program_weight_tensor_parameter_581: class Program_weight_tensor_parameter_582: name = "parameter_582" + original_name = "batch_norm2d_30.w_0" shape = [96] dtype = "float32" min_val = float("0.0243293") @@ -6413,6 +6996,7 @@ class Program_weight_tensor_parameter_582: class Program_weight_tensor_parameter_583: name = "parameter_583" + original_name = "batch_norm2d_30.w_2" shape = [96] dtype = "float32" min_val = float("6.31792e-05") @@ -6424,6 +7008,7 @@ class Program_weight_tensor_parameter_583: class Program_weight_tensor_parameter_584: name = "parameter_584" + original_name = "batch_norm2d_30.w_1" shape = [96] dtype = "float32" min_val = float("-0.0383061") @@ -6435,6 +7020,7 @@ class Program_weight_tensor_parameter_584: class Program_weight_tensor_parameter_585: name = "parameter_585" + original_name = "conv2d_31.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.0448708") @@ -6446,6 +7032,7 @@ class Program_weight_tensor_parameter_585: class Program_weight_tensor_parameter_586: name = "parameter_586" + original_name = "batch_norm2d_29.b_0" shape = [96] dtype = "float32" min_val = float("-1.24942") @@ -6457,6 +7044,7 @@ class Program_weight_tensor_parameter_586: class Program_weight_tensor_parameter_587: name = "parameter_587" + original_name = "batch_norm2d_29.w_0" shape = [96] dtype = "float32" min_val = float("0.315495") @@ -6468,6 +7056,7 @@ class Program_weight_tensor_parameter_587: class Program_weight_tensor_parameter_588: name = "parameter_588" + original_name = "batch_norm2d_29.w_2" shape = [96] dtype = "float32" min_val = float("0.00339766") @@ -6479,6 +7068,7 @@ class Program_weight_tensor_parameter_588: class Program_weight_tensor_parameter_589: name = "parameter_589" + original_name = "batch_norm2d_29.w_1" shape = [96] dtype = "float32" min_val = float("-0.0545808") @@ -6490,6 +7080,7 @@ class Program_weight_tensor_parameter_589: class Program_weight_tensor_parameter_590: name = "parameter_590" + original_name = "conv2d_30.w_0" shape = [96, 96, 3, 3] dtype = "float32" min_val = float("-0.065253") @@ -6501,6 +7092,7 @@ class Program_weight_tensor_parameter_590: class Program_weight_tensor_parameter_591: name = "parameter_591" + original_name = "batch_norm2d_28.b_0" shape = [96] dtype = "float32" min_val = float("-3.58296") @@ -6512,6 +7104,7 @@ class Program_weight_tensor_parameter_591: class Program_weight_tensor_parameter_592: name = "parameter_592" + original_name = "batch_norm2d_28.w_0" shape = [96] dtype = "float32" min_val = float("0.511106") @@ -6523,6 +7116,7 @@ class Program_weight_tensor_parameter_592: class Program_weight_tensor_parameter_593: name = "parameter_593" + original_name = "batch_norm2d_28.w_2" shape = [96] dtype = "float32" min_val = float("0.0202502") @@ -6534,6 +7128,7 @@ class Program_weight_tensor_parameter_593: class Program_weight_tensor_parameter_594: name = "parameter_594" + original_name = "batch_norm2d_28.w_1" shape = [96] dtype = "float32" min_val = float("-0.823777") @@ -6545,6 +7140,7 @@ class Program_weight_tensor_parameter_594: class Program_weight_tensor_parameter_595: name = "parameter_595" + original_name = "conv2d_29.w_0" shape = [96, 96, 3, 3] dtype = "float32" min_val = float("-0.0973524") @@ -6556,6 +7152,7 @@ class Program_weight_tensor_parameter_595: class Program_weight_tensor_parameter_596: name = "parameter_596" + original_name = "batch_norm2d_27.b_0" shape = [96] dtype = "float32" min_val = float("-0.892064") @@ -6567,6 +7164,7 @@ class Program_weight_tensor_parameter_596: class Program_weight_tensor_parameter_597: name = "parameter_597" + original_name = "batch_norm2d_27.w_0" shape = [96] dtype = "float32" min_val = float("0.0191223") @@ -6578,6 +7176,7 @@ class Program_weight_tensor_parameter_597: class Program_weight_tensor_parameter_598: name = "parameter_598" + original_name = "batch_norm2d_27.w_2" shape = [96] dtype = "float32" min_val = float("4.82579e-05") @@ -6589,6 +7188,7 @@ class Program_weight_tensor_parameter_598: class Program_weight_tensor_parameter_599: name = "parameter_599" + original_name = "batch_norm2d_27.w_1" shape = [96] dtype = "float32" min_val = float("-0.0327526") @@ -6600,6 +7200,7 @@ class Program_weight_tensor_parameter_599: class Program_weight_tensor_parameter_600: name = "parameter_600" + original_name = "conv2d_28.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.0499906") @@ -6611,6 +7212,7 @@ class Program_weight_tensor_parameter_600: class Program_weight_tensor_parameter_601: name = "parameter_601" + original_name = "batch_norm2d_26.b_0" shape = [96] dtype = "float32" min_val = float("-0.891955") @@ -6622,6 +7224,7 @@ class Program_weight_tensor_parameter_601: class Program_weight_tensor_parameter_602: name = "parameter_602" + original_name = "batch_norm2d_26.w_0" shape = [96] dtype = "float32" min_val = float("0.17446") @@ -6633,6 +7236,7 @@ class Program_weight_tensor_parameter_602: class Program_weight_tensor_parameter_603: name = "parameter_603" + original_name = "batch_norm2d_26.w_2" shape = [96] dtype = "float32" min_val = float("0.00236192") @@ -6644,6 +7248,7 @@ class Program_weight_tensor_parameter_603: class Program_weight_tensor_parameter_604: name = "parameter_604" + original_name = "batch_norm2d_26.w_1" shape = [96] dtype = "float32" min_val = float("-0.0582992") @@ -6655,6 +7260,7 @@ class Program_weight_tensor_parameter_604: class Program_weight_tensor_parameter_605: name = "parameter_605" + original_name = "conv2d_27.w_0" shape = [96, 96, 3, 3] dtype = "float32" min_val = float("-0.057305") @@ -6666,6 +7272,7 @@ class Program_weight_tensor_parameter_605: class Program_weight_tensor_parameter_606: name = "parameter_606" + original_name = "batch_norm2d_25.b_0" shape = [96] dtype = "float32" min_val = float("-2.65777") @@ -6677,6 +7284,7 @@ class Program_weight_tensor_parameter_606: class Program_weight_tensor_parameter_607: name = "parameter_607" + original_name = "batch_norm2d_25.w_0" shape = [96] dtype = "float32" min_val = float("0.512951") @@ -6688,6 +7296,7 @@ class Program_weight_tensor_parameter_607: class Program_weight_tensor_parameter_608: name = "parameter_608" + original_name = "batch_norm2d_25.w_2" shape = [96] dtype = "float32" min_val = float("0.0176905") @@ -6699,6 +7308,7 @@ class Program_weight_tensor_parameter_608: class Program_weight_tensor_parameter_609: name = "parameter_609" + original_name = "batch_norm2d_25.w_1" shape = [96] dtype = "float32" min_val = float("-0.762613") @@ -6710,6 +7320,7 @@ class Program_weight_tensor_parameter_609: class Program_weight_tensor_parameter_610: name = "parameter_610" + original_name = "conv2d_26.w_0" shape = [96, 96, 3, 3] dtype = "float32" min_val = float("-0.0738037") @@ -6721,6 +7332,7 @@ class Program_weight_tensor_parameter_610: class Program_weight_tensor_parameter_611: name = "parameter_611" + original_name = "batch_norm2d_24.b_0" shape = [96] dtype = "float32" min_val = float("-0.978262") @@ -6732,6 +7344,7 @@ class Program_weight_tensor_parameter_611: class Program_weight_tensor_parameter_612: name = "parameter_612" + original_name = "batch_norm2d_24.w_0" shape = [96] dtype = "float32" min_val = float("0.0498074") @@ -6743,6 +7356,7 @@ class Program_weight_tensor_parameter_612: class Program_weight_tensor_parameter_613: name = "parameter_613" + original_name = "batch_norm2d_24.w_2" shape = [96] dtype = "float32" min_val = float("0.000185263") @@ -6754,6 +7368,7 @@ class Program_weight_tensor_parameter_613: class Program_weight_tensor_parameter_614: name = "parameter_614" + original_name = "batch_norm2d_24.w_1" shape = [96] dtype = "float32" min_val = float("-0.041706") @@ -6765,6 +7380,7 @@ class Program_weight_tensor_parameter_614: class Program_weight_tensor_parameter_615: name = "parameter_615" + original_name = "conv2d_25.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.0731207") @@ -6776,6 +7392,7 @@ class Program_weight_tensor_parameter_615: class Program_weight_tensor_parameter_616: name = "parameter_616" + original_name = "batch_norm2d_23.b_0" shape = [96] dtype = "float32" min_val = float("-0.978083") @@ -6787,6 +7404,7 @@ class Program_weight_tensor_parameter_616: class Program_weight_tensor_parameter_617: name = "parameter_617" + original_name = "batch_norm2d_23.w_0" shape = [96] dtype = "float32" min_val = float("0.236133") @@ -6798,6 +7416,7 @@ class Program_weight_tensor_parameter_617: class Program_weight_tensor_parameter_618: name = "parameter_618" + original_name = "batch_norm2d_23.w_2" shape = [96] dtype = "float32" min_val = float("0.00623834") @@ -6809,6 +7428,7 @@ class Program_weight_tensor_parameter_618: class Program_weight_tensor_parameter_619: name = "parameter_619" + original_name = "batch_norm2d_23.w_1" shape = [96] dtype = "float32" min_val = float("-0.0709982") @@ -6820,6 +7440,7 @@ class Program_weight_tensor_parameter_619: class Program_weight_tensor_parameter_620: name = "parameter_620" + original_name = "conv2d_24.w_0" shape = [96, 96, 3, 3] dtype = "float32" min_val = float("-0.0654835") @@ -6831,6 +7452,7 @@ class Program_weight_tensor_parameter_620: class Program_weight_tensor_parameter_621: name = "parameter_621" + original_name = "batch_norm2d_22.b_0" shape = [96] dtype = "float32" min_val = float("-3.46434") @@ -6842,6 +7464,7 @@ class Program_weight_tensor_parameter_621: class Program_weight_tensor_parameter_622: name = "parameter_622" + original_name = "batch_norm2d_22.w_0" shape = [96] dtype = "float32" min_val = float("0.686506") @@ -6853,6 +7476,7 @@ class Program_weight_tensor_parameter_622: class Program_weight_tensor_parameter_623: name = "parameter_623" + original_name = "batch_norm2d_22.w_2" shape = [96] dtype = "float32" min_val = float("0.0132607") @@ -6864,6 +7488,7 @@ class Program_weight_tensor_parameter_623: class Program_weight_tensor_parameter_624: name = "parameter_624" + original_name = "batch_norm2d_22.w_1" shape = [96] dtype = "float32" min_val = float("-0.483153") @@ -6875,6 +7500,7 @@ class Program_weight_tensor_parameter_624: class Program_weight_tensor_parameter_625: name = "parameter_625" + original_name = "conv2d_23.w_0" shape = [96, 96, 3, 3] dtype = "float32" min_val = float("-0.0824841") @@ -6886,6 +7512,7 @@ class Program_weight_tensor_parameter_625: class Program_weight_tensor_parameter_626: name = "parameter_626" + original_name = "batch_norm2d_21.b_0" shape = [96] dtype = "float32" min_val = float("-0.625302") @@ -6897,6 +7524,7 @@ class Program_weight_tensor_parameter_626: class Program_weight_tensor_parameter_627: name = "parameter_627" + original_name = "batch_norm2d_21.w_0" shape = [96] dtype = "float32" min_val = float("0.0910018") @@ -6908,6 +7536,7 @@ class Program_weight_tensor_parameter_627: class Program_weight_tensor_parameter_628: name = "parameter_628" + original_name = "batch_norm2d_21.w_2" shape = [96] dtype = "float32" min_val = float("0.000380277") @@ -6919,6 +7548,7 @@ class Program_weight_tensor_parameter_628: class Program_weight_tensor_parameter_629: name = "parameter_629" + original_name = "batch_norm2d_21.w_1" shape = [96] dtype = "float32" min_val = float("-0.035932") @@ -6930,6 +7560,7 @@ class Program_weight_tensor_parameter_629: class Program_weight_tensor_parameter_630: name = "parameter_630" + original_name = "conv2d_22.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.0925016") @@ -6941,6 +7572,7 @@ class Program_weight_tensor_parameter_630: class Program_weight_tensor_parameter_631: name = "parameter_631" + original_name = "batch_norm2d_20.b_0" shape = [96] dtype = "float32" min_val = float("-0.625183") @@ -6952,6 +7584,7 @@ class Program_weight_tensor_parameter_631: class Program_weight_tensor_parameter_632: name = "parameter_632" + original_name = "batch_norm2d_20.w_0" shape = [96] dtype = "float32" min_val = float("0.210658") @@ -6963,6 +7596,7 @@ class Program_weight_tensor_parameter_632: class Program_weight_tensor_parameter_633: name = "parameter_633" + original_name = "batch_norm2d_20.w_2" shape = [96] dtype = "float32" min_val = float("0.0103464") @@ -6974,6 +7608,7 @@ class Program_weight_tensor_parameter_633: class Program_weight_tensor_parameter_634: name = "parameter_634" + original_name = "batch_norm2d_20.w_1" shape = [96] dtype = "float32" min_val = float("-0.108648") @@ -6985,6 +7620,7 @@ class Program_weight_tensor_parameter_634: class Program_weight_tensor_parameter_635: name = "parameter_635" + original_name = "conv2d_21.w_0" shape = [96, 96, 3, 3] dtype = "float32" min_val = float("-0.0885375") @@ -6996,6 +7632,7 @@ class Program_weight_tensor_parameter_635: class Program_weight_tensor_parameter_636: name = "parameter_636" + original_name = "batch_norm2d_19.b_0" shape = [96] dtype = "float32" min_val = float("-2.40893") @@ -7007,6 +7644,7 @@ class Program_weight_tensor_parameter_636: class Program_weight_tensor_parameter_637: name = "parameter_637" + original_name = "batch_norm2d_19.w_0" shape = [96] dtype = "float32" min_val = float("0.853968") @@ -7018,6 +7656,7 @@ class Program_weight_tensor_parameter_637: class Program_weight_tensor_parameter_638: name = "parameter_638" + original_name = "batch_norm2d_19.w_2" shape = [96] dtype = "float32" min_val = float("0.0115478") @@ -7029,6 +7668,7 @@ class Program_weight_tensor_parameter_638: class Program_weight_tensor_parameter_639: name = "parameter_639" + original_name = "batch_norm2d_19.w_1" shape = [96] dtype = "float32" min_val = float("-0.571223") @@ -7040,6 +7680,7 @@ class Program_weight_tensor_parameter_639: class Program_weight_tensor_parameter_640: name = "parameter_640" + original_name = "conv2d_20.w_0" shape = [96, 96, 3, 3] dtype = "float32" min_val = float("-0.15411") @@ -7051,6 +7692,7 @@ class Program_weight_tensor_parameter_640: class Program_weight_tensor_parameter_641: name = "parameter_641" + original_name = "batch_norm2d_18.b_0" shape = [96] dtype = "float32" min_val = float("-3.16609") @@ -7062,6 +7704,7 @@ class Program_weight_tensor_parameter_641: class Program_weight_tensor_parameter_642: name = "parameter_642" + original_name = "batch_norm2d_18.w_0" shape = [96] dtype = "float32" min_val = float("0.214988") @@ -7073,6 +7716,7 @@ class Program_weight_tensor_parameter_642: class Program_weight_tensor_parameter_643: name = "parameter_643" + original_name = "batch_norm2d_18.w_2" shape = [96] dtype = "float32" min_val = float("0.00741665") @@ -7084,6 +7728,7 @@ class Program_weight_tensor_parameter_643: class Program_weight_tensor_parameter_644: name = "parameter_644" + original_name = "batch_norm2d_18.w_1" shape = [96] dtype = "float32" min_val = float("-0.27197") @@ -7095,6 +7740,7 @@ class Program_weight_tensor_parameter_644: class Program_weight_tensor_parameter_645: name = "parameter_645" + original_name = "conv2d_19.w_0" shape = [96, 192, 1, 1] dtype = "float32" min_val = float("-0.186901") @@ -7106,6 +7752,7 @@ class Program_weight_tensor_parameter_645: class Program_weight_tensor_parameter_646: name = "parameter_646" + original_name = "batch_norm2d_17.b_0" shape = [96] dtype = "float32" min_val = float("-4.92284") @@ -7117,6 +7764,7 @@ class Program_weight_tensor_parameter_646: class Program_weight_tensor_parameter_647: name = "parameter_647" + original_name = "batch_norm2d_17.w_0" shape = [96] dtype = "float32" min_val = float("0.414126") @@ -7128,6 +7776,7 @@ class Program_weight_tensor_parameter_647: class Program_weight_tensor_parameter_648: name = "parameter_648" + original_name = "batch_norm2d_17.w_2" shape = [96] dtype = "float32" min_val = float("0.00527536") @@ -7139,6 +7788,7 @@ class Program_weight_tensor_parameter_648: class Program_weight_tensor_parameter_649: name = "parameter_649" + original_name = "batch_norm2d_17.w_1" shape = [96] dtype = "float32" min_val = float("-0.171845") @@ -7150,6 +7800,7 @@ class Program_weight_tensor_parameter_649: class Program_weight_tensor_parameter_650: name = "parameter_650" + original_name = "conv2d_18.w_0" shape = [96, 192, 1, 1] dtype = "float32" min_val = float("-0.116975") @@ -7161,6 +7812,7 @@ class Program_weight_tensor_parameter_650: class Program_weight_tensor_parameter_651: name = "parameter_651" + original_name = "batch_norm2d_16.b_0" shape = [192] dtype = "float32" min_val = float("-2.27475") @@ -7172,6 +7824,7 @@ class Program_weight_tensor_parameter_651: class Program_weight_tensor_parameter_652: name = "parameter_652" + original_name = "batch_norm2d_16.w_0" shape = [192] dtype = "float32" min_val = float("0.632268") @@ -7183,6 +7836,7 @@ class Program_weight_tensor_parameter_652: class Program_weight_tensor_parameter_653: name = "parameter_653" + original_name = "batch_norm2d_16.w_2" shape = [192] dtype = "float32" min_val = float("0.0110312") @@ -7194,6 +7848,7 @@ class Program_weight_tensor_parameter_653: class Program_weight_tensor_parameter_654: name = "parameter_654" + original_name = "batch_norm2d_16.w_1" shape = [192] dtype = "float32" min_val = float("-0.578422") @@ -7205,6 +7860,7 @@ class Program_weight_tensor_parameter_654: class Program_weight_tensor_parameter_655: name = "parameter_655" + original_name = "conv2d_17.w_0" shape = [192, 128, 3, 3] dtype = "float32" min_val = float("-0.0856428") @@ -7216,6 +7872,7 @@ class Program_weight_tensor_parameter_655: class Program_weight_tensor_parameter_656: name = "parameter_656" + original_name = "batch_norm2d_15.b_0" shape = [128] dtype = "float32" min_val = float("-2.81597") @@ -7227,6 +7884,7 @@ class Program_weight_tensor_parameter_656: class Program_weight_tensor_parameter_657: name = "parameter_657" + original_name = "batch_norm2d_15.w_0" shape = [128] dtype = "float32" min_val = float("0.311227") @@ -7238,6 +7896,7 @@ class Program_weight_tensor_parameter_657: class Program_weight_tensor_parameter_658: name = "parameter_658" + original_name = "batch_norm2d_15.w_2" shape = [128] dtype = "float32" min_val = float("0.000859604") @@ -7249,6 +7908,7 @@ class Program_weight_tensor_parameter_658: class Program_weight_tensor_parameter_659: name = "parameter_659" + original_name = "batch_norm2d_15.w_1" shape = [128] dtype = "float32" min_val = float("-0.237838") @@ -7260,6 +7920,7 @@ class Program_weight_tensor_parameter_659: class Program_weight_tensor_parameter_660: name = "parameter_660" + original_name = "conv2d_16.w_0" shape = [128, 96, 1, 1] dtype = "float32" min_val = float("-0.171773") @@ -7271,6 +7932,7 @@ class Program_weight_tensor_parameter_660: class Program_weight_tensor_parameter_661: name = "parameter_661" + original_name = "conv2d_15.b_0" shape = [96] dtype = "float32" min_val = float("-0.0180386") @@ -7282,6 +7944,7 @@ class Program_weight_tensor_parameter_661: class Program_weight_tensor_parameter_662: name = "parameter_662" + original_name = "conv2d_15.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.30281") @@ -7293,6 +7956,7 @@ class Program_weight_tensor_parameter_662: class Program_weight_tensor_parameter_663: name = "parameter_663" + original_name = "batch_norm2d_14.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -7302,6 +7966,7 @@ class Program_weight_tensor_parameter_663: class Program_weight_tensor_parameter_664: name = "parameter_664" + original_name = "batch_norm2d_14.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -7311,6 +7976,7 @@ class Program_weight_tensor_parameter_664: class Program_weight_tensor_parameter_665: name = "parameter_665" + original_name = "batch_norm2d_14.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -7320,6 +7986,7 @@ class Program_weight_tensor_parameter_665: class Program_weight_tensor_parameter_666: name = "parameter_666" + original_name = "batch_norm2d_14.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -7329,6 +7996,7 @@ class Program_weight_tensor_parameter_666: class Program_weight_tensor_parameter_667: name = "parameter_667" + original_name = "conv2d_14.w_0" shape = [48, 48, 1, 1] dtype = "float32" min_val = float("-0.0501789") @@ -7340,6 +8008,7 @@ class Program_weight_tensor_parameter_667: class Program_weight_tensor_parameter_668: name = "parameter_668" + original_name = "batch_norm2d_13.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -7349,6 +8018,7 @@ class Program_weight_tensor_parameter_668: class Program_weight_tensor_parameter_669: name = "parameter_669" + original_name = "batch_norm2d_13.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -7358,6 +8028,7 @@ class Program_weight_tensor_parameter_669: class Program_weight_tensor_parameter_670: name = "parameter_670" + original_name = "batch_norm2d_13.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -7367,6 +8038,7 @@ class Program_weight_tensor_parameter_670: class Program_weight_tensor_parameter_671: name = "parameter_671" + original_name = "batch_norm2d_13.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -7376,6 +8048,7 @@ class Program_weight_tensor_parameter_671: class Program_weight_tensor_parameter_672: name = "parameter_672" + original_name = "conv2d_13.w_0" shape = [48, 48, 3, 3] dtype = "float32" min_val = float("-0.0578676") @@ -7387,6 +8060,7 @@ class Program_weight_tensor_parameter_672: class Program_weight_tensor_parameter_673: name = "parameter_673" + original_name = "batch_norm2d_12.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -7396,6 +8070,7 @@ class Program_weight_tensor_parameter_673: class Program_weight_tensor_parameter_674: name = "parameter_674" + original_name = "batch_norm2d_12.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -7405,6 +8080,7 @@ class Program_weight_tensor_parameter_674: class Program_weight_tensor_parameter_675: name = "parameter_675" + original_name = "batch_norm2d_12.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -7414,6 +8090,7 @@ class Program_weight_tensor_parameter_675: class Program_weight_tensor_parameter_676: name = "parameter_676" + original_name = "batch_norm2d_12.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -7423,6 +8100,7 @@ class Program_weight_tensor_parameter_676: class Program_weight_tensor_parameter_677: name = "parameter_677" + original_name = "conv2d_12.w_0" shape = [48, 48, 3, 3] dtype = "float32" min_val = float("-0.0925274") @@ -7434,6 +8112,7 @@ class Program_weight_tensor_parameter_677: class Program_weight_tensor_parameter_678: name = "parameter_678" + original_name = "batch_norm2d_11.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -7443,6 +8122,7 @@ class Program_weight_tensor_parameter_678: class Program_weight_tensor_parameter_679: name = "parameter_679" + original_name = "batch_norm2d_11.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -7452,6 +8132,7 @@ class Program_weight_tensor_parameter_679: class Program_weight_tensor_parameter_680: name = "parameter_680" + original_name = "batch_norm2d_11.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -7461,6 +8142,7 @@ class Program_weight_tensor_parameter_680: class Program_weight_tensor_parameter_681: name = "parameter_681" + original_name = "batch_norm2d_11.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -7470,6 +8152,7 @@ class Program_weight_tensor_parameter_681: class Program_weight_tensor_parameter_682: name = "parameter_682" + original_name = "conv2d_11.w_0" shape = [48, 48, 1, 1] dtype = "float32" min_val = float("-0.0727088") @@ -7481,6 +8164,7 @@ class Program_weight_tensor_parameter_682: class Program_weight_tensor_parameter_683: name = "parameter_683" + original_name = "batch_norm2d_10.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -7490,6 +8174,7 @@ class Program_weight_tensor_parameter_683: class Program_weight_tensor_parameter_684: name = "parameter_684" + original_name = "batch_norm2d_10.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -7499,6 +8184,7 @@ class Program_weight_tensor_parameter_684: class Program_weight_tensor_parameter_685: name = "parameter_685" + original_name = "batch_norm2d_10.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -7508,6 +8194,7 @@ class Program_weight_tensor_parameter_685: class Program_weight_tensor_parameter_686: name = "parameter_686" + original_name = "batch_norm2d_10.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -7517,6 +8204,7 @@ class Program_weight_tensor_parameter_686: class Program_weight_tensor_parameter_687: name = "parameter_687" + original_name = "conv2d_10.w_0" shape = [48, 48, 3, 3] dtype = "float32" min_val = float("-0.0621898") @@ -7528,6 +8216,7 @@ class Program_weight_tensor_parameter_687: class Program_weight_tensor_parameter_688: name = "parameter_688" + original_name = "batch_norm2d_9.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -7537,6 +8226,7 @@ class Program_weight_tensor_parameter_688: class Program_weight_tensor_parameter_689: name = "parameter_689" + original_name = "batch_norm2d_9.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -7546,6 +8236,7 @@ class Program_weight_tensor_parameter_689: class Program_weight_tensor_parameter_690: name = "parameter_690" + original_name = "batch_norm2d_9.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -7555,6 +8246,7 @@ class Program_weight_tensor_parameter_690: class Program_weight_tensor_parameter_691: name = "parameter_691" + original_name = "batch_norm2d_9.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -7564,6 +8256,7 @@ class Program_weight_tensor_parameter_691: class Program_weight_tensor_parameter_692: name = "parameter_692" + original_name = "conv2d_9.w_0" shape = [48, 48, 3, 3] dtype = "float32" min_val = float("-0.11162") @@ -7575,6 +8268,7 @@ class Program_weight_tensor_parameter_692: class Program_weight_tensor_parameter_693: name = "parameter_693" + original_name = "batch_norm2d_8.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -7584,6 +8278,7 @@ class Program_weight_tensor_parameter_693: class Program_weight_tensor_parameter_694: name = "parameter_694" + original_name = "batch_norm2d_8.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -7593,6 +8288,7 @@ class Program_weight_tensor_parameter_694: class Program_weight_tensor_parameter_695: name = "parameter_695" + original_name = "batch_norm2d_8.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -7602,6 +8298,7 @@ class Program_weight_tensor_parameter_695: class Program_weight_tensor_parameter_696: name = "parameter_696" + original_name = "batch_norm2d_8.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -7611,6 +8308,7 @@ class Program_weight_tensor_parameter_696: class Program_weight_tensor_parameter_697: name = "parameter_697" + original_name = "conv2d_8.w_0" shape = [48, 48, 1, 1] dtype = "float32" min_val = float("-0.0944494") @@ -7622,6 +8320,7 @@ class Program_weight_tensor_parameter_697: class Program_weight_tensor_parameter_698: name = "parameter_698" + original_name = "batch_norm2d_7.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -7631,6 +8330,7 @@ class Program_weight_tensor_parameter_698: class Program_weight_tensor_parameter_699: name = "parameter_699" + original_name = "batch_norm2d_7.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -7640,6 +8340,7 @@ class Program_weight_tensor_parameter_699: class Program_weight_tensor_parameter_700: name = "parameter_700" + original_name = "batch_norm2d_7.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -7649,6 +8350,7 @@ class Program_weight_tensor_parameter_700: class Program_weight_tensor_parameter_701: name = "parameter_701" + original_name = "batch_norm2d_7.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -7658,6 +8360,7 @@ class Program_weight_tensor_parameter_701: class Program_weight_tensor_parameter_702: name = "parameter_702" + original_name = "conv2d_7.w_0" shape = [48, 48, 3, 3] dtype = "float32" min_val = float("-0.0691644") @@ -7669,6 +8372,7 @@ class Program_weight_tensor_parameter_702: class Program_weight_tensor_parameter_703: name = "parameter_703" + original_name = "batch_norm2d_6.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -7678,6 +8382,7 @@ class Program_weight_tensor_parameter_703: class Program_weight_tensor_parameter_704: name = "parameter_704" + original_name = "batch_norm2d_6.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -7687,6 +8392,7 @@ class Program_weight_tensor_parameter_704: class Program_weight_tensor_parameter_705: name = "parameter_705" + original_name = "batch_norm2d_6.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -7696,6 +8402,7 @@ class Program_weight_tensor_parameter_705: class Program_weight_tensor_parameter_706: name = "parameter_706" + original_name = "batch_norm2d_6.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -7705,6 +8412,7 @@ class Program_weight_tensor_parameter_706: class Program_weight_tensor_parameter_707: name = "parameter_707" + original_name = "conv2d_6.w_0" shape = [48, 48, 3, 3] dtype = "float32" min_val = float("-0.133213") @@ -7716,6 +8424,7 @@ class Program_weight_tensor_parameter_707: class Program_weight_tensor_parameter_708: name = "parameter_708" + original_name = "batch_norm2d_5.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -7725,6 +8434,7 @@ class Program_weight_tensor_parameter_708: class Program_weight_tensor_parameter_709: name = "parameter_709" + original_name = "batch_norm2d_5.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -7734,6 +8444,7 @@ class Program_weight_tensor_parameter_709: class Program_weight_tensor_parameter_710: name = "parameter_710" + original_name = "batch_norm2d_5.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -7743,6 +8454,7 @@ class Program_weight_tensor_parameter_710: class Program_weight_tensor_parameter_711: name = "parameter_711" + original_name = "batch_norm2d_5.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -7752,6 +8464,7 @@ class Program_weight_tensor_parameter_711: class Program_weight_tensor_parameter_712: name = "parameter_712" + original_name = "conv2d_5.w_0" shape = [48, 96, 1, 1] dtype = "float32" min_val = float("-0.17806") @@ -7763,6 +8476,7 @@ class Program_weight_tensor_parameter_712: class Program_weight_tensor_parameter_713: name = "parameter_713" + original_name = "batch_norm2d_4.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -7772,6 +8486,7 @@ class Program_weight_tensor_parameter_713: class Program_weight_tensor_parameter_714: name = "parameter_714" + original_name = "batch_norm2d_4.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -7781,6 +8496,7 @@ class Program_weight_tensor_parameter_714: class Program_weight_tensor_parameter_715: name = "parameter_715" + original_name = "batch_norm2d_4.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -7790,6 +8506,7 @@ class Program_weight_tensor_parameter_715: class Program_weight_tensor_parameter_716: name = "parameter_716" + original_name = "batch_norm2d_4.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -7799,6 +8516,7 @@ class Program_weight_tensor_parameter_716: class Program_weight_tensor_parameter_717: name = "parameter_717" + original_name = "conv2d_4.w_0" shape = [48, 96, 1, 1] dtype = "float32" min_val = float("-0.135012") @@ -7810,6 +8528,7 @@ class Program_weight_tensor_parameter_717: class Program_weight_tensor_parameter_718: name = "parameter_718" + original_name = "batch_norm2d_3.b_0" shape = [96] dtype = "float32" min_val = float("-3.40701") @@ -7821,6 +8540,7 @@ class Program_weight_tensor_parameter_718: class Program_weight_tensor_parameter_719: name = "parameter_719" + original_name = "batch_norm2d_3.w_0" shape = [96] dtype = "float32" min_val = float("0.865919") @@ -7832,6 +8552,7 @@ class Program_weight_tensor_parameter_719: class Program_weight_tensor_parameter_720: name = "parameter_720" + original_name = "batch_norm2d_3.w_2" shape = [96] dtype = "float32" min_val = float("0.704881") @@ -7843,6 +8564,7 @@ class Program_weight_tensor_parameter_720: class Program_weight_tensor_parameter_721: name = "parameter_721" + original_name = "batch_norm2d_3.w_1" shape = [96] dtype = "float32" min_val = float("-1.47461") @@ -7854,6 +8576,7 @@ class Program_weight_tensor_parameter_721: class Program_weight_tensor_parameter_722: name = "parameter_722" + original_name = "conv2d_3.w_0" shape = [96, 64, 3, 3] dtype = "float32" min_val = float("-0.110689") @@ -7865,6 +8588,7 @@ class Program_weight_tensor_parameter_722: class Program_weight_tensor_parameter_723: name = "parameter_723" + original_name = "batch_norm2d_2.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -7874,6 +8598,7 @@ class Program_weight_tensor_parameter_723: class Program_weight_tensor_parameter_724: name = "parameter_724" + original_name = "batch_norm2d_2.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -7883,6 +8608,7 @@ class Program_weight_tensor_parameter_724: class Program_weight_tensor_parameter_725: name = "parameter_725" + original_name = "batch_norm2d_2.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -7892,6 +8618,7 @@ class Program_weight_tensor_parameter_725: class Program_weight_tensor_parameter_726: name = "parameter_726" + original_name = "batch_norm2d_2.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -7901,6 +8628,7 @@ class Program_weight_tensor_parameter_726: class Program_weight_tensor_parameter_727: name = "parameter_727" + original_name = "conv2d_2.w_0" shape = [64, 32, 3, 3] dtype = "float32" min_val = float("-0.179264") @@ -7912,6 +8640,7 @@ class Program_weight_tensor_parameter_727: class Program_weight_tensor_parameter_728: name = "parameter_728" + original_name = "batch_norm2d_1.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -7921,6 +8650,7 @@ class Program_weight_tensor_parameter_728: class Program_weight_tensor_parameter_729: name = "parameter_729" + original_name = "batch_norm2d_1.w_0" shape = [32] dtype = "float32" min_val = float("0") @@ -7930,6 +8660,7 @@ class Program_weight_tensor_parameter_729: class Program_weight_tensor_parameter_730: name = "parameter_730" + original_name = "batch_norm2d_1.w_2" shape = [32] dtype = "float32" min_val = float("0") @@ -7939,6 +8670,7 @@ class Program_weight_tensor_parameter_730: class Program_weight_tensor_parameter_731: name = "parameter_731" + original_name = "batch_norm2d_1.w_1" shape = [32] dtype = "float32" min_val = float("0") @@ -7948,6 +8680,7 @@ class Program_weight_tensor_parameter_731: class Program_weight_tensor_parameter_732: name = "parameter_732" + original_name = "conv2d_1.w_0" shape = [32, 32, 3, 3] dtype = "float32" min_val = float("-0.347786") @@ -7959,6 +8692,7 @@ class Program_weight_tensor_parameter_732: class Program_weight_tensor_parameter_733: name = "parameter_733" + original_name = "batch_norm2d_0.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -7968,6 +8702,7 @@ class Program_weight_tensor_parameter_733: class Program_weight_tensor_parameter_734: name = "parameter_734" + original_name = "batch_norm2d_0.w_0" shape = [32] dtype = "float32" min_val = float("0") @@ -7977,6 +8712,7 @@ class Program_weight_tensor_parameter_734: class Program_weight_tensor_parameter_735: name = "parameter_735" + original_name = "batch_norm2d_0.w_2" shape = [32] dtype = "float32" min_val = float("0") @@ -7986,6 +8722,7 @@ class Program_weight_tensor_parameter_735: class Program_weight_tensor_parameter_736: name = "parameter_736" + original_name = "batch_norm2d_0.w_1" shape = [32] dtype = "float32" min_val = float("0") @@ -7995,6 +8732,7 @@ class Program_weight_tensor_parameter_736: class Program_weight_tensor_parameter_737: name = "parameter_737" + original_name = "conv2d_0.w_0" shape = [32, 3, 3, 3] dtype = "float32" min_val = float("-0.317155") diff --git a/paddle_samples/PaddleX/PP-YOLOE_plus_SOD-S/subgraph_0/input_meta.py b/paddle_samples/PaddleX/PP-YOLOE_plus_SOD-S/subgraph_0/input_meta.py index e095af688..e483b1486 100644 --- a/paddle_samples/PaddleX/PP-YOLOE_plus_SOD-S/subgraph_0/input_meta.py +++ b/paddle_samples/PaddleX/PP-YOLOE_plus_SOD-S/subgraph_0/input_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_data_0: name = "data_0" + original_name = "param_88" shape = [512, 1536] dtype = "float32" min_val = float("-0.0470195") @@ -11,6 +12,7 @@ class Program_weight_tensor_data_0: class Program_weight_tensor_data_1: name = "data_1" + original_name = "param_89" shape = [1536] dtype = "float32" min_val = float("-0.00375843") @@ -22,6 +24,7 @@ class Program_weight_tensor_data_1: class Program_weight_tensor_data_2: name = "data_2" + original_name = "param_90" shape = [512, 1536] dtype = "float32" min_val = float("-0.0453469") @@ -33,6 +36,7 @@ class Program_weight_tensor_data_2: class Program_weight_tensor_data_3: name = "data_3" + original_name = "param_91" shape = [1536] dtype = "float32" min_val = float("-0.00084003") @@ -44,6 +48,7 @@ class Program_weight_tensor_data_3: class Program_weight_tensor_data_4: name = "data_4" + original_name = "param_92" shape = [512, 1536] dtype = "float32" min_val = float("-0.0444956") @@ -55,6 +60,7 @@ class Program_weight_tensor_data_4: class Program_weight_tensor_data_5: name = "data_5" + original_name = "param_93" shape = [1536] dtype = "float32" min_val = float("-0.000738378") @@ -66,6 +72,7 @@ class Program_weight_tensor_data_5: class Program_weight_tensor_data_6: name = "data_6" + original_name = "param_94" shape = [512, 1536] dtype = "float32" min_val = float("-0.0439514") @@ -77,6 +84,7 @@ class Program_weight_tensor_data_6: class Program_weight_tensor_data_7: name = "data_7" + original_name = "param_95" shape = [1536] dtype = "float32" min_val = float("-0.000790993") @@ -88,6 +96,7 @@ class Program_weight_tensor_data_7: class Program_weight_tensor_data_8: name = "data_8" + original_name = "var_5934" shape = [2, 128, 56, 56] dtype = "float32" min_val = float("-0.278465") @@ -99,6 +108,7 @@ class Program_weight_tensor_data_8: class Program_weight_tensor_data_9: name = "data_9" + original_name = "var_5935" shape = [2, 256, 28, 28] dtype = "float32" min_val = float("-0.278465") @@ -110,6 +120,7 @@ class Program_weight_tensor_data_9: class Program_weight_tensor_data_10: name = "data_10" + original_name = "var_5936" shape = [2, 512, 14, 14] dtype = "float32" min_val = float("-0.278465") diff --git a/paddle_samples/PaddleX/PP-YOLOE_plus_SOD-S/subgraph_0/weight_meta.py b/paddle_samples/PaddleX/PP-YOLOE_plus_SOD-S/subgraph_0/weight_meta.py index b881dcc4c..bdea6c15a 100644 --- a/paddle_samples/PaddleX/PP-YOLOE_plus_SOD-S/subgraph_0/weight_meta.py +++ b/paddle_samples/PaddleX/PP-YOLOE_plus_SOD-S/subgraph_0/weight_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_parameter_0: name = "parameter_0" + original_name = "batch_norm2d_71.b_0" shape = [384] dtype = "float32" min_val = float("-0.452683") @@ -11,6 +12,7 @@ class Program_weight_tensor_parameter_0: class Program_weight_tensor_parameter_1: name = "parameter_1" + original_name = "batch_norm2d_71.w_0" shape = [384] dtype = "float32" min_val = float("0.925767") @@ -22,6 +24,7 @@ class Program_weight_tensor_parameter_1: class Program_weight_tensor_parameter_2: name = "parameter_2" + original_name = "batch_norm2d_71.w_2" shape = [384] dtype = "float32" min_val = float("0.00208755") @@ -33,6 +36,7 @@ class Program_weight_tensor_parameter_2: class Program_weight_tensor_parameter_3: name = "parameter_3" + original_name = "batch_norm2d_71.w_1" shape = [384] dtype = "float32" min_val = float("-0.119095") @@ -44,6 +48,7 @@ class Program_weight_tensor_parameter_3: class Program_weight_tensor_parameter_4: name = "parameter_4" + original_name = "conv2d_75.w_0" shape = [384, 384, 1, 1] dtype = "float32" min_val = float("-0.0715528") @@ -55,6 +60,7 @@ class Program_weight_tensor_parameter_4: class Program_weight_tensor_parameter_5: name = "parameter_5" + original_name = "batch_norm2d_70.b_0" shape = [192] dtype = "float32" min_val = float("-0.285295") @@ -66,6 +72,7 @@ class Program_weight_tensor_parameter_5: class Program_weight_tensor_parameter_6: name = "parameter_6" + original_name = "batch_norm2d_70.w_0" shape = [192] dtype = "float32" min_val = float("0.874212") @@ -77,6 +84,7 @@ class Program_weight_tensor_parameter_6: class Program_weight_tensor_parameter_7: name = "parameter_7" + original_name = "batch_norm2d_70.w_2" shape = [192] dtype = "float32" min_val = float("0.00225351") @@ -88,6 +96,7 @@ class Program_weight_tensor_parameter_7: class Program_weight_tensor_parameter_8: name = "parameter_8" + original_name = "batch_norm2d_70.w_1" shape = [192] dtype = "float32" min_val = float("-0.0454866") @@ -99,6 +108,7 @@ class Program_weight_tensor_parameter_8: class Program_weight_tensor_parameter_9: name = "parameter_9" + original_name = "conv2d_74.w_0" shape = [192, 192, 1, 1] dtype = "float32" min_val = float("-0.0396186") @@ -110,6 +120,7 @@ class Program_weight_tensor_parameter_9: class Program_weight_tensor_parameter_10: name = "parameter_10" + original_name = "batch_norm2d_69.b_0" shape = [192] dtype = "float32" min_val = float("-0.285295") @@ -121,6 +132,7 @@ class Program_weight_tensor_parameter_10: class Program_weight_tensor_parameter_11: name = "parameter_11" + original_name = "batch_norm2d_69.w_0" shape = [192] dtype = "float32" min_val = float("0.955327") @@ -132,6 +144,7 @@ class Program_weight_tensor_parameter_11: class Program_weight_tensor_parameter_12: name = "parameter_12" + original_name = "batch_norm2d_69.w_2" shape = [192] dtype = "float32" min_val = float("0.00377482") @@ -143,6 +156,7 @@ class Program_weight_tensor_parameter_12: class Program_weight_tensor_parameter_13: name = "parameter_13" + original_name = "batch_norm2d_69.w_1" shape = [192] dtype = "float32" min_val = float("-0.113549") @@ -154,6 +168,7 @@ class Program_weight_tensor_parameter_13: class Program_weight_tensor_parameter_14: name = "parameter_14" + original_name = "conv2d_73.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.0682235") @@ -165,6 +180,7 @@ class Program_weight_tensor_parameter_14: class Program_weight_tensor_parameter_15: name = "parameter_15" + original_name = "batch_norm2d_68.b_0" shape = [192] dtype = "float32" min_val = float("-0.36677") @@ -176,6 +192,7 @@ class Program_weight_tensor_parameter_15: class Program_weight_tensor_parameter_16: name = "parameter_16" + original_name = "batch_norm2d_68.w_0" shape = [192] dtype = "float32" min_val = float("0.867888") @@ -187,6 +204,7 @@ class Program_weight_tensor_parameter_16: class Program_weight_tensor_parameter_17: name = "parameter_17" + original_name = "batch_norm2d_68.w_2" shape = [192] dtype = "float32" min_val = float("0.00703681") @@ -198,6 +216,7 @@ class Program_weight_tensor_parameter_17: class Program_weight_tensor_parameter_18: name = "parameter_18" + original_name = "batch_norm2d_68.w_1" shape = [192] dtype = "float32" min_val = float("-0.19484") @@ -209,6 +228,7 @@ class Program_weight_tensor_parameter_18: class Program_weight_tensor_parameter_19: name = "parameter_19" + original_name = "conv2d_72.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.0761584") @@ -220,6 +240,7 @@ class Program_weight_tensor_parameter_19: class Program_weight_tensor_parameter_20: name = "parameter_20" + original_name = "batch_norm2d_67.b_0" shape = [192] dtype = "float32" min_val = float("-0.271966") @@ -231,6 +252,7 @@ class Program_weight_tensor_parameter_20: class Program_weight_tensor_parameter_21: name = "parameter_21" + original_name = "batch_norm2d_67.w_0" shape = [192] dtype = "float32" min_val = float("0.903154") @@ -242,6 +264,7 @@ class Program_weight_tensor_parameter_21: class Program_weight_tensor_parameter_22: name = "parameter_22" + original_name = "batch_norm2d_67.w_2" shape = [192] dtype = "float32" min_val = float("0.00362619") @@ -253,6 +276,7 @@ class Program_weight_tensor_parameter_22: class Program_weight_tensor_parameter_23: name = "parameter_23" + original_name = "batch_norm2d_67.w_1" shape = [192] dtype = "float32" min_val = float("-0.108991") @@ -264,6 +288,7 @@ class Program_weight_tensor_parameter_23: class Program_weight_tensor_parameter_24: name = "parameter_24" + original_name = "conv2d_71.w_0" shape = [192, 576, 1, 1] dtype = "float32" min_val = float("-0.114156") @@ -275,6 +300,7 @@ class Program_weight_tensor_parameter_24: class Program_weight_tensor_parameter_25: name = "parameter_25" + original_name = "batch_norm2d_66.b_0" shape = [192] dtype = "float32" min_val = float("-0.110844") @@ -286,6 +312,7 @@ class Program_weight_tensor_parameter_25: class Program_weight_tensor_parameter_26: name = "parameter_26" + original_name = "batch_norm2d_66.w_0" shape = [192] dtype = "float32" min_val = float("0.825233") @@ -297,6 +324,7 @@ class Program_weight_tensor_parameter_26: class Program_weight_tensor_parameter_27: name = "parameter_27" + original_name = "batch_norm2d_66.w_2" shape = [192] dtype = "float32" min_val = float("0.0022236") @@ -308,6 +336,7 @@ class Program_weight_tensor_parameter_27: class Program_weight_tensor_parameter_28: name = "parameter_28" + original_name = "batch_norm2d_66.w_1" shape = [192] dtype = "float32" min_val = float("-0.095126") @@ -319,6 +348,7 @@ class Program_weight_tensor_parameter_28: class Program_weight_tensor_parameter_29: name = "parameter_29" + original_name = "conv2d_70.w_0" shape = [192, 576, 1, 1] dtype = "float32" min_val = float("-0.0338268") @@ -330,6 +360,7 @@ class Program_weight_tensor_parameter_29: class Program_weight_tensor_parameter_30: name = "parameter_30" + original_name = "batch_norm2d_65.b_0" shape = [192] dtype = "float32" min_val = float("-0.187741") @@ -341,6 +372,7 @@ class Program_weight_tensor_parameter_30: class Program_weight_tensor_parameter_31: name = "parameter_31" + original_name = "batch_norm2d_65.w_0" shape = [192] dtype = "float32" min_val = float("0.811184") @@ -352,6 +384,7 @@ class Program_weight_tensor_parameter_31: class Program_weight_tensor_parameter_32: name = "parameter_32" + original_name = "batch_norm2d_65.w_2" shape = [192] dtype = "float32" min_val = float("0.00769277") @@ -363,6 +396,7 @@ class Program_weight_tensor_parameter_32: class Program_weight_tensor_parameter_33: name = "parameter_33" + original_name = "batch_norm2d_65.w_1" shape = [192] dtype = "float32" min_val = float("-0.460153") @@ -374,6 +408,7 @@ class Program_weight_tensor_parameter_33: class Program_weight_tensor_parameter_34: name = "parameter_34" + original_name = "conv2d_69.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.0422034") @@ -385,6 +420,7 @@ class Program_weight_tensor_parameter_34: class Program_weight_tensor_parameter_35: name = "parameter_35" + original_name = "batch_norm2d_64.b_0" shape = [192] dtype = "float32" min_val = float("-0.42846") @@ -396,6 +432,7 @@ class Program_weight_tensor_parameter_35: class Program_weight_tensor_parameter_36: name = "parameter_36" + original_name = "batch_norm2d_64.w_0" shape = [192] dtype = "float32" min_val = float("0.675626") @@ -407,6 +444,7 @@ class Program_weight_tensor_parameter_36: class Program_weight_tensor_parameter_37: name = "parameter_37" + original_name = "batch_norm2d_64.w_2" shape = [192] dtype = "float32" min_val = float("0.00528809") @@ -418,6 +456,7 @@ class Program_weight_tensor_parameter_37: class Program_weight_tensor_parameter_38: name = "parameter_38" + original_name = "batch_norm2d_64.w_1" shape = [192] dtype = "float32" min_val = float("-0.151869") @@ -429,6 +468,7 @@ class Program_weight_tensor_parameter_38: class Program_weight_tensor_parameter_39: name = "parameter_39" + original_name = "conv2d_68.w_0" shape = [192, 192, 1, 1] dtype = "float32" min_val = float("-0.130795") @@ -440,6 +480,7 @@ class Program_weight_tensor_parameter_39: class Program_weight_tensor_parameter_40: name = "parameter_40" + original_name = "batch_norm2d_63.b_0" shape = [96] dtype = "float32" min_val = float("-0.288894") @@ -451,6 +492,7 @@ class Program_weight_tensor_parameter_40: class Program_weight_tensor_parameter_41: name = "parameter_41" + original_name = "batch_norm2d_63.w_0" shape = [96] dtype = "float32" min_val = float("0.803374") @@ -462,6 +504,7 @@ class Program_weight_tensor_parameter_41: class Program_weight_tensor_parameter_42: name = "parameter_42" + original_name = "batch_norm2d_63.w_2" shape = [96] dtype = "float32" min_val = float("0.00123928") @@ -473,6 +516,7 @@ class Program_weight_tensor_parameter_42: class Program_weight_tensor_parameter_43: name = "parameter_43" + original_name = "batch_norm2d_63.w_1" shape = [96] dtype = "float32" min_val = float("-0.0524226") @@ -484,6 +528,7 @@ class Program_weight_tensor_parameter_43: class Program_weight_tensor_parameter_44: name = "parameter_44" + original_name = "conv2d_67.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.0689645") @@ -495,6 +540,7 @@ class Program_weight_tensor_parameter_44: class Program_weight_tensor_parameter_45: name = "parameter_45" + original_name = "batch_norm2d_62.b_0" shape = [96] dtype = "float32" min_val = float("-0.288894") @@ -506,6 +552,7 @@ class Program_weight_tensor_parameter_45: class Program_weight_tensor_parameter_46: name = "parameter_46" + original_name = "batch_norm2d_62.w_0" shape = [96] dtype = "float32" min_val = float("0.915165") @@ -517,6 +564,7 @@ class Program_weight_tensor_parameter_46: class Program_weight_tensor_parameter_47: name = "parameter_47" + original_name = "batch_norm2d_62.w_2" shape = [96] dtype = "float32" min_val = float("0.00571053") @@ -528,6 +576,7 @@ class Program_weight_tensor_parameter_47: class Program_weight_tensor_parameter_48: name = "parameter_48" + original_name = "batch_norm2d_62.w_1" shape = [96] dtype = "float32" min_val = float("-0.126333") @@ -539,6 +588,7 @@ class Program_weight_tensor_parameter_48: class Program_weight_tensor_parameter_49: name = "parameter_49" + original_name = "conv2d_66.w_0" shape = [96, 96, 3, 3] dtype = "float32" min_val = float("-0.0904939") @@ -550,6 +600,7 @@ class Program_weight_tensor_parameter_49: class Program_weight_tensor_parameter_50: name = "parameter_50" + original_name = "batch_norm2d_61.b_0" shape = [96] dtype = "float32" min_val = float("-0.598498") @@ -561,6 +612,7 @@ class Program_weight_tensor_parameter_50: class Program_weight_tensor_parameter_51: name = "parameter_51" + original_name = "batch_norm2d_61.w_0" shape = [96] dtype = "float32" min_val = float("0.748158") @@ -572,6 +624,7 @@ class Program_weight_tensor_parameter_51: class Program_weight_tensor_parameter_52: name = "parameter_52" + original_name = "batch_norm2d_61.w_2" shape = [96] dtype = "float32" min_val = float("0.0139572") @@ -583,6 +636,7 @@ class Program_weight_tensor_parameter_52: class Program_weight_tensor_parameter_53: name = "parameter_53" + original_name = "batch_norm2d_61.w_1" shape = [96] dtype = "float32" min_val = float("-0.0925774") @@ -594,6 +648,7 @@ class Program_weight_tensor_parameter_53: class Program_weight_tensor_parameter_54: name = "parameter_54" + original_name = "conv2d_65.w_0" shape = [96, 96, 3, 3] dtype = "float32" min_val = float("-0.0863241") @@ -605,6 +660,7 @@ class Program_weight_tensor_parameter_54: class Program_weight_tensor_parameter_55: name = "parameter_55" + original_name = "batch_norm2d_60.b_0" shape = [96] dtype = "float32" min_val = float("-0.492263") @@ -616,6 +672,7 @@ class Program_weight_tensor_parameter_55: class Program_weight_tensor_parameter_56: name = "parameter_56" + original_name = "batch_norm2d_60.w_0" shape = [96] dtype = "float32" min_val = float("0.846758") @@ -627,6 +684,7 @@ class Program_weight_tensor_parameter_56: class Program_weight_tensor_parameter_57: name = "parameter_57" + original_name = "batch_norm2d_60.w_2" shape = [96] dtype = "float32" min_val = float("0.00479462") @@ -638,6 +696,7 @@ class Program_weight_tensor_parameter_57: class Program_weight_tensor_parameter_58: name = "parameter_58" + original_name = "batch_norm2d_60.w_1" shape = [96] dtype = "float32" min_val = float("-0.152602") @@ -649,6 +708,7 @@ class Program_weight_tensor_parameter_58: class Program_weight_tensor_parameter_59: name = "parameter_59" + original_name = "conv2d_64.w_0" shape = [96, 288, 1, 1] dtype = "float32" min_val = float("-0.0816609") @@ -660,6 +720,7 @@ class Program_weight_tensor_parameter_59: class Program_weight_tensor_parameter_60: name = "parameter_60" + original_name = "batch_norm2d_59.b_0" shape = [96] dtype = "float32" min_val = float("-0.13292") @@ -671,6 +732,7 @@ class Program_weight_tensor_parameter_60: class Program_weight_tensor_parameter_61: name = "parameter_61" + original_name = "batch_norm2d_59.w_0" shape = [96] dtype = "float32" min_val = float("0.816688") @@ -682,6 +744,7 @@ class Program_weight_tensor_parameter_61: class Program_weight_tensor_parameter_62: name = "parameter_62" + original_name = "batch_norm2d_59.w_2" shape = [96] dtype = "float32" min_val = float("0.00533424") @@ -693,6 +756,7 @@ class Program_weight_tensor_parameter_62: class Program_weight_tensor_parameter_63: name = "parameter_63" + original_name = "batch_norm2d_59.w_1" shape = [96] dtype = "float32" min_val = float("-0.08722") @@ -704,6 +768,7 @@ class Program_weight_tensor_parameter_63: class Program_weight_tensor_parameter_64: name = "parameter_64" + original_name = "conv2d_63.w_0" shape = [96, 288, 1, 1] dtype = "float32" min_val = float("-0.0877099") @@ -715,6 +780,7 @@ class Program_weight_tensor_parameter_64: class Program_weight_tensor_parameter_65: name = "parameter_65" + original_name = "batch_norm2d_58.b_0" shape = [96] dtype = "float32" min_val = float("-0.275331") @@ -726,6 +792,7 @@ class Program_weight_tensor_parameter_65: class Program_weight_tensor_parameter_66: name = "parameter_66" + original_name = "batch_norm2d_58.w_0" shape = [96] dtype = "float32" min_val = float("0.710157") @@ -737,6 +804,7 @@ class Program_weight_tensor_parameter_66: class Program_weight_tensor_parameter_67: name = "parameter_67" + original_name = "batch_norm2d_58.w_2" shape = [96] dtype = "float32" min_val = float("0.0117676") @@ -748,6 +816,7 @@ class Program_weight_tensor_parameter_67: class Program_weight_tensor_parameter_68: name = "parameter_68" + original_name = "batch_norm2d_58.w_1" shape = [96] dtype = "float32" min_val = float("-0.444541") @@ -759,6 +828,7 @@ class Program_weight_tensor_parameter_68: class Program_weight_tensor_parameter_69: name = "parameter_69" + original_name = "conv2d_62.w_0" shape = [96, 96, 3, 3] dtype = "float32" min_val = float("-0.059943") @@ -770,6 +840,7 @@ class Program_weight_tensor_parameter_69: class Program_weight_tensor_parameter_70: name = "parameter_70" + original_name = "batch_norm2d_57.b_0" shape = [96] dtype = "float32" min_val = float("-0.725935") @@ -781,6 +852,7 @@ class Program_weight_tensor_parameter_70: class Program_weight_tensor_parameter_71: name = "parameter_71" + original_name = "batch_norm2d_57.w_0" shape = [96] dtype = "float32" min_val = float("0.498479") @@ -792,6 +864,7 @@ class Program_weight_tensor_parameter_71: class Program_weight_tensor_parameter_72: name = "parameter_72" + original_name = "batch_norm2d_57.w_2" shape = [96] dtype = "float32" min_val = float("0.00945594") @@ -803,6 +876,7 @@ class Program_weight_tensor_parameter_72: class Program_weight_tensor_parameter_73: name = "parameter_73" + original_name = "batch_norm2d_57.w_1" shape = [96] dtype = "float32" min_val = float("-0.243464") @@ -814,6 +888,7 @@ class Program_weight_tensor_parameter_73: class Program_weight_tensor_parameter_74: name = "parameter_74" + original_name = "conv2d_61.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.213183") @@ -825,6 +900,7 @@ class Program_weight_tensor_parameter_74: class Program_weight_tensor_parameter_75: name = "parameter_75" + original_name = "batch_norm2d_56.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -834,6 +910,7 @@ class Program_weight_tensor_parameter_75: class Program_weight_tensor_parameter_76: name = "parameter_76" + original_name = "batch_norm2d_56.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -843,6 +920,7 @@ class Program_weight_tensor_parameter_76: class Program_weight_tensor_parameter_77: name = "parameter_77" + original_name = "batch_norm2d_56.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -852,6 +930,7 @@ class Program_weight_tensor_parameter_77: class Program_weight_tensor_parameter_78: name = "parameter_78" + original_name = "batch_norm2d_56.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -861,6 +940,7 @@ class Program_weight_tensor_parameter_78: class Program_weight_tensor_parameter_79: name = "parameter_79" + original_name = "conv2d_60.w_0" shape = [48, 48, 1, 1] dtype = "float32" min_val = float("-0.0883355") @@ -872,6 +952,7 @@ class Program_weight_tensor_parameter_79: class Program_weight_tensor_parameter_80: name = "parameter_80" + original_name = "batch_norm2d_55.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -881,6 +962,7 @@ class Program_weight_tensor_parameter_80: class Program_weight_tensor_parameter_81: name = "parameter_81" + original_name = "batch_norm2d_55.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -890,6 +972,7 @@ class Program_weight_tensor_parameter_81: class Program_weight_tensor_parameter_82: name = "parameter_82" + original_name = "batch_norm2d_55.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -899,6 +982,7 @@ class Program_weight_tensor_parameter_82: class Program_weight_tensor_parameter_83: name = "parameter_83" + original_name = "batch_norm2d_55.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -908,6 +992,7 @@ class Program_weight_tensor_parameter_83: class Program_weight_tensor_parameter_84: name = "parameter_84" + original_name = "conv2d_59.w_0" shape = [48, 48, 3, 3] dtype = "float32" min_val = float("-0.146307") @@ -919,6 +1004,7 @@ class Program_weight_tensor_parameter_84: class Program_weight_tensor_parameter_85: name = "parameter_85" + original_name = "batch_norm2d_54.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -928,6 +1014,7 @@ class Program_weight_tensor_parameter_85: class Program_weight_tensor_parameter_86: name = "parameter_86" + original_name = "batch_norm2d_54.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -937,6 +1024,7 @@ class Program_weight_tensor_parameter_86: class Program_weight_tensor_parameter_87: name = "parameter_87" + original_name = "batch_norm2d_54.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -946,6 +1034,7 @@ class Program_weight_tensor_parameter_87: class Program_weight_tensor_parameter_88: name = "parameter_88" + original_name = "batch_norm2d_54.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -955,6 +1044,7 @@ class Program_weight_tensor_parameter_88: class Program_weight_tensor_parameter_89: name = "parameter_89" + original_name = "conv2d_58.w_0" shape = [48, 48, 3, 3] dtype = "float32" min_val = float("-0.156302") @@ -966,6 +1056,7 @@ class Program_weight_tensor_parameter_89: class Program_weight_tensor_parameter_90: name = "parameter_90" + original_name = "batch_norm2d_53.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -975,6 +1066,7 @@ class Program_weight_tensor_parameter_90: class Program_weight_tensor_parameter_91: name = "parameter_91" + original_name = "batch_norm2d_53.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -984,6 +1076,7 @@ class Program_weight_tensor_parameter_91: class Program_weight_tensor_parameter_92: name = "parameter_92" + original_name = "batch_norm2d_53.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -993,6 +1086,7 @@ class Program_weight_tensor_parameter_92: class Program_weight_tensor_parameter_93: name = "parameter_93" + original_name = "batch_norm2d_53.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -1002,6 +1096,7 @@ class Program_weight_tensor_parameter_93: class Program_weight_tensor_parameter_94: name = "parameter_94" + original_name = "conv2d_57.w_0" shape = [48, 224, 1, 1] dtype = "float32" min_val = float("-0.204906") @@ -1013,6 +1108,7 @@ class Program_weight_tensor_parameter_94: class Program_weight_tensor_parameter_95: name = "parameter_95" + original_name = "batch_norm2d_52.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -1022,6 +1118,7 @@ class Program_weight_tensor_parameter_95: class Program_weight_tensor_parameter_96: name = "parameter_96" + original_name = "batch_norm2d_52.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -1031,6 +1128,7 @@ class Program_weight_tensor_parameter_96: class Program_weight_tensor_parameter_97: name = "parameter_97" + original_name = "batch_norm2d_52.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -1040,6 +1138,7 @@ class Program_weight_tensor_parameter_97: class Program_weight_tensor_parameter_98: name = "parameter_98" + original_name = "batch_norm2d_52.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -1049,6 +1148,7 @@ class Program_weight_tensor_parameter_98: class Program_weight_tensor_parameter_99: name = "parameter_99" + original_name = "conv2d_56.w_0" shape = [48, 224, 1, 1] dtype = "float32" min_val = float("-0.117432") @@ -1060,6 +1160,7 @@ class Program_weight_tensor_parameter_99: class Program_weight_tensor_parameter_100: name = "parameter_100" + original_name = "batch_norm2d_51.b_0" shape = [96] dtype = "float32" min_val = float("-0.279666") @@ -1071,6 +1172,7 @@ class Program_weight_tensor_parameter_100: class Program_weight_tensor_parameter_101: name = "parameter_101" + original_name = "batch_norm2d_51.w_0" shape = [96] dtype = "float32" min_val = float("0.570144") @@ -1082,6 +1184,7 @@ class Program_weight_tensor_parameter_101: class Program_weight_tensor_parameter_102: name = "parameter_102" + original_name = "batch_norm2d_51.w_2" shape = [96] dtype = "float32" min_val = float("0.0106917") @@ -1093,6 +1196,7 @@ class Program_weight_tensor_parameter_102: class Program_weight_tensor_parameter_103: name = "parameter_103" + original_name = "batch_norm2d_51.w_1" shape = [96] dtype = "float32" min_val = float("-0.179085") @@ -1104,6 +1208,7 @@ class Program_weight_tensor_parameter_103: class Program_weight_tensor_parameter_104: name = "parameter_104" + original_name = "conv2d_55.w_0" shape = [96, 192, 1, 1] dtype = "float32" min_val = float("-0.10276") @@ -1115,6 +1220,7 @@ class Program_weight_tensor_parameter_104: class Program_weight_tensor_parameter_105: name = "parameter_105" + original_name = "batch_norm2d_50.b_0" shape = [192] dtype = "float32" min_val = float("-0.406643") @@ -1126,6 +1232,7 @@ class Program_weight_tensor_parameter_105: class Program_weight_tensor_parameter_106: name = "parameter_106" + original_name = "batch_norm2d_50.w_0" shape = [192] dtype = "float32" min_val = float("0.710043") @@ -1137,6 +1244,7 @@ class Program_weight_tensor_parameter_106: class Program_weight_tensor_parameter_107: name = "parameter_107" + original_name = "batch_norm2d_50.w_2" shape = [192] dtype = "float32" min_val = float("0.00807578") @@ -1148,6 +1256,7 @@ class Program_weight_tensor_parameter_107: class Program_weight_tensor_parameter_108: name = "parameter_108" + original_name = "batch_norm2d_50.w_1" shape = [192] dtype = "float32" min_val = float("-0.249401") @@ -1159,6 +1268,7 @@ class Program_weight_tensor_parameter_108: class Program_weight_tensor_parameter_109: name = "parameter_109" + original_name = "conv2d_54.w_0" shape = [192, 192, 1, 1] dtype = "float32" min_val = float("-0.108991") @@ -1170,6 +1280,7 @@ class Program_weight_tensor_parameter_109: class Program_weight_tensor_parameter_110: name = "parameter_110" + original_name = "batch_norm2d_49.b_0" shape = [96] dtype = "float32" min_val = float("-0.329532") @@ -1181,6 +1292,7 @@ class Program_weight_tensor_parameter_110: class Program_weight_tensor_parameter_111: name = "parameter_111" + original_name = "batch_norm2d_49.w_0" shape = [96] dtype = "float32" min_val = float("0.675243") @@ -1192,6 +1304,7 @@ class Program_weight_tensor_parameter_111: class Program_weight_tensor_parameter_112: name = "parameter_112" + original_name = "batch_norm2d_49.w_2" shape = [96] dtype = "float32" min_val = float("0.00447042") @@ -1203,6 +1316,7 @@ class Program_weight_tensor_parameter_112: class Program_weight_tensor_parameter_113: name = "parameter_113" + original_name = "batch_norm2d_49.w_1" shape = [96] dtype = "float32" min_val = float("-0.054565") @@ -1214,6 +1328,7 @@ class Program_weight_tensor_parameter_113: class Program_weight_tensor_parameter_114: name = "parameter_114" + original_name = "conv2d_53.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.051869") @@ -1225,6 +1340,7 @@ class Program_weight_tensor_parameter_114: class Program_weight_tensor_parameter_115: name = "parameter_115" + original_name = "batch_norm2d_48.b_0" shape = [96] dtype = "float32" min_val = float("-0.329532") @@ -1236,6 +1352,7 @@ class Program_weight_tensor_parameter_115: class Program_weight_tensor_parameter_116: name = "parameter_116" + original_name = "batch_norm2d_48.w_0" shape = [96] dtype = "float32" min_val = float("0.879109") @@ -1247,6 +1364,7 @@ class Program_weight_tensor_parameter_116: class Program_weight_tensor_parameter_117: name = "parameter_117" + original_name = "batch_norm2d_48.w_2" shape = [96] dtype = "float32" min_val = float("0.014723") @@ -1258,6 +1376,7 @@ class Program_weight_tensor_parameter_117: class Program_weight_tensor_parameter_118: name = "parameter_118" + original_name = "batch_norm2d_48.w_1" shape = [96] dtype = "float32" min_val = float("-0.119395") @@ -1269,6 +1388,7 @@ class Program_weight_tensor_parameter_118: class Program_weight_tensor_parameter_119: name = "parameter_119" + original_name = "conv2d_52.w_0" shape = [96, 96, 3, 3] dtype = "float32" min_val = float("-0.119863") @@ -1280,6 +1400,7 @@ class Program_weight_tensor_parameter_119: class Program_weight_tensor_parameter_120: name = "parameter_120" + original_name = "batch_norm2d_47.b_0" shape = [96] dtype = "float32" min_val = float("-0.684479") @@ -1291,6 +1412,7 @@ class Program_weight_tensor_parameter_120: class Program_weight_tensor_parameter_121: name = "parameter_121" + original_name = "batch_norm2d_47.w_0" shape = [96] dtype = "float32" min_val = float("0.647154") @@ -1302,6 +1424,7 @@ class Program_weight_tensor_parameter_121: class Program_weight_tensor_parameter_122: name = "parameter_122" + original_name = "batch_norm2d_47.w_2" shape = [96] dtype = "float32" min_val = float("0.0179775") @@ -1313,6 +1436,7 @@ class Program_weight_tensor_parameter_122: class Program_weight_tensor_parameter_123: name = "parameter_123" + original_name = "batch_norm2d_47.w_1" shape = [96] dtype = "float32" min_val = float("-0.143275") @@ -1324,6 +1448,7 @@ class Program_weight_tensor_parameter_123: class Program_weight_tensor_parameter_124: name = "parameter_124" + original_name = "conv2d_51.w_0" shape = [96, 96, 3, 3] dtype = "float32" min_val = float("-0.150701") @@ -1335,6 +1460,7 @@ class Program_weight_tensor_parameter_124: class Program_weight_tensor_parameter_125: name = "parameter_125" + original_name = "batch_norm2d_46.b_0" shape = [96] dtype = "float32" min_val = float("-0.629138") @@ -1346,6 +1472,7 @@ class Program_weight_tensor_parameter_125: class Program_weight_tensor_parameter_126: name = "parameter_126" + original_name = "batch_norm2d_46.w_0" shape = [96] dtype = "float32" min_val = float("0.719559") @@ -1357,6 +1484,7 @@ class Program_weight_tensor_parameter_126: class Program_weight_tensor_parameter_127: name = "parameter_127" + original_name = "batch_norm2d_46.w_2" shape = [96] dtype = "float32" min_val = float("0.00837206") @@ -1368,6 +1496,7 @@ class Program_weight_tensor_parameter_127: class Program_weight_tensor_parameter_128: name = "parameter_128" + original_name = "batch_norm2d_46.w_1" shape = [96] dtype = "float32" min_val = float("-0.1407") @@ -1379,6 +1508,7 @@ class Program_weight_tensor_parameter_128: class Program_weight_tensor_parameter_129: name = "parameter_129" + original_name = "conv2d_50.w_0" shape = [96, 448, 1, 1] dtype = "float32" min_val = float("-0.232396") @@ -1390,6 +1520,7 @@ class Program_weight_tensor_parameter_129: class Program_weight_tensor_parameter_130: name = "parameter_130" + original_name = "batch_norm2d_45.b_0" shape = [96] dtype = "float32" min_val = float("-0.188629") @@ -1401,6 +1532,7 @@ class Program_weight_tensor_parameter_130: class Program_weight_tensor_parameter_131: name = "parameter_131" + original_name = "batch_norm2d_45.w_0" shape = [96] dtype = "float32" min_val = float("0.811507") @@ -1412,6 +1544,7 @@ class Program_weight_tensor_parameter_131: class Program_weight_tensor_parameter_132: name = "parameter_132" + original_name = "batch_norm2d_45.w_2" shape = [96] dtype = "float32" min_val = float("0.00601664") @@ -1423,6 +1556,7 @@ class Program_weight_tensor_parameter_132: class Program_weight_tensor_parameter_133: name = "parameter_133" + original_name = "batch_norm2d_45.w_1" shape = [96] dtype = "float32" min_val = float("-0.108405") @@ -1434,6 +1568,7 @@ class Program_weight_tensor_parameter_133: class Program_weight_tensor_parameter_134: name = "parameter_134" + original_name = "conv2d_49.w_0" shape = [96, 448, 1, 1] dtype = "float32" min_val = float("-0.13482") @@ -1445,6 +1580,7 @@ class Program_weight_tensor_parameter_134: class Program_weight_tensor_parameter_135: name = "parameter_135" + original_name = "batch_norm2d_44.b_0" shape = [192] dtype = "float32" min_val = float("-0.428995") @@ -1456,6 +1592,7 @@ class Program_weight_tensor_parameter_135: class Program_weight_tensor_parameter_136: name = "parameter_136" + original_name = "batch_norm2d_44.w_0" shape = [192] dtype = "float32" min_val = float("0.677275") @@ -1467,6 +1604,7 @@ class Program_weight_tensor_parameter_136: class Program_weight_tensor_parameter_137: name = "parameter_137" + original_name = "batch_norm2d_44.w_2" shape = [192] dtype = "float32" min_val = float("0.00937355") @@ -1478,6 +1616,7 @@ class Program_weight_tensor_parameter_137: class Program_weight_tensor_parameter_138: name = "parameter_138" + original_name = "batch_norm2d_44.w_1" shape = [192] dtype = "float32" min_val = float("-0.132564") @@ -1489,6 +1628,7 @@ class Program_weight_tensor_parameter_138: class Program_weight_tensor_parameter_139: name = "parameter_139" + original_name = "conv2d_48.w_0" shape = [192, 384, 1, 1] dtype = "float32" min_val = float("-0.0860279") @@ -1500,6 +1640,7 @@ class Program_weight_tensor_parameter_139: class Program_weight_tensor_parameter_140: name = "parameter_140" + original_name = "batch_norm2d_43.b_0" shape = [384] dtype = "float32" min_val = float("-0.34678") @@ -1511,6 +1652,7 @@ class Program_weight_tensor_parameter_140: class Program_weight_tensor_parameter_141: name = "parameter_141" + original_name = "batch_norm2d_43.w_0" shape = [384] dtype = "float32" min_val = float("0.85173") @@ -1522,6 +1664,7 @@ class Program_weight_tensor_parameter_141: class Program_weight_tensor_parameter_142: name = "parameter_142" + original_name = "batch_norm2d_43.w_2" shape = [384] dtype = "float32" min_val = float("0.0064039") @@ -1533,6 +1676,7 @@ class Program_weight_tensor_parameter_142: class Program_weight_tensor_parameter_143: name = "parameter_143" + original_name = "batch_norm2d_43.w_1" shape = [384] dtype = "float32" min_val = float("-0.140946") @@ -1544,6 +1688,7 @@ class Program_weight_tensor_parameter_143: class Program_weight_tensor_parameter_144: name = "parameter_144" + original_name = "conv2d_47.w_0" shape = [384, 384, 1, 1] dtype = "float32" min_val = float("-0.12987") @@ -1555,6 +1700,7 @@ class Program_weight_tensor_parameter_144: class Program_weight_tensor_parameter_145: name = "parameter_145" + original_name = "batch_norm2d_42.b_0" shape = [192] dtype = "float32" min_val = float("-0.341339") @@ -1566,6 +1712,7 @@ class Program_weight_tensor_parameter_145: class Program_weight_tensor_parameter_146: name = "parameter_146" + original_name = "batch_norm2d_42.w_0" shape = [192] dtype = "float32" min_val = float("0.847019") @@ -1577,6 +1724,7 @@ class Program_weight_tensor_parameter_146: class Program_weight_tensor_parameter_147: name = "parameter_147" + original_name = "batch_norm2d_42.w_2" shape = [192] dtype = "float32" min_val = float("0.0589707") @@ -1588,6 +1736,7 @@ class Program_weight_tensor_parameter_147: class Program_weight_tensor_parameter_148: name = "parameter_148" + original_name = "batch_norm2d_42.w_1" shape = [192] dtype = "float32" min_val = float("-1.61611") @@ -1599,6 +1748,7 @@ class Program_weight_tensor_parameter_148: class Program_weight_tensor_parameter_149: name = "parameter_149" + original_name = "conv2d_46.w_0" shape = [192, 768, 1, 1] dtype = "float32" min_val = float("-0.114669") @@ -1610,6 +1760,7 @@ class Program_weight_tensor_parameter_149: class Program_weight_tensor_parameter_150: name = "parameter_150" + original_name = "batch_norm2d_41.b_0" shape = [192] dtype = "float32" min_val = float("-0.0804559") @@ -1621,6 +1772,7 @@ class Program_weight_tensor_parameter_150: class Program_weight_tensor_parameter_151: name = "parameter_151" + original_name = "batch_norm2d_41.w_0" shape = [192] dtype = "float32" min_val = float("0.807653") @@ -1632,6 +1784,7 @@ class Program_weight_tensor_parameter_151: class Program_weight_tensor_parameter_152: name = "parameter_152" + original_name = "batch_norm2d_41.w_2" shape = [192] dtype = "float32" min_val = float("0.00516858") @@ -1643,6 +1796,7 @@ class Program_weight_tensor_parameter_152: class Program_weight_tensor_parameter_153: name = "parameter_153" + original_name = "batch_norm2d_41.w_1" shape = [192] dtype = "float32" min_val = float("-0.109864") @@ -1654,6 +1808,7 @@ class Program_weight_tensor_parameter_153: class Program_weight_tensor_parameter_154: name = "parameter_154" + original_name = "conv2d_45.w_0" shape = [192, 192, 1, 1] dtype = "float32" min_val = float("-0.0362525") @@ -1665,6 +1820,7 @@ class Program_weight_tensor_parameter_154: class Program_weight_tensor_parameter_155: name = "parameter_155" + original_name = "batch_norm2d_40.b_0" shape = [192] dtype = "float32" min_val = float("-0.0804559") @@ -1676,6 +1832,7 @@ class Program_weight_tensor_parameter_155: class Program_weight_tensor_parameter_156: name = "parameter_156" + original_name = "batch_norm2d_40.w_0" shape = [192] dtype = "float32" min_val = float("0.898348") @@ -1687,6 +1844,7 @@ class Program_weight_tensor_parameter_156: class Program_weight_tensor_parameter_157: name = "parameter_157" + original_name = "batch_norm2d_40.w_2" shape = [192] dtype = "float32" min_val = float("0.0238668") @@ -1698,6 +1856,7 @@ class Program_weight_tensor_parameter_157: class Program_weight_tensor_parameter_158: name = "parameter_158" + original_name = "batch_norm2d_40.w_1" shape = [192] dtype = "float32" min_val = float("-0.338547") @@ -1709,6 +1868,7 @@ class Program_weight_tensor_parameter_158: class Program_weight_tensor_parameter_159: name = "parameter_159" + original_name = "conv2d_44.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.0682399") @@ -1720,6 +1880,7 @@ class Program_weight_tensor_parameter_159: class Program_weight_tensor_parameter_160: name = "parameter_160" + original_name = "batch_norm2d_39.b_0" shape = [192] dtype = "float32" min_val = float("-0.283115") @@ -1731,6 +1892,7 @@ class Program_weight_tensor_parameter_160: class Program_weight_tensor_parameter_161: name = "parameter_161" + original_name = "batch_norm2d_39.w_0" shape = [192] dtype = "float32" min_val = float("0.866579") @@ -1742,6 +1904,7 @@ class Program_weight_tensor_parameter_161: class Program_weight_tensor_parameter_162: name = "parameter_162" + original_name = "batch_norm2d_39.w_2" shape = [192] dtype = "float32" min_val = float("0.0281452") @@ -1753,6 +1916,7 @@ class Program_weight_tensor_parameter_162: class Program_weight_tensor_parameter_163: name = "parameter_163" + original_name = "batch_norm2d_39.w_1" shape = [192] dtype = "float32" min_val = float("-0.464178") @@ -1764,6 +1928,7 @@ class Program_weight_tensor_parameter_163: class Program_weight_tensor_parameter_164: name = "parameter_164" + original_name = "conv2d_43.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.0784709") @@ -1775,6 +1940,7 @@ class Program_weight_tensor_parameter_164: class Program_weight_tensor_parameter_165: name = "parameter_165" + original_name = "batch_norm2d_38.b_0" shape = [192] dtype = "float32" min_val = float("-0.309798") @@ -1786,6 +1952,7 @@ class Program_weight_tensor_parameter_165: class Program_weight_tensor_parameter_166: name = "parameter_166" + original_name = "batch_norm2d_38.w_0" shape = [192] dtype = "float32" min_val = float("0.917672") @@ -1797,6 +1964,7 @@ class Program_weight_tensor_parameter_166: class Program_weight_tensor_parameter_167: name = "parameter_167" + original_name = "batch_norm2d_38.w_2" shape = [192] dtype = "float32" min_val = float("0.0152609") @@ -1808,6 +1976,7 @@ class Program_weight_tensor_parameter_167: class Program_weight_tensor_parameter_168: name = "parameter_168" + original_name = "batch_norm2d_38.w_1" shape = [192] dtype = "float32" min_val = float("-0.209367") @@ -1819,6 +1988,7 @@ class Program_weight_tensor_parameter_168: class Program_weight_tensor_parameter_169: name = "parameter_169" + original_name = "conv2d_42.w_0" shape = [192, 512, 1, 1] dtype = "float32" min_val = float("-0.111587") @@ -1830,6 +2000,7 @@ class Program_weight_tensor_parameter_169: class Program_weight_tensor_parameter_170: name = "parameter_170" + original_name = "batch_norm2d_37.b_0" shape = [192] dtype = "float32" min_val = float("-0.12447") @@ -1841,6 +2012,7 @@ class Program_weight_tensor_parameter_170: class Program_weight_tensor_parameter_171: name = "parameter_171" + original_name = "batch_norm2d_37.w_0" shape = [192] dtype = "float32" min_val = float("0.874474") @@ -1852,6 +2024,7 @@ class Program_weight_tensor_parameter_171: class Program_weight_tensor_parameter_172: name = "parameter_172" + original_name = "batch_norm2d_37.w_2" shape = [192] dtype = "float32" min_val = float("0.0106412") @@ -1863,6 +2036,7 @@ class Program_weight_tensor_parameter_172: class Program_weight_tensor_parameter_173: name = "parameter_173" + original_name = "batch_norm2d_37.w_1" shape = [192] dtype = "float32" min_val = float("-0.18243") @@ -1874,6 +2048,7 @@ class Program_weight_tensor_parameter_173: class Program_weight_tensor_parameter_174: name = "parameter_174" + original_name = "conv2d_41.w_0" shape = [192, 512, 1, 1] dtype = "float32" min_val = float("-0.0353564") @@ -1885,6 +2060,7 @@ class Program_weight_tensor_parameter_174: class Program_weight_tensor_parameter_175: name = "parameter_175" + original_name = "layer_norm_1.b_0_deepcopy_47" shape = [512] dtype = "float32" min_val = float("-4.43684e-10") @@ -1896,6 +2072,7 @@ class Program_weight_tensor_parameter_175: class Program_weight_tensor_parameter_176: name = "parameter_176" + original_name = "layer_norm_1.w_0_deepcopy_46" shape = [512] dtype = "float32" min_val = float("0.777531") @@ -1907,6 +2084,7 @@ class Program_weight_tensor_parameter_176: class Program_weight_tensor_parameter_177: name = "parameter_177" + original_name = "linear_2.b_0_deepcopy_43" shape = [512] dtype = "float32" min_val = float("-0.0178149") @@ -1918,6 +2096,7 @@ class Program_weight_tensor_parameter_177: class Program_weight_tensor_parameter_178: name = "parameter_178" + original_name = "linear_2.w_0_deepcopy_42" shape = [2048, 512] dtype = "float32" min_val = float("-0.0183338") @@ -1929,6 +2108,7 @@ class Program_weight_tensor_parameter_178: class Program_weight_tensor_parameter_179: name = "parameter_179" + original_name = "linear_1.b_0_deepcopy_41" shape = [2048] dtype = "float32" min_val = float("-0.0350612") @@ -1940,6 +2120,7 @@ class Program_weight_tensor_parameter_179: class Program_weight_tensor_parameter_180: name = "parameter_180" + original_name = "linear_1.w_0_deepcopy_40" shape = [512, 2048] dtype = "float32" min_val = float("-0.0353432") @@ -1951,6 +2132,7 @@ class Program_weight_tensor_parameter_180: class Program_weight_tensor_parameter_181: name = "parameter_181" + original_name = "layer_norm_0.b_0_deepcopy_45" shape = [512] dtype = "float32" min_val = float("-0.00108203") @@ -1962,6 +2144,7 @@ class Program_weight_tensor_parameter_181: class Program_weight_tensor_parameter_182: name = "parameter_182" + original_name = "layer_norm_0.w_0_deepcopy_44" shape = [512] dtype = "float32" min_val = float("0.787457") @@ -1973,6 +2156,7 @@ class Program_weight_tensor_parameter_182: class Program_weight_tensor_parameter_183: name = "parameter_183" + original_name = "linear_0.b_0_deepcopy_39" shape = [512] dtype = "float32" min_val = float("-0.000754274") @@ -1984,6 +2168,7 @@ class Program_weight_tensor_parameter_183: class Program_weight_tensor_parameter_184: name = "parameter_184" + original_name = "linear_0.w_0_deepcopy_38" shape = [512, 512] dtype = "float32" min_val = float("-0.0612193") @@ -1995,6 +2180,7 @@ class Program_weight_tensor_parameter_184: class Program_weight_tensor_parameter_185: name = "parameter_185" + original_name = "layer_norm_1.b_0_deepcopy_35" shape = [512] dtype = "float32" min_val = float("-0.000854187") @@ -2006,6 +2192,7 @@ class Program_weight_tensor_parameter_185: class Program_weight_tensor_parameter_186: name = "parameter_186" + original_name = "layer_norm_1.w_0_deepcopy_34" shape = [512] dtype = "float32" min_val = float("0.787903") @@ -2017,6 +2204,7 @@ class Program_weight_tensor_parameter_186: class Program_weight_tensor_parameter_187: name = "parameter_187" + original_name = "linear_2.b_0_deepcopy_31" shape = [512] dtype = "float32" min_val = float("-0.018138") @@ -2028,6 +2216,7 @@ class Program_weight_tensor_parameter_187: class Program_weight_tensor_parameter_188: name = "parameter_188" + original_name = "linear_2.w_0_deepcopy_30" shape = [2048, 512] dtype = "float32" min_val = float("-0.0182018") @@ -2039,6 +2228,7 @@ class Program_weight_tensor_parameter_188: class Program_weight_tensor_parameter_189: name = "parameter_189" + original_name = "linear_1.b_0_deepcopy_29" shape = [2048] dtype = "float32" min_val = float("-0.0349371") @@ -2050,6 +2240,7 @@ class Program_weight_tensor_parameter_189: class Program_weight_tensor_parameter_190: name = "parameter_190" + original_name = "linear_1.w_0_deepcopy_28" shape = [512, 2048] dtype = "float32" min_val = float("-0.0352741") @@ -2061,6 +2252,7 @@ class Program_weight_tensor_parameter_190: class Program_weight_tensor_parameter_191: name = "parameter_191" + original_name = "layer_norm_0.b_0_deepcopy_33" shape = [512] dtype = "float32" min_val = float("-0.000903085") @@ -2072,6 +2264,7 @@ class Program_weight_tensor_parameter_191: class Program_weight_tensor_parameter_192: name = "parameter_192" + original_name = "layer_norm_0.w_0_deepcopy_32" shape = [512] dtype = "float32" min_val = float("0.787892") @@ -2083,6 +2276,7 @@ class Program_weight_tensor_parameter_192: class Program_weight_tensor_parameter_193: name = "parameter_193" + original_name = "linear_0.b_0_deepcopy_27" shape = [512] dtype = "float32" min_val = float("-0.000858256") @@ -2094,6 +2288,7 @@ class Program_weight_tensor_parameter_193: class Program_weight_tensor_parameter_194: name = "parameter_194" + original_name = "linear_0.w_0_deepcopy_26" shape = [512, 512] dtype = "float32" min_val = float("-0.0609822") @@ -2105,6 +2300,7 @@ class Program_weight_tensor_parameter_194: class Program_weight_tensor_parameter_195: name = "parameter_195" + original_name = "layer_norm_1.b_0_deepcopy_23" shape = [512] dtype = "float32" min_val = float("-0.00117359") @@ -2116,6 +2312,7 @@ class Program_weight_tensor_parameter_195: class Program_weight_tensor_parameter_196: name = "parameter_196" + original_name = "layer_norm_1.w_0_deepcopy_22" shape = [512] dtype = "float32" min_val = float("0.788455") @@ -2127,6 +2324,7 @@ class Program_weight_tensor_parameter_196: class Program_weight_tensor_parameter_197: name = "parameter_197" + original_name = "linear_2.b_0_deepcopy_19" shape = [512] dtype = "float32" min_val = float("-0.018283") @@ -2138,6 +2336,7 @@ class Program_weight_tensor_parameter_197: class Program_weight_tensor_parameter_198: name = "parameter_198" + original_name = "linear_2.w_0_deepcopy_18" shape = [2048, 512] dtype = "float32" min_val = float("-0.0182138") @@ -2149,6 +2348,7 @@ class Program_weight_tensor_parameter_198: class Program_weight_tensor_parameter_199: name = "parameter_199" + original_name = "linear_1.b_0_deepcopy_17" shape = [2048] dtype = "float32" min_val = float("-0.0349256") @@ -2160,6 +2360,7 @@ class Program_weight_tensor_parameter_199: class Program_weight_tensor_parameter_200: name = "parameter_200" + original_name = "linear_1.w_0_deepcopy_16" shape = [512, 2048] dtype = "float32" min_val = float("-0.0353219") @@ -2171,6 +2372,7 @@ class Program_weight_tensor_parameter_200: class Program_weight_tensor_parameter_201: name = "parameter_201" + original_name = "layer_norm_0.b_0_deepcopy_21" shape = [512] dtype = "float32" min_val = float("-0.0012677") @@ -2182,6 +2384,7 @@ class Program_weight_tensor_parameter_201: class Program_weight_tensor_parameter_202: name = "parameter_202" + original_name = "layer_norm_0.w_0_deepcopy_20" shape = [512] dtype = "float32" min_val = float("0.787729") @@ -2193,6 +2396,7 @@ class Program_weight_tensor_parameter_202: class Program_weight_tensor_parameter_203: name = "parameter_203" + original_name = "linear_0.b_0_deepcopy_15" shape = [512] dtype = "float32" min_val = float("-0.00109096") @@ -2204,6 +2408,7 @@ class Program_weight_tensor_parameter_203: class Program_weight_tensor_parameter_204: name = "parameter_204" + original_name = "linear_0.w_0_deepcopy_14" shape = [512, 512] dtype = "float32" min_val = float("-0.0611516") @@ -2215,6 +2420,7 @@ class Program_weight_tensor_parameter_204: class Program_weight_tensor_parameter_205: name = "parameter_205" + original_name = "layer_norm_1.b_0_deepcopy_11" shape = [512] dtype = "float32" min_val = float("-0.00181093") @@ -2226,6 +2432,7 @@ class Program_weight_tensor_parameter_205: class Program_weight_tensor_parameter_206: name = "parameter_206" + original_name = "layer_norm_1.w_0_deepcopy_10" shape = [512] dtype = "float32" min_val = float("0.787859") @@ -2237,6 +2444,7 @@ class Program_weight_tensor_parameter_206: class Program_weight_tensor_parameter_207: name = "parameter_207" + original_name = "linear_2.b_0_deepcopy_7" shape = [512] dtype = "float32" min_val = float("-0.018331") @@ -2248,6 +2456,7 @@ class Program_weight_tensor_parameter_207: class Program_weight_tensor_parameter_208: name = "parameter_208" + original_name = "linear_2.w_0_deepcopy_6" shape = [2048, 512] dtype = "float32" min_val = float("-0.0182466") @@ -2259,6 +2468,7 @@ class Program_weight_tensor_parameter_208: class Program_weight_tensor_parameter_209: name = "parameter_209" + original_name = "linear_1.b_0_deepcopy_5" shape = [2048] dtype = "float32" min_val = float("-0.0349322") @@ -2270,6 +2480,7 @@ class Program_weight_tensor_parameter_209: class Program_weight_tensor_parameter_210: name = "parameter_210" + original_name = "linear_1.w_0_deepcopy_4" shape = [512, 2048] dtype = "float32" min_val = float("-0.0353813") @@ -2281,6 +2492,7 @@ class Program_weight_tensor_parameter_210: class Program_weight_tensor_parameter_211: name = "parameter_211" + original_name = "layer_norm_0.b_0_deepcopy_9" shape = [512] dtype = "float32" min_val = float("-0.0019486") @@ -2292,6 +2504,7 @@ class Program_weight_tensor_parameter_211: class Program_weight_tensor_parameter_212: name = "parameter_212" + original_name = "layer_norm_0.w_0_deepcopy_8" shape = [512] dtype = "float32" min_val = float("0.786434") @@ -2303,6 +2516,7 @@ class Program_weight_tensor_parameter_212: class Program_weight_tensor_parameter_213: name = "parameter_213" + original_name = "linear_0.b_0_deepcopy_3" shape = [512] dtype = "float32" min_val = float("-0.00303392") @@ -2314,6 +2528,7 @@ class Program_weight_tensor_parameter_213: class Program_weight_tensor_parameter_214: name = "parameter_214" + original_name = "linear_0.w_0_deepcopy_2" shape = [512, 512] dtype = "float32" min_val = float("-0.0615593") diff --git a/paddle_samples/PaddleX/PP-YOLOE_plus_SOD-S/subgraph_3/input_meta.py b/paddle_samples/PaddleX/PP-YOLOE_plus_SOD-S/subgraph_3/input_meta.py index 7c46d6c4a..901145ba5 100644 --- a/paddle_samples/PaddleX/PP-YOLOE_plus_SOD-S/subgraph_3/input_meta.py +++ b/paddle_samples/PaddleX/PP-YOLOE_plus_SOD-S/subgraph_3/input_meta.py @@ -147,7 +147,6 @@ class Program_weight_tensor_data_14: original_name = "var_1336" shape = [2, 3, 672, 672] dtype = "float32" - min_val = float("0.0") max_val = float("0.995113") mean = float("0.375674") std = float("0.170865") diff --git a/paddle_samples/PaddleX/PP-YOLOE_plus_SOD-largesize-L/subgraph_0/input_meta.py b/paddle_samples/PaddleX/PP-YOLOE_plus_SOD-largesize-L/subgraph_0/input_meta.py index 6b8c591d4..235199918 100644 --- a/paddle_samples/PaddleX/PP-YOLOE_plus_SOD-largesize-L/subgraph_0/input_meta.py +++ b/paddle_samples/PaddleX/PP-YOLOE_plus_SOD-largesize-L/subgraph_0/input_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_data_0: name = "data_0" + original_name = "param_120" shape = [1024, 3072] dtype = "float32" min_val = float("-0.0337707") @@ -11,6 +12,7 @@ class Program_weight_tensor_data_0: class Program_weight_tensor_data_1: name = "data_1" + original_name = "param_121" shape = [3072] dtype = "float32" min_val = float("-0.000858163") @@ -22,6 +24,7 @@ class Program_weight_tensor_data_1: class Program_weight_tensor_data_2: name = "data_2" + original_name = "param_122" shape = [1024, 3072] dtype = "float32" min_val = float("-0.0324395") @@ -33,6 +36,7 @@ class Program_weight_tensor_data_2: class Program_weight_tensor_data_3: name = "data_3" + original_name = "param_123" shape = [3072] dtype = "float32" min_val = float("-0.000630245") @@ -44,6 +48,7 @@ class Program_weight_tensor_data_3: class Program_weight_tensor_data_4: name = "data_4" + original_name = "param_124" shape = [1024, 3072] dtype = "float32" min_val = float("-0.0321875") @@ -55,6 +60,7 @@ class Program_weight_tensor_data_4: class Program_weight_tensor_data_5: name = "data_5" + original_name = "param_125" shape = [3072] dtype = "float32" min_val = float("-0.000429817") @@ -66,6 +72,7 @@ class Program_weight_tensor_data_5: class Program_weight_tensor_data_6: name = "data_6" + original_name = "param_126" shape = [1024, 3072] dtype = "float32" min_val = float("-0.0321313") @@ -77,6 +84,7 @@ class Program_weight_tensor_data_6: class Program_weight_tensor_data_7: name = "data_7" + original_name = "param_127" shape = [3072] dtype = "float32" min_val = float("-0.000397465") @@ -88,6 +96,7 @@ class Program_weight_tensor_data_7: class Program_weight_tensor_data_8: name = "data_8" + original_name = "var_6489" shape = [1, 256, 240, 240] dtype = "float32" min_val = float("-0.278465") @@ -99,6 +108,7 @@ class Program_weight_tensor_data_8: class Program_weight_tensor_data_9: name = "data_9" + original_name = "var_6490" shape = [1, 512, 120, 120] dtype = "float32" min_val = float("-0.278465") @@ -110,6 +120,7 @@ class Program_weight_tensor_data_9: class Program_weight_tensor_data_10: name = "data_10" + original_name = "var_6491" shape = [1, 1024, 60, 60] dtype = "float32" min_val = float("-0.278465") diff --git a/paddle_samples/PaddleX/PP-YOLOE_plus_SOD-largesize-L/subgraph_0/weight_meta.py b/paddle_samples/PaddleX/PP-YOLOE_plus_SOD-largesize-L/subgraph_0/weight_meta.py index bd82badb0..503ce58b3 100644 --- a/paddle_samples/PaddleX/PP-YOLOE_plus_SOD-largesize-L/subgraph_0/weight_meta.py +++ b/paddle_samples/PaddleX/PP-YOLOE_plus_SOD-largesize-L/subgraph_0/weight_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_parameter_0: name = "parameter_0" + original_name = "batch_norm2d_137.b_0" shape = [768] dtype = "float32" min_val = float("-0.175875") @@ -11,6 +12,7 @@ class Program_weight_tensor_parameter_0: class Program_weight_tensor_parameter_1: name = "parameter_1" + original_name = "batch_norm2d_137.w_0" shape = [768] dtype = "float32" min_val = float("0.939895") @@ -22,6 +24,7 @@ class Program_weight_tensor_parameter_1: class Program_weight_tensor_parameter_2: name = "parameter_2" + original_name = "batch_norm2d_137.w_2" shape = [768] dtype = "float32" min_val = float("0.00114665") @@ -33,6 +36,7 @@ class Program_weight_tensor_parameter_2: class Program_weight_tensor_parameter_3: name = "parameter_3" + original_name = "batch_norm2d_137.w_1" shape = [768] dtype = "float32" min_val = float("-0.134835") @@ -44,6 +48,7 @@ class Program_weight_tensor_parameter_3: class Program_weight_tensor_parameter_4: name = "parameter_4" + original_name = "conv2d_141.w_0" shape = [768, 768, 1, 1] dtype = "float32" min_val = float("-0.0548133") @@ -55,6 +60,7 @@ class Program_weight_tensor_parameter_4: class Program_weight_tensor_parameter_5: name = "parameter_5" + original_name = "batch_norm2d_136.b_0" shape = [384] dtype = "float32" min_val = float("-0.14169") @@ -66,6 +72,7 @@ class Program_weight_tensor_parameter_5: class Program_weight_tensor_parameter_6: name = "parameter_6" + original_name = "batch_norm2d_136.w_0" shape = [384] dtype = "float32" min_val = float("0.945748") @@ -77,6 +84,7 @@ class Program_weight_tensor_parameter_6: class Program_weight_tensor_parameter_7: name = "parameter_7" + original_name = "batch_norm2d_136.w_2" shape = [384] dtype = "float32" min_val = float("0.000803951") @@ -88,6 +96,7 @@ class Program_weight_tensor_parameter_7: class Program_weight_tensor_parameter_8: name = "parameter_8" + original_name = "batch_norm2d_136.w_1" shape = [384] dtype = "float32" min_val = float("-0.0551849") @@ -99,6 +108,7 @@ class Program_weight_tensor_parameter_8: class Program_weight_tensor_parameter_9: name = "parameter_9" + original_name = "conv2d_140.w_0" shape = [384, 384, 1, 1] dtype = "float32" min_val = float("-0.0299323") @@ -110,6 +120,7 @@ class Program_weight_tensor_parameter_9: class Program_weight_tensor_parameter_10: name = "parameter_10" + original_name = "batch_norm2d_135.b_0" shape = [384] dtype = "float32" min_val = float("-0.14169") @@ -121,6 +132,7 @@ class Program_weight_tensor_parameter_10: class Program_weight_tensor_parameter_11: name = "parameter_11" + original_name = "batch_norm2d_135.w_0" shape = [384] dtype = "float32" min_val = float("0.968039") @@ -132,6 +144,7 @@ class Program_weight_tensor_parameter_11: class Program_weight_tensor_parameter_12: name = "parameter_12" + original_name = "batch_norm2d_135.w_2" shape = [384] dtype = "float32" min_val = float("0.00197981") @@ -143,6 +156,7 @@ class Program_weight_tensor_parameter_12: class Program_weight_tensor_parameter_13: name = "parameter_13" + original_name = "batch_norm2d_135.w_1" shape = [384] dtype = "float32" min_val = float("-0.202706") @@ -154,6 +168,7 @@ class Program_weight_tensor_parameter_13: class Program_weight_tensor_parameter_14: name = "parameter_14" + original_name = "conv2d_139.w_0" shape = [384, 384, 3, 3] dtype = "float32" min_val = float("-0.029908") @@ -165,6 +180,7 @@ class Program_weight_tensor_parameter_14: class Program_weight_tensor_parameter_15: name = "parameter_15" + original_name = "batch_norm2d_134.b_0" shape = [384] dtype = "float32" min_val = float("-0.170219") @@ -176,6 +192,7 @@ class Program_weight_tensor_parameter_15: class Program_weight_tensor_parameter_16: name = "parameter_16" + original_name = "batch_norm2d_134.w_0" shape = [384] dtype = "float32" min_val = float("0.975222") @@ -187,6 +204,7 @@ class Program_weight_tensor_parameter_16: class Program_weight_tensor_parameter_17: name = "parameter_17" + original_name = "batch_norm2d_134.w_2" shape = [384] dtype = "float32" min_val = float("0.00530043") @@ -198,6 +216,7 @@ class Program_weight_tensor_parameter_17: class Program_weight_tensor_parameter_18: name = "parameter_18" + original_name = "batch_norm2d_134.w_1" shape = [384] dtype = "float32" min_val = float("-0.265254") @@ -209,6 +228,7 @@ class Program_weight_tensor_parameter_18: class Program_weight_tensor_parameter_19: name = "parameter_19" + original_name = "conv2d_138.w_0" shape = [384, 384, 3, 3] dtype = "float32" min_val = float("-0.0331338") @@ -220,6 +240,7 @@ class Program_weight_tensor_parameter_19: class Program_weight_tensor_parameter_20: name = "parameter_20" + original_name = "batch_norm2d_133.b_0" shape = [384] dtype = "float32" min_val = float("-0.105219") @@ -231,6 +252,7 @@ class Program_weight_tensor_parameter_20: class Program_weight_tensor_parameter_21: name = "parameter_21" + original_name = "batch_norm2d_133.w_0" shape = [384] dtype = "float32" min_val = float("0.945357") @@ -242,6 +264,7 @@ class Program_weight_tensor_parameter_21: class Program_weight_tensor_parameter_22: name = "parameter_22" + original_name = "batch_norm2d_133.w_2" shape = [384] dtype = "float32" min_val = float("0.000690331") @@ -253,6 +276,7 @@ class Program_weight_tensor_parameter_22: class Program_weight_tensor_parameter_23: name = "parameter_23" + original_name = "batch_norm2d_133.w_1" shape = [384] dtype = "float32" min_val = float("-0.0849168") @@ -264,6 +288,7 @@ class Program_weight_tensor_parameter_23: class Program_weight_tensor_parameter_24: name = "parameter_24" + original_name = "conv2d_137.w_0" shape = [384, 384, 1, 1] dtype = "float32" min_val = float("-0.0267959") @@ -275,6 +300,7 @@ class Program_weight_tensor_parameter_24: class Program_weight_tensor_parameter_25: name = "parameter_25" + original_name = "batch_norm2d_132.b_0" shape = [384] dtype = "float32" min_val = float("-0.105219") @@ -286,6 +312,7 @@ class Program_weight_tensor_parameter_25: class Program_weight_tensor_parameter_26: name = "parameter_26" + original_name = "batch_norm2d_132.w_0" shape = [384] dtype = "float32" min_val = float("0.959852") @@ -297,6 +324,7 @@ class Program_weight_tensor_parameter_26: class Program_weight_tensor_parameter_27: name = "parameter_27" + original_name = "batch_norm2d_132.w_2" shape = [384] dtype = "float32" min_val = float("0.00248164") @@ -308,6 +336,7 @@ class Program_weight_tensor_parameter_27: class Program_weight_tensor_parameter_28: name = "parameter_28" + original_name = "batch_norm2d_132.w_1" shape = [384] dtype = "float32" min_val = float("-0.215332") @@ -319,6 +348,7 @@ class Program_weight_tensor_parameter_28: class Program_weight_tensor_parameter_29: name = "parameter_29" + original_name = "conv2d_136.w_0" shape = [384, 384, 3, 3] dtype = "float32" min_val = float("-0.0363929") @@ -330,6 +360,7 @@ class Program_weight_tensor_parameter_29: class Program_weight_tensor_parameter_30: name = "parameter_30" + original_name = "batch_norm2d_131.b_0" shape = [384] dtype = "float32" min_val = float("-0.0896627") @@ -341,6 +372,7 @@ class Program_weight_tensor_parameter_30: class Program_weight_tensor_parameter_31: name = "parameter_31" + original_name = "batch_norm2d_131.w_0" shape = [384] dtype = "float32" min_val = float("0.933291") @@ -352,6 +384,7 @@ class Program_weight_tensor_parameter_31: class Program_weight_tensor_parameter_32: name = "parameter_32" + original_name = "batch_norm2d_131.w_2" shape = [384] dtype = "float32" min_val = float("0.00571017") @@ -363,6 +396,7 @@ class Program_weight_tensor_parameter_32: class Program_weight_tensor_parameter_33: name = "parameter_33" + original_name = "batch_norm2d_131.w_1" shape = [384] dtype = "float32" min_val = float("-0.241704") @@ -374,6 +408,7 @@ class Program_weight_tensor_parameter_33: class Program_weight_tensor_parameter_34: name = "parameter_34" + original_name = "conv2d_135.w_0" shape = [384, 384, 3, 3] dtype = "float32" min_val = float("-0.0397047") @@ -385,6 +420,7 @@ class Program_weight_tensor_parameter_34: class Program_weight_tensor_parameter_35: name = "parameter_35" + original_name = "batch_norm2d_130.b_0" shape = [384] dtype = "float32" min_val = float("-0.116341") @@ -396,6 +432,7 @@ class Program_weight_tensor_parameter_35: class Program_weight_tensor_parameter_36: name = "parameter_36" + original_name = "batch_norm2d_130.w_0" shape = [384] dtype = "float32" min_val = float("0.929383") @@ -407,6 +444,7 @@ class Program_weight_tensor_parameter_36: class Program_weight_tensor_parameter_37: name = "parameter_37" + original_name = "batch_norm2d_130.w_2" shape = [384] dtype = "float32" min_val = float("0.00125121") @@ -418,6 +456,7 @@ class Program_weight_tensor_parameter_37: class Program_weight_tensor_parameter_38: name = "parameter_38" + original_name = "batch_norm2d_130.w_1" shape = [384] dtype = "float32" min_val = float("-0.0558903") @@ -429,6 +468,7 @@ class Program_weight_tensor_parameter_38: class Program_weight_tensor_parameter_39: name = "parameter_39" + original_name = "conv2d_134.w_0" shape = [384, 384, 1, 1] dtype = "float32" min_val = float("-0.0386337") @@ -440,6 +480,7 @@ class Program_weight_tensor_parameter_39: class Program_weight_tensor_parameter_40: name = "parameter_40" + original_name = "batch_norm2d_129.b_0" shape = [384] dtype = "float32" min_val = float("-0.116341") @@ -451,6 +492,7 @@ class Program_weight_tensor_parameter_40: class Program_weight_tensor_parameter_41: name = "parameter_41" + original_name = "batch_norm2d_129.w_0" shape = [384] dtype = "float32" min_val = float("0.981354") @@ -462,6 +504,7 @@ class Program_weight_tensor_parameter_41: class Program_weight_tensor_parameter_42: name = "parameter_42" + original_name = "batch_norm2d_129.w_2" shape = [384] dtype = "float32" min_val = float("0.00487818") @@ -473,6 +516,7 @@ class Program_weight_tensor_parameter_42: class Program_weight_tensor_parameter_43: name = "parameter_43" + original_name = "batch_norm2d_129.w_1" shape = [384] dtype = "float32" min_val = float("-0.191838") @@ -484,6 +528,7 @@ class Program_weight_tensor_parameter_43: class Program_weight_tensor_parameter_44: name = "parameter_44" + original_name = "conv2d_133.w_0" shape = [384, 384, 3, 3] dtype = "float32" min_val = float("-0.0360859") @@ -495,6 +540,7 @@ class Program_weight_tensor_parameter_44: class Program_weight_tensor_parameter_45: name = "parameter_45" + original_name = "batch_norm2d_128.b_0" shape = [384] dtype = "float32" min_val = float("-0.107113") @@ -506,6 +552,7 @@ class Program_weight_tensor_parameter_45: class Program_weight_tensor_parameter_46: name = "parameter_46" + original_name = "batch_norm2d_128.w_0" shape = [384] dtype = "float32" min_val = float("0.944795") @@ -517,6 +564,7 @@ class Program_weight_tensor_parameter_46: class Program_weight_tensor_parameter_47: name = "parameter_47" + original_name = "batch_norm2d_128.w_2" shape = [384] dtype = "float32" min_val = float("0.00539047") @@ -528,6 +576,7 @@ class Program_weight_tensor_parameter_47: class Program_weight_tensor_parameter_48: name = "parameter_48" + original_name = "batch_norm2d_128.w_1" shape = [384] dtype = "float32" min_val = float("-0.154913") @@ -539,6 +588,7 @@ class Program_weight_tensor_parameter_48: class Program_weight_tensor_parameter_49: name = "parameter_49" + original_name = "conv2d_132.w_0" shape = [384, 384, 3, 3] dtype = "float32" min_val = float("-0.0279281") @@ -550,6 +600,7 @@ class Program_weight_tensor_parameter_49: class Program_weight_tensor_parameter_50: name = "parameter_50" + original_name = "batch_norm2d_127.b_0" shape = [384] dtype = "float32" min_val = float("-0.10674") @@ -561,6 +612,7 @@ class Program_weight_tensor_parameter_50: class Program_weight_tensor_parameter_51: name = "parameter_51" + original_name = "batch_norm2d_127.w_0" shape = [384] dtype = "float32" min_val = float("0.973756") @@ -572,6 +624,7 @@ class Program_weight_tensor_parameter_51: class Program_weight_tensor_parameter_52: name = "parameter_52" + original_name = "batch_norm2d_127.w_2" shape = [384] dtype = "float32" min_val = float("0.00231428") @@ -583,6 +636,7 @@ class Program_weight_tensor_parameter_52: class Program_weight_tensor_parameter_53: name = "parameter_53" + original_name = "batch_norm2d_127.w_1" shape = [384] dtype = "float32" min_val = float("-0.10048") @@ -594,6 +648,7 @@ class Program_weight_tensor_parameter_53: class Program_weight_tensor_parameter_54: name = "parameter_54" + original_name = "conv2d_131.w_0" shape = [384, 1152, 1, 1] dtype = "float32" min_val = float("-0.0619005") @@ -605,6 +660,7 @@ class Program_weight_tensor_parameter_54: class Program_weight_tensor_parameter_55: name = "parameter_55" + original_name = "batch_norm2d_126.b_0" shape = [384] dtype = "float32" min_val = float("-0.0424904") @@ -616,6 +672,7 @@ class Program_weight_tensor_parameter_55: class Program_weight_tensor_parameter_56: name = "parameter_56" + original_name = "batch_norm2d_126.w_0" shape = [384] dtype = "float32" min_val = float("0.959519") @@ -627,6 +684,7 @@ class Program_weight_tensor_parameter_56: class Program_weight_tensor_parameter_57: name = "parameter_57" + original_name = "batch_norm2d_126.w_2" shape = [384] dtype = "float32" min_val = float("0.00124549") @@ -638,6 +696,7 @@ class Program_weight_tensor_parameter_57: class Program_weight_tensor_parameter_58: name = "parameter_58" + original_name = "batch_norm2d_126.w_1" shape = [384] dtype = "float32" min_val = float("-0.110999") @@ -649,6 +708,7 @@ class Program_weight_tensor_parameter_58: class Program_weight_tensor_parameter_59: name = "parameter_59" + original_name = "conv2d_130.w_0" shape = [384, 1152, 1, 1] dtype = "float32" min_val = float("-0.0245473") @@ -660,6 +720,7 @@ class Program_weight_tensor_parameter_59: class Program_weight_tensor_parameter_60: name = "parameter_60" + original_name = "batch_norm2d_125.b_0" shape = [384] dtype = "float32" min_val = float("-0.0529748") @@ -671,6 +732,7 @@ class Program_weight_tensor_parameter_60: class Program_weight_tensor_parameter_61: name = "parameter_61" + original_name = "batch_norm2d_125.w_0" shape = [384] dtype = "float32" min_val = float("0.988678") @@ -682,6 +744,7 @@ class Program_weight_tensor_parameter_61: class Program_weight_tensor_parameter_62: name = "parameter_62" + original_name = "batch_norm2d_125.w_2" shape = [384] dtype = "float32" min_val = float("0.00468338") @@ -693,6 +756,7 @@ class Program_weight_tensor_parameter_62: class Program_weight_tensor_parameter_63: name = "parameter_63" + original_name = "batch_norm2d_125.w_1" shape = [384] dtype = "float32" min_val = float("-0.44327") @@ -704,6 +768,7 @@ class Program_weight_tensor_parameter_63: class Program_weight_tensor_parameter_64: name = "parameter_64" + original_name = "conv2d_129.w_0" shape = [384, 384, 3, 3] dtype = "float32" min_val = float("-0.0212973") @@ -715,6 +780,7 @@ class Program_weight_tensor_parameter_64: class Program_weight_tensor_parameter_65: name = "parameter_65" + original_name = "batch_norm2d_124.b_0" shape = [384] dtype = "float32" min_val = float("-0.222314") @@ -726,6 +792,7 @@ class Program_weight_tensor_parameter_65: class Program_weight_tensor_parameter_66: name = "parameter_66" + original_name = "batch_norm2d_124.w_0" shape = [384] dtype = "float32" min_val = float("0.919259") @@ -737,6 +804,7 @@ class Program_weight_tensor_parameter_66: class Program_weight_tensor_parameter_67: name = "parameter_67" + original_name = "batch_norm2d_124.w_2" shape = [384] dtype = "float32" min_val = float("0.00374913") @@ -748,6 +816,7 @@ class Program_weight_tensor_parameter_67: class Program_weight_tensor_parameter_68: name = "parameter_68" + original_name = "batch_norm2d_124.w_1" shape = [384] dtype = "float32" min_val = float("-0.129657") @@ -759,6 +828,7 @@ class Program_weight_tensor_parameter_68: class Program_weight_tensor_parameter_69: name = "parameter_69" + original_name = "conv2d_128.w_0" shape = [384, 384, 1, 1] dtype = "float32" min_val = float("-0.0788092") @@ -770,6 +840,7 @@ class Program_weight_tensor_parameter_69: class Program_weight_tensor_parameter_70: name = "parameter_70" + original_name = "batch_norm2d_123.b_0" shape = [192] dtype = "float32" min_val = float("-0.165903") @@ -781,6 +852,7 @@ class Program_weight_tensor_parameter_70: class Program_weight_tensor_parameter_71: name = "parameter_71" + original_name = "batch_norm2d_123.w_0" shape = [192] dtype = "float32" min_val = float("0.841187") @@ -792,6 +864,7 @@ class Program_weight_tensor_parameter_71: class Program_weight_tensor_parameter_72: name = "parameter_72" + original_name = "batch_norm2d_123.w_2" shape = [192] dtype = "float32" min_val = float("0.00140171") @@ -803,6 +876,7 @@ class Program_weight_tensor_parameter_72: class Program_weight_tensor_parameter_73: name = "parameter_73" + original_name = "batch_norm2d_123.w_1" shape = [192] dtype = "float32" min_val = float("-0.0638207") @@ -814,6 +888,7 @@ class Program_weight_tensor_parameter_73: class Program_weight_tensor_parameter_74: name = "parameter_74" + original_name = "conv2d_127.w_0" shape = [192, 192, 1, 1] dtype = "float32" min_val = float("-0.0496362") @@ -825,6 +900,7 @@ class Program_weight_tensor_parameter_74: class Program_weight_tensor_parameter_75: name = "parameter_75" + original_name = "batch_norm2d_122.b_0" shape = [192] dtype = "float32" min_val = float("-0.165903") @@ -836,6 +912,7 @@ class Program_weight_tensor_parameter_75: class Program_weight_tensor_parameter_76: name = "parameter_76" + original_name = "batch_norm2d_122.w_0" shape = [192] dtype = "float32" min_val = float("0.729841") @@ -847,6 +924,7 @@ class Program_weight_tensor_parameter_76: class Program_weight_tensor_parameter_77: name = "parameter_77" + original_name = "batch_norm2d_122.w_2" shape = [192] dtype = "float32" min_val = float("0.00540078") @@ -858,6 +936,7 @@ class Program_weight_tensor_parameter_77: class Program_weight_tensor_parameter_78: name = "parameter_78" + original_name = "batch_norm2d_122.w_1" shape = [192] dtype = "float32" min_val = float("-0.219646") @@ -869,6 +948,7 @@ class Program_weight_tensor_parameter_78: class Program_weight_tensor_parameter_79: name = "parameter_79" + original_name = "conv2d_126.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.0430374") @@ -880,6 +960,7 @@ class Program_weight_tensor_parameter_79: class Program_weight_tensor_parameter_80: name = "parameter_80" + original_name = "batch_norm2d_121.b_0" shape = [192] dtype = "float32" min_val = float("-0.191344") @@ -891,6 +972,7 @@ class Program_weight_tensor_parameter_80: class Program_weight_tensor_parameter_81: name = "parameter_81" + original_name = "batch_norm2d_121.w_0" shape = [192] dtype = "float32" min_val = float("0.897737") @@ -902,6 +984,7 @@ class Program_weight_tensor_parameter_81: class Program_weight_tensor_parameter_82: name = "parameter_82" + original_name = "batch_norm2d_121.w_2" shape = [192] dtype = "float32" min_val = float("0.010406") @@ -913,6 +996,7 @@ class Program_weight_tensor_parameter_82: class Program_weight_tensor_parameter_83: name = "parameter_83" + original_name = "batch_norm2d_121.w_1" shape = [192] dtype = "float32" min_val = float("-0.296379") @@ -924,6 +1008,7 @@ class Program_weight_tensor_parameter_83: class Program_weight_tensor_parameter_84: name = "parameter_84" + original_name = "conv2d_125.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.0473781") @@ -935,6 +1020,7 @@ class Program_weight_tensor_parameter_84: class Program_weight_tensor_parameter_85: name = "parameter_85" + original_name = "batch_norm2d_120.b_0" shape = [192] dtype = "float32" min_val = float("-0.191632") @@ -946,6 +1032,7 @@ class Program_weight_tensor_parameter_85: class Program_weight_tensor_parameter_86: name = "parameter_86" + original_name = "batch_norm2d_120.w_0" shape = [192] dtype = "float32" min_val = float("0.922153") @@ -957,6 +1044,7 @@ class Program_weight_tensor_parameter_86: class Program_weight_tensor_parameter_87: name = "parameter_87" + original_name = "batch_norm2d_120.w_2" shape = [192] dtype = "float32" min_val = float("0.00109672") @@ -968,6 +1056,7 @@ class Program_weight_tensor_parameter_87: class Program_weight_tensor_parameter_88: name = "parameter_88" + original_name = "batch_norm2d_120.w_1" shape = [192] dtype = "float32" min_val = float("-0.0707542") @@ -979,6 +1068,7 @@ class Program_weight_tensor_parameter_88: class Program_weight_tensor_parameter_89: name = "parameter_89" + original_name = "conv2d_124.w_0" shape = [192, 192, 1, 1] dtype = "float32" min_val = float("-0.0386532") @@ -990,6 +1080,7 @@ class Program_weight_tensor_parameter_89: class Program_weight_tensor_parameter_90: name = "parameter_90" + original_name = "batch_norm2d_119.b_0" shape = [192] dtype = "float32" min_val = float("-0.191632") @@ -1001,6 +1092,7 @@ class Program_weight_tensor_parameter_90: class Program_weight_tensor_parameter_91: name = "parameter_91" + original_name = "batch_norm2d_119.w_0" shape = [192] dtype = "float32" min_val = float("0.968104") @@ -1012,6 +1104,7 @@ class Program_weight_tensor_parameter_91: class Program_weight_tensor_parameter_92: name = "parameter_92" + original_name = "batch_norm2d_119.w_2" shape = [192] dtype = "float32" min_val = float("0.00424155") @@ -1023,6 +1116,7 @@ class Program_weight_tensor_parameter_92: class Program_weight_tensor_parameter_93: name = "parameter_93" + original_name = "batch_norm2d_119.w_1" shape = [192] dtype = "float32" min_val = float("-0.186844") @@ -1034,6 +1128,7 @@ class Program_weight_tensor_parameter_93: class Program_weight_tensor_parameter_94: name = "parameter_94" + original_name = "conv2d_123.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.0471115") @@ -1045,6 +1140,7 @@ class Program_weight_tensor_parameter_94: class Program_weight_tensor_parameter_95: name = "parameter_95" + original_name = "batch_norm2d_118.b_0" shape = [192] dtype = "float32" min_val = float("-0.188926") @@ -1056,6 +1152,7 @@ class Program_weight_tensor_parameter_95: class Program_weight_tensor_parameter_96: name = "parameter_96" + original_name = "batch_norm2d_118.w_0" shape = [192] dtype = "float32" min_val = float("0.880419") @@ -1067,6 +1164,7 @@ class Program_weight_tensor_parameter_96: class Program_weight_tensor_parameter_97: name = "parameter_97" + original_name = "batch_norm2d_118.w_2" shape = [192] dtype = "float32" min_val = float("0.00813457") @@ -1078,6 +1176,7 @@ class Program_weight_tensor_parameter_97: class Program_weight_tensor_parameter_98: name = "parameter_98" + original_name = "batch_norm2d_118.w_1" shape = [192] dtype = "float32" min_val = float("-0.123249") @@ -1089,6 +1188,7 @@ class Program_weight_tensor_parameter_98: class Program_weight_tensor_parameter_99: name = "parameter_99" + original_name = "conv2d_122.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.0427988") @@ -1100,6 +1200,7 @@ class Program_weight_tensor_parameter_99: class Program_weight_tensor_parameter_100: name = "parameter_100" + original_name = "batch_norm2d_117.b_0" shape = [192] dtype = "float32" min_val = float("-0.229476") @@ -1111,6 +1212,7 @@ class Program_weight_tensor_parameter_100: class Program_weight_tensor_parameter_101: name = "parameter_101" + original_name = "batch_norm2d_117.w_0" shape = [192] dtype = "float32" min_val = float("0.900428") @@ -1122,6 +1224,7 @@ class Program_weight_tensor_parameter_101: class Program_weight_tensor_parameter_102: name = "parameter_102" + original_name = "batch_norm2d_117.w_2" shape = [192] dtype = "float32" min_val = float("0.00166927") @@ -1133,6 +1236,7 @@ class Program_weight_tensor_parameter_102: class Program_weight_tensor_parameter_103: name = "parameter_103" + original_name = "batch_norm2d_117.w_1" shape = [192] dtype = "float32" min_val = float("-0.0393056") @@ -1144,6 +1248,7 @@ class Program_weight_tensor_parameter_103: class Program_weight_tensor_parameter_104: name = "parameter_104" + original_name = "conv2d_121.w_0" shape = [192, 192, 1, 1] dtype = "float32" min_val = float("-0.0436514") @@ -1155,6 +1260,7 @@ class Program_weight_tensor_parameter_104: class Program_weight_tensor_parameter_105: name = "parameter_105" + original_name = "batch_norm2d_116.b_0" shape = [192] dtype = "float32" min_val = float("-0.229476") @@ -1166,6 +1272,7 @@ class Program_weight_tensor_parameter_105: class Program_weight_tensor_parameter_106: name = "parameter_106" + original_name = "batch_norm2d_116.w_0" shape = [192] dtype = "float32" min_val = float("0.947654") @@ -1177,6 +1284,7 @@ class Program_weight_tensor_parameter_106: class Program_weight_tensor_parameter_107: name = "parameter_107" + original_name = "batch_norm2d_116.w_2" shape = [192] dtype = "float32" min_val = float("0.00719786") @@ -1188,6 +1296,7 @@ class Program_weight_tensor_parameter_107: class Program_weight_tensor_parameter_108: name = "parameter_108" + original_name = "batch_norm2d_116.w_1" shape = [192] dtype = "float32" min_val = float("-0.131118") @@ -1199,6 +1308,7 @@ class Program_weight_tensor_parameter_108: class Program_weight_tensor_parameter_109: name = "parameter_109" + original_name = "conv2d_120.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.0485053") @@ -1210,6 +1320,7 @@ class Program_weight_tensor_parameter_109: class Program_weight_tensor_parameter_110: name = "parameter_110" + original_name = "batch_norm2d_115.b_0" shape = [192] dtype = "float32" min_val = float("-0.234305") @@ -1221,6 +1332,7 @@ class Program_weight_tensor_parameter_110: class Program_weight_tensor_parameter_111: name = "parameter_111" + original_name = "batch_norm2d_115.w_0" shape = [192] dtype = "float32" min_val = float("0.886145") @@ -1232,6 +1344,7 @@ class Program_weight_tensor_parameter_111: class Program_weight_tensor_parameter_112: name = "parameter_112" + original_name = "batch_norm2d_115.w_2" shape = [192] dtype = "float32" min_val = float("0.00868237") @@ -1243,6 +1356,7 @@ class Program_weight_tensor_parameter_112: class Program_weight_tensor_parameter_113: name = "parameter_113" + original_name = "batch_norm2d_115.w_1" shape = [192] dtype = "float32" min_val = float("-0.181299") @@ -1254,6 +1368,7 @@ class Program_weight_tensor_parameter_113: class Program_weight_tensor_parameter_114: name = "parameter_114" + original_name = "conv2d_119.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.0410138") @@ -1265,6 +1380,7 @@ class Program_weight_tensor_parameter_114: class Program_weight_tensor_parameter_115: name = "parameter_115" + original_name = "batch_norm2d_114.b_0" shape = [192] dtype = "float32" min_val = float("-0.199948") @@ -1276,6 +1392,7 @@ class Program_weight_tensor_parameter_115: class Program_weight_tensor_parameter_116: name = "parameter_116" + original_name = "batch_norm2d_114.w_0" shape = [192] dtype = "float32" min_val = float("0.925493") @@ -1287,6 +1404,7 @@ class Program_weight_tensor_parameter_116: class Program_weight_tensor_parameter_117: name = "parameter_117" + original_name = "batch_norm2d_114.w_2" shape = [192] dtype = "float32" min_val = float("0.00445121") @@ -1298,6 +1416,7 @@ class Program_weight_tensor_parameter_117: class Program_weight_tensor_parameter_118: name = "parameter_118" + original_name = "batch_norm2d_114.w_1" shape = [192] dtype = "float32" min_val = float("-0.0890928") @@ -1309,6 +1428,7 @@ class Program_weight_tensor_parameter_118: class Program_weight_tensor_parameter_119: name = "parameter_119" + original_name = "conv2d_118.w_0" shape = [192, 576, 1, 1] dtype = "float32" min_val = float("-0.0628757") @@ -1320,6 +1440,7 @@ class Program_weight_tensor_parameter_119: class Program_weight_tensor_parameter_120: name = "parameter_120" + original_name = "batch_norm2d_113.b_0" shape = [192] dtype = "float32" min_val = float("-0.099963") @@ -1331,6 +1452,7 @@ class Program_weight_tensor_parameter_120: class Program_weight_tensor_parameter_121: name = "parameter_121" + original_name = "batch_norm2d_113.w_0" shape = [192] dtype = "float32" min_val = float("0.923856") @@ -1342,6 +1464,7 @@ class Program_weight_tensor_parameter_121: class Program_weight_tensor_parameter_122: name = "parameter_122" + original_name = "batch_norm2d_113.w_2" shape = [192] dtype = "float32" min_val = float("0.00335112") @@ -1353,6 +1476,7 @@ class Program_weight_tensor_parameter_122: class Program_weight_tensor_parameter_123: name = "parameter_123" + original_name = "batch_norm2d_113.w_1" shape = [192] dtype = "float32" min_val = float("-0.0728879") @@ -1364,6 +1488,7 @@ class Program_weight_tensor_parameter_123: class Program_weight_tensor_parameter_124: name = "parameter_124" + original_name = "conv2d_117.w_0" shape = [192, 576, 1, 1] dtype = "float32" min_val = float("-0.0557051") @@ -1375,6 +1500,7 @@ class Program_weight_tensor_parameter_124: class Program_weight_tensor_parameter_125: name = "parameter_125" + original_name = "batch_norm2d_112.b_0" shape = [192] dtype = "float32" min_val = float("-0.15908") @@ -1386,6 +1512,7 @@ class Program_weight_tensor_parameter_125: class Program_weight_tensor_parameter_126: name = "parameter_126" + original_name = "batch_norm2d_112.w_0" shape = [192] dtype = "float32" min_val = float("0.921159") @@ -1397,6 +1524,7 @@ class Program_weight_tensor_parameter_126: class Program_weight_tensor_parameter_127: name = "parameter_127" + original_name = "batch_norm2d_112.w_2" shape = [192] dtype = "float32" min_val = float("0.00450918") @@ -1408,6 +1536,7 @@ class Program_weight_tensor_parameter_127: class Program_weight_tensor_parameter_128: name = "parameter_128" + original_name = "batch_norm2d_112.w_1" shape = [192] dtype = "float32" min_val = float("-0.400402") @@ -1419,6 +1548,7 @@ class Program_weight_tensor_parameter_128: class Program_weight_tensor_parameter_129: name = "parameter_129" + original_name = "conv2d_116.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.0350247") @@ -1430,6 +1560,7 @@ class Program_weight_tensor_parameter_129: class Program_weight_tensor_parameter_130: name = "parameter_130" + original_name = "batch_norm2d_111.b_0" shape = [192] dtype = "float32" min_val = float("-0.552249") @@ -1441,6 +1572,7 @@ class Program_weight_tensor_parameter_130: class Program_weight_tensor_parameter_131: name = "parameter_131" + original_name = "batch_norm2d_111.w_0" shape = [192] dtype = "float32" min_val = float("0.541478") @@ -1452,6 +1584,7 @@ class Program_weight_tensor_parameter_131: class Program_weight_tensor_parameter_132: name = "parameter_132" + original_name = "batch_norm2d_111.w_2" shape = [192] dtype = "float32" min_val = float("0.00575627") @@ -1463,6 +1596,7 @@ class Program_weight_tensor_parameter_132: class Program_weight_tensor_parameter_133: name = "parameter_133" + original_name = "batch_norm2d_111.w_1" shape = [192] dtype = "float32" min_val = float("-0.182786") @@ -1474,6 +1608,7 @@ class Program_weight_tensor_parameter_133: class Program_weight_tensor_parameter_134: name = "parameter_134" + original_name = "conv2d_115.w_0" shape = [192, 192, 1, 1] dtype = "float32" min_val = float("-0.14008") @@ -1485,6 +1620,7 @@ class Program_weight_tensor_parameter_134: class Program_weight_tensor_parameter_135: name = "parameter_135" + original_name = "batch_norm2d_110.b_0" shape = [96] dtype = "float32" min_val = float("-0.457965") @@ -1496,6 +1632,7 @@ class Program_weight_tensor_parameter_135: class Program_weight_tensor_parameter_136: name = "parameter_136" + original_name = "batch_norm2d_110.w_0" shape = [96] dtype = "float32" min_val = float("0.762871") @@ -1507,6 +1644,7 @@ class Program_weight_tensor_parameter_136: class Program_weight_tensor_parameter_137: name = "parameter_137" + original_name = "batch_norm2d_110.w_2" shape = [96] dtype = "float32" min_val = float("0.00291973") @@ -1518,6 +1656,7 @@ class Program_weight_tensor_parameter_137: class Program_weight_tensor_parameter_138: name = "parameter_138" + original_name = "batch_norm2d_110.w_1" shape = [96] dtype = "float32" min_val = float("-0.0589327") @@ -1529,6 +1668,7 @@ class Program_weight_tensor_parameter_138: class Program_weight_tensor_parameter_139: name = "parameter_139" + original_name = "conv2d_114.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.0753795") @@ -1540,6 +1680,7 @@ class Program_weight_tensor_parameter_139: class Program_weight_tensor_parameter_140: name = "parameter_140" + original_name = "batch_norm2d_109.b_0" shape = [96] dtype = "float32" min_val = float("-0.457965") @@ -1551,6 +1692,7 @@ class Program_weight_tensor_parameter_140: class Program_weight_tensor_parameter_141: name = "parameter_141" + original_name = "batch_norm2d_109.w_0" shape = [96] dtype = "float32" min_val = float("0.505008") @@ -1562,6 +1704,7 @@ class Program_weight_tensor_parameter_141: class Program_weight_tensor_parameter_142: name = "parameter_142" + original_name = "batch_norm2d_109.w_2" shape = [96] dtype = "float32" min_val = float("0.00864928") @@ -1573,6 +1716,7 @@ class Program_weight_tensor_parameter_142: class Program_weight_tensor_parameter_143: name = "parameter_143" + original_name = "batch_norm2d_109.w_1" shape = [96] dtype = "float32" min_val = float("-0.235962") @@ -1584,6 +1728,7 @@ class Program_weight_tensor_parameter_143: class Program_weight_tensor_parameter_144: name = "parameter_144" + original_name = "conv2d_113.w_0" shape = [96, 96, 3, 3] dtype = "float32" min_val = float("-0.0934278") @@ -1595,6 +1740,7 @@ class Program_weight_tensor_parameter_144: class Program_weight_tensor_parameter_145: name = "parameter_145" + original_name = "batch_norm2d_108.b_0" shape = [96] dtype = "float32" min_val = float("-0.703686") @@ -1606,6 +1752,7 @@ class Program_weight_tensor_parameter_145: class Program_weight_tensor_parameter_146: name = "parameter_146" + original_name = "batch_norm2d_108.w_0" shape = [96] dtype = "float32" min_val = float("0.723215") @@ -1617,6 +1764,7 @@ class Program_weight_tensor_parameter_146: class Program_weight_tensor_parameter_147: name = "parameter_147" + original_name = "batch_norm2d_108.w_2" shape = [96] dtype = "float32" min_val = float("0.01304") @@ -1628,6 +1776,7 @@ class Program_weight_tensor_parameter_147: class Program_weight_tensor_parameter_148: name = "parameter_148" + original_name = "batch_norm2d_108.w_1" shape = [96] dtype = "float32" min_val = float("-0.214376") @@ -1639,6 +1788,7 @@ class Program_weight_tensor_parameter_148: class Program_weight_tensor_parameter_149: name = "parameter_149" + original_name = "conv2d_112.w_0" shape = [96, 96, 3, 3] dtype = "float32" min_val = float("-0.0919381") @@ -1650,6 +1800,7 @@ class Program_weight_tensor_parameter_149: class Program_weight_tensor_parameter_150: name = "parameter_150" + original_name = "batch_norm2d_107.b_0" shape = [96] dtype = "float32" min_val = float("-0.36415") @@ -1661,6 +1812,7 @@ class Program_weight_tensor_parameter_150: class Program_weight_tensor_parameter_151: name = "parameter_151" + original_name = "batch_norm2d_107.w_0" shape = [96] dtype = "float32" min_val = float("0.626999") @@ -1672,6 +1824,7 @@ class Program_weight_tensor_parameter_151: class Program_weight_tensor_parameter_152: name = "parameter_152" + original_name = "batch_norm2d_107.w_2" shape = [96] dtype = "float32" min_val = float("0.0031041") @@ -1683,6 +1836,7 @@ class Program_weight_tensor_parameter_152: class Program_weight_tensor_parameter_153: name = "parameter_153" + original_name = "batch_norm2d_107.w_1" shape = [96] dtype = "float32" min_val = float("-0.0660527") @@ -1694,6 +1848,7 @@ class Program_weight_tensor_parameter_153: class Program_weight_tensor_parameter_154: name = "parameter_154" + original_name = "conv2d_111.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.0710399") @@ -1705,6 +1860,7 @@ class Program_weight_tensor_parameter_154: class Program_weight_tensor_parameter_155: name = "parameter_155" + original_name = "batch_norm2d_106.b_0" shape = [96] dtype = "float32" min_val = float("-0.36415") @@ -1716,6 +1872,7 @@ class Program_weight_tensor_parameter_155: class Program_weight_tensor_parameter_156: name = "parameter_156" + original_name = "batch_norm2d_106.w_0" shape = [96] dtype = "float32" min_val = float("0.811164") @@ -1727,6 +1884,7 @@ class Program_weight_tensor_parameter_156: class Program_weight_tensor_parameter_157: name = "parameter_157" + original_name = "batch_norm2d_106.w_2" shape = [96] dtype = "float32" min_val = float("0.0111709") @@ -1738,6 +1896,7 @@ class Program_weight_tensor_parameter_157: class Program_weight_tensor_parameter_158: name = "parameter_158" + original_name = "batch_norm2d_106.w_1" shape = [96] dtype = "float32" min_val = float("-0.164873") @@ -1749,6 +1908,7 @@ class Program_weight_tensor_parameter_158: class Program_weight_tensor_parameter_159: name = "parameter_159" + original_name = "conv2d_110.w_0" shape = [96, 96, 3, 3] dtype = "float32" min_val = float("-0.0811913") @@ -1760,6 +1920,7 @@ class Program_weight_tensor_parameter_159: class Program_weight_tensor_parameter_160: name = "parameter_160" + original_name = "batch_norm2d_105.b_0" shape = [96] dtype = "float32" min_val = float("-0.486488") @@ -1771,6 +1932,7 @@ class Program_weight_tensor_parameter_160: class Program_weight_tensor_parameter_161: name = "parameter_161" + original_name = "batch_norm2d_105.w_0" shape = [96] dtype = "float32" min_val = float("0.777448") @@ -1782,6 +1944,7 @@ class Program_weight_tensor_parameter_161: class Program_weight_tensor_parameter_162: name = "parameter_162" + original_name = "batch_norm2d_105.w_2" shape = [96] dtype = "float32" min_val = float("0.00992224") @@ -1793,6 +1956,7 @@ class Program_weight_tensor_parameter_162: class Program_weight_tensor_parameter_163: name = "parameter_163" + original_name = "batch_norm2d_105.w_1" shape = [96] dtype = "float32" min_val = float("-0.154927") @@ -1804,6 +1968,7 @@ class Program_weight_tensor_parameter_163: class Program_weight_tensor_parameter_164: name = "parameter_164" + original_name = "conv2d_109.w_0" shape = [96, 96, 3, 3] dtype = "float32" min_val = float("-0.0993817") @@ -1815,6 +1980,7 @@ class Program_weight_tensor_parameter_164: class Program_weight_tensor_parameter_165: name = "parameter_165" + original_name = "batch_norm2d_104.b_0" shape = [96] dtype = "float32" min_val = float("-0.489705") @@ -1826,6 +1992,7 @@ class Program_weight_tensor_parameter_165: class Program_weight_tensor_parameter_166: name = "parameter_166" + original_name = "batch_norm2d_104.w_0" shape = [96] dtype = "float32" min_val = float("0.722939") @@ -1837,6 +2004,7 @@ class Program_weight_tensor_parameter_166: class Program_weight_tensor_parameter_167: name = "parameter_167" + original_name = "batch_norm2d_104.w_2" shape = [96] dtype = "float32" min_val = float("0.00775018") @@ -1848,6 +2016,7 @@ class Program_weight_tensor_parameter_167: class Program_weight_tensor_parameter_168: name = "parameter_168" + original_name = "batch_norm2d_104.w_1" shape = [96] dtype = "float32" min_val = float("-0.0570203") @@ -1859,6 +2028,7 @@ class Program_weight_tensor_parameter_168: class Program_weight_tensor_parameter_169: name = "parameter_169" + original_name = "conv2d_108.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.103947") @@ -1870,6 +2040,7 @@ class Program_weight_tensor_parameter_169: class Program_weight_tensor_parameter_170: name = "parameter_170" + original_name = "batch_norm2d_103.b_0" shape = [96] dtype = "float32" min_val = float("-0.489705") @@ -1881,6 +2052,7 @@ class Program_weight_tensor_parameter_170: class Program_weight_tensor_parameter_171: name = "parameter_171" + original_name = "batch_norm2d_103.w_0" shape = [96] dtype = "float32" min_val = float("0.766535") @@ -1892,6 +2064,7 @@ class Program_weight_tensor_parameter_171: class Program_weight_tensor_parameter_172: name = "parameter_172" + original_name = "batch_norm2d_103.w_2" shape = [96] dtype = "float32" min_val = float("0.0171361") @@ -1903,6 +2076,7 @@ class Program_weight_tensor_parameter_172: class Program_weight_tensor_parameter_173: name = "parameter_173" + original_name = "batch_norm2d_103.w_1" shape = [96] dtype = "float32" min_val = float("-0.199015") @@ -1914,6 +2088,7 @@ class Program_weight_tensor_parameter_173: class Program_weight_tensor_parameter_174: name = "parameter_174" + original_name = "conv2d_107.w_0" shape = [96, 96, 3, 3] dtype = "float32" min_val = float("-0.099267") @@ -1925,6 +2100,7 @@ class Program_weight_tensor_parameter_174: class Program_weight_tensor_parameter_175: name = "parameter_175" + original_name = "batch_norm2d_102.b_0" shape = [96] dtype = "float32" min_val = float("-0.564609") @@ -1936,6 +2112,7 @@ class Program_weight_tensor_parameter_175: class Program_weight_tensor_parameter_176: name = "parameter_176" + original_name = "batch_norm2d_102.w_0" shape = [96] dtype = "float32" min_val = float("0.764463") @@ -1947,6 +2124,7 @@ class Program_weight_tensor_parameter_176: class Program_weight_tensor_parameter_177: name = "parameter_177" + original_name = "batch_norm2d_102.w_2" shape = [96] dtype = "float32" min_val = float("0.0148204") @@ -1958,6 +2136,7 @@ class Program_weight_tensor_parameter_177: class Program_weight_tensor_parameter_178: name = "parameter_178" + original_name = "batch_norm2d_102.w_1" shape = [96] dtype = "float32" min_val = float("-0.172672") @@ -1969,6 +2148,7 @@ class Program_weight_tensor_parameter_178: class Program_weight_tensor_parameter_179: name = "parameter_179" + original_name = "conv2d_106.w_0" shape = [96, 96, 3, 3] dtype = "float32" min_val = float("-0.142387") @@ -1980,6 +2160,7 @@ class Program_weight_tensor_parameter_179: class Program_weight_tensor_parameter_180: name = "parameter_180" + original_name = "batch_norm2d_101.b_0" shape = [96] dtype = "float32" min_val = float("-0.625413") @@ -1991,6 +2172,7 @@ class Program_weight_tensor_parameter_180: class Program_weight_tensor_parameter_181: name = "parameter_181" + original_name = "batch_norm2d_101.w_0" shape = [96] dtype = "float32" min_val = float("0.647481") @@ -2002,6 +2184,7 @@ class Program_weight_tensor_parameter_181: class Program_weight_tensor_parameter_182: name = "parameter_182" + original_name = "batch_norm2d_101.w_2" shape = [96] dtype = "float32" min_val = float("0.0117598") @@ -2013,6 +2196,7 @@ class Program_weight_tensor_parameter_182: class Program_weight_tensor_parameter_183: name = "parameter_183" + original_name = "batch_norm2d_101.w_1" shape = [96] dtype = "float32" min_val = float("-0.116141") @@ -2024,6 +2208,7 @@ class Program_weight_tensor_parameter_183: class Program_weight_tensor_parameter_184: name = "parameter_184" + original_name = "conv2d_105.w_0" shape = [96, 448, 1, 1] dtype = "float32" min_val = float("-0.149031") @@ -2035,6 +2220,7 @@ class Program_weight_tensor_parameter_184: class Program_weight_tensor_parameter_185: name = "parameter_185" + original_name = "batch_norm2d_100.b_0" shape = [96] dtype = "float32" min_val = float("-0.0986348") @@ -2046,6 +2232,7 @@ class Program_weight_tensor_parameter_185: class Program_weight_tensor_parameter_186: name = "parameter_186" + original_name = "batch_norm2d_100.w_0" shape = [96] dtype = "float32" min_val = float("0.703927") @@ -2057,6 +2244,7 @@ class Program_weight_tensor_parameter_186: class Program_weight_tensor_parameter_187: name = "parameter_187" + original_name = "batch_norm2d_100.w_2" shape = [96] dtype = "float32" min_val = float("0.00537723") @@ -2068,6 +2256,7 @@ class Program_weight_tensor_parameter_187: class Program_weight_tensor_parameter_188: name = "parameter_188" + original_name = "batch_norm2d_100.w_1" shape = [96] dtype = "float32" min_val = float("-0.0902684") @@ -2079,6 +2268,7 @@ class Program_weight_tensor_parameter_188: class Program_weight_tensor_parameter_189: name = "parameter_189" + original_name = "conv2d_104.w_0" shape = [96, 448, 1, 1] dtype = "float32" min_val = float("-0.0952113") @@ -2090,6 +2280,7 @@ class Program_weight_tensor_parameter_189: class Program_weight_tensor_parameter_190: name = "parameter_190" + original_name = "batch_norm2d_99.b_0" shape = [192] dtype = "float32" min_val = float("-0.295368") @@ -2101,6 +2292,7 @@ class Program_weight_tensor_parameter_190: class Program_weight_tensor_parameter_191: name = "parameter_191" + original_name = "batch_norm2d_99.w_0" shape = [192] dtype = "float32" min_val = float("0.670697") @@ -2112,6 +2304,7 @@ class Program_weight_tensor_parameter_191: class Program_weight_tensor_parameter_192: name = "parameter_192" + original_name = "batch_norm2d_99.w_2" shape = [192] dtype = "float32" min_val = float("0.00790397") @@ -2123,6 +2316,7 @@ class Program_weight_tensor_parameter_192: class Program_weight_tensor_parameter_193: name = "parameter_193" + original_name = "batch_norm2d_99.w_1" shape = [192] dtype = "float32" min_val = float("-0.148651") @@ -2134,6 +2328,7 @@ class Program_weight_tensor_parameter_193: class Program_weight_tensor_parameter_194: name = "parameter_194" + original_name = "conv2d_103.w_0" shape = [192, 384, 1, 1] dtype = "float32" min_val = float("-0.0959148") @@ -2145,6 +2340,7 @@ class Program_weight_tensor_parameter_194: class Program_weight_tensor_parameter_195: name = "parameter_195" + original_name = "batch_norm2d_98.b_0" shape = [384] dtype = "float32" min_val = float("-0.201782") @@ -2156,6 +2352,7 @@ class Program_weight_tensor_parameter_195: class Program_weight_tensor_parameter_196: name = "parameter_196" + original_name = "batch_norm2d_98.w_0" shape = [384] dtype = "float32" min_val = float("0.87318") @@ -2167,6 +2364,7 @@ class Program_weight_tensor_parameter_196: class Program_weight_tensor_parameter_197: name = "parameter_197" + original_name = "batch_norm2d_98.w_2" shape = [384] dtype = "float32" min_val = float("0.00737718") @@ -2178,6 +2376,7 @@ class Program_weight_tensor_parameter_197: class Program_weight_tensor_parameter_198: name = "parameter_198" + original_name = "batch_norm2d_98.w_1" shape = [384] dtype = "float32" min_val = float("-0.339572") @@ -2189,6 +2388,7 @@ class Program_weight_tensor_parameter_198: class Program_weight_tensor_parameter_199: name = "parameter_199" + original_name = "conv2d_102.w_0" shape = [384, 384, 1, 1] dtype = "float32" min_val = float("-0.104641") @@ -2200,6 +2400,7 @@ class Program_weight_tensor_parameter_199: class Program_weight_tensor_parameter_200: name = "parameter_200" + original_name = "batch_norm2d_97.b_0" shape = [192] dtype = "float32" min_val = float("-0.176949") @@ -2211,6 +2412,7 @@ class Program_weight_tensor_parameter_200: class Program_weight_tensor_parameter_201: name = "parameter_201" + original_name = "batch_norm2d_97.w_0" shape = [192] dtype = "float32" min_val = float("0.884903") @@ -2222,6 +2424,7 @@ class Program_weight_tensor_parameter_201: class Program_weight_tensor_parameter_202: name = "parameter_202" + original_name = "batch_norm2d_97.w_2" shape = [192] dtype = "float32" min_val = float("0.00353105") @@ -2233,6 +2436,7 @@ class Program_weight_tensor_parameter_202: class Program_weight_tensor_parameter_203: name = "parameter_203" + original_name = "batch_norm2d_97.w_1" shape = [192] dtype = "float32" min_val = float("-0.0792437") @@ -2244,6 +2448,7 @@ class Program_weight_tensor_parameter_203: class Program_weight_tensor_parameter_204: name = "parameter_204" + original_name = "conv2d_101.w_0" shape = [192, 192, 1, 1] dtype = "float32" min_val = float("-0.0569077") @@ -2255,6 +2460,7 @@ class Program_weight_tensor_parameter_204: class Program_weight_tensor_parameter_205: name = "parameter_205" + original_name = "batch_norm2d_96.b_0" shape = [192] dtype = "float32" min_val = float("-0.176949") @@ -2266,6 +2472,7 @@ class Program_weight_tensor_parameter_205: class Program_weight_tensor_parameter_206: name = "parameter_206" + original_name = "batch_norm2d_96.w_0" shape = [192] dtype = "float32" min_val = float("0.945936") @@ -2277,6 +2484,7 @@ class Program_weight_tensor_parameter_206: class Program_weight_tensor_parameter_207: name = "parameter_207" + original_name = "batch_norm2d_96.w_2" shape = [192] dtype = "float32" min_val = float("0.0150858") @@ -2288,6 +2496,7 @@ class Program_weight_tensor_parameter_207: class Program_weight_tensor_parameter_208: name = "parameter_208" + original_name = "batch_norm2d_96.w_1" shape = [192] dtype = "float32" min_val = float("-0.177971") @@ -2299,6 +2508,7 @@ class Program_weight_tensor_parameter_208: class Program_weight_tensor_parameter_209: name = "parameter_209" + original_name = "conv2d_100.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.0444125") @@ -2310,6 +2520,7 @@ class Program_weight_tensor_parameter_209: class Program_weight_tensor_parameter_210: name = "parameter_210" + original_name = "batch_norm2d_95.b_0" shape = [192] dtype = "float32" min_val = float("-0.217095") @@ -2321,6 +2532,7 @@ class Program_weight_tensor_parameter_210: class Program_weight_tensor_parameter_211: name = "parameter_211" + original_name = "batch_norm2d_95.w_0" shape = [192] dtype = "float32" min_val = float("0.939032") @@ -2332,6 +2544,7 @@ class Program_weight_tensor_parameter_211: class Program_weight_tensor_parameter_212: name = "parameter_212" + original_name = "batch_norm2d_95.w_2" shape = [192] dtype = "float32" min_val = float("0.0360784") @@ -2343,6 +2556,7 @@ class Program_weight_tensor_parameter_212: class Program_weight_tensor_parameter_213: name = "parameter_213" + original_name = "batch_norm2d_95.w_1" shape = [192] dtype = "float32" min_val = float("-0.264891") @@ -2354,6 +2568,7 @@ class Program_weight_tensor_parameter_213: class Program_weight_tensor_parameter_214: name = "parameter_214" + original_name = "conv2d_99.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.0622024") @@ -2365,6 +2580,7 @@ class Program_weight_tensor_parameter_214: class Program_weight_tensor_parameter_215: name = "parameter_215" + original_name = "batch_norm2d_94.b_0" shape = [192] dtype = "float32" min_val = float("-0.196618") @@ -2376,6 +2592,7 @@ class Program_weight_tensor_parameter_215: class Program_weight_tensor_parameter_216: name = "parameter_216" + original_name = "batch_norm2d_94.w_0" shape = [192] dtype = "float32" min_val = float("0.94411") @@ -2387,6 +2604,7 @@ class Program_weight_tensor_parameter_216: class Program_weight_tensor_parameter_217: name = "parameter_217" + original_name = "batch_norm2d_94.w_2" shape = [192] dtype = "float32" min_val = float("0.00227428") @@ -2398,6 +2616,7 @@ class Program_weight_tensor_parameter_217: class Program_weight_tensor_parameter_218: name = "parameter_218" + original_name = "batch_norm2d_94.w_1" shape = [192] dtype = "float32" min_val = float("-0.0961104") @@ -2409,6 +2628,7 @@ class Program_weight_tensor_parameter_218: class Program_weight_tensor_parameter_219: name = "parameter_219" + original_name = "conv2d_98.w_0" shape = [192, 192, 1, 1] dtype = "float32" min_val = float("-0.0313925") @@ -2420,6 +2640,7 @@ class Program_weight_tensor_parameter_219: class Program_weight_tensor_parameter_220: name = "parameter_220" + original_name = "batch_norm2d_93.b_0" shape = [192] dtype = "float32" min_val = float("-0.196618") @@ -2431,6 +2652,7 @@ class Program_weight_tensor_parameter_220: class Program_weight_tensor_parameter_221: name = "parameter_221" + original_name = "batch_norm2d_93.w_0" shape = [192] dtype = "float32" min_val = float("0.953711") @@ -2442,6 +2664,7 @@ class Program_weight_tensor_parameter_221: class Program_weight_tensor_parameter_222: name = "parameter_222" + original_name = "batch_norm2d_93.w_2" shape = [192] dtype = "float32" min_val = float("0.010102") @@ -2453,6 +2676,7 @@ class Program_weight_tensor_parameter_222: class Program_weight_tensor_parameter_223: name = "parameter_223" + original_name = "batch_norm2d_93.w_1" shape = [192] dtype = "float32" min_val = float("-0.189335") @@ -2464,6 +2688,7 @@ class Program_weight_tensor_parameter_223: class Program_weight_tensor_parameter_224: name = "parameter_224" + original_name = "conv2d_97.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.0484375") @@ -2475,6 +2700,7 @@ class Program_weight_tensor_parameter_224: class Program_weight_tensor_parameter_225: name = "parameter_225" + original_name = "batch_norm2d_92.b_0" shape = [192] dtype = "float32" min_val = float("-0.232846") @@ -2486,6 +2712,7 @@ class Program_weight_tensor_parameter_225: class Program_weight_tensor_parameter_226: name = "parameter_226" + original_name = "batch_norm2d_92.w_0" shape = [192] dtype = "float32" min_val = float("0.946521") @@ -2497,6 +2724,7 @@ class Program_weight_tensor_parameter_226: class Program_weight_tensor_parameter_227: name = "parameter_227" + original_name = "batch_norm2d_92.w_2" shape = [192] dtype = "float32" min_val = float("0.0359278") @@ -2508,6 +2736,7 @@ class Program_weight_tensor_parameter_227: class Program_weight_tensor_parameter_228: name = "parameter_228" + original_name = "batch_norm2d_92.w_1" shape = [192] dtype = "float32" min_val = float("-0.353487") @@ -2519,6 +2748,7 @@ class Program_weight_tensor_parameter_228: class Program_weight_tensor_parameter_229: name = "parameter_229" + original_name = "conv2d_96.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.0611518") @@ -2530,6 +2760,7 @@ class Program_weight_tensor_parameter_229: class Program_weight_tensor_parameter_230: name = "parameter_230" + original_name = "batch_norm2d_91.b_0" shape = [192] dtype = "float32" min_val = float("-0.154886") @@ -2541,6 +2772,7 @@ class Program_weight_tensor_parameter_230: class Program_weight_tensor_parameter_231: name = "parameter_231" + original_name = "batch_norm2d_91.w_0" shape = [192] dtype = "float32" min_val = float("0.932342") @@ -2552,6 +2784,7 @@ class Program_weight_tensor_parameter_231: class Program_weight_tensor_parameter_232: name = "parameter_232" + original_name = "batch_norm2d_91.w_2" shape = [192] dtype = "float32" min_val = float("0.00201715") @@ -2563,6 +2796,7 @@ class Program_weight_tensor_parameter_232: class Program_weight_tensor_parameter_233: name = "parameter_233" + original_name = "batch_norm2d_91.w_1" shape = [192] dtype = "float32" min_val = float("-0.0829276") @@ -2574,6 +2808,7 @@ class Program_weight_tensor_parameter_233: class Program_weight_tensor_parameter_234: name = "parameter_234" + original_name = "conv2d_95.w_0" shape = [192, 192, 1, 1] dtype = "float32" min_val = float("-0.0348639") @@ -2585,6 +2820,7 @@ class Program_weight_tensor_parameter_234: class Program_weight_tensor_parameter_235: name = "parameter_235" + original_name = "batch_norm2d_90.b_0" shape = [192] dtype = "float32" min_val = float("-0.154886") @@ -2596,6 +2832,7 @@ class Program_weight_tensor_parameter_235: class Program_weight_tensor_parameter_236: name = "parameter_236" + original_name = "batch_norm2d_90.w_0" shape = [192] dtype = "float32" min_val = float("0.936173") @@ -2607,6 +2844,7 @@ class Program_weight_tensor_parameter_236: class Program_weight_tensor_parameter_237: name = "parameter_237" + original_name = "batch_norm2d_90.w_2" shape = [192] dtype = "float32" min_val = float("0.00897904") @@ -2618,6 +2856,7 @@ class Program_weight_tensor_parameter_237: class Program_weight_tensor_parameter_238: name = "parameter_238" + original_name = "batch_norm2d_90.w_1" shape = [192] dtype = "float32" min_val = float("-0.280815") @@ -2629,6 +2868,7 @@ class Program_weight_tensor_parameter_238: class Program_weight_tensor_parameter_239: name = "parameter_239" + original_name = "conv2d_94.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.0372381") @@ -2640,6 +2880,7 @@ class Program_weight_tensor_parameter_239: class Program_weight_tensor_parameter_240: name = "parameter_240" + original_name = "batch_norm2d_89.b_0" shape = [192] dtype = "float32" min_val = float("-0.289029") @@ -2651,6 +2892,7 @@ class Program_weight_tensor_parameter_240: class Program_weight_tensor_parameter_241: name = "parameter_241" + original_name = "batch_norm2d_89.w_0" shape = [192] dtype = "float32" min_val = float("0.943873") @@ -2662,6 +2904,7 @@ class Program_weight_tensor_parameter_241: class Program_weight_tensor_parameter_242: name = "parameter_242" + original_name = "batch_norm2d_89.w_2" shape = [192] dtype = "float32" min_val = float("0.0146575") @@ -2673,6 +2916,7 @@ class Program_weight_tensor_parameter_242: class Program_weight_tensor_parameter_243: name = "parameter_243" + original_name = "batch_norm2d_89.w_1" shape = [192] dtype = "float32" min_val = float("-0.381573") @@ -2684,6 +2928,7 @@ class Program_weight_tensor_parameter_243: class Program_weight_tensor_parameter_244: name = "parameter_244" + original_name = "conv2d_93.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.0566325") @@ -2695,6 +2940,7 @@ class Program_weight_tensor_parameter_244: class Program_weight_tensor_parameter_245: name = "parameter_245" + original_name = "batch_norm2d_88.b_0" shape = [192] dtype = "float32" min_val = float("-0.257034") @@ -2706,6 +2952,7 @@ class Program_weight_tensor_parameter_245: class Program_weight_tensor_parameter_246: name = "parameter_246" + original_name = "batch_norm2d_88.w_0" shape = [192] dtype = "float32" min_val = float("0.916942") @@ -2717,6 +2964,7 @@ class Program_weight_tensor_parameter_246: class Program_weight_tensor_parameter_247: name = "parameter_247" + original_name = "batch_norm2d_88.w_2" shape = [192] dtype = "float32" min_val = float("0.00558811") @@ -2728,6 +2976,7 @@ class Program_weight_tensor_parameter_247: class Program_weight_tensor_parameter_248: name = "parameter_248" + original_name = "batch_norm2d_88.w_1" shape = [192] dtype = "float32" min_val = float("-0.120986") @@ -2739,6 +2988,7 @@ class Program_weight_tensor_parameter_248: class Program_weight_tensor_parameter_249: name = "parameter_249" + original_name = "conv2d_92.w_0" shape = [192, 896, 1, 1] dtype = "float32" min_val = float("-0.0812543") @@ -2750,6 +3000,7 @@ class Program_weight_tensor_parameter_249: class Program_weight_tensor_parameter_250: name = "parameter_250" + original_name = "batch_norm2d_87.b_0" shape = [192] dtype = "float32" min_val = float("-0.176609") @@ -2761,6 +3012,7 @@ class Program_weight_tensor_parameter_250: class Program_weight_tensor_parameter_251: name = "parameter_251" + original_name = "batch_norm2d_87.w_0" shape = [192] dtype = "float32" min_val = float("0.951166") @@ -2772,6 +3024,7 @@ class Program_weight_tensor_parameter_251: class Program_weight_tensor_parameter_252: name = "parameter_252" + original_name = "batch_norm2d_87.w_2" shape = [192] dtype = "float32" min_val = float("0.0068407") @@ -2783,6 +3036,7 @@ class Program_weight_tensor_parameter_252: class Program_weight_tensor_parameter_253: name = "parameter_253" + original_name = "batch_norm2d_87.w_1" shape = [192] dtype = "float32" min_val = float("-0.076614") @@ -2794,6 +3048,7 @@ class Program_weight_tensor_parameter_253: class Program_weight_tensor_parameter_254: name = "parameter_254" + original_name = "conv2d_91.w_0" shape = [192, 896, 1, 1] dtype = "float32" min_val = float("-0.0552042") @@ -2805,6 +3060,7 @@ class Program_weight_tensor_parameter_254: class Program_weight_tensor_parameter_255: name = "parameter_255" + original_name = "batch_norm2d_86.b_0" shape = [384] dtype = "float32" min_val = float("-0.249775") @@ -2816,6 +3072,7 @@ class Program_weight_tensor_parameter_255: class Program_weight_tensor_parameter_256: name = "parameter_256" + original_name = "batch_norm2d_86.w_0" shape = [384] dtype = "float32" min_val = float("0.814907") @@ -2827,6 +3084,7 @@ class Program_weight_tensor_parameter_256: class Program_weight_tensor_parameter_257: name = "parameter_257" + original_name = "batch_norm2d_86.w_2" shape = [384] dtype = "float32" min_val = float("0.00972713") @@ -2838,6 +3096,7 @@ class Program_weight_tensor_parameter_257: class Program_weight_tensor_parameter_258: name = "parameter_258" + original_name = "batch_norm2d_86.w_1" shape = [384] dtype = "float32" min_val = float("-0.146609") @@ -2849,6 +3108,7 @@ class Program_weight_tensor_parameter_258: class Program_weight_tensor_parameter_259: name = "parameter_259" + original_name = "conv2d_90.w_0" shape = [384, 768, 1, 1] dtype = "float32" min_val = float("-0.0364879") @@ -2860,6 +3120,7 @@ class Program_weight_tensor_parameter_259: class Program_weight_tensor_parameter_260: name = "parameter_260" + original_name = "batch_norm2d_85.b_0" shape = [768] dtype = "float32" min_val = float("-0.104277") @@ -2871,6 +3132,7 @@ class Program_weight_tensor_parameter_260: class Program_weight_tensor_parameter_261: name = "parameter_261" + original_name = "batch_norm2d_85.w_0" shape = [768] dtype = "float32" min_val = float("0.9523") @@ -2882,6 +3144,7 @@ class Program_weight_tensor_parameter_261: class Program_weight_tensor_parameter_262: name = "parameter_262" + original_name = "batch_norm2d_85.w_2" shape = [768] dtype = "float32" min_val = float("0.00431744") @@ -2893,6 +3156,7 @@ class Program_weight_tensor_parameter_262: class Program_weight_tensor_parameter_263: name = "parameter_263" + original_name = "batch_norm2d_85.w_1" shape = [768] dtype = "float32" min_val = float("-0.103952") @@ -2904,6 +3168,7 @@ class Program_weight_tensor_parameter_263: class Program_weight_tensor_parameter_264: name = "parameter_264" + original_name = "conv2d_89.w_0" shape = [768, 768, 1, 1] dtype = "float32" min_val = float("-0.0581811") @@ -2915,6 +3180,7 @@ class Program_weight_tensor_parameter_264: class Program_weight_tensor_parameter_265: name = "parameter_265" + original_name = "batch_norm2d_84.b_0" shape = [384] dtype = "float32" min_val = float("-0.158167") @@ -2926,6 +3192,7 @@ class Program_weight_tensor_parameter_265: class Program_weight_tensor_parameter_266: name = "parameter_266" + original_name = "batch_norm2d_84.w_0" shape = [384] dtype = "float32" min_val = float("0.888577") @@ -2937,6 +3204,7 @@ class Program_weight_tensor_parameter_266: class Program_weight_tensor_parameter_267: name = "parameter_267" + original_name = "batch_norm2d_84.w_2" shape = [384] dtype = "float32" min_val = float("0.005659") @@ -2948,6 +3216,7 @@ class Program_weight_tensor_parameter_267: class Program_weight_tensor_parameter_268: name = "parameter_268" + original_name = "batch_norm2d_84.w_1" shape = [384] dtype = "float32" min_val = float("-0.0685013") @@ -2959,6 +3228,7 @@ class Program_weight_tensor_parameter_268: class Program_weight_tensor_parameter_269: name = "parameter_269" + original_name = "conv2d_88.w_0" shape = [384, 384, 1, 1] dtype = "float32" min_val = float("-0.0396793") @@ -2970,6 +3240,7 @@ class Program_weight_tensor_parameter_269: class Program_weight_tensor_parameter_270: name = "parameter_270" + original_name = "batch_norm2d_83.b_0" shape = [384] dtype = "float32" min_val = float("-0.158167") @@ -2981,6 +3252,7 @@ class Program_weight_tensor_parameter_270: class Program_weight_tensor_parameter_271: name = "parameter_271" + original_name = "batch_norm2d_83.w_0" shape = [384] dtype = "float32" min_val = float("0.879914") @@ -2992,6 +3264,7 @@ class Program_weight_tensor_parameter_271: class Program_weight_tensor_parameter_272: name = "parameter_272" + original_name = "batch_norm2d_83.w_2" shape = [384] dtype = "float32" min_val = float("0.0237413") @@ -3003,6 +3276,7 @@ class Program_weight_tensor_parameter_272: class Program_weight_tensor_parameter_273: name = "parameter_273" + original_name = "batch_norm2d_83.w_1" shape = [384] dtype = "float32" min_val = float("-0.277276") @@ -3014,6 +3288,7 @@ class Program_weight_tensor_parameter_273: class Program_weight_tensor_parameter_274: name = "parameter_274" + original_name = "conv2d_87.w_0" shape = [384, 384, 3, 3] dtype = "float32" min_val = float("-0.0424879") @@ -3025,6 +3300,7 @@ class Program_weight_tensor_parameter_274: class Program_weight_tensor_parameter_275: name = "parameter_275" + original_name = "batch_norm2d_82.b_0" shape = [384] dtype = "float32" min_val = float("-0.0801146") @@ -3036,6 +3312,7 @@ class Program_weight_tensor_parameter_275: class Program_weight_tensor_parameter_276: name = "parameter_276" + original_name = "batch_norm2d_82.w_0" shape = [384] dtype = "float32" min_val = float("0.920205") @@ -3047,6 +3324,7 @@ class Program_weight_tensor_parameter_276: class Program_weight_tensor_parameter_277: name = "parameter_277" + original_name = "batch_norm2d_82.w_2" shape = [384] dtype = "float32" min_val = float("0.0222636") @@ -3058,6 +3336,7 @@ class Program_weight_tensor_parameter_277: class Program_weight_tensor_parameter_278: name = "parameter_278" + original_name = "batch_norm2d_82.w_1" shape = [384] dtype = "float32" min_val = float("-0.233438") @@ -3069,6 +3348,7 @@ class Program_weight_tensor_parameter_278: class Program_weight_tensor_parameter_279: name = "parameter_279" + original_name = "conv2d_86.w_0" shape = [384, 384, 3, 3] dtype = "float32" min_val = float("-0.0274571") @@ -3080,6 +3360,7 @@ class Program_weight_tensor_parameter_279: class Program_weight_tensor_parameter_280: name = "parameter_280" + original_name = "batch_norm2d_81.b_0" shape = [384] dtype = "float32" min_val = float("-0.0739507") @@ -3091,6 +3372,7 @@ class Program_weight_tensor_parameter_280: class Program_weight_tensor_parameter_281: name = "parameter_281" + original_name = "batch_norm2d_81.w_0" shape = [384] dtype = "float32" min_val = float("0.946312") @@ -3102,6 +3384,7 @@ class Program_weight_tensor_parameter_281: class Program_weight_tensor_parameter_282: name = "parameter_282" + original_name = "batch_norm2d_81.w_2" shape = [384] dtype = "float32" min_val = float("0.0631221") @@ -3113,6 +3396,7 @@ class Program_weight_tensor_parameter_282: class Program_weight_tensor_parameter_283: name = "parameter_283" + original_name = "batch_norm2d_81.w_1" shape = [384] dtype = "float32" min_val = float("-1.66842") @@ -3124,6 +3408,7 @@ class Program_weight_tensor_parameter_283: class Program_weight_tensor_parameter_284: name = "parameter_284" + original_name = "conv2d_85.w_0" shape = [384, 1536, 1, 1] dtype = "float32" min_val = float("-0.0467314") @@ -3135,6 +3420,7 @@ class Program_weight_tensor_parameter_284: class Program_weight_tensor_parameter_285: name = "parameter_285" + original_name = "batch_norm2d_80.b_0" shape = [384] dtype = "float32" min_val = float("-0.0183803") @@ -3146,6 +3432,7 @@ class Program_weight_tensor_parameter_285: class Program_weight_tensor_parameter_286: name = "parameter_286" + original_name = "batch_norm2d_80.w_0" shape = [384] dtype = "float32" min_val = float("0.969538") @@ -3157,6 +3444,7 @@ class Program_weight_tensor_parameter_286: class Program_weight_tensor_parameter_287: name = "parameter_287" + original_name = "batch_norm2d_80.w_2" shape = [384] dtype = "float32" min_val = float("0.00292151") @@ -3168,6 +3456,7 @@ class Program_weight_tensor_parameter_287: class Program_weight_tensor_parameter_288: name = "parameter_288" + original_name = "batch_norm2d_80.w_1" shape = [384] dtype = "float32" min_val = float("-0.0974501") @@ -3179,6 +3468,7 @@ class Program_weight_tensor_parameter_288: class Program_weight_tensor_parameter_289: name = "parameter_289" + original_name = "conv2d_84.w_0" shape = [384, 384, 1, 1] dtype = "float32" min_val = float("-0.0333324") @@ -3190,6 +3480,7 @@ class Program_weight_tensor_parameter_289: class Program_weight_tensor_parameter_290: name = "parameter_290" + original_name = "batch_norm2d_79.b_0" shape = [384] dtype = "float32" min_val = float("-0.0183803") @@ -3201,6 +3492,7 @@ class Program_weight_tensor_parameter_290: class Program_weight_tensor_parameter_291: name = "parameter_291" + original_name = "batch_norm2d_79.w_0" shape = [384] dtype = "float32" min_val = float("0.972046") @@ -3212,6 +3504,7 @@ class Program_weight_tensor_parameter_291: class Program_weight_tensor_parameter_292: name = "parameter_292" + original_name = "batch_norm2d_79.w_2" shape = [384] dtype = "float32" min_val = float("0.0170031") @@ -3223,6 +3516,7 @@ class Program_weight_tensor_parameter_292: class Program_weight_tensor_parameter_293: name = "parameter_293" + original_name = "batch_norm2d_79.w_1" shape = [384] dtype = "float32" min_val = float("-0.324472") @@ -3234,6 +3528,7 @@ class Program_weight_tensor_parameter_293: class Program_weight_tensor_parameter_294: name = "parameter_294" + original_name = "conv2d_83.w_0" shape = [384, 384, 3, 3] dtype = "float32" min_val = float("-0.0285762") @@ -3245,6 +3540,7 @@ class Program_weight_tensor_parameter_294: class Program_weight_tensor_parameter_295: name = "parameter_295" + original_name = "batch_norm2d_78.b_0" shape = [384] dtype = "float32" min_val = float("-0.0498104") @@ -3256,6 +3552,7 @@ class Program_weight_tensor_parameter_295: class Program_weight_tensor_parameter_296: name = "parameter_296" + original_name = "batch_norm2d_78.w_0" shape = [384] dtype = "float32" min_val = float("0.953878") @@ -3267,6 +3564,7 @@ class Program_weight_tensor_parameter_296: class Program_weight_tensor_parameter_297: name = "parameter_297" + original_name = "batch_norm2d_78.w_2" shape = [384] dtype = "float32" min_val = float("0.0734313") @@ -3278,6 +3576,7 @@ class Program_weight_tensor_parameter_297: class Program_weight_tensor_parameter_298: name = "parameter_298" + original_name = "batch_norm2d_78.w_1" shape = [384] dtype = "float32" min_val = float("-1.24929") @@ -3289,6 +3588,7 @@ class Program_weight_tensor_parameter_298: class Program_weight_tensor_parameter_299: name = "parameter_299" + original_name = "conv2d_82.w_0" shape = [384, 384, 3, 3] dtype = "float32" min_val = float("-0.0242525") @@ -3300,6 +3600,7 @@ class Program_weight_tensor_parameter_299: class Program_weight_tensor_parameter_300: name = "parameter_300" + original_name = "batch_norm2d_77.b_0" shape = [384] dtype = "float32" min_val = float("-0.0360838") @@ -3311,6 +3612,7 @@ class Program_weight_tensor_parameter_300: class Program_weight_tensor_parameter_301: name = "parameter_301" + original_name = "batch_norm2d_77.w_0" shape = [384] dtype = "float32" min_val = float("0.984179") @@ -3322,6 +3624,7 @@ class Program_weight_tensor_parameter_301: class Program_weight_tensor_parameter_302: name = "parameter_302" + original_name = "batch_norm2d_77.w_2" shape = [384] dtype = "float32" min_val = float("0.00218605") @@ -3333,6 +3636,7 @@ class Program_weight_tensor_parameter_302: class Program_weight_tensor_parameter_303: name = "parameter_303" + original_name = "batch_norm2d_77.w_1" shape = [384] dtype = "float32" min_val = float("-0.0778225") @@ -3344,6 +3648,7 @@ class Program_weight_tensor_parameter_303: class Program_weight_tensor_parameter_304: name = "parameter_304" + original_name = "conv2d_81.w_0" shape = [384, 384, 1, 1] dtype = "float32" min_val = float("-0.020915") @@ -3355,6 +3660,7 @@ class Program_weight_tensor_parameter_304: class Program_weight_tensor_parameter_305: name = "parameter_305" + original_name = "batch_norm2d_76.b_0" shape = [384] dtype = "float32" min_val = float("-0.0360838") @@ -3366,6 +3672,7 @@ class Program_weight_tensor_parameter_305: class Program_weight_tensor_parameter_306: name = "parameter_306" + original_name = "batch_norm2d_76.w_0" shape = [384] dtype = "float32" min_val = float("0.982136") @@ -3377,6 +3684,7 @@ class Program_weight_tensor_parameter_306: class Program_weight_tensor_parameter_307: name = "parameter_307" + original_name = "batch_norm2d_76.w_2" shape = [384] dtype = "float32" min_val = float("0.009878") @@ -3388,6 +3696,7 @@ class Program_weight_tensor_parameter_307: class Program_weight_tensor_parameter_308: name = "parameter_308" + original_name = "batch_norm2d_76.w_1" shape = [384] dtype = "float32" min_val = float("-0.234125") @@ -3399,6 +3708,7 @@ class Program_weight_tensor_parameter_308: class Program_weight_tensor_parameter_309: name = "parameter_309" + original_name = "conv2d_80.w_0" shape = [384, 384, 3, 3] dtype = "float32" min_val = float("-0.0111228") @@ -3410,6 +3720,7 @@ class Program_weight_tensor_parameter_309: class Program_weight_tensor_parameter_310: name = "parameter_310" + original_name = "batch_norm2d_75.b_0" shape = [384] dtype = "float32" min_val = float("-0.0529908") @@ -3421,6 +3732,7 @@ class Program_weight_tensor_parameter_310: class Program_weight_tensor_parameter_311: name = "parameter_311" + original_name = "batch_norm2d_75.w_0" shape = [384] dtype = "float32" min_val = float("0.976061") @@ -3432,6 +3744,7 @@ class Program_weight_tensor_parameter_311: class Program_weight_tensor_parameter_312: name = "parameter_312" + original_name = "batch_norm2d_75.w_2" shape = [384] dtype = "float32" min_val = float("0.0131757") @@ -3443,6 +3756,7 @@ class Program_weight_tensor_parameter_312: class Program_weight_tensor_parameter_313: name = "parameter_313" + original_name = "batch_norm2d_75.w_1" shape = [384] dtype = "float32" min_val = float("-0.181599") @@ -3454,6 +3768,7 @@ class Program_weight_tensor_parameter_313: class Program_weight_tensor_parameter_314: name = "parameter_314" + original_name = "conv2d_79.w_0" shape = [384, 384, 3, 3] dtype = "float32" min_val = float("-0.0155422") @@ -3465,6 +3780,7 @@ class Program_weight_tensor_parameter_314: class Program_weight_tensor_parameter_315: name = "parameter_315" + original_name = "batch_norm2d_74.b_0" shape = [384] dtype = "float32" min_val = float("-0.0699577") @@ -3476,6 +3792,7 @@ class Program_weight_tensor_parameter_315: class Program_weight_tensor_parameter_316: name = "parameter_316" + original_name = "batch_norm2d_74.w_0" shape = [384] dtype = "float32" min_val = float("0.981937") @@ -3487,6 +3804,7 @@ class Program_weight_tensor_parameter_316: class Program_weight_tensor_parameter_317: name = "parameter_317" + original_name = "batch_norm2d_74.w_2" shape = [384] dtype = "float32" min_val = float("0.00841399") @@ -3498,6 +3816,7 @@ class Program_weight_tensor_parameter_317: class Program_weight_tensor_parameter_318: name = "parameter_318" + original_name = "batch_norm2d_74.w_1" shape = [384] dtype = "float32" min_val = float("-0.130353") @@ -3509,6 +3828,7 @@ class Program_weight_tensor_parameter_318: class Program_weight_tensor_parameter_319: name = "parameter_319" + original_name = "conv2d_78.w_0" shape = [384, 1024, 1, 1] dtype = "float32" min_val = float("-0.0187221") @@ -3520,6 +3840,7 @@ class Program_weight_tensor_parameter_319: class Program_weight_tensor_parameter_320: name = "parameter_320" + original_name = "batch_norm2d_73.b_0" shape = [384] dtype = "float32" min_val = float("-0.0240994") @@ -3531,6 +3852,7 @@ class Program_weight_tensor_parameter_320: class Program_weight_tensor_parameter_321: name = "parameter_321" + original_name = "batch_norm2d_73.w_0" shape = [384] dtype = "float32" min_val = float("0.994047") @@ -3542,6 +3864,7 @@ class Program_weight_tensor_parameter_321: class Program_weight_tensor_parameter_322: name = "parameter_322" + original_name = "batch_norm2d_73.w_2" shape = [384] dtype = "float32" min_val = float("0.0107661") @@ -3553,6 +3876,7 @@ class Program_weight_tensor_parameter_322: class Program_weight_tensor_parameter_323: name = "parameter_323" + original_name = "batch_norm2d_73.w_1" shape = [384] dtype = "float32" min_val = float("-0.201399") @@ -3564,6 +3888,7 @@ class Program_weight_tensor_parameter_323: class Program_weight_tensor_parameter_324: name = "parameter_324" + original_name = "conv2d_77.w_0" shape = [384, 1024, 1, 1] dtype = "float32" min_val = float("-0.038196") @@ -3575,6 +3900,7 @@ class Program_weight_tensor_parameter_324: class Program_weight_tensor_parameter_325: name = "parameter_325" + original_name = "layer_norm_1.b_0_deepcopy_47" shape = [1024] dtype = "float32" min_val = float("-3.19613e-10") @@ -3586,6 +3912,7 @@ class Program_weight_tensor_parameter_325: class Program_weight_tensor_parameter_326: name = "parameter_326" + original_name = "layer_norm_1.w_0_deepcopy_46" shape = [1024] dtype = "float32" min_val = float("0.826158") @@ -3597,6 +3924,7 @@ class Program_weight_tensor_parameter_326: class Program_weight_tensor_parameter_327: name = "parameter_327" + original_name = "linear_2.b_0_deepcopy_43" shape = [1024] dtype = "float32" min_val = float("-0.0184723") @@ -3608,6 +3936,7 @@ class Program_weight_tensor_parameter_327: class Program_weight_tensor_parameter_328: name = "parameter_328" + original_name = "linear_2.w_0_deepcopy_42" shape = [2048, 1024] dtype = "float32" min_val = float("-0.0186692") @@ -3619,6 +3948,7 @@ class Program_weight_tensor_parameter_328: class Program_weight_tensor_parameter_329: name = "parameter_329" + original_name = "linear_1.b_0_deepcopy_41" shape = [2048] dtype = "float32" min_val = float("-0.0258373") @@ -3630,6 +3960,7 @@ class Program_weight_tensor_parameter_329: class Program_weight_tensor_parameter_330: name = "parameter_330" + original_name = "linear_1.w_0_deepcopy_40" shape = [1024, 2048] dtype = "float32" min_val = float("-0.0261231") @@ -3641,6 +3972,7 @@ class Program_weight_tensor_parameter_330: class Program_weight_tensor_parameter_331: name = "parameter_331" + original_name = "layer_norm_0.b_0_deepcopy_45" shape = [1024] dtype = "float32" min_val = float("-0.000644044") @@ -3652,6 +3984,7 @@ class Program_weight_tensor_parameter_331: class Program_weight_tensor_parameter_332: name = "parameter_332" + original_name = "layer_norm_0.w_0_deepcopy_44" shape = [1024] dtype = "float32" min_val = float("0.825074") @@ -3663,6 +3996,7 @@ class Program_weight_tensor_parameter_332: class Program_weight_tensor_parameter_333: name = "parameter_333" + original_name = "linear_0.b_0_deepcopy_39" shape = [1024] dtype = "float32" min_val = float("-0.00057158") @@ -3674,6 +4008,7 @@ class Program_weight_tensor_parameter_333: class Program_weight_tensor_parameter_334: name = "parameter_334" + original_name = "linear_0.w_0_deepcopy_38" shape = [1024, 1024] dtype = "float32" min_val = float("-0.0452304") @@ -3685,6 +4020,7 @@ class Program_weight_tensor_parameter_334: class Program_weight_tensor_parameter_335: name = "parameter_335" + original_name = "layer_norm_1.b_0_deepcopy_35" shape = [1024] dtype = "float32" min_val = float("-0.000495877") @@ -3696,6 +4032,7 @@ class Program_weight_tensor_parameter_335: class Program_weight_tensor_parameter_336: name = "parameter_336" + original_name = "layer_norm_1.w_0_deepcopy_34" shape = [1024] dtype = "float32" min_val = float("0.825239") @@ -3707,6 +4044,7 @@ class Program_weight_tensor_parameter_336: class Program_weight_tensor_parameter_337: name = "parameter_337" + original_name = "linear_2.b_0_deepcopy_31" shape = [1024] dtype = "float32" min_val = float("-0.0182543") @@ -3718,6 +4056,7 @@ class Program_weight_tensor_parameter_337: class Program_weight_tensor_parameter_338: name = "parameter_338" + original_name = "linear_2.w_0_deepcopy_30" shape = [2048, 1024] dtype = "float32" min_val = float("-0.0185874") @@ -3729,6 +4068,7 @@ class Program_weight_tensor_parameter_338: class Program_weight_tensor_parameter_339: name = "parameter_339" + original_name = "linear_1.b_0_deepcopy_29" shape = [2048] dtype = "float32" min_val = float("-0.0258718") @@ -3740,6 +4080,7 @@ class Program_weight_tensor_parameter_339: class Program_weight_tensor_parameter_340: name = "parameter_340" + original_name = "linear_1.w_0_deepcopy_28" shape = [1024, 2048] dtype = "float32" min_val = float("-0.0260955") @@ -3751,6 +4092,7 @@ class Program_weight_tensor_parameter_340: class Program_weight_tensor_parameter_341: name = "parameter_341" + original_name = "layer_norm_0.b_0_deepcopy_33" shape = [1024] dtype = "float32" min_val = float("-0.000468906") @@ -3762,6 +4104,7 @@ class Program_weight_tensor_parameter_341: class Program_weight_tensor_parameter_342: name = "parameter_342" + original_name = "layer_norm_0.w_0_deepcopy_32" shape = [1024] dtype = "float32" min_val = float("0.825682") @@ -3773,6 +4116,7 @@ class Program_weight_tensor_parameter_342: class Program_weight_tensor_parameter_343: name = "parameter_343" + original_name = "linear_0.b_0_deepcopy_27" shape = [1024] dtype = "float32" min_val = float("-0.000528327") @@ -3784,6 +4128,7 @@ class Program_weight_tensor_parameter_343: class Program_weight_tensor_parameter_344: name = "parameter_344" + original_name = "linear_0.w_0_deepcopy_26" shape = [1024, 1024] dtype = "float32" min_val = float("-0.0450293") @@ -3795,6 +4140,7 @@ class Program_weight_tensor_parameter_344: class Program_weight_tensor_parameter_345: name = "parameter_345" + original_name = "layer_norm_1.b_0_deepcopy_23" shape = [1024] dtype = "float32" min_val = float("-0.000544222") @@ -3806,6 +4152,7 @@ class Program_weight_tensor_parameter_345: class Program_weight_tensor_parameter_346: name = "parameter_346" + original_name = "layer_norm_1.w_0_deepcopy_22" shape = [1024] dtype = "float32" min_val = float("0.825946") @@ -3817,6 +4164,7 @@ class Program_weight_tensor_parameter_346: class Program_weight_tensor_parameter_347: name = "parameter_347" + original_name = "linear_2.b_0_deepcopy_19" shape = [1024] dtype = "float32" min_val = float("-0.0184486") @@ -3828,6 +4176,7 @@ class Program_weight_tensor_parameter_347: class Program_weight_tensor_parameter_348: name = "parameter_348" + original_name = "linear_2.w_0_deepcopy_18" shape = [2048, 1024] dtype = "float32" min_val = float("-0.0185587") @@ -3839,6 +4188,7 @@ class Program_weight_tensor_parameter_348: class Program_weight_tensor_parameter_349: name = "parameter_349" + original_name = "linear_1.b_0_deepcopy_17" shape = [2048] dtype = "float32" min_val = float("-0.0259392") @@ -3850,6 +4200,7 @@ class Program_weight_tensor_parameter_349: class Program_weight_tensor_parameter_350: name = "parameter_350" + original_name = "linear_1.w_0_deepcopy_16" shape = [1024, 2048] dtype = "float32" min_val = float("-0.0261446") @@ -3861,6 +4212,7 @@ class Program_weight_tensor_parameter_350: class Program_weight_tensor_parameter_351: name = "parameter_351" + original_name = "layer_norm_0.b_0_deepcopy_21" shape = [1024] dtype = "float32" min_val = float("-0.000525085") @@ -3872,6 +4224,7 @@ class Program_weight_tensor_parameter_351: class Program_weight_tensor_parameter_352: name = "parameter_352" + original_name = "layer_norm_0.w_0_deepcopy_20" shape = [1024] dtype = "float32" min_val = float("0.826325") @@ -3883,6 +4236,7 @@ class Program_weight_tensor_parameter_352: class Program_weight_tensor_parameter_353: name = "parameter_353" + original_name = "linear_0.b_0_deepcopy_15" shape = [1024] dtype = "float32" min_val = float("-0.000560374") @@ -3894,6 +4248,7 @@ class Program_weight_tensor_parameter_353: class Program_weight_tensor_parameter_354: name = "parameter_354" + original_name = "linear_0.w_0_deepcopy_14" shape = [1024, 1024] dtype = "float32" min_val = float("-0.0451116") @@ -3905,6 +4260,7 @@ class Program_weight_tensor_parameter_354: class Program_weight_tensor_parameter_355: name = "parameter_355" + original_name = "layer_norm_1.b_0_deepcopy_11" shape = [1024] dtype = "float32" min_val = float("-0.000823618") @@ -3916,6 +4272,7 @@ class Program_weight_tensor_parameter_355: class Program_weight_tensor_parameter_356: name = "parameter_356" + original_name = "layer_norm_1.w_0_deepcopy_10" shape = [1024] dtype = "float32" min_val = float("0.826282") @@ -3927,6 +4284,7 @@ class Program_weight_tensor_parameter_356: class Program_weight_tensor_parameter_357: name = "parameter_357" + original_name = "linear_2.b_0_deepcopy_7" shape = [1024] dtype = "float32" min_val = float("-0.0185659") @@ -3938,6 +4296,7 @@ class Program_weight_tensor_parameter_357: class Program_weight_tensor_parameter_358: name = "parameter_358" + original_name = "linear_2.w_0_deepcopy_6" shape = [2048, 1024] dtype = "float32" min_val = float("-0.0186583") @@ -3949,6 +4308,7 @@ class Program_weight_tensor_parameter_358: class Program_weight_tensor_parameter_359: name = "parameter_359" + original_name = "linear_1.b_0_deepcopy_5" shape = [2048] dtype = "float32" min_val = float("-0.0260157") @@ -3960,6 +4320,7 @@ class Program_weight_tensor_parameter_359: class Program_weight_tensor_parameter_360: name = "parameter_360" + original_name = "linear_1.w_0_deepcopy_4" shape = [1024, 2048] dtype = "float32" min_val = float("-0.0261391") @@ -3971,6 +4332,7 @@ class Program_weight_tensor_parameter_360: class Program_weight_tensor_parameter_361: name = "parameter_361" + original_name = "layer_norm_0.b_0_deepcopy_9" shape = [1024] dtype = "float32" min_val = float("-0.000912874") @@ -3982,6 +4344,7 @@ class Program_weight_tensor_parameter_361: class Program_weight_tensor_parameter_362: name = "parameter_362" + original_name = "layer_norm_0.w_0_deepcopy_8" shape = [1024] dtype = "float32" min_val = float("0.826227") @@ -3993,6 +4356,7 @@ class Program_weight_tensor_parameter_362: class Program_weight_tensor_parameter_363: name = "parameter_363" + original_name = "linear_0.b_0_deepcopy_3" shape = [1024] dtype = "float32" min_val = float("-0.000894026") @@ -4004,6 +4368,7 @@ class Program_weight_tensor_parameter_363: class Program_weight_tensor_parameter_364: name = "parameter_364" + original_name = "linear_0.w_0_deepcopy_2" shape = [1024, 1024] dtype = "float32" min_val = float("-0.0456631") diff --git a/paddle_samples/PaddleX/PP-YOLOE_plus_SOD-largesize-L/subgraph_6/input_meta.py b/paddle_samples/PaddleX/PP-YOLOE_plus_SOD-largesize-L/subgraph_6/input_meta.py index 66d018686..bdb78768f 100644 --- a/paddle_samples/PaddleX/PP-YOLOE_plus_SOD-largesize-L/subgraph_6/input_meta.py +++ b/paddle_samples/PaddleX/PP-YOLOE_plus_SOD-largesize-L/subgraph_6/input_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_data_0: name = "data_0" + original_name = "param_52" shape = [1] dtype = "float32" data = [0.724553] @@ -7,6 +8,7 @@ class Program_weight_tensor_data_0: class Program_weight_tensor_data_1: name = "data_1" + original_name = "param_53" shape = [1] dtype = "float32" data = [0.710696] @@ -14,6 +16,7 @@ class Program_weight_tensor_data_1: class Program_weight_tensor_data_2: name = "data_2" + original_name = "param_54" shape = [1] dtype = "float32" data = [0.69274] @@ -21,6 +24,7 @@ class Program_weight_tensor_data_2: class Program_weight_tensor_data_3: name = "data_3" + original_name = "param_55" shape = [1] dtype = "float32" data = [0.697763] @@ -28,6 +32,7 @@ class Program_weight_tensor_data_3: class Program_weight_tensor_data_4: name = "data_4" + original_name = "param_56" shape = [1] dtype = "float32" data = [0.67767] @@ -35,6 +40,7 @@ class Program_weight_tensor_data_4: class Program_weight_tensor_data_5: name = "data_5" + original_name = "param_57" shape = [1] dtype = "float32" data = [0.628229] @@ -42,6 +48,7 @@ class Program_weight_tensor_data_5: class Program_weight_tensor_data_6: name = "data_6" + original_name = "param_58" shape = [1] dtype = "float32" data = [0.643942] @@ -49,6 +56,7 @@ class Program_weight_tensor_data_6: class Program_weight_tensor_data_7: name = "data_7" + original_name = "param_59" shape = [1] dtype = "float32" data = [0.633569] @@ -56,6 +64,7 @@ class Program_weight_tensor_data_7: class Program_weight_tensor_data_8: name = "data_8" + original_name = "param_60" shape = [1] dtype = "float32" data = [0.801205] @@ -63,6 +72,7 @@ class Program_weight_tensor_data_8: class Program_weight_tensor_data_9: name = "data_9" + original_name = "param_61" shape = [1] dtype = "float32" data = [0.652613] @@ -70,6 +80,7 @@ class Program_weight_tensor_data_9: class Program_weight_tensor_data_10: name = "data_10" + original_name = "param_62" shape = [1] dtype = "float32" data = [0.636874] @@ -77,6 +88,7 @@ class Program_weight_tensor_data_10: class Program_weight_tensor_data_11: name = "data_11" + original_name = "param_63" shape = [1] dtype = "float32" data = [0.631148] @@ -84,6 +96,7 @@ class Program_weight_tensor_data_11: class Program_weight_tensor_data_12: name = "data_12" + original_name = "param_64" shape = [1] dtype = "float32" data = [0.635341] @@ -91,6 +104,7 @@ class Program_weight_tensor_data_12: class Program_weight_tensor_data_13: name = "data_13" + original_name = "param_65" shape = [1] dtype = "float32" data = [0.640054] @@ -98,6 +112,7 @@ class Program_weight_tensor_data_13: class Program_weight_tensor_data_14: name = "data_14" + original_name = "param_66" shape = [1] dtype = "float32" data = [0.755822] @@ -105,6 +120,7 @@ class Program_weight_tensor_data_14: class Program_weight_tensor_data_15: name = "data_15" + original_name = "param_67" shape = [1] dtype = "float32" data = [0.575326] @@ -112,6 +128,7 @@ class Program_weight_tensor_data_15: class Program_weight_tensor_data_16: name = "data_16" + original_name = "param_68" shape = [1] dtype = "float32" data = [0.59257] @@ -119,6 +136,7 @@ class Program_weight_tensor_data_16: class Program_weight_tensor_data_17: name = "data_17" + original_name = "param_69" shape = [1] dtype = "float32" data = [0.72331] @@ -126,6 +144,7 @@ class Program_weight_tensor_data_17: class Program_weight_tensor_data_18: name = "data_18" + original_name = "param_70" shape = [1024, 3072] dtype = "float32" min_val = float("-0.033771") @@ -137,6 +156,7 @@ class Program_weight_tensor_data_18: class Program_weight_tensor_data_19: name = "data_19" + original_name = "param_71" shape = [3072] dtype = "float32" min_val = float("-0.000858009") @@ -148,6 +168,7 @@ class Program_weight_tensor_data_19: class Program_weight_tensor_data_20: name = "data_20" + original_name = "param_72" shape = [1024, 3072] dtype = "float32" min_val = float("-0.0324395") @@ -159,6 +180,7 @@ class Program_weight_tensor_data_20: class Program_weight_tensor_data_21: name = "data_21" + original_name = "param_73" shape = [3072] dtype = "float32" min_val = float("-0.000630237") @@ -170,6 +192,7 @@ class Program_weight_tensor_data_21: class Program_weight_tensor_data_22: name = "data_22" + original_name = "param_74" shape = [1024, 3072] dtype = "float32" min_val = float("-0.0321875") @@ -181,6 +204,7 @@ class Program_weight_tensor_data_22: class Program_weight_tensor_data_23: name = "data_23" + original_name = "param_75" shape = [3072] dtype = "float32" min_val = float("-0.000429784") @@ -192,6 +216,7 @@ class Program_weight_tensor_data_23: class Program_weight_tensor_data_24: name = "data_24" + original_name = "param_76" shape = [1024, 3072] dtype = "float32" min_val = float("-0.0321313") @@ -203,6 +228,7 @@ class Program_weight_tensor_data_24: class Program_weight_tensor_data_25: name = "data_25" + original_name = "param_77" shape = [3072] dtype = "float32" min_val = float("-0.000397408") @@ -214,6 +240,7 @@ class Program_weight_tensor_data_25: class Program_weight_tensor_data_26: name = "data_26" + original_name = "var_1784" shape = [1, 3, 1088, 1088] dtype = "float32" max_val = float("1.0") diff --git a/paddle_samples/PaddleX/PP-YOLOE_plus_SOD-largesize-L/subgraph_6/weight_meta.py b/paddle_samples/PaddleX/PP-YOLOE_plus_SOD-largesize-L/subgraph_6/weight_meta.py index 7bb3e9a3b..e18ff7454 100644 --- a/paddle_samples/PaddleX/PP-YOLOE_plus_SOD-largesize-L/subgraph_6/weight_meta.py +++ b/paddle_samples/PaddleX/PP-YOLOE_plus_SOD-largesize-L/subgraph_6/weight_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_parameter_0: name = "parameter_0" + original_name = "batch_norm2d_137.b_0" shape = [768] dtype = "float32" min_val = float("-0.175875") @@ -11,6 +12,7 @@ class Program_weight_tensor_parameter_0: class Program_weight_tensor_parameter_1: name = "parameter_1" + original_name = "batch_norm2d_137.w_0" shape = [768] dtype = "float32" min_val = float("0.939895") @@ -22,6 +24,7 @@ class Program_weight_tensor_parameter_1: class Program_weight_tensor_parameter_2: name = "parameter_2" + original_name = "batch_norm2d_137.w_2" shape = [768] dtype = "float32" min_val = float("0.00117681") @@ -33,6 +36,7 @@ class Program_weight_tensor_parameter_2: class Program_weight_tensor_parameter_3: name = "parameter_3" + original_name = "batch_norm2d_137.w_1" shape = [768] dtype = "float32" min_val = float("-0.13338") @@ -44,6 +48,7 @@ class Program_weight_tensor_parameter_3: class Program_weight_tensor_parameter_4: name = "parameter_4" + original_name = "conv2d_141.w_0" shape = [768, 768, 1, 1] dtype = "float32" min_val = float("-0.0548133") @@ -55,6 +60,7 @@ class Program_weight_tensor_parameter_4: class Program_weight_tensor_parameter_5: name = "parameter_5" + original_name = "batch_norm2d_136.b_0" shape = [384] dtype = "float32" min_val = float("-0.14169") @@ -66,6 +72,7 @@ class Program_weight_tensor_parameter_5: class Program_weight_tensor_parameter_6: name = "parameter_6" + original_name = "batch_norm2d_136.w_0" shape = [384] dtype = "float32" min_val = float("0.945748") @@ -77,6 +84,7 @@ class Program_weight_tensor_parameter_6: class Program_weight_tensor_parameter_7: name = "parameter_7" + original_name = "batch_norm2d_136.w_2" shape = [384] dtype = "float32" min_val = float("0.000824752") @@ -88,6 +96,7 @@ class Program_weight_tensor_parameter_7: class Program_weight_tensor_parameter_8: name = "parameter_8" + original_name = "batch_norm2d_136.w_1" shape = [384] dtype = "float32" min_val = float("-0.0551455") @@ -99,6 +108,7 @@ class Program_weight_tensor_parameter_8: class Program_weight_tensor_parameter_9: name = "parameter_9" + original_name = "conv2d_140.w_0" shape = [384, 384, 1, 1] dtype = "float32" min_val = float("-0.0299324") @@ -110,6 +120,7 @@ class Program_weight_tensor_parameter_9: class Program_weight_tensor_parameter_10: name = "parameter_10" + original_name = "batch_norm2d_135.b_0" shape = [384] dtype = "float32" min_val = float("-0.14169") @@ -121,6 +132,7 @@ class Program_weight_tensor_parameter_10: class Program_weight_tensor_parameter_11: name = "parameter_11" + original_name = "batch_norm2d_135.w_0" shape = [384] dtype = "float32" min_val = float("0.968039") @@ -132,6 +144,7 @@ class Program_weight_tensor_parameter_11: class Program_weight_tensor_parameter_12: name = "parameter_12" + original_name = "batch_norm2d_135.w_2" shape = [384] dtype = "float32" min_val = float("0.00204949") @@ -143,6 +156,7 @@ class Program_weight_tensor_parameter_12: class Program_weight_tensor_parameter_13: name = "parameter_13" + original_name = "batch_norm2d_135.w_1" shape = [384] dtype = "float32" min_val = float("-0.203105") @@ -154,6 +168,7 @@ class Program_weight_tensor_parameter_13: class Program_weight_tensor_parameter_14: name = "parameter_14" + original_name = "conv2d_139.w_0" shape = [384, 384, 3, 3] dtype = "float32" min_val = float("-0.029908") @@ -165,6 +180,7 @@ class Program_weight_tensor_parameter_14: class Program_weight_tensor_parameter_15: name = "parameter_15" + original_name = "batch_norm2d_134.b_0" shape = [384] dtype = "float32" min_val = float("-0.170219") @@ -176,6 +192,7 @@ class Program_weight_tensor_parameter_15: class Program_weight_tensor_parameter_16: name = "parameter_16" + original_name = "batch_norm2d_134.w_0" shape = [384] dtype = "float32" min_val = float("0.975222") @@ -187,6 +204,7 @@ class Program_weight_tensor_parameter_16: class Program_weight_tensor_parameter_17: name = "parameter_17" + original_name = "batch_norm2d_134.w_2" shape = [384] dtype = "float32" min_val = float("0.00543045") @@ -198,6 +216,7 @@ class Program_weight_tensor_parameter_17: class Program_weight_tensor_parameter_18: name = "parameter_18" + original_name = "batch_norm2d_134.w_1" shape = [384] dtype = "float32" min_val = float("-0.266668") @@ -209,6 +228,7 @@ class Program_weight_tensor_parameter_18: class Program_weight_tensor_parameter_19: name = "parameter_19" + original_name = "conv2d_138.w_0" shape = [384, 384, 3, 3] dtype = "float32" min_val = float("-0.0331339") @@ -220,6 +240,7 @@ class Program_weight_tensor_parameter_19: class Program_weight_tensor_parameter_20: name = "parameter_20" + original_name = "batch_norm2d_133.b_0" shape = [384] dtype = "float32" min_val = float("-0.105219") @@ -231,6 +252,7 @@ class Program_weight_tensor_parameter_20: class Program_weight_tensor_parameter_21: name = "parameter_21" + original_name = "batch_norm2d_133.w_0" shape = [384] dtype = "float32" min_val = float("0.945357") @@ -242,6 +264,7 @@ class Program_weight_tensor_parameter_21: class Program_weight_tensor_parameter_22: name = "parameter_22" + original_name = "batch_norm2d_133.w_2" shape = [384] dtype = "float32" min_val = float("0.000719695") @@ -253,6 +276,7 @@ class Program_weight_tensor_parameter_22: class Program_weight_tensor_parameter_23: name = "parameter_23" + original_name = "batch_norm2d_133.w_1" shape = [384] dtype = "float32" min_val = float("-0.0846979") @@ -264,6 +288,7 @@ class Program_weight_tensor_parameter_23: class Program_weight_tensor_parameter_24: name = "parameter_24" + original_name = "conv2d_137.w_0" shape = [384, 384, 1, 1] dtype = "float32" min_val = float("-0.0267958") @@ -275,6 +300,7 @@ class Program_weight_tensor_parameter_24: class Program_weight_tensor_parameter_25: name = "parameter_25" + original_name = "batch_norm2d_132.b_0" shape = [384] dtype = "float32" min_val = float("-0.105219") @@ -286,6 +312,7 @@ class Program_weight_tensor_parameter_25: class Program_weight_tensor_parameter_26: name = "parameter_26" + original_name = "batch_norm2d_132.w_0" shape = [384] dtype = "float32" min_val = float("0.959852") @@ -297,6 +324,7 @@ class Program_weight_tensor_parameter_26: class Program_weight_tensor_parameter_27: name = "parameter_27" + original_name = "batch_norm2d_132.w_2" shape = [384] dtype = "float32" min_val = float("0.0024927") @@ -308,6 +336,7 @@ class Program_weight_tensor_parameter_27: class Program_weight_tensor_parameter_28: name = "parameter_28" + original_name = "batch_norm2d_132.w_1" shape = [384] dtype = "float32" min_val = float("-0.215525") @@ -319,6 +348,7 @@ class Program_weight_tensor_parameter_28: class Program_weight_tensor_parameter_29: name = "parameter_29" + original_name = "conv2d_136.w_0" shape = [384, 384, 3, 3] dtype = "float32" min_val = float("-0.0363929") @@ -330,6 +360,7 @@ class Program_weight_tensor_parameter_29: class Program_weight_tensor_parameter_30: name = "parameter_30" + original_name = "batch_norm2d_131.b_0" shape = [384] dtype = "float32" min_val = float("-0.0896627") @@ -341,6 +372,7 @@ class Program_weight_tensor_parameter_30: class Program_weight_tensor_parameter_31: name = "parameter_31" + original_name = "batch_norm2d_131.w_0" shape = [384] dtype = "float32" min_val = float("0.933291") @@ -352,6 +384,7 @@ class Program_weight_tensor_parameter_31: class Program_weight_tensor_parameter_32: name = "parameter_32" + original_name = "batch_norm2d_131.w_2" shape = [384] dtype = "float32" min_val = float("0.00555425") @@ -363,6 +396,7 @@ class Program_weight_tensor_parameter_32: class Program_weight_tensor_parameter_33: name = "parameter_33" + original_name = "batch_norm2d_131.w_1" shape = [384] dtype = "float32" min_val = float("-0.23163") @@ -374,6 +408,7 @@ class Program_weight_tensor_parameter_33: class Program_weight_tensor_parameter_34: name = "parameter_34" + original_name = "conv2d_135.w_0" shape = [384, 384, 3, 3] dtype = "float32" min_val = float("-0.0397048") @@ -385,6 +420,7 @@ class Program_weight_tensor_parameter_34: class Program_weight_tensor_parameter_35: name = "parameter_35" + original_name = "batch_norm2d_130.b_0" shape = [384] dtype = "float32" min_val = float("-0.116341") @@ -396,6 +432,7 @@ class Program_weight_tensor_parameter_35: class Program_weight_tensor_parameter_36: name = "parameter_36" + original_name = "batch_norm2d_130.w_0" shape = [384] dtype = "float32" min_val = float("0.929383") @@ -407,6 +444,7 @@ class Program_weight_tensor_parameter_36: class Program_weight_tensor_parameter_37: name = "parameter_37" + original_name = "batch_norm2d_130.w_2" shape = [384] dtype = "float32" min_val = float("0.00126812") @@ -418,6 +456,7 @@ class Program_weight_tensor_parameter_37: class Program_weight_tensor_parameter_38: name = "parameter_38" + original_name = "batch_norm2d_130.w_1" shape = [384] dtype = "float32" min_val = float("-0.0553816") @@ -429,6 +468,7 @@ class Program_weight_tensor_parameter_38: class Program_weight_tensor_parameter_39: name = "parameter_39" + original_name = "conv2d_134.w_0" shape = [384, 384, 1, 1] dtype = "float32" min_val = float("-0.0386337") @@ -440,6 +480,7 @@ class Program_weight_tensor_parameter_39: class Program_weight_tensor_parameter_40: name = "parameter_40" + original_name = "batch_norm2d_129.b_0" shape = [384] dtype = "float32" min_val = float("-0.116341") @@ -451,6 +492,7 @@ class Program_weight_tensor_parameter_40: class Program_weight_tensor_parameter_41: name = "parameter_41" + original_name = "batch_norm2d_129.w_0" shape = [384] dtype = "float32" min_val = float("0.981354") @@ -462,6 +504,7 @@ class Program_weight_tensor_parameter_41: class Program_weight_tensor_parameter_42: name = "parameter_42" + original_name = "batch_norm2d_129.w_2" shape = [384] dtype = "float32" min_val = float("0.00525993") @@ -473,6 +516,7 @@ class Program_weight_tensor_parameter_42: class Program_weight_tensor_parameter_43: name = "parameter_43" + original_name = "batch_norm2d_129.w_1" shape = [384] dtype = "float32" min_val = float("-0.189916") @@ -484,6 +528,7 @@ class Program_weight_tensor_parameter_43: class Program_weight_tensor_parameter_44: name = "parameter_44" + original_name = "conv2d_133.w_0" shape = [384, 384, 3, 3] dtype = "float32" min_val = float("-0.0360859") @@ -495,6 +540,7 @@ class Program_weight_tensor_parameter_44: class Program_weight_tensor_parameter_45: name = "parameter_45" + original_name = "batch_norm2d_128.b_0" shape = [384] dtype = "float32" min_val = float("-0.107113") @@ -506,6 +552,7 @@ class Program_weight_tensor_parameter_45: class Program_weight_tensor_parameter_46: name = "parameter_46" + original_name = "batch_norm2d_128.w_0" shape = [384] dtype = "float32" min_val = float("0.944795") @@ -517,6 +564,7 @@ class Program_weight_tensor_parameter_46: class Program_weight_tensor_parameter_47: name = "parameter_47" + original_name = "batch_norm2d_128.w_2" shape = [384] dtype = "float32" min_val = float("0.0055232") @@ -528,6 +576,7 @@ class Program_weight_tensor_parameter_47: class Program_weight_tensor_parameter_48: name = "parameter_48" + original_name = "batch_norm2d_128.w_1" shape = [384] dtype = "float32" min_val = float("-0.154281") @@ -539,6 +588,7 @@ class Program_weight_tensor_parameter_48: class Program_weight_tensor_parameter_49: name = "parameter_49" + original_name = "conv2d_132.w_0" shape = [384, 384, 3, 3] dtype = "float32" min_val = float("-0.0279282") @@ -550,6 +600,7 @@ class Program_weight_tensor_parameter_49: class Program_weight_tensor_parameter_50: name = "parameter_50" + original_name = "batch_norm2d_127.b_0" shape = [384] dtype = "float32" min_val = float("-0.10674") @@ -561,6 +612,7 @@ class Program_weight_tensor_parameter_50: class Program_weight_tensor_parameter_51: name = "parameter_51" + original_name = "batch_norm2d_127.w_0" shape = [384] dtype = "float32" min_val = float("0.973756") @@ -572,6 +624,7 @@ class Program_weight_tensor_parameter_51: class Program_weight_tensor_parameter_52: name = "parameter_52" + original_name = "batch_norm2d_127.w_2" shape = [384] dtype = "float32" min_val = float("0.00240361") @@ -583,6 +636,7 @@ class Program_weight_tensor_parameter_52: class Program_weight_tensor_parameter_53: name = "parameter_53" + original_name = "batch_norm2d_127.w_1" shape = [384] dtype = "float32" min_val = float("-0.100425") @@ -594,6 +648,7 @@ class Program_weight_tensor_parameter_53: class Program_weight_tensor_parameter_54: name = "parameter_54" + original_name = "conv2d_131.w_0" shape = [384, 1152, 1, 1] dtype = "float32" min_val = float("-0.0619005") @@ -605,6 +660,7 @@ class Program_weight_tensor_parameter_54: class Program_weight_tensor_parameter_55: name = "parameter_55" + original_name = "batch_norm2d_126.b_0" shape = [384] dtype = "float32" min_val = float("-0.0424904") @@ -616,6 +672,7 @@ class Program_weight_tensor_parameter_55: class Program_weight_tensor_parameter_56: name = "parameter_56" + original_name = "batch_norm2d_126.w_0" shape = [384] dtype = "float32" min_val = float("0.959519") @@ -627,6 +684,7 @@ class Program_weight_tensor_parameter_56: class Program_weight_tensor_parameter_57: name = "parameter_57" + original_name = "batch_norm2d_126.w_2" shape = [384] dtype = "float32" min_val = float("0.00126033") @@ -638,6 +696,7 @@ class Program_weight_tensor_parameter_57: class Program_weight_tensor_parameter_58: name = "parameter_58" + original_name = "batch_norm2d_126.w_1" shape = [384] dtype = "float32" min_val = float("-0.110881") @@ -649,6 +708,7 @@ class Program_weight_tensor_parameter_58: class Program_weight_tensor_parameter_59: name = "parameter_59" + original_name = "conv2d_130.w_0" shape = [384, 1152, 1, 1] dtype = "float32" min_val = float("-0.0245473") @@ -660,6 +720,7 @@ class Program_weight_tensor_parameter_59: class Program_weight_tensor_parameter_60: name = "parameter_60" + original_name = "batch_norm2d_125.b_0" shape = [384] dtype = "float32" min_val = float("-0.0529748") @@ -671,6 +732,7 @@ class Program_weight_tensor_parameter_60: class Program_weight_tensor_parameter_61: name = "parameter_61" + original_name = "batch_norm2d_125.w_0" shape = [384] dtype = "float32" min_val = float("0.988678") @@ -682,6 +744,7 @@ class Program_weight_tensor_parameter_61: class Program_weight_tensor_parameter_62: name = "parameter_62" + original_name = "batch_norm2d_125.w_2" shape = [384] dtype = "float32" min_val = float("0.00462734") @@ -693,6 +756,7 @@ class Program_weight_tensor_parameter_62: class Program_weight_tensor_parameter_63: name = "parameter_63" + original_name = "batch_norm2d_125.w_1" shape = [384] dtype = "float32" min_val = float("-0.443847") @@ -704,6 +768,7 @@ class Program_weight_tensor_parameter_63: class Program_weight_tensor_parameter_64: name = "parameter_64" + original_name = "conv2d_129.w_0" shape = [384, 384, 3, 3] dtype = "float32" min_val = float("-0.0212973") @@ -715,6 +780,7 @@ class Program_weight_tensor_parameter_64: class Program_weight_tensor_parameter_65: name = "parameter_65" + original_name = "batch_norm2d_124.b_0" shape = [384] dtype = "float32" min_val = float("-0.222314") @@ -726,6 +792,7 @@ class Program_weight_tensor_parameter_65: class Program_weight_tensor_parameter_66: name = "parameter_66" + original_name = "batch_norm2d_124.w_0" shape = [384] dtype = "float32" min_val = float("0.919258") @@ -737,6 +804,7 @@ class Program_weight_tensor_parameter_66: class Program_weight_tensor_parameter_67: name = "parameter_67" + original_name = "batch_norm2d_124.w_2" shape = [384] dtype = "float32" min_val = float("0.00377408") @@ -748,6 +816,7 @@ class Program_weight_tensor_parameter_67: class Program_weight_tensor_parameter_68: name = "parameter_68" + original_name = "batch_norm2d_124.w_1" shape = [384] dtype = "float32" min_val = float("-0.129308") @@ -759,6 +828,7 @@ class Program_weight_tensor_parameter_68: class Program_weight_tensor_parameter_69: name = "parameter_69" + original_name = "conv2d_128.w_0" shape = [384, 384, 1, 1] dtype = "float32" min_val = float("-0.0788092") @@ -770,6 +840,7 @@ class Program_weight_tensor_parameter_69: class Program_weight_tensor_parameter_70: name = "parameter_70" + original_name = "batch_norm2d_123.b_0" shape = [192] dtype = "float32" min_val = float("-0.165903") @@ -781,6 +852,7 @@ class Program_weight_tensor_parameter_70: class Program_weight_tensor_parameter_71: name = "parameter_71" + original_name = "batch_norm2d_123.w_0" shape = [192] dtype = "float32" min_val = float("0.841187") @@ -792,6 +864,7 @@ class Program_weight_tensor_parameter_71: class Program_weight_tensor_parameter_72: name = "parameter_72" + original_name = "batch_norm2d_123.w_2" shape = [192] dtype = "float32" min_val = float("0.00135964") @@ -803,6 +876,7 @@ class Program_weight_tensor_parameter_72: class Program_weight_tensor_parameter_73: name = "parameter_73" + original_name = "batch_norm2d_123.w_1" shape = [192] dtype = "float32" min_val = float("-0.0635033") @@ -814,6 +888,7 @@ class Program_weight_tensor_parameter_73: class Program_weight_tensor_parameter_74: name = "parameter_74" + original_name = "conv2d_127.w_0" shape = [192, 192, 1, 1] dtype = "float32" min_val = float("-0.0496361") @@ -825,6 +900,7 @@ class Program_weight_tensor_parameter_74: class Program_weight_tensor_parameter_75: name = "parameter_75" + original_name = "batch_norm2d_122.b_0" shape = [192] dtype = "float32" min_val = float("-0.165903") @@ -836,6 +912,7 @@ class Program_weight_tensor_parameter_75: class Program_weight_tensor_parameter_76: name = "parameter_76" + original_name = "batch_norm2d_122.w_0" shape = [192] dtype = "float32" min_val = float("0.72984") @@ -847,6 +924,7 @@ class Program_weight_tensor_parameter_76: class Program_weight_tensor_parameter_77: name = "parameter_77" + original_name = "batch_norm2d_122.w_2" shape = [192] dtype = "float32" min_val = float("0.00534163") @@ -858,6 +936,7 @@ class Program_weight_tensor_parameter_77: class Program_weight_tensor_parameter_78: name = "parameter_78" + original_name = "batch_norm2d_122.w_1" shape = [192] dtype = "float32" min_val = float("-0.219364") @@ -869,6 +948,7 @@ class Program_weight_tensor_parameter_78: class Program_weight_tensor_parameter_79: name = "parameter_79" + original_name = "conv2d_126.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.0430374") @@ -880,6 +960,7 @@ class Program_weight_tensor_parameter_79: class Program_weight_tensor_parameter_80: name = "parameter_80" + original_name = "batch_norm2d_121.b_0" shape = [192] dtype = "float32" min_val = float("-0.191344") @@ -891,6 +972,7 @@ class Program_weight_tensor_parameter_80: class Program_weight_tensor_parameter_81: name = "parameter_81" + original_name = "batch_norm2d_121.w_0" shape = [192] dtype = "float32" min_val = float("0.897737") @@ -902,6 +984,7 @@ class Program_weight_tensor_parameter_81: class Program_weight_tensor_parameter_82: name = "parameter_82" + original_name = "batch_norm2d_121.w_2" shape = [192] dtype = "float32" min_val = float("0.0105665") @@ -913,6 +996,7 @@ class Program_weight_tensor_parameter_82: class Program_weight_tensor_parameter_83: name = "parameter_83" + original_name = "batch_norm2d_121.w_1" shape = [192] dtype = "float32" min_val = float("-0.295941") @@ -924,6 +1008,7 @@ class Program_weight_tensor_parameter_83: class Program_weight_tensor_parameter_84: name = "parameter_84" + original_name = "conv2d_125.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.047378") @@ -935,6 +1020,7 @@ class Program_weight_tensor_parameter_84: class Program_weight_tensor_parameter_85: name = "parameter_85" + original_name = "batch_norm2d_120.b_0" shape = [192] dtype = "float32" min_val = float("-0.191632") @@ -946,6 +1032,7 @@ class Program_weight_tensor_parameter_85: class Program_weight_tensor_parameter_86: name = "parameter_86" + original_name = "batch_norm2d_120.w_0" shape = [192] dtype = "float32" min_val = float("0.922153") @@ -957,6 +1044,7 @@ class Program_weight_tensor_parameter_86: class Program_weight_tensor_parameter_87: name = "parameter_87" + original_name = "batch_norm2d_120.w_2" shape = [192] dtype = "float32" min_val = float("0.00111369") @@ -968,6 +1056,7 @@ class Program_weight_tensor_parameter_87: class Program_weight_tensor_parameter_88: name = "parameter_88" + original_name = "batch_norm2d_120.w_1" shape = [192] dtype = "float32" min_val = float("-0.0705634") @@ -979,6 +1068,7 @@ class Program_weight_tensor_parameter_88: class Program_weight_tensor_parameter_89: name = "parameter_89" + original_name = "conv2d_124.w_0" shape = [192, 192, 1, 1] dtype = "float32" min_val = float("-0.0386531") @@ -990,6 +1080,7 @@ class Program_weight_tensor_parameter_89: class Program_weight_tensor_parameter_90: name = "parameter_90" + original_name = "batch_norm2d_119.b_0" shape = [192] dtype = "float32" min_val = float("-0.191632") @@ -1001,6 +1092,7 @@ class Program_weight_tensor_parameter_90: class Program_weight_tensor_parameter_91: name = "parameter_91" + original_name = "batch_norm2d_119.w_0" shape = [192] dtype = "float32" min_val = float("0.968104") @@ -1012,6 +1104,7 @@ class Program_weight_tensor_parameter_91: class Program_weight_tensor_parameter_92: name = "parameter_92" + original_name = "batch_norm2d_119.w_2" shape = [192] dtype = "float32" min_val = float("0.00435298") @@ -1023,6 +1116,7 @@ class Program_weight_tensor_parameter_92: class Program_weight_tensor_parameter_93: name = "parameter_93" + original_name = "batch_norm2d_119.w_1" shape = [192] dtype = "float32" min_val = float("-0.186388") @@ -1034,6 +1128,7 @@ class Program_weight_tensor_parameter_93: class Program_weight_tensor_parameter_94: name = "parameter_94" + original_name = "conv2d_123.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.0471114") @@ -1045,6 +1140,7 @@ class Program_weight_tensor_parameter_94: class Program_weight_tensor_parameter_95: name = "parameter_95" + original_name = "batch_norm2d_118.b_0" shape = [192] dtype = "float32" min_val = float("-0.188926") @@ -1056,6 +1152,7 @@ class Program_weight_tensor_parameter_95: class Program_weight_tensor_parameter_96: name = "parameter_96" + original_name = "batch_norm2d_118.w_0" shape = [192] dtype = "float32" min_val = float("0.880419") @@ -1067,6 +1164,7 @@ class Program_weight_tensor_parameter_96: class Program_weight_tensor_parameter_97: name = "parameter_97" + original_name = "batch_norm2d_118.w_2" shape = [192] dtype = "float32" min_val = float("0.00811769") @@ -1078,6 +1176,7 @@ class Program_weight_tensor_parameter_97: class Program_weight_tensor_parameter_98: name = "parameter_98" + original_name = "batch_norm2d_118.w_1" shape = [192] dtype = "float32" min_val = float("-0.117015") @@ -1089,6 +1188,7 @@ class Program_weight_tensor_parameter_98: class Program_weight_tensor_parameter_99: name = "parameter_99" + original_name = "conv2d_122.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.0427984") @@ -1100,6 +1200,7 @@ class Program_weight_tensor_parameter_99: class Program_weight_tensor_parameter_100: name = "parameter_100" + original_name = "batch_norm2d_117.b_0" shape = [192] dtype = "float32" min_val = float("-0.229476") @@ -1111,6 +1212,7 @@ class Program_weight_tensor_parameter_100: class Program_weight_tensor_parameter_101: name = "parameter_101" + original_name = "batch_norm2d_117.w_0" shape = [192] dtype = "float32" min_val = float("0.900428") @@ -1122,6 +1224,7 @@ class Program_weight_tensor_parameter_101: class Program_weight_tensor_parameter_102: name = "parameter_102" + original_name = "batch_norm2d_117.w_2" shape = [192] dtype = "float32" min_val = float("0.00171033") @@ -1133,6 +1236,7 @@ class Program_weight_tensor_parameter_102: class Program_weight_tensor_parameter_103: name = "parameter_103" + original_name = "batch_norm2d_117.w_1" shape = [192] dtype = "float32" min_val = float("-0.0390528") @@ -1144,6 +1248,7 @@ class Program_weight_tensor_parameter_103: class Program_weight_tensor_parameter_104: name = "parameter_104" + original_name = "conv2d_121.w_0" shape = [192, 192, 1, 1] dtype = "float32" min_val = float("-0.0436521") @@ -1155,6 +1260,7 @@ class Program_weight_tensor_parameter_104: class Program_weight_tensor_parameter_105: name = "parameter_105" + original_name = "batch_norm2d_116.b_0" shape = [192] dtype = "float32" min_val = float("-0.229476") @@ -1166,6 +1272,7 @@ class Program_weight_tensor_parameter_105: class Program_weight_tensor_parameter_106: name = "parameter_106" + original_name = "batch_norm2d_116.w_0" shape = [192] dtype = "float32" min_val = float("0.947654") @@ -1177,6 +1284,7 @@ class Program_weight_tensor_parameter_106: class Program_weight_tensor_parameter_107: name = "parameter_107" + original_name = "batch_norm2d_116.w_2" shape = [192] dtype = "float32" min_val = float("0.00720341") @@ -1188,6 +1296,7 @@ class Program_weight_tensor_parameter_107: class Program_weight_tensor_parameter_108: name = "parameter_108" + original_name = "batch_norm2d_116.w_1" shape = [192] dtype = "float32" min_val = float("-0.129987") @@ -1199,6 +1308,7 @@ class Program_weight_tensor_parameter_108: class Program_weight_tensor_parameter_109: name = "parameter_109" + original_name = "conv2d_120.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.0485052") @@ -1210,6 +1320,7 @@ class Program_weight_tensor_parameter_109: class Program_weight_tensor_parameter_110: name = "parameter_110" + original_name = "batch_norm2d_115.b_0" shape = [192] dtype = "float32" min_val = float("-0.234305") @@ -1221,6 +1332,7 @@ class Program_weight_tensor_parameter_110: class Program_weight_tensor_parameter_111: name = "parameter_111" + original_name = "batch_norm2d_115.w_0" shape = [192] dtype = "float32" min_val = float("0.886145") @@ -1232,6 +1344,7 @@ class Program_weight_tensor_parameter_111: class Program_weight_tensor_parameter_112: name = "parameter_112" + original_name = "batch_norm2d_115.w_2" shape = [192] dtype = "float32" min_val = float("0.00887124") @@ -1243,6 +1356,7 @@ class Program_weight_tensor_parameter_112: class Program_weight_tensor_parameter_113: name = "parameter_113" + original_name = "batch_norm2d_115.w_1" shape = [192] dtype = "float32" min_val = float("-0.180694") @@ -1254,6 +1368,7 @@ class Program_weight_tensor_parameter_113: class Program_weight_tensor_parameter_114: name = "parameter_114" + original_name = "conv2d_119.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.0410156") @@ -1265,6 +1380,7 @@ class Program_weight_tensor_parameter_114: class Program_weight_tensor_parameter_115: name = "parameter_115" + original_name = "batch_norm2d_114.b_0" shape = [192] dtype = "float32" min_val = float("-0.199948") @@ -1276,6 +1392,7 @@ class Program_weight_tensor_parameter_115: class Program_weight_tensor_parameter_116: name = "parameter_116" + original_name = "batch_norm2d_114.w_0" shape = [192] dtype = "float32" min_val = float("0.925493") @@ -1287,6 +1404,7 @@ class Program_weight_tensor_parameter_116: class Program_weight_tensor_parameter_117: name = "parameter_117" + original_name = "batch_norm2d_114.w_2" shape = [192] dtype = "float32" min_val = float("0.0044011") @@ -1298,6 +1416,7 @@ class Program_weight_tensor_parameter_117: class Program_weight_tensor_parameter_118: name = "parameter_118" + original_name = "batch_norm2d_114.w_1" shape = [192] dtype = "float32" min_val = float("-0.0887579") @@ -1309,6 +1428,7 @@ class Program_weight_tensor_parameter_118: class Program_weight_tensor_parameter_119: name = "parameter_119" + original_name = "conv2d_118.w_0" shape = [192, 576, 1, 1] dtype = "float32" min_val = float("-0.0628757") @@ -1320,6 +1440,7 @@ class Program_weight_tensor_parameter_119: class Program_weight_tensor_parameter_120: name = "parameter_120" + original_name = "batch_norm2d_113.b_0" shape = [192] dtype = "float32" min_val = float("-0.0999632") @@ -1331,6 +1452,7 @@ class Program_weight_tensor_parameter_120: class Program_weight_tensor_parameter_121: name = "parameter_121" + original_name = "batch_norm2d_113.w_0" shape = [192] dtype = "float32" min_val = float("0.923856") @@ -1342,6 +1464,7 @@ class Program_weight_tensor_parameter_121: class Program_weight_tensor_parameter_122: name = "parameter_122" + original_name = "batch_norm2d_113.w_2" shape = [192] dtype = "float32" min_val = float("0.00336859") @@ -1353,6 +1476,7 @@ class Program_weight_tensor_parameter_122: class Program_weight_tensor_parameter_123: name = "parameter_123" + original_name = "batch_norm2d_113.w_1" shape = [192] dtype = "float32" min_val = float("-0.0725677") @@ -1364,6 +1488,7 @@ class Program_weight_tensor_parameter_123: class Program_weight_tensor_parameter_124: name = "parameter_124" + original_name = "conv2d_117.w_0" shape = [192, 576, 1, 1] dtype = "float32" min_val = float("-0.0557051") @@ -1375,6 +1500,7 @@ class Program_weight_tensor_parameter_124: class Program_weight_tensor_parameter_125: name = "parameter_125" + original_name = "batch_norm2d_112.b_0" shape = [192] dtype = "float32" min_val = float("-0.15908") @@ -1386,6 +1512,7 @@ class Program_weight_tensor_parameter_125: class Program_weight_tensor_parameter_126: name = "parameter_126" + original_name = "batch_norm2d_112.w_0" shape = [192] dtype = "float32" min_val = float("0.921159") @@ -1397,6 +1524,7 @@ class Program_weight_tensor_parameter_126: class Program_weight_tensor_parameter_127: name = "parameter_127" + original_name = "batch_norm2d_112.w_2" shape = [192] dtype = "float32" min_val = float("0.00433237") @@ -1408,6 +1536,7 @@ class Program_weight_tensor_parameter_127: class Program_weight_tensor_parameter_128: name = "parameter_128" + original_name = "batch_norm2d_112.w_1" shape = [192] dtype = "float32" min_val = float("-0.396741") @@ -1419,6 +1548,7 @@ class Program_weight_tensor_parameter_128: class Program_weight_tensor_parameter_129: name = "parameter_129" + original_name = "conv2d_116.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.0350237") @@ -1430,6 +1560,7 @@ class Program_weight_tensor_parameter_129: class Program_weight_tensor_parameter_130: name = "parameter_130" + original_name = "batch_norm2d_111.b_0" shape = [192] dtype = "float32" min_val = float("-0.552248") @@ -1441,6 +1572,7 @@ class Program_weight_tensor_parameter_130: class Program_weight_tensor_parameter_131: name = "parameter_131" + original_name = "batch_norm2d_111.w_0" shape = [192] dtype = "float32" min_val = float("0.541472") @@ -1452,6 +1584,7 @@ class Program_weight_tensor_parameter_131: class Program_weight_tensor_parameter_132: name = "parameter_132" + original_name = "batch_norm2d_111.w_2" shape = [192] dtype = "float32" min_val = float("0.00561378") @@ -1463,6 +1596,7 @@ class Program_weight_tensor_parameter_132: class Program_weight_tensor_parameter_133: name = "parameter_133" + original_name = "batch_norm2d_111.w_1" shape = [192] dtype = "float32" min_val = float("-0.181834") @@ -1474,6 +1608,7 @@ class Program_weight_tensor_parameter_133: class Program_weight_tensor_parameter_134: name = "parameter_134" + original_name = "conv2d_115.w_0" shape = [192, 192, 1, 1] dtype = "float32" min_val = float("-0.140077") @@ -1485,6 +1620,7 @@ class Program_weight_tensor_parameter_134: class Program_weight_tensor_parameter_135: name = "parameter_135" + original_name = "batch_norm2d_110.b_0" shape = [96] dtype = "float32" min_val = float("-0.457965") @@ -1496,6 +1632,7 @@ class Program_weight_tensor_parameter_135: class Program_weight_tensor_parameter_136: name = "parameter_136" + original_name = "batch_norm2d_110.w_0" shape = [96] dtype = "float32" min_val = float("0.762871") @@ -1507,6 +1644,7 @@ class Program_weight_tensor_parameter_136: class Program_weight_tensor_parameter_137: name = "parameter_137" + original_name = "batch_norm2d_110.w_2" shape = [96] dtype = "float32" min_val = float("0.00298878") @@ -1518,6 +1656,7 @@ class Program_weight_tensor_parameter_137: class Program_weight_tensor_parameter_138: name = "parameter_138" + original_name = "batch_norm2d_110.w_1" shape = [96] dtype = "float32" min_val = float("-0.0584852") @@ -1529,6 +1668,7 @@ class Program_weight_tensor_parameter_138: class Program_weight_tensor_parameter_139: name = "parameter_139" + original_name = "conv2d_114.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.0753814") @@ -1540,6 +1680,7 @@ class Program_weight_tensor_parameter_139: class Program_weight_tensor_parameter_140: name = "parameter_140" + original_name = "batch_norm2d_109.b_0" shape = [96] dtype = "float32" min_val = float("-0.457965") @@ -1551,6 +1692,7 @@ class Program_weight_tensor_parameter_140: class Program_weight_tensor_parameter_141: name = "parameter_141" + original_name = "batch_norm2d_109.w_0" shape = [96] dtype = "float32" min_val = float("0.505007") @@ -1562,6 +1704,7 @@ class Program_weight_tensor_parameter_141: class Program_weight_tensor_parameter_142: name = "parameter_142" + original_name = "batch_norm2d_109.w_2" shape = [96] dtype = "float32" min_val = float("0.00864011") @@ -1573,6 +1716,7 @@ class Program_weight_tensor_parameter_142: class Program_weight_tensor_parameter_143: name = "parameter_143" + original_name = "batch_norm2d_109.w_1" shape = [96] dtype = "float32" min_val = float("-0.235686") @@ -1584,6 +1728,7 @@ class Program_weight_tensor_parameter_143: class Program_weight_tensor_parameter_144: name = "parameter_144" + original_name = "conv2d_113.w_0" shape = [96, 96, 3, 3] dtype = "float32" min_val = float("-0.0934252") @@ -1595,6 +1740,7 @@ class Program_weight_tensor_parameter_144: class Program_weight_tensor_parameter_145: name = "parameter_145" + original_name = "batch_norm2d_108.b_0" shape = [96] dtype = "float32" min_val = float("-0.703685") @@ -1606,6 +1752,7 @@ class Program_weight_tensor_parameter_145: class Program_weight_tensor_parameter_146: name = "parameter_146" + original_name = "batch_norm2d_108.w_0" shape = [96] dtype = "float32" min_val = float("0.723217") @@ -1617,6 +1764,7 @@ class Program_weight_tensor_parameter_146: class Program_weight_tensor_parameter_147: name = "parameter_147" + original_name = "batch_norm2d_108.w_2" shape = [96] dtype = "float32" min_val = float("0.0126242") @@ -1628,6 +1776,7 @@ class Program_weight_tensor_parameter_147: class Program_weight_tensor_parameter_148: name = "parameter_148" + original_name = "batch_norm2d_108.w_1" shape = [96] dtype = "float32" min_val = float("-0.211007") @@ -1639,6 +1788,7 @@ class Program_weight_tensor_parameter_148: class Program_weight_tensor_parameter_149: name = "parameter_149" + original_name = "conv2d_112.w_0" shape = [96, 96, 3, 3] dtype = "float32" min_val = float("-0.0919405") @@ -1650,6 +1800,7 @@ class Program_weight_tensor_parameter_149: class Program_weight_tensor_parameter_150: name = "parameter_150" + original_name = "batch_norm2d_107.b_0" shape = [96] dtype = "float32" min_val = float("-0.364151") @@ -1661,6 +1812,7 @@ class Program_weight_tensor_parameter_150: class Program_weight_tensor_parameter_151: name = "parameter_151" + original_name = "batch_norm2d_107.w_0" shape = [96] dtype = "float32" min_val = float("0.626997") @@ -1672,6 +1824,7 @@ class Program_weight_tensor_parameter_151: class Program_weight_tensor_parameter_152: name = "parameter_152" + original_name = "batch_norm2d_107.w_2" shape = [96] dtype = "float32" min_val = float("0.00320113") @@ -1683,6 +1836,7 @@ class Program_weight_tensor_parameter_152: class Program_weight_tensor_parameter_153: name = "parameter_153" + original_name = "batch_norm2d_107.w_1" shape = [96] dtype = "float32" min_val = float("-0.0654375") @@ -1694,6 +1848,7 @@ class Program_weight_tensor_parameter_153: class Program_weight_tensor_parameter_154: name = "parameter_154" + original_name = "conv2d_111.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.0710343") @@ -1705,6 +1860,7 @@ class Program_weight_tensor_parameter_154: class Program_weight_tensor_parameter_155: name = "parameter_155" + original_name = "batch_norm2d_106.b_0" shape = [96] dtype = "float32" min_val = float("-0.364151") @@ -1716,6 +1872,7 @@ class Program_weight_tensor_parameter_155: class Program_weight_tensor_parameter_156: name = "parameter_156" + original_name = "batch_norm2d_106.w_0" shape = [96] dtype = "float32" min_val = float("0.811163") @@ -1727,6 +1884,7 @@ class Program_weight_tensor_parameter_156: class Program_weight_tensor_parameter_157: name = "parameter_157" + original_name = "batch_norm2d_106.w_2" shape = [96] dtype = "float32" min_val = float("0.0112523") @@ -1738,6 +1896,7 @@ class Program_weight_tensor_parameter_157: class Program_weight_tensor_parameter_158: name = "parameter_158" + original_name = "batch_norm2d_106.w_1" shape = [96] dtype = "float32" min_val = float("-0.163525") @@ -1749,6 +1908,7 @@ class Program_weight_tensor_parameter_158: class Program_weight_tensor_parameter_159: name = "parameter_159" + original_name = "conv2d_110.w_0" shape = [96, 96, 3, 3] dtype = "float32" min_val = float("-0.0811922") @@ -1760,6 +1920,7 @@ class Program_weight_tensor_parameter_159: class Program_weight_tensor_parameter_160: name = "parameter_160" + original_name = "batch_norm2d_105.b_0" shape = [96] dtype = "float32" min_val = float("-0.486488") @@ -1771,6 +1932,7 @@ class Program_weight_tensor_parameter_160: class Program_weight_tensor_parameter_161: name = "parameter_161" + original_name = "batch_norm2d_105.w_0" shape = [96] dtype = "float32" min_val = float("0.77745") @@ -1782,6 +1944,7 @@ class Program_weight_tensor_parameter_161: class Program_weight_tensor_parameter_162: name = "parameter_162" + original_name = "batch_norm2d_105.w_2" shape = [96] dtype = "float32" min_val = float("0.0097702") @@ -1793,6 +1956,7 @@ class Program_weight_tensor_parameter_162: class Program_weight_tensor_parameter_163: name = "parameter_163" + original_name = "batch_norm2d_105.w_1" shape = [96] dtype = "float32" min_val = float("-0.150372") @@ -1804,6 +1968,7 @@ class Program_weight_tensor_parameter_163: class Program_weight_tensor_parameter_164: name = "parameter_164" + original_name = "conv2d_109.w_0" shape = [96, 96, 3, 3] dtype = "float32" min_val = float("-0.0993825") @@ -1815,6 +1980,7 @@ class Program_weight_tensor_parameter_164: class Program_weight_tensor_parameter_165: name = "parameter_165" + original_name = "batch_norm2d_104.b_0" shape = [96] dtype = "float32" min_val = float("-0.489705") @@ -1826,6 +1992,7 @@ class Program_weight_tensor_parameter_165: class Program_weight_tensor_parameter_166: name = "parameter_166" + original_name = "batch_norm2d_104.w_0" shape = [96] dtype = "float32" min_val = float("0.722939") @@ -1837,6 +2004,7 @@ class Program_weight_tensor_parameter_166: class Program_weight_tensor_parameter_167: name = "parameter_167" + original_name = "batch_norm2d_104.w_2" shape = [96] dtype = "float32" min_val = float("0.00758991") @@ -1848,6 +2016,7 @@ class Program_weight_tensor_parameter_167: class Program_weight_tensor_parameter_168: name = "parameter_168" + original_name = "batch_norm2d_104.w_1" shape = [96] dtype = "float32" min_val = float("-0.0561772") @@ -1859,6 +2028,7 @@ class Program_weight_tensor_parameter_168: class Program_weight_tensor_parameter_169: name = "parameter_169" + original_name = "conv2d_108.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.103953") @@ -1870,6 +2040,7 @@ class Program_weight_tensor_parameter_169: class Program_weight_tensor_parameter_170: name = "parameter_170" + original_name = "batch_norm2d_103.b_0" shape = [96] dtype = "float32" min_val = float("-0.489705") @@ -1881,6 +2052,7 @@ class Program_weight_tensor_parameter_170: class Program_weight_tensor_parameter_171: name = "parameter_171" + original_name = "batch_norm2d_103.w_0" shape = [96] dtype = "float32" min_val = float("0.766539") @@ -1892,6 +2064,7 @@ class Program_weight_tensor_parameter_171: class Program_weight_tensor_parameter_172: name = "parameter_172" + original_name = "batch_norm2d_103.w_2" shape = [96] dtype = "float32" min_val = float("0.0171857") @@ -1903,6 +2076,7 @@ class Program_weight_tensor_parameter_172: class Program_weight_tensor_parameter_173: name = "parameter_173" + original_name = "batch_norm2d_103.w_1" shape = [96] dtype = "float32" min_val = float("-0.196226") @@ -1914,6 +2088,7 @@ class Program_weight_tensor_parameter_173: class Program_weight_tensor_parameter_174: name = "parameter_174" + original_name = "conv2d_107.w_0" shape = [96, 96, 3, 3] dtype = "float32" min_val = float("-0.0992682") @@ -1925,6 +2100,7 @@ class Program_weight_tensor_parameter_174: class Program_weight_tensor_parameter_175: name = "parameter_175" + original_name = "batch_norm2d_102.b_0" shape = [96] dtype = "float32" min_val = float("-0.564609") @@ -1936,6 +2112,7 @@ class Program_weight_tensor_parameter_175: class Program_weight_tensor_parameter_176: name = "parameter_176" + original_name = "batch_norm2d_102.w_0" shape = [96] dtype = "float32" min_val = float("0.764459") @@ -1947,6 +2124,7 @@ class Program_weight_tensor_parameter_176: class Program_weight_tensor_parameter_177: name = "parameter_177" + original_name = "batch_norm2d_102.w_2" shape = [96] dtype = "float32" min_val = float("0.0145544") @@ -1958,6 +2136,7 @@ class Program_weight_tensor_parameter_177: class Program_weight_tensor_parameter_178: name = "parameter_178" + original_name = "batch_norm2d_102.w_1" shape = [96] dtype = "float32" min_val = float("-0.17302") @@ -1969,6 +2148,7 @@ class Program_weight_tensor_parameter_178: class Program_weight_tensor_parameter_179: name = "parameter_179" + original_name = "conv2d_106.w_0" shape = [96, 96, 3, 3] dtype = "float32" min_val = float("-0.142383") @@ -1980,6 +2160,7 @@ class Program_weight_tensor_parameter_179: class Program_weight_tensor_parameter_180: name = "parameter_180" + original_name = "batch_norm2d_101.b_0" shape = [96] dtype = "float32" min_val = float("-0.625413") @@ -1991,6 +2172,7 @@ class Program_weight_tensor_parameter_180: class Program_weight_tensor_parameter_181: name = "parameter_181" + original_name = "batch_norm2d_101.w_0" shape = [96] dtype = "float32" min_val = float("0.647479") @@ -2002,6 +2184,7 @@ class Program_weight_tensor_parameter_181: class Program_weight_tensor_parameter_182: name = "parameter_182" + original_name = "batch_norm2d_101.w_2" shape = [96] dtype = "float32" min_val = float("0.0115736") @@ -2013,6 +2196,7 @@ class Program_weight_tensor_parameter_182: class Program_weight_tensor_parameter_183: name = "parameter_183" + original_name = "batch_norm2d_101.w_1" shape = [96] dtype = "float32" min_val = float("-0.112932") @@ -2024,6 +2208,7 @@ class Program_weight_tensor_parameter_183: class Program_weight_tensor_parameter_184: name = "parameter_184" + original_name = "conv2d_105.w_0" shape = [96, 448, 1, 1] dtype = "float32" min_val = float("-0.14903") @@ -2035,6 +2220,7 @@ class Program_weight_tensor_parameter_184: class Program_weight_tensor_parameter_185: name = "parameter_185" + original_name = "batch_norm2d_100.b_0" shape = [96] dtype = "float32" min_val = float("-0.0986349") @@ -2046,6 +2232,7 @@ class Program_weight_tensor_parameter_185: class Program_weight_tensor_parameter_186: name = "parameter_186" + original_name = "batch_norm2d_100.w_0" shape = [96] dtype = "float32" min_val = float("0.703928") @@ -2057,6 +2244,7 @@ class Program_weight_tensor_parameter_186: class Program_weight_tensor_parameter_187: name = "parameter_187" + original_name = "batch_norm2d_100.w_2" shape = [96] dtype = "float32" min_val = float("0.00519295") @@ -2068,6 +2256,7 @@ class Program_weight_tensor_parameter_187: class Program_weight_tensor_parameter_188: name = "parameter_188" + original_name = "batch_norm2d_100.w_1" shape = [96] dtype = "float32" min_val = float("-0.0889976") @@ -2079,6 +2268,7 @@ class Program_weight_tensor_parameter_188: class Program_weight_tensor_parameter_189: name = "parameter_189" + original_name = "conv2d_104.w_0" shape = [96, 448, 1, 1] dtype = "float32" min_val = float("-0.0952125") @@ -2090,6 +2280,7 @@ class Program_weight_tensor_parameter_189: class Program_weight_tensor_parameter_190: name = "parameter_190" + original_name = "batch_norm2d_99.b_0" shape = [192] dtype = "float32" min_val = float("-0.295367") @@ -2101,6 +2292,7 @@ class Program_weight_tensor_parameter_190: class Program_weight_tensor_parameter_191: name = "parameter_191" + original_name = "batch_norm2d_99.w_0" shape = [192] dtype = "float32" min_val = float("0.670697") @@ -2112,6 +2304,7 @@ class Program_weight_tensor_parameter_191: class Program_weight_tensor_parameter_192: name = "parameter_192" + original_name = "batch_norm2d_99.w_2" shape = [192] dtype = "float32" min_val = float("0.00810165") @@ -2123,6 +2316,7 @@ class Program_weight_tensor_parameter_192: class Program_weight_tensor_parameter_193: name = "parameter_193" + original_name = "batch_norm2d_99.w_1" shape = [192] dtype = "float32" min_val = float("-0.147256") @@ -2134,6 +2328,7 @@ class Program_weight_tensor_parameter_193: class Program_weight_tensor_parameter_194: name = "parameter_194" + original_name = "conv2d_103.w_0" shape = [192, 384, 1, 1] dtype = "float32" min_val = float("-0.0959126") @@ -2145,6 +2340,7 @@ class Program_weight_tensor_parameter_194: class Program_weight_tensor_parameter_195: name = "parameter_195" + original_name = "batch_norm2d_98.b_0" shape = [384] dtype = "float32" min_val = float("-0.201782") @@ -2156,6 +2352,7 @@ class Program_weight_tensor_parameter_195: class Program_weight_tensor_parameter_196: name = "parameter_196" + original_name = "batch_norm2d_98.w_0" shape = [384] dtype = "float32" min_val = float("0.873178") @@ -2167,6 +2364,7 @@ class Program_weight_tensor_parameter_196: class Program_weight_tensor_parameter_197: name = "parameter_197" + original_name = "batch_norm2d_98.w_2" shape = [384] dtype = "float32" min_val = float("0.00754976") @@ -2178,6 +2376,7 @@ class Program_weight_tensor_parameter_197: class Program_weight_tensor_parameter_198: name = "parameter_198" + original_name = "batch_norm2d_98.w_1" shape = [384] dtype = "float32" min_val = float("-0.335306") @@ -2189,6 +2388,7 @@ class Program_weight_tensor_parameter_198: class Program_weight_tensor_parameter_199: name = "parameter_199" + original_name = "conv2d_102.w_0" shape = [384, 384, 1, 1] dtype = "float32" min_val = float("-0.104641") @@ -2200,6 +2400,7 @@ class Program_weight_tensor_parameter_199: class Program_weight_tensor_parameter_200: name = "parameter_200" + original_name = "batch_norm2d_97.b_0" shape = [192] dtype = "float32" min_val = float("-0.176949") @@ -2211,6 +2412,7 @@ class Program_weight_tensor_parameter_200: class Program_weight_tensor_parameter_201: name = "parameter_201" + original_name = "batch_norm2d_97.w_0" shape = [192] dtype = "float32" min_val = float("0.884903") @@ -2222,6 +2424,7 @@ class Program_weight_tensor_parameter_201: class Program_weight_tensor_parameter_202: name = "parameter_202" + original_name = "batch_norm2d_97.w_2" shape = [192] dtype = "float32" min_val = float("0.00345864") @@ -2233,6 +2436,7 @@ class Program_weight_tensor_parameter_202: class Program_weight_tensor_parameter_203: name = "parameter_203" + original_name = "batch_norm2d_97.w_1" shape = [192] dtype = "float32" min_val = float("-0.0784978") @@ -2244,6 +2448,7 @@ class Program_weight_tensor_parameter_203: class Program_weight_tensor_parameter_204: name = "parameter_204" + original_name = "conv2d_101.w_0" shape = [192, 192, 1, 1] dtype = "float32" min_val = float("-0.0569075") @@ -2255,6 +2460,7 @@ class Program_weight_tensor_parameter_204: class Program_weight_tensor_parameter_205: name = "parameter_205" + original_name = "batch_norm2d_96.b_0" shape = [192] dtype = "float32" min_val = float("-0.176949") @@ -2266,6 +2472,7 @@ class Program_weight_tensor_parameter_205: class Program_weight_tensor_parameter_206: name = "parameter_206" + original_name = "batch_norm2d_96.w_0" shape = [192] dtype = "float32" min_val = float("0.945936") @@ -2277,6 +2484,7 @@ class Program_weight_tensor_parameter_206: class Program_weight_tensor_parameter_207: name = "parameter_207" + original_name = "batch_norm2d_96.w_2" shape = [192] dtype = "float32" min_val = float("0.0155344") @@ -2288,6 +2496,7 @@ class Program_weight_tensor_parameter_207: class Program_weight_tensor_parameter_208: name = "parameter_208" + original_name = "batch_norm2d_96.w_1" shape = [192] dtype = "float32" min_val = float("-0.176762") @@ -2299,6 +2508,7 @@ class Program_weight_tensor_parameter_208: class Program_weight_tensor_parameter_209: name = "parameter_209" + original_name = "conv2d_100.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.0444131") @@ -2310,6 +2520,7 @@ class Program_weight_tensor_parameter_209: class Program_weight_tensor_parameter_210: name = "parameter_210" + original_name = "batch_norm2d_95.b_0" shape = [192] dtype = "float32" min_val = float("-0.217095") @@ -2321,6 +2532,7 @@ class Program_weight_tensor_parameter_210: class Program_weight_tensor_parameter_211: name = "parameter_211" + original_name = "batch_norm2d_95.w_0" shape = [192] dtype = "float32" min_val = float("0.939031") @@ -2332,6 +2544,7 @@ class Program_weight_tensor_parameter_211: class Program_weight_tensor_parameter_212: name = "parameter_212" + original_name = "batch_norm2d_95.w_2" shape = [192] dtype = "float32" min_val = float("0.0364266") @@ -2343,6 +2556,7 @@ class Program_weight_tensor_parameter_212: class Program_weight_tensor_parameter_213: name = "parameter_213" + original_name = "batch_norm2d_95.w_1" shape = [192] dtype = "float32" min_val = float("-0.262217") @@ -2354,6 +2568,7 @@ class Program_weight_tensor_parameter_213: class Program_weight_tensor_parameter_214: name = "parameter_214" + original_name = "conv2d_99.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.0622017") @@ -2365,6 +2580,7 @@ class Program_weight_tensor_parameter_214: class Program_weight_tensor_parameter_215: name = "parameter_215" + original_name = "batch_norm2d_94.b_0" shape = [192] dtype = "float32" min_val = float("-0.196617") @@ -2376,6 +2592,7 @@ class Program_weight_tensor_parameter_215: class Program_weight_tensor_parameter_216: name = "parameter_216" + original_name = "batch_norm2d_94.w_0" shape = [192] dtype = "float32" min_val = float("0.94411") @@ -2387,6 +2604,7 @@ class Program_weight_tensor_parameter_216: class Program_weight_tensor_parameter_217: name = "parameter_217" + original_name = "batch_norm2d_94.w_2" shape = [192] dtype = "float32" min_val = float("0.00228676") @@ -2398,6 +2616,7 @@ class Program_weight_tensor_parameter_217: class Program_weight_tensor_parameter_218: name = "parameter_218" + original_name = "batch_norm2d_94.w_1" shape = [192] dtype = "float32" min_val = float("-0.0953901") @@ -2409,6 +2628,7 @@ class Program_weight_tensor_parameter_218: class Program_weight_tensor_parameter_219: name = "parameter_219" + original_name = "conv2d_98.w_0" shape = [192, 192, 1, 1] dtype = "float32" min_val = float("-0.0313923") @@ -2420,6 +2640,7 @@ class Program_weight_tensor_parameter_219: class Program_weight_tensor_parameter_220: name = "parameter_220" + original_name = "batch_norm2d_93.b_0" shape = [192] dtype = "float32" min_val = float("-0.196617") @@ -2431,6 +2652,7 @@ class Program_weight_tensor_parameter_220: class Program_weight_tensor_parameter_221: name = "parameter_221" + original_name = "batch_norm2d_93.w_0" shape = [192] dtype = "float32" min_val = float("0.953711") @@ -2442,6 +2664,7 @@ class Program_weight_tensor_parameter_221: class Program_weight_tensor_parameter_222: name = "parameter_222" + original_name = "batch_norm2d_93.w_2" shape = [192] dtype = "float32" min_val = float("0.0101684") @@ -2453,6 +2676,7 @@ class Program_weight_tensor_parameter_222: class Program_weight_tensor_parameter_223: name = "parameter_223" + original_name = "batch_norm2d_93.w_1" shape = [192] dtype = "float32" min_val = float("-0.187973") @@ -2464,6 +2688,7 @@ class Program_weight_tensor_parameter_223: class Program_weight_tensor_parameter_224: name = "parameter_224" + original_name = "conv2d_97.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.0484376") @@ -2475,6 +2700,7 @@ class Program_weight_tensor_parameter_224: class Program_weight_tensor_parameter_225: name = "parameter_225" + original_name = "batch_norm2d_92.b_0" shape = [192] dtype = "float32" min_val = float("-0.232846") @@ -2486,6 +2712,7 @@ class Program_weight_tensor_parameter_225: class Program_weight_tensor_parameter_226: name = "parameter_226" + original_name = "batch_norm2d_92.w_0" shape = [192] dtype = "float32" min_val = float("0.946521") @@ -2497,6 +2724,7 @@ class Program_weight_tensor_parameter_226: class Program_weight_tensor_parameter_227: name = "parameter_227" + original_name = "batch_norm2d_92.w_2" shape = [192] dtype = "float32" min_val = float("0.0361899") @@ -2508,6 +2736,7 @@ class Program_weight_tensor_parameter_227: class Program_weight_tensor_parameter_228: name = "parameter_228" + original_name = "batch_norm2d_92.w_1" shape = [192] dtype = "float32" min_val = float("-0.350006") @@ -2519,6 +2748,7 @@ class Program_weight_tensor_parameter_228: class Program_weight_tensor_parameter_229: name = "parameter_229" + original_name = "conv2d_96.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.0611519") @@ -2530,6 +2760,7 @@ class Program_weight_tensor_parameter_229: class Program_weight_tensor_parameter_230: name = "parameter_230" + original_name = "batch_norm2d_91.b_0" shape = [192] dtype = "float32" min_val = float("-0.154886") @@ -2541,6 +2772,7 @@ class Program_weight_tensor_parameter_230: class Program_weight_tensor_parameter_231: name = "parameter_231" + original_name = "batch_norm2d_91.w_0" shape = [192] dtype = "float32" min_val = float("0.932342") @@ -2552,6 +2784,7 @@ class Program_weight_tensor_parameter_231: class Program_weight_tensor_parameter_232: name = "parameter_232" + original_name = "batch_norm2d_91.w_2" shape = [192] dtype = "float32" min_val = float("0.0020288") @@ -2563,6 +2796,7 @@ class Program_weight_tensor_parameter_232: class Program_weight_tensor_parameter_233: name = "parameter_233" + original_name = "batch_norm2d_91.w_1" shape = [192] dtype = "float32" min_val = float("-0.0826953") @@ -2574,6 +2808,7 @@ class Program_weight_tensor_parameter_233: class Program_weight_tensor_parameter_234: name = "parameter_234" + original_name = "conv2d_95.w_0" shape = [192, 192, 1, 1] dtype = "float32" min_val = float("-0.0348635") @@ -2585,6 +2820,7 @@ class Program_weight_tensor_parameter_234: class Program_weight_tensor_parameter_235: name = "parameter_235" + original_name = "batch_norm2d_90.b_0" shape = [192] dtype = "float32" min_val = float("-0.154886") @@ -2596,6 +2832,7 @@ class Program_weight_tensor_parameter_235: class Program_weight_tensor_parameter_236: name = "parameter_236" + original_name = "batch_norm2d_90.w_0" shape = [192] dtype = "float32" min_val = float("0.936172") @@ -2607,6 +2844,7 @@ class Program_weight_tensor_parameter_236: class Program_weight_tensor_parameter_237: name = "parameter_237" + original_name = "batch_norm2d_90.w_2" shape = [192] dtype = "float32" min_val = float("0.0092625") @@ -2618,6 +2856,7 @@ class Program_weight_tensor_parameter_237: class Program_weight_tensor_parameter_238: name = "parameter_238" + original_name = "batch_norm2d_90.w_1" shape = [192] dtype = "float32" min_val = float("-0.280981") @@ -2629,6 +2868,7 @@ class Program_weight_tensor_parameter_238: class Program_weight_tensor_parameter_239: name = "parameter_239" + original_name = "conv2d_94.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.0372398") @@ -2640,6 +2880,7 @@ class Program_weight_tensor_parameter_239: class Program_weight_tensor_parameter_240: name = "parameter_240" + original_name = "batch_norm2d_89.b_0" shape = [192] dtype = "float32" min_val = float("-0.289028") @@ -2651,6 +2892,7 @@ class Program_weight_tensor_parameter_240: class Program_weight_tensor_parameter_241: name = "parameter_241" + original_name = "batch_norm2d_89.w_0" shape = [192] dtype = "float32" min_val = float("0.943873") @@ -2662,6 +2904,7 @@ class Program_weight_tensor_parameter_241: class Program_weight_tensor_parameter_242: name = "parameter_242" + original_name = "batch_norm2d_89.w_2" shape = [192] dtype = "float32" min_val = float("0.0146559") @@ -2673,6 +2916,7 @@ class Program_weight_tensor_parameter_242: class Program_weight_tensor_parameter_243: name = "parameter_243" + original_name = "batch_norm2d_89.w_1" shape = [192] dtype = "float32" min_val = float("-0.381607") @@ -2684,6 +2928,7 @@ class Program_weight_tensor_parameter_243: class Program_weight_tensor_parameter_244: name = "parameter_244" + original_name = "conv2d_93.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.0566363") @@ -2695,6 +2940,7 @@ class Program_weight_tensor_parameter_244: class Program_weight_tensor_parameter_245: name = "parameter_245" + original_name = "batch_norm2d_88.b_0" shape = [192] dtype = "float32" min_val = float("-0.257034") @@ -2706,6 +2952,7 @@ class Program_weight_tensor_parameter_245: class Program_weight_tensor_parameter_246: name = "parameter_246" + original_name = "batch_norm2d_88.w_0" shape = [192] dtype = "float32" min_val = float("0.916939") @@ -2717,6 +2964,7 @@ class Program_weight_tensor_parameter_246: class Program_weight_tensor_parameter_247: name = "parameter_247" + original_name = "batch_norm2d_88.w_2" shape = [192] dtype = "float32" min_val = float("0.00558617") @@ -2728,6 +2976,7 @@ class Program_weight_tensor_parameter_247: class Program_weight_tensor_parameter_248: name = "parameter_248" + original_name = "batch_norm2d_88.w_1" shape = [192] dtype = "float32" min_val = float("-0.121693") @@ -2739,6 +2988,7 @@ class Program_weight_tensor_parameter_248: class Program_weight_tensor_parameter_249: name = "parameter_249" + original_name = "conv2d_92.w_0" shape = [192, 896, 1, 1] dtype = "float32" min_val = float("-0.0812553") @@ -2750,6 +3000,7 @@ class Program_weight_tensor_parameter_249: class Program_weight_tensor_parameter_250: name = "parameter_250" + original_name = "batch_norm2d_87.b_0" shape = [192] dtype = "float32" min_val = float("-0.176608") @@ -2761,6 +3012,7 @@ class Program_weight_tensor_parameter_250: class Program_weight_tensor_parameter_251: name = "parameter_251" + original_name = "batch_norm2d_87.w_0" shape = [192] dtype = "float32" min_val = float("0.951166") @@ -2772,6 +3024,7 @@ class Program_weight_tensor_parameter_251: class Program_weight_tensor_parameter_252: name = "parameter_252" + original_name = "batch_norm2d_87.w_2" shape = [192] dtype = "float32" min_val = float("0.00700942") @@ -2783,6 +3036,7 @@ class Program_weight_tensor_parameter_252: class Program_weight_tensor_parameter_253: name = "parameter_253" + original_name = "batch_norm2d_87.w_1" shape = [192] dtype = "float32" min_val = float("-0.0747975") @@ -2794,6 +3048,7 @@ class Program_weight_tensor_parameter_253: class Program_weight_tensor_parameter_254: name = "parameter_254" + original_name = "conv2d_91.w_0" shape = [192, 896, 1, 1] dtype = "float32" min_val = float("-0.055207") @@ -2805,6 +3060,7 @@ class Program_weight_tensor_parameter_254: class Program_weight_tensor_parameter_255: name = "parameter_255" + original_name = "batch_norm2d_86.b_0" shape = [384] dtype = "float32" min_val = float("-0.249775") @@ -2816,6 +3072,7 @@ class Program_weight_tensor_parameter_255: class Program_weight_tensor_parameter_256: name = "parameter_256" + original_name = "batch_norm2d_86.w_0" shape = [384] dtype = "float32" min_val = float("0.814907") @@ -2827,6 +3084,7 @@ class Program_weight_tensor_parameter_256: class Program_weight_tensor_parameter_257: name = "parameter_257" + original_name = "batch_norm2d_86.w_2" shape = [384] dtype = "float32" min_val = float("0.0100937") @@ -2838,6 +3096,7 @@ class Program_weight_tensor_parameter_257: class Program_weight_tensor_parameter_258: name = "parameter_258" + original_name = "batch_norm2d_86.w_1" shape = [384] dtype = "float32" min_val = float("-0.146181") @@ -2849,6 +3108,7 @@ class Program_weight_tensor_parameter_258: class Program_weight_tensor_parameter_259: name = "parameter_259" + original_name = "conv2d_90.w_0" shape = [384, 768, 1, 1] dtype = "float32" min_val = float("-0.0364868") @@ -2860,6 +3120,7 @@ class Program_weight_tensor_parameter_259: class Program_weight_tensor_parameter_260: name = "parameter_260" + original_name = "batch_norm2d_85.b_0" shape = [768] dtype = "float32" min_val = float("-0.104276") @@ -2871,6 +3132,7 @@ class Program_weight_tensor_parameter_260: class Program_weight_tensor_parameter_261: name = "parameter_261" + original_name = "batch_norm2d_85.w_0" shape = [768] dtype = "float32" min_val = float("0.9523") @@ -2882,6 +3144,7 @@ class Program_weight_tensor_parameter_261: class Program_weight_tensor_parameter_262: name = "parameter_262" + original_name = "batch_norm2d_85.w_2" shape = [768] dtype = "float32" min_val = float("0.00433515") @@ -2893,6 +3156,7 @@ class Program_weight_tensor_parameter_262: class Program_weight_tensor_parameter_263: name = "parameter_263" + original_name = "batch_norm2d_85.w_1" shape = [768] dtype = "float32" min_val = float("-0.103914") @@ -2904,6 +3168,7 @@ class Program_weight_tensor_parameter_263: class Program_weight_tensor_parameter_264: name = "parameter_264" + original_name = "conv2d_89.w_0" shape = [768, 768, 1, 1] dtype = "float32" min_val = float("-0.0581812") @@ -2915,6 +3180,7 @@ class Program_weight_tensor_parameter_264: class Program_weight_tensor_parameter_265: name = "parameter_265" + original_name = "batch_norm2d_84.b_0" shape = [384] dtype = "float32" min_val = float("-0.158166") @@ -2926,6 +3192,7 @@ class Program_weight_tensor_parameter_265: class Program_weight_tensor_parameter_266: name = "parameter_266" + original_name = "batch_norm2d_84.w_0" shape = [384] dtype = "float32" min_val = float("0.888577") @@ -2937,6 +3204,7 @@ class Program_weight_tensor_parameter_266: class Program_weight_tensor_parameter_267: name = "parameter_267" + original_name = "batch_norm2d_84.w_2" shape = [384] dtype = "float32" min_val = float("0.00600496") @@ -2948,6 +3216,7 @@ class Program_weight_tensor_parameter_267: class Program_weight_tensor_parameter_268: name = "parameter_268" + original_name = "batch_norm2d_84.w_1" shape = [384] dtype = "float32" min_val = float("-0.0681594") @@ -2959,6 +3228,7 @@ class Program_weight_tensor_parameter_268: class Program_weight_tensor_parameter_269: name = "parameter_269" + original_name = "conv2d_88.w_0" shape = [384, 384, 1, 1] dtype = "float32" min_val = float("-0.0396804") @@ -2970,6 +3240,7 @@ class Program_weight_tensor_parameter_269: class Program_weight_tensor_parameter_270: name = "parameter_270" + original_name = "batch_norm2d_83.b_0" shape = [384] dtype = "float32" min_val = float("-0.158166") @@ -2981,6 +3252,7 @@ class Program_weight_tensor_parameter_270: class Program_weight_tensor_parameter_271: name = "parameter_271" + original_name = "batch_norm2d_83.w_0" shape = [384] dtype = "float32" min_val = float("0.879914") @@ -2992,6 +3264,7 @@ class Program_weight_tensor_parameter_271: class Program_weight_tensor_parameter_272: name = "parameter_272" + original_name = "batch_norm2d_83.w_2" shape = [384] dtype = "float32" min_val = float("0.0282422") @@ -3003,6 +3276,7 @@ class Program_weight_tensor_parameter_272: class Program_weight_tensor_parameter_273: name = "parameter_273" + original_name = "batch_norm2d_83.w_1" shape = [384] dtype = "float32" min_val = float("-0.276936") @@ -3014,6 +3288,7 @@ class Program_weight_tensor_parameter_273: class Program_weight_tensor_parameter_274: name = "parameter_274" + original_name = "conv2d_87.w_0" shape = [384, 384, 3, 3] dtype = "float32" min_val = float("-0.0424878") @@ -3025,6 +3300,7 @@ class Program_weight_tensor_parameter_274: class Program_weight_tensor_parameter_275: name = "parameter_275" + original_name = "batch_norm2d_82.b_0" shape = [384] dtype = "float32" min_val = float("-0.0801146") @@ -3036,6 +3312,7 @@ class Program_weight_tensor_parameter_275: class Program_weight_tensor_parameter_276: name = "parameter_276" + original_name = "batch_norm2d_82.w_0" shape = [384] dtype = "float32" min_val = float("0.920205") @@ -3047,6 +3324,7 @@ class Program_weight_tensor_parameter_276: class Program_weight_tensor_parameter_277: name = "parameter_277" + original_name = "batch_norm2d_82.w_2" shape = [384] dtype = "float32" min_val = float("0.0223031") @@ -3058,6 +3336,7 @@ class Program_weight_tensor_parameter_277: class Program_weight_tensor_parameter_278: name = "parameter_278" + original_name = "batch_norm2d_82.w_1" shape = [384] dtype = "float32" min_val = float("-0.235368") @@ -3069,6 +3348,7 @@ class Program_weight_tensor_parameter_278: class Program_weight_tensor_parameter_279: name = "parameter_279" + original_name = "conv2d_86.w_0" shape = [384, 384, 3, 3] dtype = "float32" min_val = float("-0.0274578") @@ -3080,6 +3360,7 @@ class Program_weight_tensor_parameter_279: class Program_weight_tensor_parameter_280: name = "parameter_280" + original_name = "batch_norm2d_81.b_0" shape = [384] dtype = "float32" min_val = float("-0.0739505") @@ -3091,6 +3372,7 @@ class Program_weight_tensor_parameter_280: class Program_weight_tensor_parameter_281: name = "parameter_281" + original_name = "batch_norm2d_81.w_0" shape = [384] dtype = "float32" min_val = float("0.946312") @@ -3102,6 +3384,7 @@ class Program_weight_tensor_parameter_281: class Program_weight_tensor_parameter_282: name = "parameter_282" + original_name = "batch_norm2d_81.w_2" shape = [384] dtype = "float32" min_val = float("0.0666339") @@ -3113,6 +3396,7 @@ class Program_weight_tensor_parameter_282: class Program_weight_tensor_parameter_283: name = "parameter_283" + original_name = "batch_norm2d_81.w_1" shape = [384] dtype = "float32" min_val = float("-1.5811") @@ -3124,6 +3408,7 @@ class Program_weight_tensor_parameter_283: class Program_weight_tensor_parameter_284: name = "parameter_284" + original_name = "conv2d_85.w_0" shape = [384, 1536, 1, 1] dtype = "float32" min_val = float("-0.0467316") @@ -3135,6 +3420,7 @@ class Program_weight_tensor_parameter_284: class Program_weight_tensor_parameter_285: name = "parameter_285" + original_name = "batch_norm2d_80.b_0" shape = [384] dtype = "float32" min_val = float("-0.0183804") @@ -3146,6 +3432,7 @@ class Program_weight_tensor_parameter_285: class Program_weight_tensor_parameter_286: name = "parameter_286" + original_name = "batch_norm2d_80.w_0" shape = [384] dtype = "float32" min_val = float("0.969538") @@ -3157,6 +3444,7 @@ class Program_weight_tensor_parameter_286: class Program_weight_tensor_parameter_287: name = "parameter_287" + original_name = "batch_norm2d_80.w_2" shape = [384] dtype = "float32" min_val = float("0.00292493") @@ -3168,6 +3456,7 @@ class Program_weight_tensor_parameter_287: class Program_weight_tensor_parameter_288: name = "parameter_288" + original_name = "batch_norm2d_80.w_1" shape = [384] dtype = "float32" min_val = float("-0.0970828") @@ -3179,6 +3468,7 @@ class Program_weight_tensor_parameter_288: class Program_weight_tensor_parameter_289: name = "parameter_289" + original_name = "conv2d_84.w_0" shape = [384, 384, 1, 1] dtype = "float32" min_val = float("-0.0333324") @@ -3190,6 +3480,7 @@ class Program_weight_tensor_parameter_289: class Program_weight_tensor_parameter_290: name = "parameter_290" + original_name = "batch_norm2d_79.b_0" shape = [384] dtype = "float32" min_val = float("-0.0183804") @@ -3201,6 +3492,7 @@ class Program_weight_tensor_parameter_290: class Program_weight_tensor_parameter_291: name = "parameter_291" + original_name = "batch_norm2d_79.w_0" shape = [384] dtype = "float32" min_val = float("0.972046") @@ -3212,6 +3504,7 @@ class Program_weight_tensor_parameter_291: class Program_weight_tensor_parameter_292: name = "parameter_292" + original_name = "batch_norm2d_79.w_2" shape = [384] dtype = "float32" min_val = float("0.0169031") @@ -3223,6 +3516,7 @@ class Program_weight_tensor_parameter_292: class Program_weight_tensor_parameter_293: name = "parameter_293" + original_name = "batch_norm2d_79.w_1" shape = [384] dtype = "float32" min_val = float("-0.322591") @@ -3234,6 +3528,7 @@ class Program_weight_tensor_parameter_293: class Program_weight_tensor_parameter_294: name = "parameter_294" + original_name = "conv2d_83.w_0" shape = [384, 384, 3, 3] dtype = "float32" min_val = float("-0.0285775") @@ -3245,6 +3540,7 @@ class Program_weight_tensor_parameter_294: class Program_weight_tensor_parameter_295: name = "parameter_295" + original_name = "batch_norm2d_78.b_0" shape = [384] dtype = "float32" min_val = float("-0.0498105") @@ -3256,6 +3552,7 @@ class Program_weight_tensor_parameter_295: class Program_weight_tensor_parameter_296: name = "parameter_296" + original_name = "batch_norm2d_78.w_0" shape = [384] dtype = "float32" min_val = float("0.953878") @@ -3267,6 +3564,7 @@ class Program_weight_tensor_parameter_296: class Program_weight_tensor_parameter_297: name = "parameter_297" + original_name = "batch_norm2d_78.w_2" shape = [384] dtype = "float32" min_val = float("0.072012") @@ -3278,6 +3576,7 @@ class Program_weight_tensor_parameter_297: class Program_weight_tensor_parameter_298: name = "parameter_298" + original_name = "batch_norm2d_78.w_1" shape = [384] dtype = "float32" min_val = float("-1.24821") @@ -3289,6 +3588,7 @@ class Program_weight_tensor_parameter_298: class Program_weight_tensor_parameter_299: name = "parameter_299" + original_name = "conv2d_82.w_0" shape = [384, 384, 3, 3] dtype = "float32" min_val = float("-0.024251") @@ -3300,6 +3600,7 @@ class Program_weight_tensor_parameter_299: class Program_weight_tensor_parameter_300: name = "parameter_300" + original_name = "batch_norm2d_77.b_0" shape = [384] dtype = "float32" min_val = float("-0.0360838") @@ -3311,6 +3612,7 @@ class Program_weight_tensor_parameter_300: class Program_weight_tensor_parameter_301: name = "parameter_301" + original_name = "batch_norm2d_77.w_0" shape = [384] dtype = "float32" min_val = float("0.984179") @@ -3322,6 +3624,7 @@ class Program_weight_tensor_parameter_301: class Program_weight_tensor_parameter_302: name = "parameter_302" + original_name = "batch_norm2d_77.w_2" shape = [384] dtype = "float32" min_val = float("0.00227712") @@ -3333,6 +3636,7 @@ class Program_weight_tensor_parameter_302: class Program_weight_tensor_parameter_303: name = "parameter_303" + original_name = "batch_norm2d_77.w_1" shape = [384] dtype = "float32" min_val = float("-0.0776737") @@ -3344,6 +3648,7 @@ class Program_weight_tensor_parameter_303: class Program_weight_tensor_parameter_304: name = "parameter_304" + original_name = "conv2d_81.w_0" shape = [384, 384, 1, 1] dtype = "float32" min_val = float("-0.0209147") @@ -3355,6 +3660,7 @@ class Program_weight_tensor_parameter_304: class Program_weight_tensor_parameter_305: name = "parameter_305" + original_name = "batch_norm2d_76.b_0" shape = [384] dtype = "float32" min_val = float("-0.0360838") @@ -3366,6 +3672,7 @@ class Program_weight_tensor_parameter_305: class Program_weight_tensor_parameter_306: name = "parameter_306" + original_name = "batch_norm2d_76.w_0" shape = [384] dtype = "float32" min_val = float("0.982136") @@ -3377,6 +3684,7 @@ class Program_weight_tensor_parameter_306: class Program_weight_tensor_parameter_307: name = "parameter_307" + original_name = "batch_norm2d_76.w_2" shape = [384] dtype = "float32" min_val = float("0.00988707") @@ -3388,6 +3696,7 @@ class Program_weight_tensor_parameter_307: class Program_weight_tensor_parameter_308: name = "parameter_308" + original_name = "batch_norm2d_76.w_1" shape = [384] dtype = "float32" min_val = float("-0.234406") @@ -3399,6 +3708,7 @@ class Program_weight_tensor_parameter_308: class Program_weight_tensor_parameter_309: name = "parameter_309" + original_name = "conv2d_80.w_0" shape = [384, 384, 3, 3] dtype = "float32" min_val = float("-0.0111228") @@ -3410,6 +3720,7 @@ class Program_weight_tensor_parameter_309: class Program_weight_tensor_parameter_310: name = "parameter_310" + original_name = "batch_norm2d_75.b_0" shape = [384] dtype = "float32" min_val = float("-0.0529908") @@ -3421,6 +3732,7 @@ class Program_weight_tensor_parameter_310: class Program_weight_tensor_parameter_311: name = "parameter_311" + original_name = "batch_norm2d_75.w_0" shape = [384] dtype = "float32" min_val = float("0.976061") @@ -3432,6 +3744,7 @@ class Program_weight_tensor_parameter_311: class Program_weight_tensor_parameter_312: name = "parameter_312" + original_name = "batch_norm2d_75.w_2" shape = [384] dtype = "float32" min_val = float("0.0127448") @@ -3443,6 +3756,7 @@ class Program_weight_tensor_parameter_312: class Program_weight_tensor_parameter_313: name = "parameter_313" + original_name = "batch_norm2d_75.w_1" shape = [384] dtype = "float32" min_val = float("-0.182538") @@ -3454,6 +3768,7 @@ class Program_weight_tensor_parameter_313: class Program_weight_tensor_parameter_314: name = "parameter_314" + original_name = "conv2d_79.w_0" shape = [384, 384, 3, 3] dtype = "float32" min_val = float("-0.0155426") @@ -3465,6 +3780,7 @@ class Program_weight_tensor_parameter_314: class Program_weight_tensor_parameter_315: name = "parameter_315" + original_name = "batch_norm2d_74.b_0" shape = [384] dtype = "float32" min_val = float("-0.0699578") @@ -3476,6 +3792,7 @@ class Program_weight_tensor_parameter_315: class Program_weight_tensor_parameter_316: name = "parameter_316" + original_name = "batch_norm2d_74.w_0" shape = [384] dtype = "float32" min_val = float("0.981937") @@ -3487,6 +3804,7 @@ class Program_weight_tensor_parameter_316: class Program_weight_tensor_parameter_317: name = "parameter_317" + original_name = "batch_norm2d_74.w_2" shape = [384] dtype = "float32" min_val = float("0.00885101") @@ -3498,6 +3816,7 @@ class Program_weight_tensor_parameter_317: class Program_weight_tensor_parameter_318: name = "parameter_318" + original_name = "batch_norm2d_74.w_1" shape = [384] dtype = "float32" min_val = float("-0.11759") @@ -3509,6 +3828,7 @@ class Program_weight_tensor_parameter_318: class Program_weight_tensor_parameter_319: name = "parameter_319" + original_name = "conv2d_78.w_0" shape = [384, 1024, 1, 1] dtype = "float32" min_val = float("-0.0187213") @@ -3520,6 +3840,7 @@ class Program_weight_tensor_parameter_319: class Program_weight_tensor_parameter_320: name = "parameter_320" + original_name = "batch_norm2d_73.b_0" shape = [384] dtype = "float32" min_val = float("-0.024099") @@ -3531,6 +3852,7 @@ class Program_weight_tensor_parameter_320: class Program_weight_tensor_parameter_321: name = "parameter_321" + original_name = "batch_norm2d_73.w_0" shape = [384] dtype = "float32" min_val = float("0.994048") @@ -3542,6 +3864,7 @@ class Program_weight_tensor_parameter_321: class Program_weight_tensor_parameter_322: name = "parameter_322" + original_name = "batch_norm2d_73.w_2" shape = [384] dtype = "float32" min_val = float("0.0117538") @@ -3553,6 +3876,7 @@ class Program_weight_tensor_parameter_322: class Program_weight_tensor_parameter_323: name = "parameter_323" + original_name = "batch_norm2d_73.w_1" shape = [384] dtype = "float32" min_val = float("-0.154192") @@ -3564,6 +3888,7 @@ class Program_weight_tensor_parameter_323: class Program_weight_tensor_parameter_324: name = "parameter_324" + original_name = "conv2d_77.w_0" shape = [384, 1024, 1, 1] dtype = "float32" min_val = float("-0.0381973") @@ -3575,6 +3900,7 @@ class Program_weight_tensor_parameter_324: class Program_weight_tensor_parameter_325: name = "parameter_325" + original_name = "layer_norm_1.b_0_deepcopy_47" shape = [1024] dtype = "float32" min_val = float("-3.19596e-10") @@ -3586,6 +3912,7 @@ class Program_weight_tensor_parameter_325: class Program_weight_tensor_parameter_326: name = "parameter_326" + original_name = "layer_norm_1.w_0_deepcopy_46" shape = [1024] dtype = "float32" min_val = float("0.826159") @@ -3597,6 +3924,7 @@ class Program_weight_tensor_parameter_326: class Program_weight_tensor_parameter_327: name = "parameter_327" + original_name = "linear_2.b_0_deepcopy_43" shape = [1024] dtype = "float32" min_val = float("-0.0184725") @@ -3608,6 +3936,7 @@ class Program_weight_tensor_parameter_327: class Program_weight_tensor_parameter_328: name = "parameter_328" + original_name = "linear_2.w_0_deepcopy_42" shape = [2048, 1024] dtype = "float32" min_val = float("-0.0186694") @@ -3619,6 +3948,7 @@ class Program_weight_tensor_parameter_328: class Program_weight_tensor_parameter_329: name = "parameter_329" + original_name = "linear_1.b_0_deepcopy_41" shape = [2048] dtype = "float32" min_val = float("-0.0258373") @@ -3630,6 +3960,7 @@ class Program_weight_tensor_parameter_329: class Program_weight_tensor_parameter_330: name = "parameter_330" + original_name = "linear_1.w_0_deepcopy_40" shape = [1024, 2048] dtype = "float32" min_val = float("-0.0261231") @@ -3641,6 +3972,7 @@ class Program_weight_tensor_parameter_330: class Program_weight_tensor_parameter_331: name = "parameter_331" + original_name = "layer_norm_0.b_0_deepcopy_45" shape = [1024] dtype = "float32" min_val = float("-0.000644078") @@ -3652,6 +3984,7 @@ class Program_weight_tensor_parameter_331: class Program_weight_tensor_parameter_332: name = "parameter_332" + original_name = "layer_norm_0.w_0_deepcopy_44" shape = [1024] dtype = "float32" min_val = float("0.825075") @@ -3663,6 +3996,7 @@ class Program_weight_tensor_parameter_332: class Program_weight_tensor_parameter_333: name = "parameter_333" + original_name = "linear_0.b_0_deepcopy_39" shape = [1024] dtype = "float32" min_val = float("-0.000571568") @@ -3674,6 +4008,7 @@ class Program_weight_tensor_parameter_333: class Program_weight_tensor_parameter_334: name = "parameter_334" + original_name = "linear_0.w_0_deepcopy_38" shape = [1024, 1024] dtype = "float32" min_val = float("-0.0452304") @@ -3685,6 +4020,7 @@ class Program_weight_tensor_parameter_334: class Program_weight_tensor_parameter_335: name = "parameter_335" + original_name = "layer_norm_1.b_0_deepcopy_35" shape = [1024] dtype = "float32" min_val = float("-0.000495841") @@ -3696,6 +4032,7 @@ class Program_weight_tensor_parameter_335: class Program_weight_tensor_parameter_336: name = "parameter_336" + original_name = "layer_norm_1.w_0_deepcopy_34" shape = [1024] dtype = "float32" min_val = float("0.825239") @@ -3707,6 +4044,7 @@ class Program_weight_tensor_parameter_336: class Program_weight_tensor_parameter_337: name = "parameter_337" + original_name = "linear_2.b_0_deepcopy_31" shape = [1024] dtype = "float32" min_val = float("-0.0182544") @@ -3718,6 +4056,7 @@ class Program_weight_tensor_parameter_337: class Program_weight_tensor_parameter_338: name = "parameter_338" + original_name = "linear_2.w_0_deepcopy_30" shape = [2048, 1024] dtype = "float32" min_val = float("-0.0185876") @@ -3729,6 +4068,7 @@ class Program_weight_tensor_parameter_338: class Program_weight_tensor_parameter_339: name = "parameter_339" + original_name = "linear_1.b_0_deepcopy_29" shape = [2048] dtype = "float32" min_val = float("-0.0258719") @@ -3740,6 +4080,7 @@ class Program_weight_tensor_parameter_339: class Program_weight_tensor_parameter_340: name = "parameter_340" + original_name = "linear_1.w_0_deepcopy_28" shape = [1024, 2048] dtype = "float32" min_val = float("-0.0260955") @@ -3751,6 +4092,7 @@ class Program_weight_tensor_parameter_340: class Program_weight_tensor_parameter_341: name = "parameter_341" + original_name = "layer_norm_0.b_0_deepcopy_33" shape = [1024] dtype = "float32" min_val = float("-0.000468908") @@ -3762,6 +4104,7 @@ class Program_weight_tensor_parameter_341: class Program_weight_tensor_parameter_342: name = "parameter_342" + original_name = "layer_norm_0.w_0_deepcopy_32" shape = [1024] dtype = "float32" min_val = float("0.825683") @@ -3773,6 +4116,7 @@ class Program_weight_tensor_parameter_342: class Program_weight_tensor_parameter_343: name = "parameter_343" + original_name = "linear_0.b_0_deepcopy_27" shape = [1024] dtype = "float32" min_val = float("-0.000528411") @@ -3784,6 +4128,7 @@ class Program_weight_tensor_parameter_343: class Program_weight_tensor_parameter_344: name = "parameter_344" + original_name = "linear_0.w_0_deepcopy_26" shape = [1024, 1024] dtype = "float32" min_val = float("-0.0450293") @@ -3795,6 +4140,7 @@ class Program_weight_tensor_parameter_344: class Program_weight_tensor_parameter_345: name = "parameter_345" + original_name = "layer_norm_1.b_0_deepcopy_23" shape = [1024] dtype = "float32" min_val = float("-0.000544272") @@ -3806,6 +4152,7 @@ class Program_weight_tensor_parameter_345: class Program_weight_tensor_parameter_346: name = "parameter_346" + original_name = "layer_norm_1.w_0_deepcopy_22" shape = [1024] dtype = "float32" min_val = float("0.825946") @@ -3817,6 +4164,7 @@ class Program_weight_tensor_parameter_346: class Program_weight_tensor_parameter_347: name = "parameter_347" + original_name = "linear_2.b_0_deepcopy_19" shape = [1024] dtype = "float32" min_val = float("-0.0184487") @@ -3828,6 +4176,7 @@ class Program_weight_tensor_parameter_347: class Program_weight_tensor_parameter_348: name = "parameter_348" + original_name = "linear_2.w_0_deepcopy_18" shape = [2048, 1024] dtype = "float32" min_val = float("-0.0185587") @@ -3839,6 +4188,7 @@ class Program_weight_tensor_parameter_348: class Program_weight_tensor_parameter_349: name = "parameter_349" + original_name = "linear_1.b_0_deepcopy_17" shape = [2048] dtype = "float32" min_val = float("-0.0259392") @@ -3850,6 +4200,7 @@ class Program_weight_tensor_parameter_349: class Program_weight_tensor_parameter_350: name = "parameter_350" + original_name = "linear_1.w_0_deepcopy_16" shape = [1024, 2048] dtype = "float32" min_val = float("-0.0261446") @@ -3861,6 +4212,7 @@ class Program_weight_tensor_parameter_350: class Program_weight_tensor_parameter_351: name = "parameter_351" + original_name = "layer_norm_0.b_0_deepcopy_21" shape = [1024] dtype = "float32" min_val = float("-0.000525158") @@ -3872,6 +4224,7 @@ class Program_weight_tensor_parameter_351: class Program_weight_tensor_parameter_352: name = "parameter_352" + original_name = "layer_norm_0.w_0_deepcopy_20" shape = [1024] dtype = "float32" min_val = float("0.826325") @@ -3883,6 +4236,7 @@ class Program_weight_tensor_parameter_352: class Program_weight_tensor_parameter_353: name = "parameter_353" + original_name = "linear_0.b_0_deepcopy_15" shape = [1024] dtype = "float32" min_val = float("-0.000560432") @@ -3894,6 +4248,7 @@ class Program_weight_tensor_parameter_353: class Program_weight_tensor_parameter_354: name = "parameter_354" + original_name = "linear_0.w_0_deepcopy_14" shape = [1024, 1024] dtype = "float32" min_val = float("-0.0451116") @@ -3905,6 +4260,7 @@ class Program_weight_tensor_parameter_354: class Program_weight_tensor_parameter_355: name = "parameter_355" + original_name = "layer_norm_1.b_0_deepcopy_11" shape = [1024] dtype = "float32" min_val = float("-0.000823759") @@ -3916,6 +4272,7 @@ class Program_weight_tensor_parameter_355: class Program_weight_tensor_parameter_356: name = "parameter_356" + original_name = "layer_norm_1.w_0_deepcopy_10" shape = [1024] dtype = "float32" min_val = float("0.826283") @@ -3927,6 +4284,7 @@ class Program_weight_tensor_parameter_356: class Program_weight_tensor_parameter_357: name = "parameter_357" + original_name = "linear_2.b_0_deepcopy_7" shape = [1024] dtype = "float32" min_val = float("-0.0185659") @@ -3938,6 +4296,7 @@ class Program_weight_tensor_parameter_357: class Program_weight_tensor_parameter_358: name = "parameter_358" + original_name = "linear_2.w_0_deepcopy_6" shape = [2048, 1024] dtype = "float32" min_val = float("-0.0186584") @@ -3949,6 +4308,7 @@ class Program_weight_tensor_parameter_358: class Program_weight_tensor_parameter_359: name = "parameter_359" + original_name = "linear_1.b_0_deepcopy_5" shape = [2048] dtype = "float32" min_val = float("-0.0260158") @@ -3960,6 +4320,7 @@ class Program_weight_tensor_parameter_359: class Program_weight_tensor_parameter_360: name = "parameter_360" + original_name = "linear_1.w_0_deepcopy_4" shape = [1024, 2048] dtype = "float32" min_val = float("-0.026139") @@ -3971,6 +4332,7 @@ class Program_weight_tensor_parameter_360: class Program_weight_tensor_parameter_361: name = "parameter_361" + original_name = "layer_norm_0.b_0_deepcopy_9" shape = [1024] dtype = "float32" min_val = float("-0.000913026") @@ -3982,6 +4344,7 @@ class Program_weight_tensor_parameter_361: class Program_weight_tensor_parameter_362: name = "parameter_362" + original_name = "layer_norm_0.w_0_deepcopy_8" shape = [1024] dtype = "float32" min_val = float("0.826227") @@ -3993,6 +4356,7 @@ class Program_weight_tensor_parameter_362: class Program_weight_tensor_parameter_363: name = "parameter_363" + original_name = "linear_0.b_0_deepcopy_3" shape = [1024] dtype = "float32" min_val = float("-0.00089386") @@ -4004,6 +4368,7 @@ class Program_weight_tensor_parameter_363: class Program_weight_tensor_parameter_364: name = "parameter_364" + original_name = "linear_0.w_0_deepcopy_2" shape = [1024, 1024] dtype = "float32" min_val = float("-0.0456631") @@ -4015,6 +4380,7 @@ class Program_weight_tensor_parameter_364: class Program_weight_tensor_parameter_365: name = "parameter_365" + original_name = "batch_norm2d_72.b_0" shape = [1024] dtype = "float32" min_val = float("-3.75937") @@ -4026,6 +4392,7 @@ class Program_weight_tensor_parameter_365: class Program_weight_tensor_parameter_366: name = "parameter_366" + original_name = "batch_norm2d_72.w_0" shape = [1024] dtype = "float32" min_val = float("1.61944") @@ -4037,6 +4404,7 @@ class Program_weight_tensor_parameter_366: class Program_weight_tensor_parameter_367: name = "parameter_367" + original_name = "batch_norm2d_72.w_2" shape = [1024] dtype = "float32" min_val = float("0.00515514") @@ -4048,6 +4416,7 @@ class Program_weight_tensor_parameter_367: class Program_weight_tensor_parameter_368: name = "parameter_368" + original_name = "batch_norm2d_72.w_1" shape = [1024] dtype = "float32" min_val = float("-0.173492") @@ -4059,6 +4428,7 @@ class Program_weight_tensor_parameter_368: class Program_weight_tensor_parameter_369: name = "parameter_369" + original_name = "conv2d_76.w_0" shape = [1024, 768, 1, 1] dtype = "float32" min_val = float("-0.0420016") @@ -4070,6 +4440,7 @@ class Program_weight_tensor_parameter_369: class Program_weight_tensor_parameter_370: name = "parameter_370" + original_name = "conv2d_75.b_0" shape = [768] dtype = "float32" min_val = float("-0.0144958") @@ -4081,6 +4452,7 @@ class Program_weight_tensor_parameter_370: class Program_weight_tensor_parameter_371: name = "parameter_371" + original_name = "conv2d_75.w_0" shape = [768, 768, 1, 1] dtype = "float32" min_val = float("-0.0809974") @@ -4092,6 +4464,7 @@ class Program_weight_tensor_parameter_371: class Program_weight_tensor_parameter_372: name = "parameter_372" + original_name = "batch_norm2d_71.b_0" shape = [384] dtype = "float32" min_val = float("-1.77404") @@ -4103,6 +4476,7 @@ class Program_weight_tensor_parameter_372: class Program_weight_tensor_parameter_373: name = "parameter_373" + original_name = "batch_norm2d_71.w_0" shape = [384] dtype = "float32" min_val = float("0.188368") @@ -4114,6 +4488,7 @@ class Program_weight_tensor_parameter_373: class Program_weight_tensor_parameter_374: name = "parameter_374" + original_name = "batch_norm2d_71.w_2" shape = [384] dtype = "float32" min_val = float("7.69323e-05") @@ -4125,6 +4500,7 @@ class Program_weight_tensor_parameter_374: class Program_weight_tensor_parameter_375: name = "parameter_375" + original_name = "batch_norm2d_71.w_1" shape = [384] dtype = "float32" min_val = float("-0.0656167") @@ -4136,6 +4512,7 @@ class Program_weight_tensor_parameter_375: class Program_weight_tensor_parameter_376: name = "parameter_376" + original_name = "conv2d_74.w_0" shape = [384, 384, 1, 1] dtype = "float32" min_val = float("-0.020871") @@ -4147,6 +4524,7 @@ class Program_weight_tensor_parameter_376: class Program_weight_tensor_parameter_377: name = "parameter_377" + original_name = "batch_norm2d_70.b_0" shape = [384] dtype = "float32" min_val = float("-1.77405") @@ -4158,6 +4536,7 @@ class Program_weight_tensor_parameter_377: class Program_weight_tensor_parameter_378: name = "parameter_378" + original_name = "batch_norm2d_70.w_0" shape = [384] dtype = "float32" min_val = float("0.335122") @@ -4169,6 +4548,7 @@ class Program_weight_tensor_parameter_378: class Program_weight_tensor_parameter_379: name = "parameter_379" + original_name = "batch_norm2d_70.w_2" shape = [384] dtype = "float32" min_val = float("0.000764026") @@ -4180,6 +4560,7 @@ class Program_weight_tensor_parameter_379: class Program_weight_tensor_parameter_380: name = "parameter_380" + original_name = "batch_norm2d_70.w_1" shape = [384] dtype = "float32" min_val = float("-0.229833") @@ -4191,6 +4572,7 @@ class Program_weight_tensor_parameter_380: class Program_weight_tensor_parameter_381: name = "parameter_381" + original_name = "conv2d_73.w_0" shape = [384, 384, 3, 3] dtype = "float32" min_val = float("-0.0185255") @@ -4202,6 +4584,7 @@ class Program_weight_tensor_parameter_381: class Program_weight_tensor_parameter_382: name = "parameter_382" + original_name = "batch_norm2d_69.b_0" shape = [384] dtype = "float32" min_val = float("-2.58205") @@ -4213,6 +4596,7 @@ class Program_weight_tensor_parameter_382: class Program_weight_tensor_parameter_383: name = "parameter_383" + original_name = "batch_norm2d_69.w_0" shape = [384] dtype = "float32" min_val = float("0.51894") @@ -4224,6 +4608,7 @@ class Program_weight_tensor_parameter_383: class Program_weight_tensor_parameter_384: name = "parameter_384" + original_name = "batch_norm2d_69.w_2" shape = [384] dtype = "float32" min_val = float("0.0445179") @@ -4235,6 +4620,7 @@ class Program_weight_tensor_parameter_384: class Program_weight_tensor_parameter_385: name = "parameter_385" + original_name = "batch_norm2d_69.w_1" shape = [384] dtype = "float32" min_val = float("-1.05877") @@ -4246,6 +4632,7 @@ class Program_weight_tensor_parameter_385: class Program_weight_tensor_parameter_386: name = "parameter_386" + original_name = "conv2d_72.w_0" shape = [384, 384, 3, 3] dtype = "float32" min_val = float("-0.0217847") @@ -4257,6 +4644,7 @@ class Program_weight_tensor_parameter_386: class Program_weight_tensor_parameter_387: name = "parameter_387" + original_name = "batch_norm2d_68.b_0" shape = [384] dtype = "float32" min_val = float("-1.93932") @@ -4268,6 +4656,7 @@ class Program_weight_tensor_parameter_387: class Program_weight_tensor_parameter_388: name = "parameter_388" + original_name = "batch_norm2d_68.w_0" shape = [384] dtype = "float32" min_val = float("0.163976") @@ -4279,6 +4668,7 @@ class Program_weight_tensor_parameter_388: class Program_weight_tensor_parameter_389: name = "parameter_389" + original_name = "batch_norm2d_68.w_2" shape = [384] dtype = "float32" min_val = float("8.46446e-05") @@ -4290,6 +4680,7 @@ class Program_weight_tensor_parameter_389: class Program_weight_tensor_parameter_390: name = "parameter_390" + original_name = "batch_norm2d_68.w_1" shape = [384] dtype = "float32" min_val = float("-0.0395058") @@ -4301,6 +4692,7 @@ class Program_weight_tensor_parameter_390: class Program_weight_tensor_parameter_391: name = "parameter_391" + original_name = "conv2d_71.w_0" shape = [384, 384, 1, 1] dtype = "float32" min_val = float("-0.0311026") @@ -4312,6 +4704,7 @@ class Program_weight_tensor_parameter_391: class Program_weight_tensor_parameter_392: name = "parameter_392" + original_name = "batch_norm2d_67.b_0" shape = [384] dtype = "float32" min_val = float("-1.9394") @@ -4323,6 +4716,7 @@ class Program_weight_tensor_parameter_392: class Program_weight_tensor_parameter_393: name = "parameter_393" + original_name = "batch_norm2d_67.w_0" shape = [384] dtype = "float32" min_val = float("0.583818") @@ -4334,6 +4728,7 @@ class Program_weight_tensor_parameter_393: class Program_weight_tensor_parameter_394: name = "parameter_394" + original_name = "batch_norm2d_67.w_2" shape = [384] dtype = "float32" min_val = float("0.00151649") @@ -4345,6 +4740,7 @@ class Program_weight_tensor_parameter_394: class Program_weight_tensor_parameter_395: name = "parameter_395" + original_name = "batch_norm2d_67.w_1" shape = [384] dtype = "float32" min_val = float("-0.114817") @@ -4356,6 +4752,7 @@ class Program_weight_tensor_parameter_395: class Program_weight_tensor_parameter_396: name = "parameter_396" + original_name = "conv2d_70.w_0" shape = [384, 384, 3, 3] dtype = "float32" min_val = float("-0.0211861") @@ -4367,6 +4764,7 @@ class Program_weight_tensor_parameter_396: class Program_weight_tensor_parameter_397: name = "parameter_397" + original_name = "batch_norm2d_66.b_0" shape = [384] dtype = "float32" min_val = float("-2.39618") @@ -4378,6 +4776,7 @@ class Program_weight_tensor_parameter_397: class Program_weight_tensor_parameter_398: name = "parameter_398" + original_name = "batch_norm2d_66.w_0" shape = [384] dtype = "float32" min_val = float("0.454223") @@ -4389,6 +4788,7 @@ class Program_weight_tensor_parameter_398: class Program_weight_tensor_parameter_399: name = "parameter_399" + original_name = "batch_norm2d_66.w_2" shape = [384] dtype = "float32" min_val = float("0.0369914") @@ -4400,6 +4800,7 @@ class Program_weight_tensor_parameter_399: class Program_weight_tensor_parameter_400: name = "parameter_400" + original_name = "batch_norm2d_66.w_1" shape = [384] dtype = "float32" min_val = float("-0.916864") @@ -4411,6 +4812,7 @@ class Program_weight_tensor_parameter_400: class Program_weight_tensor_parameter_401: name = "parameter_401" + original_name = "conv2d_69.w_0" shape = [384, 384, 3, 3] dtype = "float32" min_val = float("-0.0304568") @@ -4422,6 +4824,7 @@ class Program_weight_tensor_parameter_401: class Program_weight_tensor_parameter_402: name = "parameter_402" + original_name = "batch_norm2d_65.b_0" shape = [384] dtype = "float32" min_val = float("-1.87628") @@ -4433,6 +4836,7 @@ class Program_weight_tensor_parameter_402: class Program_weight_tensor_parameter_403: name = "parameter_403" + original_name = "batch_norm2d_65.w_0" shape = [384] dtype = "float32" min_val = float("0.0771953") @@ -4444,6 +4848,7 @@ class Program_weight_tensor_parameter_403: class Program_weight_tensor_parameter_404: name = "parameter_404" + original_name = "batch_norm2d_65.w_2" shape = [384] dtype = "float32" min_val = float("7.57603e-05") @@ -4455,6 +4860,7 @@ class Program_weight_tensor_parameter_404: class Program_weight_tensor_parameter_405: name = "parameter_405" + original_name = "batch_norm2d_65.w_1" shape = [384] dtype = "float32" min_val = float("-0.0528798") @@ -4466,6 +4872,7 @@ class Program_weight_tensor_parameter_405: class Program_weight_tensor_parameter_406: name = "parameter_406" + original_name = "conv2d_68.w_0" shape = [384, 384, 1, 1] dtype = "float32" min_val = float("-0.0213328") @@ -4477,6 +4884,7 @@ class Program_weight_tensor_parameter_406: class Program_weight_tensor_parameter_407: name = "parameter_407" + original_name = "batch_norm2d_64.b_0" shape = [384] dtype = "float32" min_val = float("-1.87669") @@ -4488,6 +4896,7 @@ class Program_weight_tensor_parameter_407: class Program_weight_tensor_parameter_408: name = "parameter_408" + original_name = "batch_norm2d_64.w_0" shape = [384] dtype = "float32" min_val = float("0.522977") @@ -4499,6 +4908,7 @@ class Program_weight_tensor_parameter_408: class Program_weight_tensor_parameter_409: name = "parameter_409" + original_name = "batch_norm2d_64.w_2" shape = [384] dtype = "float32" min_val = float("0.00214087") @@ -4510,6 +4920,7 @@ class Program_weight_tensor_parameter_409: class Program_weight_tensor_parameter_410: name = "parameter_410" + original_name = "batch_norm2d_64.w_1" shape = [384] dtype = "float32" min_val = float("-0.272097") @@ -4521,6 +4932,7 @@ class Program_weight_tensor_parameter_410: class Program_weight_tensor_parameter_411: name = "parameter_411" + original_name = "conv2d_67.w_0" shape = [384, 384, 3, 3] dtype = "float32" min_val = float("-0.0214852") @@ -4532,6 +4944,7 @@ class Program_weight_tensor_parameter_411: class Program_weight_tensor_parameter_412: name = "parameter_412" + original_name = "batch_norm2d_63.b_0" shape = [384] dtype = "float32" min_val = float("-2.1565") @@ -4543,6 +4956,7 @@ class Program_weight_tensor_parameter_412: class Program_weight_tensor_parameter_413: name = "parameter_413" + original_name = "batch_norm2d_63.w_0" shape = [384] dtype = "float32" min_val = float("0.707119") @@ -4554,6 +4968,7 @@ class Program_weight_tensor_parameter_413: class Program_weight_tensor_parameter_414: name = "parameter_414" + original_name = "batch_norm2d_63.w_2" shape = [384] dtype = "float32" min_val = float("0.0267598") @@ -4565,6 +4980,7 @@ class Program_weight_tensor_parameter_414: class Program_weight_tensor_parameter_415: name = "parameter_415" + original_name = "batch_norm2d_63.w_1" shape = [384] dtype = "float32" min_val = float("-0.737016") @@ -4576,6 +4992,7 @@ class Program_weight_tensor_parameter_415: class Program_weight_tensor_parameter_416: name = "parameter_416" + original_name = "conv2d_66.w_0" shape = [384, 384, 3, 3] dtype = "float32" min_val = float("-0.0300983") @@ -4587,6 +5004,7 @@ class Program_weight_tensor_parameter_416: class Program_weight_tensor_parameter_417: name = "parameter_417" + original_name = "batch_norm2d_62.b_0" shape = [384] dtype = "float32" min_val = float("-2.92344") @@ -4598,6 +5016,7 @@ class Program_weight_tensor_parameter_417: class Program_weight_tensor_parameter_418: name = "parameter_418" + original_name = "batch_norm2d_62.w_0" shape = [384] dtype = "float32" min_val = float("0.953228") @@ -4609,6 +5028,7 @@ class Program_weight_tensor_parameter_418: class Program_weight_tensor_parameter_419: name = "parameter_419" + original_name = "batch_norm2d_62.w_2" shape = [384] dtype = "float32" min_val = float("0.00273562") @@ -4620,6 +5040,7 @@ class Program_weight_tensor_parameter_419: class Program_weight_tensor_parameter_420: name = "parameter_420" + original_name = "batch_norm2d_62.w_1" shape = [384] dtype = "float32" min_val = float("-0.279172") @@ -4631,6 +5052,7 @@ class Program_weight_tensor_parameter_420: class Program_weight_tensor_parameter_421: name = "parameter_421" + original_name = "conv2d_65.w_0" shape = [384, 768, 1, 1] dtype = "float32" min_val = float("-0.0411036") @@ -4642,6 +5064,7 @@ class Program_weight_tensor_parameter_421: class Program_weight_tensor_parameter_422: name = "parameter_422" + original_name = "batch_norm2d_61.b_0" shape = [384] dtype = "float32" min_val = float("-2.24702") @@ -4653,6 +5076,7 @@ class Program_weight_tensor_parameter_422: class Program_weight_tensor_parameter_423: name = "parameter_423" + original_name = "batch_norm2d_61.w_0" shape = [384] dtype = "float32" min_val = float("0.965876") @@ -4664,6 +5088,7 @@ class Program_weight_tensor_parameter_423: class Program_weight_tensor_parameter_424: name = "parameter_424" + original_name = "batch_norm2d_61.w_2" shape = [384] dtype = "float32" min_val = float("0.000839665") @@ -4675,6 +5100,7 @@ class Program_weight_tensor_parameter_424: class Program_weight_tensor_parameter_425: name = "parameter_425" + original_name = "batch_norm2d_61.w_1" shape = [384] dtype = "float32" min_val = float("-0.0182533") @@ -4686,6 +5112,7 @@ class Program_weight_tensor_parameter_425: class Program_weight_tensor_parameter_426: name = "parameter_426" + original_name = "conv2d_64.w_0" shape = [384, 768, 1, 1] dtype = "float32" min_val = float("-0.0837021") @@ -4697,6 +5124,7 @@ class Program_weight_tensor_parameter_426: class Program_weight_tensor_parameter_427: name = "parameter_427" + original_name = "batch_norm2d_60.b_0" shape = [768] dtype = "float32" min_val = float("-2.40194") @@ -4708,6 +5136,7 @@ class Program_weight_tensor_parameter_427: class Program_weight_tensor_parameter_428: name = "parameter_428" + original_name = "batch_norm2d_60.w_0" shape = [768] dtype = "float32" min_val = float("0.53146") @@ -4719,6 +5148,7 @@ class Program_weight_tensor_parameter_428: class Program_weight_tensor_parameter_429: name = "parameter_429" + original_name = "batch_norm2d_60.w_2" shape = [768] dtype = "float32" min_val = float("0.00745832") @@ -4730,6 +5160,7 @@ class Program_weight_tensor_parameter_429: class Program_weight_tensor_parameter_430: name = "parameter_430" + original_name = "batch_norm2d_60.w_1" shape = [768] dtype = "float32" min_val = float("-0.236023") @@ -4741,6 +5172,7 @@ class Program_weight_tensor_parameter_430: class Program_weight_tensor_parameter_431: name = "parameter_431" + original_name = "conv2d_63.w_0" shape = [768, 512, 3, 3] dtype = "float32" min_val = float("-0.0383779") @@ -4752,6 +5184,7 @@ class Program_weight_tensor_parameter_431: class Program_weight_tensor_parameter_432: name = "parameter_432" + original_name = "batch_norm2d_59.b_0" shape = [512] dtype = "float32" min_val = float("-3.39029") @@ -4763,6 +5196,7 @@ class Program_weight_tensor_parameter_432: class Program_weight_tensor_parameter_433: name = "parameter_433" + original_name = "batch_norm2d_59.w_0" shape = [512] dtype = "float32" min_val = float("0.520928") @@ -4774,6 +5208,7 @@ class Program_weight_tensor_parameter_433: class Program_weight_tensor_parameter_434: name = "parameter_434" + original_name = "batch_norm2d_59.w_2" shape = [512] dtype = "float32" min_val = float("0.00230842") @@ -4785,6 +5220,7 @@ class Program_weight_tensor_parameter_434: class Program_weight_tensor_parameter_435: name = "parameter_435" + original_name = "batch_norm2d_59.w_1" shape = [512] dtype = "float32" min_val = float("-0.159179") @@ -4796,6 +5232,7 @@ class Program_weight_tensor_parameter_435: class Program_weight_tensor_parameter_436: name = "parameter_436" + original_name = "conv2d_62.w_0" shape = [512, 384, 1, 1] dtype = "float32" min_val = float("-0.208779") @@ -4807,6 +5244,7 @@ class Program_weight_tensor_parameter_436: class Program_weight_tensor_parameter_437: name = "parameter_437" + original_name = "conv2d_61.b_0" shape = [384] dtype = "float32" min_val = float("-0.0103559") @@ -4818,6 +5256,7 @@ class Program_weight_tensor_parameter_437: class Program_weight_tensor_parameter_438: name = "parameter_438" + original_name = "conv2d_61.w_0" shape = [384, 384, 1, 1] dtype = "float32" min_val = float("-0.204999") @@ -4829,6 +5268,7 @@ class Program_weight_tensor_parameter_438: class Program_weight_tensor_parameter_439: name = "parameter_439" + original_name = "batch_norm2d_58.b_0" shape = [192] dtype = "float32" min_val = float("-1.97063") @@ -4840,6 +5280,7 @@ class Program_weight_tensor_parameter_439: class Program_weight_tensor_parameter_440: name = "parameter_440" + original_name = "batch_norm2d_58.w_0" shape = [192] dtype = "float32" min_val = float("0.0528508") @@ -4851,6 +5292,7 @@ class Program_weight_tensor_parameter_440: class Program_weight_tensor_parameter_441: name = "parameter_441" + original_name = "batch_norm2d_58.w_2" shape = [192] dtype = "float32" min_val = float("9.94453e-05") @@ -4862,6 +5304,7 @@ class Program_weight_tensor_parameter_441: class Program_weight_tensor_parameter_442: name = "parameter_442" + original_name = "batch_norm2d_58.w_1" shape = [192] dtype = "float32" min_val = float("-0.0376085") @@ -4873,6 +5316,7 @@ class Program_weight_tensor_parameter_442: class Program_weight_tensor_parameter_443: name = "parameter_443" + original_name = "conv2d_60.w_0" shape = [192, 192, 1, 1] dtype = "float32" min_val = float("-0.0210389") @@ -4884,6 +5328,7 @@ class Program_weight_tensor_parameter_443: class Program_weight_tensor_parameter_444: name = "parameter_444" + original_name = "batch_norm2d_57.b_0" shape = [192] dtype = "float32" min_val = float("-1.97059") @@ -4895,6 +5340,7 @@ class Program_weight_tensor_parameter_444: class Program_weight_tensor_parameter_445: name = "parameter_445" + original_name = "batch_norm2d_57.w_0" shape = [192] dtype = "float32" min_val = float("0.372764") @@ -4906,6 +5352,7 @@ class Program_weight_tensor_parameter_445: class Program_weight_tensor_parameter_446: name = "parameter_446" + original_name = "batch_norm2d_57.w_2" shape = [192] dtype = "float32" min_val = float("0.0014863") @@ -4917,6 +5364,7 @@ class Program_weight_tensor_parameter_446: class Program_weight_tensor_parameter_447: name = "parameter_447" + original_name = "batch_norm2d_57.w_1" shape = [192] dtype = "float32" min_val = float("-0.115289") @@ -4928,6 +5376,7 @@ class Program_weight_tensor_parameter_447: class Program_weight_tensor_parameter_448: name = "parameter_448" + original_name = "conv2d_59.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.031927") @@ -4939,6 +5388,7 @@ class Program_weight_tensor_parameter_448: class Program_weight_tensor_parameter_449: name = "parameter_449" + original_name = "batch_norm2d_56.b_0" shape = [192] dtype = "float32" min_val = float("-2.89054") @@ -4950,6 +5400,7 @@ class Program_weight_tensor_parameter_449: class Program_weight_tensor_parameter_450: name = "parameter_450" + original_name = "batch_norm2d_56.w_0" shape = [192] dtype = "float32" min_val = float("0.695074") @@ -4961,6 +5412,7 @@ class Program_weight_tensor_parameter_450: class Program_weight_tensor_parameter_451: name = "parameter_451" + original_name = "batch_norm2d_56.w_2" shape = [192] dtype = "float32" min_val = float("0.0658237") @@ -4972,6 +5424,7 @@ class Program_weight_tensor_parameter_451: class Program_weight_tensor_parameter_452: name = "parameter_452" + original_name = "batch_norm2d_56.w_1" shape = [192] dtype = "float32" min_val = float("-2.47032") @@ -4983,6 +5436,7 @@ class Program_weight_tensor_parameter_452: class Program_weight_tensor_parameter_453: name = "parameter_453" + original_name = "conv2d_58.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.0350379") @@ -4994,6 +5448,7 @@ class Program_weight_tensor_parameter_453: class Program_weight_tensor_parameter_454: name = "parameter_454" + original_name = "batch_norm2d_55.b_0" shape = [192] dtype = "float32" min_val = float("-1.94031") @@ -5005,6 +5460,7 @@ class Program_weight_tensor_parameter_454: class Program_weight_tensor_parameter_455: name = "parameter_455" + original_name = "batch_norm2d_55.w_0" shape = [192] dtype = "float32" min_val = float("0.0449424") @@ -5016,6 +5472,7 @@ class Program_weight_tensor_parameter_455: class Program_weight_tensor_parameter_456: name = "parameter_456" + original_name = "batch_norm2d_55.w_2" shape = [192] dtype = "float32" min_val = float("7.91667e-05") @@ -5027,6 +5484,7 @@ class Program_weight_tensor_parameter_456: class Program_weight_tensor_parameter_457: name = "parameter_457" + original_name = "batch_norm2d_55.w_1" shape = [192] dtype = "float32" min_val = float("-0.0363552") @@ -5038,6 +5496,7 @@ class Program_weight_tensor_parameter_457: class Program_weight_tensor_parameter_458: name = "parameter_458" + original_name = "conv2d_57.w_0" shape = [192, 192, 1, 1] dtype = "float32" min_val = float("-0.02483") @@ -5049,6 +5508,7 @@ class Program_weight_tensor_parameter_458: class Program_weight_tensor_parameter_459: name = "parameter_459" + original_name = "batch_norm2d_54.b_0" shape = [192] dtype = "float32" min_val = float("-1.94031") @@ -5060,6 +5520,7 @@ class Program_weight_tensor_parameter_459: class Program_weight_tensor_parameter_460: name = "parameter_460" + original_name = "batch_norm2d_54.w_0" shape = [192] dtype = "float32" min_val = float("0.481654") @@ -5071,6 +5532,7 @@ class Program_weight_tensor_parameter_460: class Program_weight_tensor_parameter_461: name = "parameter_461" + original_name = "batch_norm2d_54.w_2" shape = [192] dtype = "float32" min_val = float("0.00303177") @@ -5082,6 +5544,7 @@ class Program_weight_tensor_parameter_461: class Program_weight_tensor_parameter_462: name = "parameter_462" + original_name = "batch_norm2d_54.w_1" shape = [192] dtype = "float32" min_val = float("-0.0803161") @@ -5093,6 +5556,7 @@ class Program_weight_tensor_parameter_462: class Program_weight_tensor_parameter_463: name = "parameter_463" + original_name = "conv2d_56.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.0229799") @@ -5104,6 +5568,7 @@ class Program_weight_tensor_parameter_463: class Program_weight_tensor_parameter_464: name = "parameter_464" + original_name = "batch_norm2d_53.b_0" shape = [192] dtype = "float32" min_val = float("-2.50826") @@ -5115,6 +5580,7 @@ class Program_weight_tensor_parameter_464: class Program_weight_tensor_parameter_465: name = "parameter_465" + original_name = "batch_norm2d_53.w_0" shape = [192] dtype = "float32" min_val = float("0.653803") @@ -5126,6 +5592,7 @@ class Program_weight_tensor_parameter_465: class Program_weight_tensor_parameter_466: name = "parameter_466" + original_name = "batch_norm2d_53.w_2" shape = [192] dtype = "float32" min_val = float("0.0475951") @@ -5137,6 +5604,7 @@ class Program_weight_tensor_parameter_466: class Program_weight_tensor_parameter_467: name = "parameter_467" + original_name = "batch_norm2d_53.w_1" shape = [192] dtype = "float32" min_val = float("-2.16167") @@ -5148,6 +5616,7 @@ class Program_weight_tensor_parameter_467: class Program_weight_tensor_parameter_468: name = "parameter_468" + original_name = "conv2d_55.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.038582") @@ -5159,6 +5628,7 @@ class Program_weight_tensor_parameter_468: class Program_weight_tensor_parameter_469: name = "parameter_469" + original_name = "batch_norm2d_52.b_0" shape = [192] dtype = "float32" min_val = float("-1.75738") @@ -5170,6 +5640,7 @@ class Program_weight_tensor_parameter_469: class Program_weight_tensor_parameter_470: name = "parameter_470" + original_name = "batch_norm2d_52.w_0" shape = [192] dtype = "float32" min_val = float("0.00305103") @@ -5181,6 +5652,7 @@ class Program_weight_tensor_parameter_470: class Program_weight_tensor_parameter_471: name = "parameter_471" + original_name = "batch_norm2d_52.w_2" shape = [192] dtype = "float32" min_val = float("1.02293e-06") @@ -5192,6 +5664,7 @@ class Program_weight_tensor_parameter_471: class Program_weight_tensor_parameter_472: name = "parameter_472" + original_name = "batch_norm2d_52.w_1" shape = [192] dtype = "float32" min_val = float("-0.031609") @@ -5203,6 +5676,7 @@ class Program_weight_tensor_parameter_472: class Program_weight_tensor_parameter_473: name = "parameter_473" + original_name = "conv2d_54.w_0" shape = [192, 192, 1, 1] dtype = "float32" min_val = float("-0.0307534") @@ -5214,6 +5688,7 @@ class Program_weight_tensor_parameter_473: class Program_weight_tensor_parameter_474: name = "parameter_474" + original_name = "batch_norm2d_51.b_0" shape = [192] dtype = "float32" min_val = float("-1.75744") @@ -5225,6 +5700,7 @@ class Program_weight_tensor_parameter_474: class Program_weight_tensor_parameter_475: name = "parameter_475" + original_name = "batch_norm2d_51.w_0" shape = [192] dtype = "float32" min_val = float("0.405457") @@ -5236,6 +5712,7 @@ class Program_weight_tensor_parameter_475: class Program_weight_tensor_parameter_476: name = "parameter_476" + original_name = "batch_norm2d_51.w_2" shape = [192] dtype = "float32" min_val = float("0.00267969") @@ -5247,6 +5724,7 @@ class Program_weight_tensor_parameter_476: class Program_weight_tensor_parameter_477: name = "parameter_477" + original_name = "batch_norm2d_51.w_1" shape = [192] dtype = "float32" min_val = float("-0.0881741") @@ -5258,6 +5736,7 @@ class Program_weight_tensor_parameter_477: class Program_weight_tensor_parameter_478: name = "parameter_478" + original_name = "conv2d_53.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.0336081") @@ -5269,6 +5748,7 @@ class Program_weight_tensor_parameter_478: class Program_weight_tensor_parameter_479: name = "parameter_479" + original_name = "batch_norm2d_50.b_0" shape = [192] dtype = "float32" min_val = float("-2.49703") @@ -5280,6 +5760,7 @@ class Program_weight_tensor_parameter_479: class Program_weight_tensor_parameter_480: name = "parameter_480" + original_name = "batch_norm2d_50.w_0" shape = [192] dtype = "float32" min_val = float("0.652493") @@ -5291,6 +5772,7 @@ class Program_weight_tensor_parameter_480: class Program_weight_tensor_parameter_481: name = "parameter_481" + original_name = "batch_norm2d_50.w_2" shape = [192] dtype = "float32" min_val = float("0.0304637") @@ -5302,6 +5784,7 @@ class Program_weight_tensor_parameter_481: class Program_weight_tensor_parameter_482: name = "parameter_482" + original_name = "batch_norm2d_50.w_1" shape = [192] dtype = "float32" min_val = float("-1.70097") @@ -5313,6 +5796,7 @@ class Program_weight_tensor_parameter_482: class Program_weight_tensor_parameter_483: name = "parameter_483" + original_name = "conv2d_52.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.0472912") @@ -5324,6 +5808,7 @@ class Program_weight_tensor_parameter_483: class Program_weight_tensor_parameter_484: name = "parameter_484" + original_name = "batch_norm2d_49.b_0" shape = [192] dtype = "float32" min_val = float("-2.07915") @@ -5335,6 +5820,7 @@ class Program_weight_tensor_parameter_484: class Program_weight_tensor_parameter_485: name = "parameter_485" + original_name = "batch_norm2d_49.w_0" shape = [192] dtype = "float32" min_val = float("0.000522804") @@ -5346,6 +5832,7 @@ class Program_weight_tensor_parameter_485: class Program_weight_tensor_parameter_486: name = "parameter_486" + original_name = "batch_norm2d_49.w_2" shape = [192] dtype = "float32" min_val = float("5.96543e-08") @@ -5357,6 +5844,7 @@ class Program_weight_tensor_parameter_486: class Program_weight_tensor_parameter_487: name = "parameter_487" + original_name = "batch_norm2d_49.w_1" shape = [192] dtype = "float32" min_val = float("-0.0266706") @@ -5368,6 +5856,7 @@ class Program_weight_tensor_parameter_487: class Program_weight_tensor_parameter_488: name = "parameter_488" + original_name = "conv2d_51.w_0" shape = [192, 192, 1, 1] dtype = "float32" min_val = float("-0.0207564") @@ -5379,6 +5868,7 @@ class Program_weight_tensor_parameter_488: class Program_weight_tensor_parameter_489: name = "parameter_489" + original_name = "batch_norm2d_48.b_0" shape = [192] dtype = "float32" min_val = float("-2.07924") @@ -5390,6 +5880,7 @@ class Program_weight_tensor_parameter_489: class Program_weight_tensor_parameter_490: name = "parameter_490" + original_name = "batch_norm2d_48.w_0" shape = [192] dtype = "float32" min_val = float("0.395086") @@ -5401,6 +5892,7 @@ class Program_weight_tensor_parameter_490: class Program_weight_tensor_parameter_491: name = "parameter_491" + original_name = "batch_norm2d_48.w_2" shape = [192] dtype = "float32" min_val = float("0.00302737") @@ -5412,6 +5904,7 @@ class Program_weight_tensor_parameter_491: class Program_weight_tensor_parameter_492: name = "parameter_492" + original_name = "batch_norm2d_48.w_1" shape = [192] dtype = "float32" min_val = float("-0.0788482") @@ -5423,6 +5916,7 @@ class Program_weight_tensor_parameter_492: class Program_weight_tensor_parameter_493: name = "parameter_493" + original_name = "conv2d_50.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.0340016") @@ -5434,6 +5928,7 @@ class Program_weight_tensor_parameter_493: class Program_weight_tensor_parameter_494: name = "parameter_494" + original_name = "batch_norm2d_47.b_0" shape = [192] dtype = "float32" min_val = float("-2.74084") @@ -5445,6 +5940,7 @@ class Program_weight_tensor_parameter_494: class Program_weight_tensor_parameter_495: name = "parameter_495" + original_name = "batch_norm2d_47.w_0" shape = [192] dtype = "float32" min_val = float("0.761952") @@ -5456,6 +5952,7 @@ class Program_weight_tensor_parameter_495: class Program_weight_tensor_parameter_496: name = "parameter_496" + original_name = "batch_norm2d_47.w_2" shape = [192] dtype = "float32" min_val = float("0.0278922") @@ -5467,6 +5964,7 @@ class Program_weight_tensor_parameter_496: class Program_weight_tensor_parameter_497: name = "parameter_497" + original_name = "batch_norm2d_47.w_1" shape = [192] dtype = "float32" min_val = float("-1.39522") @@ -5478,6 +5976,7 @@ class Program_weight_tensor_parameter_497: class Program_weight_tensor_parameter_498: name = "parameter_498" + original_name = "conv2d_49.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.0589398") @@ -5489,6 +5988,7 @@ class Program_weight_tensor_parameter_498: class Program_weight_tensor_parameter_499: name = "parameter_499" + original_name = "batch_norm2d_46.b_0" shape = [192] dtype = "float32" min_val = float("-1.212") @@ -5500,6 +6000,7 @@ class Program_weight_tensor_parameter_499: class Program_weight_tensor_parameter_500: name = "parameter_500" + original_name = "batch_norm2d_46.w_0" shape = [192] dtype = "float32" min_val = float("-9.43381e-05") @@ -5511,6 +6012,7 @@ class Program_weight_tensor_parameter_500: class Program_weight_tensor_parameter_501: name = "parameter_501" + original_name = "batch_norm2d_46.w_2" shape = [192] dtype = "float32" min_val = float("2.50564e-10") @@ -5522,6 +6024,7 @@ class Program_weight_tensor_parameter_501: class Program_weight_tensor_parameter_502: name = "parameter_502" + original_name = "batch_norm2d_46.w_1" shape = [192] dtype = "float32" min_val = float("-0.0445459") @@ -5533,6 +6036,7 @@ class Program_weight_tensor_parameter_502: class Program_weight_tensor_parameter_503: name = "parameter_503" + original_name = "conv2d_48.w_0" shape = [192, 192, 1, 1] dtype = "float32" min_val = float("-0.0374404") @@ -5544,6 +6048,7 @@ class Program_weight_tensor_parameter_503: class Program_weight_tensor_parameter_504: name = "parameter_504" + original_name = "batch_norm2d_45.b_0" shape = [192] dtype = "float32" min_val = float("-1.21197") @@ -5555,6 +6060,7 @@ class Program_weight_tensor_parameter_504: class Program_weight_tensor_parameter_505: name = "parameter_505" + original_name = "batch_norm2d_45.w_0" shape = [192] dtype = "float32" min_val = float("0.382853") @@ -5566,6 +6072,7 @@ class Program_weight_tensor_parameter_505: class Program_weight_tensor_parameter_506: name = "parameter_506" + original_name = "batch_norm2d_45.w_2" shape = [192] dtype = "float32" min_val = float("0.00286492") @@ -5577,6 +6084,7 @@ class Program_weight_tensor_parameter_506: class Program_weight_tensor_parameter_507: name = "parameter_507" + original_name = "batch_norm2d_45.w_1" shape = [192] dtype = "float32" min_val = float("-0.0776134") @@ -5588,6 +6096,7 @@ class Program_weight_tensor_parameter_507: class Program_weight_tensor_parameter_508: name = "parameter_508" + original_name = "conv2d_47.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.0368355") @@ -5599,6 +6108,7 @@ class Program_weight_tensor_parameter_508: class Program_weight_tensor_parameter_509: name = "parameter_509" + original_name = "batch_norm2d_44.b_0" shape = [192] dtype = "float32" min_val = float("-2.48699") @@ -5610,6 +6120,7 @@ class Program_weight_tensor_parameter_509: class Program_weight_tensor_parameter_510: name = "parameter_510" + original_name = "batch_norm2d_44.w_0" shape = [192] dtype = "float32" min_val = float("0.689021") @@ -5621,6 +6132,7 @@ class Program_weight_tensor_parameter_510: class Program_weight_tensor_parameter_511: name = "parameter_511" + original_name = "batch_norm2d_44.w_2" shape = [192] dtype = "float32" min_val = float("0.0195954") @@ -5632,6 +6144,7 @@ class Program_weight_tensor_parameter_511: class Program_weight_tensor_parameter_512: name = "parameter_512" + original_name = "batch_norm2d_44.w_1" shape = [192] dtype = "float32" min_val = float("-0.841336") @@ -5643,6 +6156,7 @@ class Program_weight_tensor_parameter_512: class Program_weight_tensor_parameter_513: name = "parameter_513" + original_name = "conv2d_46.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.0647608") @@ -5654,6 +6168,7 @@ class Program_weight_tensor_parameter_513: class Program_weight_tensor_parameter_514: name = "parameter_514" + original_name = "batch_norm2d_43.b_0" shape = [192] dtype = "float32" min_val = float("-1.21773") @@ -5665,6 +6180,7 @@ class Program_weight_tensor_parameter_514: class Program_weight_tensor_parameter_515: name = "parameter_515" + original_name = "batch_norm2d_43.w_0" shape = [192] dtype = "float32" min_val = float("0.00864435") @@ -5676,6 +6192,7 @@ class Program_weight_tensor_parameter_515: class Program_weight_tensor_parameter_516: name = "parameter_516" + original_name = "batch_norm2d_43.w_2" shape = [192] dtype = "float32" min_val = float("2.28062e-05") @@ -5687,6 +6204,7 @@ class Program_weight_tensor_parameter_516: class Program_weight_tensor_parameter_517: name = "parameter_517" + original_name = "batch_norm2d_43.w_1" shape = [192] dtype = "float32" min_val = float("-0.0690564") @@ -5698,6 +6216,7 @@ class Program_weight_tensor_parameter_517: class Program_weight_tensor_parameter_518: name = "parameter_518" + original_name = "conv2d_45.w_0" shape = [192, 192, 1, 1] dtype = "float32" min_val = float("-0.0626678") @@ -5709,6 +6228,7 @@ class Program_weight_tensor_parameter_518: class Program_weight_tensor_parameter_519: name = "parameter_519" + original_name = "batch_norm2d_42.b_0" shape = [192] dtype = "float32" min_val = float("-1.21774") @@ -5720,6 +6240,7 @@ class Program_weight_tensor_parameter_519: class Program_weight_tensor_parameter_520: name = "parameter_520" + original_name = "batch_norm2d_42.w_0" shape = [192] dtype = "float32" min_val = float("0.353208") @@ -5731,6 +6252,7 @@ class Program_weight_tensor_parameter_520: class Program_weight_tensor_parameter_521: name = "parameter_521" + original_name = "batch_norm2d_42.w_2" shape = [192] dtype = "float32" min_val = float("0.00474286") @@ -5742,6 +6264,7 @@ class Program_weight_tensor_parameter_521: class Program_weight_tensor_parameter_522: name = "parameter_522" + original_name = "batch_norm2d_42.w_1" shape = [192] dtype = "float32" min_val = float("-0.102625") @@ -5753,6 +6276,7 @@ class Program_weight_tensor_parameter_522: class Program_weight_tensor_parameter_523: name = "parameter_523" + original_name = "conv2d_44.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.0712483") @@ -5764,6 +6288,7 @@ class Program_weight_tensor_parameter_523: class Program_weight_tensor_parameter_524: name = "parameter_524" + original_name = "batch_norm2d_41.b_0" shape = [192] dtype = "float32" min_val = float("-1.87984") @@ -5775,6 +6300,7 @@ class Program_weight_tensor_parameter_524: class Program_weight_tensor_parameter_525: name = "parameter_525" + original_name = "batch_norm2d_41.w_0" shape = [192] dtype = "float32" min_val = float("0.790161") @@ -5786,6 +6312,7 @@ class Program_weight_tensor_parameter_525: class Program_weight_tensor_parameter_526: name = "parameter_526" + original_name = "batch_norm2d_41.w_2" shape = [192] dtype = "float32" min_val = float("0.0175701") @@ -5797,6 +6324,7 @@ class Program_weight_tensor_parameter_526: class Program_weight_tensor_parameter_527: name = "parameter_527" + original_name = "batch_norm2d_41.w_1" shape = [192] dtype = "float32" min_val = float("-0.857673") @@ -5808,6 +6336,7 @@ class Program_weight_tensor_parameter_527: class Program_weight_tensor_parameter_528: name = "parameter_528" + original_name = "conv2d_43.w_0" shape = [192, 192, 3, 3] dtype = "float32" min_val = float("-0.0680887") @@ -5819,6 +6348,7 @@ class Program_weight_tensor_parameter_528: class Program_weight_tensor_parameter_529: name = "parameter_529" + original_name = "batch_norm2d_40.b_0" shape = [192] dtype = "float32" min_val = float("-2.86208") @@ -5830,6 +6360,7 @@ class Program_weight_tensor_parameter_529: class Program_weight_tensor_parameter_530: name = "parameter_530" + original_name = "batch_norm2d_40.w_0" shape = [192] dtype = "float32" min_val = float("0.490153") @@ -5841,6 +6372,7 @@ class Program_weight_tensor_parameter_530: class Program_weight_tensor_parameter_531: name = "parameter_531" + original_name = "batch_norm2d_40.w_2" shape = [192] dtype = "float32" min_val = float("0.0121565") @@ -5852,6 +6384,7 @@ class Program_weight_tensor_parameter_531: class Program_weight_tensor_parameter_532: name = "parameter_532" + original_name = "batch_norm2d_40.w_1" shape = [192] dtype = "float32" min_val = float("-0.232643") @@ -5863,6 +6396,7 @@ class Program_weight_tensor_parameter_532: class Program_weight_tensor_parameter_533: name = "parameter_533" + original_name = "conv2d_42.w_0" shape = [192, 384, 1, 1] dtype = "float32" min_val = float("-0.112904") @@ -5874,6 +6408,7 @@ class Program_weight_tensor_parameter_533: class Program_weight_tensor_parameter_534: name = "parameter_534" + original_name = "batch_norm2d_39.b_0" shape = [192] dtype = "float32" min_val = float("-2.96795") @@ -5885,6 +6420,7 @@ class Program_weight_tensor_parameter_534: class Program_weight_tensor_parameter_535: name = "parameter_535" + original_name = "batch_norm2d_39.w_0" shape = [192] dtype = "float32" min_val = float("0.830405") @@ -5896,6 +6432,7 @@ class Program_weight_tensor_parameter_535: class Program_weight_tensor_parameter_536: name = "parameter_536" + original_name = "batch_norm2d_39.w_2" shape = [192] dtype = "float32" min_val = float("0.00635322") @@ -5907,6 +6444,7 @@ class Program_weight_tensor_parameter_536: class Program_weight_tensor_parameter_537: name = "parameter_537" + original_name = "batch_norm2d_39.w_1" shape = [192] dtype = "float32" min_val = float("-0.144806") @@ -5918,6 +6456,7 @@ class Program_weight_tensor_parameter_537: class Program_weight_tensor_parameter_538: name = "parameter_538" + original_name = "conv2d_41.w_0" shape = [192, 384, 1, 1] dtype = "float32" min_val = float("-0.100414") @@ -5929,6 +6468,7 @@ class Program_weight_tensor_parameter_538: class Program_weight_tensor_parameter_539: name = "parameter_539" + original_name = "batch_norm2d_38.b_0" shape = [384] dtype = "float32" min_val = float("-2.9234") @@ -5940,6 +6480,7 @@ class Program_weight_tensor_parameter_539: class Program_weight_tensor_parameter_540: name = "parameter_540" + original_name = "batch_norm2d_38.w_0" shape = [384] dtype = "float32" min_val = float("0.633896") @@ -5951,6 +6492,7 @@ class Program_weight_tensor_parameter_540: class Program_weight_tensor_parameter_541: name = "parameter_541" + original_name = "batch_norm2d_38.w_2" shape = [384] dtype = "float32" min_val = float("0.0117343") @@ -5962,6 +6504,7 @@ class Program_weight_tensor_parameter_541: class Program_weight_tensor_parameter_542: name = "parameter_542" + original_name = "batch_norm2d_38.w_1" shape = [384] dtype = "float32" min_val = float("-0.269172") @@ -5973,6 +6516,7 @@ class Program_weight_tensor_parameter_542: class Program_weight_tensor_parameter_543: name = "parameter_543" + original_name = "conv2d_40.w_0" shape = [384, 256, 3, 3] dtype = "float32" min_val = float("-0.0777711") @@ -5984,6 +6528,7 @@ class Program_weight_tensor_parameter_543: class Program_weight_tensor_parameter_544: name = "parameter_544" + original_name = "batch_norm2d_37.b_0" shape = [256] dtype = "float32" min_val = float("-2.04675") @@ -5995,6 +6540,7 @@ class Program_weight_tensor_parameter_544: class Program_weight_tensor_parameter_545: name = "parameter_545" + original_name = "batch_norm2d_37.w_0" shape = [256] dtype = "float32" min_val = float("0.509654") @@ -6006,6 +6552,7 @@ class Program_weight_tensor_parameter_545: class Program_weight_tensor_parameter_546: name = "parameter_546" + original_name = "batch_norm2d_37.w_2" shape = [256] dtype = "float32" min_val = float("0.00164958") @@ -6017,6 +6564,7 @@ class Program_weight_tensor_parameter_546: class Program_weight_tensor_parameter_547: name = "parameter_547" + original_name = "batch_norm2d_37.w_1" shape = [256] dtype = "float32" min_val = float("-0.248048") @@ -6028,6 +6576,7 @@ class Program_weight_tensor_parameter_547: class Program_weight_tensor_parameter_548: name = "parameter_548" + original_name = "conv2d_39.w_0" shape = [256, 192, 1, 1] dtype = "float32" min_val = float("-0.211445") @@ -6039,6 +6588,7 @@ class Program_weight_tensor_parameter_548: class Program_weight_tensor_parameter_549: name = "parameter_549" + original_name = "conv2d_38.b_0" shape = [192] dtype = "float32" min_val = float("-0.0146056") @@ -6050,6 +6600,7 @@ class Program_weight_tensor_parameter_549: class Program_weight_tensor_parameter_550: name = "parameter_550" + original_name = "conv2d_38.w_0" shape = [192, 192, 1, 1] dtype = "float32" min_val = float("-0.340895") @@ -6061,6 +6612,7 @@ class Program_weight_tensor_parameter_550: class Program_weight_tensor_parameter_551: name = "parameter_551" + original_name = "batch_norm2d_36.b_0" shape = [96] dtype = "float32" min_val = float("-1.9141") @@ -6072,6 +6624,7 @@ class Program_weight_tensor_parameter_551: class Program_weight_tensor_parameter_552: name = "parameter_552" + original_name = "batch_norm2d_36.w_0" shape = [96] dtype = "float32" min_val = float("0.139627") @@ -6083,6 +6636,7 @@ class Program_weight_tensor_parameter_552: class Program_weight_tensor_parameter_553: name = "parameter_553" + original_name = "batch_norm2d_36.w_2" shape = [96] dtype = "float32" min_val = float("9.44925e-05") @@ -6094,6 +6648,7 @@ class Program_weight_tensor_parameter_553: class Program_weight_tensor_parameter_554: name = "parameter_554" + original_name = "batch_norm2d_36.w_1" shape = [96] dtype = "float32" min_val = float("-0.0508301") @@ -6105,6 +6660,7 @@ class Program_weight_tensor_parameter_554: class Program_weight_tensor_parameter_555: name = "parameter_555" + original_name = "conv2d_37.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.0529209") @@ -6116,6 +6672,7 @@ class Program_weight_tensor_parameter_555: class Program_weight_tensor_parameter_556: name = "parameter_556" + original_name = "batch_norm2d_35.b_0" shape = [96] dtype = "float32" min_val = float("-1.91385") @@ -6127,6 +6684,7 @@ class Program_weight_tensor_parameter_556: class Program_weight_tensor_parameter_557: name = "parameter_557" + original_name = "batch_norm2d_35.w_0" shape = [96] dtype = "float32" min_val = float("0.343945") @@ -6138,6 +6696,7 @@ class Program_weight_tensor_parameter_557: class Program_weight_tensor_parameter_558: name = "parameter_558" + original_name = "batch_norm2d_35.w_2" shape = [96] dtype = "float32" min_val = float("0.000831351") @@ -6149,6 +6708,7 @@ class Program_weight_tensor_parameter_558: class Program_weight_tensor_parameter_559: name = "parameter_559" + original_name = "batch_norm2d_35.w_1" shape = [96] dtype = "float32" min_val = float("-0.135563") @@ -6160,6 +6720,7 @@ class Program_weight_tensor_parameter_559: class Program_weight_tensor_parameter_560: name = "parameter_560" + original_name = "conv2d_36.w_0" shape = [96, 96, 3, 3] dtype = "float32" min_val = float("-0.0417476") @@ -6171,6 +6732,7 @@ class Program_weight_tensor_parameter_560: class Program_weight_tensor_parameter_561: name = "parameter_561" + original_name = "batch_norm2d_34.b_0" shape = [96] dtype = "float32" min_val = float("-2.46669") @@ -6182,6 +6744,7 @@ class Program_weight_tensor_parameter_561: class Program_weight_tensor_parameter_562: name = "parameter_562" + original_name = "batch_norm2d_34.w_0" shape = [96] dtype = "float32" min_val = float("0.540095") @@ -6193,6 +6756,7 @@ class Program_weight_tensor_parameter_562: class Program_weight_tensor_parameter_563: name = "parameter_563" + original_name = "batch_norm2d_34.w_2" shape = [96] dtype = "float32" min_val = float("0.0343629") @@ -6204,6 +6768,7 @@ class Program_weight_tensor_parameter_563: class Program_weight_tensor_parameter_564: name = "parameter_564" + original_name = "batch_norm2d_34.w_1" shape = [96] dtype = "float32" min_val = float("-2.59911") @@ -6215,6 +6780,7 @@ class Program_weight_tensor_parameter_564: class Program_weight_tensor_parameter_565: name = "parameter_565" + original_name = "conv2d_35.w_0" shape = [96, 96, 3, 3] dtype = "float32" min_val = float("-0.159603") @@ -6226,6 +6792,7 @@ class Program_weight_tensor_parameter_565: class Program_weight_tensor_parameter_566: name = "parameter_566" + original_name = "batch_norm2d_33.b_0" shape = [96] dtype = "float32" min_val = float("-1.38744") @@ -6237,6 +6804,7 @@ class Program_weight_tensor_parameter_566: class Program_weight_tensor_parameter_567: name = "parameter_567" + original_name = "batch_norm2d_33.w_0" shape = [96] dtype = "float32" min_val = float("0.0452771") @@ -6248,6 +6816,7 @@ class Program_weight_tensor_parameter_567: class Program_weight_tensor_parameter_568: name = "parameter_568" + original_name = "batch_norm2d_33.w_2" shape = [96] dtype = "float32" min_val = float("7.60148e-05") @@ -6259,6 +6828,7 @@ class Program_weight_tensor_parameter_568: class Program_weight_tensor_parameter_569: name = "parameter_569" + original_name = "batch_norm2d_33.w_1" shape = [96] dtype = "float32" min_val = float("-0.0497884") @@ -6270,6 +6840,7 @@ class Program_weight_tensor_parameter_569: class Program_weight_tensor_parameter_570: name = "parameter_570" + original_name = "conv2d_34.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.0484855") @@ -6281,6 +6852,7 @@ class Program_weight_tensor_parameter_570: class Program_weight_tensor_parameter_571: name = "parameter_571" + original_name = "batch_norm2d_32.b_0" shape = [96] dtype = "float32" min_val = float("-1.38716") @@ -6292,6 +6864,7 @@ class Program_weight_tensor_parameter_571: class Program_weight_tensor_parameter_572: name = "parameter_572" + original_name = "batch_norm2d_32.w_0" shape = [96] dtype = "float32" min_val = float("0.373276") @@ -6303,6 +6876,7 @@ class Program_weight_tensor_parameter_572: class Program_weight_tensor_parameter_573: name = "parameter_573" + original_name = "batch_norm2d_32.w_2" shape = [96] dtype = "float32" min_val = float("0.00302635") @@ -6314,6 +6888,7 @@ class Program_weight_tensor_parameter_573: class Program_weight_tensor_parameter_574: name = "parameter_574" + original_name = "batch_norm2d_32.w_1" shape = [96] dtype = "float32" min_val = float("-0.106151") @@ -6325,6 +6900,7 @@ class Program_weight_tensor_parameter_574: class Program_weight_tensor_parameter_575: name = "parameter_575" + original_name = "conv2d_33.w_0" shape = [96, 96, 3, 3] dtype = "float32" min_val = float("-0.0601192") @@ -6336,6 +6912,7 @@ class Program_weight_tensor_parameter_575: class Program_weight_tensor_parameter_576: name = "parameter_576" + original_name = "batch_norm2d_31.b_0" shape = [96] dtype = "float32" min_val = float("-3.32059") @@ -6347,6 +6924,7 @@ class Program_weight_tensor_parameter_576: class Program_weight_tensor_parameter_577: name = "parameter_577" + original_name = "batch_norm2d_31.w_0" shape = [96] dtype = "float32" min_val = float("0.470758") @@ -6358,6 +6936,7 @@ class Program_weight_tensor_parameter_577: class Program_weight_tensor_parameter_578: name = "parameter_578" + original_name = "batch_norm2d_31.w_2" shape = [96] dtype = "float32" min_val = float("0.0279788") @@ -6369,6 +6948,7 @@ class Program_weight_tensor_parameter_578: class Program_weight_tensor_parameter_579: name = "parameter_579" + original_name = "batch_norm2d_31.w_1" shape = [96] dtype = "float32" min_val = float("-1.05837") @@ -6380,6 +6960,7 @@ class Program_weight_tensor_parameter_579: class Program_weight_tensor_parameter_580: name = "parameter_580" + original_name = "conv2d_32.w_0" shape = [96, 96, 3, 3] dtype = "float32" min_val = float("-0.152735") @@ -6391,6 +6972,7 @@ class Program_weight_tensor_parameter_580: class Program_weight_tensor_parameter_581: name = "parameter_581" + original_name = "batch_norm2d_30.b_0" shape = [96] dtype = "float32" min_val = float("-1.24949") @@ -6402,6 +6984,7 @@ class Program_weight_tensor_parameter_581: class Program_weight_tensor_parameter_582: name = "parameter_582" + original_name = "batch_norm2d_30.w_0" shape = [96] dtype = "float32" min_val = float("0.0224878") @@ -6413,6 +6996,7 @@ class Program_weight_tensor_parameter_582: class Program_weight_tensor_parameter_583: name = "parameter_583" + original_name = "batch_norm2d_30.w_2" shape = [96] dtype = "float32" min_val = float("2.50107e-05") @@ -6424,6 +7008,7 @@ class Program_weight_tensor_parameter_583: class Program_weight_tensor_parameter_584: name = "parameter_584" + original_name = "batch_norm2d_30.w_1" shape = [96] dtype = "float32" min_val = float("-0.0398841") @@ -6435,6 +7020,7 @@ class Program_weight_tensor_parameter_584: class Program_weight_tensor_parameter_585: name = "parameter_585" + original_name = "conv2d_31.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.0406747") @@ -6446,6 +7032,7 @@ class Program_weight_tensor_parameter_585: class Program_weight_tensor_parameter_586: name = "parameter_586" + original_name = "batch_norm2d_29.b_0" shape = [96] dtype = "float32" min_val = float("-1.24929") @@ -6457,6 +7044,7 @@ class Program_weight_tensor_parameter_586: class Program_weight_tensor_parameter_587: name = "parameter_587" + original_name = "batch_norm2d_29.w_0" shape = [96] dtype = "float32" min_val = float("0.311326") @@ -6468,6 +7056,7 @@ class Program_weight_tensor_parameter_587: class Program_weight_tensor_parameter_588: name = "parameter_588" + original_name = "batch_norm2d_29.w_2" shape = [96] dtype = "float32" min_val = float("0.00299069") @@ -6479,6 +7068,7 @@ class Program_weight_tensor_parameter_588: class Program_weight_tensor_parameter_589: name = "parameter_589" + original_name = "batch_norm2d_29.w_1" shape = [96] dtype = "float32" min_val = float("-0.104806") @@ -6490,6 +7080,7 @@ class Program_weight_tensor_parameter_589: class Program_weight_tensor_parameter_590: name = "parameter_590" + original_name = "conv2d_30.w_0" shape = [96, 96, 3, 3] dtype = "float32" min_val = float("-0.0728298") @@ -6501,6 +7092,7 @@ class Program_weight_tensor_parameter_590: class Program_weight_tensor_parameter_591: name = "parameter_591" + original_name = "batch_norm2d_28.b_0" shape = [96] dtype = "float32" min_val = float("-3.5826") @@ -6512,6 +7104,7 @@ class Program_weight_tensor_parameter_591: class Program_weight_tensor_parameter_592: name = "parameter_592" + original_name = "batch_norm2d_28.w_0" shape = [96] dtype = "float32" min_val = float("0.511064") @@ -6523,6 +7116,7 @@ class Program_weight_tensor_parameter_592: class Program_weight_tensor_parameter_593: name = "parameter_593" + original_name = "batch_norm2d_28.w_2" shape = [96] dtype = "float32" min_val = float("0.021583") @@ -6534,6 +7128,7 @@ class Program_weight_tensor_parameter_593: class Program_weight_tensor_parameter_594: name = "parameter_594" + original_name = "batch_norm2d_28.w_1" shape = [96] dtype = "float32" min_val = float("-0.95654") @@ -6545,6 +7140,7 @@ class Program_weight_tensor_parameter_594: class Program_weight_tensor_parameter_595: name = "parameter_595" + original_name = "conv2d_29.w_0" shape = [96, 96, 3, 3] dtype = "float32" min_val = float("-0.0984925") @@ -6556,6 +7152,7 @@ class Program_weight_tensor_parameter_595: class Program_weight_tensor_parameter_596: name = "parameter_596" + original_name = "batch_norm2d_27.b_0" shape = [96] dtype = "float32" min_val = float("-0.891765") @@ -6567,6 +7164,7 @@ class Program_weight_tensor_parameter_596: class Program_weight_tensor_parameter_597: name = "parameter_597" + original_name = "batch_norm2d_27.w_0" shape = [96] dtype = "float32" min_val = float("0.0202036") @@ -6578,6 +7176,7 @@ class Program_weight_tensor_parameter_597: class Program_weight_tensor_parameter_598: name = "parameter_598" + original_name = "batch_norm2d_27.w_2" shape = [96] dtype = "float32" min_val = float("5.2419e-05") @@ -6589,6 +7188,7 @@ class Program_weight_tensor_parameter_598: class Program_weight_tensor_parameter_599: name = "parameter_599" + original_name = "batch_norm2d_27.w_1" shape = [96] dtype = "float32" min_val = float("-0.0353761") @@ -6600,6 +7200,7 @@ class Program_weight_tensor_parameter_599: class Program_weight_tensor_parameter_600: name = "parameter_600" + original_name = "conv2d_28.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.050403") @@ -6611,6 +7212,7 @@ class Program_weight_tensor_parameter_600: class Program_weight_tensor_parameter_601: name = "parameter_601" + original_name = "batch_norm2d_26.b_0" shape = [96] dtype = "float32" min_val = float("-0.891522") @@ -6622,6 +7224,7 @@ class Program_weight_tensor_parameter_601: class Program_weight_tensor_parameter_602: name = "parameter_602" + original_name = "batch_norm2d_26.w_0" shape = [96] dtype = "float32" min_val = float("0.170998") @@ -6633,6 +7236,7 @@ class Program_weight_tensor_parameter_602: class Program_weight_tensor_parameter_603: name = "parameter_603" + original_name = "batch_norm2d_26.w_2" shape = [96] dtype = "float32" min_val = float("0.00186209") @@ -6644,6 +7248,7 @@ class Program_weight_tensor_parameter_603: class Program_weight_tensor_parameter_604: name = "parameter_604" + original_name = "batch_norm2d_26.w_1" shape = [96] dtype = "float32" min_val = float("-0.0317255") @@ -6655,6 +7260,7 @@ class Program_weight_tensor_parameter_604: class Program_weight_tensor_parameter_605: name = "parameter_605" + original_name = "conv2d_27.w_0" shape = [96, 96, 3, 3] dtype = "float32" min_val = float("-0.0673552") @@ -6666,6 +7272,7 @@ class Program_weight_tensor_parameter_605: class Program_weight_tensor_parameter_606: name = "parameter_606" + original_name = "batch_norm2d_25.b_0" shape = [96] dtype = "float32" min_val = float("-2.65797") @@ -6677,6 +7284,7 @@ class Program_weight_tensor_parameter_606: class Program_weight_tensor_parameter_607: name = "parameter_607" + original_name = "batch_norm2d_25.w_0" shape = [96] dtype = "float32" min_val = float("0.510122") @@ -6688,6 +7296,7 @@ class Program_weight_tensor_parameter_607: class Program_weight_tensor_parameter_608: name = "parameter_608" + original_name = "batch_norm2d_25.w_2" shape = [96] dtype = "float32" min_val = float("0.0170441") @@ -6699,6 +7308,7 @@ class Program_weight_tensor_parameter_608: class Program_weight_tensor_parameter_609: name = "parameter_609" + original_name = "batch_norm2d_25.w_1" shape = [96] dtype = "float32" min_val = float("-0.802591") @@ -6710,6 +7320,7 @@ class Program_weight_tensor_parameter_609: class Program_weight_tensor_parameter_610: name = "parameter_610" + original_name = "conv2d_26.w_0" shape = [96, 96, 3, 3] dtype = "float32" min_val = float("-0.0799583") @@ -6721,6 +7332,7 @@ class Program_weight_tensor_parameter_610: class Program_weight_tensor_parameter_611: name = "parameter_611" + original_name = "batch_norm2d_24.b_0" shape = [96] dtype = "float32" min_val = float("-0.979363") @@ -6732,6 +7344,7 @@ class Program_weight_tensor_parameter_611: class Program_weight_tensor_parameter_612: name = "parameter_612" + original_name = "batch_norm2d_24.w_0" shape = [96] dtype = "float32" min_val = float("0.0499672") @@ -6743,6 +7356,7 @@ class Program_weight_tensor_parameter_612: class Program_weight_tensor_parameter_613: name = "parameter_613" + original_name = "batch_norm2d_24.w_2" shape = [96] dtype = "float32" min_val = float("0.000124848") @@ -6754,6 +7368,7 @@ class Program_weight_tensor_parameter_613: class Program_weight_tensor_parameter_614: name = "parameter_614" + original_name = "batch_norm2d_24.w_1" shape = [96] dtype = "float32" min_val = float("-0.0427797") @@ -6765,6 +7380,7 @@ class Program_weight_tensor_parameter_614: class Program_weight_tensor_parameter_615: name = "parameter_615" + original_name = "conv2d_25.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.0730409") @@ -6776,6 +7392,7 @@ class Program_weight_tensor_parameter_615: class Program_weight_tensor_parameter_616: name = "parameter_616" + original_name = "batch_norm2d_23.b_0" shape = [96] dtype = "float32" min_val = float("-0.979598") @@ -6787,6 +7404,7 @@ class Program_weight_tensor_parameter_616: class Program_weight_tensor_parameter_617: name = "parameter_617" + original_name = "batch_norm2d_23.w_0" shape = [96] dtype = "float32" min_val = float("0.240111") @@ -6798,6 +7416,7 @@ class Program_weight_tensor_parameter_617: class Program_weight_tensor_parameter_618: name = "parameter_618" + original_name = "batch_norm2d_23.w_2" shape = [96] dtype = "float32" min_val = float("0.00479228") @@ -6809,6 +7428,7 @@ class Program_weight_tensor_parameter_618: class Program_weight_tensor_parameter_619: name = "parameter_619" + original_name = "batch_norm2d_23.w_1" shape = [96] dtype = "float32" min_val = float("-0.0884025") @@ -6820,6 +7440,7 @@ class Program_weight_tensor_parameter_619: class Program_weight_tensor_parameter_620: name = "parameter_620" + original_name = "conv2d_24.w_0" shape = [96, 96, 3, 3] dtype = "float32" min_val = float("-0.070586") @@ -6831,6 +7452,7 @@ class Program_weight_tensor_parameter_620: class Program_weight_tensor_parameter_621: name = "parameter_621" + original_name = "batch_norm2d_22.b_0" shape = [96] dtype = "float32" min_val = float("-3.46749") @@ -6842,6 +7464,7 @@ class Program_weight_tensor_parameter_621: class Program_weight_tensor_parameter_622: name = "parameter_622" + original_name = "batch_norm2d_22.w_0" shape = [96] dtype = "float32" min_val = float("0.68469") @@ -6853,6 +7476,7 @@ class Program_weight_tensor_parameter_622: class Program_weight_tensor_parameter_623: name = "parameter_623" + original_name = "batch_norm2d_22.w_2" shape = [96] dtype = "float32" min_val = float("0.0126502") @@ -6864,6 +7488,7 @@ class Program_weight_tensor_parameter_623: class Program_weight_tensor_parameter_624: name = "parameter_624" + original_name = "batch_norm2d_22.w_1" shape = [96] dtype = "float32" min_val = float("-0.59646") @@ -6875,6 +7500,7 @@ class Program_weight_tensor_parameter_624: class Program_weight_tensor_parameter_625: name = "parameter_625" + original_name = "conv2d_23.w_0" shape = [96, 96, 3, 3] dtype = "float32" min_val = float("-0.0875016") @@ -6886,6 +7512,7 @@ class Program_weight_tensor_parameter_625: class Program_weight_tensor_parameter_626: name = "parameter_626" + original_name = "batch_norm2d_21.b_0" shape = [96] dtype = "float32" min_val = float("-0.623249") @@ -6897,6 +7524,7 @@ class Program_weight_tensor_parameter_626: class Program_weight_tensor_parameter_627: name = "parameter_627" + original_name = "batch_norm2d_21.w_0" shape = [96] dtype = "float32" min_val = float("0.0905173") @@ -6908,6 +7536,7 @@ class Program_weight_tensor_parameter_627: class Program_weight_tensor_parameter_628: name = "parameter_628" + original_name = "batch_norm2d_21.w_2" shape = [96] dtype = "float32" min_val = float("0.000486077") @@ -6919,6 +7548,7 @@ class Program_weight_tensor_parameter_628: class Program_weight_tensor_parameter_629: name = "parameter_629" + original_name = "batch_norm2d_21.w_1" shape = [96] dtype = "float32" min_val = float("-0.0378971") @@ -6930,6 +7560,7 @@ class Program_weight_tensor_parameter_629: class Program_weight_tensor_parameter_630: name = "parameter_630" + original_name = "conv2d_22.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.0967686") @@ -6941,6 +7572,7 @@ class Program_weight_tensor_parameter_630: class Program_weight_tensor_parameter_631: name = "parameter_631" + original_name = "batch_norm2d_20.b_0" shape = [96] dtype = "float32" min_val = float("-0.62253") @@ -6952,6 +7584,7 @@ class Program_weight_tensor_parameter_631: class Program_weight_tensor_parameter_632: name = "parameter_632" + original_name = "batch_norm2d_20.w_0" shape = [96] dtype = "float32" min_val = float("0.210918") @@ -6963,6 +7596,7 @@ class Program_weight_tensor_parameter_632: class Program_weight_tensor_parameter_633: name = "parameter_633" + original_name = "batch_norm2d_20.w_2" shape = [96] dtype = "float32" min_val = float("0.0110923") @@ -6974,6 +7608,7 @@ class Program_weight_tensor_parameter_633: class Program_weight_tensor_parameter_634: name = "parameter_634" + original_name = "batch_norm2d_20.w_1" shape = [96] dtype = "float32" min_val = float("-0.105783") @@ -6985,6 +7620,7 @@ class Program_weight_tensor_parameter_634: class Program_weight_tensor_parameter_635: name = "parameter_635" + original_name = "conv2d_21.w_0" shape = [96, 96, 3, 3] dtype = "float32" min_val = float("-0.0996365") @@ -6996,6 +7632,7 @@ class Program_weight_tensor_parameter_635: class Program_weight_tensor_parameter_636: name = "parameter_636" + original_name = "batch_norm2d_19.b_0" shape = [96] dtype = "float32" min_val = float("-2.4099") @@ -7007,6 +7644,7 @@ class Program_weight_tensor_parameter_636: class Program_weight_tensor_parameter_637: name = "parameter_637" + original_name = "batch_norm2d_19.w_0" shape = [96] dtype = "float32" min_val = float("0.855439") @@ -7018,6 +7656,7 @@ class Program_weight_tensor_parameter_637: class Program_weight_tensor_parameter_638: name = "parameter_638" + original_name = "batch_norm2d_19.w_2" shape = [96] dtype = "float32" min_val = float("0.0104439") @@ -7029,6 +7668,7 @@ class Program_weight_tensor_parameter_638: class Program_weight_tensor_parameter_639: name = "parameter_639" + original_name = "batch_norm2d_19.w_1" shape = [96] dtype = "float32" min_val = float("-0.780626") @@ -7040,6 +7680,7 @@ class Program_weight_tensor_parameter_639: class Program_weight_tensor_parameter_640: name = "parameter_640" + original_name = "conv2d_20.w_0" shape = [96, 96, 3, 3] dtype = "float32" min_val = float("-0.154701") @@ -7051,6 +7692,7 @@ class Program_weight_tensor_parameter_640: class Program_weight_tensor_parameter_641: name = "parameter_641" + original_name = "batch_norm2d_18.b_0" shape = [96] dtype = "float32" min_val = float("-3.15956") @@ -7062,6 +7704,7 @@ class Program_weight_tensor_parameter_641: class Program_weight_tensor_parameter_642: name = "parameter_642" + original_name = "batch_norm2d_18.w_0" shape = [96] dtype = "float32" min_val = float("0.209789") @@ -7073,6 +7716,7 @@ class Program_weight_tensor_parameter_642: class Program_weight_tensor_parameter_643: name = "parameter_643" + original_name = "batch_norm2d_18.w_2" shape = [96] dtype = "float32" min_val = float("0.00949005") @@ -7084,6 +7728,7 @@ class Program_weight_tensor_parameter_643: class Program_weight_tensor_parameter_644: name = "parameter_644" + original_name = "batch_norm2d_18.w_1" shape = [96] dtype = "float32" min_val = float("-0.272514") @@ -7095,6 +7740,7 @@ class Program_weight_tensor_parameter_644: class Program_weight_tensor_parameter_645: name = "parameter_645" + original_name = "conv2d_19.w_0" shape = [96, 192, 1, 1] dtype = "float32" min_val = float("-0.190092") @@ -7106,6 +7752,7 @@ class Program_weight_tensor_parameter_645: class Program_weight_tensor_parameter_646: name = "parameter_646" + original_name = "batch_norm2d_17.b_0" shape = [96] dtype = "float32" min_val = float("-4.92412") @@ -7117,6 +7764,7 @@ class Program_weight_tensor_parameter_646: class Program_weight_tensor_parameter_647: name = "parameter_647" + original_name = "batch_norm2d_17.w_0" shape = [96] dtype = "float32" min_val = float("0.411425") @@ -7128,6 +7776,7 @@ class Program_weight_tensor_parameter_647: class Program_weight_tensor_parameter_648: name = "parameter_648" + original_name = "batch_norm2d_17.w_2" shape = [96] dtype = "float32" min_val = float("0.00569395") @@ -7139,6 +7788,7 @@ class Program_weight_tensor_parameter_648: class Program_weight_tensor_parameter_649: name = "parameter_649" + original_name = "batch_norm2d_17.w_1" shape = [96] dtype = "float32" min_val = float("-0.123122") @@ -7150,6 +7800,7 @@ class Program_weight_tensor_parameter_649: class Program_weight_tensor_parameter_650: name = "parameter_650" + original_name = "conv2d_18.w_0" shape = [96, 192, 1, 1] dtype = "float32" min_val = float("-0.115428") @@ -7161,6 +7812,7 @@ class Program_weight_tensor_parameter_650: class Program_weight_tensor_parameter_651: name = "parameter_651" + original_name = "batch_norm2d_16.b_0" shape = [192] dtype = "float32" min_val = float("-2.27512") @@ -7172,6 +7824,7 @@ class Program_weight_tensor_parameter_651: class Program_weight_tensor_parameter_652: name = "parameter_652" + original_name = "batch_norm2d_16.w_0" shape = [192] dtype = "float32" min_val = float("0.632726") @@ -7183,6 +7836,7 @@ class Program_weight_tensor_parameter_652: class Program_weight_tensor_parameter_653: name = "parameter_653" + original_name = "batch_norm2d_16.w_2" shape = [192] dtype = "float32" min_val = float("0.0128887") @@ -7194,6 +7848,7 @@ class Program_weight_tensor_parameter_653: class Program_weight_tensor_parameter_654: name = "parameter_654" + original_name = "batch_norm2d_16.w_1" shape = [192] dtype = "float32" min_val = float("-0.476717") @@ -7205,6 +7860,7 @@ class Program_weight_tensor_parameter_654: class Program_weight_tensor_parameter_655: name = "parameter_655" + original_name = "conv2d_17.w_0" shape = [192, 128, 3, 3] dtype = "float32" min_val = float("-0.0811233") @@ -7216,6 +7872,7 @@ class Program_weight_tensor_parameter_655: class Program_weight_tensor_parameter_656: name = "parameter_656" + original_name = "batch_norm2d_15.b_0" shape = [128] dtype = "float32" min_val = float("-2.81253") @@ -7227,6 +7884,7 @@ class Program_weight_tensor_parameter_656: class Program_weight_tensor_parameter_657: name = "parameter_657" + original_name = "batch_norm2d_15.w_0" shape = [128] dtype = "float32" min_val = float("0.302011") @@ -7238,6 +7896,7 @@ class Program_weight_tensor_parameter_657: class Program_weight_tensor_parameter_658: name = "parameter_658" + original_name = "batch_norm2d_15.w_2" shape = [128] dtype = "float32" min_val = float("0.000683803") @@ -7249,6 +7908,7 @@ class Program_weight_tensor_parameter_658: class Program_weight_tensor_parameter_659: name = "parameter_659" + original_name = "batch_norm2d_15.w_1" shape = [128] dtype = "float32" min_val = float("-0.241007") @@ -7260,6 +7920,7 @@ class Program_weight_tensor_parameter_659: class Program_weight_tensor_parameter_660: name = "parameter_660" + original_name = "conv2d_16.w_0" shape = [128, 96, 1, 1] dtype = "float32" min_val = float("-0.16828") @@ -7271,6 +7932,7 @@ class Program_weight_tensor_parameter_660: class Program_weight_tensor_parameter_661: name = "parameter_661" + original_name = "conv2d_15.b_0" shape = [96] dtype = "float32" min_val = float("-0.0182017") @@ -7282,6 +7944,7 @@ class Program_weight_tensor_parameter_661: class Program_weight_tensor_parameter_662: name = "parameter_662" + original_name = "conv2d_15.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.297058") @@ -7293,6 +7956,7 @@ class Program_weight_tensor_parameter_662: class Program_weight_tensor_parameter_663: name = "parameter_663" + original_name = "batch_norm2d_14.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -7302,6 +7966,7 @@ class Program_weight_tensor_parameter_663: class Program_weight_tensor_parameter_664: name = "parameter_664" + original_name = "batch_norm2d_14.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -7311,6 +7976,7 @@ class Program_weight_tensor_parameter_664: class Program_weight_tensor_parameter_665: name = "parameter_665" + original_name = "batch_norm2d_14.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -7320,6 +7986,7 @@ class Program_weight_tensor_parameter_665: class Program_weight_tensor_parameter_666: name = "parameter_666" + original_name = "batch_norm2d_14.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -7329,6 +7996,7 @@ class Program_weight_tensor_parameter_666: class Program_weight_tensor_parameter_667: name = "parameter_667" + original_name = "conv2d_14.w_0" shape = [48, 48, 1, 1] dtype = "float32" min_val = float("-0.0524219") @@ -7340,6 +8008,7 @@ class Program_weight_tensor_parameter_667: class Program_weight_tensor_parameter_668: name = "parameter_668" + original_name = "batch_norm2d_13.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -7349,6 +8018,7 @@ class Program_weight_tensor_parameter_668: class Program_weight_tensor_parameter_669: name = "parameter_669" + original_name = "batch_norm2d_13.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -7358,6 +8028,7 @@ class Program_weight_tensor_parameter_669: class Program_weight_tensor_parameter_670: name = "parameter_670" + original_name = "batch_norm2d_13.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -7367,6 +8038,7 @@ class Program_weight_tensor_parameter_670: class Program_weight_tensor_parameter_671: name = "parameter_671" + original_name = "batch_norm2d_13.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -7376,6 +8048,7 @@ class Program_weight_tensor_parameter_671: class Program_weight_tensor_parameter_672: name = "parameter_672" + original_name = "conv2d_13.w_0" shape = [48, 48, 3, 3] dtype = "float32" min_val = float("-0.053396") @@ -7387,6 +8060,7 @@ class Program_weight_tensor_parameter_672: class Program_weight_tensor_parameter_673: name = "parameter_673" + original_name = "batch_norm2d_12.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -7396,6 +8070,7 @@ class Program_weight_tensor_parameter_673: class Program_weight_tensor_parameter_674: name = "parameter_674" + original_name = "batch_norm2d_12.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -7405,6 +8080,7 @@ class Program_weight_tensor_parameter_674: class Program_weight_tensor_parameter_675: name = "parameter_675" + original_name = "batch_norm2d_12.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -7414,6 +8090,7 @@ class Program_weight_tensor_parameter_675: class Program_weight_tensor_parameter_676: name = "parameter_676" + original_name = "batch_norm2d_12.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -7423,6 +8100,7 @@ class Program_weight_tensor_parameter_676: class Program_weight_tensor_parameter_677: name = "parameter_677" + original_name = "conv2d_12.w_0" shape = [48, 48, 3, 3] dtype = "float32" min_val = float("-0.0907736") @@ -7434,6 +8112,7 @@ class Program_weight_tensor_parameter_677: class Program_weight_tensor_parameter_678: name = "parameter_678" + original_name = "batch_norm2d_11.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -7443,6 +8122,7 @@ class Program_weight_tensor_parameter_678: class Program_weight_tensor_parameter_679: name = "parameter_679" + original_name = "batch_norm2d_11.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -7452,6 +8132,7 @@ class Program_weight_tensor_parameter_679: class Program_weight_tensor_parameter_680: name = "parameter_680" + original_name = "batch_norm2d_11.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -7461,6 +8142,7 @@ class Program_weight_tensor_parameter_680: class Program_weight_tensor_parameter_681: name = "parameter_681" + original_name = "batch_norm2d_11.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -7470,6 +8152,7 @@ class Program_weight_tensor_parameter_681: class Program_weight_tensor_parameter_682: name = "parameter_682" + original_name = "conv2d_11.w_0" shape = [48, 48, 1, 1] dtype = "float32" min_val = float("-0.0701343") @@ -7481,6 +8164,7 @@ class Program_weight_tensor_parameter_682: class Program_weight_tensor_parameter_683: name = "parameter_683" + original_name = "batch_norm2d_10.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -7490,6 +8174,7 @@ class Program_weight_tensor_parameter_683: class Program_weight_tensor_parameter_684: name = "parameter_684" + original_name = "batch_norm2d_10.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -7499,6 +8184,7 @@ class Program_weight_tensor_parameter_684: class Program_weight_tensor_parameter_685: name = "parameter_685" + original_name = "batch_norm2d_10.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -7508,6 +8194,7 @@ class Program_weight_tensor_parameter_685: class Program_weight_tensor_parameter_686: name = "parameter_686" + original_name = "batch_norm2d_10.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -7517,6 +8204,7 @@ class Program_weight_tensor_parameter_686: class Program_weight_tensor_parameter_687: name = "parameter_687" + original_name = "conv2d_10.w_0" shape = [48, 48, 3, 3] dtype = "float32" min_val = float("-0.0625249") @@ -7528,6 +8216,7 @@ class Program_weight_tensor_parameter_687: class Program_weight_tensor_parameter_688: name = "parameter_688" + original_name = "batch_norm2d_9.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -7537,6 +8226,7 @@ class Program_weight_tensor_parameter_688: class Program_weight_tensor_parameter_689: name = "parameter_689" + original_name = "batch_norm2d_9.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -7546,6 +8236,7 @@ class Program_weight_tensor_parameter_689: class Program_weight_tensor_parameter_690: name = "parameter_690" + original_name = "batch_norm2d_9.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -7555,6 +8246,7 @@ class Program_weight_tensor_parameter_690: class Program_weight_tensor_parameter_691: name = "parameter_691" + original_name = "batch_norm2d_9.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -7564,6 +8256,7 @@ class Program_weight_tensor_parameter_691: class Program_weight_tensor_parameter_692: name = "parameter_692" + original_name = "conv2d_9.w_0" shape = [48, 48, 3, 3] dtype = "float32" min_val = float("-0.105534") @@ -7575,6 +8268,7 @@ class Program_weight_tensor_parameter_692: class Program_weight_tensor_parameter_693: name = "parameter_693" + original_name = "batch_norm2d_8.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -7584,6 +8278,7 @@ class Program_weight_tensor_parameter_693: class Program_weight_tensor_parameter_694: name = "parameter_694" + original_name = "batch_norm2d_8.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -7593,6 +8288,7 @@ class Program_weight_tensor_parameter_694: class Program_weight_tensor_parameter_695: name = "parameter_695" + original_name = "batch_norm2d_8.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -7602,6 +8298,7 @@ class Program_weight_tensor_parameter_695: class Program_weight_tensor_parameter_696: name = "parameter_696" + original_name = "batch_norm2d_8.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -7611,6 +8308,7 @@ class Program_weight_tensor_parameter_696: class Program_weight_tensor_parameter_697: name = "parameter_697" + original_name = "conv2d_8.w_0" shape = [48, 48, 1, 1] dtype = "float32" min_val = float("-0.0927544") @@ -7622,6 +8320,7 @@ class Program_weight_tensor_parameter_697: class Program_weight_tensor_parameter_698: name = "parameter_698" + original_name = "batch_norm2d_7.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -7631,6 +8330,7 @@ class Program_weight_tensor_parameter_698: class Program_weight_tensor_parameter_699: name = "parameter_699" + original_name = "batch_norm2d_7.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -7640,6 +8340,7 @@ class Program_weight_tensor_parameter_699: class Program_weight_tensor_parameter_700: name = "parameter_700" + original_name = "batch_norm2d_7.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -7649,6 +8350,7 @@ class Program_weight_tensor_parameter_700: class Program_weight_tensor_parameter_701: name = "parameter_701" + original_name = "batch_norm2d_7.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -7658,6 +8360,7 @@ class Program_weight_tensor_parameter_701: class Program_weight_tensor_parameter_702: name = "parameter_702" + original_name = "conv2d_7.w_0" shape = [48, 48, 3, 3] dtype = "float32" min_val = float("-0.0662936") @@ -7669,6 +8372,7 @@ class Program_weight_tensor_parameter_702: class Program_weight_tensor_parameter_703: name = "parameter_703" + original_name = "batch_norm2d_6.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -7678,6 +8382,7 @@ class Program_weight_tensor_parameter_703: class Program_weight_tensor_parameter_704: name = "parameter_704" + original_name = "batch_norm2d_6.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -7687,6 +8392,7 @@ class Program_weight_tensor_parameter_704: class Program_weight_tensor_parameter_705: name = "parameter_705" + original_name = "batch_norm2d_6.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -7696,6 +8402,7 @@ class Program_weight_tensor_parameter_705: class Program_weight_tensor_parameter_706: name = "parameter_706" + original_name = "batch_norm2d_6.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -7705,6 +8412,7 @@ class Program_weight_tensor_parameter_706: class Program_weight_tensor_parameter_707: name = "parameter_707" + original_name = "conv2d_6.w_0" shape = [48, 48, 3, 3] dtype = "float32" min_val = float("-0.115861") @@ -7716,6 +8424,7 @@ class Program_weight_tensor_parameter_707: class Program_weight_tensor_parameter_708: name = "parameter_708" + original_name = "batch_norm2d_5.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -7725,6 +8434,7 @@ class Program_weight_tensor_parameter_708: class Program_weight_tensor_parameter_709: name = "parameter_709" + original_name = "batch_norm2d_5.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -7734,6 +8444,7 @@ class Program_weight_tensor_parameter_709: class Program_weight_tensor_parameter_710: name = "parameter_710" + original_name = "batch_norm2d_5.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -7743,6 +8454,7 @@ class Program_weight_tensor_parameter_710: class Program_weight_tensor_parameter_711: name = "parameter_711" + original_name = "batch_norm2d_5.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -7752,6 +8464,7 @@ class Program_weight_tensor_parameter_711: class Program_weight_tensor_parameter_712: name = "parameter_712" + original_name = "conv2d_5.w_0" shape = [48, 96, 1, 1] dtype = "float32" min_val = float("-0.156722") @@ -7763,6 +8476,7 @@ class Program_weight_tensor_parameter_712: class Program_weight_tensor_parameter_713: name = "parameter_713" + original_name = "batch_norm2d_4.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -7772,6 +8486,7 @@ class Program_weight_tensor_parameter_713: class Program_weight_tensor_parameter_714: name = "parameter_714" + original_name = "batch_norm2d_4.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -7781,6 +8496,7 @@ class Program_weight_tensor_parameter_714: class Program_weight_tensor_parameter_715: name = "parameter_715" + original_name = "batch_norm2d_4.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -7790,6 +8506,7 @@ class Program_weight_tensor_parameter_715: class Program_weight_tensor_parameter_716: name = "parameter_716" + original_name = "batch_norm2d_4.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -7799,6 +8516,7 @@ class Program_weight_tensor_parameter_716: class Program_weight_tensor_parameter_717: name = "parameter_717" + original_name = "conv2d_4.w_0" shape = [48, 96, 1, 1] dtype = "float32" min_val = float("-0.133366") @@ -7810,6 +8528,7 @@ class Program_weight_tensor_parameter_717: class Program_weight_tensor_parameter_718: name = "parameter_718" + original_name = "batch_norm2d_3.b_0" shape = [96] dtype = "float32" min_val = float("-3.40388") @@ -7821,6 +8540,7 @@ class Program_weight_tensor_parameter_718: class Program_weight_tensor_parameter_719: name = "parameter_719" + original_name = "batch_norm2d_3.w_0" shape = [96] dtype = "float32" min_val = float("0.861639") @@ -7832,6 +8552,7 @@ class Program_weight_tensor_parameter_719: class Program_weight_tensor_parameter_720: name = "parameter_720" + original_name = "batch_norm2d_3.w_2" shape = [96] dtype = "float32" min_val = float("0.68512") @@ -7843,6 +8564,7 @@ class Program_weight_tensor_parameter_720: class Program_weight_tensor_parameter_721: name = "parameter_721" + original_name = "batch_norm2d_3.w_1" shape = [96] dtype = "float32" min_val = float("-1.44893") @@ -7854,6 +8576,7 @@ class Program_weight_tensor_parameter_721: class Program_weight_tensor_parameter_722: name = "parameter_722" + original_name = "conv2d_3.w_0" shape = [96, 64, 3, 3] dtype = "float32" min_val = float("-0.115845") @@ -7865,6 +8588,7 @@ class Program_weight_tensor_parameter_722: class Program_weight_tensor_parameter_723: name = "parameter_723" + original_name = "batch_norm2d_2.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -7874,6 +8598,7 @@ class Program_weight_tensor_parameter_723: class Program_weight_tensor_parameter_724: name = "parameter_724" + original_name = "batch_norm2d_2.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -7883,6 +8608,7 @@ class Program_weight_tensor_parameter_724: class Program_weight_tensor_parameter_725: name = "parameter_725" + original_name = "batch_norm2d_2.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -7892,6 +8618,7 @@ class Program_weight_tensor_parameter_725: class Program_weight_tensor_parameter_726: name = "parameter_726" + original_name = "batch_norm2d_2.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -7901,6 +8628,7 @@ class Program_weight_tensor_parameter_726: class Program_weight_tensor_parameter_727: name = "parameter_727" + original_name = "conv2d_2.w_0" shape = [64, 32, 3, 3] dtype = "float32" min_val = float("-0.153743") @@ -7912,6 +8640,7 @@ class Program_weight_tensor_parameter_727: class Program_weight_tensor_parameter_728: name = "parameter_728" + original_name = "batch_norm2d_1.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -7921,6 +8650,7 @@ class Program_weight_tensor_parameter_728: class Program_weight_tensor_parameter_729: name = "parameter_729" + original_name = "batch_norm2d_1.w_0" shape = [32] dtype = "float32" min_val = float("0") @@ -7930,6 +8660,7 @@ class Program_weight_tensor_parameter_729: class Program_weight_tensor_parameter_730: name = "parameter_730" + original_name = "batch_norm2d_1.w_2" shape = [32] dtype = "float32" min_val = float("0") @@ -7939,6 +8670,7 @@ class Program_weight_tensor_parameter_730: class Program_weight_tensor_parameter_731: name = "parameter_731" + original_name = "batch_norm2d_1.w_1" shape = [32] dtype = "float32" min_val = float("0") @@ -7948,6 +8680,7 @@ class Program_weight_tensor_parameter_731: class Program_weight_tensor_parameter_732: name = "parameter_732" + original_name = "conv2d_1.w_0" shape = [32, 32, 3, 3] dtype = "float32" min_val = float("-0.307002") @@ -7959,6 +8692,7 @@ class Program_weight_tensor_parameter_732: class Program_weight_tensor_parameter_733: name = "parameter_733" + original_name = "batch_norm2d_0.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -7968,6 +8702,7 @@ class Program_weight_tensor_parameter_733: class Program_weight_tensor_parameter_734: name = "parameter_734" + original_name = "batch_norm2d_0.w_0" shape = [32] dtype = "float32" min_val = float("0") @@ -7977,6 +8712,7 @@ class Program_weight_tensor_parameter_734: class Program_weight_tensor_parameter_735: name = "parameter_735" + original_name = "batch_norm2d_0.w_2" shape = [32] dtype = "float32" min_val = float("0") @@ -7986,6 +8722,7 @@ class Program_weight_tensor_parameter_735: class Program_weight_tensor_parameter_736: name = "parameter_736" + original_name = "batch_norm2d_0.w_1" shape = [32] dtype = "float32" min_val = float("0") @@ -7995,6 +8732,7 @@ class Program_weight_tensor_parameter_736: class Program_weight_tensor_parameter_737: name = "parameter_737" + original_name = "conv2d_0.w_0" shape = [32, 3, 3, 3] dtype = "float32" min_val = float("-0.297631") diff --git a/paddle_samples/PaddleX/PicoDet-L_layout_17cls/subgraph_0/input_meta.py b/paddle_samples/PaddleX/PicoDet-L_layout_17cls/subgraph_0/input_meta.py index a6d4643da..7a209db6f 100644 --- a/paddle_samples/PaddleX/PicoDet-L_layout_17cls/subgraph_0/input_meta.py +++ b/paddle_samples/PaddleX/PicoDet-L_layout_17cls/subgraph_0/input_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_data_0: name = "data_0" + original_name = "var_658" shape = [2, 3, 576, 576] dtype = "float32" min_val = float("-2.1179") @@ -11,6 +12,7 @@ class Program_weight_tensor_data_0: class Program_weight_tensor_data_1: name = "data_1" + original_name = "var_907" shape = [8] dtype = "float32" data = [0.0, 1.00025, 2.00049, 3.00076, 4.00099, 5.00125, 6.00153, 7.00177] diff --git a/paddle_samples/PaddleX/PicoDet-L_layout_17cls/subgraph_0/weight_meta.py b/paddle_samples/PaddleX/PicoDet-L_layout_17cls/subgraph_0/weight_meta.py index 5ae979a14..ab1d539c8 100644 --- a/paddle_samples/PaddleX/PicoDet-L_layout_17cls/subgraph_0/weight_meta.py +++ b/paddle_samples/PaddleX/PicoDet-L_layout_17cls/subgraph_0/weight_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_parameter_0: name = "parameter_0" + original_name = "batch_norm2d_97.b_0" shape = [1] dtype = "float32" min_val = float("0") @@ -9,6 +10,7 @@ class Program_weight_tensor_parameter_0: class Program_weight_tensor_parameter_1: name = "parameter_1" + original_name = "batch_norm2d_97.w_0" shape = [1] dtype = "float32" min_val = float("0") @@ -18,6 +20,7 @@ class Program_weight_tensor_parameter_1: class Program_weight_tensor_parameter_2: name = "parameter_2" + original_name = "batch_norm2d_97.w_2" shape = [1] dtype = "float32" min_val = float("0") @@ -27,6 +30,7 @@ class Program_weight_tensor_parameter_2: class Program_weight_tensor_parameter_3: name = "parameter_3" + original_name = "batch_norm2d_97.w_1" shape = [1] dtype = "float32" min_val = float("0") @@ -36,6 +40,7 @@ class Program_weight_tensor_parameter_3: class Program_weight_tensor_parameter_4: name = "parameter_4" + original_name = "conv2d_115.w_0" shape = [1, 1, 1, 1] dtype = "float32" min_val = float("0") @@ -45,6 +50,7 @@ class Program_weight_tensor_parameter_4: class Program_weight_tensor_parameter_5: name = "parameter_5" + original_name = "batch_norm2d_96.b_0" shape = [1] dtype = "float32" min_val = float("0") @@ -54,6 +60,7 @@ class Program_weight_tensor_parameter_5: class Program_weight_tensor_parameter_6: name = "parameter_6" + original_name = "batch_norm2d_96.w_0" shape = [1] dtype = "float32" min_val = float("0") @@ -63,6 +70,7 @@ class Program_weight_tensor_parameter_6: class Program_weight_tensor_parameter_7: name = "parameter_7" + original_name = "batch_norm2d_96.w_2" shape = [1] dtype = "float32" min_val = float("0") @@ -72,6 +80,7 @@ class Program_weight_tensor_parameter_7: class Program_weight_tensor_parameter_8: name = "parameter_8" + original_name = "batch_norm2d_96.w_1" shape = [1] dtype = "float32" min_val = float("0") @@ -81,6 +90,7 @@ class Program_weight_tensor_parameter_8: class Program_weight_tensor_parameter_9: name = "parameter_9" + original_name = "conv2d_114.w_0" shape = [1, 160, 5, 5] dtype = "float32" min_val = float("-0.600657") @@ -92,6 +102,7 @@ class Program_weight_tensor_parameter_9: class Program_weight_tensor_parameter_10: name = "parameter_10" + original_name = "conv2d_113.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -101,6 +112,7 @@ class Program_weight_tensor_parameter_10: class Program_weight_tensor_parameter_11: name = "parameter_11" + original_name = "conv2d_113.w_0" shape = [32, 160, 1, 1] dtype = "float32" min_val = float("-0.365975") @@ -112,6 +124,7 @@ class Program_weight_tensor_parameter_11: class Program_weight_tensor_parameter_12: name = "parameter_12" + original_name = "conv2d_112.b_0" shape = [11] dtype = "float32" min_val = float("0") @@ -121,6 +134,7 @@ class Program_weight_tensor_parameter_12: class Program_weight_tensor_parameter_13: name = "parameter_13" + original_name = "conv2d_112.w_0" shape = [11, 160, 1, 1] dtype = "float32" min_val = float("-0.0347038") @@ -132,6 +146,7 @@ class Program_weight_tensor_parameter_13: class Program_weight_tensor_parameter_14: name = "parameter_14" + original_name = "batch_norm2d_89.b_0" shape = [160] dtype = "float32" min_val = float("0.0428933") @@ -143,6 +158,7 @@ class Program_weight_tensor_parameter_14: class Program_weight_tensor_parameter_15: name = "parameter_15" + original_name = "batch_norm2d_89.w_0" shape = [160] dtype = "float32" min_val = float("1.00671") @@ -154,6 +170,7 @@ class Program_weight_tensor_parameter_15: class Program_weight_tensor_parameter_16: name = "parameter_16" + original_name = "batch_norm2d_89.w_2" shape = [160] dtype = "float32" min_val = float("0.00400654") @@ -165,6 +182,7 @@ class Program_weight_tensor_parameter_16: class Program_weight_tensor_parameter_17: name = "parameter_17" + original_name = "batch_norm2d_89.w_1" shape = [160] dtype = "float32" min_val = float("-0.215883") @@ -176,6 +194,7 @@ class Program_weight_tensor_parameter_17: class Program_weight_tensor_parameter_18: name = "parameter_18" + original_name = "conv2d_97.w_0" shape = [160, 160, 1, 1] dtype = "float32" min_val = float("-0.201925") @@ -187,6 +206,7 @@ class Program_weight_tensor_parameter_18: class Program_weight_tensor_parameter_19: name = "parameter_19" + original_name = "conv2d_96.b_0" shape = [160] dtype = "float32" min_val = float("-0.0283452") @@ -198,6 +218,7 @@ class Program_weight_tensor_parameter_19: class Program_weight_tensor_parameter_20: name = "parameter_20" + original_name = "conv2d_96.w_0" shape = [160, 160, 1, 1] dtype = "float32" min_val = float("-0.0356003") @@ -209,6 +230,7 @@ class Program_weight_tensor_parameter_20: class Program_weight_tensor_parameter_21: name = "parameter_21" + original_name = "batch_norm2d_88.b_0" shape = [160] dtype = "float32" min_val = float("-0.137252") @@ -220,6 +242,7 @@ class Program_weight_tensor_parameter_21: class Program_weight_tensor_parameter_22: name = "parameter_22" + original_name = "batch_norm2d_88.w_0" shape = [160] dtype = "float32" min_val = float("0.828274") @@ -231,6 +254,7 @@ class Program_weight_tensor_parameter_22: class Program_weight_tensor_parameter_23: name = "parameter_23" + original_name = "batch_norm2d_88.w_2" shape = [160] dtype = "float32" min_val = float("0.0538522") @@ -242,6 +266,7 @@ class Program_weight_tensor_parameter_23: class Program_weight_tensor_parameter_24: name = "parameter_24" + original_name = "batch_norm2d_88.w_1" shape = [160] dtype = "float32" min_val = float("-0.37207") @@ -253,6 +278,7 @@ class Program_weight_tensor_parameter_24: class Program_weight_tensor_parameter_25: name = "parameter_25" + original_name = "conv2d_95.w_0" shape = [160, 160, 1, 1] dtype = "float32" min_val = float("-0.307729") @@ -264,6 +290,7 @@ class Program_weight_tensor_parameter_25: class Program_weight_tensor_parameter_26: name = "parameter_26" + original_name = "batch_norm2d_87.b_0" shape = [160] dtype = "float32" min_val = float("-0.0835023") @@ -275,6 +302,7 @@ class Program_weight_tensor_parameter_26: class Program_weight_tensor_parameter_27: name = "parameter_27" + original_name = "batch_norm2d_87.w_0" shape = [160] dtype = "float32" min_val = float("0.803378") @@ -286,6 +314,7 @@ class Program_weight_tensor_parameter_27: class Program_weight_tensor_parameter_28: name = "parameter_28" + original_name = "batch_norm2d_87.w_2" shape = [160] dtype = "float32" min_val = float("0.00790948") @@ -297,6 +326,7 @@ class Program_weight_tensor_parameter_28: class Program_weight_tensor_parameter_29: name = "parameter_29" + original_name = "batch_norm2d_87.w_1" shape = [160] dtype = "float32" min_val = float("-0.127175") @@ -308,6 +338,7 @@ class Program_weight_tensor_parameter_29: class Program_weight_tensor_parameter_30: name = "parameter_30" + original_name = "conv2d_94.w_0" shape = [160, 1, 5, 5] dtype = "float32" min_val = float("-0.192778") @@ -319,6 +350,7 @@ class Program_weight_tensor_parameter_30: class Program_weight_tensor_parameter_31: name = "parameter_31" + original_name = "batch_norm2d_86.b_0" shape = [160] dtype = "float32" min_val = float("-0.164956") @@ -330,6 +362,7 @@ class Program_weight_tensor_parameter_31: class Program_weight_tensor_parameter_32: name = "parameter_32" + original_name = "batch_norm2d_86.w_0" shape = [160] dtype = "float32" min_val = float("0.932487") @@ -341,6 +374,7 @@ class Program_weight_tensor_parameter_32: class Program_weight_tensor_parameter_33: name = "parameter_33" + original_name = "batch_norm2d_86.w_2" shape = [160] dtype = "float32" min_val = float("0.0552324") @@ -352,6 +386,7 @@ class Program_weight_tensor_parameter_33: class Program_weight_tensor_parameter_34: name = "parameter_34" + original_name = "batch_norm2d_86.w_1" shape = [160] dtype = "float32" min_val = float("-0.329545") @@ -363,6 +398,7 @@ class Program_weight_tensor_parameter_34: class Program_weight_tensor_parameter_35: name = "parameter_35" + original_name = "conv2d_93.w_0" shape = [160, 160, 1, 1] dtype = "float32" min_val = float("-0.290404") @@ -374,6 +410,7 @@ class Program_weight_tensor_parameter_35: class Program_weight_tensor_parameter_36: name = "parameter_36" + original_name = "batch_norm2d_85.b_0" shape = [160] dtype = "float32" min_val = float("-0.043837") @@ -385,6 +422,7 @@ class Program_weight_tensor_parameter_36: class Program_weight_tensor_parameter_37: name = "parameter_37" + original_name = "batch_norm2d_85.w_0" shape = [160] dtype = "float32" min_val = float("0.847469") @@ -396,6 +434,7 @@ class Program_weight_tensor_parameter_37: class Program_weight_tensor_parameter_38: name = "parameter_38" + original_name = "batch_norm2d_85.w_2" shape = [160] dtype = "float32" min_val = float("0.00895768") @@ -407,6 +446,7 @@ class Program_weight_tensor_parameter_38: class Program_weight_tensor_parameter_39: name = "parameter_39" + original_name = "batch_norm2d_85.w_1" shape = [160] dtype = "float32" min_val = float("-0.136901") @@ -418,6 +458,7 @@ class Program_weight_tensor_parameter_39: class Program_weight_tensor_parameter_40: name = "parameter_40" + original_name = "conv2d_92.w_0" shape = [160, 1, 5, 5] dtype = "float32" min_val = float("-0.216146") @@ -429,6 +470,7 @@ class Program_weight_tensor_parameter_40: class Program_weight_tensor_parameter_41: name = "parameter_41" + original_name = "batch_norm2d_84.b_0" shape = [160] dtype = "float32" min_val = float("-0.0564147") @@ -440,6 +482,7 @@ class Program_weight_tensor_parameter_41: class Program_weight_tensor_parameter_42: name = "parameter_42" + original_name = "batch_norm2d_84.w_0" shape = [160] dtype = "float32" min_val = float("0.951573") @@ -451,6 +494,7 @@ class Program_weight_tensor_parameter_42: class Program_weight_tensor_parameter_43: name = "parameter_43" + original_name = "batch_norm2d_84.w_2" shape = [160] dtype = "float32" min_val = float("0.029289") @@ -462,6 +506,7 @@ class Program_weight_tensor_parameter_43: class Program_weight_tensor_parameter_44: name = "parameter_44" + original_name = "batch_norm2d_84.w_1" shape = [160] dtype = "float32" min_val = float("-0.506876") @@ -473,6 +518,7 @@ class Program_weight_tensor_parameter_44: class Program_weight_tensor_parameter_45: name = "parameter_45" + original_name = "conv2d_91.w_0" shape = [160, 160, 1, 1] dtype = "float32" min_val = float("-0.27225") @@ -484,6 +530,7 @@ class Program_weight_tensor_parameter_45: class Program_weight_tensor_parameter_46: name = "parameter_46" + original_name = "batch_norm2d_83.b_0" shape = [160] dtype = "float32" min_val = float("-0.0523509") @@ -495,6 +542,7 @@ class Program_weight_tensor_parameter_46: class Program_weight_tensor_parameter_47: name = "parameter_47" + original_name = "batch_norm2d_83.w_0" shape = [160] dtype = "float32" min_val = float("0.859555") @@ -506,6 +554,7 @@ class Program_weight_tensor_parameter_47: class Program_weight_tensor_parameter_48: name = "parameter_48" + original_name = "batch_norm2d_83.w_2" shape = [160] dtype = "float32" min_val = float("0.0077849") @@ -517,6 +566,7 @@ class Program_weight_tensor_parameter_48: class Program_weight_tensor_parameter_49: name = "parameter_49" + original_name = "batch_norm2d_83.w_1" shape = [160] dtype = "float32" min_val = float("-0.151267") @@ -528,6 +578,7 @@ class Program_weight_tensor_parameter_49: class Program_weight_tensor_parameter_50: name = "parameter_50" + original_name = "conv2d_90.w_0" shape = [160, 1, 5, 5] dtype = "float32" min_val = float("-0.185484") @@ -539,6 +590,7 @@ class Program_weight_tensor_parameter_50: class Program_weight_tensor_parameter_51: name = "parameter_51" + original_name = "batch_norm2d_82.b_0" shape = [160] dtype = "float32" min_val = float("-0.0460681") @@ -550,6 +602,7 @@ class Program_weight_tensor_parameter_51: class Program_weight_tensor_parameter_52: name = "parameter_52" + original_name = "batch_norm2d_82.w_0" shape = [160] dtype = "float32" min_val = float("0.96892") @@ -561,6 +614,7 @@ class Program_weight_tensor_parameter_52: class Program_weight_tensor_parameter_53: name = "parameter_53" + original_name = "batch_norm2d_82.w_2" shape = [160] dtype = "float32" min_val = float("0.023065") @@ -572,6 +626,7 @@ class Program_weight_tensor_parameter_53: class Program_weight_tensor_parameter_54: name = "parameter_54" + original_name = "batch_norm2d_82.w_1" shape = [160] dtype = "float32" min_val = float("-0.391033") @@ -583,6 +638,7 @@ class Program_weight_tensor_parameter_54: class Program_weight_tensor_parameter_55: name = "parameter_55" + original_name = "conv2d_89.w_0" shape = [160, 160, 1, 1] dtype = "float32" min_val = float("-0.240512") @@ -594,6 +650,7 @@ class Program_weight_tensor_parameter_55: class Program_weight_tensor_parameter_56: name = "parameter_56" + original_name = "batch_norm2d_81.b_0" shape = [160] dtype = "float32" min_val = float("-0.0267738") @@ -605,6 +662,7 @@ class Program_weight_tensor_parameter_56: class Program_weight_tensor_parameter_57: name = "parameter_57" + original_name = "batch_norm2d_81.w_0" shape = [160] dtype = "float32" min_val = float("0.880486") @@ -616,6 +674,7 @@ class Program_weight_tensor_parameter_57: class Program_weight_tensor_parameter_58: name = "parameter_58" + original_name = "batch_norm2d_81.w_2" shape = [160] dtype = "float32" min_val = float("0.00813666") @@ -627,6 +686,7 @@ class Program_weight_tensor_parameter_58: class Program_weight_tensor_parameter_59: name = "parameter_59" + original_name = "batch_norm2d_81.w_1" shape = [160] dtype = "float32" min_val = float("-0.1155") @@ -638,6 +698,7 @@ class Program_weight_tensor_parameter_59: class Program_weight_tensor_parameter_60: name = "parameter_60" + original_name = "conv2d_88.w_0" shape = [160, 1, 5, 5] dtype = "float32" min_val = float("-0.239544") @@ -649,6 +710,7 @@ class Program_weight_tensor_parameter_60: class Program_weight_tensor_parameter_61: name = "parameter_61" + original_name = "batch_norm2d_95.b_0" shape = [1] dtype = "float32" min_val = float("0") @@ -658,6 +720,7 @@ class Program_weight_tensor_parameter_61: class Program_weight_tensor_parameter_62: name = "parameter_62" + original_name = "batch_norm2d_95.w_0" shape = [1] dtype = "float32" min_val = float("0") @@ -667,6 +730,7 @@ class Program_weight_tensor_parameter_62: class Program_weight_tensor_parameter_63: name = "parameter_63" + original_name = "batch_norm2d_95.w_2" shape = [1] dtype = "float32" min_val = float("0") @@ -676,6 +740,7 @@ class Program_weight_tensor_parameter_63: class Program_weight_tensor_parameter_64: name = "parameter_64" + original_name = "batch_norm2d_95.w_1" shape = [1] dtype = "float32" min_val = float("0") @@ -685,6 +750,7 @@ class Program_weight_tensor_parameter_64: class Program_weight_tensor_parameter_65: name = "parameter_65" + original_name = "conv2d_111.w_0" shape = [1, 1, 1, 1] dtype = "float32" min_val = float("0") @@ -694,6 +760,7 @@ class Program_weight_tensor_parameter_65: class Program_weight_tensor_parameter_66: name = "parameter_66" + original_name = "batch_norm2d_94.b_0" shape = [1] dtype = "float32" min_val = float("0") @@ -703,6 +770,7 @@ class Program_weight_tensor_parameter_66: class Program_weight_tensor_parameter_67: name = "parameter_67" + original_name = "batch_norm2d_94.w_0" shape = [1] dtype = "float32" min_val = float("0") @@ -712,6 +780,7 @@ class Program_weight_tensor_parameter_67: class Program_weight_tensor_parameter_68: name = "parameter_68" + original_name = "batch_norm2d_94.w_2" shape = [1] dtype = "float32" min_val = float("0") @@ -721,6 +790,7 @@ class Program_weight_tensor_parameter_68: class Program_weight_tensor_parameter_69: name = "parameter_69" + original_name = "batch_norm2d_94.w_1" shape = [1] dtype = "float32" min_val = float("0") @@ -730,6 +800,7 @@ class Program_weight_tensor_parameter_69: class Program_weight_tensor_parameter_70: name = "parameter_70" + original_name = "conv2d_110.w_0" shape = [1, 160, 5, 5] dtype = "float32" min_val = float("-0.536144") @@ -741,6 +812,7 @@ class Program_weight_tensor_parameter_70: class Program_weight_tensor_parameter_71: name = "parameter_71" + original_name = "conv2d_109.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -750,6 +822,7 @@ class Program_weight_tensor_parameter_71: class Program_weight_tensor_parameter_72: name = "parameter_72" + original_name = "conv2d_109.w_0" shape = [32, 160, 1, 1] dtype = "float32" min_val = float("-0.427286") @@ -761,6 +834,7 @@ class Program_weight_tensor_parameter_72: class Program_weight_tensor_parameter_73: name = "parameter_73" + original_name = "conv2d_108.b_0" shape = [11] dtype = "float32" min_val = float("0") @@ -770,6 +844,7 @@ class Program_weight_tensor_parameter_73: class Program_weight_tensor_parameter_74: name = "parameter_74" + original_name = "conv2d_108.w_0" shape = [11, 160, 1, 1] dtype = "float32" min_val = float("-0.0351808") @@ -781,6 +856,7 @@ class Program_weight_tensor_parameter_74: class Program_weight_tensor_parameter_75: name = "parameter_75" + original_name = "batch_norm2d_80.b_0" shape = [160] dtype = "float32" min_val = float("-0.0346443") @@ -792,6 +868,7 @@ class Program_weight_tensor_parameter_75: class Program_weight_tensor_parameter_76: name = "parameter_76" + original_name = "batch_norm2d_80.w_0" shape = [160] dtype = "float32" min_val = float("0.907205") @@ -803,6 +880,7 @@ class Program_weight_tensor_parameter_76: class Program_weight_tensor_parameter_77: name = "parameter_77" + original_name = "batch_norm2d_80.w_2" shape = [160] dtype = "float32" min_val = float("0.0068585") @@ -814,6 +892,7 @@ class Program_weight_tensor_parameter_77: class Program_weight_tensor_parameter_78: name = "parameter_78" + original_name = "batch_norm2d_80.w_1" shape = [160] dtype = "float32" min_val = float("-0.222326") @@ -825,6 +904,7 @@ class Program_weight_tensor_parameter_78: class Program_weight_tensor_parameter_79: name = "parameter_79" + original_name = "conv2d_87.w_0" shape = [160, 160, 1, 1] dtype = "float32" min_val = float("-0.261689") @@ -836,6 +916,7 @@ class Program_weight_tensor_parameter_79: class Program_weight_tensor_parameter_80: name = "parameter_80" + original_name = "conv2d_86.b_0" shape = [160] dtype = "float32" min_val = float("-0.0552408") @@ -847,6 +928,7 @@ class Program_weight_tensor_parameter_80: class Program_weight_tensor_parameter_81: name = "parameter_81" + original_name = "conv2d_86.w_0" shape = [160, 160, 1, 1] dtype = "float32" min_val = float("-0.0766068") @@ -858,6 +940,7 @@ class Program_weight_tensor_parameter_81: class Program_weight_tensor_parameter_82: name = "parameter_82" + original_name = "batch_norm2d_79.b_0" shape = [160] dtype = "float32" min_val = float("-0.0966589") @@ -869,6 +952,7 @@ class Program_weight_tensor_parameter_82: class Program_weight_tensor_parameter_83: name = "parameter_83" + original_name = "batch_norm2d_79.w_0" shape = [160] dtype = "float32" min_val = float("0.84403") @@ -880,6 +964,7 @@ class Program_weight_tensor_parameter_83: class Program_weight_tensor_parameter_84: name = "parameter_84" + original_name = "batch_norm2d_79.w_2" shape = [160] dtype = "float32" min_val = float("0.0517221") @@ -891,6 +976,7 @@ class Program_weight_tensor_parameter_84: class Program_weight_tensor_parameter_85: name = "parameter_85" + original_name = "batch_norm2d_79.w_1" shape = [160] dtype = "float32" min_val = float("-0.343387") @@ -902,6 +988,7 @@ class Program_weight_tensor_parameter_85: class Program_weight_tensor_parameter_86: name = "parameter_86" + original_name = "conv2d_85.w_0" shape = [160, 160, 1, 1] dtype = "float32" min_val = float("-0.405545") @@ -913,6 +1000,7 @@ class Program_weight_tensor_parameter_86: class Program_weight_tensor_parameter_87: name = "parameter_87" + original_name = "batch_norm2d_78.b_0" shape = [160] dtype = "float32" min_val = float("-0.0526603") @@ -924,6 +1012,7 @@ class Program_weight_tensor_parameter_87: class Program_weight_tensor_parameter_88: name = "parameter_88" + original_name = "batch_norm2d_78.w_0" shape = [160] dtype = "float32" min_val = float("0.819372") @@ -935,6 +1024,7 @@ class Program_weight_tensor_parameter_88: class Program_weight_tensor_parameter_89: name = "parameter_89" + original_name = "batch_norm2d_78.w_2" shape = [160] dtype = "float32" min_val = float("0.0102514") @@ -946,6 +1036,7 @@ class Program_weight_tensor_parameter_89: class Program_weight_tensor_parameter_90: name = "parameter_90" + original_name = "batch_norm2d_78.w_1" shape = [160] dtype = "float32" min_val = float("-0.127795") @@ -957,6 +1048,7 @@ class Program_weight_tensor_parameter_90: class Program_weight_tensor_parameter_91: name = "parameter_91" + original_name = "conv2d_84.w_0" shape = [160, 1, 5, 5] dtype = "float32" min_val = float("-0.228746") @@ -968,6 +1060,7 @@ class Program_weight_tensor_parameter_91: class Program_weight_tensor_parameter_92: name = "parameter_92" + original_name = "batch_norm2d_77.b_0" shape = [160] dtype = "float32" min_val = float("-0.11581") @@ -979,6 +1072,7 @@ class Program_weight_tensor_parameter_92: class Program_weight_tensor_parameter_93: name = "parameter_93" + original_name = "batch_norm2d_77.w_0" shape = [160] dtype = "float32" min_val = float("0.893374") @@ -990,6 +1084,7 @@ class Program_weight_tensor_parameter_93: class Program_weight_tensor_parameter_94: name = "parameter_94" + original_name = "batch_norm2d_77.w_2" shape = [160] dtype = "float32" min_val = float("0.0517208") @@ -1001,6 +1096,7 @@ class Program_weight_tensor_parameter_94: class Program_weight_tensor_parameter_95: name = "parameter_95" + original_name = "batch_norm2d_77.w_1" shape = [160] dtype = "float32" min_val = float("-0.357514") @@ -1012,6 +1108,7 @@ class Program_weight_tensor_parameter_95: class Program_weight_tensor_parameter_96: name = "parameter_96" + original_name = "conv2d_83.w_0" shape = [160, 160, 1, 1] dtype = "float32" min_val = float("-0.258233") @@ -1023,6 +1120,7 @@ class Program_weight_tensor_parameter_96: class Program_weight_tensor_parameter_97: name = "parameter_97" + original_name = "batch_norm2d_76.b_0" shape = [160] dtype = "float32" min_val = float("-0.029207") @@ -1034,6 +1132,7 @@ class Program_weight_tensor_parameter_97: class Program_weight_tensor_parameter_98: name = "parameter_98" + original_name = "batch_norm2d_76.w_0" shape = [160] dtype = "float32" min_val = float("0.823624") @@ -1045,6 +1144,7 @@ class Program_weight_tensor_parameter_98: class Program_weight_tensor_parameter_99: name = "parameter_99" + original_name = "batch_norm2d_76.w_2" shape = [160] dtype = "float32" min_val = float("0.00796652") @@ -1056,6 +1156,7 @@ class Program_weight_tensor_parameter_99: class Program_weight_tensor_parameter_100: name = "parameter_100" + original_name = "batch_norm2d_76.w_1" shape = [160] dtype = "float32" min_val = float("-0.117892") @@ -1067,6 +1168,7 @@ class Program_weight_tensor_parameter_100: class Program_weight_tensor_parameter_101: name = "parameter_101" + original_name = "conv2d_82.w_0" shape = [160, 1, 5, 5] dtype = "float32" min_val = float("-0.275915") @@ -1078,6 +1180,7 @@ class Program_weight_tensor_parameter_101: class Program_weight_tensor_parameter_102: name = "parameter_102" + original_name = "batch_norm2d_75.b_0" shape = [160] dtype = "float32" min_val = float("-0.044958") @@ -1089,6 +1192,7 @@ class Program_weight_tensor_parameter_102: class Program_weight_tensor_parameter_103: name = "parameter_103" + original_name = "batch_norm2d_75.w_0" shape = [160] dtype = "float32" min_val = float("0.950211") @@ -1100,6 +1204,7 @@ class Program_weight_tensor_parameter_103: class Program_weight_tensor_parameter_104: name = "parameter_104" + original_name = "batch_norm2d_75.w_2" shape = [160] dtype = "float32" min_val = float("0.0307128") @@ -1111,6 +1216,7 @@ class Program_weight_tensor_parameter_104: class Program_weight_tensor_parameter_105: name = "parameter_105" + original_name = "batch_norm2d_75.w_1" shape = [160] dtype = "float32" min_val = float("-0.513411") @@ -1122,6 +1228,7 @@ class Program_weight_tensor_parameter_105: class Program_weight_tensor_parameter_106: name = "parameter_106" + original_name = "conv2d_81.w_0" shape = [160, 160, 1, 1] dtype = "float32" min_val = float("-0.271779") @@ -1133,6 +1240,7 @@ class Program_weight_tensor_parameter_106: class Program_weight_tensor_parameter_107: name = "parameter_107" + original_name = "batch_norm2d_74.b_0" shape = [160] dtype = "float32" min_val = float("-0.0244513") @@ -1144,6 +1252,7 @@ class Program_weight_tensor_parameter_107: class Program_weight_tensor_parameter_108: name = "parameter_108" + original_name = "batch_norm2d_74.w_0" shape = [160] dtype = "float32" min_val = float("0.855091") @@ -1155,6 +1264,7 @@ class Program_weight_tensor_parameter_108: class Program_weight_tensor_parameter_109: name = "parameter_109" + original_name = "batch_norm2d_74.w_2" shape = [160] dtype = "float32" min_val = float("0.00248565") @@ -1166,6 +1276,7 @@ class Program_weight_tensor_parameter_109: class Program_weight_tensor_parameter_110: name = "parameter_110" + original_name = "batch_norm2d_74.w_1" shape = [160] dtype = "float32" min_val = float("-0.136606") @@ -1177,6 +1288,7 @@ class Program_weight_tensor_parameter_110: class Program_weight_tensor_parameter_111: name = "parameter_111" + original_name = "conv2d_80.w_0" shape = [160, 1, 5, 5] dtype = "float32" min_val = float("-0.372922") @@ -1188,6 +1300,7 @@ class Program_weight_tensor_parameter_111: class Program_weight_tensor_parameter_112: name = "parameter_112" + original_name = "batch_norm2d_73.b_0" shape = [160] dtype = "float32" min_val = float("-0.0311749") @@ -1199,6 +1312,7 @@ class Program_weight_tensor_parameter_112: class Program_weight_tensor_parameter_113: name = "parameter_113" + original_name = "batch_norm2d_73.w_0" shape = [160] dtype = "float32" min_val = float("0.91028") @@ -1210,6 +1324,7 @@ class Program_weight_tensor_parameter_113: class Program_weight_tensor_parameter_114: name = "parameter_114" + original_name = "batch_norm2d_73.w_2" shape = [160] dtype = "float32" min_val = float("0.0144636") @@ -1221,6 +1336,7 @@ class Program_weight_tensor_parameter_114: class Program_weight_tensor_parameter_115: name = "parameter_115" + original_name = "batch_norm2d_73.w_1" shape = [160] dtype = "float32" min_val = float("-0.274015") @@ -1232,6 +1348,7 @@ class Program_weight_tensor_parameter_115: class Program_weight_tensor_parameter_116: name = "parameter_116" + original_name = "conv2d_79.w_0" shape = [160, 160, 1, 1] dtype = "float32" min_val = float("-0.222651") @@ -1243,6 +1360,7 @@ class Program_weight_tensor_parameter_116: class Program_weight_tensor_parameter_117: name = "parameter_117" + original_name = "batch_norm2d_72.b_0" shape = [160] dtype = "float32" min_val = float("-0.00309718") @@ -1254,6 +1372,7 @@ class Program_weight_tensor_parameter_117: class Program_weight_tensor_parameter_118: name = "parameter_118" + original_name = "batch_norm2d_72.w_0" shape = [160] dtype = "float32" min_val = float("0.796547") @@ -1265,6 +1384,7 @@ class Program_weight_tensor_parameter_118: class Program_weight_tensor_parameter_119: name = "parameter_119" + original_name = "batch_norm2d_72.w_2" shape = [160] dtype = "float32" min_val = float("0.00359146") @@ -1276,6 +1396,7 @@ class Program_weight_tensor_parameter_119: class Program_weight_tensor_parameter_120: name = "parameter_120" + original_name = "batch_norm2d_72.w_1" shape = [160] dtype = "float32" min_val = float("-0.0924018") @@ -1287,6 +1408,7 @@ class Program_weight_tensor_parameter_120: class Program_weight_tensor_parameter_121: name = "parameter_121" + original_name = "conv2d_78.w_0" shape = [160, 1, 5, 5] dtype = "float32" min_val = float("-0.296177") @@ -1298,6 +1420,7 @@ class Program_weight_tensor_parameter_121: class Program_weight_tensor_parameter_122: name = "parameter_122" + original_name = "batch_norm2d_93.b_0" shape = [1] dtype = "float32" min_val = float("0") @@ -1307,6 +1430,7 @@ class Program_weight_tensor_parameter_122: class Program_weight_tensor_parameter_123: name = "parameter_123" + original_name = "batch_norm2d_93.w_0" shape = [1] dtype = "float32" min_val = float("0") @@ -1316,6 +1440,7 @@ class Program_weight_tensor_parameter_123: class Program_weight_tensor_parameter_124: name = "parameter_124" + original_name = "batch_norm2d_93.w_2" shape = [1] dtype = "float32" min_val = float("0") @@ -1325,6 +1450,7 @@ class Program_weight_tensor_parameter_124: class Program_weight_tensor_parameter_125: name = "parameter_125" + original_name = "batch_norm2d_93.w_1" shape = [1] dtype = "float32" min_val = float("0") @@ -1334,6 +1460,7 @@ class Program_weight_tensor_parameter_125: class Program_weight_tensor_parameter_126: name = "parameter_126" + original_name = "conv2d_107.w_0" shape = [1, 1, 1, 1] dtype = "float32" min_val = float("0") @@ -1343,6 +1470,7 @@ class Program_weight_tensor_parameter_126: class Program_weight_tensor_parameter_127: name = "parameter_127" + original_name = "batch_norm2d_92.b_0" shape = [1] dtype = "float32" min_val = float("0") @@ -1352,6 +1480,7 @@ class Program_weight_tensor_parameter_127: class Program_weight_tensor_parameter_128: name = "parameter_128" + original_name = "batch_norm2d_92.w_0" shape = [1] dtype = "float32" min_val = float("0") @@ -1361,6 +1490,7 @@ class Program_weight_tensor_parameter_128: class Program_weight_tensor_parameter_129: name = "parameter_129" + original_name = "batch_norm2d_92.w_2" shape = [1] dtype = "float32" min_val = float("0") @@ -1370,6 +1500,7 @@ class Program_weight_tensor_parameter_129: class Program_weight_tensor_parameter_130: name = "parameter_130" + original_name = "batch_norm2d_92.w_1" shape = [1] dtype = "float32" min_val = float("0") @@ -1379,6 +1510,7 @@ class Program_weight_tensor_parameter_130: class Program_weight_tensor_parameter_131: name = "parameter_131" + original_name = "conv2d_106.w_0" shape = [1, 160, 5, 5] dtype = "float32" min_val = float("-0.2583") @@ -1390,6 +1522,7 @@ class Program_weight_tensor_parameter_131: class Program_weight_tensor_parameter_132: name = "parameter_132" + original_name = "conv2d_105.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -1399,6 +1532,7 @@ class Program_weight_tensor_parameter_132: class Program_weight_tensor_parameter_133: name = "parameter_133" + original_name = "conv2d_105.w_0" shape = [32, 160, 1, 1] dtype = "float32" min_val = float("-0.503078") @@ -1410,6 +1544,7 @@ class Program_weight_tensor_parameter_133: class Program_weight_tensor_parameter_134: name = "parameter_134" + original_name = "conv2d_104.b_0" shape = [11] dtype = "float32" min_val = float("0") @@ -1419,6 +1554,7 @@ class Program_weight_tensor_parameter_134: class Program_weight_tensor_parameter_135: name = "parameter_135" + original_name = "conv2d_104.w_0" shape = [11, 160, 1, 1] dtype = "float32" min_val = float("-0.0356441") @@ -1430,6 +1566,7 @@ class Program_weight_tensor_parameter_135: class Program_weight_tensor_parameter_136: name = "parameter_136" + original_name = "batch_norm2d_71.b_0" shape = [160] dtype = "float32" min_val = float("-0.099936") @@ -1441,6 +1578,7 @@ class Program_weight_tensor_parameter_136: class Program_weight_tensor_parameter_137: name = "parameter_137" + original_name = "batch_norm2d_71.w_0" shape = [160] dtype = "float32" min_val = float("0.77299") @@ -1452,6 +1590,7 @@ class Program_weight_tensor_parameter_137: class Program_weight_tensor_parameter_138: name = "parameter_138" + original_name = "batch_norm2d_71.w_2" shape = [160] dtype = "float32" min_val = float("0.00340977") @@ -1463,6 +1602,7 @@ class Program_weight_tensor_parameter_138: class Program_weight_tensor_parameter_139: name = "parameter_139" + original_name = "batch_norm2d_71.w_1" shape = [160] dtype = "float32" min_val = float("-0.27917") @@ -1474,6 +1614,7 @@ class Program_weight_tensor_parameter_139: class Program_weight_tensor_parameter_140: name = "parameter_140" + original_name = "conv2d_77.w_0" shape = [160, 160, 1, 1] dtype = "float32" min_val = float("-0.267203") @@ -1485,6 +1626,7 @@ class Program_weight_tensor_parameter_140: class Program_weight_tensor_parameter_141: name = "parameter_141" + original_name = "conv2d_76.b_0" shape = [160] dtype = "float32" min_val = float("-0.0846574") @@ -1496,6 +1638,7 @@ class Program_weight_tensor_parameter_141: class Program_weight_tensor_parameter_142: name = "parameter_142" + original_name = "conv2d_76.w_0" shape = [160, 160, 1, 1] dtype = "float32" min_val = float("-0.131849") @@ -1507,6 +1650,7 @@ class Program_weight_tensor_parameter_142: class Program_weight_tensor_parameter_143: name = "parameter_143" + original_name = "batch_norm2d_70.b_0" shape = [160] dtype = "float32" min_val = float("-0.0307217") @@ -1518,6 +1662,7 @@ class Program_weight_tensor_parameter_143: class Program_weight_tensor_parameter_144: name = "parameter_144" + original_name = "batch_norm2d_70.w_0" shape = [160] dtype = "float32" min_val = float("0.829736") @@ -1529,6 +1674,7 @@ class Program_weight_tensor_parameter_144: class Program_weight_tensor_parameter_145: name = "parameter_145" + original_name = "batch_norm2d_70.w_2" shape = [160] dtype = "float32" min_val = float("0.0431634") @@ -1540,6 +1686,7 @@ class Program_weight_tensor_parameter_145: class Program_weight_tensor_parameter_146: name = "parameter_146" + original_name = "batch_norm2d_70.w_1" shape = [160] dtype = "float32" min_val = float("-0.423523") @@ -1551,6 +1698,7 @@ class Program_weight_tensor_parameter_146: class Program_weight_tensor_parameter_147: name = "parameter_147" + original_name = "conv2d_75.w_0" shape = [160, 160, 1, 1] dtype = "float32" min_val = float("-0.281977") @@ -1562,6 +1710,7 @@ class Program_weight_tensor_parameter_147: class Program_weight_tensor_parameter_148: name = "parameter_148" + original_name = "batch_norm2d_69.b_0" shape = [160] dtype = "float32" min_val = float("-0.0267559") @@ -1573,6 +1722,7 @@ class Program_weight_tensor_parameter_148: class Program_weight_tensor_parameter_149: name = "parameter_149" + original_name = "batch_norm2d_69.w_0" shape = [160] dtype = "float32" min_val = float("0.827276") @@ -1584,6 +1734,7 @@ class Program_weight_tensor_parameter_149: class Program_weight_tensor_parameter_150: name = "parameter_150" + original_name = "batch_norm2d_69.w_2" shape = [160] dtype = "float32" min_val = float("0.00872923") @@ -1595,6 +1746,7 @@ class Program_weight_tensor_parameter_150: class Program_weight_tensor_parameter_151: name = "parameter_151" + original_name = "batch_norm2d_69.w_1" shape = [160] dtype = "float32" min_val = float("-0.113989") @@ -1606,6 +1758,7 @@ class Program_weight_tensor_parameter_151: class Program_weight_tensor_parameter_152: name = "parameter_152" + original_name = "conv2d_74.w_0" shape = [160, 1, 5, 5] dtype = "float32" min_val = float("-0.209792") @@ -1617,6 +1770,7 @@ class Program_weight_tensor_parameter_152: class Program_weight_tensor_parameter_153: name = "parameter_153" + original_name = "batch_norm2d_68.b_0" shape = [160] dtype = "float32" min_val = float("-0.118698") @@ -1628,6 +1782,7 @@ class Program_weight_tensor_parameter_153: class Program_weight_tensor_parameter_154: name = "parameter_154" + original_name = "batch_norm2d_68.w_0" shape = [160] dtype = "float32" min_val = float("0.865875") @@ -1639,6 +1794,7 @@ class Program_weight_tensor_parameter_154: class Program_weight_tensor_parameter_155: name = "parameter_155" + original_name = "batch_norm2d_68.w_2" shape = [160] dtype = "float32" min_val = float("0.0349899") @@ -1650,6 +1806,7 @@ class Program_weight_tensor_parameter_155: class Program_weight_tensor_parameter_156: name = "parameter_156" + original_name = "batch_norm2d_68.w_1" shape = [160] dtype = "float32" min_val = float("-0.536005") @@ -1661,6 +1818,7 @@ class Program_weight_tensor_parameter_156: class Program_weight_tensor_parameter_157: name = "parameter_157" + original_name = "conv2d_73.w_0" shape = [160, 160, 1, 1] dtype = "float32" min_val = float("-0.356653") @@ -1672,6 +1830,7 @@ class Program_weight_tensor_parameter_157: class Program_weight_tensor_parameter_158: name = "parameter_158" + original_name = "batch_norm2d_67.b_0" shape = [160] dtype = "float32" min_val = float("-0.0289716") @@ -1683,6 +1842,7 @@ class Program_weight_tensor_parameter_158: class Program_weight_tensor_parameter_159: name = "parameter_159" + original_name = "batch_norm2d_67.w_0" shape = [160] dtype = "float32" min_val = float("0.838799") @@ -1694,6 +1854,7 @@ class Program_weight_tensor_parameter_159: class Program_weight_tensor_parameter_160: name = "parameter_160" + original_name = "batch_norm2d_67.w_2" shape = [160] dtype = "float32" min_val = float("0.00206484") @@ -1705,6 +1866,7 @@ class Program_weight_tensor_parameter_160: class Program_weight_tensor_parameter_161: name = "parameter_161" + original_name = "batch_norm2d_67.w_1" shape = [160] dtype = "float32" min_val = float("-0.0973601") @@ -1716,6 +1878,7 @@ class Program_weight_tensor_parameter_161: class Program_weight_tensor_parameter_162: name = "parameter_162" + original_name = "conv2d_72.w_0" shape = [160, 1, 5, 5] dtype = "float32" min_val = float("-0.244048") @@ -1727,6 +1890,7 @@ class Program_weight_tensor_parameter_162: class Program_weight_tensor_parameter_163: name = "parameter_163" + original_name = "batch_norm2d_66.b_0" shape = [160] dtype = "float32" min_val = float("-0.0330543") @@ -1738,6 +1902,7 @@ class Program_weight_tensor_parameter_163: class Program_weight_tensor_parameter_164: name = "parameter_164" + original_name = "batch_norm2d_66.w_0" shape = [160] dtype = "float32" min_val = float("0.935846") @@ -1749,6 +1914,7 @@ class Program_weight_tensor_parameter_164: class Program_weight_tensor_parameter_165: name = "parameter_165" + original_name = "batch_norm2d_66.w_2" shape = [160] dtype = "float32" min_val = float("0.0460605") @@ -1760,6 +1926,7 @@ class Program_weight_tensor_parameter_165: class Program_weight_tensor_parameter_166: name = "parameter_166" + original_name = "batch_norm2d_66.w_1" shape = [160] dtype = "float32" min_val = float("-0.588218") @@ -1771,6 +1938,7 @@ class Program_weight_tensor_parameter_166: class Program_weight_tensor_parameter_167: name = "parameter_167" + original_name = "conv2d_71.w_0" shape = [160, 160, 1, 1] dtype = "float32" min_val = float("-0.315328") @@ -1782,6 +1950,7 @@ class Program_weight_tensor_parameter_167: class Program_weight_tensor_parameter_168: name = "parameter_168" + original_name = "batch_norm2d_65.b_0" shape = [160] dtype = "float32" min_val = float("-0.0269026") @@ -1793,6 +1962,7 @@ class Program_weight_tensor_parameter_168: class Program_weight_tensor_parameter_169: name = "parameter_169" + original_name = "batch_norm2d_65.w_0" shape = [160] dtype = "float32" min_val = float("0.841916") @@ -1804,6 +1974,7 @@ class Program_weight_tensor_parameter_169: class Program_weight_tensor_parameter_170: name = "parameter_170" + original_name = "batch_norm2d_65.w_2" shape = [160] dtype = "float32" min_val = float("0.00226701") @@ -1815,6 +1986,7 @@ class Program_weight_tensor_parameter_170: class Program_weight_tensor_parameter_171: name = "parameter_171" + original_name = "batch_norm2d_65.w_1" shape = [160] dtype = "float32" min_val = float("-0.107582") @@ -1826,6 +1998,7 @@ class Program_weight_tensor_parameter_171: class Program_weight_tensor_parameter_172: name = "parameter_172" + original_name = "conv2d_70.w_0" shape = [160, 1, 5, 5] dtype = "float32" min_val = float("-0.292634") @@ -1837,6 +2010,7 @@ class Program_weight_tensor_parameter_172: class Program_weight_tensor_parameter_173: name = "parameter_173" + original_name = "batch_norm2d_64.b_0" shape = [160] dtype = "float32" min_val = float("-0.04156") @@ -1848,6 +2022,7 @@ class Program_weight_tensor_parameter_173: class Program_weight_tensor_parameter_174: name = "parameter_174" + original_name = "batch_norm2d_64.w_0" shape = [160] dtype = "float32" min_val = float("0.912712") @@ -1859,6 +2034,7 @@ class Program_weight_tensor_parameter_174: class Program_weight_tensor_parameter_175: name = "parameter_175" + original_name = "batch_norm2d_64.w_2" shape = [160] dtype = "float32" min_val = float("0.0172795") @@ -1870,6 +2046,7 @@ class Program_weight_tensor_parameter_175: class Program_weight_tensor_parameter_176: name = "parameter_176" + original_name = "batch_norm2d_64.w_1" shape = [160] dtype = "float32" min_val = float("-0.37759") @@ -1881,6 +2058,7 @@ class Program_weight_tensor_parameter_176: class Program_weight_tensor_parameter_177: name = "parameter_177" + original_name = "conv2d_69.w_0" shape = [160, 160, 1, 1] dtype = "float32" min_val = float("-0.27331") @@ -1892,6 +2070,7 @@ class Program_weight_tensor_parameter_177: class Program_weight_tensor_parameter_178: name = "parameter_178" + original_name = "batch_norm2d_63.b_0" shape = [160] dtype = "float32" min_val = float("-0.0202709") @@ -1903,6 +2082,7 @@ class Program_weight_tensor_parameter_178: class Program_weight_tensor_parameter_179: name = "parameter_179" + original_name = "batch_norm2d_63.w_0" shape = [160] dtype = "float32" min_val = float("0.823561") @@ -1914,6 +2094,7 @@ class Program_weight_tensor_parameter_179: class Program_weight_tensor_parameter_180: name = "parameter_180" + original_name = "batch_norm2d_63.w_2" shape = [160] dtype = "float32" min_val = float("0.00166747") @@ -1925,6 +2106,7 @@ class Program_weight_tensor_parameter_180: class Program_weight_tensor_parameter_181: name = "parameter_181" + original_name = "batch_norm2d_63.w_1" shape = [160] dtype = "float32" min_val = float("-0.113725") @@ -1936,6 +2118,7 @@ class Program_weight_tensor_parameter_181: class Program_weight_tensor_parameter_182: name = "parameter_182" + original_name = "conv2d_68.w_0" shape = [160, 1, 5, 5] dtype = "float32" min_val = float("-0.231076") @@ -1947,6 +2130,7 @@ class Program_weight_tensor_parameter_182: class Program_weight_tensor_parameter_183: name = "parameter_183" + original_name = "batch_norm2d_91.b_0" shape = [1] dtype = "float32" min_val = float("0") @@ -1956,6 +2140,7 @@ class Program_weight_tensor_parameter_183: class Program_weight_tensor_parameter_184: name = "parameter_184" + original_name = "batch_norm2d_91.w_0" shape = [1] dtype = "float32" min_val = float("0") @@ -1965,6 +2150,7 @@ class Program_weight_tensor_parameter_184: class Program_weight_tensor_parameter_185: name = "parameter_185" + original_name = "batch_norm2d_91.w_2" shape = [1] dtype = "float32" min_val = float("0") @@ -1974,6 +2160,7 @@ class Program_weight_tensor_parameter_185: class Program_weight_tensor_parameter_186: name = "parameter_186" + original_name = "batch_norm2d_91.w_1" shape = [1] dtype = "float32" min_val = float("0") @@ -1983,6 +2170,7 @@ class Program_weight_tensor_parameter_186: class Program_weight_tensor_parameter_187: name = "parameter_187" + original_name = "conv2d_103.w_0" shape = [1, 1, 1, 1] dtype = "float32" min_val = float("0") @@ -1992,6 +2180,7 @@ class Program_weight_tensor_parameter_187: class Program_weight_tensor_parameter_188: name = "parameter_188" + original_name = "batch_norm2d_90.b_0" shape = [1] dtype = "float32" min_val = float("0") @@ -2001,6 +2190,7 @@ class Program_weight_tensor_parameter_188: class Program_weight_tensor_parameter_189: name = "parameter_189" + original_name = "batch_norm2d_90.w_0" shape = [1] dtype = "float32" min_val = float("0") @@ -2010,6 +2200,7 @@ class Program_weight_tensor_parameter_189: class Program_weight_tensor_parameter_190: name = "parameter_190" + original_name = "batch_norm2d_90.w_2" shape = [1] dtype = "float32" min_val = float("0") @@ -2019,6 +2210,7 @@ class Program_weight_tensor_parameter_190: class Program_weight_tensor_parameter_191: name = "parameter_191" + original_name = "batch_norm2d_90.w_1" shape = [1] dtype = "float32" min_val = float("0") @@ -2028,6 +2220,7 @@ class Program_weight_tensor_parameter_191: class Program_weight_tensor_parameter_192: name = "parameter_192" + original_name = "conv2d_102.w_0" shape = [1, 160, 5, 5] dtype = "float32" min_val = float("-0.422775") @@ -2039,6 +2232,7 @@ class Program_weight_tensor_parameter_192: class Program_weight_tensor_parameter_193: name = "parameter_193" + original_name = "conv2d_101.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -2048,6 +2242,7 @@ class Program_weight_tensor_parameter_193: class Program_weight_tensor_parameter_194: name = "parameter_194" + original_name = "conv2d_101.w_0" shape = [32, 160, 1, 1] dtype = "float32" min_val = float("-0.45371") @@ -2059,6 +2254,7 @@ class Program_weight_tensor_parameter_194: class Program_weight_tensor_parameter_195: name = "parameter_195" + original_name = "conv2d_100.b_0" shape = [11] dtype = "float32" min_val = float("0") @@ -2068,6 +2264,7 @@ class Program_weight_tensor_parameter_195: class Program_weight_tensor_parameter_196: name = "parameter_196" + original_name = "conv2d_100.w_0" shape = [11, 160, 1, 1] dtype = "float32" min_val = float("-0.0317066") @@ -2079,6 +2276,7 @@ class Program_weight_tensor_parameter_196: class Program_weight_tensor_parameter_197: name = "parameter_197" + original_name = "batch_norm2d_62.b_0" shape = [160] dtype = "float32" min_val = float("-0.0179209") @@ -2090,6 +2288,7 @@ class Program_weight_tensor_parameter_197: class Program_weight_tensor_parameter_198: name = "parameter_198" + original_name = "batch_norm2d_62.w_0" shape = [160] dtype = "float32" min_val = float("0.697781") @@ -2101,6 +2300,7 @@ class Program_weight_tensor_parameter_198: class Program_weight_tensor_parameter_199: name = "parameter_199" + original_name = "batch_norm2d_62.w_2" shape = [160] dtype = "float32" min_val = float("0.0109077") @@ -2112,6 +2312,7 @@ class Program_weight_tensor_parameter_199: class Program_weight_tensor_parameter_200: name = "parameter_200" + original_name = "batch_norm2d_62.w_1" shape = [160] dtype = "float32" min_val = float("-0.318866") @@ -2123,6 +2324,7 @@ class Program_weight_tensor_parameter_200: class Program_weight_tensor_parameter_201: name = "parameter_201" + original_name = "conv2d_67.w_0" shape = [160, 160, 1, 1] dtype = "float32" min_val = float("-0.427231") @@ -2134,6 +2336,7 @@ class Program_weight_tensor_parameter_201: class Program_weight_tensor_parameter_202: name = "parameter_202" + original_name = "conv2d_66.b_0" shape = [160] dtype = "float32" min_val = float("-0.106943") @@ -2145,6 +2348,7 @@ class Program_weight_tensor_parameter_202: class Program_weight_tensor_parameter_203: name = "parameter_203" + original_name = "conv2d_66.w_0" shape = [160, 160, 1, 1] dtype = "float32" min_val = float("-0.251195") @@ -2156,6 +2360,7 @@ class Program_weight_tensor_parameter_203: class Program_weight_tensor_parameter_204: name = "parameter_204" + original_name = "batch_norm2d_61.b_0" shape = [160] dtype = "float32" min_val = float("-0.0755261") @@ -2167,6 +2372,7 @@ class Program_weight_tensor_parameter_204: class Program_weight_tensor_parameter_205: name = "parameter_205" + original_name = "batch_norm2d_61.w_0" shape = [160] dtype = "float32" min_val = float("0.866127") @@ -2178,6 +2384,7 @@ class Program_weight_tensor_parameter_205: class Program_weight_tensor_parameter_206: name = "parameter_206" + original_name = "batch_norm2d_61.w_2" shape = [160] dtype = "float32" min_val = float("0.0647528") @@ -2189,6 +2396,7 @@ class Program_weight_tensor_parameter_206: class Program_weight_tensor_parameter_207: name = "parameter_207" + original_name = "batch_norm2d_61.w_1" shape = [160] dtype = "float32" min_val = float("-0.520049") @@ -2200,6 +2408,7 @@ class Program_weight_tensor_parameter_207: class Program_weight_tensor_parameter_208: name = "parameter_208" + original_name = "conv2d_65.w_0" shape = [160, 160, 1, 1] dtype = "float32" min_val = float("-0.39414") @@ -2211,6 +2420,7 @@ class Program_weight_tensor_parameter_208: class Program_weight_tensor_parameter_209: name = "parameter_209" + original_name = "batch_norm2d_60.b_0" shape = [160] dtype = "float32" min_val = float("-0.0329387") @@ -2222,6 +2432,7 @@ class Program_weight_tensor_parameter_209: class Program_weight_tensor_parameter_210: name = "parameter_210" + original_name = "batch_norm2d_60.w_0" shape = [160] dtype = "float32" min_val = float("0.553015") @@ -2233,6 +2444,7 @@ class Program_weight_tensor_parameter_210: class Program_weight_tensor_parameter_211: name = "parameter_211" + original_name = "batch_norm2d_60.w_2" shape = [160] dtype = "float32" min_val = float("0.00158392") @@ -2244,6 +2456,7 @@ class Program_weight_tensor_parameter_211: class Program_weight_tensor_parameter_212: name = "parameter_212" + original_name = "batch_norm2d_60.w_1" shape = [160] dtype = "float32" min_val = float("-0.126551") @@ -2255,6 +2468,7 @@ class Program_weight_tensor_parameter_212: class Program_weight_tensor_parameter_213: name = "parameter_213" + original_name = "conv2d_64.w_0" shape = [160, 1, 5, 5] dtype = "float32" min_val = float("-0.304104") @@ -2266,6 +2480,7 @@ class Program_weight_tensor_parameter_213: class Program_weight_tensor_parameter_214: name = "parameter_214" + original_name = "batch_norm2d_59.b_0" shape = [160] dtype = "float32" min_val = float("-0.0688978") @@ -2277,6 +2492,7 @@ class Program_weight_tensor_parameter_214: class Program_weight_tensor_parameter_215: name = "parameter_215" + original_name = "batch_norm2d_59.w_0" shape = [160] dtype = "float32" min_val = float("0.839806") @@ -2288,6 +2504,7 @@ class Program_weight_tensor_parameter_215: class Program_weight_tensor_parameter_216: name = "parameter_216" + original_name = "batch_norm2d_59.w_2" shape = [160] dtype = "float32" min_val = float("0.0364732") @@ -2299,6 +2516,7 @@ class Program_weight_tensor_parameter_216: class Program_weight_tensor_parameter_217: name = "parameter_217" + original_name = "batch_norm2d_59.w_1" shape = [160] dtype = "float32" min_val = float("-0.350011") @@ -2310,6 +2528,7 @@ class Program_weight_tensor_parameter_217: class Program_weight_tensor_parameter_218: name = "parameter_218" + original_name = "conv2d_63.w_0" shape = [160, 160, 1, 1] dtype = "float32" min_val = float("-0.31198") @@ -2321,6 +2540,7 @@ class Program_weight_tensor_parameter_218: class Program_weight_tensor_parameter_219: name = "parameter_219" + original_name = "batch_norm2d_58.b_0" shape = [160] dtype = "float32" min_val = float("-0.0472676") @@ -2332,6 +2552,7 @@ class Program_weight_tensor_parameter_219: class Program_weight_tensor_parameter_220: name = "parameter_220" + original_name = "batch_norm2d_58.w_0" shape = [160] dtype = "float32" min_val = float("0.74702") @@ -2343,6 +2564,7 @@ class Program_weight_tensor_parameter_220: class Program_weight_tensor_parameter_221: name = "parameter_221" + original_name = "batch_norm2d_58.w_2" shape = [160] dtype = "float32" min_val = float("0.00476241") @@ -2354,6 +2576,7 @@ class Program_weight_tensor_parameter_221: class Program_weight_tensor_parameter_222: name = "parameter_222" + original_name = "batch_norm2d_58.w_1" shape = [160] dtype = "float32" min_val = float("-0.154024") @@ -2365,6 +2588,7 @@ class Program_weight_tensor_parameter_222: class Program_weight_tensor_parameter_223: name = "parameter_223" + original_name = "conv2d_62.w_0" shape = [160, 1, 5, 5] dtype = "float32" min_val = float("-0.270579") @@ -2376,6 +2600,7 @@ class Program_weight_tensor_parameter_223: class Program_weight_tensor_parameter_224: name = "parameter_224" + original_name = "batch_norm2d_57.b_0" shape = [160] dtype = "float32" min_val = float("-0.0427445") @@ -2387,6 +2612,7 @@ class Program_weight_tensor_parameter_224: class Program_weight_tensor_parameter_225: name = "parameter_225" + original_name = "batch_norm2d_57.w_0" shape = [160] dtype = "float32" min_val = float("0.879859") @@ -2398,6 +2624,7 @@ class Program_weight_tensor_parameter_225: class Program_weight_tensor_parameter_226: name = "parameter_226" + original_name = "batch_norm2d_57.w_2" shape = [160] dtype = "float32" min_val = float("0.0321212") @@ -2409,6 +2636,7 @@ class Program_weight_tensor_parameter_226: class Program_weight_tensor_parameter_227: name = "parameter_227" + original_name = "batch_norm2d_57.w_1" shape = [160] dtype = "float32" min_val = float("-0.514584") @@ -2420,6 +2648,7 @@ class Program_weight_tensor_parameter_227: class Program_weight_tensor_parameter_228: name = "parameter_228" + original_name = "conv2d_61.w_0" shape = [160, 160, 1, 1] dtype = "float32" min_val = float("-0.323038") @@ -2431,6 +2660,7 @@ class Program_weight_tensor_parameter_228: class Program_weight_tensor_parameter_229: name = "parameter_229" + original_name = "batch_norm2d_56.b_0" shape = [160] dtype = "float32" min_val = float("-0.0546701") @@ -2442,6 +2672,7 @@ class Program_weight_tensor_parameter_229: class Program_weight_tensor_parameter_230: name = "parameter_230" + original_name = "batch_norm2d_56.w_0" shape = [160] dtype = "float32" min_val = float("0.631356") @@ -2453,6 +2684,7 @@ class Program_weight_tensor_parameter_230: class Program_weight_tensor_parameter_231: name = "parameter_231" + original_name = "batch_norm2d_56.w_2" shape = [160] dtype = "float32" min_val = float("0.00300946") @@ -2464,6 +2696,7 @@ class Program_weight_tensor_parameter_231: class Program_weight_tensor_parameter_232: name = "parameter_232" + original_name = "batch_norm2d_56.w_1" shape = [160] dtype = "float32" min_val = float("-0.122947") @@ -2475,6 +2708,7 @@ class Program_weight_tensor_parameter_232: class Program_weight_tensor_parameter_233: name = "parameter_233" + original_name = "conv2d_60.w_0" shape = [160, 1, 5, 5] dtype = "float32" min_val = float("-0.259897") @@ -2486,6 +2720,7 @@ class Program_weight_tensor_parameter_233: class Program_weight_tensor_parameter_234: name = "parameter_234" + original_name = "batch_norm2d_55.b_0" shape = [160] dtype = "float32" min_val = float("-0.0468143") @@ -2497,6 +2732,7 @@ class Program_weight_tensor_parameter_234: class Program_weight_tensor_parameter_235: name = "parameter_235" + original_name = "batch_norm2d_55.w_0" shape = [160] dtype = "float32" min_val = float("0.910259") @@ -2508,6 +2744,7 @@ class Program_weight_tensor_parameter_235: class Program_weight_tensor_parameter_236: name = "parameter_236" + original_name = "batch_norm2d_55.w_2" shape = [160] dtype = "float32" min_val = float("0.0450476") @@ -2519,6 +2756,7 @@ class Program_weight_tensor_parameter_236: class Program_weight_tensor_parameter_237: name = "parameter_237" + original_name = "batch_norm2d_55.w_1" shape = [160] dtype = "float32" min_val = float("-0.372936") @@ -2530,6 +2768,7 @@ class Program_weight_tensor_parameter_237: class Program_weight_tensor_parameter_238: name = "parameter_238" + original_name = "conv2d_59.w_0" shape = [160, 160, 1, 1] dtype = "float32" min_val = float("-0.23484") @@ -2541,6 +2780,7 @@ class Program_weight_tensor_parameter_238: class Program_weight_tensor_parameter_239: name = "parameter_239" + original_name = "batch_norm2d_54.b_0" shape = [160] dtype = "float32" min_val = float("-0.034552") @@ -2552,6 +2792,7 @@ class Program_weight_tensor_parameter_239: class Program_weight_tensor_parameter_240: name = "parameter_240" + original_name = "batch_norm2d_54.w_0" shape = [160] dtype = "float32" min_val = float("0.780116") @@ -2563,6 +2804,7 @@ class Program_weight_tensor_parameter_240: class Program_weight_tensor_parameter_241: name = "parameter_241" + original_name = "batch_norm2d_54.w_2" shape = [160] dtype = "float32" min_val = float("0.00603962") @@ -2574,6 +2816,7 @@ class Program_weight_tensor_parameter_241: class Program_weight_tensor_parameter_242: name = "parameter_242" + original_name = "batch_norm2d_54.w_1" shape = [160] dtype = "float32" min_val = float("-0.131625") @@ -2585,6 +2828,7 @@ class Program_weight_tensor_parameter_242: class Program_weight_tensor_parameter_243: name = "parameter_243" + original_name = "conv2d_58.w_0" shape = [160, 1, 5, 5] dtype = "float32" min_val = float("-0.250969") @@ -2596,6 +2840,7 @@ class Program_weight_tensor_parameter_243: class Program_weight_tensor_parameter_244: name = "parameter_244" + original_name = "batch_norm2d_33.b_0" shape = [160] dtype = "float32" min_val = float("-0.0291199") @@ -2607,6 +2852,7 @@ class Program_weight_tensor_parameter_244: class Program_weight_tensor_parameter_245: name = "parameter_245" + original_name = "batch_norm2d_33.w_0" shape = [160] dtype = "float32" min_val = float("0.626687") @@ -2618,6 +2864,7 @@ class Program_weight_tensor_parameter_245: class Program_weight_tensor_parameter_246: name = "parameter_246" + original_name = "batch_norm2d_33.w_2" shape = [160] dtype = "float32" min_val = float("0.114298") @@ -2629,6 +2876,7 @@ class Program_weight_tensor_parameter_246: class Program_weight_tensor_parameter_247: name = "parameter_247" + original_name = "batch_norm2d_33.w_1" shape = [160] dtype = "float32" min_val = float("-0.228379") @@ -2640,6 +2888,7 @@ class Program_weight_tensor_parameter_247: class Program_weight_tensor_parameter_248: name = "parameter_248" + original_name = "conv2d_37.w_0" shape = [160, 160, 1, 1] dtype = "float32" min_val = float("-0.285233") @@ -2651,6 +2900,7 @@ class Program_weight_tensor_parameter_248: class Program_weight_tensor_parameter_249: name = "parameter_249" + original_name = "batch_norm2d_32.b_0" shape = [160] dtype = "float32" min_val = float("-0.0325565") @@ -2662,6 +2912,7 @@ class Program_weight_tensor_parameter_249: class Program_weight_tensor_parameter_250: name = "parameter_250" + original_name = "batch_norm2d_32.w_0" shape = [160] dtype = "float32" min_val = float("0.61755") @@ -2673,6 +2924,7 @@ class Program_weight_tensor_parameter_250: class Program_weight_tensor_parameter_251: name = "parameter_251" + original_name = "batch_norm2d_32.w_2" shape = [160] dtype = "float32" min_val = float("0.0151111") @@ -2684,6 +2936,7 @@ class Program_weight_tensor_parameter_251: class Program_weight_tensor_parameter_252: name = "parameter_252" + original_name = "batch_norm2d_32.w_1" shape = [160] dtype = "float32" min_val = float("-0.418314") @@ -2695,6 +2948,7 @@ class Program_weight_tensor_parameter_252: class Program_weight_tensor_parameter_253: name = "parameter_253" + original_name = "conv2d_36.w_0" shape = [160, 1, 5, 5] dtype = "float32" min_val = float("-0.388543") @@ -2706,6 +2960,7 @@ class Program_weight_tensor_parameter_253: class Program_weight_tensor_parameter_254: name = "parameter_254" + original_name = "batch_norm2d_31.b_0" shape = [160] dtype = "float32" min_val = float("-0.0376141") @@ -2717,6 +2972,7 @@ class Program_weight_tensor_parameter_254: class Program_weight_tensor_parameter_255: name = "parameter_255" + original_name = "batch_norm2d_31.w_0" shape = [160] dtype = "float32" min_val = float("0.572496") @@ -2728,6 +2984,7 @@ class Program_weight_tensor_parameter_255: class Program_weight_tensor_parameter_256: name = "parameter_256" + original_name = "batch_norm2d_31.w_2" shape = [160] dtype = "float32" min_val = float("0.100631") @@ -2739,6 +2996,7 @@ class Program_weight_tensor_parameter_256: class Program_weight_tensor_parameter_257: name = "parameter_257" + original_name = "batch_norm2d_31.w_1" shape = [160] dtype = "float32" min_val = float("-0.241801") @@ -2750,6 +3008,7 @@ class Program_weight_tensor_parameter_257: class Program_weight_tensor_parameter_258: name = "parameter_258" + original_name = "conv2d_35.w_0" shape = [160, 160, 1, 1] dtype = "float32" min_val = float("-0.190079") @@ -2761,6 +3020,7 @@ class Program_weight_tensor_parameter_258: class Program_weight_tensor_parameter_259: name = "parameter_259" + original_name = "batch_norm2d_30.b_0" shape = [160] dtype = "float32" min_val = float("-0.0332988") @@ -2772,6 +3032,7 @@ class Program_weight_tensor_parameter_259: class Program_weight_tensor_parameter_260: name = "parameter_260" + original_name = "batch_norm2d_30.w_0" shape = [160] dtype = "float32" min_val = float("0.585353") @@ -2783,6 +3044,7 @@ class Program_weight_tensor_parameter_260: class Program_weight_tensor_parameter_261: name = "parameter_261" + original_name = "batch_norm2d_30.w_2" shape = [160] dtype = "float32" min_val = float("0.00364763") @@ -2794,6 +3056,7 @@ class Program_weight_tensor_parameter_261: class Program_weight_tensor_parameter_262: name = "parameter_262" + original_name = "batch_norm2d_30.w_1" shape = [160] dtype = "float32" min_val = float("-0.197793") @@ -2805,6 +3068,7 @@ class Program_weight_tensor_parameter_262: class Program_weight_tensor_parameter_263: name = "parameter_263" + original_name = "conv2d_34.w_0" shape = [160, 1, 5, 5] dtype = "float32" min_val = float("-0.381731") @@ -2816,6 +3080,7 @@ class Program_weight_tensor_parameter_263: class Program_weight_tensor_parameter_264: name = "parameter_264" + original_name = "batch_norm2d_53.b_0" shape = [160] dtype = "float32" min_val = float("-0.0248212") @@ -2827,6 +3092,7 @@ class Program_weight_tensor_parameter_264: class Program_weight_tensor_parameter_265: name = "parameter_265" + original_name = "batch_norm2d_53.w_0" shape = [160] dtype = "float32" min_val = float("0.862371") @@ -2838,6 +3104,7 @@ class Program_weight_tensor_parameter_265: class Program_weight_tensor_parameter_266: name = "parameter_266" + original_name = "batch_norm2d_53.w_2" shape = [160] dtype = "float32" min_val = float("0.280771") @@ -2849,6 +3116,7 @@ class Program_weight_tensor_parameter_266: class Program_weight_tensor_parameter_267: name = "parameter_267" + original_name = "batch_norm2d_53.w_1" shape = [160] dtype = "float32" min_val = float("-0.584582") @@ -2860,6 +3128,7 @@ class Program_weight_tensor_parameter_267: class Program_weight_tensor_parameter_268: name = "parameter_268" + original_name = "conv2d_57.w_0" shape = [160, 320, 1, 1] dtype = "float32" min_val = float("-0.252617") @@ -2871,6 +3140,7 @@ class Program_weight_tensor_parameter_268: class Program_weight_tensor_parameter_269: name = "parameter_269" + original_name = "batch_norm2d_52.b_0" shape = [320] dtype = "float32" min_val = float("-0.0232896") @@ -2882,6 +3152,7 @@ class Program_weight_tensor_parameter_269: class Program_weight_tensor_parameter_270: name = "parameter_270" + original_name = "batch_norm2d_52.w_0" shape = [320] dtype = "float32" min_val = float("0.842911") @@ -2893,6 +3164,7 @@ class Program_weight_tensor_parameter_270: class Program_weight_tensor_parameter_271: name = "parameter_271" + original_name = "batch_norm2d_52.w_2" shape = [320] dtype = "float32" min_val = float("0.0228347") @@ -2904,6 +3176,7 @@ class Program_weight_tensor_parameter_271: class Program_weight_tensor_parameter_272: name = "parameter_272" + original_name = "batch_norm2d_52.w_1" shape = [320] dtype = "float32" min_val = float("-0.262674") @@ -2915,6 +3188,7 @@ class Program_weight_tensor_parameter_272: class Program_weight_tensor_parameter_273: name = "parameter_273" + original_name = "conv2d_56.w_0" shape = [320, 1, 5, 5] dtype = "float32" min_val = float("-0.827294") @@ -2926,6 +3200,7 @@ class Program_weight_tensor_parameter_273: class Program_weight_tensor_parameter_274: name = "parameter_274" + original_name = "batch_norm2d_51.b_0" shape = [320] dtype = "float32" min_val = float("-0.0264182") @@ -2937,6 +3212,7 @@ class Program_weight_tensor_parameter_274: class Program_weight_tensor_parameter_275: name = "parameter_275" + original_name = "batch_norm2d_51.w_0" shape = [320] dtype = "float32" min_val = float("0.925001") @@ -2948,6 +3224,7 @@ class Program_weight_tensor_parameter_275: class Program_weight_tensor_parameter_276: name = "parameter_276" + original_name = "batch_norm2d_51.w_2" shape = [320] dtype = "float32" min_val = float("0.262418") @@ -2959,6 +3236,7 @@ class Program_weight_tensor_parameter_276: class Program_weight_tensor_parameter_277: name = "parameter_277" + original_name = "batch_norm2d_51.w_1" shape = [320] dtype = "float32" min_val = float("-0.345992") @@ -2970,6 +3248,7 @@ class Program_weight_tensor_parameter_277: class Program_weight_tensor_parameter_278: name = "parameter_278" + original_name = "conv2d_55.w_0" shape = [320, 320, 1, 1] dtype = "float32" min_val = float("-0.239955") @@ -2981,6 +3260,7 @@ class Program_weight_tensor_parameter_278: class Program_weight_tensor_parameter_279: name = "parameter_279" + original_name = "batch_norm2d_50.b_0" shape = [320] dtype = "float32" min_val = float("-0.0523676") @@ -2992,6 +3272,7 @@ class Program_weight_tensor_parameter_279: class Program_weight_tensor_parameter_280: name = "parameter_280" + original_name = "batch_norm2d_50.w_0" shape = [320] dtype = "float32" min_val = float("0.819276") @@ -3003,6 +3284,7 @@ class Program_weight_tensor_parameter_280: class Program_weight_tensor_parameter_281: name = "parameter_281" + original_name = "batch_norm2d_50.w_2" shape = [320] dtype = "float32" min_val = float("0.00449247") @@ -3014,6 +3296,7 @@ class Program_weight_tensor_parameter_281: class Program_weight_tensor_parameter_282: name = "parameter_282" + original_name = "batch_norm2d_50.w_1" shape = [320] dtype = "float32" min_val = float("-0.153285") @@ -3025,6 +3308,7 @@ class Program_weight_tensor_parameter_282: class Program_weight_tensor_parameter_283: name = "parameter_283" + original_name = "conv2d_54.w_0" shape = [320, 1, 5, 5] dtype = "float32" min_val = float("-0.724057") @@ -3036,6 +3320,7 @@ class Program_weight_tensor_parameter_283: class Program_weight_tensor_parameter_284: name = "parameter_284" + original_name = "batch_norm2d_49.b_0" shape = [160] dtype = "float32" min_val = float("-0.0106541") @@ -3047,6 +3332,7 @@ class Program_weight_tensor_parameter_284: class Program_weight_tensor_parameter_285: name = "parameter_285" + original_name = "batch_norm2d_49.w_0" shape = [160] dtype = "float32" min_val = float("0.639752") @@ -3058,6 +3344,7 @@ class Program_weight_tensor_parameter_285: class Program_weight_tensor_parameter_286: name = "parameter_286" + original_name = "batch_norm2d_49.w_2" shape = [160] dtype = "float32" min_val = float("0.0824698") @@ -3069,6 +3356,7 @@ class Program_weight_tensor_parameter_286: class Program_weight_tensor_parameter_287: name = "parameter_287" + original_name = "batch_norm2d_49.w_1" shape = [160] dtype = "float32" min_val = float("-0.243841") @@ -3080,6 +3368,7 @@ class Program_weight_tensor_parameter_287: class Program_weight_tensor_parameter_288: name = "parameter_288" + original_name = "conv2d_53.w_0" shape = [160, 160, 1, 1] dtype = "float32" min_val = float("-0.246128") @@ -3091,6 +3380,7 @@ class Program_weight_tensor_parameter_288: class Program_weight_tensor_parameter_289: name = "parameter_289" + original_name = "batch_norm2d_48.b_0" shape = [160] dtype = "float32" min_val = float("-0.0530809") @@ -3102,6 +3392,7 @@ class Program_weight_tensor_parameter_289: class Program_weight_tensor_parameter_290: name = "parameter_290" + original_name = "batch_norm2d_48.w_0" shape = [160] dtype = "float32" min_val = float("0.539444") @@ -3113,6 +3404,7 @@ class Program_weight_tensor_parameter_290: class Program_weight_tensor_parameter_291: name = "parameter_291" + original_name = "batch_norm2d_48.w_2" shape = [160] dtype = "float32" min_val = float("0.0110039") @@ -3124,6 +3416,7 @@ class Program_weight_tensor_parameter_291: class Program_weight_tensor_parameter_292: name = "parameter_292" + original_name = "batch_norm2d_48.w_1" shape = [160] dtype = "float32" min_val = float("-0.281357") @@ -3135,6 +3428,7 @@ class Program_weight_tensor_parameter_292: class Program_weight_tensor_parameter_293: name = "parameter_293" + original_name = "conv2d_52.w_0" shape = [160, 1, 5, 5] dtype = "float32" min_val = float("-0.398614") @@ -3146,6 +3440,7 @@ class Program_weight_tensor_parameter_293: class Program_weight_tensor_parameter_294: name = "parameter_294" + original_name = "batch_norm2d_47.b_0" shape = [160] dtype = "float32" min_val = float("-0.0419701") @@ -3157,6 +3452,7 @@ class Program_weight_tensor_parameter_294: class Program_weight_tensor_parameter_295: name = "parameter_295" + original_name = "batch_norm2d_47.w_0" shape = [160] dtype = "float32" min_val = float("0.87882") @@ -3168,6 +3464,7 @@ class Program_weight_tensor_parameter_295: class Program_weight_tensor_parameter_296: name = "parameter_296" + original_name = "batch_norm2d_47.w_2" shape = [160] dtype = "float32" min_val = float("0.351053") @@ -3179,6 +3476,7 @@ class Program_weight_tensor_parameter_296: class Program_weight_tensor_parameter_297: name = "parameter_297" + original_name = "batch_norm2d_47.w_1" shape = [160] dtype = "float32" min_val = float("-0.518443") @@ -3190,6 +3488,7 @@ class Program_weight_tensor_parameter_297: class Program_weight_tensor_parameter_298: name = "parameter_298" + original_name = "conv2d_51.w_0" shape = [160, 320, 1, 1] dtype = "float32" min_val = float("-0.26603") @@ -3201,6 +3500,7 @@ class Program_weight_tensor_parameter_298: class Program_weight_tensor_parameter_299: name = "parameter_299" + original_name = "batch_norm2d_46.b_0" shape = [320] dtype = "float32" min_val = float("-0.0273029") @@ -3212,6 +3512,7 @@ class Program_weight_tensor_parameter_299: class Program_weight_tensor_parameter_300: name = "parameter_300" + original_name = "batch_norm2d_46.w_0" shape = [320] dtype = "float32" min_val = float("0.801888") @@ -3223,6 +3524,7 @@ class Program_weight_tensor_parameter_300: class Program_weight_tensor_parameter_301: name = "parameter_301" + original_name = "batch_norm2d_46.w_2" shape = [320] dtype = "float32" min_val = float("0.0102874") @@ -3234,6 +3536,7 @@ class Program_weight_tensor_parameter_301: class Program_weight_tensor_parameter_302: name = "parameter_302" + original_name = "batch_norm2d_46.w_1" shape = [320] dtype = "float32" min_val = float("-0.379929") @@ -3245,6 +3548,7 @@ class Program_weight_tensor_parameter_302: class Program_weight_tensor_parameter_303: name = "parameter_303" + original_name = "conv2d_50.w_0" shape = [320, 1, 5, 5] dtype = "float32" min_val = float("-0.748418") @@ -3256,6 +3560,7 @@ class Program_weight_tensor_parameter_303: class Program_weight_tensor_parameter_304: name = "parameter_304" + original_name = "batch_norm2d_45.b_0" shape = [320] dtype = "float32" min_val = float("-0.0390889") @@ -3267,6 +3572,7 @@ class Program_weight_tensor_parameter_304: class Program_weight_tensor_parameter_305: name = "parameter_305" + original_name = "batch_norm2d_45.w_0" shape = [320] dtype = "float32" min_val = float("0.928824") @@ -3278,6 +3584,7 @@ class Program_weight_tensor_parameter_305: class Program_weight_tensor_parameter_306: name = "parameter_306" + original_name = "batch_norm2d_45.w_2" shape = [320] dtype = "float32" min_val = float("0.263508") @@ -3289,6 +3596,7 @@ class Program_weight_tensor_parameter_306: class Program_weight_tensor_parameter_307: name = "parameter_307" + original_name = "batch_norm2d_45.w_1" shape = [320] dtype = "float32" min_val = float("-0.500478") @@ -3300,6 +3608,7 @@ class Program_weight_tensor_parameter_307: class Program_weight_tensor_parameter_308: name = "parameter_308" + original_name = "conv2d_49.w_0" shape = [320, 320, 1, 1] dtype = "float32" min_val = float("-0.33573") @@ -3311,6 +3620,7 @@ class Program_weight_tensor_parameter_308: class Program_weight_tensor_parameter_309: name = "parameter_309" + original_name = "batch_norm2d_44.b_0" shape = [320] dtype = "float32" min_val = float("-0.0379724") @@ -3322,6 +3632,7 @@ class Program_weight_tensor_parameter_309: class Program_weight_tensor_parameter_310: name = "parameter_310" + original_name = "batch_norm2d_44.w_0" shape = [320] dtype = "float32" min_val = float("0.772452") @@ -3333,6 +3644,7 @@ class Program_weight_tensor_parameter_310: class Program_weight_tensor_parameter_311: name = "parameter_311" + original_name = "batch_norm2d_44.w_2" shape = [320] dtype = "float32" min_val = float("0.0105535") @@ -3344,6 +3656,7 @@ class Program_weight_tensor_parameter_311: class Program_weight_tensor_parameter_312: name = "parameter_312" + original_name = "batch_norm2d_44.w_1" shape = [320] dtype = "float32" min_val = float("-0.337992") @@ -3355,6 +3668,7 @@ class Program_weight_tensor_parameter_312: class Program_weight_tensor_parameter_313: name = "parameter_313" + original_name = "conv2d_48.w_0" shape = [320, 1, 5, 5] dtype = "float32" min_val = float("-0.700472") @@ -3366,6 +3680,7 @@ class Program_weight_tensor_parameter_313: class Program_weight_tensor_parameter_314: name = "parameter_314" + original_name = "batch_norm2d_43.b_0" shape = [160] dtype = "float32" min_val = float("-0.0308389") @@ -3377,6 +3692,7 @@ class Program_weight_tensor_parameter_314: class Program_weight_tensor_parameter_315: name = "parameter_315" + original_name = "batch_norm2d_43.w_0" shape = [160] dtype = "float32" min_val = float("0.599795") @@ -3388,6 +3704,7 @@ class Program_weight_tensor_parameter_315: class Program_weight_tensor_parameter_316: name = "parameter_316" + original_name = "batch_norm2d_43.w_2" shape = [160] dtype = "float32" min_val = float("0.120822") @@ -3399,6 +3716,7 @@ class Program_weight_tensor_parameter_316: class Program_weight_tensor_parameter_317: name = "parameter_317" + original_name = "batch_norm2d_43.w_1" shape = [160] dtype = "float32" min_val = float("-0.148185") @@ -3410,6 +3728,7 @@ class Program_weight_tensor_parameter_317: class Program_weight_tensor_parameter_318: name = "parameter_318" + original_name = "conv2d_47.w_0" shape = [160, 160, 1, 1] dtype = "float32" min_val = float("-0.23499") @@ -3421,6 +3740,7 @@ class Program_weight_tensor_parameter_318: class Program_weight_tensor_parameter_319: name = "parameter_319" + original_name = "batch_norm2d_42.b_0" shape = [160] dtype = "float32" min_val = float("-0.0407756") @@ -3432,6 +3752,7 @@ class Program_weight_tensor_parameter_319: class Program_weight_tensor_parameter_320: name = "parameter_320" + original_name = "batch_norm2d_42.w_0" shape = [160] dtype = "float32" min_val = float("0.521142") @@ -3443,6 +3764,7 @@ class Program_weight_tensor_parameter_320: class Program_weight_tensor_parameter_321: name = "parameter_321" + original_name = "batch_norm2d_42.w_2" shape = [160] dtype = "float32" min_val = float("0.0361932") @@ -3454,6 +3776,7 @@ class Program_weight_tensor_parameter_321: class Program_weight_tensor_parameter_322: name = "parameter_322" + original_name = "batch_norm2d_42.w_1" shape = [160] dtype = "float32" min_val = float("-0.488548") @@ -3465,6 +3788,7 @@ class Program_weight_tensor_parameter_322: class Program_weight_tensor_parameter_323: name = "parameter_323" + original_name = "conv2d_46.w_0" shape = [160, 1, 5, 5] dtype = "float32" min_val = float("-0.394379") @@ -3476,6 +3800,7 @@ class Program_weight_tensor_parameter_323: class Program_weight_tensor_parameter_324: name = "parameter_324" + original_name = "batch_norm2d_41.b_0" shape = [160] dtype = "float32" min_val = float("-0.00991972") @@ -3487,6 +3812,7 @@ class Program_weight_tensor_parameter_324: class Program_weight_tensor_parameter_325: name = "parameter_325" + original_name = "batch_norm2d_41.w_0" shape = [160] dtype = "float32" min_val = float("0.843879") @@ -3498,6 +3824,7 @@ class Program_weight_tensor_parameter_325: class Program_weight_tensor_parameter_326: name = "parameter_326" + original_name = "batch_norm2d_41.w_2" shape = [160] dtype = "float32" min_val = float("0.331154") @@ -3509,6 +3836,7 @@ class Program_weight_tensor_parameter_326: class Program_weight_tensor_parameter_327: name = "parameter_327" + original_name = "batch_norm2d_41.w_1" shape = [160] dtype = "float32" min_val = float("-0.668181") @@ -3520,6 +3848,7 @@ class Program_weight_tensor_parameter_327: class Program_weight_tensor_parameter_328: name = "parameter_328" + original_name = "conv2d_45.w_0" shape = [160, 320, 1, 1] dtype = "float32" min_val = float("-0.292215") @@ -3531,6 +3860,7 @@ class Program_weight_tensor_parameter_328: class Program_weight_tensor_parameter_329: name = "parameter_329" + original_name = "batch_norm2d_40.b_0" shape = [320] dtype = "float32" min_val = float("-0.0448017") @@ -3542,6 +3872,7 @@ class Program_weight_tensor_parameter_329: class Program_weight_tensor_parameter_330: name = "parameter_330" + original_name = "batch_norm2d_40.w_0" shape = [320] dtype = "float32" min_val = float("0.772261") @@ -3553,6 +3884,7 @@ class Program_weight_tensor_parameter_330: class Program_weight_tensor_parameter_331: name = "parameter_331" + original_name = "batch_norm2d_40.w_2" shape = [320] dtype = "float32" min_val = float("0.0455745") @@ -3564,6 +3896,7 @@ class Program_weight_tensor_parameter_331: class Program_weight_tensor_parameter_332: name = "parameter_332" + original_name = "batch_norm2d_40.w_1" shape = [320] dtype = "float32" min_val = float("-0.469359") @@ -3575,6 +3908,7 @@ class Program_weight_tensor_parameter_332: class Program_weight_tensor_parameter_333: name = "parameter_333" + original_name = "conv2d_44.w_0" shape = [320, 1, 5, 5] dtype = "float32" min_val = float("-0.767494") @@ -3586,6 +3920,7 @@ class Program_weight_tensor_parameter_333: class Program_weight_tensor_parameter_334: name = "parameter_334" + original_name = "batch_norm2d_39.b_0" shape = [320] dtype = "float32" min_val = float("-0.0794769") @@ -3597,6 +3932,7 @@ class Program_weight_tensor_parameter_334: class Program_weight_tensor_parameter_335: name = "parameter_335" + original_name = "batch_norm2d_39.w_0" shape = [320] dtype = "float32" min_val = float("0.878525") @@ -3608,6 +3944,7 @@ class Program_weight_tensor_parameter_335: class Program_weight_tensor_parameter_336: name = "parameter_336" + original_name = "batch_norm2d_39.w_2" shape = [320] dtype = "float32" min_val = float("0.270631") @@ -3619,6 +3956,7 @@ class Program_weight_tensor_parameter_336: class Program_weight_tensor_parameter_337: name = "parameter_337" + original_name = "batch_norm2d_39.w_1" shape = [320] dtype = "float32" min_val = float("-0.571073") @@ -3630,6 +3968,7 @@ class Program_weight_tensor_parameter_337: class Program_weight_tensor_parameter_338: name = "parameter_338" + original_name = "conv2d_43.w_0" shape = [320, 320, 1, 1] dtype = "float32" min_val = float("-0.308295") @@ -3641,6 +3980,7 @@ class Program_weight_tensor_parameter_338: class Program_weight_tensor_parameter_339: name = "parameter_339" + original_name = "batch_norm2d_38.b_0" shape = [320] dtype = "float32" min_val = float("-0.0387463") @@ -3652,6 +3992,7 @@ class Program_weight_tensor_parameter_339: class Program_weight_tensor_parameter_340: name = "parameter_340" + original_name = "batch_norm2d_38.w_0" shape = [320] dtype = "float32" min_val = float("0.732848") @@ -3663,6 +4004,7 @@ class Program_weight_tensor_parameter_340: class Program_weight_tensor_parameter_341: name = "parameter_341" + original_name = "batch_norm2d_38.w_2" shape = [320] dtype = "float32" min_val = float("0.018494") @@ -3674,6 +4016,7 @@ class Program_weight_tensor_parameter_341: class Program_weight_tensor_parameter_342: name = "parameter_342" + original_name = "batch_norm2d_38.w_1" shape = [320] dtype = "float32" min_val = float("-0.385606") @@ -3685,6 +4028,7 @@ class Program_weight_tensor_parameter_342: class Program_weight_tensor_parameter_343: name = "parameter_343" + original_name = "conv2d_42.w_0" shape = [320, 1, 5, 5] dtype = "float32" min_val = float("-0.73223") @@ -3696,6 +4040,7 @@ class Program_weight_tensor_parameter_343: class Program_weight_tensor_parameter_344: name = "parameter_344" + original_name = "batch_norm2d_37.b_0" shape = [160] dtype = "float32" min_val = float("-0.0217557") @@ -3707,6 +4052,7 @@ class Program_weight_tensor_parameter_344: class Program_weight_tensor_parameter_345: name = "parameter_345" + original_name = "batch_norm2d_37.w_0" shape = [160] dtype = "float32" min_val = float("0.87303") @@ -3718,6 +4064,7 @@ class Program_weight_tensor_parameter_345: class Program_weight_tensor_parameter_346: name = "parameter_346" + original_name = "batch_norm2d_37.w_2" shape = [160] dtype = "float32" min_val = float("0.351834") @@ -3729,6 +4076,7 @@ class Program_weight_tensor_parameter_346: class Program_weight_tensor_parameter_347: name = "parameter_347" + original_name = "batch_norm2d_37.w_1" shape = [160] dtype = "float32" min_val = float("-0.633124") @@ -3740,6 +4088,7 @@ class Program_weight_tensor_parameter_347: class Program_weight_tensor_parameter_348: name = "parameter_348" + original_name = "conv2d_41.w_0" shape = [160, 320, 1, 1] dtype = "float32" min_val = float("-0.247984") @@ -3751,6 +4100,7 @@ class Program_weight_tensor_parameter_348: class Program_weight_tensor_parameter_349: name = "parameter_349" + original_name = "batch_norm2d_36.b_0" shape = [320] dtype = "float32" min_val = float("-0.0226518") @@ -3762,6 +4112,7 @@ class Program_weight_tensor_parameter_349: class Program_weight_tensor_parameter_350: name = "parameter_350" + original_name = "batch_norm2d_36.w_0" shape = [320] dtype = "float32" min_val = float("0.823101") @@ -3773,6 +4124,7 @@ class Program_weight_tensor_parameter_350: class Program_weight_tensor_parameter_351: name = "parameter_351" + original_name = "batch_norm2d_36.w_2" shape = [320] dtype = "float32" min_val = float("0.0309845") @@ -3784,6 +4136,7 @@ class Program_weight_tensor_parameter_351: class Program_weight_tensor_parameter_352: name = "parameter_352" + original_name = "batch_norm2d_36.w_1" shape = [320] dtype = "float32" min_val = float("-0.323988") @@ -3795,6 +4148,7 @@ class Program_weight_tensor_parameter_352: class Program_weight_tensor_parameter_353: name = "parameter_353" + original_name = "conv2d_40.w_0" shape = [320, 1, 5, 5] dtype = "float32" min_val = float("-0.743855") @@ -3806,6 +4160,7 @@ class Program_weight_tensor_parameter_353: class Program_weight_tensor_parameter_354: name = "parameter_354" + original_name = "batch_norm2d_35.b_0" shape = [320] dtype = "float32" min_val = float("-0.0293611") @@ -3817,6 +4172,7 @@ class Program_weight_tensor_parameter_354: class Program_weight_tensor_parameter_355: name = "parameter_355" + original_name = "batch_norm2d_35.w_0" shape = [320] dtype = "float32" min_val = float("0.913355") @@ -3828,6 +4184,7 @@ class Program_weight_tensor_parameter_355: class Program_weight_tensor_parameter_356: name = "parameter_356" + original_name = "batch_norm2d_35.w_2" shape = [320] dtype = "float32" min_val = float("0.236992") @@ -3839,6 +4196,7 @@ class Program_weight_tensor_parameter_356: class Program_weight_tensor_parameter_357: name = "parameter_357" + original_name = "batch_norm2d_35.w_1" shape = [320] dtype = "float32" min_val = float("-0.579431") @@ -3850,6 +4208,7 @@ class Program_weight_tensor_parameter_357: class Program_weight_tensor_parameter_358: name = "parameter_358" + original_name = "conv2d_39.w_0" shape = [320, 320, 1, 1] dtype = "float32" min_val = float("-0.277896") @@ -3861,6 +4220,7 @@ class Program_weight_tensor_parameter_358: class Program_weight_tensor_parameter_359: name = "parameter_359" + original_name = "batch_norm2d_34.b_0" shape = [320] dtype = "float32" min_val = float("-0.0456389") @@ -3872,6 +4232,7 @@ class Program_weight_tensor_parameter_359: class Program_weight_tensor_parameter_360: name = "parameter_360" + original_name = "batch_norm2d_34.w_0" shape = [320] dtype = "float32" min_val = float("0.742537") @@ -3883,6 +4244,7 @@ class Program_weight_tensor_parameter_360: class Program_weight_tensor_parameter_361: name = "parameter_361" + original_name = "batch_norm2d_34.w_2" shape = [320] dtype = "float32" min_val = float("0.0111846") @@ -3894,6 +4256,7 @@ class Program_weight_tensor_parameter_361: class Program_weight_tensor_parameter_362: name = "parameter_362" + original_name = "batch_norm2d_34.w_1" shape = [320] dtype = "float32" min_val = float("-0.167909") @@ -3905,6 +4268,7 @@ class Program_weight_tensor_parameter_362: class Program_weight_tensor_parameter_363: name = "parameter_363" + original_name = "conv2d_38.w_0" shape = [320, 1, 5, 5] dtype = "float32" min_val = float("-0.790535") @@ -3916,6 +4280,7 @@ class Program_weight_tensor_parameter_363: class Program_weight_tensor_parameter_364: name = "parameter_364" + original_name = "batch_norm2d_29.b_0" shape = [160] dtype = "float32" min_val = float("-0.0585648") @@ -3927,6 +4292,7 @@ class Program_weight_tensor_parameter_364: class Program_weight_tensor_parameter_365: name = "parameter_365" + original_name = "batch_norm2d_29.w_0" shape = [160] dtype = "float32" min_val = float("0.628471") @@ -3938,6 +4304,7 @@ class Program_weight_tensor_parameter_365: class Program_weight_tensor_parameter_366: name = "parameter_366" + original_name = "batch_norm2d_29.w_2" shape = [160] dtype = "float32" min_val = float("3.48361") @@ -3949,6 +4316,7 @@ class Program_weight_tensor_parameter_366: class Program_weight_tensor_parameter_367: name = "parameter_367" + original_name = "batch_norm2d_29.w_1" shape = [160] dtype = "float32" min_val = float("-2.555") @@ -3960,6 +4328,7 @@ class Program_weight_tensor_parameter_367: class Program_weight_tensor_parameter_368: name = "parameter_368" + original_name = "conv2d_33.w_0" shape = [160, 1024, 1, 1] dtype = "float32" min_val = float("-0.169979") @@ -3971,6 +4340,7 @@ class Program_weight_tensor_parameter_368: class Program_weight_tensor_parameter_369: name = "parameter_369" + original_name = "batch_norm2d_28.b_0" shape = [160] dtype = "float32" min_val = float("-0.0314462") @@ -3982,6 +4352,7 @@ class Program_weight_tensor_parameter_369: class Program_weight_tensor_parameter_370: name = "parameter_370" + original_name = "batch_norm2d_28.w_0" shape = [160] dtype = "float32" min_val = float("0.630305") @@ -3993,6 +4364,7 @@ class Program_weight_tensor_parameter_370: class Program_weight_tensor_parameter_371: name = "parameter_371" + original_name = "batch_norm2d_28.w_2" shape = [160] dtype = "float32" min_val = float("0.303377") @@ -4004,6 +4376,7 @@ class Program_weight_tensor_parameter_371: class Program_weight_tensor_parameter_372: name = "parameter_372" + original_name = "batch_norm2d_28.w_1" shape = [160] dtype = "float32" min_val = float("-1.10512") @@ -4015,6 +4388,7 @@ class Program_weight_tensor_parameter_372: class Program_weight_tensor_parameter_373: name = "parameter_373" + original_name = "conv2d_32.w_0" shape = [160, 512, 1, 1] dtype = "float32" min_val = float("-0.225878") @@ -4026,6 +4400,7 @@ class Program_weight_tensor_parameter_373: class Program_weight_tensor_parameter_374: name = "parameter_374" + original_name = "batch_norm2d_27.b_0" shape = [160] dtype = "float32" min_val = float("-0.205287") @@ -4037,6 +4412,7 @@ class Program_weight_tensor_parameter_374: class Program_weight_tensor_parameter_375: name = "parameter_375" + original_name = "batch_norm2d_27.w_0" shape = [160] dtype = "float32" min_val = float("0.589941") @@ -4048,6 +4424,7 @@ class Program_weight_tensor_parameter_375: class Program_weight_tensor_parameter_376: name = "parameter_376" + original_name = "batch_norm2d_27.w_2" shape = [160] dtype = "float32" min_val = float("0.380835") @@ -4059,6 +4436,7 @@ class Program_weight_tensor_parameter_376: class Program_weight_tensor_parameter_377: name = "parameter_377" + original_name = "batch_norm2d_27.w_1" shape = [160] dtype = "float32" min_val = float("-1.28333") @@ -4070,6 +4448,7 @@ class Program_weight_tensor_parameter_377: class Program_weight_tensor_parameter_378: name = "parameter_378" + original_name = "conv2d_31.w_0" shape = [160, 256, 1, 1] dtype = "float32" min_val = float("-0.269517") @@ -4081,6 +4460,7 @@ class Program_weight_tensor_parameter_378: class Program_weight_tensor_parameter_379: name = "parameter_379" + original_name = "batch_norm2d_26.b_0" shape = [1024] dtype = "float32" min_val = float("-4.6212") @@ -4092,6 +4472,7 @@ class Program_weight_tensor_parameter_379: class Program_weight_tensor_parameter_380: name = "parameter_380" + original_name = "batch_norm2d_26.w_0" shape = [1024] dtype = "float32" min_val = float("2.57045") @@ -4103,6 +4484,7 @@ class Program_weight_tensor_parameter_380: class Program_weight_tensor_parameter_381: name = "parameter_381" + original_name = "batch_norm2d_26.w_2" shape = [1024] dtype = "float32" min_val = float("0.0637851") @@ -4114,6 +4496,7 @@ class Program_weight_tensor_parameter_381: class Program_weight_tensor_parameter_382: name = "parameter_382" + original_name = "batch_norm2d_26.w_1" shape = [1024] dtype = "float32" min_val = float("-0.69245") @@ -4125,6 +4508,7 @@ class Program_weight_tensor_parameter_382: class Program_weight_tensor_parameter_383: name = "parameter_383" + original_name = "conv2d_30.w_0" shape = [1024, 1024, 1, 1] dtype = "float32" min_val = float("-0.913178") @@ -4136,6 +4520,7 @@ class Program_weight_tensor_parameter_383: class Program_weight_tensor_parameter_384: name = "parameter_384" + original_name = "conv2d_29.b_0" shape = [1024] dtype = "float32" min_val = float("-0.292223") @@ -4147,6 +4532,7 @@ class Program_weight_tensor_parameter_384: class Program_weight_tensor_parameter_385: name = "parameter_385" + original_name = "conv2d_29.w_0" shape = [1024, 256, 1, 1] dtype = "float32" min_val = float("-0.469521") @@ -4158,6 +4544,7 @@ class Program_weight_tensor_parameter_385: class Program_weight_tensor_parameter_386: name = "parameter_386" + original_name = "conv2d_28.b_0" shape = [256] dtype = "float32" min_val = float("-0.0212755") @@ -4169,6 +4556,7 @@ class Program_weight_tensor_parameter_386: class Program_weight_tensor_parameter_387: name = "parameter_387" + original_name = "conv2d_28.w_0" shape = [256, 1024, 1, 1] dtype = "float32" min_val = float("-0.206432") @@ -4180,6 +4568,7 @@ class Program_weight_tensor_parameter_387: class Program_weight_tensor_parameter_388: name = "parameter_388" + original_name = "batch_norm2d_25.b_0" shape = [1024] dtype = "float32" min_val = float("-3.58965") @@ -4191,6 +4580,7 @@ class Program_weight_tensor_parameter_388: class Program_weight_tensor_parameter_389: name = "parameter_389" + original_name = "batch_norm2d_25.w_0" shape = [1024] dtype = "float32" min_val = float("0.187217") @@ -4202,6 +4592,7 @@ class Program_weight_tensor_parameter_389: class Program_weight_tensor_parameter_390: name = "parameter_390" + original_name = "batch_norm2d_25.w_2" shape = [1024] dtype = "float32" min_val = float("6.66772e-05") @@ -4213,6 +4604,7 @@ class Program_weight_tensor_parameter_390: class Program_weight_tensor_parameter_391: name = "parameter_391" + original_name = "batch_norm2d_25.w_1" shape = [1024] dtype = "float32" min_val = float("-1.78135") @@ -4224,6 +4616,7 @@ class Program_weight_tensor_parameter_391: class Program_weight_tensor_parameter_392: name = "parameter_392" + original_name = "conv2d_27.w_0" shape = [1024, 1, 5, 5] dtype = "float32" min_val = float("-0.36244") @@ -4235,6 +4628,7 @@ class Program_weight_tensor_parameter_392: class Program_weight_tensor_parameter_393: name = "parameter_393" + original_name = "batch_norm2d_24.b_0" shape = [1024] dtype = "float32" min_val = float("-2.85367") @@ -4246,6 +4640,7 @@ class Program_weight_tensor_parameter_393: class Program_weight_tensor_parameter_394: name = "parameter_394" + original_name = "batch_norm2d_24.w_0" shape = [1024] dtype = "float32" min_val = float("-0.104229") @@ -4257,6 +4652,7 @@ class Program_weight_tensor_parameter_394: class Program_weight_tensor_parameter_395: name = "parameter_395" + original_name = "batch_norm2d_24.w_2" shape = [1024] dtype = "float32" min_val = float("0.026801") @@ -4268,6 +4664,7 @@ class Program_weight_tensor_parameter_395: class Program_weight_tensor_parameter_396: name = "parameter_396" + original_name = "batch_norm2d_24.w_1" shape = [1024] dtype = "float32" min_val = float("-0.586885") @@ -4279,6 +4676,7 @@ class Program_weight_tensor_parameter_396: class Program_weight_tensor_parameter_397: name = "parameter_397" + original_name = "conv2d_26.w_0" shape = [1024, 512, 1, 1] dtype = "float32" min_val = float("-0.409152") @@ -4290,6 +4688,7 @@ class Program_weight_tensor_parameter_397: class Program_weight_tensor_parameter_398: name = "parameter_398" + original_name = "conv2d_25.b_0" shape = [512] dtype = "float32" min_val = float("-0.176917") @@ -4301,6 +4700,7 @@ class Program_weight_tensor_parameter_398: class Program_weight_tensor_parameter_399: name = "parameter_399" + original_name = "conv2d_25.w_0" shape = [512, 128, 1, 1] dtype = "float32" min_val = float("-0.343035") @@ -4312,6 +4712,7 @@ class Program_weight_tensor_parameter_399: class Program_weight_tensor_parameter_400: name = "parameter_400" + original_name = "conv2d_24.b_0" shape = [128] dtype = "float32" min_val = float("-0.000576574") @@ -4323,6 +4724,7 @@ class Program_weight_tensor_parameter_400: class Program_weight_tensor_parameter_401: name = "parameter_401" + original_name = "conv2d_24.w_0" shape = [128, 512, 1, 1] dtype = "float32" min_val = float("-0.426451") @@ -4334,6 +4736,7 @@ class Program_weight_tensor_parameter_401: class Program_weight_tensor_parameter_402: name = "parameter_402" + original_name = "batch_norm2d_23.b_0" shape = [512] dtype = "float32" min_val = float("-1.74285") @@ -4345,6 +4748,7 @@ class Program_weight_tensor_parameter_402: class Program_weight_tensor_parameter_403: name = "parameter_403" + original_name = "batch_norm2d_23.w_0" shape = [512] dtype = "float32" min_val = float("0.404693") @@ -4356,6 +4760,7 @@ class Program_weight_tensor_parameter_403: class Program_weight_tensor_parameter_404: name = "parameter_404" + original_name = "batch_norm2d_23.w_2" shape = [512] dtype = "float32" min_val = float("0.000129238") @@ -4367,6 +4772,7 @@ class Program_weight_tensor_parameter_404: class Program_weight_tensor_parameter_405: name = "parameter_405" + original_name = "batch_norm2d_23.w_1" shape = [512] dtype = "float32" min_val = float("-1.57023") @@ -4378,6 +4784,7 @@ class Program_weight_tensor_parameter_405: class Program_weight_tensor_parameter_406: name = "parameter_406" + original_name = "conv2d_23.w_0" shape = [512, 1, 5, 5] dtype = "float32" min_val = float("-0.233605") @@ -4389,6 +4796,7 @@ class Program_weight_tensor_parameter_406: class Program_weight_tensor_parameter_407: name = "parameter_407" + original_name = "batch_norm2d_22.b_0" shape = [512] dtype = "float32" min_val = float("-2.48149") @@ -4400,6 +4808,7 @@ class Program_weight_tensor_parameter_407: class Program_weight_tensor_parameter_408: name = "parameter_408" + original_name = "batch_norm2d_22.w_0" shape = [512] dtype = "float32" min_val = float("-0.853402") @@ -4411,6 +4820,7 @@ class Program_weight_tensor_parameter_408: class Program_weight_tensor_parameter_409: name = "parameter_409" + original_name = "batch_norm2d_22.w_2" shape = [512] dtype = "float32" min_val = float("0.326067") @@ -4422,6 +4832,7 @@ class Program_weight_tensor_parameter_409: class Program_weight_tensor_parameter_410: name = "parameter_410" + original_name = "batch_norm2d_22.w_1" shape = [512] dtype = "float32" min_val = float("-3.61779") @@ -4433,6 +4844,7 @@ class Program_weight_tensor_parameter_410: class Program_weight_tensor_parameter_411: name = "parameter_411" + original_name = "conv2d_22.w_0" shape = [512, 512, 1, 1] dtype = "float32" min_val = float("-0.347137") @@ -4444,6 +4856,7 @@ class Program_weight_tensor_parameter_411: class Program_weight_tensor_parameter_412: name = "parameter_412" + original_name = "batch_norm2d_21.b_0" shape = [512] dtype = "float32" min_val = float("-4.77095") @@ -4455,6 +4868,7 @@ class Program_weight_tensor_parameter_412: class Program_weight_tensor_parameter_413: name = "parameter_413" + original_name = "batch_norm2d_21.w_0" shape = [512] dtype = "float32" min_val = float("0.552657") @@ -4466,6 +4880,7 @@ class Program_weight_tensor_parameter_413: class Program_weight_tensor_parameter_414: name = "parameter_414" + original_name = "batch_norm2d_21.w_2" shape = [512] dtype = "float32" min_val = float("5.27582e-05") @@ -4477,6 +4892,7 @@ class Program_weight_tensor_parameter_414: class Program_weight_tensor_parameter_415: name = "parameter_415" + original_name = "batch_norm2d_21.w_1" shape = [512] dtype = "float32" min_val = float("-2.02709") @@ -4488,6 +4904,7 @@ class Program_weight_tensor_parameter_415: class Program_weight_tensor_parameter_416: name = "parameter_416" + original_name = "conv2d_21.w_0" shape = [512, 1, 5, 5] dtype = "float32" min_val = float("-0.601614") @@ -4499,6 +4916,7 @@ class Program_weight_tensor_parameter_416: class Program_weight_tensor_parameter_417: name = "parameter_417" + original_name = "batch_norm2d_20.b_0" shape = [512] dtype = "float32" min_val = float("-3.38228") @@ -4510,6 +4928,7 @@ class Program_weight_tensor_parameter_417: class Program_weight_tensor_parameter_418: name = "parameter_418" + original_name = "batch_norm2d_20.w_0" shape = [512] dtype = "float32" min_val = float("-0.140665") @@ -4521,6 +4940,7 @@ class Program_weight_tensor_parameter_418: class Program_weight_tensor_parameter_419: name = "parameter_419" + original_name = "batch_norm2d_20.w_2" shape = [512] dtype = "float32" min_val = float("0.384093") @@ -4532,6 +4952,7 @@ class Program_weight_tensor_parameter_419: class Program_weight_tensor_parameter_420: name = "parameter_420" + original_name = "batch_norm2d_20.w_1" shape = [512] dtype = "float32" min_val = float("-3.50081") @@ -4543,6 +4964,7 @@ class Program_weight_tensor_parameter_420: class Program_weight_tensor_parameter_421: name = "parameter_421" + original_name = "conv2d_20.w_0" shape = [512, 512, 1, 1] dtype = "float32" min_val = float("-0.428498") @@ -4554,6 +4976,7 @@ class Program_weight_tensor_parameter_421: class Program_weight_tensor_parameter_422: name = "parameter_422" + original_name = "batch_norm2d_19.b_0" shape = [512] dtype = "float32" min_val = float("-3.56486") @@ -4565,6 +4988,7 @@ class Program_weight_tensor_parameter_422: class Program_weight_tensor_parameter_423: name = "parameter_423" + original_name = "batch_norm2d_19.w_0" shape = [512] dtype = "float32" min_val = float("0.45626") @@ -4576,6 +5000,7 @@ class Program_weight_tensor_parameter_423: class Program_weight_tensor_parameter_424: name = "parameter_424" + original_name = "batch_norm2d_19.w_2" shape = [512] dtype = "float32" min_val = float("7.73909e-05") @@ -4587,6 +5012,7 @@ class Program_weight_tensor_parameter_424: class Program_weight_tensor_parameter_425: name = "parameter_425" + original_name = "batch_norm2d_19.w_1" shape = [512] dtype = "float32" min_val = float("-5.80061") @@ -4598,6 +5024,7 @@ class Program_weight_tensor_parameter_425: class Program_weight_tensor_parameter_426: name = "parameter_426" + original_name = "conv2d_19.w_0" shape = [512, 1, 5, 5] dtype = "float32" min_val = float("-0.575256") @@ -4609,6 +5036,7 @@ class Program_weight_tensor_parameter_426: class Program_weight_tensor_parameter_427: name = "parameter_427" + original_name = "batch_norm2d_18.b_0" shape = [512] dtype = "float32" min_val = float("-3.54693") @@ -4620,6 +5048,7 @@ class Program_weight_tensor_parameter_427: class Program_weight_tensor_parameter_428: name = "parameter_428" + original_name = "batch_norm2d_18.w_0" shape = [512] dtype = "float32" min_val = float("0.0327633") @@ -4631,6 +5060,7 @@ class Program_weight_tensor_parameter_428: class Program_weight_tensor_parameter_429: name = "parameter_429" + original_name = "batch_norm2d_18.w_2" shape = [512] dtype = "float32" min_val = float("0.388966") @@ -4642,6 +5072,7 @@ class Program_weight_tensor_parameter_429: class Program_weight_tensor_parameter_430: name = "parameter_430" + original_name = "batch_norm2d_18.w_1" shape = [512] dtype = "float32" min_val = float("-4.17615") @@ -4653,6 +5084,7 @@ class Program_weight_tensor_parameter_430: class Program_weight_tensor_parameter_431: name = "parameter_431" + original_name = "conv2d_18.w_0" shape = [512, 512, 1, 1] dtype = "float32" min_val = float("-0.488104") @@ -4664,6 +5096,7 @@ class Program_weight_tensor_parameter_431: class Program_weight_tensor_parameter_432: name = "parameter_432" + original_name = "batch_norm2d_17.b_0" shape = [512] dtype = "float32" min_val = float("-2.86678") @@ -4675,6 +5108,7 @@ class Program_weight_tensor_parameter_432: class Program_weight_tensor_parameter_433: name = "parameter_433" + original_name = "batch_norm2d_17.w_0" shape = [512] dtype = "float32" min_val = float("0.538258") @@ -4686,6 +5120,7 @@ class Program_weight_tensor_parameter_433: class Program_weight_tensor_parameter_434: name = "parameter_434" + original_name = "batch_norm2d_17.w_2" shape = [512] dtype = "float32" min_val = float("0.000102648") @@ -4697,6 +5132,7 @@ class Program_weight_tensor_parameter_434: class Program_weight_tensor_parameter_435: name = "parameter_435" + original_name = "batch_norm2d_17.w_1" shape = [512] dtype = "float32" min_val = float("-5.27332") @@ -4708,6 +5144,7 @@ class Program_weight_tensor_parameter_435: class Program_weight_tensor_parameter_436: name = "parameter_436" + original_name = "conv2d_17.w_0" shape = [512, 1, 5, 5] dtype = "float32" min_val = float("-0.570171") @@ -4719,6 +5156,7 @@ class Program_weight_tensor_parameter_436: class Program_weight_tensor_parameter_437: name = "parameter_437" + original_name = "batch_norm2d_16.b_0" shape = [512] dtype = "float32" min_val = float("-3.34727") @@ -4730,6 +5168,7 @@ class Program_weight_tensor_parameter_437: class Program_weight_tensor_parameter_438: name = "parameter_438" + original_name = "batch_norm2d_16.w_0" shape = [512] dtype = "float32" min_val = float("0.0455797") @@ -4741,6 +5180,7 @@ class Program_weight_tensor_parameter_438: class Program_weight_tensor_parameter_439: name = "parameter_439" + original_name = "batch_norm2d_16.w_2" shape = [512] dtype = "float32" min_val = float("0.398938") @@ -4752,6 +5192,7 @@ class Program_weight_tensor_parameter_439: class Program_weight_tensor_parameter_440: name = "parameter_440" + original_name = "batch_norm2d_16.w_1" shape = [512] dtype = "float32" min_val = float("-3.77985") @@ -4763,6 +5204,7 @@ class Program_weight_tensor_parameter_440: class Program_weight_tensor_parameter_441: name = "parameter_441" + original_name = "conv2d_16.w_0" shape = [512, 512, 1, 1] dtype = "float32" min_val = float("-0.504453") @@ -4774,6 +5216,7 @@ class Program_weight_tensor_parameter_441: class Program_weight_tensor_parameter_442: name = "parameter_442" + original_name = "batch_norm2d_15.b_0" shape = [512] dtype = "float32" min_val = float("-3.64556") @@ -4785,6 +5228,7 @@ class Program_weight_tensor_parameter_442: class Program_weight_tensor_parameter_443: name = "parameter_443" + original_name = "batch_norm2d_15.w_0" shape = [512] dtype = "float32" min_val = float("0.566612") @@ -4796,6 +5240,7 @@ class Program_weight_tensor_parameter_443: class Program_weight_tensor_parameter_444: name = "parameter_444" + original_name = "batch_norm2d_15.w_2" shape = [512] dtype = "float32" min_val = float("0.000178209") @@ -4807,6 +5252,7 @@ class Program_weight_tensor_parameter_444: class Program_weight_tensor_parameter_445: name = "parameter_445" + original_name = "batch_norm2d_15.w_1" shape = [512] dtype = "float32" min_val = float("-1.99605") @@ -4818,6 +5264,7 @@ class Program_weight_tensor_parameter_445: class Program_weight_tensor_parameter_446: name = "parameter_446" + original_name = "conv2d_15.w_0" shape = [512, 1, 5, 5] dtype = "float32" min_val = float("-0.502033") @@ -4829,6 +5276,7 @@ class Program_weight_tensor_parameter_446: class Program_weight_tensor_parameter_447: name = "parameter_447" + original_name = "batch_norm2d_14.b_0" shape = [512] dtype = "float32" min_val = float("-2.64711") @@ -4840,6 +5288,7 @@ class Program_weight_tensor_parameter_447: class Program_weight_tensor_parameter_448: name = "parameter_448" + original_name = "batch_norm2d_14.w_0" shape = [512] dtype = "float32" min_val = float("0.0564457") @@ -4851,6 +5300,7 @@ class Program_weight_tensor_parameter_448: class Program_weight_tensor_parameter_449: name = "parameter_449" + original_name = "batch_norm2d_14.w_2" shape = [512] dtype = "float32" min_val = float("0.466178") @@ -4862,6 +5312,7 @@ class Program_weight_tensor_parameter_449: class Program_weight_tensor_parameter_450: name = "parameter_450" + original_name = "batch_norm2d_14.w_1" shape = [512] dtype = "float32" min_val = float("-3.80985") @@ -4873,6 +5324,7 @@ class Program_weight_tensor_parameter_450: class Program_weight_tensor_parameter_451: name = "parameter_451" + original_name = "conv2d_14.w_0" shape = [512, 512, 1, 1] dtype = "float32" min_val = float("-0.517498") @@ -4884,6 +5336,7 @@ class Program_weight_tensor_parameter_451: class Program_weight_tensor_parameter_452: name = "parameter_452" + original_name = "batch_norm2d_13.b_0" shape = [512] dtype = "float32" min_val = float("-6.0471") @@ -4895,6 +5348,7 @@ class Program_weight_tensor_parameter_452: class Program_weight_tensor_parameter_453: name = "parameter_453" + original_name = "batch_norm2d_13.w_0" shape = [512] dtype = "float32" min_val = float("0.50849") @@ -4906,6 +5360,7 @@ class Program_weight_tensor_parameter_453: class Program_weight_tensor_parameter_454: name = "parameter_454" + original_name = "batch_norm2d_13.w_2" shape = [512] dtype = "float32" min_val = float("0.000132926") @@ -4917,6 +5372,7 @@ class Program_weight_tensor_parameter_454: class Program_weight_tensor_parameter_455: name = "parameter_455" + original_name = "batch_norm2d_13.w_1" shape = [512] dtype = "float32" min_val = float("-1.91192") @@ -4928,6 +5384,7 @@ class Program_weight_tensor_parameter_455: class Program_weight_tensor_parameter_456: name = "parameter_456" + original_name = "conv2d_13.w_0" shape = [512, 1, 5, 5] dtype = "float32" min_val = float("-0.585066") @@ -4939,6 +5396,7 @@ class Program_weight_tensor_parameter_456: class Program_weight_tensor_parameter_457: name = "parameter_457" + original_name = "batch_norm2d_12.b_0" shape = [512] dtype = "float32" min_val = float("-1.90384") @@ -4950,6 +5408,7 @@ class Program_weight_tensor_parameter_457: class Program_weight_tensor_parameter_458: name = "parameter_458" + original_name = "batch_norm2d_12.w_0" shape = [512] dtype = "float32" min_val = float("-0.15877") @@ -4961,6 +5420,7 @@ class Program_weight_tensor_parameter_458: class Program_weight_tensor_parameter_459: name = "parameter_459" + original_name = "batch_norm2d_12.w_2" shape = [512] dtype = "float32" min_val = float("0.188812") @@ -4972,6 +5432,7 @@ class Program_weight_tensor_parameter_459: class Program_weight_tensor_parameter_460: name = "parameter_460" + original_name = "batch_norm2d_12.w_1" shape = [512] dtype = "float32" min_val = float("-4.64945") @@ -4983,6 +5444,7 @@ class Program_weight_tensor_parameter_460: class Program_weight_tensor_parameter_461: name = "parameter_461" + original_name = "conv2d_12.w_0" shape = [512, 256, 1, 1] dtype = "float32" min_val = float("-0.509109") @@ -4994,6 +5456,7 @@ class Program_weight_tensor_parameter_461: class Program_weight_tensor_parameter_462: name = "parameter_462" + original_name = "batch_norm2d_11.b_0" shape = [256] dtype = "float32" min_val = float("-2.05957") @@ -5005,6 +5468,7 @@ class Program_weight_tensor_parameter_462: class Program_weight_tensor_parameter_463: name = "parameter_463" + original_name = "batch_norm2d_11.w_0" shape = [256] dtype = "float32" min_val = float("0.497247") @@ -5016,6 +5480,7 @@ class Program_weight_tensor_parameter_463: class Program_weight_tensor_parameter_464: name = "parameter_464" + original_name = "batch_norm2d_11.w_2" shape = [256] dtype = "float32" min_val = float("0.000182634") @@ -5027,6 +5492,7 @@ class Program_weight_tensor_parameter_464: class Program_weight_tensor_parameter_465: name = "parameter_465" + original_name = "batch_norm2d_11.w_1" shape = [256] dtype = "float32" min_val = float("-1.50292") @@ -5038,6 +5504,7 @@ class Program_weight_tensor_parameter_465: class Program_weight_tensor_parameter_466: name = "parameter_466" + original_name = "conv2d_11.w_0" shape = [256, 1, 3, 3] dtype = "float32" min_val = float("-0.269443") @@ -5049,6 +5516,7 @@ class Program_weight_tensor_parameter_466: class Program_weight_tensor_parameter_467: name = "parameter_467" + original_name = "batch_norm2d_10.b_0" shape = [256] dtype = "float32" min_val = float("-2.84023") @@ -5060,6 +5528,7 @@ class Program_weight_tensor_parameter_467: class Program_weight_tensor_parameter_468: name = "parameter_468" + original_name = "batch_norm2d_10.w_0" shape = [256] dtype = "float32" min_val = float("0.059048") @@ -5071,6 +5540,7 @@ class Program_weight_tensor_parameter_468: class Program_weight_tensor_parameter_469: name = "parameter_469" + original_name = "batch_norm2d_10.w_2" shape = [256] dtype = "float32" min_val = float("0.183628") @@ -5082,6 +5552,7 @@ class Program_weight_tensor_parameter_469: class Program_weight_tensor_parameter_470: name = "parameter_470" + original_name = "batch_norm2d_10.w_1" shape = [256] dtype = "float32" min_val = float("-4.02082") @@ -5093,6 +5564,7 @@ class Program_weight_tensor_parameter_470: class Program_weight_tensor_parameter_471: name = "parameter_471" + original_name = "conv2d_10.w_0" shape = [256, 256, 1, 1] dtype = "float32" min_val = float("-0.460632") @@ -5104,6 +5576,7 @@ class Program_weight_tensor_parameter_471: class Program_weight_tensor_parameter_472: name = "parameter_472" + original_name = "batch_norm2d_9.b_0" shape = [256] dtype = "float32" min_val = float("-2.07573") @@ -5115,6 +5588,7 @@ class Program_weight_tensor_parameter_472: class Program_weight_tensor_parameter_473: name = "parameter_473" + original_name = "batch_norm2d_9.w_0" shape = [256] dtype = "float32" min_val = float("0.369252") @@ -5126,6 +5600,7 @@ class Program_weight_tensor_parameter_473: class Program_weight_tensor_parameter_474: name = "parameter_474" + original_name = "batch_norm2d_9.w_2" shape = [256] dtype = "float32" min_val = float("4.68259e-05") @@ -5137,6 +5612,7 @@ class Program_weight_tensor_parameter_474: class Program_weight_tensor_parameter_475: name = "parameter_475" + original_name = "batch_norm2d_9.w_1" shape = [256] dtype = "float32" min_val = float("-4.35021") @@ -5148,6 +5624,7 @@ class Program_weight_tensor_parameter_475: class Program_weight_tensor_parameter_476: name = "parameter_476" + original_name = "conv2d_9.w_0" shape = [256, 1, 3, 3] dtype = "float32" min_val = float("-0.661158") @@ -5159,6 +5636,7 @@ class Program_weight_tensor_parameter_476: class Program_weight_tensor_parameter_477: name = "parameter_477" + original_name = "batch_norm2d_8.b_0" shape = [256] dtype = "float32" min_val = float("-1.84558") @@ -5170,6 +5648,7 @@ class Program_weight_tensor_parameter_477: class Program_weight_tensor_parameter_478: name = "parameter_478" + original_name = "batch_norm2d_8.w_0" shape = [256] dtype = "float32" min_val = float("-0.52954") @@ -5181,6 +5660,7 @@ class Program_weight_tensor_parameter_478: class Program_weight_tensor_parameter_479: name = "parameter_479" + original_name = "batch_norm2d_8.w_2" shape = [256] dtype = "float32" min_val = float("0.19213") @@ -5192,6 +5672,7 @@ class Program_weight_tensor_parameter_479: class Program_weight_tensor_parameter_480: name = "parameter_480" + original_name = "batch_norm2d_8.w_1" shape = [256] dtype = "float32" min_val = float("-6.54047") @@ -5203,6 +5684,7 @@ class Program_weight_tensor_parameter_480: class Program_weight_tensor_parameter_481: name = "parameter_481" + original_name = "conv2d_8.w_0" shape = [256, 128, 1, 1] dtype = "float32" min_val = float("-0.823371") @@ -5214,6 +5696,7 @@ class Program_weight_tensor_parameter_481: class Program_weight_tensor_parameter_482: name = "parameter_482" + original_name = "batch_norm2d_7.b_0" shape = [128] dtype = "float32" min_val = float("-4.42021") @@ -5225,6 +5708,7 @@ class Program_weight_tensor_parameter_482: class Program_weight_tensor_parameter_483: name = "parameter_483" + original_name = "batch_norm2d_7.w_0" shape = [128] dtype = "float32" min_val = float("0.522047") @@ -5236,6 +5720,7 @@ class Program_weight_tensor_parameter_483: class Program_weight_tensor_parameter_484: name = "parameter_484" + original_name = "batch_norm2d_7.w_2" shape = [128] dtype = "float32" min_val = float("7.05128e-05") @@ -5247,6 +5732,7 @@ class Program_weight_tensor_parameter_484: class Program_weight_tensor_parameter_485: name = "parameter_485" + original_name = "batch_norm2d_7.w_1" shape = [128] dtype = "float32" min_val = float("-1.87888") @@ -5258,6 +5744,7 @@ class Program_weight_tensor_parameter_485: class Program_weight_tensor_parameter_486: name = "parameter_486" + original_name = "conv2d_7.w_0" shape = [128, 1, 3, 3] dtype = "float32" min_val = float("-0.339252") @@ -5269,6 +5756,7 @@ class Program_weight_tensor_parameter_486: class Program_weight_tensor_parameter_487: name = "parameter_487" + original_name = "batch_norm2d_6.b_0" shape = [128] dtype = "float32" min_val = float("-2.29716") @@ -5280,6 +5768,7 @@ class Program_weight_tensor_parameter_487: class Program_weight_tensor_parameter_488: name = "parameter_488" + original_name = "batch_norm2d_6.w_0" shape = [128] dtype = "float32" min_val = float("-0.768895") @@ -5291,6 +5780,7 @@ class Program_weight_tensor_parameter_488: class Program_weight_tensor_parameter_489: name = "parameter_489" + original_name = "batch_norm2d_6.w_2" shape = [128] dtype = "float32" min_val = float("1.04615") @@ -5302,6 +5792,7 @@ class Program_weight_tensor_parameter_489: class Program_weight_tensor_parameter_490: name = "parameter_490" + original_name = "batch_norm2d_6.w_1" shape = [128] dtype = "float32" min_val = float("-6.41995") @@ -5313,6 +5804,7 @@ class Program_weight_tensor_parameter_490: class Program_weight_tensor_parameter_491: name = "parameter_491" + original_name = "conv2d_6.w_0" shape = [128, 128, 1, 1] dtype = "float32" min_val = float("-1.05901") @@ -5324,6 +5816,7 @@ class Program_weight_tensor_parameter_491: class Program_weight_tensor_parameter_492: name = "parameter_492" + original_name = "batch_norm2d_5.b_0" shape = [128] dtype = "float32" min_val = float("-1.45664") @@ -5335,6 +5828,7 @@ class Program_weight_tensor_parameter_492: class Program_weight_tensor_parameter_493: name = "parameter_493" + original_name = "batch_norm2d_5.w_0" shape = [128] dtype = "float32" min_val = float("0.929322") @@ -5346,6 +5840,7 @@ class Program_weight_tensor_parameter_493: class Program_weight_tensor_parameter_494: name = "parameter_494" + original_name = "batch_norm2d_5.w_2" shape = [128] dtype = "float32" min_val = float("0.000392915") @@ -5357,6 +5852,7 @@ class Program_weight_tensor_parameter_494: class Program_weight_tensor_parameter_495: name = "parameter_495" + original_name = "batch_norm2d_5.w_1" shape = [128] dtype = "float32" min_val = float("-3.14563") @@ -5368,6 +5864,7 @@ class Program_weight_tensor_parameter_495: class Program_weight_tensor_parameter_496: name = "parameter_496" + original_name = "conv2d_5.w_0" shape = [128, 1, 3, 3] dtype = "float32" min_val = float("-1.33615") @@ -5379,6 +5876,7 @@ class Program_weight_tensor_parameter_496: class Program_weight_tensor_parameter_497: name = "parameter_497" + original_name = "batch_norm2d_4.b_0" shape = [128] dtype = "float32" min_val = float("-3.68101") @@ -5390,6 +5888,7 @@ class Program_weight_tensor_parameter_497: class Program_weight_tensor_parameter_498: name = "parameter_498" + original_name = "batch_norm2d_4.w_0" shape = [128] dtype = "float32" min_val = float("-0.323351") @@ -5401,6 +5900,7 @@ class Program_weight_tensor_parameter_498: class Program_weight_tensor_parameter_499: name = "parameter_499" + original_name = "batch_norm2d_4.w_2" shape = [128] dtype = "float32" min_val = float("0.882378") @@ -5412,6 +5912,7 @@ class Program_weight_tensor_parameter_499: class Program_weight_tensor_parameter_500: name = "parameter_500" + original_name = "batch_norm2d_4.w_1" shape = [128] dtype = "float32" min_val = float("-10.0289") @@ -5423,6 +5924,7 @@ class Program_weight_tensor_parameter_500: class Program_weight_tensor_parameter_501: name = "parameter_501" + original_name = "conv2d_4.w_0" shape = [128, 64, 1, 1] dtype = "float32" min_val = float("-0.803442") @@ -5434,6 +5936,7 @@ class Program_weight_tensor_parameter_501: class Program_weight_tensor_parameter_502: name = "parameter_502" + original_name = "batch_norm2d_3.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -5443,6 +5946,7 @@ class Program_weight_tensor_parameter_502: class Program_weight_tensor_parameter_503: name = "parameter_503" + original_name = "batch_norm2d_3.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -5452,6 +5956,7 @@ class Program_weight_tensor_parameter_503: class Program_weight_tensor_parameter_504: name = "parameter_504" + original_name = "batch_norm2d_3.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -5461,6 +5966,7 @@ class Program_weight_tensor_parameter_504: class Program_weight_tensor_parameter_505: name = "parameter_505" + original_name = "batch_norm2d_3.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -5470,6 +5976,7 @@ class Program_weight_tensor_parameter_505: class Program_weight_tensor_parameter_506: name = "parameter_506" + original_name = "conv2d_3.w_0" shape = [64, 1, 3, 3] dtype = "float32" min_val = float("-0.352457") @@ -5481,6 +5988,7 @@ class Program_weight_tensor_parameter_506: class Program_weight_tensor_parameter_507: name = "parameter_507" + original_name = "batch_norm2d_2.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -5490,6 +5998,7 @@ class Program_weight_tensor_parameter_507: class Program_weight_tensor_parameter_508: name = "parameter_508" + original_name = "batch_norm2d_2.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -5499,6 +6008,7 @@ class Program_weight_tensor_parameter_508: class Program_weight_tensor_parameter_509: name = "parameter_509" + original_name = "batch_norm2d_2.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -5508,6 +6018,7 @@ class Program_weight_tensor_parameter_509: class Program_weight_tensor_parameter_510: name = "parameter_510" + original_name = "batch_norm2d_2.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -5517,6 +6028,7 @@ class Program_weight_tensor_parameter_510: class Program_weight_tensor_parameter_511: name = "parameter_511" + original_name = "conv2d_2.w_0" shape = [64, 32, 1, 1] dtype = "float32" min_val = float("-0.93893") @@ -5528,6 +6040,7 @@ class Program_weight_tensor_parameter_511: class Program_weight_tensor_parameter_512: name = "parameter_512" + original_name = "batch_norm2d_1.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -5537,6 +6050,7 @@ class Program_weight_tensor_parameter_512: class Program_weight_tensor_parameter_513: name = "parameter_513" + original_name = "batch_norm2d_1.w_0" shape = [32] dtype = "float32" min_val = float("0") @@ -5546,6 +6060,7 @@ class Program_weight_tensor_parameter_513: class Program_weight_tensor_parameter_514: name = "parameter_514" + original_name = "batch_norm2d_1.w_2" shape = [32] dtype = "float32" min_val = float("0") @@ -5555,6 +6070,7 @@ class Program_weight_tensor_parameter_514: class Program_weight_tensor_parameter_515: name = "parameter_515" + original_name = "batch_norm2d_1.w_1" shape = [32] dtype = "float32" min_val = float("0") @@ -5564,6 +6080,7 @@ class Program_weight_tensor_parameter_515: class Program_weight_tensor_parameter_516: name = "parameter_516" + original_name = "conv2d_1.w_0" shape = [32, 1, 3, 3] dtype = "float32" min_val = float("-1.33777") @@ -5575,6 +6092,7 @@ class Program_weight_tensor_parameter_516: class Program_weight_tensor_parameter_517: name = "parameter_517" + original_name = "batch_norm2d_0.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -5584,6 +6102,7 @@ class Program_weight_tensor_parameter_517: class Program_weight_tensor_parameter_518: name = "parameter_518" + original_name = "batch_norm2d_0.w_0" shape = [32] dtype = "float32" min_val = float("0") @@ -5593,6 +6112,7 @@ class Program_weight_tensor_parameter_518: class Program_weight_tensor_parameter_519: name = "parameter_519" + original_name = "batch_norm2d_0.w_2" shape = [32] dtype = "float32" min_val = float("0") @@ -5602,6 +6122,7 @@ class Program_weight_tensor_parameter_519: class Program_weight_tensor_parameter_520: name = "parameter_520" + original_name = "batch_norm2d_0.w_1" shape = [32] dtype = "float32" min_val = float("0") @@ -5611,6 +6132,7 @@ class Program_weight_tensor_parameter_520: class Program_weight_tensor_parameter_521: name = "parameter_521" + original_name = "conv2d_0.w_0" shape = [32, 3, 3, 3] dtype = "float32" min_val = float("-0.587999") diff --git a/paddle_samples/PaddleX/PicoDet-S/subgraph_12/input_meta.py b/paddle_samples/PaddleX/PicoDet-S/subgraph_12/input_meta.py index 0cd0693dc..11364565b 100644 --- a/paddle_samples/PaddleX/PicoDet-S/subgraph_12/input_meta.py +++ b/paddle_samples/PaddleX/PicoDet-S/subgraph_12/input_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_data_0: name = "data_0" + original_name = "var_2146" shape = [16, 3, 288, 288] dtype = "float32" min_val = float("-2.67097") @@ -11,6 +12,7 @@ class Program_weight_tensor_data_0: class Program_weight_tensor_data_1: name = "data_1" + original_name = "var_2387" shape = [8] dtype = "float32" data = [0.0, 1.0, 2.0, 3.0003, 4.0, 5.0, 6.0006, 7.0] diff --git a/paddle_samples/PaddleX/PicoDet-S/subgraph_12/weight_meta.py b/paddle_samples/PaddleX/PicoDet-S/subgraph_12/weight_meta.py index af5b40bcd..24bdeb113 100644 --- a/paddle_samples/PaddleX/PicoDet-S/subgraph_12/weight_meta.py +++ b/paddle_samples/PaddleX/PicoDet-S/subgraph_12/weight_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_parameter_0: name = "parameter_0" + original_name = "batch_norm2d_81.b_0" shape = [1] dtype = "float32" min_val = float("0") @@ -9,6 +10,7 @@ class Program_weight_tensor_parameter_0: class Program_weight_tensor_parameter_1: name = "parameter_1" + original_name = "batch_norm2d_81.w_0" shape = [1] dtype = "float32" min_val = float("0") @@ -18,6 +20,7 @@ class Program_weight_tensor_parameter_1: class Program_weight_tensor_parameter_2: name = "parameter_2" + original_name = "batch_norm2d_81.w_2" shape = [1] dtype = "float32" min_val = float("0") @@ -27,6 +30,7 @@ class Program_weight_tensor_parameter_2: class Program_weight_tensor_parameter_3: name = "parameter_3" + original_name = "batch_norm2d_81.w_1" shape = [1] dtype = "float32" min_val = float("0") @@ -36,6 +40,7 @@ class Program_weight_tensor_parameter_3: class Program_weight_tensor_parameter_4: name = "parameter_4" + original_name = "conv2d_99.w_0" shape = [1, 1, 1, 1] dtype = "float32" min_val = float("0") @@ -45,6 +50,7 @@ class Program_weight_tensor_parameter_4: class Program_weight_tensor_parameter_5: name = "parameter_5" + original_name = "batch_norm2d_80.b_0" shape = [1] dtype = "float32" min_val = float("0") @@ -54,6 +60,7 @@ class Program_weight_tensor_parameter_5: class Program_weight_tensor_parameter_6: name = "parameter_6" + original_name = "batch_norm2d_80.w_0" shape = [1] dtype = "float32" min_val = float("0") @@ -63,6 +70,7 @@ class Program_weight_tensor_parameter_6: class Program_weight_tensor_parameter_7: name = "parameter_7" + original_name = "batch_norm2d_80.w_2" shape = [1] dtype = "float32" min_val = float("0") @@ -72,6 +80,7 @@ class Program_weight_tensor_parameter_7: class Program_weight_tensor_parameter_8: name = "parameter_8" + original_name = "batch_norm2d_80.w_1" shape = [1] dtype = "float32" min_val = float("0") @@ -81,6 +90,7 @@ class Program_weight_tensor_parameter_8: class Program_weight_tensor_parameter_9: name = "parameter_9" + original_name = "conv2d_98.w_0" shape = [1, 96, 5, 5] dtype = "float32" min_val = float("-0.269445") @@ -92,6 +102,7 @@ class Program_weight_tensor_parameter_9: class Program_weight_tensor_parameter_10: name = "parameter_10" + original_name = "conv2d_97.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -101,6 +112,7 @@ class Program_weight_tensor_parameter_10: class Program_weight_tensor_parameter_11: name = "parameter_11" + original_name = "conv2d_97.w_0" shape = [32, 96, 1, 1] dtype = "float32" min_val = float("-0.482696") @@ -112,6 +124,7 @@ class Program_weight_tensor_parameter_11: class Program_weight_tensor_parameter_12: name = "parameter_12" + original_name = "conv2d_96.b_0" shape = [4] dtype = "float32" min_val = float("0") @@ -121,6 +134,7 @@ class Program_weight_tensor_parameter_12: class Program_weight_tensor_parameter_13: name = "parameter_13" + original_name = "conv2d_96.w_0" shape = [4, 96, 1, 1] dtype = "float32" min_val = float("-0.026241") @@ -132,6 +146,7 @@ class Program_weight_tensor_parameter_13: class Program_weight_tensor_parameter_14: name = "parameter_14" + original_name = "batch_norm2d_73.b_0" shape = [96] dtype = "float32" min_val = float("-0.0386539") @@ -143,6 +158,7 @@ class Program_weight_tensor_parameter_14: class Program_weight_tensor_parameter_15: name = "parameter_15" + original_name = "batch_norm2d_73.w_0" shape = [96] dtype = "float32" min_val = float("0.828459") @@ -154,6 +170,7 @@ class Program_weight_tensor_parameter_15: class Program_weight_tensor_parameter_16: name = "parameter_16" + original_name = "batch_norm2d_73.w_2" shape = [96] dtype = "float32" min_val = float("0.0041083") @@ -165,6 +182,7 @@ class Program_weight_tensor_parameter_16: class Program_weight_tensor_parameter_17: name = "parameter_17" + original_name = "batch_norm2d_73.w_1" shape = [96] dtype = "float32" min_val = float("-0.135119") @@ -176,6 +194,7 @@ class Program_weight_tensor_parameter_17: class Program_weight_tensor_parameter_18: name = "parameter_18" + original_name = "conv2d_81.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.23216") @@ -187,6 +206,7 @@ class Program_weight_tensor_parameter_18: class Program_weight_tensor_parameter_19: name = "parameter_19" + original_name = "conv2d_80.b_0" shape = [96] dtype = "float32" min_val = float("-0.0389156") @@ -198,6 +218,7 @@ class Program_weight_tensor_parameter_19: class Program_weight_tensor_parameter_20: name = "parameter_20" + original_name = "conv2d_80.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.161448") @@ -209,6 +230,7 @@ class Program_weight_tensor_parameter_20: class Program_weight_tensor_parameter_21: name = "parameter_21" + original_name = "batch_norm2d_72.b_0" shape = [96] dtype = "float32" min_val = float("-1.39206") @@ -220,6 +242,7 @@ class Program_weight_tensor_parameter_21: class Program_weight_tensor_parameter_22: name = "parameter_22" + original_name = "batch_norm2d_72.w_0" shape = [96] dtype = "float32" min_val = float("0.579293") @@ -231,6 +254,7 @@ class Program_weight_tensor_parameter_22: class Program_weight_tensor_parameter_23: name = "parameter_23" + original_name = "batch_norm2d_72.w_2" shape = [96] dtype = "float32" min_val = float("0.0255919") @@ -242,6 +266,7 @@ class Program_weight_tensor_parameter_23: class Program_weight_tensor_parameter_24: name = "parameter_24" + original_name = "batch_norm2d_72.w_1" shape = [96] dtype = "float32" min_val = float("-0.627332") @@ -253,6 +278,7 @@ class Program_weight_tensor_parameter_24: class Program_weight_tensor_parameter_25: name = "parameter_25" + original_name = "conv2d_79.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.331948") @@ -264,6 +290,7 @@ class Program_weight_tensor_parameter_25: class Program_weight_tensor_parameter_26: name = "parameter_26" + original_name = "batch_norm2d_71.b_0" shape = [96] dtype = "float32" min_val = float("-0.930941") @@ -275,6 +302,7 @@ class Program_weight_tensor_parameter_26: class Program_weight_tensor_parameter_27: name = "parameter_27" + original_name = "batch_norm2d_71.w_0" shape = [96] dtype = "float32" min_val = float("0.534449") @@ -286,6 +314,7 @@ class Program_weight_tensor_parameter_27: class Program_weight_tensor_parameter_28: name = "parameter_28" + original_name = "batch_norm2d_71.w_2" shape = [96] dtype = "float32" min_val = float("0.0017011") @@ -297,6 +326,7 @@ class Program_weight_tensor_parameter_28: class Program_weight_tensor_parameter_29: name = "parameter_29" + original_name = "batch_norm2d_71.w_1" shape = [96] dtype = "float32" min_val = float("-0.167339") @@ -308,6 +338,7 @@ class Program_weight_tensor_parameter_29: class Program_weight_tensor_parameter_30: name = "parameter_30" + original_name = "conv2d_78.w_0" shape = [96, 1, 5, 5] dtype = "float32" min_val = float("-0.150452") @@ -319,6 +350,7 @@ class Program_weight_tensor_parameter_30: class Program_weight_tensor_parameter_31: name = "parameter_31" + original_name = "batch_norm2d_70.b_0" shape = [96] dtype = "float32" min_val = float("-0.97352") @@ -330,6 +362,7 @@ class Program_weight_tensor_parameter_31: class Program_weight_tensor_parameter_32: name = "parameter_32" + original_name = "batch_norm2d_70.w_0" shape = [96] dtype = "float32" min_val = float("0.21796") @@ -341,6 +374,7 @@ class Program_weight_tensor_parameter_32: class Program_weight_tensor_parameter_33: name = "parameter_33" + original_name = "batch_norm2d_70.w_2" shape = [96] dtype = "float32" min_val = float("0.0234018") @@ -352,6 +386,7 @@ class Program_weight_tensor_parameter_33: class Program_weight_tensor_parameter_34: name = "parameter_34" + original_name = "batch_norm2d_70.w_1" shape = [96] dtype = "float32" min_val = float("-0.267") @@ -363,6 +398,7 @@ class Program_weight_tensor_parameter_34: class Program_weight_tensor_parameter_35: name = "parameter_35" + original_name = "conv2d_77.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.258249") @@ -374,6 +410,7 @@ class Program_weight_tensor_parameter_35: class Program_weight_tensor_parameter_36: name = "parameter_36" + original_name = "batch_norm2d_69.b_0" shape = [96] dtype = "float32" min_val = float("-0.708856") @@ -385,6 +422,7 @@ class Program_weight_tensor_parameter_36: class Program_weight_tensor_parameter_37: name = "parameter_37" + original_name = "batch_norm2d_69.w_0" shape = [96] dtype = "float32" min_val = float("0.821682") @@ -396,6 +434,7 @@ class Program_weight_tensor_parameter_37: class Program_weight_tensor_parameter_38: name = "parameter_38" + original_name = "batch_norm2d_69.w_2" shape = [96] dtype = "float32" min_val = float("4.76891e-05") @@ -407,6 +446,7 @@ class Program_weight_tensor_parameter_38: class Program_weight_tensor_parameter_39: name = "parameter_39" + original_name = "batch_norm2d_69.w_1" shape = [96] dtype = "float32" min_val = float("-0.0278458") @@ -418,6 +458,7 @@ class Program_weight_tensor_parameter_39: class Program_weight_tensor_parameter_40: name = "parameter_40" + original_name = "conv2d_76.w_0" shape = [96, 1, 5, 5] dtype = "float32" min_val = float("-0.286588") @@ -429,6 +470,7 @@ class Program_weight_tensor_parameter_40: class Program_weight_tensor_parameter_41: name = "parameter_41" + original_name = "batch_norm2d_79.b_0" shape = [1] dtype = "float32" min_val = float("0") @@ -438,6 +480,7 @@ class Program_weight_tensor_parameter_41: class Program_weight_tensor_parameter_42: name = "parameter_42" + original_name = "batch_norm2d_79.w_0" shape = [1] dtype = "float32" min_val = float("0") @@ -447,6 +490,7 @@ class Program_weight_tensor_parameter_42: class Program_weight_tensor_parameter_43: name = "parameter_43" + original_name = "batch_norm2d_79.w_2" shape = [1] dtype = "float32" min_val = float("0") @@ -456,6 +500,7 @@ class Program_weight_tensor_parameter_43: class Program_weight_tensor_parameter_44: name = "parameter_44" + original_name = "batch_norm2d_79.w_1" shape = [1] dtype = "float32" min_val = float("0") @@ -465,6 +510,7 @@ class Program_weight_tensor_parameter_44: class Program_weight_tensor_parameter_45: name = "parameter_45" + original_name = "conv2d_95.w_0" shape = [1, 1, 1, 1] dtype = "float32" min_val = float("0") @@ -474,6 +520,7 @@ class Program_weight_tensor_parameter_45: class Program_weight_tensor_parameter_46: name = "parameter_46" + original_name = "batch_norm2d_78.b_0" shape = [1] dtype = "float32" min_val = float("0") @@ -483,6 +530,7 @@ class Program_weight_tensor_parameter_46: class Program_weight_tensor_parameter_47: name = "parameter_47" + original_name = "batch_norm2d_78.w_0" shape = [1] dtype = "float32" min_val = float("0") @@ -492,6 +540,7 @@ class Program_weight_tensor_parameter_47: class Program_weight_tensor_parameter_48: name = "parameter_48" + original_name = "batch_norm2d_78.w_2" shape = [1] dtype = "float32" min_val = float("0") @@ -501,6 +550,7 @@ class Program_weight_tensor_parameter_48: class Program_weight_tensor_parameter_49: name = "parameter_49" + original_name = "batch_norm2d_78.w_1" shape = [1] dtype = "float32" min_val = float("0") @@ -510,6 +560,7 @@ class Program_weight_tensor_parameter_49: class Program_weight_tensor_parameter_50: name = "parameter_50" + original_name = "conv2d_94.w_0" shape = [1, 96, 5, 5] dtype = "float32" min_val = float("-0.738362") @@ -521,6 +572,7 @@ class Program_weight_tensor_parameter_50: class Program_weight_tensor_parameter_51: name = "parameter_51" + original_name = "conv2d_93.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -530,6 +582,7 @@ class Program_weight_tensor_parameter_51: class Program_weight_tensor_parameter_52: name = "parameter_52" + original_name = "conv2d_93.w_0" shape = [32, 96, 1, 1] dtype = "float32" min_val = float("-0.59612") @@ -541,6 +594,7 @@ class Program_weight_tensor_parameter_52: class Program_weight_tensor_parameter_53: name = "parameter_53" + original_name = "conv2d_92.b_0" shape = [4] dtype = "float32" min_val = float("0") @@ -550,6 +604,7 @@ class Program_weight_tensor_parameter_53: class Program_weight_tensor_parameter_54: name = "parameter_54" + original_name = "conv2d_92.w_0" shape = [4, 96, 1, 1] dtype = "float32" min_val = float("-0.0255437") @@ -561,6 +616,7 @@ class Program_weight_tensor_parameter_54: class Program_weight_tensor_parameter_55: name = "parameter_55" + original_name = "batch_norm2d_68.b_0" shape = [96] dtype = "float32" min_val = float("-0.0989313") @@ -572,6 +628,7 @@ class Program_weight_tensor_parameter_55: class Program_weight_tensor_parameter_56: name = "parameter_56" + original_name = "batch_norm2d_68.w_0" shape = [96] dtype = "float32" min_val = float("0.609421") @@ -583,6 +640,7 @@ class Program_weight_tensor_parameter_56: class Program_weight_tensor_parameter_57: name = "parameter_57" + original_name = "batch_norm2d_68.w_2" shape = [96] dtype = "float32" min_val = float("0.00821664") @@ -594,6 +652,7 @@ class Program_weight_tensor_parameter_57: class Program_weight_tensor_parameter_58: name = "parameter_58" + original_name = "batch_norm2d_68.w_1" shape = [96] dtype = "float32" min_val = float("-0.221309") @@ -605,6 +664,7 @@ class Program_weight_tensor_parameter_58: class Program_weight_tensor_parameter_59: name = "parameter_59" + original_name = "conv2d_75.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.432803") @@ -616,6 +676,7 @@ class Program_weight_tensor_parameter_59: class Program_weight_tensor_parameter_60: name = "parameter_60" + original_name = "conv2d_74.b_0" shape = [96] dtype = "float32" min_val = float("-0.0660739") @@ -627,6 +688,7 @@ class Program_weight_tensor_parameter_60: class Program_weight_tensor_parameter_61: name = "parameter_61" + original_name = "conv2d_74.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.190331") @@ -638,6 +700,7 @@ class Program_weight_tensor_parameter_61: class Program_weight_tensor_parameter_62: name = "parameter_62" + original_name = "batch_norm2d_67.b_0" shape = [96] dtype = "float32" min_val = float("-1.42296") @@ -649,6 +712,7 @@ class Program_weight_tensor_parameter_62: class Program_weight_tensor_parameter_63: name = "parameter_63" + original_name = "batch_norm2d_67.w_0" shape = [96] dtype = "float32" min_val = float("0.386897") @@ -660,6 +724,7 @@ class Program_weight_tensor_parameter_63: class Program_weight_tensor_parameter_64: name = "parameter_64" + original_name = "batch_norm2d_67.w_2" shape = [96] dtype = "float32" min_val = float("0.051474") @@ -671,6 +736,7 @@ class Program_weight_tensor_parameter_64: class Program_weight_tensor_parameter_65: name = "parameter_65" + original_name = "batch_norm2d_67.w_1" shape = [96] dtype = "float32" min_val = float("-0.954577") @@ -682,6 +748,7 @@ class Program_weight_tensor_parameter_65: class Program_weight_tensor_parameter_66: name = "parameter_66" + original_name = "conv2d_73.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.47819") @@ -693,6 +760,7 @@ class Program_weight_tensor_parameter_66: class Program_weight_tensor_parameter_67: name = "parameter_67" + original_name = "batch_norm2d_66.b_0" shape = [96] dtype = "float32" min_val = float("-1.82671") @@ -704,6 +772,7 @@ class Program_weight_tensor_parameter_67: class Program_weight_tensor_parameter_68: name = "parameter_68" + original_name = "batch_norm2d_66.w_0" shape = [96] dtype = "float32" min_val = float("0.471409") @@ -715,6 +784,7 @@ class Program_weight_tensor_parameter_68: class Program_weight_tensor_parameter_69: name = "parameter_69" + original_name = "batch_norm2d_66.w_2" shape = [96] dtype = "float32" min_val = float("0.000139263") @@ -726,6 +796,7 @@ class Program_weight_tensor_parameter_69: class Program_weight_tensor_parameter_70: name = "parameter_70" + original_name = "batch_norm2d_66.w_1" shape = [96] dtype = "float32" min_val = float("-0.41986") @@ -737,6 +808,7 @@ class Program_weight_tensor_parameter_70: class Program_weight_tensor_parameter_71: name = "parameter_71" + original_name = "conv2d_72.w_0" shape = [96, 1, 5, 5] dtype = "float32" min_val = float("-0.353687") @@ -748,6 +820,7 @@ class Program_weight_tensor_parameter_71: class Program_weight_tensor_parameter_72: name = "parameter_72" + original_name = "batch_norm2d_65.b_0" shape = [96] dtype = "float32" min_val = float("-1.28068") @@ -759,6 +832,7 @@ class Program_weight_tensor_parameter_72: class Program_weight_tensor_parameter_73: name = "parameter_73" + original_name = "batch_norm2d_65.w_0" shape = [96] dtype = "float32" min_val = float("0.0654328") @@ -770,6 +844,7 @@ class Program_weight_tensor_parameter_73: class Program_weight_tensor_parameter_74: name = "parameter_74" + original_name = "batch_norm2d_65.w_2" shape = [96] dtype = "float32" min_val = float("0.0304701") @@ -781,6 +856,7 @@ class Program_weight_tensor_parameter_74: class Program_weight_tensor_parameter_75: name = "parameter_75" + original_name = "batch_norm2d_65.w_1" shape = [96] dtype = "float32" min_val = float("-0.773945") @@ -792,6 +868,7 @@ class Program_weight_tensor_parameter_75: class Program_weight_tensor_parameter_76: name = "parameter_76" + original_name = "conv2d_71.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.428556") @@ -803,6 +880,7 @@ class Program_weight_tensor_parameter_76: class Program_weight_tensor_parameter_77: name = "parameter_77" + original_name = "batch_norm2d_64.b_0" shape = [96] dtype = "float32" min_val = float("-1.10008") @@ -814,6 +892,7 @@ class Program_weight_tensor_parameter_77: class Program_weight_tensor_parameter_78: name = "parameter_78" + original_name = "batch_norm2d_64.w_0" shape = [96] dtype = "float32" min_val = float("0.4353") @@ -825,6 +904,7 @@ class Program_weight_tensor_parameter_78: class Program_weight_tensor_parameter_79: name = "parameter_79" + original_name = "batch_norm2d_64.w_2" shape = [96] dtype = "float32" min_val = float("9.29295e-05") @@ -836,6 +916,7 @@ class Program_weight_tensor_parameter_79: class Program_weight_tensor_parameter_80: name = "parameter_80" + original_name = "batch_norm2d_64.w_1" shape = [96] dtype = "float32" min_val = float("-0.612495") @@ -847,6 +928,7 @@ class Program_weight_tensor_parameter_80: class Program_weight_tensor_parameter_81: name = "parameter_81" + original_name = "conv2d_70.w_0" shape = [96, 1, 5, 5] dtype = "float32" min_val = float("-0.297489") @@ -858,6 +940,7 @@ class Program_weight_tensor_parameter_81: class Program_weight_tensor_parameter_82: name = "parameter_82" + original_name = "batch_norm2d_77.b_0" shape = [1] dtype = "float32" min_val = float("0") @@ -867,6 +950,7 @@ class Program_weight_tensor_parameter_82: class Program_weight_tensor_parameter_83: name = "parameter_83" + original_name = "batch_norm2d_77.w_0" shape = [1] dtype = "float32" min_val = float("0") @@ -876,6 +960,7 @@ class Program_weight_tensor_parameter_83: class Program_weight_tensor_parameter_84: name = "parameter_84" + original_name = "batch_norm2d_77.w_2" shape = [1] dtype = "float32" min_val = float("0") @@ -885,6 +970,7 @@ class Program_weight_tensor_parameter_84: class Program_weight_tensor_parameter_85: name = "parameter_85" + original_name = "batch_norm2d_77.w_1" shape = [1] dtype = "float32" min_val = float("0") @@ -894,6 +980,7 @@ class Program_weight_tensor_parameter_85: class Program_weight_tensor_parameter_86: name = "parameter_86" + original_name = "conv2d_91.w_0" shape = [1, 1, 1, 1] dtype = "float32" min_val = float("0") @@ -903,6 +990,7 @@ class Program_weight_tensor_parameter_86: class Program_weight_tensor_parameter_87: name = "parameter_87" + original_name = "batch_norm2d_76.b_0" shape = [1] dtype = "float32" min_val = float("0") @@ -912,6 +1000,7 @@ class Program_weight_tensor_parameter_87: class Program_weight_tensor_parameter_88: name = "parameter_88" + original_name = "batch_norm2d_76.w_0" shape = [1] dtype = "float32" min_val = float("0") @@ -921,6 +1010,7 @@ class Program_weight_tensor_parameter_88: class Program_weight_tensor_parameter_89: name = "parameter_89" + original_name = "batch_norm2d_76.w_2" shape = [1] dtype = "float32" min_val = float("0") @@ -930,6 +1020,7 @@ class Program_weight_tensor_parameter_89: class Program_weight_tensor_parameter_90: name = "parameter_90" + original_name = "batch_norm2d_76.w_1" shape = [1] dtype = "float32" min_val = float("0") @@ -939,6 +1030,7 @@ class Program_weight_tensor_parameter_90: class Program_weight_tensor_parameter_91: name = "parameter_91" + original_name = "conv2d_90.w_0" shape = [1, 96, 5, 5] dtype = "float32" min_val = float("-0.889754") @@ -950,6 +1042,7 @@ class Program_weight_tensor_parameter_91: class Program_weight_tensor_parameter_92: name = "parameter_92" + original_name = "conv2d_89.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -959,6 +1052,7 @@ class Program_weight_tensor_parameter_92: class Program_weight_tensor_parameter_93: name = "parameter_93" + original_name = "conv2d_89.w_0" shape = [32, 96, 1, 1] dtype = "float32" min_val = float("-0.581373") @@ -970,6 +1064,7 @@ class Program_weight_tensor_parameter_93: class Program_weight_tensor_parameter_94: name = "parameter_94" + original_name = "conv2d_88.b_0" shape = [4] dtype = "float32" min_val = float("0") @@ -979,6 +1074,7 @@ class Program_weight_tensor_parameter_94: class Program_weight_tensor_parameter_95: name = "parameter_95" + original_name = "conv2d_88.w_0" shape = [4, 96, 1, 1] dtype = "float32" min_val = float("-0.0255968") @@ -990,6 +1086,7 @@ class Program_weight_tensor_parameter_95: class Program_weight_tensor_parameter_96: name = "parameter_96" + original_name = "batch_norm2d_63.b_0" shape = [96] dtype = "float32" min_val = float("-0.94137") @@ -1001,6 +1098,7 @@ class Program_weight_tensor_parameter_96: class Program_weight_tensor_parameter_97: name = "parameter_97" + original_name = "batch_norm2d_63.w_0" shape = [96] dtype = "float32" min_val = float("0.366447") @@ -1012,6 +1110,7 @@ class Program_weight_tensor_parameter_97: class Program_weight_tensor_parameter_98: name = "parameter_98" + original_name = "batch_norm2d_63.w_2" shape = [96] dtype = "float32" min_val = float("0.0287144") @@ -1023,6 +1122,7 @@ class Program_weight_tensor_parameter_98: class Program_weight_tensor_parameter_99: name = "parameter_99" + original_name = "batch_norm2d_63.w_1" shape = [96] dtype = "float32" min_val = float("-0.615769") @@ -1034,6 +1134,7 @@ class Program_weight_tensor_parameter_99: class Program_weight_tensor_parameter_100: name = "parameter_100" + original_name = "conv2d_69.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.560049") @@ -1045,6 +1146,7 @@ class Program_weight_tensor_parameter_100: class Program_weight_tensor_parameter_101: name = "parameter_101" + original_name = "conv2d_68.b_0" shape = [96] dtype = "float32" min_val = float("-0.039443") @@ -1056,6 +1158,7 @@ class Program_weight_tensor_parameter_101: class Program_weight_tensor_parameter_102: name = "parameter_102" + original_name = "conv2d_68.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.297396") @@ -1067,6 +1170,7 @@ class Program_weight_tensor_parameter_102: class Program_weight_tensor_parameter_103: name = "parameter_103" + original_name = "batch_norm2d_62.b_0" shape = [96] dtype = "float32" min_val = float("-2.18491") @@ -1078,6 +1182,7 @@ class Program_weight_tensor_parameter_103: class Program_weight_tensor_parameter_104: name = "parameter_104" + original_name = "batch_norm2d_62.w_0" shape = [96] dtype = "float32" min_val = float("-0.783288") @@ -1089,6 +1194,7 @@ class Program_weight_tensor_parameter_104: class Program_weight_tensor_parameter_105: name = "parameter_105" + original_name = "batch_norm2d_62.w_2" shape = [96] dtype = "float32" min_val = float("0.0704238") @@ -1100,6 +1206,7 @@ class Program_weight_tensor_parameter_105: class Program_weight_tensor_parameter_106: name = "parameter_106" + original_name = "batch_norm2d_62.w_1" shape = [96] dtype = "float32" min_val = float("-1.50569") @@ -1111,6 +1218,7 @@ class Program_weight_tensor_parameter_106: class Program_weight_tensor_parameter_107: name = "parameter_107" + original_name = "conv2d_67.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.519273") @@ -1122,6 +1230,7 @@ class Program_weight_tensor_parameter_107: class Program_weight_tensor_parameter_108: name = "parameter_108" + original_name = "batch_norm2d_61.b_0" shape = [96] dtype = "float32" min_val = float("-1.05647") @@ -1133,6 +1242,7 @@ class Program_weight_tensor_parameter_108: class Program_weight_tensor_parameter_109: name = "parameter_109" + original_name = "batch_norm2d_61.w_0" shape = [96] dtype = "float32" min_val = float("0.431147") @@ -1144,6 +1254,7 @@ class Program_weight_tensor_parameter_109: class Program_weight_tensor_parameter_110: name = "parameter_110" + original_name = "batch_norm2d_61.w_2" shape = [96] dtype = "float32" min_val = float("6.94141e-05") @@ -1155,6 +1266,7 @@ class Program_weight_tensor_parameter_110: class Program_weight_tensor_parameter_111: name = "parameter_111" + original_name = "batch_norm2d_61.w_1" shape = [96] dtype = "float32" min_val = float("-0.382427") @@ -1166,6 +1278,7 @@ class Program_weight_tensor_parameter_111: class Program_weight_tensor_parameter_112: name = "parameter_112" + original_name = "conv2d_66.w_0" shape = [96, 1, 5, 5] dtype = "float32" min_val = float("-0.494021") @@ -1177,6 +1290,7 @@ class Program_weight_tensor_parameter_112: class Program_weight_tensor_parameter_113: name = "parameter_113" + original_name = "batch_norm2d_60.b_0" shape = [96] dtype = "float32" min_val = float("-0.918264") @@ -1188,6 +1302,7 @@ class Program_weight_tensor_parameter_113: class Program_weight_tensor_parameter_114: name = "parameter_114" + original_name = "batch_norm2d_60.w_0" shape = [96] dtype = "float32" min_val = float("0.0373496") @@ -1199,6 +1314,7 @@ class Program_weight_tensor_parameter_114: class Program_weight_tensor_parameter_115: name = "parameter_115" + original_name = "batch_norm2d_60.w_2" shape = [96] dtype = "float32" min_val = float("0.0524067") @@ -1210,6 +1326,7 @@ class Program_weight_tensor_parameter_115: class Program_weight_tensor_parameter_116: name = "parameter_116" + original_name = "batch_norm2d_60.w_1" shape = [96] dtype = "float32" min_val = float("-1.11667") @@ -1221,6 +1338,7 @@ class Program_weight_tensor_parameter_116: class Program_weight_tensor_parameter_117: name = "parameter_117" + original_name = "conv2d_65.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.581063") @@ -1232,6 +1350,7 @@ class Program_weight_tensor_parameter_117: class Program_weight_tensor_parameter_118: name = "parameter_118" + original_name = "batch_norm2d_59.b_0" shape = [96] dtype = "float32" min_val = float("-0.738662") @@ -1243,6 +1362,7 @@ class Program_weight_tensor_parameter_118: class Program_weight_tensor_parameter_119: name = "parameter_119" + original_name = "batch_norm2d_59.w_0" shape = [96] dtype = "float32" min_val = float("0.452267") @@ -1254,6 +1374,7 @@ class Program_weight_tensor_parameter_119: class Program_weight_tensor_parameter_120: name = "parameter_120" + original_name = "batch_norm2d_59.w_2" shape = [96] dtype = "float32" min_val = float("5.93424e-05") @@ -1265,6 +1386,7 @@ class Program_weight_tensor_parameter_120: class Program_weight_tensor_parameter_121: name = "parameter_121" + original_name = "batch_norm2d_59.w_1" shape = [96] dtype = "float32" min_val = float("-0.956372") @@ -1276,6 +1398,7 @@ class Program_weight_tensor_parameter_121: class Program_weight_tensor_parameter_122: name = "parameter_122" + original_name = "conv2d_64.w_0" shape = [96, 1, 5, 5] dtype = "float32" min_val = float("-0.387297") @@ -1287,6 +1410,7 @@ class Program_weight_tensor_parameter_122: class Program_weight_tensor_parameter_123: name = "parameter_123" + original_name = "batch_norm2d_75.b_0" shape = [1] dtype = "float32" min_val = float("0") @@ -1296,6 +1420,7 @@ class Program_weight_tensor_parameter_123: class Program_weight_tensor_parameter_124: name = "parameter_124" + original_name = "batch_norm2d_75.w_0" shape = [1] dtype = "float32" min_val = float("0") @@ -1305,6 +1430,7 @@ class Program_weight_tensor_parameter_124: class Program_weight_tensor_parameter_125: name = "parameter_125" + original_name = "batch_norm2d_75.w_2" shape = [1] dtype = "float32" min_val = float("0") @@ -1314,6 +1440,7 @@ class Program_weight_tensor_parameter_125: class Program_weight_tensor_parameter_126: name = "parameter_126" + original_name = "batch_norm2d_75.w_1" shape = [1] dtype = "float32" min_val = float("0") @@ -1323,6 +1450,7 @@ class Program_weight_tensor_parameter_126: class Program_weight_tensor_parameter_127: name = "parameter_127" + original_name = "conv2d_87.w_0" shape = [1, 1, 1, 1] dtype = "float32" min_val = float("0") @@ -1332,6 +1460,7 @@ class Program_weight_tensor_parameter_127: class Program_weight_tensor_parameter_128: name = "parameter_128" + original_name = "batch_norm2d_74.b_0" shape = [1] dtype = "float32" min_val = float("0") @@ -1341,6 +1470,7 @@ class Program_weight_tensor_parameter_128: class Program_weight_tensor_parameter_129: name = "parameter_129" + original_name = "batch_norm2d_74.w_0" shape = [1] dtype = "float32" min_val = float("0") @@ -1350,6 +1480,7 @@ class Program_weight_tensor_parameter_129: class Program_weight_tensor_parameter_130: name = "parameter_130" + original_name = "batch_norm2d_74.w_2" shape = [1] dtype = "float32" min_val = float("0") @@ -1359,6 +1490,7 @@ class Program_weight_tensor_parameter_130: class Program_weight_tensor_parameter_131: name = "parameter_131" + original_name = "batch_norm2d_74.w_1" shape = [1] dtype = "float32" min_val = float("0") @@ -1368,6 +1500,7 @@ class Program_weight_tensor_parameter_131: class Program_weight_tensor_parameter_132: name = "parameter_132" + original_name = "conv2d_86.w_0" shape = [1, 96, 5, 5] dtype = "float32" min_val = float("-0.541901") @@ -1379,6 +1512,7 @@ class Program_weight_tensor_parameter_132: class Program_weight_tensor_parameter_133: name = "parameter_133" + original_name = "conv2d_85.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -1388,6 +1522,7 @@ class Program_weight_tensor_parameter_133: class Program_weight_tensor_parameter_134: name = "parameter_134" + original_name = "conv2d_85.w_0" shape = [32, 96, 1, 1] dtype = "float32" min_val = float("-0.603402") @@ -1399,6 +1534,7 @@ class Program_weight_tensor_parameter_134: class Program_weight_tensor_parameter_135: name = "parameter_135" + original_name = "conv2d_84.b_0" shape = [4] dtype = "float32" min_val = float("0") @@ -1408,6 +1544,7 @@ class Program_weight_tensor_parameter_135: class Program_weight_tensor_parameter_136: name = "parameter_136" + original_name = "conv2d_84.w_0" shape = [4, 96, 1, 1] dtype = "float32" min_val = float("-0.0282628") @@ -1419,6 +1556,7 @@ class Program_weight_tensor_parameter_136: class Program_weight_tensor_parameter_137: name = "parameter_137" + original_name = "batch_norm2d_58.b_0" shape = [96] dtype = "float32" min_val = float("-0.517474") @@ -1430,6 +1568,7 @@ class Program_weight_tensor_parameter_137: class Program_weight_tensor_parameter_138: name = "parameter_138" + original_name = "batch_norm2d_58.w_0" shape = [96] dtype = "float32" min_val = float("0.626348") @@ -1441,6 +1580,7 @@ class Program_weight_tensor_parameter_138: class Program_weight_tensor_parameter_139: name = "parameter_139" + original_name = "batch_norm2d_58.w_2" shape = [96] dtype = "float32" min_val = float("0.024957") @@ -1452,6 +1592,7 @@ class Program_weight_tensor_parameter_139: class Program_weight_tensor_parameter_140: name = "parameter_140" + original_name = "batch_norm2d_58.w_1" shape = [96] dtype = "float32" min_val = float("-0.894422") @@ -1463,6 +1604,7 @@ class Program_weight_tensor_parameter_140: class Program_weight_tensor_parameter_141: name = "parameter_141" + original_name = "conv2d_63.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.777963") @@ -1474,6 +1616,7 @@ class Program_weight_tensor_parameter_141: class Program_weight_tensor_parameter_142: name = "parameter_142" + original_name = "conv2d_62.b_0" shape = [96] dtype = "float32" min_val = float("-0.0296252") @@ -1485,6 +1628,7 @@ class Program_weight_tensor_parameter_142: class Program_weight_tensor_parameter_143: name = "parameter_143" + original_name = "conv2d_62.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.493271") @@ -1496,6 +1640,7 @@ class Program_weight_tensor_parameter_143: class Program_weight_tensor_parameter_144: name = "parameter_144" + original_name = "batch_norm2d_57.b_0" shape = [96] dtype = "float32" min_val = float("-1.82311") @@ -1507,6 +1652,7 @@ class Program_weight_tensor_parameter_144: class Program_weight_tensor_parameter_145: name = "parameter_145" + original_name = "batch_norm2d_57.w_0" shape = [96] dtype = "float32" min_val = float("0.360213") @@ -1518,6 +1664,7 @@ class Program_weight_tensor_parameter_145: class Program_weight_tensor_parameter_146: name = "parameter_146" + original_name = "batch_norm2d_57.w_2" shape = [96] dtype = "float32" min_val = float("0.0472455") @@ -1529,6 +1676,7 @@ class Program_weight_tensor_parameter_146: class Program_weight_tensor_parameter_147: name = "parameter_147" + original_name = "batch_norm2d_57.w_1" shape = [96] dtype = "float32" min_val = float("-2.16312") @@ -1540,6 +1688,7 @@ class Program_weight_tensor_parameter_147: class Program_weight_tensor_parameter_148: name = "parameter_148" + original_name = "conv2d_61.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.553904") @@ -1551,6 +1700,7 @@ class Program_weight_tensor_parameter_148: class Program_weight_tensor_parameter_149: name = "parameter_149" + original_name = "batch_norm2d_56.b_0" shape = [96] dtype = "float32" min_val = float("-0.939243") @@ -1562,6 +1712,7 @@ class Program_weight_tensor_parameter_149: class Program_weight_tensor_parameter_150: name = "parameter_150" + original_name = "batch_norm2d_56.w_0" shape = [96] dtype = "float32" min_val = float("0.266135") @@ -1573,6 +1724,7 @@ class Program_weight_tensor_parameter_150: class Program_weight_tensor_parameter_151: name = "parameter_151" + original_name = "batch_norm2d_56.w_2" shape = [96] dtype = "float32" min_val = float("0.000136714") @@ -1584,6 +1736,7 @@ class Program_weight_tensor_parameter_151: class Program_weight_tensor_parameter_152: name = "parameter_152" + original_name = "batch_norm2d_56.w_1" shape = [96] dtype = "float32" min_val = float("-0.788481") @@ -1595,6 +1748,7 @@ class Program_weight_tensor_parameter_152: class Program_weight_tensor_parameter_153: name = "parameter_153" + original_name = "conv2d_60.w_0" shape = [96, 1, 5, 5] dtype = "float32" min_val = float("-0.562266") @@ -1606,6 +1760,7 @@ class Program_weight_tensor_parameter_153: class Program_weight_tensor_parameter_154: name = "parameter_154" + original_name = "batch_norm2d_55.b_0" shape = [96] dtype = "float32" min_val = float("-1.46556") @@ -1617,6 +1772,7 @@ class Program_weight_tensor_parameter_154: class Program_weight_tensor_parameter_155: name = "parameter_155" + original_name = "batch_norm2d_55.w_0" shape = [96] dtype = "float32" min_val = float("0.0418698") @@ -1628,6 +1784,7 @@ class Program_weight_tensor_parameter_155: class Program_weight_tensor_parameter_156: name = "parameter_156" + original_name = "batch_norm2d_55.w_2" shape = [96] dtype = "float32" min_val = float("0.0644678") @@ -1639,6 +1796,7 @@ class Program_weight_tensor_parameter_156: class Program_weight_tensor_parameter_157: name = "parameter_157" + original_name = "batch_norm2d_55.w_1" shape = [96] dtype = "float32" min_val = float("-1.5239") @@ -1650,6 +1808,7 @@ class Program_weight_tensor_parameter_157: class Program_weight_tensor_parameter_158: name = "parameter_158" + original_name = "conv2d_59.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.554408") @@ -1661,6 +1820,7 @@ class Program_weight_tensor_parameter_158: class Program_weight_tensor_parameter_159: name = "parameter_159" + original_name = "batch_norm2d_54.b_0" shape = [96] dtype = "float32" min_val = float("-1.21138") @@ -1672,6 +1832,7 @@ class Program_weight_tensor_parameter_159: class Program_weight_tensor_parameter_160: name = "parameter_160" + original_name = "batch_norm2d_54.w_0" shape = [96] dtype = "float32" min_val = float("0.391629") @@ -1683,6 +1844,7 @@ class Program_weight_tensor_parameter_160: class Program_weight_tensor_parameter_161: name = "parameter_161" + original_name = "batch_norm2d_54.w_2" shape = [96] dtype = "float32" min_val = float("0.000114896") @@ -1694,6 +1856,7 @@ class Program_weight_tensor_parameter_161: class Program_weight_tensor_parameter_162: name = "parameter_162" + original_name = "batch_norm2d_54.w_1" shape = [96] dtype = "float32" min_val = float("-0.643511") @@ -1705,6 +1868,7 @@ class Program_weight_tensor_parameter_162: class Program_weight_tensor_parameter_163: name = "parameter_163" + original_name = "conv2d_58.w_0" shape = [96, 1, 5, 5] dtype = "float32" min_val = float("-0.432745") @@ -1716,6 +1880,7 @@ class Program_weight_tensor_parameter_163: class Program_weight_tensor_parameter_164: name = "parameter_164" + original_name = "batch_norm2d_33.b_0" shape = [96] dtype = "float32" min_val = float("-0.0700948") @@ -1727,6 +1892,7 @@ class Program_weight_tensor_parameter_164: class Program_weight_tensor_parameter_165: name = "parameter_165" + original_name = "batch_norm2d_33.w_0" shape = [96] dtype = "float32" min_val = float("0.0293734") @@ -1738,6 +1904,7 @@ class Program_weight_tensor_parameter_165: class Program_weight_tensor_parameter_166: name = "parameter_166" + original_name = "batch_norm2d_33.w_2" shape = [96] dtype = "float32" min_val = float("0.000103696") @@ -1749,6 +1916,7 @@ class Program_weight_tensor_parameter_166: class Program_weight_tensor_parameter_167: name = "parameter_167" + original_name = "batch_norm2d_33.w_1" shape = [96] dtype = "float32" min_val = float("-0.00417754") @@ -1760,6 +1928,7 @@ class Program_weight_tensor_parameter_167: class Program_weight_tensor_parameter_168: name = "parameter_168" + original_name = "conv2d_37.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.146401") @@ -1771,6 +1940,7 @@ class Program_weight_tensor_parameter_168: class Program_weight_tensor_parameter_169: name = "parameter_169" + original_name = "batch_norm2d_32.b_0" shape = [96] dtype = "float32" min_val = float("-0.0372991") @@ -1782,6 +1952,7 @@ class Program_weight_tensor_parameter_169: class Program_weight_tensor_parameter_170: name = "parameter_170" + original_name = "batch_norm2d_32.w_0" shape = [96] dtype = "float32" min_val = float("0.0611877") @@ -1793,6 +1964,7 @@ class Program_weight_tensor_parameter_170: class Program_weight_tensor_parameter_171: name = "parameter_171" + original_name = "batch_norm2d_32.w_2" shape = [96] dtype = "float32" min_val = float("7.53554e-05") @@ -1804,6 +1976,7 @@ class Program_weight_tensor_parameter_171: class Program_weight_tensor_parameter_172: name = "parameter_172" + original_name = "batch_norm2d_32.w_1" shape = [96] dtype = "float32" min_val = float("-0.23033") @@ -1815,6 +1988,7 @@ class Program_weight_tensor_parameter_172: class Program_weight_tensor_parameter_173: name = "parameter_173" + original_name = "conv2d_36.w_0" shape = [96, 1, 5, 5] dtype = "float32" min_val = float("-0.127158") @@ -1826,6 +2000,7 @@ class Program_weight_tensor_parameter_173: class Program_weight_tensor_parameter_174: name = "parameter_174" + original_name = "batch_norm2d_31.b_0" shape = [96] dtype = "float32" min_val = float("-0.103094") @@ -1837,6 +2012,7 @@ class Program_weight_tensor_parameter_174: class Program_weight_tensor_parameter_175: name = "parameter_175" + original_name = "batch_norm2d_31.w_0" shape = [96] dtype = "float32" min_val = float("0.00959989") @@ -1848,6 +2024,7 @@ class Program_weight_tensor_parameter_175: class Program_weight_tensor_parameter_176: name = "parameter_176" + original_name = "batch_norm2d_31.w_2" shape = [96] dtype = "float32" min_val = float("2.15479e-05") @@ -1859,6 +2036,7 @@ class Program_weight_tensor_parameter_176: class Program_weight_tensor_parameter_177: name = "parameter_177" + original_name = "batch_norm2d_31.w_1" shape = [96] dtype = "float32" min_val = float("-0.00382952") @@ -1870,6 +2048,7 @@ class Program_weight_tensor_parameter_177: class Program_weight_tensor_parameter_178: name = "parameter_178" + original_name = "conv2d_35.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.20591") @@ -1881,6 +2060,7 @@ class Program_weight_tensor_parameter_178: class Program_weight_tensor_parameter_179: name = "parameter_179" + original_name = "batch_norm2d_30.b_0" shape = [96] dtype = "float32" min_val = float("-0.0245617") @@ -1892,6 +2072,7 @@ class Program_weight_tensor_parameter_179: class Program_weight_tensor_parameter_180: name = "parameter_180" + original_name = "batch_norm2d_30.w_0" shape = [96] dtype = "float32" min_val = float("0.0737232") @@ -1903,6 +2084,7 @@ class Program_weight_tensor_parameter_180: class Program_weight_tensor_parameter_181: name = "parameter_181" + original_name = "batch_norm2d_30.w_2" shape = [96] dtype = "float32" min_val = float("2.61084e-05") @@ -1914,6 +2096,7 @@ class Program_weight_tensor_parameter_181: class Program_weight_tensor_parameter_182: name = "parameter_182" + original_name = "batch_norm2d_30.w_1" shape = [96] dtype = "float32" min_val = float("-0.0173724") @@ -1925,6 +2108,7 @@ class Program_weight_tensor_parameter_182: class Program_weight_tensor_parameter_183: name = "parameter_183" + original_name = "conv2d_34.w_0" shape = [96, 1, 5, 5] dtype = "float32" min_val = float("-0.0847993") @@ -1936,6 +2120,7 @@ class Program_weight_tensor_parameter_183: class Program_weight_tensor_parameter_184: name = "parameter_184" + original_name = "batch_norm2d_53.b_0" shape = [96] dtype = "float32" min_val = float("-0.825361") @@ -1947,6 +2132,7 @@ class Program_weight_tensor_parameter_184: class Program_weight_tensor_parameter_185: name = "parameter_185" + original_name = "batch_norm2d_53.w_0" shape = [96] dtype = "float32" min_val = float("0.0648042") @@ -1958,6 +2144,7 @@ class Program_weight_tensor_parameter_185: class Program_weight_tensor_parameter_186: name = "parameter_186" + original_name = "batch_norm2d_53.w_2" shape = [96] dtype = "float32" min_val = float("0.07505") @@ -1969,6 +2156,7 @@ class Program_weight_tensor_parameter_186: class Program_weight_tensor_parameter_187: name = "parameter_187" + original_name = "batch_norm2d_53.w_1" shape = [96] dtype = "float32" min_val = float("-0.571573") @@ -1980,6 +2168,7 @@ class Program_weight_tensor_parameter_187: class Program_weight_tensor_parameter_188: name = "parameter_188" + original_name = "conv2d_57.w_0" shape = [96, 192, 1, 1] dtype = "float32" min_val = float("-0.37051") @@ -1991,6 +2180,7 @@ class Program_weight_tensor_parameter_188: class Program_weight_tensor_parameter_189: name = "parameter_189" + original_name = "batch_norm2d_52.b_0" shape = [192] dtype = "float32" min_val = float("-0.617106") @@ -2002,6 +2192,7 @@ class Program_weight_tensor_parameter_189: class Program_weight_tensor_parameter_190: name = "parameter_190" + original_name = "batch_norm2d_52.w_0" shape = [192] dtype = "float32" min_val = float("0.584145") @@ -2013,6 +2204,7 @@ class Program_weight_tensor_parameter_190: class Program_weight_tensor_parameter_191: name = "parameter_191" + original_name = "batch_norm2d_52.w_2" shape = [192] dtype = "float32" min_val = float("0.00043962") @@ -2024,6 +2216,7 @@ class Program_weight_tensor_parameter_191: class Program_weight_tensor_parameter_192: name = "parameter_192" + original_name = "batch_norm2d_52.w_1" shape = [192] dtype = "float32" min_val = float("-0.270378") @@ -2035,6 +2228,7 @@ class Program_weight_tensor_parameter_192: class Program_weight_tensor_parameter_193: name = "parameter_193" + original_name = "conv2d_56.w_0" shape = [192, 1, 5, 5] dtype = "float32" min_val = float("-0.217718") @@ -2046,6 +2240,7 @@ class Program_weight_tensor_parameter_193: class Program_weight_tensor_parameter_194: name = "parameter_194" + original_name = "batch_norm2d_51.b_0" shape = [192] dtype = "float32" min_val = float("-1.1353") @@ -2057,6 +2252,7 @@ class Program_weight_tensor_parameter_194: class Program_weight_tensor_parameter_195: name = "parameter_195" + original_name = "batch_norm2d_51.w_0" shape = [192] dtype = "float32" min_val = float("0.109504") @@ -2068,6 +2264,7 @@ class Program_weight_tensor_parameter_195: class Program_weight_tensor_parameter_196: name = "parameter_196" + original_name = "batch_norm2d_51.w_2" shape = [192] dtype = "float32" min_val = float("0.0703491") @@ -2079,6 +2276,7 @@ class Program_weight_tensor_parameter_196: class Program_weight_tensor_parameter_197: name = "parameter_197" + original_name = "batch_norm2d_51.w_1" shape = [192] dtype = "float32" min_val = float("-0.749884") @@ -2090,6 +2288,7 @@ class Program_weight_tensor_parameter_197: class Program_weight_tensor_parameter_198: name = "parameter_198" + original_name = "conv2d_55.w_0" shape = [192, 192, 1, 1] dtype = "float32" min_val = float("-0.339782") @@ -2101,6 +2300,7 @@ class Program_weight_tensor_parameter_198: class Program_weight_tensor_parameter_199: name = "parameter_199" + original_name = "batch_norm2d_50.b_0" shape = [192] dtype = "float32" min_val = float("-0.799302") @@ -2112,6 +2312,7 @@ class Program_weight_tensor_parameter_199: class Program_weight_tensor_parameter_200: name = "parameter_200" + original_name = "batch_norm2d_50.w_0" shape = [192] dtype = "float32" min_val = float("0.576676") @@ -2123,6 +2324,7 @@ class Program_weight_tensor_parameter_200: class Program_weight_tensor_parameter_201: name = "parameter_201" + original_name = "batch_norm2d_50.w_2" shape = [192] dtype = "float32" min_val = float("4.29064e-05") @@ -2134,6 +2336,7 @@ class Program_weight_tensor_parameter_201: class Program_weight_tensor_parameter_202: name = "parameter_202" + original_name = "batch_norm2d_50.w_1" shape = [192] dtype = "float32" min_val = float("-0.138592") @@ -2145,6 +2348,7 @@ class Program_weight_tensor_parameter_202: class Program_weight_tensor_parameter_203: name = "parameter_203" + original_name = "conv2d_54.w_0" shape = [192, 1, 5, 5] dtype = "float32" min_val = float("-0.333901") @@ -2156,6 +2360,7 @@ class Program_weight_tensor_parameter_203: class Program_weight_tensor_parameter_204: name = "parameter_204" + original_name = "batch_norm2d_49.b_0" shape = [96] dtype = "float32" min_val = float("-0.271221") @@ -2167,6 +2372,7 @@ class Program_weight_tensor_parameter_204: class Program_weight_tensor_parameter_205: name = "parameter_205" + original_name = "batch_norm2d_49.w_0" shape = [96] dtype = "float32" min_val = float("0.0421079") @@ -2178,6 +2384,7 @@ class Program_weight_tensor_parameter_205: class Program_weight_tensor_parameter_206: name = "parameter_206" + original_name = "batch_norm2d_49.w_2" shape = [96] dtype = "float32" min_val = float("0.000647954") @@ -2189,6 +2396,7 @@ class Program_weight_tensor_parameter_206: class Program_weight_tensor_parameter_207: name = "parameter_207" + original_name = "batch_norm2d_49.w_1" shape = [96] dtype = "float32" min_val = float("-0.0215934") @@ -2200,6 +2408,7 @@ class Program_weight_tensor_parameter_207: class Program_weight_tensor_parameter_208: name = "parameter_208" + original_name = "conv2d_53.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.333797") @@ -2211,6 +2420,7 @@ class Program_weight_tensor_parameter_208: class Program_weight_tensor_parameter_209: name = "parameter_209" + original_name = "batch_norm2d_48.b_0" shape = [96] dtype = "float32" min_val = float("-0.29402") @@ -2222,6 +2432,7 @@ class Program_weight_tensor_parameter_209: class Program_weight_tensor_parameter_210: name = "parameter_210" + original_name = "batch_norm2d_48.w_0" shape = [96] dtype = "float32" min_val = float("0.0711968") @@ -2233,6 +2444,7 @@ class Program_weight_tensor_parameter_210: class Program_weight_tensor_parameter_211: name = "parameter_211" + original_name = "batch_norm2d_48.w_2" shape = [96] dtype = "float32" min_val = float("2.90939e-05") @@ -2244,6 +2456,7 @@ class Program_weight_tensor_parameter_211: class Program_weight_tensor_parameter_212: name = "parameter_212" + original_name = "batch_norm2d_48.w_1" shape = [96] dtype = "float32" min_val = float("-0.431962") @@ -2255,6 +2468,7 @@ class Program_weight_tensor_parameter_212: class Program_weight_tensor_parameter_213: name = "parameter_213" + original_name = "conv2d_52.w_0" shape = [96, 1, 5, 5] dtype = "float32" min_val = float("-0.22857") @@ -2266,6 +2480,7 @@ class Program_weight_tensor_parameter_213: class Program_weight_tensor_parameter_214: name = "parameter_214" + original_name = "batch_norm2d_47.b_0" shape = [96] dtype = "float32" min_val = float("-1.37667") @@ -2277,6 +2492,7 @@ class Program_weight_tensor_parameter_214: class Program_weight_tensor_parameter_215: name = "parameter_215" + original_name = "batch_norm2d_47.w_0" shape = [96] dtype = "float32" min_val = float("0.0395503") @@ -2288,6 +2504,7 @@ class Program_weight_tensor_parameter_215: class Program_weight_tensor_parameter_216: name = "parameter_216" + original_name = "batch_norm2d_47.w_2" shape = [96] dtype = "float32" min_val = float("0.136279") @@ -2299,6 +2516,7 @@ class Program_weight_tensor_parameter_216: class Program_weight_tensor_parameter_217: name = "parameter_217" + original_name = "batch_norm2d_47.w_1" shape = [96] dtype = "float32" min_val = float("-0.665498") @@ -2310,6 +2528,7 @@ class Program_weight_tensor_parameter_217: class Program_weight_tensor_parameter_218: name = "parameter_218" + original_name = "conv2d_51.w_0" shape = [96, 192, 1, 1] dtype = "float32" min_val = float("-0.513298") @@ -2321,6 +2540,7 @@ class Program_weight_tensor_parameter_218: class Program_weight_tensor_parameter_219: name = "parameter_219" + original_name = "batch_norm2d_46.b_0" shape = [192] dtype = "float32" min_val = float("-0.951625") @@ -2332,6 +2552,7 @@ class Program_weight_tensor_parameter_219: class Program_weight_tensor_parameter_220: name = "parameter_220" + original_name = "batch_norm2d_46.w_0" shape = [192] dtype = "float32" min_val = float("0.520597") @@ -2343,6 +2564,7 @@ class Program_weight_tensor_parameter_220: class Program_weight_tensor_parameter_221: name = "parameter_221" + original_name = "batch_norm2d_46.w_2" shape = [192] dtype = "float32" min_val = float("0.000116402") @@ -2354,6 +2576,7 @@ class Program_weight_tensor_parameter_221: class Program_weight_tensor_parameter_222: name = "parameter_222" + original_name = "batch_norm2d_46.w_1" shape = [192] dtype = "float32" min_val = float("-0.520661") @@ -2365,6 +2588,7 @@ class Program_weight_tensor_parameter_222: class Program_weight_tensor_parameter_223: name = "parameter_223" + original_name = "conv2d_50.w_0" shape = [192, 1, 5, 5] dtype = "float32" min_val = float("-0.276373") @@ -2376,6 +2600,7 @@ class Program_weight_tensor_parameter_223: class Program_weight_tensor_parameter_224: name = "parameter_224" + original_name = "batch_norm2d_45.b_0" shape = [192] dtype = "float32" min_val = float("-1.23521") @@ -2387,6 +2612,7 @@ class Program_weight_tensor_parameter_224: class Program_weight_tensor_parameter_225: name = "parameter_225" + original_name = "batch_norm2d_45.w_0" shape = [192] dtype = "float32" min_val = float("0.0514489") @@ -2398,6 +2624,7 @@ class Program_weight_tensor_parameter_225: class Program_weight_tensor_parameter_226: name = "parameter_226" + original_name = "batch_norm2d_45.w_2" shape = [192] dtype = "float32" min_val = float("0.0917646") @@ -2409,6 +2636,7 @@ class Program_weight_tensor_parameter_226: class Program_weight_tensor_parameter_227: name = "parameter_227" + original_name = "batch_norm2d_45.w_1" shape = [192] dtype = "float32" min_val = float("-1.0441") @@ -2420,6 +2648,7 @@ class Program_weight_tensor_parameter_227: class Program_weight_tensor_parameter_228: name = "parameter_228" + original_name = "conv2d_49.w_0" shape = [192, 192, 1, 1] dtype = "float32" min_val = float("-0.435505") @@ -2431,6 +2660,7 @@ class Program_weight_tensor_parameter_228: class Program_weight_tensor_parameter_229: name = "parameter_229" + original_name = "batch_norm2d_44.b_0" shape = [192] dtype = "float32" min_val = float("-1.01181") @@ -2442,6 +2672,7 @@ class Program_weight_tensor_parameter_229: class Program_weight_tensor_parameter_230: name = "parameter_230" + original_name = "batch_norm2d_44.w_0" shape = [192] dtype = "float32" min_val = float("0.593966") @@ -2453,6 +2684,7 @@ class Program_weight_tensor_parameter_230: class Program_weight_tensor_parameter_231: name = "parameter_231" + original_name = "batch_norm2d_44.w_2" shape = [192] dtype = "float32" min_val = float("6.03816e-05") @@ -2464,6 +2696,7 @@ class Program_weight_tensor_parameter_231: class Program_weight_tensor_parameter_232: name = "parameter_232" + original_name = "batch_norm2d_44.w_1" shape = [192] dtype = "float32" min_val = float("-0.454907") @@ -2475,6 +2708,7 @@ class Program_weight_tensor_parameter_232: class Program_weight_tensor_parameter_233: name = "parameter_233" + original_name = "conv2d_48.w_0" shape = [192, 1, 5, 5] dtype = "float32" min_val = float("-0.399144") @@ -2486,6 +2720,7 @@ class Program_weight_tensor_parameter_233: class Program_weight_tensor_parameter_234: name = "parameter_234" + original_name = "batch_norm2d_43.b_0" shape = [96] dtype = "float32" min_val = float("-0.185989") @@ -2497,6 +2732,7 @@ class Program_weight_tensor_parameter_234: class Program_weight_tensor_parameter_235: name = "parameter_235" + original_name = "batch_norm2d_43.w_0" shape = [96] dtype = "float32" min_val = float("0.0381173") @@ -2508,6 +2744,7 @@ class Program_weight_tensor_parameter_235: class Program_weight_tensor_parameter_236: name = "parameter_236" + original_name = "batch_norm2d_43.w_2" shape = [96] dtype = "float32" min_val = float("0.00100778") @@ -2519,6 +2756,7 @@ class Program_weight_tensor_parameter_236: class Program_weight_tensor_parameter_237: name = "parameter_237" + original_name = "batch_norm2d_43.w_1" shape = [96] dtype = "float32" min_val = float("-0.0603913") @@ -2530,6 +2768,7 @@ class Program_weight_tensor_parameter_237: class Program_weight_tensor_parameter_238: name = "parameter_238" + original_name = "conv2d_47.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.417148") @@ -2541,6 +2780,7 @@ class Program_weight_tensor_parameter_238: class Program_weight_tensor_parameter_239: name = "parameter_239" + original_name = "batch_norm2d_42.b_0" shape = [96] dtype = "float32" min_val = float("-0.368603") @@ -2552,6 +2792,7 @@ class Program_weight_tensor_parameter_239: class Program_weight_tensor_parameter_240: name = "parameter_240" + original_name = "batch_norm2d_42.w_0" shape = [96] dtype = "float32" min_val = float("0.0912661") @@ -2563,6 +2804,7 @@ class Program_weight_tensor_parameter_240: class Program_weight_tensor_parameter_241: name = "parameter_241" + original_name = "batch_norm2d_42.w_2" shape = [96] dtype = "float32" min_val = float("2.39915e-05") @@ -2574,6 +2816,7 @@ class Program_weight_tensor_parameter_241: class Program_weight_tensor_parameter_242: name = "parameter_242" + original_name = "batch_norm2d_42.w_1" shape = [96] dtype = "float32" min_val = float("-0.413515") @@ -2585,6 +2828,7 @@ class Program_weight_tensor_parameter_242: class Program_weight_tensor_parameter_243: name = "parameter_243" + original_name = "conv2d_46.w_0" shape = [96, 1, 5, 5] dtype = "float32" min_val = float("-0.280871") @@ -2596,6 +2840,7 @@ class Program_weight_tensor_parameter_243: class Program_weight_tensor_parameter_244: name = "parameter_244" + original_name = "batch_norm2d_41.b_0" shape = [96] dtype = "float32" min_val = float("-1.88184") @@ -2607,6 +2852,7 @@ class Program_weight_tensor_parameter_244: class Program_weight_tensor_parameter_245: name = "parameter_245" + original_name = "batch_norm2d_41.w_0" shape = [96] dtype = "float32" min_val = float("0.0410807") @@ -2618,6 +2864,7 @@ class Program_weight_tensor_parameter_245: class Program_weight_tensor_parameter_246: name = "parameter_246" + original_name = "batch_norm2d_41.w_2" shape = [96] dtype = "float32" min_val = float("0.107559") @@ -2629,6 +2876,7 @@ class Program_weight_tensor_parameter_246: class Program_weight_tensor_parameter_247: name = "parameter_247" + original_name = "batch_norm2d_41.w_1" shape = [96] dtype = "float32" min_val = float("-1.32878") @@ -2640,6 +2888,7 @@ class Program_weight_tensor_parameter_247: class Program_weight_tensor_parameter_248: name = "parameter_248" + original_name = "conv2d_45.w_0" shape = [96, 192, 1, 1] dtype = "float32" min_val = float("-0.488177") @@ -2651,6 +2900,7 @@ class Program_weight_tensor_parameter_248: class Program_weight_tensor_parameter_249: name = "parameter_249" + original_name = "batch_norm2d_40.b_0" shape = [192] dtype = "float32" min_val = float("-1.55766") @@ -2662,6 +2912,7 @@ class Program_weight_tensor_parameter_249: class Program_weight_tensor_parameter_250: name = "parameter_250" + original_name = "batch_norm2d_40.w_0" shape = [192] dtype = "float32" min_val = float("0.400469") @@ -2673,6 +2924,7 @@ class Program_weight_tensor_parameter_250: class Program_weight_tensor_parameter_251: name = "parameter_251" + original_name = "batch_norm2d_40.w_2" shape = [192] dtype = "float32" min_val = float("0.000224153") @@ -2684,6 +2936,7 @@ class Program_weight_tensor_parameter_251: class Program_weight_tensor_parameter_252: name = "parameter_252" + original_name = "batch_norm2d_40.w_1" shape = [192] dtype = "float32" min_val = float("-0.72552") @@ -2695,6 +2948,7 @@ class Program_weight_tensor_parameter_252: class Program_weight_tensor_parameter_253: name = "parameter_253" + original_name = "conv2d_44.w_0" shape = [192, 1, 5, 5] dtype = "float32" min_val = float("-0.565116") @@ -2706,6 +2960,7 @@ class Program_weight_tensor_parameter_253: class Program_weight_tensor_parameter_254: name = "parameter_254" + original_name = "batch_norm2d_39.b_0" shape = [192] dtype = "float32" min_val = float("-2.28197") @@ -2717,6 +2972,7 @@ class Program_weight_tensor_parameter_254: class Program_weight_tensor_parameter_255: name = "parameter_255" + original_name = "batch_norm2d_39.w_0" shape = [192] dtype = "float32" min_val = float("0.0692399") @@ -2728,6 +2984,7 @@ class Program_weight_tensor_parameter_255: class Program_weight_tensor_parameter_256: name = "parameter_256" + original_name = "batch_norm2d_39.w_2" shape = [192] dtype = "float32" min_val = float("0.185356") @@ -2739,6 +2996,7 @@ class Program_weight_tensor_parameter_256: class Program_weight_tensor_parameter_257: name = "parameter_257" + original_name = "batch_norm2d_39.w_1" shape = [192] dtype = "float32" min_val = float("-2.12608") @@ -2750,6 +3008,7 @@ class Program_weight_tensor_parameter_257: class Program_weight_tensor_parameter_258: name = "parameter_258" + original_name = "conv2d_43.w_0" shape = [192, 192, 1, 1] dtype = "float32" min_val = float("-0.436971") @@ -2761,6 +3020,7 @@ class Program_weight_tensor_parameter_258: class Program_weight_tensor_parameter_259: name = "parameter_259" + original_name = "batch_norm2d_38.b_0" shape = [192] dtype = "float32" min_val = float("-0.84437") @@ -2772,6 +3032,7 @@ class Program_weight_tensor_parameter_259: class Program_weight_tensor_parameter_260: name = "parameter_260" + original_name = "batch_norm2d_38.w_0" shape = [192] dtype = "float32" min_val = float("0.467892") @@ -2783,6 +3044,7 @@ class Program_weight_tensor_parameter_260: class Program_weight_tensor_parameter_261: name = "parameter_261" + original_name = "batch_norm2d_38.w_2" shape = [192] dtype = "float32" min_val = float("5.73361e-05") @@ -2794,6 +3056,7 @@ class Program_weight_tensor_parameter_261: class Program_weight_tensor_parameter_262: name = "parameter_262" + original_name = "batch_norm2d_38.w_1" shape = [192] dtype = "float32" min_val = float("-1.01965") @@ -2805,6 +3068,7 @@ class Program_weight_tensor_parameter_262: class Program_weight_tensor_parameter_263: name = "parameter_263" + original_name = "conv2d_42.w_0" shape = [192, 1, 5, 5] dtype = "float32" min_val = float("-0.617473") @@ -2816,6 +3080,7 @@ class Program_weight_tensor_parameter_263: class Program_weight_tensor_parameter_264: name = "parameter_264" + original_name = "batch_norm2d_37.b_0" shape = [96] dtype = "float32" min_val = float("-0.818426") @@ -2827,6 +3092,7 @@ class Program_weight_tensor_parameter_264: class Program_weight_tensor_parameter_265: name = "parameter_265" + original_name = "batch_norm2d_37.w_0" shape = [96] dtype = "float32" min_val = float("0.0514596") @@ -2838,6 +3104,7 @@ class Program_weight_tensor_parameter_265: class Program_weight_tensor_parameter_266: name = "parameter_266" + original_name = "batch_norm2d_37.w_2" shape = [96] dtype = "float32" min_val = float("0.292986") @@ -2849,6 +3116,7 @@ class Program_weight_tensor_parameter_266: class Program_weight_tensor_parameter_267: name = "parameter_267" + original_name = "batch_norm2d_37.w_1" shape = [96] dtype = "float32" min_val = float("-1.14216") @@ -2860,6 +3128,7 @@ class Program_weight_tensor_parameter_267: class Program_weight_tensor_parameter_268: name = "parameter_268" + original_name = "conv2d_41.w_0" shape = [96, 192, 1, 1] dtype = "float32" min_val = float("-0.639696") @@ -2871,6 +3140,7 @@ class Program_weight_tensor_parameter_268: class Program_weight_tensor_parameter_269: name = "parameter_269" + original_name = "batch_norm2d_36.b_0" shape = [192] dtype = "float32" min_val = float("-1.02538") @@ -2882,6 +3152,7 @@ class Program_weight_tensor_parameter_269: class Program_weight_tensor_parameter_270: name = "parameter_270" + original_name = "batch_norm2d_36.w_0" shape = [192] dtype = "float32" min_val = float("0.569226") @@ -2893,6 +3164,7 @@ class Program_weight_tensor_parameter_270: class Program_weight_tensor_parameter_271: name = "parameter_271" + original_name = "batch_norm2d_36.w_2" shape = [192] dtype = "float32" min_val = float("0.000213398") @@ -2904,6 +3176,7 @@ class Program_weight_tensor_parameter_271: class Program_weight_tensor_parameter_272: name = "parameter_272" + original_name = "batch_norm2d_36.w_1" shape = [192] dtype = "float32" min_val = float("-0.502444") @@ -2915,6 +3188,7 @@ class Program_weight_tensor_parameter_272: class Program_weight_tensor_parameter_273: name = "parameter_273" + original_name = "conv2d_40.w_0" shape = [192, 1, 5, 5] dtype = "float32" min_val = float("-0.392068") @@ -2926,6 +3200,7 @@ class Program_weight_tensor_parameter_273: class Program_weight_tensor_parameter_274: name = "parameter_274" + original_name = "batch_norm2d_35.b_0" shape = [192] dtype = "float32" min_val = float("-1.18876") @@ -2937,6 +3212,7 @@ class Program_weight_tensor_parameter_274: class Program_weight_tensor_parameter_275: name = "parameter_275" + original_name = "batch_norm2d_35.w_0" shape = [192] dtype = "float32" min_val = float("0.0733476") @@ -2948,6 +3224,7 @@ class Program_weight_tensor_parameter_275: class Program_weight_tensor_parameter_276: name = "parameter_276" + original_name = "batch_norm2d_35.w_2" shape = [192] dtype = "float32" min_val = float("0.249782") @@ -2959,6 +3236,7 @@ class Program_weight_tensor_parameter_276: class Program_weight_tensor_parameter_277: name = "parameter_277" + original_name = "batch_norm2d_35.w_1" shape = [192] dtype = "float32" min_val = float("-2.02858") @@ -2970,6 +3248,7 @@ class Program_weight_tensor_parameter_277: class Program_weight_tensor_parameter_278: name = "parameter_278" + original_name = "conv2d_39.w_0" shape = [192, 192, 1, 1] dtype = "float32" min_val = float("-0.371773") @@ -2981,6 +3260,7 @@ class Program_weight_tensor_parameter_278: class Program_weight_tensor_parameter_279: name = "parameter_279" + original_name = "batch_norm2d_34.b_0" shape = [192] dtype = "float32" min_val = float("-1.2481") @@ -2992,6 +3272,7 @@ class Program_weight_tensor_parameter_279: class Program_weight_tensor_parameter_280: name = "parameter_280" + original_name = "batch_norm2d_34.w_0" shape = [192] dtype = "float32" min_val = float("0.797116") @@ -3003,6 +3284,7 @@ class Program_weight_tensor_parameter_280: class Program_weight_tensor_parameter_281: name = "parameter_281" + original_name = "batch_norm2d_34.w_2" shape = [192] dtype = "float32" min_val = float("4.30112e-05") @@ -3014,6 +3296,7 @@ class Program_weight_tensor_parameter_281: class Program_weight_tensor_parameter_282: name = "parameter_282" + original_name = "batch_norm2d_34.w_1" shape = [192] dtype = "float32" min_val = float("-0.0568171") @@ -3025,6 +3308,7 @@ class Program_weight_tensor_parameter_282: class Program_weight_tensor_parameter_283: name = "parameter_283" + original_name = "conv2d_38.w_0" shape = [192, 1, 5, 5] dtype = "float32" min_val = float("-0.451177") @@ -3036,6 +3320,7 @@ class Program_weight_tensor_parameter_283: class Program_weight_tensor_parameter_284: name = "parameter_284" + original_name = "batch_norm2d_29.b_0" shape = [96] dtype = "float32" min_val = float("-0.258106") @@ -3047,6 +3332,7 @@ class Program_weight_tensor_parameter_284: class Program_weight_tensor_parameter_285: name = "parameter_285" + original_name = "batch_norm2d_29.w_0" shape = [96] dtype = "float32" min_val = float("0.0573399") @@ -3058,6 +3344,7 @@ class Program_weight_tensor_parameter_285: class Program_weight_tensor_parameter_286: name = "parameter_286" + original_name = "batch_norm2d_29.w_2" shape = [96] dtype = "float32" min_val = float("1.55584") @@ -3069,6 +3356,7 @@ class Program_weight_tensor_parameter_286: class Program_weight_tensor_parameter_287: name = "parameter_287" + original_name = "batch_norm2d_29.w_1" shape = [96] dtype = "float32" min_val = float("-2.10238") @@ -3080,6 +3368,7 @@ class Program_weight_tensor_parameter_287: class Program_weight_tensor_parameter_288: name = "parameter_288" + original_name = "conv2d_33.w_0" shape = [96, 384, 1, 1] dtype = "float32" min_val = float("-0.275955") @@ -3091,6 +3380,7 @@ class Program_weight_tensor_parameter_288: class Program_weight_tensor_parameter_289: name = "parameter_289" + original_name = "batch_norm2d_28.b_0" shape = [96] dtype = "float32" min_val = float("-0.109792") @@ -3102,6 +3392,7 @@ class Program_weight_tensor_parameter_289: class Program_weight_tensor_parameter_290: name = "parameter_290" + original_name = "batch_norm2d_28.w_0" shape = [96] dtype = "float32" min_val = float("0.0418628") @@ -3113,6 +3404,7 @@ class Program_weight_tensor_parameter_290: class Program_weight_tensor_parameter_291: name = "parameter_291" + original_name = "batch_norm2d_28.w_2" shape = [96] dtype = "float32" min_val = float("0.0506477") @@ -3124,6 +3416,7 @@ class Program_weight_tensor_parameter_291: class Program_weight_tensor_parameter_292: name = "parameter_292" + original_name = "batch_norm2d_28.w_1" shape = [96] dtype = "float32" min_val = float("-0.836163") @@ -3135,6 +3428,7 @@ class Program_weight_tensor_parameter_292: class Program_weight_tensor_parameter_293: name = "parameter_293" + original_name = "conv2d_32.w_0" shape = [96, 192, 1, 1] dtype = "float32" min_val = float("-0.196536") @@ -3146,6 +3440,7 @@ class Program_weight_tensor_parameter_293: class Program_weight_tensor_parameter_294: name = "parameter_294" + original_name = "batch_norm2d_27.b_0" shape = [96] dtype = "float32" min_val = float("-0.328918") @@ -3157,6 +3452,7 @@ class Program_weight_tensor_parameter_294: class Program_weight_tensor_parameter_295: name = "parameter_295" + original_name = "batch_norm2d_27.w_0" shape = [96] dtype = "float32" min_val = float("0.0323956") @@ -3168,6 +3464,7 @@ class Program_weight_tensor_parameter_295: class Program_weight_tensor_parameter_296: name = "parameter_296" + original_name = "batch_norm2d_27.w_2" shape = [96] dtype = "float32" min_val = float("0.0107128") @@ -3179,6 +3476,7 @@ class Program_weight_tensor_parameter_296: class Program_weight_tensor_parameter_297: name = "parameter_297" + original_name = "batch_norm2d_27.w_1" shape = [96] dtype = "float32" min_val = float("-0.884357") @@ -3190,6 +3488,7 @@ class Program_weight_tensor_parameter_297: class Program_weight_tensor_parameter_298: name = "parameter_298" + original_name = "conv2d_31.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.434625") @@ -3201,6 +3500,7 @@ class Program_weight_tensor_parameter_298: class Program_weight_tensor_parameter_299: name = "parameter_299" + original_name = "batch_norm2d_26.b_0" shape = [384] dtype = "float32" min_val = float("-6.11027") @@ -3212,6 +3512,7 @@ class Program_weight_tensor_parameter_299: class Program_weight_tensor_parameter_300: name = "parameter_300" + original_name = "batch_norm2d_26.w_0" shape = [384] dtype = "float32" min_val = float("2.67792") @@ -3223,6 +3524,7 @@ class Program_weight_tensor_parameter_300: class Program_weight_tensor_parameter_301: name = "parameter_301" + original_name = "batch_norm2d_26.w_2" shape = [384] dtype = "float32" min_val = float("0.0258297") @@ -3234,6 +3536,7 @@ class Program_weight_tensor_parameter_301: class Program_weight_tensor_parameter_302: name = "parameter_302" + original_name = "batch_norm2d_26.w_1" shape = [384] dtype = "float32" min_val = float("-0.499005") @@ -3245,6 +3548,7 @@ class Program_weight_tensor_parameter_302: class Program_weight_tensor_parameter_303: name = "parameter_303" + original_name = "conv2d_30.w_0" shape = [384, 384, 1, 1] dtype = "float32" min_val = float("-0.286031") @@ -3256,6 +3560,7 @@ class Program_weight_tensor_parameter_303: class Program_weight_tensor_parameter_304: name = "parameter_304" + original_name = "conv2d_29.b_0" shape = [384] dtype = "float32" min_val = float("-0.117643") @@ -3267,6 +3572,7 @@ class Program_weight_tensor_parameter_304: class Program_weight_tensor_parameter_305: name = "parameter_305" + original_name = "conv2d_29.w_0" shape = [384, 96, 1, 1] dtype = "float32" min_val = float("-0.428594") @@ -3278,6 +3584,7 @@ class Program_weight_tensor_parameter_305: class Program_weight_tensor_parameter_306: name = "parameter_306" + original_name = "conv2d_28.b_0" shape = [96] dtype = "float32" min_val = float("-0.0236603") @@ -3289,6 +3596,7 @@ class Program_weight_tensor_parameter_306: class Program_weight_tensor_parameter_307: name = "parameter_307" + original_name = "conv2d_28.w_0" shape = [96, 384, 1, 1] dtype = "float32" min_val = float("-0.339184") @@ -3300,6 +3608,7 @@ class Program_weight_tensor_parameter_307: class Program_weight_tensor_parameter_308: name = "parameter_308" + original_name = "batch_norm2d_25.b_0" shape = [384] dtype = "float32" min_val = float("-2.61335") @@ -3311,6 +3620,7 @@ class Program_weight_tensor_parameter_308: class Program_weight_tensor_parameter_309: name = "parameter_309" + original_name = "batch_norm2d_25.w_0" shape = [384] dtype = "float32" min_val = float("0.288345") @@ -3322,6 +3632,7 @@ class Program_weight_tensor_parameter_309: class Program_weight_tensor_parameter_310: name = "parameter_310" + original_name = "batch_norm2d_25.w_2" shape = [384] dtype = "float32" min_val = float("9.35599e-05") @@ -3333,6 +3644,7 @@ class Program_weight_tensor_parameter_310: class Program_weight_tensor_parameter_311: name = "parameter_311" + original_name = "batch_norm2d_25.w_1" shape = [384] dtype = "float32" min_val = float("-2.41253") @@ -3344,6 +3656,7 @@ class Program_weight_tensor_parameter_311: class Program_weight_tensor_parameter_312: name = "parameter_312" + original_name = "conv2d_27.w_0" shape = [384, 1, 5, 5] dtype = "float32" min_val = float("-0.513285") @@ -3355,6 +3668,7 @@ class Program_weight_tensor_parameter_312: class Program_weight_tensor_parameter_313: name = "parameter_313" + original_name = "batch_norm2d_24.b_0" shape = [384] dtype = "float32" min_val = float("-2.56748") @@ -3366,6 +3680,7 @@ class Program_weight_tensor_parameter_313: class Program_weight_tensor_parameter_314: name = "parameter_314" + original_name = "batch_norm2d_24.w_0" shape = [384] dtype = "float32" min_val = float("-2.95141") @@ -3377,6 +3692,7 @@ class Program_weight_tensor_parameter_314: class Program_weight_tensor_parameter_315: name = "parameter_315" + original_name = "batch_norm2d_24.w_2" shape = [384] dtype = "float32" min_val = float("0.0351946") @@ -3388,6 +3704,7 @@ class Program_weight_tensor_parameter_315: class Program_weight_tensor_parameter_316: name = "parameter_316" + original_name = "batch_norm2d_24.w_1" shape = [384] dtype = "float32" min_val = float("-0.914949") @@ -3399,6 +3716,7 @@ class Program_weight_tensor_parameter_316: class Program_weight_tensor_parameter_317: name = "parameter_317" + original_name = "conv2d_26.w_0" shape = [384, 192, 1, 1] dtype = "float32" min_val = float("-0.339303") @@ -3410,6 +3728,7 @@ class Program_weight_tensor_parameter_317: class Program_weight_tensor_parameter_318: name = "parameter_318" + original_name = "conv2d_25.b_0" shape = [192] dtype = "float32" min_val = float("-0.13279") @@ -3421,6 +3740,7 @@ class Program_weight_tensor_parameter_318: class Program_weight_tensor_parameter_319: name = "parameter_319" + original_name = "conv2d_25.w_0" shape = [192, 48, 1, 1] dtype = "float32" min_val = float("-0.526314") @@ -3432,6 +3752,7 @@ class Program_weight_tensor_parameter_319: class Program_weight_tensor_parameter_320: name = "parameter_320" + original_name = "conv2d_24.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -3441,6 +3762,7 @@ class Program_weight_tensor_parameter_320: class Program_weight_tensor_parameter_321: name = "parameter_321" + original_name = "conv2d_24.w_0" shape = [48, 192, 1, 1] dtype = "float32" min_val = float("-0.656151") @@ -3452,6 +3774,7 @@ class Program_weight_tensor_parameter_321: class Program_weight_tensor_parameter_322: name = "parameter_322" + original_name = "batch_norm2d_23.b_0" shape = [192] dtype = "float32" min_val = float("-1.41872") @@ -3463,6 +3786,7 @@ class Program_weight_tensor_parameter_322: class Program_weight_tensor_parameter_323: name = "parameter_323" + original_name = "batch_norm2d_23.w_0" shape = [192] dtype = "float32" min_val = float("0.527996") @@ -3474,6 +3798,7 @@ class Program_weight_tensor_parameter_323: class Program_weight_tensor_parameter_324: name = "parameter_324" + original_name = "batch_norm2d_23.w_2" shape = [192] dtype = "float32" min_val = float("0.000307674") @@ -3485,6 +3810,7 @@ class Program_weight_tensor_parameter_324: class Program_weight_tensor_parameter_325: name = "parameter_325" + original_name = "batch_norm2d_23.w_1" shape = [192] dtype = "float32" min_val = float("-2.86015") @@ -3496,6 +3822,7 @@ class Program_weight_tensor_parameter_325: class Program_weight_tensor_parameter_326: name = "parameter_326" + original_name = "conv2d_23.w_0" shape = [192, 1, 5, 5] dtype = "float32" min_val = float("-0.267879") @@ -3507,6 +3834,7 @@ class Program_weight_tensor_parameter_326: class Program_weight_tensor_parameter_327: name = "parameter_327" + original_name = "batch_norm2d_22.b_0" shape = [192] dtype = "float32" min_val = float("-2.45728") @@ -3518,6 +3846,7 @@ class Program_weight_tensor_parameter_327: class Program_weight_tensor_parameter_328: name = "parameter_328" + original_name = "batch_norm2d_22.w_0" shape = [192] dtype = "float32" min_val = float("0.0542238") @@ -3529,6 +3858,7 @@ class Program_weight_tensor_parameter_328: class Program_weight_tensor_parameter_329: name = "parameter_329" + original_name = "batch_norm2d_22.w_2" shape = [192] dtype = "float32" min_val = float("0.278447") @@ -3540,6 +3870,7 @@ class Program_weight_tensor_parameter_329: class Program_weight_tensor_parameter_330: name = "parameter_330" + original_name = "batch_norm2d_22.w_1" shape = [192] dtype = "float32" min_val = float("-3.26965") @@ -3551,6 +3882,7 @@ class Program_weight_tensor_parameter_330: class Program_weight_tensor_parameter_331: name = "parameter_331" + original_name = "conv2d_22.w_0" shape = [192, 192, 1, 1] dtype = "float32" min_val = float("-0.994633") @@ -3562,6 +3894,7 @@ class Program_weight_tensor_parameter_331: class Program_weight_tensor_parameter_332: name = "parameter_332" + original_name = "batch_norm2d_21.b_0" shape = [192] dtype = "float32" min_val = float("-1.65492") @@ -3573,6 +3906,7 @@ class Program_weight_tensor_parameter_332: class Program_weight_tensor_parameter_333: name = "parameter_333" + original_name = "batch_norm2d_21.w_0" shape = [192] dtype = "float32" min_val = float("0.409057") @@ -3584,6 +3918,7 @@ class Program_weight_tensor_parameter_333: class Program_weight_tensor_parameter_334: name = "parameter_334" + original_name = "batch_norm2d_21.w_2" shape = [192] dtype = "float32" min_val = float("0.000162837") @@ -3595,6 +3930,7 @@ class Program_weight_tensor_parameter_334: class Program_weight_tensor_parameter_335: name = "parameter_335" + original_name = "batch_norm2d_21.w_1" shape = [192] dtype = "float32" min_val = float("-2.24603") @@ -3606,6 +3942,7 @@ class Program_weight_tensor_parameter_335: class Program_weight_tensor_parameter_336: name = "parameter_336" + original_name = "conv2d_21.w_0" shape = [192, 1, 5, 5] dtype = "float32" min_val = float("-0.806895") @@ -3617,6 +3954,7 @@ class Program_weight_tensor_parameter_336: class Program_weight_tensor_parameter_337: name = "parameter_337" + original_name = "batch_norm2d_20.b_0" shape = [192] dtype = "float32" min_val = float("-3.95405") @@ -3628,6 +3966,7 @@ class Program_weight_tensor_parameter_337: class Program_weight_tensor_parameter_338: name = "parameter_338" + original_name = "batch_norm2d_20.w_0" shape = [192] dtype = "float32" min_val = float("-0.869903") @@ -3639,6 +3978,7 @@ class Program_weight_tensor_parameter_338: class Program_weight_tensor_parameter_339: name = "parameter_339" + original_name = "batch_norm2d_20.w_2" shape = [192] dtype = "float32" min_val = float("0.322026") @@ -3650,6 +3990,7 @@ class Program_weight_tensor_parameter_339: class Program_weight_tensor_parameter_340: name = "parameter_340" + original_name = "batch_norm2d_20.w_1" shape = [192] dtype = "float32" min_val = float("-4.11024") @@ -3661,6 +4002,7 @@ class Program_weight_tensor_parameter_340: class Program_weight_tensor_parameter_341: name = "parameter_341" + original_name = "conv2d_20.w_0" shape = [192, 192, 1, 1] dtype = "float32" min_val = float("-0.493816") @@ -3672,6 +4014,7 @@ class Program_weight_tensor_parameter_341: class Program_weight_tensor_parameter_342: name = "parameter_342" + original_name = "batch_norm2d_19.b_0" shape = [192] dtype = "float32" min_val = float("-3.73867") @@ -3683,6 +4026,7 @@ class Program_weight_tensor_parameter_342: class Program_weight_tensor_parameter_343: name = "parameter_343" + original_name = "batch_norm2d_19.w_0" shape = [192] dtype = "float32" min_val = float("0.456435") @@ -3694,6 +4038,7 @@ class Program_weight_tensor_parameter_343: class Program_weight_tensor_parameter_344: name = "parameter_344" + original_name = "batch_norm2d_19.w_2" shape = [192] dtype = "float32" min_val = float("0.000142246") @@ -3705,6 +4050,7 @@ class Program_weight_tensor_parameter_344: class Program_weight_tensor_parameter_345: name = "parameter_345" + original_name = "batch_norm2d_19.w_1" shape = [192] dtype = "float32" min_val = float("-7.16568") @@ -3716,6 +4062,7 @@ class Program_weight_tensor_parameter_345: class Program_weight_tensor_parameter_346: name = "parameter_346" + original_name = "conv2d_19.w_0" shape = [192, 1, 5, 5] dtype = "float32" min_val = float("-0.649686") @@ -3727,6 +4074,7 @@ class Program_weight_tensor_parameter_346: class Program_weight_tensor_parameter_347: name = "parameter_347" + original_name = "batch_norm2d_18.b_0" shape = [192] dtype = "float32" min_val = float("-2.31648") @@ -3738,6 +4086,7 @@ class Program_weight_tensor_parameter_347: class Program_weight_tensor_parameter_348: name = "parameter_348" + original_name = "batch_norm2d_18.w_0" shape = [192] dtype = "float32" min_val = float("-0.767004") @@ -3749,6 +4098,7 @@ class Program_weight_tensor_parameter_348: class Program_weight_tensor_parameter_349: name = "parameter_349" + original_name = "batch_norm2d_18.w_2" shape = [192] dtype = "float32" min_val = float("0.261792") @@ -3760,6 +4110,7 @@ class Program_weight_tensor_parameter_349: class Program_weight_tensor_parameter_350: name = "parameter_350" + original_name = "batch_norm2d_18.w_1" shape = [192] dtype = "float32" min_val = float("-4.86867") @@ -3771,6 +4122,7 @@ class Program_weight_tensor_parameter_350: class Program_weight_tensor_parameter_351: name = "parameter_351" + original_name = "conv2d_18.w_0" shape = [192, 192, 1, 1] dtype = "float32" min_val = float("-0.452378") @@ -3782,6 +4134,7 @@ class Program_weight_tensor_parameter_351: class Program_weight_tensor_parameter_352: name = "parameter_352" + original_name = "batch_norm2d_17.b_0" shape = [192] dtype = "float32" min_val = float("-2.75603") @@ -3793,6 +4146,7 @@ class Program_weight_tensor_parameter_352: class Program_weight_tensor_parameter_353: name = "parameter_353" + original_name = "batch_norm2d_17.w_0" shape = [192] dtype = "float32" min_val = float("0.477941") @@ -3804,6 +4158,7 @@ class Program_weight_tensor_parameter_353: class Program_weight_tensor_parameter_354: name = "parameter_354" + original_name = "batch_norm2d_17.w_2" shape = [192] dtype = "float32" min_val = float("0.000129202") @@ -3815,6 +4170,7 @@ class Program_weight_tensor_parameter_354: class Program_weight_tensor_parameter_355: name = "parameter_355" + original_name = "batch_norm2d_17.w_1" shape = [192] dtype = "float32" min_val = float("-3.79941") @@ -3826,6 +4182,7 @@ class Program_weight_tensor_parameter_355: class Program_weight_tensor_parameter_356: name = "parameter_356" + original_name = "conv2d_17.w_0" shape = [192, 1, 5, 5] dtype = "float32" min_val = float("-0.779576") @@ -3837,6 +4194,7 @@ class Program_weight_tensor_parameter_356: class Program_weight_tensor_parameter_357: name = "parameter_357" + original_name = "batch_norm2d_16.b_0" shape = [192] dtype = "float32" min_val = float("-2.80439") @@ -3848,6 +4206,7 @@ class Program_weight_tensor_parameter_357: class Program_weight_tensor_parameter_358: name = "parameter_358" + original_name = "batch_norm2d_16.w_0" shape = [192] dtype = "float32" min_val = float("-0.838245") @@ -3859,6 +4218,7 @@ class Program_weight_tensor_parameter_358: class Program_weight_tensor_parameter_359: name = "parameter_359" + original_name = "batch_norm2d_16.w_2" shape = [192] dtype = "float32" min_val = float("0.284666") @@ -3870,6 +4230,7 @@ class Program_weight_tensor_parameter_359: class Program_weight_tensor_parameter_360: name = "parameter_360" + original_name = "batch_norm2d_16.w_1" shape = [192] dtype = "float32" min_val = float("-5.90195") @@ -3881,6 +4242,7 @@ class Program_weight_tensor_parameter_360: class Program_weight_tensor_parameter_361: name = "parameter_361" + original_name = "conv2d_16.w_0" shape = [192, 192, 1, 1] dtype = "float32" min_val = float("-0.485193") @@ -3892,6 +4254,7 @@ class Program_weight_tensor_parameter_361: class Program_weight_tensor_parameter_362: name = "parameter_362" + original_name = "batch_norm2d_15.b_0" shape = [192] dtype = "float32" min_val = float("-1.64011") @@ -3903,6 +4266,7 @@ class Program_weight_tensor_parameter_362: class Program_weight_tensor_parameter_363: name = "parameter_363" + original_name = "batch_norm2d_15.w_0" shape = [192] dtype = "float32" min_val = float("0.451585") @@ -3914,6 +4278,7 @@ class Program_weight_tensor_parameter_363: class Program_weight_tensor_parameter_364: name = "parameter_364" + original_name = "batch_norm2d_15.w_2" shape = [192] dtype = "float32" min_val = float("0.000168566") @@ -3925,6 +4290,7 @@ class Program_weight_tensor_parameter_364: class Program_weight_tensor_parameter_365: name = "parameter_365" + original_name = "batch_norm2d_15.w_1" shape = [192] dtype = "float32" min_val = float("-2.94921") @@ -3936,6 +4302,7 @@ class Program_weight_tensor_parameter_365: class Program_weight_tensor_parameter_366: name = "parameter_366" + original_name = "conv2d_15.w_0" shape = [192, 1, 5, 5] dtype = "float32" min_val = float("-0.814876") @@ -3947,6 +4314,7 @@ class Program_weight_tensor_parameter_366: class Program_weight_tensor_parameter_367: name = "parameter_367" + original_name = "batch_norm2d_14.b_0" shape = [192] dtype = "float32" min_val = float("-1.76006") @@ -3958,6 +4326,7 @@ class Program_weight_tensor_parameter_367: class Program_weight_tensor_parameter_368: name = "parameter_368" + original_name = "batch_norm2d_14.w_0" shape = [192] dtype = "float32" min_val = float("-1.20688") @@ -3969,6 +4338,7 @@ class Program_weight_tensor_parameter_368: class Program_weight_tensor_parameter_369: name = "parameter_369" + original_name = "batch_norm2d_14.w_2" shape = [192] dtype = "float32" min_val = float("0.404626") @@ -3980,6 +4350,7 @@ class Program_weight_tensor_parameter_369: class Program_weight_tensor_parameter_370: name = "parameter_370" + original_name = "batch_norm2d_14.w_1" shape = [192] dtype = "float32" min_val = float("-5.09013") @@ -3991,6 +4362,7 @@ class Program_weight_tensor_parameter_370: class Program_weight_tensor_parameter_371: name = "parameter_371" + original_name = "conv2d_14.w_0" shape = [192, 192, 1, 1] dtype = "float32" min_val = float("-0.802113") @@ -4002,6 +4374,7 @@ class Program_weight_tensor_parameter_371: class Program_weight_tensor_parameter_372: name = "parameter_372" + original_name = "batch_norm2d_13.b_0" shape = [192] dtype = "float32" min_val = float("-3.91487") @@ -4013,6 +4386,7 @@ class Program_weight_tensor_parameter_372: class Program_weight_tensor_parameter_373: name = "parameter_373" + original_name = "batch_norm2d_13.w_0" shape = [192] dtype = "float32" min_val = float("0.456656") @@ -4024,6 +4398,7 @@ class Program_weight_tensor_parameter_373: class Program_weight_tensor_parameter_374: name = "parameter_374" + original_name = "batch_norm2d_13.w_2" shape = [192] dtype = "float32" min_val = float("0.000108652") @@ -4035,6 +4410,7 @@ class Program_weight_tensor_parameter_374: class Program_weight_tensor_parameter_375: name = "parameter_375" + original_name = "batch_norm2d_13.w_1" shape = [192] dtype = "float32" min_val = float("-4.84769") @@ -4046,6 +4422,7 @@ class Program_weight_tensor_parameter_375: class Program_weight_tensor_parameter_376: name = "parameter_376" + original_name = "conv2d_13.w_0" shape = [192, 1, 5, 5] dtype = "float32" min_val = float("-0.821266") @@ -4057,6 +4434,7 @@ class Program_weight_tensor_parameter_376: class Program_weight_tensor_parameter_377: name = "parameter_377" + original_name = "batch_norm2d_12.b_0" shape = [192] dtype = "float32" min_val = float("-1.52828") @@ -4068,6 +4446,7 @@ class Program_weight_tensor_parameter_377: class Program_weight_tensor_parameter_378: name = "parameter_378" + original_name = "batch_norm2d_12.w_0" shape = [192] dtype = "float32" min_val = float("-0.423076") @@ -4079,6 +4458,7 @@ class Program_weight_tensor_parameter_378: class Program_weight_tensor_parameter_379: name = "parameter_379" + original_name = "batch_norm2d_12.w_2" shape = [192] dtype = "float32" min_val = float("0.0954261") @@ -4090,6 +4470,7 @@ class Program_weight_tensor_parameter_379: class Program_weight_tensor_parameter_380: name = "parameter_380" + original_name = "batch_norm2d_12.w_1" shape = [192] dtype = "float32" min_val = float("-5.90915") @@ -4101,6 +4482,7 @@ class Program_weight_tensor_parameter_380: class Program_weight_tensor_parameter_381: name = "parameter_381" + original_name = "conv2d_12.w_0" shape = [192, 96, 1, 1] dtype = "float32" min_val = float("-0.661111") @@ -4112,6 +4494,7 @@ class Program_weight_tensor_parameter_381: class Program_weight_tensor_parameter_382: name = "parameter_382" + original_name = "batch_norm2d_11.b_0" shape = [96] dtype = "float32" min_val = float("-1.08049") @@ -4123,6 +4506,7 @@ class Program_weight_tensor_parameter_382: class Program_weight_tensor_parameter_383: name = "parameter_383" + original_name = "batch_norm2d_11.w_0" shape = [96] dtype = "float32" min_val = float("0.352709") @@ -4134,6 +4518,7 @@ class Program_weight_tensor_parameter_383: class Program_weight_tensor_parameter_384: name = "parameter_384" + original_name = "batch_norm2d_11.w_2" shape = [96] dtype = "float32" min_val = float("0.00165767") @@ -4145,6 +4530,7 @@ class Program_weight_tensor_parameter_384: class Program_weight_tensor_parameter_385: name = "parameter_385" + original_name = "batch_norm2d_11.w_1" shape = [96] dtype = "float32" min_val = float("-2.73023") @@ -4156,6 +4542,7 @@ class Program_weight_tensor_parameter_385: class Program_weight_tensor_parameter_386: name = "parameter_386" + original_name = "conv2d_11.w_0" shape = [96, 1, 3, 3] dtype = "float32" min_val = float("-0.29978") @@ -4167,6 +4554,7 @@ class Program_weight_tensor_parameter_386: class Program_weight_tensor_parameter_387: name = "parameter_387" + original_name = "batch_norm2d_10.b_0" shape = [96] dtype = "float32" min_val = float("-1.63841") @@ -4178,6 +4566,7 @@ class Program_weight_tensor_parameter_387: class Program_weight_tensor_parameter_388: name = "parameter_388" + original_name = "batch_norm2d_10.w_0" shape = [96] dtype = "float32" min_val = float("0.237488") @@ -4189,6 +4578,7 @@ class Program_weight_tensor_parameter_388: class Program_weight_tensor_parameter_389: name = "parameter_389" + original_name = "batch_norm2d_10.w_2" shape = [96] dtype = "float32" min_val = float("0.634253") @@ -4200,6 +4590,7 @@ class Program_weight_tensor_parameter_389: class Program_weight_tensor_parameter_390: name = "parameter_390" + original_name = "batch_norm2d_10.w_1" shape = [96] dtype = "float32" min_val = float("-10.3317") @@ -4211,6 +4602,7 @@ class Program_weight_tensor_parameter_390: class Program_weight_tensor_parameter_391: name = "parameter_391" + original_name = "conv2d_10.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.96274") @@ -4222,6 +4614,7 @@ class Program_weight_tensor_parameter_391: class Program_weight_tensor_parameter_392: name = "parameter_392" + original_name = "batch_norm2d_9.b_0" shape = [96] dtype = "float32" min_val = float("-1.19896") @@ -4233,6 +4626,7 @@ class Program_weight_tensor_parameter_392: class Program_weight_tensor_parameter_393: name = "parameter_393" + original_name = "batch_norm2d_9.w_0" shape = [96] dtype = "float32" min_val = float("0.550015") @@ -4244,6 +4638,7 @@ class Program_weight_tensor_parameter_393: class Program_weight_tensor_parameter_394: name = "parameter_394" + original_name = "batch_norm2d_9.w_2" shape = [96] dtype = "float32" min_val = float("8.92012e-05") @@ -4255,6 +4650,7 @@ class Program_weight_tensor_parameter_394: class Program_weight_tensor_parameter_395: name = "parameter_395" + original_name = "batch_norm2d_9.w_1" shape = [96] dtype = "float32" min_val = float("-8.79784") @@ -4266,6 +4662,7 @@ class Program_weight_tensor_parameter_395: class Program_weight_tensor_parameter_396: name = "parameter_396" + original_name = "conv2d_9.w_0" shape = [96, 1, 3, 3] dtype = "float32" min_val = float("-1.10303") @@ -4277,6 +4674,7 @@ class Program_weight_tensor_parameter_396: class Program_weight_tensor_parameter_397: name = "parameter_397" + original_name = "batch_norm2d_8.b_0" shape = [96] dtype = "float32" min_val = float("-1.31793") @@ -4288,6 +4686,7 @@ class Program_weight_tensor_parameter_397: class Program_weight_tensor_parameter_398: name = "parameter_398" + original_name = "batch_norm2d_8.w_0" shape = [96] dtype = "float32" min_val = float("-0.366671") @@ -4299,6 +4698,7 @@ class Program_weight_tensor_parameter_398: class Program_weight_tensor_parameter_399: name = "parameter_399" + original_name = "batch_norm2d_8.w_2" shape = [96] dtype = "float32" min_val = float("0.142347") @@ -4310,6 +4710,7 @@ class Program_weight_tensor_parameter_399: class Program_weight_tensor_parameter_400: name = "parameter_400" + original_name = "batch_norm2d_8.w_1" shape = [96] dtype = "float32" min_val = float("-9.11412") @@ -4321,6 +4722,7 @@ class Program_weight_tensor_parameter_400: class Program_weight_tensor_parameter_401: name = "parameter_401" + original_name = "conv2d_8.w_0" shape = [96, 48, 1, 1] dtype = "float32" min_val = float("-1.11613") @@ -4332,6 +4734,7 @@ class Program_weight_tensor_parameter_401: class Program_weight_tensor_parameter_402: name = "parameter_402" + original_name = "batch_norm2d_7.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -4341,6 +4744,7 @@ class Program_weight_tensor_parameter_402: class Program_weight_tensor_parameter_403: name = "parameter_403" + original_name = "batch_norm2d_7.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -4350,6 +4754,7 @@ class Program_weight_tensor_parameter_403: class Program_weight_tensor_parameter_404: name = "parameter_404" + original_name = "batch_norm2d_7.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -4359,6 +4764,7 @@ class Program_weight_tensor_parameter_404: class Program_weight_tensor_parameter_405: name = "parameter_405" + original_name = "batch_norm2d_7.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -4368,6 +4774,7 @@ class Program_weight_tensor_parameter_405: class Program_weight_tensor_parameter_406: name = "parameter_406" + original_name = "conv2d_7.w_0" shape = [48, 1, 3, 3] dtype = "float32" min_val = float("-0.529541") @@ -4379,6 +4786,7 @@ class Program_weight_tensor_parameter_406: class Program_weight_tensor_parameter_407: name = "parameter_407" + original_name = "batch_norm2d_6.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -4388,6 +4796,7 @@ class Program_weight_tensor_parameter_407: class Program_weight_tensor_parameter_408: name = "parameter_408" + original_name = "batch_norm2d_6.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -4397,6 +4806,7 @@ class Program_weight_tensor_parameter_408: class Program_weight_tensor_parameter_409: name = "parameter_409" + original_name = "batch_norm2d_6.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -4406,6 +4816,7 @@ class Program_weight_tensor_parameter_409: class Program_weight_tensor_parameter_410: name = "parameter_410" + original_name = "batch_norm2d_6.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -4415,6 +4826,7 @@ class Program_weight_tensor_parameter_410: class Program_weight_tensor_parameter_411: name = "parameter_411" + original_name = "conv2d_6.w_0" shape = [48, 48, 1, 1] dtype = "float32" min_val = float("-1.20793") @@ -4426,6 +4838,7 @@ class Program_weight_tensor_parameter_411: class Program_weight_tensor_parameter_412: name = "parameter_412" + original_name = "batch_norm2d_5.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -4435,6 +4848,7 @@ class Program_weight_tensor_parameter_412: class Program_weight_tensor_parameter_413: name = "parameter_413" + original_name = "batch_norm2d_5.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -4444,6 +4858,7 @@ class Program_weight_tensor_parameter_413: class Program_weight_tensor_parameter_414: name = "parameter_414" + original_name = "batch_norm2d_5.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -4453,6 +4868,7 @@ class Program_weight_tensor_parameter_414: class Program_weight_tensor_parameter_415: name = "parameter_415" + original_name = "batch_norm2d_5.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -4462,6 +4878,7 @@ class Program_weight_tensor_parameter_415: class Program_weight_tensor_parameter_416: name = "parameter_416" + original_name = "conv2d_5.w_0" shape = [48, 1, 3, 3] dtype = "float32" min_val = float("-1.50698") @@ -4473,6 +4890,7 @@ class Program_weight_tensor_parameter_416: class Program_weight_tensor_parameter_417: name = "parameter_417" + original_name = "batch_norm2d_4.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -4482,6 +4900,7 @@ class Program_weight_tensor_parameter_417: class Program_weight_tensor_parameter_418: name = "parameter_418" + original_name = "batch_norm2d_4.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -4491,6 +4910,7 @@ class Program_weight_tensor_parameter_418: class Program_weight_tensor_parameter_419: name = "parameter_419" + original_name = "batch_norm2d_4.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -4500,6 +4920,7 @@ class Program_weight_tensor_parameter_419: class Program_weight_tensor_parameter_420: name = "parameter_420" + original_name = "batch_norm2d_4.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -4509,6 +4930,7 @@ class Program_weight_tensor_parameter_420: class Program_weight_tensor_parameter_421: name = "parameter_421" + original_name = "conv2d_4.w_0" shape = [48, 24, 1, 1] dtype = "float32" min_val = float("-1.37878") @@ -4520,6 +4942,7 @@ class Program_weight_tensor_parameter_421: class Program_weight_tensor_parameter_422: name = "parameter_422" + original_name = "batch_norm2d_3.b_0" shape = [24] dtype = "float32" min_val = float("0") @@ -4529,6 +4952,7 @@ class Program_weight_tensor_parameter_422: class Program_weight_tensor_parameter_423: name = "parameter_423" + original_name = "batch_norm2d_3.w_0" shape = [24] dtype = "float32" min_val = float("0") @@ -4538,6 +4962,7 @@ class Program_weight_tensor_parameter_423: class Program_weight_tensor_parameter_424: name = "parameter_424" + original_name = "batch_norm2d_3.w_2" shape = [24] dtype = "float32" min_val = float("0") @@ -4547,6 +4972,7 @@ class Program_weight_tensor_parameter_424: class Program_weight_tensor_parameter_425: name = "parameter_425" + original_name = "batch_norm2d_3.w_1" shape = [24] dtype = "float32" min_val = float("0") @@ -4556,6 +4982,7 @@ class Program_weight_tensor_parameter_425: class Program_weight_tensor_parameter_426: name = "parameter_426" + original_name = "conv2d_3.w_0" shape = [24, 1, 3, 3] dtype = "float32" min_val = float("-0.801611") @@ -4567,6 +4994,7 @@ class Program_weight_tensor_parameter_426: class Program_weight_tensor_parameter_427: name = "parameter_427" + original_name = "batch_norm2d_2.b_0" shape = [24] dtype = "float32" min_val = float("0") @@ -4576,6 +5004,7 @@ class Program_weight_tensor_parameter_427: class Program_weight_tensor_parameter_428: name = "parameter_428" + original_name = "batch_norm2d_2.w_0" shape = [24] dtype = "float32" min_val = float("0") @@ -4585,6 +5014,7 @@ class Program_weight_tensor_parameter_428: class Program_weight_tensor_parameter_429: name = "parameter_429" + original_name = "batch_norm2d_2.w_2" shape = [24] dtype = "float32" min_val = float("0") @@ -4594,6 +5024,7 @@ class Program_weight_tensor_parameter_429: class Program_weight_tensor_parameter_430: name = "parameter_430" + original_name = "batch_norm2d_2.w_1" shape = [24] dtype = "float32" min_val = float("0") @@ -4603,6 +5034,7 @@ class Program_weight_tensor_parameter_430: class Program_weight_tensor_parameter_431: name = "parameter_431" + original_name = "conv2d_2.w_0" shape = [24, 16, 1, 1] dtype = "float32" min_val = float("-1.47472") @@ -4614,6 +5046,7 @@ class Program_weight_tensor_parameter_431: class Program_weight_tensor_parameter_432: name = "parameter_432" + original_name = "batch_norm2d_1.b_0" shape = [16] dtype = "float32" min_val = float("0") @@ -4623,6 +5056,7 @@ class Program_weight_tensor_parameter_432: class Program_weight_tensor_parameter_433: name = "parameter_433" + original_name = "batch_norm2d_1.w_0" shape = [16] dtype = "float32" min_val = float("0") @@ -4632,6 +5066,7 @@ class Program_weight_tensor_parameter_433: class Program_weight_tensor_parameter_434: name = "parameter_434" + original_name = "batch_norm2d_1.w_2" shape = [16] dtype = "float32" min_val = float("0") @@ -4641,6 +5076,7 @@ class Program_weight_tensor_parameter_434: class Program_weight_tensor_parameter_435: name = "parameter_435" + original_name = "batch_norm2d_1.w_1" shape = [16] dtype = "float32" min_val = float("0") @@ -4650,6 +5086,7 @@ class Program_weight_tensor_parameter_435: class Program_weight_tensor_parameter_436: name = "parameter_436" + original_name = "conv2d_1.w_0" shape = [16, 1, 3, 3] dtype = "float32" min_val = float("-1.82883") @@ -4661,6 +5098,7 @@ class Program_weight_tensor_parameter_436: class Program_weight_tensor_parameter_437: name = "parameter_437" + original_name = "batch_norm2d_0.b_0" shape = [16] dtype = "float32" min_val = float("0") @@ -4670,6 +5108,7 @@ class Program_weight_tensor_parameter_437: class Program_weight_tensor_parameter_438: name = "parameter_438" + original_name = "batch_norm2d_0.w_0" shape = [16] dtype = "float32" min_val = float("0") @@ -4679,6 +5118,7 @@ class Program_weight_tensor_parameter_438: class Program_weight_tensor_parameter_439: name = "parameter_439" + original_name = "batch_norm2d_0.w_2" shape = [16] dtype = "float32" min_val = float("0") @@ -4688,6 +5128,7 @@ class Program_weight_tensor_parameter_439: class Program_weight_tensor_parameter_440: name = "parameter_440" + original_name = "batch_norm2d_0.w_1" shape = [16] dtype = "float32" min_val = float("0") @@ -4697,6 +5138,7 @@ class Program_weight_tensor_parameter_440: class Program_weight_tensor_parameter_441: name = "parameter_441" + original_name = "conv2d_0.w_0" shape = [16, 3, 3, 3] dtype = "float32" min_val = float("-1.19149") diff --git a/paddle_samples/PaddleX/PicoDet-S/subgraph_13/input_meta.py b/paddle_samples/PaddleX/PicoDet-S/subgraph_13/input_meta.py index 79822ac1e..ece307651 100644 --- a/paddle_samples/PaddleX/PicoDet-S/subgraph_13/input_meta.py +++ b/paddle_samples/PaddleX/PicoDet-S/subgraph_13/input_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_data_0: name = "data_0" + original_name = "var_610" shape = [16, 3, 384, 384] dtype = "float32" min_val = float("-2.1179") @@ -11,6 +12,7 @@ class Program_weight_tensor_data_0: class Program_weight_tensor_data_1: name = "data_1" + original_name = "var_847" shape = [8] dtype = "float32" data = [0.0, 1.0, 2.0, 3.0003, 4.0, 5.0, 6.0006, 7.0] diff --git a/paddle_samples/PaddleX/PicoDet-S/subgraph_13/weight_meta.py b/paddle_samples/PaddleX/PicoDet-S/subgraph_13/weight_meta.py index cb84b5be1..22f560ba3 100644 --- a/paddle_samples/PaddleX/PicoDet-S/subgraph_13/weight_meta.py +++ b/paddle_samples/PaddleX/PicoDet-S/subgraph_13/weight_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_parameter_0: name = "parameter_0" + original_name = "batch_norm2d_81.b_0" shape = [1] dtype = "float32" min_val = float("0") @@ -9,6 +10,7 @@ class Program_weight_tensor_parameter_0: class Program_weight_tensor_parameter_1: name = "parameter_1" + original_name = "batch_norm2d_81.w_0" shape = [1] dtype = "float32" min_val = float("0") @@ -18,6 +20,7 @@ class Program_weight_tensor_parameter_1: class Program_weight_tensor_parameter_2: name = "parameter_2" + original_name = "batch_norm2d_81.w_2" shape = [1] dtype = "float32" min_val = float("0") @@ -27,6 +30,7 @@ class Program_weight_tensor_parameter_2: class Program_weight_tensor_parameter_3: name = "parameter_3" + original_name = "batch_norm2d_81.w_1" shape = [1] dtype = "float32" min_val = float("0") @@ -36,6 +40,7 @@ class Program_weight_tensor_parameter_3: class Program_weight_tensor_parameter_4: name = "parameter_4" + original_name = "conv2d_99.w_0" shape = [1, 1, 1, 1] dtype = "float32" min_val = float("0") @@ -45,6 +50,7 @@ class Program_weight_tensor_parameter_4: class Program_weight_tensor_parameter_5: name = "parameter_5" + original_name = "batch_norm2d_80.b_0" shape = [1] dtype = "float32" min_val = float("0") @@ -54,6 +60,7 @@ class Program_weight_tensor_parameter_5: class Program_weight_tensor_parameter_6: name = "parameter_6" + original_name = "batch_norm2d_80.w_0" shape = [1] dtype = "float32" min_val = float("0") @@ -63,6 +70,7 @@ class Program_weight_tensor_parameter_6: class Program_weight_tensor_parameter_7: name = "parameter_7" + original_name = "batch_norm2d_80.w_2" shape = [1] dtype = "float32" min_val = float("0") @@ -72,6 +80,7 @@ class Program_weight_tensor_parameter_7: class Program_weight_tensor_parameter_8: name = "parameter_8" + original_name = "batch_norm2d_80.w_1" shape = [1] dtype = "float32" min_val = float("0") @@ -81,6 +90,7 @@ class Program_weight_tensor_parameter_8: class Program_weight_tensor_parameter_9: name = "parameter_9" + original_name = "conv2d_98.w_0" shape = [1, 96, 5, 5] dtype = "float32" min_val = float("-0.269415") @@ -92,6 +102,7 @@ class Program_weight_tensor_parameter_9: class Program_weight_tensor_parameter_10: name = "parameter_10" + original_name = "conv2d_97.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -101,6 +112,7 @@ class Program_weight_tensor_parameter_10: class Program_weight_tensor_parameter_11: name = "parameter_11" + original_name = "conv2d_97.w_0" shape = [32, 96, 1, 1] dtype = "float32" min_val = float("-0.482699") @@ -112,6 +124,7 @@ class Program_weight_tensor_parameter_11: class Program_weight_tensor_parameter_12: name = "parameter_12" + original_name = "conv2d_96.b_0" shape = [4] dtype = "float32" min_val = float("0") @@ -121,6 +134,7 @@ class Program_weight_tensor_parameter_12: class Program_weight_tensor_parameter_13: name = "parameter_13" + original_name = "conv2d_96.w_0" shape = [4, 96, 1, 1] dtype = "float32" min_val = float("-0.0262188") @@ -132,6 +146,7 @@ class Program_weight_tensor_parameter_13: class Program_weight_tensor_parameter_14: name = "parameter_14" + original_name = "batch_norm2d_73.b_0" shape = [96] dtype = "float32" min_val = float("-0.0386536") @@ -143,6 +158,7 @@ class Program_weight_tensor_parameter_14: class Program_weight_tensor_parameter_15: name = "parameter_15" + original_name = "batch_norm2d_73.w_0" shape = [96] dtype = "float32" min_val = float("0.828462") @@ -154,6 +170,7 @@ class Program_weight_tensor_parameter_15: class Program_weight_tensor_parameter_16: name = "parameter_16" + original_name = "batch_norm2d_73.w_2" shape = [96] dtype = "float32" min_val = float("0.0041606") @@ -165,6 +182,7 @@ class Program_weight_tensor_parameter_16: class Program_weight_tensor_parameter_17: name = "parameter_17" + original_name = "batch_norm2d_73.w_1" shape = [96] dtype = "float32" min_val = float("-0.135106") @@ -176,6 +194,7 @@ class Program_weight_tensor_parameter_17: class Program_weight_tensor_parameter_18: name = "parameter_18" + original_name = "conv2d_81.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.232158") @@ -187,6 +206,7 @@ class Program_weight_tensor_parameter_18: class Program_weight_tensor_parameter_19: name = "parameter_19" + original_name = "conv2d_80.b_0" shape = [96] dtype = "float32" min_val = float("-0.0389175") @@ -198,6 +218,7 @@ class Program_weight_tensor_parameter_19: class Program_weight_tensor_parameter_20: name = "parameter_20" + original_name = "conv2d_80.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.161451") @@ -209,6 +230,7 @@ class Program_weight_tensor_parameter_20: class Program_weight_tensor_parameter_21: name = "parameter_21" + original_name = "batch_norm2d_72.b_0" shape = [96] dtype = "float32" min_val = float("-1.39206") @@ -220,6 +242,7 @@ class Program_weight_tensor_parameter_21: class Program_weight_tensor_parameter_22: name = "parameter_22" + original_name = "batch_norm2d_72.w_0" shape = [96] dtype = "float32" min_val = float("0.579307") @@ -231,6 +254,7 @@ class Program_weight_tensor_parameter_22: class Program_weight_tensor_parameter_23: name = "parameter_23" + original_name = "batch_norm2d_72.w_2" shape = [96] dtype = "float32" min_val = float("0.0251631") @@ -242,6 +266,7 @@ class Program_weight_tensor_parameter_23: class Program_weight_tensor_parameter_24: name = "parameter_24" + original_name = "batch_norm2d_72.w_1" shape = [96] dtype = "float32" min_val = float("-0.625812") @@ -253,6 +278,7 @@ class Program_weight_tensor_parameter_24: class Program_weight_tensor_parameter_25: name = "parameter_25" + original_name = "conv2d_79.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.331948") @@ -264,6 +290,7 @@ class Program_weight_tensor_parameter_25: class Program_weight_tensor_parameter_26: name = "parameter_26" + original_name = "batch_norm2d_71.b_0" shape = [96] dtype = "float32" min_val = float("-0.930938") @@ -275,6 +302,7 @@ class Program_weight_tensor_parameter_26: class Program_weight_tensor_parameter_27: name = "parameter_27" + original_name = "batch_norm2d_71.w_0" shape = [96] dtype = "float32" min_val = float("0.534431") @@ -286,6 +314,7 @@ class Program_weight_tensor_parameter_27: class Program_weight_tensor_parameter_28: name = "parameter_28" + original_name = "batch_norm2d_71.w_2" shape = [96] dtype = "float32" min_val = float("0.00172193") @@ -297,6 +326,7 @@ class Program_weight_tensor_parameter_28: class Program_weight_tensor_parameter_29: name = "parameter_29" + original_name = "batch_norm2d_71.w_1" shape = [96] dtype = "float32" min_val = float("-0.166487") @@ -308,6 +338,7 @@ class Program_weight_tensor_parameter_29: class Program_weight_tensor_parameter_30: name = "parameter_30" + original_name = "conv2d_78.w_0" shape = [96, 1, 5, 5] dtype = "float32" min_val = float("-0.150457") @@ -319,6 +350,7 @@ class Program_weight_tensor_parameter_30: class Program_weight_tensor_parameter_31: name = "parameter_31" + original_name = "batch_norm2d_70.b_0" shape = [96] dtype = "float32" min_val = float("-0.973499") @@ -330,6 +362,7 @@ class Program_weight_tensor_parameter_31: class Program_weight_tensor_parameter_32: name = "parameter_32" + original_name = "batch_norm2d_70.w_0" shape = [96] dtype = "float32" min_val = float("0.217951") @@ -341,6 +374,7 @@ class Program_weight_tensor_parameter_32: class Program_weight_tensor_parameter_33: name = "parameter_33" + original_name = "batch_norm2d_70.w_2" shape = [96] dtype = "float32" min_val = float("0.0236934") @@ -352,6 +386,7 @@ class Program_weight_tensor_parameter_33: class Program_weight_tensor_parameter_34: name = "parameter_34" + original_name = "batch_norm2d_70.w_1" shape = [96] dtype = "float32" min_val = float("-0.266494") @@ -363,6 +398,7 @@ class Program_weight_tensor_parameter_34: class Program_weight_tensor_parameter_35: name = "parameter_35" + original_name = "conv2d_77.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.258252") @@ -374,6 +410,7 @@ class Program_weight_tensor_parameter_35: class Program_weight_tensor_parameter_36: name = "parameter_36" + original_name = "batch_norm2d_69.b_0" shape = [96] dtype = "float32" min_val = float("-0.70886") @@ -385,6 +422,7 @@ class Program_weight_tensor_parameter_36: class Program_weight_tensor_parameter_37: name = "parameter_37" + original_name = "batch_norm2d_69.w_0" shape = [96] dtype = "float32" min_val = float("0.821693") @@ -396,6 +434,7 @@ class Program_weight_tensor_parameter_37: class Program_weight_tensor_parameter_38: name = "parameter_38" + original_name = "batch_norm2d_69.w_2" shape = [96] dtype = "float32" min_val = float("4.86209e-05") @@ -407,6 +446,7 @@ class Program_weight_tensor_parameter_38: class Program_weight_tensor_parameter_39: name = "parameter_39" + original_name = "batch_norm2d_69.w_1" shape = [96] dtype = "float32" min_val = float("-0.0273101") @@ -418,6 +458,7 @@ class Program_weight_tensor_parameter_39: class Program_weight_tensor_parameter_40: name = "parameter_40" + original_name = "conv2d_76.w_0" shape = [96, 1, 5, 5] dtype = "float32" min_val = float("-0.286588") @@ -429,6 +470,7 @@ class Program_weight_tensor_parameter_40: class Program_weight_tensor_parameter_41: name = "parameter_41" + original_name = "batch_norm2d_79.b_0" shape = [1] dtype = "float32" min_val = float("0") @@ -438,6 +480,7 @@ class Program_weight_tensor_parameter_41: class Program_weight_tensor_parameter_42: name = "parameter_42" + original_name = "batch_norm2d_79.w_0" shape = [1] dtype = "float32" min_val = float("0") @@ -447,6 +490,7 @@ class Program_weight_tensor_parameter_42: class Program_weight_tensor_parameter_43: name = "parameter_43" + original_name = "batch_norm2d_79.w_2" shape = [1] dtype = "float32" min_val = float("0") @@ -456,6 +500,7 @@ class Program_weight_tensor_parameter_43: class Program_weight_tensor_parameter_44: name = "parameter_44" + original_name = "batch_norm2d_79.w_1" shape = [1] dtype = "float32" min_val = float("0") @@ -465,6 +510,7 @@ class Program_weight_tensor_parameter_44: class Program_weight_tensor_parameter_45: name = "parameter_45" + original_name = "conv2d_95.w_0" shape = [1, 1, 1, 1] dtype = "float32" min_val = float("0") @@ -474,6 +520,7 @@ class Program_weight_tensor_parameter_45: class Program_weight_tensor_parameter_46: name = "parameter_46" + original_name = "batch_norm2d_78.b_0" shape = [1] dtype = "float32" min_val = float("0") @@ -483,6 +530,7 @@ class Program_weight_tensor_parameter_46: class Program_weight_tensor_parameter_47: name = "parameter_47" + original_name = "batch_norm2d_78.w_0" shape = [1] dtype = "float32" min_val = float("0") @@ -492,6 +540,7 @@ class Program_weight_tensor_parameter_47: class Program_weight_tensor_parameter_48: name = "parameter_48" + original_name = "batch_norm2d_78.w_2" shape = [1] dtype = "float32" min_val = float("0") @@ -501,6 +550,7 @@ class Program_weight_tensor_parameter_48: class Program_weight_tensor_parameter_49: name = "parameter_49" + original_name = "batch_norm2d_78.w_1" shape = [1] dtype = "float32" min_val = float("0") @@ -510,6 +560,7 @@ class Program_weight_tensor_parameter_49: class Program_weight_tensor_parameter_50: name = "parameter_50" + original_name = "conv2d_94.w_0" shape = [1, 96, 5, 5] dtype = "float32" min_val = float("-0.738675") @@ -521,6 +572,7 @@ class Program_weight_tensor_parameter_50: class Program_weight_tensor_parameter_51: name = "parameter_51" + original_name = "conv2d_93.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -530,6 +582,7 @@ class Program_weight_tensor_parameter_51: class Program_weight_tensor_parameter_52: name = "parameter_52" + original_name = "conv2d_93.w_0" shape = [32, 96, 1, 1] dtype = "float32" min_val = float("-0.596263") @@ -541,6 +594,7 @@ class Program_weight_tensor_parameter_52: class Program_weight_tensor_parameter_53: name = "parameter_53" + original_name = "conv2d_92.b_0" shape = [4] dtype = "float32" min_val = float("0") @@ -550,6 +604,7 @@ class Program_weight_tensor_parameter_53: class Program_weight_tensor_parameter_54: name = "parameter_54" + original_name = "conv2d_92.w_0" shape = [4, 96, 1, 1] dtype = "float32" min_val = float("-0.0256892") @@ -561,6 +616,7 @@ class Program_weight_tensor_parameter_54: class Program_weight_tensor_parameter_55: name = "parameter_55" + original_name = "batch_norm2d_68.b_0" shape = [96] dtype = "float32" min_val = float("-0.0989288") @@ -572,6 +628,7 @@ class Program_weight_tensor_parameter_55: class Program_weight_tensor_parameter_56: name = "parameter_56" + original_name = "batch_norm2d_68.w_0" shape = [96] dtype = "float32" min_val = float("0.609417") @@ -583,6 +640,7 @@ class Program_weight_tensor_parameter_56: class Program_weight_tensor_parameter_57: name = "parameter_57" + original_name = "batch_norm2d_68.w_2" shape = [96] dtype = "float32" min_val = float("0.00843896") @@ -594,6 +652,7 @@ class Program_weight_tensor_parameter_57: class Program_weight_tensor_parameter_58: name = "parameter_58" + original_name = "batch_norm2d_68.w_1" shape = [96] dtype = "float32" min_val = float("-0.220598") @@ -605,6 +664,7 @@ class Program_weight_tensor_parameter_58: class Program_weight_tensor_parameter_59: name = "parameter_59" + original_name = "conv2d_75.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.432795") @@ -616,6 +676,7 @@ class Program_weight_tensor_parameter_59: class Program_weight_tensor_parameter_60: name = "parameter_60" + original_name = "conv2d_74.b_0" shape = [96] dtype = "float32" min_val = float("-0.0660859") @@ -627,6 +688,7 @@ class Program_weight_tensor_parameter_60: class Program_weight_tensor_parameter_61: name = "parameter_61" + original_name = "conv2d_74.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.190344") @@ -638,6 +700,7 @@ class Program_weight_tensor_parameter_61: class Program_weight_tensor_parameter_62: name = "parameter_62" + original_name = "batch_norm2d_67.b_0" shape = [96] dtype = "float32" min_val = float("-1.4229") @@ -649,6 +712,7 @@ class Program_weight_tensor_parameter_62: class Program_weight_tensor_parameter_63: name = "parameter_63" + original_name = "batch_norm2d_67.w_0" shape = [96] dtype = "float32" min_val = float("0.386881") @@ -660,6 +724,7 @@ class Program_weight_tensor_parameter_63: class Program_weight_tensor_parameter_64: name = "parameter_64" + original_name = "batch_norm2d_67.w_2" shape = [96] dtype = "float32" min_val = float("0.0508518") @@ -671,6 +736,7 @@ class Program_weight_tensor_parameter_64: class Program_weight_tensor_parameter_65: name = "parameter_65" + original_name = "batch_norm2d_67.w_1" shape = [96] dtype = "float32" min_val = float("-0.952262") @@ -682,6 +748,7 @@ class Program_weight_tensor_parameter_65: class Program_weight_tensor_parameter_66: name = "parameter_66" + original_name = "conv2d_73.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.478185") @@ -693,6 +760,7 @@ class Program_weight_tensor_parameter_66: class Program_weight_tensor_parameter_67: name = "parameter_67" + original_name = "batch_norm2d_66.b_0" shape = [96] dtype = "float32" min_val = float("-1.82675") @@ -704,6 +772,7 @@ class Program_weight_tensor_parameter_67: class Program_weight_tensor_parameter_68: name = "parameter_68" + original_name = "batch_norm2d_66.w_0" shape = [96] dtype = "float32" min_val = float("0.471464") @@ -715,6 +784,7 @@ class Program_weight_tensor_parameter_68: class Program_weight_tensor_parameter_69: name = "parameter_69" + original_name = "batch_norm2d_66.w_2" shape = [96] dtype = "float32" min_val = float("0.000139979") @@ -726,6 +796,7 @@ class Program_weight_tensor_parameter_69: class Program_weight_tensor_parameter_70: name = "parameter_70" + original_name = "batch_norm2d_66.w_1" shape = [96] dtype = "float32" min_val = float("-0.41869") @@ -737,6 +808,7 @@ class Program_weight_tensor_parameter_70: class Program_weight_tensor_parameter_71: name = "parameter_71" + original_name = "conv2d_72.w_0" shape = [96, 1, 5, 5] dtype = "float32" min_val = float("-0.353651") @@ -748,6 +820,7 @@ class Program_weight_tensor_parameter_71: class Program_weight_tensor_parameter_72: name = "parameter_72" + original_name = "batch_norm2d_65.b_0" shape = [96] dtype = "float32" min_val = float("-1.2807") @@ -759,6 +832,7 @@ class Program_weight_tensor_parameter_72: class Program_weight_tensor_parameter_73: name = "parameter_73" + original_name = "batch_norm2d_65.w_0" shape = [96] dtype = "float32" min_val = float("0.0654939") @@ -770,6 +844,7 @@ class Program_weight_tensor_parameter_73: class Program_weight_tensor_parameter_74: name = "parameter_74" + original_name = "batch_norm2d_65.w_2" shape = [96] dtype = "float32" min_val = float("0.0298232") @@ -781,6 +856,7 @@ class Program_weight_tensor_parameter_74: class Program_weight_tensor_parameter_75: name = "parameter_75" + original_name = "batch_norm2d_65.w_1" shape = [96] dtype = "float32" min_val = float("-0.773746") @@ -792,6 +868,7 @@ class Program_weight_tensor_parameter_75: class Program_weight_tensor_parameter_76: name = "parameter_76" + original_name = "conv2d_71.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.428566") @@ -803,6 +880,7 @@ class Program_weight_tensor_parameter_76: class Program_weight_tensor_parameter_77: name = "parameter_77" + original_name = "batch_norm2d_64.b_0" shape = [96] dtype = "float32" min_val = float("-1.10005") @@ -814,6 +892,7 @@ class Program_weight_tensor_parameter_77: class Program_weight_tensor_parameter_78: name = "parameter_78" + original_name = "batch_norm2d_64.w_0" shape = [96] dtype = "float32" min_val = float("0.435313") @@ -825,6 +904,7 @@ class Program_weight_tensor_parameter_78: class Program_weight_tensor_parameter_79: name = "parameter_79" + original_name = "batch_norm2d_64.w_2" shape = [96] dtype = "float32" min_val = float("9.27917e-05") @@ -836,6 +916,7 @@ class Program_weight_tensor_parameter_79: class Program_weight_tensor_parameter_80: name = "parameter_80" + original_name = "batch_norm2d_64.w_1" shape = [96] dtype = "float32" min_val = float("-0.609336") @@ -847,6 +928,7 @@ class Program_weight_tensor_parameter_80: class Program_weight_tensor_parameter_81: name = "parameter_81" + original_name = "conv2d_70.w_0" shape = [96, 1, 5, 5] dtype = "float32" min_val = float("-0.297561") @@ -858,6 +940,7 @@ class Program_weight_tensor_parameter_81: class Program_weight_tensor_parameter_82: name = "parameter_82" + original_name = "batch_norm2d_77.b_0" shape = [1] dtype = "float32" min_val = float("0") @@ -867,6 +950,7 @@ class Program_weight_tensor_parameter_82: class Program_weight_tensor_parameter_83: name = "parameter_83" + original_name = "batch_norm2d_77.w_0" shape = [1] dtype = "float32" min_val = float("0") @@ -876,6 +960,7 @@ class Program_weight_tensor_parameter_83: class Program_weight_tensor_parameter_84: name = "parameter_84" + original_name = "batch_norm2d_77.w_2" shape = [1] dtype = "float32" min_val = float("0") @@ -885,6 +970,7 @@ class Program_weight_tensor_parameter_84: class Program_weight_tensor_parameter_85: name = "parameter_85" + original_name = "batch_norm2d_77.w_1" shape = [1] dtype = "float32" min_val = float("0") @@ -894,6 +980,7 @@ class Program_weight_tensor_parameter_85: class Program_weight_tensor_parameter_86: name = "parameter_86" + original_name = "conv2d_91.w_0" shape = [1, 1, 1, 1] dtype = "float32" min_val = float("0") @@ -903,6 +990,7 @@ class Program_weight_tensor_parameter_86: class Program_weight_tensor_parameter_87: name = "parameter_87" + original_name = "batch_norm2d_76.b_0" shape = [1] dtype = "float32" min_val = float("0") @@ -912,6 +1000,7 @@ class Program_weight_tensor_parameter_87: class Program_weight_tensor_parameter_88: name = "parameter_88" + original_name = "batch_norm2d_76.w_0" shape = [1] dtype = "float32" min_val = float("0") @@ -921,6 +1010,7 @@ class Program_weight_tensor_parameter_88: class Program_weight_tensor_parameter_89: name = "parameter_89" + original_name = "batch_norm2d_76.w_2" shape = [1] dtype = "float32" min_val = float("0") @@ -930,6 +1020,7 @@ class Program_weight_tensor_parameter_89: class Program_weight_tensor_parameter_90: name = "parameter_90" + original_name = "batch_norm2d_76.w_1" shape = [1] dtype = "float32" min_val = float("0") @@ -939,6 +1030,7 @@ class Program_weight_tensor_parameter_90: class Program_weight_tensor_parameter_91: name = "parameter_91" + original_name = "conv2d_90.w_0" shape = [1, 96, 5, 5] dtype = "float32" min_val = float("-0.889734") @@ -950,6 +1042,7 @@ class Program_weight_tensor_parameter_91: class Program_weight_tensor_parameter_92: name = "parameter_92" + original_name = "conv2d_89.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -959,6 +1052,7 @@ class Program_weight_tensor_parameter_92: class Program_weight_tensor_parameter_93: name = "parameter_93" + original_name = "conv2d_89.w_0" shape = [32, 96, 1, 1] dtype = "float32" min_val = float("-0.581374") @@ -970,6 +1064,7 @@ class Program_weight_tensor_parameter_93: class Program_weight_tensor_parameter_94: name = "parameter_94" + original_name = "conv2d_88.b_0" shape = [4] dtype = "float32" min_val = float("0") @@ -979,6 +1074,7 @@ class Program_weight_tensor_parameter_94: class Program_weight_tensor_parameter_95: name = "parameter_95" + original_name = "conv2d_88.w_0" shape = [4, 96, 1, 1] dtype = "float32" min_val = float("-0.0246186") @@ -990,6 +1086,7 @@ class Program_weight_tensor_parameter_95: class Program_weight_tensor_parameter_96: name = "parameter_96" + original_name = "batch_norm2d_63.b_0" shape = [96] dtype = "float32" min_val = float("-0.941358") @@ -1001,6 +1098,7 @@ class Program_weight_tensor_parameter_96: class Program_weight_tensor_parameter_97: name = "parameter_97" + original_name = "batch_norm2d_63.w_0" shape = [96] dtype = "float32" min_val = float("0.36645") @@ -1012,6 +1110,7 @@ class Program_weight_tensor_parameter_97: class Program_weight_tensor_parameter_98: name = "parameter_98" + original_name = "batch_norm2d_63.w_2" shape = [96] dtype = "float32" min_val = float("0.0295765") @@ -1023,6 +1122,7 @@ class Program_weight_tensor_parameter_98: class Program_weight_tensor_parameter_99: name = "parameter_99" + original_name = "batch_norm2d_63.w_1" shape = [96] dtype = "float32" min_val = float("-0.615738") @@ -1034,6 +1134,7 @@ class Program_weight_tensor_parameter_99: class Program_weight_tensor_parameter_100: name = "parameter_100" + original_name = "conv2d_69.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.560056") @@ -1045,6 +1146,7 @@ class Program_weight_tensor_parameter_100: class Program_weight_tensor_parameter_101: name = "parameter_101" + original_name = "conv2d_68.b_0" shape = [96] dtype = "float32" min_val = float("-0.0394424") @@ -1056,6 +1158,7 @@ class Program_weight_tensor_parameter_101: class Program_weight_tensor_parameter_102: name = "parameter_102" + original_name = "conv2d_68.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.297412") @@ -1067,6 +1170,7 @@ class Program_weight_tensor_parameter_102: class Program_weight_tensor_parameter_103: name = "parameter_103" + original_name = "batch_norm2d_62.b_0" shape = [96] dtype = "float32" min_val = float("-2.1848") @@ -1078,6 +1182,7 @@ class Program_weight_tensor_parameter_103: class Program_weight_tensor_parameter_104: name = "parameter_104" + original_name = "batch_norm2d_62.w_0" shape = [96] dtype = "float32" min_val = float("-0.783256") @@ -1089,6 +1194,7 @@ class Program_weight_tensor_parameter_104: class Program_weight_tensor_parameter_105: name = "parameter_105" + original_name = "batch_norm2d_62.w_2" shape = [96] dtype = "float32" min_val = float("0.0703577") @@ -1100,6 +1206,7 @@ class Program_weight_tensor_parameter_105: class Program_weight_tensor_parameter_106: name = "parameter_106" + original_name = "batch_norm2d_62.w_1" shape = [96] dtype = "float32" min_val = float("-1.50515") @@ -1111,6 +1218,7 @@ class Program_weight_tensor_parameter_106: class Program_weight_tensor_parameter_107: name = "parameter_107" + original_name = "conv2d_67.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.519282") @@ -1122,6 +1230,7 @@ class Program_weight_tensor_parameter_107: class Program_weight_tensor_parameter_108: name = "parameter_108" + original_name = "batch_norm2d_61.b_0" shape = [96] dtype = "float32" min_val = float("-1.05674") @@ -1133,6 +1242,7 @@ class Program_weight_tensor_parameter_108: class Program_weight_tensor_parameter_109: name = "parameter_109" + original_name = "batch_norm2d_61.w_0" shape = [96] dtype = "float32" min_val = float("0.431292") @@ -1144,6 +1254,7 @@ class Program_weight_tensor_parameter_109: class Program_weight_tensor_parameter_110: name = "parameter_110" + original_name = "batch_norm2d_61.w_2" shape = [96] dtype = "float32" min_val = float("6.95828e-05") @@ -1155,6 +1266,7 @@ class Program_weight_tensor_parameter_110: class Program_weight_tensor_parameter_111: name = "parameter_111" + original_name = "batch_norm2d_61.w_1" shape = [96] dtype = "float32" min_val = float("-0.380892") @@ -1166,6 +1278,7 @@ class Program_weight_tensor_parameter_111: class Program_weight_tensor_parameter_112: name = "parameter_112" + original_name = "conv2d_66.w_0" shape = [96, 1, 5, 5] dtype = "float32" min_val = float("-0.49401") @@ -1177,6 +1290,7 @@ class Program_weight_tensor_parameter_112: class Program_weight_tensor_parameter_113: name = "parameter_113" + original_name = "batch_norm2d_60.b_0" shape = [96] dtype = "float32" min_val = float("-0.918283") @@ -1188,6 +1302,7 @@ class Program_weight_tensor_parameter_113: class Program_weight_tensor_parameter_114: name = "parameter_114" + original_name = "batch_norm2d_60.w_0" shape = [96] dtype = "float32" min_val = float("0.0374061") @@ -1199,6 +1314,7 @@ class Program_weight_tensor_parameter_114: class Program_weight_tensor_parameter_115: name = "parameter_115" + original_name = "batch_norm2d_60.w_2" shape = [96] dtype = "float32" min_val = float("0.0527503") @@ -1210,6 +1326,7 @@ class Program_weight_tensor_parameter_115: class Program_weight_tensor_parameter_116: name = "parameter_116" + original_name = "batch_norm2d_60.w_1" shape = [96] dtype = "float32" min_val = float("-1.11653") @@ -1221,6 +1338,7 @@ class Program_weight_tensor_parameter_116: class Program_weight_tensor_parameter_117: name = "parameter_117" + original_name = "conv2d_65.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.580984") @@ -1232,6 +1350,7 @@ class Program_weight_tensor_parameter_117: class Program_weight_tensor_parameter_118: name = "parameter_118" + original_name = "batch_norm2d_59.b_0" shape = [96] dtype = "float32" min_val = float("-0.738653") @@ -1243,6 +1362,7 @@ class Program_weight_tensor_parameter_118: class Program_weight_tensor_parameter_119: name = "parameter_119" + original_name = "batch_norm2d_59.w_0" shape = [96] dtype = "float32" min_val = float("0.452223") @@ -1254,6 +1374,7 @@ class Program_weight_tensor_parameter_119: class Program_weight_tensor_parameter_120: name = "parameter_120" + original_name = "batch_norm2d_59.w_2" shape = [96] dtype = "float32" min_val = float("5.96621e-05") @@ -1265,6 +1386,7 @@ class Program_weight_tensor_parameter_120: class Program_weight_tensor_parameter_121: name = "parameter_121" + original_name = "batch_norm2d_59.w_1" shape = [96] dtype = "float32" min_val = float("-0.953254") @@ -1276,6 +1398,7 @@ class Program_weight_tensor_parameter_121: class Program_weight_tensor_parameter_122: name = "parameter_122" + original_name = "conv2d_64.w_0" shape = [96, 1, 5, 5] dtype = "float32" min_val = float("-0.38733") @@ -1287,6 +1410,7 @@ class Program_weight_tensor_parameter_122: class Program_weight_tensor_parameter_123: name = "parameter_123" + original_name = "batch_norm2d_75.b_0" shape = [1] dtype = "float32" min_val = float("0") @@ -1296,6 +1420,7 @@ class Program_weight_tensor_parameter_123: class Program_weight_tensor_parameter_124: name = "parameter_124" + original_name = "batch_norm2d_75.w_0" shape = [1] dtype = "float32" min_val = float("0") @@ -1305,6 +1430,7 @@ class Program_weight_tensor_parameter_124: class Program_weight_tensor_parameter_125: name = "parameter_125" + original_name = "batch_norm2d_75.w_2" shape = [1] dtype = "float32" min_val = float("0") @@ -1314,6 +1440,7 @@ class Program_weight_tensor_parameter_125: class Program_weight_tensor_parameter_126: name = "parameter_126" + original_name = "batch_norm2d_75.w_1" shape = [1] dtype = "float32" min_val = float("0") @@ -1323,6 +1450,7 @@ class Program_weight_tensor_parameter_126: class Program_weight_tensor_parameter_127: name = "parameter_127" + original_name = "conv2d_87.w_0" shape = [1, 1, 1, 1] dtype = "float32" min_val = float("0") @@ -1332,6 +1460,7 @@ class Program_weight_tensor_parameter_127: class Program_weight_tensor_parameter_128: name = "parameter_128" + original_name = "batch_norm2d_74.b_0" shape = [1] dtype = "float32" min_val = float("0") @@ -1341,6 +1470,7 @@ class Program_weight_tensor_parameter_128: class Program_weight_tensor_parameter_129: name = "parameter_129" + original_name = "batch_norm2d_74.w_0" shape = [1] dtype = "float32" min_val = float("0") @@ -1350,6 +1480,7 @@ class Program_weight_tensor_parameter_129: class Program_weight_tensor_parameter_130: name = "parameter_130" + original_name = "batch_norm2d_74.w_2" shape = [1] dtype = "float32" min_val = float("0") @@ -1359,6 +1490,7 @@ class Program_weight_tensor_parameter_130: class Program_weight_tensor_parameter_131: name = "parameter_131" + original_name = "batch_norm2d_74.w_1" shape = [1] dtype = "float32" min_val = float("0") @@ -1368,6 +1500,7 @@ class Program_weight_tensor_parameter_131: class Program_weight_tensor_parameter_132: name = "parameter_132" + original_name = "conv2d_86.w_0" shape = [1, 96, 5, 5] dtype = "float32" min_val = float("-0.542044") @@ -1379,6 +1512,7 @@ class Program_weight_tensor_parameter_132: class Program_weight_tensor_parameter_133: name = "parameter_133" + original_name = "conv2d_85.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -1388,6 +1522,7 @@ class Program_weight_tensor_parameter_133: class Program_weight_tensor_parameter_134: name = "parameter_134" + original_name = "conv2d_85.w_0" shape = [32, 96, 1, 1] dtype = "float32" min_val = float("-0.603401") @@ -1399,6 +1534,7 @@ class Program_weight_tensor_parameter_134: class Program_weight_tensor_parameter_135: name = "parameter_135" + original_name = "conv2d_84.b_0" shape = [4] dtype = "float32" min_val = float("0") @@ -1408,6 +1544,7 @@ class Program_weight_tensor_parameter_135: class Program_weight_tensor_parameter_136: name = "parameter_136" + original_name = "conv2d_84.w_0" shape = [4, 96, 1, 1] dtype = "float32" min_val = float("-0.0277417") @@ -1419,6 +1556,7 @@ class Program_weight_tensor_parameter_136: class Program_weight_tensor_parameter_137: name = "parameter_137" + original_name = "batch_norm2d_58.b_0" shape = [96] dtype = "float32" min_val = float("-0.517484") @@ -1430,6 +1568,7 @@ class Program_weight_tensor_parameter_137: class Program_weight_tensor_parameter_138: name = "parameter_138" + original_name = "batch_norm2d_58.w_0" shape = [96] dtype = "float32" min_val = float("0.626411") @@ -1441,6 +1580,7 @@ class Program_weight_tensor_parameter_138: class Program_weight_tensor_parameter_139: name = "parameter_139" + original_name = "batch_norm2d_58.w_2" shape = [96] dtype = "float32" min_val = float("0.0251703") @@ -1452,6 +1592,7 @@ class Program_weight_tensor_parameter_139: class Program_weight_tensor_parameter_140: name = "parameter_140" + original_name = "batch_norm2d_58.w_1" shape = [96] dtype = "float32" min_val = float("-0.893") @@ -1463,6 +1604,7 @@ class Program_weight_tensor_parameter_140: class Program_weight_tensor_parameter_141: name = "parameter_141" + original_name = "conv2d_63.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.777995") @@ -1474,6 +1616,7 @@ class Program_weight_tensor_parameter_141: class Program_weight_tensor_parameter_142: name = "parameter_142" + original_name = "conv2d_62.b_0" shape = [96] dtype = "float32" min_val = float("-0.0296698") @@ -1485,6 +1628,7 @@ class Program_weight_tensor_parameter_142: class Program_weight_tensor_parameter_143: name = "parameter_143" + original_name = "conv2d_62.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.493352") @@ -1496,6 +1640,7 @@ class Program_weight_tensor_parameter_143: class Program_weight_tensor_parameter_144: name = "parameter_144" + original_name = "batch_norm2d_57.b_0" shape = [96] dtype = "float32" min_val = float("-1.82312") @@ -1507,6 +1652,7 @@ class Program_weight_tensor_parameter_144: class Program_weight_tensor_parameter_145: name = "parameter_145" + original_name = "batch_norm2d_57.w_0" shape = [96] dtype = "float32" min_val = float("0.360541") @@ -1518,6 +1664,7 @@ class Program_weight_tensor_parameter_145: class Program_weight_tensor_parameter_146: name = "parameter_146" + original_name = "batch_norm2d_57.w_2" shape = [96] dtype = "float32" min_val = float("0.0472078") @@ -1529,6 +1676,7 @@ class Program_weight_tensor_parameter_146: class Program_weight_tensor_parameter_147: name = "parameter_147" + original_name = "batch_norm2d_57.w_1" shape = [96] dtype = "float32" min_val = float("-2.16335") @@ -1540,6 +1688,7 @@ class Program_weight_tensor_parameter_147: class Program_weight_tensor_parameter_148: name = "parameter_148" + original_name = "conv2d_61.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.553919") @@ -1551,6 +1700,7 @@ class Program_weight_tensor_parameter_148: class Program_weight_tensor_parameter_149: name = "parameter_149" + original_name = "batch_norm2d_56.b_0" shape = [96] dtype = "float32" min_val = float("-0.939077") @@ -1562,6 +1712,7 @@ class Program_weight_tensor_parameter_149: class Program_weight_tensor_parameter_150: name = "parameter_150" + original_name = "batch_norm2d_56.w_0" shape = [96] dtype = "float32" min_val = float("0.266239") @@ -1573,6 +1724,7 @@ class Program_weight_tensor_parameter_150: class Program_weight_tensor_parameter_151: name = "parameter_151" + original_name = "batch_norm2d_56.w_2" shape = [96] dtype = "float32" min_val = float("0.000138683") @@ -1584,6 +1736,7 @@ class Program_weight_tensor_parameter_151: class Program_weight_tensor_parameter_152: name = "parameter_152" + original_name = "batch_norm2d_56.w_1" shape = [96] dtype = "float32" min_val = float("-0.786412") @@ -1595,6 +1748,7 @@ class Program_weight_tensor_parameter_152: class Program_weight_tensor_parameter_153: name = "parameter_153" + original_name = "conv2d_60.w_0" shape = [96, 1, 5, 5] dtype = "float32" min_val = float("-0.562465") @@ -1606,6 +1760,7 @@ class Program_weight_tensor_parameter_153: class Program_weight_tensor_parameter_154: name = "parameter_154" + original_name = "batch_norm2d_55.b_0" shape = [96] dtype = "float32" min_val = float("-1.46572") @@ -1617,6 +1772,7 @@ class Program_weight_tensor_parameter_154: class Program_weight_tensor_parameter_155: name = "parameter_155" + original_name = "batch_norm2d_55.w_0" shape = [96] dtype = "float32" min_val = float("0.0419306") @@ -1628,6 +1784,7 @@ class Program_weight_tensor_parameter_155: class Program_weight_tensor_parameter_156: name = "parameter_156" + original_name = "batch_norm2d_55.w_2" shape = [96] dtype = "float32" min_val = float("0.0648206") @@ -1639,6 +1796,7 @@ class Program_weight_tensor_parameter_156: class Program_weight_tensor_parameter_157: name = "parameter_157" + original_name = "batch_norm2d_55.w_1" shape = [96] dtype = "float32" min_val = float("-1.52351") @@ -1650,6 +1808,7 @@ class Program_weight_tensor_parameter_157: class Program_weight_tensor_parameter_158: name = "parameter_158" + original_name = "conv2d_59.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.554379") @@ -1661,6 +1820,7 @@ class Program_weight_tensor_parameter_158: class Program_weight_tensor_parameter_159: name = "parameter_159" + original_name = "batch_norm2d_54.b_0" shape = [96] dtype = "float32" min_val = float("-1.21131") @@ -1672,6 +1832,7 @@ class Program_weight_tensor_parameter_159: class Program_weight_tensor_parameter_160: name = "parameter_160" + original_name = "batch_norm2d_54.w_0" shape = [96] dtype = "float32" min_val = float("0.391779") @@ -1683,6 +1844,7 @@ class Program_weight_tensor_parameter_160: class Program_weight_tensor_parameter_161: name = "parameter_161" + original_name = "batch_norm2d_54.w_2" shape = [96] dtype = "float32" min_val = float("0.000115675") @@ -1694,6 +1856,7 @@ class Program_weight_tensor_parameter_161: class Program_weight_tensor_parameter_162: name = "parameter_162" + original_name = "batch_norm2d_54.w_1" shape = [96] dtype = "float32" min_val = float("-0.64249") @@ -1705,6 +1868,7 @@ class Program_weight_tensor_parameter_162: class Program_weight_tensor_parameter_163: name = "parameter_163" + original_name = "conv2d_58.w_0" shape = [96, 1, 5, 5] dtype = "float32" min_val = float("-0.432752") @@ -1716,6 +1880,7 @@ class Program_weight_tensor_parameter_163: class Program_weight_tensor_parameter_164: name = "parameter_164" + original_name = "batch_norm2d_33.b_0" shape = [96] dtype = "float32" min_val = float("-0.0701015") @@ -1727,6 +1892,7 @@ class Program_weight_tensor_parameter_164: class Program_weight_tensor_parameter_165: name = "parameter_165" + original_name = "batch_norm2d_33.w_0" shape = [96] dtype = "float32" min_val = float("0.0293757") @@ -1738,6 +1904,7 @@ class Program_weight_tensor_parameter_165: class Program_weight_tensor_parameter_166: name = "parameter_166" + original_name = "batch_norm2d_33.w_2" shape = [96] dtype = "float32" min_val = float("0.000101999") @@ -1749,6 +1916,7 @@ class Program_weight_tensor_parameter_166: class Program_weight_tensor_parameter_167: name = "parameter_167" + original_name = "batch_norm2d_33.w_1" shape = [96] dtype = "float32" min_val = float("-0.00417801") @@ -1760,6 +1928,7 @@ class Program_weight_tensor_parameter_167: class Program_weight_tensor_parameter_168: name = "parameter_168" + original_name = "conv2d_37.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.146404") @@ -1771,6 +1940,7 @@ class Program_weight_tensor_parameter_168: class Program_weight_tensor_parameter_169: name = "parameter_169" + original_name = "batch_norm2d_32.b_0" shape = [96] dtype = "float32" min_val = float("-0.0372937") @@ -1782,6 +1952,7 @@ class Program_weight_tensor_parameter_169: class Program_weight_tensor_parameter_170: name = "parameter_170" + original_name = "batch_norm2d_32.w_0" shape = [96] dtype = "float32" min_val = float("0.0612304") @@ -1793,6 +1964,7 @@ class Program_weight_tensor_parameter_170: class Program_weight_tensor_parameter_171: name = "parameter_171" + original_name = "batch_norm2d_32.w_2" shape = [96] dtype = "float32" min_val = float("7.3868e-05") @@ -1804,6 +1976,7 @@ class Program_weight_tensor_parameter_171: class Program_weight_tensor_parameter_172: name = "parameter_172" + original_name = "batch_norm2d_32.w_1" shape = [96] dtype = "float32" min_val = float("-0.22886") @@ -1815,6 +1988,7 @@ class Program_weight_tensor_parameter_172: class Program_weight_tensor_parameter_173: name = "parameter_173" + original_name = "conv2d_36.w_0" shape = [96, 1, 5, 5] dtype = "float32" min_val = float("-0.127131") @@ -1826,6 +2000,7 @@ class Program_weight_tensor_parameter_173: class Program_weight_tensor_parameter_174: name = "parameter_174" + original_name = "batch_norm2d_31.b_0" shape = [96] dtype = "float32" min_val = float("-0.103095") @@ -1837,6 +2012,7 @@ class Program_weight_tensor_parameter_174: class Program_weight_tensor_parameter_175: name = "parameter_175" + original_name = "batch_norm2d_31.w_0" shape = [96] dtype = "float32" min_val = float("0.00939856") @@ -1848,6 +2024,7 @@ class Program_weight_tensor_parameter_175: class Program_weight_tensor_parameter_176: name = "parameter_176" + original_name = "batch_norm2d_31.w_2" shape = [96] dtype = "float32" min_val = float("2.12615e-05") @@ -1859,6 +2036,7 @@ class Program_weight_tensor_parameter_176: class Program_weight_tensor_parameter_177: name = "parameter_177" + original_name = "batch_norm2d_31.w_1" shape = [96] dtype = "float32" min_val = float("-0.0038288") @@ -1870,6 +2048,7 @@ class Program_weight_tensor_parameter_177: class Program_weight_tensor_parameter_178: name = "parameter_178" + original_name = "conv2d_35.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.205892") @@ -1881,6 +2060,7 @@ class Program_weight_tensor_parameter_178: class Program_weight_tensor_parameter_179: name = "parameter_179" + original_name = "batch_norm2d_30.b_0" shape = [96] dtype = "float32" min_val = float("-0.0245624") @@ -1892,6 +2072,7 @@ class Program_weight_tensor_parameter_179: class Program_weight_tensor_parameter_180: name = "parameter_180" + original_name = "batch_norm2d_30.w_0" shape = [96] dtype = "float32" min_val = float("0.0736913") @@ -1903,6 +2084,7 @@ class Program_weight_tensor_parameter_180: class Program_weight_tensor_parameter_181: name = "parameter_181" + original_name = "batch_norm2d_30.w_2" shape = [96] dtype = "float32" min_val = float("2.69926e-05") @@ -1914,6 +2096,7 @@ class Program_weight_tensor_parameter_181: class Program_weight_tensor_parameter_182: name = "parameter_182" + original_name = "batch_norm2d_30.w_1" shape = [96] dtype = "float32" min_val = float("-0.0172389") @@ -1925,6 +2108,7 @@ class Program_weight_tensor_parameter_182: class Program_weight_tensor_parameter_183: name = "parameter_183" + original_name = "conv2d_34.w_0" shape = [96, 1, 5, 5] dtype = "float32" min_val = float("-0.0847947") @@ -1936,6 +2120,7 @@ class Program_weight_tensor_parameter_183: class Program_weight_tensor_parameter_184: name = "parameter_184" + original_name = "batch_norm2d_53.b_0" shape = [96] dtype = "float32" min_val = float("-0.825351") @@ -1947,6 +2132,7 @@ class Program_weight_tensor_parameter_184: class Program_weight_tensor_parameter_185: name = "parameter_185" + original_name = "batch_norm2d_53.w_0" shape = [96] dtype = "float32" min_val = float("0.0648123") @@ -1958,6 +2144,7 @@ class Program_weight_tensor_parameter_185: class Program_weight_tensor_parameter_186: name = "parameter_186" + original_name = "batch_norm2d_53.w_2" shape = [96] dtype = "float32" min_val = float("0.0752648") @@ -1969,6 +2156,7 @@ class Program_weight_tensor_parameter_186: class Program_weight_tensor_parameter_187: name = "parameter_187" + original_name = "batch_norm2d_53.w_1" shape = [96] dtype = "float32" min_val = float("-0.570523") @@ -1980,6 +2168,7 @@ class Program_weight_tensor_parameter_187: class Program_weight_tensor_parameter_188: name = "parameter_188" + original_name = "conv2d_57.w_0" shape = [96, 192, 1, 1] dtype = "float32" min_val = float("-0.370531") @@ -1991,6 +2180,7 @@ class Program_weight_tensor_parameter_188: class Program_weight_tensor_parameter_189: name = "parameter_189" + original_name = "batch_norm2d_52.b_0" shape = [192] dtype = "float32" min_val = float("-0.617094") @@ -2002,6 +2192,7 @@ class Program_weight_tensor_parameter_189: class Program_weight_tensor_parameter_190: name = "parameter_190" + original_name = "batch_norm2d_52.w_0" shape = [192] dtype = "float32" min_val = float("0.584169") @@ -2013,6 +2204,7 @@ class Program_weight_tensor_parameter_190: class Program_weight_tensor_parameter_191: name = "parameter_191" + original_name = "batch_norm2d_52.w_2" shape = [192] dtype = "float32" min_val = float("0.000442546") @@ -2024,6 +2216,7 @@ class Program_weight_tensor_parameter_191: class Program_weight_tensor_parameter_192: name = "parameter_192" + original_name = "batch_norm2d_52.w_1" shape = [192] dtype = "float32" min_val = float("-0.269408") @@ -2035,6 +2228,7 @@ class Program_weight_tensor_parameter_192: class Program_weight_tensor_parameter_193: name = "parameter_193" + original_name = "conv2d_56.w_0" shape = [192, 1, 5, 5] dtype = "float32" min_val = float("-0.217672") @@ -2046,6 +2240,7 @@ class Program_weight_tensor_parameter_193: class Program_weight_tensor_parameter_194: name = "parameter_194" + original_name = "batch_norm2d_51.b_0" shape = [192] dtype = "float32" min_val = float("-1.13531") @@ -2057,6 +2252,7 @@ class Program_weight_tensor_parameter_194: class Program_weight_tensor_parameter_195: name = "parameter_195" + original_name = "batch_norm2d_51.w_0" shape = [192] dtype = "float32" min_val = float("0.109496") @@ -2068,6 +2264,7 @@ class Program_weight_tensor_parameter_195: class Program_weight_tensor_parameter_196: name = "parameter_196" + original_name = "batch_norm2d_51.w_2" shape = [192] dtype = "float32" min_val = float("0.0710206") @@ -2079,6 +2276,7 @@ class Program_weight_tensor_parameter_196: class Program_weight_tensor_parameter_197: name = "parameter_197" + original_name = "batch_norm2d_51.w_1" shape = [192] dtype = "float32" min_val = float("-0.748504") @@ -2090,6 +2288,7 @@ class Program_weight_tensor_parameter_197: class Program_weight_tensor_parameter_198: name = "parameter_198" + original_name = "conv2d_55.w_0" shape = [192, 192, 1, 1] dtype = "float32" min_val = float("-0.339804") @@ -2101,6 +2300,7 @@ class Program_weight_tensor_parameter_198: class Program_weight_tensor_parameter_199: name = "parameter_199" + original_name = "batch_norm2d_50.b_0" shape = [192] dtype = "float32" min_val = float("-0.799298") @@ -2112,6 +2312,7 @@ class Program_weight_tensor_parameter_199: class Program_weight_tensor_parameter_200: name = "parameter_200" + original_name = "batch_norm2d_50.w_0" shape = [192] dtype = "float32" min_val = float("0.576681") @@ -2123,6 +2324,7 @@ class Program_weight_tensor_parameter_200: class Program_weight_tensor_parameter_201: name = "parameter_201" + original_name = "batch_norm2d_50.w_2" shape = [192] dtype = "float32" min_val = float("4.39845e-05") @@ -2134,6 +2336,7 @@ class Program_weight_tensor_parameter_201: class Program_weight_tensor_parameter_202: name = "parameter_202" + original_name = "batch_norm2d_50.w_1" shape = [192] dtype = "float32" min_val = float("-0.137086") @@ -2145,6 +2348,7 @@ class Program_weight_tensor_parameter_202: class Program_weight_tensor_parameter_203: name = "parameter_203" + original_name = "conv2d_54.w_0" shape = [192, 1, 5, 5] dtype = "float32" min_val = float("-0.333959") @@ -2156,6 +2360,7 @@ class Program_weight_tensor_parameter_203: class Program_weight_tensor_parameter_204: name = "parameter_204" + original_name = "batch_norm2d_49.b_0" shape = [96] dtype = "float32" min_val = float("-0.271305") @@ -2167,6 +2372,7 @@ class Program_weight_tensor_parameter_204: class Program_weight_tensor_parameter_205: name = "parameter_205" + original_name = "batch_norm2d_49.w_0" shape = [96] dtype = "float32" min_val = float("0.0421347") @@ -2178,6 +2384,7 @@ class Program_weight_tensor_parameter_205: class Program_weight_tensor_parameter_206: name = "parameter_206" + original_name = "batch_norm2d_49.w_2" shape = [96] dtype = "float32" min_val = float("0.000664014") @@ -2189,6 +2396,7 @@ class Program_weight_tensor_parameter_206: class Program_weight_tensor_parameter_207: name = "parameter_207" + original_name = "batch_norm2d_49.w_1" shape = [96] dtype = "float32" min_val = float("-0.0215895") @@ -2200,6 +2408,7 @@ class Program_weight_tensor_parameter_207: class Program_weight_tensor_parameter_208: name = "parameter_208" + original_name = "conv2d_53.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.333792") @@ -2211,6 +2420,7 @@ class Program_weight_tensor_parameter_208: class Program_weight_tensor_parameter_209: name = "parameter_209" + original_name = "batch_norm2d_48.b_0" shape = [96] dtype = "float32" min_val = float("-0.294127") @@ -2222,6 +2432,7 @@ class Program_weight_tensor_parameter_209: class Program_weight_tensor_parameter_210: name = "parameter_210" + original_name = "batch_norm2d_48.w_0" shape = [96] dtype = "float32" min_val = float("0.0711855") @@ -2233,6 +2444,7 @@ class Program_weight_tensor_parameter_210: class Program_weight_tensor_parameter_211: name = "parameter_211" + original_name = "batch_norm2d_48.w_2" shape = [96] dtype = "float32" min_val = float("2.8924e-05") @@ -2244,6 +2456,7 @@ class Program_weight_tensor_parameter_211: class Program_weight_tensor_parameter_212: name = "parameter_212" + original_name = "batch_norm2d_48.w_1" shape = [96] dtype = "float32" min_val = float("-0.43146") @@ -2255,6 +2468,7 @@ class Program_weight_tensor_parameter_212: class Program_weight_tensor_parameter_213: name = "parameter_213" + original_name = "conv2d_52.w_0" shape = [96, 1, 5, 5] dtype = "float32" min_val = float("-0.228678") @@ -2266,6 +2480,7 @@ class Program_weight_tensor_parameter_213: class Program_weight_tensor_parameter_214: name = "parameter_214" + original_name = "batch_norm2d_47.b_0" shape = [96] dtype = "float32" min_val = float("-1.37664") @@ -2277,6 +2492,7 @@ class Program_weight_tensor_parameter_214: class Program_weight_tensor_parameter_215: name = "parameter_215" + original_name = "batch_norm2d_47.w_0" shape = [96] dtype = "float32" min_val = float("0.0399133") @@ -2288,6 +2504,7 @@ class Program_weight_tensor_parameter_215: class Program_weight_tensor_parameter_216: name = "parameter_216" + original_name = "batch_norm2d_47.w_2" shape = [96] dtype = "float32" min_val = float("0.13754") @@ -2299,6 +2516,7 @@ class Program_weight_tensor_parameter_216: class Program_weight_tensor_parameter_217: name = "parameter_217" + original_name = "batch_norm2d_47.w_1" shape = [96] dtype = "float32" min_val = float("-0.665858") @@ -2310,6 +2528,7 @@ class Program_weight_tensor_parameter_217: class Program_weight_tensor_parameter_218: name = "parameter_218" + original_name = "conv2d_51.w_0" shape = [96, 192, 1, 1] dtype = "float32" min_val = float("-0.513346") @@ -2321,6 +2540,7 @@ class Program_weight_tensor_parameter_218: class Program_weight_tensor_parameter_219: name = "parameter_219" + original_name = "batch_norm2d_46.b_0" shape = [192] dtype = "float32" min_val = float("-0.951624") @@ -2332,6 +2552,7 @@ class Program_weight_tensor_parameter_219: class Program_weight_tensor_parameter_220: name = "parameter_220" + original_name = "batch_norm2d_46.w_0" shape = [192] dtype = "float32" min_val = float("0.520269") @@ -2343,6 +2564,7 @@ class Program_weight_tensor_parameter_220: class Program_weight_tensor_parameter_221: name = "parameter_221" + original_name = "batch_norm2d_46.w_2" shape = [192] dtype = "float32" min_val = float("0.000116139") @@ -2354,6 +2576,7 @@ class Program_weight_tensor_parameter_221: class Program_weight_tensor_parameter_222: name = "parameter_222" + original_name = "batch_norm2d_46.w_1" shape = [192] dtype = "float32" min_val = float("-0.518874") @@ -2365,6 +2588,7 @@ class Program_weight_tensor_parameter_222: class Program_weight_tensor_parameter_223: name = "parameter_223" + original_name = "conv2d_50.w_0" shape = [192, 1, 5, 5] dtype = "float32" min_val = float("-0.27626") @@ -2376,6 +2600,7 @@ class Program_weight_tensor_parameter_223: class Program_weight_tensor_parameter_224: name = "parameter_224" + original_name = "batch_norm2d_45.b_0" shape = [192] dtype = "float32" min_val = float("-1.23519") @@ -2387,6 +2612,7 @@ class Program_weight_tensor_parameter_224: class Program_weight_tensor_parameter_225: name = "parameter_225" + original_name = "batch_norm2d_45.w_0" shape = [192] dtype = "float32" min_val = float("0.0513988") @@ -2398,6 +2624,7 @@ class Program_weight_tensor_parameter_225: class Program_weight_tensor_parameter_226: name = "parameter_226" + original_name = "batch_norm2d_45.w_2" shape = [192] dtype = "float32" min_val = float("0.0929235") @@ -2409,6 +2636,7 @@ class Program_weight_tensor_parameter_226: class Program_weight_tensor_parameter_227: name = "parameter_227" + original_name = "batch_norm2d_45.w_1" shape = [192] dtype = "float32" min_val = float("-1.04517") @@ -2420,6 +2648,7 @@ class Program_weight_tensor_parameter_227: class Program_weight_tensor_parameter_228: name = "parameter_228" + original_name = "conv2d_49.w_0" shape = [192, 192, 1, 1] dtype = "float32" min_val = float("-0.4355") @@ -2431,6 +2660,7 @@ class Program_weight_tensor_parameter_228: class Program_weight_tensor_parameter_229: name = "parameter_229" + original_name = "batch_norm2d_44.b_0" shape = [192] dtype = "float32" min_val = float("-1.01181") @@ -2442,6 +2672,7 @@ class Program_weight_tensor_parameter_229: class Program_weight_tensor_parameter_230: name = "parameter_230" + original_name = "batch_norm2d_44.w_0" shape = [192] dtype = "float32" min_val = float("0.593978") @@ -2453,6 +2684,7 @@ class Program_weight_tensor_parameter_230: class Program_weight_tensor_parameter_231: name = "parameter_231" + original_name = "batch_norm2d_44.w_2" shape = [192] dtype = "float32" min_val = float("5.98893e-05") @@ -2464,6 +2696,7 @@ class Program_weight_tensor_parameter_231: class Program_weight_tensor_parameter_232: name = "parameter_232" + original_name = "batch_norm2d_44.w_1" shape = [192] dtype = "float32" min_val = float("-0.453752") @@ -2475,6 +2708,7 @@ class Program_weight_tensor_parameter_232: class Program_weight_tensor_parameter_233: name = "parameter_233" + original_name = "conv2d_48.w_0" shape = [192, 1, 5, 5] dtype = "float32" min_val = float("-0.399178") @@ -2486,6 +2720,7 @@ class Program_weight_tensor_parameter_233: class Program_weight_tensor_parameter_234: name = "parameter_234" + original_name = "batch_norm2d_43.b_0" shape = [96] dtype = "float32" min_val = float("-0.185986") @@ -2497,6 +2732,7 @@ class Program_weight_tensor_parameter_234: class Program_weight_tensor_parameter_235: name = "parameter_235" + original_name = "batch_norm2d_43.w_0" shape = [96] dtype = "float32" min_val = float("0.0379374") @@ -2508,6 +2744,7 @@ class Program_weight_tensor_parameter_235: class Program_weight_tensor_parameter_236: name = "parameter_236" + original_name = "batch_norm2d_43.w_2" shape = [96] dtype = "float32" min_val = float("0.00100173") @@ -2519,6 +2756,7 @@ class Program_weight_tensor_parameter_236: class Program_weight_tensor_parameter_237: name = "parameter_237" + original_name = "batch_norm2d_43.w_1" shape = [96] dtype = "float32" min_val = float("-0.0604006") @@ -2530,6 +2768,7 @@ class Program_weight_tensor_parameter_237: class Program_weight_tensor_parameter_238: name = "parameter_238" + original_name = "conv2d_47.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.417191") @@ -2541,6 +2780,7 @@ class Program_weight_tensor_parameter_238: class Program_weight_tensor_parameter_239: name = "parameter_239" + original_name = "batch_norm2d_42.b_0" shape = [96] dtype = "float32" min_val = float("-0.368555") @@ -2552,6 +2792,7 @@ class Program_weight_tensor_parameter_239: class Program_weight_tensor_parameter_240: name = "parameter_240" + original_name = "batch_norm2d_42.w_0" shape = [96] dtype = "float32" min_val = float("0.0912847") @@ -2563,6 +2804,7 @@ class Program_weight_tensor_parameter_240: class Program_weight_tensor_parameter_241: name = "parameter_241" + original_name = "batch_norm2d_42.w_2" shape = [96] dtype = "float32" min_val = float("2.40738e-05") @@ -2574,6 +2816,7 @@ class Program_weight_tensor_parameter_241: class Program_weight_tensor_parameter_242: name = "parameter_242" + original_name = "batch_norm2d_42.w_1" shape = [96] dtype = "float32" min_val = float("-0.413149") @@ -2585,6 +2828,7 @@ class Program_weight_tensor_parameter_242: class Program_weight_tensor_parameter_243: name = "parameter_243" + original_name = "conv2d_46.w_0" shape = [96, 1, 5, 5] dtype = "float32" min_val = float("-0.280859") @@ -2596,6 +2840,7 @@ class Program_weight_tensor_parameter_243: class Program_weight_tensor_parameter_244: name = "parameter_244" + original_name = "batch_norm2d_41.b_0" shape = [96] dtype = "float32" min_val = float("-1.88173") @@ -2607,6 +2852,7 @@ class Program_weight_tensor_parameter_244: class Program_weight_tensor_parameter_245: name = "parameter_245" + original_name = "batch_norm2d_41.w_0" shape = [96] dtype = "float32" min_val = float("0.0413252") @@ -2618,6 +2864,7 @@ class Program_weight_tensor_parameter_245: class Program_weight_tensor_parameter_246: name = "parameter_246" + original_name = "batch_norm2d_41.w_2" shape = [96] dtype = "float32" min_val = float("0.107094") @@ -2629,6 +2876,7 @@ class Program_weight_tensor_parameter_246: class Program_weight_tensor_parameter_247: name = "parameter_247" + original_name = "batch_norm2d_41.w_1" shape = [96] dtype = "float32" min_val = float("-1.32941") @@ -2640,6 +2888,7 @@ class Program_weight_tensor_parameter_247: class Program_weight_tensor_parameter_248: name = "parameter_248" + original_name = "conv2d_45.w_0" shape = [96, 192, 1, 1] dtype = "float32" min_val = float("-0.488235") @@ -2651,6 +2900,7 @@ class Program_weight_tensor_parameter_248: class Program_weight_tensor_parameter_249: name = "parameter_249" + original_name = "batch_norm2d_40.b_0" shape = [192] dtype = "float32" min_val = float("-1.5577") @@ -2662,6 +2912,7 @@ class Program_weight_tensor_parameter_249: class Program_weight_tensor_parameter_250: name = "parameter_250" + original_name = "batch_norm2d_40.w_0" shape = [192] dtype = "float32" min_val = float("0.400465") @@ -2673,6 +2924,7 @@ class Program_weight_tensor_parameter_250: class Program_weight_tensor_parameter_251: name = "parameter_251" + original_name = "batch_norm2d_40.w_2" shape = [192] dtype = "float32" min_val = float("0.00022517") @@ -2684,6 +2936,7 @@ class Program_weight_tensor_parameter_251: class Program_weight_tensor_parameter_252: name = "parameter_252" + original_name = "batch_norm2d_40.w_1" shape = [192] dtype = "float32" min_val = float("-0.725367") @@ -2695,6 +2948,7 @@ class Program_weight_tensor_parameter_252: class Program_weight_tensor_parameter_253: name = "parameter_253" + original_name = "conv2d_44.w_0" shape = [192, 1, 5, 5] dtype = "float32" min_val = float("-0.56515") @@ -2706,6 +2960,7 @@ class Program_weight_tensor_parameter_253: class Program_weight_tensor_parameter_254: name = "parameter_254" + original_name = "batch_norm2d_39.b_0" shape = [192] dtype = "float32" min_val = float("-2.28197") @@ -2717,6 +2972,7 @@ class Program_weight_tensor_parameter_254: class Program_weight_tensor_parameter_255: name = "parameter_255" + original_name = "batch_norm2d_39.w_0" shape = [192] dtype = "float32" min_val = float("0.0694161") @@ -2728,6 +2984,7 @@ class Program_weight_tensor_parameter_255: class Program_weight_tensor_parameter_256: name = "parameter_256" + original_name = "batch_norm2d_39.w_2" shape = [192] dtype = "float32" min_val = float("0.184104") @@ -2739,6 +2996,7 @@ class Program_weight_tensor_parameter_256: class Program_weight_tensor_parameter_257: name = "parameter_257" + original_name = "batch_norm2d_39.w_1" shape = [192] dtype = "float32" min_val = float("-2.12706") @@ -2750,6 +3008,7 @@ class Program_weight_tensor_parameter_257: class Program_weight_tensor_parameter_258: name = "parameter_258" + original_name = "conv2d_43.w_0" shape = [192, 192, 1, 1] dtype = "float32" min_val = float("-0.436888") @@ -2761,6 +3020,7 @@ class Program_weight_tensor_parameter_258: class Program_weight_tensor_parameter_259: name = "parameter_259" + original_name = "batch_norm2d_38.b_0" shape = [192] dtype = "float32" min_val = float("-0.844411") @@ -2772,6 +3032,7 @@ class Program_weight_tensor_parameter_259: class Program_weight_tensor_parameter_260: name = "parameter_260" + original_name = "batch_norm2d_38.w_0" shape = [192] dtype = "float32" min_val = float("0.46783") @@ -2783,6 +3044,7 @@ class Program_weight_tensor_parameter_260: class Program_weight_tensor_parameter_261: name = "parameter_261" + original_name = "batch_norm2d_38.w_2" shape = [192] dtype = "float32" min_val = float("5.86023e-05") @@ -2794,6 +3056,7 @@ class Program_weight_tensor_parameter_261: class Program_weight_tensor_parameter_262: name = "parameter_262" + original_name = "batch_norm2d_38.w_1" shape = [192] dtype = "float32" min_val = float("-1.01767") @@ -2805,6 +3068,7 @@ class Program_weight_tensor_parameter_262: class Program_weight_tensor_parameter_263: name = "parameter_263" + original_name = "conv2d_42.w_0" shape = [192, 1, 5, 5] dtype = "float32" min_val = float("-0.617472") @@ -2816,6 +3080,7 @@ class Program_weight_tensor_parameter_263: class Program_weight_tensor_parameter_264: name = "parameter_264" + original_name = "batch_norm2d_37.b_0" shape = [96] dtype = "float32" min_val = float("-0.818492") @@ -2827,6 +3092,7 @@ class Program_weight_tensor_parameter_264: class Program_weight_tensor_parameter_265: name = "parameter_265" + original_name = "batch_norm2d_37.w_0" shape = [96] dtype = "float32" min_val = float("0.0519763") @@ -2838,6 +3104,7 @@ class Program_weight_tensor_parameter_265: class Program_weight_tensor_parameter_266: name = "parameter_266" + original_name = "batch_norm2d_37.w_2" shape = [96] dtype = "float32" min_val = float("0.29829") @@ -2849,6 +3116,7 @@ class Program_weight_tensor_parameter_266: class Program_weight_tensor_parameter_267: name = "parameter_267" + original_name = "batch_norm2d_37.w_1" shape = [96] dtype = "float32" min_val = float("-1.14147") @@ -2860,6 +3128,7 @@ class Program_weight_tensor_parameter_267: class Program_weight_tensor_parameter_268: name = "parameter_268" + original_name = "conv2d_41.w_0" shape = [96, 192, 1, 1] dtype = "float32" min_val = float("-0.639691") @@ -2871,6 +3140,7 @@ class Program_weight_tensor_parameter_268: class Program_weight_tensor_parameter_269: name = "parameter_269" + original_name = "batch_norm2d_36.b_0" shape = [192] dtype = "float32" min_val = float("-1.0254") @@ -2882,6 +3152,7 @@ class Program_weight_tensor_parameter_269: class Program_weight_tensor_parameter_270: name = "parameter_270" + original_name = "batch_norm2d_36.w_0" shape = [192] dtype = "float32" min_val = float("0.569151") @@ -2893,6 +3164,7 @@ class Program_weight_tensor_parameter_270: class Program_weight_tensor_parameter_271: name = "parameter_271" + original_name = "batch_norm2d_36.w_2" shape = [192] dtype = "float32" min_val = float("0.000211807") @@ -2904,6 +3176,7 @@ class Program_weight_tensor_parameter_271: class Program_weight_tensor_parameter_272: name = "parameter_272" + original_name = "batch_norm2d_36.w_1" shape = [192] dtype = "float32" min_val = float("-0.499987") @@ -2915,6 +3188,7 @@ class Program_weight_tensor_parameter_272: class Program_weight_tensor_parameter_273: name = "parameter_273" + original_name = "conv2d_40.w_0" shape = [192, 1, 5, 5] dtype = "float32" min_val = float("-0.391759") @@ -2926,6 +3200,7 @@ class Program_weight_tensor_parameter_273: class Program_weight_tensor_parameter_274: name = "parameter_274" + original_name = "batch_norm2d_35.b_0" shape = [192] dtype = "float32" min_val = float("-1.18885") @@ -2937,6 +3212,7 @@ class Program_weight_tensor_parameter_274: class Program_weight_tensor_parameter_275: name = "parameter_275" + original_name = "batch_norm2d_35.w_0" shape = [192] dtype = "float32" min_val = float("0.0734486") @@ -2948,6 +3224,7 @@ class Program_weight_tensor_parameter_275: class Program_weight_tensor_parameter_276: name = "parameter_276" + original_name = "batch_norm2d_35.w_2" shape = [192] dtype = "float32" min_val = float("0.245635") @@ -2959,6 +3236,7 @@ class Program_weight_tensor_parameter_276: class Program_weight_tensor_parameter_277: name = "parameter_277" + original_name = "batch_norm2d_35.w_1" shape = [192] dtype = "float32" min_val = float("-2.02518") @@ -2970,6 +3248,7 @@ class Program_weight_tensor_parameter_277: class Program_weight_tensor_parameter_278: name = "parameter_278" + original_name = "conv2d_39.w_0" shape = [192, 192, 1, 1] dtype = "float32" min_val = float("-0.371777") @@ -2981,6 +3260,7 @@ class Program_weight_tensor_parameter_278: class Program_weight_tensor_parameter_279: name = "parameter_279" + original_name = "batch_norm2d_34.b_0" shape = [192] dtype = "float32" min_val = float("-1.24809") @@ -2992,6 +3272,7 @@ class Program_weight_tensor_parameter_279: class Program_weight_tensor_parameter_280: name = "parameter_280" + original_name = "batch_norm2d_34.w_0" shape = [192] dtype = "float32" min_val = float("0.797257") @@ -3003,6 +3284,7 @@ class Program_weight_tensor_parameter_280: class Program_weight_tensor_parameter_281: name = "parameter_281" + original_name = "batch_norm2d_34.w_2" shape = [192] dtype = "float32" min_val = float("4.37763e-05") @@ -3014,6 +3296,7 @@ class Program_weight_tensor_parameter_281: class Program_weight_tensor_parameter_282: name = "parameter_282" + original_name = "batch_norm2d_34.w_1" shape = [192] dtype = "float32" min_val = float("-0.0569166") @@ -3025,6 +3308,7 @@ class Program_weight_tensor_parameter_282: class Program_weight_tensor_parameter_283: name = "parameter_283" + original_name = "conv2d_38.w_0" shape = [192, 1, 5, 5] dtype = "float32" min_val = float("-0.451088") @@ -3036,6 +3320,7 @@ class Program_weight_tensor_parameter_283: class Program_weight_tensor_parameter_284: name = "parameter_284" + original_name = "batch_norm2d_29.b_0" shape = [96] dtype = "float32" min_val = float("-0.258102") @@ -3047,6 +3332,7 @@ class Program_weight_tensor_parameter_284: class Program_weight_tensor_parameter_285: name = "parameter_285" + original_name = "batch_norm2d_29.w_0" shape = [96] dtype = "float32" min_val = float("0.0573719") @@ -3058,6 +3344,7 @@ class Program_weight_tensor_parameter_285: class Program_weight_tensor_parameter_286: name = "parameter_286" + original_name = "batch_norm2d_29.w_2" shape = [96] dtype = "float32" min_val = float("1.64779") @@ -3069,6 +3356,7 @@ class Program_weight_tensor_parameter_286: class Program_weight_tensor_parameter_287: name = "parameter_287" + original_name = "batch_norm2d_29.w_1" shape = [96] dtype = "float32" min_val = float("-2.08205") @@ -3080,6 +3368,7 @@ class Program_weight_tensor_parameter_287: class Program_weight_tensor_parameter_288: name = "parameter_288" + original_name = "conv2d_33.w_0" shape = [96, 384, 1, 1] dtype = "float32" min_val = float("-0.27613") @@ -3091,6 +3380,7 @@ class Program_weight_tensor_parameter_288: class Program_weight_tensor_parameter_289: name = "parameter_289" + original_name = "batch_norm2d_28.b_0" shape = [96] dtype = "float32" min_val = float("-0.109487") @@ -3102,6 +3392,7 @@ class Program_weight_tensor_parameter_289: class Program_weight_tensor_parameter_290: name = "parameter_290" + original_name = "batch_norm2d_28.w_0" shape = [96] dtype = "float32" min_val = float("0.0418336") @@ -3113,6 +3404,7 @@ class Program_weight_tensor_parameter_290: class Program_weight_tensor_parameter_291: name = "parameter_291" + original_name = "batch_norm2d_28.w_2" shape = [96] dtype = "float32" min_val = float("0.0508983") @@ -3124,6 +3416,7 @@ class Program_weight_tensor_parameter_291: class Program_weight_tensor_parameter_292: name = "parameter_292" + original_name = "batch_norm2d_28.w_1" shape = [96] dtype = "float32" min_val = float("-0.834536") @@ -3135,6 +3428,7 @@ class Program_weight_tensor_parameter_292: class Program_weight_tensor_parameter_293: name = "parameter_293" + original_name = "conv2d_32.w_0" shape = [96, 192, 1, 1] dtype = "float32" min_val = float("-0.196545") @@ -3146,6 +3440,7 @@ class Program_weight_tensor_parameter_293: class Program_weight_tensor_parameter_294: name = "parameter_294" + original_name = "batch_norm2d_27.b_0" shape = [96] dtype = "float32" min_val = float("-0.328893") @@ -3157,6 +3452,7 @@ class Program_weight_tensor_parameter_294: class Program_weight_tensor_parameter_295: name = "parameter_295" + original_name = "batch_norm2d_27.w_0" shape = [96] dtype = "float32" min_val = float("0.032189") @@ -3168,6 +3464,7 @@ class Program_weight_tensor_parameter_295: class Program_weight_tensor_parameter_296: name = "parameter_296" + original_name = "batch_norm2d_27.w_2" shape = [96] dtype = "float32" min_val = float("0.0106813") @@ -3179,6 +3476,7 @@ class Program_weight_tensor_parameter_296: class Program_weight_tensor_parameter_297: name = "parameter_297" + original_name = "batch_norm2d_27.w_1" shape = [96] dtype = "float32" min_val = float("-0.884092") @@ -3190,6 +3488,7 @@ class Program_weight_tensor_parameter_297: class Program_weight_tensor_parameter_298: name = "parameter_298" + original_name = "conv2d_31.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.434595") @@ -3201,6 +3500,7 @@ class Program_weight_tensor_parameter_298: class Program_weight_tensor_parameter_299: name = "parameter_299" + original_name = "batch_norm2d_26.b_0" shape = [384] dtype = "float32" min_val = float("-6.11028") @@ -3212,6 +3512,7 @@ class Program_weight_tensor_parameter_299: class Program_weight_tensor_parameter_300: name = "parameter_300" + original_name = "batch_norm2d_26.w_0" shape = [384] dtype = "float32" min_val = float("2.67797") @@ -3223,6 +3524,7 @@ class Program_weight_tensor_parameter_300: class Program_weight_tensor_parameter_301: name = "parameter_301" + original_name = "batch_norm2d_26.w_2" shape = [384] dtype = "float32" min_val = float("0.0266098") @@ -3234,6 +3536,7 @@ class Program_weight_tensor_parameter_301: class Program_weight_tensor_parameter_302: name = "parameter_302" + original_name = "batch_norm2d_26.w_1" shape = [384] dtype = "float32" min_val = float("-0.496267") @@ -3245,6 +3548,7 @@ class Program_weight_tensor_parameter_302: class Program_weight_tensor_parameter_303: name = "parameter_303" + original_name = "conv2d_30.w_0" shape = [384, 384, 1, 1] dtype = "float32" min_val = float("-0.285842") @@ -3256,6 +3560,7 @@ class Program_weight_tensor_parameter_303: class Program_weight_tensor_parameter_304: name = "parameter_304" + original_name = "conv2d_29.b_0" shape = [384] dtype = "float32" min_val = float("-0.117827") @@ -3267,6 +3572,7 @@ class Program_weight_tensor_parameter_304: class Program_weight_tensor_parameter_305: name = "parameter_305" + original_name = "conv2d_29.w_0" shape = [384, 96, 1, 1] dtype = "float32" min_val = float("-0.428589") @@ -3278,6 +3584,7 @@ class Program_weight_tensor_parameter_305: class Program_weight_tensor_parameter_306: name = "parameter_306" + original_name = "conv2d_28.b_0" shape = [96] dtype = "float32" min_val = float("-0.0236603") @@ -3289,6 +3596,7 @@ class Program_weight_tensor_parameter_306: class Program_weight_tensor_parameter_307: name = "parameter_307" + original_name = "conv2d_28.w_0" shape = [96, 384, 1, 1] dtype = "float32" min_val = float("-0.33906") @@ -3300,6 +3608,7 @@ class Program_weight_tensor_parameter_307: class Program_weight_tensor_parameter_308: name = "parameter_308" + original_name = "batch_norm2d_25.b_0" shape = [384] dtype = "float32" min_val = float("-2.61319") @@ -3311,6 +3620,7 @@ class Program_weight_tensor_parameter_308: class Program_weight_tensor_parameter_309: name = "parameter_309" + original_name = "batch_norm2d_25.w_0" shape = [384] dtype = "float32" min_val = float("0.287978") @@ -3322,6 +3632,7 @@ class Program_weight_tensor_parameter_309: class Program_weight_tensor_parameter_310: name = "parameter_310" + original_name = "batch_norm2d_25.w_2" shape = [384] dtype = "float32" min_val = float("9.44173e-05") @@ -3333,6 +3644,7 @@ class Program_weight_tensor_parameter_310: class Program_weight_tensor_parameter_311: name = "parameter_311" + original_name = "batch_norm2d_25.w_1" shape = [384] dtype = "float32" min_val = float("-2.40261") @@ -3344,6 +3656,7 @@ class Program_weight_tensor_parameter_311: class Program_weight_tensor_parameter_312: name = "parameter_312" + original_name = "conv2d_27.w_0" shape = [384, 1, 5, 5] dtype = "float32" min_val = float("-0.513294") @@ -3355,6 +3668,7 @@ class Program_weight_tensor_parameter_312: class Program_weight_tensor_parameter_313: name = "parameter_313" + original_name = "batch_norm2d_24.b_0" shape = [384] dtype = "float32" min_val = float("-2.56751") @@ -3366,6 +3680,7 @@ class Program_weight_tensor_parameter_313: class Program_weight_tensor_parameter_314: name = "parameter_314" + original_name = "batch_norm2d_24.w_0" shape = [384] dtype = "float32" min_val = float("-2.95142") @@ -3377,6 +3692,7 @@ class Program_weight_tensor_parameter_314: class Program_weight_tensor_parameter_315: name = "parameter_315" + original_name = "batch_norm2d_24.w_2" shape = [384] dtype = "float32" min_val = float("0.0348909") @@ -3388,6 +3704,7 @@ class Program_weight_tensor_parameter_315: class Program_weight_tensor_parameter_316: name = "parameter_316" + original_name = "batch_norm2d_24.w_1" shape = [384] dtype = "float32" min_val = float("-0.914385") @@ -3399,6 +3716,7 @@ class Program_weight_tensor_parameter_316: class Program_weight_tensor_parameter_317: name = "parameter_317" + original_name = "conv2d_26.w_0" shape = [384, 192, 1, 1] dtype = "float32" min_val = float("-0.339063") @@ -3410,6 +3728,7 @@ class Program_weight_tensor_parameter_317: class Program_weight_tensor_parameter_318: name = "parameter_318" + original_name = "conv2d_25.b_0" shape = [192] dtype = "float32" min_val = float("-0.132946") @@ -3421,6 +3740,7 @@ class Program_weight_tensor_parameter_318: class Program_weight_tensor_parameter_319: name = "parameter_319" + original_name = "conv2d_25.w_0" shape = [192, 48, 1, 1] dtype = "float32" min_val = float("-0.526327") @@ -3432,6 +3752,7 @@ class Program_weight_tensor_parameter_319: class Program_weight_tensor_parameter_320: name = "parameter_320" + original_name = "conv2d_24.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -3441,6 +3762,7 @@ class Program_weight_tensor_parameter_320: class Program_weight_tensor_parameter_321: name = "parameter_321" + original_name = "conv2d_24.w_0" shape = [48, 192, 1, 1] dtype = "float32" min_val = float("-0.656176") @@ -3452,6 +3774,7 @@ class Program_weight_tensor_parameter_321: class Program_weight_tensor_parameter_322: name = "parameter_322" + original_name = "batch_norm2d_23.b_0" shape = [192] dtype = "float32" min_val = float("-1.41873") @@ -3463,6 +3786,7 @@ class Program_weight_tensor_parameter_322: class Program_weight_tensor_parameter_323: name = "parameter_323" + original_name = "batch_norm2d_23.w_0" shape = [192] dtype = "float32" min_val = float("0.528064") @@ -3474,6 +3798,7 @@ class Program_weight_tensor_parameter_323: class Program_weight_tensor_parameter_324: name = "parameter_324" + original_name = "batch_norm2d_23.w_2" shape = [192] dtype = "float32" min_val = float("0.000308839") @@ -3485,6 +3810,7 @@ class Program_weight_tensor_parameter_324: class Program_weight_tensor_parameter_325: name = "parameter_325" + original_name = "batch_norm2d_23.w_1" shape = [192] dtype = "float32" min_val = float("-2.85341") @@ -3496,6 +3822,7 @@ class Program_weight_tensor_parameter_325: class Program_weight_tensor_parameter_326: name = "parameter_326" + original_name = "conv2d_23.w_0" shape = [192, 1, 5, 5] dtype = "float32" min_val = float("-0.268057") @@ -3507,6 +3834,7 @@ class Program_weight_tensor_parameter_326: class Program_weight_tensor_parameter_327: name = "parameter_327" + original_name = "batch_norm2d_22.b_0" shape = [192] dtype = "float32" min_val = float("-2.45743") @@ -3518,6 +3846,7 @@ class Program_weight_tensor_parameter_327: class Program_weight_tensor_parameter_328: name = "parameter_328" + original_name = "batch_norm2d_22.w_0" shape = [192] dtype = "float32" min_val = float("0.0544244") @@ -3529,6 +3858,7 @@ class Program_weight_tensor_parameter_328: class Program_weight_tensor_parameter_329: name = "parameter_329" + original_name = "batch_norm2d_22.w_2" shape = [192] dtype = "float32" min_val = float("0.278671") @@ -3540,6 +3870,7 @@ class Program_weight_tensor_parameter_329: class Program_weight_tensor_parameter_330: name = "parameter_330" + original_name = "batch_norm2d_22.w_1" shape = [192] dtype = "float32" min_val = float("-3.26784") @@ -3551,6 +3882,7 @@ class Program_weight_tensor_parameter_330: class Program_weight_tensor_parameter_331: name = "parameter_331" + original_name = "conv2d_22.w_0" shape = [192, 192, 1, 1] dtype = "float32" min_val = float("-0.994642") @@ -3562,6 +3894,7 @@ class Program_weight_tensor_parameter_331: class Program_weight_tensor_parameter_332: name = "parameter_332" + original_name = "batch_norm2d_21.b_0" shape = [192] dtype = "float32" min_val = float("-1.65501") @@ -3573,6 +3906,7 @@ class Program_weight_tensor_parameter_332: class Program_weight_tensor_parameter_333: name = "parameter_333" + original_name = "batch_norm2d_21.w_0" shape = [192] dtype = "float32" min_val = float("0.40918") @@ -3584,6 +3918,7 @@ class Program_weight_tensor_parameter_333: class Program_weight_tensor_parameter_334: name = "parameter_334" + original_name = "batch_norm2d_21.w_2" shape = [192] dtype = "float32" min_val = float("0.000161923") @@ -3595,6 +3930,7 @@ class Program_weight_tensor_parameter_334: class Program_weight_tensor_parameter_335: name = "parameter_335" + original_name = "batch_norm2d_21.w_1" shape = [192] dtype = "float32" min_val = float("-2.22993") @@ -3606,6 +3942,7 @@ class Program_weight_tensor_parameter_335: class Program_weight_tensor_parameter_336: name = "parameter_336" + original_name = "conv2d_21.w_0" shape = [192, 1, 5, 5] dtype = "float32" min_val = float("-0.806665") @@ -3617,6 +3954,7 @@ class Program_weight_tensor_parameter_336: class Program_weight_tensor_parameter_337: name = "parameter_337" + original_name = "batch_norm2d_20.b_0" shape = [192] dtype = "float32" min_val = float("-3.95421") @@ -3628,6 +3966,7 @@ class Program_weight_tensor_parameter_337: class Program_weight_tensor_parameter_338: name = "parameter_338" + original_name = "batch_norm2d_20.w_0" shape = [192] dtype = "float32" min_val = float("-0.869917") @@ -3639,6 +3978,7 @@ class Program_weight_tensor_parameter_338: class Program_weight_tensor_parameter_339: name = "parameter_339" + original_name = "batch_norm2d_20.w_2" shape = [192] dtype = "float32" min_val = float("0.328032") @@ -3650,6 +3990,7 @@ class Program_weight_tensor_parameter_339: class Program_weight_tensor_parameter_340: name = "parameter_340" + original_name = "batch_norm2d_20.w_1" shape = [192] dtype = "float32" min_val = float("-4.11241") @@ -3661,6 +4002,7 @@ class Program_weight_tensor_parameter_340: class Program_weight_tensor_parameter_341: name = "parameter_341" + original_name = "conv2d_20.w_0" shape = [192, 192, 1, 1] dtype = "float32" min_val = float("-0.493359") @@ -3672,6 +4014,7 @@ class Program_weight_tensor_parameter_341: class Program_weight_tensor_parameter_342: name = "parameter_342" + original_name = "batch_norm2d_19.b_0" shape = [192] dtype = "float32" min_val = float("-3.73864") @@ -3683,6 +4026,7 @@ class Program_weight_tensor_parameter_342: class Program_weight_tensor_parameter_343: name = "parameter_343" + original_name = "batch_norm2d_19.w_0" shape = [192] dtype = "float32" min_val = float("0.456652") @@ -3694,6 +4038,7 @@ class Program_weight_tensor_parameter_343: class Program_weight_tensor_parameter_344: name = "parameter_344" + original_name = "batch_norm2d_19.w_2" shape = [192] dtype = "float32" min_val = float("0.000142939") @@ -3705,6 +4050,7 @@ class Program_weight_tensor_parameter_344: class Program_weight_tensor_parameter_345: name = "parameter_345" + original_name = "batch_norm2d_19.w_1" shape = [192] dtype = "float32" min_val = float("-7.14772") @@ -3716,6 +4062,7 @@ class Program_weight_tensor_parameter_345: class Program_weight_tensor_parameter_346: name = "parameter_346" + original_name = "conv2d_19.w_0" shape = [192, 1, 5, 5] dtype = "float32" min_val = float("-0.649713") @@ -3727,6 +4074,7 @@ class Program_weight_tensor_parameter_346: class Program_weight_tensor_parameter_347: name = "parameter_347" + original_name = "batch_norm2d_18.b_0" shape = [192] dtype = "float32" min_val = float("-2.31635") @@ -3738,6 +4086,7 @@ class Program_weight_tensor_parameter_347: class Program_weight_tensor_parameter_348: name = "parameter_348" + original_name = "batch_norm2d_18.w_0" shape = [192] dtype = "float32" min_val = float("-0.767129") @@ -3749,6 +4098,7 @@ class Program_weight_tensor_parameter_348: class Program_weight_tensor_parameter_349: name = "parameter_349" + original_name = "batch_norm2d_18.w_2" shape = [192] dtype = "float32" min_val = float("0.268817") @@ -3760,6 +4110,7 @@ class Program_weight_tensor_parameter_349: class Program_weight_tensor_parameter_350: name = "parameter_350" + original_name = "batch_norm2d_18.w_1" shape = [192] dtype = "float32" min_val = float("-4.86791") @@ -3771,6 +4122,7 @@ class Program_weight_tensor_parameter_350: class Program_weight_tensor_parameter_351: name = "parameter_351" + original_name = "conv2d_18.w_0" shape = [192, 192, 1, 1] dtype = "float32" min_val = float("-0.452247") @@ -3782,6 +4134,7 @@ class Program_weight_tensor_parameter_351: class Program_weight_tensor_parameter_352: name = "parameter_352" + original_name = "batch_norm2d_17.b_0" shape = [192] dtype = "float32" min_val = float("-2.75591") @@ -3793,6 +4146,7 @@ class Program_weight_tensor_parameter_352: class Program_weight_tensor_parameter_353: name = "parameter_353" + original_name = "batch_norm2d_17.w_0" shape = [192] dtype = "float32" min_val = float("0.477652") @@ -3804,6 +4158,7 @@ class Program_weight_tensor_parameter_353: class Program_weight_tensor_parameter_354: name = "parameter_354" + original_name = "batch_norm2d_17.w_2" shape = [192] dtype = "float32" min_val = float("0.000129095") @@ -3815,6 +4170,7 @@ class Program_weight_tensor_parameter_354: class Program_weight_tensor_parameter_355: name = "parameter_355" + original_name = "batch_norm2d_17.w_1" shape = [192] dtype = "float32" min_val = float("-3.78359") @@ -3826,6 +4182,7 @@ class Program_weight_tensor_parameter_355: class Program_weight_tensor_parameter_356: name = "parameter_356" + original_name = "conv2d_17.w_0" shape = [192, 1, 5, 5] dtype = "float32" min_val = float("-0.779566") @@ -3837,6 +4194,7 @@ class Program_weight_tensor_parameter_356: class Program_weight_tensor_parameter_357: name = "parameter_357" + original_name = "batch_norm2d_16.b_0" shape = [192] dtype = "float32" min_val = float("-2.8044") @@ -3848,6 +4206,7 @@ class Program_weight_tensor_parameter_357: class Program_weight_tensor_parameter_358: name = "parameter_358" + original_name = "batch_norm2d_16.w_0" shape = [192] dtype = "float32" min_val = float("-0.837985") @@ -3859,6 +4218,7 @@ class Program_weight_tensor_parameter_358: class Program_weight_tensor_parameter_359: name = "parameter_359" + original_name = "batch_norm2d_16.w_2" shape = [192] dtype = "float32" min_val = float("0.282522") @@ -3870,6 +4230,7 @@ class Program_weight_tensor_parameter_359: class Program_weight_tensor_parameter_360: name = "parameter_360" + original_name = "batch_norm2d_16.w_1" shape = [192] dtype = "float32" min_val = float("-5.90192") @@ -3881,6 +4242,7 @@ class Program_weight_tensor_parameter_360: class Program_weight_tensor_parameter_361: name = "parameter_361" + original_name = "conv2d_16.w_0" shape = [192, 192, 1, 1] dtype = "float32" min_val = float("-0.48512") @@ -3892,6 +4254,7 @@ class Program_weight_tensor_parameter_361: class Program_weight_tensor_parameter_362: name = "parameter_362" + original_name = "batch_norm2d_15.b_0" shape = [192] dtype = "float32" min_val = float("-1.64006") @@ -3903,6 +4266,7 @@ class Program_weight_tensor_parameter_362: class Program_weight_tensor_parameter_363: name = "parameter_363" + original_name = "batch_norm2d_15.w_0" shape = [192] dtype = "float32" min_val = float("0.451191") @@ -3914,6 +4278,7 @@ class Program_weight_tensor_parameter_363: class Program_weight_tensor_parameter_364: name = "parameter_364" + original_name = "batch_norm2d_15.w_2" shape = [192] dtype = "float32" min_val = float("0.000169649") @@ -3925,6 +4290,7 @@ class Program_weight_tensor_parameter_364: class Program_weight_tensor_parameter_365: name = "parameter_365" + original_name = "batch_norm2d_15.w_1" shape = [192] dtype = "float32" min_val = float("-2.95443") @@ -3936,6 +4302,7 @@ class Program_weight_tensor_parameter_365: class Program_weight_tensor_parameter_366: name = "parameter_366" + original_name = "conv2d_15.w_0" shape = [192, 1, 5, 5] dtype = "float32" min_val = float("-0.815335") @@ -3947,6 +4314,7 @@ class Program_weight_tensor_parameter_366: class Program_weight_tensor_parameter_367: name = "parameter_367" + original_name = "batch_norm2d_14.b_0" shape = [192] dtype = "float32" min_val = float("-1.76007") @@ -3958,6 +4326,7 @@ class Program_weight_tensor_parameter_367: class Program_weight_tensor_parameter_368: name = "parameter_368" + original_name = "batch_norm2d_14.w_0" shape = [192] dtype = "float32" min_val = float("-1.20686") @@ -3969,6 +4338,7 @@ class Program_weight_tensor_parameter_368: class Program_weight_tensor_parameter_369: name = "parameter_369" + original_name = "batch_norm2d_14.w_2" shape = [192] dtype = "float32" min_val = float("0.400428") @@ -3980,6 +4350,7 @@ class Program_weight_tensor_parameter_369: class Program_weight_tensor_parameter_370: name = "parameter_370" + original_name = "batch_norm2d_14.w_1" shape = [192] dtype = "float32" min_val = float("-5.09375") @@ -3991,6 +4362,7 @@ class Program_weight_tensor_parameter_370: class Program_weight_tensor_parameter_371: name = "parameter_371" + original_name = "conv2d_14.w_0" shape = [192, 192, 1, 1] dtype = "float32" min_val = float("-0.802088") @@ -4002,6 +4374,7 @@ class Program_weight_tensor_parameter_371: class Program_weight_tensor_parameter_372: name = "parameter_372" + original_name = "batch_norm2d_13.b_0" shape = [192] dtype = "float32" min_val = float("-3.91483") @@ -4013,6 +4386,7 @@ class Program_weight_tensor_parameter_372: class Program_weight_tensor_parameter_373: name = "parameter_373" + original_name = "batch_norm2d_13.w_0" shape = [192] dtype = "float32" min_val = float("0.456819") @@ -4024,6 +4398,7 @@ class Program_weight_tensor_parameter_373: class Program_weight_tensor_parameter_374: name = "parameter_374" + original_name = "batch_norm2d_13.w_2" shape = [192] dtype = "float32" min_val = float("0.000109378") @@ -4035,6 +4410,7 @@ class Program_weight_tensor_parameter_374: class Program_weight_tensor_parameter_375: name = "parameter_375" + original_name = "batch_norm2d_13.w_1" shape = [192] dtype = "float32" min_val = float("-4.82902") @@ -4046,6 +4422,7 @@ class Program_weight_tensor_parameter_375: class Program_weight_tensor_parameter_376: name = "parameter_376" + original_name = "conv2d_13.w_0" shape = [192, 1, 5, 5] dtype = "float32" min_val = float("-0.821296") @@ -4057,6 +4434,7 @@ class Program_weight_tensor_parameter_376: class Program_weight_tensor_parameter_377: name = "parameter_377" + original_name = "batch_norm2d_12.b_0" shape = [192] dtype = "float32" min_val = float("-1.5284") @@ -4068,6 +4446,7 @@ class Program_weight_tensor_parameter_377: class Program_weight_tensor_parameter_378: name = "parameter_378" + original_name = "batch_norm2d_12.w_0" shape = [192] dtype = "float32" min_val = float("-0.422661") @@ -4079,6 +4458,7 @@ class Program_weight_tensor_parameter_378: class Program_weight_tensor_parameter_379: name = "parameter_379" + original_name = "batch_norm2d_12.w_2" shape = [192] dtype = "float32" min_val = float("0.0950696") @@ -4090,6 +4470,7 @@ class Program_weight_tensor_parameter_379: class Program_weight_tensor_parameter_380: name = "parameter_380" + original_name = "batch_norm2d_12.w_1" shape = [192] dtype = "float32" min_val = float("-5.90959") @@ -4101,6 +4482,7 @@ class Program_weight_tensor_parameter_380: class Program_weight_tensor_parameter_381: name = "parameter_381" + original_name = "conv2d_12.w_0" shape = [192, 96, 1, 1] dtype = "float32" min_val = float("-0.660693") @@ -4112,6 +4494,7 @@ class Program_weight_tensor_parameter_381: class Program_weight_tensor_parameter_382: name = "parameter_382" + original_name = "batch_norm2d_11.b_0" shape = [96] dtype = "float32" min_val = float("-1.08052") @@ -4123,6 +4506,7 @@ class Program_weight_tensor_parameter_382: class Program_weight_tensor_parameter_383: name = "parameter_383" + original_name = "batch_norm2d_11.w_0" shape = [96] dtype = "float32" min_val = float("0.352514") @@ -4134,6 +4518,7 @@ class Program_weight_tensor_parameter_383: class Program_weight_tensor_parameter_384: name = "parameter_384" + original_name = "batch_norm2d_11.w_2" shape = [96] dtype = "float32" min_val = float("0.00156682") @@ -4145,6 +4530,7 @@ class Program_weight_tensor_parameter_384: class Program_weight_tensor_parameter_385: name = "parameter_385" + original_name = "batch_norm2d_11.w_1" shape = [96] dtype = "float32" min_val = float("-2.72779") @@ -4156,6 +4542,7 @@ class Program_weight_tensor_parameter_385: class Program_weight_tensor_parameter_386: name = "parameter_386" + original_name = "conv2d_11.w_0" shape = [96, 1, 3, 3] dtype = "float32" min_val = float("-0.299602") @@ -4167,6 +4554,7 @@ class Program_weight_tensor_parameter_386: class Program_weight_tensor_parameter_387: name = "parameter_387" + original_name = "batch_norm2d_10.b_0" shape = [96] dtype = "float32" min_val = float("-1.63831") @@ -4178,6 +4566,7 @@ class Program_weight_tensor_parameter_387: class Program_weight_tensor_parameter_388: name = "parameter_388" + original_name = "batch_norm2d_10.w_0" shape = [96] dtype = "float32" min_val = float("0.237994") @@ -4189,6 +4578,7 @@ class Program_weight_tensor_parameter_388: class Program_weight_tensor_parameter_389: name = "parameter_389" + original_name = "batch_norm2d_10.w_2" shape = [96] dtype = "float32" min_val = float("0.643119") @@ -4200,6 +4590,7 @@ class Program_weight_tensor_parameter_389: class Program_weight_tensor_parameter_390: name = "parameter_390" + original_name = "batch_norm2d_10.w_1" shape = [96] dtype = "float32" min_val = float("-10.337") @@ -4211,6 +4602,7 @@ class Program_weight_tensor_parameter_390: class Program_weight_tensor_parameter_391: name = "parameter_391" + original_name = "conv2d_10.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.962915") @@ -4222,6 +4614,7 @@ class Program_weight_tensor_parameter_391: class Program_weight_tensor_parameter_392: name = "parameter_392" + original_name = "batch_norm2d_9.b_0" shape = [96] dtype = "float32" min_val = float("-1.19898") @@ -4233,6 +4626,7 @@ class Program_weight_tensor_parameter_392: class Program_weight_tensor_parameter_393: name = "parameter_393" + original_name = "batch_norm2d_9.w_0" shape = [96] dtype = "float32" min_val = float("0.549857") @@ -4244,6 +4638,7 @@ class Program_weight_tensor_parameter_393: class Program_weight_tensor_parameter_394: name = "parameter_394" + original_name = "batch_norm2d_9.w_2" shape = [96] dtype = "float32" min_val = float("9.31117e-05") @@ -4255,6 +4650,7 @@ class Program_weight_tensor_parameter_394: class Program_weight_tensor_parameter_395: name = "parameter_395" + original_name = "batch_norm2d_9.w_1" shape = [96] dtype = "float32" min_val = float("-8.79184") @@ -4266,6 +4662,7 @@ class Program_weight_tensor_parameter_395: class Program_weight_tensor_parameter_396: name = "parameter_396" + original_name = "conv2d_9.w_0" shape = [96, 1, 3, 3] dtype = "float32" min_val = float("-1.10347") @@ -4277,6 +4674,7 @@ class Program_weight_tensor_parameter_396: class Program_weight_tensor_parameter_397: name = "parameter_397" + original_name = "batch_norm2d_8.b_0" shape = [96] dtype = "float32" min_val = float("-1.31776") @@ -4288,6 +4686,7 @@ class Program_weight_tensor_parameter_397: class Program_weight_tensor_parameter_398: name = "parameter_398" + original_name = "batch_norm2d_8.w_0" shape = [96] dtype = "float32" min_val = float("-0.365664") @@ -4299,6 +4698,7 @@ class Program_weight_tensor_parameter_398: class Program_weight_tensor_parameter_399: name = "parameter_399" + original_name = "batch_norm2d_8.w_2" shape = [96] dtype = "float32" min_val = float("0.144575") @@ -4310,6 +4710,7 @@ class Program_weight_tensor_parameter_399: class Program_weight_tensor_parameter_400: name = "parameter_400" + original_name = "batch_norm2d_8.w_1" shape = [96] dtype = "float32" min_val = float("-9.12511") @@ -4321,6 +4722,7 @@ class Program_weight_tensor_parameter_400: class Program_weight_tensor_parameter_401: name = "parameter_401" + original_name = "conv2d_8.w_0" shape = [96, 48, 1, 1] dtype = "float32" min_val = float("-1.11604") @@ -4332,6 +4734,7 @@ class Program_weight_tensor_parameter_401: class Program_weight_tensor_parameter_402: name = "parameter_402" + original_name = "batch_norm2d_7.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -4341,6 +4744,7 @@ class Program_weight_tensor_parameter_402: class Program_weight_tensor_parameter_403: name = "parameter_403" + original_name = "batch_norm2d_7.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -4350,6 +4754,7 @@ class Program_weight_tensor_parameter_403: class Program_weight_tensor_parameter_404: name = "parameter_404" + original_name = "batch_norm2d_7.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -4359,6 +4764,7 @@ class Program_weight_tensor_parameter_404: class Program_weight_tensor_parameter_405: name = "parameter_405" + original_name = "batch_norm2d_7.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -4368,6 +4774,7 @@ class Program_weight_tensor_parameter_405: class Program_weight_tensor_parameter_406: name = "parameter_406" + original_name = "conv2d_7.w_0" shape = [48, 1, 3, 3] dtype = "float32" min_val = float("-0.529469") @@ -4379,6 +4786,7 @@ class Program_weight_tensor_parameter_406: class Program_weight_tensor_parameter_407: name = "parameter_407" + original_name = "batch_norm2d_6.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -4388,6 +4796,7 @@ class Program_weight_tensor_parameter_407: class Program_weight_tensor_parameter_408: name = "parameter_408" + original_name = "batch_norm2d_6.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -4397,6 +4806,7 @@ class Program_weight_tensor_parameter_408: class Program_weight_tensor_parameter_409: name = "parameter_409" + original_name = "batch_norm2d_6.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -4406,6 +4816,7 @@ class Program_weight_tensor_parameter_409: class Program_weight_tensor_parameter_410: name = "parameter_410" + original_name = "batch_norm2d_6.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -4415,6 +4826,7 @@ class Program_weight_tensor_parameter_410: class Program_weight_tensor_parameter_411: name = "parameter_411" + original_name = "conv2d_6.w_0" shape = [48, 48, 1, 1] dtype = "float32" min_val = float("-1.20776") @@ -4426,6 +4838,7 @@ class Program_weight_tensor_parameter_411: class Program_weight_tensor_parameter_412: name = "parameter_412" + original_name = "batch_norm2d_5.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -4435,6 +4848,7 @@ class Program_weight_tensor_parameter_412: class Program_weight_tensor_parameter_413: name = "parameter_413" + original_name = "batch_norm2d_5.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -4444,6 +4858,7 @@ class Program_weight_tensor_parameter_413: class Program_weight_tensor_parameter_414: name = "parameter_414" + original_name = "batch_norm2d_5.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -4453,6 +4868,7 @@ class Program_weight_tensor_parameter_414: class Program_weight_tensor_parameter_415: name = "parameter_415" + original_name = "batch_norm2d_5.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -4462,6 +4878,7 @@ class Program_weight_tensor_parameter_415: class Program_weight_tensor_parameter_416: name = "parameter_416" + original_name = "conv2d_5.w_0" shape = [48, 1, 3, 3] dtype = "float32" min_val = float("-1.50645") @@ -4473,6 +4890,7 @@ class Program_weight_tensor_parameter_416: class Program_weight_tensor_parameter_417: name = "parameter_417" + original_name = "batch_norm2d_4.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -4482,6 +4900,7 @@ class Program_weight_tensor_parameter_417: class Program_weight_tensor_parameter_418: name = "parameter_418" + original_name = "batch_norm2d_4.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -4491,6 +4910,7 @@ class Program_weight_tensor_parameter_418: class Program_weight_tensor_parameter_419: name = "parameter_419" + original_name = "batch_norm2d_4.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -4500,6 +4920,7 @@ class Program_weight_tensor_parameter_419: class Program_weight_tensor_parameter_420: name = "parameter_420" + original_name = "batch_norm2d_4.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -4509,6 +4930,7 @@ class Program_weight_tensor_parameter_420: class Program_weight_tensor_parameter_421: name = "parameter_421" + original_name = "conv2d_4.w_0" shape = [48, 24, 1, 1] dtype = "float32" min_val = float("-1.37888") @@ -4520,6 +4942,7 @@ class Program_weight_tensor_parameter_421: class Program_weight_tensor_parameter_422: name = "parameter_422" + original_name = "batch_norm2d_3.b_0" shape = [24] dtype = "float32" min_val = float("0") @@ -4529,6 +4952,7 @@ class Program_weight_tensor_parameter_422: class Program_weight_tensor_parameter_423: name = "parameter_423" + original_name = "batch_norm2d_3.w_0" shape = [24] dtype = "float32" min_val = float("0") @@ -4538,6 +4962,7 @@ class Program_weight_tensor_parameter_423: class Program_weight_tensor_parameter_424: name = "parameter_424" + original_name = "batch_norm2d_3.w_2" shape = [24] dtype = "float32" min_val = float("0") @@ -4547,6 +4972,7 @@ class Program_weight_tensor_parameter_424: class Program_weight_tensor_parameter_425: name = "parameter_425" + original_name = "batch_norm2d_3.w_1" shape = [24] dtype = "float32" min_val = float("0") @@ -4556,6 +4982,7 @@ class Program_weight_tensor_parameter_425: class Program_weight_tensor_parameter_426: name = "parameter_426" + original_name = "conv2d_3.w_0" shape = [24, 1, 3, 3] dtype = "float32" min_val = float("-0.801316") @@ -4567,6 +4994,7 @@ class Program_weight_tensor_parameter_426: class Program_weight_tensor_parameter_427: name = "parameter_427" + original_name = "batch_norm2d_2.b_0" shape = [24] dtype = "float32" min_val = float("0") @@ -4576,6 +5004,7 @@ class Program_weight_tensor_parameter_427: class Program_weight_tensor_parameter_428: name = "parameter_428" + original_name = "batch_norm2d_2.w_0" shape = [24] dtype = "float32" min_val = float("0") @@ -4585,6 +5014,7 @@ class Program_weight_tensor_parameter_428: class Program_weight_tensor_parameter_429: name = "parameter_429" + original_name = "batch_norm2d_2.w_2" shape = [24] dtype = "float32" min_val = float("0") @@ -4594,6 +5024,7 @@ class Program_weight_tensor_parameter_429: class Program_weight_tensor_parameter_430: name = "parameter_430" + original_name = "batch_norm2d_2.w_1" shape = [24] dtype = "float32" min_val = float("0") @@ -4603,6 +5034,7 @@ class Program_weight_tensor_parameter_430: class Program_weight_tensor_parameter_431: name = "parameter_431" + original_name = "conv2d_2.w_0" shape = [24, 16, 1, 1] dtype = "float32" min_val = float("-1.47479") @@ -4614,6 +5046,7 @@ class Program_weight_tensor_parameter_431: class Program_weight_tensor_parameter_432: name = "parameter_432" + original_name = "batch_norm2d_1.b_0" shape = [16] dtype = "float32" min_val = float("0") @@ -4623,6 +5056,7 @@ class Program_weight_tensor_parameter_432: class Program_weight_tensor_parameter_433: name = "parameter_433" + original_name = "batch_norm2d_1.w_0" shape = [16] dtype = "float32" min_val = float("0") @@ -4632,6 +5066,7 @@ class Program_weight_tensor_parameter_433: class Program_weight_tensor_parameter_434: name = "parameter_434" + original_name = "batch_norm2d_1.w_2" shape = [16] dtype = "float32" min_val = float("0") @@ -4641,6 +5076,7 @@ class Program_weight_tensor_parameter_434: class Program_weight_tensor_parameter_435: name = "parameter_435" + original_name = "batch_norm2d_1.w_1" shape = [16] dtype = "float32" min_val = float("0") @@ -4650,6 +5086,7 @@ class Program_weight_tensor_parameter_435: class Program_weight_tensor_parameter_436: name = "parameter_436" + original_name = "conv2d_1.w_0" shape = [16, 1, 3, 3] dtype = "float32" min_val = float("-1.83202") @@ -4661,6 +5098,7 @@ class Program_weight_tensor_parameter_436: class Program_weight_tensor_parameter_437: name = "parameter_437" + original_name = "batch_norm2d_0.b_0" shape = [16] dtype = "float32" min_val = float("0") @@ -4670,6 +5108,7 @@ class Program_weight_tensor_parameter_437: class Program_weight_tensor_parameter_438: name = "parameter_438" + original_name = "batch_norm2d_0.w_0" shape = [16] dtype = "float32" min_val = float("0") @@ -4679,6 +5118,7 @@ class Program_weight_tensor_parameter_438: class Program_weight_tensor_parameter_439: name = "parameter_439" + original_name = "batch_norm2d_0.w_2" shape = [16] dtype = "float32" min_val = float("0") @@ -4688,6 +5128,7 @@ class Program_weight_tensor_parameter_439: class Program_weight_tensor_parameter_440: name = "parameter_440" + original_name = "batch_norm2d_0.w_1" shape = [16] dtype = "float32" min_val = float("0") @@ -4697,6 +5138,7 @@ class Program_weight_tensor_parameter_440: class Program_weight_tensor_parameter_441: name = "parameter_441" + original_name = "conv2d_0.w_0" shape = [16, 3, 3, 3] dtype = "float32" min_val = float("-1.19271") diff --git a/paddle_samples/PaddleX/PicoDet-XS/subgraph_4/input_meta.py b/paddle_samples/PaddleX/PicoDet-XS/subgraph_4/input_meta.py index 0bc98c18a..1c881b243 100644 --- a/paddle_samples/PaddleX/PicoDet-XS/subgraph_4/input_meta.py +++ b/paddle_samples/PaddleX/PicoDet-XS/subgraph_4/input_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_data_0: name = "data_0" + original_name = "var_610" shape = [4, 3, 480, 480] dtype = "float32" min_val = float("-2.1179") @@ -11,6 +12,7 @@ class Program_weight_tensor_data_0: class Program_weight_tensor_data_1: name = "data_1" + original_name = "var_847" shape = [8] dtype = "float32" data = [0.0, 1.0, 2.0, 3.0003, 4.0, 5.0, 6.0006, 7.0] diff --git a/paddle_samples/PaddleX/PicoDet-XS/subgraph_4/weight_meta.py b/paddle_samples/PaddleX/PicoDet-XS/subgraph_4/weight_meta.py index 554e46541..61956318e 100644 --- a/paddle_samples/PaddleX/PicoDet-XS/subgraph_4/weight_meta.py +++ b/paddle_samples/PaddleX/PicoDet-XS/subgraph_4/weight_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_parameter_0: name = "parameter_0" + original_name = "batch_norm2d_81.b_0" shape = [1] dtype = "float32" min_val = float("0") @@ -9,6 +10,7 @@ class Program_weight_tensor_parameter_0: class Program_weight_tensor_parameter_1: name = "parameter_1" + original_name = "batch_norm2d_81.w_0" shape = [1] dtype = "float32" min_val = float("0") @@ -18,6 +20,7 @@ class Program_weight_tensor_parameter_1: class Program_weight_tensor_parameter_2: name = "parameter_2" + original_name = "batch_norm2d_81.w_2" shape = [1] dtype = "float32" min_val = float("0") @@ -27,6 +30,7 @@ class Program_weight_tensor_parameter_2: class Program_weight_tensor_parameter_3: name = "parameter_3" + original_name = "batch_norm2d_81.w_1" shape = [1] dtype = "float32" min_val = float("0") @@ -36,6 +40,7 @@ class Program_weight_tensor_parameter_3: class Program_weight_tensor_parameter_4: name = "parameter_4" + original_name = "conv2d_99.w_0" shape = [1, 1, 1, 1] dtype = "float32" min_val = float("0") @@ -45,6 +50,7 @@ class Program_weight_tensor_parameter_4: class Program_weight_tensor_parameter_5: name = "parameter_5" + original_name = "batch_norm2d_80.b_0" shape = [1] dtype = "float32" min_val = float("0") @@ -54,6 +60,7 @@ class Program_weight_tensor_parameter_5: class Program_weight_tensor_parameter_6: name = "parameter_6" + original_name = "batch_norm2d_80.w_0" shape = [1] dtype = "float32" min_val = float("0") @@ -63,6 +70,7 @@ class Program_weight_tensor_parameter_6: class Program_weight_tensor_parameter_7: name = "parameter_7" + original_name = "batch_norm2d_80.w_2" shape = [1] dtype = "float32" min_val = float("0") @@ -72,6 +80,7 @@ class Program_weight_tensor_parameter_7: class Program_weight_tensor_parameter_8: name = "parameter_8" + original_name = "batch_norm2d_80.w_1" shape = [1] dtype = "float32" min_val = float("0") @@ -81,6 +90,7 @@ class Program_weight_tensor_parameter_8: class Program_weight_tensor_parameter_9: name = "parameter_9" + original_name = "conv2d_98.w_0" shape = [1, 96, 5, 5] dtype = "float32" min_val = float("-0.370391") @@ -92,6 +102,7 @@ class Program_weight_tensor_parameter_9: class Program_weight_tensor_parameter_10: name = "parameter_10" + original_name = "conv2d_97.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -101,6 +112,7 @@ class Program_weight_tensor_parameter_10: class Program_weight_tensor_parameter_11: name = "parameter_11" + original_name = "conv2d_97.w_0" shape = [32, 96, 1, 1] dtype = "float32" min_val = float("-0.596619") @@ -112,6 +124,7 @@ class Program_weight_tensor_parameter_11: class Program_weight_tensor_parameter_12: name = "parameter_12" + original_name = "conv2d_96.b_0" shape = [4] dtype = "float32" min_val = float("0") @@ -121,6 +134,7 @@ class Program_weight_tensor_parameter_12: class Program_weight_tensor_parameter_13: name = "parameter_13" + original_name = "conv2d_96.w_0" shape = [4, 96, 1, 1] dtype = "float32" min_val = float("-0.0247338") @@ -132,6 +146,7 @@ class Program_weight_tensor_parameter_13: class Program_weight_tensor_parameter_14: name = "parameter_14" + original_name = "batch_norm2d_73.b_0" shape = [96] dtype = "float32" min_val = float("-0.365648") @@ -143,6 +158,7 @@ class Program_weight_tensor_parameter_14: class Program_weight_tensor_parameter_15: name = "parameter_15" + original_name = "batch_norm2d_73.w_0" shape = [96] dtype = "float32" min_val = float("0.533085") @@ -154,6 +170,7 @@ class Program_weight_tensor_parameter_15: class Program_weight_tensor_parameter_16: name = "parameter_16" + original_name = "batch_norm2d_73.w_2" shape = [96] dtype = "float32" min_val = float("0.00946462") @@ -165,6 +182,7 @@ class Program_weight_tensor_parameter_16: class Program_weight_tensor_parameter_17: name = "parameter_17" + original_name = "batch_norm2d_73.w_1" shape = [96] dtype = "float32" min_val = float("-0.206108") @@ -176,6 +194,7 @@ class Program_weight_tensor_parameter_17: class Program_weight_tensor_parameter_18: name = "parameter_18" + original_name = "conv2d_81.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.274541") @@ -187,6 +206,7 @@ class Program_weight_tensor_parameter_18: class Program_weight_tensor_parameter_19: name = "parameter_19" + original_name = "conv2d_80.b_0" shape = [96] dtype = "float32" min_val = float("-0.0581132") @@ -198,6 +218,7 @@ class Program_weight_tensor_parameter_19: class Program_weight_tensor_parameter_20: name = "parameter_20" + original_name = "conv2d_80.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.189324") @@ -209,6 +230,7 @@ class Program_weight_tensor_parameter_20: class Program_weight_tensor_parameter_21: name = "parameter_21" + original_name = "batch_norm2d_72.b_0" shape = [96] dtype = "float32" min_val = float("-1.04876") @@ -220,6 +242,7 @@ class Program_weight_tensor_parameter_21: class Program_weight_tensor_parameter_22: name = "parameter_22" + original_name = "batch_norm2d_72.w_0" shape = [96] dtype = "float32" min_val = float("0.549376") @@ -231,6 +254,7 @@ class Program_weight_tensor_parameter_22: class Program_weight_tensor_parameter_23: name = "parameter_23" + original_name = "batch_norm2d_72.w_2" shape = [96] dtype = "float32" min_val = float("0.0395664") @@ -242,6 +266,7 @@ class Program_weight_tensor_parameter_23: class Program_weight_tensor_parameter_24: name = "parameter_24" + original_name = "batch_norm2d_72.w_1" shape = [96] dtype = "float32" min_val = float("-0.743622") @@ -253,6 +278,7 @@ class Program_weight_tensor_parameter_24: class Program_weight_tensor_parameter_25: name = "parameter_25" + original_name = "conv2d_79.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.387566") @@ -264,6 +290,7 @@ class Program_weight_tensor_parameter_25: class Program_weight_tensor_parameter_26: name = "parameter_26" + original_name = "batch_norm2d_71.b_0" shape = [96] dtype = "float32" min_val = float("-0.988036") @@ -275,6 +302,7 @@ class Program_weight_tensor_parameter_26: class Program_weight_tensor_parameter_27: name = "parameter_27" + original_name = "batch_norm2d_71.w_0" shape = [96] dtype = "float32" min_val = float("0.57921") @@ -286,6 +314,7 @@ class Program_weight_tensor_parameter_27: class Program_weight_tensor_parameter_28: name = "parameter_28" + original_name = "batch_norm2d_71.w_2" shape = [96] dtype = "float32" min_val = float("7.04189e-05") @@ -297,6 +326,7 @@ class Program_weight_tensor_parameter_28: class Program_weight_tensor_parameter_29: name = "parameter_29" + original_name = "batch_norm2d_71.w_1" shape = [96] dtype = "float32" min_val = float("-0.386604") @@ -308,6 +338,7 @@ class Program_weight_tensor_parameter_29: class Program_weight_tensor_parameter_30: name = "parameter_30" + original_name = "conv2d_78.w_0" shape = [96, 1, 5, 5] dtype = "float32" min_val = float("-0.209991") @@ -319,6 +350,7 @@ class Program_weight_tensor_parameter_30: class Program_weight_tensor_parameter_31: name = "parameter_31" + original_name = "batch_norm2d_70.b_0" shape = [96] dtype = "float32" min_val = float("-1.28817") @@ -330,6 +362,7 @@ class Program_weight_tensor_parameter_31: class Program_weight_tensor_parameter_32: name = "parameter_32" + original_name = "batch_norm2d_70.w_0" shape = [96] dtype = "float32" min_val = float("0.0495081") @@ -341,6 +374,7 @@ class Program_weight_tensor_parameter_32: class Program_weight_tensor_parameter_33: name = "parameter_33" + original_name = "batch_norm2d_70.w_2" shape = [96] dtype = "float32" min_val = float("0.0364859") @@ -352,6 +386,7 @@ class Program_weight_tensor_parameter_33: class Program_weight_tensor_parameter_34: name = "parameter_34" + original_name = "batch_norm2d_70.w_1" shape = [96] dtype = "float32" min_val = float("-0.706492") @@ -363,6 +398,7 @@ class Program_weight_tensor_parameter_34: class Program_weight_tensor_parameter_35: name = "parameter_35" + original_name = "conv2d_77.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.297062") @@ -374,6 +410,7 @@ class Program_weight_tensor_parameter_35: class Program_weight_tensor_parameter_36: name = "parameter_36" + original_name = "batch_norm2d_69.b_0" shape = [96] dtype = "float32" min_val = float("-1.44232") @@ -385,6 +422,7 @@ class Program_weight_tensor_parameter_36: class Program_weight_tensor_parameter_37: name = "parameter_37" + original_name = "batch_norm2d_69.w_0" shape = [96] dtype = "float32" min_val = float("0.741795") @@ -396,6 +434,7 @@ class Program_weight_tensor_parameter_37: class Program_weight_tensor_parameter_38: name = "parameter_38" + original_name = "batch_norm2d_69.w_2" shape = [96] dtype = "float32" min_val = float("0.000148002") @@ -407,6 +446,7 @@ class Program_weight_tensor_parameter_38: class Program_weight_tensor_parameter_39: name = "parameter_39" + original_name = "batch_norm2d_69.w_1" shape = [96] dtype = "float32" min_val = float("-0.0720141") @@ -418,6 +458,7 @@ class Program_weight_tensor_parameter_39: class Program_weight_tensor_parameter_40: name = "parameter_40" + original_name = "conv2d_76.w_0" shape = [96, 1, 5, 5] dtype = "float32" min_val = float("-0.285311") @@ -429,6 +470,7 @@ class Program_weight_tensor_parameter_40: class Program_weight_tensor_parameter_41: name = "parameter_41" + original_name = "batch_norm2d_79.b_0" shape = [1] dtype = "float32" min_val = float("0") @@ -438,6 +480,7 @@ class Program_weight_tensor_parameter_41: class Program_weight_tensor_parameter_42: name = "parameter_42" + original_name = "batch_norm2d_79.w_0" shape = [1] dtype = "float32" min_val = float("0") @@ -447,6 +490,7 @@ class Program_weight_tensor_parameter_42: class Program_weight_tensor_parameter_43: name = "parameter_43" + original_name = "batch_norm2d_79.w_2" shape = [1] dtype = "float32" min_val = float("0") @@ -456,6 +500,7 @@ class Program_weight_tensor_parameter_43: class Program_weight_tensor_parameter_44: name = "parameter_44" + original_name = "batch_norm2d_79.w_1" shape = [1] dtype = "float32" min_val = float("0") @@ -465,6 +510,7 @@ class Program_weight_tensor_parameter_44: class Program_weight_tensor_parameter_45: name = "parameter_45" + original_name = "conv2d_95.w_0" shape = [1, 1, 1, 1] dtype = "float32" min_val = float("0") @@ -474,6 +520,7 @@ class Program_weight_tensor_parameter_45: class Program_weight_tensor_parameter_46: name = "parameter_46" + original_name = "batch_norm2d_78.b_0" shape = [1] dtype = "float32" min_val = float("0") @@ -483,6 +530,7 @@ class Program_weight_tensor_parameter_46: class Program_weight_tensor_parameter_47: name = "parameter_47" + original_name = "batch_norm2d_78.w_0" shape = [1] dtype = "float32" min_val = float("0") @@ -492,6 +540,7 @@ class Program_weight_tensor_parameter_47: class Program_weight_tensor_parameter_48: name = "parameter_48" + original_name = "batch_norm2d_78.w_2" shape = [1] dtype = "float32" min_val = float("0") @@ -501,6 +550,7 @@ class Program_weight_tensor_parameter_48: class Program_weight_tensor_parameter_49: name = "parameter_49" + original_name = "batch_norm2d_78.w_1" shape = [1] dtype = "float32" min_val = float("0") @@ -510,6 +560,7 @@ class Program_weight_tensor_parameter_49: class Program_weight_tensor_parameter_50: name = "parameter_50" + original_name = "conv2d_94.w_0" shape = [1, 96, 5, 5] dtype = "float32" min_val = float("-0.526895") @@ -521,6 +572,7 @@ class Program_weight_tensor_parameter_50: class Program_weight_tensor_parameter_51: name = "parameter_51" + original_name = "conv2d_93.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -530,6 +582,7 @@ class Program_weight_tensor_parameter_51: class Program_weight_tensor_parameter_52: name = "parameter_52" + original_name = "conv2d_93.w_0" shape = [32, 96, 1, 1] dtype = "float32" min_val = float("-0.556513") @@ -541,6 +594,7 @@ class Program_weight_tensor_parameter_52: class Program_weight_tensor_parameter_53: name = "parameter_53" + original_name = "conv2d_92.b_0" shape = [4] dtype = "float32" min_val = float("0") @@ -550,6 +604,7 @@ class Program_weight_tensor_parameter_53: class Program_weight_tensor_parameter_54: name = "parameter_54" + original_name = "conv2d_92.w_0" shape = [4, 96, 1, 1] dtype = "float32" min_val = float("-0.0293143") @@ -561,6 +616,7 @@ class Program_weight_tensor_parameter_54: class Program_weight_tensor_parameter_55: name = "parameter_55" + original_name = "batch_norm2d_68.b_0" shape = [96] dtype = "float32" min_val = float("-0.515496") @@ -572,6 +628,7 @@ class Program_weight_tensor_parameter_55: class Program_weight_tensor_parameter_56: name = "parameter_56" + original_name = "batch_norm2d_68.w_0" shape = [96] dtype = "float32" min_val = float("0.516979") @@ -583,6 +640,7 @@ class Program_weight_tensor_parameter_56: class Program_weight_tensor_parameter_57: name = "parameter_57" + original_name = "batch_norm2d_68.w_2" shape = [96] dtype = "float32" min_val = float("0.0112997") @@ -594,6 +652,7 @@ class Program_weight_tensor_parameter_57: class Program_weight_tensor_parameter_58: name = "parameter_58" + original_name = "batch_norm2d_68.w_1" shape = [96] dtype = "float32" min_val = float("-0.332276") @@ -605,6 +664,7 @@ class Program_weight_tensor_parameter_58: class Program_weight_tensor_parameter_59: name = "parameter_59" + original_name = "conv2d_75.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.418305") @@ -616,6 +676,7 @@ class Program_weight_tensor_parameter_59: class Program_weight_tensor_parameter_60: name = "parameter_60" + original_name = "conv2d_74.b_0" shape = [96] dtype = "float32" min_val = float("-0.0361324") @@ -627,6 +688,7 @@ class Program_weight_tensor_parameter_60: class Program_weight_tensor_parameter_61: name = "parameter_61" + original_name = "conv2d_74.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.35548") @@ -638,6 +700,7 @@ class Program_weight_tensor_parameter_61: class Program_weight_tensor_parameter_62: name = "parameter_62" + original_name = "batch_norm2d_67.b_0" shape = [96] dtype = "float32" min_val = float("-1.9389") @@ -649,6 +712,7 @@ class Program_weight_tensor_parameter_62: class Program_weight_tensor_parameter_63: name = "parameter_63" + original_name = "batch_norm2d_67.w_0" shape = [96] dtype = "float32" min_val = float("0.382919") @@ -660,6 +724,7 @@ class Program_weight_tensor_parameter_63: class Program_weight_tensor_parameter_64: name = "parameter_64" + original_name = "batch_norm2d_67.w_2" shape = [96] dtype = "float32" min_val = float("0.0557326") @@ -671,6 +736,7 @@ class Program_weight_tensor_parameter_64: class Program_weight_tensor_parameter_65: name = "parameter_65" + original_name = "batch_norm2d_67.w_1" shape = [96] dtype = "float32" min_val = float("-1.73652") @@ -682,6 +748,7 @@ class Program_weight_tensor_parameter_65: class Program_weight_tensor_parameter_66: name = "parameter_66" + original_name = "conv2d_73.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.521325") @@ -693,6 +760,7 @@ class Program_weight_tensor_parameter_66: class Program_weight_tensor_parameter_67: name = "parameter_67" + original_name = "batch_norm2d_66.b_0" shape = [96] dtype = "float32" min_val = float("-0.981454") @@ -704,6 +772,7 @@ class Program_weight_tensor_parameter_67: class Program_weight_tensor_parameter_68: name = "parameter_68" + original_name = "batch_norm2d_66.w_0" shape = [96] dtype = "float32" min_val = float("0.497136") @@ -715,6 +784,7 @@ class Program_weight_tensor_parameter_68: class Program_weight_tensor_parameter_69: name = "parameter_69" + original_name = "batch_norm2d_66.w_2" shape = [96] dtype = "float32" min_val = float("0.00239578") @@ -726,6 +796,7 @@ class Program_weight_tensor_parameter_69: class Program_weight_tensor_parameter_70: name = "parameter_70" + original_name = "batch_norm2d_66.w_1" shape = [96] dtype = "float32" min_val = float("-0.362203") @@ -737,6 +808,7 @@ class Program_weight_tensor_parameter_70: class Program_weight_tensor_parameter_71: name = "parameter_71" + original_name = "conv2d_72.w_0" shape = [96, 1, 5, 5] dtype = "float32" min_val = float("-0.383866") @@ -748,6 +820,7 @@ class Program_weight_tensor_parameter_71: class Program_weight_tensor_parameter_72: name = "parameter_72" + original_name = "batch_norm2d_65.b_0" shape = [96] dtype = "float32" min_val = float("-1.27052") @@ -759,6 +832,7 @@ class Program_weight_tensor_parameter_72: class Program_weight_tensor_parameter_73: name = "parameter_73" + original_name = "batch_norm2d_65.w_0" shape = [96] dtype = "float32" min_val = float("0.143835") @@ -770,6 +844,7 @@ class Program_weight_tensor_parameter_73: class Program_weight_tensor_parameter_74: name = "parameter_74" + original_name = "batch_norm2d_65.w_2" shape = [96] dtype = "float32" min_val = float("0.0328354") @@ -781,6 +856,7 @@ class Program_weight_tensor_parameter_74: class Program_weight_tensor_parameter_75: name = "parameter_75" + original_name = "batch_norm2d_65.w_1" shape = [96] dtype = "float32" min_val = float("-1.18394") @@ -792,6 +868,7 @@ class Program_weight_tensor_parameter_75: class Program_weight_tensor_parameter_76: name = "parameter_76" + original_name = "conv2d_71.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.411144") @@ -803,6 +880,7 @@ class Program_weight_tensor_parameter_76: class Program_weight_tensor_parameter_77: name = "parameter_77" + original_name = "batch_norm2d_64.b_0" shape = [96] dtype = "float32" min_val = float("-0.951711") @@ -814,6 +892,7 @@ class Program_weight_tensor_parameter_77: class Program_weight_tensor_parameter_78: name = "parameter_78" + original_name = "batch_norm2d_64.w_0" shape = [96] dtype = "float32" min_val = float("0.421821") @@ -825,6 +904,7 @@ class Program_weight_tensor_parameter_78: class Program_weight_tensor_parameter_79: name = "parameter_79" + original_name = "batch_norm2d_64.w_2" shape = [96] dtype = "float32" min_val = float("0.000407052") @@ -836,6 +916,7 @@ class Program_weight_tensor_parameter_79: class Program_weight_tensor_parameter_80: name = "parameter_80" + original_name = "batch_norm2d_64.w_1" shape = [96] dtype = "float32" min_val = float("-0.638224") @@ -847,6 +928,7 @@ class Program_weight_tensor_parameter_80: class Program_weight_tensor_parameter_81: name = "parameter_81" + original_name = "conv2d_70.w_0" shape = [96, 1, 5, 5] dtype = "float32" min_val = float("-0.266356") @@ -858,6 +940,7 @@ class Program_weight_tensor_parameter_81: class Program_weight_tensor_parameter_82: name = "parameter_82" + original_name = "batch_norm2d_77.b_0" shape = [1] dtype = "float32" min_val = float("0") @@ -867,6 +950,7 @@ class Program_weight_tensor_parameter_82: class Program_weight_tensor_parameter_83: name = "parameter_83" + original_name = "batch_norm2d_77.w_0" shape = [1] dtype = "float32" min_val = float("0") @@ -876,6 +960,7 @@ class Program_weight_tensor_parameter_83: class Program_weight_tensor_parameter_84: name = "parameter_84" + original_name = "batch_norm2d_77.w_2" shape = [1] dtype = "float32" min_val = float("0") @@ -885,6 +970,7 @@ class Program_weight_tensor_parameter_84: class Program_weight_tensor_parameter_85: name = "parameter_85" + original_name = "batch_norm2d_77.w_1" shape = [1] dtype = "float32" min_val = float("0") @@ -894,6 +980,7 @@ class Program_weight_tensor_parameter_85: class Program_weight_tensor_parameter_86: name = "parameter_86" + original_name = "conv2d_91.w_0" shape = [1, 1, 1, 1] dtype = "float32" min_val = float("0") @@ -903,6 +990,7 @@ class Program_weight_tensor_parameter_86: class Program_weight_tensor_parameter_87: name = "parameter_87" + original_name = "batch_norm2d_76.b_0" shape = [1] dtype = "float32" min_val = float("0") @@ -912,6 +1000,7 @@ class Program_weight_tensor_parameter_87: class Program_weight_tensor_parameter_88: name = "parameter_88" + original_name = "batch_norm2d_76.w_0" shape = [1] dtype = "float32" min_val = float("0") @@ -921,6 +1010,7 @@ class Program_weight_tensor_parameter_88: class Program_weight_tensor_parameter_89: name = "parameter_89" + original_name = "batch_norm2d_76.w_2" shape = [1] dtype = "float32" min_val = float("0") @@ -930,6 +1020,7 @@ class Program_weight_tensor_parameter_89: class Program_weight_tensor_parameter_90: name = "parameter_90" + original_name = "batch_norm2d_76.w_1" shape = [1] dtype = "float32" min_val = float("0") @@ -939,6 +1030,7 @@ class Program_weight_tensor_parameter_90: class Program_weight_tensor_parameter_91: name = "parameter_91" + original_name = "conv2d_90.w_0" shape = [1, 96, 5, 5] dtype = "float32" min_val = float("-0.149375") @@ -950,6 +1042,7 @@ class Program_weight_tensor_parameter_91: class Program_weight_tensor_parameter_92: name = "parameter_92" + original_name = "conv2d_89.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -959,6 +1052,7 @@ class Program_weight_tensor_parameter_92: class Program_weight_tensor_parameter_93: name = "parameter_93" + original_name = "conv2d_89.w_0" shape = [32, 96, 1, 1] dtype = "float32" min_val = float("-0.563531") @@ -970,6 +1064,7 @@ class Program_weight_tensor_parameter_93: class Program_weight_tensor_parameter_94: name = "parameter_94" + original_name = "conv2d_88.b_0" shape = [4] dtype = "float32" min_val = float("0") @@ -979,6 +1074,7 @@ class Program_weight_tensor_parameter_94: class Program_weight_tensor_parameter_95: name = "parameter_95" + original_name = "conv2d_88.w_0" shape = [4, 96, 1, 1] dtype = "float32" min_val = float("-0.0279582") @@ -990,6 +1086,7 @@ class Program_weight_tensor_parameter_95: class Program_weight_tensor_parameter_96: name = "parameter_96" + original_name = "batch_norm2d_63.b_0" shape = [96] dtype = "float32" min_val = float("-1.01967") @@ -1001,6 +1098,7 @@ class Program_weight_tensor_parameter_96: class Program_weight_tensor_parameter_97: name = "parameter_97" + original_name = "batch_norm2d_63.w_0" shape = [96] dtype = "float32" min_val = float("0.727034") @@ -1012,6 +1110,7 @@ class Program_weight_tensor_parameter_97: class Program_weight_tensor_parameter_98: name = "parameter_98" + original_name = "batch_norm2d_63.w_2" shape = [96] dtype = "float32" min_val = float("0.0295035") @@ -1023,6 +1122,7 @@ class Program_weight_tensor_parameter_98: class Program_weight_tensor_parameter_99: name = "parameter_99" + original_name = "batch_norm2d_63.w_1" shape = [96] dtype = "float32" min_val = float("-0.960749") @@ -1034,6 +1134,7 @@ class Program_weight_tensor_parameter_99: class Program_weight_tensor_parameter_100: name = "parameter_100" + original_name = "conv2d_69.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.509252") @@ -1045,6 +1146,7 @@ class Program_weight_tensor_parameter_100: class Program_weight_tensor_parameter_101: name = "parameter_101" + original_name = "conv2d_68.b_0" shape = [96] dtype = "float32" min_val = float("-0.0391753") @@ -1056,6 +1158,7 @@ class Program_weight_tensor_parameter_101: class Program_weight_tensor_parameter_102: name = "parameter_102" + original_name = "conv2d_68.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.420426") @@ -1067,6 +1170,7 @@ class Program_weight_tensor_parameter_102: class Program_weight_tensor_parameter_103: name = "parameter_103" + original_name = "batch_norm2d_62.b_0" shape = [96] dtype = "float32" min_val = float("-1.50395") @@ -1078,6 +1182,7 @@ class Program_weight_tensor_parameter_103: class Program_weight_tensor_parameter_104: name = "parameter_104" + original_name = "batch_norm2d_62.w_0" shape = [96] dtype = "float32" min_val = float("0.305909") @@ -1089,6 +1194,7 @@ class Program_weight_tensor_parameter_104: class Program_weight_tensor_parameter_105: name = "parameter_105" + original_name = "batch_norm2d_62.w_2" shape = [96] dtype = "float32" min_val = float("0.0659031") @@ -1100,6 +1206,7 @@ class Program_weight_tensor_parameter_105: class Program_weight_tensor_parameter_106: name = "parameter_106" + original_name = "batch_norm2d_62.w_1" shape = [96] dtype = "float32" min_val = float("-1.48733") @@ -1111,6 +1218,7 @@ class Program_weight_tensor_parameter_106: class Program_weight_tensor_parameter_107: name = "parameter_107" + original_name = "conv2d_67.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.44399") @@ -1122,6 +1230,7 @@ class Program_weight_tensor_parameter_107: class Program_weight_tensor_parameter_108: name = "parameter_108" + original_name = "batch_norm2d_61.b_0" shape = [96] dtype = "float32" min_val = float("-0.748428") @@ -1133,6 +1242,7 @@ class Program_weight_tensor_parameter_108: class Program_weight_tensor_parameter_109: name = "parameter_109" + original_name = "batch_norm2d_61.w_0" shape = [96] dtype = "float32" min_val = float("0.391552") @@ -1144,6 +1254,7 @@ class Program_weight_tensor_parameter_109: class Program_weight_tensor_parameter_110: name = "parameter_110" + original_name = "batch_norm2d_61.w_2" shape = [96] dtype = "float32" min_val = float("8.30656e-05") @@ -1155,6 +1266,7 @@ class Program_weight_tensor_parameter_110: class Program_weight_tensor_parameter_111: name = "parameter_111" + original_name = "batch_norm2d_61.w_1" shape = [96] dtype = "float32" min_val = float("-0.493572") @@ -1166,6 +1278,7 @@ class Program_weight_tensor_parameter_111: class Program_weight_tensor_parameter_112: name = "parameter_112" + original_name = "conv2d_66.w_0" shape = [96, 1, 5, 5] dtype = "float32" min_val = float("-0.382507") @@ -1177,6 +1290,7 @@ class Program_weight_tensor_parameter_112: class Program_weight_tensor_parameter_113: name = "parameter_113" + original_name = "batch_norm2d_60.b_0" shape = [96] dtype = "float32" min_val = float("-0.928614") @@ -1188,6 +1302,7 @@ class Program_weight_tensor_parameter_113: class Program_weight_tensor_parameter_114: name = "parameter_114" + original_name = "batch_norm2d_60.w_0" shape = [96] dtype = "float32" min_val = float("0.0268282") @@ -1199,6 +1314,7 @@ class Program_weight_tensor_parameter_114: class Program_weight_tensor_parameter_115: name = "parameter_115" + original_name = "batch_norm2d_60.w_2" shape = [96] dtype = "float32" min_val = float("0.0555829") @@ -1210,6 +1326,7 @@ class Program_weight_tensor_parameter_115: class Program_weight_tensor_parameter_116: name = "parameter_116" + original_name = "batch_norm2d_60.w_1" shape = [96] dtype = "float32" min_val = float("-0.957678") @@ -1221,6 +1338,7 @@ class Program_weight_tensor_parameter_116: class Program_weight_tensor_parameter_117: name = "parameter_117" + original_name = "conv2d_65.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.494034") @@ -1232,6 +1350,7 @@ class Program_weight_tensor_parameter_117: class Program_weight_tensor_parameter_118: name = "parameter_118" + original_name = "batch_norm2d_59.b_0" shape = [96] dtype = "float32" min_val = float("-0.255533") @@ -1243,6 +1362,7 @@ class Program_weight_tensor_parameter_118: class Program_weight_tensor_parameter_119: name = "parameter_119" + original_name = "batch_norm2d_59.w_0" shape = [96] dtype = "float32" min_val = float("0.337748") @@ -1254,6 +1374,7 @@ class Program_weight_tensor_parameter_119: class Program_weight_tensor_parameter_120: name = "parameter_120" + original_name = "batch_norm2d_59.w_2" shape = [96] dtype = "float32" min_val = float("4.89575e-05") @@ -1265,6 +1386,7 @@ class Program_weight_tensor_parameter_120: class Program_weight_tensor_parameter_121: name = "parameter_121" + original_name = "batch_norm2d_59.w_1" shape = [96] dtype = "float32" min_val = float("-0.924544") @@ -1276,6 +1398,7 @@ class Program_weight_tensor_parameter_121: class Program_weight_tensor_parameter_122: name = "parameter_122" + original_name = "conv2d_64.w_0" shape = [96, 1, 5, 5] dtype = "float32" min_val = float("-0.355186") @@ -1287,6 +1410,7 @@ class Program_weight_tensor_parameter_122: class Program_weight_tensor_parameter_123: name = "parameter_123" + original_name = "batch_norm2d_75.b_0" shape = [1] dtype = "float32" min_val = float("0") @@ -1296,6 +1420,7 @@ class Program_weight_tensor_parameter_123: class Program_weight_tensor_parameter_124: name = "parameter_124" + original_name = "batch_norm2d_75.w_0" shape = [1] dtype = "float32" min_val = float("0") @@ -1305,6 +1430,7 @@ class Program_weight_tensor_parameter_124: class Program_weight_tensor_parameter_125: name = "parameter_125" + original_name = "batch_norm2d_75.w_2" shape = [1] dtype = "float32" min_val = float("0") @@ -1314,6 +1440,7 @@ class Program_weight_tensor_parameter_125: class Program_weight_tensor_parameter_126: name = "parameter_126" + original_name = "batch_norm2d_75.w_1" shape = [1] dtype = "float32" min_val = float("0") @@ -1323,6 +1450,7 @@ class Program_weight_tensor_parameter_126: class Program_weight_tensor_parameter_127: name = "parameter_127" + original_name = "conv2d_87.w_0" shape = [1, 1, 1, 1] dtype = "float32" min_val = float("0") @@ -1332,6 +1460,7 @@ class Program_weight_tensor_parameter_127: class Program_weight_tensor_parameter_128: name = "parameter_128" + original_name = "batch_norm2d_74.b_0" shape = [1] dtype = "float32" min_val = float("0") @@ -1341,6 +1470,7 @@ class Program_weight_tensor_parameter_128: class Program_weight_tensor_parameter_129: name = "parameter_129" + original_name = "batch_norm2d_74.w_0" shape = [1] dtype = "float32" min_val = float("0") @@ -1350,6 +1480,7 @@ class Program_weight_tensor_parameter_129: class Program_weight_tensor_parameter_130: name = "parameter_130" + original_name = "batch_norm2d_74.w_2" shape = [1] dtype = "float32" min_val = float("0") @@ -1359,6 +1490,7 @@ class Program_weight_tensor_parameter_130: class Program_weight_tensor_parameter_131: name = "parameter_131" + original_name = "batch_norm2d_74.w_1" shape = [1] dtype = "float32" min_val = float("0") @@ -1368,6 +1500,7 @@ class Program_weight_tensor_parameter_131: class Program_weight_tensor_parameter_132: name = "parameter_132" + original_name = "conv2d_86.w_0" shape = [1, 96, 5, 5] dtype = "float32" min_val = float("-0.309517") @@ -1379,6 +1512,7 @@ class Program_weight_tensor_parameter_132: class Program_weight_tensor_parameter_133: name = "parameter_133" + original_name = "conv2d_85.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -1388,6 +1522,7 @@ class Program_weight_tensor_parameter_133: class Program_weight_tensor_parameter_134: name = "parameter_134" + original_name = "conv2d_85.w_0" shape = [32, 96, 1, 1] dtype = "float32" min_val = float("-0.557687") @@ -1399,6 +1534,7 @@ class Program_weight_tensor_parameter_134: class Program_weight_tensor_parameter_135: name = "parameter_135" + original_name = "conv2d_84.b_0" shape = [4] dtype = "float32" min_val = float("0") @@ -1408,6 +1544,7 @@ class Program_weight_tensor_parameter_135: class Program_weight_tensor_parameter_136: name = "parameter_136" + original_name = "conv2d_84.w_0" shape = [4, 96, 1, 1] dtype = "float32" min_val = float("-0.0253596") @@ -1419,6 +1556,7 @@ class Program_weight_tensor_parameter_136: class Program_weight_tensor_parameter_137: name = "parameter_137" + original_name = "batch_norm2d_58.b_0" shape = [96] dtype = "float32" min_val = float("-0.897616") @@ -1430,6 +1568,7 @@ class Program_weight_tensor_parameter_137: class Program_weight_tensor_parameter_138: name = "parameter_138" + original_name = "batch_norm2d_58.w_0" shape = [96] dtype = "float32" min_val = float("0.691672") @@ -1441,6 +1580,7 @@ class Program_weight_tensor_parameter_138: class Program_weight_tensor_parameter_139: name = "parameter_139" + original_name = "batch_norm2d_58.w_2" shape = [96] dtype = "float32" min_val = float("0.00702492") @@ -1452,6 +1592,7 @@ class Program_weight_tensor_parameter_139: class Program_weight_tensor_parameter_140: name = "parameter_140" + original_name = "batch_norm2d_58.w_1" shape = [96] dtype = "float32" min_val = float("-0.433618") @@ -1463,6 +1604,7 @@ class Program_weight_tensor_parameter_140: class Program_weight_tensor_parameter_141: name = "parameter_141" + original_name = "conv2d_63.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.664341") @@ -1474,6 +1616,7 @@ class Program_weight_tensor_parameter_141: class Program_weight_tensor_parameter_142: name = "parameter_142" + original_name = "conv2d_62.b_0" shape = [96] dtype = "float32" min_val = float("-0.0408503") @@ -1485,6 +1628,7 @@ class Program_weight_tensor_parameter_142: class Program_weight_tensor_parameter_143: name = "parameter_143" + original_name = "conv2d_62.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.641514") @@ -1496,6 +1640,7 @@ class Program_weight_tensor_parameter_143: class Program_weight_tensor_parameter_144: name = "parameter_144" + original_name = "batch_norm2d_57.b_0" shape = [96] dtype = "float32" min_val = float("-1.86159") @@ -1507,6 +1652,7 @@ class Program_weight_tensor_parameter_144: class Program_weight_tensor_parameter_145: name = "parameter_145" + original_name = "batch_norm2d_57.w_0" shape = [96] dtype = "float32" min_val = float("0.140902") @@ -1518,6 +1664,7 @@ class Program_weight_tensor_parameter_145: class Program_weight_tensor_parameter_146: name = "parameter_146" + original_name = "batch_norm2d_57.w_2" shape = [96] dtype = "float32" min_val = float("0.0575649") @@ -1529,6 +1676,7 @@ class Program_weight_tensor_parameter_146: class Program_weight_tensor_parameter_147: name = "parameter_147" + original_name = "batch_norm2d_57.w_1" shape = [96] dtype = "float32" min_val = float("-3.02587") @@ -1540,6 +1688,7 @@ class Program_weight_tensor_parameter_147: class Program_weight_tensor_parameter_148: name = "parameter_148" + original_name = "conv2d_61.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.611152") @@ -1551,6 +1700,7 @@ class Program_weight_tensor_parameter_148: class Program_weight_tensor_parameter_149: name = "parameter_149" + original_name = "batch_norm2d_56.b_0" shape = [96] dtype = "float32" min_val = float("-1.22176") @@ -1562,6 +1712,7 @@ class Program_weight_tensor_parameter_149: class Program_weight_tensor_parameter_150: name = "parameter_150" + original_name = "batch_norm2d_56.w_0" shape = [96] dtype = "float32" min_val = float("0.366806") @@ -1573,6 +1724,7 @@ class Program_weight_tensor_parameter_150: class Program_weight_tensor_parameter_151: name = "parameter_151" + original_name = "batch_norm2d_56.w_2" shape = [96] dtype = "float32" min_val = float("6.76028e-05") @@ -1584,6 +1736,7 @@ class Program_weight_tensor_parameter_151: class Program_weight_tensor_parameter_152: name = "parameter_152" + original_name = "batch_norm2d_56.w_1" shape = [96] dtype = "float32" min_val = float("-0.796705") @@ -1595,6 +1748,7 @@ class Program_weight_tensor_parameter_152: class Program_weight_tensor_parameter_153: name = "parameter_153" + original_name = "conv2d_60.w_0" shape = [96, 1, 5, 5] dtype = "float32" min_val = float("-0.522492") @@ -1606,6 +1760,7 @@ class Program_weight_tensor_parameter_153: class Program_weight_tensor_parameter_154: name = "parameter_154" + original_name = "batch_norm2d_55.b_0" shape = [96] dtype = "float32" min_val = float("-1.54069") @@ -1617,6 +1772,7 @@ class Program_weight_tensor_parameter_154: class Program_weight_tensor_parameter_155: name = "parameter_155" + original_name = "batch_norm2d_55.w_0" shape = [96] dtype = "float32" min_val = float("0.0272979") @@ -1628,6 +1784,7 @@ class Program_weight_tensor_parameter_155: class Program_weight_tensor_parameter_156: name = "parameter_156" + original_name = "batch_norm2d_55.w_2" shape = [96] dtype = "float32" min_val = float("0.0455815") @@ -1639,6 +1796,7 @@ class Program_weight_tensor_parameter_156: class Program_weight_tensor_parameter_157: name = "parameter_157" + original_name = "batch_norm2d_55.w_1" shape = [96] dtype = "float32" min_val = float("-1.45937") @@ -1650,6 +1808,7 @@ class Program_weight_tensor_parameter_157: class Program_weight_tensor_parameter_158: name = "parameter_158" + original_name = "conv2d_59.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.475431") @@ -1661,6 +1820,7 @@ class Program_weight_tensor_parameter_158: class Program_weight_tensor_parameter_159: name = "parameter_159" + original_name = "batch_norm2d_54.b_0" shape = [96] dtype = "float32" min_val = float("-1.79696") @@ -1672,6 +1832,7 @@ class Program_weight_tensor_parameter_159: class Program_weight_tensor_parameter_160: name = "parameter_160" + original_name = "batch_norm2d_54.w_0" shape = [96] dtype = "float32" min_val = float("0.245582") @@ -1683,6 +1844,7 @@ class Program_weight_tensor_parameter_160: class Program_weight_tensor_parameter_161: name = "parameter_161" + original_name = "batch_norm2d_54.w_2" shape = [96] dtype = "float32" min_val = float("0.000411559") @@ -1694,6 +1856,7 @@ class Program_weight_tensor_parameter_161: class Program_weight_tensor_parameter_162: name = "parameter_162" + original_name = "batch_norm2d_54.w_1" shape = [96] dtype = "float32" min_val = float("-0.530734") @@ -1705,6 +1868,7 @@ class Program_weight_tensor_parameter_162: class Program_weight_tensor_parameter_163: name = "parameter_163" + original_name = "conv2d_58.w_0" shape = [96, 1, 5, 5] dtype = "float32" min_val = float("-0.383802") @@ -1716,6 +1880,7 @@ class Program_weight_tensor_parameter_163: class Program_weight_tensor_parameter_164: name = "parameter_164" + original_name = "batch_norm2d_33.b_0" shape = [96] dtype = "float32" min_val = float("-0.177089") @@ -1727,6 +1892,7 @@ class Program_weight_tensor_parameter_164: class Program_weight_tensor_parameter_165: name = "parameter_165" + original_name = "batch_norm2d_33.w_0" shape = [96] dtype = "float32" min_val = float("0.0512158") @@ -1738,6 +1904,7 @@ class Program_weight_tensor_parameter_165: class Program_weight_tensor_parameter_166: name = "parameter_166" + original_name = "batch_norm2d_33.w_2" shape = [96] dtype = "float32" min_val = float("0.000580428") @@ -1749,6 +1916,7 @@ class Program_weight_tensor_parameter_166: class Program_weight_tensor_parameter_167: name = "parameter_167" + original_name = "batch_norm2d_33.w_1" shape = [96] dtype = "float32" min_val = float("-0.0193415") @@ -1760,6 +1928,7 @@ class Program_weight_tensor_parameter_167: class Program_weight_tensor_parameter_168: name = "parameter_168" + original_name = "conv2d_37.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.286595") @@ -1771,6 +1940,7 @@ class Program_weight_tensor_parameter_168: class Program_weight_tensor_parameter_169: name = "parameter_169" + original_name = "batch_norm2d_32.b_0" shape = [96] dtype = "float32" min_val = float("-0.260296") @@ -1782,6 +1952,7 @@ class Program_weight_tensor_parameter_169: class Program_weight_tensor_parameter_170: name = "parameter_170" + original_name = "batch_norm2d_32.w_0" shape = [96] dtype = "float32" min_val = float("0.110983") @@ -1793,6 +1964,7 @@ class Program_weight_tensor_parameter_170: class Program_weight_tensor_parameter_171: name = "parameter_171" + original_name = "batch_norm2d_32.w_2" shape = [96] dtype = "float32" min_val = float("0.0001147") @@ -1804,6 +1976,7 @@ class Program_weight_tensor_parameter_171: class Program_weight_tensor_parameter_172: name = "parameter_172" + original_name = "batch_norm2d_32.w_1" shape = [96] dtype = "float32" min_val = float("-0.295121") @@ -1815,6 +1988,7 @@ class Program_weight_tensor_parameter_172: class Program_weight_tensor_parameter_173: name = "parameter_173" + original_name = "conv2d_36.w_0" shape = [96, 1, 5, 5] dtype = "float32" min_val = float("-0.1392") @@ -1826,6 +2000,7 @@ class Program_weight_tensor_parameter_173: class Program_weight_tensor_parameter_174: name = "parameter_174" + original_name = "batch_norm2d_31.b_0" shape = [96] dtype = "float32" min_val = float("-0.164623") @@ -1837,6 +2012,7 @@ class Program_weight_tensor_parameter_174: class Program_weight_tensor_parameter_175: name = "parameter_175" + original_name = "batch_norm2d_31.w_0" shape = [96] dtype = "float32" min_val = float("0.0174356") @@ -1848,6 +2024,7 @@ class Program_weight_tensor_parameter_175: class Program_weight_tensor_parameter_176: name = "parameter_176" + original_name = "batch_norm2d_31.w_2" shape = [96] dtype = "float32" min_val = float("5.37452e-05") @@ -1859,6 +2036,7 @@ class Program_weight_tensor_parameter_176: class Program_weight_tensor_parameter_177: name = "parameter_177" + original_name = "batch_norm2d_31.w_1" shape = [96] dtype = "float32" min_val = float("-0.00460261") @@ -1870,6 +2048,7 @@ class Program_weight_tensor_parameter_177: class Program_weight_tensor_parameter_178: name = "parameter_178" + original_name = "conv2d_35.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.13472") @@ -1881,6 +2060,7 @@ class Program_weight_tensor_parameter_178: class Program_weight_tensor_parameter_179: name = "parameter_179" + original_name = "batch_norm2d_30.b_0" shape = [96] dtype = "float32" min_val = float("-0.0326482") @@ -1892,6 +2072,7 @@ class Program_weight_tensor_parameter_179: class Program_weight_tensor_parameter_180: name = "parameter_180" + original_name = "batch_norm2d_30.w_0" shape = [96] dtype = "float32" min_val = float("0.0808171") @@ -1903,6 +2084,7 @@ class Program_weight_tensor_parameter_180: class Program_weight_tensor_parameter_181: name = "parameter_181" + original_name = "batch_norm2d_30.w_2" shape = [96] dtype = "float32" min_val = float("3.14679e-05") @@ -1914,6 +2096,7 @@ class Program_weight_tensor_parameter_181: class Program_weight_tensor_parameter_182: name = "parameter_182" + original_name = "batch_norm2d_30.w_1" shape = [96] dtype = "float32" min_val = float("-0.0180771") @@ -1925,6 +2108,7 @@ class Program_weight_tensor_parameter_182: class Program_weight_tensor_parameter_183: name = "parameter_183" + original_name = "conv2d_34.w_0" shape = [96, 1, 5, 5] dtype = "float32" min_val = float("-0.0873846") @@ -1936,6 +2120,7 @@ class Program_weight_tensor_parameter_183: class Program_weight_tensor_parameter_184: name = "parameter_184" + original_name = "batch_norm2d_53.b_0" shape = [96] dtype = "float32" min_val = float("-1.30999") @@ -1947,6 +2132,7 @@ class Program_weight_tensor_parameter_184: class Program_weight_tensor_parameter_185: name = "parameter_185" + original_name = "batch_norm2d_53.w_0" shape = [96] dtype = "float32" min_val = float("0.0866569") @@ -1958,6 +2144,7 @@ class Program_weight_tensor_parameter_185: class Program_weight_tensor_parameter_186: name = "parameter_186" + original_name = "batch_norm2d_53.w_2" shape = [96] dtype = "float32" min_val = float("0.115122") @@ -1969,6 +2156,7 @@ class Program_weight_tensor_parameter_186: class Program_weight_tensor_parameter_187: name = "parameter_187" + original_name = "batch_norm2d_53.w_1" shape = [96] dtype = "float32" min_val = float("-0.566716") @@ -1980,6 +2168,7 @@ class Program_weight_tensor_parameter_187: class Program_weight_tensor_parameter_188: name = "parameter_188" + original_name = "conv2d_57.w_0" shape = [96, 192, 1, 1] dtype = "float32" min_val = float("-0.33141") @@ -1991,6 +2180,7 @@ class Program_weight_tensor_parameter_188: class Program_weight_tensor_parameter_189: name = "parameter_189" + original_name = "batch_norm2d_52.b_0" shape = [192] dtype = "float32" min_val = float("-1.00179") @@ -2002,6 +2192,7 @@ class Program_weight_tensor_parameter_189: class Program_weight_tensor_parameter_190: name = "parameter_190" + original_name = "batch_norm2d_52.w_0" shape = [192] dtype = "float32" min_val = float("0.59987") @@ -2013,6 +2204,7 @@ class Program_weight_tensor_parameter_190: class Program_weight_tensor_parameter_191: name = "parameter_191" + original_name = "batch_norm2d_52.w_2" shape = [192] dtype = "float32" min_val = float("0.0036511") @@ -2024,6 +2216,7 @@ class Program_weight_tensor_parameter_191: class Program_weight_tensor_parameter_192: name = "parameter_192" + original_name = "batch_norm2d_52.w_1" shape = [192] dtype = "float32" min_val = float("-0.376435") @@ -2035,6 +2228,7 @@ class Program_weight_tensor_parameter_192: class Program_weight_tensor_parameter_193: name = "parameter_193" + original_name = "conv2d_56.w_0" shape = [192, 1, 5, 5] dtype = "float32" min_val = float("-0.304953") @@ -2046,6 +2240,7 @@ class Program_weight_tensor_parameter_193: class Program_weight_tensor_parameter_194: name = "parameter_194" + original_name = "batch_norm2d_51.b_0" shape = [192] dtype = "float32" min_val = float("-1.5034") @@ -2057,6 +2252,7 @@ class Program_weight_tensor_parameter_194: class Program_weight_tensor_parameter_195: name = "parameter_195" + original_name = "batch_norm2d_51.w_0" shape = [192] dtype = "float32" min_val = float("0.411181") @@ -2068,6 +2264,7 @@ class Program_weight_tensor_parameter_195: class Program_weight_tensor_parameter_196: name = "parameter_196" + original_name = "batch_norm2d_51.w_2" shape = [192] dtype = "float32" min_val = float("0.146266") @@ -2079,6 +2276,7 @@ class Program_weight_tensor_parameter_196: class Program_weight_tensor_parameter_197: name = "parameter_197" + original_name = "batch_norm2d_51.w_1" shape = [192] dtype = "float32" min_val = float("-0.949296") @@ -2090,6 +2288,7 @@ class Program_weight_tensor_parameter_197: class Program_weight_tensor_parameter_198: name = "parameter_198" + original_name = "conv2d_55.w_0" shape = [192, 192, 1, 1] dtype = "float32" min_val = float("-0.295619") @@ -2101,6 +2300,7 @@ class Program_weight_tensor_parameter_198: class Program_weight_tensor_parameter_199: name = "parameter_199" + original_name = "batch_norm2d_50.b_0" shape = [192] dtype = "float32" min_val = float("-1.31561") @@ -2112,6 +2312,7 @@ class Program_weight_tensor_parameter_199: class Program_weight_tensor_parameter_200: name = "parameter_200" + original_name = "batch_norm2d_50.w_0" shape = [192] dtype = "float32" min_val = float("0.725639") @@ -2123,6 +2324,7 @@ class Program_weight_tensor_parameter_200: class Program_weight_tensor_parameter_201: name = "parameter_201" + original_name = "batch_norm2d_50.w_2" shape = [192] dtype = "float32" min_val = float("5.99319e-05") @@ -2134,6 +2336,7 @@ class Program_weight_tensor_parameter_201: class Program_weight_tensor_parameter_202: name = "parameter_202" + original_name = "batch_norm2d_50.w_1" shape = [192] dtype = "float32" min_val = float("-0.0445685") @@ -2145,6 +2348,7 @@ class Program_weight_tensor_parameter_202: class Program_weight_tensor_parameter_203: name = "parameter_203" + original_name = "conv2d_54.w_0" shape = [192, 1, 5, 5] dtype = "float32" min_val = float("-0.417018") @@ -2156,6 +2360,7 @@ class Program_weight_tensor_parameter_203: class Program_weight_tensor_parameter_204: name = "parameter_204" + original_name = "batch_norm2d_49.b_0" shape = [96] dtype = "float32" min_val = float("-0.243076") @@ -2167,6 +2372,7 @@ class Program_weight_tensor_parameter_204: class Program_weight_tensor_parameter_205: name = "parameter_205" + original_name = "batch_norm2d_49.w_0" shape = [96] dtype = "float32" min_val = float("0.0427605") @@ -2178,6 +2384,7 @@ class Program_weight_tensor_parameter_205: class Program_weight_tensor_parameter_206: name = "parameter_206" + original_name = "batch_norm2d_49.w_2" shape = [96] dtype = "float32" min_val = float("0.00174747") @@ -2189,6 +2396,7 @@ class Program_weight_tensor_parameter_206: class Program_weight_tensor_parameter_207: name = "parameter_207" + original_name = "batch_norm2d_49.w_1" shape = [96] dtype = "float32" min_val = float("-0.0498505") @@ -2200,6 +2408,7 @@ class Program_weight_tensor_parameter_207: class Program_weight_tensor_parameter_208: name = "parameter_208" + original_name = "conv2d_53.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.369655") @@ -2211,6 +2420,7 @@ class Program_weight_tensor_parameter_208: class Program_weight_tensor_parameter_209: name = "parameter_209" + original_name = "batch_norm2d_48.b_0" shape = [96] dtype = "float32" min_val = float("-0.424992") @@ -2222,6 +2432,7 @@ class Program_weight_tensor_parameter_209: class Program_weight_tensor_parameter_210: name = "parameter_210" + original_name = "batch_norm2d_48.w_0" shape = [96] dtype = "float32" min_val = float("0.114337") @@ -2233,6 +2444,7 @@ class Program_weight_tensor_parameter_210: class Program_weight_tensor_parameter_211: name = "parameter_211" + original_name = "batch_norm2d_48.w_2" shape = [96] dtype = "float32" min_val = float("0.000137999") @@ -2244,6 +2456,7 @@ class Program_weight_tensor_parameter_211: class Program_weight_tensor_parameter_212: name = "parameter_212" + original_name = "batch_norm2d_48.w_1" shape = [96] dtype = "float32" min_val = float("-0.719778") @@ -2255,6 +2468,7 @@ class Program_weight_tensor_parameter_212: class Program_weight_tensor_parameter_213: name = "parameter_213" + original_name = "conv2d_52.w_0" shape = [96, 1, 5, 5] dtype = "float32" min_val = float("-0.242408") @@ -2266,6 +2480,7 @@ class Program_weight_tensor_parameter_213: class Program_weight_tensor_parameter_214: name = "parameter_214" + original_name = "batch_norm2d_47.b_0" shape = [96] dtype = "float32" min_val = float("-1.55991") @@ -2277,6 +2492,7 @@ class Program_weight_tensor_parameter_214: class Program_weight_tensor_parameter_215: name = "parameter_215" + original_name = "batch_norm2d_47.w_0" shape = [96] dtype = "float32" min_val = float("0.0558965") @@ -2288,6 +2504,7 @@ class Program_weight_tensor_parameter_215: class Program_weight_tensor_parameter_216: name = "parameter_216" + original_name = "batch_norm2d_47.w_2" shape = [96] dtype = "float32" min_val = float("0.217084") @@ -2299,6 +2516,7 @@ class Program_weight_tensor_parameter_216: class Program_weight_tensor_parameter_217: name = "parameter_217" + original_name = "batch_norm2d_47.w_1" shape = [96] dtype = "float32" min_val = float("-1.09006") @@ -2310,6 +2528,7 @@ class Program_weight_tensor_parameter_217: class Program_weight_tensor_parameter_218: name = "parameter_218" + original_name = "conv2d_51.w_0" shape = [96, 192, 1, 1] dtype = "float32" min_val = float("-0.360032") @@ -2321,6 +2540,7 @@ class Program_weight_tensor_parameter_218: class Program_weight_tensor_parameter_219: name = "parameter_219" + original_name = "batch_norm2d_46.b_0" shape = [192] dtype = "float32" min_val = float("-1.12196") @@ -2332,6 +2552,7 @@ class Program_weight_tensor_parameter_219: class Program_weight_tensor_parameter_220: name = "parameter_220" + original_name = "batch_norm2d_46.w_0" shape = [192] dtype = "float32" min_val = float("0.550621") @@ -2343,6 +2564,7 @@ class Program_weight_tensor_parameter_220: class Program_weight_tensor_parameter_221: name = "parameter_221" + original_name = "batch_norm2d_46.w_2" shape = [192] dtype = "float32" min_val = float("0.00263449") @@ -2354,6 +2576,7 @@ class Program_weight_tensor_parameter_221: class Program_weight_tensor_parameter_222: name = "parameter_222" + original_name = "batch_norm2d_46.w_1" shape = [192] dtype = "float32" min_val = float("-0.472504") @@ -2365,6 +2588,7 @@ class Program_weight_tensor_parameter_222: class Program_weight_tensor_parameter_223: name = "parameter_223" + original_name = "conv2d_50.w_0" shape = [192, 1, 5, 5] dtype = "float32" min_val = float("-0.351348") @@ -2376,6 +2600,7 @@ class Program_weight_tensor_parameter_223: class Program_weight_tensor_parameter_224: name = "parameter_224" + original_name = "batch_norm2d_45.b_0" shape = [192] dtype = "float32" min_val = float("-1.39522") @@ -2387,6 +2612,7 @@ class Program_weight_tensor_parameter_224: class Program_weight_tensor_parameter_225: name = "parameter_225" + original_name = "batch_norm2d_45.w_0" shape = [192] dtype = "float32" min_val = float("0.262594") @@ -2398,6 +2624,7 @@ class Program_weight_tensor_parameter_225: class Program_weight_tensor_parameter_226: name = "parameter_226" + original_name = "batch_norm2d_45.w_2" shape = [192] dtype = "float32" min_val = float("0.108356") @@ -2409,6 +2636,7 @@ class Program_weight_tensor_parameter_226: class Program_weight_tensor_parameter_227: name = "parameter_227" + original_name = "batch_norm2d_45.w_1" shape = [192] dtype = "float32" min_val = float("-1.00023") @@ -2420,6 +2648,7 @@ class Program_weight_tensor_parameter_227: class Program_weight_tensor_parameter_228: name = "parameter_228" + original_name = "conv2d_49.w_0" shape = [192, 192, 1, 1] dtype = "float32" min_val = float("-0.394264") @@ -2431,6 +2660,7 @@ class Program_weight_tensor_parameter_228: class Program_weight_tensor_parameter_229: name = "parameter_229" + original_name = "batch_norm2d_44.b_0" shape = [192] dtype = "float32" min_val = float("-1.42668") @@ -2442,6 +2672,7 @@ class Program_weight_tensor_parameter_229: class Program_weight_tensor_parameter_230: name = "parameter_230" + original_name = "batch_norm2d_44.w_0" shape = [192] dtype = "float32" min_val = float("0.618592") @@ -2453,6 +2684,7 @@ class Program_weight_tensor_parameter_230: class Program_weight_tensor_parameter_231: name = "parameter_231" + original_name = "batch_norm2d_44.w_2" shape = [192] dtype = "float32" min_val = float("6.7304e-05") @@ -2464,6 +2696,7 @@ class Program_weight_tensor_parameter_231: class Program_weight_tensor_parameter_232: name = "parameter_232" + original_name = "batch_norm2d_44.w_1" shape = [192] dtype = "float32" min_val = float("-0.524774") @@ -2475,6 +2708,7 @@ class Program_weight_tensor_parameter_232: class Program_weight_tensor_parameter_233: name = "parameter_233" + original_name = "conv2d_48.w_0" shape = [192, 1, 5, 5] dtype = "float32" min_val = float("-0.470739") @@ -2486,6 +2720,7 @@ class Program_weight_tensor_parameter_233: class Program_weight_tensor_parameter_234: name = "parameter_234" + original_name = "batch_norm2d_43.b_0" shape = [96] dtype = "float32" min_val = float("-0.221691") @@ -2497,6 +2732,7 @@ class Program_weight_tensor_parameter_234: class Program_weight_tensor_parameter_235: name = "parameter_235" + original_name = "batch_norm2d_43.w_0" shape = [96] dtype = "float32" min_val = float("0.0333748") @@ -2508,6 +2744,7 @@ class Program_weight_tensor_parameter_235: class Program_weight_tensor_parameter_236: name = "parameter_236" + original_name = "batch_norm2d_43.w_2" shape = [96] dtype = "float32" min_val = float("0.00247242") @@ -2519,6 +2756,7 @@ class Program_weight_tensor_parameter_236: class Program_weight_tensor_parameter_237: name = "parameter_237" + original_name = "batch_norm2d_43.w_1" shape = [96] dtype = "float32" min_val = float("-0.0628314") @@ -2530,6 +2768,7 @@ class Program_weight_tensor_parameter_237: class Program_weight_tensor_parameter_238: name = "parameter_238" + original_name = "conv2d_47.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.429102") @@ -2541,6 +2780,7 @@ class Program_weight_tensor_parameter_238: class Program_weight_tensor_parameter_239: name = "parameter_239" + original_name = "batch_norm2d_42.b_0" shape = [96] dtype = "float32" min_val = float("-0.294481") @@ -2552,6 +2792,7 @@ class Program_weight_tensor_parameter_239: class Program_weight_tensor_parameter_240: name = "parameter_240" + original_name = "batch_norm2d_42.w_0" shape = [96] dtype = "float32" min_val = float("0.126566") @@ -2563,6 +2804,7 @@ class Program_weight_tensor_parameter_240: class Program_weight_tensor_parameter_241: name = "parameter_241" + original_name = "batch_norm2d_42.w_2" shape = [96] dtype = "float32" min_val = float("0.000638177") @@ -2574,6 +2816,7 @@ class Program_weight_tensor_parameter_241: class Program_weight_tensor_parameter_242: name = "parameter_242" + original_name = "batch_norm2d_42.w_1" shape = [96] dtype = "float32" min_val = float("-0.477663") @@ -2585,6 +2828,7 @@ class Program_weight_tensor_parameter_242: class Program_weight_tensor_parameter_243: name = "parameter_243" + original_name = "conv2d_46.w_0" shape = [96, 1, 5, 5] dtype = "float32" min_val = float("-0.227033") @@ -2596,6 +2840,7 @@ class Program_weight_tensor_parameter_243: class Program_weight_tensor_parameter_244: name = "parameter_244" + original_name = "batch_norm2d_41.b_0" shape = [96] dtype = "float32" min_val = float("-2.11923") @@ -2607,6 +2852,7 @@ class Program_weight_tensor_parameter_244: class Program_weight_tensor_parameter_245: name = "parameter_245" + original_name = "batch_norm2d_41.w_0" shape = [96] dtype = "float32" min_val = float("0.143089") @@ -2618,6 +2864,7 @@ class Program_weight_tensor_parameter_245: class Program_weight_tensor_parameter_246: name = "parameter_246" + original_name = "batch_norm2d_41.w_2" shape = [96] dtype = "float32" min_val = float("0.111723") @@ -2629,6 +2876,7 @@ class Program_weight_tensor_parameter_246: class Program_weight_tensor_parameter_247: name = "parameter_247" + original_name = "batch_norm2d_41.w_1" shape = [96] dtype = "float32" min_val = float("-1.01914") @@ -2640,6 +2888,7 @@ class Program_weight_tensor_parameter_247: class Program_weight_tensor_parameter_248: name = "parameter_248" + original_name = "conv2d_45.w_0" shape = [96, 192, 1, 1] dtype = "float32" min_val = float("-0.509249") @@ -2651,6 +2900,7 @@ class Program_weight_tensor_parameter_248: class Program_weight_tensor_parameter_249: name = "parameter_249" + original_name = "batch_norm2d_40.b_0" shape = [192] dtype = "float32" min_val = float("-1.44112") @@ -2662,6 +2912,7 @@ class Program_weight_tensor_parameter_249: class Program_weight_tensor_parameter_250: name = "parameter_250" + original_name = "batch_norm2d_40.w_0" shape = [192] dtype = "float32" min_val = float("0.311378") @@ -2673,6 +2924,7 @@ class Program_weight_tensor_parameter_250: class Program_weight_tensor_parameter_251: name = "parameter_251" + original_name = "batch_norm2d_40.w_2" shape = [192] dtype = "float32" min_val = float("7.05688e-05") @@ -2684,6 +2936,7 @@ class Program_weight_tensor_parameter_251: class Program_weight_tensor_parameter_252: name = "parameter_252" + original_name = "batch_norm2d_40.w_1" shape = [192] dtype = "float32" min_val = float("-0.477052") @@ -2695,6 +2948,7 @@ class Program_weight_tensor_parameter_252: class Program_weight_tensor_parameter_253: name = "parameter_253" + original_name = "conv2d_44.w_0" shape = [192, 1, 5, 5] dtype = "float32" min_val = float("-0.547887") @@ -2706,6 +2960,7 @@ class Program_weight_tensor_parameter_253: class Program_weight_tensor_parameter_254: name = "parameter_254" + original_name = "batch_norm2d_39.b_0" shape = [192] dtype = "float32" min_val = float("-2.03338") @@ -2717,6 +2972,7 @@ class Program_weight_tensor_parameter_254: class Program_weight_tensor_parameter_255: name = "parameter_255" + original_name = "batch_norm2d_39.w_0" shape = [192] dtype = "float32" min_val = float("0.0334618") @@ -2728,6 +2984,7 @@ class Program_weight_tensor_parameter_255: class Program_weight_tensor_parameter_256: name = "parameter_256" + original_name = "batch_norm2d_39.w_2" shape = [192] dtype = "float32" min_val = float("0.215802") @@ -2739,6 +2996,7 @@ class Program_weight_tensor_parameter_256: class Program_weight_tensor_parameter_257: name = "parameter_257" + original_name = "batch_norm2d_39.w_1" shape = [192] dtype = "float32" min_val = float("-1.80015") @@ -2750,6 +3008,7 @@ class Program_weight_tensor_parameter_257: class Program_weight_tensor_parameter_258: name = "parameter_258" + original_name = "conv2d_43.w_0" shape = [192, 192, 1, 1] dtype = "float32" min_val = float("-0.4358") @@ -2761,6 +3020,7 @@ class Program_weight_tensor_parameter_258: class Program_weight_tensor_parameter_259: name = "parameter_259" + original_name = "batch_norm2d_38.b_0" shape = [192] dtype = "float32" min_val = float("-1.22928") @@ -2772,6 +3032,7 @@ class Program_weight_tensor_parameter_259: class Program_weight_tensor_parameter_260: name = "parameter_260" + original_name = "batch_norm2d_38.w_0" shape = [192] dtype = "float32" min_val = float("0.53516") @@ -2783,6 +3044,7 @@ class Program_weight_tensor_parameter_260: class Program_weight_tensor_parameter_261: name = "parameter_261" + original_name = "batch_norm2d_38.w_2" shape = [192] dtype = "float32" min_val = float("3.95907e-05") @@ -2794,6 +3056,7 @@ class Program_weight_tensor_parameter_261: class Program_weight_tensor_parameter_262: name = "parameter_262" + original_name = "batch_norm2d_38.w_1" shape = [192] dtype = "float32" min_val = float("-0.782828") @@ -2805,6 +3068,7 @@ class Program_weight_tensor_parameter_262: class Program_weight_tensor_parameter_263: name = "parameter_263" + original_name = "conv2d_42.w_0" shape = [192, 1, 5, 5] dtype = "float32" min_val = float("-0.856061") @@ -2816,6 +3080,7 @@ class Program_weight_tensor_parameter_263: class Program_weight_tensor_parameter_264: name = "parameter_264" + original_name = "batch_norm2d_37.b_0" shape = [96] dtype = "float32" min_val = float("-0.837604") @@ -2827,6 +3092,7 @@ class Program_weight_tensor_parameter_264: class Program_weight_tensor_parameter_265: name = "parameter_265" + original_name = "batch_norm2d_37.w_0" shape = [96] dtype = "float32" min_val = float("0.0703397") @@ -2838,6 +3104,7 @@ class Program_weight_tensor_parameter_265: class Program_weight_tensor_parameter_266: name = "parameter_266" + original_name = "batch_norm2d_37.w_2" shape = [96] dtype = "float32" min_val = float("0.258882") @@ -2849,6 +3116,7 @@ class Program_weight_tensor_parameter_266: class Program_weight_tensor_parameter_267: name = "parameter_267" + original_name = "batch_norm2d_37.w_1" shape = [96] dtype = "float32" min_val = float("-1.00585") @@ -2860,6 +3128,7 @@ class Program_weight_tensor_parameter_267: class Program_weight_tensor_parameter_268: name = "parameter_268" + original_name = "conv2d_41.w_0" shape = [96, 192, 1, 1] dtype = "float32" min_val = float("-0.289291") @@ -2871,6 +3140,7 @@ class Program_weight_tensor_parameter_268: class Program_weight_tensor_parameter_269: name = "parameter_269" + original_name = "batch_norm2d_36.b_0" shape = [192] dtype = "float32" min_val = float("-1.08735") @@ -2882,6 +3152,7 @@ class Program_weight_tensor_parameter_269: class Program_weight_tensor_parameter_270: name = "parameter_270" + original_name = "batch_norm2d_36.w_0" shape = [192] dtype = "float32" min_val = float("0.554844") @@ -2893,6 +3164,7 @@ class Program_weight_tensor_parameter_270: class Program_weight_tensor_parameter_271: name = "parameter_271" + original_name = "batch_norm2d_36.w_2" shape = [192] dtype = "float32" min_val = float("0.00146287") @@ -2904,6 +3176,7 @@ class Program_weight_tensor_parameter_271: class Program_weight_tensor_parameter_272: name = "parameter_272" + original_name = "batch_norm2d_36.w_1" shape = [192] dtype = "float32" min_val = float("-0.570966") @@ -2915,6 +3188,7 @@ class Program_weight_tensor_parameter_272: class Program_weight_tensor_parameter_273: name = "parameter_273" + original_name = "conv2d_40.w_0" shape = [192, 1, 5, 5] dtype = "float32" min_val = float("-0.485553") @@ -2926,6 +3200,7 @@ class Program_weight_tensor_parameter_273: class Program_weight_tensor_parameter_274: name = "parameter_274" + original_name = "batch_norm2d_35.b_0" shape = [192] dtype = "float32" min_val = float("-1.31088") @@ -2937,6 +3212,7 @@ class Program_weight_tensor_parameter_274: class Program_weight_tensor_parameter_275: name = "parameter_275" + original_name = "batch_norm2d_35.w_0" shape = [192] dtype = "float32" min_val = float("0.13446") @@ -2948,6 +3224,7 @@ class Program_weight_tensor_parameter_275: class Program_weight_tensor_parameter_276: name = "parameter_276" + original_name = "batch_norm2d_35.w_2" shape = [192] dtype = "float32" min_val = float("0.257016") @@ -2959,6 +3236,7 @@ class Program_weight_tensor_parameter_276: class Program_weight_tensor_parameter_277: name = "parameter_277" + original_name = "batch_norm2d_35.w_1" shape = [192] dtype = "float32" min_val = float("-1.8917") @@ -2970,6 +3248,7 @@ class Program_weight_tensor_parameter_277: class Program_weight_tensor_parameter_278: name = "parameter_278" + original_name = "conv2d_39.w_0" shape = [192, 192, 1, 1] dtype = "float32" min_val = float("-0.354697") @@ -2981,6 +3260,7 @@ class Program_weight_tensor_parameter_278: class Program_weight_tensor_parameter_279: name = "parameter_279" + original_name = "batch_norm2d_34.b_0" shape = [192] dtype = "float32" min_val = float("-1.26396") @@ -2992,6 +3272,7 @@ class Program_weight_tensor_parameter_279: class Program_weight_tensor_parameter_280: name = "parameter_280" + original_name = "batch_norm2d_34.w_0" shape = [192] dtype = "float32" min_val = float("0.621223") @@ -3003,6 +3284,7 @@ class Program_weight_tensor_parameter_280: class Program_weight_tensor_parameter_281: name = "parameter_281" + original_name = "batch_norm2d_34.w_2" shape = [192] dtype = "float32" min_val = float("3.68007e-05") @@ -3014,6 +3296,7 @@ class Program_weight_tensor_parameter_281: class Program_weight_tensor_parameter_282: name = "parameter_282" + original_name = "batch_norm2d_34.w_1" shape = [192] dtype = "float32" min_val = float("-0.102347") @@ -3025,6 +3308,7 @@ class Program_weight_tensor_parameter_282: class Program_weight_tensor_parameter_283: name = "parameter_283" + original_name = "conv2d_38.w_0" shape = [192, 1, 5, 5] dtype = "float32" min_val = float("-0.51549") @@ -3036,6 +3320,7 @@ class Program_weight_tensor_parameter_283: class Program_weight_tensor_parameter_284: name = "parameter_284" + original_name = "batch_norm2d_29.b_0" shape = [96] dtype = "float32" min_val = float("-0.150208") @@ -3047,6 +3332,7 @@ class Program_weight_tensor_parameter_284: class Program_weight_tensor_parameter_285: name = "parameter_285" + original_name = "batch_norm2d_29.w_0" shape = [96] dtype = "float32" min_val = float("0.0505571") @@ -3058,6 +3344,7 @@ class Program_weight_tensor_parameter_285: class Program_weight_tensor_parameter_286: name = "parameter_286" + original_name = "batch_norm2d_29.w_2" shape = [96] dtype = "float32" min_val = float("3.41699") @@ -3069,6 +3356,7 @@ class Program_weight_tensor_parameter_286: class Program_weight_tensor_parameter_287: name = "parameter_287" + original_name = "batch_norm2d_29.w_1" shape = [96] dtype = "float32" min_val = float("-3.6074") @@ -3080,6 +3368,7 @@ class Program_weight_tensor_parameter_287: class Program_weight_tensor_parameter_288: name = "parameter_288" + original_name = "conv2d_33.w_0" shape = [96, 176, 1, 1] dtype = "float32" min_val = float("-0.321192") @@ -3091,6 +3380,7 @@ class Program_weight_tensor_parameter_288: class Program_weight_tensor_parameter_289: name = "parameter_289" + original_name = "batch_norm2d_28.b_0" shape = [96] dtype = "float32" min_val = float("-0.0887339") @@ -3102,6 +3392,7 @@ class Program_weight_tensor_parameter_289: class Program_weight_tensor_parameter_290: name = "parameter_290" + original_name = "batch_norm2d_28.w_0" shape = [96] dtype = "float32" min_val = float("0.0361377") @@ -3113,6 +3404,7 @@ class Program_weight_tensor_parameter_290: class Program_weight_tensor_parameter_291: name = "parameter_291" + original_name = "batch_norm2d_28.w_2" shape = [96] dtype = "float32" min_val = float("0.088848") @@ -3124,6 +3416,7 @@ class Program_weight_tensor_parameter_291: class Program_weight_tensor_parameter_292: name = "parameter_292" + original_name = "batch_norm2d_28.w_1" shape = [96] dtype = "float32" min_val = float("-1.36184") @@ -3135,6 +3428,7 @@ class Program_weight_tensor_parameter_292: class Program_weight_tensor_parameter_293: name = "parameter_293" + original_name = "conv2d_32.w_0" shape = [96, 88, 1, 1] dtype = "float32" min_val = float("-0.291685") @@ -3146,6 +3440,7 @@ class Program_weight_tensor_parameter_293: class Program_weight_tensor_parameter_294: name = "parameter_294" + original_name = "batch_norm2d_27.b_0" shape = [96] dtype = "float32" min_val = float("-0.470952") @@ -3157,6 +3452,7 @@ class Program_weight_tensor_parameter_294: class Program_weight_tensor_parameter_295: name = "parameter_295" + original_name = "batch_norm2d_27.w_0" shape = [96] dtype = "float32" min_val = float("0.0274791") @@ -3168,6 +3464,7 @@ class Program_weight_tensor_parameter_295: class Program_weight_tensor_parameter_296: name = "parameter_296" + original_name = "batch_norm2d_27.w_2" shape = [96] dtype = "float32" min_val = float("0.0874587") @@ -3179,6 +3476,7 @@ class Program_weight_tensor_parameter_296: class Program_weight_tensor_parameter_297: name = "parameter_297" + original_name = "batch_norm2d_27.w_1" shape = [96] dtype = "float32" min_val = float("-1.75999") @@ -3190,6 +3488,7 @@ class Program_weight_tensor_parameter_297: class Program_weight_tensor_parameter_298: name = "parameter_298" + original_name = "conv2d_31.w_0" shape = [96, 48, 1, 1] dtype = "float32" min_val = float("-0.363586") @@ -3201,6 +3500,7 @@ class Program_weight_tensor_parameter_298: class Program_weight_tensor_parameter_299: name = "parameter_299" + original_name = "batch_norm2d_26.b_0" shape = [176] dtype = "float32" min_val = float("-9.11118") @@ -3212,6 +3512,7 @@ class Program_weight_tensor_parameter_299: class Program_weight_tensor_parameter_300: name = "parameter_300" + original_name = "batch_norm2d_26.w_0" shape = [176] dtype = "float32" min_val = float("3.5604") @@ -3223,6 +3524,7 @@ class Program_weight_tensor_parameter_300: class Program_weight_tensor_parameter_301: name = "parameter_301" + original_name = "batch_norm2d_26.w_2" shape = [176] dtype = "float32" min_val = float("0.0246136") @@ -3234,6 +3536,7 @@ class Program_weight_tensor_parameter_301: class Program_weight_tensor_parameter_302: name = "parameter_302" + original_name = "batch_norm2d_26.w_1" shape = [176] dtype = "float32" min_val = float("-0.446209") @@ -3245,6 +3548,7 @@ class Program_weight_tensor_parameter_302: class Program_weight_tensor_parameter_303: name = "parameter_303" + original_name = "conv2d_30.w_0" shape = [176, 176, 1, 1] dtype = "float32" min_val = float("-0.687882") @@ -3256,6 +3560,7 @@ class Program_weight_tensor_parameter_303: class Program_weight_tensor_parameter_304: name = "parameter_304" + original_name = "conv2d_29.b_0" shape = [176] dtype = "float32" min_val = float("-0.150748") @@ -3267,6 +3572,7 @@ class Program_weight_tensor_parameter_304: class Program_weight_tensor_parameter_305: name = "parameter_305" + original_name = "conv2d_29.w_0" shape = [176, 44, 1, 1] dtype = "float32" min_val = float("-0.621958") @@ -3278,6 +3584,7 @@ class Program_weight_tensor_parameter_305: class Program_weight_tensor_parameter_306: name = "parameter_306" + original_name = "conv2d_28.b_0" shape = [44] dtype = "float32" min_val = float("0") @@ -3287,6 +3594,7 @@ class Program_weight_tensor_parameter_306: class Program_weight_tensor_parameter_307: name = "parameter_307" + original_name = "conv2d_28.w_0" shape = [44, 176, 1, 1] dtype = "float32" min_val = float("-0.625018") @@ -3298,6 +3606,7 @@ class Program_weight_tensor_parameter_307: class Program_weight_tensor_parameter_308: name = "parameter_308" + original_name = "batch_norm2d_25.b_0" shape = [176] dtype = "float32" min_val = float("-1.71231") @@ -3309,6 +3618,7 @@ class Program_weight_tensor_parameter_308: class Program_weight_tensor_parameter_309: name = "parameter_309" + original_name = "batch_norm2d_25.w_0" shape = [176] dtype = "float32" min_val = float("-0.843924") @@ -3320,6 +3630,7 @@ class Program_weight_tensor_parameter_309: class Program_weight_tensor_parameter_310: name = "parameter_310" + original_name = "batch_norm2d_25.w_2" shape = [176] dtype = "float32" min_val = float("0.000142262") @@ -3331,6 +3642,7 @@ class Program_weight_tensor_parameter_310: class Program_weight_tensor_parameter_311: name = "parameter_311" + original_name = "batch_norm2d_25.w_1" shape = [176] dtype = "float32" min_val = float("-5.35414") @@ -3342,6 +3654,7 @@ class Program_weight_tensor_parameter_311: class Program_weight_tensor_parameter_312: name = "parameter_312" + original_name = "conv2d_27.w_0" shape = [176, 1, 5, 5] dtype = "float32" min_val = float("-0.586937") @@ -3353,6 +3666,7 @@ class Program_weight_tensor_parameter_312: class Program_weight_tensor_parameter_313: name = "parameter_313" + original_name = "batch_norm2d_24.b_0" shape = [176] dtype = "float32" min_val = float("-2.35887") @@ -3364,6 +3678,7 @@ class Program_weight_tensor_parameter_313: class Program_weight_tensor_parameter_314: name = "parameter_314" + original_name = "batch_norm2d_24.w_0" shape = [176] dtype = "float32" min_val = float("-0.301818") @@ -3375,6 +3690,7 @@ class Program_weight_tensor_parameter_314: class Program_weight_tensor_parameter_315: name = "parameter_315" + original_name = "batch_norm2d_24.w_2" shape = [176] dtype = "float32" min_val = float("0.0379402") @@ -3386,6 +3702,7 @@ class Program_weight_tensor_parameter_315: class Program_weight_tensor_parameter_316: name = "parameter_316" + original_name = "batch_norm2d_24.w_1" shape = [176] dtype = "float32" min_val = float("-0.978587") @@ -3397,6 +3714,7 @@ class Program_weight_tensor_parameter_316: class Program_weight_tensor_parameter_317: name = "parameter_317" + original_name = "conv2d_26.w_0" shape = [176, 88, 1, 1] dtype = "float32" min_val = float("-0.488273") @@ -3408,6 +3726,7 @@ class Program_weight_tensor_parameter_317: class Program_weight_tensor_parameter_318: name = "parameter_318" + original_name = "conv2d_25.b_0" shape = [88] dtype = "float32" min_val = float("-0.140884") @@ -3419,6 +3738,7 @@ class Program_weight_tensor_parameter_318: class Program_weight_tensor_parameter_319: name = "parameter_319" + original_name = "conv2d_25.w_0" shape = [88, 22, 1, 1] dtype = "float32" min_val = float("-0.597084") @@ -3430,6 +3750,7 @@ class Program_weight_tensor_parameter_319: class Program_weight_tensor_parameter_320: name = "parameter_320" + original_name = "conv2d_24.b_0" shape = [22] dtype = "float32" min_val = float("0") @@ -3439,6 +3760,7 @@ class Program_weight_tensor_parameter_320: class Program_weight_tensor_parameter_321: name = "parameter_321" + original_name = "conv2d_24.w_0" shape = [22, 88, 1, 1] dtype = "float32" min_val = float("-0.548668") @@ -3450,6 +3772,7 @@ class Program_weight_tensor_parameter_321: class Program_weight_tensor_parameter_322: name = "parameter_322" + original_name = "batch_norm2d_23.b_0" shape = [88] dtype = "float32" min_val = float("-1.24324") @@ -3461,6 +3784,7 @@ class Program_weight_tensor_parameter_322: class Program_weight_tensor_parameter_323: name = "parameter_323" + original_name = "batch_norm2d_23.w_0" shape = [88] dtype = "float32" min_val = float("-1.13631") @@ -3472,6 +3796,7 @@ class Program_weight_tensor_parameter_323: class Program_weight_tensor_parameter_324: name = "parameter_324" + original_name = "batch_norm2d_23.w_2" shape = [88] dtype = "float32" min_val = float("0.000182247") @@ -3483,6 +3808,7 @@ class Program_weight_tensor_parameter_324: class Program_weight_tensor_parameter_325: name = "parameter_325" + original_name = "batch_norm2d_23.w_1" shape = [88] dtype = "float32" min_val = float("-6.44318") @@ -3494,6 +3820,7 @@ class Program_weight_tensor_parameter_325: class Program_weight_tensor_parameter_326: name = "parameter_326" + original_name = "conv2d_23.w_0" shape = [88, 1, 5, 5] dtype = "float32" min_val = float("-0.238176") @@ -3505,6 +3832,7 @@ class Program_weight_tensor_parameter_326: class Program_weight_tensor_parameter_327: name = "parameter_327" + original_name = "batch_norm2d_22.b_0" shape = [88] dtype = "float32" min_val = float("-2.06879") @@ -3516,6 +3844,7 @@ class Program_weight_tensor_parameter_327: class Program_weight_tensor_parameter_328: name = "parameter_328" + original_name = "batch_norm2d_22.w_0" shape = [88] dtype = "float32" min_val = float("-1.68314") @@ -3527,6 +3856,7 @@ class Program_weight_tensor_parameter_328: class Program_weight_tensor_parameter_329: name = "parameter_329" + original_name = "batch_norm2d_22.w_2" shape = [88] dtype = "float32" min_val = float("0.338223") @@ -3538,6 +3868,7 @@ class Program_weight_tensor_parameter_329: class Program_weight_tensor_parameter_330: name = "parameter_330" + original_name = "batch_norm2d_22.w_1" shape = [88] dtype = "float32" min_val = float("-4.8749") @@ -3549,6 +3880,7 @@ class Program_weight_tensor_parameter_330: class Program_weight_tensor_parameter_331: name = "parameter_331" + original_name = "conv2d_22.w_0" shape = [88, 88, 1, 1] dtype = "float32" min_val = float("-0.945255") @@ -3560,6 +3892,7 @@ class Program_weight_tensor_parameter_331: class Program_weight_tensor_parameter_332: name = "parameter_332" + original_name = "batch_norm2d_21.b_0" shape = [88] dtype = "float32" min_val = float("-0.899953") @@ -3571,6 +3904,7 @@ class Program_weight_tensor_parameter_332: class Program_weight_tensor_parameter_333: name = "parameter_333" + original_name = "batch_norm2d_21.w_0" shape = [88] dtype = "float32" min_val = float("0.401066") @@ -3582,6 +3916,7 @@ class Program_weight_tensor_parameter_333: class Program_weight_tensor_parameter_334: name = "parameter_334" + original_name = "batch_norm2d_21.w_2" shape = [88] dtype = "float32" min_val = float("0.00026083") @@ -3593,6 +3928,7 @@ class Program_weight_tensor_parameter_334: class Program_weight_tensor_parameter_335: name = "parameter_335" + original_name = "batch_norm2d_21.w_1" shape = [88] dtype = "float32" min_val = float("-15.9765") @@ -3604,6 +3940,7 @@ class Program_weight_tensor_parameter_335: class Program_weight_tensor_parameter_336: name = "parameter_336" + original_name = "conv2d_21.w_0" shape = [88, 1, 5, 5] dtype = "float32" min_val = float("-0.828405") @@ -3615,6 +3952,7 @@ class Program_weight_tensor_parameter_336: class Program_weight_tensor_parameter_337: name = "parameter_337" + original_name = "batch_norm2d_20.b_0" shape = [88] dtype = "float32" min_val = float("-2.24684") @@ -3626,6 +3964,7 @@ class Program_weight_tensor_parameter_337: class Program_weight_tensor_parameter_338: name = "parameter_338" + original_name = "batch_norm2d_20.w_0" shape = [88] dtype = "float32" min_val = float("-0.943654") @@ -3637,6 +3976,7 @@ class Program_weight_tensor_parameter_338: class Program_weight_tensor_parameter_339: name = "parameter_339" + original_name = "batch_norm2d_20.w_2" shape = [88] dtype = "float32" min_val = float("0.486818") @@ -3648,6 +3988,7 @@ class Program_weight_tensor_parameter_339: class Program_weight_tensor_parameter_340: name = "parameter_340" + original_name = "batch_norm2d_20.w_1" shape = [88] dtype = "float32" min_val = float("-6.83759") @@ -3659,6 +4000,7 @@ class Program_weight_tensor_parameter_340: class Program_weight_tensor_parameter_341: name = "parameter_341" + original_name = "conv2d_20.w_0" shape = [88, 88, 1, 1] dtype = "float32" min_val = float("-0.592394") @@ -3670,6 +4012,7 @@ class Program_weight_tensor_parameter_341: class Program_weight_tensor_parameter_342: name = "parameter_342" + original_name = "batch_norm2d_19.b_0" shape = [88] dtype = "float32" min_val = float("-1.85347") @@ -3681,6 +4024,7 @@ class Program_weight_tensor_parameter_342: class Program_weight_tensor_parameter_343: name = "parameter_343" + original_name = "batch_norm2d_19.w_0" shape = [88] dtype = "float32" min_val = float("-1.03769") @@ -3692,6 +4036,7 @@ class Program_weight_tensor_parameter_343: class Program_weight_tensor_parameter_344: name = "parameter_344" + original_name = "batch_norm2d_19.w_2" shape = [88] dtype = "float32" min_val = float("0.000235759") @@ -3703,6 +4048,7 @@ class Program_weight_tensor_parameter_344: class Program_weight_tensor_parameter_345: name = "parameter_345" + original_name = "batch_norm2d_19.w_1" shape = [88] dtype = "float32" min_val = float("-9.86858") @@ -3714,6 +4060,7 @@ class Program_weight_tensor_parameter_345: class Program_weight_tensor_parameter_346: name = "parameter_346" + original_name = "conv2d_19.w_0" shape = [88, 1, 5, 5] dtype = "float32" min_val = float("-0.865873") @@ -3725,6 +4072,7 @@ class Program_weight_tensor_parameter_346: class Program_weight_tensor_parameter_347: name = "parameter_347" + original_name = "batch_norm2d_18.b_0" shape = [88] dtype = "float32" min_val = float("-1.75167") @@ -3736,6 +4084,7 @@ class Program_weight_tensor_parameter_347: class Program_weight_tensor_parameter_348: name = "parameter_348" + original_name = "batch_norm2d_18.w_0" shape = [88] dtype = "float32" min_val = float("-1.13239") @@ -3747,6 +4096,7 @@ class Program_weight_tensor_parameter_348: class Program_weight_tensor_parameter_349: name = "parameter_349" + original_name = "batch_norm2d_18.w_2" shape = [88] dtype = "float32" min_val = float("0.47364") @@ -3758,6 +4108,7 @@ class Program_weight_tensor_parameter_349: class Program_weight_tensor_parameter_350: name = "parameter_350" + original_name = "batch_norm2d_18.w_1" shape = [88] dtype = "float32" min_val = float("-5.66201") @@ -3769,6 +4120,7 @@ class Program_weight_tensor_parameter_350: class Program_weight_tensor_parameter_351: name = "parameter_351" + original_name = "conv2d_18.w_0" shape = [88, 88, 1, 1] dtype = "float32" min_val = float("-0.688069") @@ -3780,6 +4132,7 @@ class Program_weight_tensor_parameter_351: class Program_weight_tensor_parameter_352: name = "parameter_352" + original_name = "batch_norm2d_17.b_0" shape = [88] dtype = "float32" min_val = float("-1.46931") @@ -3791,6 +4144,7 @@ class Program_weight_tensor_parameter_352: class Program_weight_tensor_parameter_353: name = "parameter_353" + original_name = "batch_norm2d_17.w_0" shape = [88] dtype = "float32" min_val = float("0.548915") @@ -3802,6 +4156,7 @@ class Program_weight_tensor_parameter_353: class Program_weight_tensor_parameter_354: name = "parameter_354" + original_name = "batch_norm2d_17.w_2" shape = [88] dtype = "float32" min_val = float("0.000155686") @@ -3813,6 +4168,7 @@ class Program_weight_tensor_parameter_354: class Program_weight_tensor_parameter_355: name = "parameter_355" + original_name = "batch_norm2d_17.w_1" shape = [88] dtype = "float32" min_val = float("-11.3318") @@ -3824,6 +4180,7 @@ class Program_weight_tensor_parameter_355: class Program_weight_tensor_parameter_356: name = "parameter_356" + original_name = "conv2d_17.w_0" shape = [88, 1, 5, 5] dtype = "float32" min_val = float("-0.86108") @@ -3835,6 +4192,7 @@ class Program_weight_tensor_parameter_356: class Program_weight_tensor_parameter_357: name = "parameter_357" + original_name = "batch_norm2d_16.b_0" shape = [88] dtype = "float32" min_val = float("-1.97382") @@ -3846,6 +4204,7 @@ class Program_weight_tensor_parameter_357: class Program_weight_tensor_parameter_358: name = "parameter_358" + original_name = "batch_norm2d_16.w_0" shape = [88] dtype = "float32" min_val = float("-0.210855") @@ -3857,6 +4216,7 @@ class Program_weight_tensor_parameter_358: class Program_weight_tensor_parameter_359: name = "parameter_359" + original_name = "batch_norm2d_16.w_2" shape = [88] dtype = "float32" min_val = float("0.339745") @@ -3868,6 +4228,7 @@ class Program_weight_tensor_parameter_359: class Program_weight_tensor_parameter_360: name = "parameter_360" + original_name = "batch_norm2d_16.w_1" shape = [88] dtype = "float32" min_val = float("-6.64902") @@ -3879,6 +4240,7 @@ class Program_weight_tensor_parameter_360: class Program_weight_tensor_parameter_361: name = "parameter_361" + original_name = "conv2d_16.w_0" shape = [88, 88, 1, 1] dtype = "float32" min_val = float("-0.776332") @@ -3890,6 +4252,7 @@ class Program_weight_tensor_parameter_361: class Program_weight_tensor_parameter_362: name = "parameter_362" + original_name = "batch_norm2d_15.b_0" shape = [88] dtype = "float32" min_val = float("-1.53342") @@ -3901,6 +4264,7 @@ class Program_weight_tensor_parameter_362: class Program_weight_tensor_parameter_363: name = "parameter_363" + original_name = "batch_norm2d_15.w_0" shape = [88] dtype = "float32" min_val = float("0.413294") @@ -3912,6 +4276,7 @@ class Program_weight_tensor_parameter_363: class Program_weight_tensor_parameter_364: name = "parameter_364" + original_name = "batch_norm2d_15.w_2" shape = [88] dtype = "float32" min_val = float("0.000399") @@ -3923,6 +4288,7 @@ class Program_weight_tensor_parameter_364: class Program_weight_tensor_parameter_365: name = "parameter_365" + original_name = "batch_norm2d_15.w_1" shape = [88] dtype = "float32" min_val = float("-47.3967") @@ -3934,6 +4300,7 @@ class Program_weight_tensor_parameter_365: class Program_weight_tensor_parameter_366: name = "parameter_366" + original_name = "conv2d_15.w_0" shape = [88, 1, 5, 5] dtype = "float32" min_val = float("-1.18891") @@ -3945,6 +4312,7 @@ class Program_weight_tensor_parameter_366: class Program_weight_tensor_parameter_367: name = "parameter_367" + original_name = "batch_norm2d_14.b_0" shape = [88] dtype = "float32" min_val = float("-1.64683") @@ -3956,6 +4324,7 @@ class Program_weight_tensor_parameter_367: class Program_weight_tensor_parameter_368: name = "parameter_368" + original_name = "batch_norm2d_14.w_0" shape = [88] dtype = "float32" min_val = float("-1.29092") @@ -3967,6 +4336,7 @@ class Program_weight_tensor_parameter_368: class Program_weight_tensor_parameter_369: name = "parameter_369" + original_name = "batch_norm2d_14.w_2" shape = [88] dtype = "float32" min_val = float("0.48465") @@ -3978,6 +4348,7 @@ class Program_weight_tensor_parameter_369: class Program_weight_tensor_parameter_370: name = "parameter_370" + original_name = "batch_norm2d_14.w_1" shape = [88] dtype = "float32" min_val = float("-13.3093") @@ -3989,6 +4360,7 @@ class Program_weight_tensor_parameter_370: class Program_weight_tensor_parameter_371: name = "parameter_371" + original_name = "conv2d_14.w_0" shape = [88, 88, 1, 1] dtype = "float32" min_val = float("-0.988296") @@ -4000,6 +4372,7 @@ class Program_weight_tensor_parameter_371: class Program_weight_tensor_parameter_372: name = "parameter_372" + original_name = "batch_norm2d_13.b_0" shape = [88] dtype = "float32" min_val = float("-2.36244") @@ -4011,6 +4384,7 @@ class Program_weight_tensor_parameter_372: class Program_weight_tensor_parameter_373: name = "parameter_373" + original_name = "batch_norm2d_13.w_0" shape = [88] dtype = "float32" min_val = float("0.622713") @@ -4022,6 +4396,7 @@ class Program_weight_tensor_parameter_373: class Program_weight_tensor_parameter_374: name = "parameter_374" + original_name = "batch_norm2d_13.w_2" shape = [88] dtype = "float32" min_val = float("0.000162941") @@ -4033,6 +4408,7 @@ class Program_weight_tensor_parameter_374: class Program_weight_tensor_parameter_375: name = "parameter_375" + original_name = "batch_norm2d_13.w_1" shape = [88] dtype = "float32" min_val = float("-5.74536") @@ -4044,6 +4420,7 @@ class Program_weight_tensor_parameter_375: class Program_weight_tensor_parameter_376: name = "parameter_376" + original_name = "conv2d_13.w_0" shape = [88, 1, 5, 5] dtype = "float32" min_val = float("-1.04311") @@ -4055,6 +4432,7 @@ class Program_weight_tensor_parameter_376: class Program_weight_tensor_parameter_377: name = "parameter_377" + original_name = "batch_norm2d_12.b_0" shape = [88] dtype = "float32" min_val = float("-1.61996") @@ -4066,6 +4444,7 @@ class Program_weight_tensor_parameter_377: class Program_weight_tensor_parameter_378: name = "parameter_378" + original_name = "batch_norm2d_12.w_0" shape = [88] dtype = "float32" min_val = float("-0.905796") @@ -4077,6 +4456,7 @@ class Program_weight_tensor_parameter_378: class Program_weight_tensor_parameter_379: name = "parameter_379" + original_name = "batch_norm2d_12.w_2" shape = [88] dtype = "float32" min_val = float("0.1438") @@ -4088,6 +4468,7 @@ class Program_weight_tensor_parameter_379: class Program_weight_tensor_parameter_380: name = "parameter_380" + original_name = "batch_norm2d_12.w_1" shape = [88] dtype = "float32" min_val = float("-9.26231") @@ -4099,6 +4480,7 @@ class Program_weight_tensor_parameter_380: class Program_weight_tensor_parameter_381: name = "parameter_381" + original_name = "conv2d_12.w_0" shape = [88, 48, 1, 1] dtype = "float32" min_val = float("-0.848459") @@ -4110,6 +4492,7 @@ class Program_weight_tensor_parameter_381: class Program_weight_tensor_parameter_382: name = "parameter_382" + original_name = "batch_norm2d_11.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -4119,6 +4502,7 @@ class Program_weight_tensor_parameter_382: class Program_weight_tensor_parameter_383: name = "parameter_383" + original_name = "batch_norm2d_11.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -4128,6 +4512,7 @@ class Program_weight_tensor_parameter_383: class Program_weight_tensor_parameter_384: name = "parameter_384" + original_name = "batch_norm2d_11.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -4137,6 +4522,7 @@ class Program_weight_tensor_parameter_384: class Program_weight_tensor_parameter_385: name = "parameter_385" + original_name = "batch_norm2d_11.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -4146,6 +4532,7 @@ class Program_weight_tensor_parameter_385: class Program_weight_tensor_parameter_386: name = "parameter_386" + original_name = "conv2d_11.w_0" shape = [48, 1, 3, 3] dtype = "float32" min_val = float("-0.299736") @@ -4157,6 +4544,7 @@ class Program_weight_tensor_parameter_386: class Program_weight_tensor_parameter_387: name = "parameter_387" + original_name = "batch_norm2d_10.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -4166,6 +4554,7 @@ class Program_weight_tensor_parameter_387: class Program_weight_tensor_parameter_388: name = "parameter_388" + original_name = "batch_norm2d_10.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -4175,6 +4564,7 @@ class Program_weight_tensor_parameter_388: class Program_weight_tensor_parameter_389: name = "parameter_389" + original_name = "batch_norm2d_10.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -4184,6 +4574,7 @@ class Program_weight_tensor_parameter_389: class Program_weight_tensor_parameter_390: name = "parameter_390" + original_name = "batch_norm2d_10.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -4193,6 +4584,7 @@ class Program_weight_tensor_parameter_390: class Program_weight_tensor_parameter_391: name = "parameter_391" + original_name = "conv2d_10.w_0" shape = [48, 48, 1, 1] dtype = "float32" min_val = float("-1.09538") @@ -4204,6 +4596,7 @@ class Program_weight_tensor_parameter_391: class Program_weight_tensor_parameter_392: name = "parameter_392" + original_name = "batch_norm2d_9.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -4213,6 +4606,7 @@ class Program_weight_tensor_parameter_392: class Program_weight_tensor_parameter_393: name = "parameter_393" + original_name = "batch_norm2d_9.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -4222,6 +4616,7 @@ class Program_weight_tensor_parameter_393: class Program_weight_tensor_parameter_394: name = "parameter_394" + original_name = "batch_norm2d_9.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -4231,6 +4626,7 @@ class Program_weight_tensor_parameter_394: class Program_weight_tensor_parameter_395: name = "parameter_395" + original_name = "batch_norm2d_9.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -4240,6 +4636,7 @@ class Program_weight_tensor_parameter_395: class Program_weight_tensor_parameter_396: name = "parameter_396" + original_name = "conv2d_9.w_0" shape = [48, 1, 3, 3] dtype = "float32" min_val = float("-1.13349") @@ -4251,6 +4648,7 @@ class Program_weight_tensor_parameter_396: class Program_weight_tensor_parameter_397: name = "parameter_397" + original_name = "batch_norm2d_8.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -4260,6 +4658,7 @@ class Program_weight_tensor_parameter_397: class Program_weight_tensor_parameter_398: name = "parameter_398" + original_name = "batch_norm2d_8.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -4269,6 +4668,7 @@ class Program_weight_tensor_parameter_398: class Program_weight_tensor_parameter_399: name = "parameter_399" + original_name = "batch_norm2d_8.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -4278,6 +4678,7 @@ class Program_weight_tensor_parameter_399: class Program_weight_tensor_parameter_400: name = "parameter_400" + original_name = "batch_norm2d_8.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -4287,6 +4688,7 @@ class Program_weight_tensor_parameter_400: class Program_weight_tensor_parameter_401: name = "parameter_401" + original_name = "conv2d_8.w_0" shape = [48, 24, 1, 1] dtype = "float32" min_val = float("-1.19438") @@ -4298,6 +4700,7 @@ class Program_weight_tensor_parameter_401: class Program_weight_tensor_parameter_402: name = "parameter_402" + original_name = "batch_norm2d_7.b_0" shape = [24] dtype = "float32" min_val = float("0") @@ -4307,6 +4710,7 @@ class Program_weight_tensor_parameter_402: class Program_weight_tensor_parameter_403: name = "parameter_403" + original_name = "batch_norm2d_7.w_0" shape = [24] dtype = "float32" min_val = float("0") @@ -4316,6 +4720,7 @@ class Program_weight_tensor_parameter_403: class Program_weight_tensor_parameter_404: name = "parameter_404" + original_name = "batch_norm2d_7.w_2" shape = [24] dtype = "float32" min_val = float("0") @@ -4325,6 +4730,7 @@ class Program_weight_tensor_parameter_404: class Program_weight_tensor_parameter_405: name = "parameter_405" + original_name = "batch_norm2d_7.w_1" shape = [24] dtype = "float32" min_val = float("0") @@ -4334,6 +4740,7 @@ class Program_weight_tensor_parameter_405: class Program_weight_tensor_parameter_406: name = "parameter_406" + original_name = "conv2d_7.w_0" shape = [24, 1, 3, 3] dtype = "float32" min_val = float("-0.445812") @@ -4345,6 +4752,7 @@ class Program_weight_tensor_parameter_406: class Program_weight_tensor_parameter_407: name = "parameter_407" + original_name = "batch_norm2d_6.b_0" shape = [24] dtype = "float32" min_val = float("0") @@ -4354,6 +4762,7 @@ class Program_weight_tensor_parameter_407: class Program_weight_tensor_parameter_408: name = "parameter_408" + original_name = "batch_norm2d_6.w_0" shape = [24] dtype = "float32" min_val = float("0") @@ -4363,6 +4772,7 @@ class Program_weight_tensor_parameter_408: class Program_weight_tensor_parameter_409: name = "parameter_409" + original_name = "batch_norm2d_6.w_2" shape = [24] dtype = "float32" min_val = float("0") @@ -4372,6 +4782,7 @@ class Program_weight_tensor_parameter_409: class Program_weight_tensor_parameter_410: name = "parameter_410" + original_name = "batch_norm2d_6.w_1" shape = [24] dtype = "float32" min_val = float("0") @@ -4381,6 +4792,7 @@ class Program_weight_tensor_parameter_410: class Program_weight_tensor_parameter_411: name = "parameter_411" + original_name = "conv2d_6.w_0" shape = [24, 24, 1, 1] dtype = "float32" min_val = float("-1.34901") @@ -4392,6 +4804,7 @@ class Program_weight_tensor_parameter_411: class Program_weight_tensor_parameter_412: name = "parameter_412" + original_name = "batch_norm2d_5.b_0" shape = [24] dtype = "float32" min_val = float("0") @@ -4401,6 +4814,7 @@ class Program_weight_tensor_parameter_412: class Program_weight_tensor_parameter_413: name = "parameter_413" + original_name = "batch_norm2d_5.w_0" shape = [24] dtype = "float32" min_val = float("0") @@ -4410,6 +4824,7 @@ class Program_weight_tensor_parameter_413: class Program_weight_tensor_parameter_414: name = "parameter_414" + original_name = "batch_norm2d_5.w_2" shape = [24] dtype = "float32" min_val = float("0") @@ -4419,6 +4834,7 @@ class Program_weight_tensor_parameter_414: class Program_weight_tensor_parameter_415: name = "parameter_415" + original_name = "batch_norm2d_5.w_1" shape = [24] dtype = "float32" min_val = float("0") @@ -4428,6 +4844,7 @@ class Program_weight_tensor_parameter_415: class Program_weight_tensor_parameter_416: name = "parameter_416" + original_name = "conv2d_5.w_0" shape = [24, 1, 3, 3] dtype = "float32" min_val = float("-1.18106") @@ -4439,6 +4856,7 @@ class Program_weight_tensor_parameter_416: class Program_weight_tensor_parameter_417: name = "parameter_417" + original_name = "batch_norm2d_4.b_0" shape = [24] dtype = "float32" min_val = float("0") @@ -4448,6 +4866,7 @@ class Program_weight_tensor_parameter_417: class Program_weight_tensor_parameter_418: name = "parameter_418" + original_name = "batch_norm2d_4.w_0" shape = [24] dtype = "float32" min_val = float("0") @@ -4457,6 +4876,7 @@ class Program_weight_tensor_parameter_418: class Program_weight_tensor_parameter_419: name = "parameter_419" + original_name = "batch_norm2d_4.w_2" shape = [24] dtype = "float32" min_val = float("0") @@ -4466,6 +4886,7 @@ class Program_weight_tensor_parameter_419: class Program_weight_tensor_parameter_420: name = "parameter_420" + original_name = "batch_norm2d_4.w_1" shape = [24] dtype = "float32" min_val = float("0") @@ -4475,6 +4896,7 @@ class Program_weight_tensor_parameter_420: class Program_weight_tensor_parameter_421: name = "parameter_421" + original_name = "conv2d_4.w_0" shape = [24, 16, 1, 1] dtype = "float32" min_val = float("-1.82105") @@ -4486,6 +4908,7 @@ class Program_weight_tensor_parameter_421: class Program_weight_tensor_parameter_422: name = "parameter_422" + original_name = "batch_norm2d_3.b_0" shape = [16] dtype = "float32" min_val = float("0") @@ -4495,6 +4918,7 @@ class Program_weight_tensor_parameter_422: class Program_weight_tensor_parameter_423: name = "parameter_423" + original_name = "batch_norm2d_3.w_0" shape = [16] dtype = "float32" min_val = float("0") @@ -4504,6 +4928,7 @@ class Program_weight_tensor_parameter_423: class Program_weight_tensor_parameter_424: name = "parameter_424" + original_name = "batch_norm2d_3.w_2" shape = [16] dtype = "float32" min_val = float("0") @@ -4513,6 +4938,7 @@ class Program_weight_tensor_parameter_424: class Program_weight_tensor_parameter_425: name = "parameter_425" + original_name = "batch_norm2d_3.w_1" shape = [16] dtype = "float32" min_val = float("0") @@ -4522,6 +4948,7 @@ class Program_weight_tensor_parameter_425: class Program_weight_tensor_parameter_426: name = "parameter_426" + original_name = "conv2d_3.w_0" shape = [16, 1, 3, 3] dtype = "float32" min_val = float("-0.654964") @@ -4533,6 +4960,7 @@ class Program_weight_tensor_parameter_426: class Program_weight_tensor_parameter_427: name = "parameter_427" + original_name = "batch_norm2d_2.b_0" shape = [16] dtype = "float32" min_val = float("0") @@ -4542,6 +4970,7 @@ class Program_weight_tensor_parameter_427: class Program_weight_tensor_parameter_428: name = "parameter_428" + original_name = "batch_norm2d_2.w_0" shape = [16] dtype = "float32" min_val = float("0") @@ -4551,6 +4980,7 @@ class Program_weight_tensor_parameter_428: class Program_weight_tensor_parameter_429: name = "parameter_429" + original_name = "batch_norm2d_2.w_2" shape = [16] dtype = "float32" min_val = float("0") @@ -4560,6 +4990,7 @@ class Program_weight_tensor_parameter_429: class Program_weight_tensor_parameter_430: name = "parameter_430" + original_name = "batch_norm2d_2.w_1" shape = [16] dtype = "float32" min_val = float("0") @@ -4569,6 +5000,7 @@ class Program_weight_tensor_parameter_430: class Program_weight_tensor_parameter_431: name = "parameter_431" + original_name = "conv2d_2.w_0" shape = [16, 8, 1, 1] dtype = "float32" min_val = float("-1.50342") @@ -4580,6 +5012,7 @@ class Program_weight_tensor_parameter_431: class Program_weight_tensor_parameter_432: name = "parameter_432" + original_name = "batch_norm2d_1.b_0" shape = [8] dtype = "float32" min_val = float("0") @@ -4589,6 +5022,7 @@ class Program_weight_tensor_parameter_432: class Program_weight_tensor_parameter_433: name = "parameter_433" + original_name = "batch_norm2d_1.w_0" shape = [8] dtype = "float32" min_val = float("0") @@ -4598,6 +5032,7 @@ class Program_weight_tensor_parameter_433: class Program_weight_tensor_parameter_434: name = "parameter_434" + original_name = "batch_norm2d_1.w_2" shape = [8] dtype = "float32" min_val = float("0") @@ -4607,6 +5042,7 @@ class Program_weight_tensor_parameter_434: class Program_weight_tensor_parameter_435: name = "parameter_435" + original_name = "batch_norm2d_1.w_1" shape = [8] dtype = "float32" min_val = float("0") @@ -4616,6 +5052,7 @@ class Program_weight_tensor_parameter_435: class Program_weight_tensor_parameter_436: name = "parameter_436" + original_name = "conv2d_1.w_0" shape = [8, 1, 3, 3] dtype = "float32" min_val = float("-2.12639") @@ -4627,6 +5064,7 @@ class Program_weight_tensor_parameter_436: class Program_weight_tensor_parameter_437: name = "parameter_437" + original_name = "batch_norm2d_0.b_0" shape = [8] dtype = "float32" min_val = float("0") @@ -4636,6 +5074,7 @@ class Program_weight_tensor_parameter_437: class Program_weight_tensor_parameter_438: name = "parameter_438" + original_name = "batch_norm2d_0.w_0" shape = [8] dtype = "float32" min_val = float("0") @@ -4645,6 +5084,7 @@ class Program_weight_tensor_parameter_438: class Program_weight_tensor_parameter_439: name = "parameter_439" + original_name = "batch_norm2d_0.w_2" shape = [8] dtype = "float32" min_val = float("0") @@ -4654,6 +5094,7 @@ class Program_weight_tensor_parameter_439: class Program_weight_tensor_parameter_440: name = "parameter_440" + original_name = "batch_norm2d_0.w_1" shape = [8] dtype = "float32" min_val = float("0") @@ -4663,6 +5104,7 @@ class Program_weight_tensor_parameter_440: class Program_weight_tensor_parameter_441: name = "parameter_441" + original_name = "conv2d_0.w_0" shape = [8, 3, 3, 3] dtype = "float32" min_val = float("-1.697") diff --git a/paddle_samples/PaddleX/PicoDet-XS/subgraph_6/input_meta.py b/paddle_samples/PaddleX/PicoDet-XS/subgraph_6/input_meta.py index 2fc35b65b..2758c5ae6 100644 --- a/paddle_samples/PaddleX/PicoDet-XS/subgraph_6/input_meta.py +++ b/paddle_samples/PaddleX/PicoDet-XS/subgraph_6/input_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_data_0: name = "data_0" + original_name = "var_1708" shape = [4, 3, 448, 448] dtype = "float32" min_val = float("-2.62534") @@ -11,6 +12,7 @@ class Program_weight_tensor_data_0: class Program_weight_tensor_data_1: name = "data_1" + original_name = "var_1949" shape = [8] dtype = "float32" data = [0.0, 1.0, 2.0, 3.0003, 4.0, 5.0, 6.0006, 7.0] diff --git a/paddle_samples/PaddleX/PicoDet-XS/subgraph_6/weight_meta.py b/paddle_samples/PaddleX/PicoDet-XS/subgraph_6/weight_meta.py index ab626761b..42d60161b 100644 --- a/paddle_samples/PaddleX/PicoDet-XS/subgraph_6/weight_meta.py +++ b/paddle_samples/PaddleX/PicoDet-XS/subgraph_6/weight_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_parameter_0: name = "parameter_0" + original_name = "batch_norm2d_81.b_0" shape = [1] dtype = "float32" min_val = float("0") @@ -9,6 +10,7 @@ class Program_weight_tensor_parameter_0: class Program_weight_tensor_parameter_1: name = "parameter_1" + original_name = "batch_norm2d_81.w_0" shape = [1] dtype = "float32" min_val = float("0") @@ -18,6 +20,7 @@ class Program_weight_tensor_parameter_1: class Program_weight_tensor_parameter_2: name = "parameter_2" + original_name = "batch_norm2d_81.w_2" shape = [1] dtype = "float32" min_val = float("0") @@ -27,6 +30,7 @@ class Program_weight_tensor_parameter_2: class Program_weight_tensor_parameter_3: name = "parameter_3" + original_name = "batch_norm2d_81.w_1" shape = [1] dtype = "float32" min_val = float("0") @@ -36,6 +40,7 @@ class Program_weight_tensor_parameter_3: class Program_weight_tensor_parameter_4: name = "parameter_4" + original_name = "conv2d_99.w_0" shape = [1, 1, 1, 1] dtype = "float32" min_val = float("0") @@ -45,6 +50,7 @@ class Program_weight_tensor_parameter_4: class Program_weight_tensor_parameter_5: name = "parameter_5" + original_name = "batch_norm2d_80.b_0" shape = [1] dtype = "float32" min_val = float("0") @@ -54,6 +60,7 @@ class Program_weight_tensor_parameter_5: class Program_weight_tensor_parameter_6: name = "parameter_6" + original_name = "batch_norm2d_80.w_0" shape = [1] dtype = "float32" min_val = float("0") @@ -63,6 +70,7 @@ class Program_weight_tensor_parameter_6: class Program_weight_tensor_parameter_7: name = "parameter_7" + original_name = "batch_norm2d_80.w_2" shape = [1] dtype = "float32" min_val = float("0") @@ -72,6 +80,7 @@ class Program_weight_tensor_parameter_7: class Program_weight_tensor_parameter_8: name = "parameter_8" + original_name = "batch_norm2d_80.w_1" shape = [1] dtype = "float32" min_val = float("0") @@ -81,6 +90,7 @@ class Program_weight_tensor_parameter_8: class Program_weight_tensor_parameter_9: name = "parameter_9" + original_name = "conv2d_98.w_0" shape = [1, 96, 5, 5] dtype = "float32" min_val = float("-0.370393") @@ -92,6 +102,7 @@ class Program_weight_tensor_parameter_9: class Program_weight_tensor_parameter_10: name = "parameter_10" + original_name = "conv2d_97.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -101,6 +112,7 @@ class Program_weight_tensor_parameter_10: class Program_weight_tensor_parameter_11: name = "parameter_11" + original_name = "conv2d_97.w_0" shape = [32, 96, 1, 1] dtype = "float32" min_val = float("-0.596619") @@ -112,6 +124,7 @@ class Program_weight_tensor_parameter_11: class Program_weight_tensor_parameter_12: name = "parameter_12" + original_name = "conv2d_96.b_0" shape = [4] dtype = "float32" min_val = float("0") @@ -121,6 +134,7 @@ class Program_weight_tensor_parameter_12: class Program_weight_tensor_parameter_13: name = "parameter_13" + original_name = "conv2d_96.w_0" shape = [4, 96, 1, 1] dtype = "float32" min_val = float("-0.0247682") @@ -132,6 +146,7 @@ class Program_weight_tensor_parameter_13: class Program_weight_tensor_parameter_14: name = "parameter_14" + original_name = "batch_norm2d_73.b_0" shape = [96] dtype = "float32" min_val = float("-0.365648") @@ -143,6 +158,7 @@ class Program_weight_tensor_parameter_14: class Program_weight_tensor_parameter_15: name = "parameter_15" + original_name = "batch_norm2d_73.w_0" shape = [96] dtype = "float32" min_val = float("0.533085") @@ -154,6 +170,7 @@ class Program_weight_tensor_parameter_15: class Program_weight_tensor_parameter_16: name = "parameter_16" + original_name = "batch_norm2d_73.w_2" shape = [96] dtype = "float32" min_val = float("0.00912969") @@ -165,6 +182,7 @@ class Program_weight_tensor_parameter_16: class Program_weight_tensor_parameter_17: name = "parameter_17" + original_name = "batch_norm2d_73.w_1" shape = [96] dtype = "float32" min_val = float("-0.2067") @@ -176,6 +194,7 @@ class Program_weight_tensor_parameter_17: class Program_weight_tensor_parameter_18: name = "parameter_18" + original_name = "conv2d_81.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.27454") @@ -187,6 +206,7 @@ class Program_weight_tensor_parameter_18: class Program_weight_tensor_parameter_19: name = "parameter_19" + original_name = "conv2d_80.b_0" shape = [96] dtype = "float32" min_val = float("-0.0581129") @@ -198,6 +218,7 @@ class Program_weight_tensor_parameter_19: class Program_weight_tensor_parameter_20: name = "parameter_20" + original_name = "conv2d_80.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.189324") @@ -209,6 +230,7 @@ class Program_weight_tensor_parameter_20: class Program_weight_tensor_parameter_21: name = "parameter_21" + original_name = "batch_norm2d_72.b_0" shape = [96] dtype = "float32" min_val = float("-1.04876") @@ -220,6 +242,7 @@ class Program_weight_tensor_parameter_21: class Program_weight_tensor_parameter_22: name = "parameter_22" + original_name = "batch_norm2d_72.w_0" shape = [96] dtype = "float32" min_val = float("0.549376") @@ -231,6 +254,7 @@ class Program_weight_tensor_parameter_22: class Program_weight_tensor_parameter_23: name = "parameter_23" + original_name = "batch_norm2d_72.w_2" shape = [96] dtype = "float32" min_val = float("0.0394639") @@ -242,6 +266,7 @@ class Program_weight_tensor_parameter_23: class Program_weight_tensor_parameter_24: name = "parameter_24" + original_name = "batch_norm2d_72.w_1" shape = [96] dtype = "float32" min_val = float("-0.74367") @@ -253,6 +278,7 @@ class Program_weight_tensor_parameter_24: class Program_weight_tensor_parameter_25: name = "parameter_25" + original_name = "conv2d_79.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.387566") @@ -264,6 +290,7 @@ class Program_weight_tensor_parameter_25: class Program_weight_tensor_parameter_26: name = "parameter_26" + original_name = "batch_norm2d_71.b_0" shape = [96] dtype = "float32" min_val = float("-0.988036") @@ -275,6 +302,7 @@ class Program_weight_tensor_parameter_26: class Program_weight_tensor_parameter_27: name = "parameter_27" + original_name = "batch_norm2d_71.w_0" shape = [96] dtype = "float32" min_val = float("0.57921") @@ -286,6 +314,7 @@ class Program_weight_tensor_parameter_27: class Program_weight_tensor_parameter_28: name = "parameter_28" + original_name = "batch_norm2d_71.w_2" shape = [96] dtype = "float32" min_val = float("7.01779e-05") @@ -297,6 +326,7 @@ class Program_weight_tensor_parameter_28: class Program_weight_tensor_parameter_29: name = "parameter_29" + original_name = "batch_norm2d_71.w_1" shape = [96] dtype = "float32" min_val = float("-0.391475") @@ -308,6 +338,7 @@ class Program_weight_tensor_parameter_29: class Program_weight_tensor_parameter_30: name = "parameter_30" + original_name = "conv2d_78.w_0" shape = [96, 1, 5, 5] dtype = "float32" min_val = float("-0.209992") @@ -319,6 +350,7 @@ class Program_weight_tensor_parameter_30: class Program_weight_tensor_parameter_31: name = "parameter_31" + original_name = "batch_norm2d_70.b_0" shape = [96] dtype = "float32" min_val = float("-1.28817") @@ -330,6 +362,7 @@ class Program_weight_tensor_parameter_31: class Program_weight_tensor_parameter_32: name = "parameter_32" + original_name = "batch_norm2d_70.w_0" shape = [96] dtype = "float32" min_val = float("0.0495093") @@ -341,6 +374,7 @@ class Program_weight_tensor_parameter_32: class Program_weight_tensor_parameter_33: name = "parameter_33" + original_name = "batch_norm2d_70.w_2" shape = [96] dtype = "float32" min_val = float("0.0364532") @@ -352,6 +386,7 @@ class Program_weight_tensor_parameter_33: class Program_weight_tensor_parameter_34: name = "parameter_34" + original_name = "batch_norm2d_70.w_1" shape = [96] dtype = "float32" min_val = float("-0.706397") @@ -363,6 +398,7 @@ class Program_weight_tensor_parameter_34: class Program_weight_tensor_parameter_35: name = "parameter_35" + original_name = "conv2d_77.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.297062") @@ -374,6 +410,7 @@ class Program_weight_tensor_parameter_35: class Program_weight_tensor_parameter_36: name = "parameter_36" + original_name = "batch_norm2d_69.b_0" shape = [96] dtype = "float32" min_val = float("-1.44232") @@ -385,6 +422,7 @@ class Program_weight_tensor_parameter_36: class Program_weight_tensor_parameter_37: name = "parameter_37" + original_name = "batch_norm2d_69.w_0" shape = [96] dtype = "float32" min_val = float("0.741794") @@ -396,6 +434,7 @@ class Program_weight_tensor_parameter_37: class Program_weight_tensor_parameter_38: name = "parameter_38" + original_name = "batch_norm2d_69.w_2" shape = [96] dtype = "float32" min_val = float("0.000148336") @@ -407,6 +446,7 @@ class Program_weight_tensor_parameter_38: class Program_weight_tensor_parameter_39: name = "parameter_39" + original_name = "batch_norm2d_69.w_1" shape = [96] dtype = "float32" min_val = float("-0.0726466") @@ -418,6 +458,7 @@ class Program_weight_tensor_parameter_39: class Program_weight_tensor_parameter_40: name = "parameter_40" + original_name = "conv2d_76.w_0" shape = [96, 1, 5, 5] dtype = "float32" min_val = float("-0.285311") @@ -429,6 +470,7 @@ class Program_weight_tensor_parameter_40: class Program_weight_tensor_parameter_41: name = "parameter_41" + original_name = "batch_norm2d_79.b_0" shape = [1] dtype = "float32" min_val = float("0") @@ -438,6 +480,7 @@ class Program_weight_tensor_parameter_41: class Program_weight_tensor_parameter_42: name = "parameter_42" + original_name = "batch_norm2d_79.w_0" shape = [1] dtype = "float32" min_val = float("0") @@ -447,6 +490,7 @@ class Program_weight_tensor_parameter_42: class Program_weight_tensor_parameter_43: name = "parameter_43" + original_name = "batch_norm2d_79.w_2" shape = [1] dtype = "float32" min_val = float("0") @@ -456,6 +500,7 @@ class Program_weight_tensor_parameter_43: class Program_weight_tensor_parameter_44: name = "parameter_44" + original_name = "batch_norm2d_79.w_1" shape = [1] dtype = "float32" min_val = float("0") @@ -465,6 +510,7 @@ class Program_weight_tensor_parameter_44: class Program_weight_tensor_parameter_45: name = "parameter_45" + original_name = "conv2d_95.w_0" shape = [1, 1, 1, 1] dtype = "float32" min_val = float("0") @@ -474,6 +520,7 @@ class Program_weight_tensor_parameter_45: class Program_weight_tensor_parameter_46: name = "parameter_46" + original_name = "batch_norm2d_78.b_0" shape = [1] dtype = "float32" min_val = float("0") @@ -483,6 +530,7 @@ class Program_weight_tensor_parameter_46: class Program_weight_tensor_parameter_47: name = "parameter_47" + original_name = "batch_norm2d_78.w_0" shape = [1] dtype = "float32" min_val = float("0") @@ -492,6 +540,7 @@ class Program_weight_tensor_parameter_47: class Program_weight_tensor_parameter_48: name = "parameter_48" + original_name = "batch_norm2d_78.w_2" shape = [1] dtype = "float32" min_val = float("0") @@ -501,6 +550,7 @@ class Program_weight_tensor_parameter_48: class Program_weight_tensor_parameter_49: name = "parameter_49" + original_name = "batch_norm2d_78.w_1" shape = [1] dtype = "float32" min_val = float("0") @@ -510,6 +560,7 @@ class Program_weight_tensor_parameter_49: class Program_weight_tensor_parameter_50: name = "parameter_50" + original_name = "conv2d_94.w_0" shape = [1, 96, 5, 5] dtype = "float32" min_val = float("-0.526908") @@ -521,6 +572,7 @@ class Program_weight_tensor_parameter_50: class Program_weight_tensor_parameter_51: name = "parameter_51" + original_name = "conv2d_93.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -530,6 +582,7 @@ class Program_weight_tensor_parameter_51: class Program_weight_tensor_parameter_52: name = "parameter_52" + original_name = "conv2d_93.w_0" shape = [32, 96, 1, 1] dtype = "float32" min_val = float("-0.556513") @@ -541,6 +594,7 @@ class Program_weight_tensor_parameter_52: class Program_weight_tensor_parameter_53: name = "parameter_53" + original_name = "conv2d_92.b_0" shape = [4] dtype = "float32" min_val = float("0") @@ -550,6 +604,7 @@ class Program_weight_tensor_parameter_53: class Program_weight_tensor_parameter_54: name = "parameter_54" + original_name = "conv2d_92.w_0" shape = [4, 96, 1, 1] dtype = "float32" min_val = float("-0.0295013") @@ -561,6 +616,7 @@ class Program_weight_tensor_parameter_54: class Program_weight_tensor_parameter_55: name = "parameter_55" + original_name = "batch_norm2d_68.b_0" shape = [96] dtype = "float32" min_val = float("-0.515496") @@ -572,6 +628,7 @@ class Program_weight_tensor_parameter_55: class Program_weight_tensor_parameter_56: name = "parameter_56" + original_name = "batch_norm2d_68.w_0" shape = [96] dtype = "float32" min_val = float("0.516979") @@ -583,6 +640,7 @@ class Program_weight_tensor_parameter_56: class Program_weight_tensor_parameter_57: name = "parameter_57" + original_name = "batch_norm2d_68.w_2" shape = [96] dtype = "float32" min_val = float("0.0111976") @@ -594,6 +652,7 @@ class Program_weight_tensor_parameter_57: class Program_weight_tensor_parameter_58: name = "parameter_58" + original_name = "batch_norm2d_68.w_1" shape = [96] dtype = "float32" min_val = float("-0.332197") @@ -605,6 +664,7 @@ class Program_weight_tensor_parameter_58: class Program_weight_tensor_parameter_59: name = "parameter_59" + original_name = "conv2d_75.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.418305") @@ -616,6 +676,7 @@ class Program_weight_tensor_parameter_59: class Program_weight_tensor_parameter_60: name = "parameter_60" + original_name = "conv2d_74.b_0" shape = [96] dtype = "float32" min_val = float("-0.0361323") @@ -627,6 +688,7 @@ class Program_weight_tensor_parameter_60: class Program_weight_tensor_parameter_61: name = "parameter_61" + original_name = "conv2d_74.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.355478") @@ -638,6 +700,7 @@ class Program_weight_tensor_parameter_61: class Program_weight_tensor_parameter_62: name = "parameter_62" + original_name = "batch_norm2d_67.b_0" shape = [96] dtype = "float32" min_val = float("-1.9389") @@ -649,6 +712,7 @@ class Program_weight_tensor_parameter_62: class Program_weight_tensor_parameter_63: name = "parameter_63" + original_name = "batch_norm2d_67.w_0" shape = [96] dtype = "float32" min_val = float("0.382919") @@ -660,6 +724,7 @@ class Program_weight_tensor_parameter_63: class Program_weight_tensor_parameter_64: name = "parameter_64" + original_name = "batch_norm2d_67.w_2" shape = [96] dtype = "float32" min_val = float("0.054957") @@ -671,6 +736,7 @@ class Program_weight_tensor_parameter_64: class Program_weight_tensor_parameter_65: name = "parameter_65" + original_name = "batch_norm2d_67.w_1" shape = [96] dtype = "float32" min_val = float("-1.73683") @@ -682,6 +748,7 @@ class Program_weight_tensor_parameter_65: class Program_weight_tensor_parameter_66: name = "parameter_66" + original_name = "conv2d_73.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.521321") @@ -693,6 +760,7 @@ class Program_weight_tensor_parameter_66: class Program_weight_tensor_parameter_67: name = "parameter_67" + original_name = "batch_norm2d_66.b_0" shape = [96] dtype = "float32" min_val = float("-0.981454") @@ -704,6 +772,7 @@ class Program_weight_tensor_parameter_67: class Program_weight_tensor_parameter_68: name = "parameter_68" + original_name = "batch_norm2d_66.w_0" shape = [96] dtype = "float32" min_val = float("0.497134") @@ -715,6 +784,7 @@ class Program_weight_tensor_parameter_68: class Program_weight_tensor_parameter_69: name = "parameter_69" + original_name = "batch_norm2d_66.w_2" shape = [96] dtype = "float32" min_val = float("0.00239658") @@ -726,6 +796,7 @@ class Program_weight_tensor_parameter_69: class Program_weight_tensor_parameter_70: name = "parameter_70" + original_name = "batch_norm2d_66.w_1" shape = [96] dtype = "float32" min_val = float("-0.364066") @@ -737,6 +808,7 @@ class Program_weight_tensor_parameter_70: class Program_weight_tensor_parameter_71: name = "parameter_71" + original_name = "conv2d_72.w_0" shape = [96, 1, 5, 5] dtype = "float32" min_val = float("-0.383865") @@ -748,6 +820,7 @@ class Program_weight_tensor_parameter_71: class Program_weight_tensor_parameter_72: name = "parameter_72" + original_name = "batch_norm2d_65.b_0" shape = [96] dtype = "float32" min_val = float("-1.27053") @@ -759,6 +832,7 @@ class Program_weight_tensor_parameter_72: class Program_weight_tensor_parameter_73: name = "parameter_73" + original_name = "batch_norm2d_65.w_0" shape = [96] dtype = "float32" min_val = float("0.143827") @@ -770,6 +844,7 @@ class Program_weight_tensor_parameter_73: class Program_weight_tensor_parameter_74: name = "parameter_74" + original_name = "batch_norm2d_65.w_2" shape = [96] dtype = "float32" min_val = float("0.0325401") @@ -781,6 +856,7 @@ class Program_weight_tensor_parameter_74: class Program_weight_tensor_parameter_75: name = "parameter_75" + original_name = "batch_norm2d_65.w_1" shape = [96] dtype = "float32" min_val = float("-1.18403") @@ -792,6 +868,7 @@ class Program_weight_tensor_parameter_75: class Program_weight_tensor_parameter_76: name = "parameter_76" + original_name = "conv2d_71.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.411147") @@ -803,6 +880,7 @@ class Program_weight_tensor_parameter_76: class Program_weight_tensor_parameter_77: name = "parameter_77" + original_name = "batch_norm2d_64.b_0" shape = [96] dtype = "float32" min_val = float("-0.951711") @@ -814,6 +892,7 @@ class Program_weight_tensor_parameter_77: class Program_weight_tensor_parameter_78: name = "parameter_78" + original_name = "batch_norm2d_64.w_0" shape = [96] dtype = "float32" min_val = float("0.421821") @@ -825,6 +904,7 @@ class Program_weight_tensor_parameter_78: class Program_weight_tensor_parameter_79: name = "parameter_79" + original_name = "batch_norm2d_64.w_2" shape = [96] dtype = "float32" min_val = float("0.00040577") @@ -836,6 +916,7 @@ class Program_weight_tensor_parameter_79: class Program_weight_tensor_parameter_80: name = "parameter_80" + original_name = "batch_norm2d_64.w_1" shape = [96] dtype = "float32" min_val = float("-0.638333") @@ -847,6 +928,7 @@ class Program_weight_tensor_parameter_80: class Program_weight_tensor_parameter_81: name = "parameter_81" + original_name = "conv2d_70.w_0" shape = [96, 1, 5, 5] dtype = "float32" min_val = float("-0.266353") @@ -858,6 +940,7 @@ class Program_weight_tensor_parameter_81: class Program_weight_tensor_parameter_82: name = "parameter_82" + original_name = "batch_norm2d_77.b_0" shape = [1] dtype = "float32" min_val = float("0") @@ -867,6 +950,7 @@ class Program_weight_tensor_parameter_82: class Program_weight_tensor_parameter_83: name = "parameter_83" + original_name = "batch_norm2d_77.w_0" shape = [1] dtype = "float32" min_val = float("0") @@ -876,6 +960,7 @@ class Program_weight_tensor_parameter_83: class Program_weight_tensor_parameter_84: name = "parameter_84" + original_name = "batch_norm2d_77.w_2" shape = [1] dtype = "float32" min_val = float("0") @@ -885,6 +970,7 @@ class Program_weight_tensor_parameter_84: class Program_weight_tensor_parameter_85: name = "parameter_85" + original_name = "batch_norm2d_77.w_1" shape = [1] dtype = "float32" min_val = float("0") @@ -894,6 +980,7 @@ class Program_weight_tensor_parameter_85: class Program_weight_tensor_parameter_86: name = "parameter_86" + original_name = "conv2d_91.w_0" shape = [1, 1, 1, 1] dtype = "float32" min_val = float("0") @@ -903,6 +990,7 @@ class Program_weight_tensor_parameter_86: class Program_weight_tensor_parameter_87: name = "parameter_87" + original_name = "batch_norm2d_76.b_0" shape = [1] dtype = "float32" min_val = float("0") @@ -912,6 +1000,7 @@ class Program_weight_tensor_parameter_87: class Program_weight_tensor_parameter_88: name = "parameter_88" + original_name = "batch_norm2d_76.w_0" shape = [1] dtype = "float32" min_val = float("0") @@ -921,6 +1010,7 @@ class Program_weight_tensor_parameter_88: class Program_weight_tensor_parameter_89: name = "parameter_89" + original_name = "batch_norm2d_76.w_2" shape = [1] dtype = "float32" min_val = float("0") @@ -930,6 +1020,7 @@ class Program_weight_tensor_parameter_89: class Program_weight_tensor_parameter_90: name = "parameter_90" + original_name = "batch_norm2d_76.w_1" shape = [1] dtype = "float32" min_val = float("0") @@ -939,6 +1030,7 @@ class Program_weight_tensor_parameter_90: class Program_weight_tensor_parameter_91: name = "parameter_91" + original_name = "conv2d_90.w_0" shape = [1, 96, 5, 5] dtype = "float32" min_val = float("-0.148738") @@ -950,6 +1042,7 @@ class Program_weight_tensor_parameter_91: class Program_weight_tensor_parameter_92: name = "parameter_92" + original_name = "conv2d_89.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -959,6 +1052,7 @@ class Program_weight_tensor_parameter_92: class Program_weight_tensor_parameter_93: name = "parameter_93" + original_name = "conv2d_89.w_0" shape = [32, 96, 1, 1] dtype = "float32" min_val = float("-0.563531") @@ -970,6 +1064,7 @@ class Program_weight_tensor_parameter_93: class Program_weight_tensor_parameter_94: name = "parameter_94" + original_name = "conv2d_88.b_0" shape = [4] dtype = "float32" min_val = float("0") @@ -979,6 +1074,7 @@ class Program_weight_tensor_parameter_94: class Program_weight_tensor_parameter_95: name = "parameter_95" + original_name = "conv2d_88.w_0" shape = [4, 96, 1, 1] dtype = "float32" min_val = float("-0.0315677") @@ -990,6 +1086,7 @@ class Program_weight_tensor_parameter_95: class Program_weight_tensor_parameter_96: name = "parameter_96" + original_name = "batch_norm2d_63.b_0" shape = [96] dtype = "float32" min_val = float("-1.01976") @@ -1001,6 +1098,7 @@ class Program_weight_tensor_parameter_96: class Program_weight_tensor_parameter_97: name = "parameter_97" + original_name = "batch_norm2d_63.w_0" shape = [96] dtype = "float32" min_val = float("0.727114") @@ -1012,6 +1110,7 @@ class Program_weight_tensor_parameter_97: class Program_weight_tensor_parameter_98: name = "parameter_98" + original_name = "batch_norm2d_63.w_2" shape = [96] dtype = "float32" min_val = float("0.029274") @@ -1023,6 +1122,7 @@ class Program_weight_tensor_parameter_98: class Program_weight_tensor_parameter_99: name = "parameter_99" + original_name = "batch_norm2d_63.w_1" shape = [96] dtype = "float32" min_val = float("-0.960717") @@ -1034,6 +1134,7 @@ class Program_weight_tensor_parameter_99: class Program_weight_tensor_parameter_100: name = "parameter_100" + original_name = "conv2d_69.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.509126") @@ -1045,6 +1146,7 @@ class Program_weight_tensor_parameter_100: class Program_weight_tensor_parameter_101: name = "parameter_101" + original_name = "conv2d_68.b_0" shape = [96] dtype = "float32" min_val = float("-0.038882") @@ -1056,6 +1158,7 @@ class Program_weight_tensor_parameter_101: class Program_weight_tensor_parameter_102: name = "parameter_102" + original_name = "conv2d_68.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.420278") @@ -1067,6 +1170,7 @@ class Program_weight_tensor_parameter_102: class Program_weight_tensor_parameter_103: name = "parameter_103" + original_name = "batch_norm2d_62.b_0" shape = [96] dtype = "float32" min_val = float("-1.50389") @@ -1078,6 +1182,7 @@ class Program_weight_tensor_parameter_103: class Program_weight_tensor_parameter_104: name = "parameter_104" + original_name = "batch_norm2d_62.w_0" shape = [96] dtype = "float32" min_val = float("0.305781") @@ -1089,6 +1194,7 @@ class Program_weight_tensor_parameter_104: class Program_weight_tensor_parameter_105: name = "parameter_105" + original_name = "batch_norm2d_62.w_2" shape = [96] dtype = "float32" min_val = float("0.0652354") @@ -1100,6 +1206,7 @@ class Program_weight_tensor_parameter_105: class Program_weight_tensor_parameter_106: name = "parameter_106" + original_name = "batch_norm2d_62.w_1" shape = [96] dtype = "float32" min_val = float("-1.48741") @@ -1111,6 +1218,7 @@ class Program_weight_tensor_parameter_106: class Program_weight_tensor_parameter_107: name = "parameter_107" + original_name = "conv2d_67.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.44424") @@ -1122,6 +1230,7 @@ class Program_weight_tensor_parameter_107: class Program_weight_tensor_parameter_108: name = "parameter_108" + original_name = "batch_norm2d_61.b_0" shape = [96] dtype = "float32" min_val = float("-0.74844") @@ -1133,6 +1242,7 @@ class Program_weight_tensor_parameter_108: class Program_weight_tensor_parameter_109: name = "parameter_109" + original_name = "batch_norm2d_61.w_0" shape = [96] dtype = "float32" min_val = float("0.392191") @@ -1144,6 +1254,7 @@ class Program_weight_tensor_parameter_109: class Program_weight_tensor_parameter_110: name = "parameter_110" + original_name = "batch_norm2d_61.w_2" shape = [96] dtype = "float32" min_val = float("8.3464e-05") @@ -1155,6 +1266,7 @@ class Program_weight_tensor_parameter_110: class Program_weight_tensor_parameter_111: name = "parameter_111" + original_name = "batch_norm2d_61.w_1" shape = [96] dtype = "float32" min_val = float("-0.493965") @@ -1166,6 +1278,7 @@ class Program_weight_tensor_parameter_111: class Program_weight_tensor_parameter_112: name = "parameter_112" + original_name = "conv2d_66.w_0" shape = [96, 1, 5, 5] dtype = "float32" min_val = float("-0.382529") @@ -1177,6 +1290,7 @@ class Program_weight_tensor_parameter_112: class Program_weight_tensor_parameter_113: name = "parameter_113" + original_name = "batch_norm2d_60.b_0" shape = [96] dtype = "float32" min_val = float("-0.928622") @@ -1188,6 +1302,7 @@ class Program_weight_tensor_parameter_113: class Program_weight_tensor_parameter_114: name = "parameter_114" + original_name = "batch_norm2d_60.w_0" shape = [96] dtype = "float32" min_val = float("0.0258385") @@ -1199,6 +1314,7 @@ class Program_weight_tensor_parameter_114: class Program_weight_tensor_parameter_115: name = "parameter_115" + original_name = "batch_norm2d_60.w_2" shape = [96] dtype = "float32" min_val = float("0.0557636") @@ -1210,6 +1326,7 @@ class Program_weight_tensor_parameter_115: class Program_weight_tensor_parameter_116: name = "parameter_116" + original_name = "batch_norm2d_60.w_1" shape = [96] dtype = "float32" min_val = float("-0.958153") @@ -1221,6 +1338,7 @@ class Program_weight_tensor_parameter_116: class Program_weight_tensor_parameter_117: name = "parameter_117" + original_name = "conv2d_65.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.494345") @@ -1232,6 +1350,7 @@ class Program_weight_tensor_parameter_117: class Program_weight_tensor_parameter_118: name = "parameter_118" + original_name = "batch_norm2d_59.b_0" shape = [96] dtype = "float32" min_val = float("-0.255534") @@ -1243,6 +1362,7 @@ class Program_weight_tensor_parameter_118: class Program_weight_tensor_parameter_119: name = "parameter_119" + original_name = "batch_norm2d_59.w_0" shape = [96] dtype = "float32" min_val = float("0.337752") @@ -1254,6 +1374,7 @@ class Program_weight_tensor_parameter_119: class Program_weight_tensor_parameter_120: name = "parameter_120" + original_name = "batch_norm2d_59.w_2" shape = [96] dtype = "float32" min_val = float("4.90439e-05") @@ -1265,6 +1386,7 @@ class Program_weight_tensor_parameter_120: class Program_weight_tensor_parameter_121: name = "parameter_121" + original_name = "batch_norm2d_59.w_1" shape = [96] dtype = "float32" min_val = float("-0.924828") @@ -1276,6 +1398,7 @@ class Program_weight_tensor_parameter_121: class Program_weight_tensor_parameter_122: name = "parameter_122" + original_name = "conv2d_64.w_0" shape = [96, 1, 5, 5] dtype = "float32" min_val = float("-0.355184") @@ -1287,6 +1410,7 @@ class Program_weight_tensor_parameter_122: class Program_weight_tensor_parameter_123: name = "parameter_123" + original_name = "batch_norm2d_75.b_0" shape = [1] dtype = "float32" min_val = float("0") @@ -1296,6 +1420,7 @@ class Program_weight_tensor_parameter_123: class Program_weight_tensor_parameter_124: name = "parameter_124" + original_name = "batch_norm2d_75.w_0" shape = [1] dtype = "float32" min_val = float("0") @@ -1305,6 +1430,7 @@ class Program_weight_tensor_parameter_124: class Program_weight_tensor_parameter_125: name = "parameter_125" + original_name = "batch_norm2d_75.w_2" shape = [1] dtype = "float32" min_val = float("0") @@ -1314,6 +1440,7 @@ class Program_weight_tensor_parameter_125: class Program_weight_tensor_parameter_126: name = "parameter_126" + original_name = "batch_norm2d_75.w_1" shape = [1] dtype = "float32" min_val = float("0") @@ -1323,6 +1450,7 @@ class Program_weight_tensor_parameter_126: class Program_weight_tensor_parameter_127: name = "parameter_127" + original_name = "conv2d_87.w_0" shape = [1, 1, 1, 1] dtype = "float32" min_val = float("0") @@ -1332,6 +1460,7 @@ class Program_weight_tensor_parameter_127: class Program_weight_tensor_parameter_128: name = "parameter_128" + original_name = "batch_norm2d_74.b_0" shape = [1] dtype = "float32" min_val = float("0") @@ -1341,6 +1470,7 @@ class Program_weight_tensor_parameter_128: class Program_weight_tensor_parameter_129: name = "parameter_129" + original_name = "batch_norm2d_74.w_0" shape = [1] dtype = "float32" min_val = float("0") @@ -1350,6 +1480,7 @@ class Program_weight_tensor_parameter_129: class Program_weight_tensor_parameter_130: name = "parameter_130" + original_name = "batch_norm2d_74.w_2" shape = [1] dtype = "float32" min_val = float("0") @@ -1359,6 +1490,7 @@ class Program_weight_tensor_parameter_130: class Program_weight_tensor_parameter_131: name = "parameter_131" + original_name = "batch_norm2d_74.w_1" shape = [1] dtype = "float32" min_val = float("0") @@ -1368,6 +1500,7 @@ class Program_weight_tensor_parameter_131: class Program_weight_tensor_parameter_132: name = "parameter_132" + original_name = "conv2d_86.w_0" shape = [1, 96, 5, 5] dtype = "float32" min_val = float("-0.309523") @@ -1379,6 +1512,7 @@ class Program_weight_tensor_parameter_132: class Program_weight_tensor_parameter_133: name = "parameter_133" + original_name = "conv2d_85.b_0" shape = [32] dtype = "float32" min_val = float("0") @@ -1388,6 +1522,7 @@ class Program_weight_tensor_parameter_133: class Program_weight_tensor_parameter_134: name = "parameter_134" + original_name = "conv2d_85.w_0" shape = [32, 96, 1, 1] dtype = "float32" min_val = float("-0.557687") @@ -1399,6 +1534,7 @@ class Program_weight_tensor_parameter_134: class Program_weight_tensor_parameter_135: name = "parameter_135" + original_name = "conv2d_84.b_0" shape = [4] dtype = "float32" min_val = float("0") @@ -1408,6 +1544,7 @@ class Program_weight_tensor_parameter_135: class Program_weight_tensor_parameter_136: name = "parameter_136" + original_name = "conv2d_84.w_0" shape = [4, 96, 1, 1] dtype = "float32" min_val = float("-0.0253732") @@ -1419,6 +1556,7 @@ class Program_weight_tensor_parameter_136: class Program_weight_tensor_parameter_137: name = "parameter_137" + original_name = "batch_norm2d_58.b_0" shape = [96] dtype = "float32" min_val = float("-0.897614") @@ -1430,6 +1568,7 @@ class Program_weight_tensor_parameter_137: class Program_weight_tensor_parameter_138: name = "parameter_138" + original_name = "batch_norm2d_58.w_0" shape = [96] dtype = "float32" min_val = float("0.691672") @@ -1441,6 +1580,7 @@ class Program_weight_tensor_parameter_138: class Program_weight_tensor_parameter_139: name = "parameter_139" + original_name = "batch_norm2d_58.w_2" shape = [96] dtype = "float32" min_val = float("0.00683363") @@ -1452,6 +1592,7 @@ class Program_weight_tensor_parameter_139: class Program_weight_tensor_parameter_140: name = "parameter_140" + original_name = "batch_norm2d_58.w_1" shape = [96] dtype = "float32" min_val = float("-0.433153") @@ -1463,6 +1604,7 @@ class Program_weight_tensor_parameter_140: class Program_weight_tensor_parameter_141: name = "parameter_141" + original_name = "conv2d_63.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.66434") @@ -1474,6 +1616,7 @@ class Program_weight_tensor_parameter_141: class Program_weight_tensor_parameter_142: name = "parameter_142" + original_name = "conv2d_62.b_0" shape = [96] dtype = "float32" min_val = float("-0.0408518") @@ -1485,6 +1628,7 @@ class Program_weight_tensor_parameter_142: class Program_weight_tensor_parameter_143: name = "parameter_143" + original_name = "conv2d_62.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.641514") @@ -1496,6 +1640,7 @@ class Program_weight_tensor_parameter_143: class Program_weight_tensor_parameter_144: name = "parameter_144" + original_name = "batch_norm2d_57.b_0" shape = [96] dtype = "float32" min_val = float("-1.86159") @@ -1507,6 +1652,7 @@ class Program_weight_tensor_parameter_144: class Program_weight_tensor_parameter_145: name = "parameter_145" + original_name = "batch_norm2d_57.w_0" shape = [96] dtype = "float32" min_val = float("0.140917") @@ -1518,6 +1664,7 @@ class Program_weight_tensor_parameter_145: class Program_weight_tensor_parameter_146: name = "parameter_146" + original_name = "batch_norm2d_57.w_2" shape = [96] dtype = "float32" min_val = float("0.0585159") @@ -1529,6 +1676,7 @@ class Program_weight_tensor_parameter_146: class Program_weight_tensor_parameter_147: name = "parameter_147" + original_name = "batch_norm2d_57.w_1" shape = [96] dtype = "float32" min_val = float("-3.02577") @@ -1540,6 +1688,7 @@ class Program_weight_tensor_parameter_147: class Program_weight_tensor_parameter_148: name = "parameter_148" + original_name = "conv2d_61.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.611152") @@ -1551,6 +1700,7 @@ class Program_weight_tensor_parameter_148: class Program_weight_tensor_parameter_149: name = "parameter_149" + original_name = "batch_norm2d_56.b_0" shape = [96] dtype = "float32" min_val = float("-1.22176") @@ -1562,6 +1712,7 @@ class Program_weight_tensor_parameter_149: class Program_weight_tensor_parameter_150: name = "parameter_150" + original_name = "batch_norm2d_56.w_0" shape = [96] dtype = "float32" min_val = float("0.366803") @@ -1573,6 +1724,7 @@ class Program_weight_tensor_parameter_150: class Program_weight_tensor_parameter_151: name = "parameter_151" + original_name = "batch_norm2d_56.w_2" shape = [96] dtype = "float32" min_val = float("6.5594e-05") @@ -1584,6 +1736,7 @@ class Program_weight_tensor_parameter_151: class Program_weight_tensor_parameter_152: name = "parameter_152" + original_name = "batch_norm2d_56.w_1" shape = [96] dtype = "float32" min_val = float("-0.79717") @@ -1595,6 +1748,7 @@ class Program_weight_tensor_parameter_152: class Program_weight_tensor_parameter_153: name = "parameter_153" + original_name = "conv2d_60.w_0" shape = [96, 1, 5, 5] dtype = "float32" min_val = float("-0.522492") @@ -1606,6 +1760,7 @@ class Program_weight_tensor_parameter_153: class Program_weight_tensor_parameter_154: name = "parameter_154" + original_name = "batch_norm2d_55.b_0" shape = [96] dtype = "float32" min_val = float("-1.54069") @@ -1617,6 +1772,7 @@ class Program_weight_tensor_parameter_154: class Program_weight_tensor_parameter_155: name = "parameter_155" + original_name = "batch_norm2d_55.w_0" shape = [96] dtype = "float32" min_val = float("0.0273143") @@ -1628,6 +1784,7 @@ class Program_weight_tensor_parameter_155: class Program_weight_tensor_parameter_156: name = "parameter_156" + original_name = "batch_norm2d_55.w_2" shape = [96] dtype = "float32" min_val = float("0.0457279") @@ -1639,6 +1796,7 @@ class Program_weight_tensor_parameter_156: class Program_weight_tensor_parameter_157: name = "parameter_157" + original_name = "batch_norm2d_55.w_1" shape = [96] dtype = "float32" min_val = float("-1.45932") @@ -1650,6 +1808,7 @@ class Program_weight_tensor_parameter_157: class Program_weight_tensor_parameter_158: name = "parameter_158" + original_name = "conv2d_59.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.475433") @@ -1661,6 +1820,7 @@ class Program_weight_tensor_parameter_158: class Program_weight_tensor_parameter_159: name = "parameter_159" + original_name = "batch_norm2d_54.b_0" shape = [96] dtype = "float32" min_val = float("-1.79696") @@ -1672,6 +1832,7 @@ class Program_weight_tensor_parameter_159: class Program_weight_tensor_parameter_160: name = "parameter_160" + original_name = "batch_norm2d_54.w_0" shape = [96] dtype = "float32" min_val = float("0.24558") @@ -1683,6 +1844,7 @@ class Program_weight_tensor_parameter_160: class Program_weight_tensor_parameter_161: name = "parameter_161" + original_name = "batch_norm2d_54.w_2" shape = [96] dtype = "float32" min_val = float("0.00040884") @@ -1694,6 +1856,7 @@ class Program_weight_tensor_parameter_161: class Program_weight_tensor_parameter_162: name = "parameter_162" + original_name = "batch_norm2d_54.w_1" shape = [96] dtype = "float32" min_val = float("-0.532114") @@ -1705,6 +1868,7 @@ class Program_weight_tensor_parameter_162: class Program_weight_tensor_parameter_163: name = "parameter_163" + original_name = "conv2d_58.w_0" shape = [96, 1, 5, 5] dtype = "float32" min_val = float("-0.383801") @@ -1716,6 +1880,7 @@ class Program_weight_tensor_parameter_163: class Program_weight_tensor_parameter_164: name = "parameter_164" + original_name = "batch_norm2d_33.b_0" shape = [96] dtype = "float32" min_val = float("-0.17709") @@ -1727,6 +1892,7 @@ class Program_weight_tensor_parameter_164: class Program_weight_tensor_parameter_165: name = "parameter_165" + original_name = "batch_norm2d_33.w_0" shape = [96] dtype = "float32" min_val = float("0.0512167") @@ -1738,6 +1904,7 @@ class Program_weight_tensor_parameter_165: class Program_weight_tensor_parameter_166: name = "parameter_166" + original_name = "batch_norm2d_33.w_2" shape = [96] dtype = "float32" min_val = float("0.000577371") @@ -1749,6 +1916,7 @@ class Program_weight_tensor_parameter_166: class Program_weight_tensor_parameter_167: name = "parameter_167" + original_name = "batch_norm2d_33.w_1" shape = [96] dtype = "float32" min_val = float("-0.0193416") @@ -1760,6 +1928,7 @@ class Program_weight_tensor_parameter_167: class Program_weight_tensor_parameter_168: name = "parameter_168" + original_name = "conv2d_37.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.286597") @@ -1771,6 +1940,7 @@ class Program_weight_tensor_parameter_168: class Program_weight_tensor_parameter_169: name = "parameter_169" + original_name = "batch_norm2d_32.b_0" shape = [96] dtype = "float32" min_val = float("-0.260295") @@ -1782,6 +1952,7 @@ class Program_weight_tensor_parameter_169: class Program_weight_tensor_parameter_170: name = "parameter_170" + original_name = "batch_norm2d_32.w_0" shape = [96] dtype = "float32" min_val = float("0.110979") @@ -1793,6 +1964,7 @@ class Program_weight_tensor_parameter_170: class Program_weight_tensor_parameter_171: name = "parameter_171" + original_name = "batch_norm2d_32.w_2" shape = [96] dtype = "float32" min_val = float("0.000112504") @@ -1804,6 +1976,7 @@ class Program_weight_tensor_parameter_171: class Program_weight_tensor_parameter_172: name = "parameter_172" + original_name = "batch_norm2d_32.w_1" shape = [96] dtype = "float32" min_val = float("-0.295345") @@ -1815,6 +1988,7 @@ class Program_weight_tensor_parameter_172: class Program_weight_tensor_parameter_173: name = "parameter_173" + original_name = "conv2d_36.w_0" shape = [96, 1, 5, 5] dtype = "float32" min_val = float("-0.139199") @@ -1826,6 +2000,7 @@ class Program_weight_tensor_parameter_173: class Program_weight_tensor_parameter_174: name = "parameter_174" + original_name = "batch_norm2d_31.b_0" shape = [96] dtype = "float32" min_val = float("-0.164622") @@ -1837,6 +2012,7 @@ class Program_weight_tensor_parameter_174: class Program_weight_tensor_parameter_175: name = "parameter_175" + original_name = "batch_norm2d_31.w_0" shape = [96] dtype = "float32" min_val = float("0.0174356") @@ -1848,6 +2024,7 @@ class Program_weight_tensor_parameter_175: class Program_weight_tensor_parameter_176: name = "parameter_176" + original_name = "batch_norm2d_31.w_2" shape = [96] dtype = "float32" min_val = float("5.3156e-05") @@ -1859,6 +2036,7 @@ class Program_weight_tensor_parameter_176: class Program_weight_tensor_parameter_177: name = "parameter_177" + original_name = "batch_norm2d_31.w_1" shape = [96] dtype = "float32" min_val = float("-0.00460117") @@ -1870,6 +2048,7 @@ class Program_weight_tensor_parameter_177: class Program_weight_tensor_parameter_178: name = "parameter_178" + original_name = "conv2d_35.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.13472") @@ -1881,6 +2060,7 @@ class Program_weight_tensor_parameter_178: class Program_weight_tensor_parameter_179: name = "parameter_179" + original_name = "batch_norm2d_30.b_0" shape = [96] dtype = "float32" min_val = float("-0.0326481") @@ -1892,6 +2072,7 @@ class Program_weight_tensor_parameter_179: class Program_weight_tensor_parameter_180: name = "parameter_180" + original_name = "batch_norm2d_30.w_0" shape = [96] dtype = "float32" min_val = float("0.0808183") @@ -1903,6 +2084,7 @@ class Program_weight_tensor_parameter_180: class Program_weight_tensor_parameter_181: name = "parameter_181" + original_name = "batch_norm2d_30.w_2" shape = [96] dtype = "float32" min_val = float("3.06399e-05") @@ -1914,6 +2096,7 @@ class Program_weight_tensor_parameter_181: class Program_weight_tensor_parameter_182: name = "parameter_182" + original_name = "batch_norm2d_30.w_1" shape = [96] dtype = "float32" min_val = float("-0.0181096") @@ -1925,6 +2108,7 @@ class Program_weight_tensor_parameter_182: class Program_weight_tensor_parameter_183: name = "parameter_183" + original_name = "conv2d_34.w_0" shape = [96, 1, 5, 5] dtype = "float32" min_val = float("-0.0873842") @@ -1936,6 +2120,7 @@ class Program_weight_tensor_parameter_183: class Program_weight_tensor_parameter_184: name = "parameter_184" + original_name = "batch_norm2d_53.b_0" shape = [96] dtype = "float32" min_val = float("-1.30998") @@ -1947,6 +2132,7 @@ class Program_weight_tensor_parameter_184: class Program_weight_tensor_parameter_185: name = "parameter_185" + original_name = "batch_norm2d_53.w_0" shape = [96] dtype = "float32" min_val = float("0.0866571") @@ -1958,6 +2144,7 @@ class Program_weight_tensor_parameter_185: class Program_weight_tensor_parameter_186: name = "parameter_186" + original_name = "batch_norm2d_53.w_2" shape = [96] dtype = "float32" min_val = float("0.116575") @@ -1969,6 +2156,7 @@ class Program_weight_tensor_parameter_186: class Program_weight_tensor_parameter_187: name = "parameter_187" + original_name = "batch_norm2d_53.w_1" shape = [96] dtype = "float32" min_val = float("-0.568116") @@ -1980,6 +2168,7 @@ class Program_weight_tensor_parameter_187: class Program_weight_tensor_parameter_188: name = "parameter_188" + original_name = "conv2d_57.w_0" shape = [96, 192, 1, 1] dtype = "float32" min_val = float("-0.33141") @@ -1991,6 +2180,7 @@ class Program_weight_tensor_parameter_188: class Program_weight_tensor_parameter_189: name = "parameter_189" + original_name = "batch_norm2d_52.b_0" shape = [192] dtype = "float32" min_val = float("-1.00179") @@ -2002,6 +2192,7 @@ class Program_weight_tensor_parameter_189: class Program_weight_tensor_parameter_190: name = "parameter_190" + original_name = "batch_norm2d_52.w_0" shape = [192] dtype = "float32" min_val = float("0.59987") @@ -2013,6 +2204,7 @@ class Program_weight_tensor_parameter_190: class Program_weight_tensor_parameter_191: name = "parameter_191" + original_name = "batch_norm2d_52.w_2" shape = [192] dtype = "float32" min_val = float("0.00374903") @@ -2024,6 +2216,7 @@ class Program_weight_tensor_parameter_191: class Program_weight_tensor_parameter_192: name = "parameter_192" + original_name = "batch_norm2d_52.w_1" shape = [192] dtype = "float32" min_val = float("-0.378435") @@ -2035,6 +2228,7 @@ class Program_weight_tensor_parameter_192: class Program_weight_tensor_parameter_193: name = "parameter_193" + original_name = "conv2d_56.w_0" shape = [192, 1, 5, 5] dtype = "float32" min_val = float("-0.304951") @@ -2046,6 +2240,7 @@ class Program_weight_tensor_parameter_193: class Program_weight_tensor_parameter_194: name = "parameter_194" + original_name = "batch_norm2d_51.b_0" shape = [192] dtype = "float32" min_val = float("-1.5034") @@ -2057,6 +2252,7 @@ class Program_weight_tensor_parameter_194: class Program_weight_tensor_parameter_195: name = "parameter_195" + original_name = "batch_norm2d_51.w_0" shape = [192] dtype = "float32" min_val = float("0.411181") @@ -2068,6 +2264,7 @@ class Program_weight_tensor_parameter_195: class Program_weight_tensor_parameter_196: name = "parameter_196" + original_name = "batch_norm2d_51.w_2" shape = [192] dtype = "float32" min_val = float("0.14646") @@ -2079,6 +2276,7 @@ class Program_weight_tensor_parameter_196: class Program_weight_tensor_parameter_197: name = "parameter_197" + original_name = "batch_norm2d_51.w_1" shape = [192] dtype = "float32" min_val = float("-0.950145") @@ -2090,6 +2288,7 @@ class Program_weight_tensor_parameter_197: class Program_weight_tensor_parameter_198: name = "parameter_198" + original_name = "conv2d_55.w_0" shape = [192, 192, 1, 1] dtype = "float32" min_val = float("-0.295618") @@ -2101,6 +2300,7 @@ class Program_weight_tensor_parameter_198: class Program_weight_tensor_parameter_199: name = "parameter_199" + original_name = "batch_norm2d_50.b_0" shape = [192] dtype = "float32" min_val = float("-1.31561") @@ -2112,6 +2312,7 @@ class Program_weight_tensor_parameter_199: class Program_weight_tensor_parameter_200: name = "parameter_200" + original_name = "batch_norm2d_50.w_0" shape = [192] dtype = "float32" min_val = float("0.725639") @@ -2123,6 +2324,7 @@ class Program_weight_tensor_parameter_200: class Program_weight_tensor_parameter_201: name = "parameter_201" + original_name = "batch_norm2d_50.w_2" shape = [192] dtype = "float32" min_val = float("6.06505e-05") @@ -2134,6 +2336,7 @@ class Program_weight_tensor_parameter_201: class Program_weight_tensor_parameter_202: name = "parameter_202" + original_name = "batch_norm2d_50.w_1" shape = [192] dtype = "float32" min_val = float("-0.0449533") @@ -2145,6 +2348,7 @@ class Program_weight_tensor_parameter_202: class Program_weight_tensor_parameter_203: name = "parameter_203" + original_name = "conv2d_54.w_0" shape = [192, 1, 5, 5] dtype = "float32" min_val = float("-0.417017") @@ -2156,6 +2360,7 @@ class Program_weight_tensor_parameter_203: class Program_weight_tensor_parameter_204: name = "parameter_204" + original_name = "batch_norm2d_49.b_0" shape = [96] dtype = "float32" min_val = float("-0.243075") @@ -2167,6 +2372,7 @@ class Program_weight_tensor_parameter_204: class Program_weight_tensor_parameter_205: name = "parameter_205" + original_name = "batch_norm2d_49.w_0" shape = [96] dtype = "float32" min_val = float("0.0427695") @@ -2178,6 +2384,7 @@ class Program_weight_tensor_parameter_205: class Program_weight_tensor_parameter_206: name = "parameter_206" + original_name = "batch_norm2d_49.w_2" shape = [96] dtype = "float32" min_val = float("0.00175132") @@ -2189,6 +2396,7 @@ class Program_weight_tensor_parameter_206: class Program_weight_tensor_parameter_207: name = "parameter_207" + original_name = "batch_norm2d_49.w_1" shape = [96] dtype = "float32" min_val = float("-0.0498497") @@ -2200,6 +2408,7 @@ class Program_weight_tensor_parameter_207: class Program_weight_tensor_parameter_208: name = "parameter_208" + original_name = "conv2d_53.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.369653") @@ -2211,6 +2420,7 @@ class Program_weight_tensor_parameter_208: class Program_weight_tensor_parameter_209: name = "parameter_209" + original_name = "batch_norm2d_48.b_0" shape = [96] dtype = "float32" min_val = float("-0.424991") @@ -2222,6 +2432,7 @@ class Program_weight_tensor_parameter_209: class Program_weight_tensor_parameter_210: name = "parameter_210" + original_name = "batch_norm2d_48.w_0" shape = [96] dtype = "float32" min_val = float("0.114334") @@ -2233,6 +2444,7 @@ class Program_weight_tensor_parameter_210: class Program_weight_tensor_parameter_211: name = "parameter_211" + original_name = "batch_norm2d_48.w_2" shape = [96] dtype = "float32" min_val = float("0.000138033") @@ -2244,6 +2456,7 @@ class Program_weight_tensor_parameter_211: class Program_weight_tensor_parameter_212: name = "parameter_212" + original_name = "batch_norm2d_48.w_1" shape = [96] dtype = "float32" min_val = float("-0.722059") @@ -2255,6 +2468,7 @@ class Program_weight_tensor_parameter_212: class Program_weight_tensor_parameter_213: name = "parameter_213" + original_name = "conv2d_52.w_0" shape = [96, 1, 5, 5] dtype = "float32" min_val = float("-0.242406") @@ -2266,6 +2480,7 @@ class Program_weight_tensor_parameter_213: class Program_weight_tensor_parameter_214: name = "parameter_214" + original_name = "batch_norm2d_47.b_0" shape = [96] dtype = "float32" min_val = float("-1.55992") @@ -2277,6 +2492,7 @@ class Program_weight_tensor_parameter_214: class Program_weight_tensor_parameter_215: name = "parameter_215" + original_name = "batch_norm2d_47.w_0" shape = [96] dtype = "float32" min_val = float("0.0556648") @@ -2288,6 +2504,7 @@ class Program_weight_tensor_parameter_215: class Program_weight_tensor_parameter_216: name = "parameter_216" + original_name = "batch_norm2d_47.w_2" shape = [96] dtype = "float32" min_val = float("0.21517") @@ -2299,6 +2516,7 @@ class Program_weight_tensor_parameter_216: class Program_weight_tensor_parameter_217: name = "parameter_217" + original_name = "batch_norm2d_47.w_1" shape = [96] dtype = "float32" min_val = float("-1.08931") @@ -2310,6 +2528,7 @@ class Program_weight_tensor_parameter_217: class Program_weight_tensor_parameter_218: name = "parameter_218" + original_name = "conv2d_51.w_0" shape = [96, 192, 1, 1] dtype = "float32" min_val = float("-0.360031") @@ -2321,6 +2540,7 @@ class Program_weight_tensor_parameter_218: class Program_weight_tensor_parameter_219: name = "parameter_219" + original_name = "batch_norm2d_46.b_0" shape = [192] dtype = "float32" min_val = float("-1.12198") @@ -2332,6 +2552,7 @@ class Program_weight_tensor_parameter_219: class Program_weight_tensor_parameter_220: name = "parameter_220" + original_name = "batch_norm2d_46.w_0" shape = [192] dtype = "float32" min_val = float("0.550447") @@ -2343,6 +2564,7 @@ class Program_weight_tensor_parameter_220: class Program_weight_tensor_parameter_221: name = "parameter_221" + original_name = "batch_norm2d_46.w_2" shape = [192] dtype = "float32" min_val = float("0.00258404") @@ -2354,6 +2576,7 @@ class Program_weight_tensor_parameter_221: class Program_weight_tensor_parameter_222: name = "parameter_222" + original_name = "batch_norm2d_46.w_1" shape = [192] dtype = "float32" min_val = float("-0.472993") @@ -2365,6 +2588,7 @@ class Program_weight_tensor_parameter_222: class Program_weight_tensor_parameter_223: name = "parameter_223" + original_name = "conv2d_50.w_0" shape = [192, 1, 5, 5] dtype = "float32" min_val = float("-0.351402") @@ -2376,6 +2600,7 @@ class Program_weight_tensor_parameter_223: class Program_weight_tensor_parameter_224: name = "parameter_224" + original_name = "batch_norm2d_45.b_0" shape = [192] dtype = "float32" min_val = float("-1.39524") @@ -2387,6 +2612,7 @@ class Program_weight_tensor_parameter_224: class Program_weight_tensor_parameter_225: name = "parameter_225" + original_name = "batch_norm2d_45.w_0" shape = [192] dtype = "float32" min_val = float("0.262573") @@ -2398,6 +2624,7 @@ class Program_weight_tensor_parameter_225: class Program_weight_tensor_parameter_226: name = "parameter_226" + original_name = "batch_norm2d_45.w_2" shape = [192] dtype = "float32" min_val = float("0.108548") @@ -2409,6 +2636,7 @@ class Program_weight_tensor_parameter_226: class Program_weight_tensor_parameter_227: name = "parameter_227" + original_name = "batch_norm2d_45.w_1" shape = [192] dtype = "float32" min_val = float("-0.999839") @@ -2420,6 +2648,7 @@ class Program_weight_tensor_parameter_227: class Program_weight_tensor_parameter_228: name = "parameter_228" + original_name = "conv2d_49.w_0" shape = [192, 192, 1, 1] dtype = "float32" min_val = float("-0.394252") @@ -2431,6 +2660,7 @@ class Program_weight_tensor_parameter_228: class Program_weight_tensor_parameter_229: name = "parameter_229" + original_name = "batch_norm2d_44.b_0" shape = [192] dtype = "float32" min_val = float("-1.42666") @@ -2442,6 +2672,7 @@ class Program_weight_tensor_parameter_229: class Program_weight_tensor_parameter_230: name = "parameter_230" + original_name = "batch_norm2d_44.w_0" shape = [192] dtype = "float32" min_val = float("0.618557") @@ -2453,6 +2684,7 @@ class Program_weight_tensor_parameter_230: class Program_weight_tensor_parameter_231: name = "parameter_231" + original_name = "batch_norm2d_44.w_2" shape = [192] dtype = "float32" min_val = float("6.6943e-05") @@ -2464,6 +2696,7 @@ class Program_weight_tensor_parameter_231: class Program_weight_tensor_parameter_232: name = "parameter_232" + original_name = "batch_norm2d_44.w_1" shape = [192] dtype = "float32" min_val = float("-0.525134") @@ -2475,6 +2708,7 @@ class Program_weight_tensor_parameter_232: class Program_weight_tensor_parameter_233: name = "parameter_233" + original_name = "conv2d_48.w_0" shape = [192, 1, 5, 5] dtype = "float32" min_val = float("-0.470918") @@ -2486,6 +2720,7 @@ class Program_weight_tensor_parameter_233: class Program_weight_tensor_parameter_234: name = "parameter_234" + original_name = "batch_norm2d_43.b_0" shape = [96] dtype = "float32" min_val = float("-0.221583") @@ -2497,6 +2732,7 @@ class Program_weight_tensor_parameter_234: class Program_weight_tensor_parameter_235: name = "parameter_235" + original_name = "batch_norm2d_43.w_0" shape = [96] dtype = "float32" min_val = float("0.0332592") @@ -2508,6 +2744,7 @@ class Program_weight_tensor_parameter_235: class Program_weight_tensor_parameter_236: name = "parameter_236" + original_name = "batch_norm2d_43.w_2" shape = [96] dtype = "float32" min_val = float("0.00243744") @@ -2519,6 +2756,7 @@ class Program_weight_tensor_parameter_236: class Program_weight_tensor_parameter_237: name = "parameter_237" + original_name = "batch_norm2d_43.w_1" shape = [96] dtype = "float32" min_val = float("-0.0628324") @@ -2530,6 +2768,7 @@ class Program_weight_tensor_parameter_237: class Program_weight_tensor_parameter_238: name = "parameter_238" + original_name = "conv2d_47.w_0" shape = [96, 96, 1, 1] dtype = "float32" min_val = float("-0.429183") @@ -2541,6 +2780,7 @@ class Program_weight_tensor_parameter_238: class Program_weight_tensor_parameter_239: name = "parameter_239" + original_name = "batch_norm2d_42.b_0" shape = [96] dtype = "float32" min_val = float("-0.294481") @@ -2552,6 +2792,7 @@ class Program_weight_tensor_parameter_239: class Program_weight_tensor_parameter_240: name = "parameter_240" + original_name = "batch_norm2d_42.w_0" shape = [96] dtype = "float32" min_val = float("0.127512") @@ -2563,6 +2804,7 @@ class Program_weight_tensor_parameter_240: class Program_weight_tensor_parameter_241: name = "parameter_241" + original_name = "batch_norm2d_42.w_2" shape = [96] dtype = "float32" min_val = float("0.000637375") @@ -2574,6 +2816,7 @@ class Program_weight_tensor_parameter_241: class Program_weight_tensor_parameter_242: name = "parameter_242" + original_name = "batch_norm2d_42.w_1" shape = [96] dtype = "float32" min_val = float("-0.47781") @@ -2585,6 +2828,7 @@ class Program_weight_tensor_parameter_242: class Program_weight_tensor_parameter_243: name = "parameter_243" + original_name = "conv2d_46.w_0" shape = [96, 1, 5, 5] dtype = "float32" min_val = float("-0.227017") @@ -2596,6 +2840,7 @@ class Program_weight_tensor_parameter_243: class Program_weight_tensor_parameter_244: name = "parameter_244" + original_name = "batch_norm2d_41.b_0" shape = [96] dtype = "float32" min_val = float("-2.11926") @@ -2607,6 +2852,7 @@ class Program_weight_tensor_parameter_244: class Program_weight_tensor_parameter_245: name = "parameter_245" + original_name = "batch_norm2d_41.w_0" shape = [96] dtype = "float32" min_val = float("0.142937") @@ -2618,6 +2864,7 @@ class Program_weight_tensor_parameter_245: class Program_weight_tensor_parameter_246: name = "parameter_246" + original_name = "batch_norm2d_41.w_2" shape = [96] dtype = "float32" min_val = float("0.111924") @@ -2629,6 +2876,7 @@ class Program_weight_tensor_parameter_246: class Program_weight_tensor_parameter_247: name = "parameter_247" + original_name = "batch_norm2d_41.w_1" shape = [96] dtype = "float32" min_val = float("-1.01852") @@ -2640,6 +2888,7 @@ class Program_weight_tensor_parameter_247: class Program_weight_tensor_parameter_248: name = "parameter_248" + original_name = "conv2d_45.w_0" shape = [96, 192, 1, 1] dtype = "float32" min_val = float("-0.509217") @@ -2651,6 +2900,7 @@ class Program_weight_tensor_parameter_248: class Program_weight_tensor_parameter_249: name = "parameter_249" + original_name = "batch_norm2d_40.b_0" shape = [192] dtype = "float32" min_val = float("-1.44113") @@ -2662,6 +2912,7 @@ class Program_weight_tensor_parameter_249: class Program_weight_tensor_parameter_250: name = "parameter_250" + original_name = "batch_norm2d_40.w_0" shape = [192] dtype = "float32" min_val = float("0.311368") @@ -2673,6 +2924,7 @@ class Program_weight_tensor_parameter_250: class Program_weight_tensor_parameter_251: name = "parameter_251" + original_name = "batch_norm2d_40.w_2" shape = [192] dtype = "float32" min_val = float("7.05707e-05") @@ -2684,6 +2936,7 @@ class Program_weight_tensor_parameter_251: class Program_weight_tensor_parameter_252: name = "parameter_252" + original_name = "batch_norm2d_40.w_1" shape = [192] dtype = "float32" min_val = float("-0.475896") @@ -2695,6 +2948,7 @@ class Program_weight_tensor_parameter_252: class Program_weight_tensor_parameter_253: name = "parameter_253" + original_name = "conv2d_44.w_0" shape = [192, 1, 5, 5] dtype = "float32" min_val = float("-0.547906") @@ -2706,6 +2960,7 @@ class Program_weight_tensor_parameter_253: class Program_weight_tensor_parameter_254: name = "parameter_254" + original_name = "batch_norm2d_39.b_0" shape = [192] dtype = "float32" min_val = float("-2.03343") @@ -2717,6 +2972,7 @@ class Program_weight_tensor_parameter_254: class Program_weight_tensor_parameter_255: name = "parameter_255" + original_name = "batch_norm2d_39.w_0" shape = [192] dtype = "float32" min_val = float("0.0329333") @@ -2728,6 +2984,7 @@ class Program_weight_tensor_parameter_255: class Program_weight_tensor_parameter_256: name = "parameter_256" + original_name = "batch_norm2d_39.w_2" shape = [192] dtype = "float32" min_val = float("0.216146") @@ -2739,6 +2996,7 @@ class Program_weight_tensor_parameter_256: class Program_weight_tensor_parameter_257: name = "parameter_257" + original_name = "batch_norm2d_39.w_1" shape = [192] dtype = "float32" min_val = float("-1.80093") @@ -2750,6 +3008,7 @@ class Program_weight_tensor_parameter_257: class Program_weight_tensor_parameter_258: name = "parameter_258" + original_name = "conv2d_43.w_0" shape = [192, 192, 1, 1] dtype = "float32" min_val = float("-0.435787") @@ -2761,6 +3020,7 @@ class Program_weight_tensor_parameter_258: class Program_weight_tensor_parameter_259: name = "parameter_259" + original_name = "batch_norm2d_38.b_0" shape = [192] dtype = "float32" min_val = float("-1.22927") @@ -2772,6 +3032,7 @@ class Program_weight_tensor_parameter_259: class Program_weight_tensor_parameter_260: name = "parameter_260" + original_name = "batch_norm2d_38.w_0" shape = [192] dtype = "float32" min_val = float("0.535146") @@ -2783,6 +3044,7 @@ class Program_weight_tensor_parameter_260: class Program_weight_tensor_parameter_261: name = "parameter_261" + original_name = "batch_norm2d_38.w_2" shape = [192] dtype = "float32" min_val = float("3.89587e-05") @@ -2794,6 +3056,7 @@ class Program_weight_tensor_parameter_261: class Program_weight_tensor_parameter_262: name = "parameter_262" + original_name = "batch_norm2d_38.w_1" shape = [192] dtype = "float32" min_val = float("-0.783396") @@ -2805,6 +3068,7 @@ class Program_weight_tensor_parameter_262: class Program_weight_tensor_parameter_263: name = "parameter_263" + original_name = "conv2d_42.w_0" shape = [192, 1, 5, 5] dtype = "float32" min_val = float("-0.85639") @@ -2816,6 +3080,7 @@ class Program_weight_tensor_parameter_263: class Program_weight_tensor_parameter_264: name = "parameter_264" + original_name = "batch_norm2d_37.b_0" shape = [96] dtype = "float32" min_val = float("-0.837607") @@ -2827,6 +3092,7 @@ class Program_weight_tensor_parameter_264: class Program_weight_tensor_parameter_265: name = "parameter_265" + original_name = "batch_norm2d_37.w_0" shape = [96] dtype = "float32" min_val = float("0.0703274") @@ -2838,6 +3104,7 @@ class Program_weight_tensor_parameter_265: class Program_weight_tensor_parameter_266: name = "parameter_266" + original_name = "batch_norm2d_37.w_2" shape = [96] dtype = "float32" min_val = float("0.257464") @@ -2849,6 +3116,7 @@ class Program_weight_tensor_parameter_266: class Program_weight_tensor_parameter_267: name = "parameter_267" + original_name = "batch_norm2d_37.w_1" shape = [96] dtype = "float32" min_val = float("-1.00687") @@ -2860,6 +3128,7 @@ class Program_weight_tensor_parameter_267: class Program_weight_tensor_parameter_268: name = "parameter_268" + original_name = "conv2d_41.w_0" shape = [96, 192, 1, 1] dtype = "float32" min_val = float("-0.289324") @@ -2871,6 +3140,7 @@ class Program_weight_tensor_parameter_268: class Program_weight_tensor_parameter_269: name = "parameter_269" + original_name = "batch_norm2d_36.b_0" shape = [192] dtype = "float32" min_val = float("-1.08735") @@ -2882,6 +3152,7 @@ class Program_weight_tensor_parameter_269: class Program_weight_tensor_parameter_270: name = "parameter_270" + original_name = "batch_norm2d_36.w_0" shape = [192] dtype = "float32" min_val = float("0.554966") @@ -2893,6 +3164,7 @@ class Program_weight_tensor_parameter_270: class Program_weight_tensor_parameter_271: name = "parameter_271" + original_name = "batch_norm2d_36.w_2" shape = [192] dtype = "float32" min_val = float("0.00144024") @@ -2904,6 +3176,7 @@ class Program_weight_tensor_parameter_271: class Program_weight_tensor_parameter_272: name = "parameter_272" + original_name = "batch_norm2d_36.w_1" shape = [192] dtype = "float32" min_val = float("-0.572725") @@ -2915,6 +3188,7 @@ class Program_weight_tensor_parameter_272: class Program_weight_tensor_parameter_273: name = "parameter_273" + original_name = "conv2d_40.w_0" shape = [192, 1, 5, 5] dtype = "float32" min_val = float("-0.485615") @@ -2926,6 +3200,7 @@ class Program_weight_tensor_parameter_273: class Program_weight_tensor_parameter_274: name = "parameter_274" + original_name = "batch_norm2d_35.b_0" shape = [192] dtype = "float32" min_val = float("-1.31086") @@ -2937,6 +3212,7 @@ class Program_weight_tensor_parameter_274: class Program_weight_tensor_parameter_275: name = "parameter_275" + original_name = "batch_norm2d_35.w_0" shape = [192] dtype = "float32" min_val = float("0.134485") @@ -2948,6 +3224,7 @@ class Program_weight_tensor_parameter_275: class Program_weight_tensor_parameter_276: name = "parameter_276" + original_name = "batch_norm2d_35.w_2" shape = [192] dtype = "float32" min_val = float("0.249372") @@ -2959,6 +3236,7 @@ class Program_weight_tensor_parameter_276: class Program_weight_tensor_parameter_277: name = "parameter_277" + original_name = "batch_norm2d_35.w_1" shape = [192] dtype = "float32" min_val = float("-1.89341") @@ -2970,6 +3248,7 @@ class Program_weight_tensor_parameter_277: class Program_weight_tensor_parameter_278: name = "parameter_278" + original_name = "conv2d_39.w_0" shape = [192, 192, 1, 1] dtype = "float32" min_val = float("-0.354787") @@ -2981,6 +3260,7 @@ class Program_weight_tensor_parameter_278: class Program_weight_tensor_parameter_279: name = "parameter_279" + original_name = "batch_norm2d_34.b_0" shape = [192] dtype = "float32" min_val = float("-1.26398") @@ -2992,6 +3272,7 @@ class Program_weight_tensor_parameter_279: class Program_weight_tensor_parameter_280: name = "parameter_280" + original_name = "batch_norm2d_34.w_0" shape = [192] dtype = "float32" min_val = float("0.621238") @@ -3003,6 +3284,7 @@ class Program_weight_tensor_parameter_280: class Program_weight_tensor_parameter_281: name = "parameter_281" + original_name = "batch_norm2d_34.w_2" shape = [192] dtype = "float32" min_val = float("3.72148e-05") @@ -3014,6 +3296,7 @@ class Program_weight_tensor_parameter_281: class Program_weight_tensor_parameter_282: name = "parameter_282" + original_name = "batch_norm2d_34.w_1" shape = [192] dtype = "float32" min_val = float("-0.102497") @@ -3025,6 +3308,7 @@ class Program_weight_tensor_parameter_282: class Program_weight_tensor_parameter_283: name = "parameter_283" + original_name = "conv2d_38.w_0" shape = [192, 1, 5, 5] dtype = "float32" min_val = float("-0.515542") @@ -3036,6 +3320,7 @@ class Program_weight_tensor_parameter_283: class Program_weight_tensor_parameter_284: name = "parameter_284" + original_name = "batch_norm2d_29.b_0" shape = [96] dtype = "float32" min_val = float("-0.150142") @@ -3047,6 +3332,7 @@ class Program_weight_tensor_parameter_284: class Program_weight_tensor_parameter_285: name = "parameter_285" + original_name = "batch_norm2d_29.w_0" shape = [96] dtype = "float32" min_val = float("0.0505305") @@ -3058,6 +3344,7 @@ class Program_weight_tensor_parameter_285: class Program_weight_tensor_parameter_286: name = "parameter_286" + original_name = "batch_norm2d_29.w_2" shape = [96] dtype = "float32" min_val = float("3.40472") @@ -3069,6 +3356,7 @@ class Program_weight_tensor_parameter_286: class Program_weight_tensor_parameter_287: name = "parameter_287" + original_name = "batch_norm2d_29.w_1" shape = [96] dtype = "float32" min_val = float("-3.6046") @@ -3080,6 +3368,7 @@ class Program_weight_tensor_parameter_287: class Program_weight_tensor_parameter_288: name = "parameter_288" + original_name = "conv2d_33.w_0" shape = [96, 176, 1, 1] dtype = "float32" min_val = float("-0.321227") @@ -3091,6 +3380,7 @@ class Program_weight_tensor_parameter_288: class Program_weight_tensor_parameter_289: name = "parameter_289" + original_name = "batch_norm2d_28.b_0" shape = [96] dtype = "float32" min_val = float("-0.0887356") @@ -3102,6 +3392,7 @@ class Program_weight_tensor_parameter_289: class Program_weight_tensor_parameter_290: name = "parameter_290" + original_name = "batch_norm2d_28.w_0" shape = [96] dtype = "float32" min_val = float("0.0358205") @@ -3113,6 +3404,7 @@ class Program_weight_tensor_parameter_290: class Program_weight_tensor_parameter_291: name = "parameter_291" + original_name = "batch_norm2d_28.w_2" shape = [96] dtype = "float32" min_val = float("0.0883915") @@ -3124,6 +3416,7 @@ class Program_weight_tensor_parameter_291: class Program_weight_tensor_parameter_292: name = "parameter_292" + original_name = "batch_norm2d_28.w_1" shape = [96] dtype = "float32" min_val = float("-1.36234") @@ -3135,6 +3428,7 @@ class Program_weight_tensor_parameter_292: class Program_weight_tensor_parameter_293: name = "parameter_293" + original_name = "conv2d_32.w_0" shape = [96, 88, 1, 1] dtype = "float32" min_val = float("-0.291628") @@ -3146,6 +3440,7 @@ class Program_weight_tensor_parameter_293: class Program_weight_tensor_parameter_294: name = "parameter_294" + original_name = "batch_norm2d_27.b_0" shape = [96] dtype = "float32" min_val = float("-0.470974") @@ -3157,6 +3452,7 @@ class Program_weight_tensor_parameter_294: class Program_weight_tensor_parameter_295: name = "parameter_295" + original_name = "batch_norm2d_27.w_0" shape = [96] dtype = "float32" min_val = float("0.0277848") @@ -3168,6 +3464,7 @@ class Program_weight_tensor_parameter_295: class Program_weight_tensor_parameter_296: name = "parameter_296" + original_name = "batch_norm2d_27.w_2" shape = [96] dtype = "float32" min_val = float("0.0882263") @@ -3179,6 +3476,7 @@ class Program_weight_tensor_parameter_296: class Program_weight_tensor_parameter_297: name = "parameter_297" + original_name = "batch_norm2d_27.w_1" shape = [96] dtype = "float32" min_val = float("-1.75989") @@ -3190,6 +3488,7 @@ class Program_weight_tensor_parameter_297: class Program_weight_tensor_parameter_298: name = "parameter_298" + original_name = "conv2d_31.w_0" shape = [96, 48, 1, 1] dtype = "float32" min_val = float("-0.363594") @@ -3201,6 +3500,7 @@ class Program_weight_tensor_parameter_298: class Program_weight_tensor_parameter_299: name = "parameter_299" + original_name = "batch_norm2d_26.b_0" shape = [176] dtype = "float32" min_val = float("-9.11117") @@ -3212,6 +3512,7 @@ class Program_weight_tensor_parameter_299: class Program_weight_tensor_parameter_300: name = "parameter_300" + original_name = "batch_norm2d_26.w_0" shape = [176] dtype = "float32" min_val = float("3.5604") @@ -3223,6 +3524,7 @@ class Program_weight_tensor_parameter_300: class Program_weight_tensor_parameter_301: name = "parameter_301" + original_name = "batch_norm2d_26.w_2" shape = [176] dtype = "float32" min_val = float("0.0240968") @@ -3234,6 +3536,7 @@ class Program_weight_tensor_parameter_301: class Program_weight_tensor_parameter_302: name = "parameter_302" + original_name = "batch_norm2d_26.w_1" shape = [176] dtype = "float32" min_val = float("-0.447343") @@ -3245,6 +3548,7 @@ class Program_weight_tensor_parameter_302: class Program_weight_tensor_parameter_303: name = "parameter_303" + original_name = "conv2d_30.w_0" shape = [176, 176, 1, 1] dtype = "float32" min_val = float("-0.687812") @@ -3256,6 +3560,7 @@ class Program_weight_tensor_parameter_303: class Program_weight_tensor_parameter_304: name = "parameter_304" + original_name = "conv2d_29.b_0" shape = [176] dtype = "float32" min_val = float("-0.150743") @@ -3267,6 +3572,7 @@ class Program_weight_tensor_parameter_304: class Program_weight_tensor_parameter_305: name = "parameter_305" + original_name = "conv2d_29.w_0" shape = [176, 44, 1, 1] dtype = "float32" min_val = float("-0.621957") @@ -3278,6 +3584,7 @@ class Program_weight_tensor_parameter_305: class Program_weight_tensor_parameter_306: name = "parameter_306" + original_name = "conv2d_28.b_0" shape = [44] dtype = "float32" min_val = float("0") @@ -3287,6 +3594,7 @@ class Program_weight_tensor_parameter_306: class Program_weight_tensor_parameter_307: name = "parameter_307" + original_name = "conv2d_28.w_0" shape = [44, 176, 1, 1] dtype = "float32" min_val = float("-0.624865") @@ -3298,6 +3606,7 @@ class Program_weight_tensor_parameter_307: class Program_weight_tensor_parameter_308: name = "parameter_308" + original_name = "batch_norm2d_25.b_0" shape = [176] dtype = "float32" min_val = float("-1.71231") @@ -3309,6 +3618,7 @@ class Program_weight_tensor_parameter_308: class Program_weight_tensor_parameter_309: name = "parameter_309" + original_name = "batch_norm2d_25.w_0" shape = [176] dtype = "float32" min_val = float("-0.843936") @@ -3320,6 +3630,7 @@ class Program_weight_tensor_parameter_309: class Program_weight_tensor_parameter_310: name = "parameter_310" + original_name = "batch_norm2d_25.w_2" shape = [176] dtype = "float32" min_val = float("0.000142641") @@ -3331,6 +3642,7 @@ class Program_weight_tensor_parameter_310: class Program_weight_tensor_parameter_311: name = "parameter_311" + original_name = "batch_norm2d_25.w_1" shape = [176] dtype = "float32" min_val = float("-5.35663") @@ -3342,6 +3654,7 @@ class Program_weight_tensor_parameter_311: class Program_weight_tensor_parameter_312: name = "parameter_312" + original_name = "conv2d_27.w_0" shape = [176, 1, 5, 5] dtype = "float32" min_val = float("-0.58706") @@ -3353,6 +3666,7 @@ class Program_weight_tensor_parameter_312: class Program_weight_tensor_parameter_313: name = "parameter_313" + original_name = "batch_norm2d_24.b_0" shape = [176] dtype = "float32" min_val = float("-2.35882") @@ -3364,6 +3678,7 @@ class Program_weight_tensor_parameter_313: class Program_weight_tensor_parameter_314: name = "parameter_314" + original_name = "batch_norm2d_24.w_0" shape = [176] dtype = "float32" min_val = float("-0.301828") @@ -3375,6 +3690,7 @@ class Program_weight_tensor_parameter_314: class Program_weight_tensor_parameter_315: name = "parameter_315" + original_name = "batch_norm2d_24.w_2" shape = [176] dtype = "float32" min_val = float("0.0384658") @@ -3386,6 +3702,7 @@ class Program_weight_tensor_parameter_315: class Program_weight_tensor_parameter_316: name = "parameter_316" + original_name = "batch_norm2d_24.w_1" shape = [176] dtype = "float32" min_val = float("-0.979925") @@ -3397,6 +3714,7 @@ class Program_weight_tensor_parameter_316: class Program_weight_tensor_parameter_317: name = "parameter_317" + original_name = "conv2d_26.w_0" shape = [176, 88, 1, 1] dtype = "float32" min_val = float("-0.488285") @@ -3408,6 +3726,7 @@ class Program_weight_tensor_parameter_317: class Program_weight_tensor_parameter_318: name = "parameter_318" + original_name = "conv2d_25.b_0" shape = [88] dtype = "float32" min_val = float("-0.141012") @@ -3419,6 +3738,7 @@ class Program_weight_tensor_parameter_318: class Program_weight_tensor_parameter_319: name = "parameter_319" + original_name = "conv2d_25.w_0" shape = [88, 22, 1, 1] dtype = "float32" min_val = float("-0.597058") @@ -3430,6 +3750,7 @@ class Program_weight_tensor_parameter_319: class Program_weight_tensor_parameter_320: name = "parameter_320" + original_name = "conv2d_24.b_0" shape = [22] dtype = "float32" min_val = float("0") @@ -3439,6 +3760,7 @@ class Program_weight_tensor_parameter_320: class Program_weight_tensor_parameter_321: name = "parameter_321" + original_name = "conv2d_24.w_0" shape = [22, 88, 1, 1] dtype = "float32" min_val = float("-0.548615") @@ -3450,6 +3772,7 @@ class Program_weight_tensor_parameter_321: class Program_weight_tensor_parameter_322: name = "parameter_322" + original_name = "batch_norm2d_23.b_0" shape = [88] dtype = "float32" min_val = float("-1.24322") @@ -3461,6 +3784,7 @@ class Program_weight_tensor_parameter_322: class Program_weight_tensor_parameter_323: name = "parameter_323" + original_name = "batch_norm2d_23.w_0" shape = [88] dtype = "float32" min_val = float("-1.1364") @@ -3472,6 +3796,7 @@ class Program_weight_tensor_parameter_323: class Program_weight_tensor_parameter_324: name = "parameter_324" + original_name = "batch_norm2d_23.w_2" shape = [88] dtype = "float32" min_val = float("0.000182405") @@ -3483,6 +3808,7 @@ class Program_weight_tensor_parameter_324: class Program_weight_tensor_parameter_325: name = "parameter_325" + original_name = "batch_norm2d_23.w_1" shape = [88] dtype = "float32" min_val = float("-6.44533") @@ -3494,6 +3820,7 @@ class Program_weight_tensor_parameter_325: class Program_weight_tensor_parameter_326: name = "parameter_326" + original_name = "conv2d_23.w_0" shape = [88, 1, 5, 5] dtype = "float32" min_val = float("-0.238231") @@ -3505,6 +3832,7 @@ class Program_weight_tensor_parameter_326: class Program_weight_tensor_parameter_327: name = "parameter_327" + original_name = "batch_norm2d_22.b_0" shape = [88] dtype = "float32" min_val = float("-2.06878") @@ -3516,6 +3844,7 @@ class Program_weight_tensor_parameter_327: class Program_weight_tensor_parameter_328: name = "parameter_328" + original_name = "batch_norm2d_22.w_0" shape = [88] dtype = "float32" min_val = float("-1.68307") @@ -3527,6 +3856,7 @@ class Program_weight_tensor_parameter_328: class Program_weight_tensor_parameter_329: name = "parameter_329" + original_name = "batch_norm2d_22.w_2" shape = [88] dtype = "float32" min_val = float("0.337157") @@ -3538,6 +3868,7 @@ class Program_weight_tensor_parameter_329: class Program_weight_tensor_parameter_330: name = "parameter_330" + original_name = "batch_norm2d_22.w_1" shape = [88] dtype = "float32" min_val = float("-4.87676") @@ -3549,6 +3880,7 @@ class Program_weight_tensor_parameter_330: class Program_weight_tensor_parameter_331: name = "parameter_331" + original_name = "conv2d_22.w_0" shape = [88, 88, 1, 1] dtype = "float32" min_val = float("-0.945244") @@ -3560,6 +3892,7 @@ class Program_weight_tensor_parameter_331: class Program_weight_tensor_parameter_332: name = "parameter_332" + original_name = "batch_norm2d_21.b_0" shape = [88] dtype = "float32" min_val = float("-0.900002") @@ -3571,6 +3904,7 @@ class Program_weight_tensor_parameter_332: class Program_weight_tensor_parameter_333: name = "parameter_333" + original_name = "batch_norm2d_21.w_0" shape = [88] dtype = "float32" min_val = float("0.401021") @@ -3582,6 +3916,7 @@ class Program_weight_tensor_parameter_333: class Program_weight_tensor_parameter_334: name = "parameter_334" + original_name = "batch_norm2d_21.w_2" shape = [88] dtype = "float32" min_val = float("0.000260068") @@ -3593,6 +3928,7 @@ class Program_weight_tensor_parameter_334: class Program_weight_tensor_parameter_335: name = "parameter_335" + original_name = "batch_norm2d_21.w_1" shape = [88] dtype = "float32" min_val = float("-15.9912") @@ -3604,6 +3940,7 @@ class Program_weight_tensor_parameter_335: class Program_weight_tensor_parameter_336: name = "parameter_336" + original_name = "conv2d_21.w_0" shape = [88, 1, 5, 5] dtype = "float32" min_val = float("-0.828417") @@ -3615,6 +3952,7 @@ class Program_weight_tensor_parameter_336: class Program_weight_tensor_parameter_337: name = "parameter_337" + original_name = "batch_norm2d_20.b_0" shape = [88] dtype = "float32" min_val = float("-2.24673") @@ -3626,6 +3964,7 @@ class Program_weight_tensor_parameter_337: class Program_weight_tensor_parameter_338: name = "parameter_338" + original_name = "batch_norm2d_20.w_0" shape = [88] dtype = "float32" min_val = float("-0.943668") @@ -3637,6 +3976,7 @@ class Program_weight_tensor_parameter_338: class Program_weight_tensor_parameter_339: name = "parameter_339" + original_name = "batch_norm2d_20.w_2" shape = [88] dtype = "float32" min_val = float("0.477999") @@ -3648,6 +3988,7 @@ class Program_weight_tensor_parameter_339: class Program_weight_tensor_parameter_340: name = "parameter_340" + original_name = "batch_norm2d_20.w_1" shape = [88] dtype = "float32" min_val = float("-6.83725") @@ -3659,6 +4000,7 @@ class Program_weight_tensor_parameter_340: class Program_weight_tensor_parameter_341: name = "parameter_341" + original_name = "conv2d_20.w_0" shape = [88, 88, 1, 1] dtype = "float32" min_val = float("-0.592371") @@ -3670,6 +4012,7 @@ class Program_weight_tensor_parameter_341: class Program_weight_tensor_parameter_342: name = "parameter_342" + original_name = "batch_norm2d_19.b_0" shape = [88] dtype = "float32" min_val = float("-1.8535") @@ -3681,6 +4024,7 @@ class Program_weight_tensor_parameter_342: class Program_weight_tensor_parameter_343: name = "parameter_343" + original_name = "batch_norm2d_19.w_0" shape = [88] dtype = "float32" min_val = float("-1.03791") @@ -3692,6 +4036,7 @@ class Program_weight_tensor_parameter_343: class Program_weight_tensor_parameter_344: name = "parameter_344" + original_name = "batch_norm2d_19.w_2" shape = [88] dtype = "float32" min_val = float("0.000237056") @@ -3703,6 +4048,7 @@ class Program_weight_tensor_parameter_344: class Program_weight_tensor_parameter_345: name = "parameter_345" + original_name = "batch_norm2d_19.w_1" shape = [88] dtype = "float32" min_val = float("-9.88008") @@ -3714,6 +4060,7 @@ class Program_weight_tensor_parameter_345: class Program_weight_tensor_parameter_346: name = "parameter_346" + original_name = "conv2d_19.w_0" shape = [88, 1, 5, 5] dtype = "float32" min_val = float("-0.86593") @@ -3725,6 +4072,7 @@ class Program_weight_tensor_parameter_346: class Program_weight_tensor_parameter_347: name = "parameter_347" + original_name = "batch_norm2d_18.b_0" shape = [88] dtype = "float32" min_val = float("-1.75172") @@ -3736,6 +4084,7 @@ class Program_weight_tensor_parameter_347: class Program_weight_tensor_parameter_348: name = "parameter_348" + original_name = "batch_norm2d_18.w_0" shape = [88] dtype = "float32" min_val = float("-1.13233") @@ -3747,6 +4096,7 @@ class Program_weight_tensor_parameter_348: class Program_weight_tensor_parameter_349: name = "parameter_349" + original_name = "batch_norm2d_18.w_2" shape = [88] dtype = "float32" min_val = float("0.47541") @@ -3758,6 +4108,7 @@ class Program_weight_tensor_parameter_349: class Program_weight_tensor_parameter_350: name = "parameter_350" + original_name = "batch_norm2d_18.w_1" shape = [88] dtype = "float32" min_val = float("-5.66211") @@ -3769,6 +4120,7 @@ class Program_weight_tensor_parameter_350: class Program_weight_tensor_parameter_351: name = "parameter_351" + original_name = "conv2d_18.w_0" shape = [88, 88, 1, 1] dtype = "float32" min_val = float("-0.688036") @@ -3780,6 +4132,7 @@ class Program_weight_tensor_parameter_351: class Program_weight_tensor_parameter_352: name = "parameter_352" + original_name = "batch_norm2d_17.b_0" shape = [88] dtype = "float32" min_val = float("-1.46937") @@ -3791,6 +4144,7 @@ class Program_weight_tensor_parameter_352: class Program_weight_tensor_parameter_353: name = "parameter_353" + original_name = "batch_norm2d_17.w_0" shape = [88] dtype = "float32" min_val = float("0.548747") @@ -3802,6 +4156,7 @@ class Program_weight_tensor_parameter_353: class Program_weight_tensor_parameter_354: name = "parameter_354" + original_name = "batch_norm2d_17.w_2" shape = [88] dtype = "float32" min_val = float("0.000154021") @@ -3813,6 +4168,7 @@ class Program_weight_tensor_parameter_354: class Program_weight_tensor_parameter_355: name = "parameter_355" + original_name = "batch_norm2d_17.w_1" shape = [88] dtype = "float32" min_val = float("-11.3024") @@ -3824,6 +4180,7 @@ class Program_weight_tensor_parameter_355: class Program_weight_tensor_parameter_356: name = "parameter_356" + original_name = "conv2d_17.w_0" shape = [88, 1, 5, 5] dtype = "float32" min_val = float("-0.861084") @@ -3835,6 +4192,7 @@ class Program_weight_tensor_parameter_356: class Program_weight_tensor_parameter_357: name = "parameter_357" + original_name = "batch_norm2d_16.b_0" shape = [88] dtype = "float32" min_val = float("-1.97382") @@ -3846,6 +4204,7 @@ class Program_weight_tensor_parameter_357: class Program_weight_tensor_parameter_358: name = "parameter_358" + original_name = "batch_norm2d_16.w_0" shape = [88] dtype = "float32" min_val = float("-0.210887") @@ -3857,6 +4216,7 @@ class Program_weight_tensor_parameter_358: class Program_weight_tensor_parameter_359: name = "parameter_359" + original_name = "batch_norm2d_16.w_2" shape = [88] dtype = "float32" min_val = float("0.336744") @@ -3868,6 +4228,7 @@ class Program_weight_tensor_parameter_359: class Program_weight_tensor_parameter_360: name = "parameter_360" + original_name = "batch_norm2d_16.w_1" shape = [88] dtype = "float32" min_val = float("-6.64915") @@ -3879,6 +4240,7 @@ class Program_weight_tensor_parameter_360: class Program_weight_tensor_parameter_361: name = "parameter_361" + original_name = "conv2d_16.w_0" shape = [88, 88, 1, 1] dtype = "float32" min_val = float("-0.776418") @@ -3890,6 +4252,7 @@ class Program_weight_tensor_parameter_361: class Program_weight_tensor_parameter_362: name = "parameter_362" + original_name = "batch_norm2d_15.b_0" shape = [88] dtype = "float32" min_val = float("-1.53344") @@ -3901,6 +4264,7 @@ class Program_weight_tensor_parameter_362: class Program_weight_tensor_parameter_363: name = "parameter_363" + original_name = "batch_norm2d_15.w_0" shape = [88] dtype = "float32" min_val = float("0.413397") @@ -3912,6 +4276,7 @@ class Program_weight_tensor_parameter_363: class Program_weight_tensor_parameter_364: name = "parameter_364" + original_name = "batch_norm2d_15.w_2" shape = [88] dtype = "float32" min_val = float("0.000399659") @@ -3923,6 +4288,7 @@ class Program_weight_tensor_parameter_364: class Program_weight_tensor_parameter_365: name = "parameter_365" + original_name = "batch_norm2d_15.w_1" shape = [88] dtype = "float32" min_val = float("-47.3395") @@ -3934,6 +4300,7 @@ class Program_weight_tensor_parameter_365: class Program_weight_tensor_parameter_366: name = "parameter_366" + original_name = "conv2d_15.w_0" shape = [88, 1, 5, 5] dtype = "float32" min_val = float("-1.1889") @@ -3945,6 +4312,7 @@ class Program_weight_tensor_parameter_366: class Program_weight_tensor_parameter_367: name = "parameter_367" + original_name = "batch_norm2d_14.b_0" shape = [88] dtype = "float32" min_val = float("-1.64676") @@ -3956,6 +4324,7 @@ class Program_weight_tensor_parameter_367: class Program_weight_tensor_parameter_368: name = "parameter_368" + original_name = "batch_norm2d_14.w_0" shape = [88] dtype = "float32" min_val = float("-1.29096") @@ -3967,6 +4336,7 @@ class Program_weight_tensor_parameter_368: class Program_weight_tensor_parameter_369: name = "parameter_369" + original_name = "batch_norm2d_14.w_2" shape = [88] dtype = "float32" min_val = float("0.477523") @@ -3978,6 +4348,7 @@ class Program_weight_tensor_parameter_369: class Program_weight_tensor_parameter_370: name = "parameter_370" + original_name = "batch_norm2d_14.w_1" shape = [88] dtype = "float32" min_val = float("-13.3074") @@ -3989,6 +4360,7 @@ class Program_weight_tensor_parameter_370: class Program_weight_tensor_parameter_371: name = "parameter_371" + original_name = "conv2d_14.w_0" shape = [88, 88, 1, 1] dtype = "float32" min_val = float("-0.988244") @@ -4000,6 +4372,7 @@ class Program_weight_tensor_parameter_371: class Program_weight_tensor_parameter_372: name = "parameter_372" + original_name = "batch_norm2d_13.b_0" shape = [88] dtype = "float32" min_val = float("-2.36244") @@ -4011,6 +4384,7 @@ class Program_weight_tensor_parameter_372: class Program_weight_tensor_parameter_373: name = "parameter_373" + original_name = "batch_norm2d_13.w_0" shape = [88] dtype = "float32" min_val = float("0.622812") @@ -4022,6 +4396,7 @@ class Program_weight_tensor_parameter_373: class Program_weight_tensor_parameter_374: name = "parameter_374" + original_name = "batch_norm2d_13.w_2" shape = [88] dtype = "float32" min_val = float("0.000163461") @@ -4033,6 +4408,7 @@ class Program_weight_tensor_parameter_374: class Program_weight_tensor_parameter_375: name = "parameter_375" + original_name = "batch_norm2d_13.w_1" shape = [88] dtype = "float32" min_val = float("-5.71862") @@ -4044,6 +4420,7 @@ class Program_weight_tensor_parameter_375: class Program_weight_tensor_parameter_376: name = "parameter_376" + original_name = "conv2d_13.w_0" shape = [88, 1, 5, 5] dtype = "float32" min_val = float("-1.04337") @@ -4055,6 +4432,7 @@ class Program_weight_tensor_parameter_376: class Program_weight_tensor_parameter_377: name = "parameter_377" + original_name = "batch_norm2d_12.b_0" shape = [88] dtype = "float32" min_val = float("-1.62014") @@ -4066,6 +4444,7 @@ class Program_weight_tensor_parameter_377: class Program_weight_tensor_parameter_378: name = "parameter_378" + original_name = "batch_norm2d_12.w_0" shape = [88] dtype = "float32" min_val = float("-0.905831") @@ -4077,6 +4456,7 @@ class Program_weight_tensor_parameter_378: class Program_weight_tensor_parameter_379: name = "parameter_379" + original_name = "batch_norm2d_12.w_2" shape = [88] dtype = "float32" min_val = float("0.146372") @@ -4088,6 +4468,7 @@ class Program_weight_tensor_parameter_379: class Program_weight_tensor_parameter_380: name = "parameter_380" + original_name = "batch_norm2d_12.w_1" shape = [88] dtype = "float32" min_val = float("-9.26248") @@ -4099,6 +4480,7 @@ class Program_weight_tensor_parameter_380: class Program_weight_tensor_parameter_381: name = "parameter_381" + original_name = "conv2d_12.w_0" shape = [88, 48, 1, 1] dtype = "float32" min_val = float("-0.848392") @@ -4110,6 +4492,7 @@ class Program_weight_tensor_parameter_381: class Program_weight_tensor_parameter_382: name = "parameter_382" + original_name = "batch_norm2d_11.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -4119,6 +4502,7 @@ class Program_weight_tensor_parameter_382: class Program_weight_tensor_parameter_383: name = "parameter_383" + original_name = "batch_norm2d_11.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -4128,6 +4512,7 @@ class Program_weight_tensor_parameter_383: class Program_weight_tensor_parameter_384: name = "parameter_384" + original_name = "batch_norm2d_11.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -4137,6 +4522,7 @@ class Program_weight_tensor_parameter_384: class Program_weight_tensor_parameter_385: name = "parameter_385" + original_name = "batch_norm2d_11.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -4146,6 +4532,7 @@ class Program_weight_tensor_parameter_385: class Program_weight_tensor_parameter_386: name = "parameter_386" + original_name = "conv2d_11.w_0" shape = [48, 1, 3, 3] dtype = "float32" min_val = float("-0.299872") @@ -4157,6 +4544,7 @@ class Program_weight_tensor_parameter_386: class Program_weight_tensor_parameter_387: name = "parameter_387" + original_name = "batch_norm2d_10.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -4166,6 +4554,7 @@ class Program_weight_tensor_parameter_387: class Program_weight_tensor_parameter_388: name = "parameter_388" + original_name = "batch_norm2d_10.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -4175,6 +4564,7 @@ class Program_weight_tensor_parameter_388: class Program_weight_tensor_parameter_389: name = "parameter_389" + original_name = "batch_norm2d_10.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -4184,6 +4574,7 @@ class Program_weight_tensor_parameter_389: class Program_weight_tensor_parameter_390: name = "parameter_390" + original_name = "batch_norm2d_10.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -4193,6 +4584,7 @@ class Program_weight_tensor_parameter_390: class Program_weight_tensor_parameter_391: name = "parameter_391" + original_name = "conv2d_10.w_0" shape = [48, 48, 1, 1] dtype = "float32" min_val = float("-1.09539") @@ -4204,6 +4596,7 @@ class Program_weight_tensor_parameter_391: class Program_weight_tensor_parameter_392: name = "parameter_392" + original_name = "batch_norm2d_9.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -4213,6 +4606,7 @@ class Program_weight_tensor_parameter_392: class Program_weight_tensor_parameter_393: name = "parameter_393" + original_name = "batch_norm2d_9.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -4222,6 +4616,7 @@ class Program_weight_tensor_parameter_393: class Program_weight_tensor_parameter_394: name = "parameter_394" + original_name = "batch_norm2d_9.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -4231,6 +4626,7 @@ class Program_weight_tensor_parameter_394: class Program_weight_tensor_parameter_395: name = "parameter_395" + original_name = "batch_norm2d_9.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -4240,6 +4636,7 @@ class Program_weight_tensor_parameter_395: class Program_weight_tensor_parameter_396: name = "parameter_396" + original_name = "conv2d_9.w_0" shape = [48, 1, 3, 3] dtype = "float32" min_val = float("-1.13346") @@ -4251,6 +4648,7 @@ class Program_weight_tensor_parameter_396: class Program_weight_tensor_parameter_397: name = "parameter_397" + original_name = "batch_norm2d_8.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -4260,6 +4658,7 @@ class Program_weight_tensor_parameter_397: class Program_weight_tensor_parameter_398: name = "parameter_398" + original_name = "batch_norm2d_8.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -4269,6 +4668,7 @@ class Program_weight_tensor_parameter_398: class Program_weight_tensor_parameter_399: name = "parameter_399" + original_name = "batch_norm2d_8.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -4278,6 +4678,7 @@ class Program_weight_tensor_parameter_399: class Program_weight_tensor_parameter_400: name = "parameter_400" + original_name = "batch_norm2d_8.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -4287,6 +4688,7 @@ class Program_weight_tensor_parameter_400: class Program_weight_tensor_parameter_401: name = "parameter_401" + original_name = "conv2d_8.w_0" shape = [48, 24, 1, 1] dtype = "float32" min_val = float("-1.19459") @@ -4298,6 +4700,7 @@ class Program_weight_tensor_parameter_401: class Program_weight_tensor_parameter_402: name = "parameter_402" + original_name = "batch_norm2d_7.b_0" shape = [24] dtype = "float32" min_val = float("0") @@ -4307,6 +4710,7 @@ class Program_weight_tensor_parameter_402: class Program_weight_tensor_parameter_403: name = "parameter_403" + original_name = "batch_norm2d_7.w_0" shape = [24] dtype = "float32" min_val = float("0") @@ -4316,6 +4720,7 @@ class Program_weight_tensor_parameter_403: class Program_weight_tensor_parameter_404: name = "parameter_404" + original_name = "batch_norm2d_7.w_2" shape = [24] dtype = "float32" min_val = float("0") @@ -4325,6 +4730,7 @@ class Program_weight_tensor_parameter_404: class Program_weight_tensor_parameter_405: name = "parameter_405" + original_name = "batch_norm2d_7.w_1" shape = [24] dtype = "float32" min_val = float("0") @@ -4334,6 +4740,7 @@ class Program_weight_tensor_parameter_405: class Program_weight_tensor_parameter_406: name = "parameter_406" + original_name = "conv2d_7.w_0" shape = [24, 1, 3, 3] dtype = "float32" min_val = float("-0.445645") @@ -4345,6 +4752,7 @@ class Program_weight_tensor_parameter_406: class Program_weight_tensor_parameter_407: name = "parameter_407" + original_name = "batch_norm2d_6.b_0" shape = [24] dtype = "float32" min_val = float("0") @@ -4354,6 +4762,7 @@ class Program_weight_tensor_parameter_407: class Program_weight_tensor_parameter_408: name = "parameter_408" + original_name = "batch_norm2d_6.w_0" shape = [24] dtype = "float32" min_val = float("0") @@ -4363,6 +4772,7 @@ class Program_weight_tensor_parameter_408: class Program_weight_tensor_parameter_409: name = "parameter_409" + original_name = "batch_norm2d_6.w_2" shape = [24] dtype = "float32" min_val = float("0") @@ -4372,6 +4782,7 @@ class Program_weight_tensor_parameter_409: class Program_weight_tensor_parameter_410: name = "parameter_410" + original_name = "batch_norm2d_6.w_1" shape = [24] dtype = "float32" min_val = float("0") @@ -4381,6 +4792,7 @@ class Program_weight_tensor_parameter_410: class Program_weight_tensor_parameter_411: name = "parameter_411" + original_name = "conv2d_6.w_0" shape = [24, 24, 1, 1] dtype = "float32" min_val = float("-1.34903") @@ -4392,6 +4804,7 @@ class Program_weight_tensor_parameter_411: class Program_weight_tensor_parameter_412: name = "parameter_412" + original_name = "batch_norm2d_5.b_0" shape = [24] dtype = "float32" min_val = float("0") @@ -4401,6 +4814,7 @@ class Program_weight_tensor_parameter_412: class Program_weight_tensor_parameter_413: name = "parameter_413" + original_name = "batch_norm2d_5.w_0" shape = [24] dtype = "float32" min_val = float("0") @@ -4410,6 +4824,7 @@ class Program_weight_tensor_parameter_413: class Program_weight_tensor_parameter_414: name = "parameter_414" + original_name = "batch_norm2d_5.w_2" shape = [24] dtype = "float32" min_val = float("0") @@ -4419,6 +4834,7 @@ class Program_weight_tensor_parameter_414: class Program_weight_tensor_parameter_415: name = "parameter_415" + original_name = "batch_norm2d_5.w_1" shape = [24] dtype = "float32" min_val = float("0") @@ -4428,6 +4844,7 @@ class Program_weight_tensor_parameter_415: class Program_weight_tensor_parameter_416: name = "parameter_416" + original_name = "conv2d_5.w_0" shape = [24, 1, 3, 3] dtype = "float32" min_val = float("-1.18107") @@ -4439,6 +4856,7 @@ class Program_weight_tensor_parameter_416: class Program_weight_tensor_parameter_417: name = "parameter_417" + original_name = "batch_norm2d_4.b_0" shape = [24] dtype = "float32" min_val = float("0") @@ -4448,6 +4866,7 @@ class Program_weight_tensor_parameter_417: class Program_weight_tensor_parameter_418: name = "parameter_418" + original_name = "batch_norm2d_4.w_0" shape = [24] dtype = "float32" min_val = float("0") @@ -4457,6 +4876,7 @@ class Program_weight_tensor_parameter_418: class Program_weight_tensor_parameter_419: name = "parameter_419" + original_name = "batch_norm2d_4.w_2" shape = [24] dtype = "float32" min_val = float("0") @@ -4466,6 +4886,7 @@ class Program_weight_tensor_parameter_419: class Program_weight_tensor_parameter_420: name = "parameter_420" + original_name = "batch_norm2d_4.w_1" shape = [24] dtype = "float32" min_val = float("0") @@ -4475,6 +4896,7 @@ class Program_weight_tensor_parameter_420: class Program_weight_tensor_parameter_421: name = "parameter_421" + original_name = "conv2d_4.w_0" shape = [24, 16, 1, 1] dtype = "float32" min_val = float("-1.82103") @@ -4486,6 +4908,7 @@ class Program_weight_tensor_parameter_421: class Program_weight_tensor_parameter_422: name = "parameter_422" + original_name = "batch_norm2d_3.b_0" shape = [16] dtype = "float32" min_val = float("0") @@ -4495,6 +4918,7 @@ class Program_weight_tensor_parameter_422: class Program_weight_tensor_parameter_423: name = "parameter_423" + original_name = "batch_norm2d_3.w_0" shape = [16] dtype = "float32" min_val = float("0") @@ -4504,6 +4928,7 @@ class Program_weight_tensor_parameter_423: class Program_weight_tensor_parameter_424: name = "parameter_424" + original_name = "batch_norm2d_3.w_2" shape = [16] dtype = "float32" min_val = float("0") @@ -4513,6 +4938,7 @@ class Program_weight_tensor_parameter_424: class Program_weight_tensor_parameter_425: name = "parameter_425" + original_name = "batch_norm2d_3.w_1" shape = [16] dtype = "float32" min_val = float("0") @@ -4522,6 +4948,7 @@ class Program_weight_tensor_parameter_425: class Program_weight_tensor_parameter_426: name = "parameter_426" + original_name = "conv2d_3.w_0" shape = [16, 1, 3, 3] dtype = "float32" min_val = float("-0.655494") @@ -4533,6 +4960,7 @@ class Program_weight_tensor_parameter_426: class Program_weight_tensor_parameter_427: name = "parameter_427" + original_name = "batch_norm2d_2.b_0" shape = [16] dtype = "float32" min_val = float("0") @@ -4542,6 +4970,7 @@ class Program_weight_tensor_parameter_427: class Program_weight_tensor_parameter_428: name = "parameter_428" + original_name = "batch_norm2d_2.w_0" shape = [16] dtype = "float32" min_val = float("0") @@ -4551,6 +4980,7 @@ class Program_weight_tensor_parameter_428: class Program_weight_tensor_parameter_429: name = "parameter_429" + original_name = "batch_norm2d_2.w_2" shape = [16] dtype = "float32" min_val = float("0") @@ -4560,6 +4990,7 @@ class Program_weight_tensor_parameter_429: class Program_weight_tensor_parameter_430: name = "parameter_430" + original_name = "batch_norm2d_2.w_1" shape = [16] dtype = "float32" min_val = float("0") @@ -4569,6 +5000,7 @@ class Program_weight_tensor_parameter_430: class Program_weight_tensor_parameter_431: name = "parameter_431" + original_name = "conv2d_2.w_0" shape = [16, 8, 1, 1] dtype = "float32" min_val = float("-1.5035") @@ -4580,6 +5012,7 @@ class Program_weight_tensor_parameter_431: class Program_weight_tensor_parameter_432: name = "parameter_432" + original_name = "batch_norm2d_1.b_0" shape = [8] dtype = "float32" min_val = float("0") @@ -4589,6 +5022,7 @@ class Program_weight_tensor_parameter_432: class Program_weight_tensor_parameter_433: name = "parameter_433" + original_name = "batch_norm2d_1.w_0" shape = [8] dtype = "float32" min_val = float("0") @@ -4598,6 +5032,7 @@ class Program_weight_tensor_parameter_433: class Program_weight_tensor_parameter_434: name = "parameter_434" + original_name = "batch_norm2d_1.w_2" shape = [8] dtype = "float32" min_val = float("0") @@ -4607,6 +5042,7 @@ class Program_weight_tensor_parameter_434: class Program_weight_tensor_parameter_435: name = "parameter_435" + original_name = "batch_norm2d_1.w_1" shape = [8] dtype = "float32" min_val = float("0") @@ -4616,6 +5052,7 @@ class Program_weight_tensor_parameter_435: class Program_weight_tensor_parameter_436: name = "parameter_436" + original_name = "conv2d_1.w_0" shape = [8, 1, 3, 3] dtype = "float32" min_val = float("-2.12633") @@ -4627,6 +5064,7 @@ class Program_weight_tensor_parameter_436: class Program_weight_tensor_parameter_437: name = "parameter_437" + original_name = "batch_norm2d_0.b_0" shape = [8] dtype = "float32" min_val = float("0") @@ -4636,6 +5074,7 @@ class Program_weight_tensor_parameter_437: class Program_weight_tensor_parameter_438: name = "parameter_438" + original_name = "batch_norm2d_0.w_0" shape = [8] dtype = "float32" min_val = float("0") @@ -4645,6 +5084,7 @@ class Program_weight_tensor_parameter_438: class Program_weight_tensor_parameter_439: name = "parameter_439" + original_name = "batch_norm2d_0.w_2" shape = [8] dtype = "float32" min_val = float("0") @@ -4654,6 +5094,7 @@ class Program_weight_tensor_parameter_439: class Program_weight_tensor_parameter_440: name = "parameter_440" + original_name = "batch_norm2d_0.w_1" shape = [8] dtype = "float32" min_val = float("0") @@ -4663,6 +5104,7 @@ class Program_weight_tensor_parameter_440: class Program_weight_tensor_parameter_441: name = "parameter_441" + original_name = "conv2d_0.w_0" shape = [8, 3, 3, 3] dtype = "float32" min_val = float("-1.69757") diff --git a/paddle_samples/PaddleX/ResNet34/input_meta.py b/paddle_samples/PaddleX/ResNet34/input_meta.py index 22400d235..b419d6f62 100644 --- a/paddle_samples/PaddleX/ResNet34/input_meta.py +++ b/paddle_samples/PaddleX/ResNet34/input_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_data_0: name = "data_0" + original_name = "var_257" shape = [64, 3, 224, 224] dtype = "float32" min_val = float("-2.1179") diff --git a/paddle_samples/PaddleX/ResNet34/weight_meta.py b/paddle_samples/PaddleX/ResNet34/weight_meta.py index 36f5c0dfd..dce677fde 100644 --- a/paddle_samples/PaddleX/ResNet34/weight_meta.py +++ b/paddle_samples/PaddleX/ResNet34/weight_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_parameter_0: name = "parameter_0" + original_name = "linear_0.b_0" shape = [102] dtype = "float32" min_val = float("-0.00732447") @@ -11,6 +12,7 @@ class Program_weight_tensor_parameter_0: class Program_weight_tensor_parameter_1: name = "parameter_1" + original_name = "linear_0.w_0" shape = [512, 102] dtype = "float32" min_val = float("-0.0653061") @@ -22,6 +24,7 @@ class Program_weight_tensor_parameter_1: class Program_weight_tensor_parameter_2: name = "parameter_2" + original_name = "batch_norm_36.b_0" shape = [512] dtype = "float32" min_val = float("0.0619078") @@ -33,6 +36,7 @@ class Program_weight_tensor_parameter_2: class Program_weight_tensor_parameter_3: name = "parameter_3" + original_name = "batch_norm_36.w_0" shape = [512] dtype = "float32" min_val = float("1.2881") @@ -44,6 +48,7 @@ class Program_weight_tensor_parameter_3: class Program_weight_tensor_parameter_4: name = "parameter_4" + original_name = "batch_norm_36.w_2" shape = [512] dtype = "float32" min_val = float("0.00750688") @@ -55,6 +60,7 @@ class Program_weight_tensor_parameter_4: class Program_weight_tensor_parameter_5: name = "parameter_5" + original_name = "batch_norm_36.w_1" shape = [512] dtype = "float32" min_val = float("-0.186616") @@ -66,6 +72,7 @@ class Program_weight_tensor_parameter_5: class Program_weight_tensor_parameter_6: name = "parameter_6" + original_name = "conv2d_36.w_0" shape = [512, 512, 3, 3] dtype = "float32" min_val = float("-0.298925") @@ -77,6 +84,7 @@ class Program_weight_tensor_parameter_6: class Program_weight_tensor_parameter_7: name = "parameter_7" + original_name = "batch_norm_35.b_0" shape = [512] dtype = "float32" min_val = float("-0.425072") @@ -88,6 +96,7 @@ class Program_weight_tensor_parameter_7: class Program_weight_tensor_parameter_8: name = "parameter_8" + original_name = "batch_norm_35.w_0" shape = [512] dtype = "float32" min_val = float("0.0835765") @@ -99,6 +108,7 @@ class Program_weight_tensor_parameter_8: class Program_weight_tensor_parameter_9: name = "parameter_9" + original_name = "batch_norm_35.w_2" shape = [512] dtype = "float32" min_val = float("0.187859") @@ -110,6 +120,7 @@ class Program_weight_tensor_parameter_9: class Program_weight_tensor_parameter_10: name = "parameter_10" + original_name = "batch_norm_35.w_1" shape = [512] dtype = "float32" min_val = float("-1.1747") @@ -121,6 +132,7 @@ class Program_weight_tensor_parameter_10: class Program_weight_tensor_parameter_11: name = "parameter_11" + original_name = "conv2d_35.w_0" shape = [512, 512, 3, 3] dtype = "float32" min_val = float("-0.321108") @@ -132,6 +144,7 @@ class Program_weight_tensor_parameter_11: class Program_weight_tensor_parameter_12: name = "parameter_12" + original_name = "batch_norm_34.b_0" shape = [512] dtype = "float32" min_val = float("-0.43092") @@ -143,6 +156,7 @@ class Program_weight_tensor_parameter_12: class Program_weight_tensor_parameter_13: name = "parameter_13" + original_name = "batch_norm_34.w_0" shape = [512] dtype = "float32" min_val = float("0.0606334") @@ -154,6 +168,7 @@ class Program_weight_tensor_parameter_13: class Program_weight_tensor_parameter_14: name = "parameter_14" + original_name = "batch_norm_34.w_2" shape = [512] dtype = "float32" min_val = float("0.00230152") @@ -165,6 +180,7 @@ class Program_weight_tensor_parameter_14: class Program_weight_tensor_parameter_15: name = "parameter_15" + original_name = "batch_norm_34.w_1" shape = [512] dtype = "float32" min_val = float("-0.235395") @@ -176,6 +192,7 @@ class Program_weight_tensor_parameter_15: class Program_weight_tensor_parameter_16: name = "parameter_16" + original_name = "conv2d_34.w_0" shape = [512, 512, 3, 3] dtype = "float32" min_val = float("-0.178384") @@ -187,6 +204,7 @@ class Program_weight_tensor_parameter_16: class Program_weight_tensor_parameter_17: name = "parameter_17" + original_name = "batch_norm_33.b_0" shape = [512] dtype = "float32" min_val = float("-0.448914") @@ -198,6 +216,7 @@ class Program_weight_tensor_parameter_17: class Program_weight_tensor_parameter_18: name = "parameter_18" + original_name = "batch_norm_33.w_0" shape = [512] dtype = "float32" min_val = float("0.0905492") @@ -209,6 +228,7 @@ class Program_weight_tensor_parameter_18: class Program_weight_tensor_parameter_19: name = "parameter_19" + original_name = "batch_norm_33.w_2" shape = [512] dtype = "float32" min_val = float("0.114285") @@ -220,6 +240,7 @@ class Program_weight_tensor_parameter_19: class Program_weight_tensor_parameter_20: name = "parameter_20" + original_name = "batch_norm_33.w_1" shape = [512] dtype = "float32" min_val = float("-1.38396") @@ -231,6 +252,7 @@ class Program_weight_tensor_parameter_20: class Program_weight_tensor_parameter_21: name = "parameter_21" + original_name = "conv2d_33.w_0" shape = [512, 512, 3, 3] dtype = "float32" min_val = float("-0.202471") @@ -242,6 +264,7 @@ class Program_weight_tensor_parameter_21: class Program_weight_tensor_parameter_22: name = "parameter_22" + original_name = "batch_norm_32.b_0" shape = [512] dtype = "float32" min_val = float("-0.283236") @@ -253,6 +276,7 @@ class Program_weight_tensor_parameter_22: class Program_weight_tensor_parameter_23: name = "parameter_23" + original_name = "batch_norm_32.w_0" shape = [512] dtype = "float32" min_val = float("0.0759205") @@ -264,6 +288,7 @@ class Program_weight_tensor_parameter_23: class Program_weight_tensor_parameter_24: name = "parameter_24" + original_name = "batch_norm_32.w_2" shape = [512] dtype = "float32" min_val = float("0.00417477") @@ -275,6 +300,7 @@ class Program_weight_tensor_parameter_24: class Program_weight_tensor_parameter_25: name = "parameter_25" + original_name = "batch_norm_32.w_1" shape = [512] dtype = "float32" min_val = float("-0.722916") @@ -286,6 +312,7 @@ class Program_weight_tensor_parameter_25: class Program_weight_tensor_parameter_26: name = "parameter_26" + original_name = "conv2d_32.w_0" shape = [512, 256, 1, 1] dtype = "float32" min_val = float("-0.362304") @@ -297,6 +324,7 @@ class Program_weight_tensor_parameter_26: class Program_weight_tensor_parameter_27: name = "parameter_27" + original_name = "batch_norm_31.b_0" shape = [512] dtype = "float32" min_val = float("-0.283236") @@ -308,6 +336,7 @@ class Program_weight_tensor_parameter_27: class Program_weight_tensor_parameter_28: name = "parameter_28" + original_name = "batch_norm_31.w_0" shape = [512] dtype = "float32" min_val = float("0.13423") @@ -319,6 +348,7 @@ class Program_weight_tensor_parameter_28: class Program_weight_tensor_parameter_29: name = "parameter_29" + original_name = "batch_norm_31.w_2" shape = [512] dtype = "float32" min_val = float("0.00775836") @@ -330,6 +360,7 @@ class Program_weight_tensor_parameter_29: class Program_weight_tensor_parameter_30: name = "parameter_30" + original_name = "batch_norm_31.w_1" shape = [512] dtype = "float32" min_val = float("-0.280942") @@ -341,6 +372,7 @@ class Program_weight_tensor_parameter_30: class Program_weight_tensor_parameter_31: name = "parameter_31" + original_name = "conv2d_31.w_0" shape = [512, 512, 3, 3] dtype = "float32" min_val = float("-0.185563") @@ -352,6 +384,7 @@ class Program_weight_tensor_parameter_31: class Program_weight_tensor_parameter_32: name = "parameter_32" + original_name = "batch_norm_30.b_0" shape = [512] dtype = "float32" min_val = float("-0.380015") @@ -363,6 +396,7 @@ class Program_weight_tensor_parameter_32: class Program_weight_tensor_parameter_33: name = "parameter_33" + original_name = "batch_norm_30.w_0" shape = [512] dtype = "float32" min_val = float("0.143348") @@ -374,6 +408,7 @@ class Program_weight_tensor_parameter_33: class Program_weight_tensor_parameter_34: name = "parameter_34" + original_name = "batch_norm_30.w_2" shape = [512] dtype = "float32" min_val = float("0.0719699") @@ -385,6 +420,7 @@ class Program_weight_tensor_parameter_34: class Program_weight_tensor_parameter_35: name = "parameter_35" + original_name = "batch_norm_30.w_1" shape = [512] dtype = "float32" min_val = float("-0.605227") @@ -396,6 +432,7 @@ class Program_weight_tensor_parameter_35: class Program_weight_tensor_parameter_36: name = "parameter_36" + original_name = "conv2d_30.w_0" shape = [512, 256, 3, 3] dtype = "float32" min_val = float("-0.21273") @@ -407,6 +444,7 @@ class Program_weight_tensor_parameter_36: class Program_weight_tensor_parameter_37: name = "parameter_37" + original_name = "batch_norm_29.b_0" shape = [256] dtype = "float32" min_val = float("-0.411975") @@ -418,6 +456,7 @@ class Program_weight_tensor_parameter_37: class Program_weight_tensor_parameter_38: name = "parameter_38" + original_name = "batch_norm_29.w_0" shape = [256] dtype = "float32" min_val = float("-0.0436469") @@ -429,6 +468,7 @@ class Program_weight_tensor_parameter_38: class Program_weight_tensor_parameter_39: name = "parameter_39" + original_name = "batch_norm_29.w_2" shape = [256] dtype = "float32" min_val = float("0.000852981") @@ -440,6 +480,7 @@ class Program_weight_tensor_parameter_39: class Program_weight_tensor_parameter_40: name = "parameter_40" + original_name = "batch_norm_29.w_1" shape = [256] dtype = "float32" min_val = float("-0.122012") @@ -451,6 +492,7 @@ class Program_weight_tensor_parameter_40: class Program_weight_tensor_parameter_41: name = "parameter_41" + original_name = "conv2d_29.w_0" shape = [256, 256, 3, 3] dtype = "float32" min_val = float("-0.196337") @@ -462,6 +504,7 @@ class Program_weight_tensor_parameter_41: class Program_weight_tensor_parameter_42: name = "parameter_42" + original_name = "batch_norm_28.b_0" shape = [256] dtype = "float32" min_val = float("-0.464183") @@ -473,6 +516,7 @@ class Program_weight_tensor_parameter_42: class Program_weight_tensor_parameter_43: name = "parameter_43" + original_name = "batch_norm_28.w_0" shape = [256] dtype = "float32" min_val = float("0.12464") @@ -484,6 +528,7 @@ class Program_weight_tensor_parameter_43: class Program_weight_tensor_parameter_44: name = "parameter_44" + original_name = "batch_norm_28.w_2" shape = [256] dtype = "float32" min_val = float("0.0678341") @@ -495,6 +540,7 @@ class Program_weight_tensor_parameter_44: class Program_weight_tensor_parameter_45: name = "parameter_45" + original_name = "batch_norm_28.w_1" shape = [256] dtype = "float32" min_val = float("-0.950659") @@ -506,6 +552,7 @@ class Program_weight_tensor_parameter_45: class Program_weight_tensor_parameter_46: name = "parameter_46" + original_name = "conv2d_28.w_0" shape = [256, 256, 3, 3] dtype = "float32" min_val = float("-0.142044") @@ -517,6 +564,7 @@ class Program_weight_tensor_parameter_46: class Program_weight_tensor_parameter_47: name = "parameter_47" + original_name = "batch_norm_27.b_0" shape = [256] dtype = "float32" min_val = float("-0.43194") @@ -528,6 +576,7 @@ class Program_weight_tensor_parameter_47: class Program_weight_tensor_parameter_48: name = "parameter_48" + original_name = "batch_norm_27.w_0" shape = [256] dtype = "float32" min_val = float("-0.0179444") @@ -539,6 +588,7 @@ class Program_weight_tensor_parameter_48: class Program_weight_tensor_parameter_49: name = "parameter_49" + original_name = "batch_norm_27.w_2" shape = [256] dtype = "float32" min_val = float("0.000798367") @@ -550,6 +600,7 @@ class Program_weight_tensor_parameter_49: class Program_weight_tensor_parameter_50: name = "parameter_50" + original_name = "batch_norm_27.w_1" shape = [256] dtype = "float32" min_val = float("-0.116323") @@ -561,6 +612,7 @@ class Program_weight_tensor_parameter_50: class Program_weight_tensor_parameter_51: name = "parameter_51" + original_name = "conv2d_27.w_0" shape = [256, 256, 3, 3] dtype = "float32" min_val = float("-0.159809") @@ -572,6 +624,7 @@ class Program_weight_tensor_parameter_51: class Program_weight_tensor_parameter_52: name = "parameter_52" + original_name = "batch_norm_26.b_0" shape = [256] dtype = "float32" min_val = float("-0.614435") @@ -583,6 +636,7 @@ class Program_weight_tensor_parameter_52: class Program_weight_tensor_parameter_53: name = "parameter_53" + original_name = "batch_norm_26.w_0" shape = [256] dtype = "float32" min_val = float("0.0882068") @@ -594,6 +648,7 @@ class Program_weight_tensor_parameter_53: class Program_weight_tensor_parameter_54: name = "parameter_54" + original_name = "batch_norm_26.w_2" shape = [256] dtype = "float32" min_val = float("0.0646289") @@ -605,6 +660,7 @@ class Program_weight_tensor_parameter_54: class Program_weight_tensor_parameter_55: name = "parameter_55" + original_name = "batch_norm_26.w_1" shape = [256] dtype = "float32" min_val = float("-1.27113") @@ -616,6 +672,7 @@ class Program_weight_tensor_parameter_55: class Program_weight_tensor_parameter_56: name = "parameter_56" + original_name = "conv2d_26.w_0" shape = [256, 256, 3, 3] dtype = "float32" min_val = float("-0.196362") @@ -627,6 +684,7 @@ class Program_weight_tensor_parameter_56: class Program_weight_tensor_parameter_57: name = "parameter_57" + original_name = "batch_norm_25.b_0" shape = [256] dtype = "float32" min_val = float("-0.334026") @@ -638,6 +696,7 @@ class Program_weight_tensor_parameter_57: class Program_weight_tensor_parameter_58: name = "parameter_58" + original_name = "batch_norm_25.w_0" shape = [256] dtype = "float32" min_val = float("-0.0349018") @@ -649,6 +708,7 @@ class Program_weight_tensor_parameter_58: class Program_weight_tensor_parameter_59: name = "parameter_59" + original_name = "batch_norm_25.w_2" shape = [256] dtype = "float32" min_val = float("0.000437978") @@ -660,6 +720,7 @@ class Program_weight_tensor_parameter_59: class Program_weight_tensor_parameter_60: name = "parameter_60" + original_name = "batch_norm_25.w_1" shape = [256] dtype = "float32" min_val = float("-0.106792") @@ -671,6 +732,7 @@ class Program_weight_tensor_parameter_60: class Program_weight_tensor_parameter_61: name = "parameter_61" + original_name = "conv2d_25.w_0" shape = [256, 256, 3, 3] dtype = "float32" min_val = float("-0.180331") @@ -682,6 +744,7 @@ class Program_weight_tensor_parameter_61: class Program_weight_tensor_parameter_62: name = "parameter_62" + original_name = "batch_norm_24.b_0" shape = [256] dtype = "float32" min_val = float("-0.499412") @@ -693,6 +756,7 @@ class Program_weight_tensor_parameter_62: class Program_weight_tensor_parameter_63: name = "parameter_63" + original_name = "batch_norm_24.w_0" shape = [256] dtype = "float32" min_val = float("0.0969858") @@ -704,6 +768,7 @@ class Program_weight_tensor_parameter_63: class Program_weight_tensor_parameter_64: name = "parameter_64" + original_name = "batch_norm_24.w_2" shape = [256] dtype = "float32" min_val = float("0.0597269") @@ -715,6 +780,7 @@ class Program_weight_tensor_parameter_64: class Program_weight_tensor_parameter_65: name = "parameter_65" + original_name = "batch_norm_24.w_1" shape = [256] dtype = "float32" min_val = float("-1.11839") @@ -726,6 +792,7 @@ class Program_weight_tensor_parameter_65: class Program_weight_tensor_parameter_66: name = "parameter_66" + original_name = "conv2d_24.w_0" shape = [256, 256, 3, 3] dtype = "float32" min_val = float("-0.203343") @@ -737,6 +804,7 @@ class Program_weight_tensor_parameter_66: class Program_weight_tensor_parameter_67: name = "parameter_67" + original_name = "batch_norm_23.b_0" shape = [256] dtype = "float32" min_val = float("-0.261182") @@ -748,6 +816,7 @@ class Program_weight_tensor_parameter_67: class Program_weight_tensor_parameter_68: name = "parameter_68" + original_name = "batch_norm_23.w_0" shape = [256] dtype = "float32" min_val = float("-0.0527634") @@ -759,6 +828,7 @@ class Program_weight_tensor_parameter_68: class Program_weight_tensor_parameter_69: name = "parameter_69" + original_name = "batch_norm_23.w_2" shape = [256] dtype = "float32" min_val = float("0.000791365") @@ -770,6 +840,7 @@ class Program_weight_tensor_parameter_69: class Program_weight_tensor_parameter_70: name = "parameter_70" + original_name = "batch_norm_23.w_1" shape = [256] dtype = "float32" min_val = float("-0.142533") @@ -781,6 +852,7 @@ class Program_weight_tensor_parameter_70: class Program_weight_tensor_parameter_71: name = "parameter_71" + original_name = "conv2d_23.w_0" shape = [256, 256, 3, 3] dtype = "float32" min_val = float("-0.197163") @@ -792,6 +864,7 @@ class Program_weight_tensor_parameter_71: class Program_weight_tensor_parameter_72: name = "parameter_72" + original_name = "batch_norm_22.b_0" shape = [256] dtype = "float32" min_val = float("-0.430694") @@ -803,6 +876,7 @@ class Program_weight_tensor_parameter_72: class Program_weight_tensor_parameter_73: name = "parameter_73" + original_name = "batch_norm_22.w_0" shape = [256] dtype = "float32" min_val = float("0.0896191") @@ -814,6 +888,7 @@ class Program_weight_tensor_parameter_73: class Program_weight_tensor_parameter_74: name = "parameter_74" + original_name = "batch_norm_22.w_2" shape = [256] dtype = "float32" min_val = float("0.0579574") @@ -825,6 +900,7 @@ class Program_weight_tensor_parameter_74: class Program_weight_tensor_parameter_75: name = "parameter_75" + original_name = "batch_norm_22.w_1" shape = [256] dtype = "float32" min_val = float("-1.40902") @@ -836,6 +912,7 @@ class Program_weight_tensor_parameter_75: class Program_weight_tensor_parameter_76: name = "parameter_76" + original_name = "conv2d_22.w_0" shape = [256, 256, 3, 3] dtype = "float32" min_val = float("-0.332837") @@ -847,6 +924,7 @@ class Program_weight_tensor_parameter_76: class Program_weight_tensor_parameter_77: name = "parameter_77" + original_name = "batch_norm_21.b_0" shape = [256] dtype = "float32" min_val = float("-0.373112") @@ -858,6 +936,7 @@ class Program_weight_tensor_parameter_77: class Program_weight_tensor_parameter_78: name = "parameter_78" + original_name = "batch_norm_21.w_0" shape = [256] dtype = "float32" min_val = float("0.0604544") @@ -869,6 +948,7 @@ class Program_weight_tensor_parameter_78: class Program_weight_tensor_parameter_79: name = "parameter_79" + original_name = "batch_norm_21.w_2" shape = [256] dtype = "float32" min_val = float("0.00314414") @@ -880,6 +960,7 @@ class Program_weight_tensor_parameter_79: class Program_weight_tensor_parameter_80: name = "parameter_80" + original_name = "batch_norm_21.w_1" shape = [256] dtype = "float32" min_val = float("-0.4374") @@ -891,6 +972,7 @@ class Program_weight_tensor_parameter_80: class Program_weight_tensor_parameter_81: name = "parameter_81" + original_name = "conv2d_21.w_0" shape = [256, 256, 3, 3] dtype = "float32" min_val = float("-0.415132") @@ -902,6 +984,7 @@ class Program_weight_tensor_parameter_81: class Program_weight_tensor_parameter_82: name = "parameter_82" + original_name = "batch_norm_20.b_0" shape = [256] dtype = "float32" min_val = float("-0.352913") @@ -913,6 +996,7 @@ class Program_weight_tensor_parameter_82: class Program_weight_tensor_parameter_83: name = "parameter_83" + original_name = "batch_norm_20.w_0" shape = [256] dtype = "float32" min_val = float("0.12402") @@ -924,6 +1008,7 @@ class Program_weight_tensor_parameter_83: class Program_weight_tensor_parameter_84: name = "parameter_84" + original_name = "batch_norm_20.w_2" shape = [256] dtype = "float32" min_val = float("0.0848397") @@ -935,6 +1020,7 @@ class Program_weight_tensor_parameter_84: class Program_weight_tensor_parameter_85: name = "parameter_85" + original_name = "batch_norm_20.w_1" shape = [256] dtype = "float32" min_val = float("-3.97347") @@ -946,6 +1032,7 @@ class Program_weight_tensor_parameter_85: class Program_weight_tensor_parameter_86: name = "parameter_86" + original_name = "conv2d_20.w_0" shape = [256, 256, 3, 3] dtype = "float32" min_val = float("-0.311461") @@ -957,6 +1044,7 @@ class Program_weight_tensor_parameter_86: class Program_weight_tensor_parameter_87: name = "parameter_87" + original_name = "batch_norm_19.b_0" shape = [256] dtype = "float32" min_val = float("-0.192546") @@ -968,6 +1056,7 @@ class Program_weight_tensor_parameter_87: class Program_weight_tensor_parameter_88: name = "parameter_88" + original_name = "batch_norm_19.w_0" shape = [256] dtype = "float32" min_val = float("0.000829896") @@ -979,6 +1068,7 @@ class Program_weight_tensor_parameter_88: class Program_weight_tensor_parameter_89: name = "parameter_89" + original_name = "batch_norm_19.w_2" shape = [256] dtype = "float32" min_val = float("0.00182802") @@ -990,6 +1080,7 @@ class Program_weight_tensor_parameter_89: class Program_weight_tensor_parameter_90: name = "parameter_90" + original_name = "batch_norm_19.w_1" shape = [256] dtype = "float32" min_val = float("-0.558411") @@ -1001,6 +1092,7 @@ class Program_weight_tensor_parameter_90: class Program_weight_tensor_parameter_91: name = "parameter_91" + original_name = "conv2d_19.w_0" shape = [256, 128, 1, 1] dtype = "float32" min_val = float("-0.619603") @@ -1012,6 +1104,7 @@ class Program_weight_tensor_parameter_91: class Program_weight_tensor_parameter_92: name = "parameter_92" + original_name = "batch_norm_18.b_0" shape = [256] dtype = "float32" min_val = float("-0.192546") @@ -1023,6 +1116,7 @@ class Program_weight_tensor_parameter_92: class Program_weight_tensor_parameter_93: name = "parameter_93" + original_name = "batch_norm_18.w_0" shape = [256] dtype = "float32" min_val = float("0.164904") @@ -1034,6 +1128,7 @@ class Program_weight_tensor_parameter_93: class Program_weight_tensor_parameter_94: name = "parameter_94" + original_name = "batch_norm_18.w_2" shape = [256] dtype = "float32" min_val = float("0.0146515") @@ -1045,6 +1140,7 @@ class Program_weight_tensor_parameter_94: class Program_weight_tensor_parameter_95: name = "parameter_95" + original_name = "batch_norm_18.w_1" shape = [256] dtype = "float32" min_val = float("-0.634483") @@ -1056,6 +1152,7 @@ class Program_weight_tensor_parameter_95: class Program_weight_tensor_parameter_96: name = "parameter_96" + original_name = "conv2d_18.w_0" shape = [256, 256, 3, 3] dtype = "float32" min_val = float("-0.36627") @@ -1067,6 +1164,7 @@ class Program_weight_tensor_parameter_96: class Program_weight_tensor_parameter_97: name = "parameter_97" + original_name = "batch_norm_17.b_0" shape = [256] dtype = "float32" min_val = float("-0.327893") @@ -1078,6 +1176,7 @@ class Program_weight_tensor_parameter_97: class Program_weight_tensor_parameter_98: name = "parameter_98" + original_name = "batch_norm_17.w_0" shape = [256] dtype = "float32" min_val = float("0.17553") @@ -1089,6 +1188,7 @@ class Program_weight_tensor_parameter_98: class Program_weight_tensor_parameter_99: name = "parameter_99" + original_name = "batch_norm_17.w_2" shape = [256] dtype = "float32" min_val = float("0.108307") @@ -1100,6 +1200,7 @@ class Program_weight_tensor_parameter_99: class Program_weight_tensor_parameter_100: name = "parameter_100" + original_name = "batch_norm_17.w_1" shape = [256] dtype = "float32" min_val = float("-1.72096") @@ -1111,6 +1212,7 @@ class Program_weight_tensor_parameter_100: class Program_weight_tensor_parameter_101: name = "parameter_101" + original_name = "conv2d_17.w_0" shape = [256, 128, 3, 3] dtype = "float32" min_val = float("-0.382215") @@ -1122,6 +1224,7 @@ class Program_weight_tensor_parameter_101: class Program_weight_tensor_parameter_102: name = "parameter_102" + original_name = "batch_norm_16.b_0" shape = [128] dtype = "float32" min_val = float("-0.45121") @@ -1133,6 +1236,7 @@ class Program_weight_tensor_parameter_102: class Program_weight_tensor_parameter_103: name = "parameter_103" + original_name = "batch_norm_16.w_0" shape = [128] dtype = "float32" min_val = float("-0.0248204") @@ -1144,6 +1248,7 @@ class Program_weight_tensor_parameter_103: class Program_weight_tensor_parameter_104: name = "parameter_104" + original_name = "batch_norm_16.w_2" shape = [128] dtype = "float32" min_val = float("0.000772585") @@ -1155,6 +1260,7 @@ class Program_weight_tensor_parameter_104: class Program_weight_tensor_parameter_105: name = "parameter_105" + original_name = "batch_norm_16.w_1" shape = [128] dtype = "float32" min_val = float("-0.101866") @@ -1166,6 +1272,7 @@ class Program_weight_tensor_parameter_105: class Program_weight_tensor_parameter_106: name = "parameter_106" + original_name = "conv2d_16.w_0" shape = [128, 128, 3, 3] dtype = "float32" min_val = float("-0.179024") @@ -1177,6 +1284,7 @@ class Program_weight_tensor_parameter_106: class Program_weight_tensor_parameter_107: name = "parameter_107" + original_name = "batch_norm_15.b_0" shape = [128] dtype = "float32" min_val = float("-0.578049") @@ -1188,6 +1296,7 @@ class Program_weight_tensor_parameter_107: class Program_weight_tensor_parameter_108: name = "parameter_108" + original_name = "batch_norm_15.w_0" shape = [128] dtype = "float32" min_val = float("0.166776") @@ -1199,6 +1308,7 @@ class Program_weight_tensor_parameter_108: class Program_weight_tensor_parameter_109: name = "parameter_109" + original_name = "batch_norm_15.w_2" shape = [128] dtype = "float32" min_val = float("0.0850269") @@ -1210,6 +1320,7 @@ class Program_weight_tensor_parameter_109: class Program_weight_tensor_parameter_110: name = "parameter_110" + original_name = "batch_norm_15.w_1" shape = [128] dtype = "float32" min_val = float("-1.39037") @@ -1221,6 +1332,7 @@ class Program_weight_tensor_parameter_110: class Program_weight_tensor_parameter_111: name = "parameter_111" + original_name = "conv2d_15.w_0" shape = [128, 128, 3, 3] dtype = "float32" min_val = float("-0.225883") @@ -1232,6 +1344,7 @@ class Program_weight_tensor_parameter_111: class Program_weight_tensor_parameter_112: name = "parameter_112" + original_name = "batch_norm_14.b_0" shape = [128] dtype = "float32" min_val = float("-0.294601") @@ -1243,6 +1356,7 @@ class Program_weight_tensor_parameter_112: class Program_weight_tensor_parameter_113: name = "parameter_113" + original_name = "batch_norm_14.w_0" shape = [128] dtype = "float32" min_val = float("-0.0100518") @@ -1254,6 +1368,7 @@ class Program_weight_tensor_parameter_113: class Program_weight_tensor_parameter_114: name = "parameter_114" + original_name = "batch_norm_14.w_2" shape = [128] dtype = "float32" min_val = float("0.00102113") @@ -1265,6 +1380,7 @@ class Program_weight_tensor_parameter_114: class Program_weight_tensor_parameter_115: name = "parameter_115" + original_name = "batch_norm_14.w_1" shape = [128] dtype = "float32" min_val = float("-0.193841") @@ -1276,6 +1392,7 @@ class Program_weight_tensor_parameter_115: class Program_weight_tensor_parameter_116: name = "parameter_116" + original_name = "conv2d_14.w_0" shape = [128, 128, 3, 3] dtype = "float32" min_val = float("-0.216981") @@ -1287,6 +1404,7 @@ class Program_weight_tensor_parameter_116: class Program_weight_tensor_parameter_117: name = "parameter_117" + original_name = "batch_norm_13.b_0" shape = [128] dtype = "float32" min_val = float("-0.459543") @@ -1298,6 +1416,7 @@ class Program_weight_tensor_parameter_117: class Program_weight_tensor_parameter_118: name = "parameter_118" + original_name = "batch_norm_13.w_0" shape = [128] dtype = "float32" min_val = float("0.193833") @@ -1309,6 +1428,7 @@ class Program_weight_tensor_parameter_118: class Program_weight_tensor_parameter_119: name = "parameter_119" + original_name = "batch_norm_13.w_2" shape = [128] dtype = "float32" min_val = float("0.0841521") @@ -1320,6 +1440,7 @@ class Program_weight_tensor_parameter_119: class Program_weight_tensor_parameter_120: name = "parameter_120" + original_name = "batch_norm_13.w_1" shape = [128] dtype = "float32" min_val = float("-1.45398") @@ -1331,6 +1452,7 @@ class Program_weight_tensor_parameter_120: class Program_weight_tensor_parameter_121: name = "parameter_121" + original_name = "conv2d_13.w_0" shape = [128, 128, 3, 3] dtype = "float32" min_val = float("-0.246662") @@ -1342,6 +1464,7 @@ class Program_weight_tensor_parameter_121: class Program_weight_tensor_parameter_122: name = "parameter_122" + original_name = "batch_norm_12.b_0" shape = [128] dtype = "float32" min_val = float("-0.39714") @@ -1353,6 +1476,7 @@ class Program_weight_tensor_parameter_122: class Program_weight_tensor_parameter_123: name = "parameter_123" + original_name = "batch_norm_12.w_0" shape = [128] dtype = "float32" min_val = float("0.0117418") @@ -1364,6 +1488,7 @@ class Program_weight_tensor_parameter_123: class Program_weight_tensor_parameter_124: name = "parameter_124" + original_name = "batch_norm_12.w_2" shape = [128] dtype = "float32" min_val = float("0.00219838") @@ -1375,6 +1500,7 @@ class Program_weight_tensor_parameter_124: class Program_weight_tensor_parameter_125: name = "parameter_125" + original_name = "batch_norm_12.w_1" shape = [128] dtype = "float32" min_val = float("-0.326518") @@ -1386,6 +1512,7 @@ class Program_weight_tensor_parameter_125: class Program_weight_tensor_parameter_126: name = "parameter_126" + original_name = "conv2d_12.w_0" shape = [128, 128, 3, 3] dtype = "float32" min_val = float("-0.312315") @@ -1397,6 +1524,7 @@ class Program_weight_tensor_parameter_126: class Program_weight_tensor_parameter_127: name = "parameter_127" + original_name = "batch_norm_11.b_0" shape = [128] dtype = "float32" min_val = float("-0.470488") @@ -1408,6 +1536,7 @@ class Program_weight_tensor_parameter_127: class Program_weight_tensor_parameter_128: name = "parameter_128" + original_name = "batch_norm_11.w_0" shape = [128] dtype = "float32" min_val = float("0.148687") @@ -1419,6 +1548,7 @@ class Program_weight_tensor_parameter_128: class Program_weight_tensor_parameter_129: name = "parameter_129" + original_name = "batch_norm_11.w_2" shape = [128] dtype = "float32" min_val = float("0.0981113") @@ -1430,6 +1560,7 @@ class Program_weight_tensor_parameter_129: class Program_weight_tensor_parameter_130: name = "parameter_130" + original_name = "batch_norm_11.w_1" shape = [128] dtype = "float32" min_val = float("-1.96872") @@ -1441,6 +1572,7 @@ class Program_weight_tensor_parameter_130: class Program_weight_tensor_parameter_131: name = "parameter_131" + original_name = "conv2d_11.w_0" shape = [128, 128, 3, 3] dtype = "float32" min_val = float("-0.282095") @@ -1452,6 +1584,7 @@ class Program_weight_tensor_parameter_131: class Program_weight_tensor_parameter_132: name = "parameter_132" + original_name = "batch_norm_10.b_0" shape = [128] dtype = "float32" min_val = float("-0.157064") @@ -1463,6 +1596,7 @@ class Program_weight_tensor_parameter_132: class Program_weight_tensor_parameter_133: name = "parameter_133" + original_name = "batch_norm_10.w_0" shape = [128] dtype = "float32" min_val = float("-0.0254056") @@ -1474,6 +1608,7 @@ class Program_weight_tensor_parameter_133: class Program_weight_tensor_parameter_134: name = "parameter_134" + original_name = "batch_norm_10.w_2" shape = [128] dtype = "float32" min_val = float("0.0032392") @@ -1485,6 +1620,7 @@ class Program_weight_tensor_parameter_134: class Program_weight_tensor_parameter_135: name = "parameter_135" + original_name = "batch_norm_10.w_1" shape = [128] dtype = "float32" min_val = float("-0.420115") @@ -1496,6 +1632,7 @@ class Program_weight_tensor_parameter_135: class Program_weight_tensor_parameter_136: name = "parameter_136" + original_name = "conv2d_10.w_0" shape = [128, 64, 1, 1] dtype = "float32" min_val = float("-0.575807") @@ -1507,6 +1644,7 @@ class Program_weight_tensor_parameter_136: class Program_weight_tensor_parameter_137: name = "parameter_137" + original_name = "batch_norm_9.b_0" shape = [128] dtype = "float32" min_val = float("-0.157064") @@ -1518,6 +1656,7 @@ class Program_weight_tensor_parameter_137: class Program_weight_tensor_parameter_138: name = "parameter_138" + original_name = "batch_norm_9.w_0" shape = [128] dtype = "float32" min_val = float("0.0610506") @@ -1529,6 +1668,7 @@ class Program_weight_tensor_parameter_138: class Program_weight_tensor_parameter_139: name = "parameter_139" + original_name = "batch_norm_9.w_2" shape = [128] dtype = "float32" min_val = float("0.015735") @@ -1540,6 +1680,7 @@ class Program_weight_tensor_parameter_139: class Program_weight_tensor_parameter_140: name = "parameter_140" + original_name = "batch_norm_9.w_1" shape = [128] dtype = "float32" min_val = float("-0.885638") @@ -1551,6 +1692,7 @@ class Program_weight_tensor_parameter_140: class Program_weight_tensor_parameter_141: name = "parameter_141" + original_name = "conv2d_9.w_0" shape = [128, 128, 3, 3] dtype = "float32" min_val = float("-0.300911") @@ -1562,6 +1704,7 @@ class Program_weight_tensor_parameter_141: class Program_weight_tensor_parameter_142: name = "parameter_142" + original_name = "batch_norm_8.b_0" shape = [128] dtype = "float32" min_val = float("-0.323633") @@ -1573,6 +1716,7 @@ class Program_weight_tensor_parameter_142: class Program_weight_tensor_parameter_143: name = "parameter_143" + original_name = "batch_norm_8.w_0" shape = [128] dtype = "float32" min_val = float("0.203252") @@ -1584,6 +1728,7 @@ class Program_weight_tensor_parameter_143: class Program_weight_tensor_parameter_144: name = "parameter_144" + original_name = "batch_norm_8.w_2" shape = [128] dtype = "float32" min_val = float("0.167567") @@ -1595,6 +1740,7 @@ class Program_weight_tensor_parameter_144: class Program_weight_tensor_parameter_145: name = "parameter_145" + original_name = "batch_norm_8.w_1" shape = [128] dtype = "float32" min_val = float("-1.23064") @@ -1606,6 +1752,7 @@ class Program_weight_tensor_parameter_145: class Program_weight_tensor_parameter_146: name = "parameter_146" + original_name = "conv2d_8.w_0" shape = [128, 64, 3, 3] dtype = "float32" min_val = float("-0.318123") @@ -1617,6 +1764,7 @@ class Program_weight_tensor_parameter_146: class Program_weight_tensor_parameter_147: name = "parameter_147" + original_name = "batch_norm_7.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -1626,6 +1774,7 @@ class Program_weight_tensor_parameter_147: class Program_weight_tensor_parameter_148: name = "parameter_148" + original_name = "batch_norm_7.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -1635,6 +1784,7 @@ class Program_weight_tensor_parameter_148: class Program_weight_tensor_parameter_149: name = "parameter_149" + original_name = "batch_norm_7.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -1644,6 +1794,7 @@ class Program_weight_tensor_parameter_149: class Program_weight_tensor_parameter_150: name = "parameter_150" + original_name = "batch_norm_7.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -1653,6 +1804,7 @@ class Program_weight_tensor_parameter_150: class Program_weight_tensor_parameter_151: name = "parameter_151" + original_name = "conv2d_7.w_0" shape = [64, 64, 3, 3] dtype = "float32" min_val = float("-0.241938") @@ -1664,6 +1816,7 @@ class Program_weight_tensor_parameter_151: class Program_weight_tensor_parameter_152: name = "parameter_152" + original_name = "batch_norm_6.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -1673,6 +1826,7 @@ class Program_weight_tensor_parameter_152: class Program_weight_tensor_parameter_153: name = "parameter_153" + original_name = "batch_norm_6.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -1682,6 +1836,7 @@ class Program_weight_tensor_parameter_153: class Program_weight_tensor_parameter_154: name = "parameter_154" + original_name = "batch_norm_6.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -1691,6 +1846,7 @@ class Program_weight_tensor_parameter_154: class Program_weight_tensor_parameter_155: name = "parameter_155" + original_name = "batch_norm_6.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -1700,6 +1856,7 @@ class Program_weight_tensor_parameter_155: class Program_weight_tensor_parameter_156: name = "parameter_156" + original_name = "conv2d_6.w_0" shape = [64, 64, 3, 3] dtype = "float32" min_val = float("-0.288299") @@ -1711,6 +1868,7 @@ class Program_weight_tensor_parameter_156: class Program_weight_tensor_parameter_157: name = "parameter_157" + original_name = "batch_norm_5.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -1720,6 +1878,7 @@ class Program_weight_tensor_parameter_157: class Program_weight_tensor_parameter_158: name = "parameter_158" + original_name = "batch_norm_5.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -1729,6 +1888,7 @@ class Program_weight_tensor_parameter_158: class Program_weight_tensor_parameter_159: name = "parameter_159" + original_name = "batch_norm_5.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -1738,6 +1898,7 @@ class Program_weight_tensor_parameter_159: class Program_weight_tensor_parameter_160: name = "parameter_160" + original_name = "batch_norm_5.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -1747,6 +1908,7 @@ class Program_weight_tensor_parameter_160: class Program_weight_tensor_parameter_161: name = "parameter_161" + original_name = "conv2d_5.w_0" shape = [64, 64, 3, 3] dtype = "float32" min_val = float("-0.328851") @@ -1758,6 +1920,7 @@ class Program_weight_tensor_parameter_161: class Program_weight_tensor_parameter_162: name = "parameter_162" + original_name = "batch_norm_4.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -1767,6 +1930,7 @@ class Program_weight_tensor_parameter_162: class Program_weight_tensor_parameter_163: name = "parameter_163" + original_name = "batch_norm_4.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -1776,6 +1940,7 @@ class Program_weight_tensor_parameter_163: class Program_weight_tensor_parameter_164: name = "parameter_164" + original_name = "batch_norm_4.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -1785,6 +1950,7 @@ class Program_weight_tensor_parameter_164: class Program_weight_tensor_parameter_165: name = "parameter_165" + original_name = "batch_norm_4.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -1794,6 +1960,7 @@ class Program_weight_tensor_parameter_165: class Program_weight_tensor_parameter_166: name = "parameter_166" + original_name = "conv2d_4.w_0" shape = [64, 64, 3, 3] dtype = "float32" min_val = float("-0.40449") @@ -1805,6 +1972,7 @@ class Program_weight_tensor_parameter_166: class Program_weight_tensor_parameter_167: name = "parameter_167" + original_name = "batch_norm_3.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -1814,6 +1982,7 @@ class Program_weight_tensor_parameter_167: class Program_weight_tensor_parameter_168: name = "parameter_168" + original_name = "batch_norm_3.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -1823,6 +1992,7 @@ class Program_weight_tensor_parameter_168: class Program_weight_tensor_parameter_169: name = "parameter_169" + original_name = "batch_norm_3.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -1832,6 +2002,7 @@ class Program_weight_tensor_parameter_169: class Program_weight_tensor_parameter_170: name = "parameter_170" + original_name = "batch_norm_3.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -1841,6 +2012,7 @@ class Program_weight_tensor_parameter_170: class Program_weight_tensor_parameter_171: name = "parameter_171" + original_name = "conv2d_3.w_0" shape = [64, 64, 1, 1] dtype = "float32" min_val = float("-0.846643") @@ -1852,6 +2024,7 @@ class Program_weight_tensor_parameter_171: class Program_weight_tensor_parameter_172: name = "parameter_172" + original_name = "batch_norm_2.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -1861,6 +2034,7 @@ class Program_weight_tensor_parameter_172: class Program_weight_tensor_parameter_173: name = "parameter_173" + original_name = "batch_norm_2.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -1870,6 +2044,7 @@ class Program_weight_tensor_parameter_173: class Program_weight_tensor_parameter_174: name = "parameter_174" + original_name = "batch_norm_2.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -1879,6 +2054,7 @@ class Program_weight_tensor_parameter_174: class Program_weight_tensor_parameter_175: name = "parameter_175" + original_name = "batch_norm_2.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -1888,6 +2064,7 @@ class Program_weight_tensor_parameter_175: class Program_weight_tensor_parameter_176: name = "parameter_176" + original_name = "conv2d_2.w_0" shape = [64, 64, 3, 3] dtype = "float32" min_val = float("-0.305449") @@ -1899,6 +2076,7 @@ class Program_weight_tensor_parameter_176: class Program_weight_tensor_parameter_177: name = "parameter_177" + original_name = "batch_norm_1.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -1908,6 +2086,7 @@ class Program_weight_tensor_parameter_177: class Program_weight_tensor_parameter_178: name = "parameter_178" + original_name = "batch_norm_1.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -1917,6 +2096,7 @@ class Program_weight_tensor_parameter_178: class Program_weight_tensor_parameter_179: name = "parameter_179" + original_name = "batch_norm_1.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -1926,6 +2106,7 @@ class Program_weight_tensor_parameter_179: class Program_weight_tensor_parameter_180: name = "parameter_180" + original_name = "batch_norm_1.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -1935,6 +2116,7 @@ class Program_weight_tensor_parameter_180: class Program_weight_tensor_parameter_181: name = "parameter_181" + original_name = "conv2d_1.w_0" shape = [64, 64, 3, 3] dtype = "float32" min_val = float("-0.523378") @@ -1946,6 +2128,7 @@ class Program_weight_tensor_parameter_181: class Program_weight_tensor_parameter_182: name = "parameter_182" + original_name = "batch_norm_0.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -1955,6 +2138,7 @@ class Program_weight_tensor_parameter_182: class Program_weight_tensor_parameter_183: name = "parameter_183" + original_name = "batch_norm_0.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -1964,6 +2148,7 @@ class Program_weight_tensor_parameter_183: class Program_weight_tensor_parameter_184: name = "parameter_184" + original_name = "batch_norm_0.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -1973,6 +2158,7 @@ class Program_weight_tensor_parameter_184: class Program_weight_tensor_parameter_185: name = "parameter_185" + original_name = "batch_norm_0.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -1982,6 +2168,7 @@ class Program_weight_tensor_parameter_185: class Program_weight_tensor_parameter_186: name = "parameter_186" + original_name = "conv2d_0.w_0" shape = [64, 3, 7, 7] dtype = "float32" min_val = float("-0.761036") diff --git a/paddle_samples/PaddleX/SOLOv2/subgraph_0/input_meta.py b/paddle_samples/PaddleX/SOLOv2/subgraph_0/input_meta.py index 8950b9023..0516f0eb3 100644 --- a/paddle_samples/PaddleX/SOLOv2/subgraph_0/input_meta.py +++ b/paddle_samples/PaddleX/SOLOv2/subgraph_0/input_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_data_0: name = "data_0" + original_name = "var_10" shape = [1, 3, 800, 928] dtype = "float32" min_val = float("-2.1179") diff --git a/paddle_samples/PaddleX/SOLOv2/subgraph_0/weight_meta.py b/paddle_samples/PaddleX/SOLOv2/subgraph_0/weight_meta.py index f4353b95a..809b5692d 100644 --- a/paddle_samples/PaddleX/SOLOv2/subgraph_0/weight_meta.py +++ b/paddle_samples/PaddleX/SOLOv2/subgraph_0/weight_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_parameter_0: name = "parameter_0" + original_name = "conv2d_70.b_0" shape = [2] dtype = "float32" min_val = float("0") @@ -9,6 +10,7 @@ class Program_weight_tensor_parameter_0: class Program_weight_tensor_parameter_1: name = "parameter_1" + original_name = "conv2d_70.w_0" shape = [2, 512, 3, 3] dtype = "float32" min_val = float("-0.0349261") @@ -20,6 +22,7 @@ class Program_weight_tensor_parameter_1: class Program_weight_tensor_parameter_2: name = "parameter_2" + original_name = "group_norm_7.b_0" shape = [512] dtype = "float32" min_val = float("-0.164361") @@ -31,6 +34,7 @@ class Program_weight_tensor_parameter_2: class Program_weight_tensor_parameter_3: name = "parameter_3" + original_name = "group_norm_7.w_0" shape = [512] dtype = "float32" min_val = float("0.819882") @@ -42,6 +46,7 @@ class Program_weight_tensor_parameter_3: class Program_weight_tensor_parameter_4: name = "parameter_4" + original_name = "conv2d_68.w_0" shape = [512, 512, 3, 3] dtype = "float32" min_val = float("-0.0714306") @@ -53,6 +58,7 @@ class Program_weight_tensor_parameter_4: class Program_weight_tensor_parameter_5: name = "parameter_5" + original_name = "group_norm_5.b_0" shape = [512] dtype = "float32" min_val = float("-0.140229") @@ -64,6 +70,7 @@ class Program_weight_tensor_parameter_5: class Program_weight_tensor_parameter_6: name = "parameter_6" + original_name = "group_norm_5.w_0" shape = [512] dtype = "float32" min_val = float("0.85017") @@ -75,6 +82,7 @@ class Program_weight_tensor_parameter_6: class Program_weight_tensor_parameter_7: name = "parameter_7" + original_name = "conv2d_66.w_0" shape = [512, 512, 3, 3] dtype = "float32" min_val = float("-0.0881653") @@ -86,6 +94,7 @@ class Program_weight_tensor_parameter_7: class Program_weight_tensor_parameter_8: name = "parameter_8" + original_name = "group_norm_3.b_0" shape = [512] dtype = "float32" min_val = float("-0.133168") @@ -97,6 +106,7 @@ class Program_weight_tensor_parameter_8: class Program_weight_tensor_parameter_9: name = "parameter_9" + original_name = "group_norm_3.w_0" shape = [512] dtype = "float32" min_val = float("0.896954") @@ -108,6 +118,7 @@ class Program_weight_tensor_parameter_9: class Program_weight_tensor_parameter_10: name = "parameter_10" + original_name = "conv2d_64.w_0" shape = [512, 512, 3, 3] dtype = "float32" min_val = float("-0.0769052") @@ -119,6 +130,7 @@ class Program_weight_tensor_parameter_10: class Program_weight_tensor_parameter_11: name = "parameter_11" + original_name = "group_norm_1.b_0" shape = [512] dtype = "float32" min_val = float("-0.0838783") @@ -130,6 +142,7 @@ class Program_weight_tensor_parameter_11: class Program_weight_tensor_parameter_12: name = "parameter_12" + original_name = "group_norm_1.w_0" shape = [512] dtype = "float32" min_val = float("0.930605") @@ -141,6 +154,7 @@ class Program_weight_tensor_parameter_12: class Program_weight_tensor_parameter_13: name = "parameter_13" + original_name = "conv2d_62.w_0" shape = [512, 256, 3, 3] dtype = "float32" min_val = float("-0.0528755") @@ -152,6 +166,7 @@ class Program_weight_tensor_parameter_13: class Program_weight_tensor_parameter_14: name = "parameter_14" + original_name = "conv2d_69.b_0" shape = [256] dtype = "float32" min_val = float("-0.108389") @@ -163,6 +178,7 @@ class Program_weight_tensor_parameter_14: class Program_weight_tensor_parameter_15: name = "parameter_15" + original_name = "conv2d_69.w_0" shape = [256, 512, 3, 3] dtype = "float32" min_val = float("-0.114531") @@ -174,6 +190,7 @@ class Program_weight_tensor_parameter_15: class Program_weight_tensor_parameter_16: name = "parameter_16" + original_name = "group_norm_6.b_0" shape = [512] dtype = "float32" min_val = float("-0.167083") @@ -185,6 +202,7 @@ class Program_weight_tensor_parameter_16: class Program_weight_tensor_parameter_17: name = "parameter_17" + original_name = "group_norm_6.w_0" shape = [512] dtype = "float32" min_val = float("0.527187") @@ -196,6 +214,7 @@ class Program_weight_tensor_parameter_17: class Program_weight_tensor_parameter_18: name = "parameter_18" + original_name = "conv2d_67.w_0" shape = [512, 512, 3, 3] dtype = "float32" min_val = float("-0.077059") @@ -207,6 +226,7 @@ class Program_weight_tensor_parameter_18: class Program_weight_tensor_parameter_19: name = "parameter_19" + original_name = "group_norm_4.b_0" shape = [512] dtype = "float32" min_val = float("-0.109315") @@ -218,6 +238,7 @@ class Program_weight_tensor_parameter_19: class Program_weight_tensor_parameter_20: name = "parameter_20" + original_name = "group_norm_4.w_0" shape = [512] dtype = "float32" min_val = float("0.56948") @@ -229,6 +250,7 @@ class Program_weight_tensor_parameter_20: class Program_weight_tensor_parameter_21: name = "parameter_21" + original_name = "conv2d_65.w_0" shape = [512, 512, 3, 3] dtype = "float32" min_val = float("-0.0796294") @@ -240,6 +262,7 @@ class Program_weight_tensor_parameter_21: class Program_weight_tensor_parameter_22: name = "parameter_22" + original_name = "group_norm_2.b_0" shape = [512] dtype = "float32" min_val = float("-0.140072") @@ -251,6 +274,7 @@ class Program_weight_tensor_parameter_22: class Program_weight_tensor_parameter_23: name = "parameter_23" + original_name = "group_norm_2.w_0" shape = [512] dtype = "float32" min_val = float("0.447613") @@ -262,6 +286,7 @@ class Program_weight_tensor_parameter_23: class Program_weight_tensor_parameter_24: name = "parameter_24" + original_name = "conv2d_63.w_0" shape = [512, 512, 3, 3] dtype = "float32" min_val = float("-0.0713139") @@ -273,6 +298,7 @@ class Program_weight_tensor_parameter_24: class Program_weight_tensor_parameter_25: name = "parameter_25" + original_name = "group_norm_0.b_0" shape = [512] dtype = "float32" min_val = float("-0.1152") @@ -284,6 +310,7 @@ class Program_weight_tensor_parameter_25: class Program_weight_tensor_parameter_26: name = "parameter_26" + original_name = "group_norm_0.w_0" shape = [512] dtype = "float32" min_val = float("0.890313") @@ -295,6 +322,7 @@ class Program_weight_tensor_parameter_26: class Program_weight_tensor_parameter_27: name = "parameter_27" + original_name = "conv2d_61.w_0" shape = [512, 258, 3, 3] dtype = "float32" min_val = float("-0.177044") @@ -306,6 +334,7 @@ class Program_weight_tensor_parameter_27: class Program_weight_tensor_parameter_28: name = "parameter_28" + original_name = "group_norm_15.b_0" shape = [256] dtype = "float32" min_val = float("-0.260234") @@ -317,6 +346,7 @@ class Program_weight_tensor_parameter_28: class Program_weight_tensor_parameter_29: name = "parameter_29" + original_name = "group_norm_15.w_0" shape = [256] dtype = "float32" min_val = float("0.363452") @@ -328,6 +358,7 @@ class Program_weight_tensor_parameter_29: class Program_weight_tensor_parameter_30: name = "parameter_30" + original_name = "conv2d_78.w_0" shape = [256, 128, 1, 1] dtype = "float32" min_val = float("-0.235533") @@ -339,6 +370,7 @@ class Program_weight_tensor_parameter_30: class Program_weight_tensor_parameter_31: name = "parameter_31" + original_name = "group_norm_14.b_0" shape = [128] dtype = "float32" min_val = float("-0.118527") @@ -350,6 +382,7 @@ class Program_weight_tensor_parameter_31: class Program_weight_tensor_parameter_32: name = "parameter_32" + original_name = "group_norm_14.w_0" shape = [128] dtype = "float32" min_val = float("0.836141") @@ -361,6 +394,7 @@ class Program_weight_tensor_parameter_32: class Program_weight_tensor_parameter_33: name = "parameter_33" + original_name = "conv2d_77.w_0" shape = [128, 128, 3, 3] dtype = "float32" min_val = float("-0.0902413") @@ -372,6 +406,7 @@ class Program_weight_tensor_parameter_33: class Program_weight_tensor_parameter_34: name = "parameter_34" + original_name = "group_norm_13.b_0" shape = [128] dtype = "float32" min_val = float("-0.118766") @@ -383,6 +418,7 @@ class Program_weight_tensor_parameter_34: class Program_weight_tensor_parameter_35: name = "parameter_35" + original_name = "group_norm_13.w_0" shape = [128] dtype = "float32" min_val = float("0.807154") @@ -394,6 +430,7 @@ class Program_weight_tensor_parameter_35: class Program_weight_tensor_parameter_36: name = "parameter_36" + original_name = "conv2d_76.w_0" shape = [128, 128, 3, 3] dtype = "float32" min_val = float("-0.092151") @@ -405,6 +442,7 @@ class Program_weight_tensor_parameter_36: class Program_weight_tensor_parameter_37: name = "parameter_37" + original_name = "group_norm_12.b_0" shape = [128] dtype = "float32" min_val = float("-0.183982") @@ -416,6 +454,7 @@ class Program_weight_tensor_parameter_37: class Program_weight_tensor_parameter_38: name = "parameter_38" + original_name = "group_norm_12.w_0" shape = [128] dtype = "float32" min_val = float("0.819542") @@ -427,6 +466,7 @@ class Program_weight_tensor_parameter_38: class Program_weight_tensor_parameter_39: name = "parameter_39" + original_name = "conv2d_75.w_0" shape = [128, 258, 3, 3] dtype = "float32" min_val = float("-0.335542") @@ -438,6 +478,7 @@ class Program_weight_tensor_parameter_39: class Program_weight_tensor_parameter_40: name = "parameter_40" + original_name = "group_norm_11.b_0" shape = [128] dtype = "float32" min_val = float("-0.130034") @@ -449,6 +490,7 @@ class Program_weight_tensor_parameter_40: class Program_weight_tensor_parameter_41: name = "parameter_41" + original_name = "group_norm_11.w_0" shape = [128] dtype = "float32" min_val = float("0.532729") @@ -460,6 +502,7 @@ class Program_weight_tensor_parameter_41: class Program_weight_tensor_parameter_42: name = "parameter_42" + original_name = "conv2d_74.w_0" shape = [128, 128, 3, 3] dtype = "float32" min_val = float("-0.0706972") @@ -471,6 +514,7 @@ class Program_weight_tensor_parameter_42: class Program_weight_tensor_parameter_43: name = "parameter_43" + original_name = "group_norm_10.b_0" shape = [128] dtype = "float32" min_val = float("-0.0981304") @@ -482,6 +526,7 @@ class Program_weight_tensor_parameter_43: class Program_weight_tensor_parameter_44: name = "parameter_44" + original_name = "group_norm_10.w_0" shape = [128] dtype = "float32" min_val = float("0.796152") @@ -493,6 +538,7 @@ class Program_weight_tensor_parameter_44: class Program_weight_tensor_parameter_45: name = "parameter_45" + original_name = "conv2d_73.w_0" shape = [128, 256, 3, 3] dtype = "float32" min_val = float("-0.0437951") @@ -504,6 +550,7 @@ class Program_weight_tensor_parameter_45: class Program_weight_tensor_parameter_46: name = "parameter_46" + original_name = "group_norm_9.b_0" shape = [128] dtype = "float32" min_val = float("-0.128166") @@ -515,6 +562,7 @@ class Program_weight_tensor_parameter_46: class Program_weight_tensor_parameter_47: name = "parameter_47" + original_name = "group_norm_9.w_0" shape = [128] dtype = "float32" min_val = float("0.534046") @@ -526,6 +574,7 @@ class Program_weight_tensor_parameter_47: class Program_weight_tensor_parameter_48: name = "parameter_48" + original_name = "conv2d_72.w_0" shape = [128, 256, 3, 3] dtype = "float32" min_val = float("-0.0504267") @@ -537,6 +586,7 @@ class Program_weight_tensor_parameter_48: class Program_weight_tensor_parameter_49: name = "parameter_49" + original_name = "group_norm_8.b_0" shape = [128] dtype = "float32" min_val = float("-0.135621") @@ -548,6 +598,7 @@ class Program_weight_tensor_parameter_49: class Program_weight_tensor_parameter_50: name = "parameter_50" + original_name = "group_norm_8.w_0" shape = [128] dtype = "float32" min_val = float("0.885653") @@ -559,6 +610,7 @@ class Program_weight_tensor_parameter_50: class Program_weight_tensor_parameter_51: name = "parameter_51" + original_name = "conv2d_71.w_0" shape = [128, 256, 3, 3] dtype = "float32" min_val = float("-0.051786") @@ -570,6 +622,7 @@ class Program_weight_tensor_parameter_51: class Program_weight_tensor_parameter_52: name = "parameter_52" + original_name = "conv2d_60.b_0" shape = [256] dtype = "float32" min_val = float("-0.140687") @@ -581,6 +634,7 @@ class Program_weight_tensor_parameter_52: class Program_weight_tensor_parameter_53: name = "parameter_53" + original_name = "conv2d_60.w_0" shape = [256, 256, 3, 3] dtype = "float32" min_val = float("-0.0449519") @@ -592,6 +646,7 @@ class Program_weight_tensor_parameter_53: class Program_weight_tensor_parameter_54: name = "parameter_54" + original_name = "conv2d_58.b_0" shape = [256] dtype = "float32" min_val = float("-0.0684487") @@ -603,6 +658,7 @@ class Program_weight_tensor_parameter_54: class Program_weight_tensor_parameter_55: name = "parameter_55" + original_name = "conv2d_58.w_0" shape = [256, 256, 3, 3] dtype = "float32" min_val = float("-0.0372709") @@ -614,6 +670,7 @@ class Program_weight_tensor_parameter_55: class Program_weight_tensor_parameter_56: name = "parameter_56" + original_name = "conv2d_56.b_0" shape = [256] dtype = "float32" min_val = float("-0.0925719") @@ -625,6 +682,7 @@ class Program_weight_tensor_parameter_56: class Program_weight_tensor_parameter_57: name = "parameter_57" + original_name = "conv2d_56.w_0" shape = [256, 256, 3, 3] dtype = "float32" min_val = float("-0.0363277") @@ -636,6 +694,7 @@ class Program_weight_tensor_parameter_57: class Program_weight_tensor_parameter_58: name = "parameter_58" + original_name = "conv2d_54.b_0" shape = [256] dtype = "float32" min_val = float("-0.0806452") @@ -647,6 +706,7 @@ class Program_weight_tensor_parameter_58: class Program_weight_tensor_parameter_59: name = "parameter_59" + original_name = "conv2d_54.w_0" shape = [256, 256, 3, 3] dtype = "float32" min_val = float("-0.0385173") @@ -658,6 +718,7 @@ class Program_weight_tensor_parameter_59: class Program_weight_tensor_parameter_60: name = "parameter_60" + original_name = "conv2d_59.b_0" shape = [256] dtype = "float32" min_val = float("-0.119148") @@ -669,6 +730,7 @@ class Program_weight_tensor_parameter_60: class Program_weight_tensor_parameter_61: name = "parameter_61" + original_name = "conv2d_59.w_0" shape = [256, 2048, 1, 1] dtype = "float32" min_val = float("-0.0718746") @@ -680,6 +742,7 @@ class Program_weight_tensor_parameter_61: class Program_weight_tensor_parameter_62: name = "parameter_62" + original_name = "conv2d_57.b_0" shape = [256] dtype = "float32" min_val = float("-0.0679942") @@ -691,6 +754,7 @@ class Program_weight_tensor_parameter_62: class Program_weight_tensor_parameter_63: name = "parameter_63" + original_name = "conv2d_57.w_0" shape = [256, 1024, 1, 1] dtype = "float32" min_val = float("-0.110228") @@ -702,6 +766,7 @@ class Program_weight_tensor_parameter_63: class Program_weight_tensor_parameter_64: name = "parameter_64" + original_name = "conv2d_55.b_0" shape = [256] dtype = "float32" min_val = float("-0.0655844") @@ -713,6 +778,7 @@ class Program_weight_tensor_parameter_64: class Program_weight_tensor_parameter_65: name = "parameter_65" + original_name = "conv2d_55.w_0" shape = [256, 512, 1, 1] dtype = "float32" min_val = float("-0.0781448") @@ -724,6 +790,7 @@ class Program_weight_tensor_parameter_65: class Program_weight_tensor_parameter_66: name = "parameter_66" + original_name = "conv2d_53.b_0" shape = [256] dtype = "float32" min_val = float("-0.072923") @@ -735,6 +802,7 @@ class Program_weight_tensor_parameter_66: class Program_weight_tensor_parameter_67: name = "parameter_67" + original_name = "conv2d_53.w_0" shape = [256, 256, 1, 1] dtype = "float32" min_val = float("-0.0820079") @@ -746,6 +814,7 @@ class Program_weight_tensor_parameter_67: class Program_weight_tensor_parameter_68: name = "parameter_68" + original_name = "batch_norm2d_52.b_0" shape = [2048] dtype = "float32" min_val = float("-0.0151986") @@ -757,6 +826,7 @@ class Program_weight_tensor_parameter_68: class Program_weight_tensor_parameter_69: name = "parameter_69" + original_name = "batch_norm2d_52.w_0" shape = [2048] dtype = "float32" min_val = float("0.0697239") @@ -768,6 +838,7 @@ class Program_weight_tensor_parameter_69: class Program_weight_tensor_parameter_70: name = "parameter_70" + original_name = "batch_norm2d_52.w_2" shape = [2048] dtype = "float32" min_val = float("1.81634e-05") @@ -779,6 +850,7 @@ class Program_weight_tensor_parameter_70: class Program_weight_tensor_parameter_71: name = "parameter_71" + original_name = "batch_norm2d_52.w_1" shape = [2048] dtype = "float32" min_val = float("-0.0611881") @@ -790,6 +862,7 @@ class Program_weight_tensor_parameter_71: class Program_weight_tensor_parameter_72: name = "parameter_72" + original_name = "conv2d_52.w_0" shape = [2048, 512, 1, 1] dtype = "float32" min_val = float("-0.175753") @@ -801,6 +874,7 @@ class Program_weight_tensor_parameter_72: class Program_weight_tensor_parameter_73: name = "parameter_73" + original_name = "batch_norm2d_51.b_0" shape = [512] dtype = "float32" min_val = float("-0.256637") @@ -812,6 +886,7 @@ class Program_weight_tensor_parameter_73: class Program_weight_tensor_parameter_74: name = "parameter_74" + original_name = "batch_norm2d_51.w_0" shape = [512] dtype = "float32" min_val = float("0.142703") @@ -823,6 +898,7 @@ class Program_weight_tensor_parameter_74: class Program_weight_tensor_parameter_75: name = "parameter_75" + original_name = "batch_norm2d_51.w_2" shape = [512] dtype = "float32" min_val = float("0.00782461") @@ -834,6 +910,7 @@ class Program_weight_tensor_parameter_75: class Program_weight_tensor_parameter_76: name = "parameter_76" + original_name = "batch_norm2d_51.w_1" shape = [512] dtype = "float32" min_val = float("-0.213289") @@ -845,6 +922,7 @@ class Program_weight_tensor_parameter_76: class Program_weight_tensor_parameter_77: name = "parameter_77" + original_name = "conv2d_51.w_0" shape = [512, 512, 3, 3] dtype = "float32" min_val = float("-0.154763") @@ -856,6 +934,7 @@ class Program_weight_tensor_parameter_77: class Program_weight_tensor_parameter_78: name = "parameter_78" + original_name = "batch_norm2d_50.b_0" shape = [512] dtype = "float32" min_val = float("-0.37403") @@ -867,6 +946,7 @@ class Program_weight_tensor_parameter_78: class Program_weight_tensor_parameter_79: name = "parameter_79" + original_name = "batch_norm2d_50.w_0" shape = [512] dtype = "float32" min_val = float("0.112228") @@ -878,6 +958,7 @@ class Program_weight_tensor_parameter_79: class Program_weight_tensor_parameter_80: name = "parameter_80" + original_name = "batch_norm2d_50.w_2" shape = [512] dtype = "float32" min_val = float("0.00368759") @@ -889,6 +970,7 @@ class Program_weight_tensor_parameter_80: class Program_weight_tensor_parameter_81: name = "parameter_81" + original_name = "batch_norm2d_50.w_1" shape = [512] dtype = "float32" min_val = float("-0.165443") @@ -900,6 +982,7 @@ class Program_weight_tensor_parameter_81: class Program_weight_tensor_parameter_82: name = "parameter_82" + original_name = "conv2d_50.w_0" shape = [512, 2048, 1, 1] dtype = "float32" min_val = float("-0.311529") @@ -911,6 +994,7 @@ class Program_weight_tensor_parameter_82: class Program_weight_tensor_parameter_83: name = "parameter_83" + original_name = "batch_norm2d_49.b_0" shape = [2048] dtype = "float32" min_val = float("-0.344631") @@ -922,6 +1006,7 @@ class Program_weight_tensor_parameter_83: class Program_weight_tensor_parameter_84: name = "parameter_84" + original_name = "batch_norm2d_49.w_0" shape = [2048] dtype = "float32" min_val = float("-0.0556075") @@ -933,6 +1018,7 @@ class Program_weight_tensor_parameter_84: class Program_weight_tensor_parameter_85: name = "parameter_85" + original_name = "batch_norm2d_49.w_2" shape = [2048] dtype = "float32" min_val = float("0.000374117") @@ -944,6 +1030,7 @@ class Program_weight_tensor_parameter_85: class Program_weight_tensor_parameter_86: name = "parameter_86" + original_name = "batch_norm2d_49.w_1" shape = [2048] dtype = "float32" min_val = float("-0.143769") @@ -955,6 +1042,7 @@ class Program_weight_tensor_parameter_86: class Program_weight_tensor_parameter_87: name = "parameter_87" + original_name = "conv2d_49.w_0" shape = [2048, 512, 1, 1] dtype = "float32" min_val = float("-0.262652") @@ -966,6 +1054,7 @@ class Program_weight_tensor_parameter_87: class Program_weight_tensor_parameter_88: name = "parameter_88" + original_name = "batch_norm2d_48.b_0" shape = [512] dtype = "float32" min_val = float("-0.281487") @@ -977,6 +1066,7 @@ class Program_weight_tensor_parameter_88: class Program_weight_tensor_parameter_89: name = "parameter_89" + original_name = "batch_norm2d_48.w_0" shape = [512] dtype = "float32" min_val = float("0.103009") @@ -988,6 +1078,7 @@ class Program_weight_tensor_parameter_89: class Program_weight_tensor_parameter_90: name = "parameter_90" + original_name = "batch_norm2d_48.w_2" shape = [512] dtype = "float32" min_val = float("0.00667127") @@ -999,6 +1090,7 @@ class Program_weight_tensor_parameter_90: class Program_weight_tensor_parameter_91: name = "parameter_91" + original_name = "batch_norm2d_48.w_1" shape = [512] dtype = "float32" min_val = float("-0.333268") @@ -1010,6 +1102,7 @@ class Program_weight_tensor_parameter_91: class Program_weight_tensor_parameter_92: name = "parameter_92" + original_name = "conv2d_48.w_0" shape = [512, 512, 3, 3] dtype = "float32" min_val = float("-0.16425") @@ -1021,6 +1114,7 @@ class Program_weight_tensor_parameter_92: class Program_weight_tensor_parameter_93: name = "parameter_93" + original_name = "batch_norm2d_47.b_0" shape = [512] dtype = "float32" min_val = float("-0.204657") @@ -1032,6 +1126,7 @@ class Program_weight_tensor_parameter_93: class Program_weight_tensor_parameter_94: name = "parameter_94" + original_name = "batch_norm2d_47.w_0" shape = [512] dtype = "float32" min_val = float("0.110153") @@ -1043,6 +1138,7 @@ class Program_weight_tensor_parameter_94: class Program_weight_tensor_parameter_95: name = "parameter_95" + original_name = "batch_norm2d_47.w_2" shape = [512] dtype = "float32" min_val = float("0.00384403") @@ -1054,6 +1150,7 @@ class Program_weight_tensor_parameter_95: class Program_weight_tensor_parameter_96: name = "parameter_96" + original_name = "batch_norm2d_47.w_1" shape = [512] dtype = "float32" min_val = float("-0.156824") @@ -1065,6 +1162,7 @@ class Program_weight_tensor_parameter_96: class Program_weight_tensor_parameter_97: name = "parameter_97" + original_name = "conv2d_47.w_0" shape = [512, 2048, 1, 1] dtype = "float32" min_val = float("-0.199091") @@ -1076,6 +1174,7 @@ class Program_weight_tensor_parameter_97: class Program_weight_tensor_parameter_98: name = "parameter_98" + original_name = "batch_norm2d_46.b_0" shape = [2048] dtype = "float32" min_val = float("-0.175843") @@ -1087,6 +1186,7 @@ class Program_weight_tensor_parameter_98: class Program_weight_tensor_parameter_99: name = "parameter_99" + original_name = "batch_norm2d_46.w_0" shape = [2048] dtype = "float32" min_val = float("0.0445195") @@ -1098,6 +1198,7 @@ class Program_weight_tensor_parameter_99: class Program_weight_tensor_parameter_100: name = "parameter_100" + original_name = "batch_norm2d_46.w_2" shape = [2048] dtype = "float32" min_val = float("0.00110241") @@ -1109,6 +1210,7 @@ class Program_weight_tensor_parameter_100: class Program_weight_tensor_parameter_101: name = "parameter_101" + original_name = "batch_norm2d_46.w_1" shape = [2048] dtype = "float32" min_val = float("-0.149371") @@ -1120,6 +1222,7 @@ class Program_weight_tensor_parameter_101: class Program_weight_tensor_parameter_102: name = "parameter_102" + original_name = "conv2d_46.w_0" shape = [2048, 1024, 1, 1] dtype = "float32" min_val = float("-0.230573") @@ -1131,6 +1234,7 @@ class Program_weight_tensor_parameter_102: class Program_weight_tensor_parameter_103: name = "parameter_103" + original_name = "batch_norm2d_45.b_0" shape = [2048] dtype = "float32" min_val = float("-0.175843") @@ -1142,6 +1246,7 @@ class Program_weight_tensor_parameter_103: class Program_weight_tensor_parameter_104: name = "parameter_104" + original_name = "batch_norm2d_45.w_0" shape = [2048] dtype = "float32" min_val = float("-0.0299624") @@ -1153,6 +1258,7 @@ class Program_weight_tensor_parameter_104: class Program_weight_tensor_parameter_105: name = "parameter_105" + original_name = "batch_norm2d_45.w_2" shape = [2048] dtype = "float32" min_val = float("0.000316897") @@ -1164,6 +1270,7 @@ class Program_weight_tensor_parameter_105: class Program_weight_tensor_parameter_106: name = "parameter_106" + original_name = "batch_norm2d_45.w_1" shape = [2048] dtype = "float32" min_val = float("-0.119339") @@ -1175,6 +1282,7 @@ class Program_weight_tensor_parameter_106: class Program_weight_tensor_parameter_107: name = "parameter_107" + original_name = "conv2d_45.w_0" shape = [2048, 512, 1, 1] dtype = "float32" min_val = float("-0.282002") @@ -1186,6 +1294,7 @@ class Program_weight_tensor_parameter_107: class Program_weight_tensor_parameter_108: name = "parameter_108" + original_name = "batch_norm2d_44.b_0" shape = [512] dtype = "float32" min_val = float("-0.166324") @@ -1197,6 +1306,7 @@ class Program_weight_tensor_parameter_108: class Program_weight_tensor_parameter_109: name = "parameter_109" + original_name = "batch_norm2d_44.w_0" shape = [512] dtype = "float32" min_val = float("0.122111") @@ -1208,6 +1318,7 @@ class Program_weight_tensor_parameter_109: class Program_weight_tensor_parameter_110: name = "parameter_110" + original_name = "batch_norm2d_44.w_2" shape = [512] dtype = "float32" min_val = float("0.00556363") @@ -1219,6 +1330,7 @@ class Program_weight_tensor_parameter_110: class Program_weight_tensor_parameter_111: name = "parameter_111" + original_name = "batch_norm2d_44.w_1" shape = [512] dtype = "float32" min_val = float("-0.114129") @@ -1230,6 +1342,7 @@ class Program_weight_tensor_parameter_111: class Program_weight_tensor_parameter_112: name = "parameter_112" + original_name = "conv2d_44.w_0" shape = [512, 512, 3, 3] dtype = "float32" min_val = float("-0.323367") @@ -1241,6 +1354,7 @@ class Program_weight_tensor_parameter_112: class Program_weight_tensor_parameter_113: name = "parameter_113" + original_name = "batch_norm2d_43.b_0" shape = [512] dtype = "float32" min_val = float("-0.323683") @@ -1252,6 +1366,7 @@ class Program_weight_tensor_parameter_113: class Program_weight_tensor_parameter_114: name = "parameter_114" + original_name = "batch_norm2d_43.w_0" shape = [512] dtype = "float32" min_val = float("0.103589") @@ -1263,6 +1378,7 @@ class Program_weight_tensor_parameter_114: class Program_weight_tensor_parameter_115: name = "parameter_115" + original_name = "batch_norm2d_43.w_2" shape = [512] dtype = "float32" min_val = float("0.00626156") @@ -1274,6 +1390,7 @@ class Program_weight_tensor_parameter_115: class Program_weight_tensor_parameter_116: name = "parameter_116" + original_name = "batch_norm2d_43.w_1" shape = [512] dtype = "float32" min_val = float("-0.154333") @@ -1285,6 +1402,7 @@ class Program_weight_tensor_parameter_116: class Program_weight_tensor_parameter_117: name = "parameter_117" + original_name = "conv2d_43.w_0" shape = [512, 1024, 1, 1] dtype = "float32" min_val = float("-0.276323") @@ -1296,6 +1414,7 @@ class Program_weight_tensor_parameter_117: class Program_weight_tensor_parameter_118: name = "parameter_118" + original_name = "batch_norm2d_42.b_0" shape = [1024] dtype = "float32" min_val = float("-0.350923") @@ -1307,6 +1426,7 @@ class Program_weight_tensor_parameter_118: class Program_weight_tensor_parameter_119: name = "parameter_119" + original_name = "batch_norm2d_42.w_0" shape = [1024] dtype = "float32" min_val = float("-0.0704466") @@ -1318,6 +1438,7 @@ class Program_weight_tensor_parameter_119: class Program_weight_tensor_parameter_120: name = "parameter_120" + original_name = "batch_norm2d_42.w_2" shape = [1024] dtype = "float32" min_val = float("7.2606e-05") @@ -1329,6 +1450,7 @@ class Program_weight_tensor_parameter_120: class Program_weight_tensor_parameter_121: name = "parameter_121" + original_name = "batch_norm2d_42.w_1" shape = [1024] dtype = "float32" min_val = float("-0.14425") @@ -1340,6 +1462,7 @@ class Program_weight_tensor_parameter_121: class Program_weight_tensor_parameter_122: name = "parameter_122" + original_name = "conv2d_42.w_0" shape = [1024, 256, 1, 1] dtype = "float32" min_val = float("-0.53185") @@ -1351,6 +1474,7 @@ class Program_weight_tensor_parameter_122: class Program_weight_tensor_parameter_123: name = "parameter_123" + original_name = "batch_norm2d_41.b_0" shape = [256] dtype = "float32" min_val = float("-0.21119") @@ -1362,6 +1486,7 @@ class Program_weight_tensor_parameter_123: class Program_weight_tensor_parameter_124: name = "parameter_124" + original_name = "batch_norm2d_41.w_0" shape = [256] dtype = "float32" min_val = float("0.11773") @@ -1373,6 +1498,7 @@ class Program_weight_tensor_parameter_124: class Program_weight_tensor_parameter_125: name = "parameter_125" + original_name = "batch_norm2d_41.w_2" shape = [256] dtype = "float32" min_val = float("0.0034165") @@ -1384,6 +1510,7 @@ class Program_weight_tensor_parameter_125: class Program_weight_tensor_parameter_126: name = "parameter_126" + original_name = "batch_norm2d_41.w_1" shape = [256] dtype = "float32" min_val = float("-0.17006") @@ -1395,6 +1522,7 @@ class Program_weight_tensor_parameter_126: class Program_weight_tensor_parameter_127: name = "parameter_127" + original_name = "conv2d_41.w_0" shape = [256, 256, 3, 3] dtype = "float32" min_val = float("-0.237766") @@ -1406,6 +1534,7 @@ class Program_weight_tensor_parameter_127: class Program_weight_tensor_parameter_128: name = "parameter_128" + original_name = "batch_norm2d_40.b_0" shape = [256] dtype = "float32" min_val = float("-0.311719") @@ -1417,6 +1546,7 @@ class Program_weight_tensor_parameter_128: class Program_weight_tensor_parameter_129: name = "parameter_129" + original_name = "batch_norm2d_40.w_0" shape = [256] dtype = "float32" min_val = float("0.0999653") @@ -1428,6 +1558,7 @@ class Program_weight_tensor_parameter_129: class Program_weight_tensor_parameter_130: name = "parameter_130" + original_name = "batch_norm2d_40.w_2" shape = [256] dtype = "float32" min_val = float("0.00541324") @@ -1439,6 +1570,7 @@ class Program_weight_tensor_parameter_130: class Program_weight_tensor_parameter_131: name = "parameter_131" + original_name = "batch_norm2d_40.w_1" shape = [256] dtype = "float32" min_val = float("-0.217744") @@ -1450,6 +1582,7 @@ class Program_weight_tensor_parameter_131: class Program_weight_tensor_parameter_132: name = "parameter_132" + original_name = "conv2d_40.w_0" shape = [256, 1024, 1, 1] dtype = "float32" min_val = float("-0.232746") @@ -1461,6 +1594,7 @@ class Program_weight_tensor_parameter_132: class Program_weight_tensor_parameter_133: name = "parameter_133" + original_name = "batch_norm2d_39.b_0" shape = [1024] dtype = "float32" min_val = float("-0.257451") @@ -1472,6 +1606,7 @@ class Program_weight_tensor_parameter_133: class Program_weight_tensor_parameter_134: name = "parameter_134" + original_name = "batch_norm2d_39.w_0" shape = [1024] dtype = "float32" min_val = float("-0.0934567") @@ -1483,6 +1618,7 @@ class Program_weight_tensor_parameter_134: class Program_weight_tensor_parameter_135: name = "parameter_135" + original_name = "batch_norm2d_39.w_2" shape = [1024] dtype = "float32" min_val = float("3.37401e-05") @@ -1494,6 +1630,7 @@ class Program_weight_tensor_parameter_135: class Program_weight_tensor_parameter_136: name = "parameter_136" + original_name = "batch_norm2d_39.w_1" shape = [1024] dtype = "float32" min_val = float("-0.0803722") @@ -1505,6 +1642,7 @@ class Program_weight_tensor_parameter_136: class Program_weight_tensor_parameter_137: name = "parameter_137" + original_name = "conv2d_39.w_0" shape = [1024, 256, 1, 1] dtype = "float32" min_val = float("-0.227086") @@ -1516,6 +1654,7 @@ class Program_weight_tensor_parameter_137: class Program_weight_tensor_parameter_138: name = "parameter_138" + original_name = "batch_norm2d_38.b_0" shape = [256] dtype = "float32" min_val = float("-0.230738") @@ -1527,6 +1666,7 @@ class Program_weight_tensor_parameter_138: class Program_weight_tensor_parameter_139: name = "parameter_139" + original_name = "batch_norm2d_38.w_0" shape = [256] dtype = "float32" min_val = float("0.0986424") @@ -1538,6 +1678,7 @@ class Program_weight_tensor_parameter_139: class Program_weight_tensor_parameter_140: name = "parameter_140" + original_name = "batch_norm2d_38.w_2" shape = [256] dtype = "float32" min_val = float("0.00330071") @@ -1549,6 +1690,7 @@ class Program_weight_tensor_parameter_140: class Program_weight_tensor_parameter_141: name = "parameter_141" + original_name = "batch_norm2d_38.w_1" shape = [256] dtype = "float32" min_val = float("-0.162212") @@ -1560,6 +1702,7 @@ class Program_weight_tensor_parameter_141: class Program_weight_tensor_parameter_142: name = "parameter_142" + original_name = "conv2d_38.w_0" shape = [256, 256, 3, 3] dtype = "float32" min_val = float("-0.151922") @@ -1571,6 +1714,7 @@ class Program_weight_tensor_parameter_142: class Program_weight_tensor_parameter_143: name = "parameter_143" + original_name = "batch_norm2d_37.b_0" shape = [256] dtype = "float32" min_val = float("-0.30878") @@ -1582,6 +1726,7 @@ class Program_weight_tensor_parameter_143: class Program_weight_tensor_parameter_144: name = "parameter_144" + original_name = "batch_norm2d_37.w_0" shape = [256] dtype = "float32" min_val = float("0.091752") @@ -1593,6 +1738,7 @@ class Program_weight_tensor_parameter_144: class Program_weight_tensor_parameter_145: name = "parameter_145" + original_name = "batch_norm2d_37.w_2" shape = [256] dtype = "float32" min_val = float("0.00507588") @@ -1604,6 +1750,7 @@ class Program_weight_tensor_parameter_145: class Program_weight_tensor_parameter_146: name = "parameter_146" + original_name = "batch_norm2d_37.w_1" shape = [256] dtype = "float32" min_val = float("-0.253355") @@ -1615,6 +1762,7 @@ class Program_weight_tensor_parameter_146: class Program_weight_tensor_parameter_147: name = "parameter_147" + original_name = "conv2d_37.w_0" shape = [256, 1024, 1, 1] dtype = "float32" min_val = float("-0.181943") @@ -1626,6 +1774,7 @@ class Program_weight_tensor_parameter_147: class Program_weight_tensor_parameter_148: name = "parameter_148" + original_name = "batch_norm2d_36.b_0" shape = [1024] dtype = "float32" min_val = float("-0.277434") @@ -1637,6 +1786,7 @@ class Program_weight_tensor_parameter_148: class Program_weight_tensor_parameter_149: name = "parameter_149" + original_name = "batch_norm2d_36.w_0" shape = [1024] dtype = "float32" min_val = float("-0.0669331") @@ -1648,6 +1798,7 @@ class Program_weight_tensor_parameter_149: class Program_weight_tensor_parameter_150: name = "parameter_150" + original_name = "batch_norm2d_36.w_2" shape = [1024] dtype = "float32" min_val = float("3.75198e-05") @@ -1659,6 +1810,7 @@ class Program_weight_tensor_parameter_150: class Program_weight_tensor_parameter_151: name = "parameter_151" + original_name = "batch_norm2d_36.w_1" shape = [1024] dtype = "float32" min_val = float("-0.0982832") @@ -1670,6 +1822,7 @@ class Program_weight_tensor_parameter_151: class Program_weight_tensor_parameter_152: name = "parameter_152" + original_name = "conv2d_36.w_0" shape = [1024, 256, 1, 1] dtype = "float32" min_val = float("-0.216755") @@ -1681,6 +1834,7 @@ class Program_weight_tensor_parameter_152: class Program_weight_tensor_parameter_153: name = "parameter_153" + original_name = "batch_norm2d_35.b_0" shape = [256] dtype = "float32" min_val = float("-0.296816") @@ -1692,6 +1846,7 @@ class Program_weight_tensor_parameter_153: class Program_weight_tensor_parameter_154: name = "parameter_154" + original_name = "batch_norm2d_35.w_0" shape = [256] dtype = "float32" min_val = float("0.107435") @@ -1703,6 +1858,7 @@ class Program_weight_tensor_parameter_154: class Program_weight_tensor_parameter_155: name = "parameter_155" + original_name = "batch_norm2d_35.w_2" shape = [256] dtype = "float32" min_val = float("0.00385646") @@ -1714,6 +1870,7 @@ class Program_weight_tensor_parameter_155: class Program_weight_tensor_parameter_156: name = "parameter_156" + original_name = "batch_norm2d_35.w_1" shape = [256] dtype = "float32" min_val = float("-0.169503") @@ -1725,6 +1882,7 @@ class Program_weight_tensor_parameter_156: class Program_weight_tensor_parameter_157: name = "parameter_157" + original_name = "conv2d_35.w_0" shape = [256, 256, 3, 3] dtype = "float32" min_val = float("-0.156506") @@ -1736,6 +1894,7 @@ class Program_weight_tensor_parameter_157: class Program_weight_tensor_parameter_158: name = "parameter_158" + original_name = "batch_norm2d_34.b_0" shape = [256] dtype = "float32" min_val = float("-0.213348") @@ -1747,6 +1906,7 @@ class Program_weight_tensor_parameter_158: class Program_weight_tensor_parameter_159: name = "parameter_159" + original_name = "batch_norm2d_34.w_0" shape = [256] dtype = "float32" min_val = float("0.0886164") @@ -1758,6 +1918,7 @@ class Program_weight_tensor_parameter_159: class Program_weight_tensor_parameter_160: name = "parameter_160" + original_name = "batch_norm2d_34.w_2" shape = [256] dtype = "float32" min_val = float("0.00512839") @@ -1769,6 +1930,7 @@ class Program_weight_tensor_parameter_160: class Program_weight_tensor_parameter_161: name = "parameter_161" + original_name = "batch_norm2d_34.w_1" shape = [256] dtype = "float32" min_val = float("-0.186114") @@ -1780,6 +1942,7 @@ class Program_weight_tensor_parameter_161: class Program_weight_tensor_parameter_162: name = "parameter_162" + original_name = "conv2d_34.w_0" shape = [256, 1024, 1, 1] dtype = "float32" min_val = float("-0.203824") @@ -1791,6 +1954,7 @@ class Program_weight_tensor_parameter_162: class Program_weight_tensor_parameter_163: name = "parameter_163" + original_name = "batch_norm2d_33.b_0" shape = [1024] dtype = "float32" min_val = float("-0.250531") @@ -1802,6 +1966,7 @@ class Program_weight_tensor_parameter_163: class Program_weight_tensor_parameter_164: name = "parameter_164" + original_name = "batch_norm2d_33.w_0" shape = [1024] dtype = "float32" min_val = float("-0.0906585") @@ -1813,6 +1978,7 @@ class Program_weight_tensor_parameter_164: class Program_weight_tensor_parameter_165: name = "parameter_165" + original_name = "batch_norm2d_33.w_2" shape = [1024] dtype = "float32" min_val = float("5.60519e-45") @@ -1824,6 +1990,7 @@ class Program_weight_tensor_parameter_165: class Program_weight_tensor_parameter_166: name = "parameter_166" + original_name = "batch_norm2d_33.w_1" shape = [1024] dtype = "float32" min_val = float("-0.167247") @@ -1835,6 +2002,7 @@ class Program_weight_tensor_parameter_166: class Program_weight_tensor_parameter_167: name = "parameter_167" + original_name = "conv2d_33.w_0" shape = [1024, 256, 1, 1] dtype = "float32" min_val = float("-0.322723") @@ -1846,6 +2014,7 @@ class Program_weight_tensor_parameter_167: class Program_weight_tensor_parameter_168: name = "parameter_168" + original_name = "batch_norm2d_32.b_0" shape = [256] dtype = "float32" min_val = float("-0.35206") @@ -1857,6 +2026,7 @@ class Program_weight_tensor_parameter_168: class Program_weight_tensor_parameter_169: name = "parameter_169" + original_name = "batch_norm2d_32.w_0" shape = [256] dtype = "float32" min_val = float("0.109761") @@ -1868,6 +2038,7 @@ class Program_weight_tensor_parameter_169: class Program_weight_tensor_parameter_170: name = "parameter_170" + original_name = "batch_norm2d_32.w_2" shape = [256] dtype = "float32" min_val = float("0.00447337") @@ -1879,6 +2050,7 @@ class Program_weight_tensor_parameter_170: class Program_weight_tensor_parameter_171: name = "parameter_171" + original_name = "batch_norm2d_32.w_1" shape = [256] dtype = "float32" min_val = float("-0.27309") @@ -1890,6 +2062,7 @@ class Program_weight_tensor_parameter_171: class Program_weight_tensor_parameter_172: name = "parameter_172" + original_name = "conv2d_32.w_0" shape = [256, 256, 3, 3] dtype = "float32" min_val = float("-0.143202") @@ -1901,6 +2074,7 @@ class Program_weight_tensor_parameter_172: class Program_weight_tensor_parameter_173: name = "parameter_173" + original_name = "batch_norm2d_31.b_0" shape = [256] dtype = "float32" min_val = float("-0.244607") @@ -1912,6 +2086,7 @@ class Program_weight_tensor_parameter_173: class Program_weight_tensor_parameter_174: name = "parameter_174" + original_name = "batch_norm2d_31.w_0" shape = [256] dtype = "float32" min_val = float("0.0940088") @@ -1923,6 +2098,7 @@ class Program_weight_tensor_parameter_174: class Program_weight_tensor_parameter_175: name = "parameter_175" + original_name = "batch_norm2d_31.w_2" shape = [256] dtype = "float32" min_val = float("0.00569095") @@ -1934,6 +2110,7 @@ class Program_weight_tensor_parameter_175: class Program_weight_tensor_parameter_176: name = "parameter_176" + original_name = "batch_norm2d_31.w_1" shape = [256] dtype = "float32" min_val = float("-0.216939") @@ -1945,6 +2122,7 @@ class Program_weight_tensor_parameter_176: class Program_weight_tensor_parameter_177: name = "parameter_177" + original_name = "conv2d_31.w_0" shape = [256, 1024, 1, 1] dtype = "float32" min_val = float("-0.194415") @@ -1956,6 +2134,7 @@ class Program_weight_tensor_parameter_177: class Program_weight_tensor_parameter_178: name = "parameter_178" + original_name = "batch_norm2d_30.b_0" shape = [1024] dtype = "float32" min_val = float("-0.28682") @@ -1967,6 +2146,7 @@ class Program_weight_tensor_parameter_178: class Program_weight_tensor_parameter_179: name = "parameter_179" + original_name = "batch_norm2d_30.w_0" shape = [1024] dtype = "float32" min_val = float("-0.095438") @@ -1978,6 +2158,7 @@ class Program_weight_tensor_parameter_179: class Program_weight_tensor_parameter_180: name = "parameter_180" + original_name = "batch_norm2d_30.w_2" shape = [1024] dtype = "float32" min_val = float("5.60519e-45") @@ -1989,6 +2170,7 @@ class Program_weight_tensor_parameter_180: class Program_weight_tensor_parameter_181: name = "parameter_181" + original_name = "batch_norm2d_30.w_1" shape = [1024] dtype = "float32" min_val = float("-0.115913") @@ -2000,6 +2182,7 @@ class Program_weight_tensor_parameter_181: class Program_weight_tensor_parameter_182: name = "parameter_182" + original_name = "conv2d_30.w_0" shape = [1024, 256, 1, 1] dtype = "float32" min_val = float("-0.237181") @@ -2011,6 +2194,7 @@ class Program_weight_tensor_parameter_182: class Program_weight_tensor_parameter_183: name = "parameter_183" + original_name = "batch_norm2d_29.b_0" shape = [256] dtype = "float32" min_val = float("-0.375338") @@ -2022,6 +2206,7 @@ class Program_weight_tensor_parameter_183: class Program_weight_tensor_parameter_184: name = "parameter_184" + original_name = "batch_norm2d_29.w_0" shape = [256] dtype = "float32" min_val = float("0.112331") @@ -2033,6 +2218,7 @@ class Program_weight_tensor_parameter_184: class Program_weight_tensor_parameter_185: name = "parameter_185" + original_name = "batch_norm2d_29.w_2" shape = [256] dtype = "float32" min_val = float("0.00590442") @@ -2044,6 +2230,7 @@ class Program_weight_tensor_parameter_185: class Program_weight_tensor_parameter_186: name = "parameter_186" + original_name = "batch_norm2d_29.w_1" shape = [256] dtype = "float32" min_val = float("-0.703403") @@ -2055,6 +2242,7 @@ class Program_weight_tensor_parameter_186: class Program_weight_tensor_parameter_187: name = "parameter_187" + original_name = "conv2d_29.w_0" shape = [256, 256, 3, 3] dtype = "float32" min_val = float("-0.278981") @@ -2066,6 +2254,7 @@ class Program_weight_tensor_parameter_187: class Program_weight_tensor_parameter_188: name = "parameter_188" + original_name = "batch_norm2d_28.b_0" shape = [256] dtype = "float32" min_val = float("-0.206504") @@ -2077,6 +2266,7 @@ class Program_weight_tensor_parameter_188: class Program_weight_tensor_parameter_189: name = "parameter_189" + original_name = "batch_norm2d_28.w_0" shape = [256] dtype = "float32" min_val = float("0.0505531") @@ -2088,6 +2278,7 @@ class Program_weight_tensor_parameter_189: class Program_weight_tensor_parameter_190: name = "parameter_190" + original_name = "batch_norm2d_28.w_2" shape = [256] dtype = "float32" min_val = float("0.00534171") @@ -2099,6 +2290,7 @@ class Program_weight_tensor_parameter_190: class Program_weight_tensor_parameter_191: name = "parameter_191" + original_name = "batch_norm2d_28.w_1" shape = [256] dtype = "float32" min_val = float("-0.874778") @@ -2110,6 +2302,7 @@ class Program_weight_tensor_parameter_191: class Program_weight_tensor_parameter_192: name = "parameter_192" + original_name = "conv2d_28.w_0" shape = [256, 1024, 1, 1] dtype = "float32" min_val = float("-0.147456") @@ -2121,6 +2314,7 @@ class Program_weight_tensor_parameter_192: class Program_weight_tensor_parameter_193: name = "parameter_193" + original_name = "batch_norm2d_27.b_0" shape = [1024] dtype = "float32" min_val = float("-0.129305") @@ -2132,6 +2326,7 @@ class Program_weight_tensor_parameter_193: class Program_weight_tensor_parameter_194: name = "parameter_194" + original_name = "batch_norm2d_27.w_0" shape = [1024] dtype = "float32" min_val = float("-0.0782405") @@ -2143,6 +2338,7 @@ class Program_weight_tensor_parameter_194: class Program_weight_tensor_parameter_195: name = "parameter_195" + original_name = "batch_norm2d_27.w_2" shape = [1024] dtype = "float32" min_val = float("5.60519e-45") @@ -2154,6 +2350,7 @@ class Program_weight_tensor_parameter_195: class Program_weight_tensor_parameter_196: name = "parameter_196" + original_name = "batch_norm2d_27.w_1" shape = [1024] dtype = "float32" min_val = float("-0.212177") @@ -2165,6 +2362,7 @@ class Program_weight_tensor_parameter_196: class Program_weight_tensor_parameter_197: name = "parameter_197" + original_name = "conv2d_27.w_0" shape = [1024, 512, 1, 1] dtype = "float32" min_val = float("-0.294114") @@ -2176,6 +2374,7 @@ class Program_weight_tensor_parameter_197: class Program_weight_tensor_parameter_198: name = "parameter_198" + original_name = "batch_norm2d_26.b_0" shape = [1024] dtype = "float32" min_val = float("-0.129305") @@ -2187,6 +2386,7 @@ class Program_weight_tensor_parameter_198: class Program_weight_tensor_parameter_199: name = "parameter_199" + original_name = "batch_norm2d_26.w_0" shape = [1024] dtype = "float32" min_val = float("-0.112497") @@ -2198,6 +2398,7 @@ class Program_weight_tensor_parameter_199: class Program_weight_tensor_parameter_200: name = "parameter_200" + original_name = "batch_norm2d_26.w_2" shape = [1024] dtype = "float32" min_val = float("5.60519e-45") @@ -2209,6 +2410,7 @@ class Program_weight_tensor_parameter_200: class Program_weight_tensor_parameter_201: name = "parameter_201" + original_name = "batch_norm2d_26.w_1" shape = [1024] dtype = "float32" min_val = float("-0.274559") @@ -2220,6 +2422,7 @@ class Program_weight_tensor_parameter_201: class Program_weight_tensor_parameter_202: name = "parameter_202" + original_name = "conv2d_26.w_0" shape = [1024, 256, 1, 1] dtype = "float32" min_val = float("-0.270319") @@ -2231,6 +2434,7 @@ class Program_weight_tensor_parameter_202: class Program_weight_tensor_parameter_203: name = "parameter_203" + original_name = "batch_norm2d_25.b_0" shape = [256] dtype = "float32" min_val = float("-0.169522") @@ -2242,6 +2446,7 @@ class Program_weight_tensor_parameter_203: class Program_weight_tensor_parameter_204: name = "parameter_204" + original_name = "batch_norm2d_25.w_0" shape = [256] dtype = "float32" min_val = float("0.134911") @@ -2253,6 +2458,7 @@ class Program_weight_tensor_parameter_204: class Program_weight_tensor_parameter_205: name = "parameter_205" + original_name = "batch_norm2d_25.w_2" shape = [256] dtype = "float32" min_val = float("0.00830091") @@ -2264,6 +2470,7 @@ class Program_weight_tensor_parameter_205: class Program_weight_tensor_parameter_206: name = "parameter_206" + original_name = "batch_norm2d_25.w_1" shape = [256] dtype = "float32" min_val = float("-0.334949") @@ -2275,6 +2482,7 @@ class Program_weight_tensor_parameter_206: class Program_weight_tensor_parameter_207: name = "parameter_207" + original_name = "conv2d_25.w_0" shape = [256, 256, 3, 3] dtype = "float32" min_val = float("-0.161917") @@ -2286,6 +2494,7 @@ class Program_weight_tensor_parameter_207: class Program_weight_tensor_parameter_208: name = "parameter_208" + original_name = "batch_norm2d_24.b_0" shape = [256] dtype = "float32" min_val = float("-0.37354") @@ -2297,6 +2506,7 @@ class Program_weight_tensor_parameter_208: class Program_weight_tensor_parameter_209: name = "parameter_209" + original_name = "batch_norm2d_24.w_0" shape = [256] dtype = "float32" min_val = float("0.131707") @@ -2308,6 +2518,7 @@ class Program_weight_tensor_parameter_209: class Program_weight_tensor_parameter_210: name = "parameter_210" + original_name = "batch_norm2d_24.w_2" shape = [256] dtype = "float32" min_val = float("0.00841896") @@ -2319,6 +2530,7 @@ class Program_weight_tensor_parameter_210: class Program_weight_tensor_parameter_211: name = "parameter_211" + original_name = "batch_norm2d_24.w_1" shape = [256] dtype = "float32" min_val = float("-0.320503") @@ -2330,6 +2542,7 @@ class Program_weight_tensor_parameter_211: class Program_weight_tensor_parameter_212: name = "parameter_212" + original_name = "conv2d_24.w_0" shape = [256, 512, 1, 1] dtype = "float32" min_val = float("-0.255429") @@ -2341,6 +2554,7 @@ class Program_weight_tensor_parameter_212: class Program_weight_tensor_parameter_213: name = "parameter_213" + original_name = "batch_norm2d_23.b_0" shape = [512] dtype = "float32" min_val = float("-0.265757") @@ -2352,6 +2566,7 @@ class Program_weight_tensor_parameter_213: class Program_weight_tensor_parameter_214: name = "parameter_214" + original_name = "batch_norm2d_23.w_0" shape = [512] dtype = "float32" min_val = float("-0.118525") @@ -2363,6 +2578,7 @@ class Program_weight_tensor_parameter_214: class Program_weight_tensor_parameter_215: name = "parameter_215" + original_name = "batch_norm2d_23.w_2" shape = [512] dtype = "float32" min_val = float("6.90463e-05") @@ -2374,6 +2590,7 @@ class Program_weight_tensor_parameter_215: class Program_weight_tensor_parameter_216: name = "parameter_216" + original_name = "batch_norm2d_23.w_1" shape = [512] dtype = "float32" min_val = float("-0.118861") @@ -2385,6 +2602,7 @@ class Program_weight_tensor_parameter_216: class Program_weight_tensor_parameter_217: name = "parameter_217" + original_name = "conv2d_23.w_0" shape = [512, 128, 1, 1] dtype = "float32" min_val = float("-0.176259") @@ -2396,6 +2614,7 @@ class Program_weight_tensor_parameter_217: class Program_weight_tensor_parameter_218: name = "parameter_218" + original_name = "batch_norm2d_22.b_0" shape = [128] dtype = "float32" min_val = float("-0.232824") @@ -2407,6 +2626,7 @@ class Program_weight_tensor_parameter_218: class Program_weight_tensor_parameter_219: name = "parameter_219" + original_name = "batch_norm2d_22.w_0" shape = [128] dtype = "float32" min_val = float("0.118575") @@ -2418,6 +2638,7 @@ class Program_weight_tensor_parameter_219: class Program_weight_tensor_parameter_220: name = "parameter_220" + original_name = "batch_norm2d_22.w_2" shape = [128] dtype = "float32" min_val = float("0.00369018") @@ -2429,6 +2650,7 @@ class Program_weight_tensor_parameter_220: class Program_weight_tensor_parameter_221: name = "parameter_221" + original_name = "batch_norm2d_22.w_1" shape = [128] dtype = "float32" min_val = float("-0.258314") @@ -2440,6 +2662,7 @@ class Program_weight_tensor_parameter_221: class Program_weight_tensor_parameter_222: name = "parameter_222" + original_name = "conv2d_22.w_0" shape = [128, 128, 3, 3] dtype = "float32" min_val = float("-0.14727") @@ -2451,6 +2674,7 @@ class Program_weight_tensor_parameter_222: class Program_weight_tensor_parameter_223: name = "parameter_223" + original_name = "batch_norm2d_21.b_0" shape = [128] dtype = "float32" min_val = float("-0.217196") @@ -2462,6 +2686,7 @@ class Program_weight_tensor_parameter_223: class Program_weight_tensor_parameter_224: name = "parameter_224" + original_name = "batch_norm2d_21.w_0" shape = [128] dtype = "float32" min_val = float("0.0919349") @@ -2473,6 +2698,7 @@ class Program_weight_tensor_parameter_224: class Program_weight_tensor_parameter_225: name = "parameter_225" + original_name = "batch_norm2d_21.w_2" shape = [128] dtype = "float32" min_val = float("0.00814274") @@ -2484,6 +2710,7 @@ class Program_weight_tensor_parameter_225: class Program_weight_tensor_parameter_226: name = "parameter_226" + original_name = "batch_norm2d_21.w_1" shape = [128] dtype = "float32" min_val = float("-0.257526") @@ -2495,6 +2722,7 @@ class Program_weight_tensor_parameter_226: class Program_weight_tensor_parameter_227: name = "parameter_227" + original_name = "conv2d_21.w_0" shape = [128, 512, 1, 1] dtype = "float32" min_val = float("-0.170298") @@ -2506,6 +2734,7 @@ class Program_weight_tensor_parameter_227: class Program_weight_tensor_parameter_228: name = "parameter_228" + original_name = "batch_norm2d_20.b_0" shape = [512] dtype = "float32" min_val = float("-0.237422") @@ -2517,6 +2746,7 @@ class Program_weight_tensor_parameter_228: class Program_weight_tensor_parameter_229: name = "parameter_229" + original_name = "batch_norm2d_20.w_0" shape = [512] dtype = "float32" min_val = float("-0.139337") @@ -2528,6 +2758,7 @@ class Program_weight_tensor_parameter_229: class Program_weight_tensor_parameter_230: name = "parameter_230" + original_name = "batch_norm2d_20.w_2" shape = [512] dtype = "float32" min_val = float("5.60519e-45") @@ -2539,6 +2770,7 @@ class Program_weight_tensor_parameter_230: class Program_weight_tensor_parameter_231: name = "parameter_231" + original_name = "batch_norm2d_20.w_1" shape = [512] dtype = "float32" min_val = float("-0.117879") @@ -2550,6 +2782,7 @@ class Program_weight_tensor_parameter_231: class Program_weight_tensor_parameter_232: name = "parameter_232" + original_name = "conv2d_20.w_0" shape = [512, 128, 1, 1] dtype = "float32" min_val = float("-0.231102") @@ -2561,6 +2794,7 @@ class Program_weight_tensor_parameter_232: class Program_weight_tensor_parameter_233: name = "parameter_233" + original_name = "batch_norm2d_19.b_0" shape = [128] dtype = "float32" min_val = float("-0.242849") @@ -2572,6 +2806,7 @@ class Program_weight_tensor_parameter_233: class Program_weight_tensor_parameter_234: name = "parameter_234" + original_name = "batch_norm2d_19.w_0" shape = [128] dtype = "float32" min_val = float("0.105254") @@ -2583,6 +2818,7 @@ class Program_weight_tensor_parameter_234: class Program_weight_tensor_parameter_235: name = "parameter_235" + original_name = "batch_norm2d_19.w_2" shape = [128] dtype = "float32" min_val = float("0.00364882") @@ -2594,6 +2830,7 @@ class Program_weight_tensor_parameter_235: class Program_weight_tensor_parameter_236: name = "parameter_236" + original_name = "batch_norm2d_19.w_1" shape = [128] dtype = "float32" min_val = float("-0.254192") @@ -2605,6 +2842,7 @@ class Program_weight_tensor_parameter_236: class Program_weight_tensor_parameter_237: name = "parameter_237" + original_name = "conv2d_19.w_0" shape = [128, 128, 3, 3] dtype = "float32" min_val = float("-0.171269") @@ -2616,6 +2854,7 @@ class Program_weight_tensor_parameter_237: class Program_weight_tensor_parameter_238: name = "parameter_238" + original_name = "batch_norm2d_18.b_0" shape = [128] dtype = "float32" min_val = float("-0.193408") @@ -2627,6 +2866,7 @@ class Program_weight_tensor_parameter_238: class Program_weight_tensor_parameter_239: name = "parameter_239" + original_name = "batch_norm2d_18.w_0" shape = [128] dtype = "float32" min_val = float("0.105105") @@ -2638,6 +2878,7 @@ class Program_weight_tensor_parameter_239: class Program_weight_tensor_parameter_240: name = "parameter_240" + original_name = "batch_norm2d_18.w_2" shape = [128] dtype = "float32" min_val = float("0.00566029") @@ -2649,6 +2890,7 @@ class Program_weight_tensor_parameter_240: class Program_weight_tensor_parameter_241: name = "parameter_241" + original_name = "batch_norm2d_18.w_1" shape = [128] dtype = "float32" min_val = float("-0.218906") @@ -2660,6 +2902,7 @@ class Program_weight_tensor_parameter_241: class Program_weight_tensor_parameter_242: name = "parameter_242" + original_name = "conv2d_18.w_0" shape = [128, 512, 1, 1] dtype = "float32" min_val = float("-0.159413") @@ -2671,6 +2914,7 @@ class Program_weight_tensor_parameter_242: class Program_weight_tensor_parameter_243: name = "parameter_243" + original_name = "batch_norm2d_17.b_0" shape = [512] dtype = "float32" min_val = float("-0.24628") @@ -2682,6 +2926,7 @@ class Program_weight_tensor_parameter_243: class Program_weight_tensor_parameter_244: name = "parameter_244" + original_name = "batch_norm2d_17.w_0" shape = [512] dtype = "float32" min_val = float("-0.192908") @@ -2693,6 +2938,7 @@ class Program_weight_tensor_parameter_244: class Program_weight_tensor_parameter_245: name = "parameter_245" + original_name = "batch_norm2d_17.w_2" shape = [512] dtype = "float32" min_val = float("5.60519e-45") @@ -2704,6 +2950,7 @@ class Program_weight_tensor_parameter_245: class Program_weight_tensor_parameter_246: name = "parameter_246" + original_name = "batch_norm2d_17.w_1" shape = [512] dtype = "float32" min_val = float("-0.127209") @@ -2715,6 +2962,7 @@ class Program_weight_tensor_parameter_246: class Program_weight_tensor_parameter_247: name = "parameter_247" + original_name = "conv2d_17.w_0" shape = [512, 128, 1, 1] dtype = "float32" min_val = float("-0.322905") @@ -2726,6 +2974,7 @@ class Program_weight_tensor_parameter_247: class Program_weight_tensor_parameter_248: name = "parameter_248" + original_name = "batch_norm2d_16.b_0" shape = [128] dtype = "float32" min_val = float("-0.361119") @@ -2737,6 +2986,7 @@ class Program_weight_tensor_parameter_248: class Program_weight_tensor_parameter_249: name = "parameter_249" + original_name = "batch_norm2d_16.w_0" shape = [128] dtype = "float32" min_val = float("0.108785") @@ -2748,6 +2998,7 @@ class Program_weight_tensor_parameter_249: class Program_weight_tensor_parameter_250: name = "parameter_250" + original_name = "batch_norm2d_16.w_2" shape = [128] dtype = "float32" min_val = float("0.0020564") @@ -2759,6 +3010,7 @@ class Program_weight_tensor_parameter_250: class Program_weight_tensor_parameter_251: name = "parameter_251" + original_name = "batch_norm2d_16.w_1" shape = [128] dtype = "float32" min_val = float("-0.881857") @@ -2770,6 +3022,7 @@ class Program_weight_tensor_parameter_251: class Program_weight_tensor_parameter_252: name = "parameter_252" + original_name = "conv2d_16.w_0" shape = [128, 128, 3, 3] dtype = "float32" min_val = float("-0.243456") @@ -2781,6 +3034,7 @@ class Program_weight_tensor_parameter_252: class Program_weight_tensor_parameter_253: name = "parameter_253" + original_name = "batch_norm2d_15.b_0" shape = [128] dtype = "float32" min_val = float("-0.338885") @@ -2792,6 +3046,7 @@ class Program_weight_tensor_parameter_253: class Program_weight_tensor_parameter_254: name = "parameter_254" + original_name = "batch_norm2d_15.w_0" shape = [128] dtype = "float32" min_val = float("0.0665754") @@ -2803,6 +3058,7 @@ class Program_weight_tensor_parameter_254: class Program_weight_tensor_parameter_255: name = "parameter_255" + original_name = "batch_norm2d_15.w_2" shape = [128] dtype = "float32" min_val = float("0.00275181") @@ -2814,6 +3070,7 @@ class Program_weight_tensor_parameter_255: class Program_weight_tensor_parameter_256: name = "parameter_256" + original_name = "batch_norm2d_15.w_1" shape = [128] dtype = "float32" min_val = float("-0.271882") @@ -2825,6 +3082,7 @@ class Program_weight_tensor_parameter_256: class Program_weight_tensor_parameter_257: name = "parameter_257" + original_name = "conv2d_15.w_0" shape = [128, 512, 1, 1] dtype = "float32" min_val = float("-0.195535") @@ -2836,6 +3094,7 @@ class Program_weight_tensor_parameter_257: class Program_weight_tensor_parameter_258: name = "parameter_258" + original_name = "batch_norm2d_14.b_0" shape = [512] dtype = "float32" min_val = float("-0.173999") @@ -2847,6 +3106,7 @@ class Program_weight_tensor_parameter_258: class Program_weight_tensor_parameter_259: name = "parameter_259" + original_name = "batch_norm2d_14.w_0" shape = [512] dtype = "float32" min_val = float("-0.0847974") @@ -2858,6 +3118,7 @@ class Program_weight_tensor_parameter_259: class Program_weight_tensor_parameter_260: name = "parameter_260" + original_name = "batch_norm2d_14.w_2" shape = [512] dtype = "float32" min_val = float("5.60519e-45") @@ -2869,6 +3130,7 @@ class Program_weight_tensor_parameter_260: class Program_weight_tensor_parameter_261: name = "parameter_261" + original_name = "batch_norm2d_14.w_1" shape = [512] dtype = "float32" min_val = float("-0.532972") @@ -2880,6 +3142,7 @@ class Program_weight_tensor_parameter_261: class Program_weight_tensor_parameter_262: name = "parameter_262" + original_name = "conv2d_14.w_0" shape = [512, 256, 1, 1] dtype = "float32" min_val = float("-0.411257") @@ -2891,6 +3154,7 @@ class Program_weight_tensor_parameter_262: class Program_weight_tensor_parameter_263: name = "parameter_263" + original_name = "batch_norm2d_13.b_0" shape = [512] dtype = "float32" min_val = float("-0.173999") @@ -2902,6 +3166,7 @@ class Program_weight_tensor_parameter_263: class Program_weight_tensor_parameter_264: name = "parameter_264" + original_name = "batch_norm2d_13.w_0" shape = [512] dtype = "float32" min_val = float("-0.110676") @@ -2913,6 +3178,7 @@ class Program_weight_tensor_parameter_264: class Program_weight_tensor_parameter_265: name = "parameter_265" + original_name = "batch_norm2d_13.w_2" shape = [512] dtype = "float32" min_val = float("5.60519e-45") @@ -2924,6 +3190,7 @@ class Program_weight_tensor_parameter_265: class Program_weight_tensor_parameter_266: name = "parameter_266" + original_name = "batch_norm2d_13.w_1" shape = [512] dtype = "float32" min_val = float("-0.415464") @@ -2935,6 +3202,7 @@ class Program_weight_tensor_parameter_266: class Program_weight_tensor_parameter_267: name = "parameter_267" + original_name = "conv2d_13.w_0" shape = [512, 128, 1, 1] dtype = "float32" min_val = float("-0.306972") @@ -2946,6 +3214,7 @@ class Program_weight_tensor_parameter_267: class Program_weight_tensor_parameter_268: name = "parameter_268" + original_name = "batch_norm2d_12.b_0" shape = [128] dtype = "float32" min_val = float("-0.0749191") @@ -2957,6 +3226,7 @@ class Program_weight_tensor_parameter_268: class Program_weight_tensor_parameter_269: name = "parameter_269" + original_name = "batch_norm2d_12.w_0" shape = [128] dtype = "float32" min_val = float("0.151982") @@ -2968,6 +3238,7 @@ class Program_weight_tensor_parameter_269: class Program_weight_tensor_parameter_270: name = "parameter_270" + original_name = "batch_norm2d_12.w_2" shape = [128] dtype = "float32" min_val = float("0.011571") @@ -2979,6 +3250,7 @@ class Program_weight_tensor_parameter_270: class Program_weight_tensor_parameter_271: name = "parameter_271" + original_name = "batch_norm2d_12.w_1" shape = [128] dtype = "float32" min_val = float("-0.199453") @@ -2990,6 +3262,7 @@ class Program_weight_tensor_parameter_271: class Program_weight_tensor_parameter_272: name = "parameter_272" + original_name = "conv2d_12.w_0" shape = [128, 128, 3, 3] dtype = "float32" min_val = float("-0.152539") @@ -3001,6 +3274,7 @@ class Program_weight_tensor_parameter_272: class Program_weight_tensor_parameter_273: name = "parameter_273" + original_name = "batch_norm2d_11.b_0" shape = [128] dtype = "float32" min_val = float("-0.254357") @@ -3012,6 +3286,7 @@ class Program_weight_tensor_parameter_273: class Program_weight_tensor_parameter_274: name = "parameter_274" + original_name = "batch_norm2d_11.w_0" shape = [128] dtype = "float32" min_val = float("7.24973e-19") @@ -3023,6 +3298,7 @@ class Program_weight_tensor_parameter_274: class Program_weight_tensor_parameter_275: name = "parameter_275" + original_name = "batch_norm2d_11.w_2" shape = [128] dtype = "float32" min_val = float("5.60519e-45") @@ -3034,6 +3310,7 @@ class Program_weight_tensor_parameter_275: class Program_weight_tensor_parameter_276: name = "parameter_276" + original_name = "batch_norm2d_11.w_1" shape = [128] dtype = "float32" min_val = float("-0.268525") @@ -3045,6 +3322,7 @@ class Program_weight_tensor_parameter_276: class Program_weight_tensor_parameter_277: name = "parameter_277" + original_name = "conv2d_11.w_0" shape = [128, 256, 1, 1] dtype = "float32" min_val = float("-0.224242") @@ -3056,6 +3334,7 @@ class Program_weight_tensor_parameter_277: class Program_weight_tensor_parameter_278: name = "parameter_278" + original_name = "batch_norm2d_10.b_0" shape = [256] dtype = "float32" min_val = float("-0.164026") @@ -3067,6 +3346,7 @@ class Program_weight_tensor_parameter_278: class Program_weight_tensor_parameter_279: name = "parameter_279" + original_name = "batch_norm2d_10.w_0" shape = [256] dtype = "float32" min_val = float("-0.255219") @@ -3078,6 +3358,7 @@ class Program_weight_tensor_parameter_279: class Program_weight_tensor_parameter_280: name = "parameter_280" + original_name = "batch_norm2d_10.w_2" shape = [256] dtype = "float32" min_val = float("7.12898e-05") @@ -3089,6 +3370,7 @@ class Program_weight_tensor_parameter_280: class Program_weight_tensor_parameter_281: name = "parameter_281" + original_name = "batch_norm2d_10.w_1" shape = [256] dtype = "float32" min_val = float("-0.133956") @@ -3100,6 +3382,7 @@ class Program_weight_tensor_parameter_281: class Program_weight_tensor_parameter_282: name = "parameter_282" + original_name = "conv2d_10.w_0" shape = [256, 64, 1, 1] dtype = "float32" min_val = float("-0.325233") @@ -3111,6 +3394,7 @@ class Program_weight_tensor_parameter_282: class Program_weight_tensor_parameter_283: name = "parameter_283" + original_name = "batch_norm2d_9.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -3120,6 +3404,7 @@ class Program_weight_tensor_parameter_283: class Program_weight_tensor_parameter_284: name = "parameter_284" + original_name = "batch_norm2d_9.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -3129,6 +3414,7 @@ class Program_weight_tensor_parameter_284: class Program_weight_tensor_parameter_285: name = "parameter_285" + original_name = "batch_norm2d_9.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -3138,6 +3424,7 @@ class Program_weight_tensor_parameter_285: class Program_weight_tensor_parameter_286: name = "parameter_286" + original_name = "batch_norm2d_9.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -3147,6 +3434,7 @@ class Program_weight_tensor_parameter_286: class Program_weight_tensor_parameter_287: name = "parameter_287" + original_name = "conv2d_9.w_0" shape = [64, 64, 3, 3] dtype = "float32" min_val = float("-0.259976") @@ -3158,6 +3446,7 @@ class Program_weight_tensor_parameter_287: class Program_weight_tensor_parameter_288: name = "parameter_288" + original_name = "batch_norm2d_8.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -3167,6 +3456,7 @@ class Program_weight_tensor_parameter_288: class Program_weight_tensor_parameter_289: name = "parameter_289" + original_name = "batch_norm2d_8.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -3176,6 +3466,7 @@ class Program_weight_tensor_parameter_289: class Program_weight_tensor_parameter_290: name = "parameter_290" + original_name = "batch_norm2d_8.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -3185,6 +3476,7 @@ class Program_weight_tensor_parameter_290: class Program_weight_tensor_parameter_291: name = "parameter_291" + original_name = "batch_norm2d_8.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -3194,6 +3486,7 @@ class Program_weight_tensor_parameter_291: class Program_weight_tensor_parameter_292: name = "parameter_292" + original_name = "conv2d_8.w_0" shape = [64, 256, 1, 1] dtype = "float32" min_val = float("-0.188924") @@ -3205,6 +3498,7 @@ class Program_weight_tensor_parameter_292: class Program_weight_tensor_parameter_293: name = "parameter_293" + original_name = "batch_norm2d_7.b_0" shape = [256] dtype = "float32" min_val = float("-0.177785") @@ -3216,6 +3510,7 @@ class Program_weight_tensor_parameter_293: class Program_weight_tensor_parameter_294: name = "parameter_294" + original_name = "batch_norm2d_7.w_0" shape = [256] dtype = "float32" min_val = float("-0.203582") @@ -3227,6 +3522,7 @@ class Program_weight_tensor_parameter_294: class Program_weight_tensor_parameter_295: name = "parameter_295" + original_name = "batch_norm2d_7.w_2" shape = [256] dtype = "float32" min_val = float("5.60519e-45") @@ -3238,6 +3534,7 @@ class Program_weight_tensor_parameter_295: class Program_weight_tensor_parameter_296: name = "parameter_296" + original_name = "batch_norm2d_7.w_1" shape = [256] dtype = "float32" min_val = float("-0.0838559") @@ -3249,6 +3546,7 @@ class Program_weight_tensor_parameter_296: class Program_weight_tensor_parameter_297: name = "parameter_297" + original_name = "conv2d_7.w_0" shape = [256, 64, 1, 1] dtype = "float32" min_val = float("-0.25193") @@ -3260,6 +3558,7 @@ class Program_weight_tensor_parameter_297: class Program_weight_tensor_parameter_298: name = "parameter_298" + original_name = "batch_norm2d_6.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -3269,6 +3568,7 @@ class Program_weight_tensor_parameter_298: class Program_weight_tensor_parameter_299: name = "parameter_299" + original_name = "batch_norm2d_6.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -3278,6 +3578,7 @@ class Program_weight_tensor_parameter_299: class Program_weight_tensor_parameter_300: name = "parameter_300" + original_name = "batch_norm2d_6.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -3287,6 +3588,7 @@ class Program_weight_tensor_parameter_300: class Program_weight_tensor_parameter_301: name = "parameter_301" + original_name = "batch_norm2d_6.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -3296,6 +3598,7 @@ class Program_weight_tensor_parameter_301: class Program_weight_tensor_parameter_302: name = "parameter_302" + original_name = "conv2d_6.w_0" shape = [64, 64, 3, 3] dtype = "float32" min_val = float("-0.277854") @@ -3307,6 +3610,7 @@ class Program_weight_tensor_parameter_302: class Program_weight_tensor_parameter_303: name = "parameter_303" + original_name = "batch_norm2d_5.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -3316,6 +3620,7 @@ class Program_weight_tensor_parameter_303: class Program_weight_tensor_parameter_304: name = "parameter_304" + original_name = "batch_norm2d_5.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -3325,6 +3630,7 @@ class Program_weight_tensor_parameter_304: class Program_weight_tensor_parameter_305: name = "parameter_305" + original_name = "batch_norm2d_5.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -3334,6 +3640,7 @@ class Program_weight_tensor_parameter_305: class Program_weight_tensor_parameter_306: name = "parameter_306" + original_name = "batch_norm2d_5.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -3343,6 +3650,7 @@ class Program_weight_tensor_parameter_306: class Program_weight_tensor_parameter_307: name = "parameter_307" + original_name = "conv2d_5.w_0" shape = [64, 256, 1, 1] dtype = "float32" min_val = float("-0.340224") @@ -3354,6 +3662,7 @@ class Program_weight_tensor_parameter_307: class Program_weight_tensor_parameter_308: name = "parameter_308" + original_name = "batch_norm2d_4.b_0" shape = [256] dtype = "float32" min_val = float("-0.214949") @@ -3365,6 +3674,7 @@ class Program_weight_tensor_parameter_308: class Program_weight_tensor_parameter_309: name = "parameter_309" + original_name = "batch_norm2d_4.w_0" shape = [256] dtype = "float32" min_val = float("-0.140797") @@ -3376,6 +3686,7 @@ class Program_weight_tensor_parameter_309: class Program_weight_tensor_parameter_310: name = "parameter_310" + original_name = "batch_norm2d_4.w_2" shape = [256] dtype = "float32" min_val = float("5.60519e-45") @@ -3387,6 +3698,7 @@ class Program_weight_tensor_parameter_310: class Program_weight_tensor_parameter_311: name = "parameter_311" + original_name = "batch_norm2d_4.w_1" shape = [256] dtype = "float32" min_val = float("-0.788633") @@ -3398,6 +3710,7 @@ class Program_weight_tensor_parameter_311: class Program_weight_tensor_parameter_312: name = "parameter_312" + original_name = "conv2d_4.w_0" shape = [256, 64, 1, 1] dtype = "float32" min_val = float("-0.636505") @@ -3409,6 +3722,7 @@ class Program_weight_tensor_parameter_312: class Program_weight_tensor_parameter_313: name = "parameter_313" + original_name = "batch_norm2d_3.b_0" shape = [256] dtype = "float32" min_val = float("-0.214949") @@ -3420,6 +3734,7 @@ class Program_weight_tensor_parameter_313: class Program_weight_tensor_parameter_314: name = "parameter_314" + original_name = "batch_norm2d_3.w_0" shape = [256] dtype = "float32" min_val = float("-0.351382") @@ -3431,6 +3746,7 @@ class Program_weight_tensor_parameter_314: class Program_weight_tensor_parameter_315: name = "parameter_315" + original_name = "batch_norm2d_3.w_2" shape = [256] dtype = "float32" min_val = float("5.60519e-45") @@ -3442,6 +3758,7 @@ class Program_weight_tensor_parameter_315: class Program_weight_tensor_parameter_316: name = "parameter_316" + original_name = "batch_norm2d_3.w_1" shape = [256] dtype = "float32" min_val = float("-0.221928") @@ -3453,6 +3770,7 @@ class Program_weight_tensor_parameter_316: class Program_weight_tensor_parameter_317: name = "parameter_317" + original_name = "conv2d_3.w_0" shape = [256, 64, 1, 1] dtype = "float32" min_val = float("-0.416398") @@ -3464,6 +3782,7 @@ class Program_weight_tensor_parameter_317: class Program_weight_tensor_parameter_318: name = "parameter_318" + original_name = "batch_norm2d_2.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -3473,6 +3792,7 @@ class Program_weight_tensor_parameter_318: class Program_weight_tensor_parameter_319: name = "parameter_319" + original_name = "batch_norm2d_2.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -3482,6 +3802,7 @@ class Program_weight_tensor_parameter_319: class Program_weight_tensor_parameter_320: name = "parameter_320" + original_name = "batch_norm2d_2.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -3491,6 +3812,7 @@ class Program_weight_tensor_parameter_320: class Program_weight_tensor_parameter_321: name = "parameter_321" + original_name = "batch_norm2d_2.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -3500,6 +3822,7 @@ class Program_weight_tensor_parameter_321: class Program_weight_tensor_parameter_322: name = "parameter_322" + original_name = "conv2d_2.w_0" shape = [64, 64, 3, 3] dtype = "float32" min_val = float("-0.367549") @@ -3511,6 +3834,7 @@ class Program_weight_tensor_parameter_322: class Program_weight_tensor_parameter_323: name = "parameter_323" + original_name = "batch_norm2d_1.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -3520,6 +3844,7 @@ class Program_weight_tensor_parameter_323: class Program_weight_tensor_parameter_324: name = "parameter_324" + original_name = "batch_norm2d_1.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -3529,6 +3854,7 @@ class Program_weight_tensor_parameter_324: class Program_weight_tensor_parameter_325: name = "parameter_325" + original_name = "batch_norm2d_1.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -3538,6 +3864,7 @@ class Program_weight_tensor_parameter_325: class Program_weight_tensor_parameter_326: name = "parameter_326" + original_name = "batch_norm2d_1.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -3547,6 +3874,7 @@ class Program_weight_tensor_parameter_326: class Program_weight_tensor_parameter_327: name = "parameter_327" + original_name = "conv2d_1.w_0" shape = [64, 64, 1, 1] dtype = "float32" min_val = float("-0.628423") @@ -3558,6 +3886,7 @@ class Program_weight_tensor_parameter_327: class Program_weight_tensor_parameter_328: name = "parameter_328" + original_name = "batch_norm2d_0.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -3567,6 +3896,7 @@ class Program_weight_tensor_parameter_328: class Program_weight_tensor_parameter_329: name = "parameter_329" + original_name = "batch_norm2d_0.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -3576,6 +3906,7 @@ class Program_weight_tensor_parameter_329: class Program_weight_tensor_parameter_330: name = "parameter_330" + original_name = "batch_norm2d_0.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -3585,6 +3916,7 @@ class Program_weight_tensor_parameter_330: class Program_weight_tensor_parameter_331: name = "parameter_331" + original_name = "batch_norm2d_0.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -3594,6 +3926,7 @@ class Program_weight_tensor_parameter_331: class Program_weight_tensor_parameter_332: name = "parameter_332" + original_name = "conv2d_0.w_0" shape = [64, 3, 7, 7] dtype = "float32" min_val = float("-0.593645") diff --git a/paddle_samples/PaddleX/SOLOv2/subgraph_2/input_meta.py b/paddle_samples/PaddleX/SOLOv2/subgraph_2/input_meta.py index 8950b9023..578896fe3 100644 --- a/paddle_samples/PaddleX/SOLOv2/subgraph_2/input_meta.py +++ b/paddle_samples/PaddleX/SOLOv2/subgraph_2/input_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_data_0: name = "data_0" + original_name = "var_2302" shape = [1, 3, 800, 928] dtype = "float32" min_val = float("-2.1179") diff --git a/paddle_samples/PaddleX/SOLOv2/subgraph_2/weight_meta.py b/paddle_samples/PaddleX/SOLOv2/subgraph_2/weight_meta.py index 8e12d2616..66985301b 100644 --- a/paddle_samples/PaddleX/SOLOv2/subgraph_2/weight_meta.py +++ b/paddle_samples/PaddleX/SOLOv2/subgraph_2/weight_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_parameter_0: name = "parameter_0" + original_name = "conv2d_70.b_0" shape = [2] dtype = "float32" min_val = float("0") @@ -9,6 +10,7 @@ class Program_weight_tensor_parameter_0: class Program_weight_tensor_parameter_1: name = "parameter_1" + original_name = "conv2d_70.w_0" shape = [2, 512, 3, 3] dtype = "float32" min_val = float("-0.0408139") @@ -20,6 +22,7 @@ class Program_weight_tensor_parameter_1: class Program_weight_tensor_parameter_2: name = "parameter_2" + original_name = "group_norm_7.b_0" shape = [512] dtype = "float32" min_val = float("-0.164576") @@ -31,6 +34,7 @@ class Program_weight_tensor_parameter_2: class Program_weight_tensor_parameter_3: name = "parameter_3" + original_name = "group_norm_7.w_0" shape = [512] dtype = "float32" min_val = float("0.820262") @@ -42,6 +46,7 @@ class Program_weight_tensor_parameter_3: class Program_weight_tensor_parameter_4: name = "parameter_4" + original_name = "conv2d_68.w_0" shape = [512, 512, 3, 3] dtype = "float32" min_val = float("-0.0716389") @@ -53,6 +58,7 @@ class Program_weight_tensor_parameter_4: class Program_weight_tensor_parameter_5: name = "parameter_5" + original_name = "group_norm_5.b_0" shape = [512] dtype = "float32" min_val = float("-0.140273") @@ -64,6 +70,7 @@ class Program_weight_tensor_parameter_5: class Program_weight_tensor_parameter_6: name = "parameter_6" + original_name = "group_norm_5.w_0" shape = [512] dtype = "float32" min_val = float("0.849924") @@ -75,6 +82,7 @@ class Program_weight_tensor_parameter_6: class Program_weight_tensor_parameter_7: name = "parameter_7" + original_name = "conv2d_66.w_0" shape = [512, 512, 3, 3] dtype = "float32" min_val = float("-0.0881508") @@ -86,6 +94,7 @@ class Program_weight_tensor_parameter_7: class Program_weight_tensor_parameter_8: name = "parameter_8" + original_name = "group_norm_3.b_0" shape = [512] dtype = "float32" min_val = float("-0.133432") @@ -97,6 +106,7 @@ class Program_weight_tensor_parameter_8: class Program_weight_tensor_parameter_9: name = "parameter_9" + original_name = "group_norm_3.w_0" shape = [512] dtype = "float32" min_val = float("0.896877") @@ -108,6 +118,7 @@ class Program_weight_tensor_parameter_9: class Program_weight_tensor_parameter_10: name = "parameter_10" + original_name = "conv2d_64.w_0" shape = [512, 512, 3, 3] dtype = "float32" min_val = float("-0.0774497") @@ -119,6 +130,7 @@ class Program_weight_tensor_parameter_10: class Program_weight_tensor_parameter_11: name = "parameter_11" + original_name = "group_norm_1.b_0" shape = [512] dtype = "float32" min_val = float("-0.0835772") @@ -130,6 +142,7 @@ class Program_weight_tensor_parameter_11: class Program_weight_tensor_parameter_12: name = "parameter_12" + original_name = "group_norm_1.w_0" shape = [512] dtype = "float32" min_val = float("0.930588") @@ -141,6 +154,7 @@ class Program_weight_tensor_parameter_12: class Program_weight_tensor_parameter_13: name = "parameter_13" + original_name = "conv2d_62.w_0" shape = [512, 256, 3, 3] dtype = "float32" min_val = float("-0.0527954") @@ -152,6 +166,7 @@ class Program_weight_tensor_parameter_13: class Program_weight_tensor_parameter_14: name = "parameter_14" + original_name = "conv2d_69.b_0" shape = [256] dtype = "float32" min_val = float("-0.108704") @@ -163,6 +178,7 @@ class Program_weight_tensor_parameter_14: class Program_weight_tensor_parameter_15: name = "parameter_15" + original_name = "conv2d_69.w_0" shape = [256, 512, 3, 3] dtype = "float32" min_val = float("-0.114528") @@ -174,6 +190,7 @@ class Program_weight_tensor_parameter_15: class Program_weight_tensor_parameter_16: name = "parameter_16" + original_name = "group_norm_6.b_0" shape = [512] dtype = "float32" min_val = float("-0.167228") @@ -185,6 +202,7 @@ class Program_weight_tensor_parameter_16: class Program_weight_tensor_parameter_17: name = "parameter_17" + original_name = "group_norm_6.w_0" shape = [512] dtype = "float32" min_val = float("0.527656") @@ -196,6 +214,7 @@ class Program_weight_tensor_parameter_17: class Program_weight_tensor_parameter_18: name = "parameter_18" + original_name = "conv2d_67.w_0" shape = [512, 512, 3, 3] dtype = "float32" min_val = float("-0.0770595") @@ -207,6 +226,7 @@ class Program_weight_tensor_parameter_18: class Program_weight_tensor_parameter_19: name = "parameter_19" + original_name = "group_norm_4.b_0" shape = [512] dtype = "float32" min_val = float("-0.109454") @@ -218,6 +238,7 @@ class Program_weight_tensor_parameter_19: class Program_weight_tensor_parameter_20: name = "parameter_20" + original_name = "group_norm_4.w_0" shape = [512] dtype = "float32" min_val = float("0.568624") @@ -229,6 +250,7 @@ class Program_weight_tensor_parameter_20: class Program_weight_tensor_parameter_21: name = "parameter_21" + original_name = "conv2d_65.w_0" shape = [512, 512, 3, 3] dtype = "float32" min_val = float("-0.0796277") @@ -240,6 +262,7 @@ class Program_weight_tensor_parameter_21: class Program_weight_tensor_parameter_22: name = "parameter_22" + original_name = "group_norm_2.b_0" shape = [512] dtype = "float32" min_val = float("-0.140032") @@ -251,6 +274,7 @@ class Program_weight_tensor_parameter_22: class Program_weight_tensor_parameter_23: name = "parameter_23" + original_name = "group_norm_2.w_0" shape = [512] dtype = "float32" min_val = float("0.447633") @@ -262,6 +286,7 @@ class Program_weight_tensor_parameter_23: class Program_weight_tensor_parameter_24: name = "parameter_24" + original_name = "conv2d_63.w_0" shape = [512, 512, 3, 3] dtype = "float32" min_val = float("-0.0713307") @@ -273,6 +298,7 @@ class Program_weight_tensor_parameter_24: class Program_weight_tensor_parameter_25: name = "parameter_25" + original_name = "group_norm_0.b_0" shape = [512] dtype = "float32" min_val = float("-0.115167") @@ -284,6 +310,7 @@ class Program_weight_tensor_parameter_25: class Program_weight_tensor_parameter_26: name = "parameter_26" + original_name = "group_norm_0.w_0" shape = [512] dtype = "float32" min_val = float("0.889679") @@ -295,6 +322,7 @@ class Program_weight_tensor_parameter_26: class Program_weight_tensor_parameter_27: name = "parameter_27" + original_name = "conv2d_61.w_0" shape = [512, 258, 3, 3] dtype = "float32" min_val = float("-0.176995") @@ -306,6 +334,7 @@ class Program_weight_tensor_parameter_27: class Program_weight_tensor_parameter_28: name = "parameter_28" + original_name = "group_norm_15.b_0" shape = [256] dtype = "float32" min_val = float("-0.260207") @@ -317,6 +346,7 @@ class Program_weight_tensor_parameter_28: class Program_weight_tensor_parameter_29: name = "parameter_29" + original_name = "group_norm_15.w_0" shape = [256] dtype = "float32" min_val = float("0.362334") @@ -328,6 +358,7 @@ class Program_weight_tensor_parameter_29: class Program_weight_tensor_parameter_30: name = "parameter_30" + original_name = "conv2d_78.w_0" shape = [256, 128, 1, 1] dtype = "float32" min_val = float("-0.235548") @@ -339,6 +370,7 @@ class Program_weight_tensor_parameter_30: class Program_weight_tensor_parameter_31: name = "parameter_31" + original_name = "group_norm_14.b_0" shape = [128] dtype = "float32" min_val = float("-0.118527") @@ -350,6 +382,7 @@ class Program_weight_tensor_parameter_31: class Program_weight_tensor_parameter_32: name = "parameter_32" + original_name = "group_norm_14.w_0" shape = [128] dtype = "float32" min_val = float("0.835934") @@ -361,6 +394,7 @@ class Program_weight_tensor_parameter_32: class Program_weight_tensor_parameter_33: name = "parameter_33" + original_name = "conv2d_77.w_0" shape = [128, 128, 3, 3] dtype = "float32" min_val = float("-0.0902691") @@ -372,6 +406,7 @@ class Program_weight_tensor_parameter_33: class Program_weight_tensor_parameter_34: name = "parameter_34" + original_name = "group_norm_13.b_0" shape = [128] dtype = "float32" min_val = float("-0.118716") @@ -383,6 +418,7 @@ class Program_weight_tensor_parameter_34: class Program_weight_tensor_parameter_35: name = "parameter_35" + original_name = "group_norm_13.w_0" shape = [128] dtype = "float32" min_val = float("0.807064") @@ -394,6 +430,7 @@ class Program_weight_tensor_parameter_35: class Program_weight_tensor_parameter_36: name = "parameter_36" + original_name = "conv2d_76.w_0" shape = [128, 128, 3, 3] dtype = "float32" min_val = float("-0.0921701") @@ -405,6 +442,7 @@ class Program_weight_tensor_parameter_36: class Program_weight_tensor_parameter_37: name = "parameter_37" + original_name = "group_norm_12.b_0" shape = [128] dtype = "float32" min_val = float("-0.18393") @@ -416,6 +454,7 @@ class Program_weight_tensor_parameter_37: class Program_weight_tensor_parameter_38: name = "parameter_38" + original_name = "group_norm_12.w_0" shape = [128] dtype = "float32" min_val = float("0.819504") @@ -427,6 +466,7 @@ class Program_weight_tensor_parameter_38: class Program_weight_tensor_parameter_39: name = "parameter_39" + original_name = "conv2d_75.w_0" shape = [128, 258, 3, 3] dtype = "float32" min_val = float("-0.335587") @@ -438,6 +478,7 @@ class Program_weight_tensor_parameter_39: class Program_weight_tensor_parameter_40: name = "parameter_40" + original_name = "group_norm_11.b_0" shape = [128] dtype = "float32" min_val = float("-0.129893") @@ -449,6 +490,7 @@ class Program_weight_tensor_parameter_40: class Program_weight_tensor_parameter_41: name = "parameter_41" + original_name = "group_norm_11.w_0" shape = [128] dtype = "float32" min_val = float("0.532926") @@ -460,6 +502,7 @@ class Program_weight_tensor_parameter_41: class Program_weight_tensor_parameter_42: name = "parameter_42" + original_name = "conv2d_74.w_0" shape = [128, 128, 3, 3] dtype = "float32" min_val = float("-0.0707103") @@ -471,6 +514,7 @@ class Program_weight_tensor_parameter_42: class Program_weight_tensor_parameter_43: name = "parameter_43" + original_name = "group_norm_10.b_0" shape = [128] dtype = "float32" min_val = float("-0.09812") @@ -482,6 +526,7 @@ class Program_weight_tensor_parameter_43: class Program_weight_tensor_parameter_44: name = "parameter_44" + original_name = "group_norm_10.w_0" shape = [128] dtype = "float32" min_val = float("0.796179") @@ -493,6 +538,7 @@ class Program_weight_tensor_parameter_44: class Program_weight_tensor_parameter_45: name = "parameter_45" + original_name = "conv2d_73.w_0" shape = [128, 256, 3, 3] dtype = "float32" min_val = float("-0.043773") @@ -504,6 +550,7 @@ class Program_weight_tensor_parameter_45: class Program_weight_tensor_parameter_46: name = "parameter_46" + original_name = "group_norm_9.b_0" shape = [128] dtype = "float32" min_val = float("-0.128115") @@ -515,6 +562,7 @@ class Program_weight_tensor_parameter_46: class Program_weight_tensor_parameter_47: name = "parameter_47" + original_name = "group_norm_9.w_0" shape = [128] dtype = "float32" min_val = float("0.533942") @@ -526,6 +574,7 @@ class Program_weight_tensor_parameter_47: class Program_weight_tensor_parameter_48: name = "parameter_48" + original_name = "conv2d_72.w_0" shape = [128, 256, 3, 3] dtype = "float32" min_val = float("-0.0504074") @@ -537,6 +586,7 @@ class Program_weight_tensor_parameter_48: class Program_weight_tensor_parameter_49: name = "parameter_49" + original_name = "group_norm_8.b_0" shape = [128] dtype = "float32" min_val = float("-0.135607") @@ -548,6 +598,7 @@ class Program_weight_tensor_parameter_49: class Program_weight_tensor_parameter_50: name = "parameter_50" + original_name = "group_norm_8.w_0" shape = [128] dtype = "float32" min_val = float("0.885637") @@ -559,6 +610,7 @@ class Program_weight_tensor_parameter_50: class Program_weight_tensor_parameter_51: name = "parameter_51" + original_name = "conv2d_71.w_0" shape = [128, 256, 3, 3] dtype = "float32" min_val = float("-0.0518155") @@ -570,6 +622,7 @@ class Program_weight_tensor_parameter_51: class Program_weight_tensor_parameter_52: name = "parameter_52" + original_name = "conv2d_60.b_0" shape = [256] dtype = "float32" min_val = float("-0.141143") @@ -581,6 +634,7 @@ class Program_weight_tensor_parameter_52: class Program_weight_tensor_parameter_53: name = "parameter_53" + original_name = "conv2d_60.w_0" shape = [256, 256, 3, 3] dtype = "float32" min_val = float("-0.0449099") @@ -592,6 +646,7 @@ class Program_weight_tensor_parameter_53: class Program_weight_tensor_parameter_54: name = "parameter_54" + original_name = "conv2d_58.b_0" shape = [256] dtype = "float32" min_val = float("-0.0684464") @@ -603,6 +658,7 @@ class Program_weight_tensor_parameter_54: class Program_weight_tensor_parameter_55: name = "parameter_55" + original_name = "conv2d_58.w_0" shape = [256, 256, 3, 3] dtype = "float32" min_val = float("-0.0372319") @@ -614,6 +670,7 @@ class Program_weight_tensor_parameter_55: class Program_weight_tensor_parameter_56: name = "parameter_56" + original_name = "conv2d_56.b_0" shape = [256] dtype = "float32" min_val = float("-0.0926705") @@ -625,6 +682,7 @@ class Program_weight_tensor_parameter_56: class Program_weight_tensor_parameter_57: name = "parameter_57" + original_name = "conv2d_56.w_0" shape = [256, 256, 3, 3] dtype = "float32" min_val = float("-0.0363118") @@ -636,6 +694,7 @@ class Program_weight_tensor_parameter_57: class Program_weight_tensor_parameter_58: name = "parameter_58" + original_name = "conv2d_54.b_0" shape = [256] dtype = "float32" min_val = float("-0.0809814") @@ -647,6 +706,7 @@ class Program_weight_tensor_parameter_58: class Program_weight_tensor_parameter_59: name = "parameter_59" + original_name = "conv2d_54.w_0" shape = [256, 256, 3, 3] dtype = "float32" min_val = float("-0.0385208") @@ -658,6 +718,7 @@ class Program_weight_tensor_parameter_59: class Program_weight_tensor_parameter_60: name = "parameter_60" + original_name = "conv2d_59.b_0" shape = [256] dtype = "float32" min_val = float("-0.120511") @@ -669,6 +730,7 @@ class Program_weight_tensor_parameter_60: class Program_weight_tensor_parameter_61: name = "parameter_61" + original_name = "conv2d_59.w_0" shape = [256, 2048, 1, 1] dtype = "float32" min_val = float("-0.0718854") @@ -680,6 +742,7 @@ class Program_weight_tensor_parameter_61: class Program_weight_tensor_parameter_62: name = "parameter_62" + original_name = "conv2d_57.b_0" shape = [256] dtype = "float32" min_val = float("-0.0679209") @@ -691,6 +754,7 @@ class Program_weight_tensor_parameter_62: class Program_weight_tensor_parameter_63: name = "parameter_63" + original_name = "conv2d_57.w_0" shape = [256, 1024, 1, 1] dtype = "float32" min_val = float("-0.11025") @@ -702,6 +766,7 @@ class Program_weight_tensor_parameter_63: class Program_weight_tensor_parameter_64: name = "parameter_64" + original_name = "conv2d_55.b_0" shape = [256] dtype = "float32" min_val = float("-0.0655969") @@ -713,6 +778,7 @@ class Program_weight_tensor_parameter_64: class Program_weight_tensor_parameter_65: name = "parameter_65" + original_name = "conv2d_55.w_0" shape = [256, 512, 1, 1] dtype = "float32" min_val = float("-0.0782085") @@ -724,6 +790,7 @@ class Program_weight_tensor_parameter_65: class Program_weight_tensor_parameter_66: name = "parameter_66" + original_name = "conv2d_53.b_0" shape = [256] dtype = "float32" min_val = float("-0.0731329") @@ -735,6 +802,7 @@ class Program_weight_tensor_parameter_66: class Program_weight_tensor_parameter_67: name = "parameter_67" + original_name = "conv2d_53.w_0" shape = [256, 256, 1, 1] dtype = "float32" min_val = float("-0.0819963") @@ -746,6 +814,7 @@ class Program_weight_tensor_parameter_67: class Program_weight_tensor_parameter_68: name = "parameter_68" + original_name = "batch_norm2d_52.b_0" shape = [2048] dtype = "float32" min_val = float("-0.0151986") @@ -757,6 +826,7 @@ class Program_weight_tensor_parameter_68: class Program_weight_tensor_parameter_69: name = "parameter_69" + original_name = "batch_norm2d_52.w_0" shape = [2048] dtype = "float32" min_val = float("0.0697239") @@ -768,6 +838,7 @@ class Program_weight_tensor_parameter_69: class Program_weight_tensor_parameter_70: name = "parameter_70" + original_name = "batch_norm2d_52.w_2" shape = [2048] dtype = "float32" min_val = float("1.81634e-05") @@ -779,6 +850,7 @@ class Program_weight_tensor_parameter_70: class Program_weight_tensor_parameter_71: name = "parameter_71" + original_name = "batch_norm2d_52.w_1" shape = [2048] dtype = "float32" min_val = float("-0.0611881") @@ -790,6 +862,7 @@ class Program_weight_tensor_parameter_71: class Program_weight_tensor_parameter_72: name = "parameter_72" + original_name = "conv2d_52.w_0" shape = [2048, 512, 1, 1] dtype = "float32" min_val = float("-0.175748") @@ -801,6 +874,7 @@ class Program_weight_tensor_parameter_72: class Program_weight_tensor_parameter_73: name = "parameter_73" + original_name = "batch_norm2d_51.b_0" shape = [512] dtype = "float32" min_val = float("-0.256637") @@ -812,6 +886,7 @@ class Program_weight_tensor_parameter_73: class Program_weight_tensor_parameter_74: name = "parameter_74" + original_name = "batch_norm2d_51.w_0" shape = [512] dtype = "float32" min_val = float("0.142703") @@ -823,6 +898,7 @@ class Program_weight_tensor_parameter_74: class Program_weight_tensor_parameter_75: name = "parameter_75" + original_name = "batch_norm2d_51.w_2" shape = [512] dtype = "float32" min_val = float("0.00782461") @@ -834,6 +910,7 @@ class Program_weight_tensor_parameter_75: class Program_weight_tensor_parameter_76: name = "parameter_76" + original_name = "batch_norm2d_51.w_1" shape = [512] dtype = "float32" min_val = float("-0.213289") @@ -845,6 +922,7 @@ class Program_weight_tensor_parameter_76: class Program_weight_tensor_parameter_77: name = "parameter_77" + original_name = "conv2d_51.w_0" shape = [512, 512, 3, 3] dtype = "float32" min_val = float("-0.15478") @@ -856,6 +934,7 @@ class Program_weight_tensor_parameter_77: class Program_weight_tensor_parameter_78: name = "parameter_78" + original_name = "batch_norm2d_50.b_0" shape = [512] dtype = "float32" min_val = float("-0.37403") @@ -867,6 +946,7 @@ class Program_weight_tensor_parameter_78: class Program_weight_tensor_parameter_79: name = "parameter_79" + original_name = "batch_norm2d_50.w_0" shape = [512] dtype = "float32" min_val = float("0.112228") @@ -878,6 +958,7 @@ class Program_weight_tensor_parameter_79: class Program_weight_tensor_parameter_80: name = "parameter_80" + original_name = "batch_norm2d_50.w_2" shape = [512] dtype = "float32" min_val = float("0.00368759") @@ -889,6 +970,7 @@ class Program_weight_tensor_parameter_80: class Program_weight_tensor_parameter_81: name = "parameter_81" + original_name = "batch_norm2d_50.w_1" shape = [512] dtype = "float32" min_val = float("-0.165443") @@ -900,6 +982,7 @@ class Program_weight_tensor_parameter_81: class Program_weight_tensor_parameter_82: name = "parameter_82" + original_name = "conv2d_50.w_0" shape = [512, 2048, 1, 1] dtype = "float32" min_val = float("-0.308366") @@ -911,6 +994,7 @@ class Program_weight_tensor_parameter_82: class Program_weight_tensor_parameter_83: name = "parameter_83" + original_name = "batch_norm2d_49.b_0" shape = [2048] dtype = "float32" min_val = float("-0.344631") @@ -922,6 +1006,7 @@ class Program_weight_tensor_parameter_83: class Program_weight_tensor_parameter_84: name = "parameter_84" + original_name = "batch_norm2d_49.w_0" shape = [2048] dtype = "float32" min_val = float("-0.0556075") @@ -933,6 +1018,7 @@ class Program_weight_tensor_parameter_84: class Program_weight_tensor_parameter_85: name = "parameter_85" + original_name = "batch_norm2d_49.w_2" shape = [2048] dtype = "float32" min_val = float("0.000374117") @@ -944,6 +1030,7 @@ class Program_weight_tensor_parameter_85: class Program_weight_tensor_parameter_86: name = "parameter_86" + original_name = "batch_norm2d_49.w_1" shape = [2048] dtype = "float32" min_val = float("-0.143769") @@ -955,6 +1042,7 @@ class Program_weight_tensor_parameter_86: class Program_weight_tensor_parameter_87: name = "parameter_87" + original_name = "conv2d_49.w_0" shape = [2048, 512, 1, 1] dtype = "float32" min_val = float("-0.262648") @@ -966,6 +1054,7 @@ class Program_weight_tensor_parameter_87: class Program_weight_tensor_parameter_88: name = "parameter_88" + original_name = "batch_norm2d_48.b_0" shape = [512] dtype = "float32" min_val = float("-0.281487") @@ -977,6 +1066,7 @@ class Program_weight_tensor_parameter_88: class Program_weight_tensor_parameter_89: name = "parameter_89" + original_name = "batch_norm2d_48.w_0" shape = [512] dtype = "float32" min_val = float("0.103009") @@ -988,6 +1078,7 @@ class Program_weight_tensor_parameter_89: class Program_weight_tensor_parameter_90: name = "parameter_90" + original_name = "batch_norm2d_48.w_2" shape = [512] dtype = "float32" min_val = float("0.00667127") @@ -999,6 +1090,7 @@ class Program_weight_tensor_parameter_90: class Program_weight_tensor_parameter_91: name = "parameter_91" + original_name = "batch_norm2d_48.w_1" shape = [512] dtype = "float32" min_val = float("-0.333268") @@ -1010,6 +1102,7 @@ class Program_weight_tensor_parameter_91: class Program_weight_tensor_parameter_92: name = "parameter_92" + original_name = "conv2d_48.w_0" shape = [512, 512, 3, 3] dtype = "float32" min_val = float("-0.164359") @@ -1021,6 +1114,7 @@ class Program_weight_tensor_parameter_92: class Program_weight_tensor_parameter_93: name = "parameter_93" + original_name = "batch_norm2d_47.b_0" shape = [512] dtype = "float32" min_val = float("-0.204657") @@ -1032,6 +1126,7 @@ class Program_weight_tensor_parameter_93: class Program_weight_tensor_parameter_94: name = "parameter_94" + original_name = "batch_norm2d_47.w_0" shape = [512] dtype = "float32" min_val = float("0.110153") @@ -1043,6 +1138,7 @@ class Program_weight_tensor_parameter_94: class Program_weight_tensor_parameter_95: name = "parameter_95" + original_name = "batch_norm2d_47.w_2" shape = [512] dtype = "float32" min_val = float("0.00384403") @@ -1054,6 +1150,7 @@ class Program_weight_tensor_parameter_95: class Program_weight_tensor_parameter_96: name = "parameter_96" + original_name = "batch_norm2d_47.w_1" shape = [512] dtype = "float32" min_val = float("-0.156824") @@ -1065,6 +1162,7 @@ class Program_weight_tensor_parameter_96: class Program_weight_tensor_parameter_97: name = "parameter_97" + original_name = "conv2d_47.w_0" shape = [512, 2048, 1, 1] dtype = "float32" min_val = float("-0.199223") @@ -1076,6 +1174,7 @@ class Program_weight_tensor_parameter_97: class Program_weight_tensor_parameter_98: name = "parameter_98" + original_name = "batch_norm2d_46.b_0" shape = [2048] dtype = "float32" min_val = float("-0.175843") @@ -1087,6 +1186,7 @@ class Program_weight_tensor_parameter_98: class Program_weight_tensor_parameter_99: name = "parameter_99" + original_name = "batch_norm2d_46.w_0" shape = [2048] dtype = "float32" min_val = float("0.0445195") @@ -1098,6 +1198,7 @@ class Program_weight_tensor_parameter_99: class Program_weight_tensor_parameter_100: name = "parameter_100" + original_name = "batch_norm2d_46.w_2" shape = [2048] dtype = "float32" min_val = float("0.00110241") @@ -1109,6 +1210,7 @@ class Program_weight_tensor_parameter_100: class Program_weight_tensor_parameter_101: name = "parameter_101" + original_name = "batch_norm2d_46.w_1" shape = [2048] dtype = "float32" min_val = float("-0.149371") @@ -1120,6 +1222,7 @@ class Program_weight_tensor_parameter_101: class Program_weight_tensor_parameter_102: name = "parameter_102" + original_name = "conv2d_46.w_0" shape = [2048, 1024, 1, 1] dtype = "float32" min_val = float("-0.230213") @@ -1131,6 +1234,7 @@ class Program_weight_tensor_parameter_102: class Program_weight_tensor_parameter_103: name = "parameter_103" + original_name = "batch_norm2d_45.b_0" shape = [2048] dtype = "float32" min_val = float("-0.175843") @@ -1142,6 +1246,7 @@ class Program_weight_tensor_parameter_103: class Program_weight_tensor_parameter_104: name = "parameter_104" + original_name = "batch_norm2d_45.w_0" shape = [2048] dtype = "float32" min_val = float("-0.0299624") @@ -1153,6 +1258,7 @@ class Program_weight_tensor_parameter_104: class Program_weight_tensor_parameter_105: name = "parameter_105" + original_name = "batch_norm2d_45.w_2" shape = [2048] dtype = "float32" min_val = float("0.000316897") @@ -1164,6 +1270,7 @@ class Program_weight_tensor_parameter_105: class Program_weight_tensor_parameter_106: name = "parameter_106" + original_name = "batch_norm2d_45.w_1" shape = [2048] dtype = "float32" min_val = float("-0.119339") @@ -1175,6 +1282,7 @@ class Program_weight_tensor_parameter_106: class Program_weight_tensor_parameter_107: name = "parameter_107" + original_name = "conv2d_45.w_0" shape = [2048, 512, 1, 1] dtype = "float32" min_val = float("-0.281965") @@ -1186,6 +1294,7 @@ class Program_weight_tensor_parameter_107: class Program_weight_tensor_parameter_108: name = "parameter_108" + original_name = "batch_norm2d_44.b_0" shape = [512] dtype = "float32" min_val = float("-0.166324") @@ -1197,6 +1306,7 @@ class Program_weight_tensor_parameter_108: class Program_weight_tensor_parameter_109: name = "parameter_109" + original_name = "batch_norm2d_44.w_0" shape = [512] dtype = "float32" min_val = float("0.122111") @@ -1208,6 +1318,7 @@ class Program_weight_tensor_parameter_109: class Program_weight_tensor_parameter_110: name = "parameter_110" + original_name = "batch_norm2d_44.w_2" shape = [512] dtype = "float32" min_val = float("0.00556363") @@ -1219,6 +1330,7 @@ class Program_weight_tensor_parameter_110: class Program_weight_tensor_parameter_111: name = "parameter_111" + original_name = "batch_norm2d_44.w_1" shape = [512] dtype = "float32" min_val = float("-0.114129") @@ -1230,6 +1342,7 @@ class Program_weight_tensor_parameter_111: class Program_weight_tensor_parameter_112: name = "parameter_112" + original_name = "conv2d_44.w_0" shape = [512, 512, 3, 3] dtype = "float32" min_val = float("-0.32344") @@ -1241,6 +1354,7 @@ class Program_weight_tensor_parameter_112: class Program_weight_tensor_parameter_113: name = "parameter_113" + original_name = "batch_norm2d_43.b_0" shape = [512] dtype = "float32" min_val = float("-0.323683") @@ -1252,6 +1366,7 @@ class Program_weight_tensor_parameter_113: class Program_weight_tensor_parameter_114: name = "parameter_114" + original_name = "batch_norm2d_43.w_0" shape = [512] dtype = "float32" min_val = float("0.103589") @@ -1263,6 +1378,7 @@ class Program_weight_tensor_parameter_114: class Program_weight_tensor_parameter_115: name = "parameter_115" + original_name = "batch_norm2d_43.w_2" shape = [512] dtype = "float32" min_val = float("0.00626156") @@ -1274,6 +1390,7 @@ class Program_weight_tensor_parameter_115: class Program_weight_tensor_parameter_116: name = "parameter_116" + original_name = "batch_norm2d_43.w_1" shape = [512] dtype = "float32" min_val = float("-0.154333") @@ -1285,6 +1402,7 @@ class Program_weight_tensor_parameter_116: class Program_weight_tensor_parameter_117: name = "parameter_117" + original_name = "conv2d_43.w_0" shape = [512, 1024, 1, 1] dtype = "float32" min_val = float("-0.276262") @@ -1296,6 +1414,7 @@ class Program_weight_tensor_parameter_117: class Program_weight_tensor_parameter_118: name = "parameter_118" + original_name = "batch_norm2d_42.b_0" shape = [1024] dtype = "float32" min_val = float("-0.350923") @@ -1307,6 +1426,7 @@ class Program_weight_tensor_parameter_118: class Program_weight_tensor_parameter_119: name = "parameter_119" + original_name = "batch_norm2d_42.w_0" shape = [1024] dtype = "float32" min_val = float("-0.0704466") @@ -1318,6 +1438,7 @@ class Program_weight_tensor_parameter_119: class Program_weight_tensor_parameter_120: name = "parameter_120" + original_name = "batch_norm2d_42.w_2" shape = [1024] dtype = "float32" min_val = float("7.2606e-05") @@ -1329,6 +1450,7 @@ class Program_weight_tensor_parameter_120: class Program_weight_tensor_parameter_121: name = "parameter_121" + original_name = "batch_norm2d_42.w_1" shape = [1024] dtype = "float32" min_val = float("-0.14425") @@ -1340,6 +1462,7 @@ class Program_weight_tensor_parameter_121: class Program_weight_tensor_parameter_122: name = "parameter_122" + original_name = "conv2d_42.w_0" shape = [1024, 256, 1, 1] dtype = "float32" min_val = float("-0.531898") @@ -1351,6 +1474,7 @@ class Program_weight_tensor_parameter_122: class Program_weight_tensor_parameter_123: name = "parameter_123" + original_name = "batch_norm2d_41.b_0" shape = [256] dtype = "float32" min_val = float("-0.21119") @@ -1362,6 +1486,7 @@ class Program_weight_tensor_parameter_123: class Program_weight_tensor_parameter_124: name = "parameter_124" + original_name = "batch_norm2d_41.w_0" shape = [256] dtype = "float32" min_val = float("0.11773") @@ -1373,6 +1498,7 @@ class Program_weight_tensor_parameter_124: class Program_weight_tensor_parameter_125: name = "parameter_125" + original_name = "batch_norm2d_41.w_2" shape = [256] dtype = "float32" min_val = float("0.0034165") @@ -1384,6 +1510,7 @@ class Program_weight_tensor_parameter_125: class Program_weight_tensor_parameter_126: name = "parameter_126" + original_name = "batch_norm2d_41.w_1" shape = [256] dtype = "float32" min_val = float("-0.17006") @@ -1395,6 +1522,7 @@ class Program_weight_tensor_parameter_126: class Program_weight_tensor_parameter_127: name = "parameter_127" + original_name = "conv2d_41.w_0" shape = [256, 256, 3, 3] dtype = "float32" min_val = float("-0.237546") @@ -1406,6 +1534,7 @@ class Program_weight_tensor_parameter_127: class Program_weight_tensor_parameter_128: name = "parameter_128" + original_name = "batch_norm2d_40.b_0" shape = [256] dtype = "float32" min_val = float("-0.311719") @@ -1417,6 +1546,7 @@ class Program_weight_tensor_parameter_128: class Program_weight_tensor_parameter_129: name = "parameter_129" + original_name = "batch_norm2d_40.w_0" shape = [256] dtype = "float32" min_val = float("0.0999653") @@ -1428,6 +1558,7 @@ class Program_weight_tensor_parameter_129: class Program_weight_tensor_parameter_130: name = "parameter_130" + original_name = "batch_norm2d_40.w_2" shape = [256] dtype = "float32" min_val = float("0.00541324") @@ -1439,6 +1570,7 @@ class Program_weight_tensor_parameter_130: class Program_weight_tensor_parameter_131: name = "parameter_131" + original_name = "batch_norm2d_40.w_1" shape = [256] dtype = "float32" min_val = float("-0.217744") @@ -1450,6 +1582,7 @@ class Program_weight_tensor_parameter_131: class Program_weight_tensor_parameter_132: name = "parameter_132" + original_name = "conv2d_40.w_0" shape = [256, 1024, 1, 1] dtype = "float32" min_val = float("-0.232663") @@ -1461,6 +1594,7 @@ class Program_weight_tensor_parameter_132: class Program_weight_tensor_parameter_133: name = "parameter_133" + original_name = "batch_norm2d_39.b_0" shape = [1024] dtype = "float32" min_val = float("-0.257451") @@ -1472,6 +1606,7 @@ class Program_weight_tensor_parameter_133: class Program_weight_tensor_parameter_134: name = "parameter_134" + original_name = "batch_norm2d_39.w_0" shape = [1024] dtype = "float32" min_val = float("-0.0934567") @@ -1483,6 +1618,7 @@ class Program_weight_tensor_parameter_134: class Program_weight_tensor_parameter_135: name = "parameter_135" + original_name = "batch_norm2d_39.w_2" shape = [1024] dtype = "float32" min_val = float("3.37401e-05") @@ -1494,6 +1630,7 @@ class Program_weight_tensor_parameter_135: class Program_weight_tensor_parameter_136: name = "parameter_136" + original_name = "batch_norm2d_39.w_1" shape = [1024] dtype = "float32" min_val = float("-0.0803722") @@ -1505,6 +1642,7 @@ class Program_weight_tensor_parameter_136: class Program_weight_tensor_parameter_137: name = "parameter_137" + original_name = "conv2d_39.w_0" shape = [1024, 256, 1, 1] dtype = "float32" min_val = float("-0.227153") @@ -1516,6 +1654,7 @@ class Program_weight_tensor_parameter_137: class Program_weight_tensor_parameter_138: name = "parameter_138" + original_name = "batch_norm2d_38.b_0" shape = [256] dtype = "float32" min_val = float("-0.230738") @@ -1527,6 +1666,7 @@ class Program_weight_tensor_parameter_138: class Program_weight_tensor_parameter_139: name = "parameter_139" + original_name = "batch_norm2d_38.w_0" shape = [256] dtype = "float32" min_val = float("0.0986424") @@ -1538,6 +1678,7 @@ class Program_weight_tensor_parameter_139: class Program_weight_tensor_parameter_140: name = "parameter_140" + original_name = "batch_norm2d_38.w_2" shape = [256] dtype = "float32" min_val = float("0.00330071") @@ -1549,6 +1690,7 @@ class Program_weight_tensor_parameter_140: class Program_weight_tensor_parameter_141: name = "parameter_141" + original_name = "batch_norm2d_38.w_1" shape = [256] dtype = "float32" min_val = float("-0.162212") @@ -1560,6 +1702,7 @@ class Program_weight_tensor_parameter_141: class Program_weight_tensor_parameter_142: name = "parameter_142" + original_name = "conv2d_38.w_0" shape = [256, 256, 3, 3] dtype = "float32" min_val = float("-0.151859") @@ -1571,6 +1714,7 @@ class Program_weight_tensor_parameter_142: class Program_weight_tensor_parameter_143: name = "parameter_143" + original_name = "batch_norm2d_37.b_0" shape = [256] dtype = "float32" min_val = float("-0.30878") @@ -1582,6 +1726,7 @@ class Program_weight_tensor_parameter_143: class Program_weight_tensor_parameter_144: name = "parameter_144" + original_name = "batch_norm2d_37.w_0" shape = [256] dtype = "float32" min_val = float("0.091752") @@ -1593,6 +1738,7 @@ class Program_weight_tensor_parameter_144: class Program_weight_tensor_parameter_145: name = "parameter_145" + original_name = "batch_norm2d_37.w_2" shape = [256] dtype = "float32" min_val = float("0.00507588") @@ -1604,6 +1750,7 @@ class Program_weight_tensor_parameter_145: class Program_weight_tensor_parameter_146: name = "parameter_146" + original_name = "batch_norm2d_37.w_1" shape = [256] dtype = "float32" min_val = float("-0.253355") @@ -1615,6 +1762,7 @@ class Program_weight_tensor_parameter_146: class Program_weight_tensor_parameter_147: name = "parameter_147" + original_name = "conv2d_37.w_0" shape = [256, 1024, 1, 1] dtype = "float32" min_val = float("-0.181888") @@ -1626,6 +1774,7 @@ class Program_weight_tensor_parameter_147: class Program_weight_tensor_parameter_148: name = "parameter_148" + original_name = "batch_norm2d_36.b_0" shape = [1024] dtype = "float32" min_val = float("-0.277434") @@ -1637,6 +1786,7 @@ class Program_weight_tensor_parameter_148: class Program_weight_tensor_parameter_149: name = "parameter_149" + original_name = "batch_norm2d_36.w_0" shape = [1024] dtype = "float32" min_val = float("-0.0669331") @@ -1648,6 +1798,7 @@ class Program_weight_tensor_parameter_149: class Program_weight_tensor_parameter_150: name = "parameter_150" + original_name = "batch_norm2d_36.w_2" shape = [1024] dtype = "float32" min_val = float("3.75198e-05") @@ -1659,6 +1810,7 @@ class Program_weight_tensor_parameter_150: class Program_weight_tensor_parameter_151: name = "parameter_151" + original_name = "batch_norm2d_36.w_1" shape = [1024] dtype = "float32" min_val = float("-0.0982832") @@ -1670,6 +1822,7 @@ class Program_weight_tensor_parameter_151: class Program_weight_tensor_parameter_152: name = "parameter_152" + original_name = "conv2d_36.w_0" shape = [1024, 256, 1, 1] dtype = "float32" min_val = float("-0.216702") @@ -1681,6 +1834,7 @@ class Program_weight_tensor_parameter_152: class Program_weight_tensor_parameter_153: name = "parameter_153" + original_name = "batch_norm2d_35.b_0" shape = [256] dtype = "float32" min_val = float("-0.296816") @@ -1692,6 +1846,7 @@ class Program_weight_tensor_parameter_153: class Program_weight_tensor_parameter_154: name = "parameter_154" + original_name = "batch_norm2d_35.w_0" shape = [256] dtype = "float32" min_val = float("0.107435") @@ -1703,6 +1858,7 @@ class Program_weight_tensor_parameter_154: class Program_weight_tensor_parameter_155: name = "parameter_155" + original_name = "batch_norm2d_35.w_2" shape = [256] dtype = "float32" min_val = float("0.00385646") @@ -1714,6 +1870,7 @@ class Program_weight_tensor_parameter_155: class Program_weight_tensor_parameter_156: name = "parameter_156" + original_name = "batch_norm2d_35.w_1" shape = [256] dtype = "float32" min_val = float("-0.169503") @@ -1725,6 +1882,7 @@ class Program_weight_tensor_parameter_156: class Program_weight_tensor_parameter_157: name = "parameter_157" + original_name = "conv2d_35.w_0" shape = [256, 256, 3, 3] dtype = "float32" min_val = float("-0.156512") @@ -1736,6 +1894,7 @@ class Program_weight_tensor_parameter_157: class Program_weight_tensor_parameter_158: name = "parameter_158" + original_name = "batch_norm2d_34.b_0" shape = [256] dtype = "float32" min_val = float("-0.213348") @@ -1747,6 +1906,7 @@ class Program_weight_tensor_parameter_158: class Program_weight_tensor_parameter_159: name = "parameter_159" + original_name = "batch_norm2d_34.w_0" shape = [256] dtype = "float32" min_val = float("0.0886164") @@ -1758,6 +1918,7 @@ class Program_weight_tensor_parameter_159: class Program_weight_tensor_parameter_160: name = "parameter_160" + original_name = "batch_norm2d_34.w_2" shape = [256] dtype = "float32" min_val = float("0.00512839") @@ -1769,6 +1930,7 @@ class Program_weight_tensor_parameter_160: class Program_weight_tensor_parameter_161: name = "parameter_161" + original_name = "batch_norm2d_34.w_1" shape = [256] dtype = "float32" min_val = float("-0.186114") @@ -1780,6 +1942,7 @@ class Program_weight_tensor_parameter_161: class Program_weight_tensor_parameter_162: name = "parameter_162" + original_name = "conv2d_34.w_0" shape = [256, 1024, 1, 1] dtype = "float32" min_val = float("-0.203743") @@ -1791,6 +1954,7 @@ class Program_weight_tensor_parameter_162: class Program_weight_tensor_parameter_163: name = "parameter_163" + original_name = "batch_norm2d_33.b_0" shape = [1024] dtype = "float32" min_val = float("-0.250531") @@ -1802,6 +1966,7 @@ class Program_weight_tensor_parameter_163: class Program_weight_tensor_parameter_164: name = "parameter_164" + original_name = "batch_norm2d_33.w_0" shape = [1024] dtype = "float32" min_val = float("-0.0906585") @@ -1813,6 +1978,7 @@ class Program_weight_tensor_parameter_164: class Program_weight_tensor_parameter_165: name = "parameter_165" + original_name = "batch_norm2d_33.w_2" shape = [1024] dtype = "float32" min_val = float("5.60519e-45") @@ -1824,6 +1990,7 @@ class Program_weight_tensor_parameter_165: class Program_weight_tensor_parameter_166: name = "parameter_166" + original_name = "batch_norm2d_33.w_1" shape = [1024] dtype = "float32" min_val = float("-0.167247") @@ -1835,6 +2002,7 @@ class Program_weight_tensor_parameter_166: class Program_weight_tensor_parameter_167: name = "parameter_167" + original_name = "conv2d_33.w_0" shape = [1024, 256, 1, 1] dtype = "float32" min_val = float("-0.322799") @@ -1846,6 +2014,7 @@ class Program_weight_tensor_parameter_167: class Program_weight_tensor_parameter_168: name = "parameter_168" + original_name = "batch_norm2d_32.b_0" shape = [256] dtype = "float32" min_val = float("-0.35206") @@ -1857,6 +2026,7 @@ class Program_weight_tensor_parameter_168: class Program_weight_tensor_parameter_169: name = "parameter_169" + original_name = "batch_norm2d_32.w_0" shape = [256] dtype = "float32" min_val = float("0.109761") @@ -1868,6 +2038,7 @@ class Program_weight_tensor_parameter_169: class Program_weight_tensor_parameter_170: name = "parameter_170" + original_name = "batch_norm2d_32.w_2" shape = [256] dtype = "float32" min_val = float("0.00447337") @@ -1879,6 +2050,7 @@ class Program_weight_tensor_parameter_170: class Program_weight_tensor_parameter_171: name = "parameter_171" + original_name = "batch_norm2d_32.w_1" shape = [256] dtype = "float32" min_val = float("-0.27309") @@ -1890,6 +2062,7 @@ class Program_weight_tensor_parameter_171: class Program_weight_tensor_parameter_172: name = "parameter_172" + original_name = "conv2d_32.w_0" shape = [256, 256, 3, 3] dtype = "float32" min_val = float("-0.1432") @@ -1901,6 +2074,7 @@ class Program_weight_tensor_parameter_172: class Program_weight_tensor_parameter_173: name = "parameter_173" + original_name = "batch_norm2d_31.b_0" shape = [256] dtype = "float32" min_val = float("-0.244607") @@ -1912,6 +2086,7 @@ class Program_weight_tensor_parameter_173: class Program_weight_tensor_parameter_174: name = "parameter_174" + original_name = "batch_norm2d_31.w_0" shape = [256] dtype = "float32" min_val = float("0.0940088") @@ -1923,6 +2098,7 @@ class Program_weight_tensor_parameter_174: class Program_weight_tensor_parameter_175: name = "parameter_175" + original_name = "batch_norm2d_31.w_2" shape = [256] dtype = "float32" min_val = float("0.00569095") @@ -1934,6 +2110,7 @@ class Program_weight_tensor_parameter_175: class Program_weight_tensor_parameter_176: name = "parameter_176" + original_name = "batch_norm2d_31.w_1" shape = [256] dtype = "float32" min_val = float("-0.216939") @@ -1945,6 +2122,7 @@ class Program_weight_tensor_parameter_176: class Program_weight_tensor_parameter_177: name = "parameter_177" + original_name = "conv2d_31.w_0" shape = [256, 1024, 1, 1] dtype = "float32" min_val = float("-0.194384") @@ -1956,6 +2134,7 @@ class Program_weight_tensor_parameter_177: class Program_weight_tensor_parameter_178: name = "parameter_178" + original_name = "batch_norm2d_30.b_0" shape = [1024] dtype = "float32" min_val = float("-0.28682") @@ -1967,6 +2146,7 @@ class Program_weight_tensor_parameter_178: class Program_weight_tensor_parameter_179: name = "parameter_179" + original_name = "batch_norm2d_30.w_0" shape = [1024] dtype = "float32" min_val = float("-0.095438") @@ -1978,6 +2158,7 @@ class Program_weight_tensor_parameter_179: class Program_weight_tensor_parameter_180: name = "parameter_180" + original_name = "batch_norm2d_30.w_2" shape = [1024] dtype = "float32" min_val = float("5.60519e-45") @@ -1989,6 +2170,7 @@ class Program_weight_tensor_parameter_180: class Program_weight_tensor_parameter_181: name = "parameter_181" + original_name = "batch_norm2d_30.w_1" shape = [1024] dtype = "float32" min_val = float("-0.115913") @@ -2000,6 +2182,7 @@ class Program_weight_tensor_parameter_181: class Program_weight_tensor_parameter_182: name = "parameter_182" + original_name = "conv2d_30.w_0" shape = [1024, 256, 1, 1] dtype = "float32" min_val = float("-0.237178") @@ -2011,6 +2194,7 @@ class Program_weight_tensor_parameter_182: class Program_weight_tensor_parameter_183: name = "parameter_183" + original_name = "batch_norm2d_29.b_0" shape = [256] dtype = "float32" min_val = float("-0.375338") @@ -2022,6 +2206,7 @@ class Program_weight_tensor_parameter_183: class Program_weight_tensor_parameter_184: name = "parameter_184" + original_name = "batch_norm2d_29.w_0" shape = [256] dtype = "float32" min_val = float("0.112331") @@ -2033,6 +2218,7 @@ class Program_weight_tensor_parameter_184: class Program_weight_tensor_parameter_185: name = "parameter_185" + original_name = "batch_norm2d_29.w_2" shape = [256] dtype = "float32" min_val = float("0.00590442") @@ -2044,6 +2230,7 @@ class Program_weight_tensor_parameter_185: class Program_weight_tensor_parameter_186: name = "parameter_186" + original_name = "batch_norm2d_29.w_1" shape = [256] dtype = "float32" min_val = float("-0.703403") @@ -2055,6 +2242,7 @@ class Program_weight_tensor_parameter_186: class Program_weight_tensor_parameter_187: name = "parameter_187" + original_name = "conv2d_29.w_0" shape = [256, 256, 3, 3] dtype = "float32" min_val = float("-0.278966") @@ -2066,6 +2254,7 @@ class Program_weight_tensor_parameter_187: class Program_weight_tensor_parameter_188: name = "parameter_188" + original_name = "batch_norm2d_28.b_0" shape = [256] dtype = "float32" min_val = float("-0.206504") @@ -2077,6 +2266,7 @@ class Program_weight_tensor_parameter_188: class Program_weight_tensor_parameter_189: name = "parameter_189" + original_name = "batch_norm2d_28.w_0" shape = [256] dtype = "float32" min_val = float("0.0505531") @@ -2088,6 +2278,7 @@ class Program_weight_tensor_parameter_189: class Program_weight_tensor_parameter_190: name = "parameter_190" + original_name = "batch_norm2d_28.w_2" shape = [256] dtype = "float32" min_val = float("0.00534171") @@ -2099,6 +2290,7 @@ class Program_weight_tensor_parameter_190: class Program_weight_tensor_parameter_191: name = "parameter_191" + original_name = "batch_norm2d_28.w_1" shape = [256] dtype = "float32" min_val = float("-0.874778") @@ -2110,6 +2302,7 @@ class Program_weight_tensor_parameter_191: class Program_weight_tensor_parameter_192: name = "parameter_192" + original_name = "conv2d_28.w_0" shape = [256, 1024, 1, 1] dtype = "float32" min_val = float("-0.147392") @@ -2121,6 +2314,7 @@ class Program_weight_tensor_parameter_192: class Program_weight_tensor_parameter_193: name = "parameter_193" + original_name = "batch_norm2d_27.b_0" shape = [1024] dtype = "float32" min_val = float("-0.129305") @@ -2132,6 +2326,7 @@ class Program_weight_tensor_parameter_193: class Program_weight_tensor_parameter_194: name = "parameter_194" + original_name = "batch_norm2d_27.w_0" shape = [1024] dtype = "float32" min_val = float("-0.0782405") @@ -2143,6 +2338,7 @@ class Program_weight_tensor_parameter_194: class Program_weight_tensor_parameter_195: name = "parameter_195" + original_name = "batch_norm2d_27.w_2" shape = [1024] dtype = "float32" min_val = float("5.60519e-45") @@ -2154,6 +2350,7 @@ class Program_weight_tensor_parameter_195: class Program_weight_tensor_parameter_196: name = "parameter_196" + original_name = "batch_norm2d_27.w_1" shape = [1024] dtype = "float32" min_val = float("-0.212177") @@ -2165,6 +2362,7 @@ class Program_weight_tensor_parameter_196: class Program_weight_tensor_parameter_197: name = "parameter_197" + original_name = "conv2d_27.w_0" shape = [1024, 512, 1, 1] dtype = "float32" min_val = float("-0.294154") @@ -2176,6 +2374,7 @@ class Program_weight_tensor_parameter_197: class Program_weight_tensor_parameter_198: name = "parameter_198" + original_name = "batch_norm2d_26.b_0" shape = [1024] dtype = "float32" min_val = float("-0.129305") @@ -2187,6 +2386,7 @@ class Program_weight_tensor_parameter_198: class Program_weight_tensor_parameter_199: name = "parameter_199" + original_name = "batch_norm2d_26.w_0" shape = [1024] dtype = "float32" min_val = float("-0.112497") @@ -2198,6 +2398,7 @@ class Program_weight_tensor_parameter_199: class Program_weight_tensor_parameter_200: name = "parameter_200" + original_name = "batch_norm2d_26.w_2" shape = [1024] dtype = "float32" min_val = float("5.60519e-45") @@ -2209,6 +2410,7 @@ class Program_weight_tensor_parameter_200: class Program_weight_tensor_parameter_201: name = "parameter_201" + original_name = "batch_norm2d_26.w_1" shape = [1024] dtype = "float32" min_val = float("-0.274559") @@ -2220,6 +2422,7 @@ class Program_weight_tensor_parameter_201: class Program_weight_tensor_parameter_202: name = "parameter_202" + original_name = "conv2d_26.w_0" shape = [1024, 256, 1, 1] dtype = "float32" min_val = float("-0.270378") @@ -2231,6 +2434,7 @@ class Program_weight_tensor_parameter_202: class Program_weight_tensor_parameter_203: name = "parameter_203" + original_name = "batch_norm2d_25.b_0" shape = [256] dtype = "float32" min_val = float("-0.169522") @@ -2242,6 +2446,7 @@ class Program_weight_tensor_parameter_203: class Program_weight_tensor_parameter_204: name = "parameter_204" + original_name = "batch_norm2d_25.w_0" shape = [256] dtype = "float32" min_val = float("0.134911") @@ -2253,6 +2458,7 @@ class Program_weight_tensor_parameter_204: class Program_weight_tensor_parameter_205: name = "parameter_205" + original_name = "batch_norm2d_25.w_2" shape = [256] dtype = "float32" min_val = float("0.00830091") @@ -2264,6 +2470,7 @@ class Program_weight_tensor_parameter_205: class Program_weight_tensor_parameter_206: name = "parameter_206" + original_name = "batch_norm2d_25.w_1" shape = [256] dtype = "float32" min_val = float("-0.334949") @@ -2275,6 +2482,7 @@ class Program_weight_tensor_parameter_206: class Program_weight_tensor_parameter_207: name = "parameter_207" + original_name = "conv2d_25.w_0" shape = [256, 256, 3, 3] dtype = "float32" min_val = float("-0.161937") @@ -2286,6 +2494,7 @@ class Program_weight_tensor_parameter_207: class Program_weight_tensor_parameter_208: name = "parameter_208" + original_name = "batch_norm2d_24.b_0" shape = [256] dtype = "float32" min_val = float("-0.37354") @@ -2297,6 +2506,7 @@ class Program_weight_tensor_parameter_208: class Program_weight_tensor_parameter_209: name = "parameter_209" + original_name = "batch_norm2d_24.w_0" shape = [256] dtype = "float32" min_val = float("0.131707") @@ -2308,6 +2518,7 @@ class Program_weight_tensor_parameter_209: class Program_weight_tensor_parameter_210: name = "parameter_210" + original_name = "batch_norm2d_24.w_2" shape = [256] dtype = "float32" min_val = float("0.00841896") @@ -2319,6 +2530,7 @@ class Program_weight_tensor_parameter_210: class Program_weight_tensor_parameter_211: name = "parameter_211" + original_name = "batch_norm2d_24.w_1" shape = [256] dtype = "float32" min_val = float("-0.320503") @@ -2330,6 +2542,7 @@ class Program_weight_tensor_parameter_211: class Program_weight_tensor_parameter_212: name = "parameter_212" + original_name = "conv2d_24.w_0" shape = [256, 512, 1, 1] dtype = "float32" min_val = float("-0.255573") @@ -2341,6 +2554,7 @@ class Program_weight_tensor_parameter_212: class Program_weight_tensor_parameter_213: name = "parameter_213" + original_name = "batch_norm2d_23.b_0" shape = [512] dtype = "float32" min_val = float("-0.265757") @@ -2352,6 +2566,7 @@ class Program_weight_tensor_parameter_213: class Program_weight_tensor_parameter_214: name = "parameter_214" + original_name = "batch_norm2d_23.w_0" shape = [512] dtype = "float32" min_val = float("-0.118525") @@ -2363,6 +2578,7 @@ class Program_weight_tensor_parameter_214: class Program_weight_tensor_parameter_215: name = "parameter_215" + original_name = "batch_norm2d_23.w_2" shape = [512] dtype = "float32" min_val = float("6.90463e-05") @@ -2374,6 +2590,7 @@ class Program_weight_tensor_parameter_215: class Program_weight_tensor_parameter_216: name = "parameter_216" + original_name = "batch_norm2d_23.w_1" shape = [512] dtype = "float32" min_val = float("-0.118861") @@ -2385,6 +2602,7 @@ class Program_weight_tensor_parameter_216: class Program_weight_tensor_parameter_217: name = "parameter_217" + original_name = "conv2d_23.w_0" shape = [512, 128, 1, 1] dtype = "float32" min_val = float("-0.176336") @@ -2396,6 +2614,7 @@ class Program_weight_tensor_parameter_217: class Program_weight_tensor_parameter_218: name = "parameter_218" + original_name = "batch_norm2d_22.b_0" shape = [128] dtype = "float32" min_val = float("-0.232824") @@ -2407,6 +2626,7 @@ class Program_weight_tensor_parameter_218: class Program_weight_tensor_parameter_219: name = "parameter_219" + original_name = "batch_norm2d_22.w_0" shape = [128] dtype = "float32" min_val = float("0.118575") @@ -2418,6 +2638,7 @@ class Program_weight_tensor_parameter_219: class Program_weight_tensor_parameter_220: name = "parameter_220" + original_name = "batch_norm2d_22.w_2" shape = [128] dtype = "float32" min_val = float("0.00369018") @@ -2429,6 +2650,7 @@ class Program_weight_tensor_parameter_220: class Program_weight_tensor_parameter_221: name = "parameter_221" + original_name = "batch_norm2d_22.w_1" shape = [128] dtype = "float32" min_val = float("-0.258314") @@ -2440,6 +2662,7 @@ class Program_weight_tensor_parameter_221: class Program_weight_tensor_parameter_222: name = "parameter_222" + original_name = "conv2d_22.w_0" shape = [128, 128, 3, 3] dtype = "float32" min_val = float("-0.147398") @@ -2451,6 +2674,7 @@ class Program_weight_tensor_parameter_222: class Program_weight_tensor_parameter_223: name = "parameter_223" + original_name = "batch_norm2d_21.b_0" shape = [128] dtype = "float32" min_val = float("-0.217196") @@ -2462,6 +2686,7 @@ class Program_weight_tensor_parameter_223: class Program_weight_tensor_parameter_224: name = "parameter_224" + original_name = "batch_norm2d_21.w_0" shape = [128] dtype = "float32" min_val = float("0.0919349") @@ -2473,6 +2698,7 @@ class Program_weight_tensor_parameter_224: class Program_weight_tensor_parameter_225: name = "parameter_225" + original_name = "batch_norm2d_21.w_2" shape = [128] dtype = "float32" min_val = float("0.00814274") @@ -2484,6 +2710,7 @@ class Program_weight_tensor_parameter_225: class Program_weight_tensor_parameter_226: name = "parameter_226" + original_name = "batch_norm2d_21.w_1" shape = [128] dtype = "float32" min_val = float("-0.257526") @@ -2495,6 +2722,7 @@ class Program_weight_tensor_parameter_226: class Program_weight_tensor_parameter_227: name = "parameter_227" + original_name = "conv2d_21.w_0" shape = [128, 512, 1, 1] dtype = "float32" min_val = float("-0.170314") @@ -2506,6 +2734,7 @@ class Program_weight_tensor_parameter_227: class Program_weight_tensor_parameter_228: name = "parameter_228" + original_name = "batch_norm2d_20.b_0" shape = [512] dtype = "float32" min_val = float("-0.237422") @@ -2517,6 +2746,7 @@ class Program_weight_tensor_parameter_228: class Program_weight_tensor_parameter_229: name = "parameter_229" + original_name = "batch_norm2d_20.w_0" shape = [512] dtype = "float32" min_val = float("-0.139337") @@ -2528,6 +2758,7 @@ class Program_weight_tensor_parameter_229: class Program_weight_tensor_parameter_230: name = "parameter_230" + original_name = "batch_norm2d_20.w_2" shape = [512] dtype = "float32" min_val = float("5.60519e-45") @@ -2539,6 +2770,7 @@ class Program_weight_tensor_parameter_230: class Program_weight_tensor_parameter_231: name = "parameter_231" + original_name = "batch_norm2d_20.w_1" shape = [512] dtype = "float32" min_val = float("-0.117879") @@ -2550,6 +2782,7 @@ class Program_weight_tensor_parameter_231: class Program_weight_tensor_parameter_232: name = "parameter_232" + original_name = "conv2d_20.w_0" shape = [512, 128, 1, 1] dtype = "float32" min_val = float("-0.230988") @@ -2561,6 +2794,7 @@ class Program_weight_tensor_parameter_232: class Program_weight_tensor_parameter_233: name = "parameter_233" + original_name = "batch_norm2d_19.b_0" shape = [128] dtype = "float32" min_val = float("-0.242849") @@ -2572,6 +2806,7 @@ class Program_weight_tensor_parameter_233: class Program_weight_tensor_parameter_234: name = "parameter_234" + original_name = "batch_norm2d_19.w_0" shape = [128] dtype = "float32" min_val = float("0.105254") @@ -2583,6 +2818,7 @@ class Program_weight_tensor_parameter_234: class Program_weight_tensor_parameter_235: name = "parameter_235" + original_name = "batch_norm2d_19.w_2" shape = [128] dtype = "float32" min_val = float("0.00364882") @@ -2594,6 +2830,7 @@ class Program_weight_tensor_parameter_235: class Program_weight_tensor_parameter_236: name = "parameter_236" + original_name = "batch_norm2d_19.w_1" shape = [128] dtype = "float32" min_val = float("-0.254192") @@ -2605,6 +2842,7 @@ class Program_weight_tensor_parameter_236: class Program_weight_tensor_parameter_237: name = "parameter_237" + original_name = "conv2d_19.w_0" shape = [128, 128, 3, 3] dtype = "float32" min_val = float("-0.171305") @@ -2616,6 +2854,7 @@ class Program_weight_tensor_parameter_237: class Program_weight_tensor_parameter_238: name = "parameter_238" + original_name = "batch_norm2d_18.b_0" shape = [128] dtype = "float32" min_val = float("-0.193408") @@ -2627,6 +2866,7 @@ class Program_weight_tensor_parameter_238: class Program_weight_tensor_parameter_239: name = "parameter_239" + original_name = "batch_norm2d_18.w_0" shape = [128] dtype = "float32" min_val = float("0.105105") @@ -2638,6 +2878,7 @@ class Program_weight_tensor_parameter_239: class Program_weight_tensor_parameter_240: name = "parameter_240" + original_name = "batch_norm2d_18.w_2" shape = [128] dtype = "float32" min_val = float("0.00566029") @@ -2649,6 +2890,7 @@ class Program_weight_tensor_parameter_240: class Program_weight_tensor_parameter_241: name = "parameter_241" + original_name = "batch_norm2d_18.w_1" shape = [128] dtype = "float32" min_val = float("-0.218906") @@ -2660,6 +2902,7 @@ class Program_weight_tensor_parameter_241: class Program_weight_tensor_parameter_242: name = "parameter_242" + original_name = "conv2d_18.w_0" shape = [128, 512, 1, 1] dtype = "float32" min_val = float("-0.159404") @@ -2671,6 +2914,7 @@ class Program_weight_tensor_parameter_242: class Program_weight_tensor_parameter_243: name = "parameter_243" + original_name = "batch_norm2d_17.b_0" shape = [512] dtype = "float32" min_val = float("-0.24628") @@ -2682,6 +2926,7 @@ class Program_weight_tensor_parameter_243: class Program_weight_tensor_parameter_244: name = "parameter_244" + original_name = "batch_norm2d_17.w_0" shape = [512] dtype = "float32" min_val = float("-0.192908") @@ -2693,6 +2938,7 @@ class Program_weight_tensor_parameter_244: class Program_weight_tensor_parameter_245: name = "parameter_245" + original_name = "batch_norm2d_17.w_2" shape = [512] dtype = "float32" min_val = float("5.60519e-45") @@ -2704,6 +2950,7 @@ class Program_weight_tensor_parameter_245: class Program_weight_tensor_parameter_246: name = "parameter_246" + original_name = "batch_norm2d_17.w_1" shape = [512] dtype = "float32" min_val = float("-0.127209") @@ -2715,6 +2962,7 @@ class Program_weight_tensor_parameter_246: class Program_weight_tensor_parameter_247: name = "parameter_247" + original_name = "conv2d_17.w_0" shape = [512, 128, 1, 1] dtype = "float32" min_val = float("-0.322991") @@ -2726,6 +2974,7 @@ class Program_weight_tensor_parameter_247: class Program_weight_tensor_parameter_248: name = "parameter_248" + original_name = "batch_norm2d_16.b_0" shape = [128] dtype = "float32" min_val = float("-0.361119") @@ -2737,6 +2986,7 @@ class Program_weight_tensor_parameter_248: class Program_weight_tensor_parameter_249: name = "parameter_249" + original_name = "batch_norm2d_16.w_0" shape = [128] dtype = "float32" min_val = float("0.108785") @@ -2748,6 +2998,7 @@ class Program_weight_tensor_parameter_249: class Program_weight_tensor_parameter_250: name = "parameter_250" + original_name = "batch_norm2d_16.w_2" shape = [128] dtype = "float32" min_val = float("0.0020564") @@ -2759,6 +3010,7 @@ class Program_weight_tensor_parameter_250: class Program_weight_tensor_parameter_251: name = "parameter_251" + original_name = "batch_norm2d_16.w_1" shape = [128] dtype = "float32" min_val = float("-0.881857") @@ -2770,6 +3022,7 @@ class Program_weight_tensor_parameter_251: class Program_weight_tensor_parameter_252: name = "parameter_252" + original_name = "conv2d_16.w_0" shape = [128, 128, 3, 3] dtype = "float32" min_val = float("-0.243431") @@ -2781,6 +3034,7 @@ class Program_weight_tensor_parameter_252: class Program_weight_tensor_parameter_253: name = "parameter_253" + original_name = "batch_norm2d_15.b_0" shape = [128] dtype = "float32" min_val = float("-0.338885") @@ -2792,6 +3046,7 @@ class Program_weight_tensor_parameter_253: class Program_weight_tensor_parameter_254: name = "parameter_254" + original_name = "batch_norm2d_15.w_0" shape = [128] dtype = "float32" min_val = float("0.0665754") @@ -2803,6 +3058,7 @@ class Program_weight_tensor_parameter_254: class Program_weight_tensor_parameter_255: name = "parameter_255" + original_name = "batch_norm2d_15.w_2" shape = [128] dtype = "float32" min_val = float("0.00275181") @@ -2814,6 +3070,7 @@ class Program_weight_tensor_parameter_255: class Program_weight_tensor_parameter_256: name = "parameter_256" + original_name = "batch_norm2d_15.w_1" shape = [128] dtype = "float32" min_val = float("-0.271882") @@ -2825,6 +3082,7 @@ class Program_weight_tensor_parameter_256: class Program_weight_tensor_parameter_257: name = "parameter_257" + original_name = "conv2d_15.w_0" shape = [128, 512, 1, 1] dtype = "float32" min_val = float("-0.195609") @@ -2836,6 +3094,7 @@ class Program_weight_tensor_parameter_257: class Program_weight_tensor_parameter_258: name = "parameter_258" + original_name = "batch_norm2d_14.b_0" shape = [512] dtype = "float32" min_val = float("-0.173999") @@ -2847,6 +3106,7 @@ class Program_weight_tensor_parameter_258: class Program_weight_tensor_parameter_259: name = "parameter_259" + original_name = "batch_norm2d_14.w_0" shape = [512] dtype = "float32" min_val = float("-0.0847974") @@ -2858,6 +3118,7 @@ class Program_weight_tensor_parameter_259: class Program_weight_tensor_parameter_260: name = "parameter_260" + original_name = "batch_norm2d_14.w_2" shape = [512] dtype = "float32" min_val = float("5.60519e-45") @@ -2869,6 +3130,7 @@ class Program_weight_tensor_parameter_260: class Program_weight_tensor_parameter_261: name = "parameter_261" + original_name = "batch_norm2d_14.w_1" shape = [512] dtype = "float32" min_val = float("-0.532972") @@ -2880,6 +3142,7 @@ class Program_weight_tensor_parameter_261: class Program_weight_tensor_parameter_262: name = "parameter_262" + original_name = "conv2d_14.w_0" shape = [512, 256, 1, 1] dtype = "float32" min_val = float("-0.41124") @@ -2891,6 +3154,7 @@ class Program_weight_tensor_parameter_262: class Program_weight_tensor_parameter_263: name = "parameter_263" + original_name = "batch_norm2d_13.b_0" shape = [512] dtype = "float32" min_val = float("-0.173999") @@ -2902,6 +3166,7 @@ class Program_weight_tensor_parameter_263: class Program_weight_tensor_parameter_264: name = "parameter_264" + original_name = "batch_norm2d_13.w_0" shape = [512] dtype = "float32" min_val = float("-0.110676") @@ -2913,6 +3178,7 @@ class Program_weight_tensor_parameter_264: class Program_weight_tensor_parameter_265: name = "parameter_265" + original_name = "batch_norm2d_13.w_2" shape = [512] dtype = "float32" min_val = float("5.60519e-45") @@ -2924,6 +3190,7 @@ class Program_weight_tensor_parameter_265: class Program_weight_tensor_parameter_266: name = "parameter_266" + original_name = "batch_norm2d_13.w_1" shape = [512] dtype = "float32" min_val = float("-0.415464") @@ -2935,6 +3202,7 @@ class Program_weight_tensor_parameter_266: class Program_weight_tensor_parameter_267: name = "parameter_267" + original_name = "conv2d_13.w_0" shape = [512, 128, 1, 1] dtype = "float32" min_val = float("-0.307118") @@ -2946,6 +3214,7 @@ class Program_weight_tensor_parameter_267: class Program_weight_tensor_parameter_268: name = "parameter_268" + original_name = "batch_norm2d_12.b_0" shape = [128] dtype = "float32" min_val = float("-0.0749191") @@ -2957,6 +3226,7 @@ class Program_weight_tensor_parameter_268: class Program_weight_tensor_parameter_269: name = "parameter_269" + original_name = "batch_norm2d_12.w_0" shape = [128] dtype = "float32" min_val = float("0.151982") @@ -2968,6 +3238,7 @@ class Program_weight_tensor_parameter_269: class Program_weight_tensor_parameter_270: name = "parameter_270" + original_name = "batch_norm2d_12.w_2" shape = [128] dtype = "float32" min_val = float("0.011571") @@ -2979,6 +3250,7 @@ class Program_weight_tensor_parameter_270: class Program_weight_tensor_parameter_271: name = "parameter_271" + original_name = "batch_norm2d_12.w_1" shape = [128] dtype = "float32" min_val = float("-0.199453") @@ -2990,6 +3262,7 @@ class Program_weight_tensor_parameter_271: class Program_weight_tensor_parameter_272: name = "parameter_272" + original_name = "conv2d_12.w_0" shape = [128, 128, 3, 3] dtype = "float32" min_val = float("-0.15259") @@ -3001,6 +3274,7 @@ class Program_weight_tensor_parameter_272: class Program_weight_tensor_parameter_273: name = "parameter_273" + original_name = "batch_norm2d_11.b_0" shape = [128] dtype = "float32" min_val = float("-0.254357") @@ -3012,6 +3286,7 @@ class Program_weight_tensor_parameter_273: class Program_weight_tensor_parameter_274: name = "parameter_274" + original_name = "batch_norm2d_11.w_0" shape = [128] dtype = "float32" min_val = float("7.24973e-19") @@ -3023,6 +3298,7 @@ class Program_weight_tensor_parameter_274: class Program_weight_tensor_parameter_275: name = "parameter_275" + original_name = "batch_norm2d_11.w_2" shape = [128] dtype = "float32" min_val = float("5.60519e-45") @@ -3034,6 +3310,7 @@ class Program_weight_tensor_parameter_275: class Program_weight_tensor_parameter_276: name = "parameter_276" + original_name = "batch_norm2d_11.w_1" shape = [128] dtype = "float32" min_val = float("-0.268525") @@ -3045,6 +3322,7 @@ class Program_weight_tensor_parameter_276: class Program_weight_tensor_parameter_277: name = "parameter_277" + original_name = "conv2d_11.w_0" shape = [128, 256, 1, 1] dtype = "float32" min_val = float("-0.224228") @@ -3056,6 +3334,7 @@ class Program_weight_tensor_parameter_277: class Program_weight_tensor_parameter_278: name = "parameter_278" + original_name = "batch_norm2d_10.b_0" shape = [256] dtype = "float32" min_val = float("-0.164026") @@ -3067,6 +3346,7 @@ class Program_weight_tensor_parameter_278: class Program_weight_tensor_parameter_279: name = "parameter_279" + original_name = "batch_norm2d_10.w_0" shape = [256] dtype = "float32" min_val = float("-0.255219") @@ -3078,6 +3358,7 @@ class Program_weight_tensor_parameter_279: class Program_weight_tensor_parameter_280: name = "parameter_280" + original_name = "batch_norm2d_10.w_2" shape = [256] dtype = "float32" min_val = float("7.12898e-05") @@ -3089,6 +3370,7 @@ class Program_weight_tensor_parameter_280: class Program_weight_tensor_parameter_281: name = "parameter_281" + original_name = "batch_norm2d_10.w_1" shape = [256] dtype = "float32" min_val = float("-0.133956") @@ -3100,6 +3382,7 @@ class Program_weight_tensor_parameter_281: class Program_weight_tensor_parameter_282: name = "parameter_282" + original_name = "conv2d_10.w_0" shape = [256, 64, 1, 1] dtype = "float32" min_val = float("-0.325233") @@ -3111,6 +3394,7 @@ class Program_weight_tensor_parameter_282: class Program_weight_tensor_parameter_283: name = "parameter_283" + original_name = "batch_norm2d_9.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -3120,6 +3404,7 @@ class Program_weight_tensor_parameter_283: class Program_weight_tensor_parameter_284: name = "parameter_284" + original_name = "batch_norm2d_9.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -3129,6 +3414,7 @@ class Program_weight_tensor_parameter_284: class Program_weight_tensor_parameter_285: name = "parameter_285" + original_name = "batch_norm2d_9.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -3138,6 +3424,7 @@ class Program_weight_tensor_parameter_285: class Program_weight_tensor_parameter_286: name = "parameter_286" + original_name = "batch_norm2d_9.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -3147,6 +3434,7 @@ class Program_weight_tensor_parameter_286: class Program_weight_tensor_parameter_287: name = "parameter_287" + original_name = "conv2d_9.w_0" shape = [64, 64, 3, 3] dtype = "float32" min_val = float("-0.259976") @@ -3158,6 +3446,7 @@ class Program_weight_tensor_parameter_287: class Program_weight_tensor_parameter_288: name = "parameter_288" + original_name = "batch_norm2d_8.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -3167,6 +3456,7 @@ class Program_weight_tensor_parameter_288: class Program_weight_tensor_parameter_289: name = "parameter_289" + original_name = "batch_norm2d_8.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -3176,6 +3466,7 @@ class Program_weight_tensor_parameter_289: class Program_weight_tensor_parameter_290: name = "parameter_290" + original_name = "batch_norm2d_8.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -3185,6 +3476,7 @@ class Program_weight_tensor_parameter_290: class Program_weight_tensor_parameter_291: name = "parameter_291" + original_name = "batch_norm2d_8.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -3194,6 +3486,7 @@ class Program_weight_tensor_parameter_291: class Program_weight_tensor_parameter_292: name = "parameter_292" + original_name = "conv2d_8.w_0" shape = [64, 256, 1, 1] dtype = "float32" min_val = float("-0.188924") @@ -3205,6 +3498,7 @@ class Program_weight_tensor_parameter_292: class Program_weight_tensor_parameter_293: name = "parameter_293" + original_name = "batch_norm2d_7.b_0" shape = [256] dtype = "float32" min_val = float("-0.177785") @@ -3216,6 +3510,7 @@ class Program_weight_tensor_parameter_293: class Program_weight_tensor_parameter_294: name = "parameter_294" + original_name = "batch_norm2d_7.w_0" shape = [256] dtype = "float32" min_val = float("-0.203582") @@ -3227,6 +3522,7 @@ class Program_weight_tensor_parameter_294: class Program_weight_tensor_parameter_295: name = "parameter_295" + original_name = "batch_norm2d_7.w_2" shape = [256] dtype = "float32" min_val = float("5.60519e-45") @@ -3238,6 +3534,7 @@ class Program_weight_tensor_parameter_295: class Program_weight_tensor_parameter_296: name = "parameter_296" + original_name = "batch_norm2d_7.w_1" shape = [256] dtype = "float32" min_val = float("-0.0838559") @@ -3249,6 +3546,7 @@ class Program_weight_tensor_parameter_296: class Program_weight_tensor_parameter_297: name = "parameter_297" + original_name = "conv2d_7.w_0" shape = [256, 64, 1, 1] dtype = "float32" min_val = float("-0.25193") @@ -3260,6 +3558,7 @@ class Program_weight_tensor_parameter_297: class Program_weight_tensor_parameter_298: name = "parameter_298" + original_name = "batch_norm2d_6.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -3269,6 +3568,7 @@ class Program_weight_tensor_parameter_298: class Program_weight_tensor_parameter_299: name = "parameter_299" + original_name = "batch_norm2d_6.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -3278,6 +3578,7 @@ class Program_weight_tensor_parameter_299: class Program_weight_tensor_parameter_300: name = "parameter_300" + original_name = "batch_norm2d_6.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -3287,6 +3588,7 @@ class Program_weight_tensor_parameter_300: class Program_weight_tensor_parameter_301: name = "parameter_301" + original_name = "batch_norm2d_6.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -3296,6 +3598,7 @@ class Program_weight_tensor_parameter_301: class Program_weight_tensor_parameter_302: name = "parameter_302" + original_name = "conv2d_6.w_0" shape = [64, 64, 3, 3] dtype = "float32" min_val = float("-0.277854") @@ -3307,6 +3610,7 @@ class Program_weight_tensor_parameter_302: class Program_weight_tensor_parameter_303: name = "parameter_303" + original_name = "batch_norm2d_5.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -3316,6 +3620,7 @@ class Program_weight_tensor_parameter_303: class Program_weight_tensor_parameter_304: name = "parameter_304" + original_name = "batch_norm2d_5.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -3325,6 +3630,7 @@ class Program_weight_tensor_parameter_304: class Program_weight_tensor_parameter_305: name = "parameter_305" + original_name = "batch_norm2d_5.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -3334,6 +3640,7 @@ class Program_weight_tensor_parameter_305: class Program_weight_tensor_parameter_306: name = "parameter_306" + original_name = "batch_norm2d_5.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -3343,6 +3650,7 @@ class Program_weight_tensor_parameter_306: class Program_weight_tensor_parameter_307: name = "parameter_307" + original_name = "conv2d_5.w_0" shape = [64, 256, 1, 1] dtype = "float32" min_val = float("-0.340224") @@ -3354,6 +3662,7 @@ class Program_weight_tensor_parameter_307: class Program_weight_tensor_parameter_308: name = "parameter_308" + original_name = "batch_norm2d_4.b_0" shape = [256] dtype = "float32" min_val = float("-0.214949") @@ -3365,6 +3674,7 @@ class Program_weight_tensor_parameter_308: class Program_weight_tensor_parameter_309: name = "parameter_309" + original_name = "batch_norm2d_4.w_0" shape = [256] dtype = "float32" min_val = float("-0.140797") @@ -3376,6 +3686,7 @@ class Program_weight_tensor_parameter_309: class Program_weight_tensor_parameter_310: name = "parameter_310" + original_name = "batch_norm2d_4.w_2" shape = [256] dtype = "float32" min_val = float("5.60519e-45") @@ -3387,6 +3698,7 @@ class Program_weight_tensor_parameter_310: class Program_weight_tensor_parameter_311: name = "parameter_311" + original_name = "batch_norm2d_4.w_1" shape = [256] dtype = "float32" min_val = float("-0.788633") @@ -3398,6 +3710,7 @@ class Program_weight_tensor_parameter_311: class Program_weight_tensor_parameter_312: name = "parameter_312" + original_name = "conv2d_4.w_0" shape = [256, 64, 1, 1] dtype = "float32" min_val = float("-0.636505") @@ -3409,6 +3722,7 @@ class Program_weight_tensor_parameter_312: class Program_weight_tensor_parameter_313: name = "parameter_313" + original_name = "batch_norm2d_3.b_0" shape = [256] dtype = "float32" min_val = float("-0.214949") @@ -3420,6 +3734,7 @@ class Program_weight_tensor_parameter_313: class Program_weight_tensor_parameter_314: name = "parameter_314" + original_name = "batch_norm2d_3.w_0" shape = [256] dtype = "float32" min_val = float("-0.351382") @@ -3431,6 +3746,7 @@ class Program_weight_tensor_parameter_314: class Program_weight_tensor_parameter_315: name = "parameter_315" + original_name = "batch_norm2d_3.w_2" shape = [256] dtype = "float32" min_val = float("5.60519e-45") @@ -3442,6 +3758,7 @@ class Program_weight_tensor_parameter_315: class Program_weight_tensor_parameter_316: name = "parameter_316" + original_name = "batch_norm2d_3.w_1" shape = [256] dtype = "float32" min_val = float("-0.221928") @@ -3453,6 +3770,7 @@ class Program_weight_tensor_parameter_316: class Program_weight_tensor_parameter_317: name = "parameter_317" + original_name = "conv2d_3.w_0" shape = [256, 64, 1, 1] dtype = "float32" min_val = float("-0.416398") @@ -3464,6 +3782,7 @@ class Program_weight_tensor_parameter_317: class Program_weight_tensor_parameter_318: name = "parameter_318" + original_name = "batch_norm2d_2.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -3473,6 +3792,7 @@ class Program_weight_tensor_parameter_318: class Program_weight_tensor_parameter_319: name = "parameter_319" + original_name = "batch_norm2d_2.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -3482,6 +3802,7 @@ class Program_weight_tensor_parameter_319: class Program_weight_tensor_parameter_320: name = "parameter_320" + original_name = "batch_norm2d_2.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -3491,6 +3812,7 @@ class Program_weight_tensor_parameter_320: class Program_weight_tensor_parameter_321: name = "parameter_321" + original_name = "batch_norm2d_2.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -3500,6 +3822,7 @@ class Program_weight_tensor_parameter_321: class Program_weight_tensor_parameter_322: name = "parameter_322" + original_name = "conv2d_2.w_0" shape = [64, 64, 3, 3] dtype = "float32" min_val = float("-0.367549") @@ -3511,6 +3834,7 @@ class Program_weight_tensor_parameter_322: class Program_weight_tensor_parameter_323: name = "parameter_323" + original_name = "batch_norm2d_1.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -3520,6 +3844,7 @@ class Program_weight_tensor_parameter_323: class Program_weight_tensor_parameter_324: name = "parameter_324" + original_name = "batch_norm2d_1.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -3529,6 +3854,7 @@ class Program_weight_tensor_parameter_324: class Program_weight_tensor_parameter_325: name = "parameter_325" + original_name = "batch_norm2d_1.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -3538,6 +3864,7 @@ class Program_weight_tensor_parameter_325: class Program_weight_tensor_parameter_326: name = "parameter_326" + original_name = "batch_norm2d_1.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -3547,6 +3874,7 @@ class Program_weight_tensor_parameter_326: class Program_weight_tensor_parameter_327: name = "parameter_327" + original_name = "conv2d_1.w_0" shape = [64, 64, 1, 1] dtype = "float32" min_val = float("-0.628423") @@ -3558,6 +3886,7 @@ class Program_weight_tensor_parameter_327: class Program_weight_tensor_parameter_328: name = "parameter_328" + original_name = "batch_norm2d_0.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -3567,6 +3896,7 @@ class Program_weight_tensor_parameter_328: class Program_weight_tensor_parameter_329: name = "parameter_329" + original_name = "batch_norm2d_0.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -3576,6 +3906,7 @@ class Program_weight_tensor_parameter_329: class Program_weight_tensor_parameter_330: name = "parameter_330" + original_name = "batch_norm2d_0.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -3585,6 +3916,7 @@ class Program_weight_tensor_parameter_330: class Program_weight_tensor_parameter_331: name = "parameter_331" + original_name = "batch_norm2d_0.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -3594,6 +3926,7 @@ class Program_weight_tensor_parameter_331: class Program_weight_tensor_parameter_332: name = "parameter_332" + original_name = "conv2d_0.w_0" shape = [64, 3, 7, 7] dtype = "float32" min_val = float("-0.593645") diff --git a/paddle_samples/PaddleX/SegFormer-B1/subgraph_0/input_meta.py b/paddle_samples/PaddleX/SegFormer-B1/subgraph_0/input_meta.py index 92d9bbc06..44275f73d 100644 --- a/paddle_samples/PaddleX/SegFormer-B1/subgraph_0/input_meta.py +++ b/paddle_samples/PaddleX/SegFormer-B1/subgraph_0/input_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_data_0: name = "data_0" + original_name = "var_0" shape = [2, 3, 512, 1024] dtype = "float32" min_val = float("-1.0") diff --git a/paddle_samples/PaddleX/SegFormer-B1/subgraph_0/weight_meta.py b/paddle_samples/PaddleX/SegFormer-B1/subgraph_0/weight_meta.py index 54ed63bcc..c1940872a 100644 --- a/paddle_samples/PaddleX/SegFormer-B1/subgraph_0/weight_meta.py +++ b/paddle_samples/PaddleX/SegFormer-B1/subgraph_0/weight_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_parameter_0: name = "parameter_0" + original_name = "conv2d_19.b_0" shape = [2] dtype = "float32" min_val = float("0") @@ -9,6 +10,7 @@ class Program_weight_tensor_parameter_0: class Program_weight_tensor_parameter_1: name = "parameter_1" + original_name = "conv2d_19.w_0" shape = [2, 256, 1, 1] dtype = "float32" min_val = float("-0.285621") @@ -20,6 +22,7 @@ class Program_weight_tensor_parameter_1: class Program_weight_tensor_parameter_2: name = "parameter_2" + original_name = "batch_norm2d_0.b_0" shape = [256] dtype = "float32" data = None @@ -27,6 +30,7 @@ class Program_weight_tensor_parameter_2: class Program_weight_tensor_parameter_3: name = "parameter_3" + original_name = "batch_norm2d_0.w_0" shape = [256] dtype = "float32" min_val = float("1.0") @@ -37,6 +41,7 @@ class Program_weight_tensor_parameter_3: class Program_weight_tensor_parameter_4: name = "parameter_4" + original_name = "batch_norm2d_0.w_2" shape = [256] dtype = "float32" min_val = float("1.0") @@ -47,6 +52,7 @@ class Program_weight_tensor_parameter_4: class Program_weight_tensor_parameter_5: name = "parameter_5" + original_name = "batch_norm2d_0.w_1" shape = [256] dtype = "float32" data = None @@ -54,6 +60,7 @@ class Program_weight_tensor_parameter_5: class Program_weight_tensor_parameter_6: name = "parameter_6" + original_name = "conv2d_18.w_0" shape = [256, 1024, 1, 1] dtype = "float32" min_val = float("-0.193104") @@ -65,6 +72,7 @@ class Program_weight_tensor_parameter_6: class Program_weight_tensor_parameter_7: name = "parameter_7" + original_name = "linear_43.b_0" shape = [256] dtype = "float32" data = None @@ -72,6 +80,7 @@ class Program_weight_tensor_parameter_7: class Program_weight_tensor_parameter_8: name = "parameter_8" + original_name = "linear_43.w_0" shape = [64, 256] dtype = "float32" min_val = float("-0.136913") @@ -83,6 +92,7 @@ class Program_weight_tensor_parameter_8: class Program_weight_tensor_parameter_9: name = "parameter_9" + original_name = "linear_42.b_0" shape = [256] dtype = "float32" data = None @@ -90,6 +100,7 @@ class Program_weight_tensor_parameter_9: class Program_weight_tensor_parameter_10: name = "parameter_10" + original_name = "linear_42.w_0" shape = [128, 256] dtype = "float32" min_val = float("-0.124996") @@ -101,6 +112,7 @@ class Program_weight_tensor_parameter_10: class Program_weight_tensor_parameter_11: name = "parameter_11" + original_name = "linear_41.b_0" shape = [256] dtype = "float32" data = None @@ -108,6 +120,7 @@ class Program_weight_tensor_parameter_11: class Program_weight_tensor_parameter_12: name = "parameter_12" + original_name = "linear_41.w_0" shape = [320, 256] dtype = "float32" min_val = float("-0.102061") @@ -119,6 +132,7 @@ class Program_weight_tensor_parameter_12: class Program_weight_tensor_parameter_13: name = "parameter_13" + original_name = "linear_40.b_0" shape = [256] dtype = "float32" data = None @@ -126,6 +140,7 @@ class Program_weight_tensor_parameter_13: class Program_weight_tensor_parameter_14: name = "parameter_14" + original_name = "linear_40.w_0" shape = [512, 256] dtype = "float32" min_val = float("-0.0883872") @@ -137,6 +152,7 @@ class Program_weight_tensor_parameter_14: class Program_weight_tensor_parameter_15: name = "parameter_15" + original_name = "layer_norm_29.b_0" shape = [512] dtype = "float32" min_val = float("-0.640523") @@ -148,6 +164,7 @@ class Program_weight_tensor_parameter_15: class Program_weight_tensor_parameter_16: name = "parameter_16" + original_name = "layer_norm_29.w_0" shape = [512] dtype = "float32" min_val = float("0.0226028") @@ -159,6 +176,7 @@ class Program_weight_tensor_parameter_16: class Program_weight_tensor_parameter_17: name = "parameter_17" + original_name = "linear_39.b_0" shape = [512] dtype = "float32" min_val = float("-8.28303") @@ -170,6 +188,7 @@ class Program_weight_tensor_parameter_17: class Program_weight_tensor_parameter_18: name = "parameter_18" + original_name = "linear_39.w_0" shape = [2048, 512] dtype = "float32" min_val = float("-1.61183") @@ -181,6 +200,7 @@ class Program_weight_tensor_parameter_18: class Program_weight_tensor_parameter_19: name = "parameter_19" + original_name = "conv2d_17.b_0" shape = [2048] dtype = "float32" min_val = float("-1.79191") @@ -192,6 +212,7 @@ class Program_weight_tensor_parameter_19: class Program_weight_tensor_parameter_20: name = "parameter_20" + original_name = "conv2d_17.w_0" shape = [2048, 1, 3, 3] dtype = "float32" min_val = float("-0.441248") @@ -203,6 +224,7 @@ class Program_weight_tensor_parameter_20: class Program_weight_tensor_parameter_21: name = "parameter_21" + original_name = "linear_38.b_0" shape = [2048] dtype = "float32" min_val = float("-3.16905") @@ -214,6 +236,7 @@ class Program_weight_tensor_parameter_21: class Program_weight_tensor_parameter_22: name = "parameter_22" + original_name = "linear_38.w_0" shape = [512, 2048] dtype = "float32" min_val = float("-0.429152") @@ -225,6 +248,7 @@ class Program_weight_tensor_parameter_22: class Program_weight_tensor_parameter_23: name = "parameter_23" + original_name = "layer_norm_28.b_0" shape = [512] dtype = "float32" min_val = float("-2.52545") @@ -236,6 +260,7 @@ class Program_weight_tensor_parameter_23: class Program_weight_tensor_parameter_24: name = "parameter_24" + original_name = "layer_norm_28.w_0" shape = [512] dtype = "float32" min_val = float("0.236685") @@ -247,6 +272,7 @@ class Program_weight_tensor_parameter_24: class Program_weight_tensor_parameter_25: name = "parameter_25" + original_name = "linear_37.b_0" shape = [512] dtype = "float32" min_val = float("-4.65309") @@ -258,6 +284,7 @@ class Program_weight_tensor_parameter_25: class Program_weight_tensor_parameter_26: name = "parameter_26" + original_name = "linear_37.w_0" shape = [512, 512] dtype = "float32" min_val = float("-0.837202") @@ -269,6 +296,7 @@ class Program_weight_tensor_parameter_26: class Program_weight_tensor_parameter_27: name = "parameter_27" + original_name = "linear_36.b_0" shape = [1024] dtype = "float32" min_val = float("-0.965832") @@ -280,6 +308,7 @@ class Program_weight_tensor_parameter_27: class Program_weight_tensor_parameter_28: name = "parameter_28" + original_name = "linear_36.w_0" shape = [512, 1024] dtype = "float32" min_val = float("-0.300903") @@ -291,6 +320,7 @@ class Program_weight_tensor_parameter_28: class Program_weight_tensor_parameter_29: name = "parameter_29" + original_name = "linear_35.b_0" shape = [512] dtype = "float32" min_val = float("-2.9636") @@ -302,6 +332,7 @@ class Program_weight_tensor_parameter_29: class Program_weight_tensor_parameter_30: name = "parameter_30" + original_name = "linear_35.w_0" shape = [512, 512] dtype = "float32" min_val = float("-0.28009") @@ -313,6 +344,7 @@ class Program_weight_tensor_parameter_30: class Program_weight_tensor_parameter_31: name = "parameter_31" + original_name = "layer_norm_27.b_0" shape = [512] dtype = "float32" min_val = float("-1.86386") @@ -324,6 +356,7 @@ class Program_weight_tensor_parameter_31: class Program_weight_tensor_parameter_32: name = "parameter_32" + original_name = "layer_norm_27.w_0" shape = [512] dtype = "float32" min_val = float("0.159688") @@ -335,6 +368,7 @@ class Program_weight_tensor_parameter_32: class Program_weight_tensor_parameter_33: name = "parameter_33" + original_name = "linear_34.b_0" shape = [512] dtype = "float32" min_val = float("-2.45413") @@ -346,6 +380,7 @@ class Program_weight_tensor_parameter_33: class Program_weight_tensor_parameter_34: name = "parameter_34" + original_name = "linear_34.w_0" shape = [2048, 512] dtype = "float32" min_val = float("-0.677935") @@ -357,6 +392,7 @@ class Program_weight_tensor_parameter_34: class Program_weight_tensor_parameter_35: name = "parameter_35" + original_name = "conv2d_16.b_0" shape = [2048] dtype = "float32" min_val = float("-1.60345") @@ -368,6 +404,7 @@ class Program_weight_tensor_parameter_35: class Program_weight_tensor_parameter_36: name = "parameter_36" + original_name = "conv2d_16.w_0" shape = [2048, 1, 3, 3] dtype = "float32" min_val = float("-0.362334") @@ -379,6 +416,7 @@ class Program_weight_tensor_parameter_36: class Program_weight_tensor_parameter_37: name = "parameter_37" + original_name = "linear_33.b_0" shape = [2048] dtype = "float32" min_val = float("-1.99874") @@ -390,6 +428,7 @@ class Program_weight_tensor_parameter_37: class Program_weight_tensor_parameter_38: name = "parameter_38" + original_name = "linear_33.w_0" shape = [512, 2048] dtype = "float32" min_val = float("-0.379622") @@ -401,6 +440,7 @@ class Program_weight_tensor_parameter_38: class Program_weight_tensor_parameter_39: name = "parameter_39" + original_name = "layer_norm_26.b_0" shape = [512] dtype = "float32" min_val = float("-2.88469") @@ -412,6 +452,7 @@ class Program_weight_tensor_parameter_39: class Program_weight_tensor_parameter_40: name = "parameter_40" + original_name = "layer_norm_26.w_0" shape = [512] dtype = "float32" min_val = float("0.880221") @@ -423,6 +464,7 @@ class Program_weight_tensor_parameter_40: class Program_weight_tensor_parameter_41: name = "parameter_41" + original_name = "linear_32.b_0" shape = [512] dtype = "float32" min_val = float("-0.41496") @@ -434,6 +476,7 @@ class Program_weight_tensor_parameter_41: class Program_weight_tensor_parameter_42: name = "parameter_42" + original_name = "linear_32.w_0" shape = [512, 512] dtype = "float32" min_val = float("-0.334796") @@ -445,6 +488,7 @@ class Program_weight_tensor_parameter_42: class Program_weight_tensor_parameter_43: name = "parameter_43" + original_name = "linear_31.b_0" shape = [1024] dtype = "float32" min_val = float("-0.804213") @@ -456,6 +500,7 @@ class Program_weight_tensor_parameter_43: class Program_weight_tensor_parameter_44: name = "parameter_44" + original_name = "linear_31.w_0" shape = [512, 1024] dtype = "float32" min_val = float("-0.360982") @@ -467,6 +512,7 @@ class Program_weight_tensor_parameter_44: class Program_weight_tensor_parameter_45: name = "parameter_45" + original_name = "linear_30.b_0" shape = [512] dtype = "float32" min_val = float("-1.84558") @@ -478,6 +524,7 @@ class Program_weight_tensor_parameter_45: class Program_weight_tensor_parameter_46: name = "parameter_46" + original_name = "linear_30.w_0" shape = [512, 512] dtype = "float32" min_val = float("-0.318339") @@ -489,6 +536,7 @@ class Program_weight_tensor_parameter_46: class Program_weight_tensor_parameter_47: name = "parameter_47" + original_name = "layer_norm_25.b_0" shape = [512] dtype = "float32" min_val = float("-1.38388") @@ -500,6 +548,7 @@ class Program_weight_tensor_parameter_47: class Program_weight_tensor_parameter_48: name = "parameter_48" + original_name = "layer_norm_25.w_0" shape = [512] dtype = "float32" min_val = float("-0.00056076") @@ -511,6 +560,7 @@ class Program_weight_tensor_parameter_48: class Program_weight_tensor_parameter_49: name = "parameter_49" + original_name = "layer_norm_3.b_0" shape = [512] dtype = "float32" min_val = float("-0.799406") @@ -522,6 +572,7 @@ class Program_weight_tensor_parameter_49: class Program_weight_tensor_parameter_50: name = "parameter_50" + original_name = "layer_norm_3.w_0" shape = [512] dtype = "float32" min_val = float("0.038996") @@ -533,6 +584,7 @@ class Program_weight_tensor_parameter_50: class Program_weight_tensor_parameter_51: name = "parameter_51" + original_name = "conv2d_3.b_0" shape = [512] dtype = "float32" min_val = float("-1.59531") @@ -544,6 +596,7 @@ class Program_weight_tensor_parameter_51: class Program_weight_tensor_parameter_52: name = "parameter_52" + original_name = "conv2d_3.w_0" shape = [512, 320, 3, 3] dtype = "float32" min_val = float("-1.48096") @@ -555,6 +608,7 @@ class Program_weight_tensor_parameter_52: class Program_weight_tensor_parameter_53: name = "parameter_53" + original_name = "layer_norm_24.b_0" shape = [320] dtype = "float32" min_val = float("-0.324093") @@ -566,6 +620,7 @@ class Program_weight_tensor_parameter_53: class Program_weight_tensor_parameter_54: name = "parameter_54" + original_name = "layer_norm_24.w_0" shape = [320] dtype = "float32" min_val = float("0.168841") @@ -577,6 +632,7 @@ class Program_weight_tensor_parameter_54: class Program_weight_tensor_parameter_55: name = "parameter_55" + original_name = "linear_29.b_0" shape = [320] dtype = "float32" min_val = float("-23.5264") @@ -588,6 +644,7 @@ class Program_weight_tensor_parameter_55: class Program_weight_tensor_parameter_56: name = "parameter_56" + original_name = "linear_29.w_0" shape = [1280, 320] dtype = "float32" min_val = float("-3.64432") @@ -599,6 +656,7 @@ class Program_weight_tensor_parameter_56: class Program_weight_tensor_parameter_57: name = "parameter_57" + original_name = "conv2d_15.b_0" shape = [1280] dtype = "float32" min_val = float("-1.56084") @@ -610,6 +668,7 @@ class Program_weight_tensor_parameter_57: class Program_weight_tensor_parameter_58: name = "parameter_58" + original_name = "conv2d_15.w_0" shape = [1280, 1, 3, 3] dtype = "float32" min_val = float("-0.614649") @@ -621,6 +680,7 @@ class Program_weight_tensor_parameter_58: class Program_weight_tensor_parameter_59: name = "parameter_59" + original_name = "linear_28.b_0" shape = [1280] dtype = "float32" min_val = float("-2.15391") @@ -632,6 +692,7 @@ class Program_weight_tensor_parameter_59: class Program_weight_tensor_parameter_60: name = "parameter_60" + original_name = "linear_28.w_0" shape = [320, 1280] dtype = "float32" min_val = float("-0.412544") @@ -643,6 +704,7 @@ class Program_weight_tensor_parameter_60: class Program_weight_tensor_parameter_61: name = "parameter_61" + original_name = "layer_norm_23.b_0" shape = [320] dtype = "float32" min_val = float("-1.27748") @@ -654,6 +716,7 @@ class Program_weight_tensor_parameter_61: class Program_weight_tensor_parameter_62: name = "parameter_62" + original_name = "layer_norm_23.w_0" shape = [320] dtype = "float32" min_val = float("0.369745") @@ -665,6 +728,7 @@ class Program_weight_tensor_parameter_62: class Program_weight_tensor_parameter_63: name = "parameter_63" + original_name = "linear_27.b_0" shape = [320] dtype = "float32" min_val = float("-1.47338") @@ -676,6 +740,7 @@ class Program_weight_tensor_parameter_63: class Program_weight_tensor_parameter_64: name = "parameter_64" + original_name = "linear_27.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.960873") @@ -687,6 +752,7 @@ class Program_weight_tensor_parameter_64: class Program_weight_tensor_parameter_65: name = "parameter_65" + original_name = "linear_26.b_0" shape = [640] dtype = "float32" min_val = float("-1.10372") @@ -698,6 +764,7 @@ class Program_weight_tensor_parameter_65: class Program_weight_tensor_parameter_66: name = "parameter_66" + original_name = "linear_26.w_0" shape = [320, 640] dtype = "float32" min_val = float("-0.343218") @@ -709,6 +776,7 @@ class Program_weight_tensor_parameter_66: class Program_weight_tensor_parameter_67: name = "parameter_67" + original_name = "layer_norm_22.b_0" shape = [320] dtype = "float32" min_val = float("-0.283099") @@ -720,6 +788,7 @@ class Program_weight_tensor_parameter_67: class Program_weight_tensor_parameter_68: name = "parameter_68" + original_name = "layer_norm_22.w_0" shape = [320] dtype = "float32" min_val = float("-0.000472527") @@ -731,6 +800,7 @@ class Program_weight_tensor_parameter_68: class Program_weight_tensor_parameter_69: name = "parameter_69" + original_name = "conv2d_14.b_0" shape = [320] dtype = "float32" min_val = float("-5.18353") @@ -742,6 +812,7 @@ class Program_weight_tensor_parameter_69: class Program_weight_tensor_parameter_70: name = "parameter_70" + original_name = "conv2d_14.w_0" shape = [320, 320, 2, 2] dtype = "float32" min_val = float("-0.386955") @@ -753,6 +824,7 @@ class Program_weight_tensor_parameter_70: class Program_weight_tensor_parameter_71: name = "parameter_71" + original_name = "linear_25.b_0" shape = [320] dtype = "float32" min_val = float("-3.88819") @@ -764,6 +836,7 @@ class Program_weight_tensor_parameter_71: class Program_weight_tensor_parameter_72: name = "parameter_72" + original_name = "linear_25.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.306673") @@ -775,6 +848,7 @@ class Program_weight_tensor_parameter_72: class Program_weight_tensor_parameter_73: name = "parameter_73" + original_name = "layer_norm_21.b_0" shape = [320] dtype = "float32" min_val = float("-0.823653") @@ -786,6 +860,7 @@ class Program_weight_tensor_parameter_73: class Program_weight_tensor_parameter_74: name = "parameter_74" + original_name = "layer_norm_21.w_0" shape = [320] dtype = "float32" min_val = float("0.182725") @@ -797,6 +872,7 @@ class Program_weight_tensor_parameter_74: class Program_weight_tensor_parameter_75: name = "parameter_75" + original_name = "linear_24.b_0" shape = [320] dtype = "float32" min_val = float("-4.38171") @@ -808,6 +884,7 @@ class Program_weight_tensor_parameter_75: class Program_weight_tensor_parameter_76: name = "parameter_76" + original_name = "linear_24.w_0" shape = [1280, 320] dtype = "float32" min_val = float("-1.46264") @@ -819,6 +896,7 @@ class Program_weight_tensor_parameter_76: class Program_weight_tensor_parameter_77: name = "parameter_77" + original_name = "conv2d_13.b_0" shape = [1280] dtype = "float32" min_val = float("-2.23645") @@ -830,6 +908,7 @@ class Program_weight_tensor_parameter_77: class Program_weight_tensor_parameter_78: name = "parameter_78" + original_name = "conv2d_13.w_0" shape = [1280, 1, 3, 3] dtype = "float32" min_val = float("-0.410158") @@ -841,6 +920,7 @@ class Program_weight_tensor_parameter_78: class Program_weight_tensor_parameter_79: name = "parameter_79" + original_name = "linear_23.b_0" shape = [1280] dtype = "float32" min_val = float("-3.76004") @@ -852,6 +932,7 @@ class Program_weight_tensor_parameter_79: class Program_weight_tensor_parameter_80: name = "parameter_80" + original_name = "linear_23.w_0" shape = [320, 1280] dtype = "float32" min_val = float("-0.496973") @@ -863,6 +944,7 @@ class Program_weight_tensor_parameter_80: class Program_weight_tensor_parameter_81: name = "parameter_81" + original_name = "layer_norm_20.b_0" shape = [320] dtype = "float32" min_val = float("-1.1992") @@ -874,6 +956,7 @@ class Program_weight_tensor_parameter_81: class Program_weight_tensor_parameter_82: name = "parameter_82" + original_name = "layer_norm_20.w_0" shape = [320] dtype = "float32" min_val = float("0.586359") @@ -885,6 +968,7 @@ class Program_weight_tensor_parameter_82: class Program_weight_tensor_parameter_83: name = "parameter_83" + original_name = "linear_22.b_0" shape = [320] dtype = "float32" min_val = float("-0.261722") @@ -896,6 +980,7 @@ class Program_weight_tensor_parameter_83: class Program_weight_tensor_parameter_84: name = "parameter_84" + original_name = "linear_22.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.328254") @@ -907,6 +992,7 @@ class Program_weight_tensor_parameter_84: class Program_weight_tensor_parameter_85: name = "parameter_85" + original_name = "linear_21.b_0" shape = [640] dtype = "float32" min_val = float("-1.0092") @@ -918,6 +1004,7 @@ class Program_weight_tensor_parameter_85: class Program_weight_tensor_parameter_86: name = "parameter_86" + original_name = "linear_21.w_0" shape = [320, 640] dtype = "float32" min_val = float("-0.384673") @@ -929,6 +1016,7 @@ class Program_weight_tensor_parameter_86: class Program_weight_tensor_parameter_87: name = "parameter_87" + original_name = "layer_norm_19.b_0" shape = [320] dtype = "float32" min_val = float("-0.30413") @@ -940,6 +1028,7 @@ class Program_weight_tensor_parameter_87: class Program_weight_tensor_parameter_88: name = "parameter_88" + original_name = "layer_norm_19.w_0" shape = [320] dtype = "float32" min_val = float("-0.00397075") @@ -951,6 +1040,7 @@ class Program_weight_tensor_parameter_88: class Program_weight_tensor_parameter_89: name = "parameter_89" + original_name = "conv2d_12.b_0" shape = [320] dtype = "float32" min_val = float("-3.21966") @@ -962,6 +1052,7 @@ class Program_weight_tensor_parameter_89: class Program_weight_tensor_parameter_90: name = "parameter_90" + original_name = "conv2d_12.w_0" shape = [320, 320, 2, 2] dtype = "float32" min_val = float("-0.277805") @@ -973,6 +1064,7 @@ class Program_weight_tensor_parameter_90: class Program_weight_tensor_parameter_91: name = "parameter_91" + original_name = "linear_20.b_0" shape = [320] dtype = "float32" min_val = float("-2.99991") @@ -984,6 +1076,7 @@ class Program_weight_tensor_parameter_91: class Program_weight_tensor_parameter_92: name = "parameter_92" + original_name = "linear_20.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.298048") @@ -995,6 +1088,7 @@ class Program_weight_tensor_parameter_92: class Program_weight_tensor_parameter_93: name = "parameter_93" + original_name = "layer_norm_18.b_0" shape = [320] dtype = "float32" min_val = float("-0.657501") @@ -1006,6 +1100,7 @@ class Program_weight_tensor_parameter_93: class Program_weight_tensor_parameter_94: name = "parameter_94" + original_name = "layer_norm_18.w_0" shape = [320] dtype = "float32" min_val = float("-0.000616667") @@ -1017,6 +1112,7 @@ class Program_weight_tensor_parameter_94: class Program_weight_tensor_parameter_95: name = "parameter_95" + original_name = "layer_norm_2.b_0" shape = [320] dtype = "float32" min_val = float("-1.09885") @@ -1028,6 +1124,7 @@ class Program_weight_tensor_parameter_95: class Program_weight_tensor_parameter_96: name = "parameter_96" + original_name = "layer_norm_2.w_0" shape = [320] dtype = "float32" min_val = float("0.0946223") @@ -1039,6 +1136,7 @@ class Program_weight_tensor_parameter_96: class Program_weight_tensor_parameter_97: name = "parameter_97" + original_name = "conv2d_2.b_0" shape = [320] dtype = "float32" min_val = float("-12.5054") @@ -1050,6 +1148,7 @@ class Program_weight_tensor_parameter_97: class Program_weight_tensor_parameter_98: name = "parameter_98" + original_name = "conv2d_2.w_0" shape = [320, 128, 3, 3] dtype = "float32" min_val = float("-1.70219") @@ -1061,6 +1160,7 @@ class Program_weight_tensor_parameter_98: class Program_weight_tensor_parameter_99: name = "parameter_99" + original_name = "layer_norm_17.b_0" shape = [128] dtype = "float32" min_val = float("-0.456984") @@ -1072,6 +1172,7 @@ class Program_weight_tensor_parameter_99: class Program_weight_tensor_parameter_100: name = "parameter_100" + original_name = "layer_norm_17.w_0" shape = [128] dtype = "float32" min_val = float("0.0780124") @@ -1083,6 +1184,7 @@ class Program_weight_tensor_parameter_100: class Program_weight_tensor_parameter_101: name = "parameter_101" + original_name = "linear_19.b_0" shape = [128] dtype = "float32" min_val = float("-9.29685") @@ -1094,6 +1196,7 @@ class Program_weight_tensor_parameter_101: class Program_weight_tensor_parameter_102: name = "parameter_102" + original_name = "linear_19.w_0" shape = [512, 128] dtype = "float32" min_val = float("-1.72144") @@ -1105,6 +1208,7 @@ class Program_weight_tensor_parameter_102: class Program_weight_tensor_parameter_103: name = "parameter_103" + original_name = "conv2d_11.b_0" shape = [512] dtype = "float32" min_val = float("-2.67068") @@ -1116,6 +1220,7 @@ class Program_weight_tensor_parameter_103: class Program_weight_tensor_parameter_104: name = "parameter_104" + original_name = "conv2d_11.w_0" shape = [512, 1, 3, 3] dtype = "float32" min_val = float("-0.63324") @@ -1127,6 +1232,7 @@ class Program_weight_tensor_parameter_104: class Program_weight_tensor_parameter_105: name = "parameter_105" + original_name = "linear_18.b_0" shape = [512] dtype = "float32" min_val = float("-5.14361") @@ -1138,6 +1244,7 @@ class Program_weight_tensor_parameter_105: class Program_weight_tensor_parameter_106: name = "parameter_106" + original_name = "linear_18.w_0" shape = [128, 512] dtype = "float32" min_val = float("-0.376379") @@ -1149,6 +1256,7 @@ class Program_weight_tensor_parameter_106: class Program_weight_tensor_parameter_107: name = "parameter_107" + original_name = "layer_norm_16.b_0" shape = [128] dtype = "float32" min_val = float("-2.09152") @@ -1160,6 +1268,7 @@ class Program_weight_tensor_parameter_107: class Program_weight_tensor_parameter_108: name = "parameter_108" + original_name = "layer_norm_16.w_0" shape = [128] dtype = "float32" min_val = float("1.28643") @@ -1171,6 +1280,7 @@ class Program_weight_tensor_parameter_108: class Program_weight_tensor_parameter_109: name = "parameter_109" + original_name = "linear_17.b_0" shape = [128] dtype = "float32" min_val = float("-0.124438") @@ -1182,6 +1292,7 @@ class Program_weight_tensor_parameter_109: class Program_weight_tensor_parameter_110: name = "parameter_110" + original_name = "linear_17.w_0" shape = [128, 128] dtype = "float32" min_val = float("-0.464365") @@ -1193,6 +1304,7 @@ class Program_weight_tensor_parameter_110: class Program_weight_tensor_parameter_111: name = "parameter_111" + original_name = "linear_16.b_0" shape = [256] dtype = "float32" min_val = float("-0.983209") @@ -1204,6 +1316,7 @@ class Program_weight_tensor_parameter_111: class Program_weight_tensor_parameter_112: name = "parameter_112" + original_name = "linear_16.w_0" shape = [128, 256] dtype = "float32" min_val = float("-0.386233") @@ -1215,6 +1328,7 @@ class Program_weight_tensor_parameter_112: class Program_weight_tensor_parameter_113: name = "parameter_113" + original_name = "layer_norm_15.b_0" shape = [128] dtype = "float32" min_val = float("-0.513053") @@ -1226,6 +1340,7 @@ class Program_weight_tensor_parameter_113: class Program_weight_tensor_parameter_114: name = "parameter_114" + original_name = "layer_norm_15.w_0" shape = [128] dtype = "float32" min_val = float("-0.00157539") @@ -1237,6 +1352,7 @@ class Program_weight_tensor_parameter_114: class Program_weight_tensor_parameter_115: name = "parameter_115" + original_name = "conv2d_10.b_0" shape = [128] dtype = "float32" min_val = float("-2.95376") @@ -1248,6 +1364,7 @@ class Program_weight_tensor_parameter_115: class Program_weight_tensor_parameter_116: name = "parameter_116" + original_name = "conv2d_10.w_0" shape = [128, 128, 4, 4] dtype = "float32" min_val = float("-0.281817") @@ -1259,6 +1376,7 @@ class Program_weight_tensor_parameter_116: class Program_weight_tensor_parameter_117: name = "parameter_117" + original_name = "linear_15.b_0" shape = [128] dtype = "float32" min_val = float("-3.01601") @@ -1270,6 +1388,7 @@ class Program_weight_tensor_parameter_117: class Program_weight_tensor_parameter_118: name = "parameter_118" + original_name = "linear_15.w_0" shape = [128, 128] dtype = "float32" min_val = float("-0.308206") @@ -1281,6 +1400,7 @@ class Program_weight_tensor_parameter_118: class Program_weight_tensor_parameter_119: name = "parameter_119" + original_name = "layer_norm_14.b_0" shape = [128] dtype = "float32" min_val = float("-0.790887") @@ -1292,6 +1412,7 @@ class Program_weight_tensor_parameter_119: class Program_weight_tensor_parameter_120: name = "parameter_120" + original_name = "layer_norm_14.w_0" shape = [128] dtype = "float32" min_val = float("0.315586") @@ -1303,6 +1424,7 @@ class Program_weight_tensor_parameter_120: class Program_weight_tensor_parameter_121: name = "parameter_121" + original_name = "linear_14.b_0" shape = [128] dtype = "float32" min_val = float("-1.51189") @@ -1314,6 +1436,7 @@ class Program_weight_tensor_parameter_121: class Program_weight_tensor_parameter_122: name = "parameter_122" + original_name = "linear_14.w_0" shape = [512, 128] dtype = "float32" min_val = float("-0.607708") @@ -1325,6 +1448,7 @@ class Program_weight_tensor_parameter_122: class Program_weight_tensor_parameter_123: name = "parameter_123" + original_name = "conv2d_9.b_0" shape = [512] dtype = "float32" min_val = float("-1.9238") @@ -1336,6 +1460,7 @@ class Program_weight_tensor_parameter_123: class Program_weight_tensor_parameter_124: name = "parameter_124" + original_name = "conv2d_9.w_0" shape = [512, 1, 3, 3] dtype = "float32" min_val = float("-0.301527") @@ -1347,6 +1472,7 @@ class Program_weight_tensor_parameter_124: class Program_weight_tensor_parameter_125: name = "parameter_125" + original_name = "linear_13.b_0" shape = [512] dtype = "float32" min_val = float("-3.80511") @@ -1358,6 +1484,7 @@ class Program_weight_tensor_parameter_125: class Program_weight_tensor_parameter_126: name = "parameter_126" + original_name = "linear_13.w_0" shape = [128, 512] dtype = "float32" min_val = float("-0.301925") @@ -1369,6 +1496,7 @@ class Program_weight_tensor_parameter_126: class Program_weight_tensor_parameter_127: name = "parameter_127" + original_name = "layer_norm_13.b_0" shape = [128] dtype = "float32" min_val = float("-1.88493") @@ -1380,6 +1508,7 @@ class Program_weight_tensor_parameter_127: class Program_weight_tensor_parameter_128: name = "parameter_128" + original_name = "layer_norm_13.w_0" shape = [128] dtype = "float32" min_val = float("2.02607") @@ -1391,6 +1520,7 @@ class Program_weight_tensor_parameter_128: class Program_weight_tensor_parameter_129: name = "parameter_129" + original_name = "linear_12.b_0" shape = [128] dtype = "float32" min_val = float("-0.239518") @@ -1402,6 +1532,7 @@ class Program_weight_tensor_parameter_129: class Program_weight_tensor_parameter_130: name = "parameter_130" + original_name = "linear_12.w_0" shape = [128, 128] dtype = "float32" min_val = float("-0.200575") @@ -1413,6 +1544,7 @@ class Program_weight_tensor_parameter_130: class Program_weight_tensor_parameter_131: name = "parameter_131" + original_name = "linear_11.b_0" shape = [256] dtype = "float32" min_val = float("-1.67485") @@ -1424,6 +1556,7 @@ class Program_weight_tensor_parameter_131: class Program_weight_tensor_parameter_132: name = "parameter_132" + original_name = "linear_11.w_0" shape = [128, 256] dtype = "float32" min_val = float("-0.42649") @@ -1435,6 +1568,7 @@ class Program_weight_tensor_parameter_132: class Program_weight_tensor_parameter_133: name = "parameter_133" + original_name = "layer_norm_12.b_0" shape = [128] dtype = "float32" min_val = float("-0.316753") @@ -1446,6 +1580,7 @@ class Program_weight_tensor_parameter_133: class Program_weight_tensor_parameter_134: name = "parameter_134" + original_name = "layer_norm_12.w_0" shape = [128] dtype = "float32" min_val = float("-0.00520847") @@ -1457,6 +1592,7 @@ class Program_weight_tensor_parameter_134: class Program_weight_tensor_parameter_135: name = "parameter_135" + original_name = "conv2d_8.b_0" shape = [128] dtype = "float32" min_val = float("-2.12598") @@ -1468,6 +1604,7 @@ class Program_weight_tensor_parameter_135: class Program_weight_tensor_parameter_136: name = "parameter_136" + original_name = "conv2d_8.w_0" shape = [128, 128, 4, 4] dtype = "float32" min_val = float("-0.291574") @@ -1479,6 +1616,7 @@ class Program_weight_tensor_parameter_136: class Program_weight_tensor_parameter_137: name = "parameter_137" + original_name = "linear_10.b_0" shape = [128] dtype = "float32" min_val = float("-1.7361") @@ -1490,6 +1628,7 @@ class Program_weight_tensor_parameter_137: class Program_weight_tensor_parameter_138: name = "parameter_138" + original_name = "linear_10.w_0" shape = [128, 128] dtype = "float32" min_val = float("-0.353118") @@ -1501,6 +1640,7 @@ class Program_weight_tensor_parameter_138: class Program_weight_tensor_parameter_139: name = "parameter_139" + original_name = "layer_norm_11.b_0" shape = [128] dtype = "float32" min_val = float("-0.97893") @@ -1512,6 +1652,7 @@ class Program_weight_tensor_parameter_139: class Program_weight_tensor_parameter_140: name = "parameter_140" + original_name = "layer_norm_11.w_0" shape = [128] dtype = "float32" min_val = float("0.000274388") @@ -1523,6 +1664,7 @@ class Program_weight_tensor_parameter_140: class Program_weight_tensor_parameter_141: name = "parameter_141" + original_name = "layer_norm_1.b_0" shape = [128] dtype = "float32" min_val = float("-0.56144") @@ -1534,6 +1676,7 @@ class Program_weight_tensor_parameter_141: class Program_weight_tensor_parameter_142: name = "parameter_142" + original_name = "layer_norm_1.w_0" shape = [128] dtype = "float32" min_val = float("0.16681") @@ -1545,6 +1688,7 @@ class Program_weight_tensor_parameter_142: class Program_weight_tensor_parameter_143: name = "parameter_143" + original_name = "conv2d_1.b_0" shape = [128] dtype = "float32" min_val = float("-1.24974") @@ -1556,6 +1700,7 @@ class Program_weight_tensor_parameter_143: class Program_weight_tensor_parameter_144: name = "parameter_144" + original_name = "conv2d_1.w_0" shape = [128, 64, 3, 3] dtype = "float32" min_val = float("-0.490309") @@ -1567,6 +1712,7 @@ class Program_weight_tensor_parameter_144: class Program_weight_tensor_parameter_145: name = "parameter_145" + original_name = "layer_norm_10.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -1576,6 +1722,7 @@ class Program_weight_tensor_parameter_145: class Program_weight_tensor_parameter_146: name = "parameter_146" + original_name = "layer_norm_10.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -1585,6 +1732,7 @@ class Program_weight_tensor_parameter_146: class Program_weight_tensor_parameter_147: name = "parameter_147" + original_name = "linear_9.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -1594,6 +1742,7 @@ class Program_weight_tensor_parameter_147: class Program_weight_tensor_parameter_148: name = "parameter_148" + original_name = "linear_9.w_0" shape = [256, 64] dtype = "float32" min_val = float("-0.66447") @@ -1605,6 +1754,7 @@ class Program_weight_tensor_parameter_148: class Program_weight_tensor_parameter_149: name = "parameter_149" + original_name = "conv2d_7.b_0" shape = [256] dtype = "float32" min_val = float("-1.78404") @@ -1616,6 +1766,7 @@ class Program_weight_tensor_parameter_149: class Program_weight_tensor_parameter_150: name = "parameter_150" + original_name = "conv2d_7.w_0" shape = [256, 1, 3, 3] dtype = "float32" min_val = float("-0.401445") @@ -1627,6 +1778,7 @@ class Program_weight_tensor_parameter_150: class Program_weight_tensor_parameter_151: name = "parameter_151" + original_name = "linear_8.b_0" shape = [256] dtype = "float32" min_val = float("-2.55396") @@ -1638,6 +1790,7 @@ class Program_weight_tensor_parameter_151: class Program_weight_tensor_parameter_152: name = "parameter_152" + original_name = "linear_8.w_0" shape = [64, 256] dtype = "float32" min_val = float("-0.35554") @@ -1649,6 +1802,7 @@ class Program_weight_tensor_parameter_152: class Program_weight_tensor_parameter_153: name = "parameter_153" + original_name = "layer_norm_9.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -1658,6 +1812,7 @@ class Program_weight_tensor_parameter_153: class Program_weight_tensor_parameter_154: name = "parameter_154" + original_name = "layer_norm_9.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -1667,6 +1822,7 @@ class Program_weight_tensor_parameter_154: class Program_weight_tensor_parameter_155: name = "parameter_155" + original_name = "linear_7.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -1676,6 +1832,7 @@ class Program_weight_tensor_parameter_155: class Program_weight_tensor_parameter_156: name = "parameter_156" + original_name = "linear_7.w_0" shape = [64, 64] dtype = "float32" min_val = float("-0.342726") @@ -1687,6 +1844,7 @@ class Program_weight_tensor_parameter_156: class Program_weight_tensor_parameter_157: name = "parameter_157" + original_name = "linear_6.b_0" shape = [128] dtype = "float32" min_val = float("-2.95555") @@ -1698,6 +1856,7 @@ class Program_weight_tensor_parameter_157: class Program_weight_tensor_parameter_158: name = "parameter_158" + original_name = "linear_6.w_0" shape = [64, 128] dtype = "float32" min_val = float("-0.492233") @@ -1709,6 +1868,7 @@ class Program_weight_tensor_parameter_158: class Program_weight_tensor_parameter_159: name = "parameter_159" + original_name = "layer_norm_8.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -1718,6 +1878,7 @@ class Program_weight_tensor_parameter_159: class Program_weight_tensor_parameter_160: name = "parameter_160" + original_name = "layer_norm_8.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -1727,6 +1888,7 @@ class Program_weight_tensor_parameter_160: class Program_weight_tensor_parameter_161: name = "parameter_161" + original_name = "conv2d_6.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -1736,6 +1898,7 @@ class Program_weight_tensor_parameter_161: class Program_weight_tensor_parameter_162: name = "parameter_162" + original_name = "conv2d_6.w_0" shape = [64, 64, 8, 8] dtype = "float32" min_val = float("-0.445696") @@ -1747,6 +1910,7 @@ class Program_weight_tensor_parameter_162: class Program_weight_tensor_parameter_163: name = "parameter_163" + original_name = "linear_5.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -1756,6 +1920,7 @@ class Program_weight_tensor_parameter_163: class Program_weight_tensor_parameter_164: name = "parameter_164" + original_name = "linear_5.w_0" shape = [64, 64] dtype = "float32" min_val = float("-0.464383") @@ -1767,6 +1932,7 @@ class Program_weight_tensor_parameter_164: class Program_weight_tensor_parameter_165: name = "parameter_165" + original_name = "layer_norm_7.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -1776,6 +1942,7 @@ class Program_weight_tensor_parameter_165: class Program_weight_tensor_parameter_166: name = "parameter_166" + original_name = "layer_norm_7.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -1785,6 +1952,7 @@ class Program_weight_tensor_parameter_166: class Program_weight_tensor_parameter_167: name = "parameter_167" + original_name = "linear_4.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -1794,6 +1962,7 @@ class Program_weight_tensor_parameter_167: class Program_weight_tensor_parameter_168: name = "parameter_168" + original_name = "linear_4.w_0" shape = [256, 64] dtype = "float32" min_val = float("-0.376997") @@ -1805,6 +1974,7 @@ class Program_weight_tensor_parameter_168: class Program_weight_tensor_parameter_169: name = "parameter_169" + original_name = "conv2d_5.b_0" shape = [256] dtype = "float32" min_val = float("-1.99132") @@ -1816,6 +1986,7 @@ class Program_weight_tensor_parameter_169: class Program_weight_tensor_parameter_170: name = "parameter_170" + original_name = "conv2d_5.w_0" shape = [256, 1, 3, 3] dtype = "float32" min_val = float("-0.638545") @@ -1827,6 +1998,7 @@ class Program_weight_tensor_parameter_170: class Program_weight_tensor_parameter_171: name = "parameter_171" + original_name = "linear_3.b_0" shape = [256] dtype = "float32" min_val = float("-6.93562") @@ -1838,6 +2010,7 @@ class Program_weight_tensor_parameter_171: class Program_weight_tensor_parameter_172: name = "parameter_172" + original_name = "linear_3.w_0" shape = [64, 256] dtype = "float32" min_val = float("-0.640579") @@ -1849,6 +2022,7 @@ class Program_weight_tensor_parameter_172: class Program_weight_tensor_parameter_173: name = "parameter_173" + original_name = "layer_norm_6.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -1858,6 +2032,7 @@ class Program_weight_tensor_parameter_173: class Program_weight_tensor_parameter_174: name = "parameter_174" + original_name = "layer_norm_6.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -1867,6 +2042,7 @@ class Program_weight_tensor_parameter_174: class Program_weight_tensor_parameter_175: name = "parameter_175" + original_name = "linear_2.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -1876,6 +2052,7 @@ class Program_weight_tensor_parameter_175: class Program_weight_tensor_parameter_176: name = "parameter_176" + original_name = "linear_2.w_0" shape = [64, 64] dtype = "float32" min_val = float("-0.190437") @@ -1887,6 +2064,7 @@ class Program_weight_tensor_parameter_176: class Program_weight_tensor_parameter_177: name = "parameter_177" + original_name = "linear_1.b_0" shape = [128] dtype = "float32" min_val = float("-1.99398") @@ -1898,6 +2076,7 @@ class Program_weight_tensor_parameter_177: class Program_weight_tensor_parameter_178: name = "parameter_178" + original_name = "linear_1.w_0" shape = [64, 128] dtype = "float32" min_val = float("-0.550134") @@ -1909,6 +2088,7 @@ class Program_weight_tensor_parameter_178: class Program_weight_tensor_parameter_179: name = "parameter_179" + original_name = "layer_norm_5.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -1918,6 +2098,7 @@ class Program_weight_tensor_parameter_179: class Program_weight_tensor_parameter_180: name = "parameter_180" + original_name = "layer_norm_5.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -1927,6 +2108,7 @@ class Program_weight_tensor_parameter_180: class Program_weight_tensor_parameter_181: name = "parameter_181" + original_name = "conv2d_4.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -1936,6 +2118,7 @@ class Program_weight_tensor_parameter_181: class Program_weight_tensor_parameter_182: name = "parameter_182" + original_name = "conv2d_4.w_0" shape = [64, 64, 8, 8] dtype = "float32" min_val = float("-0.385525") @@ -1947,6 +2130,7 @@ class Program_weight_tensor_parameter_182: class Program_weight_tensor_parameter_183: name = "parameter_183" + original_name = "linear_0.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -1956,6 +2140,7 @@ class Program_weight_tensor_parameter_183: class Program_weight_tensor_parameter_184: name = "parameter_184" + original_name = "linear_0.w_0" shape = [64, 64] dtype = "float32" min_val = float("-0.589345") @@ -1967,6 +2152,7 @@ class Program_weight_tensor_parameter_184: class Program_weight_tensor_parameter_185: name = "parameter_185" + original_name = "layer_norm_4.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -1976,6 +2162,7 @@ class Program_weight_tensor_parameter_185: class Program_weight_tensor_parameter_186: name = "parameter_186" + original_name = "layer_norm_4.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -1985,6 +2172,7 @@ class Program_weight_tensor_parameter_186: class Program_weight_tensor_parameter_187: name = "parameter_187" + original_name = "layer_norm_0.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -1994,6 +2182,7 @@ class Program_weight_tensor_parameter_187: class Program_weight_tensor_parameter_188: name = "parameter_188" + original_name = "layer_norm_0.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -2003,6 +2192,7 @@ class Program_weight_tensor_parameter_188: class Program_weight_tensor_parameter_189: name = "parameter_189" + original_name = "conv2d_0.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -2012,6 +2202,7 @@ class Program_weight_tensor_parameter_189: class Program_weight_tensor_parameter_190: name = "parameter_190" + original_name = "conv2d_0.w_0" shape = [64, 3, 7, 7] dtype = "float32" min_val = float("-0.306084") diff --git a/paddle_samples/PaddleX/SegFormer-B2/subgraph_1/input_meta.py b/paddle_samples/PaddleX/SegFormer-B2/subgraph_1/input_meta.py index 9f3e35589..e6cfafd3e 100644 --- a/paddle_samples/PaddleX/SegFormer-B2/subgraph_1/input_meta.py +++ b/paddle_samples/PaddleX/SegFormer-B2/subgraph_1/input_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_data_0: name = "data_0" + original_name = "var_0" shape = [2, 3, 512, 1024] dtype = "float32" min_val = float("-0.843137") diff --git a/paddle_samples/PaddleX/SegFormer-B2/subgraph_1/weight_meta.py b/paddle_samples/PaddleX/SegFormer-B2/subgraph_1/weight_meta.py index 03b1e974f..b70b70d21 100644 --- a/paddle_samples/PaddleX/SegFormer-B2/subgraph_1/weight_meta.py +++ b/paddle_samples/PaddleX/SegFormer-B2/subgraph_1/weight_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_parameter_0: name = "parameter_0" + original_name = "conv2d_34.b_0" shape = [2] dtype = "float32" min_val = float("0") @@ -9,6 +10,7 @@ class Program_weight_tensor_parameter_0: class Program_weight_tensor_parameter_1: name = "parameter_1" + original_name = "conv2d_34.w_0" shape = [2, 768, 1, 1] dtype = "float32" min_val = float("-0.154972") @@ -20,6 +22,7 @@ class Program_weight_tensor_parameter_1: class Program_weight_tensor_parameter_2: name = "parameter_2" + original_name = "batch_norm2d_0.b_0" shape = [768] dtype = "float32" data = None @@ -27,6 +30,7 @@ class Program_weight_tensor_parameter_2: class Program_weight_tensor_parameter_3: name = "parameter_3" + original_name = "batch_norm2d_0.w_0" shape = [768] dtype = "float32" min_val = float("1.0") @@ -37,6 +41,7 @@ class Program_weight_tensor_parameter_3: class Program_weight_tensor_parameter_4: name = "parameter_4" + original_name = "batch_norm2d_0.w_2" shape = [768] dtype = "float32" min_val = float("1.0") @@ -47,6 +52,7 @@ class Program_weight_tensor_parameter_4: class Program_weight_tensor_parameter_5: name = "parameter_5" + original_name = "batch_norm2d_0.w_1" shape = [768] dtype = "float32" data = None @@ -54,6 +60,7 @@ class Program_weight_tensor_parameter_5: class Program_weight_tensor_parameter_6: name = "parameter_6" + original_name = "conv2d_33.w_0" shape = [768, 3072, 1, 1] dtype = "float32" min_val = float("-0.123054") @@ -65,6 +72,7 @@ class Program_weight_tensor_parameter_6: class Program_weight_tensor_parameter_7: name = "parameter_7" + original_name = "linear_83.b_0" shape = [768] dtype = "float32" data = None @@ -72,6 +80,7 @@ class Program_weight_tensor_parameter_7: class Program_weight_tensor_parameter_8: name = "parameter_8" + original_name = "linear_83.w_0" shape = [64, 768] dtype = "float32" min_val = float("-0.0849193") @@ -83,6 +92,7 @@ class Program_weight_tensor_parameter_8: class Program_weight_tensor_parameter_9: name = "parameter_9" + original_name = "linear_82.b_0" shape = [768] dtype = "float32" data = None @@ -90,6 +100,7 @@ class Program_weight_tensor_parameter_9: class Program_weight_tensor_parameter_10: name = "parameter_10" + original_name = "linear_82.w_0" shape = [128, 768] dtype = "float32" min_val = float("-0.0818299") @@ -101,6 +112,7 @@ class Program_weight_tensor_parameter_10: class Program_weight_tensor_parameter_11: name = "parameter_11" + original_name = "linear_81.b_0" shape = [768] dtype = "float32" data = None @@ -108,6 +120,7 @@ class Program_weight_tensor_parameter_11: class Program_weight_tensor_parameter_12: name = "parameter_12" + original_name = "linear_81.w_0" shape = [320, 768] dtype = "float32" min_val = float("-0.0742597") @@ -119,6 +132,7 @@ class Program_weight_tensor_parameter_12: class Program_weight_tensor_parameter_13: name = "parameter_13" + original_name = "linear_80.b_0" shape = [768] dtype = "float32" data = None @@ -126,6 +140,7 @@ class Program_weight_tensor_parameter_13: class Program_weight_tensor_parameter_14: name = "parameter_14" + original_name = "linear_80.w_0" shape = [512, 768] dtype = "float32" min_val = float("-0.0684648") @@ -137,6 +152,7 @@ class Program_weight_tensor_parameter_14: class Program_weight_tensor_parameter_15: name = "parameter_15" + original_name = "layer_norm_52.b_0" shape = [512] dtype = "float32" min_val = float("-0.691189") @@ -148,6 +164,7 @@ class Program_weight_tensor_parameter_15: class Program_weight_tensor_parameter_16: name = "parameter_16" + original_name = "layer_norm_52.w_0" shape = [512] dtype = "float32" min_val = float("0.0493847") @@ -159,6 +176,7 @@ class Program_weight_tensor_parameter_16: class Program_weight_tensor_parameter_17: name = "parameter_17" + original_name = "linear_79.b_0" shape = [512] dtype = "float32" min_val = float("-4.66977") @@ -170,6 +188,7 @@ class Program_weight_tensor_parameter_17: class Program_weight_tensor_parameter_18: name = "parameter_18" + original_name = "linear_79.w_0" shape = [2048, 512] dtype = "float32" min_val = float("-1.61258") @@ -181,6 +200,7 @@ class Program_weight_tensor_parameter_18: class Program_weight_tensor_parameter_19: name = "parameter_19" + original_name = "conv2d_32.b_0" shape = [2048] dtype = "float32" min_val = float("-1.66115") @@ -192,6 +212,7 @@ class Program_weight_tensor_parameter_19: class Program_weight_tensor_parameter_20: name = "parameter_20" + original_name = "conv2d_32.w_0" shape = [2048, 1, 3, 3] dtype = "float32" min_val = float("-0.600288") @@ -203,6 +224,7 @@ class Program_weight_tensor_parameter_20: class Program_weight_tensor_parameter_21: name = "parameter_21" + original_name = "linear_78.b_0" shape = [2048] dtype = "float32" min_val = float("-5.67372") @@ -214,6 +236,7 @@ class Program_weight_tensor_parameter_21: class Program_weight_tensor_parameter_22: name = "parameter_22" + original_name = "linear_78.w_0" shape = [512, 2048] dtype = "float32" min_val = float("-0.611694") @@ -225,6 +248,7 @@ class Program_weight_tensor_parameter_22: class Program_weight_tensor_parameter_23: name = "parameter_23" + original_name = "layer_norm_51.b_0" shape = [512] dtype = "float32" min_val = float("-1.46193") @@ -236,6 +260,7 @@ class Program_weight_tensor_parameter_23: class Program_weight_tensor_parameter_24: name = "parameter_24" + original_name = "layer_norm_51.w_0" shape = [512] dtype = "float32" min_val = float("0.1782") @@ -247,6 +272,7 @@ class Program_weight_tensor_parameter_24: class Program_weight_tensor_parameter_25: name = "parameter_25" + original_name = "linear_77.b_0" shape = [512] dtype = "float32" min_val = float("-5.8902") @@ -258,6 +284,7 @@ class Program_weight_tensor_parameter_25: class Program_weight_tensor_parameter_26: name = "parameter_26" + original_name = "linear_77.w_0" shape = [512, 512] dtype = "float32" min_val = float("-1.08059") @@ -269,6 +296,7 @@ class Program_weight_tensor_parameter_26: class Program_weight_tensor_parameter_27: name = "parameter_27" + original_name = "linear_76.b_0" shape = [1024] dtype = "float32" min_val = float("-0.977369") @@ -280,6 +308,7 @@ class Program_weight_tensor_parameter_27: class Program_weight_tensor_parameter_28: name = "parameter_28" + original_name = "linear_76.w_0" shape = [512, 1024] dtype = "float32" min_val = float("-0.372655") @@ -291,6 +320,7 @@ class Program_weight_tensor_parameter_28: class Program_weight_tensor_parameter_29: name = "parameter_29" + original_name = "linear_75.b_0" shape = [512] dtype = "float32" min_val = float("-2.27125") @@ -302,6 +332,7 @@ class Program_weight_tensor_parameter_29: class Program_weight_tensor_parameter_30: name = "parameter_30" + original_name = "linear_75.w_0" shape = [512, 512] dtype = "float32" min_val = float("-0.268387") @@ -313,6 +344,7 @@ class Program_weight_tensor_parameter_30: class Program_weight_tensor_parameter_31: name = "parameter_31" + original_name = "layer_norm_50.b_0" shape = [512] dtype = "float32" min_val = float("-1.8682") @@ -324,6 +356,7 @@ class Program_weight_tensor_parameter_31: class Program_weight_tensor_parameter_32: name = "parameter_32" + original_name = "layer_norm_50.w_0" shape = [512] dtype = "float32" min_val = float("-0.000327602") @@ -335,6 +368,7 @@ class Program_weight_tensor_parameter_32: class Program_weight_tensor_parameter_33: name = "parameter_33" + original_name = "linear_74.b_0" shape = [512] dtype = "float32" min_val = float("-6.11314") @@ -346,6 +380,7 @@ class Program_weight_tensor_parameter_33: class Program_weight_tensor_parameter_34: name = "parameter_34" + original_name = "linear_74.w_0" shape = [2048, 512] dtype = "float32" min_val = float("-1.7186") @@ -357,6 +392,7 @@ class Program_weight_tensor_parameter_34: class Program_weight_tensor_parameter_35: name = "parameter_35" + original_name = "conv2d_31.b_0" shape = [2048] dtype = "float32" min_val = float("-2.22147") @@ -368,6 +404,7 @@ class Program_weight_tensor_parameter_35: class Program_weight_tensor_parameter_36: name = "parameter_36" + original_name = "conv2d_31.w_0" shape = [2048, 1, 3, 3] dtype = "float32" min_val = float("-0.788859") @@ -379,6 +416,7 @@ class Program_weight_tensor_parameter_36: class Program_weight_tensor_parameter_37: name = "parameter_37" + original_name = "linear_73.b_0" shape = [2048] dtype = "float32" min_val = float("-9.83464") @@ -390,6 +428,7 @@ class Program_weight_tensor_parameter_37: class Program_weight_tensor_parameter_38: name = "parameter_38" + original_name = "linear_73.w_0" shape = [512, 2048] dtype = "float32" min_val = float("-0.489352") @@ -401,6 +440,7 @@ class Program_weight_tensor_parameter_38: class Program_weight_tensor_parameter_39: name = "parameter_39" + original_name = "layer_norm_49.b_0" shape = [512] dtype = "float32" min_val = float("-2.64313") @@ -412,6 +452,7 @@ class Program_weight_tensor_parameter_39: class Program_weight_tensor_parameter_40: name = "parameter_40" + original_name = "layer_norm_49.w_0" shape = [512] dtype = "float32" min_val = float("-0.000440737") @@ -423,6 +464,7 @@ class Program_weight_tensor_parameter_40: class Program_weight_tensor_parameter_41: name = "parameter_41" + original_name = "linear_72.b_0" shape = [512] dtype = "float32" min_val = float("-2.27875") @@ -434,6 +476,7 @@ class Program_weight_tensor_parameter_41: class Program_weight_tensor_parameter_42: name = "parameter_42" + original_name = "linear_72.w_0" shape = [512, 512] dtype = "float32" min_val = float("-0.565352") @@ -445,6 +488,7 @@ class Program_weight_tensor_parameter_42: class Program_weight_tensor_parameter_43: name = "parameter_43" + original_name = "linear_71.b_0" shape = [1024] dtype = "float32" min_val = float("-0.778176") @@ -456,6 +500,7 @@ class Program_weight_tensor_parameter_43: class Program_weight_tensor_parameter_44: name = "parameter_44" + original_name = "linear_71.w_0" shape = [512, 1024] dtype = "float32" min_val = float("-0.313423") @@ -467,6 +512,7 @@ class Program_weight_tensor_parameter_44: class Program_weight_tensor_parameter_45: name = "parameter_45" + original_name = "linear_70.b_0" shape = [512] dtype = "float32" min_val = float("-2.54147") @@ -478,6 +524,7 @@ class Program_weight_tensor_parameter_45: class Program_weight_tensor_parameter_46: name = "parameter_46" + original_name = "linear_70.w_0" shape = [512, 512] dtype = "float32" min_val = float("-0.328666") @@ -489,6 +536,7 @@ class Program_weight_tensor_parameter_46: class Program_weight_tensor_parameter_47: name = "parameter_47" + original_name = "layer_norm_48.b_0" shape = [512] dtype = "float32" min_val = float("-1.07251") @@ -500,6 +548,7 @@ class Program_weight_tensor_parameter_47: class Program_weight_tensor_parameter_48: name = "parameter_48" + original_name = "layer_norm_48.w_0" shape = [512] dtype = "float32" min_val = float("-0.0874226") @@ -511,6 +560,7 @@ class Program_weight_tensor_parameter_48: class Program_weight_tensor_parameter_49: name = "parameter_49" + original_name = "linear_69.b_0" shape = [512] dtype = "float32" min_val = float("-1.16334") @@ -522,6 +572,7 @@ class Program_weight_tensor_parameter_49: class Program_weight_tensor_parameter_50: name = "parameter_50" + original_name = "linear_69.w_0" shape = [2048, 512] dtype = "float32" min_val = float("-1.29887") @@ -533,6 +584,7 @@ class Program_weight_tensor_parameter_50: class Program_weight_tensor_parameter_51: name = "parameter_51" + original_name = "conv2d_30.b_0" shape = [2048] dtype = "float32" min_val = float("-2.74508") @@ -544,6 +596,7 @@ class Program_weight_tensor_parameter_51: class Program_weight_tensor_parameter_52: name = "parameter_52" + original_name = "conv2d_30.w_0" shape = [2048, 1, 3, 3] dtype = "float32" min_val = float("-0.436297") @@ -555,6 +608,7 @@ class Program_weight_tensor_parameter_52: class Program_weight_tensor_parameter_53: name = "parameter_53" + original_name = "linear_68.b_0" shape = [2048] dtype = "float32" min_val = float("-1.96272") @@ -566,6 +620,7 @@ class Program_weight_tensor_parameter_53: class Program_weight_tensor_parameter_54: name = "parameter_54" + original_name = "linear_68.w_0" shape = [512, 2048] dtype = "float32" min_val = float("-0.411464") @@ -577,6 +632,7 @@ class Program_weight_tensor_parameter_54: class Program_weight_tensor_parameter_55: name = "parameter_55" + original_name = "layer_norm_47.b_0" shape = [512] dtype = "float32" min_val = float("-1.66992") @@ -588,6 +644,7 @@ class Program_weight_tensor_parameter_55: class Program_weight_tensor_parameter_56: name = "parameter_56" + original_name = "layer_norm_47.w_0" shape = [512] dtype = "float32" min_val = float("0.00120023") @@ -599,6 +656,7 @@ class Program_weight_tensor_parameter_56: class Program_weight_tensor_parameter_57: name = "parameter_57" + original_name = "linear_67.b_0" shape = [512] dtype = "float32" min_val = float("-0.860894") @@ -610,6 +668,7 @@ class Program_weight_tensor_parameter_57: class Program_weight_tensor_parameter_58: name = "parameter_58" + original_name = "linear_67.w_0" shape = [512, 512] dtype = "float32" min_val = float("-0.412764") @@ -621,6 +680,7 @@ class Program_weight_tensor_parameter_58: class Program_weight_tensor_parameter_59: name = "parameter_59" + original_name = "linear_66.b_0" shape = [1024] dtype = "float32" min_val = float("-0.786169") @@ -632,6 +692,7 @@ class Program_weight_tensor_parameter_59: class Program_weight_tensor_parameter_60: name = "parameter_60" + original_name = "linear_66.w_0" shape = [512, 1024] dtype = "float32" min_val = float("-0.315709") @@ -643,6 +704,7 @@ class Program_weight_tensor_parameter_60: class Program_weight_tensor_parameter_61: name = "parameter_61" + original_name = "linear_65.b_0" shape = [512] dtype = "float32" min_val = float("-2.42356") @@ -654,6 +716,7 @@ class Program_weight_tensor_parameter_61: class Program_weight_tensor_parameter_62: name = "parameter_62" + original_name = "linear_65.w_0" shape = [512, 512] dtype = "float32" min_val = float("-0.339588") @@ -665,6 +728,7 @@ class Program_weight_tensor_parameter_62: class Program_weight_tensor_parameter_63: name = "parameter_63" + original_name = "layer_norm_46.b_0" shape = [512] dtype = "float32" min_val = float("-0.614378") @@ -676,6 +740,7 @@ class Program_weight_tensor_parameter_63: class Program_weight_tensor_parameter_64: name = "parameter_64" + original_name = "layer_norm_46.w_0" shape = [512] dtype = "float32" min_val = float("-0.000546054") @@ -687,6 +752,7 @@ class Program_weight_tensor_parameter_64: class Program_weight_tensor_parameter_65: name = "parameter_65" + original_name = "layer_norm_3.b_0" shape = [512] dtype = "float32" min_val = float("-2.42019") @@ -698,6 +764,7 @@ class Program_weight_tensor_parameter_65: class Program_weight_tensor_parameter_66: name = "parameter_66" + original_name = "layer_norm_3.w_0" shape = [512] dtype = "float32" min_val = float("0.0291935") @@ -709,6 +776,7 @@ class Program_weight_tensor_parameter_66: class Program_weight_tensor_parameter_67: name = "parameter_67" + original_name = "conv2d_3.b_0" shape = [512] dtype = "float32" min_val = float("-22.5292") @@ -720,6 +788,7 @@ class Program_weight_tensor_parameter_67: class Program_weight_tensor_parameter_68: name = "parameter_68" + original_name = "conv2d_3.w_0" shape = [512, 320, 3, 3] dtype = "float32" min_val = float("-0.817274") @@ -731,6 +800,7 @@ class Program_weight_tensor_parameter_68: class Program_weight_tensor_parameter_69: name = "parameter_69" + original_name = "layer_norm_45.b_0" shape = [320] dtype = "float32" min_val = float("-1.32957") @@ -742,6 +812,7 @@ class Program_weight_tensor_parameter_69: class Program_weight_tensor_parameter_70: name = "parameter_70" + original_name = "layer_norm_45.w_0" shape = [320] dtype = "float32" min_val = float("0.0452703") @@ -753,6 +824,7 @@ class Program_weight_tensor_parameter_70: class Program_weight_tensor_parameter_71: name = "parameter_71" + original_name = "linear_64.b_0" shape = [320] dtype = "float32" min_val = float("-0.541573") @@ -764,6 +836,7 @@ class Program_weight_tensor_parameter_71: class Program_weight_tensor_parameter_72: name = "parameter_72" + original_name = "linear_64.w_0" shape = [1280, 320] dtype = "float32" min_val = float("-2.09295") @@ -775,6 +848,7 @@ class Program_weight_tensor_parameter_72: class Program_weight_tensor_parameter_73: name = "parameter_73" + original_name = "conv2d_29.b_0" shape = [1280] dtype = "float32" min_val = float("-2.67916") @@ -786,6 +860,7 @@ class Program_weight_tensor_parameter_73: class Program_weight_tensor_parameter_74: name = "parameter_74" + original_name = "conv2d_29.w_0" shape = [1280, 1, 3, 3] dtype = "float32" min_val = float("-1.67916") @@ -797,6 +872,7 @@ class Program_weight_tensor_parameter_74: class Program_weight_tensor_parameter_75: name = "parameter_75" + original_name = "linear_63.b_0" shape = [1280] dtype = "float32" min_val = float("-2.78884") @@ -808,6 +884,7 @@ class Program_weight_tensor_parameter_75: class Program_weight_tensor_parameter_76: name = "parameter_76" + original_name = "linear_63.w_0" shape = [320, 1280] dtype = "float32" min_val = float("-0.650852") @@ -819,6 +896,7 @@ class Program_weight_tensor_parameter_76: class Program_weight_tensor_parameter_77: name = "parameter_77" + original_name = "layer_norm_44.b_0" shape = [320] dtype = "float32" min_val = float("-2.24247") @@ -830,6 +908,7 @@ class Program_weight_tensor_parameter_77: class Program_weight_tensor_parameter_78: name = "parameter_78" + original_name = "layer_norm_44.w_0" shape = [320] dtype = "float32" min_val = float("0.257647") @@ -841,6 +920,7 @@ class Program_weight_tensor_parameter_78: class Program_weight_tensor_parameter_79: name = "parameter_79" + original_name = "linear_62.b_0" shape = [320] dtype = "float32" min_val = float("-0.270934") @@ -852,6 +932,7 @@ class Program_weight_tensor_parameter_79: class Program_weight_tensor_parameter_80: name = "parameter_80" + original_name = "linear_62.w_0" shape = [320, 320] dtype = "float32" min_val = float("-1.35486") @@ -863,6 +944,7 @@ class Program_weight_tensor_parameter_80: class Program_weight_tensor_parameter_81: name = "parameter_81" + original_name = "linear_61.b_0" shape = [640] dtype = "float32" min_val = float("-2.35506") @@ -874,6 +956,7 @@ class Program_weight_tensor_parameter_81: class Program_weight_tensor_parameter_82: name = "parameter_82" + original_name = "linear_61.w_0" shape = [320, 640] dtype = "float32" min_val = float("-0.400663") @@ -885,6 +968,7 @@ class Program_weight_tensor_parameter_82: class Program_weight_tensor_parameter_83: name = "parameter_83" + original_name = "layer_norm_43.b_0" shape = [320] dtype = "float32" min_val = float("-0.27989") @@ -896,6 +980,7 @@ class Program_weight_tensor_parameter_83: class Program_weight_tensor_parameter_84: name = "parameter_84" + original_name = "layer_norm_43.w_0" shape = [320] dtype = "float32" min_val = float("-0.0126834") @@ -907,6 +992,7 @@ class Program_weight_tensor_parameter_84: class Program_weight_tensor_parameter_85: name = "parameter_85" + original_name = "conv2d_28.b_0" shape = [320] dtype = "float32" min_val = float("-4.88612") @@ -918,6 +1004,7 @@ class Program_weight_tensor_parameter_85: class Program_weight_tensor_parameter_86: name = "parameter_86" + original_name = "conv2d_28.w_0" shape = [320, 320, 2, 2] dtype = "float32" min_val = float("-0.302937") @@ -929,6 +1016,7 @@ class Program_weight_tensor_parameter_86: class Program_weight_tensor_parameter_87: name = "parameter_87" + original_name = "linear_60.b_0" shape = [320] dtype = "float32" min_val = float("-2.07501") @@ -940,6 +1028,7 @@ class Program_weight_tensor_parameter_87: class Program_weight_tensor_parameter_88: name = "parameter_88" + original_name = "linear_60.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.295472") @@ -951,6 +1040,7 @@ class Program_weight_tensor_parameter_88: class Program_weight_tensor_parameter_89: name = "parameter_89" + original_name = "layer_norm_42.b_0" shape = [320] dtype = "float32" min_val = float("-2.05835") @@ -962,6 +1052,7 @@ class Program_weight_tensor_parameter_89: class Program_weight_tensor_parameter_90: name = "parameter_90" + original_name = "layer_norm_42.w_0" shape = [320] dtype = "float32" min_val = float("0.181273") @@ -973,6 +1064,7 @@ class Program_weight_tensor_parameter_90: class Program_weight_tensor_parameter_91: name = "parameter_91" + original_name = "linear_59.b_0" shape = [320] dtype = "float32" min_val = float("-0.746436") @@ -984,6 +1076,7 @@ class Program_weight_tensor_parameter_91: class Program_weight_tensor_parameter_92: name = "parameter_92" + original_name = "linear_59.w_0" shape = [1280, 320] dtype = "float32" min_val = float("-1.09337") @@ -995,6 +1088,7 @@ class Program_weight_tensor_parameter_92: class Program_weight_tensor_parameter_93: name = "parameter_93" + original_name = "conv2d_27.b_0" shape = [1280] dtype = "float32" min_val = float("-2.43138") @@ -1006,6 +1100,7 @@ class Program_weight_tensor_parameter_93: class Program_weight_tensor_parameter_94: name = "parameter_94" + original_name = "conv2d_27.w_0" shape = [1280, 1, 3, 3] dtype = "float32" min_val = float("-0.462057") @@ -1017,6 +1112,7 @@ class Program_weight_tensor_parameter_94: class Program_weight_tensor_parameter_95: name = "parameter_95" + original_name = "linear_58.b_0" shape = [1280] dtype = "float32" min_val = float("-2.50448") @@ -1028,6 +1124,7 @@ class Program_weight_tensor_parameter_95: class Program_weight_tensor_parameter_96: name = "parameter_96" + original_name = "linear_58.w_0" shape = [320, 1280] dtype = "float32" min_val = float("-0.308366") @@ -1039,6 +1136,7 @@ class Program_weight_tensor_parameter_96: class Program_weight_tensor_parameter_97: name = "parameter_97" + original_name = "layer_norm_41.b_0" shape = [320] dtype = "float32" min_val = float("-2.70646") @@ -1050,6 +1148,7 @@ class Program_weight_tensor_parameter_97: class Program_weight_tensor_parameter_98: name = "parameter_98" + original_name = "layer_norm_41.w_0" shape = [320] dtype = "float32" min_val = float("0.451875") @@ -1061,6 +1160,7 @@ class Program_weight_tensor_parameter_98: class Program_weight_tensor_parameter_99: name = "parameter_99" + original_name = "linear_57.b_0" shape = [320] dtype = "float32" min_val = float("-0.292624") @@ -1072,6 +1172,7 @@ class Program_weight_tensor_parameter_99: class Program_weight_tensor_parameter_100: name = "parameter_100" + original_name = "linear_57.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.38477") @@ -1083,6 +1184,7 @@ class Program_weight_tensor_parameter_100: class Program_weight_tensor_parameter_101: name = "parameter_101" + original_name = "linear_56.b_0" shape = [640] dtype = "float32" min_val = float("-0.937883") @@ -1094,6 +1196,7 @@ class Program_weight_tensor_parameter_101: class Program_weight_tensor_parameter_102: name = "parameter_102" + original_name = "linear_56.w_0" shape = [320, 640] dtype = "float32" min_val = float("-0.363069") @@ -1105,6 +1208,7 @@ class Program_weight_tensor_parameter_102: class Program_weight_tensor_parameter_103: name = "parameter_103" + original_name = "layer_norm_40.b_0" shape = [320] dtype = "float32" min_val = float("-0.32236") @@ -1116,6 +1220,7 @@ class Program_weight_tensor_parameter_103: class Program_weight_tensor_parameter_104: name = "parameter_104" + original_name = "layer_norm_40.w_0" shape = [320] dtype = "float32" min_val = float("-0.0021013") @@ -1127,6 +1232,7 @@ class Program_weight_tensor_parameter_104: class Program_weight_tensor_parameter_105: name = "parameter_105" + original_name = "conv2d_26.b_0" shape = [320] dtype = "float32" min_val = float("-6.31") @@ -1138,6 +1244,7 @@ class Program_weight_tensor_parameter_105: class Program_weight_tensor_parameter_106: name = "parameter_106" + original_name = "conv2d_26.w_0" shape = [320, 320, 2, 2] dtype = "float32" min_val = float("-0.347969") @@ -1149,6 +1256,7 @@ class Program_weight_tensor_parameter_106: class Program_weight_tensor_parameter_107: name = "parameter_107" + original_name = "linear_55.b_0" shape = [320] dtype = "float32" min_val = float("-2.46483") @@ -1160,6 +1268,7 @@ class Program_weight_tensor_parameter_107: class Program_weight_tensor_parameter_108: name = "parameter_108" + original_name = "linear_55.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.254507") @@ -1171,6 +1280,7 @@ class Program_weight_tensor_parameter_108: class Program_weight_tensor_parameter_109: name = "parameter_109" + original_name = "layer_norm_39.b_0" shape = [320] dtype = "float32" min_val = float("-1.41728") @@ -1182,6 +1292,7 @@ class Program_weight_tensor_parameter_109: class Program_weight_tensor_parameter_110: name = "parameter_110" + original_name = "layer_norm_39.w_0" shape = [320] dtype = "float32" min_val = float("0.200039") @@ -1193,6 +1304,7 @@ class Program_weight_tensor_parameter_110: class Program_weight_tensor_parameter_111: name = "parameter_111" + original_name = "linear_54.b_0" shape = [320] dtype = "float32" min_val = float("-0.742419") @@ -1204,6 +1316,7 @@ class Program_weight_tensor_parameter_111: class Program_weight_tensor_parameter_112: name = "parameter_112" + original_name = "linear_54.w_0" shape = [1280, 320] dtype = "float32" min_val = float("-0.615131") @@ -1215,6 +1328,7 @@ class Program_weight_tensor_parameter_112: class Program_weight_tensor_parameter_113: name = "parameter_113" + original_name = "conv2d_25.b_0" shape = [1280] dtype = "float32" min_val = float("-1.89819") @@ -1226,6 +1340,7 @@ class Program_weight_tensor_parameter_113: class Program_weight_tensor_parameter_114: name = "parameter_114" + original_name = "conv2d_25.w_0" shape = [1280, 1, 3, 3] dtype = "float32" min_val = float("-0.327521") @@ -1237,6 +1352,7 @@ class Program_weight_tensor_parameter_114: class Program_weight_tensor_parameter_115: name = "parameter_115" + original_name = "linear_53.b_0" shape = [1280] dtype = "float32" min_val = float("-2.08756") @@ -1248,6 +1364,7 @@ class Program_weight_tensor_parameter_115: class Program_weight_tensor_parameter_116: name = "parameter_116" + original_name = "linear_53.w_0" shape = [320, 1280] dtype = "float32" min_val = float("-0.294655") @@ -1259,6 +1376,7 @@ class Program_weight_tensor_parameter_116: class Program_weight_tensor_parameter_117: name = "parameter_117" + original_name = "layer_norm_38.b_0" shape = [320] dtype = "float32" min_val = float("-2.59604") @@ -1270,6 +1388,7 @@ class Program_weight_tensor_parameter_117: class Program_weight_tensor_parameter_118: name = "parameter_118" + original_name = "layer_norm_38.w_0" shape = [320] dtype = "float32" min_val = float("0.668575") @@ -1281,6 +1400,7 @@ class Program_weight_tensor_parameter_118: class Program_weight_tensor_parameter_119: name = "parameter_119" + original_name = "linear_52.b_0" shape = [320] dtype = "float32" min_val = float("-0.315773") @@ -1292,6 +1412,7 @@ class Program_weight_tensor_parameter_119: class Program_weight_tensor_parameter_120: name = "parameter_120" + original_name = "linear_52.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.357638") @@ -1303,6 +1424,7 @@ class Program_weight_tensor_parameter_120: class Program_weight_tensor_parameter_121: name = "parameter_121" + original_name = "linear_51.b_0" shape = [640] dtype = "float32" min_val = float("-0.900104") @@ -1314,6 +1436,7 @@ class Program_weight_tensor_parameter_121: class Program_weight_tensor_parameter_122: name = "parameter_122" + original_name = "linear_51.w_0" shape = [320, 640] dtype = "float32" min_val = float("-0.352983") @@ -1325,6 +1448,7 @@ class Program_weight_tensor_parameter_122: class Program_weight_tensor_parameter_123: name = "parameter_123" + original_name = "layer_norm_37.b_0" shape = [320] dtype = "float32" min_val = float("-0.21016") @@ -1336,6 +1460,7 @@ class Program_weight_tensor_parameter_123: class Program_weight_tensor_parameter_124: name = "parameter_124" + original_name = "layer_norm_37.w_0" shape = [320] dtype = "float32" min_val = float("-0.004402") @@ -1347,6 +1472,7 @@ class Program_weight_tensor_parameter_124: class Program_weight_tensor_parameter_125: name = "parameter_125" + original_name = "conv2d_24.b_0" shape = [320] dtype = "float32" min_val = float("-5.38347") @@ -1358,6 +1484,7 @@ class Program_weight_tensor_parameter_125: class Program_weight_tensor_parameter_126: name = "parameter_126" + original_name = "conv2d_24.w_0" shape = [320, 320, 2, 2] dtype = "float32" min_val = float("-0.39425") @@ -1369,6 +1496,7 @@ class Program_weight_tensor_parameter_126: class Program_weight_tensor_parameter_127: name = "parameter_127" + original_name = "linear_50.b_0" shape = [320] dtype = "float32" min_val = float("-3.10566") @@ -1380,6 +1508,7 @@ class Program_weight_tensor_parameter_127: class Program_weight_tensor_parameter_128: name = "parameter_128" + original_name = "linear_50.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.240393") @@ -1391,6 +1520,7 @@ class Program_weight_tensor_parameter_128: class Program_weight_tensor_parameter_129: name = "parameter_129" + original_name = "layer_norm_36.b_0" shape = [320] dtype = "float32" min_val = float("-1.20044") @@ -1402,6 +1532,7 @@ class Program_weight_tensor_parameter_129: class Program_weight_tensor_parameter_130: name = "parameter_130" + original_name = "layer_norm_36.w_0" shape = [320] dtype = "float32" min_val = float("0.276933") @@ -1413,6 +1544,7 @@ class Program_weight_tensor_parameter_130: class Program_weight_tensor_parameter_131: name = "parameter_131" + original_name = "linear_49.b_0" shape = [320] dtype = "float32" min_val = float("-0.688461") @@ -1424,6 +1556,7 @@ class Program_weight_tensor_parameter_131: class Program_weight_tensor_parameter_132: name = "parameter_132" + original_name = "linear_49.w_0" shape = [1280, 320] dtype = "float32" min_val = float("-0.615715") @@ -1435,6 +1568,7 @@ class Program_weight_tensor_parameter_132: class Program_weight_tensor_parameter_133: name = "parameter_133" + original_name = "conv2d_23.b_0" shape = [1280] dtype = "float32" min_val = float("-1.63361") @@ -1446,6 +1580,7 @@ class Program_weight_tensor_parameter_133: class Program_weight_tensor_parameter_134: name = "parameter_134" + original_name = "conv2d_23.w_0" shape = [1280, 1, 3, 3] dtype = "float32" min_val = float("-0.374118") @@ -1457,6 +1592,7 @@ class Program_weight_tensor_parameter_134: class Program_weight_tensor_parameter_135: name = "parameter_135" + original_name = "linear_48.b_0" shape = [1280] dtype = "float32" min_val = float("-2.23482") @@ -1468,6 +1604,7 @@ class Program_weight_tensor_parameter_135: class Program_weight_tensor_parameter_136: name = "parameter_136" + original_name = "linear_48.w_0" shape = [320, 1280] dtype = "float32" min_val = float("-0.286591") @@ -1479,6 +1616,7 @@ class Program_weight_tensor_parameter_136: class Program_weight_tensor_parameter_137: name = "parameter_137" + original_name = "layer_norm_35.b_0" shape = [320] dtype = "float32" min_val = float("-2.72044") @@ -1490,6 +1628,7 @@ class Program_weight_tensor_parameter_137: class Program_weight_tensor_parameter_138: name = "parameter_138" + original_name = "layer_norm_35.w_0" shape = [320] dtype = "float32" min_val = float("0.794303") @@ -1501,6 +1640,7 @@ class Program_weight_tensor_parameter_138: class Program_weight_tensor_parameter_139: name = "parameter_139" + original_name = "linear_47.b_0" shape = [320] dtype = "float32" min_val = float("-0.235514") @@ -1512,6 +1652,7 @@ class Program_weight_tensor_parameter_139: class Program_weight_tensor_parameter_140: name = "parameter_140" + original_name = "linear_47.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.302127") @@ -1523,6 +1664,7 @@ class Program_weight_tensor_parameter_140: class Program_weight_tensor_parameter_141: name = "parameter_141" + original_name = "linear_46.b_0" shape = [640] dtype = "float32" min_val = float("-0.749387") @@ -1534,6 +1676,7 @@ class Program_weight_tensor_parameter_141: class Program_weight_tensor_parameter_142: name = "parameter_142" + original_name = "linear_46.w_0" shape = [320, 640] dtype = "float32" min_val = float("-0.357105") @@ -1545,6 +1688,7 @@ class Program_weight_tensor_parameter_142: class Program_weight_tensor_parameter_143: name = "parameter_143" + original_name = "layer_norm_34.b_0" shape = [320] dtype = "float32" min_val = float("-0.278376") @@ -1556,6 +1700,7 @@ class Program_weight_tensor_parameter_143: class Program_weight_tensor_parameter_144: name = "parameter_144" + original_name = "layer_norm_34.w_0" shape = [320] dtype = "float32" min_val = float("-0.0020496") @@ -1567,6 +1712,7 @@ class Program_weight_tensor_parameter_144: class Program_weight_tensor_parameter_145: name = "parameter_145" + original_name = "conv2d_22.b_0" shape = [320] dtype = "float32" min_val = float("-6.27016") @@ -1578,6 +1724,7 @@ class Program_weight_tensor_parameter_145: class Program_weight_tensor_parameter_146: name = "parameter_146" + original_name = "conv2d_22.w_0" shape = [320, 320, 2, 2] dtype = "float32" min_val = float("-0.397881") @@ -1589,6 +1736,7 @@ class Program_weight_tensor_parameter_146: class Program_weight_tensor_parameter_147: name = "parameter_147" + original_name = "linear_45.b_0" shape = [320] dtype = "float32" min_val = float("-3.47581") @@ -1600,6 +1748,7 @@ class Program_weight_tensor_parameter_147: class Program_weight_tensor_parameter_148: name = "parameter_148" + original_name = "linear_45.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.266039") @@ -1611,6 +1760,7 @@ class Program_weight_tensor_parameter_148: class Program_weight_tensor_parameter_149: name = "parameter_149" + original_name = "layer_norm_33.b_0" shape = [320] dtype = "float32" min_val = float("-1.28293") @@ -1622,6 +1772,7 @@ class Program_weight_tensor_parameter_149: class Program_weight_tensor_parameter_150: name = "parameter_150" + original_name = "layer_norm_33.w_0" shape = [320] dtype = "float32" min_val = float("0.00312078") @@ -1633,6 +1784,7 @@ class Program_weight_tensor_parameter_150: class Program_weight_tensor_parameter_151: name = "parameter_151" + original_name = "linear_44.b_0" shape = [320] dtype = "float32" min_val = float("-0.693739") @@ -1644,6 +1796,7 @@ class Program_weight_tensor_parameter_151: class Program_weight_tensor_parameter_152: name = "parameter_152" + original_name = "linear_44.w_0" shape = [1280, 320] dtype = "float32" min_val = float("-0.349514") @@ -1655,6 +1808,7 @@ class Program_weight_tensor_parameter_152: class Program_weight_tensor_parameter_153: name = "parameter_153" + original_name = "conv2d_21.b_0" shape = [1280] dtype = "float32" min_val = float("-1.87758") @@ -1666,6 +1820,7 @@ class Program_weight_tensor_parameter_153: class Program_weight_tensor_parameter_154: name = "parameter_154" + original_name = "conv2d_21.w_0" shape = [1280, 1, 3, 3] dtype = "float32" min_val = float("-0.364022") @@ -1677,6 +1832,7 @@ class Program_weight_tensor_parameter_154: class Program_weight_tensor_parameter_155: name = "parameter_155" + original_name = "linear_43.b_0" shape = [1280] dtype = "float32" min_val = float("-4.01862") @@ -1688,6 +1844,7 @@ class Program_weight_tensor_parameter_155: class Program_weight_tensor_parameter_156: name = "parameter_156" + original_name = "linear_43.w_0" shape = [320, 1280] dtype = "float32" min_val = float("-0.331706") @@ -1699,6 +1856,7 @@ class Program_weight_tensor_parameter_156: class Program_weight_tensor_parameter_157: name = "parameter_157" + original_name = "layer_norm_32.b_0" shape = [320] dtype = "float32" min_val = float("-2.97076") @@ -1710,6 +1868,7 @@ class Program_weight_tensor_parameter_157: class Program_weight_tensor_parameter_158: name = "parameter_158" + original_name = "layer_norm_32.w_0" shape = [320] dtype = "float32" min_val = float("0.812703") @@ -1721,6 +1880,7 @@ class Program_weight_tensor_parameter_158: class Program_weight_tensor_parameter_159: name = "parameter_159" + original_name = "linear_42.b_0" shape = [320] dtype = "float32" min_val = float("-0.296969") @@ -1732,6 +1892,7 @@ class Program_weight_tensor_parameter_159: class Program_weight_tensor_parameter_160: name = "parameter_160" + original_name = "linear_42.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.276241") @@ -1743,6 +1904,7 @@ class Program_weight_tensor_parameter_160: class Program_weight_tensor_parameter_161: name = "parameter_161" + original_name = "linear_41.b_0" shape = [640] dtype = "float32" min_val = float("-0.749122") @@ -1754,6 +1916,7 @@ class Program_weight_tensor_parameter_161: class Program_weight_tensor_parameter_162: name = "parameter_162" + original_name = "linear_41.w_0" shape = [320, 640] dtype = "float32" min_val = float("-0.386116") @@ -1765,6 +1928,7 @@ class Program_weight_tensor_parameter_162: class Program_weight_tensor_parameter_163: name = "parameter_163" + original_name = "layer_norm_31.b_0" shape = [320] dtype = "float32" min_val = float("-0.316042") @@ -1776,6 +1940,7 @@ class Program_weight_tensor_parameter_163: class Program_weight_tensor_parameter_164: name = "parameter_164" + original_name = "layer_norm_31.w_0" shape = [320] dtype = "float32" min_val = float("-0.0104858") @@ -1787,6 +1952,7 @@ class Program_weight_tensor_parameter_164: class Program_weight_tensor_parameter_165: name = "parameter_165" + original_name = "conv2d_20.b_0" shape = [320] dtype = "float32" min_val = float("-4.51992") @@ -1798,6 +1964,7 @@ class Program_weight_tensor_parameter_165: class Program_weight_tensor_parameter_166: name = "parameter_166" + original_name = "conv2d_20.w_0" shape = [320, 320, 2, 2] dtype = "float32" min_val = float("-0.304811") @@ -1809,6 +1976,7 @@ class Program_weight_tensor_parameter_166: class Program_weight_tensor_parameter_167: name = "parameter_167" + original_name = "linear_40.b_0" shape = [320] dtype = "float32" min_val = float("-3.0758") @@ -1820,6 +1988,7 @@ class Program_weight_tensor_parameter_167: class Program_weight_tensor_parameter_168: name = "parameter_168" + original_name = "linear_40.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.267154") @@ -1831,6 +2000,7 @@ class Program_weight_tensor_parameter_168: class Program_weight_tensor_parameter_169: name = "parameter_169" + original_name = "layer_norm_30.b_0" shape = [320] dtype = "float32" min_val = float("-1.23931") @@ -1842,6 +2012,7 @@ class Program_weight_tensor_parameter_169: class Program_weight_tensor_parameter_170: name = "parameter_170" + original_name = "layer_norm_30.w_0" shape = [320] dtype = "float32" min_val = float("0.00102178") @@ -1853,6 +2024,7 @@ class Program_weight_tensor_parameter_170: class Program_weight_tensor_parameter_171: name = "parameter_171" + original_name = "linear_39.b_0" shape = [320] dtype = "float32" min_val = float("-0.771293") @@ -1864,6 +2036,7 @@ class Program_weight_tensor_parameter_171: class Program_weight_tensor_parameter_172: name = "parameter_172" + original_name = "linear_39.w_0" shape = [1280, 320] dtype = "float32" min_val = float("-0.393932") @@ -1875,6 +2048,7 @@ class Program_weight_tensor_parameter_172: class Program_weight_tensor_parameter_173: name = "parameter_173" + original_name = "conv2d_19.b_0" shape = [1280] dtype = "float32" min_val = float("-2.46368") @@ -1886,6 +2060,7 @@ class Program_weight_tensor_parameter_173: class Program_weight_tensor_parameter_174: name = "parameter_174" + original_name = "conv2d_19.w_0" shape = [1280, 1, 3, 3] dtype = "float32" min_val = float("-0.284703") @@ -1897,6 +2072,7 @@ class Program_weight_tensor_parameter_174: class Program_weight_tensor_parameter_175: name = "parameter_175" + original_name = "linear_38.b_0" shape = [1280] dtype = "float32" min_val = float("-4.07692") @@ -1908,6 +2084,7 @@ class Program_weight_tensor_parameter_175: class Program_weight_tensor_parameter_176: name = "parameter_176" + original_name = "linear_38.w_0" shape = [320, 1280] dtype = "float32" min_val = float("-0.432058") @@ -1919,6 +2096,7 @@ class Program_weight_tensor_parameter_176: class Program_weight_tensor_parameter_177: name = "parameter_177" + original_name = "layer_norm_29.b_0" shape = [320] dtype = "float32" min_val = float("-2.94165") @@ -1930,6 +2108,7 @@ class Program_weight_tensor_parameter_177: class Program_weight_tensor_parameter_178: name = "parameter_178" + original_name = "layer_norm_29.w_0" shape = [320] dtype = "float32" min_val = float("0.670693") @@ -1941,6 +2120,7 @@ class Program_weight_tensor_parameter_178: class Program_weight_tensor_parameter_179: name = "parameter_179" + original_name = "linear_37.b_0" shape = [320] dtype = "float32" min_val = float("-0.333423") @@ -1952,6 +2132,7 @@ class Program_weight_tensor_parameter_179: class Program_weight_tensor_parameter_180: name = "parameter_180" + original_name = "linear_37.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.23495") @@ -1963,6 +2144,7 @@ class Program_weight_tensor_parameter_180: class Program_weight_tensor_parameter_181: name = "parameter_181" + original_name = "linear_36.b_0" shape = [640] dtype = "float32" min_val = float("-0.832972") @@ -1974,6 +2156,7 @@ class Program_weight_tensor_parameter_181: class Program_weight_tensor_parameter_182: name = "parameter_182" + original_name = "linear_36.w_0" shape = [320, 640] dtype = "float32" min_val = float("-0.326697") @@ -1985,6 +2168,7 @@ class Program_weight_tensor_parameter_182: class Program_weight_tensor_parameter_183: name = "parameter_183" + original_name = "layer_norm_28.b_0" shape = [320] dtype = "float32" min_val = float("-0.216537") @@ -1996,6 +2180,7 @@ class Program_weight_tensor_parameter_183: class Program_weight_tensor_parameter_184: name = "parameter_184" + original_name = "layer_norm_28.w_0" shape = [320] dtype = "float32" min_val = float("-0.0333169") @@ -2007,6 +2192,7 @@ class Program_weight_tensor_parameter_184: class Program_weight_tensor_parameter_185: name = "parameter_185" + original_name = "conv2d_18.b_0" shape = [320] dtype = "float32" min_val = float("-4.94267") @@ -2018,6 +2204,7 @@ class Program_weight_tensor_parameter_185: class Program_weight_tensor_parameter_186: name = "parameter_186" + original_name = "conv2d_18.w_0" shape = [320, 320, 2, 2] dtype = "float32" min_val = float("-0.203247") @@ -2029,6 +2216,7 @@ class Program_weight_tensor_parameter_186: class Program_weight_tensor_parameter_187: name = "parameter_187" + original_name = "linear_35.b_0" shape = [320] dtype = "float32" min_val = float("-2.60083") @@ -2040,6 +2228,7 @@ class Program_weight_tensor_parameter_187: class Program_weight_tensor_parameter_188: name = "parameter_188" + original_name = "linear_35.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.322497") @@ -2051,6 +2240,7 @@ class Program_weight_tensor_parameter_188: class Program_weight_tensor_parameter_189: name = "parameter_189" + original_name = "layer_norm_27.b_0" shape = [320] dtype = "float32" min_val = float("-1.19437") @@ -2062,6 +2252,7 @@ class Program_weight_tensor_parameter_189: class Program_weight_tensor_parameter_190: name = "parameter_190" + original_name = "layer_norm_27.w_0" shape = [320] dtype = "float32" min_val = float("-0.000536069") @@ -2073,6 +2264,7 @@ class Program_weight_tensor_parameter_190: class Program_weight_tensor_parameter_191: name = "parameter_191" + original_name = "layer_norm_2.b_0" shape = [320] dtype = "float32" min_val = float("-0.580686") @@ -2084,6 +2276,7 @@ class Program_weight_tensor_parameter_191: class Program_weight_tensor_parameter_192: name = "parameter_192" + original_name = "layer_norm_2.w_0" shape = [320] dtype = "float32" min_val = float("0.0733955") @@ -2095,6 +2288,7 @@ class Program_weight_tensor_parameter_192: class Program_weight_tensor_parameter_193: name = "parameter_193" + original_name = "conv2d_2.b_0" shape = [320] dtype = "float32" min_val = float("-0.928759") @@ -2106,6 +2300,7 @@ class Program_weight_tensor_parameter_193: class Program_weight_tensor_parameter_194: name = "parameter_194" + original_name = "conv2d_2.w_0" shape = [320, 128, 3, 3] dtype = "float32" min_val = float("-1.05042") @@ -2117,6 +2312,7 @@ class Program_weight_tensor_parameter_194: class Program_weight_tensor_parameter_195: name = "parameter_195" + original_name = "layer_norm_26.b_0" shape = [128] dtype = "float32" min_val = float("-0.328959") @@ -2128,6 +2324,7 @@ class Program_weight_tensor_parameter_195: class Program_weight_tensor_parameter_196: name = "parameter_196" + original_name = "layer_norm_26.w_0" shape = [128] dtype = "float32" min_val = float("0.0824334") @@ -2139,6 +2336,7 @@ class Program_weight_tensor_parameter_196: class Program_weight_tensor_parameter_197: name = "parameter_197" + original_name = "linear_34.b_0" shape = [128] dtype = "float32" min_val = float("-9.9281") @@ -2150,6 +2348,7 @@ class Program_weight_tensor_parameter_197: class Program_weight_tensor_parameter_198: name = "parameter_198" + original_name = "linear_34.w_0" shape = [512, 128] dtype = "float32" min_val = float("-1.99528") @@ -2161,6 +2360,7 @@ class Program_weight_tensor_parameter_198: class Program_weight_tensor_parameter_199: name = "parameter_199" + original_name = "conv2d_17.b_0" shape = [512] dtype = "float32" min_val = float("-1.28683") @@ -2172,6 +2372,7 @@ class Program_weight_tensor_parameter_199: class Program_weight_tensor_parameter_200: name = "parameter_200" + original_name = "conv2d_17.w_0" shape = [512, 1, 3, 3] dtype = "float32" min_val = float("-0.639596") @@ -2183,6 +2384,7 @@ class Program_weight_tensor_parameter_200: class Program_weight_tensor_parameter_201: name = "parameter_201" + original_name = "linear_33.b_0" shape = [512] dtype = "float32" min_val = float("-2.63657") @@ -2194,6 +2396,7 @@ class Program_weight_tensor_parameter_201: class Program_weight_tensor_parameter_202: name = "parameter_202" + original_name = "linear_33.w_0" shape = [128, 512] dtype = "float32" min_val = float("-0.261169") @@ -2205,6 +2408,7 @@ class Program_weight_tensor_parameter_202: class Program_weight_tensor_parameter_203: name = "parameter_203" + original_name = "layer_norm_25.b_0" shape = [128] dtype = "float32" min_val = float("-1.51388") @@ -2216,6 +2420,7 @@ class Program_weight_tensor_parameter_203: class Program_weight_tensor_parameter_204: name = "parameter_204" + original_name = "layer_norm_25.w_0" shape = [128] dtype = "float32" min_val = float("1.26803") @@ -2227,6 +2432,7 @@ class Program_weight_tensor_parameter_204: class Program_weight_tensor_parameter_205: name = "parameter_205" + original_name = "linear_32.b_0" shape = [128] dtype = "float32" min_val = float("-0.361277") @@ -2238,6 +2444,7 @@ class Program_weight_tensor_parameter_205: class Program_weight_tensor_parameter_206: name = "parameter_206" + original_name = "linear_32.w_0" shape = [128, 128] dtype = "float32" min_val = float("-0.511442") @@ -2249,6 +2456,7 @@ class Program_weight_tensor_parameter_206: class Program_weight_tensor_parameter_207: name = "parameter_207" + original_name = "linear_31.b_0" shape = [256] dtype = "float32" min_val = float("-0.978002") @@ -2260,6 +2468,7 @@ class Program_weight_tensor_parameter_207: class Program_weight_tensor_parameter_208: name = "parameter_208" + original_name = "linear_31.w_0" shape = [128, 256] dtype = "float32" min_val = float("-0.286234") @@ -2271,6 +2480,7 @@ class Program_weight_tensor_parameter_208: class Program_weight_tensor_parameter_209: name = "parameter_209" + original_name = "layer_norm_24.b_0" shape = [128] dtype = "float32" min_val = float("-0.699238") @@ -2282,6 +2492,7 @@ class Program_weight_tensor_parameter_209: class Program_weight_tensor_parameter_210: name = "parameter_210" + original_name = "layer_norm_24.w_0" shape = [128] dtype = "float32" min_val = float("-0.0324536") @@ -2293,6 +2504,7 @@ class Program_weight_tensor_parameter_210: class Program_weight_tensor_parameter_211: name = "parameter_211" + original_name = "conv2d_16.b_0" shape = [128] dtype = "float32" min_val = float("-3.67294") @@ -2304,6 +2516,7 @@ class Program_weight_tensor_parameter_211: class Program_weight_tensor_parameter_212: name = "parameter_212" + original_name = "conv2d_16.w_0" shape = [128, 128, 4, 4] dtype = "float32" min_val = float("-0.290403") @@ -2315,6 +2528,7 @@ class Program_weight_tensor_parameter_212: class Program_weight_tensor_parameter_213: name = "parameter_213" + original_name = "linear_30.b_0" shape = [128] dtype = "float32" min_val = float("-2.39381") @@ -2326,6 +2540,7 @@ class Program_weight_tensor_parameter_213: class Program_weight_tensor_parameter_214: name = "parameter_214" + original_name = "linear_30.w_0" shape = [128, 128] dtype = "float32" min_val = float("-0.245529") @@ -2337,6 +2552,7 @@ class Program_weight_tensor_parameter_214: class Program_weight_tensor_parameter_215: name = "parameter_215" + original_name = "layer_norm_23.b_0" shape = [128] dtype = "float32" min_val = float("-0.872957") @@ -2348,6 +2564,7 @@ class Program_weight_tensor_parameter_215: class Program_weight_tensor_parameter_216: name = "parameter_216" + original_name = "layer_norm_23.w_0" shape = [128] dtype = "float32" min_val = float("0.456072") @@ -2359,6 +2576,7 @@ class Program_weight_tensor_parameter_216: class Program_weight_tensor_parameter_217: name = "parameter_217" + original_name = "linear_29.b_0" shape = [128] dtype = "float32" min_val = float("-2.20168") @@ -2370,6 +2588,7 @@ class Program_weight_tensor_parameter_217: class Program_weight_tensor_parameter_218: name = "parameter_218" + original_name = "linear_29.w_0" shape = [512, 128] dtype = "float32" min_val = float("-0.748773") @@ -2381,6 +2600,7 @@ class Program_weight_tensor_parameter_218: class Program_weight_tensor_parameter_219: name = "parameter_219" + original_name = "conv2d_15.b_0" shape = [512] dtype = "float32" min_val = float("-1.42284") @@ -2392,6 +2612,7 @@ class Program_weight_tensor_parameter_219: class Program_weight_tensor_parameter_220: name = "parameter_220" + original_name = "conv2d_15.w_0" shape = [512, 1, 3, 3] dtype = "float32" min_val = float("-0.283171") @@ -2403,6 +2624,7 @@ class Program_weight_tensor_parameter_220: class Program_weight_tensor_parameter_221: name = "parameter_221" + original_name = "linear_28.b_0" shape = [512] dtype = "float32" min_val = float("-5.21467") @@ -2414,6 +2636,7 @@ class Program_weight_tensor_parameter_221: class Program_weight_tensor_parameter_222: name = "parameter_222" + original_name = "linear_28.w_0" shape = [128, 512] dtype = "float32" min_val = float("-0.238951") @@ -2425,6 +2648,7 @@ class Program_weight_tensor_parameter_222: class Program_weight_tensor_parameter_223: name = "parameter_223" + original_name = "layer_norm_22.b_0" shape = [128] dtype = "float32" min_val = float("-1.05295") @@ -2436,6 +2660,7 @@ class Program_weight_tensor_parameter_223: class Program_weight_tensor_parameter_224: name = "parameter_224" + original_name = "layer_norm_22.w_0" shape = [128] dtype = "float32" min_val = float("1.60659") @@ -2447,6 +2672,7 @@ class Program_weight_tensor_parameter_224: class Program_weight_tensor_parameter_225: name = "parameter_225" + original_name = "linear_27.b_0" shape = [128] dtype = "float32" min_val = float("-0.122631") @@ -2458,6 +2684,7 @@ class Program_weight_tensor_parameter_225: class Program_weight_tensor_parameter_226: name = "parameter_226" + original_name = "linear_27.w_0" shape = [128, 128] dtype = "float32" min_val = float("-0.303581") @@ -2469,6 +2696,7 @@ class Program_weight_tensor_parameter_226: class Program_weight_tensor_parameter_227: name = "parameter_227" + original_name = "linear_26.b_0" shape = [256] dtype = "float32" min_val = float("-1.16652") @@ -2480,6 +2708,7 @@ class Program_weight_tensor_parameter_227: class Program_weight_tensor_parameter_228: name = "parameter_228" + original_name = "linear_26.w_0" shape = [128, 256] dtype = "float32" min_val = float("-0.315367") @@ -2491,6 +2720,7 @@ class Program_weight_tensor_parameter_228: class Program_weight_tensor_parameter_229: name = "parameter_229" + original_name = "layer_norm_21.b_0" shape = [128] dtype = "float32" min_val = float("-0.343842") @@ -2502,6 +2732,7 @@ class Program_weight_tensor_parameter_229: class Program_weight_tensor_parameter_230: name = "parameter_230" + original_name = "layer_norm_21.w_0" shape = [128] dtype = "float32" min_val = float("-0.00558591") @@ -2513,6 +2744,7 @@ class Program_weight_tensor_parameter_230: class Program_weight_tensor_parameter_231: name = "parameter_231" + original_name = "conv2d_14.b_0" shape = [128] dtype = "float32" min_val = float("-4.47848") @@ -2524,6 +2756,7 @@ class Program_weight_tensor_parameter_231: class Program_weight_tensor_parameter_232: name = "parameter_232" + original_name = "conv2d_14.w_0" shape = [128, 128, 4, 4] dtype = "float32" min_val = float("-0.312117") @@ -2535,6 +2768,7 @@ class Program_weight_tensor_parameter_232: class Program_weight_tensor_parameter_233: name = "parameter_233" + original_name = "linear_25.b_0" shape = [128] dtype = "float32" min_val = float("-3.64924") @@ -2546,6 +2780,7 @@ class Program_weight_tensor_parameter_233: class Program_weight_tensor_parameter_234: name = "parameter_234" + original_name = "linear_25.w_0" shape = [128, 128] dtype = "float32" min_val = float("-0.260143") @@ -2557,6 +2792,7 @@ class Program_weight_tensor_parameter_234: class Program_weight_tensor_parameter_235: name = "parameter_235" + original_name = "layer_norm_20.b_0" shape = [128] dtype = "float32" min_val = float("-1.29729") @@ -2568,6 +2804,7 @@ class Program_weight_tensor_parameter_235: class Program_weight_tensor_parameter_236: name = "parameter_236" + original_name = "layer_norm_20.w_0" shape = [128] dtype = "float32" min_val = float("0.682266") @@ -2579,6 +2816,7 @@ class Program_weight_tensor_parameter_236: class Program_weight_tensor_parameter_237: name = "parameter_237" + original_name = "linear_24.b_0" shape = [128] dtype = "float32" min_val = float("-1.17729") @@ -2590,6 +2828,7 @@ class Program_weight_tensor_parameter_237: class Program_weight_tensor_parameter_238: name = "parameter_238" + original_name = "linear_24.w_0" shape = [512, 128] dtype = "float32" min_val = float("-0.751349") @@ -2601,6 +2840,7 @@ class Program_weight_tensor_parameter_238: class Program_weight_tensor_parameter_239: name = "parameter_239" + original_name = "conv2d_13.b_0" shape = [512] dtype = "float32" min_val = float("-2.06389") @@ -2612,6 +2852,7 @@ class Program_weight_tensor_parameter_239: class Program_weight_tensor_parameter_240: name = "parameter_240" + original_name = "conv2d_13.w_0" shape = [512, 1, 3, 3] dtype = "float32" min_val = float("-0.267601") @@ -2623,6 +2864,7 @@ class Program_weight_tensor_parameter_240: class Program_weight_tensor_parameter_241: name = "parameter_241" + original_name = "linear_23.b_0" shape = [512] dtype = "float32" min_val = float("-3.38259") @@ -2634,6 +2876,7 @@ class Program_weight_tensor_parameter_241: class Program_weight_tensor_parameter_242: name = "parameter_242" + original_name = "linear_23.w_0" shape = [128, 512] dtype = "float32" min_val = float("-0.360548") @@ -2645,6 +2888,7 @@ class Program_weight_tensor_parameter_242: class Program_weight_tensor_parameter_243: name = "parameter_243" + original_name = "layer_norm_19.b_0" shape = [128] dtype = "float32" min_val = float("-1.34105") @@ -2656,6 +2900,7 @@ class Program_weight_tensor_parameter_243: class Program_weight_tensor_parameter_244: name = "parameter_244" + original_name = "layer_norm_19.w_0" shape = [128] dtype = "float32" min_val = float("1.80374") @@ -2667,6 +2912,7 @@ class Program_weight_tensor_parameter_244: class Program_weight_tensor_parameter_245: name = "parameter_245" + original_name = "linear_22.b_0" shape = [128] dtype = "float32" min_val = float("-0.123908") @@ -2678,6 +2924,7 @@ class Program_weight_tensor_parameter_245: class Program_weight_tensor_parameter_246: name = "parameter_246" + original_name = "linear_22.w_0" shape = [128, 128] dtype = "float32" min_val = float("-0.243484") @@ -2689,6 +2936,7 @@ class Program_weight_tensor_parameter_246: class Program_weight_tensor_parameter_247: name = "parameter_247" + original_name = "linear_21.b_0" shape = [256] dtype = "float32" min_val = float("-1.33269") @@ -2700,6 +2948,7 @@ class Program_weight_tensor_parameter_247: class Program_weight_tensor_parameter_248: name = "parameter_248" + original_name = "linear_21.w_0" shape = [128, 256] dtype = "float32" min_val = float("-0.455009") @@ -2711,6 +2960,7 @@ class Program_weight_tensor_parameter_248: class Program_weight_tensor_parameter_249: name = "parameter_249" + original_name = "layer_norm_18.b_0" shape = [128] dtype = "float32" min_val = float("-0.2684") @@ -2722,6 +2972,7 @@ class Program_weight_tensor_parameter_249: class Program_weight_tensor_parameter_250: name = "parameter_250" + original_name = "layer_norm_18.w_0" shape = [128] dtype = "float32" min_val = float("-0.0059907") @@ -2733,6 +2984,7 @@ class Program_weight_tensor_parameter_250: class Program_weight_tensor_parameter_251: name = "parameter_251" + original_name = "conv2d_12.b_0" shape = [128] dtype = "float32" min_val = float("-2.14358") @@ -2744,6 +2996,7 @@ class Program_weight_tensor_parameter_251: class Program_weight_tensor_parameter_252: name = "parameter_252" + original_name = "conv2d_12.w_0" shape = [128, 128, 4, 4] dtype = "float32" min_val = float("-0.281166") @@ -2755,6 +3008,7 @@ class Program_weight_tensor_parameter_252: class Program_weight_tensor_parameter_253: name = "parameter_253" + original_name = "linear_20.b_0" shape = [128] dtype = "float32" min_val = float("-2.56314") @@ -2766,6 +3020,7 @@ class Program_weight_tensor_parameter_253: class Program_weight_tensor_parameter_254: name = "parameter_254" + original_name = "linear_20.w_0" shape = [128, 128] dtype = "float32" min_val = float("-0.337617") @@ -2777,6 +3032,7 @@ class Program_weight_tensor_parameter_254: class Program_weight_tensor_parameter_255: name = "parameter_255" + original_name = "layer_norm_17.b_0" shape = [128] dtype = "float32" min_val = float("-0.591999") @@ -2788,6 +3044,7 @@ class Program_weight_tensor_parameter_255: class Program_weight_tensor_parameter_256: name = "parameter_256" + original_name = "layer_norm_17.w_0" shape = [128] dtype = "float32" min_val = float("0.339968") @@ -2799,6 +3056,7 @@ class Program_weight_tensor_parameter_256: class Program_weight_tensor_parameter_257: name = "parameter_257" + original_name = "linear_19.b_0" shape = [128] dtype = "float32" min_val = float("-0.946922") @@ -2810,6 +3068,7 @@ class Program_weight_tensor_parameter_257: class Program_weight_tensor_parameter_258: name = "parameter_258" + original_name = "linear_19.w_0" shape = [512, 128] dtype = "float32" min_val = float("-0.2982") @@ -2821,6 +3080,7 @@ class Program_weight_tensor_parameter_258: class Program_weight_tensor_parameter_259: name = "parameter_259" + original_name = "conv2d_11.b_0" shape = [512] dtype = "float32" min_val = float("-2.03982") @@ -2832,6 +3092,7 @@ class Program_weight_tensor_parameter_259: class Program_weight_tensor_parameter_260: name = "parameter_260" + original_name = "conv2d_11.w_0" shape = [512, 1, 3, 3] dtype = "float32" min_val = float("-0.295111") @@ -2843,6 +3104,7 @@ class Program_weight_tensor_parameter_260: class Program_weight_tensor_parameter_261: name = "parameter_261" + original_name = "linear_18.b_0" shape = [512] dtype = "float32" min_val = float("-2.94665") @@ -2854,6 +3116,7 @@ class Program_weight_tensor_parameter_261: class Program_weight_tensor_parameter_262: name = "parameter_262" + original_name = "linear_18.w_0" shape = [128, 512] dtype = "float32" min_val = float("-0.311634") @@ -2865,6 +3128,7 @@ class Program_weight_tensor_parameter_262: class Program_weight_tensor_parameter_263: name = "parameter_263" + original_name = "layer_norm_16.b_0" shape = [128] dtype = "float32" min_val = float("-1.14431") @@ -2876,6 +3140,7 @@ class Program_weight_tensor_parameter_263: class Program_weight_tensor_parameter_264: name = "parameter_264" + original_name = "layer_norm_16.w_0" shape = [128] dtype = "float32" min_val = float("1.3384") @@ -2887,6 +3152,7 @@ class Program_weight_tensor_parameter_264: class Program_weight_tensor_parameter_265: name = "parameter_265" + original_name = "linear_17.b_0" shape = [128] dtype = "float32" min_val = float("-0.358703") @@ -2898,6 +3164,7 @@ class Program_weight_tensor_parameter_265: class Program_weight_tensor_parameter_266: name = "parameter_266" + original_name = "linear_17.w_0" shape = [128, 128] dtype = "float32" min_val = float("-0.182639") @@ -2909,6 +3176,7 @@ class Program_weight_tensor_parameter_266: class Program_weight_tensor_parameter_267: name = "parameter_267" + original_name = "linear_16.b_0" shape = [256] dtype = "float32" min_val = float("-1.22553") @@ -2920,6 +3188,7 @@ class Program_weight_tensor_parameter_267: class Program_weight_tensor_parameter_268: name = "parameter_268" + original_name = "linear_16.w_0" shape = [128, 256] dtype = "float32" min_val = float("-0.408316") @@ -2931,6 +3200,7 @@ class Program_weight_tensor_parameter_268: class Program_weight_tensor_parameter_269: name = "parameter_269" + original_name = "layer_norm_15.b_0" shape = [128] dtype = "float32" min_val = float("-0.30067") @@ -2942,6 +3212,7 @@ class Program_weight_tensor_parameter_269: class Program_weight_tensor_parameter_270: name = "parameter_270" + original_name = "layer_norm_15.w_0" shape = [128] dtype = "float32" min_val = float("-0.0179247") @@ -2953,6 +3224,7 @@ class Program_weight_tensor_parameter_270: class Program_weight_tensor_parameter_271: name = "parameter_271" + original_name = "conv2d_10.b_0" shape = [128] dtype = "float32" min_val = float("-3.02983") @@ -2964,6 +3236,7 @@ class Program_weight_tensor_parameter_271: class Program_weight_tensor_parameter_272: name = "parameter_272" + original_name = "conv2d_10.w_0" shape = [128, 128, 4, 4] dtype = "float32" min_val = float("-0.299401") @@ -2975,6 +3248,7 @@ class Program_weight_tensor_parameter_272: class Program_weight_tensor_parameter_273: name = "parameter_273" + original_name = "linear_15.b_0" shape = [128] dtype = "float32" min_val = float("-2.16972") @@ -2986,6 +3260,7 @@ class Program_weight_tensor_parameter_273: class Program_weight_tensor_parameter_274: name = "parameter_274" + original_name = "linear_15.w_0" shape = [128, 128] dtype = "float32" min_val = float("-0.379632") @@ -2997,6 +3272,7 @@ class Program_weight_tensor_parameter_274: class Program_weight_tensor_parameter_275: name = "parameter_275" + original_name = "layer_norm_14.b_0" shape = [128] dtype = "float32" min_val = float("-0.768707") @@ -3008,6 +3284,7 @@ class Program_weight_tensor_parameter_275: class Program_weight_tensor_parameter_276: name = "parameter_276" + original_name = "layer_norm_14.w_0" shape = [128] dtype = "float32" min_val = float("-0.000327386") @@ -3019,6 +3296,7 @@ class Program_weight_tensor_parameter_276: class Program_weight_tensor_parameter_277: name = "parameter_277" + original_name = "layer_norm_1.b_0" shape = [128] dtype = "float32" min_val = float("-0.438389") @@ -3030,6 +3308,7 @@ class Program_weight_tensor_parameter_277: class Program_weight_tensor_parameter_278: name = "parameter_278" + original_name = "layer_norm_1.w_0" shape = [128] dtype = "float32" min_val = float("0.36418") @@ -3041,6 +3320,7 @@ class Program_weight_tensor_parameter_278: class Program_weight_tensor_parameter_279: name = "parameter_279" + original_name = "conv2d_1.b_0" shape = [128] dtype = "float32" min_val = float("-3.80341") @@ -3052,6 +3332,7 @@ class Program_weight_tensor_parameter_279: class Program_weight_tensor_parameter_280: name = "parameter_280" + original_name = "conv2d_1.w_0" shape = [128, 64, 3, 3] dtype = "float32" min_val = float("-0.308809") @@ -3063,6 +3344,7 @@ class Program_weight_tensor_parameter_280: class Program_weight_tensor_parameter_281: name = "parameter_281" + original_name = "layer_norm_13.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -3072,6 +3354,7 @@ class Program_weight_tensor_parameter_281: class Program_weight_tensor_parameter_282: name = "parameter_282" + original_name = "layer_norm_13.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -3081,6 +3364,7 @@ class Program_weight_tensor_parameter_282: class Program_weight_tensor_parameter_283: name = "parameter_283" + original_name = "linear_14.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -3090,6 +3374,7 @@ class Program_weight_tensor_parameter_283: class Program_weight_tensor_parameter_284: name = "parameter_284" + original_name = "linear_14.w_0" shape = [256, 64] dtype = "float32" min_val = float("-0.65183") @@ -3101,6 +3386,7 @@ class Program_weight_tensor_parameter_284: class Program_weight_tensor_parameter_285: name = "parameter_285" + original_name = "conv2d_9.b_0" shape = [256] dtype = "float32" min_val = float("-0.948557") @@ -3112,6 +3398,7 @@ class Program_weight_tensor_parameter_285: class Program_weight_tensor_parameter_286: name = "parameter_286" + original_name = "conv2d_9.w_0" shape = [256, 1, 3, 3] dtype = "float32" min_val = float("-0.331579") @@ -3123,6 +3410,7 @@ class Program_weight_tensor_parameter_286: class Program_weight_tensor_parameter_287: name = "parameter_287" + original_name = "linear_13.b_0" shape = [256] dtype = "float32" min_val = float("-4.10082") @@ -3134,6 +3422,7 @@ class Program_weight_tensor_parameter_287: class Program_weight_tensor_parameter_288: name = "parameter_288" + original_name = "linear_13.w_0" shape = [64, 256] dtype = "float32" min_val = float("-0.384294") @@ -3145,6 +3434,7 @@ class Program_weight_tensor_parameter_288: class Program_weight_tensor_parameter_289: name = "parameter_289" + original_name = "layer_norm_12.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -3154,6 +3444,7 @@ class Program_weight_tensor_parameter_289: class Program_weight_tensor_parameter_290: name = "parameter_290" + original_name = "layer_norm_12.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -3163,6 +3454,7 @@ class Program_weight_tensor_parameter_290: class Program_weight_tensor_parameter_291: name = "parameter_291" + original_name = "linear_12.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -3172,6 +3464,7 @@ class Program_weight_tensor_parameter_291: class Program_weight_tensor_parameter_292: name = "parameter_292" + original_name = "linear_12.w_0" shape = [64, 64] dtype = "float32" min_val = float("-0.435159") @@ -3183,6 +3476,7 @@ class Program_weight_tensor_parameter_292: class Program_weight_tensor_parameter_293: name = "parameter_293" + original_name = "linear_11.b_0" shape = [128] dtype = "float32" min_val = float("-1.53412") @@ -3194,6 +3488,7 @@ class Program_weight_tensor_parameter_293: class Program_weight_tensor_parameter_294: name = "parameter_294" + original_name = "linear_11.w_0" shape = [64, 128] dtype = "float32" min_val = float("-0.531219") @@ -3205,6 +3500,7 @@ class Program_weight_tensor_parameter_294: class Program_weight_tensor_parameter_295: name = "parameter_295" + original_name = "layer_norm_11.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -3214,6 +3510,7 @@ class Program_weight_tensor_parameter_295: class Program_weight_tensor_parameter_296: name = "parameter_296" + original_name = "layer_norm_11.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -3223,6 +3520,7 @@ class Program_weight_tensor_parameter_296: class Program_weight_tensor_parameter_297: name = "parameter_297" + original_name = "conv2d_8.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -3232,6 +3530,7 @@ class Program_weight_tensor_parameter_297: class Program_weight_tensor_parameter_298: name = "parameter_298" + original_name = "conv2d_8.w_0" shape = [64, 64, 8, 8] dtype = "float32" min_val = float("-0.412526") @@ -3243,6 +3542,7 @@ class Program_weight_tensor_parameter_298: class Program_weight_tensor_parameter_299: name = "parameter_299" + original_name = "linear_10.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -3252,6 +3552,7 @@ class Program_weight_tensor_parameter_299: class Program_weight_tensor_parameter_300: name = "parameter_300" + original_name = "linear_10.w_0" shape = [64, 64] dtype = "float32" min_val = float("-0.41881") @@ -3263,6 +3564,7 @@ class Program_weight_tensor_parameter_300: class Program_weight_tensor_parameter_301: name = "parameter_301" + original_name = "layer_norm_10.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -3272,6 +3574,7 @@ class Program_weight_tensor_parameter_301: class Program_weight_tensor_parameter_302: name = "parameter_302" + original_name = "layer_norm_10.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -3281,6 +3584,7 @@ class Program_weight_tensor_parameter_302: class Program_weight_tensor_parameter_303: name = "parameter_303" + original_name = "linear_9.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -3290,6 +3594,7 @@ class Program_weight_tensor_parameter_303: class Program_weight_tensor_parameter_304: name = "parameter_304" + original_name = "linear_9.w_0" shape = [256, 64] dtype = "float32" min_val = float("-0.396406") @@ -3301,6 +3606,7 @@ class Program_weight_tensor_parameter_304: class Program_weight_tensor_parameter_305: name = "parameter_305" + original_name = "conv2d_7.b_0" shape = [256] dtype = "float32" min_val = float("-1.65761") @@ -3312,6 +3618,7 @@ class Program_weight_tensor_parameter_305: class Program_weight_tensor_parameter_306: name = "parameter_306" + original_name = "conv2d_7.w_0" shape = [256, 1, 3, 3] dtype = "float32" min_val = float("-0.41694") @@ -3323,6 +3630,7 @@ class Program_weight_tensor_parameter_306: class Program_weight_tensor_parameter_307: name = "parameter_307" + original_name = "linear_8.b_0" shape = [256] dtype = "float32" min_val = float("-4.72273") @@ -3334,6 +3642,7 @@ class Program_weight_tensor_parameter_307: class Program_weight_tensor_parameter_308: name = "parameter_308" + original_name = "linear_8.w_0" shape = [64, 256] dtype = "float32" min_val = float("-0.409124") @@ -3345,6 +3654,7 @@ class Program_weight_tensor_parameter_308: class Program_weight_tensor_parameter_309: name = "parameter_309" + original_name = "layer_norm_9.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -3354,6 +3664,7 @@ class Program_weight_tensor_parameter_309: class Program_weight_tensor_parameter_310: name = "parameter_310" + original_name = "layer_norm_9.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -3363,6 +3674,7 @@ class Program_weight_tensor_parameter_310: class Program_weight_tensor_parameter_311: name = "parameter_311" + original_name = "linear_7.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -3372,6 +3684,7 @@ class Program_weight_tensor_parameter_311: class Program_weight_tensor_parameter_312: name = "parameter_312" + original_name = "linear_7.w_0" shape = [64, 64] dtype = "float32" min_val = float("-0.154474") @@ -3383,6 +3696,7 @@ class Program_weight_tensor_parameter_312: class Program_weight_tensor_parameter_313: name = "parameter_313" + original_name = "linear_6.b_0" shape = [128] dtype = "float32" min_val = float("-0.978147") @@ -3394,6 +3708,7 @@ class Program_weight_tensor_parameter_313: class Program_weight_tensor_parameter_314: name = "parameter_314" + original_name = "linear_6.w_0" shape = [64, 128] dtype = "float32" min_val = float("-0.539761") @@ -3405,6 +3720,7 @@ class Program_weight_tensor_parameter_314: class Program_weight_tensor_parameter_315: name = "parameter_315" + original_name = "layer_norm_8.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -3414,6 +3730,7 @@ class Program_weight_tensor_parameter_315: class Program_weight_tensor_parameter_316: name = "parameter_316" + original_name = "layer_norm_8.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -3423,6 +3740,7 @@ class Program_weight_tensor_parameter_316: class Program_weight_tensor_parameter_317: name = "parameter_317" + original_name = "conv2d_6.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -3432,6 +3750,7 @@ class Program_weight_tensor_parameter_317: class Program_weight_tensor_parameter_318: name = "parameter_318" + original_name = "conv2d_6.w_0" shape = [64, 64, 8, 8] dtype = "float32" min_val = float("-0.38175") @@ -3443,6 +3762,7 @@ class Program_weight_tensor_parameter_318: class Program_weight_tensor_parameter_319: name = "parameter_319" + original_name = "linear_5.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -3452,6 +3772,7 @@ class Program_weight_tensor_parameter_319: class Program_weight_tensor_parameter_320: name = "parameter_320" + original_name = "linear_5.w_0" shape = [64, 64] dtype = "float32" min_val = float("-0.433176") @@ -3463,6 +3784,7 @@ class Program_weight_tensor_parameter_320: class Program_weight_tensor_parameter_321: name = "parameter_321" + original_name = "layer_norm_7.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -3472,6 +3794,7 @@ class Program_weight_tensor_parameter_321: class Program_weight_tensor_parameter_322: name = "parameter_322" + original_name = "layer_norm_7.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -3481,6 +3804,7 @@ class Program_weight_tensor_parameter_322: class Program_weight_tensor_parameter_323: name = "parameter_323" + original_name = "linear_4.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -3490,6 +3814,7 @@ class Program_weight_tensor_parameter_323: class Program_weight_tensor_parameter_324: name = "parameter_324" + original_name = "linear_4.w_0" shape = [256, 64] dtype = "float32" min_val = float("-0.282121") @@ -3501,6 +3826,7 @@ class Program_weight_tensor_parameter_324: class Program_weight_tensor_parameter_325: name = "parameter_325" + original_name = "conv2d_5.b_0" shape = [256] dtype = "float32" min_val = float("-1.24102") @@ -3512,6 +3838,7 @@ class Program_weight_tensor_parameter_325: class Program_weight_tensor_parameter_326: name = "parameter_326" + original_name = "conv2d_5.w_0" shape = [256, 1, 3, 3] dtype = "float32" min_val = float("-0.66786") @@ -3523,6 +3850,7 @@ class Program_weight_tensor_parameter_326: class Program_weight_tensor_parameter_327: name = "parameter_327" + original_name = "linear_3.b_0" shape = [256] dtype = "float32" min_val = float("-3.62173") @@ -3534,6 +3862,7 @@ class Program_weight_tensor_parameter_327: class Program_weight_tensor_parameter_328: name = "parameter_328" + original_name = "linear_3.w_0" shape = [64, 256] dtype = "float32" min_val = float("-0.556731") @@ -3545,6 +3874,7 @@ class Program_weight_tensor_parameter_328: class Program_weight_tensor_parameter_329: name = "parameter_329" + original_name = "layer_norm_6.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -3554,6 +3884,7 @@ class Program_weight_tensor_parameter_329: class Program_weight_tensor_parameter_330: name = "parameter_330" + original_name = "layer_norm_6.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -3563,6 +3894,7 @@ class Program_weight_tensor_parameter_330: class Program_weight_tensor_parameter_331: name = "parameter_331" + original_name = "linear_2.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -3572,6 +3904,7 @@ class Program_weight_tensor_parameter_331: class Program_weight_tensor_parameter_332: name = "parameter_332" + original_name = "linear_2.w_0" shape = [64, 64] dtype = "float32" min_val = float("-0.257225") @@ -3583,6 +3916,7 @@ class Program_weight_tensor_parameter_332: class Program_weight_tensor_parameter_333: name = "parameter_333" + original_name = "linear_1.b_0" shape = [128] dtype = "float32" min_val = float("-0.555134") @@ -3594,6 +3928,7 @@ class Program_weight_tensor_parameter_333: class Program_weight_tensor_parameter_334: name = "parameter_334" + original_name = "linear_1.w_0" shape = [64, 128] dtype = "float32" min_val = float("-0.470537") @@ -3605,6 +3940,7 @@ class Program_weight_tensor_parameter_334: class Program_weight_tensor_parameter_335: name = "parameter_335" + original_name = "layer_norm_5.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -3614,6 +3950,7 @@ class Program_weight_tensor_parameter_335: class Program_weight_tensor_parameter_336: name = "parameter_336" + original_name = "layer_norm_5.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -3623,6 +3960,7 @@ class Program_weight_tensor_parameter_336: class Program_weight_tensor_parameter_337: name = "parameter_337" + original_name = "conv2d_4.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -3632,6 +3970,7 @@ class Program_weight_tensor_parameter_337: class Program_weight_tensor_parameter_338: name = "parameter_338" + original_name = "conv2d_4.w_0" shape = [64, 64, 8, 8] dtype = "float32" min_val = float("-0.87924") @@ -3643,6 +3982,7 @@ class Program_weight_tensor_parameter_338: class Program_weight_tensor_parameter_339: name = "parameter_339" + original_name = "linear_0.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -3652,6 +3992,7 @@ class Program_weight_tensor_parameter_339: class Program_weight_tensor_parameter_340: name = "parameter_340" + original_name = "linear_0.w_0" shape = [64, 64] dtype = "float32" min_val = float("-0.496433") @@ -3663,6 +4004,7 @@ class Program_weight_tensor_parameter_340: class Program_weight_tensor_parameter_341: name = "parameter_341" + original_name = "layer_norm_4.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -3672,6 +4014,7 @@ class Program_weight_tensor_parameter_341: class Program_weight_tensor_parameter_342: name = "parameter_342" + original_name = "layer_norm_4.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -3681,6 +4024,7 @@ class Program_weight_tensor_parameter_342: class Program_weight_tensor_parameter_343: name = "parameter_343" + original_name = "layer_norm_0.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -3690,6 +4034,7 @@ class Program_weight_tensor_parameter_343: class Program_weight_tensor_parameter_344: name = "parameter_344" + original_name = "layer_norm_0.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -3699,6 +4044,7 @@ class Program_weight_tensor_parameter_344: class Program_weight_tensor_parameter_345: name = "parameter_345" + original_name = "conv2d_0.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -3708,6 +4054,7 @@ class Program_weight_tensor_parameter_345: class Program_weight_tensor_parameter_346: name = "parameter_346" + original_name = "conv2d_0.w_0" shape = [64, 3, 7, 7] dtype = "float32" min_val = float("-0.433343") diff --git a/paddle_samples/PaddleX/SegFormer-B3/subgraph_0/input_meta.py b/paddle_samples/PaddleX/SegFormer-B3/subgraph_0/input_meta.py index 0b589f11e..a1c2f1891 100644 --- a/paddle_samples/PaddleX/SegFormer-B3/subgraph_0/input_meta.py +++ b/paddle_samples/PaddleX/SegFormer-B3/subgraph_0/input_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_data_0: name = "data_0" + original_name = "var_0" shape = [2, 3, 512, 1024] dtype = "float32" min_val = float("-1.0") diff --git a/paddle_samples/PaddleX/SegFormer-B3/subgraph_0/weight_meta.py b/paddle_samples/PaddleX/SegFormer-B3/subgraph_0/weight_meta.py index d29d5f7d8..f4454edae 100644 --- a/paddle_samples/PaddleX/SegFormer-B3/subgraph_0/weight_meta.py +++ b/paddle_samples/PaddleX/SegFormer-B3/subgraph_0/weight_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_parameter_0: name = "parameter_0" + original_name = "conv2d_58.b_0" shape = [2] dtype = "float32" min_val = float("0") @@ -9,6 +10,7 @@ class Program_weight_tensor_parameter_0: class Program_weight_tensor_parameter_1: name = "parameter_1" + original_name = "conv2d_58.w_0" shape = [2, 768, 1, 1] dtype = "float32" min_val = float("-0.143596") @@ -20,6 +22,7 @@ class Program_weight_tensor_parameter_1: class Program_weight_tensor_parameter_2: name = "parameter_2" + original_name = "batch_norm2d_0.b_0" shape = [768] dtype = "float32" data = None @@ -27,6 +30,7 @@ class Program_weight_tensor_parameter_2: class Program_weight_tensor_parameter_3: name = "parameter_3" + original_name = "batch_norm2d_0.w_0" shape = [768] dtype = "float32" min_val = float("1.0") @@ -37,6 +41,7 @@ class Program_weight_tensor_parameter_3: class Program_weight_tensor_parameter_4: name = "parameter_4" + original_name = "batch_norm2d_0.w_2" shape = [768] dtype = "float32" min_val = float("1.0") @@ -47,6 +52,7 @@ class Program_weight_tensor_parameter_4: class Program_weight_tensor_parameter_5: name = "parameter_5" + original_name = "batch_norm2d_0.w_1" shape = [768] dtype = "float32" data = None @@ -54,6 +60,7 @@ class Program_weight_tensor_parameter_5: class Program_weight_tensor_parameter_6: name = "parameter_6" + original_name = "conv2d_57.w_0" shape = [768, 3072, 1, 1] dtype = "float32" min_val = float("-0.132225") @@ -65,6 +72,7 @@ class Program_weight_tensor_parameter_6: class Program_weight_tensor_parameter_7: name = "parameter_7" + original_name = "linear_143.b_0" shape = [768] dtype = "float32" data = None @@ -72,6 +80,7 @@ class Program_weight_tensor_parameter_7: class Program_weight_tensor_parameter_8: name = "parameter_8" + original_name = "linear_143.w_0" shape = [64, 768] dtype = "float32" min_val = float("-0.0849109") @@ -83,6 +92,7 @@ class Program_weight_tensor_parameter_8: class Program_weight_tensor_parameter_9: name = "parameter_9" + original_name = "linear_142.b_0" shape = [768] dtype = "float32" data = None @@ -90,6 +100,7 @@ class Program_weight_tensor_parameter_9: class Program_weight_tensor_parameter_10: name = "parameter_10" + original_name = "linear_142.w_0" shape = [128, 768] dtype = "float32" min_val = float("-0.0818294") @@ -101,6 +112,7 @@ class Program_weight_tensor_parameter_10: class Program_weight_tensor_parameter_11: name = "parameter_11" + original_name = "linear_141.b_0" shape = [768] dtype = "float32" data = None @@ -108,6 +120,7 @@ class Program_weight_tensor_parameter_11: class Program_weight_tensor_parameter_12: name = "parameter_12" + original_name = "linear_141.w_0" shape = [320, 768] dtype = "float32" min_val = float("-0.0742604") @@ -119,6 +132,7 @@ class Program_weight_tensor_parameter_12: class Program_weight_tensor_parameter_13: name = "parameter_13" + original_name = "linear_140.b_0" shape = [768] dtype = "float32" data = None @@ -126,6 +140,7 @@ class Program_weight_tensor_parameter_13: class Program_weight_tensor_parameter_14: name = "parameter_14" + original_name = "linear_140.w_0" shape = [512, 768] dtype = "float32" min_val = float("-0.0684652") @@ -137,6 +152,7 @@ class Program_weight_tensor_parameter_14: class Program_weight_tensor_parameter_15: name = "parameter_15" + original_name = "layer_norm_88.b_0" shape = [512] dtype = "float32" min_val = float("-0.963053") @@ -148,6 +164,7 @@ class Program_weight_tensor_parameter_15: class Program_weight_tensor_parameter_16: name = "parameter_16" + original_name = "layer_norm_88.w_0" shape = [512] dtype = "float32" min_val = float("0.00692694") @@ -159,6 +176,7 @@ class Program_weight_tensor_parameter_16: class Program_weight_tensor_parameter_17: name = "parameter_17" + original_name = "linear_139.b_0" shape = [512] dtype = "float32" min_val = float("-3.63898") @@ -170,6 +188,7 @@ class Program_weight_tensor_parameter_17: class Program_weight_tensor_parameter_18: name = "parameter_18" + original_name = "linear_139.w_0" shape = [2048, 512] dtype = "float32" min_val = float("-1.74778") @@ -181,6 +200,7 @@ class Program_weight_tensor_parameter_18: class Program_weight_tensor_parameter_19: name = "parameter_19" + original_name = "conv2d_56.b_0" shape = [2048] dtype = "float32" min_val = float("-2.12714") @@ -192,6 +212,7 @@ class Program_weight_tensor_parameter_19: class Program_weight_tensor_parameter_20: name = "parameter_20" + original_name = "conv2d_56.w_0" shape = [2048, 1, 3, 3] dtype = "float32" min_val = float("-0.298452") @@ -203,6 +224,7 @@ class Program_weight_tensor_parameter_20: class Program_weight_tensor_parameter_21: name = "parameter_21" + original_name = "linear_138.b_0" shape = [2048] dtype = "float32" min_val = float("-2.93982") @@ -214,6 +236,7 @@ class Program_weight_tensor_parameter_21: class Program_weight_tensor_parameter_22: name = "parameter_22" + original_name = "linear_138.w_0" shape = [512, 2048] dtype = "float32" min_val = float("-0.617404") @@ -225,6 +248,7 @@ class Program_weight_tensor_parameter_22: class Program_weight_tensor_parameter_23: name = "parameter_23" + original_name = "layer_norm_87.b_0" shape = [512] dtype = "float32" min_val = float("-1.91398") @@ -236,6 +260,7 @@ class Program_weight_tensor_parameter_23: class Program_weight_tensor_parameter_24: name = "parameter_24" + original_name = "layer_norm_87.w_0" shape = [512] dtype = "float32" min_val = float("0.294649") @@ -247,6 +272,7 @@ class Program_weight_tensor_parameter_24: class Program_weight_tensor_parameter_25: name = "parameter_25" + original_name = "linear_137.b_0" shape = [512] dtype = "float32" min_val = float("-0.752427") @@ -258,6 +284,7 @@ class Program_weight_tensor_parameter_25: class Program_weight_tensor_parameter_26: name = "parameter_26" + original_name = "linear_137.w_0" shape = [512, 512] dtype = "float32" min_val = float("-0.706209") @@ -269,6 +296,7 @@ class Program_weight_tensor_parameter_26: class Program_weight_tensor_parameter_27: name = "parameter_27" + original_name = "linear_136.b_0" shape = [1024] dtype = "float32" min_val = float("-1.16606") @@ -280,6 +308,7 @@ class Program_weight_tensor_parameter_27: class Program_weight_tensor_parameter_28: name = "parameter_28" + original_name = "linear_136.w_0" shape = [512, 1024] dtype = "float32" min_val = float("-0.306008") @@ -291,6 +320,7 @@ class Program_weight_tensor_parameter_28: class Program_weight_tensor_parameter_29: name = "parameter_29" + original_name = "linear_135.b_0" shape = [512] dtype = "float32" min_val = float("-2.37844") @@ -302,6 +332,7 @@ class Program_weight_tensor_parameter_29: class Program_weight_tensor_parameter_30: name = "parameter_30" + original_name = "linear_135.w_0" shape = [512, 512] dtype = "float32" min_val = float("-0.287764") @@ -313,6 +344,7 @@ class Program_weight_tensor_parameter_30: class Program_weight_tensor_parameter_31: name = "parameter_31" + original_name = "layer_norm_86.b_0" shape = [512] dtype = "float32" min_val = float("-1.03049") @@ -324,6 +356,7 @@ class Program_weight_tensor_parameter_31: class Program_weight_tensor_parameter_32: name = "parameter_32" + original_name = "layer_norm_86.w_0" shape = [512] dtype = "float32" min_val = float("0.184856") @@ -335,6 +368,7 @@ class Program_weight_tensor_parameter_32: class Program_weight_tensor_parameter_33: name = "parameter_33" + original_name = "linear_134.b_0" shape = [512] dtype = "float32" min_val = float("-0.869683") @@ -346,6 +380,7 @@ class Program_weight_tensor_parameter_33: class Program_weight_tensor_parameter_34: name = "parameter_34" + original_name = "linear_134.w_0" shape = [2048, 512] dtype = "float32" min_val = float("-2.44937") @@ -357,6 +392,7 @@ class Program_weight_tensor_parameter_34: class Program_weight_tensor_parameter_35: name = "parameter_35" + original_name = "conv2d_55.b_0" shape = [2048] dtype = "float32" min_val = float("-1.00468") @@ -368,6 +404,7 @@ class Program_weight_tensor_parameter_35: class Program_weight_tensor_parameter_36: name = "parameter_36" + original_name = "conv2d_55.w_0" shape = [2048, 1, 3, 3] dtype = "float32" min_val = float("-0.75407") @@ -379,6 +416,7 @@ class Program_weight_tensor_parameter_36: class Program_weight_tensor_parameter_37: name = "parameter_37" + original_name = "linear_133.b_0" shape = [2048] dtype = "float32" min_val = float("-12.727") @@ -390,6 +428,7 @@ class Program_weight_tensor_parameter_37: class Program_weight_tensor_parameter_38: name = "parameter_38" + original_name = "linear_133.w_0" shape = [512, 2048] dtype = "float32" min_val = float("-0.668821") @@ -401,6 +440,7 @@ class Program_weight_tensor_parameter_38: class Program_weight_tensor_parameter_39: name = "parameter_39" + original_name = "layer_norm_85.b_0" shape = [512] dtype = "float32" min_val = float("-2.75671") @@ -412,6 +452,7 @@ class Program_weight_tensor_parameter_39: class Program_weight_tensor_parameter_40: name = "parameter_40" + original_name = "layer_norm_85.w_0" shape = [512] dtype = "float32" min_val = float("0.271118") @@ -423,6 +464,7 @@ class Program_weight_tensor_parameter_40: class Program_weight_tensor_parameter_41: name = "parameter_41" + original_name = "linear_132.b_0" shape = [512] dtype = "float32" min_val = float("-0.641285") @@ -434,6 +476,7 @@ class Program_weight_tensor_parameter_41: class Program_weight_tensor_parameter_42: name = "parameter_42" + original_name = "linear_132.w_0" shape = [512, 512] dtype = "float32" min_val = float("-0.841194") @@ -445,6 +488,7 @@ class Program_weight_tensor_parameter_42: class Program_weight_tensor_parameter_43: name = "parameter_43" + original_name = "linear_131.b_0" shape = [1024] dtype = "float32" min_val = float("-0.888809") @@ -456,6 +500,7 @@ class Program_weight_tensor_parameter_43: class Program_weight_tensor_parameter_44: name = "parameter_44" + original_name = "linear_131.w_0" shape = [512, 1024] dtype = "float32" min_val = float("-0.334751") @@ -467,6 +512,7 @@ class Program_weight_tensor_parameter_44: class Program_weight_tensor_parameter_45: name = "parameter_45" + original_name = "linear_130.b_0" shape = [512] dtype = "float32" min_val = float("-2.57159") @@ -478,6 +524,7 @@ class Program_weight_tensor_parameter_45: class Program_weight_tensor_parameter_46: name = "parameter_46" + original_name = "linear_130.w_0" shape = [512, 512] dtype = "float32" min_val = float("-0.271196") @@ -489,6 +536,7 @@ class Program_weight_tensor_parameter_46: class Program_weight_tensor_parameter_47: name = "parameter_47" + original_name = "layer_norm_84.b_0" shape = [512] dtype = "float32" min_val = float("-0.866785") @@ -500,6 +548,7 @@ class Program_weight_tensor_parameter_47: class Program_weight_tensor_parameter_48: name = "parameter_48" + original_name = "layer_norm_84.w_0" shape = [512] dtype = "float32" min_val = float("0.104234") @@ -511,6 +560,7 @@ class Program_weight_tensor_parameter_48: class Program_weight_tensor_parameter_49: name = "parameter_49" + original_name = "linear_129.b_0" shape = [512] dtype = "float32" min_val = float("-0.975446") @@ -522,6 +572,7 @@ class Program_weight_tensor_parameter_49: class Program_weight_tensor_parameter_50: name = "parameter_50" + original_name = "linear_129.w_0" shape = [2048, 512] dtype = "float32" min_val = float("-2.89467") @@ -533,6 +584,7 @@ class Program_weight_tensor_parameter_50: class Program_weight_tensor_parameter_51: name = "parameter_51" + original_name = "conv2d_54.b_0" shape = [2048] dtype = "float32" min_val = float("-1.62559") @@ -544,6 +596,7 @@ class Program_weight_tensor_parameter_51: class Program_weight_tensor_parameter_52: name = "parameter_52" + original_name = "conv2d_54.w_0" shape = [2048, 1, 3, 3] dtype = "float32" min_val = float("-0.399698") @@ -555,6 +608,7 @@ class Program_weight_tensor_parameter_52: class Program_weight_tensor_parameter_53: name = "parameter_53" + original_name = "linear_128.b_0" shape = [2048] dtype = "float32" min_val = float("-6.80921") @@ -566,6 +620,7 @@ class Program_weight_tensor_parameter_53: class Program_weight_tensor_parameter_54: name = "parameter_54" + original_name = "linear_128.w_0" shape = [512, 2048] dtype = "float32" min_val = float("-0.874816") @@ -577,6 +632,7 @@ class Program_weight_tensor_parameter_54: class Program_weight_tensor_parameter_55: name = "parameter_55" + original_name = "layer_norm_83.b_0" shape = [512] dtype = "float32" min_val = float("-3.29453") @@ -588,6 +644,7 @@ class Program_weight_tensor_parameter_55: class Program_weight_tensor_parameter_56: name = "parameter_56" + original_name = "layer_norm_83.w_0" shape = [512] dtype = "float32" min_val = float("0.0366081") @@ -599,6 +656,7 @@ class Program_weight_tensor_parameter_56: class Program_weight_tensor_parameter_57: name = "parameter_57" + original_name = "linear_127.b_0" shape = [512] dtype = "float32" min_val = float("-0.484838") @@ -610,6 +668,7 @@ class Program_weight_tensor_parameter_57: class Program_weight_tensor_parameter_58: name = "parameter_58" + original_name = "linear_127.w_0" shape = [512, 512] dtype = "float32" min_val = float("-0.357622") @@ -621,6 +680,7 @@ class Program_weight_tensor_parameter_58: class Program_weight_tensor_parameter_59: name = "parameter_59" + original_name = "linear_126.b_0" shape = [1024] dtype = "float32" min_val = float("-0.84463") @@ -632,6 +692,7 @@ class Program_weight_tensor_parameter_59: class Program_weight_tensor_parameter_60: name = "parameter_60" + original_name = "linear_126.w_0" shape = [512, 1024] dtype = "float32" min_val = float("-0.315074") @@ -643,6 +704,7 @@ class Program_weight_tensor_parameter_60: class Program_weight_tensor_parameter_61: name = "parameter_61" + original_name = "linear_125.b_0" shape = [512] dtype = "float32" min_val = float("-2.32301") @@ -654,6 +716,7 @@ class Program_weight_tensor_parameter_61: class Program_weight_tensor_parameter_62: name = "parameter_62" + original_name = "linear_125.w_0" shape = [512, 512] dtype = "float32" min_val = float("-0.357903") @@ -665,6 +728,7 @@ class Program_weight_tensor_parameter_62: class Program_weight_tensor_parameter_63: name = "parameter_63" + original_name = "layer_norm_82.b_0" shape = [512] dtype = "float32" min_val = float("-0.685981") @@ -676,6 +740,7 @@ class Program_weight_tensor_parameter_63: class Program_weight_tensor_parameter_64: name = "parameter_64" + original_name = "layer_norm_82.w_0" shape = [512] dtype = "float32" min_val = float("0.0113185") @@ -687,6 +752,7 @@ class Program_weight_tensor_parameter_64: class Program_weight_tensor_parameter_65: name = "parameter_65" + original_name = "layer_norm_3.b_0" shape = [512] dtype = "float32" min_val = float("-1.34684") @@ -698,6 +764,7 @@ class Program_weight_tensor_parameter_65: class Program_weight_tensor_parameter_66: name = "parameter_66" + original_name = "layer_norm_3.w_0" shape = [512] dtype = "float32" min_val = float("0.0546334") @@ -709,6 +776,7 @@ class Program_weight_tensor_parameter_66: class Program_weight_tensor_parameter_67: name = "parameter_67" + original_name = "conv2d_3.b_0" shape = [512] dtype = "float32" min_val = float("-2.98973") @@ -720,6 +788,7 @@ class Program_weight_tensor_parameter_67: class Program_weight_tensor_parameter_68: name = "parameter_68" + original_name = "conv2d_3.w_0" shape = [512, 320, 3, 3] dtype = "float32" min_val = float("-1.37861") @@ -731,6 +800,7 @@ class Program_weight_tensor_parameter_68: class Program_weight_tensor_parameter_69: name = "parameter_69" + original_name = "layer_norm_81.b_0" shape = [320] dtype = "float32" min_val = float("-1.62758") @@ -742,6 +812,7 @@ class Program_weight_tensor_parameter_69: class Program_weight_tensor_parameter_70: name = "parameter_70" + original_name = "layer_norm_81.w_0" shape = [320] dtype = "float32" min_val = float("0.101732") @@ -753,6 +824,7 @@ class Program_weight_tensor_parameter_70: class Program_weight_tensor_parameter_71: name = "parameter_71" + original_name = "linear_124.b_0" shape = [320] dtype = "float32" min_val = float("-0.286306") @@ -764,6 +836,7 @@ class Program_weight_tensor_parameter_71: class Program_weight_tensor_parameter_72: name = "parameter_72" + original_name = "linear_124.w_0" shape = [1280, 320] dtype = "float32" min_val = float("-1.52306") @@ -775,6 +848,7 @@ class Program_weight_tensor_parameter_72: class Program_weight_tensor_parameter_73: name = "parameter_73" + original_name = "conv2d_53.b_0" shape = [1280] dtype = "float32" min_val = float("-1.80869") @@ -786,6 +860,7 @@ class Program_weight_tensor_parameter_73: class Program_weight_tensor_parameter_74: name = "parameter_74" + original_name = "conv2d_53.w_0" shape = [1280, 1, 3, 3] dtype = "float32" min_val = float("-0.80127") @@ -797,6 +872,7 @@ class Program_weight_tensor_parameter_74: class Program_weight_tensor_parameter_75: name = "parameter_75" + original_name = "linear_123.b_0" shape = [1280] dtype = "float32" min_val = float("-1.27856") @@ -808,6 +884,7 @@ class Program_weight_tensor_parameter_75: class Program_weight_tensor_parameter_76: name = "parameter_76" + original_name = "linear_123.w_0" shape = [320, 1280] dtype = "float32" min_val = float("-0.468957") @@ -819,6 +896,7 @@ class Program_weight_tensor_parameter_76: class Program_weight_tensor_parameter_77: name = "parameter_77" + original_name = "layer_norm_80.b_0" shape = [320] dtype = "float32" min_val = float("-1.72257") @@ -830,6 +908,7 @@ class Program_weight_tensor_parameter_77: class Program_weight_tensor_parameter_78: name = "parameter_78" + original_name = "layer_norm_80.w_0" shape = [320] dtype = "float32" min_val = float("0.143325") @@ -841,6 +920,7 @@ class Program_weight_tensor_parameter_78: class Program_weight_tensor_parameter_79: name = "parameter_79" + original_name = "linear_122.b_0" shape = [320] dtype = "float32" min_val = float("-0.234191") @@ -852,6 +932,7 @@ class Program_weight_tensor_parameter_79: class Program_weight_tensor_parameter_80: name = "parameter_80" + original_name = "linear_122.w_0" shape = [320, 320] dtype = "float32" min_val = float("-1.50077") @@ -863,6 +944,7 @@ class Program_weight_tensor_parameter_80: class Program_weight_tensor_parameter_81: name = "parameter_81" + original_name = "linear_121.b_0" shape = [640] dtype = "float32" min_val = float("-8.28112") @@ -874,6 +956,7 @@ class Program_weight_tensor_parameter_81: class Program_weight_tensor_parameter_82: name = "parameter_82" + original_name = "linear_121.w_0" shape = [320, 640] dtype = "float32" min_val = float("-0.428893") @@ -885,6 +968,7 @@ class Program_weight_tensor_parameter_82: class Program_weight_tensor_parameter_83: name = "parameter_83" + original_name = "layer_norm_79.b_0" shape = [320] dtype = "float32" min_val = float("-0.592313") @@ -896,6 +980,7 @@ class Program_weight_tensor_parameter_83: class Program_weight_tensor_parameter_84: name = "parameter_84" + original_name = "layer_norm_79.w_0" shape = [320] dtype = "float32" min_val = float("-0.0075549") @@ -907,6 +992,7 @@ class Program_weight_tensor_parameter_84: class Program_weight_tensor_parameter_85: name = "parameter_85" + original_name = "conv2d_52.b_0" shape = [320] dtype = "float32" min_val = float("-3.22999") @@ -918,6 +1004,7 @@ class Program_weight_tensor_parameter_85: class Program_weight_tensor_parameter_86: name = "parameter_86" + original_name = "conv2d_52.w_0" shape = [320, 320, 2, 2] dtype = "float32" min_val = float("-0.367877") @@ -929,6 +1016,7 @@ class Program_weight_tensor_parameter_86: class Program_weight_tensor_parameter_87: name = "parameter_87" + original_name = "linear_120.b_0" shape = [320] dtype = "float32" min_val = float("-1.92459") @@ -940,6 +1028,7 @@ class Program_weight_tensor_parameter_87: class Program_weight_tensor_parameter_88: name = "parameter_88" + original_name = "linear_120.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.341789") @@ -951,6 +1040,7 @@ class Program_weight_tensor_parameter_88: class Program_weight_tensor_parameter_89: name = "parameter_89" + original_name = "layer_norm_78.b_0" shape = [320] dtype = "float32" min_val = float("-1.95456") @@ -962,6 +1052,7 @@ class Program_weight_tensor_parameter_89: class Program_weight_tensor_parameter_90: name = "parameter_90" + original_name = "layer_norm_78.w_0" shape = [320] dtype = "float32" min_val = float("0.178343") @@ -973,6 +1064,7 @@ class Program_weight_tensor_parameter_90: class Program_weight_tensor_parameter_91: name = "parameter_91" + original_name = "linear_119.b_0" shape = [320] dtype = "float32" min_val = float("-0.349769") @@ -984,6 +1076,7 @@ class Program_weight_tensor_parameter_91: class Program_weight_tensor_parameter_92: name = "parameter_92" + original_name = "linear_119.w_0" shape = [1280, 320] dtype = "float32" min_val = float("-0.940136") @@ -995,6 +1088,7 @@ class Program_weight_tensor_parameter_92: class Program_weight_tensor_parameter_93: name = "parameter_93" + original_name = "conv2d_51.b_0" shape = [1280] dtype = "float32" min_val = float("-2.21091") @@ -1006,6 +1100,7 @@ class Program_weight_tensor_parameter_93: class Program_weight_tensor_parameter_94: name = "parameter_94" + original_name = "conv2d_51.w_0" shape = [1280, 1, 3, 3] dtype = "float32" min_val = float("-0.785976") @@ -1017,6 +1112,7 @@ class Program_weight_tensor_parameter_94: class Program_weight_tensor_parameter_95: name = "parameter_95" + original_name = "linear_118.b_0" shape = [1280] dtype = "float32" min_val = float("-2.40809") @@ -1028,6 +1124,7 @@ class Program_weight_tensor_parameter_95: class Program_weight_tensor_parameter_96: name = "parameter_96" + original_name = "linear_118.w_0" shape = [320, 1280] dtype = "float32" min_val = float("-0.490331") @@ -1039,6 +1136,7 @@ class Program_weight_tensor_parameter_96: class Program_weight_tensor_parameter_97: name = "parameter_97" + original_name = "layer_norm_77.b_0" shape = [320] dtype = "float32" min_val = float("-1.23975") @@ -1050,6 +1148,7 @@ class Program_weight_tensor_parameter_97: class Program_weight_tensor_parameter_98: name = "parameter_98" + original_name = "layer_norm_77.w_0" shape = [320] dtype = "float32" min_val = float("0.118928") @@ -1061,6 +1160,7 @@ class Program_weight_tensor_parameter_98: class Program_weight_tensor_parameter_99: name = "parameter_99" + original_name = "linear_117.b_0" shape = [320] dtype = "float32" min_val = float("-0.185994") @@ -1072,6 +1172,7 @@ class Program_weight_tensor_parameter_99: class Program_weight_tensor_parameter_100: name = "parameter_100" + original_name = "linear_117.w_0" shape = [320, 320] dtype = "float32" min_val = float("-1.01232") @@ -1083,6 +1184,7 @@ class Program_weight_tensor_parameter_100: class Program_weight_tensor_parameter_101: name = "parameter_101" + original_name = "linear_116.b_0" shape = [640] dtype = "float32" min_val = float("-3.45691") @@ -1094,6 +1196,7 @@ class Program_weight_tensor_parameter_101: class Program_weight_tensor_parameter_102: name = "parameter_102" + original_name = "linear_116.w_0" shape = [320, 640] dtype = "float32" min_val = float("-0.408055") @@ -1105,6 +1208,7 @@ class Program_weight_tensor_parameter_102: class Program_weight_tensor_parameter_103: name = "parameter_103" + original_name = "layer_norm_76.b_0" shape = [320] dtype = "float32" min_val = float("-0.397654") @@ -1116,6 +1220,7 @@ class Program_weight_tensor_parameter_103: class Program_weight_tensor_parameter_104: name = "parameter_104" + original_name = "layer_norm_76.w_0" shape = [320] dtype = "float32" min_val = float("-0.0455543") @@ -1127,6 +1232,7 @@ class Program_weight_tensor_parameter_104: class Program_weight_tensor_parameter_105: name = "parameter_105" + original_name = "conv2d_50.b_0" shape = [320] dtype = "float32" min_val = float("-4.48621") @@ -1138,6 +1244,7 @@ class Program_weight_tensor_parameter_105: class Program_weight_tensor_parameter_106: name = "parameter_106" + original_name = "conv2d_50.w_0" shape = [320, 320, 2, 2] dtype = "float32" min_val = float("-0.368899") @@ -1149,6 +1256,7 @@ class Program_weight_tensor_parameter_106: class Program_weight_tensor_parameter_107: name = "parameter_107" + original_name = "linear_115.b_0" shape = [320] dtype = "float32" min_val = float("-2.66029") @@ -1160,6 +1268,7 @@ class Program_weight_tensor_parameter_107: class Program_weight_tensor_parameter_108: name = "parameter_108" + original_name = "linear_115.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.277437") @@ -1171,6 +1280,7 @@ class Program_weight_tensor_parameter_108: class Program_weight_tensor_parameter_109: name = "parameter_109" + original_name = "layer_norm_75.b_0" shape = [320] dtype = "float32" min_val = float("-1.97448") @@ -1182,6 +1292,7 @@ class Program_weight_tensor_parameter_109: class Program_weight_tensor_parameter_110: name = "parameter_110" + original_name = "layer_norm_75.w_0" shape = [320] dtype = "float32" min_val = float("0.169339") @@ -1193,6 +1304,7 @@ class Program_weight_tensor_parameter_110: class Program_weight_tensor_parameter_111: name = "parameter_111" + original_name = "linear_114.b_0" shape = [320] dtype = "float32" min_val = float("-0.34793") @@ -1204,6 +1316,7 @@ class Program_weight_tensor_parameter_111: class Program_weight_tensor_parameter_112: name = "parameter_112" + original_name = "linear_114.w_0" shape = [1280, 320] dtype = "float32" min_val = float("-0.943906") @@ -1215,6 +1328,7 @@ class Program_weight_tensor_parameter_112: class Program_weight_tensor_parameter_113: name = "parameter_113" + original_name = "conv2d_49.b_0" shape = [1280] dtype = "float32" min_val = float("-2.10186") @@ -1226,6 +1340,7 @@ class Program_weight_tensor_parameter_113: class Program_weight_tensor_parameter_114: name = "parameter_114" + original_name = "conv2d_49.w_0" shape = [1280, 1, 3, 3] dtype = "float32" min_val = float("-0.511464") @@ -1237,6 +1352,7 @@ class Program_weight_tensor_parameter_114: class Program_weight_tensor_parameter_115: name = "parameter_115" + original_name = "linear_113.b_0" shape = [1280] dtype = "float32" min_val = float("-3.2596") @@ -1248,6 +1364,7 @@ class Program_weight_tensor_parameter_115: class Program_weight_tensor_parameter_116: name = "parameter_116" + original_name = "linear_113.w_0" shape = [320, 1280] dtype = "float32" min_val = float("-0.467937") @@ -1259,6 +1376,7 @@ class Program_weight_tensor_parameter_116: class Program_weight_tensor_parameter_117: name = "parameter_117" + original_name = "layer_norm_74.b_0" shape = [320] dtype = "float32" min_val = float("-1.14677") @@ -1270,6 +1388,7 @@ class Program_weight_tensor_parameter_117: class Program_weight_tensor_parameter_118: name = "parameter_118" + original_name = "layer_norm_74.w_0" shape = [320] dtype = "float32" min_val = float("0.130399") @@ -1281,6 +1400,7 @@ class Program_weight_tensor_parameter_118: class Program_weight_tensor_parameter_119: name = "parameter_119" + original_name = "linear_112.b_0" shape = [320] dtype = "float32" min_val = float("-0.162402") @@ -1292,6 +1412,7 @@ class Program_weight_tensor_parameter_119: class Program_weight_tensor_parameter_120: name = "parameter_120" + original_name = "linear_112.w_0" shape = [320, 320] dtype = "float32" min_val = float("-1.248") @@ -1303,6 +1424,7 @@ class Program_weight_tensor_parameter_120: class Program_weight_tensor_parameter_121: name = "parameter_121" + original_name = "linear_111.b_0" shape = [640] dtype = "float32" min_val = float("-1.01313") @@ -1314,6 +1436,7 @@ class Program_weight_tensor_parameter_121: class Program_weight_tensor_parameter_122: name = "parameter_122" + original_name = "linear_111.w_0" shape = [320, 640] dtype = "float32" min_val = float("-0.41278") @@ -1325,6 +1448,7 @@ class Program_weight_tensor_parameter_122: class Program_weight_tensor_parameter_123: name = "parameter_123" + original_name = "layer_norm_73.b_0" shape = [320] dtype = "float32" min_val = float("-0.138184") @@ -1336,6 +1460,7 @@ class Program_weight_tensor_parameter_123: class Program_weight_tensor_parameter_124: name = "parameter_124" + original_name = "layer_norm_73.w_0" shape = [320] dtype = "float32" min_val = float("-0.0464694") @@ -1347,6 +1472,7 @@ class Program_weight_tensor_parameter_124: class Program_weight_tensor_parameter_125: name = "parameter_125" + original_name = "conv2d_48.b_0" shape = [320] dtype = "float32" min_val = float("-5.19216") @@ -1358,6 +1484,7 @@ class Program_weight_tensor_parameter_125: class Program_weight_tensor_parameter_126: name = "parameter_126" + original_name = "conv2d_48.w_0" shape = [320, 320, 2, 2] dtype = "float32" min_val = float("-0.393314") @@ -1369,6 +1496,7 @@ class Program_weight_tensor_parameter_126: class Program_weight_tensor_parameter_127: name = "parameter_127" + original_name = "linear_110.b_0" shape = [320] dtype = "float32" min_val = float("-2.21728") @@ -1380,6 +1508,7 @@ class Program_weight_tensor_parameter_127: class Program_weight_tensor_parameter_128: name = "parameter_128" + original_name = "linear_110.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.24402") @@ -1391,6 +1520,7 @@ class Program_weight_tensor_parameter_128: class Program_weight_tensor_parameter_129: name = "parameter_129" + original_name = "layer_norm_72.b_0" shape = [320] dtype = "float32" min_val = float("-2.1147") @@ -1402,6 +1532,7 @@ class Program_weight_tensor_parameter_129: class Program_weight_tensor_parameter_130: name = "parameter_130" + original_name = "layer_norm_72.w_0" shape = [320] dtype = "float32" min_val = float("0.193093") @@ -1413,6 +1544,7 @@ class Program_weight_tensor_parameter_130: class Program_weight_tensor_parameter_131: name = "parameter_131" + original_name = "linear_109.b_0" shape = [320] dtype = "float32" min_val = float("-0.281282") @@ -1424,6 +1556,7 @@ class Program_weight_tensor_parameter_131: class Program_weight_tensor_parameter_132: name = "parameter_132" + original_name = "linear_109.w_0" shape = [1280, 320] dtype = "float32" min_val = float("-0.982913") @@ -1435,6 +1568,7 @@ class Program_weight_tensor_parameter_132: class Program_weight_tensor_parameter_133: name = "parameter_133" + original_name = "conv2d_47.b_0" shape = [1280] dtype = "float32" min_val = float("-1.61811") @@ -1446,6 +1580,7 @@ class Program_weight_tensor_parameter_133: class Program_weight_tensor_parameter_134: name = "parameter_134" + original_name = "conv2d_47.w_0" shape = [1280, 1, 3, 3] dtype = "float32" min_val = float("-0.580233") @@ -1457,6 +1592,7 @@ class Program_weight_tensor_parameter_134: class Program_weight_tensor_parameter_135: name = "parameter_135" + original_name = "linear_108.b_0" shape = [1280] dtype = "float32" min_val = float("-2.72138") @@ -1468,6 +1604,7 @@ class Program_weight_tensor_parameter_135: class Program_weight_tensor_parameter_136: name = "parameter_136" + original_name = "linear_108.w_0" shape = [320, 1280] dtype = "float32" min_val = float("-0.605835") @@ -1479,6 +1616,7 @@ class Program_weight_tensor_parameter_136: class Program_weight_tensor_parameter_137: name = "parameter_137" + original_name = "layer_norm_71.b_0" shape = [320] dtype = "float32" min_val = float("-1.33534") @@ -1490,6 +1628,7 @@ class Program_weight_tensor_parameter_137: class Program_weight_tensor_parameter_138: name = "parameter_138" + original_name = "layer_norm_71.w_0" shape = [320] dtype = "float32" min_val = float("0.167713") @@ -1501,6 +1640,7 @@ class Program_weight_tensor_parameter_138: class Program_weight_tensor_parameter_139: name = "parameter_139" + original_name = "linear_107.b_0" shape = [320] dtype = "float32" min_val = float("-0.197496") @@ -1512,6 +1652,7 @@ class Program_weight_tensor_parameter_139: class Program_weight_tensor_parameter_140: name = "parameter_140" + original_name = "linear_107.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.608485") @@ -1523,6 +1664,7 @@ class Program_weight_tensor_parameter_140: class Program_weight_tensor_parameter_141: name = "parameter_141" + original_name = "linear_106.b_0" shape = [640] dtype = "float32" min_val = float("-1.53117") @@ -1534,6 +1676,7 @@ class Program_weight_tensor_parameter_141: class Program_weight_tensor_parameter_142: name = "parameter_142" + original_name = "linear_106.w_0" shape = [320, 640] dtype = "float32" min_val = float("-0.398146") @@ -1545,6 +1688,7 @@ class Program_weight_tensor_parameter_142: class Program_weight_tensor_parameter_143: name = "parameter_143" + original_name = "layer_norm_70.b_0" shape = [320] dtype = "float32" min_val = float("-0.606562") @@ -1556,6 +1700,7 @@ class Program_weight_tensor_parameter_143: class Program_weight_tensor_parameter_144: name = "parameter_144" + original_name = "layer_norm_70.w_0" shape = [320] dtype = "float32" min_val = float("-0.0549927") @@ -1567,6 +1712,7 @@ class Program_weight_tensor_parameter_144: class Program_weight_tensor_parameter_145: name = "parameter_145" + original_name = "conv2d_46.b_0" shape = [320] dtype = "float32" min_val = float("-4.0911") @@ -1578,6 +1724,7 @@ class Program_weight_tensor_parameter_145: class Program_weight_tensor_parameter_146: name = "parameter_146" + original_name = "conv2d_46.w_0" shape = [320, 320, 2, 2] dtype = "float32" min_val = float("-0.374492") @@ -1589,6 +1736,7 @@ class Program_weight_tensor_parameter_146: class Program_weight_tensor_parameter_147: name = "parameter_147" + original_name = "linear_105.b_0" shape = [320] dtype = "float32" min_val = float("-2.62271") @@ -1600,6 +1748,7 @@ class Program_weight_tensor_parameter_147: class Program_weight_tensor_parameter_148: name = "parameter_148" + original_name = "linear_105.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.276657") @@ -1611,6 +1760,7 @@ class Program_weight_tensor_parameter_148: class Program_weight_tensor_parameter_149: name = "parameter_149" + original_name = "layer_norm_69.b_0" shape = [320] dtype = "float32" min_val = float("-0.811771") @@ -1622,6 +1772,7 @@ class Program_weight_tensor_parameter_149: class Program_weight_tensor_parameter_150: name = "parameter_150" + original_name = "layer_norm_69.w_0" shape = [320] dtype = "float32" min_val = float("0.216041") @@ -1633,6 +1784,7 @@ class Program_weight_tensor_parameter_150: class Program_weight_tensor_parameter_151: name = "parameter_151" + original_name = "linear_104.b_0" shape = [320] dtype = "float32" min_val = float("-0.284063") @@ -1644,6 +1796,7 @@ class Program_weight_tensor_parameter_151: class Program_weight_tensor_parameter_152: name = "parameter_152" + original_name = "linear_104.w_0" shape = [1280, 320] dtype = "float32" min_val = float("-1.05095") @@ -1655,6 +1808,7 @@ class Program_weight_tensor_parameter_152: class Program_weight_tensor_parameter_153: name = "parameter_153" + original_name = "conv2d_45.b_0" shape = [1280] dtype = "float32" min_val = float("-1.81158") @@ -1666,6 +1820,7 @@ class Program_weight_tensor_parameter_153: class Program_weight_tensor_parameter_154: name = "parameter_154" + original_name = "conv2d_45.w_0" shape = [1280, 1, 3, 3] dtype = "float32" min_val = float("-0.961145") @@ -1677,6 +1832,7 @@ class Program_weight_tensor_parameter_154: class Program_weight_tensor_parameter_155: name = "parameter_155" + original_name = "linear_103.b_0" shape = [1280] dtype = "float32" min_val = float("-2.70179") @@ -1688,6 +1844,7 @@ class Program_weight_tensor_parameter_155: class Program_weight_tensor_parameter_156: name = "parameter_156" + original_name = "linear_103.w_0" shape = [320, 1280] dtype = "float32" min_val = float("-0.515404") @@ -1699,6 +1856,7 @@ class Program_weight_tensor_parameter_156: class Program_weight_tensor_parameter_157: name = "parameter_157" + original_name = "layer_norm_68.b_0" shape = [320] dtype = "float32" min_val = float("-1.61607") @@ -1710,6 +1868,7 @@ class Program_weight_tensor_parameter_157: class Program_weight_tensor_parameter_158: name = "parameter_158" + original_name = "layer_norm_68.w_0" shape = [320] dtype = "float32" min_val = float("0.244421") @@ -1721,6 +1880,7 @@ class Program_weight_tensor_parameter_158: class Program_weight_tensor_parameter_159: name = "parameter_159" + original_name = "linear_102.b_0" shape = [320] dtype = "float32" min_val = float("-0.15244") @@ -1732,6 +1892,7 @@ class Program_weight_tensor_parameter_159: class Program_weight_tensor_parameter_160: name = "parameter_160" + original_name = "linear_102.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.311542") @@ -1743,6 +1904,7 @@ class Program_weight_tensor_parameter_160: class Program_weight_tensor_parameter_161: name = "parameter_161" + original_name = "linear_101.b_0" shape = [640] dtype = "float32" min_val = float("-1.00988") @@ -1754,6 +1916,7 @@ class Program_weight_tensor_parameter_161: class Program_weight_tensor_parameter_162: name = "parameter_162" + original_name = "linear_101.w_0" shape = [320, 640] dtype = "float32" min_val = float("-0.416267") @@ -1765,6 +1928,7 @@ class Program_weight_tensor_parameter_162: class Program_weight_tensor_parameter_163: name = "parameter_163" + original_name = "layer_norm_67.b_0" shape = [320] dtype = "float32" min_val = float("-0.222455") @@ -1776,6 +1940,7 @@ class Program_weight_tensor_parameter_163: class Program_weight_tensor_parameter_164: name = "parameter_164" + original_name = "layer_norm_67.w_0" shape = [320] dtype = "float32" min_val = float("-0.00449497") @@ -1787,6 +1952,7 @@ class Program_weight_tensor_parameter_164: class Program_weight_tensor_parameter_165: name = "parameter_165" + original_name = "conv2d_44.b_0" shape = [320] dtype = "float32" min_val = float("-6.72445") @@ -1798,6 +1964,7 @@ class Program_weight_tensor_parameter_165: class Program_weight_tensor_parameter_166: name = "parameter_166" + original_name = "conv2d_44.w_0" shape = [320, 320, 2, 2] dtype = "float32" min_val = float("-0.441434") @@ -1809,6 +1976,7 @@ class Program_weight_tensor_parameter_166: class Program_weight_tensor_parameter_167: name = "parameter_167" + original_name = "linear_100.b_0" shape = [320] dtype = "float32" min_val = float("-2.84648") @@ -1820,6 +1988,7 @@ class Program_weight_tensor_parameter_167: class Program_weight_tensor_parameter_168: name = "parameter_168" + original_name = "linear_100.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.235234") @@ -1831,6 +2000,7 @@ class Program_weight_tensor_parameter_168: class Program_weight_tensor_parameter_169: name = "parameter_169" + original_name = "layer_norm_66.b_0" shape = [320] dtype = "float32" min_val = float("-1.13336") @@ -1842,6 +2012,7 @@ class Program_weight_tensor_parameter_169: class Program_weight_tensor_parameter_170: name = "parameter_170" + original_name = "layer_norm_66.w_0" shape = [320] dtype = "float32" min_val = float("0.123042") @@ -1853,6 +2024,7 @@ class Program_weight_tensor_parameter_170: class Program_weight_tensor_parameter_171: name = "parameter_171" + original_name = "linear_99.b_0" shape = [320] dtype = "float32" min_val = float("-0.310795") @@ -1864,6 +2036,7 @@ class Program_weight_tensor_parameter_171: class Program_weight_tensor_parameter_172: name = "parameter_172" + original_name = "linear_99.w_0" shape = [1280, 320] dtype = "float32" min_val = float("-0.652521") @@ -1875,6 +2048,7 @@ class Program_weight_tensor_parameter_172: class Program_weight_tensor_parameter_173: name = "parameter_173" + original_name = "conv2d_43.b_0" shape = [1280] dtype = "float32" min_val = float("-1.363") @@ -1886,6 +2060,7 @@ class Program_weight_tensor_parameter_173: class Program_weight_tensor_parameter_174: name = "parameter_174" + original_name = "conv2d_43.w_0" shape = [1280, 1, 3, 3] dtype = "float32" min_val = float("-0.707557") @@ -1897,6 +2072,7 @@ class Program_weight_tensor_parameter_174: class Program_weight_tensor_parameter_175: name = "parameter_175" + original_name = "linear_98.b_0" shape = [1280] dtype = "float32" min_val = float("-2.66978") @@ -1908,6 +2084,7 @@ class Program_weight_tensor_parameter_175: class Program_weight_tensor_parameter_176: name = "parameter_176" + original_name = "linear_98.w_0" shape = [320, 1280] dtype = "float32" min_val = float("-0.450463") @@ -1919,6 +2096,7 @@ class Program_weight_tensor_parameter_176: class Program_weight_tensor_parameter_177: name = "parameter_177" + original_name = "layer_norm_65.b_0" shape = [320] dtype = "float32" min_val = float("-1.64871") @@ -1930,6 +2108,7 @@ class Program_weight_tensor_parameter_177: class Program_weight_tensor_parameter_178: name = "parameter_178" + original_name = "layer_norm_65.w_0" shape = [320] dtype = "float32" min_val = float("0.273106") @@ -1941,6 +2120,7 @@ class Program_weight_tensor_parameter_178: class Program_weight_tensor_parameter_179: name = "parameter_179" + original_name = "linear_97.b_0" shape = [320] dtype = "float32" min_val = float("-0.24894") @@ -1952,6 +2132,7 @@ class Program_weight_tensor_parameter_179: class Program_weight_tensor_parameter_180: name = "parameter_180" + original_name = "linear_97.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.437735") @@ -1963,6 +2144,7 @@ class Program_weight_tensor_parameter_180: class Program_weight_tensor_parameter_181: name = "parameter_181" + original_name = "linear_96.b_0" shape = [640] dtype = "float32" min_val = float("-0.789568") @@ -1974,6 +2156,7 @@ class Program_weight_tensor_parameter_181: class Program_weight_tensor_parameter_182: name = "parameter_182" + original_name = "linear_96.w_0" shape = [320, 640] dtype = "float32" min_val = float("-0.397866") @@ -1985,6 +2168,7 @@ class Program_weight_tensor_parameter_182: class Program_weight_tensor_parameter_183: name = "parameter_183" + original_name = "layer_norm_64.b_0" shape = [320] dtype = "float32" min_val = float("-0.205507") @@ -1996,6 +2180,7 @@ class Program_weight_tensor_parameter_183: class Program_weight_tensor_parameter_184: name = "parameter_184" + original_name = "layer_norm_64.w_0" shape = [320] dtype = "float32" min_val = float("-0.00654609") @@ -2007,6 +2192,7 @@ class Program_weight_tensor_parameter_184: class Program_weight_tensor_parameter_185: name = "parameter_185" + original_name = "conv2d_42.b_0" shape = [320] dtype = "float32" min_val = float("-5.30833") @@ -2018,6 +2204,7 @@ class Program_weight_tensor_parameter_185: class Program_weight_tensor_parameter_186: name = "parameter_186" + original_name = "conv2d_42.w_0" shape = [320, 320, 2, 2] dtype = "float32" min_val = float("-0.452437") @@ -2029,6 +2216,7 @@ class Program_weight_tensor_parameter_186: class Program_weight_tensor_parameter_187: name = "parameter_187" + original_name = "linear_95.b_0" shape = [320] dtype = "float32" min_val = float("-2.70259") @@ -2040,6 +2228,7 @@ class Program_weight_tensor_parameter_187: class Program_weight_tensor_parameter_188: name = "parameter_188" + original_name = "linear_95.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.250459") @@ -2051,6 +2240,7 @@ class Program_weight_tensor_parameter_188: class Program_weight_tensor_parameter_189: name = "parameter_189" + original_name = "layer_norm_63.b_0" shape = [320] dtype = "float32" min_val = float("-1.18784") @@ -2062,6 +2252,7 @@ class Program_weight_tensor_parameter_189: class Program_weight_tensor_parameter_190: name = "parameter_190" + original_name = "layer_norm_63.w_0" shape = [320] dtype = "float32" min_val = float("0.147348") @@ -2073,6 +2264,7 @@ class Program_weight_tensor_parameter_190: class Program_weight_tensor_parameter_191: name = "parameter_191" + original_name = "linear_94.b_0" shape = [320] dtype = "float32" min_val = float("-0.355624") @@ -2084,6 +2276,7 @@ class Program_weight_tensor_parameter_191: class Program_weight_tensor_parameter_192: name = "parameter_192" + original_name = "linear_94.w_0" shape = [1280, 320] dtype = "float32" min_val = float("-0.70344") @@ -2095,6 +2288,7 @@ class Program_weight_tensor_parameter_192: class Program_weight_tensor_parameter_193: name = "parameter_193" + original_name = "conv2d_41.b_0" shape = [1280] dtype = "float32" min_val = float("-1.71278") @@ -2106,6 +2300,7 @@ class Program_weight_tensor_parameter_193: class Program_weight_tensor_parameter_194: name = "parameter_194" + original_name = "conv2d_41.w_0" shape = [1280, 1, 3, 3] dtype = "float32" min_val = float("-1.63773") @@ -2117,6 +2312,7 @@ class Program_weight_tensor_parameter_194: class Program_weight_tensor_parameter_195: name = "parameter_195" + original_name = "linear_93.b_0" shape = [1280] dtype = "float32" min_val = float("-2.43915") @@ -2128,6 +2324,7 @@ class Program_weight_tensor_parameter_195: class Program_weight_tensor_parameter_196: name = "parameter_196" + original_name = "linear_93.w_0" shape = [320, 1280] dtype = "float32" min_val = float("-0.447934") @@ -2139,6 +2336,7 @@ class Program_weight_tensor_parameter_196: class Program_weight_tensor_parameter_197: name = "parameter_197" + original_name = "layer_norm_62.b_0" shape = [320] dtype = "float32" min_val = float("-1.72366") @@ -2150,6 +2348,7 @@ class Program_weight_tensor_parameter_197: class Program_weight_tensor_parameter_198: name = "parameter_198" + original_name = "layer_norm_62.w_0" shape = [320] dtype = "float32" min_val = float("0.367186") @@ -2161,6 +2360,7 @@ class Program_weight_tensor_parameter_198: class Program_weight_tensor_parameter_199: name = "parameter_199" + original_name = "linear_92.b_0" shape = [320] dtype = "float32" min_val = float("-0.147982") @@ -2172,6 +2372,7 @@ class Program_weight_tensor_parameter_199: class Program_weight_tensor_parameter_200: name = "parameter_200" + original_name = "linear_92.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.315174") @@ -2183,6 +2384,7 @@ class Program_weight_tensor_parameter_200: class Program_weight_tensor_parameter_201: name = "parameter_201" + original_name = "linear_91.b_0" shape = [640] dtype = "float32" min_val = float("-1.06848") @@ -2194,6 +2396,7 @@ class Program_weight_tensor_parameter_201: class Program_weight_tensor_parameter_202: name = "parameter_202" + original_name = "linear_91.w_0" shape = [320, 640] dtype = "float32" min_val = float("-0.380639") @@ -2205,6 +2408,7 @@ class Program_weight_tensor_parameter_202: class Program_weight_tensor_parameter_203: name = "parameter_203" + original_name = "layer_norm_61.b_0" shape = [320] dtype = "float32" min_val = float("-0.332555") @@ -2216,6 +2420,7 @@ class Program_weight_tensor_parameter_203: class Program_weight_tensor_parameter_204: name = "parameter_204" + original_name = "layer_norm_61.w_0" shape = [320] dtype = "float32" min_val = float("-0.00696373") @@ -2227,6 +2432,7 @@ class Program_weight_tensor_parameter_204: class Program_weight_tensor_parameter_205: name = "parameter_205" + original_name = "conv2d_40.b_0" shape = [320] dtype = "float32" min_val = float("-5.65661") @@ -2238,6 +2444,7 @@ class Program_weight_tensor_parameter_205: class Program_weight_tensor_parameter_206: name = "parameter_206" + original_name = "conv2d_40.w_0" shape = [320, 320, 2, 2] dtype = "float32" min_val = float("-0.452337") @@ -2249,6 +2456,7 @@ class Program_weight_tensor_parameter_206: class Program_weight_tensor_parameter_207: name = "parameter_207" + original_name = "linear_90.b_0" shape = [320] dtype = "float32" min_val = float("-2.31572") @@ -2260,6 +2468,7 @@ class Program_weight_tensor_parameter_207: class Program_weight_tensor_parameter_208: name = "parameter_208" + original_name = "linear_90.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.319029") @@ -2271,6 +2480,7 @@ class Program_weight_tensor_parameter_208: class Program_weight_tensor_parameter_209: name = "parameter_209" + original_name = "layer_norm_60.b_0" shape = [320] dtype = "float32" min_val = float("-1.17876") @@ -2282,6 +2492,7 @@ class Program_weight_tensor_parameter_209: class Program_weight_tensor_parameter_210: name = "parameter_210" + original_name = "layer_norm_60.w_0" shape = [320] dtype = "float32" min_val = float("0.153098") @@ -2293,6 +2504,7 @@ class Program_weight_tensor_parameter_210: class Program_weight_tensor_parameter_211: name = "parameter_211" + original_name = "linear_89.b_0" shape = [320] dtype = "float32" min_val = float("-0.340085") @@ -2304,6 +2516,7 @@ class Program_weight_tensor_parameter_211: class Program_weight_tensor_parameter_212: name = "parameter_212" + original_name = "linear_89.w_0" shape = [1280, 320] dtype = "float32" min_val = float("-0.704058") @@ -2315,6 +2528,7 @@ class Program_weight_tensor_parameter_212: class Program_weight_tensor_parameter_213: name = "parameter_213" + original_name = "conv2d_39.b_0" shape = [1280] dtype = "float32" min_val = float("-2.06264") @@ -2326,6 +2540,7 @@ class Program_weight_tensor_parameter_213: class Program_weight_tensor_parameter_214: name = "parameter_214" + original_name = "conv2d_39.w_0" shape = [1280, 1, 3, 3] dtype = "float32" min_val = float("-1.75862") @@ -2337,6 +2552,7 @@ class Program_weight_tensor_parameter_214: class Program_weight_tensor_parameter_215: name = "parameter_215" + original_name = "linear_88.b_0" shape = [1280] dtype = "float32" min_val = float("-2.79886") @@ -2348,6 +2564,7 @@ class Program_weight_tensor_parameter_215: class Program_weight_tensor_parameter_216: name = "parameter_216" + original_name = "linear_88.w_0" shape = [320, 1280] dtype = "float32" min_val = float("-0.440331") @@ -2359,6 +2576,7 @@ class Program_weight_tensor_parameter_216: class Program_weight_tensor_parameter_217: name = "parameter_217" + original_name = "layer_norm_59.b_0" shape = [320] dtype = "float32" min_val = float("-1.79003") @@ -2370,6 +2588,7 @@ class Program_weight_tensor_parameter_217: class Program_weight_tensor_parameter_218: name = "parameter_218" + original_name = "layer_norm_59.w_0" shape = [320] dtype = "float32" min_val = float("0.47463") @@ -2381,6 +2600,7 @@ class Program_weight_tensor_parameter_218: class Program_weight_tensor_parameter_219: name = "parameter_219" + original_name = "linear_87.b_0" shape = [320] dtype = "float32" min_val = float("-0.13856") @@ -2392,6 +2612,7 @@ class Program_weight_tensor_parameter_219: class Program_weight_tensor_parameter_220: name = "parameter_220" + original_name = "linear_87.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.375189") @@ -2403,6 +2624,7 @@ class Program_weight_tensor_parameter_220: class Program_weight_tensor_parameter_221: name = "parameter_221" + original_name = "linear_86.b_0" shape = [640] dtype = "float32" min_val = float("-0.822315") @@ -2414,6 +2636,7 @@ class Program_weight_tensor_parameter_221: class Program_weight_tensor_parameter_222: name = "parameter_222" + original_name = "linear_86.w_0" shape = [320, 640] dtype = "float32" min_val = float("-0.474278") @@ -2425,6 +2648,7 @@ class Program_weight_tensor_parameter_222: class Program_weight_tensor_parameter_223: name = "parameter_223" + original_name = "layer_norm_58.b_0" shape = [320] dtype = "float32" min_val = float("-0.23353") @@ -2436,6 +2660,7 @@ class Program_weight_tensor_parameter_223: class Program_weight_tensor_parameter_224: name = "parameter_224" + original_name = "layer_norm_58.w_0" shape = [320] dtype = "float32" min_val = float("-0.115005") @@ -2447,6 +2672,7 @@ class Program_weight_tensor_parameter_224: class Program_weight_tensor_parameter_225: name = "parameter_225" + original_name = "conv2d_38.b_0" shape = [320] dtype = "float32" min_val = float("-6.68654") @@ -2458,6 +2684,7 @@ class Program_weight_tensor_parameter_225: class Program_weight_tensor_parameter_226: name = "parameter_226" + original_name = "conv2d_38.w_0" shape = [320, 320, 2, 2] dtype = "float32" min_val = float("-0.375019") @@ -2469,6 +2696,7 @@ class Program_weight_tensor_parameter_226: class Program_weight_tensor_parameter_227: name = "parameter_227" + original_name = "linear_85.b_0" shape = [320] dtype = "float32" min_val = float("-2.8188") @@ -2480,6 +2708,7 @@ class Program_weight_tensor_parameter_227: class Program_weight_tensor_parameter_228: name = "parameter_228" + original_name = "linear_85.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.305233") @@ -2491,6 +2720,7 @@ class Program_weight_tensor_parameter_228: class Program_weight_tensor_parameter_229: name = "parameter_229" + original_name = "layer_norm_57.b_0" shape = [320] dtype = "float32" min_val = float("-1.25198") @@ -2502,6 +2732,7 @@ class Program_weight_tensor_parameter_229: class Program_weight_tensor_parameter_230: name = "parameter_230" + original_name = "layer_norm_57.w_0" shape = [320] dtype = "float32" min_val = float("0.195228") @@ -2513,6 +2744,7 @@ class Program_weight_tensor_parameter_230: class Program_weight_tensor_parameter_231: name = "parameter_231" + original_name = "linear_84.b_0" shape = [320] dtype = "float32" min_val = float("-0.403075") @@ -2524,6 +2756,7 @@ class Program_weight_tensor_parameter_231: class Program_weight_tensor_parameter_232: name = "parameter_232" + original_name = "linear_84.w_0" shape = [1280, 320] dtype = "float32" min_val = float("-0.688622") @@ -2535,6 +2768,7 @@ class Program_weight_tensor_parameter_232: class Program_weight_tensor_parameter_233: name = "parameter_233" + original_name = "conv2d_37.b_0" shape = [1280] dtype = "float32" min_val = float("-1.86952") @@ -2546,6 +2780,7 @@ class Program_weight_tensor_parameter_233: class Program_weight_tensor_parameter_234: name = "parameter_234" + original_name = "conv2d_37.w_0" shape = [1280, 1, 3, 3] dtype = "float32" min_val = float("-0.339875") @@ -2557,6 +2792,7 @@ class Program_weight_tensor_parameter_234: class Program_weight_tensor_parameter_235: name = "parameter_235" + original_name = "linear_83.b_0" shape = [1280] dtype = "float32" min_val = float("-1.67288") @@ -2568,6 +2804,7 @@ class Program_weight_tensor_parameter_235: class Program_weight_tensor_parameter_236: name = "parameter_236" + original_name = "linear_83.w_0" shape = [320, 1280] dtype = "float32" min_val = float("-0.45373") @@ -2579,6 +2816,7 @@ class Program_weight_tensor_parameter_236: class Program_weight_tensor_parameter_237: name = "parameter_237" + original_name = "layer_norm_56.b_0" shape = [320] dtype = "float32" min_val = float("-1.90936") @@ -2590,6 +2828,7 @@ class Program_weight_tensor_parameter_237: class Program_weight_tensor_parameter_238: name = "parameter_238" + original_name = "layer_norm_56.w_0" shape = [320] dtype = "float32" min_val = float("0.614984") @@ -2601,6 +2840,7 @@ class Program_weight_tensor_parameter_238: class Program_weight_tensor_parameter_239: name = "parameter_239" + original_name = "linear_82.b_0" shape = [320] dtype = "float32" min_val = float("-0.15685") @@ -2612,6 +2852,7 @@ class Program_weight_tensor_parameter_239: class Program_weight_tensor_parameter_240: name = "parameter_240" + original_name = "linear_82.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.401675") @@ -2623,6 +2864,7 @@ class Program_weight_tensor_parameter_240: class Program_weight_tensor_parameter_241: name = "parameter_241" + original_name = "linear_81.b_0" shape = [640] dtype = "float32" min_val = float("-0.997232") @@ -2634,6 +2876,7 @@ class Program_weight_tensor_parameter_241: class Program_weight_tensor_parameter_242: name = "parameter_242" + original_name = "linear_81.w_0" shape = [320, 640] dtype = "float32" min_val = float("-0.439793") @@ -2645,6 +2888,7 @@ class Program_weight_tensor_parameter_242: class Program_weight_tensor_parameter_243: name = "parameter_243" + original_name = "layer_norm_55.b_0" shape = [320] dtype = "float32" min_val = float("-0.187995") @@ -2656,6 +2900,7 @@ class Program_weight_tensor_parameter_243: class Program_weight_tensor_parameter_244: name = "parameter_244" + original_name = "layer_norm_55.w_0" shape = [320] dtype = "float32" min_val = float("-0.0653846") @@ -2667,6 +2912,7 @@ class Program_weight_tensor_parameter_244: class Program_weight_tensor_parameter_245: name = "parameter_245" + original_name = "conv2d_36.b_0" shape = [320] dtype = "float32" min_val = float("-5.68068") @@ -2678,6 +2924,7 @@ class Program_weight_tensor_parameter_245: class Program_weight_tensor_parameter_246: name = "parameter_246" + original_name = "conv2d_36.w_0" shape = [320, 320, 2, 2] dtype = "float32" min_val = float("-0.410177") @@ -2689,6 +2936,7 @@ class Program_weight_tensor_parameter_246: class Program_weight_tensor_parameter_247: name = "parameter_247" + original_name = "linear_80.b_0" shape = [320] dtype = "float32" min_val = float("-2.90816") @@ -2700,6 +2948,7 @@ class Program_weight_tensor_parameter_247: class Program_weight_tensor_parameter_248: name = "parameter_248" + original_name = "linear_80.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.262042") @@ -2711,6 +2960,7 @@ class Program_weight_tensor_parameter_248: class Program_weight_tensor_parameter_249: name = "parameter_249" + original_name = "layer_norm_54.b_0" shape = [320] dtype = "float32" min_val = float("-1.20802") @@ -2722,6 +2972,7 @@ class Program_weight_tensor_parameter_249: class Program_weight_tensor_parameter_250: name = "parameter_250" + original_name = "layer_norm_54.w_0" shape = [320] dtype = "float32" min_val = float("0.210572") @@ -2733,6 +2984,7 @@ class Program_weight_tensor_parameter_250: class Program_weight_tensor_parameter_251: name = "parameter_251" + original_name = "linear_79.b_0" shape = [320] dtype = "float32" min_val = float("-0.369533") @@ -2744,6 +2996,7 @@ class Program_weight_tensor_parameter_251: class Program_weight_tensor_parameter_252: name = "parameter_252" + original_name = "linear_79.w_0" shape = [1280, 320] dtype = "float32" min_val = float("-0.597047") @@ -2755,6 +3008,7 @@ class Program_weight_tensor_parameter_252: class Program_weight_tensor_parameter_253: name = "parameter_253" + original_name = "conv2d_35.b_0" shape = [1280] dtype = "float32" min_val = float("-1.57846") @@ -2766,6 +3020,7 @@ class Program_weight_tensor_parameter_253: class Program_weight_tensor_parameter_254: name = "parameter_254" + original_name = "conv2d_35.w_0" shape = [1280, 1, 3, 3] dtype = "float32" min_val = float("-0.283197") @@ -2777,6 +3032,7 @@ class Program_weight_tensor_parameter_254: class Program_weight_tensor_parameter_255: name = "parameter_255" + original_name = "linear_78.b_0" shape = [1280] dtype = "float32" min_val = float("-1.99126") @@ -2788,6 +3044,7 @@ class Program_weight_tensor_parameter_255: class Program_weight_tensor_parameter_256: name = "parameter_256" + original_name = "linear_78.w_0" shape = [320, 1280] dtype = "float32" min_val = float("-0.470148") @@ -2799,6 +3056,7 @@ class Program_weight_tensor_parameter_256: class Program_weight_tensor_parameter_257: name = "parameter_257" + original_name = "layer_norm_53.b_0" shape = [320] dtype = "float32" min_val = float("-1.81172") @@ -2810,6 +3068,7 @@ class Program_weight_tensor_parameter_257: class Program_weight_tensor_parameter_258: name = "parameter_258" + original_name = "layer_norm_53.w_0" shape = [320] dtype = "float32" min_val = float("0.667368") @@ -2821,6 +3080,7 @@ class Program_weight_tensor_parameter_258: class Program_weight_tensor_parameter_259: name = "parameter_259" + original_name = "linear_77.b_0" shape = [320] dtype = "float32" min_val = float("-0.169811") @@ -2832,6 +3092,7 @@ class Program_weight_tensor_parameter_259: class Program_weight_tensor_parameter_260: name = "parameter_260" + original_name = "linear_77.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.372184") @@ -2843,6 +3104,7 @@ class Program_weight_tensor_parameter_260: class Program_weight_tensor_parameter_261: name = "parameter_261" + original_name = "linear_76.b_0" shape = [640] dtype = "float32" min_val = float("-1.05135") @@ -2854,6 +3116,7 @@ class Program_weight_tensor_parameter_261: class Program_weight_tensor_parameter_262: name = "parameter_262" + original_name = "linear_76.w_0" shape = [320, 640] dtype = "float32" min_val = float("-0.4159") @@ -2865,6 +3128,7 @@ class Program_weight_tensor_parameter_262: class Program_weight_tensor_parameter_263: name = "parameter_263" + original_name = "layer_norm_52.b_0" shape = [320] dtype = "float32" min_val = float("-0.150528") @@ -2876,6 +3140,7 @@ class Program_weight_tensor_parameter_263: class Program_weight_tensor_parameter_264: name = "parameter_264" + original_name = "layer_norm_52.w_0" shape = [320] dtype = "float32" min_val = float("-0.00983953") @@ -2887,6 +3152,7 @@ class Program_weight_tensor_parameter_264: class Program_weight_tensor_parameter_265: name = "parameter_265" + original_name = "conv2d_34.b_0" shape = [320] dtype = "float32" min_val = float("-6.20589") @@ -2898,6 +3164,7 @@ class Program_weight_tensor_parameter_265: class Program_weight_tensor_parameter_266: name = "parameter_266" + original_name = "conv2d_34.w_0" shape = [320, 320, 2, 2] dtype = "float32" min_val = float("-0.481813") @@ -2909,6 +3176,7 @@ class Program_weight_tensor_parameter_266: class Program_weight_tensor_parameter_267: name = "parameter_267" + original_name = "linear_75.b_0" shape = [320] dtype = "float32" min_val = float("-3.13715") @@ -2920,6 +3188,7 @@ class Program_weight_tensor_parameter_267: class Program_weight_tensor_parameter_268: name = "parameter_268" + original_name = "linear_75.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.261047") @@ -2931,6 +3200,7 @@ class Program_weight_tensor_parameter_268: class Program_weight_tensor_parameter_269: name = "parameter_269" + original_name = "layer_norm_51.b_0" shape = [320] dtype = "float32" min_val = float("-0.807362") @@ -2942,6 +3212,7 @@ class Program_weight_tensor_parameter_269: class Program_weight_tensor_parameter_270: name = "parameter_270" + original_name = "layer_norm_51.w_0" shape = [320] dtype = "float32" min_val = float("0.163395") @@ -2953,6 +3224,7 @@ class Program_weight_tensor_parameter_270: class Program_weight_tensor_parameter_271: name = "parameter_271" + original_name = "linear_74.b_0" shape = [320] dtype = "float32" min_val = float("-0.331304") @@ -2964,6 +3236,7 @@ class Program_weight_tensor_parameter_271: class Program_weight_tensor_parameter_272: name = "parameter_272" + original_name = "linear_74.w_0" shape = [1280, 320] dtype = "float32" min_val = float("-0.395048") @@ -2975,6 +3248,7 @@ class Program_weight_tensor_parameter_272: class Program_weight_tensor_parameter_273: name = "parameter_273" + original_name = "conv2d_33.b_0" shape = [1280] dtype = "float32" min_val = float("-1.45204") @@ -2986,6 +3260,7 @@ class Program_weight_tensor_parameter_273: class Program_weight_tensor_parameter_274: name = "parameter_274" + original_name = "conv2d_33.w_0" shape = [1280, 1, 3, 3] dtype = "float32" min_val = float("-0.449044") @@ -2997,6 +3272,7 @@ class Program_weight_tensor_parameter_274: class Program_weight_tensor_parameter_275: name = "parameter_275" + original_name = "linear_73.b_0" shape = [1280] dtype = "float32" min_val = float("-4.40104") @@ -3008,6 +3284,7 @@ class Program_weight_tensor_parameter_275: class Program_weight_tensor_parameter_276: name = "parameter_276" + original_name = "linear_73.w_0" shape = [320, 1280] dtype = "float32" min_val = float("-0.425955") @@ -3019,6 +3296,7 @@ class Program_weight_tensor_parameter_276: class Program_weight_tensor_parameter_277: name = "parameter_277" + original_name = "layer_norm_50.b_0" shape = [320] dtype = "float32" min_val = float("-1.81473") @@ -3030,6 +3308,7 @@ class Program_weight_tensor_parameter_277: class Program_weight_tensor_parameter_278: name = "parameter_278" + original_name = "layer_norm_50.w_0" shape = [320] dtype = "float32" min_val = float("0.757548") @@ -3041,6 +3320,7 @@ class Program_weight_tensor_parameter_278: class Program_weight_tensor_parameter_279: name = "parameter_279" + original_name = "linear_72.b_0" shape = [320] dtype = "float32" min_val = float("-0.130713") @@ -3052,6 +3332,7 @@ class Program_weight_tensor_parameter_279: class Program_weight_tensor_parameter_280: name = "parameter_280" + original_name = "linear_72.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.516144") @@ -3063,6 +3344,7 @@ class Program_weight_tensor_parameter_280: class Program_weight_tensor_parameter_281: name = "parameter_281" + original_name = "linear_71.b_0" shape = [640] dtype = "float32" min_val = float("-0.715721") @@ -3074,6 +3356,7 @@ class Program_weight_tensor_parameter_281: class Program_weight_tensor_parameter_282: name = "parameter_282" + original_name = "linear_71.w_0" shape = [320, 640] dtype = "float32" min_val = float("-0.380159") @@ -3085,6 +3368,7 @@ class Program_weight_tensor_parameter_282: class Program_weight_tensor_parameter_283: name = "parameter_283" + original_name = "layer_norm_49.b_0" shape = [320] dtype = "float32" min_val = float("-0.258705") @@ -3096,6 +3380,7 @@ class Program_weight_tensor_parameter_283: class Program_weight_tensor_parameter_284: name = "parameter_284" + original_name = "layer_norm_49.w_0" shape = [320] dtype = "float32" min_val = float("-0.0221611") @@ -3107,6 +3392,7 @@ class Program_weight_tensor_parameter_284: class Program_weight_tensor_parameter_285: name = "parameter_285" + original_name = "conv2d_32.b_0" shape = [320] dtype = "float32" min_val = float("-5.40314") @@ -3118,6 +3404,7 @@ class Program_weight_tensor_parameter_285: class Program_weight_tensor_parameter_286: name = "parameter_286" + original_name = "conv2d_32.w_0" shape = [320, 320, 2, 2] dtype = "float32" min_val = float("-0.344047") @@ -3129,6 +3416,7 @@ class Program_weight_tensor_parameter_286: class Program_weight_tensor_parameter_287: name = "parameter_287" + original_name = "linear_70.b_0" shape = [320] dtype = "float32" min_val = float("-2.77296") @@ -3140,6 +3428,7 @@ class Program_weight_tensor_parameter_287: class Program_weight_tensor_parameter_288: name = "parameter_288" + original_name = "linear_70.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.285939") @@ -3151,6 +3440,7 @@ class Program_weight_tensor_parameter_288: class Program_weight_tensor_parameter_289: name = "parameter_289" + original_name = "layer_norm_48.b_0" shape = [320] dtype = "float32" min_val = float("-0.570158") @@ -3162,6 +3452,7 @@ class Program_weight_tensor_parameter_289: class Program_weight_tensor_parameter_290: name = "parameter_290" + original_name = "layer_norm_48.w_0" shape = [320] dtype = "float32" min_val = float("0.129808") @@ -3173,6 +3464,7 @@ class Program_weight_tensor_parameter_290: class Program_weight_tensor_parameter_291: name = "parameter_291" + original_name = "linear_69.b_0" shape = [320] dtype = "float32" min_val = float("-0.338678") @@ -3184,6 +3476,7 @@ class Program_weight_tensor_parameter_291: class Program_weight_tensor_parameter_292: name = "parameter_292" + original_name = "linear_69.w_0" shape = [1280, 320] dtype = "float32" min_val = float("-0.399746") @@ -3195,6 +3488,7 @@ class Program_weight_tensor_parameter_292: class Program_weight_tensor_parameter_293: name = "parameter_293" + original_name = "conv2d_31.b_0" shape = [1280] dtype = "float32" min_val = float("-1.55261") @@ -3206,6 +3500,7 @@ class Program_weight_tensor_parameter_293: class Program_weight_tensor_parameter_294: name = "parameter_294" + original_name = "conv2d_31.w_0" shape = [1280, 1, 3, 3] dtype = "float32" min_val = float("-0.325473") @@ -3217,6 +3512,7 @@ class Program_weight_tensor_parameter_294: class Program_weight_tensor_parameter_295: name = "parameter_295" + original_name = "linear_68.b_0" shape = [1280] dtype = "float32" min_val = float("-1.57066") @@ -3228,6 +3524,7 @@ class Program_weight_tensor_parameter_295: class Program_weight_tensor_parameter_296: name = "parameter_296" + original_name = "linear_68.w_0" shape = [320, 1280] dtype = "float32" min_val = float("-0.40964") @@ -3239,6 +3536,7 @@ class Program_weight_tensor_parameter_296: class Program_weight_tensor_parameter_297: name = "parameter_297" + original_name = "layer_norm_47.b_0" shape = [320] dtype = "float32" min_val = float("-1.80023") @@ -3250,6 +3548,7 @@ class Program_weight_tensor_parameter_297: class Program_weight_tensor_parameter_298: name = "parameter_298" + original_name = "layer_norm_47.w_0" shape = [320] dtype = "float32" min_val = float("0.886209") @@ -3261,6 +3560,7 @@ class Program_weight_tensor_parameter_298: class Program_weight_tensor_parameter_299: name = "parameter_299" + original_name = "linear_67.b_0" shape = [320] dtype = "float32" min_val = float("-0.165045") @@ -3272,6 +3572,7 @@ class Program_weight_tensor_parameter_299: class Program_weight_tensor_parameter_300: name = "parameter_300" + original_name = "linear_67.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.374829") @@ -3283,6 +3584,7 @@ class Program_weight_tensor_parameter_300: class Program_weight_tensor_parameter_301: name = "parameter_301" + original_name = "linear_66.b_0" shape = [640] dtype = "float32" min_val = float("-0.663002") @@ -3294,6 +3596,7 @@ class Program_weight_tensor_parameter_301: class Program_weight_tensor_parameter_302: name = "parameter_302" + original_name = "linear_66.w_0" shape = [320, 640] dtype = "float32" min_val = float("-0.415669") @@ -3305,6 +3608,7 @@ class Program_weight_tensor_parameter_302: class Program_weight_tensor_parameter_303: name = "parameter_303" + original_name = "layer_norm_46.b_0" shape = [320] dtype = "float32" min_val = float("-0.211971") @@ -3316,6 +3620,7 @@ class Program_weight_tensor_parameter_303: class Program_weight_tensor_parameter_304: name = "parameter_304" + original_name = "layer_norm_46.w_0" shape = [320] dtype = "float32" min_val = float("-0.00619279") @@ -3327,6 +3632,7 @@ class Program_weight_tensor_parameter_304: class Program_weight_tensor_parameter_305: name = "parameter_305" + original_name = "conv2d_30.b_0" shape = [320] dtype = "float32" min_val = float("-7.18489") @@ -3338,6 +3644,7 @@ class Program_weight_tensor_parameter_305: class Program_weight_tensor_parameter_306: name = "parameter_306" + original_name = "conv2d_30.w_0" shape = [320, 320, 2, 2] dtype = "float32" min_val = float("-0.390317") @@ -3349,6 +3656,7 @@ class Program_weight_tensor_parameter_306: class Program_weight_tensor_parameter_307: name = "parameter_307" + original_name = "linear_65.b_0" shape = [320] dtype = "float32" min_val = float("-2.87293") @@ -3360,6 +3668,7 @@ class Program_weight_tensor_parameter_307: class Program_weight_tensor_parameter_308: name = "parameter_308" + original_name = "linear_65.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.268775") @@ -3371,6 +3680,7 @@ class Program_weight_tensor_parameter_308: class Program_weight_tensor_parameter_309: name = "parameter_309" + original_name = "layer_norm_45.b_0" shape = [320] dtype = "float32" min_val = float("-0.957414") @@ -3382,6 +3692,7 @@ class Program_weight_tensor_parameter_309: class Program_weight_tensor_parameter_310: name = "parameter_310" + original_name = "layer_norm_45.w_0" shape = [320] dtype = "float32" min_val = float("0.247425") @@ -3393,6 +3704,7 @@ class Program_weight_tensor_parameter_310: class Program_weight_tensor_parameter_311: name = "parameter_311" + original_name = "linear_64.b_0" shape = [320] dtype = "float32" min_val = float("-0.319988") @@ -3404,6 +3716,7 @@ class Program_weight_tensor_parameter_311: class Program_weight_tensor_parameter_312: name = "parameter_312" + original_name = "linear_64.w_0" shape = [1280, 320] dtype = "float32" min_val = float("-0.381741") @@ -3415,6 +3728,7 @@ class Program_weight_tensor_parameter_312: class Program_weight_tensor_parameter_313: name = "parameter_313" + original_name = "conv2d_29.b_0" shape = [1280] dtype = "float32" min_val = float("-1.72017") @@ -3426,6 +3740,7 @@ class Program_weight_tensor_parameter_313: class Program_weight_tensor_parameter_314: name = "parameter_314" + original_name = "conv2d_29.w_0" shape = [1280, 1, 3, 3] dtype = "float32" min_val = float("-0.307238") @@ -3437,6 +3752,7 @@ class Program_weight_tensor_parameter_314: class Program_weight_tensor_parameter_315: name = "parameter_315" + original_name = "linear_63.b_0" shape = [1280] dtype = "float32" min_val = float("-1.40428") @@ -3448,6 +3764,7 @@ class Program_weight_tensor_parameter_315: class Program_weight_tensor_parameter_316: name = "parameter_316" + original_name = "linear_63.w_0" shape = [320, 1280] dtype = "float32" min_val = float("-0.386223") @@ -3459,6 +3776,7 @@ class Program_weight_tensor_parameter_316: class Program_weight_tensor_parameter_317: name = "parameter_317" + original_name = "layer_norm_44.b_0" shape = [320] dtype = "float32" min_val = float("-1.71651") @@ -3470,6 +3788,7 @@ class Program_weight_tensor_parameter_317: class Program_weight_tensor_parameter_318: name = "parameter_318" + original_name = "layer_norm_44.w_0" shape = [320] dtype = "float32" min_val = float("0.81549") @@ -3481,6 +3800,7 @@ class Program_weight_tensor_parameter_318: class Program_weight_tensor_parameter_319: name = "parameter_319" + original_name = "linear_62.b_0" shape = [320] dtype = "float32" min_val = float("-0.255728") @@ -3492,6 +3812,7 @@ class Program_weight_tensor_parameter_319: class Program_weight_tensor_parameter_320: name = "parameter_320" + original_name = "linear_62.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.512906") @@ -3503,6 +3824,7 @@ class Program_weight_tensor_parameter_320: class Program_weight_tensor_parameter_321: name = "parameter_321" + original_name = "linear_61.b_0" shape = [640] dtype = "float32" min_val = float("-0.790195") @@ -3514,6 +3836,7 @@ class Program_weight_tensor_parameter_321: class Program_weight_tensor_parameter_322: name = "parameter_322" + original_name = "linear_61.w_0" shape = [320, 640] dtype = "float32" min_val = float("-0.411509") @@ -3525,6 +3848,7 @@ class Program_weight_tensor_parameter_322: class Program_weight_tensor_parameter_323: name = "parameter_323" + original_name = "layer_norm_43.b_0" shape = [320] dtype = "float32" min_val = float("-0.186814") @@ -3536,6 +3860,7 @@ class Program_weight_tensor_parameter_323: class Program_weight_tensor_parameter_324: name = "parameter_324" + original_name = "layer_norm_43.w_0" shape = [320] dtype = "float32" min_val = float("-0.00553587") @@ -3547,6 +3872,7 @@ class Program_weight_tensor_parameter_324: class Program_weight_tensor_parameter_325: name = "parameter_325" + original_name = "conv2d_28.b_0" shape = [320] dtype = "float32" min_val = float("-6.27505") @@ -3558,6 +3884,7 @@ class Program_weight_tensor_parameter_325: class Program_weight_tensor_parameter_326: name = "parameter_326" + original_name = "conv2d_28.w_0" shape = [320, 320, 2, 2] dtype = "float32" min_val = float("-0.32929") @@ -3569,6 +3896,7 @@ class Program_weight_tensor_parameter_326: class Program_weight_tensor_parameter_327: name = "parameter_327" + original_name = "linear_60.b_0" shape = [320] dtype = "float32" min_val = float("-2.9021") @@ -3580,6 +3908,7 @@ class Program_weight_tensor_parameter_327: class Program_weight_tensor_parameter_328: name = "parameter_328" + original_name = "linear_60.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.288858") @@ -3591,6 +3920,7 @@ class Program_weight_tensor_parameter_328: class Program_weight_tensor_parameter_329: name = "parameter_329" + original_name = "layer_norm_42.b_0" shape = [320] dtype = "float32" min_val = float("-0.664332") @@ -3602,6 +3932,7 @@ class Program_weight_tensor_parameter_329: class Program_weight_tensor_parameter_330: name = "parameter_330" + original_name = "layer_norm_42.w_0" shape = [320] dtype = "float32" min_val = float("0.167984") @@ -3613,6 +3944,7 @@ class Program_weight_tensor_parameter_330: class Program_weight_tensor_parameter_331: name = "parameter_331" + original_name = "linear_59.b_0" shape = [320] dtype = "float32" min_val = float("-0.397663") @@ -3624,6 +3956,7 @@ class Program_weight_tensor_parameter_331: class Program_weight_tensor_parameter_332: name = "parameter_332" + original_name = "linear_59.w_0" shape = [1280, 320] dtype = "float32" min_val = float("-0.333919") @@ -3635,6 +3968,7 @@ class Program_weight_tensor_parameter_332: class Program_weight_tensor_parameter_333: name = "parameter_333" + original_name = "conv2d_27.b_0" shape = [1280] dtype = "float32" min_val = float("-1.72626") @@ -3646,6 +3980,7 @@ class Program_weight_tensor_parameter_333: class Program_weight_tensor_parameter_334: name = "parameter_334" + original_name = "conv2d_27.w_0" shape = [1280, 1, 3, 3] dtype = "float32" min_val = float("-0.344388") @@ -3657,6 +3992,7 @@ class Program_weight_tensor_parameter_334: class Program_weight_tensor_parameter_335: name = "parameter_335" + original_name = "linear_58.b_0" shape = [1280] dtype = "float32" min_val = float("-1.98112") @@ -3668,6 +4004,7 @@ class Program_weight_tensor_parameter_335: class Program_weight_tensor_parameter_336: name = "parameter_336" + original_name = "linear_58.w_0" shape = [320, 1280] dtype = "float32" min_val = float("-0.331564") @@ -3679,6 +4016,7 @@ class Program_weight_tensor_parameter_336: class Program_weight_tensor_parameter_337: name = "parameter_337" + original_name = "layer_norm_41.b_0" shape = [320] dtype = "float32" min_val = float("-1.82069") @@ -3690,6 +4028,7 @@ class Program_weight_tensor_parameter_337: class Program_weight_tensor_parameter_338: name = "parameter_338" + original_name = "layer_norm_41.w_0" shape = [320] dtype = "float32" min_val = float("0.795017") @@ -3701,6 +4040,7 @@ class Program_weight_tensor_parameter_338: class Program_weight_tensor_parameter_339: name = "parameter_339" + original_name = "linear_57.b_0" shape = [320] dtype = "float32" min_val = float("-0.238886") @@ -3712,6 +4052,7 @@ class Program_weight_tensor_parameter_339: class Program_weight_tensor_parameter_340: name = "parameter_340" + original_name = "linear_57.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.458613") @@ -3723,6 +4064,7 @@ class Program_weight_tensor_parameter_340: class Program_weight_tensor_parameter_341: name = "parameter_341" + original_name = "linear_56.b_0" shape = [640] dtype = "float32" min_val = float("-0.983097") @@ -3734,6 +4076,7 @@ class Program_weight_tensor_parameter_341: class Program_weight_tensor_parameter_342: name = "parameter_342" + original_name = "linear_56.w_0" shape = [320, 640] dtype = "float32" min_val = float("-0.379278") @@ -3745,6 +4088,7 @@ class Program_weight_tensor_parameter_342: class Program_weight_tensor_parameter_343: name = "parameter_343" + original_name = "layer_norm_40.b_0" shape = [320] dtype = "float32" min_val = float("-0.210138") @@ -3756,6 +4100,7 @@ class Program_weight_tensor_parameter_343: class Program_weight_tensor_parameter_344: name = "parameter_344" + original_name = "layer_norm_40.w_0" shape = [320] dtype = "float32" min_val = float("-0.00828851") @@ -3767,6 +4112,7 @@ class Program_weight_tensor_parameter_344: class Program_weight_tensor_parameter_345: name = "parameter_345" + original_name = "conv2d_26.b_0" shape = [320] dtype = "float32" min_val = float("-5.32108") @@ -3778,6 +4124,7 @@ class Program_weight_tensor_parameter_345: class Program_weight_tensor_parameter_346: name = "parameter_346" + original_name = "conv2d_26.w_0" shape = [320, 320, 2, 2] dtype = "float32" min_val = float("-0.373719") @@ -3789,6 +4136,7 @@ class Program_weight_tensor_parameter_346: class Program_weight_tensor_parameter_347: name = "parameter_347" + original_name = "linear_55.b_0" shape = [320] dtype = "float32" min_val = float("-3.29479") @@ -3800,6 +4148,7 @@ class Program_weight_tensor_parameter_347: class Program_weight_tensor_parameter_348: name = "parameter_348" + original_name = "linear_55.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.299709") @@ -3811,6 +4160,7 @@ class Program_weight_tensor_parameter_348: class Program_weight_tensor_parameter_349: name = "parameter_349" + original_name = "layer_norm_39.b_0" shape = [320] dtype = "float32" min_val = float("-0.666155") @@ -3822,6 +4172,7 @@ class Program_weight_tensor_parameter_349: class Program_weight_tensor_parameter_350: name = "parameter_350" + original_name = "layer_norm_39.w_0" shape = [320] dtype = "float32" min_val = float("0.0638893") @@ -3833,6 +4184,7 @@ class Program_weight_tensor_parameter_350: class Program_weight_tensor_parameter_351: name = "parameter_351" + original_name = "linear_54.b_0" shape = [320] dtype = "float32" min_val = float("-0.404111") @@ -3844,6 +4196,7 @@ class Program_weight_tensor_parameter_351: class Program_weight_tensor_parameter_352: name = "parameter_352" + original_name = "linear_54.w_0" shape = [1280, 320] dtype = "float32" min_val = float("-0.287928") @@ -3855,6 +4208,7 @@ class Program_weight_tensor_parameter_352: class Program_weight_tensor_parameter_353: name = "parameter_353" + original_name = "conv2d_25.b_0" shape = [1280] dtype = "float32" min_val = float("-1.84936") @@ -3866,6 +4220,7 @@ class Program_weight_tensor_parameter_353: class Program_weight_tensor_parameter_354: name = "parameter_354" + original_name = "conv2d_25.w_0" shape = [1280, 1, 3, 3] dtype = "float32" min_val = float("-0.306933") @@ -3877,6 +4232,7 @@ class Program_weight_tensor_parameter_354: class Program_weight_tensor_parameter_355: name = "parameter_355" + original_name = "linear_53.b_0" shape = [1280] dtype = "float32" min_val = float("-2.11018") @@ -3888,6 +4244,7 @@ class Program_weight_tensor_parameter_355: class Program_weight_tensor_parameter_356: name = "parameter_356" + original_name = "linear_53.w_0" shape = [320, 1280] dtype = "float32" min_val = float("-0.306039") @@ -3899,6 +4256,7 @@ class Program_weight_tensor_parameter_356: class Program_weight_tensor_parameter_357: name = "parameter_357" + original_name = "layer_norm_38.b_0" shape = [320] dtype = "float32" min_val = float("-1.81231") @@ -3910,6 +4268,7 @@ class Program_weight_tensor_parameter_357: class Program_weight_tensor_parameter_358: name = "parameter_358" + original_name = "layer_norm_38.w_0" shape = [320] dtype = "float32" min_val = float("0.78473") @@ -3921,6 +4280,7 @@ class Program_weight_tensor_parameter_358: class Program_weight_tensor_parameter_359: name = "parameter_359" + original_name = "linear_52.b_0" shape = [320] dtype = "float32" min_val = float("-0.216306") @@ -3932,6 +4292,7 @@ class Program_weight_tensor_parameter_359: class Program_weight_tensor_parameter_360: name = "parameter_360" + original_name = "linear_52.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.333386") @@ -3943,6 +4304,7 @@ class Program_weight_tensor_parameter_360: class Program_weight_tensor_parameter_361: name = "parameter_361" + original_name = "linear_51.b_0" shape = [640] dtype = "float32" min_val = float("-0.794476") @@ -3954,6 +4316,7 @@ class Program_weight_tensor_parameter_361: class Program_weight_tensor_parameter_362: name = "parameter_362" + original_name = "linear_51.w_0" shape = [320, 640] dtype = "float32" min_val = float("-0.35342") @@ -3965,6 +4328,7 @@ class Program_weight_tensor_parameter_362: class Program_weight_tensor_parameter_363: name = "parameter_363" + original_name = "layer_norm_37.b_0" shape = [320] dtype = "float32" min_val = float("-0.217595") @@ -3976,6 +4340,7 @@ class Program_weight_tensor_parameter_363: class Program_weight_tensor_parameter_364: name = "parameter_364" + original_name = "layer_norm_37.w_0" shape = [320] dtype = "float32" min_val = float("-0.0143471") @@ -3987,6 +4352,7 @@ class Program_weight_tensor_parameter_364: class Program_weight_tensor_parameter_365: name = "parameter_365" + original_name = "conv2d_24.b_0" shape = [320] dtype = "float32" min_val = float("-5.71788") @@ -3998,6 +4364,7 @@ class Program_weight_tensor_parameter_365: class Program_weight_tensor_parameter_366: name = "parameter_366" + original_name = "conv2d_24.w_0" shape = [320, 320, 2, 2] dtype = "float32" min_val = float("-0.265807") @@ -4009,6 +4376,7 @@ class Program_weight_tensor_parameter_366: class Program_weight_tensor_parameter_367: name = "parameter_367" + original_name = "linear_50.b_0" shape = [320] dtype = "float32" min_val = float("-2.57946") @@ -4020,6 +4388,7 @@ class Program_weight_tensor_parameter_367: class Program_weight_tensor_parameter_368: name = "parameter_368" + original_name = "linear_50.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.382365") @@ -4031,6 +4400,7 @@ class Program_weight_tensor_parameter_368: class Program_weight_tensor_parameter_369: name = "parameter_369" + original_name = "layer_norm_36.b_0" shape = [320] dtype = "float32" min_val = float("-0.762153") @@ -4042,6 +4412,7 @@ class Program_weight_tensor_parameter_369: class Program_weight_tensor_parameter_370: name = "parameter_370" + original_name = "layer_norm_36.w_0" shape = [320] dtype = "float32" min_val = float("-0.012421") @@ -4053,6 +4424,7 @@ class Program_weight_tensor_parameter_370: class Program_weight_tensor_parameter_371: name = "parameter_371" + original_name = "linear_49.b_0" shape = [320] dtype = "float32" min_val = float("-0.389952") @@ -4064,6 +4436,7 @@ class Program_weight_tensor_parameter_371: class Program_weight_tensor_parameter_372: name = "parameter_372" + original_name = "linear_49.w_0" shape = [1280, 320] dtype = "float32" min_val = float("-0.271814") @@ -4075,6 +4448,7 @@ class Program_weight_tensor_parameter_372: class Program_weight_tensor_parameter_373: name = "parameter_373" + original_name = "conv2d_23.b_0" shape = [1280] dtype = "float32" min_val = float("-1.77638") @@ -4086,6 +4460,7 @@ class Program_weight_tensor_parameter_373: class Program_weight_tensor_parameter_374: name = "parameter_374" + original_name = "conv2d_23.w_0" shape = [1280, 1, 3, 3] dtype = "float32" min_val = float("-0.311052") @@ -4097,6 +4472,7 @@ class Program_weight_tensor_parameter_374: class Program_weight_tensor_parameter_375: name = "parameter_375" + original_name = "linear_48.b_0" shape = [1280] dtype = "float32" min_val = float("-2.88059") @@ -4108,6 +4484,7 @@ class Program_weight_tensor_parameter_375: class Program_weight_tensor_parameter_376: name = "parameter_376" + original_name = "linear_48.w_0" shape = [320, 1280] dtype = "float32" min_val = float("-0.30887") @@ -4119,6 +4496,7 @@ class Program_weight_tensor_parameter_376: class Program_weight_tensor_parameter_377: name = "parameter_377" + original_name = "layer_norm_35.b_0" shape = [320] dtype = "float32" min_val = float("-1.82889") @@ -4130,6 +4508,7 @@ class Program_weight_tensor_parameter_377: class Program_weight_tensor_parameter_378: name = "parameter_378" + original_name = "layer_norm_35.w_0" shape = [320] dtype = "float32" min_val = float("0.714881") @@ -4141,6 +4520,7 @@ class Program_weight_tensor_parameter_378: class Program_weight_tensor_parameter_379: name = "parameter_379" + original_name = "linear_47.b_0" shape = [320] dtype = "float32" min_val = float("-0.187008") @@ -4152,6 +4532,7 @@ class Program_weight_tensor_parameter_379: class Program_weight_tensor_parameter_380: name = "parameter_380" + original_name = "linear_47.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.215116") @@ -4163,6 +4544,7 @@ class Program_weight_tensor_parameter_380: class Program_weight_tensor_parameter_381: name = "parameter_381" + original_name = "linear_46.b_0" shape = [640] dtype = "float32" min_val = float("-0.920291") @@ -4174,6 +4556,7 @@ class Program_weight_tensor_parameter_381: class Program_weight_tensor_parameter_382: name = "parameter_382" + original_name = "linear_46.w_0" shape = [320, 640] dtype = "float32" min_val = float("-0.308666") @@ -4185,6 +4568,7 @@ class Program_weight_tensor_parameter_382: class Program_weight_tensor_parameter_383: name = "parameter_383" + original_name = "layer_norm_34.b_0" shape = [320] dtype = "float32" min_val = float("-0.150812") @@ -4196,6 +4580,7 @@ class Program_weight_tensor_parameter_383: class Program_weight_tensor_parameter_384: name = "parameter_384" + original_name = "layer_norm_34.w_0" shape = [320] dtype = "float32" min_val = float("-0.0255246") @@ -4207,6 +4592,7 @@ class Program_weight_tensor_parameter_384: class Program_weight_tensor_parameter_385: name = "parameter_385" + original_name = "conv2d_22.b_0" shape = [320] dtype = "float32" min_val = float("-4.48454") @@ -4218,6 +4604,7 @@ class Program_weight_tensor_parameter_385: class Program_weight_tensor_parameter_386: name = "parameter_386" + original_name = "conv2d_22.w_0" shape = [320, 320, 2, 2] dtype = "float32" min_val = float("-0.242553") @@ -4229,6 +4616,7 @@ class Program_weight_tensor_parameter_386: class Program_weight_tensor_parameter_387: name = "parameter_387" + original_name = "linear_45.b_0" shape = [320] dtype = "float32" min_val = float("-3.08071") @@ -4240,6 +4628,7 @@ class Program_weight_tensor_parameter_387: class Program_weight_tensor_parameter_388: name = "parameter_388" + original_name = "linear_45.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.257644") @@ -4251,6 +4640,7 @@ class Program_weight_tensor_parameter_388: class Program_weight_tensor_parameter_389: name = "parameter_389" + original_name = "layer_norm_33.b_0" shape = [320] dtype = "float32" min_val = float("-0.490992") @@ -4262,6 +4652,7 @@ class Program_weight_tensor_parameter_389: class Program_weight_tensor_parameter_390: name = "parameter_390" + original_name = "layer_norm_33.w_0" shape = [320] dtype = "float32" min_val = float("-0.00424516") @@ -4273,6 +4664,7 @@ class Program_weight_tensor_parameter_390: class Program_weight_tensor_parameter_391: name = "parameter_391" + original_name = "linear_44.b_0" shape = [320] dtype = "float32" min_val = float("-0.491151") @@ -4284,6 +4676,7 @@ class Program_weight_tensor_parameter_391: class Program_weight_tensor_parameter_392: name = "parameter_392" + original_name = "linear_44.w_0" shape = [1280, 320] dtype = "float32" min_val = float("-0.294021") @@ -4295,6 +4688,7 @@ class Program_weight_tensor_parameter_392: class Program_weight_tensor_parameter_393: name = "parameter_393" + original_name = "conv2d_21.b_0" shape = [1280] dtype = "float32" min_val = float("-1.90305") @@ -4306,6 +4700,7 @@ class Program_weight_tensor_parameter_393: class Program_weight_tensor_parameter_394: name = "parameter_394" + original_name = "conv2d_21.w_0" shape = [1280, 1, 3, 3] dtype = "float32" min_val = float("-0.260666") @@ -4317,6 +4712,7 @@ class Program_weight_tensor_parameter_394: class Program_weight_tensor_parameter_395: name = "parameter_395" + original_name = "linear_43.b_0" shape = [1280] dtype = "float32" min_val = float("-3.12463") @@ -4328,6 +4724,7 @@ class Program_weight_tensor_parameter_395: class Program_weight_tensor_parameter_396: name = "parameter_396" + original_name = "linear_43.w_0" shape = [320, 1280] dtype = "float32" min_val = float("-0.307825") @@ -4339,6 +4736,7 @@ class Program_weight_tensor_parameter_396: class Program_weight_tensor_parameter_397: name = "parameter_397" + original_name = "layer_norm_32.b_0" shape = [320] dtype = "float32" min_val = float("-1.83558") @@ -4350,6 +4748,7 @@ class Program_weight_tensor_parameter_397: class Program_weight_tensor_parameter_398: name = "parameter_398" + original_name = "layer_norm_32.w_0" shape = [320] dtype = "float32" min_val = float("0.504337") @@ -4361,6 +4760,7 @@ class Program_weight_tensor_parameter_398: class Program_weight_tensor_parameter_399: name = "parameter_399" + original_name = "linear_42.b_0" shape = [320] dtype = "float32" min_val = float("-0.13289") @@ -4372,6 +4772,7 @@ class Program_weight_tensor_parameter_399: class Program_weight_tensor_parameter_400: name = "parameter_400" + original_name = "linear_42.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.226877") @@ -4383,6 +4784,7 @@ class Program_weight_tensor_parameter_400: class Program_weight_tensor_parameter_401: name = "parameter_401" + original_name = "linear_41.b_0" shape = [640] dtype = "float32" min_val = float("-0.707951") @@ -4394,6 +4796,7 @@ class Program_weight_tensor_parameter_401: class Program_weight_tensor_parameter_402: name = "parameter_402" + original_name = "linear_41.w_0" shape = [320, 640] dtype = "float32" min_val = float("-0.334262") @@ -4405,6 +4808,7 @@ class Program_weight_tensor_parameter_402: class Program_weight_tensor_parameter_403: name = "parameter_403" + original_name = "layer_norm_31.b_0" shape = [320] dtype = "float32" min_val = float("-0.166555") @@ -4416,6 +4820,7 @@ class Program_weight_tensor_parameter_403: class Program_weight_tensor_parameter_404: name = "parameter_404" + original_name = "layer_norm_31.w_0" shape = [320] dtype = "float32" min_val = float("-0.0282381") @@ -4427,6 +4832,7 @@ class Program_weight_tensor_parameter_404: class Program_weight_tensor_parameter_405: name = "parameter_405" + original_name = "conv2d_20.b_0" shape = [320] dtype = "float32" min_val = float("-4.47267") @@ -4438,6 +4844,7 @@ class Program_weight_tensor_parameter_405: class Program_weight_tensor_parameter_406: name = "parameter_406" + original_name = "conv2d_20.w_0" shape = [320, 320, 2, 2] dtype = "float32" min_val = float("-0.303896") @@ -4449,6 +4856,7 @@ class Program_weight_tensor_parameter_406: class Program_weight_tensor_parameter_407: name = "parameter_407" + original_name = "linear_40.b_0" shape = [320] dtype = "float32" min_val = float("-3.26069") @@ -4460,6 +4868,7 @@ class Program_weight_tensor_parameter_407: class Program_weight_tensor_parameter_408: name = "parameter_408" + original_name = "linear_40.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.303904") @@ -4471,6 +4880,7 @@ class Program_weight_tensor_parameter_408: class Program_weight_tensor_parameter_409: name = "parameter_409" + original_name = "layer_norm_30.b_0" shape = [320] dtype = "float32" min_val = float("-0.823595") @@ -4482,6 +4892,7 @@ class Program_weight_tensor_parameter_409: class Program_weight_tensor_parameter_410: name = "parameter_410" + original_name = "layer_norm_30.w_0" shape = [320] dtype = "float32" min_val = float("-0.00632425") @@ -4493,6 +4904,7 @@ class Program_weight_tensor_parameter_410: class Program_weight_tensor_parameter_411: name = "parameter_411" + original_name = "linear_39.b_0" shape = [320] dtype = "float32" min_val = float("-0.616079") @@ -4504,6 +4916,7 @@ class Program_weight_tensor_parameter_411: class Program_weight_tensor_parameter_412: name = "parameter_412" + original_name = "linear_39.w_0" shape = [1280, 320] dtype = "float32" min_val = float("-0.279556") @@ -4515,6 +4928,7 @@ class Program_weight_tensor_parameter_412: class Program_weight_tensor_parameter_413: name = "parameter_413" + original_name = "conv2d_19.b_0" shape = [1280] dtype = "float32" min_val = float("-1.9688") @@ -4526,6 +4940,7 @@ class Program_weight_tensor_parameter_413: class Program_weight_tensor_parameter_414: name = "parameter_414" + original_name = "conv2d_19.w_0" shape = [1280, 1, 3, 3] dtype = "float32" min_val = float("-0.247543") @@ -4537,6 +4952,7 @@ class Program_weight_tensor_parameter_414: class Program_weight_tensor_parameter_415: name = "parameter_415" + original_name = "linear_38.b_0" shape = [1280] dtype = "float32" min_val = float("-2.18298") @@ -4548,6 +4964,7 @@ class Program_weight_tensor_parameter_415: class Program_weight_tensor_parameter_416: name = "parameter_416" + original_name = "linear_38.w_0" shape = [320, 1280] dtype = "float32" min_val = float("-0.315811") @@ -4559,6 +4976,7 @@ class Program_weight_tensor_parameter_416: class Program_weight_tensor_parameter_417: name = "parameter_417" + original_name = "layer_norm_29.b_0" shape = [320] dtype = "float32" min_val = float("-1.43853") @@ -4570,6 +4988,7 @@ class Program_weight_tensor_parameter_417: class Program_weight_tensor_parameter_418: name = "parameter_418" + original_name = "layer_norm_29.w_0" shape = [320] dtype = "float32" min_val = float("0.571226") @@ -4581,6 +5000,7 @@ class Program_weight_tensor_parameter_418: class Program_weight_tensor_parameter_419: name = "parameter_419" + original_name = "linear_37.b_0" shape = [320] dtype = "float32" min_val = float("-0.210356") @@ -4592,6 +5012,7 @@ class Program_weight_tensor_parameter_419: class Program_weight_tensor_parameter_420: name = "parameter_420" + original_name = "linear_37.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.207657") @@ -4603,6 +5024,7 @@ class Program_weight_tensor_parameter_420: class Program_weight_tensor_parameter_421: name = "parameter_421" + original_name = "linear_36.b_0" shape = [640] dtype = "float32" min_val = float("-0.719059") @@ -4614,6 +5036,7 @@ class Program_weight_tensor_parameter_421: class Program_weight_tensor_parameter_422: name = "parameter_422" + original_name = "linear_36.w_0" shape = [320, 640] dtype = "float32" min_val = float("-0.448279") @@ -4625,6 +5048,7 @@ class Program_weight_tensor_parameter_422: class Program_weight_tensor_parameter_423: name = "parameter_423" + original_name = "layer_norm_28.b_0" shape = [320] dtype = "float32" min_val = float("-0.103162") @@ -4636,6 +5060,7 @@ class Program_weight_tensor_parameter_423: class Program_weight_tensor_parameter_424: name = "parameter_424" + original_name = "layer_norm_28.w_0" shape = [320] dtype = "float32" min_val = float("-0.0993432") @@ -4647,6 +5072,7 @@ class Program_weight_tensor_parameter_424: class Program_weight_tensor_parameter_425: name = "parameter_425" + original_name = "conv2d_18.b_0" shape = [320] dtype = "float32" min_val = float("-4.11193") @@ -4658,6 +5084,7 @@ class Program_weight_tensor_parameter_425: class Program_weight_tensor_parameter_426: name = "parameter_426" + original_name = "conv2d_18.w_0" shape = [320, 320, 2, 2] dtype = "float32" min_val = float("-0.235165") @@ -4669,6 +5096,7 @@ class Program_weight_tensor_parameter_426: class Program_weight_tensor_parameter_427: name = "parameter_427" + original_name = "linear_35.b_0" shape = [320] dtype = "float32" min_val = float("-2.99429") @@ -4680,6 +5108,7 @@ class Program_weight_tensor_parameter_427: class Program_weight_tensor_parameter_428: name = "parameter_428" + original_name = "linear_35.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.310488") @@ -4691,6 +5120,7 @@ class Program_weight_tensor_parameter_428: class Program_weight_tensor_parameter_429: name = "parameter_429" + original_name = "layer_norm_27.b_0" shape = [320] dtype = "float32" min_val = float("-1.41554") @@ -4702,6 +5132,7 @@ class Program_weight_tensor_parameter_429: class Program_weight_tensor_parameter_430: name = "parameter_430" + original_name = "layer_norm_27.w_0" shape = [320] dtype = "float32" min_val = float("-0.00554277") @@ -4713,6 +5144,7 @@ class Program_weight_tensor_parameter_430: class Program_weight_tensor_parameter_431: name = "parameter_431" + original_name = "layer_norm_2.b_0" shape = [320] dtype = "float32" min_val = float("-0.570871") @@ -4724,6 +5156,7 @@ class Program_weight_tensor_parameter_431: class Program_weight_tensor_parameter_432: name = "parameter_432" + original_name = "layer_norm_2.w_0" shape = [320] dtype = "float32" min_val = float("0.112588") @@ -4735,6 +5168,7 @@ class Program_weight_tensor_parameter_432: class Program_weight_tensor_parameter_433: name = "parameter_433" + original_name = "conv2d_2.b_0" shape = [320] dtype = "float32" min_val = float("-11.6033") @@ -4746,6 +5180,7 @@ class Program_weight_tensor_parameter_433: class Program_weight_tensor_parameter_434: name = "parameter_434" + original_name = "conv2d_2.w_0" shape = [320, 128, 3, 3] dtype = "float32" min_val = float("-1.58912") @@ -4757,6 +5192,7 @@ class Program_weight_tensor_parameter_434: class Program_weight_tensor_parameter_435: name = "parameter_435" + original_name = "layer_norm_26.b_0" shape = [128] dtype = "float32" min_val = float("-0.375724") @@ -4768,6 +5204,7 @@ class Program_weight_tensor_parameter_435: class Program_weight_tensor_parameter_436: name = "parameter_436" + original_name = "layer_norm_26.w_0" shape = [128] dtype = "float32" min_val = float("0.0908502") @@ -4779,6 +5216,7 @@ class Program_weight_tensor_parameter_436: class Program_weight_tensor_parameter_437: name = "parameter_437" + original_name = "linear_34.b_0" shape = [128] dtype = "float32" min_val = float("-9.62662") @@ -4790,6 +5228,7 @@ class Program_weight_tensor_parameter_437: class Program_weight_tensor_parameter_438: name = "parameter_438" + original_name = "linear_34.w_0" shape = [512, 128] dtype = "float32" min_val = float("-1.72939") @@ -4801,6 +5240,7 @@ class Program_weight_tensor_parameter_438: class Program_weight_tensor_parameter_439: name = "parameter_439" + original_name = "conv2d_17.b_0" shape = [512] dtype = "float32" min_val = float("-1.25946") @@ -4812,6 +5252,7 @@ class Program_weight_tensor_parameter_439: class Program_weight_tensor_parameter_440: name = "parameter_440" + original_name = "conv2d_17.w_0" shape = [512, 1, 3, 3] dtype = "float32" min_val = float("-0.231996") @@ -4823,6 +5264,7 @@ class Program_weight_tensor_parameter_440: class Program_weight_tensor_parameter_441: name = "parameter_441" + original_name = "linear_33.b_0" shape = [512] dtype = "float32" min_val = float("-2.27177") @@ -4834,6 +5276,7 @@ class Program_weight_tensor_parameter_441: class Program_weight_tensor_parameter_442: name = "parameter_442" + original_name = "linear_33.w_0" shape = [128, 512] dtype = "float32" min_val = float("-0.284037") @@ -4845,6 +5288,7 @@ class Program_weight_tensor_parameter_442: class Program_weight_tensor_parameter_443: name = "parameter_443" + original_name = "layer_norm_25.b_0" shape = [128] dtype = "float32" min_val = float("-2.00562") @@ -4856,6 +5300,7 @@ class Program_weight_tensor_parameter_443: class Program_weight_tensor_parameter_444: name = "parameter_444" + original_name = "layer_norm_25.w_0" shape = [128] dtype = "float32" min_val = float("1.1314") @@ -4867,6 +5312,7 @@ class Program_weight_tensor_parameter_444: class Program_weight_tensor_parameter_445: name = "parameter_445" + original_name = "linear_32.b_0" shape = [128] dtype = "float32" min_val = float("-0.511934") @@ -4878,6 +5324,7 @@ class Program_weight_tensor_parameter_445: class Program_weight_tensor_parameter_446: name = "parameter_446" + original_name = "linear_32.w_0" shape = [128, 128] dtype = "float32" min_val = float("-0.708884") @@ -4889,6 +5336,7 @@ class Program_weight_tensor_parameter_446: class Program_weight_tensor_parameter_447: name = "parameter_447" + original_name = "linear_31.b_0" shape = [256] dtype = "float32" min_val = float("-1.02247") @@ -4900,6 +5348,7 @@ class Program_weight_tensor_parameter_447: class Program_weight_tensor_parameter_448: name = "parameter_448" + original_name = "linear_31.w_0" shape = [128, 256] dtype = "float32" min_val = float("-0.359115") @@ -4911,6 +5360,7 @@ class Program_weight_tensor_parameter_448: class Program_weight_tensor_parameter_449: name = "parameter_449" + original_name = "layer_norm_24.b_0" shape = [128] dtype = "float32" min_val = float("-0.674521") @@ -4922,6 +5372,7 @@ class Program_weight_tensor_parameter_449: class Program_weight_tensor_parameter_450: name = "parameter_450" + original_name = "layer_norm_24.w_0" shape = [128] dtype = "float32" min_val = float("-0.0213703") @@ -4933,6 +5384,7 @@ class Program_weight_tensor_parameter_450: class Program_weight_tensor_parameter_451: name = "parameter_451" + original_name = "conv2d_16.b_0" shape = [128] dtype = "float32" min_val = float("-2.52489") @@ -4944,6 +5396,7 @@ class Program_weight_tensor_parameter_451: class Program_weight_tensor_parameter_452: name = "parameter_452" + original_name = "conv2d_16.w_0" shape = [128, 128, 4, 4] dtype = "float32" min_val = float("-0.258079") @@ -4955,6 +5408,7 @@ class Program_weight_tensor_parameter_452: class Program_weight_tensor_parameter_453: name = "parameter_453" + original_name = "linear_30.b_0" shape = [128] dtype = "float32" min_val = float("-2.67692") @@ -4966,6 +5420,7 @@ class Program_weight_tensor_parameter_453: class Program_weight_tensor_parameter_454: name = "parameter_454" + original_name = "linear_30.w_0" shape = [128, 128] dtype = "float32" min_val = float("-0.247492") @@ -4977,6 +5432,7 @@ class Program_weight_tensor_parameter_454: class Program_weight_tensor_parameter_455: name = "parameter_455" + original_name = "layer_norm_23.b_0" shape = [128] dtype = "float32" min_val = float("-0.804122") @@ -4988,6 +5444,7 @@ class Program_weight_tensor_parameter_455: class Program_weight_tensor_parameter_456: name = "parameter_456" + original_name = "layer_norm_23.w_0" shape = [128] dtype = "float32" min_val = float("0.366938") @@ -4999,6 +5456,7 @@ class Program_weight_tensor_parameter_456: class Program_weight_tensor_parameter_457: name = "parameter_457" + original_name = "linear_29.b_0" shape = [128] dtype = "float32" min_val = float("-2.50052") @@ -5010,6 +5468,7 @@ class Program_weight_tensor_parameter_457: class Program_weight_tensor_parameter_458: name = "parameter_458" + original_name = "linear_29.w_0" shape = [512, 128] dtype = "float32" min_val = float("-0.875382") @@ -5021,6 +5480,7 @@ class Program_weight_tensor_parameter_458: class Program_weight_tensor_parameter_459: name = "parameter_459" + original_name = "conv2d_15.b_0" shape = [512] dtype = "float32" min_val = float("-1.24779") @@ -5032,6 +5492,7 @@ class Program_weight_tensor_parameter_459: class Program_weight_tensor_parameter_460: name = "parameter_460" + original_name = "conv2d_15.w_0" shape = [512, 1, 3, 3] dtype = "float32" min_val = float("-0.278901") @@ -5043,6 +5504,7 @@ class Program_weight_tensor_parameter_460: class Program_weight_tensor_parameter_461: name = "parameter_461" + original_name = "linear_28.b_0" shape = [512] dtype = "float32" min_val = float("-7.03747") @@ -5054,6 +5516,7 @@ class Program_weight_tensor_parameter_461: class Program_weight_tensor_parameter_462: name = "parameter_462" + original_name = "linear_28.w_0" shape = [128, 512] dtype = "float32" min_val = float("-0.308231") @@ -5065,6 +5528,7 @@ class Program_weight_tensor_parameter_462: class Program_weight_tensor_parameter_463: name = "parameter_463" + original_name = "layer_norm_22.b_0" shape = [128] dtype = "float32" min_val = float("-1.35898") @@ -5076,6 +5540,7 @@ class Program_weight_tensor_parameter_463: class Program_weight_tensor_parameter_464: name = "parameter_464" + original_name = "layer_norm_22.w_0" shape = [128] dtype = "float32" min_val = float("1.35904") @@ -5087,6 +5552,7 @@ class Program_weight_tensor_parameter_464: class Program_weight_tensor_parameter_465: name = "parameter_465" + original_name = "linear_27.b_0" shape = [128] dtype = "float32" min_val = float("-0.0737803") @@ -5098,6 +5564,7 @@ class Program_weight_tensor_parameter_465: class Program_weight_tensor_parameter_466: name = "parameter_466" + original_name = "linear_27.w_0" shape = [128, 128] dtype = "float32" min_val = float("-0.401669") @@ -5109,6 +5576,7 @@ class Program_weight_tensor_parameter_466: class Program_weight_tensor_parameter_467: name = "parameter_467" + original_name = "linear_26.b_0" shape = [256] dtype = "float32" min_val = float("-0.841314") @@ -5120,6 +5588,7 @@ class Program_weight_tensor_parameter_467: class Program_weight_tensor_parameter_468: name = "parameter_468" + original_name = "linear_26.w_0" shape = [128, 256] dtype = "float32" min_val = float("-0.300935") @@ -5131,6 +5600,7 @@ class Program_weight_tensor_parameter_468: class Program_weight_tensor_parameter_469: name = "parameter_469" + original_name = "layer_norm_21.b_0" shape = [128] dtype = "float32" min_val = float("-0.492066") @@ -5142,6 +5612,7 @@ class Program_weight_tensor_parameter_469: class Program_weight_tensor_parameter_470: name = "parameter_470" + original_name = "layer_norm_21.w_0" shape = [128] dtype = "float32" min_val = float("-0.0094024") @@ -5153,6 +5624,7 @@ class Program_weight_tensor_parameter_470: class Program_weight_tensor_parameter_471: name = "parameter_471" + original_name = "conv2d_14.b_0" shape = [128] dtype = "float32" min_val = float("-2.10385") @@ -5164,6 +5636,7 @@ class Program_weight_tensor_parameter_471: class Program_weight_tensor_parameter_472: name = "parameter_472" + original_name = "conv2d_14.w_0" shape = [128, 128, 4, 4] dtype = "float32" min_val = float("-0.232168") @@ -5175,6 +5648,7 @@ class Program_weight_tensor_parameter_472: class Program_weight_tensor_parameter_473: name = "parameter_473" + original_name = "linear_25.b_0" shape = [128] dtype = "float32" min_val = float("-2.32466") @@ -5186,6 +5660,7 @@ class Program_weight_tensor_parameter_473: class Program_weight_tensor_parameter_474: name = "parameter_474" + original_name = "linear_25.w_0" shape = [128, 128] dtype = "float32" min_val = float("-0.313488") @@ -5197,6 +5672,7 @@ class Program_weight_tensor_parameter_474: class Program_weight_tensor_parameter_475: name = "parameter_475" + original_name = "layer_norm_20.b_0" shape = [128] dtype = "float32" min_val = float("-0.604607") @@ -5208,6 +5684,7 @@ class Program_weight_tensor_parameter_475: class Program_weight_tensor_parameter_476: name = "parameter_476" + original_name = "layer_norm_20.w_0" shape = [128] dtype = "float32" min_val = float("0.406153") @@ -5219,6 +5696,7 @@ class Program_weight_tensor_parameter_476: class Program_weight_tensor_parameter_477: name = "parameter_477" + original_name = "linear_24.b_0" shape = [128] dtype = "float32" min_val = float("-1.49585") @@ -5230,6 +5708,7 @@ class Program_weight_tensor_parameter_477: class Program_weight_tensor_parameter_478: name = "parameter_478" + original_name = "linear_24.w_0" shape = [512, 128] dtype = "float32" min_val = float("-0.6684") @@ -5241,6 +5720,7 @@ class Program_weight_tensor_parameter_478: class Program_weight_tensor_parameter_479: name = "parameter_479" + original_name = "conv2d_13.b_0" shape = [512] dtype = "float32" min_val = float("-2.10741") @@ -5252,6 +5732,7 @@ class Program_weight_tensor_parameter_479: class Program_weight_tensor_parameter_480: name = "parameter_480" + original_name = "conv2d_13.w_0" shape = [512, 1, 3, 3] dtype = "float32" min_val = float("-0.270409") @@ -5263,6 +5744,7 @@ class Program_weight_tensor_parameter_480: class Program_weight_tensor_parameter_481: name = "parameter_481" + original_name = "linear_23.b_0" shape = [512] dtype = "float32" min_val = float("-3.61496") @@ -5274,6 +5756,7 @@ class Program_weight_tensor_parameter_481: class Program_weight_tensor_parameter_482: name = "parameter_482" + original_name = "linear_23.w_0" shape = [128, 512] dtype = "float32" min_val = float("-0.308736") @@ -5285,6 +5768,7 @@ class Program_weight_tensor_parameter_482: class Program_weight_tensor_parameter_483: name = "parameter_483" + original_name = "layer_norm_19.b_0" shape = [128] dtype = "float32" min_val = float("-0.851445") @@ -5296,6 +5780,7 @@ class Program_weight_tensor_parameter_483: class Program_weight_tensor_parameter_484: name = "parameter_484" + original_name = "layer_norm_19.w_0" shape = [128] dtype = "float32" min_val = float("1.72984") @@ -5307,6 +5792,7 @@ class Program_weight_tensor_parameter_484: class Program_weight_tensor_parameter_485: name = "parameter_485" + original_name = "linear_22.b_0" shape = [128] dtype = "float32" min_val = float("-0.0993888") @@ -5318,6 +5804,7 @@ class Program_weight_tensor_parameter_485: class Program_weight_tensor_parameter_486: name = "parameter_486" + original_name = "linear_22.w_0" shape = [128, 128] dtype = "float32" min_val = float("-0.20069") @@ -5329,6 +5816,7 @@ class Program_weight_tensor_parameter_486: class Program_weight_tensor_parameter_487: name = "parameter_487" + original_name = "linear_21.b_0" shape = [256] dtype = "float32" min_val = float("-0.864138") @@ -5340,6 +5828,7 @@ class Program_weight_tensor_parameter_487: class Program_weight_tensor_parameter_488: name = "parameter_488" + original_name = "linear_21.w_0" shape = [128, 256] dtype = "float32" min_val = float("-0.409152") @@ -5351,6 +5840,7 @@ class Program_weight_tensor_parameter_488: class Program_weight_tensor_parameter_489: name = "parameter_489" + original_name = "layer_norm_18.b_0" shape = [128] dtype = "float32" min_val = float("-0.344777") @@ -5362,6 +5852,7 @@ class Program_weight_tensor_parameter_489: class Program_weight_tensor_parameter_490: name = "parameter_490" + original_name = "layer_norm_18.w_0" shape = [128] dtype = "float32" min_val = float("-0.0081214") @@ -5373,6 +5864,7 @@ class Program_weight_tensor_parameter_490: class Program_weight_tensor_parameter_491: name = "parameter_491" + original_name = "conv2d_12.b_0" shape = [128] dtype = "float32" min_val = float("-3.4722") @@ -5384,6 +5876,7 @@ class Program_weight_tensor_parameter_491: class Program_weight_tensor_parameter_492: name = "parameter_492" + original_name = "conv2d_12.w_0" shape = [128, 128, 4, 4] dtype = "float32" min_val = float("-0.286959") @@ -5395,6 +5888,7 @@ class Program_weight_tensor_parameter_492: class Program_weight_tensor_parameter_493: name = "parameter_493" + original_name = "linear_20.b_0" shape = [128] dtype = "float32" min_val = float("-2.34107") @@ -5406,6 +5900,7 @@ class Program_weight_tensor_parameter_493: class Program_weight_tensor_parameter_494: name = "parameter_494" + original_name = "linear_20.w_0" shape = [128, 128] dtype = "float32" min_val = float("-0.278176") @@ -5417,6 +5912,7 @@ class Program_weight_tensor_parameter_494: class Program_weight_tensor_parameter_495: name = "parameter_495" + original_name = "layer_norm_17.b_0" shape = [128] dtype = "float32" min_val = float("-0.796629") @@ -5428,6 +5924,7 @@ class Program_weight_tensor_parameter_495: class Program_weight_tensor_parameter_496: name = "parameter_496" + original_name = "layer_norm_17.w_0" shape = [128] dtype = "float32" min_val = float("0.619536") @@ -5439,6 +5936,7 @@ class Program_weight_tensor_parameter_496: class Program_weight_tensor_parameter_497: name = "parameter_497" + original_name = "linear_19.b_0" shape = [128] dtype = "float32" min_val = float("-1.04256") @@ -5450,6 +5948,7 @@ class Program_weight_tensor_parameter_497: class Program_weight_tensor_parameter_498: name = "parameter_498" + original_name = "linear_19.w_0" shape = [512, 128] dtype = "float32" min_val = float("-0.294681") @@ -5461,6 +5960,7 @@ class Program_weight_tensor_parameter_498: class Program_weight_tensor_parameter_499: name = "parameter_499" + original_name = "conv2d_11.b_0" shape = [512] dtype = "float32" min_val = float("-2.20913") @@ -5472,6 +5972,7 @@ class Program_weight_tensor_parameter_499: class Program_weight_tensor_parameter_500: name = "parameter_500" + original_name = "conv2d_11.w_0" shape = [512, 1, 3, 3] dtype = "float32" min_val = float("-0.301789") @@ -5483,6 +5984,7 @@ class Program_weight_tensor_parameter_500: class Program_weight_tensor_parameter_501: name = "parameter_501" + original_name = "linear_18.b_0" shape = [512] dtype = "float32" min_val = float("-3.10988") @@ -5494,6 +5996,7 @@ class Program_weight_tensor_parameter_501: class Program_weight_tensor_parameter_502: name = "parameter_502" + original_name = "linear_18.w_0" shape = [128, 512] dtype = "float32" min_val = float("-0.269608") @@ -5505,6 +6008,7 @@ class Program_weight_tensor_parameter_502: class Program_weight_tensor_parameter_503: name = "parameter_503" + original_name = "layer_norm_16.b_0" shape = [128] dtype = "float32" min_val = float("-1.1237") @@ -5516,6 +6020,7 @@ class Program_weight_tensor_parameter_503: class Program_weight_tensor_parameter_504: name = "parameter_504" + original_name = "layer_norm_16.w_0" shape = [128] dtype = "float32" min_val = float("1.37913") @@ -5527,6 +6032,7 @@ class Program_weight_tensor_parameter_504: class Program_weight_tensor_parameter_505: name = "parameter_505" + original_name = "linear_17.b_0" shape = [128] dtype = "float32" min_val = float("-0.223593") @@ -5538,6 +6044,7 @@ class Program_weight_tensor_parameter_505: class Program_weight_tensor_parameter_506: name = "parameter_506" + original_name = "linear_17.w_0" shape = [128, 128] dtype = "float32" min_val = float("-0.180025") @@ -5549,6 +6056,7 @@ class Program_weight_tensor_parameter_506: class Program_weight_tensor_parameter_507: name = "parameter_507" + original_name = "linear_16.b_0" shape = [256] dtype = "float32" min_val = float("-0.678846") @@ -5560,6 +6068,7 @@ class Program_weight_tensor_parameter_507: class Program_weight_tensor_parameter_508: name = "parameter_508" + original_name = "linear_16.w_0" shape = [128, 256] dtype = "float32" min_val = float("-0.355438") @@ -5571,6 +6080,7 @@ class Program_weight_tensor_parameter_508: class Program_weight_tensor_parameter_509: name = "parameter_509" + original_name = "layer_norm_15.b_0" shape = [128] dtype = "float32" min_val = float("-0.42557") @@ -5582,6 +6092,7 @@ class Program_weight_tensor_parameter_509: class Program_weight_tensor_parameter_510: name = "parameter_510" + original_name = "layer_norm_15.w_0" shape = [128] dtype = "float32" min_val = float("-0.0254958") @@ -5593,6 +6104,7 @@ class Program_weight_tensor_parameter_510: class Program_weight_tensor_parameter_511: name = "parameter_511" + original_name = "conv2d_10.b_0" shape = [128] dtype = "float32" min_val = float("-2.61834") @@ -5604,6 +6116,7 @@ class Program_weight_tensor_parameter_511: class Program_weight_tensor_parameter_512: name = "parameter_512" + original_name = "conv2d_10.w_0" shape = [128, 128, 4, 4] dtype = "float32" min_val = float("-0.275417") @@ -5615,6 +6128,7 @@ class Program_weight_tensor_parameter_512: class Program_weight_tensor_parameter_513: name = "parameter_513" + original_name = "linear_15.b_0" shape = [128] dtype = "float32" min_val = float("-2.66879") @@ -5626,6 +6140,7 @@ class Program_weight_tensor_parameter_513: class Program_weight_tensor_parameter_514: name = "parameter_514" + original_name = "linear_15.w_0" shape = [128, 128] dtype = "float32" min_val = float("-0.289356") @@ -5637,6 +6152,7 @@ class Program_weight_tensor_parameter_514: class Program_weight_tensor_parameter_515: name = "parameter_515" + original_name = "layer_norm_14.b_0" shape = [128] dtype = "float32" min_val = float("-0.601191") @@ -5648,6 +6164,7 @@ class Program_weight_tensor_parameter_515: class Program_weight_tensor_parameter_516: name = "parameter_516" + original_name = "layer_norm_14.w_0" shape = [128] dtype = "float32" min_val = float("0.152583") @@ -5659,6 +6176,7 @@ class Program_weight_tensor_parameter_516: class Program_weight_tensor_parameter_517: name = "parameter_517" + original_name = "layer_norm_1.b_0" shape = [128] dtype = "float32" min_val = float("-0.402948") @@ -5670,6 +6188,7 @@ class Program_weight_tensor_parameter_517: class Program_weight_tensor_parameter_518: name = "parameter_518" + original_name = "layer_norm_1.w_0" shape = [128] dtype = "float32" min_val = float("0.25929") @@ -5681,6 +6200,7 @@ class Program_weight_tensor_parameter_518: class Program_weight_tensor_parameter_519: name = "parameter_519" + original_name = "conv2d_1.b_0" shape = [128] dtype = "float32" min_val = float("-3.44804") @@ -5692,6 +6212,7 @@ class Program_weight_tensor_parameter_519: class Program_weight_tensor_parameter_520: name = "parameter_520" + original_name = "conv2d_1.w_0" shape = [128, 64, 3, 3] dtype = "float32" min_val = float("-0.339731") @@ -5703,6 +6224,7 @@ class Program_weight_tensor_parameter_520: class Program_weight_tensor_parameter_521: name = "parameter_521" + original_name = "layer_norm_13.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -5712,6 +6234,7 @@ class Program_weight_tensor_parameter_521: class Program_weight_tensor_parameter_522: name = "parameter_522" + original_name = "layer_norm_13.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -5721,6 +6244,7 @@ class Program_weight_tensor_parameter_522: class Program_weight_tensor_parameter_523: name = "parameter_523" + original_name = "linear_14.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -5730,6 +6254,7 @@ class Program_weight_tensor_parameter_523: class Program_weight_tensor_parameter_524: name = "parameter_524" + original_name = "linear_14.w_0" shape = [256, 64] dtype = "float32" min_val = float("-0.614452") @@ -5741,6 +6266,7 @@ class Program_weight_tensor_parameter_524: class Program_weight_tensor_parameter_525: name = "parameter_525" + original_name = "conv2d_9.b_0" shape = [256] dtype = "float32" min_val = float("-1.27933") @@ -5752,6 +6278,7 @@ class Program_weight_tensor_parameter_525: class Program_weight_tensor_parameter_526: name = "parameter_526" + original_name = "conv2d_9.w_0" shape = [256, 1, 3, 3] dtype = "float32" min_val = float("-0.408098") @@ -5763,6 +6290,7 @@ class Program_weight_tensor_parameter_526: class Program_weight_tensor_parameter_527: name = "parameter_527" + original_name = "linear_13.b_0" shape = [256] dtype = "float32" min_val = float("-2.51212") @@ -5774,6 +6302,7 @@ class Program_weight_tensor_parameter_527: class Program_weight_tensor_parameter_528: name = "parameter_528" + original_name = "linear_13.w_0" shape = [64, 256] dtype = "float32" min_val = float("-0.306974") @@ -5785,6 +6314,7 @@ class Program_weight_tensor_parameter_528: class Program_weight_tensor_parameter_529: name = "parameter_529" + original_name = "layer_norm_12.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -5794,6 +6324,7 @@ class Program_weight_tensor_parameter_529: class Program_weight_tensor_parameter_530: name = "parameter_530" + original_name = "layer_norm_12.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -5803,6 +6334,7 @@ class Program_weight_tensor_parameter_530: class Program_weight_tensor_parameter_531: name = "parameter_531" + original_name = "linear_12.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -5812,6 +6344,7 @@ class Program_weight_tensor_parameter_531: class Program_weight_tensor_parameter_532: name = "parameter_532" + original_name = "linear_12.w_0" shape = [64, 64] dtype = "float32" min_val = float("-0.283092") @@ -5823,6 +6356,7 @@ class Program_weight_tensor_parameter_532: class Program_weight_tensor_parameter_533: name = "parameter_533" + original_name = "linear_11.b_0" shape = [128] dtype = "float32" min_val = float("-1.33098") @@ -5834,6 +6368,7 @@ class Program_weight_tensor_parameter_533: class Program_weight_tensor_parameter_534: name = "parameter_534" + original_name = "linear_11.w_0" shape = [64, 128] dtype = "float32" min_val = float("-0.406126") @@ -5845,6 +6380,7 @@ class Program_weight_tensor_parameter_534: class Program_weight_tensor_parameter_535: name = "parameter_535" + original_name = "layer_norm_11.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -5854,6 +6390,7 @@ class Program_weight_tensor_parameter_535: class Program_weight_tensor_parameter_536: name = "parameter_536" + original_name = "layer_norm_11.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -5863,6 +6400,7 @@ class Program_weight_tensor_parameter_536: class Program_weight_tensor_parameter_537: name = "parameter_537" + original_name = "conv2d_8.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -5872,6 +6410,7 @@ class Program_weight_tensor_parameter_537: class Program_weight_tensor_parameter_538: name = "parameter_538" + original_name = "conv2d_8.w_0" shape = [64, 64, 8, 8] dtype = "float32" min_val = float("-0.40897") @@ -5883,6 +6422,7 @@ class Program_weight_tensor_parameter_538: class Program_weight_tensor_parameter_539: name = "parameter_539" + original_name = "linear_10.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -5892,6 +6432,7 @@ class Program_weight_tensor_parameter_539: class Program_weight_tensor_parameter_540: name = "parameter_540" + original_name = "linear_10.w_0" shape = [64, 64] dtype = "float32" min_val = float("-0.484371") @@ -5903,6 +6444,7 @@ class Program_weight_tensor_parameter_540: class Program_weight_tensor_parameter_541: name = "parameter_541" + original_name = "layer_norm_10.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -5912,6 +6454,7 @@ class Program_weight_tensor_parameter_541: class Program_weight_tensor_parameter_542: name = "parameter_542" + original_name = "layer_norm_10.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -5921,6 +6464,7 @@ class Program_weight_tensor_parameter_542: class Program_weight_tensor_parameter_543: name = "parameter_543" + original_name = "linear_9.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -5930,6 +6474,7 @@ class Program_weight_tensor_parameter_543: class Program_weight_tensor_parameter_544: name = "parameter_544" + original_name = "linear_9.w_0" shape = [256, 64] dtype = "float32" min_val = float("-0.337552") @@ -5941,6 +6486,7 @@ class Program_weight_tensor_parameter_544: class Program_weight_tensor_parameter_545: name = "parameter_545" + original_name = "conv2d_7.b_0" shape = [256] dtype = "float32" min_val = float("-1.51092") @@ -5952,6 +6498,7 @@ class Program_weight_tensor_parameter_545: class Program_weight_tensor_parameter_546: name = "parameter_546" + original_name = "conv2d_7.w_0" shape = [256, 1, 3, 3] dtype = "float32" min_val = float("-0.469962") @@ -5963,6 +6510,7 @@ class Program_weight_tensor_parameter_546: class Program_weight_tensor_parameter_547: name = "parameter_547" + original_name = "linear_8.b_0" shape = [256] dtype = "float32" min_val = float("-4.37523") @@ -5974,6 +6522,7 @@ class Program_weight_tensor_parameter_547: class Program_weight_tensor_parameter_548: name = "parameter_548" + original_name = "linear_8.w_0" shape = [64, 256] dtype = "float32" min_val = float("-0.387788") @@ -5985,6 +6534,7 @@ class Program_weight_tensor_parameter_548: class Program_weight_tensor_parameter_549: name = "parameter_549" + original_name = "layer_norm_9.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -5994,6 +6544,7 @@ class Program_weight_tensor_parameter_549: class Program_weight_tensor_parameter_550: name = "parameter_550" + original_name = "layer_norm_9.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -6003,6 +6554,7 @@ class Program_weight_tensor_parameter_550: class Program_weight_tensor_parameter_551: name = "parameter_551" + original_name = "linear_7.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -6012,6 +6564,7 @@ class Program_weight_tensor_parameter_551: class Program_weight_tensor_parameter_552: name = "parameter_552" + original_name = "linear_7.w_0" shape = [64, 64] dtype = "float32" min_val = float("-0.122983") @@ -6023,6 +6576,7 @@ class Program_weight_tensor_parameter_552: class Program_weight_tensor_parameter_553: name = "parameter_553" + original_name = "linear_6.b_0" shape = [128] dtype = "float32" min_val = float("-1.21128") @@ -6034,6 +6588,7 @@ class Program_weight_tensor_parameter_553: class Program_weight_tensor_parameter_554: name = "parameter_554" + original_name = "linear_6.w_0" shape = [64, 128] dtype = "float32" min_val = float("-0.474528") @@ -6045,6 +6600,7 @@ class Program_weight_tensor_parameter_554: class Program_weight_tensor_parameter_555: name = "parameter_555" + original_name = "layer_norm_8.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -6054,6 +6610,7 @@ class Program_weight_tensor_parameter_555: class Program_weight_tensor_parameter_556: name = "parameter_556" + original_name = "layer_norm_8.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -6063,6 +6620,7 @@ class Program_weight_tensor_parameter_556: class Program_weight_tensor_parameter_557: name = "parameter_557" + original_name = "conv2d_6.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -6072,6 +6630,7 @@ class Program_weight_tensor_parameter_557: class Program_weight_tensor_parameter_558: name = "parameter_558" + original_name = "conv2d_6.w_0" shape = [64, 64, 8, 8] dtype = "float32" min_val = float("-0.311942") @@ -6083,6 +6642,7 @@ class Program_weight_tensor_parameter_558: class Program_weight_tensor_parameter_559: name = "parameter_559" + original_name = "linear_5.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -6092,6 +6652,7 @@ class Program_weight_tensor_parameter_559: class Program_weight_tensor_parameter_560: name = "parameter_560" + original_name = "linear_5.w_0" shape = [64, 64] dtype = "float32" min_val = float("-0.405493") @@ -6103,6 +6664,7 @@ class Program_weight_tensor_parameter_560: class Program_weight_tensor_parameter_561: name = "parameter_561" + original_name = "layer_norm_7.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -6112,6 +6674,7 @@ class Program_weight_tensor_parameter_561: class Program_weight_tensor_parameter_562: name = "parameter_562" + original_name = "layer_norm_7.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -6121,6 +6684,7 @@ class Program_weight_tensor_parameter_562: class Program_weight_tensor_parameter_563: name = "parameter_563" + original_name = "linear_4.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -6130,6 +6694,7 @@ class Program_weight_tensor_parameter_563: class Program_weight_tensor_parameter_564: name = "parameter_564" + original_name = "linear_4.w_0" shape = [256, 64] dtype = "float32" min_val = float("-0.305501") @@ -6141,6 +6706,7 @@ class Program_weight_tensor_parameter_564: class Program_weight_tensor_parameter_565: name = "parameter_565" + original_name = "conv2d_5.b_0" shape = [256] dtype = "float32" min_val = float("-1.84049") @@ -6152,6 +6718,7 @@ class Program_weight_tensor_parameter_565: class Program_weight_tensor_parameter_566: name = "parameter_566" + original_name = "conv2d_5.w_0" shape = [256, 1, 3, 3] dtype = "float32" min_val = float("-0.598833") @@ -6163,6 +6730,7 @@ class Program_weight_tensor_parameter_566: class Program_weight_tensor_parameter_567: name = "parameter_567" + original_name = "linear_3.b_0" shape = [256] dtype = "float32" min_val = float("-6.61604") @@ -6174,6 +6742,7 @@ class Program_weight_tensor_parameter_567: class Program_weight_tensor_parameter_568: name = "parameter_568" + original_name = "linear_3.w_0" shape = [64, 256] dtype = "float32" min_val = float("-0.606294") @@ -6185,6 +6754,7 @@ class Program_weight_tensor_parameter_568: class Program_weight_tensor_parameter_569: name = "parameter_569" + original_name = "layer_norm_6.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -6194,6 +6764,7 @@ class Program_weight_tensor_parameter_569: class Program_weight_tensor_parameter_570: name = "parameter_570" + original_name = "layer_norm_6.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -6203,6 +6774,7 @@ class Program_weight_tensor_parameter_570: class Program_weight_tensor_parameter_571: name = "parameter_571" + original_name = "linear_2.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -6212,6 +6784,7 @@ class Program_weight_tensor_parameter_571: class Program_weight_tensor_parameter_572: name = "parameter_572" + original_name = "linear_2.w_0" shape = [64, 64] dtype = "float32" min_val = float("-0.198358") @@ -6223,6 +6796,7 @@ class Program_weight_tensor_parameter_572: class Program_weight_tensor_parameter_573: name = "parameter_573" + original_name = "linear_1.b_0" shape = [128] dtype = "float32" min_val = float("-1.84329") @@ -6234,6 +6808,7 @@ class Program_weight_tensor_parameter_573: class Program_weight_tensor_parameter_574: name = "parameter_574" + original_name = "linear_1.w_0" shape = [64, 128] dtype = "float32" min_val = float("-0.496123") @@ -6245,6 +6820,7 @@ class Program_weight_tensor_parameter_574: class Program_weight_tensor_parameter_575: name = "parameter_575" + original_name = "layer_norm_5.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -6254,6 +6830,7 @@ class Program_weight_tensor_parameter_575: class Program_weight_tensor_parameter_576: name = "parameter_576" + original_name = "layer_norm_5.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -6263,6 +6840,7 @@ class Program_weight_tensor_parameter_576: class Program_weight_tensor_parameter_577: name = "parameter_577" + original_name = "conv2d_4.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -6272,6 +6850,7 @@ class Program_weight_tensor_parameter_577: class Program_weight_tensor_parameter_578: name = "parameter_578" + original_name = "conv2d_4.w_0" shape = [64, 64, 8, 8] dtype = "float32" min_val = float("-0.455881") @@ -6283,6 +6862,7 @@ class Program_weight_tensor_parameter_578: class Program_weight_tensor_parameter_579: name = "parameter_579" + original_name = "linear_0.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -6292,6 +6872,7 @@ class Program_weight_tensor_parameter_579: class Program_weight_tensor_parameter_580: name = "parameter_580" + original_name = "linear_0.w_0" shape = [64, 64] dtype = "float32" min_val = float("-0.46165") @@ -6303,6 +6884,7 @@ class Program_weight_tensor_parameter_580: class Program_weight_tensor_parameter_581: name = "parameter_581" + original_name = "layer_norm_4.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -6312,6 +6894,7 @@ class Program_weight_tensor_parameter_581: class Program_weight_tensor_parameter_582: name = "parameter_582" + original_name = "layer_norm_4.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -6321,6 +6904,7 @@ class Program_weight_tensor_parameter_582: class Program_weight_tensor_parameter_583: name = "parameter_583" + original_name = "layer_norm_0.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -6330,6 +6914,7 @@ class Program_weight_tensor_parameter_583: class Program_weight_tensor_parameter_584: name = "parameter_584" + original_name = "layer_norm_0.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -6339,6 +6924,7 @@ class Program_weight_tensor_parameter_584: class Program_weight_tensor_parameter_585: name = "parameter_585" + original_name = "conv2d_0.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -6348,6 +6934,7 @@ class Program_weight_tensor_parameter_585: class Program_weight_tensor_parameter_586: name = "parameter_586" + original_name = "conv2d_0.w_0" shape = [64, 3, 7, 7] dtype = "float32" min_val = float("-0.394077") diff --git a/paddle_samples/PaddleX/SegFormer-B4/subgraph_0/input_meta.py b/paddle_samples/PaddleX/SegFormer-B4/subgraph_0/input_meta.py index 3bc847dc8..0ab31ae0e 100644 --- a/paddle_samples/PaddleX/SegFormer-B4/subgraph_0/input_meta.py +++ b/paddle_samples/PaddleX/SegFormer-B4/subgraph_0/input_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_data_0: name = "data_0" + original_name = "var_0" shape = [2, 3, 512, 1024] dtype = "float32" min_val = float("-1.0") diff --git a/paddle_samples/PaddleX/SegFormer-B4/subgraph_0/weight_meta.py b/paddle_samples/PaddleX/SegFormer-B4/subgraph_0/weight_meta.py index 3d29e5265..6648e4362 100644 --- a/paddle_samples/PaddleX/SegFormer-B4/subgraph_0/weight_meta.py +++ b/paddle_samples/PaddleX/SegFormer-B4/subgraph_0/weight_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_parameter_0: name = "parameter_0" + original_name = "conv2d_84.b_0" shape = [2] dtype = "float32" min_val = float("0") @@ -9,6 +10,7 @@ class Program_weight_tensor_parameter_0: class Program_weight_tensor_parameter_1: name = "parameter_1" + original_name = "conv2d_84.w_0" shape = [2, 768, 1, 1] dtype = "float32" min_val = float("-0.165987") @@ -20,6 +22,7 @@ class Program_weight_tensor_parameter_1: class Program_weight_tensor_parameter_2: name = "parameter_2" + original_name = "batch_norm2d_0.b_0" shape = [768] dtype = "float32" data = None @@ -27,6 +30,7 @@ class Program_weight_tensor_parameter_2: class Program_weight_tensor_parameter_3: name = "parameter_3" + original_name = "batch_norm2d_0.w_0" shape = [768] dtype = "float32" min_val = float("1.0") @@ -37,6 +41,7 @@ class Program_weight_tensor_parameter_3: class Program_weight_tensor_parameter_4: name = "parameter_4" + original_name = "batch_norm2d_0.w_2" shape = [768] dtype = "float32" min_val = float("1.0") @@ -47,6 +52,7 @@ class Program_weight_tensor_parameter_4: class Program_weight_tensor_parameter_5: name = "parameter_5" + original_name = "batch_norm2d_0.w_1" shape = [768] dtype = "float32" data = None @@ -54,6 +60,7 @@ class Program_weight_tensor_parameter_5: class Program_weight_tensor_parameter_6: name = "parameter_6" + original_name = "conv2d_83.w_0" shape = [768, 3072, 1, 1] dtype = "float32" min_val = float("-0.135115") @@ -65,6 +72,7 @@ class Program_weight_tensor_parameter_6: class Program_weight_tensor_parameter_7: name = "parameter_7" + original_name = "linear_208.b_0" shape = [768] dtype = "float32" data = None @@ -72,6 +80,7 @@ class Program_weight_tensor_parameter_7: class Program_weight_tensor_parameter_8: name = "parameter_8" + original_name = "linear_208.w_0" shape = [64, 768] dtype = "float32" min_val = float("-0.0849193") @@ -83,6 +92,7 @@ class Program_weight_tensor_parameter_8: class Program_weight_tensor_parameter_9: name = "parameter_9" + original_name = "linear_207.b_0" shape = [768] dtype = "float32" data = None @@ -90,6 +100,7 @@ class Program_weight_tensor_parameter_9: class Program_weight_tensor_parameter_10: name = "parameter_10" + original_name = "linear_207.w_0" shape = [128, 768] dtype = "float32" min_val = float("-0.0818302") @@ -101,6 +112,7 @@ class Program_weight_tensor_parameter_10: class Program_weight_tensor_parameter_11: name = "parameter_11" + original_name = "linear_206.b_0" shape = [768] dtype = "float32" data = None @@ -108,6 +120,7 @@ class Program_weight_tensor_parameter_11: class Program_weight_tensor_parameter_12: name = "parameter_12" + original_name = "linear_206.w_0" shape = [320, 768] dtype = "float32" min_val = float("-0.0742597") @@ -119,6 +132,7 @@ class Program_weight_tensor_parameter_12: class Program_weight_tensor_parameter_13: name = "parameter_13" + original_name = "linear_205.b_0" shape = [768] dtype = "float32" data = None @@ -126,6 +140,7 @@ class Program_weight_tensor_parameter_13: class Program_weight_tensor_parameter_14: name = "parameter_14" + original_name = "linear_205.w_0" shape = [512, 768] dtype = "float32" min_val = float("-0.0684648") @@ -137,6 +152,7 @@ class Program_weight_tensor_parameter_14: class Program_weight_tensor_parameter_15: name = "parameter_15" + original_name = "layer_norm_127.b_0" shape = [512] dtype = "float32" min_val = float("-1.02259") @@ -148,6 +164,7 @@ class Program_weight_tensor_parameter_15: class Program_weight_tensor_parameter_16: name = "parameter_16" + original_name = "layer_norm_127.w_0" shape = [512] dtype = "float32" min_val = float("0.000871397") @@ -159,6 +176,7 @@ class Program_weight_tensor_parameter_16: class Program_weight_tensor_parameter_17: name = "parameter_17" + original_name = "linear_204.b_0" shape = [512] dtype = "float32" min_val = float("-5.85417") @@ -170,6 +188,7 @@ class Program_weight_tensor_parameter_17: class Program_weight_tensor_parameter_18: name = "parameter_18" + original_name = "linear_204.w_0" shape = [2048, 512] dtype = "float32" min_val = float("-2.20833") @@ -181,6 +200,7 @@ class Program_weight_tensor_parameter_18: class Program_weight_tensor_parameter_19: name = "parameter_19" + original_name = "conv2d_82.b_0" shape = [2048] dtype = "float32" min_val = float("-1.52763") @@ -192,6 +212,7 @@ class Program_weight_tensor_parameter_19: class Program_weight_tensor_parameter_20: name = "parameter_20" + original_name = "conv2d_82.w_0" shape = [2048, 1, 3, 3] dtype = "float32" min_val = float("-0.678162") @@ -203,6 +224,7 @@ class Program_weight_tensor_parameter_20: class Program_weight_tensor_parameter_21: name = "parameter_21" + original_name = "linear_203.b_0" shape = [2048] dtype = "float32" min_val = float("-1.6605") @@ -214,6 +236,7 @@ class Program_weight_tensor_parameter_21: class Program_weight_tensor_parameter_22: name = "parameter_22" + original_name = "linear_203.w_0" shape = [512, 2048] dtype = "float32" min_val = float("-0.648029") @@ -225,6 +248,7 @@ class Program_weight_tensor_parameter_22: class Program_weight_tensor_parameter_23: name = "parameter_23" + original_name = "layer_norm_126.b_0" shape = [512] dtype = "float32" min_val = float("-1.27822") @@ -236,6 +260,7 @@ class Program_weight_tensor_parameter_23: class Program_weight_tensor_parameter_24: name = "parameter_24" + original_name = "layer_norm_126.w_0" shape = [512] dtype = "float32" min_val = float("0.211879") @@ -247,6 +272,7 @@ class Program_weight_tensor_parameter_24: class Program_weight_tensor_parameter_25: name = "parameter_25" + original_name = "linear_202.b_0" shape = [512] dtype = "float32" min_val = float("-5.85524") @@ -258,6 +284,7 @@ class Program_weight_tensor_parameter_25: class Program_weight_tensor_parameter_26: name = "parameter_26" + original_name = "linear_202.w_0" shape = [512, 512] dtype = "float32" min_val = float("-0.915749") @@ -269,6 +296,7 @@ class Program_weight_tensor_parameter_26: class Program_weight_tensor_parameter_27: name = "parameter_27" + original_name = "linear_201.b_0" shape = [1024] dtype = "float32" min_val = float("-1.94081") @@ -280,6 +308,7 @@ class Program_weight_tensor_parameter_27: class Program_weight_tensor_parameter_28: name = "parameter_28" + original_name = "linear_201.w_0" shape = [512, 1024] dtype = "float32" min_val = float("-0.299746") @@ -291,6 +320,7 @@ class Program_weight_tensor_parameter_28: class Program_weight_tensor_parameter_29: name = "parameter_29" + original_name = "linear_200.b_0" shape = [512] dtype = "float32" min_val = float("-2.34658") @@ -302,6 +332,7 @@ class Program_weight_tensor_parameter_29: class Program_weight_tensor_parameter_30: name = "parameter_30" + original_name = "linear_200.w_0" shape = [512, 512] dtype = "float32" min_val = float("-0.308371") @@ -313,6 +344,7 @@ class Program_weight_tensor_parameter_30: class Program_weight_tensor_parameter_31: name = "parameter_31" + original_name = "layer_norm_125.b_0" shape = [512] dtype = "float32" min_val = float("-1.52776") @@ -324,6 +356,7 @@ class Program_weight_tensor_parameter_31: class Program_weight_tensor_parameter_32: name = "parameter_32" + original_name = "layer_norm_125.w_0" shape = [512] dtype = "float32" min_val = float("0.0495567") @@ -335,6 +368,7 @@ class Program_weight_tensor_parameter_32: class Program_weight_tensor_parameter_33: name = "parameter_33" + original_name = "linear_199.b_0" shape = [512] dtype = "float32" min_val = float("-7.93791") @@ -346,6 +380,7 @@ class Program_weight_tensor_parameter_33: class Program_weight_tensor_parameter_34: name = "parameter_34" + original_name = "linear_199.w_0" shape = [2048, 512] dtype = "float32" min_val = float("-1.95751") @@ -357,6 +392,7 @@ class Program_weight_tensor_parameter_34: class Program_weight_tensor_parameter_35: name = "parameter_35" + original_name = "conv2d_81.b_0" shape = [2048] dtype = "float32" min_val = float("-1.02865") @@ -368,6 +404,7 @@ class Program_weight_tensor_parameter_35: class Program_weight_tensor_parameter_36: name = "parameter_36" + original_name = "conv2d_81.w_0" shape = [2048, 1, 3, 3] dtype = "float32" min_val = float("-0.819086") @@ -379,6 +416,7 @@ class Program_weight_tensor_parameter_36: class Program_weight_tensor_parameter_37: name = "parameter_37" + original_name = "linear_198.b_0" shape = [2048] dtype = "float32" min_val = float("-2.04537") @@ -390,6 +428,7 @@ class Program_weight_tensor_parameter_37: class Program_weight_tensor_parameter_38: name = "parameter_38" + original_name = "linear_198.w_0" shape = [512, 2048] dtype = "float32" min_val = float("-0.849893") @@ -401,6 +440,7 @@ class Program_weight_tensor_parameter_38: class Program_weight_tensor_parameter_39: name = "parameter_39" + original_name = "layer_norm_124.b_0" shape = [512] dtype = "float32" min_val = float("-2.00448") @@ -412,6 +452,7 @@ class Program_weight_tensor_parameter_39: class Program_weight_tensor_parameter_40: name = "parameter_40" + original_name = "layer_norm_124.w_0" shape = [512] dtype = "float32" min_val = float("0.285106") @@ -423,6 +464,7 @@ class Program_weight_tensor_parameter_40: class Program_weight_tensor_parameter_41: name = "parameter_41" + original_name = "linear_197.b_0" shape = [512] dtype = "float32" min_val = float("-5.5123") @@ -434,6 +476,7 @@ class Program_weight_tensor_parameter_41: class Program_weight_tensor_parameter_42: name = "parameter_42" + original_name = "linear_197.w_0" shape = [512, 512] dtype = "float32" min_val = float("-0.586563") @@ -445,6 +488,7 @@ class Program_weight_tensor_parameter_42: class Program_weight_tensor_parameter_43: name = "parameter_43" + original_name = "linear_196.b_0" shape = [1024] dtype = "float32" min_val = float("-3.25562") @@ -456,6 +500,7 @@ class Program_weight_tensor_parameter_43: class Program_weight_tensor_parameter_44: name = "parameter_44" + original_name = "linear_196.w_0" shape = [512, 1024] dtype = "float32" min_val = float("-0.505312") @@ -467,6 +512,7 @@ class Program_weight_tensor_parameter_44: class Program_weight_tensor_parameter_45: name = "parameter_45" + original_name = "linear_195.b_0" shape = [512] dtype = "float32" min_val = float("-1.97175") @@ -478,6 +524,7 @@ class Program_weight_tensor_parameter_45: class Program_weight_tensor_parameter_46: name = "parameter_46" + original_name = "linear_195.w_0" shape = [512, 512] dtype = "float32" min_val = float("-0.311821") @@ -489,6 +536,7 @@ class Program_weight_tensor_parameter_46: class Program_weight_tensor_parameter_47: name = "parameter_47" + original_name = "layer_norm_123.b_0" shape = [512] dtype = "float32" min_val = float("-1.70026") @@ -500,6 +548,7 @@ class Program_weight_tensor_parameter_47: class Program_weight_tensor_parameter_48: name = "parameter_48" + original_name = "layer_norm_123.w_0" shape = [512] dtype = "float32" min_val = float("0.00612336") @@ -511,6 +560,7 @@ class Program_weight_tensor_parameter_48: class Program_weight_tensor_parameter_49: name = "parameter_49" + original_name = "linear_194.b_0" shape = [512] dtype = "float32" min_val = float("-4.72022") @@ -522,6 +572,7 @@ class Program_weight_tensor_parameter_49: class Program_weight_tensor_parameter_50: name = "parameter_50" + original_name = "linear_194.w_0" shape = [2048, 512] dtype = "float32" min_val = float("-1.84216") @@ -533,6 +584,7 @@ class Program_weight_tensor_parameter_50: class Program_weight_tensor_parameter_51: name = "parameter_51" + original_name = "conv2d_80.b_0" shape = [2048] dtype = "float32" min_val = float("-1.01104") @@ -544,6 +596,7 @@ class Program_weight_tensor_parameter_51: class Program_weight_tensor_parameter_52: name = "parameter_52" + original_name = "conv2d_80.w_0" shape = [2048, 1, 3, 3] dtype = "float32" min_val = float("-0.549333") @@ -555,6 +608,7 @@ class Program_weight_tensor_parameter_52: class Program_weight_tensor_parameter_53: name = "parameter_53" + original_name = "linear_193.b_0" shape = [2048] dtype = "float32" min_val = float("-4.31449") @@ -566,6 +620,7 @@ class Program_weight_tensor_parameter_53: class Program_weight_tensor_parameter_54: name = "parameter_54" + original_name = "linear_193.w_0" shape = [512, 2048] dtype = "float32" min_val = float("-0.5945") @@ -577,6 +632,7 @@ class Program_weight_tensor_parameter_54: class Program_weight_tensor_parameter_55: name = "parameter_55" + original_name = "layer_norm_122.b_0" shape = [512] dtype = "float32" min_val = float("-1.27165") @@ -588,6 +644,7 @@ class Program_weight_tensor_parameter_55: class Program_weight_tensor_parameter_56: name = "parameter_56" + original_name = "layer_norm_122.w_0" shape = [512] dtype = "float32" min_val = float("0.437638") @@ -599,6 +656,7 @@ class Program_weight_tensor_parameter_56: class Program_weight_tensor_parameter_57: name = "parameter_57" + original_name = "linear_192.b_0" shape = [512] dtype = "float32" min_val = float("-1.67002") @@ -610,6 +668,7 @@ class Program_weight_tensor_parameter_57: class Program_weight_tensor_parameter_58: name = "parameter_58" + original_name = "linear_192.w_0" shape = [512, 512] dtype = "float32" min_val = float("-0.403791") @@ -621,6 +680,7 @@ class Program_weight_tensor_parameter_58: class Program_weight_tensor_parameter_59: name = "parameter_59" + original_name = "linear_191.b_0" shape = [1024] dtype = "float32" min_val = float("-1.19302") @@ -632,6 +692,7 @@ class Program_weight_tensor_parameter_59: class Program_weight_tensor_parameter_60: name = "parameter_60" + original_name = "linear_191.w_0" shape = [512, 1024] dtype = "float32" min_val = float("-0.343409") @@ -643,6 +704,7 @@ class Program_weight_tensor_parameter_60: class Program_weight_tensor_parameter_61: name = "parameter_61" + original_name = "linear_190.b_0" shape = [512] dtype = "float32" min_val = float("-2.06353") @@ -654,6 +716,7 @@ class Program_weight_tensor_parameter_61: class Program_weight_tensor_parameter_62: name = "parameter_62" + original_name = "linear_190.w_0" shape = [512, 512] dtype = "float32" min_val = float("-0.280705") @@ -665,6 +728,7 @@ class Program_weight_tensor_parameter_62: class Program_weight_tensor_parameter_63: name = "parameter_63" + original_name = "layer_norm_121.b_0" shape = [512] dtype = "float32" min_val = float("-1.29877") @@ -676,6 +740,7 @@ class Program_weight_tensor_parameter_63: class Program_weight_tensor_parameter_64: name = "parameter_64" + original_name = "layer_norm_121.w_0" shape = [512] dtype = "float32" min_val = float("0.000804451") @@ -687,6 +752,7 @@ class Program_weight_tensor_parameter_64: class Program_weight_tensor_parameter_65: name = "parameter_65" + original_name = "layer_norm_3.b_0" shape = [512] dtype = "float32" min_val = float("-5.98008") @@ -698,6 +764,7 @@ class Program_weight_tensor_parameter_65: class Program_weight_tensor_parameter_66: name = "parameter_66" + original_name = "layer_norm_3.w_0" shape = [512] dtype = "float32" min_val = float("0.0686089") @@ -709,6 +776,7 @@ class Program_weight_tensor_parameter_66: class Program_weight_tensor_parameter_67: name = "parameter_67" + original_name = "conv2d_3.b_0" shape = [512] dtype = "float32" min_val = float("-6.35785") @@ -720,6 +788,7 @@ class Program_weight_tensor_parameter_67: class Program_weight_tensor_parameter_68: name = "parameter_68" + original_name = "conv2d_3.w_0" shape = [512, 320, 3, 3] dtype = "float32" min_val = float("-2.24026") @@ -731,6 +800,7 @@ class Program_weight_tensor_parameter_68: class Program_weight_tensor_parameter_69: name = "parameter_69" + original_name = "layer_norm_120.b_0" shape = [320] dtype = "float32" min_val = float("-1.35412") @@ -742,6 +812,7 @@ class Program_weight_tensor_parameter_69: class Program_weight_tensor_parameter_70: name = "parameter_70" + original_name = "layer_norm_120.w_0" shape = [320] dtype = "float32" min_val = float("0.120345") @@ -753,6 +824,7 @@ class Program_weight_tensor_parameter_70: class Program_weight_tensor_parameter_71: name = "parameter_71" + original_name = "linear_189.b_0" shape = [320] dtype = "float32" min_val = float("-0.386169") @@ -764,6 +836,7 @@ class Program_weight_tensor_parameter_71: class Program_weight_tensor_parameter_72: name = "parameter_72" + original_name = "linear_189.w_0" shape = [1280, 320] dtype = "float32" min_val = float("-1.40357") @@ -775,6 +848,7 @@ class Program_weight_tensor_parameter_72: class Program_weight_tensor_parameter_73: name = "parameter_73" + original_name = "conv2d_79.b_0" shape = [1280] dtype = "float32" min_val = float("-2.59475") @@ -786,6 +860,7 @@ class Program_weight_tensor_parameter_73: class Program_weight_tensor_parameter_74: name = "parameter_74" + original_name = "conv2d_79.w_0" shape = [1280, 1, 3, 3] dtype = "float32" min_val = float("-1.40982") @@ -797,6 +872,7 @@ class Program_weight_tensor_parameter_74: class Program_weight_tensor_parameter_75: name = "parameter_75" + original_name = "linear_188.b_0" shape = [1280] dtype = "float32" min_val = float("-1.479") @@ -808,6 +884,7 @@ class Program_weight_tensor_parameter_75: class Program_weight_tensor_parameter_76: name = "parameter_76" + original_name = "linear_188.w_0" shape = [320, 1280] dtype = "float32" min_val = float("-0.573769") @@ -819,6 +896,7 @@ class Program_weight_tensor_parameter_76: class Program_weight_tensor_parameter_77: name = "parameter_77" + original_name = "layer_norm_119.b_0" shape = [320] dtype = "float32" min_val = float("-1.47961") @@ -830,6 +908,7 @@ class Program_weight_tensor_parameter_77: class Program_weight_tensor_parameter_78: name = "parameter_78" + original_name = "layer_norm_119.w_0" shape = [320] dtype = "float32" min_val = float("0.138292") @@ -841,6 +920,7 @@ class Program_weight_tensor_parameter_78: class Program_weight_tensor_parameter_79: name = "parameter_79" + original_name = "linear_187.b_0" shape = [320] dtype = "float32" min_val = float("-0.393891") @@ -852,6 +932,7 @@ class Program_weight_tensor_parameter_79: class Program_weight_tensor_parameter_80: name = "parameter_80" + original_name = "linear_187.w_0" shape = [320, 320] dtype = "float32" min_val = float("-1.56968") @@ -863,6 +944,7 @@ class Program_weight_tensor_parameter_80: class Program_weight_tensor_parameter_81: name = "parameter_81" + original_name = "linear_186.b_0" shape = [640] dtype = "float32" min_val = float("-15.8983") @@ -874,6 +956,7 @@ class Program_weight_tensor_parameter_81: class Program_weight_tensor_parameter_82: name = "parameter_82" + original_name = "linear_186.w_0" shape = [320, 640] dtype = "float32" min_val = float("-0.401033") @@ -885,6 +968,7 @@ class Program_weight_tensor_parameter_82: class Program_weight_tensor_parameter_83: name = "parameter_83" + original_name = "layer_norm_118.b_0" shape = [320] dtype = "float32" min_val = float("-0.476741") @@ -896,6 +980,7 @@ class Program_weight_tensor_parameter_83: class Program_weight_tensor_parameter_84: name = "parameter_84" + original_name = "layer_norm_118.w_0" shape = [320] dtype = "float32" min_val = float("-0.0145891") @@ -907,6 +992,7 @@ class Program_weight_tensor_parameter_84: class Program_weight_tensor_parameter_85: name = "parameter_85" + original_name = "conv2d_78.b_0" shape = [320] dtype = "float32" min_val = float("-2.7418") @@ -918,6 +1004,7 @@ class Program_weight_tensor_parameter_85: class Program_weight_tensor_parameter_86: name = "parameter_86" + original_name = "conv2d_78.w_0" shape = [320, 320, 2, 2] dtype = "float32" min_val = float("-0.325267") @@ -929,6 +1016,7 @@ class Program_weight_tensor_parameter_86: class Program_weight_tensor_parameter_87: name = "parameter_87" + original_name = "linear_185.b_0" shape = [320] dtype = "float32" min_val = float("-2.04761") @@ -940,6 +1028,7 @@ class Program_weight_tensor_parameter_87: class Program_weight_tensor_parameter_88: name = "parameter_88" + original_name = "linear_185.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.36763") @@ -951,6 +1040,7 @@ class Program_weight_tensor_parameter_88: class Program_weight_tensor_parameter_89: name = "parameter_89" + original_name = "layer_norm_117.b_0" shape = [320] dtype = "float32" min_val = float("-1.7927") @@ -962,6 +1052,7 @@ class Program_weight_tensor_parameter_89: class Program_weight_tensor_parameter_90: name = "parameter_90" + original_name = "layer_norm_117.w_0" shape = [320] dtype = "float32" min_val = float("0.2127") @@ -973,6 +1064,7 @@ class Program_weight_tensor_parameter_90: class Program_weight_tensor_parameter_91: name = "parameter_91" + original_name = "linear_184.b_0" shape = [320] dtype = "float32" min_val = float("-0.357214") @@ -984,6 +1076,7 @@ class Program_weight_tensor_parameter_91: class Program_weight_tensor_parameter_92: name = "parameter_92" + original_name = "linear_184.w_0" shape = [1280, 320] dtype = "float32" min_val = float("-1.03298") @@ -995,6 +1088,7 @@ class Program_weight_tensor_parameter_92: class Program_weight_tensor_parameter_93: name = "parameter_93" + original_name = "conv2d_77.b_0" shape = [1280] dtype = "float32" min_val = float("-3.23837") @@ -1006,6 +1100,7 @@ class Program_weight_tensor_parameter_93: class Program_weight_tensor_parameter_94: name = "parameter_94" + original_name = "conv2d_77.w_0" shape = [1280, 1, 3, 3] dtype = "float32" min_val = float("-0.953388") @@ -1017,6 +1112,7 @@ class Program_weight_tensor_parameter_94: class Program_weight_tensor_parameter_95: name = "parameter_95" + original_name = "linear_183.b_0" shape = [1280] dtype = "float32" min_val = float("-1.73923") @@ -1028,6 +1124,7 @@ class Program_weight_tensor_parameter_95: class Program_weight_tensor_parameter_96: name = "parameter_96" + original_name = "linear_183.w_0" shape = [320, 1280] dtype = "float32" min_val = float("-0.622307") @@ -1039,6 +1136,7 @@ class Program_weight_tensor_parameter_96: class Program_weight_tensor_parameter_97: name = "parameter_97" + original_name = "layer_norm_116.b_0" shape = [320] dtype = "float32" min_val = float("-0.927853") @@ -1050,6 +1148,7 @@ class Program_weight_tensor_parameter_97: class Program_weight_tensor_parameter_98: name = "parameter_98" + original_name = "layer_norm_116.w_0" shape = [320] dtype = "float32" min_val = float("0.0847431") @@ -1061,6 +1160,7 @@ class Program_weight_tensor_parameter_98: class Program_weight_tensor_parameter_99: name = "parameter_99" + original_name = "linear_182.b_0" shape = [320] dtype = "float32" min_val = float("-0.211183") @@ -1072,6 +1172,7 @@ class Program_weight_tensor_parameter_99: class Program_weight_tensor_parameter_100: name = "parameter_100" + original_name = "linear_182.w_0" shape = [320, 320] dtype = "float32" min_val = float("-1.20802") @@ -1083,6 +1184,7 @@ class Program_weight_tensor_parameter_100: class Program_weight_tensor_parameter_101: name = "parameter_101" + original_name = "linear_181.b_0" shape = [640] dtype = "float32" min_val = float("-1.07635") @@ -1094,6 +1196,7 @@ class Program_weight_tensor_parameter_101: class Program_weight_tensor_parameter_102: name = "parameter_102" + original_name = "linear_181.w_0" shape = [320, 640] dtype = "float32" min_val = float("-0.374248") @@ -1105,6 +1208,7 @@ class Program_weight_tensor_parameter_102: class Program_weight_tensor_parameter_103: name = "parameter_103" + original_name = "layer_norm_115.b_0" shape = [320] dtype = "float32" min_val = float("-0.441734") @@ -1116,6 +1220,7 @@ class Program_weight_tensor_parameter_103: class Program_weight_tensor_parameter_104: name = "parameter_104" + original_name = "layer_norm_115.w_0" shape = [320] dtype = "float32" min_val = float("-0.163302") @@ -1127,6 +1232,7 @@ class Program_weight_tensor_parameter_104: class Program_weight_tensor_parameter_105: name = "parameter_105" + original_name = "conv2d_76.b_0" shape = [320] dtype = "float32" min_val = float("-2.88566") @@ -1138,6 +1244,7 @@ class Program_weight_tensor_parameter_105: class Program_weight_tensor_parameter_106: name = "parameter_106" + original_name = "conv2d_76.w_0" shape = [320, 320, 2, 2] dtype = "float32" min_val = float("-0.34308") @@ -1149,6 +1256,7 @@ class Program_weight_tensor_parameter_106: class Program_weight_tensor_parameter_107: name = "parameter_107" + original_name = "linear_180.b_0" shape = [320] dtype = "float32" min_val = float("-2.09821") @@ -1160,6 +1268,7 @@ class Program_weight_tensor_parameter_107: class Program_weight_tensor_parameter_108: name = "parameter_108" + original_name = "linear_180.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.292369") @@ -1171,6 +1280,7 @@ class Program_weight_tensor_parameter_108: class Program_weight_tensor_parameter_109: name = "parameter_109" + original_name = "layer_norm_114.b_0" shape = [320] dtype = "float32" min_val = float("-1.41445") @@ -1182,6 +1292,7 @@ class Program_weight_tensor_parameter_109: class Program_weight_tensor_parameter_110: name = "parameter_110" + original_name = "layer_norm_114.w_0" shape = [320] dtype = "float32" min_val = float("0.18733") @@ -1193,6 +1304,7 @@ class Program_weight_tensor_parameter_110: class Program_weight_tensor_parameter_111: name = "parameter_111" + original_name = "linear_179.b_0" shape = [320] dtype = "float32" min_val = float("-0.332993") @@ -1204,6 +1316,7 @@ class Program_weight_tensor_parameter_111: class Program_weight_tensor_parameter_112: name = "parameter_112" + original_name = "linear_179.w_0" shape = [1280, 320] dtype = "float32" min_val = float("-0.923492") @@ -1215,6 +1328,7 @@ class Program_weight_tensor_parameter_112: class Program_weight_tensor_parameter_113: name = "parameter_113" + original_name = "conv2d_75.b_0" shape = [1280] dtype = "float32" min_val = float("-2.88806") @@ -1226,6 +1340,7 @@ class Program_weight_tensor_parameter_113: class Program_weight_tensor_parameter_114: name = "parameter_114" + original_name = "conv2d_75.w_0" shape = [1280, 1, 3, 3] dtype = "float32" min_val = float("-0.831423") @@ -1237,6 +1352,7 @@ class Program_weight_tensor_parameter_114: class Program_weight_tensor_parameter_115: name = "parameter_115" + original_name = "linear_178.b_0" shape = [1280] dtype = "float32" min_val = float("-2.47642") @@ -1248,6 +1364,7 @@ class Program_weight_tensor_parameter_115: class Program_weight_tensor_parameter_116: name = "parameter_116" + original_name = "linear_178.w_0" shape = [320, 1280] dtype = "float32" min_val = float("-0.387012") @@ -1259,6 +1376,7 @@ class Program_weight_tensor_parameter_116: class Program_weight_tensor_parameter_117: name = "parameter_117" + original_name = "layer_norm_113.b_0" shape = [320] dtype = "float32" min_val = float("-0.922313") @@ -1270,6 +1388,7 @@ class Program_weight_tensor_parameter_117: class Program_weight_tensor_parameter_118: name = "parameter_118" + original_name = "layer_norm_113.w_0" shape = [320] dtype = "float32" min_val = float("0.0904113") @@ -1281,6 +1400,7 @@ class Program_weight_tensor_parameter_118: class Program_weight_tensor_parameter_119: name = "parameter_119" + original_name = "linear_177.b_0" shape = [320] dtype = "float32" min_val = float("-0.273089") @@ -1292,6 +1412,7 @@ class Program_weight_tensor_parameter_119: class Program_weight_tensor_parameter_120: name = "parameter_120" + original_name = "linear_177.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.8052") @@ -1303,6 +1424,7 @@ class Program_weight_tensor_parameter_120: class Program_weight_tensor_parameter_121: name = "parameter_121" + original_name = "linear_176.b_0" shape = [640] dtype = "float32" min_val = float("-9.58141") @@ -1314,6 +1436,7 @@ class Program_weight_tensor_parameter_121: class Program_weight_tensor_parameter_122: name = "parameter_122" + original_name = "linear_176.w_0" shape = [320, 640] dtype = "float32" min_val = float("-0.415045") @@ -1325,6 +1448,7 @@ class Program_weight_tensor_parameter_122: class Program_weight_tensor_parameter_123: name = "parameter_123" + original_name = "layer_norm_112.b_0" shape = [320] dtype = "float32" min_val = float("-0.830805") @@ -1336,6 +1460,7 @@ class Program_weight_tensor_parameter_123: class Program_weight_tensor_parameter_124: name = "parameter_124" + original_name = "layer_norm_112.w_0" shape = [320] dtype = "float32" min_val = float("-0.239324") @@ -1347,6 +1472,7 @@ class Program_weight_tensor_parameter_124: class Program_weight_tensor_parameter_125: name = "parameter_125" + original_name = "conv2d_74.b_0" shape = [320] dtype = "float32" min_val = float("-3.55682") @@ -1358,6 +1484,7 @@ class Program_weight_tensor_parameter_125: class Program_weight_tensor_parameter_126: name = "parameter_126" + original_name = "conv2d_74.w_0" shape = [320, 320, 2, 2] dtype = "float32" min_val = float("-0.280195") @@ -1369,6 +1496,7 @@ class Program_weight_tensor_parameter_126: class Program_weight_tensor_parameter_127: name = "parameter_127" + original_name = "linear_175.b_0" shape = [320] dtype = "float32" min_val = float("-2.58726") @@ -1380,6 +1508,7 @@ class Program_weight_tensor_parameter_127: class Program_weight_tensor_parameter_128: name = "parameter_128" + original_name = "linear_175.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.282214") @@ -1391,6 +1520,7 @@ class Program_weight_tensor_parameter_128: class Program_weight_tensor_parameter_129: name = "parameter_129" + original_name = "layer_norm_111.b_0" shape = [320] dtype = "float32" min_val = float("-1.25897") @@ -1402,6 +1532,7 @@ class Program_weight_tensor_parameter_129: class Program_weight_tensor_parameter_130: name = "parameter_130" + original_name = "layer_norm_111.w_0" shape = [320] dtype = "float32" min_val = float("0.217126") @@ -1413,6 +1544,7 @@ class Program_weight_tensor_parameter_130: class Program_weight_tensor_parameter_131: name = "parameter_131" + original_name = "linear_174.b_0" shape = [320] dtype = "float32" min_val = float("-0.415906") @@ -1424,6 +1556,7 @@ class Program_weight_tensor_parameter_131: class Program_weight_tensor_parameter_132: name = "parameter_132" + original_name = "linear_174.w_0" shape = [1280, 320] dtype = "float32" min_val = float("-0.723878") @@ -1435,6 +1568,7 @@ class Program_weight_tensor_parameter_132: class Program_weight_tensor_parameter_133: name = "parameter_133" + original_name = "conv2d_73.b_0" shape = [1280] dtype = "float32" min_val = float("-2.75604") @@ -1446,6 +1580,7 @@ class Program_weight_tensor_parameter_133: class Program_weight_tensor_parameter_134: name = "parameter_134" + original_name = "conv2d_73.w_0" shape = [1280, 1, 3, 3] dtype = "float32" min_val = float("-0.946538") @@ -1457,6 +1592,7 @@ class Program_weight_tensor_parameter_134: class Program_weight_tensor_parameter_135: name = "parameter_135" + original_name = "linear_173.b_0" shape = [1280] dtype = "float32" min_val = float("-2.46651") @@ -1468,6 +1604,7 @@ class Program_weight_tensor_parameter_135: class Program_weight_tensor_parameter_136: name = "parameter_136" + original_name = "linear_173.w_0" shape = [320, 1280] dtype = "float32" min_val = float("-0.519054") @@ -1479,6 +1616,7 @@ class Program_weight_tensor_parameter_136: class Program_weight_tensor_parameter_137: name = "parameter_137" + original_name = "layer_norm_110.b_0" shape = [320] dtype = "float32" min_val = float("-0.930883") @@ -1490,6 +1628,7 @@ class Program_weight_tensor_parameter_137: class Program_weight_tensor_parameter_138: name = "parameter_138" + original_name = "layer_norm_110.w_0" shape = [320] dtype = "float32" min_val = float("0.119395") @@ -1501,6 +1640,7 @@ class Program_weight_tensor_parameter_138: class Program_weight_tensor_parameter_139: name = "parameter_139" + original_name = "linear_172.b_0" shape = [320] dtype = "float32" min_val = float("-0.202187") @@ -1512,6 +1652,7 @@ class Program_weight_tensor_parameter_139: class Program_weight_tensor_parameter_140: name = "parameter_140" + original_name = "linear_172.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.634861") @@ -1523,6 +1664,7 @@ class Program_weight_tensor_parameter_140: class Program_weight_tensor_parameter_141: name = "parameter_141" + original_name = "linear_171.b_0" shape = [640] dtype = "float32" min_val = float("-6.37172") @@ -1534,6 +1676,7 @@ class Program_weight_tensor_parameter_141: class Program_weight_tensor_parameter_142: name = "parameter_142" + original_name = "linear_171.w_0" shape = [320, 640] dtype = "float32" min_val = float("-0.523848") @@ -1545,6 +1688,7 @@ class Program_weight_tensor_parameter_142: class Program_weight_tensor_parameter_143: name = "parameter_143" + original_name = "layer_norm_109.b_0" shape = [320] dtype = "float32" min_val = float("-0.325406") @@ -1556,6 +1700,7 @@ class Program_weight_tensor_parameter_143: class Program_weight_tensor_parameter_144: name = "parameter_144" + original_name = "layer_norm_109.w_0" shape = [320] dtype = "float32" min_val = float("-0.0137739") @@ -1567,6 +1712,7 @@ class Program_weight_tensor_parameter_144: class Program_weight_tensor_parameter_145: name = "parameter_145" + original_name = "conv2d_72.b_0" shape = [320] dtype = "float32" min_val = float("-4.54959") @@ -1578,6 +1724,7 @@ class Program_weight_tensor_parameter_145: class Program_weight_tensor_parameter_146: name = "parameter_146" + original_name = "conv2d_72.w_0" shape = [320, 320, 2, 2] dtype = "float32" min_val = float("-0.538833") @@ -1589,6 +1736,7 @@ class Program_weight_tensor_parameter_146: class Program_weight_tensor_parameter_147: name = "parameter_147" + original_name = "linear_170.b_0" shape = [320] dtype = "float32" min_val = float("-2.80913") @@ -1600,6 +1748,7 @@ class Program_weight_tensor_parameter_147: class Program_weight_tensor_parameter_148: name = "parameter_148" + original_name = "linear_170.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.300707") @@ -1611,6 +1760,7 @@ class Program_weight_tensor_parameter_148: class Program_weight_tensor_parameter_149: name = "parameter_149" + original_name = "layer_norm_108.b_0" shape = [320] dtype = "float32" min_val = float("-1.42616") @@ -1622,6 +1772,7 @@ class Program_weight_tensor_parameter_149: class Program_weight_tensor_parameter_150: name = "parameter_150" + original_name = "layer_norm_108.w_0" shape = [320] dtype = "float32" min_val = float("0.183767") @@ -1633,6 +1784,7 @@ class Program_weight_tensor_parameter_150: class Program_weight_tensor_parameter_151: name = "parameter_151" + original_name = "linear_169.b_0" shape = [320] dtype = "float32" min_val = float("-0.321798") @@ -1644,6 +1796,7 @@ class Program_weight_tensor_parameter_151: class Program_weight_tensor_parameter_152: name = "parameter_152" + original_name = "linear_169.w_0" shape = [1280, 320] dtype = "float32" min_val = float("-0.976216") @@ -1655,6 +1808,7 @@ class Program_weight_tensor_parameter_152: class Program_weight_tensor_parameter_153: name = "parameter_153" + original_name = "conv2d_71.b_0" shape = [1280] dtype = "float32" min_val = float("-1.68754") @@ -1666,6 +1820,7 @@ class Program_weight_tensor_parameter_153: class Program_weight_tensor_parameter_154: name = "parameter_154" + original_name = "conv2d_71.w_0" shape = [1280, 1, 3, 3] dtype = "float32" min_val = float("-0.708606") @@ -1677,6 +1832,7 @@ class Program_weight_tensor_parameter_154: class Program_weight_tensor_parameter_155: name = "parameter_155" + original_name = "linear_168.b_0" shape = [1280] dtype = "float32" min_val = float("-2.98253") @@ -1688,6 +1844,7 @@ class Program_weight_tensor_parameter_155: class Program_weight_tensor_parameter_156: name = "parameter_156" + original_name = "linear_168.w_0" shape = [320, 1280] dtype = "float32" min_val = float("-0.284315") @@ -1699,6 +1856,7 @@ class Program_weight_tensor_parameter_156: class Program_weight_tensor_parameter_157: name = "parameter_157" + original_name = "layer_norm_107.b_0" shape = [320] dtype = "float32" min_val = float("-1.02321") @@ -1710,6 +1868,7 @@ class Program_weight_tensor_parameter_157: class Program_weight_tensor_parameter_158: name = "parameter_158" + original_name = "layer_norm_107.w_0" shape = [320] dtype = "float32" min_val = float("0.137445") @@ -1721,6 +1880,7 @@ class Program_weight_tensor_parameter_158: class Program_weight_tensor_parameter_159: name = "parameter_159" + original_name = "linear_167.b_0" shape = [320] dtype = "float32" min_val = float("-0.117676") @@ -1732,6 +1892,7 @@ class Program_weight_tensor_parameter_159: class Program_weight_tensor_parameter_160: name = "parameter_160" + original_name = "linear_167.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.726305") @@ -1743,6 +1904,7 @@ class Program_weight_tensor_parameter_160: class Program_weight_tensor_parameter_161: name = "parameter_161" + original_name = "linear_166.b_0" shape = [640] dtype = "float32" min_val = float("-1.617") @@ -1754,6 +1916,7 @@ class Program_weight_tensor_parameter_161: class Program_weight_tensor_parameter_162: name = "parameter_162" + original_name = "linear_166.w_0" shape = [320, 640] dtype = "float32" min_val = float("-0.443749") @@ -1765,6 +1928,7 @@ class Program_weight_tensor_parameter_162: class Program_weight_tensor_parameter_163: name = "parameter_163" + original_name = "layer_norm_106.b_0" shape = [320] dtype = "float32" min_val = float("-0.448465") @@ -1776,6 +1940,7 @@ class Program_weight_tensor_parameter_163: class Program_weight_tensor_parameter_164: name = "parameter_164" + original_name = "layer_norm_106.w_0" shape = [320] dtype = "float32" min_val = float("-0.0456717") @@ -1787,6 +1952,7 @@ class Program_weight_tensor_parameter_164: class Program_weight_tensor_parameter_165: name = "parameter_165" + original_name = "conv2d_70.b_0" shape = [320] dtype = "float32" min_val = float("-3.48795") @@ -1798,6 +1964,7 @@ class Program_weight_tensor_parameter_165: class Program_weight_tensor_parameter_166: name = "parameter_166" + original_name = "conv2d_70.w_0" shape = [320, 320, 2, 2] dtype = "float32" min_val = float("-0.521329") @@ -1809,6 +1976,7 @@ class Program_weight_tensor_parameter_166: class Program_weight_tensor_parameter_167: name = "parameter_167" + original_name = "linear_165.b_0" shape = [320] dtype = "float32" min_val = float("-2.60367") @@ -1820,6 +1988,7 @@ class Program_weight_tensor_parameter_167: class Program_weight_tensor_parameter_168: name = "parameter_168" + original_name = "linear_165.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.30295") @@ -1831,6 +2000,7 @@ class Program_weight_tensor_parameter_168: class Program_weight_tensor_parameter_169: name = "parameter_169" + original_name = "layer_norm_105.b_0" shape = [320] dtype = "float32" min_val = float("-1.99444") @@ -1842,6 +2012,7 @@ class Program_weight_tensor_parameter_169: class Program_weight_tensor_parameter_170: name = "parameter_170" + original_name = "layer_norm_105.w_0" shape = [320] dtype = "float32" min_val = float("0.231471") @@ -1853,6 +2024,7 @@ class Program_weight_tensor_parameter_170: class Program_weight_tensor_parameter_171: name = "parameter_171" + original_name = "linear_164.b_0" shape = [320] dtype = "float32" min_val = float("-0.313673") @@ -1864,6 +2036,7 @@ class Program_weight_tensor_parameter_171: class Program_weight_tensor_parameter_172: name = "parameter_172" + original_name = "linear_164.w_0" shape = [1280, 320] dtype = "float32" min_val = float("-0.732987") @@ -1875,6 +2048,7 @@ class Program_weight_tensor_parameter_172: class Program_weight_tensor_parameter_173: name = "parameter_173" + original_name = "conv2d_69.b_0" shape = [1280] dtype = "float32" min_val = float("-1.83336") @@ -1886,6 +2060,7 @@ class Program_weight_tensor_parameter_173: class Program_weight_tensor_parameter_174: name = "parameter_174" + original_name = "conv2d_69.w_0" shape = [1280, 1, 3, 3] dtype = "float32" min_val = float("-0.574736") @@ -1897,6 +2072,7 @@ class Program_weight_tensor_parameter_174: class Program_weight_tensor_parameter_175: name = "parameter_175" + original_name = "linear_163.b_0" shape = [1280] dtype = "float32" min_val = float("-1.64122") @@ -1908,6 +2084,7 @@ class Program_weight_tensor_parameter_175: class Program_weight_tensor_parameter_176: name = "parameter_176" + original_name = "linear_163.w_0" shape = [320, 1280] dtype = "float32" min_val = float("-0.709737") @@ -1919,6 +2096,7 @@ class Program_weight_tensor_parameter_176: class Program_weight_tensor_parameter_177: name = "parameter_177" + original_name = "layer_norm_104.b_0" shape = [320] dtype = "float32" min_val = float("-1.07873") @@ -1930,6 +2108,7 @@ class Program_weight_tensor_parameter_177: class Program_weight_tensor_parameter_178: name = "parameter_178" + original_name = "layer_norm_104.w_0" shape = [320] dtype = "float32" min_val = float("0.158754") @@ -1941,6 +2120,7 @@ class Program_weight_tensor_parameter_178: class Program_weight_tensor_parameter_179: name = "parameter_179" + original_name = "linear_162.b_0" shape = [320] dtype = "float32" min_val = float("-0.267689") @@ -1952,6 +2132,7 @@ class Program_weight_tensor_parameter_179: class Program_weight_tensor_parameter_180: name = "parameter_180" + original_name = "linear_162.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.607529") @@ -1963,6 +2144,7 @@ class Program_weight_tensor_parameter_180: class Program_weight_tensor_parameter_181: name = "parameter_181" + original_name = "linear_161.b_0" shape = [640] dtype = "float32" min_val = float("-4.50914") @@ -1974,6 +2156,7 @@ class Program_weight_tensor_parameter_181: class Program_weight_tensor_parameter_182: name = "parameter_182" + original_name = "linear_161.w_0" shape = [320, 640] dtype = "float32" min_val = float("-0.413852") @@ -1985,6 +2168,7 @@ class Program_weight_tensor_parameter_182: class Program_weight_tensor_parameter_183: name = "parameter_183" + original_name = "layer_norm_103.b_0" shape = [320] dtype = "float32" min_val = float("-0.232262") @@ -1996,6 +2180,7 @@ class Program_weight_tensor_parameter_183: class Program_weight_tensor_parameter_184: name = "parameter_184" + original_name = "layer_norm_103.w_0" shape = [320] dtype = "float32" min_val = float("-0.0273772") @@ -2007,6 +2192,7 @@ class Program_weight_tensor_parameter_184: class Program_weight_tensor_parameter_185: name = "parameter_185" + original_name = "conv2d_68.b_0" shape = [320] dtype = "float32" min_val = float("-4.06267") @@ -2018,6 +2204,7 @@ class Program_weight_tensor_parameter_185: class Program_weight_tensor_parameter_186: name = "parameter_186" + original_name = "conv2d_68.w_0" shape = [320, 320, 2, 2] dtype = "float32" min_val = float("-0.462078") @@ -2029,6 +2216,7 @@ class Program_weight_tensor_parameter_186: class Program_weight_tensor_parameter_187: name = "parameter_187" + original_name = "linear_160.b_0" shape = [320] dtype = "float32" min_val = float("-2.62914") @@ -2040,6 +2228,7 @@ class Program_weight_tensor_parameter_187: class Program_weight_tensor_parameter_188: name = "parameter_188" + original_name = "linear_160.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.334351") @@ -2051,6 +2240,7 @@ class Program_weight_tensor_parameter_188: class Program_weight_tensor_parameter_189: name = "parameter_189" + original_name = "layer_norm_102.b_0" shape = [320] dtype = "float32" min_val = float("-1.69036") @@ -2062,6 +2252,7 @@ class Program_weight_tensor_parameter_189: class Program_weight_tensor_parameter_190: name = "parameter_190" + original_name = "layer_norm_102.w_0" shape = [320] dtype = "float32" min_val = float("0.209597") @@ -2073,6 +2264,7 @@ class Program_weight_tensor_parameter_190: class Program_weight_tensor_parameter_191: name = "parameter_191" + original_name = "linear_159.b_0" shape = [320] dtype = "float32" min_val = float("-0.322043") @@ -2084,6 +2276,7 @@ class Program_weight_tensor_parameter_191: class Program_weight_tensor_parameter_192: name = "parameter_192" + original_name = "linear_159.w_0" shape = [1280, 320] dtype = "float32" min_val = float("-0.906983") @@ -2095,6 +2288,7 @@ class Program_weight_tensor_parameter_192: class Program_weight_tensor_parameter_193: name = "parameter_193" + original_name = "conv2d_67.b_0" shape = [1280] dtype = "float32" min_val = float("-1.28493") @@ -2106,6 +2300,7 @@ class Program_weight_tensor_parameter_193: class Program_weight_tensor_parameter_194: name = "parameter_194" + original_name = "conv2d_67.w_0" shape = [1280, 1, 3, 3] dtype = "float32" min_val = float("-0.354917") @@ -2117,6 +2312,7 @@ class Program_weight_tensor_parameter_194: class Program_weight_tensor_parameter_195: name = "parameter_195" + original_name = "linear_158.b_0" shape = [1280] dtype = "float32" min_val = float("-2.06108") @@ -2128,6 +2324,7 @@ class Program_weight_tensor_parameter_195: class Program_weight_tensor_parameter_196: name = "parameter_196" + original_name = "linear_158.w_0" shape = [320, 1280] dtype = "float32" min_val = float("-0.357727") @@ -2139,6 +2336,7 @@ class Program_weight_tensor_parameter_196: class Program_weight_tensor_parameter_197: name = "parameter_197" + original_name = "layer_norm_101.b_0" shape = [320] dtype = "float32" min_val = float("-1.11964") @@ -2150,6 +2348,7 @@ class Program_weight_tensor_parameter_197: class Program_weight_tensor_parameter_198: name = "parameter_198" + original_name = "layer_norm_101.w_0" shape = [320] dtype = "float32" min_val = float("0.18133") @@ -2161,6 +2360,7 @@ class Program_weight_tensor_parameter_198: class Program_weight_tensor_parameter_199: name = "parameter_199" + original_name = "linear_157.b_0" shape = [320] dtype = "float32" min_val = float("-0.0947414") @@ -2172,6 +2372,7 @@ class Program_weight_tensor_parameter_199: class Program_weight_tensor_parameter_200: name = "parameter_200" + original_name = "linear_157.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.61402") @@ -2183,6 +2384,7 @@ class Program_weight_tensor_parameter_200: class Program_weight_tensor_parameter_201: name = "parameter_201" + original_name = "linear_156.b_0" shape = [640] dtype = "float32" min_val = float("-1.31202") @@ -2194,6 +2396,7 @@ class Program_weight_tensor_parameter_201: class Program_weight_tensor_parameter_202: name = "parameter_202" + original_name = "linear_156.w_0" shape = [320, 640] dtype = "float32" min_val = float("-0.416914") @@ -2205,6 +2408,7 @@ class Program_weight_tensor_parameter_202: class Program_weight_tensor_parameter_203: name = "parameter_203" + original_name = "layer_norm_100.b_0" shape = [320] dtype = "float32" min_val = float("-0.111854") @@ -2216,6 +2420,7 @@ class Program_weight_tensor_parameter_203: class Program_weight_tensor_parameter_204: name = "parameter_204" + original_name = "layer_norm_100.w_0" shape = [320] dtype = "float32" min_val = float("-0.0415021") @@ -2227,6 +2432,7 @@ class Program_weight_tensor_parameter_204: class Program_weight_tensor_parameter_205: name = "parameter_205" + original_name = "conv2d_66.b_0" shape = [320] dtype = "float32" min_val = float("-7.29373") @@ -2238,6 +2444,7 @@ class Program_weight_tensor_parameter_205: class Program_weight_tensor_parameter_206: name = "parameter_206" + original_name = "conv2d_66.w_0" shape = [320, 320, 2, 2] dtype = "float32" min_val = float("-0.53357") @@ -2249,6 +2456,7 @@ class Program_weight_tensor_parameter_206: class Program_weight_tensor_parameter_207: name = "parameter_207" + original_name = "linear_155.b_0" shape = [320] dtype = "float32" min_val = float("-2.02768") @@ -2260,6 +2468,7 @@ class Program_weight_tensor_parameter_207: class Program_weight_tensor_parameter_208: name = "parameter_208" + original_name = "linear_155.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.277326") @@ -2271,6 +2480,7 @@ class Program_weight_tensor_parameter_208: class Program_weight_tensor_parameter_209: name = "parameter_209" + original_name = "layer_norm_99.b_0" shape = [320] dtype = "float32" min_val = float("-0.894869") @@ -2282,6 +2492,7 @@ class Program_weight_tensor_parameter_209: class Program_weight_tensor_parameter_210: name = "parameter_210" + original_name = "layer_norm_99.w_0" shape = [320] dtype = "float32" min_val = float("0.102329") @@ -2293,6 +2504,7 @@ class Program_weight_tensor_parameter_210: class Program_weight_tensor_parameter_211: name = "parameter_211" + original_name = "linear_154.b_0" shape = [320] dtype = "float32" min_val = float("-0.265216") @@ -2304,6 +2516,7 @@ class Program_weight_tensor_parameter_211: class Program_weight_tensor_parameter_212: name = "parameter_212" + original_name = "linear_154.w_0" shape = [1280, 320] dtype = "float32" min_val = float("-0.974727") @@ -2315,6 +2528,7 @@ class Program_weight_tensor_parameter_212: class Program_weight_tensor_parameter_213: name = "parameter_213" + original_name = "conv2d_65.b_0" shape = [1280] dtype = "float32" min_val = float("-1.30942") @@ -2326,6 +2540,7 @@ class Program_weight_tensor_parameter_213: class Program_weight_tensor_parameter_214: name = "parameter_214" + original_name = "conv2d_65.w_0" shape = [1280, 1, 3, 3] dtype = "float32" min_val = float("-0.37841") @@ -2337,6 +2552,7 @@ class Program_weight_tensor_parameter_214: class Program_weight_tensor_parameter_215: name = "parameter_215" + original_name = "linear_153.b_0" shape = [1280] dtype = "float32" min_val = float("-1.25661") @@ -2348,6 +2564,7 @@ class Program_weight_tensor_parameter_215: class Program_weight_tensor_parameter_216: name = "parameter_216" + original_name = "linear_153.w_0" shape = [320, 1280] dtype = "float32" min_val = float("-0.332") @@ -2359,6 +2576,7 @@ class Program_weight_tensor_parameter_216: class Program_weight_tensor_parameter_217: name = "parameter_217" + original_name = "layer_norm_98.b_0" shape = [320] dtype = "float32" min_val = float("-1.21512") @@ -2370,6 +2588,7 @@ class Program_weight_tensor_parameter_217: class Program_weight_tensor_parameter_218: name = "parameter_218" + original_name = "layer_norm_98.w_0" shape = [320] dtype = "float32" min_val = float("0.201712") @@ -2381,6 +2600,7 @@ class Program_weight_tensor_parameter_218: class Program_weight_tensor_parameter_219: name = "parameter_219" + original_name = "linear_152.b_0" shape = [320] dtype = "float32" min_val = float("-0.114747") @@ -2392,6 +2612,7 @@ class Program_weight_tensor_parameter_219: class Program_weight_tensor_parameter_220: name = "parameter_220" + original_name = "linear_152.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.436594") @@ -2403,6 +2624,7 @@ class Program_weight_tensor_parameter_220: class Program_weight_tensor_parameter_221: name = "parameter_221" + original_name = "linear_151.b_0" shape = [640] dtype = "float32" min_val = float("-1.0579") @@ -2414,6 +2636,7 @@ class Program_weight_tensor_parameter_221: class Program_weight_tensor_parameter_222: name = "parameter_222" + original_name = "linear_151.w_0" shape = [320, 640] dtype = "float32" min_val = float("-0.516492") @@ -2425,6 +2648,7 @@ class Program_weight_tensor_parameter_222: class Program_weight_tensor_parameter_223: name = "parameter_223" + original_name = "layer_norm_97.b_0" shape = [320] dtype = "float32" min_val = float("-0.238922") @@ -2436,6 +2660,7 @@ class Program_weight_tensor_parameter_223: class Program_weight_tensor_parameter_224: name = "parameter_224" + original_name = "layer_norm_97.w_0" shape = [320] dtype = "float32" min_val = float("-0.0327622") @@ -2447,6 +2672,7 @@ class Program_weight_tensor_parameter_224: class Program_weight_tensor_parameter_225: name = "parameter_225" + original_name = "conv2d_64.b_0" shape = [320] dtype = "float32" min_val = float("-6.6498") @@ -2458,6 +2684,7 @@ class Program_weight_tensor_parameter_225: class Program_weight_tensor_parameter_226: name = "parameter_226" + original_name = "conv2d_64.w_0" shape = [320, 320, 2, 2] dtype = "float32" min_val = float("-0.480615") @@ -2469,6 +2696,7 @@ class Program_weight_tensor_parameter_226: class Program_weight_tensor_parameter_227: name = "parameter_227" + original_name = "linear_150.b_0" shape = [320] dtype = "float32" min_val = float("-2.32358") @@ -2480,6 +2708,7 @@ class Program_weight_tensor_parameter_227: class Program_weight_tensor_parameter_228: name = "parameter_228" + original_name = "linear_150.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.295444") @@ -2491,6 +2720,7 @@ class Program_weight_tensor_parameter_228: class Program_weight_tensor_parameter_229: name = "parameter_229" + original_name = "layer_norm_96.b_0" shape = [320] dtype = "float32" min_val = float("-1.19372") @@ -2502,6 +2732,7 @@ class Program_weight_tensor_parameter_229: class Program_weight_tensor_parameter_230: name = "parameter_230" + original_name = "layer_norm_96.w_0" shape = [320] dtype = "float32" min_val = float("0.15452") @@ -2513,6 +2744,7 @@ class Program_weight_tensor_parameter_230: class Program_weight_tensor_parameter_231: name = "parameter_231" + original_name = "linear_149.b_0" shape = [320] dtype = "float32" min_val = float("-0.341411") @@ -2524,6 +2756,7 @@ class Program_weight_tensor_parameter_231: class Program_weight_tensor_parameter_232: name = "parameter_232" + original_name = "linear_149.w_0" shape = [1280, 320] dtype = "float32" min_val = float("-0.668858") @@ -2535,6 +2768,7 @@ class Program_weight_tensor_parameter_232: class Program_weight_tensor_parameter_233: name = "parameter_233" + original_name = "conv2d_63.b_0" shape = [1280] dtype = "float32" min_val = float("-1.04738") @@ -2546,6 +2780,7 @@ class Program_weight_tensor_parameter_233: class Program_weight_tensor_parameter_234: name = "parameter_234" + original_name = "conv2d_63.w_0" shape = [1280, 1, 3, 3] dtype = "float32" min_val = float("-0.316829") @@ -2557,6 +2792,7 @@ class Program_weight_tensor_parameter_234: class Program_weight_tensor_parameter_235: name = "parameter_235" + original_name = "linear_148.b_0" shape = [1280] dtype = "float32" min_val = float("-2.48447") @@ -2568,6 +2804,7 @@ class Program_weight_tensor_parameter_235: class Program_weight_tensor_parameter_236: name = "parameter_236" + original_name = "linear_148.w_0" shape = [320, 1280] dtype = "float32" min_val = float("-0.284366") @@ -2579,6 +2816,7 @@ class Program_weight_tensor_parameter_236: class Program_weight_tensor_parameter_237: name = "parameter_237" + original_name = "layer_norm_95.b_0" shape = [320] dtype = "float32" min_val = float("-1.36582") @@ -2590,6 +2828,7 @@ class Program_weight_tensor_parameter_237: class Program_weight_tensor_parameter_238: name = "parameter_238" + original_name = "layer_norm_95.w_0" shape = [320] dtype = "float32" min_val = float("0.249821") @@ -2601,6 +2840,7 @@ class Program_weight_tensor_parameter_238: class Program_weight_tensor_parameter_239: name = "parameter_239" + original_name = "linear_147.b_0" shape = [320] dtype = "float32" min_val = float("-0.123265") @@ -2612,6 +2852,7 @@ class Program_weight_tensor_parameter_239: class Program_weight_tensor_parameter_240: name = "parameter_240" + original_name = "linear_147.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.314012") @@ -2623,6 +2864,7 @@ class Program_weight_tensor_parameter_240: class Program_weight_tensor_parameter_241: name = "parameter_241" + original_name = "linear_146.b_0" shape = [640] dtype = "float32" min_val = float("-0.844808") @@ -2634,6 +2876,7 @@ class Program_weight_tensor_parameter_241: class Program_weight_tensor_parameter_242: name = "parameter_242" + original_name = "linear_146.w_0" shape = [320, 640] dtype = "float32" min_val = float("-0.48751") @@ -2645,6 +2888,7 @@ class Program_weight_tensor_parameter_242: class Program_weight_tensor_parameter_243: name = "parameter_243" + original_name = "layer_norm_94.b_0" shape = [320] dtype = "float32" min_val = float("-0.266673") @@ -2656,6 +2900,7 @@ class Program_weight_tensor_parameter_243: class Program_weight_tensor_parameter_244: name = "parameter_244" + original_name = "layer_norm_94.w_0" shape = [320] dtype = "float32" min_val = float("-0.0928434") @@ -2667,6 +2912,7 @@ class Program_weight_tensor_parameter_244: class Program_weight_tensor_parameter_245: name = "parameter_245" + original_name = "conv2d_62.b_0" shape = [320] dtype = "float32" min_val = float("-6.06298") @@ -2678,6 +2924,7 @@ class Program_weight_tensor_parameter_245: class Program_weight_tensor_parameter_246: name = "parameter_246" + original_name = "conv2d_62.w_0" shape = [320, 320, 2, 2] dtype = "float32" min_val = float("-0.62179") @@ -2689,6 +2936,7 @@ class Program_weight_tensor_parameter_246: class Program_weight_tensor_parameter_247: name = "parameter_247" + original_name = "linear_145.b_0" shape = [320] dtype = "float32" min_val = float("-2.38848") @@ -2700,6 +2948,7 @@ class Program_weight_tensor_parameter_247: class Program_weight_tensor_parameter_248: name = "parameter_248" + original_name = "linear_145.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.29057") @@ -2711,6 +2960,7 @@ class Program_weight_tensor_parameter_248: class Program_weight_tensor_parameter_249: name = "parameter_249" + original_name = "layer_norm_93.b_0" shape = [320] dtype = "float32" min_val = float("-1.21963") @@ -2722,6 +2972,7 @@ class Program_weight_tensor_parameter_249: class Program_weight_tensor_parameter_250: name = "parameter_250" + original_name = "layer_norm_93.w_0" shape = [320] dtype = "float32" min_val = float("0.159279") @@ -2733,6 +2984,7 @@ class Program_weight_tensor_parameter_250: class Program_weight_tensor_parameter_251: name = "parameter_251" + original_name = "linear_144.b_0" shape = [320] dtype = "float32" min_val = float("-0.314957") @@ -2744,6 +2996,7 @@ class Program_weight_tensor_parameter_251: class Program_weight_tensor_parameter_252: name = "parameter_252" + original_name = "linear_144.w_0" shape = [1280, 320] dtype = "float32" min_val = float("-0.868632") @@ -2755,6 +3008,7 @@ class Program_weight_tensor_parameter_252: class Program_weight_tensor_parameter_253: name = "parameter_253" + original_name = "conv2d_61.b_0" shape = [1280] dtype = "float32" min_val = float("-1.63993") @@ -2766,6 +3020,7 @@ class Program_weight_tensor_parameter_253: class Program_weight_tensor_parameter_254: name = "parameter_254" + original_name = "conv2d_61.w_0" shape = [1280, 1, 3, 3] dtype = "float32" min_val = float("-0.385165") @@ -2777,6 +3032,7 @@ class Program_weight_tensor_parameter_254: class Program_weight_tensor_parameter_255: name = "parameter_255" + original_name = "linear_143.b_0" shape = [1280] dtype = "float32" min_val = float("-1.81705") @@ -2788,6 +3044,7 @@ class Program_weight_tensor_parameter_255: class Program_weight_tensor_parameter_256: name = "parameter_256" + original_name = "linear_143.w_0" shape = [320, 1280] dtype = "float32" min_val = float("-0.767502") @@ -2799,6 +3056,7 @@ class Program_weight_tensor_parameter_256: class Program_weight_tensor_parameter_257: name = "parameter_257" + original_name = "layer_norm_92.b_0" shape = [320] dtype = "float32" min_val = float("-1.50059") @@ -2810,6 +3068,7 @@ class Program_weight_tensor_parameter_257: class Program_weight_tensor_parameter_258: name = "parameter_258" + original_name = "layer_norm_92.w_0" shape = [320] dtype = "float32" min_val = float("0.304763") @@ -2821,6 +3080,7 @@ class Program_weight_tensor_parameter_258: class Program_weight_tensor_parameter_259: name = "parameter_259" + original_name = "linear_142.b_0" shape = [320] dtype = "float32" min_val = float("-0.155954") @@ -2832,6 +3092,7 @@ class Program_weight_tensor_parameter_259: class Program_weight_tensor_parameter_260: name = "parameter_260" + original_name = "linear_142.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.268636") @@ -2843,6 +3104,7 @@ class Program_weight_tensor_parameter_260: class Program_weight_tensor_parameter_261: name = "parameter_261" + original_name = "linear_141.b_0" shape = [640] dtype = "float32" min_val = float("-1.59446") @@ -2854,6 +3116,7 @@ class Program_weight_tensor_parameter_261: class Program_weight_tensor_parameter_262: name = "parameter_262" + original_name = "linear_141.w_0" shape = [320, 640] dtype = "float32" min_val = float("-0.494858") @@ -2865,6 +3128,7 @@ class Program_weight_tensor_parameter_262: class Program_weight_tensor_parameter_263: name = "parameter_263" + original_name = "layer_norm_91.b_0" shape = [320] dtype = "float32" min_val = float("-0.199155") @@ -2876,6 +3140,7 @@ class Program_weight_tensor_parameter_263: class Program_weight_tensor_parameter_264: name = "parameter_264" + original_name = "layer_norm_91.w_0" shape = [320] dtype = "float32" min_val = float("-0.0726563") @@ -2887,6 +3152,7 @@ class Program_weight_tensor_parameter_264: class Program_weight_tensor_parameter_265: name = "parameter_265" + original_name = "conv2d_60.b_0" shape = [320] dtype = "float32" min_val = float("-6.15728") @@ -2898,6 +3164,7 @@ class Program_weight_tensor_parameter_265: class Program_weight_tensor_parameter_266: name = "parameter_266" + original_name = "conv2d_60.w_0" shape = [320, 320, 2, 2] dtype = "float32" min_val = float("-0.475044") @@ -2909,6 +3176,7 @@ class Program_weight_tensor_parameter_266: class Program_weight_tensor_parameter_267: name = "parameter_267" + original_name = "linear_140.b_0" shape = [320] dtype = "float32" min_val = float("-2.46898") @@ -2920,6 +3188,7 @@ class Program_weight_tensor_parameter_267: class Program_weight_tensor_parameter_268: name = "parameter_268" + original_name = "linear_140.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.288684") @@ -2931,6 +3200,7 @@ class Program_weight_tensor_parameter_268: class Program_weight_tensor_parameter_269: name = "parameter_269" + original_name = "layer_norm_90.b_0" shape = [320] dtype = "float32" min_val = float("-1.14593") @@ -2942,6 +3212,7 @@ class Program_weight_tensor_parameter_269: class Program_weight_tensor_parameter_270: name = "parameter_270" + original_name = "layer_norm_90.w_0" shape = [320] dtype = "float32" min_val = float("0.164899") @@ -2953,6 +3224,7 @@ class Program_weight_tensor_parameter_270: class Program_weight_tensor_parameter_271: name = "parameter_271" + original_name = "linear_139.b_0" shape = [320] dtype = "float32" min_val = float("-0.34052") @@ -2964,6 +3236,7 @@ class Program_weight_tensor_parameter_271: class Program_weight_tensor_parameter_272: name = "parameter_272" + original_name = "linear_139.w_0" shape = [1280, 320] dtype = "float32" min_val = float("-0.840497") @@ -2975,6 +3248,7 @@ class Program_weight_tensor_parameter_272: class Program_weight_tensor_parameter_273: name = "parameter_273" + original_name = "conv2d_59.b_0" shape = [1280] dtype = "float32" min_val = float("-1.78877") @@ -2986,6 +3260,7 @@ class Program_weight_tensor_parameter_273: class Program_weight_tensor_parameter_274: name = "parameter_274" + original_name = "conv2d_59.w_0" shape = [1280, 1, 3, 3] dtype = "float32" min_val = float("-1.52953") @@ -2997,6 +3272,7 @@ class Program_weight_tensor_parameter_274: class Program_weight_tensor_parameter_275: name = "parameter_275" + original_name = "linear_138.b_0" shape = [1280] dtype = "float32" min_val = float("-1.71127") @@ -3008,6 +3284,7 @@ class Program_weight_tensor_parameter_275: class Program_weight_tensor_parameter_276: name = "parameter_276" + original_name = "linear_138.w_0" shape = [320, 1280] dtype = "float32" min_val = float("-0.723822") @@ -3019,6 +3296,7 @@ class Program_weight_tensor_parameter_276: class Program_weight_tensor_parameter_277: name = "parameter_277" + original_name = "layer_norm_89.b_0" shape = [320] dtype = "float32" min_val = float("-1.56486") @@ -3030,6 +3308,7 @@ class Program_weight_tensor_parameter_277: class Program_weight_tensor_parameter_278: name = "parameter_278" + original_name = "layer_norm_89.w_0" shape = [320] dtype = "float32" min_val = float("0.358165") @@ -3041,6 +3320,7 @@ class Program_weight_tensor_parameter_278: class Program_weight_tensor_parameter_279: name = "parameter_279" + original_name = "linear_137.b_0" shape = [320] dtype = "float32" min_val = float("-0.118528") @@ -3052,6 +3332,7 @@ class Program_weight_tensor_parameter_279: class Program_weight_tensor_parameter_280: name = "parameter_280" + original_name = "linear_137.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.264504") @@ -3063,6 +3344,7 @@ class Program_weight_tensor_parameter_280: class Program_weight_tensor_parameter_281: name = "parameter_281" + original_name = "linear_136.b_0" shape = [640] dtype = "float32" min_val = float("-0.804097") @@ -3074,6 +3356,7 @@ class Program_weight_tensor_parameter_281: class Program_weight_tensor_parameter_282: name = "parameter_282" + original_name = "linear_136.w_0" shape = [320, 640] dtype = "float32" min_val = float("-0.384582") @@ -3085,6 +3368,7 @@ class Program_weight_tensor_parameter_282: class Program_weight_tensor_parameter_283: name = "parameter_283" + original_name = "layer_norm_88.b_0" shape = [320] dtype = "float32" min_val = float("-0.14997") @@ -3096,6 +3380,7 @@ class Program_weight_tensor_parameter_283: class Program_weight_tensor_parameter_284: name = "parameter_284" + original_name = "layer_norm_88.w_0" shape = [320] dtype = "float32" min_val = float("-0.0830172") @@ -3107,6 +3392,7 @@ class Program_weight_tensor_parameter_284: class Program_weight_tensor_parameter_285: name = "parameter_285" + original_name = "conv2d_58.b_0" shape = [320] dtype = "float32" min_val = float("-6.02289") @@ -3118,6 +3404,7 @@ class Program_weight_tensor_parameter_285: class Program_weight_tensor_parameter_286: name = "parameter_286" + original_name = "conv2d_58.w_0" shape = [320, 320, 2, 2] dtype = "float32" min_val = float("-0.42281") @@ -3129,6 +3416,7 @@ class Program_weight_tensor_parameter_286: class Program_weight_tensor_parameter_287: name = "parameter_287" + original_name = "linear_135.b_0" shape = [320] dtype = "float32" min_val = float("-2.31654") @@ -3140,6 +3428,7 @@ class Program_weight_tensor_parameter_287: class Program_weight_tensor_parameter_288: name = "parameter_288" + original_name = "linear_135.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.236055") @@ -3151,6 +3440,7 @@ class Program_weight_tensor_parameter_288: class Program_weight_tensor_parameter_289: name = "parameter_289" + original_name = "layer_norm_87.b_0" shape = [320] dtype = "float32" min_val = float("-0.7576") @@ -3162,6 +3452,7 @@ class Program_weight_tensor_parameter_289: class Program_weight_tensor_parameter_290: name = "parameter_290" + original_name = "layer_norm_87.w_0" shape = [320] dtype = "float32" min_val = float("0.11233") @@ -3173,6 +3464,7 @@ class Program_weight_tensor_parameter_290: class Program_weight_tensor_parameter_291: name = "parameter_291" + original_name = "linear_134.b_0" shape = [320] dtype = "float32" min_val = float("-0.291849") @@ -3184,6 +3476,7 @@ class Program_weight_tensor_parameter_291: class Program_weight_tensor_parameter_292: name = "parameter_292" + original_name = "linear_134.w_0" shape = [1280, 320] dtype = "float32" min_val = float("-0.96494") @@ -3195,6 +3488,7 @@ class Program_weight_tensor_parameter_292: class Program_weight_tensor_parameter_293: name = "parameter_293" + original_name = "conv2d_57.b_0" shape = [1280] dtype = "float32" min_val = float("-1.42085") @@ -3206,6 +3500,7 @@ class Program_weight_tensor_parameter_293: class Program_weight_tensor_parameter_294: name = "parameter_294" + original_name = "conv2d_57.w_0" shape = [1280, 1, 3, 3] dtype = "float32" min_val = float("-1.20375") @@ -3217,6 +3512,7 @@ class Program_weight_tensor_parameter_294: class Program_weight_tensor_parameter_295: name = "parameter_295" + original_name = "linear_133.b_0" shape = [1280] dtype = "float32" min_val = float("-1.40237") @@ -3228,6 +3524,7 @@ class Program_weight_tensor_parameter_295: class Program_weight_tensor_parameter_296: name = "parameter_296" + original_name = "linear_133.w_0" shape = [320, 1280] dtype = "float32" min_val = float("-0.485896") @@ -3239,6 +3536,7 @@ class Program_weight_tensor_parameter_296: class Program_weight_tensor_parameter_297: name = "parameter_297" + original_name = "layer_norm_86.b_0" shape = [320] dtype = "float32" min_val = float("-1.74211") @@ -3250,6 +3548,7 @@ class Program_weight_tensor_parameter_297: class Program_weight_tensor_parameter_298: name = "parameter_298" + original_name = "layer_norm_86.w_0" shape = [320] dtype = "float32" min_val = float("0.483486") @@ -3261,6 +3560,7 @@ class Program_weight_tensor_parameter_298: class Program_weight_tensor_parameter_299: name = "parameter_299" + original_name = "linear_132.b_0" shape = [320] dtype = "float32" min_val = float("-0.1444") @@ -3272,6 +3572,7 @@ class Program_weight_tensor_parameter_299: class Program_weight_tensor_parameter_300: name = "parameter_300" + original_name = "linear_132.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.552184") @@ -3283,6 +3584,7 @@ class Program_weight_tensor_parameter_300: class Program_weight_tensor_parameter_301: name = "parameter_301" + original_name = "linear_131.b_0" shape = [640] dtype = "float32" min_val = float("-0.713936") @@ -3294,6 +3596,7 @@ class Program_weight_tensor_parameter_301: class Program_weight_tensor_parameter_302: name = "parameter_302" + original_name = "linear_131.w_0" shape = [320, 640] dtype = "float32" min_val = float("-0.433285") @@ -3305,6 +3608,7 @@ class Program_weight_tensor_parameter_302: class Program_weight_tensor_parameter_303: name = "parameter_303" + original_name = "layer_norm_85.b_0" shape = [320] dtype = "float32" min_val = float("-0.122442") @@ -3316,6 +3620,7 @@ class Program_weight_tensor_parameter_303: class Program_weight_tensor_parameter_304: name = "parameter_304" + original_name = "layer_norm_85.w_0" shape = [320] dtype = "float32" min_val = float("-0.00550539") @@ -3327,6 +3632,7 @@ class Program_weight_tensor_parameter_304: class Program_weight_tensor_parameter_305: name = "parameter_305" + original_name = "conv2d_56.b_0" shape = [320] dtype = "float32" min_val = float("-5.74037") @@ -3338,6 +3644,7 @@ class Program_weight_tensor_parameter_305: class Program_weight_tensor_parameter_306: name = "parameter_306" + original_name = "conv2d_56.w_0" shape = [320, 320, 2, 2] dtype = "float32" min_val = float("-0.419255") @@ -3349,6 +3656,7 @@ class Program_weight_tensor_parameter_306: class Program_weight_tensor_parameter_307: name = "parameter_307" + original_name = "linear_130.b_0" shape = [320] dtype = "float32" min_val = float("-3.17443") @@ -3360,6 +3668,7 @@ class Program_weight_tensor_parameter_307: class Program_weight_tensor_parameter_308: name = "parameter_308" + original_name = "linear_130.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.25203") @@ -3371,6 +3680,7 @@ class Program_weight_tensor_parameter_308: class Program_weight_tensor_parameter_309: name = "parameter_309" + original_name = "layer_norm_84.b_0" shape = [320] dtype = "float32" min_val = float("-0.864557") @@ -3382,6 +3692,7 @@ class Program_weight_tensor_parameter_309: class Program_weight_tensor_parameter_310: name = "parameter_310" + original_name = "layer_norm_84.w_0" shape = [320] dtype = "float32" min_val = float("0.145404") @@ -3393,6 +3704,7 @@ class Program_weight_tensor_parameter_310: class Program_weight_tensor_parameter_311: name = "parameter_311" + original_name = "linear_129.b_0" shape = [320] dtype = "float32" min_val = float("-0.301808") @@ -3404,6 +3716,7 @@ class Program_weight_tensor_parameter_311: class Program_weight_tensor_parameter_312: name = "parameter_312" + original_name = "linear_129.w_0" shape = [1280, 320] dtype = "float32" min_val = float("-0.613709") @@ -3415,6 +3728,7 @@ class Program_weight_tensor_parameter_312: class Program_weight_tensor_parameter_313: name = "parameter_313" + original_name = "conv2d_55.b_0" shape = [1280] dtype = "float32" min_val = float("-1.59302") @@ -3426,6 +3740,7 @@ class Program_weight_tensor_parameter_313: class Program_weight_tensor_parameter_314: name = "parameter_314" + original_name = "conv2d_55.w_0" shape = [1280, 1, 3, 3] dtype = "float32" min_val = float("-0.760021") @@ -3437,6 +3752,7 @@ class Program_weight_tensor_parameter_314: class Program_weight_tensor_parameter_315: name = "parameter_315" + original_name = "linear_128.b_0" shape = [1280] dtype = "float32" min_val = float("-3.27037") @@ -3448,6 +3764,7 @@ class Program_weight_tensor_parameter_315: class Program_weight_tensor_parameter_316: name = "parameter_316" + original_name = "linear_128.w_0" shape = [320, 1280] dtype = "float32" min_val = float("-0.496509") @@ -3459,6 +3776,7 @@ class Program_weight_tensor_parameter_316: class Program_weight_tensor_parameter_317: name = "parameter_317" + original_name = "layer_norm_83.b_0" shape = [320] dtype = "float32" min_val = float("-1.78432") @@ -3470,6 +3788,7 @@ class Program_weight_tensor_parameter_317: class Program_weight_tensor_parameter_318: name = "parameter_318" + original_name = "layer_norm_83.w_0" shape = [320] dtype = "float32" min_val = float("0.504915") @@ -3481,6 +3800,7 @@ class Program_weight_tensor_parameter_318: class Program_weight_tensor_parameter_319: name = "parameter_319" + original_name = "linear_127.b_0" shape = [320] dtype = "float32" min_val = float("-0.138444") @@ -3492,6 +3812,7 @@ class Program_weight_tensor_parameter_319: class Program_weight_tensor_parameter_320: name = "parameter_320" + original_name = "linear_127.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.517245") @@ -3503,6 +3824,7 @@ class Program_weight_tensor_parameter_320: class Program_weight_tensor_parameter_321: name = "parameter_321" + original_name = "linear_126.b_0" shape = [640] dtype = "float32" min_val = float("-0.724884") @@ -3514,6 +3836,7 @@ class Program_weight_tensor_parameter_321: class Program_weight_tensor_parameter_322: name = "parameter_322" + original_name = "linear_126.w_0" shape = [320, 640] dtype = "float32" min_val = float("-0.410461") @@ -3525,6 +3848,7 @@ class Program_weight_tensor_parameter_322: class Program_weight_tensor_parameter_323: name = "parameter_323" + original_name = "layer_norm_82.b_0" shape = [320] dtype = "float32" min_val = float("-0.183915") @@ -3536,6 +3860,7 @@ class Program_weight_tensor_parameter_323: class Program_weight_tensor_parameter_324: name = "parameter_324" + original_name = "layer_norm_82.w_0" shape = [320] dtype = "float32" min_val = float("-0.0233878") @@ -3547,6 +3872,7 @@ class Program_weight_tensor_parameter_324: class Program_weight_tensor_parameter_325: name = "parameter_325" + original_name = "conv2d_54.b_0" shape = [320] dtype = "float32" min_val = float("-7.56581") @@ -3558,6 +3884,7 @@ class Program_weight_tensor_parameter_325: class Program_weight_tensor_parameter_326: name = "parameter_326" + original_name = "conv2d_54.w_0" shape = [320, 320, 2, 2] dtype = "float32" min_val = float("-0.517444") @@ -3569,6 +3896,7 @@ class Program_weight_tensor_parameter_326: class Program_weight_tensor_parameter_327: name = "parameter_327" + original_name = "linear_125.b_0" shape = [320] dtype = "float32" min_val = float("-2.55769") @@ -3580,6 +3908,7 @@ class Program_weight_tensor_parameter_327: class Program_weight_tensor_parameter_328: name = "parameter_328" + original_name = "linear_125.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.266532") @@ -3591,6 +3920,7 @@ class Program_weight_tensor_parameter_328: class Program_weight_tensor_parameter_329: name = "parameter_329" + original_name = "layer_norm_81.b_0" shape = [320] dtype = "float32" min_val = float("-0.844534") @@ -3602,6 +3932,7 @@ class Program_weight_tensor_parameter_329: class Program_weight_tensor_parameter_330: name = "parameter_330" + original_name = "layer_norm_81.w_0" shape = [320] dtype = "float32" min_val = float("0.152097") @@ -3613,6 +3944,7 @@ class Program_weight_tensor_parameter_330: class Program_weight_tensor_parameter_331: name = "parameter_331" + original_name = "linear_124.b_0" shape = [320] dtype = "float32" min_val = float("-0.295306") @@ -3624,6 +3956,7 @@ class Program_weight_tensor_parameter_331: class Program_weight_tensor_parameter_332: name = "parameter_332" + original_name = "linear_124.w_0" shape = [1280, 320] dtype = "float32" min_val = float("-0.59615") @@ -3635,6 +3968,7 @@ class Program_weight_tensor_parameter_332: class Program_weight_tensor_parameter_333: name = "parameter_333" + original_name = "conv2d_53.b_0" shape = [1280] dtype = "float32" min_val = float("-1.56524") @@ -3646,6 +3980,7 @@ class Program_weight_tensor_parameter_333: class Program_weight_tensor_parameter_334: name = "parameter_334" + original_name = "conv2d_53.w_0" shape = [1280, 1, 3, 3] dtype = "float32" min_val = float("-0.322643") @@ -3657,6 +3992,7 @@ class Program_weight_tensor_parameter_334: class Program_weight_tensor_parameter_335: name = "parameter_335" + original_name = "linear_123.b_0" shape = [1280] dtype = "float32" min_val = float("-2.37229") @@ -3668,6 +4004,7 @@ class Program_weight_tensor_parameter_335: class Program_weight_tensor_parameter_336: name = "parameter_336" + original_name = "linear_123.w_0" shape = [320, 1280] dtype = "float32" min_val = float("-0.411136") @@ -3679,6 +4016,7 @@ class Program_weight_tensor_parameter_336: class Program_weight_tensor_parameter_337: name = "parameter_337" + original_name = "layer_norm_80.b_0" shape = [320] dtype = "float32" min_val = float("-1.89872") @@ -3690,6 +4028,7 @@ class Program_weight_tensor_parameter_337: class Program_weight_tensor_parameter_338: name = "parameter_338" + original_name = "layer_norm_80.w_0" shape = [320] dtype = "float32" min_val = float("0.625647") @@ -3701,6 +4040,7 @@ class Program_weight_tensor_parameter_338: class Program_weight_tensor_parameter_339: name = "parameter_339" + original_name = "linear_122.b_0" shape = [320] dtype = "float32" min_val = float("-0.141177") @@ -3712,6 +4052,7 @@ class Program_weight_tensor_parameter_339: class Program_weight_tensor_parameter_340: name = "parameter_340" + original_name = "linear_122.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.692667") @@ -3723,6 +4064,7 @@ class Program_weight_tensor_parameter_340: class Program_weight_tensor_parameter_341: name = "parameter_341" + original_name = "linear_121.b_0" shape = [640] dtype = "float32" min_val = float("-0.646605") @@ -3734,6 +4076,7 @@ class Program_weight_tensor_parameter_341: class Program_weight_tensor_parameter_342: name = "parameter_342" + original_name = "linear_121.w_0" shape = [320, 640] dtype = "float32" min_val = float("-0.391208") @@ -3745,6 +4088,7 @@ class Program_weight_tensor_parameter_342: class Program_weight_tensor_parameter_343: name = "parameter_343" + original_name = "layer_norm_79.b_0" shape = [320] dtype = "float32" min_val = float("-0.468067") @@ -3756,6 +4100,7 @@ class Program_weight_tensor_parameter_343: class Program_weight_tensor_parameter_344: name = "parameter_344" + original_name = "layer_norm_79.w_0" shape = [320] dtype = "float32" min_val = float("-0.0570338") @@ -3767,6 +4112,7 @@ class Program_weight_tensor_parameter_344: class Program_weight_tensor_parameter_345: name = "parameter_345" + original_name = "conv2d_52.b_0" shape = [320] dtype = "float32" min_val = float("-5.5669") @@ -3778,6 +4124,7 @@ class Program_weight_tensor_parameter_345: class Program_weight_tensor_parameter_346: name = "parameter_346" + original_name = "conv2d_52.w_0" shape = [320, 320, 2, 2] dtype = "float32" min_val = float("-0.316049") @@ -3789,6 +4136,7 @@ class Program_weight_tensor_parameter_346: class Program_weight_tensor_parameter_347: name = "parameter_347" + original_name = "linear_120.b_0" shape = [320] dtype = "float32" min_val = float("-2.82703") @@ -3800,6 +4148,7 @@ class Program_weight_tensor_parameter_347: class Program_weight_tensor_parameter_348: name = "parameter_348" + original_name = "linear_120.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.409982") @@ -3811,6 +4160,7 @@ class Program_weight_tensor_parameter_348: class Program_weight_tensor_parameter_349: name = "parameter_349" + original_name = "layer_norm_78.b_0" shape = [320] dtype = "float32" min_val = float("-1.05159") @@ -3822,6 +4172,7 @@ class Program_weight_tensor_parameter_349: class Program_weight_tensor_parameter_350: name = "parameter_350" + original_name = "layer_norm_78.w_0" shape = [320] dtype = "float32" min_val = float("0.223924") @@ -3833,6 +4184,7 @@ class Program_weight_tensor_parameter_350: class Program_weight_tensor_parameter_351: name = "parameter_351" + original_name = "linear_119.b_0" shape = [320] dtype = "float32" min_val = float("-0.32505") @@ -3844,6 +4196,7 @@ class Program_weight_tensor_parameter_351: class Program_weight_tensor_parameter_352: name = "parameter_352" + original_name = "linear_119.w_0" shape = [1280, 320] dtype = "float32" min_val = float("-0.660831") @@ -3855,6 +4208,7 @@ class Program_weight_tensor_parameter_352: class Program_weight_tensor_parameter_353: name = "parameter_353" + original_name = "conv2d_51.b_0" shape = [1280] dtype = "float32" min_val = float("-1.50937") @@ -3866,6 +4220,7 @@ class Program_weight_tensor_parameter_353: class Program_weight_tensor_parameter_354: name = "parameter_354" + original_name = "conv2d_51.w_0" shape = [1280, 1, 3, 3] dtype = "float32" min_val = float("-0.278985") @@ -3877,6 +4232,7 @@ class Program_weight_tensor_parameter_354: class Program_weight_tensor_parameter_355: name = "parameter_355" + original_name = "linear_118.b_0" shape = [1280] dtype = "float32" min_val = float("-3.24671") @@ -3888,6 +4244,7 @@ class Program_weight_tensor_parameter_355: class Program_weight_tensor_parameter_356: name = "parameter_356" + original_name = "linear_118.w_0" shape = [320, 1280] dtype = "float32" min_val = float("-0.523738") @@ -3899,6 +4256,7 @@ class Program_weight_tensor_parameter_356: class Program_weight_tensor_parameter_357: name = "parameter_357" + original_name = "layer_norm_77.b_0" shape = [320] dtype = "float32" min_val = float("-1.80966") @@ -3910,6 +4268,7 @@ class Program_weight_tensor_parameter_357: class Program_weight_tensor_parameter_358: name = "parameter_358" + original_name = "layer_norm_77.w_0" shape = [320] dtype = "float32" min_val = float("0.726855") @@ -3921,6 +4280,7 @@ class Program_weight_tensor_parameter_358: class Program_weight_tensor_parameter_359: name = "parameter_359" + original_name = "linear_117.b_0" shape = [320] dtype = "float32" min_val = float("-0.159628") @@ -3932,6 +4292,7 @@ class Program_weight_tensor_parameter_359: class Program_weight_tensor_parameter_360: name = "parameter_360" + original_name = "linear_117.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.615846") @@ -3943,6 +4304,7 @@ class Program_weight_tensor_parameter_360: class Program_weight_tensor_parameter_361: name = "parameter_361" + original_name = "linear_116.b_0" shape = [640] dtype = "float32" min_val = float("-0.736365") @@ -3954,6 +4316,7 @@ class Program_weight_tensor_parameter_361: class Program_weight_tensor_parameter_362: name = "parameter_362" + original_name = "linear_116.w_0" shape = [320, 640] dtype = "float32" min_val = float("-0.468391") @@ -3965,6 +4328,7 @@ class Program_weight_tensor_parameter_362: class Program_weight_tensor_parameter_363: name = "parameter_363" + original_name = "layer_norm_76.b_0" shape = [320] dtype = "float32" min_val = float("-0.179777") @@ -3976,6 +4340,7 @@ class Program_weight_tensor_parameter_363: class Program_weight_tensor_parameter_364: name = "parameter_364" + original_name = "layer_norm_76.w_0" shape = [320] dtype = "float32" min_val = float("-0.0930473") @@ -3987,6 +4352,7 @@ class Program_weight_tensor_parameter_364: class Program_weight_tensor_parameter_365: name = "parameter_365" + original_name = "conv2d_50.b_0" shape = [320] dtype = "float32" min_val = float("-5.19951") @@ -3998,6 +4364,7 @@ class Program_weight_tensor_parameter_365: class Program_weight_tensor_parameter_366: name = "parameter_366" + original_name = "conv2d_50.w_0" shape = [320, 320, 2, 2] dtype = "float32" min_val = float("-0.412921") @@ -4009,6 +4376,7 @@ class Program_weight_tensor_parameter_366: class Program_weight_tensor_parameter_367: name = "parameter_367" + original_name = "linear_115.b_0" shape = [320] dtype = "float32" min_val = float("-3.1099") @@ -4020,6 +4388,7 @@ class Program_weight_tensor_parameter_367: class Program_weight_tensor_parameter_368: name = "parameter_368" + original_name = "linear_115.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.349629") @@ -4031,6 +4400,7 @@ class Program_weight_tensor_parameter_368: class Program_weight_tensor_parameter_369: name = "parameter_369" + original_name = "layer_norm_75.b_0" shape = [320] dtype = "float32" min_val = float("-0.628736") @@ -4042,6 +4412,7 @@ class Program_weight_tensor_parameter_369: class Program_weight_tensor_parameter_370: name = "parameter_370" + original_name = "layer_norm_75.w_0" shape = [320] dtype = "float32" min_val = float("0.0907995") @@ -4053,6 +4424,7 @@ class Program_weight_tensor_parameter_370: class Program_weight_tensor_parameter_371: name = "parameter_371" + original_name = "linear_114.b_0" shape = [320] dtype = "float32" min_val = float("-0.431674") @@ -4064,6 +4436,7 @@ class Program_weight_tensor_parameter_371: class Program_weight_tensor_parameter_372: name = "parameter_372" + original_name = "linear_114.w_0" shape = [1280, 320] dtype = "float32" min_val = float("-0.777345") @@ -4075,6 +4448,7 @@ class Program_weight_tensor_parameter_372: class Program_weight_tensor_parameter_373: name = "parameter_373" + original_name = "conv2d_49.b_0" shape = [1280] dtype = "float32" min_val = float("-1.46086") @@ -4086,6 +4460,7 @@ class Program_weight_tensor_parameter_373: class Program_weight_tensor_parameter_374: name = "parameter_374" + original_name = "conv2d_49.w_0" shape = [1280, 1, 3, 3] dtype = "float32" min_val = float("-0.60767") @@ -4097,6 +4472,7 @@ class Program_weight_tensor_parameter_374: class Program_weight_tensor_parameter_375: name = "parameter_375" + original_name = "linear_113.b_0" shape = [1280] dtype = "float32" min_val = float("-3.2046") @@ -4108,6 +4484,7 @@ class Program_weight_tensor_parameter_375: class Program_weight_tensor_parameter_376: name = "parameter_376" + original_name = "linear_113.w_0" shape = [320, 1280] dtype = "float32" min_val = float("-0.537294") @@ -4119,6 +4496,7 @@ class Program_weight_tensor_parameter_376: class Program_weight_tensor_parameter_377: name = "parameter_377" + original_name = "layer_norm_74.b_0" shape = [320] dtype = "float32" min_val = float("-1.89839") @@ -4130,6 +4508,7 @@ class Program_weight_tensor_parameter_377: class Program_weight_tensor_parameter_378: name = "parameter_378" + original_name = "layer_norm_74.w_0" shape = [320] dtype = "float32" min_val = float("0.748371") @@ -4141,6 +4520,7 @@ class Program_weight_tensor_parameter_378: class Program_weight_tensor_parameter_379: name = "parameter_379" + original_name = "linear_112.b_0" shape = [320] dtype = "float32" min_val = float("-0.119465") @@ -4152,6 +4532,7 @@ class Program_weight_tensor_parameter_379: class Program_weight_tensor_parameter_380: name = "parameter_380" + original_name = "linear_112.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.662061") @@ -4163,6 +4544,7 @@ class Program_weight_tensor_parameter_380: class Program_weight_tensor_parameter_381: name = "parameter_381" + original_name = "linear_111.b_0" shape = [640] dtype = "float32" min_val = float("-0.858249") @@ -4174,6 +4556,7 @@ class Program_weight_tensor_parameter_381: class Program_weight_tensor_parameter_382: name = "parameter_382" + original_name = "linear_111.w_0" shape = [320, 640] dtype = "float32" min_val = float("-0.381867") @@ -4185,6 +4568,7 @@ class Program_weight_tensor_parameter_382: class Program_weight_tensor_parameter_383: name = "parameter_383" + original_name = "layer_norm_73.b_0" shape = [320] dtype = "float32" min_val = float("-0.17057") @@ -4196,6 +4580,7 @@ class Program_weight_tensor_parameter_383: class Program_weight_tensor_parameter_384: name = "parameter_384" + original_name = "layer_norm_73.w_0" shape = [320] dtype = "float32" min_val = float("-0.0212908") @@ -4207,6 +4592,7 @@ class Program_weight_tensor_parameter_384: class Program_weight_tensor_parameter_385: name = "parameter_385" + original_name = "conv2d_48.b_0" shape = [320] dtype = "float32" min_val = float("-5.98554") @@ -4218,6 +4604,7 @@ class Program_weight_tensor_parameter_385: class Program_weight_tensor_parameter_386: name = "parameter_386" + original_name = "conv2d_48.w_0" shape = [320, 320, 2, 2] dtype = "float32" min_val = float("-0.347335") @@ -4229,6 +4616,7 @@ class Program_weight_tensor_parameter_386: class Program_weight_tensor_parameter_387: name = "parameter_387" + original_name = "linear_110.b_0" shape = [320] dtype = "float32" min_val = float("-2.55924") @@ -4240,6 +4628,7 @@ class Program_weight_tensor_parameter_387: class Program_weight_tensor_parameter_388: name = "parameter_388" + original_name = "linear_110.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.320847") @@ -4251,6 +4640,7 @@ class Program_weight_tensor_parameter_388: class Program_weight_tensor_parameter_389: name = "parameter_389" + original_name = "layer_norm_72.b_0" shape = [320] dtype = "float32" min_val = float("-1.00207") @@ -4262,6 +4652,7 @@ class Program_weight_tensor_parameter_389: class Program_weight_tensor_parameter_390: name = "parameter_390" + original_name = "layer_norm_72.w_0" shape = [320] dtype = "float32" min_val = float("0.228424") @@ -4273,6 +4664,7 @@ class Program_weight_tensor_parameter_390: class Program_weight_tensor_parameter_391: name = "parameter_391" + original_name = "linear_109.b_0" shape = [320] dtype = "float32" min_val = float("-0.335456") @@ -4284,6 +4676,7 @@ class Program_weight_tensor_parameter_391: class Program_weight_tensor_parameter_392: name = "parameter_392" + original_name = "linear_109.w_0" shape = [1280, 320] dtype = "float32" min_val = float("-0.442903") @@ -4295,6 +4688,7 @@ class Program_weight_tensor_parameter_392: class Program_weight_tensor_parameter_393: name = "parameter_393" + original_name = "conv2d_47.b_0" shape = [1280] dtype = "float32" min_val = float("-1.94674") @@ -4306,6 +4700,7 @@ class Program_weight_tensor_parameter_393: class Program_weight_tensor_parameter_394: name = "parameter_394" + original_name = "conv2d_47.w_0" shape = [1280, 1, 3, 3] dtype = "float32" min_val = float("-0.301661") @@ -4317,6 +4712,7 @@ class Program_weight_tensor_parameter_394: class Program_weight_tensor_parameter_395: name = "parameter_395" + original_name = "linear_108.b_0" shape = [1280] dtype = "float32" min_val = float("-1.52061") @@ -4328,6 +4724,7 @@ class Program_weight_tensor_parameter_395: class Program_weight_tensor_parameter_396: name = "parameter_396" + original_name = "linear_108.w_0" shape = [320, 1280] dtype = "float32" min_val = float("-0.499755") @@ -4339,6 +4736,7 @@ class Program_weight_tensor_parameter_396: class Program_weight_tensor_parameter_397: name = "parameter_397" + original_name = "layer_norm_71.b_0" shape = [320] dtype = "float32" min_val = float("-2.06635") @@ -4350,6 +4748,7 @@ class Program_weight_tensor_parameter_397: class Program_weight_tensor_parameter_398: name = "parameter_398" + original_name = "layer_norm_71.w_0" shape = [320] dtype = "float32" min_val = float("0.865265") @@ -4361,6 +4760,7 @@ class Program_weight_tensor_parameter_398: class Program_weight_tensor_parameter_399: name = "parameter_399" + original_name = "linear_107.b_0" shape = [320] dtype = "float32" min_val = float("-0.279438") @@ -4372,6 +4772,7 @@ class Program_weight_tensor_parameter_399: class Program_weight_tensor_parameter_400: name = "parameter_400" + original_name = "linear_107.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.555429") @@ -4383,6 +4784,7 @@ class Program_weight_tensor_parameter_400: class Program_weight_tensor_parameter_401: name = "parameter_401" + original_name = "linear_106.b_0" shape = [640] dtype = "float32" min_val = float("-0.85569") @@ -4394,6 +4796,7 @@ class Program_weight_tensor_parameter_401: class Program_weight_tensor_parameter_402: name = "parameter_402" + original_name = "linear_106.w_0" shape = [320, 640] dtype = "float32" min_val = float("-0.460792") @@ -4405,6 +4808,7 @@ class Program_weight_tensor_parameter_402: class Program_weight_tensor_parameter_403: name = "parameter_403" + original_name = "layer_norm_70.b_0" shape = [320] dtype = "float32" min_val = float("-0.204199") @@ -4416,6 +4820,7 @@ class Program_weight_tensor_parameter_403: class Program_weight_tensor_parameter_404: name = "parameter_404" + original_name = "layer_norm_70.w_0" shape = [320] dtype = "float32" min_val = float("-0.00733827") @@ -4427,6 +4832,7 @@ class Program_weight_tensor_parameter_404: class Program_weight_tensor_parameter_405: name = "parameter_405" + original_name = "conv2d_46.b_0" shape = [320] dtype = "float32" min_val = float("-5.25159") @@ -4438,6 +4844,7 @@ class Program_weight_tensor_parameter_405: class Program_weight_tensor_parameter_406: name = "parameter_406" + original_name = "conv2d_46.w_0" shape = [320, 320, 2, 2] dtype = "float32" min_val = float("-0.439518") @@ -4449,6 +4856,7 @@ class Program_weight_tensor_parameter_406: class Program_weight_tensor_parameter_407: name = "parameter_407" + original_name = "linear_105.b_0" shape = [320] dtype = "float32" min_val = float("-2.12109") @@ -4460,6 +4868,7 @@ class Program_weight_tensor_parameter_407: class Program_weight_tensor_parameter_408: name = "parameter_408" + original_name = "linear_105.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.242086") @@ -4471,6 +4880,7 @@ class Program_weight_tensor_parameter_408: class Program_weight_tensor_parameter_409: name = "parameter_409" + original_name = "layer_norm_69.b_0" shape = [320] dtype = "float32" min_val = float("-0.914785") @@ -4482,6 +4892,7 @@ class Program_weight_tensor_parameter_409: class Program_weight_tensor_parameter_410: name = "parameter_410" + original_name = "layer_norm_69.w_0" shape = [320] dtype = "float32" min_val = float("0.21392") @@ -4493,6 +4904,7 @@ class Program_weight_tensor_parameter_410: class Program_weight_tensor_parameter_411: name = "parameter_411" + original_name = "linear_104.b_0" shape = [320] dtype = "float32" min_val = float("-0.363888") @@ -4504,6 +4916,7 @@ class Program_weight_tensor_parameter_411: class Program_weight_tensor_parameter_412: name = "parameter_412" + original_name = "linear_104.w_0" shape = [1280, 320] dtype = "float32" min_val = float("-0.38103") @@ -4515,6 +4928,7 @@ class Program_weight_tensor_parameter_412: class Program_weight_tensor_parameter_413: name = "parameter_413" + original_name = "conv2d_45.b_0" shape = [1280] dtype = "float32" min_val = float("-1.77555") @@ -4526,6 +4940,7 @@ class Program_weight_tensor_parameter_413: class Program_weight_tensor_parameter_414: name = "parameter_414" + original_name = "conv2d_45.w_0" shape = [1280, 1, 3, 3] dtype = "float32" min_val = float("-0.264398") @@ -4537,6 +4952,7 @@ class Program_weight_tensor_parameter_414: class Program_weight_tensor_parameter_415: name = "parameter_415" + original_name = "linear_103.b_0" shape = [1280] dtype = "float32" min_val = float("-2.9276") @@ -4548,6 +4964,7 @@ class Program_weight_tensor_parameter_415: class Program_weight_tensor_parameter_416: name = "parameter_416" + original_name = "linear_103.w_0" shape = [320, 1280] dtype = "float32" min_val = float("-0.66038") @@ -4559,6 +4976,7 @@ class Program_weight_tensor_parameter_416: class Program_weight_tensor_parameter_417: name = "parameter_417" + original_name = "layer_norm_68.b_0" shape = [320] dtype = "float32" min_val = float("-2.0844") @@ -4570,6 +4988,7 @@ class Program_weight_tensor_parameter_417: class Program_weight_tensor_parameter_418: name = "parameter_418" + original_name = "layer_norm_68.w_0" shape = [320] dtype = "float32" min_val = float("1.01034") @@ -4581,6 +5000,7 @@ class Program_weight_tensor_parameter_418: class Program_weight_tensor_parameter_419: name = "parameter_419" + original_name = "linear_102.b_0" shape = [320] dtype = "float32" min_val = float("-0.269541") @@ -4592,6 +5012,7 @@ class Program_weight_tensor_parameter_419: class Program_weight_tensor_parameter_420: name = "parameter_420" + original_name = "linear_102.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.268278") @@ -4603,6 +5024,7 @@ class Program_weight_tensor_parameter_420: class Program_weight_tensor_parameter_421: name = "parameter_421" + original_name = "linear_101.b_0" shape = [640] dtype = "float32" min_val = float("-0.688472") @@ -4614,6 +5036,7 @@ class Program_weight_tensor_parameter_421: class Program_weight_tensor_parameter_422: name = "parameter_422" + original_name = "linear_101.w_0" shape = [320, 640] dtype = "float32" min_val = float("-0.421877") @@ -4625,6 +5048,7 @@ class Program_weight_tensor_parameter_422: class Program_weight_tensor_parameter_423: name = "parameter_423" + original_name = "layer_norm_67.b_0" shape = [320] dtype = "float32" min_val = float("-0.167465") @@ -4636,6 +5060,7 @@ class Program_weight_tensor_parameter_423: class Program_weight_tensor_parameter_424: name = "parameter_424" + original_name = "layer_norm_67.w_0" shape = [320] dtype = "float32" min_val = float("-0.0124202") @@ -4647,6 +5072,7 @@ class Program_weight_tensor_parameter_424: class Program_weight_tensor_parameter_425: name = "parameter_425" + original_name = "conv2d_44.b_0" shape = [320] dtype = "float32" min_val = float("-4.95285") @@ -4658,6 +5084,7 @@ class Program_weight_tensor_parameter_425: class Program_weight_tensor_parameter_426: name = "parameter_426" + original_name = "conv2d_44.w_0" shape = [320, 320, 2, 2] dtype = "float32" min_val = float("-0.381908") @@ -4669,6 +5096,7 @@ class Program_weight_tensor_parameter_426: class Program_weight_tensor_parameter_427: name = "parameter_427" + original_name = "linear_100.b_0" shape = [320] dtype = "float32" min_val = float("-2.61862") @@ -4680,6 +5108,7 @@ class Program_weight_tensor_parameter_427: class Program_weight_tensor_parameter_428: name = "parameter_428" + original_name = "linear_100.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.243882") @@ -4691,6 +5120,7 @@ class Program_weight_tensor_parameter_428: class Program_weight_tensor_parameter_429: name = "parameter_429" + original_name = "layer_norm_66.b_0" shape = [320] dtype = "float32" min_val = float("-0.980394") @@ -4702,6 +5132,7 @@ class Program_weight_tensor_parameter_429: class Program_weight_tensor_parameter_430: name = "parameter_430" + original_name = "layer_norm_66.w_0" shape = [320] dtype = "float32" min_val = float("0.252189") @@ -4713,6 +5144,7 @@ class Program_weight_tensor_parameter_430: class Program_weight_tensor_parameter_431: name = "parameter_431" + original_name = "linear_99.b_0" shape = [320] dtype = "float32" min_val = float("-0.345725") @@ -4724,6 +5156,7 @@ class Program_weight_tensor_parameter_431: class Program_weight_tensor_parameter_432: name = "parameter_432" + original_name = "linear_99.w_0" shape = [1280, 320] dtype = "float32" min_val = float("-0.548317") @@ -4735,6 +5168,7 @@ class Program_weight_tensor_parameter_432: class Program_weight_tensor_parameter_433: name = "parameter_433" + original_name = "conv2d_43.b_0" shape = [1280] dtype = "float32" min_val = float("-1.69078") @@ -4746,6 +5180,7 @@ class Program_weight_tensor_parameter_433: class Program_weight_tensor_parameter_434: name = "parameter_434" + original_name = "conv2d_43.w_0" shape = [1280, 1, 3, 3] dtype = "float32" min_val = float("-0.303207") @@ -4757,6 +5192,7 @@ class Program_weight_tensor_parameter_434: class Program_weight_tensor_parameter_435: name = "parameter_435" + original_name = "linear_98.b_0" shape = [1280] dtype = "float32" min_val = float("-2.30794") @@ -4768,6 +5204,7 @@ class Program_weight_tensor_parameter_435: class Program_weight_tensor_parameter_436: name = "parameter_436" + original_name = "linear_98.w_0" shape = [320, 1280] dtype = "float32" min_val = float("-0.623639") @@ -4779,6 +5216,7 @@ class Program_weight_tensor_parameter_436: class Program_weight_tensor_parameter_437: name = "parameter_437" + original_name = "layer_norm_65.b_0" shape = [320] dtype = "float32" min_val = float("-1.97283") @@ -4790,6 +5228,7 @@ class Program_weight_tensor_parameter_437: class Program_weight_tensor_parameter_438: name = "parameter_438" + original_name = "layer_norm_65.w_0" shape = [320] dtype = "float32" min_val = float("1.12353") @@ -4801,6 +5240,7 @@ class Program_weight_tensor_parameter_438: class Program_weight_tensor_parameter_439: name = "parameter_439" + original_name = "linear_97.b_0" shape = [320] dtype = "float32" min_val = float("-0.238466") @@ -4812,6 +5252,7 @@ class Program_weight_tensor_parameter_439: class Program_weight_tensor_parameter_440: name = "parameter_440" + original_name = "linear_97.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.51954") @@ -4823,6 +5264,7 @@ class Program_weight_tensor_parameter_440: class Program_weight_tensor_parameter_441: name = "parameter_441" + original_name = "linear_96.b_0" shape = [640] dtype = "float32" min_val = float("-0.962348") @@ -4834,6 +5276,7 @@ class Program_weight_tensor_parameter_441: class Program_weight_tensor_parameter_442: name = "parameter_442" + original_name = "linear_96.w_0" shape = [320, 640] dtype = "float32" min_val = float("-0.347669") @@ -4845,6 +5288,7 @@ class Program_weight_tensor_parameter_442: class Program_weight_tensor_parameter_443: name = "parameter_443" + original_name = "layer_norm_64.b_0" shape = [320] dtype = "float32" min_val = float("-0.106623") @@ -4856,6 +5300,7 @@ class Program_weight_tensor_parameter_443: class Program_weight_tensor_parameter_444: name = "parameter_444" + original_name = "layer_norm_64.w_0" shape = [320] dtype = "float32" min_val = float("-0.0158363") @@ -4867,6 +5312,7 @@ class Program_weight_tensor_parameter_444: class Program_weight_tensor_parameter_445: name = "parameter_445" + original_name = "conv2d_42.b_0" shape = [320] dtype = "float32" min_val = float("-5.5619") @@ -4878,6 +5324,7 @@ class Program_weight_tensor_parameter_445: class Program_weight_tensor_parameter_446: name = "parameter_446" + original_name = "conv2d_42.w_0" shape = [320, 320, 2, 2] dtype = "float32" min_val = float("-0.515496") @@ -4889,6 +5336,7 @@ class Program_weight_tensor_parameter_446: class Program_weight_tensor_parameter_447: name = "parameter_447" + original_name = "linear_95.b_0" shape = [320] dtype = "float32" min_val = float("-2.07699") @@ -4900,6 +5348,7 @@ class Program_weight_tensor_parameter_447: class Program_weight_tensor_parameter_448: name = "parameter_448" + original_name = "linear_95.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.281898") @@ -4911,6 +5360,7 @@ class Program_weight_tensor_parameter_448: class Program_weight_tensor_parameter_449: name = "parameter_449" + original_name = "layer_norm_63.b_0" shape = [320] dtype = "float32" min_val = float("-0.888007") @@ -4922,6 +5372,7 @@ class Program_weight_tensor_parameter_449: class Program_weight_tensor_parameter_450: name = "parameter_450" + original_name = "layer_norm_63.w_0" shape = [320] dtype = "float32" min_val = float("0.205935") @@ -4933,6 +5384,7 @@ class Program_weight_tensor_parameter_450: class Program_weight_tensor_parameter_451: name = "parameter_451" + original_name = "linear_94.b_0" shape = [320] dtype = "float32" min_val = float("-0.384038") @@ -4944,6 +5396,7 @@ class Program_weight_tensor_parameter_451: class Program_weight_tensor_parameter_452: name = "parameter_452" + original_name = "linear_94.w_0" shape = [1280, 320] dtype = "float32" min_val = float("-0.414227") @@ -4955,6 +5408,7 @@ class Program_weight_tensor_parameter_452: class Program_weight_tensor_parameter_453: name = "parameter_453" + original_name = "conv2d_41.b_0" shape = [1280] dtype = "float32" min_val = float("-1.3894") @@ -4966,6 +5420,7 @@ class Program_weight_tensor_parameter_453: class Program_weight_tensor_parameter_454: name = "parameter_454" + original_name = "conv2d_41.w_0" shape = [1280, 1, 3, 3] dtype = "float32" min_val = float("-0.31159") @@ -4977,6 +5432,7 @@ class Program_weight_tensor_parameter_454: class Program_weight_tensor_parameter_455: name = "parameter_455" + original_name = "linear_93.b_0" shape = [1280] dtype = "float32" min_val = float("-1.81557") @@ -4988,6 +5444,7 @@ class Program_weight_tensor_parameter_455: class Program_weight_tensor_parameter_456: name = "parameter_456" + original_name = "linear_93.w_0" shape = [320, 1280] dtype = "float32" min_val = float("-0.440553") @@ -4999,6 +5456,7 @@ class Program_weight_tensor_parameter_456: class Program_weight_tensor_parameter_457: name = "parameter_457" + original_name = "layer_norm_62.b_0" shape = [320] dtype = "float32" min_val = float("-1.84869") @@ -5010,6 +5468,7 @@ class Program_weight_tensor_parameter_457: class Program_weight_tensor_parameter_458: name = "parameter_458" + original_name = "layer_norm_62.w_0" shape = [320] dtype = "float32" min_val = float("1.06165") @@ -5021,6 +5480,7 @@ class Program_weight_tensor_parameter_458: class Program_weight_tensor_parameter_459: name = "parameter_459" + original_name = "linear_92.b_0" shape = [320] dtype = "float32" min_val = float("-0.10052") @@ -5032,6 +5492,7 @@ class Program_weight_tensor_parameter_459: class Program_weight_tensor_parameter_460: name = "parameter_460" + original_name = "linear_92.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.574163") @@ -5043,6 +5504,7 @@ class Program_weight_tensor_parameter_460: class Program_weight_tensor_parameter_461: name = "parameter_461" + original_name = "linear_91.b_0" shape = [640] dtype = "float32" min_val = float("-0.611984") @@ -5054,6 +5516,7 @@ class Program_weight_tensor_parameter_461: class Program_weight_tensor_parameter_462: name = "parameter_462" + original_name = "linear_91.w_0" shape = [320, 640] dtype = "float32" min_val = float("-0.363361") @@ -5065,6 +5528,7 @@ class Program_weight_tensor_parameter_462: class Program_weight_tensor_parameter_463: name = "parameter_463" + original_name = "layer_norm_61.b_0" shape = [320] dtype = "float32" min_val = float("-0.147203") @@ -5076,6 +5540,7 @@ class Program_weight_tensor_parameter_463: class Program_weight_tensor_parameter_464: name = "parameter_464" + original_name = "layer_norm_61.w_0" shape = [320] dtype = "float32" min_val = float("-0.073161") @@ -5087,6 +5552,7 @@ class Program_weight_tensor_parameter_464: class Program_weight_tensor_parameter_465: name = "parameter_465" + original_name = "conv2d_40.b_0" shape = [320] dtype = "float32" min_val = float("-3.84773") @@ -5098,6 +5564,7 @@ class Program_weight_tensor_parameter_465: class Program_weight_tensor_parameter_466: name = "parameter_466" + original_name = "conv2d_40.w_0" shape = [320, 320, 2, 2] dtype = "float32" min_val = float("-0.297027") @@ -5109,6 +5576,7 @@ class Program_weight_tensor_parameter_466: class Program_weight_tensor_parameter_467: name = "parameter_467" + original_name = "linear_90.b_0" shape = [320] dtype = "float32" min_val = float("-3.16401") @@ -5120,6 +5588,7 @@ class Program_weight_tensor_parameter_467: class Program_weight_tensor_parameter_468: name = "parameter_468" + original_name = "linear_90.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.256327") @@ -5131,6 +5600,7 @@ class Program_weight_tensor_parameter_468: class Program_weight_tensor_parameter_469: name = "parameter_469" + original_name = "layer_norm_60.b_0" shape = [320] dtype = "float32" min_val = float("-0.770296") @@ -5142,6 +5612,7 @@ class Program_weight_tensor_parameter_469: class Program_weight_tensor_parameter_470: name = "parameter_470" + original_name = "layer_norm_60.w_0" shape = [320] dtype = "float32" min_val = float("-0.229292") @@ -5153,6 +5624,7 @@ class Program_weight_tensor_parameter_470: class Program_weight_tensor_parameter_471: name = "parameter_471" + original_name = "linear_89.b_0" shape = [320] dtype = "float32" min_val = float("-0.401654") @@ -5164,6 +5636,7 @@ class Program_weight_tensor_parameter_471: class Program_weight_tensor_parameter_472: name = "parameter_472" + original_name = "linear_89.w_0" shape = [1280, 320] dtype = "float32" min_val = float("-0.306615") @@ -5175,6 +5648,7 @@ class Program_weight_tensor_parameter_472: class Program_weight_tensor_parameter_473: name = "parameter_473" + original_name = "conv2d_39.b_0" shape = [1280] dtype = "float32" min_val = float("-1.56947") @@ -5186,6 +5660,7 @@ class Program_weight_tensor_parameter_473: class Program_weight_tensor_parameter_474: name = "parameter_474" + original_name = "conv2d_39.w_0" shape = [1280, 1, 3, 3] dtype = "float32" min_val = float("-0.335425") @@ -5197,6 +5672,7 @@ class Program_weight_tensor_parameter_474: class Program_weight_tensor_parameter_475: name = "parameter_475" + original_name = "linear_88.b_0" shape = [1280] dtype = "float32" min_val = float("-1.88391") @@ -5208,6 +5684,7 @@ class Program_weight_tensor_parameter_475: class Program_weight_tensor_parameter_476: name = "parameter_476" + original_name = "linear_88.w_0" shape = [320, 1280] dtype = "float32" min_val = float("-0.43191") @@ -5219,6 +5696,7 @@ class Program_weight_tensor_parameter_476: class Program_weight_tensor_parameter_477: name = "parameter_477" + original_name = "layer_norm_59.b_0" shape = [320] dtype = "float32" min_val = float("-2.03492") @@ -5230,6 +5708,7 @@ class Program_weight_tensor_parameter_477: class Program_weight_tensor_parameter_478: name = "parameter_478" + original_name = "layer_norm_59.w_0" shape = [320] dtype = "float32" min_val = float("1.01835") @@ -5241,6 +5720,7 @@ class Program_weight_tensor_parameter_478: class Program_weight_tensor_parameter_479: name = "parameter_479" + original_name = "linear_87.b_0" shape = [320] dtype = "float32" min_val = float("-0.22337") @@ -5252,6 +5732,7 @@ class Program_weight_tensor_parameter_479: class Program_weight_tensor_parameter_480: name = "parameter_480" + original_name = "linear_87.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.265972") @@ -5263,6 +5744,7 @@ class Program_weight_tensor_parameter_480: class Program_weight_tensor_parameter_481: name = "parameter_481" + original_name = "linear_86.b_0" shape = [640] dtype = "float32" min_val = float("-0.778327") @@ -5274,6 +5756,7 @@ class Program_weight_tensor_parameter_481: class Program_weight_tensor_parameter_482: name = "parameter_482" + original_name = "linear_86.w_0" shape = [320, 640] dtype = "float32" min_val = float("-0.376258") @@ -5285,6 +5768,7 @@ class Program_weight_tensor_parameter_482: class Program_weight_tensor_parameter_483: name = "parameter_483" + original_name = "layer_norm_58.b_0" shape = [320] dtype = "float32" min_val = float("-0.281155") @@ -5296,6 +5780,7 @@ class Program_weight_tensor_parameter_483: class Program_weight_tensor_parameter_484: name = "parameter_484" + original_name = "layer_norm_58.w_0" shape = [320] dtype = "float32" min_val = float("-0.0084616") @@ -5307,6 +5792,7 @@ class Program_weight_tensor_parameter_484: class Program_weight_tensor_parameter_485: name = "parameter_485" + original_name = "conv2d_38.b_0" shape = [320] dtype = "float32" min_val = float("-5.20115") @@ -5318,6 +5804,7 @@ class Program_weight_tensor_parameter_485: class Program_weight_tensor_parameter_486: name = "parameter_486" + original_name = "conv2d_38.w_0" shape = [320, 320, 2, 2] dtype = "float32" min_val = float("-0.476842") @@ -5329,6 +5816,7 @@ class Program_weight_tensor_parameter_486: class Program_weight_tensor_parameter_487: name = "parameter_487" + original_name = "linear_85.b_0" shape = [320] dtype = "float32" min_val = float("-3.47395") @@ -5340,6 +5828,7 @@ class Program_weight_tensor_parameter_487: class Program_weight_tensor_parameter_488: name = "parameter_488" + original_name = "linear_85.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.218896") @@ -5351,6 +5840,7 @@ class Program_weight_tensor_parameter_488: class Program_weight_tensor_parameter_489: name = "parameter_489" + original_name = "layer_norm_57.b_0" shape = [320] dtype = "float32" min_val = float("-0.951256") @@ -5362,6 +5852,7 @@ class Program_weight_tensor_parameter_489: class Program_weight_tensor_parameter_490: name = "parameter_490" + original_name = "layer_norm_57.w_0" shape = [320] dtype = "float32" min_val = float("0.316476") @@ -5373,6 +5864,7 @@ class Program_weight_tensor_parameter_490: class Program_weight_tensor_parameter_491: name = "parameter_491" + original_name = "linear_84.b_0" shape = [320] dtype = "float32" min_val = float("-0.388369") @@ -5384,6 +5876,7 @@ class Program_weight_tensor_parameter_491: class Program_weight_tensor_parameter_492: name = "parameter_492" + original_name = "linear_84.w_0" shape = [1280, 320] dtype = "float32" min_val = float("-0.294564") @@ -5395,6 +5888,7 @@ class Program_weight_tensor_parameter_492: class Program_weight_tensor_parameter_493: name = "parameter_493" + original_name = "conv2d_37.b_0" shape = [1280] dtype = "float32" min_val = float("-1.62033") @@ -5406,6 +5900,7 @@ class Program_weight_tensor_parameter_493: class Program_weight_tensor_parameter_494: name = "parameter_494" + original_name = "conv2d_37.w_0" shape = [1280, 1, 3, 3] dtype = "float32" min_val = float("-0.345775") @@ -5417,6 +5912,7 @@ class Program_weight_tensor_parameter_494: class Program_weight_tensor_parameter_495: name = "parameter_495" + original_name = "linear_83.b_0" shape = [1280] dtype = "float32" min_val = float("-2.19565") @@ -5428,6 +5924,7 @@ class Program_weight_tensor_parameter_495: class Program_weight_tensor_parameter_496: name = "parameter_496" + original_name = "linear_83.w_0" shape = [320, 1280] dtype = "float32" min_val = float("-0.534907") @@ -5439,6 +5936,7 @@ class Program_weight_tensor_parameter_496: class Program_weight_tensor_parameter_497: name = "parameter_497" + original_name = "layer_norm_56.b_0" shape = [320] dtype = "float32" min_val = float("-1.87303") @@ -5450,6 +5948,7 @@ class Program_weight_tensor_parameter_497: class Program_weight_tensor_parameter_498: name = "parameter_498" + original_name = "layer_norm_56.w_0" shape = [320] dtype = "float32" min_val = float("1.20951") @@ -5461,6 +5960,7 @@ class Program_weight_tensor_parameter_498: class Program_weight_tensor_parameter_499: name = "parameter_499" + original_name = "linear_82.b_0" shape = [320] dtype = "float32" min_val = float("-0.13852") @@ -5472,6 +5972,7 @@ class Program_weight_tensor_parameter_499: class Program_weight_tensor_parameter_500: name = "parameter_500" + original_name = "linear_82.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.495528") @@ -5483,6 +5984,7 @@ class Program_weight_tensor_parameter_500: class Program_weight_tensor_parameter_501: name = "parameter_501" + original_name = "linear_81.b_0" shape = [640] dtype = "float32" min_val = float("-0.706431") @@ -5494,6 +5996,7 @@ class Program_weight_tensor_parameter_501: class Program_weight_tensor_parameter_502: name = "parameter_502" + original_name = "linear_81.w_0" shape = [320, 640] dtype = "float32" min_val = float("-0.391195") @@ -5505,6 +6008,7 @@ class Program_weight_tensor_parameter_502: class Program_weight_tensor_parameter_503: name = "parameter_503" + original_name = "layer_norm_55.b_0" shape = [320] dtype = "float32" min_val = float("-0.181357") @@ -5516,6 +6020,7 @@ class Program_weight_tensor_parameter_503: class Program_weight_tensor_parameter_504: name = "parameter_504" + original_name = "layer_norm_55.w_0" shape = [320] dtype = "float32" min_val = float("-0.0334577") @@ -5527,6 +6032,7 @@ class Program_weight_tensor_parameter_504: class Program_weight_tensor_parameter_505: name = "parameter_505" + original_name = "conv2d_36.b_0" shape = [320] dtype = "float32" min_val = float("-4.3661") @@ -5538,6 +6044,7 @@ class Program_weight_tensor_parameter_505: class Program_weight_tensor_parameter_506: name = "parameter_506" + original_name = "conv2d_36.w_0" shape = [320, 320, 2, 2] dtype = "float32" min_val = float("-0.362742") @@ -5549,6 +6056,7 @@ class Program_weight_tensor_parameter_506: class Program_weight_tensor_parameter_507: name = "parameter_507" + original_name = "linear_80.b_0" shape = [320] dtype = "float32" min_val = float("-2.94725") @@ -5560,6 +6068,7 @@ class Program_weight_tensor_parameter_507: class Program_weight_tensor_parameter_508: name = "parameter_508" + original_name = "linear_80.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.313824") @@ -5571,6 +6080,7 @@ class Program_weight_tensor_parameter_508: class Program_weight_tensor_parameter_509: name = "parameter_509" + original_name = "layer_norm_54.b_0" shape = [320] dtype = "float32" min_val = float("-0.913655") @@ -5582,6 +6092,7 @@ class Program_weight_tensor_parameter_509: class Program_weight_tensor_parameter_510: name = "parameter_510" + original_name = "layer_norm_54.w_0" shape = [320] dtype = "float32" min_val = float("-0.0137031") @@ -5593,6 +6104,7 @@ class Program_weight_tensor_parameter_510: class Program_weight_tensor_parameter_511: name = "parameter_511" + original_name = "linear_79.b_0" shape = [320] dtype = "float32" min_val = float("-0.448044") @@ -5604,6 +6116,7 @@ class Program_weight_tensor_parameter_511: class Program_weight_tensor_parameter_512: name = "parameter_512" + original_name = "linear_79.w_0" shape = [1280, 320] dtype = "float32" min_val = float("-0.315247") @@ -5615,6 +6128,7 @@ class Program_weight_tensor_parameter_512: class Program_weight_tensor_parameter_513: name = "parameter_513" + original_name = "conv2d_35.b_0" shape = [1280] dtype = "float32" min_val = float("-1.71996") @@ -5626,6 +6140,7 @@ class Program_weight_tensor_parameter_513: class Program_weight_tensor_parameter_514: name = "parameter_514" + original_name = "conv2d_35.w_0" shape = [1280, 1, 3, 3] dtype = "float32" min_val = float("-0.305765") @@ -5637,6 +6152,7 @@ class Program_weight_tensor_parameter_514: class Program_weight_tensor_parameter_515: name = "parameter_515" + original_name = "linear_78.b_0" shape = [1280] dtype = "float32" min_val = float("-1.84189") @@ -5648,6 +6164,7 @@ class Program_weight_tensor_parameter_515: class Program_weight_tensor_parameter_516: name = "parameter_516" + original_name = "linear_78.w_0" shape = [320, 1280] dtype = "float32" min_val = float("-0.327231") @@ -5659,6 +6176,7 @@ class Program_weight_tensor_parameter_516: class Program_weight_tensor_parameter_517: name = "parameter_517" + original_name = "layer_norm_53.b_0" shape = [320] dtype = "float32" min_val = float("-1.90711") @@ -5670,6 +6188,7 @@ class Program_weight_tensor_parameter_517: class Program_weight_tensor_parameter_518: name = "parameter_518" + original_name = "layer_norm_53.w_0" shape = [320] dtype = "float32" min_val = float("0.877931") @@ -5681,6 +6200,7 @@ class Program_weight_tensor_parameter_518: class Program_weight_tensor_parameter_519: name = "parameter_519" + original_name = "linear_77.b_0" shape = [320] dtype = "float32" min_val = float("-0.266665") @@ -5692,6 +6212,7 @@ class Program_weight_tensor_parameter_519: class Program_weight_tensor_parameter_520: name = "parameter_520" + original_name = "linear_77.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.458296") @@ -5703,6 +6224,7 @@ class Program_weight_tensor_parameter_520: class Program_weight_tensor_parameter_521: name = "parameter_521" + original_name = "linear_76.b_0" shape = [640] dtype = "float32" min_val = float("-0.824347") @@ -5714,6 +6236,7 @@ class Program_weight_tensor_parameter_521: class Program_weight_tensor_parameter_522: name = "parameter_522" + original_name = "linear_76.w_0" shape = [320, 640] dtype = "float32" min_val = float("-0.342193") @@ -5725,6 +6248,7 @@ class Program_weight_tensor_parameter_522: class Program_weight_tensor_parameter_523: name = "parameter_523" + original_name = "layer_norm_52.b_0" shape = [320] dtype = "float32" min_val = float("-0.281601") @@ -5736,6 +6260,7 @@ class Program_weight_tensor_parameter_523: class Program_weight_tensor_parameter_524: name = "parameter_524" + original_name = "layer_norm_52.w_0" shape = [320] dtype = "float32" min_val = float("-0.0113149") @@ -5747,6 +6272,7 @@ class Program_weight_tensor_parameter_524: class Program_weight_tensor_parameter_525: name = "parameter_525" + original_name = "conv2d_34.b_0" shape = [320] dtype = "float32" min_val = float("-3.3083") @@ -5758,6 +6284,7 @@ class Program_weight_tensor_parameter_525: class Program_weight_tensor_parameter_526: name = "parameter_526" + original_name = "conv2d_34.w_0" shape = [320, 320, 2, 2] dtype = "float32" min_val = float("-0.479361") @@ -5769,6 +6296,7 @@ class Program_weight_tensor_parameter_526: class Program_weight_tensor_parameter_527: name = "parameter_527" + original_name = "linear_75.b_0" shape = [320] dtype = "float32" min_val = float("-3.14476") @@ -5780,6 +6308,7 @@ class Program_weight_tensor_parameter_527: class Program_weight_tensor_parameter_528: name = "parameter_528" + original_name = "linear_75.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.287166") @@ -5791,6 +6320,7 @@ class Program_weight_tensor_parameter_528: class Program_weight_tensor_parameter_529: name = "parameter_529" + original_name = "layer_norm_51.b_0" shape = [320] dtype = "float32" min_val = float("-0.691517") @@ -5802,6 +6332,7 @@ class Program_weight_tensor_parameter_529: class Program_weight_tensor_parameter_530: name = "parameter_530" + original_name = "layer_norm_51.w_0" shape = [320] dtype = "float32" min_val = float("0.00064667") @@ -5813,6 +6344,7 @@ class Program_weight_tensor_parameter_530: class Program_weight_tensor_parameter_531: name = "parameter_531" + original_name = "linear_74.b_0" shape = [320] dtype = "float32" min_val = float("-0.634162") @@ -5824,6 +6356,7 @@ class Program_weight_tensor_parameter_531: class Program_weight_tensor_parameter_532: name = "parameter_532" + original_name = "linear_74.w_0" shape = [1280, 320] dtype = "float32" min_val = float("-0.351534") @@ -5835,6 +6368,7 @@ class Program_weight_tensor_parameter_532: class Program_weight_tensor_parameter_533: name = "parameter_533" + original_name = "conv2d_33.b_0" shape = [1280] dtype = "float32" min_val = float("-2.00643") @@ -5846,6 +6380,7 @@ class Program_weight_tensor_parameter_533: class Program_weight_tensor_parameter_534: name = "parameter_534" + original_name = "conv2d_33.w_0" shape = [1280, 1, 3, 3] dtype = "float32" min_val = float("-0.254322") @@ -5857,6 +6392,7 @@ class Program_weight_tensor_parameter_534: class Program_weight_tensor_parameter_535: name = "parameter_535" + original_name = "linear_73.b_0" shape = [1280] dtype = "float32" min_val = float("-2.23057") @@ -5868,6 +6404,7 @@ class Program_weight_tensor_parameter_535: class Program_weight_tensor_parameter_536: name = "parameter_536" + original_name = "linear_73.w_0" shape = [320, 1280] dtype = "float32" min_val = float("-0.351788") @@ -5879,6 +6416,7 @@ class Program_weight_tensor_parameter_536: class Program_weight_tensor_parameter_537: name = "parameter_537" + original_name = "layer_norm_50.b_0" shape = [320] dtype = "float32" min_val = float("-1.80314") @@ -5890,6 +6428,7 @@ class Program_weight_tensor_parameter_537: class Program_weight_tensor_parameter_538: name = "parameter_538" + original_name = "layer_norm_50.w_0" shape = [320] dtype = "float32" min_val = float("0.684533") @@ -5901,6 +6440,7 @@ class Program_weight_tensor_parameter_538: class Program_weight_tensor_parameter_539: name = "parameter_539" + original_name = "linear_72.b_0" shape = [320] dtype = "float32" min_val = float("-0.209862") @@ -5912,6 +6452,7 @@ class Program_weight_tensor_parameter_539: class Program_weight_tensor_parameter_540: name = "parameter_540" + original_name = "linear_72.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.428339") @@ -5923,6 +6464,7 @@ class Program_weight_tensor_parameter_540: class Program_weight_tensor_parameter_541: name = "parameter_541" + original_name = "linear_71.b_0" shape = [640] dtype = "float32" min_val = float("-0.788902") @@ -5934,6 +6476,7 @@ class Program_weight_tensor_parameter_541: class Program_weight_tensor_parameter_542: name = "parameter_542" + original_name = "linear_71.w_0" shape = [320, 640] dtype = "float32" min_val = float("-0.385714") @@ -5945,6 +6488,7 @@ class Program_weight_tensor_parameter_542: class Program_weight_tensor_parameter_543: name = "parameter_543" + original_name = "layer_norm_49.b_0" shape = [320] dtype = "float32" min_val = float("-0.243747") @@ -5956,6 +6500,7 @@ class Program_weight_tensor_parameter_543: class Program_weight_tensor_parameter_544: name = "parameter_544" + original_name = "layer_norm_49.w_0" shape = [320] dtype = "float32" min_val = float("-0.0203944") @@ -5967,6 +6512,7 @@ class Program_weight_tensor_parameter_544: class Program_weight_tensor_parameter_545: name = "parameter_545" + original_name = "conv2d_32.b_0" shape = [320] dtype = "float32" min_val = float("-3.21577") @@ -5978,6 +6524,7 @@ class Program_weight_tensor_parameter_545: class Program_weight_tensor_parameter_546: name = "parameter_546" + original_name = "conv2d_32.w_0" shape = [320, 320, 2, 2] dtype = "float32" min_val = float("-0.283199") @@ -5989,6 +6536,7 @@ class Program_weight_tensor_parameter_546: class Program_weight_tensor_parameter_547: name = "parameter_547" + original_name = "linear_70.b_0" shape = [320] dtype = "float32" min_val = float("-2.50601") @@ -6000,6 +6548,7 @@ class Program_weight_tensor_parameter_547: class Program_weight_tensor_parameter_548: name = "parameter_548" + original_name = "linear_70.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.300327") @@ -6011,6 +6560,7 @@ class Program_weight_tensor_parameter_548: class Program_weight_tensor_parameter_549: name = "parameter_549" + original_name = "layer_norm_48.b_0" shape = [320] dtype = "float32" min_val = float("-0.459846") @@ -6022,6 +6572,7 @@ class Program_weight_tensor_parameter_549: class Program_weight_tensor_parameter_550: name = "parameter_550" + original_name = "layer_norm_48.w_0" shape = [320] dtype = "float32" min_val = float("-0.00696922") @@ -6033,6 +6584,7 @@ class Program_weight_tensor_parameter_550: class Program_weight_tensor_parameter_551: name = "parameter_551" + original_name = "linear_69.b_0" shape = [320] dtype = "float32" min_val = float("-1.03951") @@ -6044,6 +6596,7 @@ class Program_weight_tensor_parameter_551: class Program_weight_tensor_parameter_552: name = "parameter_552" + original_name = "linear_69.w_0" shape = [1280, 320] dtype = "float32" min_val = float("-0.382691") @@ -6055,6 +6608,7 @@ class Program_weight_tensor_parameter_552: class Program_weight_tensor_parameter_553: name = "parameter_553" + original_name = "conv2d_31.b_0" shape = [1280] dtype = "float32" min_val = float("-1.83716") @@ -6066,6 +6620,7 @@ class Program_weight_tensor_parameter_553: class Program_weight_tensor_parameter_554: name = "parameter_554" + original_name = "conv2d_31.w_0" shape = [1280, 1, 3, 3] dtype = "float32" min_val = float("-0.312131") @@ -6077,6 +6632,7 @@ class Program_weight_tensor_parameter_554: class Program_weight_tensor_parameter_555: name = "parameter_555" + original_name = "linear_68.b_0" shape = [1280] dtype = "float32" min_val = float("-2.47351") @@ -6088,6 +6644,7 @@ class Program_weight_tensor_parameter_555: class Program_weight_tensor_parameter_556: name = "parameter_556" + original_name = "linear_68.w_0" shape = [320, 1280] dtype = "float32" min_val = float("-0.371089") @@ -6099,6 +6656,7 @@ class Program_weight_tensor_parameter_556: class Program_weight_tensor_parameter_557: name = "parameter_557" + original_name = "layer_norm_47.b_0" shape = [320] dtype = "float32" min_val = float("-1.55264") @@ -6110,6 +6668,7 @@ class Program_weight_tensor_parameter_557: class Program_weight_tensor_parameter_558: name = "parameter_558" + original_name = "layer_norm_47.w_0" shape = [320] dtype = "float32" min_val = float("0.064097") @@ -6121,6 +6680,7 @@ class Program_weight_tensor_parameter_558: class Program_weight_tensor_parameter_559: name = "parameter_559" + original_name = "linear_67.b_0" shape = [320] dtype = "float32" min_val = float("-0.180081") @@ -6132,6 +6692,7 @@ class Program_weight_tensor_parameter_559: class Program_weight_tensor_parameter_560: name = "parameter_560" + original_name = "linear_67.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.333174") @@ -6143,6 +6704,7 @@ class Program_weight_tensor_parameter_560: class Program_weight_tensor_parameter_561: name = "parameter_561" + original_name = "linear_66.b_0" shape = [640] dtype = "float32" min_val = float("-0.764661") @@ -6154,6 +6716,7 @@ class Program_weight_tensor_parameter_561: class Program_weight_tensor_parameter_562: name = "parameter_562" + original_name = "linear_66.w_0" shape = [320, 640] dtype = "float32" min_val = float("-0.312826") @@ -6165,6 +6728,7 @@ class Program_weight_tensor_parameter_562: class Program_weight_tensor_parameter_563: name = "parameter_563" + original_name = "layer_norm_46.b_0" shape = [320] dtype = "float32" min_val = float("-0.171379") @@ -6176,6 +6740,7 @@ class Program_weight_tensor_parameter_563: class Program_weight_tensor_parameter_564: name = "parameter_564" + original_name = "layer_norm_46.w_0" shape = [320] dtype = "float32" min_val = float("-0.0800917") @@ -6187,6 +6752,7 @@ class Program_weight_tensor_parameter_564: class Program_weight_tensor_parameter_565: name = "parameter_565" + original_name = "conv2d_30.b_0" shape = [320] dtype = "float32" min_val = float("-4.88392") @@ -6198,6 +6764,7 @@ class Program_weight_tensor_parameter_565: class Program_weight_tensor_parameter_566: name = "parameter_566" + original_name = "conv2d_30.w_0" shape = [320, 320, 2, 2] dtype = "float32" min_val = float("-0.237464") @@ -6209,6 +6776,7 @@ class Program_weight_tensor_parameter_566: class Program_weight_tensor_parameter_567: name = "parameter_567" + original_name = "linear_65.b_0" shape = [320] dtype = "float32" min_val = float("-2.65374") @@ -6220,6 +6788,7 @@ class Program_weight_tensor_parameter_567: class Program_weight_tensor_parameter_568: name = "parameter_568" + original_name = "linear_65.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.305084") @@ -6231,6 +6800,7 @@ class Program_weight_tensor_parameter_568: class Program_weight_tensor_parameter_569: name = "parameter_569" + original_name = "layer_norm_45.b_0" shape = [320] dtype = "float32" min_val = float("-0.489885") @@ -6242,6 +6812,7 @@ class Program_weight_tensor_parameter_569: class Program_weight_tensor_parameter_570: name = "parameter_570" + original_name = "layer_norm_45.w_0" shape = [320] dtype = "float32" min_val = float("-0.00276871") @@ -6253,6 +6824,7 @@ class Program_weight_tensor_parameter_570: class Program_weight_tensor_parameter_571: name = "parameter_571" + original_name = "linear_64.b_0" shape = [320] dtype = "float32" min_val = float("-1.33905") @@ -6264,6 +6836,7 @@ class Program_weight_tensor_parameter_571: class Program_weight_tensor_parameter_572: name = "parameter_572" + original_name = "linear_64.w_0" shape = [1280, 320] dtype = "float32" min_val = float("-0.418753") @@ -6275,6 +6848,7 @@ class Program_weight_tensor_parameter_572: class Program_weight_tensor_parameter_573: name = "parameter_573" + original_name = "conv2d_29.b_0" shape = [1280] dtype = "float32" min_val = float("-1.75893") @@ -6286,6 +6860,7 @@ class Program_weight_tensor_parameter_573: class Program_weight_tensor_parameter_574: name = "parameter_574" + original_name = "conv2d_29.w_0" shape = [1280, 1, 3, 3] dtype = "float32" min_val = float("-0.265421") @@ -6297,6 +6872,7 @@ class Program_weight_tensor_parameter_574: class Program_weight_tensor_parameter_575: name = "parameter_575" + original_name = "linear_63.b_0" shape = [1280] dtype = "float32" min_val = float("-2.70545") @@ -6308,6 +6884,7 @@ class Program_weight_tensor_parameter_575: class Program_weight_tensor_parameter_576: name = "parameter_576" + original_name = "linear_63.w_0" shape = [320, 1280] dtype = "float32" min_val = float("-0.305355") @@ -6319,6 +6896,7 @@ class Program_weight_tensor_parameter_576: class Program_weight_tensor_parameter_577: name = "parameter_577" + original_name = "layer_norm_44.b_0" shape = [320] dtype = "float32" min_val = float("-1.46923") @@ -6330,6 +6908,7 @@ class Program_weight_tensor_parameter_577: class Program_weight_tensor_parameter_578: name = "parameter_578" + original_name = "layer_norm_44.w_0" shape = [320] dtype = "float32" min_val = float("0.556237") @@ -6341,6 +6920,7 @@ class Program_weight_tensor_parameter_578: class Program_weight_tensor_parameter_579: name = "parameter_579" + original_name = "linear_62.b_0" shape = [320] dtype = "float32" min_val = float("-0.127203") @@ -6352,6 +6932,7 @@ class Program_weight_tensor_parameter_579: class Program_weight_tensor_parameter_580: name = "parameter_580" + original_name = "linear_62.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.309202") @@ -6363,6 +6944,7 @@ class Program_weight_tensor_parameter_580: class Program_weight_tensor_parameter_581: name = "parameter_581" + original_name = "linear_61.b_0" shape = [640] dtype = "float32" min_val = float("-0.918587") @@ -6374,6 +6956,7 @@ class Program_weight_tensor_parameter_581: class Program_weight_tensor_parameter_582: name = "parameter_582" + original_name = "linear_61.w_0" shape = [320, 640] dtype = "float32" min_val = float("-0.281015") @@ -6385,6 +6968,7 @@ class Program_weight_tensor_parameter_582: class Program_weight_tensor_parameter_583: name = "parameter_583" + original_name = "layer_norm_43.b_0" shape = [320] dtype = "float32" min_val = float("-0.118987") @@ -6396,6 +6980,7 @@ class Program_weight_tensor_parameter_583: class Program_weight_tensor_parameter_584: name = "parameter_584" + original_name = "layer_norm_43.w_0" shape = [320] dtype = "float32" min_val = float("-0.019374") @@ -6407,6 +6992,7 @@ class Program_weight_tensor_parameter_584: class Program_weight_tensor_parameter_585: name = "parameter_585" + original_name = "conv2d_28.b_0" shape = [320] dtype = "float32" min_val = float("-3.3583") @@ -6418,6 +7004,7 @@ class Program_weight_tensor_parameter_585: class Program_weight_tensor_parameter_586: name = "parameter_586" + original_name = "conv2d_28.w_0" shape = [320, 320, 2, 2] dtype = "float32" min_val = float("-0.245141") @@ -6429,6 +7016,7 @@ class Program_weight_tensor_parameter_586: class Program_weight_tensor_parameter_587: name = "parameter_587" + original_name = "linear_60.b_0" shape = [320] dtype = "float32" min_val = float("-2.83937") @@ -6440,6 +7028,7 @@ class Program_weight_tensor_parameter_587: class Program_weight_tensor_parameter_588: name = "parameter_588" + original_name = "linear_60.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.27752") @@ -6451,6 +7040,7 @@ class Program_weight_tensor_parameter_588: class Program_weight_tensor_parameter_589: name = "parameter_589" + original_name = "layer_norm_42.b_0" shape = [320] dtype = "float32" min_val = float("-0.459076") @@ -6462,6 +7052,7 @@ class Program_weight_tensor_parameter_589: class Program_weight_tensor_parameter_590: name = "parameter_590" + original_name = "layer_norm_42.w_0" shape = [320] dtype = "float32" min_val = float("-0.00570921") @@ -6473,6 +7064,7 @@ class Program_weight_tensor_parameter_590: class Program_weight_tensor_parameter_591: name = "parameter_591" + original_name = "linear_59.b_0" shape = [320] dtype = "float32" min_val = float("-1.92716") @@ -6484,6 +7076,7 @@ class Program_weight_tensor_parameter_591: class Program_weight_tensor_parameter_592: name = "parameter_592" + original_name = "linear_59.w_0" shape = [1280, 320] dtype = "float32" min_val = float("-0.483879") @@ -6495,6 +7088,7 @@ class Program_weight_tensor_parameter_592: class Program_weight_tensor_parameter_593: name = "parameter_593" + original_name = "conv2d_27.b_0" shape = [1280] dtype = "float32" min_val = float("-1.49472") @@ -6506,6 +7100,7 @@ class Program_weight_tensor_parameter_593: class Program_weight_tensor_parameter_594: name = "parameter_594" + original_name = "conv2d_27.w_0" shape = [1280, 1, 3, 3] dtype = "float32" min_val = float("-0.245614") @@ -6517,6 +7112,7 @@ class Program_weight_tensor_parameter_594: class Program_weight_tensor_parameter_595: name = "parameter_595" + original_name = "linear_58.b_0" shape = [1280] dtype = "float32" min_val = float("-2.82751") @@ -6528,6 +7124,7 @@ class Program_weight_tensor_parameter_595: class Program_weight_tensor_parameter_596: name = "parameter_596" + original_name = "linear_58.w_0" shape = [320, 1280] dtype = "float32" min_val = float("-0.383605") @@ -6539,6 +7136,7 @@ class Program_weight_tensor_parameter_596: class Program_weight_tensor_parameter_597: name = "parameter_597" + original_name = "layer_norm_41.b_0" shape = [320] dtype = "float32" min_val = float("-1.49998") @@ -6550,6 +7148,7 @@ class Program_weight_tensor_parameter_597: class Program_weight_tensor_parameter_598: name = "parameter_598" + original_name = "layer_norm_41.w_0" shape = [320] dtype = "float32" min_val = float("0.512688") @@ -6561,6 +7160,7 @@ class Program_weight_tensor_parameter_598: class Program_weight_tensor_parameter_599: name = "parameter_599" + original_name = "linear_57.b_0" shape = [320] dtype = "float32" min_val = float("-0.208171") @@ -6572,6 +7172,7 @@ class Program_weight_tensor_parameter_599: class Program_weight_tensor_parameter_600: name = "parameter_600" + original_name = "linear_57.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.425122") @@ -6583,6 +7184,7 @@ class Program_weight_tensor_parameter_600: class Program_weight_tensor_parameter_601: name = "parameter_601" + original_name = "linear_56.b_0" shape = [640] dtype = "float32" min_val = float("-0.647617") @@ -6594,6 +7196,7 @@ class Program_weight_tensor_parameter_601: class Program_weight_tensor_parameter_602: name = "parameter_602" + original_name = "linear_56.w_0" shape = [320, 640] dtype = "float32" min_val = float("-0.336658") @@ -6605,6 +7208,7 @@ class Program_weight_tensor_parameter_602: class Program_weight_tensor_parameter_603: name = "parameter_603" + original_name = "layer_norm_40.b_0" shape = [320] dtype = "float32" min_val = float("-0.1775") @@ -6616,6 +7220,7 @@ class Program_weight_tensor_parameter_603: class Program_weight_tensor_parameter_604: name = "parameter_604" + original_name = "layer_norm_40.w_0" shape = [320] dtype = "float32" min_val = float("-0.0244371") @@ -6627,6 +7232,7 @@ class Program_weight_tensor_parameter_604: class Program_weight_tensor_parameter_605: name = "parameter_605" + original_name = "conv2d_26.b_0" shape = [320] dtype = "float32" min_val = float("-3.49194") @@ -6638,6 +7244,7 @@ class Program_weight_tensor_parameter_605: class Program_weight_tensor_parameter_606: name = "parameter_606" + original_name = "conv2d_26.w_0" shape = [320, 320, 2, 2] dtype = "float32" min_val = float("-0.232495") @@ -6649,6 +7256,7 @@ class Program_weight_tensor_parameter_606: class Program_weight_tensor_parameter_607: name = "parameter_607" + original_name = "linear_55.b_0" shape = [320] dtype = "float32" min_val = float("-3.38736") @@ -6660,6 +7268,7 @@ class Program_weight_tensor_parameter_607: class Program_weight_tensor_parameter_608: name = "parameter_608" + original_name = "linear_55.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.266856") @@ -6671,6 +7280,7 @@ class Program_weight_tensor_parameter_608: class Program_weight_tensor_parameter_609: name = "parameter_609" + original_name = "layer_norm_39.b_0" shape = [320] dtype = "float32" min_val = float("-0.806727") @@ -6682,6 +7292,7 @@ class Program_weight_tensor_parameter_609: class Program_weight_tensor_parameter_610: name = "parameter_610" + original_name = "layer_norm_39.w_0" shape = [320] dtype = "float32" min_val = float("-0.00950158") @@ -6693,6 +7304,7 @@ class Program_weight_tensor_parameter_610: class Program_weight_tensor_parameter_611: name = "parameter_611" + original_name = "layer_norm_2.b_0" shape = [320] dtype = "float32" min_val = float("-1.37683") @@ -6704,6 +7316,7 @@ class Program_weight_tensor_parameter_611: class Program_weight_tensor_parameter_612: name = "parameter_612" + original_name = "layer_norm_2.w_0" shape = [320] dtype = "float32" min_val = float("0.0983726") @@ -6715,6 +7328,7 @@ class Program_weight_tensor_parameter_612: class Program_weight_tensor_parameter_613: name = "parameter_613" + original_name = "conv2d_2.b_0" shape = [320] dtype = "float32" min_val = float("-1.90275") @@ -6726,6 +7340,7 @@ class Program_weight_tensor_parameter_613: class Program_weight_tensor_parameter_614: name = "parameter_614" + original_name = "conv2d_2.w_0" shape = [320, 128, 3, 3] dtype = "float32" min_val = float("-1.03018") @@ -6737,6 +7352,7 @@ class Program_weight_tensor_parameter_614: class Program_weight_tensor_parameter_615: name = "parameter_615" + original_name = "layer_norm_38.b_0" shape = [128] dtype = "float32" min_val = float("-0.414266") @@ -6748,6 +7364,7 @@ class Program_weight_tensor_parameter_615: class Program_weight_tensor_parameter_616: name = "parameter_616" + original_name = "layer_norm_38.w_0" shape = [128] dtype = "float32" min_val = float("0.0777218") @@ -6759,6 +7376,7 @@ class Program_weight_tensor_parameter_616: class Program_weight_tensor_parameter_617: name = "parameter_617" + original_name = "linear_54.b_0" shape = [128] dtype = "float32" min_val = float("-9.25915") @@ -6770,6 +7388,7 @@ class Program_weight_tensor_parameter_617: class Program_weight_tensor_parameter_618: name = "parameter_618" + original_name = "linear_54.w_0" shape = [512, 128] dtype = "float32" min_val = float("-1.78961") @@ -6781,6 +7400,7 @@ class Program_weight_tensor_parameter_618: class Program_weight_tensor_parameter_619: name = "parameter_619" + original_name = "conv2d_25.b_0" shape = [512] dtype = "float32" min_val = float("-1.30161") @@ -6792,6 +7412,7 @@ class Program_weight_tensor_parameter_619: class Program_weight_tensor_parameter_620: name = "parameter_620" + original_name = "conv2d_25.w_0" shape = [512, 1, 3, 3] dtype = "float32" min_val = float("-0.599383") @@ -6803,6 +7424,7 @@ class Program_weight_tensor_parameter_620: class Program_weight_tensor_parameter_621: name = "parameter_621" + original_name = "linear_53.b_0" shape = [512] dtype = "float32" min_val = float("-1.50283") @@ -6814,6 +7436,7 @@ class Program_weight_tensor_parameter_621: class Program_weight_tensor_parameter_622: name = "parameter_622" + original_name = "linear_53.w_0" shape = [128, 512] dtype = "float32" min_val = float("-0.275843") @@ -6825,6 +7448,7 @@ class Program_weight_tensor_parameter_622: class Program_weight_tensor_parameter_623: name = "parameter_623" + original_name = "layer_norm_37.b_0" shape = [128] dtype = "float32" min_val = float("-1.88747") @@ -6836,6 +7460,7 @@ class Program_weight_tensor_parameter_623: class Program_weight_tensor_parameter_624: name = "parameter_624" + original_name = "layer_norm_37.w_0" shape = [128] dtype = "float32" min_val = float("0.997765") @@ -6847,6 +7472,7 @@ class Program_weight_tensor_parameter_624: class Program_weight_tensor_parameter_625: name = "parameter_625" + original_name = "linear_52.b_0" shape = [128] dtype = "float32" min_val = float("-0.914004") @@ -6858,6 +7484,7 @@ class Program_weight_tensor_parameter_625: class Program_weight_tensor_parameter_626: name = "parameter_626" + original_name = "linear_52.w_0" shape = [128, 128] dtype = "float32" min_val = float("-0.785685") @@ -6869,6 +7496,7 @@ class Program_weight_tensor_parameter_626: class Program_weight_tensor_parameter_627: name = "parameter_627" + original_name = "linear_51.b_0" shape = [256] dtype = "float32" min_val = float("-0.470057") @@ -6880,6 +7508,7 @@ class Program_weight_tensor_parameter_627: class Program_weight_tensor_parameter_628: name = "parameter_628" + original_name = "linear_51.w_0" shape = [128, 256] dtype = "float32" min_val = float("-0.380423") @@ -6891,6 +7520,7 @@ class Program_weight_tensor_parameter_628: class Program_weight_tensor_parameter_629: name = "parameter_629" + original_name = "layer_norm_36.b_0" shape = [128] dtype = "float32" min_val = float("-0.29407") @@ -6902,6 +7532,7 @@ class Program_weight_tensor_parameter_629: class Program_weight_tensor_parameter_630: name = "parameter_630" + original_name = "layer_norm_36.w_0" shape = [128] dtype = "float32" min_val = float("-0.0145862") @@ -6913,6 +7544,7 @@ class Program_weight_tensor_parameter_630: class Program_weight_tensor_parameter_631: name = "parameter_631" + original_name = "conv2d_24.b_0" shape = [128] dtype = "float32" min_val = float("-2.76489") @@ -6924,6 +7556,7 @@ class Program_weight_tensor_parameter_631: class Program_weight_tensor_parameter_632: name = "parameter_632" + original_name = "conv2d_24.w_0" shape = [128, 128, 4, 4] dtype = "float32" min_val = float("-0.200871") @@ -6935,6 +7568,7 @@ class Program_weight_tensor_parameter_632: class Program_weight_tensor_parameter_633: name = "parameter_633" + original_name = "linear_50.b_0" shape = [128] dtype = "float32" min_val = float("-1.67559") @@ -6946,6 +7580,7 @@ class Program_weight_tensor_parameter_633: class Program_weight_tensor_parameter_634: name = "parameter_634" + original_name = "linear_50.w_0" shape = [128, 128] dtype = "float32" min_val = float("-0.26448") @@ -6957,6 +7592,7 @@ class Program_weight_tensor_parameter_634: class Program_weight_tensor_parameter_635: name = "parameter_635" + original_name = "layer_norm_35.b_0" shape = [128] dtype = "float32" min_val = float("-0.835337") @@ -6968,6 +7604,7 @@ class Program_weight_tensor_parameter_635: class Program_weight_tensor_parameter_636: name = "parameter_636" + original_name = "layer_norm_35.w_0" shape = [128] dtype = "float32" min_val = float("0.194558") @@ -6979,6 +7616,7 @@ class Program_weight_tensor_parameter_636: class Program_weight_tensor_parameter_637: name = "parameter_637" + original_name = "linear_49.b_0" shape = [128] dtype = "float32" min_val = float("-3.51845") @@ -6990,6 +7628,7 @@ class Program_weight_tensor_parameter_637: class Program_weight_tensor_parameter_638: name = "parameter_638" + original_name = "linear_49.w_0" shape = [512, 128] dtype = "float32" min_val = float("-1.19464") @@ -7001,6 +7640,7 @@ class Program_weight_tensor_parameter_638: class Program_weight_tensor_parameter_639: name = "parameter_639" + original_name = "conv2d_23.b_0" shape = [512] dtype = "float32" min_val = float("-1.45936") @@ -7012,6 +7652,7 @@ class Program_weight_tensor_parameter_639: class Program_weight_tensor_parameter_640: name = "parameter_640" + original_name = "conv2d_23.w_0" shape = [512, 1, 3, 3] dtype = "float32" min_val = float("-0.366268") @@ -7023,6 +7664,7 @@ class Program_weight_tensor_parameter_640: class Program_weight_tensor_parameter_641: name = "parameter_641" + original_name = "linear_48.b_0" shape = [512] dtype = "float32" min_val = float("-1.85231") @@ -7034,6 +7676,7 @@ class Program_weight_tensor_parameter_641: class Program_weight_tensor_parameter_642: name = "parameter_642" + original_name = "linear_48.w_0" shape = [128, 512] dtype = "float32" min_val = float("-0.346462") @@ -7045,6 +7688,7 @@ class Program_weight_tensor_parameter_642: class Program_weight_tensor_parameter_643: name = "parameter_643" + original_name = "layer_norm_34.b_0" shape = [128] dtype = "float32" min_val = float("-1.82166") @@ -7056,6 +7700,7 @@ class Program_weight_tensor_parameter_643: class Program_weight_tensor_parameter_644: name = "parameter_644" + original_name = "layer_norm_34.w_0" shape = [128] dtype = "float32" min_val = float("1.26883") @@ -7067,6 +7712,7 @@ class Program_weight_tensor_parameter_644: class Program_weight_tensor_parameter_645: name = "parameter_645" + original_name = "linear_47.b_0" shape = [128] dtype = "float32" min_val = float("-0.0555517") @@ -7078,6 +7724,7 @@ class Program_weight_tensor_parameter_645: class Program_weight_tensor_parameter_646: name = "parameter_646" + original_name = "linear_47.w_0" shape = [128, 128] dtype = "float32" min_val = float("-0.525702") @@ -7089,6 +7736,7 @@ class Program_weight_tensor_parameter_646: class Program_weight_tensor_parameter_647: name = "parameter_647" + original_name = "linear_46.b_0" shape = [256] dtype = "float32" min_val = float("-0.832025") @@ -7100,6 +7748,7 @@ class Program_weight_tensor_parameter_647: class Program_weight_tensor_parameter_648: name = "parameter_648" + original_name = "linear_46.w_0" shape = [128, 256] dtype = "float32" min_val = float("-0.32686") @@ -7111,6 +7760,7 @@ class Program_weight_tensor_parameter_648: class Program_weight_tensor_parameter_649: name = "parameter_649" + original_name = "layer_norm_33.b_0" shape = [128] dtype = "float32" min_val = float("-0.364127") @@ -7122,6 +7772,7 @@ class Program_weight_tensor_parameter_649: class Program_weight_tensor_parameter_650: name = "parameter_650" + original_name = "layer_norm_33.w_0" shape = [128] dtype = "float32" min_val = float("-0.0128896") @@ -7133,6 +7784,7 @@ class Program_weight_tensor_parameter_650: class Program_weight_tensor_parameter_651: name = "parameter_651" + original_name = "conv2d_22.b_0" shape = [128] dtype = "float32" min_val = float("-3.08894") @@ -7144,6 +7796,7 @@ class Program_weight_tensor_parameter_651: class Program_weight_tensor_parameter_652: name = "parameter_652" + original_name = "conv2d_22.w_0" shape = [128, 128, 4, 4] dtype = "float32" min_val = float("-0.205126") @@ -7155,6 +7808,7 @@ class Program_weight_tensor_parameter_652: class Program_weight_tensor_parameter_653: name = "parameter_653" + original_name = "linear_45.b_0" shape = [128] dtype = "float32" min_val = float("-1.42989") @@ -7166,6 +7820,7 @@ class Program_weight_tensor_parameter_653: class Program_weight_tensor_parameter_654: name = "parameter_654" + original_name = "linear_45.w_0" shape = [128, 128] dtype = "float32" min_val = float("-0.24844") @@ -7177,6 +7832,7 @@ class Program_weight_tensor_parameter_654: class Program_weight_tensor_parameter_655: name = "parameter_655" + original_name = "layer_norm_32.b_0" shape = [128] dtype = "float32" min_val = float("-0.834618") @@ -7188,6 +7844,7 @@ class Program_weight_tensor_parameter_655: class Program_weight_tensor_parameter_656: name = "parameter_656" + original_name = "layer_norm_32.w_0" shape = [128] dtype = "float32" min_val = float("0.229859") @@ -7199,6 +7856,7 @@ class Program_weight_tensor_parameter_656: class Program_weight_tensor_parameter_657: name = "parameter_657" + original_name = "linear_44.b_0" shape = [128] dtype = "float32" min_val = float("-1.87213") @@ -7210,6 +7868,7 @@ class Program_weight_tensor_parameter_657: class Program_weight_tensor_parameter_658: name = "parameter_658" + original_name = "linear_44.w_0" shape = [512, 128] dtype = "float32" min_val = float("-0.753826") @@ -7221,6 +7880,7 @@ class Program_weight_tensor_parameter_658: class Program_weight_tensor_parameter_659: name = "parameter_659" + original_name = "conv2d_21.b_0" shape = [512] dtype = "float32" min_val = float("-1.46297") @@ -7232,6 +7892,7 @@ class Program_weight_tensor_parameter_659: class Program_weight_tensor_parameter_660: name = "parameter_660" + original_name = "conv2d_21.w_0" shape = [512, 1, 3, 3] dtype = "float32" min_val = float("-0.263766") @@ -7243,6 +7904,7 @@ class Program_weight_tensor_parameter_660: class Program_weight_tensor_parameter_661: name = "parameter_661" + original_name = "linear_43.b_0" shape = [512] dtype = "float32" min_val = float("-4.30956") @@ -7254,6 +7916,7 @@ class Program_weight_tensor_parameter_661: class Program_weight_tensor_parameter_662: name = "parameter_662" + original_name = "linear_43.w_0" shape = [128, 512] dtype = "float32" min_val = float("-0.243908") @@ -7265,6 +7928,7 @@ class Program_weight_tensor_parameter_662: class Program_weight_tensor_parameter_663: name = "parameter_663" + original_name = "layer_norm_31.b_0" shape = [128] dtype = "float32" min_val = float("-1.31031") @@ -7276,6 +7940,7 @@ class Program_weight_tensor_parameter_663: class Program_weight_tensor_parameter_664: name = "parameter_664" + original_name = "layer_norm_31.w_0" shape = [128] dtype = "float32" min_val = float("1.53785") @@ -7287,6 +7952,7 @@ class Program_weight_tensor_parameter_664: class Program_weight_tensor_parameter_665: name = "parameter_665" + original_name = "linear_42.b_0" shape = [128] dtype = "float32" min_val = float("-0.072993") @@ -7298,6 +7964,7 @@ class Program_weight_tensor_parameter_665: class Program_weight_tensor_parameter_666: name = "parameter_666" + original_name = "linear_42.w_0" shape = [128, 128] dtype = "float32" min_val = float("-0.25129") @@ -7309,6 +7976,7 @@ class Program_weight_tensor_parameter_666: class Program_weight_tensor_parameter_667: name = "parameter_667" + original_name = "linear_41.b_0" shape = [256] dtype = "float32" min_val = float("-0.748558") @@ -7320,6 +7988,7 @@ class Program_weight_tensor_parameter_667: class Program_weight_tensor_parameter_668: name = "parameter_668" + original_name = "linear_41.w_0" shape = [128, 256] dtype = "float32" min_val = float("-0.25701") @@ -7331,6 +8000,7 @@ class Program_weight_tensor_parameter_668: class Program_weight_tensor_parameter_669: name = "parameter_669" + original_name = "layer_norm_30.b_0" shape = [128] dtype = "float32" min_val = float("-0.355538") @@ -7342,6 +8012,7 @@ class Program_weight_tensor_parameter_669: class Program_weight_tensor_parameter_670: name = "parameter_670" + original_name = "layer_norm_30.w_0" shape = [128] dtype = "float32" min_val = float("-0.01532") @@ -7353,6 +8024,7 @@ class Program_weight_tensor_parameter_670: class Program_weight_tensor_parameter_671: name = "parameter_671" + original_name = "conv2d_20.b_0" shape = [128] dtype = "float32" min_val = float("-4.25378") @@ -7364,6 +8036,7 @@ class Program_weight_tensor_parameter_671: class Program_weight_tensor_parameter_672: name = "parameter_672" + original_name = "conv2d_20.w_0" shape = [128, 128, 4, 4] dtype = "float32" min_val = float("-0.237647") @@ -7375,6 +8048,7 @@ class Program_weight_tensor_parameter_672: class Program_weight_tensor_parameter_673: name = "parameter_673" + original_name = "linear_40.b_0" shape = [128] dtype = "float32" min_val = float("-2.44111") @@ -7386,6 +8060,7 @@ class Program_weight_tensor_parameter_673: class Program_weight_tensor_parameter_674: name = "parameter_674" + original_name = "linear_40.w_0" shape = [128, 128] dtype = "float32" min_val = float("-0.23881") @@ -7397,6 +8072,7 @@ class Program_weight_tensor_parameter_674: class Program_weight_tensor_parameter_675: name = "parameter_675" + original_name = "layer_norm_29.b_0" shape = [128] dtype = "float32" min_val = float("-0.695864") @@ -7408,6 +8084,7 @@ class Program_weight_tensor_parameter_675: class Program_weight_tensor_parameter_676: name = "parameter_676" + original_name = "layer_norm_29.w_0" shape = [128] dtype = "float32" min_val = float("0.320919") @@ -7419,6 +8096,7 @@ class Program_weight_tensor_parameter_676: class Program_weight_tensor_parameter_677: name = "parameter_677" + original_name = "linear_39.b_0" shape = [128] dtype = "float32" min_val = float("-1.20562") @@ -7430,6 +8108,7 @@ class Program_weight_tensor_parameter_677: class Program_weight_tensor_parameter_678: name = "parameter_678" + original_name = "linear_39.w_0" shape = [512, 128] dtype = "float32" min_val = float("-0.532543") @@ -7441,6 +8120,7 @@ class Program_weight_tensor_parameter_678: class Program_weight_tensor_parameter_679: name = "parameter_679" + original_name = "conv2d_19.b_0" shape = [512] dtype = "float32" min_val = float("-1.10035") @@ -7452,6 +8132,7 @@ class Program_weight_tensor_parameter_679: class Program_weight_tensor_parameter_680: name = "parameter_680" + original_name = "conv2d_19.w_0" shape = [512, 1, 3, 3] dtype = "float32" min_val = float("-0.265156") @@ -7463,6 +8144,7 @@ class Program_weight_tensor_parameter_680: class Program_weight_tensor_parameter_681: name = "parameter_681" + original_name = "linear_38.b_0" shape = [512] dtype = "float32" min_val = float("-3.17358") @@ -7474,6 +8156,7 @@ class Program_weight_tensor_parameter_681: class Program_weight_tensor_parameter_682: name = "parameter_682" + original_name = "linear_38.w_0" shape = [128, 512] dtype = "float32" min_val = float("-0.299642") @@ -7485,6 +8168,7 @@ class Program_weight_tensor_parameter_682: class Program_weight_tensor_parameter_683: name = "parameter_683" + original_name = "layer_norm_28.b_0" shape = [128] dtype = "float32" min_val = float("-0.921507") @@ -7496,6 +8180,7 @@ class Program_weight_tensor_parameter_683: class Program_weight_tensor_parameter_684: name = "parameter_684" + original_name = "layer_norm_28.w_0" shape = [128] dtype = "float32" min_val = float("1.54851") @@ -7507,6 +8192,7 @@ class Program_weight_tensor_parameter_684: class Program_weight_tensor_parameter_685: name = "parameter_685" + original_name = "linear_37.b_0" shape = [128] dtype = "float32" min_val = float("-0.0646714") @@ -7518,6 +8204,7 @@ class Program_weight_tensor_parameter_685: class Program_weight_tensor_parameter_686: name = "parameter_686" + original_name = "linear_37.w_0" shape = [128, 128] dtype = "float32" min_val = float("-0.212155") @@ -7529,6 +8216,7 @@ class Program_weight_tensor_parameter_686: class Program_weight_tensor_parameter_687: name = "parameter_687" + original_name = "linear_36.b_0" shape = [256] dtype = "float32" min_val = float("-0.578473") @@ -7540,6 +8228,7 @@ class Program_weight_tensor_parameter_687: class Program_weight_tensor_parameter_688: name = "parameter_688" + original_name = "linear_36.w_0" shape = [128, 256] dtype = "float32" min_val = float("-0.283223") @@ -7551,6 +8240,7 @@ class Program_weight_tensor_parameter_688: class Program_weight_tensor_parameter_689: name = "parameter_689" + original_name = "layer_norm_27.b_0" shape = [128] dtype = "float32" min_val = float("-0.3853") @@ -7562,6 +8252,7 @@ class Program_weight_tensor_parameter_689: class Program_weight_tensor_parameter_690: name = "parameter_690" + original_name = "layer_norm_27.w_0" shape = [128] dtype = "float32" min_val = float("-0.0103786") @@ -7573,6 +8264,7 @@ class Program_weight_tensor_parameter_690: class Program_weight_tensor_parameter_691: name = "parameter_691" + original_name = "conv2d_18.b_0" shape = [128] dtype = "float32" min_val = float("-4.60672") @@ -7584,6 +8276,7 @@ class Program_weight_tensor_parameter_691: class Program_weight_tensor_parameter_692: name = "parameter_692" + original_name = "conv2d_18.w_0" shape = [128, 128, 4, 4] dtype = "float32" min_val = float("-0.262076") @@ -7595,6 +8288,7 @@ class Program_weight_tensor_parameter_692: class Program_weight_tensor_parameter_693: name = "parameter_693" + original_name = "linear_35.b_0" shape = [128] dtype = "float32" min_val = float("-2.0388") @@ -7606,6 +8300,7 @@ class Program_weight_tensor_parameter_693: class Program_weight_tensor_parameter_694: name = "parameter_694" + original_name = "linear_35.w_0" shape = [128, 128] dtype = "float32" min_val = float("-0.286252") @@ -7617,6 +8312,7 @@ class Program_weight_tensor_parameter_694: class Program_weight_tensor_parameter_695: name = "parameter_695" + original_name = "layer_norm_26.b_0" shape = [128] dtype = "float32" min_val = float("-1.06307") @@ -7628,6 +8324,7 @@ class Program_weight_tensor_parameter_695: class Program_weight_tensor_parameter_696: name = "parameter_696" + original_name = "layer_norm_26.w_0" shape = [128] dtype = "float32" min_val = float("0.539285") @@ -7639,6 +8336,7 @@ class Program_weight_tensor_parameter_696: class Program_weight_tensor_parameter_697: name = "parameter_697" + original_name = "linear_34.b_0" shape = [128] dtype = "float32" min_val = float("-0.947775") @@ -7650,6 +8348,7 @@ class Program_weight_tensor_parameter_697: class Program_weight_tensor_parameter_698: name = "parameter_698" + original_name = "linear_34.w_0" shape = [512, 128] dtype = "float32" min_val = float("-0.422859") @@ -7661,6 +8360,7 @@ class Program_weight_tensor_parameter_698: class Program_weight_tensor_parameter_699: name = "parameter_699" + original_name = "conv2d_17.b_0" shape = [512] dtype = "float32" min_val = float("-1.70367") @@ -7672,6 +8372,7 @@ class Program_weight_tensor_parameter_699: class Program_weight_tensor_parameter_700: name = "parameter_700" + original_name = "conv2d_17.w_0" shape = [512, 1, 3, 3] dtype = "float32" min_val = float("-0.310654") @@ -7683,6 +8384,7 @@ class Program_weight_tensor_parameter_700: class Program_weight_tensor_parameter_701: name = "parameter_701" + original_name = "linear_33.b_0" shape = [512] dtype = "float32" min_val = float("-2.73436") @@ -7694,6 +8396,7 @@ class Program_weight_tensor_parameter_701: class Program_weight_tensor_parameter_702: name = "parameter_702" + original_name = "linear_33.w_0" shape = [128, 512] dtype = "float32" min_val = float("-0.277734") @@ -7705,6 +8408,7 @@ class Program_weight_tensor_parameter_702: class Program_weight_tensor_parameter_703: name = "parameter_703" + original_name = "layer_norm_25.b_0" shape = [128] dtype = "float32" min_val = float("-0.946741") @@ -7716,6 +8420,7 @@ class Program_weight_tensor_parameter_703: class Program_weight_tensor_parameter_704: name = "parameter_704" + original_name = "layer_norm_25.w_0" shape = [128] dtype = "float32" min_val = float("1.47791") @@ -7727,6 +8432,7 @@ class Program_weight_tensor_parameter_704: class Program_weight_tensor_parameter_705: name = "parameter_705" + original_name = "linear_32.b_0" shape = [128] dtype = "float32" min_val = float("-0.0807209") @@ -7738,6 +8444,7 @@ class Program_weight_tensor_parameter_705: class Program_weight_tensor_parameter_706: name = "parameter_706" + original_name = "linear_32.w_0" shape = [128, 128] dtype = "float32" min_val = float("-0.20739") @@ -7749,6 +8456,7 @@ class Program_weight_tensor_parameter_706: class Program_weight_tensor_parameter_707: name = "parameter_707" + original_name = "linear_31.b_0" shape = [256] dtype = "float32" min_val = float("-0.647726") @@ -7760,6 +8468,7 @@ class Program_weight_tensor_parameter_707: class Program_weight_tensor_parameter_708: name = "parameter_708" + original_name = "linear_31.w_0" shape = [128, 256] dtype = "float32" min_val = float("-0.332014") @@ -7771,6 +8480,7 @@ class Program_weight_tensor_parameter_708: class Program_weight_tensor_parameter_709: name = "parameter_709" + original_name = "layer_norm_24.b_0" shape = [128] dtype = "float32" min_val = float("-0.360979") @@ -7782,6 +8492,7 @@ class Program_weight_tensor_parameter_709: class Program_weight_tensor_parameter_710: name = "parameter_710" + original_name = "layer_norm_24.w_0" shape = [128] dtype = "float32" min_val = float("-0.0160126") @@ -7793,6 +8504,7 @@ class Program_weight_tensor_parameter_710: class Program_weight_tensor_parameter_711: name = "parameter_711" + original_name = "conv2d_16.b_0" shape = [128] dtype = "float32" min_val = float("-3.48088") @@ -7804,6 +8516,7 @@ class Program_weight_tensor_parameter_711: class Program_weight_tensor_parameter_712: name = "parameter_712" + original_name = "conv2d_16.w_0" shape = [128, 128, 4, 4] dtype = "float32" min_val = float("-0.298548") @@ -7815,6 +8528,7 @@ class Program_weight_tensor_parameter_712: class Program_weight_tensor_parameter_713: name = "parameter_713" + original_name = "linear_30.b_0" shape = [128] dtype = "float32" min_val = float("-2.17597") @@ -7826,6 +8540,7 @@ class Program_weight_tensor_parameter_713: class Program_weight_tensor_parameter_714: name = "parameter_714" + original_name = "linear_30.w_0" shape = [128, 128] dtype = "float32" min_val = float("-0.329135") @@ -7837,6 +8552,7 @@ class Program_weight_tensor_parameter_714: class Program_weight_tensor_parameter_715: name = "parameter_715" + original_name = "layer_norm_23.b_0" shape = [128] dtype = "float32" min_val = float("-0.680658") @@ -7848,6 +8564,7 @@ class Program_weight_tensor_parameter_715: class Program_weight_tensor_parameter_716: name = "parameter_716" + original_name = "layer_norm_23.w_0" shape = [128] dtype = "float32" min_val = float("0.557702") @@ -7859,6 +8576,7 @@ class Program_weight_tensor_parameter_716: class Program_weight_tensor_parameter_717: name = "parameter_717" + original_name = "linear_29.b_0" shape = [128] dtype = "float32" min_val = float("-0.743833") @@ -7870,6 +8588,7 @@ class Program_weight_tensor_parameter_717: class Program_weight_tensor_parameter_718: name = "parameter_718" + original_name = "linear_29.w_0" shape = [512, 128] dtype = "float32" min_val = float("-0.239266") @@ -7881,6 +8600,7 @@ class Program_weight_tensor_parameter_718: class Program_weight_tensor_parameter_719: name = "parameter_719" + original_name = "conv2d_15.b_0" shape = [512] dtype = "float32" min_val = float("-1.45802") @@ -7892,6 +8612,7 @@ class Program_weight_tensor_parameter_719: class Program_weight_tensor_parameter_720: name = "parameter_720" + original_name = "conv2d_15.w_0" shape = [512, 1, 3, 3] dtype = "float32" min_val = float("-0.306124") @@ -7903,6 +8624,7 @@ class Program_weight_tensor_parameter_720: class Program_weight_tensor_parameter_721: name = "parameter_721" + original_name = "linear_28.b_0" shape = [512] dtype = "float32" min_val = float("-2.95294") @@ -7914,6 +8636,7 @@ class Program_weight_tensor_parameter_721: class Program_weight_tensor_parameter_722: name = "parameter_722" + original_name = "linear_28.w_0" shape = [128, 512] dtype = "float32" min_val = float("-0.329847") @@ -7925,6 +8648,7 @@ class Program_weight_tensor_parameter_722: class Program_weight_tensor_parameter_723: name = "parameter_723" + original_name = "layer_norm_22.b_0" shape = [128] dtype = "float32" min_val = float("-1.02939") @@ -7936,6 +8660,7 @@ class Program_weight_tensor_parameter_723: class Program_weight_tensor_parameter_724: name = "parameter_724" + original_name = "layer_norm_22.w_0" shape = [128] dtype = "float32" min_val = float("1.6659") @@ -7947,6 +8672,7 @@ class Program_weight_tensor_parameter_724: class Program_weight_tensor_parameter_725: name = "parameter_725" + original_name = "linear_27.b_0" shape = [128] dtype = "float32" min_val = float("-0.144514") @@ -7958,6 +8684,7 @@ class Program_weight_tensor_parameter_725: class Program_weight_tensor_parameter_726: name = "parameter_726" + original_name = "linear_27.w_0" shape = [128, 128] dtype = "float32" min_val = float("-0.196458") @@ -7969,6 +8696,7 @@ class Program_weight_tensor_parameter_726: class Program_weight_tensor_parameter_727: name = "parameter_727" + original_name = "linear_26.b_0" shape = [256] dtype = "float32" min_val = float("-0.77196") @@ -7980,6 +8708,7 @@ class Program_weight_tensor_parameter_727: class Program_weight_tensor_parameter_728: name = "parameter_728" + original_name = "linear_26.w_0" shape = [128, 256] dtype = "float32" min_val = float("-0.480152") @@ -7991,6 +8720,7 @@ class Program_weight_tensor_parameter_728: class Program_weight_tensor_parameter_729: name = "parameter_729" + original_name = "layer_norm_21.b_0" shape = [128] dtype = "float32" min_val = float("-0.283814") @@ -8002,6 +8732,7 @@ class Program_weight_tensor_parameter_729: class Program_weight_tensor_parameter_730: name = "parameter_730" + original_name = "layer_norm_21.w_0" shape = [128] dtype = "float32" min_val = float("-0.00659812") @@ -8013,6 +8744,7 @@ class Program_weight_tensor_parameter_730: class Program_weight_tensor_parameter_731: name = "parameter_731" + original_name = "conv2d_14.b_0" shape = [128] dtype = "float32" min_val = float("-3.43107") @@ -8024,6 +8756,7 @@ class Program_weight_tensor_parameter_731: class Program_weight_tensor_parameter_732: name = "parameter_732" + original_name = "conv2d_14.w_0" shape = [128, 128, 4, 4] dtype = "float32" min_val = float("-0.284808") @@ -8035,6 +8768,7 @@ class Program_weight_tensor_parameter_732: class Program_weight_tensor_parameter_733: name = "parameter_733" + original_name = "linear_25.b_0" shape = [128] dtype = "float32" min_val = float("-2.14207") @@ -8046,6 +8780,7 @@ class Program_weight_tensor_parameter_733: class Program_weight_tensor_parameter_734: name = "parameter_734" + original_name = "linear_25.w_0" shape = [128, 128] dtype = "float32" min_val = float("-0.287975") @@ -8057,6 +8792,7 @@ class Program_weight_tensor_parameter_734: class Program_weight_tensor_parameter_735: name = "parameter_735" + original_name = "layer_norm_20.b_0" shape = [128] dtype = "float32" min_val = float("-0.614829") @@ -8068,6 +8804,7 @@ class Program_weight_tensor_parameter_735: class Program_weight_tensor_parameter_736: name = "parameter_736" + original_name = "layer_norm_20.w_0" shape = [128] dtype = "float32" min_val = float("0.484106") @@ -8079,6 +8816,7 @@ class Program_weight_tensor_parameter_736: class Program_weight_tensor_parameter_737: name = "parameter_737" + original_name = "linear_24.b_0" shape = [128] dtype = "float32" min_val = float("-0.676873") @@ -8090,6 +8828,7 @@ class Program_weight_tensor_parameter_737: class Program_weight_tensor_parameter_738: name = "parameter_738" + original_name = "linear_24.w_0" shape = [512, 128] dtype = "float32" min_val = float("-0.2708") @@ -8101,6 +8840,7 @@ class Program_weight_tensor_parameter_738: class Program_weight_tensor_parameter_739: name = "parameter_739" + original_name = "conv2d_13.b_0" shape = [512] dtype = "float32" min_val = float("-1.45859") @@ -8112,6 +8852,7 @@ class Program_weight_tensor_parameter_739: class Program_weight_tensor_parameter_740: name = "parameter_740" + original_name = "conv2d_13.w_0" shape = [512, 1, 3, 3] dtype = "float32" min_val = float("-0.368161") @@ -8123,6 +8864,7 @@ class Program_weight_tensor_parameter_740: class Program_weight_tensor_parameter_741: name = "parameter_741" + original_name = "linear_23.b_0" shape = [512] dtype = "float32" min_val = float("-2.8581") @@ -8134,6 +8876,7 @@ class Program_weight_tensor_parameter_741: class Program_weight_tensor_parameter_742: name = "parameter_742" + original_name = "linear_23.w_0" shape = [128, 512] dtype = "float32" min_val = float("-0.313993") @@ -8145,6 +8888,7 @@ class Program_weight_tensor_parameter_742: class Program_weight_tensor_parameter_743: name = "parameter_743" + original_name = "layer_norm_19.b_0" shape = [128] dtype = "float32" min_val = float("-0.932659") @@ -8156,6 +8900,7 @@ class Program_weight_tensor_parameter_743: class Program_weight_tensor_parameter_744: name = "parameter_744" + original_name = "layer_norm_19.w_0" shape = [128] dtype = "float32" min_val = float("1.96267") @@ -8167,6 +8912,7 @@ class Program_weight_tensor_parameter_744: class Program_weight_tensor_parameter_745: name = "parameter_745" + original_name = "linear_22.b_0" shape = [128] dtype = "float32" min_val = float("-0.209583") @@ -8178,6 +8924,7 @@ class Program_weight_tensor_parameter_745: class Program_weight_tensor_parameter_746: name = "parameter_746" + original_name = "linear_22.w_0" shape = [128, 128] dtype = "float32" min_val = float("-0.182094") @@ -8189,6 +8936,7 @@ class Program_weight_tensor_parameter_746: class Program_weight_tensor_parameter_747: name = "parameter_747" + original_name = "linear_21.b_0" shape = [256] dtype = "float32" min_val = float("-0.861444") @@ -8200,6 +8948,7 @@ class Program_weight_tensor_parameter_747: class Program_weight_tensor_parameter_748: name = "parameter_748" + original_name = "linear_21.w_0" shape = [128, 256] dtype = "float32" min_val = float("-0.353396") @@ -8211,6 +8960,7 @@ class Program_weight_tensor_parameter_748: class Program_weight_tensor_parameter_749: name = "parameter_749" + original_name = "layer_norm_18.b_0" shape = [128] dtype = "float32" min_val = float("-0.421813") @@ -8222,6 +8972,7 @@ class Program_weight_tensor_parameter_749: class Program_weight_tensor_parameter_750: name = "parameter_750" + original_name = "layer_norm_18.w_0" shape = [128] dtype = "float32" min_val = float("-0.010542") @@ -8233,6 +8984,7 @@ class Program_weight_tensor_parameter_750: class Program_weight_tensor_parameter_751: name = "parameter_751" + original_name = "conv2d_12.b_0" shape = [128] dtype = "float32" min_val = float("-3.37719") @@ -8244,6 +8996,7 @@ class Program_weight_tensor_parameter_751: class Program_weight_tensor_parameter_752: name = "parameter_752" + original_name = "conv2d_12.w_0" shape = [128, 128, 4, 4] dtype = "float32" min_val = float("-0.258112") @@ -8255,6 +9008,7 @@ class Program_weight_tensor_parameter_752: class Program_weight_tensor_parameter_753: name = "parameter_753" + original_name = "linear_20.b_0" shape = [128] dtype = "float32" min_val = float("-2.06864") @@ -8266,6 +9020,7 @@ class Program_weight_tensor_parameter_753: class Program_weight_tensor_parameter_754: name = "parameter_754" + original_name = "linear_20.w_0" shape = [128, 128] dtype = "float32" min_val = float("-0.290333") @@ -8277,6 +9032,7 @@ class Program_weight_tensor_parameter_754: class Program_weight_tensor_parameter_755: name = "parameter_755" + original_name = "layer_norm_17.b_0" shape = [128] dtype = "float32" min_val = float("-0.636674") @@ -8288,6 +9044,7 @@ class Program_weight_tensor_parameter_755: class Program_weight_tensor_parameter_756: name = "parameter_756" + original_name = "layer_norm_17.w_0" shape = [128] dtype = "float32" min_val = float("0.21469") @@ -8299,6 +9056,7 @@ class Program_weight_tensor_parameter_756: class Program_weight_tensor_parameter_757: name = "parameter_757" + original_name = "linear_19.b_0" shape = [128] dtype = "float32" min_val = float("-0.601021") @@ -8310,6 +9068,7 @@ class Program_weight_tensor_parameter_757: class Program_weight_tensor_parameter_758: name = "parameter_758" + original_name = "linear_19.w_0" shape = [512, 128] dtype = "float32" min_val = float("-0.267446") @@ -8321,6 +9080,7 @@ class Program_weight_tensor_parameter_758: class Program_weight_tensor_parameter_759: name = "parameter_759" + original_name = "conv2d_11.b_0" shape = [512] dtype = "float32" min_val = float("-1.93136") @@ -8332,6 +9092,7 @@ class Program_weight_tensor_parameter_759: class Program_weight_tensor_parameter_760: name = "parameter_760" + original_name = "conv2d_11.w_0" shape = [512, 1, 3, 3] dtype = "float32" min_val = float("-0.325884") @@ -8343,6 +9104,7 @@ class Program_weight_tensor_parameter_760: class Program_weight_tensor_parameter_761: name = "parameter_761" + original_name = "linear_18.b_0" shape = [512] dtype = "float32" min_val = float("-2.58797") @@ -8354,6 +9116,7 @@ class Program_weight_tensor_parameter_761: class Program_weight_tensor_parameter_762: name = "parameter_762" + original_name = "linear_18.w_0" shape = [128, 512] dtype = "float32" min_val = float("-0.247873") @@ -8365,6 +9128,7 @@ class Program_weight_tensor_parameter_762: class Program_weight_tensor_parameter_763: name = "parameter_763" + original_name = "layer_norm_16.b_0" shape = [128] dtype = "float32" min_val = float("-1.04639") @@ -8376,6 +9140,7 @@ class Program_weight_tensor_parameter_763: class Program_weight_tensor_parameter_764: name = "parameter_764" + original_name = "layer_norm_16.w_0" shape = [128] dtype = "float32" min_val = float("2.22499") @@ -8387,6 +9152,7 @@ class Program_weight_tensor_parameter_764: class Program_weight_tensor_parameter_765: name = "parameter_765" + original_name = "linear_17.b_0" shape = [128] dtype = "float32" min_val = float("-0.354487") @@ -8398,6 +9164,7 @@ class Program_weight_tensor_parameter_765: class Program_weight_tensor_parameter_766: name = "parameter_766" + original_name = "linear_17.w_0" shape = [128, 128] dtype = "float32" min_val = float("-0.175427") @@ -8409,6 +9176,7 @@ class Program_weight_tensor_parameter_766: class Program_weight_tensor_parameter_767: name = "parameter_767" + original_name = "linear_16.b_0" shape = [256] dtype = "float32" min_val = float("-0.928573") @@ -8420,6 +9188,7 @@ class Program_weight_tensor_parameter_767: class Program_weight_tensor_parameter_768: name = "parameter_768" + original_name = "linear_16.w_0" shape = [128, 256] dtype = "float32" min_val = float("-0.342452") @@ -8431,6 +9200,7 @@ class Program_weight_tensor_parameter_768: class Program_weight_tensor_parameter_769: name = "parameter_769" + original_name = "layer_norm_15.b_0" shape = [128] dtype = "float32" min_val = float("-0.457146") @@ -8442,6 +9212,7 @@ class Program_weight_tensor_parameter_769: class Program_weight_tensor_parameter_770: name = "parameter_770" + original_name = "layer_norm_15.w_0" shape = [128] dtype = "float32" min_val = float("-0.0256019") @@ -8453,6 +9224,7 @@ class Program_weight_tensor_parameter_770: class Program_weight_tensor_parameter_771: name = "parameter_771" + original_name = "conv2d_10.b_0" shape = [128] dtype = "float32" min_val = float("-2.38085") @@ -8464,6 +9236,7 @@ class Program_weight_tensor_parameter_771: class Program_weight_tensor_parameter_772: name = "parameter_772" + original_name = "conv2d_10.w_0" shape = [128, 128, 4, 4] dtype = "float32" min_val = float("-0.25149") @@ -8475,6 +9248,7 @@ class Program_weight_tensor_parameter_772: class Program_weight_tensor_parameter_773: name = "parameter_773" + original_name = "linear_15.b_0" shape = [128] dtype = "float32" min_val = float("-3.01502") @@ -8486,6 +9260,7 @@ class Program_weight_tensor_parameter_773: class Program_weight_tensor_parameter_774: name = "parameter_774" + original_name = "linear_15.w_0" shape = [128, 128] dtype = "float32" min_val = float("-0.315197") @@ -8497,6 +9272,7 @@ class Program_weight_tensor_parameter_774: class Program_weight_tensor_parameter_775: name = "parameter_775" + original_name = "layer_norm_14.b_0" shape = [128] dtype = "float32" min_val = float("-0.564825") @@ -8508,6 +9284,7 @@ class Program_weight_tensor_parameter_775: class Program_weight_tensor_parameter_776: name = "parameter_776" + original_name = "layer_norm_14.w_0" shape = [128] dtype = "float32" min_val = float("0.274637") @@ -8519,6 +9296,7 @@ class Program_weight_tensor_parameter_776: class Program_weight_tensor_parameter_777: name = "parameter_777" + original_name = "layer_norm_1.b_0" shape = [128] dtype = "float32" min_val = float("-0.498051") @@ -8530,6 +9308,7 @@ class Program_weight_tensor_parameter_777: class Program_weight_tensor_parameter_778: name = "parameter_778" + original_name = "layer_norm_1.w_0" shape = [128] dtype = "float32" min_val = float("0.550205") @@ -8541,6 +9320,7 @@ class Program_weight_tensor_parameter_778: class Program_weight_tensor_parameter_779: name = "parameter_779" + original_name = "conv2d_1.b_0" shape = [128] dtype = "float32" min_val = float("-1.44897") @@ -8552,6 +9332,7 @@ class Program_weight_tensor_parameter_779: class Program_weight_tensor_parameter_780: name = "parameter_780" + original_name = "conv2d_1.w_0" shape = [128, 64, 3, 3] dtype = "float32" min_val = float("-0.280918") @@ -8563,6 +9344,7 @@ class Program_weight_tensor_parameter_780: class Program_weight_tensor_parameter_781: name = "parameter_781" + original_name = "layer_norm_13.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -8572,6 +9354,7 @@ class Program_weight_tensor_parameter_781: class Program_weight_tensor_parameter_782: name = "parameter_782" + original_name = "layer_norm_13.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -8581,6 +9364,7 @@ class Program_weight_tensor_parameter_782: class Program_weight_tensor_parameter_783: name = "parameter_783" + original_name = "linear_14.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -8590,6 +9374,7 @@ class Program_weight_tensor_parameter_783: class Program_weight_tensor_parameter_784: name = "parameter_784" + original_name = "linear_14.w_0" shape = [256, 64] dtype = "float32" min_val = float("-0.606199") @@ -8601,6 +9386,7 @@ class Program_weight_tensor_parameter_784: class Program_weight_tensor_parameter_785: name = "parameter_785" + original_name = "conv2d_9.b_0" shape = [256] dtype = "float32" min_val = float("-1.02246") @@ -8612,6 +9398,7 @@ class Program_weight_tensor_parameter_785: class Program_weight_tensor_parameter_786: name = "parameter_786" + original_name = "conv2d_9.w_0" shape = [256, 1, 3, 3] dtype = "float32" min_val = float("-0.301256") @@ -8623,6 +9410,7 @@ class Program_weight_tensor_parameter_786: class Program_weight_tensor_parameter_787: name = "parameter_787" + original_name = "linear_13.b_0" shape = [256] dtype = "float32" min_val = float("-4.40059") @@ -8634,6 +9422,7 @@ class Program_weight_tensor_parameter_787: class Program_weight_tensor_parameter_788: name = "parameter_788" + original_name = "linear_13.w_0" shape = [64, 256] dtype = "float32" min_val = float("-0.285198") @@ -8645,6 +9434,7 @@ class Program_weight_tensor_parameter_788: class Program_weight_tensor_parameter_789: name = "parameter_789" + original_name = "layer_norm_12.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -8654,6 +9444,7 @@ class Program_weight_tensor_parameter_789: class Program_weight_tensor_parameter_790: name = "parameter_790" + original_name = "layer_norm_12.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -8663,6 +9454,7 @@ class Program_weight_tensor_parameter_790: class Program_weight_tensor_parameter_791: name = "parameter_791" + original_name = "linear_12.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -8672,6 +9464,7 @@ class Program_weight_tensor_parameter_791: class Program_weight_tensor_parameter_792: name = "parameter_792" + original_name = "linear_12.w_0" shape = [64, 64] dtype = "float32" min_val = float("-0.514386") @@ -8683,6 +9476,7 @@ class Program_weight_tensor_parameter_792: class Program_weight_tensor_parameter_793: name = "parameter_793" + original_name = "linear_11.b_0" shape = [128] dtype = "float32" min_val = float("-1.85439") @@ -8694,6 +9488,7 @@ class Program_weight_tensor_parameter_793: class Program_weight_tensor_parameter_794: name = "parameter_794" + original_name = "linear_11.w_0" shape = [64, 128] dtype = "float32" min_val = float("-0.463118") @@ -8705,6 +9500,7 @@ class Program_weight_tensor_parameter_794: class Program_weight_tensor_parameter_795: name = "parameter_795" + original_name = "layer_norm_11.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -8714,6 +9510,7 @@ class Program_weight_tensor_parameter_795: class Program_weight_tensor_parameter_796: name = "parameter_796" + original_name = "layer_norm_11.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -8723,6 +9520,7 @@ class Program_weight_tensor_parameter_796: class Program_weight_tensor_parameter_797: name = "parameter_797" + original_name = "conv2d_8.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -8732,6 +9530,7 @@ class Program_weight_tensor_parameter_797: class Program_weight_tensor_parameter_798: name = "parameter_798" + original_name = "conv2d_8.w_0" shape = [64, 64, 8, 8] dtype = "float32" min_val = float("-0.395173") @@ -8743,6 +9542,7 @@ class Program_weight_tensor_parameter_798: class Program_weight_tensor_parameter_799: name = "parameter_799" + original_name = "linear_10.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -8752,6 +9552,7 @@ class Program_weight_tensor_parameter_799: class Program_weight_tensor_parameter_800: name = "parameter_800" + original_name = "linear_10.w_0" shape = [64, 64] dtype = "float32" min_val = float("-0.352266") @@ -8763,6 +9564,7 @@ class Program_weight_tensor_parameter_800: class Program_weight_tensor_parameter_801: name = "parameter_801" + original_name = "layer_norm_10.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -8772,6 +9574,7 @@ class Program_weight_tensor_parameter_801: class Program_weight_tensor_parameter_802: name = "parameter_802" + original_name = "layer_norm_10.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -8781,6 +9584,7 @@ class Program_weight_tensor_parameter_802: class Program_weight_tensor_parameter_803: name = "parameter_803" + original_name = "linear_9.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -8790,6 +9594,7 @@ class Program_weight_tensor_parameter_803: class Program_weight_tensor_parameter_804: name = "parameter_804" + original_name = "linear_9.w_0" shape = [256, 64] dtype = "float32" min_val = float("-0.284242") @@ -8801,6 +9606,7 @@ class Program_weight_tensor_parameter_804: class Program_weight_tensor_parameter_805: name = "parameter_805" + original_name = "conv2d_7.b_0" shape = [256] dtype = "float32" min_val = float("-1.53579") @@ -8812,6 +9618,7 @@ class Program_weight_tensor_parameter_805: class Program_weight_tensor_parameter_806: name = "parameter_806" + original_name = "conv2d_7.w_0" shape = [256, 1, 3, 3] dtype = "float32" min_val = float("-0.34041") @@ -8823,6 +9630,7 @@ class Program_weight_tensor_parameter_806: class Program_weight_tensor_parameter_807: name = "parameter_807" + original_name = "linear_8.b_0" shape = [256] dtype = "float32" min_val = float("-3.55806") @@ -8834,6 +9642,7 @@ class Program_weight_tensor_parameter_807: class Program_weight_tensor_parameter_808: name = "parameter_808" + original_name = "linear_8.w_0" shape = [64, 256] dtype = "float32" min_val = float("-0.396789") @@ -8845,6 +9654,7 @@ class Program_weight_tensor_parameter_808: class Program_weight_tensor_parameter_809: name = "parameter_809" + original_name = "layer_norm_9.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -8854,6 +9664,7 @@ class Program_weight_tensor_parameter_809: class Program_weight_tensor_parameter_810: name = "parameter_810" + original_name = "layer_norm_9.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -8863,6 +9674,7 @@ class Program_weight_tensor_parameter_810: class Program_weight_tensor_parameter_811: name = "parameter_811" + original_name = "linear_7.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -8872,6 +9684,7 @@ class Program_weight_tensor_parameter_811: class Program_weight_tensor_parameter_812: name = "parameter_812" + original_name = "linear_7.w_0" shape = [64, 64] dtype = "float32" min_val = float("-0.186179") @@ -8883,6 +9696,7 @@ class Program_weight_tensor_parameter_812: class Program_weight_tensor_parameter_813: name = "parameter_813" + original_name = "linear_6.b_0" shape = [128] dtype = "float32" min_val = float("-1.44905") @@ -8894,6 +9708,7 @@ class Program_weight_tensor_parameter_813: class Program_weight_tensor_parameter_814: name = "parameter_814" + original_name = "linear_6.w_0" shape = [64, 128] dtype = "float32" min_val = float("-0.460178") @@ -8905,6 +9720,7 @@ class Program_weight_tensor_parameter_814: class Program_weight_tensor_parameter_815: name = "parameter_815" + original_name = "layer_norm_8.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -8914,6 +9730,7 @@ class Program_weight_tensor_parameter_815: class Program_weight_tensor_parameter_816: name = "parameter_816" + original_name = "layer_norm_8.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -8923,6 +9740,7 @@ class Program_weight_tensor_parameter_816: class Program_weight_tensor_parameter_817: name = "parameter_817" + original_name = "conv2d_6.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -8932,6 +9750,7 @@ class Program_weight_tensor_parameter_817: class Program_weight_tensor_parameter_818: name = "parameter_818" + original_name = "conv2d_6.w_0" shape = [64, 64, 8, 8] dtype = "float32" min_val = float("-0.304926") @@ -8943,6 +9762,7 @@ class Program_weight_tensor_parameter_818: class Program_weight_tensor_parameter_819: name = "parameter_819" + original_name = "linear_5.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -8952,6 +9772,7 @@ class Program_weight_tensor_parameter_819: class Program_weight_tensor_parameter_820: name = "parameter_820" + original_name = "linear_5.w_0" shape = [64, 64] dtype = "float32" min_val = float("-0.412775") @@ -8963,6 +9784,7 @@ class Program_weight_tensor_parameter_820: class Program_weight_tensor_parameter_821: name = "parameter_821" + original_name = "layer_norm_7.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -8972,6 +9794,7 @@ class Program_weight_tensor_parameter_821: class Program_weight_tensor_parameter_822: name = "parameter_822" + original_name = "layer_norm_7.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -8981,6 +9804,7 @@ class Program_weight_tensor_parameter_822: class Program_weight_tensor_parameter_823: name = "parameter_823" + original_name = "linear_4.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -8990,6 +9814,7 @@ class Program_weight_tensor_parameter_823: class Program_weight_tensor_parameter_824: name = "parameter_824" + original_name = "linear_4.w_0" shape = [256, 64] dtype = "float32" min_val = float("-0.439705") @@ -9001,6 +9826,7 @@ class Program_weight_tensor_parameter_824: class Program_weight_tensor_parameter_825: name = "parameter_825" + original_name = "conv2d_5.b_0" shape = [256] dtype = "float32" min_val = float("-1.48609") @@ -9012,6 +9838,7 @@ class Program_weight_tensor_parameter_825: class Program_weight_tensor_parameter_826: name = "parameter_826" + original_name = "conv2d_5.w_0" shape = [256, 1, 3, 3] dtype = "float32" min_val = float("-0.485531") @@ -9023,6 +9850,7 @@ class Program_weight_tensor_parameter_826: class Program_weight_tensor_parameter_827: name = "parameter_827" + original_name = "linear_3.b_0" shape = [256] dtype = "float32" min_val = float("-4.91522") @@ -9034,6 +9862,7 @@ class Program_weight_tensor_parameter_827: class Program_weight_tensor_parameter_828: name = "parameter_828" + original_name = "linear_3.w_0" shape = [64, 256] dtype = "float32" min_val = float("-0.487375") @@ -9045,6 +9874,7 @@ class Program_weight_tensor_parameter_828: class Program_weight_tensor_parameter_829: name = "parameter_829" + original_name = "layer_norm_6.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -9054,6 +9884,7 @@ class Program_weight_tensor_parameter_829: class Program_weight_tensor_parameter_830: name = "parameter_830" + original_name = "layer_norm_6.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -9063,6 +9894,7 @@ class Program_weight_tensor_parameter_830: class Program_weight_tensor_parameter_831: name = "parameter_831" + original_name = "linear_2.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -9072,6 +9904,7 @@ class Program_weight_tensor_parameter_831: class Program_weight_tensor_parameter_832: name = "parameter_832" + original_name = "linear_2.w_0" shape = [64, 64] dtype = "float32" min_val = float("-0.269843") @@ -9083,6 +9916,7 @@ class Program_weight_tensor_parameter_832: class Program_weight_tensor_parameter_833: name = "parameter_833" + original_name = "linear_1.b_0" shape = [128] dtype = "float32" min_val = float("-1.39375") @@ -9094,6 +9928,7 @@ class Program_weight_tensor_parameter_833: class Program_weight_tensor_parameter_834: name = "parameter_834" + original_name = "linear_1.w_0" shape = [64, 128] dtype = "float32" min_val = float("-0.502646") @@ -9105,6 +9940,7 @@ class Program_weight_tensor_parameter_834: class Program_weight_tensor_parameter_835: name = "parameter_835" + original_name = "layer_norm_5.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -9114,6 +9950,7 @@ class Program_weight_tensor_parameter_835: class Program_weight_tensor_parameter_836: name = "parameter_836" + original_name = "layer_norm_5.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -9123,6 +9960,7 @@ class Program_weight_tensor_parameter_836: class Program_weight_tensor_parameter_837: name = "parameter_837" + original_name = "conv2d_4.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -9132,6 +9970,7 @@ class Program_weight_tensor_parameter_837: class Program_weight_tensor_parameter_838: name = "parameter_838" + original_name = "conv2d_4.w_0" shape = [64, 64, 8, 8] dtype = "float32" min_val = float("-0.601951") @@ -9143,6 +9982,7 @@ class Program_weight_tensor_parameter_838: class Program_weight_tensor_parameter_839: name = "parameter_839" + original_name = "linear_0.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -9152,6 +9992,7 @@ class Program_weight_tensor_parameter_839: class Program_weight_tensor_parameter_840: name = "parameter_840" + original_name = "linear_0.w_0" shape = [64, 64] dtype = "float32" min_val = float("-0.460391") @@ -9163,6 +10004,7 @@ class Program_weight_tensor_parameter_840: class Program_weight_tensor_parameter_841: name = "parameter_841" + original_name = "layer_norm_4.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -9172,6 +10014,7 @@ class Program_weight_tensor_parameter_841: class Program_weight_tensor_parameter_842: name = "parameter_842" + original_name = "layer_norm_4.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -9181,6 +10024,7 @@ class Program_weight_tensor_parameter_842: class Program_weight_tensor_parameter_843: name = "parameter_843" + original_name = "layer_norm_0.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -9190,6 +10034,7 @@ class Program_weight_tensor_parameter_843: class Program_weight_tensor_parameter_844: name = "parameter_844" + original_name = "layer_norm_0.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -9199,6 +10044,7 @@ class Program_weight_tensor_parameter_844: class Program_weight_tensor_parameter_845: name = "parameter_845" + original_name = "conv2d_0.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -9208,6 +10054,7 @@ class Program_weight_tensor_parameter_845: class Program_weight_tensor_parameter_846: name = "parameter_846" + original_name = "conv2d_0.w_0" shape = [64, 3, 7, 7] dtype = "float32" min_val = float("-0.383338") diff --git a/paddle_samples/PaddleX/SegFormer-B5/subgraph_1/input_meta.py b/paddle_samples/PaddleX/SegFormer-B5/subgraph_1/input_meta.py index 30797fdd8..eb34deb58 100644 --- a/paddle_samples/PaddleX/SegFormer-B5/subgraph_1/input_meta.py +++ b/paddle_samples/PaddleX/SegFormer-B5/subgraph_1/input_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_data_0: name = "data_0" + original_name = "var_0" shape = [2, 3, 512, 1024] dtype = "float32" min_val = float("-1.0") diff --git a/paddle_samples/PaddleX/SegFormer-B5/subgraph_1/weight_meta.py b/paddle_samples/PaddleX/SegFormer-B5/subgraph_1/weight_meta.py index b341753d6..edecf49c6 100644 --- a/paddle_samples/PaddleX/SegFormer-B5/subgraph_1/weight_meta.py +++ b/paddle_samples/PaddleX/SegFormer-B5/subgraph_1/weight_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_parameter_0: name = "parameter_0" + original_name = "conv2d_106.b_0" shape = [2] dtype = "float32" min_val = float("0") @@ -9,6 +10,7 @@ class Program_weight_tensor_parameter_0: class Program_weight_tensor_parameter_1: name = "parameter_1" + original_name = "conv2d_106.w_0" shape = [2, 768, 1, 1] dtype = "float32" min_val = float("-0.166187") @@ -20,6 +22,7 @@ class Program_weight_tensor_parameter_1: class Program_weight_tensor_parameter_2: name = "parameter_2" + original_name = "batch_norm2d_0.b_0" shape = [768] dtype = "float32" data = None @@ -27,6 +30,7 @@ class Program_weight_tensor_parameter_2: class Program_weight_tensor_parameter_3: name = "parameter_3" + original_name = "batch_norm2d_0.w_0" shape = [768] dtype = "float32" min_val = float("1.0") @@ -37,6 +41,7 @@ class Program_weight_tensor_parameter_3: class Program_weight_tensor_parameter_4: name = "parameter_4" + original_name = "batch_norm2d_0.w_2" shape = [768] dtype = "float32" min_val = float("1.0") @@ -47,6 +52,7 @@ class Program_weight_tensor_parameter_4: class Program_weight_tensor_parameter_5: name = "parameter_5" + original_name = "batch_norm2d_0.w_1" shape = [768] dtype = "float32" data = None @@ -54,6 +60,7 @@ class Program_weight_tensor_parameter_5: class Program_weight_tensor_parameter_6: name = "parameter_6" + original_name = "conv2d_105.w_0" shape = [768, 3072, 1, 1] dtype = "float32" min_val = float("-0.124333") @@ -65,6 +72,7 @@ class Program_weight_tensor_parameter_6: class Program_weight_tensor_parameter_7: name = "parameter_7" + original_name = "linear_263.b_0" shape = [768] dtype = "float32" data = None @@ -72,6 +80,7 @@ class Program_weight_tensor_parameter_7: class Program_weight_tensor_parameter_8: name = "parameter_8" + original_name = "linear_263.w_0" shape = [64, 768] dtype = "float32" min_val = float("-0.0849185") @@ -83,6 +92,7 @@ class Program_weight_tensor_parameter_8: class Program_weight_tensor_parameter_9: name = "parameter_9" + original_name = "linear_262.b_0" shape = [768] dtype = "float32" data = None @@ -90,6 +100,7 @@ class Program_weight_tensor_parameter_9: class Program_weight_tensor_parameter_10: name = "parameter_10" + original_name = "linear_262.w_0" shape = [128, 768] dtype = "float32" min_val = float("-0.0818316") @@ -101,6 +112,7 @@ class Program_weight_tensor_parameter_10: class Program_weight_tensor_parameter_11: name = "parameter_11" + original_name = "linear_261.b_0" shape = [768] dtype = "float32" data = None @@ -108,6 +120,7 @@ class Program_weight_tensor_parameter_11: class Program_weight_tensor_parameter_12: name = "parameter_12" + original_name = "linear_261.w_0" shape = [320, 768] dtype = "float32" min_val = float("-0.0742609") @@ -119,6 +132,7 @@ class Program_weight_tensor_parameter_12: class Program_weight_tensor_parameter_13: name = "parameter_13" + original_name = "linear_260.b_0" shape = [768] dtype = "float32" data = None @@ -126,6 +140,7 @@ class Program_weight_tensor_parameter_13: class Program_weight_tensor_parameter_14: name = "parameter_14" + original_name = "linear_260.w_0" shape = [512, 768] dtype = "float32" min_val = float("-0.068465") @@ -137,6 +152,7 @@ class Program_weight_tensor_parameter_14: class Program_weight_tensor_parameter_15: name = "parameter_15" + original_name = "layer_norm_160.b_0" shape = [512] dtype = "float32" min_val = float("-0.357887") @@ -148,6 +164,7 @@ class Program_weight_tensor_parameter_15: class Program_weight_tensor_parameter_16: name = "parameter_16" + original_name = "layer_norm_160.w_0" shape = [512] dtype = "float32" min_val = float("0.0010232") @@ -159,6 +176,7 @@ class Program_weight_tensor_parameter_16: class Program_weight_tensor_parameter_17: name = "parameter_17" + original_name = "linear_259.b_0" shape = [512] dtype = "float32" min_val = float("-4.7026") @@ -170,6 +188,7 @@ class Program_weight_tensor_parameter_17: class Program_weight_tensor_parameter_18: name = "parameter_18" + original_name = "linear_259.w_0" shape = [2048, 512] dtype = "float32" min_val = float("-1.90506") @@ -181,6 +200,7 @@ class Program_weight_tensor_parameter_18: class Program_weight_tensor_parameter_19: name = "parameter_19" + original_name = "conv2d_104.b_0" shape = [2048] dtype = "float32" min_val = float("-1.65574") @@ -192,6 +212,7 @@ class Program_weight_tensor_parameter_19: class Program_weight_tensor_parameter_20: name = "parameter_20" + original_name = "conv2d_104.w_0" shape = [2048, 1, 3, 3] dtype = "float32" min_val = float("-0.254338") @@ -203,6 +224,7 @@ class Program_weight_tensor_parameter_20: class Program_weight_tensor_parameter_21: name = "parameter_21" + original_name = "linear_258.b_0" shape = [2048] dtype = "float32" min_val = float("-2.71424") @@ -214,6 +236,7 @@ class Program_weight_tensor_parameter_21: class Program_weight_tensor_parameter_22: name = "parameter_22" + original_name = "linear_258.w_0" shape = [512, 2048] dtype = "float32" min_val = float("-0.897064") @@ -225,6 +248,7 @@ class Program_weight_tensor_parameter_22: class Program_weight_tensor_parameter_23: name = "parameter_23" + original_name = "layer_norm_159.b_0" shape = [512] dtype = "float32" min_val = float("-1.5178") @@ -236,6 +260,7 @@ class Program_weight_tensor_parameter_23: class Program_weight_tensor_parameter_24: name = "parameter_24" + original_name = "layer_norm_159.w_0" shape = [512] dtype = "float32" min_val = float("0.28629") @@ -247,6 +272,7 @@ class Program_weight_tensor_parameter_24: class Program_weight_tensor_parameter_25: name = "parameter_25" + original_name = "linear_257.b_0" shape = [512] dtype = "float32" min_val = float("-5.02836") @@ -258,6 +284,7 @@ class Program_weight_tensor_parameter_25: class Program_weight_tensor_parameter_26: name = "parameter_26" + original_name = "linear_257.w_0" shape = [512, 512] dtype = "float32" min_val = float("-0.809571") @@ -269,6 +296,7 @@ class Program_weight_tensor_parameter_26: class Program_weight_tensor_parameter_27: name = "parameter_27" + original_name = "linear_256.b_0" shape = [1024] dtype = "float32" min_val = float("-8.34825") @@ -280,6 +308,7 @@ class Program_weight_tensor_parameter_27: class Program_weight_tensor_parameter_28: name = "parameter_28" + original_name = "linear_256.w_0" shape = [512, 1024] dtype = "float32" min_val = float("-0.314331") @@ -291,6 +320,7 @@ class Program_weight_tensor_parameter_28: class Program_weight_tensor_parameter_29: name = "parameter_29" + original_name = "linear_255.b_0" shape = [512] dtype = "float32" min_val = float("-2.05464") @@ -302,6 +332,7 @@ class Program_weight_tensor_parameter_29: class Program_weight_tensor_parameter_30: name = "parameter_30" + original_name = "linear_255.w_0" shape = [512, 512] dtype = "float32" min_val = float("-0.298439") @@ -313,6 +344,7 @@ class Program_weight_tensor_parameter_30: class Program_weight_tensor_parameter_31: name = "parameter_31" + original_name = "layer_norm_158.b_0" shape = [512] dtype = "float32" min_val = float("-0.128896") @@ -324,6 +356,7 @@ class Program_weight_tensor_parameter_31: class Program_weight_tensor_parameter_32: name = "parameter_32" + original_name = "layer_norm_158.w_0" shape = [512] dtype = "float32" min_val = float("2.42261e-05") @@ -335,6 +368,7 @@ class Program_weight_tensor_parameter_32: class Program_weight_tensor_parameter_33: name = "parameter_33" + original_name = "linear_254.b_0" shape = [512] dtype = "float32" min_val = float("-7.63878") @@ -346,6 +380,7 @@ class Program_weight_tensor_parameter_33: class Program_weight_tensor_parameter_34: name = "parameter_34" + original_name = "linear_254.w_0" shape = [2048, 512] dtype = "float32" min_val = float("-2.59057") @@ -357,6 +392,7 @@ class Program_weight_tensor_parameter_34: class Program_weight_tensor_parameter_35: name = "parameter_35" + original_name = "conv2d_103.b_0" shape = [2048] dtype = "float32" min_val = float("-0.925887") @@ -368,6 +404,7 @@ class Program_weight_tensor_parameter_35: class Program_weight_tensor_parameter_36: name = "parameter_36" + original_name = "conv2d_103.w_0" shape = [2048, 1, 3, 3] dtype = "float32" min_val = float("-0.652777") @@ -379,6 +416,7 @@ class Program_weight_tensor_parameter_36: class Program_weight_tensor_parameter_37: name = "parameter_37" + original_name = "linear_253.b_0" shape = [2048] dtype = "float32" min_val = float("-3.03078") @@ -390,6 +428,7 @@ class Program_weight_tensor_parameter_37: class Program_weight_tensor_parameter_38: name = "parameter_38" + original_name = "linear_253.w_0" shape = [512, 2048] dtype = "float32" min_val = float("-0.578776") @@ -401,6 +440,7 @@ class Program_weight_tensor_parameter_38: class Program_weight_tensor_parameter_39: name = "parameter_39" + original_name = "layer_norm_157.b_0" shape = [512] dtype = "float32" min_val = float("-2.89124") @@ -412,6 +452,7 @@ class Program_weight_tensor_parameter_39: class Program_weight_tensor_parameter_40: name = "parameter_40" + original_name = "layer_norm_157.w_0" shape = [512] dtype = "float32" min_val = float("0.259754") @@ -423,6 +464,7 @@ class Program_weight_tensor_parameter_40: class Program_weight_tensor_parameter_41: name = "parameter_41" + original_name = "linear_252.b_0" shape = [512] dtype = "float32" min_val = float("-5.19747") @@ -434,6 +476,7 @@ class Program_weight_tensor_parameter_41: class Program_weight_tensor_parameter_42: name = "parameter_42" + original_name = "linear_252.w_0" shape = [512, 512] dtype = "float32" min_val = float("-0.742721") @@ -445,6 +488,7 @@ class Program_weight_tensor_parameter_42: class Program_weight_tensor_parameter_43: name = "parameter_43" + original_name = "linear_251.b_0" shape = [1024] dtype = "float32" min_val = float("-3.3482") @@ -456,6 +500,7 @@ class Program_weight_tensor_parameter_43: class Program_weight_tensor_parameter_44: name = "parameter_44" + original_name = "linear_251.w_0" shape = [512, 1024] dtype = "float32" min_val = float("-0.353797") @@ -467,6 +512,7 @@ class Program_weight_tensor_parameter_44: class Program_weight_tensor_parameter_45: name = "parameter_45" + original_name = "linear_250.b_0" shape = [512] dtype = "float32" min_val = float("-2.82219") @@ -478,6 +524,7 @@ class Program_weight_tensor_parameter_45: class Program_weight_tensor_parameter_46: name = "parameter_46" + original_name = "linear_250.w_0" shape = [512, 512] dtype = "float32" min_val = float("-0.322587") @@ -489,6 +536,7 @@ class Program_weight_tensor_parameter_46: class Program_weight_tensor_parameter_47: name = "parameter_47" + original_name = "layer_norm_156.b_0" shape = [512] dtype = "float32" min_val = float("-0.167091") @@ -500,6 +548,7 @@ class Program_weight_tensor_parameter_47: class Program_weight_tensor_parameter_48: name = "parameter_48" + original_name = "layer_norm_156.w_0" shape = [512] dtype = "float32" min_val = float("-0.149601") @@ -511,6 +560,7 @@ class Program_weight_tensor_parameter_48: class Program_weight_tensor_parameter_49: name = "parameter_49" + original_name = "linear_249.b_0" shape = [512] dtype = "float32" min_val = float("-6.2909") @@ -522,6 +572,7 @@ class Program_weight_tensor_parameter_49: class Program_weight_tensor_parameter_50: name = "parameter_50" + original_name = "linear_249.w_0" shape = [2048, 512] dtype = "float32" min_val = float("-1.98775") @@ -533,6 +584,7 @@ class Program_weight_tensor_parameter_50: class Program_weight_tensor_parameter_51: name = "parameter_51" + original_name = "conv2d_102.b_0" shape = [2048] dtype = "float32" min_val = float("-1.33575") @@ -544,6 +596,7 @@ class Program_weight_tensor_parameter_51: class Program_weight_tensor_parameter_52: name = "parameter_52" + original_name = "conv2d_102.w_0" shape = [2048, 1, 3, 3] dtype = "float32" min_val = float("-0.451967") @@ -555,6 +608,7 @@ class Program_weight_tensor_parameter_52: class Program_weight_tensor_parameter_53: name = "parameter_53" + original_name = "linear_248.b_0" shape = [2048] dtype = "float32" min_val = float("-10.6385") @@ -566,6 +620,7 @@ class Program_weight_tensor_parameter_53: class Program_weight_tensor_parameter_54: name = "parameter_54" + original_name = "linear_248.w_0" shape = [512, 2048] dtype = "float32" min_val = float("-0.503705") @@ -577,6 +632,7 @@ class Program_weight_tensor_parameter_54: class Program_weight_tensor_parameter_55: name = "parameter_55" + original_name = "layer_norm_155.b_0" shape = [512] dtype = "float32" min_val = float("-2.7108") @@ -588,6 +644,7 @@ class Program_weight_tensor_parameter_55: class Program_weight_tensor_parameter_56: name = "parameter_56" + original_name = "layer_norm_155.w_0" shape = [512] dtype = "float32" min_val = float("0.00418757") @@ -599,6 +656,7 @@ class Program_weight_tensor_parameter_56: class Program_weight_tensor_parameter_57: name = "parameter_57" + original_name = "linear_247.b_0" shape = [512] dtype = "float32" min_val = float("-2.66744") @@ -610,6 +668,7 @@ class Program_weight_tensor_parameter_57: class Program_weight_tensor_parameter_58: name = "parameter_58" + original_name = "linear_247.w_0" shape = [512, 512] dtype = "float32" min_val = float("-0.662249") @@ -621,6 +680,7 @@ class Program_weight_tensor_parameter_58: class Program_weight_tensor_parameter_59: name = "parameter_59" + original_name = "linear_246.b_0" shape = [1024] dtype = "float32" min_val = float("-3.89636") @@ -632,6 +692,7 @@ class Program_weight_tensor_parameter_59: class Program_weight_tensor_parameter_60: name = "parameter_60" + original_name = "linear_246.w_0" shape = [512, 1024] dtype = "float32" min_val = float("-0.308867") @@ -643,6 +704,7 @@ class Program_weight_tensor_parameter_60: class Program_weight_tensor_parameter_61: name = "parameter_61" + original_name = "linear_245.b_0" shape = [512] dtype = "float32" min_val = float("-3.24896") @@ -654,6 +716,7 @@ class Program_weight_tensor_parameter_61: class Program_weight_tensor_parameter_62: name = "parameter_62" + original_name = "linear_245.w_0" shape = [512, 512] dtype = "float32" min_val = float("-0.338396") @@ -665,6 +728,7 @@ class Program_weight_tensor_parameter_62: class Program_weight_tensor_parameter_63: name = "parameter_63" + original_name = "layer_norm_154.b_0" shape = [512] dtype = "float32" min_val = float("-0.139643") @@ -676,6 +740,7 @@ class Program_weight_tensor_parameter_63: class Program_weight_tensor_parameter_64: name = "parameter_64" + original_name = "layer_norm_154.w_0" shape = [512] dtype = "float32" min_val = float("-0.0010015") @@ -687,6 +752,7 @@ class Program_weight_tensor_parameter_64: class Program_weight_tensor_parameter_65: name = "parameter_65" + original_name = "layer_norm_3.b_0" shape = [512] dtype = "float32" min_val = float("-6.43594") @@ -698,6 +764,7 @@ class Program_weight_tensor_parameter_65: class Program_weight_tensor_parameter_66: name = "parameter_66" + original_name = "layer_norm_3.w_0" shape = [512] dtype = "float32" min_val = float("0.0519705") @@ -709,6 +776,7 @@ class Program_weight_tensor_parameter_66: class Program_weight_tensor_parameter_67: name = "parameter_67" + original_name = "conv2d_3.b_0" shape = [512] dtype = "float32" min_val = float("-6.87845") @@ -720,6 +788,7 @@ class Program_weight_tensor_parameter_67: class Program_weight_tensor_parameter_68: name = "parameter_68" + original_name = "conv2d_3.w_0" shape = [512, 320, 3, 3] dtype = "float32" min_val = float("-1.28306") @@ -731,6 +800,7 @@ class Program_weight_tensor_parameter_68: class Program_weight_tensor_parameter_69: name = "parameter_69" + original_name = "layer_norm_153.b_0" shape = [320] dtype = "float32" min_val = float("-1.59197") @@ -742,6 +812,7 @@ class Program_weight_tensor_parameter_69: class Program_weight_tensor_parameter_70: name = "parameter_70" + original_name = "layer_norm_153.w_0" shape = [320] dtype = "float32" min_val = float("0.0904737") @@ -753,6 +824,7 @@ class Program_weight_tensor_parameter_70: class Program_weight_tensor_parameter_71: name = "parameter_71" + original_name = "linear_244.b_0" shape = [320] dtype = "float32" min_val = float("-0.62013") @@ -764,6 +836,7 @@ class Program_weight_tensor_parameter_71: class Program_weight_tensor_parameter_72: name = "parameter_72" + original_name = "linear_244.w_0" shape = [1280, 320] dtype = "float32" min_val = float("-1.27726") @@ -775,6 +848,7 @@ class Program_weight_tensor_parameter_72: class Program_weight_tensor_parameter_73: name = "parameter_73" + original_name = "conv2d_101.b_0" shape = [1280] dtype = "float32" min_val = float("-1.82204") @@ -786,6 +860,7 @@ class Program_weight_tensor_parameter_73: class Program_weight_tensor_parameter_74: name = "parameter_74" + original_name = "conv2d_101.w_0" shape = [1280, 1, 3, 3] dtype = "float32" min_val = float("-0.614582") @@ -797,6 +872,7 @@ class Program_weight_tensor_parameter_74: class Program_weight_tensor_parameter_75: name = "parameter_75" + original_name = "linear_243.b_0" shape = [1280] dtype = "float32" min_val = float("-2.13715") @@ -808,6 +884,7 @@ class Program_weight_tensor_parameter_75: class Program_weight_tensor_parameter_76: name = "parameter_76" + original_name = "linear_243.w_0" shape = [320, 1280] dtype = "float32" min_val = float("-0.364689") @@ -819,6 +896,7 @@ class Program_weight_tensor_parameter_76: class Program_weight_tensor_parameter_77: name = "parameter_77" + original_name = "layer_norm_152.b_0" shape = [320] dtype = "float32" min_val = float("-1.89406") @@ -830,6 +908,7 @@ class Program_weight_tensor_parameter_77: class Program_weight_tensor_parameter_78: name = "parameter_78" + original_name = "layer_norm_152.w_0" shape = [320] dtype = "float32" min_val = float("-0.0168493") @@ -841,6 +920,7 @@ class Program_weight_tensor_parameter_78: class Program_weight_tensor_parameter_79: name = "parameter_79" + original_name = "linear_242.b_0" shape = [320] dtype = "float32" min_val = float("-0.434925") @@ -852,6 +932,7 @@ class Program_weight_tensor_parameter_79: class Program_weight_tensor_parameter_80: name = "parameter_80" + original_name = "linear_242.w_0" shape = [320, 320] dtype = "float32" min_val = float("-1.61625") @@ -863,6 +944,7 @@ class Program_weight_tensor_parameter_80: class Program_weight_tensor_parameter_81: name = "parameter_81" + original_name = "linear_241.b_0" shape = [640] dtype = "float32" min_val = float("-12.6986") @@ -874,6 +956,7 @@ class Program_weight_tensor_parameter_81: class Program_weight_tensor_parameter_82: name = "parameter_82" + original_name = "linear_241.w_0" shape = [320, 640] dtype = "float32" min_val = float("-0.338528") @@ -885,6 +968,7 @@ class Program_weight_tensor_parameter_82: class Program_weight_tensor_parameter_83: name = "parameter_83" + original_name = "layer_norm_151.b_0" shape = [320] dtype = "float32" min_val = float("-0.674946") @@ -896,6 +980,7 @@ class Program_weight_tensor_parameter_83: class Program_weight_tensor_parameter_84: name = "parameter_84" + original_name = "layer_norm_151.w_0" shape = [320] dtype = "float32" min_val = float("-0.107023") @@ -907,6 +992,7 @@ class Program_weight_tensor_parameter_84: class Program_weight_tensor_parameter_85: name = "parameter_85" + original_name = "conv2d_100.b_0" shape = [320] dtype = "float32" min_val = float("-2.97914") @@ -918,6 +1004,7 @@ class Program_weight_tensor_parameter_85: class Program_weight_tensor_parameter_86: name = "parameter_86" + original_name = "conv2d_100.w_0" shape = [320, 320, 2, 2] dtype = "float32" min_val = float("-0.3376") @@ -929,6 +1016,7 @@ class Program_weight_tensor_parameter_86: class Program_weight_tensor_parameter_87: name = "parameter_87" + original_name = "linear_240.b_0" shape = [320] dtype = "float32" min_val = float("-1.57554") @@ -940,6 +1028,7 @@ class Program_weight_tensor_parameter_87: class Program_weight_tensor_parameter_88: name = "parameter_88" + original_name = "linear_240.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.24823") @@ -951,6 +1040,7 @@ class Program_weight_tensor_parameter_88: class Program_weight_tensor_parameter_89: name = "parameter_89" + original_name = "layer_norm_150.b_0" shape = [320] dtype = "float32" min_val = float("-1.89752") @@ -962,6 +1052,7 @@ class Program_weight_tensor_parameter_89: class Program_weight_tensor_parameter_90: name = "parameter_90" + original_name = "layer_norm_150.w_0" shape = [320] dtype = "float32" min_val = float("-0.336237") @@ -973,6 +1064,7 @@ class Program_weight_tensor_parameter_90: class Program_weight_tensor_parameter_91: name = "parameter_91" + original_name = "linear_239.b_0" shape = [320] dtype = "float32" min_val = float("-0.298894") @@ -984,6 +1076,7 @@ class Program_weight_tensor_parameter_91: class Program_weight_tensor_parameter_92: name = "parameter_92" + original_name = "linear_239.w_0" shape = [1280, 320] dtype = "float32" min_val = float("-0.980333") @@ -995,6 +1088,7 @@ class Program_weight_tensor_parameter_92: class Program_weight_tensor_parameter_93: name = "parameter_93" + original_name = "conv2d_99.b_0" shape = [1280] dtype = "float32" min_val = float("-1.12664") @@ -1006,6 +1100,7 @@ class Program_weight_tensor_parameter_93: class Program_weight_tensor_parameter_94: name = "parameter_94" + original_name = "conv2d_99.w_0" shape = [1280, 1, 3, 3] dtype = "float32" min_val = float("-0.26769") @@ -1017,6 +1112,7 @@ class Program_weight_tensor_parameter_94: class Program_weight_tensor_parameter_95: name = "parameter_95" + original_name = "linear_238.b_0" shape = [1280] dtype = "float32" min_val = float("-1.76545") @@ -1028,6 +1124,7 @@ class Program_weight_tensor_parameter_95: class Program_weight_tensor_parameter_96: name = "parameter_96" + original_name = "linear_238.w_0" shape = [320, 1280] dtype = "float32" min_val = float("-0.583235") @@ -1039,6 +1136,7 @@ class Program_weight_tensor_parameter_96: class Program_weight_tensor_parameter_97: name = "parameter_97" + original_name = "layer_norm_149.b_0" shape = [320] dtype = "float32" min_val = float("-1.58146") @@ -1050,6 +1148,7 @@ class Program_weight_tensor_parameter_97: class Program_weight_tensor_parameter_98: name = "parameter_98" + original_name = "layer_norm_149.w_0" shape = [320] dtype = "float32" min_val = float("-0.00389179") @@ -1061,6 +1160,7 @@ class Program_weight_tensor_parameter_98: class Program_weight_tensor_parameter_99: name = "parameter_99" + original_name = "linear_237.b_0" shape = [320] dtype = "float32" min_val = float("-0.416542") @@ -1072,6 +1172,7 @@ class Program_weight_tensor_parameter_99: class Program_weight_tensor_parameter_100: name = "parameter_100" + original_name = "linear_237.w_0" shape = [320, 320] dtype = "float32" min_val = float("-1.21825") @@ -1083,6 +1184,7 @@ class Program_weight_tensor_parameter_100: class Program_weight_tensor_parameter_101: name = "parameter_101" + original_name = "linear_236.b_0" shape = [640] dtype = "float32" min_val = float("-10.283") @@ -1094,6 +1196,7 @@ class Program_weight_tensor_parameter_101: class Program_weight_tensor_parameter_102: name = "parameter_102" + original_name = "linear_236.w_0" shape = [320, 640] dtype = "float32" min_val = float("-0.456232") @@ -1105,6 +1208,7 @@ class Program_weight_tensor_parameter_102: class Program_weight_tensor_parameter_103: name = "parameter_103" + original_name = "layer_norm_148.b_0" shape = [320] dtype = "float32" min_val = float("-0.67427") @@ -1116,6 +1220,7 @@ class Program_weight_tensor_parameter_103: class Program_weight_tensor_parameter_104: name = "parameter_104" + original_name = "layer_norm_148.w_0" shape = [320] dtype = "float32" min_val = float("-0.0318242") @@ -1127,6 +1232,7 @@ class Program_weight_tensor_parameter_104: class Program_weight_tensor_parameter_105: name = "parameter_105" + original_name = "conv2d_98.b_0" shape = [320] dtype = "float32" min_val = float("-3.41759") @@ -1138,6 +1244,7 @@ class Program_weight_tensor_parameter_105: class Program_weight_tensor_parameter_106: name = "parameter_106" + original_name = "conv2d_98.w_0" shape = [320, 320, 2, 2] dtype = "float32" min_val = float("-0.521714") @@ -1149,6 +1256,7 @@ class Program_weight_tensor_parameter_106: class Program_weight_tensor_parameter_107: name = "parameter_107" + original_name = "linear_235.b_0" shape = [320] dtype = "float32" min_val = float("-1.80256") @@ -1160,6 +1268,7 @@ class Program_weight_tensor_parameter_107: class Program_weight_tensor_parameter_108: name = "parameter_108" + original_name = "linear_235.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.283359") @@ -1171,6 +1280,7 @@ class Program_weight_tensor_parameter_108: class Program_weight_tensor_parameter_109: name = "parameter_109" + original_name = "layer_norm_147.b_0" shape = [320] dtype = "float32" min_val = float("-2.03699") @@ -1182,6 +1292,7 @@ class Program_weight_tensor_parameter_109: class Program_weight_tensor_parameter_110: name = "parameter_110" + original_name = "layer_norm_147.w_0" shape = [320] dtype = "float32" min_val = float("0.178573") @@ -1193,6 +1304,7 @@ class Program_weight_tensor_parameter_110: class Program_weight_tensor_parameter_111: name = "parameter_111" + original_name = "linear_234.b_0" shape = [320] dtype = "float32" min_val = float("-0.336918") @@ -1204,6 +1316,7 @@ class Program_weight_tensor_parameter_111: class Program_weight_tensor_parameter_112: name = "parameter_112" + original_name = "linear_234.w_0" shape = [1280, 320] dtype = "float32" min_val = float("-0.897825") @@ -1215,6 +1328,7 @@ class Program_weight_tensor_parameter_112: class Program_weight_tensor_parameter_113: name = "parameter_113" + original_name = "conv2d_97.b_0" shape = [1280] dtype = "float32" min_val = float("-1.98512") @@ -1226,6 +1340,7 @@ class Program_weight_tensor_parameter_113: class Program_weight_tensor_parameter_114: name = "parameter_114" + original_name = "conv2d_97.w_0" shape = [1280, 1, 3, 3] dtype = "float32" min_val = float("-0.327017") @@ -1237,6 +1352,7 @@ class Program_weight_tensor_parameter_114: class Program_weight_tensor_parameter_115: name = "parameter_115" + original_name = "linear_233.b_0" shape = [1280] dtype = "float32" min_val = float("-2.22735") @@ -1248,6 +1364,7 @@ class Program_weight_tensor_parameter_115: class Program_weight_tensor_parameter_116: name = "parameter_116" + original_name = "linear_233.w_0" shape = [320, 1280] dtype = "float32" min_val = float("-0.355309") @@ -1259,6 +1376,7 @@ class Program_weight_tensor_parameter_116: class Program_weight_tensor_parameter_117: name = "parameter_117" + original_name = "layer_norm_146.b_0" shape = [320] dtype = "float32" min_val = float("-1.54479") @@ -1270,6 +1388,7 @@ class Program_weight_tensor_parameter_117: class Program_weight_tensor_parameter_118: name = "parameter_118" + original_name = "layer_norm_146.w_0" shape = [320] dtype = "float32" min_val = float("0.0381999") @@ -1281,6 +1400,7 @@ class Program_weight_tensor_parameter_118: class Program_weight_tensor_parameter_119: name = "parameter_119" + original_name = "linear_232.b_0" shape = [320] dtype = "float32" min_val = float("-0.406395") @@ -1292,6 +1412,7 @@ class Program_weight_tensor_parameter_119: class Program_weight_tensor_parameter_120: name = "parameter_120" + original_name = "linear_232.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.911184") @@ -1303,6 +1424,7 @@ class Program_weight_tensor_parameter_120: class Program_weight_tensor_parameter_121: name = "parameter_121" + original_name = "linear_231.b_0" shape = [640] dtype = "float32" min_val = float("-15.9853") @@ -1314,6 +1436,7 @@ class Program_weight_tensor_parameter_121: class Program_weight_tensor_parameter_122: name = "parameter_122" + original_name = "linear_231.w_0" shape = [320, 640] dtype = "float32" min_val = float("-0.396519") @@ -1325,6 +1448,7 @@ class Program_weight_tensor_parameter_122: class Program_weight_tensor_parameter_123: name = "parameter_123" + original_name = "layer_norm_145.b_0" shape = [320] dtype = "float32" min_val = float("-0.718329") @@ -1336,6 +1460,7 @@ class Program_weight_tensor_parameter_123: class Program_weight_tensor_parameter_124: name = "parameter_124" + original_name = "layer_norm_145.w_0" shape = [320] dtype = "float32" min_val = float("-0.0320265") @@ -1347,6 +1472,7 @@ class Program_weight_tensor_parameter_124: class Program_weight_tensor_parameter_125: name = "parameter_125" + original_name = "conv2d_96.b_0" shape = [320] dtype = "float32" min_val = float("-2.10584") @@ -1358,6 +1484,7 @@ class Program_weight_tensor_parameter_125: class Program_weight_tensor_parameter_126: name = "parameter_126" + original_name = "conv2d_96.w_0" shape = [320, 320, 2, 2] dtype = "float32" min_val = float("-0.428867") @@ -1369,6 +1496,7 @@ class Program_weight_tensor_parameter_126: class Program_weight_tensor_parameter_127: name = "parameter_127" + original_name = "linear_230.b_0" shape = [320] dtype = "float32" min_val = float("-2.60946") @@ -1380,6 +1508,7 @@ class Program_weight_tensor_parameter_127: class Program_weight_tensor_parameter_128: name = "parameter_128" + original_name = "linear_230.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.324627") @@ -1391,6 +1520,7 @@ class Program_weight_tensor_parameter_128: class Program_weight_tensor_parameter_129: name = "parameter_129" + original_name = "layer_norm_144.b_0" shape = [320] dtype = "float32" min_val = float("-1.17499") @@ -1402,6 +1532,7 @@ class Program_weight_tensor_parameter_129: class Program_weight_tensor_parameter_130: name = "parameter_130" + original_name = "layer_norm_144.w_0" shape = [320] dtype = "float32" min_val = float("0.27045") @@ -1413,6 +1544,7 @@ class Program_weight_tensor_parameter_130: class Program_weight_tensor_parameter_131: name = "parameter_131" + original_name = "linear_229.b_0" shape = [320] dtype = "float32" min_val = float("-0.33604") @@ -1424,6 +1556,7 @@ class Program_weight_tensor_parameter_131: class Program_weight_tensor_parameter_132: name = "parameter_132" + original_name = "linear_229.w_0" shape = [1280, 320] dtype = "float32" min_val = float("-0.801093") @@ -1435,6 +1568,7 @@ class Program_weight_tensor_parameter_132: class Program_weight_tensor_parameter_133: name = "parameter_133" + original_name = "conv2d_95.b_0" shape = [1280] dtype = "float32" min_val = float("-1.88638") @@ -1446,6 +1580,7 @@ class Program_weight_tensor_parameter_133: class Program_weight_tensor_parameter_134: name = "parameter_134" + original_name = "conv2d_95.w_0" shape = [1280, 1, 3, 3] dtype = "float32" min_val = float("-0.525723") @@ -1457,6 +1592,7 @@ class Program_weight_tensor_parameter_134: class Program_weight_tensor_parameter_135: name = "parameter_135" + original_name = "linear_228.b_0" shape = [1280] dtype = "float32" min_val = float("-2.42072") @@ -1468,6 +1604,7 @@ class Program_weight_tensor_parameter_135: class Program_weight_tensor_parameter_136: name = "parameter_136" + original_name = "linear_228.w_0" shape = [320, 1280] dtype = "float32" min_val = float("-0.332544") @@ -1479,6 +1616,7 @@ class Program_weight_tensor_parameter_136: class Program_weight_tensor_parameter_137: name = "parameter_137" + original_name = "layer_norm_143.b_0" shape = [320] dtype = "float32" min_val = float("-1.36046") @@ -1490,6 +1628,7 @@ class Program_weight_tensor_parameter_137: class Program_weight_tensor_parameter_138: name = "parameter_138" + original_name = "layer_norm_143.w_0" shape = [320] dtype = "float32" min_val = float("0.0390421") @@ -1501,6 +1640,7 @@ class Program_weight_tensor_parameter_138: class Program_weight_tensor_parameter_139: name = "parameter_139" + original_name = "linear_227.b_0" shape = [320] dtype = "float32" min_val = float("-0.483681") @@ -1512,6 +1652,7 @@ class Program_weight_tensor_parameter_139: class Program_weight_tensor_parameter_140: name = "parameter_140" + original_name = "linear_227.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.639744") @@ -1523,6 +1664,7 @@ class Program_weight_tensor_parameter_140: class Program_weight_tensor_parameter_141: name = "parameter_141" + original_name = "linear_226.b_0" shape = [640] dtype = "float32" min_val = float("-20.6281") @@ -1534,6 +1676,7 @@ class Program_weight_tensor_parameter_141: class Program_weight_tensor_parameter_142: name = "parameter_142" + original_name = "linear_226.w_0" shape = [320, 640] dtype = "float32" min_val = float("-0.414826") @@ -1545,6 +1688,7 @@ class Program_weight_tensor_parameter_142: class Program_weight_tensor_parameter_143: name = "parameter_143" + original_name = "layer_norm_142.b_0" shape = [320] dtype = "float32" min_val = float("-0.193303") @@ -1556,6 +1700,7 @@ class Program_weight_tensor_parameter_143: class Program_weight_tensor_parameter_144: name = "parameter_144" + original_name = "layer_norm_142.w_0" shape = [320] dtype = "float32" min_val = float("-0.108909") @@ -1567,6 +1712,7 @@ class Program_weight_tensor_parameter_144: class Program_weight_tensor_parameter_145: name = "parameter_145" + original_name = "conv2d_94.b_0" shape = [320] dtype = "float32" min_val = float("-1.96858") @@ -1578,6 +1724,7 @@ class Program_weight_tensor_parameter_145: class Program_weight_tensor_parameter_146: name = "parameter_146" + original_name = "conv2d_94.w_0" shape = [320, 320, 2, 2] dtype = "float32" min_val = float("-0.347373") @@ -1589,6 +1736,7 @@ class Program_weight_tensor_parameter_146: class Program_weight_tensor_parameter_147: name = "parameter_147" + original_name = "linear_225.b_0" shape = [320] dtype = "float32" min_val = float("-2.48399") @@ -1600,6 +1748,7 @@ class Program_weight_tensor_parameter_147: class Program_weight_tensor_parameter_148: name = "parameter_148" + original_name = "linear_225.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.308816") @@ -1611,6 +1760,7 @@ class Program_weight_tensor_parameter_148: class Program_weight_tensor_parameter_149: name = "parameter_149" + original_name = "layer_norm_141.b_0" shape = [320] dtype = "float32" min_val = float("-1.00136") @@ -1622,6 +1772,7 @@ class Program_weight_tensor_parameter_149: class Program_weight_tensor_parameter_150: name = "parameter_150" + original_name = "layer_norm_141.w_0" shape = [320] dtype = "float32" min_val = float("0.379627") @@ -1633,6 +1784,7 @@ class Program_weight_tensor_parameter_150: class Program_weight_tensor_parameter_151: name = "parameter_151" + original_name = "linear_224.b_0" shape = [320] dtype = "float32" min_val = float("-0.287676") @@ -1644,6 +1796,7 @@ class Program_weight_tensor_parameter_151: class Program_weight_tensor_parameter_152: name = "parameter_152" + original_name = "linear_224.w_0" shape = [1280, 320] dtype = "float32" min_val = float("-0.77788") @@ -1655,6 +1808,7 @@ class Program_weight_tensor_parameter_152: class Program_weight_tensor_parameter_153: name = "parameter_153" + original_name = "conv2d_93.b_0" shape = [1280] dtype = "float32" min_val = float("-1.5385") @@ -1666,6 +1820,7 @@ class Program_weight_tensor_parameter_153: class Program_weight_tensor_parameter_154: name = "parameter_154" + original_name = "conv2d_93.w_0" shape = [1280, 1, 3, 3] dtype = "float32" min_val = float("-0.470392") @@ -1677,6 +1832,7 @@ class Program_weight_tensor_parameter_154: class Program_weight_tensor_parameter_155: name = "parameter_155" + original_name = "linear_223.b_0" shape = [1280] dtype = "float32" min_val = float("-2.62194") @@ -1688,6 +1844,7 @@ class Program_weight_tensor_parameter_155: class Program_weight_tensor_parameter_156: name = "parameter_156" + original_name = "linear_223.w_0" shape = [320, 1280] dtype = "float32" min_val = float("-0.305521") @@ -1699,6 +1856,7 @@ class Program_weight_tensor_parameter_156: class Program_weight_tensor_parameter_157: name = "parameter_157" + original_name = "layer_norm_140.b_0" shape = [320] dtype = "float32" min_val = float("-1.23868") @@ -1710,6 +1868,7 @@ class Program_weight_tensor_parameter_157: class Program_weight_tensor_parameter_158: name = "parameter_158" + original_name = "layer_norm_140.w_0" shape = [320] dtype = "float32" min_val = float("0.0262557") @@ -1721,6 +1880,7 @@ class Program_weight_tensor_parameter_158: class Program_weight_tensor_parameter_159: name = "parameter_159" + original_name = "linear_222.b_0" shape = [320] dtype = "float32" min_val = float("-0.205749") @@ -1732,6 +1892,7 @@ class Program_weight_tensor_parameter_159: class Program_weight_tensor_parameter_160: name = "parameter_160" + original_name = "linear_222.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.834789") @@ -1743,6 +1904,7 @@ class Program_weight_tensor_parameter_160: class Program_weight_tensor_parameter_161: name = "parameter_161" + original_name = "linear_221.b_0" shape = [640] dtype = "float32" min_val = float("-2.83684") @@ -1754,6 +1916,7 @@ class Program_weight_tensor_parameter_161: class Program_weight_tensor_parameter_162: name = "parameter_162" + original_name = "linear_221.w_0" shape = [320, 640] dtype = "float32" min_val = float("-0.362715") @@ -1765,6 +1928,7 @@ class Program_weight_tensor_parameter_162: class Program_weight_tensor_parameter_163: name = "parameter_163" + original_name = "layer_norm_139.b_0" shape = [320] dtype = "float32" min_val = float("-0.418454") @@ -1776,6 +1940,7 @@ class Program_weight_tensor_parameter_163: class Program_weight_tensor_parameter_164: name = "parameter_164" + original_name = "layer_norm_139.w_0" shape = [320] dtype = "float32" min_val = float("-0.0966305") @@ -1787,6 +1952,7 @@ class Program_weight_tensor_parameter_164: class Program_weight_tensor_parameter_165: name = "parameter_165" + original_name = "conv2d_92.b_0" shape = [320] dtype = "float32" min_val = float("-3.5513") @@ -1798,6 +1964,7 @@ class Program_weight_tensor_parameter_165: class Program_weight_tensor_parameter_166: name = "parameter_166" + original_name = "conv2d_92.w_0" shape = [320, 320, 2, 2] dtype = "float32" min_val = float("-0.291294") @@ -1809,6 +1976,7 @@ class Program_weight_tensor_parameter_166: class Program_weight_tensor_parameter_167: name = "parameter_167" + original_name = "linear_220.b_0" shape = [320] dtype = "float32" min_val = float("-2.70865") @@ -1820,6 +1988,7 @@ class Program_weight_tensor_parameter_167: class Program_weight_tensor_parameter_168: name = "parameter_168" + original_name = "linear_220.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.257162") @@ -1831,6 +2000,7 @@ class Program_weight_tensor_parameter_168: class Program_weight_tensor_parameter_169: name = "parameter_169" + original_name = "layer_norm_138.b_0" shape = [320] dtype = "float32" min_val = float("-1.43355") @@ -1842,6 +2012,7 @@ class Program_weight_tensor_parameter_169: class Program_weight_tensor_parameter_170: name = "parameter_170" + original_name = "layer_norm_138.w_0" shape = [320] dtype = "float32" min_val = float("0.289137") @@ -1853,6 +2024,7 @@ class Program_weight_tensor_parameter_170: class Program_weight_tensor_parameter_171: name = "parameter_171" + original_name = "linear_219.b_0" shape = [320] dtype = "float32" min_val = float("-0.267224") @@ -1864,6 +2036,7 @@ class Program_weight_tensor_parameter_171: class Program_weight_tensor_parameter_172: name = "parameter_172" + original_name = "linear_219.w_0" shape = [1280, 320] dtype = "float32" min_val = float("-0.94624") @@ -1875,6 +2048,7 @@ class Program_weight_tensor_parameter_172: class Program_weight_tensor_parameter_173: name = "parameter_173" + original_name = "conv2d_91.b_0" shape = [1280] dtype = "float32" min_val = float("-1.31363") @@ -1886,6 +2060,7 @@ class Program_weight_tensor_parameter_173: class Program_weight_tensor_parameter_174: name = "parameter_174" + original_name = "conv2d_91.w_0" shape = [1280, 1, 3, 3] dtype = "float32" min_val = float("-0.339388") @@ -1897,6 +2072,7 @@ class Program_weight_tensor_parameter_174: class Program_weight_tensor_parameter_175: name = "parameter_175" + original_name = "linear_218.b_0" shape = [1280] dtype = "float32" min_val = float("-1.68985") @@ -1908,6 +2084,7 @@ class Program_weight_tensor_parameter_175: class Program_weight_tensor_parameter_176: name = "parameter_176" + original_name = "linear_218.w_0" shape = [320, 1280] dtype = "float32" min_val = float("-0.306238") @@ -1919,6 +2096,7 @@ class Program_weight_tensor_parameter_176: class Program_weight_tensor_parameter_177: name = "parameter_177" + original_name = "layer_norm_137.b_0" shape = [320] dtype = "float32" min_val = float("-1.16345") @@ -1930,6 +2108,7 @@ class Program_weight_tensor_parameter_177: class Program_weight_tensor_parameter_178: name = "parameter_178" + original_name = "layer_norm_137.w_0" shape = [320] dtype = "float32" min_val = float("0.00318684") @@ -1941,6 +2120,7 @@ class Program_weight_tensor_parameter_178: class Program_weight_tensor_parameter_179: name = "parameter_179" + original_name = "linear_217.b_0" shape = [320] dtype = "float32" min_val = float("-0.157115") @@ -1952,6 +2132,7 @@ class Program_weight_tensor_parameter_179: class Program_weight_tensor_parameter_180: name = "parameter_180" + original_name = "linear_217.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.598616") @@ -1963,6 +2144,7 @@ class Program_weight_tensor_parameter_180: class Program_weight_tensor_parameter_181: name = "parameter_181" + original_name = "linear_216.b_0" shape = [640] dtype = "float32" min_val = float("-4.39802") @@ -1974,6 +2156,7 @@ class Program_weight_tensor_parameter_181: class Program_weight_tensor_parameter_182: name = "parameter_182" + original_name = "linear_216.w_0" shape = [320, 640] dtype = "float32" min_val = float("-0.426193") @@ -1985,6 +2168,7 @@ class Program_weight_tensor_parameter_182: class Program_weight_tensor_parameter_183: name = "parameter_183" + original_name = "layer_norm_136.b_0" shape = [320] dtype = "float32" min_val = float("-0.253983") @@ -1996,6 +2180,7 @@ class Program_weight_tensor_parameter_183: class Program_weight_tensor_parameter_184: name = "parameter_184" + original_name = "layer_norm_136.w_0" shape = [320] dtype = "float32" min_val = float("-0.214936") @@ -2007,6 +2192,7 @@ class Program_weight_tensor_parameter_184: class Program_weight_tensor_parameter_185: name = "parameter_185" + original_name = "conv2d_90.b_0" shape = [320] dtype = "float32" min_val = float("-5.28786") @@ -2018,6 +2204,7 @@ class Program_weight_tensor_parameter_185: class Program_weight_tensor_parameter_186: name = "parameter_186" + original_name = "conv2d_90.w_0" shape = [320, 320, 2, 2] dtype = "float32" min_val = float("-0.467319") @@ -2029,6 +2216,7 @@ class Program_weight_tensor_parameter_186: class Program_weight_tensor_parameter_187: name = "parameter_187" + original_name = "linear_215.b_0" shape = [320] dtype = "float32" min_val = float("-3.11809") @@ -2040,6 +2228,7 @@ class Program_weight_tensor_parameter_187: class Program_weight_tensor_parameter_188: name = "parameter_188" + original_name = "linear_215.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.253443") @@ -2051,6 +2240,7 @@ class Program_weight_tensor_parameter_188: class Program_weight_tensor_parameter_189: name = "parameter_189" + original_name = "layer_norm_135.b_0" shape = [320] dtype = "float32" min_val = float("-1.48207") @@ -2062,6 +2252,7 @@ class Program_weight_tensor_parameter_189: class Program_weight_tensor_parameter_190: name = "parameter_190" + original_name = "layer_norm_135.w_0" shape = [320] dtype = "float32" min_val = float("0.169469") @@ -2073,6 +2264,7 @@ class Program_weight_tensor_parameter_190: class Program_weight_tensor_parameter_191: name = "parameter_191" + original_name = "linear_214.b_0" shape = [320] dtype = "float32" min_val = float("-0.255695") @@ -2084,6 +2276,7 @@ class Program_weight_tensor_parameter_191: class Program_weight_tensor_parameter_192: name = "parameter_192" + original_name = "linear_214.w_0" shape = [1280, 320] dtype = "float32" min_val = float("-1.04585") @@ -2095,6 +2288,7 @@ class Program_weight_tensor_parameter_192: class Program_weight_tensor_parameter_193: name = "parameter_193" + original_name = "conv2d_89.b_0" shape = [1280] dtype = "float32" min_val = float("-1.3552") @@ -2106,6 +2300,7 @@ class Program_weight_tensor_parameter_193: class Program_weight_tensor_parameter_194: name = "parameter_194" + original_name = "conv2d_89.w_0" shape = [1280, 1, 3, 3] dtype = "float32" min_val = float("-0.369311") @@ -2117,6 +2312,7 @@ class Program_weight_tensor_parameter_194: class Program_weight_tensor_parameter_195: name = "parameter_195" + original_name = "linear_213.b_0" shape = [1280] dtype = "float32" min_val = float("-2.91215") @@ -2128,6 +2324,7 @@ class Program_weight_tensor_parameter_195: class Program_weight_tensor_parameter_196: name = "parameter_196" + original_name = "linear_213.w_0" shape = [320, 1280] dtype = "float32" min_val = float("-1.12572") @@ -2139,6 +2336,7 @@ class Program_weight_tensor_parameter_196: class Program_weight_tensor_parameter_197: name = "parameter_197" + original_name = "layer_norm_134.b_0" shape = [320] dtype = "float32" min_val = float("-1.26212") @@ -2150,6 +2348,7 @@ class Program_weight_tensor_parameter_197: class Program_weight_tensor_parameter_198: name = "parameter_198" + original_name = "layer_norm_134.w_0" shape = [320] dtype = "float32" min_val = float("0.112685") @@ -2161,6 +2360,7 @@ class Program_weight_tensor_parameter_198: class Program_weight_tensor_parameter_199: name = "parameter_199" + original_name = "linear_212.b_0" shape = [320] dtype = "float32" min_val = float("-0.138783") @@ -2172,6 +2372,7 @@ class Program_weight_tensor_parameter_199: class Program_weight_tensor_parameter_200: name = "parameter_200" + original_name = "linear_212.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.622924") @@ -2183,6 +2384,7 @@ class Program_weight_tensor_parameter_200: class Program_weight_tensor_parameter_201: name = "parameter_201" + original_name = "linear_211.b_0" shape = [640] dtype = "float32" min_val = float("-6.16358") @@ -2194,6 +2396,7 @@ class Program_weight_tensor_parameter_201: class Program_weight_tensor_parameter_202: name = "parameter_202" + original_name = "linear_211.w_0" shape = [320, 640] dtype = "float32" min_val = float("-0.355349") @@ -2205,6 +2408,7 @@ class Program_weight_tensor_parameter_202: class Program_weight_tensor_parameter_203: name = "parameter_203" + original_name = "layer_norm_133.b_0" shape = [320] dtype = "float32" min_val = float("-0.278457") @@ -2216,6 +2420,7 @@ class Program_weight_tensor_parameter_203: class Program_weight_tensor_parameter_204: name = "parameter_204" + original_name = "layer_norm_133.w_0" shape = [320] dtype = "float32" min_val = float("-0.0954183") @@ -2227,6 +2432,7 @@ class Program_weight_tensor_parameter_204: class Program_weight_tensor_parameter_205: name = "parameter_205" + original_name = "conv2d_88.b_0" shape = [320] dtype = "float32" min_val = float("-5.2467") @@ -2238,6 +2444,7 @@ class Program_weight_tensor_parameter_205: class Program_weight_tensor_parameter_206: name = "parameter_206" + original_name = "conv2d_88.w_0" shape = [320, 320, 2, 2] dtype = "float32" min_val = float("-0.544662") @@ -2249,6 +2456,7 @@ class Program_weight_tensor_parameter_206: class Program_weight_tensor_parameter_207: name = "parameter_207" + original_name = "linear_210.b_0" shape = [320] dtype = "float32" min_val = float("-1.86372") @@ -2260,6 +2468,7 @@ class Program_weight_tensor_parameter_207: class Program_weight_tensor_parameter_208: name = "parameter_208" + original_name = "linear_210.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.260841") @@ -2271,6 +2480,7 @@ class Program_weight_tensor_parameter_208: class Program_weight_tensor_parameter_209: name = "parameter_209" + original_name = "layer_norm_132.b_0" shape = [320] dtype = "float32" min_val = float("-1.4234") @@ -2282,6 +2492,7 @@ class Program_weight_tensor_parameter_209: class Program_weight_tensor_parameter_210: name = "parameter_210" + original_name = "layer_norm_132.w_0" shape = [320] dtype = "float32" min_val = float("0.167792") @@ -2293,6 +2504,7 @@ class Program_weight_tensor_parameter_210: class Program_weight_tensor_parameter_211: name = "parameter_211" + original_name = "linear_209.b_0" shape = [320] dtype = "float32" min_val = float("-0.292387") @@ -2304,6 +2516,7 @@ class Program_weight_tensor_parameter_211: class Program_weight_tensor_parameter_212: name = "parameter_212" + original_name = "linear_209.w_0" shape = [1280, 320] dtype = "float32" min_val = float("-1.06765") @@ -2315,6 +2528,7 @@ class Program_weight_tensor_parameter_212: class Program_weight_tensor_parameter_213: name = "parameter_213" + original_name = "conv2d_87.b_0" shape = [1280] dtype = "float32" min_val = float("-1.7001") @@ -2326,6 +2540,7 @@ class Program_weight_tensor_parameter_213: class Program_weight_tensor_parameter_214: name = "parameter_214" + original_name = "conv2d_87.w_0" shape = [1280, 1, 3, 3] dtype = "float32" min_val = float("-0.284687") @@ -2337,6 +2552,7 @@ class Program_weight_tensor_parameter_214: class Program_weight_tensor_parameter_215: name = "parameter_215" + original_name = "linear_208.b_0" shape = [1280] dtype = "float32" min_val = float("-2.66598") @@ -2348,6 +2564,7 @@ class Program_weight_tensor_parameter_215: class Program_weight_tensor_parameter_216: name = "parameter_216" + original_name = "linear_208.w_0" shape = [320, 1280] dtype = "float32" min_val = float("-0.437568") @@ -2359,6 +2576,7 @@ class Program_weight_tensor_parameter_216: class Program_weight_tensor_parameter_217: name = "parameter_217" + original_name = "layer_norm_131.b_0" shape = [320] dtype = "float32" min_val = float("-1.12555") @@ -2370,6 +2588,7 @@ class Program_weight_tensor_parameter_217: class Program_weight_tensor_parameter_218: name = "parameter_218" + original_name = "layer_norm_131.w_0" shape = [320] dtype = "float32" min_val = float("-0.00272376") @@ -2381,6 +2600,7 @@ class Program_weight_tensor_parameter_218: class Program_weight_tensor_parameter_219: name = "parameter_219" + original_name = "linear_207.b_0" shape = [320] dtype = "float32" min_val = float("-0.267413") @@ -2392,6 +2612,7 @@ class Program_weight_tensor_parameter_219: class Program_weight_tensor_parameter_220: name = "parameter_220" + original_name = "linear_207.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.779295") @@ -2403,6 +2624,7 @@ class Program_weight_tensor_parameter_220: class Program_weight_tensor_parameter_221: name = "parameter_221" + original_name = "linear_206.b_0" shape = [640] dtype = "float32" min_val = float("-7.56718") @@ -2414,6 +2636,7 @@ class Program_weight_tensor_parameter_221: class Program_weight_tensor_parameter_222: name = "parameter_222" + original_name = "linear_206.w_0" shape = [320, 640] dtype = "float32" min_val = float("-0.486972") @@ -2425,6 +2648,7 @@ class Program_weight_tensor_parameter_222: class Program_weight_tensor_parameter_223: name = "parameter_223" + original_name = "layer_norm_130.b_0" shape = [320] dtype = "float32" min_val = float("-0.434674") @@ -2436,6 +2660,7 @@ class Program_weight_tensor_parameter_223: class Program_weight_tensor_parameter_224: name = "parameter_224" + original_name = "layer_norm_130.w_0" shape = [320] dtype = "float32" min_val = float("-0.0791383") @@ -2447,6 +2672,7 @@ class Program_weight_tensor_parameter_224: class Program_weight_tensor_parameter_225: name = "parameter_225" + original_name = "conv2d_86.b_0" shape = [320] dtype = "float32" min_val = float("-3.47912") @@ -2458,6 +2684,7 @@ class Program_weight_tensor_parameter_225: class Program_weight_tensor_parameter_226: name = "parameter_226" + original_name = "conv2d_86.w_0" shape = [320, 320, 2, 2] dtype = "float32" min_val = float("-0.226725") @@ -2469,6 +2696,7 @@ class Program_weight_tensor_parameter_226: class Program_weight_tensor_parameter_227: name = "parameter_227" + original_name = "linear_205.b_0" shape = [320] dtype = "float32" min_val = float("-1.82421") @@ -2480,6 +2708,7 @@ class Program_weight_tensor_parameter_227: class Program_weight_tensor_parameter_228: name = "parameter_228" + original_name = "linear_205.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.413882") @@ -2491,6 +2720,7 @@ class Program_weight_tensor_parameter_228: class Program_weight_tensor_parameter_229: name = "parameter_229" + original_name = "layer_norm_129.b_0" shape = [320] dtype = "float32" min_val = float("-0.789545") @@ -2502,6 +2732,7 @@ class Program_weight_tensor_parameter_229: class Program_weight_tensor_parameter_230: name = "parameter_230" + original_name = "layer_norm_129.w_0" shape = [320] dtype = "float32" min_val = float("-0.199647") @@ -2513,6 +2744,7 @@ class Program_weight_tensor_parameter_230: class Program_weight_tensor_parameter_231: name = "parameter_231" + original_name = "linear_204.b_0" shape = [320] dtype = "float32" min_val = float("-0.226323") @@ -2524,6 +2756,7 @@ class Program_weight_tensor_parameter_231: class Program_weight_tensor_parameter_232: name = "parameter_232" + original_name = "linear_204.w_0" shape = [1280, 320] dtype = "float32" min_val = float("-0.74588") @@ -2535,6 +2768,7 @@ class Program_weight_tensor_parameter_232: class Program_weight_tensor_parameter_233: name = "parameter_233" + original_name = "conv2d_85.b_0" shape = [1280] dtype = "float32" min_val = float("-1.7268") @@ -2546,6 +2780,7 @@ class Program_weight_tensor_parameter_233: class Program_weight_tensor_parameter_234: name = "parameter_234" + original_name = "conv2d_85.w_0" shape = [1280, 1, 3, 3] dtype = "float32" min_val = float("-0.719857") @@ -2557,6 +2792,7 @@ class Program_weight_tensor_parameter_234: class Program_weight_tensor_parameter_235: name = "parameter_235" + original_name = "linear_203.b_0" shape = [1280] dtype = "float32" min_val = float("-2.92474") @@ -2568,6 +2804,7 @@ class Program_weight_tensor_parameter_235: class Program_weight_tensor_parameter_236: name = "parameter_236" + original_name = "linear_203.w_0" shape = [320, 1280] dtype = "float32" min_val = float("-0.472637") @@ -2579,6 +2816,7 @@ class Program_weight_tensor_parameter_236: class Program_weight_tensor_parameter_237: name = "parameter_237" + original_name = "layer_norm_128.b_0" shape = [320] dtype = "float32" min_val = float("-1.10633") @@ -2590,6 +2828,7 @@ class Program_weight_tensor_parameter_237: class Program_weight_tensor_parameter_238: name = "parameter_238" + original_name = "layer_norm_128.w_0" shape = [320] dtype = "float32" min_val = float("0.00304593") @@ -2601,6 +2840,7 @@ class Program_weight_tensor_parameter_238: class Program_weight_tensor_parameter_239: name = "parameter_239" + original_name = "linear_202.b_0" shape = [320] dtype = "float32" min_val = float("-0.137469") @@ -2612,6 +2852,7 @@ class Program_weight_tensor_parameter_239: class Program_weight_tensor_parameter_240: name = "parameter_240" + original_name = "linear_202.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.734359") @@ -2623,6 +2864,7 @@ class Program_weight_tensor_parameter_240: class Program_weight_tensor_parameter_241: name = "parameter_241" + original_name = "linear_201.b_0" shape = [640] dtype = "float32" min_val = float("-8.11013") @@ -2634,6 +2876,7 @@ class Program_weight_tensor_parameter_241: class Program_weight_tensor_parameter_242: name = "parameter_242" + original_name = "linear_201.w_0" shape = [320, 640] dtype = "float32" min_val = float("-0.331209") @@ -2645,6 +2888,7 @@ class Program_weight_tensor_parameter_242: class Program_weight_tensor_parameter_243: name = "parameter_243" + original_name = "layer_norm_127.b_0" shape = [320] dtype = "float32" min_val = float("-0.199389") @@ -2656,6 +2900,7 @@ class Program_weight_tensor_parameter_243: class Program_weight_tensor_parameter_244: name = "parameter_244" + original_name = "layer_norm_127.w_0" shape = [320] dtype = "float32" min_val = float("-0.0296447") @@ -2667,6 +2912,7 @@ class Program_weight_tensor_parameter_244: class Program_weight_tensor_parameter_245: name = "parameter_245" + original_name = "conv2d_84.b_0" shape = [320] dtype = "float32" min_val = float("-3.7846") @@ -2678,6 +2924,7 @@ class Program_weight_tensor_parameter_245: class Program_weight_tensor_parameter_246: name = "parameter_246" + original_name = "conv2d_84.w_0" shape = [320, 320, 2, 2] dtype = "float32" min_val = float("-0.430166") @@ -2689,6 +2936,7 @@ class Program_weight_tensor_parameter_246: class Program_weight_tensor_parameter_247: name = "parameter_247" + original_name = "linear_200.b_0" shape = [320] dtype = "float32" min_val = float("-1.6438") @@ -2700,6 +2948,7 @@ class Program_weight_tensor_parameter_247: class Program_weight_tensor_parameter_248: name = "parameter_248" + original_name = "linear_200.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.290477") @@ -2711,6 +2960,7 @@ class Program_weight_tensor_parameter_248: class Program_weight_tensor_parameter_249: name = "parameter_249" + original_name = "layer_norm_126.b_0" shape = [320] dtype = "float32" min_val = float("-1.62617") @@ -2722,6 +2972,7 @@ class Program_weight_tensor_parameter_249: class Program_weight_tensor_parameter_250: name = "parameter_250" + original_name = "layer_norm_126.w_0" shape = [320] dtype = "float32" min_val = float("-0.0302405") @@ -2733,6 +2984,7 @@ class Program_weight_tensor_parameter_250: class Program_weight_tensor_parameter_251: name = "parameter_251" + original_name = "linear_199.b_0" shape = [320] dtype = "float32" min_val = float("-0.234558") @@ -2744,6 +2996,7 @@ class Program_weight_tensor_parameter_251: class Program_weight_tensor_parameter_252: name = "parameter_252" + original_name = "linear_199.w_0" shape = [1280, 320] dtype = "float32" min_val = float("-0.727356") @@ -2755,6 +3008,7 @@ class Program_weight_tensor_parameter_252: class Program_weight_tensor_parameter_253: name = "parameter_253" + original_name = "conv2d_83.b_0" shape = [1280] dtype = "float32" min_val = float("-1.36569") @@ -2766,6 +3020,7 @@ class Program_weight_tensor_parameter_253: class Program_weight_tensor_parameter_254: name = "parameter_254" + original_name = "conv2d_83.w_0" shape = [1280, 1, 3, 3] dtype = "float32" min_val = float("-0.254647") @@ -2777,6 +3032,7 @@ class Program_weight_tensor_parameter_254: class Program_weight_tensor_parameter_255: name = "parameter_255" + original_name = "linear_198.b_0" shape = [1280] dtype = "float32" min_val = float("-2.44043") @@ -2788,6 +3044,7 @@ class Program_weight_tensor_parameter_255: class Program_weight_tensor_parameter_256: name = "parameter_256" + original_name = "linear_198.w_0" shape = [320, 1280] dtype = "float32" min_val = float("-0.423588") @@ -2799,6 +3056,7 @@ class Program_weight_tensor_parameter_256: class Program_weight_tensor_parameter_257: name = "parameter_257" + original_name = "layer_norm_125.b_0" shape = [320] dtype = "float32" min_val = float("-1.09194") @@ -2810,6 +3068,7 @@ class Program_weight_tensor_parameter_257: class Program_weight_tensor_parameter_258: name = "parameter_258" + original_name = "layer_norm_125.w_0" shape = [320] dtype = "float32" min_val = float("-0.00209891") @@ -2821,6 +3080,7 @@ class Program_weight_tensor_parameter_258: class Program_weight_tensor_parameter_259: name = "parameter_259" + original_name = "linear_197.b_0" shape = [320] dtype = "float32" min_val = float("-0.251095") @@ -2832,6 +3092,7 @@ class Program_weight_tensor_parameter_259: class Program_weight_tensor_parameter_260: name = "parameter_260" + original_name = "linear_197.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.698944") @@ -2843,6 +3104,7 @@ class Program_weight_tensor_parameter_260: class Program_weight_tensor_parameter_261: name = "parameter_261" + original_name = "linear_196.b_0" shape = [640] dtype = "float32" min_val = float("-5.95613") @@ -2854,6 +3116,7 @@ class Program_weight_tensor_parameter_261: class Program_weight_tensor_parameter_262: name = "parameter_262" + original_name = "linear_196.w_0" shape = [320, 640] dtype = "float32" min_val = float("-0.389783") @@ -2865,6 +3128,7 @@ class Program_weight_tensor_parameter_262: class Program_weight_tensor_parameter_263: name = "parameter_263" + original_name = "layer_norm_124.b_0" shape = [320] dtype = "float32" min_val = float("-0.238718") @@ -2876,6 +3140,7 @@ class Program_weight_tensor_parameter_263: class Program_weight_tensor_parameter_264: name = "parameter_264" + original_name = "layer_norm_124.w_0" shape = [320] dtype = "float32" min_val = float("-0.0092566") @@ -2887,6 +3152,7 @@ class Program_weight_tensor_parameter_264: class Program_weight_tensor_parameter_265: name = "parameter_265" + original_name = "conv2d_82.b_0" shape = [320] dtype = "float32" min_val = float("-4.6069") @@ -2898,6 +3164,7 @@ class Program_weight_tensor_parameter_265: class Program_weight_tensor_parameter_266: name = "parameter_266" + original_name = "conv2d_82.w_0" shape = [320, 320, 2, 2] dtype = "float32" min_val = float("-0.423275") @@ -2909,6 +3176,7 @@ class Program_weight_tensor_parameter_266: class Program_weight_tensor_parameter_267: name = "parameter_267" + original_name = "linear_195.b_0" shape = [320] dtype = "float32" min_val = float("-1.38311") @@ -2920,6 +3188,7 @@ class Program_weight_tensor_parameter_267: class Program_weight_tensor_parameter_268: name = "parameter_268" + original_name = "linear_195.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.316663") @@ -2931,6 +3200,7 @@ class Program_weight_tensor_parameter_268: class Program_weight_tensor_parameter_269: name = "parameter_269" + original_name = "layer_norm_123.b_0" shape = [320] dtype = "float32" min_val = float("-1.51941") @@ -2942,6 +3212,7 @@ class Program_weight_tensor_parameter_269: class Program_weight_tensor_parameter_270: name = "parameter_270" + original_name = "layer_norm_123.w_0" shape = [320] dtype = "float32" min_val = float("0.195294") @@ -2953,6 +3224,7 @@ class Program_weight_tensor_parameter_270: class Program_weight_tensor_parameter_271: name = "parameter_271" + original_name = "linear_194.b_0" shape = [320] dtype = "float32" min_val = float("-0.212451") @@ -2964,6 +3236,7 @@ class Program_weight_tensor_parameter_271: class Program_weight_tensor_parameter_272: name = "parameter_272" + original_name = "linear_194.w_0" shape = [1280, 320] dtype = "float32" min_val = float("-0.762894") @@ -2975,6 +3248,7 @@ class Program_weight_tensor_parameter_272: class Program_weight_tensor_parameter_273: name = "parameter_273" + original_name = "conv2d_81.b_0" shape = [1280] dtype = "float32" min_val = float("-1.42303") @@ -2986,6 +3260,7 @@ class Program_weight_tensor_parameter_273: class Program_weight_tensor_parameter_274: name = "parameter_274" + original_name = "conv2d_81.w_0" shape = [1280, 1, 3, 3] dtype = "float32" min_val = float("-0.291364") @@ -2997,6 +3272,7 @@ class Program_weight_tensor_parameter_274: class Program_weight_tensor_parameter_275: name = "parameter_275" + original_name = "linear_193.b_0" shape = [1280] dtype = "float32" min_val = float("-2.32766") @@ -3008,6 +3284,7 @@ class Program_weight_tensor_parameter_275: class Program_weight_tensor_parameter_276: name = "parameter_276" + original_name = "linear_193.w_0" shape = [320, 1280] dtype = "float32" min_val = float("-0.237654") @@ -3019,6 +3296,7 @@ class Program_weight_tensor_parameter_276: class Program_weight_tensor_parameter_277: name = "parameter_277" + original_name = "layer_norm_122.b_0" shape = [320] dtype = "float32" min_val = float("-1.32246") @@ -3030,6 +3308,7 @@ class Program_weight_tensor_parameter_277: class Program_weight_tensor_parameter_278: name = "parameter_278" + original_name = "layer_norm_122.w_0" shape = [320] dtype = "float32" min_val = float("0.0118884") @@ -3041,6 +3320,7 @@ class Program_weight_tensor_parameter_278: class Program_weight_tensor_parameter_279: name = "parameter_279" + original_name = "linear_192.b_0" shape = [320] dtype = "float32" min_val = float("-0.158567") @@ -3052,6 +3332,7 @@ class Program_weight_tensor_parameter_279: class Program_weight_tensor_parameter_280: name = "parameter_280" + original_name = "linear_192.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.230686") @@ -3063,6 +3344,7 @@ class Program_weight_tensor_parameter_280: class Program_weight_tensor_parameter_281: name = "parameter_281" + original_name = "linear_191.b_0" shape = [640] dtype = "float32" min_val = float("-5.45671") @@ -3074,6 +3356,7 @@ class Program_weight_tensor_parameter_281: class Program_weight_tensor_parameter_282: name = "parameter_282" + original_name = "linear_191.w_0" shape = [320, 640] dtype = "float32" min_val = float("-0.430799") @@ -3085,6 +3368,7 @@ class Program_weight_tensor_parameter_282: class Program_weight_tensor_parameter_283: name = "parameter_283" + original_name = "layer_norm_121.b_0" shape = [320] dtype = "float32" min_val = float("-0.157535") @@ -3096,6 +3380,7 @@ class Program_weight_tensor_parameter_283: class Program_weight_tensor_parameter_284: name = "parameter_284" + original_name = "layer_norm_121.w_0" shape = [320] dtype = "float32" min_val = float("-0.0826056") @@ -3107,6 +3392,7 @@ class Program_weight_tensor_parameter_284: class Program_weight_tensor_parameter_285: name = "parameter_285" + original_name = "conv2d_80.b_0" shape = [320] dtype = "float32" min_val = float("-5.25") @@ -3118,6 +3404,7 @@ class Program_weight_tensor_parameter_285: class Program_weight_tensor_parameter_286: name = "parameter_286" + original_name = "conv2d_80.w_0" shape = [320, 320, 2, 2] dtype = "float32" min_val = float("-0.421888") @@ -3129,6 +3416,7 @@ class Program_weight_tensor_parameter_286: class Program_weight_tensor_parameter_287: name = "parameter_287" + original_name = "linear_190.b_0" shape = [320] dtype = "float32" min_val = float("-2.49861") @@ -3140,6 +3428,7 @@ class Program_weight_tensor_parameter_287: class Program_weight_tensor_parameter_288: name = "parameter_288" + original_name = "linear_190.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.249187") @@ -3151,6 +3440,7 @@ class Program_weight_tensor_parameter_288: class Program_weight_tensor_parameter_289: name = "parameter_289" + original_name = "layer_norm_120.b_0" shape = [320] dtype = "float32" min_val = float("-1.01736") @@ -3162,6 +3452,7 @@ class Program_weight_tensor_parameter_289: class Program_weight_tensor_parameter_290: name = "parameter_290" + original_name = "layer_norm_120.w_0" shape = [320] dtype = "float32" min_val = float("0.0820517") @@ -3173,6 +3464,7 @@ class Program_weight_tensor_parameter_290: class Program_weight_tensor_parameter_291: name = "parameter_291" + original_name = "linear_189.b_0" shape = [320] dtype = "float32" min_val = float("-0.182801") @@ -3184,6 +3476,7 @@ class Program_weight_tensor_parameter_291: class Program_weight_tensor_parameter_292: name = "parameter_292" + original_name = "linear_189.w_0" shape = [1280, 320] dtype = "float32" min_val = float("-0.603279") @@ -3195,6 +3488,7 @@ class Program_weight_tensor_parameter_292: class Program_weight_tensor_parameter_293: name = "parameter_293" + original_name = "conv2d_79.b_0" shape = [1280] dtype = "float32" min_val = float("-1.86319") @@ -3206,6 +3500,7 @@ class Program_weight_tensor_parameter_293: class Program_weight_tensor_parameter_294: name = "parameter_294" + original_name = "conv2d_79.w_0" shape = [1280, 1, 3, 3] dtype = "float32" min_val = float("-0.360099") @@ -3217,6 +3512,7 @@ class Program_weight_tensor_parameter_294: class Program_weight_tensor_parameter_295: name = "parameter_295" + original_name = "linear_188.b_0" shape = [1280] dtype = "float32" min_val = float("-1.21291") @@ -3228,6 +3524,7 @@ class Program_weight_tensor_parameter_295: class Program_weight_tensor_parameter_296: name = "parameter_296" + original_name = "linear_188.w_0" shape = [320, 1280] dtype = "float32" min_val = float("-0.329059") @@ -3239,6 +3536,7 @@ class Program_weight_tensor_parameter_296: class Program_weight_tensor_parameter_297: name = "parameter_297" + original_name = "layer_norm_119.b_0" shape = [320] dtype = "float32" min_val = float("-1.13621") @@ -3250,6 +3548,7 @@ class Program_weight_tensor_parameter_297: class Program_weight_tensor_parameter_298: name = "parameter_298" + original_name = "layer_norm_119.w_0" shape = [320] dtype = "float32" min_val = float("-0.00518586") @@ -3261,6 +3560,7 @@ class Program_weight_tensor_parameter_298: class Program_weight_tensor_parameter_299: name = "parameter_299" + original_name = "linear_187.b_0" shape = [320] dtype = "float32" min_val = float("-0.143441") @@ -3272,6 +3572,7 @@ class Program_weight_tensor_parameter_299: class Program_weight_tensor_parameter_300: name = "parameter_300" + original_name = "linear_187.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.208241") @@ -3283,6 +3584,7 @@ class Program_weight_tensor_parameter_300: class Program_weight_tensor_parameter_301: name = "parameter_301" + original_name = "linear_186.b_0" shape = [640] dtype = "float32" min_val = float("-3.60592") @@ -3294,6 +3596,7 @@ class Program_weight_tensor_parameter_301: class Program_weight_tensor_parameter_302: name = "parameter_302" + original_name = "linear_186.w_0" shape = [320, 640] dtype = "float32" min_val = float("-0.436491") @@ -3305,6 +3608,7 @@ class Program_weight_tensor_parameter_302: class Program_weight_tensor_parameter_303: name = "parameter_303" + original_name = "layer_norm_118.b_0" shape = [320] dtype = "float32" min_val = float("-0.207984") @@ -3316,6 +3620,7 @@ class Program_weight_tensor_parameter_303: class Program_weight_tensor_parameter_304: name = "parameter_304" + original_name = "layer_norm_118.w_0" shape = [320] dtype = "float32" min_val = float("-0.00776677") @@ -3327,6 +3632,7 @@ class Program_weight_tensor_parameter_304: class Program_weight_tensor_parameter_305: name = "parameter_305" + original_name = "conv2d_78.b_0" shape = [320] dtype = "float32" min_val = float("-6.24008") @@ -3338,6 +3644,7 @@ class Program_weight_tensor_parameter_305: class Program_weight_tensor_parameter_306: name = "parameter_306" + original_name = "conv2d_78.w_0" shape = [320, 320, 2, 2] dtype = "float32" min_val = float("-0.558594") @@ -3349,6 +3656,7 @@ class Program_weight_tensor_parameter_306: class Program_weight_tensor_parameter_307: name = "parameter_307" + original_name = "linear_185.b_0" shape = [320] dtype = "float32" min_val = float("-2.10799") @@ -3360,6 +3668,7 @@ class Program_weight_tensor_parameter_307: class Program_weight_tensor_parameter_308: name = "parameter_308" + original_name = "linear_185.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.261043") @@ -3371,6 +3680,7 @@ class Program_weight_tensor_parameter_308: class Program_weight_tensor_parameter_309: name = "parameter_309" + original_name = "layer_norm_117.b_0" shape = [320] dtype = "float32" min_val = float("-0.937679") @@ -3382,6 +3692,7 @@ class Program_weight_tensor_parameter_309: class Program_weight_tensor_parameter_310: name = "parameter_310" + original_name = "layer_norm_117.w_0" shape = [320] dtype = "float32" min_val = float("0.0650364") @@ -3393,6 +3704,7 @@ class Program_weight_tensor_parameter_310: class Program_weight_tensor_parameter_311: name = "parameter_311" + original_name = "linear_184.b_0" shape = [320] dtype = "float32" min_val = float("-0.195398") @@ -3404,6 +3716,7 @@ class Program_weight_tensor_parameter_311: class Program_weight_tensor_parameter_312: name = "parameter_312" + original_name = "linear_184.w_0" shape = [1280, 320] dtype = "float32" min_val = float("-0.432497") @@ -3415,6 +3728,7 @@ class Program_weight_tensor_parameter_312: class Program_weight_tensor_parameter_313: name = "parameter_313" + original_name = "conv2d_77.b_0" shape = [1280] dtype = "float32" min_val = float("-1.02516") @@ -3426,6 +3740,7 @@ class Program_weight_tensor_parameter_313: class Program_weight_tensor_parameter_314: name = "parameter_314" + original_name = "conv2d_77.w_0" shape = [1280, 1, 3, 3] dtype = "float32" min_val = float("-0.270221") @@ -3437,6 +3752,7 @@ class Program_weight_tensor_parameter_314: class Program_weight_tensor_parameter_315: name = "parameter_315" + original_name = "linear_183.b_0" shape = [1280] dtype = "float32" min_val = float("-3.03239") @@ -3448,6 +3764,7 @@ class Program_weight_tensor_parameter_315: class Program_weight_tensor_parameter_316: name = "parameter_316" + original_name = "linear_183.w_0" shape = [320, 1280] dtype = "float32" min_val = float("-0.27316") @@ -3459,6 +3776,7 @@ class Program_weight_tensor_parameter_316: class Program_weight_tensor_parameter_317: name = "parameter_317" + original_name = "layer_norm_116.b_0" shape = [320] dtype = "float32" min_val = float("-1.13776") @@ -3470,6 +3788,7 @@ class Program_weight_tensor_parameter_317: class Program_weight_tensor_parameter_318: name = "parameter_318" + original_name = "layer_norm_116.w_0" shape = [320] dtype = "float32" min_val = float("-0.00666182") @@ -3481,6 +3800,7 @@ class Program_weight_tensor_parameter_318: class Program_weight_tensor_parameter_319: name = "parameter_319" + original_name = "linear_182.b_0" shape = [320] dtype = "float32" min_val = float("-0.0833429") @@ -3492,6 +3812,7 @@ class Program_weight_tensor_parameter_319: class Program_weight_tensor_parameter_320: name = "parameter_320" + original_name = "linear_182.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.21217") @@ -3503,6 +3824,7 @@ class Program_weight_tensor_parameter_320: class Program_weight_tensor_parameter_321: name = "parameter_321" + original_name = "linear_181.b_0" shape = [640] dtype = "float32" min_val = float("-1.94433") @@ -3514,6 +3836,7 @@ class Program_weight_tensor_parameter_321: class Program_weight_tensor_parameter_322: name = "parameter_322" + original_name = "linear_181.w_0" shape = [320, 640] dtype = "float32" min_val = float("-0.476529") @@ -3525,6 +3848,7 @@ class Program_weight_tensor_parameter_322: class Program_weight_tensor_parameter_323: name = "parameter_323" + original_name = "layer_norm_115.b_0" shape = [320] dtype = "float32" min_val = float("-0.158993") @@ -3536,6 +3860,7 @@ class Program_weight_tensor_parameter_323: class Program_weight_tensor_parameter_324: name = "parameter_324" + original_name = "layer_norm_115.w_0" shape = [320] dtype = "float32" min_val = float("-0.174354") @@ -3547,6 +3872,7 @@ class Program_weight_tensor_parameter_324: class Program_weight_tensor_parameter_325: name = "parameter_325" + original_name = "conv2d_76.b_0" shape = [320] dtype = "float32" min_val = float("-5.15309") @@ -3558,6 +3884,7 @@ class Program_weight_tensor_parameter_325: class Program_weight_tensor_parameter_326: name = "parameter_326" + original_name = "conv2d_76.w_0" shape = [320, 320, 2, 2] dtype = "float32" min_val = float("-0.441333") @@ -3569,6 +3896,7 @@ class Program_weight_tensor_parameter_326: class Program_weight_tensor_parameter_327: name = "parameter_327" + original_name = "linear_180.b_0" shape = [320] dtype = "float32" min_val = float("-1.8142") @@ -3580,6 +3908,7 @@ class Program_weight_tensor_parameter_327: class Program_weight_tensor_parameter_328: name = "parameter_328" + original_name = "linear_180.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.227791") @@ -3591,6 +3920,7 @@ class Program_weight_tensor_parameter_328: class Program_weight_tensor_parameter_329: name = "parameter_329" + original_name = "layer_norm_114.b_0" shape = [320] dtype = "float32" min_val = float("-0.974117") @@ -3602,6 +3932,7 @@ class Program_weight_tensor_parameter_329: class Program_weight_tensor_parameter_330: name = "parameter_330" + original_name = "layer_norm_114.w_0" shape = [320] dtype = "float32" min_val = float("0.0527976") @@ -3613,6 +3944,7 @@ class Program_weight_tensor_parameter_330: class Program_weight_tensor_parameter_331: name = "parameter_331" + original_name = "linear_179.b_0" shape = [320] dtype = "float32" min_val = float("-0.226132") @@ -3624,6 +3956,7 @@ class Program_weight_tensor_parameter_331: class Program_weight_tensor_parameter_332: name = "parameter_332" + original_name = "linear_179.w_0" shape = [1280, 320] dtype = "float32" min_val = float("-0.95069") @@ -3635,6 +3968,7 @@ class Program_weight_tensor_parameter_332: class Program_weight_tensor_parameter_333: name = "parameter_333" + original_name = "conv2d_75.b_0" shape = [1280] dtype = "float32" min_val = float("-1.21542") @@ -3646,6 +3980,7 @@ class Program_weight_tensor_parameter_333: class Program_weight_tensor_parameter_334: name = "parameter_334" + original_name = "conv2d_75.w_0" shape = [1280, 1, 3, 3] dtype = "float32" min_val = float("-0.296052") @@ -3657,6 +3992,7 @@ class Program_weight_tensor_parameter_334: class Program_weight_tensor_parameter_335: name = "parameter_335" + original_name = "linear_178.b_0" shape = [1280] dtype = "float32" min_val = float("-4.22012") @@ -3668,6 +4004,7 @@ class Program_weight_tensor_parameter_335: class Program_weight_tensor_parameter_336: name = "parameter_336" + original_name = "linear_178.w_0" shape = [320, 1280] dtype = "float32" min_val = float("-0.364201") @@ -3679,6 +4016,7 @@ class Program_weight_tensor_parameter_336: class Program_weight_tensor_parameter_337: name = "parameter_337" + original_name = "layer_norm_113.b_0" shape = [320] dtype = "float32" min_val = float("-1.24758") @@ -3690,6 +4028,7 @@ class Program_weight_tensor_parameter_337: class Program_weight_tensor_parameter_338: name = "parameter_338" + original_name = "layer_norm_113.w_0" shape = [320] dtype = "float32" min_val = float("0.00797244") @@ -3701,6 +4040,7 @@ class Program_weight_tensor_parameter_338: class Program_weight_tensor_parameter_339: name = "parameter_339" + original_name = "linear_177.b_0" shape = [320] dtype = "float32" min_val = float("-0.147043") @@ -3712,6 +4052,7 @@ class Program_weight_tensor_parameter_339: class Program_weight_tensor_parameter_340: name = "parameter_340" + original_name = "linear_177.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.69547") @@ -3723,6 +4064,7 @@ class Program_weight_tensor_parameter_340: class Program_weight_tensor_parameter_341: name = "parameter_341" + original_name = "linear_176.b_0" shape = [640] dtype = "float32" min_val = float("-1.58093") @@ -3734,6 +4076,7 @@ class Program_weight_tensor_parameter_341: class Program_weight_tensor_parameter_342: name = "parameter_342" + original_name = "linear_176.w_0" shape = [320, 640] dtype = "float32" min_val = float("-0.449067") @@ -3745,6 +4088,7 @@ class Program_weight_tensor_parameter_342: class Program_weight_tensor_parameter_343: name = "parameter_343" + original_name = "layer_norm_112.b_0" shape = [320] dtype = "float32" min_val = float("-0.154203") @@ -3756,6 +4100,7 @@ class Program_weight_tensor_parameter_343: class Program_weight_tensor_parameter_344: name = "parameter_344" + original_name = "layer_norm_112.w_0" shape = [320] dtype = "float32" min_val = float("-0.0408572") @@ -3767,6 +4112,7 @@ class Program_weight_tensor_parameter_344: class Program_weight_tensor_parameter_345: name = "parameter_345" + original_name = "conv2d_74.b_0" shape = [320] dtype = "float32" min_val = float("-5.84315") @@ -3778,6 +4124,7 @@ class Program_weight_tensor_parameter_345: class Program_weight_tensor_parameter_346: name = "parameter_346" + original_name = "conv2d_74.w_0" shape = [320, 320, 2, 2] dtype = "float32" min_val = float("-0.388103") @@ -3789,6 +4136,7 @@ class Program_weight_tensor_parameter_346: class Program_weight_tensor_parameter_347: name = "parameter_347" + original_name = "linear_175.b_0" shape = [320] dtype = "float32" min_val = float("-2.1344") @@ -3800,6 +4148,7 @@ class Program_weight_tensor_parameter_347: class Program_weight_tensor_parameter_348: name = "parameter_348" + original_name = "linear_175.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.281586") @@ -3811,6 +4160,7 @@ class Program_weight_tensor_parameter_348: class Program_weight_tensor_parameter_349: name = "parameter_349" + original_name = "layer_norm_111.b_0" shape = [320] dtype = "float32" min_val = float("-0.750333") @@ -3822,6 +4172,7 @@ class Program_weight_tensor_parameter_349: class Program_weight_tensor_parameter_350: name = "parameter_350" + original_name = "layer_norm_111.w_0" shape = [320] dtype = "float32" min_val = float("0.103254") @@ -3833,6 +4184,7 @@ class Program_weight_tensor_parameter_350: class Program_weight_tensor_parameter_351: name = "parameter_351" + original_name = "linear_174.b_0" shape = [320] dtype = "float32" min_val = float("-0.193571") @@ -3844,6 +4196,7 @@ class Program_weight_tensor_parameter_351: class Program_weight_tensor_parameter_352: name = "parameter_352" + original_name = "linear_174.w_0" shape = [1280, 320] dtype = "float32" min_val = float("-0.741013") @@ -3855,6 +4208,7 @@ class Program_weight_tensor_parameter_352: class Program_weight_tensor_parameter_353: name = "parameter_353" + original_name = "conv2d_73.b_0" shape = [1280] dtype = "float32" min_val = float("-1.1876") @@ -3866,6 +4220,7 @@ class Program_weight_tensor_parameter_353: class Program_weight_tensor_parameter_354: name = "parameter_354" + original_name = "conv2d_73.w_0" shape = [1280, 1, 3, 3] dtype = "float32" min_val = float("-0.307003") @@ -3877,6 +4232,7 @@ class Program_weight_tensor_parameter_354: class Program_weight_tensor_parameter_355: name = "parameter_355" + original_name = "linear_173.b_0" shape = [1280] dtype = "float32" min_val = float("-1.96735") @@ -3888,6 +4244,7 @@ class Program_weight_tensor_parameter_355: class Program_weight_tensor_parameter_356: name = "parameter_356" + original_name = "linear_173.w_0" shape = [320, 1280] dtype = "float32" min_val = float("-0.389249") @@ -3899,6 +4256,7 @@ class Program_weight_tensor_parameter_356: class Program_weight_tensor_parameter_357: name = "parameter_357" + original_name = "layer_norm_110.b_0" shape = [320] dtype = "float32" min_val = float("-1.20314") @@ -3910,6 +4268,7 @@ class Program_weight_tensor_parameter_357: class Program_weight_tensor_parameter_358: name = "parameter_358" + original_name = "layer_norm_110.w_0" shape = [320] dtype = "float32" min_val = float("0.00972839") @@ -3921,6 +4280,7 @@ class Program_weight_tensor_parameter_358: class Program_weight_tensor_parameter_359: name = "parameter_359" + original_name = "linear_172.b_0" shape = [320] dtype = "float32" min_val = float("-0.215166") @@ -3932,6 +4292,7 @@ class Program_weight_tensor_parameter_359: class Program_weight_tensor_parameter_360: name = "parameter_360" + original_name = "linear_172.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.315986") @@ -3943,6 +4304,7 @@ class Program_weight_tensor_parameter_360: class Program_weight_tensor_parameter_361: name = "parameter_361" + original_name = "linear_171.b_0" shape = [640] dtype = "float32" min_val = float("-2.15602") @@ -3954,6 +4316,7 @@ class Program_weight_tensor_parameter_361: class Program_weight_tensor_parameter_362: name = "parameter_362" + original_name = "linear_171.w_0" shape = [320, 640] dtype = "float32" min_val = float("-0.406242") @@ -3965,6 +4328,7 @@ class Program_weight_tensor_parameter_362: class Program_weight_tensor_parameter_363: name = "parameter_363" + original_name = "layer_norm_109.b_0" shape = [320] dtype = "float32" min_val = float("-0.192503") @@ -3976,6 +4340,7 @@ class Program_weight_tensor_parameter_363: class Program_weight_tensor_parameter_364: name = "parameter_364" + original_name = "layer_norm_109.w_0" shape = [320] dtype = "float32" min_val = float("-0.00619446") @@ -3987,6 +4352,7 @@ class Program_weight_tensor_parameter_364: class Program_weight_tensor_parameter_365: name = "parameter_365" + original_name = "conv2d_72.b_0" shape = [320] dtype = "float32" min_val = float("-6.30788") @@ -3998,6 +4364,7 @@ class Program_weight_tensor_parameter_365: class Program_weight_tensor_parameter_366: name = "parameter_366" + original_name = "conv2d_72.w_0" shape = [320, 320, 2, 2] dtype = "float32" min_val = float("-0.659098") @@ -4009,6 +4376,7 @@ class Program_weight_tensor_parameter_366: class Program_weight_tensor_parameter_367: name = "parameter_367" + original_name = "linear_170.b_0" shape = [320] dtype = "float32" min_val = float("-2.25676") @@ -4020,6 +4388,7 @@ class Program_weight_tensor_parameter_367: class Program_weight_tensor_parameter_368: name = "parameter_368" + original_name = "linear_170.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.276674") @@ -4031,6 +4400,7 @@ class Program_weight_tensor_parameter_368: class Program_weight_tensor_parameter_369: name = "parameter_369" + original_name = "layer_norm_108.b_0" shape = [320] dtype = "float32" min_val = float("-0.975506") @@ -4042,6 +4412,7 @@ class Program_weight_tensor_parameter_369: class Program_weight_tensor_parameter_370: name = "parameter_370" + original_name = "layer_norm_108.w_0" shape = [320] dtype = "float32" min_val = float("0.0376605") @@ -4053,6 +4424,7 @@ class Program_weight_tensor_parameter_370: class Program_weight_tensor_parameter_371: name = "parameter_371" + original_name = "linear_169.b_0" shape = [320] dtype = "float32" min_val = float("-0.229479") @@ -4064,6 +4436,7 @@ class Program_weight_tensor_parameter_371: class Program_weight_tensor_parameter_372: name = "parameter_372" + original_name = "linear_169.w_0" shape = [1280, 320] dtype = "float32" min_val = float("-0.645613") @@ -4075,6 +4448,7 @@ class Program_weight_tensor_parameter_372: class Program_weight_tensor_parameter_373: name = "parameter_373" + original_name = "conv2d_71.b_0" shape = [1280] dtype = "float32" min_val = float("-1.26381") @@ -4086,6 +4460,7 @@ class Program_weight_tensor_parameter_373: class Program_weight_tensor_parameter_374: name = "parameter_374" + original_name = "conv2d_71.w_0" shape = [1280, 1, 3, 3] dtype = "float32" min_val = float("-0.334479") @@ -4097,6 +4472,7 @@ class Program_weight_tensor_parameter_374: class Program_weight_tensor_parameter_375: name = "parameter_375" + original_name = "linear_168.b_0" shape = [1280] dtype = "float32" min_val = float("-2.61992") @@ -4108,6 +4484,7 @@ class Program_weight_tensor_parameter_375: class Program_weight_tensor_parameter_376: name = "parameter_376" + original_name = "linear_168.w_0" shape = [320, 1280] dtype = "float32" min_val = float("-0.643768") @@ -4119,6 +4496,7 @@ class Program_weight_tensor_parameter_376: class Program_weight_tensor_parameter_377: name = "parameter_377" + original_name = "layer_norm_107.b_0" shape = [320] dtype = "float32" min_val = float("-1.35127") @@ -4130,6 +4508,7 @@ class Program_weight_tensor_parameter_377: class Program_weight_tensor_parameter_378: name = "parameter_378" + original_name = "layer_norm_107.w_0" shape = [320] dtype = "float32" min_val = float("0.229739") @@ -4141,6 +4520,7 @@ class Program_weight_tensor_parameter_378: class Program_weight_tensor_parameter_379: name = "parameter_379" + original_name = "linear_167.b_0" shape = [320] dtype = "float32" min_val = float("-0.188353") @@ -4152,6 +4532,7 @@ class Program_weight_tensor_parameter_379: class Program_weight_tensor_parameter_380: name = "parameter_380" + original_name = "linear_167.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.699592") @@ -4163,6 +4544,7 @@ class Program_weight_tensor_parameter_380: class Program_weight_tensor_parameter_381: name = "parameter_381" + original_name = "linear_166.b_0" shape = [640] dtype = "float32" min_val = float("-2.82314") @@ -4174,6 +4556,7 @@ class Program_weight_tensor_parameter_381: class Program_weight_tensor_parameter_382: name = "parameter_382" + original_name = "linear_166.w_0" shape = [320, 640] dtype = "float32" min_val = float("-0.416528") @@ -4185,6 +4568,7 @@ class Program_weight_tensor_parameter_382: class Program_weight_tensor_parameter_383: name = "parameter_383" + original_name = "layer_norm_106.b_0" shape = [320] dtype = "float32" min_val = float("-0.243611") @@ -4196,6 +4580,7 @@ class Program_weight_tensor_parameter_383: class Program_weight_tensor_parameter_384: name = "parameter_384" + original_name = "layer_norm_106.w_0" shape = [320] dtype = "float32" min_val = float("-0.00955912") @@ -4207,6 +4592,7 @@ class Program_weight_tensor_parameter_384: class Program_weight_tensor_parameter_385: name = "parameter_385" + original_name = "conv2d_70.b_0" shape = [320] dtype = "float32" min_val = float("-4.63872") @@ -4218,6 +4604,7 @@ class Program_weight_tensor_parameter_385: class Program_weight_tensor_parameter_386: name = "parameter_386" + original_name = "conv2d_70.w_0" shape = [320, 320, 2, 2] dtype = "float32" min_val = float("-0.473943") @@ -4229,6 +4616,7 @@ class Program_weight_tensor_parameter_386: class Program_weight_tensor_parameter_387: name = "parameter_387" + original_name = "linear_165.b_0" shape = [320] dtype = "float32" min_val = float("-2.2453") @@ -4240,6 +4628,7 @@ class Program_weight_tensor_parameter_387: class Program_weight_tensor_parameter_388: name = "parameter_388" + original_name = "linear_165.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.354477") @@ -4251,6 +4640,7 @@ class Program_weight_tensor_parameter_388: class Program_weight_tensor_parameter_389: name = "parameter_389" + original_name = "layer_norm_105.b_0" shape = [320] dtype = "float32" min_val = float("-1.01505") @@ -4262,6 +4652,7 @@ class Program_weight_tensor_parameter_389: class Program_weight_tensor_parameter_390: name = "parameter_390" + original_name = "layer_norm_105.w_0" shape = [320] dtype = "float32" min_val = float("0.0826408") @@ -4273,6 +4664,7 @@ class Program_weight_tensor_parameter_390: class Program_weight_tensor_parameter_391: name = "parameter_391" + original_name = "linear_164.b_0" shape = [320] dtype = "float32" min_val = float("-0.192114") @@ -4284,6 +4676,7 @@ class Program_weight_tensor_parameter_391: class Program_weight_tensor_parameter_392: name = "parameter_392" + original_name = "linear_164.w_0" shape = [1280, 320] dtype = "float32" min_val = float("-0.994763") @@ -4295,6 +4688,7 @@ class Program_weight_tensor_parameter_392: class Program_weight_tensor_parameter_393: name = "parameter_393" + original_name = "conv2d_69.b_0" shape = [1280] dtype = "float32" min_val = float("-1.35907") @@ -4306,6 +4700,7 @@ class Program_weight_tensor_parameter_393: class Program_weight_tensor_parameter_394: name = "parameter_394" + original_name = "conv2d_69.w_0" shape = [1280, 1, 3, 3] dtype = "float32" min_val = float("-1.19975") @@ -4317,6 +4712,7 @@ class Program_weight_tensor_parameter_394: class Program_weight_tensor_parameter_395: name = "parameter_395" + original_name = "linear_163.b_0" shape = [1280] dtype = "float32" min_val = float("-3.30917") @@ -4328,6 +4724,7 @@ class Program_weight_tensor_parameter_395: class Program_weight_tensor_parameter_396: name = "parameter_396" + original_name = "linear_163.w_0" shape = [320, 1280] dtype = "float32" min_val = float("-0.710841") @@ -4339,6 +4736,7 @@ class Program_weight_tensor_parameter_396: class Program_weight_tensor_parameter_397: name = "parameter_397" + original_name = "layer_norm_104.b_0" shape = [320] dtype = "float32" min_val = float("-1.43508") @@ -4350,6 +4748,7 @@ class Program_weight_tensor_parameter_397: class Program_weight_tensor_parameter_398: name = "parameter_398" + original_name = "layer_norm_104.w_0" shape = [320] dtype = "float32" min_val = float("0.322102") @@ -4361,6 +4760,7 @@ class Program_weight_tensor_parameter_398: class Program_weight_tensor_parameter_399: name = "parameter_399" + original_name = "linear_162.b_0" shape = [320] dtype = "float32" min_val = float("-0.213885") @@ -4372,6 +4772,7 @@ class Program_weight_tensor_parameter_399: class Program_weight_tensor_parameter_400: name = "parameter_400" + original_name = "linear_162.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.759889") @@ -4383,6 +4784,7 @@ class Program_weight_tensor_parameter_400: class Program_weight_tensor_parameter_401: name = "parameter_401" + original_name = "linear_161.b_0" shape = [640] dtype = "float32" min_val = float("-3.08885") @@ -4394,6 +4796,7 @@ class Program_weight_tensor_parameter_401: class Program_weight_tensor_parameter_402: name = "parameter_402" + original_name = "linear_161.w_0" shape = [320, 640] dtype = "float32" min_val = float("-0.497132") @@ -4405,6 +4808,7 @@ class Program_weight_tensor_parameter_402: class Program_weight_tensor_parameter_403: name = "parameter_403" + original_name = "layer_norm_103.b_0" shape = [320] dtype = "float32" min_val = float("-0.255236") @@ -4416,6 +4820,7 @@ class Program_weight_tensor_parameter_403: class Program_weight_tensor_parameter_404: name = "parameter_404" + original_name = "layer_norm_103.w_0" shape = [320] dtype = "float32" min_val = float("-0.00836209") @@ -4427,6 +4832,7 @@ class Program_weight_tensor_parameter_404: class Program_weight_tensor_parameter_405: name = "parameter_405" + original_name = "conv2d_68.b_0" shape = [320] dtype = "float32" min_val = float("-5.75582") @@ -4438,6 +4844,7 @@ class Program_weight_tensor_parameter_405: class Program_weight_tensor_parameter_406: name = "parameter_406" + original_name = "conv2d_68.w_0" shape = [320, 320, 2, 2] dtype = "float32" min_val = float("-0.5956") @@ -4449,6 +4856,7 @@ class Program_weight_tensor_parameter_406: class Program_weight_tensor_parameter_407: name = "parameter_407" + original_name = "linear_160.b_0" shape = [320] dtype = "float32" min_val = float("-2.87259") @@ -4460,6 +4868,7 @@ class Program_weight_tensor_parameter_407: class Program_weight_tensor_parameter_408: name = "parameter_408" + original_name = "linear_160.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.501985") @@ -4471,6 +4880,7 @@ class Program_weight_tensor_parameter_408: class Program_weight_tensor_parameter_409: name = "parameter_409" + original_name = "layer_norm_102.b_0" shape = [320] dtype = "float32" min_val = float("-0.873343") @@ -4482,6 +4892,7 @@ class Program_weight_tensor_parameter_409: class Program_weight_tensor_parameter_410: name = "parameter_410" + original_name = "layer_norm_102.w_0" shape = [320] dtype = "float32" min_val = float("0.15589") @@ -4493,6 +4904,7 @@ class Program_weight_tensor_parameter_410: class Program_weight_tensor_parameter_411: name = "parameter_411" + original_name = "linear_159.b_0" shape = [320] dtype = "float32" min_val = float("-0.261939") @@ -4504,6 +4916,7 @@ class Program_weight_tensor_parameter_411: class Program_weight_tensor_parameter_412: name = "parameter_412" + original_name = "linear_159.w_0" shape = [1280, 320] dtype = "float32" min_val = float("-0.779811") @@ -4515,6 +4928,7 @@ class Program_weight_tensor_parameter_412: class Program_weight_tensor_parameter_413: name = "parameter_413" + original_name = "conv2d_67.b_0" shape = [1280] dtype = "float32" min_val = float("-1.36533") @@ -4526,6 +4940,7 @@ class Program_weight_tensor_parameter_413: class Program_weight_tensor_parameter_414: name = "parameter_414" + original_name = "conv2d_67.w_0" shape = [1280, 1, 3, 3] dtype = "float32" min_val = float("-0.977371") @@ -4537,6 +4952,7 @@ class Program_weight_tensor_parameter_414: class Program_weight_tensor_parameter_415: name = "parameter_415" + original_name = "linear_158.b_0" shape = [1280] dtype = "float32" min_val = float("-4.06971") @@ -4548,6 +4964,7 @@ class Program_weight_tensor_parameter_415: class Program_weight_tensor_parameter_416: name = "parameter_416" + original_name = "linear_158.w_0" shape = [320, 1280] dtype = "float32" min_val = float("-0.493097") @@ -4559,6 +4976,7 @@ class Program_weight_tensor_parameter_416: class Program_weight_tensor_parameter_417: name = "parameter_417" + original_name = "layer_norm_101.b_0" shape = [320] dtype = "float32" min_val = float("-1.2547") @@ -4570,6 +4988,7 @@ class Program_weight_tensor_parameter_417: class Program_weight_tensor_parameter_418: name = "parameter_418" + original_name = "layer_norm_101.w_0" shape = [320] dtype = "float32" min_val = float("0.417345") @@ -4581,6 +5000,7 @@ class Program_weight_tensor_parameter_418: class Program_weight_tensor_parameter_419: name = "parameter_419" + original_name = "linear_157.b_0" shape = [320] dtype = "float32" min_val = float("-0.180316") @@ -4592,6 +5012,7 @@ class Program_weight_tensor_parameter_419: class Program_weight_tensor_parameter_420: name = "parameter_420" + original_name = "linear_157.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.678071") @@ -4603,6 +5024,7 @@ class Program_weight_tensor_parameter_420: class Program_weight_tensor_parameter_421: name = "parameter_421" + original_name = "linear_156.b_0" shape = [640] dtype = "float32" min_val = float("-2.07446") @@ -4614,6 +5036,7 @@ class Program_weight_tensor_parameter_421: class Program_weight_tensor_parameter_422: name = "parameter_422" + original_name = "linear_156.w_0" shape = [320, 640] dtype = "float32" min_val = float("-0.411652") @@ -4625,6 +5048,7 @@ class Program_weight_tensor_parameter_422: class Program_weight_tensor_parameter_423: name = "parameter_423" + original_name = "layer_norm_100.b_0" shape = [320] dtype = "float32" min_val = float("-0.191782") @@ -4636,6 +5060,7 @@ class Program_weight_tensor_parameter_423: class Program_weight_tensor_parameter_424: name = "parameter_424" + original_name = "layer_norm_100.w_0" shape = [320] dtype = "float32" min_val = float("-0.0909807") @@ -4647,6 +5072,7 @@ class Program_weight_tensor_parameter_424: class Program_weight_tensor_parameter_425: name = "parameter_425" + original_name = "conv2d_66.b_0" shape = [320] dtype = "float32" min_val = float("-4.71109") @@ -4658,6 +5084,7 @@ class Program_weight_tensor_parameter_425: class Program_weight_tensor_parameter_426: name = "parameter_426" + original_name = "conv2d_66.w_0" shape = [320, 320, 2, 2] dtype = "float32" min_val = float("-0.424253") @@ -4669,6 +5096,7 @@ class Program_weight_tensor_parameter_426: class Program_weight_tensor_parameter_427: name = "parameter_427" + original_name = "linear_155.b_0" shape = [320] dtype = "float32" min_val = float("-2.25098") @@ -4680,6 +5108,7 @@ class Program_weight_tensor_parameter_427: class Program_weight_tensor_parameter_428: name = "parameter_428" + original_name = "linear_155.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.456915") @@ -4691,6 +5120,7 @@ class Program_weight_tensor_parameter_428: class Program_weight_tensor_parameter_429: name = "parameter_429" + original_name = "layer_norm_99.b_0" shape = [320] dtype = "float32" min_val = float("-0.997252") @@ -4702,6 +5132,7 @@ class Program_weight_tensor_parameter_429: class Program_weight_tensor_parameter_430: name = "parameter_430" + original_name = "layer_norm_99.w_0" shape = [320] dtype = "float32" min_val = float("0.154874") @@ -4713,6 +5144,7 @@ class Program_weight_tensor_parameter_430: class Program_weight_tensor_parameter_431: name = "parameter_431" + original_name = "linear_154.b_0" shape = [320] dtype = "float32" min_val = float("-0.233491") @@ -4724,6 +5156,7 @@ class Program_weight_tensor_parameter_431: class Program_weight_tensor_parameter_432: name = "parameter_432" + original_name = "linear_154.w_0" shape = [1280, 320] dtype = "float32" min_val = float("-0.866717") @@ -4735,6 +5168,7 @@ class Program_weight_tensor_parameter_432: class Program_weight_tensor_parameter_433: name = "parameter_433" + original_name = "conv2d_65.b_0" shape = [1280] dtype = "float32" min_val = float("-1.5845") @@ -4746,6 +5180,7 @@ class Program_weight_tensor_parameter_433: class Program_weight_tensor_parameter_434: name = "parameter_434" + original_name = "conv2d_65.w_0" shape = [1280, 1, 3, 3] dtype = "float32" min_val = float("-0.947925") @@ -4757,6 +5192,7 @@ class Program_weight_tensor_parameter_434: class Program_weight_tensor_parameter_435: name = "parameter_435" + original_name = "linear_153.b_0" shape = [1280] dtype = "float32" min_val = float("-1.08056") @@ -4768,6 +5204,7 @@ class Program_weight_tensor_parameter_435: class Program_weight_tensor_parameter_436: name = "parameter_436" + original_name = "linear_153.w_0" shape = [320, 1280] dtype = "float32" min_val = float("-0.706538") @@ -4779,6 +5216,7 @@ class Program_weight_tensor_parameter_436: class Program_weight_tensor_parameter_437: name = "parameter_437" + original_name = "layer_norm_98.b_0" shape = [320] dtype = "float32" min_val = float("-1.2775") @@ -4790,6 +5228,7 @@ class Program_weight_tensor_parameter_437: class Program_weight_tensor_parameter_438: name = "parameter_438" + original_name = "layer_norm_98.w_0" shape = [320] dtype = "float32" min_val = float("0.414722") @@ -4801,6 +5240,7 @@ class Program_weight_tensor_parameter_438: class Program_weight_tensor_parameter_439: name = "parameter_439" + original_name = "linear_152.b_0" shape = [320] dtype = "float32" min_val = float("-0.17718") @@ -4812,6 +5252,7 @@ class Program_weight_tensor_parameter_439: class Program_weight_tensor_parameter_440: name = "parameter_440" + original_name = "linear_152.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.713343") @@ -4823,6 +5264,7 @@ class Program_weight_tensor_parameter_440: class Program_weight_tensor_parameter_441: name = "parameter_441" + original_name = "linear_151.b_0" shape = [640] dtype = "float32" min_val = float("-1.672") @@ -4834,6 +5276,7 @@ class Program_weight_tensor_parameter_441: class Program_weight_tensor_parameter_442: name = "parameter_442" + original_name = "linear_151.w_0" shape = [320, 640] dtype = "float32" min_val = float("-0.441513") @@ -4845,6 +5288,7 @@ class Program_weight_tensor_parameter_442: class Program_weight_tensor_parameter_443: name = "parameter_443" + original_name = "layer_norm_97.b_0" shape = [320] dtype = "float32" min_val = float("-0.250176") @@ -4856,6 +5300,7 @@ class Program_weight_tensor_parameter_443: class Program_weight_tensor_parameter_444: name = "parameter_444" + original_name = "layer_norm_97.w_0" shape = [320] dtype = "float32" min_val = float("-0.0776005") @@ -4867,6 +5312,7 @@ class Program_weight_tensor_parameter_444: class Program_weight_tensor_parameter_445: name = "parameter_445" + original_name = "conv2d_64.b_0" shape = [320] dtype = "float32" min_val = float("-5.25639") @@ -4878,6 +5324,7 @@ class Program_weight_tensor_parameter_445: class Program_weight_tensor_parameter_446: name = "parameter_446" + original_name = "conv2d_64.w_0" shape = [320, 320, 2, 2] dtype = "float32" min_val = float("-0.390908") @@ -4889,6 +5336,7 @@ class Program_weight_tensor_parameter_446: class Program_weight_tensor_parameter_447: name = "parameter_447" + original_name = "linear_150.b_0" shape = [320] dtype = "float32" min_val = float("-2.27542") @@ -4900,6 +5348,7 @@ class Program_weight_tensor_parameter_447: class Program_weight_tensor_parameter_448: name = "parameter_448" + original_name = "linear_150.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.300445") @@ -4911,6 +5360,7 @@ class Program_weight_tensor_parameter_448: class Program_weight_tensor_parameter_449: name = "parameter_449" + original_name = "layer_norm_96.b_0" shape = [320] dtype = "float32" min_val = float("-0.995535") @@ -4922,6 +5372,7 @@ class Program_weight_tensor_parameter_449: class Program_weight_tensor_parameter_450: name = "parameter_450" + original_name = "layer_norm_96.w_0" shape = [320] dtype = "float32" min_val = float("0.185146") @@ -4933,6 +5384,7 @@ class Program_weight_tensor_parameter_450: class Program_weight_tensor_parameter_451: name = "parameter_451" + original_name = "linear_149.b_0" shape = [320] dtype = "float32" min_val = float("-0.291246") @@ -4944,6 +5396,7 @@ class Program_weight_tensor_parameter_451: class Program_weight_tensor_parameter_452: name = "parameter_452" + original_name = "linear_149.w_0" shape = [1280, 320] dtype = "float32" min_val = float("-0.622104") @@ -4955,6 +5408,7 @@ class Program_weight_tensor_parameter_452: class Program_weight_tensor_parameter_453: name = "parameter_453" + original_name = "conv2d_63.b_0" shape = [1280] dtype = "float32" min_val = float("-1.97231") @@ -4966,6 +5420,7 @@ class Program_weight_tensor_parameter_453: class Program_weight_tensor_parameter_454: name = "parameter_454" + original_name = "conv2d_63.w_0" shape = [1280, 1, 3, 3] dtype = "float32" min_val = float("-0.286697") @@ -4977,6 +5432,7 @@ class Program_weight_tensor_parameter_454: class Program_weight_tensor_parameter_455: name = "parameter_455" + original_name = "linear_148.b_0" shape = [1280] dtype = "float32" min_val = float("-1.91434") @@ -4988,6 +5444,7 @@ class Program_weight_tensor_parameter_455: class Program_weight_tensor_parameter_456: name = "parameter_456" + original_name = "linear_148.w_0" shape = [320, 1280] dtype = "float32" min_val = float("-0.552608") @@ -4999,6 +5456,7 @@ class Program_weight_tensor_parameter_456: class Program_weight_tensor_parameter_457: name = "parameter_457" + original_name = "layer_norm_95.b_0" shape = [320] dtype = "float32" min_val = float("-1.46989") @@ -5010,6 +5468,7 @@ class Program_weight_tensor_parameter_457: class Program_weight_tensor_parameter_458: name = "parameter_458" + original_name = "layer_norm_95.w_0" shape = [320] dtype = "float32" min_val = float("0.541299") @@ -5021,6 +5480,7 @@ class Program_weight_tensor_parameter_458: class Program_weight_tensor_parameter_459: name = "parameter_459" + original_name = "linear_147.b_0" shape = [320] dtype = "float32" min_val = float("-0.200693") @@ -5032,6 +5492,7 @@ class Program_weight_tensor_parameter_459: class Program_weight_tensor_parameter_460: name = "parameter_460" + original_name = "linear_147.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.469849") @@ -5043,6 +5504,7 @@ class Program_weight_tensor_parameter_460: class Program_weight_tensor_parameter_461: name = "parameter_461" + original_name = "linear_146.b_0" shape = [640] dtype = "float32" min_val = float("-1.46257") @@ -5054,6 +5516,7 @@ class Program_weight_tensor_parameter_461: class Program_weight_tensor_parameter_462: name = "parameter_462" + original_name = "linear_146.w_0" shape = [320, 640] dtype = "float32" min_val = float("-0.438637") @@ -5065,6 +5528,7 @@ class Program_weight_tensor_parameter_462: class Program_weight_tensor_parameter_463: name = "parameter_463" + original_name = "layer_norm_94.b_0" shape = [320] dtype = "float32" min_val = float("-0.185637") @@ -5076,6 +5540,7 @@ class Program_weight_tensor_parameter_463: class Program_weight_tensor_parameter_464: name = "parameter_464" + original_name = "layer_norm_94.w_0" shape = [320] dtype = "float32" min_val = float("-0.0149624") @@ -5087,6 +5552,7 @@ class Program_weight_tensor_parameter_464: class Program_weight_tensor_parameter_465: name = "parameter_465" + original_name = "conv2d_62.b_0" shape = [320] dtype = "float32" min_val = float("-5.61338") @@ -5098,6 +5564,7 @@ class Program_weight_tensor_parameter_465: class Program_weight_tensor_parameter_466: name = "parameter_466" + original_name = "conv2d_62.w_0" shape = [320, 320, 2, 2] dtype = "float32" min_val = float("-0.670348") @@ -5109,6 +5576,7 @@ class Program_weight_tensor_parameter_466: class Program_weight_tensor_parameter_467: name = "parameter_467" + original_name = "linear_145.b_0" shape = [320] dtype = "float32" min_val = float("-2.74093") @@ -5120,6 +5588,7 @@ class Program_weight_tensor_parameter_467: class Program_weight_tensor_parameter_468: name = "parameter_468" + original_name = "linear_145.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.292831") @@ -5131,6 +5600,7 @@ class Program_weight_tensor_parameter_468: class Program_weight_tensor_parameter_469: name = "parameter_469" + original_name = "layer_norm_93.b_0" shape = [320] dtype = "float32" min_val = float("-0.956971") @@ -5142,6 +5612,7 @@ class Program_weight_tensor_parameter_469: class Program_weight_tensor_parameter_470: name = "parameter_470" + original_name = "layer_norm_93.w_0" shape = [320] dtype = "float32" min_val = float("0.00380617") @@ -5153,6 +5624,7 @@ class Program_weight_tensor_parameter_470: class Program_weight_tensor_parameter_471: name = "parameter_471" + original_name = "linear_144.b_0" shape = [320] dtype = "float32" min_val = float("-0.259801") @@ -5164,6 +5636,7 @@ class Program_weight_tensor_parameter_471: class Program_weight_tensor_parameter_472: name = "parameter_472" + original_name = "linear_144.w_0" shape = [1280, 320] dtype = "float32" min_val = float("-0.71566") @@ -5175,6 +5648,7 @@ class Program_weight_tensor_parameter_472: class Program_weight_tensor_parameter_473: name = "parameter_473" + original_name = "conv2d_61.b_0" shape = [1280] dtype = "float32" min_val = float("-1.37339") @@ -5186,6 +5660,7 @@ class Program_weight_tensor_parameter_473: class Program_weight_tensor_parameter_474: name = "parameter_474" + original_name = "conv2d_61.w_0" shape = [1280, 1, 3, 3] dtype = "float32" min_val = float("-0.525885") @@ -5197,6 +5672,7 @@ class Program_weight_tensor_parameter_474: class Program_weight_tensor_parameter_475: name = "parameter_475" + original_name = "linear_143.b_0" shape = [1280] dtype = "float32" min_val = float("-1.66202") @@ -5208,6 +5684,7 @@ class Program_weight_tensor_parameter_475: class Program_weight_tensor_parameter_476: name = "parameter_476" + original_name = "linear_143.w_0" shape = [320, 1280] dtype = "float32" min_val = float("-0.584605") @@ -5219,6 +5696,7 @@ class Program_weight_tensor_parameter_476: class Program_weight_tensor_parameter_477: name = "parameter_477" + original_name = "layer_norm_92.b_0" shape = [320] dtype = "float32" min_val = float("-1.49156") @@ -5230,6 +5708,7 @@ class Program_weight_tensor_parameter_477: class Program_weight_tensor_parameter_478: name = "parameter_478" + original_name = "layer_norm_92.w_0" shape = [320] dtype = "float32" min_val = float("0.636852") @@ -5241,6 +5720,7 @@ class Program_weight_tensor_parameter_478: class Program_weight_tensor_parameter_479: name = "parameter_479" + original_name = "linear_142.b_0" shape = [320] dtype = "float32" min_val = float("-0.201149") @@ -5252,6 +5732,7 @@ class Program_weight_tensor_parameter_479: class Program_weight_tensor_parameter_480: name = "parameter_480" + original_name = "linear_142.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.64879") @@ -5263,6 +5744,7 @@ class Program_weight_tensor_parameter_480: class Program_weight_tensor_parameter_481: name = "parameter_481" + original_name = "linear_141.b_0" shape = [640] dtype = "float32" min_val = float("-0.82927") @@ -5274,6 +5756,7 @@ class Program_weight_tensor_parameter_481: class Program_weight_tensor_parameter_482: name = "parameter_482" + original_name = "linear_141.w_0" shape = [320, 640] dtype = "float32" min_val = float("-0.383766") @@ -5285,6 +5768,7 @@ class Program_weight_tensor_parameter_482: class Program_weight_tensor_parameter_483: name = "parameter_483" + original_name = "layer_norm_91.b_0" shape = [320] dtype = "float32" min_val = float("-0.304469") @@ -5296,6 +5780,7 @@ class Program_weight_tensor_parameter_483: class Program_weight_tensor_parameter_484: name = "parameter_484" + original_name = "layer_norm_91.w_0" shape = [320] dtype = "float32" min_val = float("-0.121533") @@ -5307,6 +5792,7 @@ class Program_weight_tensor_parameter_484: class Program_weight_tensor_parameter_485: name = "parameter_485" + original_name = "conv2d_60.b_0" shape = [320] dtype = "float32" min_val = float("-4.97177") @@ -5318,6 +5804,7 @@ class Program_weight_tensor_parameter_485: class Program_weight_tensor_parameter_486: name = "parameter_486" + original_name = "conv2d_60.w_0" shape = [320, 320, 2, 2] dtype = "float32" min_val = float("-0.360474") @@ -5329,6 +5816,7 @@ class Program_weight_tensor_parameter_486: class Program_weight_tensor_parameter_487: name = "parameter_487" + original_name = "linear_140.b_0" shape = [320] dtype = "float32" min_val = float("-2.67235") @@ -5340,6 +5828,7 @@ class Program_weight_tensor_parameter_487: class Program_weight_tensor_parameter_488: name = "parameter_488" + original_name = "linear_140.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.299987") @@ -5351,6 +5840,7 @@ class Program_weight_tensor_parameter_488: class Program_weight_tensor_parameter_489: name = "parameter_489" + original_name = "layer_norm_90.b_0" shape = [320] dtype = "float32" min_val = float("-0.930084") @@ -5362,6 +5852,7 @@ class Program_weight_tensor_parameter_489: class Program_weight_tensor_parameter_490: name = "parameter_490" + original_name = "layer_norm_90.w_0" shape = [320] dtype = "float32" min_val = float("-0.389296") @@ -5373,6 +5864,7 @@ class Program_weight_tensor_parameter_490: class Program_weight_tensor_parameter_491: name = "parameter_491" + original_name = "linear_139.b_0" shape = [320] dtype = "float32" min_val = float("-0.24837") @@ -5384,6 +5876,7 @@ class Program_weight_tensor_parameter_491: class Program_weight_tensor_parameter_492: name = "parameter_492" + original_name = "linear_139.w_0" shape = [1280, 320] dtype = "float32" min_val = float("-0.797422") @@ -5395,6 +5888,7 @@ class Program_weight_tensor_parameter_492: class Program_weight_tensor_parameter_493: name = "parameter_493" + original_name = "conv2d_59.b_0" shape = [1280] dtype = "float32" min_val = float("-1.49328") @@ -5406,6 +5900,7 @@ class Program_weight_tensor_parameter_493: class Program_weight_tensor_parameter_494: name = "parameter_494" + original_name = "conv2d_59.w_0" shape = [1280, 1, 3, 3] dtype = "float32" min_val = float("-0.27773") @@ -5417,6 +5912,7 @@ class Program_weight_tensor_parameter_494: class Program_weight_tensor_parameter_495: name = "parameter_495" + original_name = "linear_138.b_0" shape = [1280] dtype = "float32" min_val = float("-1.30934") @@ -5428,6 +5924,7 @@ class Program_weight_tensor_parameter_495: class Program_weight_tensor_parameter_496: name = "parameter_496" + original_name = "linear_138.w_0" shape = [320, 1280] dtype = "float32" min_val = float("-0.604131") @@ -5439,6 +5936,7 @@ class Program_weight_tensor_parameter_496: class Program_weight_tensor_parameter_497: name = "parameter_497" + original_name = "layer_norm_89.b_0" shape = [320] dtype = "float32" min_val = float("-1.3826") @@ -5450,6 +5948,7 @@ class Program_weight_tensor_parameter_497: class Program_weight_tensor_parameter_498: name = "parameter_498" + original_name = "layer_norm_89.w_0" shape = [320] dtype = "float32" min_val = float("0.572649") @@ -5461,6 +5960,7 @@ class Program_weight_tensor_parameter_498: class Program_weight_tensor_parameter_499: name = "parameter_499" + original_name = "linear_137.b_0" shape = [320] dtype = "float32" min_val = float("-0.182027") @@ -5472,6 +5972,7 @@ class Program_weight_tensor_parameter_499: class Program_weight_tensor_parameter_500: name = "parameter_500" + original_name = "linear_137.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.849283") @@ -5483,6 +5984,7 @@ class Program_weight_tensor_parameter_500: class Program_weight_tensor_parameter_501: name = "parameter_501" + original_name = "linear_136.b_0" shape = [640] dtype = "float32" min_val = float("-1.14027") @@ -5494,6 +5996,7 @@ class Program_weight_tensor_parameter_501: class Program_weight_tensor_parameter_502: name = "parameter_502" + original_name = "linear_136.w_0" shape = [320, 640] dtype = "float32" min_val = float("-0.446958") @@ -5505,6 +6008,7 @@ class Program_weight_tensor_parameter_502: class Program_weight_tensor_parameter_503: name = "parameter_503" + original_name = "layer_norm_88.b_0" shape = [320] dtype = "float32" min_val = float("-0.154141") @@ -5516,6 +6020,7 @@ class Program_weight_tensor_parameter_503: class Program_weight_tensor_parameter_504: name = "parameter_504" + original_name = "layer_norm_88.w_0" shape = [320] dtype = "float32" min_val = float("-0.100034") @@ -5527,6 +6032,7 @@ class Program_weight_tensor_parameter_504: class Program_weight_tensor_parameter_505: name = "parameter_505" + original_name = "conv2d_58.b_0" shape = [320] dtype = "float32" min_val = float("-6.38854") @@ -5538,6 +6044,7 @@ class Program_weight_tensor_parameter_505: class Program_weight_tensor_parameter_506: name = "parameter_506" + original_name = "conv2d_58.w_0" shape = [320, 320, 2, 2] dtype = "float32" min_val = float("-0.640865") @@ -5549,6 +6056,7 @@ class Program_weight_tensor_parameter_506: class Program_weight_tensor_parameter_507: name = "parameter_507" + original_name = "linear_135.b_0" shape = [320] dtype = "float32" min_val = float("-3.13877") @@ -5560,6 +6068,7 @@ class Program_weight_tensor_parameter_507: class Program_weight_tensor_parameter_508: name = "parameter_508" + original_name = "linear_135.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.389922") @@ -5571,6 +6080,7 @@ class Program_weight_tensor_parameter_508: class Program_weight_tensor_parameter_509: name = "parameter_509" + original_name = "layer_norm_87.b_0" shape = [320] dtype = "float32" min_val = float("-0.909575") @@ -5582,6 +6092,7 @@ class Program_weight_tensor_parameter_509: class Program_weight_tensor_parameter_510: name = "parameter_510" + original_name = "layer_norm_87.w_0" shape = [320] dtype = "float32" min_val = float("0.126343") @@ -5593,6 +6104,7 @@ class Program_weight_tensor_parameter_510: class Program_weight_tensor_parameter_511: name = "parameter_511" + original_name = "linear_134.b_0" shape = [320] dtype = "float32" min_val = float("-0.282001") @@ -5604,6 +6116,7 @@ class Program_weight_tensor_parameter_511: class Program_weight_tensor_parameter_512: name = "parameter_512" + original_name = "linear_134.w_0" shape = [1280, 320] dtype = "float32" min_val = float("-0.775353") @@ -5615,6 +6128,7 @@ class Program_weight_tensor_parameter_512: class Program_weight_tensor_parameter_513: name = "parameter_513" + original_name = "conv2d_57.b_0" shape = [1280] dtype = "float32" min_val = float("-1.48315") @@ -5626,6 +6140,7 @@ class Program_weight_tensor_parameter_513: class Program_weight_tensor_parameter_514: name = "parameter_514" + original_name = "conv2d_57.w_0" shape = [1280, 1, 3, 3] dtype = "float32" min_val = float("-0.288438") @@ -5637,6 +6152,7 @@ class Program_weight_tensor_parameter_514: class Program_weight_tensor_parameter_515: name = "parameter_515" + original_name = "linear_133.b_0" shape = [1280] dtype = "float32" min_val = float("-3.35542") @@ -5648,6 +6164,7 @@ class Program_weight_tensor_parameter_515: class Program_weight_tensor_parameter_516: name = "parameter_516" + original_name = "linear_133.w_0" shape = [320, 1280] dtype = "float32" min_val = float("-0.655022") @@ -5659,6 +6176,7 @@ class Program_weight_tensor_parameter_516: class Program_weight_tensor_parameter_517: name = "parameter_517" + original_name = "layer_norm_86.b_0" shape = [320] dtype = "float32" min_val = float("-1.54124") @@ -5670,6 +6188,7 @@ class Program_weight_tensor_parameter_517: class Program_weight_tensor_parameter_518: name = "parameter_518" + original_name = "layer_norm_86.w_0" shape = [320] dtype = "float32" min_val = float("0.706658") @@ -5681,6 +6200,7 @@ class Program_weight_tensor_parameter_518: class Program_weight_tensor_parameter_519: name = "parameter_519" + original_name = "linear_132.b_0" shape = [320] dtype = "float32" min_val = float("-0.213499") @@ -5692,6 +6212,7 @@ class Program_weight_tensor_parameter_519: class Program_weight_tensor_parameter_520: name = "parameter_520" + original_name = "linear_132.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.718055") @@ -5703,6 +6224,7 @@ class Program_weight_tensor_parameter_520: class Program_weight_tensor_parameter_521: name = "parameter_521" + original_name = "linear_131.b_0" shape = [640] dtype = "float32" min_val = float("-1.3878") @@ -5714,6 +6236,7 @@ class Program_weight_tensor_parameter_521: class Program_weight_tensor_parameter_522: name = "parameter_522" + original_name = "linear_131.w_0" shape = [320, 640] dtype = "float32" min_val = float("-0.435789") @@ -5725,6 +6248,7 @@ class Program_weight_tensor_parameter_522: class Program_weight_tensor_parameter_523: name = "parameter_523" + original_name = "layer_norm_85.b_0" shape = [320] dtype = "float32" min_val = float("-0.220717") @@ -5736,6 +6260,7 @@ class Program_weight_tensor_parameter_523: class Program_weight_tensor_parameter_524: name = "parameter_524" + original_name = "layer_norm_85.w_0" shape = [320] dtype = "float32" min_val = float("-0.0865395") @@ -5747,6 +6272,7 @@ class Program_weight_tensor_parameter_524: class Program_weight_tensor_parameter_525: name = "parameter_525" + original_name = "conv2d_56.b_0" shape = [320] dtype = "float32" min_val = float("-4.11539") @@ -5758,6 +6284,7 @@ class Program_weight_tensor_parameter_525: class Program_weight_tensor_parameter_526: name = "parameter_526" + original_name = "conv2d_56.w_0" shape = [320, 320, 2, 2] dtype = "float32" min_val = float("-0.378583") @@ -5769,6 +6296,7 @@ class Program_weight_tensor_parameter_526: class Program_weight_tensor_parameter_527: name = "parameter_527" + original_name = "linear_130.b_0" shape = [320] dtype = "float32" min_val = float("-2.23921") @@ -5780,6 +6308,7 @@ class Program_weight_tensor_parameter_527: class Program_weight_tensor_parameter_528: name = "parameter_528" + original_name = "linear_130.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.372668") @@ -5791,6 +6320,7 @@ class Program_weight_tensor_parameter_528: class Program_weight_tensor_parameter_529: name = "parameter_529" + original_name = "layer_norm_84.b_0" shape = [320] dtype = "float32" min_val = float("-0.43466") @@ -5802,6 +6332,7 @@ class Program_weight_tensor_parameter_529: class Program_weight_tensor_parameter_530: name = "parameter_530" + original_name = "layer_norm_84.w_0" shape = [320] dtype = "float32" min_val = float("0.0990338") @@ -5813,6 +6344,7 @@ class Program_weight_tensor_parameter_530: class Program_weight_tensor_parameter_531: name = "parameter_531" + original_name = "linear_129.b_0" shape = [320] dtype = "float32" min_val = float("-0.252991") @@ -5824,6 +6356,7 @@ class Program_weight_tensor_parameter_531: class Program_weight_tensor_parameter_532: name = "parameter_532" + original_name = "linear_129.w_0" shape = [1280, 320] dtype = "float32" min_val = float("-0.592251") @@ -5835,6 +6368,7 @@ class Program_weight_tensor_parameter_532: class Program_weight_tensor_parameter_533: name = "parameter_533" + original_name = "conv2d_55.b_0" shape = [1280] dtype = "float32" min_val = float("-1.62274") @@ -5846,6 +6380,7 @@ class Program_weight_tensor_parameter_533: class Program_weight_tensor_parameter_534: name = "parameter_534" + original_name = "conv2d_55.w_0" shape = [1280, 1, 3, 3] dtype = "float32" min_val = float("-0.267969") @@ -5857,6 +6392,7 @@ class Program_weight_tensor_parameter_534: class Program_weight_tensor_parameter_535: name = "parameter_535" + original_name = "linear_128.b_0" shape = [1280] dtype = "float32" min_val = float("-1.82177") @@ -5868,6 +6404,7 @@ class Program_weight_tensor_parameter_535: class Program_weight_tensor_parameter_536: name = "parameter_536" + original_name = "linear_128.w_0" shape = [320, 1280] dtype = "float32" min_val = float("-0.679399") @@ -5879,6 +6416,7 @@ class Program_weight_tensor_parameter_536: class Program_weight_tensor_parameter_537: name = "parameter_537" + original_name = "layer_norm_83.b_0" shape = [320] dtype = "float32" min_val = float("-1.5251") @@ -5890,6 +6428,7 @@ class Program_weight_tensor_parameter_537: class Program_weight_tensor_parameter_538: name = "parameter_538" + original_name = "layer_norm_83.w_0" shape = [320] dtype = "float32" min_val = float("0.712872") @@ -5901,6 +6440,7 @@ class Program_weight_tensor_parameter_538: class Program_weight_tensor_parameter_539: name = "parameter_539" + original_name = "linear_127.b_0" shape = [320] dtype = "float32" min_val = float("-0.0930134") @@ -5912,6 +6452,7 @@ class Program_weight_tensor_parameter_539: class Program_weight_tensor_parameter_540: name = "parameter_540" + original_name = "linear_127.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.575761") @@ -5923,6 +6464,7 @@ class Program_weight_tensor_parameter_540: class Program_weight_tensor_parameter_541: name = "parameter_541" + original_name = "linear_126.b_0" shape = [640] dtype = "float32" min_val = float("-0.658989") @@ -5934,6 +6476,7 @@ class Program_weight_tensor_parameter_541: class Program_weight_tensor_parameter_542: name = "parameter_542" + original_name = "linear_126.w_0" shape = [320, 640] dtype = "float32" min_val = float("-0.378011") @@ -5945,6 +6488,7 @@ class Program_weight_tensor_parameter_542: class Program_weight_tensor_parameter_543: name = "parameter_543" + original_name = "layer_norm_82.b_0" shape = [320] dtype = "float32" min_val = float("-0.121292") @@ -5956,6 +6500,7 @@ class Program_weight_tensor_parameter_543: class Program_weight_tensor_parameter_544: name = "parameter_544" + original_name = "layer_norm_82.w_0" shape = [320] dtype = "float32" min_val = float("-0.0911821") @@ -5967,6 +6512,7 @@ class Program_weight_tensor_parameter_544: class Program_weight_tensor_parameter_545: name = "parameter_545" + original_name = "conv2d_54.b_0" shape = [320] dtype = "float32" min_val = float("-4.20868") @@ -5978,6 +6524,7 @@ class Program_weight_tensor_parameter_545: class Program_weight_tensor_parameter_546: name = "parameter_546" + original_name = "conv2d_54.w_0" shape = [320, 320, 2, 2] dtype = "float32" min_val = float("-0.347552") @@ -5989,6 +6536,7 @@ class Program_weight_tensor_parameter_546: class Program_weight_tensor_parameter_547: name = "parameter_547" + original_name = "linear_125.b_0" shape = [320] dtype = "float32" min_val = float("-3.29505") @@ -6000,6 +6548,7 @@ class Program_weight_tensor_parameter_547: class Program_weight_tensor_parameter_548: name = "parameter_548" + original_name = "linear_125.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.268003") @@ -6011,6 +6560,7 @@ class Program_weight_tensor_parameter_548: class Program_weight_tensor_parameter_549: name = "parameter_549" + original_name = "layer_norm_81.b_0" shape = [320] dtype = "float32" min_val = float("-0.589203") @@ -6022,6 +6572,7 @@ class Program_weight_tensor_parameter_549: class Program_weight_tensor_parameter_550: name = "parameter_550" + original_name = "layer_norm_81.w_0" shape = [320] dtype = "float32" min_val = float("0.130318") @@ -6033,6 +6584,7 @@ class Program_weight_tensor_parameter_550: class Program_weight_tensor_parameter_551: name = "parameter_551" + original_name = "linear_124.b_0" shape = [320] dtype = "float32" min_val = float("-0.226436") @@ -6044,6 +6596,7 @@ class Program_weight_tensor_parameter_551: class Program_weight_tensor_parameter_552: name = "parameter_552" + original_name = "linear_124.w_0" shape = [1280, 320] dtype = "float32" min_val = float("-0.338409") @@ -6055,6 +6608,7 @@ class Program_weight_tensor_parameter_552: class Program_weight_tensor_parameter_553: name = "parameter_553" + original_name = "conv2d_53.b_0" shape = [1280] dtype = "float32" min_val = float("-1.62772") @@ -6066,6 +6620,7 @@ class Program_weight_tensor_parameter_553: class Program_weight_tensor_parameter_554: name = "parameter_554" + original_name = "conv2d_53.w_0" shape = [1280, 1, 3, 3] dtype = "float32" min_val = float("-0.255513") @@ -6077,6 +6632,7 @@ class Program_weight_tensor_parameter_554: class Program_weight_tensor_parameter_555: name = "parameter_555" + original_name = "linear_123.b_0" shape = [1280] dtype = "float32" min_val = float("-1.9157") @@ -6088,6 +6644,7 @@ class Program_weight_tensor_parameter_555: class Program_weight_tensor_parameter_556: name = "parameter_556" + original_name = "linear_123.w_0" shape = [320, 1280] dtype = "float32" min_val = float("-0.413284") @@ -6099,6 +6656,7 @@ class Program_weight_tensor_parameter_556: class Program_weight_tensor_parameter_557: name = "parameter_557" + original_name = "layer_norm_80.b_0" shape = [320] dtype = "float32" min_val = float("-1.35845") @@ -6110,6 +6668,7 @@ class Program_weight_tensor_parameter_557: class Program_weight_tensor_parameter_558: name = "parameter_558" + original_name = "layer_norm_80.w_0" shape = [320] dtype = "float32" min_val = float("0.240199") @@ -6121,6 +6680,7 @@ class Program_weight_tensor_parameter_558: class Program_weight_tensor_parameter_559: name = "parameter_559" + original_name = "linear_122.b_0" shape = [320] dtype = "float32" min_val = float("-0.0885255") @@ -6132,6 +6692,7 @@ class Program_weight_tensor_parameter_559: class Program_weight_tensor_parameter_560: name = "parameter_560" + original_name = "linear_122.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.55872") @@ -6143,6 +6704,7 @@ class Program_weight_tensor_parameter_560: class Program_weight_tensor_parameter_561: name = "parameter_561" + original_name = "linear_121.b_0" shape = [640] dtype = "float32" min_val = float("-0.630756") @@ -6154,6 +6716,7 @@ class Program_weight_tensor_parameter_561: class Program_weight_tensor_parameter_562: name = "parameter_562" + original_name = "linear_121.w_0" shape = [320, 640] dtype = "float32" min_val = float("-0.392607") @@ -6165,6 +6728,7 @@ class Program_weight_tensor_parameter_562: class Program_weight_tensor_parameter_563: name = "parameter_563" + original_name = "layer_norm_79.b_0" shape = [320] dtype = "float32" min_val = float("-0.0884986") @@ -6176,6 +6740,7 @@ class Program_weight_tensor_parameter_563: class Program_weight_tensor_parameter_564: name = "parameter_564" + original_name = "layer_norm_79.w_0" shape = [320] dtype = "float32" min_val = float("-0.0244792") @@ -6187,6 +6752,7 @@ class Program_weight_tensor_parameter_564: class Program_weight_tensor_parameter_565: name = "parameter_565" + original_name = "conv2d_52.b_0" shape = [320] dtype = "float32" min_val = float("-5.53786") @@ -6198,6 +6764,7 @@ class Program_weight_tensor_parameter_565: class Program_weight_tensor_parameter_566: name = "parameter_566" + original_name = "conv2d_52.w_0" shape = [320, 320, 2, 2] dtype = "float32" min_val = float("-0.414349") @@ -6209,6 +6776,7 @@ class Program_weight_tensor_parameter_566: class Program_weight_tensor_parameter_567: name = "parameter_567" + original_name = "linear_120.b_0" shape = [320] dtype = "float32" min_val = float("-2.45115") @@ -6220,6 +6788,7 @@ class Program_weight_tensor_parameter_567: class Program_weight_tensor_parameter_568: name = "parameter_568" + original_name = "linear_120.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.238391") @@ -6231,6 +6800,7 @@ class Program_weight_tensor_parameter_568: class Program_weight_tensor_parameter_569: name = "parameter_569" + original_name = "layer_norm_78.b_0" shape = [320] dtype = "float32" min_val = float("-0.835308") @@ -6242,6 +6812,7 @@ class Program_weight_tensor_parameter_569: class Program_weight_tensor_parameter_570: name = "parameter_570" + original_name = "layer_norm_78.w_0" shape = [320] dtype = "float32" min_val = float("0.164732") @@ -6253,6 +6824,7 @@ class Program_weight_tensor_parameter_570: class Program_weight_tensor_parameter_571: name = "parameter_571" + original_name = "linear_119.b_0" shape = [320] dtype = "float32" min_val = float("-0.305216") @@ -6264,6 +6836,7 @@ class Program_weight_tensor_parameter_571: class Program_weight_tensor_parameter_572: name = "parameter_572" + original_name = "linear_119.w_0" shape = [1280, 320] dtype = "float32" min_val = float("-0.300276") @@ -6275,6 +6848,7 @@ class Program_weight_tensor_parameter_572: class Program_weight_tensor_parameter_573: name = "parameter_573" + original_name = "conv2d_51.b_0" shape = [1280] dtype = "float32" min_val = float("-1.58121") @@ -6286,6 +6860,7 @@ class Program_weight_tensor_parameter_573: class Program_weight_tensor_parameter_574: name = "parameter_574" + original_name = "conv2d_51.w_0" shape = [1280, 1, 3, 3] dtype = "float32" min_val = float("-0.270609") @@ -6297,6 +6872,7 @@ class Program_weight_tensor_parameter_574: class Program_weight_tensor_parameter_575: name = "parameter_575" + original_name = "linear_118.b_0" shape = [1280] dtype = "float32" min_val = float("-2.26423") @@ -6308,6 +6884,7 @@ class Program_weight_tensor_parameter_575: class Program_weight_tensor_parameter_576: name = "parameter_576" + original_name = "linear_118.w_0" shape = [320, 1280] dtype = "float32" min_val = float("-0.526174") @@ -6319,6 +6896,7 @@ class Program_weight_tensor_parameter_576: class Program_weight_tensor_parameter_577: name = "parameter_577" + original_name = "layer_norm_77.b_0" shape = [320] dtype = "float32" min_val = float("-1.86852") @@ -6330,6 +6908,7 @@ class Program_weight_tensor_parameter_577: class Program_weight_tensor_parameter_578: name = "parameter_578" + original_name = "layer_norm_77.w_0" shape = [320] dtype = "float32" min_val = float("0.439614") @@ -6341,6 +6920,7 @@ class Program_weight_tensor_parameter_578: class Program_weight_tensor_parameter_579: name = "parameter_579" + original_name = "linear_117.b_0" shape = [320] dtype = "float32" min_val = float("-0.168556") @@ -6352,6 +6932,7 @@ class Program_weight_tensor_parameter_579: class Program_weight_tensor_parameter_580: name = "parameter_580" + original_name = "linear_117.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.631964") @@ -6363,6 +6944,7 @@ class Program_weight_tensor_parameter_580: class Program_weight_tensor_parameter_581: name = "parameter_581" + original_name = "linear_116.b_0" shape = [640] dtype = "float32" min_val = float("-0.950498") @@ -6374,6 +6956,7 @@ class Program_weight_tensor_parameter_581: class Program_weight_tensor_parameter_582: name = "parameter_582" + original_name = "linear_116.w_0" shape = [320, 640] dtype = "float32" min_val = float("-0.457362") @@ -6385,6 +6968,7 @@ class Program_weight_tensor_parameter_582: class Program_weight_tensor_parameter_583: name = "parameter_583" + original_name = "layer_norm_76.b_0" shape = [320] dtype = "float32" min_val = float("-0.174717") @@ -6396,6 +6980,7 @@ class Program_weight_tensor_parameter_583: class Program_weight_tensor_parameter_584: name = "parameter_584" + original_name = "layer_norm_76.w_0" shape = [320] dtype = "float32" min_val = float("-0.0460412") @@ -6407,6 +6992,7 @@ class Program_weight_tensor_parameter_584: class Program_weight_tensor_parameter_585: name = "parameter_585" + original_name = "conv2d_50.b_0" shape = [320] dtype = "float32" min_val = float("-4.45403") @@ -6418,6 +7004,7 @@ class Program_weight_tensor_parameter_585: class Program_weight_tensor_parameter_586: name = "parameter_586" + original_name = "conv2d_50.w_0" shape = [320, 320, 2, 2] dtype = "float32" min_val = float("-0.457572") @@ -6429,6 +7016,7 @@ class Program_weight_tensor_parameter_586: class Program_weight_tensor_parameter_587: name = "parameter_587" + original_name = "linear_115.b_0" shape = [320] dtype = "float32" min_val = float("-2.87655") @@ -6440,6 +7028,7 @@ class Program_weight_tensor_parameter_587: class Program_weight_tensor_parameter_588: name = "parameter_588" + original_name = "linear_115.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.245825") @@ -6451,6 +7040,7 @@ class Program_weight_tensor_parameter_588: class Program_weight_tensor_parameter_589: name = "parameter_589" + original_name = "layer_norm_75.b_0" shape = [320] dtype = "float32" min_val = float("-0.745473") @@ -6462,6 +7052,7 @@ class Program_weight_tensor_parameter_589: class Program_weight_tensor_parameter_590: name = "parameter_590" + original_name = "layer_norm_75.w_0" shape = [320] dtype = "float32" min_val = float("0.167462") @@ -6473,6 +7064,7 @@ class Program_weight_tensor_parameter_590: class Program_weight_tensor_parameter_591: name = "parameter_591" + original_name = "linear_114.b_0" shape = [320] dtype = "float32" min_val = float("-0.303229") @@ -6484,6 +7076,7 @@ class Program_weight_tensor_parameter_591: class Program_weight_tensor_parameter_592: name = "parameter_592" + original_name = "linear_114.w_0" shape = [1280, 320] dtype = "float32" min_val = float("-0.293138") @@ -6495,6 +7088,7 @@ class Program_weight_tensor_parameter_592: class Program_weight_tensor_parameter_593: name = "parameter_593" + original_name = "conv2d_49.b_0" shape = [1280] dtype = "float32" min_val = float("-1.4623") @@ -6506,6 +7100,7 @@ class Program_weight_tensor_parameter_593: class Program_weight_tensor_parameter_594: name = "parameter_594" + original_name = "conv2d_49.w_0" shape = [1280, 1, 3, 3] dtype = "float32" min_val = float("-0.277999") @@ -6517,6 +7112,7 @@ class Program_weight_tensor_parameter_594: class Program_weight_tensor_parameter_595: name = "parameter_595" + original_name = "linear_113.b_0" shape = [1280] dtype = "float32" min_val = float("-1.71075") @@ -6528,6 +7124,7 @@ class Program_weight_tensor_parameter_595: class Program_weight_tensor_parameter_596: name = "parameter_596" + original_name = "linear_113.w_0" shape = [320, 1280] dtype = "float32" min_val = float("-0.369439") @@ -6539,6 +7136,7 @@ class Program_weight_tensor_parameter_596: class Program_weight_tensor_parameter_597: name = "parameter_597" + original_name = "layer_norm_74.b_0" shape = [320] dtype = "float32" min_val = float("-1.61103") @@ -6550,6 +7148,7 @@ class Program_weight_tensor_parameter_597: class Program_weight_tensor_parameter_598: name = "parameter_598" + original_name = "layer_norm_74.w_0" shape = [320] dtype = "float32" min_val = float("0.249431") @@ -6561,6 +7160,7 @@ class Program_weight_tensor_parameter_598: class Program_weight_tensor_parameter_599: name = "parameter_599" + original_name = "linear_112.b_0" shape = [320] dtype = "float32" min_val = float("-0.187648") @@ -6572,6 +7172,7 @@ class Program_weight_tensor_parameter_599: class Program_weight_tensor_parameter_600: name = "parameter_600" + original_name = "linear_112.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.324247") @@ -6583,6 +7184,7 @@ class Program_weight_tensor_parameter_600: class Program_weight_tensor_parameter_601: name = "parameter_601" + original_name = "linear_111.b_0" shape = [640] dtype = "float32" min_val = float("-1.28881") @@ -6594,6 +7196,7 @@ class Program_weight_tensor_parameter_601: class Program_weight_tensor_parameter_602: name = "parameter_602" + original_name = "linear_111.w_0" shape = [320, 640] dtype = "float32" min_val = float("-0.432806") @@ -6605,6 +7208,7 @@ class Program_weight_tensor_parameter_602: class Program_weight_tensor_parameter_603: name = "parameter_603" + original_name = "layer_norm_73.b_0" shape = [320] dtype = "float32" min_val = float("-0.170525") @@ -6616,6 +7220,7 @@ class Program_weight_tensor_parameter_603: class Program_weight_tensor_parameter_604: name = "parameter_604" + original_name = "layer_norm_73.w_0" shape = [320] dtype = "float32" min_val = float("-0.0581484") @@ -6627,6 +7232,7 @@ class Program_weight_tensor_parameter_604: class Program_weight_tensor_parameter_605: name = "parameter_605" + original_name = "conv2d_48.b_0" shape = [320] dtype = "float32" min_val = float("-4.26459") @@ -6638,6 +7244,7 @@ class Program_weight_tensor_parameter_605: class Program_weight_tensor_parameter_606: name = "parameter_606" + original_name = "conv2d_48.w_0" shape = [320, 320, 2, 2] dtype = "float32" min_val = float("-0.354049") @@ -6649,6 +7256,7 @@ class Program_weight_tensor_parameter_606: class Program_weight_tensor_parameter_607: name = "parameter_607" + original_name = "linear_110.b_0" shape = [320] dtype = "float32" min_val = float("-3.39124") @@ -6660,6 +7268,7 @@ class Program_weight_tensor_parameter_607: class Program_weight_tensor_parameter_608: name = "parameter_608" + original_name = "linear_110.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.245829") @@ -6671,6 +7280,7 @@ class Program_weight_tensor_parameter_608: class Program_weight_tensor_parameter_609: name = "parameter_609" + original_name = "layer_norm_72.b_0" shape = [320] dtype = "float32" min_val = float("-0.803417") @@ -6682,6 +7292,7 @@ class Program_weight_tensor_parameter_609: class Program_weight_tensor_parameter_610: name = "parameter_610" + original_name = "layer_norm_72.w_0" shape = [320] dtype = "float32" min_val = float("0.0732666") @@ -6693,6 +7304,7 @@ class Program_weight_tensor_parameter_610: class Program_weight_tensor_parameter_611: name = "parameter_611" + original_name = "linear_109.b_0" shape = [320] dtype = "float32" min_val = float("-0.288361") @@ -6704,6 +7316,7 @@ class Program_weight_tensor_parameter_611: class Program_weight_tensor_parameter_612: name = "parameter_612" + original_name = "linear_109.w_0" shape = [1280, 320] dtype = "float32" min_val = float("-0.385943") @@ -6715,6 +7328,7 @@ class Program_weight_tensor_parameter_612: class Program_weight_tensor_parameter_613: name = "parameter_613" + original_name = "conv2d_47.b_0" shape = [1280] dtype = "float32" min_val = float("-1.72149") @@ -6726,6 +7340,7 @@ class Program_weight_tensor_parameter_613: class Program_weight_tensor_parameter_614: name = "parameter_614" + original_name = "conv2d_47.w_0" shape = [1280, 1, 3, 3] dtype = "float32" min_val = float("-0.259731") @@ -6737,6 +7352,7 @@ class Program_weight_tensor_parameter_614: class Program_weight_tensor_parameter_615: name = "parameter_615" + original_name = "linear_108.b_0" shape = [1280] dtype = "float32" min_val = float("-1.97522") @@ -6748,6 +7364,7 @@ class Program_weight_tensor_parameter_615: class Program_weight_tensor_parameter_616: name = "parameter_616" + original_name = "linear_108.w_0" shape = [320, 1280] dtype = "float32" min_val = float("-0.598189") @@ -6759,6 +7376,7 @@ class Program_weight_tensor_parameter_616: class Program_weight_tensor_parameter_617: name = "parameter_617" + original_name = "layer_norm_71.b_0" shape = [320] dtype = "float32" min_val = float("-1.56237") @@ -6770,6 +7388,7 @@ class Program_weight_tensor_parameter_617: class Program_weight_tensor_parameter_618: name = "parameter_618" + original_name = "layer_norm_71.w_0" shape = [320] dtype = "float32" min_val = float("0.668453") @@ -6781,6 +7400,7 @@ class Program_weight_tensor_parameter_618: class Program_weight_tensor_parameter_619: name = "parameter_619" + original_name = "linear_107.b_0" shape = [320] dtype = "float32" min_val = float("-0.250741") @@ -6792,6 +7412,7 @@ class Program_weight_tensor_parameter_619: class Program_weight_tensor_parameter_620: name = "parameter_620" + original_name = "linear_107.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.624308") @@ -6803,6 +7424,7 @@ class Program_weight_tensor_parameter_620: class Program_weight_tensor_parameter_621: name = "parameter_621" + original_name = "linear_106.b_0" shape = [640] dtype = "float32" min_val = float("-1.08966") @@ -6814,6 +7436,7 @@ class Program_weight_tensor_parameter_621: class Program_weight_tensor_parameter_622: name = "parameter_622" + original_name = "linear_106.w_0" shape = [320, 640] dtype = "float32" min_val = float("-0.396377") @@ -6825,6 +7448,7 @@ class Program_weight_tensor_parameter_622: class Program_weight_tensor_parameter_623: name = "parameter_623" + original_name = "layer_norm_70.b_0" shape = [320] dtype = "float32" min_val = float("-0.192083") @@ -6836,6 +7460,7 @@ class Program_weight_tensor_parameter_623: class Program_weight_tensor_parameter_624: name = "parameter_624" + original_name = "layer_norm_70.w_0" shape = [320] dtype = "float32" min_val = float("-0.0360734") @@ -6847,6 +7472,7 @@ class Program_weight_tensor_parameter_624: class Program_weight_tensor_parameter_625: name = "parameter_625" + original_name = "conv2d_46.b_0" shape = [320] dtype = "float32" min_val = float("-4.21307") @@ -6858,6 +7484,7 @@ class Program_weight_tensor_parameter_625: class Program_weight_tensor_parameter_626: name = "parameter_626" + original_name = "conv2d_46.w_0" shape = [320, 320, 2, 2] dtype = "float32" min_val = float("-0.368987") @@ -6869,6 +7496,7 @@ class Program_weight_tensor_parameter_626: class Program_weight_tensor_parameter_627: name = "parameter_627" + original_name = "linear_105.b_0" shape = [320] dtype = "float32" min_val = float("-2.44485") @@ -6880,6 +7508,7 @@ class Program_weight_tensor_parameter_627: class Program_weight_tensor_parameter_628: name = "parameter_628" + original_name = "linear_105.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.289487") @@ -6891,6 +7520,7 @@ class Program_weight_tensor_parameter_628: class Program_weight_tensor_parameter_629: name = "parameter_629" + original_name = "layer_norm_69.b_0" shape = [320] dtype = "float32" min_val = float("-0.747517") @@ -6902,6 +7532,7 @@ class Program_weight_tensor_parameter_629: class Program_weight_tensor_parameter_630: name = "parameter_630" + original_name = "layer_norm_69.w_0" shape = [320] dtype = "float32" min_val = float("0.134345") @@ -6913,6 +7544,7 @@ class Program_weight_tensor_parameter_630: class Program_weight_tensor_parameter_631: name = "parameter_631" + original_name = "linear_104.b_0" shape = [320] dtype = "float32" min_val = float("-0.31892") @@ -6924,6 +7556,7 @@ class Program_weight_tensor_parameter_631: class Program_weight_tensor_parameter_632: name = "parameter_632" + original_name = "linear_104.w_0" shape = [1280, 320] dtype = "float32" min_val = float("-0.459561") @@ -6935,6 +7568,7 @@ class Program_weight_tensor_parameter_632: class Program_weight_tensor_parameter_633: name = "parameter_633" + original_name = "conv2d_45.b_0" shape = [1280] dtype = "float32" min_val = float("-1.72777") @@ -6946,6 +7580,7 @@ class Program_weight_tensor_parameter_633: class Program_weight_tensor_parameter_634: name = "parameter_634" + original_name = "conv2d_45.w_0" shape = [1280, 1, 3, 3] dtype = "float32" min_val = float("-0.260941") @@ -6957,6 +7592,7 @@ class Program_weight_tensor_parameter_634: class Program_weight_tensor_parameter_635: name = "parameter_635" + original_name = "linear_103.b_0" shape = [1280] dtype = "float32" min_val = float("-1.35074") @@ -6968,6 +7604,7 @@ class Program_weight_tensor_parameter_635: class Program_weight_tensor_parameter_636: name = "parameter_636" + original_name = "linear_103.w_0" shape = [320, 1280] dtype = "float32" min_val = float("-0.505778") @@ -6979,6 +7616,7 @@ class Program_weight_tensor_parameter_636: class Program_weight_tensor_parameter_637: name = "parameter_637" + original_name = "layer_norm_68.b_0" shape = [320] dtype = "float32" min_val = float("-1.78835") @@ -6990,6 +7628,7 @@ class Program_weight_tensor_parameter_637: class Program_weight_tensor_parameter_638: name = "parameter_638" + original_name = "layer_norm_68.w_0" shape = [320] dtype = "float32" min_val = float("0.618298") @@ -7001,6 +7640,7 @@ class Program_weight_tensor_parameter_638: class Program_weight_tensor_parameter_639: name = "parameter_639" + original_name = "linear_102.b_0" shape = [320] dtype = "float32" min_val = float("-0.304568") @@ -7012,6 +7652,7 @@ class Program_weight_tensor_parameter_639: class Program_weight_tensor_parameter_640: name = "parameter_640" + original_name = "linear_102.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.713428") @@ -7023,6 +7664,7 @@ class Program_weight_tensor_parameter_640: class Program_weight_tensor_parameter_641: name = "parameter_641" + original_name = "linear_101.b_0" shape = [640] dtype = "float32" min_val = float("-0.728708") @@ -7034,6 +7676,7 @@ class Program_weight_tensor_parameter_641: class Program_weight_tensor_parameter_642: name = "parameter_642" + original_name = "linear_101.w_0" shape = [320, 640] dtype = "float32" min_val = float("-0.468803") @@ -7045,6 +7688,7 @@ class Program_weight_tensor_parameter_642: class Program_weight_tensor_parameter_643: name = "parameter_643" + original_name = "layer_norm_67.b_0" shape = [320] dtype = "float32" min_val = float("-0.156031") @@ -7056,6 +7700,7 @@ class Program_weight_tensor_parameter_643: class Program_weight_tensor_parameter_644: name = "parameter_644" + original_name = "layer_norm_67.w_0" shape = [320] dtype = "float32" min_val = float("-0.0096201") @@ -7067,6 +7712,7 @@ class Program_weight_tensor_parameter_644: class Program_weight_tensor_parameter_645: name = "parameter_645" + original_name = "conv2d_44.b_0" shape = [320] dtype = "float32" min_val = float("-5.13965") @@ -7078,6 +7724,7 @@ class Program_weight_tensor_parameter_645: class Program_weight_tensor_parameter_646: name = "parameter_646" + original_name = "conv2d_44.w_0" shape = [320, 320, 2, 2] dtype = "float32" min_val = float("-0.435027") @@ -7089,6 +7736,7 @@ class Program_weight_tensor_parameter_646: class Program_weight_tensor_parameter_647: name = "parameter_647" + original_name = "linear_100.b_0" shape = [320] dtype = "float32" min_val = float("-3.09182") @@ -7100,6 +7748,7 @@ class Program_weight_tensor_parameter_647: class Program_weight_tensor_parameter_648: name = "parameter_648" + original_name = "linear_100.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.245764") @@ -7111,6 +7760,7 @@ class Program_weight_tensor_parameter_648: class Program_weight_tensor_parameter_649: name = "parameter_649" + original_name = "layer_norm_66.b_0" shape = [320] dtype = "float32" min_val = float("-0.849572") @@ -7122,6 +7772,7 @@ class Program_weight_tensor_parameter_649: class Program_weight_tensor_parameter_650: name = "parameter_650" + original_name = "layer_norm_66.w_0" shape = [320] dtype = "float32" min_val = float("0.171056") @@ -7133,6 +7784,7 @@ class Program_weight_tensor_parameter_650: class Program_weight_tensor_parameter_651: name = "parameter_651" + original_name = "linear_99.b_0" shape = [320] dtype = "float32" min_val = float("-0.310266") @@ -7144,6 +7796,7 @@ class Program_weight_tensor_parameter_651: class Program_weight_tensor_parameter_652: name = "parameter_652" + original_name = "linear_99.w_0" shape = [1280, 320] dtype = "float32" min_val = float("-0.432067") @@ -7155,6 +7808,7 @@ class Program_weight_tensor_parameter_652: class Program_weight_tensor_parameter_653: name = "parameter_653" + original_name = "conv2d_43.b_0" shape = [1280] dtype = "float32" min_val = float("-1.51093") @@ -7166,6 +7820,7 @@ class Program_weight_tensor_parameter_653: class Program_weight_tensor_parameter_654: name = "parameter_654" + original_name = "conv2d_43.w_0" shape = [1280, 1, 3, 3] dtype = "float32" min_val = float("-0.295504") @@ -7177,6 +7832,7 @@ class Program_weight_tensor_parameter_654: class Program_weight_tensor_parameter_655: name = "parameter_655" + original_name = "linear_98.b_0" shape = [1280] dtype = "float32" min_val = float("-1.31516") @@ -7188,6 +7844,7 @@ class Program_weight_tensor_parameter_655: class Program_weight_tensor_parameter_656: name = "parameter_656" + original_name = "linear_98.w_0" shape = [320, 1280] dtype = "float32" min_val = float("-0.5051") @@ -7199,6 +7856,7 @@ class Program_weight_tensor_parameter_656: class Program_weight_tensor_parameter_657: name = "parameter_657" + original_name = "layer_norm_65.b_0" shape = [320] dtype = "float32" min_val = float("-2.01004") @@ -7210,6 +7868,7 @@ class Program_weight_tensor_parameter_657: class Program_weight_tensor_parameter_658: name = "parameter_658" + original_name = "layer_norm_65.w_0" shape = [320] dtype = "float32" min_val = float("0.670875") @@ -7221,6 +7880,7 @@ class Program_weight_tensor_parameter_658: class Program_weight_tensor_parameter_659: name = "parameter_659" + original_name = "linear_97.b_0" shape = [320] dtype = "float32" min_val = float("-0.113715") @@ -7232,6 +7892,7 @@ class Program_weight_tensor_parameter_659: class Program_weight_tensor_parameter_660: name = "parameter_660" + original_name = "linear_97.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.429307") @@ -7243,6 +7904,7 @@ class Program_weight_tensor_parameter_660: class Program_weight_tensor_parameter_661: name = "parameter_661" + original_name = "linear_96.b_0" shape = [640] dtype = "float32" min_val = float("-0.707022") @@ -7254,6 +7916,7 @@ class Program_weight_tensor_parameter_661: class Program_weight_tensor_parameter_662: name = "parameter_662" + original_name = "linear_96.w_0" shape = [320, 640] dtype = "float32" min_val = float("-0.375588") @@ -7265,6 +7928,7 @@ class Program_weight_tensor_parameter_662: class Program_weight_tensor_parameter_663: name = "parameter_663" + original_name = "layer_norm_64.b_0" shape = [320] dtype = "float32" min_val = float("-0.156251") @@ -7276,6 +7940,7 @@ class Program_weight_tensor_parameter_663: class Program_weight_tensor_parameter_664: name = "parameter_664" + original_name = "layer_norm_64.w_0" shape = [320] dtype = "float32" min_val = float("-0.111312") @@ -7287,6 +7952,7 @@ class Program_weight_tensor_parameter_664: class Program_weight_tensor_parameter_665: name = "parameter_665" + original_name = "conv2d_42.b_0" shape = [320] dtype = "float32" min_val = float("-2.57925") @@ -7298,6 +7964,7 @@ class Program_weight_tensor_parameter_665: class Program_weight_tensor_parameter_666: name = "parameter_666" + original_name = "conv2d_42.w_0" shape = [320, 320, 2, 2] dtype = "float32" min_val = float("-0.226743") @@ -7309,6 +7976,7 @@ class Program_weight_tensor_parameter_666: class Program_weight_tensor_parameter_667: name = "parameter_667" + original_name = "linear_95.b_0" shape = [320] dtype = "float32" min_val = float("-2.32781") @@ -7320,6 +7988,7 @@ class Program_weight_tensor_parameter_667: class Program_weight_tensor_parameter_668: name = "parameter_668" + original_name = "linear_95.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.278578") @@ -7331,6 +8000,7 @@ class Program_weight_tensor_parameter_668: class Program_weight_tensor_parameter_669: name = "parameter_669" + original_name = "layer_norm_63.b_0" shape = [320] dtype = "float32" min_val = float("-0.977649") @@ -7342,6 +8012,7 @@ class Program_weight_tensor_parameter_669: class Program_weight_tensor_parameter_670: name = "parameter_670" + original_name = "layer_norm_63.w_0" shape = [320] dtype = "float32" min_val = float("-0.156425") @@ -7353,6 +8024,7 @@ class Program_weight_tensor_parameter_670: class Program_weight_tensor_parameter_671: name = "parameter_671" + original_name = "linear_94.b_0" shape = [320] dtype = "float32" min_val = float("-0.344163") @@ -7364,6 +8036,7 @@ class Program_weight_tensor_parameter_671: class Program_weight_tensor_parameter_672: name = "parameter_672" + original_name = "linear_94.w_0" shape = [1280, 320] dtype = "float32" min_val = float("-0.285557") @@ -7375,6 +8048,7 @@ class Program_weight_tensor_parameter_672: class Program_weight_tensor_parameter_673: name = "parameter_673" + original_name = "conv2d_41.b_0" shape = [1280] dtype = "float32" min_val = float("-1.77335") @@ -7386,6 +8060,7 @@ class Program_weight_tensor_parameter_673: class Program_weight_tensor_parameter_674: name = "parameter_674" + original_name = "conv2d_41.w_0" shape = [1280, 1, 3, 3] dtype = "float32" min_val = float("-0.315814") @@ -7397,6 +8072,7 @@ class Program_weight_tensor_parameter_674: class Program_weight_tensor_parameter_675: name = "parameter_675" + original_name = "linear_93.b_0" shape = [1280] dtype = "float32" min_val = float("-1.54796") @@ -7408,6 +8084,7 @@ class Program_weight_tensor_parameter_675: class Program_weight_tensor_parameter_676: name = "parameter_676" + original_name = "linear_93.w_0" shape = [320, 1280] dtype = "float32" min_val = float("-0.387985") @@ -7419,6 +8096,7 @@ class Program_weight_tensor_parameter_676: class Program_weight_tensor_parameter_677: name = "parameter_677" + original_name = "layer_norm_62.b_0" shape = [320] dtype = "float32" min_val = float("-1.55908") @@ -7430,6 +8108,7 @@ class Program_weight_tensor_parameter_677: class Program_weight_tensor_parameter_678: name = "parameter_678" + original_name = "layer_norm_62.w_0" shape = [320] dtype = "float32" min_val = float("0.653681") @@ -7441,6 +8120,7 @@ class Program_weight_tensor_parameter_678: class Program_weight_tensor_parameter_679: name = "parameter_679" + original_name = "linear_92.b_0" shape = [320] dtype = "float32" min_val = float("-0.238974") @@ -7452,6 +8132,7 @@ class Program_weight_tensor_parameter_679: class Program_weight_tensor_parameter_680: name = "parameter_680" + original_name = "linear_92.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.300052") @@ -7463,6 +8144,7 @@ class Program_weight_tensor_parameter_680: class Program_weight_tensor_parameter_681: name = "parameter_681" + original_name = "linear_91.b_0" shape = [640] dtype = "float32" min_val = float("-0.834231") @@ -7474,6 +8156,7 @@ class Program_weight_tensor_parameter_681: class Program_weight_tensor_parameter_682: name = "parameter_682" + original_name = "linear_91.w_0" shape = [320, 640] dtype = "float32" min_val = float("-0.341502") @@ -7485,6 +8168,7 @@ class Program_weight_tensor_parameter_682: class Program_weight_tensor_parameter_683: name = "parameter_683" + original_name = "layer_norm_61.b_0" shape = [320] dtype = "float32" min_val = float("-0.378668") @@ -7496,6 +8180,7 @@ class Program_weight_tensor_parameter_683: class Program_weight_tensor_parameter_684: name = "parameter_684" + original_name = "layer_norm_61.w_0" shape = [320] dtype = "float32" min_val = float("-0.020289") @@ -7507,6 +8192,7 @@ class Program_weight_tensor_parameter_684: class Program_weight_tensor_parameter_685: name = "parameter_685" + original_name = "conv2d_40.b_0" shape = [320] dtype = "float32" min_val = float("-3.21852") @@ -7518,6 +8204,7 @@ class Program_weight_tensor_parameter_685: class Program_weight_tensor_parameter_686: name = "parameter_686" + original_name = "conv2d_40.w_0" shape = [320, 320, 2, 2] dtype = "float32" min_val = float("-0.367259") @@ -7529,6 +8216,7 @@ class Program_weight_tensor_parameter_686: class Program_weight_tensor_parameter_687: name = "parameter_687" + original_name = "linear_90.b_0" shape = [320] dtype = "float32" min_val = float("-3.26289") @@ -7540,6 +8228,7 @@ class Program_weight_tensor_parameter_687: class Program_weight_tensor_parameter_688: name = "parameter_688" + original_name = "linear_90.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.224539") @@ -7551,6 +8240,7 @@ class Program_weight_tensor_parameter_688: class Program_weight_tensor_parameter_689: name = "parameter_689" + original_name = "layer_norm_60.b_0" shape = [320] dtype = "float32" min_val = float("-0.825606") @@ -7562,6 +8252,7 @@ class Program_weight_tensor_parameter_689: class Program_weight_tensor_parameter_690: name = "parameter_690" + original_name = "layer_norm_60.w_0" shape = [320] dtype = "float32" min_val = float("0.232073") @@ -7573,6 +8264,7 @@ class Program_weight_tensor_parameter_690: class Program_weight_tensor_parameter_691: name = "parameter_691" + original_name = "linear_89.b_0" shape = [320] dtype = "float32" min_val = float("-0.345987") @@ -7584,6 +8276,7 @@ class Program_weight_tensor_parameter_691: class Program_weight_tensor_parameter_692: name = "parameter_692" + original_name = "linear_89.w_0" shape = [1280, 320] dtype = "float32" min_val = float("-0.279983") @@ -7595,6 +8288,7 @@ class Program_weight_tensor_parameter_692: class Program_weight_tensor_parameter_693: name = "parameter_693" + original_name = "conv2d_39.b_0" shape = [1280] dtype = "float32" min_val = float("-1.54822") @@ -7606,6 +8300,7 @@ class Program_weight_tensor_parameter_693: class Program_weight_tensor_parameter_694: name = "parameter_694" + original_name = "conv2d_39.w_0" shape = [1280, 1, 3, 3] dtype = "float32" min_val = float("-0.324463") @@ -7617,6 +8312,7 @@ class Program_weight_tensor_parameter_694: class Program_weight_tensor_parameter_695: name = "parameter_695" + original_name = "linear_88.b_0" shape = [1280] dtype = "float32" min_val = float("-1.7673") @@ -7628,6 +8324,7 @@ class Program_weight_tensor_parameter_695: class Program_weight_tensor_parameter_696: name = "parameter_696" + original_name = "linear_88.w_0" shape = [320, 1280] dtype = "float32" min_val = float("-0.301027") @@ -7639,6 +8336,7 @@ class Program_weight_tensor_parameter_696: class Program_weight_tensor_parameter_697: name = "parameter_697" + original_name = "layer_norm_59.b_0" shape = [320] dtype = "float32" min_val = float("-1.55767") @@ -7650,6 +8348,7 @@ class Program_weight_tensor_parameter_697: class Program_weight_tensor_parameter_698: name = "parameter_698" + original_name = "layer_norm_59.w_0" shape = [320] dtype = "float32" min_val = float("0.708795") @@ -7661,6 +8360,7 @@ class Program_weight_tensor_parameter_698: class Program_weight_tensor_parameter_699: name = "parameter_699" + original_name = "linear_87.b_0" shape = [320] dtype = "float32" min_val = float("-0.249204") @@ -7672,6 +8372,7 @@ class Program_weight_tensor_parameter_699: class Program_weight_tensor_parameter_700: name = "parameter_700" + original_name = "linear_87.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.244382") @@ -7683,6 +8384,7 @@ class Program_weight_tensor_parameter_700: class Program_weight_tensor_parameter_701: name = "parameter_701" + original_name = "linear_86.b_0" shape = [640] dtype = "float32" min_val = float("-0.68596") @@ -7694,6 +8396,7 @@ class Program_weight_tensor_parameter_701: class Program_weight_tensor_parameter_702: name = "parameter_702" + original_name = "linear_86.w_0" shape = [320, 640] dtype = "float32" min_val = float("-0.390033") @@ -7705,6 +8408,7 @@ class Program_weight_tensor_parameter_702: class Program_weight_tensor_parameter_703: name = "parameter_703" + original_name = "layer_norm_58.b_0" shape = [320] dtype = "float32" min_val = float("-0.162135") @@ -7716,6 +8420,7 @@ class Program_weight_tensor_parameter_703: class Program_weight_tensor_parameter_704: name = "parameter_704" + original_name = "layer_norm_58.w_0" shape = [320] dtype = "float32" min_val = float("-0.0184501") @@ -7727,6 +8432,7 @@ class Program_weight_tensor_parameter_704: class Program_weight_tensor_parameter_705: name = "parameter_705" + original_name = "conv2d_38.b_0" shape = [320] dtype = "float32" min_val = float("-3.44579") @@ -7738,6 +8444,7 @@ class Program_weight_tensor_parameter_705: class Program_weight_tensor_parameter_706: name = "parameter_706" + original_name = "conv2d_38.w_0" shape = [320, 320, 2, 2] dtype = "float32" min_val = float("-0.520433") @@ -7749,6 +8456,7 @@ class Program_weight_tensor_parameter_706: class Program_weight_tensor_parameter_707: name = "parameter_707" + original_name = "linear_85.b_0" shape = [320] dtype = "float32" min_val = float("-3.06335") @@ -7760,6 +8468,7 @@ class Program_weight_tensor_parameter_707: class Program_weight_tensor_parameter_708: name = "parameter_708" + original_name = "linear_85.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.24765") @@ -7771,6 +8480,7 @@ class Program_weight_tensor_parameter_708: class Program_weight_tensor_parameter_709: name = "parameter_709" + original_name = "layer_norm_57.b_0" shape = [320] dtype = "float32" min_val = float("-1.00251") @@ -7782,6 +8492,7 @@ class Program_weight_tensor_parameter_709: class Program_weight_tensor_parameter_710: name = "parameter_710" + original_name = "layer_norm_57.w_0" shape = [320] dtype = "float32" min_val = float("0.286313") @@ -7793,6 +8504,7 @@ class Program_weight_tensor_parameter_710: class Program_weight_tensor_parameter_711: name = "parameter_711" + original_name = "linear_84.b_0" shape = [320] dtype = "float32" min_val = float("-0.458921") @@ -7804,6 +8516,7 @@ class Program_weight_tensor_parameter_711: class Program_weight_tensor_parameter_712: name = "parameter_712" + original_name = "linear_84.w_0" shape = [1280, 320] dtype = "float32" min_val = float("-0.310352") @@ -7815,6 +8528,7 @@ class Program_weight_tensor_parameter_712: class Program_weight_tensor_parameter_713: name = "parameter_713" + original_name = "conv2d_37.b_0" shape = [1280] dtype = "float32" min_val = float("-1.6509") @@ -7826,6 +8540,7 @@ class Program_weight_tensor_parameter_713: class Program_weight_tensor_parameter_714: name = "parameter_714" + original_name = "conv2d_37.w_0" shape = [1280, 1, 3, 3] dtype = "float32" min_val = float("-0.351299") @@ -7837,6 +8552,7 @@ class Program_weight_tensor_parameter_714: class Program_weight_tensor_parameter_715: name = "parameter_715" + original_name = "linear_83.b_0" shape = [1280] dtype = "float32" min_val = float("-1.7766") @@ -7848,6 +8564,7 @@ class Program_weight_tensor_parameter_715: class Program_weight_tensor_parameter_716: name = "parameter_716" + original_name = "linear_83.w_0" shape = [320, 1280] dtype = "float32" min_val = float("-0.310995") @@ -7859,6 +8576,7 @@ class Program_weight_tensor_parameter_716: class Program_weight_tensor_parameter_717: name = "parameter_717" + original_name = "layer_norm_56.b_0" shape = [320] dtype = "float32" min_val = float("-1.47415") @@ -7870,6 +8588,7 @@ class Program_weight_tensor_parameter_717: class Program_weight_tensor_parameter_718: name = "parameter_718" + original_name = "layer_norm_56.w_0" shape = [320] dtype = "float32" min_val = float("0.533458") @@ -7881,6 +8600,7 @@ class Program_weight_tensor_parameter_718: class Program_weight_tensor_parameter_719: name = "parameter_719" + original_name = "linear_82.b_0" shape = [320] dtype = "float32" min_val = float("-0.325561") @@ -7892,6 +8612,7 @@ class Program_weight_tensor_parameter_719: class Program_weight_tensor_parameter_720: name = "parameter_720" + original_name = "linear_82.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.327479") @@ -7903,6 +8624,7 @@ class Program_weight_tensor_parameter_720: class Program_weight_tensor_parameter_721: name = "parameter_721" + original_name = "linear_81.b_0" shape = [640] dtype = "float32" min_val = float("-0.755022") @@ -7914,6 +8636,7 @@ class Program_weight_tensor_parameter_721: class Program_weight_tensor_parameter_722: name = "parameter_722" + original_name = "linear_81.w_0" shape = [320, 640] dtype = "float32" min_val = float("-0.371094") @@ -7925,6 +8648,7 @@ class Program_weight_tensor_parameter_722: class Program_weight_tensor_parameter_723: name = "parameter_723" + original_name = "layer_norm_55.b_0" shape = [320] dtype = "float32" min_val = float("-0.245808") @@ -7936,6 +8660,7 @@ class Program_weight_tensor_parameter_723: class Program_weight_tensor_parameter_724: name = "parameter_724" + original_name = "layer_norm_55.w_0" shape = [320] dtype = "float32" min_val = float("-0.0196085") @@ -7947,6 +8672,7 @@ class Program_weight_tensor_parameter_724: class Program_weight_tensor_parameter_725: name = "parameter_725" + original_name = "conv2d_36.b_0" shape = [320] dtype = "float32" min_val = float("-4.48554") @@ -7958,6 +8684,7 @@ class Program_weight_tensor_parameter_725: class Program_weight_tensor_parameter_726: name = "parameter_726" + original_name = "conv2d_36.w_0" shape = [320, 320, 2, 2] dtype = "float32" min_val = float("-0.245503") @@ -7969,6 +8696,7 @@ class Program_weight_tensor_parameter_726: class Program_weight_tensor_parameter_727: name = "parameter_727" + original_name = "linear_80.b_0" shape = [320] dtype = "float32" min_val = float("-3.06795") @@ -7980,6 +8708,7 @@ class Program_weight_tensor_parameter_727: class Program_weight_tensor_parameter_728: name = "parameter_728" + original_name = "linear_80.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.241762") @@ -7991,6 +8720,7 @@ class Program_weight_tensor_parameter_728: class Program_weight_tensor_parameter_729: name = "parameter_729" + original_name = "layer_norm_54.b_0" shape = [320] dtype = "float32" min_val = float("-0.633756") @@ -8002,6 +8732,7 @@ class Program_weight_tensor_parameter_729: class Program_weight_tensor_parameter_730: name = "parameter_730" + original_name = "layer_norm_54.w_0" shape = [320] dtype = "float32" min_val = float("-0.00551442") @@ -8013,6 +8744,7 @@ class Program_weight_tensor_parameter_730: class Program_weight_tensor_parameter_731: name = "parameter_731" + original_name = "linear_79.b_0" shape = [320] dtype = "float32" min_val = float("-0.323807") @@ -8024,6 +8756,7 @@ class Program_weight_tensor_parameter_731: class Program_weight_tensor_parameter_732: name = "parameter_732" + original_name = "linear_79.w_0" shape = [1280, 320] dtype = "float32" min_val = float("-0.273271") @@ -8035,6 +8768,7 @@ class Program_weight_tensor_parameter_732: class Program_weight_tensor_parameter_733: name = "parameter_733" + original_name = "conv2d_35.b_0" shape = [1280] dtype = "float32" min_val = float("-1.83857") @@ -8046,6 +8780,7 @@ class Program_weight_tensor_parameter_733: class Program_weight_tensor_parameter_734: name = "parameter_734" + original_name = "conv2d_35.w_0" shape = [1280, 1, 3, 3] dtype = "float32" min_val = float("-0.325945") @@ -8057,6 +8792,7 @@ class Program_weight_tensor_parameter_734: class Program_weight_tensor_parameter_735: name = "parameter_735" + original_name = "linear_78.b_0" shape = [1280] dtype = "float32" min_val = float("-1.86429") @@ -8068,6 +8804,7 @@ class Program_weight_tensor_parameter_735: class Program_weight_tensor_parameter_736: name = "parameter_736" + original_name = "linear_78.w_0" shape = [320, 1280] dtype = "float32" min_val = float("-0.292427") @@ -8079,6 +8816,7 @@ class Program_weight_tensor_parameter_736: class Program_weight_tensor_parameter_737: name = "parameter_737" + original_name = "layer_norm_53.b_0" shape = [320] dtype = "float32" min_val = float("-1.49528") @@ -8090,6 +8828,7 @@ class Program_weight_tensor_parameter_737: class Program_weight_tensor_parameter_738: name = "parameter_738" + original_name = "layer_norm_53.w_0" shape = [320] dtype = "float32" min_val = float("0.552866") @@ -8101,6 +8840,7 @@ class Program_weight_tensor_parameter_738: class Program_weight_tensor_parameter_739: name = "parameter_739" + original_name = "linear_77.b_0" shape = [320] dtype = "float32" min_val = float("-0.239403") @@ -8112,6 +8852,7 @@ class Program_weight_tensor_parameter_739: class Program_weight_tensor_parameter_740: name = "parameter_740" + original_name = "linear_77.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.435518") @@ -8123,6 +8864,7 @@ class Program_weight_tensor_parameter_740: class Program_weight_tensor_parameter_741: name = "parameter_741" + original_name = "linear_76.b_0" shape = [640] dtype = "float32" min_val = float("-0.929426") @@ -8134,6 +8876,7 @@ class Program_weight_tensor_parameter_741: class Program_weight_tensor_parameter_742: name = "parameter_742" + original_name = "linear_76.w_0" shape = [320, 640] dtype = "float32" min_val = float("-0.344342") @@ -8145,6 +8888,7 @@ class Program_weight_tensor_parameter_742: class Program_weight_tensor_parameter_743: name = "parameter_743" + original_name = "layer_norm_52.b_0" shape = [320] dtype = "float32" min_val = float("-0.612875") @@ -8156,6 +8900,7 @@ class Program_weight_tensor_parameter_743: class Program_weight_tensor_parameter_744: name = "parameter_744" + original_name = "layer_norm_52.w_0" shape = [320] dtype = "float32" min_val = float("-0.0304679") @@ -8167,6 +8912,7 @@ class Program_weight_tensor_parameter_744: class Program_weight_tensor_parameter_745: name = "parameter_745" + original_name = "conv2d_34.b_0" shape = [320] dtype = "float32" min_val = float("-4.2544") @@ -8178,6 +8924,7 @@ class Program_weight_tensor_parameter_745: class Program_weight_tensor_parameter_746: name = "parameter_746" + original_name = "conv2d_34.w_0" shape = [320, 320, 2, 2] dtype = "float32" min_val = float("-0.343695") @@ -8189,6 +8936,7 @@ class Program_weight_tensor_parameter_746: class Program_weight_tensor_parameter_747: name = "parameter_747" + original_name = "linear_75.b_0" shape = [320] dtype = "float32" min_val = float("-3.21412") @@ -8200,6 +8948,7 @@ class Program_weight_tensor_parameter_747: class Program_weight_tensor_parameter_748: name = "parameter_748" + original_name = "linear_75.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.285465") @@ -8211,6 +8960,7 @@ class Program_weight_tensor_parameter_748: class Program_weight_tensor_parameter_749: name = "parameter_749" + original_name = "layer_norm_51.b_0" shape = [320] dtype = "float32" min_val = float("-0.596413") @@ -8222,6 +8972,7 @@ class Program_weight_tensor_parameter_749: class Program_weight_tensor_parameter_750: name = "parameter_750" + original_name = "layer_norm_51.w_0" shape = [320] dtype = "float32" min_val = float("-0.0141684") @@ -8233,6 +8984,7 @@ class Program_weight_tensor_parameter_750: class Program_weight_tensor_parameter_751: name = "parameter_751" + original_name = "linear_74.b_0" shape = [320] dtype = "float32" min_val = float("-0.351025") @@ -8244,6 +8996,7 @@ class Program_weight_tensor_parameter_751: class Program_weight_tensor_parameter_752: name = "parameter_752" + original_name = "linear_74.w_0" shape = [1280, 320] dtype = "float32" min_val = float("-0.270787") @@ -8255,6 +9008,7 @@ class Program_weight_tensor_parameter_752: class Program_weight_tensor_parameter_753: name = "parameter_753" + original_name = "conv2d_33.b_0" shape = [1280] dtype = "float32" min_val = float("-1.73419") @@ -8266,6 +9020,7 @@ class Program_weight_tensor_parameter_753: class Program_weight_tensor_parameter_754: name = "parameter_754" + original_name = "conv2d_33.w_0" shape = [1280, 1, 3, 3] dtype = "float32" min_val = float("-0.322499") @@ -8277,6 +9032,7 @@ class Program_weight_tensor_parameter_754: class Program_weight_tensor_parameter_755: name = "parameter_755" + original_name = "linear_73.b_0" shape = [1280] dtype = "float32" min_val = float("-1.85134") @@ -8288,6 +9044,7 @@ class Program_weight_tensor_parameter_755: class Program_weight_tensor_parameter_756: name = "parameter_756" + original_name = "linear_73.w_0" shape = [320, 1280] dtype = "float32" min_val = float("-0.343489") @@ -8299,6 +9056,7 @@ class Program_weight_tensor_parameter_756: class Program_weight_tensor_parameter_757: name = "parameter_757" + original_name = "layer_norm_50.b_0" shape = [320] dtype = "float32" min_val = float("-1.24976") @@ -8310,6 +9068,7 @@ class Program_weight_tensor_parameter_757: class Program_weight_tensor_parameter_758: name = "parameter_758" + original_name = "layer_norm_50.w_0" shape = [320] dtype = "float32" min_val = float("0.0173939") @@ -8321,6 +9080,7 @@ class Program_weight_tensor_parameter_758: class Program_weight_tensor_parameter_759: name = "parameter_759" + original_name = "linear_72.b_0" shape = [320] dtype = "float32" min_val = float("-0.281464") @@ -8332,6 +9092,7 @@ class Program_weight_tensor_parameter_759: class Program_weight_tensor_parameter_760: name = "parameter_760" + original_name = "linear_72.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.343912") @@ -8343,6 +9104,7 @@ class Program_weight_tensor_parameter_760: class Program_weight_tensor_parameter_761: name = "parameter_761" + original_name = "linear_71.b_0" shape = [640] dtype = "float32" min_val = float("-0.720195") @@ -8354,6 +9116,7 @@ class Program_weight_tensor_parameter_761: class Program_weight_tensor_parameter_762: name = "parameter_762" + original_name = "linear_71.w_0" shape = [320, 640] dtype = "float32" min_val = float("-0.277681") @@ -8365,6 +9128,7 @@ class Program_weight_tensor_parameter_762: class Program_weight_tensor_parameter_763: name = "parameter_763" + original_name = "layer_norm_49.b_0" shape = [320] dtype = "float32" min_val = float("-0.370313") @@ -8376,6 +9140,7 @@ class Program_weight_tensor_parameter_763: class Program_weight_tensor_parameter_764: name = "parameter_764" + original_name = "layer_norm_49.w_0" shape = [320] dtype = "float32" min_val = float("-0.0398942") @@ -8387,6 +9152,7 @@ class Program_weight_tensor_parameter_764: class Program_weight_tensor_parameter_765: name = "parameter_765" + original_name = "conv2d_32.b_0" shape = [320] dtype = "float32" min_val = float("-2.46804") @@ -8398,6 +9164,7 @@ class Program_weight_tensor_parameter_765: class Program_weight_tensor_parameter_766: name = "parameter_766" + original_name = "conv2d_32.w_0" shape = [320, 320, 2, 2] dtype = "float32" min_val = float("-0.233878") @@ -8409,6 +9176,7 @@ class Program_weight_tensor_parameter_766: class Program_weight_tensor_parameter_767: name = "parameter_767" + original_name = "linear_70.b_0" shape = [320] dtype = "float32" min_val = float("-2.14641") @@ -8420,6 +9188,7 @@ class Program_weight_tensor_parameter_767: class Program_weight_tensor_parameter_768: name = "parameter_768" + original_name = "linear_70.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.224681") @@ -8431,6 +9200,7 @@ class Program_weight_tensor_parameter_768: class Program_weight_tensor_parameter_769: name = "parameter_769" + original_name = "layer_norm_48.b_0" shape = [320] dtype = "float32" min_val = float("-0.427028") @@ -8442,6 +9212,7 @@ class Program_weight_tensor_parameter_769: class Program_weight_tensor_parameter_770: name = "parameter_770" + original_name = "layer_norm_48.w_0" shape = [320] dtype = "float32" min_val = float("-0.00800239") @@ -8453,6 +9224,7 @@ class Program_weight_tensor_parameter_770: class Program_weight_tensor_parameter_771: name = "parameter_771" + original_name = "linear_69.b_0" shape = [320] dtype = "float32" min_val = float("-0.321276") @@ -8464,6 +9236,7 @@ class Program_weight_tensor_parameter_771: class Program_weight_tensor_parameter_772: name = "parameter_772" + original_name = "linear_69.w_0" shape = [1280, 320] dtype = "float32" min_val = float("-0.289418") @@ -8475,6 +9248,7 @@ class Program_weight_tensor_parameter_772: class Program_weight_tensor_parameter_773: name = "parameter_773" + original_name = "conv2d_31.b_0" shape = [1280] dtype = "float32" min_val = float("-1.6611") @@ -8486,6 +9260,7 @@ class Program_weight_tensor_parameter_773: class Program_weight_tensor_parameter_774: name = "parameter_774" + original_name = "conv2d_31.w_0" shape = [1280, 1, 3, 3] dtype = "float32" min_val = float("-0.338334") @@ -8497,6 +9272,7 @@ class Program_weight_tensor_parameter_774: class Program_weight_tensor_parameter_775: name = "parameter_775" + original_name = "linear_68.b_0" shape = [1280] dtype = "float32" min_val = float("-1.58606") @@ -8508,6 +9284,7 @@ class Program_weight_tensor_parameter_775: class Program_weight_tensor_parameter_776: name = "parameter_776" + original_name = "linear_68.w_0" shape = [320, 1280] dtype = "float32" min_val = float("-0.340049") @@ -8519,6 +9296,7 @@ class Program_weight_tensor_parameter_776: class Program_weight_tensor_parameter_777: name = "parameter_777" + original_name = "layer_norm_47.b_0" shape = [320] dtype = "float32" min_val = float("-1.41946") @@ -8530,6 +9308,7 @@ class Program_weight_tensor_parameter_777: class Program_weight_tensor_parameter_778: name = "parameter_778" + original_name = "layer_norm_47.w_0" shape = [320] dtype = "float32" min_val = float("0.000992939") @@ -8541,6 +9320,7 @@ class Program_weight_tensor_parameter_778: class Program_weight_tensor_parameter_779: name = "parameter_779" + original_name = "linear_67.b_0" shape = [320] dtype = "float32" min_val = float("-0.259159") @@ -8552,6 +9332,7 @@ class Program_weight_tensor_parameter_779: class Program_weight_tensor_parameter_780: name = "parameter_780" + original_name = "linear_67.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.252015") @@ -8563,6 +9344,7 @@ class Program_weight_tensor_parameter_780: class Program_weight_tensor_parameter_781: name = "parameter_781" + original_name = "linear_66.b_0" shape = [640] dtype = "float32" min_val = float("-0.823204") @@ -8574,6 +9356,7 @@ class Program_weight_tensor_parameter_781: class Program_weight_tensor_parameter_782: name = "parameter_782" + original_name = "linear_66.w_0" shape = [320, 640] dtype = "float32" min_val = float("-0.277735") @@ -8585,6 +9368,7 @@ class Program_weight_tensor_parameter_782: class Program_weight_tensor_parameter_783: name = "parameter_783" + original_name = "layer_norm_46.b_0" shape = [320] dtype = "float32" min_val = float("-0.347667") @@ -8596,6 +9380,7 @@ class Program_weight_tensor_parameter_783: class Program_weight_tensor_parameter_784: name = "parameter_784" + original_name = "layer_norm_46.w_0" shape = [320] dtype = "float32" min_val = float("-0.0349864") @@ -8607,6 +9392,7 @@ class Program_weight_tensor_parameter_784: class Program_weight_tensor_parameter_785: name = "parameter_785" + original_name = "conv2d_30.b_0" shape = [320] dtype = "float32" min_val = float("-2.95946") @@ -8618,6 +9404,7 @@ class Program_weight_tensor_parameter_785: class Program_weight_tensor_parameter_786: name = "parameter_786" + original_name = "conv2d_30.w_0" shape = [320, 320, 2, 2] dtype = "float32" min_val = float("-0.242017") @@ -8629,6 +9416,7 @@ class Program_weight_tensor_parameter_786: class Program_weight_tensor_parameter_787: name = "parameter_787" + original_name = "linear_65.b_0" shape = [320] dtype = "float32" min_val = float("-2.32294") @@ -8640,6 +9428,7 @@ class Program_weight_tensor_parameter_787: class Program_weight_tensor_parameter_788: name = "parameter_788" + original_name = "linear_65.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.270416") @@ -8651,6 +9440,7 @@ class Program_weight_tensor_parameter_788: class Program_weight_tensor_parameter_789: name = "parameter_789" + original_name = "layer_norm_45.b_0" shape = [320] dtype = "float32" min_val = float("-0.651834") @@ -8662,6 +9452,7 @@ class Program_weight_tensor_parameter_789: class Program_weight_tensor_parameter_790: name = "parameter_790" + original_name = "layer_norm_45.w_0" shape = [320] dtype = "float32" min_val = float("-0.00631106") @@ -8673,6 +9464,7 @@ class Program_weight_tensor_parameter_790: class Program_weight_tensor_parameter_791: name = "parameter_791" + original_name = "linear_64.b_0" shape = [320] dtype = "float32" min_val = float("-0.342766") @@ -8684,6 +9476,7 @@ class Program_weight_tensor_parameter_791: class Program_weight_tensor_parameter_792: name = "parameter_792" + original_name = "linear_64.w_0" shape = [1280, 320] dtype = "float32" min_val = float("-0.234695") @@ -8695,6 +9488,7 @@ class Program_weight_tensor_parameter_792: class Program_weight_tensor_parameter_793: name = "parameter_793" + original_name = "conv2d_29.b_0" shape = [1280] dtype = "float32" min_val = float("-1.64185") @@ -8706,6 +9500,7 @@ class Program_weight_tensor_parameter_793: class Program_weight_tensor_parameter_794: name = "parameter_794" + original_name = "conv2d_29.w_0" shape = [1280, 1, 3, 3] dtype = "float32" min_val = float("-0.285239") @@ -8717,6 +9512,7 @@ class Program_weight_tensor_parameter_794: class Program_weight_tensor_parameter_795: name = "parameter_795" + original_name = "linear_63.b_0" shape = [1280] dtype = "float32" min_val = float("-2.49557") @@ -8728,6 +9524,7 @@ class Program_weight_tensor_parameter_795: class Program_weight_tensor_parameter_796: name = "parameter_796" + original_name = "linear_63.w_0" shape = [320, 1280] dtype = "float32" min_val = float("-0.290203") @@ -8739,6 +9536,7 @@ class Program_weight_tensor_parameter_796: class Program_weight_tensor_parameter_797: name = "parameter_797" + original_name = "layer_norm_44.b_0" shape = [320] dtype = "float32" min_val = float("-1.51485") @@ -8750,6 +9548,7 @@ class Program_weight_tensor_parameter_797: class Program_weight_tensor_parameter_798: name = "parameter_798" + original_name = "layer_norm_44.w_0" shape = [320] dtype = "float32" min_val = float("0.00436043") @@ -8761,6 +9560,7 @@ class Program_weight_tensor_parameter_798: class Program_weight_tensor_parameter_799: name = "parameter_799" + original_name = "linear_62.b_0" shape = [320] dtype = "float32" min_val = float("-0.159094") @@ -8772,6 +9572,7 @@ class Program_weight_tensor_parameter_799: class Program_weight_tensor_parameter_800: name = "parameter_800" + original_name = "linear_62.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.246641") @@ -8783,6 +9584,7 @@ class Program_weight_tensor_parameter_800: class Program_weight_tensor_parameter_801: name = "parameter_801" + original_name = "linear_61.b_0" shape = [640] dtype = "float32" min_val = float("-0.845385") @@ -8794,6 +9596,7 @@ class Program_weight_tensor_parameter_801: class Program_weight_tensor_parameter_802: name = "parameter_802" + original_name = "linear_61.w_0" shape = [320, 640] dtype = "float32" min_val = float("-0.279825") @@ -8805,6 +9608,7 @@ class Program_weight_tensor_parameter_802: class Program_weight_tensor_parameter_803: name = "parameter_803" + original_name = "layer_norm_43.b_0" shape = [320] dtype = "float32" min_val = float("-0.217396") @@ -8816,6 +9620,7 @@ class Program_weight_tensor_parameter_803: class Program_weight_tensor_parameter_804: name = "parameter_804" + original_name = "layer_norm_43.w_0" shape = [320] dtype = "float32" min_val = float("-0.0153662") @@ -8827,6 +9632,7 @@ class Program_weight_tensor_parameter_804: class Program_weight_tensor_parameter_805: name = "parameter_805" + original_name = "conv2d_28.b_0" shape = [320] dtype = "float32" min_val = float("-3.81986") @@ -8838,6 +9644,7 @@ class Program_weight_tensor_parameter_805: class Program_weight_tensor_parameter_806: name = "parameter_806" + original_name = "conv2d_28.w_0" shape = [320, 320, 2, 2] dtype = "float32" min_val = float("-0.280749") @@ -8849,6 +9656,7 @@ class Program_weight_tensor_parameter_806: class Program_weight_tensor_parameter_807: name = "parameter_807" + original_name = "linear_60.b_0" shape = [320] dtype = "float32" min_val = float("-2.56962") @@ -8860,6 +9668,7 @@ class Program_weight_tensor_parameter_807: class Program_weight_tensor_parameter_808: name = "parameter_808" + original_name = "linear_60.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.234366") @@ -8871,6 +9680,7 @@ class Program_weight_tensor_parameter_808: class Program_weight_tensor_parameter_809: name = "parameter_809" + original_name = "layer_norm_42.b_0" shape = [320] dtype = "float32" min_val = float("-0.640326") @@ -8882,6 +9692,7 @@ class Program_weight_tensor_parameter_809: class Program_weight_tensor_parameter_810: name = "parameter_810" + original_name = "layer_norm_42.w_0" shape = [320] dtype = "float32" min_val = float("0.000632062") @@ -8893,6 +9704,7 @@ class Program_weight_tensor_parameter_810: class Program_weight_tensor_parameter_811: name = "parameter_811" + original_name = "linear_59.b_0" shape = [320] dtype = "float32" min_val = float("-0.554298") @@ -8904,6 +9716,7 @@ class Program_weight_tensor_parameter_811: class Program_weight_tensor_parameter_812: name = "parameter_812" + original_name = "linear_59.w_0" shape = [1280, 320] dtype = "float32" min_val = float("-0.230125") @@ -8915,6 +9728,7 @@ class Program_weight_tensor_parameter_812: class Program_weight_tensor_parameter_813: name = "parameter_813" + original_name = "conv2d_27.b_0" shape = [1280] dtype = "float32" min_val = float("-1.89319") @@ -8926,6 +9740,7 @@ class Program_weight_tensor_parameter_813: class Program_weight_tensor_parameter_814: name = "parameter_814" + original_name = "conv2d_27.w_0" shape = [1280, 1, 3, 3] dtype = "float32" min_val = float("-0.255679") @@ -8937,6 +9752,7 @@ class Program_weight_tensor_parameter_814: class Program_weight_tensor_parameter_815: name = "parameter_815" + original_name = "linear_58.b_0" shape = [1280] dtype = "float32" min_val = float("-1.8712") @@ -8948,6 +9764,7 @@ class Program_weight_tensor_parameter_815: class Program_weight_tensor_parameter_816: name = "parameter_816" + original_name = "linear_58.w_0" shape = [320, 1280] dtype = "float32" min_val = float("-0.261501") @@ -8959,6 +9776,7 @@ class Program_weight_tensor_parameter_816: class Program_weight_tensor_parameter_817: name = "parameter_817" + original_name = "layer_norm_41.b_0" shape = [320] dtype = "float32" min_val = float("-1.78701") @@ -8970,6 +9788,7 @@ class Program_weight_tensor_parameter_817: class Program_weight_tensor_parameter_818: name = "parameter_818" + original_name = "layer_norm_41.w_0" shape = [320] dtype = "float32" min_val = float("0.00698113") @@ -8981,6 +9800,7 @@ class Program_weight_tensor_parameter_818: class Program_weight_tensor_parameter_819: name = "parameter_819" + original_name = "linear_57.b_0" shape = [320] dtype = "float32" min_val = float("-0.120175") @@ -8992,6 +9812,7 @@ class Program_weight_tensor_parameter_819: class Program_weight_tensor_parameter_820: name = "parameter_820" + original_name = "linear_57.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.197257") @@ -9003,6 +9824,7 @@ class Program_weight_tensor_parameter_820: class Program_weight_tensor_parameter_821: name = "parameter_821" + original_name = "linear_56.b_0" shape = [640] dtype = "float32" min_val = float("-0.508992") @@ -9014,6 +9836,7 @@ class Program_weight_tensor_parameter_821: class Program_weight_tensor_parameter_822: name = "parameter_822" + original_name = "linear_56.w_0" shape = [320, 640] dtype = "float32" min_val = float("-0.332737") @@ -9025,6 +9848,7 @@ class Program_weight_tensor_parameter_822: class Program_weight_tensor_parameter_823: name = "parameter_823" + original_name = "layer_norm_40.b_0" shape = [320] dtype = "float32" min_val = float("-0.122372") @@ -9036,6 +9860,7 @@ class Program_weight_tensor_parameter_823: class Program_weight_tensor_parameter_824: name = "parameter_824" + original_name = "layer_norm_40.w_0" shape = [320] dtype = "float32" min_val = float("-0.0118366") @@ -9047,6 +9872,7 @@ class Program_weight_tensor_parameter_824: class Program_weight_tensor_parameter_825: name = "parameter_825" + original_name = "conv2d_26.b_0" shape = [320] dtype = "float32" min_val = float("-2.82795") @@ -9058,6 +9884,7 @@ class Program_weight_tensor_parameter_825: class Program_weight_tensor_parameter_826: name = "parameter_826" + original_name = "conv2d_26.w_0" shape = [320, 320, 2, 2] dtype = "float32" min_val = float("-0.238056") @@ -9069,6 +9896,7 @@ class Program_weight_tensor_parameter_826: class Program_weight_tensor_parameter_827: name = "parameter_827" + original_name = "linear_55.b_0" shape = [320] dtype = "float32" min_val = float("-2.78906") @@ -9080,6 +9908,7 @@ class Program_weight_tensor_parameter_827: class Program_weight_tensor_parameter_828: name = "parameter_828" + original_name = "linear_55.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.279395") @@ -9091,6 +9920,7 @@ class Program_weight_tensor_parameter_828: class Program_weight_tensor_parameter_829: name = "parameter_829" + original_name = "layer_norm_39.b_0" shape = [320] dtype = "float32" min_val = float("-0.701993") @@ -9102,6 +9932,7 @@ class Program_weight_tensor_parameter_829: class Program_weight_tensor_parameter_830: name = "parameter_830" + original_name = "layer_norm_39.w_0" shape = [320] dtype = "float32" min_val = float("-0.00840128") @@ -9113,6 +9944,7 @@ class Program_weight_tensor_parameter_830: class Program_weight_tensor_parameter_831: name = "parameter_831" + original_name = "linear_54.b_0" shape = [320] dtype = "float32" min_val = float("-0.858084") @@ -9124,6 +9956,7 @@ class Program_weight_tensor_parameter_831: class Program_weight_tensor_parameter_832: name = "parameter_832" + original_name = "linear_54.w_0" shape = [1280, 320] dtype = "float32" min_val = float("-0.249266") @@ -9135,6 +9968,7 @@ class Program_weight_tensor_parameter_832: class Program_weight_tensor_parameter_833: name = "parameter_833" + original_name = "conv2d_25.b_0" shape = [1280] dtype = "float32" min_val = float("-1.88723") @@ -9146,6 +9980,7 @@ class Program_weight_tensor_parameter_833: class Program_weight_tensor_parameter_834: name = "parameter_834" + original_name = "conv2d_25.w_0" shape = [1280, 1, 3, 3] dtype = "float32" min_val = float("-0.254703") @@ -9157,6 +9992,7 @@ class Program_weight_tensor_parameter_834: class Program_weight_tensor_parameter_835: name = "parameter_835" + original_name = "linear_53.b_0" shape = [1280] dtype = "float32" min_val = float("-3.42029") @@ -9168,6 +10004,7 @@ class Program_weight_tensor_parameter_835: class Program_weight_tensor_parameter_836: name = "parameter_836" + original_name = "linear_53.w_0" shape = [320, 1280] dtype = "float32" min_val = float("-0.292305") @@ -9179,6 +10016,7 @@ class Program_weight_tensor_parameter_836: class Program_weight_tensor_parameter_837: name = "parameter_837" + original_name = "layer_norm_38.b_0" shape = [320] dtype = "float32" min_val = float("-1.62864") @@ -9190,6 +10028,7 @@ class Program_weight_tensor_parameter_837: class Program_weight_tensor_parameter_838: name = "parameter_838" + original_name = "layer_norm_38.w_0" shape = [320] dtype = "float32" min_val = float("0.00519936") @@ -9201,6 +10040,7 @@ class Program_weight_tensor_parameter_838: class Program_weight_tensor_parameter_839: name = "parameter_839" + original_name = "linear_52.b_0" shape = [320] dtype = "float32" min_val = float("-0.222602") @@ -9212,6 +10052,7 @@ class Program_weight_tensor_parameter_839: class Program_weight_tensor_parameter_840: name = "parameter_840" + original_name = "linear_52.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.215186") @@ -9223,6 +10064,7 @@ class Program_weight_tensor_parameter_840: class Program_weight_tensor_parameter_841: name = "parameter_841" + original_name = "linear_51.b_0" shape = [640] dtype = "float32" min_val = float("-0.8417") @@ -9234,6 +10076,7 @@ class Program_weight_tensor_parameter_841: class Program_weight_tensor_parameter_842: name = "parameter_842" + original_name = "linear_51.w_0" shape = [320, 640] dtype = "float32" min_val = float("-0.325541") @@ -9245,6 +10088,7 @@ class Program_weight_tensor_parameter_842: class Program_weight_tensor_parameter_843: name = "parameter_843" + original_name = "layer_norm_37.b_0" shape = [320] dtype = "float32" min_val = float("-0.153755") @@ -9256,6 +10100,7 @@ class Program_weight_tensor_parameter_843: class Program_weight_tensor_parameter_844: name = "parameter_844" + original_name = "layer_norm_37.w_0" shape = [320] dtype = "float32" min_val = float("-0.0198821") @@ -9267,6 +10112,7 @@ class Program_weight_tensor_parameter_844: class Program_weight_tensor_parameter_845: name = "parameter_845" + original_name = "conv2d_24.b_0" shape = [320] dtype = "float32" min_val = float("-3.45501") @@ -9278,6 +10124,7 @@ class Program_weight_tensor_parameter_845: class Program_weight_tensor_parameter_846: name = "parameter_846" + original_name = "conv2d_24.w_0" shape = [320, 320, 2, 2] dtype = "float32" min_val = float("-0.245286") @@ -9289,6 +10136,7 @@ class Program_weight_tensor_parameter_846: class Program_weight_tensor_parameter_847: name = "parameter_847" + original_name = "linear_50.b_0" shape = [320] dtype = "float32" min_val = float("-3.08291") @@ -9300,6 +10148,7 @@ class Program_weight_tensor_parameter_847: class Program_weight_tensor_parameter_848: name = "parameter_848" + original_name = "linear_50.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.246531") @@ -9311,6 +10160,7 @@ class Program_weight_tensor_parameter_848: class Program_weight_tensor_parameter_849: name = "parameter_849" + original_name = "layer_norm_36.b_0" shape = [320] dtype = "float32" min_val = float("-0.604563") @@ -9322,6 +10172,7 @@ class Program_weight_tensor_parameter_849: class Program_weight_tensor_parameter_850: name = "parameter_850" + original_name = "layer_norm_36.w_0" shape = [320] dtype = "float32" min_val = float("-0.0028891") @@ -9333,6 +10184,7 @@ class Program_weight_tensor_parameter_850: class Program_weight_tensor_parameter_851: name = "parameter_851" + original_name = "linear_49.b_0" shape = [320] dtype = "float32" min_val = float("-1.16453") @@ -9344,6 +10196,7 @@ class Program_weight_tensor_parameter_851: class Program_weight_tensor_parameter_852: name = "parameter_852" + original_name = "linear_49.w_0" shape = [1280, 320] dtype = "float32" min_val = float("-0.229612") @@ -9355,6 +10208,7 @@ class Program_weight_tensor_parameter_852: class Program_weight_tensor_parameter_853: name = "parameter_853" + original_name = "conv2d_23.b_0" shape = [1280] dtype = "float32" min_val = float("-1.67457") @@ -9366,6 +10220,7 @@ class Program_weight_tensor_parameter_853: class Program_weight_tensor_parameter_854: name = "parameter_854" + original_name = "conv2d_23.w_0" shape = [1280, 1, 3, 3] dtype = "float32" min_val = float("-0.237992") @@ -9377,6 +10232,7 @@ class Program_weight_tensor_parameter_854: class Program_weight_tensor_parameter_855: name = "parameter_855" + original_name = "linear_48.b_0" shape = [1280] dtype = "float32" min_val = float("-2.34032") @@ -9388,6 +10244,7 @@ class Program_weight_tensor_parameter_855: class Program_weight_tensor_parameter_856: name = "parameter_856" + original_name = "linear_48.w_0" shape = [320, 1280] dtype = "float32" min_val = float("-0.526324") @@ -9399,6 +10256,7 @@ class Program_weight_tensor_parameter_856: class Program_weight_tensor_parameter_857: name = "parameter_857" + original_name = "layer_norm_35.b_0" shape = [320] dtype = "float32" min_val = float("-1.11349") @@ -9410,6 +10268,7 @@ class Program_weight_tensor_parameter_857: class Program_weight_tensor_parameter_858: name = "parameter_858" + original_name = "layer_norm_35.w_0" shape = [320] dtype = "float32" min_val = float("0.000103087") @@ -9421,6 +10280,7 @@ class Program_weight_tensor_parameter_858: class Program_weight_tensor_parameter_859: name = "parameter_859" + original_name = "linear_47.b_0" shape = [320] dtype = "float32" min_val = float("-0.17885") @@ -9432,6 +10292,7 @@ class Program_weight_tensor_parameter_859: class Program_weight_tensor_parameter_860: name = "parameter_860" + original_name = "linear_47.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.210498") @@ -9443,6 +10304,7 @@ class Program_weight_tensor_parameter_860: class Program_weight_tensor_parameter_861: name = "parameter_861" + original_name = "linear_46.b_0" shape = [640] dtype = "float32" min_val = float("-1.05757") @@ -9454,6 +10316,7 @@ class Program_weight_tensor_parameter_861: class Program_weight_tensor_parameter_862: name = "parameter_862" + original_name = "linear_46.w_0" shape = [320, 640] dtype = "float32" min_val = float("-0.274117") @@ -9465,6 +10328,7 @@ class Program_weight_tensor_parameter_862: class Program_weight_tensor_parameter_863: name = "parameter_863" + original_name = "layer_norm_34.b_0" shape = [320] dtype = "float32" min_val = float("-0.0716366") @@ -9476,6 +10340,7 @@ class Program_weight_tensor_parameter_863: class Program_weight_tensor_parameter_864: name = "parameter_864" + original_name = "layer_norm_34.w_0" shape = [320] dtype = "float32" min_val = float("-0.0904465") @@ -9487,6 +10352,7 @@ class Program_weight_tensor_parameter_864: class Program_weight_tensor_parameter_865: name = "parameter_865" + original_name = "conv2d_22.b_0" shape = [320] dtype = "float32" min_val = float("-2.94481") @@ -9498,6 +10364,7 @@ class Program_weight_tensor_parameter_865: class Program_weight_tensor_parameter_866: name = "parameter_866" + original_name = "conv2d_22.w_0" shape = [320, 320, 2, 2] dtype = "float32" min_val = float("-0.218656") @@ -9509,6 +10376,7 @@ class Program_weight_tensor_parameter_866: class Program_weight_tensor_parameter_867: name = "parameter_867" + original_name = "linear_45.b_0" shape = [320] dtype = "float32" min_val = float("-2.98089") @@ -9520,6 +10388,7 @@ class Program_weight_tensor_parameter_867: class Program_weight_tensor_parameter_868: name = "parameter_868" + original_name = "linear_45.w_0" shape = [320, 320] dtype = "float32" min_val = float("-0.342989") @@ -9531,6 +10400,7 @@ class Program_weight_tensor_parameter_868: class Program_weight_tensor_parameter_869: name = "parameter_869" + original_name = "layer_norm_33.b_0" shape = [320] dtype = "float32" min_val = float("-0.944898") @@ -9542,6 +10412,7 @@ class Program_weight_tensor_parameter_869: class Program_weight_tensor_parameter_870: name = "parameter_870" + original_name = "layer_norm_33.w_0" shape = [320] dtype = "float32" min_val = float("-0.00649783") @@ -9553,6 +10424,7 @@ class Program_weight_tensor_parameter_870: class Program_weight_tensor_parameter_871: name = "parameter_871" + original_name = "layer_norm_2.b_0" shape = [320] dtype = "float32" min_val = float("-0.862682") @@ -9564,6 +10436,7 @@ class Program_weight_tensor_parameter_871: class Program_weight_tensor_parameter_872: name = "parameter_872" + original_name = "layer_norm_2.w_0" shape = [320] dtype = "float32" min_val = float("0.168801") @@ -9575,6 +10448,7 @@ class Program_weight_tensor_parameter_872: class Program_weight_tensor_parameter_873: name = "parameter_873" + original_name = "conv2d_2.b_0" shape = [320] dtype = "float32" min_val = float("-10.2868") @@ -9586,6 +10460,7 @@ class Program_weight_tensor_parameter_873: class Program_weight_tensor_parameter_874: name = "parameter_874" + original_name = "conv2d_2.w_0" shape = [320, 128, 3, 3] dtype = "float32" min_val = float("-1.5811") @@ -9597,6 +10472,7 @@ class Program_weight_tensor_parameter_874: class Program_weight_tensor_parameter_875: name = "parameter_875" + original_name = "layer_norm_32.b_0" shape = [128] dtype = "float32" min_val = float("-0.599351") @@ -9608,6 +10484,7 @@ class Program_weight_tensor_parameter_875: class Program_weight_tensor_parameter_876: name = "parameter_876" + original_name = "layer_norm_32.w_0" shape = [128] dtype = "float32" min_val = float("0.064086") @@ -9619,6 +10496,7 @@ class Program_weight_tensor_parameter_876: class Program_weight_tensor_parameter_877: name = "parameter_877" + original_name = "linear_44.b_0" shape = [128] dtype = "float32" min_val = float("-8.87212") @@ -9630,6 +10508,7 @@ class Program_weight_tensor_parameter_877: class Program_weight_tensor_parameter_878: name = "parameter_878" + original_name = "linear_44.w_0" shape = [512, 128] dtype = "float32" min_val = float("-1.53555") @@ -9641,6 +10520,7 @@ class Program_weight_tensor_parameter_878: class Program_weight_tensor_parameter_879: name = "parameter_879" + original_name = "conv2d_21.b_0" shape = [512] dtype = "float32" min_val = float("-1.02268") @@ -9652,6 +10532,7 @@ class Program_weight_tensor_parameter_879: class Program_weight_tensor_parameter_880: name = "parameter_880" + original_name = "conv2d_21.w_0" shape = [512, 1, 3, 3] dtype = "float32" min_val = float("-0.682178") @@ -9663,6 +10544,7 @@ class Program_weight_tensor_parameter_880: class Program_weight_tensor_parameter_881: name = "parameter_881" + original_name = "linear_43.b_0" shape = [512] dtype = "float32" min_val = float("-2.09688") @@ -9674,6 +10556,7 @@ class Program_weight_tensor_parameter_881: class Program_weight_tensor_parameter_882: name = "parameter_882" + original_name = "linear_43.w_0" shape = [128, 512] dtype = "float32" min_val = float("-0.272621") @@ -9685,6 +10568,7 @@ class Program_weight_tensor_parameter_882: class Program_weight_tensor_parameter_883: name = "parameter_883" + original_name = "layer_norm_31.b_0" shape = [128] dtype = "float32" min_val = float("-1.93573") @@ -9696,6 +10580,7 @@ class Program_weight_tensor_parameter_883: class Program_weight_tensor_parameter_884: name = "parameter_884" + original_name = "layer_norm_31.w_0" shape = [128] dtype = "float32" min_val = float("1.37241") @@ -9707,6 +10592,7 @@ class Program_weight_tensor_parameter_884: class Program_weight_tensor_parameter_885: name = "parameter_885" + original_name = "linear_42.b_0" shape = [128] dtype = "float32" min_val = float("-0.857048") @@ -9718,6 +10604,7 @@ class Program_weight_tensor_parameter_885: class Program_weight_tensor_parameter_886: name = "parameter_886" + original_name = "linear_42.w_0" shape = [128, 128] dtype = "float32" min_val = float("-0.633571") @@ -9729,6 +10616,7 @@ class Program_weight_tensor_parameter_886: class Program_weight_tensor_parameter_887: name = "parameter_887" + original_name = "linear_41.b_0" shape = [256] dtype = "float32" min_val = float("-0.842905") @@ -9740,6 +10628,7 @@ class Program_weight_tensor_parameter_887: class Program_weight_tensor_parameter_888: name = "parameter_888" + original_name = "linear_41.w_0" shape = [128, 256] dtype = "float32" min_val = float("-0.374109") @@ -9751,6 +10640,7 @@ class Program_weight_tensor_parameter_888: class Program_weight_tensor_parameter_889: name = "parameter_889" + original_name = "layer_norm_30.b_0" shape = [128] dtype = "float32" min_val = float("-0.22314") @@ -9762,6 +10652,7 @@ class Program_weight_tensor_parameter_889: class Program_weight_tensor_parameter_890: name = "parameter_890" + original_name = "layer_norm_30.w_0" shape = [128] dtype = "float32" min_val = float("-0.0170242") @@ -9773,6 +10664,7 @@ class Program_weight_tensor_parameter_890: class Program_weight_tensor_parameter_891: name = "parameter_891" + original_name = "conv2d_20.b_0" shape = [128] dtype = "float32" min_val = float("-4.08218") @@ -9784,6 +10676,7 @@ class Program_weight_tensor_parameter_891: class Program_weight_tensor_parameter_892: name = "parameter_892" + original_name = "conv2d_20.w_0" shape = [128, 128, 4, 4] dtype = "float32" min_val = float("-0.231279") @@ -9795,6 +10688,7 @@ class Program_weight_tensor_parameter_892: class Program_weight_tensor_parameter_893: name = "parameter_893" + original_name = "linear_40.b_0" shape = [128] dtype = "float32" min_val = float("-2.0309") @@ -9806,6 +10700,7 @@ class Program_weight_tensor_parameter_893: class Program_weight_tensor_parameter_894: name = "parameter_894" + original_name = "linear_40.w_0" shape = [128, 128] dtype = "float32" min_val = float("-0.272538") @@ -9817,6 +10712,7 @@ class Program_weight_tensor_parameter_894: class Program_weight_tensor_parameter_895: name = "parameter_895" + original_name = "layer_norm_29.b_0" shape = [128] dtype = "float32" min_val = float("-0.871294") @@ -9828,6 +10724,7 @@ class Program_weight_tensor_parameter_895: class Program_weight_tensor_parameter_896: name = "parameter_896" + original_name = "layer_norm_29.w_0" shape = [128] dtype = "float32" min_val = float("0.206867") @@ -9839,6 +10736,7 @@ class Program_weight_tensor_parameter_896: class Program_weight_tensor_parameter_897: name = "parameter_897" + original_name = "linear_39.b_0" shape = [128] dtype = "float32" min_val = float("-3.33076") @@ -9850,6 +10748,7 @@ class Program_weight_tensor_parameter_897: class Program_weight_tensor_parameter_898: name = "parameter_898" + original_name = "linear_39.w_0" shape = [512, 128] dtype = "float32" min_val = float("-1.12953") @@ -9861,6 +10760,7 @@ class Program_weight_tensor_parameter_898: class Program_weight_tensor_parameter_899: name = "parameter_899" + original_name = "conv2d_19.b_0" shape = [512] dtype = "float32" min_val = float("-1.15296") @@ -9872,6 +10772,7 @@ class Program_weight_tensor_parameter_899: class Program_weight_tensor_parameter_900: name = "parameter_900" + original_name = "conv2d_19.w_0" shape = [512, 1, 3, 3] dtype = "float32" min_val = float("-0.284919") @@ -9883,6 +10784,7 @@ class Program_weight_tensor_parameter_900: class Program_weight_tensor_parameter_901: name = "parameter_901" + original_name = "linear_38.b_0" shape = [512] dtype = "float32" min_val = float("-1.8062") @@ -9894,6 +10796,7 @@ class Program_weight_tensor_parameter_901: class Program_weight_tensor_parameter_902: name = "parameter_902" + original_name = "linear_38.w_0" shape = [128, 512] dtype = "float32" min_val = float("-0.286118") @@ -9905,6 +10808,7 @@ class Program_weight_tensor_parameter_902: class Program_weight_tensor_parameter_903: name = "parameter_903" + original_name = "layer_norm_28.b_0" shape = [128] dtype = "float32" min_val = float("-1.1835") @@ -9916,6 +10820,7 @@ class Program_weight_tensor_parameter_903: class Program_weight_tensor_parameter_904: name = "parameter_904" + original_name = "layer_norm_28.w_0" shape = [128] dtype = "float32" min_val = float("1.7787") @@ -9927,6 +10832,7 @@ class Program_weight_tensor_parameter_904: class Program_weight_tensor_parameter_905: name = "parameter_905" + original_name = "linear_37.b_0" shape = [128] dtype = "float32" min_val = float("-0.0781033") @@ -9938,6 +10844,7 @@ class Program_weight_tensor_parameter_905: class Program_weight_tensor_parameter_906: name = "parameter_906" + original_name = "linear_37.w_0" shape = [128, 128] dtype = "float32" min_val = float("-0.555115") @@ -9949,6 +10856,7 @@ class Program_weight_tensor_parameter_906: class Program_weight_tensor_parameter_907: name = "parameter_907" + original_name = "linear_36.b_0" shape = [256] dtype = "float32" min_val = float("-1.38446") @@ -9960,6 +10868,7 @@ class Program_weight_tensor_parameter_907: class Program_weight_tensor_parameter_908: name = "parameter_908" + original_name = "linear_36.w_0" shape = [128, 256] dtype = "float32" min_val = float("-0.391075") @@ -9971,6 +10880,7 @@ class Program_weight_tensor_parameter_908: class Program_weight_tensor_parameter_909: name = "parameter_909" + original_name = "layer_norm_27.b_0" shape = [128] dtype = "float32" min_val = float("-0.407634") @@ -9982,6 +10892,7 @@ class Program_weight_tensor_parameter_909: class Program_weight_tensor_parameter_910: name = "parameter_910" + original_name = "layer_norm_27.w_0" shape = [128] dtype = "float32" min_val = float("-0.0471777") @@ -9993,6 +10904,7 @@ class Program_weight_tensor_parameter_910: class Program_weight_tensor_parameter_911: name = "parameter_911" + original_name = "conv2d_18.b_0" shape = [128] dtype = "float32" min_val = float("-3.19268") @@ -10004,6 +10916,7 @@ class Program_weight_tensor_parameter_911: class Program_weight_tensor_parameter_912: name = "parameter_912" + original_name = "conv2d_18.w_0" shape = [128, 128, 4, 4] dtype = "float32" min_val = float("-0.257424") @@ -10015,6 +10928,7 @@ class Program_weight_tensor_parameter_912: class Program_weight_tensor_parameter_913: name = "parameter_913" + original_name = "linear_35.b_0" shape = [128] dtype = "float32" min_val = float("-2.77905") @@ -10026,6 +10940,7 @@ class Program_weight_tensor_parameter_913: class Program_weight_tensor_parameter_914: name = "parameter_914" + original_name = "linear_35.w_0" shape = [128, 128] dtype = "float32" min_val = float("-0.343239") @@ -10037,6 +10952,7 @@ class Program_weight_tensor_parameter_914: class Program_weight_tensor_parameter_915: name = "parameter_915" + original_name = "layer_norm_26.b_0" shape = [128] dtype = "float32" min_val = float("-0.647584") @@ -10048,6 +10964,7 @@ class Program_weight_tensor_parameter_915: class Program_weight_tensor_parameter_916: name = "parameter_916" + original_name = "layer_norm_26.w_0" shape = [128] dtype = "float32" min_val = float("0.355097") @@ -10059,6 +10976,7 @@ class Program_weight_tensor_parameter_916: class Program_weight_tensor_parameter_917: name = "parameter_917" + original_name = "linear_34.b_0" shape = [128] dtype = "float32" min_val = float("-1.5894") @@ -10070,6 +10988,7 @@ class Program_weight_tensor_parameter_917: class Program_weight_tensor_parameter_918: name = "parameter_918" + original_name = "linear_34.w_0" shape = [512, 128] dtype = "float32" min_val = float("-1.13667") @@ -10081,6 +11000,7 @@ class Program_weight_tensor_parameter_918: class Program_weight_tensor_parameter_919: name = "parameter_919" + original_name = "conv2d_17.b_0" shape = [512] dtype = "float32" min_val = float("-1.55001") @@ -10092,6 +11012,7 @@ class Program_weight_tensor_parameter_919: class Program_weight_tensor_parameter_920: name = "parameter_920" + original_name = "conv2d_17.w_0" shape = [512, 1, 3, 3] dtype = "float32" min_val = float("-0.269756") @@ -10103,6 +11024,7 @@ class Program_weight_tensor_parameter_920: class Program_weight_tensor_parameter_921: name = "parameter_921" + original_name = "linear_33.b_0" shape = [512] dtype = "float32" min_val = float("-2.59853") @@ -10114,6 +11036,7 @@ class Program_weight_tensor_parameter_921: class Program_weight_tensor_parameter_922: name = "parameter_922" + original_name = "linear_33.w_0" shape = [128, 512] dtype = "float32" min_val = float("-0.233559") @@ -10125,6 +11048,7 @@ class Program_weight_tensor_parameter_922: class Program_weight_tensor_parameter_923: name = "parameter_923" + original_name = "layer_norm_25.b_0" shape = [128] dtype = "float32" min_val = float("-0.87305") @@ -10136,6 +11060,7 @@ class Program_weight_tensor_parameter_923: class Program_weight_tensor_parameter_924: name = "parameter_924" + original_name = "layer_norm_25.w_0" shape = [128] dtype = "float32" min_val = float("1.41796") @@ -10147,6 +11072,7 @@ class Program_weight_tensor_parameter_924: class Program_weight_tensor_parameter_925: name = "parameter_925" + original_name = "linear_32.b_0" shape = [128] dtype = "float32" min_val = float("-0.0893535") @@ -10158,6 +11084,7 @@ class Program_weight_tensor_parameter_925: class Program_weight_tensor_parameter_926: name = "parameter_926" + original_name = "linear_32.w_0" shape = [128, 128] dtype = "float32" min_val = float("-0.313284") @@ -10169,6 +11096,7 @@ class Program_weight_tensor_parameter_926: class Program_weight_tensor_parameter_927: name = "parameter_927" + original_name = "linear_31.b_0" shape = [256] dtype = "float32" min_val = float("-0.761704") @@ -10180,6 +11108,7 @@ class Program_weight_tensor_parameter_927: class Program_weight_tensor_parameter_928: name = "parameter_928" + original_name = "linear_31.w_0" shape = [128, 256] dtype = "float32" min_val = float("-0.271858") @@ -10191,6 +11120,7 @@ class Program_weight_tensor_parameter_928: class Program_weight_tensor_parameter_929: name = "parameter_929" + original_name = "layer_norm_24.b_0" shape = [128] dtype = "float32" min_val = float("-0.421239") @@ -10202,6 +11132,7 @@ class Program_weight_tensor_parameter_929: class Program_weight_tensor_parameter_930: name = "parameter_930" + original_name = "layer_norm_24.w_0" shape = [128] dtype = "float32" min_val = float("-0.0161549") @@ -10213,6 +11144,7 @@ class Program_weight_tensor_parameter_930: class Program_weight_tensor_parameter_931: name = "parameter_931" + original_name = "conv2d_16.b_0" shape = [128] dtype = "float32" min_val = float("-4.25486") @@ -10224,6 +11156,7 @@ class Program_weight_tensor_parameter_931: class Program_weight_tensor_parameter_932: name = "parameter_932" + original_name = "conv2d_16.w_0" shape = [128, 128, 4, 4] dtype = "float32" min_val = float("-0.283271") @@ -10235,6 +11168,7 @@ class Program_weight_tensor_parameter_932: class Program_weight_tensor_parameter_933: name = "parameter_933" + original_name = "linear_30.b_0" shape = [128] dtype = "float32" min_val = float("-2.67671") @@ -10246,6 +11180,7 @@ class Program_weight_tensor_parameter_933: class Program_weight_tensor_parameter_934: name = "parameter_934" + original_name = "linear_30.w_0" shape = [128, 128] dtype = "float32" min_val = float("-0.243441") @@ -10257,6 +11192,7 @@ class Program_weight_tensor_parameter_934: class Program_weight_tensor_parameter_935: name = "parameter_935" + original_name = "layer_norm_23.b_0" shape = [128] dtype = "float32" min_val = float("-0.741706") @@ -10268,6 +11204,7 @@ class Program_weight_tensor_parameter_935: class Program_weight_tensor_parameter_936: name = "parameter_936" + original_name = "layer_norm_23.w_0" shape = [128] dtype = "float32" min_val = float("0.497877") @@ -10279,6 +11216,7 @@ class Program_weight_tensor_parameter_936: class Program_weight_tensor_parameter_937: name = "parameter_937" + original_name = "linear_29.b_0" shape = [128] dtype = "float32" min_val = float("-0.94379") @@ -10290,6 +11228,7 @@ class Program_weight_tensor_parameter_937: class Program_weight_tensor_parameter_938: name = "parameter_938" + original_name = "linear_29.w_0" shape = [512, 128] dtype = "float32" min_val = float("-0.435286") @@ -10301,6 +11240,7 @@ class Program_weight_tensor_parameter_938: class Program_weight_tensor_parameter_939: name = "parameter_939" + original_name = "conv2d_15.b_0" shape = [512] dtype = "float32" min_val = float("-1.07732") @@ -10312,6 +11252,7 @@ class Program_weight_tensor_parameter_939: class Program_weight_tensor_parameter_940: name = "parameter_940" + original_name = "conv2d_15.w_0" shape = [512, 1, 3, 3] dtype = "float32" min_val = float("-0.311389") @@ -10323,6 +11264,7 @@ class Program_weight_tensor_parameter_940: class Program_weight_tensor_parameter_941: name = "parameter_941" + original_name = "linear_28.b_0" shape = [512] dtype = "float32" min_val = float("-2.61495") @@ -10334,6 +11276,7 @@ class Program_weight_tensor_parameter_941: class Program_weight_tensor_parameter_942: name = "parameter_942" + original_name = "linear_28.w_0" shape = [128, 512] dtype = "float32" min_val = float("-0.329316") @@ -10345,6 +11288,7 @@ class Program_weight_tensor_parameter_942: class Program_weight_tensor_parameter_943: name = "parameter_943" + original_name = "layer_norm_22.b_0" shape = [128] dtype = "float32" min_val = float("-0.838124") @@ -10356,6 +11300,7 @@ class Program_weight_tensor_parameter_943: class Program_weight_tensor_parameter_944: name = "parameter_944" + original_name = "layer_norm_22.w_0" shape = [128] dtype = "float32" min_val = float("1.39675") @@ -10367,6 +11312,7 @@ class Program_weight_tensor_parameter_944: class Program_weight_tensor_parameter_945: name = "parameter_945" + original_name = "linear_27.b_0" shape = [128] dtype = "float32" min_val = float("-0.096383") @@ -10378,6 +11324,7 @@ class Program_weight_tensor_parameter_945: class Program_weight_tensor_parameter_946: name = "parameter_946" + original_name = "linear_27.w_0" shape = [128, 128] dtype = "float32" min_val = float("-0.205983") @@ -10389,6 +11336,7 @@ class Program_weight_tensor_parameter_946: class Program_weight_tensor_parameter_947: name = "parameter_947" + original_name = "linear_26.b_0" shape = [256] dtype = "float32" min_val = float("-0.911494") @@ -10400,6 +11348,7 @@ class Program_weight_tensor_parameter_947: class Program_weight_tensor_parameter_948: name = "parameter_948" + original_name = "linear_26.w_0" shape = [128, 256] dtype = "float32" min_val = float("-0.423463") @@ -10411,6 +11360,7 @@ class Program_weight_tensor_parameter_948: class Program_weight_tensor_parameter_949: name = "parameter_949" + original_name = "layer_norm_21.b_0" shape = [128] dtype = "float32" min_val = float("-0.234736") @@ -10422,6 +11372,7 @@ class Program_weight_tensor_parameter_949: class Program_weight_tensor_parameter_950: name = "parameter_950" + original_name = "layer_norm_21.w_0" shape = [128] dtype = "float32" min_val = float("-0.0113353") @@ -10433,6 +11384,7 @@ class Program_weight_tensor_parameter_950: class Program_weight_tensor_parameter_951: name = "parameter_951" + original_name = "conv2d_14.b_0" shape = [128] dtype = "float32" min_val = float("-2.14563") @@ -10444,6 +11396,7 @@ class Program_weight_tensor_parameter_951: class Program_weight_tensor_parameter_952: name = "parameter_952" + original_name = "conv2d_14.w_0" shape = [128, 128, 4, 4] dtype = "float32" min_val = float("-0.223322") @@ -10455,6 +11408,7 @@ class Program_weight_tensor_parameter_952: class Program_weight_tensor_parameter_953: name = "parameter_953" + original_name = "linear_25.b_0" shape = [128] dtype = "float32" min_val = float("-1.47736") @@ -10466,6 +11420,7 @@ class Program_weight_tensor_parameter_953: class Program_weight_tensor_parameter_954: name = "parameter_954" + original_name = "linear_25.w_0" shape = [128, 128] dtype = "float32" min_val = float("-0.334935") @@ -10477,6 +11432,7 @@ class Program_weight_tensor_parameter_954: class Program_weight_tensor_parameter_955: name = "parameter_955" + original_name = "layer_norm_20.b_0" shape = [128] dtype = "float32" min_val = float("-0.772115") @@ -10488,6 +11444,7 @@ class Program_weight_tensor_parameter_955: class Program_weight_tensor_parameter_956: name = "parameter_956" + original_name = "layer_norm_20.w_0" shape = [128] dtype = "float32" min_val = float("0.357539") @@ -10499,6 +11456,7 @@ class Program_weight_tensor_parameter_956: class Program_weight_tensor_parameter_957: name = "parameter_957" + original_name = "linear_24.b_0" shape = [128] dtype = "float32" min_val = float("-0.418293") @@ -10510,6 +11468,7 @@ class Program_weight_tensor_parameter_957: class Program_weight_tensor_parameter_958: name = "parameter_958" + original_name = "linear_24.w_0" shape = [512, 128] dtype = "float32" min_val = float("-0.649722") @@ -10521,6 +11480,7 @@ class Program_weight_tensor_parameter_958: class Program_weight_tensor_parameter_959: name = "parameter_959" + original_name = "conv2d_13.b_0" shape = [512] dtype = "float32" min_val = float("-1.78176") @@ -10532,6 +11492,7 @@ class Program_weight_tensor_parameter_959: class Program_weight_tensor_parameter_960: name = "parameter_960" + original_name = "conv2d_13.w_0" shape = [512, 1, 3, 3] dtype = "float32" min_val = float("-0.282022") @@ -10543,6 +11504,7 @@ class Program_weight_tensor_parameter_960: class Program_weight_tensor_parameter_961: name = "parameter_961" + original_name = "linear_23.b_0" shape = [512] dtype = "float32" min_val = float("-1.95747") @@ -10554,6 +11516,7 @@ class Program_weight_tensor_parameter_961: class Program_weight_tensor_parameter_962: name = "parameter_962" + original_name = "linear_23.w_0" shape = [128, 512] dtype = "float32" min_val = float("-0.308691") @@ -10565,6 +11528,7 @@ class Program_weight_tensor_parameter_962: class Program_weight_tensor_parameter_963: name = "parameter_963" + original_name = "layer_norm_19.b_0" shape = [128] dtype = "float32" min_val = float("-0.928904") @@ -10576,6 +11540,7 @@ class Program_weight_tensor_parameter_963: class Program_weight_tensor_parameter_964: name = "parameter_964" + original_name = "layer_norm_19.w_0" shape = [128] dtype = "float32" min_val = float("1.79897") @@ -10587,6 +11552,7 @@ class Program_weight_tensor_parameter_964: class Program_weight_tensor_parameter_965: name = "parameter_965" + original_name = "linear_22.b_0" shape = [128] dtype = "float32" min_val = float("-0.269601") @@ -10598,6 +11564,7 @@ class Program_weight_tensor_parameter_965: class Program_weight_tensor_parameter_966: name = "parameter_966" + original_name = "linear_22.w_0" shape = [128, 128] dtype = "float32" min_val = float("-0.190373") @@ -10609,6 +11576,7 @@ class Program_weight_tensor_parameter_966: class Program_weight_tensor_parameter_967: name = "parameter_967" + original_name = "linear_21.b_0" shape = [256] dtype = "float32" min_val = float("-0.99607") @@ -10620,6 +11588,7 @@ class Program_weight_tensor_parameter_967: class Program_weight_tensor_parameter_968: name = "parameter_968" + original_name = "linear_21.w_0" shape = [128, 256] dtype = "float32" min_val = float("-0.32454") @@ -10631,6 +11600,7 @@ class Program_weight_tensor_parameter_968: class Program_weight_tensor_parameter_969: name = "parameter_969" + original_name = "layer_norm_18.b_0" shape = [128] dtype = "float32" min_val = float("-0.227354") @@ -10642,6 +11612,7 @@ class Program_weight_tensor_parameter_969: class Program_weight_tensor_parameter_970: name = "parameter_970" + original_name = "layer_norm_18.w_0" shape = [128] dtype = "float32" min_val = float("-0.00725169") @@ -10653,6 +11624,7 @@ class Program_weight_tensor_parameter_970: class Program_weight_tensor_parameter_971: name = "parameter_971" + original_name = "conv2d_12.b_0" shape = [128] dtype = "float32" min_val = float("-3.19564") @@ -10664,6 +11636,7 @@ class Program_weight_tensor_parameter_971: class Program_weight_tensor_parameter_972: name = "parameter_972" + original_name = "conv2d_12.w_0" shape = [128, 128, 4, 4] dtype = "float32" min_val = float("-0.255212") @@ -10675,6 +11648,7 @@ class Program_weight_tensor_parameter_972: class Program_weight_tensor_parameter_973: name = "parameter_973" + original_name = "linear_20.b_0" shape = [128] dtype = "float32" min_val = float("-1.18171") @@ -10686,6 +11660,7 @@ class Program_weight_tensor_parameter_973: class Program_weight_tensor_parameter_974: name = "parameter_974" + original_name = "linear_20.w_0" shape = [128, 128] dtype = "float32" min_val = float("-0.291488") @@ -10697,6 +11672,7 @@ class Program_weight_tensor_parameter_974: class Program_weight_tensor_parameter_975: name = "parameter_975" + original_name = "layer_norm_17.b_0" shape = [128] dtype = "float32" min_val = float("-0.762324") @@ -10708,6 +11684,7 @@ class Program_weight_tensor_parameter_975: class Program_weight_tensor_parameter_976: name = "parameter_976" + original_name = "layer_norm_17.w_0" shape = [128] dtype = "float32" min_val = float("0.35513") @@ -10719,6 +11696,7 @@ class Program_weight_tensor_parameter_976: class Program_weight_tensor_parameter_977: name = "parameter_977" + original_name = "linear_19.b_0" shape = [128] dtype = "float32" min_val = float("-0.347453") @@ -10730,6 +11708,7 @@ class Program_weight_tensor_parameter_977: class Program_weight_tensor_parameter_978: name = "parameter_978" + original_name = "linear_19.w_0" shape = [512, 128] dtype = "float32" min_val = float("-0.25642") @@ -10741,6 +11720,7 @@ class Program_weight_tensor_parameter_978: class Program_weight_tensor_parameter_979: name = "parameter_979" + original_name = "conv2d_11.b_0" shape = [512] dtype = "float32" min_val = float("-1.6541") @@ -10752,6 +11732,7 @@ class Program_weight_tensor_parameter_979: class Program_weight_tensor_parameter_980: name = "parameter_980" + original_name = "conv2d_11.w_0" shape = [512, 1, 3, 3] dtype = "float32" min_val = float("-0.32945") @@ -10763,6 +11744,7 @@ class Program_weight_tensor_parameter_980: class Program_weight_tensor_parameter_981: name = "parameter_981" + original_name = "linear_18.b_0" shape = [512] dtype = "float32" min_val = float("-3.35571") @@ -10774,6 +11756,7 @@ class Program_weight_tensor_parameter_981: class Program_weight_tensor_parameter_982: name = "parameter_982" + original_name = "linear_18.w_0" shape = [128, 512] dtype = "float32" min_val = float("-0.281701") @@ -10785,6 +11768,7 @@ class Program_weight_tensor_parameter_982: class Program_weight_tensor_parameter_983: name = "parameter_983" + original_name = "layer_norm_16.b_0" shape = [128] dtype = "float32" min_val = float("-1.2706") @@ -10796,6 +11780,7 @@ class Program_weight_tensor_parameter_983: class Program_weight_tensor_parameter_984: name = "parameter_984" + original_name = "layer_norm_16.w_0" shape = [128] dtype = "float32" min_val = float("2.20307") @@ -10807,6 +11792,7 @@ class Program_weight_tensor_parameter_984: class Program_weight_tensor_parameter_985: name = "parameter_985" + original_name = "linear_17.b_0" shape = [128] dtype = "float32" min_val = float("-0.361161") @@ -10818,6 +11804,7 @@ class Program_weight_tensor_parameter_985: class Program_weight_tensor_parameter_986: name = "parameter_986" + original_name = "linear_17.w_0" shape = [128, 128] dtype = "float32" min_val = float("-0.236201") @@ -10829,6 +11816,7 @@ class Program_weight_tensor_parameter_986: class Program_weight_tensor_parameter_987: name = "parameter_987" + original_name = "linear_16.b_0" shape = [256] dtype = "float32" min_val = float("-2.1011") @@ -10840,6 +11828,7 @@ class Program_weight_tensor_parameter_987: class Program_weight_tensor_parameter_988: name = "parameter_988" + original_name = "linear_16.w_0" shape = [128, 256] dtype = "float32" min_val = float("-0.497821") @@ -10851,6 +11840,7 @@ class Program_weight_tensor_parameter_988: class Program_weight_tensor_parameter_989: name = "parameter_989" + original_name = "layer_norm_15.b_0" shape = [128] dtype = "float32" min_val = float("-0.373667") @@ -10862,6 +11852,7 @@ class Program_weight_tensor_parameter_989: class Program_weight_tensor_parameter_990: name = "parameter_990" + original_name = "layer_norm_15.w_0" shape = [128] dtype = "float32" min_val = float("-0.021346") @@ -10873,6 +11864,7 @@ class Program_weight_tensor_parameter_990: class Program_weight_tensor_parameter_991: name = "parameter_991" + original_name = "conv2d_10.b_0" shape = [128] dtype = "float32" min_val = float("-1.97567") @@ -10884,6 +11876,7 @@ class Program_weight_tensor_parameter_991: class Program_weight_tensor_parameter_992: name = "parameter_992" + original_name = "conv2d_10.w_0" shape = [128, 128, 4, 4] dtype = "float32" min_val = float("-0.28185") @@ -10895,6 +11888,7 @@ class Program_weight_tensor_parameter_992: class Program_weight_tensor_parameter_993: name = "parameter_993" + original_name = "linear_15.b_0" shape = [128] dtype = "float32" min_val = float("-2.66027") @@ -10906,6 +11900,7 @@ class Program_weight_tensor_parameter_993: class Program_weight_tensor_parameter_994: name = "parameter_994" + original_name = "linear_15.w_0" shape = [128, 128] dtype = "float32" min_val = float("-0.327791") @@ -10917,6 +11912,7 @@ class Program_weight_tensor_parameter_994: class Program_weight_tensor_parameter_995: name = "parameter_995" + original_name = "layer_norm_14.b_0" shape = [128] dtype = "float32" min_val = float("-0.503285") @@ -10928,6 +11924,7 @@ class Program_weight_tensor_parameter_995: class Program_weight_tensor_parameter_996: name = "parameter_996" + original_name = "layer_norm_14.w_0" shape = [128] dtype = "float32" min_val = float("0.00332943") @@ -10939,6 +11936,7 @@ class Program_weight_tensor_parameter_996: class Program_weight_tensor_parameter_997: name = "parameter_997" + original_name = "layer_norm_1.b_0" shape = [128] dtype = "float32" min_val = float("-0.407577") @@ -10950,6 +11948,7 @@ class Program_weight_tensor_parameter_997: class Program_weight_tensor_parameter_998: name = "parameter_998" + original_name = "layer_norm_1.w_0" shape = [128] dtype = "float32" min_val = float("0.513143") @@ -10961,6 +11960,7 @@ class Program_weight_tensor_parameter_998: class Program_weight_tensor_parameter_999: name = "parameter_999" + original_name = "conv2d_1.b_0" shape = [128] dtype = "float32" min_val = float("-1.52709") @@ -10972,6 +11972,7 @@ class Program_weight_tensor_parameter_999: class Program_weight_tensor_parameter_1000: name = "parameter_1000" + original_name = "conv2d_1.w_0" shape = [128, 64, 3, 3] dtype = "float32" min_val = float("-0.340706") @@ -10983,6 +11984,7 @@ class Program_weight_tensor_parameter_1000: class Program_weight_tensor_parameter_1001: name = "parameter_1001" + original_name = "layer_norm_13.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -10992,6 +11994,7 @@ class Program_weight_tensor_parameter_1001: class Program_weight_tensor_parameter_1002: name = "parameter_1002" + original_name = "layer_norm_13.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -11001,6 +12004,7 @@ class Program_weight_tensor_parameter_1002: class Program_weight_tensor_parameter_1003: name = "parameter_1003" + original_name = "linear_14.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -11010,6 +12014,7 @@ class Program_weight_tensor_parameter_1003: class Program_weight_tensor_parameter_1004: name = "parameter_1004" + original_name = "linear_14.w_0" shape = [256, 64] dtype = "float32" min_val = float("-0.630357") @@ -11021,6 +12026,7 @@ class Program_weight_tensor_parameter_1004: class Program_weight_tensor_parameter_1005: name = "parameter_1005" + original_name = "conv2d_9.b_0" shape = [256] dtype = "float32" min_val = float("-0.988269") @@ -11032,6 +12038,7 @@ class Program_weight_tensor_parameter_1005: class Program_weight_tensor_parameter_1006: name = "parameter_1006" + original_name = "conv2d_9.w_0" shape = [256, 1, 3, 3] dtype = "float32" min_val = float("-0.301169") @@ -11043,6 +12050,7 @@ class Program_weight_tensor_parameter_1006: class Program_weight_tensor_parameter_1007: name = "parameter_1007" + original_name = "linear_13.b_0" shape = [256] dtype = "float32" min_val = float("-4.32536") @@ -11054,6 +12062,7 @@ class Program_weight_tensor_parameter_1007: class Program_weight_tensor_parameter_1008: name = "parameter_1008" + original_name = "linear_13.w_0" shape = [64, 256] dtype = "float32" min_val = float("-0.304823") @@ -11065,6 +12074,7 @@ class Program_weight_tensor_parameter_1008: class Program_weight_tensor_parameter_1009: name = "parameter_1009" + original_name = "layer_norm_12.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -11074,6 +12084,7 @@ class Program_weight_tensor_parameter_1009: class Program_weight_tensor_parameter_1010: name = "parameter_1010" + original_name = "layer_norm_12.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -11083,6 +12094,7 @@ class Program_weight_tensor_parameter_1010: class Program_weight_tensor_parameter_1011: name = "parameter_1011" + original_name = "linear_12.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -11092,6 +12104,7 @@ class Program_weight_tensor_parameter_1011: class Program_weight_tensor_parameter_1012: name = "parameter_1012" + original_name = "linear_12.w_0" shape = [64, 64] dtype = "float32" min_val = float("-0.277164") @@ -11103,6 +12116,7 @@ class Program_weight_tensor_parameter_1012: class Program_weight_tensor_parameter_1013: name = "parameter_1013" + original_name = "linear_11.b_0" shape = [128] dtype = "float32" min_val = float("-2.56786") @@ -11114,6 +12128,7 @@ class Program_weight_tensor_parameter_1013: class Program_weight_tensor_parameter_1014: name = "parameter_1014" + original_name = "linear_11.w_0" shape = [64, 128] dtype = "float32" min_val = float("-0.420686") @@ -11125,6 +12140,7 @@ class Program_weight_tensor_parameter_1014: class Program_weight_tensor_parameter_1015: name = "parameter_1015" + original_name = "layer_norm_11.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -11134,6 +12150,7 @@ class Program_weight_tensor_parameter_1015: class Program_weight_tensor_parameter_1016: name = "parameter_1016" + original_name = "layer_norm_11.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -11143,6 +12160,7 @@ class Program_weight_tensor_parameter_1016: class Program_weight_tensor_parameter_1017: name = "parameter_1017" + original_name = "conv2d_8.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -11152,6 +12170,7 @@ class Program_weight_tensor_parameter_1017: class Program_weight_tensor_parameter_1018: name = "parameter_1018" + original_name = "conv2d_8.w_0" shape = [64, 64, 8, 8] dtype = "float32" min_val = float("-0.382685") @@ -11163,6 +12182,7 @@ class Program_weight_tensor_parameter_1018: class Program_weight_tensor_parameter_1019: name = "parameter_1019" + original_name = "linear_10.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -11172,6 +12192,7 @@ class Program_weight_tensor_parameter_1019: class Program_weight_tensor_parameter_1020: name = "parameter_1020" + original_name = "linear_10.w_0" shape = [64, 64] dtype = "float32" min_val = float("-0.373109") @@ -11183,6 +12204,7 @@ class Program_weight_tensor_parameter_1020: class Program_weight_tensor_parameter_1021: name = "parameter_1021" + original_name = "layer_norm_10.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -11192,6 +12214,7 @@ class Program_weight_tensor_parameter_1021: class Program_weight_tensor_parameter_1022: name = "parameter_1022" + original_name = "layer_norm_10.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -11201,6 +12224,7 @@ class Program_weight_tensor_parameter_1022: class Program_weight_tensor_parameter_1023: name = "parameter_1023" + original_name = "linear_9.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -11210,6 +12234,7 @@ class Program_weight_tensor_parameter_1023: class Program_weight_tensor_parameter_1024: name = "parameter_1024" + original_name = "linear_9.w_0" shape = [256, 64] dtype = "float32" min_val = float("-0.393177") @@ -11221,6 +12246,7 @@ class Program_weight_tensor_parameter_1024: class Program_weight_tensor_parameter_1025: name = "parameter_1025" + original_name = "conv2d_7.b_0" shape = [256] dtype = "float32" min_val = float("-1.48205") @@ -11232,6 +12258,7 @@ class Program_weight_tensor_parameter_1025: class Program_weight_tensor_parameter_1026: name = "parameter_1026" + original_name = "conv2d_7.w_0" shape = [256, 1, 3, 3] dtype = "float32" min_val = float("-0.344454") @@ -11243,6 +12270,7 @@ class Program_weight_tensor_parameter_1026: class Program_weight_tensor_parameter_1027: name = "parameter_1027" + original_name = "linear_8.b_0" shape = [256] dtype = "float32" min_val = float("-3.06851") @@ -11254,6 +12282,7 @@ class Program_weight_tensor_parameter_1027: class Program_weight_tensor_parameter_1028: name = "parameter_1028" + original_name = "linear_8.w_0" shape = [64, 256] dtype = "float32" min_val = float("-0.390596") @@ -11265,6 +12294,7 @@ class Program_weight_tensor_parameter_1028: class Program_weight_tensor_parameter_1029: name = "parameter_1029" + original_name = "layer_norm_9.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -11274,6 +12304,7 @@ class Program_weight_tensor_parameter_1029: class Program_weight_tensor_parameter_1030: name = "parameter_1030" + original_name = "layer_norm_9.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -11283,6 +12314,7 @@ class Program_weight_tensor_parameter_1030: class Program_weight_tensor_parameter_1031: name = "parameter_1031" + original_name = "linear_7.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -11292,6 +12324,7 @@ class Program_weight_tensor_parameter_1031: class Program_weight_tensor_parameter_1032: name = "parameter_1032" + original_name = "linear_7.w_0" shape = [64, 64] dtype = "float32" min_val = float("-0.132042") @@ -11303,6 +12336,7 @@ class Program_weight_tensor_parameter_1032: class Program_weight_tensor_parameter_1033: name = "parameter_1033" + original_name = "linear_6.b_0" shape = [128] dtype = "float32" min_val = float("-0.733266") @@ -11314,6 +12348,7 @@ class Program_weight_tensor_parameter_1033: class Program_weight_tensor_parameter_1034: name = "parameter_1034" + original_name = "linear_6.w_0" shape = [64, 128] dtype = "float32" min_val = float("-0.4776") @@ -11325,6 +12360,7 @@ class Program_weight_tensor_parameter_1034: class Program_weight_tensor_parameter_1035: name = "parameter_1035" + original_name = "layer_norm_8.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -11334,6 +12370,7 @@ class Program_weight_tensor_parameter_1035: class Program_weight_tensor_parameter_1036: name = "parameter_1036" + original_name = "layer_norm_8.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -11343,6 +12380,7 @@ class Program_weight_tensor_parameter_1036: class Program_weight_tensor_parameter_1037: name = "parameter_1037" + original_name = "conv2d_6.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -11352,6 +12390,7 @@ class Program_weight_tensor_parameter_1037: class Program_weight_tensor_parameter_1038: name = "parameter_1038" + original_name = "conv2d_6.w_0" shape = [64, 64, 8, 8] dtype = "float32" min_val = float("-0.313897") @@ -11363,6 +12402,7 @@ class Program_weight_tensor_parameter_1038: class Program_weight_tensor_parameter_1039: name = "parameter_1039" + original_name = "linear_5.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -11372,6 +12412,7 @@ class Program_weight_tensor_parameter_1039: class Program_weight_tensor_parameter_1040: name = "parameter_1040" + original_name = "linear_5.w_0" shape = [64, 64] dtype = "float32" min_val = float("-0.444973") @@ -11383,6 +12424,7 @@ class Program_weight_tensor_parameter_1040: class Program_weight_tensor_parameter_1041: name = "parameter_1041" + original_name = "layer_norm_7.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -11392,6 +12434,7 @@ class Program_weight_tensor_parameter_1041: class Program_weight_tensor_parameter_1042: name = "parameter_1042" + original_name = "layer_norm_7.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -11401,6 +12444,7 @@ class Program_weight_tensor_parameter_1042: class Program_weight_tensor_parameter_1043: name = "parameter_1043" + original_name = "linear_4.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -11410,6 +12454,7 @@ class Program_weight_tensor_parameter_1043: class Program_weight_tensor_parameter_1044: name = "parameter_1044" + original_name = "linear_4.w_0" shape = [256, 64] dtype = "float32" min_val = float("-0.327109") @@ -11421,6 +12466,7 @@ class Program_weight_tensor_parameter_1044: class Program_weight_tensor_parameter_1045: name = "parameter_1045" + original_name = "conv2d_5.b_0" shape = [256] dtype = "float32" min_val = float("-1.84261") @@ -11432,6 +12478,7 @@ class Program_weight_tensor_parameter_1045: class Program_weight_tensor_parameter_1046: name = "parameter_1046" + original_name = "conv2d_5.w_0" shape = [256, 1, 3, 3] dtype = "float32" min_val = float("-0.514648") @@ -11443,6 +12490,7 @@ class Program_weight_tensor_parameter_1046: class Program_weight_tensor_parameter_1047: name = "parameter_1047" + original_name = "linear_3.b_0" shape = [256] dtype = "float32" min_val = float("-5.83809") @@ -11454,6 +12502,7 @@ class Program_weight_tensor_parameter_1047: class Program_weight_tensor_parameter_1048: name = "parameter_1048" + original_name = "linear_3.w_0" shape = [64, 256] dtype = "float32" min_val = float("-0.44435") @@ -11465,6 +12514,7 @@ class Program_weight_tensor_parameter_1048: class Program_weight_tensor_parameter_1049: name = "parameter_1049" + original_name = "layer_norm_6.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -11474,6 +12524,7 @@ class Program_weight_tensor_parameter_1049: class Program_weight_tensor_parameter_1050: name = "parameter_1050" + original_name = "layer_norm_6.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -11483,6 +12534,7 @@ class Program_weight_tensor_parameter_1050: class Program_weight_tensor_parameter_1051: name = "parameter_1051" + original_name = "linear_2.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -11492,6 +12544,7 @@ class Program_weight_tensor_parameter_1051: class Program_weight_tensor_parameter_1052: name = "parameter_1052" + original_name = "linear_2.w_0" shape = [64, 64] dtype = "float32" min_val = float("-0.194526") @@ -11503,6 +12556,7 @@ class Program_weight_tensor_parameter_1052: class Program_weight_tensor_parameter_1053: name = "parameter_1053" + original_name = "linear_1.b_0" shape = [128] dtype = "float32" min_val = float("-1.09372") @@ -11514,6 +12568,7 @@ class Program_weight_tensor_parameter_1053: class Program_weight_tensor_parameter_1054: name = "parameter_1054" + original_name = "linear_1.w_0" shape = [64, 128] dtype = "float32" min_val = float("-0.419287") @@ -11525,6 +12580,7 @@ class Program_weight_tensor_parameter_1054: class Program_weight_tensor_parameter_1055: name = "parameter_1055" + original_name = "layer_norm_5.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -11534,6 +12590,7 @@ class Program_weight_tensor_parameter_1055: class Program_weight_tensor_parameter_1056: name = "parameter_1056" + original_name = "layer_norm_5.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -11543,6 +12600,7 @@ class Program_weight_tensor_parameter_1056: class Program_weight_tensor_parameter_1057: name = "parameter_1057" + original_name = "conv2d_4.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -11552,6 +12610,7 @@ class Program_weight_tensor_parameter_1057: class Program_weight_tensor_parameter_1058: name = "parameter_1058" + original_name = "conv2d_4.w_0" shape = [64, 64, 8, 8] dtype = "float32" min_val = float("-0.620883") @@ -11563,6 +12622,7 @@ class Program_weight_tensor_parameter_1058: class Program_weight_tensor_parameter_1059: name = "parameter_1059" + original_name = "linear_0.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -11572,6 +12632,7 @@ class Program_weight_tensor_parameter_1059: class Program_weight_tensor_parameter_1060: name = "parameter_1060" + original_name = "linear_0.w_0" shape = [64, 64] dtype = "float32" min_val = float("-0.510337") @@ -11583,6 +12644,7 @@ class Program_weight_tensor_parameter_1060: class Program_weight_tensor_parameter_1061: name = "parameter_1061" + original_name = "layer_norm_4.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -11592,6 +12654,7 @@ class Program_weight_tensor_parameter_1061: class Program_weight_tensor_parameter_1062: name = "parameter_1062" + original_name = "layer_norm_4.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -11601,6 +12664,7 @@ class Program_weight_tensor_parameter_1062: class Program_weight_tensor_parameter_1063: name = "parameter_1063" + original_name = "layer_norm_0.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -11610,6 +12674,7 @@ class Program_weight_tensor_parameter_1063: class Program_weight_tensor_parameter_1064: name = "parameter_1064" + original_name = "layer_norm_0.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -11619,6 +12684,7 @@ class Program_weight_tensor_parameter_1064: class Program_weight_tensor_parameter_1065: name = "parameter_1065" + original_name = "conv2d_0.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -11628,6 +12694,7 @@ class Program_weight_tensor_parameter_1065: class Program_weight_tensor_parameter_1066: name = "parameter_1066" + original_name = "conv2d_0.w_0" shape = [64, 3, 7, 7] dtype = "float32" min_val = float("-0.380031") diff --git a/paddle_samples/PaddleX/SwinTransformer_base_patch4_window12_384/subgraph_0/input_meta.py b/paddle_samples/PaddleX/SwinTransformer_base_patch4_window12_384/subgraph_0/input_meta.py index 901a8453c..f56537877 100644 --- a/paddle_samples/PaddleX/SwinTransformer_base_patch4_window12_384/subgraph_0/input_meta.py +++ b/paddle_samples/PaddleX/SwinTransformer_base_patch4_window12_384/subgraph_0/input_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_data_0: name = "data_0" + original_name = "param_0" shape = [529, 4] dtype = "float32" min_val = float("-8.30816") @@ -11,6 +12,7 @@ class Program_weight_tensor_data_0: class Program_weight_tensor_data_1: name = "data_1" + original_name = "param_1" shape = [529, 4] dtype = "float32" min_val = float("-6.25859") @@ -22,6 +24,7 @@ class Program_weight_tensor_data_1: class Program_weight_tensor_data_2: name = "data_2" + original_name = "param_10" shape = [529, 16] dtype = "float32" min_val = float("-9.32549") @@ -33,6 +36,7 @@ class Program_weight_tensor_data_2: class Program_weight_tensor_data_3: name = "data_3" + original_name = "param_11" shape = [529, 16] dtype = "float32" min_val = float("-8.43712") @@ -44,6 +48,7 @@ class Program_weight_tensor_data_3: class Program_weight_tensor_data_4: name = "data_4" + original_name = "param_12" shape = [529, 16] dtype = "float32" min_val = float("-8.95809") @@ -55,6 +60,7 @@ class Program_weight_tensor_data_4: class Program_weight_tensor_data_5: name = "data_5" + original_name = "param_13" shape = [529, 16] dtype = "float32" min_val = float("-4.71207") @@ -66,6 +72,7 @@ class Program_weight_tensor_data_5: class Program_weight_tensor_data_6: name = "data_6" + original_name = "param_14" shape = [529, 16] dtype = "float32" min_val = float("-8.97278") @@ -77,6 +84,7 @@ class Program_weight_tensor_data_6: class Program_weight_tensor_data_7: name = "data_7" + original_name = "param_15" shape = [529, 16] dtype = "float32" min_val = float("-5.29824") @@ -88,6 +96,7 @@ class Program_weight_tensor_data_7: class Program_weight_tensor_data_8: name = "data_8" + original_name = "param_16" shape = [529, 16] dtype = "float32" min_val = float("-11.8762") @@ -99,6 +108,7 @@ class Program_weight_tensor_data_8: class Program_weight_tensor_data_9: name = "data_9" + original_name = "param_17" shape = [529, 16] dtype = "float32" min_val = float("-8.84712") @@ -110,6 +120,7 @@ class Program_weight_tensor_data_9: class Program_weight_tensor_data_10: name = "data_10" + original_name = "param_18" shape = [529, 16] dtype = "float32" min_val = float("-10.4539") @@ -121,6 +132,7 @@ class Program_weight_tensor_data_10: class Program_weight_tensor_data_11: name = "data_11" + original_name = "param_19" shape = [529, 16] dtype = "float32" min_val = float("-10.8027") @@ -132,6 +144,7 @@ class Program_weight_tensor_data_11: class Program_weight_tensor_data_12: name = "data_12" + original_name = "param_2" shape = [529, 8] dtype = "float32" min_val = float("-5.98832") @@ -143,6 +156,7 @@ class Program_weight_tensor_data_12: class Program_weight_tensor_data_13: name = "data_13" + original_name = "param_20" shape = [529, 16] dtype = "float32" min_val = float("-14.7827") @@ -154,6 +168,7 @@ class Program_weight_tensor_data_13: class Program_weight_tensor_data_14: name = "data_14" + original_name = "param_21" shape = [529, 16] dtype = "float32" min_val = float("-14.5049") @@ -165,6 +180,7 @@ class Program_weight_tensor_data_14: class Program_weight_tensor_data_15: name = "data_15" + original_name = "param_22" shape = [529, 32] dtype = "float32" min_val = float("-15.7035") @@ -176,6 +192,7 @@ class Program_weight_tensor_data_15: class Program_weight_tensor_data_16: name = "data_16" + original_name = "param_23" shape = [529, 32] dtype = "float32" min_val = float("-15.7504") @@ -187,6 +204,7 @@ class Program_weight_tensor_data_16: class Program_weight_tensor_data_17: name = "data_17" + original_name = "param_3" shape = [529, 8] dtype = "float32" min_val = float("-8.14151") @@ -198,6 +216,7 @@ class Program_weight_tensor_data_17: class Program_weight_tensor_data_18: name = "data_18" + original_name = "param_4" shape = [529, 16] dtype = "float32" min_val = float("-11.037") @@ -209,6 +228,7 @@ class Program_weight_tensor_data_18: class Program_weight_tensor_data_19: name = "data_19" + original_name = "param_5" shape = [529, 16] dtype = "float32" min_val = float("-10.1083") @@ -220,6 +240,7 @@ class Program_weight_tensor_data_19: class Program_weight_tensor_data_20: name = "data_20" + original_name = "param_6" shape = [529, 16] dtype = "float32" min_val = float("-10.8027") @@ -231,6 +252,7 @@ class Program_weight_tensor_data_20: class Program_weight_tensor_data_21: name = "data_21" + original_name = "param_7" shape = [529, 16] dtype = "float32" min_val = float("-7.44028") @@ -242,6 +264,7 @@ class Program_weight_tensor_data_21: class Program_weight_tensor_data_22: name = "data_22" + original_name = "param_8" shape = [529, 16] dtype = "float32" min_val = float("-13.3957") @@ -253,6 +276,7 @@ class Program_weight_tensor_data_22: class Program_weight_tensor_data_23: name = "data_23" + original_name = "param_9" shape = [529, 16] dtype = "float32" min_val = float("-9.89985") @@ -264,6 +288,7 @@ class Program_weight_tensor_data_23: class Program_weight_tensor_data_24: name = "data_24" + original_name = "var_0" shape = [4, 3, 384, 384] dtype = "float32" min_val = float("-4.29068") @@ -275,6 +300,7 @@ class Program_weight_tensor_data_24: class Program_weight_tensor_data_25: name = "data_25" + original_name = "var_1021" shape = [144, 144] dtype = "int64" min_val = 0 @@ -284,6 +310,7 @@ class Program_weight_tensor_data_25: class Program_weight_tensor_data_26: name = "data_26" + original_name = "var_1101" shape = [144, 144] dtype = "int64" min_val = 0 @@ -293,6 +320,7 @@ class Program_weight_tensor_data_26: class Program_weight_tensor_data_27: name = "data_27" + original_name = "var_1163" shape = [144, 144] dtype = "int64" min_val = 0 @@ -302,6 +330,7 @@ class Program_weight_tensor_data_27: class Program_weight_tensor_data_28: name = "data_28" + original_name = "var_1243" shape = [144, 144] dtype = "int64" min_val = 0 @@ -311,6 +340,7 @@ class Program_weight_tensor_data_28: class Program_weight_tensor_data_29: name = "data_29" + original_name = "var_1305" shape = [144, 144] dtype = "int64" min_val = 0 @@ -320,6 +350,7 @@ class Program_weight_tensor_data_29: class Program_weight_tensor_data_30: name = "data_30" + original_name = "var_1385" shape = [144, 144] dtype = "int64" min_val = 0 @@ -329,6 +360,7 @@ class Program_weight_tensor_data_30: class Program_weight_tensor_data_31: name = "data_31" + original_name = "var_1447" shape = [144, 144] dtype = "int64" min_val = 0 @@ -338,6 +370,7 @@ class Program_weight_tensor_data_31: class Program_weight_tensor_data_32: name = "data_32" + original_name = "var_1527" shape = [144, 144] dtype = "int64" min_val = 0 @@ -347,6 +380,7 @@ class Program_weight_tensor_data_32: class Program_weight_tensor_data_33: name = "data_33" + original_name = "var_159" shape = [144, 144] dtype = "int64" min_val = 0 @@ -356,6 +390,7 @@ class Program_weight_tensor_data_33: class Program_weight_tensor_data_34: name = "data_34" + original_name = "var_1599" shape = [144, 144] dtype = "int64" min_val = 0 @@ -365,6 +400,7 @@ class Program_weight_tensor_data_34: class Program_weight_tensor_data_35: name = "data_35" + original_name = "var_1679" shape = [144, 144] dtype = "int64" min_val = 0 @@ -374,6 +410,7 @@ class Program_weight_tensor_data_35: class Program_weight_tensor_data_36: name = "data_36" + original_name = "var_21" shape = [144, 144] dtype = "int64" min_val = 0 @@ -383,6 +420,7 @@ class Program_weight_tensor_data_36: class Program_weight_tensor_data_37: name = "data_37" + original_name = "var_239" shape = [144, 144] dtype = "int64" min_val = 0 @@ -392,6 +430,7 @@ class Program_weight_tensor_data_37: class Program_weight_tensor_data_38: name = "data_38" + original_name = "var_311" shape = [144, 144] dtype = "int64" min_val = 0 @@ -401,6 +440,7 @@ class Program_weight_tensor_data_38: class Program_weight_tensor_data_39: name = "data_39" + original_name = "var_391" shape = [144, 144] dtype = "int64" min_val = 0 @@ -410,6 +450,7 @@ class Program_weight_tensor_data_39: class Program_weight_tensor_data_40: name = "data_40" + original_name = "var_453" shape = [144, 144] dtype = "int64" min_val = 0 @@ -419,6 +460,7 @@ class Program_weight_tensor_data_40: class Program_weight_tensor_data_41: name = "data_41" + original_name = "var_533" shape = [144, 144] dtype = "int64" min_val = 0 @@ -428,6 +470,7 @@ class Program_weight_tensor_data_41: class Program_weight_tensor_data_42: name = "data_42" + original_name = "var_595" shape = [144, 144] dtype = "int64" min_val = 0 @@ -437,6 +480,7 @@ class Program_weight_tensor_data_42: class Program_weight_tensor_data_43: name = "data_43" + original_name = "var_675" shape = [144, 144] dtype = "int64" min_val = 0 @@ -446,6 +490,7 @@ class Program_weight_tensor_data_43: class Program_weight_tensor_data_44: name = "data_44" + original_name = "var_737" shape = [144, 144] dtype = "int64" min_val = 0 @@ -455,6 +500,7 @@ class Program_weight_tensor_data_44: class Program_weight_tensor_data_45: name = "data_45" + original_name = "var_817" shape = [144, 144] dtype = "int64" min_val = 0 @@ -464,6 +510,7 @@ class Program_weight_tensor_data_45: class Program_weight_tensor_data_46: name = "data_46" + original_name = "var_87" shape = [144, 144] dtype = "int64" min_val = 0 @@ -473,6 +520,7 @@ class Program_weight_tensor_data_46: class Program_weight_tensor_data_47: name = "data_47" + original_name = "var_879" shape = [144, 144] dtype = "int64" min_val = 0 @@ -482,6 +530,7 @@ class Program_weight_tensor_data_47: class Program_weight_tensor_data_48: name = "data_48" + original_name = "var_959" shape = [144, 144] dtype = "int64" min_val = 0 diff --git a/paddle_samples/PaddleX/SwinTransformer_base_patch4_window12_384/subgraph_0/weight_meta.py b/paddle_samples/PaddleX/SwinTransformer_base_patch4_window12_384/subgraph_0/weight_meta.py index f8ddb5aed..c5e647813 100644 --- a/paddle_samples/PaddleX/SwinTransformer_base_patch4_window12_384/subgraph_0/weight_meta.py +++ b/paddle_samples/PaddleX/SwinTransformer_base_patch4_window12_384/subgraph_0/weight_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_parameter_0: name = "parameter_0" + original_name = "linear_99.b_0" shape = [102] dtype = "float32" data = None @@ -7,6 +8,7 @@ class Program_weight_tensor_parameter_0: class Program_weight_tensor_parameter_1: name = "parameter_1" + original_name = "linear_99.w_0" shape = [1024, 102] dtype = "float32" min_val = float("-0.0825595") @@ -18,6 +20,7 @@ class Program_weight_tensor_parameter_1: class Program_weight_tensor_parameter_2: name = "parameter_2" + original_name = "layer_norm_52.b_0" shape = [1024] dtype = "float32" min_val = float("-0.301759") @@ -29,6 +32,7 @@ class Program_weight_tensor_parameter_2: class Program_weight_tensor_parameter_3: name = "parameter_3" + original_name = "layer_norm_52.w_0" shape = [1024] dtype = "float32" min_val = float("-0.000381063") @@ -40,6 +44,7 @@ class Program_weight_tensor_parameter_3: class Program_weight_tensor_parameter_4: name = "parameter_4" + original_name = "linear_98.b_0" shape = [1024] dtype = "float32" min_val = float("-5.6785") @@ -51,6 +56,7 @@ class Program_weight_tensor_parameter_4: class Program_weight_tensor_parameter_5: name = "parameter_5" + original_name = "linear_98.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-1.78028") @@ -62,6 +68,7 @@ class Program_weight_tensor_parameter_5: class Program_weight_tensor_parameter_6: name = "parameter_6" + original_name = "linear_97.b_0" shape = [4096] dtype = "float32" min_val = float("-2.68819") @@ -73,6 +80,7 @@ class Program_weight_tensor_parameter_6: class Program_weight_tensor_parameter_7: name = "parameter_7" + original_name = "linear_97.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.725077") @@ -84,6 +92,7 @@ class Program_weight_tensor_parameter_7: class Program_weight_tensor_parameter_8: name = "parameter_8" + original_name = "layer_norm_51.b_0" shape = [1024] dtype = "float32" min_val = float("-1.01782") @@ -95,6 +104,7 @@ class Program_weight_tensor_parameter_8: class Program_weight_tensor_parameter_9: name = "parameter_9" + original_name = "layer_norm_51.w_0" shape = [1024] dtype = "float32" min_val = float("-0.000119732") @@ -106,6 +116,7 @@ class Program_weight_tensor_parameter_9: class Program_weight_tensor_parameter_10: name = "parameter_10" + original_name = "linear_96.b_0" shape = [1024] dtype = "float32" min_val = float("-5.61236") @@ -117,6 +128,7 @@ class Program_weight_tensor_parameter_10: class Program_weight_tensor_parameter_11: name = "parameter_11" + original_name = "linear_96.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-1.24825") @@ -128,6 +140,7 @@ class Program_weight_tensor_parameter_11: class Program_weight_tensor_parameter_12: name = "parameter_12" + original_name = "linear_95.b_0" shape = [3072] dtype = "float32" min_val = float("-2.51729") @@ -139,6 +152,7 @@ class Program_weight_tensor_parameter_12: class Program_weight_tensor_parameter_13: name = "parameter_13" + original_name = "linear_95.w_0" shape = [1024, 3072] dtype = "float32" min_val = float("-0.348306") @@ -150,6 +164,7 @@ class Program_weight_tensor_parameter_13: class Program_weight_tensor_parameter_14: name = "parameter_14" + original_name = "layer_norm_50.b_0" shape = [1024] dtype = "float32" min_val = float("-2.02456") @@ -161,6 +176,7 @@ class Program_weight_tensor_parameter_14: class Program_weight_tensor_parameter_15: name = "parameter_15" + original_name = "layer_norm_50.w_0" shape = [1024] dtype = "float32" min_val = float("0.000730434") @@ -172,6 +188,7 @@ class Program_weight_tensor_parameter_15: class Program_weight_tensor_parameter_16: name = "parameter_16" + original_name = "linear_94.b_0" shape = [1024] dtype = "float32" min_val = float("-5.8421") @@ -183,6 +200,7 @@ class Program_weight_tensor_parameter_16: class Program_weight_tensor_parameter_17: name = "parameter_17" + original_name = "linear_94.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-1.53009") @@ -194,6 +212,7 @@ class Program_weight_tensor_parameter_17: class Program_weight_tensor_parameter_18: name = "parameter_18" + original_name = "linear_93.b_0" shape = [4096] dtype = "float32" min_val = float("-1.18142") @@ -205,6 +224,7 @@ class Program_weight_tensor_parameter_18: class Program_weight_tensor_parameter_19: name = "parameter_19" + original_name = "linear_93.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.641736") @@ -216,6 +236,7 @@ class Program_weight_tensor_parameter_19: class Program_weight_tensor_parameter_20: name = "parameter_20" + original_name = "layer_norm_49.b_0" shape = [1024] dtype = "float32" min_val = float("-0.973838") @@ -227,6 +248,7 @@ class Program_weight_tensor_parameter_20: class Program_weight_tensor_parameter_21: name = "parameter_21" + original_name = "layer_norm_49.w_0" shape = [1024] dtype = "float32" min_val = float("-0.0202965") @@ -238,6 +260,7 @@ class Program_weight_tensor_parameter_21: class Program_weight_tensor_parameter_22: name = "parameter_22" + original_name = "linear_92.b_0" shape = [1024] dtype = "float32" min_val = float("-5.34395") @@ -249,6 +272,7 @@ class Program_weight_tensor_parameter_22: class Program_weight_tensor_parameter_23: name = "parameter_23" + original_name = "linear_92.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-1.63388") @@ -260,6 +284,7 @@ class Program_weight_tensor_parameter_23: class Program_weight_tensor_parameter_24: name = "parameter_24" + original_name = "linear_91.b_0" shape = [3072] dtype = "float32" min_val = float("-2.67258") @@ -271,6 +296,7 @@ class Program_weight_tensor_parameter_24: class Program_weight_tensor_parameter_25: name = "parameter_25" + original_name = "linear_91.w_0" shape = [1024, 3072] dtype = "float32" min_val = float("-0.315973") @@ -282,6 +308,7 @@ class Program_weight_tensor_parameter_25: class Program_weight_tensor_parameter_26: name = "parameter_26" + original_name = "layer_norm_48.b_0" shape = [1024] dtype = "float32" min_val = float("-1.88685") @@ -293,6 +320,7 @@ class Program_weight_tensor_parameter_26: class Program_weight_tensor_parameter_27: name = "parameter_27" + original_name = "layer_norm_48.w_0" shape = [1024] dtype = "float32" min_val = float("0.000778533") @@ -304,6 +332,7 @@ class Program_weight_tensor_parameter_27: class Program_weight_tensor_parameter_28: name = "parameter_28" + original_name = "linear_90.w_0" shape = [2048, 1024] dtype = "float32" min_val = float("-1.11939") @@ -315,6 +344,7 @@ class Program_weight_tensor_parameter_28: class Program_weight_tensor_parameter_29: name = "parameter_29" + original_name = "layer_norm_47.b_0" shape = [2048] dtype = "float32" min_val = float("-1.55898") @@ -326,6 +356,7 @@ class Program_weight_tensor_parameter_29: class Program_weight_tensor_parameter_30: name = "parameter_30" + original_name = "layer_norm_47.w_0" shape = [2048] dtype = "float32" min_val = float("0.083125") @@ -337,6 +368,7 @@ class Program_weight_tensor_parameter_30: class Program_weight_tensor_parameter_31: name = "parameter_31" + original_name = "linear_89.b_0" shape = [512] dtype = "float32" min_val = float("-1.84993") @@ -348,6 +380,7 @@ class Program_weight_tensor_parameter_31: class Program_weight_tensor_parameter_32: name = "parameter_32" + original_name = "linear_89.w_0" shape = [2048, 512] dtype = "float32" min_val = float("-1.93296") @@ -359,6 +392,7 @@ class Program_weight_tensor_parameter_32: class Program_weight_tensor_parameter_33: name = "parameter_33" + original_name = "linear_88.b_0" shape = [2048] dtype = "float32" min_val = float("-2.4898") @@ -370,6 +404,7 @@ class Program_weight_tensor_parameter_33: class Program_weight_tensor_parameter_34: name = "parameter_34" + original_name = "linear_88.w_0" shape = [512, 2048] dtype = "float32" min_val = float("-0.785506") @@ -381,6 +416,7 @@ class Program_weight_tensor_parameter_34: class Program_weight_tensor_parameter_35: name = "parameter_35" + original_name = "layer_norm_46.b_0" shape = [512] dtype = "float32" min_val = float("-1.90426") @@ -392,6 +428,7 @@ class Program_weight_tensor_parameter_35: class Program_weight_tensor_parameter_36: name = "parameter_36" + original_name = "layer_norm_46.w_0" shape = [512] dtype = "float32" min_val = float("0.282539") @@ -403,6 +440,7 @@ class Program_weight_tensor_parameter_36: class Program_weight_tensor_parameter_37: name = "parameter_37" + original_name = "linear_87.b_0" shape = [512] dtype = "float32" min_val = float("-2.20901") @@ -414,6 +452,7 @@ class Program_weight_tensor_parameter_37: class Program_weight_tensor_parameter_38: name = "parameter_38" + original_name = "linear_87.w_0" shape = [512, 512] dtype = "float32" min_val = float("-1.61236") @@ -425,6 +464,7 @@ class Program_weight_tensor_parameter_38: class Program_weight_tensor_parameter_39: name = "parameter_39" + original_name = "linear_86.b_0" shape = [1536] dtype = "float32" min_val = float("-2.49562") @@ -436,6 +476,7 @@ class Program_weight_tensor_parameter_39: class Program_weight_tensor_parameter_40: name = "parameter_40" + original_name = "linear_86.w_0" shape = [512, 1536] dtype = "float32" min_val = float("-0.353463") @@ -447,6 +488,7 @@ class Program_weight_tensor_parameter_40: class Program_weight_tensor_parameter_41: name = "parameter_41" + original_name = "layer_norm_45.b_0" shape = [512] dtype = "float32" min_val = float("-1.40524") @@ -458,6 +500,7 @@ class Program_weight_tensor_parameter_41: class Program_weight_tensor_parameter_42: name = "parameter_42" + original_name = "layer_norm_45.w_0" shape = [512] dtype = "float32" min_val = float("-0.0275907") @@ -469,6 +512,7 @@ class Program_weight_tensor_parameter_42: class Program_weight_tensor_parameter_43: name = "parameter_43" + original_name = "linear_85.b_0" shape = [512] dtype = "float32" min_val = float("-1.75148") @@ -480,6 +524,7 @@ class Program_weight_tensor_parameter_43: class Program_weight_tensor_parameter_44: name = "parameter_44" + original_name = "linear_85.w_0" shape = [2048, 512] dtype = "float32" min_val = float("-1.32386") @@ -491,6 +536,7 @@ class Program_weight_tensor_parameter_44: class Program_weight_tensor_parameter_45: name = "parameter_45" + original_name = "linear_84.b_0" shape = [2048] dtype = "float32" min_val = float("-2.11087") @@ -502,6 +548,7 @@ class Program_weight_tensor_parameter_45: class Program_weight_tensor_parameter_46: name = "parameter_46" + original_name = "linear_84.w_0" shape = [512, 2048] dtype = "float32" min_val = float("-0.553849") @@ -513,6 +560,7 @@ class Program_weight_tensor_parameter_46: class Program_weight_tensor_parameter_47: name = "parameter_47" + original_name = "layer_norm_44.b_0" shape = [512] dtype = "float32" min_val = float("-2.26105") @@ -524,6 +572,7 @@ class Program_weight_tensor_parameter_47: class Program_weight_tensor_parameter_48: name = "parameter_48" + original_name = "layer_norm_44.w_0" shape = [512] dtype = "float32" min_val = float("0.33021") @@ -535,6 +584,7 @@ class Program_weight_tensor_parameter_48: class Program_weight_tensor_parameter_49: name = "parameter_49" + original_name = "linear_83.b_0" shape = [512] dtype = "float32" min_val = float("-2.15291") @@ -546,6 +596,7 @@ class Program_weight_tensor_parameter_49: class Program_weight_tensor_parameter_50: name = "parameter_50" + original_name = "linear_83.w_0" shape = [512, 512] dtype = "float32" min_val = float("-0.907019") @@ -557,6 +608,7 @@ class Program_weight_tensor_parameter_50: class Program_weight_tensor_parameter_51: name = "parameter_51" + original_name = "linear_82.b_0" shape = [1536] dtype = "float32" min_val = float("-2.49229") @@ -568,6 +620,7 @@ class Program_weight_tensor_parameter_51: class Program_weight_tensor_parameter_52: name = "parameter_52" + original_name = "linear_82.w_0" shape = [512, 1536] dtype = "float32" min_val = float("-0.378789") @@ -579,6 +632,7 @@ class Program_weight_tensor_parameter_52: class Program_weight_tensor_parameter_53: name = "parameter_53" + original_name = "layer_norm_43.b_0" shape = [512] dtype = "float32" min_val = float("-1.12132") @@ -590,6 +644,7 @@ class Program_weight_tensor_parameter_53: class Program_weight_tensor_parameter_54: name = "parameter_54" + original_name = "layer_norm_43.w_0" shape = [512] dtype = "float32" min_val = float("-0.477141") @@ -601,6 +656,7 @@ class Program_weight_tensor_parameter_54: class Program_weight_tensor_parameter_55: name = "parameter_55" + original_name = "linear_81.b_0" shape = [512] dtype = "float32" min_val = float("-1.66971") @@ -612,6 +668,7 @@ class Program_weight_tensor_parameter_55: class Program_weight_tensor_parameter_56: name = "parameter_56" + original_name = "linear_81.w_0" shape = [2048, 512] dtype = "float32" min_val = float("-1.45878") @@ -623,6 +680,7 @@ class Program_weight_tensor_parameter_56: class Program_weight_tensor_parameter_57: name = "parameter_57" + original_name = "linear_80.b_0" shape = [2048] dtype = "float32" min_val = float("-1.28136") @@ -634,6 +692,7 @@ class Program_weight_tensor_parameter_57: class Program_weight_tensor_parameter_58: name = "parameter_58" + original_name = "linear_80.w_0" shape = [512, 2048] dtype = "float32" min_val = float("-0.510794") @@ -645,6 +704,7 @@ class Program_weight_tensor_parameter_58: class Program_weight_tensor_parameter_59: name = "parameter_59" + original_name = "layer_norm_42.b_0" shape = [512] dtype = "float32" min_val = float("-2.22279") @@ -656,6 +716,7 @@ class Program_weight_tensor_parameter_59: class Program_weight_tensor_parameter_60: name = "parameter_60" + original_name = "layer_norm_42.w_0" shape = [512] dtype = "float32" min_val = float("0.286975") @@ -667,6 +728,7 @@ class Program_weight_tensor_parameter_60: class Program_weight_tensor_parameter_61: name = "parameter_61" + original_name = "linear_79.b_0" shape = [512] dtype = "float32" min_val = float("-1.40334") @@ -678,6 +740,7 @@ class Program_weight_tensor_parameter_61: class Program_weight_tensor_parameter_62: name = "parameter_62" + original_name = "linear_79.w_0" shape = [512, 512] dtype = "float32" min_val = float("-0.679723") @@ -689,6 +752,7 @@ class Program_weight_tensor_parameter_62: class Program_weight_tensor_parameter_63: name = "parameter_63" + original_name = "linear_78.b_0" shape = [1536] dtype = "float32" min_val = float("-2.72811") @@ -700,6 +764,7 @@ class Program_weight_tensor_parameter_63: class Program_weight_tensor_parameter_64: name = "parameter_64" + original_name = "linear_78.w_0" shape = [512, 1536] dtype = "float32" min_val = float("-0.365878") @@ -711,6 +776,7 @@ class Program_weight_tensor_parameter_64: class Program_weight_tensor_parameter_65: name = "parameter_65" + original_name = "layer_norm_41.b_0" shape = [512] dtype = "float32" min_val = float("-0.945871") @@ -722,6 +788,7 @@ class Program_weight_tensor_parameter_65: class Program_weight_tensor_parameter_66: name = "parameter_66" + original_name = "layer_norm_41.w_0" shape = [512] dtype = "float32" min_val = float("-0.000476146") @@ -733,6 +800,7 @@ class Program_weight_tensor_parameter_66: class Program_weight_tensor_parameter_67: name = "parameter_67" + original_name = "linear_77.b_0" shape = [512] dtype = "float32" min_val = float("-1.32087") @@ -744,6 +812,7 @@ class Program_weight_tensor_parameter_67: class Program_weight_tensor_parameter_68: name = "parameter_68" + original_name = "linear_77.w_0" shape = [2048, 512] dtype = "float32" min_val = float("-1.15338") @@ -755,6 +824,7 @@ class Program_weight_tensor_parameter_68: class Program_weight_tensor_parameter_69: name = "parameter_69" + original_name = "linear_76.b_0" shape = [2048] dtype = "float32" min_val = float("-0.905111") @@ -766,6 +836,7 @@ class Program_weight_tensor_parameter_69: class Program_weight_tensor_parameter_70: name = "parameter_70" + original_name = "linear_76.w_0" shape = [512, 2048] dtype = "float32" min_val = float("-0.442241") @@ -777,6 +848,7 @@ class Program_weight_tensor_parameter_70: class Program_weight_tensor_parameter_71: name = "parameter_71" + original_name = "layer_norm_40.b_0" shape = [512] dtype = "float32" min_val = float("-1.97073") @@ -788,6 +860,7 @@ class Program_weight_tensor_parameter_71: class Program_weight_tensor_parameter_72: name = "parameter_72" + original_name = "layer_norm_40.w_0" shape = [512] dtype = "float32" min_val = float("0.257333") @@ -799,6 +872,7 @@ class Program_weight_tensor_parameter_72: class Program_weight_tensor_parameter_73: name = "parameter_73" + original_name = "linear_75.b_0" shape = [512] dtype = "float32" min_val = float("-1.53453") @@ -810,6 +884,7 @@ class Program_weight_tensor_parameter_73: class Program_weight_tensor_parameter_74: name = "parameter_74" + original_name = "linear_75.w_0" shape = [512, 512] dtype = "float32" min_val = float("-0.678975") @@ -821,6 +896,7 @@ class Program_weight_tensor_parameter_74: class Program_weight_tensor_parameter_75: name = "parameter_75" + original_name = "linear_74.b_0" shape = [1536] dtype = "float32" min_val = float("-2.84914") @@ -832,6 +908,7 @@ class Program_weight_tensor_parameter_75: class Program_weight_tensor_parameter_76: name = "parameter_76" + original_name = "linear_74.w_0" shape = [512, 1536] dtype = "float32" min_val = float("-0.310674") @@ -843,6 +920,7 @@ class Program_weight_tensor_parameter_76: class Program_weight_tensor_parameter_77: name = "parameter_77" + original_name = "layer_norm_39.b_0" shape = [512] dtype = "float32" min_val = float("-0.951871") @@ -854,6 +932,7 @@ class Program_weight_tensor_parameter_77: class Program_weight_tensor_parameter_78: name = "parameter_78" + original_name = "layer_norm_39.w_0" shape = [512] dtype = "float32" min_val = float("-0.000225911") @@ -865,6 +944,7 @@ class Program_weight_tensor_parameter_78: class Program_weight_tensor_parameter_79: name = "parameter_79" + original_name = "linear_73.b_0" shape = [512] dtype = "float32" min_val = float("-0.992691") @@ -876,6 +956,7 @@ class Program_weight_tensor_parameter_79: class Program_weight_tensor_parameter_80: name = "parameter_80" + original_name = "linear_73.w_0" shape = [2048, 512] dtype = "float32" min_val = float("-0.476424") @@ -887,6 +968,7 @@ class Program_weight_tensor_parameter_80: class Program_weight_tensor_parameter_81: name = "parameter_81" + original_name = "linear_72.b_0" shape = [2048] dtype = "float32" min_val = float("-0.886278") @@ -898,6 +980,7 @@ class Program_weight_tensor_parameter_81: class Program_weight_tensor_parameter_82: name = "parameter_82" + original_name = "linear_72.w_0" shape = [512, 2048] dtype = "float32" min_val = float("-0.470579") @@ -909,6 +992,7 @@ class Program_weight_tensor_parameter_82: class Program_weight_tensor_parameter_83: name = "parameter_83" + original_name = "layer_norm_38.b_0" shape = [512] dtype = "float32" min_val = float("-1.60415") @@ -920,6 +1004,7 @@ class Program_weight_tensor_parameter_83: class Program_weight_tensor_parameter_84: name = "parameter_84" + original_name = "layer_norm_38.w_0" shape = [512] dtype = "float32" min_val = float("0.24575") @@ -931,6 +1016,7 @@ class Program_weight_tensor_parameter_84: class Program_weight_tensor_parameter_85: name = "parameter_85" + original_name = "linear_71.b_0" shape = [512] dtype = "float32" min_val = float("-1.10666") @@ -942,6 +1028,7 @@ class Program_weight_tensor_parameter_85: class Program_weight_tensor_parameter_86: name = "parameter_86" + original_name = "linear_71.w_0" shape = [512, 512] dtype = "float32" min_val = float("-0.590809") @@ -953,6 +1040,7 @@ class Program_weight_tensor_parameter_86: class Program_weight_tensor_parameter_87: name = "parameter_87" + original_name = "linear_70.b_0" shape = [1536] dtype = "float32" min_val = float("-2.5101") @@ -964,6 +1052,7 @@ class Program_weight_tensor_parameter_87: class Program_weight_tensor_parameter_88: name = "parameter_88" + original_name = "linear_70.w_0" shape = [512, 1536] dtype = "float32" min_val = float("-0.276231") @@ -975,6 +1064,7 @@ class Program_weight_tensor_parameter_88: class Program_weight_tensor_parameter_89: name = "parameter_89" + original_name = "layer_norm_37.b_0" shape = [512] dtype = "float32" min_val = float("-0.958495") @@ -986,6 +1076,7 @@ class Program_weight_tensor_parameter_89: class Program_weight_tensor_parameter_90: name = "parameter_90" + original_name = "layer_norm_37.w_0" shape = [512] dtype = "float32" min_val = float("-0.000570339") @@ -997,6 +1088,7 @@ class Program_weight_tensor_parameter_90: class Program_weight_tensor_parameter_91: name = "parameter_91" + original_name = "linear_69.b_0" shape = [512] dtype = "float32" min_val = float("-0.91261") @@ -1008,6 +1100,7 @@ class Program_weight_tensor_parameter_91: class Program_weight_tensor_parameter_92: name = "parameter_92" + original_name = "linear_69.w_0" shape = [2048, 512] dtype = "float32" min_val = float("-0.434777") @@ -1019,6 +1112,7 @@ class Program_weight_tensor_parameter_92: class Program_weight_tensor_parameter_93: name = "parameter_93" + original_name = "linear_68.b_0" shape = [2048] dtype = "float32" min_val = float("-0.936417") @@ -1030,6 +1124,7 @@ class Program_weight_tensor_parameter_93: class Program_weight_tensor_parameter_94: name = "parameter_94" + original_name = "linear_68.w_0" shape = [512, 2048] dtype = "float32" min_val = float("-0.572416") @@ -1041,6 +1136,7 @@ class Program_weight_tensor_parameter_94: class Program_weight_tensor_parameter_95: name = "parameter_95" + original_name = "layer_norm_36.b_0" shape = [512] dtype = "float32" min_val = float("-1.2439") @@ -1052,6 +1148,7 @@ class Program_weight_tensor_parameter_95: class Program_weight_tensor_parameter_96: name = "parameter_96" + original_name = "layer_norm_36.w_0" shape = [512] dtype = "float32" min_val = float("0.214692") @@ -1063,6 +1160,7 @@ class Program_weight_tensor_parameter_96: class Program_weight_tensor_parameter_97: name = "parameter_97" + original_name = "linear_67.b_0" shape = [512] dtype = "float32" min_val = float("-1.1577") @@ -1074,6 +1172,7 @@ class Program_weight_tensor_parameter_97: class Program_weight_tensor_parameter_98: name = "parameter_98" + original_name = "linear_67.w_0" shape = [512, 512] dtype = "float32" min_val = float("-1.02098") @@ -1085,6 +1184,7 @@ class Program_weight_tensor_parameter_98: class Program_weight_tensor_parameter_99: name = "parameter_99" + original_name = "linear_66.b_0" shape = [1536] dtype = "float32" min_val = float("-2.76691") @@ -1096,6 +1196,7 @@ class Program_weight_tensor_parameter_99: class Program_weight_tensor_parameter_100: name = "parameter_100" + original_name = "linear_66.w_0" shape = [512, 1536] dtype = "float32" min_val = float("-0.34595") @@ -1107,6 +1208,7 @@ class Program_weight_tensor_parameter_100: class Program_weight_tensor_parameter_101: name = "parameter_101" + original_name = "layer_norm_35.b_0" shape = [512] dtype = "float32" min_val = float("-1.1916") @@ -1118,6 +1220,7 @@ class Program_weight_tensor_parameter_101: class Program_weight_tensor_parameter_102: name = "parameter_102" + original_name = "layer_norm_35.w_0" shape = [512] dtype = "float32" min_val = float("0.000370217") @@ -1129,6 +1232,7 @@ class Program_weight_tensor_parameter_102: class Program_weight_tensor_parameter_103: name = "parameter_103" + original_name = "linear_65.b_0" shape = [512] dtype = "float32" min_val = float("-0.860114") @@ -1140,6 +1244,7 @@ class Program_weight_tensor_parameter_103: class Program_weight_tensor_parameter_104: name = "parameter_104" + original_name = "linear_65.w_0" shape = [2048, 512] dtype = "float32" min_val = float("-0.446003") @@ -1151,6 +1256,7 @@ class Program_weight_tensor_parameter_104: class Program_weight_tensor_parameter_105: name = "parameter_105" + original_name = "linear_64.b_0" shape = [2048] dtype = "float32" min_val = float("-0.901727") @@ -1162,6 +1268,7 @@ class Program_weight_tensor_parameter_105: class Program_weight_tensor_parameter_106: name = "parameter_106" + original_name = "linear_64.w_0" shape = [512, 2048] dtype = "float32" min_val = float("-0.421261") @@ -1173,6 +1280,7 @@ class Program_weight_tensor_parameter_106: class Program_weight_tensor_parameter_107: name = "parameter_107" + original_name = "layer_norm_34.b_0" shape = [512] dtype = "float32" min_val = float("-1.00594") @@ -1184,6 +1292,7 @@ class Program_weight_tensor_parameter_107: class Program_weight_tensor_parameter_108: name = "parameter_108" + original_name = "layer_norm_34.w_0" shape = [512] dtype = "float32" min_val = float("0.0170856") @@ -1195,6 +1304,7 @@ class Program_weight_tensor_parameter_108: class Program_weight_tensor_parameter_109: name = "parameter_109" + original_name = "linear_63.b_0" shape = [512] dtype = "float32" min_val = float("-0.561736") @@ -1206,6 +1316,7 @@ class Program_weight_tensor_parameter_109: class Program_weight_tensor_parameter_110: name = "parameter_110" + original_name = "linear_63.w_0" shape = [512, 512] dtype = "float32" min_val = float("-0.579609") @@ -1217,6 +1328,7 @@ class Program_weight_tensor_parameter_110: class Program_weight_tensor_parameter_111: name = "parameter_111" + original_name = "linear_62.b_0" shape = [1536] dtype = "float32" min_val = float("-2.78053") @@ -1228,6 +1340,7 @@ class Program_weight_tensor_parameter_111: class Program_weight_tensor_parameter_112: name = "parameter_112" + original_name = "linear_62.w_0" shape = [512, 1536] dtype = "float32" min_val = float("-0.310384") @@ -1239,6 +1352,7 @@ class Program_weight_tensor_parameter_112: class Program_weight_tensor_parameter_113: name = "parameter_113" + original_name = "layer_norm_33.b_0" shape = [512] dtype = "float32" min_val = float("-1.23634") @@ -1250,6 +1364,7 @@ class Program_weight_tensor_parameter_113: class Program_weight_tensor_parameter_114: name = "parameter_114" + original_name = "layer_norm_33.w_0" shape = [512] dtype = "float32" min_val = float("0.113772") @@ -1261,6 +1376,7 @@ class Program_weight_tensor_parameter_114: class Program_weight_tensor_parameter_115: name = "parameter_115" + original_name = "linear_61.b_0" shape = [512] dtype = "float32" min_val = float("-0.985331") @@ -1272,6 +1388,7 @@ class Program_weight_tensor_parameter_115: class Program_weight_tensor_parameter_116: name = "parameter_116" + original_name = "linear_61.w_0" shape = [2048, 512] dtype = "float32" min_val = float("-0.517439") @@ -1283,6 +1400,7 @@ class Program_weight_tensor_parameter_116: class Program_weight_tensor_parameter_117: name = "parameter_117" + original_name = "linear_60.b_0" shape = [2048] dtype = "float32" min_val = float("-1.04455") @@ -1294,6 +1412,7 @@ class Program_weight_tensor_parameter_117: class Program_weight_tensor_parameter_118: name = "parameter_118" + original_name = "linear_60.w_0" shape = [512, 2048] dtype = "float32" min_val = float("-0.379833") @@ -1305,6 +1424,7 @@ class Program_weight_tensor_parameter_118: class Program_weight_tensor_parameter_119: name = "parameter_119" + original_name = "layer_norm_32.b_0" shape = [512] dtype = "float32" min_val = float("-1.10337") @@ -1316,6 +1436,7 @@ class Program_weight_tensor_parameter_119: class Program_weight_tensor_parameter_120: name = "parameter_120" + original_name = "layer_norm_32.w_0" shape = [512] dtype = "float32" min_val = float("0.0459248") @@ -1327,6 +1448,7 @@ class Program_weight_tensor_parameter_120: class Program_weight_tensor_parameter_121: name = "parameter_121" + original_name = "linear_59.b_0" shape = [512] dtype = "float32" min_val = float("-0.958492") @@ -1338,6 +1460,7 @@ class Program_weight_tensor_parameter_121: class Program_weight_tensor_parameter_122: name = "parameter_122" + original_name = "linear_59.w_0" shape = [512, 512] dtype = "float32" min_val = float("-0.440129") @@ -1349,6 +1472,7 @@ class Program_weight_tensor_parameter_122: class Program_weight_tensor_parameter_123: name = "parameter_123" + original_name = "linear_58.b_0" shape = [1536] dtype = "float32" min_val = float("-2.9546") @@ -1360,6 +1484,7 @@ class Program_weight_tensor_parameter_123: class Program_weight_tensor_parameter_124: name = "parameter_124" + original_name = "linear_58.w_0" shape = [512, 1536] dtype = "float32" min_val = float("-0.304006") @@ -1371,6 +1496,7 @@ class Program_weight_tensor_parameter_124: class Program_weight_tensor_parameter_125: name = "parameter_125" + original_name = "layer_norm_31.b_0" shape = [512] dtype = "float32" min_val = float("-1.67884") @@ -1382,6 +1508,7 @@ class Program_weight_tensor_parameter_125: class Program_weight_tensor_parameter_126: name = "parameter_126" + original_name = "layer_norm_31.w_0" shape = [512] dtype = "float32" min_val = float("0.199841") @@ -1393,6 +1520,7 @@ class Program_weight_tensor_parameter_126: class Program_weight_tensor_parameter_127: name = "parameter_127" + original_name = "linear_57.b_0" shape = [512] dtype = "float32" min_val = float("-1.16323") @@ -1404,6 +1532,7 @@ class Program_weight_tensor_parameter_127: class Program_weight_tensor_parameter_128: name = "parameter_128" + original_name = "linear_57.w_0" shape = [2048, 512] dtype = "float32" min_val = float("-0.496643") @@ -1415,6 +1544,7 @@ class Program_weight_tensor_parameter_128: class Program_weight_tensor_parameter_129: name = "parameter_129" + original_name = "linear_56.b_0" shape = [2048] dtype = "float32" min_val = float("-0.925101") @@ -1426,6 +1556,7 @@ class Program_weight_tensor_parameter_129: class Program_weight_tensor_parameter_130: name = "parameter_130" + original_name = "linear_56.w_0" shape = [512, 2048] dtype = "float32" min_val = float("-0.981278") @@ -1437,6 +1568,7 @@ class Program_weight_tensor_parameter_130: class Program_weight_tensor_parameter_131: name = "parameter_131" + original_name = "layer_norm_30.b_0" shape = [512] dtype = "float32" min_val = float("-1.16816") @@ -1448,6 +1580,7 @@ class Program_weight_tensor_parameter_131: class Program_weight_tensor_parameter_132: name = "parameter_132" + original_name = "layer_norm_30.w_0" shape = [512] dtype = "float32" min_val = float("0.275234") @@ -1459,6 +1592,7 @@ class Program_weight_tensor_parameter_132: class Program_weight_tensor_parameter_133: name = "parameter_133" + original_name = "linear_55.b_0" shape = [512] dtype = "float32" min_val = float("-0.583252") @@ -1470,6 +1604,7 @@ class Program_weight_tensor_parameter_133: class Program_weight_tensor_parameter_134: name = "parameter_134" + original_name = "linear_55.w_0" shape = [512, 512] dtype = "float32" min_val = float("-0.432782") @@ -1481,6 +1616,7 @@ class Program_weight_tensor_parameter_134: class Program_weight_tensor_parameter_135: name = "parameter_135" + original_name = "linear_54.b_0" shape = [1536] dtype = "float32" min_val = float("-2.58128") @@ -1492,6 +1628,7 @@ class Program_weight_tensor_parameter_135: class Program_weight_tensor_parameter_136: name = "parameter_136" + original_name = "linear_54.w_0" shape = [512, 1536] dtype = "float32" min_val = float("-0.306801") @@ -1503,6 +1640,7 @@ class Program_weight_tensor_parameter_136: class Program_weight_tensor_parameter_137: name = "parameter_137" + original_name = "layer_norm_29.b_0" shape = [512] dtype = "float32" min_val = float("-1.94602") @@ -1514,6 +1652,7 @@ class Program_weight_tensor_parameter_137: class Program_weight_tensor_parameter_138: name = "parameter_138" + original_name = "layer_norm_29.w_0" shape = [512] dtype = "float32" min_val = float("0.215205") @@ -1525,6 +1664,7 @@ class Program_weight_tensor_parameter_138: class Program_weight_tensor_parameter_139: name = "parameter_139" + original_name = "linear_53.b_0" shape = [512] dtype = "float32" min_val = float("-1.48191") @@ -1536,6 +1676,7 @@ class Program_weight_tensor_parameter_139: class Program_weight_tensor_parameter_140: name = "parameter_140" + original_name = "linear_53.w_0" shape = [2048, 512] dtype = "float32" min_val = float("-0.604879") @@ -1547,6 +1688,7 @@ class Program_weight_tensor_parameter_140: class Program_weight_tensor_parameter_141: name = "parameter_141" + original_name = "linear_52.b_0" shape = [2048] dtype = "float32" min_val = float("-1.00745") @@ -1558,6 +1700,7 @@ class Program_weight_tensor_parameter_141: class Program_weight_tensor_parameter_142: name = "parameter_142" + original_name = "linear_52.w_0" shape = [512, 2048] dtype = "float32" min_val = float("-0.707159") @@ -1569,6 +1712,7 @@ class Program_weight_tensor_parameter_142: class Program_weight_tensor_parameter_143: name = "parameter_143" + original_name = "layer_norm_28.b_0" shape = [512] dtype = "float32" min_val = float("-1.19401") @@ -1580,6 +1724,7 @@ class Program_weight_tensor_parameter_143: class Program_weight_tensor_parameter_144: name = "parameter_144" + original_name = "layer_norm_28.w_0" shape = [512] dtype = "float32" min_val = float("0.36519") @@ -1591,6 +1736,7 @@ class Program_weight_tensor_parameter_144: class Program_weight_tensor_parameter_145: name = "parameter_145" + original_name = "linear_51.b_0" shape = [512] dtype = "float32" min_val = float("-0.855767") @@ -1602,6 +1748,7 @@ class Program_weight_tensor_parameter_145: class Program_weight_tensor_parameter_146: name = "parameter_146" + original_name = "linear_51.w_0" shape = [512, 512] dtype = "float32" min_val = float("-0.418717") @@ -1613,6 +1760,7 @@ class Program_weight_tensor_parameter_146: class Program_weight_tensor_parameter_147: name = "parameter_147" + original_name = "linear_50.b_0" shape = [1536] dtype = "float32" min_val = float("-2.91982") @@ -1624,6 +1772,7 @@ class Program_weight_tensor_parameter_147: class Program_weight_tensor_parameter_148: name = "parameter_148" + original_name = "linear_50.w_0" shape = [512, 1536] dtype = "float32" min_val = float("-0.358488") @@ -1635,6 +1784,7 @@ class Program_weight_tensor_parameter_148: class Program_weight_tensor_parameter_149: name = "parameter_149" + original_name = "layer_norm_27.b_0" shape = [512] dtype = "float32" min_val = float("-1.8155") @@ -1646,6 +1796,7 @@ class Program_weight_tensor_parameter_149: class Program_weight_tensor_parameter_150: name = "parameter_150" + original_name = "layer_norm_27.w_0" shape = [512] dtype = "float32" min_val = float("0.261267") @@ -1657,6 +1808,7 @@ class Program_weight_tensor_parameter_150: class Program_weight_tensor_parameter_151: name = "parameter_151" + original_name = "linear_49.b_0" shape = [512] dtype = "float32" min_val = float("-1.38603") @@ -1668,6 +1820,7 @@ class Program_weight_tensor_parameter_151: class Program_weight_tensor_parameter_152: name = "parameter_152" + original_name = "linear_49.w_0" shape = [2048, 512] dtype = "float32" min_val = float("-0.598741") @@ -1679,6 +1832,7 @@ class Program_weight_tensor_parameter_152: class Program_weight_tensor_parameter_153: name = "parameter_153" + original_name = "linear_48.b_0" shape = [2048] dtype = "float32" min_val = float("-1.14302") @@ -1690,6 +1844,7 @@ class Program_weight_tensor_parameter_153: class Program_weight_tensor_parameter_154: name = "parameter_154" + original_name = "linear_48.w_0" shape = [512, 2048] dtype = "float32" min_val = float("-0.602115") @@ -1701,6 +1856,7 @@ class Program_weight_tensor_parameter_154: class Program_weight_tensor_parameter_155: name = "parameter_155" + original_name = "layer_norm_26.b_0" shape = [512] dtype = "float32" min_val = float("-1.13957") @@ -1712,6 +1868,7 @@ class Program_weight_tensor_parameter_155: class Program_weight_tensor_parameter_156: name = "parameter_156" + original_name = "layer_norm_26.w_0" shape = [512] dtype = "float32" min_val = float("0.453767") @@ -1723,6 +1880,7 @@ class Program_weight_tensor_parameter_156: class Program_weight_tensor_parameter_157: name = "parameter_157" + original_name = "linear_47.b_0" shape = [512] dtype = "float32" min_val = float("-0.565682") @@ -1734,6 +1892,7 @@ class Program_weight_tensor_parameter_157: class Program_weight_tensor_parameter_158: name = "parameter_158" + original_name = "linear_47.w_0" shape = [512, 512] dtype = "float32" min_val = float("-0.39262") @@ -1745,6 +1904,7 @@ class Program_weight_tensor_parameter_158: class Program_weight_tensor_parameter_159: name = "parameter_159" + original_name = "linear_46.b_0" shape = [1536] dtype = "float32" min_val = float("-2.72397") @@ -1756,6 +1916,7 @@ class Program_weight_tensor_parameter_159: class Program_weight_tensor_parameter_160: name = "parameter_160" + original_name = "linear_46.w_0" shape = [512, 1536] dtype = "float32" min_val = float("-0.311578") @@ -1767,6 +1928,7 @@ class Program_weight_tensor_parameter_160: class Program_weight_tensor_parameter_161: name = "parameter_161" + original_name = "layer_norm_25.b_0" shape = [512] dtype = "float32" min_val = float("-1.76243") @@ -1778,6 +1940,7 @@ class Program_weight_tensor_parameter_161: class Program_weight_tensor_parameter_162: name = "parameter_162" + original_name = "layer_norm_25.w_0" shape = [512] dtype = "float32" min_val = float("0.243") @@ -1789,6 +1952,7 @@ class Program_weight_tensor_parameter_162: class Program_weight_tensor_parameter_163: name = "parameter_163" + original_name = "linear_45.b_0" shape = [512] dtype = "float32" min_val = float("-1.44657") @@ -1800,6 +1964,7 @@ class Program_weight_tensor_parameter_163: class Program_weight_tensor_parameter_164: name = "parameter_164" + original_name = "linear_45.w_0" shape = [2048, 512] dtype = "float32" min_val = float("-0.547257") @@ -1811,6 +1976,7 @@ class Program_weight_tensor_parameter_164: class Program_weight_tensor_parameter_165: name = "parameter_165" + original_name = "linear_44.b_0" shape = [2048] dtype = "float32" min_val = float("-1.23541") @@ -1822,6 +1988,7 @@ class Program_weight_tensor_parameter_165: class Program_weight_tensor_parameter_166: name = "parameter_166" + original_name = "linear_44.w_0" shape = [512, 2048] dtype = "float32" min_val = float("-0.539725") @@ -1833,6 +2000,7 @@ class Program_weight_tensor_parameter_166: class Program_weight_tensor_parameter_167: name = "parameter_167" + original_name = "layer_norm_24.b_0" shape = [512] dtype = "float32" min_val = float("-1.12037") @@ -1844,6 +2012,7 @@ class Program_weight_tensor_parameter_167: class Program_weight_tensor_parameter_168: name = "parameter_168" + original_name = "layer_norm_24.w_0" shape = [512] dtype = "float32" min_val = float("0.403264") @@ -1855,6 +2024,7 @@ class Program_weight_tensor_parameter_168: class Program_weight_tensor_parameter_169: name = "parameter_169" + original_name = "linear_43.b_0" shape = [512] dtype = "float32" min_val = float("-0.774936") @@ -1866,6 +2036,7 @@ class Program_weight_tensor_parameter_169: class Program_weight_tensor_parameter_170: name = "parameter_170" + original_name = "linear_43.w_0" shape = [512, 512] dtype = "float32" min_val = float("-0.499257") @@ -1877,6 +2048,7 @@ class Program_weight_tensor_parameter_170: class Program_weight_tensor_parameter_171: name = "parameter_171" + original_name = "linear_42.b_0" shape = [1536] dtype = "float32" min_val = float("-2.76909") @@ -1888,6 +2060,7 @@ class Program_weight_tensor_parameter_171: class Program_weight_tensor_parameter_172: name = "parameter_172" + original_name = "linear_42.w_0" shape = [512, 1536] dtype = "float32" min_val = float("-0.433279") @@ -1899,6 +2072,7 @@ class Program_weight_tensor_parameter_172: class Program_weight_tensor_parameter_173: name = "parameter_173" + original_name = "layer_norm_23.b_0" shape = [512] dtype = "float32" min_val = float("-1.63342") @@ -1910,6 +2084,7 @@ class Program_weight_tensor_parameter_173: class Program_weight_tensor_parameter_174: name = "parameter_174" + original_name = "layer_norm_23.w_0" shape = [512] dtype = "float32" min_val = float("0.199521") @@ -1921,6 +2096,7 @@ class Program_weight_tensor_parameter_174: class Program_weight_tensor_parameter_175: name = "parameter_175" + original_name = "linear_41.b_0" shape = [512] dtype = "float32" min_val = float("-0.60455") @@ -1932,6 +2108,7 @@ class Program_weight_tensor_parameter_175: class Program_weight_tensor_parameter_176: name = "parameter_176" + original_name = "linear_41.w_0" shape = [2048, 512] dtype = "float32" min_val = float("-0.406433") @@ -1943,6 +2120,7 @@ class Program_weight_tensor_parameter_176: class Program_weight_tensor_parameter_177: name = "parameter_177" + original_name = "linear_40.b_0" shape = [2048] dtype = "float32" min_val = float("-1.27595") @@ -1954,6 +2132,7 @@ class Program_weight_tensor_parameter_177: class Program_weight_tensor_parameter_178: name = "parameter_178" + original_name = "linear_40.w_0" shape = [512, 2048] dtype = "float32" min_val = float("-0.48605") @@ -1965,6 +2144,7 @@ class Program_weight_tensor_parameter_178: class Program_weight_tensor_parameter_179: name = "parameter_179" + original_name = "layer_norm_22.b_0" shape = [512] dtype = "float32" min_val = float("-1.04133") @@ -1976,6 +2156,7 @@ class Program_weight_tensor_parameter_179: class Program_weight_tensor_parameter_180: name = "parameter_180" + original_name = "layer_norm_22.w_0" shape = [512] dtype = "float32" min_val = float("0.3741") @@ -1987,6 +2168,7 @@ class Program_weight_tensor_parameter_180: class Program_weight_tensor_parameter_181: name = "parameter_181" + original_name = "linear_39.b_0" shape = [512] dtype = "float32" min_val = float("-0.550611") @@ -1998,6 +2180,7 @@ class Program_weight_tensor_parameter_181: class Program_weight_tensor_parameter_182: name = "parameter_182" + original_name = "linear_39.w_0" shape = [512, 512] dtype = "float32" min_val = float("-0.415624") @@ -2009,6 +2192,7 @@ class Program_weight_tensor_parameter_182: class Program_weight_tensor_parameter_183: name = "parameter_183" + original_name = "linear_38.b_0" shape = [1536] dtype = "float32" min_val = float("-2.0251") @@ -2020,6 +2204,7 @@ class Program_weight_tensor_parameter_183: class Program_weight_tensor_parameter_184: name = "parameter_184" + original_name = "linear_38.w_0" shape = [512, 1536] dtype = "float32" min_val = float("-0.472765") @@ -2031,6 +2216,7 @@ class Program_weight_tensor_parameter_184: class Program_weight_tensor_parameter_185: name = "parameter_185" + original_name = "layer_norm_21.b_0" shape = [512] dtype = "float32" min_val = float("-1.49627") @@ -2042,6 +2228,7 @@ class Program_weight_tensor_parameter_185: class Program_weight_tensor_parameter_186: name = "parameter_186" + original_name = "layer_norm_21.w_0" shape = [512] dtype = "float32" min_val = float("0.187682") @@ -2053,6 +2240,7 @@ class Program_weight_tensor_parameter_186: class Program_weight_tensor_parameter_187: name = "parameter_187" + original_name = "linear_37.b_0" shape = [512] dtype = "float32" min_val = float("-0.669698") @@ -2064,6 +2252,7 @@ class Program_weight_tensor_parameter_187: class Program_weight_tensor_parameter_188: name = "parameter_188" + original_name = "linear_37.w_0" shape = [2048, 512] dtype = "float32" min_val = float("-0.462316") @@ -2075,6 +2264,7 @@ class Program_weight_tensor_parameter_188: class Program_weight_tensor_parameter_189: name = "parameter_189" + original_name = "linear_36.b_0" shape = [2048] dtype = "float32" min_val = float("-1.34516") @@ -2086,6 +2276,7 @@ class Program_weight_tensor_parameter_189: class Program_weight_tensor_parameter_190: name = "parameter_190" + original_name = "linear_36.w_0" shape = [512, 2048] dtype = "float32" min_val = float("-0.490179") @@ -2097,6 +2288,7 @@ class Program_weight_tensor_parameter_190: class Program_weight_tensor_parameter_191: name = "parameter_191" + original_name = "layer_norm_20.b_0" shape = [512] dtype = "float32" min_val = float("-1.11991") @@ -2108,6 +2300,7 @@ class Program_weight_tensor_parameter_191: class Program_weight_tensor_parameter_192: name = "parameter_192" + original_name = "layer_norm_20.w_0" shape = [512] dtype = "float32" min_val = float("0.28138") @@ -2119,6 +2312,7 @@ class Program_weight_tensor_parameter_192: class Program_weight_tensor_parameter_193: name = "parameter_193" + original_name = "linear_35.b_0" shape = [512] dtype = "float32" min_val = float("-0.567245") @@ -2130,6 +2324,7 @@ class Program_weight_tensor_parameter_193: class Program_weight_tensor_parameter_194: name = "parameter_194" + original_name = "linear_35.w_0" shape = [512, 512] dtype = "float32" min_val = float("-0.312369") @@ -2141,6 +2336,7 @@ class Program_weight_tensor_parameter_194: class Program_weight_tensor_parameter_195: name = "parameter_195" + original_name = "linear_34.b_0" shape = [1536] dtype = "float32" min_val = float("-1.86803") @@ -2152,6 +2348,7 @@ class Program_weight_tensor_parameter_195: class Program_weight_tensor_parameter_196: name = "parameter_196" + original_name = "linear_34.w_0" shape = [512, 1536] dtype = "float32" min_val = float("-0.484798") @@ -2163,6 +2360,7 @@ class Program_weight_tensor_parameter_196: class Program_weight_tensor_parameter_197: name = "parameter_197" + original_name = "layer_norm_19.b_0" shape = [512] dtype = "float32" min_val = float("-1.04798") @@ -2174,6 +2372,7 @@ class Program_weight_tensor_parameter_197: class Program_weight_tensor_parameter_198: name = "parameter_198" + original_name = "layer_norm_19.w_0" shape = [512] dtype = "float32" min_val = float("0.12878") @@ -2185,6 +2384,7 @@ class Program_weight_tensor_parameter_198: class Program_weight_tensor_parameter_199: name = "parameter_199" + original_name = "linear_33.b_0" shape = [512] dtype = "float32" min_val = float("-0.613189") @@ -2196,6 +2396,7 @@ class Program_weight_tensor_parameter_199: class Program_weight_tensor_parameter_200: name = "parameter_200" + original_name = "linear_33.w_0" shape = [2048, 512] dtype = "float32" min_val = float("-0.536131") @@ -2207,6 +2408,7 @@ class Program_weight_tensor_parameter_200: class Program_weight_tensor_parameter_201: name = "parameter_201" + original_name = "linear_32.b_0" shape = [2048] dtype = "float32" min_val = float("-1.38762") @@ -2218,6 +2420,7 @@ class Program_weight_tensor_parameter_201: class Program_weight_tensor_parameter_202: name = "parameter_202" + original_name = "linear_32.w_0" shape = [512, 2048] dtype = "float32" min_val = float("-0.572844") @@ -2229,6 +2432,7 @@ class Program_weight_tensor_parameter_202: class Program_weight_tensor_parameter_203: name = "parameter_203" + original_name = "layer_norm_18.b_0" shape = [512] dtype = "float32" min_val = float("-1.17605") @@ -2240,6 +2444,7 @@ class Program_weight_tensor_parameter_203: class Program_weight_tensor_parameter_204: name = "parameter_204" + original_name = "layer_norm_18.w_0" shape = [512] dtype = "float32" min_val = float("0.193377") @@ -2251,6 +2456,7 @@ class Program_weight_tensor_parameter_204: class Program_weight_tensor_parameter_205: name = "parameter_205" + original_name = "linear_31.b_0" shape = [512] dtype = "float32" min_val = float("-0.571203") @@ -2262,6 +2468,7 @@ class Program_weight_tensor_parameter_205: class Program_weight_tensor_parameter_206: name = "parameter_206" + original_name = "linear_31.w_0" shape = [512, 512] dtype = "float32" min_val = float("-0.312326") @@ -2273,6 +2480,7 @@ class Program_weight_tensor_parameter_206: class Program_weight_tensor_parameter_207: name = "parameter_207" + original_name = "linear_30.b_0" shape = [1536] dtype = "float32" min_val = float("-2.46424") @@ -2284,6 +2492,7 @@ class Program_weight_tensor_parameter_207: class Program_weight_tensor_parameter_208: name = "parameter_208" + original_name = "linear_30.w_0" shape = [512, 1536] dtype = "float32" min_val = float("-0.494378") @@ -2295,6 +2504,7 @@ class Program_weight_tensor_parameter_208: class Program_weight_tensor_parameter_209: name = "parameter_209" + original_name = "layer_norm_17.b_0" shape = [512] dtype = "float32" min_val = float("-0.977111") @@ -2306,6 +2516,7 @@ class Program_weight_tensor_parameter_209: class Program_weight_tensor_parameter_210: name = "parameter_210" + original_name = "layer_norm_17.w_0" shape = [512] dtype = "float32" min_val = float("0.145346") @@ -2317,6 +2528,7 @@ class Program_weight_tensor_parameter_210: class Program_weight_tensor_parameter_211: name = "parameter_211" + original_name = "linear_29.b_0" shape = [512] dtype = "float32" min_val = float("-0.746403") @@ -2328,6 +2540,7 @@ class Program_weight_tensor_parameter_211: class Program_weight_tensor_parameter_212: name = "parameter_212" + original_name = "linear_29.w_0" shape = [2048, 512] dtype = "float32" min_val = float("-0.476859") @@ -2339,6 +2552,7 @@ class Program_weight_tensor_parameter_212: class Program_weight_tensor_parameter_213: name = "parameter_213" + original_name = "linear_28.b_0" shape = [2048] dtype = "float32" min_val = float("-1.29133") @@ -2350,6 +2564,7 @@ class Program_weight_tensor_parameter_213: class Program_weight_tensor_parameter_214: name = "parameter_214" + original_name = "linear_28.w_0" shape = [512, 2048] dtype = "float32" min_val = float("-0.346308") @@ -2361,6 +2576,7 @@ class Program_weight_tensor_parameter_214: class Program_weight_tensor_parameter_215: name = "parameter_215" + original_name = "layer_norm_16.b_0" shape = [512] dtype = "float32" min_val = float("-1.32993") @@ -2372,6 +2588,7 @@ class Program_weight_tensor_parameter_215: class Program_weight_tensor_parameter_216: name = "parameter_216" + original_name = "layer_norm_16.w_0" shape = [512] dtype = "float32" min_val = float("0.00483879") @@ -2383,6 +2600,7 @@ class Program_weight_tensor_parameter_216: class Program_weight_tensor_parameter_217: name = "parameter_217" + original_name = "linear_27.b_0" shape = [512] dtype = "float32" min_val = float("-0.743886") @@ -2394,6 +2612,7 @@ class Program_weight_tensor_parameter_217: class Program_weight_tensor_parameter_218: name = "parameter_218" + original_name = "linear_27.w_0" shape = [512, 512] dtype = "float32" min_val = float("-0.346557") @@ -2405,6 +2624,7 @@ class Program_weight_tensor_parameter_218: class Program_weight_tensor_parameter_219: name = "parameter_219" + original_name = "linear_26.b_0" shape = [1536] dtype = "float32" min_val = float("-2.40824") @@ -2416,6 +2636,7 @@ class Program_weight_tensor_parameter_219: class Program_weight_tensor_parameter_220: name = "parameter_220" + original_name = "linear_26.w_0" shape = [512, 1536] dtype = "float32" min_val = float("-0.366561") @@ -2427,6 +2648,7 @@ class Program_weight_tensor_parameter_220: class Program_weight_tensor_parameter_221: name = "parameter_221" + original_name = "layer_norm_15.b_0" shape = [512] dtype = "float32" min_val = float("-0.990461") @@ -2438,6 +2660,7 @@ class Program_weight_tensor_parameter_221: class Program_weight_tensor_parameter_222: name = "parameter_222" + original_name = "layer_norm_15.w_0" shape = [512] dtype = "float32" min_val = float("0.00666621") @@ -2449,6 +2672,7 @@ class Program_weight_tensor_parameter_222: class Program_weight_tensor_parameter_223: name = "parameter_223" + original_name = "linear_25.b_0" shape = [512] dtype = "float32" min_val = float("-0.675623") @@ -2460,6 +2684,7 @@ class Program_weight_tensor_parameter_223: class Program_weight_tensor_parameter_224: name = "parameter_224" + original_name = "linear_25.w_0" shape = [2048, 512] dtype = "float32" min_val = float("-0.345576") @@ -2471,6 +2696,7 @@ class Program_weight_tensor_parameter_224: class Program_weight_tensor_parameter_225: name = "parameter_225" + original_name = "linear_24.b_0" shape = [2048] dtype = "float32" min_val = float("-1.41396") @@ -2482,6 +2708,7 @@ class Program_weight_tensor_parameter_225: class Program_weight_tensor_parameter_226: name = "parameter_226" + original_name = "linear_24.w_0" shape = [512, 2048] dtype = "float32" min_val = float("-0.296496") @@ -2493,6 +2720,7 @@ class Program_weight_tensor_parameter_226: class Program_weight_tensor_parameter_227: name = "parameter_227" + original_name = "layer_norm_14.b_0" shape = [512] dtype = "float32" min_val = float("-1.19303") @@ -2504,6 +2732,7 @@ class Program_weight_tensor_parameter_227: class Program_weight_tensor_parameter_228: name = "parameter_228" + original_name = "layer_norm_14.w_0" shape = [512] dtype = "float32" min_val = float("0.000131497") @@ -2515,6 +2744,7 @@ class Program_weight_tensor_parameter_228: class Program_weight_tensor_parameter_229: name = "parameter_229" + original_name = "linear_23.b_0" shape = [512] dtype = "float32" min_val = float("-0.544079") @@ -2526,6 +2756,7 @@ class Program_weight_tensor_parameter_229: class Program_weight_tensor_parameter_230: name = "parameter_230" + original_name = "linear_23.w_0" shape = [512, 512] dtype = "float32" min_val = float("-0.291838") @@ -2537,6 +2768,7 @@ class Program_weight_tensor_parameter_230: class Program_weight_tensor_parameter_231: name = "parameter_231" + original_name = "linear_22.b_0" shape = [1536] dtype = "float32" min_val = float("-2.59543") @@ -2548,6 +2780,7 @@ class Program_weight_tensor_parameter_231: class Program_weight_tensor_parameter_232: name = "parameter_232" + original_name = "linear_22.w_0" shape = [512, 1536] dtype = "float32" min_val = float("-0.516349") @@ -2559,6 +2792,7 @@ class Program_weight_tensor_parameter_232: class Program_weight_tensor_parameter_233: name = "parameter_233" + original_name = "layer_norm_13.b_0" shape = [512] dtype = "float32" min_val = float("-0.965711") @@ -2570,6 +2804,7 @@ class Program_weight_tensor_parameter_233: class Program_weight_tensor_parameter_234: name = "parameter_234" + original_name = "layer_norm_13.w_0" shape = [512] dtype = "float32" min_val = float("0.00208257") @@ -2581,6 +2816,7 @@ class Program_weight_tensor_parameter_234: class Program_weight_tensor_parameter_235: name = "parameter_235" + original_name = "linear_21.b_0" shape = [512] dtype = "float32" min_val = float("-0.737248") @@ -2592,6 +2828,7 @@ class Program_weight_tensor_parameter_235: class Program_weight_tensor_parameter_236: name = "parameter_236" + original_name = "linear_21.w_0" shape = [2048, 512] dtype = "float32" min_val = float("-0.442399") @@ -2603,6 +2840,7 @@ class Program_weight_tensor_parameter_236: class Program_weight_tensor_parameter_237: name = "parameter_237" + original_name = "linear_20.b_0" shape = [2048] dtype = "float32" min_val = float("-1.71163") @@ -2614,6 +2852,7 @@ class Program_weight_tensor_parameter_237: class Program_weight_tensor_parameter_238: name = "parameter_238" + original_name = "linear_20.w_0" shape = [512, 2048] dtype = "float32" min_val = float("-0.353305") @@ -2625,6 +2864,7 @@ class Program_weight_tensor_parameter_238: class Program_weight_tensor_parameter_239: name = "parameter_239" + original_name = "layer_norm_12.b_0" shape = [512] dtype = "float32" min_val = float("-1.01993") @@ -2636,6 +2876,7 @@ class Program_weight_tensor_parameter_239: class Program_weight_tensor_parameter_240: name = "parameter_240" + original_name = "layer_norm_12.w_0" shape = [512] dtype = "float32" min_val = float("-0.00454127") @@ -2647,6 +2888,7 @@ class Program_weight_tensor_parameter_240: class Program_weight_tensor_parameter_241: name = "parameter_241" + original_name = "linear_19.b_0" shape = [512] dtype = "float32" min_val = float("-0.814813") @@ -2658,6 +2900,7 @@ class Program_weight_tensor_parameter_241: class Program_weight_tensor_parameter_242: name = "parameter_242" + original_name = "linear_19.w_0" shape = [512, 512] dtype = "float32" min_val = float("-0.270312") @@ -2669,6 +2912,7 @@ class Program_weight_tensor_parameter_242: class Program_weight_tensor_parameter_243: name = "parameter_243" + original_name = "linear_18.b_0" shape = [1536] dtype = "float32" min_val = float("-2.69002") @@ -2680,6 +2924,7 @@ class Program_weight_tensor_parameter_243: class Program_weight_tensor_parameter_244: name = "parameter_244" + original_name = "linear_18.w_0" shape = [512, 1536] dtype = "float32" min_val = float("-0.395219") @@ -2691,6 +2936,7 @@ class Program_weight_tensor_parameter_244: class Program_weight_tensor_parameter_245: name = "parameter_245" + original_name = "layer_norm_11.b_0" shape = [512] dtype = "float32" min_val = float("-0.873189") @@ -2702,6 +2948,7 @@ class Program_weight_tensor_parameter_245: class Program_weight_tensor_parameter_246: name = "parameter_246" + original_name = "layer_norm_11.w_0" shape = [512] dtype = "float32" min_val = float("-0.0117842") @@ -2713,6 +2960,7 @@ class Program_weight_tensor_parameter_246: class Program_weight_tensor_parameter_247: name = "parameter_247" + original_name = "linear_17.w_0" shape = [1024, 512] dtype = "float32" min_val = float("-0.25114") @@ -2724,6 +2972,7 @@ class Program_weight_tensor_parameter_247: class Program_weight_tensor_parameter_248: name = "parameter_248" + original_name = "layer_norm_10.b_0" shape = [1024] dtype = "float32" min_val = float("-0.941112") @@ -2735,6 +2984,7 @@ class Program_weight_tensor_parameter_248: class Program_weight_tensor_parameter_249: name = "parameter_249" + original_name = "layer_norm_10.w_0" shape = [1024] dtype = "float32" min_val = float("0.148363") @@ -2746,6 +2996,7 @@ class Program_weight_tensor_parameter_249: class Program_weight_tensor_parameter_250: name = "parameter_250" + original_name = "linear_16.b_0" shape = [256] dtype = "float32" min_val = float("-2.98842") @@ -2757,6 +3008,7 @@ class Program_weight_tensor_parameter_250: class Program_weight_tensor_parameter_251: name = "parameter_251" + original_name = "linear_16.w_0" shape = [1024, 256] dtype = "float32" min_val = float("-0.653492") @@ -2768,6 +3020,7 @@ class Program_weight_tensor_parameter_251: class Program_weight_tensor_parameter_252: name = "parameter_252" + original_name = "linear_15.b_0" shape = [1024] dtype = "float32" min_val = float("-1.41519") @@ -2779,6 +3032,7 @@ class Program_weight_tensor_parameter_252: class Program_weight_tensor_parameter_253: name = "parameter_253" + original_name = "linear_15.w_0" shape = [256, 1024] dtype = "float32" min_val = float("-0.302812") @@ -2790,6 +3044,7 @@ class Program_weight_tensor_parameter_253: class Program_weight_tensor_parameter_254: name = "parameter_254" + original_name = "layer_norm_9.b_0" shape = [256] dtype = "float32" min_val = float("-1.0555") @@ -2801,6 +3056,7 @@ class Program_weight_tensor_parameter_254: class Program_weight_tensor_parameter_255: name = "parameter_255" + original_name = "layer_norm_9.w_0" shape = [256] dtype = "float32" min_val = float("0.244394") @@ -2812,6 +3068,7 @@ class Program_weight_tensor_parameter_255: class Program_weight_tensor_parameter_256: name = "parameter_256" + original_name = "linear_14.b_0" shape = [256] dtype = "float32" min_val = float("-2.06598") @@ -2823,6 +3080,7 @@ class Program_weight_tensor_parameter_256: class Program_weight_tensor_parameter_257: name = "parameter_257" + original_name = "linear_14.w_0" shape = [256, 256] dtype = "float32" min_val = float("-0.215266") @@ -2834,6 +3092,7 @@ class Program_weight_tensor_parameter_257: class Program_weight_tensor_parameter_258: name = "parameter_258" + original_name = "linear_13.b_0" shape = [768] dtype = "float32" min_val = float("-1.52523") @@ -2845,6 +3104,7 @@ class Program_weight_tensor_parameter_258: class Program_weight_tensor_parameter_259: name = "parameter_259" + original_name = "linear_13.w_0" shape = [256, 768] dtype = "float32" min_val = float("-0.469426") @@ -2856,6 +3116,7 @@ class Program_weight_tensor_parameter_259: class Program_weight_tensor_parameter_260: name = "parameter_260" + original_name = "layer_norm_8.b_0" shape = [256] dtype = "float32" min_val = float("-1.00206") @@ -2867,6 +3128,7 @@ class Program_weight_tensor_parameter_260: class Program_weight_tensor_parameter_261: name = "parameter_261" + original_name = "layer_norm_8.w_0" shape = [256] dtype = "float32" min_val = float("0.386884") @@ -2878,6 +3140,7 @@ class Program_weight_tensor_parameter_261: class Program_weight_tensor_parameter_262: name = "parameter_262" + original_name = "linear_12.b_0" shape = [256] dtype = "float32" min_val = float("-1.89517") @@ -2889,6 +3152,7 @@ class Program_weight_tensor_parameter_262: class Program_weight_tensor_parameter_263: name = "parameter_263" + original_name = "linear_12.w_0" shape = [1024, 256] dtype = "float32" min_val = float("-0.375393") @@ -2900,6 +3164,7 @@ class Program_weight_tensor_parameter_263: class Program_weight_tensor_parameter_264: name = "parameter_264" + original_name = "linear_11.b_0" shape = [1024] dtype = "float32" min_val = float("-1.47422") @@ -2911,6 +3176,7 @@ class Program_weight_tensor_parameter_264: class Program_weight_tensor_parameter_265: name = "parameter_265" + original_name = "linear_11.w_0" shape = [256, 1024] dtype = "float32" min_val = float("-0.52185") @@ -2922,6 +3188,7 @@ class Program_weight_tensor_parameter_265: class Program_weight_tensor_parameter_266: name = "parameter_266" + original_name = "layer_norm_7.b_0" shape = [256] dtype = "float32" min_val = float("-0.535491") @@ -2933,6 +3200,7 @@ class Program_weight_tensor_parameter_266: class Program_weight_tensor_parameter_267: name = "parameter_267" + original_name = "layer_norm_7.w_0" shape = [256] dtype = "float32" min_val = float("0.000320881") @@ -2944,6 +3212,7 @@ class Program_weight_tensor_parameter_267: class Program_weight_tensor_parameter_268: name = "parameter_268" + original_name = "linear_10.b_0" shape = [256] dtype = "float32" min_val = float("-1.61703") @@ -2955,6 +3224,7 @@ class Program_weight_tensor_parameter_268: class Program_weight_tensor_parameter_269: name = "parameter_269" + original_name = "linear_10.w_0" shape = [256, 256] dtype = "float32" min_val = float("-0.28634") @@ -2966,6 +3236,7 @@ class Program_weight_tensor_parameter_269: class Program_weight_tensor_parameter_270: name = "parameter_270" + original_name = "linear_9.b_0" shape = [768] dtype = "float32" min_val = float("-1.82638") @@ -2977,6 +3248,7 @@ class Program_weight_tensor_parameter_270: class Program_weight_tensor_parameter_271: name = "parameter_271" + original_name = "linear_9.w_0" shape = [256, 768] dtype = "float32" min_val = float("-0.403714") @@ -2988,6 +3260,7 @@ class Program_weight_tensor_parameter_271: class Program_weight_tensor_parameter_272: name = "parameter_272" + original_name = "layer_norm_6.b_0" shape = [256] dtype = "float32" min_val = float("-0.347607") @@ -2999,6 +3272,7 @@ class Program_weight_tensor_parameter_272: class Program_weight_tensor_parameter_273: name = "parameter_273" + original_name = "layer_norm_6.w_0" shape = [256] dtype = "float32" min_val = float("0.178241") @@ -3010,6 +3284,7 @@ class Program_weight_tensor_parameter_273: class Program_weight_tensor_parameter_274: name = "parameter_274" + original_name = "linear_8.w_0" shape = [512, 256] dtype = "float32" min_val = float("-0.256307") @@ -3021,6 +3296,7 @@ class Program_weight_tensor_parameter_274: class Program_weight_tensor_parameter_275: name = "parameter_275" + original_name = "layer_norm_5.b_0" shape = [512] dtype = "float32" min_val = float("-1.40055") @@ -3032,6 +3308,7 @@ class Program_weight_tensor_parameter_275: class Program_weight_tensor_parameter_276: name = "parameter_276" + original_name = "layer_norm_5.w_0" shape = [512] dtype = "float32" min_val = float("0.214824") @@ -3043,6 +3320,7 @@ class Program_weight_tensor_parameter_276: class Program_weight_tensor_parameter_277: name = "parameter_277" + original_name = "linear_7.b_0" shape = [128] dtype = "float32" min_val = float("-1.5635") @@ -3054,6 +3332,7 @@ class Program_weight_tensor_parameter_277: class Program_weight_tensor_parameter_278: name = "parameter_278" + original_name = "linear_7.w_0" shape = [512, 128] dtype = "float32" min_val = float("-0.51444") @@ -3065,6 +3344,7 @@ class Program_weight_tensor_parameter_278: class Program_weight_tensor_parameter_279: name = "parameter_279" + original_name = "linear_6.b_0" shape = [512] dtype = "float32" min_val = float("-1.71915") @@ -3076,6 +3356,7 @@ class Program_weight_tensor_parameter_279: class Program_weight_tensor_parameter_280: name = "parameter_280" + original_name = "linear_6.w_0" shape = [128, 512] dtype = "float32" min_val = float("-0.37718") @@ -3087,6 +3368,7 @@ class Program_weight_tensor_parameter_280: class Program_weight_tensor_parameter_281: name = "parameter_281" + original_name = "layer_norm_4.b_0" shape = [128] dtype = "float32" min_val = float("-1.00194") @@ -3098,6 +3380,7 @@ class Program_weight_tensor_parameter_281: class Program_weight_tensor_parameter_282: name = "parameter_282" + original_name = "layer_norm_4.w_0" shape = [128] dtype = "float32" min_val = float("0.00034077") @@ -3109,6 +3392,7 @@ class Program_weight_tensor_parameter_282: class Program_weight_tensor_parameter_283: name = "parameter_283" + original_name = "linear_5.b_0" shape = [128] dtype = "float32" min_val = float("-0.700907") @@ -3120,6 +3404,7 @@ class Program_weight_tensor_parameter_283: class Program_weight_tensor_parameter_284: name = "parameter_284" + original_name = "linear_5.w_0" shape = [128, 128] dtype = "float32" min_val = float("-0.32846") @@ -3131,6 +3416,7 @@ class Program_weight_tensor_parameter_284: class Program_weight_tensor_parameter_285: name = "parameter_285" + original_name = "linear_4.b_0" shape = [384] dtype = "float32" min_val = float("-1.54238") @@ -3142,6 +3428,7 @@ class Program_weight_tensor_parameter_285: class Program_weight_tensor_parameter_286: name = "parameter_286" + original_name = "linear_4.w_0" shape = [128, 384] dtype = "float32" min_val = float("-0.425403") @@ -3153,6 +3440,7 @@ class Program_weight_tensor_parameter_286: class Program_weight_tensor_parameter_287: name = "parameter_287" + original_name = "layer_norm_3.b_0" shape = [128] dtype = "float32" min_val = float("-1.11156") @@ -3164,6 +3452,7 @@ class Program_weight_tensor_parameter_287: class Program_weight_tensor_parameter_288: name = "parameter_288" + original_name = "layer_norm_3.w_0" shape = [128] dtype = "float32" min_val = float("0.252732") @@ -3175,6 +3464,7 @@ class Program_weight_tensor_parameter_288: class Program_weight_tensor_parameter_289: name = "parameter_289" + original_name = "linear_3.b_0" shape = [128] dtype = "float32" min_val = float("-2.63589") @@ -3186,6 +3476,7 @@ class Program_weight_tensor_parameter_289: class Program_weight_tensor_parameter_290: name = "parameter_290" + original_name = "linear_3.w_0" shape = [512, 128] dtype = "float32" min_val = float("-0.325922") @@ -3197,6 +3488,7 @@ class Program_weight_tensor_parameter_290: class Program_weight_tensor_parameter_291: name = "parameter_291" + original_name = "linear_2.b_0" shape = [512] dtype = "float32" min_val = float("-1.455") @@ -3208,6 +3500,7 @@ class Program_weight_tensor_parameter_291: class Program_weight_tensor_parameter_292: name = "parameter_292" + original_name = "linear_2.w_0" shape = [128, 512] dtype = "float32" min_val = float("-0.459232") @@ -3219,6 +3512,7 @@ class Program_weight_tensor_parameter_292: class Program_weight_tensor_parameter_293: name = "parameter_293" + original_name = "layer_norm_2.b_0" shape = [128] dtype = "float32" min_val = float("-1.28569") @@ -3230,6 +3524,7 @@ class Program_weight_tensor_parameter_293: class Program_weight_tensor_parameter_294: name = "parameter_294" + original_name = "layer_norm_2.w_0" shape = [128] dtype = "float32" min_val = float("0.359776") @@ -3241,6 +3536,7 @@ class Program_weight_tensor_parameter_294: class Program_weight_tensor_parameter_295: name = "parameter_295" + original_name = "linear_1.b_0" shape = [128] dtype = "float32" min_val = float("-0.853647") @@ -3252,6 +3548,7 @@ class Program_weight_tensor_parameter_295: class Program_weight_tensor_parameter_296: name = "parameter_296" + original_name = "linear_1.w_0" shape = [128, 128] dtype = "float32" min_val = float("-0.284258") @@ -3263,6 +3560,7 @@ class Program_weight_tensor_parameter_296: class Program_weight_tensor_parameter_297: name = "parameter_297" + original_name = "linear_0.b_0" shape = [384] dtype = "float32" min_val = float("-2.0302") @@ -3274,6 +3572,7 @@ class Program_weight_tensor_parameter_297: class Program_weight_tensor_parameter_298: name = "parameter_298" + original_name = "linear_0.w_0" shape = [128, 384] dtype = "float32" min_val = float("-0.530585") @@ -3285,6 +3584,7 @@ class Program_weight_tensor_parameter_298: class Program_weight_tensor_parameter_299: name = "parameter_299" + original_name = "layer_norm_1.b_0" shape = [128] dtype = "float32" min_val = float("-0.862344") @@ -3296,6 +3596,7 @@ class Program_weight_tensor_parameter_299: class Program_weight_tensor_parameter_300: name = "parameter_300" + original_name = "layer_norm_1.w_0" shape = [128] dtype = "float32" min_val = float("-0.000493348") @@ -3307,6 +3608,7 @@ class Program_weight_tensor_parameter_300: class Program_weight_tensor_parameter_301: name = "parameter_301" + original_name = "layer_norm_0.b_0" shape = [128] dtype = "float32" min_val = float("-1.25937") @@ -3318,6 +3620,7 @@ class Program_weight_tensor_parameter_301: class Program_weight_tensor_parameter_302: name = "parameter_302" + original_name = "layer_norm_0.w_0" shape = [128] dtype = "float32" min_val = float("-0.0173378") @@ -3329,6 +3632,7 @@ class Program_weight_tensor_parameter_302: class Program_weight_tensor_parameter_303: name = "parameter_303" + original_name = "conv2d_0.b_0" shape = [128] dtype = "float32" min_val = float("-1.13759") @@ -3340,6 +3644,7 @@ class Program_weight_tensor_parameter_303: class Program_weight_tensor_parameter_304: name = "parameter_304" + original_name = "conv2d_0.w_0" shape = [128, 3, 4, 4] dtype = "float32" min_val = float("-0.315612") diff --git a/paddle_samples/PaddleX/SwinTransformer_base_patch4_window7_224/subgraph_0/input_meta.py b/paddle_samples/PaddleX/SwinTransformer_base_patch4_window7_224/subgraph_0/input_meta.py index 93e586cff..f5052ce0d 100644 --- a/paddle_samples/PaddleX/SwinTransformer_base_patch4_window7_224/subgraph_0/input_meta.py +++ b/paddle_samples/PaddleX/SwinTransformer_base_patch4_window7_224/subgraph_0/input_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_data_0: name = "data_0" + original_name = "param_0" shape = [169, 4] dtype = "float32" min_val = float("-8.34796") @@ -11,6 +12,7 @@ class Program_weight_tensor_data_0: class Program_weight_tensor_data_1: name = "data_1" + original_name = "param_1" shape = [169, 4] dtype = "float32" min_val = float("-6.26646") @@ -22,6 +24,7 @@ class Program_weight_tensor_data_1: class Program_weight_tensor_data_2: name = "data_2" + original_name = "param_10" shape = [169, 16] dtype = "float32" min_val = float("-9.44072") @@ -33,6 +36,7 @@ class Program_weight_tensor_data_2: class Program_weight_tensor_data_3: name = "data_3" + original_name = "param_11" shape = [169, 16] dtype = "float32" min_val = float("-8.36854") @@ -44,6 +48,7 @@ class Program_weight_tensor_data_3: class Program_weight_tensor_data_4: name = "data_4" + original_name = "param_12" shape = [169, 16] dtype = "float32" min_val = float("-8.9458") @@ -55,6 +60,7 @@ class Program_weight_tensor_data_4: class Program_weight_tensor_data_5: name = "data_5" + original_name = "param_13" shape = [169, 16] dtype = "float32" min_val = float("-4.65111") @@ -66,6 +72,7 @@ class Program_weight_tensor_data_5: class Program_weight_tensor_data_6: name = "data_6" + original_name = "param_14" shape = [169, 16] dtype = "float32" min_val = float("-8.62345") @@ -77,6 +84,7 @@ class Program_weight_tensor_data_6: class Program_weight_tensor_data_7: name = "data_7" + original_name = "param_15" shape = [169, 16] dtype = "float32" min_val = float("-5.26338") @@ -88,6 +96,7 @@ class Program_weight_tensor_data_7: class Program_weight_tensor_data_8: name = "data_8" + original_name = "param_16" shape = [169, 16] dtype = "float32" min_val = float("-11.8762") @@ -99,6 +108,7 @@ class Program_weight_tensor_data_8: class Program_weight_tensor_data_9: name = "data_9" + original_name = "param_17" shape = [169, 16] dtype = "float32" min_val = float("-8.85042") @@ -110,6 +120,7 @@ class Program_weight_tensor_data_9: class Program_weight_tensor_data_10: name = "data_10" + original_name = "param_18" shape = [169, 16] dtype = "float32" min_val = float("-10.4529") @@ -121,6 +132,7 @@ class Program_weight_tensor_data_10: class Program_weight_tensor_data_11: name = "data_11" + original_name = "param_19" shape = [169, 16] dtype = "float32" min_val = float("-10.8006") @@ -132,6 +144,7 @@ class Program_weight_tensor_data_11: class Program_weight_tensor_data_12: name = "data_12" + original_name = "param_2" shape = [169, 8] dtype = "float32" min_val = float("-6.02162") @@ -143,6 +156,7 @@ class Program_weight_tensor_data_12: class Program_weight_tensor_data_13: name = "data_13" + original_name = "param_20" shape = [169, 16] dtype = "float32" min_val = float("-14.7827") @@ -154,6 +168,7 @@ class Program_weight_tensor_data_13: class Program_weight_tensor_data_14: name = "data_14" + original_name = "param_21" shape = [169, 16] dtype = "float32" min_val = float("-14.5049") @@ -165,6 +180,7 @@ class Program_weight_tensor_data_14: class Program_weight_tensor_data_15: name = "data_15" + original_name = "param_22" shape = [169, 32] dtype = "float32" min_val = float("-15.6921") @@ -176,6 +192,7 @@ class Program_weight_tensor_data_15: class Program_weight_tensor_data_16: name = "data_16" + original_name = "param_23" shape = [169, 32] dtype = "float32" min_val = float("-15.7445") @@ -187,6 +204,7 @@ class Program_weight_tensor_data_16: class Program_weight_tensor_data_17: name = "data_17" + original_name = "param_3" shape = [169, 8] dtype = "float32" min_val = float("-7.64378") @@ -198,6 +216,7 @@ class Program_weight_tensor_data_17: class Program_weight_tensor_data_18: name = "data_18" + original_name = "param_4" shape = [169, 16] dtype = "float32" min_val = float("-10.9591") @@ -209,6 +228,7 @@ class Program_weight_tensor_data_18: class Program_weight_tensor_data_19: name = "data_19" + original_name = "param_5" shape = [169, 16] dtype = "float32" min_val = float("-10.1732") @@ -220,6 +240,7 @@ class Program_weight_tensor_data_19: class Program_weight_tensor_data_20: name = "data_20" + original_name = "param_6" shape = [169, 16] dtype = "float32" min_val = float("-10.5645") @@ -231,6 +252,7 @@ class Program_weight_tensor_data_20: class Program_weight_tensor_data_21: name = "data_21" + original_name = "param_7" shape = [169, 16] dtype = "float32" min_val = float("-7.65788") @@ -242,6 +264,7 @@ class Program_weight_tensor_data_21: class Program_weight_tensor_data_22: name = "data_22" + original_name = "param_8" shape = [169, 16] dtype = "float32" min_val = float("-12.4201") @@ -253,6 +276,7 @@ class Program_weight_tensor_data_22: class Program_weight_tensor_data_23: name = "data_23" + original_name = "param_9" shape = [169, 16] dtype = "float32" min_val = float("-9.87954") @@ -264,6 +288,7 @@ class Program_weight_tensor_data_23: class Program_weight_tensor_data_24: name = "data_24" + original_name = "var_0" shape = [16, 3, 224, 224] dtype = "float32" min_val = float("-4.7145") @@ -275,6 +300,7 @@ class Program_weight_tensor_data_24: class Program_weight_tensor_data_25: name = "data_25" + original_name = "var_1021" shape = [49, 49] dtype = "int64" min_val = 0 @@ -284,6 +310,7 @@ class Program_weight_tensor_data_25: class Program_weight_tensor_data_26: name = "data_26" + original_name = "var_1101" shape = [49, 49] dtype = "int64" min_val = 0 @@ -293,6 +320,7 @@ class Program_weight_tensor_data_26: class Program_weight_tensor_data_27: name = "data_27" + original_name = "var_1163" shape = [49, 49] dtype = "int64" min_val = 0 @@ -302,6 +330,7 @@ class Program_weight_tensor_data_27: class Program_weight_tensor_data_28: name = "data_28" + original_name = "var_1243" shape = [49, 49] dtype = "int64" min_val = 0 @@ -311,6 +340,7 @@ class Program_weight_tensor_data_28: class Program_weight_tensor_data_29: name = "data_29" + original_name = "var_1305" shape = [49, 49] dtype = "int64" min_val = 0 @@ -320,6 +350,7 @@ class Program_weight_tensor_data_29: class Program_weight_tensor_data_30: name = "data_30" + original_name = "var_1385" shape = [49, 49] dtype = "int64" min_val = 0 @@ -329,6 +360,7 @@ class Program_weight_tensor_data_30: class Program_weight_tensor_data_31: name = "data_31" + original_name = "var_1447" shape = [49, 49] dtype = "int64" min_val = 0 @@ -338,6 +370,7 @@ class Program_weight_tensor_data_31: class Program_weight_tensor_data_32: name = "data_32" + original_name = "var_1527" shape = [49, 49] dtype = "int64" min_val = 0 @@ -347,6 +380,7 @@ class Program_weight_tensor_data_32: class Program_weight_tensor_data_33: name = "data_33" + original_name = "var_159" shape = [49, 49] dtype = "int64" min_val = 0 @@ -356,6 +390,7 @@ class Program_weight_tensor_data_33: class Program_weight_tensor_data_34: name = "data_34" + original_name = "var_1599" shape = [49, 49] dtype = "int64" min_val = 0 @@ -365,6 +400,7 @@ class Program_weight_tensor_data_34: class Program_weight_tensor_data_35: name = "data_35" + original_name = "var_1679" shape = [49, 49] dtype = "int64" min_val = 0 @@ -374,6 +410,7 @@ class Program_weight_tensor_data_35: class Program_weight_tensor_data_36: name = "data_36" + original_name = "var_21" shape = [49, 49] dtype = "int64" min_val = 0 @@ -383,6 +420,7 @@ class Program_weight_tensor_data_36: class Program_weight_tensor_data_37: name = "data_37" + original_name = "var_239" shape = [49, 49] dtype = "int64" min_val = 0 @@ -392,6 +430,7 @@ class Program_weight_tensor_data_37: class Program_weight_tensor_data_38: name = "data_38" + original_name = "var_311" shape = [49, 49] dtype = "int64" min_val = 0 @@ -401,6 +440,7 @@ class Program_weight_tensor_data_38: class Program_weight_tensor_data_39: name = "data_39" + original_name = "var_391" shape = [49, 49] dtype = "int64" min_val = 0 @@ -410,6 +450,7 @@ class Program_weight_tensor_data_39: class Program_weight_tensor_data_40: name = "data_40" + original_name = "var_453" shape = [49, 49] dtype = "int64" min_val = 0 @@ -419,6 +460,7 @@ class Program_weight_tensor_data_40: class Program_weight_tensor_data_41: name = "data_41" + original_name = "var_533" shape = [49, 49] dtype = "int64" min_val = 0 @@ -428,6 +470,7 @@ class Program_weight_tensor_data_41: class Program_weight_tensor_data_42: name = "data_42" + original_name = "var_595" shape = [49, 49] dtype = "int64" min_val = 0 @@ -437,6 +480,7 @@ class Program_weight_tensor_data_42: class Program_weight_tensor_data_43: name = "data_43" + original_name = "var_675" shape = [49, 49] dtype = "int64" min_val = 0 @@ -446,6 +490,7 @@ class Program_weight_tensor_data_43: class Program_weight_tensor_data_44: name = "data_44" + original_name = "var_737" shape = [49, 49] dtype = "int64" min_val = 0 @@ -455,6 +500,7 @@ class Program_weight_tensor_data_44: class Program_weight_tensor_data_45: name = "data_45" + original_name = "var_817" shape = [49, 49] dtype = "int64" min_val = 0 @@ -464,6 +510,7 @@ class Program_weight_tensor_data_45: class Program_weight_tensor_data_46: name = "data_46" + original_name = "var_87" shape = [49, 49] dtype = "int64" min_val = 0 @@ -473,6 +520,7 @@ class Program_weight_tensor_data_46: class Program_weight_tensor_data_47: name = "data_47" + original_name = "var_879" shape = [49, 49] dtype = "int64" min_val = 0 @@ -482,6 +530,7 @@ class Program_weight_tensor_data_47: class Program_weight_tensor_data_48: name = "data_48" + original_name = "var_959" shape = [49, 49] dtype = "int64" min_val = 0 diff --git a/paddle_samples/PaddleX/SwinTransformer_base_patch4_window7_224/subgraph_0/weight_meta.py b/paddle_samples/PaddleX/SwinTransformer_base_patch4_window7_224/subgraph_0/weight_meta.py index 722f18fcc..7534953ab 100644 --- a/paddle_samples/PaddleX/SwinTransformer_base_patch4_window7_224/subgraph_0/weight_meta.py +++ b/paddle_samples/PaddleX/SwinTransformer_base_patch4_window7_224/subgraph_0/weight_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_parameter_0: name = "parameter_0" + original_name = "linear_99.b_0" shape = [102] dtype = "float32" data = None @@ -7,6 +8,7 @@ class Program_weight_tensor_parameter_0: class Program_weight_tensor_parameter_1: name = "parameter_1" + original_name = "linear_99.w_0" shape = [1024, 102] dtype = "float32" min_val = float("-0.0825595") @@ -18,6 +20,7 @@ class Program_weight_tensor_parameter_1: class Program_weight_tensor_parameter_2: name = "parameter_2" + original_name = "layer_norm_52.b_0" shape = [1024] dtype = "float32" min_val = float("-0.300978") @@ -29,6 +32,7 @@ class Program_weight_tensor_parameter_2: class Program_weight_tensor_parameter_3: name = "parameter_3" + original_name = "layer_norm_52.w_0" shape = [1024] dtype = "float32" min_val = float("-0.00031831") @@ -40,6 +44,7 @@ class Program_weight_tensor_parameter_3: class Program_weight_tensor_parameter_4: name = "parameter_4" + original_name = "linear_98.b_0" shape = [1024] dtype = "float32" min_val = float("-5.67573") @@ -51,6 +56,7 @@ class Program_weight_tensor_parameter_4: class Program_weight_tensor_parameter_5: name = "parameter_5" + original_name = "linear_98.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-1.74188") @@ -62,6 +68,7 @@ class Program_weight_tensor_parameter_5: class Program_weight_tensor_parameter_6: name = "parameter_6" + original_name = "linear_97.b_0" shape = [4096] dtype = "float32" min_val = float("-2.69186") @@ -73,6 +80,7 @@ class Program_weight_tensor_parameter_6: class Program_weight_tensor_parameter_7: name = "parameter_7" + original_name = "linear_97.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.701426") @@ -84,6 +92,7 @@ class Program_weight_tensor_parameter_7: class Program_weight_tensor_parameter_8: name = "parameter_8" + original_name = "layer_norm_51.b_0" shape = [1024] dtype = "float32" min_val = float("-1.02085") @@ -95,6 +104,7 @@ class Program_weight_tensor_parameter_8: class Program_weight_tensor_parameter_9: name = "parameter_9" + original_name = "layer_norm_51.w_0" shape = [1024] dtype = "float32" min_val = float("-0.00359901") @@ -106,6 +116,7 @@ class Program_weight_tensor_parameter_9: class Program_weight_tensor_parameter_10: name = "parameter_10" + original_name = "linear_96.b_0" shape = [1024] dtype = "float32" min_val = float("-5.60048") @@ -117,6 +128,7 @@ class Program_weight_tensor_parameter_10: class Program_weight_tensor_parameter_11: name = "parameter_11" + original_name = "linear_96.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-1.21512") @@ -128,6 +140,7 @@ class Program_weight_tensor_parameter_11: class Program_weight_tensor_parameter_12: name = "parameter_12" + original_name = "linear_95.b_0" shape = [3072] dtype = "float32" min_val = float("-2.51441") @@ -139,6 +152,7 @@ class Program_weight_tensor_parameter_12: class Program_weight_tensor_parameter_13: name = "parameter_13" + original_name = "linear_95.w_0" shape = [1024, 3072] dtype = "float32" min_val = float("-0.343697") @@ -150,6 +164,7 @@ class Program_weight_tensor_parameter_13: class Program_weight_tensor_parameter_14: name = "parameter_14" + original_name = "layer_norm_50.b_0" shape = [1024] dtype = "float32" min_val = float("-2.02304") @@ -161,6 +176,7 @@ class Program_weight_tensor_parameter_14: class Program_weight_tensor_parameter_15: name = "parameter_15" + original_name = "layer_norm_50.w_0" shape = [1024] dtype = "float32" min_val = float("-0.000157167") @@ -172,6 +188,7 @@ class Program_weight_tensor_parameter_15: class Program_weight_tensor_parameter_16: name = "parameter_16" + original_name = "linear_94.b_0" shape = [1024] dtype = "float32" min_val = float("-5.83856") @@ -183,6 +200,7 @@ class Program_weight_tensor_parameter_16: class Program_weight_tensor_parameter_17: name = "parameter_17" + original_name = "linear_94.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-1.44268") @@ -194,6 +212,7 @@ class Program_weight_tensor_parameter_17: class Program_weight_tensor_parameter_18: name = "parameter_18" + original_name = "linear_93.b_0" shape = [4096] dtype = "float32" min_val = float("-1.17796") @@ -205,6 +224,7 @@ class Program_weight_tensor_parameter_18: class Program_weight_tensor_parameter_19: name = "parameter_19" + original_name = "linear_93.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.630848") @@ -216,6 +236,7 @@ class Program_weight_tensor_parameter_19: class Program_weight_tensor_parameter_20: name = "parameter_20" + original_name = "layer_norm_49.b_0" shape = [1024] dtype = "float32" min_val = float("-0.97426") @@ -227,6 +248,7 @@ class Program_weight_tensor_parameter_20: class Program_weight_tensor_parameter_21: name = "parameter_21" + original_name = "layer_norm_49.w_0" shape = [1024] dtype = "float32" min_val = float("-0.0164068") @@ -238,6 +260,7 @@ class Program_weight_tensor_parameter_21: class Program_weight_tensor_parameter_22: name = "parameter_22" + original_name = "linear_92.b_0" shape = [1024] dtype = "float32" min_val = float("-5.33624") @@ -249,6 +272,7 @@ class Program_weight_tensor_parameter_22: class Program_weight_tensor_parameter_23: name = "parameter_23" + original_name = "linear_92.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-1.6303") @@ -260,6 +284,7 @@ class Program_weight_tensor_parameter_23: class Program_weight_tensor_parameter_24: name = "parameter_24" + original_name = "linear_91.b_0" shape = [3072] dtype = "float32" min_val = float("-2.67215") @@ -271,6 +296,7 @@ class Program_weight_tensor_parameter_24: class Program_weight_tensor_parameter_25: name = "parameter_25" + original_name = "linear_91.w_0" shape = [1024, 3072] dtype = "float32" min_val = float("-0.325901") @@ -282,6 +308,7 @@ class Program_weight_tensor_parameter_25: class Program_weight_tensor_parameter_26: name = "parameter_26" + original_name = "layer_norm_48.b_0" shape = [1024] dtype = "float32" min_val = float("-1.88525") @@ -293,6 +320,7 @@ class Program_weight_tensor_parameter_26: class Program_weight_tensor_parameter_27: name = "parameter_27" + original_name = "layer_norm_48.w_0" shape = [1024] dtype = "float32" min_val = float("0.000125914") @@ -304,6 +332,7 @@ class Program_weight_tensor_parameter_27: class Program_weight_tensor_parameter_28: name = "parameter_28" + original_name = "linear_90.w_0" shape = [2048, 1024] dtype = "float32" min_val = float("-1.13124") @@ -315,6 +344,7 @@ class Program_weight_tensor_parameter_28: class Program_weight_tensor_parameter_29: name = "parameter_29" + original_name = "layer_norm_47.b_0" shape = [2048] dtype = "float32" min_val = float("-1.56112") @@ -326,6 +356,7 @@ class Program_weight_tensor_parameter_29: class Program_weight_tensor_parameter_30: name = "parameter_30" + original_name = "layer_norm_47.w_0" shape = [2048] dtype = "float32" min_val = float("0.0890273") @@ -337,6 +368,7 @@ class Program_weight_tensor_parameter_30: class Program_weight_tensor_parameter_31: name = "parameter_31" + original_name = "linear_89.b_0" shape = [512] dtype = "float32" min_val = float("-1.8465") @@ -348,6 +380,7 @@ class Program_weight_tensor_parameter_31: class Program_weight_tensor_parameter_32: name = "parameter_32" + original_name = "linear_89.w_0" shape = [2048, 512] dtype = "float32" min_val = float("-1.91128") @@ -359,6 +392,7 @@ class Program_weight_tensor_parameter_32: class Program_weight_tensor_parameter_33: name = "parameter_33" + original_name = "linear_88.b_0" shape = [2048] dtype = "float32" min_val = float("-2.48175") @@ -370,6 +404,7 @@ class Program_weight_tensor_parameter_33: class Program_weight_tensor_parameter_34: name = "parameter_34" + original_name = "linear_88.w_0" shape = [512, 2048] dtype = "float32" min_val = float("-0.784368") @@ -381,6 +416,7 @@ class Program_weight_tensor_parameter_34: class Program_weight_tensor_parameter_35: name = "parameter_35" + original_name = "layer_norm_46.b_0" shape = [512] dtype = "float32" min_val = float("-1.89666") @@ -392,6 +428,7 @@ class Program_weight_tensor_parameter_35: class Program_weight_tensor_parameter_36: name = "parameter_36" + original_name = "layer_norm_46.w_0" shape = [512] dtype = "float32" min_val = float("0.282367") @@ -403,6 +440,7 @@ class Program_weight_tensor_parameter_36: class Program_weight_tensor_parameter_37: name = "parameter_37" + original_name = "linear_87.b_0" shape = [512] dtype = "float32" min_val = float("-2.21121") @@ -414,6 +452,7 @@ class Program_weight_tensor_parameter_37: class Program_weight_tensor_parameter_38: name = "parameter_38" + original_name = "linear_87.w_0" shape = [512, 512] dtype = "float32" min_val = float("-1.57379") @@ -425,6 +464,7 @@ class Program_weight_tensor_parameter_38: class Program_weight_tensor_parameter_39: name = "parameter_39" + original_name = "linear_86.b_0" shape = [1536] dtype = "float32" min_val = float("-2.49419") @@ -436,6 +476,7 @@ class Program_weight_tensor_parameter_39: class Program_weight_tensor_parameter_40: name = "parameter_40" + original_name = "linear_86.w_0" shape = [512, 1536] dtype = "float32" min_val = float("-0.350373") @@ -447,6 +488,7 @@ class Program_weight_tensor_parameter_40: class Program_weight_tensor_parameter_41: name = "parameter_41" + original_name = "layer_norm_45.b_0" shape = [512] dtype = "float32" min_val = float("-1.39961") @@ -458,6 +500,7 @@ class Program_weight_tensor_parameter_41: class Program_weight_tensor_parameter_42: name = "parameter_42" + original_name = "layer_norm_45.w_0" shape = [512] dtype = "float32" min_val = float("-0.0125483") @@ -469,6 +512,7 @@ class Program_weight_tensor_parameter_42: class Program_weight_tensor_parameter_43: name = "parameter_43" + original_name = "linear_85.b_0" shape = [512] dtype = "float32" min_val = float("-1.76007") @@ -480,6 +524,7 @@ class Program_weight_tensor_parameter_43: class Program_weight_tensor_parameter_44: name = "parameter_44" + original_name = "linear_85.w_0" shape = [2048, 512] dtype = "float32" min_val = float("-1.31572") @@ -491,6 +536,7 @@ class Program_weight_tensor_parameter_44: class Program_weight_tensor_parameter_45: name = "parameter_45" + original_name = "linear_84.b_0" shape = [2048] dtype = "float32" min_val = float("-2.10307") @@ -502,6 +548,7 @@ class Program_weight_tensor_parameter_45: class Program_weight_tensor_parameter_46: name = "parameter_46" + original_name = "linear_84.w_0" shape = [512, 2048] dtype = "float32" min_val = float("-0.560885") @@ -513,6 +560,7 @@ class Program_weight_tensor_parameter_46: class Program_weight_tensor_parameter_47: name = "parameter_47" + original_name = "layer_norm_44.b_0" shape = [512] dtype = "float32" min_val = float("-2.25157") @@ -524,6 +572,7 @@ class Program_weight_tensor_parameter_47: class Program_weight_tensor_parameter_48: name = "parameter_48" + original_name = "layer_norm_44.w_0" shape = [512] dtype = "float32" min_val = float("0.329498") @@ -535,6 +584,7 @@ class Program_weight_tensor_parameter_48: class Program_weight_tensor_parameter_49: name = "parameter_49" + original_name = "linear_83.b_0" shape = [512] dtype = "float32" min_val = float("-2.1653") @@ -546,6 +596,7 @@ class Program_weight_tensor_parameter_49: class Program_weight_tensor_parameter_50: name = "parameter_50" + original_name = "linear_83.w_0" shape = [512, 512] dtype = "float32" min_val = float("-0.883493") @@ -557,6 +608,7 @@ class Program_weight_tensor_parameter_50: class Program_weight_tensor_parameter_51: name = "parameter_51" + original_name = "linear_82.b_0" shape = [1536] dtype = "float32" min_val = float("-2.48684") @@ -568,6 +620,7 @@ class Program_weight_tensor_parameter_51: class Program_weight_tensor_parameter_52: name = "parameter_52" + original_name = "linear_82.w_0" shape = [512, 1536] dtype = "float32" min_val = float("-0.375647") @@ -579,6 +632,7 @@ class Program_weight_tensor_parameter_52: class Program_weight_tensor_parameter_53: name = "parameter_53" + original_name = "layer_norm_43.b_0" shape = [512] dtype = "float32" min_val = float("-1.1173") @@ -590,6 +644,7 @@ class Program_weight_tensor_parameter_53: class Program_weight_tensor_parameter_54: name = "parameter_54" + original_name = "layer_norm_43.w_0" shape = [512] dtype = "float32" min_val = float("-0.462549") @@ -601,6 +656,7 @@ class Program_weight_tensor_parameter_54: class Program_weight_tensor_parameter_55: name = "parameter_55" + original_name = "linear_81.b_0" shape = [512] dtype = "float32" min_val = float("-1.68048") @@ -612,6 +668,7 @@ class Program_weight_tensor_parameter_55: class Program_weight_tensor_parameter_56: name = "parameter_56" + original_name = "linear_81.w_0" shape = [2048, 512] dtype = "float32" min_val = float("-1.44987") @@ -623,6 +680,7 @@ class Program_weight_tensor_parameter_56: class Program_weight_tensor_parameter_57: name = "parameter_57" + original_name = "linear_80.b_0" shape = [2048] dtype = "float32" min_val = float("-1.27964") @@ -634,6 +692,7 @@ class Program_weight_tensor_parameter_57: class Program_weight_tensor_parameter_58: name = "parameter_58" + original_name = "linear_80.w_0" shape = [512, 2048] dtype = "float32" min_val = float("-0.501075") @@ -645,6 +704,7 @@ class Program_weight_tensor_parameter_58: class Program_weight_tensor_parameter_59: name = "parameter_59" + original_name = "layer_norm_42.b_0" shape = [512] dtype = "float32" min_val = float("-2.21216") @@ -656,6 +716,7 @@ class Program_weight_tensor_parameter_59: class Program_weight_tensor_parameter_60: name = "parameter_60" + original_name = "layer_norm_42.w_0" shape = [512] dtype = "float32" min_val = float("0.28918") @@ -667,6 +728,7 @@ class Program_weight_tensor_parameter_60: class Program_weight_tensor_parameter_61: name = "parameter_61" + original_name = "linear_79.b_0" shape = [512] dtype = "float32" min_val = float("-1.41871") @@ -678,6 +740,7 @@ class Program_weight_tensor_parameter_61: class Program_weight_tensor_parameter_62: name = "parameter_62" + original_name = "linear_79.w_0" shape = [512, 512] dtype = "float32" min_val = float("-0.695611") @@ -689,6 +752,7 @@ class Program_weight_tensor_parameter_62: class Program_weight_tensor_parameter_63: name = "parameter_63" + original_name = "linear_78.b_0" shape = [1536] dtype = "float32" min_val = float("-2.72392") @@ -700,6 +764,7 @@ class Program_weight_tensor_parameter_63: class Program_weight_tensor_parameter_64: name = "parameter_64" + original_name = "linear_78.w_0" shape = [512, 1536] dtype = "float32" min_val = float("-0.363268") @@ -711,6 +776,7 @@ class Program_weight_tensor_parameter_64: class Program_weight_tensor_parameter_65: name = "parameter_65" + original_name = "layer_norm_41.b_0" shape = [512] dtype = "float32" min_val = float("-0.943073") @@ -722,6 +788,7 @@ class Program_weight_tensor_parameter_65: class Program_weight_tensor_parameter_66: name = "parameter_66" + original_name = "layer_norm_41.w_0" shape = [512] dtype = "float32" min_val = float("-0.000762541") @@ -733,6 +800,7 @@ class Program_weight_tensor_parameter_66: class Program_weight_tensor_parameter_67: name = "parameter_67" + original_name = "linear_77.b_0" shape = [512] dtype = "float32" min_val = float("-1.33425") @@ -744,6 +812,7 @@ class Program_weight_tensor_parameter_67: class Program_weight_tensor_parameter_68: name = "parameter_68" + original_name = "linear_77.w_0" shape = [2048, 512] dtype = "float32" min_val = float("-1.14728") @@ -755,6 +824,7 @@ class Program_weight_tensor_parameter_68: class Program_weight_tensor_parameter_69: name = "parameter_69" + original_name = "linear_76.b_0" shape = [2048] dtype = "float32" min_val = float("-0.905062") @@ -766,6 +836,7 @@ class Program_weight_tensor_parameter_69: class Program_weight_tensor_parameter_70: name = "parameter_70" + original_name = "linear_76.w_0" shape = [512, 2048] dtype = "float32" min_val = float("-0.446582") @@ -777,6 +848,7 @@ class Program_weight_tensor_parameter_70: class Program_weight_tensor_parameter_71: name = "parameter_71" + original_name = "layer_norm_40.b_0" shape = [512] dtype = "float32" min_val = float("-1.9656") @@ -788,6 +860,7 @@ class Program_weight_tensor_parameter_71: class Program_weight_tensor_parameter_72: name = "parameter_72" + original_name = "layer_norm_40.w_0" shape = [512] dtype = "float32" min_val = float("0.251532") @@ -799,6 +872,7 @@ class Program_weight_tensor_parameter_72: class Program_weight_tensor_parameter_73: name = "parameter_73" + original_name = "linear_75.b_0" shape = [512] dtype = "float32" min_val = float("-1.54941") @@ -810,6 +884,7 @@ class Program_weight_tensor_parameter_73: class Program_weight_tensor_parameter_74: name = "parameter_74" + original_name = "linear_75.w_0" shape = [512, 512] dtype = "float32" min_val = float("-0.686293") @@ -821,6 +896,7 @@ class Program_weight_tensor_parameter_74: class Program_weight_tensor_parameter_75: name = "parameter_75" + original_name = "linear_74.b_0" shape = [1536] dtype = "float32" min_val = float("-2.84464") @@ -832,6 +908,7 @@ class Program_weight_tensor_parameter_75: class Program_weight_tensor_parameter_76: name = "parameter_76" + original_name = "linear_74.w_0" shape = [512, 1536] dtype = "float32" min_val = float("-0.30592") @@ -843,6 +920,7 @@ class Program_weight_tensor_parameter_76: class Program_weight_tensor_parameter_77: name = "parameter_77" + original_name = "layer_norm_39.b_0" shape = [512] dtype = "float32" min_val = float("-0.94841") @@ -854,6 +932,7 @@ class Program_weight_tensor_parameter_77: class Program_weight_tensor_parameter_78: name = "parameter_78" + original_name = "layer_norm_39.w_0" shape = [512] dtype = "float32" min_val = float("0.000829767") @@ -865,6 +944,7 @@ class Program_weight_tensor_parameter_78: class Program_weight_tensor_parameter_79: name = "parameter_79" + original_name = "linear_73.b_0" shape = [512] dtype = "float32" min_val = float("-1.00717") @@ -876,6 +956,7 @@ class Program_weight_tensor_parameter_79: class Program_weight_tensor_parameter_80: name = "parameter_80" + original_name = "linear_73.w_0" shape = [2048, 512] dtype = "float32" min_val = float("-0.470336") @@ -887,6 +968,7 @@ class Program_weight_tensor_parameter_80: class Program_weight_tensor_parameter_81: name = "parameter_81" + original_name = "linear_72.b_0" shape = [2048] dtype = "float32" min_val = float("-0.884727") @@ -898,6 +980,7 @@ class Program_weight_tensor_parameter_81: class Program_weight_tensor_parameter_82: name = "parameter_82" + original_name = "linear_72.w_0" shape = [512, 2048] dtype = "float32" min_val = float("-0.464842") @@ -909,6 +992,7 @@ class Program_weight_tensor_parameter_82: class Program_weight_tensor_parameter_83: name = "parameter_83" + original_name = "layer_norm_38.b_0" shape = [512] dtype = "float32" min_val = float("-1.60215") @@ -920,6 +1004,7 @@ class Program_weight_tensor_parameter_83: class Program_weight_tensor_parameter_84: name = "parameter_84" + original_name = "layer_norm_38.w_0" shape = [512] dtype = "float32" min_val = float("0.237596") @@ -931,6 +1016,7 @@ class Program_weight_tensor_parameter_84: class Program_weight_tensor_parameter_85: name = "parameter_85" + original_name = "linear_71.b_0" shape = [512] dtype = "float32" min_val = float("-1.11673") @@ -942,6 +1028,7 @@ class Program_weight_tensor_parameter_85: class Program_weight_tensor_parameter_86: name = "parameter_86" + original_name = "linear_71.w_0" shape = [512, 512] dtype = "float32" min_val = float("-0.592227") @@ -953,6 +1040,7 @@ class Program_weight_tensor_parameter_86: class Program_weight_tensor_parameter_87: name = "parameter_87" + original_name = "linear_70.b_0" shape = [1536] dtype = "float32" min_val = float("-2.50662") @@ -964,6 +1052,7 @@ class Program_weight_tensor_parameter_87: class Program_weight_tensor_parameter_88: name = "parameter_88" + original_name = "linear_70.w_0" shape = [512, 1536] dtype = "float32" min_val = float("-0.267111") @@ -975,6 +1064,7 @@ class Program_weight_tensor_parameter_88: class Program_weight_tensor_parameter_89: name = "parameter_89" + original_name = "layer_norm_37.b_0" shape = [512] dtype = "float32" min_val = float("-0.956097") @@ -986,6 +1076,7 @@ class Program_weight_tensor_parameter_89: class Program_weight_tensor_parameter_90: name = "parameter_90" + original_name = "layer_norm_37.w_0" shape = [512] dtype = "float32" min_val = float("-8.65305e-06") @@ -997,6 +1088,7 @@ class Program_weight_tensor_parameter_90: class Program_weight_tensor_parameter_91: name = "parameter_91" + original_name = "linear_69.b_0" shape = [512] dtype = "float32" min_val = float("-0.930186") @@ -1008,6 +1100,7 @@ class Program_weight_tensor_parameter_91: class Program_weight_tensor_parameter_92: name = "parameter_92" + original_name = "linear_69.w_0" shape = [2048, 512] dtype = "float32" min_val = float("-0.423554") @@ -1019,6 +1112,7 @@ class Program_weight_tensor_parameter_92: class Program_weight_tensor_parameter_93: name = "parameter_93" + original_name = "linear_68.b_0" shape = [2048] dtype = "float32" min_val = float("-0.935431") @@ -1030,6 +1124,7 @@ class Program_weight_tensor_parameter_93: class Program_weight_tensor_parameter_94: name = "parameter_94" + original_name = "linear_68.w_0" shape = [512, 2048] dtype = "float32" min_val = float("-0.559658") @@ -1041,6 +1136,7 @@ class Program_weight_tensor_parameter_94: class Program_weight_tensor_parameter_95: name = "parameter_95" + original_name = "layer_norm_36.b_0" shape = [512] dtype = "float32" min_val = float("-1.24264") @@ -1052,6 +1148,7 @@ class Program_weight_tensor_parameter_95: class Program_weight_tensor_parameter_96: name = "parameter_96" + original_name = "layer_norm_36.w_0" shape = [512] dtype = "float32" min_val = float("0.212479") @@ -1063,6 +1160,7 @@ class Program_weight_tensor_parameter_96: class Program_weight_tensor_parameter_97: name = "parameter_97" + original_name = "linear_67.b_0" shape = [512] dtype = "float32" min_val = float("-1.16708") @@ -1074,6 +1172,7 @@ class Program_weight_tensor_parameter_97: class Program_weight_tensor_parameter_98: name = "parameter_98" + original_name = "linear_67.w_0" shape = [512, 512] dtype = "float32" min_val = float("-1.01944") @@ -1085,6 +1184,7 @@ class Program_weight_tensor_parameter_98: class Program_weight_tensor_parameter_99: name = "parameter_99" + original_name = "linear_66.b_0" shape = [1536] dtype = "float32" min_val = float("-2.76412") @@ -1096,6 +1196,7 @@ class Program_weight_tensor_parameter_99: class Program_weight_tensor_parameter_100: name = "parameter_100" + original_name = "linear_66.w_0" shape = [512, 1536] dtype = "float32" min_val = float("-0.340762") @@ -1107,6 +1208,7 @@ class Program_weight_tensor_parameter_100: class Program_weight_tensor_parameter_101: name = "parameter_101" + original_name = "layer_norm_35.b_0" shape = [512] dtype = "float32" min_val = float("-1.19003") @@ -1118,6 +1220,7 @@ class Program_weight_tensor_parameter_101: class Program_weight_tensor_parameter_102: name = "parameter_102" + original_name = "layer_norm_35.w_0" shape = [512] dtype = "float32" min_val = float("-0.000454755") @@ -1129,6 +1232,7 @@ class Program_weight_tensor_parameter_102: class Program_weight_tensor_parameter_103: name = "parameter_103" + original_name = "linear_65.b_0" shape = [512] dtype = "float32" min_val = float("-0.866029") @@ -1140,6 +1244,7 @@ class Program_weight_tensor_parameter_103: class Program_weight_tensor_parameter_104: name = "parameter_104" + original_name = "linear_65.w_0" shape = [2048, 512] dtype = "float32" min_val = float("-0.432976") @@ -1151,6 +1256,7 @@ class Program_weight_tensor_parameter_104: class Program_weight_tensor_parameter_105: name = "parameter_105" + original_name = "linear_64.b_0" shape = [2048] dtype = "float32" min_val = float("-0.900483") @@ -1162,6 +1268,7 @@ class Program_weight_tensor_parameter_105: class Program_weight_tensor_parameter_106: name = "parameter_106" + original_name = "linear_64.w_0" shape = [512, 2048] dtype = "float32" min_val = float("-0.400918") @@ -1173,6 +1280,7 @@ class Program_weight_tensor_parameter_106: class Program_weight_tensor_parameter_107: name = "parameter_107" + original_name = "layer_norm_34.b_0" shape = [512] dtype = "float32" min_val = float("-1.01327") @@ -1184,6 +1292,7 @@ class Program_weight_tensor_parameter_107: class Program_weight_tensor_parameter_108: name = "parameter_108" + original_name = "layer_norm_34.w_0" shape = [512] dtype = "float32" min_val = float("0.0112991") @@ -1195,6 +1304,7 @@ class Program_weight_tensor_parameter_108: class Program_weight_tensor_parameter_109: name = "parameter_109" + original_name = "linear_63.b_0" shape = [512] dtype = "float32" min_val = float("-0.568835") @@ -1206,6 +1316,7 @@ class Program_weight_tensor_parameter_109: class Program_weight_tensor_parameter_110: name = "parameter_110" + original_name = "linear_63.w_0" shape = [512, 512] dtype = "float32" min_val = float("-0.578201") @@ -1217,6 +1328,7 @@ class Program_weight_tensor_parameter_110: class Program_weight_tensor_parameter_111: name = "parameter_111" + original_name = "linear_62.b_0" shape = [1536] dtype = "float32" min_val = float("-2.77626") @@ -1228,6 +1340,7 @@ class Program_weight_tensor_parameter_111: class Program_weight_tensor_parameter_112: name = "parameter_112" + original_name = "linear_62.w_0" shape = [512, 1536] dtype = "float32" min_val = float("-0.304097") @@ -1239,6 +1352,7 @@ class Program_weight_tensor_parameter_112: class Program_weight_tensor_parameter_113: name = "parameter_113" + original_name = "layer_norm_33.b_0" shape = [512] dtype = "float32" min_val = float("-1.23827") @@ -1250,6 +1364,7 @@ class Program_weight_tensor_parameter_113: class Program_weight_tensor_parameter_114: name = "parameter_114" + original_name = "layer_norm_33.w_0" shape = [512] dtype = "float32" min_val = float("0.113256") @@ -1261,6 +1376,7 @@ class Program_weight_tensor_parameter_114: class Program_weight_tensor_parameter_115: name = "parameter_115" + original_name = "linear_61.b_0" shape = [512] dtype = "float32" min_val = float("-0.987302") @@ -1272,6 +1388,7 @@ class Program_weight_tensor_parameter_115: class Program_weight_tensor_parameter_116: name = "parameter_116" + original_name = "linear_61.w_0" shape = [2048, 512] dtype = "float32" min_val = float("-0.512766") @@ -1283,6 +1400,7 @@ class Program_weight_tensor_parameter_116: class Program_weight_tensor_parameter_117: name = "parameter_117" + original_name = "linear_60.b_0" shape = [2048] dtype = "float32" min_val = float("-1.0441") @@ -1294,6 +1412,7 @@ class Program_weight_tensor_parameter_117: class Program_weight_tensor_parameter_118: name = "parameter_118" + original_name = "linear_60.w_0" shape = [512, 2048] dtype = "float32" min_val = float("-0.374075") @@ -1305,6 +1424,7 @@ class Program_weight_tensor_parameter_118: class Program_weight_tensor_parameter_119: name = "parameter_119" + original_name = "layer_norm_32.b_0" shape = [512] dtype = "float32" min_val = float("-1.1091") @@ -1316,6 +1436,7 @@ class Program_weight_tensor_parameter_119: class Program_weight_tensor_parameter_120: name = "parameter_120" + original_name = "layer_norm_32.w_0" shape = [512] dtype = "float32" min_val = float("0.0403643") @@ -1327,6 +1448,7 @@ class Program_weight_tensor_parameter_120: class Program_weight_tensor_parameter_121: name = "parameter_121" + original_name = "linear_59.b_0" shape = [512] dtype = "float32" min_val = float("-0.961821") @@ -1338,6 +1460,7 @@ class Program_weight_tensor_parameter_121: class Program_weight_tensor_parameter_122: name = "parameter_122" + original_name = "linear_59.w_0" shape = [512, 512] dtype = "float32" min_val = float("-0.458651") @@ -1349,6 +1472,7 @@ class Program_weight_tensor_parameter_122: class Program_weight_tensor_parameter_123: name = "parameter_123" + original_name = "linear_58.b_0" shape = [1536] dtype = "float32" min_val = float("-2.94872") @@ -1360,6 +1484,7 @@ class Program_weight_tensor_parameter_123: class Program_weight_tensor_parameter_124: name = "parameter_124" + original_name = "linear_58.w_0" shape = [512, 1536] dtype = "float32" min_val = float("-0.298488") @@ -1371,6 +1496,7 @@ class Program_weight_tensor_parameter_124: class Program_weight_tensor_parameter_125: name = "parameter_125" + original_name = "layer_norm_31.b_0" shape = [512] dtype = "float32" min_val = float("-1.681") @@ -1382,6 +1508,7 @@ class Program_weight_tensor_parameter_125: class Program_weight_tensor_parameter_126: name = "parameter_126" + original_name = "layer_norm_31.w_0" shape = [512] dtype = "float32" min_val = float("0.197773") @@ -1393,6 +1520,7 @@ class Program_weight_tensor_parameter_126: class Program_weight_tensor_parameter_127: name = "parameter_127" + original_name = "linear_57.b_0" shape = [512] dtype = "float32" min_val = float("-1.16849") @@ -1404,6 +1532,7 @@ class Program_weight_tensor_parameter_127: class Program_weight_tensor_parameter_128: name = "parameter_128" + original_name = "linear_57.w_0" shape = [2048, 512] dtype = "float32" min_val = float("-0.502254") @@ -1415,6 +1544,7 @@ class Program_weight_tensor_parameter_128: class Program_weight_tensor_parameter_129: name = "parameter_129" + original_name = "linear_56.b_0" shape = [2048] dtype = "float32" min_val = float("-0.923182") @@ -1426,6 +1556,7 @@ class Program_weight_tensor_parameter_129: class Program_weight_tensor_parameter_130: name = "parameter_130" + original_name = "linear_56.w_0" shape = [512, 2048] dtype = "float32" min_val = float("-0.962015") @@ -1437,6 +1568,7 @@ class Program_weight_tensor_parameter_130: class Program_weight_tensor_parameter_131: name = "parameter_131" + original_name = "layer_norm_30.b_0" shape = [512] dtype = "float32" min_val = float("-1.17428") @@ -1448,6 +1580,7 @@ class Program_weight_tensor_parameter_131: class Program_weight_tensor_parameter_132: name = "parameter_132" + original_name = "layer_norm_30.w_0" shape = [512] dtype = "float32" min_val = float("0.27494") @@ -1459,6 +1592,7 @@ class Program_weight_tensor_parameter_132: class Program_weight_tensor_parameter_133: name = "parameter_133" + original_name = "linear_55.b_0" shape = [512] dtype = "float32" min_val = float("-0.591835") @@ -1470,6 +1604,7 @@ class Program_weight_tensor_parameter_133: class Program_weight_tensor_parameter_134: name = "parameter_134" + original_name = "linear_55.w_0" shape = [512, 512] dtype = "float32" min_val = float("-0.42562") @@ -1481,6 +1616,7 @@ class Program_weight_tensor_parameter_134: class Program_weight_tensor_parameter_135: name = "parameter_135" + original_name = "linear_54.b_0" shape = [1536] dtype = "float32" min_val = float("-2.57656") @@ -1492,6 +1628,7 @@ class Program_weight_tensor_parameter_135: class Program_weight_tensor_parameter_136: name = "parameter_136" + original_name = "linear_54.w_0" shape = [512, 1536] dtype = "float32" min_val = float("-0.304571") @@ -1503,6 +1640,7 @@ class Program_weight_tensor_parameter_136: class Program_weight_tensor_parameter_137: name = "parameter_137" + original_name = "layer_norm_29.b_0" shape = [512] dtype = "float32" min_val = float("-1.94558") @@ -1514,6 +1652,7 @@ class Program_weight_tensor_parameter_137: class Program_weight_tensor_parameter_138: name = "parameter_138" + original_name = "layer_norm_29.w_0" shape = [512] dtype = "float32" min_val = float("0.215969") @@ -1525,6 +1664,7 @@ class Program_weight_tensor_parameter_138: class Program_weight_tensor_parameter_139: name = "parameter_139" + original_name = "linear_53.b_0" shape = [512] dtype = "float32" min_val = float("-1.48821") @@ -1536,6 +1676,7 @@ class Program_weight_tensor_parameter_139: class Program_weight_tensor_parameter_140: name = "parameter_140" + original_name = "linear_53.w_0" shape = [2048, 512] dtype = "float32" min_val = float("-0.599004") @@ -1547,6 +1688,7 @@ class Program_weight_tensor_parameter_140: class Program_weight_tensor_parameter_141: name = "parameter_141" + original_name = "linear_52.b_0" shape = [2048] dtype = "float32" min_val = float("-1.00643") @@ -1558,6 +1700,7 @@ class Program_weight_tensor_parameter_141: class Program_weight_tensor_parameter_142: name = "parameter_142" + original_name = "linear_52.w_0" shape = [512, 2048] dtype = "float32" min_val = float("-0.704527") @@ -1569,6 +1712,7 @@ class Program_weight_tensor_parameter_142: class Program_weight_tensor_parameter_143: name = "parameter_143" + original_name = "layer_norm_28.b_0" shape = [512] dtype = "float32" min_val = float("-1.19963") @@ -1580,6 +1724,7 @@ class Program_weight_tensor_parameter_143: class Program_weight_tensor_parameter_144: name = "parameter_144" + original_name = "layer_norm_28.w_0" shape = [512] dtype = "float32" min_val = float("0.368851") @@ -1591,6 +1736,7 @@ class Program_weight_tensor_parameter_144: class Program_weight_tensor_parameter_145: name = "parameter_145" + original_name = "linear_51.b_0" shape = [512] dtype = "float32" min_val = float("-0.860647") @@ -1602,6 +1748,7 @@ class Program_weight_tensor_parameter_145: class Program_weight_tensor_parameter_146: name = "parameter_146" + original_name = "linear_51.w_0" shape = [512, 512] dtype = "float32" min_val = float("-0.41629") @@ -1613,6 +1760,7 @@ class Program_weight_tensor_parameter_146: class Program_weight_tensor_parameter_147: name = "parameter_147" + original_name = "linear_50.b_0" shape = [1536] dtype = "float32" min_val = float("-2.91229") @@ -1624,6 +1772,7 @@ class Program_weight_tensor_parameter_147: class Program_weight_tensor_parameter_148: name = "parameter_148" + original_name = "linear_50.w_0" shape = [512, 1536] dtype = "float32" min_val = float("-0.349808") @@ -1635,6 +1784,7 @@ class Program_weight_tensor_parameter_148: class Program_weight_tensor_parameter_149: name = "parameter_149" + original_name = "layer_norm_27.b_0" shape = [512] dtype = "float32" min_val = float("-1.82153") @@ -1646,6 +1796,7 @@ class Program_weight_tensor_parameter_149: class Program_weight_tensor_parameter_150: name = "parameter_150" + original_name = "layer_norm_27.w_0" shape = [512] dtype = "float32" min_val = float("0.253831") @@ -1657,6 +1808,7 @@ class Program_weight_tensor_parameter_150: class Program_weight_tensor_parameter_151: name = "parameter_151" + original_name = "linear_49.b_0" shape = [512] dtype = "float32" min_val = float("-1.39136") @@ -1668,6 +1820,7 @@ class Program_weight_tensor_parameter_151: class Program_weight_tensor_parameter_152: name = "parameter_152" + original_name = "linear_49.w_0" shape = [2048, 512] dtype = "float32" min_val = float("-0.596602") @@ -1679,6 +1832,7 @@ class Program_weight_tensor_parameter_152: class Program_weight_tensor_parameter_153: name = "parameter_153" + original_name = "linear_48.b_0" shape = [2048] dtype = "float32" min_val = float("-1.14342") @@ -1690,6 +1844,7 @@ class Program_weight_tensor_parameter_153: class Program_weight_tensor_parameter_154: name = "parameter_154" + original_name = "linear_48.w_0" shape = [512, 2048] dtype = "float32" min_val = float("-0.599486") @@ -1701,6 +1856,7 @@ class Program_weight_tensor_parameter_154: class Program_weight_tensor_parameter_155: name = "parameter_155" + original_name = "layer_norm_26.b_0" shape = [512] dtype = "float32" min_val = float("-1.14258") @@ -1712,6 +1868,7 @@ class Program_weight_tensor_parameter_155: class Program_weight_tensor_parameter_156: name = "parameter_156" + original_name = "layer_norm_26.w_0" shape = [512] dtype = "float32" min_val = float("0.443705") @@ -1723,6 +1880,7 @@ class Program_weight_tensor_parameter_156: class Program_weight_tensor_parameter_157: name = "parameter_157" + original_name = "linear_47.b_0" shape = [512] dtype = "float32" min_val = float("-0.573817") @@ -1734,6 +1892,7 @@ class Program_weight_tensor_parameter_157: class Program_weight_tensor_parameter_158: name = "parameter_158" + original_name = "linear_47.w_0" shape = [512, 512] dtype = "float32" min_val = float("-0.386993") @@ -1745,6 +1904,7 @@ class Program_weight_tensor_parameter_158: class Program_weight_tensor_parameter_159: name = "parameter_159" + original_name = "linear_46.b_0" shape = [1536] dtype = "float32" min_val = float("-2.72148") @@ -1756,6 +1916,7 @@ class Program_weight_tensor_parameter_159: class Program_weight_tensor_parameter_160: name = "parameter_160" + original_name = "linear_46.w_0" shape = [512, 1536] dtype = "float32" min_val = float("-0.307471") @@ -1767,6 +1928,7 @@ class Program_weight_tensor_parameter_160: class Program_weight_tensor_parameter_161: name = "parameter_161" + original_name = "layer_norm_25.b_0" shape = [512] dtype = "float32" min_val = float("-1.76529") @@ -1778,6 +1940,7 @@ class Program_weight_tensor_parameter_161: class Program_weight_tensor_parameter_162: name = "parameter_162" + original_name = "layer_norm_25.w_0" shape = [512] dtype = "float32" min_val = float("0.237231") @@ -1789,6 +1952,7 @@ class Program_weight_tensor_parameter_162: class Program_weight_tensor_parameter_163: name = "parameter_163" + original_name = "linear_45.b_0" shape = [512] dtype = "float32" min_val = float("-1.45125") @@ -1800,6 +1964,7 @@ class Program_weight_tensor_parameter_163: class Program_weight_tensor_parameter_164: name = "parameter_164" + original_name = "linear_45.w_0" shape = [2048, 512] dtype = "float32" min_val = float("-0.531308") @@ -1811,6 +1976,7 @@ class Program_weight_tensor_parameter_164: class Program_weight_tensor_parameter_165: name = "parameter_165" + original_name = "linear_44.b_0" shape = [2048] dtype = "float32" min_val = float("-1.2324") @@ -1822,6 +1988,7 @@ class Program_weight_tensor_parameter_165: class Program_weight_tensor_parameter_166: name = "parameter_166" + original_name = "linear_44.w_0" shape = [512, 2048] dtype = "float32" min_val = float("-0.538401") @@ -1833,6 +2000,7 @@ class Program_weight_tensor_parameter_166: class Program_weight_tensor_parameter_167: name = "parameter_167" + original_name = "layer_norm_24.b_0" shape = [512] dtype = "float32" min_val = float("-1.12218") @@ -1844,6 +2012,7 @@ class Program_weight_tensor_parameter_167: class Program_weight_tensor_parameter_168: name = "parameter_168" + original_name = "layer_norm_24.w_0" shape = [512] dtype = "float32" min_val = float("0.400276") @@ -1855,6 +2024,7 @@ class Program_weight_tensor_parameter_168: class Program_weight_tensor_parameter_169: name = "parameter_169" + original_name = "linear_43.b_0" shape = [512] dtype = "float32" min_val = float("-0.779718") @@ -1866,6 +2036,7 @@ class Program_weight_tensor_parameter_169: class Program_weight_tensor_parameter_170: name = "parameter_170" + original_name = "linear_43.w_0" shape = [512, 512] dtype = "float32" min_val = float("-0.506302") @@ -1877,6 +2048,7 @@ class Program_weight_tensor_parameter_170: class Program_weight_tensor_parameter_171: name = "parameter_171" + original_name = "linear_42.b_0" shape = [1536] dtype = "float32" min_val = float("-2.76195") @@ -1888,6 +2060,7 @@ class Program_weight_tensor_parameter_171: class Program_weight_tensor_parameter_172: name = "parameter_172" + original_name = "linear_42.w_0" shape = [512, 1536] dtype = "float32" min_val = float("-0.427909") @@ -1899,6 +2072,7 @@ class Program_weight_tensor_parameter_172: class Program_weight_tensor_parameter_173: name = "parameter_173" + original_name = "layer_norm_23.b_0" shape = [512] dtype = "float32" min_val = float("-1.636") @@ -1910,6 +2084,7 @@ class Program_weight_tensor_parameter_173: class Program_weight_tensor_parameter_174: name = "parameter_174" + original_name = "layer_norm_23.w_0" shape = [512] dtype = "float32" min_val = float("0.191401") @@ -1921,6 +2096,7 @@ class Program_weight_tensor_parameter_174: class Program_weight_tensor_parameter_175: name = "parameter_175" + original_name = "linear_41.b_0" shape = [512] dtype = "float32" min_val = float("-0.608384") @@ -1932,6 +2108,7 @@ class Program_weight_tensor_parameter_175: class Program_weight_tensor_parameter_176: name = "parameter_176" + original_name = "linear_41.w_0" shape = [2048, 512] dtype = "float32" min_val = float("-0.391701") @@ -1943,6 +2120,7 @@ class Program_weight_tensor_parameter_176: class Program_weight_tensor_parameter_177: name = "parameter_177" + original_name = "linear_40.b_0" shape = [2048] dtype = "float32" min_val = float("-1.27122") @@ -1954,6 +2132,7 @@ class Program_weight_tensor_parameter_177: class Program_weight_tensor_parameter_178: name = "parameter_178" + original_name = "linear_40.w_0" shape = [512, 2048] dtype = "float32" min_val = float("-0.486424") @@ -1965,6 +2144,7 @@ class Program_weight_tensor_parameter_178: class Program_weight_tensor_parameter_179: name = "parameter_179" + original_name = "layer_norm_22.b_0" shape = [512] dtype = "float32" min_val = float("-1.04588") @@ -1976,6 +2156,7 @@ class Program_weight_tensor_parameter_179: class Program_weight_tensor_parameter_180: name = "parameter_180" + original_name = "layer_norm_22.w_0" shape = [512] dtype = "float32" min_val = float("0.367185") @@ -1987,6 +2168,7 @@ class Program_weight_tensor_parameter_180: class Program_weight_tensor_parameter_181: name = "parameter_181" + original_name = "linear_39.b_0" shape = [512] dtype = "float32" min_val = float("-0.549837") @@ -1998,6 +2180,7 @@ class Program_weight_tensor_parameter_181: class Program_weight_tensor_parameter_182: name = "parameter_182" + original_name = "linear_39.w_0" shape = [512, 512] dtype = "float32" min_val = float("-0.408881") @@ -2009,6 +2192,7 @@ class Program_weight_tensor_parameter_182: class Program_weight_tensor_parameter_183: name = "parameter_183" + original_name = "linear_38.b_0" shape = [1536] dtype = "float32" min_val = float("-2.02407") @@ -2020,6 +2204,7 @@ class Program_weight_tensor_parameter_183: class Program_weight_tensor_parameter_184: name = "parameter_184" + original_name = "linear_38.w_0" shape = [512, 1536] dtype = "float32" min_val = float("-0.465174") @@ -2031,6 +2216,7 @@ class Program_weight_tensor_parameter_184: class Program_weight_tensor_parameter_185: name = "parameter_185" + original_name = "layer_norm_21.b_0" shape = [512] dtype = "float32" min_val = float("-1.50078") @@ -2042,6 +2228,7 @@ class Program_weight_tensor_parameter_185: class Program_weight_tensor_parameter_186: name = "parameter_186" + original_name = "layer_norm_21.w_0" shape = [512] dtype = "float32" min_val = float("0.174192") @@ -2053,6 +2240,7 @@ class Program_weight_tensor_parameter_186: class Program_weight_tensor_parameter_187: name = "parameter_187" + original_name = "linear_37.b_0" shape = [512] dtype = "float32" min_val = float("-0.669539") @@ -2064,6 +2252,7 @@ class Program_weight_tensor_parameter_187: class Program_weight_tensor_parameter_188: name = "parameter_188" + original_name = "linear_37.w_0" shape = [2048, 512] dtype = "float32" min_val = float("-0.446392") @@ -2075,6 +2264,7 @@ class Program_weight_tensor_parameter_188: class Program_weight_tensor_parameter_189: name = "parameter_189" + original_name = "linear_36.b_0" shape = [2048] dtype = "float32" min_val = float("-1.34166") @@ -2086,6 +2276,7 @@ class Program_weight_tensor_parameter_189: class Program_weight_tensor_parameter_190: name = "parameter_190" + original_name = "linear_36.w_0" shape = [512, 2048] dtype = "float32" min_val = float("-0.483534") @@ -2097,6 +2288,7 @@ class Program_weight_tensor_parameter_190: class Program_weight_tensor_parameter_191: name = "parameter_191" + original_name = "layer_norm_20.b_0" shape = [512] dtype = "float32" min_val = float("-1.12218") @@ -2108,6 +2300,7 @@ class Program_weight_tensor_parameter_191: class Program_weight_tensor_parameter_192: name = "parameter_192" + original_name = "layer_norm_20.w_0" shape = [512] dtype = "float32" min_val = float("0.276464") @@ -2119,6 +2312,7 @@ class Program_weight_tensor_parameter_192: class Program_weight_tensor_parameter_193: name = "parameter_193" + original_name = "linear_35.b_0" shape = [512] dtype = "float32" min_val = float("-0.570258") @@ -2130,6 +2324,7 @@ class Program_weight_tensor_parameter_193: class Program_weight_tensor_parameter_194: name = "parameter_194" + original_name = "linear_35.w_0" shape = [512, 512] dtype = "float32" min_val = float("-0.31601") @@ -2141,6 +2336,7 @@ class Program_weight_tensor_parameter_194: class Program_weight_tensor_parameter_195: name = "parameter_195" + original_name = "linear_34.b_0" shape = [1536] dtype = "float32" min_val = float("-1.85033") @@ -2152,6 +2348,7 @@ class Program_weight_tensor_parameter_195: class Program_weight_tensor_parameter_196: name = "parameter_196" + original_name = "linear_34.w_0" shape = [512, 1536] dtype = "float32" min_val = float("-0.473473") @@ -2163,6 +2360,7 @@ class Program_weight_tensor_parameter_196: class Program_weight_tensor_parameter_197: name = "parameter_197" + original_name = "layer_norm_19.b_0" shape = [512] dtype = "float32" min_val = float("-1.05393") @@ -2174,6 +2372,7 @@ class Program_weight_tensor_parameter_197: class Program_weight_tensor_parameter_198: name = "parameter_198" + original_name = "layer_norm_19.w_0" shape = [512] dtype = "float32" min_val = float("0.12018") @@ -2185,6 +2384,7 @@ class Program_weight_tensor_parameter_198: class Program_weight_tensor_parameter_199: name = "parameter_199" + original_name = "linear_33.b_0" shape = [512] dtype = "float32" min_val = float("-0.613383") @@ -2196,6 +2396,7 @@ class Program_weight_tensor_parameter_199: class Program_weight_tensor_parameter_200: name = "parameter_200" + original_name = "linear_33.w_0" shape = [2048, 512] dtype = "float32" min_val = float("-0.539175") @@ -2207,6 +2408,7 @@ class Program_weight_tensor_parameter_200: class Program_weight_tensor_parameter_201: name = "parameter_201" + original_name = "linear_32.b_0" shape = [2048] dtype = "float32" min_val = float("-1.38495") @@ -2218,6 +2420,7 @@ class Program_weight_tensor_parameter_201: class Program_weight_tensor_parameter_202: name = "parameter_202" + original_name = "linear_32.w_0" shape = [512, 2048] dtype = "float32" min_val = float("-0.569038") @@ -2229,6 +2432,7 @@ class Program_weight_tensor_parameter_202: class Program_weight_tensor_parameter_203: name = "parameter_203" + original_name = "layer_norm_18.b_0" shape = [512] dtype = "float32" min_val = float("-1.17571") @@ -2240,6 +2444,7 @@ class Program_weight_tensor_parameter_203: class Program_weight_tensor_parameter_204: name = "parameter_204" + original_name = "layer_norm_18.w_0" shape = [512] dtype = "float32" min_val = float("0.196253") @@ -2251,6 +2456,7 @@ class Program_weight_tensor_parameter_204: class Program_weight_tensor_parameter_205: name = "parameter_205" + original_name = "linear_31.b_0" shape = [512] dtype = "float32" min_val = float("-0.574377") @@ -2262,6 +2468,7 @@ class Program_weight_tensor_parameter_205: class Program_weight_tensor_parameter_206: name = "parameter_206" + original_name = "linear_31.w_0" shape = [512, 512] dtype = "float32" min_val = float("-0.29636") @@ -2273,6 +2480,7 @@ class Program_weight_tensor_parameter_206: class Program_weight_tensor_parameter_207: name = "parameter_207" + original_name = "linear_30.b_0" shape = [1536] dtype = "float32" min_val = float("-2.432") @@ -2284,6 +2492,7 @@ class Program_weight_tensor_parameter_207: class Program_weight_tensor_parameter_208: name = "parameter_208" + original_name = "linear_30.w_0" shape = [512, 1536] dtype = "float32" min_val = float("-0.484242") @@ -2295,6 +2504,7 @@ class Program_weight_tensor_parameter_208: class Program_weight_tensor_parameter_209: name = "parameter_209" + original_name = "layer_norm_17.b_0" shape = [512] dtype = "float32" min_val = float("-0.98432") @@ -2306,6 +2516,7 @@ class Program_weight_tensor_parameter_209: class Program_weight_tensor_parameter_210: name = "parameter_210" + original_name = "layer_norm_17.w_0" shape = [512] dtype = "float32" min_val = float("0.139695") @@ -2317,6 +2528,7 @@ class Program_weight_tensor_parameter_210: class Program_weight_tensor_parameter_211: name = "parameter_211" + original_name = "linear_29.b_0" shape = [512] dtype = "float32" min_val = float("-0.747032") @@ -2328,6 +2540,7 @@ class Program_weight_tensor_parameter_211: class Program_weight_tensor_parameter_212: name = "parameter_212" + original_name = "linear_29.w_0" shape = [2048, 512] dtype = "float32" min_val = float("-0.469845") @@ -2339,6 +2552,7 @@ class Program_weight_tensor_parameter_212: class Program_weight_tensor_parameter_213: name = "parameter_213" + original_name = "linear_28.b_0" shape = [2048] dtype = "float32" min_val = float("-1.29449") @@ -2350,6 +2564,7 @@ class Program_weight_tensor_parameter_213: class Program_weight_tensor_parameter_214: name = "parameter_214" + original_name = "linear_28.w_0" shape = [512, 2048] dtype = "float32" min_val = float("-0.327133") @@ -2361,6 +2576,7 @@ class Program_weight_tensor_parameter_214: class Program_weight_tensor_parameter_215: name = "parameter_215" + original_name = "layer_norm_16.b_0" shape = [512] dtype = "float32" min_val = float("-1.32674") @@ -2372,6 +2588,7 @@ class Program_weight_tensor_parameter_215: class Program_weight_tensor_parameter_216: name = "parameter_216" + original_name = "layer_norm_16.w_0" shape = [512] dtype = "float32" min_val = float("0.00171956") @@ -2383,6 +2600,7 @@ class Program_weight_tensor_parameter_216: class Program_weight_tensor_parameter_217: name = "parameter_217" + original_name = "linear_27.b_0" shape = [512] dtype = "float32" min_val = float("-0.747817") @@ -2394,6 +2612,7 @@ class Program_weight_tensor_parameter_217: class Program_weight_tensor_parameter_218: name = "parameter_218" + original_name = "linear_27.w_0" shape = [512, 512] dtype = "float32" min_val = float("-0.344125") @@ -2405,6 +2624,7 @@ class Program_weight_tensor_parameter_218: class Program_weight_tensor_parameter_219: name = "parameter_219" + original_name = "linear_26.b_0" shape = [1536] dtype = "float32" min_val = float("-2.38855") @@ -2416,6 +2636,7 @@ class Program_weight_tensor_parameter_219: class Program_weight_tensor_parameter_220: name = "parameter_220" + original_name = "linear_26.w_0" shape = [512, 1536] dtype = "float32" min_val = float("-0.351298") @@ -2427,6 +2648,7 @@ class Program_weight_tensor_parameter_220: class Program_weight_tensor_parameter_221: name = "parameter_221" + original_name = "layer_norm_15.b_0" shape = [512] dtype = "float32" min_val = float("-0.99012") @@ -2438,6 +2660,7 @@ class Program_weight_tensor_parameter_221: class Program_weight_tensor_parameter_222: name = "parameter_222" + original_name = "layer_norm_15.w_0" shape = [512] dtype = "float32" min_val = float("0.00247525") @@ -2449,6 +2672,7 @@ class Program_weight_tensor_parameter_222: class Program_weight_tensor_parameter_223: name = "parameter_223" + original_name = "linear_25.b_0" shape = [512] dtype = "float32" min_val = float("-0.675325") @@ -2460,6 +2684,7 @@ class Program_weight_tensor_parameter_223: class Program_weight_tensor_parameter_224: name = "parameter_224" + original_name = "linear_25.w_0" shape = [2048, 512] dtype = "float32" min_val = float("-0.338762") @@ -2471,6 +2696,7 @@ class Program_weight_tensor_parameter_224: class Program_weight_tensor_parameter_225: name = "parameter_225" + original_name = "linear_24.b_0" shape = [2048] dtype = "float32" min_val = float("-1.41262") @@ -2482,6 +2708,7 @@ class Program_weight_tensor_parameter_225: class Program_weight_tensor_parameter_226: name = "parameter_226" + original_name = "linear_24.w_0" shape = [512, 2048] dtype = "float32" min_val = float("-0.293502") @@ -2493,6 +2720,7 @@ class Program_weight_tensor_parameter_226: class Program_weight_tensor_parameter_227: name = "parameter_227" + original_name = "layer_norm_14.b_0" shape = [512] dtype = "float32" min_val = float("-1.1875") @@ -2504,6 +2732,7 @@ class Program_weight_tensor_parameter_227: class Program_weight_tensor_parameter_228: name = "parameter_228" + original_name = "layer_norm_14.w_0" shape = [512] dtype = "float32" min_val = float("0.00124824") @@ -2515,6 +2744,7 @@ class Program_weight_tensor_parameter_228: class Program_weight_tensor_parameter_229: name = "parameter_229" + original_name = "linear_23.b_0" shape = [512] dtype = "float32" min_val = float("-0.54819") @@ -2526,6 +2756,7 @@ class Program_weight_tensor_parameter_229: class Program_weight_tensor_parameter_230: name = "parameter_230" + original_name = "linear_23.w_0" shape = [512, 512] dtype = "float32" min_val = float("-0.295063") @@ -2537,6 +2768,7 @@ class Program_weight_tensor_parameter_230: class Program_weight_tensor_parameter_231: name = "parameter_231" + original_name = "linear_22.b_0" shape = [1536] dtype = "float32" min_val = float("-2.58799") @@ -2548,6 +2780,7 @@ class Program_weight_tensor_parameter_231: class Program_weight_tensor_parameter_232: name = "parameter_232" + original_name = "linear_22.w_0" shape = [512, 1536] dtype = "float32" min_val = float("-0.506295") @@ -2559,6 +2792,7 @@ class Program_weight_tensor_parameter_232: class Program_weight_tensor_parameter_233: name = "parameter_233" + original_name = "layer_norm_13.b_0" shape = [512] dtype = "float32" min_val = float("-0.961602") @@ -2570,6 +2804,7 @@ class Program_weight_tensor_parameter_233: class Program_weight_tensor_parameter_234: name = "parameter_234" + original_name = "layer_norm_13.w_0" shape = [512] dtype = "float32" min_val = float("0.000444557") @@ -2581,6 +2816,7 @@ class Program_weight_tensor_parameter_234: class Program_weight_tensor_parameter_235: name = "parameter_235" + original_name = "linear_21.b_0" shape = [512] dtype = "float32" min_val = float("-0.737795") @@ -2592,6 +2828,7 @@ class Program_weight_tensor_parameter_235: class Program_weight_tensor_parameter_236: name = "parameter_236" + original_name = "linear_21.w_0" shape = [2048, 512] dtype = "float32" min_val = float("-0.434825") @@ -2603,6 +2840,7 @@ class Program_weight_tensor_parameter_236: class Program_weight_tensor_parameter_237: name = "parameter_237" + original_name = "linear_20.b_0" shape = [2048] dtype = "float32" min_val = float("-1.71279") @@ -2614,6 +2852,7 @@ class Program_weight_tensor_parameter_237: class Program_weight_tensor_parameter_238: name = "parameter_238" + original_name = "linear_20.w_0" shape = [512, 2048] dtype = "float32" min_val = float("-0.349905") @@ -2625,6 +2864,7 @@ class Program_weight_tensor_parameter_238: class Program_weight_tensor_parameter_239: name = "parameter_239" + original_name = "layer_norm_12.b_0" shape = [512] dtype = "float32" min_val = float("-1.01503") @@ -2636,6 +2876,7 @@ class Program_weight_tensor_parameter_239: class Program_weight_tensor_parameter_240: name = "parameter_240" + original_name = "layer_norm_12.w_0" shape = [512] dtype = "float32" min_val = float("-0.00309318") @@ -2647,6 +2888,7 @@ class Program_weight_tensor_parameter_240: class Program_weight_tensor_parameter_241: name = "parameter_241" + original_name = "linear_19.b_0" shape = [512] dtype = "float32" min_val = float("-0.821854") @@ -2658,6 +2900,7 @@ class Program_weight_tensor_parameter_241: class Program_weight_tensor_parameter_242: name = "parameter_242" + original_name = "linear_19.w_0" shape = [512, 512] dtype = "float32" min_val = float("-0.273471") @@ -2669,6 +2912,7 @@ class Program_weight_tensor_parameter_242: class Program_weight_tensor_parameter_243: name = "parameter_243" + original_name = "linear_18.b_0" shape = [1536] dtype = "float32" min_val = float("-2.68392") @@ -2680,6 +2924,7 @@ class Program_weight_tensor_parameter_243: class Program_weight_tensor_parameter_244: name = "parameter_244" + original_name = "linear_18.w_0" shape = [512, 1536] dtype = "float32" min_val = float("-0.388922") @@ -2691,6 +2936,7 @@ class Program_weight_tensor_parameter_244: class Program_weight_tensor_parameter_245: name = "parameter_245" + original_name = "layer_norm_11.b_0" shape = [512] dtype = "float32" min_val = float("-0.869193") @@ -2702,6 +2948,7 @@ class Program_weight_tensor_parameter_245: class Program_weight_tensor_parameter_246: name = "parameter_246" + original_name = "layer_norm_11.w_0" shape = [512] dtype = "float32" min_val = float("-0.00131842") @@ -2713,6 +2960,7 @@ class Program_weight_tensor_parameter_246: class Program_weight_tensor_parameter_247: name = "parameter_247" + original_name = "linear_17.w_0" shape = [1024, 512] dtype = "float32" min_val = float("-0.254725") @@ -2724,6 +2972,7 @@ class Program_weight_tensor_parameter_247: class Program_weight_tensor_parameter_248: name = "parameter_248" + original_name = "layer_norm_10.b_0" shape = [1024] dtype = "float32" min_val = float("-0.938848") @@ -2735,6 +2984,7 @@ class Program_weight_tensor_parameter_248: class Program_weight_tensor_parameter_249: name = "parameter_249" + original_name = "layer_norm_10.w_0" shape = [1024] dtype = "float32" min_val = float("0.151589") @@ -2746,6 +2996,7 @@ class Program_weight_tensor_parameter_249: class Program_weight_tensor_parameter_250: name = "parameter_250" + original_name = "linear_16.b_0" shape = [256] dtype = "float32" min_val = float("-2.98083") @@ -2757,6 +3008,7 @@ class Program_weight_tensor_parameter_250: class Program_weight_tensor_parameter_251: name = "parameter_251" + original_name = "linear_16.w_0" shape = [1024, 256] dtype = "float32" min_val = float("-0.628008") @@ -2768,6 +3020,7 @@ class Program_weight_tensor_parameter_251: class Program_weight_tensor_parameter_252: name = "parameter_252" + original_name = "linear_15.b_0" shape = [1024] dtype = "float32" min_val = float("-1.41386") @@ -2779,6 +3032,7 @@ class Program_weight_tensor_parameter_252: class Program_weight_tensor_parameter_253: name = "parameter_253" + original_name = "linear_15.w_0" shape = [256, 1024] dtype = "float32" min_val = float("-0.298711") @@ -2790,6 +3044,7 @@ class Program_weight_tensor_parameter_253: class Program_weight_tensor_parameter_254: name = "parameter_254" + original_name = "layer_norm_9.b_0" shape = [256] dtype = "float32" min_val = float("-1.05058") @@ -2801,6 +3056,7 @@ class Program_weight_tensor_parameter_254: class Program_weight_tensor_parameter_255: name = "parameter_255" + original_name = "layer_norm_9.w_0" shape = [256] dtype = "float32" min_val = float("0.239457") @@ -2812,6 +3068,7 @@ class Program_weight_tensor_parameter_255: class Program_weight_tensor_parameter_256: name = "parameter_256" + original_name = "linear_14.b_0" shape = [256] dtype = "float32" min_val = float("-2.04351") @@ -2823,6 +3080,7 @@ class Program_weight_tensor_parameter_256: class Program_weight_tensor_parameter_257: name = "parameter_257" + original_name = "linear_14.w_0" shape = [256, 256] dtype = "float32" min_val = float("-0.216044") @@ -2834,6 +3092,7 @@ class Program_weight_tensor_parameter_257: class Program_weight_tensor_parameter_258: name = "parameter_258" + original_name = "linear_13.b_0" shape = [768] dtype = "float32" min_val = float("-1.52453") @@ -2845,6 +3104,7 @@ class Program_weight_tensor_parameter_258: class Program_weight_tensor_parameter_259: name = "parameter_259" + original_name = "linear_13.w_0" shape = [256, 768] dtype = "float32" min_val = float("-0.459792") @@ -2856,6 +3116,7 @@ class Program_weight_tensor_parameter_259: class Program_weight_tensor_parameter_260: name = "parameter_260" + original_name = "layer_norm_8.b_0" shape = [256] dtype = "float32" min_val = float("-1.00334") @@ -2867,6 +3128,7 @@ class Program_weight_tensor_parameter_260: class Program_weight_tensor_parameter_261: name = "parameter_261" + original_name = "layer_norm_8.w_0" shape = [256] dtype = "float32" min_val = float("0.386137") @@ -2878,6 +3140,7 @@ class Program_weight_tensor_parameter_261: class Program_weight_tensor_parameter_262: name = "parameter_262" + original_name = "linear_12.b_0" shape = [256] dtype = "float32" min_val = float("-1.91259") @@ -2889,6 +3152,7 @@ class Program_weight_tensor_parameter_262: class Program_weight_tensor_parameter_263: name = "parameter_263" + original_name = "linear_12.w_0" shape = [1024, 256] dtype = "float32" min_val = float("-0.36279") @@ -2900,6 +3164,7 @@ class Program_weight_tensor_parameter_263: class Program_weight_tensor_parameter_264: name = "parameter_264" + original_name = "linear_11.b_0" shape = [1024] dtype = "float32" min_val = float("-1.47212") @@ -2911,6 +3176,7 @@ class Program_weight_tensor_parameter_264: class Program_weight_tensor_parameter_265: name = "parameter_265" + original_name = "linear_11.w_0" shape = [256, 1024] dtype = "float32" min_val = float("-0.51498") @@ -2922,6 +3188,7 @@ class Program_weight_tensor_parameter_265: class Program_weight_tensor_parameter_266: name = "parameter_266" + original_name = "layer_norm_7.b_0" shape = [256] dtype = "float32" min_val = float("-0.537193") @@ -2933,6 +3200,7 @@ class Program_weight_tensor_parameter_266: class Program_weight_tensor_parameter_267: name = "parameter_267" + original_name = "layer_norm_7.w_0" shape = [256] dtype = "float32" min_val = float("-0.00119768") @@ -2944,6 +3212,7 @@ class Program_weight_tensor_parameter_267: class Program_weight_tensor_parameter_268: name = "parameter_268" + original_name = "linear_10.b_0" shape = [256] dtype = "float32" min_val = float("-1.6185") @@ -2955,6 +3224,7 @@ class Program_weight_tensor_parameter_268: class Program_weight_tensor_parameter_269: name = "parameter_269" + original_name = "linear_10.w_0" shape = [256, 256] dtype = "float32" min_val = float("-0.294549") @@ -2966,6 +3236,7 @@ class Program_weight_tensor_parameter_269: class Program_weight_tensor_parameter_270: name = "parameter_270" + original_name = "linear_9.b_0" shape = [768] dtype = "float32" min_val = float("-1.83694") @@ -2977,6 +3248,7 @@ class Program_weight_tensor_parameter_270: class Program_weight_tensor_parameter_271: name = "parameter_271" + original_name = "linear_9.w_0" shape = [256, 768] dtype = "float32" min_val = float("-0.389052") @@ -2988,6 +3260,7 @@ class Program_weight_tensor_parameter_271: class Program_weight_tensor_parameter_272: name = "parameter_272" + original_name = "layer_norm_6.b_0" shape = [256] dtype = "float32" min_val = float("-0.343515") @@ -2999,6 +3272,7 @@ class Program_weight_tensor_parameter_272: class Program_weight_tensor_parameter_273: name = "parameter_273" + original_name = "layer_norm_6.w_0" shape = [256] dtype = "float32" min_val = float("0.16752") @@ -3010,6 +3284,7 @@ class Program_weight_tensor_parameter_273: class Program_weight_tensor_parameter_274: name = "parameter_274" + original_name = "linear_8.w_0" shape = [512, 256] dtype = "float32" min_val = float("-0.254516") @@ -3021,6 +3296,7 @@ class Program_weight_tensor_parameter_274: class Program_weight_tensor_parameter_275: name = "parameter_275" + original_name = "layer_norm_5.b_0" shape = [512] dtype = "float32" min_val = float("-1.39598") @@ -3032,6 +3308,7 @@ class Program_weight_tensor_parameter_275: class Program_weight_tensor_parameter_276: name = "parameter_276" + original_name = "layer_norm_5.w_0" shape = [512] dtype = "float32" min_val = float("0.218499") @@ -3043,6 +3320,7 @@ class Program_weight_tensor_parameter_276: class Program_weight_tensor_parameter_277: name = "parameter_277" + original_name = "linear_7.b_0" shape = [128] dtype = "float32" min_val = float("-1.56516") @@ -3054,6 +3332,7 @@ class Program_weight_tensor_parameter_277: class Program_weight_tensor_parameter_278: name = "parameter_278" + original_name = "linear_7.w_0" shape = [512, 128] dtype = "float32" min_val = float("-0.509274") @@ -3065,6 +3344,7 @@ class Program_weight_tensor_parameter_278: class Program_weight_tensor_parameter_279: name = "parameter_279" + original_name = "linear_6.b_0" shape = [512] dtype = "float32" min_val = float("-1.72128") @@ -3076,6 +3356,7 @@ class Program_weight_tensor_parameter_279: class Program_weight_tensor_parameter_280: name = "parameter_280" + original_name = "linear_6.w_0" shape = [128, 512] dtype = "float32" min_val = float("-0.367164") @@ -3087,6 +3368,7 @@ class Program_weight_tensor_parameter_280: class Program_weight_tensor_parameter_281: name = "parameter_281" + original_name = "layer_norm_4.b_0" shape = [128] dtype = "float32" min_val = float("-1.00559") @@ -3098,6 +3380,7 @@ class Program_weight_tensor_parameter_281: class Program_weight_tensor_parameter_282: name = "parameter_282" + original_name = "layer_norm_4.w_0" shape = [128] dtype = "float32" min_val = float("0.000973343") @@ -3109,6 +3392,7 @@ class Program_weight_tensor_parameter_282: class Program_weight_tensor_parameter_283: name = "parameter_283" + original_name = "linear_5.b_0" shape = [128] dtype = "float32" min_val = float("-0.702735") @@ -3120,6 +3404,7 @@ class Program_weight_tensor_parameter_283: class Program_weight_tensor_parameter_284: name = "parameter_284" + original_name = "linear_5.w_0" shape = [128, 128] dtype = "float32" min_val = float("-0.327238") @@ -3131,6 +3416,7 @@ class Program_weight_tensor_parameter_284: class Program_weight_tensor_parameter_285: name = "parameter_285" + original_name = "linear_4.b_0" shape = [384] dtype = "float32" min_val = float("-1.55252") @@ -3142,6 +3428,7 @@ class Program_weight_tensor_parameter_285: class Program_weight_tensor_parameter_286: name = "parameter_286" + original_name = "linear_4.w_0" shape = [128, 384] dtype = "float32" min_val = float("-0.414749") @@ -3153,6 +3440,7 @@ class Program_weight_tensor_parameter_286: class Program_weight_tensor_parameter_287: name = "parameter_287" + original_name = "layer_norm_3.b_0" shape = [128] dtype = "float32" min_val = float("-1.10819") @@ -3164,6 +3452,7 @@ class Program_weight_tensor_parameter_287: class Program_weight_tensor_parameter_288: name = "parameter_288" + original_name = "layer_norm_3.w_0" shape = [128] dtype = "float32" min_val = float("0.253696") @@ -3175,6 +3464,7 @@ class Program_weight_tensor_parameter_288: class Program_weight_tensor_parameter_289: name = "parameter_289" + original_name = "linear_3.b_0" shape = [128] dtype = "float32" min_val = float("-2.64049") @@ -3186,6 +3476,7 @@ class Program_weight_tensor_parameter_289: class Program_weight_tensor_parameter_290: name = "parameter_290" + original_name = "linear_3.w_0" shape = [512, 128] dtype = "float32" min_val = float("-0.314923") @@ -3197,6 +3488,7 @@ class Program_weight_tensor_parameter_290: class Program_weight_tensor_parameter_291: name = "parameter_291" + original_name = "linear_2.b_0" shape = [512] dtype = "float32" min_val = float("-1.45571") @@ -3208,6 +3500,7 @@ class Program_weight_tensor_parameter_291: class Program_weight_tensor_parameter_292: name = "parameter_292" + original_name = "linear_2.w_0" shape = [128, 512] dtype = "float32" min_val = float("-0.451241") @@ -3219,6 +3512,7 @@ class Program_weight_tensor_parameter_292: class Program_weight_tensor_parameter_293: name = "parameter_293" + original_name = "layer_norm_2.b_0" shape = [128] dtype = "float32" min_val = float("-1.28867") @@ -3230,6 +3524,7 @@ class Program_weight_tensor_parameter_293: class Program_weight_tensor_parameter_294: name = "parameter_294" + original_name = "layer_norm_2.w_0" shape = [128] dtype = "float32" min_val = float("0.357004") @@ -3241,6 +3536,7 @@ class Program_weight_tensor_parameter_294: class Program_weight_tensor_parameter_295: name = "parameter_295" + original_name = "linear_1.b_0" shape = [128] dtype = "float32" min_val = float("-0.85365") @@ -3252,6 +3548,7 @@ class Program_weight_tensor_parameter_295: class Program_weight_tensor_parameter_296: name = "parameter_296" + original_name = "linear_1.w_0" shape = [128, 128] dtype = "float32" min_val = float("-0.288925") @@ -3263,6 +3560,7 @@ class Program_weight_tensor_parameter_296: class Program_weight_tensor_parameter_297: name = "parameter_297" + original_name = "linear_0.b_0" shape = [384] dtype = "float32" min_val = float("-2.01378") @@ -3274,6 +3572,7 @@ class Program_weight_tensor_parameter_297: class Program_weight_tensor_parameter_298: name = "parameter_298" + original_name = "linear_0.w_0" shape = [128, 384] dtype = "float32" min_val = float("-0.513282") @@ -3285,6 +3584,7 @@ class Program_weight_tensor_parameter_298: class Program_weight_tensor_parameter_299: name = "parameter_299" + original_name = "layer_norm_1.b_0" shape = [128] dtype = "float32" min_val = float("-0.857635") @@ -3296,6 +3596,7 @@ class Program_weight_tensor_parameter_299: class Program_weight_tensor_parameter_300: name = "parameter_300" + original_name = "layer_norm_1.w_0" shape = [128] dtype = "float32" min_val = float("-0.000575064") @@ -3307,6 +3608,7 @@ class Program_weight_tensor_parameter_300: class Program_weight_tensor_parameter_301: name = "parameter_301" + original_name = "layer_norm_0.b_0" shape = [128] dtype = "float32" min_val = float("-1.25275") @@ -3318,6 +3620,7 @@ class Program_weight_tensor_parameter_301: class Program_weight_tensor_parameter_302: name = "parameter_302" + original_name = "layer_norm_0.w_0" shape = [128] dtype = "float32" min_val = float("-0.0179978") @@ -3329,6 +3632,7 @@ class Program_weight_tensor_parameter_302: class Program_weight_tensor_parameter_303: name = "parameter_303" + original_name = "conv2d_0.b_0" shape = [128] dtype = "float32" min_val = float("-1.14447") @@ -3340,6 +3644,7 @@ class Program_weight_tensor_parameter_303: class Program_weight_tensor_parameter_304: name = "parameter_304" + original_name = "conv2d_0.w_0" shape = [128, 3, 4, 4] dtype = "float32" min_val = float("-0.309677") diff --git a/paddle_samples/PaddleX/SwinTransformer_base_patch4_window7_224/subgraph_1/input_meta.py b/paddle_samples/PaddleX/SwinTransformer_base_patch4_window7_224/subgraph_1/input_meta.py index 9869273a0..3e77ba87a 100644 --- a/paddle_samples/PaddleX/SwinTransformer_base_patch4_window7_224/subgraph_1/input_meta.py +++ b/paddle_samples/PaddleX/SwinTransformer_base_patch4_window7_224/subgraph_1/input_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_data_0: name = "data_0" + original_name = "param_24" shape = [169, 4] dtype = "float32" min_val = float("-8.35157") @@ -11,6 +12,7 @@ class Program_weight_tensor_data_0: class Program_weight_tensor_data_1: name = "data_1" + original_name = "param_25" shape = [169, 4] dtype = "float32" min_val = float("-6.26919") @@ -22,6 +24,7 @@ class Program_weight_tensor_data_1: class Program_weight_tensor_data_2: name = "data_2" + original_name = "param_26" shape = [169, 8] dtype = "float32" min_val = float("-6.02229") @@ -33,6 +36,7 @@ class Program_weight_tensor_data_2: class Program_weight_tensor_data_3: name = "data_3" + original_name = "param_27" shape = [169, 8] dtype = "float32" min_val = float("-7.6461") @@ -44,6 +48,7 @@ class Program_weight_tensor_data_3: class Program_weight_tensor_data_4: name = "data_4" + original_name = "param_28" shape = [169, 16] dtype = "float32" min_val = float("-10.9588") @@ -55,6 +60,7 @@ class Program_weight_tensor_data_4: class Program_weight_tensor_data_5: name = "data_5" + original_name = "param_29" shape = [169, 16] dtype = "float32" min_val = float("-10.1747") @@ -66,6 +72,7 @@ class Program_weight_tensor_data_5: class Program_weight_tensor_data_6: name = "data_6" + original_name = "param_30" shape = [169, 16] dtype = "float32" min_val = float("-10.5637") @@ -77,6 +84,7 @@ class Program_weight_tensor_data_6: class Program_weight_tensor_data_7: name = "data_7" + original_name = "param_31" shape = [169, 16] dtype = "float32" min_val = float("-7.65857") @@ -88,6 +96,7 @@ class Program_weight_tensor_data_7: class Program_weight_tensor_data_8: name = "data_8" + original_name = "param_32" shape = [169, 16] dtype = "float32" min_val = float("-12.4205") @@ -99,6 +108,7 @@ class Program_weight_tensor_data_8: class Program_weight_tensor_data_9: name = "data_9" + original_name = "param_33" shape = [169, 16] dtype = "float32" min_val = float("-9.87814") @@ -110,6 +120,7 @@ class Program_weight_tensor_data_9: class Program_weight_tensor_data_10: name = "data_10" + original_name = "param_34" shape = [169, 16] dtype = "float32" min_val = float("-9.44169") @@ -121,6 +132,7 @@ class Program_weight_tensor_data_10: class Program_weight_tensor_data_11: name = "data_11" + original_name = "param_35" shape = [169, 16] dtype = "float32" min_val = float("-8.3709") @@ -132,6 +144,7 @@ class Program_weight_tensor_data_11: class Program_weight_tensor_data_12: name = "data_12" + original_name = "param_36" shape = [169, 16] dtype = "float32" min_val = float("-8.94788") @@ -143,6 +156,7 @@ class Program_weight_tensor_data_12: class Program_weight_tensor_data_13: name = "data_13" + original_name = "param_37" shape = [169, 16] dtype = "float32" min_val = float("-4.65554") @@ -154,6 +168,7 @@ class Program_weight_tensor_data_13: class Program_weight_tensor_data_14: name = "data_14" + original_name = "param_38" shape = [169, 16] dtype = "float32" min_val = float("-8.62036") @@ -165,6 +180,7 @@ class Program_weight_tensor_data_14: class Program_weight_tensor_data_15: name = "data_15" + original_name = "param_39" shape = [169, 16] dtype = "float32" min_val = float("-5.26244") @@ -176,6 +192,7 @@ class Program_weight_tensor_data_15: class Program_weight_tensor_data_16: name = "data_16" + original_name = "param_40" shape = [169, 16] dtype = "float32" min_val = float("-11.877") @@ -187,6 +204,7 @@ class Program_weight_tensor_data_16: class Program_weight_tensor_data_17: name = "data_17" + original_name = "param_41" shape = [169, 16] dtype = "float32" min_val = float("-8.8516") @@ -198,6 +216,7 @@ class Program_weight_tensor_data_17: class Program_weight_tensor_data_18: name = "data_18" + original_name = "param_42" shape = [169, 16] dtype = "float32" min_val = float("-10.4541") @@ -209,6 +228,7 @@ class Program_weight_tensor_data_18: class Program_weight_tensor_data_19: name = "data_19" + original_name = "param_43" shape = [169, 16] dtype = "float32" min_val = float("-10.8017") @@ -220,6 +240,7 @@ class Program_weight_tensor_data_19: class Program_weight_tensor_data_20: name = "data_20" + original_name = "param_44" shape = [169, 16] dtype = "float32" min_val = float("-14.7827") @@ -231,6 +252,7 @@ class Program_weight_tensor_data_20: class Program_weight_tensor_data_21: name = "data_21" + original_name = "param_45" shape = [169, 16] dtype = "float32" min_val = float("-14.5046") @@ -242,6 +264,7 @@ class Program_weight_tensor_data_21: class Program_weight_tensor_data_22: name = "data_22" + original_name = "param_46" shape = [169, 32] dtype = "float32" min_val = float("-15.6922") @@ -253,6 +276,7 @@ class Program_weight_tensor_data_22: class Program_weight_tensor_data_23: name = "data_23" + original_name = "param_47" shape = [169, 32] dtype = "float32" min_val = float("-15.7419") @@ -264,6 +288,7 @@ class Program_weight_tensor_data_23: class Program_weight_tensor_data_24: name = "data_24" + original_name = "var_1731" shape = [12, 3, 224, 224] dtype = "float32" min_val = float("-3.84403") @@ -275,6 +300,7 @@ class Program_weight_tensor_data_24: class Program_weight_tensor_data_25: name = "data_25" + original_name = "var_1760" shape = [49, 49] dtype = "int64" min_val = 0 @@ -284,6 +310,7 @@ class Program_weight_tensor_data_25: class Program_weight_tensor_data_26: name = "data_26" + original_name = "var_1834" shape = [49, 49] dtype = "int64" min_val = 0 @@ -293,6 +320,7 @@ class Program_weight_tensor_data_26: class Program_weight_tensor_data_27: name = "data_27" + original_name = "var_1922" shape = [49, 49] dtype = "int64" min_val = 0 @@ -302,6 +330,7 @@ class Program_weight_tensor_data_27: class Program_weight_tensor_data_28: name = "data_28" + original_name = "var_2014" shape = [49, 49] dtype = "int64" min_val = 0 @@ -311,6 +340,7 @@ class Program_weight_tensor_data_28: class Program_weight_tensor_data_29: name = "data_29" + original_name = "var_2102" shape = [49, 49] dtype = "int64" min_val = 0 @@ -320,6 +350,7 @@ class Program_weight_tensor_data_29: class Program_weight_tensor_data_30: name = "data_30" + original_name = "var_2194" shape = [49, 49] dtype = "int64" min_val = 0 @@ -329,6 +360,7 @@ class Program_weight_tensor_data_30: class Program_weight_tensor_data_31: name = "data_31" + original_name = "var_2266" shape = [49, 49] dtype = "int64" min_val = 0 @@ -338,6 +370,7 @@ class Program_weight_tensor_data_31: class Program_weight_tensor_data_32: name = "data_32" + original_name = "var_2358" shape = [49, 49] dtype = "int64" min_val = 0 @@ -347,6 +380,7 @@ class Program_weight_tensor_data_32: class Program_weight_tensor_data_33: name = "data_33" + original_name = "var_2430" shape = [49, 49] dtype = "int64" min_val = 0 @@ -356,6 +390,7 @@ class Program_weight_tensor_data_33: class Program_weight_tensor_data_34: name = "data_34" + original_name = "var_2522" shape = [49, 49] dtype = "int64" min_val = 0 @@ -365,6 +400,7 @@ class Program_weight_tensor_data_34: class Program_weight_tensor_data_35: name = "data_35" + original_name = "var_2594" shape = [49, 49] dtype = "int64" min_val = 0 @@ -374,6 +410,7 @@ class Program_weight_tensor_data_35: class Program_weight_tensor_data_36: name = "data_36" + original_name = "var_2686" shape = [49, 49] dtype = "int64" min_val = 0 @@ -383,6 +420,7 @@ class Program_weight_tensor_data_36: class Program_weight_tensor_data_37: name = "data_37" + original_name = "var_2758" shape = [49, 49] dtype = "int64" min_val = 0 @@ -392,6 +430,7 @@ class Program_weight_tensor_data_37: class Program_weight_tensor_data_38: name = "data_38" + original_name = "var_2850" shape = [49, 49] dtype = "int64" min_val = 0 @@ -401,6 +440,7 @@ class Program_weight_tensor_data_38: class Program_weight_tensor_data_39: name = "data_39" + original_name = "var_2922" shape = [49, 49] dtype = "int64" min_val = 0 @@ -410,6 +450,7 @@ class Program_weight_tensor_data_39: class Program_weight_tensor_data_40: name = "data_40" + original_name = "var_3014" shape = [49, 49] dtype = "int64" min_val = 0 @@ -419,6 +460,7 @@ class Program_weight_tensor_data_40: class Program_weight_tensor_data_41: name = "data_41" + original_name = "var_3086" shape = [49, 49] dtype = "int64" min_val = 0 @@ -428,6 +470,7 @@ class Program_weight_tensor_data_41: class Program_weight_tensor_data_42: name = "data_42" + original_name = "var_3178" shape = [49, 49] dtype = "int64" min_val = 0 @@ -437,6 +480,7 @@ class Program_weight_tensor_data_42: class Program_weight_tensor_data_43: name = "data_43" + original_name = "var_3250" shape = [49, 49] dtype = "int64" min_val = 0 @@ -446,6 +490,7 @@ class Program_weight_tensor_data_43: class Program_weight_tensor_data_44: name = "data_44" + original_name = "var_3342" shape = [49, 49] dtype = "int64" min_val = 0 @@ -455,6 +500,7 @@ class Program_weight_tensor_data_44: class Program_weight_tensor_data_45: name = "data_45" + original_name = "var_3414" shape = [49, 49] dtype = "int64" min_val = 0 @@ -464,6 +510,7 @@ class Program_weight_tensor_data_45: class Program_weight_tensor_data_46: name = "data_46" + original_name = "var_3506" shape = [49, 49] dtype = "int64" min_val = 0 @@ -473,6 +520,7 @@ class Program_weight_tensor_data_46: class Program_weight_tensor_data_47: name = "data_47" + original_name = "var_3594" shape = [49, 49] dtype = "int64" min_val = 0 @@ -482,6 +530,7 @@ class Program_weight_tensor_data_47: class Program_weight_tensor_data_48: name = "data_48" + original_name = "var_3686" shape = [49, 49] dtype = "int64" min_val = 0 diff --git a/paddle_samples/PaddleX/SwinTransformer_base_patch4_window7_224/subgraph_1/weight_meta.py b/paddle_samples/PaddleX/SwinTransformer_base_patch4_window7_224/subgraph_1/weight_meta.py index f544c1540..42a9ec8cb 100644 --- a/paddle_samples/PaddleX/SwinTransformer_base_patch4_window7_224/subgraph_1/weight_meta.py +++ b/paddle_samples/PaddleX/SwinTransformer_base_patch4_window7_224/subgraph_1/weight_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_parameter_0: name = "parameter_0" + original_name = "linear_99.b_0" shape = [102] dtype = "float32" min_val = float("-0.00339081") @@ -11,6 +12,7 @@ class Program_weight_tensor_parameter_0: class Program_weight_tensor_parameter_1: name = "parameter_1" + original_name = "linear_99.w_0" shape = [1024, 102] dtype = "float32" min_val = float("-0.0852928") @@ -22,6 +24,7 @@ class Program_weight_tensor_parameter_1: class Program_weight_tensor_parameter_2: name = "parameter_2" + original_name = "layer_norm_52.b_0" shape = [1024] dtype = "float32" min_val = float("-0.298413") @@ -33,6 +36,7 @@ class Program_weight_tensor_parameter_2: class Program_weight_tensor_parameter_3: name = "parameter_3" + original_name = "layer_norm_52.w_0" shape = [1024] dtype = "float32" min_val = float("-0.00322359") @@ -44,6 +48,7 @@ class Program_weight_tensor_parameter_3: class Program_weight_tensor_parameter_4: name = "parameter_4" + original_name = "linear_98.b_0" shape = [1024] dtype = "float32" min_val = float("-5.68281") @@ -55,6 +60,7 @@ class Program_weight_tensor_parameter_4: class Program_weight_tensor_parameter_5: name = "parameter_5" + original_name = "linear_98.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-1.74127") @@ -66,6 +72,7 @@ class Program_weight_tensor_parameter_5: class Program_weight_tensor_parameter_6: name = "parameter_6" + original_name = "linear_97.b_0" shape = [4096] dtype = "float32" min_val = float("-2.69856") @@ -77,6 +84,7 @@ class Program_weight_tensor_parameter_6: class Program_weight_tensor_parameter_7: name = "parameter_7" + original_name = "linear_97.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.695921") @@ -88,6 +96,7 @@ class Program_weight_tensor_parameter_7: class Program_weight_tensor_parameter_8: name = "parameter_8" + original_name = "layer_norm_51.b_0" shape = [1024] dtype = "float32" min_val = float("-1.03297") @@ -99,6 +108,7 @@ class Program_weight_tensor_parameter_8: class Program_weight_tensor_parameter_9: name = "parameter_9" + original_name = "layer_norm_51.w_0" shape = [1024] dtype = "float32" min_val = float("-0.00559091") @@ -110,6 +120,7 @@ class Program_weight_tensor_parameter_9: class Program_weight_tensor_parameter_10: name = "parameter_10" + original_name = "linear_96.b_0" shape = [1024] dtype = "float32" min_val = float("-5.61142") @@ -121,6 +132,7 @@ class Program_weight_tensor_parameter_10: class Program_weight_tensor_parameter_11: name = "parameter_11" + original_name = "linear_96.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-1.20795") @@ -132,6 +144,7 @@ class Program_weight_tensor_parameter_11: class Program_weight_tensor_parameter_12: name = "parameter_12" + original_name = "linear_95.b_0" shape = [3072] dtype = "float32" min_val = float("-2.51679") @@ -143,6 +156,7 @@ class Program_weight_tensor_parameter_12: class Program_weight_tensor_parameter_13: name = "parameter_13" + original_name = "linear_95.w_0" shape = [1024, 3072] dtype = "float32" min_val = float("-0.343214") @@ -154,6 +168,7 @@ class Program_weight_tensor_parameter_13: class Program_weight_tensor_parameter_14: name = "parameter_14" + original_name = "layer_norm_50.b_0" shape = [1024] dtype = "float32" min_val = float("-2.03064") @@ -165,6 +180,7 @@ class Program_weight_tensor_parameter_14: class Program_weight_tensor_parameter_15: name = "parameter_15" + original_name = "layer_norm_50.w_0" shape = [1024] dtype = "float32" min_val = float("0.000999921") @@ -176,6 +192,7 @@ class Program_weight_tensor_parameter_15: class Program_weight_tensor_parameter_16: name = "parameter_16" + original_name = "linear_94.b_0" shape = [1024] dtype = "float32" min_val = float("-5.84628") @@ -187,6 +204,7 @@ class Program_weight_tensor_parameter_16: class Program_weight_tensor_parameter_17: name = "parameter_17" + original_name = "linear_94.w_0" shape = [4096, 1024] dtype = "float32" min_val = float("-1.43575") @@ -198,6 +216,7 @@ class Program_weight_tensor_parameter_17: class Program_weight_tensor_parameter_18: name = "parameter_18" + original_name = "linear_93.b_0" shape = [4096] dtype = "float32" min_val = float("-1.1791") @@ -209,6 +228,7 @@ class Program_weight_tensor_parameter_18: class Program_weight_tensor_parameter_19: name = "parameter_19" + original_name = "linear_93.w_0" shape = [1024, 4096] dtype = "float32" min_val = float("-0.628782") @@ -220,6 +240,7 @@ class Program_weight_tensor_parameter_19: class Program_weight_tensor_parameter_20: name = "parameter_20" + original_name = "layer_norm_49.b_0" shape = [1024] dtype = "float32" min_val = float("-0.973225") @@ -231,6 +252,7 @@ class Program_weight_tensor_parameter_20: class Program_weight_tensor_parameter_21: name = "parameter_21" + original_name = "layer_norm_49.w_0" shape = [1024] dtype = "float32" min_val = float("-0.0186439") @@ -242,6 +264,7 @@ class Program_weight_tensor_parameter_21: class Program_weight_tensor_parameter_22: name = "parameter_22" + original_name = "linear_92.b_0" shape = [1024] dtype = "float32" min_val = float("-5.33786") @@ -253,6 +276,7 @@ class Program_weight_tensor_parameter_22: class Program_weight_tensor_parameter_23: name = "parameter_23" + original_name = "linear_92.w_0" shape = [1024, 1024] dtype = "float32" min_val = float("-1.63194") @@ -264,6 +288,7 @@ class Program_weight_tensor_parameter_23: class Program_weight_tensor_parameter_24: name = "parameter_24" + original_name = "linear_91.b_0" shape = [3072] dtype = "float32" min_val = float("-2.67093") @@ -275,6 +300,7 @@ class Program_weight_tensor_parameter_24: class Program_weight_tensor_parameter_25: name = "parameter_25" + original_name = "linear_91.w_0" shape = [1024, 3072] dtype = "float32" min_val = float("-0.329921") @@ -286,6 +312,7 @@ class Program_weight_tensor_parameter_25: class Program_weight_tensor_parameter_26: name = "parameter_26" + original_name = "layer_norm_48.b_0" shape = [1024] dtype = "float32" min_val = float("-1.88798") @@ -297,6 +324,7 @@ class Program_weight_tensor_parameter_26: class Program_weight_tensor_parameter_27: name = "parameter_27" + original_name = "layer_norm_48.w_0" shape = [1024] dtype = "float32" min_val = float("0.00173846") @@ -308,6 +336,7 @@ class Program_weight_tensor_parameter_27: class Program_weight_tensor_parameter_28: name = "parameter_28" + original_name = "linear_90.w_0" shape = [2048, 1024] dtype = "float32" min_val = float("-1.14029") @@ -319,6 +348,7 @@ class Program_weight_tensor_parameter_28: class Program_weight_tensor_parameter_29: name = "parameter_29" + original_name = "layer_norm_47.b_0" shape = [2048] dtype = "float32" min_val = float("-1.56509") @@ -330,6 +360,7 @@ class Program_weight_tensor_parameter_29: class Program_weight_tensor_parameter_30: name = "parameter_30" + original_name = "layer_norm_47.w_0" shape = [2048] dtype = "float32" min_val = float("0.0916285") @@ -341,6 +372,7 @@ class Program_weight_tensor_parameter_30: class Program_weight_tensor_parameter_31: name = "parameter_31" + original_name = "linear_89.b_0" shape = [512] dtype = "float32" min_val = float("-1.8465") @@ -352,6 +384,7 @@ class Program_weight_tensor_parameter_31: class Program_weight_tensor_parameter_32: name = "parameter_32" + original_name = "linear_89.w_0" shape = [2048, 512] dtype = "float32" min_val = float("-1.90565") @@ -363,6 +396,7 @@ class Program_weight_tensor_parameter_32: class Program_weight_tensor_parameter_33: name = "parameter_33" + original_name = "linear_88.b_0" shape = [2048] dtype = "float32" min_val = float("-2.48127") @@ -374,6 +408,7 @@ class Program_weight_tensor_parameter_33: class Program_weight_tensor_parameter_34: name = "parameter_34" + original_name = "linear_88.w_0" shape = [512, 2048] dtype = "float32" min_val = float("-0.785721") @@ -385,6 +420,7 @@ class Program_weight_tensor_parameter_34: class Program_weight_tensor_parameter_35: name = "parameter_35" + original_name = "layer_norm_46.b_0" shape = [512] dtype = "float32" min_val = float("-1.89228") @@ -396,6 +432,7 @@ class Program_weight_tensor_parameter_35: class Program_weight_tensor_parameter_36: name = "parameter_36" + original_name = "layer_norm_46.w_0" shape = [512] dtype = "float32" min_val = float("0.282493") @@ -407,6 +444,7 @@ class Program_weight_tensor_parameter_36: class Program_weight_tensor_parameter_37: name = "parameter_37" + original_name = "linear_87.b_0" shape = [512] dtype = "float32" min_val = float("-2.21345") @@ -418,6 +456,7 @@ class Program_weight_tensor_parameter_37: class Program_weight_tensor_parameter_38: name = "parameter_38" + original_name = "linear_87.w_0" shape = [512, 512] dtype = "float32" min_val = float("-1.57026") @@ -429,6 +468,7 @@ class Program_weight_tensor_parameter_38: class Program_weight_tensor_parameter_39: name = "parameter_39" + original_name = "linear_86.b_0" shape = [1536] dtype = "float32" min_val = float("-2.4888") @@ -440,6 +480,7 @@ class Program_weight_tensor_parameter_39: class Program_weight_tensor_parameter_40: name = "parameter_40" + original_name = "linear_86.w_0" shape = [512, 1536] dtype = "float32" min_val = float("-0.344807") @@ -451,6 +492,7 @@ class Program_weight_tensor_parameter_40: class Program_weight_tensor_parameter_41: name = "parameter_41" + original_name = "layer_norm_45.b_0" shape = [512] dtype = "float32" min_val = float("-1.39856") @@ -462,6 +504,7 @@ class Program_weight_tensor_parameter_41: class Program_weight_tensor_parameter_42: name = "parameter_42" + original_name = "layer_norm_45.w_0" shape = [512] dtype = "float32" min_val = float("-0.0106274") @@ -473,6 +516,7 @@ class Program_weight_tensor_parameter_42: class Program_weight_tensor_parameter_43: name = "parameter_43" + original_name = "linear_85.b_0" shape = [512] dtype = "float32" min_val = float("-1.76037") @@ -484,6 +528,7 @@ class Program_weight_tensor_parameter_43: class Program_weight_tensor_parameter_44: name = "parameter_44" + original_name = "linear_85.w_0" shape = [2048, 512] dtype = "float32" min_val = float("-1.31361") @@ -495,6 +540,7 @@ class Program_weight_tensor_parameter_44: class Program_weight_tensor_parameter_45: name = "parameter_45" + original_name = "linear_84.b_0" shape = [2048] dtype = "float32" min_val = float("-2.10443") @@ -506,6 +552,7 @@ class Program_weight_tensor_parameter_45: class Program_weight_tensor_parameter_46: name = "parameter_46" + original_name = "linear_84.w_0" shape = [512, 2048] dtype = "float32" min_val = float("-0.562919") @@ -517,6 +564,7 @@ class Program_weight_tensor_parameter_46: class Program_weight_tensor_parameter_47: name = "parameter_47" + original_name = "layer_norm_44.b_0" shape = [512] dtype = "float32" min_val = float("-2.24666") @@ -528,6 +576,7 @@ class Program_weight_tensor_parameter_47: class Program_weight_tensor_parameter_48: name = "parameter_48" + original_name = "layer_norm_44.w_0" shape = [512] dtype = "float32" min_val = float("0.329326") @@ -539,6 +588,7 @@ class Program_weight_tensor_parameter_48: class Program_weight_tensor_parameter_49: name = "parameter_49" + original_name = "linear_83.b_0" shape = [512] dtype = "float32" min_val = float("-2.16465") @@ -550,6 +600,7 @@ class Program_weight_tensor_parameter_49: class Program_weight_tensor_parameter_50: name = "parameter_50" + original_name = "linear_83.w_0" shape = [512, 512] dtype = "float32" min_val = float("-0.882134") @@ -561,6 +612,7 @@ class Program_weight_tensor_parameter_50: class Program_weight_tensor_parameter_51: name = "parameter_51" + original_name = "linear_82.b_0" shape = [1536] dtype = "float32" min_val = float("-2.48699") @@ -572,6 +624,7 @@ class Program_weight_tensor_parameter_51: class Program_weight_tensor_parameter_52: name = "parameter_52" + original_name = "linear_82.w_0" shape = [512, 1536] dtype = "float32" min_val = float("-0.375673") @@ -583,6 +636,7 @@ class Program_weight_tensor_parameter_52: class Program_weight_tensor_parameter_53: name = "parameter_53" + original_name = "layer_norm_43.b_0" shape = [512] dtype = "float32" min_val = float("-1.11794") @@ -594,6 +648,7 @@ class Program_weight_tensor_parameter_53: class Program_weight_tensor_parameter_54: name = "parameter_54" + original_name = "layer_norm_43.w_0" shape = [512] dtype = "float32" min_val = float("-0.460366") @@ -605,6 +660,7 @@ class Program_weight_tensor_parameter_54: class Program_weight_tensor_parameter_55: name = "parameter_55" + original_name = "linear_81.b_0" shape = [512] dtype = "float32" min_val = float("-1.68029") @@ -616,6 +672,7 @@ class Program_weight_tensor_parameter_55: class Program_weight_tensor_parameter_56: name = "parameter_56" + original_name = "linear_81.w_0" shape = [2048, 512] dtype = "float32" min_val = float("-1.45227") @@ -627,6 +684,7 @@ class Program_weight_tensor_parameter_56: class Program_weight_tensor_parameter_57: name = "parameter_57" + original_name = "linear_80.b_0" shape = [2048] dtype = "float32" min_val = float("-1.28021") @@ -638,6 +696,7 @@ class Program_weight_tensor_parameter_57: class Program_weight_tensor_parameter_58: name = "parameter_58" + original_name = "linear_80.w_0" shape = [512, 2048] dtype = "float32" min_val = float("-0.503439") @@ -649,6 +708,7 @@ class Program_weight_tensor_parameter_58: class Program_weight_tensor_parameter_59: name = "parameter_59" + original_name = "layer_norm_42.b_0" shape = [512] dtype = "float32" min_val = float("-2.20653") @@ -660,6 +720,7 @@ class Program_weight_tensor_parameter_59: class Program_weight_tensor_parameter_60: name = "parameter_60" + original_name = "layer_norm_42.w_0" shape = [512] dtype = "float32" min_val = float("0.289682") @@ -671,6 +732,7 @@ class Program_weight_tensor_parameter_60: class Program_weight_tensor_parameter_61: name = "parameter_61" + original_name = "linear_79.b_0" shape = [512] dtype = "float32" min_val = float("-1.4201") @@ -682,6 +744,7 @@ class Program_weight_tensor_parameter_61: class Program_weight_tensor_parameter_62: name = "parameter_62" + original_name = "linear_79.w_0" shape = [512, 512] dtype = "float32" min_val = float("-0.69512") @@ -693,6 +756,7 @@ class Program_weight_tensor_parameter_62: class Program_weight_tensor_parameter_63: name = "parameter_63" + original_name = "linear_78.b_0" shape = [1536] dtype = "float32" min_val = float("-2.72652") @@ -704,6 +768,7 @@ class Program_weight_tensor_parameter_63: class Program_weight_tensor_parameter_64: name = "parameter_64" + original_name = "linear_78.w_0" shape = [512, 1536] dtype = "float32" min_val = float("-0.36301") @@ -715,6 +780,7 @@ class Program_weight_tensor_parameter_64: class Program_weight_tensor_parameter_65: name = "parameter_65" + original_name = "layer_norm_41.b_0" shape = [512] dtype = "float32" min_val = float("-0.945434") @@ -726,6 +792,7 @@ class Program_weight_tensor_parameter_65: class Program_weight_tensor_parameter_66: name = "parameter_66" + original_name = "layer_norm_41.w_0" shape = [512] dtype = "float32" min_val = float("0.000129176") @@ -737,6 +804,7 @@ class Program_weight_tensor_parameter_66: class Program_weight_tensor_parameter_67: name = "parameter_67" + original_name = "linear_77.b_0" shape = [512] dtype = "float32" min_val = float("-1.33491") @@ -748,6 +816,7 @@ class Program_weight_tensor_parameter_67: class Program_weight_tensor_parameter_68: name = "parameter_68" + original_name = "linear_77.w_0" shape = [2048, 512] dtype = "float32" min_val = float("-1.14651") @@ -759,6 +828,7 @@ class Program_weight_tensor_parameter_68: class Program_weight_tensor_parameter_69: name = "parameter_69" + original_name = "linear_76.b_0" shape = [2048] dtype = "float32" min_val = float("-0.898149") @@ -770,6 +840,7 @@ class Program_weight_tensor_parameter_69: class Program_weight_tensor_parameter_70: name = "parameter_70" + original_name = "linear_76.w_0" shape = [512, 2048] dtype = "float32" min_val = float("-0.449572") @@ -781,6 +852,7 @@ class Program_weight_tensor_parameter_70: class Program_weight_tensor_parameter_71: name = "parameter_71" + original_name = "layer_norm_40.b_0" shape = [512] dtype = "float32" min_val = float("-1.9677") @@ -792,6 +864,7 @@ class Program_weight_tensor_parameter_71: class Program_weight_tensor_parameter_72: name = "parameter_72" + original_name = "layer_norm_40.w_0" shape = [512] dtype = "float32" min_val = float("0.250668") @@ -803,6 +876,7 @@ class Program_weight_tensor_parameter_72: class Program_weight_tensor_parameter_73: name = "parameter_73" + original_name = "linear_75.b_0" shape = [512] dtype = "float32" min_val = float("-1.5505") @@ -814,6 +888,7 @@ class Program_weight_tensor_parameter_73: class Program_weight_tensor_parameter_74: name = "parameter_74" + original_name = "linear_75.w_0" shape = [512, 512] dtype = "float32" min_val = float("-0.689045") @@ -825,6 +900,7 @@ class Program_weight_tensor_parameter_74: class Program_weight_tensor_parameter_75: name = "parameter_75" + original_name = "linear_74.b_0" shape = [1536] dtype = "float32" min_val = float("-2.84621") @@ -836,6 +912,7 @@ class Program_weight_tensor_parameter_75: class Program_weight_tensor_parameter_76: name = "parameter_76" + original_name = "linear_74.w_0" shape = [512, 1536] dtype = "float32" min_val = float("-0.306847") @@ -847,6 +924,7 @@ class Program_weight_tensor_parameter_76: class Program_weight_tensor_parameter_77: name = "parameter_77" + original_name = "layer_norm_39.b_0" shape = [512] dtype = "float32" min_val = float("-0.947213") @@ -858,6 +936,7 @@ class Program_weight_tensor_parameter_77: class Program_weight_tensor_parameter_78: name = "parameter_78" + original_name = "layer_norm_39.w_0" shape = [512] dtype = "float32" min_val = float("0.000809803") @@ -869,6 +948,7 @@ class Program_weight_tensor_parameter_78: class Program_weight_tensor_parameter_79: name = "parameter_79" + original_name = "linear_73.b_0" shape = [512] dtype = "float32" min_val = float("-1.00469") @@ -880,6 +960,7 @@ class Program_weight_tensor_parameter_79: class Program_weight_tensor_parameter_80: name = "parameter_80" + original_name = "linear_73.w_0" shape = [2048, 512] dtype = "float32" min_val = float("-0.47156") @@ -891,6 +972,7 @@ class Program_weight_tensor_parameter_80: class Program_weight_tensor_parameter_81: name = "parameter_81" + original_name = "linear_72.b_0" shape = [2048] dtype = "float32" min_val = float("-0.880635") @@ -902,6 +984,7 @@ class Program_weight_tensor_parameter_81: class Program_weight_tensor_parameter_82: name = "parameter_82" + original_name = "linear_72.w_0" shape = [512, 2048] dtype = "float32" min_val = float("-0.464778") @@ -913,6 +996,7 @@ class Program_weight_tensor_parameter_82: class Program_weight_tensor_parameter_83: name = "parameter_83" + original_name = "layer_norm_38.b_0" shape = [512] dtype = "float32" min_val = float("-1.60249") @@ -924,6 +1008,7 @@ class Program_weight_tensor_parameter_83: class Program_weight_tensor_parameter_84: name = "parameter_84" + original_name = "layer_norm_38.w_0" shape = [512] dtype = "float32" min_val = float("0.235195") @@ -935,6 +1020,7 @@ class Program_weight_tensor_parameter_84: class Program_weight_tensor_parameter_85: name = "parameter_85" + original_name = "linear_71.b_0" shape = [512] dtype = "float32" min_val = float("-1.11752") @@ -946,6 +1032,7 @@ class Program_weight_tensor_parameter_85: class Program_weight_tensor_parameter_86: name = "parameter_86" + original_name = "linear_71.w_0" shape = [512, 512] dtype = "float32" min_val = float("-0.591295") @@ -957,6 +1044,7 @@ class Program_weight_tensor_parameter_86: class Program_weight_tensor_parameter_87: name = "parameter_87" + original_name = "linear_70.b_0" shape = [1536] dtype = "float32" min_val = float("-2.50997") @@ -968,6 +1056,7 @@ class Program_weight_tensor_parameter_87: class Program_weight_tensor_parameter_88: name = "parameter_88" + original_name = "linear_70.w_0" shape = [512, 1536] dtype = "float32" min_val = float("-0.268922") @@ -979,6 +1068,7 @@ class Program_weight_tensor_parameter_88: class Program_weight_tensor_parameter_89: name = "parameter_89" + original_name = "layer_norm_37.b_0" shape = [512] dtype = "float32" min_val = float("-0.957363") @@ -990,6 +1080,7 @@ class Program_weight_tensor_parameter_89: class Program_weight_tensor_parameter_90: name = "parameter_90" + original_name = "layer_norm_37.w_0" shape = [512] dtype = "float32" min_val = float("-0.00190594") @@ -1001,6 +1092,7 @@ class Program_weight_tensor_parameter_90: class Program_weight_tensor_parameter_91: name = "parameter_91" + original_name = "linear_69.b_0" shape = [512] dtype = "float32" min_val = float("-0.931152") @@ -1012,6 +1104,7 @@ class Program_weight_tensor_parameter_91: class Program_weight_tensor_parameter_92: name = "parameter_92" + original_name = "linear_69.w_0" shape = [2048, 512] dtype = "float32" min_val = float("-0.426197") @@ -1023,6 +1116,7 @@ class Program_weight_tensor_parameter_92: class Program_weight_tensor_parameter_93: name = "parameter_93" + original_name = "linear_68.b_0" shape = [2048] dtype = "float32" min_val = float("-0.936706") @@ -1034,6 +1128,7 @@ class Program_weight_tensor_parameter_93: class Program_weight_tensor_parameter_94: name = "parameter_94" + original_name = "linear_68.w_0" shape = [512, 2048] dtype = "float32" min_val = float("-0.558033") @@ -1045,6 +1140,7 @@ class Program_weight_tensor_parameter_94: class Program_weight_tensor_parameter_95: name = "parameter_95" + original_name = "layer_norm_36.b_0" shape = [512] dtype = "float32" min_val = float("-1.24315") @@ -1056,6 +1152,7 @@ class Program_weight_tensor_parameter_95: class Program_weight_tensor_parameter_96: name = "parameter_96" + original_name = "layer_norm_36.w_0" shape = [512] dtype = "float32" min_val = float("0.213141") @@ -1067,6 +1164,7 @@ class Program_weight_tensor_parameter_96: class Program_weight_tensor_parameter_97: name = "parameter_97" + original_name = "linear_67.b_0" shape = [512] dtype = "float32" min_val = float("-1.16917") @@ -1078,6 +1176,7 @@ class Program_weight_tensor_parameter_97: class Program_weight_tensor_parameter_98: name = "parameter_98" + original_name = "linear_67.w_0" shape = [512, 512] dtype = "float32" min_val = float("-1.02109") @@ -1089,6 +1188,7 @@ class Program_weight_tensor_parameter_98: class Program_weight_tensor_parameter_99: name = "parameter_99" + original_name = "linear_66.b_0" shape = [1536] dtype = "float32" min_val = float("-2.76792") @@ -1100,6 +1200,7 @@ class Program_weight_tensor_parameter_99: class Program_weight_tensor_parameter_100: name = "parameter_100" + original_name = "linear_66.w_0" shape = [512, 1536] dtype = "float32" min_val = float("-0.339498") @@ -1111,6 +1212,7 @@ class Program_weight_tensor_parameter_100: class Program_weight_tensor_parameter_101: name = "parameter_101" + original_name = "layer_norm_35.b_0" shape = [512] dtype = "float32" min_val = float("-1.18932") @@ -1122,6 +1224,7 @@ class Program_weight_tensor_parameter_101: class Program_weight_tensor_parameter_102: name = "parameter_102" + original_name = "layer_norm_35.w_0" shape = [512] dtype = "float32" min_val = float("0.00258438") @@ -1133,6 +1236,7 @@ class Program_weight_tensor_parameter_102: class Program_weight_tensor_parameter_103: name = "parameter_103" + original_name = "linear_65.b_0" shape = [512] dtype = "float32" min_val = float("-0.866276") @@ -1144,6 +1248,7 @@ class Program_weight_tensor_parameter_103: class Program_weight_tensor_parameter_104: name = "parameter_104" + original_name = "linear_65.w_0" shape = [2048, 512] dtype = "float32" min_val = float("-0.432479") @@ -1155,6 +1260,7 @@ class Program_weight_tensor_parameter_104: class Program_weight_tensor_parameter_105: name = "parameter_105" + original_name = "linear_64.b_0" shape = [2048] dtype = "float32" min_val = float("-0.900603") @@ -1166,6 +1272,7 @@ class Program_weight_tensor_parameter_105: class Program_weight_tensor_parameter_106: name = "parameter_106" + original_name = "linear_64.w_0" shape = [512, 2048] dtype = "float32" min_val = float("-0.406714") @@ -1177,6 +1284,7 @@ class Program_weight_tensor_parameter_106: class Program_weight_tensor_parameter_107: name = "parameter_107" + original_name = "layer_norm_34.b_0" shape = [512] dtype = "float32" min_val = float("-1.0145") @@ -1188,6 +1296,7 @@ class Program_weight_tensor_parameter_107: class Program_weight_tensor_parameter_108: name = "parameter_108" + original_name = "layer_norm_34.w_0" shape = [512] dtype = "float32" min_val = float("0.0117145") @@ -1199,6 +1308,7 @@ class Program_weight_tensor_parameter_108: class Program_weight_tensor_parameter_109: name = "parameter_109" + original_name = "linear_63.b_0" shape = [512] dtype = "float32" min_val = float("-0.570489") @@ -1210,6 +1320,7 @@ class Program_weight_tensor_parameter_109: class Program_weight_tensor_parameter_110: name = "parameter_110" + original_name = "linear_63.w_0" shape = [512, 512] dtype = "float32" min_val = float("-0.573036") @@ -1221,6 +1332,7 @@ class Program_weight_tensor_parameter_110: class Program_weight_tensor_parameter_111: name = "parameter_111" + original_name = "linear_62.b_0" shape = [1536] dtype = "float32" min_val = float("-2.7765") @@ -1232,6 +1344,7 @@ class Program_weight_tensor_parameter_111: class Program_weight_tensor_parameter_112: name = "parameter_112" + original_name = "linear_62.w_0" shape = [512, 1536] dtype = "float32" min_val = float("-0.301955") @@ -1243,6 +1356,7 @@ class Program_weight_tensor_parameter_112: class Program_weight_tensor_parameter_113: name = "parameter_113" + original_name = "layer_norm_33.b_0" shape = [512] dtype = "float32" min_val = float("-1.23658") @@ -1254,6 +1368,7 @@ class Program_weight_tensor_parameter_113: class Program_weight_tensor_parameter_114: name = "parameter_114" + original_name = "layer_norm_33.w_0" shape = [512] dtype = "float32" min_val = float("0.113181") @@ -1265,6 +1380,7 @@ class Program_weight_tensor_parameter_114: class Program_weight_tensor_parameter_115: name = "parameter_115" + original_name = "linear_61.b_0" shape = [512] dtype = "float32" min_val = float("-0.988776") @@ -1276,6 +1392,7 @@ class Program_weight_tensor_parameter_115: class Program_weight_tensor_parameter_116: name = "parameter_116" + original_name = "linear_61.w_0" shape = [2048, 512] dtype = "float32" min_val = float("-0.507395") @@ -1287,6 +1404,7 @@ class Program_weight_tensor_parameter_116: class Program_weight_tensor_parameter_117: name = "parameter_117" + original_name = "linear_60.b_0" shape = [2048] dtype = "float32" min_val = float("-1.04561") @@ -1298,6 +1416,7 @@ class Program_weight_tensor_parameter_117: class Program_weight_tensor_parameter_118: name = "parameter_118" + original_name = "linear_60.w_0" shape = [512, 2048] dtype = "float32" min_val = float("-0.376831") @@ -1309,6 +1428,7 @@ class Program_weight_tensor_parameter_118: class Program_weight_tensor_parameter_119: name = "parameter_119" + original_name = "layer_norm_32.b_0" shape = [512] dtype = "float32" min_val = float("-1.11014") @@ -1320,6 +1440,7 @@ class Program_weight_tensor_parameter_119: class Program_weight_tensor_parameter_120: name = "parameter_120" + original_name = "layer_norm_32.w_0" shape = [512] dtype = "float32" min_val = float("0.0412977") @@ -1331,6 +1452,7 @@ class Program_weight_tensor_parameter_120: class Program_weight_tensor_parameter_121: name = "parameter_121" + original_name = "linear_59.b_0" shape = [512] dtype = "float32" min_val = float("-0.961211") @@ -1342,6 +1464,7 @@ class Program_weight_tensor_parameter_121: class Program_weight_tensor_parameter_122: name = "parameter_122" + original_name = "linear_59.w_0" shape = [512, 512] dtype = "float32" min_val = float("-0.456277") @@ -1353,6 +1476,7 @@ class Program_weight_tensor_parameter_122: class Program_weight_tensor_parameter_123: name = "parameter_123" + original_name = "linear_58.b_0" shape = [1536] dtype = "float32" min_val = float("-2.94983") @@ -1364,6 +1488,7 @@ class Program_weight_tensor_parameter_123: class Program_weight_tensor_parameter_124: name = "parameter_124" + original_name = "linear_58.w_0" shape = [512, 1536] dtype = "float32" min_val = float("-0.297322") @@ -1375,6 +1500,7 @@ class Program_weight_tensor_parameter_124: class Program_weight_tensor_parameter_125: name = "parameter_125" + original_name = "layer_norm_31.b_0" shape = [512] dtype = "float32" min_val = float("-1.68091") @@ -1386,6 +1512,7 @@ class Program_weight_tensor_parameter_125: class Program_weight_tensor_parameter_126: name = "parameter_126" + original_name = "layer_norm_31.w_0" shape = [512] dtype = "float32" min_val = float("0.197783") @@ -1397,6 +1524,7 @@ class Program_weight_tensor_parameter_126: class Program_weight_tensor_parameter_127: name = "parameter_127" + original_name = "linear_57.b_0" shape = [512] dtype = "float32" min_val = float("-1.17085") @@ -1408,6 +1536,7 @@ class Program_weight_tensor_parameter_127: class Program_weight_tensor_parameter_128: name = "parameter_128" + original_name = "linear_57.w_0" shape = [2048, 512] dtype = "float32" min_val = float("-0.506696") @@ -1419,6 +1548,7 @@ class Program_weight_tensor_parameter_128: class Program_weight_tensor_parameter_129: name = "parameter_129" + original_name = "linear_56.b_0" shape = [2048] dtype = "float32" min_val = float("-0.922793") @@ -1430,6 +1560,7 @@ class Program_weight_tensor_parameter_129: class Program_weight_tensor_parameter_130: name = "parameter_130" + original_name = "linear_56.w_0" shape = [512, 2048] dtype = "float32" min_val = float("-0.959411") @@ -1441,6 +1572,7 @@ class Program_weight_tensor_parameter_130: class Program_weight_tensor_parameter_131: name = "parameter_131" + original_name = "layer_norm_30.b_0" shape = [512] dtype = "float32" min_val = float("-1.17093") @@ -1452,6 +1584,7 @@ class Program_weight_tensor_parameter_131: class Program_weight_tensor_parameter_132: name = "parameter_132" + original_name = "layer_norm_30.w_0" shape = [512] dtype = "float32" min_val = float("0.274545") @@ -1463,6 +1596,7 @@ class Program_weight_tensor_parameter_132: class Program_weight_tensor_parameter_133: name = "parameter_133" + original_name = "linear_55.b_0" shape = [512] dtype = "float32" min_val = float("-0.592245") @@ -1474,6 +1608,7 @@ class Program_weight_tensor_parameter_133: class Program_weight_tensor_parameter_134: name = "parameter_134" + original_name = "linear_55.w_0" shape = [512, 512] dtype = "float32" min_val = float("-0.426103") @@ -1485,6 +1620,7 @@ class Program_weight_tensor_parameter_134: class Program_weight_tensor_parameter_135: name = "parameter_135" + original_name = "linear_54.b_0" shape = [1536] dtype = "float32" min_val = float("-2.57754") @@ -1496,6 +1632,7 @@ class Program_weight_tensor_parameter_135: class Program_weight_tensor_parameter_136: name = "parameter_136" + original_name = "linear_54.w_0" shape = [512, 1536] dtype = "float32" min_val = float("-0.302804") @@ -1507,6 +1644,7 @@ class Program_weight_tensor_parameter_136: class Program_weight_tensor_parameter_137: name = "parameter_137" + original_name = "layer_norm_29.b_0" shape = [512] dtype = "float32" min_val = float("-1.94345") @@ -1518,6 +1656,7 @@ class Program_weight_tensor_parameter_137: class Program_weight_tensor_parameter_138: name = "parameter_138" + original_name = "layer_norm_29.w_0" shape = [512] dtype = "float32" min_val = float("0.218373") @@ -1529,6 +1668,7 @@ class Program_weight_tensor_parameter_138: class Program_weight_tensor_parameter_139: name = "parameter_139" + original_name = "linear_53.b_0" shape = [512] dtype = "float32" min_val = float("-1.49068") @@ -1540,6 +1680,7 @@ class Program_weight_tensor_parameter_139: class Program_weight_tensor_parameter_140: name = "parameter_140" + original_name = "linear_53.w_0" shape = [2048, 512] dtype = "float32" min_val = float("-0.594898") @@ -1551,6 +1692,7 @@ class Program_weight_tensor_parameter_140: class Program_weight_tensor_parameter_141: name = "parameter_141" + original_name = "linear_52.b_0" shape = [2048] dtype = "float32" min_val = float("-1.01011") @@ -1562,6 +1704,7 @@ class Program_weight_tensor_parameter_141: class Program_weight_tensor_parameter_142: name = "parameter_142" + original_name = "linear_52.w_0" shape = [512, 2048] dtype = "float32" min_val = float("-0.700529") @@ -1573,6 +1716,7 @@ class Program_weight_tensor_parameter_142: class Program_weight_tensor_parameter_143: name = "parameter_143" + original_name = "layer_norm_28.b_0" shape = [512] dtype = "float32" min_val = float("-1.19381") @@ -1584,6 +1728,7 @@ class Program_weight_tensor_parameter_143: class Program_weight_tensor_parameter_144: name = "parameter_144" + original_name = "layer_norm_28.w_0" shape = [512] dtype = "float32" min_val = float("0.367199") @@ -1595,6 +1740,7 @@ class Program_weight_tensor_parameter_144: class Program_weight_tensor_parameter_145: name = "parameter_145" + original_name = "linear_51.b_0" shape = [512] dtype = "float32" min_val = float("-0.858971") @@ -1606,6 +1752,7 @@ class Program_weight_tensor_parameter_145: class Program_weight_tensor_parameter_146: name = "parameter_146" + original_name = "linear_51.w_0" shape = [512, 512] dtype = "float32" min_val = float("-0.415129") @@ -1617,6 +1764,7 @@ class Program_weight_tensor_parameter_146: class Program_weight_tensor_parameter_147: name = "parameter_147" + original_name = "linear_50.b_0" shape = [1536] dtype = "float32" min_val = float("-2.91731") @@ -1628,6 +1776,7 @@ class Program_weight_tensor_parameter_147: class Program_weight_tensor_parameter_148: name = "parameter_148" + original_name = "linear_50.w_0" shape = [512, 1536] dtype = "float32" min_val = float("-0.349039") @@ -1639,6 +1788,7 @@ class Program_weight_tensor_parameter_148: class Program_weight_tensor_parameter_149: name = "parameter_149" + original_name = "layer_norm_27.b_0" shape = [512] dtype = "float32" min_val = float("-1.82394") @@ -1650,6 +1800,7 @@ class Program_weight_tensor_parameter_149: class Program_weight_tensor_parameter_150: name = "parameter_150" + original_name = "layer_norm_27.w_0" shape = [512] dtype = "float32" min_val = float("0.249513") @@ -1661,6 +1812,7 @@ class Program_weight_tensor_parameter_150: class Program_weight_tensor_parameter_151: name = "parameter_151" + original_name = "linear_49.b_0" shape = [512] dtype = "float32" min_val = float("-1.3932") @@ -1672,6 +1824,7 @@ class Program_weight_tensor_parameter_151: class Program_weight_tensor_parameter_152: name = "parameter_152" + original_name = "linear_49.w_0" shape = [2048, 512] dtype = "float32" min_val = float("-0.59764") @@ -1683,6 +1836,7 @@ class Program_weight_tensor_parameter_152: class Program_weight_tensor_parameter_153: name = "parameter_153" + original_name = "linear_48.b_0" shape = [2048] dtype = "float32" min_val = float("-1.14458") @@ -1694,6 +1848,7 @@ class Program_weight_tensor_parameter_153: class Program_weight_tensor_parameter_154: name = "parameter_154" + original_name = "linear_48.w_0" shape = [512, 2048] dtype = "float32" min_val = float("-0.599303") @@ -1705,6 +1860,7 @@ class Program_weight_tensor_parameter_154: class Program_weight_tensor_parameter_155: name = "parameter_155" + original_name = "layer_norm_26.b_0" shape = [512] dtype = "float32" min_val = float("-1.14038") @@ -1716,6 +1872,7 @@ class Program_weight_tensor_parameter_155: class Program_weight_tensor_parameter_156: name = "parameter_156" + original_name = "layer_norm_26.w_0" shape = [512] dtype = "float32" min_val = float("0.446159") @@ -1727,6 +1884,7 @@ class Program_weight_tensor_parameter_156: class Program_weight_tensor_parameter_157: name = "parameter_157" + original_name = "linear_47.b_0" shape = [512] dtype = "float32" min_val = float("-0.577975") @@ -1738,6 +1896,7 @@ class Program_weight_tensor_parameter_157: class Program_weight_tensor_parameter_158: name = "parameter_158" + original_name = "linear_47.w_0" shape = [512, 512] dtype = "float32" min_val = float("-0.386004") @@ -1749,6 +1908,7 @@ class Program_weight_tensor_parameter_158: class Program_weight_tensor_parameter_159: name = "parameter_159" + original_name = "linear_46.b_0" shape = [1536] dtype = "float32" min_val = float("-2.72338") @@ -1760,6 +1920,7 @@ class Program_weight_tensor_parameter_159: class Program_weight_tensor_parameter_160: name = "parameter_160" + original_name = "linear_46.w_0" shape = [512, 1536] dtype = "float32" min_val = float("-0.310438") @@ -1771,6 +1932,7 @@ class Program_weight_tensor_parameter_160: class Program_weight_tensor_parameter_161: name = "parameter_161" + original_name = "layer_norm_25.b_0" shape = [512] dtype = "float32" min_val = float("-1.7652") @@ -1782,6 +1944,7 @@ class Program_weight_tensor_parameter_161: class Program_weight_tensor_parameter_162: name = "parameter_162" + original_name = "layer_norm_25.w_0" shape = [512] dtype = "float32" min_val = float("0.23496") @@ -1793,6 +1956,7 @@ class Program_weight_tensor_parameter_162: class Program_weight_tensor_parameter_163: name = "parameter_163" + original_name = "linear_45.b_0" shape = [512] dtype = "float32" min_val = float("-1.4547") @@ -1804,6 +1968,7 @@ class Program_weight_tensor_parameter_163: class Program_weight_tensor_parameter_164: name = "parameter_164" + original_name = "linear_45.w_0" shape = [2048, 512] dtype = "float32" min_val = float("-0.529154") @@ -1815,6 +1980,7 @@ class Program_weight_tensor_parameter_164: class Program_weight_tensor_parameter_165: name = "parameter_165" + original_name = "linear_44.b_0" shape = [2048] dtype = "float32" min_val = float("-1.22367") @@ -1826,6 +1992,7 @@ class Program_weight_tensor_parameter_165: class Program_weight_tensor_parameter_166: name = "parameter_166" + original_name = "linear_44.w_0" shape = [512, 2048] dtype = "float32" min_val = float("-0.533354") @@ -1837,6 +2004,7 @@ class Program_weight_tensor_parameter_166: class Program_weight_tensor_parameter_167: name = "parameter_167" + original_name = "layer_norm_24.b_0" shape = [512] dtype = "float32" min_val = float("-1.12307") @@ -1848,6 +2016,7 @@ class Program_weight_tensor_parameter_167: class Program_weight_tensor_parameter_168: name = "parameter_168" + original_name = "layer_norm_24.w_0" shape = [512] dtype = "float32" min_val = float("0.401248") @@ -1859,6 +2028,7 @@ class Program_weight_tensor_parameter_168: class Program_weight_tensor_parameter_169: name = "parameter_169" + original_name = "linear_43.b_0" shape = [512] dtype = "float32" min_val = float("-0.777867") @@ -1870,6 +2040,7 @@ class Program_weight_tensor_parameter_169: class Program_weight_tensor_parameter_170: name = "parameter_170" + original_name = "linear_43.w_0" shape = [512, 512] dtype = "float32" min_val = float("-0.504021") @@ -1881,6 +2052,7 @@ class Program_weight_tensor_parameter_170: class Program_weight_tensor_parameter_171: name = "parameter_171" + original_name = "linear_42.b_0" shape = [1536] dtype = "float32" min_val = float("-2.76991") @@ -1892,6 +2064,7 @@ class Program_weight_tensor_parameter_171: class Program_weight_tensor_parameter_172: name = "parameter_172" + original_name = "linear_42.w_0" shape = [512, 1536] dtype = "float32" min_val = float("-0.426341") @@ -1903,6 +2076,7 @@ class Program_weight_tensor_parameter_172: class Program_weight_tensor_parameter_173: name = "parameter_173" + original_name = "layer_norm_23.b_0" shape = [512] dtype = "float32" min_val = float("-1.63316") @@ -1914,6 +2088,7 @@ class Program_weight_tensor_parameter_173: class Program_weight_tensor_parameter_174: name = "parameter_174" + original_name = "layer_norm_23.w_0" shape = [512] dtype = "float32" min_val = float("0.187892") @@ -1925,6 +2100,7 @@ class Program_weight_tensor_parameter_174: class Program_weight_tensor_parameter_175: name = "parameter_175" + original_name = "linear_41.b_0" shape = [512] dtype = "float32" min_val = float("-0.613175") @@ -1936,6 +2112,7 @@ class Program_weight_tensor_parameter_175: class Program_weight_tensor_parameter_176: name = "parameter_176" + original_name = "linear_41.w_0" shape = [2048, 512] dtype = "float32" min_val = float("-0.390706") @@ -1947,6 +2124,7 @@ class Program_weight_tensor_parameter_176: class Program_weight_tensor_parameter_177: name = "parameter_177" + original_name = "linear_40.b_0" shape = [2048] dtype = "float32" min_val = float("-1.27482") @@ -1958,6 +2136,7 @@ class Program_weight_tensor_parameter_177: class Program_weight_tensor_parameter_178: name = "parameter_178" + original_name = "linear_40.w_0" shape = [512, 2048] dtype = "float32" min_val = float("-0.484037") @@ -1969,6 +2148,7 @@ class Program_weight_tensor_parameter_178: class Program_weight_tensor_parameter_179: name = "parameter_179" + original_name = "layer_norm_22.b_0" shape = [512] dtype = "float32" min_val = float("-1.04238") @@ -1980,6 +2160,7 @@ class Program_weight_tensor_parameter_179: class Program_weight_tensor_parameter_180: name = "parameter_180" + original_name = "layer_norm_22.w_0" shape = [512] dtype = "float32" min_val = float("0.370907") @@ -1991,6 +2172,7 @@ class Program_weight_tensor_parameter_180: class Program_weight_tensor_parameter_181: name = "parameter_181" + original_name = "linear_39.b_0" shape = [512] dtype = "float32" min_val = float("-0.552353") @@ -2002,6 +2184,7 @@ class Program_weight_tensor_parameter_181: class Program_weight_tensor_parameter_182: name = "parameter_182" + original_name = "linear_39.w_0" shape = [512, 512] dtype = "float32" min_val = float("-0.414368") @@ -2013,6 +2196,7 @@ class Program_weight_tensor_parameter_182: class Program_weight_tensor_parameter_183: name = "parameter_183" + original_name = "linear_38.b_0" shape = [1536] dtype = "float32" min_val = float("-2.02891") @@ -2024,6 +2208,7 @@ class Program_weight_tensor_parameter_183: class Program_weight_tensor_parameter_184: name = "parameter_184" + original_name = "linear_38.w_0" shape = [512, 1536] dtype = "float32" min_val = float("-0.468538") @@ -2035,6 +2220,7 @@ class Program_weight_tensor_parameter_184: class Program_weight_tensor_parameter_185: name = "parameter_185" + original_name = "layer_norm_21.b_0" shape = [512] dtype = "float32" min_val = float("-1.49665") @@ -2046,6 +2232,7 @@ class Program_weight_tensor_parameter_185: class Program_weight_tensor_parameter_186: name = "parameter_186" + original_name = "layer_norm_21.w_0" shape = [512] dtype = "float32" min_val = float("0.169749") @@ -2057,6 +2244,7 @@ class Program_weight_tensor_parameter_186: class Program_weight_tensor_parameter_187: name = "parameter_187" + original_name = "linear_37.b_0" shape = [512] dtype = "float32" min_val = float("-0.66809") @@ -2068,6 +2256,7 @@ class Program_weight_tensor_parameter_187: class Program_weight_tensor_parameter_188: name = "parameter_188" + original_name = "linear_37.w_0" shape = [2048, 512] dtype = "float32" min_val = float("-0.44651") @@ -2079,6 +2268,7 @@ class Program_weight_tensor_parameter_188: class Program_weight_tensor_parameter_189: name = "parameter_189" + original_name = "linear_36.b_0" shape = [2048] dtype = "float32" min_val = float("-1.34416") @@ -2090,6 +2280,7 @@ class Program_weight_tensor_parameter_189: class Program_weight_tensor_parameter_190: name = "parameter_190" + original_name = "linear_36.w_0" shape = [512, 2048] dtype = "float32" min_val = float("-0.482229") @@ -2101,6 +2292,7 @@ class Program_weight_tensor_parameter_190: class Program_weight_tensor_parameter_191: name = "parameter_191" + original_name = "layer_norm_20.b_0" shape = [512] dtype = "float32" min_val = float("-1.12325") @@ -2112,6 +2304,7 @@ class Program_weight_tensor_parameter_191: class Program_weight_tensor_parameter_192: name = "parameter_192" + original_name = "layer_norm_20.w_0" shape = [512] dtype = "float32" min_val = float("0.276635") @@ -2123,6 +2316,7 @@ class Program_weight_tensor_parameter_192: class Program_weight_tensor_parameter_193: name = "parameter_193" + original_name = "linear_35.b_0" shape = [512] dtype = "float32" min_val = float("-0.569741") @@ -2134,6 +2328,7 @@ class Program_weight_tensor_parameter_193: class Program_weight_tensor_parameter_194: name = "parameter_194" + original_name = "linear_35.w_0" shape = [512, 512] dtype = "float32" min_val = float("-0.314703") @@ -2145,6 +2340,7 @@ class Program_weight_tensor_parameter_194: class Program_weight_tensor_parameter_195: name = "parameter_195" + original_name = "linear_34.b_0" shape = [1536] dtype = "float32" min_val = float("-1.85512") @@ -2156,6 +2352,7 @@ class Program_weight_tensor_parameter_195: class Program_weight_tensor_parameter_196: name = "parameter_196" + original_name = "linear_34.w_0" shape = [512, 1536] dtype = "float32" min_val = float("-0.477888") @@ -2167,6 +2364,7 @@ class Program_weight_tensor_parameter_196: class Program_weight_tensor_parameter_197: name = "parameter_197" + original_name = "layer_norm_19.b_0" shape = [512] dtype = "float32" min_val = float("-1.0507") @@ -2178,6 +2376,7 @@ class Program_weight_tensor_parameter_197: class Program_weight_tensor_parameter_198: name = "parameter_198" + original_name = "layer_norm_19.w_0" shape = [512] dtype = "float32" min_val = float("0.122625") @@ -2189,6 +2388,7 @@ class Program_weight_tensor_parameter_198: class Program_weight_tensor_parameter_199: name = "parameter_199" + original_name = "linear_33.b_0" shape = [512] dtype = "float32" min_val = float("-0.611639") @@ -2200,6 +2400,7 @@ class Program_weight_tensor_parameter_199: class Program_weight_tensor_parameter_200: name = "parameter_200" + original_name = "linear_33.w_0" shape = [2048, 512] dtype = "float32" min_val = float("-0.539138") @@ -2211,6 +2412,7 @@ class Program_weight_tensor_parameter_200: class Program_weight_tensor_parameter_201: name = "parameter_201" + original_name = "linear_32.b_0" shape = [2048] dtype = "float32" min_val = float("-1.38504") @@ -2222,6 +2424,7 @@ class Program_weight_tensor_parameter_201: class Program_weight_tensor_parameter_202: name = "parameter_202" + original_name = "linear_32.w_0" shape = [512, 2048] dtype = "float32" min_val = float("-0.571152") @@ -2233,6 +2436,7 @@ class Program_weight_tensor_parameter_202: class Program_weight_tensor_parameter_203: name = "parameter_203" + original_name = "layer_norm_18.b_0" shape = [512] dtype = "float32" min_val = float("-1.171") @@ -2244,6 +2448,7 @@ class Program_weight_tensor_parameter_203: class Program_weight_tensor_parameter_204: name = "parameter_204" + original_name = "layer_norm_18.w_0" shape = [512] dtype = "float32" min_val = float("0.198497") @@ -2255,6 +2460,7 @@ class Program_weight_tensor_parameter_204: class Program_weight_tensor_parameter_205: name = "parameter_205" + original_name = "linear_31.b_0" shape = [512] dtype = "float32" min_val = float("-0.575145") @@ -2266,6 +2472,7 @@ class Program_weight_tensor_parameter_205: class Program_weight_tensor_parameter_206: name = "parameter_206" + original_name = "linear_31.w_0" shape = [512, 512] dtype = "float32" min_val = float("-0.294475") @@ -2277,6 +2484,7 @@ class Program_weight_tensor_parameter_206: class Program_weight_tensor_parameter_207: name = "parameter_207" + original_name = "linear_30.b_0" shape = [1536] dtype = "float32" min_val = float("-2.43613") @@ -2288,6 +2496,7 @@ class Program_weight_tensor_parameter_207: class Program_weight_tensor_parameter_208: name = "parameter_208" + original_name = "linear_30.w_0" shape = [512, 1536] dtype = "float32" min_val = float("-0.482794") @@ -2299,6 +2508,7 @@ class Program_weight_tensor_parameter_208: class Program_weight_tensor_parameter_209: name = "parameter_209" + original_name = "layer_norm_17.b_0" shape = [512] dtype = "float32" min_val = float("-0.983449") @@ -2310,6 +2520,7 @@ class Program_weight_tensor_parameter_209: class Program_weight_tensor_parameter_210: name = "parameter_210" + original_name = "layer_norm_17.w_0" shape = [512] dtype = "float32" min_val = float("0.138266") @@ -2321,6 +2532,7 @@ class Program_weight_tensor_parameter_210: class Program_weight_tensor_parameter_211: name = "parameter_211" + original_name = "linear_29.b_0" shape = [512] dtype = "float32" min_val = float("-0.745349") @@ -2332,6 +2544,7 @@ class Program_weight_tensor_parameter_211: class Program_weight_tensor_parameter_212: name = "parameter_212" + original_name = "linear_29.w_0" shape = [2048, 512] dtype = "float32" min_val = float("-0.469036") @@ -2343,6 +2556,7 @@ class Program_weight_tensor_parameter_212: class Program_weight_tensor_parameter_213: name = "parameter_213" + original_name = "linear_28.b_0" shape = [2048] dtype = "float32" min_val = float("-1.29033") @@ -2354,6 +2568,7 @@ class Program_weight_tensor_parameter_213: class Program_weight_tensor_parameter_214: name = "parameter_214" + original_name = "linear_28.w_0" shape = [512, 2048] dtype = "float32" min_val = float("-0.326713") @@ -2365,6 +2580,7 @@ class Program_weight_tensor_parameter_214: class Program_weight_tensor_parameter_215: name = "parameter_215" + original_name = "layer_norm_16.b_0" shape = [512] dtype = "float32" min_val = float("-1.3187") @@ -2376,6 +2592,7 @@ class Program_weight_tensor_parameter_215: class Program_weight_tensor_parameter_216: name = "parameter_216" + original_name = "layer_norm_16.w_0" shape = [512] dtype = "float32" min_val = float("0.000379424") @@ -2387,6 +2604,7 @@ class Program_weight_tensor_parameter_216: class Program_weight_tensor_parameter_217: name = "parameter_217" + original_name = "linear_27.b_0" shape = [512] dtype = "float32" min_val = float("-0.752592") @@ -2398,6 +2616,7 @@ class Program_weight_tensor_parameter_217: class Program_weight_tensor_parameter_218: name = "parameter_218" + original_name = "linear_27.w_0" shape = [512, 512] dtype = "float32" min_val = float("-0.346008") @@ -2409,6 +2628,7 @@ class Program_weight_tensor_parameter_218: class Program_weight_tensor_parameter_219: name = "parameter_219" + original_name = "linear_26.b_0" shape = [1536] dtype = "float32" min_val = float("-2.39234") @@ -2420,6 +2640,7 @@ class Program_weight_tensor_parameter_219: class Program_weight_tensor_parameter_220: name = "parameter_220" + original_name = "linear_26.w_0" shape = [512, 1536] dtype = "float32" min_val = float("-0.354986") @@ -2431,6 +2652,7 @@ class Program_weight_tensor_parameter_220: class Program_weight_tensor_parameter_221: name = "parameter_221" + original_name = "layer_norm_15.b_0" shape = [512] dtype = "float32" min_val = float("-0.991129") @@ -2442,6 +2664,7 @@ class Program_weight_tensor_parameter_221: class Program_weight_tensor_parameter_222: name = "parameter_222" + original_name = "layer_norm_15.w_0" shape = [512] dtype = "float32" min_val = float("0.00229") @@ -2453,6 +2676,7 @@ class Program_weight_tensor_parameter_222: class Program_weight_tensor_parameter_223: name = "parameter_223" + original_name = "linear_25.b_0" shape = [512] dtype = "float32" min_val = float("-0.671195") @@ -2464,6 +2688,7 @@ class Program_weight_tensor_parameter_223: class Program_weight_tensor_parameter_224: name = "parameter_224" + original_name = "linear_25.w_0" shape = [2048, 512] dtype = "float32" min_val = float("-0.340001") @@ -2475,6 +2700,7 @@ class Program_weight_tensor_parameter_224: class Program_weight_tensor_parameter_225: name = "parameter_225" + original_name = "linear_24.b_0" shape = [2048] dtype = "float32" min_val = float("-1.41103") @@ -2486,6 +2712,7 @@ class Program_weight_tensor_parameter_225: class Program_weight_tensor_parameter_226: name = "parameter_226" + original_name = "linear_24.w_0" shape = [512, 2048] dtype = "float32" min_val = float("-0.292941") @@ -2497,6 +2724,7 @@ class Program_weight_tensor_parameter_226: class Program_weight_tensor_parameter_227: name = "parameter_227" + original_name = "layer_norm_14.b_0" shape = [512] dtype = "float32" min_val = float("-1.18587") @@ -2508,6 +2736,7 @@ class Program_weight_tensor_parameter_227: class Program_weight_tensor_parameter_228: name = "parameter_228" + original_name = "layer_norm_14.w_0" shape = [512] dtype = "float32" min_val = float("0.000472873") @@ -2519,6 +2748,7 @@ class Program_weight_tensor_parameter_228: class Program_weight_tensor_parameter_229: name = "parameter_229" + original_name = "linear_23.b_0" shape = [512] dtype = "float32" min_val = float("-0.544169") @@ -2530,6 +2760,7 @@ class Program_weight_tensor_parameter_229: class Program_weight_tensor_parameter_230: name = "parameter_230" + original_name = "linear_23.w_0" shape = [512, 512] dtype = "float32" min_val = float("-0.291579") @@ -2541,6 +2772,7 @@ class Program_weight_tensor_parameter_230: class Program_weight_tensor_parameter_231: name = "parameter_231" + original_name = "linear_22.b_0" shape = [1536] dtype = "float32" min_val = float("-2.5873") @@ -2552,6 +2784,7 @@ class Program_weight_tensor_parameter_231: class Program_weight_tensor_parameter_232: name = "parameter_232" + original_name = "linear_22.w_0" shape = [512, 1536] dtype = "float32" min_val = float("-0.509198") @@ -2563,6 +2796,7 @@ class Program_weight_tensor_parameter_232: class Program_weight_tensor_parameter_233: name = "parameter_233" + original_name = "layer_norm_13.b_0" shape = [512] dtype = "float32" min_val = float("-0.961899") @@ -2574,6 +2808,7 @@ class Program_weight_tensor_parameter_233: class Program_weight_tensor_parameter_234: name = "parameter_234" + original_name = "layer_norm_13.w_0" shape = [512] dtype = "float32" min_val = float("0.00115605") @@ -2585,6 +2820,7 @@ class Program_weight_tensor_parameter_234: class Program_weight_tensor_parameter_235: name = "parameter_235" + original_name = "linear_21.b_0" shape = [512] dtype = "float32" min_val = float("-0.736192") @@ -2596,6 +2832,7 @@ class Program_weight_tensor_parameter_235: class Program_weight_tensor_parameter_236: name = "parameter_236" + original_name = "linear_21.w_0" shape = [2048, 512] dtype = "float32" min_val = float("-0.440924") @@ -2607,6 +2844,7 @@ class Program_weight_tensor_parameter_236: class Program_weight_tensor_parameter_237: name = "parameter_237" + original_name = "linear_20.b_0" shape = [2048] dtype = "float32" min_val = float("-1.7101") @@ -2618,6 +2856,7 @@ class Program_weight_tensor_parameter_237: class Program_weight_tensor_parameter_238: name = "parameter_238" + original_name = "linear_20.w_0" shape = [512, 2048] dtype = "float32" min_val = float("-0.351908") @@ -2629,6 +2868,7 @@ class Program_weight_tensor_parameter_238: class Program_weight_tensor_parameter_239: name = "parameter_239" + original_name = "layer_norm_12.b_0" shape = [512] dtype = "float32" min_val = float("-1.01456") @@ -2640,6 +2880,7 @@ class Program_weight_tensor_parameter_239: class Program_weight_tensor_parameter_240: name = "parameter_240" + original_name = "layer_norm_12.w_0" shape = [512] dtype = "float32" min_val = float("-0.00322139") @@ -2651,6 +2892,7 @@ class Program_weight_tensor_parameter_240: class Program_weight_tensor_parameter_241: name = "parameter_241" + original_name = "linear_19.b_0" shape = [512] dtype = "float32" min_val = float("-0.819815") @@ -2662,6 +2904,7 @@ class Program_weight_tensor_parameter_241: class Program_weight_tensor_parameter_242: name = "parameter_242" + original_name = "linear_19.w_0" shape = [512, 512] dtype = "float32" min_val = float("-0.271166") @@ -2673,6 +2916,7 @@ class Program_weight_tensor_parameter_242: class Program_weight_tensor_parameter_243: name = "parameter_243" + original_name = "linear_18.b_0" shape = [1536] dtype = "float32" min_val = float("-2.68309") @@ -2684,6 +2928,7 @@ class Program_weight_tensor_parameter_243: class Program_weight_tensor_parameter_244: name = "parameter_244" + original_name = "linear_18.w_0" shape = [512, 1536] dtype = "float32" min_val = float("-0.391106") @@ -2695,6 +2940,7 @@ class Program_weight_tensor_parameter_244: class Program_weight_tensor_parameter_245: name = "parameter_245" + original_name = "layer_norm_11.b_0" shape = [512] dtype = "float32" min_val = float("-0.868753") @@ -2706,6 +2952,7 @@ class Program_weight_tensor_parameter_245: class Program_weight_tensor_parameter_246: name = "parameter_246" + original_name = "layer_norm_11.w_0" shape = [512] dtype = "float32" min_val = float("-0.00608563") @@ -2717,6 +2964,7 @@ class Program_weight_tensor_parameter_246: class Program_weight_tensor_parameter_247: name = "parameter_247" + original_name = "linear_17.w_0" shape = [1024, 512] dtype = "float32" min_val = float("-0.255117") @@ -2728,6 +2976,7 @@ class Program_weight_tensor_parameter_247: class Program_weight_tensor_parameter_248: name = "parameter_248" + original_name = "layer_norm_10.b_0" shape = [1024] dtype = "float32" min_val = float("-0.935262") @@ -2739,6 +2988,7 @@ class Program_weight_tensor_parameter_248: class Program_weight_tensor_parameter_249: name = "parameter_249" + original_name = "layer_norm_10.w_0" shape = [1024] dtype = "float32" min_val = float("0.154335") @@ -2750,6 +3000,7 @@ class Program_weight_tensor_parameter_249: class Program_weight_tensor_parameter_250: name = "parameter_250" + original_name = "linear_16.b_0" shape = [256] dtype = "float32" min_val = float("-2.97465") @@ -2761,6 +3012,7 @@ class Program_weight_tensor_parameter_250: class Program_weight_tensor_parameter_251: name = "parameter_251" + original_name = "linear_16.w_0" shape = [1024, 256] dtype = "float32" min_val = float("-0.6238") @@ -2772,6 +3024,7 @@ class Program_weight_tensor_parameter_251: class Program_weight_tensor_parameter_252: name = "parameter_252" + original_name = "linear_15.b_0" shape = [1024] dtype = "float32" min_val = float("-1.40923") @@ -2783,6 +3036,7 @@ class Program_weight_tensor_parameter_252: class Program_weight_tensor_parameter_253: name = "parameter_253" + original_name = "linear_15.w_0" shape = [256, 1024] dtype = "float32" min_val = float("-0.296465") @@ -2794,6 +3048,7 @@ class Program_weight_tensor_parameter_253: class Program_weight_tensor_parameter_254: name = "parameter_254" + original_name = "layer_norm_9.b_0" shape = [256] dtype = "float32" min_val = float("-1.04576") @@ -2805,6 +3060,7 @@ class Program_weight_tensor_parameter_254: class Program_weight_tensor_parameter_255: name = "parameter_255" + original_name = "layer_norm_9.w_0" shape = [256] dtype = "float32" min_val = float("0.237415") @@ -2816,6 +3072,7 @@ class Program_weight_tensor_parameter_255: class Program_weight_tensor_parameter_256: name = "parameter_256" + original_name = "linear_14.b_0" shape = [256] dtype = "float32" min_val = float("-2.0389") @@ -2827,6 +3084,7 @@ class Program_weight_tensor_parameter_256: class Program_weight_tensor_parameter_257: name = "parameter_257" + original_name = "linear_14.w_0" shape = [256, 256] dtype = "float32" min_val = float("-0.212481") @@ -2838,6 +3096,7 @@ class Program_weight_tensor_parameter_257: class Program_weight_tensor_parameter_258: name = "parameter_258" + original_name = "linear_13.b_0" shape = [768] dtype = "float32" min_val = float("-1.52152") @@ -2849,6 +3108,7 @@ class Program_weight_tensor_parameter_258: class Program_weight_tensor_parameter_259: name = "parameter_259" + original_name = "linear_13.w_0" shape = [256, 768] dtype = "float32" min_val = float("-0.457258") @@ -2860,6 +3120,7 @@ class Program_weight_tensor_parameter_259: class Program_weight_tensor_parameter_260: name = "parameter_260" + original_name = "layer_norm_8.b_0" shape = [256] dtype = "float32" min_val = float("-1.00091") @@ -2871,6 +3132,7 @@ class Program_weight_tensor_parameter_260: class Program_weight_tensor_parameter_261: name = "parameter_261" + original_name = "layer_norm_8.w_0" shape = [256] dtype = "float32" min_val = float("0.384928") @@ -2882,6 +3144,7 @@ class Program_weight_tensor_parameter_261: class Program_weight_tensor_parameter_262: name = "parameter_262" + original_name = "linear_12.b_0" shape = [256] dtype = "float32" min_val = float("-1.90921") @@ -2893,6 +3156,7 @@ class Program_weight_tensor_parameter_262: class Program_weight_tensor_parameter_263: name = "parameter_263" + original_name = "linear_12.w_0" shape = [1024, 256] dtype = "float32" min_val = float("-0.358945") @@ -2904,6 +3168,7 @@ class Program_weight_tensor_parameter_263: class Program_weight_tensor_parameter_264: name = "parameter_264" + original_name = "linear_11.b_0" shape = [1024] dtype = "float32" min_val = float("-1.4739") @@ -2915,6 +3180,7 @@ class Program_weight_tensor_parameter_264: class Program_weight_tensor_parameter_265: name = "parameter_265" + original_name = "linear_11.w_0" shape = [256, 1024] dtype = "float32" min_val = float("-0.514122") @@ -2926,6 +3192,7 @@ class Program_weight_tensor_parameter_265: class Program_weight_tensor_parameter_266: name = "parameter_266" + original_name = "layer_norm_7.b_0" shape = [256] dtype = "float32" min_val = float("-0.538291") @@ -2937,6 +3204,7 @@ class Program_weight_tensor_parameter_266: class Program_weight_tensor_parameter_267: name = "parameter_267" + original_name = "layer_norm_7.w_0" shape = [256] dtype = "float32" min_val = float("-0.0020934") @@ -2948,6 +3216,7 @@ class Program_weight_tensor_parameter_267: class Program_weight_tensor_parameter_268: name = "parameter_268" + original_name = "linear_10.b_0" shape = [256] dtype = "float32" min_val = float("-1.61499") @@ -2959,6 +3228,7 @@ class Program_weight_tensor_parameter_268: class Program_weight_tensor_parameter_269: name = "parameter_269" + original_name = "linear_10.w_0" shape = [256, 256] dtype = "float32" min_val = float("-0.294283") @@ -2970,6 +3240,7 @@ class Program_weight_tensor_parameter_269: class Program_weight_tensor_parameter_270: name = "parameter_270" + original_name = "linear_9.b_0" shape = [768] dtype = "float32" min_val = float("-1.83538") @@ -2981,6 +3252,7 @@ class Program_weight_tensor_parameter_270: class Program_weight_tensor_parameter_271: name = "parameter_271" + original_name = "linear_9.w_0" shape = [256, 768] dtype = "float32" min_val = float("-0.387162") @@ -2992,6 +3264,7 @@ class Program_weight_tensor_parameter_271: class Program_weight_tensor_parameter_272: name = "parameter_272" + original_name = "layer_norm_6.b_0" shape = [256] dtype = "float32" min_val = float("-0.343488") @@ -3003,6 +3276,7 @@ class Program_weight_tensor_parameter_272: class Program_weight_tensor_parameter_273: name = "parameter_273" + original_name = "layer_norm_6.w_0" shape = [256] dtype = "float32" min_val = float("0.166703") @@ -3014,6 +3288,7 @@ class Program_weight_tensor_parameter_273: class Program_weight_tensor_parameter_274: name = "parameter_274" + original_name = "linear_8.w_0" shape = [512, 256] dtype = "float32" min_val = float("-0.252898") @@ -3025,6 +3300,7 @@ class Program_weight_tensor_parameter_274: class Program_weight_tensor_parameter_275: name = "parameter_275" + original_name = "layer_norm_5.b_0" shape = [512] dtype = "float32" min_val = float("-1.39532") @@ -3036,6 +3312,7 @@ class Program_weight_tensor_parameter_275: class Program_weight_tensor_parameter_276: name = "parameter_276" + original_name = "layer_norm_5.w_0" shape = [512] dtype = "float32" min_val = float("0.217922") @@ -3047,6 +3324,7 @@ class Program_weight_tensor_parameter_276: class Program_weight_tensor_parameter_277: name = "parameter_277" + original_name = "linear_7.b_0" shape = [128] dtype = "float32" min_val = float("-1.56366") @@ -3058,6 +3336,7 @@ class Program_weight_tensor_parameter_277: class Program_weight_tensor_parameter_278: name = "parameter_278" + original_name = "linear_7.w_0" shape = [512, 128] dtype = "float32" min_val = float("-0.505823") @@ -3069,6 +3348,7 @@ class Program_weight_tensor_parameter_278: class Program_weight_tensor_parameter_279: name = "parameter_279" + original_name = "linear_6.b_0" shape = [512] dtype = "float32" min_val = float("-1.72246") @@ -3080,6 +3360,7 @@ class Program_weight_tensor_parameter_279: class Program_weight_tensor_parameter_280: name = "parameter_280" + original_name = "linear_6.w_0" shape = [128, 512] dtype = "float32" min_val = float("-0.361328") @@ -3091,6 +3372,7 @@ class Program_weight_tensor_parameter_280: class Program_weight_tensor_parameter_281: name = "parameter_281" + original_name = "layer_norm_4.b_0" shape = [128] dtype = "float32" min_val = float("-1.00183") @@ -3102,6 +3384,7 @@ class Program_weight_tensor_parameter_281: class Program_weight_tensor_parameter_282: name = "parameter_282" + original_name = "layer_norm_4.w_0" shape = [128] dtype = "float32" min_val = float("-0.00133267") @@ -3113,6 +3396,7 @@ class Program_weight_tensor_parameter_282: class Program_weight_tensor_parameter_283: name = "parameter_283" + original_name = "linear_5.b_0" shape = [128] dtype = "float32" min_val = float("-0.699082") @@ -3124,6 +3408,7 @@ class Program_weight_tensor_parameter_283: class Program_weight_tensor_parameter_284: name = "parameter_284" + original_name = "linear_5.w_0" shape = [128, 128] dtype = "float32" min_val = float("-0.322705") @@ -3135,6 +3420,7 @@ class Program_weight_tensor_parameter_284: class Program_weight_tensor_parameter_285: name = "parameter_285" + original_name = "linear_4.b_0" shape = [384] dtype = "float32" min_val = float("-1.55027") @@ -3146,6 +3432,7 @@ class Program_weight_tensor_parameter_285: class Program_weight_tensor_parameter_286: name = "parameter_286" + original_name = "linear_4.w_0" shape = [128, 384] dtype = "float32" min_val = float("-0.411067") @@ -3157,6 +3444,7 @@ class Program_weight_tensor_parameter_286: class Program_weight_tensor_parameter_287: name = "parameter_287" + original_name = "layer_norm_3.b_0" shape = [128] dtype = "float32" min_val = float("-1.10826") @@ -3168,6 +3456,7 @@ class Program_weight_tensor_parameter_287: class Program_weight_tensor_parameter_288: name = "parameter_288" + original_name = "layer_norm_3.w_0" shape = [128] dtype = "float32" min_val = float("0.253007") @@ -3179,6 +3468,7 @@ class Program_weight_tensor_parameter_288: class Program_weight_tensor_parameter_289: name = "parameter_289" + original_name = "linear_3.b_0" shape = [128] dtype = "float32" min_val = float("-2.64179") @@ -3190,6 +3480,7 @@ class Program_weight_tensor_parameter_289: class Program_weight_tensor_parameter_290: name = "parameter_290" + original_name = "linear_3.w_0" shape = [512, 128] dtype = "float32" min_val = float("-0.313234") @@ -3201,6 +3492,7 @@ class Program_weight_tensor_parameter_290: class Program_weight_tensor_parameter_291: name = "parameter_291" + original_name = "linear_2.b_0" shape = [512] dtype = "float32" min_val = float("-1.45618") @@ -3212,6 +3504,7 @@ class Program_weight_tensor_parameter_291: class Program_weight_tensor_parameter_292: name = "parameter_292" + original_name = "linear_2.w_0" shape = [128, 512] dtype = "float32" min_val = float("-0.452436") @@ -3223,6 +3516,7 @@ class Program_weight_tensor_parameter_292: class Program_weight_tensor_parameter_293: name = "parameter_293" + original_name = "layer_norm_2.b_0" shape = [128] dtype = "float32" min_val = float("-1.28898") @@ -3234,6 +3528,7 @@ class Program_weight_tensor_parameter_293: class Program_weight_tensor_parameter_294: name = "parameter_294" + original_name = "layer_norm_2.w_0" shape = [128] dtype = "float32" min_val = float("0.357388") @@ -3245,6 +3540,7 @@ class Program_weight_tensor_parameter_294: class Program_weight_tensor_parameter_295: name = "parameter_295" + original_name = "linear_1.b_0" shape = [128] dtype = "float32" min_val = float("-0.852365") @@ -3256,6 +3552,7 @@ class Program_weight_tensor_parameter_295: class Program_weight_tensor_parameter_296: name = "parameter_296" + original_name = "linear_1.w_0" shape = [128, 128] dtype = "float32" min_val = float("-0.289313") @@ -3267,6 +3564,7 @@ class Program_weight_tensor_parameter_296: class Program_weight_tensor_parameter_297: name = "parameter_297" + original_name = "linear_0.b_0" shape = [384] dtype = "float32" min_val = float("-2.00579") @@ -3278,6 +3576,7 @@ class Program_weight_tensor_parameter_297: class Program_weight_tensor_parameter_298: name = "parameter_298" + original_name = "linear_0.w_0" shape = [128, 384] dtype = "float32" min_val = float("-0.512804") @@ -3289,6 +3588,7 @@ class Program_weight_tensor_parameter_298: class Program_weight_tensor_parameter_299: name = "parameter_299" + original_name = "layer_norm_1.b_0" shape = [128] dtype = "float32" min_val = float("-0.85838") @@ -3300,6 +3600,7 @@ class Program_weight_tensor_parameter_299: class Program_weight_tensor_parameter_300: name = "parameter_300" + original_name = "layer_norm_1.w_0" shape = [128] dtype = "float32" min_val = float("-0.00292675") @@ -3311,6 +3612,7 @@ class Program_weight_tensor_parameter_300: class Program_weight_tensor_parameter_301: name = "parameter_301" + original_name = "layer_norm_0.b_0" shape = [128] dtype = "float32" min_val = float("-1.25223") @@ -3322,6 +3624,7 @@ class Program_weight_tensor_parameter_301: class Program_weight_tensor_parameter_302: name = "parameter_302" + original_name = "layer_norm_0.w_0" shape = [128] dtype = "float32" min_val = float("-0.018946") @@ -3333,6 +3636,7 @@ class Program_weight_tensor_parameter_302: class Program_weight_tensor_parameter_303: name = "parameter_303" + original_name = "conv2d_0.b_0" shape = [128] dtype = "float32" min_val = float("-1.14507") @@ -3344,6 +3648,7 @@ class Program_weight_tensor_parameter_303: class Program_weight_tensor_parameter_304: name = "parameter_304" + original_name = "conv2d_0.w_0" shape = [128, 3, 4, 4] dtype = "float32" min_val = float("-0.310082") diff --git a/paddle_samples/PaddleX/SwinTransformer_large_patch4_window12_384/subgraph_1/input_meta.py b/paddle_samples/PaddleX/SwinTransformer_large_patch4_window12_384/subgraph_1/input_meta.py index 5bdcf6106..d3d70e38f 100644 --- a/paddle_samples/PaddleX/SwinTransformer_large_patch4_window12_384/subgraph_1/input_meta.py +++ b/paddle_samples/PaddleX/SwinTransformer_large_patch4_window12_384/subgraph_1/input_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_data_0: name = "data_0" + original_name = "param_0" shape = [529, 6] dtype = "float32" min_val = float("-10.4622") @@ -11,6 +12,7 @@ class Program_weight_tensor_data_0: class Program_weight_tensor_data_1: name = "data_1" + original_name = "param_1" shape = [529, 6] dtype = "float32" min_val = float("-9.03289") @@ -22,6 +24,7 @@ class Program_weight_tensor_data_1: class Program_weight_tensor_data_2: name = "data_2" + original_name = "param_10" shape = [529, 24] dtype = "float32" min_val = float("-10.8818") @@ -33,6 +36,7 @@ class Program_weight_tensor_data_2: class Program_weight_tensor_data_3: name = "data_3" + original_name = "param_11" shape = [529, 24] dtype = "float32" min_val = float("-9.20977") @@ -44,6 +48,7 @@ class Program_weight_tensor_data_3: class Program_weight_tensor_data_4: name = "data_4" + original_name = "param_12" shape = [529, 24] dtype = "float32" min_val = float("-7.68711") @@ -55,6 +60,7 @@ class Program_weight_tensor_data_4: class Program_weight_tensor_data_5: name = "data_5" + original_name = "param_13" shape = [529, 24] dtype = "float32" min_val = float("-12.0255") @@ -66,6 +72,7 @@ class Program_weight_tensor_data_5: class Program_weight_tensor_data_6: name = "data_6" + original_name = "param_14" shape = [529, 24] dtype = "float32" min_val = float("-15.3361") @@ -77,6 +84,7 @@ class Program_weight_tensor_data_6: class Program_weight_tensor_data_7: name = "data_7" + original_name = "param_15" shape = [529, 24] dtype = "float32" min_val = float("-9.30855") @@ -88,6 +96,7 @@ class Program_weight_tensor_data_7: class Program_weight_tensor_data_8: name = "data_8" + original_name = "param_16" shape = [529, 24] dtype = "float32" min_val = float("-11.8211") @@ -99,6 +108,7 @@ class Program_weight_tensor_data_8: class Program_weight_tensor_data_9: name = "data_9" + original_name = "param_17" shape = [529, 24] dtype = "float32" min_val = float("-5.89451") @@ -110,6 +120,7 @@ class Program_weight_tensor_data_9: class Program_weight_tensor_data_10: name = "data_10" + original_name = "param_18" shape = [529, 24] dtype = "float32" min_val = float("-11.3616") @@ -121,6 +132,7 @@ class Program_weight_tensor_data_10: class Program_weight_tensor_data_11: name = "data_11" + original_name = "param_19" shape = [529, 24] dtype = "float32" min_val = float("-13.2805") @@ -132,6 +144,7 @@ class Program_weight_tensor_data_11: class Program_weight_tensor_data_12: name = "data_12" + original_name = "param_2" shape = [529, 12] dtype = "float32" min_val = float("-10.1626") @@ -143,6 +156,7 @@ class Program_weight_tensor_data_12: class Program_weight_tensor_data_13: name = "data_13" + original_name = "param_20" shape = [529, 24] dtype = "float32" min_val = float("-16.8128") @@ -154,6 +168,7 @@ class Program_weight_tensor_data_13: class Program_weight_tensor_data_14: name = "data_14" + original_name = "param_21" shape = [529, 24] dtype = "float32" min_val = float("-13.9336") @@ -165,6 +180,7 @@ class Program_weight_tensor_data_14: class Program_weight_tensor_data_15: name = "data_15" + original_name = "param_22" shape = [529, 48] dtype = "float32" min_val = float("-20.3918") @@ -176,6 +192,7 @@ class Program_weight_tensor_data_15: class Program_weight_tensor_data_16: name = "data_16" + original_name = "param_23" shape = [529, 48] dtype = "float32" min_val = float("-22.1903") @@ -187,6 +204,7 @@ class Program_weight_tensor_data_16: class Program_weight_tensor_data_17: name = "data_17" + original_name = "param_3" shape = [529, 12] dtype = "float32" min_val = float("-8.61095") @@ -198,6 +216,7 @@ class Program_weight_tensor_data_17: class Program_weight_tensor_data_18: name = "data_18" + original_name = "param_4" shape = [529, 24] dtype = "float32" min_val = float("-10.9481") @@ -209,6 +228,7 @@ class Program_weight_tensor_data_18: class Program_weight_tensor_data_19: name = "data_19" + original_name = "param_5" shape = [529, 24] dtype = "float32" min_val = float("-12.1641") @@ -220,6 +240,7 @@ class Program_weight_tensor_data_19: class Program_weight_tensor_data_20: name = "data_20" + original_name = "param_6" shape = [529, 24] dtype = "float32" min_val = float("-12.3952") @@ -231,6 +252,7 @@ class Program_weight_tensor_data_20: class Program_weight_tensor_data_21: name = "data_21" + original_name = "param_7" shape = [529, 24] dtype = "float32" min_val = float("-8.56044") @@ -242,6 +264,7 @@ class Program_weight_tensor_data_21: class Program_weight_tensor_data_22: name = "data_22" + original_name = "param_8" shape = [529, 24] dtype = "float32" min_val = float("-13.0328") @@ -253,6 +276,7 @@ class Program_weight_tensor_data_22: class Program_weight_tensor_data_23: name = "data_23" + original_name = "param_9" shape = [529, 24] dtype = "float32" min_val = float("-13.3212") @@ -264,6 +288,7 @@ class Program_weight_tensor_data_23: class Program_weight_tensor_data_24: name = "data_24" + original_name = "var_0" shape = [4, 3, 384, 384] dtype = "float32" min_val = float("-4.29068") @@ -275,6 +300,7 @@ class Program_weight_tensor_data_24: class Program_weight_tensor_data_25: name = "data_25" + original_name = "var_1021" shape = [144, 144] dtype = "int64" min_val = 0 @@ -284,6 +310,7 @@ class Program_weight_tensor_data_25: class Program_weight_tensor_data_26: name = "data_26" + original_name = "var_1101" shape = [144, 144] dtype = "int64" min_val = 0 @@ -293,6 +320,7 @@ class Program_weight_tensor_data_26: class Program_weight_tensor_data_27: name = "data_27" + original_name = "var_1163" shape = [144, 144] dtype = "int64" min_val = 0 @@ -302,6 +330,7 @@ class Program_weight_tensor_data_27: class Program_weight_tensor_data_28: name = "data_28" + original_name = "var_1243" shape = [144, 144] dtype = "int64" min_val = 0 @@ -311,6 +340,7 @@ class Program_weight_tensor_data_28: class Program_weight_tensor_data_29: name = "data_29" + original_name = "var_1305" shape = [144, 144] dtype = "int64" min_val = 0 @@ -320,6 +350,7 @@ class Program_weight_tensor_data_29: class Program_weight_tensor_data_30: name = "data_30" + original_name = "var_1385" shape = [144, 144] dtype = "int64" min_val = 0 @@ -329,6 +360,7 @@ class Program_weight_tensor_data_30: class Program_weight_tensor_data_31: name = "data_31" + original_name = "var_1447" shape = [144, 144] dtype = "int64" min_val = 0 @@ -338,6 +370,7 @@ class Program_weight_tensor_data_31: class Program_weight_tensor_data_32: name = "data_32" + original_name = "var_1527" shape = [144, 144] dtype = "int64" min_val = 0 @@ -347,6 +380,7 @@ class Program_weight_tensor_data_32: class Program_weight_tensor_data_33: name = "data_33" + original_name = "var_159" shape = [144, 144] dtype = "int64" min_val = 0 @@ -356,6 +390,7 @@ class Program_weight_tensor_data_33: class Program_weight_tensor_data_34: name = "data_34" + original_name = "var_1599" shape = [144, 144] dtype = "int64" min_val = 0 @@ -365,6 +400,7 @@ class Program_weight_tensor_data_34: class Program_weight_tensor_data_35: name = "data_35" + original_name = "var_1679" shape = [144, 144] dtype = "int64" min_val = 0 @@ -374,6 +410,7 @@ class Program_weight_tensor_data_35: class Program_weight_tensor_data_36: name = "data_36" + original_name = "var_21" shape = [144, 144] dtype = "int64" min_val = 0 @@ -383,6 +420,7 @@ class Program_weight_tensor_data_36: class Program_weight_tensor_data_37: name = "data_37" + original_name = "var_239" shape = [144, 144] dtype = "int64" min_val = 0 @@ -392,6 +430,7 @@ class Program_weight_tensor_data_37: class Program_weight_tensor_data_38: name = "data_38" + original_name = "var_311" shape = [144, 144] dtype = "int64" min_val = 0 @@ -401,6 +440,7 @@ class Program_weight_tensor_data_38: class Program_weight_tensor_data_39: name = "data_39" + original_name = "var_391" shape = [144, 144] dtype = "int64" min_val = 0 @@ -410,6 +450,7 @@ class Program_weight_tensor_data_39: class Program_weight_tensor_data_40: name = "data_40" + original_name = "var_453" shape = [144, 144] dtype = "int64" min_val = 0 @@ -419,6 +460,7 @@ class Program_weight_tensor_data_40: class Program_weight_tensor_data_41: name = "data_41" + original_name = "var_533" shape = [144, 144] dtype = "int64" min_val = 0 @@ -428,6 +470,7 @@ class Program_weight_tensor_data_41: class Program_weight_tensor_data_42: name = "data_42" + original_name = "var_595" shape = [144, 144] dtype = "int64" min_val = 0 @@ -437,6 +480,7 @@ class Program_weight_tensor_data_42: class Program_weight_tensor_data_43: name = "data_43" + original_name = "var_675" shape = [144, 144] dtype = "int64" min_val = 0 @@ -446,6 +490,7 @@ class Program_weight_tensor_data_43: class Program_weight_tensor_data_44: name = "data_44" + original_name = "var_737" shape = [144, 144] dtype = "int64" min_val = 0 @@ -455,6 +500,7 @@ class Program_weight_tensor_data_44: class Program_weight_tensor_data_45: name = "data_45" + original_name = "var_817" shape = [144, 144] dtype = "int64" min_val = 0 @@ -464,6 +510,7 @@ class Program_weight_tensor_data_45: class Program_weight_tensor_data_46: name = "data_46" + original_name = "var_87" shape = [144, 144] dtype = "int64" min_val = 0 @@ -473,6 +520,7 @@ class Program_weight_tensor_data_46: class Program_weight_tensor_data_47: name = "data_47" + original_name = "var_879" shape = [144, 144] dtype = "int64" min_val = 0 @@ -482,6 +530,7 @@ class Program_weight_tensor_data_47: class Program_weight_tensor_data_48: name = "data_48" + original_name = "var_959" shape = [144, 144] dtype = "int64" min_val = 0 diff --git a/paddle_samples/PaddleX/SwinTransformer_large_patch4_window12_384/subgraph_1/weight_meta.py b/paddle_samples/PaddleX/SwinTransformer_large_patch4_window12_384/subgraph_1/weight_meta.py index 852762a3f..f005edceb 100644 --- a/paddle_samples/PaddleX/SwinTransformer_large_patch4_window12_384/subgraph_1/weight_meta.py +++ b/paddle_samples/PaddleX/SwinTransformer_large_patch4_window12_384/subgraph_1/weight_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_parameter_0: name = "parameter_0" + original_name = "linear_99.b_0" shape = [102] dtype = "float32" data = None @@ -7,6 +8,7 @@ class Program_weight_tensor_parameter_0: class Program_weight_tensor_parameter_1: name = "parameter_1" + original_name = "linear_99.w_0" shape = [1536, 102] dtype = "float32" min_val = float("-0.0921121") @@ -18,6 +20,7 @@ class Program_weight_tensor_parameter_1: class Program_weight_tensor_parameter_2: name = "parameter_2" + original_name = "layer_norm_52.b_0" shape = [1536] dtype = "float32" min_val = float("-0.716581") @@ -29,6 +32,7 @@ class Program_weight_tensor_parameter_2: class Program_weight_tensor_parameter_3: name = "parameter_3" + original_name = "layer_norm_52.w_0" shape = [1536] dtype = "float32" min_val = float("-0.00244394") @@ -40,6 +44,7 @@ class Program_weight_tensor_parameter_3: class Program_weight_tensor_parameter_4: name = "parameter_4" + original_name = "linear_98.b_0" shape = [1536] dtype = "float32" min_val = float("-5.64898") @@ -51,6 +56,7 @@ class Program_weight_tensor_parameter_4: class Program_weight_tensor_parameter_5: name = "parameter_5" + original_name = "linear_98.w_0" shape = [6144, 1536] dtype = "float32" min_val = float("-3.13181") @@ -62,6 +68,7 @@ class Program_weight_tensor_parameter_5: class Program_weight_tensor_parameter_6: name = "parameter_6" + original_name = "linear_97.b_0" shape = [6144] dtype = "float32" min_val = float("-9.77266") @@ -73,6 +80,7 @@ class Program_weight_tensor_parameter_6: class Program_weight_tensor_parameter_7: name = "parameter_7" + original_name = "linear_97.w_0" shape = [1536, 6144] dtype = "float32" min_val = float("-0.885293") @@ -84,6 +92,7 @@ class Program_weight_tensor_parameter_7: class Program_weight_tensor_parameter_8: name = "parameter_8" + original_name = "layer_norm_51.b_0" shape = [1536] dtype = "float32" min_val = float("-2.86633") @@ -95,6 +104,7 @@ class Program_weight_tensor_parameter_8: class Program_weight_tensor_parameter_9: name = "parameter_9" + original_name = "layer_norm_51.w_0" shape = [1536] dtype = "float32" min_val = float("-0.0162185") @@ -106,6 +116,7 @@ class Program_weight_tensor_parameter_9: class Program_weight_tensor_parameter_10: name = "parameter_10" + original_name = "linear_96.b_0" shape = [1536] dtype = "float32" min_val = float("-23.0972") @@ -117,6 +128,7 @@ class Program_weight_tensor_parameter_10: class Program_weight_tensor_parameter_11: name = "parameter_11" + original_name = "linear_96.w_0" shape = [1536, 1536] dtype = "float32" min_val = float("-1.51103") @@ -128,6 +140,7 @@ class Program_weight_tensor_parameter_11: class Program_weight_tensor_parameter_12: name = "parameter_12" + original_name = "linear_95.b_0" shape = [4608] dtype = "float32" min_val = float("-4.74244") @@ -139,6 +152,7 @@ class Program_weight_tensor_parameter_12: class Program_weight_tensor_parameter_13: name = "parameter_13" + original_name = "linear_95.w_0" shape = [1536, 4608] dtype = "float32" min_val = float("-0.583445") @@ -150,6 +164,7 @@ class Program_weight_tensor_parameter_13: class Program_weight_tensor_parameter_14: name = "parameter_14" + original_name = "layer_norm_50.b_0" shape = [1536] dtype = "float32" min_val = float("-4.94677") @@ -161,6 +176,7 @@ class Program_weight_tensor_parameter_14: class Program_weight_tensor_parameter_15: name = "parameter_15" + original_name = "layer_norm_50.w_0" shape = [1536] dtype = "float32" min_val = float("0.000812601") @@ -172,6 +188,7 @@ class Program_weight_tensor_parameter_15: class Program_weight_tensor_parameter_16: name = "parameter_16" + original_name = "linear_94.b_0" shape = [1536] dtype = "float32" min_val = float("-22.7806") @@ -183,6 +200,7 @@ class Program_weight_tensor_parameter_16: class Program_weight_tensor_parameter_17: name = "parameter_17" + original_name = "linear_94.w_0" shape = [6144, 1536] dtype = "float32" min_val = float("-1.88539") @@ -194,6 +212,7 @@ class Program_weight_tensor_parameter_17: class Program_weight_tensor_parameter_18: name = "parameter_18" + original_name = "linear_93.b_0" shape = [6144] dtype = "float32" min_val = float("-2.30546") @@ -205,6 +224,7 @@ class Program_weight_tensor_parameter_18: class Program_weight_tensor_parameter_19: name = "parameter_19" + original_name = "linear_93.w_0" shape = [1536, 6144] dtype = "float32" min_val = float("-1.22039") @@ -216,6 +236,7 @@ class Program_weight_tensor_parameter_19: class Program_weight_tensor_parameter_20: name = "parameter_20" + original_name = "layer_norm_49.b_0" shape = [1536] dtype = "float32" min_val = float("-6.87468") @@ -227,6 +248,7 @@ class Program_weight_tensor_parameter_20: class Program_weight_tensor_parameter_21: name = "parameter_21" + original_name = "layer_norm_49.w_0" shape = [1536] dtype = "float32" min_val = float("-0.0305479") @@ -238,6 +260,7 @@ class Program_weight_tensor_parameter_21: class Program_weight_tensor_parameter_22: name = "parameter_22" + original_name = "linear_92.b_0" shape = [1536] dtype = "float32" min_val = float("-37.5411") @@ -249,6 +272,7 @@ class Program_weight_tensor_parameter_22: class Program_weight_tensor_parameter_23: name = "parameter_23" + original_name = "linear_92.w_0" shape = [1536, 1536] dtype = "float32" min_val = float("-1.48884") @@ -260,6 +284,7 @@ class Program_weight_tensor_parameter_23: class Program_weight_tensor_parameter_24: name = "parameter_24" + original_name = "linear_91.b_0" shape = [4608] dtype = "float32" min_val = float("-5.23909") @@ -271,6 +296,7 @@ class Program_weight_tensor_parameter_24: class Program_weight_tensor_parameter_25: name = "parameter_25" + original_name = "linear_91.w_0" shape = [1536, 4608] dtype = "float32" min_val = float("-0.596428") @@ -282,6 +308,7 @@ class Program_weight_tensor_parameter_25: class Program_weight_tensor_parameter_26: name = "parameter_26" + original_name = "layer_norm_48.b_0" shape = [1536] dtype = "float32" min_val = float("-3.0933") @@ -293,6 +320,7 @@ class Program_weight_tensor_parameter_26: class Program_weight_tensor_parameter_27: name = "parameter_27" + original_name = "layer_norm_48.w_0" shape = [1536] dtype = "float32" min_val = float("-0.00616813") @@ -304,6 +332,7 @@ class Program_weight_tensor_parameter_27: class Program_weight_tensor_parameter_28: name = "parameter_28" + original_name = "linear_90.w_0" shape = [3072, 1536] dtype = "float32" min_val = float("-1.17477") @@ -315,6 +344,7 @@ class Program_weight_tensor_parameter_28: class Program_weight_tensor_parameter_29: name = "parameter_29" + original_name = "layer_norm_47.b_0" shape = [3072] dtype = "float32" min_val = float("-1.34421") @@ -326,6 +356,7 @@ class Program_weight_tensor_parameter_29: class Program_weight_tensor_parameter_30: name = "parameter_30" + original_name = "layer_norm_47.w_0" shape = [3072] dtype = "float32" min_val = float("0.0492393") @@ -337,6 +368,7 @@ class Program_weight_tensor_parameter_30: class Program_weight_tensor_parameter_31: name = "parameter_31" + original_name = "linear_89.b_0" shape = [768] dtype = "float32" min_val = float("-5.24934") @@ -348,6 +380,7 @@ class Program_weight_tensor_parameter_31: class Program_weight_tensor_parameter_32: name = "parameter_32" + original_name = "linear_89.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-3.21582") @@ -359,6 +392,7 @@ class Program_weight_tensor_parameter_32: class Program_weight_tensor_parameter_33: name = "parameter_33" + original_name = "linear_88.b_0" shape = [3072] dtype = "float32" min_val = float("-2.12648") @@ -370,6 +404,7 @@ class Program_weight_tensor_parameter_33: class Program_weight_tensor_parameter_34: name = "parameter_34" + original_name = "linear_88.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.719778") @@ -381,6 +416,7 @@ class Program_weight_tensor_parameter_34: class Program_weight_tensor_parameter_35: name = "parameter_35" + original_name = "layer_norm_46.b_0" shape = [768] dtype = "float32" min_val = float("-4.05213") @@ -392,6 +428,7 @@ class Program_weight_tensor_parameter_35: class Program_weight_tensor_parameter_36: name = "parameter_36" + original_name = "layer_norm_46.w_0" shape = [768] dtype = "float32" min_val = float("0.470305") @@ -403,6 +440,7 @@ class Program_weight_tensor_parameter_36: class Program_weight_tensor_parameter_37: name = "parameter_37" + original_name = "linear_87.b_0" shape = [768] dtype = "float32" min_val = float("-4.19687") @@ -414,6 +452,7 @@ class Program_weight_tensor_parameter_37: class Program_weight_tensor_parameter_38: name = "parameter_38" + original_name = "linear_87.w_0" shape = [768, 768] dtype = "float32" min_val = float("-2.30984") @@ -425,6 +464,7 @@ class Program_weight_tensor_parameter_38: class Program_weight_tensor_parameter_39: name = "parameter_39" + original_name = "linear_86.b_0" shape = [2304] dtype = "float32" min_val = float("-5.37136") @@ -436,6 +476,7 @@ class Program_weight_tensor_parameter_39: class Program_weight_tensor_parameter_40: name = "parameter_40" + original_name = "linear_86.w_0" shape = [768, 2304] dtype = "float32" min_val = float("-0.568128") @@ -447,6 +488,7 @@ class Program_weight_tensor_parameter_40: class Program_weight_tensor_parameter_41: name = "parameter_41" + original_name = "layer_norm_45.b_0" shape = [768] dtype = "float32" min_val = float("-3.72405") @@ -458,6 +500,7 @@ class Program_weight_tensor_parameter_41: class Program_weight_tensor_parameter_42: name = "parameter_42" + original_name = "layer_norm_45.w_0" shape = [768] dtype = "float32" min_val = float("0.0237492") @@ -469,6 +512,7 @@ class Program_weight_tensor_parameter_42: class Program_weight_tensor_parameter_43: name = "parameter_43" + original_name = "linear_85.b_0" shape = [768] dtype = "float32" min_val = float("-1.72669") @@ -480,6 +524,7 @@ class Program_weight_tensor_parameter_43: class Program_weight_tensor_parameter_44: name = "parameter_44" + original_name = "linear_85.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-2.23477") @@ -491,6 +536,7 @@ class Program_weight_tensor_parameter_44: class Program_weight_tensor_parameter_45: name = "parameter_45" + original_name = "linear_84.b_0" shape = [3072] dtype = "float32" min_val = float("-2.14509") @@ -502,6 +548,7 @@ class Program_weight_tensor_parameter_45: class Program_weight_tensor_parameter_46: name = "parameter_46" + original_name = "linear_84.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.993865") @@ -513,6 +560,7 @@ class Program_weight_tensor_parameter_46: class Program_weight_tensor_parameter_47: name = "parameter_47" + original_name = "layer_norm_44.b_0" shape = [768] dtype = "float32" min_val = float("-5.06618") @@ -524,6 +572,7 @@ class Program_weight_tensor_parameter_47: class Program_weight_tensor_parameter_48: name = "parameter_48" + original_name = "layer_norm_44.w_0" shape = [768] dtype = "float32" min_val = float("0.346082") @@ -535,6 +584,7 @@ class Program_weight_tensor_parameter_48: class Program_weight_tensor_parameter_49: name = "parameter_49" + original_name = "linear_83.b_0" shape = [768] dtype = "float32" min_val = float("-3.13555") @@ -546,6 +596,7 @@ class Program_weight_tensor_parameter_49: class Program_weight_tensor_parameter_50: name = "parameter_50" + original_name = "linear_83.w_0" shape = [768, 768] dtype = "float32" min_val = float("-1.62348") @@ -557,6 +608,7 @@ class Program_weight_tensor_parameter_50: class Program_weight_tensor_parameter_51: name = "parameter_51" + original_name = "linear_82.b_0" shape = [2304] dtype = "float32" min_val = float("-5.25444") @@ -568,6 +620,7 @@ class Program_weight_tensor_parameter_51: class Program_weight_tensor_parameter_52: name = "parameter_52" + original_name = "linear_82.w_0" shape = [768, 2304] dtype = "float32" min_val = float("-0.628736") @@ -579,6 +632,7 @@ class Program_weight_tensor_parameter_52: class Program_weight_tensor_parameter_53: name = "parameter_53" + original_name = "layer_norm_43.b_0" shape = [768] dtype = "float32" min_val = float("-2.83966") @@ -590,6 +644,7 @@ class Program_weight_tensor_parameter_53: class Program_weight_tensor_parameter_54: name = "parameter_54" + original_name = "layer_norm_43.w_0" shape = [768] dtype = "float32" min_val = float("0.0947514") @@ -601,6 +656,7 @@ class Program_weight_tensor_parameter_54: class Program_weight_tensor_parameter_55: name = "parameter_55" + original_name = "linear_81.b_0" shape = [768] dtype = "float32" min_val = float("-1.47697") @@ -612,6 +668,7 @@ class Program_weight_tensor_parameter_55: class Program_weight_tensor_parameter_56: name = "parameter_56" + original_name = "linear_81.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-1.83876") @@ -623,6 +680,7 @@ class Program_weight_tensor_parameter_56: class Program_weight_tensor_parameter_57: name = "parameter_57" + original_name = "linear_80.b_0" shape = [3072] dtype = "float32" min_val = float("-2.18119") @@ -634,6 +692,7 @@ class Program_weight_tensor_parameter_57: class Program_weight_tensor_parameter_58: name = "parameter_58" + original_name = "linear_80.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.591351") @@ -645,6 +704,7 @@ class Program_weight_tensor_parameter_58: class Program_weight_tensor_parameter_59: name = "parameter_59" + original_name = "layer_norm_42.b_0" shape = [768] dtype = "float32" min_val = float("-4.7084") @@ -656,6 +716,7 @@ class Program_weight_tensor_parameter_59: class Program_weight_tensor_parameter_60: name = "parameter_60" + original_name = "layer_norm_42.w_0" shape = [768] dtype = "float32" min_val = float("0.00217702") @@ -667,6 +728,7 @@ class Program_weight_tensor_parameter_60: class Program_weight_tensor_parameter_61: name = "parameter_61" + original_name = "linear_79.b_0" shape = [768] dtype = "float32" min_val = float("-1.99749") @@ -678,6 +740,7 @@ class Program_weight_tensor_parameter_61: class Program_weight_tensor_parameter_62: name = "parameter_62" + original_name = "linear_79.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.803538") @@ -689,6 +752,7 @@ class Program_weight_tensor_parameter_62: class Program_weight_tensor_parameter_63: name = "parameter_63" + original_name = "linear_78.b_0" shape = [2304] dtype = "float32" min_val = float("-5.65047") @@ -700,6 +764,7 @@ class Program_weight_tensor_parameter_63: class Program_weight_tensor_parameter_64: name = "parameter_64" + original_name = "linear_78.w_0" shape = [768, 2304] dtype = "float32" min_val = float("-0.600243") @@ -711,6 +776,7 @@ class Program_weight_tensor_parameter_64: class Program_weight_tensor_parameter_65: name = "parameter_65" + original_name = "layer_norm_41.b_0" shape = [768] dtype = "float32" min_val = float("-2.04346") @@ -722,6 +788,7 @@ class Program_weight_tensor_parameter_65: class Program_weight_tensor_parameter_66: name = "parameter_66" + original_name = "layer_norm_41.w_0" shape = [768] dtype = "float32" min_val = float("0.0121386") @@ -733,6 +800,7 @@ class Program_weight_tensor_parameter_66: class Program_weight_tensor_parameter_67: name = "parameter_67" + original_name = "linear_77.b_0" shape = [768] dtype = "float32" min_val = float("-1.81119") @@ -744,6 +812,7 @@ class Program_weight_tensor_parameter_67: class Program_weight_tensor_parameter_68: name = "parameter_68" + original_name = "linear_77.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-0.752024") @@ -755,6 +824,7 @@ class Program_weight_tensor_parameter_68: class Program_weight_tensor_parameter_69: name = "parameter_69" + original_name = "linear_76.b_0" shape = [3072] dtype = "float32" min_val = float("-2.10917") @@ -766,6 +836,7 @@ class Program_weight_tensor_parameter_69: class Program_weight_tensor_parameter_70: name = "parameter_70" + original_name = "linear_76.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.511362") @@ -777,6 +848,7 @@ class Program_weight_tensor_parameter_70: class Program_weight_tensor_parameter_71: name = "parameter_71" + original_name = "layer_norm_40.b_0" shape = [768] dtype = "float32" min_val = float("-3.90636") @@ -788,6 +860,7 @@ class Program_weight_tensor_parameter_71: class Program_weight_tensor_parameter_72: name = "parameter_72" + original_name = "layer_norm_40.w_0" shape = [768] dtype = "float32" min_val = float("-0.0278045") @@ -799,6 +872,7 @@ class Program_weight_tensor_parameter_72: class Program_weight_tensor_parameter_73: name = "parameter_73" + original_name = "linear_75.b_0" shape = [768] dtype = "float32" min_val = float("-3.51185") @@ -810,6 +884,7 @@ class Program_weight_tensor_parameter_73: class Program_weight_tensor_parameter_74: name = "parameter_74" + original_name = "linear_75.w_0" shape = [768, 768] dtype = "float32" min_val = float("-1.57004") @@ -821,6 +896,7 @@ class Program_weight_tensor_parameter_74: class Program_weight_tensor_parameter_75: name = "parameter_75" + original_name = "linear_74.b_0" shape = [2304] dtype = "float32" min_val = float("-4.51336") @@ -832,6 +908,7 @@ class Program_weight_tensor_parameter_75: class Program_weight_tensor_parameter_76: name = "parameter_76" + original_name = "linear_74.w_0" shape = [768, 2304] dtype = "float32" min_val = float("-0.577265") @@ -843,6 +920,7 @@ class Program_weight_tensor_parameter_76: class Program_weight_tensor_parameter_77: name = "parameter_77" + original_name = "layer_norm_39.b_0" shape = [768] dtype = "float32" min_val = float("-1.61055") @@ -854,6 +932,7 @@ class Program_weight_tensor_parameter_77: class Program_weight_tensor_parameter_78: name = "parameter_78" + original_name = "layer_norm_39.w_0" shape = [768] dtype = "float32" min_val = float("0.154485") @@ -865,6 +944,7 @@ class Program_weight_tensor_parameter_78: class Program_weight_tensor_parameter_79: name = "parameter_79" + original_name = "linear_73.b_0" shape = [768] dtype = "float32" min_val = float("-3.10472") @@ -876,6 +956,7 @@ class Program_weight_tensor_parameter_79: class Program_weight_tensor_parameter_80: name = "parameter_80" + original_name = "linear_73.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-0.530554") @@ -887,6 +968,7 @@ class Program_weight_tensor_parameter_80: class Program_weight_tensor_parameter_81: name = "parameter_81" + original_name = "linear_72.b_0" shape = [3072] dtype = "float32" min_val = float("-2.23439") @@ -898,6 +980,7 @@ class Program_weight_tensor_parameter_81: class Program_weight_tensor_parameter_82: name = "parameter_82" + original_name = "linear_72.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-1.29977") @@ -909,6 +992,7 @@ class Program_weight_tensor_parameter_82: class Program_weight_tensor_parameter_83: name = "parameter_83" + original_name = "layer_norm_38.b_0" shape = [768] dtype = "float32" min_val = float("-2.9066") @@ -920,6 +1004,7 @@ class Program_weight_tensor_parameter_83: class Program_weight_tensor_parameter_84: name = "parameter_84" + original_name = "layer_norm_38.w_0" shape = [768] dtype = "float32" min_val = float("-0.0777251") @@ -931,6 +1016,7 @@ class Program_weight_tensor_parameter_84: class Program_weight_tensor_parameter_85: name = "parameter_85" + original_name = "linear_71.b_0" shape = [768] dtype = "float32" min_val = float("-1.99945") @@ -942,6 +1028,7 @@ class Program_weight_tensor_parameter_85: class Program_weight_tensor_parameter_86: name = "parameter_86" + original_name = "linear_71.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.837059") @@ -953,6 +1040,7 @@ class Program_weight_tensor_parameter_86: class Program_weight_tensor_parameter_87: name = "parameter_87" + original_name = "linear_70.b_0" shape = [2304] dtype = "float32" min_val = float("-4.94734") @@ -964,6 +1052,7 @@ class Program_weight_tensor_parameter_87: class Program_weight_tensor_parameter_88: name = "parameter_88" + original_name = "linear_70.w_0" shape = [768, 2304] dtype = "float32" min_val = float("-0.598978") @@ -975,6 +1064,7 @@ class Program_weight_tensor_parameter_88: class Program_weight_tensor_parameter_89: name = "parameter_89" + original_name = "layer_norm_37.b_0" shape = [768] dtype = "float32" min_val = float("-2.58423") @@ -986,6 +1076,7 @@ class Program_weight_tensor_parameter_89: class Program_weight_tensor_parameter_90: name = "parameter_90" + original_name = "layer_norm_37.w_0" shape = [768] dtype = "float32" min_val = float("0.138124") @@ -997,6 +1088,7 @@ class Program_weight_tensor_parameter_90: class Program_weight_tensor_parameter_91: name = "parameter_91" + original_name = "linear_69.b_0" shape = [768] dtype = "float32" min_val = float("-4.30309") @@ -1008,6 +1100,7 @@ class Program_weight_tensor_parameter_91: class Program_weight_tensor_parameter_92: name = "parameter_92" + original_name = "linear_69.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-0.935346") @@ -1019,6 +1112,7 @@ class Program_weight_tensor_parameter_92: class Program_weight_tensor_parameter_93: name = "parameter_93" + original_name = "linear_68.b_0" shape = [3072] dtype = "float32" min_val = float("-2.18108") @@ -1030,6 +1124,7 @@ class Program_weight_tensor_parameter_93: class Program_weight_tensor_parameter_94: name = "parameter_94" + original_name = "linear_68.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.740791") @@ -1041,6 +1136,7 @@ class Program_weight_tensor_parameter_94: class Program_weight_tensor_parameter_95: name = "parameter_95" + original_name = "layer_norm_36.b_0" shape = [768] dtype = "float32" min_val = float("-1.77707") @@ -1052,6 +1148,7 @@ class Program_weight_tensor_parameter_95: class Program_weight_tensor_parameter_96: name = "parameter_96" + original_name = "layer_norm_36.w_0" shape = [768] dtype = "float32" min_val = float("-0.0403654") @@ -1063,6 +1160,7 @@ class Program_weight_tensor_parameter_96: class Program_weight_tensor_parameter_97: name = "parameter_97" + original_name = "linear_67.b_0" shape = [768] dtype = "float32" min_val = float("-2.74302") @@ -1074,6 +1172,7 @@ class Program_weight_tensor_parameter_97: class Program_weight_tensor_parameter_98: name = "parameter_98" + original_name = "linear_67.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.607407") @@ -1085,6 +1184,7 @@ class Program_weight_tensor_parameter_98: class Program_weight_tensor_parameter_99: name = "parameter_99" + original_name = "linear_66.b_0" shape = [2304] dtype = "float32" min_val = float("-3.84958") @@ -1096,6 +1196,7 @@ class Program_weight_tensor_parameter_99: class Program_weight_tensor_parameter_100: name = "parameter_100" + original_name = "linear_66.w_0" shape = [768, 2304] dtype = "float32" min_val = float("-0.552831") @@ -1107,6 +1208,7 @@ class Program_weight_tensor_parameter_100: class Program_weight_tensor_parameter_101: name = "parameter_101" + original_name = "layer_norm_35.b_0" shape = [768] dtype = "float32" min_val = float("-3.04251") @@ -1118,6 +1220,7 @@ class Program_weight_tensor_parameter_101: class Program_weight_tensor_parameter_102: name = "parameter_102" + original_name = "layer_norm_35.w_0" shape = [768] dtype = "float32" min_val = float("0.0779379") @@ -1129,6 +1232,7 @@ class Program_weight_tensor_parameter_102: class Program_weight_tensor_parameter_103: name = "parameter_103" + original_name = "linear_65.b_0" shape = [768] dtype = "float32" min_val = float("-4.68522") @@ -1140,6 +1244,7 @@ class Program_weight_tensor_parameter_103: class Program_weight_tensor_parameter_104: name = "parameter_104" + original_name = "linear_65.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-1.14152") @@ -1151,6 +1256,7 @@ class Program_weight_tensor_parameter_104: class Program_weight_tensor_parameter_105: name = "parameter_105" + original_name = "linear_64.b_0" shape = [3072] dtype = "float32" min_val = float("-2.45307") @@ -1162,6 +1268,7 @@ class Program_weight_tensor_parameter_105: class Program_weight_tensor_parameter_106: name = "parameter_106" + original_name = "linear_64.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.420597") @@ -1173,6 +1280,7 @@ class Program_weight_tensor_parameter_106: class Program_weight_tensor_parameter_107: name = "parameter_107" + original_name = "layer_norm_34.b_0" shape = [768] dtype = "float32" min_val = float("-1.68546") @@ -1184,6 +1292,7 @@ class Program_weight_tensor_parameter_107: class Program_weight_tensor_parameter_108: name = "parameter_108" + original_name = "layer_norm_34.w_0" shape = [768] dtype = "float32" min_val = float("0.00270205") @@ -1195,6 +1304,7 @@ class Program_weight_tensor_parameter_108: class Program_weight_tensor_parameter_109: name = "parameter_109" + original_name = "linear_63.b_0" shape = [768] dtype = "float32" min_val = float("-1.59992") @@ -1206,6 +1316,7 @@ class Program_weight_tensor_parameter_109: class Program_weight_tensor_parameter_110: name = "parameter_110" + original_name = "linear_63.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.434936") @@ -1217,6 +1328,7 @@ class Program_weight_tensor_parameter_110: class Program_weight_tensor_parameter_111: name = "parameter_111" + original_name = "linear_62.b_0" shape = [2304] dtype = "float32" min_val = float("-3.34384") @@ -1228,6 +1340,7 @@ class Program_weight_tensor_parameter_111: class Program_weight_tensor_parameter_112: name = "parameter_112" + original_name = "linear_62.w_0" shape = [768, 2304] dtype = "float32" min_val = float("-0.584971") @@ -1239,6 +1352,7 @@ class Program_weight_tensor_parameter_112: class Program_weight_tensor_parameter_113: name = "parameter_113" + original_name = "layer_norm_33.b_0" shape = [768] dtype = "float32" min_val = float("-3.36681") @@ -1250,6 +1364,7 @@ class Program_weight_tensor_parameter_113: class Program_weight_tensor_parameter_114: name = "parameter_114" + original_name = "layer_norm_33.w_0" shape = [768] dtype = "float32" min_val = float("0.143708") @@ -1261,6 +1376,7 @@ class Program_weight_tensor_parameter_114: class Program_weight_tensor_parameter_115: name = "parameter_115" + original_name = "linear_61.b_0" shape = [768] dtype = "float32" min_val = float("-4.88005") @@ -1272,6 +1388,7 @@ class Program_weight_tensor_parameter_115: class Program_weight_tensor_parameter_116: name = "parameter_116" + original_name = "linear_61.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-1.00052") @@ -1283,6 +1400,7 @@ class Program_weight_tensor_parameter_116: class Program_weight_tensor_parameter_117: name = "parameter_117" + original_name = "linear_60.b_0" shape = [3072] dtype = "float32" min_val = float("-2.8426") @@ -1294,6 +1412,7 @@ class Program_weight_tensor_parameter_117: class Program_weight_tensor_parameter_118: name = "parameter_118" + original_name = "linear_60.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.72303") @@ -1305,6 +1424,7 @@ class Program_weight_tensor_parameter_118: class Program_weight_tensor_parameter_119: name = "parameter_119" + original_name = "layer_norm_32.b_0" shape = [768] dtype = "float32" min_val = float("-1.73376") @@ -1316,6 +1436,7 @@ class Program_weight_tensor_parameter_119: class Program_weight_tensor_parameter_120: name = "parameter_120" + original_name = "layer_norm_32.w_0" shape = [768] dtype = "float32" min_val = float("-0.000380825") @@ -1327,6 +1448,7 @@ class Program_weight_tensor_parameter_120: class Program_weight_tensor_parameter_121: name = "parameter_121" + original_name = "linear_59.b_0" shape = [768] dtype = "float32" min_val = float("-1.86918") @@ -1338,6 +1460,7 @@ class Program_weight_tensor_parameter_121: class Program_weight_tensor_parameter_122: name = "parameter_122" + original_name = "linear_59.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.595401") @@ -1349,6 +1472,7 @@ class Program_weight_tensor_parameter_122: class Program_weight_tensor_parameter_123: name = "parameter_123" + original_name = "linear_58.b_0" shape = [2304] dtype = "float32" min_val = float("-4.92381") @@ -1360,6 +1484,7 @@ class Program_weight_tensor_parameter_123: class Program_weight_tensor_parameter_124: name = "parameter_124" + original_name = "linear_58.w_0" shape = [768, 2304] dtype = "float32" min_val = float("-0.58939") @@ -1371,6 +1496,7 @@ class Program_weight_tensor_parameter_124: class Program_weight_tensor_parameter_125: name = "parameter_125" + original_name = "layer_norm_31.b_0" shape = [768] dtype = "float32" min_val = float("-3.32971") @@ -1382,6 +1508,7 @@ class Program_weight_tensor_parameter_125: class Program_weight_tensor_parameter_126: name = "parameter_126" + original_name = "layer_norm_31.w_0" shape = [768] dtype = "float32" min_val = float("0.0118237") @@ -1393,6 +1520,7 @@ class Program_weight_tensor_parameter_126: class Program_weight_tensor_parameter_127: name = "parameter_127" + original_name = "linear_57.b_0" shape = [768] dtype = "float32" min_val = float("-2.51702") @@ -1404,6 +1532,7 @@ class Program_weight_tensor_parameter_127: class Program_weight_tensor_parameter_128: name = "parameter_128" + original_name = "linear_57.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-0.911382") @@ -1415,6 +1544,7 @@ class Program_weight_tensor_parameter_128: class Program_weight_tensor_parameter_129: name = "parameter_129" + original_name = "linear_56.b_0" shape = [3072] dtype = "float32" min_val = float("-3.35308") @@ -1426,6 +1556,7 @@ class Program_weight_tensor_parameter_129: class Program_weight_tensor_parameter_130: name = "parameter_130" + original_name = "linear_56.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.588494") @@ -1437,6 +1568,7 @@ class Program_weight_tensor_parameter_130: class Program_weight_tensor_parameter_131: name = "parameter_131" + original_name = "layer_norm_30.b_0" shape = [768] dtype = "float32" min_val = float("-1.7974") @@ -1448,6 +1580,7 @@ class Program_weight_tensor_parameter_131: class Program_weight_tensor_parameter_132: name = "parameter_132" + original_name = "layer_norm_30.w_0" shape = [768] dtype = "float32" min_val = float("-0.00010268") @@ -1459,6 +1592,7 @@ class Program_weight_tensor_parameter_132: class Program_weight_tensor_parameter_133: name = "parameter_133" + original_name = "linear_55.b_0" shape = [768] dtype = "float32" min_val = float("-1.02682") @@ -1470,6 +1604,7 @@ class Program_weight_tensor_parameter_133: class Program_weight_tensor_parameter_134: name = "parameter_134" + original_name = "linear_55.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.69856") @@ -1481,6 +1616,7 @@ class Program_weight_tensor_parameter_134: class Program_weight_tensor_parameter_135: name = "parameter_135" + original_name = "linear_54.b_0" shape = [2304] dtype = "float32" min_val = float("-5.09908") @@ -1492,6 +1628,7 @@ class Program_weight_tensor_parameter_135: class Program_weight_tensor_parameter_136: name = "parameter_136" + original_name = "linear_54.w_0" shape = [768, 2304] dtype = "float32" min_val = float("-0.696862") @@ -1503,6 +1640,7 @@ class Program_weight_tensor_parameter_136: class Program_weight_tensor_parameter_137: name = "parameter_137" + original_name = "layer_norm_29.b_0" shape = [768] dtype = "float32" min_val = float("-3.25767") @@ -1514,6 +1652,7 @@ class Program_weight_tensor_parameter_137: class Program_weight_tensor_parameter_138: name = "parameter_138" + original_name = "layer_norm_29.w_0" shape = [768] dtype = "float32" min_val = float("-0.0241309") @@ -1525,6 +1664,7 @@ class Program_weight_tensor_parameter_138: class Program_weight_tensor_parameter_139: name = "parameter_139" + original_name = "linear_53.b_0" shape = [768] dtype = "float32" min_val = float("-1.43559") @@ -1536,6 +1676,7 @@ class Program_weight_tensor_parameter_139: class Program_weight_tensor_parameter_140: name = "parameter_140" + original_name = "linear_53.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-0.969465") @@ -1547,6 +1688,7 @@ class Program_weight_tensor_parameter_140: class Program_weight_tensor_parameter_141: name = "parameter_141" + original_name = "linear_52.b_0" shape = [3072] dtype = "float32" min_val = float("-4.18946") @@ -1558,6 +1700,7 @@ class Program_weight_tensor_parameter_141: class Program_weight_tensor_parameter_142: name = "parameter_142" + original_name = "linear_52.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.509378") @@ -1569,6 +1712,7 @@ class Program_weight_tensor_parameter_142: class Program_weight_tensor_parameter_143: name = "parameter_143" + original_name = "layer_norm_28.b_0" shape = [768] dtype = "float32" min_val = float("-1.72607") @@ -1580,6 +1724,7 @@ class Program_weight_tensor_parameter_143: class Program_weight_tensor_parameter_144: name = "parameter_144" + original_name = "layer_norm_28.w_0" shape = [768] dtype = "float32" min_val = float("-0.00155866") @@ -1591,6 +1736,7 @@ class Program_weight_tensor_parameter_144: class Program_weight_tensor_parameter_145: name = "parameter_145" + original_name = "linear_51.b_0" shape = [768] dtype = "float32" min_val = float("-1.77895") @@ -1602,6 +1748,7 @@ class Program_weight_tensor_parameter_145: class Program_weight_tensor_parameter_146: name = "parameter_146" + original_name = "linear_51.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.728201") @@ -1613,6 +1760,7 @@ class Program_weight_tensor_parameter_146: class Program_weight_tensor_parameter_147: name = "parameter_147" + original_name = "linear_50.b_0" shape = [2304] dtype = "float32" min_val = float("-3.99468") @@ -1624,6 +1772,7 @@ class Program_weight_tensor_parameter_147: class Program_weight_tensor_parameter_148: name = "parameter_148" + original_name = "linear_50.w_0" shape = [768, 2304] dtype = "float32" min_val = float("-0.656733") @@ -1635,6 +1784,7 @@ class Program_weight_tensor_parameter_148: class Program_weight_tensor_parameter_149: name = "parameter_149" + original_name = "layer_norm_27.b_0" shape = [768] dtype = "float32" min_val = float("-3.15767") @@ -1646,6 +1796,7 @@ class Program_weight_tensor_parameter_149: class Program_weight_tensor_parameter_150: name = "parameter_150" + original_name = "layer_norm_27.w_0" shape = [768] dtype = "float32" min_val = float("0.106784") @@ -1657,6 +1808,7 @@ class Program_weight_tensor_parameter_150: class Program_weight_tensor_parameter_151: name = "parameter_151" + original_name = "linear_49.b_0" shape = [768] dtype = "float32" min_val = float("-1.0326") @@ -1668,6 +1820,7 @@ class Program_weight_tensor_parameter_151: class Program_weight_tensor_parameter_152: name = "parameter_152" + original_name = "linear_49.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-0.95433") @@ -1679,6 +1832,7 @@ class Program_weight_tensor_parameter_152: class Program_weight_tensor_parameter_153: name = "parameter_153" + original_name = "linear_48.b_0" shape = [3072] dtype = "float32" min_val = float("-3.58638") @@ -1690,6 +1844,7 @@ class Program_weight_tensor_parameter_153: class Program_weight_tensor_parameter_154: name = "parameter_154" + original_name = "linear_48.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.564804") @@ -1701,6 +1856,7 @@ class Program_weight_tensor_parameter_154: class Program_weight_tensor_parameter_155: name = "parameter_155" + original_name = "layer_norm_26.b_0" shape = [768] dtype = "float32" min_val = float("-1.38338") @@ -1712,6 +1868,7 @@ class Program_weight_tensor_parameter_155: class Program_weight_tensor_parameter_156: name = "parameter_156" + original_name = "layer_norm_26.w_0" shape = [768] dtype = "float32" min_val = float("0.00136458") @@ -1723,6 +1880,7 @@ class Program_weight_tensor_parameter_156: class Program_weight_tensor_parameter_157: name = "parameter_157" + original_name = "linear_47.b_0" shape = [768] dtype = "float32" min_val = float("-1.38308") @@ -1734,6 +1892,7 @@ class Program_weight_tensor_parameter_157: class Program_weight_tensor_parameter_158: name = "parameter_158" + original_name = "linear_47.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.52797") @@ -1745,6 +1904,7 @@ class Program_weight_tensor_parameter_158: class Program_weight_tensor_parameter_159: name = "parameter_159" + original_name = "linear_46.b_0" shape = [2304] dtype = "float32" min_val = float("-3.65044") @@ -1756,6 +1916,7 @@ class Program_weight_tensor_parameter_159: class Program_weight_tensor_parameter_160: name = "parameter_160" + original_name = "linear_46.w_0" shape = [768, 2304] dtype = "float32" min_val = float("-0.631119") @@ -1767,6 +1928,7 @@ class Program_weight_tensor_parameter_160: class Program_weight_tensor_parameter_161: name = "parameter_161" + original_name = "layer_norm_25.b_0" shape = [768] dtype = "float32" min_val = float("-2.59769") @@ -1778,6 +1940,7 @@ class Program_weight_tensor_parameter_161: class Program_weight_tensor_parameter_162: name = "parameter_162" + original_name = "layer_norm_25.w_0" shape = [768] dtype = "float32" min_val = float("0.0312813") @@ -1789,6 +1952,7 @@ class Program_weight_tensor_parameter_162: class Program_weight_tensor_parameter_163: name = "parameter_163" + original_name = "linear_45.b_0" shape = [768] dtype = "float32" min_val = float("-1.08458") @@ -1800,6 +1964,7 @@ class Program_weight_tensor_parameter_163: class Program_weight_tensor_parameter_164: name = "parameter_164" + original_name = "linear_45.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-0.66015") @@ -1811,6 +1976,7 @@ class Program_weight_tensor_parameter_164: class Program_weight_tensor_parameter_165: name = "parameter_165" + original_name = "linear_44.b_0" shape = [3072] dtype = "float32" min_val = float("-3.28796") @@ -1822,6 +1988,7 @@ class Program_weight_tensor_parameter_165: class Program_weight_tensor_parameter_166: name = "parameter_166" + original_name = "linear_44.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.495555") @@ -1833,6 +2000,7 @@ class Program_weight_tensor_parameter_166: class Program_weight_tensor_parameter_167: name = "parameter_167" + original_name = "layer_norm_24.b_0" shape = [768] dtype = "float32" min_val = float("-1.56072") @@ -1844,6 +2012,7 @@ class Program_weight_tensor_parameter_167: class Program_weight_tensor_parameter_168: name = "parameter_168" + original_name = "layer_norm_24.w_0" shape = [768] dtype = "float32" min_val = float("0.015498") @@ -1855,6 +2024,7 @@ class Program_weight_tensor_parameter_168: class Program_weight_tensor_parameter_169: name = "parameter_169" + original_name = "linear_43.b_0" shape = [768] dtype = "float32" min_val = float("-1.43805") @@ -1866,6 +2036,7 @@ class Program_weight_tensor_parameter_169: class Program_weight_tensor_parameter_170: name = "parameter_170" + original_name = "linear_43.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.629227") @@ -1877,6 +2048,7 @@ class Program_weight_tensor_parameter_170: class Program_weight_tensor_parameter_171: name = "parameter_171" + original_name = "linear_42.b_0" shape = [2304] dtype = "float32" min_val = float("-3.43687") @@ -1888,6 +2060,7 @@ class Program_weight_tensor_parameter_171: class Program_weight_tensor_parameter_172: name = "parameter_172" + original_name = "linear_42.w_0" shape = [768, 2304] dtype = "float32" min_val = float("-0.59105") @@ -1899,6 +2072,7 @@ class Program_weight_tensor_parameter_172: class Program_weight_tensor_parameter_173: name = "parameter_173" + original_name = "layer_norm_23.b_0" shape = [768] dtype = "float32" min_val = float("-2.32193") @@ -1910,6 +2084,7 @@ class Program_weight_tensor_parameter_173: class Program_weight_tensor_parameter_174: name = "parameter_174" + original_name = "layer_norm_23.w_0" shape = [768] dtype = "float32" min_val = float("0.0336639") @@ -1921,6 +2096,7 @@ class Program_weight_tensor_parameter_174: class Program_weight_tensor_parameter_175: name = "parameter_175" + original_name = "linear_41.b_0" shape = [768] dtype = "float32" min_val = float("-0.922227") @@ -1932,6 +2108,7 @@ class Program_weight_tensor_parameter_175: class Program_weight_tensor_parameter_176: name = "parameter_176" + original_name = "linear_41.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-0.617216") @@ -1943,6 +2120,7 @@ class Program_weight_tensor_parameter_176: class Program_weight_tensor_parameter_177: name = "parameter_177" + original_name = "linear_40.b_0" shape = [3072] dtype = "float32" min_val = float("-3.43824") @@ -1954,6 +2132,7 @@ class Program_weight_tensor_parameter_177: class Program_weight_tensor_parameter_178: name = "parameter_178" + original_name = "linear_40.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.725919") @@ -1965,6 +2144,7 @@ class Program_weight_tensor_parameter_178: class Program_weight_tensor_parameter_179: name = "parameter_179" + original_name = "layer_norm_22.b_0" shape = [768] dtype = "float32" min_val = float("-1.8907") @@ -1976,6 +2156,7 @@ class Program_weight_tensor_parameter_179: class Program_weight_tensor_parameter_180: name = "parameter_180" + original_name = "layer_norm_22.w_0" shape = [768] dtype = "float32" min_val = float("0.0160568") @@ -1987,6 +2168,7 @@ class Program_weight_tensor_parameter_180: class Program_weight_tensor_parameter_181: name = "parameter_181" + original_name = "linear_39.b_0" shape = [768] dtype = "float32" min_val = float("-1.45785") @@ -1998,6 +2180,7 @@ class Program_weight_tensor_parameter_181: class Program_weight_tensor_parameter_182: name = "parameter_182" + original_name = "linear_39.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.601762") @@ -2009,6 +2192,7 @@ class Program_weight_tensor_parameter_182: class Program_weight_tensor_parameter_183: name = "parameter_183" + original_name = "linear_38.b_0" shape = [2304] dtype = "float32" min_val = float("-3.35239") @@ -2020,6 +2204,7 @@ class Program_weight_tensor_parameter_183: class Program_weight_tensor_parameter_184: name = "parameter_184" + original_name = "linear_38.w_0" shape = [768, 2304] dtype = "float32" min_val = float("-0.790326") @@ -2031,6 +2216,7 @@ class Program_weight_tensor_parameter_184: class Program_weight_tensor_parameter_185: name = "parameter_185" + original_name = "layer_norm_21.b_0" shape = [768] dtype = "float32" min_val = float("-2.41981") @@ -2042,6 +2228,7 @@ class Program_weight_tensor_parameter_185: class Program_weight_tensor_parameter_186: name = "parameter_186" + original_name = "layer_norm_21.w_0" shape = [768] dtype = "float32" min_val = float("-0.00740415") @@ -2053,6 +2240,7 @@ class Program_weight_tensor_parameter_186: class Program_weight_tensor_parameter_187: name = "parameter_187" + original_name = "linear_37.b_0" shape = [768] dtype = "float32" min_val = float("-0.966731") @@ -2064,6 +2252,7 @@ class Program_weight_tensor_parameter_187: class Program_weight_tensor_parameter_188: name = "parameter_188" + original_name = "linear_37.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-0.836169") @@ -2075,6 +2264,7 @@ class Program_weight_tensor_parameter_188: class Program_weight_tensor_parameter_189: name = "parameter_189" + original_name = "linear_36.b_0" shape = [3072] dtype = "float32" min_val = float("-3.52899") @@ -2086,6 +2276,7 @@ class Program_weight_tensor_parameter_189: class Program_weight_tensor_parameter_190: name = "parameter_190" + original_name = "linear_36.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.569905") @@ -2097,6 +2288,7 @@ class Program_weight_tensor_parameter_190: class Program_weight_tensor_parameter_191: name = "parameter_191" + original_name = "layer_norm_20.b_0" shape = [768] dtype = "float32" min_val = float("-1.83979") @@ -2108,6 +2300,7 @@ class Program_weight_tensor_parameter_191: class Program_weight_tensor_parameter_192: name = "parameter_192" + original_name = "layer_norm_20.w_0" shape = [768] dtype = "float32" min_val = float("-0.00426636") @@ -2119,6 +2312,7 @@ class Program_weight_tensor_parameter_192: class Program_weight_tensor_parameter_193: name = "parameter_193" + original_name = "linear_35.b_0" shape = [768] dtype = "float32" min_val = float("-1.65689") @@ -2130,6 +2324,7 @@ class Program_weight_tensor_parameter_193: class Program_weight_tensor_parameter_194: name = "parameter_194" + original_name = "linear_35.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.772793") @@ -2141,6 +2336,7 @@ class Program_weight_tensor_parameter_194: class Program_weight_tensor_parameter_195: name = "parameter_195" + original_name = "linear_34.b_0" shape = [2304] dtype = "float32" min_val = float("-4.24147") @@ -2152,6 +2348,7 @@ class Program_weight_tensor_parameter_195: class Program_weight_tensor_parameter_196: name = "parameter_196" + original_name = "linear_34.w_0" shape = [768, 2304] dtype = "float32" min_val = float("-0.936083") @@ -2163,6 +2360,7 @@ class Program_weight_tensor_parameter_196: class Program_weight_tensor_parameter_197: name = "parameter_197" + original_name = "layer_norm_19.b_0" shape = [768] dtype = "float32" min_val = float("-2.56759") @@ -2174,6 +2372,7 @@ class Program_weight_tensor_parameter_197: class Program_weight_tensor_parameter_198: name = "parameter_198" + original_name = "layer_norm_19.w_0" shape = [768] dtype = "float32" min_val = float("0.0035046") @@ -2185,6 +2384,7 @@ class Program_weight_tensor_parameter_198: class Program_weight_tensor_parameter_199: name = "parameter_199" + original_name = "linear_33.b_0" shape = [768] dtype = "float32" min_val = float("-1.31184") @@ -2196,6 +2396,7 @@ class Program_weight_tensor_parameter_199: class Program_weight_tensor_parameter_200: name = "parameter_200" + original_name = "linear_33.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-0.599418") @@ -2207,6 +2408,7 @@ class Program_weight_tensor_parameter_200: class Program_weight_tensor_parameter_201: name = "parameter_201" + original_name = "linear_32.b_0" shape = [3072] dtype = "float32" min_val = float("-3.68772") @@ -2218,6 +2420,7 @@ class Program_weight_tensor_parameter_201: class Program_weight_tensor_parameter_202: name = "parameter_202" + original_name = "linear_32.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.506968") @@ -2229,6 +2432,7 @@ class Program_weight_tensor_parameter_202: class Program_weight_tensor_parameter_203: name = "parameter_203" + original_name = "layer_norm_18.b_0" shape = [768] dtype = "float32" min_val = float("-1.84777") @@ -2240,6 +2444,7 @@ class Program_weight_tensor_parameter_203: class Program_weight_tensor_parameter_204: name = "parameter_204" + original_name = "layer_norm_18.w_0" shape = [768] dtype = "float32" min_val = float("-0.00319528") @@ -2251,6 +2456,7 @@ class Program_weight_tensor_parameter_204: class Program_weight_tensor_parameter_205: name = "parameter_205" + original_name = "linear_31.b_0" shape = [768] dtype = "float32" min_val = float("-1.01244") @@ -2262,6 +2468,7 @@ class Program_weight_tensor_parameter_205: class Program_weight_tensor_parameter_206: name = "parameter_206" + original_name = "linear_31.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.56894") @@ -2273,6 +2480,7 @@ class Program_weight_tensor_parameter_206: class Program_weight_tensor_parameter_207: name = "parameter_207" + original_name = "linear_30.b_0" shape = [2304] dtype = "float32" min_val = float("-3.49749") @@ -2284,6 +2492,7 @@ class Program_weight_tensor_parameter_207: class Program_weight_tensor_parameter_208: name = "parameter_208" + original_name = "linear_30.w_0" shape = [768, 2304] dtype = "float32" min_val = float("-0.509087") @@ -2295,6 +2504,7 @@ class Program_weight_tensor_parameter_208: class Program_weight_tensor_parameter_209: name = "parameter_209" + original_name = "layer_norm_17.b_0" shape = [768] dtype = "float32" min_val = float("-2.19114") @@ -2306,6 +2516,7 @@ class Program_weight_tensor_parameter_209: class Program_weight_tensor_parameter_210: name = "parameter_210" + original_name = "layer_norm_17.w_0" shape = [768] dtype = "float32" min_val = float("-0.00231333") @@ -2317,6 +2528,7 @@ class Program_weight_tensor_parameter_210: class Program_weight_tensor_parameter_211: name = "parameter_211" + original_name = "linear_29.b_0" shape = [768] dtype = "float32" min_val = float("-1.2669") @@ -2328,6 +2540,7 @@ class Program_weight_tensor_parameter_211: class Program_weight_tensor_parameter_212: name = "parameter_212" + original_name = "linear_29.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-0.578491") @@ -2339,6 +2552,7 @@ class Program_weight_tensor_parameter_212: class Program_weight_tensor_parameter_213: name = "parameter_213" + original_name = "linear_28.b_0" shape = [3072] dtype = "float32" min_val = float("-3.75726") @@ -2350,6 +2564,7 @@ class Program_weight_tensor_parameter_213: class Program_weight_tensor_parameter_214: name = "parameter_214" + original_name = "linear_28.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.584685") @@ -2361,6 +2576,7 @@ class Program_weight_tensor_parameter_214: class Program_weight_tensor_parameter_215: name = "parameter_215" + original_name = "layer_norm_16.b_0" shape = [768] dtype = "float32" min_val = float("-1.85112") @@ -2372,6 +2588,7 @@ class Program_weight_tensor_parameter_215: class Program_weight_tensor_parameter_216: name = "parameter_216" + original_name = "layer_norm_16.w_0" shape = [768] dtype = "float32" min_val = float("-0.00447146") @@ -2383,6 +2600,7 @@ class Program_weight_tensor_parameter_216: class Program_weight_tensor_parameter_217: name = "parameter_217" + original_name = "linear_27.b_0" shape = [768] dtype = "float32" min_val = float("-1.45518") @@ -2394,6 +2612,7 @@ class Program_weight_tensor_parameter_217: class Program_weight_tensor_parameter_218: name = "parameter_218" + original_name = "linear_27.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.400837") @@ -2405,6 +2624,7 @@ class Program_weight_tensor_parameter_218: class Program_weight_tensor_parameter_219: name = "parameter_219" + original_name = "linear_26.b_0" shape = [2304] dtype = "float32" min_val = float("-4.16843") @@ -2416,6 +2636,7 @@ class Program_weight_tensor_parameter_219: class Program_weight_tensor_parameter_220: name = "parameter_220" + original_name = "linear_26.w_0" shape = [768, 2304] dtype = "float32" min_val = float("-0.608986") @@ -2427,6 +2648,7 @@ class Program_weight_tensor_parameter_220: class Program_weight_tensor_parameter_221: name = "parameter_221" + original_name = "layer_norm_15.b_0" shape = [768] dtype = "float32" min_val = float("-2.56146") @@ -2438,6 +2660,7 @@ class Program_weight_tensor_parameter_221: class Program_weight_tensor_parameter_222: name = "parameter_222" + original_name = "layer_norm_15.w_0" shape = [768] dtype = "float32" min_val = float("-0.00468997") @@ -2449,6 +2672,7 @@ class Program_weight_tensor_parameter_222: class Program_weight_tensor_parameter_223: name = "parameter_223" + original_name = "linear_25.b_0" shape = [768] dtype = "float32" min_val = float("-1.00877") @@ -2460,6 +2684,7 @@ class Program_weight_tensor_parameter_223: class Program_weight_tensor_parameter_224: name = "parameter_224" + original_name = "linear_25.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-0.655552") @@ -2471,6 +2696,7 @@ class Program_weight_tensor_parameter_224: class Program_weight_tensor_parameter_225: name = "parameter_225" + original_name = "linear_24.b_0" shape = [3072] dtype = "float32" min_val = float("-4.15263") @@ -2482,6 +2708,7 @@ class Program_weight_tensor_parameter_225: class Program_weight_tensor_parameter_226: name = "parameter_226" + original_name = "linear_24.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.464216") @@ -2493,6 +2720,7 @@ class Program_weight_tensor_parameter_226: class Program_weight_tensor_parameter_227: name = "parameter_227" + original_name = "layer_norm_14.b_0" shape = [768] dtype = "float32" min_val = float("-1.96702") @@ -2504,6 +2732,7 @@ class Program_weight_tensor_parameter_227: class Program_weight_tensor_parameter_228: name = "parameter_228" + original_name = "layer_norm_14.w_0" shape = [768] dtype = "float32" min_val = float("-0.00605676") @@ -2515,6 +2744,7 @@ class Program_weight_tensor_parameter_228: class Program_weight_tensor_parameter_229: name = "parameter_229" + original_name = "linear_23.b_0" shape = [768] dtype = "float32" min_val = float("-1.33569") @@ -2526,6 +2756,7 @@ class Program_weight_tensor_parameter_229: class Program_weight_tensor_parameter_230: name = "parameter_230" + original_name = "linear_23.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.591529") @@ -2537,6 +2768,7 @@ class Program_weight_tensor_parameter_230: class Program_weight_tensor_parameter_231: name = "parameter_231" + original_name = "linear_22.b_0" shape = [2304] dtype = "float32" min_val = float("-3.72622") @@ -2548,6 +2780,7 @@ class Program_weight_tensor_parameter_231: class Program_weight_tensor_parameter_232: name = "parameter_232" + original_name = "linear_22.w_0" shape = [768, 2304] dtype = "float32" min_val = float("-0.540851") @@ -2559,6 +2792,7 @@ class Program_weight_tensor_parameter_232: class Program_weight_tensor_parameter_233: name = "parameter_233" + original_name = "layer_norm_13.b_0" shape = [768] dtype = "float32" min_val = float("-1.36338") @@ -2570,6 +2804,7 @@ class Program_weight_tensor_parameter_233: class Program_weight_tensor_parameter_234: name = "parameter_234" + original_name = "layer_norm_13.w_0" shape = [768] dtype = "float32" min_val = float("-0.0098726") @@ -2581,6 +2816,7 @@ class Program_weight_tensor_parameter_234: class Program_weight_tensor_parameter_235: name = "parameter_235" + original_name = "linear_21.b_0" shape = [768] dtype = "float32" min_val = float("-1.3665") @@ -2592,6 +2828,7 @@ class Program_weight_tensor_parameter_235: class Program_weight_tensor_parameter_236: name = "parameter_236" + original_name = "linear_21.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-0.509148") @@ -2603,6 +2840,7 @@ class Program_weight_tensor_parameter_236: class Program_weight_tensor_parameter_237: name = "parameter_237" + original_name = "linear_20.b_0" shape = [3072] dtype = "float32" min_val = float("-4.37637") @@ -2614,6 +2852,7 @@ class Program_weight_tensor_parameter_237: class Program_weight_tensor_parameter_238: name = "parameter_238" + original_name = "linear_20.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.408345") @@ -2625,6 +2864,7 @@ class Program_weight_tensor_parameter_238: class Program_weight_tensor_parameter_239: name = "parameter_239" + original_name = "layer_norm_12.b_0" shape = [768] dtype = "float32" min_val = float("-1.94687") @@ -2636,6 +2876,7 @@ class Program_weight_tensor_parameter_239: class Program_weight_tensor_parameter_240: name = "parameter_240" + original_name = "layer_norm_12.w_0" shape = [768] dtype = "float32" min_val = float("-0.0222912") @@ -2647,6 +2888,7 @@ class Program_weight_tensor_parameter_240: class Program_weight_tensor_parameter_241: name = "parameter_241" + original_name = "linear_19.b_0" shape = [768] dtype = "float32" min_val = float("-1.52958") @@ -2658,6 +2900,7 @@ class Program_weight_tensor_parameter_241: class Program_weight_tensor_parameter_242: name = "parameter_242" + original_name = "linear_19.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.540496") @@ -2669,6 +2912,7 @@ class Program_weight_tensor_parameter_242: class Program_weight_tensor_parameter_243: name = "parameter_243" + original_name = "linear_18.b_0" shape = [2304] dtype = "float32" min_val = float("-3.88688") @@ -2680,6 +2924,7 @@ class Program_weight_tensor_parameter_243: class Program_weight_tensor_parameter_244: name = "parameter_244" + original_name = "linear_18.w_0" shape = [768, 2304] dtype = "float32" min_val = float("-0.565362") @@ -2691,6 +2936,7 @@ class Program_weight_tensor_parameter_244: class Program_weight_tensor_parameter_245: name = "parameter_245" + original_name = "layer_norm_11.b_0" shape = [768] dtype = "float32" min_val = float("-0.582757") @@ -2702,6 +2948,7 @@ class Program_weight_tensor_parameter_245: class Program_weight_tensor_parameter_246: name = "parameter_246" + original_name = "layer_norm_11.w_0" shape = [768] dtype = "float32" min_val = float("-0.0170591") @@ -2713,6 +2960,7 @@ class Program_weight_tensor_parameter_246: class Program_weight_tensor_parameter_247: name = "parameter_247" + original_name = "linear_17.w_0" shape = [1536, 768] dtype = "float32" min_val = float("-0.389307") @@ -2724,6 +2972,7 @@ class Program_weight_tensor_parameter_247: class Program_weight_tensor_parameter_248: name = "parameter_248" + original_name = "layer_norm_10.b_0" shape = [1536] dtype = "float32" min_val = float("-1.22106") @@ -2735,6 +2984,7 @@ class Program_weight_tensor_parameter_248: class Program_weight_tensor_parameter_249: name = "parameter_249" + original_name = "layer_norm_10.w_0" shape = [1536] dtype = "float32" min_val = float("0.139853") @@ -2746,6 +2996,7 @@ class Program_weight_tensor_parameter_249: class Program_weight_tensor_parameter_250: name = "parameter_250" + original_name = "linear_16.b_0" shape = [384] dtype = "float32" min_val = float("-7.43347") @@ -2757,6 +3008,7 @@ class Program_weight_tensor_parameter_250: class Program_weight_tensor_parameter_251: name = "parameter_251" + original_name = "linear_16.w_0" shape = [1536, 384] dtype = "float32" min_val = float("-0.96922") @@ -2768,6 +3020,7 @@ class Program_weight_tensor_parameter_251: class Program_weight_tensor_parameter_252: name = "parameter_252" + original_name = "linear_15.b_0" shape = [1536] dtype = "float32" min_val = float("-2.63013") @@ -2779,6 +3032,7 @@ class Program_weight_tensor_parameter_252: class Program_weight_tensor_parameter_253: name = "parameter_253" + original_name = "linear_15.w_0" shape = [384, 1536] dtype = "float32" min_val = float("-0.481854") @@ -2790,6 +3044,7 @@ class Program_weight_tensor_parameter_253: class Program_weight_tensor_parameter_254: name = "parameter_254" + original_name = "layer_norm_9.b_0" shape = [384] dtype = "float32" min_val = float("-1.92433") @@ -2801,6 +3056,7 @@ class Program_weight_tensor_parameter_254: class Program_weight_tensor_parameter_255: name = "parameter_255" + original_name = "layer_norm_9.w_0" shape = [384] dtype = "float32" min_val = float("-0.00313667") @@ -2812,6 +3068,7 @@ class Program_weight_tensor_parameter_255: class Program_weight_tensor_parameter_256: name = "parameter_256" + original_name = "linear_14.b_0" shape = [384] dtype = "float32" min_val = float("-6.39685") @@ -2823,6 +3080,7 @@ class Program_weight_tensor_parameter_256: class Program_weight_tensor_parameter_257: name = "parameter_257" + original_name = "linear_14.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.3267") @@ -2834,6 +3092,7 @@ class Program_weight_tensor_parameter_257: class Program_weight_tensor_parameter_258: name = "parameter_258" + original_name = "linear_13.b_0" shape = [1152] dtype = "float32" min_val = float("-3.22718") @@ -2845,6 +3104,7 @@ class Program_weight_tensor_parameter_258: class Program_weight_tensor_parameter_259: name = "parameter_259" + original_name = "linear_13.w_0" shape = [384, 1152] dtype = "float32" min_val = float("-0.73738") @@ -2856,6 +3116,7 @@ class Program_weight_tensor_parameter_259: class Program_weight_tensor_parameter_260: name = "parameter_260" + original_name = "layer_norm_8.b_0" shape = [384] dtype = "float32" min_val = float("-1.48498") @@ -2867,6 +3128,7 @@ class Program_weight_tensor_parameter_260: class Program_weight_tensor_parameter_261: name = "parameter_261" + original_name = "layer_norm_8.w_0" shape = [384] dtype = "float32" min_val = float("0.139982") @@ -2878,6 +3140,7 @@ class Program_weight_tensor_parameter_261: class Program_weight_tensor_parameter_262: name = "parameter_262" + original_name = "linear_12.b_0" shape = [384] dtype = "float32" min_val = float("-5.82281") @@ -2889,6 +3152,7 @@ class Program_weight_tensor_parameter_262: class Program_weight_tensor_parameter_263: name = "parameter_263" + original_name = "linear_12.w_0" shape = [1536, 384] dtype = "float32" min_val = float("-0.437198") @@ -2900,6 +3164,7 @@ class Program_weight_tensor_parameter_263: class Program_weight_tensor_parameter_264: name = "parameter_264" + original_name = "linear_11.b_0" shape = [1536] dtype = "float32" min_val = float("-2.79129") @@ -2911,6 +3176,7 @@ class Program_weight_tensor_parameter_264: class Program_weight_tensor_parameter_265: name = "parameter_265" + original_name = "linear_11.w_0" shape = [384, 1536] dtype = "float32" min_val = float("-0.414644") @@ -2922,6 +3188,7 @@ class Program_weight_tensor_parameter_265: class Program_weight_tensor_parameter_266: name = "parameter_266" + original_name = "layer_norm_7.b_0" shape = [384] dtype = "float32" min_val = float("-0.822266") @@ -2933,6 +3200,7 @@ class Program_weight_tensor_parameter_266: class Program_weight_tensor_parameter_267: name = "parameter_267" + original_name = "layer_norm_7.w_0" shape = [384] dtype = "float32" min_val = float("-0.00684004") @@ -2944,6 +3212,7 @@ class Program_weight_tensor_parameter_267: class Program_weight_tensor_parameter_268: name = "parameter_268" + original_name = "linear_10.b_0" shape = [384] dtype = "float32" min_val = float("-5.74186") @@ -2955,6 +3224,7 @@ class Program_weight_tensor_parameter_268: class Program_weight_tensor_parameter_269: name = "parameter_269" + original_name = "linear_10.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.411367") @@ -2966,6 +3236,7 @@ class Program_weight_tensor_parameter_269: class Program_weight_tensor_parameter_270: name = "parameter_270" + original_name = "linear_9.b_0" shape = [1152] dtype = "float32" min_val = float("-2.78212") @@ -2977,6 +3248,7 @@ class Program_weight_tensor_parameter_270: class Program_weight_tensor_parameter_271: name = "parameter_271" + original_name = "linear_9.w_0" shape = [384, 1152] dtype = "float32" min_val = float("-0.587265") @@ -2988,6 +3260,7 @@ class Program_weight_tensor_parameter_271: class Program_weight_tensor_parameter_272: name = "parameter_272" + original_name = "layer_norm_6.b_0" shape = [384] dtype = "float32" min_val = float("-0.550254") @@ -2999,6 +3272,7 @@ class Program_weight_tensor_parameter_272: class Program_weight_tensor_parameter_273: name = "parameter_273" + original_name = "layer_norm_6.w_0" shape = [384] dtype = "float32" min_val = float("-0.0114319") @@ -3010,6 +3284,7 @@ class Program_weight_tensor_parameter_273: class Program_weight_tensor_parameter_274: name = "parameter_274" + original_name = "linear_8.w_0" shape = [768, 384] dtype = "float32" min_val = float("-0.429916") @@ -3021,6 +3296,7 @@ class Program_weight_tensor_parameter_274: class Program_weight_tensor_parameter_275: name = "parameter_275" + original_name = "layer_norm_5.b_0" shape = [768] dtype = "float32" min_val = float("-1.4646") @@ -3032,6 +3308,7 @@ class Program_weight_tensor_parameter_275: class Program_weight_tensor_parameter_276: name = "parameter_276" + original_name = "layer_norm_5.w_0" shape = [768] dtype = "float32" min_val = float("0.143799") @@ -3043,6 +3320,7 @@ class Program_weight_tensor_parameter_276: class Program_weight_tensor_parameter_277: name = "parameter_277" + original_name = "linear_7.b_0" shape = [192] dtype = "float32" min_val = float("-6.27268") @@ -3054,6 +3332,7 @@ class Program_weight_tensor_parameter_277: class Program_weight_tensor_parameter_278: name = "parameter_278" + original_name = "linear_7.w_0" shape = [768, 192] dtype = "float32" min_val = float("-0.859501") @@ -3065,6 +3344,7 @@ class Program_weight_tensor_parameter_278: class Program_weight_tensor_parameter_279: name = "parameter_279" + original_name = "linear_6.b_0" shape = [768] dtype = "float32" min_val = float("-2.60647") @@ -3076,6 +3356,7 @@ class Program_weight_tensor_parameter_279: class Program_weight_tensor_parameter_280: name = "parameter_280" + original_name = "linear_6.w_0" shape = [192, 768] dtype = "float32" min_val = float("-0.457411") @@ -3087,6 +3368,7 @@ class Program_weight_tensor_parameter_280: class Program_weight_tensor_parameter_281: name = "parameter_281" + original_name = "layer_norm_4.b_0" shape = [192] dtype = "float32" min_val = float("-1.13713") @@ -3098,6 +3380,7 @@ class Program_weight_tensor_parameter_281: class Program_weight_tensor_parameter_282: name = "parameter_282" + original_name = "layer_norm_4.w_0" shape = [192] dtype = "float32" min_val = float("-0.000186107") @@ -3109,6 +3392,7 @@ class Program_weight_tensor_parameter_282: class Program_weight_tensor_parameter_283: name = "parameter_283" + original_name = "linear_5.b_0" shape = [192] dtype = "float32" min_val = float("-1.4287") @@ -3120,6 +3404,7 @@ class Program_weight_tensor_parameter_283: class Program_weight_tensor_parameter_284: name = "parameter_284" + original_name = "linear_5.w_0" shape = [192, 192] dtype = "float32" min_val = float("-0.394143") @@ -3131,6 +3416,7 @@ class Program_weight_tensor_parameter_284: class Program_weight_tensor_parameter_285: name = "parameter_285" + original_name = "linear_4.b_0" shape = [576] dtype = "float32" min_val = float("-2.94483") @@ -3142,6 +3428,7 @@ class Program_weight_tensor_parameter_285: class Program_weight_tensor_parameter_286: name = "parameter_286" + original_name = "linear_4.w_0" shape = [192, 576] dtype = "float32" min_val = float("-0.439545") @@ -3153,6 +3440,7 @@ class Program_weight_tensor_parameter_286: class Program_weight_tensor_parameter_287: name = "parameter_287" + original_name = "layer_norm_3.b_0" shape = [192] dtype = "float32" min_val = float("-0.876003") @@ -3164,6 +3452,7 @@ class Program_weight_tensor_parameter_287: class Program_weight_tensor_parameter_288: name = "parameter_288" + original_name = "layer_norm_3.w_0" shape = [192] dtype = "float32" min_val = float("-0.000824008") @@ -3175,6 +3464,7 @@ class Program_weight_tensor_parameter_288: class Program_weight_tensor_parameter_289: name = "parameter_289" + original_name = "linear_3.b_0" shape = [192] dtype = "float32" min_val = float("-6.03491") @@ -3186,6 +3476,7 @@ class Program_weight_tensor_parameter_289: class Program_weight_tensor_parameter_290: name = "parameter_290" + original_name = "linear_3.w_0" shape = [768, 192] dtype = "float32" min_val = float("-0.370342") @@ -3197,6 +3488,7 @@ class Program_weight_tensor_parameter_290: class Program_weight_tensor_parameter_291: name = "parameter_291" + original_name = "linear_2.b_0" shape = [768] dtype = "float32" min_val = float("-3.08096") @@ -3208,6 +3500,7 @@ class Program_weight_tensor_parameter_291: class Program_weight_tensor_parameter_292: name = "parameter_292" + original_name = "linear_2.w_0" shape = [192, 768] dtype = "float32" min_val = float("-0.564003") @@ -3219,6 +3512,7 @@ class Program_weight_tensor_parameter_292: class Program_weight_tensor_parameter_293: name = "parameter_293" + original_name = "layer_norm_2.b_0" shape = [192] dtype = "float32" min_val = float("-1.22505") @@ -3230,6 +3524,7 @@ class Program_weight_tensor_parameter_293: class Program_weight_tensor_parameter_294: name = "parameter_294" + original_name = "layer_norm_2.w_0" shape = [192] dtype = "float32" min_val = float("0.00255137") @@ -3241,6 +3536,7 @@ class Program_weight_tensor_parameter_294: class Program_weight_tensor_parameter_295: name = "parameter_295" + original_name = "linear_1.b_0" shape = [192] dtype = "float32" min_val = float("-2.30063") @@ -3252,6 +3548,7 @@ class Program_weight_tensor_parameter_295: class Program_weight_tensor_parameter_296: name = "parameter_296" + original_name = "linear_1.w_0" shape = [192, 192] dtype = "float32" min_val = float("-0.407784") @@ -3263,6 +3560,7 @@ class Program_weight_tensor_parameter_296: class Program_weight_tensor_parameter_297: name = "parameter_297" + original_name = "linear_0.b_0" shape = [576] dtype = "float32" min_val = float("-3.99358") @@ -3274,6 +3572,7 @@ class Program_weight_tensor_parameter_297: class Program_weight_tensor_parameter_298: name = "parameter_298" + original_name = "linear_0.w_0" shape = [192, 576] dtype = "float32" min_val = float("-0.480755") @@ -3285,6 +3584,7 @@ class Program_weight_tensor_parameter_298: class Program_weight_tensor_parameter_299: name = "parameter_299" + original_name = "layer_norm_1.b_0" shape = [192] dtype = "float32" min_val = float("-0.989179") @@ -3296,6 +3596,7 @@ class Program_weight_tensor_parameter_299: class Program_weight_tensor_parameter_300: name = "parameter_300" + original_name = "layer_norm_1.w_0" shape = [192] dtype = "float32" min_val = float("-0.0232513") @@ -3307,6 +3608,7 @@ class Program_weight_tensor_parameter_300: class Program_weight_tensor_parameter_301: name = "parameter_301" + original_name = "layer_norm_0.b_0" shape = [192] dtype = "float32" min_val = float("-2.32856") @@ -3318,6 +3620,7 @@ class Program_weight_tensor_parameter_301: class Program_weight_tensor_parameter_302: name = "parameter_302" + original_name = "layer_norm_0.w_0" shape = [192] dtype = "float32" min_val = float("-0.0169884") @@ -3329,6 +3632,7 @@ class Program_weight_tensor_parameter_302: class Program_weight_tensor_parameter_303: name = "parameter_303" + original_name = "conv2d_0.b_0" shape = [192] dtype = "float32" min_val = float("-0.85125") @@ -3340,6 +3644,7 @@ class Program_weight_tensor_parameter_303: class Program_weight_tensor_parameter_304: name = "parameter_304" + original_name = "conv2d_0.w_0" shape = [192, 3, 4, 4] dtype = "float32" min_val = float("-0.34511") diff --git a/paddle_samples/PaddleX/SwinTransformer_large_patch4_window7_224/subgraph_0/input_meta.py b/paddle_samples/PaddleX/SwinTransformer_large_patch4_window7_224/subgraph_0/input_meta.py index ab3b49b66..91c29ab1f 100644 --- a/paddle_samples/PaddleX/SwinTransformer_large_patch4_window7_224/subgraph_0/input_meta.py +++ b/paddle_samples/PaddleX/SwinTransformer_large_patch4_window7_224/subgraph_0/input_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_data_0: name = "data_0" + original_name = "param_0" shape = [169, 6] dtype = "float32" min_val = float("-10.3241") @@ -11,6 +12,7 @@ class Program_weight_tensor_data_0: class Program_weight_tensor_data_1: name = "data_1" + original_name = "param_1" shape = [169, 6] dtype = "float32" min_val = float("-8.58318") @@ -22,6 +24,7 @@ class Program_weight_tensor_data_1: class Program_weight_tensor_data_2: name = "data_2" + original_name = "param_10" shape = [169, 24] dtype = "float32" min_val = float("-10.7504") @@ -33,6 +36,7 @@ class Program_weight_tensor_data_2: class Program_weight_tensor_data_3: name = "data_3" + original_name = "param_11" shape = [169, 24] dtype = "float32" min_val = float("-8.67962") @@ -44,6 +48,7 @@ class Program_weight_tensor_data_3: class Program_weight_tensor_data_4: name = "data_4" + original_name = "param_12" shape = [169, 24] dtype = "float32" min_val = float("-7.57958") @@ -55,6 +60,7 @@ class Program_weight_tensor_data_4: class Program_weight_tensor_data_5: name = "data_5" + original_name = "param_13" shape = [169, 24] dtype = "float32" min_val = float("-12.0178") @@ -66,6 +72,7 @@ class Program_weight_tensor_data_5: class Program_weight_tensor_data_6: name = "data_6" + original_name = "param_14" shape = [169, 24] dtype = "float32" min_val = float("-14.5763") @@ -77,6 +84,7 @@ class Program_weight_tensor_data_6: class Program_weight_tensor_data_7: name = "data_7" + original_name = "param_15" shape = [169, 24] dtype = "float32" min_val = float("-9.31443") @@ -88,6 +96,7 @@ class Program_weight_tensor_data_7: class Program_weight_tensor_data_8: name = "data_8" + original_name = "param_16" shape = [169, 24] dtype = "float32" min_val = float("-10.532") @@ -99,6 +108,7 @@ class Program_weight_tensor_data_8: class Program_weight_tensor_data_9: name = "data_9" + original_name = "param_17" shape = [169, 24] dtype = "float32" min_val = float("-5.79538") @@ -110,6 +120,7 @@ class Program_weight_tensor_data_9: class Program_weight_tensor_data_10: name = "data_10" + original_name = "param_18" shape = [169, 24] dtype = "float32" min_val = float("-11.2408") @@ -121,6 +132,7 @@ class Program_weight_tensor_data_10: class Program_weight_tensor_data_11: name = "data_11" + original_name = "param_19" shape = [169, 24] dtype = "float32" min_val = float("-11.7032") @@ -132,6 +144,7 @@ class Program_weight_tensor_data_11: class Program_weight_tensor_data_12: name = "data_12" + original_name = "param_2" shape = [169, 12] dtype = "float32" min_val = float("-9.90303") @@ -143,6 +156,7 @@ class Program_weight_tensor_data_12: class Program_weight_tensor_data_13: name = "data_13" + original_name = "param_20" shape = [169, 24] dtype = "float32" min_val = float("-16.8217") @@ -154,6 +168,7 @@ class Program_weight_tensor_data_13: class Program_weight_tensor_data_14: name = "data_14" + original_name = "param_21" shape = [169, 24] dtype = "float32" min_val = float("-13.9222") @@ -165,6 +180,7 @@ class Program_weight_tensor_data_14: class Program_weight_tensor_data_15: name = "data_15" + original_name = "param_22" shape = [169, 48] dtype = "float32" min_val = float("-20.1905") @@ -176,6 +192,7 @@ class Program_weight_tensor_data_15: class Program_weight_tensor_data_16: name = "data_16" + original_name = "param_23" shape = [169, 48] dtype = "float32" min_val = float("-21.7737") @@ -187,6 +204,7 @@ class Program_weight_tensor_data_16: class Program_weight_tensor_data_17: name = "data_17" + original_name = "param_3" shape = [169, 12] dtype = "float32" min_val = float("-8.17582") @@ -198,6 +216,7 @@ class Program_weight_tensor_data_17: class Program_weight_tensor_data_18: name = "data_18" + original_name = "param_4" shape = [169, 24] dtype = "float32" min_val = float("-10.7597") @@ -209,6 +228,7 @@ class Program_weight_tensor_data_18: class Program_weight_tensor_data_19: name = "data_19" + original_name = "param_5" shape = [169, 24] dtype = "float32" min_val = float("-12.1126") @@ -220,6 +240,7 @@ class Program_weight_tensor_data_19: class Program_weight_tensor_data_20: name = "data_20" + original_name = "param_6" shape = [169, 24] dtype = "float32" min_val = float("-11.8414") @@ -231,6 +252,7 @@ class Program_weight_tensor_data_20: class Program_weight_tensor_data_21: name = "data_21" + original_name = "param_7" shape = [169, 24] dtype = "float32" min_val = float("-8.60576") @@ -242,6 +264,7 @@ class Program_weight_tensor_data_21: class Program_weight_tensor_data_22: name = "data_22" + original_name = "param_8" shape = [169, 24] dtype = "float32" min_val = float("-13.0313") @@ -253,6 +276,7 @@ class Program_weight_tensor_data_22: class Program_weight_tensor_data_23: name = "data_23" + original_name = "param_9" shape = [169, 24] dtype = "float32" min_val = float("-13.3256") @@ -264,6 +288,7 @@ class Program_weight_tensor_data_23: class Program_weight_tensor_data_24: name = "data_24" + original_name = "var_0" shape = [8, 3, 224, 224] dtype = "float32" min_val = float("-4.29068") @@ -275,6 +300,7 @@ class Program_weight_tensor_data_24: class Program_weight_tensor_data_25: name = "data_25" + original_name = "var_1021" shape = [49, 49] dtype = "int64" min_val = 0 @@ -284,6 +310,7 @@ class Program_weight_tensor_data_25: class Program_weight_tensor_data_26: name = "data_26" + original_name = "var_1101" shape = [49, 49] dtype = "int64" min_val = 0 @@ -293,6 +320,7 @@ class Program_weight_tensor_data_26: class Program_weight_tensor_data_27: name = "data_27" + original_name = "var_1163" shape = [49, 49] dtype = "int64" min_val = 0 @@ -302,6 +330,7 @@ class Program_weight_tensor_data_27: class Program_weight_tensor_data_28: name = "data_28" + original_name = "var_1243" shape = [49, 49] dtype = "int64" min_val = 0 @@ -311,6 +340,7 @@ class Program_weight_tensor_data_28: class Program_weight_tensor_data_29: name = "data_29" + original_name = "var_1305" shape = [49, 49] dtype = "int64" min_val = 0 @@ -320,6 +350,7 @@ class Program_weight_tensor_data_29: class Program_weight_tensor_data_30: name = "data_30" + original_name = "var_1385" shape = [49, 49] dtype = "int64" min_val = 0 @@ -329,6 +360,7 @@ class Program_weight_tensor_data_30: class Program_weight_tensor_data_31: name = "data_31" + original_name = "var_1447" shape = [49, 49] dtype = "int64" min_val = 0 @@ -338,6 +370,7 @@ class Program_weight_tensor_data_31: class Program_weight_tensor_data_32: name = "data_32" + original_name = "var_1527" shape = [49, 49] dtype = "int64" min_val = 0 @@ -347,6 +380,7 @@ class Program_weight_tensor_data_32: class Program_weight_tensor_data_33: name = "data_33" + original_name = "var_159" shape = [49, 49] dtype = "int64" min_val = 0 @@ -356,6 +390,7 @@ class Program_weight_tensor_data_33: class Program_weight_tensor_data_34: name = "data_34" + original_name = "var_1599" shape = [49, 49] dtype = "int64" min_val = 0 @@ -365,6 +400,7 @@ class Program_weight_tensor_data_34: class Program_weight_tensor_data_35: name = "data_35" + original_name = "var_1679" shape = [49, 49] dtype = "int64" min_val = 0 @@ -374,6 +410,7 @@ class Program_weight_tensor_data_35: class Program_weight_tensor_data_36: name = "data_36" + original_name = "var_21" shape = [49, 49] dtype = "int64" min_val = 0 @@ -383,6 +420,7 @@ class Program_weight_tensor_data_36: class Program_weight_tensor_data_37: name = "data_37" + original_name = "var_239" shape = [49, 49] dtype = "int64" min_val = 0 @@ -392,6 +430,7 @@ class Program_weight_tensor_data_37: class Program_weight_tensor_data_38: name = "data_38" + original_name = "var_311" shape = [49, 49] dtype = "int64" min_val = 0 @@ -401,6 +440,7 @@ class Program_weight_tensor_data_38: class Program_weight_tensor_data_39: name = "data_39" + original_name = "var_391" shape = [49, 49] dtype = "int64" min_val = 0 @@ -410,6 +450,7 @@ class Program_weight_tensor_data_39: class Program_weight_tensor_data_40: name = "data_40" + original_name = "var_453" shape = [49, 49] dtype = "int64" min_val = 0 @@ -419,6 +460,7 @@ class Program_weight_tensor_data_40: class Program_weight_tensor_data_41: name = "data_41" + original_name = "var_533" shape = [49, 49] dtype = "int64" min_val = 0 @@ -428,6 +470,7 @@ class Program_weight_tensor_data_41: class Program_weight_tensor_data_42: name = "data_42" + original_name = "var_595" shape = [49, 49] dtype = "int64" min_val = 0 @@ -437,6 +480,7 @@ class Program_weight_tensor_data_42: class Program_weight_tensor_data_43: name = "data_43" + original_name = "var_675" shape = [49, 49] dtype = "int64" min_val = 0 @@ -446,6 +490,7 @@ class Program_weight_tensor_data_43: class Program_weight_tensor_data_44: name = "data_44" + original_name = "var_737" shape = [49, 49] dtype = "int64" min_val = 0 @@ -455,6 +500,7 @@ class Program_weight_tensor_data_44: class Program_weight_tensor_data_45: name = "data_45" + original_name = "var_817" shape = [49, 49] dtype = "int64" min_val = 0 @@ -464,6 +510,7 @@ class Program_weight_tensor_data_45: class Program_weight_tensor_data_46: name = "data_46" + original_name = "var_87" shape = [49, 49] dtype = "int64" min_val = 0 @@ -473,6 +520,7 @@ class Program_weight_tensor_data_46: class Program_weight_tensor_data_47: name = "data_47" + original_name = "var_879" shape = [49, 49] dtype = "int64" min_val = 0 @@ -482,6 +530,7 @@ class Program_weight_tensor_data_47: class Program_weight_tensor_data_48: name = "data_48" + original_name = "var_959" shape = [49, 49] dtype = "int64" min_val = 0 diff --git a/paddle_samples/PaddleX/SwinTransformer_large_patch4_window7_224/subgraph_0/weight_meta.py b/paddle_samples/PaddleX/SwinTransformer_large_patch4_window7_224/subgraph_0/weight_meta.py index 562df5fce..e33038dba 100644 --- a/paddle_samples/PaddleX/SwinTransformer_large_patch4_window7_224/subgraph_0/weight_meta.py +++ b/paddle_samples/PaddleX/SwinTransformer_large_patch4_window7_224/subgraph_0/weight_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_parameter_0: name = "parameter_0" + original_name = "linear_99.b_0" shape = [102] dtype = "float32" data = None @@ -7,6 +8,7 @@ class Program_weight_tensor_parameter_0: class Program_weight_tensor_parameter_1: name = "parameter_1" + original_name = "linear_99.w_0" shape = [1536, 102] dtype = "float32" min_val = float("-0.0921121") @@ -18,6 +20,7 @@ class Program_weight_tensor_parameter_1: class Program_weight_tensor_parameter_2: name = "parameter_2" + original_name = "layer_norm_52.b_0" shape = [1536] dtype = "float32" min_val = float("-0.704584") @@ -29,6 +32,7 @@ class Program_weight_tensor_parameter_2: class Program_weight_tensor_parameter_3: name = "parameter_3" + original_name = "layer_norm_52.w_0" shape = [1536] dtype = "float32" min_val = float("-0.00448912") @@ -40,6 +44,7 @@ class Program_weight_tensor_parameter_3: class Program_weight_tensor_parameter_4: name = "parameter_4" + original_name = "linear_98.b_0" shape = [1536] dtype = "float32" min_val = float("-5.64997") @@ -51,6 +56,7 @@ class Program_weight_tensor_parameter_4: class Program_weight_tensor_parameter_5: name = "parameter_5" + original_name = "linear_98.w_0" shape = [6144, 1536] dtype = "float32" min_val = float("-3.13825") @@ -62,6 +68,7 @@ class Program_weight_tensor_parameter_5: class Program_weight_tensor_parameter_6: name = "parameter_6" + original_name = "linear_97.b_0" shape = [6144] dtype = "float32" min_val = float("-9.76716") @@ -73,6 +80,7 @@ class Program_weight_tensor_parameter_6: class Program_weight_tensor_parameter_7: name = "parameter_7" + original_name = "linear_97.w_0" shape = [1536, 6144] dtype = "float32" min_val = float("-0.907898") @@ -84,6 +92,7 @@ class Program_weight_tensor_parameter_7: class Program_weight_tensor_parameter_8: name = "parameter_8" + original_name = "layer_norm_51.b_0" shape = [1536] dtype = "float32" min_val = float("-2.864") @@ -95,6 +104,7 @@ class Program_weight_tensor_parameter_8: class Program_weight_tensor_parameter_9: name = "parameter_9" + original_name = "layer_norm_51.w_0" shape = [1536] dtype = "float32" min_val = float("-0.0188047") @@ -106,6 +116,7 @@ class Program_weight_tensor_parameter_9: class Program_weight_tensor_parameter_10: name = "parameter_10" + original_name = "linear_96.b_0" shape = [1536] dtype = "float32" min_val = float("-23.0995") @@ -117,6 +128,7 @@ class Program_weight_tensor_parameter_10: class Program_weight_tensor_parameter_11: name = "parameter_11" + original_name = "linear_96.w_0" shape = [1536, 1536] dtype = "float32" min_val = float("-1.48139") @@ -128,6 +140,7 @@ class Program_weight_tensor_parameter_11: class Program_weight_tensor_parameter_12: name = "parameter_12" + original_name = "linear_95.b_0" shape = [4608] dtype = "float32" min_val = float("-4.73898") @@ -139,6 +152,7 @@ class Program_weight_tensor_parameter_12: class Program_weight_tensor_parameter_13: name = "parameter_13" + original_name = "linear_95.w_0" shape = [1536, 4608] dtype = "float32" min_val = float("-0.585035") @@ -150,6 +164,7 @@ class Program_weight_tensor_parameter_13: class Program_weight_tensor_parameter_14: name = "parameter_14" + original_name = "layer_norm_50.b_0" shape = [1536] dtype = "float32" min_val = float("-4.93766") @@ -161,6 +176,7 @@ class Program_weight_tensor_parameter_14: class Program_weight_tensor_parameter_15: name = "parameter_15" + original_name = "layer_norm_50.w_0" shape = [1536] dtype = "float32" min_val = float("0.000171371") @@ -172,6 +188,7 @@ class Program_weight_tensor_parameter_15: class Program_weight_tensor_parameter_16: name = "parameter_16" + original_name = "linear_94.b_0" shape = [1536] dtype = "float32" min_val = float("-22.8212") @@ -183,6 +200,7 @@ class Program_weight_tensor_parameter_16: class Program_weight_tensor_parameter_17: name = "parameter_17" + original_name = "linear_94.w_0" shape = [6144, 1536] dtype = "float32" min_val = float("-1.90444") @@ -194,6 +212,7 @@ class Program_weight_tensor_parameter_17: class Program_weight_tensor_parameter_18: name = "parameter_18" + original_name = "linear_93.b_0" shape = [6144] dtype = "float32" min_val = float("-2.30416") @@ -205,6 +224,7 @@ class Program_weight_tensor_parameter_18: class Program_weight_tensor_parameter_19: name = "parameter_19" + original_name = "linear_93.w_0" shape = [1536, 6144] dtype = "float32" min_val = float("-1.21126") @@ -216,6 +236,7 @@ class Program_weight_tensor_parameter_19: class Program_weight_tensor_parameter_20: name = "parameter_20" + original_name = "layer_norm_49.b_0" shape = [1536] dtype = "float32" min_val = float("-6.87228") @@ -227,6 +248,7 @@ class Program_weight_tensor_parameter_20: class Program_weight_tensor_parameter_21: name = "parameter_21" + original_name = "layer_norm_49.w_0" shape = [1536] dtype = "float32" min_val = float("-0.019426") @@ -238,6 +260,7 @@ class Program_weight_tensor_parameter_21: class Program_weight_tensor_parameter_22: name = "parameter_22" + original_name = "linear_92.b_0" shape = [1536] dtype = "float32" min_val = float("-37.56") @@ -249,6 +272,7 @@ class Program_weight_tensor_parameter_22: class Program_weight_tensor_parameter_23: name = "parameter_23" + original_name = "linear_92.w_0" shape = [1536, 1536] dtype = "float32" min_val = float("-1.48508") @@ -260,6 +284,7 @@ class Program_weight_tensor_parameter_23: class Program_weight_tensor_parameter_24: name = "parameter_24" + original_name = "linear_91.b_0" shape = [4608] dtype = "float32" min_val = float("-5.23648") @@ -271,6 +296,7 @@ class Program_weight_tensor_parameter_24: class Program_weight_tensor_parameter_25: name = "parameter_25" + original_name = "linear_91.w_0" shape = [1536, 4608] dtype = "float32" min_val = float("-0.598569") @@ -282,6 +308,7 @@ class Program_weight_tensor_parameter_25: class Program_weight_tensor_parameter_26: name = "parameter_26" + original_name = "layer_norm_48.b_0" shape = [1536] dtype = "float32" min_val = float("-3.10297") @@ -293,6 +320,7 @@ class Program_weight_tensor_parameter_26: class Program_weight_tensor_parameter_27: name = "parameter_27" + original_name = "layer_norm_48.w_0" shape = [1536] dtype = "float32" min_val = float("-0.0112058") @@ -304,6 +332,7 @@ class Program_weight_tensor_parameter_27: class Program_weight_tensor_parameter_28: name = "parameter_28" + original_name = "linear_90.w_0" shape = [3072, 1536] dtype = "float32" min_val = float("-1.18301") @@ -315,6 +344,7 @@ class Program_weight_tensor_parameter_28: class Program_weight_tensor_parameter_29: name = "parameter_29" + original_name = "layer_norm_47.b_0" shape = [3072] dtype = "float32" min_val = float("-1.35163") @@ -326,6 +356,7 @@ class Program_weight_tensor_parameter_29: class Program_weight_tensor_parameter_30: name = "parameter_30" + original_name = "layer_norm_47.w_0" shape = [3072] dtype = "float32" min_val = float("0.0537523") @@ -337,6 +368,7 @@ class Program_weight_tensor_parameter_30: class Program_weight_tensor_parameter_31: name = "parameter_31" + original_name = "linear_89.b_0" shape = [768] dtype = "float32" min_val = float("-5.25286") @@ -348,6 +380,7 @@ class Program_weight_tensor_parameter_31: class Program_weight_tensor_parameter_32: name = "parameter_32" + original_name = "linear_89.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-3.22187") @@ -359,6 +392,7 @@ class Program_weight_tensor_parameter_32: class Program_weight_tensor_parameter_33: name = "parameter_33" + original_name = "linear_88.b_0" shape = [3072] dtype = "float32" min_val = float("-2.12441") @@ -370,6 +404,7 @@ class Program_weight_tensor_parameter_33: class Program_weight_tensor_parameter_34: name = "parameter_34" + original_name = "linear_88.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.715185") @@ -381,6 +416,7 @@ class Program_weight_tensor_parameter_34: class Program_weight_tensor_parameter_35: name = "parameter_35" + original_name = "layer_norm_46.b_0" shape = [768] dtype = "float32" min_val = float("-4.03642") @@ -392,6 +428,7 @@ class Program_weight_tensor_parameter_35: class Program_weight_tensor_parameter_36: name = "parameter_36" + original_name = "layer_norm_46.w_0" shape = [768] dtype = "float32" min_val = float("0.471934") @@ -403,6 +440,7 @@ class Program_weight_tensor_parameter_36: class Program_weight_tensor_parameter_37: name = "parameter_37" + original_name = "linear_87.b_0" shape = [768] dtype = "float32" min_val = float("-4.24913") @@ -414,6 +452,7 @@ class Program_weight_tensor_parameter_37: class Program_weight_tensor_parameter_38: name = "parameter_38" + original_name = "linear_87.w_0" shape = [768, 768] dtype = "float32" min_val = float("-2.26162") @@ -425,6 +464,7 @@ class Program_weight_tensor_parameter_38: class Program_weight_tensor_parameter_39: name = "parameter_39" + original_name = "linear_86.b_0" shape = [2304] dtype = "float32" min_val = float("-5.37301") @@ -436,6 +476,7 @@ class Program_weight_tensor_parameter_39: class Program_weight_tensor_parameter_40: name = "parameter_40" + original_name = "linear_86.w_0" shape = [768, 2304] dtype = "float32" min_val = float("-0.563909") @@ -447,6 +488,7 @@ class Program_weight_tensor_parameter_40: class Program_weight_tensor_parameter_41: name = "parameter_41" + original_name = "layer_norm_45.b_0" shape = [768] dtype = "float32" min_val = float("-3.72592") @@ -458,6 +500,7 @@ class Program_weight_tensor_parameter_41: class Program_weight_tensor_parameter_42: name = "parameter_42" + original_name = "layer_norm_45.w_0" shape = [768] dtype = "float32" min_val = float("0.0204089") @@ -469,6 +512,7 @@ class Program_weight_tensor_parameter_42: class Program_weight_tensor_parameter_43: name = "parameter_43" + original_name = "linear_85.b_0" shape = [768] dtype = "float32" min_val = float("-1.73044") @@ -480,6 +524,7 @@ class Program_weight_tensor_parameter_43: class Program_weight_tensor_parameter_44: name = "parameter_44" + original_name = "linear_85.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-2.24018") @@ -491,6 +536,7 @@ class Program_weight_tensor_parameter_44: class Program_weight_tensor_parameter_45: name = "parameter_45" + original_name = "linear_84.b_0" shape = [3072] dtype = "float32" min_val = float("-2.14574") @@ -502,6 +548,7 @@ class Program_weight_tensor_parameter_45: class Program_weight_tensor_parameter_46: name = "parameter_46" + original_name = "linear_84.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.98417") @@ -513,6 +560,7 @@ class Program_weight_tensor_parameter_46: class Program_weight_tensor_parameter_47: name = "parameter_47" + original_name = "layer_norm_44.b_0" shape = [768] dtype = "float32" min_val = float("-5.06209") @@ -524,6 +572,7 @@ class Program_weight_tensor_parameter_47: class Program_weight_tensor_parameter_48: name = "parameter_48" + original_name = "layer_norm_44.w_0" shape = [768] dtype = "float32" min_val = float("0.351817") @@ -535,6 +584,7 @@ class Program_weight_tensor_parameter_48: class Program_weight_tensor_parameter_49: name = "parameter_49" + original_name = "linear_83.b_0" shape = [768] dtype = "float32" min_val = float("-3.14406") @@ -546,6 +596,7 @@ class Program_weight_tensor_parameter_49: class Program_weight_tensor_parameter_50: name = "parameter_50" + original_name = "linear_83.w_0" shape = [768, 768] dtype = "float32" min_val = float("-1.60981") @@ -557,6 +608,7 @@ class Program_weight_tensor_parameter_50: class Program_weight_tensor_parameter_51: name = "parameter_51" + original_name = "linear_82.b_0" shape = [2304] dtype = "float32" min_val = float("-5.25393") @@ -568,6 +620,7 @@ class Program_weight_tensor_parameter_51: class Program_weight_tensor_parameter_52: name = "parameter_52" + original_name = "linear_82.w_0" shape = [768, 2304] dtype = "float32" min_val = float("-0.620573") @@ -579,6 +632,7 @@ class Program_weight_tensor_parameter_52: class Program_weight_tensor_parameter_53: name = "parameter_53" + original_name = "layer_norm_43.b_0" shape = [768] dtype = "float32" min_val = float("-2.84017") @@ -590,6 +644,7 @@ class Program_weight_tensor_parameter_53: class Program_weight_tensor_parameter_54: name = "parameter_54" + original_name = "layer_norm_43.w_0" shape = [768] dtype = "float32" min_val = float("0.0951353") @@ -601,6 +656,7 @@ class Program_weight_tensor_parameter_54: class Program_weight_tensor_parameter_55: name = "parameter_55" + original_name = "linear_81.b_0" shape = [768] dtype = "float32" min_val = float("-1.4788") @@ -612,6 +668,7 @@ class Program_weight_tensor_parameter_55: class Program_weight_tensor_parameter_56: name = "parameter_56" + original_name = "linear_81.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-1.84029") @@ -623,6 +680,7 @@ class Program_weight_tensor_parameter_56: class Program_weight_tensor_parameter_57: name = "parameter_57" + original_name = "linear_80.b_0" shape = [3072] dtype = "float32" min_val = float("-2.17839") @@ -634,6 +692,7 @@ class Program_weight_tensor_parameter_57: class Program_weight_tensor_parameter_58: name = "parameter_58" + original_name = "linear_80.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.582682") @@ -645,6 +704,7 @@ class Program_weight_tensor_parameter_58: class Program_weight_tensor_parameter_59: name = "parameter_59" + original_name = "layer_norm_42.b_0" shape = [768] dtype = "float32" min_val = float("-4.70468") @@ -656,6 +716,7 @@ class Program_weight_tensor_parameter_59: class Program_weight_tensor_parameter_60: name = "parameter_60" + original_name = "layer_norm_42.w_0" shape = [768] dtype = "float32" min_val = float("-0.00580218") @@ -667,6 +728,7 @@ class Program_weight_tensor_parameter_60: class Program_weight_tensor_parameter_61: name = "parameter_61" + original_name = "linear_79.b_0" shape = [768] dtype = "float32" min_val = float("-2.0011") @@ -678,6 +740,7 @@ class Program_weight_tensor_parameter_61: class Program_weight_tensor_parameter_62: name = "parameter_62" + original_name = "linear_79.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.807821") @@ -689,6 +752,7 @@ class Program_weight_tensor_parameter_62: class Program_weight_tensor_parameter_63: name = "parameter_63" + original_name = "linear_78.b_0" shape = [2304] dtype = "float32" min_val = float("-5.64921") @@ -700,6 +764,7 @@ class Program_weight_tensor_parameter_63: class Program_weight_tensor_parameter_64: name = "parameter_64" + original_name = "linear_78.w_0" shape = [768, 2304] dtype = "float32" min_val = float("-0.591616") @@ -711,6 +776,7 @@ class Program_weight_tensor_parameter_64: class Program_weight_tensor_parameter_65: name = "parameter_65" + original_name = "layer_norm_41.b_0" shape = [768] dtype = "float32" min_val = float("-2.04281") @@ -722,6 +788,7 @@ class Program_weight_tensor_parameter_65: class Program_weight_tensor_parameter_66: name = "parameter_66" + original_name = "layer_norm_41.w_0" shape = [768] dtype = "float32" min_val = float("0.0139667") @@ -733,6 +800,7 @@ class Program_weight_tensor_parameter_66: class Program_weight_tensor_parameter_67: name = "parameter_67" + original_name = "linear_77.b_0" shape = [768] dtype = "float32" min_val = float("-1.81579") @@ -744,6 +812,7 @@ class Program_weight_tensor_parameter_67: class Program_weight_tensor_parameter_68: name = "parameter_68" + original_name = "linear_77.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-0.737792") @@ -755,6 +824,7 @@ class Program_weight_tensor_parameter_68: class Program_weight_tensor_parameter_69: name = "parameter_69" + original_name = "linear_76.b_0" shape = [3072] dtype = "float32" min_val = float("-2.11139") @@ -766,6 +836,7 @@ class Program_weight_tensor_parameter_69: class Program_weight_tensor_parameter_70: name = "parameter_70" + original_name = "linear_76.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.517444") @@ -777,6 +848,7 @@ class Program_weight_tensor_parameter_70: class Program_weight_tensor_parameter_71: name = "parameter_71" + original_name = "layer_norm_40.b_0" shape = [768] dtype = "float32" min_val = float("-3.9029") @@ -788,6 +860,7 @@ class Program_weight_tensor_parameter_71: class Program_weight_tensor_parameter_72: name = "parameter_72" + original_name = "layer_norm_40.w_0" shape = [768] dtype = "float32" min_val = float("-0.0172421") @@ -799,6 +872,7 @@ class Program_weight_tensor_parameter_72: class Program_weight_tensor_parameter_73: name = "parameter_73" + original_name = "linear_75.b_0" shape = [768] dtype = "float32" min_val = float("-3.51606") @@ -810,6 +884,7 @@ class Program_weight_tensor_parameter_73: class Program_weight_tensor_parameter_74: name = "parameter_74" + original_name = "linear_75.w_0" shape = [768, 768] dtype = "float32" min_val = float("-1.55506") @@ -821,6 +896,7 @@ class Program_weight_tensor_parameter_74: class Program_weight_tensor_parameter_75: name = "parameter_75" + original_name = "linear_74.b_0" shape = [2304] dtype = "float32" min_val = float("-4.51314") @@ -832,6 +908,7 @@ class Program_weight_tensor_parameter_75: class Program_weight_tensor_parameter_76: name = "parameter_76" + original_name = "linear_74.w_0" shape = [768, 2304] dtype = "float32" min_val = float("-0.571176") @@ -843,6 +920,7 @@ class Program_weight_tensor_parameter_76: class Program_weight_tensor_parameter_77: name = "parameter_77" + original_name = "layer_norm_39.b_0" shape = [768] dtype = "float32" min_val = float("-1.61426") @@ -854,6 +932,7 @@ class Program_weight_tensor_parameter_77: class Program_weight_tensor_parameter_78: name = "parameter_78" + original_name = "layer_norm_39.w_0" shape = [768] dtype = "float32" min_val = float("0.148214") @@ -865,6 +944,7 @@ class Program_weight_tensor_parameter_78: class Program_weight_tensor_parameter_79: name = "parameter_79" + original_name = "linear_73.b_0" shape = [768] dtype = "float32" min_val = float("-3.11466") @@ -876,6 +956,7 @@ class Program_weight_tensor_parameter_79: class Program_weight_tensor_parameter_80: name = "parameter_80" + original_name = "linear_73.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-0.523968") @@ -887,6 +968,7 @@ class Program_weight_tensor_parameter_80: class Program_weight_tensor_parameter_81: name = "parameter_81" + original_name = "linear_72.b_0" shape = [3072] dtype = "float32" min_val = float("-2.23677") @@ -898,6 +980,7 @@ class Program_weight_tensor_parameter_81: class Program_weight_tensor_parameter_82: name = "parameter_82" + original_name = "linear_72.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-1.30491") @@ -909,6 +992,7 @@ class Program_weight_tensor_parameter_82: class Program_weight_tensor_parameter_83: name = "parameter_83" + original_name = "layer_norm_38.b_0" shape = [768] dtype = "float32" min_val = float("-2.90604") @@ -920,6 +1004,7 @@ class Program_weight_tensor_parameter_83: class Program_weight_tensor_parameter_84: name = "parameter_84" + original_name = "layer_norm_38.w_0" shape = [768] dtype = "float32" min_val = float("-0.0588838") @@ -931,6 +1016,7 @@ class Program_weight_tensor_parameter_84: class Program_weight_tensor_parameter_85: name = "parameter_85" + original_name = "linear_71.b_0" shape = [768] dtype = "float32" min_val = float("-2.00626") @@ -942,6 +1028,7 @@ class Program_weight_tensor_parameter_85: class Program_weight_tensor_parameter_86: name = "parameter_86" + original_name = "linear_71.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.838486") @@ -953,6 +1040,7 @@ class Program_weight_tensor_parameter_86: class Program_weight_tensor_parameter_87: name = "parameter_87" + original_name = "linear_70.b_0" shape = [2304] dtype = "float32" min_val = float("-4.94644") @@ -964,6 +1052,7 @@ class Program_weight_tensor_parameter_87: class Program_weight_tensor_parameter_88: name = "parameter_88" + original_name = "linear_70.w_0" shape = [768, 2304] dtype = "float32" min_val = float("-0.592913") @@ -975,6 +1064,7 @@ class Program_weight_tensor_parameter_88: class Program_weight_tensor_parameter_89: name = "parameter_89" + original_name = "layer_norm_37.b_0" shape = [768] dtype = "float32" min_val = float("-2.59161") @@ -986,6 +1076,7 @@ class Program_weight_tensor_parameter_89: class Program_weight_tensor_parameter_90: name = "parameter_90" + original_name = "layer_norm_37.w_0" shape = [768] dtype = "float32" min_val = float("0.12626") @@ -997,6 +1088,7 @@ class Program_weight_tensor_parameter_90: class Program_weight_tensor_parameter_91: name = "parameter_91" + original_name = "linear_69.b_0" shape = [768] dtype = "float32" min_val = float("-4.31392") @@ -1008,6 +1100,7 @@ class Program_weight_tensor_parameter_91: class Program_weight_tensor_parameter_92: name = "parameter_92" + original_name = "linear_69.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-0.93226") @@ -1019,6 +1112,7 @@ class Program_weight_tensor_parameter_92: class Program_weight_tensor_parameter_93: name = "parameter_93" + original_name = "linear_68.b_0" shape = [3072] dtype = "float32" min_val = float("-2.17812") @@ -1030,6 +1124,7 @@ class Program_weight_tensor_parameter_93: class Program_weight_tensor_parameter_94: name = "parameter_94" + original_name = "linear_68.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.716976") @@ -1041,6 +1136,7 @@ class Program_weight_tensor_parameter_94: class Program_weight_tensor_parameter_95: name = "parameter_95" + original_name = "layer_norm_36.b_0" shape = [768] dtype = "float32" min_val = float("-1.77128") @@ -1052,6 +1148,7 @@ class Program_weight_tensor_parameter_95: class Program_weight_tensor_parameter_96: name = "parameter_96" + original_name = "layer_norm_36.w_0" shape = [768] dtype = "float32" min_val = float("-0.0234062") @@ -1063,6 +1160,7 @@ class Program_weight_tensor_parameter_96: class Program_weight_tensor_parameter_97: name = "parameter_97" + original_name = "linear_67.b_0" shape = [768] dtype = "float32" min_val = float("-2.74816") @@ -1074,6 +1172,7 @@ class Program_weight_tensor_parameter_97: class Program_weight_tensor_parameter_98: name = "parameter_98" + original_name = "linear_67.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.611027") @@ -1085,6 +1184,7 @@ class Program_weight_tensor_parameter_98: class Program_weight_tensor_parameter_99: name = "parameter_99" + original_name = "linear_66.b_0" shape = [2304] dtype = "float32" min_val = float("-3.84781") @@ -1096,6 +1196,7 @@ class Program_weight_tensor_parameter_99: class Program_weight_tensor_parameter_100: name = "parameter_100" + original_name = "linear_66.w_0" shape = [768, 2304] dtype = "float32" min_val = float("-0.547712") @@ -1107,6 +1208,7 @@ class Program_weight_tensor_parameter_100: class Program_weight_tensor_parameter_101: name = "parameter_101" + original_name = "layer_norm_35.b_0" shape = [768] dtype = "float32" min_val = float("-3.05004") @@ -1118,6 +1220,7 @@ class Program_weight_tensor_parameter_101: class Program_weight_tensor_parameter_102: name = "parameter_102" + original_name = "layer_norm_35.w_0" shape = [768] dtype = "float32" min_val = float("0.0686034") @@ -1129,6 +1232,7 @@ class Program_weight_tensor_parameter_102: class Program_weight_tensor_parameter_103: name = "parameter_103" + original_name = "linear_65.b_0" shape = [768] dtype = "float32" min_val = float("-4.68687") @@ -1140,6 +1244,7 @@ class Program_weight_tensor_parameter_103: class Program_weight_tensor_parameter_104: name = "parameter_104" + original_name = "linear_65.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-1.14085") @@ -1151,6 +1256,7 @@ class Program_weight_tensor_parameter_104: class Program_weight_tensor_parameter_105: name = "parameter_105" + original_name = "linear_64.b_0" shape = [3072] dtype = "float32" min_val = float("-2.4522") @@ -1162,6 +1268,7 @@ class Program_weight_tensor_parameter_105: class Program_weight_tensor_parameter_106: name = "parameter_106" + original_name = "linear_64.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.419692") @@ -1173,6 +1280,7 @@ class Program_weight_tensor_parameter_106: class Program_weight_tensor_parameter_107: name = "parameter_107" + original_name = "layer_norm_34.b_0" shape = [768] dtype = "float32" min_val = float("-1.68316") @@ -1184,6 +1292,7 @@ class Program_weight_tensor_parameter_107: class Program_weight_tensor_parameter_108: name = "parameter_108" + original_name = "layer_norm_34.w_0" shape = [768] dtype = "float32" min_val = float("0.00202248") @@ -1195,6 +1304,7 @@ class Program_weight_tensor_parameter_108: class Program_weight_tensor_parameter_109: name = "parameter_109" + original_name = "linear_63.b_0" shape = [768] dtype = "float32" min_val = float("-1.60206") @@ -1206,6 +1316,7 @@ class Program_weight_tensor_parameter_109: class Program_weight_tensor_parameter_110: name = "parameter_110" + original_name = "linear_63.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.440633") @@ -1217,6 +1328,7 @@ class Program_weight_tensor_parameter_110: class Program_weight_tensor_parameter_111: name = "parameter_111" + original_name = "linear_62.b_0" shape = [2304] dtype = "float32" min_val = float("-3.34215") @@ -1228,6 +1340,7 @@ class Program_weight_tensor_parameter_111: class Program_weight_tensor_parameter_112: name = "parameter_112" + original_name = "linear_62.w_0" shape = [768, 2304] dtype = "float32" min_val = float("-0.581978") @@ -1239,6 +1352,7 @@ class Program_weight_tensor_parameter_112: class Program_weight_tensor_parameter_113: name = "parameter_113" + original_name = "layer_norm_33.b_0" shape = [768] dtype = "float32" min_val = float("-3.3748") @@ -1250,6 +1364,7 @@ class Program_weight_tensor_parameter_113: class Program_weight_tensor_parameter_114: name = "parameter_114" + original_name = "layer_norm_33.w_0" shape = [768] dtype = "float32" min_val = float("0.147299") @@ -1261,6 +1376,7 @@ class Program_weight_tensor_parameter_114: class Program_weight_tensor_parameter_115: name = "parameter_115" + original_name = "linear_61.b_0" shape = [768] dtype = "float32" min_val = float("-4.87313") @@ -1272,6 +1388,7 @@ class Program_weight_tensor_parameter_115: class Program_weight_tensor_parameter_116: name = "parameter_116" + original_name = "linear_61.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-1.01224") @@ -1283,6 +1400,7 @@ class Program_weight_tensor_parameter_116: class Program_weight_tensor_parameter_117: name = "parameter_117" + original_name = "linear_60.b_0" shape = [3072] dtype = "float32" min_val = float("-2.83841") @@ -1294,6 +1412,7 @@ class Program_weight_tensor_parameter_117: class Program_weight_tensor_parameter_118: name = "parameter_118" + original_name = "linear_60.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.706444") @@ -1305,6 +1424,7 @@ class Program_weight_tensor_parameter_118: class Program_weight_tensor_parameter_119: name = "parameter_119" + original_name = "layer_norm_32.b_0" shape = [768] dtype = "float32" min_val = float("-1.72978") @@ -1316,6 +1436,7 @@ class Program_weight_tensor_parameter_119: class Program_weight_tensor_parameter_120: name = "parameter_120" + original_name = "layer_norm_32.w_0" shape = [768] dtype = "float32" min_val = float("0.000515942") @@ -1327,6 +1448,7 @@ class Program_weight_tensor_parameter_120: class Program_weight_tensor_parameter_121: name = "parameter_121" + original_name = "linear_59.b_0" shape = [768] dtype = "float32" min_val = float("-1.87285") @@ -1338,6 +1460,7 @@ class Program_weight_tensor_parameter_121: class Program_weight_tensor_parameter_122: name = "parameter_122" + original_name = "linear_59.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.581905") @@ -1349,6 +1472,7 @@ class Program_weight_tensor_parameter_122: class Program_weight_tensor_parameter_123: name = "parameter_123" + original_name = "linear_58.b_0" shape = [2304] dtype = "float32" min_val = float("-4.91428") @@ -1360,6 +1484,7 @@ class Program_weight_tensor_parameter_123: class Program_weight_tensor_parameter_124: name = "parameter_124" + original_name = "linear_58.w_0" shape = [768, 2304] dtype = "float32" min_val = float("-0.588718") @@ -1371,6 +1496,7 @@ class Program_weight_tensor_parameter_124: class Program_weight_tensor_parameter_125: name = "parameter_125" + original_name = "layer_norm_31.b_0" shape = [768] dtype = "float32" min_val = float("-3.34201") @@ -1382,6 +1508,7 @@ class Program_weight_tensor_parameter_125: class Program_weight_tensor_parameter_126: name = "parameter_126" + original_name = "layer_norm_31.w_0" shape = [768] dtype = "float32" min_val = float("0.00529684") @@ -1393,6 +1520,7 @@ class Program_weight_tensor_parameter_126: class Program_weight_tensor_parameter_127: name = "parameter_127" + original_name = "linear_57.b_0" shape = [768] dtype = "float32" min_val = float("-2.51448") @@ -1404,6 +1532,7 @@ class Program_weight_tensor_parameter_127: class Program_weight_tensor_parameter_128: name = "parameter_128" + original_name = "linear_57.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-0.906987") @@ -1415,6 +1544,7 @@ class Program_weight_tensor_parameter_128: class Program_weight_tensor_parameter_129: name = "parameter_129" + original_name = "linear_56.b_0" shape = [3072] dtype = "float32" min_val = float("-3.34815") @@ -1426,6 +1556,7 @@ class Program_weight_tensor_parameter_129: class Program_weight_tensor_parameter_130: name = "parameter_130" + original_name = "linear_56.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.581071") @@ -1437,6 +1568,7 @@ class Program_weight_tensor_parameter_130: class Program_weight_tensor_parameter_131: name = "parameter_131" + original_name = "layer_norm_30.b_0" shape = [768] dtype = "float32" min_val = float("-1.78847") @@ -1448,6 +1580,7 @@ class Program_weight_tensor_parameter_131: class Program_weight_tensor_parameter_132: name = "parameter_132" + original_name = "layer_norm_30.w_0" shape = [768] dtype = "float32" min_val = float("-0.000879829") @@ -1459,6 +1592,7 @@ class Program_weight_tensor_parameter_132: class Program_weight_tensor_parameter_133: name = "parameter_133" + original_name = "linear_55.b_0" shape = [768] dtype = "float32" min_val = float("-1.02785") @@ -1470,6 +1604,7 @@ class Program_weight_tensor_parameter_133: class Program_weight_tensor_parameter_134: name = "parameter_134" + original_name = "linear_55.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.679053") @@ -1481,6 +1616,7 @@ class Program_weight_tensor_parameter_134: class Program_weight_tensor_parameter_135: name = "parameter_135" + original_name = "linear_54.b_0" shape = [2304] dtype = "float32" min_val = float("-5.13289") @@ -1492,6 +1628,7 @@ class Program_weight_tensor_parameter_135: class Program_weight_tensor_parameter_136: name = "parameter_136" + original_name = "linear_54.w_0" shape = [768, 2304] dtype = "float32" min_val = float("-0.699801") @@ -1503,6 +1640,7 @@ class Program_weight_tensor_parameter_136: class Program_weight_tensor_parameter_137: name = "parameter_137" + original_name = "layer_norm_29.b_0" shape = [768] dtype = "float32" min_val = float("-3.26285") @@ -1514,6 +1652,7 @@ class Program_weight_tensor_parameter_137: class Program_weight_tensor_parameter_138: name = "parameter_138" + original_name = "layer_norm_29.w_0" shape = [768] dtype = "float32" min_val = float("-0.0133322") @@ -1525,6 +1664,7 @@ class Program_weight_tensor_parameter_138: class Program_weight_tensor_parameter_139: name = "parameter_139" + original_name = "linear_53.b_0" shape = [768] dtype = "float32" min_val = float("-1.42826") @@ -1536,6 +1676,7 @@ class Program_weight_tensor_parameter_139: class Program_weight_tensor_parameter_140: name = "parameter_140" + original_name = "linear_53.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-0.943281") @@ -1547,6 +1688,7 @@ class Program_weight_tensor_parameter_140: class Program_weight_tensor_parameter_141: name = "parameter_141" + original_name = "linear_52.b_0" shape = [3072] dtype = "float32" min_val = float("-4.2036") @@ -1558,6 +1700,7 @@ class Program_weight_tensor_parameter_141: class Program_weight_tensor_parameter_142: name = "parameter_142" + original_name = "linear_52.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.487991") @@ -1569,6 +1712,7 @@ class Program_weight_tensor_parameter_142: class Program_weight_tensor_parameter_143: name = "parameter_143" + original_name = "layer_norm_28.b_0" shape = [768] dtype = "float32" min_val = float("-1.72023") @@ -1580,6 +1724,7 @@ class Program_weight_tensor_parameter_143: class Program_weight_tensor_parameter_144: name = "parameter_144" + original_name = "layer_norm_28.w_0" shape = [768] dtype = "float32" min_val = float("0.000837677") @@ -1591,6 +1736,7 @@ class Program_weight_tensor_parameter_144: class Program_weight_tensor_parameter_145: name = "parameter_145" + original_name = "linear_51.b_0" shape = [768] dtype = "float32" min_val = float("-1.78231") @@ -1602,6 +1748,7 @@ class Program_weight_tensor_parameter_145: class Program_weight_tensor_parameter_146: name = "parameter_146" + original_name = "linear_51.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.726244") @@ -1613,6 +1760,7 @@ class Program_weight_tensor_parameter_146: class Program_weight_tensor_parameter_147: name = "parameter_147" + original_name = "linear_50.b_0" shape = [2304] dtype = "float32" min_val = float("-3.99063") @@ -1624,6 +1772,7 @@ class Program_weight_tensor_parameter_147: class Program_weight_tensor_parameter_148: name = "parameter_148" + original_name = "linear_50.w_0" shape = [768, 2304] dtype = "float32" min_val = float("-0.665501") @@ -1635,6 +1784,7 @@ class Program_weight_tensor_parameter_148: class Program_weight_tensor_parameter_149: name = "parameter_149" + original_name = "layer_norm_27.b_0" shape = [768] dtype = "float32" min_val = float("-3.16356") @@ -1646,6 +1796,7 @@ class Program_weight_tensor_parameter_149: class Program_weight_tensor_parameter_150: name = "parameter_150" + original_name = "layer_norm_27.w_0" shape = [768] dtype = "float32" min_val = float("0.102714") @@ -1657,6 +1808,7 @@ class Program_weight_tensor_parameter_150: class Program_weight_tensor_parameter_151: name = "parameter_151" + original_name = "linear_49.b_0" shape = [768] dtype = "float32" min_val = float("-1.03716") @@ -1668,6 +1820,7 @@ class Program_weight_tensor_parameter_151: class Program_weight_tensor_parameter_152: name = "parameter_152" + original_name = "linear_49.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-0.951307") @@ -1679,6 +1832,7 @@ class Program_weight_tensor_parameter_152: class Program_weight_tensor_parameter_153: name = "parameter_153" + original_name = "linear_48.b_0" shape = [3072] dtype = "float32" min_val = float("-3.58636") @@ -1690,6 +1844,7 @@ class Program_weight_tensor_parameter_153: class Program_weight_tensor_parameter_154: name = "parameter_154" + original_name = "linear_48.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.560069") @@ -1701,6 +1856,7 @@ class Program_weight_tensor_parameter_154: class Program_weight_tensor_parameter_155: name = "parameter_155" + original_name = "layer_norm_26.b_0" shape = [768] dtype = "float32" min_val = float("-1.37913") @@ -1712,6 +1868,7 @@ class Program_weight_tensor_parameter_155: class Program_weight_tensor_parameter_156: name = "parameter_156" + original_name = "layer_norm_26.w_0" shape = [768] dtype = "float32" min_val = float("0.00111656") @@ -1723,6 +1880,7 @@ class Program_weight_tensor_parameter_156: class Program_weight_tensor_parameter_157: name = "parameter_157" + original_name = "linear_47.b_0" shape = [768] dtype = "float32" min_val = float("-1.38636") @@ -1734,6 +1892,7 @@ class Program_weight_tensor_parameter_157: class Program_weight_tensor_parameter_158: name = "parameter_158" + original_name = "linear_47.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.520222") @@ -1745,6 +1904,7 @@ class Program_weight_tensor_parameter_158: class Program_weight_tensor_parameter_159: name = "parameter_159" + original_name = "linear_46.b_0" shape = [2304] dtype = "float32" min_val = float("-3.63535") @@ -1756,6 +1916,7 @@ class Program_weight_tensor_parameter_159: class Program_weight_tensor_parameter_160: name = "parameter_160" + original_name = "linear_46.w_0" shape = [768, 2304] dtype = "float32" min_val = float("-0.617318") @@ -1767,6 +1928,7 @@ class Program_weight_tensor_parameter_160: class Program_weight_tensor_parameter_161: name = "parameter_161" + original_name = "layer_norm_25.b_0" shape = [768] dtype = "float32" min_val = float("-2.60325") @@ -1778,6 +1940,7 @@ class Program_weight_tensor_parameter_161: class Program_weight_tensor_parameter_162: name = "parameter_162" + original_name = "layer_norm_25.w_0" shape = [768] dtype = "float32" min_val = float("0.0274543") @@ -1789,6 +1952,7 @@ class Program_weight_tensor_parameter_162: class Program_weight_tensor_parameter_163: name = "parameter_163" + original_name = "linear_45.b_0" shape = [768] dtype = "float32" min_val = float("-1.08602") @@ -1800,6 +1964,7 @@ class Program_weight_tensor_parameter_163: class Program_weight_tensor_parameter_164: name = "parameter_164" + original_name = "linear_45.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-0.666764") @@ -1811,6 +1976,7 @@ class Program_weight_tensor_parameter_164: class Program_weight_tensor_parameter_165: name = "parameter_165" + original_name = "linear_44.b_0" shape = [3072] dtype = "float32" min_val = float("-3.28747") @@ -1822,6 +1988,7 @@ class Program_weight_tensor_parameter_165: class Program_weight_tensor_parameter_166: name = "parameter_166" + original_name = "linear_44.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.4876") @@ -1833,6 +2000,7 @@ class Program_weight_tensor_parameter_166: class Program_weight_tensor_parameter_167: name = "parameter_167" + original_name = "layer_norm_24.b_0" shape = [768] dtype = "float32" min_val = float("-1.5611") @@ -1844,6 +2012,7 @@ class Program_weight_tensor_parameter_167: class Program_weight_tensor_parameter_168: name = "parameter_168" + original_name = "layer_norm_24.w_0" shape = [768] dtype = "float32" min_val = float("0.014641") @@ -1855,6 +2024,7 @@ class Program_weight_tensor_parameter_168: class Program_weight_tensor_parameter_169: name = "parameter_169" + original_name = "linear_43.b_0" shape = [768] dtype = "float32" min_val = float("-1.43785") @@ -1866,6 +2036,7 @@ class Program_weight_tensor_parameter_169: class Program_weight_tensor_parameter_170: name = "parameter_170" + original_name = "linear_43.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.627548") @@ -1877,6 +2048,7 @@ class Program_weight_tensor_parameter_170: class Program_weight_tensor_parameter_171: name = "parameter_171" + original_name = "linear_42.b_0" shape = [2304] dtype = "float32" min_val = float("-3.43145") @@ -1888,6 +2060,7 @@ class Program_weight_tensor_parameter_171: class Program_weight_tensor_parameter_172: name = "parameter_172" + original_name = "linear_42.w_0" shape = [768, 2304] dtype = "float32" min_val = float("-0.581227") @@ -1899,6 +2072,7 @@ class Program_weight_tensor_parameter_172: class Program_weight_tensor_parameter_173: name = "parameter_173" + original_name = "layer_norm_23.b_0" shape = [768] dtype = "float32" min_val = float("-2.3427") @@ -1910,6 +2084,7 @@ class Program_weight_tensor_parameter_173: class Program_weight_tensor_parameter_174: name = "parameter_174" + original_name = "layer_norm_23.w_0" shape = [768] dtype = "float32" min_val = float("0.0227158") @@ -1921,6 +2096,7 @@ class Program_weight_tensor_parameter_174: class Program_weight_tensor_parameter_175: name = "parameter_175" + original_name = "linear_41.b_0" shape = [768] dtype = "float32" min_val = float("-0.922316") @@ -1932,6 +2108,7 @@ class Program_weight_tensor_parameter_175: class Program_weight_tensor_parameter_176: name = "parameter_176" + original_name = "linear_41.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-0.623398") @@ -1943,6 +2120,7 @@ class Program_weight_tensor_parameter_176: class Program_weight_tensor_parameter_177: name = "parameter_177" + original_name = "linear_40.b_0" shape = [3072] dtype = "float32" min_val = float("-3.43297") @@ -1954,6 +2132,7 @@ class Program_weight_tensor_parameter_177: class Program_weight_tensor_parameter_178: name = "parameter_178" + original_name = "linear_40.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.722896") @@ -1965,6 +2144,7 @@ class Program_weight_tensor_parameter_178: class Program_weight_tensor_parameter_179: name = "parameter_179" + original_name = "layer_norm_22.b_0" shape = [768] dtype = "float32" min_val = float("-1.89052") @@ -1976,6 +2156,7 @@ class Program_weight_tensor_parameter_179: class Program_weight_tensor_parameter_180: name = "parameter_180" + original_name = "layer_norm_22.w_0" shape = [768] dtype = "float32" min_val = float("0.0171377") @@ -1987,6 +2168,7 @@ class Program_weight_tensor_parameter_180: class Program_weight_tensor_parameter_181: name = "parameter_181" + original_name = "linear_39.b_0" shape = [768] dtype = "float32" min_val = float("-1.46043") @@ -1998,6 +2180,7 @@ class Program_weight_tensor_parameter_181: class Program_weight_tensor_parameter_182: name = "parameter_182" + original_name = "linear_39.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.598944") @@ -2009,6 +2192,7 @@ class Program_weight_tensor_parameter_182: class Program_weight_tensor_parameter_183: name = "parameter_183" + original_name = "linear_38.b_0" shape = [2304] dtype = "float32" min_val = float("-3.38464") @@ -2020,6 +2204,7 @@ class Program_weight_tensor_parameter_183: class Program_weight_tensor_parameter_184: name = "parameter_184" + original_name = "linear_38.w_0" shape = [768, 2304] dtype = "float32" min_val = float("-0.790775") @@ -2031,6 +2216,7 @@ class Program_weight_tensor_parameter_184: class Program_weight_tensor_parameter_185: name = "parameter_185" + original_name = "layer_norm_21.b_0" shape = [768] dtype = "float32" min_val = float("-2.41823") @@ -2042,6 +2228,7 @@ class Program_weight_tensor_parameter_185: class Program_weight_tensor_parameter_186: name = "parameter_186" + original_name = "layer_norm_21.w_0" shape = [768] dtype = "float32" min_val = float("-0.00125545") @@ -2053,6 +2240,7 @@ class Program_weight_tensor_parameter_186: class Program_weight_tensor_parameter_187: name = "parameter_187" + original_name = "linear_37.b_0" shape = [768] dtype = "float32" min_val = float("-0.966673") @@ -2064,6 +2252,7 @@ class Program_weight_tensor_parameter_187: class Program_weight_tensor_parameter_188: name = "parameter_188" + original_name = "linear_37.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-0.788093") @@ -2075,6 +2264,7 @@ class Program_weight_tensor_parameter_188: class Program_weight_tensor_parameter_189: name = "parameter_189" + original_name = "linear_36.b_0" shape = [3072] dtype = "float32" min_val = float("-3.52668") @@ -2086,6 +2276,7 @@ class Program_weight_tensor_parameter_189: class Program_weight_tensor_parameter_190: name = "parameter_190" + original_name = "linear_36.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.548297") @@ -2097,6 +2288,7 @@ class Program_weight_tensor_parameter_190: class Program_weight_tensor_parameter_191: name = "parameter_191" + original_name = "layer_norm_20.b_0" shape = [768] dtype = "float32" min_val = float("-1.84113") @@ -2108,6 +2300,7 @@ class Program_weight_tensor_parameter_191: class Program_weight_tensor_parameter_192: name = "parameter_192" + original_name = "layer_norm_20.w_0" shape = [768] dtype = "float32" min_val = float("0.00036599") @@ -2119,6 +2312,7 @@ class Program_weight_tensor_parameter_192: class Program_weight_tensor_parameter_193: name = "parameter_193" + original_name = "linear_35.b_0" shape = [768] dtype = "float32" min_val = float("-1.66019") @@ -2130,6 +2324,7 @@ class Program_weight_tensor_parameter_193: class Program_weight_tensor_parameter_194: name = "parameter_194" + original_name = "linear_35.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.764658") @@ -2141,6 +2336,7 @@ class Program_weight_tensor_parameter_194: class Program_weight_tensor_parameter_195: name = "parameter_195" + original_name = "linear_34.b_0" shape = [2304] dtype = "float32" min_val = float("-4.22564") @@ -2152,6 +2348,7 @@ class Program_weight_tensor_parameter_195: class Program_weight_tensor_parameter_196: name = "parameter_196" + original_name = "linear_34.w_0" shape = [768, 2304] dtype = "float32" min_val = float("-0.928681") @@ -2163,6 +2360,7 @@ class Program_weight_tensor_parameter_196: class Program_weight_tensor_parameter_197: name = "parameter_197" + original_name = "layer_norm_19.b_0" shape = [768] dtype = "float32" min_val = float("-2.57912") @@ -2174,6 +2372,7 @@ class Program_weight_tensor_parameter_197: class Program_weight_tensor_parameter_198: name = "parameter_198" + original_name = "layer_norm_19.w_0" shape = [768] dtype = "float32" min_val = float("0.00145275") @@ -2185,6 +2384,7 @@ class Program_weight_tensor_parameter_198: class Program_weight_tensor_parameter_199: name = "parameter_199" + original_name = "linear_33.b_0" shape = [768] dtype = "float32" min_val = float("-1.311") @@ -2196,6 +2396,7 @@ class Program_weight_tensor_parameter_199: class Program_weight_tensor_parameter_200: name = "parameter_200" + original_name = "linear_33.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-0.603177") @@ -2207,6 +2408,7 @@ class Program_weight_tensor_parameter_200: class Program_weight_tensor_parameter_201: name = "parameter_201" + original_name = "linear_32.b_0" shape = [3072] dtype = "float32" min_val = float("-3.68704") @@ -2218,6 +2420,7 @@ class Program_weight_tensor_parameter_201: class Program_weight_tensor_parameter_202: name = "parameter_202" + original_name = "linear_32.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.492515") @@ -2229,6 +2432,7 @@ class Program_weight_tensor_parameter_202: class Program_weight_tensor_parameter_203: name = "parameter_203" + original_name = "layer_norm_18.b_0" shape = [768] dtype = "float32" min_val = float("-1.85442") @@ -2240,6 +2444,7 @@ class Program_weight_tensor_parameter_203: class Program_weight_tensor_parameter_204: name = "parameter_204" + original_name = "layer_norm_18.w_0" shape = [768] dtype = "float32" min_val = float("-0.00586149") @@ -2251,6 +2456,7 @@ class Program_weight_tensor_parameter_204: class Program_weight_tensor_parameter_205: name = "parameter_205" + original_name = "linear_31.b_0" shape = [768] dtype = "float32" min_val = float("-1.01112") @@ -2262,6 +2468,7 @@ class Program_weight_tensor_parameter_205: class Program_weight_tensor_parameter_206: name = "parameter_206" + original_name = "linear_31.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.563765") @@ -2273,6 +2480,7 @@ class Program_weight_tensor_parameter_206: class Program_weight_tensor_parameter_207: name = "parameter_207" + original_name = "linear_30.b_0" shape = [2304] dtype = "float32" min_val = float("-3.49776") @@ -2284,6 +2492,7 @@ class Program_weight_tensor_parameter_207: class Program_weight_tensor_parameter_208: name = "parameter_208" + original_name = "linear_30.w_0" shape = [768, 2304] dtype = "float32" min_val = float("-0.513315") @@ -2295,6 +2504,7 @@ class Program_weight_tensor_parameter_208: class Program_weight_tensor_parameter_209: name = "parameter_209" + original_name = "layer_norm_17.b_0" shape = [768] dtype = "float32" min_val = float("-2.19543") @@ -2306,6 +2516,7 @@ class Program_weight_tensor_parameter_209: class Program_weight_tensor_parameter_210: name = "parameter_210" + original_name = "layer_norm_17.w_0" shape = [768] dtype = "float32" min_val = float("-0.00423051") @@ -2317,6 +2528,7 @@ class Program_weight_tensor_parameter_210: class Program_weight_tensor_parameter_211: name = "parameter_211" + original_name = "linear_29.b_0" shape = [768] dtype = "float32" min_val = float("-1.26723") @@ -2328,6 +2540,7 @@ class Program_weight_tensor_parameter_211: class Program_weight_tensor_parameter_212: name = "parameter_212" + original_name = "linear_29.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-0.569463") @@ -2339,6 +2552,7 @@ class Program_weight_tensor_parameter_212: class Program_weight_tensor_parameter_213: name = "parameter_213" + original_name = "linear_28.b_0" shape = [3072] dtype = "float32" min_val = float("-3.76881") @@ -2350,6 +2564,7 @@ class Program_weight_tensor_parameter_213: class Program_weight_tensor_parameter_214: name = "parameter_214" + original_name = "linear_28.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.581989") @@ -2361,6 +2576,7 @@ class Program_weight_tensor_parameter_214: class Program_weight_tensor_parameter_215: name = "parameter_215" + original_name = "layer_norm_16.b_0" shape = [768] dtype = "float32" min_val = float("-1.85229") @@ -2372,6 +2588,7 @@ class Program_weight_tensor_parameter_215: class Program_weight_tensor_parameter_216: name = "parameter_216" + original_name = "layer_norm_16.w_0" shape = [768] dtype = "float32" min_val = float("-0.0113052") @@ -2383,6 +2600,7 @@ class Program_weight_tensor_parameter_216: class Program_weight_tensor_parameter_217: name = "parameter_217" + original_name = "linear_27.b_0" shape = [768] dtype = "float32" min_val = float("-1.45671") @@ -2394,6 +2612,7 @@ class Program_weight_tensor_parameter_217: class Program_weight_tensor_parameter_218: name = "parameter_218" + original_name = "linear_27.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.399091") @@ -2405,6 +2624,7 @@ class Program_weight_tensor_parameter_218: class Program_weight_tensor_parameter_219: name = "parameter_219" + original_name = "linear_26.b_0" shape = [2304] dtype = "float32" min_val = float("-4.17066") @@ -2416,6 +2636,7 @@ class Program_weight_tensor_parameter_219: class Program_weight_tensor_parameter_220: name = "parameter_220" + original_name = "linear_26.w_0" shape = [768, 2304] dtype = "float32" min_val = float("-0.623775") @@ -2427,6 +2648,7 @@ class Program_weight_tensor_parameter_220: class Program_weight_tensor_parameter_221: name = "parameter_221" + original_name = "layer_norm_15.b_0" shape = [768] dtype = "float32" min_val = float("-2.56506") @@ -2438,6 +2660,7 @@ class Program_weight_tensor_parameter_221: class Program_weight_tensor_parameter_222: name = "parameter_222" + original_name = "layer_norm_15.w_0" shape = [768] dtype = "float32" min_val = float("-0.00336164") @@ -2449,6 +2672,7 @@ class Program_weight_tensor_parameter_222: class Program_weight_tensor_parameter_223: name = "parameter_223" + original_name = "linear_25.b_0" shape = [768] dtype = "float32" min_val = float("-1.00791") @@ -2460,6 +2684,7 @@ class Program_weight_tensor_parameter_223: class Program_weight_tensor_parameter_224: name = "parameter_224" + original_name = "linear_25.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-0.640575") @@ -2471,6 +2696,7 @@ class Program_weight_tensor_parameter_224: class Program_weight_tensor_parameter_225: name = "parameter_225" + original_name = "linear_24.b_0" shape = [3072] dtype = "float32" min_val = float("-4.14445") @@ -2482,6 +2708,7 @@ class Program_weight_tensor_parameter_225: class Program_weight_tensor_parameter_226: name = "parameter_226" + original_name = "linear_24.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.462229") @@ -2493,6 +2720,7 @@ class Program_weight_tensor_parameter_226: class Program_weight_tensor_parameter_227: name = "parameter_227" + original_name = "layer_norm_14.b_0" shape = [768] dtype = "float32" min_val = float("-1.9657") @@ -2504,6 +2732,7 @@ class Program_weight_tensor_parameter_227: class Program_weight_tensor_parameter_228: name = "parameter_228" + original_name = "layer_norm_14.w_0" shape = [768] dtype = "float32" min_val = float("-0.0108586") @@ -2515,6 +2744,7 @@ class Program_weight_tensor_parameter_228: class Program_weight_tensor_parameter_229: name = "parameter_229" + original_name = "linear_23.b_0" shape = [768] dtype = "float32" min_val = float("-1.33926") @@ -2526,6 +2756,7 @@ class Program_weight_tensor_parameter_229: class Program_weight_tensor_parameter_230: name = "parameter_230" + original_name = "linear_23.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.599678") @@ -2537,6 +2768,7 @@ class Program_weight_tensor_parameter_230: class Program_weight_tensor_parameter_231: name = "parameter_231" + original_name = "linear_22.b_0" shape = [2304] dtype = "float32" min_val = float("-3.70782") @@ -2548,6 +2780,7 @@ class Program_weight_tensor_parameter_231: class Program_weight_tensor_parameter_232: name = "parameter_232" + original_name = "linear_22.w_0" shape = [768, 2304] dtype = "float32" min_val = float("-0.537585") @@ -2559,6 +2792,7 @@ class Program_weight_tensor_parameter_232: class Program_weight_tensor_parameter_233: name = "parameter_233" + original_name = "layer_norm_13.b_0" shape = [768] dtype = "float32" min_val = float("-1.36292") @@ -2570,6 +2804,7 @@ class Program_weight_tensor_parameter_233: class Program_weight_tensor_parameter_234: name = "parameter_234" + original_name = "layer_norm_13.w_0" shape = [768] dtype = "float32" min_val = float("-0.00689248") @@ -2581,6 +2816,7 @@ class Program_weight_tensor_parameter_234: class Program_weight_tensor_parameter_235: name = "parameter_235" + original_name = "linear_21.b_0" shape = [768] dtype = "float32" min_val = float("-1.36428") @@ -2592,6 +2828,7 @@ class Program_weight_tensor_parameter_235: class Program_weight_tensor_parameter_236: name = "parameter_236" + original_name = "linear_21.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-0.493194") @@ -2603,6 +2840,7 @@ class Program_weight_tensor_parameter_236: class Program_weight_tensor_parameter_237: name = "parameter_237" + original_name = "linear_20.b_0" shape = [3072] dtype = "float32" min_val = float("-4.37402") @@ -2614,6 +2852,7 @@ class Program_weight_tensor_parameter_237: class Program_weight_tensor_parameter_238: name = "parameter_238" + original_name = "linear_20.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.403948") @@ -2625,6 +2864,7 @@ class Program_weight_tensor_parameter_238: class Program_weight_tensor_parameter_239: name = "parameter_239" + original_name = "layer_norm_12.b_0" shape = [768] dtype = "float32" min_val = float("-1.93683") @@ -2636,6 +2876,7 @@ class Program_weight_tensor_parameter_239: class Program_weight_tensor_parameter_240: name = "parameter_240" + original_name = "layer_norm_12.w_0" shape = [768] dtype = "float32" min_val = float("-0.0158466") @@ -2647,6 +2888,7 @@ class Program_weight_tensor_parameter_240: class Program_weight_tensor_parameter_241: name = "parameter_241" + original_name = "linear_19.b_0" shape = [768] dtype = "float32" min_val = float("-1.53027") @@ -2658,6 +2900,7 @@ class Program_weight_tensor_parameter_241: class Program_weight_tensor_parameter_242: name = "parameter_242" + original_name = "linear_19.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.5205") @@ -2669,6 +2912,7 @@ class Program_weight_tensor_parameter_242: class Program_weight_tensor_parameter_243: name = "parameter_243" + original_name = "linear_18.b_0" shape = [2304] dtype = "float32" min_val = float("-3.89015") @@ -2680,6 +2924,7 @@ class Program_weight_tensor_parameter_243: class Program_weight_tensor_parameter_244: name = "parameter_244" + original_name = "linear_18.w_0" shape = [768, 2304] dtype = "float32" min_val = float("-0.541329") @@ -2691,6 +2936,7 @@ class Program_weight_tensor_parameter_244: class Program_weight_tensor_parameter_245: name = "parameter_245" + original_name = "layer_norm_11.b_0" shape = [768] dtype = "float32" min_val = float("-0.582355") @@ -2702,6 +2948,7 @@ class Program_weight_tensor_parameter_245: class Program_weight_tensor_parameter_246: name = "parameter_246" + original_name = "layer_norm_11.w_0" shape = [768] dtype = "float32" min_val = float("-0.00701112") @@ -2713,6 +2960,7 @@ class Program_weight_tensor_parameter_246: class Program_weight_tensor_parameter_247: name = "parameter_247" + original_name = "linear_17.w_0" shape = [1536, 768] dtype = "float32" min_val = float("-0.387579") @@ -2724,6 +2972,7 @@ class Program_weight_tensor_parameter_247: class Program_weight_tensor_parameter_248: name = "parameter_248" + original_name = "layer_norm_10.b_0" shape = [1536] dtype = "float32" min_val = float("-1.21632") @@ -2735,6 +2984,7 @@ class Program_weight_tensor_parameter_248: class Program_weight_tensor_parameter_249: name = "parameter_249" + original_name = "layer_norm_10.w_0" shape = [1536] dtype = "float32" min_val = float("0.138717") @@ -2746,6 +2996,7 @@ class Program_weight_tensor_parameter_249: class Program_weight_tensor_parameter_250: name = "parameter_250" + original_name = "linear_16.b_0" shape = [384] dtype = "float32" min_val = float("-7.41405") @@ -2757,6 +3008,7 @@ class Program_weight_tensor_parameter_250: class Program_weight_tensor_parameter_251: name = "parameter_251" + original_name = "linear_16.w_0" shape = [1536, 384] dtype = "float32" min_val = float("-0.941645") @@ -2768,6 +3020,7 @@ class Program_weight_tensor_parameter_251: class Program_weight_tensor_parameter_252: name = "parameter_252" + original_name = "linear_15.b_0" shape = [1536] dtype = "float32" min_val = float("-2.62767") @@ -2779,6 +3032,7 @@ class Program_weight_tensor_parameter_252: class Program_weight_tensor_parameter_253: name = "parameter_253" + original_name = "linear_15.w_0" shape = [384, 1536] dtype = "float32" min_val = float("-0.483183") @@ -2790,6 +3044,7 @@ class Program_weight_tensor_parameter_253: class Program_weight_tensor_parameter_254: name = "parameter_254" + original_name = "layer_norm_9.b_0" shape = [384] dtype = "float32" min_val = float("-1.92285") @@ -2801,6 +3056,7 @@ class Program_weight_tensor_parameter_254: class Program_weight_tensor_parameter_255: name = "parameter_255" + original_name = "layer_norm_9.w_0" shape = [384] dtype = "float32" min_val = float("0.000511112") @@ -2812,6 +3068,7 @@ class Program_weight_tensor_parameter_255: class Program_weight_tensor_parameter_256: name = "parameter_256" + original_name = "linear_14.b_0" shape = [384] dtype = "float32" min_val = float("-6.35704") @@ -2823,6 +3080,7 @@ class Program_weight_tensor_parameter_256: class Program_weight_tensor_parameter_257: name = "parameter_257" + original_name = "linear_14.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.331492") @@ -2834,6 +3092,7 @@ class Program_weight_tensor_parameter_257: class Program_weight_tensor_parameter_258: name = "parameter_258" + original_name = "linear_13.b_0" shape = [1152] dtype = "float32" min_val = float("-3.23835") @@ -2845,6 +3104,7 @@ class Program_weight_tensor_parameter_258: class Program_weight_tensor_parameter_259: name = "parameter_259" + original_name = "linear_13.w_0" shape = [384, 1152] dtype = "float32" min_val = float("-0.725733") @@ -2856,6 +3116,7 @@ class Program_weight_tensor_parameter_259: class Program_weight_tensor_parameter_260: name = "parameter_260" + original_name = "layer_norm_8.b_0" shape = [384] dtype = "float32" min_val = float("-1.48137") @@ -2867,6 +3128,7 @@ class Program_weight_tensor_parameter_260: class Program_weight_tensor_parameter_261: name = "parameter_261" + original_name = "layer_norm_8.w_0" shape = [384] dtype = "float32" min_val = float("0.13403") @@ -2878,6 +3140,7 @@ class Program_weight_tensor_parameter_261: class Program_weight_tensor_parameter_262: name = "parameter_262" + original_name = "linear_12.b_0" shape = [384] dtype = "float32" min_val = float("-5.84017") @@ -2889,6 +3152,7 @@ class Program_weight_tensor_parameter_262: class Program_weight_tensor_parameter_263: name = "parameter_263" + original_name = "linear_12.w_0" shape = [1536, 384] dtype = "float32" min_val = float("-0.452618") @@ -2900,6 +3164,7 @@ class Program_weight_tensor_parameter_263: class Program_weight_tensor_parameter_264: name = "parameter_264" + original_name = "linear_11.b_0" shape = [1536] dtype = "float32" min_val = float("-2.79265") @@ -2911,6 +3176,7 @@ class Program_weight_tensor_parameter_264: class Program_weight_tensor_parameter_265: name = "parameter_265" + original_name = "linear_11.w_0" shape = [384, 1536] dtype = "float32" min_val = float("-0.426987") @@ -2922,6 +3188,7 @@ class Program_weight_tensor_parameter_265: class Program_weight_tensor_parameter_266: name = "parameter_266" + original_name = "layer_norm_7.b_0" shape = [384] dtype = "float32" min_val = float("-0.818317") @@ -2933,6 +3200,7 @@ class Program_weight_tensor_parameter_266: class Program_weight_tensor_parameter_267: name = "parameter_267" + original_name = "layer_norm_7.w_0" shape = [384] dtype = "float32" min_val = float("-0.00464777") @@ -2944,6 +3212,7 @@ class Program_weight_tensor_parameter_267: class Program_weight_tensor_parameter_268: name = "parameter_268" + original_name = "linear_10.b_0" shape = [384] dtype = "float32" min_val = float("-5.74942") @@ -2955,6 +3224,7 @@ class Program_weight_tensor_parameter_268: class Program_weight_tensor_parameter_269: name = "parameter_269" + original_name = "linear_10.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.419491") @@ -2966,6 +3236,7 @@ class Program_weight_tensor_parameter_269: class Program_weight_tensor_parameter_270: name = "parameter_270" + original_name = "linear_9.b_0" shape = [1152] dtype = "float32" min_val = float("-2.7924") @@ -2977,6 +3248,7 @@ class Program_weight_tensor_parameter_270: class Program_weight_tensor_parameter_271: name = "parameter_271" + original_name = "linear_9.w_0" shape = [384, 1152] dtype = "float32" min_val = float("-0.561166") @@ -2988,6 +3260,7 @@ class Program_weight_tensor_parameter_271: class Program_weight_tensor_parameter_272: name = "parameter_272" + original_name = "layer_norm_6.b_0" shape = [384] dtype = "float32" min_val = float("-0.545485") @@ -2999,6 +3272,7 @@ class Program_weight_tensor_parameter_272: class Program_weight_tensor_parameter_273: name = "parameter_273" + original_name = "layer_norm_6.w_0" shape = [384] dtype = "float32" min_val = float("-0.00340301") @@ -3010,6 +3284,7 @@ class Program_weight_tensor_parameter_273: class Program_weight_tensor_parameter_274: name = "parameter_274" + original_name = "linear_8.w_0" shape = [768, 384] dtype = "float32" min_val = float("-0.431339") @@ -3021,6 +3296,7 @@ class Program_weight_tensor_parameter_274: class Program_weight_tensor_parameter_275: name = "parameter_275" + original_name = "layer_norm_5.b_0" shape = [768] dtype = "float32" min_val = float("-1.46172") @@ -3032,6 +3308,7 @@ class Program_weight_tensor_parameter_275: class Program_weight_tensor_parameter_276: name = "parameter_276" + original_name = "layer_norm_5.w_0" shape = [768] dtype = "float32" min_val = float("0.150141") @@ -3043,6 +3320,7 @@ class Program_weight_tensor_parameter_276: class Program_weight_tensor_parameter_277: name = "parameter_277" + original_name = "linear_7.b_0" shape = [192] dtype = "float32" min_val = float("-6.27238") @@ -3054,6 +3332,7 @@ class Program_weight_tensor_parameter_277: class Program_weight_tensor_parameter_278: name = "parameter_278" + original_name = "linear_7.w_0" shape = [768, 192] dtype = "float32" min_val = float("-0.837592") @@ -3065,6 +3344,7 @@ class Program_weight_tensor_parameter_278: class Program_weight_tensor_parameter_279: name = "parameter_279" + original_name = "linear_6.b_0" shape = [768] dtype = "float32" min_val = float("-2.60515") @@ -3076,6 +3356,7 @@ class Program_weight_tensor_parameter_279: class Program_weight_tensor_parameter_280: name = "parameter_280" + original_name = "linear_6.w_0" shape = [192, 768] dtype = "float32" min_val = float("-0.451936") @@ -3087,6 +3368,7 @@ class Program_weight_tensor_parameter_280: class Program_weight_tensor_parameter_281: name = "parameter_281" + original_name = "layer_norm_4.b_0" shape = [192] dtype = "float32" min_val = float("-1.13738") @@ -3098,6 +3380,7 @@ class Program_weight_tensor_parameter_281: class Program_weight_tensor_parameter_282: name = "parameter_282" + original_name = "layer_norm_4.w_0" shape = [192] dtype = "float32" min_val = float("-0.00152619") @@ -3109,6 +3392,7 @@ class Program_weight_tensor_parameter_282: class Program_weight_tensor_parameter_283: name = "parameter_283" + original_name = "linear_5.b_0" shape = [192] dtype = "float32" min_val = float("-1.44722") @@ -3120,6 +3404,7 @@ class Program_weight_tensor_parameter_283: class Program_weight_tensor_parameter_284: name = "parameter_284" + original_name = "linear_5.w_0" shape = [192, 192] dtype = "float32" min_val = float("-0.403793") @@ -3131,6 +3416,7 @@ class Program_weight_tensor_parameter_284: class Program_weight_tensor_parameter_285: name = "parameter_285" + original_name = "linear_4.b_0" shape = [576] dtype = "float32" min_val = float("-2.94962") @@ -3142,6 +3428,7 @@ class Program_weight_tensor_parameter_285: class Program_weight_tensor_parameter_286: name = "parameter_286" + original_name = "linear_4.w_0" shape = [192, 576] dtype = "float32" min_val = float("-0.422497") @@ -3153,6 +3440,7 @@ class Program_weight_tensor_parameter_286: class Program_weight_tensor_parameter_287: name = "parameter_287" + original_name = "layer_norm_3.b_0" shape = [192] dtype = "float32" min_val = float("-0.876222") @@ -3164,6 +3452,7 @@ class Program_weight_tensor_parameter_287: class Program_weight_tensor_parameter_288: name = "parameter_288" + original_name = "layer_norm_3.w_0" shape = [192] dtype = "float32" min_val = float("-0.00195363") @@ -3175,6 +3464,7 @@ class Program_weight_tensor_parameter_288: class Program_weight_tensor_parameter_289: name = "parameter_289" + original_name = "linear_3.b_0" shape = [192] dtype = "float32" min_val = float("-6.04373") @@ -3186,6 +3476,7 @@ class Program_weight_tensor_parameter_289: class Program_weight_tensor_parameter_290: name = "parameter_290" + original_name = "linear_3.w_0" shape = [768, 192] dtype = "float32" min_val = float("-0.368389") @@ -3197,6 +3488,7 @@ class Program_weight_tensor_parameter_290: class Program_weight_tensor_parameter_291: name = "parameter_291" + original_name = "linear_2.b_0" shape = [768] dtype = "float32" min_val = float("-3.07803") @@ -3208,6 +3500,7 @@ class Program_weight_tensor_parameter_291: class Program_weight_tensor_parameter_292: name = "parameter_292" + original_name = "linear_2.w_0" shape = [192, 768] dtype = "float32" min_val = float("-0.56164") @@ -3219,6 +3512,7 @@ class Program_weight_tensor_parameter_292: class Program_weight_tensor_parameter_293: name = "parameter_293" + original_name = "layer_norm_2.b_0" shape = [192] dtype = "float32" min_val = float("-1.22395") @@ -3230,6 +3524,7 @@ class Program_weight_tensor_parameter_293: class Program_weight_tensor_parameter_294: name = "parameter_294" + original_name = "layer_norm_2.w_0" shape = [192] dtype = "float32" min_val = float("-0.000611421") @@ -3241,6 +3536,7 @@ class Program_weight_tensor_parameter_294: class Program_weight_tensor_parameter_295: name = "parameter_295" + original_name = "linear_1.b_0" shape = [192] dtype = "float32" min_val = float("-2.30202") @@ -3252,6 +3548,7 @@ class Program_weight_tensor_parameter_295: class Program_weight_tensor_parameter_296: name = "parameter_296" + original_name = "linear_1.w_0" shape = [192, 192] dtype = "float32" min_val = float("-0.409367") @@ -3263,6 +3560,7 @@ class Program_weight_tensor_parameter_296: class Program_weight_tensor_parameter_297: name = "parameter_297" + original_name = "linear_0.b_0" shape = [576] dtype = "float32" min_val = float("-3.97172") @@ -3274,6 +3572,7 @@ class Program_weight_tensor_parameter_297: class Program_weight_tensor_parameter_298: name = "parameter_298" + original_name = "linear_0.w_0" shape = [192, 576] dtype = "float32" min_val = float("-0.470962") @@ -3285,6 +3584,7 @@ class Program_weight_tensor_parameter_298: class Program_weight_tensor_parameter_299: name = "parameter_299" + original_name = "layer_norm_1.b_0" shape = [192] dtype = "float32" min_val = float("-0.983902") @@ -3296,6 +3596,7 @@ class Program_weight_tensor_parameter_299: class Program_weight_tensor_parameter_300: name = "parameter_300" + original_name = "layer_norm_1.w_0" shape = [192] dtype = "float32" min_val = float("-0.00194015") @@ -3307,6 +3608,7 @@ class Program_weight_tensor_parameter_300: class Program_weight_tensor_parameter_301: name = "parameter_301" + original_name = "layer_norm_0.b_0" shape = [192] dtype = "float32" min_val = float("-2.32116") @@ -3318,6 +3620,7 @@ class Program_weight_tensor_parameter_301: class Program_weight_tensor_parameter_302: name = "parameter_302" + original_name = "layer_norm_0.w_0" shape = [192] dtype = "float32" min_val = float("-0.0130715") @@ -3329,6 +3632,7 @@ class Program_weight_tensor_parameter_302: class Program_weight_tensor_parameter_303: name = "parameter_303" + original_name = "conv2d_0.b_0" shape = [192] dtype = "float32" min_val = float("-0.871267") @@ -3340,6 +3644,7 @@ class Program_weight_tensor_parameter_303: class Program_weight_tensor_parameter_304: name = "parameter_304" + original_name = "conv2d_0.w_0" shape = [192, 3, 4, 4] dtype = "float32" min_val = float("-0.345054") diff --git a/paddle_samples/PaddleX/SwinTransformer_large_patch4_window7_224/subgraph_2/input_meta.py b/paddle_samples/PaddleX/SwinTransformer_large_patch4_window7_224/subgraph_2/input_meta.py index d3ecba47e..64f03fa2f 100644 --- a/paddle_samples/PaddleX/SwinTransformer_large_patch4_window7_224/subgraph_2/input_meta.py +++ b/paddle_samples/PaddleX/SwinTransformer_large_patch4_window7_224/subgraph_2/input_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_data_0: name = "data_0" + original_name = "param_24" shape = [169, 6] dtype = "float32" min_val = float("-10.3296") @@ -11,6 +12,7 @@ class Program_weight_tensor_data_0: class Program_weight_tensor_data_1: name = "data_1" + original_name = "param_25" shape = [169, 6] dtype = "float32" min_val = float("-8.58887") @@ -22,6 +24,7 @@ class Program_weight_tensor_data_1: class Program_weight_tensor_data_2: name = "data_2" + original_name = "param_26" shape = [169, 12] dtype = "float32" min_val = float("-9.90518") @@ -33,6 +36,7 @@ class Program_weight_tensor_data_2: class Program_weight_tensor_data_3: name = "data_3" + original_name = "param_27" shape = [169, 12] dtype = "float32" min_val = float("-8.16882") @@ -44,6 +48,7 @@ class Program_weight_tensor_data_3: class Program_weight_tensor_data_4: name = "data_4" + original_name = "param_28" shape = [169, 24] dtype = "float32" min_val = float("-10.7626") @@ -55,6 +60,7 @@ class Program_weight_tensor_data_4: class Program_weight_tensor_data_5: name = "data_5" + original_name = "param_29" shape = [169, 24] dtype = "float32" min_val = float("-12.1173") @@ -66,6 +72,7 @@ class Program_weight_tensor_data_5: class Program_weight_tensor_data_6: name = "data_6" + original_name = "param_30" shape = [169, 24] dtype = "float32" min_val = float("-11.8484") @@ -77,6 +84,7 @@ class Program_weight_tensor_data_6: class Program_weight_tensor_data_7: name = "data_7" + original_name = "param_31" shape = [169, 24] dtype = "float32" min_val = float("-8.61177") @@ -88,6 +96,7 @@ class Program_weight_tensor_data_7: class Program_weight_tensor_data_8: name = "data_8" + original_name = "param_32" shape = [169, 24] dtype = "float32" min_val = float("-13.0287") @@ -99,6 +108,7 @@ class Program_weight_tensor_data_8: class Program_weight_tensor_data_9: name = "data_9" + original_name = "param_33" shape = [169, 24] dtype = "float32" min_val = float("-13.3304") @@ -110,6 +120,7 @@ class Program_weight_tensor_data_9: class Program_weight_tensor_data_10: name = "data_10" + original_name = "param_34" shape = [169, 24] dtype = "float32" min_val = float("-10.7532") @@ -121,6 +132,7 @@ class Program_weight_tensor_data_10: class Program_weight_tensor_data_11: name = "data_11" + original_name = "param_35" shape = [169, 24] dtype = "float32" min_val = float("-8.68031") @@ -132,6 +144,7 @@ class Program_weight_tensor_data_11: class Program_weight_tensor_data_12: name = "data_12" + original_name = "param_36" shape = [169, 24] dtype = "float32" min_val = float("-7.57641") @@ -143,6 +156,7 @@ class Program_weight_tensor_data_12: class Program_weight_tensor_data_13: name = "data_13" + original_name = "param_37" shape = [169, 24] dtype = "float32" min_val = float("-12.0265") @@ -154,6 +168,7 @@ class Program_weight_tensor_data_13: class Program_weight_tensor_data_14: name = "data_14" + original_name = "param_38" shape = [169, 24] dtype = "float32" min_val = float("-14.5738") @@ -165,6 +180,7 @@ class Program_weight_tensor_data_14: class Program_weight_tensor_data_15: name = "data_15" + original_name = "param_39" shape = [169, 24] dtype = "float32" min_val = float("-9.32066") @@ -176,6 +192,7 @@ class Program_weight_tensor_data_15: class Program_weight_tensor_data_16: name = "data_16" + original_name = "param_40" shape = [169, 24] dtype = "float32" min_val = float("-10.5179") @@ -187,6 +204,7 @@ class Program_weight_tensor_data_16: class Program_weight_tensor_data_17: name = "data_17" + original_name = "param_41" shape = [169, 24] dtype = "float32" min_val = float("-5.79733") @@ -198,6 +216,7 @@ class Program_weight_tensor_data_17: class Program_weight_tensor_data_18: name = "data_18" + original_name = "param_42" shape = [169, 24] dtype = "float32" min_val = float("-11.2457") @@ -209,6 +228,7 @@ class Program_weight_tensor_data_18: class Program_weight_tensor_data_19: name = "data_19" + original_name = "param_43" shape = [169, 24] dtype = "float32" min_val = float("-11.7029") @@ -220,6 +240,7 @@ class Program_weight_tensor_data_19: class Program_weight_tensor_data_20: name = "data_20" + original_name = "param_44" shape = [169, 24] dtype = "float32" min_val = float("-16.8219") @@ -231,6 +252,7 @@ class Program_weight_tensor_data_20: class Program_weight_tensor_data_21: name = "data_21" + original_name = "param_45" shape = [169, 24] dtype = "float32" min_val = float("-13.9228") @@ -242,6 +264,7 @@ class Program_weight_tensor_data_21: class Program_weight_tensor_data_22: name = "data_22" + original_name = "param_46" shape = [169, 48] dtype = "float32" min_val = float("-20.1903") @@ -253,6 +276,7 @@ class Program_weight_tensor_data_22: class Program_weight_tensor_data_23: name = "data_23" + original_name = "param_47" shape = [169, 48] dtype = "float32" min_val = float("-21.7741") @@ -264,6 +288,7 @@ class Program_weight_tensor_data_23: class Program_weight_tensor_data_24: name = "data_24" + original_name = "var_1731" shape = [4, 3, 224, 224] dtype = "float32" min_val = float("-4.90711") @@ -275,6 +300,7 @@ class Program_weight_tensor_data_24: class Program_weight_tensor_data_25: name = "data_25" + original_name = "var_1760" shape = [49, 49] dtype = "int64" min_val = 0 @@ -284,6 +310,7 @@ class Program_weight_tensor_data_25: class Program_weight_tensor_data_26: name = "data_26" + original_name = "var_1834" shape = [49, 49] dtype = "int64" min_val = 0 @@ -293,6 +320,7 @@ class Program_weight_tensor_data_26: class Program_weight_tensor_data_27: name = "data_27" + original_name = "var_1922" shape = [49, 49] dtype = "int64" min_val = 0 @@ -302,6 +330,7 @@ class Program_weight_tensor_data_27: class Program_weight_tensor_data_28: name = "data_28" + original_name = "var_2014" shape = [49, 49] dtype = "int64" min_val = 0 @@ -311,6 +340,7 @@ class Program_weight_tensor_data_28: class Program_weight_tensor_data_29: name = "data_29" + original_name = "var_2102" shape = [49, 49] dtype = "int64" min_val = 0 @@ -320,6 +350,7 @@ class Program_weight_tensor_data_29: class Program_weight_tensor_data_30: name = "data_30" + original_name = "var_2194" shape = [49, 49] dtype = "int64" min_val = 0 @@ -329,6 +360,7 @@ class Program_weight_tensor_data_30: class Program_weight_tensor_data_31: name = "data_31" + original_name = "var_2266" shape = [49, 49] dtype = "int64" min_val = 0 @@ -338,6 +370,7 @@ class Program_weight_tensor_data_31: class Program_weight_tensor_data_32: name = "data_32" + original_name = "var_2358" shape = [49, 49] dtype = "int64" min_val = 0 @@ -347,6 +380,7 @@ class Program_weight_tensor_data_32: class Program_weight_tensor_data_33: name = "data_33" + original_name = "var_2430" shape = [49, 49] dtype = "int64" min_val = 0 @@ -356,6 +390,7 @@ class Program_weight_tensor_data_33: class Program_weight_tensor_data_34: name = "data_34" + original_name = "var_2522" shape = [49, 49] dtype = "int64" min_val = 0 @@ -365,6 +400,7 @@ class Program_weight_tensor_data_34: class Program_weight_tensor_data_35: name = "data_35" + original_name = "var_2594" shape = [49, 49] dtype = "int64" min_val = 0 @@ -374,6 +410,7 @@ class Program_weight_tensor_data_35: class Program_weight_tensor_data_36: name = "data_36" + original_name = "var_2686" shape = [49, 49] dtype = "int64" min_val = 0 @@ -383,6 +420,7 @@ class Program_weight_tensor_data_36: class Program_weight_tensor_data_37: name = "data_37" + original_name = "var_2758" shape = [49, 49] dtype = "int64" min_val = 0 @@ -392,6 +430,7 @@ class Program_weight_tensor_data_37: class Program_weight_tensor_data_38: name = "data_38" + original_name = "var_2850" shape = [49, 49] dtype = "int64" min_val = 0 @@ -401,6 +440,7 @@ class Program_weight_tensor_data_38: class Program_weight_tensor_data_39: name = "data_39" + original_name = "var_2922" shape = [49, 49] dtype = "int64" min_val = 0 @@ -410,6 +450,7 @@ class Program_weight_tensor_data_39: class Program_weight_tensor_data_40: name = "data_40" + original_name = "var_3014" shape = [49, 49] dtype = "int64" min_val = 0 @@ -419,6 +460,7 @@ class Program_weight_tensor_data_40: class Program_weight_tensor_data_41: name = "data_41" + original_name = "var_3086" shape = [49, 49] dtype = "int64" min_val = 0 @@ -428,6 +470,7 @@ class Program_weight_tensor_data_41: class Program_weight_tensor_data_42: name = "data_42" + original_name = "var_3178" shape = [49, 49] dtype = "int64" min_val = 0 @@ -437,6 +480,7 @@ class Program_weight_tensor_data_42: class Program_weight_tensor_data_43: name = "data_43" + original_name = "var_3250" shape = [49, 49] dtype = "int64" min_val = 0 @@ -446,6 +490,7 @@ class Program_weight_tensor_data_43: class Program_weight_tensor_data_44: name = "data_44" + original_name = "var_3342" shape = [49, 49] dtype = "int64" min_val = 0 @@ -455,6 +500,7 @@ class Program_weight_tensor_data_44: class Program_weight_tensor_data_45: name = "data_45" + original_name = "var_3414" shape = [49, 49] dtype = "int64" min_val = 0 @@ -464,6 +510,7 @@ class Program_weight_tensor_data_45: class Program_weight_tensor_data_46: name = "data_46" + original_name = "var_3506" shape = [49, 49] dtype = "int64" min_val = 0 @@ -473,6 +520,7 @@ class Program_weight_tensor_data_46: class Program_weight_tensor_data_47: name = "data_47" + original_name = "var_3594" shape = [49, 49] dtype = "int64" min_val = 0 @@ -482,6 +530,7 @@ class Program_weight_tensor_data_47: class Program_weight_tensor_data_48: name = "data_48" + original_name = "var_3686" shape = [49, 49] dtype = "int64" min_val = 0 diff --git a/paddle_samples/PaddleX/SwinTransformer_large_patch4_window7_224/subgraph_2/weight_meta.py b/paddle_samples/PaddleX/SwinTransformer_large_patch4_window7_224/subgraph_2/weight_meta.py index ca484ab89..0ff5458bc 100644 --- a/paddle_samples/PaddleX/SwinTransformer_large_patch4_window7_224/subgraph_2/weight_meta.py +++ b/paddle_samples/PaddleX/SwinTransformer_large_patch4_window7_224/subgraph_2/weight_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_parameter_0: name = "parameter_0" + original_name = "linear_99.b_0" shape = [102] dtype = "float32" min_val = float("-0.0039639") @@ -11,6 +12,7 @@ class Program_weight_tensor_parameter_0: class Program_weight_tensor_parameter_1: name = "parameter_1" + original_name = "linear_99.w_0" shape = [1536, 102] dtype = "float32" min_val = float("-0.0970433") @@ -22,6 +24,7 @@ class Program_weight_tensor_parameter_1: class Program_weight_tensor_parameter_2: name = "parameter_2" + original_name = "layer_norm_52.b_0" shape = [1536] dtype = "float32" min_val = float("-0.704576") @@ -33,6 +36,7 @@ class Program_weight_tensor_parameter_2: class Program_weight_tensor_parameter_3: name = "parameter_3" + original_name = "layer_norm_52.w_0" shape = [1536] dtype = "float32" min_val = float("-0.00489932") @@ -44,6 +48,7 @@ class Program_weight_tensor_parameter_3: class Program_weight_tensor_parameter_4: name = "parameter_4" + original_name = "linear_98.b_0" shape = [1536] dtype = "float32" min_val = float("-5.63161") @@ -55,6 +60,7 @@ class Program_weight_tensor_parameter_4: class Program_weight_tensor_parameter_5: name = "parameter_5" + original_name = "linear_98.w_0" shape = [6144, 1536] dtype = "float32" min_val = float("-3.14346") @@ -66,6 +72,7 @@ class Program_weight_tensor_parameter_5: class Program_weight_tensor_parameter_6: name = "parameter_6" + original_name = "linear_97.b_0" shape = [6144] dtype = "float32" min_val = float("-9.77432") @@ -77,6 +84,7 @@ class Program_weight_tensor_parameter_6: class Program_weight_tensor_parameter_7: name = "parameter_7" + original_name = "linear_97.w_0" shape = [1536, 6144] dtype = "float32" min_val = float("-0.903425") @@ -88,6 +96,7 @@ class Program_weight_tensor_parameter_7: class Program_weight_tensor_parameter_8: name = "parameter_8" + original_name = "layer_norm_51.b_0" shape = [1536] dtype = "float32" min_val = float("-2.85305") @@ -99,6 +108,7 @@ class Program_weight_tensor_parameter_8: class Program_weight_tensor_parameter_9: name = "parameter_9" + original_name = "layer_norm_51.w_0" shape = [1536] dtype = "float32" min_val = float("-0.0158043") @@ -110,6 +120,7 @@ class Program_weight_tensor_parameter_9: class Program_weight_tensor_parameter_10: name = "parameter_10" + original_name = "linear_96.b_0" shape = [1536] dtype = "float32" min_val = float("-23.0898") @@ -121,6 +132,7 @@ class Program_weight_tensor_parameter_10: class Program_weight_tensor_parameter_11: name = "parameter_11" + original_name = "linear_96.w_0" shape = [1536, 1536] dtype = "float32" min_val = float("-1.48036") @@ -132,6 +144,7 @@ class Program_weight_tensor_parameter_11: class Program_weight_tensor_parameter_12: name = "parameter_12" + original_name = "linear_95.b_0" shape = [4608] dtype = "float32" min_val = float("-4.73746") @@ -143,6 +156,7 @@ class Program_weight_tensor_parameter_12: class Program_weight_tensor_parameter_13: name = "parameter_13" + original_name = "linear_95.w_0" shape = [1536, 4608] dtype = "float32" min_val = float("-0.582973") @@ -154,6 +168,7 @@ class Program_weight_tensor_parameter_13: class Program_weight_tensor_parameter_14: name = "parameter_14" + original_name = "layer_norm_50.b_0" shape = [1536] dtype = "float32" min_val = float("-4.9264") @@ -165,6 +180,7 @@ class Program_weight_tensor_parameter_14: class Program_weight_tensor_parameter_15: name = "parameter_15" + original_name = "layer_norm_50.w_0" shape = [1536] dtype = "float32" min_val = float("0.00100379") @@ -176,6 +192,7 @@ class Program_weight_tensor_parameter_15: class Program_weight_tensor_parameter_16: name = "parameter_16" + original_name = "linear_94.b_0" shape = [1536] dtype = "float32" min_val = float("-22.8166") @@ -187,6 +204,7 @@ class Program_weight_tensor_parameter_16: class Program_weight_tensor_parameter_17: name = "parameter_17" + original_name = "linear_94.w_0" shape = [6144, 1536] dtype = "float32" min_val = float("-1.89919") @@ -198,6 +216,7 @@ class Program_weight_tensor_parameter_17: class Program_weight_tensor_parameter_18: name = "parameter_18" + original_name = "linear_93.b_0" shape = [6144] dtype = "float32" min_val = float("-2.30786") @@ -209,6 +228,7 @@ class Program_weight_tensor_parameter_18: class Program_weight_tensor_parameter_19: name = "parameter_19" + original_name = "linear_93.w_0" shape = [1536, 6144] dtype = "float32" min_val = float("-1.20285") @@ -220,6 +240,7 @@ class Program_weight_tensor_parameter_19: class Program_weight_tensor_parameter_20: name = "parameter_20" + original_name = "layer_norm_49.b_0" shape = [1536] dtype = "float32" min_val = float("-6.85899") @@ -231,6 +252,7 @@ class Program_weight_tensor_parameter_20: class Program_weight_tensor_parameter_21: name = "parameter_21" + original_name = "layer_norm_49.w_0" shape = [1536] dtype = "float32" min_val = float("-0.0158562") @@ -242,6 +264,7 @@ class Program_weight_tensor_parameter_21: class Program_weight_tensor_parameter_22: name = "parameter_22" + original_name = "linear_92.b_0" shape = [1536] dtype = "float32" min_val = float("-37.5633") @@ -253,6 +276,7 @@ class Program_weight_tensor_parameter_22: class Program_weight_tensor_parameter_23: name = "parameter_23" + original_name = "linear_92.w_0" shape = [1536, 1536] dtype = "float32" min_val = float("-1.48352") @@ -264,6 +288,7 @@ class Program_weight_tensor_parameter_23: class Program_weight_tensor_parameter_24: name = "parameter_24" + original_name = "linear_91.b_0" shape = [4608] dtype = "float32" min_val = float("-5.23715") @@ -275,6 +300,7 @@ class Program_weight_tensor_parameter_24: class Program_weight_tensor_parameter_25: name = "parameter_25" + original_name = "linear_91.w_0" shape = [1536, 4608] dtype = "float32" min_val = float("-0.602173") @@ -286,6 +312,7 @@ class Program_weight_tensor_parameter_25: class Program_weight_tensor_parameter_26: name = "parameter_26" + original_name = "layer_norm_48.b_0" shape = [1536] dtype = "float32" min_val = float("-3.10674") @@ -297,6 +324,7 @@ class Program_weight_tensor_parameter_26: class Program_weight_tensor_parameter_27: name = "parameter_27" + original_name = "layer_norm_48.w_0" shape = [1536] dtype = "float32" min_val = float("-0.0104172") @@ -308,6 +336,7 @@ class Program_weight_tensor_parameter_27: class Program_weight_tensor_parameter_28: name = "parameter_28" + original_name = "linear_90.w_0" shape = [3072, 1536] dtype = "float32" min_val = float("-1.18391") @@ -319,6 +348,7 @@ class Program_weight_tensor_parameter_28: class Program_weight_tensor_parameter_29: name = "parameter_29" + original_name = "layer_norm_47.b_0" shape = [3072] dtype = "float32" min_val = float("-1.3487") @@ -330,6 +360,7 @@ class Program_weight_tensor_parameter_29: class Program_weight_tensor_parameter_30: name = "parameter_30" + original_name = "layer_norm_47.w_0" shape = [3072] dtype = "float32" min_val = float("0.059087") @@ -341,6 +372,7 @@ class Program_weight_tensor_parameter_30: class Program_weight_tensor_parameter_31: name = "parameter_31" + original_name = "linear_89.b_0" shape = [768] dtype = "float32" min_val = float("-5.24392") @@ -352,6 +384,7 @@ class Program_weight_tensor_parameter_31: class Program_weight_tensor_parameter_32: name = "parameter_32" + original_name = "linear_89.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-3.20854") @@ -363,6 +396,7 @@ class Program_weight_tensor_parameter_32: class Program_weight_tensor_parameter_33: name = "parameter_33" + original_name = "linear_88.b_0" shape = [3072] dtype = "float32" min_val = float("-2.12913") @@ -374,6 +408,7 @@ class Program_weight_tensor_parameter_33: class Program_weight_tensor_parameter_34: name = "parameter_34" + original_name = "linear_88.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.720142") @@ -385,6 +420,7 @@ class Program_weight_tensor_parameter_34: class Program_weight_tensor_parameter_35: name = "parameter_35" + original_name = "layer_norm_46.b_0" shape = [768] dtype = "float32" min_val = float("-4.03911") @@ -396,6 +432,7 @@ class Program_weight_tensor_parameter_35: class Program_weight_tensor_parameter_36: name = "parameter_36" + original_name = "layer_norm_46.w_0" shape = [768] dtype = "float32" min_val = float("0.476403") @@ -407,6 +444,7 @@ class Program_weight_tensor_parameter_36: class Program_weight_tensor_parameter_37: name = "parameter_37" + original_name = "linear_87.b_0" shape = [768] dtype = "float32" min_val = float("-4.25525") @@ -418,6 +456,7 @@ class Program_weight_tensor_parameter_37: class Program_weight_tensor_parameter_38: name = "parameter_38" + original_name = "linear_87.w_0" shape = [768, 768] dtype = "float32" min_val = float("-2.25515") @@ -429,6 +468,7 @@ class Program_weight_tensor_parameter_38: class Program_weight_tensor_parameter_39: name = "parameter_39" + original_name = "linear_86.b_0" shape = [2304] dtype = "float32" min_val = float("-5.37212") @@ -440,6 +480,7 @@ class Program_weight_tensor_parameter_39: class Program_weight_tensor_parameter_40: name = "parameter_40" + original_name = "linear_86.w_0" shape = [768, 2304] dtype = "float32" min_val = float("-0.570551") @@ -451,6 +492,7 @@ class Program_weight_tensor_parameter_40: class Program_weight_tensor_parameter_41: name = "parameter_41" + original_name = "layer_norm_45.b_0" shape = [768] dtype = "float32" min_val = float("-3.72464") @@ -462,6 +504,7 @@ class Program_weight_tensor_parameter_41: class Program_weight_tensor_parameter_42: name = "parameter_42" + original_name = "layer_norm_45.w_0" shape = [768] dtype = "float32" min_val = float("0.0158203") @@ -473,6 +516,7 @@ class Program_weight_tensor_parameter_42: class Program_weight_tensor_parameter_43: name = "parameter_43" + original_name = "linear_85.b_0" shape = [768] dtype = "float32" min_val = float("-1.73056") @@ -484,6 +528,7 @@ class Program_weight_tensor_parameter_43: class Program_weight_tensor_parameter_44: name = "parameter_44" + original_name = "linear_85.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-2.23708") @@ -495,6 +540,7 @@ class Program_weight_tensor_parameter_44: class Program_weight_tensor_parameter_45: name = "parameter_45" + original_name = "linear_84.b_0" shape = [3072] dtype = "float32" min_val = float("-2.14582") @@ -506,6 +552,7 @@ class Program_weight_tensor_parameter_45: class Program_weight_tensor_parameter_46: name = "parameter_46" + original_name = "linear_84.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.979467") @@ -517,6 +564,7 @@ class Program_weight_tensor_parameter_46: class Program_weight_tensor_parameter_47: name = "parameter_47" + original_name = "layer_norm_44.b_0" shape = [768] dtype = "float32" min_val = float("-5.0521") @@ -528,6 +576,7 @@ class Program_weight_tensor_parameter_47: class Program_weight_tensor_parameter_48: name = "parameter_48" + original_name = "layer_norm_44.w_0" shape = [768] dtype = "float32" min_val = float("0.352626") @@ -539,6 +588,7 @@ class Program_weight_tensor_parameter_48: class Program_weight_tensor_parameter_49: name = "parameter_49" + original_name = "linear_83.b_0" shape = [768] dtype = "float32" min_val = float("-3.14897") @@ -550,6 +600,7 @@ class Program_weight_tensor_parameter_49: class Program_weight_tensor_parameter_50: name = "parameter_50" + original_name = "linear_83.w_0" shape = [768, 768] dtype = "float32" min_val = float("-1.61014") @@ -561,6 +612,7 @@ class Program_weight_tensor_parameter_50: class Program_weight_tensor_parameter_51: name = "parameter_51" + original_name = "linear_82.b_0" shape = [2304] dtype = "float32" min_val = float("-5.25879") @@ -572,6 +624,7 @@ class Program_weight_tensor_parameter_51: class Program_weight_tensor_parameter_52: name = "parameter_52" + original_name = "linear_82.w_0" shape = [768, 2304] dtype = "float32" min_val = float("-0.621077") @@ -583,6 +636,7 @@ class Program_weight_tensor_parameter_52: class Program_weight_tensor_parameter_53: name = "parameter_53" + original_name = "layer_norm_43.b_0" shape = [768] dtype = "float32" min_val = float("-2.84151") @@ -594,6 +648,7 @@ class Program_weight_tensor_parameter_53: class Program_weight_tensor_parameter_54: name = "parameter_54" + original_name = "layer_norm_43.w_0" shape = [768] dtype = "float32" min_val = float("0.0929597") @@ -605,6 +660,7 @@ class Program_weight_tensor_parameter_54: class Program_weight_tensor_parameter_55: name = "parameter_55" + original_name = "linear_81.b_0" shape = [768] dtype = "float32" min_val = float("-1.4788") @@ -616,6 +672,7 @@ class Program_weight_tensor_parameter_55: class Program_weight_tensor_parameter_56: name = "parameter_56" + original_name = "linear_81.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-1.83843") @@ -627,6 +684,7 @@ class Program_weight_tensor_parameter_56: class Program_weight_tensor_parameter_57: name = "parameter_57" + original_name = "linear_80.b_0" shape = [3072] dtype = "float32" min_val = float("-2.17436") @@ -638,6 +696,7 @@ class Program_weight_tensor_parameter_57: class Program_weight_tensor_parameter_58: name = "parameter_58" + original_name = "linear_80.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.581173") @@ -649,6 +708,7 @@ class Program_weight_tensor_parameter_58: class Program_weight_tensor_parameter_59: name = "parameter_59" + original_name = "layer_norm_42.b_0" shape = [768] dtype = "float32" min_val = float("-4.70335") @@ -660,6 +720,7 @@ class Program_weight_tensor_parameter_59: class Program_weight_tensor_parameter_60: name = "parameter_60" + original_name = "layer_norm_42.w_0" shape = [768] dtype = "float32" min_val = float("-0.00725026") @@ -671,6 +732,7 @@ class Program_weight_tensor_parameter_60: class Program_weight_tensor_parameter_61: name = "parameter_61" + original_name = "linear_79.b_0" shape = [768] dtype = "float32" min_val = float("-2.00442") @@ -682,6 +744,7 @@ class Program_weight_tensor_parameter_61: class Program_weight_tensor_parameter_62: name = "parameter_62" + original_name = "linear_79.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.807033") @@ -693,6 +756,7 @@ class Program_weight_tensor_parameter_62: class Program_weight_tensor_parameter_63: name = "parameter_63" + original_name = "linear_78.b_0" shape = [2304] dtype = "float32" min_val = float("-5.65652") @@ -704,6 +768,7 @@ class Program_weight_tensor_parameter_63: class Program_weight_tensor_parameter_64: name = "parameter_64" + original_name = "linear_78.w_0" shape = [768, 2304] dtype = "float32" min_val = float("-0.589095") @@ -715,6 +780,7 @@ class Program_weight_tensor_parameter_64: class Program_weight_tensor_parameter_65: name = "parameter_65" + original_name = "layer_norm_41.b_0" shape = [768] dtype = "float32" min_val = float("-2.03777") @@ -726,6 +792,7 @@ class Program_weight_tensor_parameter_65: class Program_weight_tensor_parameter_66: name = "parameter_66" + original_name = "layer_norm_41.w_0" shape = [768] dtype = "float32" min_val = float("0.005589") @@ -737,6 +804,7 @@ class Program_weight_tensor_parameter_66: class Program_weight_tensor_parameter_67: name = "parameter_67" + original_name = "linear_77.b_0" shape = [768] dtype = "float32" min_val = float("-1.81673") @@ -748,6 +816,7 @@ class Program_weight_tensor_parameter_67: class Program_weight_tensor_parameter_68: name = "parameter_68" + original_name = "linear_77.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-0.72681") @@ -759,6 +828,7 @@ class Program_weight_tensor_parameter_68: class Program_weight_tensor_parameter_69: name = "parameter_69" + original_name = "linear_76.b_0" shape = [3072] dtype = "float32" min_val = float("-2.10295") @@ -770,6 +840,7 @@ class Program_weight_tensor_parameter_69: class Program_weight_tensor_parameter_70: name = "parameter_70" + original_name = "linear_76.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.515325") @@ -781,6 +852,7 @@ class Program_weight_tensor_parameter_70: class Program_weight_tensor_parameter_71: name = "parameter_71" + original_name = "layer_norm_40.b_0" shape = [768] dtype = "float32" min_val = float("-3.89756") @@ -792,6 +864,7 @@ class Program_weight_tensor_parameter_71: class Program_weight_tensor_parameter_72: name = "parameter_72" + original_name = "layer_norm_40.w_0" shape = [768] dtype = "float32" min_val = float("-0.0137726") @@ -803,6 +876,7 @@ class Program_weight_tensor_parameter_72: class Program_weight_tensor_parameter_73: name = "parameter_73" + original_name = "linear_75.b_0" shape = [768] dtype = "float32" min_val = float("-3.52008") @@ -814,6 +888,7 @@ class Program_weight_tensor_parameter_73: class Program_weight_tensor_parameter_74: name = "parameter_74" + original_name = "linear_75.w_0" shape = [768, 768] dtype = "float32" min_val = float("-1.55377") @@ -825,6 +900,7 @@ class Program_weight_tensor_parameter_74: class Program_weight_tensor_parameter_75: name = "parameter_75" + original_name = "linear_74.b_0" shape = [2304] dtype = "float32" min_val = float("-4.51817") @@ -836,6 +912,7 @@ class Program_weight_tensor_parameter_75: class Program_weight_tensor_parameter_76: name = "parameter_76" + original_name = "linear_74.w_0" shape = [768, 2304] dtype = "float32" min_val = float("-0.570347") @@ -847,6 +924,7 @@ class Program_weight_tensor_parameter_76: class Program_weight_tensor_parameter_77: name = "parameter_77" + original_name = "layer_norm_39.b_0" shape = [768] dtype = "float32" min_val = float("-1.62365") @@ -858,6 +936,7 @@ class Program_weight_tensor_parameter_77: class Program_weight_tensor_parameter_78: name = "parameter_78" + original_name = "layer_norm_39.w_0" shape = [768] dtype = "float32" min_val = float("0.138169") @@ -869,6 +948,7 @@ class Program_weight_tensor_parameter_78: class Program_weight_tensor_parameter_79: name = "parameter_79" + original_name = "linear_73.b_0" shape = [768] dtype = "float32" min_val = float("-3.10498") @@ -880,6 +960,7 @@ class Program_weight_tensor_parameter_79: class Program_weight_tensor_parameter_80: name = "parameter_80" + original_name = "linear_73.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-0.522027") @@ -891,6 +972,7 @@ class Program_weight_tensor_parameter_80: class Program_weight_tensor_parameter_81: name = "parameter_81" + original_name = "linear_72.b_0" shape = [3072] dtype = "float32" min_val = float("-2.24362") @@ -902,6 +984,7 @@ class Program_weight_tensor_parameter_81: class Program_weight_tensor_parameter_82: name = "parameter_82" + original_name = "linear_72.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-1.3036") @@ -913,6 +996,7 @@ class Program_weight_tensor_parameter_82: class Program_weight_tensor_parameter_83: name = "parameter_83" + original_name = "layer_norm_38.b_0" shape = [768] dtype = "float32" min_val = float("-2.90207") @@ -924,6 +1008,7 @@ class Program_weight_tensor_parameter_83: class Program_weight_tensor_parameter_84: name = "parameter_84" + original_name = "layer_norm_38.w_0" shape = [768] dtype = "float32" min_val = float("-0.0593156") @@ -935,6 +1020,7 @@ class Program_weight_tensor_parameter_84: class Program_weight_tensor_parameter_85: name = "parameter_85" + original_name = "linear_71.b_0" shape = [768] dtype = "float32" min_val = float("-2.00778") @@ -946,6 +1032,7 @@ class Program_weight_tensor_parameter_85: class Program_weight_tensor_parameter_86: name = "parameter_86" + original_name = "linear_71.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.842035") @@ -957,6 +1044,7 @@ class Program_weight_tensor_parameter_86: class Program_weight_tensor_parameter_87: name = "parameter_87" + original_name = "linear_70.b_0" shape = [2304] dtype = "float32" min_val = float("-4.94835") @@ -968,6 +1056,7 @@ class Program_weight_tensor_parameter_87: class Program_weight_tensor_parameter_88: name = "parameter_88" + original_name = "linear_70.w_0" shape = [768, 2304] dtype = "float32" min_val = float("-0.584122") @@ -979,6 +1068,7 @@ class Program_weight_tensor_parameter_88: class Program_weight_tensor_parameter_89: name = "parameter_89" + original_name = "layer_norm_37.b_0" shape = [768] dtype = "float32" min_val = float("-2.60455") @@ -990,6 +1080,7 @@ class Program_weight_tensor_parameter_89: class Program_weight_tensor_parameter_90: name = "parameter_90" + original_name = "layer_norm_37.w_0" shape = [768] dtype = "float32" min_val = float("0.115943") @@ -1001,6 +1092,7 @@ class Program_weight_tensor_parameter_90: class Program_weight_tensor_parameter_91: name = "parameter_91" + original_name = "linear_69.b_0" shape = [768] dtype = "float32" min_val = float("-4.30648") @@ -1012,6 +1104,7 @@ class Program_weight_tensor_parameter_91: class Program_weight_tensor_parameter_92: name = "parameter_92" + original_name = "linear_69.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-0.927432") @@ -1023,6 +1116,7 @@ class Program_weight_tensor_parameter_92: class Program_weight_tensor_parameter_93: name = "parameter_93" + original_name = "linear_68.b_0" shape = [3072] dtype = "float32" min_val = float("-2.17972") @@ -1034,6 +1128,7 @@ class Program_weight_tensor_parameter_93: class Program_weight_tensor_parameter_94: name = "parameter_94" + original_name = "linear_68.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.717324") @@ -1045,6 +1140,7 @@ class Program_weight_tensor_parameter_94: class Program_weight_tensor_parameter_95: name = "parameter_95" + original_name = "layer_norm_36.b_0" shape = [768] dtype = "float32" min_val = float("-1.75699") @@ -1056,6 +1152,7 @@ class Program_weight_tensor_parameter_95: class Program_weight_tensor_parameter_96: name = "parameter_96" + original_name = "layer_norm_36.w_0" shape = [768] dtype = "float32" min_val = float("-0.0289944") @@ -1067,6 +1164,7 @@ class Program_weight_tensor_parameter_96: class Program_weight_tensor_parameter_97: name = "parameter_97" + original_name = "linear_67.b_0" shape = [768] dtype = "float32" min_val = float("-2.75054") @@ -1078,6 +1176,7 @@ class Program_weight_tensor_parameter_97: class Program_weight_tensor_parameter_98: name = "parameter_98" + original_name = "linear_67.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.60569") @@ -1089,6 +1188,7 @@ class Program_weight_tensor_parameter_98: class Program_weight_tensor_parameter_99: name = "parameter_99" + original_name = "linear_66.b_0" shape = [2304] dtype = "float32" min_val = float("-3.85882") @@ -1100,6 +1200,7 @@ class Program_weight_tensor_parameter_99: class Program_weight_tensor_parameter_100: name = "parameter_100" + original_name = "linear_66.w_0" shape = [768, 2304] dtype = "float32" min_val = float("-0.55069") @@ -1111,6 +1212,7 @@ class Program_weight_tensor_parameter_100: class Program_weight_tensor_parameter_101: name = "parameter_101" + original_name = "layer_norm_35.b_0" shape = [768] dtype = "float32" min_val = float("-3.05861") @@ -1122,6 +1224,7 @@ class Program_weight_tensor_parameter_101: class Program_weight_tensor_parameter_102: name = "parameter_102" + original_name = "layer_norm_35.w_0" shape = [768] dtype = "float32" min_val = float("0.067056") @@ -1133,6 +1236,7 @@ class Program_weight_tensor_parameter_102: class Program_weight_tensor_parameter_103: name = "parameter_103" + original_name = "linear_65.b_0" shape = [768] dtype = "float32" min_val = float("-4.68059") @@ -1144,6 +1248,7 @@ class Program_weight_tensor_parameter_103: class Program_weight_tensor_parameter_104: name = "parameter_104" + original_name = "linear_65.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-1.14693") @@ -1155,6 +1260,7 @@ class Program_weight_tensor_parameter_104: class Program_weight_tensor_parameter_105: name = "parameter_105" + original_name = "linear_64.b_0" shape = [3072] dtype = "float32" min_val = float("-2.45807") @@ -1166,6 +1272,7 @@ class Program_weight_tensor_parameter_105: class Program_weight_tensor_parameter_106: name = "parameter_106" + original_name = "linear_64.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.424229") @@ -1177,6 +1284,7 @@ class Program_weight_tensor_parameter_106: class Program_weight_tensor_parameter_107: name = "parameter_107" + original_name = "layer_norm_34.b_0" shape = [768] dtype = "float32" min_val = float("-1.69456") @@ -1188,6 +1296,7 @@ class Program_weight_tensor_parameter_107: class Program_weight_tensor_parameter_108: name = "parameter_108" + original_name = "layer_norm_34.w_0" shape = [768] dtype = "float32" min_val = float("0.00494335") @@ -1199,6 +1308,7 @@ class Program_weight_tensor_parameter_108: class Program_weight_tensor_parameter_109: name = "parameter_109" + original_name = "linear_63.b_0" shape = [768] dtype = "float32" min_val = float("-1.6044") @@ -1210,6 +1320,7 @@ class Program_weight_tensor_parameter_109: class Program_weight_tensor_parameter_110: name = "parameter_110" + original_name = "linear_63.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.440926") @@ -1221,6 +1332,7 @@ class Program_weight_tensor_parameter_110: class Program_weight_tensor_parameter_111: name = "parameter_111" + original_name = "linear_62.b_0" shape = [2304] dtype = "float32" min_val = float("-3.34438") @@ -1232,6 +1344,7 @@ class Program_weight_tensor_parameter_111: class Program_weight_tensor_parameter_112: name = "parameter_112" + original_name = "linear_62.w_0" shape = [768, 2304] dtype = "float32" min_val = float("-0.57376") @@ -1243,6 +1356,7 @@ class Program_weight_tensor_parameter_112: class Program_weight_tensor_parameter_113: name = "parameter_113" + original_name = "layer_norm_33.b_0" shape = [768] dtype = "float32" min_val = float("-3.38745") @@ -1254,6 +1368,7 @@ class Program_weight_tensor_parameter_113: class Program_weight_tensor_parameter_114: name = "parameter_114" + original_name = "layer_norm_33.w_0" shape = [768] dtype = "float32" min_val = float("0.145274") @@ -1265,6 +1380,7 @@ class Program_weight_tensor_parameter_114: class Program_weight_tensor_parameter_115: name = "parameter_115" + original_name = "linear_61.b_0" shape = [768] dtype = "float32" min_val = float("-4.88506") @@ -1276,6 +1392,7 @@ class Program_weight_tensor_parameter_115: class Program_weight_tensor_parameter_116: name = "parameter_116" + original_name = "linear_61.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-1.00765") @@ -1287,6 +1404,7 @@ class Program_weight_tensor_parameter_116: class Program_weight_tensor_parameter_117: name = "parameter_117" + original_name = "linear_60.b_0" shape = [3072] dtype = "float32" min_val = float("-2.84458") @@ -1298,6 +1416,7 @@ class Program_weight_tensor_parameter_117: class Program_weight_tensor_parameter_118: name = "parameter_118" + original_name = "linear_60.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.705352") @@ -1309,6 +1428,7 @@ class Program_weight_tensor_parameter_118: class Program_weight_tensor_parameter_119: name = "parameter_119" + original_name = "layer_norm_32.b_0" shape = [768] dtype = "float32" min_val = float("-1.73371") @@ -1320,6 +1440,7 @@ class Program_weight_tensor_parameter_119: class Program_weight_tensor_parameter_120: name = "parameter_120" + original_name = "layer_norm_32.w_0" shape = [768] dtype = "float32" min_val = float("0.00373687") @@ -1331,6 +1452,7 @@ class Program_weight_tensor_parameter_120: class Program_weight_tensor_parameter_121: name = "parameter_121" + original_name = "linear_59.b_0" shape = [768] dtype = "float32" min_val = float("-1.87609") @@ -1342,6 +1464,7 @@ class Program_weight_tensor_parameter_121: class Program_weight_tensor_parameter_122: name = "parameter_122" + original_name = "linear_59.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.592065") @@ -1353,6 +1476,7 @@ class Program_weight_tensor_parameter_122: class Program_weight_tensor_parameter_123: name = "parameter_123" + original_name = "linear_58.b_0" shape = [2304] dtype = "float32" min_val = float("-4.9161") @@ -1364,6 +1488,7 @@ class Program_weight_tensor_parameter_123: class Program_weight_tensor_parameter_124: name = "parameter_124" + original_name = "linear_58.w_0" shape = [768, 2304] dtype = "float32" min_val = float("-0.590309") @@ -1375,6 +1500,7 @@ class Program_weight_tensor_parameter_124: class Program_weight_tensor_parameter_125: name = "parameter_125" + original_name = "layer_norm_31.b_0" shape = [768] dtype = "float32" min_val = float("-3.35951") @@ -1386,6 +1512,7 @@ class Program_weight_tensor_parameter_125: class Program_weight_tensor_parameter_126: name = "parameter_126" + original_name = "layer_norm_31.w_0" shape = [768] dtype = "float32" min_val = float("0.000981313") @@ -1397,6 +1524,7 @@ class Program_weight_tensor_parameter_126: class Program_weight_tensor_parameter_127: name = "parameter_127" + original_name = "linear_57.b_0" shape = [768] dtype = "float32" min_val = float("-2.50776") @@ -1408,6 +1536,7 @@ class Program_weight_tensor_parameter_127: class Program_weight_tensor_parameter_128: name = "parameter_128" + original_name = "linear_57.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-0.906142") @@ -1419,6 +1548,7 @@ class Program_weight_tensor_parameter_128: class Program_weight_tensor_parameter_129: name = "parameter_129" + original_name = "linear_56.b_0" shape = [3072] dtype = "float32" min_val = float("-3.35046") @@ -1430,6 +1560,7 @@ class Program_weight_tensor_parameter_129: class Program_weight_tensor_parameter_130: name = "parameter_130" + original_name = "linear_56.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.577671") @@ -1441,6 +1572,7 @@ class Program_weight_tensor_parameter_130: class Program_weight_tensor_parameter_131: name = "parameter_131" + original_name = "layer_norm_30.b_0" shape = [768] dtype = "float32" min_val = float("-1.79581") @@ -1452,6 +1584,7 @@ class Program_weight_tensor_parameter_131: class Program_weight_tensor_parameter_132: name = "parameter_132" + original_name = "layer_norm_30.w_0" shape = [768] dtype = "float32" min_val = float("0.00188213") @@ -1463,6 +1596,7 @@ class Program_weight_tensor_parameter_132: class Program_weight_tensor_parameter_133: name = "parameter_133" + original_name = "linear_55.b_0" shape = [768] dtype = "float32" min_val = float("-1.02941") @@ -1474,6 +1608,7 @@ class Program_weight_tensor_parameter_133: class Program_weight_tensor_parameter_134: name = "parameter_134" + original_name = "linear_55.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.681398") @@ -1485,6 +1620,7 @@ class Program_weight_tensor_parameter_134: class Program_weight_tensor_parameter_135: name = "parameter_135" + original_name = "linear_54.b_0" shape = [2304] dtype = "float32" min_val = float("-5.13255") @@ -1496,6 +1632,7 @@ class Program_weight_tensor_parameter_135: class Program_weight_tensor_parameter_136: name = "parameter_136" + original_name = "linear_54.w_0" shape = [768, 2304] dtype = "float32" min_val = float("-0.708263") @@ -1507,6 +1644,7 @@ class Program_weight_tensor_parameter_136: class Program_weight_tensor_parameter_137: name = "parameter_137" + original_name = "layer_norm_29.b_0" shape = [768] dtype = "float32" min_val = float("-3.26672") @@ -1518,6 +1656,7 @@ class Program_weight_tensor_parameter_137: class Program_weight_tensor_parameter_138: name = "parameter_138" + original_name = "layer_norm_29.w_0" shape = [768] dtype = "float32" min_val = float("-0.00720201") @@ -1529,6 +1668,7 @@ class Program_weight_tensor_parameter_138: class Program_weight_tensor_parameter_139: name = "parameter_139" + original_name = "linear_53.b_0" shape = [768] dtype = "float32" min_val = float("-1.42344") @@ -1540,6 +1680,7 @@ class Program_weight_tensor_parameter_139: class Program_weight_tensor_parameter_140: name = "parameter_140" + original_name = "linear_53.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-0.944526") @@ -1551,6 +1692,7 @@ class Program_weight_tensor_parameter_140: class Program_weight_tensor_parameter_141: name = "parameter_141" + original_name = "linear_52.b_0" shape = [3072] dtype = "float32" min_val = float("-4.2021") @@ -1562,6 +1704,7 @@ class Program_weight_tensor_parameter_141: class Program_weight_tensor_parameter_142: name = "parameter_142" + original_name = "linear_52.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.489216") @@ -1573,6 +1716,7 @@ class Program_weight_tensor_parameter_142: class Program_weight_tensor_parameter_143: name = "parameter_143" + original_name = "layer_norm_28.b_0" shape = [768] dtype = "float32" min_val = float("-1.71688") @@ -1584,6 +1728,7 @@ class Program_weight_tensor_parameter_143: class Program_weight_tensor_parameter_144: name = "parameter_144" + original_name = "layer_norm_28.w_0" shape = [768] dtype = "float32" min_val = float("0.00891985") @@ -1595,6 +1740,7 @@ class Program_weight_tensor_parameter_144: class Program_weight_tensor_parameter_145: name = "parameter_145" + original_name = "linear_51.b_0" shape = [768] dtype = "float32" min_val = float("-1.78121") @@ -1606,6 +1752,7 @@ class Program_weight_tensor_parameter_145: class Program_weight_tensor_parameter_146: name = "parameter_146" + original_name = "linear_51.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.719495") @@ -1617,6 +1764,7 @@ class Program_weight_tensor_parameter_146: class Program_weight_tensor_parameter_147: name = "parameter_147" + original_name = "linear_50.b_0" shape = [2304] dtype = "float32" min_val = float("-4.00149") @@ -1628,6 +1776,7 @@ class Program_weight_tensor_parameter_147: class Program_weight_tensor_parameter_148: name = "parameter_148" + original_name = "linear_50.w_0" shape = [768, 2304] dtype = "float32" min_val = float("-0.668832") @@ -1639,6 +1788,7 @@ class Program_weight_tensor_parameter_148: class Program_weight_tensor_parameter_149: name = "parameter_149" + original_name = "layer_norm_27.b_0" shape = [768] dtype = "float32" min_val = float("-3.16254") @@ -1650,6 +1800,7 @@ class Program_weight_tensor_parameter_149: class Program_weight_tensor_parameter_150: name = "parameter_150" + original_name = "layer_norm_27.w_0" shape = [768] dtype = "float32" min_val = float("0.0975965") @@ -1661,6 +1812,7 @@ class Program_weight_tensor_parameter_150: class Program_weight_tensor_parameter_151: name = "parameter_151" + original_name = "linear_49.b_0" shape = [768] dtype = "float32" min_val = float("-1.04035") @@ -1672,6 +1824,7 @@ class Program_weight_tensor_parameter_151: class Program_weight_tensor_parameter_152: name = "parameter_152" + original_name = "linear_49.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-0.952044") @@ -1683,6 +1836,7 @@ class Program_weight_tensor_parameter_152: class Program_weight_tensor_parameter_153: name = "parameter_153" + original_name = "linear_48.b_0" shape = [3072] dtype = "float32" min_val = float("-3.59193") @@ -1694,6 +1848,7 @@ class Program_weight_tensor_parameter_153: class Program_weight_tensor_parameter_154: name = "parameter_154" + original_name = "linear_48.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.560203") @@ -1705,6 +1860,7 @@ class Program_weight_tensor_parameter_154: class Program_weight_tensor_parameter_155: name = "parameter_155" + original_name = "layer_norm_26.b_0" shape = [768] dtype = "float32" min_val = float("-1.38039") @@ -1716,6 +1872,7 @@ class Program_weight_tensor_parameter_155: class Program_weight_tensor_parameter_156: name = "parameter_156" + original_name = "layer_norm_26.w_0" shape = [768] dtype = "float32" min_val = float("-0.00160366") @@ -1727,6 +1884,7 @@ class Program_weight_tensor_parameter_156: class Program_weight_tensor_parameter_157: name = "parameter_157" + original_name = "linear_47.b_0" shape = [768] dtype = "float32" min_val = float("-1.38299") @@ -1738,6 +1896,7 @@ class Program_weight_tensor_parameter_157: class Program_weight_tensor_parameter_158: name = "parameter_158" + original_name = "linear_47.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.527948") @@ -1749,6 +1908,7 @@ class Program_weight_tensor_parameter_158: class Program_weight_tensor_parameter_159: name = "parameter_159" + original_name = "linear_46.b_0" shape = [2304] dtype = "float32" min_val = float("-3.63507") @@ -1760,6 +1920,7 @@ class Program_weight_tensor_parameter_159: class Program_weight_tensor_parameter_160: name = "parameter_160" + original_name = "linear_46.w_0" shape = [768, 2304] dtype = "float32" min_val = float("-0.617454") @@ -1771,6 +1932,7 @@ class Program_weight_tensor_parameter_160: class Program_weight_tensor_parameter_161: name = "parameter_161" + original_name = "layer_norm_25.b_0" shape = [768] dtype = "float32" min_val = float("-2.60017") @@ -1782,6 +1944,7 @@ class Program_weight_tensor_parameter_161: class Program_weight_tensor_parameter_162: name = "parameter_162" + original_name = "layer_norm_25.w_0" shape = [768] dtype = "float32" min_val = float("0.0308432") @@ -1793,6 +1956,7 @@ class Program_weight_tensor_parameter_162: class Program_weight_tensor_parameter_163: name = "parameter_163" + original_name = "linear_45.b_0" shape = [768] dtype = "float32" min_val = float("-1.08633") @@ -1804,6 +1968,7 @@ class Program_weight_tensor_parameter_163: class Program_weight_tensor_parameter_164: name = "parameter_164" + original_name = "linear_45.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-0.665865") @@ -1815,6 +1980,7 @@ class Program_weight_tensor_parameter_164: class Program_weight_tensor_parameter_165: name = "parameter_165" + original_name = "linear_44.b_0" shape = [3072] dtype = "float32" min_val = float("-3.28988") @@ -1826,6 +1992,7 @@ class Program_weight_tensor_parameter_165: class Program_weight_tensor_parameter_166: name = "parameter_166" + original_name = "linear_44.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.491601") @@ -1837,6 +2004,7 @@ class Program_weight_tensor_parameter_166: class Program_weight_tensor_parameter_167: name = "parameter_167" + original_name = "layer_norm_24.b_0" shape = [768] dtype = "float32" min_val = float("-1.56267") @@ -1848,6 +2016,7 @@ class Program_weight_tensor_parameter_167: class Program_weight_tensor_parameter_168: name = "parameter_168" + original_name = "layer_norm_24.w_0" shape = [768] dtype = "float32" min_val = float("0.0186232") @@ -1859,6 +2028,7 @@ class Program_weight_tensor_parameter_168: class Program_weight_tensor_parameter_169: name = "parameter_169" + original_name = "linear_43.b_0" shape = [768] dtype = "float32" min_val = float("-1.44051") @@ -1870,6 +2040,7 @@ class Program_weight_tensor_parameter_169: class Program_weight_tensor_parameter_170: name = "parameter_170" + original_name = "linear_43.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.617444") @@ -1881,6 +2052,7 @@ class Program_weight_tensor_parameter_170: class Program_weight_tensor_parameter_171: name = "parameter_171" + original_name = "linear_42.b_0" shape = [2304] dtype = "float32" min_val = float("-3.43671") @@ -1892,6 +2064,7 @@ class Program_weight_tensor_parameter_171: class Program_weight_tensor_parameter_172: name = "parameter_172" + original_name = "linear_42.w_0" shape = [768, 2304] dtype = "float32" min_val = float("-0.574165") @@ -1903,6 +2076,7 @@ class Program_weight_tensor_parameter_172: class Program_weight_tensor_parameter_173: name = "parameter_173" + original_name = "layer_norm_23.b_0" shape = [768] dtype = "float32" min_val = float("-2.33937") @@ -1914,6 +2088,7 @@ class Program_weight_tensor_parameter_173: class Program_weight_tensor_parameter_174: name = "parameter_174" + original_name = "layer_norm_23.w_0" shape = [768] dtype = "float32" min_val = float("0.0177852") @@ -1925,6 +2100,7 @@ class Program_weight_tensor_parameter_174: class Program_weight_tensor_parameter_175: name = "parameter_175" + original_name = "linear_41.b_0" shape = [768] dtype = "float32" min_val = float("-0.921831") @@ -1936,6 +2112,7 @@ class Program_weight_tensor_parameter_175: class Program_weight_tensor_parameter_176: name = "parameter_176" + original_name = "linear_41.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-0.628339") @@ -1947,6 +2124,7 @@ class Program_weight_tensor_parameter_176: class Program_weight_tensor_parameter_177: name = "parameter_177" + original_name = "linear_40.b_0" shape = [3072] dtype = "float32" min_val = float("-3.43158") @@ -1958,6 +2136,7 @@ class Program_weight_tensor_parameter_177: class Program_weight_tensor_parameter_178: name = "parameter_178" + original_name = "linear_40.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.724141") @@ -1969,6 +2148,7 @@ class Program_weight_tensor_parameter_178: class Program_weight_tensor_parameter_179: name = "parameter_179" + original_name = "layer_norm_22.b_0" shape = [768] dtype = "float32" min_val = float("-1.88324") @@ -1980,6 +2160,7 @@ class Program_weight_tensor_parameter_179: class Program_weight_tensor_parameter_180: name = "parameter_180" + original_name = "layer_norm_22.w_0" shape = [768] dtype = "float32" min_val = float("0.012379") @@ -1991,6 +2172,7 @@ class Program_weight_tensor_parameter_180: class Program_weight_tensor_parameter_181: name = "parameter_181" + original_name = "linear_39.b_0" shape = [768] dtype = "float32" min_val = float("-1.45943") @@ -2002,6 +2184,7 @@ class Program_weight_tensor_parameter_181: class Program_weight_tensor_parameter_182: name = "parameter_182" + original_name = "linear_39.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.596271") @@ -2013,6 +2196,7 @@ class Program_weight_tensor_parameter_182: class Program_weight_tensor_parameter_183: name = "parameter_183" + original_name = "linear_38.b_0" shape = [2304] dtype = "float32" min_val = float("-3.38891") @@ -2024,6 +2208,7 @@ class Program_weight_tensor_parameter_183: class Program_weight_tensor_parameter_184: name = "parameter_184" + original_name = "linear_38.w_0" shape = [768, 2304] dtype = "float32" min_val = float("-0.79643") @@ -2035,6 +2220,7 @@ class Program_weight_tensor_parameter_184: class Program_weight_tensor_parameter_185: name = "parameter_185" + original_name = "layer_norm_21.b_0" shape = [768] dtype = "float32" min_val = float("-2.41399") @@ -2046,6 +2232,7 @@ class Program_weight_tensor_parameter_185: class Program_weight_tensor_parameter_186: name = "parameter_186" + original_name = "layer_norm_21.w_0" shape = [768] dtype = "float32" min_val = float("0.00641553") @@ -2057,6 +2244,7 @@ class Program_weight_tensor_parameter_186: class Program_weight_tensor_parameter_187: name = "parameter_187" + original_name = "linear_37.b_0" shape = [768] dtype = "float32" min_val = float("-0.965009") @@ -2068,6 +2256,7 @@ class Program_weight_tensor_parameter_187: class Program_weight_tensor_parameter_188: name = "parameter_188" + original_name = "linear_37.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-0.793557") @@ -2079,6 +2268,7 @@ class Program_weight_tensor_parameter_188: class Program_weight_tensor_parameter_189: name = "parameter_189" + original_name = "linear_36.b_0" shape = [3072] dtype = "float32" min_val = float("-3.51897") @@ -2090,6 +2280,7 @@ class Program_weight_tensor_parameter_189: class Program_weight_tensor_parameter_190: name = "parameter_190" + original_name = "linear_36.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.547397") @@ -2101,6 +2292,7 @@ class Program_weight_tensor_parameter_190: class Program_weight_tensor_parameter_191: name = "parameter_191" + original_name = "layer_norm_20.b_0" shape = [768] dtype = "float32" min_val = float("-1.83248") @@ -2112,6 +2304,7 @@ class Program_weight_tensor_parameter_191: class Program_weight_tensor_parameter_192: name = "parameter_192" + original_name = "layer_norm_20.w_0" shape = [768] dtype = "float32" min_val = float("0.00346064") @@ -2123,6 +2316,7 @@ class Program_weight_tensor_parameter_192: class Program_weight_tensor_parameter_193: name = "parameter_193" + original_name = "linear_35.b_0" shape = [768] dtype = "float32" min_val = float("-1.65959") @@ -2134,6 +2328,7 @@ class Program_weight_tensor_parameter_193: class Program_weight_tensor_parameter_194: name = "parameter_194" + original_name = "linear_35.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.765161") @@ -2145,6 +2340,7 @@ class Program_weight_tensor_parameter_194: class Program_weight_tensor_parameter_195: name = "parameter_195" + original_name = "linear_34.b_0" shape = [2304] dtype = "float32" min_val = float("-4.22842") @@ -2156,6 +2352,7 @@ class Program_weight_tensor_parameter_195: class Program_weight_tensor_parameter_196: name = "parameter_196" + original_name = "linear_34.w_0" shape = [768, 2304] dtype = "float32" min_val = float("-0.936553") @@ -2167,6 +2364,7 @@ class Program_weight_tensor_parameter_196: class Program_weight_tensor_parameter_197: name = "parameter_197" + original_name = "layer_norm_19.b_0" shape = [768] dtype = "float32" min_val = float("-2.57213") @@ -2178,6 +2376,7 @@ class Program_weight_tensor_parameter_197: class Program_weight_tensor_parameter_198: name = "parameter_198" + original_name = "layer_norm_19.w_0" shape = [768] dtype = "float32" min_val = float("-0.00137983") @@ -2189,6 +2388,7 @@ class Program_weight_tensor_parameter_198: class Program_weight_tensor_parameter_199: name = "parameter_199" + original_name = "linear_33.b_0" shape = [768] dtype = "float32" min_val = float("-1.30937") @@ -2200,6 +2400,7 @@ class Program_weight_tensor_parameter_199: class Program_weight_tensor_parameter_200: name = "parameter_200" + original_name = "linear_33.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-0.60018") @@ -2211,6 +2412,7 @@ class Program_weight_tensor_parameter_200: class Program_weight_tensor_parameter_201: name = "parameter_201" + original_name = "linear_32.b_0" shape = [3072] dtype = "float32" min_val = float("-3.68851") @@ -2222,6 +2424,7 @@ class Program_weight_tensor_parameter_201: class Program_weight_tensor_parameter_202: name = "parameter_202" + original_name = "linear_32.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.489597") @@ -2233,6 +2436,7 @@ class Program_weight_tensor_parameter_202: class Program_weight_tensor_parameter_203: name = "parameter_203" + original_name = "layer_norm_18.b_0" shape = [768] dtype = "float32" min_val = float("-1.84991") @@ -2244,6 +2448,7 @@ class Program_weight_tensor_parameter_203: class Program_weight_tensor_parameter_204: name = "parameter_204" + original_name = "layer_norm_18.w_0" shape = [768] dtype = "float32" min_val = float("-0.00775934") @@ -2255,6 +2460,7 @@ class Program_weight_tensor_parameter_204: class Program_weight_tensor_parameter_205: name = "parameter_205" + original_name = "linear_31.b_0" shape = [768] dtype = "float32" min_val = float("-1.0151") @@ -2266,6 +2472,7 @@ class Program_weight_tensor_parameter_205: class Program_weight_tensor_parameter_206: name = "parameter_206" + original_name = "linear_31.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.571353") @@ -2277,6 +2484,7 @@ class Program_weight_tensor_parameter_206: class Program_weight_tensor_parameter_207: name = "parameter_207" + original_name = "linear_30.b_0" shape = [2304] dtype = "float32" min_val = float("-3.49578") @@ -2288,6 +2496,7 @@ class Program_weight_tensor_parameter_207: class Program_weight_tensor_parameter_208: name = "parameter_208" + original_name = "linear_30.w_0" shape = [768, 2304] dtype = "float32" min_val = float("-0.520641") @@ -2299,6 +2508,7 @@ class Program_weight_tensor_parameter_208: class Program_weight_tensor_parameter_209: name = "parameter_209" + original_name = "layer_norm_17.b_0" shape = [768] dtype = "float32" min_val = float("-2.19212") @@ -2310,6 +2520,7 @@ class Program_weight_tensor_parameter_209: class Program_weight_tensor_parameter_210: name = "parameter_210" + original_name = "layer_norm_17.w_0" shape = [768] dtype = "float32" min_val = float("-0.00594886") @@ -2321,6 +2532,7 @@ class Program_weight_tensor_parameter_210: class Program_weight_tensor_parameter_211: name = "parameter_211" + original_name = "linear_29.b_0" shape = [768] dtype = "float32" min_val = float("-1.26396") @@ -2332,6 +2544,7 @@ class Program_weight_tensor_parameter_211: class Program_weight_tensor_parameter_212: name = "parameter_212" + original_name = "linear_29.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-0.557577") @@ -2343,6 +2556,7 @@ class Program_weight_tensor_parameter_212: class Program_weight_tensor_parameter_213: name = "parameter_213" + original_name = "linear_28.b_0" shape = [3072] dtype = "float32" min_val = float("-3.76427") @@ -2354,6 +2568,7 @@ class Program_weight_tensor_parameter_213: class Program_weight_tensor_parameter_214: name = "parameter_214" + original_name = "linear_28.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.572513") @@ -2365,6 +2580,7 @@ class Program_weight_tensor_parameter_214: class Program_weight_tensor_parameter_215: name = "parameter_215" + original_name = "layer_norm_16.b_0" shape = [768] dtype = "float32" min_val = float("-1.84786") @@ -2376,6 +2592,7 @@ class Program_weight_tensor_parameter_215: class Program_weight_tensor_parameter_216: name = "parameter_216" + original_name = "layer_norm_16.w_0" shape = [768] dtype = "float32" min_val = float("-0.0173741") @@ -2387,6 +2604,7 @@ class Program_weight_tensor_parameter_216: class Program_weight_tensor_parameter_217: name = "parameter_217" + original_name = "linear_27.b_0" shape = [768] dtype = "float32" min_val = float("-1.45167") @@ -2398,6 +2616,7 @@ class Program_weight_tensor_parameter_217: class Program_weight_tensor_parameter_218: name = "parameter_218" + original_name = "linear_27.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.402874") @@ -2409,6 +2628,7 @@ class Program_weight_tensor_parameter_218: class Program_weight_tensor_parameter_219: name = "parameter_219" + original_name = "linear_26.b_0" shape = [2304] dtype = "float32" min_val = float("-4.17524") @@ -2420,6 +2640,7 @@ class Program_weight_tensor_parameter_219: class Program_weight_tensor_parameter_220: name = "parameter_220" + original_name = "linear_26.w_0" shape = [768, 2304] dtype = "float32" min_val = float("-0.622484") @@ -2431,6 +2652,7 @@ class Program_weight_tensor_parameter_220: class Program_weight_tensor_parameter_221: name = "parameter_221" + original_name = "layer_norm_15.b_0" shape = [768] dtype = "float32" min_val = float("-2.56081") @@ -2442,6 +2664,7 @@ class Program_weight_tensor_parameter_221: class Program_weight_tensor_parameter_222: name = "parameter_222" + original_name = "layer_norm_15.w_0" shape = [768] dtype = "float32" min_val = float("-0.00879466") @@ -2453,6 +2676,7 @@ class Program_weight_tensor_parameter_222: class Program_weight_tensor_parameter_223: name = "parameter_223" + original_name = "linear_25.b_0" shape = [768] dtype = "float32" min_val = float("-1.01241") @@ -2464,6 +2688,7 @@ class Program_weight_tensor_parameter_223: class Program_weight_tensor_parameter_224: name = "parameter_224" + original_name = "linear_25.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-0.649975") @@ -2475,6 +2700,7 @@ class Program_weight_tensor_parameter_224: class Program_weight_tensor_parameter_225: name = "parameter_225" + original_name = "linear_24.b_0" shape = [3072] dtype = "float32" min_val = float("-4.14149") @@ -2486,6 +2712,7 @@ class Program_weight_tensor_parameter_225: class Program_weight_tensor_parameter_226: name = "parameter_226" + original_name = "linear_24.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.458694") @@ -2497,6 +2724,7 @@ class Program_weight_tensor_parameter_226: class Program_weight_tensor_parameter_227: name = "parameter_227" + original_name = "layer_norm_14.b_0" shape = [768] dtype = "float32" min_val = float("-1.95884") @@ -2508,6 +2736,7 @@ class Program_weight_tensor_parameter_227: class Program_weight_tensor_parameter_228: name = "parameter_228" + original_name = "layer_norm_14.w_0" shape = [768] dtype = "float32" min_val = float("-0.0202645") @@ -2519,6 +2748,7 @@ class Program_weight_tensor_parameter_228: class Program_weight_tensor_parameter_229: name = "parameter_229" + original_name = "linear_23.b_0" shape = [768] dtype = "float32" min_val = float("-1.33536") @@ -2530,6 +2760,7 @@ class Program_weight_tensor_parameter_229: class Program_weight_tensor_parameter_230: name = "parameter_230" + original_name = "linear_23.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.596133") @@ -2541,6 +2772,7 @@ class Program_weight_tensor_parameter_230: class Program_weight_tensor_parameter_231: name = "parameter_231" + original_name = "linear_22.b_0" shape = [2304] dtype = "float32" min_val = float("-3.70939") @@ -2552,6 +2784,7 @@ class Program_weight_tensor_parameter_231: class Program_weight_tensor_parameter_232: name = "parameter_232" + original_name = "linear_22.w_0" shape = [768, 2304] dtype = "float32" min_val = float("-0.534558") @@ -2563,6 +2796,7 @@ class Program_weight_tensor_parameter_232: class Program_weight_tensor_parameter_233: name = "parameter_233" + original_name = "layer_norm_13.b_0" shape = [768] dtype = "float32" min_val = float("-1.36202") @@ -2574,6 +2808,7 @@ class Program_weight_tensor_parameter_233: class Program_weight_tensor_parameter_234: name = "parameter_234" + original_name = "layer_norm_13.w_0" shape = [768] dtype = "float32" min_val = float("-0.0166836") @@ -2585,6 +2820,7 @@ class Program_weight_tensor_parameter_234: class Program_weight_tensor_parameter_235: name = "parameter_235" + original_name = "linear_21.b_0" shape = [768] dtype = "float32" min_val = float("-1.36359") @@ -2596,6 +2832,7 @@ class Program_weight_tensor_parameter_235: class Program_weight_tensor_parameter_236: name = "parameter_236" + original_name = "linear_21.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-0.494304") @@ -2607,6 +2844,7 @@ class Program_weight_tensor_parameter_236: class Program_weight_tensor_parameter_237: name = "parameter_237" + original_name = "linear_20.b_0" shape = [3072] dtype = "float32" min_val = float("-4.37069") @@ -2618,6 +2856,7 @@ class Program_weight_tensor_parameter_237: class Program_weight_tensor_parameter_238: name = "parameter_238" + original_name = "linear_20.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.40097") @@ -2629,6 +2868,7 @@ class Program_weight_tensor_parameter_238: class Program_weight_tensor_parameter_239: name = "parameter_239" + original_name = "layer_norm_12.b_0" shape = [768] dtype = "float32" min_val = float("-1.94018") @@ -2640,6 +2880,7 @@ class Program_weight_tensor_parameter_239: class Program_weight_tensor_parameter_240: name = "parameter_240" + original_name = "layer_norm_12.w_0" shape = [768] dtype = "float32" min_val = float("-0.0241882") @@ -2651,6 +2892,7 @@ class Program_weight_tensor_parameter_240: class Program_weight_tensor_parameter_241: name = "parameter_241" + original_name = "linear_19.b_0" shape = [768] dtype = "float32" min_val = float("-1.52935") @@ -2662,6 +2904,7 @@ class Program_weight_tensor_parameter_241: class Program_weight_tensor_parameter_242: name = "parameter_242" + original_name = "linear_19.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.527912") @@ -2673,6 +2916,7 @@ class Program_weight_tensor_parameter_242: class Program_weight_tensor_parameter_243: name = "parameter_243" + original_name = "linear_18.b_0" shape = [2304] dtype = "float32" min_val = float("-3.9033") @@ -2684,6 +2928,7 @@ class Program_weight_tensor_parameter_243: class Program_weight_tensor_parameter_244: name = "parameter_244" + original_name = "linear_18.w_0" shape = [768, 2304] dtype = "float32" min_val = float("-0.538067") @@ -2695,6 +2940,7 @@ class Program_weight_tensor_parameter_244: class Program_weight_tensor_parameter_245: name = "parameter_245" + original_name = "layer_norm_11.b_0" shape = [768] dtype = "float32" min_val = float("-0.585253") @@ -2706,6 +2952,7 @@ class Program_weight_tensor_parameter_245: class Program_weight_tensor_parameter_246: name = "parameter_246" + original_name = "layer_norm_11.w_0" shape = [768] dtype = "float32" min_val = float("-0.0088233") @@ -2717,6 +2964,7 @@ class Program_weight_tensor_parameter_246: class Program_weight_tensor_parameter_247: name = "parameter_247" + original_name = "linear_17.w_0" shape = [1536, 768] dtype = "float32" min_val = float("-0.389695") @@ -2728,6 +2976,7 @@ class Program_weight_tensor_parameter_247: class Program_weight_tensor_parameter_248: name = "parameter_248" + original_name = "layer_norm_10.b_0" shape = [1536] dtype = "float32" min_val = float("-1.20127") @@ -2739,6 +2988,7 @@ class Program_weight_tensor_parameter_248: class Program_weight_tensor_parameter_249: name = "parameter_249" + original_name = "layer_norm_10.w_0" shape = [1536] dtype = "float32" min_val = float("0.147333") @@ -2750,6 +3000,7 @@ class Program_weight_tensor_parameter_249: class Program_weight_tensor_parameter_250: name = "parameter_250" + original_name = "linear_16.b_0" shape = [384] dtype = "float32" min_val = float("-7.38438") @@ -2761,6 +3012,7 @@ class Program_weight_tensor_parameter_250: class Program_weight_tensor_parameter_251: name = "parameter_251" + original_name = "linear_16.w_0" shape = [1536, 384] dtype = "float32" min_val = float("-0.93667") @@ -2772,6 +3024,7 @@ class Program_weight_tensor_parameter_251: class Program_weight_tensor_parameter_252: name = "parameter_252" + original_name = "linear_15.b_0" shape = [1536] dtype = "float32" min_val = float("-2.63423") @@ -2783,6 +3036,7 @@ class Program_weight_tensor_parameter_252: class Program_weight_tensor_parameter_253: name = "parameter_253" + original_name = "linear_15.w_0" shape = [384, 1536] dtype = "float32" min_val = float("-0.47853") @@ -2794,6 +3048,7 @@ class Program_weight_tensor_parameter_253: class Program_weight_tensor_parameter_254: name = "parameter_254" + original_name = "layer_norm_9.b_0" shape = [384] dtype = "float32" min_val = float("-1.92586") @@ -2805,6 +3060,7 @@ class Program_weight_tensor_parameter_254: class Program_weight_tensor_parameter_255: name = "parameter_255" + original_name = "layer_norm_9.w_0" shape = [384] dtype = "float32" min_val = float("0.0040841") @@ -2816,6 +3072,7 @@ class Program_weight_tensor_parameter_255: class Program_weight_tensor_parameter_256: name = "parameter_256" + original_name = "linear_14.b_0" shape = [384] dtype = "float32" min_val = float("-6.33508") @@ -2827,6 +3084,7 @@ class Program_weight_tensor_parameter_256: class Program_weight_tensor_parameter_257: name = "parameter_257" + original_name = "linear_14.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.337339") @@ -2838,6 +3096,7 @@ class Program_weight_tensor_parameter_257: class Program_weight_tensor_parameter_258: name = "parameter_258" + original_name = "linear_13.b_0" shape = [1152] dtype = "float32" min_val = float("-3.23663") @@ -2849,6 +3108,7 @@ class Program_weight_tensor_parameter_258: class Program_weight_tensor_parameter_259: name = "parameter_259" + original_name = "linear_13.w_0" shape = [384, 1152] dtype = "float32" min_val = float("-0.714906") @@ -2860,6 +3120,7 @@ class Program_weight_tensor_parameter_259: class Program_weight_tensor_parameter_260: name = "parameter_260" + original_name = "layer_norm_8.b_0" shape = [384] dtype = "float32" min_val = float("-1.47529") @@ -2871,6 +3132,7 @@ class Program_weight_tensor_parameter_260: class Program_weight_tensor_parameter_261: name = "parameter_261" + original_name = "layer_norm_8.w_0" shape = [384] dtype = "float32" min_val = float("0.133417") @@ -2882,6 +3144,7 @@ class Program_weight_tensor_parameter_261: class Program_weight_tensor_parameter_262: name = "parameter_262" + original_name = "linear_12.b_0" shape = [384] dtype = "float32" min_val = float("-5.82546") @@ -2893,6 +3156,7 @@ class Program_weight_tensor_parameter_262: class Program_weight_tensor_parameter_263: name = "parameter_263" + original_name = "linear_12.w_0" shape = [1536, 384] dtype = "float32" min_val = float("-0.457301") @@ -2904,6 +3168,7 @@ class Program_weight_tensor_parameter_263: class Program_weight_tensor_parameter_264: name = "parameter_264" + original_name = "linear_11.b_0" shape = [1536] dtype = "float32" min_val = float("-2.80096") @@ -2915,6 +3180,7 @@ class Program_weight_tensor_parameter_264: class Program_weight_tensor_parameter_265: name = "parameter_265" + original_name = "linear_11.w_0" shape = [384, 1536] dtype = "float32" min_val = float("-0.422604") @@ -2926,6 +3192,7 @@ class Program_weight_tensor_parameter_265: class Program_weight_tensor_parameter_266: name = "parameter_266" + original_name = "layer_norm_7.b_0" shape = [384] dtype = "float32" min_val = float("-0.819488") @@ -2937,6 +3204,7 @@ class Program_weight_tensor_parameter_266: class Program_weight_tensor_parameter_267: name = "parameter_267" + original_name = "layer_norm_7.w_0" shape = [384] dtype = "float32" min_val = float("-0.00822958") @@ -2948,6 +3216,7 @@ class Program_weight_tensor_parameter_267: class Program_weight_tensor_parameter_268: name = "parameter_268" + original_name = "linear_10.b_0" shape = [384] dtype = "float32" min_val = float("-5.73781") @@ -2959,6 +3228,7 @@ class Program_weight_tensor_parameter_268: class Program_weight_tensor_parameter_269: name = "parameter_269" + original_name = "linear_10.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.413975") @@ -2970,6 +3240,7 @@ class Program_weight_tensor_parameter_269: class Program_weight_tensor_parameter_270: name = "parameter_270" + original_name = "linear_9.b_0" shape = [1152] dtype = "float32" min_val = float("-2.78681") @@ -2981,6 +3252,7 @@ class Program_weight_tensor_parameter_270: class Program_weight_tensor_parameter_271: name = "parameter_271" + original_name = "linear_9.w_0" shape = [384, 1152] dtype = "float32" min_val = float("-0.561243") @@ -2992,6 +3264,7 @@ class Program_weight_tensor_parameter_271: class Program_weight_tensor_parameter_272: name = "parameter_272" + original_name = "layer_norm_6.b_0" shape = [384] dtype = "float32" min_val = float("-0.536764") @@ -3003,6 +3276,7 @@ class Program_weight_tensor_parameter_272: class Program_weight_tensor_parameter_273: name = "parameter_273" + original_name = "layer_norm_6.w_0" shape = [384] dtype = "float32" min_val = float("-0.00844302") @@ -3014,6 +3288,7 @@ class Program_weight_tensor_parameter_273: class Program_weight_tensor_parameter_274: name = "parameter_274" + original_name = "linear_8.w_0" shape = [768, 384] dtype = "float32" min_val = float("-0.435344") @@ -3025,6 +3300,7 @@ class Program_weight_tensor_parameter_274: class Program_weight_tensor_parameter_275: name = "parameter_275" + original_name = "layer_norm_5.b_0" shape = [768] dtype = "float32" min_val = float("-1.45818") @@ -3036,6 +3312,7 @@ class Program_weight_tensor_parameter_275: class Program_weight_tensor_parameter_276: name = "parameter_276" + original_name = "layer_norm_5.w_0" shape = [768] dtype = "float32" min_val = float("0.151278") @@ -3047,6 +3324,7 @@ class Program_weight_tensor_parameter_276: class Program_weight_tensor_parameter_277: name = "parameter_277" + original_name = "linear_7.b_0" shape = [192] dtype = "float32" min_val = float("-6.26595") @@ -3058,6 +3336,7 @@ class Program_weight_tensor_parameter_277: class Program_weight_tensor_parameter_278: name = "parameter_278" + original_name = "linear_7.w_0" shape = [768, 192] dtype = "float32" min_val = float("-0.834608") @@ -3069,6 +3348,7 @@ class Program_weight_tensor_parameter_278: class Program_weight_tensor_parameter_279: name = "parameter_279" + original_name = "linear_6.b_0" shape = [768] dtype = "float32" min_val = float("-2.60017") @@ -3080,6 +3360,7 @@ class Program_weight_tensor_parameter_279: class Program_weight_tensor_parameter_280: name = "parameter_280" + original_name = "linear_6.w_0" shape = [192, 768] dtype = "float32" min_val = float("-0.45152") @@ -3091,6 +3372,7 @@ class Program_weight_tensor_parameter_280: class Program_weight_tensor_parameter_281: name = "parameter_281" + original_name = "layer_norm_4.b_0" shape = [192] dtype = "float32" min_val = float("-1.14375") @@ -3102,6 +3384,7 @@ class Program_weight_tensor_parameter_281: class Program_weight_tensor_parameter_282: name = "parameter_282" + original_name = "layer_norm_4.w_0" shape = [192] dtype = "float32" min_val = float("-0.001608") @@ -3113,6 +3396,7 @@ class Program_weight_tensor_parameter_282: class Program_weight_tensor_parameter_283: name = "parameter_283" + original_name = "linear_5.b_0" shape = [192] dtype = "float32" min_val = float("-1.44116") @@ -3124,6 +3408,7 @@ class Program_weight_tensor_parameter_283: class Program_weight_tensor_parameter_284: name = "parameter_284" + original_name = "linear_5.w_0" shape = [192, 192] dtype = "float32" min_val = float("-0.402529") @@ -3135,6 +3420,7 @@ class Program_weight_tensor_parameter_284: class Program_weight_tensor_parameter_285: name = "parameter_285" + original_name = "linear_4.b_0" shape = [576] dtype = "float32" min_val = float("-2.9529") @@ -3146,6 +3432,7 @@ class Program_weight_tensor_parameter_285: class Program_weight_tensor_parameter_286: name = "parameter_286" + original_name = "linear_4.w_0" shape = [192, 576] dtype = "float32" min_val = float("-0.420929") @@ -3157,6 +3444,7 @@ class Program_weight_tensor_parameter_286: class Program_weight_tensor_parameter_287: name = "parameter_287" + original_name = "layer_norm_3.b_0" shape = [192] dtype = "float32" min_val = float("-0.873584") @@ -3168,6 +3456,7 @@ class Program_weight_tensor_parameter_287: class Program_weight_tensor_parameter_288: name = "parameter_288" + original_name = "layer_norm_3.w_0" shape = [192] dtype = "float32" min_val = float("-0.00601319") @@ -3179,6 +3468,7 @@ class Program_weight_tensor_parameter_288: class Program_weight_tensor_parameter_289: name = "parameter_289" + original_name = "linear_3.b_0" shape = [192] dtype = "float32" min_val = float("-6.04454") @@ -3190,6 +3480,7 @@ class Program_weight_tensor_parameter_289: class Program_weight_tensor_parameter_290: name = "parameter_290" + original_name = "linear_3.w_0" shape = [768, 192] dtype = "float32" min_val = float("-0.364098") @@ -3201,6 +3492,7 @@ class Program_weight_tensor_parameter_290: class Program_weight_tensor_parameter_291: name = "parameter_291" + original_name = "linear_2.b_0" shape = [768] dtype = "float32" min_val = float("-3.07776") @@ -3212,6 +3504,7 @@ class Program_weight_tensor_parameter_291: class Program_weight_tensor_parameter_292: name = "parameter_292" + original_name = "linear_2.w_0" shape = [192, 768] dtype = "float32" min_val = float("-0.560387") @@ -3223,6 +3516,7 @@ class Program_weight_tensor_parameter_292: class Program_weight_tensor_parameter_293: name = "parameter_293" + original_name = "layer_norm_2.b_0" shape = [192] dtype = "float32" min_val = float("-1.22102") @@ -3234,6 +3528,7 @@ class Program_weight_tensor_parameter_293: class Program_weight_tensor_parameter_294: name = "parameter_294" + original_name = "layer_norm_2.w_0" shape = [192] dtype = "float32" min_val = float("-0.0034553") @@ -3245,6 +3540,7 @@ class Program_weight_tensor_parameter_294: class Program_weight_tensor_parameter_295: name = "parameter_295" + original_name = "linear_1.b_0" shape = [192] dtype = "float32" min_val = float("-2.29587") @@ -3256,6 +3552,7 @@ class Program_weight_tensor_parameter_295: class Program_weight_tensor_parameter_296: name = "parameter_296" + original_name = "linear_1.w_0" shape = [192, 192] dtype = "float32" min_val = float("-0.40913") @@ -3267,6 +3564,7 @@ class Program_weight_tensor_parameter_296: class Program_weight_tensor_parameter_297: name = "parameter_297" + original_name = "linear_0.b_0" shape = [576] dtype = "float32" min_val = float("-3.96597") @@ -3278,6 +3576,7 @@ class Program_weight_tensor_parameter_297: class Program_weight_tensor_parameter_298: name = "parameter_298" + original_name = "linear_0.w_0" shape = [192, 576] dtype = "float32" min_val = float("-0.482814") @@ -3289,6 +3588,7 @@ class Program_weight_tensor_parameter_298: class Program_weight_tensor_parameter_299: name = "parameter_299" + original_name = "layer_norm_1.b_0" shape = [192] dtype = "float32" min_val = float("-0.977463") @@ -3300,6 +3600,7 @@ class Program_weight_tensor_parameter_299: class Program_weight_tensor_parameter_300: name = "parameter_300" + original_name = "layer_norm_1.w_0" shape = [192] dtype = "float32" min_val = float("-0.0141446") @@ -3311,6 +3612,7 @@ class Program_weight_tensor_parameter_300: class Program_weight_tensor_parameter_301: name = "parameter_301" + original_name = "layer_norm_0.b_0" shape = [192] dtype = "float32" min_val = float("-2.32457") @@ -3322,6 +3624,7 @@ class Program_weight_tensor_parameter_301: class Program_weight_tensor_parameter_302: name = "parameter_302" + original_name = "layer_norm_0.w_0" shape = [192] dtype = "float32" min_val = float("-0.0164242") @@ -3333,6 +3636,7 @@ class Program_weight_tensor_parameter_302: class Program_weight_tensor_parameter_303: name = "parameter_303" + original_name = "conv2d_0.b_0" shape = [192] dtype = "float32" min_val = float("-0.870857") @@ -3344,6 +3648,7 @@ class Program_weight_tensor_parameter_303: class Program_weight_tensor_parameter_304: name = "parameter_304" + original_name = "conv2d_0.w_0" shape = [192, 3, 4, 4] dtype = "float32" min_val = float("-0.345763") diff --git a/paddle_samples/PaddleX/SwinTransformer_small_patch4_window7_224/subgraph_0/input_meta.py b/paddle_samples/PaddleX/SwinTransformer_small_patch4_window7_224/subgraph_0/input_meta.py index f55cbc01b..f442dea94 100644 --- a/paddle_samples/PaddleX/SwinTransformer_small_patch4_window7_224/subgraph_0/input_meta.py +++ b/paddle_samples/PaddleX/SwinTransformer_small_patch4_window7_224/subgraph_0/input_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_data_0: name = "data_0" + original_name = "param_0" shape = [169, 3] dtype = "float32" min_val = float("-10.8562") @@ -11,6 +12,7 @@ class Program_weight_tensor_data_0: class Program_weight_tensor_data_1: name = "data_1" + original_name = "param_1" shape = [169, 3] dtype = "float32" min_val = float("-10.721") @@ -22,6 +24,7 @@ class Program_weight_tensor_data_1: class Program_weight_tensor_data_2: name = "data_2" + original_name = "param_10" shape = [169, 12] dtype = "float32" min_val = float("-10.131") @@ -33,6 +36,7 @@ class Program_weight_tensor_data_2: class Program_weight_tensor_data_3: name = "data_3" + original_name = "param_11" shape = [169, 12] dtype = "float32" min_val = float("-5.80667") @@ -44,6 +48,7 @@ class Program_weight_tensor_data_3: class Program_weight_tensor_data_4: name = "data_4" + original_name = "param_12" shape = [169, 12] dtype = "float32" min_val = float("-10.4214") @@ -55,6 +60,7 @@ class Program_weight_tensor_data_4: class Program_weight_tensor_data_5: name = "data_5" + original_name = "param_13" shape = [169, 12] dtype = "float32" min_val = float("-8.23036") @@ -66,6 +72,7 @@ class Program_weight_tensor_data_5: class Program_weight_tensor_data_6: name = "data_6" + original_name = "param_14" shape = [169, 12] dtype = "float32" min_val = float("-9.32025") @@ -77,6 +84,7 @@ class Program_weight_tensor_data_6: class Program_weight_tensor_data_7: name = "data_7" + original_name = "param_15" shape = [169, 12] dtype = "float32" min_val = float("-4.72291") @@ -88,6 +96,7 @@ class Program_weight_tensor_data_7: class Program_weight_tensor_data_8: name = "data_8" + original_name = "param_16" shape = [169, 12] dtype = "float32" min_val = float("-9.57262") @@ -99,6 +108,7 @@ class Program_weight_tensor_data_8: class Program_weight_tensor_data_9: name = "data_9" + original_name = "param_17" shape = [169, 12] dtype = "float32" min_val = float("-5.86546") @@ -110,6 +120,7 @@ class Program_weight_tensor_data_9: class Program_weight_tensor_data_10: name = "data_10" + original_name = "param_18" shape = [169, 12] dtype = "float32" min_val = float("-11.2008") @@ -121,6 +132,7 @@ class Program_weight_tensor_data_10: class Program_weight_tensor_data_11: name = "data_11" + original_name = "param_19" shape = [169, 12] dtype = "float32" min_val = float("-11.4129") @@ -132,6 +144,7 @@ class Program_weight_tensor_data_11: class Program_weight_tensor_data_12: name = "data_12" + original_name = "param_2" shape = [169, 6] dtype = "float32" min_val = float("-4.59034") @@ -143,6 +156,7 @@ class Program_weight_tensor_data_12: class Program_weight_tensor_data_13: name = "data_13" + original_name = "param_20" shape = [169, 12] dtype = "float32" min_val = float("-12.6651") @@ -154,6 +168,7 @@ class Program_weight_tensor_data_13: class Program_weight_tensor_data_14: name = "data_14" + original_name = "param_21" shape = [169, 12] dtype = "float32" min_val = float("-6.02371") @@ -165,6 +180,7 @@ class Program_weight_tensor_data_14: class Program_weight_tensor_data_15: name = "data_15" + original_name = "param_22" shape = [169, 24] dtype = "float32" min_val = float("-16.8576") @@ -176,6 +192,7 @@ class Program_weight_tensor_data_15: class Program_weight_tensor_data_16: name = "data_16" + original_name = "param_23" shape = [169, 24] dtype = "float32" min_val = float("-18.213") @@ -187,6 +204,7 @@ class Program_weight_tensor_data_16: class Program_weight_tensor_data_17: name = "data_17" + original_name = "param_3" shape = [169, 6] dtype = "float32" min_val = float("-4.7555") @@ -198,6 +216,7 @@ class Program_weight_tensor_data_17: class Program_weight_tensor_data_18: name = "data_18" + original_name = "param_4" shape = [169, 12] dtype = "float32" min_val = float("-8.4496") @@ -209,6 +228,7 @@ class Program_weight_tensor_data_18: class Program_weight_tensor_data_19: name = "data_19" + original_name = "param_5" shape = [169, 12] dtype = "float32" min_val = float("-11.1686") @@ -220,6 +240,7 @@ class Program_weight_tensor_data_19: class Program_weight_tensor_data_20: name = "data_20" + original_name = "param_6" shape = [169, 12] dtype = "float32" min_val = float("-9.9087") @@ -231,6 +252,7 @@ class Program_weight_tensor_data_20: class Program_weight_tensor_data_21: name = "data_21" + original_name = "param_7" shape = [169, 12] dtype = "float32" min_val = float("-8.18147") @@ -242,6 +264,7 @@ class Program_weight_tensor_data_21: class Program_weight_tensor_data_22: name = "data_22" + original_name = "param_8" shape = [169, 12] dtype = "float32" min_val = float("-5.04052") @@ -253,6 +276,7 @@ class Program_weight_tensor_data_22: class Program_weight_tensor_data_23: name = "data_23" + original_name = "param_9" shape = [169, 12] dtype = "float32" min_val = float("-6.80778") @@ -264,6 +288,7 @@ class Program_weight_tensor_data_23: class Program_weight_tensor_data_24: name = "data_24" + original_name = "var_0" shape = [64, 3, 224, 224] dtype = "float32" min_val = float("-3.87123") @@ -275,6 +300,7 @@ class Program_weight_tensor_data_24: class Program_weight_tensor_data_25: name = "data_25" + original_name = "var_1021" shape = [49, 49] dtype = "int64" min_val = 0 @@ -284,6 +310,7 @@ class Program_weight_tensor_data_25: class Program_weight_tensor_data_26: name = "data_26" + original_name = "var_1101" shape = [49, 49] dtype = "int64" min_val = 0 @@ -293,6 +320,7 @@ class Program_weight_tensor_data_26: class Program_weight_tensor_data_27: name = "data_27" + original_name = "var_1163" shape = [49, 49] dtype = "int64" min_val = 0 @@ -302,6 +330,7 @@ class Program_weight_tensor_data_27: class Program_weight_tensor_data_28: name = "data_28" + original_name = "var_1243" shape = [49, 49] dtype = "int64" min_val = 0 @@ -311,6 +340,7 @@ class Program_weight_tensor_data_28: class Program_weight_tensor_data_29: name = "data_29" + original_name = "var_1305" shape = [49, 49] dtype = "int64" min_val = 0 @@ -320,6 +350,7 @@ class Program_weight_tensor_data_29: class Program_weight_tensor_data_30: name = "data_30" + original_name = "var_1385" shape = [49, 49] dtype = "int64" min_val = 0 @@ -329,6 +360,7 @@ class Program_weight_tensor_data_30: class Program_weight_tensor_data_31: name = "data_31" + original_name = "var_1447" shape = [49, 49] dtype = "int64" min_val = 0 @@ -338,6 +370,7 @@ class Program_weight_tensor_data_31: class Program_weight_tensor_data_32: name = "data_32" + original_name = "var_1527" shape = [49, 49] dtype = "int64" min_val = 0 @@ -347,6 +380,7 @@ class Program_weight_tensor_data_32: class Program_weight_tensor_data_33: name = "data_33" + original_name = "var_159" shape = [49, 49] dtype = "int64" min_val = 0 @@ -356,6 +390,7 @@ class Program_weight_tensor_data_33: class Program_weight_tensor_data_34: name = "data_34" + original_name = "var_1599" shape = [49, 49] dtype = "int64" min_val = 0 @@ -365,6 +400,7 @@ class Program_weight_tensor_data_34: class Program_weight_tensor_data_35: name = "data_35" + original_name = "var_1679" shape = [49, 49] dtype = "int64" min_val = 0 @@ -374,6 +410,7 @@ class Program_weight_tensor_data_35: class Program_weight_tensor_data_36: name = "data_36" + original_name = "var_21" shape = [49, 49] dtype = "int64" min_val = 0 @@ -383,6 +420,7 @@ class Program_weight_tensor_data_36: class Program_weight_tensor_data_37: name = "data_37" + original_name = "var_239" shape = [49, 49] dtype = "int64" min_val = 0 @@ -392,6 +430,7 @@ class Program_weight_tensor_data_37: class Program_weight_tensor_data_38: name = "data_38" + original_name = "var_311" shape = [49, 49] dtype = "int64" min_val = 0 @@ -401,6 +440,7 @@ class Program_weight_tensor_data_38: class Program_weight_tensor_data_39: name = "data_39" + original_name = "var_391" shape = [49, 49] dtype = "int64" min_val = 0 @@ -410,6 +450,7 @@ class Program_weight_tensor_data_39: class Program_weight_tensor_data_40: name = "data_40" + original_name = "var_453" shape = [49, 49] dtype = "int64" min_val = 0 @@ -419,6 +460,7 @@ class Program_weight_tensor_data_40: class Program_weight_tensor_data_41: name = "data_41" + original_name = "var_533" shape = [49, 49] dtype = "int64" min_val = 0 @@ -428,6 +470,7 @@ class Program_weight_tensor_data_41: class Program_weight_tensor_data_42: name = "data_42" + original_name = "var_595" shape = [49, 49] dtype = "int64" min_val = 0 @@ -437,6 +480,7 @@ class Program_weight_tensor_data_42: class Program_weight_tensor_data_43: name = "data_43" + original_name = "var_675" shape = [49, 49] dtype = "int64" min_val = 0 @@ -446,6 +490,7 @@ class Program_weight_tensor_data_43: class Program_weight_tensor_data_44: name = "data_44" + original_name = "var_737" shape = [49, 49] dtype = "int64" min_val = 0 @@ -455,6 +500,7 @@ class Program_weight_tensor_data_44: class Program_weight_tensor_data_45: name = "data_45" + original_name = "var_817" shape = [49, 49] dtype = "int64" min_val = 0 @@ -464,6 +510,7 @@ class Program_weight_tensor_data_45: class Program_weight_tensor_data_46: name = "data_46" + original_name = "var_87" shape = [49, 49] dtype = "int64" min_val = 0 @@ -473,6 +520,7 @@ class Program_weight_tensor_data_46: class Program_weight_tensor_data_47: name = "data_47" + original_name = "var_879" shape = [49, 49] dtype = "int64" min_val = 0 @@ -482,6 +530,7 @@ class Program_weight_tensor_data_47: class Program_weight_tensor_data_48: name = "data_48" + original_name = "var_959" shape = [49, 49] dtype = "int64" min_val = 0 diff --git a/paddle_samples/PaddleX/SwinTransformer_small_patch4_window7_224/subgraph_0/weight_meta.py b/paddle_samples/PaddleX/SwinTransformer_small_patch4_window7_224/subgraph_0/weight_meta.py index db6e60ea8..4bedb5cf2 100644 --- a/paddle_samples/PaddleX/SwinTransformer_small_patch4_window7_224/subgraph_0/weight_meta.py +++ b/paddle_samples/PaddleX/SwinTransformer_small_patch4_window7_224/subgraph_0/weight_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_parameter_0: name = "parameter_0" + original_name = "linear_99.b_0" shape = [102] dtype = "float32" data = None @@ -7,6 +8,7 @@ class Program_weight_tensor_parameter_0: class Program_weight_tensor_parameter_1: name = "parameter_1" + original_name = "linear_99.w_0" shape = [768, 102] dtype = "float32" min_val = float("-0.0887979") @@ -18,6 +20,7 @@ class Program_weight_tensor_parameter_1: class Program_weight_tensor_parameter_2: name = "parameter_2" + original_name = "layer_norm_52.b_0" shape = [768] dtype = "float32" min_val = float("-0.590235") @@ -29,6 +32,7 @@ class Program_weight_tensor_parameter_2: class Program_weight_tensor_parameter_3: name = "parameter_3" + original_name = "layer_norm_52.w_0" shape = [768] dtype = "float32" min_val = float("-0.000240898") @@ -40,6 +44,7 @@ class Program_weight_tensor_parameter_3: class Program_weight_tensor_parameter_4: name = "parameter_4" + original_name = "linear_98.b_0" shape = [768] dtype = "float32" min_val = float("-10.7023") @@ -51,6 +56,7 @@ class Program_weight_tensor_parameter_4: class Program_weight_tensor_parameter_5: name = "parameter_5" + original_name = "linear_98.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-1.74835") @@ -62,6 +68,7 @@ class Program_weight_tensor_parameter_5: class Program_weight_tensor_parameter_6: name = "parameter_6" + original_name = "linear_97.b_0" shape = [3072] dtype = "float32" min_val = float("-2.04066") @@ -73,6 +80,7 @@ class Program_weight_tensor_parameter_6: class Program_weight_tensor_parameter_7: name = "parameter_7" + original_name = "linear_97.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.719578") @@ -84,6 +92,7 @@ class Program_weight_tensor_parameter_7: class Program_weight_tensor_parameter_8: name = "parameter_8" + original_name = "layer_norm_51.b_0" shape = [768] dtype = "float32" min_val = float("-1.48326") @@ -95,6 +104,7 @@ class Program_weight_tensor_parameter_8: class Program_weight_tensor_parameter_9: name = "parameter_9" + original_name = "layer_norm_51.w_0" shape = [768] dtype = "float32" min_val = float("-0.0104547") @@ -106,6 +116,7 @@ class Program_weight_tensor_parameter_9: class Program_weight_tensor_parameter_10: name = "parameter_10" + original_name = "linear_96.b_0" shape = [768] dtype = "float32" min_val = float("-10.5599") @@ -117,6 +128,7 @@ class Program_weight_tensor_parameter_10: class Program_weight_tensor_parameter_11: name = "parameter_11" + original_name = "linear_96.w_0" shape = [768, 768] dtype = "float32" min_val = float("-1.15759") @@ -128,6 +140,7 @@ class Program_weight_tensor_parameter_11: class Program_weight_tensor_parameter_12: name = "parameter_12" + original_name = "linear_95.b_0" shape = [2304] dtype = "float32" min_val = float("-3.08885") @@ -139,6 +152,7 @@ class Program_weight_tensor_parameter_12: class Program_weight_tensor_parameter_13: name = "parameter_13" + original_name = "linear_95.w_0" shape = [768, 2304] dtype = "float32" min_val = float("-0.496372") @@ -150,6 +164,7 @@ class Program_weight_tensor_parameter_13: class Program_weight_tensor_parameter_14: name = "parameter_14" + original_name = "layer_norm_50.b_0" shape = [768] dtype = "float32" min_val = float("-0.610759") @@ -161,6 +176,7 @@ class Program_weight_tensor_parameter_14: class Program_weight_tensor_parameter_15: name = "parameter_15" + original_name = "layer_norm_50.w_0" shape = [768] dtype = "float32" min_val = float("0.116551") @@ -172,6 +188,7 @@ class Program_weight_tensor_parameter_15: class Program_weight_tensor_parameter_16: name = "parameter_16" + original_name = "linear_94.b_0" shape = [768] dtype = "float32" min_val = float("-8.89402") @@ -183,6 +200,7 @@ class Program_weight_tensor_parameter_16: class Program_weight_tensor_parameter_17: name = "parameter_17" + original_name = "linear_94.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-1.61058") @@ -194,6 +212,7 @@ class Program_weight_tensor_parameter_17: class Program_weight_tensor_parameter_18: name = "parameter_18" + original_name = "linear_93.b_0" shape = [3072] dtype = "float32" min_val = float("-1.22006") @@ -205,6 +224,7 @@ class Program_weight_tensor_parameter_18: class Program_weight_tensor_parameter_19: name = "parameter_19" + original_name = "linear_93.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.430694") @@ -216,6 +236,7 @@ class Program_weight_tensor_parameter_19: class Program_weight_tensor_parameter_20: name = "parameter_20" + original_name = "layer_norm_49.b_0" shape = [768] dtype = "float32" min_val = float("-1.90376") @@ -227,6 +248,7 @@ class Program_weight_tensor_parameter_20: class Program_weight_tensor_parameter_21: name = "parameter_21" + original_name = "layer_norm_49.w_0" shape = [768] dtype = "float32" min_val = float("0.157451") @@ -238,6 +260,7 @@ class Program_weight_tensor_parameter_21: class Program_weight_tensor_parameter_22: name = "parameter_22" + original_name = "linear_92.b_0" shape = [768] dtype = "float32" min_val = float("-11.9637") @@ -249,6 +272,7 @@ class Program_weight_tensor_parameter_22: class Program_weight_tensor_parameter_23: name = "parameter_23" + original_name = "linear_92.w_0" shape = [768, 768] dtype = "float32" min_val = float("-1.00479") @@ -260,6 +284,7 @@ class Program_weight_tensor_parameter_23: class Program_weight_tensor_parameter_24: name = "parameter_24" + original_name = "linear_91.b_0" shape = [2304] dtype = "float32" min_val = float("-2.55967") @@ -271,6 +296,7 @@ class Program_weight_tensor_parameter_24: class Program_weight_tensor_parameter_25: name = "parameter_25" + original_name = "linear_91.w_0" shape = [768, 2304] dtype = "float32" min_val = float("-0.378148") @@ -282,6 +308,7 @@ class Program_weight_tensor_parameter_25: class Program_weight_tensor_parameter_26: name = "parameter_26" + original_name = "layer_norm_48.b_0" shape = [768] dtype = "float32" min_val = float("-0.483067") @@ -293,6 +320,7 @@ class Program_weight_tensor_parameter_26: class Program_weight_tensor_parameter_27: name = "parameter_27" + original_name = "layer_norm_48.w_0" shape = [768] dtype = "float32" min_val = float("0.00175781") @@ -304,6 +332,7 @@ class Program_weight_tensor_parameter_27: class Program_weight_tensor_parameter_28: name = "parameter_28" + original_name = "linear_90.w_0" shape = [1536, 768] dtype = "float32" min_val = float("-0.807711") @@ -315,6 +344,7 @@ class Program_weight_tensor_parameter_28: class Program_weight_tensor_parameter_29: name = "parameter_29" + original_name = "layer_norm_47.b_0" shape = [1536] dtype = "float32" min_val = float("-0.856513") @@ -326,6 +356,7 @@ class Program_weight_tensor_parameter_29: class Program_weight_tensor_parameter_30: name = "parameter_30" + original_name = "layer_norm_47.w_0" shape = [1536] dtype = "float32" min_val = float("0.113077") @@ -337,6 +368,7 @@ class Program_weight_tensor_parameter_30: class Program_weight_tensor_parameter_31: name = "parameter_31" + original_name = "linear_89.b_0" shape = [384] dtype = "float32" min_val = float("-1.56504") @@ -348,6 +380,7 @@ class Program_weight_tensor_parameter_31: class Program_weight_tensor_parameter_32: name = "parameter_32" + original_name = "linear_89.w_0" shape = [1536, 384] dtype = "float32" min_val = float("-2.02084") @@ -359,6 +392,7 @@ class Program_weight_tensor_parameter_32: class Program_weight_tensor_parameter_33: name = "parameter_33" + original_name = "linear_88.b_0" shape = [1536] dtype = "float32" min_val = float("-1.12341") @@ -370,6 +404,7 @@ class Program_weight_tensor_parameter_33: class Program_weight_tensor_parameter_34: name = "parameter_34" + original_name = "linear_88.w_0" shape = [384, 1536] dtype = "float32" min_val = float("-0.589133") @@ -381,6 +416,7 @@ class Program_weight_tensor_parameter_34: class Program_weight_tensor_parameter_35: name = "parameter_35" + original_name = "layer_norm_46.b_0" shape = [384] dtype = "float32" min_val = float("-1.5281") @@ -392,6 +428,7 @@ class Program_weight_tensor_parameter_35: class Program_weight_tensor_parameter_36: name = "parameter_36" + original_name = "layer_norm_46.w_0" shape = [384] dtype = "float32" min_val = float("0.676636") @@ -403,6 +440,7 @@ class Program_weight_tensor_parameter_36: class Program_weight_tensor_parameter_37: name = "parameter_37" + original_name = "linear_87.b_0" shape = [384] dtype = "float32" min_val = float("-1.39128") @@ -414,6 +452,7 @@ class Program_weight_tensor_parameter_37: class Program_weight_tensor_parameter_38: name = "parameter_38" + original_name = "linear_87.w_0" shape = [384, 384] dtype = "float32" min_val = float("-1.18998") @@ -425,6 +464,7 @@ class Program_weight_tensor_parameter_38: class Program_weight_tensor_parameter_39: name = "parameter_39" + original_name = "linear_86.b_0" shape = [1152] dtype = "float32" min_val = float("-2.74119") @@ -436,6 +476,7 @@ class Program_weight_tensor_parameter_39: class Program_weight_tensor_parameter_40: name = "parameter_40" + original_name = "linear_86.w_0" shape = [384, 1152] dtype = "float32" min_val = float("-0.275169") @@ -447,6 +488,7 @@ class Program_weight_tensor_parameter_40: class Program_weight_tensor_parameter_41: name = "parameter_41" + original_name = "layer_norm_45.b_0" shape = [384] dtype = "float32" min_val = float("-1.10933") @@ -458,6 +500,7 @@ class Program_weight_tensor_parameter_41: class Program_weight_tensor_parameter_42: name = "parameter_42" + original_name = "layer_norm_45.w_0" shape = [384] dtype = "float32" min_val = float("0.0934882") @@ -469,6 +512,7 @@ class Program_weight_tensor_parameter_42: class Program_weight_tensor_parameter_43: name = "parameter_43" + original_name = "linear_85.b_0" shape = [384] dtype = "float32" min_val = float("-1.80482") @@ -480,6 +524,7 @@ class Program_weight_tensor_parameter_43: class Program_weight_tensor_parameter_44: name = "parameter_44" + original_name = "linear_85.w_0" shape = [1536, 384] dtype = "float32" min_val = float("-1.22573") @@ -491,6 +536,7 @@ class Program_weight_tensor_parameter_44: class Program_weight_tensor_parameter_45: name = "parameter_45" + original_name = "linear_84.b_0" shape = [1536] dtype = "float32" min_val = float("-1.14126") @@ -502,6 +548,7 @@ class Program_weight_tensor_parameter_45: class Program_weight_tensor_parameter_46: name = "parameter_46" + original_name = "linear_84.w_0" shape = [384, 1536] dtype = "float32" min_val = float("-0.733478") @@ -513,6 +560,7 @@ class Program_weight_tensor_parameter_46: class Program_weight_tensor_parameter_47: name = "parameter_47" + original_name = "layer_norm_44.b_0" shape = [384] dtype = "float32" min_val = float("-1.68308") @@ -524,6 +572,7 @@ class Program_weight_tensor_parameter_47: class Program_weight_tensor_parameter_48: name = "parameter_48" + original_name = "layer_norm_44.w_0" shape = [384] dtype = "float32" min_val = float("0.502989") @@ -535,6 +584,7 @@ class Program_weight_tensor_parameter_48: class Program_weight_tensor_parameter_49: name = "parameter_49" + original_name = "linear_83.b_0" shape = [384] dtype = "float32" min_val = float("-1.4626") @@ -546,6 +596,7 @@ class Program_weight_tensor_parameter_49: class Program_weight_tensor_parameter_50: name = "parameter_50" + original_name = "linear_83.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.533361") @@ -557,6 +608,7 @@ class Program_weight_tensor_parameter_50: class Program_weight_tensor_parameter_51: name = "parameter_51" + original_name = "linear_82.b_0" shape = [1152] dtype = "float32" min_val = float("-3.10019") @@ -568,6 +620,7 @@ class Program_weight_tensor_parameter_51: class Program_weight_tensor_parameter_52: name = "parameter_52" + original_name = "linear_82.w_0" shape = [384, 1152] dtype = "float32" min_val = float("-0.389517") @@ -579,6 +632,7 @@ class Program_weight_tensor_parameter_52: class Program_weight_tensor_parameter_53: name = "parameter_53" + original_name = "layer_norm_43.b_0" shape = [384] dtype = "float32" min_val = float("-1.19679") @@ -590,6 +644,7 @@ class Program_weight_tensor_parameter_53: class Program_weight_tensor_parameter_54: name = "parameter_54" + original_name = "layer_norm_43.w_0" shape = [384] dtype = "float32" min_val = float("0.124975") @@ -601,6 +656,7 @@ class Program_weight_tensor_parameter_54: class Program_weight_tensor_parameter_55: name = "parameter_55" + original_name = "linear_81.b_0" shape = [384] dtype = "float32" min_val = float("-1.32268") @@ -612,6 +668,7 @@ class Program_weight_tensor_parameter_55: class Program_weight_tensor_parameter_56: name = "parameter_56" + original_name = "linear_81.w_0" shape = [1536, 384] dtype = "float32" min_val = float("-1.15946") @@ -623,6 +680,7 @@ class Program_weight_tensor_parameter_56: class Program_weight_tensor_parameter_57: name = "parameter_57" + original_name = "linear_80.b_0" shape = [1536] dtype = "float32" min_val = float("-1.15229") @@ -634,6 +692,7 @@ class Program_weight_tensor_parameter_57: class Program_weight_tensor_parameter_58: name = "parameter_58" + original_name = "linear_80.w_0" shape = [384, 1536] dtype = "float32" min_val = float("-0.35147") @@ -645,6 +704,7 @@ class Program_weight_tensor_parameter_58: class Program_weight_tensor_parameter_59: name = "parameter_59" + original_name = "layer_norm_42.b_0" shape = [384] dtype = "float32" min_val = float("-1.70545") @@ -656,6 +716,7 @@ class Program_weight_tensor_parameter_59: class Program_weight_tensor_parameter_60: name = "parameter_60" + original_name = "layer_norm_42.w_0" shape = [384] dtype = "float32" min_val = float("0.334615") @@ -667,6 +728,7 @@ class Program_weight_tensor_parameter_60: class Program_weight_tensor_parameter_61: name = "parameter_61" + original_name = "linear_79.b_0" shape = [384] dtype = "float32" min_val = float("-1.58625") @@ -678,6 +740,7 @@ class Program_weight_tensor_parameter_61: class Program_weight_tensor_parameter_62: name = "parameter_62" + original_name = "linear_79.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.354384") @@ -689,6 +752,7 @@ class Program_weight_tensor_parameter_62: class Program_weight_tensor_parameter_63: name = "parameter_63" + original_name = "linear_78.b_0" shape = [1152] dtype = "float32" min_val = float("-2.80981") @@ -700,6 +764,7 @@ class Program_weight_tensor_parameter_63: class Program_weight_tensor_parameter_64: name = "parameter_64" + original_name = "linear_78.w_0" shape = [384, 1152] dtype = "float32" min_val = float("-0.264833") @@ -711,6 +776,7 @@ class Program_weight_tensor_parameter_64: class Program_weight_tensor_parameter_65: name = "parameter_65" + original_name = "layer_norm_41.b_0" shape = [384] dtype = "float32" min_val = float("-1.17739") @@ -722,6 +788,7 @@ class Program_weight_tensor_parameter_65: class Program_weight_tensor_parameter_66: name = "parameter_66" + original_name = "layer_norm_41.w_0" shape = [384] dtype = "float32" min_val = float("0.078106") @@ -733,6 +800,7 @@ class Program_weight_tensor_parameter_66: class Program_weight_tensor_parameter_67: name = "parameter_67" + original_name = "linear_77.b_0" shape = [384] dtype = "float32" min_val = float("-1.34313") @@ -744,6 +812,7 @@ class Program_weight_tensor_parameter_67: class Program_weight_tensor_parameter_68: name = "parameter_68" + original_name = "linear_77.w_0" shape = [1536, 384] dtype = "float32" min_val = float("-1.01683") @@ -755,6 +824,7 @@ class Program_weight_tensor_parameter_68: class Program_weight_tensor_parameter_69: name = "parameter_69" + original_name = "linear_76.b_0" shape = [1536] dtype = "float32" min_val = float("-1.24906") @@ -766,6 +836,7 @@ class Program_weight_tensor_parameter_69: class Program_weight_tensor_parameter_70: name = "parameter_70" + original_name = "linear_76.w_0" shape = [384, 1536] dtype = "float32" min_val = float("-0.242821") @@ -777,6 +848,7 @@ class Program_weight_tensor_parameter_70: class Program_weight_tensor_parameter_71: name = "parameter_71" + original_name = "layer_norm_40.b_0" shape = [384] dtype = "float32" min_val = float("-1.41929") @@ -788,6 +860,7 @@ class Program_weight_tensor_parameter_71: class Program_weight_tensor_parameter_72: name = "parameter_72" + original_name = "layer_norm_40.w_0" shape = [384] dtype = "float32" min_val = float("0.362523") @@ -799,6 +872,7 @@ class Program_weight_tensor_parameter_72: class Program_weight_tensor_parameter_73: name = "parameter_73" + original_name = "linear_75.b_0" shape = [384] dtype = "float32" min_val = float("-2.20402") @@ -810,6 +884,7 @@ class Program_weight_tensor_parameter_73: class Program_weight_tensor_parameter_74: name = "parameter_74" + original_name = "linear_75.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.369846") @@ -821,6 +896,7 @@ class Program_weight_tensor_parameter_74: class Program_weight_tensor_parameter_75: name = "parameter_75" + original_name = "linear_74.b_0" shape = [1152] dtype = "float32" min_val = float("-2.83003") @@ -832,6 +908,7 @@ class Program_weight_tensor_parameter_75: class Program_weight_tensor_parameter_76: name = "parameter_76" + original_name = "linear_74.w_0" shape = [384, 1152] dtype = "float32" min_val = float("-0.258563") @@ -843,6 +920,7 @@ class Program_weight_tensor_parameter_76: class Program_weight_tensor_parameter_77: name = "parameter_77" + original_name = "layer_norm_39.b_0" shape = [384] dtype = "float32" min_val = float("-1.42898") @@ -854,6 +932,7 @@ class Program_weight_tensor_parameter_77: class Program_weight_tensor_parameter_78: name = "parameter_78" + original_name = "layer_norm_39.w_0" shape = [384] dtype = "float32" min_val = float("0.14976") @@ -865,6 +944,7 @@ class Program_weight_tensor_parameter_78: class Program_weight_tensor_parameter_79: name = "parameter_79" + original_name = "linear_73.b_0" shape = [384] dtype = "float32" min_val = float("-0.790021") @@ -876,6 +956,7 @@ class Program_weight_tensor_parameter_79: class Program_weight_tensor_parameter_80: name = "parameter_80" + original_name = "linear_73.w_0" shape = [1536, 384] dtype = "float32" min_val = float("-1.60798") @@ -887,6 +968,7 @@ class Program_weight_tensor_parameter_80: class Program_weight_tensor_parameter_81: name = "parameter_81" + original_name = "linear_72.b_0" shape = [1536] dtype = "float32" min_val = float("-1.27135") @@ -898,6 +980,7 @@ class Program_weight_tensor_parameter_81: class Program_weight_tensor_parameter_82: name = "parameter_82" + original_name = "linear_72.w_0" shape = [384, 1536] dtype = "float32" min_val = float("-0.302482") @@ -909,6 +992,7 @@ class Program_weight_tensor_parameter_82: class Program_weight_tensor_parameter_83: name = "parameter_83" + original_name = "layer_norm_38.b_0" shape = [384] dtype = "float32" min_val = float("-1.05285") @@ -920,6 +1004,7 @@ class Program_weight_tensor_parameter_83: class Program_weight_tensor_parameter_84: name = "parameter_84" + original_name = "layer_norm_38.w_0" shape = [384] dtype = "float32" min_val = float("0.462327") @@ -931,6 +1016,7 @@ class Program_weight_tensor_parameter_84: class Program_weight_tensor_parameter_85: name = "parameter_85" + original_name = "linear_71.b_0" shape = [384] dtype = "float32" min_val = float("-2.20092") @@ -942,6 +1028,7 @@ class Program_weight_tensor_parameter_85: class Program_weight_tensor_parameter_86: name = "parameter_86" + original_name = "linear_71.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.252379") @@ -953,6 +1040,7 @@ class Program_weight_tensor_parameter_86: class Program_weight_tensor_parameter_87: name = "parameter_87" + original_name = "linear_70.b_0" shape = [1152] dtype = "float32" min_val = float("-2.46138") @@ -964,6 +1052,7 @@ class Program_weight_tensor_parameter_87: class Program_weight_tensor_parameter_88: name = "parameter_88" + original_name = "linear_70.w_0" shape = [384, 1152] dtype = "float32" min_val = float("-0.267963") @@ -975,6 +1064,7 @@ class Program_weight_tensor_parameter_88: class Program_weight_tensor_parameter_89: name = "parameter_89" + original_name = "layer_norm_37.b_0" shape = [384] dtype = "float32" min_val = float("-1.19319") @@ -986,6 +1076,7 @@ class Program_weight_tensor_parameter_89: class Program_weight_tensor_parameter_90: name = "parameter_90" + original_name = "layer_norm_37.w_0" shape = [384] dtype = "float32" min_val = float("0.168957") @@ -997,6 +1088,7 @@ class Program_weight_tensor_parameter_90: class Program_weight_tensor_parameter_91: name = "parameter_91" + original_name = "linear_69.b_0" shape = [384] dtype = "float32" min_val = float("-0.974824") @@ -1008,6 +1100,7 @@ class Program_weight_tensor_parameter_91: class Program_weight_tensor_parameter_92: name = "parameter_92" + original_name = "linear_69.w_0" shape = [1536, 384] dtype = "float32" min_val = float("-2.68611") @@ -1019,6 +1112,7 @@ class Program_weight_tensor_parameter_92: class Program_weight_tensor_parameter_93: name = "parameter_93" + original_name = "linear_68.b_0" shape = [1536] dtype = "float32" min_val = float("-1.19099") @@ -1030,6 +1124,7 @@ class Program_weight_tensor_parameter_93: class Program_weight_tensor_parameter_94: name = "parameter_94" + original_name = "linear_68.w_0" shape = [384, 1536] dtype = "float32" min_val = float("-0.520867") @@ -1041,6 +1136,7 @@ class Program_weight_tensor_parameter_94: class Program_weight_tensor_parameter_95: name = "parameter_95" + original_name = "layer_norm_36.b_0" shape = [384] dtype = "float32" min_val = float("-1.1552") @@ -1052,6 +1148,7 @@ class Program_weight_tensor_parameter_95: class Program_weight_tensor_parameter_96: name = "parameter_96" + original_name = "layer_norm_36.w_0" shape = [384] dtype = "float32" min_val = float("0.496143") @@ -1063,6 +1160,7 @@ class Program_weight_tensor_parameter_96: class Program_weight_tensor_parameter_97: name = "parameter_97" + original_name = "linear_67.b_0" shape = [384] dtype = "float32" min_val = float("-3.34233") @@ -1074,6 +1172,7 @@ class Program_weight_tensor_parameter_97: class Program_weight_tensor_parameter_98: name = "parameter_98" + original_name = "linear_67.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.313256") @@ -1085,6 +1184,7 @@ class Program_weight_tensor_parameter_98: class Program_weight_tensor_parameter_99: name = "parameter_99" + original_name = "linear_66.b_0" shape = [1152] dtype = "float32" min_val = float("-2.85035") @@ -1096,6 +1196,7 @@ class Program_weight_tensor_parameter_99: class Program_weight_tensor_parameter_100: name = "parameter_100" + original_name = "linear_66.w_0" shape = [384, 1152] dtype = "float32" min_val = float("-0.367999") @@ -1107,6 +1208,7 @@ class Program_weight_tensor_parameter_100: class Program_weight_tensor_parameter_101: name = "parameter_101" + original_name = "layer_norm_35.b_0" shape = [384] dtype = "float32" min_val = float("-1.57556") @@ -1118,6 +1220,7 @@ class Program_weight_tensor_parameter_101: class Program_weight_tensor_parameter_102: name = "parameter_102" + original_name = "layer_norm_35.w_0" shape = [384] dtype = "float32" min_val = float("0.266772") @@ -1129,6 +1232,7 @@ class Program_weight_tensor_parameter_102: class Program_weight_tensor_parameter_103: name = "parameter_103" + original_name = "linear_65.b_0" shape = [384] dtype = "float32" min_val = float("-0.862904") @@ -1140,6 +1244,7 @@ class Program_weight_tensor_parameter_103: class Program_weight_tensor_parameter_104: name = "parameter_104" + original_name = "linear_65.w_0" shape = [1536, 384] dtype = "float32" min_val = float("-2.37199") @@ -1151,6 +1256,7 @@ class Program_weight_tensor_parameter_104: class Program_weight_tensor_parameter_105: name = "parameter_105" + original_name = "linear_64.b_0" shape = [1536] dtype = "float32" min_val = float("-1.25322") @@ -1162,6 +1268,7 @@ class Program_weight_tensor_parameter_105: class Program_weight_tensor_parameter_106: name = "parameter_106" + original_name = "linear_64.w_0" shape = [384, 1536] dtype = "float32" min_val = float("-0.460864") @@ -1173,6 +1280,7 @@ class Program_weight_tensor_parameter_106: class Program_weight_tensor_parameter_107: name = "parameter_107" + original_name = "layer_norm_34.b_0" shape = [384] dtype = "float32" min_val = float("-1.24607") @@ -1184,6 +1292,7 @@ class Program_weight_tensor_parameter_107: class Program_weight_tensor_parameter_108: name = "parameter_108" + original_name = "layer_norm_34.w_0" shape = [384] dtype = "float32" min_val = float("0.524774") @@ -1195,6 +1304,7 @@ class Program_weight_tensor_parameter_108: class Program_weight_tensor_parameter_109: name = "parameter_109" + original_name = "linear_63.b_0" shape = [384] dtype = "float32" min_val = float("-2.26615") @@ -1206,6 +1316,7 @@ class Program_weight_tensor_parameter_109: class Program_weight_tensor_parameter_110: name = "parameter_110" + original_name = "linear_63.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.240297") @@ -1217,6 +1328,7 @@ class Program_weight_tensor_parameter_110: class Program_weight_tensor_parameter_111: name = "parameter_111" + original_name = "linear_62.b_0" shape = [1152] dtype = "float32" min_val = float("-2.12624") @@ -1228,6 +1340,7 @@ class Program_weight_tensor_parameter_111: class Program_weight_tensor_parameter_112: name = "parameter_112" + original_name = "linear_62.w_0" shape = [384, 1152] dtype = "float32" min_val = float("-0.306684") @@ -1239,6 +1352,7 @@ class Program_weight_tensor_parameter_112: class Program_weight_tensor_parameter_113: name = "parameter_113" + original_name = "layer_norm_33.b_0" shape = [384] dtype = "float32" min_val = float("-1.26648") @@ -1250,6 +1364,7 @@ class Program_weight_tensor_parameter_113: class Program_weight_tensor_parameter_114: name = "parameter_114" + original_name = "layer_norm_33.w_0" shape = [384] dtype = "float32" min_val = float("0.262089") @@ -1261,6 +1376,7 @@ class Program_weight_tensor_parameter_114: class Program_weight_tensor_parameter_115: name = "parameter_115" + original_name = "linear_61.b_0" shape = [384] dtype = "float32" min_val = float("-0.729608") @@ -1272,6 +1388,7 @@ class Program_weight_tensor_parameter_115: class Program_weight_tensor_parameter_116: name = "parameter_116" + original_name = "linear_61.w_0" shape = [1536, 384] dtype = "float32" min_val = float("-2.15775") @@ -1283,6 +1400,7 @@ class Program_weight_tensor_parameter_116: class Program_weight_tensor_parameter_117: name = "parameter_117" + original_name = "linear_60.b_0" shape = [1536] dtype = "float32" min_val = float("-1.28577") @@ -1294,6 +1412,7 @@ class Program_weight_tensor_parameter_117: class Program_weight_tensor_parameter_118: name = "parameter_118" + original_name = "linear_60.w_0" shape = [384, 1536] dtype = "float32" min_val = float("-0.442118") @@ -1305,6 +1424,7 @@ class Program_weight_tensor_parameter_118: class Program_weight_tensor_parameter_119: name = "parameter_119" + original_name = "layer_norm_32.b_0" shape = [384] dtype = "float32" min_val = float("-1.15263") @@ -1316,6 +1436,7 @@ class Program_weight_tensor_parameter_119: class Program_weight_tensor_parameter_120: name = "parameter_120" + original_name = "layer_norm_32.w_0" shape = [384] dtype = "float32" min_val = float("0.531976") @@ -1327,6 +1448,7 @@ class Program_weight_tensor_parameter_120: class Program_weight_tensor_parameter_121: name = "parameter_121" + original_name = "linear_59.b_0" shape = [384] dtype = "float32" min_val = float("-3.36602") @@ -1338,6 +1460,7 @@ class Program_weight_tensor_parameter_121: class Program_weight_tensor_parameter_122: name = "parameter_122" + original_name = "linear_59.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.343028") @@ -1349,6 +1472,7 @@ class Program_weight_tensor_parameter_122: class Program_weight_tensor_parameter_123: name = "parameter_123" + original_name = "linear_58.b_0" shape = [1152] dtype = "float32" min_val = float("-2.35094") @@ -1360,6 +1484,7 @@ class Program_weight_tensor_parameter_123: class Program_weight_tensor_parameter_124: name = "parameter_124" + original_name = "linear_58.w_0" shape = [384, 1152] dtype = "float32" min_val = float("-0.323511") @@ -1371,6 +1496,7 @@ class Program_weight_tensor_parameter_124: class Program_weight_tensor_parameter_125: name = "parameter_125" + original_name = "layer_norm_31.b_0" shape = [384] dtype = "float32" min_val = float("-1.52181") @@ -1382,6 +1508,7 @@ class Program_weight_tensor_parameter_125: class Program_weight_tensor_parameter_126: name = "parameter_126" + original_name = "layer_norm_31.w_0" shape = [384] dtype = "float32" min_val = float("0.348682") @@ -1393,6 +1520,7 @@ class Program_weight_tensor_parameter_126: class Program_weight_tensor_parameter_127: name = "parameter_127" + original_name = "linear_57.b_0" shape = [384] dtype = "float32" min_val = float("-0.416716") @@ -1404,6 +1532,7 @@ class Program_weight_tensor_parameter_127: class Program_weight_tensor_parameter_128: name = "parameter_128" + original_name = "linear_57.w_0" shape = [1536, 384] dtype = "float32" min_val = float("-1.86654") @@ -1415,6 +1544,7 @@ class Program_weight_tensor_parameter_128: class Program_weight_tensor_parameter_129: name = "parameter_129" + original_name = "linear_56.b_0" shape = [1536] dtype = "float32" min_val = float("-1.29436") @@ -1426,6 +1556,7 @@ class Program_weight_tensor_parameter_129: class Program_weight_tensor_parameter_130: name = "parameter_130" + original_name = "linear_56.w_0" shape = [384, 1536] dtype = "float32" min_val = float("-0.29442") @@ -1437,6 +1568,7 @@ class Program_weight_tensor_parameter_130: class Program_weight_tensor_parameter_131: name = "parameter_131" + original_name = "layer_norm_30.b_0" shape = [384] dtype = "float32" min_val = float("-1.12232") @@ -1448,6 +1580,7 @@ class Program_weight_tensor_parameter_131: class Program_weight_tensor_parameter_132: name = "parameter_132" + original_name = "layer_norm_30.w_0" shape = [384] dtype = "float32" min_val = float("0.505876") @@ -1459,6 +1592,7 @@ class Program_weight_tensor_parameter_132: class Program_weight_tensor_parameter_133: name = "parameter_133" + original_name = "linear_55.b_0" shape = [384] dtype = "float32" min_val = float("-1.99321") @@ -1470,6 +1604,7 @@ class Program_weight_tensor_parameter_133: class Program_weight_tensor_parameter_134: name = "parameter_134" + original_name = "linear_55.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.287255") @@ -1481,6 +1616,7 @@ class Program_weight_tensor_parameter_134: class Program_weight_tensor_parameter_135: name = "parameter_135" + original_name = "linear_54.b_0" shape = [1152] dtype = "float32" min_val = float("-2.07283") @@ -1492,6 +1628,7 @@ class Program_weight_tensor_parameter_135: class Program_weight_tensor_parameter_136: name = "parameter_136" + original_name = "linear_54.w_0" shape = [384, 1152] dtype = "float32" min_val = float("-0.352515") @@ -1503,6 +1640,7 @@ class Program_weight_tensor_parameter_136: class Program_weight_tensor_parameter_137: name = "parameter_137" + original_name = "layer_norm_29.b_0" shape = [384] dtype = "float32" min_val = float("-1.17167") @@ -1514,6 +1652,7 @@ class Program_weight_tensor_parameter_137: class Program_weight_tensor_parameter_138: name = "parameter_138" + original_name = "layer_norm_29.w_0" shape = [384] dtype = "float32" min_val = float("0.300293") @@ -1525,6 +1664,7 @@ class Program_weight_tensor_parameter_138: class Program_weight_tensor_parameter_139: name = "parameter_139" + original_name = "linear_53.b_0" shape = [384] dtype = "float32" min_val = float("-0.521446") @@ -1536,6 +1676,7 @@ class Program_weight_tensor_parameter_139: class Program_weight_tensor_parameter_140: name = "parameter_140" + original_name = "linear_53.w_0" shape = [1536, 384] dtype = "float32" min_val = float("-1.75861") @@ -1547,6 +1688,7 @@ class Program_weight_tensor_parameter_140: class Program_weight_tensor_parameter_141: name = "parameter_141" + original_name = "linear_52.b_0" shape = [1536] dtype = "float32" min_val = float("-1.25935") @@ -1558,6 +1700,7 @@ class Program_weight_tensor_parameter_141: class Program_weight_tensor_parameter_142: name = "parameter_142" + original_name = "linear_52.w_0" shape = [384, 1536] dtype = "float32" min_val = float("-0.310243") @@ -1569,6 +1712,7 @@ class Program_weight_tensor_parameter_142: class Program_weight_tensor_parameter_143: name = "parameter_143" + original_name = "layer_norm_28.b_0" shape = [384] dtype = "float32" min_val = float("-0.969815") @@ -1580,6 +1724,7 @@ class Program_weight_tensor_parameter_143: class Program_weight_tensor_parameter_144: name = "parameter_144" + original_name = "layer_norm_28.w_0" shape = [384] dtype = "float32" min_val = float("0.450808") @@ -1591,6 +1736,7 @@ class Program_weight_tensor_parameter_144: class Program_weight_tensor_parameter_145: name = "parameter_145" + original_name = "linear_51.b_0" shape = [384] dtype = "float32" min_val = float("-2.87286") @@ -1602,6 +1748,7 @@ class Program_weight_tensor_parameter_145: class Program_weight_tensor_parameter_146: name = "parameter_146" + original_name = "linear_51.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.266421") @@ -1613,6 +1760,7 @@ class Program_weight_tensor_parameter_146: class Program_weight_tensor_parameter_147: name = "parameter_147" + original_name = "linear_50.b_0" shape = [1152] dtype = "float32" min_val = float("-2.61134") @@ -1624,6 +1772,7 @@ class Program_weight_tensor_parameter_147: class Program_weight_tensor_parameter_148: name = "parameter_148" + original_name = "linear_50.w_0" shape = [384, 1152] dtype = "float32" min_val = float("-0.40764") @@ -1635,6 +1784,7 @@ class Program_weight_tensor_parameter_148: class Program_weight_tensor_parameter_149: name = "parameter_149" + original_name = "layer_norm_27.b_0" shape = [384] dtype = "float32" min_val = float("-1.42045") @@ -1646,6 +1796,7 @@ class Program_weight_tensor_parameter_149: class Program_weight_tensor_parameter_150: name = "parameter_150" + original_name = "layer_norm_27.w_0" shape = [384] dtype = "float32" min_val = float("0.313364") @@ -1657,6 +1808,7 @@ class Program_weight_tensor_parameter_150: class Program_weight_tensor_parameter_151: name = "parameter_151" + original_name = "linear_49.b_0" shape = [384] dtype = "float32" min_val = float("-0.493829") @@ -1668,6 +1820,7 @@ class Program_weight_tensor_parameter_151: class Program_weight_tensor_parameter_152: name = "parameter_152" + original_name = "linear_49.w_0" shape = [1536, 384] dtype = "float32" min_val = float("-1.28735") @@ -1679,6 +1832,7 @@ class Program_weight_tensor_parameter_152: class Program_weight_tensor_parameter_153: name = "parameter_153" + original_name = "linear_48.b_0" shape = [1536] dtype = "float32" min_val = float("-1.38945") @@ -1690,6 +1844,7 @@ class Program_weight_tensor_parameter_153: class Program_weight_tensor_parameter_154: name = "parameter_154" + original_name = "linear_48.w_0" shape = [384, 1536] dtype = "float32" min_val = float("-0.303699") @@ -1701,6 +1856,7 @@ class Program_weight_tensor_parameter_154: class Program_weight_tensor_parameter_155: name = "parameter_155" + original_name = "layer_norm_26.b_0" shape = [384] dtype = "float32" min_val = float("-0.860962") @@ -1712,6 +1868,7 @@ class Program_weight_tensor_parameter_155: class Program_weight_tensor_parameter_156: name = "parameter_156" + original_name = "layer_norm_26.w_0" shape = [384] dtype = "float32" min_val = float("0.395832") @@ -1723,6 +1880,7 @@ class Program_weight_tensor_parameter_156: class Program_weight_tensor_parameter_157: name = "parameter_157" + original_name = "linear_47.b_0" shape = [384] dtype = "float32" min_val = float("-2.1996") @@ -1734,6 +1892,7 @@ class Program_weight_tensor_parameter_157: class Program_weight_tensor_parameter_158: name = "parameter_158" + original_name = "linear_47.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.207965") @@ -1745,6 +1904,7 @@ class Program_weight_tensor_parameter_158: class Program_weight_tensor_parameter_159: name = "parameter_159" + original_name = "linear_46.b_0" shape = [1152] dtype = "float32" min_val = float("-2.07564") @@ -1756,6 +1916,7 @@ class Program_weight_tensor_parameter_159: class Program_weight_tensor_parameter_160: name = "parameter_160" + original_name = "linear_46.w_0" shape = [384, 1152] dtype = "float32" min_val = float("-0.35825") @@ -1767,6 +1928,7 @@ class Program_weight_tensor_parameter_160: class Program_weight_tensor_parameter_161: name = "parameter_161" + original_name = "layer_norm_25.b_0" shape = [384] dtype = "float32" min_val = float("-1.06654") @@ -1778,6 +1940,7 @@ class Program_weight_tensor_parameter_161: class Program_weight_tensor_parameter_162: name = "parameter_162" + original_name = "layer_norm_25.w_0" shape = [384] dtype = "float32" min_val = float("0.246651") @@ -1789,6 +1952,7 @@ class Program_weight_tensor_parameter_162: class Program_weight_tensor_parameter_163: name = "parameter_163" + original_name = "linear_45.b_0" shape = [384] dtype = "float32" min_val = float("-0.533534") @@ -1800,6 +1964,7 @@ class Program_weight_tensor_parameter_163: class Program_weight_tensor_parameter_164: name = "parameter_164" + original_name = "linear_45.w_0" shape = [1536, 384] dtype = "float32" min_val = float("-0.945635") @@ -1811,6 +1976,7 @@ class Program_weight_tensor_parameter_164: class Program_weight_tensor_parameter_165: name = "parameter_165" + original_name = "linear_44.b_0" shape = [1536] dtype = "float32" min_val = float("-1.49697") @@ -1822,6 +1988,7 @@ class Program_weight_tensor_parameter_165: class Program_weight_tensor_parameter_166: name = "parameter_166" + original_name = "linear_44.w_0" shape = [384, 1536] dtype = "float32" min_val = float("-0.320538") @@ -1833,6 +2000,7 @@ class Program_weight_tensor_parameter_166: class Program_weight_tensor_parameter_167: name = "parameter_167" + original_name = "layer_norm_24.b_0" shape = [384] dtype = "float32" min_val = float("-0.657233") @@ -1844,6 +2012,7 @@ class Program_weight_tensor_parameter_167: class Program_weight_tensor_parameter_168: name = "parameter_168" + original_name = "layer_norm_24.w_0" shape = [384] dtype = "float32" min_val = float("0.351283") @@ -1855,6 +2024,7 @@ class Program_weight_tensor_parameter_168: class Program_weight_tensor_parameter_169: name = "parameter_169" + original_name = "linear_43.b_0" shape = [384] dtype = "float32" min_val = float("-2.89821") @@ -1866,6 +2036,7 @@ class Program_weight_tensor_parameter_169: class Program_weight_tensor_parameter_170: name = "parameter_170" + original_name = "linear_43.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.213457") @@ -1877,6 +2048,7 @@ class Program_weight_tensor_parameter_170: class Program_weight_tensor_parameter_171: name = "parameter_171" + original_name = "linear_42.b_0" shape = [1152] dtype = "float32" min_val = float("-2.72827") @@ -1888,6 +2060,7 @@ class Program_weight_tensor_parameter_171: class Program_weight_tensor_parameter_172: name = "parameter_172" + original_name = "linear_42.w_0" shape = [384, 1152] dtype = "float32" min_val = float("-0.369358") @@ -1899,6 +2072,7 @@ class Program_weight_tensor_parameter_172: class Program_weight_tensor_parameter_173: name = "parameter_173" + original_name = "layer_norm_23.b_0" shape = [384] dtype = "float32" min_val = float("-0.975335") @@ -1910,6 +2084,7 @@ class Program_weight_tensor_parameter_173: class Program_weight_tensor_parameter_174: name = "parameter_174" + original_name = "layer_norm_23.w_0" shape = [384] dtype = "float32" min_val = float("0.227436") @@ -1921,6 +2096,7 @@ class Program_weight_tensor_parameter_174: class Program_weight_tensor_parameter_175: name = "parameter_175" + original_name = "linear_41.b_0" shape = [384] dtype = "float32" min_val = float("-0.516349") @@ -1932,6 +2108,7 @@ class Program_weight_tensor_parameter_175: class Program_weight_tensor_parameter_176: name = "parameter_176" + original_name = "linear_41.w_0" shape = [1536, 384] dtype = "float32" min_val = float("-0.626911") @@ -1943,6 +2120,7 @@ class Program_weight_tensor_parameter_176: class Program_weight_tensor_parameter_177: name = "parameter_177" + original_name = "linear_40.b_0" shape = [1536] dtype = "float32" min_val = float("-1.26535") @@ -1954,6 +2132,7 @@ class Program_weight_tensor_parameter_177: class Program_weight_tensor_parameter_178: name = "parameter_178" + original_name = "linear_40.w_0" shape = [384, 1536] dtype = "float32" min_val = float("-0.328861") @@ -1965,6 +2144,7 @@ class Program_weight_tensor_parameter_178: class Program_weight_tensor_parameter_179: name = "parameter_179" + original_name = "layer_norm_22.b_0" shape = [384] dtype = "float32" min_val = float("-0.756583") @@ -1976,6 +2156,7 @@ class Program_weight_tensor_parameter_179: class Program_weight_tensor_parameter_180: name = "parameter_180" + original_name = "layer_norm_22.w_0" shape = [384] dtype = "float32" min_val = float("0.36361") @@ -1987,6 +2168,7 @@ class Program_weight_tensor_parameter_180: class Program_weight_tensor_parameter_181: name = "parameter_181" + original_name = "linear_39.b_0" shape = [384] dtype = "float32" min_val = float("-2.20466") @@ -1998,6 +2180,7 @@ class Program_weight_tensor_parameter_181: class Program_weight_tensor_parameter_182: name = "parameter_182" + original_name = "linear_39.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.186932") @@ -2009,6 +2192,7 @@ class Program_weight_tensor_parameter_182: class Program_weight_tensor_parameter_183: name = "parameter_183" + original_name = "linear_38.b_0" shape = [1152] dtype = "float32" min_val = float("-2.16194") @@ -2020,6 +2204,7 @@ class Program_weight_tensor_parameter_183: class Program_weight_tensor_parameter_184: name = "parameter_184" + original_name = "linear_38.w_0" shape = [384, 1152] dtype = "float32" min_val = float("-0.520332") @@ -2031,6 +2216,7 @@ class Program_weight_tensor_parameter_184: class Program_weight_tensor_parameter_185: name = "parameter_185" + original_name = "layer_norm_21.b_0" shape = [384] dtype = "float32" min_val = float("-0.826969") @@ -2042,6 +2228,7 @@ class Program_weight_tensor_parameter_185: class Program_weight_tensor_parameter_186: name = "parameter_186" + original_name = "layer_norm_21.w_0" shape = [384] dtype = "float32" min_val = float("0.218688") @@ -2053,6 +2240,7 @@ class Program_weight_tensor_parameter_186: class Program_weight_tensor_parameter_187: name = "parameter_187" + original_name = "linear_37.b_0" shape = [384] dtype = "float32" min_val = float("-0.576411") @@ -2064,6 +2252,7 @@ class Program_weight_tensor_parameter_187: class Program_weight_tensor_parameter_188: name = "parameter_188" + original_name = "linear_37.w_0" shape = [1536, 384] dtype = "float32" min_val = float("-0.469745") @@ -2075,6 +2264,7 @@ class Program_weight_tensor_parameter_188: class Program_weight_tensor_parameter_189: name = "parameter_189" + original_name = "linear_36.b_0" shape = [1536] dtype = "float32" min_val = float("-1.20274") @@ -2086,6 +2276,7 @@ class Program_weight_tensor_parameter_189: class Program_weight_tensor_parameter_190: name = "parameter_190" + original_name = "linear_36.w_0" shape = [384, 1536] dtype = "float32" min_val = float("-0.371918") @@ -2097,6 +2288,7 @@ class Program_weight_tensor_parameter_190: class Program_weight_tensor_parameter_191: name = "parameter_191" + original_name = "layer_norm_20.b_0" shape = [384] dtype = "float32" min_val = float("-0.797306") @@ -2108,6 +2300,7 @@ class Program_weight_tensor_parameter_191: class Program_weight_tensor_parameter_192: name = "parameter_192" + original_name = "layer_norm_20.w_0" shape = [384] dtype = "float32" min_val = float("0.342003") @@ -2119,6 +2312,7 @@ class Program_weight_tensor_parameter_192: class Program_weight_tensor_parameter_193: name = "parameter_193" + original_name = "linear_35.b_0" shape = [384] dtype = "float32" min_val = float("-2.63692") @@ -2130,6 +2324,7 @@ class Program_weight_tensor_parameter_193: class Program_weight_tensor_parameter_194: name = "parameter_194" + original_name = "linear_35.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.203452") @@ -2141,6 +2336,7 @@ class Program_weight_tensor_parameter_194: class Program_weight_tensor_parameter_195: name = "parameter_195" + original_name = "linear_34.b_0" shape = [1152] dtype = "float32" min_val = float("-1.95274") @@ -2152,6 +2348,7 @@ class Program_weight_tensor_parameter_195: class Program_weight_tensor_parameter_196: name = "parameter_196" + original_name = "linear_34.w_0" shape = [384, 1152] dtype = "float32" min_val = float("-0.360782") @@ -2163,6 +2360,7 @@ class Program_weight_tensor_parameter_196: class Program_weight_tensor_parameter_197: name = "parameter_197" + original_name = "layer_norm_19.b_0" shape = [384] dtype = "float32" min_val = float("-0.901936") @@ -2174,6 +2372,7 @@ class Program_weight_tensor_parameter_197: class Program_weight_tensor_parameter_198: name = "parameter_198" + original_name = "layer_norm_19.w_0" shape = [384] dtype = "float32" min_val = float("0.122888") @@ -2185,6 +2384,7 @@ class Program_weight_tensor_parameter_198: class Program_weight_tensor_parameter_199: name = "parameter_199" + original_name = "linear_33.b_0" shape = [384] dtype = "float32" min_val = float("-1.083") @@ -2196,6 +2396,7 @@ class Program_weight_tensor_parameter_199: class Program_weight_tensor_parameter_200: name = "parameter_200" + original_name = "linear_33.w_0" shape = [1536, 384] dtype = "float32" min_val = float("-0.323959") @@ -2207,6 +2408,7 @@ class Program_weight_tensor_parameter_200: class Program_weight_tensor_parameter_201: name = "parameter_201" + original_name = "linear_32.b_0" shape = [1536] dtype = "float32" min_val = float("-1.38391") @@ -2218,6 +2420,7 @@ class Program_weight_tensor_parameter_201: class Program_weight_tensor_parameter_202: name = "parameter_202" + original_name = "linear_32.w_0" shape = [384, 1536] dtype = "float32" min_val = float("-0.308758") @@ -2229,6 +2432,7 @@ class Program_weight_tensor_parameter_202: class Program_weight_tensor_parameter_203: name = "parameter_203" + original_name = "layer_norm_18.b_0" shape = [384] dtype = "float32" min_val = float("-0.82231") @@ -2240,6 +2444,7 @@ class Program_weight_tensor_parameter_203: class Program_weight_tensor_parameter_204: name = "parameter_204" + original_name = "layer_norm_18.w_0" shape = [384] dtype = "float32" min_val = float("0.372362") @@ -2251,6 +2456,7 @@ class Program_weight_tensor_parameter_204: class Program_weight_tensor_parameter_205: name = "parameter_205" + original_name = "linear_31.b_0" shape = [384] dtype = "float32" min_val = float("-1.84355") @@ -2262,6 +2468,7 @@ class Program_weight_tensor_parameter_205: class Program_weight_tensor_parameter_206: name = "parameter_206" + original_name = "linear_31.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.202916") @@ -2273,6 +2480,7 @@ class Program_weight_tensor_parameter_206: class Program_weight_tensor_parameter_207: name = "parameter_207" + original_name = "linear_30.b_0" shape = [1152] dtype = "float32" min_val = float("-2.64987") @@ -2284,6 +2492,7 @@ class Program_weight_tensor_parameter_207: class Program_weight_tensor_parameter_208: name = "parameter_208" + original_name = "linear_30.w_0" shape = [384, 1152] dtype = "float32" min_val = float("-0.372791") @@ -2295,6 +2504,7 @@ class Program_weight_tensor_parameter_208: class Program_weight_tensor_parameter_209: name = "parameter_209" + original_name = "layer_norm_17.b_0" shape = [384] dtype = "float32" min_val = float("-0.814568") @@ -2306,6 +2516,7 @@ class Program_weight_tensor_parameter_209: class Program_weight_tensor_parameter_210: name = "parameter_210" + original_name = "layer_norm_17.w_0" shape = [384] dtype = "float32" min_val = float("0.00263237") @@ -2317,6 +2528,7 @@ class Program_weight_tensor_parameter_210: class Program_weight_tensor_parameter_211: name = "parameter_211" + original_name = "linear_29.b_0" shape = [384] dtype = "float32" min_val = float("-1.3743") @@ -2328,6 +2540,7 @@ class Program_weight_tensor_parameter_211: class Program_weight_tensor_parameter_212: name = "parameter_212" + original_name = "linear_29.w_0" shape = [1536, 384] dtype = "float32" min_val = float("-0.278414") @@ -2339,6 +2552,7 @@ class Program_weight_tensor_parameter_212: class Program_weight_tensor_parameter_213: name = "parameter_213" + original_name = "linear_28.b_0" shape = [1536] dtype = "float32" min_val = float("-1.48736") @@ -2350,6 +2564,7 @@ class Program_weight_tensor_parameter_213: class Program_weight_tensor_parameter_214: name = "parameter_214" + original_name = "linear_28.w_0" shape = [384, 1536] dtype = "float32" min_val = float("-0.35682") @@ -2361,6 +2576,7 @@ class Program_weight_tensor_parameter_214: class Program_weight_tensor_parameter_215: name = "parameter_215" + original_name = "layer_norm_16.b_0" shape = [384] dtype = "float32" min_val = float("-0.638386") @@ -2372,6 +2588,7 @@ class Program_weight_tensor_parameter_215: class Program_weight_tensor_parameter_216: name = "parameter_216" + original_name = "layer_norm_16.w_0" shape = [384] dtype = "float32" min_val = float("0.279277") @@ -2383,6 +2600,7 @@ class Program_weight_tensor_parameter_216: class Program_weight_tensor_parameter_217: name = "parameter_217" + original_name = "linear_27.b_0" shape = [384] dtype = "float32" min_val = float("-2.01219") @@ -2394,6 +2612,7 @@ class Program_weight_tensor_parameter_217: class Program_weight_tensor_parameter_218: name = "parameter_218" + original_name = "linear_27.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.20099") @@ -2405,6 +2624,7 @@ class Program_weight_tensor_parameter_218: class Program_weight_tensor_parameter_219: name = "parameter_219" + original_name = "linear_26.b_0" shape = [1152] dtype = "float32" min_val = float("-1.92649") @@ -2416,6 +2636,7 @@ class Program_weight_tensor_parameter_219: class Program_weight_tensor_parameter_220: name = "parameter_220" + original_name = "linear_26.w_0" shape = [384, 1152] dtype = "float32" min_val = float("-0.373671") @@ -2427,6 +2648,7 @@ class Program_weight_tensor_parameter_220: class Program_weight_tensor_parameter_221: name = "parameter_221" + original_name = "layer_norm_15.b_0" shape = [384] dtype = "float32" min_val = float("-0.736396") @@ -2438,6 +2660,7 @@ class Program_weight_tensor_parameter_221: class Program_weight_tensor_parameter_222: name = "parameter_222" + original_name = "layer_norm_15.w_0" shape = [384] dtype = "float32" min_val = float("-0.000770239") @@ -2449,6 +2672,7 @@ class Program_weight_tensor_parameter_222: class Program_weight_tensor_parameter_223: name = "parameter_223" + original_name = "linear_25.b_0" shape = [384] dtype = "float32" min_val = float("-1.69024") @@ -2460,6 +2684,7 @@ class Program_weight_tensor_parameter_223: class Program_weight_tensor_parameter_224: name = "parameter_224" + original_name = "linear_25.w_0" shape = [1536, 384] dtype = "float32" min_val = float("-0.416184") @@ -2471,6 +2696,7 @@ class Program_weight_tensor_parameter_224: class Program_weight_tensor_parameter_225: name = "parameter_225" + original_name = "linear_24.b_0" shape = [1536] dtype = "float32" min_val = float("-1.74762") @@ -2482,6 +2708,7 @@ class Program_weight_tensor_parameter_225: class Program_weight_tensor_parameter_226: name = "parameter_226" + original_name = "linear_24.w_0" shape = [384, 1536] dtype = "float32" min_val = float("-0.429046") @@ -2493,6 +2720,7 @@ class Program_weight_tensor_parameter_226: class Program_weight_tensor_parameter_227: name = "parameter_227" + original_name = "layer_norm_14.b_0" shape = [384] dtype = "float32" min_val = float("-0.587658") @@ -2504,6 +2732,7 @@ class Program_weight_tensor_parameter_227: class Program_weight_tensor_parameter_228: name = "parameter_228" + original_name = "layer_norm_14.w_0" shape = [384] dtype = "float32" min_val = float("-0.00338338") @@ -2515,6 +2744,7 @@ class Program_weight_tensor_parameter_228: class Program_weight_tensor_parameter_229: name = "parameter_229" + original_name = "linear_23.b_0" shape = [384] dtype = "float32" min_val = float("-1.78596") @@ -2526,6 +2756,7 @@ class Program_weight_tensor_parameter_229: class Program_weight_tensor_parameter_230: name = "parameter_230" + original_name = "linear_23.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.226251") @@ -2537,6 +2768,7 @@ class Program_weight_tensor_parameter_230: class Program_weight_tensor_parameter_231: name = "parameter_231" + original_name = "linear_22.b_0" shape = [1152] dtype = "float32" min_val = float("-1.71609") @@ -2548,6 +2780,7 @@ class Program_weight_tensor_parameter_231: class Program_weight_tensor_parameter_232: name = "parameter_232" + original_name = "linear_22.w_0" shape = [384, 1152] dtype = "float32" min_val = float("-0.412568") @@ -2559,6 +2792,7 @@ class Program_weight_tensor_parameter_232: class Program_weight_tensor_parameter_233: name = "parameter_233" + original_name = "layer_norm_13.b_0" shape = [384] dtype = "float32" min_val = float("-0.516615") @@ -2570,6 +2804,7 @@ class Program_weight_tensor_parameter_233: class Program_weight_tensor_parameter_234: name = "parameter_234" + original_name = "layer_norm_13.w_0" shape = [384] dtype = "float32" min_val = float("-0.00143289") @@ -2581,6 +2816,7 @@ class Program_weight_tensor_parameter_234: class Program_weight_tensor_parameter_235: name = "parameter_235" + original_name = "linear_21.b_0" shape = [384] dtype = "float32" min_val = float("-2.10452") @@ -2592,6 +2828,7 @@ class Program_weight_tensor_parameter_235: class Program_weight_tensor_parameter_236: name = "parameter_236" + original_name = "linear_21.w_0" shape = [1536, 384] dtype = "float32" min_val = float("-0.341343") @@ -2603,6 +2840,7 @@ class Program_weight_tensor_parameter_236: class Program_weight_tensor_parameter_237: name = "parameter_237" + original_name = "linear_20.b_0" shape = [1536] dtype = "float32" min_val = float("-1.73402") @@ -2614,6 +2852,7 @@ class Program_weight_tensor_parameter_237: class Program_weight_tensor_parameter_238: name = "parameter_238" + original_name = "linear_20.w_0" shape = [384, 1536] dtype = "float32" min_val = float("-0.343444") @@ -2625,6 +2864,7 @@ class Program_weight_tensor_parameter_238: class Program_weight_tensor_parameter_239: name = "parameter_239" + original_name = "layer_norm_12.b_0" shape = [384] dtype = "float32" min_val = float("-1.26545") @@ -2636,6 +2876,7 @@ class Program_weight_tensor_parameter_239: class Program_weight_tensor_parameter_240: name = "parameter_240" + original_name = "layer_norm_12.w_0" shape = [384] dtype = "float32" min_val = float("-0.00358661") @@ -2647,6 +2888,7 @@ class Program_weight_tensor_parameter_240: class Program_weight_tensor_parameter_241: name = "parameter_241" + original_name = "linear_19.b_0" shape = [384] dtype = "float32" min_val = float("-2.76872") @@ -2658,6 +2900,7 @@ class Program_weight_tensor_parameter_241: class Program_weight_tensor_parameter_242: name = "parameter_242" + original_name = "linear_19.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.265729") @@ -2669,6 +2912,7 @@ class Program_weight_tensor_parameter_242: class Program_weight_tensor_parameter_243: name = "parameter_243" + original_name = "linear_18.b_0" shape = [1152] dtype = "float32" min_val = float("-2.44061") @@ -2680,6 +2924,7 @@ class Program_weight_tensor_parameter_243: class Program_weight_tensor_parameter_244: name = "parameter_244" + original_name = "linear_18.w_0" shape = [384, 1152] dtype = "float32" min_val = float("-0.389007") @@ -2691,6 +2936,7 @@ class Program_weight_tensor_parameter_244: class Program_weight_tensor_parameter_245: name = "parameter_245" + original_name = "layer_norm_11.b_0" shape = [384] dtype = "float32" min_val = float("-0.617775") @@ -2702,6 +2948,7 @@ class Program_weight_tensor_parameter_245: class Program_weight_tensor_parameter_246: name = "parameter_246" + original_name = "layer_norm_11.w_0" shape = [384] dtype = "float32" min_val = float("-0.00255328") @@ -2713,6 +2960,7 @@ class Program_weight_tensor_parameter_246: class Program_weight_tensor_parameter_247: name = "parameter_247" + original_name = "linear_17.w_0" shape = [768, 384] dtype = "float32" min_val = float("-0.217088") @@ -2724,6 +2972,7 @@ class Program_weight_tensor_parameter_247: class Program_weight_tensor_parameter_248: name = "parameter_248" + original_name = "layer_norm_10.b_0" shape = [768] dtype = "float32" min_val = float("-1.00131") @@ -2735,6 +2984,7 @@ class Program_weight_tensor_parameter_248: class Program_weight_tensor_parameter_249: name = "parameter_249" + original_name = "layer_norm_10.w_0" shape = [768] dtype = "float32" min_val = float("0.260274") @@ -2746,6 +2996,7 @@ class Program_weight_tensor_parameter_249: class Program_weight_tensor_parameter_250: name = "parameter_250" + original_name = "linear_16.b_0" shape = [192] dtype = "float32" min_val = float("-2.58312") @@ -2757,6 +3008,7 @@ class Program_weight_tensor_parameter_250: class Program_weight_tensor_parameter_251: name = "parameter_251" + original_name = "linear_16.w_0" shape = [768, 192] dtype = "float32" min_val = float("-0.582268") @@ -2768,6 +3020,7 @@ class Program_weight_tensor_parameter_251: class Program_weight_tensor_parameter_252: name = "parameter_252" + original_name = "linear_15.b_0" shape = [768] dtype = "float32" min_val = float("-1.67661") @@ -2779,6 +3032,7 @@ class Program_weight_tensor_parameter_252: class Program_weight_tensor_parameter_253: name = "parameter_253" + original_name = "linear_15.w_0" shape = [192, 768] dtype = "float32" min_val = float("-0.424849") @@ -2790,6 +3044,7 @@ class Program_weight_tensor_parameter_253: class Program_weight_tensor_parameter_254: name = "parameter_254" + original_name = "layer_norm_9.b_0" shape = [192] dtype = "float32" min_val = float("-0.85163") @@ -2801,6 +3056,7 @@ class Program_weight_tensor_parameter_254: class Program_weight_tensor_parameter_255: name = "parameter_255" + original_name = "layer_norm_9.w_0" shape = [192] dtype = "float32" min_val = float("0.284885") @@ -2812,6 +3068,7 @@ class Program_weight_tensor_parameter_255: class Program_weight_tensor_parameter_256: name = "parameter_256" + original_name = "linear_14.b_0" shape = [192] dtype = "float32" min_val = float("-1.57499") @@ -2823,6 +3080,7 @@ class Program_weight_tensor_parameter_256: class Program_weight_tensor_parameter_257: name = "parameter_257" + original_name = "linear_14.w_0" shape = [192, 192] dtype = "float32" min_val = float("-0.217442") @@ -2834,6 +3092,7 @@ class Program_weight_tensor_parameter_257: class Program_weight_tensor_parameter_258: name = "parameter_258" + original_name = "linear_13.b_0" shape = [576] dtype = "float32" min_val = float("-1.66302") @@ -2845,6 +3104,7 @@ class Program_weight_tensor_parameter_258: class Program_weight_tensor_parameter_259: name = "parameter_259" + original_name = "linear_13.w_0" shape = [192, 576] dtype = "float32" min_val = float("-0.613112") @@ -2856,6 +3116,7 @@ class Program_weight_tensor_parameter_259: class Program_weight_tensor_parameter_260: name = "parameter_260" + original_name = "layer_norm_8.b_0" shape = [192] dtype = "float32" min_val = float("-0.914876") @@ -2867,6 +3128,7 @@ class Program_weight_tensor_parameter_260: class Program_weight_tensor_parameter_261: name = "parameter_261" + original_name = "layer_norm_8.w_0" shape = [192] dtype = "float32" min_val = float("0.445142") @@ -2878,6 +3140,7 @@ class Program_weight_tensor_parameter_261: class Program_weight_tensor_parameter_262: name = "parameter_262" + original_name = "linear_12.b_0" shape = [192] dtype = "float32" min_val = float("-1.835") @@ -2889,6 +3152,7 @@ class Program_weight_tensor_parameter_262: class Program_weight_tensor_parameter_263: name = "parameter_263" + original_name = "linear_12.w_0" shape = [768, 192] dtype = "float32" min_val = float("-0.279653") @@ -2900,6 +3164,7 @@ class Program_weight_tensor_parameter_263: class Program_weight_tensor_parameter_264: name = "parameter_264" + original_name = "linear_11.b_0" shape = [768] dtype = "float32" min_val = float("-1.66467") @@ -2911,6 +3176,7 @@ class Program_weight_tensor_parameter_264: class Program_weight_tensor_parameter_265: name = "parameter_265" + original_name = "linear_11.w_0" shape = [192, 768] dtype = "float32" min_val = float("-0.467728") @@ -2922,6 +3188,7 @@ class Program_weight_tensor_parameter_265: class Program_weight_tensor_parameter_266: name = "parameter_266" + original_name = "layer_norm_7.b_0" shape = [192] dtype = "float32" min_val = float("-0.379318") @@ -2933,6 +3200,7 @@ class Program_weight_tensor_parameter_266: class Program_weight_tensor_parameter_267: name = "parameter_267" + original_name = "layer_norm_7.w_0" shape = [192] dtype = "float32" min_val = float("0.00771357") @@ -2944,6 +3212,7 @@ class Program_weight_tensor_parameter_267: class Program_weight_tensor_parameter_268: name = "parameter_268" + original_name = "linear_10.b_0" shape = [192] dtype = "float32" min_val = float("-1.55051") @@ -2955,6 +3224,7 @@ class Program_weight_tensor_parameter_268: class Program_weight_tensor_parameter_269: name = "parameter_269" + original_name = "linear_10.w_0" shape = [192, 192] dtype = "float32" min_val = float("-0.19929") @@ -2966,6 +3236,7 @@ class Program_weight_tensor_parameter_269: class Program_weight_tensor_parameter_270: name = "parameter_270" + original_name = "linear_9.b_0" shape = [576] dtype = "float32" min_val = float("-1.6874") @@ -2977,6 +3248,7 @@ class Program_weight_tensor_parameter_270: class Program_weight_tensor_parameter_271: name = "parameter_271" + original_name = "linear_9.w_0" shape = [192, 576] dtype = "float32" min_val = float("-0.411495") @@ -2988,6 +3260,7 @@ class Program_weight_tensor_parameter_271: class Program_weight_tensor_parameter_272: name = "parameter_272" + original_name = "layer_norm_6.b_0" shape = [192] dtype = "float32" min_val = float("-0.368167") @@ -2999,6 +3272,7 @@ class Program_weight_tensor_parameter_272: class Program_weight_tensor_parameter_273: name = "parameter_273" + original_name = "layer_norm_6.w_0" shape = [192] dtype = "float32" min_val = float("0.271828") @@ -3010,6 +3284,7 @@ class Program_weight_tensor_parameter_273: class Program_weight_tensor_parameter_274: name = "parameter_274" + original_name = "linear_8.w_0" shape = [384, 192] dtype = "float32" min_val = float("-0.223629") @@ -3021,6 +3296,7 @@ class Program_weight_tensor_parameter_274: class Program_weight_tensor_parameter_275: name = "parameter_275" + original_name = "layer_norm_5.b_0" shape = [384] dtype = "float32" min_val = float("-1.57101") @@ -3032,6 +3308,7 @@ class Program_weight_tensor_parameter_275: class Program_weight_tensor_parameter_276: name = "parameter_276" + original_name = "layer_norm_5.w_0" shape = [384] dtype = "float32" min_val = float("0.315696") @@ -3043,6 +3320,7 @@ class Program_weight_tensor_parameter_276: class Program_weight_tensor_parameter_277: name = "parameter_277" + original_name = "linear_7.b_0" shape = [96] dtype = "float32" min_val = float("-1.06289") @@ -3054,6 +3332,7 @@ class Program_weight_tensor_parameter_277: class Program_weight_tensor_parameter_278: name = "parameter_278" + original_name = "linear_7.w_0" shape = [384, 96] dtype = "float32" min_val = float("-0.377557") @@ -3065,6 +3344,7 @@ class Program_weight_tensor_parameter_278: class Program_weight_tensor_parameter_279: name = "parameter_279" + original_name = "linear_6.b_0" shape = [384] dtype = "float32" min_val = float("-1.52346") @@ -3076,6 +3356,7 @@ class Program_weight_tensor_parameter_279: class Program_weight_tensor_parameter_280: name = "parameter_280" + original_name = "linear_6.w_0" shape = [96, 384] dtype = "float32" min_val = float("-0.405638") @@ -3087,6 +3368,7 @@ class Program_weight_tensor_parameter_280: class Program_weight_tensor_parameter_281: name = "parameter_281" + original_name = "layer_norm_4.b_0" shape = [96] dtype = "float32" min_val = float("-0.615737") @@ -3098,6 +3380,7 @@ class Program_weight_tensor_parameter_281: class Program_weight_tensor_parameter_282: name = "parameter_282" + original_name = "layer_norm_4.w_0" shape = [96] dtype = "float32" min_val = float("0.356227") @@ -3109,6 +3392,7 @@ class Program_weight_tensor_parameter_282: class Program_weight_tensor_parameter_283: name = "parameter_283" + original_name = "linear_5.b_0" shape = [96] dtype = "float32" min_val = float("-0.436536") @@ -3120,6 +3404,7 @@ class Program_weight_tensor_parameter_283: class Program_weight_tensor_parameter_284: name = "parameter_284" + original_name = "linear_5.w_0" shape = [96, 96] dtype = "float32" min_val = float("-0.223395") @@ -3131,6 +3416,7 @@ class Program_weight_tensor_parameter_284: class Program_weight_tensor_parameter_285: name = "parameter_285" + original_name = "linear_4.b_0" shape = [288] dtype = "float32" min_val = float("-1.52489") @@ -3142,6 +3428,7 @@ class Program_weight_tensor_parameter_285: class Program_weight_tensor_parameter_286: name = "parameter_286" + original_name = "linear_4.w_0" shape = [96, 288] dtype = "float32" min_val = float("-0.397901") @@ -3153,6 +3440,7 @@ class Program_weight_tensor_parameter_286: class Program_weight_tensor_parameter_287: name = "parameter_287" + original_name = "layer_norm_3.b_0" shape = [96] dtype = "float32" min_val = float("-1.33122") @@ -3164,6 +3452,7 @@ class Program_weight_tensor_parameter_287: class Program_weight_tensor_parameter_288: name = "parameter_288" + original_name = "layer_norm_3.w_0" shape = [96] dtype = "float32" min_val = float("0.24892") @@ -3175,6 +3464,7 @@ class Program_weight_tensor_parameter_288: class Program_weight_tensor_parameter_289: name = "parameter_289" + original_name = "linear_3.b_0" shape = [96] dtype = "float32" min_val = float("-1.45175") @@ -3186,6 +3476,7 @@ class Program_weight_tensor_parameter_289: class Program_weight_tensor_parameter_290: name = "parameter_290" + original_name = "linear_3.w_0" shape = [384, 96] dtype = "float32" min_val = float("-0.334583") @@ -3197,6 +3488,7 @@ class Program_weight_tensor_parameter_290: class Program_weight_tensor_parameter_291: name = "parameter_291" + original_name = "linear_2.b_0" shape = [384] dtype = "float32" min_val = float("-1.62608") @@ -3208,6 +3500,7 @@ class Program_weight_tensor_parameter_291: class Program_weight_tensor_parameter_292: name = "parameter_292" + original_name = "linear_2.w_0" shape = [96, 384] dtype = "float32" min_val = float("-0.531675") @@ -3219,6 +3512,7 @@ class Program_weight_tensor_parameter_292: class Program_weight_tensor_parameter_293: name = "parameter_293" + original_name = "layer_norm_2.b_0" shape = [96] dtype = "float32" min_val = float("-1.07733") @@ -3230,6 +3524,7 @@ class Program_weight_tensor_parameter_293: class Program_weight_tensor_parameter_294: name = "parameter_294" + original_name = "layer_norm_2.w_0" shape = [96] dtype = "float32" min_val = float("0.671312") @@ -3241,6 +3536,7 @@ class Program_weight_tensor_parameter_294: class Program_weight_tensor_parameter_295: name = "parameter_295" + original_name = "linear_1.b_0" shape = [96] dtype = "float32" min_val = float("-0.79019") @@ -3252,6 +3548,7 @@ class Program_weight_tensor_parameter_295: class Program_weight_tensor_parameter_296: name = "parameter_296" + original_name = "linear_1.w_0" shape = [96, 96] dtype = "float32" min_val = float("-0.241572") @@ -3263,6 +3560,7 @@ class Program_weight_tensor_parameter_296: class Program_weight_tensor_parameter_297: name = "parameter_297" + original_name = "linear_0.b_0" shape = [288] dtype = "float32" min_val = float("-1.56233") @@ -3274,6 +3572,7 @@ class Program_weight_tensor_parameter_297: class Program_weight_tensor_parameter_298: name = "parameter_298" + original_name = "linear_0.w_0" shape = [96, 288] dtype = "float32" min_val = float("-0.755634") @@ -3285,6 +3584,7 @@ class Program_weight_tensor_parameter_298: class Program_weight_tensor_parameter_299: name = "parameter_299" + original_name = "layer_norm_1.b_0" shape = [96] dtype = "float32" min_val = float("-0.81511") @@ -3296,6 +3596,7 @@ class Program_weight_tensor_parameter_299: class Program_weight_tensor_parameter_300: name = "parameter_300" + original_name = "layer_norm_1.w_0" shape = [96] dtype = "float32" min_val = float("-0.00042576") @@ -3307,6 +3608,7 @@ class Program_weight_tensor_parameter_300: class Program_weight_tensor_parameter_301: name = "parameter_301" + original_name = "layer_norm_0.b_0" shape = [96] dtype = "float32" min_val = float("-1.47464") @@ -3318,6 +3620,7 @@ class Program_weight_tensor_parameter_301: class Program_weight_tensor_parameter_302: name = "parameter_302" + original_name = "layer_norm_0.w_0" shape = [96] dtype = "float32" min_val = float("-0.00955986") @@ -3329,6 +3632,7 @@ class Program_weight_tensor_parameter_302: class Program_weight_tensor_parameter_303: name = "parameter_303" + original_name = "conv2d_0.b_0" shape = [96] dtype = "float32" min_val = float("-1.35301") @@ -3340,6 +3644,7 @@ class Program_weight_tensor_parameter_303: class Program_weight_tensor_parameter_304: name = "parameter_304" + original_name = "conv2d_0.w_0" shape = [96, 3, 4, 4] dtype = "float32" min_val = float("-0.290897") diff --git a/paddle_samples/PaddleX/SwinTransformer_small_patch4_window7_224/subgraph_2/input_meta.py b/paddle_samples/PaddleX/SwinTransformer_small_patch4_window7_224/subgraph_2/input_meta.py index b6841ed94..4049f0250 100644 --- a/paddle_samples/PaddleX/SwinTransformer_small_patch4_window7_224/subgraph_2/input_meta.py +++ b/paddle_samples/PaddleX/SwinTransformer_small_patch4_window7_224/subgraph_2/input_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_data_0: name = "data_0" + original_name = "param_24" shape = [169, 3] dtype = "float32" min_val = float("-10.8571") @@ -11,6 +12,7 @@ class Program_weight_tensor_data_0: class Program_weight_tensor_data_1: name = "data_1" + original_name = "param_25" shape = [169, 3] dtype = "float32" min_val = float("-10.7214") @@ -22,6 +24,7 @@ class Program_weight_tensor_data_1: class Program_weight_tensor_data_2: name = "data_2" + original_name = "param_26" shape = [169, 6] dtype = "float32" min_val = float("-4.5908") @@ -33,6 +36,7 @@ class Program_weight_tensor_data_2: class Program_weight_tensor_data_3: name = "data_3" + original_name = "param_27" shape = [169, 6] dtype = "float32" min_val = float("-4.75664") @@ -44,6 +48,7 @@ class Program_weight_tensor_data_3: class Program_weight_tensor_data_4: name = "data_4" + original_name = "param_28" shape = [169, 12] dtype = "float32" min_val = float("-8.45031") @@ -55,6 +60,7 @@ class Program_weight_tensor_data_4: class Program_weight_tensor_data_5: name = "data_5" + original_name = "param_29" shape = [169, 12] dtype = "float32" min_val = float("-11.1694") @@ -66,6 +72,7 @@ class Program_weight_tensor_data_5: class Program_weight_tensor_data_6: name = "data_6" + original_name = "param_30" shape = [169, 12] dtype = "float32" min_val = float("-9.90787") @@ -77,6 +84,7 @@ class Program_weight_tensor_data_6: class Program_weight_tensor_data_7: name = "data_7" + original_name = "param_31" shape = [169, 12] dtype = "float32" min_val = float("-8.18133") @@ -88,6 +96,7 @@ class Program_weight_tensor_data_7: class Program_weight_tensor_data_8: name = "data_8" + original_name = "param_32" shape = [169, 12] dtype = "float32" min_val = float("-5.03937") @@ -99,6 +108,7 @@ class Program_weight_tensor_data_8: class Program_weight_tensor_data_9: name = "data_9" + original_name = "param_33" shape = [169, 12] dtype = "float32" min_val = float("-6.80829") @@ -110,6 +120,7 @@ class Program_weight_tensor_data_9: class Program_weight_tensor_data_10: name = "data_10" + original_name = "param_34" shape = [169, 12] dtype = "float32" min_val = float("-10.1306") @@ -121,6 +132,7 @@ class Program_weight_tensor_data_10: class Program_weight_tensor_data_11: name = "data_11" + original_name = "param_35" shape = [169, 12] dtype = "float32" min_val = float("-5.80442") @@ -132,6 +144,7 @@ class Program_weight_tensor_data_11: class Program_weight_tensor_data_12: name = "data_12" + original_name = "param_36" shape = [169, 12] dtype = "float32" min_val = float("-10.4214") @@ -143,6 +156,7 @@ class Program_weight_tensor_data_12: class Program_weight_tensor_data_13: name = "data_13" + original_name = "param_37" shape = [169, 12] dtype = "float32" min_val = float("-8.23471") @@ -154,6 +168,7 @@ class Program_weight_tensor_data_13: class Program_weight_tensor_data_14: name = "data_14" + original_name = "param_38" shape = [169, 12] dtype = "float32" min_val = float("-9.31986") @@ -165,6 +180,7 @@ class Program_weight_tensor_data_14: class Program_weight_tensor_data_15: name = "data_15" + original_name = "param_39" shape = [169, 12] dtype = "float32" min_val = float("-4.71889") @@ -176,6 +192,7 @@ class Program_weight_tensor_data_15: class Program_weight_tensor_data_16: name = "data_16" + original_name = "param_40" shape = [169, 12] dtype = "float32" min_val = float("-9.57186") @@ -187,6 +204,7 @@ class Program_weight_tensor_data_16: class Program_weight_tensor_data_17: name = "data_17" + original_name = "param_41" shape = [169, 12] dtype = "float32" min_val = float("-5.86106") @@ -198,6 +216,7 @@ class Program_weight_tensor_data_17: class Program_weight_tensor_data_18: name = "data_18" + original_name = "param_42" shape = [169, 12] dtype = "float32" min_val = float("-11.2027") @@ -209,6 +228,7 @@ class Program_weight_tensor_data_18: class Program_weight_tensor_data_19: name = "data_19" + original_name = "param_43" shape = [169, 12] dtype = "float32" min_val = float("-11.4153") @@ -220,6 +240,7 @@ class Program_weight_tensor_data_19: class Program_weight_tensor_data_20: name = "data_20" + original_name = "param_44" shape = [169, 12] dtype = "float32" min_val = float("-12.6649") @@ -231,6 +252,7 @@ class Program_weight_tensor_data_20: class Program_weight_tensor_data_21: name = "data_21" + original_name = "param_45" shape = [169, 12] dtype = "float32" min_val = float("-6.02263") @@ -242,6 +264,7 @@ class Program_weight_tensor_data_21: class Program_weight_tensor_data_22: name = "data_22" + original_name = "param_46" shape = [169, 24] dtype = "float32" min_val = float("-16.8576") @@ -253,6 +276,7 @@ class Program_weight_tensor_data_22: class Program_weight_tensor_data_23: name = "data_23" + original_name = "param_47" shape = [169, 24] dtype = "float32" min_val = float("-18.2114") @@ -264,6 +288,7 @@ class Program_weight_tensor_data_23: class Program_weight_tensor_data_24: name = "data_24" + original_name = "var_1731" shape = [60, 3, 224, 224] dtype = "float32" min_val = float("-5.29778") @@ -275,6 +300,7 @@ class Program_weight_tensor_data_24: class Program_weight_tensor_data_25: name = "data_25" + original_name = "var_1760" shape = [49, 49] dtype = "int64" min_val = 0 @@ -284,6 +310,7 @@ class Program_weight_tensor_data_25: class Program_weight_tensor_data_26: name = "data_26" + original_name = "var_1834" shape = [49, 49] dtype = "int64" min_val = 0 @@ -293,6 +320,7 @@ class Program_weight_tensor_data_26: class Program_weight_tensor_data_27: name = "data_27" + original_name = "var_1922" shape = [49, 49] dtype = "int64" min_val = 0 @@ -302,6 +330,7 @@ class Program_weight_tensor_data_27: class Program_weight_tensor_data_28: name = "data_28" + original_name = "var_2014" shape = [49, 49] dtype = "int64" min_val = 0 @@ -311,6 +340,7 @@ class Program_weight_tensor_data_28: class Program_weight_tensor_data_29: name = "data_29" + original_name = "var_2102" shape = [49, 49] dtype = "int64" min_val = 0 @@ -320,6 +350,7 @@ class Program_weight_tensor_data_29: class Program_weight_tensor_data_30: name = "data_30" + original_name = "var_2194" shape = [49, 49] dtype = "int64" min_val = 0 @@ -329,6 +360,7 @@ class Program_weight_tensor_data_30: class Program_weight_tensor_data_31: name = "data_31" + original_name = "var_2266" shape = [49, 49] dtype = "int64" min_val = 0 @@ -338,6 +370,7 @@ class Program_weight_tensor_data_31: class Program_weight_tensor_data_32: name = "data_32" + original_name = "var_2358" shape = [49, 49] dtype = "int64" min_val = 0 @@ -347,6 +380,7 @@ class Program_weight_tensor_data_32: class Program_weight_tensor_data_33: name = "data_33" + original_name = "var_2430" shape = [49, 49] dtype = "int64" min_val = 0 @@ -356,6 +390,7 @@ class Program_weight_tensor_data_33: class Program_weight_tensor_data_34: name = "data_34" + original_name = "var_2522" shape = [49, 49] dtype = "int64" min_val = 0 @@ -365,6 +400,7 @@ class Program_weight_tensor_data_34: class Program_weight_tensor_data_35: name = "data_35" + original_name = "var_2594" shape = [49, 49] dtype = "int64" min_val = 0 @@ -374,6 +410,7 @@ class Program_weight_tensor_data_35: class Program_weight_tensor_data_36: name = "data_36" + original_name = "var_2686" shape = [49, 49] dtype = "int64" min_val = 0 @@ -383,6 +420,7 @@ class Program_weight_tensor_data_36: class Program_weight_tensor_data_37: name = "data_37" + original_name = "var_2758" shape = [49, 49] dtype = "int64" min_val = 0 @@ -392,6 +430,7 @@ class Program_weight_tensor_data_37: class Program_weight_tensor_data_38: name = "data_38" + original_name = "var_2850" shape = [49, 49] dtype = "int64" min_val = 0 @@ -401,6 +440,7 @@ class Program_weight_tensor_data_38: class Program_weight_tensor_data_39: name = "data_39" + original_name = "var_2922" shape = [49, 49] dtype = "int64" min_val = 0 @@ -410,6 +450,7 @@ class Program_weight_tensor_data_39: class Program_weight_tensor_data_40: name = "data_40" + original_name = "var_3014" shape = [49, 49] dtype = "int64" min_val = 0 @@ -419,6 +460,7 @@ class Program_weight_tensor_data_40: class Program_weight_tensor_data_41: name = "data_41" + original_name = "var_3086" shape = [49, 49] dtype = "int64" min_val = 0 @@ -428,6 +470,7 @@ class Program_weight_tensor_data_41: class Program_weight_tensor_data_42: name = "data_42" + original_name = "var_3178" shape = [49, 49] dtype = "int64" min_val = 0 @@ -437,6 +480,7 @@ class Program_weight_tensor_data_42: class Program_weight_tensor_data_43: name = "data_43" + original_name = "var_3250" shape = [49, 49] dtype = "int64" min_val = 0 @@ -446,6 +490,7 @@ class Program_weight_tensor_data_43: class Program_weight_tensor_data_44: name = "data_44" + original_name = "var_3342" shape = [49, 49] dtype = "int64" min_val = 0 @@ -455,6 +500,7 @@ class Program_weight_tensor_data_44: class Program_weight_tensor_data_45: name = "data_45" + original_name = "var_3414" shape = [49, 49] dtype = "int64" min_val = 0 @@ -464,6 +510,7 @@ class Program_weight_tensor_data_45: class Program_weight_tensor_data_46: name = "data_46" + original_name = "var_3506" shape = [49, 49] dtype = "int64" min_val = 0 @@ -473,6 +520,7 @@ class Program_weight_tensor_data_46: class Program_weight_tensor_data_47: name = "data_47" + original_name = "var_3594" shape = [49, 49] dtype = "int64" min_val = 0 @@ -482,6 +530,7 @@ class Program_weight_tensor_data_47: class Program_weight_tensor_data_48: name = "data_48" + original_name = "var_3686" shape = [49, 49] dtype = "int64" min_val = 0 diff --git a/paddle_samples/PaddleX/SwinTransformer_small_patch4_window7_224/subgraph_2/weight_meta.py b/paddle_samples/PaddleX/SwinTransformer_small_patch4_window7_224/subgraph_2/weight_meta.py index 6d4e8bafa..28a670d37 100644 --- a/paddle_samples/PaddleX/SwinTransformer_small_patch4_window7_224/subgraph_2/weight_meta.py +++ b/paddle_samples/PaddleX/SwinTransformer_small_patch4_window7_224/subgraph_2/weight_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_parameter_0: name = "parameter_0" + original_name = "linear_99.b_0" shape = [102] dtype = "float32" min_val = float("-0.00275691") @@ -11,6 +12,7 @@ class Program_weight_tensor_parameter_0: class Program_weight_tensor_parameter_1: name = "parameter_1" + original_name = "linear_99.w_0" shape = [768, 102] dtype = "float32" min_val = float("-0.0907997") @@ -22,6 +24,7 @@ class Program_weight_tensor_parameter_1: class Program_weight_tensor_parameter_2: name = "parameter_2" + original_name = "layer_norm_52.b_0" shape = [768] dtype = "float32" min_val = float("-0.58873") @@ -33,6 +36,7 @@ class Program_weight_tensor_parameter_2: class Program_weight_tensor_parameter_3: name = "parameter_3" + original_name = "layer_norm_52.w_0" shape = [768] dtype = "float32" min_val = float("0.000842425") @@ -44,6 +48,7 @@ class Program_weight_tensor_parameter_3: class Program_weight_tensor_parameter_4: name = "parameter_4" + original_name = "linear_98.b_0" shape = [768] dtype = "float32" min_val = float("-10.7096") @@ -55,6 +60,7 @@ class Program_weight_tensor_parameter_4: class Program_weight_tensor_parameter_5: name = "parameter_5" + original_name = "linear_98.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-1.75495") @@ -66,6 +72,7 @@ class Program_weight_tensor_parameter_5: class Program_weight_tensor_parameter_6: name = "parameter_6" + original_name = "linear_97.b_0" shape = [3072] dtype = "float32" min_val = float("-2.04607") @@ -77,6 +84,7 @@ class Program_weight_tensor_parameter_6: class Program_weight_tensor_parameter_7: name = "parameter_7" + original_name = "linear_97.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.714835") @@ -88,6 +96,7 @@ class Program_weight_tensor_parameter_7: class Program_weight_tensor_parameter_8: name = "parameter_8" + original_name = "layer_norm_51.b_0" shape = [768] dtype = "float32" min_val = float("-1.48995") @@ -99,6 +108,7 @@ class Program_weight_tensor_parameter_8: class Program_weight_tensor_parameter_9: name = "parameter_9" + original_name = "layer_norm_51.w_0" shape = [768] dtype = "float32" min_val = float("-0.00512947") @@ -110,6 +120,7 @@ class Program_weight_tensor_parameter_9: class Program_weight_tensor_parameter_10: name = "parameter_10" + original_name = "linear_96.b_0" shape = [768] dtype = "float32" min_val = float("-10.5675") @@ -121,6 +132,7 @@ class Program_weight_tensor_parameter_10: class Program_weight_tensor_parameter_11: name = "parameter_11" + original_name = "linear_96.w_0" shape = [768, 768] dtype = "float32" min_val = float("-1.16212") @@ -132,6 +144,7 @@ class Program_weight_tensor_parameter_11: class Program_weight_tensor_parameter_12: name = "parameter_12" + original_name = "linear_95.b_0" shape = [2304] dtype = "float32" min_val = float("-3.09462") @@ -143,6 +156,7 @@ class Program_weight_tensor_parameter_12: class Program_weight_tensor_parameter_13: name = "parameter_13" + original_name = "linear_95.w_0" shape = [768, 2304] dtype = "float32" min_val = float("-0.49224") @@ -154,6 +168,7 @@ class Program_weight_tensor_parameter_13: class Program_weight_tensor_parameter_14: name = "parameter_14" + original_name = "layer_norm_50.b_0" shape = [768] dtype = "float32" min_val = float("-0.619258") @@ -165,6 +180,7 @@ class Program_weight_tensor_parameter_14: class Program_weight_tensor_parameter_15: name = "parameter_15" + original_name = "layer_norm_50.w_0" shape = [768] dtype = "float32" min_val = float("0.119699") @@ -176,6 +192,7 @@ class Program_weight_tensor_parameter_15: class Program_weight_tensor_parameter_16: name = "parameter_16" + original_name = "linear_94.b_0" shape = [768] dtype = "float32" min_val = float("-8.90211") @@ -187,6 +204,7 @@ class Program_weight_tensor_parameter_16: class Program_weight_tensor_parameter_17: name = "parameter_17" + original_name = "linear_94.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-1.61759") @@ -198,6 +216,7 @@ class Program_weight_tensor_parameter_17: class Program_weight_tensor_parameter_18: name = "parameter_18" + original_name = "linear_93.b_0" shape = [3072] dtype = "float32" min_val = float("-1.22323") @@ -209,6 +228,7 @@ class Program_weight_tensor_parameter_18: class Program_weight_tensor_parameter_19: name = "parameter_19" + original_name = "linear_93.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.428492") @@ -220,6 +240,7 @@ class Program_weight_tensor_parameter_19: class Program_weight_tensor_parameter_20: name = "parameter_20" + original_name = "layer_norm_49.b_0" shape = [768] dtype = "float32" min_val = float("-1.90147") @@ -231,6 +252,7 @@ class Program_weight_tensor_parameter_20: class Program_weight_tensor_parameter_21: name = "parameter_21" + original_name = "layer_norm_49.w_0" shape = [768] dtype = "float32" min_val = float("0.157915") @@ -242,6 +264,7 @@ class Program_weight_tensor_parameter_21: class Program_weight_tensor_parameter_22: name = "parameter_22" + original_name = "linear_92.b_0" shape = [768] dtype = "float32" min_val = float("-11.9685") @@ -253,6 +276,7 @@ class Program_weight_tensor_parameter_22: class Program_weight_tensor_parameter_23: name = "parameter_23" + original_name = "linear_92.w_0" shape = [768, 768] dtype = "float32" min_val = float("-1.0039") @@ -264,6 +288,7 @@ class Program_weight_tensor_parameter_23: class Program_weight_tensor_parameter_24: name = "parameter_24" + original_name = "linear_91.b_0" shape = [2304] dtype = "float32" min_val = float("-2.55998") @@ -275,6 +300,7 @@ class Program_weight_tensor_parameter_24: class Program_weight_tensor_parameter_25: name = "parameter_25" + original_name = "linear_91.w_0" shape = [768, 2304] dtype = "float32" min_val = float("-0.373395") @@ -286,6 +312,7 @@ class Program_weight_tensor_parameter_25: class Program_weight_tensor_parameter_26: name = "parameter_26" + original_name = "layer_norm_48.b_0" shape = [768] dtype = "float32" min_val = float("-0.478576") @@ -297,6 +324,7 @@ class Program_weight_tensor_parameter_26: class Program_weight_tensor_parameter_27: name = "parameter_27" + original_name = "layer_norm_48.w_0" shape = [768] dtype = "float32" min_val = float("0.000916656") @@ -308,6 +336,7 @@ class Program_weight_tensor_parameter_27: class Program_weight_tensor_parameter_28: name = "parameter_28" + original_name = "linear_90.w_0" shape = [1536, 768] dtype = "float32" min_val = float("-0.810195") @@ -319,6 +348,7 @@ class Program_weight_tensor_parameter_28: class Program_weight_tensor_parameter_29: name = "parameter_29" + original_name = "layer_norm_47.b_0" shape = [1536] dtype = "float32" min_val = float("-0.858392") @@ -330,6 +360,7 @@ class Program_weight_tensor_parameter_29: class Program_weight_tensor_parameter_30: name = "parameter_30" + original_name = "layer_norm_47.w_0" shape = [1536] dtype = "float32" min_val = float("0.109746") @@ -341,6 +372,7 @@ class Program_weight_tensor_parameter_30: class Program_weight_tensor_parameter_31: name = "parameter_31" + original_name = "linear_89.b_0" shape = [384] dtype = "float32" min_val = float("-1.56942") @@ -352,6 +384,7 @@ class Program_weight_tensor_parameter_31: class Program_weight_tensor_parameter_32: name = "parameter_32" + original_name = "linear_89.w_0" shape = [1536, 384] dtype = "float32" min_val = float("-2.019") @@ -363,6 +396,7 @@ class Program_weight_tensor_parameter_32: class Program_weight_tensor_parameter_33: name = "parameter_33" + original_name = "linear_88.b_0" shape = [1536] dtype = "float32" min_val = float("-1.12146") @@ -374,6 +408,7 @@ class Program_weight_tensor_parameter_33: class Program_weight_tensor_parameter_34: name = "parameter_34" + original_name = "linear_88.w_0" shape = [384, 1536] dtype = "float32" min_val = float("-0.585791") @@ -385,6 +420,7 @@ class Program_weight_tensor_parameter_34: class Program_weight_tensor_parameter_35: name = "parameter_35" + original_name = "layer_norm_46.b_0" shape = [384] dtype = "float32" min_val = float("-1.53089") @@ -396,6 +432,7 @@ class Program_weight_tensor_parameter_35: class Program_weight_tensor_parameter_36: name = "parameter_36" + original_name = "layer_norm_46.w_0" shape = [384] dtype = "float32" min_val = float("0.674541") @@ -407,6 +444,7 @@ class Program_weight_tensor_parameter_36: class Program_weight_tensor_parameter_37: name = "parameter_37" + original_name = "linear_87.b_0" shape = [384] dtype = "float32" min_val = float("-1.38952") @@ -418,6 +456,7 @@ class Program_weight_tensor_parameter_37: class Program_weight_tensor_parameter_38: name = "parameter_38" + original_name = "linear_87.w_0" shape = [384, 384] dtype = "float32" min_val = float("-1.18469") @@ -429,6 +468,7 @@ class Program_weight_tensor_parameter_38: class Program_weight_tensor_parameter_39: name = "parameter_39" + original_name = "linear_86.b_0" shape = [1152] dtype = "float32" min_val = float("-2.74369") @@ -440,6 +480,7 @@ class Program_weight_tensor_parameter_39: class Program_weight_tensor_parameter_40: name = "parameter_40" + original_name = "linear_86.w_0" shape = [384, 1152] dtype = "float32" min_val = float("-0.274347") @@ -451,6 +492,7 @@ class Program_weight_tensor_parameter_40: class Program_weight_tensor_parameter_41: name = "parameter_41" + original_name = "layer_norm_45.b_0" shape = [384] dtype = "float32" min_val = float("-1.11057") @@ -462,6 +504,7 @@ class Program_weight_tensor_parameter_41: class Program_weight_tensor_parameter_42: name = "parameter_42" + original_name = "layer_norm_45.w_0" shape = [384] dtype = "float32" min_val = float("0.0948987") @@ -473,6 +516,7 @@ class Program_weight_tensor_parameter_42: class Program_weight_tensor_parameter_43: name = "parameter_43" + original_name = "linear_85.b_0" shape = [384] dtype = "float32" min_val = float("-1.80932") @@ -484,6 +528,7 @@ class Program_weight_tensor_parameter_43: class Program_weight_tensor_parameter_44: name = "parameter_44" + original_name = "linear_85.w_0" shape = [1536, 384] dtype = "float32" min_val = float("-1.22833") @@ -495,6 +540,7 @@ class Program_weight_tensor_parameter_44: class Program_weight_tensor_parameter_45: name = "parameter_45" + original_name = "linear_84.b_0" shape = [1536] dtype = "float32" min_val = float("-1.1406") @@ -506,6 +552,7 @@ class Program_weight_tensor_parameter_45: class Program_weight_tensor_parameter_46: name = "parameter_46" + original_name = "linear_84.w_0" shape = [384, 1536] dtype = "float32" min_val = float("-0.733011") @@ -517,6 +564,7 @@ class Program_weight_tensor_parameter_46: class Program_weight_tensor_parameter_47: name = "parameter_47" + original_name = "layer_norm_44.b_0" shape = [384] dtype = "float32" min_val = float("-1.68204") @@ -528,6 +576,7 @@ class Program_weight_tensor_parameter_47: class Program_weight_tensor_parameter_48: name = "parameter_48" + original_name = "layer_norm_44.w_0" shape = [384] dtype = "float32" min_val = float("0.503251") @@ -539,6 +588,7 @@ class Program_weight_tensor_parameter_48: class Program_weight_tensor_parameter_49: name = "parameter_49" + original_name = "linear_83.b_0" shape = [384] dtype = "float32" min_val = float("-1.46483") @@ -550,6 +600,7 @@ class Program_weight_tensor_parameter_49: class Program_weight_tensor_parameter_50: name = "parameter_50" + original_name = "linear_83.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.53044") @@ -561,6 +612,7 @@ class Program_weight_tensor_parameter_50: class Program_weight_tensor_parameter_51: name = "parameter_51" + original_name = "linear_82.b_0" shape = [1152] dtype = "float32" min_val = float("-3.09941") @@ -572,6 +624,7 @@ class Program_weight_tensor_parameter_51: class Program_weight_tensor_parameter_52: name = "parameter_52" + original_name = "linear_82.w_0" shape = [384, 1152] dtype = "float32" min_val = float("-0.391412") @@ -583,6 +636,7 @@ class Program_weight_tensor_parameter_52: class Program_weight_tensor_parameter_53: name = "parameter_53" + original_name = "layer_norm_43.b_0" shape = [384] dtype = "float32" min_val = float("-1.19468") @@ -594,6 +648,7 @@ class Program_weight_tensor_parameter_53: class Program_weight_tensor_parameter_54: name = "parameter_54" + original_name = "layer_norm_43.w_0" shape = [384] dtype = "float32" min_val = float("0.130782") @@ -605,6 +660,7 @@ class Program_weight_tensor_parameter_54: class Program_weight_tensor_parameter_55: name = "parameter_55" + original_name = "linear_81.b_0" shape = [384] dtype = "float32" min_val = float("-1.32465") @@ -616,6 +672,7 @@ class Program_weight_tensor_parameter_55: class Program_weight_tensor_parameter_56: name = "parameter_56" + original_name = "linear_81.w_0" shape = [1536, 384] dtype = "float32" min_val = float("-1.1547") @@ -627,6 +684,7 @@ class Program_weight_tensor_parameter_56: class Program_weight_tensor_parameter_57: name = "parameter_57" + original_name = "linear_80.b_0" shape = [1536] dtype = "float32" min_val = float("-1.14948") @@ -638,6 +696,7 @@ class Program_weight_tensor_parameter_57: class Program_weight_tensor_parameter_58: name = "parameter_58" + original_name = "linear_80.w_0" shape = [384, 1536] dtype = "float32" min_val = float("-0.351994") @@ -649,6 +708,7 @@ class Program_weight_tensor_parameter_58: class Program_weight_tensor_parameter_59: name = "parameter_59" + original_name = "layer_norm_42.b_0" shape = [384] dtype = "float32" min_val = float("-1.70646") @@ -660,6 +720,7 @@ class Program_weight_tensor_parameter_59: class Program_weight_tensor_parameter_60: name = "parameter_60" + original_name = "layer_norm_42.w_0" shape = [384] dtype = "float32" min_val = float("0.332374") @@ -671,6 +732,7 @@ class Program_weight_tensor_parameter_60: class Program_weight_tensor_parameter_61: name = "parameter_61" + original_name = "linear_79.b_0" shape = [384] dtype = "float32" min_val = float("-1.58876") @@ -682,6 +744,7 @@ class Program_weight_tensor_parameter_61: class Program_weight_tensor_parameter_62: name = "parameter_62" + original_name = "linear_79.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.356805") @@ -693,6 +756,7 @@ class Program_weight_tensor_parameter_62: class Program_weight_tensor_parameter_63: name = "parameter_63" + original_name = "linear_78.b_0" shape = [1152] dtype = "float32" min_val = float("-2.80721") @@ -704,6 +768,7 @@ class Program_weight_tensor_parameter_63: class Program_weight_tensor_parameter_64: name = "parameter_64" + original_name = "linear_78.w_0" shape = [384, 1152] dtype = "float32" min_val = float("-0.26204") @@ -715,6 +780,7 @@ class Program_weight_tensor_parameter_64: class Program_weight_tensor_parameter_65: name = "parameter_65" + original_name = "layer_norm_41.b_0" shape = [384] dtype = "float32" min_val = float("-1.17464") @@ -726,6 +792,7 @@ class Program_weight_tensor_parameter_65: class Program_weight_tensor_parameter_66: name = "parameter_66" + original_name = "layer_norm_41.w_0" shape = [384] dtype = "float32" min_val = float("0.0800777") @@ -737,6 +804,7 @@ class Program_weight_tensor_parameter_66: class Program_weight_tensor_parameter_67: name = "parameter_67" + original_name = "linear_77.b_0" shape = [384] dtype = "float32" min_val = float("-1.34456") @@ -748,6 +816,7 @@ class Program_weight_tensor_parameter_67: class Program_weight_tensor_parameter_68: name = "parameter_68" + original_name = "linear_77.w_0" shape = [1536, 384] dtype = "float32" min_val = float("-1.01622") @@ -759,6 +828,7 @@ class Program_weight_tensor_parameter_68: class Program_weight_tensor_parameter_69: name = "parameter_69" + original_name = "linear_76.b_0" shape = [1536] dtype = "float32" min_val = float("-1.25084") @@ -770,6 +840,7 @@ class Program_weight_tensor_parameter_69: class Program_weight_tensor_parameter_70: name = "parameter_70" + original_name = "linear_76.w_0" shape = [384, 1536] dtype = "float32" min_val = float("-0.247665") @@ -781,6 +852,7 @@ class Program_weight_tensor_parameter_70: class Program_weight_tensor_parameter_71: name = "parameter_71" + original_name = "layer_norm_40.b_0" shape = [384] dtype = "float32" min_val = float("-1.42262") @@ -792,6 +864,7 @@ class Program_weight_tensor_parameter_71: class Program_weight_tensor_parameter_72: name = "parameter_72" + original_name = "layer_norm_40.w_0" shape = [384] dtype = "float32" min_val = float("0.362027") @@ -803,6 +876,7 @@ class Program_weight_tensor_parameter_72: class Program_weight_tensor_parameter_73: name = "parameter_73" + original_name = "linear_75.b_0" shape = [384] dtype = "float32" min_val = float("-2.20534") @@ -814,6 +888,7 @@ class Program_weight_tensor_parameter_73: class Program_weight_tensor_parameter_74: name = "parameter_74" + original_name = "linear_75.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.365255") @@ -825,6 +900,7 @@ class Program_weight_tensor_parameter_74: class Program_weight_tensor_parameter_75: name = "parameter_75" + original_name = "linear_74.b_0" shape = [1152] dtype = "float32" min_val = float("-2.82746") @@ -836,6 +912,7 @@ class Program_weight_tensor_parameter_75: class Program_weight_tensor_parameter_76: name = "parameter_76" + original_name = "linear_74.w_0" shape = [384, 1152] dtype = "float32" min_val = float("-0.260112") @@ -847,6 +924,7 @@ class Program_weight_tensor_parameter_76: class Program_weight_tensor_parameter_77: name = "parameter_77" + original_name = "layer_norm_39.b_0" shape = [384] dtype = "float32" min_val = float("-1.42799") @@ -858,6 +936,7 @@ class Program_weight_tensor_parameter_77: class Program_weight_tensor_parameter_78: name = "parameter_78" + original_name = "layer_norm_39.w_0" shape = [384] dtype = "float32" min_val = float("0.150719") @@ -869,6 +948,7 @@ class Program_weight_tensor_parameter_78: class Program_weight_tensor_parameter_79: name = "parameter_79" + original_name = "linear_73.b_0" shape = [384] dtype = "float32" min_val = float("-0.78658") @@ -880,6 +960,7 @@ class Program_weight_tensor_parameter_79: class Program_weight_tensor_parameter_80: name = "parameter_80" + original_name = "linear_73.w_0" shape = [1536, 384] dtype = "float32" min_val = float("-1.60364") @@ -891,6 +972,7 @@ class Program_weight_tensor_parameter_80: class Program_weight_tensor_parameter_81: name = "parameter_81" + original_name = "linear_72.b_0" shape = [1536] dtype = "float32" min_val = float("-1.26863") @@ -902,6 +984,7 @@ class Program_weight_tensor_parameter_81: class Program_weight_tensor_parameter_82: name = "parameter_82" + original_name = "linear_72.w_0" shape = [384, 1536] dtype = "float32" min_val = float("-0.307197") @@ -913,6 +996,7 @@ class Program_weight_tensor_parameter_82: class Program_weight_tensor_parameter_83: name = "parameter_83" + original_name = "layer_norm_38.b_0" shape = [384] dtype = "float32" min_val = float("-1.0534") @@ -924,6 +1008,7 @@ class Program_weight_tensor_parameter_83: class Program_weight_tensor_parameter_84: name = "parameter_84" + original_name = "layer_norm_38.w_0" shape = [384] dtype = "float32" min_val = float("0.457603") @@ -935,6 +1020,7 @@ class Program_weight_tensor_parameter_84: class Program_weight_tensor_parameter_85: name = "parameter_85" + original_name = "linear_71.b_0" shape = [384] dtype = "float32" min_val = float("-2.20153") @@ -946,6 +1032,7 @@ class Program_weight_tensor_parameter_85: class Program_weight_tensor_parameter_86: name = "parameter_86" + original_name = "linear_71.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.248154") @@ -957,6 +1044,7 @@ class Program_weight_tensor_parameter_86: class Program_weight_tensor_parameter_87: name = "parameter_87" + original_name = "linear_70.b_0" shape = [1152] dtype = "float32" min_val = float("-2.46186") @@ -968,6 +1056,7 @@ class Program_weight_tensor_parameter_87: class Program_weight_tensor_parameter_88: name = "parameter_88" + original_name = "linear_70.w_0" shape = [384, 1152] dtype = "float32" min_val = float("-0.267893") @@ -979,6 +1068,7 @@ class Program_weight_tensor_parameter_88: class Program_weight_tensor_parameter_89: name = "parameter_89" + original_name = "layer_norm_37.b_0" shape = [384] dtype = "float32" min_val = float("-1.19284") @@ -990,6 +1080,7 @@ class Program_weight_tensor_parameter_89: class Program_weight_tensor_parameter_90: name = "parameter_90" + original_name = "layer_norm_37.w_0" shape = [384] dtype = "float32" min_val = float("0.170703") @@ -1001,6 +1092,7 @@ class Program_weight_tensor_parameter_90: class Program_weight_tensor_parameter_91: name = "parameter_91" + original_name = "linear_69.b_0" shape = [384] dtype = "float32" min_val = float("-0.974824") @@ -1012,6 +1104,7 @@ class Program_weight_tensor_parameter_91: class Program_weight_tensor_parameter_92: name = "parameter_92" + original_name = "linear_69.w_0" shape = [1536, 384] dtype = "float32" min_val = float("-2.68387") @@ -1023,6 +1116,7 @@ class Program_weight_tensor_parameter_92: class Program_weight_tensor_parameter_93: name = "parameter_93" + original_name = "linear_68.b_0" shape = [1536] dtype = "float32" min_val = float("-1.18748") @@ -1034,6 +1128,7 @@ class Program_weight_tensor_parameter_93: class Program_weight_tensor_parameter_94: name = "parameter_94" + original_name = "linear_68.w_0" shape = [384, 1536] dtype = "float32" min_val = float("-0.518484") @@ -1045,6 +1140,7 @@ class Program_weight_tensor_parameter_94: class Program_weight_tensor_parameter_95: name = "parameter_95" + original_name = "layer_norm_36.b_0" shape = [384] dtype = "float32" min_val = float("-1.16081") @@ -1056,6 +1152,7 @@ class Program_weight_tensor_parameter_95: class Program_weight_tensor_parameter_96: name = "parameter_96" + original_name = "layer_norm_36.w_0" shape = [384] dtype = "float32" min_val = float("0.494575") @@ -1067,6 +1164,7 @@ class Program_weight_tensor_parameter_96: class Program_weight_tensor_parameter_97: name = "parameter_97" + original_name = "linear_67.b_0" shape = [384] dtype = "float32" min_val = float("-3.34295") @@ -1078,6 +1176,7 @@ class Program_weight_tensor_parameter_97: class Program_weight_tensor_parameter_98: name = "parameter_98" + original_name = "linear_67.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.308424") @@ -1089,6 +1188,7 @@ class Program_weight_tensor_parameter_98: class Program_weight_tensor_parameter_99: name = "parameter_99" + original_name = "linear_66.b_0" shape = [1152] dtype = "float32" min_val = float("-2.8504") @@ -1100,6 +1200,7 @@ class Program_weight_tensor_parameter_99: class Program_weight_tensor_parameter_100: name = "parameter_100" + original_name = "linear_66.w_0" shape = [384, 1152] dtype = "float32" min_val = float("-0.364949") @@ -1111,6 +1212,7 @@ class Program_weight_tensor_parameter_100: class Program_weight_tensor_parameter_101: name = "parameter_101" + original_name = "layer_norm_35.b_0" shape = [384] dtype = "float32" min_val = float("-1.573") @@ -1122,6 +1224,7 @@ class Program_weight_tensor_parameter_101: class Program_weight_tensor_parameter_102: name = "parameter_102" + original_name = "layer_norm_35.w_0" shape = [384] dtype = "float32" min_val = float("0.264343") @@ -1133,6 +1236,7 @@ class Program_weight_tensor_parameter_102: class Program_weight_tensor_parameter_103: name = "parameter_103" + original_name = "linear_65.b_0" shape = [384] dtype = "float32" min_val = float("-0.862299") @@ -1144,6 +1248,7 @@ class Program_weight_tensor_parameter_103: class Program_weight_tensor_parameter_104: name = "parameter_104" + original_name = "linear_65.w_0" shape = [1536, 384] dtype = "float32" min_val = float("-2.37274") @@ -1155,6 +1260,7 @@ class Program_weight_tensor_parameter_104: class Program_weight_tensor_parameter_105: name = "parameter_105" + original_name = "linear_64.b_0" shape = [1536] dtype = "float32" min_val = float("-1.25132") @@ -1166,6 +1272,7 @@ class Program_weight_tensor_parameter_105: class Program_weight_tensor_parameter_106: name = "parameter_106" + original_name = "linear_64.w_0" shape = [384, 1536] dtype = "float32" min_val = float("-0.45676") @@ -1177,6 +1284,7 @@ class Program_weight_tensor_parameter_106: class Program_weight_tensor_parameter_107: name = "parameter_107" + original_name = "layer_norm_34.b_0" shape = [384] dtype = "float32" min_val = float("-1.24593") @@ -1188,6 +1296,7 @@ class Program_weight_tensor_parameter_107: class Program_weight_tensor_parameter_108: name = "parameter_108" + original_name = "layer_norm_34.w_0" shape = [384] dtype = "float32" min_val = float("0.522802") @@ -1199,6 +1308,7 @@ class Program_weight_tensor_parameter_108: class Program_weight_tensor_parameter_109: name = "parameter_109" + original_name = "linear_63.b_0" shape = [384] dtype = "float32" min_val = float("-2.26855") @@ -1210,6 +1320,7 @@ class Program_weight_tensor_parameter_109: class Program_weight_tensor_parameter_110: name = "parameter_110" + original_name = "linear_63.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.238243") @@ -1221,6 +1332,7 @@ class Program_weight_tensor_parameter_110: class Program_weight_tensor_parameter_111: name = "parameter_111" + original_name = "linear_62.b_0" shape = [1152] dtype = "float32" min_val = float("-2.12382") @@ -1232,6 +1344,7 @@ class Program_weight_tensor_parameter_111: class Program_weight_tensor_parameter_112: name = "parameter_112" + original_name = "linear_62.w_0" shape = [384, 1152] dtype = "float32" min_val = float("-0.308471") @@ -1243,6 +1356,7 @@ class Program_weight_tensor_parameter_112: class Program_weight_tensor_parameter_113: name = "parameter_113" + original_name = "layer_norm_33.b_0" shape = [384] dtype = "float32" min_val = float("-1.2656") @@ -1254,6 +1368,7 @@ class Program_weight_tensor_parameter_113: class Program_weight_tensor_parameter_114: name = "parameter_114" + original_name = "layer_norm_33.w_0" shape = [384] dtype = "float32" min_val = float("0.262784") @@ -1265,6 +1380,7 @@ class Program_weight_tensor_parameter_114: class Program_weight_tensor_parameter_115: name = "parameter_115" + original_name = "linear_61.b_0" shape = [384] dtype = "float32" min_val = float("-0.726691") @@ -1276,6 +1392,7 @@ class Program_weight_tensor_parameter_115: class Program_weight_tensor_parameter_116: name = "parameter_116" + original_name = "linear_61.w_0" shape = [1536, 384] dtype = "float32" min_val = float("-2.15743") @@ -1287,6 +1404,7 @@ class Program_weight_tensor_parameter_116: class Program_weight_tensor_parameter_117: name = "parameter_117" + original_name = "linear_60.b_0" shape = [1536] dtype = "float32" min_val = float("-1.28725") @@ -1298,6 +1416,7 @@ class Program_weight_tensor_parameter_117: class Program_weight_tensor_parameter_118: name = "parameter_118" + original_name = "linear_60.w_0" shape = [384, 1536] dtype = "float32" min_val = float("-0.440707") @@ -1309,6 +1428,7 @@ class Program_weight_tensor_parameter_118: class Program_weight_tensor_parameter_119: name = "parameter_119" + original_name = "layer_norm_32.b_0" shape = [384] dtype = "float32" min_val = float("-1.15318") @@ -1320,6 +1440,7 @@ class Program_weight_tensor_parameter_119: class Program_weight_tensor_parameter_120: name = "parameter_120" + original_name = "layer_norm_32.w_0" shape = [384] dtype = "float32" min_val = float("0.533149") @@ -1331,6 +1452,7 @@ class Program_weight_tensor_parameter_120: class Program_weight_tensor_parameter_121: name = "parameter_121" + original_name = "linear_59.b_0" shape = [384] dtype = "float32" min_val = float("-3.36633") @@ -1342,6 +1464,7 @@ class Program_weight_tensor_parameter_121: class Program_weight_tensor_parameter_122: name = "parameter_122" + original_name = "linear_59.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.340351") @@ -1353,6 +1476,7 @@ class Program_weight_tensor_parameter_122: class Program_weight_tensor_parameter_123: name = "parameter_123" + original_name = "linear_58.b_0" shape = [1152] dtype = "float32" min_val = float("-2.35512") @@ -1364,6 +1488,7 @@ class Program_weight_tensor_parameter_123: class Program_weight_tensor_parameter_124: name = "parameter_124" + original_name = "linear_58.w_0" shape = [384, 1152] dtype = "float32" min_val = float("-0.323251") @@ -1375,6 +1500,7 @@ class Program_weight_tensor_parameter_124: class Program_weight_tensor_parameter_125: name = "parameter_125" + original_name = "layer_norm_31.b_0" shape = [384] dtype = "float32" min_val = float("-1.52138") @@ -1386,6 +1512,7 @@ class Program_weight_tensor_parameter_125: class Program_weight_tensor_parameter_126: name = "parameter_126" + original_name = "layer_norm_31.w_0" shape = [384] dtype = "float32" min_val = float("0.34705") @@ -1397,6 +1524,7 @@ class Program_weight_tensor_parameter_126: class Program_weight_tensor_parameter_127: name = "parameter_127" + original_name = "linear_57.b_0" shape = [384] dtype = "float32" min_val = float("-0.416331") @@ -1408,6 +1536,7 @@ class Program_weight_tensor_parameter_127: class Program_weight_tensor_parameter_128: name = "parameter_128" + original_name = "linear_57.w_0" shape = [1536, 384] dtype = "float32" min_val = float("-1.86256") @@ -1419,6 +1548,7 @@ class Program_weight_tensor_parameter_128: class Program_weight_tensor_parameter_129: name = "parameter_129" + original_name = "linear_56.b_0" shape = [1536] dtype = "float32" min_val = float("-1.29568") @@ -1430,6 +1560,7 @@ class Program_weight_tensor_parameter_129: class Program_weight_tensor_parameter_130: name = "parameter_130" + original_name = "linear_56.w_0" shape = [384, 1536] dtype = "float32" min_val = float("-0.293086") @@ -1441,6 +1572,7 @@ class Program_weight_tensor_parameter_130: class Program_weight_tensor_parameter_131: name = "parameter_131" + original_name = "layer_norm_30.b_0" shape = [384] dtype = "float32" min_val = float("-1.12291") @@ -1452,6 +1584,7 @@ class Program_weight_tensor_parameter_131: class Program_weight_tensor_parameter_132: name = "parameter_132" + original_name = "layer_norm_30.w_0" shape = [384] dtype = "float32" min_val = float("0.505238") @@ -1463,6 +1596,7 @@ class Program_weight_tensor_parameter_132: class Program_weight_tensor_parameter_133: name = "parameter_133" + original_name = "linear_55.b_0" shape = [384] dtype = "float32" min_val = float("-1.99331") @@ -1474,6 +1608,7 @@ class Program_weight_tensor_parameter_133: class Program_weight_tensor_parameter_134: name = "parameter_134" + original_name = "linear_55.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.286958") @@ -1485,6 +1620,7 @@ class Program_weight_tensor_parameter_134: class Program_weight_tensor_parameter_135: name = "parameter_135" + original_name = "linear_54.b_0" shape = [1152] dtype = "float32" min_val = float("-2.07719") @@ -1496,6 +1632,7 @@ class Program_weight_tensor_parameter_135: class Program_weight_tensor_parameter_136: name = "parameter_136" + original_name = "linear_54.w_0" shape = [384, 1152] dtype = "float32" min_val = float("-0.357524") @@ -1507,6 +1644,7 @@ class Program_weight_tensor_parameter_136: class Program_weight_tensor_parameter_137: name = "parameter_137" + original_name = "layer_norm_29.b_0" shape = [384] dtype = "float32" min_val = float("-1.17257") @@ -1518,6 +1656,7 @@ class Program_weight_tensor_parameter_137: class Program_weight_tensor_parameter_138: name = "parameter_138" + original_name = "layer_norm_29.w_0" shape = [384] dtype = "float32" min_val = float("0.299071") @@ -1529,6 +1668,7 @@ class Program_weight_tensor_parameter_138: class Program_weight_tensor_parameter_139: name = "parameter_139" + original_name = "linear_53.b_0" shape = [384] dtype = "float32" min_val = float("-0.519723") @@ -1540,6 +1680,7 @@ class Program_weight_tensor_parameter_139: class Program_weight_tensor_parameter_140: name = "parameter_140" + original_name = "linear_53.w_0" shape = [1536, 384] dtype = "float32" min_val = float("-1.75449") @@ -1551,6 +1692,7 @@ class Program_weight_tensor_parameter_140: class Program_weight_tensor_parameter_141: name = "parameter_141" + original_name = "linear_52.b_0" shape = [1536] dtype = "float32" min_val = float("-1.25938") @@ -1562,6 +1704,7 @@ class Program_weight_tensor_parameter_141: class Program_weight_tensor_parameter_142: name = "parameter_142" + original_name = "linear_52.w_0" shape = [384, 1536] dtype = "float32" min_val = float("-0.311043") @@ -1573,6 +1716,7 @@ class Program_weight_tensor_parameter_142: class Program_weight_tensor_parameter_143: name = "parameter_143" + original_name = "layer_norm_28.b_0" shape = [384] dtype = "float32" min_val = float("-0.967654") @@ -1584,6 +1728,7 @@ class Program_weight_tensor_parameter_143: class Program_weight_tensor_parameter_144: name = "parameter_144" + original_name = "layer_norm_28.w_0" shape = [384] dtype = "float32" min_val = float("0.449948") @@ -1595,6 +1740,7 @@ class Program_weight_tensor_parameter_144: class Program_weight_tensor_parameter_145: name = "parameter_145" + original_name = "linear_51.b_0" shape = [384] dtype = "float32" min_val = float("-2.87205") @@ -1606,6 +1752,7 @@ class Program_weight_tensor_parameter_145: class Program_weight_tensor_parameter_146: name = "parameter_146" + original_name = "linear_51.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.267408") @@ -1617,6 +1764,7 @@ class Program_weight_tensor_parameter_146: class Program_weight_tensor_parameter_147: name = "parameter_147" + original_name = "linear_50.b_0" shape = [1152] dtype = "float32" min_val = float("-2.61079") @@ -1628,6 +1776,7 @@ class Program_weight_tensor_parameter_147: class Program_weight_tensor_parameter_148: name = "parameter_148" + original_name = "linear_50.w_0" shape = [384, 1152] dtype = "float32" min_val = float("-0.40697") @@ -1639,6 +1788,7 @@ class Program_weight_tensor_parameter_148: class Program_weight_tensor_parameter_149: name = "parameter_149" + original_name = "layer_norm_27.b_0" shape = [384] dtype = "float32" min_val = float("-1.42047") @@ -1650,6 +1800,7 @@ class Program_weight_tensor_parameter_149: class Program_weight_tensor_parameter_150: name = "parameter_150" + original_name = "layer_norm_27.w_0" shape = [384] dtype = "float32" min_val = float("0.317132") @@ -1661,6 +1812,7 @@ class Program_weight_tensor_parameter_150: class Program_weight_tensor_parameter_151: name = "parameter_151" + original_name = "linear_49.b_0" shape = [384] dtype = "float32" min_val = float("-0.494448") @@ -1672,6 +1824,7 @@ class Program_weight_tensor_parameter_151: class Program_weight_tensor_parameter_152: name = "parameter_152" + original_name = "linear_49.w_0" shape = [1536, 384] dtype = "float32" min_val = float("-1.28837") @@ -1683,6 +1836,7 @@ class Program_weight_tensor_parameter_152: class Program_weight_tensor_parameter_153: name = "parameter_153" + original_name = "linear_48.b_0" shape = [1536] dtype = "float32" min_val = float("-1.38971") @@ -1694,6 +1848,7 @@ class Program_weight_tensor_parameter_153: class Program_weight_tensor_parameter_154: name = "parameter_154" + original_name = "linear_48.w_0" shape = [384, 1536] dtype = "float32" min_val = float("-0.304324") @@ -1705,6 +1860,7 @@ class Program_weight_tensor_parameter_154: class Program_weight_tensor_parameter_155: name = "parameter_155" + original_name = "layer_norm_26.b_0" shape = [384] dtype = "float32" min_val = float("-0.856574") @@ -1716,6 +1872,7 @@ class Program_weight_tensor_parameter_155: class Program_weight_tensor_parameter_156: name = "parameter_156" + original_name = "layer_norm_26.w_0" shape = [384] dtype = "float32" min_val = float("0.396521") @@ -1727,6 +1884,7 @@ class Program_weight_tensor_parameter_156: class Program_weight_tensor_parameter_157: name = "parameter_157" + original_name = "linear_47.b_0" shape = [384] dtype = "float32" min_val = float("-2.20311") @@ -1738,6 +1896,7 @@ class Program_weight_tensor_parameter_157: class Program_weight_tensor_parameter_158: name = "parameter_158" + original_name = "linear_47.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.208128") @@ -1749,6 +1908,7 @@ class Program_weight_tensor_parameter_158: class Program_weight_tensor_parameter_159: name = "parameter_159" + original_name = "linear_46.b_0" shape = [1152] dtype = "float32" min_val = float("-2.07782") @@ -1760,6 +1920,7 @@ class Program_weight_tensor_parameter_159: class Program_weight_tensor_parameter_160: name = "parameter_160" + original_name = "linear_46.w_0" shape = [384, 1152] dtype = "float32" min_val = float("-0.363361") @@ -1771,6 +1932,7 @@ class Program_weight_tensor_parameter_160: class Program_weight_tensor_parameter_161: name = "parameter_161" + original_name = "layer_norm_25.b_0" shape = [384] dtype = "float32" min_val = float("-1.06754") @@ -1782,6 +1944,7 @@ class Program_weight_tensor_parameter_161: class Program_weight_tensor_parameter_162: name = "parameter_162" + original_name = "layer_norm_25.w_0" shape = [384] dtype = "float32" min_val = float("0.246312") @@ -1793,6 +1956,7 @@ class Program_weight_tensor_parameter_162: class Program_weight_tensor_parameter_163: name = "parameter_163" + original_name = "linear_45.b_0" shape = [384] dtype = "float32" min_val = float("-0.533334") @@ -1804,6 +1968,7 @@ class Program_weight_tensor_parameter_163: class Program_weight_tensor_parameter_164: name = "parameter_164" + original_name = "linear_45.w_0" shape = [1536, 384] dtype = "float32" min_val = float("-0.940089") @@ -1815,6 +1980,7 @@ class Program_weight_tensor_parameter_164: class Program_weight_tensor_parameter_165: name = "parameter_165" + original_name = "linear_44.b_0" shape = [1536] dtype = "float32" min_val = float("-1.49855") @@ -1826,6 +1992,7 @@ class Program_weight_tensor_parameter_165: class Program_weight_tensor_parameter_166: name = "parameter_166" + original_name = "linear_44.w_0" shape = [384, 1536] dtype = "float32" min_val = float("-0.319334") @@ -1837,6 +2004,7 @@ class Program_weight_tensor_parameter_166: class Program_weight_tensor_parameter_167: name = "parameter_167" + original_name = "layer_norm_24.b_0" shape = [384] dtype = "float32" min_val = float("-0.66243") @@ -1848,6 +2016,7 @@ class Program_weight_tensor_parameter_167: class Program_weight_tensor_parameter_168: name = "parameter_168" + original_name = "layer_norm_24.w_0" shape = [384] dtype = "float32" min_val = float("0.355244") @@ -1859,6 +2028,7 @@ class Program_weight_tensor_parameter_168: class Program_weight_tensor_parameter_169: name = "parameter_169" + original_name = "linear_43.b_0" shape = [384] dtype = "float32" min_val = float("-2.90232") @@ -1870,6 +2040,7 @@ class Program_weight_tensor_parameter_169: class Program_weight_tensor_parameter_170: name = "parameter_170" + original_name = "linear_43.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.214937") @@ -1881,6 +2052,7 @@ class Program_weight_tensor_parameter_170: class Program_weight_tensor_parameter_171: name = "parameter_171" + original_name = "linear_42.b_0" shape = [1152] dtype = "float32" min_val = float("-2.72652") @@ -1892,6 +2064,7 @@ class Program_weight_tensor_parameter_171: class Program_weight_tensor_parameter_172: name = "parameter_172" + original_name = "linear_42.w_0" shape = [384, 1152] dtype = "float32" min_val = float("-0.36774") @@ -1903,6 +2076,7 @@ class Program_weight_tensor_parameter_172: class Program_weight_tensor_parameter_173: name = "parameter_173" + original_name = "layer_norm_23.b_0" shape = [384] dtype = "float32" min_val = float("-0.975482") @@ -1914,6 +2088,7 @@ class Program_weight_tensor_parameter_173: class Program_weight_tensor_parameter_174: name = "parameter_174" + original_name = "layer_norm_23.w_0" shape = [384] dtype = "float32" min_val = float("0.23081") @@ -1925,6 +2100,7 @@ class Program_weight_tensor_parameter_174: class Program_weight_tensor_parameter_175: name = "parameter_175" + original_name = "linear_41.b_0" shape = [384] dtype = "float32" min_val = float("-0.51648") @@ -1936,6 +2112,7 @@ class Program_weight_tensor_parameter_175: class Program_weight_tensor_parameter_176: name = "parameter_176" + original_name = "linear_41.w_0" shape = [1536, 384] dtype = "float32" min_val = float("-0.622994") @@ -1947,6 +2124,7 @@ class Program_weight_tensor_parameter_176: class Program_weight_tensor_parameter_177: name = "parameter_177" + original_name = "linear_40.b_0" shape = [1536] dtype = "float32" min_val = float("-1.26961") @@ -1958,6 +2136,7 @@ class Program_weight_tensor_parameter_177: class Program_weight_tensor_parameter_178: name = "parameter_178" + original_name = "linear_40.w_0" shape = [384, 1536] dtype = "float32" min_val = float("-0.330742") @@ -1969,6 +2148,7 @@ class Program_weight_tensor_parameter_178: class Program_weight_tensor_parameter_179: name = "parameter_179" + original_name = "layer_norm_22.b_0" shape = [384] dtype = "float32" min_val = float("-0.758784") @@ -1980,6 +2160,7 @@ class Program_weight_tensor_parameter_179: class Program_weight_tensor_parameter_180: name = "parameter_180" + original_name = "layer_norm_22.w_0" shape = [384] dtype = "float32" min_val = float("0.364318") @@ -1991,6 +2172,7 @@ class Program_weight_tensor_parameter_180: class Program_weight_tensor_parameter_181: name = "parameter_181" + original_name = "linear_39.b_0" shape = [384] dtype = "float32" min_val = float("-2.20638") @@ -2002,6 +2184,7 @@ class Program_weight_tensor_parameter_181: class Program_weight_tensor_parameter_182: name = "parameter_182" + original_name = "linear_39.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.185258") @@ -2013,6 +2196,7 @@ class Program_weight_tensor_parameter_182: class Program_weight_tensor_parameter_183: name = "parameter_183" + original_name = "linear_38.b_0" shape = [1152] dtype = "float32" min_val = float("-2.16134") @@ -2024,6 +2208,7 @@ class Program_weight_tensor_parameter_183: class Program_weight_tensor_parameter_184: name = "parameter_184" + original_name = "linear_38.w_0" shape = [384, 1152] dtype = "float32" min_val = float("-0.520512") @@ -2035,6 +2220,7 @@ class Program_weight_tensor_parameter_184: class Program_weight_tensor_parameter_185: name = "parameter_185" + original_name = "layer_norm_21.b_0" shape = [384] dtype = "float32" min_val = float("-0.83032") @@ -2046,6 +2232,7 @@ class Program_weight_tensor_parameter_185: class Program_weight_tensor_parameter_186: name = "parameter_186" + original_name = "layer_norm_21.w_0" shape = [384] dtype = "float32" min_val = float("0.219294") @@ -2057,6 +2244,7 @@ class Program_weight_tensor_parameter_186: class Program_weight_tensor_parameter_187: name = "parameter_187" + original_name = "linear_37.b_0" shape = [384] dtype = "float32" min_val = float("-0.577857") @@ -2068,6 +2256,7 @@ class Program_weight_tensor_parameter_187: class Program_weight_tensor_parameter_188: name = "parameter_188" + original_name = "linear_37.w_0" shape = [1536, 384] dtype = "float32" min_val = float("-0.466754") @@ -2079,6 +2268,7 @@ class Program_weight_tensor_parameter_188: class Program_weight_tensor_parameter_189: name = "parameter_189" + original_name = "linear_36.b_0" shape = [1536] dtype = "float32" min_val = float("-1.19879") @@ -2090,6 +2280,7 @@ class Program_weight_tensor_parameter_189: class Program_weight_tensor_parameter_190: name = "parameter_190" + original_name = "linear_36.w_0" shape = [384, 1536] dtype = "float32" min_val = float("-0.370268") @@ -2101,6 +2292,7 @@ class Program_weight_tensor_parameter_190: class Program_weight_tensor_parameter_191: name = "parameter_191" + original_name = "layer_norm_20.b_0" shape = [384] dtype = "float32" min_val = float("-0.793246") @@ -2112,6 +2304,7 @@ class Program_weight_tensor_parameter_191: class Program_weight_tensor_parameter_192: name = "parameter_192" + original_name = "layer_norm_20.w_0" shape = [384] dtype = "float32" min_val = float("0.339409") @@ -2123,6 +2316,7 @@ class Program_weight_tensor_parameter_192: class Program_weight_tensor_parameter_193: name = "parameter_193" + original_name = "linear_35.b_0" shape = [384] dtype = "float32" min_val = float("-2.63428") @@ -2134,6 +2328,7 @@ class Program_weight_tensor_parameter_193: class Program_weight_tensor_parameter_194: name = "parameter_194" + original_name = "linear_35.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.199169") @@ -2145,6 +2340,7 @@ class Program_weight_tensor_parameter_194: class Program_weight_tensor_parameter_195: name = "parameter_195" + original_name = "linear_34.b_0" shape = [1152] dtype = "float32" min_val = float("-1.9523") @@ -2156,6 +2352,7 @@ class Program_weight_tensor_parameter_195: class Program_weight_tensor_parameter_196: name = "parameter_196" + original_name = "linear_34.w_0" shape = [384, 1152] dtype = "float32" min_val = float("-0.358211") @@ -2167,6 +2364,7 @@ class Program_weight_tensor_parameter_196: class Program_weight_tensor_parameter_197: name = "parameter_197" + original_name = "layer_norm_19.b_0" shape = [384] dtype = "float32" min_val = float("-0.904172") @@ -2178,6 +2376,7 @@ class Program_weight_tensor_parameter_197: class Program_weight_tensor_parameter_198: name = "parameter_198" + original_name = "layer_norm_19.w_0" shape = [384] dtype = "float32" min_val = float("0.126169") @@ -2189,6 +2388,7 @@ class Program_weight_tensor_parameter_198: class Program_weight_tensor_parameter_199: name = "parameter_199" + original_name = "linear_33.b_0" shape = [384] dtype = "float32" min_val = float("-1.08311") @@ -2200,6 +2400,7 @@ class Program_weight_tensor_parameter_199: class Program_weight_tensor_parameter_200: name = "parameter_200" + original_name = "linear_33.w_0" shape = [1536, 384] dtype = "float32" min_val = float("-0.322753") @@ -2211,6 +2412,7 @@ class Program_weight_tensor_parameter_200: class Program_weight_tensor_parameter_201: name = "parameter_201" + original_name = "linear_32.b_0" shape = [1536] dtype = "float32" min_val = float("-1.38701") @@ -2222,6 +2424,7 @@ class Program_weight_tensor_parameter_201: class Program_weight_tensor_parameter_202: name = "parameter_202" + original_name = "linear_32.w_0" shape = [384, 1536] dtype = "float32" min_val = float("-0.309614") @@ -2233,6 +2436,7 @@ class Program_weight_tensor_parameter_202: class Program_weight_tensor_parameter_203: name = "parameter_203" + original_name = "layer_norm_18.b_0" shape = [384] dtype = "float32" min_val = float("-0.818967") @@ -2244,6 +2448,7 @@ class Program_weight_tensor_parameter_203: class Program_weight_tensor_parameter_204: name = "parameter_204" + original_name = "layer_norm_18.w_0" shape = [384] dtype = "float32" min_val = float("0.37349") @@ -2255,6 +2460,7 @@ class Program_weight_tensor_parameter_204: class Program_weight_tensor_parameter_205: name = "parameter_205" + original_name = "linear_31.b_0" shape = [384] dtype = "float32" min_val = float("-1.84358") @@ -2266,6 +2472,7 @@ class Program_weight_tensor_parameter_205: class Program_weight_tensor_parameter_206: name = "parameter_206" + original_name = "linear_31.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.20102") @@ -2277,6 +2484,7 @@ class Program_weight_tensor_parameter_206: class Program_weight_tensor_parameter_207: name = "parameter_207" + original_name = "linear_30.b_0" shape = [1152] dtype = "float32" min_val = float("-2.64589") @@ -2288,6 +2496,7 @@ class Program_weight_tensor_parameter_207: class Program_weight_tensor_parameter_208: name = "parameter_208" + original_name = "linear_30.w_0" shape = [384, 1152] dtype = "float32" min_val = float("-0.370053") @@ -2299,6 +2508,7 @@ class Program_weight_tensor_parameter_208: class Program_weight_tensor_parameter_209: name = "parameter_209" + original_name = "layer_norm_17.b_0" shape = [384] dtype = "float32" min_val = float("-0.813461") @@ -2310,6 +2520,7 @@ class Program_weight_tensor_parameter_209: class Program_weight_tensor_parameter_210: name = "parameter_210" + original_name = "layer_norm_17.w_0" shape = [384] dtype = "float32" min_val = float("0.00376829") @@ -2321,6 +2532,7 @@ class Program_weight_tensor_parameter_210: class Program_weight_tensor_parameter_211: name = "parameter_211" + original_name = "linear_29.b_0" shape = [384] dtype = "float32" min_val = float("-1.37918") @@ -2332,6 +2544,7 @@ class Program_weight_tensor_parameter_211: class Program_weight_tensor_parameter_212: name = "parameter_212" + original_name = "linear_29.w_0" shape = [1536, 384] dtype = "float32" min_val = float("-0.277969") @@ -2343,6 +2556,7 @@ class Program_weight_tensor_parameter_212: class Program_weight_tensor_parameter_213: name = "parameter_213" + original_name = "linear_28.b_0" shape = [1536] dtype = "float32" min_val = float("-1.48976") @@ -2354,6 +2568,7 @@ class Program_weight_tensor_parameter_213: class Program_weight_tensor_parameter_214: name = "parameter_214" + original_name = "linear_28.w_0" shape = [384, 1536] dtype = "float32" min_val = float("-0.357655") @@ -2365,6 +2580,7 @@ class Program_weight_tensor_parameter_214: class Program_weight_tensor_parameter_215: name = "parameter_215" + original_name = "layer_norm_16.b_0" shape = [384] dtype = "float32" min_val = float("-0.638835") @@ -2376,6 +2592,7 @@ class Program_weight_tensor_parameter_215: class Program_weight_tensor_parameter_216: name = "parameter_216" + original_name = "layer_norm_16.w_0" shape = [384] dtype = "float32" min_val = float("0.276511") @@ -2387,6 +2604,7 @@ class Program_weight_tensor_parameter_216: class Program_weight_tensor_parameter_217: name = "parameter_217" + original_name = "linear_27.b_0" shape = [384] dtype = "float32" min_val = float("-2.01559") @@ -2398,6 +2616,7 @@ class Program_weight_tensor_parameter_217: class Program_weight_tensor_parameter_218: name = "parameter_218" + original_name = "linear_27.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.20403") @@ -2409,6 +2628,7 @@ class Program_weight_tensor_parameter_218: class Program_weight_tensor_parameter_219: name = "parameter_219" + original_name = "linear_26.b_0" shape = [1152] dtype = "float32" min_val = float("-1.92642") @@ -2420,6 +2640,7 @@ class Program_weight_tensor_parameter_219: class Program_weight_tensor_parameter_220: name = "parameter_220" + original_name = "linear_26.w_0" shape = [384, 1152] dtype = "float32" min_val = float("-0.370651") @@ -2431,6 +2652,7 @@ class Program_weight_tensor_parameter_220: class Program_weight_tensor_parameter_221: name = "parameter_221" + original_name = "layer_norm_15.b_0" shape = [384] dtype = "float32" min_val = float("-0.736367") @@ -2442,6 +2664,7 @@ class Program_weight_tensor_parameter_221: class Program_weight_tensor_parameter_222: name = "parameter_222" + original_name = "layer_norm_15.w_0" shape = [384] dtype = "float32" min_val = float("-0.00224043") @@ -2453,6 +2676,7 @@ class Program_weight_tensor_parameter_222: class Program_weight_tensor_parameter_223: name = "parameter_223" + original_name = "linear_25.b_0" shape = [384] dtype = "float32" min_val = float("-1.69229") @@ -2464,6 +2688,7 @@ class Program_weight_tensor_parameter_223: class Program_weight_tensor_parameter_224: name = "parameter_224" + original_name = "linear_25.w_0" shape = [1536, 384] dtype = "float32" min_val = float("-0.414156") @@ -2475,6 +2700,7 @@ class Program_weight_tensor_parameter_224: class Program_weight_tensor_parameter_225: name = "parameter_225" + original_name = "linear_24.b_0" shape = [1536] dtype = "float32" min_val = float("-1.74456") @@ -2486,6 +2712,7 @@ class Program_weight_tensor_parameter_225: class Program_weight_tensor_parameter_226: name = "parameter_226" + original_name = "linear_24.w_0" shape = [384, 1536] dtype = "float32" min_val = float("-0.428931") @@ -2497,6 +2724,7 @@ class Program_weight_tensor_parameter_226: class Program_weight_tensor_parameter_227: name = "parameter_227" + original_name = "layer_norm_14.b_0" shape = [384] dtype = "float32" min_val = float("-0.583404") @@ -2508,6 +2736,7 @@ class Program_weight_tensor_parameter_227: class Program_weight_tensor_parameter_228: name = "parameter_228" + original_name = "layer_norm_14.w_0" shape = [384] dtype = "float32" min_val = float("-0.00464864") @@ -2519,6 +2748,7 @@ class Program_weight_tensor_parameter_228: class Program_weight_tensor_parameter_229: name = "parameter_229" + original_name = "linear_23.b_0" shape = [384] dtype = "float32" min_val = float("-1.78814") @@ -2530,6 +2760,7 @@ class Program_weight_tensor_parameter_229: class Program_weight_tensor_parameter_230: name = "parameter_230" + original_name = "linear_23.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.226848") @@ -2541,6 +2772,7 @@ class Program_weight_tensor_parameter_230: class Program_weight_tensor_parameter_231: name = "parameter_231" + original_name = "linear_22.b_0" shape = [1152] dtype = "float32" min_val = float("-1.71662") @@ -2552,6 +2784,7 @@ class Program_weight_tensor_parameter_231: class Program_weight_tensor_parameter_232: name = "parameter_232" + original_name = "linear_22.w_0" shape = [384, 1152] dtype = "float32" min_val = float("-0.411785") @@ -2563,6 +2796,7 @@ class Program_weight_tensor_parameter_232: class Program_weight_tensor_parameter_233: name = "parameter_233" + original_name = "layer_norm_13.b_0" shape = [384] dtype = "float32" min_val = float("-0.51403") @@ -2574,6 +2808,7 @@ class Program_weight_tensor_parameter_233: class Program_weight_tensor_parameter_234: name = "parameter_234" + original_name = "layer_norm_13.w_0" shape = [384] dtype = "float32" min_val = float("-0.00327181") @@ -2585,6 +2820,7 @@ class Program_weight_tensor_parameter_234: class Program_weight_tensor_parameter_235: name = "parameter_235" + original_name = "linear_21.b_0" shape = [384] dtype = "float32" min_val = float("-2.10407") @@ -2596,6 +2832,7 @@ class Program_weight_tensor_parameter_235: class Program_weight_tensor_parameter_236: name = "parameter_236" + original_name = "linear_21.w_0" shape = [1536, 384] dtype = "float32" min_val = float("-0.340821") @@ -2607,6 +2844,7 @@ class Program_weight_tensor_parameter_236: class Program_weight_tensor_parameter_237: name = "parameter_237" + original_name = "linear_20.b_0" shape = [1536] dtype = "float32" min_val = float("-1.73513") @@ -2618,6 +2856,7 @@ class Program_weight_tensor_parameter_237: class Program_weight_tensor_parameter_238: name = "parameter_238" + original_name = "linear_20.w_0" shape = [384, 1536] dtype = "float32" min_val = float("-0.341437") @@ -2629,6 +2868,7 @@ class Program_weight_tensor_parameter_238: class Program_weight_tensor_parameter_239: name = "parameter_239" + original_name = "layer_norm_12.b_0" shape = [384] dtype = "float32" min_val = float("-1.26775") @@ -2640,6 +2880,7 @@ class Program_weight_tensor_parameter_239: class Program_weight_tensor_parameter_240: name = "parameter_240" + original_name = "layer_norm_12.w_0" shape = [384] dtype = "float32" min_val = float("-0.00322608") @@ -2651,6 +2892,7 @@ class Program_weight_tensor_parameter_240: class Program_weight_tensor_parameter_241: name = "parameter_241" + original_name = "linear_19.b_0" shape = [384] dtype = "float32" min_val = float("-2.76799") @@ -2662,6 +2904,7 @@ class Program_weight_tensor_parameter_241: class Program_weight_tensor_parameter_242: name = "parameter_242" + original_name = "linear_19.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.261978") @@ -2673,6 +2916,7 @@ class Program_weight_tensor_parameter_242: class Program_weight_tensor_parameter_243: name = "parameter_243" + original_name = "linear_18.b_0" shape = [1152] dtype = "float32" min_val = float("-2.4435") @@ -2684,6 +2928,7 @@ class Program_weight_tensor_parameter_243: class Program_weight_tensor_parameter_244: name = "parameter_244" + original_name = "linear_18.w_0" shape = [384, 1152] dtype = "float32" min_val = float("-0.386209") @@ -2695,6 +2940,7 @@ class Program_weight_tensor_parameter_244: class Program_weight_tensor_parameter_245: name = "parameter_245" + original_name = "layer_norm_11.b_0" shape = [384] dtype = "float32" min_val = float("-0.613914") @@ -2706,6 +2952,7 @@ class Program_weight_tensor_parameter_245: class Program_weight_tensor_parameter_246: name = "parameter_246" + original_name = "layer_norm_11.w_0" shape = [384] dtype = "float32" min_val = float("-0.00361083") @@ -2717,6 +2964,7 @@ class Program_weight_tensor_parameter_246: class Program_weight_tensor_parameter_247: name = "parameter_247" + original_name = "linear_17.w_0" shape = [768, 384] dtype = "float32" min_val = float("-0.215659") @@ -2728,6 +2976,7 @@ class Program_weight_tensor_parameter_247: class Program_weight_tensor_parameter_248: name = "parameter_248" + original_name = "layer_norm_10.b_0" shape = [768] dtype = "float32" min_val = float("-1.00034") @@ -2739,6 +2988,7 @@ class Program_weight_tensor_parameter_248: class Program_weight_tensor_parameter_249: name = "parameter_249" + original_name = "layer_norm_10.w_0" shape = [768] dtype = "float32" min_val = float("0.259619") @@ -2750,6 +3000,7 @@ class Program_weight_tensor_parameter_249: class Program_weight_tensor_parameter_250: name = "parameter_250" + original_name = "linear_16.b_0" shape = [192] dtype = "float32" min_val = float("-2.58342") @@ -2761,6 +3012,7 @@ class Program_weight_tensor_parameter_250: class Program_weight_tensor_parameter_251: name = "parameter_251" + original_name = "linear_16.w_0" shape = [768, 192] dtype = "float32" min_val = float("-0.581234") @@ -2772,6 +3024,7 @@ class Program_weight_tensor_parameter_251: class Program_weight_tensor_parameter_252: name = "parameter_252" + original_name = "linear_15.b_0" shape = [768] dtype = "float32" min_val = float("-1.67623") @@ -2783,6 +3036,7 @@ class Program_weight_tensor_parameter_252: class Program_weight_tensor_parameter_253: name = "parameter_253" + original_name = "linear_15.w_0" shape = [192, 768] dtype = "float32" min_val = float("-0.426344") @@ -2794,6 +3048,7 @@ class Program_weight_tensor_parameter_253: class Program_weight_tensor_parameter_254: name = "parameter_254" + original_name = "layer_norm_9.b_0" shape = [192] dtype = "float32" min_val = float("-0.853561") @@ -2805,6 +3060,7 @@ class Program_weight_tensor_parameter_254: class Program_weight_tensor_parameter_255: name = "parameter_255" + original_name = "layer_norm_9.w_0" shape = [192] dtype = "float32" min_val = float("0.283782") @@ -2816,6 +3072,7 @@ class Program_weight_tensor_parameter_255: class Program_weight_tensor_parameter_256: name = "parameter_256" + original_name = "linear_14.b_0" shape = [192] dtype = "float32" min_val = float("-1.57542") @@ -2827,6 +3084,7 @@ class Program_weight_tensor_parameter_256: class Program_weight_tensor_parameter_257: name = "parameter_257" + original_name = "linear_14.w_0" shape = [192, 192] dtype = "float32" min_val = float("-0.216351") @@ -2838,6 +3096,7 @@ class Program_weight_tensor_parameter_257: class Program_weight_tensor_parameter_258: name = "parameter_258" + original_name = "linear_13.b_0" shape = [576] dtype = "float32" min_val = float("-1.66682") @@ -2849,6 +3108,7 @@ class Program_weight_tensor_parameter_258: class Program_weight_tensor_parameter_259: name = "parameter_259" + original_name = "linear_13.w_0" shape = [192, 576] dtype = "float32" min_val = float("-0.610695") @@ -2860,6 +3120,7 @@ class Program_weight_tensor_parameter_259: class Program_weight_tensor_parameter_260: name = "parameter_260" + original_name = "layer_norm_8.b_0" shape = [192] dtype = "float32" min_val = float("-0.9126") @@ -2871,6 +3132,7 @@ class Program_weight_tensor_parameter_260: class Program_weight_tensor_parameter_261: name = "parameter_261" + original_name = "layer_norm_8.w_0" shape = [192] dtype = "float32" min_val = float("0.450615") @@ -2882,6 +3144,7 @@ class Program_weight_tensor_parameter_261: class Program_weight_tensor_parameter_262: name = "parameter_262" + original_name = "linear_12.b_0" shape = [192] dtype = "float32" min_val = float("-1.83702") @@ -2893,6 +3156,7 @@ class Program_weight_tensor_parameter_262: class Program_weight_tensor_parameter_263: name = "parameter_263" + original_name = "linear_12.w_0" shape = [768, 192] dtype = "float32" min_val = float("-0.272332") @@ -2904,6 +3168,7 @@ class Program_weight_tensor_parameter_263: class Program_weight_tensor_parameter_264: name = "parameter_264" + original_name = "linear_11.b_0" shape = [768] dtype = "float32" min_val = float("-1.66589") @@ -2915,6 +3180,7 @@ class Program_weight_tensor_parameter_264: class Program_weight_tensor_parameter_265: name = "parameter_265" + original_name = "linear_11.w_0" shape = [192, 768] dtype = "float32" min_val = float("-0.463241") @@ -2926,6 +3192,7 @@ class Program_weight_tensor_parameter_265: class Program_weight_tensor_parameter_266: name = "parameter_266" + original_name = "layer_norm_7.b_0" shape = [192] dtype = "float32" min_val = float("-0.375106") @@ -2937,6 +3204,7 @@ class Program_weight_tensor_parameter_266: class Program_weight_tensor_parameter_267: name = "parameter_267" + original_name = "layer_norm_7.w_0" shape = [192] dtype = "float32" min_val = float("0.00794437") @@ -2948,6 +3216,7 @@ class Program_weight_tensor_parameter_267: class Program_weight_tensor_parameter_268: name = "parameter_268" + original_name = "linear_10.b_0" shape = [192] dtype = "float32" min_val = float("-1.55291") @@ -2959,6 +3228,7 @@ class Program_weight_tensor_parameter_268: class Program_weight_tensor_parameter_269: name = "parameter_269" + original_name = "linear_10.w_0" shape = [192, 192] dtype = "float32" min_val = float("-0.203105") @@ -2970,6 +3240,7 @@ class Program_weight_tensor_parameter_269: class Program_weight_tensor_parameter_270: name = "parameter_270" + original_name = "linear_9.b_0" shape = [576] dtype = "float32" min_val = float("-1.68921") @@ -2981,6 +3252,7 @@ class Program_weight_tensor_parameter_270: class Program_weight_tensor_parameter_271: name = "parameter_271" + original_name = "linear_9.w_0" shape = [192, 576] dtype = "float32" min_val = float("-0.406492") @@ -2992,6 +3264,7 @@ class Program_weight_tensor_parameter_271: class Program_weight_tensor_parameter_272: name = "parameter_272" + original_name = "layer_norm_6.b_0" shape = [192] dtype = "float32" min_val = float("-0.36987") @@ -3003,6 +3276,7 @@ class Program_weight_tensor_parameter_272: class Program_weight_tensor_parameter_273: name = "parameter_273" + original_name = "layer_norm_6.w_0" shape = [192] dtype = "float32" min_val = float("0.270099") @@ -3014,6 +3288,7 @@ class Program_weight_tensor_parameter_273: class Program_weight_tensor_parameter_274: name = "parameter_274" + original_name = "linear_8.w_0" shape = [384, 192] dtype = "float32" min_val = float("-0.223585") @@ -3025,6 +3300,7 @@ class Program_weight_tensor_parameter_274: class Program_weight_tensor_parameter_275: name = "parameter_275" + original_name = "layer_norm_5.b_0" shape = [384] dtype = "float32" min_val = float("-1.56708") @@ -3036,6 +3312,7 @@ class Program_weight_tensor_parameter_275: class Program_weight_tensor_parameter_276: name = "parameter_276" + original_name = "layer_norm_5.w_0" shape = [384] dtype = "float32" min_val = float("0.313979") @@ -3047,6 +3324,7 @@ class Program_weight_tensor_parameter_276: class Program_weight_tensor_parameter_277: name = "parameter_277" + original_name = "linear_7.b_0" shape = [96] dtype = "float32" min_val = float("-1.06347") @@ -3058,6 +3336,7 @@ class Program_weight_tensor_parameter_277: class Program_weight_tensor_parameter_278: name = "parameter_278" + original_name = "linear_7.w_0" shape = [384, 96] dtype = "float32" min_val = float("-0.377621") @@ -3069,6 +3348,7 @@ class Program_weight_tensor_parameter_278: class Program_weight_tensor_parameter_279: name = "parameter_279" + original_name = "linear_6.b_0" shape = [384] dtype = "float32" min_val = float("-1.52681") @@ -3080,6 +3360,7 @@ class Program_weight_tensor_parameter_279: class Program_weight_tensor_parameter_280: name = "parameter_280" + original_name = "linear_6.w_0" shape = [96, 384] dtype = "float32" min_val = float("-0.402032") @@ -3091,6 +3372,7 @@ class Program_weight_tensor_parameter_280: class Program_weight_tensor_parameter_281: name = "parameter_281" + original_name = "layer_norm_4.b_0" shape = [96] dtype = "float32" min_val = float("-0.614218") @@ -3102,6 +3384,7 @@ class Program_weight_tensor_parameter_281: class Program_weight_tensor_parameter_282: name = "parameter_282" + original_name = "layer_norm_4.w_0" shape = [96] dtype = "float32" min_val = float("0.355034") @@ -3113,6 +3396,7 @@ class Program_weight_tensor_parameter_282: class Program_weight_tensor_parameter_283: name = "parameter_283" + original_name = "linear_5.b_0" shape = [96] dtype = "float32" min_val = float("-0.435478") @@ -3124,6 +3408,7 @@ class Program_weight_tensor_parameter_283: class Program_weight_tensor_parameter_284: name = "parameter_284" + original_name = "linear_5.w_0" shape = [96, 96] dtype = "float32" min_val = float("-0.225737") @@ -3135,6 +3420,7 @@ class Program_weight_tensor_parameter_284: class Program_weight_tensor_parameter_285: name = "parameter_285" + original_name = "linear_4.b_0" shape = [288] dtype = "float32" min_val = float("-1.52394") @@ -3146,6 +3432,7 @@ class Program_weight_tensor_parameter_285: class Program_weight_tensor_parameter_286: name = "parameter_286" + original_name = "linear_4.w_0" shape = [96, 288] dtype = "float32" min_val = float("-0.398989") @@ -3157,6 +3444,7 @@ class Program_weight_tensor_parameter_286: class Program_weight_tensor_parameter_287: name = "parameter_287" + original_name = "layer_norm_3.b_0" shape = [96] dtype = "float32" min_val = float("-1.33408") @@ -3168,6 +3456,7 @@ class Program_weight_tensor_parameter_287: class Program_weight_tensor_parameter_288: name = "parameter_288" + original_name = "layer_norm_3.w_0" shape = [96] dtype = "float32" min_val = float("0.252539") @@ -3179,6 +3468,7 @@ class Program_weight_tensor_parameter_288: class Program_weight_tensor_parameter_289: name = "parameter_289" + original_name = "linear_3.b_0" shape = [96] dtype = "float32" min_val = float("-1.45123") @@ -3190,6 +3480,7 @@ class Program_weight_tensor_parameter_289: class Program_weight_tensor_parameter_290: name = "parameter_290" + original_name = "linear_3.w_0" shape = [384, 96] dtype = "float32" min_val = float("-0.336166") @@ -3201,6 +3492,7 @@ class Program_weight_tensor_parameter_290: class Program_weight_tensor_parameter_291: name = "parameter_291" + original_name = "linear_2.b_0" shape = [384] dtype = "float32" min_val = float("-1.62895") @@ -3212,6 +3504,7 @@ class Program_weight_tensor_parameter_291: class Program_weight_tensor_parameter_292: name = "parameter_292" + original_name = "linear_2.w_0" shape = [96, 384] dtype = "float32" min_val = float("-0.532516") @@ -3223,6 +3516,7 @@ class Program_weight_tensor_parameter_292: class Program_weight_tensor_parameter_293: name = "parameter_293" + original_name = "layer_norm_2.b_0" shape = [96] dtype = "float32" min_val = float("-1.0797") @@ -3234,6 +3528,7 @@ class Program_weight_tensor_parameter_293: class Program_weight_tensor_parameter_294: name = "parameter_294" + original_name = "layer_norm_2.w_0" shape = [96] dtype = "float32" min_val = float("0.670134") @@ -3245,6 +3540,7 @@ class Program_weight_tensor_parameter_294: class Program_weight_tensor_parameter_295: name = "parameter_295" + original_name = "linear_1.b_0" shape = [96] dtype = "float32" min_val = float("-0.792595") @@ -3256,6 +3552,7 @@ class Program_weight_tensor_parameter_295: class Program_weight_tensor_parameter_296: name = "parameter_296" + original_name = "linear_1.w_0" shape = [96, 96] dtype = "float32" min_val = float("-0.241791") @@ -3267,6 +3564,7 @@ class Program_weight_tensor_parameter_296: class Program_weight_tensor_parameter_297: name = "parameter_297" + original_name = "linear_0.b_0" shape = [288] dtype = "float32" min_val = float("-1.56109") @@ -3278,6 +3576,7 @@ class Program_weight_tensor_parameter_297: class Program_weight_tensor_parameter_298: name = "parameter_298" + original_name = "linear_0.w_0" shape = [96, 288] dtype = "float32" min_val = float("-0.759299") @@ -3289,6 +3588,7 @@ class Program_weight_tensor_parameter_298: class Program_weight_tensor_parameter_299: name = "parameter_299" + original_name = "layer_norm_1.b_0" shape = [96] dtype = "float32" min_val = float("-0.811588") @@ -3300,6 +3600,7 @@ class Program_weight_tensor_parameter_299: class Program_weight_tensor_parameter_300: name = "parameter_300" + original_name = "layer_norm_1.w_0" shape = [96] dtype = "float32" min_val = float("-0.0061428") @@ -3311,6 +3612,7 @@ class Program_weight_tensor_parameter_300: class Program_weight_tensor_parameter_301: name = "parameter_301" + original_name = "layer_norm_0.b_0" shape = [96] dtype = "float32" min_val = float("-1.47602") @@ -3322,6 +3624,7 @@ class Program_weight_tensor_parameter_301: class Program_weight_tensor_parameter_302: name = "parameter_302" + original_name = "layer_norm_0.w_0" shape = [96] dtype = "float32" min_val = float("-0.00987474") @@ -3333,6 +3636,7 @@ class Program_weight_tensor_parameter_302: class Program_weight_tensor_parameter_303: name = "parameter_303" + original_name = "conv2d_0.b_0" shape = [96] dtype = "float32" min_val = float("-1.35514") @@ -3344,6 +3648,7 @@ class Program_weight_tensor_parameter_303: class Program_weight_tensor_parameter_304: name = "parameter_304" + original_name = "conv2d_0.w_0" shape = [96, 3, 4, 4] dtype = "float32" min_val = float("-0.294284") diff --git a/paddle_samples/PaddleX/SwinTransformer_tiny_patch4_window7_224/subgraph_1/input_meta.py b/paddle_samples/PaddleX/SwinTransformer_tiny_patch4_window7_224/subgraph_1/input_meta.py index f0dd0a8be..6d859c2ae 100644 --- a/paddle_samples/PaddleX/SwinTransformer_tiny_patch4_window7_224/subgraph_1/input_meta.py +++ b/paddle_samples/PaddleX/SwinTransformer_tiny_patch4_window7_224/subgraph_1/input_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_data_0: name = "data_0" + original_name = "param_0" shape = [169, 3] dtype = "float32" min_val = float("-10.7112") @@ -11,6 +12,7 @@ class Program_weight_tensor_data_0: class Program_weight_tensor_data_1: name = "data_1" + original_name = "param_1" shape = [169, 3] dtype = "float32" min_val = float("-8.27699") @@ -22,6 +24,7 @@ class Program_weight_tensor_data_1: class Program_weight_tensor_data_2: name = "data_2" + original_name = "param_10" shape = [169, 24] dtype = "float32" min_val = float("-15.1434") @@ -33,6 +36,7 @@ class Program_weight_tensor_data_2: class Program_weight_tensor_data_3: name = "data_3" + original_name = "param_11" shape = [169, 24] dtype = "float32" min_val = float("-15.5121") @@ -44,6 +48,7 @@ class Program_weight_tensor_data_3: class Program_weight_tensor_data_4: name = "data_4" + original_name = "param_2" shape = [169, 6] dtype = "float32" min_val = float("-6.35951") @@ -55,6 +60,7 @@ class Program_weight_tensor_data_4: class Program_weight_tensor_data_5: name = "data_5" + original_name = "param_3" shape = [169, 6] dtype = "float32" min_val = float("-9.1058") @@ -66,6 +72,7 @@ class Program_weight_tensor_data_5: class Program_weight_tensor_data_6: name = "data_6" + original_name = "param_4" shape = [169, 12] dtype = "float32" min_val = float("-10.2288") @@ -77,6 +84,7 @@ class Program_weight_tensor_data_6: class Program_weight_tensor_data_7: name = "data_7" + original_name = "param_5" shape = [169, 12] dtype = "float32" min_val = float("-10.5655") @@ -88,6 +96,7 @@ class Program_weight_tensor_data_7: class Program_weight_tensor_data_8: name = "data_8" + original_name = "param_6" shape = [169, 12] dtype = "float32" min_val = float("-11.5152") @@ -99,6 +108,7 @@ class Program_weight_tensor_data_8: class Program_weight_tensor_data_9: name = "data_9" + original_name = "param_7" shape = [169, 12] dtype = "float32" min_val = float("-7.50729") @@ -110,6 +120,7 @@ class Program_weight_tensor_data_9: class Program_weight_tensor_data_10: name = "data_10" + original_name = "param_8" shape = [169, 12] dtype = "float32" min_val = float("-10.5941") @@ -121,6 +132,7 @@ class Program_weight_tensor_data_10: class Program_weight_tensor_data_11: name = "data_11" + original_name = "param_9" shape = [169, 12] dtype = "float32" min_val = float("-13.1412") @@ -132,6 +144,7 @@ class Program_weight_tensor_data_11: class Program_weight_tensor_data_12: name = "data_12" + original_name = "var_0" shape = [64, 3, 224, 224] dtype = "float32" min_val = float("-3.87123") @@ -143,6 +156,7 @@ class Program_weight_tensor_data_12: class Program_weight_tensor_data_13: name = "data_13" + original_name = "var_159" shape = [49, 49] dtype = "int64" min_val = 0 @@ -152,6 +166,7 @@ class Program_weight_tensor_data_13: class Program_weight_tensor_data_14: name = "data_14" + original_name = "var_21" shape = [49, 49] dtype = "int64" min_val = 0 @@ -161,6 +176,7 @@ class Program_weight_tensor_data_14: class Program_weight_tensor_data_15: name = "data_15" + original_name = "var_239" shape = [49, 49] dtype = "int64" min_val = 0 @@ -170,6 +186,7 @@ class Program_weight_tensor_data_15: class Program_weight_tensor_data_16: name = "data_16" + original_name = "var_311" shape = [49, 49] dtype = "int64" min_val = 0 @@ -179,6 +196,7 @@ class Program_weight_tensor_data_16: class Program_weight_tensor_data_17: name = "data_17" + original_name = "var_391" shape = [49, 49] dtype = "int64" min_val = 0 @@ -188,6 +206,7 @@ class Program_weight_tensor_data_17: class Program_weight_tensor_data_18: name = "data_18" + original_name = "var_453" shape = [49, 49] dtype = "int64" min_val = 0 @@ -197,6 +216,7 @@ class Program_weight_tensor_data_18: class Program_weight_tensor_data_19: name = "data_19" + original_name = "var_533" shape = [49, 49] dtype = "int64" min_val = 0 @@ -206,6 +226,7 @@ class Program_weight_tensor_data_19: class Program_weight_tensor_data_20: name = "data_20" + original_name = "var_595" shape = [49, 49] dtype = "int64" min_val = 0 @@ -215,6 +236,7 @@ class Program_weight_tensor_data_20: class Program_weight_tensor_data_21: name = "data_21" + original_name = "var_675" shape = [49, 49] dtype = "int64" min_val = 0 @@ -224,6 +246,7 @@ class Program_weight_tensor_data_21: class Program_weight_tensor_data_22: name = "data_22" + original_name = "var_747" shape = [49, 49] dtype = "int64" min_val = 0 @@ -233,6 +256,7 @@ class Program_weight_tensor_data_22: class Program_weight_tensor_data_23: name = "data_23" + original_name = "var_827" shape = [49, 49] dtype = "int64" min_val = 0 @@ -242,6 +266,7 @@ class Program_weight_tensor_data_23: class Program_weight_tensor_data_24: name = "data_24" + original_name = "var_87" shape = [49, 49] dtype = "int64" min_val = 0 diff --git a/paddle_samples/PaddleX/SwinTransformer_tiny_patch4_window7_224/subgraph_1/weight_meta.py b/paddle_samples/PaddleX/SwinTransformer_tiny_patch4_window7_224/subgraph_1/weight_meta.py index e1ed88af7..5c36f1fd5 100644 --- a/paddle_samples/PaddleX/SwinTransformer_tiny_patch4_window7_224/subgraph_1/weight_meta.py +++ b/paddle_samples/PaddleX/SwinTransformer_tiny_patch4_window7_224/subgraph_1/weight_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_parameter_0: name = "parameter_0" + original_name = "linear_51.b_0" shape = [102] dtype = "float32" data = None @@ -7,6 +8,7 @@ class Program_weight_tensor_parameter_0: class Program_weight_tensor_parameter_1: name = "parameter_1" + original_name = "linear_51.w_0" shape = [768, 102] dtype = "float32" min_val = float("-0.0795314") @@ -18,6 +20,7 @@ class Program_weight_tensor_parameter_1: class Program_weight_tensor_parameter_2: name = "parameter_2" + original_name = "layer_norm_28.b_0" shape = [768] dtype = "float32" min_val = float("-0.860357") @@ -29,6 +32,7 @@ class Program_weight_tensor_parameter_2: class Program_weight_tensor_parameter_3: name = "parameter_3" + original_name = "layer_norm_28.w_0" shape = [768] dtype = "float32" min_val = float("-0.000382801") @@ -40,6 +44,7 @@ class Program_weight_tensor_parameter_3: class Program_weight_tensor_parameter_4: name = "parameter_4" + original_name = "linear_50.b_0" shape = [768] dtype = "float32" min_val = float("-7.72635") @@ -51,6 +56,7 @@ class Program_weight_tensor_parameter_4: class Program_weight_tensor_parameter_5: name = "parameter_5" + original_name = "linear_50.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-1.78251") @@ -62,6 +68,7 @@ class Program_weight_tensor_parameter_5: class Program_weight_tensor_parameter_6: name = "parameter_6" + original_name = "linear_49.b_0" shape = [3072] dtype = "float32" min_val = float("-1.61974") @@ -73,6 +80,7 @@ class Program_weight_tensor_parameter_6: class Program_weight_tensor_parameter_7: name = "parameter_7" + original_name = "linear_49.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.470737") @@ -84,6 +92,7 @@ class Program_weight_tensor_parameter_7: class Program_weight_tensor_parameter_8: name = "parameter_8" + original_name = "layer_norm_27.b_0" shape = [768] dtype = "float32" min_val = float("-1.98279") @@ -95,6 +104,7 @@ class Program_weight_tensor_parameter_8: class Program_weight_tensor_parameter_9: name = "parameter_9" + original_name = "layer_norm_27.w_0" shape = [768] dtype = "float32" min_val = float("0.00201348") @@ -106,6 +116,7 @@ class Program_weight_tensor_parameter_9: class Program_weight_tensor_parameter_10: name = "parameter_10" + original_name = "linear_48.b_0" shape = [768] dtype = "float32" min_val = float("-11.8361") @@ -117,6 +128,7 @@ class Program_weight_tensor_parameter_10: class Program_weight_tensor_parameter_11: name = "parameter_11" + original_name = "linear_48.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.91964") @@ -128,6 +140,7 @@ class Program_weight_tensor_parameter_11: class Program_weight_tensor_parameter_12: name = "parameter_12" + original_name = "linear_47.b_0" shape = [2304] dtype = "float32" min_val = float("-2.85719") @@ -139,6 +152,7 @@ class Program_weight_tensor_parameter_12: class Program_weight_tensor_parameter_13: name = "parameter_13" + original_name = "linear_47.w_0" shape = [768, 2304] dtype = "float32" min_val = float("-0.418896") @@ -150,6 +164,7 @@ class Program_weight_tensor_parameter_13: class Program_weight_tensor_parameter_14: name = "parameter_14" + original_name = "layer_norm_26.b_0" shape = [768] dtype = "float32" min_val = float("-2.96136") @@ -161,6 +176,7 @@ class Program_weight_tensor_parameter_14: class Program_weight_tensor_parameter_15: name = "parameter_15" + original_name = "layer_norm_26.w_0" shape = [768] dtype = "float32" min_val = float("0.0248893") @@ -172,6 +188,7 @@ class Program_weight_tensor_parameter_15: class Program_weight_tensor_parameter_16: name = "parameter_16" + original_name = "linear_46.b_0" shape = [768] dtype = "float32" min_val = float("-2.18999") @@ -183,6 +200,7 @@ class Program_weight_tensor_parameter_16: class Program_weight_tensor_parameter_17: name = "parameter_17" + original_name = "linear_46.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-1.70459") @@ -194,6 +212,7 @@ class Program_weight_tensor_parameter_17: class Program_weight_tensor_parameter_18: name = "parameter_18" + original_name = "linear_45.b_0" shape = [3072] dtype = "float32" min_val = float("-1.07362") @@ -205,6 +224,7 @@ class Program_weight_tensor_parameter_18: class Program_weight_tensor_parameter_19: name = "parameter_19" + original_name = "linear_45.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.539595") @@ -216,6 +236,7 @@ class Program_weight_tensor_parameter_19: class Program_weight_tensor_parameter_20: name = "parameter_20" + original_name = "layer_norm_25.b_0" shape = [768] dtype = "float32" min_val = float("-3.41985") @@ -227,6 +248,7 @@ class Program_weight_tensor_parameter_20: class Program_weight_tensor_parameter_21: name = "parameter_21" + original_name = "layer_norm_25.w_0" shape = [768] dtype = "float32" min_val = float("-0.32046") @@ -238,6 +260,7 @@ class Program_weight_tensor_parameter_21: class Program_weight_tensor_parameter_22: name = "parameter_22" + original_name = "linear_44.b_0" shape = [768] dtype = "float32" min_val = float("-4.92739") @@ -249,6 +272,7 @@ class Program_weight_tensor_parameter_22: class Program_weight_tensor_parameter_23: name = "parameter_23" + original_name = "linear_44.w_0" shape = [768, 768] dtype = "float32" min_val = float("-1.07593") @@ -260,6 +284,7 @@ class Program_weight_tensor_parameter_23: class Program_weight_tensor_parameter_24: name = "parameter_24" + original_name = "linear_43.b_0" shape = [2304] dtype = "float32" min_val = float("-2.82658") @@ -271,6 +296,7 @@ class Program_weight_tensor_parameter_24: class Program_weight_tensor_parameter_25: name = "parameter_25" + original_name = "linear_43.w_0" shape = [768, 2304] dtype = "float32" min_val = float("-0.26931") @@ -282,6 +308,7 @@ class Program_weight_tensor_parameter_25: class Program_weight_tensor_parameter_26: name = "parameter_26" + original_name = "layer_norm_24.b_0" shape = [768] dtype = "float32" min_val = float("-1.83303") @@ -293,6 +320,7 @@ class Program_weight_tensor_parameter_26: class Program_weight_tensor_parameter_27: name = "parameter_27" + original_name = "layer_norm_24.w_0" shape = [768] dtype = "float32" min_val = float("-0.000242999") @@ -304,6 +332,7 @@ class Program_weight_tensor_parameter_27: class Program_weight_tensor_parameter_28: name = "parameter_28" + original_name = "linear_42.w_0" shape = [1536, 768] dtype = "float32" min_val = float("-0.84658") @@ -315,6 +344,7 @@ class Program_weight_tensor_parameter_28: class Program_weight_tensor_parameter_29: name = "parameter_29" + original_name = "layer_norm_23.b_0" shape = [1536] dtype = "float32" min_val = float("-0.655395") @@ -326,6 +356,7 @@ class Program_weight_tensor_parameter_29: class Program_weight_tensor_parameter_30: name = "parameter_30" + original_name = "layer_norm_23.w_0" shape = [1536] dtype = "float32" min_val = float("0.0579618") @@ -337,6 +368,7 @@ class Program_weight_tensor_parameter_30: class Program_weight_tensor_parameter_31: name = "parameter_31" + original_name = "linear_41.b_0" shape = [384] dtype = "float32" min_val = float("-0.771314") @@ -348,6 +380,7 @@ class Program_weight_tensor_parameter_31: class Program_weight_tensor_parameter_32: name = "parameter_32" + original_name = "linear_41.w_0" shape = [1536, 384] dtype = "float32" min_val = float("-1.54683") @@ -359,6 +392,7 @@ class Program_weight_tensor_parameter_32: class Program_weight_tensor_parameter_33: name = "parameter_33" + original_name = "linear_40.b_0" shape = [1536] dtype = "float32" min_val = float("-3.79974") @@ -370,6 +404,7 @@ class Program_weight_tensor_parameter_33: class Program_weight_tensor_parameter_34: name = "parameter_34" + original_name = "linear_40.w_0" shape = [384, 1536] dtype = "float32" min_val = float("-0.767617") @@ -381,6 +416,7 @@ class Program_weight_tensor_parameter_34: class Program_weight_tensor_parameter_35: name = "parameter_35" + original_name = "layer_norm_22.b_0" shape = [384] dtype = "float32" min_val = float("-2.23482") @@ -392,6 +428,7 @@ class Program_weight_tensor_parameter_35: class Program_weight_tensor_parameter_36: name = "parameter_36" + original_name = "layer_norm_22.w_0" shape = [384] dtype = "float32" min_val = float("0.25017") @@ -403,6 +440,7 @@ class Program_weight_tensor_parameter_36: class Program_weight_tensor_parameter_37: name = "parameter_37" + original_name = "linear_39.b_0" shape = [384] dtype = "float32" min_val = float("-0.870492") @@ -414,6 +452,7 @@ class Program_weight_tensor_parameter_37: class Program_weight_tensor_parameter_38: name = "parameter_38" + original_name = "linear_39.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.48025") @@ -425,6 +464,7 @@ class Program_weight_tensor_parameter_38: class Program_weight_tensor_parameter_39: name = "parameter_39" + original_name = "linear_38.b_0" shape = [1152] dtype = "float32" min_val = float("-1.84331") @@ -436,6 +476,7 @@ class Program_weight_tensor_parameter_39: class Program_weight_tensor_parameter_40: name = "parameter_40" + original_name = "linear_38.w_0" shape = [384, 1152] dtype = "float32" min_val = float("-0.364836") @@ -447,6 +488,7 @@ class Program_weight_tensor_parameter_40: class Program_weight_tensor_parameter_41: name = "parameter_41" + original_name = "layer_norm_21.b_0" shape = [384] dtype = "float32" min_val = float("-1.96051") @@ -458,6 +500,7 @@ class Program_weight_tensor_parameter_41: class Program_weight_tensor_parameter_42: name = "parameter_42" + original_name = "layer_norm_21.w_0" shape = [384] dtype = "float32" min_val = float("0.224939") @@ -469,6 +512,7 @@ class Program_weight_tensor_parameter_42: class Program_weight_tensor_parameter_43: name = "parameter_43" + original_name = "linear_37.b_0" shape = [384] dtype = "float32" min_val = float("-0.79635") @@ -480,6 +524,7 @@ class Program_weight_tensor_parameter_43: class Program_weight_tensor_parameter_44: name = "parameter_44" + original_name = "linear_37.w_0" shape = [1536, 384] dtype = "float32" min_val = float("-2.91672") @@ -491,6 +536,7 @@ class Program_weight_tensor_parameter_44: class Program_weight_tensor_parameter_45: name = "parameter_45" + original_name = "linear_36.b_0" shape = [1536] dtype = "float32" min_val = float("-2.76259") @@ -502,6 +548,7 @@ class Program_weight_tensor_parameter_45: class Program_weight_tensor_parameter_46: name = "parameter_46" + original_name = "linear_36.w_0" shape = [384, 1536] dtype = "float32" min_val = float("-0.520382") @@ -513,6 +560,7 @@ class Program_weight_tensor_parameter_46: class Program_weight_tensor_parameter_47: name = "parameter_47" + original_name = "layer_norm_20.b_0" shape = [384] dtype = "float32" min_val = float("-1.83383") @@ -524,6 +572,7 @@ class Program_weight_tensor_parameter_47: class Program_weight_tensor_parameter_48: name = "parameter_48" + original_name = "layer_norm_20.w_0" shape = [384] dtype = "float32" min_val = float("0.238976") @@ -535,6 +584,7 @@ class Program_weight_tensor_parameter_48: class Program_weight_tensor_parameter_49: name = "parameter_49" + original_name = "linear_35.b_0" shape = [384] dtype = "float32" min_val = float("-1.03139") @@ -546,6 +596,7 @@ class Program_weight_tensor_parameter_49: class Program_weight_tensor_parameter_50: name = "parameter_50" + original_name = "linear_35.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.405709") @@ -557,6 +608,7 @@ class Program_weight_tensor_parameter_50: class Program_weight_tensor_parameter_51: name = "parameter_51" + original_name = "linear_34.b_0" shape = [1152] dtype = "float32" min_val = float("-2.86243") @@ -568,6 +620,7 @@ class Program_weight_tensor_parameter_51: class Program_weight_tensor_parameter_52: name = "parameter_52" + original_name = "linear_34.w_0" shape = [384, 1152] dtype = "float32" min_val = float("-0.359465") @@ -579,6 +632,7 @@ class Program_weight_tensor_parameter_52: class Program_weight_tensor_parameter_53: name = "parameter_53" + original_name = "layer_norm_19.b_0" shape = [384] dtype = "float32" min_val = float("-1.56537") @@ -590,6 +644,7 @@ class Program_weight_tensor_parameter_53: class Program_weight_tensor_parameter_54: name = "parameter_54" + original_name = "layer_norm_19.w_0" shape = [384] dtype = "float32" min_val = float("0.254116") @@ -601,6 +656,7 @@ class Program_weight_tensor_parameter_54: class Program_weight_tensor_parameter_55: name = "parameter_55" + original_name = "linear_33.b_0" shape = [384] dtype = "float32" min_val = float("-0.586321") @@ -612,6 +668,7 @@ class Program_weight_tensor_parameter_55: class Program_weight_tensor_parameter_56: name = "parameter_56" + original_name = "linear_33.w_0" shape = [1536, 384] dtype = "float32" min_val = float("-2.38849") @@ -623,6 +680,7 @@ class Program_weight_tensor_parameter_56: class Program_weight_tensor_parameter_57: name = "parameter_57" + original_name = "linear_32.b_0" shape = [1536] dtype = "float32" min_val = float("-2.04056") @@ -634,6 +692,7 @@ class Program_weight_tensor_parameter_57: class Program_weight_tensor_parameter_58: name = "parameter_58" + original_name = "linear_32.w_0" shape = [384, 1536] dtype = "float32" min_val = float("-0.268319") @@ -645,6 +704,7 @@ class Program_weight_tensor_parameter_58: class Program_weight_tensor_parameter_59: name = "parameter_59" + original_name = "layer_norm_18.b_0" shape = [384] dtype = "float32" min_val = float("-1.62598") @@ -656,6 +716,7 @@ class Program_weight_tensor_parameter_59: class Program_weight_tensor_parameter_60: name = "parameter_60" + original_name = "layer_norm_18.w_0" shape = [384] dtype = "float32" min_val = float("0.250654") @@ -667,6 +728,7 @@ class Program_weight_tensor_parameter_60: class Program_weight_tensor_parameter_61: name = "parameter_61" + original_name = "linear_31.b_0" shape = [384] dtype = "float32" min_val = float("-0.547021") @@ -678,6 +740,7 @@ class Program_weight_tensor_parameter_61: class Program_weight_tensor_parameter_62: name = "parameter_62" + original_name = "linear_31.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.279195") @@ -689,6 +752,7 @@ class Program_weight_tensor_parameter_62: class Program_weight_tensor_parameter_63: name = "parameter_63" + original_name = "linear_30.b_0" shape = [1152] dtype = "float32" min_val = float("-1.96776") @@ -700,6 +764,7 @@ class Program_weight_tensor_parameter_63: class Program_weight_tensor_parameter_64: name = "parameter_64" + original_name = "linear_30.w_0" shape = [384, 1152] dtype = "float32" min_val = float("-0.46368") @@ -711,6 +776,7 @@ class Program_weight_tensor_parameter_64: class Program_weight_tensor_parameter_65: name = "parameter_65" + original_name = "layer_norm_17.b_0" shape = [384] dtype = "float32" min_val = float("-1.45836") @@ -722,6 +788,7 @@ class Program_weight_tensor_parameter_65: class Program_weight_tensor_parameter_66: name = "parameter_66" + original_name = "layer_norm_17.w_0" shape = [384] dtype = "float32" min_val = float("0.291836") @@ -733,6 +800,7 @@ class Program_weight_tensor_parameter_66: class Program_weight_tensor_parameter_67: name = "parameter_67" + original_name = "linear_29.b_0" shape = [384] dtype = "float32" min_val = float("-0.479116") @@ -744,6 +812,7 @@ class Program_weight_tensor_parameter_67: class Program_weight_tensor_parameter_68: name = "parameter_68" + original_name = "linear_29.w_0" shape = [1536, 384] dtype = "float32" min_val = float("-1.45846") @@ -755,6 +824,7 @@ class Program_weight_tensor_parameter_68: class Program_weight_tensor_parameter_69: name = "parameter_69" + original_name = "linear_28.b_0" shape = [1536] dtype = "float32" min_val = float("-1.45225") @@ -766,6 +836,7 @@ class Program_weight_tensor_parameter_69: class Program_weight_tensor_parameter_70: name = "parameter_70" + original_name = "linear_28.w_0" shape = [384, 1536] dtype = "float32" min_val = float("-0.276098") @@ -777,6 +848,7 @@ class Program_weight_tensor_parameter_70: class Program_weight_tensor_parameter_71: name = "parameter_71" + original_name = "layer_norm_16.b_0" shape = [384] dtype = "float32" min_val = float("-1.70773") @@ -788,6 +860,7 @@ class Program_weight_tensor_parameter_71: class Program_weight_tensor_parameter_72: name = "parameter_72" + original_name = "layer_norm_16.w_0" shape = [384] dtype = "float32" min_val = float("0.282538") @@ -799,6 +872,7 @@ class Program_weight_tensor_parameter_72: class Program_weight_tensor_parameter_73: name = "parameter_73" + original_name = "linear_27.b_0" shape = [384] dtype = "float32" min_val = float("-0.892584") @@ -810,6 +884,7 @@ class Program_weight_tensor_parameter_73: class Program_weight_tensor_parameter_74: name = "parameter_74" + original_name = "linear_27.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.223531") @@ -821,6 +896,7 @@ class Program_weight_tensor_parameter_74: class Program_weight_tensor_parameter_75: name = "parameter_75" + original_name = "linear_26.b_0" shape = [1152] dtype = "float32" min_val = float("-2.01052") @@ -832,6 +908,7 @@ class Program_weight_tensor_parameter_75: class Program_weight_tensor_parameter_76: name = "parameter_76" + original_name = "linear_26.w_0" shape = [384, 1152] dtype = "float32" min_val = float("-0.448833") @@ -843,6 +920,7 @@ class Program_weight_tensor_parameter_76: class Program_weight_tensor_parameter_77: name = "parameter_77" + original_name = "layer_norm_15.b_0" shape = [384] dtype = "float32" min_val = float("-1.17555") @@ -854,6 +932,7 @@ class Program_weight_tensor_parameter_77: class Program_weight_tensor_parameter_78: name = "parameter_78" + original_name = "layer_norm_15.w_0" shape = [384] dtype = "float32" min_val = float("0.298143") @@ -865,6 +944,7 @@ class Program_weight_tensor_parameter_78: class Program_weight_tensor_parameter_79: name = "parameter_79" + original_name = "linear_25.b_0" shape = [384] dtype = "float32" min_val = float("-0.526826") @@ -876,6 +956,7 @@ class Program_weight_tensor_parameter_79: class Program_weight_tensor_parameter_80: name = "parameter_80" + original_name = "linear_25.w_0" shape = [1536, 384] dtype = "float32" min_val = float("-0.722768") @@ -887,6 +968,7 @@ class Program_weight_tensor_parameter_80: class Program_weight_tensor_parameter_81: name = "parameter_81" + original_name = "linear_24.b_0" shape = [1536] dtype = "float32" min_val = float("-1.57168") @@ -898,6 +980,7 @@ class Program_weight_tensor_parameter_81: class Program_weight_tensor_parameter_82: name = "parameter_82" + original_name = "linear_24.w_0" shape = [384, 1536] dtype = "float32" min_val = float("-0.296794") @@ -909,6 +992,7 @@ class Program_weight_tensor_parameter_82: class Program_weight_tensor_parameter_83: name = "parameter_83" + original_name = "layer_norm_14.b_0" shape = [384] dtype = "float32" min_val = float("-1.62605") @@ -920,6 +1004,7 @@ class Program_weight_tensor_parameter_83: class Program_weight_tensor_parameter_84: name = "parameter_84" + original_name = "layer_norm_14.w_0" shape = [384] dtype = "float32" min_val = float("0.29581") @@ -931,6 +1016,7 @@ class Program_weight_tensor_parameter_84: class Program_weight_tensor_parameter_85: name = "parameter_85" + original_name = "linear_23.b_0" shape = [384] dtype = "float32" min_val = float("-0.545845") @@ -942,6 +1028,7 @@ class Program_weight_tensor_parameter_85: class Program_weight_tensor_parameter_86: name = "parameter_86" + original_name = "linear_23.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.231587") @@ -953,6 +1040,7 @@ class Program_weight_tensor_parameter_86: class Program_weight_tensor_parameter_87: name = "parameter_87" + original_name = "linear_22.b_0" shape = [1152] dtype = "float32" min_val = float("-2.03367") @@ -964,6 +1052,7 @@ class Program_weight_tensor_parameter_87: class Program_weight_tensor_parameter_88: name = "parameter_88" + original_name = "linear_22.w_0" shape = [384, 1152] dtype = "float32" min_val = float("-0.535869") @@ -975,6 +1064,7 @@ class Program_weight_tensor_parameter_88: class Program_weight_tensor_parameter_89: name = "parameter_89" + original_name = "layer_norm_13.b_0" shape = [384] dtype = "float32" min_val = float("-1.10407") @@ -986,6 +1076,7 @@ class Program_weight_tensor_parameter_89: class Program_weight_tensor_parameter_90: name = "parameter_90" + original_name = "layer_norm_13.w_0" shape = [384] dtype = "float32" min_val = float("0.273017") @@ -997,6 +1088,7 @@ class Program_weight_tensor_parameter_90: class Program_weight_tensor_parameter_91: name = "parameter_91" + original_name = "linear_21.b_0" shape = [384] dtype = "float32" min_val = float("-1.02754") @@ -1008,6 +1100,7 @@ class Program_weight_tensor_parameter_91: class Program_weight_tensor_parameter_92: name = "parameter_92" + original_name = "linear_21.w_0" shape = [1536, 384] dtype = "float32" min_val = float("-0.28713") @@ -1019,6 +1112,7 @@ class Program_weight_tensor_parameter_92: class Program_weight_tensor_parameter_93: name = "parameter_93" + original_name = "linear_20.b_0" shape = [1536] dtype = "float32" min_val = float("-1.71602") @@ -1030,6 +1124,7 @@ class Program_weight_tensor_parameter_93: class Program_weight_tensor_parameter_94: name = "parameter_94" + original_name = "linear_20.w_0" shape = [384, 1536] dtype = "float32" min_val = float("-0.291663") @@ -1041,6 +1136,7 @@ class Program_weight_tensor_parameter_94: class Program_weight_tensor_parameter_95: name = "parameter_95" + original_name = "layer_norm_12.b_0" shape = [384] dtype = "float32" min_val = float("-1.61508") @@ -1052,6 +1148,7 @@ class Program_weight_tensor_parameter_95: class Program_weight_tensor_parameter_96: name = "parameter_96" + original_name = "layer_norm_12.w_0" shape = [384] dtype = "float32" min_val = float("-0.00197534") @@ -1063,6 +1160,7 @@ class Program_weight_tensor_parameter_96: class Program_weight_tensor_parameter_97: name = "parameter_97" + original_name = "linear_19.b_0" shape = [384] dtype = "float32" min_val = float("-1.08827") @@ -1074,6 +1172,7 @@ class Program_weight_tensor_parameter_97: class Program_weight_tensor_parameter_98: name = "parameter_98" + original_name = "linear_19.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.371053") @@ -1085,6 +1184,7 @@ class Program_weight_tensor_parameter_98: class Program_weight_tensor_parameter_99: name = "parameter_99" + original_name = "linear_18.b_0" shape = [1152] dtype = "float32" min_val = float("-2.23004") @@ -1096,6 +1196,7 @@ class Program_weight_tensor_parameter_99: class Program_weight_tensor_parameter_100: name = "parameter_100" + original_name = "linear_18.w_0" shape = [384, 1152] dtype = "float32" min_val = float("-0.356278") @@ -1107,6 +1208,7 @@ class Program_weight_tensor_parameter_100: class Program_weight_tensor_parameter_101: name = "parameter_101" + original_name = "layer_norm_11.b_0" shape = [384] dtype = "float32" min_val = float("-0.763318") @@ -1118,6 +1220,7 @@ class Program_weight_tensor_parameter_101: class Program_weight_tensor_parameter_102: name = "parameter_102" + original_name = "layer_norm_11.w_0" shape = [384] dtype = "float32" min_val = float("-0.000217926") @@ -1129,6 +1232,7 @@ class Program_weight_tensor_parameter_102: class Program_weight_tensor_parameter_103: name = "parameter_103" + original_name = "linear_17.w_0" shape = [768, 384] dtype = "float32" min_val = float("-0.227342") @@ -1140,6 +1244,7 @@ class Program_weight_tensor_parameter_103: class Program_weight_tensor_parameter_104: name = "parameter_104" + original_name = "layer_norm_10.b_0" shape = [768] dtype = "float32" min_val = float("-1.05595") @@ -1151,6 +1256,7 @@ class Program_weight_tensor_parameter_104: class Program_weight_tensor_parameter_105: name = "parameter_105" + original_name = "layer_norm_10.w_0" shape = [768] dtype = "float32" min_val = float("0.194633") @@ -1162,6 +1268,7 @@ class Program_weight_tensor_parameter_105: class Program_weight_tensor_parameter_106: name = "parameter_106" + original_name = "linear_16.b_0" shape = [192] dtype = "float32" min_val = float("-0.723563") @@ -1173,6 +1280,7 @@ class Program_weight_tensor_parameter_106: class Program_weight_tensor_parameter_107: name = "parameter_107" + original_name = "linear_16.w_0" shape = [768, 192] dtype = "float32" min_val = float("-0.412632") @@ -1184,6 +1292,7 @@ class Program_weight_tensor_parameter_107: class Program_weight_tensor_parameter_108: name = "parameter_108" + original_name = "linear_15.b_0" shape = [768] dtype = "float32" min_val = float("-1.63937") @@ -1195,6 +1304,7 @@ class Program_weight_tensor_parameter_108: class Program_weight_tensor_parameter_109: name = "parameter_109" + original_name = "linear_15.w_0" shape = [192, 768] dtype = "float32" min_val = float("-0.455203") @@ -1206,6 +1316,7 @@ class Program_weight_tensor_parameter_109: class Program_weight_tensor_parameter_110: name = "parameter_110" + original_name = "layer_norm_9.b_0" shape = [192] dtype = "float32" min_val = float("-1.70434") @@ -1217,6 +1328,7 @@ class Program_weight_tensor_parameter_110: class Program_weight_tensor_parameter_111: name = "parameter_111" + original_name = "layer_norm_9.w_0" shape = [192] dtype = "float32" min_val = float("0.740197") @@ -1228,6 +1340,7 @@ class Program_weight_tensor_parameter_111: class Program_weight_tensor_parameter_112: name = "parameter_112" + original_name = "linear_14.b_0" shape = [192] dtype = "float32" min_val = float("-0.574692") @@ -1239,6 +1352,7 @@ class Program_weight_tensor_parameter_112: class Program_weight_tensor_parameter_113: name = "parameter_113" + original_name = "linear_14.w_0" shape = [192, 192] dtype = "float32" min_val = float("-0.215523") @@ -1250,6 +1364,7 @@ class Program_weight_tensor_parameter_113: class Program_weight_tensor_parameter_114: name = "parameter_114" + original_name = "linear_13.b_0" shape = [576] dtype = "float32" min_val = float("-1.88681") @@ -1261,6 +1376,7 @@ class Program_weight_tensor_parameter_114: class Program_weight_tensor_parameter_115: name = "parameter_115" + original_name = "linear_13.w_0" shape = [192, 576] dtype = "float32" min_val = float("-0.437366") @@ -1272,6 +1388,7 @@ class Program_weight_tensor_parameter_115: class Program_weight_tensor_parameter_116: name = "parameter_116" + original_name = "layer_norm_8.b_0" shape = [192] dtype = "float32" min_val = float("-1.88825") @@ -1283,6 +1400,7 @@ class Program_weight_tensor_parameter_116: class Program_weight_tensor_parameter_117: name = "parameter_117" + original_name = "layer_norm_8.w_0" shape = [192] dtype = "float32" min_val = float("0.349115") @@ -1294,6 +1412,7 @@ class Program_weight_tensor_parameter_117: class Program_weight_tensor_parameter_118: name = "parameter_118" + original_name = "linear_12.b_0" shape = [192] dtype = "float32" min_val = float("-0.596898") @@ -1305,6 +1424,7 @@ class Program_weight_tensor_parameter_118: class Program_weight_tensor_parameter_119: name = "parameter_119" + original_name = "linear_12.w_0" shape = [768, 192] dtype = "float32" min_val = float("-0.324093") @@ -1316,6 +1436,7 @@ class Program_weight_tensor_parameter_119: class Program_weight_tensor_parameter_120: name = "parameter_120" + original_name = "linear_11.b_0" shape = [768] dtype = "float32" min_val = float("-1.65501") @@ -1327,6 +1448,7 @@ class Program_weight_tensor_parameter_120: class Program_weight_tensor_parameter_121: name = "parameter_121" + original_name = "linear_11.w_0" shape = [192, 768] dtype = "float32" min_val = float("-0.624714") @@ -1338,6 +1460,7 @@ class Program_weight_tensor_parameter_121: class Program_weight_tensor_parameter_122: name = "parameter_122" + original_name = "layer_norm_7.b_0" shape = [192] dtype = "float32" min_val = float("-1.24351") @@ -1349,6 +1472,7 @@ class Program_weight_tensor_parameter_122: class Program_weight_tensor_parameter_123: name = "parameter_123" + original_name = "layer_norm_7.w_0" shape = [192] dtype = "float32" min_val = float("0.28282") @@ -1360,6 +1484,7 @@ class Program_weight_tensor_parameter_123: class Program_weight_tensor_parameter_124: name = "parameter_124" + original_name = "linear_10.b_0" shape = [192] dtype = "float32" min_val = float("-0.540177") @@ -1371,6 +1496,7 @@ class Program_weight_tensor_parameter_124: class Program_weight_tensor_parameter_125: name = "parameter_125" + original_name = "linear_10.w_0" shape = [192, 192] dtype = "float32" min_val = float("-0.205249") @@ -1382,6 +1508,7 @@ class Program_weight_tensor_parameter_125: class Program_weight_tensor_parameter_126: name = "parameter_126" + original_name = "linear_9.b_0" shape = [576] dtype = "float32" min_val = float("-1.93105") @@ -1393,6 +1520,7 @@ class Program_weight_tensor_parameter_126: class Program_weight_tensor_parameter_127: name = "parameter_127" + original_name = "linear_9.w_0" shape = [192, 576] dtype = "float32" min_val = float("-0.347082") @@ -1404,6 +1532,7 @@ class Program_weight_tensor_parameter_127: class Program_weight_tensor_parameter_128: name = "parameter_128" + original_name = "layer_norm_6.b_0" shape = [192] dtype = "float32" min_val = float("-1.14355") @@ -1415,6 +1544,7 @@ class Program_weight_tensor_parameter_128: class Program_weight_tensor_parameter_129: name = "parameter_129" + original_name = "layer_norm_6.w_0" shape = [192] dtype = "float32" min_val = float("0.27789") @@ -1426,6 +1556,7 @@ class Program_weight_tensor_parameter_129: class Program_weight_tensor_parameter_130: name = "parameter_130" + original_name = "linear_8.w_0" shape = [384, 192] dtype = "float32" min_val = float("-0.250886") @@ -1437,6 +1568,7 @@ class Program_weight_tensor_parameter_130: class Program_weight_tensor_parameter_131: name = "parameter_131" + original_name = "layer_norm_5.b_0" shape = [384] dtype = "float32" min_val = float("-0.829282") @@ -1448,6 +1580,7 @@ class Program_weight_tensor_parameter_131: class Program_weight_tensor_parameter_132: name = "parameter_132" + original_name = "layer_norm_5.w_0" shape = [384] dtype = "float32" min_val = float("0.313923") @@ -1459,6 +1592,7 @@ class Program_weight_tensor_parameter_132: class Program_weight_tensor_parameter_133: name = "parameter_133" + original_name = "linear_7.b_0" shape = [96] dtype = "float32" min_val = float("-1.35868") @@ -1470,6 +1604,7 @@ class Program_weight_tensor_parameter_133: class Program_weight_tensor_parameter_134: name = "parameter_134" + original_name = "linear_7.w_0" shape = [384, 96] dtype = "float32" min_val = float("-0.529694") @@ -1481,6 +1616,7 @@ class Program_weight_tensor_parameter_134: class Program_weight_tensor_parameter_135: name = "parameter_135" + original_name = "linear_6.b_0" shape = [384] dtype = "float32" min_val = float("-1.51572") @@ -1492,6 +1628,7 @@ class Program_weight_tensor_parameter_135: class Program_weight_tensor_parameter_136: name = "parameter_136" + original_name = "linear_6.w_0" shape = [96, 384] dtype = "float32" min_val = float("-0.377043") @@ -1503,6 +1640,7 @@ class Program_weight_tensor_parameter_136: class Program_weight_tensor_parameter_137: name = "parameter_137" + original_name = "layer_norm_4.b_0" shape = [96] dtype = "float32" min_val = float("-0.849974") @@ -1514,6 +1652,7 @@ class Program_weight_tensor_parameter_137: class Program_weight_tensor_parameter_138: name = "parameter_138" + original_name = "layer_norm_4.w_0" shape = [96] dtype = "float32" min_val = float("0.496521") @@ -1525,6 +1664,7 @@ class Program_weight_tensor_parameter_138: class Program_weight_tensor_parameter_139: name = "parameter_139" + original_name = "linear_5.b_0" shape = [96] dtype = "float32" min_val = float("-0.431304") @@ -1536,6 +1676,7 @@ class Program_weight_tensor_parameter_139: class Program_weight_tensor_parameter_140: name = "parameter_140" + original_name = "linear_5.w_0" shape = [96, 96] dtype = "float32" min_val = float("-0.286262") @@ -1547,6 +1688,7 @@ class Program_weight_tensor_parameter_140: class Program_weight_tensor_parameter_141: name = "parameter_141" + original_name = "linear_4.b_0" shape = [288] dtype = "float32" min_val = float("-1.68118") @@ -1558,6 +1700,7 @@ class Program_weight_tensor_parameter_141: class Program_weight_tensor_parameter_142: name = "parameter_142" + original_name = "linear_4.w_0" shape = [96, 288] dtype = "float32" min_val = float("-0.394128") @@ -1569,6 +1712,7 @@ class Program_weight_tensor_parameter_142: class Program_weight_tensor_parameter_143: name = "parameter_143" + original_name = "layer_norm_3.b_0" shape = [96] dtype = "float32" min_val = float("-1.2989") @@ -1580,6 +1724,7 @@ class Program_weight_tensor_parameter_143: class Program_weight_tensor_parameter_144: name = "parameter_144" + original_name = "layer_norm_3.w_0" shape = [96] dtype = "float32" min_val = float("0.31154") @@ -1591,6 +1736,7 @@ class Program_weight_tensor_parameter_144: class Program_weight_tensor_parameter_145: name = "parameter_145" + original_name = "linear_3.b_0" shape = [96] dtype = "float32" min_val = float("-1.97801") @@ -1602,6 +1748,7 @@ class Program_weight_tensor_parameter_145: class Program_weight_tensor_parameter_146: name = "parameter_146" + original_name = "linear_3.w_0" shape = [384, 96] dtype = "float32" min_val = float("-0.414184") @@ -1613,6 +1760,7 @@ class Program_weight_tensor_parameter_146: class Program_weight_tensor_parameter_147: name = "parameter_147" + original_name = "linear_2.b_0" shape = [384] dtype = "float32" min_val = float("-1.88312") @@ -1624,6 +1772,7 @@ class Program_weight_tensor_parameter_147: class Program_weight_tensor_parameter_148: name = "parameter_148" + original_name = "linear_2.w_0" shape = [96, 384] dtype = "float32" min_val = float("-0.454946") @@ -1635,6 +1784,7 @@ class Program_weight_tensor_parameter_148: class Program_weight_tensor_parameter_149: name = "parameter_149" + original_name = "layer_norm_2.b_0" shape = [96] dtype = "float32" min_val = float("-1.07612") @@ -1646,6 +1796,7 @@ class Program_weight_tensor_parameter_149: class Program_weight_tensor_parameter_150: name = "parameter_150" + original_name = "layer_norm_2.w_0" shape = [96] dtype = "float32" min_val = float("0.538878") @@ -1657,6 +1808,7 @@ class Program_weight_tensor_parameter_150: class Program_weight_tensor_parameter_151: name = "parameter_151" + original_name = "linear_1.b_0" shape = [96] dtype = "float32" min_val = float("-0.648792") @@ -1668,6 +1820,7 @@ class Program_weight_tensor_parameter_151: class Program_weight_tensor_parameter_152: name = "parameter_152" + original_name = "linear_1.w_0" shape = [96, 96] dtype = "float32" min_val = float("-0.28441") @@ -1679,6 +1832,7 @@ class Program_weight_tensor_parameter_152: class Program_weight_tensor_parameter_153: name = "parameter_153" + original_name = "linear_0.b_0" shape = [288] dtype = "float32" min_val = float("-1.78417") @@ -1690,6 +1844,7 @@ class Program_weight_tensor_parameter_153: class Program_weight_tensor_parameter_154: name = "parameter_154" + original_name = "linear_0.w_0" shape = [96, 288] dtype = "float32" min_val = float("-0.67121") @@ -1701,6 +1856,7 @@ class Program_weight_tensor_parameter_154: class Program_weight_tensor_parameter_155: name = "parameter_155" + original_name = "layer_norm_1.b_0" shape = [96] dtype = "float32" min_val = float("-0.770537") @@ -1712,6 +1868,7 @@ class Program_weight_tensor_parameter_155: class Program_weight_tensor_parameter_156: name = "parameter_156" + original_name = "layer_norm_1.w_0" shape = [96] dtype = "float32" min_val = float("-0.000584983") @@ -1723,6 +1880,7 @@ class Program_weight_tensor_parameter_156: class Program_weight_tensor_parameter_157: name = "parameter_157" + original_name = "layer_norm_0.b_0" shape = [96] dtype = "float32" min_val = float("-0.971627") @@ -1734,6 +1892,7 @@ class Program_weight_tensor_parameter_157: class Program_weight_tensor_parameter_158: name = "parameter_158" + original_name = "layer_norm_0.w_0" shape = [96] dtype = "float32" min_val = float("-0.0248192") @@ -1745,6 +1904,7 @@ class Program_weight_tensor_parameter_158: class Program_weight_tensor_parameter_159: name = "parameter_159" + original_name = "conv2d_0.b_0" shape = [96] dtype = "float32" min_val = float("-1.2024") @@ -1756,6 +1916,7 @@ class Program_weight_tensor_parameter_159: class Program_weight_tensor_parameter_160: name = "parameter_160" + original_name = "conv2d_0.w_0" shape = [96, 3, 4, 4] dtype = "float32" min_val = float("-0.313436") diff --git a/paddle_samples/PaddleX/SwinTransformer_tiny_patch4_window7_224/subgraph_2/input_meta.py b/paddle_samples/PaddleX/SwinTransformer_tiny_patch4_window7_224/subgraph_2/input_meta.py index e5908636d..556569aaf 100644 --- a/paddle_samples/PaddleX/SwinTransformer_tiny_patch4_window7_224/subgraph_2/input_meta.py +++ b/paddle_samples/PaddleX/SwinTransformer_tiny_patch4_window7_224/subgraph_2/input_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_data_0: name = "data_0" + original_name = "param_12" shape = [169, 3] dtype = "float32" min_val = float("-10.7121") @@ -11,6 +12,7 @@ class Program_weight_tensor_data_0: class Program_weight_tensor_data_1: name = "data_1" + original_name = "param_13" shape = [169, 3] dtype = "float32" min_val = float("-8.27859") @@ -22,6 +24,7 @@ class Program_weight_tensor_data_1: class Program_weight_tensor_data_2: name = "data_2" + original_name = "param_14" shape = [169, 6] dtype = "float32" min_val = float("-6.35784") @@ -33,6 +36,7 @@ class Program_weight_tensor_data_2: class Program_weight_tensor_data_3: name = "data_3" + original_name = "param_15" shape = [169, 6] dtype = "float32" min_val = float("-9.10723") @@ -44,6 +48,7 @@ class Program_weight_tensor_data_3: class Program_weight_tensor_data_4: name = "data_4" + original_name = "param_16" shape = [169, 12] dtype = "float32" min_val = float("-10.2272") @@ -55,6 +60,7 @@ class Program_weight_tensor_data_4: class Program_weight_tensor_data_5: name = "data_5" + original_name = "param_17" shape = [169, 12] dtype = "float32" min_val = float("-10.5667") @@ -66,6 +72,7 @@ class Program_weight_tensor_data_5: class Program_weight_tensor_data_6: name = "data_6" + original_name = "param_18" shape = [169, 12] dtype = "float32" min_val = float("-11.5155") @@ -77,6 +84,7 @@ class Program_weight_tensor_data_6: class Program_weight_tensor_data_7: name = "data_7" + original_name = "param_19" shape = [169, 12] dtype = "float32" min_val = float("-7.51039") @@ -88,6 +96,7 @@ class Program_weight_tensor_data_7: class Program_weight_tensor_data_8: name = "data_8" + original_name = "param_20" shape = [169, 12] dtype = "float32" min_val = float("-10.5974") @@ -99,6 +108,7 @@ class Program_weight_tensor_data_8: class Program_weight_tensor_data_9: name = "data_9" + original_name = "param_21" shape = [169, 12] dtype = "float32" min_val = float("-13.1414") @@ -110,6 +120,7 @@ class Program_weight_tensor_data_9: class Program_weight_tensor_data_10: name = "data_10" + original_name = "param_22" shape = [169, 24] dtype = "float32" min_val = float("-15.1444") @@ -121,6 +132,7 @@ class Program_weight_tensor_data_10: class Program_weight_tensor_data_11: name = "data_11" + original_name = "param_23" shape = [169, 24] dtype = "float32" min_val = float("-15.511") @@ -132,6 +144,7 @@ class Program_weight_tensor_data_11: class Program_weight_tensor_data_12: name = "data_12" + original_name = "var_1070" shape = [49, 49] dtype = "int64" min_val = 0 @@ -141,6 +154,7 @@ class Program_weight_tensor_data_12: class Program_weight_tensor_data_13: name = "data_13" + original_name = "var_1162" shape = [49, 49] dtype = "int64" min_val = 0 @@ -150,6 +164,7 @@ class Program_weight_tensor_data_13: class Program_weight_tensor_data_14: name = "data_14" + original_name = "var_1250" shape = [49, 49] dtype = "int64" min_val = 0 @@ -159,6 +174,7 @@ class Program_weight_tensor_data_14: class Program_weight_tensor_data_15: name = "data_15" + original_name = "var_1342" shape = [49, 49] dtype = "int64" min_val = 0 @@ -168,6 +184,7 @@ class Program_weight_tensor_data_15: class Program_weight_tensor_data_16: name = "data_16" + original_name = "var_1414" shape = [49, 49] dtype = "int64" min_val = 0 @@ -177,6 +194,7 @@ class Program_weight_tensor_data_16: class Program_weight_tensor_data_17: name = "data_17" + original_name = "var_1506" shape = [49, 49] dtype = "int64" min_val = 0 @@ -186,6 +204,7 @@ class Program_weight_tensor_data_17: class Program_weight_tensor_data_18: name = "data_18" + original_name = "var_1578" shape = [49, 49] dtype = "int64" min_val = 0 @@ -195,6 +214,7 @@ class Program_weight_tensor_data_18: class Program_weight_tensor_data_19: name = "data_19" + original_name = "var_1670" shape = [49, 49] dtype = "int64" min_val = 0 @@ -204,6 +224,7 @@ class Program_weight_tensor_data_19: class Program_weight_tensor_data_20: name = "data_20" + original_name = "var_1758" shape = [49, 49] dtype = "int64" min_val = 0 @@ -213,6 +234,7 @@ class Program_weight_tensor_data_20: class Program_weight_tensor_data_21: name = "data_21" + original_name = "var_1850" shape = [49, 49] dtype = "int64" min_val = 0 @@ -222,6 +244,7 @@ class Program_weight_tensor_data_21: class Program_weight_tensor_data_22: name = "data_22" + original_name = "var_879" shape = [60, 3, 224, 224] dtype = "float32" min_val = float("-5.29778") @@ -233,6 +256,7 @@ class Program_weight_tensor_data_22: class Program_weight_tensor_data_23: name = "data_23" + original_name = "var_908" shape = [49, 49] dtype = "int64" min_val = 0 @@ -242,6 +266,7 @@ class Program_weight_tensor_data_23: class Program_weight_tensor_data_24: name = "data_24" + original_name = "var_982" shape = [49, 49] dtype = "int64" min_val = 0 diff --git a/paddle_samples/PaddleX/SwinTransformer_tiny_patch4_window7_224/subgraph_2/weight_meta.py b/paddle_samples/PaddleX/SwinTransformer_tiny_patch4_window7_224/subgraph_2/weight_meta.py index 38e70029d..29661a435 100644 --- a/paddle_samples/PaddleX/SwinTransformer_tiny_patch4_window7_224/subgraph_2/weight_meta.py +++ b/paddle_samples/PaddleX/SwinTransformer_tiny_patch4_window7_224/subgraph_2/weight_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_parameter_0: name = "parameter_0" + original_name = "linear_51.b_0" shape = [102] dtype = "float32" min_val = float("-0.0032138") @@ -11,6 +12,7 @@ class Program_weight_tensor_parameter_0: class Program_weight_tensor_parameter_1: name = "parameter_1" + original_name = "linear_51.w_0" shape = [768, 102] dtype = "float32" min_val = float("-0.0793346") @@ -22,6 +24,7 @@ class Program_weight_tensor_parameter_1: class Program_weight_tensor_parameter_2: name = "parameter_2" + original_name = "layer_norm_28.b_0" shape = [768] dtype = "float32" min_val = float("-0.859228") @@ -33,6 +36,7 @@ class Program_weight_tensor_parameter_2: class Program_weight_tensor_parameter_3: name = "parameter_3" + original_name = "layer_norm_28.w_0" shape = [768] dtype = "float32" min_val = float("0.000842456") @@ -44,6 +48,7 @@ class Program_weight_tensor_parameter_3: class Program_weight_tensor_parameter_4: name = "parameter_4" + original_name = "linear_50.b_0" shape = [768] dtype = "float32" min_val = float("-7.73233") @@ -55,6 +60,7 @@ class Program_weight_tensor_parameter_4: class Program_weight_tensor_parameter_5: name = "parameter_5" + original_name = "linear_50.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-1.78837") @@ -66,6 +72,7 @@ class Program_weight_tensor_parameter_5: class Program_weight_tensor_parameter_6: name = "parameter_6" + original_name = "linear_49.b_0" shape = [3072] dtype = "float32" min_val = float("-1.61695") @@ -77,6 +84,7 @@ class Program_weight_tensor_parameter_6: class Program_weight_tensor_parameter_7: name = "parameter_7" + original_name = "linear_49.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.472912") @@ -88,6 +96,7 @@ class Program_weight_tensor_parameter_7: class Program_weight_tensor_parameter_8: name = "parameter_8" + original_name = "layer_norm_27.b_0" shape = [768] dtype = "float32" min_val = float("-1.98572") @@ -99,6 +108,7 @@ class Program_weight_tensor_parameter_8: class Program_weight_tensor_parameter_9: name = "parameter_9" + original_name = "layer_norm_27.w_0" shape = [768] dtype = "float32" min_val = float("0.0012819") @@ -110,6 +120,7 @@ class Program_weight_tensor_parameter_9: class Program_weight_tensor_parameter_10: name = "parameter_10" + original_name = "linear_48.b_0" shape = [768] dtype = "float32" min_val = float("-11.8407") @@ -121,6 +132,7 @@ class Program_weight_tensor_parameter_10: class Program_weight_tensor_parameter_11: name = "parameter_11" + original_name = "linear_48.w_0" shape = [768, 768] dtype = "float32" min_val = float("-0.919007") @@ -132,6 +144,7 @@ class Program_weight_tensor_parameter_11: class Program_weight_tensor_parameter_12: name = "parameter_12" + original_name = "linear_47.b_0" shape = [2304] dtype = "float32" min_val = float("-2.85028") @@ -143,6 +156,7 @@ class Program_weight_tensor_parameter_12: class Program_weight_tensor_parameter_13: name = "parameter_13" + original_name = "linear_47.w_0" shape = [768, 2304] dtype = "float32" min_val = float("-0.414737") @@ -154,6 +168,7 @@ class Program_weight_tensor_parameter_13: class Program_weight_tensor_parameter_14: name = "parameter_14" + original_name = "layer_norm_26.b_0" shape = [768] dtype = "float32" min_val = float("-2.96593") @@ -165,6 +180,7 @@ class Program_weight_tensor_parameter_14: class Program_weight_tensor_parameter_15: name = "parameter_15" + original_name = "layer_norm_26.w_0" shape = [768] dtype = "float32" min_val = float("0.0226107") @@ -176,6 +192,7 @@ class Program_weight_tensor_parameter_15: class Program_weight_tensor_parameter_16: name = "parameter_16" + original_name = "linear_46.b_0" shape = [768] dtype = "float32" min_val = float("-2.19378") @@ -187,6 +204,7 @@ class Program_weight_tensor_parameter_16: class Program_weight_tensor_parameter_17: name = "parameter_17" + original_name = "linear_46.w_0" shape = [3072, 768] dtype = "float32" min_val = float("-1.70617") @@ -198,6 +216,7 @@ class Program_weight_tensor_parameter_17: class Program_weight_tensor_parameter_18: name = "parameter_18" + original_name = "linear_45.b_0" shape = [3072] dtype = "float32" min_val = float("-1.07659") @@ -209,6 +228,7 @@ class Program_weight_tensor_parameter_18: class Program_weight_tensor_parameter_19: name = "parameter_19" + original_name = "linear_45.w_0" shape = [768, 3072] dtype = "float32" min_val = float("-0.538574") @@ -220,6 +240,7 @@ class Program_weight_tensor_parameter_19: class Program_weight_tensor_parameter_20: name = "parameter_20" + original_name = "layer_norm_25.b_0" shape = [768] dtype = "float32" min_val = float("-3.4174") @@ -231,6 +252,7 @@ class Program_weight_tensor_parameter_20: class Program_weight_tensor_parameter_21: name = "parameter_21" + original_name = "layer_norm_25.w_0" shape = [768] dtype = "float32" min_val = float("-0.320864") @@ -242,6 +264,7 @@ class Program_weight_tensor_parameter_21: class Program_weight_tensor_parameter_22: name = "parameter_22" + original_name = "linear_44.b_0" shape = [768] dtype = "float32" min_val = float("-4.92857") @@ -253,6 +276,7 @@ class Program_weight_tensor_parameter_22: class Program_weight_tensor_parameter_23: name = "parameter_23" + original_name = "linear_44.w_0" shape = [768, 768] dtype = "float32" min_val = float("-1.07384") @@ -264,6 +288,7 @@ class Program_weight_tensor_parameter_23: class Program_weight_tensor_parameter_24: name = "parameter_24" + original_name = "linear_43.b_0" shape = [2304] dtype = "float32" min_val = float("-2.8255") @@ -275,6 +300,7 @@ class Program_weight_tensor_parameter_24: class Program_weight_tensor_parameter_25: name = "parameter_25" + original_name = "linear_43.w_0" shape = [768, 2304] dtype = "float32" min_val = float("-0.271991") @@ -286,6 +312,7 @@ class Program_weight_tensor_parameter_25: class Program_weight_tensor_parameter_26: name = "parameter_26" + original_name = "layer_norm_24.b_0" shape = [768] dtype = "float32" min_val = float("-1.83482") @@ -297,6 +324,7 @@ class Program_weight_tensor_parameter_26: class Program_weight_tensor_parameter_27: name = "parameter_27" + original_name = "layer_norm_24.w_0" shape = [768] dtype = "float32" min_val = float("-0.00482071") @@ -308,6 +336,7 @@ class Program_weight_tensor_parameter_27: class Program_weight_tensor_parameter_28: name = "parameter_28" + original_name = "linear_42.w_0" shape = [1536, 768] dtype = "float32" min_val = float("-0.851016") @@ -319,6 +348,7 @@ class Program_weight_tensor_parameter_28: class Program_weight_tensor_parameter_29: name = "parameter_29" + original_name = "layer_norm_23.b_0" shape = [1536] dtype = "float32" min_val = float("-0.656577") @@ -330,6 +360,7 @@ class Program_weight_tensor_parameter_29: class Program_weight_tensor_parameter_30: name = "parameter_30" + original_name = "layer_norm_23.w_0" shape = [1536] dtype = "float32" min_val = float("0.0537253") @@ -341,6 +372,7 @@ class Program_weight_tensor_parameter_30: class Program_weight_tensor_parameter_31: name = "parameter_31" + original_name = "linear_41.b_0" shape = [384] dtype = "float32" min_val = float("-0.774784") @@ -352,6 +384,7 @@ class Program_weight_tensor_parameter_31: class Program_weight_tensor_parameter_32: name = "parameter_32" + original_name = "linear_41.w_0" shape = [1536, 384] dtype = "float32" min_val = float("-1.54791") @@ -363,6 +396,7 @@ class Program_weight_tensor_parameter_32: class Program_weight_tensor_parameter_33: name = "parameter_33" + original_name = "linear_40.b_0" shape = [1536] dtype = "float32" min_val = float("-3.79721") @@ -374,6 +408,7 @@ class Program_weight_tensor_parameter_33: class Program_weight_tensor_parameter_34: name = "parameter_34" + original_name = "linear_40.w_0" shape = [384, 1536] dtype = "float32" min_val = float("-0.771719") @@ -385,6 +420,7 @@ class Program_weight_tensor_parameter_34: class Program_weight_tensor_parameter_35: name = "parameter_35" + original_name = "layer_norm_22.b_0" shape = [384] dtype = "float32" min_val = float("-2.23504") @@ -396,6 +432,7 @@ class Program_weight_tensor_parameter_35: class Program_weight_tensor_parameter_36: name = "parameter_36" + original_name = "layer_norm_22.w_0" shape = [384] dtype = "float32" min_val = float("0.250031") @@ -407,6 +444,7 @@ class Program_weight_tensor_parameter_36: class Program_weight_tensor_parameter_37: name = "parameter_37" + original_name = "linear_39.b_0" shape = [384] dtype = "float32" min_val = float("-0.872449") @@ -418,6 +456,7 @@ class Program_weight_tensor_parameter_37: class Program_weight_tensor_parameter_38: name = "parameter_38" + original_name = "linear_39.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.474598") @@ -429,6 +468,7 @@ class Program_weight_tensor_parameter_38: class Program_weight_tensor_parameter_39: name = "parameter_39" + original_name = "linear_38.b_0" shape = [1152] dtype = "float32" min_val = float("-1.8448") @@ -440,6 +480,7 @@ class Program_weight_tensor_parameter_39: class Program_weight_tensor_parameter_40: name = "parameter_40" + original_name = "linear_38.w_0" shape = [384, 1152] dtype = "float32" min_val = float("-0.369862") @@ -451,6 +492,7 @@ class Program_weight_tensor_parameter_40: class Program_weight_tensor_parameter_41: name = "parameter_41" + original_name = "layer_norm_21.b_0" shape = [384] dtype = "float32" min_val = float("-1.96083") @@ -462,6 +504,7 @@ class Program_weight_tensor_parameter_41: class Program_weight_tensor_parameter_42: name = "parameter_42" + original_name = "layer_norm_21.w_0" shape = [384] dtype = "float32" min_val = float("0.224316") @@ -473,6 +516,7 @@ class Program_weight_tensor_parameter_42: class Program_weight_tensor_parameter_43: name = "parameter_43" + original_name = "linear_37.b_0" shape = [384] dtype = "float32" min_val = float("-0.797691") @@ -484,6 +528,7 @@ class Program_weight_tensor_parameter_43: class Program_weight_tensor_parameter_44: name = "parameter_44" + original_name = "linear_37.w_0" shape = [1536, 384] dtype = "float32" min_val = float("-2.91803") @@ -495,6 +540,7 @@ class Program_weight_tensor_parameter_44: class Program_weight_tensor_parameter_45: name = "parameter_45" + original_name = "linear_36.b_0" shape = [1536] dtype = "float32" min_val = float("-2.75858") @@ -506,6 +552,7 @@ class Program_weight_tensor_parameter_45: class Program_weight_tensor_parameter_46: name = "parameter_46" + original_name = "linear_36.w_0" shape = [384, 1536] dtype = "float32" min_val = float("-0.519521") @@ -517,6 +564,7 @@ class Program_weight_tensor_parameter_46: class Program_weight_tensor_parameter_47: name = "parameter_47" + original_name = "layer_norm_20.b_0" shape = [384] dtype = "float32" min_val = float("-1.83595") @@ -528,6 +576,7 @@ class Program_weight_tensor_parameter_47: class Program_weight_tensor_parameter_48: name = "parameter_48" + original_name = "layer_norm_20.w_0" shape = [384] dtype = "float32" min_val = float("0.236963") @@ -539,6 +588,7 @@ class Program_weight_tensor_parameter_48: class Program_weight_tensor_parameter_49: name = "parameter_49" + original_name = "linear_35.b_0" shape = [384] dtype = "float32" min_val = float("-1.0287") @@ -550,6 +600,7 @@ class Program_weight_tensor_parameter_49: class Program_weight_tensor_parameter_50: name = "parameter_50" + original_name = "linear_35.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.401614") @@ -561,6 +612,7 @@ class Program_weight_tensor_parameter_50: class Program_weight_tensor_parameter_51: name = "parameter_51" + original_name = "linear_34.b_0" shape = [1152] dtype = "float32" min_val = float("-2.85986") @@ -572,6 +624,7 @@ class Program_weight_tensor_parameter_51: class Program_weight_tensor_parameter_52: name = "parameter_52" + original_name = "linear_34.w_0" shape = [384, 1152] dtype = "float32" min_val = float("-0.361912") @@ -583,6 +636,7 @@ class Program_weight_tensor_parameter_52: class Program_weight_tensor_parameter_53: name = "parameter_53" + original_name = "layer_norm_19.b_0" shape = [384] dtype = "float32" min_val = float("-1.56268") @@ -594,6 +648,7 @@ class Program_weight_tensor_parameter_53: class Program_weight_tensor_parameter_54: name = "parameter_54" + original_name = "layer_norm_19.w_0" shape = [384] dtype = "float32" min_val = float("0.25592") @@ -605,6 +660,7 @@ class Program_weight_tensor_parameter_54: class Program_weight_tensor_parameter_55: name = "parameter_55" + original_name = "linear_33.b_0" shape = [384] dtype = "float32" min_val = float("-0.585054") @@ -616,6 +672,7 @@ class Program_weight_tensor_parameter_55: class Program_weight_tensor_parameter_56: name = "parameter_56" + original_name = "linear_33.w_0" shape = [1536, 384] dtype = "float32" min_val = float("-2.38867") @@ -627,6 +684,7 @@ class Program_weight_tensor_parameter_56: class Program_weight_tensor_parameter_57: name = "parameter_57" + original_name = "linear_32.b_0" shape = [1536] dtype = "float32" min_val = float("-2.03548") @@ -638,6 +696,7 @@ class Program_weight_tensor_parameter_57: class Program_weight_tensor_parameter_58: name = "parameter_58" + original_name = "linear_32.w_0" shape = [384, 1536] dtype = "float32" min_val = float("-0.26874") @@ -649,6 +708,7 @@ class Program_weight_tensor_parameter_58: class Program_weight_tensor_parameter_59: name = "parameter_59" + original_name = "layer_norm_18.b_0" shape = [384] dtype = "float32" min_val = float("-1.62054") @@ -660,6 +720,7 @@ class Program_weight_tensor_parameter_59: class Program_weight_tensor_parameter_60: name = "parameter_60" + original_name = "layer_norm_18.w_0" shape = [384] dtype = "float32" min_val = float("0.255076") @@ -671,6 +732,7 @@ class Program_weight_tensor_parameter_60: class Program_weight_tensor_parameter_61: name = "parameter_61" + original_name = "linear_31.b_0" shape = [384] dtype = "float32" min_val = float("-0.547175") @@ -682,6 +744,7 @@ class Program_weight_tensor_parameter_61: class Program_weight_tensor_parameter_62: name = "parameter_62" + original_name = "linear_31.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.276896") @@ -693,6 +756,7 @@ class Program_weight_tensor_parameter_62: class Program_weight_tensor_parameter_63: name = "parameter_63" + original_name = "linear_30.b_0" shape = [1152] dtype = "float32" min_val = float("-1.96785") @@ -704,6 +768,7 @@ class Program_weight_tensor_parameter_63: class Program_weight_tensor_parameter_64: name = "parameter_64" + original_name = "linear_30.w_0" shape = [384, 1152] dtype = "float32" min_val = float("-0.464305") @@ -715,6 +780,7 @@ class Program_weight_tensor_parameter_64: class Program_weight_tensor_parameter_65: name = "parameter_65" + original_name = "layer_norm_17.b_0" shape = [384] dtype = "float32" min_val = float("-1.45786") @@ -726,6 +792,7 @@ class Program_weight_tensor_parameter_65: class Program_weight_tensor_parameter_66: name = "parameter_66" + original_name = "layer_norm_17.w_0" shape = [384] dtype = "float32" min_val = float("0.290031") @@ -737,6 +804,7 @@ class Program_weight_tensor_parameter_66: class Program_weight_tensor_parameter_67: name = "parameter_67" + original_name = "linear_29.b_0" shape = [384] dtype = "float32" min_val = float("-0.477529") @@ -748,6 +816,7 @@ class Program_weight_tensor_parameter_67: class Program_weight_tensor_parameter_68: name = "parameter_68" + original_name = "linear_29.w_0" shape = [1536, 384] dtype = "float32" min_val = float("-1.45931") @@ -759,6 +828,7 @@ class Program_weight_tensor_parameter_68: class Program_weight_tensor_parameter_69: name = "parameter_69" + original_name = "linear_28.b_0" shape = [1536] dtype = "float32" min_val = float("-1.44845") @@ -770,6 +840,7 @@ class Program_weight_tensor_parameter_69: class Program_weight_tensor_parameter_70: name = "parameter_70" + original_name = "linear_28.w_0" shape = [384, 1536] dtype = "float32" min_val = float("-0.278611") @@ -781,6 +852,7 @@ class Program_weight_tensor_parameter_70: class Program_weight_tensor_parameter_71: name = "parameter_71" + original_name = "layer_norm_16.b_0" shape = [384] dtype = "float32" min_val = float("-1.70663") @@ -792,6 +864,7 @@ class Program_weight_tensor_parameter_71: class Program_weight_tensor_parameter_72: name = "parameter_72" + original_name = "layer_norm_16.w_0" shape = [384] dtype = "float32" min_val = float("0.283466") @@ -803,6 +876,7 @@ class Program_weight_tensor_parameter_72: class Program_weight_tensor_parameter_73: name = "parameter_73" + original_name = "linear_27.b_0" shape = [384] dtype = "float32" min_val = float("-0.892802") @@ -814,6 +888,7 @@ class Program_weight_tensor_parameter_73: class Program_weight_tensor_parameter_74: name = "parameter_74" + original_name = "linear_27.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.222939") @@ -825,6 +900,7 @@ class Program_weight_tensor_parameter_74: class Program_weight_tensor_parameter_75: name = "parameter_75" + original_name = "linear_26.b_0" shape = [1152] dtype = "float32" min_val = float("-2.01265") @@ -836,6 +912,7 @@ class Program_weight_tensor_parameter_75: class Program_weight_tensor_parameter_76: name = "parameter_76" + original_name = "linear_26.w_0" shape = [384, 1152] dtype = "float32" min_val = float("-0.446319") @@ -847,6 +924,7 @@ class Program_weight_tensor_parameter_76: class Program_weight_tensor_parameter_77: name = "parameter_77" + original_name = "layer_norm_15.b_0" shape = [384] dtype = "float32" min_val = float("-1.17689") @@ -858,6 +936,7 @@ class Program_weight_tensor_parameter_77: class Program_weight_tensor_parameter_78: name = "parameter_78" + original_name = "layer_norm_15.w_0" shape = [384] dtype = "float32" min_val = float("0.29653") @@ -869,6 +948,7 @@ class Program_weight_tensor_parameter_78: class Program_weight_tensor_parameter_79: name = "parameter_79" + original_name = "linear_25.b_0" shape = [384] dtype = "float32" min_val = float("-0.526399") @@ -880,6 +960,7 @@ class Program_weight_tensor_parameter_79: class Program_weight_tensor_parameter_80: name = "parameter_80" + original_name = "linear_25.w_0" shape = [1536, 384] dtype = "float32" min_val = float("-0.723073") @@ -891,6 +972,7 @@ class Program_weight_tensor_parameter_80: class Program_weight_tensor_parameter_81: name = "parameter_81" + original_name = "linear_24.b_0" shape = [1536] dtype = "float32" min_val = float("-1.57331") @@ -902,6 +984,7 @@ class Program_weight_tensor_parameter_81: class Program_weight_tensor_parameter_82: name = "parameter_82" + original_name = "linear_24.w_0" shape = [384, 1536] dtype = "float32" min_val = float("-0.297295") @@ -913,6 +996,7 @@ class Program_weight_tensor_parameter_82: class Program_weight_tensor_parameter_83: name = "parameter_83" + original_name = "layer_norm_14.b_0" shape = [384] dtype = "float32" min_val = float("-1.62657") @@ -924,6 +1008,7 @@ class Program_weight_tensor_parameter_83: class Program_weight_tensor_parameter_84: name = "parameter_84" + original_name = "layer_norm_14.w_0" shape = [384] dtype = "float32" min_val = float("0.295009") @@ -935,6 +1020,7 @@ class Program_weight_tensor_parameter_84: class Program_weight_tensor_parameter_85: name = "parameter_85" + original_name = "linear_23.b_0" shape = [384] dtype = "float32" min_val = float("-0.544806") @@ -946,6 +1032,7 @@ class Program_weight_tensor_parameter_85: class Program_weight_tensor_parameter_86: name = "parameter_86" + original_name = "linear_23.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.230512") @@ -957,6 +1044,7 @@ class Program_weight_tensor_parameter_86: class Program_weight_tensor_parameter_87: name = "parameter_87" + original_name = "linear_22.b_0" shape = [1152] dtype = "float32" min_val = float("-2.03808") @@ -968,6 +1056,7 @@ class Program_weight_tensor_parameter_87: class Program_weight_tensor_parameter_88: name = "parameter_88" + original_name = "linear_22.w_0" shape = [384, 1152] dtype = "float32" min_val = float("-0.531712") @@ -979,6 +1068,7 @@ class Program_weight_tensor_parameter_88: class Program_weight_tensor_parameter_89: name = "parameter_89" + original_name = "layer_norm_13.b_0" shape = [384] dtype = "float32" min_val = float("-1.10242") @@ -990,6 +1080,7 @@ class Program_weight_tensor_parameter_89: class Program_weight_tensor_parameter_90: name = "parameter_90" + original_name = "layer_norm_13.w_0" shape = [384] dtype = "float32" min_val = float("0.273649") @@ -1001,6 +1092,7 @@ class Program_weight_tensor_parameter_90: class Program_weight_tensor_parameter_91: name = "parameter_91" + original_name = "linear_21.b_0" shape = [384] dtype = "float32" min_val = float("-1.02864") @@ -1012,6 +1104,7 @@ class Program_weight_tensor_parameter_91: class Program_weight_tensor_parameter_92: name = "parameter_92" + original_name = "linear_21.w_0" shape = [1536, 384] dtype = "float32" min_val = float("-0.28693") @@ -1023,6 +1116,7 @@ class Program_weight_tensor_parameter_92: class Program_weight_tensor_parameter_93: name = "parameter_93" + original_name = "linear_20.b_0" shape = [1536] dtype = "float32" min_val = float("-1.71781") @@ -1034,6 +1128,7 @@ class Program_weight_tensor_parameter_93: class Program_weight_tensor_parameter_94: name = "parameter_94" + original_name = "linear_20.w_0" shape = [384, 1536] dtype = "float32" min_val = float("-0.295722") @@ -1045,6 +1140,7 @@ class Program_weight_tensor_parameter_94: class Program_weight_tensor_parameter_95: name = "parameter_95" + original_name = "layer_norm_12.b_0" shape = [384] dtype = "float32" min_val = float("-1.61483") @@ -1056,6 +1152,7 @@ class Program_weight_tensor_parameter_95: class Program_weight_tensor_parameter_96: name = "parameter_96" + original_name = "layer_norm_12.w_0" shape = [384] dtype = "float32" min_val = float("-0.00559005") @@ -1067,6 +1164,7 @@ class Program_weight_tensor_parameter_96: class Program_weight_tensor_parameter_97: name = "parameter_97" + original_name = "linear_19.b_0" shape = [384] dtype = "float32" min_val = float("-1.09016") @@ -1078,6 +1176,7 @@ class Program_weight_tensor_parameter_97: class Program_weight_tensor_parameter_98: name = "parameter_98" + original_name = "linear_19.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.375988") @@ -1089,6 +1188,7 @@ class Program_weight_tensor_parameter_98: class Program_weight_tensor_parameter_99: name = "parameter_99" + original_name = "linear_18.b_0" shape = [1152] dtype = "float32" min_val = float("-2.23186") @@ -1100,6 +1200,7 @@ class Program_weight_tensor_parameter_99: class Program_weight_tensor_parameter_100: name = "parameter_100" + original_name = "linear_18.w_0" shape = [384, 1152] dtype = "float32" min_val = float("-0.355518") @@ -1111,6 +1212,7 @@ class Program_weight_tensor_parameter_100: class Program_weight_tensor_parameter_101: name = "parameter_101" + original_name = "layer_norm_11.b_0" shape = [384] dtype = "float32" min_val = float("-0.760609") @@ -1122,6 +1224,7 @@ class Program_weight_tensor_parameter_101: class Program_weight_tensor_parameter_102: name = "parameter_102" + original_name = "layer_norm_11.w_0" shape = [384] dtype = "float32" min_val = float("-0.000638352") @@ -1133,6 +1236,7 @@ class Program_weight_tensor_parameter_102: class Program_weight_tensor_parameter_103: name = "parameter_103" + original_name = "linear_17.w_0" shape = [768, 384] dtype = "float32" min_val = float("-0.227441") @@ -1144,6 +1248,7 @@ class Program_weight_tensor_parameter_103: class Program_weight_tensor_parameter_104: name = "parameter_104" + original_name = "layer_norm_10.b_0" shape = [768] dtype = "float32" min_val = float("-1.0546") @@ -1155,6 +1260,7 @@ class Program_weight_tensor_parameter_104: class Program_weight_tensor_parameter_105: name = "parameter_105" + original_name = "layer_norm_10.w_0" shape = [768] dtype = "float32" min_val = float("0.196858") @@ -1166,6 +1272,7 @@ class Program_weight_tensor_parameter_105: class Program_weight_tensor_parameter_106: name = "parameter_106" + original_name = "linear_16.b_0" shape = [192] dtype = "float32" min_val = float("-0.725312") @@ -1177,6 +1284,7 @@ class Program_weight_tensor_parameter_106: class Program_weight_tensor_parameter_107: name = "parameter_107" + original_name = "linear_16.w_0" shape = [768, 192] dtype = "float32" min_val = float("-0.411441") @@ -1188,6 +1296,7 @@ class Program_weight_tensor_parameter_107: class Program_weight_tensor_parameter_108: name = "parameter_108" + original_name = "linear_15.b_0" shape = [768] dtype = "float32" min_val = float("-1.63752") @@ -1199,6 +1308,7 @@ class Program_weight_tensor_parameter_108: class Program_weight_tensor_parameter_109: name = "parameter_109" + original_name = "linear_15.w_0" shape = [192, 768] dtype = "float32" min_val = float("-0.456257") @@ -1210,6 +1320,7 @@ class Program_weight_tensor_parameter_109: class Program_weight_tensor_parameter_110: name = "parameter_110" + original_name = "layer_norm_9.b_0" shape = [192] dtype = "float32" min_val = float("-1.70154") @@ -1221,6 +1332,7 @@ class Program_weight_tensor_parameter_110: class Program_weight_tensor_parameter_111: name = "parameter_111" + original_name = "layer_norm_9.w_0" shape = [192] dtype = "float32" min_val = float("0.743944") @@ -1232,6 +1344,7 @@ class Program_weight_tensor_parameter_111: class Program_weight_tensor_parameter_112: name = "parameter_112" + original_name = "linear_14.b_0" shape = [192] dtype = "float32" min_val = float("-0.579052") @@ -1243,6 +1356,7 @@ class Program_weight_tensor_parameter_112: class Program_weight_tensor_parameter_113: name = "parameter_113" + original_name = "linear_14.w_0" shape = [192, 192] dtype = "float32" min_val = float("-0.214324") @@ -1254,6 +1368,7 @@ class Program_weight_tensor_parameter_113: class Program_weight_tensor_parameter_114: name = "parameter_114" + original_name = "linear_13.b_0" shape = [576] dtype = "float32" min_val = float("-1.88235") @@ -1265,6 +1380,7 @@ class Program_weight_tensor_parameter_114: class Program_weight_tensor_parameter_115: name = "parameter_115" + original_name = "linear_13.w_0" shape = [192, 576] dtype = "float32" min_val = float("-0.441325") @@ -1276,6 +1392,7 @@ class Program_weight_tensor_parameter_115: class Program_weight_tensor_parameter_116: name = "parameter_116" + original_name = "layer_norm_8.b_0" shape = [192] dtype = "float32" min_val = float("-1.88954") @@ -1287,6 +1404,7 @@ class Program_weight_tensor_parameter_116: class Program_weight_tensor_parameter_117: name = "parameter_117" + original_name = "layer_norm_8.w_0" shape = [192] dtype = "float32" min_val = float("0.34702") @@ -1298,6 +1416,7 @@ class Program_weight_tensor_parameter_117: class Program_weight_tensor_parameter_118: name = "parameter_118" + original_name = "linear_12.b_0" shape = [192] dtype = "float32" min_val = float("-0.591236") @@ -1309,6 +1428,7 @@ class Program_weight_tensor_parameter_118: class Program_weight_tensor_parameter_119: name = "parameter_119" + original_name = "linear_12.w_0" shape = [768, 192] dtype = "float32" min_val = float("-0.322185") @@ -1320,6 +1440,7 @@ class Program_weight_tensor_parameter_119: class Program_weight_tensor_parameter_120: name = "parameter_120" + original_name = "linear_11.b_0" shape = [768] dtype = "float32" min_val = float("-1.65778") @@ -1331,6 +1452,7 @@ class Program_weight_tensor_parameter_120: class Program_weight_tensor_parameter_121: name = "parameter_121" + original_name = "linear_11.w_0" shape = [192, 768] dtype = "float32" min_val = float("-0.623798") @@ -1342,6 +1464,7 @@ class Program_weight_tensor_parameter_121: class Program_weight_tensor_parameter_122: name = "parameter_122" + original_name = "layer_norm_7.b_0" shape = [192] dtype = "float32" min_val = float("-1.24444") @@ -1353,6 +1476,7 @@ class Program_weight_tensor_parameter_122: class Program_weight_tensor_parameter_123: name = "parameter_123" + original_name = "layer_norm_7.w_0" shape = [192] dtype = "float32" min_val = float("0.282978") @@ -1364,6 +1488,7 @@ class Program_weight_tensor_parameter_123: class Program_weight_tensor_parameter_124: name = "parameter_124" + original_name = "linear_10.b_0" shape = [192] dtype = "float32" min_val = float("-0.541021") @@ -1375,6 +1500,7 @@ class Program_weight_tensor_parameter_124: class Program_weight_tensor_parameter_125: name = "parameter_125" + original_name = "linear_10.w_0" shape = [192, 192] dtype = "float32" min_val = float("-0.206341") @@ -1386,6 +1512,7 @@ class Program_weight_tensor_parameter_125: class Program_weight_tensor_parameter_126: name = "parameter_126" + original_name = "linear_9.b_0" shape = [576] dtype = "float32" min_val = float("-1.93005") @@ -1397,6 +1524,7 @@ class Program_weight_tensor_parameter_126: class Program_weight_tensor_parameter_127: name = "parameter_127" + original_name = "linear_9.w_0" shape = [192, 576] dtype = "float32" min_val = float("-0.35002") @@ -1408,6 +1536,7 @@ class Program_weight_tensor_parameter_127: class Program_weight_tensor_parameter_128: name = "parameter_128" + original_name = "layer_norm_6.b_0" shape = [192] dtype = "float32" min_val = float("-1.14175") @@ -1419,6 +1548,7 @@ class Program_weight_tensor_parameter_128: class Program_weight_tensor_parameter_129: name = "parameter_129" + original_name = "layer_norm_6.w_0" shape = [192] dtype = "float32" min_val = float("0.277555") @@ -1430,6 +1560,7 @@ class Program_weight_tensor_parameter_129: class Program_weight_tensor_parameter_130: name = "parameter_130" + original_name = "linear_8.w_0" shape = [384, 192] dtype = "float32" min_val = float("-0.250659") @@ -1441,6 +1572,7 @@ class Program_weight_tensor_parameter_130: class Program_weight_tensor_parameter_131: name = "parameter_131" + original_name = "layer_norm_5.b_0" shape = [384] dtype = "float32" min_val = float("-0.830428") @@ -1452,6 +1584,7 @@ class Program_weight_tensor_parameter_131: class Program_weight_tensor_parameter_132: name = "parameter_132" + original_name = "layer_norm_5.w_0" shape = [384] dtype = "float32" min_val = float("0.313936") @@ -1463,6 +1596,7 @@ class Program_weight_tensor_parameter_132: class Program_weight_tensor_parameter_133: name = "parameter_133" + original_name = "linear_7.b_0" shape = [96] dtype = "float32" min_val = float("-1.35734") @@ -1474,6 +1608,7 @@ class Program_weight_tensor_parameter_133: class Program_weight_tensor_parameter_134: name = "parameter_134" + original_name = "linear_7.w_0" shape = [384, 96] dtype = "float32" min_val = float("-0.526145") @@ -1485,6 +1620,7 @@ class Program_weight_tensor_parameter_134: class Program_weight_tensor_parameter_135: name = "parameter_135" + original_name = "linear_6.b_0" shape = [384] dtype = "float32" min_val = float("-1.51315") @@ -1496,6 +1632,7 @@ class Program_weight_tensor_parameter_135: class Program_weight_tensor_parameter_136: name = "parameter_136" + original_name = "linear_6.w_0" shape = [96, 384] dtype = "float32" min_val = float("-0.375394") @@ -1507,6 +1644,7 @@ class Program_weight_tensor_parameter_136: class Program_weight_tensor_parameter_137: name = "parameter_137" + original_name = "layer_norm_4.b_0" shape = [96] dtype = "float32" min_val = float("-0.847986") @@ -1518,6 +1656,7 @@ class Program_weight_tensor_parameter_137: class Program_weight_tensor_parameter_138: name = "parameter_138" + original_name = "layer_norm_4.w_0" shape = [96] dtype = "float32" min_val = float("0.497374") @@ -1529,6 +1668,7 @@ class Program_weight_tensor_parameter_138: class Program_weight_tensor_parameter_139: name = "parameter_139" + original_name = "linear_5.b_0" shape = [96] dtype = "float32" min_val = float("-0.430658") @@ -1540,6 +1680,7 @@ class Program_weight_tensor_parameter_139: class Program_weight_tensor_parameter_140: name = "parameter_140" + original_name = "linear_5.w_0" shape = [96, 96] dtype = "float32" min_val = float("-0.282965") @@ -1551,6 +1692,7 @@ class Program_weight_tensor_parameter_140: class Program_weight_tensor_parameter_141: name = "parameter_141" + original_name = "linear_4.b_0" shape = [288] dtype = "float32" min_val = float("-1.68133") @@ -1562,6 +1704,7 @@ class Program_weight_tensor_parameter_141: class Program_weight_tensor_parameter_142: name = "parameter_142" + original_name = "linear_4.w_0" shape = [96, 288] dtype = "float32" min_val = float("-0.393352") @@ -1573,6 +1716,7 @@ class Program_weight_tensor_parameter_142: class Program_weight_tensor_parameter_143: name = "parameter_143" + original_name = "layer_norm_3.b_0" shape = [96] dtype = "float32" min_val = float("-1.30032") @@ -1584,6 +1728,7 @@ class Program_weight_tensor_parameter_143: class Program_weight_tensor_parameter_144: name = "parameter_144" + original_name = "layer_norm_3.w_0" shape = [96] dtype = "float32" min_val = float("0.307771") @@ -1595,6 +1740,7 @@ class Program_weight_tensor_parameter_144: class Program_weight_tensor_parameter_145: name = "parameter_145" + original_name = "linear_3.b_0" shape = [96] dtype = "float32" min_val = float("-1.97687") @@ -1606,6 +1752,7 @@ class Program_weight_tensor_parameter_145: class Program_weight_tensor_parameter_146: name = "parameter_146" + original_name = "linear_3.w_0" shape = [384, 96] dtype = "float32" min_val = float("-0.41369") @@ -1617,6 +1764,7 @@ class Program_weight_tensor_parameter_146: class Program_weight_tensor_parameter_147: name = "parameter_147" + original_name = "linear_2.b_0" shape = [384] dtype = "float32" min_val = float("-1.8828") @@ -1628,6 +1776,7 @@ class Program_weight_tensor_parameter_147: class Program_weight_tensor_parameter_148: name = "parameter_148" + original_name = "linear_2.w_0" shape = [96, 384] dtype = "float32" min_val = float("-0.453137") @@ -1639,6 +1788,7 @@ class Program_weight_tensor_parameter_148: class Program_weight_tensor_parameter_149: name = "parameter_149" + original_name = "layer_norm_2.b_0" shape = [96] dtype = "float32" min_val = float("-1.07431") @@ -1650,6 +1800,7 @@ class Program_weight_tensor_parameter_149: class Program_weight_tensor_parameter_150: name = "parameter_150" + original_name = "layer_norm_2.w_0" shape = [96] dtype = "float32" min_val = float("0.540165") @@ -1661,6 +1812,7 @@ class Program_weight_tensor_parameter_150: class Program_weight_tensor_parameter_151: name = "parameter_151" + original_name = "linear_1.b_0" shape = [96] dtype = "float32" min_val = float("-0.644725") @@ -1672,6 +1824,7 @@ class Program_weight_tensor_parameter_151: class Program_weight_tensor_parameter_152: name = "parameter_152" + original_name = "linear_1.w_0" shape = [96, 96] dtype = "float32" min_val = float("-0.280586") @@ -1683,6 +1836,7 @@ class Program_weight_tensor_parameter_152: class Program_weight_tensor_parameter_153: name = "parameter_153" + original_name = "linear_0.b_0" shape = [288] dtype = "float32" min_val = float("-1.78915") @@ -1694,6 +1848,7 @@ class Program_weight_tensor_parameter_153: class Program_weight_tensor_parameter_154: name = "parameter_154" + original_name = "linear_0.w_0" shape = [96, 288] dtype = "float32" min_val = float("-0.670016") @@ -1705,6 +1860,7 @@ class Program_weight_tensor_parameter_154: class Program_weight_tensor_parameter_155: name = "parameter_155" + original_name = "layer_norm_1.b_0" shape = [96] dtype = "float32" min_val = float("-0.775255") @@ -1716,6 +1872,7 @@ class Program_weight_tensor_parameter_155: class Program_weight_tensor_parameter_156: name = "parameter_156" + original_name = "layer_norm_1.w_0" shape = [96] dtype = "float32" min_val = float("-0.00252875") @@ -1727,6 +1884,7 @@ class Program_weight_tensor_parameter_156: class Program_weight_tensor_parameter_157: name = "parameter_157" + original_name = "layer_norm_0.b_0" shape = [96] dtype = "float32" min_val = float("-0.969281") @@ -1738,6 +1896,7 @@ class Program_weight_tensor_parameter_157: class Program_weight_tensor_parameter_158: name = "parameter_158" + original_name = "layer_norm_0.w_0" shape = [96] dtype = "float32" min_val = float("-0.0293355") @@ -1749,6 +1908,7 @@ class Program_weight_tensor_parameter_158: class Program_weight_tensor_parameter_159: name = "parameter_159" + original_name = "conv2d_0.b_0" shape = [96] dtype = "float32" min_val = float("-1.20207") @@ -1760,6 +1920,7 @@ class Program_weight_tensor_parameter_159: class Program_weight_tensor_parameter_160: name = "parameter_160" + original_name = "conv2d_0.w_0" shape = [96, 3, 4, 4] dtype = "float32" min_val = float("-0.312104") diff --git a/paddle_samples/PaddleX/ch_SVTRv2_rec/subgraph_10/input_meta.py b/paddle_samples/PaddleX/ch_SVTRv2_rec/subgraph_10/input_meta.py index 5c45bb35f..42abe955f 100644 --- a/paddle_samples/PaddleX/ch_SVTRv2_rec/subgraph_10/input_meta.py +++ b/paddle_samples/PaddleX/ch_SVTRv2_rec/subgraph_10/input_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_data_0: name = "data_0" + original_name = "var_2404" shape = [5, 384, 1, 40] dtype = "float32" min_val = float("-8.03003") diff --git a/paddle_samples/PaddleX/ch_SVTRv2_rec/subgraph_10/weight_meta.py b/paddle_samples/PaddleX/ch_SVTRv2_rec/subgraph_10/weight_meta.py index a2310ee69..ec06d75b0 100644 --- a/paddle_samples/PaddleX/ch_SVTRv2_rec/subgraph_10/weight_meta.py +++ b/paddle_samples/PaddleX/ch_SVTRv2_rec/subgraph_10/weight_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_parameter_0: name = "parameter_0" + original_name = "linear_64.b_0" shape = [6625] dtype = "float32" min_val = float("-1.46606") @@ -11,6 +12,7 @@ class Program_weight_tensor_parameter_0: class Program_weight_tensor_parameter_1: name = "parameter_1" + original_name = "linear_64.w_0" shape = [256, 6625] dtype = "float32" min_val = float("-0.842937") @@ -22,6 +24,7 @@ class Program_weight_tensor_parameter_1: class Program_weight_tensor_parameter_2: name = "parameter_2" + original_name = "batch_norm2d_6.b_0" shape = [256] dtype = "float32" min_val = float("-3.44711") @@ -33,6 +36,7 @@ class Program_weight_tensor_parameter_2: class Program_weight_tensor_parameter_3: name = "parameter_3" + original_name = "batch_norm2d_6.w_0" shape = [256] dtype = "float32" min_val = float("1.46858") @@ -44,6 +48,7 @@ class Program_weight_tensor_parameter_3: class Program_weight_tensor_parameter_4: name = "parameter_4" + original_name = "batch_norm2d_6.w_2" shape = [256] dtype = "float32" min_val = float("0.0225336") @@ -55,6 +60,7 @@ class Program_weight_tensor_parameter_4: class Program_weight_tensor_parameter_5: name = "parameter_5" + original_name = "batch_norm2d_6.w_1" shape = [256] dtype = "float32" min_val = float("-1.63944") @@ -66,6 +72,7 @@ class Program_weight_tensor_parameter_5: class Program_weight_tensor_parameter_6: name = "parameter_6" + original_name = "conv2d_16.w_0" shape = [256, 48, 1, 1] dtype = "float32" min_val = float("-0.346652") @@ -77,6 +84,7 @@ class Program_weight_tensor_parameter_6: class Program_weight_tensor_parameter_7: name = "parameter_7" + original_name = "batch_norm2d_5.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -86,6 +94,7 @@ class Program_weight_tensor_parameter_7: class Program_weight_tensor_parameter_8: name = "parameter_8" + original_name = "batch_norm2d_5.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -95,6 +104,7 @@ class Program_weight_tensor_parameter_8: class Program_weight_tensor_parameter_9: name = "parameter_9" + original_name = "batch_norm2d_5.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -104,6 +114,7 @@ class Program_weight_tensor_parameter_9: class Program_weight_tensor_parameter_10: name = "parameter_10" + original_name = "batch_norm2d_5.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -113,6 +124,7 @@ class Program_weight_tensor_parameter_10: class Program_weight_tensor_parameter_11: name = "parameter_11" + original_name = "conv2d_15.w_0" shape = [48, 768, 1, 3] dtype = "float32" min_val = float("-0.411824") @@ -124,6 +136,7 @@ class Program_weight_tensor_parameter_11: class Program_weight_tensor_parameter_12: name = "parameter_12" + original_name = "batch_norm2d_4.b_0" shape = [384] dtype = "float32" min_val = float("-2.40135") @@ -135,6 +148,7 @@ class Program_weight_tensor_parameter_12: class Program_weight_tensor_parameter_13: name = "parameter_13" + original_name = "batch_norm2d_4.w_0" shape = [384] dtype = "float32" min_val = float("0.392586") @@ -146,6 +160,7 @@ class Program_weight_tensor_parameter_13: class Program_weight_tensor_parameter_14: name = "parameter_14" + original_name = "batch_norm2d_4.w_2" shape = [384] dtype = "float32" min_val = float("0.185625") @@ -157,6 +172,7 @@ class Program_weight_tensor_parameter_14: class Program_weight_tensor_parameter_15: name = "parameter_15" + original_name = "batch_norm2d_4.w_1" shape = [384] dtype = "float32" min_val = float("-1.33882") @@ -168,6 +184,7 @@ class Program_weight_tensor_parameter_15: class Program_weight_tensor_parameter_16: name = "parameter_16" + original_name = "conv2d_14.w_0" shape = [384, 256, 1, 1] dtype = "float32" min_val = float("-0.334565") @@ -179,6 +196,7 @@ class Program_weight_tensor_parameter_16: class Program_weight_tensor_parameter_17: name = "parameter_17" + original_name = "layer_norm_42.b_0" shape = [256] dtype = "float32" min_val = float("-1.32656") @@ -190,6 +208,7 @@ class Program_weight_tensor_parameter_17: class Program_weight_tensor_parameter_18: name = "parameter_18" + original_name = "layer_norm_42.w_0" shape = [256] dtype = "float32" min_val = float("0.24478") @@ -201,6 +220,7 @@ class Program_weight_tensor_parameter_18: class Program_weight_tensor_parameter_19: name = "parameter_19" + original_name = "linear_63.b_0" shape = [256] dtype = "float32" min_val = float("-2.38568") @@ -212,6 +232,7 @@ class Program_weight_tensor_parameter_19: class Program_weight_tensor_parameter_20: name = "parameter_20" + original_name = "linear_63.w_0" shape = [512, 256] dtype = "float32" min_val = float("-0.770555") @@ -223,6 +244,7 @@ class Program_weight_tensor_parameter_20: class Program_weight_tensor_parameter_21: name = "parameter_21" + original_name = "linear_62.b_0" shape = [512] dtype = "float32" min_val = float("-1.72071") @@ -234,6 +256,7 @@ class Program_weight_tensor_parameter_21: class Program_weight_tensor_parameter_22: name = "parameter_22" + original_name = "linear_62.w_0" shape = [256, 512] dtype = "float32" min_val = float("-0.486813") @@ -245,6 +268,7 @@ class Program_weight_tensor_parameter_22: class Program_weight_tensor_parameter_23: name = "parameter_23" + original_name = "layer_norm_41.b_0" shape = [256] dtype = "float32" min_val = float("-0.830129") @@ -256,6 +280,7 @@ class Program_weight_tensor_parameter_23: class Program_weight_tensor_parameter_24: name = "parameter_24" + original_name = "layer_norm_41.w_0" shape = [256] dtype = "float32" min_val = float("0.515229") @@ -267,6 +292,7 @@ class Program_weight_tensor_parameter_24: class Program_weight_tensor_parameter_25: name = "parameter_25" + original_name = "linear_61.b_0" shape = [256] dtype = "float32" min_val = float("-0.910073") @@ -278,6 +304,7 @@ class Program_weight_tensor_parameter_25: class Program_weight_tensor_parameter_26: name = "parameter_26" + original_name = "linear_61.w_0" shape = [256, 256] dtype = "float32" min_val = float("-0.336543") @@ -289,6 +316,7 @@ class Program_weight_tensor_parameter_26: class Program_weight_tensor_parameter_27: name = "parameter_27" + original_name = "linear_60.b_0" shape = [768] dtype = "float32" min_val = float("-2.12229") @@ -300,6 +328,7 @@ class Program_weight_tensor_parameter_27: class Program_weight_tensor_parameter_28: name = "parameter_28" + original_name = "linear_60.w_0" shape = [256, 768] dtype = "float32" min_val = float("-0.52327") @@ -311,6 +340,7 @@ class Program_weight_tensor_parameter_28: class Program_weight_tensor_parameter_29: name = "parameter_29" + original_name = "layer_norm_40.b_0" shape = [256] dtype = "float32" min_val = float("-1.36519") @@ -322,6 +352,7 @@ class Program_weight_tensor_parameter_29: class Program_weight_tensor_parameter_30: name = "parameter_30" + original_name = "layer_norm_40.w_0" shape = [256] dtype = "float32" min_val = float("0.0883802") @@ -333,6 +364,7 @@ class Program_weight_tensor_parameter_30: class Program_weight_tensor_parameter_31: name = "parameter_31" + original_name = "linear_59.b_0" shape = [256] dtype = "float32" min_val = float("-1.50972") @@ -344,6 +376,7 @@ class Program_weight_tensor_parameter_31: class Program_weight_tensor_parameter_32: name = "parameter_32" + original_name = "linear_59.w_0" shape = [512, 256] dtype = "float32" min_val = float("-0.57403") @@ -355,6 +388,7 @@ class Program_weight_tensor_parameter_32: class Program_weight_tensor_parameter_33: name = "parameter_33" + original_name = "linear_58.b_0" shape = [512] dtype = "float32" min_val = float("-1.37099") @@ -366,6 +400,7 @@ class Program_weight_tensor_parameter_33: class Program_weight_tensor_parameter_34: name = "parameter_34" + original_name = "linear_58.w_0" shape = [256, 512] dtype = "float32" min_val = float("-0.809903") @@ -377,6 +412,7 @@ class Program_weight_tensor_parameter_34: class Program_weight_tensor_parameter_35: name = "parameter_35" + original_name = "layer_norm_39.b_0" shape = [256] dtype = "float32" min_val = float("-1.37513") @@ -388,6 +424,7 @@ class Program_weight_tensor_parameter_35: class Program_weight_tensor_parameter_36: name = "parameter_36" + original_name = "layer_norm_39.w_0" shape = [256] dtype = "float32" min_val = float("-0.00444758") @@ -399,6 +436,7 @@ class Program_weight_tensor_parameter_36: class Program_weight_tensor_parameter_37: name = "parameter_37" + original_name = "linear_57.b_0" shape = [256] dtype = "float32" min_val = float("-0.889222") @@ -410,6 +448,7 @@ class Program_weight_tensor_parameter_37: class Program_weight_tensor_parameter_38: name = "parameter_38" + original_name = "linear_57.w_0" shape = [256, 256] dtype = "float32" min_val = float("-0.295902") @@ -421,6 +460,7 @@ class Program_weight_tensor_parameter_38: class Program_weight_tensor_parameter_39: name = "parameter_39" + original_name = "linear_56.b_0" shape = [768] dtype = "float32" min_val = float("-1.65431") @@ -432,6 +472,7 @@ class Program_weight_tensor_parameter_39: class Program_weight_tensor_parameter_40: name = "parameter_40" + original_name = "linear_56.w_0" shape = [256, 768] dtype = "float32" min_val = float("-0.330457") @@ -443,6 +484,7 @@ class Program_weight_tensor_parameter_40: class Program_weight_tensor_parameter_41: name = "parameter_41" + original_name = "layer_norm_38.b_0" shape = [256] dtype = "float32" min_val = float("-0.627416") @@ -454,6 +496,7 @@ class Program_weight_tensor_parameter_41: class Program_weight_tensor_parameter_42: name = "parameter_42" + original_name = "layer_norm_38.w_0" shape = [256] dtype = "float32" min_val = float("-0.239407") @@ -465,6 +508,7 @@ class Program_weight_tensor_parameter_42: class Program_weight_tensor_parameter_43: name = "parameter_43" + original_name = "batch_norm2d_3.b_0" shape = [256] dtype = "float32" min_val = float("-1.8788") @@ -476,6 +520,7 @@ class Program_weight_tensor_parameter_43: class Program_weight_tensor_parameter_44: name = "parameter_44" + original_name = "batch_norm2d_3.w_0" shape = [256] dtype = "float32" min_val = float("0.545192") @@ -487,6 +532,7 @@ class Program_weight_tensor_parameter_44: class Program_weight_tensor_parameter_45: name = "parameter_45" + original_name = "batch_norm2d_3.w_2" shape = [256] dtype = "float32" min_val = float("0.0283431") @@ -498,6 +544,7 @@ class Program_weight_tensor_parameter_45: class Program_weight_tensor_parameter_46: name = "parameter_46" + original_name = "batch_norm2d_3.w_1" shape = [256] dtype = "float32" min_val = float("-2.46669") @@ -509,6 +556,7 @@ class Program_weight_tensor_parameter_46: class Program_weight_tensor_parameter_47: name = "parameter_47" + original_name = "conv2d_13.w_0" shape = [256, 48, 1, 1] dtype = "float32" min_val = float("-0.397741") @@ -520,6 +568,7 @@ class Program_weight_tensor_parameter_47: class Program_weight_tensor_parameter_48: name = "parameter_48" + original_name = "batch_norm2d_2.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -529,6 +578,7 @@ class Program_weight_tensor_parameter_48: class Program_weight_tensor_parameter_49: name = "parameter_49" + original_name = "batch_norm2d_2.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -538,6 +588,7 @@ class Program_weight_tensor_parameter_49: class Program_weight_tensor_parameter_50: name = "parameter_50" + original_name = "batch_norm2d_2.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -547,6 +598,7 @@ class Program_weight_tensor_parameter_50: class Program_weight_tensor_parameter_51: name = "parameter_51" + original_name = "batch_norm2d_2.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -556,6 +608,7 @@ class Program_weight_tensor_parameter_51: class Program_weight_tensor_parameter_52: name = "parameter_52" + original_name = "conv2d_12.w_0" shape = [48, 384, 1, 3] dtype = "float32" min_val = float("-0.496082") diff --git a/paddle_samples/PaddleX/ch_SVTRv2_rec/subgraph_11/input_meta.py b/paddle_samples/PaddleX/ch_SVTRv2_rec/subgraph_11/input_meta.py index 4493a6925..6b2a93886 100644 --- a/paddle_samples/PaddleX/ch_SVTRv2_rec/subgraph_11/input_meta.py +++ b/paddle_samples/PaddleX/ch_SVTRv2_rec/subgraph_11/input_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_data_0: name = "data_0" + original_name = "args_0" shape = [] dtype = "int64" data = [4] @@ -7,6 +8,7 @@ class Program_weight_tensor_data_0: class Program_weight_tensor_data_1: name = "data_1" + original_name = "args_1" shape = [] dtype = "int64" data = [256] @@ -14,6 +16,7 @@ class Program_weight_tensor_data_1: class Program_weight_tensor_data_2: name = "data_2" + original_name = "args_2" shape = [] dtype = "int64" data = [4] @@ -21,6 +24,7 @@ class Program_weight_tensor_data_2: class Program_weight_tensor_data_3: name = "data_3" + original_name = "var_3516" shape = [8, 256, 4, 80] dtype = "float32" min_val = float("-5.68937") diff --git a/paddle_samples/PaddleX/ch_SVTRv2_rec/subgraph_11/weight_meta.py b/paddle_samples/PaddleX/ch_SVTRv2_rec/subgraph_11/weight_meta.py index 2dcb644c9..ee3f9a4ab 100644 --- a/paddle_samples/PaddleX/ch_SVTRv2_rec/subgraph_11/weight_meta.py +++ b/paddle_samples/PaddleX/ch_SVTRv2_rec/subgraph_11/weight_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_parameter_0: name = "parameter_0" + original_name = "layer_norm_25.b_0" shape = [384] dtype = "float32" min_val = float("-0.611573") @@ -11,6 +12,7 @@ class Program_weight_tensor_parameter_0: class Program_weight_tensor_parameter_1: name = "parameter_1" + original_name = "layer_norm_25.w_0" shape = [384] dtype = "float32" min_val = float("-0.0176942") @@ -22,6 +24,7 @@ class Program_weight_tensor_parameter_1: class Program_weight_tensor_parameter_2: name = "parameter_2" + original_name = "conv2d_11.b_0" shape = [384] dtype = "float32" min_val = float("-2.13167") @@ -33,6 +36,7 @@ class Program_weight_tensor_parameter_2: class Program_weight_tensor_parameter_3: name = "parameter_3" + original_name = "conv2d_11.w_0" shape = [384, 256, 3, 3] dtype = "float32" min_val = float("-4.81624") @@ -44,6 +48,7 @@ class Program_weight_tensor_parameter_3: class Program_weight_tensor_parameter_4: name = "parameter_4" + original_name = "layer_norm_24.b_0" shape = [256] dtype = "float32" min_val = float("-1.00302") @@ -55,6 +60,7 @@ class Program_weight_tensor_parameter_4: class Program_weight_tensor_parameter_5: name = "parameter_5" + original_name = "layer_norm_24.w_0" shape = [256] dtype = "float32" min_val = float("0.22568") @@ -66,6 +72,7 @@ class Program_weight_tensor_parameter_5: class Program_weight_tensor_parameter_6: name = "parameter_6" + original_name = "linear_31.b_0" shape = [256] dtype = "float32" min_val = float("-8.09258") @@ -77,6 +84,7 @@ class Program_weight_tensor_parameter_6: class Program_weight_tensor_parameter_7: name = "parameter_7" + original_name = "linear_31.w_0" shape = [1024, 256] dtype = "float32" min_val = float("-1.43969") @@ -88,6 +96,7 @@ class Program_weight_tensor_parameter_7: class Program_weight_tensor_parameter_8: name = "parameter_8" + original_name = "linear_30.b_0" shape = [1024] dtype = "float32" min_val = float("-1.46189") @@ -99,6 +108,7 @@ class Program_weight_tensor_parameter_8: class Program_weight_tensor_parameter_9: name = "parameter_9" + original_name = "linear_30.w_0" shape = [256, 1024] dtype = "float32" min_val = float("-0.347586") @@ -110,6 +120,7 @@ class Program_weight_tensor_parameter_9: class Program_weight_tensor_parameter_10: name = "parameter_10" + original_name = "layer_norm_23.b_0" shape = [256] dtype = "float32" min_val = float("-3.06699") @@ -121,6 +132,7 @@ class Program_weight_tensor_parameter_10: class Program_weight_tensor_parameter_11: name = "parameter_11" + original_name = "layer_norm_23.w_0" shape = [256] dtype = "float32" min_val = float("0.112453") @@ -132,6 +144,7 @@ class Program_weight_tensor_parameter_11: class Program_weight_tensor_parameter_12: name = "parameter_12" + original_name = "linear_29.b_0" shape = [256] dtype = "float32" min_val = float("-1.23851") @@ -143,6 +156,7 @@ class Program_weight_tensor_parameter_12: class Program_weight_tensor_parameter_13: name = "parameter_13" + original_name = "linear_29.w_0" shape = [256, 256] dtype = "float32" min_val = float("-0.324324") @@ -154,6 +168,7 @@ class Program_weight_tensor_parameter_13: class Program_weight_tensor_parameter_14: name = "parameter_14" + original_name = "linear_28.b_0" shape = [768] dtype = "float32" min_val = float("-3.23557") @@ -165,6 +180,7 @@ class Program_weight_tensor_parameter_14: class Program_weight_tensor_parameter_15: name = "parameter_15" + original_name = "linear_28.w_0" shape = [256, 768] dtype = "float32" min_val = float("-0.604016") @@ -176,6 +192,7 @@ class Program_weight_tensor_parameter_15: class Program_weight_tensor_parameter_16: name = "parameter_16" + original_name = "layer_norm_22.b_0" shape = [256] dtype = "float32" min_val = float("-1.03089") @@ -187,6 +204,7 @@ class Program_weight_tensor_parameter_16: class Program_weight_tensor_parameter_17: name = "parameter_17" + original_name = "layer_norm_22.w_0" shape = [256] dtype = "float32" min_val = float("0.287937") @@ -198,6 +216,7 @@ class Program_weight_tensor_parameter_17: class Program_weight_tensor_parameter_18: name = "parameter_18" + original_name = "linear_27.b_0" shape = [256] dtype = "float32" min_val = float("-2.05805") @@ -209,6 +228,7 @@ class Program_weight_tensor_parameter_18: class Program_weight_tensor_parameter_19: name = "parameter_19" + original_name = "linear_27.w_0" shape = [1024, 256] dtype = "float32" min_val = float("-0.562583") @@ -220,6 +240,7 @@ class Program_weight_tensor_parameter_19: class Program_weight_tensor_parameter_20: name = "parameter_20" + original_name = "linear_26.b_0" shape = [1024] dtype = "float32" min_val = float("-1.33321") @@ -231,6 +252,7 @@ class Program_weight_tensor_parameter_20: class Program_weight_tensor_parameter_21: name = "parameter_21" + original_name = "linear_26.w_0" shape = [256, 1024] dtype = "float32" min_val = float("-0.381275") @@ -242,6 +264,7 @@ class Program_weight_tensor_parameter_21: class Program_weight_tensor_parameter_22: name = "parameter_22" + original_name = "layer_norm_21.b_0" shape = [256] dtype = "float32" min_val = float("-2.22319") @@ -253,6 +276,7 @@ class Program_weight_tensor_parameter_22: class Program_weight_tensor_parameter_23: name = "parameter_23" + original_name = "layer_norm_21.w_0" shape = [256] dtype = "float32" min_val = float("0.081885") @@ -264,6 +288,7 @@ class Program_weight_tensor_parameter_23: class Program_weight_tensor_parameter_24: name = "parameter_24" + original_name = "linear_25.b_0" shape = [256] dtype = "float32" min_val = float("-1.62126") @@ -275,6 +300,7 @@ class Program_weight_tensor_parameter_24: class Program_weight_tensor_parameter_25: name = "parameter_25" + original_name = "linear_25.w_0" shape = [256, 256] dtype = "float32" min_val = float("-0.486573") @@ -286,6 +312,7 @@ class Program_weight_tensor_parameter_25: class Program_weight_tensor_parameter_26: name = "parameter_26" + original_name = "linear_24.b_0" shape = [768] dtype = "float32" min_val = float("-2.94557") @@ -297,6 +324,7 @@ class Program_weight_tensor_parameter_26: class Program_weight_tensor_parameter_27: name = "parameter_27" + original_name = "linear_24.w_0" shape = [256, 768] dtype = "float32" min_val = float("-0.415111") @@ -308,6 +336,7 @@ class Program_weight_tensor_parameter_27: class Program_weight_tensor_parameter_28: name = "parameter_28" + original_name = "layer_norm_20.b_0" shape = [256] dtype = "float32" min_val = float("-1.52143") @@ -319,6 +348,7 @@ class Program_weight_tensor_parameter_28: class Program_weight_tensor_parameter_29: name = "parameter_29" + original_name = "layer_norm_20.w_0" shape = [256] dtype = "float32" min_val = float("0.263449") @@ -330,6 +360,7 @@ class Program_weight_tensor_parameter_29: class Program_weight_tensor_parameter_30: name = "parameter_30" + original_name = "linear_23.b_0" shape = [256] dtype = "float32" min_val = float("-0.74461") @@ -341,6 +372,7 @@ class Program_weight_tensor_parameter_30: class Program_weight_tensor_parameter_31: name = "parameter_31" + original_name = "linear_23.w_0" shape = [1024, 256] dtype = "float32" min_val = float("-0.504825") @@ -352,6 +384,7 @@ class Program_weight_tensor_parameter_31: class Program_weight_tensor_parameter_32: name = "parameter_32" + original_name = "linear_22.b_0" shape = [1024] dtype = "float32" min_val = float("-1.1179") @@ -363,6 +396,7 @@ class Program_weight_tensor_parameter_32: class Program_weight_tensor_parameter_33: name = "parameter_33" + original_name = "linear_22.w_0" shape = [256, 1024] dtype = "float32" min_val = float("-0.477723") @@ -374,6 +408,7 @@ class Program_weight_tensor_parameter_33: class Program_weight_tensor_parameter_34: name = "parameter_34" + original_name = "layer_norm_19.b_0" shape = [256] dtype = "float32" min_val = float("-2.66275") @@ -385,6 +420,7 @@ class Program_weight_tensor_parameter_34: class Program_weight_tensor_parameter_35: name = "parameter_35" + original_name = "layer_norm_19.w_0" shape = [256] dtype = "float32" min_val = float("0.0184295") @@ -396,6 +432,7 @@ class Program_weight_tensor_parameter_35: class Program_weight_tensor_parameter_36: name = "parameter_36" + original_name = "linear_21.b_0" shape = [256] dtype = "float32" min_val = float("-1.94903") @@ -407,6 +444,7 @@ class Program_weight_tensor_parameter_36: class Program_weight_tensor_parameter_37: name = "parameter_37" + original_name = "linear_21.w_0" shape = [256, 256] dtype = "float32" min_val = float("-0.334037") @@ -418,6 +456,7 @@ class Program_weight_tensor_parameter_37: class Program_weight_tensor_parameter_38: name = "parameter_38" + original_name = "linear_20.b_0" shape = [768] dtype = "float32" min_val = float("-3.24135") @@ -429,6 +468,7 @@ class Program_weight_tensor_parameter_38: class Program_weight_tensor_parameter_39: name = "parameter_39" + original_name = "linear_20.w_0" shape = [256, 768] dtype = "float32" min_val = float("-0.484463") @@ -440,6 +480,7 @@ class Program_weight_tensor_parameter_39: class Program_weight_tensor_parameter_40: name = "parameter_40" + original_name = "layer_norm_18.b_0" shape = [256] dtype = "float32" min_val = float("-1.22621") @@ -451,6 +492,7 @@ class Program_weight_tensor_parameter_40: class Program_weight_tensor_parameter_41: name = "parameter_41" + original_name = "layer_norm_18.w_0" shape = [256] dtype = "float32" min_val = float("0.0741299") @@ -462,6 +504,7 @@ class Program_weight_tensor_parameter_41: class Program_weight_tensor_parameter_42: name = "parameter_42" + original_name = "linear_19.b_0" shape = [256] dtype = "float32" min_val = float("-0.594821") @@ -473,6 +516,7 @@ class Program_weight_tensor_parameter_42: class Program_weight_tensor_parameter_43: name = "parameter_43" + original_name = "linear_19.w_0" shape = [1024, 256] dtype = "float32" min_val = float("-0.448166") @@ -484,6 +528,7 @@ class Program_weight_tensor_parameter_43: class Program_weight_tensor_parameter_44: name = "parameter_44" + original_name = "linear_18.b_0" shape = [1024] dtype = "float32" min_val = float("-1.13055") @@ -495,6 +540,7 @@ class Program_weight_tensor_parameter_44: class Program_weight_tensor_parameter_45: name = "parameter_45" + original_name = "linear_18.w_0" shape = [256, 1024] dtype = "float32" min_val = float("-0.338649") @@ -506,6 +552,7 @@ class Program_weight_tensor_parameter_45: class Program_weight_tensor_parameter_46: name = "parameter_46" + original_name = "layer_norm_17.b_0" shape = [256] dtype = "float32" min_val = float("-2.27003") @@ -517,6 +564,7 @@ class Program_weight_tensor_parameter_46: class Program_weight_tensor_parameter_47: name = "parameter_47" + original_name = "layer_norm_17.w_0" shape = [256] dtype = "float32" min_val = float("0.085421") @@ -528,6 +576,7 @@ class Program_weight_tensor_parameter_47: class Program_weight_tensor_parameter_48: name = "parameter_48" + original_name = "linear_17.b_0" shape = [256] dtype = "float32" min_val = float("-1.86822") @@ -539,6 +588,7 @@ class Program_weight_tensor_parameter_48: class Program_weight_tensor_parameter_49: name = "parameter_49" + original_name = "linear_17.w_0" shape = [256, 256] dtype = "float32" min_val = float("-0.429145") @@ -550,6 +600,7 @@ class Program_weight_tensor_parameter_49: class Program_weight_tensor_parameter_50: name = "parameter_50" + original_name = "linear_16.b_0" shape = [768] dtype = "float32" min_val = float("-2.90161") @@ -561,6 +612,7 @@ class Program_weight_tensor_parameter_50: class Program_weight_tensor_parameter_51: name = "parameter_51" + original_name = "linear_16.w_0" shape = [256, 768] dtype = "float32" min_val = float("-0.307903") @@ -572,6 +624,7 @@ class Program_weight_tensor_parameter_51: class Program_weight_tensor_parameter_52: name = "parameter_52" + original_name = "layer_norm_16.b_0" shape = [256] dtype = "float32" min_val = float("-1.72865") @@ -583,6 +636,7 @@ class Program_weight_tensor_parameter_52: class Program_weight_tensor_parameter_53: name = "parameter_53" + original_name = "layer_norm_16.w_0" shape = [256] dtype = "float32" min_val = float("0.295198") @@ -594,6 +648,7 @@ class Program_weight_tensor_parameter_53: class Program_weight_tensor_parameter_54: name = "parameter_54" + original_name = "linear_15.b_0" shape = [256] dtype = "float32" min_val = float("-0.541449") @@ -605,6 +660,7 @@ class Program_weight_tensor_parameter_54: class Program_weight_tensor_parameter_55: name = "parameter_55" + original_name = "linear_15.w_0" shape = [1024, 256] dtype = "float32" min_val = float("-0.438234") @@ -616,6 +672,7 @@ class Program_weight_tensor_parameter_55: class Program_weight_tensor_parameter_56: name = "parameter_56" + original_name = "linear_14.b_0" shape = [1024] dtype = "float32" min_val = float("-0.869132") @@ -627,6 +684,7 @@ class Program_weight_tensor_parameter_56: class Program_weight_tensor_parameter_57: name = "parameter_57" + original_name = "linear_14.w_0" shape = [256, 1024] dtype = "float32" min_val = float("-0.622617") @@ -638,6 +696,7 @@ class Program_weight_tensor_parameter_57: class Program_weight_tensor_parameter_58: name = "parameter_58" + original_name = "layer_norm_15.b_0" shape = [256] dtype = "float32" min_val = float("-1.54389") @@ -649,6 +708,7 @@ class Program_weight_tensor_parameter_58: class Program_weight_tensor_parameter_59: name = "parameter_59" + original_name = "layer_norm_15.w_0" shape = [256] dtype = "float32" min_val = float("0.25324") @@ -660,6 +720,7 @@ class Program_weight_tensor_parameter_59: class Program_weight_tensor_parameter_60: name = "parameter_60" + original_name = "conv2d_10.b_0" shape = [256] dtype = "float32" min_val = float("-2.27709") @@ -671,6 +732,7 @@ class Program_weight_tensor_parameter_60: class Program_weight_tensor_parameter_61: name = "parameter_61" + original_name = "conv2d_10.w_0" shape = [256, 32, 5, 5] dtype = "float32" min_val = float("-0.271916") @@ -682,6 +744,7 @@ class Program_weight_tensor_parameter_61: class Program_weight_tensor_parameter_62: name = "parameter_62" + original_name = "layer_norm_14.b_0" shape = [256] dtype = "float32" min_val = float("-0.633171") @@ -693,6 +756,7 @@ class Program_weight_tensor_parameter_62: class Program_weight_tensor_parameter_63: name = "parameter_63" + original_name = "layer_norm_14.w_0" shape = [256] dtype = "float32" min_val = float("0.0638512") @@ -704,6 +768,7 @@ class Program_weight_tensor_parameter_63: class Program_weight_tensor_parameter_64: name = "parameter_64" + original_name = "linear_13.b_0" shape = [256] dtype = "float32" min_val = float("-8.93611") @@ -715,6 +780,7 @@ class Program_weight_tensor_parameter_64: class Program_weight_tensor_parameter_65: name = "parameter_65" + original_name = "linear_13.w_0" shape = [1024, 256] dtype = "float32" min_val = float("-1.32594") @@ -726,6 +792,7 @@ class Program_weight_tensor_parameter_65: class Program_weight_tensor_parameter_66: name = "parameter_66" + original_name = "linear_12.b_0" shape = [1024] dtype = "float32" min_val = float("-1.46672") @@ -737,6 +804,7 @@ class Program_weight_tensor_parameter_66: class Program_weight_tensor_parameter_67: name = "parameter_67" + original_name = "linear_12.w_0" shape = [256, 1024] dtype = "float32" min_val = float("-0.324049") @@ -748,6 +816,7 @@ class Program_weight_tensor_parameter_67: class Program_weight_tensor_parameter_68: name = "parameter_68" + original_name = "layer_norm_13.b_0" shape = [256] dtype = "float32" min_val = float("-0.661923") @@ -759,6 +828,7 @@ class Program_weight_tensor_parameter_68: class Program_weight_tensor_parameter_69: name = "parameter_69" + original_name = "layer_norm_13.w_0" shape = [256] dtype = "float32" min_val = float("0.404883") @@ -770,6 +840,7 @@ class Program_weight_tensor_parameter_69: class Program_weight_tensor_parameter_70: name = "parameter_70" + original_name = "conv2d_9.b_0" shape = [256] dtype = "float32" min_val = float("-2.16229") @@ -781,6 +852,7 @@ class Program_weight_tensor_parameter_70: class Program_weight_tensor_parameter_71: name = "parameter_71" + original_name = "conv2d_9.w_0" shape = [256, 32, 5, 5] dtype = "float32" min_val = float("-0.265171") diff --git a/paddle_samples/PaddleX/ch_SVTRv2_rec/subgraph_13/input_meta.py b/paddle_samples/PaddleX/ch_SVTRv2_rec/subgraph_13/input_meta.py index 24f33cd65..462580a96 100644 --- a/paddle_samples/PaddleX/ch_SVTRv2_rec/subgraph_13/input_meta.py +++ b/paddle_samples/PaddleX/ch_SVTRv2_rec/subgraph_13/input_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_data_0: name = "data_0" + original_name = "var_737" shape = [4, 384, 1, 40] dtype = "float32" min_val = float("-6.75135") diff --git a/paddle_samples/PaddleX/ch_SVTRv2_rec/subgraph_13/weight_meta.py b/paddle_samples/PaddleX/ch_SVTRv2_rec/subgraph_13/weight_meta.py index 16a4561a2..ec01882a5 100644 --- a/paddle_samples/PaddleX/ch_SVTRv2_rec/subgraph_13/weight_meta.py +++ b/paddle_samples/PaddleX/ch_SVTRv2_rec/subgraph_13/weight_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_parameter_0: name = "parameter_0" + original_name = "linear_64.b_0" shape = [6625] dtype = "float32" min_val = float("-1.46606") @@ -11,6 +12,7 @@ class Program_weight_tensor_parameter_0: class Program_weight_tensor_parameter_1: name = "parameter_1" + original_name = "linear_64.w_0" shape = [256, 6625] dtype = "float32" min_val = float("-0.842937") @@ -22,6 +24,7 @@ class Program_weight_tensor_parameter_1: class Program_weight_tensor_parameter_2: name = "parameter_2" + original_name = "batch_norm2d_6.b_0" shape = [256] dtype = "float32" min_val = float("-3.44711") @@ -33,6 +36,7 @@ class Program_weight_tensor_parameter_2: class Program_weight_tensor_parameter_3: name = "parameter_3" + original_name = "batch_norm2d_6.w_0" shape = [256] dtype = "float32" min_val = float("1.46858") @@ -44,6 +48,7 @@ class Program_weight_tensor_parameter_3: class Program_weight_tensor_parameter_4: name = "parameter_4" + original_name = "batch_norm2d_6.w_2" shape = [256] dtype = "float32" min_val = float("0.0231144") @@ -55,6 +60,7 @@ class Program_weight_tensor_parameter_4: class Program_weight_tensor_parameter_5: name = "parameter_5" + original_name = "batch_norm2d_6.w_1" shape = [256] dtype = "float32" min_val = float("-1.63971") @@ -66,6 +72,7 @@ class Program_weight_tensor_parameter_5: class Program_weight_tensor_parameter_6: name = "parameter_6" + original_name = "conv2d_16.w_0" shape = [256, 48, 1, 1] dtype = "float32" min_val = float("-0.346652") @@ -77,6 +84,7 @@ class Program_weight_tensor_parameter_6: class Program_weight_tensor_parameter_7: name = "parameter_7" + original_name = "batch_norm2d_5.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -86,6 +94,7 @@ class Program_weight_tensor_parameter_7: class Program_weight_tensor_parameter_8: name = "parameter_8" + original_name = "batch_norm2d_5.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -95,6 +104,7 @@ class Program_weight_tensor_parameter_8: class Program_weight_tensor_parameter_9: name = "parameter_9" + original_name = "batch_norm2d_5.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -104,6 +114,7 @@ class Program_weight_tensor_parameter_9: class Program_weight_tensor_parameter_10: name = "parameter_10" + original_name = "batch_norm2d_5.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -113,6 +124,7 @@ class Program_weight_tensor_parameter_10: class Program_weight_tensor_parameter_11: name = "parameter_11" + original_name = "conv2d_15.w_0" shape = [48, 768, 1, 3] dtype = "float32" min_val = float("-0.411824") @@ -124,6 +136,7 @@ class Program_weight_tensor_parameter_11: class Program_weight_tensor_parameter_12: name = "parameter_12" + original_name = "batch_norm2d_4.b_0" shape = [384] dtype = "float32" min_val = float("-2.40135") @@ -135,6 +148,7 @@ class Program_weight_tensor_parameter_12: class Program_weight_tensor_parameter_13: name = "parameter_13" + original_name = "batch_norm2d_4.w_0" shape = [384] dtype = "float32" min_val = float("0.392586") @@ -146,6 +160,7 @@ class Program_weight_tensor_parameter_13: class Program_weight_tensor_parameter_14: name = "parameter_14" + original_name = "batch_norm2d_4.w_2" shape = [384] dtype = "float32" min_val = float("0.186257") @@ -157,6 +172,7 @@ class Program_weight_tensor_parameter_14: class Program_weight_tensor_parameter_15: name = "parameter_15" + original_name = "batch_norm2d_4.w_1" shape = [384] dtype = "float32" min_val = float("-1.35379") @@ -168,6 +184,7 @@ class Program_weight_tensor_parameter_15: class Program_weight_tensor_parameter_16: name = "parameter_16" + original_name = "conv2d_14.w_0" shape = [384, 256, 1, 1] dtype = "float32" min_val = float("-0.334565") @@ -179,6 +196,7 @@ class Program_weight_tensor_parameter_16: class Program_weight_tensor_parameter_17: name = "parameter_17" + original_name = "layer_norm_42.b_0" shape = [256] dtype = "float32" min_val = float("-1.32656") @@ -190,6 +208,7 @@ class Program_weight_tensor_parameter_17: class Program_weight_tensor_parameter_18: name = "parameter_18" + original_name = "layer_norm_42.w_0" shape = [256] dtype = "float32" min_val = float("0.24478") @@ -201,6 +220,7 @@ class Program_weight_tensor_parameter_18: class Program_weight_tensor_parameter_19: name = "parameter_19" + original_name = "linear_63.b_0" shape = [256] dtype = "float32" min_val = float("-2.38568") @@ -212,6 +232,7 @@ class Program_weight_tensor_parameter_19: class Program_weight_tensor_parameter_20: name = "parameter_20" + original_name = "linear_63.w_0" shape = [512, 256] dtype = "float32" min_val = float("-0.770555") @@ -223,6 +244,7 @@ class Program_weight_tensor_parameter_20: class Program_weight_tensor_parameter_21: name = "parameter_21" + original_name = "linear_62.b_0" shape = [512] dtype = "float32" min_val = float("-1.72071") @@ -234,6 +256,7 @@ class Program_weight_tensor_parameter_21: class Program_weight_tensor_parameter_22: name = "parameter_22" + original_name = "linear_62.w_0" shape = [256, 512] dtype = "float32" min_val = float("-0.486813") @@ -245,6 +268,7 @@ class Program_weight_tensor_parameter_22: class Program_weight_tensor_parameter_23: name = "parameter_23" + original_name = "layer_norm_41.b_0" shape = [256] dtype = "float32" min_val = float("-0.830129") @@ -256,6 +280,7 @@ class Program_weight_tensor_parameter_23: class Program_weight_tensor_parameter_24: name = "parameter_24" + original_name = "layer_norm_41.w_0" shape = [256] dtype = "float32" min_val = float("0.515229") @@ -267,6 +292,7 @@ class Program_weight_tensor_parameter_24: class Program_weight_tensor_parameter_25: name = "parameter_25" + original_name = "linear_61.b_0" shape = [256] dtype = "float32" min_val = float("-0.910073") @@ -278,6 +304,7 @@ class Program_weight_tensor_parameter_25: class Program_weight_tensor_parameter_26: name = "parameter_26" + original_name = "linear_61.w_0" shape = [256, 256] dtype = "float32" min_val = float("-0.336543") @@ -289,6 +316,7 @@ class Program_weight_tensor_parameter_26: class Program_weight_tensor_parameter_27: name = "parameter_27" + original_name = "linear_60.b_0" shape = [768] dtype = "float32" min_val = float("-2.12229") @@ -300,6 +328,7 @@ class Program_weight_tensor_parameter_27: class Program_weight_tensor_parameter_28: name = "parameter_28" + original_name = "linear_60.w_0" shape = [256, 768] dtype = "float32" min_val = float("-0.52327") @@ -311,6 +340,7 @@ class Program_weight_tensor_parameter_28: class Program_weight_tensor_parameter_29: name = "parameter_29" + original_name = "layer_norm_40.b_0" shape = [256] dtype = "float32" min_val = float("-1.36519") @@ -322,6 +352,7 @@ class Program_weight_tensor_parameter_29: class Program_weight_tensor_parameter_30: name = "parameter_30" + original_name = "layer_norm_40.w_0" shape = [256] dtype = "float32" min_val = float("0.0883802") @@ -333,6 +364,7 @@ class Program_weight_tensor_parameter_30: class Program_weight_tensor_parameter_31: name = "parameter_31" + original_name = "linear_59.b_0" shape = [256] dtype = "float32" min_val = float("-1.50972") @@ -344,6 +376,7 @@ class Program_weight_tensor_parameter_31: class Program_weight_tensor_parameter_32: name = "parameter_32" + original_name = "linear_59.w_0" shape = [512, 256] dtype = "float32" min_val = float("-0.57403") @@ -355,6 +388,7 @@ class Program_weight_tensor_parameter_32: class Program_weight_tensor_parameter_33: name = "parameter_33" + original_name = "linear_58.b_0" shape = [512] dtype = "float32" min_val = float("-1.37099") @@ -366,6 +400,7 @@ class Program_weight_tensor_parameter_33: class Program_weight_tensor_parameter_34: name = "parameter_34" + original_name = "linear_58.w_0" shape = [256, 512] dtype = "float32" min_val = float("-0.809903") @@ -377,6 +412,7 @@ class Program_weight_tensor_parameter_34: class Program_weight_tensor_parameter_35: name = "parameter_35" + original_name = "layer_norm_39.b_0" shape = [256] dtype = "float32" min_val = float("-1.37513") @@ -388,6 +424,7 @@ class Program_weight_tensor_parameter_35: class Program_weight_tensor_parameter_36: name = "parameter_36" + original_name = "layer_norm_39.w_0" shape = [256] dtype = "float32" min_val = float("-0.00444758") @@ -399,6 +436,7 @@ class Program_weight_tensor_parameter_36: class Program_weight_tensor_parameter_37: name = "parameter_37" + original_name = "linear_57.b_0" shape = [256] dtype = "float32" min_val = float("-0.889222") @@ -410,6 +448,7 @@ class Program_weight_tensor_parameter_37: class Program_weight_tensor_parameter_38: name = "parameter_38" + original_name = "linear_57.w_0" shape = [256, 256] dtype = "float32" min_val = float("-0.295902") @@ -421,6 +460,7 @@ class Program_weight_tensor_parameter_38: class Program_weight_tensor_parameter_39: name = "parameter_39" + original_name = "linear_56.b_0" shape = [768] dtype = "float32" min_val = float("-1.65431") @@ -432,6 +472,7 @@ class Program_weight_tensor_parameter_39: class Program_weight_tensor_parameter_40: name = "parameter_40" + original_name = "linear_56.w_0" shape = [256, 768] dtype = "float32" min_val = float("-0.330457") @@ -443,6 +484,7 @@ class Program_weight_tensor_parameter_40: class Program_weight_tensor_parameter_41: name = "parameter_41" + original_name = "layer_norm_38.b_0" shape = [256] dtype = "float32" min_val = float("-0.627416") @@ -454,6 +496,7 @@ class Program_weight_tensor_parameter_41: class Program_weight_tensor_parameter_42: name = "parameter_42" + original_name = "layer_norm_38.w_0" shape = [256] dtype = "float32" min_val = float("-0.239407") @@ -465,6 +508,7 @@ class Program_weight_tensor_parameter_42: class Program_weight_tensor_parameter_43: name = "parameter_43" + original_name = "batch_norm2d_3.b_0" shape = [256] dtype = "float32" min_val = float("-1.8788") @@ -476,6 +520,7 @@ class Program_weight_tensor_parameter_43: class Program_weight_tensor_parameter_44: name = "parameter_44" + original_name = "batch_norm2d_3.w_0" shape = [256] dtype = "float32" min_val = float("0.545192") @@ -487,6 +532,7 @@ class Program_weight_tensor_parameter_44: class Program_weight_tensor_parameter_45: name = "parameter_45" + original_name = "batch_norm2d_3.w_2" shape = [256] dtype = "float32" min_val = float("0.0291677") @@ -498,6 +544,7 @@ class Program_weight_tensor_parameter_45: class Program_weight_tensor_parameter_46: name = "parameter_46" + original_name = "batch_norm2d_3.w_1" shape = [256] dtype = "float32" min_val = float("-2.4665") @@ -509,6 +556,7 @@ class Program_weight_tensor_parameter_46: class Program_weight_tensor_parameter_47: name = "parameter_47" + original_name = "conv2d_13.w_0" shape = [256, 48, 1, 1] dtype = "float32" min_val = float("-0.397741") @@ -520,6 +568,7 @@ class Program_weight_tensor_parameter_47: class Program_weight_tensor_parameter_48: name = "parameter_48" + original_name = "batch_norm2d_2.b_0" shape = [48] dtype = "float32" min_val = float("0") @@ -529,6 +578,7 @@ class Program_weight_tensor_parameter_48: class Program_weight_tensor_parameter_49: name = "parameter_49" + original_name = "batch_norm2d_2.w_0" shape = [48] dtype = "float32" min_val = float("0") @@ -538,6 +588,7 @@ class Program_weight_tensor_parameter_49: class Program_weight_tensor_parameter_50: name = "parameter_50" + original_name = "batch_norm2d_2.w_2" shape = [48] dtype = "float32" min_val = float("0") @@ -547,6 +598,7 @@ class Program_weight_tensor_parameter_50: class Program_weight_tensor_parameter_51: name = "parameter_51" + original_name = "batch_norm2d_2.w_1" shape = [48] dtype = "float32" min_val = float("0") @@ -556,6 +608,7 @@ class Program_weight_tensor_parameter_51: class Program_weight_tensor_parameter_52: name = "parameter_52" + original_name = "conv2d_12.w_0" shape = [48, 384, 1, 3] dtype = "float32" min_val = float("-0.496082") diff --git a/paddle_samples/PaddleX/ch_SVTRv2_rec/subgraph_3/input_meta.py b/paddle_samples/PaddleX/ch_SVTRv2_rec/subgraph_3/input_meta.py index 763e77ebf..5e855df11 100644 --- a/paddle_samples/PaddleX/ch_SVTRv2_rec/subgraph_3/input_meta.py +++ b/paddle_samples/PaddleX/ch_SVTRv2_rec/subgraph_3/input_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_data_0: name = "data_0" + original_name = "var_0" shape = [4, 3, 64, 320] dtype = "float32" min_val = float("-1.0") diff --git a/paddle_samples/PaddleX/ch_SVTRv2_rec/subgraph_3/weight_meta.py b/paddle_samples/PaddleX/ch_SVTRv2_rec/subgraph_3/weight_meta.py index 9653d49e8..3ea61de65 100644 --- a/paddle_samples/PaddleX/ch_SVTRv2_rec/subgraph_3/weight_meta.py +++ b/paddle_samples/PaddleX/ch_SVTRv2_rec/subgraph_3/weight_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_parameter_0: name = "parameter_0" + original_name = "layer_norm_37.b_0" shape = [384] dtype = "float32" min_val = float("-0.391234") @@ -11,6 +12,7 @@ class Program_weight_tensor_parameter_0: class Program_weight_tensor_parameter_1: name = "parameter_1" + original_name = "layer_norm_37.w_0" shape = [384] dtype = "float32" min_val = float("-0.00273431") @@ -22,6 +24,7 @@ class Program_weight_tensor_parameter_1: class Program_weight_tensor_parameter_2: name = "parameter_2" + original_name = "linear_55.b_0" shape = [384] dtype = "float32" min_val = float("-6.21156") @@ -33,6 +36,7 @@ class Program_weight_tensor_parameter_2: class Program_weight_tensor_parameter_3: name = "parameter_3" + original_name = "linear_55.w_0" shape = [1536, 384] dtype = "float32" min_val = float("-1.26586") @@ -44,6 +48,7 @@ class Program_weight_tensor_parameter_3: class Program_weight_tensor_parameter_4: name = "parameter_4" + original_name = "linear_54.b_0" shape = [1536] dtype = "float32" min_val = float("-1.33895") @@ -55,6 +60,7 @@ class Program_weight_tensor_parameter_4: class Program_weight_tensor_parameter_5: name = "parameter_5" + original_name = "linear_54.w_0" shape = [384, 1536] dtype = "float32" min_val = float("-0.294714") @@ -66,6 +72,7 @@ class Program_weight_tensor_parameter_5: class Program_weight_tensor_parameter_6: name = "parameter_6" + original_name = "layer_norm_36.b_0" shape = [384] dtype = "float32" min_val = float("-5.66332") @@ -77,6 +84,7 @@ class Program_weight_tensor_parameter_6: class Program_weight_tensor_parameter_7: name = "parameter_7" + original_name = "layer_norm_36.w_0" shape = [384] dtype = "float32" min_val = float("-0.738694") @@ -88,6 +96,7 @@ class Program_weight_tensor_parameter_7: class Program_weight_tensor_parameter_8: name = "parameter_8" + original_name = "linear_53.b_0" shape = [384] dtype = "float32" min_val = float("-0.828408") @@ -99,6 +108,7 @@ class Program_weight_tensor_parameter_8: class Program_weight_tensor_parameter_9: name = "parameter_9" + original_name = "linear_53.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.36161") @@ -110,6 +120,7 @@ class Program_weight_tensor_parameter_9: class Program_weight_tensor_parameter_10: name = "parameter_10" + original_name = "linear_52.b_0" shape = [1152] dtype = "float32" min_val = float("-2.96073") @@ -121,6 +132,7 @@ class Program_weight_tensor_parameter_10: class Program_weight_tensor_parameter_11: name = "parameter_11" + original_name = "linear_52.w_0" shape = [384, 1152] dtype = "float32" min_val = float("-0.359938") @@ -132,6 +144,7 @@ class Program_weight_tensor_parameter_11: class Program_weight_tensor_parameter_12: name = "parameter_12" + original_name = "layer_norm_35.b_0" shape = [384] dtype = "float32" min_val = float("-0.966974") @@ -143,6 +156,7 @@ class Program_weight_tensor_parameter_12: class Program_weight_tensor_parameter_13: name = "parameter_13" + original_name = "layer_norm_35.w_0" shape = [384] dtype = "float32" min_val = float("-0.00660634") @@ -154,6 +168,7 @@ class Program_weight_tensor_parameter_13: class Program_weight_tensor_parameter_14: name = "parameter_14" + original_name = "linear_51.b_0" shape = [384] dtype = "float32" min_val = float("-1.16391") @@ -165,6 +180,7 @@ class Program_weight_tensor_parameter_14: class Program_weight_tensor_parameter_15: name = "parameter_15" + original_name = "linear_51.w_0" shape = [1536, 384] dtype = "float32" min_val = float("-0.644514") @@ -176,6 +192,7 @@ class Program_weight_tensor_parameter_15: class Program_weight_tensor_parameter_16: name = "parameter_16" + original_name = "linear_50.b_0" shape = [1536] dtype = "float32" min_val = float("-1.29829") @@ -187,6 +204,7 @@ class Program_weight_tensor_parameter_16: class Program_weight_tensor_parameter_17: name = "parameter_17" + original_name = "linear_50.w_0" shape = [384, 1536] dtype = "float32" min_val = float("-0.371481") @@ -198,6 +216,7 @@ class Program_weight_tensor_parameter_17: class Program_weight_tensor_parameter_18: name = "parameter_18" + original_name = "layer_norm_34.b_0" shape = [384] dtype = "float32" min_val = float("-3.00944") @@ -209,6 +228,7 @@ class Program_weight_tensor_parameter_18: class Program_weight_tensor_parameter_19: name = "parameter_19" + original_name = "layer_norm_34.w_0" shape = [384] dtype = "float32" min_val = float("-0.252943") @@ -220,6 +240,7 @@ class Program_weight_tensor_parameter_19: class Program_weight_tensor_parameter_20: name = "parameter_20" + original_name = "linear_49.b_0" shape = [384] dtype = "float32" min_val = float("-1.02449") @@ -231,6 +252,7 @@ class Program_weight_tensor_parameter_20: class Program_weight_tensor_parameter_21: name = "parameter_21" + original_name = "linear_49.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.32742") @@ -242,6 +264,7 @@ class Program_weight_tensor_parameter_21: class Program_weight_tensor_parameter_22: name = "parameter_22" + original_name = "linear_48.b_0" shape = [1152] dtype = "float32" min_val = float("-2.91644") @@ -253,6 +276,7 @@ class Program_weight_tensor_parameter_22: class Program_weight_tensor_parameter_23: name = "parameter_23" + original_name = "linear_48.w_0" shape = [384, 1152] dtype = "float32" min_val = float("-0.36594") @@ -264,6 +288,7 @@ class Program_weight_tensor_parameter_23: class Program_weight_tensor_parameter_24: name = "parameter_24" + original_name = "layer_norm_33.b_0" shape = [384] dtype = "float32" min_val = float("-0.957574") @@ -275,6 +300,7 @@ class Program_weight_tensor_parameter_24: class Program_weight_tensor_parameter_25: name = "parameter_25" + original_name = "layer_norm_33.w_0" shape = [384] dtype = "float32" min_val = float("-0.00498034") @@ -286,6 +312,7 @@ class Program_weight_tensor_parameter_25: class Program_weight_tensor_parameter_26: name = "parameter_26" + original_name = "linear_47.b_0" shape = [384] dtype = "float32" min_val = float("-1.07558") @@ -297,6 +324,7 @@ class Program_weight_tensor_parameter_26: class Program_weight_tensor_parameter_27: name = "parameter_27" + original_name = "linear_47.w_0" shape = [1536, 384] dtype = "float32" min_val = float("-0.659897") @@ -308,6 +336,7 @@ class Program_weight_tensor_parameter_27: class Program_weight_tensor_parameter_28: name = "parameter_28" + original_name = "linear_46.b_0" shape = [1536] dtype = "float32" min_val = float("-1.18677") @@ -319,6 +348,7 @@ class Program_weight_tensor_parameter_28: class Program_weight_tensor_parameter_29: name = "parameter_29" + original_name = "linear_46.w_0" shape = [384, 1536] dtype = "float32" min_val = float("-0.445094") @@ -330,6 +360,7 @@ class Program_weight_tensor_parameter_29: class Program_weight_tensor_parameter_30: name = "parameter_30" + original_name = "layer_norm_32.b_0" shape = [384] dtype = "float32" min_val = float("-2.24688") @@ -341,6 +372,7 @@ class Program_weight_tensor_parameter_30: class Program_weight_tensor_parameter_31: name = "parameter_31" + original_name = "layer_norm_32.w_0" shape = [384] dtype = "float32" min_val = float("-1.69377") @@ -352,6 +384,7 @@ class Program_weight_tensor_parameter_31: class Program_weight_tensor_parameter_32: name = "parameter_32" + original_name = "linear_45.b_0" shape = [384] dtype = "float32" min_val = float("-0.90244") @@ -363,6 +396,7 @@ class Program_weight_tensor_parameter_32: class Program_weight_tensor_parameter_33: name = "parameter_33" + original_name = "linear_45.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.382145") @@ -374,6 +408,7 @@ class Program_weight_tensor_parameter_33: class Program_weight_tensor_parameter_34: name = "parameter_34" + original_name = "linear_44.b_0" shape = [1152] dtype = "float32" min_val = float("-2.91195") @@ -385,6 +420,7 @@ class Program_weight_tensor_parameter_34: class Program_weight_tensor_parameter_35: name = "parameter_35" + original_name = "linear_44.w_0" shape = [384, 1152] dtype = "float32" min_val = float("-0.430739") @@ -396,6 +432,7 @@ class Program_weight_tensor_parameter_35: class Program_weight_tensor_parameter_36: name = "parameter_36" + original_name = "layer_norm_31.b_0" shape = [384] dtype = "float32" min_val = float("-0.916805") @@ -407,6 +444,7 @@ class Program_weight_tensor_parameter_36: class Program_weight_tensor_parameter_37: name = "parameter_37" + original_name = "layer_norm_31.w_0" shape = [384] dtype = "float32" min_val = float("-0.726943") @@ -418,6 +456,7 @@ class Program_weight_tensor_parameter_37: class Program_weight_tensor_parameter_38: name = "parameter_38" + original_name = "linear_43.b_0" shape = [384] dtype = "float32" min_val = float("-1.78986") @@ -429,6 +468,7 @@ class Program_weight_tensor_parameter_38: class Program_weight_tensor_parameter_39: name = "parameter_39" + original_name = "linear_43.w_0" shape = [1536, 384] dtype = "float32" min_val = float("-0.560883") @@ -440,6 +480,7 @@ class Program_weight_tensor_parameter_39: class Program_weight_tensor_parameter_40: name = "parameter_40" + original_name = "linear_42.b_0" shape = [1536] dtype = "float32" min_val = float("-1.17196") @@ -451,6 +492,7 @@ class Program_weight_tensor_parameter_40: class Program_weight_tensor_parameter_41: name = "parameter_41" + original_name = "linear_42.w_0" shape = [384, 1536] dtype = "float32" min_val = float("-0.754343") @@ -462,6 +504,7 @@ class Program_weight_tensor_parameter_41: class Program_weight_tensor_parameter_42: name = "parameter_42" + original_name = "layer_norm_30.b_0" shape = [384] dtype = "float32" min_val = float("-4.64774") @@ -473,6 +516,7 @@ class Program_weight_tensor_parameter_42: class Program_weight_tensor_parameter_43: name = "parameter_43" + original_name = "layer_norm_30.w_0" shape = [384] dtype = "float32" min_val = float("-1.46338") @@ -484,6 +528,7 @@ class Program_weight_tensor_parameter_43: class Program_weight_tensor_parameter_44: name = "parameter_44" + original_name = "linear_41.b_0" shape = [384] dtype = "float32" min_val = float("-0.820121") @@ -495,6 +540,7 @@ class Program_weight_tensor_parameter_44: class Program_weight_tensor_parameter_45: name = "parameter_45" + original_name = "linear_41.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.453601") @@ -506,6 +552,7 @@ class Program_weight_tensor_parameter_45: class Program_weight_tensor_parameter_46: name = "parameter_46" + original_name = "linear_40.b_0" shape = [1152] dtype = "float32" min_val = float("-3.22004") @@ -517,6 +564,7 @@ class Program_weight_tensor_parameter_46: class Program_weight_tensor_parameter_47: name = "parameter_47" + original_name = "linear_40.w_0" shape = [384, 1152] dtype = "float32" min_val = float("-0.413394") @@ -528,6 +576,7 @@ class Program_weight_tensor_parameter_47: class Program_weight_tensor_parameter_48: name = "parameter_48" + original_name = "layer_norm_29.b_0" shape = [384] dtype = "float32" min_val = float("-0.991885") @@ -539,6 +588,7 @@ class Program_weight_tensor_parameter_48: class Program_weight_tensor_parameter_49: name = "parameter_49" + original_name = "layer_norm_29.w_0" shape = [384] dtype = "float32" min_val = float("-0.0522129") @@ -550,6 +600,7 @@ class Program_weight_tensor_parameter_49: class Program_weight_tensor_parameter_50: name = "parameter_50" + original_name = "linear_39.b_0" shape = [384] dtype = "float32" min_val = float("-1.50586") @@ -561,6 +612,7 @@ class Program_weight_tensor_parameter_50: class Program_weight_tensor_parameter_51: name = "parameter_51" + original_name = "linear_39.w_0" shape = [1536, 384] dtype = "float32" min_val = float("-0.673087") @@ -572,6 +624,7 @@ class Program_weight_tensor_parameter_51: class Program_weight_tensor_parameter_52: name = "parameter_52" + original_name = "linear_38.b_0" shape = [1536] dtype = "float32" min_val = float("-0.743186") @@ -583,6 +636,7 @@ class Program_weight_tensor_parameter_52: class Program_weight_tensor_parameter_53: name = "parameter_53" + original_name = "linear_38.w_0" shape = [384, 1536] dtype = "float32" min_val = float("-0.299899") @@ -594,6 +648,7 @@ class Program_weight_tensor_parameter_53: class Program_weight_tensor_parameter_54: name = "parameter_54" + original_name = "layer_norm_28.b_0" shape = [384] dtype = "float32" min_val = float("-5.35791") @@ -605,6 +660,7 @@ class Program_weight_tensor_parameter_54: class Program_weight_tensor_parameter_55: name = "parameter_55" + original_name = "layer_norm_28.w_0" shape = [384] dtype = "float32" min_val = float("-0.0899557") @@ -616,6 +672,7 @@ class Program_weight_tensor_parameter_55: class Program_weight_tensor_parameter_56: name = "parameter_56" + original_name = "linear_37.b_0" shape = [384] dtype = "float32" min_val = float("-1.34634") @@ -627,6 +684,7 @@ class Program_weight_tensor_parameter_56: class Program_weight_tensor_parameter_57: name = "parameter_57" + original_name = "linear_37.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.301549") @@ -638,6 +696,7 @@ class Program_weight_tensor_parameter_57: class Program_weight_tensor_parameter_58: name = "parameter_58" + original_name = "linear_36.b_0" shape = [1152] dtype = "float32" min_val = float("-2.09165") @@ -649,6 +708,7 @@ class Program_weight_tensor_parameter_58: class Program_weight_tensor_parameter_59: name = "parameter_59" + original_name = "linear_36.w_0" shape = [384, 1152] dtype = "float32" min_val = float("-0.403793") @@ -660,6 +720,7 @@ class Program_weight_tensor_parameter_59: class Program_weight_tensor_parameter_60: name = "parameter_60" + original_name = "layer_norm_27.b_0" shape = [384] dtype = "float32" min_val = float("-1.11619") @@ -671,6 +732,7 @@ class Program_weight_tensor_parameter_60: class Program_weight_tensor_parameter_61: name = "parameter_61" + original_name = "layer_norm_27.w_0" shape = [384] dtype = "float32" min_val = float("-0.0581639") @@ -682,6 +744,7 @@ class Program_weight_tensor_parameter_61: class Program_weight_tensor_parameter_62: name = "parameter_62" + original_name = "linear_35.b_0" shape = [384] dtype = "float32" min_val = float("-1.44023") @@ -693,6 +756,7 @@ class Program_weight_tensor_parameter_62: class Program_weight_tensor_parameter_63: name = "parameter_63" + original_name = "linear_35.w_0" shape = [1536, 384] dtype = "float32" min_val = float("-0.925741") @@ -704,6 +768,7 @@ class Program_weight_tensor_parameter_63: class Program_weight_tensor_parameter_64: name = "parameter_64" + original_name = "linear_34.b_0" shape = [1536] dtype = "float32" min_val = float("-0.688288") @@ -715,6 +780,7 @@ class Program_weight_tensor_parameter_64: class Program_weight_tensor_parameter_65: name = "parameter_65" + original_name = "linear_34.w_0" shape = [384, 1536] dtype = "float32" min_val = float("-0.503913") @@ -726,6 +792,7 @@ class Program_weight_tensor_parameter_65: class Program_weight_tensor_parameter_66: name = "parameter_66" + original_name = "layer_norm_26.b_0" shape = [384] dtype = "float32" min_val = float("-3.63391") @@ -737,6 +804,7 @@ class Program_weight_tensor_parameter_66: class Program_weight_tensor_parameter_67: name = "parameter_67" + original_name = "layer_norm_26.w_0" shape = [384] dtype = "float32" min_val = float("-0.0887419") @@ -748,6 +816,7 @@ class Program_weight_tensor_parameter_67: class Program_weight_tensor_parameter_68: name = "parameter_68" + original_name = "linear_33.b_0" shape = [384] dtype = "float32" min_val = float("-0.674133") @@ -759,6 +828,7 @@ class Program_weight_tensor_parameter_68: class Program_weight_tensor_parameter_69: name = "parameter_69" + original_name = "linear_33.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.363554") @@ -770,6 +840,7 @@ class Program_weight_tensor_parameter_69: class Program_weight_tensor_parameter_70: name = "parameter_70" + original_name = "linear_32.b_0" shape = [1152] dtype = "float32" min_val = float("-2.58158") @@ -781,6 +852,7 @@ class Program_weight_tensor_parameter_70: class Program_weight_tensor_parameter_71: name = "parameter_71" + original_name = "linear_32.w_0" shape = [384, 1152] dtype = "float32" min_val = float("-0.30861") @@ -792,6 +864,7 @@ class Program_weight_tensor_parameter_71: class Program_weight_tensor_parameter_72: name = "parameter_72" + original_name = "layer_norm_25.b_0" shape = [384] dtype = "float32" min_val = float("-0.611573") @@ -803,6 +876,7 @@ class Program_weight_tensor_parameter_72: class Program_weight_tensor_parameter_73: name = "parameter_73" + original_name = "layer_norm_25.w_0" shape = [384] dtype = "float32" min_val = float("-0.0176943") @@ -814,6 +888,7 @@ class Program_weight_tensor_parameter_73: class Program_weight_tensor_parameter_74: name = "parameter_74" + original_name = "conv2d_11.b_0" shape = [384] dtype = "float32" min_val = float("-2.13167") @@ -825,6 +900,7 @@ class Program_weight_tensor_parameter_74: class Program_weight_tensor_parameter_75: name = "parameter_75" + original_name = "conv2d_11.w_0" shape = [384, 256, 3, 3] dtype = "float32" min_val = float("-4.81624") @@ -836,6 +912,7 @@ class Program_weight_tensor_parameter_75: class Program_weight_tensor_parameter_76: name = "parameter_76" + original_name = "layer_norm_24.b_0" shape = [256] dtype = "float32" min_val = float("-1.00302") @@ -847,6 +924,7 @@ class Program_weight_tensor_parameter_76: class Program_weight_tensor_parameter_77: name = "parameter_77" + original_name = "layer_norm_24.w_0" shape = [256] dtype = "float32" min_val = float("0.22568") @@ -858,6 +936,7 @@ class Program_weight_tensor_parameter_77: class Program_weight_tensor_parameter_78: name = "parameter_78" + original_name = "linear_31.b_0" shape = [256] dtype = "float32" min_val = float("-8.09258") @@ -869,6 +948,7 @@ class Program_weight_tensor_parameter_78: class Program_weight_tensor_parameter_79: name = "parameter_79" + original_name = "linear_31.w_0" shape = [1024, 256] dtype = "float32" min_val = float("-1.43969") @@ -880,6 +960,7 @@ class Program_weight_tensor_parameter_79: class Program_weight_tensor_parameter_80: name = "parameter_80" + original_name = "linear_30.b_0" shape = [1024] dtype = "float32" min_val = float("-1.46189") @@ -891,6 +972,7 @@ class Program_weight_tensor_parameter_80: class Program_weight_tensor_parameter_81: name = "parameter_81" + original_name = "linear_30.w_0" shape = [256, 1024] dtype = "float32" min_val = float("-0.347586") @@ -902,6 +984,7 @@ class Program_weight_tensor_parameter_81: class Program_weight_tensor_parameter_82: name = "parameter_82" + original_name = "layer_norm_23.b_0" shape = [256] dtype = "float32" min_val = float("-3.06699") @@ -913,6 +996,7 @@ class Program_weight_tensor_parameter_82: class Program_weight_tensor_parameter_83: name = "parameter_83" + original_name = "layer_norm_23.w_0" shape = [256] dtype = "float32" min_val = float("0.112453") @@ -924,6 +1008,7 @@ class Program_weight_tensor_parameter_83: class Program_weight_tensor_parameter_84: name = "parameter_84" + original_name = "linear_29.b_0" shape = [256] dtype = "float32" min_val = float("-1.23851") @@ -935,6 +1020,7 @@ class Program_weight_tensor_parameter_84: class Program_weight_tensor_parameter_85: name = "parameter_85" + original_name = "linear_29.w_0" shape = [256, 256] dtype = "float32" min_val = float("-0.324324") @@ -946,6 +1032,7 @@ class Program_weight_tensor_parameter_85: class Program_weight_tensor_parameter_86: name = "parameter_86" + original_name = "linear_28.b_0" shape = [768] dtype = "float32" min_val = float("-3.23557") @@ -957,6 +1044,7 @@ class Program_weight_tensor_parameter_86: class Program_weight_tensor_parameter_87: name = "parameter_87" + original_name = "linear_28.w_0" shape = [256, 768] dtype = "float32" min_val = float("-0.604016") @@ -968,6 +1056,7 @@ class Program_weight_tensor_parameter_87: class Program_weight_tensor_parameter_88: name = "parameter_88" + original_name = "layer_norm_22.b_0" shape = [256] dtype = "float32" min_val = float("-1.03089") @@ -979,6 +1068,7 @@ class Program_weight_tensor_parameter_88: class Program_weight_tensor_parameter_89: name = "parameter_89" + original_name = "layer_norm_22.w_0" shape = [256] dtype = "float32" min_val = float("0.287937") @@ -990,6 +1080,7 @@ class Program_weight_tensor_parameter_89: class Program_weight_tensor_parameter_90: name = "parameter_90" + original_name = "linear_27.b_0" shape = [256] dtype = "float32" min_val = float("-2.05805") @@ -1001,6 +1092,7 @@ class Program_weight_tensor_parameter_90: class Program_weight_tensor_parameter_91: name = "parameter_91" + original_name = "linear_27.w_0" shape = [1024, 256] dtype = "float32" min_val = float("-0.562583") @@ -1012,6 +1104,7 @@ class Program_weight_tensor_parameter_91: class Program_weight_tensor_parameter_92: name = "parameter_92" + original_name = "linear_26.b_0" shape = [1024] dtype = "float32" min_val = float("-1.33321") @@ -1023,6 +1116,7 @@ class Program_weight_tensor_parameter_92: class Program_weight_tensor_parameter_93: name = "parameter_93" + original_name = "linear_26.w_0" shape = [256, 1024] dtype = "float32" min_val = float("-0.381275") @@ -1034,6 +1128,7 @@ class Program_weight_tensor_parameter_93: class Program_weight_tensor_parameter_94: name = "parameter_94" + original_name = "layer_norm_21.b_0" shape = [256] dtype = "float32" min_val = float("-2.22319") @@ -1045,6 +1140,7 @@ class Program_weight_tensor_parameter_94: class Program_weight_tensor_parameter_95: name = "parameter_95" + original_name = "layer_norm_21.w_0" shape = [256] dtype = "float32" min_val = float("0.0818852") @@ -1056,6 +1152,7 @@ class Program_weight_tensor_parameter_95: class Program_weight_tensor_parameter_96: name = "parameter_96" + original_name = "linear_25.b_0" shape = [256] dtype = "float32" min_val = float("-1.62126") @@ -1067,6 +1164,7 @@ class Program_weight_tensor_parameter_96: class Program_weight_tensor_parameter_97: name = "parameter_97" + original_name = "linear_25.w_0" shape = [256, 256] dtype = "float32" min_val = float("-0.486574") @@ -1078,6 +1176,7 @@ class Program_weight_tensor_parameter_97: class Program_weight_tensor_parameter_98: name = "parameter_98" + original_name = "linear_24.b_0" shape = [768] dtype = "float32" min_val = float("-2.94557") @@ -1089,6 +1188,7 @@ class Program_weight_tensor_parameter_98: class Program_weight_tensor_parameter_99: name = "parameter_99" + original_name = "linear_24.w_0" shape = [256, 768] dtype = "float32" min_val = float("-0.415111") @@ -1100,6 +1200,7 @@ class Program_weight_tensor_parameter_99: class Program_weight_tensor_parameter_100: name = "parameter_100" + original_name = "layer_norm_20.b_0" shape = [256] dtype = "float32" min_val = float("-1.52143") @@ -1111,6 +1212,7 @@ class Program_weight_tensor_parameter_100: class Program_weight_tensor_parameter_101: name = "parameter_101" + original_name = "layer_norm_20.w_0" shape = [256] dtype = "float32" min_val = float("0.263449") @@ -1122,6 +1224,7 @@ class Program_weight_tensor_parameter_101: class Program_weight_tensor_parameter_102: name = "parameter_102" + original_name = "linear_23.b_0" shape = [256] dtype = "float32" min_val = float("-0.74461") @@ -1133,6 +1236,7 @@ class Program_weight_tensor_parameter_102: class Program_weight_tensor_parameter_103: name = "parameter_103" + original_name = "linear_23.w_0" shape = [1024, 256] dtype = "float32" min_val = float("-0.504825") @@ -1144,6 +1248,7 @@ class Program_weight_tensor_parameter_103: class Program_weight_tensor_parameter_104: name = "parameter_104" + original_name = "linear_22.b_0" shape = [1024] dtype = "float32" min_val = float("-1.1179") @@ -1155,6 +1260,7 @@ class Program_weight_tensor_parameter_104: class Program_weight_tensor_parameter_105: name = "parameter_105" + original_name = "linear_22.w_0" shape = [256, 1024] dtype = "float32" min_val = float("-0.477724") @@ -1166,6 +1272,7 @@ class Program_weight_tensor_parameter_105: class Program_weight_tensor_parameter_106: name = "parameter_106" + original_name = "layer_norm_19.b_0" shape = [256] dtype = "float32" min_val = float("-2.66275") @@ -1177,6 +1284,7 @@ class Program_weight_tensor_parameter_106: class Program_weight_tensor_parameter_107: name = "parameter_107" + original_name = "layer_norm_19.w_0" shape = [256] dtype = "float32" min_val = float("0.0184293") @@ -1188,6 +1296,7 @@ class Program_weight_tensor_parameter_107: class Program_weight_tensor_parameter_108: name = "parameter_108" + original_name = "linear_21.b_0" shape = [256] dtype = "float32" min_val = float("-1.94903") @@ -1199,6 +1308,7 @@ class Program_weight_tensor_parameter_108: class Program_weight_tensor_parameter_109: name = "parameter_109" + original_name = "linear_21.w_0" shape = [256, 256] dtype = "float32" min_val = float("-0.334037") @@ -1210,6 +1320,7 @@ class Program_weight_tensor_parameter_109: class Program_weight_tensor_parameter_110: name = "parameter_110" + original_name = "linear_20.b_0" shape = [768] dtype = "float32" min_val = float("-3.24135") @@ -1221,6 +1332,7 @@ class Program_weight_tensor_parameter_110: class Program_weight_tensor_parameter_111: name = "parameter_111" + original_name = "linear_20.w_0" shape = [256, 768] dtype = "float32" min_val = float("-0.484464") @@ -1232,6 +1344,7 @@ class Program_weight_tensor_parameter_111: class Program_weight_tensor_parameter_112: name = "parameter_112" + original_name = "layer_norm_18.b_0" shape = [256] dtype = "float32" min_val = float("-1.22621") @@ -1243,6 +1356,7 @@ class Program_weight_tensor_parameter_112: class Program_weight_tensor_parameter_113: name = "parameter_113" + original_name = "layer_norm_18.w_0" shape = [256] dtype = "float32" min_val = float("0.0741301") @@ -1254,6 +1368,7 @@ class Program_weight_tensor_parameter_113: class Program_weight_tensor_parameter_114: name = "parameter_114" + original_name = "linear_19.b_0" shape = [256] dtype = "float32" min_val = float("-0.594822") @@ -1265,6 +1380,7 @@ class Program_weight_tensor_parameter_114: class Program_weight_tensor_parameter_115: name = "parameter_115" + original_name = "linear_19.w_0" shape = [1024, 256] dtype = "float32" min_val = float("-0.448166") @@ -1276,6 +1392,7 @@ class Program_weight_tensor_parameter_115: class Program_weight_tensor_parameter_116: name = "parameter_116" + original_name = "linear_18.b_0" shape = [1024] dtype = "float32" min_val = float("-1.13055") @@ -1287,6 +1404,7 @@ class Program_weight_tensor_parameter_116: class Program_weight_tensor_parameter_117: name = "parameter_117" + original_name = "linear_18.w_0" shape = [256, 1024] dtype = "float32" min_val = float("-0.338649") @@ -1298,6 +1416,7 @@ class Program_weight_tensor_parameter_117: class Program_weight_tensor_parameter_118: name = "parameter_118" + original_name = "layer_norm_17.b_0" shape = [256] dtype = "float32" min_val = float("-2.27003") @@ -1309,6 +1428,7 @@ class Program_weight_tensor_parameter_118: class Program_weight_tensor_parameter_119: name = "parameter_119" + original_name = "layer_norm_17.w_0" shape = [256] dtype = "float32" min_val = float("0.0854208") @@ -1320,6 +1440,7 @@ class Program_weight_tensor_parameter_119: class Program_weight_tensor_parameter_120: name = "parameter_120" + original_name = "linear_17.b_0" shape = [256] dtype = "float32" min_val = float("-1.86822") @@ -1331,6 +1452,7 @@ class Program_weight_tensor_parameter_120: class Program_weight_tensor_parameter_121: name = "parameter_121" + original_name = "linear_17.w_0" shape = [256, 256] dtype = "float32" min_val = float("-0.429145") @@ -1342,6 +1464,7 @@ class Program_weight_tensor_parameter_121: class Program_weight_tensor_parameter_122: name = "parameter_122" + original_name = "linear_16.b_0" shape = [768] dtype = "float32" min_val = float("-2.90161") @@ -1353,6 +1476,7 @@ class Program_weight_tensor_parameter_122: class Program_weight_tensor_parameter_123: name = "parameter_123" + original_name = "linear_16.w_0" shape = [256, 768] dtype = "float32" min_val = float("-0.307903") @@ -1364,6 +1488,7 @@ class Program_weight_tensor_parameter_123: class Program_weight_tensor_parameter_124: name = "parameter_124" + original_name = "layer_norm_16.b_0" shape = [256] dtype = "float32" min_val = float("-1.72865") @@ -1375,6 +1500,7 @@ class Program_weight_tensor_parameter_124: class Program_weight_tensor_parameter_125: name = "parameter_125" + original_name = "layer_norm_16.w_0" shape = [256] dtype = "float32" min_val = float("0.295198") @@ -1386,6 +1512,7 @@ class Program_weight_tensor_parameter_125: class Program_weight_tensor_parameter_126: name = "parameter_126" + original_name = "linear_15.b_0" shape = [256] dtype = "float32" min_val = float("-0.541449") @@ -1397,6 +1524,7 @@ class Program_weight_tensor_parameter_126: class Program_weight_tensor_parameter_127: name = "parameter_127" + original_name = "linear_15.w_0" shape = [1024, 256] dtype = "float32" min_val = float("-0.438235") @@ -1408,6 +1536,7 @@ class Program_weight_tensor_parameter_127: class Program_weight_tensor_parameter_128: name = "parameter_128" + original_name = "linear_14.b_0" shape = [1024] dtype = "float32" min_val = float("-0.869132") @@ -1419,6 +1548,7 @@ class Program_weight_tensor_parameter_128: class Program_weight_tensor_parameter_129: name = "parameter_129" + original_name = "linear_14.w_0" shape = [256, 1024] dtype = "float32" min_val = float("-0.622617") @@ -1430,6 +1560,7 @@ class Program_weight_tensor_parameter_129: class Program_weight_tensor_parameter_130: name = "parameter_130" + original_name = "layer_norm_15.b_0" shape = [256] dtype = "float32" min_val = float("-1.54389") @@ -1441,6 +1572,7 @@ class Program_weight_tensor_parameter_130: class Program_weight_tensor_parameter_131: name = "parameter_131" + original_name = "layer_norm_15.w_0" shape = [256] dtype = "float32" min_val = float("0.253241") @@ -1452,6 +1584,7 @@ class Program_weight_tensor_parameter_131: class Program_weight_tensor_parameter_132: name = "parameter_132" + original_name = "conv2d_10.b_0" shape = [256] dtype = "float32" min_val = float("-2.27709") @@ -1463,6 +1596,7 @@ class Program_weight_tensor_parameter_132: class Program_weight_tensor_parameter_133: name = "parameter_133" + original_name = "conv2d_10.w_0" shape = [256, 32, 5, 5] dtype = "float32" min_val = float("-0.271916") @@ -1474,6 +1608,7 @@ class Program_weight_tensor_parameter_133: class Program_weight_tensor_parameter_134: name = "parameter_134" + original_name = "layer_norm_14.b_0" shape = [256] dtype = "float32" min_val = float("-0.633171") @@ -1485,6 +1620,7 @@ class Program_weight_tensor_parameter_134: class Program_weight_tensor_parameter_135: name = "parameter_135" + original_name = "layer_norm_14.w_0" shape = [256] dtype = "float32" min_val = float("0.0638513") @@ -1496,6 +1632,7 @@ class Program_weight_tensor_parameter_135: class Program_weight_tensor_parameter_136: name = "parameter_136" + original_name = "linear_13.b_0" shape = [256] dtype = "float32" min_val = float("-8.93611") @@ -1507,6 +1644,7 @@ class Program_weight_tensor_parameter_136: class Program_weight_tensor_parameter_137: name = "parameter_137" + original_name = "linear_13.w_0" shape = [1024, 256] dtype = "float32" min_val = float("-1.32594") @@ -1518,6 +1656,7 @@ class Program_weight_tensor_parameter_137: class Program_weight_tensor_parameter_138: name = "parameter_138" + original_name = "linear_12.b_0" shape = [1024] dtype = "float32" min_val = float("-1.46672") @@ -1529,6 +1668,7 @@ class Program_weight_tensor_parameter_138: class Program_weight_tensor_parameter_139: name = "parameter_139" + original_name = "linear_12.w_0" shape = [256, 1024] dtype = "float32" min_val = float("-0.324049") @@ -1540,6 +1680,7 @@ class Program_weight_tensor_parameter_139: class Program_weight_tensor_parameter_140: name = "parameter_140" + original_name = "layer_norm_13.b_0" shape = [256] dtype = "float32" min_val = float("-0.661923") @@ -1551,6 +1692,7 @@ class Program_weight_tensor_parameter_140: class Program_weight_tensor_parameter_141: name = "parameter_141" + original_name = "layer_norm_13.w_0" shape = [256] dtype = "float32" min_val = float("0.404883") @@ -1562,6 +1704,7 @@ class Program_weight_tensor_parameter_141: class Program_weight_tensor_parameter_142: name = "parameter_142" + original_name = "conv2d_9.b_0" shape = [256] dtype = "float32" min_val = float("-2.16229") @@ -1573,6 +1716,7 @@ class Program_weight_tensor_parameter_142: class Program_weight_tensor_parameter_143: name = "parameter_143" + original_name = "conv2d_9.w_0" shape = [256, 32, 5, 5] dtype = "float32" min_val = float("-0.265172") @@ -1584,6 +1728,7 @@ class Program_weight_tensor_parameter_143: class Program_weight_tensor_parameter_144: name = "parameter_144" + original_name = "layer_norm_12.b_0" shape = [256] dtype = "float32" min_val = float("-0.668521") @@ -1595,6 +1740,7 @@ class Program_weight_tensor_parameter_144: class Program_weight_tensor_parameter_145: name = "parameter_145" + original_name = "layer_norm_12.w_0" shape = [256] dtype = "float32" min_val = float("0.187715") @@ -1606,6 +1752,7 @@ class Program_weight_tensor_parameter_145: class Program_weight_tensor_parameter_146: name = "parameter_146" + original_name = "conv2d_8.b_0" shape = [256] dtype = "float32" min_val = float("-8.25526") @@ -1617,6 +1764,7 @@ class Program_weight_tensor_parameter_146: class Program_weight_tensor_parameter_147: name = "parameter_147" + original_name = "conv2d_8.w_0" shape = [256, 128, 3, 3] dtype = "float32" min_val = float("-6.30351") @@ -1628,6 +1776,7 @@ class Program_weight_tensor_parameter_147: class Program_weight_tensor_parameter_148: name = "parameter_148" + original_name = "layer_norm_11.b_0" shape = [128] dtype = "float32" min_val = float("-0.71621") @@ -1639,6 +1788,7 @@ class Program_weight_tensor_parameter_148: class Program_weight_tensor_parameter_149: name = "parameter_149" + original_name = "layer_norm_11.w_0" shape = [128] dtype = "float32" min_val = float("0.40875") @@ -1650,6 +1800,7 @@ class Program_weight_tensor_parameter_149: class Program_weight_tensor_parameter_150: name = "parameter_150" + original_name = "linear_11.b_0" shape = [128] dtype = "float32" min_val = float("-4.08424") @@ -1661,6 +1812,7 @@ class Program_weight_tensor_parameter_150: class Program_weight_tensor_parameter_151: name = "parameter_151" + original_name = "linear_11.w_0" shape = [512, 128] dtype = "float32" min_val = float("-0.471752") @@ -1672,6 +1824,7 @@ class Program_weight_tensor_parameter_151: class Program_weight_tensor_parameter_152: name = "parameter_152" + original_name = "linear_10.b_0" shape = [512] dtype = "float32" min_val = float("-1.89779") @@ -1683,6 +1836,7 @@ class Program_weight_tensor_parameter_152: class Program_weight_tensor_parameter_153: name = "parameter_153" + original_name = "linear_10.w_0" shape = [128, 512] dtype = "float32" min_val = float("-0.322236") @@ -1694,6 +1848,7 @@ class Program_weight_tensor_parameter_153: class Program_weight_tensor_parameter_154: name = "parameter_154" + original_name = "layer_norm_10.b_0" shape = [128] dtype = "float32" min_val = float("-0.681697") @@ -1705,6 +1860,7 @@ class Program_weight_tensor_parameter_154: class Program_weight_tensor_parameter_155: name = "parameter_155" + original_name = "layer_norm_10.w_0" shape = [128] dtype = "float32" min_val = float("0.572885") @@ -1716,6 +1872,7 @@ class Program_weight_tensor_parameter_155: class Program_weight_tensor_parameter_156: name = "parameter_156" + original_name = "conv2d_7.b_0" shape = [128] dtype = "float32" min_val = float("-0.7091") @@ -1727,6 +1884,7 @@ class Program_weight_tensor_parameter_156: class Program_weight_tensor_parameter_157: name = "parameter_157" + original_name = "conv2d_7.w_0" shape = [128, 32, 5, 5] dtype = "float32" min_val = float("-0.399124") @@ -1738,6 +1896,7 @@ class Program_weight_tensor_parameter_157: class Program_weight_tensor_parameter_158: name = "parameter_158" + original_name = "layer_norm_9.b_0" shape = [128] dtype = "float32" min_val = float("-0.594308") @@ -1749,6 +1908,7 @@ class Program_weight_tensor_parameter_158: class Program_weight_tensor_parameter_159: name = "parameter_159" + original_name = "layer_norm_9.w_0" shape = [128] dtype = "float32" min_val = float("0.27402") @@ -1760,6 +1920,7 @@ class Program_weight_tensor_parameter_159: class Program_weight_tensor_parameter_160: name = "parameter_160" + original_name = "linear_9.b_0" shape = [128] dtype = "float32" min_val = float("-5.1138") @@ -1771,6 +1932,7 @@ class Program_weight_tensor_parameter_160: class Program_weight_tensor_parameter_161: name = "parameter_161" + original_name = "linear_9.w_0" shape = [512, 128] dtype = "float32" min_val = float("-0.818937") @@ -1782,6 +1944,7 @@ class Program_weight_tensor_parameter_161: class Program_weight_tensor_parameter_162: name = "parameter_162" + original_name = "linear_8.b_0" shape = [512] dtype = "float32" min_val = float("-1.37765") @@ -1793,6 +1956,7 @@ class Program_weight_tensor_parameter_162: class Program_weight_tensor_parameter_163: name = "parameter_163" + original_name = "linear_8.w_0" shape = [128, 512] dtype = "float32" min_val = float("-0.405105") @@ -1804,6 +1968,7 @@ class Program_weight_tensor_parameter_163: class Program_weight_tensor_parameter_164: name = "parameter_164" + original_name = "layer_norm_8.b_0" shape = [128] dtype = "float32" min_val = float("-0.815685") @@ -1815,6 +1980,7 @@ class Program_weight_tensor_parameter_164: class Program_weight_tensor_parameter_165: name = "parameter_165" + original_name = "layer_norm_8.w_0" shape = [128] dtype = "float32" min_val = float("0.685492") @@ -1826,6 +1992,7 @@ class Program_weight_tensor_parameter_165: class Program_weight_tensor_parameter_166: name = "parameter_166" + original_name = "conv2d_6.b_0" shape = [128] dtype = "float32" min_val = float("-0.885409") @@ -1837,6 +2004,7 @@ class Program_weight_tensor_parameter_166: class Program_weight_tensor_parameter_167: name = "parameter_167" + original_name = "conv2d_6.w_0" shape = [128, 32, 5, 5] dtype = "float32" min_val = float("-0.317962") @@ -1848,6 +2016,7 @@ class Program_weight_tensor_parameter_167: class Program_weight_tensor_parameter_168: name = "parameter_168" + original_name = "layer_norm_7.b_0" shape = [128] dtype = "float32" min_val = float("-0.55219") @@ -1859,6 +2028,7 @@ class Program_weight_tensor_parameter_168: class Program_weight_tensor_parameter_169: name = "parameter_169" + original_name = "layer_norm_7.w_0" shape = [128] dtype = "float32" min_val = float("0.0629771") @@ -1870,6 +2040,7 @@ class Program_weight_tensor_parameter_169: class Program_weight_tensor_parameter_170: name = "parameter_170" + original_name = "linear_7.b_0" shape = [128] dtype = "float32" min_val = float("-0.830581") @@ -1881,6 +2052,7 @@ class Program_weight_tensor_parameter_170: class Program_weight_tensor_parameter_171: name = "parameter_171" + original_name = "linear_7.w_0" shape = [512, 128] dtype = "float32" min_val = float("-0.988649") @@ -1892,6 +2064,7 @@ class Program_weight_tensor_parameter_171: class Program_weight_tensor_parameter_172: name = "parameter_172" + original_name = "linear_6.b_0" shape = [512] dtype = "float32" min_val = float("-1.56353") @@ -1903,6 +2076,7 @@ class Program_weight_tensor_parameter_172: class Program_weight_tensor_parameter_173: name = "parameter_173" + original_name = "linear_6.w_0" shape = [128, 512] dtype = "float32" min_val = float("-0.414822") @@ -1914,6 +2088,7 @@ class Program_weight_tensor_parameter_173: class Program_weight_tensor_parameter_174: name = "parameter_174" + original_name = "layer_norm_6.b_0" shape = [128] dtype = "float32" min_val = float("-0.706417") @@ -1925,6 +2100,7 @@ class Program_weight_tensor_parameter_174: class Program_weight_tensor_parameter_175: name = "parameter_175" + original_name = "layer_norm_6.w_0" shape = [128] dtype = "float32" min_val = float("0.782251") @@ -1936,6 +2112,7 @@ class Program_weight_tensor_parameter_175: class Program_weight_tensor_parameter_176: name = "parameter_176" + original_name = "conv2d_5.b_0" shape = [128] dtype = "float32" min_val = float("-0.829148") @@ -1947,6 +2124,7 @@ class Program_weight_tensor_parameter_176: class Program_weight_tensor_parameter_177: name = "parameter_177" + original_name = "conv2d_5.w_0" shape = [128, 32, 5, 5] dtype = "float32" min_val = float("-0.276089") @@ -1958,6 +2136,7 @@ class Program_weight_tensor_parameter_177: class Program_weight_tensor_parameter_178: name = "parameter_178" + original_name = "layer_norm_5.b_0" shape = [128] dtype = "float32" min_val = float("-1.43839") @@ -1969,6 +2148,7 @@ class Program_weight_tensor_parameter_178: class Program_weight_tensor_parameter_179: name = "parameter_179" + original_name = "layer_norm_5.w_0" shape = [128] dtype = "float32" min_val = float("0.2699") @@ -1980,6 +2160,7 @@ class Program_weight_tensor_parameter_179: class Program_weight_tensor_parameter_180: name = "parameter_180" + original_name = "linear_5.b_0" shape = [128] dtype = "float32" min_val = float("-4.80893") @@ -1991,6 +2172,7 @@ class Program_weight_tensor_parameter_180: class Program_weight_tensor_parameter_181: name = "parameter_181" + original_name = "linear_5.w_0" shape = [512, 128] dtype = "float32" min_val = float("-0.663739") @@ -2002,6 +2184,7 @@ class Program_weight_tensor_parameter_181: class Program_weight_tensor_parameter_182: name = "parameter_182" + original_name = "linear_4.b_0" shape = [512] dtype = "float32" min_val = float("-1.38975") @@ -2013,6 +2196,7 @@ class Program_weight_tensor_parameter_182: class Program_weight_tensor_parameter_183: name = "parameter_183" + original_name = "linear_4.w_0" shape = [128, 512] dtype = "float32" min_val = float("-0.337775") @@ -2024,6 +2208,7 @@ class Program_weight_tensor_parameter_183: class Program_weight_tensor_parameter_184: name = "parameter_184" + original_name = "layer_norm_4.b_0" shape = [128] dtype = "float32" min_val = float("-1.21058") @@ -2035,6 +2220,7 @@ class Program_weight_tensor_parameter_184: class Program_weight_tensor_parameter_185: name = "parameter_185" + original_name = "layer_norm_4.w_0" shape = [128] dtype = "float32" min_val = float("0.69899") @@ -2046,6 +2232,7 @@ class Program_weight_tensor_parameter_185: class Program_weight_tensor_parameter_186: name = "parameter_186" + original_name = "conv2d_4.b_0" shape = [128] dtype = "float32" min_val = float("-1.06") @@ -2057,6 +2244,7 @@ class Program_weight_tensor_parameter_186: class Program_weight_tensor_parameter_187: name = "parameter_187" + original_name = "conv2d_4.w_0" shape = [128, 32, 5, 5] dtype = "float32" min_val = float("-0.351555") @@ -2068,6 +2256,7 @@ class Program_weight_tensor_parameter_187: class Program_weight_tensor_parameter_188: name = "parameter_188" + original_name = "layer_norm_3.b_0" shape = [128] dtype = "float32" min_val = float("-0.978142") @@ -2079,6 +2268,7 @@ class Program_weight_tensor_parameter_188: class Program_weight_tensor_parameter_189: name = "parameter_189" + original_name = "layer_norm_3.w_0" shape = [128] dtype = "float32" min_val = float("0.301307") @@ -2090,6 +2280,7 @@ class Program_weight_tensor_parameter_189: class Program_weight_tensor_parameter_190: name = "parameter_190" + original_name = "linear_3.b_0" shape = [128] dtype = "float32" min_val = float("-4.93517") @@ -2101,6 +2292,7 @@ class Program_weight_tensor_parameter_190: class Program_weight_tensor_parameter_191: name = "parameter_191" + original_name = "linear_3.w_0" shape = [512, 128] dtype = "float32" min_val = float("-0.634287") @@ -2112,6 +2304,7 @@ class Program_weight_tensor_parameter_191: class Program_weight_tensor_parameter_192: name = "parameter_192" + original_name = "linear_2.b_0" shape = [512] dtype = "float32" min_val = float("-1.76981") @@ -2123,6 +2316,7 @@ class Program_weight_tensor_parameter_192: class Program_weight_tensor_parameter_193: name = "parameter_193" + original_name = "linear_2.w_0" shape = [128, 512] dtype = "float32" min_val = float("-0.369617") @@ -2134,6 +2328,7 @@ class Program_weight_tensor_parameter_193: class Program_weight_tensor_parameter_194: name = "parameter_194" + original_name = "layer_norm_2.b_0" shape = [128] dtype = "float32" min_val = float("-0.941485") @@ -2145,6 +2340,7 @@ class Program_weight_tensor_parameter_194: class Program_weight_tensor_parameter_195: name = "parameter_195" + original_name = "layer_norm_2.w_0" shape = [128] dtype = "float32" min_val = float("0.631475") @@ -2156,6 +2352,7 @@ class Program_weight_tensor_parameter_195: class Program_weight_tensor_parameter_196: name = "parameter_196" + original_name = "conv2d_3.b_0" shape = [128] dtype = "float32" min_val = float("-1.6476") @@ -2167,6 +2364,7 @@ class Program_weight_tensor_parameter_196: class Program_weight_tensor_parameter_197: name = "parameter_197" + original_name = "conv2d_3.w_0" shape = [128, 32, 5, 5] dtype = "float32" min_val = float("-0.423287") @@ -2178,6 +2376,7 @@ class Program_weight_tensor_parameter_197: class Program_weight_tensor_parameter_198: name = "parameter_198" + original_name = "layer_norm_1.b_0" shape = [128] dtype = "float32" min_val = float("-1.06966") @@ -2189,6 +2388,7 @@ class Program_weight_tensor_parameter_198: class Program_weight_tensor_parameter_199: name = "parameter_199" + original_name = "layer_norm_1.w_0" shape = [128] dtype = "float32" min_val = float("0.476229") @@ -2200,6 +2400,7 @@ class Program_weight_tensor_parameter_199: class Program_weight_tensor_parameter_200: name = "parameter_200" + original_name = "linear_1.b_0" shape = [128] dtype = "float32" min_val = float("-0.885903") @@ -2211,6 +2412,7 @@ class Program_weight_tensor_parameter_200: class Program_weight_tensor_parameter_201: name = "parameter_201" + original_name = "linear_1.w_0" shape = [512, 128] dtype = "float32" min_val = float("-0.397077") @@ -2222,6 +2424,7 @@ class Program_weight_tensor_parameter_201: class Program_weight_tensor_parameter_202: name = "parameter_202" + original_name = "linear_0.b_0" shape = [512] dtype = "float32" min_val = float("-1.34719") @@ -2233,6 +2436,7 @@ class Program_weight_tensor_parameter_202: class Program_weight_tensor_parameter_203: name = "parameter_203" + original_name = "linear_0.w_0" shape = [128, 512] dtype = "float32" min_val = float("-0.549137") @@ -2244,6 +2448,7 @@ class Program_weight_tensor_parameter_203: class Program_weight_tensor_parameter_204: name = "parameter_204" + original_name = "layer_norm_0.b_0" shape = [128] dtype = "float32" min_val = float("-0.778871") @@ -2255,6 +2460,7 @@ class Program_weight_tensor_parameter_204: class Program_weight_tensor_parameter_205: name = "parameter_205" + original_name = "layer_norm_0.w_0" shape = [128] dtype = "float32" min_val = float("0.533038") @@ -2266,6 +2472,7 @@ class Program_weight_tensor_parameter_205: class Program_weight_tensor_parameter_206: name = "parameter_206" + original_name = "conv2d_2.b_0" shape = [128] dtype = "float32" min_val = float("-0.822867") @@ -2277,6 +2484,7 @@ class Program_weight_tensor_parameter_206: class Program_weight_tensor_parameter_207: name = "parameter_207" + original_name = "conv2d_2.w_0" shape = [128, 32, 5, 5] dtype = "float32" min_val = float("-0.486222") @@ -2288,6 +2496,7 @@ class Program_weight_tensor_parameter_207: class Program_weight_tensor_parameter_208: name = "parameter_208" + original_name = "batch_norm2d_1.b_0" shape = [128] dtype = "float32" min_val = float("-1.26996") @@ -2299,6 +2508,7 @@ class Program_weight_tensor_parameter_208: class Program_weight_tensor_parameter_209: name = "parameter_209" + original_name = "batch_norm2d_1.w_0" shape = [128] dtype = "float32" min_val = float("1.05904") @@ -2310,6 +2520,7 @@ class Program_weight_tensor_parameter_209: class Program_weight_tensor_parameter_210: name = "parameter_210" + original_name = "batch_norm2d_1.w_2" shape = [128] dtype = "float32" min_val = float("63.4499") @@ -2321,6 +2532,7 @@ class Program_weight_tensor_parameter_210: class Program_weight_tensor_parameter_211: name = "parameter_211" + original_name = "batch_norm2d_1.w_1" shape = [128] dtype = "float32" min_val = float("-45.0605") @@ -2332,6 +2544,7 @@ class Program_weight_tensor_parameter_211: class Program_weight_tensor_parameter_212: name = "parameter_212" + original_name = "conv2d_1.b_0" shape = [128] dtype = "float32" min_val = float("-0.0297299") @@ -2343,6 +2556,7 @@ class Program_weight_tensor_parameter_212: class Program_weight_tensor_parameter_213: name = "parameter_213" + original_name = "conv2d_1.w_0" shape = [128, 64, 3, 3] dtype = "float32" min_val = float("-5.5775") @@ -2354,6 +2568,7 @@ class Program_weight_tensor_parameter_213: class Program_weight_tensor_parameter_214: name = "parameter_214" + original_name = "batch_norm2d_0.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -2363,6 +2578,7 @@ class Program_weight_tensor_parameter_214: class Program_weight_tensor_parameter_215: name = "parameter_215" + original_name = "batch_norm2d_0.w_0" shape = [64] dtype = "float32" min_val = float("0") @@ -2372,6 +2588,7 @@ class Program_weight_tensor_parameter_215: class Program_weight_tensor_parameter_216: name = "parameter_216" + original_name = "batch_norm2d_0.w_2" shape = [64] dtype = "float32" min_val = float("0") @@ -2381,6 +2598,7 @@ class Program_weight_tensor_parameter_216: class Program_weight_tensor_parameter_217: name = "parameter_217" + original_name = "batch_norm2d_0.w_1" shape = [64] dtype = "float32" min_val = float("0") @@ -2390,6 +2608,7 @@ class Program_weight_tensor_parameter_217: class Program_weight_tensor_parameter_218: name = "parameter_218" + original_name = "conv2d_0.b_0" shape = [64] dtype = "float32" min_val = float("0") @@ -2399,6 +2618,7 @@ class Program_weight_tensor_parameter_218: class Program_weight_tensor_parameter_219: name = "parameter_219" + original_name = "conv2d_0.w_0" shape = [64, 3, 3, 3] dtype = "float32" min_val = float("-1.86415") diff --git a/paddle_samples/PaddleX/ch_SVTRv2_rec/subgraph_7/input_meta.py b/paddle_samples/PaddleX/ch_SVTRv2_rec/subgraph_7/input_meta.py index 629175ff7..0f79c791f 100644 --- a/paddle_samples/PaddleX/ch_SVTRv2_rec/subgraph_7/input_meta.py +++ b/paddle_samples/PaddleX/ch_SVTRv2_rec/subgraph_7/input_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_data_0: name = "data_0" + original_name = "args_0" shape = [] dtype = "int64" data = [12] @@ -7,6 +8,7 @@ class Program_weight_tensor_data_0: class Program_weight_tensor_data_1: name = "data_1" + original_name = "args_1" shape = [] dtype = "int64" data = [384] @@ -14,6 +16,7 @@ class Program_weight_tensor_data_1: class Program_weight_tensor_data_2: name = "data_2" + original_name = "args_2" shape = [] dtype = "int64" data = [12] @@ -21,6 +24,7 @@ class Program_weight_tensor_data_2: class Program_weight_tensor_data_3: name = "data_3" + original_name = "args_3" shape = [] dtype = "int64" data = [384] @@ -28,6 +32,7 @@ class Program_weight_tensor_data_3: class Program_weight_tensor_data_4: name = "data_4" + original_name = "args_4" shape = [] dtype = "int64" data = [12] @@ -35,6 +40,7 @@ class Program_weight_tensor_data_4: class Program_weight_tensor_data_5: name = "data_5" + original_name = "args_5" shape = [] dtype = "int64" data = [384] @@ -42,6 +48,7 @@ class Program_weight_tensor_data_5: class Program_weight_tensor_data_6: name = "data_6" + original_name = "var_3757" shape = [8, 160, 384] dtype = "float32" min_val = float("-5.65986") diff --git a/paddle_samples/PaddleX/ch_SVTRv2_rec/subgraph_7/weight_meta.py b/paddle_samples/PaddleX/ch_SVTRv2_rec/subgraph_7/weight_meta.py index 323191e4d..9633e500f 100644 --- a/paddle_samples/PaddleX/ch_SVTRv2_rec/subgraph_7/weight_meta.py +++ b/paddle_samples/PaddleX/ch_SVTRv2_rec/subgraph_7/weight_meta.py @@ -1,5 +1,6 @@ class Program_weight_tensor_parameter_0: name = "parameter_0" + original_name = "layer_norm_37.b_0" shape = [384] dtype = "float32" min_val = float("-0.391234") @@ -11,6 +12,7 @@ class Program_weight_tensor_parameter_0: class Program_weight_tensor_parameter_1: name = "parameter_1" + original_name = "layer_norm_37.w_0" shape = [384] dtype = "float32" min_val = float("-0.00273413") @@ -22,6 +24,7 @@ class Program_weight_tensor_parameter_1: class Program_weight_tensor_parameter_2: name = "parameter_2" + original_name = "linear_55.b_0" shape = [384] dtype = "float32" min_val = float("-6.21156") @@ -33,6 +36,7 @@ class Program_weight_tensor_parameter_2: class Program_weight_tensor_parameter_3: name = "parameter_3" + original_name = "linear_55.w_0" shape = [1536, 384] dtype = "float32" min_val = float("-1.26586") @@ -44,6 +48,7 @@ class Program_weight_tensor_parameter_3: class Program_weight_tensor_parameter_4: name = "parameter_4" + original_name = "linear_54.b_0" shape = [1536] dtype = "float32" min_val = float("-1.33895") @@ -55,6 +60,7 @@ class Program_weight_tensor_parameter_4: class Program_weight_tensor_parameter_5: name = "parameter_5" + original_name = "linear_54.w_0" shape = [384, 1536] dtype = "float32" min_val = float("-0.294714") @@ -66,6 +72,7 @@ class Program_weight_tensor_parameter_5: class Program_weight_tensor_parameter_6: name = "parameter_6" + original_name = "layer_norm_36.b_0" shape = [384] dtype = "float32" min_val = float("-5.66332") @@ -77,6 +84,7 @@ class Program_weight_tensor_parameter_6: class Program_weight_tensor_parameter_7: name = "parameter_7" + original_name = "layer_norm_36.w_0" shape = [384] dtype = "float32" min_val = float("-0.738694") @@ -88,6 +96,7 @@ class Program_weight_tensor_parameter_7: class Program_weight_tensor_parameter_8: name = "parameter_8" + original_name = "linear_53.b_0" shape = [384] dtype = "float32" min_val = float("-0.828408") @@ -99,6 +108,7 @@ class Program_weight_tensor_parameter_8: class Program_weight_tensor_parameter_9: name = "parameter_9" + original_name = "linear_53.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.361609") @@ -110,6 +120,7 @@ class Program_weight_tensor_parameter_9: class Program_weight_tensor_parameter_10: name = "parameter_10" + original_name = "linear_52.b_0" shape = [1152] dtype = "float32" min_val = float("-2.96073") @@ -121,6 +132,7 @@ class Program_weight_tensor_parameter_10: class Program_weight_tensor_parameter_11: name = "parameter_11" + original_name = "linear_52.w_0" shape = [384, 1152] dtype = "float32" min_val = float("-0.359938") @@ -132,6 +144,7 @@ class Program_weight_tensor_parameter_11: class Program_weight_tensor_parameter_12: name = "parameter_12" + original_name = "layer_norm_35.b_0" shape = [384] dtype = "float32" min_val = float("-0.966974") @@ -143,6 +156,7 @@ class Program_weight_tensor_parameter_12: class Program_weight_tensor_parameter_13: name = "parameter_13" + original_name = "layer_norm_35.w_0" shape = [384] dtype = "float32" min_val = float("-0.00660637") @@ -154,6 +168,7 @@ class Program_weight_tensor_parameter_13: class Program_weight_tensor_parameter_14: name = "parameter_14" + original_name = "linear_51.b_0" shape = [384] dtype = "float32" min_val = float("-1.16391") @@ -165,6 +180,7 @@ class Program_weight_tensor_parameter_14: class Program_weight_tensor_parameter_15: name = "parameter_15" + original_name = "linear_51.w_0" shape = [1536, 384] dtype = "float32" min_val = float("-0.644514") @@ -176,6 +192,7 @@ class Program_weight_tensor_parameter_15: class Program_weight_tensor_parameter_16: name = "parameter_16" + original_name = "linear_50.b_0" shape = [1536] dtype = "float32" min_val = float("-1.29829") @@ -187,6 +204,7 @@ class Program_weight_tensor_parameter_16: class Program_weight_tensor_parameter_17: name = "parameter_17" + original_name = "linear_50.w_0" shape = [384, 1536] dtype = "float32" min_val = float("-0.371481") @@ -198,6 +216,7 @@ class Program_weight_tensor_parameter_17: class Program_weight_tensor_parameter_18: name = "parameter_18" + original_name = "layer_norm_34.b_0" shape = [384] dtype = "float32" min_val = float("-3.00944") @@ -209,6 +228,7 @@ class Program_weight_tensor_parameter_18: class Program_weight_tensor_parameter_19: name = "parameter_19" + original_name = "layer_norm_34.w_0" shape = [384] dtype = "float32" min_val = float("-0.252943") @@ -220,6 +240,7 @@ class Program_weight_tensor_parameter_19: class Program_weight_tensor_parameter_20: name = "parameter_20" + original_name = "linear_49.b_0" shape = [384] dtype = "float32" min_val = float("-1.02449") @@ -231,6 +252,7 @@ class Program_weight_tensor_parameter_20: class Program_weight_tensor_parameter_21: name = "parameter_21" + original_name = "linear_49.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.32742") @@ -242,6 +264,7 @@ class Program_weight_tensor_parameter_21: class Program_weight_tensor_parameter_22: name = "parameter_22" + original_name = "linear_48.b_0" shape = [1152] dtype = "float32" min_val = float("-2.91644") @@ -253,6 +276,7 @@ class Program_weight_tensor_parameter_22: class Program_weight_tensor_parameter_23: name = "parameter_23" + original_name = "linear_48.w_0" shape = [384, 1152] dtype = "float32" min_val = float("-0.36594") @@ -264,6 +288,7 @@ class Program_weight_tensor_parameter_23: class Program_weight_tensor_parameter_24: name = "parameter_24" + original_name = "layer_norm_33.b_0" shape = [384] dtype = "float32" min_val = float("-0.957574") @@ -275,6 +300,7 @@ class Program_weight_tensor_parameter_24: class Program_weight_tensor_parameter_25: name = "parameter_25" + original_name = "layer_norm_33.w_0" shape = [384] dtype = "float32" min_val = float("-0.00498009") @@ -286,6 +312,7 @@ class Program_weight_tensor_parameter_25: class Program_weight_tensor_parameter_26: name = "parameter_26" + original_name = "linear_47.b_0" shape = [384] dtype = "float32" min_val = float("-1.07558") @@ -297,6 +324,7 @@ class Program_weight_tensor_parameter_26: class Program_weight_tensor_parameter_27: name = "parameter_27" + original_name = "linear_47.w_0" shape = [1536, 384] dtype = "float32" min_val = float("-0.659897") @@ -308,6 +336,7 @@ class Program_weight_tensor_parameter_27: class Program_weight_tensor_parameter_28: name = "parameter_28" + original_name = "linear_46.b_0" shape = [1536] dtype = "float32" min_val = float("-1.18677") @@ -319,6 +348,7 @@ class Program_weight_tensor_parameter_28: class Program_weight_tensor_parameter_29: name = "parameter_29" + original_name = "linear_46.w_0" shape = [384, 1536] dtype = "float32" min_val = float("-0.445094") @@ -330,6 +360,7 @@ class Program_weight_tensor_parameter_29: class Program_weight_tensor_parameter_30: name = "parameter_30" + original_name = "layer_norm_32.b_0" shape = [384] dtype = "float32" min_val = float("-2.24688") @@ -341,6 +372,7 @@ class Program_weight_tensor_parameter_30: class Program_weight_tensor_parameter_31: name = "parameter_31" + original_name = "layer_norm_32.w_0" shape = [384] dtype = "float32" min_val = float("-1.69377") @@ -352,6 +384,7 @@ class Program_weight_tensor_parameter_31: class Program_weight_tensor_parameter_32: name = "parameter_32" + original_name = "linear_45.b_0" shape = [384] dtype = "float32" min_val = float("-0.90244") @@ -363,6 +396,7 @@ class Program_weight_tensor_parameter_32: class Program_weight_tensor_parameter_33: name = "parameter_33" + original_name = "linear_45.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.382146") @@ -374,6 +408,7 @@ class Program_weight_tensor_parameter_33: class Program_weight_tensor_parameter_34: name = "parameter_34" + original_name = "linear_44.b_0" shape = [1152] dtype = "float32" min_val = float("-2.91195") @@ -385,6 +420,7 @@ class Program_weight_tensor_parameter_34: class Program_weight_tensor_parameter_35: name = "parameter_35" + original_name = "linear_44.w_0" shape = [384, 1152] dtype = "float32" min_val = float("-0.430739") @@ -396,6 +432,7 @@ class Program_weight_tensor_parameter_35: class Program_weight_tensor_parameter_36: name = "parameter_36" + original_name = "layer_norm_31.b_0" shape = [384] dtype = "float32" min_val = float("-0.916805") @@ -407,6 +444,7 @@ class Program_weight_tensor_parameter_36: class Program_weight_tensor_parameter_37: name = "parameter_37" + original_name = "layer_norm_31.w_0" shape = [384] dtype = "float32" min_val = float("-0.726943") @@ -418,6 +456,7 @@ class Program_weight_tensor_parameter_37: class Program_weight_tensor_parameter_38: name = "parameter_38" + original_name = "linear_43.b_0" shape = [384] dtype = "float32" min_val = float("-1.78986") @@ -429,6 +468,7 @@ class Program_weight_tensor_parameter_38: class Program_weight_tensor_parameter_39: name = "parameter_39" + original_name = "linear_43.w_0" shape = [1536, 384] dtype = "float32" min_val = float("-0.560883") @@ -440,6 +480,7 @@ class Program_weight_tensor_parameter_39: class Program_weight_tensor_parameter_40: name = "parameter_40" + original_name = "linear_42.b_0" shape = [1536] dtype = "float32" min_val = float("-1.17196") @@ -451,6 +492,7 @@ class Program_weight_tensor_parameter_40: class Program_weight_tensor_parameter_41: name = "parameter_41" + original_name = "linear_42.w_0" shape = [384, 1536] dtype = "float32" min_val = float("-0.754343") @@ -462,6 +504,7 @@ class Program_weight_tensor_parameter_41: class Program_weight_tensor_parameter_42: name = "parameter_42" + original_name = "layer_norm_30.b_0" shape = [384] dtype = "float32" min_val = float("-4.64774") @@ -473,6 +516,7 @@ class Program_weight_tensor_parameter_42: class Program_weight_tensor_parameter_43: name = "parameter_43" + original_name = "layer_norm_30.w_0" shape = [384] dtype = "float32" min_val = float("-1.46338") @@ -484,6 +528,7 @@ class Program_weight_tensor_parameter_43: class Program_weight_tensor_parameter_44: name = "parameter_44" + original_name = "linear_41.b_0" shape = [384] dtype = "float32" min_val = float("-0.820121") @@ -495,6 +540,7 @@ class Program_weight_tensor_parameter_44: class Program_weight_tensor_parameter_45: name = "parameter_45" + original_name = "linear_41.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.453601") @@ -506,6 +552,7 @@ class Program_weight_tensor_parameter_45: class Program_weight_tensor_parameter_46: name = "parameter_46" + original_name = "linear_40.b_0" shape = [1152] dtype = "float32" min_val = float("-3.22004") @@ -517,6 +564,7 @@ class Program_weight_tensor_parameter_46: class Program_weight_tensor_parameter_47: name = "parameter_47" + original_name = "linear_40.w_0" shape = [384, 1152] dtype = "float32" min_val = float("-0.413393") @@ -528,6 +576,7 @@ class Program_weight_tensor_parameter_47: class Program_weight_tensor_parameter_48: name = "parameter_48" + original_name = "layer_norm_29.b_0" shape = [384] dtype = "float32" min_val = float("-0.991885") @@ -539,6 +588,7 @@ class Program_weight_tensor_parameter_48: class Program_weight_tensor_parameter_49: name = "parameter_49" + original_name = "layer_norm_29.w_0" shape = [384] dtype = "float32" min_val = float("-0.0522131") @@ -550,6 +600,7 @@ class Program_weight_tensor_parameter_49: class Program_weight_tensor_parameter_50: name = "parameter_50" + original_name = "linear_39.b_0" shape = [384] dtype = "float32" min_val = float("-1.50586") @@ -561,6 +612,7 @@ class Program_weight_tensor_parameter_50: class Program_weight_tensor_parameter_51: name = "parameter_51" + original_name = "linear_39.w_0" shape = [1536, 384] dtype = "float32" min_val = float("-0.673087") @@ -572,6 +624,7 @@ class Program_weight_tensor_parameter_51: class Program_weight_tensor_parameter_52: name = "parameter_52" + original_name = "linear_38.b_0" shape = [1536] dtype = "float32" min_val = float("-0.743186") @@ -583,6 +636,7 @@ class Program_weight_tensor_parameter_52: class Program_weight_tensor_parameter_53: name = "parameter_53" + original_name = "linear_38.w_0" shape = [384, 1536] dtype = "float32" min_val = float("-0.299899") @@ -594,6 +648,7 @@ class Program_weight_tensor_parameter_53: class Program_weight_tensor_parameter_54: name = "parameter_54" + original_name = "layer_norm_28.b_0" shape = [384] dtype = "float32" min_val = float("-5.35791") @@ -605,6 +660,7 @@ class Program_weight_tensor_parameter_54: class Program_weight_tensor_parameter_55: name = "parameter_55" + original_name = "layer_norm_28.w_0" shape = [384] dtype = "float32" min_val = float("-0.0899559") @@ -616,6 +672,7 @@ class Program_weight_tensor_parameter_55: class Program_weight_tensor_parameter_56: name = "parameter_56" + original_name = "linear_37.b_0" shape = [384] dtype = "float32" min_val = float("-1.34634") @@ -627,6 +684,7 @@ class Program_weight_tensor_parameter_56: class Program_weight_tensor_parameter_57: name = "parameter_57" + original_name = "linear_37.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.301549") @@ -638,6 +696,7 @@ class Program_weight_tensor_parameter_57: class Program_weight_tensor_parameter_58: name = "parameter_58" + original_name = "linear_36.b_0" shape = [1152] dtype = "float32" min_val = float("-2.09165") @@ -649,6 +708,7 @@ class Program_weight_tensor_parameter_58: class Program_weight_tensor_parameter_59: name = "parameter_59" + original_name = "linear_36.w_0" shape = [384, 1152] dtype = "float32" min_val = float("-0.403793") @@ -660,6 +720,7 @@ class Program_weight_tensor_parameter_59: class Program_weight_tensor_parameter_60: name = "parameter_60" + original_name = "layer_norm_27.b_0" shape = [384] dtype = "float32" min_val = float("-1.11619") @@ -671,6 +732,7 @@ class Program_weight_tensor_parameter_60: class Program_weight_tensor_parameter_61: name = "parameter_61" + original_name = "layer_norm_27.w_0" shape = [384] dtype = "float32" min_val = float("-0.0581641") @@ -682,6 +744,7 @@ class Program_weight_tensor_parameter_61: class Program_weight_tensor_parameter_62: name = "parameter_62" + original_name = "linear_35.b_0" shape = [384] dtype = "float32" min_val = float("-1.44023") @@ -693,6 +756,7 @@ class Program_weight_tensor_parameter_62: class Program_weight_tensor_parameter_63: name = "parameter_63" + original_name = "linear_35.w_0" shape = [1536, 384] dtype = "float32" min_val = float("-0.925741") @@ -704,6 +768,7 @@ class Program_weight_tensor_parameter_63: class Program_weight_tensor_parameter_64: name = "parameter_64" + original_name = "linear_34.b_0" shape = [1536] dtype = "float32" min_val = float("-0.688288") @@ -715,6 +780,7 @@ class Program_weight_tensor_parameter_64: class Program_weight_tensor_parameter_65: name = "parameter_65" + original_name = "linear_34.w_0" shape = [384, 1536] dtype = "float32" min_val = float("-0.503913") @@ -726,6 +792,7 @@ class Program_weight_tensor_parameter_65: class Program_weight_tensor_parameter_66: name = "parameter_66" + original_name = "layer_norm_26.b_0" shape = [384] dtype = "float32" min_val = float("-3.63391") @@ -737,6 +804,7 @@ class Program_weight_tensor_parameter_66: class Program_weight_tensor_parameter_67: name = "parameter_67" + original_name = "layer_norm_26.w_0" shape = [384] dtype = "float32" min_val = float("-0.0887417") @@ -748,6 +816,7 @@ class Program_weight_tensor_parameter_67: class Program_weight_tensor_parameter_68: name = "parameter_68" + original_name = "linear_33.b_0" shape = [384] dtype = "float32" min_val = float("-0.674133") @@ -759,6 +828,7 @@ class Program_weight_tensor_parameter_68: class Program_weight_tensor_parameter_69: name = "parameter_69" + original_name = "linear_33.w_0" shape = [384, 384] dtype = "float32" min_val = float("-0.363553") @@ -770,6 +840,7 @@ class Program_weight_tensor_parameter_69: class Program_weight_tensor_parameter_70: name = "parameter_70" + original_name = "linear_32.b_0" shape = [1152] dtype = "float32" min_val = float("-2.58158") @@ -781,6 +852,7 @@ class Program_weight_tensor_parameter_70: class Program_weight_tensor_parameter_71: name = "parameter_71" + original_name = "linear_32.w_0" shape = [384, 1152] dtype = "float32" min_val = float("-0.30861") From fb8404cb6116c5e2c195734fd326816a8eeac261 Mon Sep 17 00:00:00 2001 From: Liu Yiqun Date: Tue, 20 Jan 2026 20:33:56 +0800 Subject: [PATCH 06/10] Update graph_hash. --- paddle_samples/PaddleNLP/bert-base-cased/graph_hash.txt | 2 +- paddle_samples/PaddleNLP/bert-large-cased/graph_hash.txt | 2 +- paddle_samples/PaddleNLP/ernie-1.0/graph_hash.txt | 2 +- paddle_samples/PaddleNLP/ernie-2.0-base-zh/graph_hash.txt | 2 +- paddle_samples/PaddleNLP/ernie-2.0-large-zh/graph_hash.txt | 2 +- paddle_samples/PaddleNLP/ernie-3.0-base-zh/graph_hash.txt | 2 +- paddle_samples/PaddleNLP/ernie-3.0-medium-zh/graph_hash.txt | 2 +- paddle_samples/PaddleNLP/ernie-3.0-micro-zh/graph_hash.txt | 2 +- paddle_samples/PaddleNLP/ernie-3.0-xbase-zh/graph_hash.txt | 2 +- paddle_samples/PaddleNLP/ernie-m-large/graph_hash.txt | 2 +- .../ernie-search-large-cross-encoder-marco-en/graph_hash.txt | 2 +- paddle_samples/PaddleNLP/ernie-tiny/graph_hash.txt | 2 +- paddle_samples/PaddleNLP/ppminilm-6l-768h/graph_hash.txt | 2 +- .../PaddleNLP/prophetnet-large-uncased/graph_hash.txt | 2 +- .../PaddleNLP/rocketqa-base-cross-encoder/graph_hash.txt | 2 +- .../PaddleNLP/rocketqa-medium-cross-encoder/graph_hash.txt | 2 +- .../PaddleNLP/rocketqa-mini-cross-encoder/graph_hash.txt | 2 +- paddle_samples/PaddleNLP/skep_ernie_1.0_large_ch/graph_hash.txt | 2 +- paddle_samples/PaddleNLP/skep_ernie_2.0_large_en/graph_hash.txt | 2 +- paddle_samples/PaddleNLP/utc-large/graph_hash.txt | 2 +- paddle_samples/PaddleNLP/utc-xbase/graph_hash.txt | 2 +- 21 files changed, 21 insertions(+), 21 deletions(-) diff --git a/paddle_samples/PaddleNLP/bert-base-cased/graph_hash.txt b/paddle_samples/PaddleNLP/bert-base-cased/graph_hash.txt index 0c116bc18..43bf28fa6 100644 --- a/paddle_samples/PaddleNLP/bert-base-cased/graph_hash.txt +++ b/paddle_samples/PaddleNLP/bert-base-cased/graph_hash.txt @@ -1 +1 @@ -517608d4d2699e09c6171648da38a4f924556cf25abd97875599acfdda5807e4 \ No newline at end of file +a354c7a9af04a38be394ae238e9a62e2595ecb8743b22059be566ce6ae3d04e3 \ No newline at end of file diff --git a/paddle_samples/PaddleNLP/bert-large-cased/graph_hash.txt b/paddle_samples/PaddleNLP/bert-large-cased/graph_hash.txt index fc6adfa90..ad3495c1b 100644 --- a/paddle_samples/PaddleNLP/bert-large-cased/graph_hash.txt +++ b/paddle_samples/PaddleNLP/bert-large-cased/graph_hash.txt @@ -1 +1 @@ -2a46a550da3ca0bd5aa6157a26aff525a3bc69ff8f67fe35b4424303c12e2820 \ No newline at end of file +0466903aebf44985bfceefc28a584e3cb3a02dff9b5f97ece86036f9c028699f \ No newline at end of file diff --git a/paddle_samples/PaddleNLP/ernie-1.0/graph_hash.txt b/paddle_samples/PaddleNLP/ernie-1.0/graph_hash.txt index b1f3221de..0cc4c5e30 100644 --- a/paddle_samples/PaddleNLP/ernie-1.0/graph_hash.txt +++ b/paddle_samples/PaddleNLP/ernie-1.0/graph_hash.txt @@ -1 +1 @@ -b23ce390b79f214cdbd74ea52c32d6dc141d93b179a7bf75f94bb12e8bd91561 \ No newline at end of file +c9ed800499d66c07b5ca6c96890ce11a9817fdf0e4035df835ad70a7c5ad17a8 \ No newline at end of file diff --git a/paddle_samples/PaddleNLP/ernie-2.0-base-zh/graph_hash.txt b/paddle_samples/PaddleNLP/ernie-2.0-base-zh/graph_hash.txt index c2d79682d..fe4beeac0 100644 --- a/paddle_samples/PaddleNLP/ernie-2.0-base-zh/graph_hash.txt +++ b/paddle_samples/PaddleNLP/ernie-2.0-base-zh/graph_hash.txt @@ -1 +1 @@ -94a3256e834ecd7e836da57b44da751d75ef9e095b04ac00abc37a5e18a01390 \ No newline at end of file +ef5611bb26e673e2ee92c8ad010b0c487bfb12607a79f9fea860787480738fed \ No newline at end of file diff --git a/paddle_samples/PaddleNLP/ernie-2.0-large-zh/graph_hash.txt b/paddle_samples/PaddleNLP/ernie-2.0-large-zh/graph_hash.txt index 1d720dfb1..bba486c8a 100644 --- a/paddle_samples/PaddleNLP/ernie-2.0-large-zh/graph_hash.txt +++ b/paddle_samples/PaddleNLP/ernie-2.0-large-zh/graph_hash.txt @@ -1 +1 @@ -fa8d65ab7f6feea97d152f37822ac8b05f79c08caf95b37ad92f9e2d817e7c72 \ No newline at end of file +902b9d175dd1676f04ba6e8f5b2ad87211abfd9a24c4f08d4b76ed3f520ba07b \ No newline at end of file diff --git a/paddle_samples/PaddleNLP/ernie-3.0-base-zh/graph_hash.txt b/paddle_samples/PaddleNLP/ernie-3.0-base-zh/graph_hash.txt index c9a64382c..a673c4476 100644 --- a/paddle_samples/PaddleNLP/ernie-3.0-base-zh/graph_hash.txt +++ b/paddle_samples/PaddleNLP/ernie-3.0-base-zh/graph_hash.txt @@ -1 +1 @@ -548e0e93a810037492943c7eb60592dbd3cdab50b07662af2ccb1ecd723c2e6f \ No newline at end of file +acc9f3385881ba07aa4e73927e23f75ccca7b40cab1ff80852ea457e4afc130f \ No newline at end of file diff --git a/paddle_samples/PaddleNLP/ernie-3.0-medium-zh/graph_hash.txt b/paddle_samples/PaddleNLP/ernie-3.0-medium-zh/graph_hash.txt index 5621f2978..7e31296fc 100644 --- a/paddle_samples/PaddleNLP/ernie-3.0-medium-zh/graph_hash.txt +++ b/paddle_samples/PaddleNLP/ernie-3.0-medium-zh/graph_hash.txt @@ -1 +1 @@ -db93f51be889fecf20bbd7cab1124b61fd36d550f434c0aa7a70b25c651d09ae \ No newline at end of file +a7942a9fd296f68327163c1e60c4b16107d12bc83a9c4e7c87d75c054c7f28f1 \ No newline at end of file diff --git a/paddle_samples/PaddleNLP/ernie-3.0-micro-zh/graph_hash.txt b/paddle_samples/PaddleNLP/ernie-3.0-micro-zh/graph_hash.txt index 9bf8cb337..37537b642 100644 --- a/paddle_samples/PaddleNLP/ernie-3.0-micro-zh/graph_hash.txt +++ b/paddle_samples/PaddleNLP/ernie-3.0-micro-zh/graph_hash.txt @@ -1 +1 @@ -1bdf89e29eaab83b48f9def1a32e5de256cee5d80cf0d0eee3d69eb0e5fc918a \ No newline at end of file +8573fb4034792407139258086ed806a651c13264dcb1e316f59f32e208a72b58 \ No newline at end of file diff --git a/paddle_samples/PaddleNLP/ernie-3.0-xbase-zh/graph_hash.txt b/paddle_samples/PaddleNLP/ernie-3.0-xbase-zh/graph_hash.txt index 2d0445e20..910b731b6 100644 --- a/paddle_samples/PaddleNLP/ernie-3.0-xbase-zh/graph_hash.txt +++ b/paddle_samples/PaddleNLP/ernie-3.0-xbase-zh/graph_hash.txt @@ -1 +1 @@ -c6f960fc9a50eeb396e6a394b8994fd1ecd1bcb49fc2ca1c3623edbff2cb0eb3 \ No newline at end of file +383d9094b90db2217fffc63380bca4a0ceda563ebd12c5ca468b4047c29c0df0 \ No newline at end of file diff --git a/paddle_samples/PaddleNLP/ernie-m-large/graph_hash.txt b/paddle_samples/PaddleNLP/ernie-m-large/graph_hash.txt index 84471eac9..84b11a52b 100644 --- a/paddle_samples/PaddleNLP/ernie-m-large/graph_hash.txt +++ b/paddle_samples/PaddleNLP/ernie-m-large/graph_hash.txt @@ -1 +1 @@ -6ec2b7a241cf5af899ead45f59561d8c09fc3198aae4733d77f6365d641a6b5c \ No newline at end of file +75a3d8dc450f6da36d42b5ab92490dca71318c1e5377035819e60cf5462a1d60 \ No newline at end of file diff --git a/paddle_samples/PaddleNLP/ernie-search-large-cross-encoder-marco-en/graph_hash.txt b/paddle_samples/PaddleNLP/ernie-search-large-cross-encoder-marco-en/graph_hash.txt index 6b7168392..212f99670 100644 --- a/paddle_samples/PaddleNLP/ernie-search-large-cross-encoder-marco-en/graph_hash.txt +++ b/paddle_samples/PaddleNLP/ernie-search-large-cross-encoder-marco-en/graph_hash.txt @@ -1 +1 @@ -9fd685e22bd085cb63b9491c25d9a42839c343546790add95924e300a1bfc094 \ No newline at end of file +aed756307ed1fa57719df17a7008b2bb0a02d84e786c567a7c3956edca9d9a53 \ No newline at end of file diff --git a/paddle_samples/PaddleNLP/ernie-tiny/graph_hash.txt b/paddle_samples/PaddleNLP/ernie-tiny/graph_hash.txt index 26a639797..210123180 100644 --- a/paddle_samples/PaddleNLP/ernie-tiny/graph_hash.txt +++ b/paddle_samples/PaddleNLP/ernie-tiny/graph_hash.txt @@ -1 +1 @@ -1ab1e873d42514b1fca6adc397e45033099ec5966dc03082c2c497203373f824 \ No newline at end of file +55159c6079c55a97e545090c0c3655aba5f92e181f5cfd6a1cb859fbc4666d9e \ No newline at end of file diff --git a/paddle_samples/PaddleNLP/ppminilm-6l-768h/graph_hash.txt b/paddle_samples/PaddleNLP/ppminilm-6l-768h/graph_hash.txt index eff266d16..7a4d2552b 100644 --- a/paddle_samples/PaddleNLP/ppminilm-6l-768h/graph_hash.txt +++ b/paddle_samples/PaddleNLP/ppminilm-6l-768h/graph_hash.txt @@ -1 +1 @@ -91520475abb0fcb88270eac4fd1b785ca926bbd5d6e4fb2364a1539d764620e4 \ No newline at end of file +ac4cd9ff971013b6eb27cba7622c2815500a33152a6738066e0acab11e77cef2 \ No newline at end of file diff --git a/paddle_samples/PaddleNLP/prophetnet-large-uncased/graph_hash.txt b/paddle_samples/PaddleNLP/prophetnet-large-uncased/graph_hash.txt index c6956af49..ff75f3629 100644 --- a/paddle_samples/PaddleNLP/prophetnet-large-uncased/graph_hash.txt +++ b/paddle_samples/PaddleNLP/prophetnet-large-uncased/graph_hash.txt @@ -1 +1 @@ -1785e57ebc48ba7295e3010e981dbdef953c11353ab44106aa0b2389da66a524 \ No newline at end of file +116a8a8f3c3bb1c2052f6ef9bb8340824db84e11315c8984ee5674ef6d0929ee \ No newline at end of file diff --git a/paddle_samples/PaddleNLP/rocketqa-base-cross-encoder/graph_hash.txt b/paddle_samples/PaddleNLP/rocketqa-base-cross-encoder/graph_hash.txt index aea163ede..531b4023c 100644 --- a/paddle_samples/PaddleNLP/rocketqa-base-cross-encoder/graph_hash.txt +++ b/paddle_samples/PaddleNLP/rocketqa-base-cross-encoder/graph_hash.txt @@ -1 +1 @@ -c8a7a60f61d78d967326f0907d58699e883998e68c9958aca32b46d1828831d5 \ No newline at end of file +1512b21895c2b5abad945f2f92dc06d2d3849767fea3c829091f221d3cfd6ab0 \ No newline at end of file diff --git a/paddle_samples/PaddleNLP/rocketqa-medium-cross-encoder/graph_hash.txt b/paddle_samples/PaddleNLP/rocketqa-medium-cross-encoder/graph_hash.txt index 8d56525eb..8c491fd39 100644 --- a/paddle_samples/PaddleNLP/rocketqa-medium-cross-encoder/graph_hash.txt +++ b/paddle_samples/PaddleNLP/rocketqa-medium-cross-encoder/graph_hash.txt @@ -1 +1 @@ -3b6f4591bccb1c69034d3d8ac92a4dc437903c7abcf585efb76e544507e56af3 \ No newline at end of file +c1e830f8a155dd7fc2d13311651a6050375fb7a735828107357e1ac0c9cb4c4e \ No newline at end of file diff --git a/paddle_samples/PaddleNLP/rocketqa-mini-cross-encoder/graph_hash.txt b/paddle_samples/PaddleNLP/rocketqa-mini-cross-encoder/graph_hash.txt index ca2e13c87..e883abbad 100644 --- a/paddle_samples/PaddleNLP/rocketqa-mini-cross-encoder/graph_hash.txt +++ b/paddle_samples/PaddleNLP/rocketqa-mini-cross-encoder/graph_hash.txt @@ -1 +1 @@ -e478da94c8f176161f7e73ee5ecd9e885afc48a8c86af7554df492d87e5ba9ac \ No newline at end of file +44e3d3ff6fdc2e7e0977fbb7516679f3f34ab79620962047aec37f9b0e1d5d4e \ No newline at end of file diff --git a/paddle_samples/PaddleNLP/skep_ernie_1.0_large_ch/graph_hash.txt b/paddle_samples/PaddleNLP/skep_ernie_1.0_large_ch/graph_hash.txt index 4b2add4a2..eece85432 100644 --- a/paddle_samples/PaddleNLP/skep_ernie_1.0_large_ch/graph_hash.txt +++ b/paddle_samples/PaddleNLP/skep_ernie_1.0_large_ch/graph_hash.txt @@ -1 +1 @@ -2a1a38318cef4b3970cd82bd45509354f5a002c3b59fc4c7f6afe215529b412c \ No newline at end of file +068a0a67e149965ea21b160a07c37c9ce66cbe43c5578d091bef028f79f49750 \ No newline at end of file diff --git a/paddle_samples/PaddleNLP/skep_ernie_2.0_large_en/graph_hash.txt b/paddle_samples/PaddleNLP/skep_ernie_2.0_large_en/graph_hash.txt index c453e6643..a0196cc39 100644 --- a/paddle_samples/PaddleNLP/skep_ernie_2.0_large_en/graph_hash.txt +++ b/paddle_samples/PaddleNLP/skep_ernie_2.0_large_en/graph_hash.txt @@ -1 +1 @@ -4a787b67f80f202bb4fca25275211cc35035ea9c6f4921e941b53bf591852527 \ No newline at end of file +d1e3cd0e88a307852394956e5b0f7018852ddecf371bc0880875df594c2bea31 \ No newline at end of file diff --git a/paddle_samples/PaddleNLP/utc-large/graph_hash.txt b/paddle_samples/PaddleNLP/utc-large/graph_hash.txt index 2f531b703..eaa331df4 100644 --- a/paddle_samples/PaddleNLP/utc-large/graph_hash.txt +++ b/paddle_samples/PaddleNLP/utc-large/graph_hash.txt @@ -1 +1 @@ -67222ab4e269a401723327ed54c55ec3ab6cf90029364ba472cfb69efc9645bb \ No newline at end of file +1314b44148e5f4b0ba88b9dae6a7981253520037cb99de0319995b6f1066fddd \ No newline at end of file diff --git a/paddle_samples/PaddleNLP/utc-xbase/graph_hash.txt b/paddle_samples/PaddleNLP/utc-xbase/graph_hash.txt index a30635948..9474567aa 100644 --- a/paddle_samples/PaddleNLP/utc-xbase/graph_hash.txt +++ b/paddle_samples/PaddleNLP/utc-xbase/graph_hash.txt @@ -1 +1 @@ -9f21603d1a47f6af22d1f0548310f43165409cb300e1f8026a7164b34125fd43 \ No newline at end of file +08b0549205d542e9fa66c2ed443095330438bf2f271f3f575153d794ef0c4a56 \ No newline at end of file From 961e06ddc53cdfd9c936ea760ef0b7c4061ad1fc Mon Sep 17 00:00:00 2001 From: Liu Yiqun Date: Tue, 20 Jan 2026 20:47:28 +0800 Subject: [PATCH 07/10] Update pre-commit. --- .pre-commit-config.yaml | 3 ++- 1 file changed, 2 insertions(+), 1 deletion(-) diff --git a/.pre-commit-config.yaml b/.pre-commit-config.yaml index 42a5c9b25..4a3cd4fd2 100755 --- a/.pre-commit-config.yaml +++ b/.pre-commit-config.yaml @@ -9,7 +9,8 @@ repos: rev: v0.14.4 hooks: - id: ruff-check - args: [--fix, --exit-non-zero-on-fix, --no-cache, --exclude=samples] + args: [--fix, --exit-non-zero-on-fix, --no-cache] + exclude: (^|/)(samples|paddle_samples)/ - repo: https://github.com/Lucas-C/pre-commit-hooks.git rev: v1.5.1 From b245ff0a56082905e2d4f53628ed9d32c955b2d0 Mon Sep 17 00:00:00 2001 From: Liu Yiqun Date: Wed, 21 Jan 2026 11:05:06 +0800 Subject: [PATCH 08/10] Minor optimization. --- graph_net/test/get_incorrect_models_test.sh | 9 +++++---- graph_net_rpc/sample_remote_executor.py | 6 +++--- 2 files changed, 8 insertions(+), 7 deletions(-) diff --git a/graph_net/test/get_incorrect_models_test.sh b/graph_net/test/get_incorrect_models_test.sh index cb62d99d4..49d269db9 100755 --- a/graph_net/test/get_incorrect_models_test.sh +++ b/graph_net/test/get_incorrect_models_test.sh @@ -8,14 +8,15 @@ PROJECT_ROOT=$(dirname "$GRAPH_NET_DIR") # 将项目根目录加入Python路径 export PYTHONPATH="$PROJECT_ROOT:$PYTHONPATH" -TOLERANCE_LIST=(-2 -1 0 1 2) +TOLERANCE_LIST=(-5 -4 -3 -2 -1 0 1) LOG_FILE_PATH="log_file_for_test.txt" python3 - < Date: Wed, 21 Jan 2026 11:14:06 +0800 Subject: [PATCH 09/10] Fix a sample to remove inf and avoid big number. --- .../PP-ShiTuV2_det/subgraph_1/weight_meta.py | 702 ++++++++---------- 1 file changed, 312 insertions(+), 390 deletions(-) diff --git a/paddle_samples/PaddleX/PP-ShiTuV2_det/subgraph_1/weight_meta.py b/paddle_samples/PaddleX/PP-ShiTuV2_det/subgraph_1/weight_meta.py index 00e09a96c..65ffb9c96 100644 --- a/paddle_samples/PaddleX/PP-ShiTuV2_det/subgraph_1/weight_meta.py +++ b/paddle_samples/PaddleX/PP-ShiTuV2_det/subgraph_1/weight_meta.py @@ -551,10 +551,10 @@ class Program_weight_tensor_parameter_46: original_name = "batch_norm2d_85.w_2" shape = [128] dtype = "float32" - min_val = float("58247.6") - max_val = float("20544400000.0") - mean = float("675237000.0") - std = float("2138970000.0") + min_val = float("-582.476") + max_val = float("2054.44") + mean = float("675.237") + std = float("213.897") data = None @@ -612,9 +612,9 @@ class Program_weight_tensor_parameter_51: shape = [128] dtype = "float32" min_val = float("0.000582512") - max_val = float("3249090000.0") - mean = float("60770300.0") - std = float("377372000.0") + max_val = float("3249.09") + mean = float("607.703") + std = float("377.372") data = None @@ -671,10 +671,10 @@ class Program_weight_tensor_parameter_56: original_name = "batch_norm2d_83.w_2" shape = [128] dtype = "float32" - min_val = float("2857.87") - max_val = float("412905000.0") - mean = float("15818900.0") - std = float("54657900.0") + min_val = float("-28.5787") + max_val = float("412.905") + mean = float("158.189") + std = float("54.6579") data = None @@ -732,9 +732,9 @@ class Program_weight_tensor_parameter_61: shape = [128] dtype = "float32" min_val = float("0.00900684") - max_val = float("3556900000.0") - mean = float("43511400.0") - std = float("330022000.0") + max_val = float("3556.90") + mean = float("435.114") + std = float("330.022") data = None @@ -792,9 +792,9 @@ class Program_weight_tensor_parameter_66: shape = [128] dtype = "float32" min_val = float("1872.14") - max_val = float("47439700.0") - mean = float("2879840.0") - std = float("7270230.0") + max_val = float("4743.97") + mean = float("2879.84") + std = float("727.023") data = None @@ -852,9 +852,9 @@ class Program_weight_tensor_parameter_71: shape = [128] dtype = "float32" min_val = float("0.219354") - max_val = float("10336600000.0") - mean = float("92021900.0") - std = float("912436000.0") + max_val = float("1033.66") + mean = float("920.219") + std = float("91.2436") data = None @@ -911,10 +911,10 @@ class Program_weight_tensor_parameter_76: original_name = "batch_norm2d_79.w_2" shape = [128] dtype = "float32" - min_val = float("9659.32") - max_val = float("2853920000.0") - mean = float("67779800.0") - std = float("311856000.0") + min_val = float("96.5932") + max_val = float("2853.92") + mean = float("677.798") + std = float("311.856") data = None @@ -972,9 +972,9 @@ class Program_weight_tensor_parameter_81: shape = [128] dtype = "float32" min_val = float("0.00619647") - max_val = float("8816430000000.0") - mean = float("70110200000.0") - std = float("776208000000.0") + max_val = float("8816.43") + mean = float("701.102") + std = float("776.208") data = None @@ -983,10 +983,10 @@ class Program_weight_tensor_parameter_82: original_name = "batch_norm2d_78.w_1" shape = [128] dtype = "float32" - min_val = float("-121377.0") - max_val = float("1669060.0") - mean = float("15548.7") - std = float("148946.0") + min_val = float("-12137.7") + max_val = float("16690.6") + mean = float("1554.87") + std = float("1489.46") data = None @@ -1053,10 +1053,8 @@ class Program_weight_tensor_parameter_88: original_name = "batch_norm2d_77.w_2" shape = [128] dtype = "float32" - min_val = float("1256560000.0") - max_val = float("2.75381e+20") - mean = float("6.70377e+18") - std = float("inf") + min_val = float("125.656") + max_val = float("2753.81") data = None @@ -1065,10 +1063,10 @@ class Program_weight_tensor_parameter_89: original_name = "batch_norm2d_77.w_1" shape = [128] dtype = "float32" - min_val = float("-18448400000.0") - max_val = float("13078200000.0") - mean = float("-31256100.0") - std = float("2219470000.0") + min_val = float("-1844.84") + max_val = float("1307.82") + mean = float("-312.561") + std = float("221.947") data = None @@ -1089,10 +1087,10 @@ class Program_weight_tensor_parameter_91: original_name = "batch_norm2d_76.b_0" shape = [128] dtype = "float32" - min_val = float("-274863.0") - max_val = float("103392.0") - mean = float("-3753.37") - std = float("32628.5") + min_val = float("-2748.63") + max_val = float("1033.92") + mean = float("-375.337") + std = float("326.285") data = None @@ -1101,10 +1099,10 @@ class Program_weight_tensor_parameter_92: original_name = "batch_norm2d_76.w_0" shape = [128] dtype = "float32" - min_val = float("-116014.0") - max_val = float("105656.0") - mean = float("-1460.97") - std = float("22152.4") + min_val = float("-1160.14") + max_val = float("1056.56") + mean = float("-146.097") + std = float("221.524") data = None @@ -1114,9 +1112,7 @@ class Program_weight_tensor_parameter_93: shape = [128] dtype = "float32" min_val = float("1.40568e-05") - max_val = float("4.05002e+20") - mean = float("1.03691e+19") - std = float("inf") + max_val = float("4050.02") data = None @@ -1125,10 +1121,10 @@ class Program_weight_tensor_parameter_94: original_name = "batch_norm2d_76.w_1" shape = [128] dtype = "float32" - min_val = float("-33518000000.0") - max_val = float("9649450000.0") - mean = float("-209845000.0") - std = float("3196110000.0") + min_val = float("-3351.80") + max_val = float("9649.45") + mean = float("-209.845") + std = float("319.611") data = None @@ -1173,10 +1169,8 @@ class Program_weight_tensor_parameter_98: original_name = "batch_norm2d_75.w_2" shape = [128] dtype = "float32" - min_val = float("37547100000.0") - max_val = float("2.3025e+21") - mean = float("4.36871e+19") - std = float("inf") + min_val = float("375.471") + max_val = float("2302.50") data = None @@ -1185,10 +1179,10 @@ class Program_weight_tensor_parameter_99: original_name = "batch_norm2d_75.w_1" shape = [128] dtype = "float32" - min_val = float("-3777800000.0") - max_val = float("25793100000.0") - mean = float("647896000.0") - std = float("3109750000.0") + min_val = float("-3777.80") + max_val = float("2579.31") + mean = float("647.896") + std = float("310.975") data = None @@ -1234,9 +1228,7 @@ class Program_weight_tensor_parameter_103: shape = [128] dtype = "float32" min_val = float("4.68528e-06") - max_val = float("6.52868e+22") - mean = float("5.83095e+20") - std = float("inf") + max_val = float("6528.68") data = None @@ -1245,10 +1237,10 @@ class Program_weight_tensor_parameter_104: original_name = "batch_norm2d_74.w_1" shape = [128] dtype = "float32" - min_val = float("-36490500000.0") - max_val = float("482921000000.0") - mean = float("3732870000.0") - std = float("42791300000.0") + min_val = float("-3649.05") + max_val = float("4829.21") + mean = float("373.287") + std = float("427.913") data = None @@ -1293,10 +1285,8 @@ class Program_weight_tensor_parameter_108: original_name = "batch_norm2d_73.w_2" shape = [128] dtype = "float32" - min_val = float("39533200000.0") - max_val = float("2.72523e+20") - mean = float("6.79091e+18") - std = float("inf") + min_val = float("395.332") + max_val = float("2725.23") data = None @@ -1305,10 +1295,10 @@ class Program_weight_tensor_parameter_109: original_name = "batch_norm2d_73.w_1" shape = [128] dtype = "float32" - min_val = float("-4719580000.0") - max_val = float("9380190000.0") - mean = float("36117900.0") - std = float("1408010000.0") + min_val = float("-4719.58") + max_val = float("9380.19") + mean = float("361.179") + std = float("1408.01") data = None @@ -1354,9 +1344,7 @@ class Program_weight_tensor_parameter_113: shape = [128] dtype = "float32" min_val = float("4.3711e-05") - max_val = float("6.69868e+20") - mean = float("1.90444e+19") - std = float("inf") + max_val = float("6698.68") data = None @@ -1365,10 +1353,10 @@ class Program_weight_tensor_parameter_114: original_name = "batch_norm2d_72.w_1" shape = [128] dtype = "float32" - min_val = float("-32210600000.0") - max_val = float("19296400000.0") - mean = float("-222346000.0") - std = float("3712630000.0") + min_val = float("-3221.06") + max_val = float("1929.64") + mean = float("-222.346") + std = float("371.263") data = None @@ -1413,10 +1401,8 @@ class Program_weight_tensor_parameter_118: original_name = "batch_norm2d_71.w_2" shape = [128] dtype = "float32" - min_val = float("131404000000.0") - max_val = float("6.8141e+21") - mean = float("1.01263e+20") - std = float("inf") + min_val = float("131.404") + max_val = float("6814.10") data = None @@ -1425,10 +1411,10 @@ class Program_weight_tensor_parameter_119: original_name = "batch_norm2d_71.w_1" shape = [128] dtype = "float32" - min_val = float("-44285100000.0") - max_val = float("3857900000.0") - mean = float("-921017000.0") - std = float("5922380000.0") + min_val = float("-4428.51") + max_val = float("3857.90") + mean = float("-921.017") + std = float("592.238") data = None @@ -1474,9 +1460,7 @@ class Program_weight_tensor_parameter_123: shape = [128] dtype = "float32" min_val = float("3.88607e-06") - max_val = float("3.93957e+20") - mean = float("1.01325e+19") - std = float("inf") + max_val = float("3939.57") data = None @@ -1485,10 +1469,10 @@ class Program_weight_tensor_parameter_124: original_name = "batch_norm2d_70.w_1" shape = [128] dtype = "float32" - min_val = float("-30152100000.0") - max_val = float("10858700000.0") - mean = float("-434912000.0") - std = float("4031290000.0") + min_val = float("-3015.21") + max_val = float("1085.87") + mean = float("-434.912") + std = float("403.129") data = None @@ -1556,9 +1540,9 @@ class Program_weight_tensor_parameter_130: shape = [128] dtype = "float32" min_val = float("3228.36") - max_val = float("101589000.0") - mean = float("4930090.0") - std = float("14493200.0") + max_val = float("10158.9") + mean = float("4930.090") + std = float("1449.32") data = None @@ -1616,9 +1600,9 @@ class Program_weight_tensor_parameter_135: shape = [128] dtype = "float32" min_val = float("0.003342") - max_val = float("260327000.0") - mean = float("5286810.0") - std = float("29247200.0") + max_val = float("2603.27") + mean = float("528.681") + std = float("2924.72") data = None @@ -1676,9 +1660,9 @@ class Program_weight_tensor_parameter_140: shape = [128] dtype = "float32" min_val = float("818.739") - max_val = float("44282300.0") - mean = float("1670460.0") - std = float("6291510.0") + max_val = float("4428.23") + mean = float("1670.46") + std = float("629.151") data = None @@ -1736,9 +1720,9 @@ class Program_weight_tensor_parameter_145: shape = [128] dtype = "float32" min_val = float("0.00133377") - max_val = float("6346100000.0") - mean = float("54220400.0") - std = float("558920000.0") + max_val = float("6346.10") + mean = float("542.204") + std = float("5589.20") data = None @@ -1796,9 +1780,9 @@ class Program_weight_tensor_parameter_150: shape = [128] dtype = "float32" min_val = float("473.663") - max_val = float("19018900.0") - mean = float("1142690.0") - std = float("3006780.0") + max_val = float("19018.9") + mean = float("1142.69") + std = float("3006.78") data = None @@ -1856,9 +1840,9 @@ class Program_weight_tensor_parameter_155: shape = [128] dtype = "float32" min_val = float("0.00890485") - max_val = float("924151000.0") - mean = float("10677200.0") - std = float("84650300.0") + max_val = float("9241.51") + mean = float("1067.72") + std = float("8465.03") data = None @@ -1916,9 +1900,9 @@ class Program_weight_tensor_parameter_160: shape = [128] dtype = "float32" min_val = float("287.46") - max_val = float("42264800.0") - mean = float("830575.0") - std = float("4131720.0") + max_val = float("4226.48") + mean = float("830.575") + std = float("413.172") data = None @@ -1976,9 +1960,9 @@ class Program_weight_tensor_parameter_165: shape = [128] dtype = "float32" min_val = float("0.000128292") - max_val = float("7539160000000.0") - mean = float("120339000000.0") - std = float("872477000000.0") + max_val = float("7539.16") + mean = float("1203.39") + std = float("872.477") data = None @@ -1987,10 +1971,10 @@ class Program_weight_tensor_parameter_166: original_name = "batch_norm2d_62.w_1" shape = [128] dtype = "float32" - min_val = float("-5927410.0") - max_val = float("3255460.0") - mean = float("-21456.2") - std = float("605831.0") + min_val = float("-5927.41") + max_val = float("3255.46") + mean = float("-2145.62") + std = float("605.831") data = None @@ -2096,9 +2080,9 @@ class Program_weight_tensor_parameter_175: shape = [128] dtype = "float32" min_val = float("0.000156346") - max_val = float("391558000.0") - mean = float("3072800.0") - std = float("34472500.0") + max_val = float("3915.58") + mean = float("307.28") + std = float("344.725") data = None @@ -2216,9 +2200,9 @@ class Program_weight_tensor_parameter_185: shape = [128] dtype = "float32" min_val = float("5.86101e-05") - max_val = float("12261400000000.0") - mean = float("105002000000.0") - std = float("1081960000000.0") + max_val = float("12261.4") + mean = float("1050.02") + std = float("10819.6") data = None @@ -2227,10 +2211,10 @@ class Program_weight_tensor_parameter_186: original_name = "batch_norm2d_30.w_1" shape = [128] dtype = "float32" - min_val = float("-528399.0") - max_val = float("1345590.0") + min_val = float("-5283.99") + max_val = float("1345.59") mean = float("-1085.89") - std = float("134658.0") + std = float("1346.58") data = None @@ -2275,10 +2259,10 @@ class Program_weight_tensor_parameter_190: original_name = "batch_norm2d_58.w_2" shape = [128] dtype = "float32" - min_val = float("14777.6") - max_val = float("6550930000.0") - mean = float("152320000.0") - std = float("760843000.0") + min_val = float("147.776") + max_val = float("6550.93") + mean = float("1523.20") + std = float("760.843") data = None @@ -2656,9 +2640,9 @@ class Program_weight_tensor_parameter_225: shape = [128] dtype = "float32" min_val = float("3.53324e-11") - max_val = float("61754200000.0") - mean = float("882636000.0") - std = float("5662340000.0") + max_val = float("6175.42") + mean = float("882.636") + std = float("5662.34") data = None @@ -2715,10 +2699,8 @@ class Program_weight_tensor_parameter_230: original_name = "batch_norm2d_50.w_2" shape = [128] dtype = "float32" - min_val = float("10737900000.0") - max_val = float("7.12963e+20") - mean = float("9.05393e+18") - std = float("inf") + min_val = float("1073.79") + max_val = float("7129.63") data = None @@ -2727,10 +2709,10 @@ class Program_weight_tensor_parameter_231: original_name = "batch_norm2d_50.w_1" shape = [128] dtype = "float32" - min_val = float("-6512240000.0") - max_val = float("32799100000.0") - mean = float("166414000.0") - std = float("3188900000.0") + min_val = float("-6512.24") + max_val = float("3279.91") + mean = float("1664.14") + std = float("3188.90") data = None @@ -3035,10 +3017,10 @@ class Program_weight_tensor_parameter_260: original_name = "batch_norm2d_47.w_2" shape = [128] dtype = "float32" - min_val = float("1289980000.0") - max_val = float("1.01952e+19") - mean = float("1.77883e+17") - std = float("1.09482e+18") + min_val = float("128.998") + max_val = float("10195.2") + mean = float("1778.83") + std = float("109.482") data = None @@ -3047,10 +3029,10 @@ class Program_weight_tensor_parameter_261: original_name = "batch_norm2d_47.w_1" shape = [128] dtype = "float32" - min_val = float("-969451000.0") - max_val = float("1719590000.0") - mean = float("19588100.0") - std = float("230945000.0") + min_val = float("-9694.51") + max_val = float("17195.9") + mean = float("1958.81") + std = float("2309.45") data = None @@ -3096,9 +3078,7 @@ class Program_weight_tensor_parameter_265: shape = [128] dtype = "float32" min_val = float("0.000180454") - max_val = float("2.04952e+21") - mean = float("1.68005e+19") - std = float("inf") + max_val = float("2049.52") data = None @@ -3107,10 +3087,10 @@ class Program_weight_tensor_parameter_266: original_name = "batch_norm2d_46.w_1" shape = [128] dtype = "float32" - min_val = float("-760651000.0") - max_val = float("48053000000.0") - mean = float("482595000.0") - std = float("4368170000.0") + min_val = float("-7606.51") + max_val = float("4805.30") + mean = float("482.595") + std = float("4368.17") data = None @@ -3155,10 +3135,8 @@ class Program_weight_tensor_parameter_270: original_name = "batch_norm2d_42.w_2" shape = [128] dtype = "float32" - min_val = float("29624000000.0") - max_val = float("3.50963e+20") - mean = float("8.11164e+18") - std = float("inf") + min_val = float("296.240") + max_val = float("3509.63") data = None @@ -3167,10 +3145,10 @@ class Program_weight_tensor_parameter_271: original_name = "batch_norm2d_42.w_1" shape = [128] dtype = "float32" - min_val = float("-25856000000.0") - max_val = float("31882100000.0") - mean = float("-168986000.0") - std = float("3911860000.0") + min_val = float("-2585.60") + max_val = float("3188.21") + mean = float("-1689.86") + std = float("3911.86") data = None @@ -3487,10 +3465,10 @@ class Program_weight_tensor_parameter_301: original_name = "batch_norm2d_36.w_1" shape = [128] dtype = "float32" - min_val = float("-10862500000.0") - max_val = float("21049400000.0") - mean = float("215672000.0") - std = float("2658170000.0") + min_val = float("-1086.25") + max_val = float("21049.4") + mean = float("2156.72") + std = float("2658.17") data = None @@ -3795,10 +3773,8 @@ class Program_weight_tensor_parameter_330: original_name = "batch_norm2d_29.w_2" shape = [128] dtype = "float32" - min_val = float("34782800000.0") - max_val = float("1.29377e+22") - mean = float("1.87313e+20") - std = float("inf") + min_val = float("347.828") + max_val = float("12937.7") data = None @@ -3807,10 +3783,10 @@ class Program_weight_tensor_parameter_331: original_name = "batch_norm2d_29.w_1" shape = [128] dtype = "float32" - min_val = float("-196325000000.0") - max_val = float("43131900000.0") - mean = float("-956323000.0") - std = float("18669900000.0") + min_val = float("-1963.25") + max_val = float("4313.19") + mean = float("-956.323") + std = float("1866.99") data = None @@ -3855,10 +3831,8 @@ class Program_weight_tensor_parameter_335: original_name = "batch_norm2d_28.w_2" shape = [128] dtype = "float32" - min_val = float("12554200000.0") - max_val = float("4.2572e+20") - mean = float("7.03993e+18") - std = float("inf") + min_val = float("125.542") + max_val = float("4257.20") data = None @@ -3867,10 +3841,10 @@ class Program_weight_tensor_parameter_336: original_name = "batch_norm2d_28.w_1" shape = [128] dtype = "float32" - min_val = float("-6273210000.0") - max_val = float("5488470000.0") - mean = float("-65105300.0") - std = float("1089890000.0") + min_val = float("-6273.21") + max_val = float("5488.47") + mean = float("-651.053") + std = float("1089.89") data = None @@ -3915,10 +3889,8 @@ class Program_weight_tensor_parameter_340: original_name = "batch_norm2d_27.w_2" shape = [128] dtype = "float32" - min_val = float("20311000000.0") - max_val = float("1.84851e+21") - mean = float("3.34316e+19") - std = float("inf") + min_val = float("20.3110") + max_val = float("1848.51") data = None @@ -3927,10 +3899,10 @@ class Program_weight_tensor_parameter_341: original_name = "batch_norm2d_27.w_1" shape = [128] dtype = "float32" - min_val = float("-32024200000.0") - max_val = float("21390200000.0") - mean = float("129123000.0") - std = float("3778280000.0") + min_val = float("-3202.42") + max_val = float("21390.2") + mean = float("1291.23") + std = float("3778.28") data = None @@ -3975,10 +3947,8 @@ class Program_weight_tensor_parameter_345: original_name = "batch_norm2d_26.w_2" shape = [1280] dtype = "float32" - min_val = float("6039170000.0") - max_val = float("6.78926e+20") - mean = float("2.28536e+18") - std = float("inf") + min_val = float("60.3917") + max_val = float("6789.26") data = None @@ -3987,10 +3957,10 @@ class Program_weight_tensor_parameter_346: original_name = "batch_norm2d_26.w_1" shape = [1280] dtype = "float32" - min_val = float("-14069300000.0") - max_val = float("7056590000.0") - mean = float("-8576950.0") - std = float("746888000.0") + min_val = float("-1406.93") + max_val = float("7056.59") + mean = float("-857.695") + std = float("7468.88") data = None @@ -4084,9 +4054,7 @@ class Program_weight_tensor_parameter_354: shape = [1280] dtype = "float32" min_val = float("3.36739e-06") - max_val = float("6.40361e+20") - mean = float("1.15563e+18") - std = float("inf") + max_val = float("6403.61") data = None @@ -4095,10 +4063,10 @@ class Program_weight_tensor_parameter_355: original_name = "batch_norm2d_25.w_1" shape = [1280] dtype = "float32" - min_val = float("-5005020000.0") - max_val = float("11081300000.0") - mean = float("100970.0") - std = float("432630000.0") + min_val = float("-5005.02") + max_val = float("11081.3") + mean = float("1009.70") + std = float("4326.30") data = None @@ -4143,10 +4111,8 @@ class Program_weight_tensor_parameter_359: original_name = "batch_norm2d_24.w_2" shape = [1280] dtype = "float32" - min_val = float("1812810.0") - max_val = float("3.21017e+21") - mean = float("6.6749e+18") - std = float("inf") + min_val = float("18.12810") + max_val = float("3210.17") data = None @@ -4155,10 +4121,10 @@ class Program_weight_tensor_parameter_360: original_name = "batch_norm2d_24.w_1" shape = [1280] dtype = "float32" - min_val = float("-16396500000.0") - max_val = float("20826700000.0") - mean = float("36435300.0") - std = float("1141100000.0") + min_val = float("-1639.65") + max_val = float("20826.7") + mean = float("3643.53") + std = float("11411.0") data = None @@ -4252,9 +4218,7 @@ class Program_weight_tensor_parameter_368: shape = [640] dtype = "float32" min_val = float("1.61123e-07") - max_val = float("1.73247e+22") - mean = float("2.78801e+19") - std = float("inf") + max_val = float("1732.47") data = None @@ -4263,10 +4227,10 @@ class Program_weight_tensor_parameter_369: original_name = "batch_norm2d_23.w_1" shape = [640] dtype = "float32" - min_val = float("-42998100000.0") - max_val = float("39908100000.0") - mean = float("-17340100.0") - std = float("2350230000.0") + min_val = float("-4299.81") + max_val = float("3990.81") + mean = float("-1734.01") + std = float("2350.23") data = None @@ -4311,10 +4275,8 @@ class Program_weight_tensor_parameter_373: original_name = "batch_norm2d_22.w_2" shape = [640] dtype = "float32" - min_val = float("1290730000.0") - max_val = float("8.78844e+20") - mean = float("2.72846e+18") - std = float("inf") + min_val = float("12.9073") + max_val = float("8788.44") data = None @@ -4323,10 +4285,10 @@ class Program_weight_tensor_parameter_374: original_name = "batch_norm2d_22.w_1" shape = [640] dtype = "float32" - min_val = float("-24967400000.0") - max_val = float("15440700000.0") - mean = float("-87857400.0") - std = float("1512910000.0") + min_val = float("-2496.74") + max_val = float("15440.7") + mean = float("-878.574") + std = float("15129.1") data = None @@ -4372,9 +4334,7 @@ class Program_weight_tensor_parameter_378: shape = [640] dtype = "float32" min_val = float("2.85619e-06") - max_val = float("1.25829e+23") - mean = float("1.97856e+20") - std = float("inf") + max_val = float("12582.9") data = None @@ -4383,10 +4343,10 @@ class Program_weight_tensor_parameter_379: original_name = "batch_norm2d_21.w_1" shape = [640] dtype = "float32" - min_val = float("-11284800000.0") - max_val = float("103528000000.0") - mean = float("205898000.0") - std = float("4183220000.0") + min_val = float("-1128.48") + max_val = float("10352.8") + mean = float("2058.98") + std = float("4183.22") data = None @@ -4431,10 +4391,8 @@ class Program_weight_tensor_parameter_383: original_name = "batch_norm2d_20.w_2" shape = [640] dtype = "float32" - min_val = float("313117000.0") - max_val = float("3.12437e+22") - mean = float("6.42489e+19") - std = float("inf") + min_val = float("31.3117") + max_val = float("3124.37") data = None @@ -4443,10 +4401,10 @@ class Program_weight_tensor_parameter_384: original_name = "batch_norm2d_20.w_1" shape = [640] dtype = "float32" - min_val = float("-13702700000.0") - max_val = float("70694600000.0") - mean = float("181567000.0") - std = float("3102160000.0") + min_val = float("-1370.27") + max_val = float("7069.46") + mean = float("1815.67") + std = float("3102.16") data = None @@ -4492,9 +4450,7 @@ class Program_weight_tensor_parameter_388: shape = [640] dtype = "float32" min_val = float("7.95316e-07") - max_val = float("2.20108e+21") - mean = float("5.93378e+18") - std = float("inf") + max_val = float("2201.08") data = None @@ -4503,10 +4459,10 @@ class Program_weight_tensor_parameter_389: original_name = "batch_norm2d_19.w_1" shape = [640] dtype = "float32" - min_val = float("-14660200000.0") - max_val = float("41376300000.0") - mean = float("3122360.0") - std = float("1908930000.0") + min_val = float("-1466.02") + max_val = float("4137.63") + mean = float("312.236") + std = float("1908.93") data = None @@ -4551,10 +4507,8 @@ class Program_weight_tensor_parameter_393: original_name = "batch_norm2d_18.w_2" shape = [640] dtype = "float32" - min_val = float("1108370000.0") - max_val = float("5.8681e+20") - mean = float("3.0167e+18") - std = float("inf") + min_val = float("11.0837") + max_val = float("5868.10") data = None @@ -4563,10 +4517,10 @@ class Program_weight_tensor_parameter_394: original_name = "batch_norm2d_18.w_1" shape = [640] dtype = "float32" - min_val = float("-23696700000.0") - max_val = float("11167400000.0") - mean = float("17698800.0") - std = float("1192040000.0") + min_val = float("-2369.67") + max_val = float("11167.4") + mean = float("1769.88") + std = float("11920.4") data = None @@ -4612,9 +4566,7 @@ class Program_weight_tensor_parameter_398: shape = [640] dtype = "float32" min_val = float("1.64366e-05") - max_val = float("2.62001e+20") - mean = float("1.07981e+18") - std = float("inf") + max_val = float("2620.01") data = None @@ -4623,10 +4575,10 @@ class Program_weight_tensor_parameter_399: original_name = "batch_norm2d_17.w_1" shape = [640] dtype = "float32" - min_val = float("-41680700000.0") - max_val = float("3437810000.0") - mean = float("-89032600.0") - std = float("1735380000.0") + min_val = float("-4168.07") + max_val = float("3437.81") + mean = float("-890.326") + std = float("1735.38") data = None @@ -4671,10 +4623,8 @@ class Program_weight_tensor_parameter_403: original_name = "batch_norm2d_16.w_2" shape = [640] dtype = "float32" - min_val = float("698022000.0") - max_val = float("8.75483e+19") - mean = float("9.29516e+17") - std = float("inf") + min_val = float("69.8022") + max_val = float("8754.83") data = None @@ -4683,10 +4633,10 @@ class Program_weight_tensor_parameter_404: original_name = "batch_norm2d_16.w_1" shape = [640] dtype = "float32" - min_val = float("-6894760000.0") - max_val = float("4684640000.0") - mean = float("-33198100.0") - std = float("540811000.0") + min_val = float("-6894.76") + max_val = float("4684.64") + mean = float("-331.981") + std = float("5408.11") data = None @@ -4732,9 +4682,7 @@ class Program_weight_tensor_parameter_408: shape = [640] dtype = "float32" min_val = float("1.1929e-06") - max_val = float("3.93504e+20") - mean = float("1.88509e+18") - std = float("inf") + max_val = float("3935.04") data = None @@ -4743,10 +4691,10 @@ class Program_weight_tensor_parameter_409: original_name = "batch_norm2d_15.w_1" shape = [640] dtype = "float32" - min_val = float("-9042340000.0") - max_val = float("16873600000.0") - mean = float("51932200.0") - std = float("1074300000.0") + min_val = float("-9042.34") + max_val = float("16873.6") + mean = float("5193.22") + std = float("10743.0") data = None @@ -4791,10 +4739,8 @@ class Program_weight_tensor_parameter_413: original_name = "batch_norm2d_14.w_2" shape = [640] dtype = "float32" - min_val = float("1423200000.0") - max_val = float("4.42497e+20") - mean = float("1.82884e+18") - std = float("inf") + min_val = float("14.2320") + max_val = float("4424.97") data = None @@ -4803,10 +4749,10 @@ class Program_weight_tensor_parameter_414: original_name = "batch_norm2d_14.w_1" shape = [640] dtype = "float32" - min_val = float("-11531700000.0") - max_val = float("10098100000.0") - mean = float("-4629380.0") - std = float("1126350000.0") + min_val = float("-1153.17") + max_val = float("10098.1") + mean = float("-462.938") + std = float("11263.5") data = None @@ -4852,9 +4798,7 @@ class Program_weight_tensor_parameter_418: shape = [640] dtype = "float32" min_val = float("0.000284452") - max_val = float("6.8432e+21") - mean = float("1.23178e+19") - std = float("inf") + max_val = float("6843.20") data = None @@ -4863,10 +4807,10 @@ class Program_weight_tensor_parameter_419: original_name = "batch_norm2d_13.w_1" shape = [640] dtype = "float32" - min_val = float("-52727100000.0") - max_val = float("32822800000.0") - mean = float("-74949700.0") - std = float("2727810000.0") + min_val = float("-5272.71") + max_val = float("3282.28") + mean = float("-749.497") + std = float("2727.81") data = None @@ -4911,10 +4855,8 @@ class Program_weight_tensor_parameter_423: original_name = "batch_norm2d_12.w_2" shape = [640] dtype = "float32" - min_val = float("5812160000.0") - max_val = float("8.35144e+20") - mean = float("4.67478e+18") - std = float("inf") + min_val = float("58.1216") + max_val = float("8351.44") data = None @@ -4923,10 +4865,10 @@ class Program_weight_tensor_parameter_424: original_name = "batch_norm2d_12.w_1" shape = [640] dtype = "float32" - min_val = float("-21357400000.0") - max_val = float("19106800000.0") - mean = float("-14194600.0") - std = float("1680160000.0") + min_val = float("-2135.74") + max_val = float("1910.68") + mean = float("-141.946") + std = float("1680.16") data = None @@ -4972,9 +4914,7 @@ class Program_weight_tensor_parameter_428: shape = [320] dtype = "float32" min_val = float("4.15695e-08") - max_val = float("5.23365e+20") - mean = float("3.4446e+18") - std = float("inf") + max_val = float("5233.65") data = None @@ -4983,10 +4923,10 @@ class Program_weight_tensor_parameter_429: original_name = "batch_norm2d_11.w_1" shape = [320] dtype = "float32" - min_val = float("-14413500000.0") - max_val = float("10180900000.0") - mean = float("30528800.0") - std = float("1377360000.0") + min_val = float("-1441.35") + max_val = float("10180.9") + mean = float("3052.88") + std = float("13773.6") data = None @@ -5031,10 +4971,8 @@ class Program_weight_tensor_parameter_433: original_name = "batch_norm2d_10.w_2" shape = [320] dtype = "float32" - min_val = float("1546490000.0") - max_val = float("1.64206e+23") - mean = float("6.03107e+20") - std = float("inf") + min_val = float("15.4649") + max_val = float("1642.06") data = None @@ -5043,10 +4981,10 @@ class Program_weight_tensor_parameter_434: original_name = "batch_norm2d_10.w_1" shape = [320] dtype = "float32" - min_val = float("-416065000000.0") - max_val = float("99026300000.0") - mean = float("-1221850000.0") - std = float("24383200000.0") + min_val = float("-4160.65") + max_val = float("9902.63") + mean = float("-1221.85") + std = float("2438.32") data = None @@ -5092,9 +5030,7 @@ class Program_weight_tensor_parameter_438: shape = [320] dtype = "float32" min_val = float("0.110365") - max_val = float("3.00155e+21") - mean = float("2.22167e+19") - std = float("inf") + max_val = float("3001.55") data = None @@ -5103,10 +5039,10 @@ class Program_weight_tensor_parameter_439: original_name = "batch_norm2d_9.w_1" shape = [320] dtype = "float32" - min_val = float("-7671760000.0") - max_val = float("90259200000.0") - mean = float("604382000.0") - std = float("6808880000.0") + min_val = float("-7671.76") + max_val = float("9025.92") + mean = float("604.382") + std = float("6808.88") data = None @@ -5151,10 +5087,8 @@ class Program_weight_tensor_parameter_443: original_name = "batch_norm2d_8.w_2" shape = [320] dtype = "float32" - min_val = float("24902000000.0") - max_val = float("6.12093e+21") - mean = float("4.71114e+19") - std = float("inf") + min_val = float("24.9020") + max_val = float("6120.93") data = None @@ -5163,10 +5097,10 @@ class Program_weight_tensor_parameter_444: original_name = "batch_norm2d_8.w_1" shape = [320] dtype = "float32" - min_val = float("-15461400000.0") - max_val = float("57228900000.0") - mean = float("276379000.0") - std = float("4405570000.0") + min_val = float("-1546.14") + max_val = float("5722.89") + mean = float("276.379") + std = float("4405.57") data = None @@ -5212,9 +5146,7 @@ class Program_weight_tensor_parameter_448: shape = [160] dtype = "float32" min_val = float("0.00149326") - max_val = float("1.16217e+23") - mean = float("7.49654e+20") - std = float("inf") + max_val = float("1162.17") data = None @@ -5223,10 +5155,10 @@ class Program_weight_tensor_parameter_449: original_name = "batch_norm2d_7.w_1" shape = [160] dtype = "float32" - min_val = float("-179041000000.0") - max_val = float("5518870000.0") - mean = float("-1501810000.0") - std = float("14392900000.0") + min_val = float("-1790.41") + max_val = float("5518.87") + mean = float("-1501.81") + std = float("14392.9") data = None @@ -5271,10 +5203,8 @@ class Program_weight_tensor_parameter_453: original_name = "batch_norm2d_6.w_2" shape = [160] dtype = "float32" - min_val = float("37523100000.0") - max_val = float("7.09867e+21") - mean = float("8.02912e+19") - std = float("inf") + min_val = float("37.5231") + max_val = float("7098.67") data = None @@ -5283,10 +5213,10 @@ class Program_weight_tensor_parameter_454: original_name = "batch_norm2d_6.w_1" shape = [160] dtype = "float32" - min_val = float("-70317900000.0") - max_val = float("40189100000.0") - mean = float("-286138000.0") - std = float("7667220000.0") + min_val = float("-7031.79") + max_val = float("4018.91") + mean = float("-2861.38") + std = float("7667.22") data = None @@ -5332,9 +5262,7 @@ class Program_weight_tensor_parameter_458: shape = [160] dtype = "float32" min_val = float("5.88099") - max_val = float("4.35754e+23") - mean = float("3.37242e+21") - std = float("inf") + max_val = float("4357.54") data = None @@ -5343,10 +5271,10 @@ class Program_weight_tensor_parameter_459: original_name = "batch_norm2d_5.w_1" shape = [160] dtype = "float32" - min_val = float("-1002460000000.0") - max_val = float("138202000000.0") - mean = float("-7099590000.0") - std = float("81248300000.0") + min_val = float("-10024.6") + max_val = float("1382.02") + mean = float("-7099.59") + std = float("8124.83") data = None @@ -5391,10 +5319,8 @@ class Program_weight_tensor_parameter_463: original_name = "batch_norm2d_4.w_2" shape = [160] dtype = "float32" - min_val = float("10782800000.0") - max_val = float("8.59511e+21") - mean = float("2.0735e+20") - std = float("inf") + min_val = float("10.7828") + max_val = float("8595.11") data = None @@ -5403,10 +5329,10 @@ class Program_weight_tensor_parameter_464: original_name = "batch_norm2d_4.w_1" shape = [160] dtype = "float32" - min_val = float("-55967000000.0") - max_val = float("24376500000.0") - mean = float("-1380990000.0") - std = float("7539840000.0") + min_val = float("-5596.70") + max_val = float("2437.65") + mean = float("-1380.99") + std = float("7539.84") data = None @@ -5452,9 +5378,7 @@ class Program_weight_tensor_parameter_468: shape = [80] dtype = "float32" min_val = float("0.00285509") - max_val = float("6.28874e+23") - mean = float("8.72561e+21") - std = float("inf") + max_val = float("6288.74") data = None @@ -5463,10 +5387,10 @@ class Program_weight_tensor_parameter_469: original_name = "batch_norm2d_3.w_1" shape = [80] dtype = "float32" - min_val = float("-824200000000.0") - max_val = float("116249000000.0") - mean = float("-10789600000.0") - std = float("93473500000.0") + min_val = float("-8242.00") + max_val = float("11624.9") + mean = float("-1078.96") + std = float("9347.35") data = None @@ -5511,10 +5435,8 @@ class Program_weight_tensor_parameter_473: original_name = "batch_norm2d_2.w_2" shape = [80] dtype = "float32" - min_val = float("29321900000.0") - max_val = float("7.11039e+23") - mean = float("9.04831e+21") - std = float("inf") + min_val = float("29.3219") + max_val = float("7110.39") data = None @@ -5523,10 +5445,10 @@ class Program_weight_tensor_parameter_474: original_name = "batch_norm2d_2.w_1" shape = [80] dtype = "float32" - min_val = float("-925988000000.0") - max_val = float("53176100000.0") - mean = float("-12275300000.0") - std = float("103743000000.0") + min_val = float("-9259.88") + max_val = float("5317.61") + mean = float("-1227.53") + std = float("10374.3") data = None From 76211b8fe12ef904340d8ccb8e149e133d13eac0 Mon Sep 17 00:00:00 2001 From: Liu Yiqun Date: Wed, 21 Jan 2026 17:41:44 +0800 Subject: [PATCH 10/10] Minor fix. --- graph_net/test/get_incorrect_models_test.sh | 15 ++++++++++----- .../PP-ShiTuV2_det/subgraph_1/weight_meta.py | 6 ++---- 2 files changed, 12 insertions(+), 9 deletions(-) diff --git a/graph_net/test/get_incorrect_models_test.sh b/graph_net/test/get_incorrect_models_test.sh index 49d269db9..bbb37b637 100755 --- a/graph_net/test/get_incorrect_models_test.sh +++ b/graph_net/test/get_incorrect_models_test.sh @@ -1,6 +1,5 @@ #!/bin/bash - SCRIPT_DIR=$(cd "$(dirname "${BASH_SOURCE[0]}")" && pwd) GRAPH_NET_DIR=$(dirname "$SCRIPT_DIR") PROJECT_ROOT=$(dirname "$GRAPH_NET_DIR") @@ -8,15 +7,21 @@ PROJECT_ROOT=$(dirname "$GRAPH_NET_DIR") # 将项目根目录加入Python路径 export PYTHONPATH="$PROJECT_ROOT:$PYTHONPATH" -TOLERANCE_LIST=(-5 -4 -3 -2 -1 0 1) LOG_FILE_PATH="log_file_for_test.txt" python3 - <